{"id": 163840, "image": "COCO_train2014_000000163840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with black tank top and multicolored skirt hugging an elephant ' s trunk\"."}], "task": "refering", "answer_template": "The \"a woman with black tank top and multicolored skirt hugging an elephant ' s trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 164, 165, 166, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 193, 201, 202, 203, 214, 215, 216, 226, 227, 228, 229, 239, 240, 241, 252, 253, 254, 265, 267], "bbox": [0.4269080779944289, 0.48367187500000003, 0.8825069637883008, 0.88325], "iscrowd": 0, "area": 14526.273250000004, "rle": {"size": [640, 359], "counts": "UUP32nc08C9C:F9G9J6I7J6J7H7]MaM_Ce2`<^MWCi2h<ZMPCl2o<WMiBm2X=VM_Bm2d=VMTBl2o=WMiAk2Z>XMaAh2a>[MZAh2e>[MVAg2j>\\MQAf2o>]Ml@f2S?S101O010O00010O0100O0101O1O001O1O2N1O2N2N2N1N3N2N1O1O001O1O001M3N1N2O000O]OQBXKo=e4[BUKc=i4gBPKY=n4l0N3M2N3N2M2N3L3M4L3M4L3M4L3M4M3L3M5K5K4L5K5K@bMh_OY2[`0c03N2O01000000O1001O1O1O1N2O0O2YOk^ObNVa0R22O1XOg^OgN[a0W1e^OjN\\a0T1d^OlN_a0Q1b^OoN`a0n0`^OROba0l0^^OUOda0h0]^OWOea0h0Z^OYOha0d0X^O\\Oja0b0W^O^Oka0?U^OBla0<T^OEoa05T^OKQb0KS^O6Rb0@R^Oa0Rb0VOU^Oh0bb0O1O1O1O2O0O1O2N3M3M3M3M2Neh=`0iVB1011N3M2O2M5K5LQli0"}, "label": "a woman with black tank top and multicolored skirt hugging an elephant ' s trunk"}]}
{"id": 163840, "image": "COCO_train2014_000000163840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman hugging the trunk of an elephant\"."}], "task": "refering", "answer_template": "The \"a woman hugging the trunk of an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 164, 165, 166, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 193, 201, 202, 203, 214, 215, 216, 226, 227, 228, 229, 239, 240, 241, 252, 253, 254, 265, 267], "bbox": [0.4269080779944289, 0.48367187500000003, 0.8825069637883008, 0.88325], "iscrowd": 0, "area": 14526.273250000004, "rle": {"size": [640, 359], "counts": "UUP32nc08C9C:F9G9J6I7J6J7H7]MaM_Ce2`<^MWCi2h<ZMPCl2o<WMiBm2X=VM_Bm2d=VMTBl2o=WMiAk2Z>XMaAh2a>[MZAh2e>[MVAg2j>\\MQAf2o>]Ml@f2S?S101O010O00010O0100O0101O1O001O1O2N1O2N2N2N1N3N2N1O1O001O1O001M3N1N2O000O]OQBXKo=e4[BUKc=i4gBPKY=n4l0N3M2N3N2M2N3L3M4L3M4L3M4L3M4M3L3M5K5K4L5K5K@bMh_OY2[`0c03N2O01000000O1001O1O1O1N2O0O2YOk^ObNVa0R22O1XOg^OgN[a0W1e^OjN\\a0T1d^OlN_a0Q1b^OoN`a0n0`^OROba0l0^^OUOda0h0]^OWOea0h0Z^OYOha0d0X^O\\Oja0b0W^O^Oka0?U^OBla0<T^OEoa05T^OKQb0KS^O6Rb0@R^Oa0Rb0VOU^Oh0bb0O1O1O1O2O0O1O2N3M3M3M3M2Neh=`0iVB1011N3M2O2M5K5LQli0"}, "label": "a woman hugging the trunk of an elephant"}]}
{"id": 303111, "image": "COCO_train2014_000000303111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant that has yellow shirted man on it\"."}], "task": "refering", "answer_template": "The \"an elephant that has yellow shirted man on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 271, 288, 289, 292, 311], "bbox": [0.5330781250000001, 0.47085416666666663, 0.802125, 0.7937291666666667], "iscrowd": 0, "area": 11116.328600000003, "rle": {"size": [480, 640], "counts": "_VP51m>3N2M4L3M3M3M3M3M3M3M3M2OO01O00001O00000hN[OWDf0h;MeC3\\<P1`CPN`<X20O0L`CjMa<V2_CjM`<W2`ChMa<X2_CgMa<Z211010O1O001O1O001O1O010O10O10000O10000O100000O010000O10000O10000O11O000000001N100O10000O2O000O100O2O000O10001N100O10000O2O0O10000O20O2N2N3M2O1N2N2N2N2O0O1O0O2O0O2O0O2N101NO1O2N1O101N1O2N1O101N00000001O0000I7H8H8H8O03N2O1000001N100000000O1000000O2O000000l0SO3N3M3O010O00N3K4L5J5H9G8Ha0_OQYk1"}, "label": "an elephant that has yellow shirted man on it"}]}
{"id": 303111, "image": "COCO_train2014_000000303111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with a man riding atop it\"."}], "task": "refering", "answer_template": "The \"an elephant with a man riding atop it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 271, 288, 289, 292, 311], "bbox": [0.5330781250000001, 0.47085416666666663, 0.802125, 0.7937291666666667], "iscrowd": 0, "area": 11116.328600000003, "rle": {"size": [480, 640], "counts": "_VP51m>3N2M4L3M3M3M3M3M3M3M3M2OO01O00001O00000hN[OWDf0h;MeC3\\<P1`CPN`<X20O0L`CjMa<V2_CjM`<W2`ChMa<X2_CgMa<Z211010O1O001O1O001O1O010O10O10000O10000O100000O010000O10000O10000O11O000000001N100O10000O2O000O100O2O000O10001N100O10000O2O0O10000O20O2N2N3M2O1N2N2N2N2O0O1O0O2O0O2O0O2N101NO1O2N1O101N1O2N1O101N00000001O0000I7H8H8H8O03N2O1000001N100000000O1000000O2O000000l0SO3N3M3O010O00N3K4L5J5H9G8Ha0_OQYk1"}, "label": "an elephant with a man riding atop it"}]}
{"id": 368647, "image": "COCO_train2014_000000368647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair farther away from the armoire\"."}], "task": "refering", "answer_template": "The \"the chair farther away from the armoire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 297, 298, 320], "bbox": [0.827609375, 0.5738875878220141, 1.0, 0.8946370023419203], "iscrowd": 0, "area": 8190.455050000002, "rle": {"size": [427, 640], "counts": "jZm62W=4iNOnD4P;T10M2O1I8G8OO3L5L3N3M201N1O1O2N1O2O001O001O001O00001O001O001O001O001O00001O001O001O0O2O001O00001O001N101N1_Ob0N1N3N1O2M2O2O0O2O001O<TFiLl8k3O101O00EkFaLT9[3?M4L4L4L4O0010000000O0100000001N1000001O000O10001O000O101O00000O2O00000@fH"}, "label": "the chair farther away from the armoire"}]}
{"id": 368647, "image": "COCO_train2014_000000368647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the right of the room\"."}], "task": "refering", "answer_template": "The \"the chair on the right of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 297, 298, 320], "bbox": [0.827609375, 0.5738875878220141, 1.0, 0.8946370023419203], "iscrowd": 0, "area": 8190.455050000002, "rle": {"size": [427, 640], "counts": "jZm62W=4iNOnD4P;T10M2O1I8G8OO3L5L3N3M201N1O1O2N1O2O001O001O001O00001O001O001O001O001O00001O001O001O0O2O001O00001O001N101N1_Ob0N1N3N1O2M2O2O0O2O001O<TFiLl8k3O101O00EkFaLT9[3?M4L4L4L4O0010000000O0100000001N1000001O000O10001O000O101O00000O2O00000@fH"}, "label": "the chair on the right of the room"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy wearing blue shirt holding plastic cup\"."}], "task": "refering", "answer_template": "The \"young boy wearing blue shirt holding plastic cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 116, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281], "bbox": [0.02103125, 0.3254566744730679, 0.24493750000000003, 0.8367213114754097], "iscrowd": 0, "area": 22644.54299999999, "rle": {"size": [427, 640], "counts": "dg5`0_<c0^Oa0F5K4N2N2N2N2O1N2N2N2N3M2N2N2O1N2O1N2O1N2O2M2O1]NoLkHR3e6bMVI_2Z6UNaIl1\\6ZN`Ig1\\6^NbIc1[6bNbI_1Z6fNdI[1Z6iNcIX1\\6jNbIW1]6kNbIU1\\6nNcIR1\\6QOaIP1^6ROaIn0^6TO`Im0_6UO`Ik0_6WO_Ij0`6YO^Ig0a6]O[Id0e6@WI`0h6R3O1O100O1O100O1000000O10000O1000000000000000000000000000000O1000000000000001O001O001O00001N101O00001O001O001O001O001O0O2O0VNSIUMo6h2UIVMl6h2XITMj6j2XITMj6j2XITMj6k2VITMl6j2UIVMl6i2TIVMn6h2TIWMm6U1dHRO?In6S1eHRO>JP7P1dHVO<JQ7m0eHXO<JQ7i0gH\\O8KR7f0hH^O7LS7<nHH0KT77QIMKLV72RI2HLX7MSI7FKY7IUI;BL]95bFK^95cFJ^95bFK^95bFK_94bFK^95cFJ^95cFJ^95cFI^97bFH`97aFG`9[OiEb0f02Q:LPFLWO^Ol:e0mEMZ:1gEM`:M`E3n;N1O6IlRY6"}, "label": "young boy wearing blue shirt holding plastic cup"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy wearing a blue shirt an licking something off his lips\"."}], "task": "refering", "answer_template": "The \"a young boy wearing a blue shirt an licking something off his lips\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 116, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281], "bbox": [0.02103125, 0.3254566744730679, 0.24493750000000003, 0.8367213114754097], "iscrowd": 0, "area": 22644.54299999999, "rle": {"size": [427, 640], "counts": "dg5`0_<c0^Oa0F5K4N2N2N2N2O1N2N2N2N3M2N2N2O1N2O1N2O1N2O2M2O1]NoLkHR3e6bMVI_2Z6UNaIl1\\6ZN`Ig1\\6^NbIc1[6bNbI_1Z6fNdI[1Z6iNcIX1\\6jNbIW1]6kNbIU1\\6nNcIR1\\6QOaIP1^6ROaIn0^6TO`Im0_6UO`Ik0_6WO_Ij0`6YO^Ig0a6]O[Id0e6@WI`0h6R3O1O100O1O100O1000000O10000O1000000000000000000000000000000O1000000000000001O001O001O00001N101O00001O001O001O001O001O0O2O0VNSIUMo6h2UIVMl6h2XITMj6j2XITMj6j2XITMj6k2VITMl6j2UIVMl6i2TIVMn6h2TIWMm6U1dHRO?In6S1eHRO>JP7P1dHVO<JQ7m0eHXO<JQ7i0gH\\O8KR7f0hH^O7LS7<nHH0KT77QIMKLV72RI2HLX7MSI7FKY7IUI;BL]95bFK^95cFJ^95bFK^95bFK_94bFK^95cFJ^95cFJ^95cFI^97bFH`97aFG`9[OiEb0f02Q:LPFLWO^Ol:e0mEMZ:1gEM`:M`E3n;N1O6IlRY6"}, "label": "a young boy wearing a blue shirt an licking something off his lips"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with two white cotton puffs in her hands\"."}], "task": "refering", "answer_template": "The \"a girl with two white cotton puffs in her hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251], "bbox": [0.52925, 0.14381733021077284, 0.9295625000000001, 0.7348243559718969], "iscrowd": 0, "area": 36924.85290000001, "rle": {"size": [427, 640], "counts": "ga]41Y=2N2N1O2N2N2M3N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N101cDgNe:[1VEjNi:W1REnNm:S1oDQOP;o0lDUOT;b1O1O1O1O001O1N2O001O1O1O1O001O1O1O0O2O1O1O00001O000011N2N1O2O1N2N2N2O0O2N2N2O1N1O2N2O1N2N2O0O2N2N2O1N2N1O10N1N2N2O2M2O1N2O1N1O010O10O01O100L3L5L4K4M4K5L4L3XGoLT7V3gHPMU7S3gHRMT7R3hHSMS7P3iHVMS7m2iHXMR7l2iHZMS7h2jH]MQ7g2kH^MP7f2kH`MQ7c2kHbMP7a2mHdMn6_4L4K5L4K4L5L31O001O001O10O01O001O1O001O1O1O1N2O1O1N2O1O1N2O1O1O12N2N3M2N2N3M2N2N3M3M2N3M3M1O1O001O0O1000001O01O00000001O000000001O01O000001O000000001O0001O0001O0000000010O2M2K6J6J6M3000O0100O100O1N2N101N2N2O1N3N3L3M4M2M3N3L3M3N3L3N3L3L4M4L3M4L3M3L5L3M3M5K:F:E;F<Da0_O>B8G9H8H9\\Om_b0"}, "label": "a girl with two white cotton puffs in her hands"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl standing at a table\"."}], "task": "refering", "answer_template": "The \"a little girl standing at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251], "bbox": [0.52925, 0.14381733021077284, 0.9295625000000001, 0.7348243559718969], "iscrowd": 0, "area": 36924.85290000001, "rle": {"size": [427, 640], "counts": "ga]41Y=2N2N1O2N2N2M3N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N101cDgNe:[1VEjNi:W1REnNm:S1oDQOP;o0lDUOT;b1O1O1O1O001O1N2O001O1O1O1O001O1O1O0O2O1O1O00001O000011N2N1O2O1N2N2N2O0O2N2N2O1N1O2N2O1N2N2O0O2N2N2O1N2N1O10N1N2N2O2M2O1N2O1N1O010O10O01O100L3L5L4K4M4K5L4L3XGoLT7V3gHPMU7S3gHRMT7R3hHSMS7P3iHVMS7m2iHXMR7l2iHZMS7h2jH]MQ7g2kH^MP7f2kH`MQ7c2kHbMP7a2mHdMn6_4L4K5L4K4L5L31O001O001O10O01O001O1O001O1O1O1N2O1O1N2O1O1N2O1O1O12N2N3M2N2N3M2N2N3M3M2N3M3M1O1O001O0O1000001O01O00000001O000000001O01O000001O000000001O0001O0001O0000000010O2M2K6J6J6M3000O0100O100O1N2N101N2N2O1N3N3L3M4M2M3N3L3M3N3L3N3L3L4M4L3M4L3M3L5L3M3M5K:F:E;F<Da0_O>B8G9H8H9\\Om_b0"}, "label": "a little girl standing at a table"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chair in which boy is sitting with black dress\"."}], "task": "refering", "answer_template": "The \"wooden chair in which boy is sitting with black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 165, 166, 184, 185, 188, 189, 207, 230, 253], "bbox": [0.0, 0.46056206088992974, 0.23590624999999998, 0.7653629976580796], "iscrowd": 0, "area": 3301.77145, "rle": {"size": [427, 640], "counts": "V8V1c:e1M4L3M4LQNmEi0P:WOTFh0i9WO[Fj0a9VOcFi0Z9WOjFc0X9]OkF>W9BmF8U9HnF3T9MPGMR93RGHo87UGDm8<WG]Ol8c0WGXOk8h0YGQOj8o0i10000O100000000000000O100000000000000O10001O2N2N1O2M2O2N2N1O2N1O1O2M2O2N2N2N1010gmk0LZRTO1O1O1O1O1O1F:M3N4L:F2N3F900O2O00001O010=SOkCHc<IdC1Q=K[e[6"}, "label": "wooden chair in which boy is sitting with black dress"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden chair that the boy in black is sitting in\"."}], "task": "refering", "answer_template": "The \"the wooden chair that the boy in black is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 165, 166, 184, 185, 188, 189, 207, 230, 253], "bbox": [0.0, 0.46056206088992974, 0.23590624999999998, 0.7653629976580796], "iscrowd": 0, "area": 3301.77145, "rle": {"size": [427, 640], "counts": "V8V1c:e1M4L3M4LQNmEi0P:WOTFh0i9WO[Fj0a9VOcFi0Z9WOjFc0X9]OkF>W9BmF8U9HnF3T9MPGMR93RGHo87UGDm8<WG]Ol8c0WGXOk8h0YGQOj8o0i10000O100000000000000O100000000000000O10001O2N2N1O2M2O2N2N1O2N1O1O2M2O2N2N2N1010gmk0LZRTO1O1O1O1O1O1F:M3N4L:F2N3F900O2O00001O010=SOkCHc<IdC1Q=K[e[6"}, "label": "the wooden chair that the boy in black is sitting in"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown - haired boy in a red shirt\"."}], "task": "refering", "answer_template": "The \"a brown - haired boy in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263], "bbox": [0.235921875, 0.25562060889929744, 0.47484374999999995, 0.7781498829039812], "iscrowd": 0, "area": 21539.060599999993, "rle": {"size": [427, 640], "counts": "eVo17R=2TCN]<<mC_Od;k08:F:F5K3M4L3N3M2O2M5L5J5L2`MYMWJi2a5dMYJ^2^5oM[JT2^5WN\\Jj1c5\\NVJg1g5_NSJd1j5_NTJb1k5_NSJd1j5_NTJb1j5aNSJb1k5_NTJa1k5bNRJ_1n5cNPJ^1n5dNPJ]1o5fNoIZ1Q6hNlIY1S6jNkIW1S6kNkIV1U6mNhIS1W6QOeIP1[6SO_IP1`6n2O100O1O100O10000O100O10000O100O10000O11O00001O001N10001O0O2O00001O0O2O00001N101O000O2O001O0O1O2O0O2O1N2O2M2bLlHb0W7YOlHg0U7VOmHi0V7SOlHm0U7oNnHP1T7mNnHR1U7jNmHT1W7fNnHW1U7cNPI\\1S7^NQIb1P7ZNSIf1o6WNSIh1o6UNRIj1P7TNRIk1P7RNQIn1P7PNRIo1P7nMRIQ2P7lMQIU2o6iMSIV2Q7eMPI\\2S7_MoH`2V7ZMlHe2m810O010O100O1O100O001O100O1O2N1O1O2O0O1O2N1O100O1mNoDIS;5oDIT;4oDIS;5PEHR;6QEGQ;7REFP;8SEEo:9TEDo:8TEFn:8UEEn:8UEEm:8V1M2N3LUk[4"}, "label": "a brown - haired boy in a red shirt"}]}
{"id": 147466, "image": "COCO_train2014_000000147466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with brown hair and red shirt with gray sleeves\"."}], "task": "refering", "answer_template": "The \"a boy with brown hair and red shirt with gray sleeves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263], "bbox": [0.235921875, 0.25562060889929744, 0.47484374999999995, 0.7781498829039812], "iscrowd": 0, "area": 21539.060599999993, "rle": {"size": [427, 640], "counts": "eVo17R=2TCN]<<mC_Od;k08:F:F5K3M4L3N3M2O2M5L5J5L2`MYMWJi2a5dMYJ^2^5oM[JT2^5WN\\Jj1c5\\NVJg1g5_NSJd1j5_NTJb1k5_NSJd1j5_NTJb1j5aNSJb1k5_NTJa1k5bNRJ_1n5cNPJ^1n5dNPJ]1o5fNoIZ1Q6hNlIY1S6jNkIW1S6kNkIV1U6mNhIS1W6QOeIP1[6SO_IP1`6n2O100O1O100O10000O100O10000O100O10000O11O00001O001N10001O0O2O00001O0O2O00001N101O000O2O001O0O1O2O0O2O1N2O2M2bLlHb0W7YOlHg0U7VOmHi0V7SOlHm0U7oNnHP1T7mNnHR1U7jNmHT1W7fNnHW1U7cNPI\\1S7^NQIb1P7ZNSIf1o6WNSIh1o6UNRIj1P7TNRIk1P7RNQIn1P7PNRIo1P7nMRIQ2P7lMQIU2o6iMSIV2Q7eMPI\\2S7_MoH`2V7ZMlHe2m810O010O100O1O100O001O100O1O2N1O1O2O0O1O2N1O100O1mNoDIS;5oDIT;4oDIS;5PEHR;6QEGQ;7REFP;8SEEo:9TEDo:8TEFn:8UEEn:8UEEm:8V1M2N3LUk[4"}, "label": "a boy with brown hair and red shirt with gray sleeves"}]}
{"id": 458762, "image": "COCO_train2014_000000458762.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white cat with it ' s right paw in the air\"."}], "task": "refering", "answer_template": "The \"black and white cat with it ' s right paw in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 5, 6, 7, 8, 10, 11, 12, 23, 24, 25, 27, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 96, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.01203125, 0.013114754098360656, 0.6453125, 0.9988056206088993], "iscrowd": 0, "area": 114558.36745000002, "rle": {"size": [427, 640], "counts": "d\\34U=5K6J6H8H8I7J7I6J6K6I7I7J6J6J6J6K5J6J6J6K6I6J7J5J6J5L5K4M3L5K4L4L5K4M3L5K4M3M3L3N3M2M3N3M2M2O2N1O101N101N101N101N1O1010O100O2O1N3N2M3M3M3M3M3M4L:F9F9H001O1N2N2N2O1N2N2O1N2N2N01O01O01O0001O01O01O01O00100O2N2O0O2N1O2O1N1iK[Je0g5jNkJS1W5hNnJW1Q6fMUJW2j5hM[JU2d5iMaJV2]5hMhJW2V5hMmJX2Q5gMSKX2k4fMYKZ2e4eM^K[2`4cMdK]2Z4bMjK]2T4bMoK_2n3_MVLa2h3^M[Lb2c3\\MlMY1R2fNTNW1j1hN[NV1c1hNbNW1\\1iNeNX1Y1hNiNW1V1hNmNV1S1jNnNU1R1kNPOS1P1lNROR1o0nNSOP1m0oNUOP1k0POVOo0j0QOXOl0i0SOYOk0h0UOYOg0j0XOXOe0j0[OWOa0l0_OUO=n0CSO:o0GQO7P1IRO4o0LRO2o0NRO0o00RONo03QOKP15QOIP17ROFo0:RODo0<RO@Q1a0oN[OT1e0mNWOV1i0f5001O000000001OO100O100O10000O100O100O10000O100O100O10000O100000000000000001O0000000000000000001O0000000000000000001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O001O001O1O001O00O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O100O1O100O1O1O100O1OgFEd5;ZJM`52^J5^5JkIJhM`0[8FjI1eM=_8AkI8`M;c8\\OkI`1S6`NkId1S6[NlIi1R6VNlIo1R6QNlIS2R6mMkIW2T6iMjI[2T6eMjI_2T6aMiId2U6\\MiIh2U6YMhIk2V6VMfIP3W6RMeIS3X6nLdIX3Y6iLdI[3Z6hLaI]3\\6eL_Ia3]6dL]Ia3`6iLSI]3j6nLhHY3T7c1L5K4L4L4L1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M3M3M3M3M4L4L4L4L4L4L3M4L4L4L4L4L2N2N3M2N2N2N3M2N2N3M2N2N2N4L8H7I8H8H7I8H7I8H9G9G6J6JjWn2"}, "label": "black and white cat with it ' s right paw in the air"}]}
{"id": 458762, "image": "COCO_train2014_000000458762.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat with its paw in the air\"."}], "task": "refering", "answer_template": "The \"a black and white cat with its paw in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 5, 6, 7, 8, 10, 11, 12, 23, 24, 25, 27, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 96, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.01203125, 0.013114754098360656, 0.6453125, 0.9988056206088993], "iscrowd": 0, "area": 114558.36745000002, "rle": {"size": [427, 640], "counts": "d\\34U=5K6J6H8H8I7J7I6J6K6I7I7J6J6J6J6K5J6J6J6K6I6J7J5J6J5L5K4M3L5K4L4L5K4M3L5K4M3M3L3N3M2M3N3M2M2O2N1O101N101N101N101N1O1010O100O2O1N3N2M3M3M3M3M3M4L:F9F9H001O1N2N2N2O1N2N2O1N2N2N01O01O01O0001O01O01O01O00100O2N2O0O2N1O2O1N1iK[Je0g5jNkJS1W5hNnJW1Q6fMUJW2j5hM[JU2d5iMaJV2]5hMhJW2V5hMmJX2Q5gMSKX2k4fMYKZ2e4eM^K[2`4cMdK]2Z4bMjK]2T4bMoK_2n3_MVLa2h3^M[Lb2c3\\MlMY1R2fNTNW1j1hN[NV1c1hNbNW1\\1iNeNX1Y1hNiNW1V1hNmNV1S1jNnNU1R1kNPOS1P1lNROR1o0nNSOP1m0oNUOP1k0POVOo0j0QOXOl0i0SOYOk0h0UOYOg0j0XOXOe0j0[OWOa0l0_OUO=n0CSO:o0GQO7P1IRO4o0LRO2o0NRO0o00RONo03QOKP15QOIP17ROFo0:RODo0<RO@Q1a0oN[OT1e0mNWOV1i0f5001O000000001OO100O100O10000O100O100O10000O100O100O10000O100000000000000001O0000000000000000001O0000000000000000001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O001O001O1O001O00O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O100O1O100O1O1O100O1OgFEd5;ZJM`52^J5^5JkIJhM`0[8FjI1eM=_8AkI8`M;c8\\OkI`1S6`NkId1S6[NlIi1R6VNlIo1R6QNlIS2R6mMkIW2T6iMjI[2T6eMjI_2T6aMiId2U6\\MiIh2U6YMhIk2V6VMfIP3W6RMeIS3X6nLdIX3Y6iLdI[3Z6hLaI]3\\6eL_Ia3]6dL]Ia3`6iLSI]3j6nLhHY3T7c1L5K4L4L4L1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M3M3M3M3M4L4L4L4L4L4L3M4L4L4L4L4L2N2N3M2N2N2N3M2N2N3M2N2N2N4L8H7I8H8H7I8H7I8H9G9G6J6JjWn2"}, "label": "a black and white cat with its paw in the air"}]}
{"id": 458762, "image": "COCO_train2014_000000458762.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tabby cat with green eyes\"."}], "task": "refering", "answer_template": "The \"a tabby cat with green eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.58825, 0.4100468384074942, 0.93478125, 1.0], "iscrowd": 0, "area": 36354.181950000006, "rle": {"size": [427, 640], "counts": "SPm41V=a0@<E;F9G6J7H7J7H8I5J5K6I7J5J7F:I7J7I8J7J6J5K3M3M2N3M3M2N2N3N1N9G;D2O001O0O10000000O0100O010O010O0010O01O01O01O010O0010O01O010O0001O00000000000001O000000000000000000000000000000000O1O1O100O1O1O1O100O1O1O1O100O1O100O1O1O100O100O100O100O100O100O10VK\\Jm1c5mMfJQ2Z5gMPKW2o4cMYK\\2g4]M`Kc2_4WMhKi2X4PMoKP3Q4iLVLW3k3cL[L\\3g3^L]Lb3e3WL`Li3d3nKaLR4b3fKcLZ4W600000001O0000000000001O0000000000001O000O1000001O00000O2O000000lFkKS9W4O000001O01O01O0000100O1O1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O4L3M4L4L3M4L4L4L4L4L4L3M4L4K5L3M3M3M2N3M3M4J:]OXTa0"}, "label": "a tabby cat with green eyes"}]}
{"id": 458762, "image": "COCO_train2014_000000458762.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and black cat looking upwards\"."}], "task": "refering", "answer_template": "The \"a brown and black cat looking upwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.58825, 0.4100468384074942, 0.93478125, 1.0], "iscrowd": 0, "area": 36354.181950000006, "rle": {"size": [427, 640], "counts": "SPm41V=a0@<E;F9G6J7H7J7H8I5J5K6I7J5J7F:I7J7I8J7J6J5K3M3M2N3M3M2N2N3N1N9G;D2O001O0O10000000O0100O010O010O0010O01O01O01O010O0010O01O010O0001O00000000000001O000000000000000000000000000000000O1O1O100O1O1O1O100O1O1O1O100O1O100O1O1O100O100O100O100O100O100O10VK\\Jm1c5mMfJQ2Z5gMPKW2o4cMYK\\2g4]M`Kc2_4WMhKi2X4PMoKP3Q4iLVLW3k3cL[L\\3g3^L]Lb3e3WL`Li3d3nKaLR4b3fKcLZ4W600000001O0000000000001O0000000000001O000O1000001O00000O2O000000lFkKS9W4O000001O01O01O0000100O1O1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O4L3M4L4L3M4L4L4L4L4L4L3M4L4K5L3M3M3M2N3M3M4J:]OXTa0"}, "label": "a brown and black cat looking upwards"}]}
{"id": 213003, "image": "COCO_train2014_000000213003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"keyboard on the laptop\"."}], "task": "refering", "answer_template": "The \"keyboard on the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 308, 309, 310, 311, 312, 313, 329, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 380, 381], "bbox": [0.300046875, 0.7692738589211618, 0.62575, 0.9520746887966806], "iscrowd": 0, "area": 10848.761649999997, "rle": {"size": [482, 640], "counts": "\\Xk21Q?0O1O1O100O1O10O01O100O1O100O1O1jNFcC;]<EcC;\\<GbC9^<HaC9_<GaC9^<HaC9^<HaC9_<H`C8_<I`C8_<I`C8`<H_C9a<G_C9a<H]C8d<H[C9e<G[C9e<GZC:f<FZC:g<FXC:h<FXC:h<FXC:h<GWC9i<HVC8j<HVC8j<IUC6l<JTCb0`<_O_Ca0b<^O^Cc0a<^O^Cb0b<_O]Ca0c<_O]Ca0c<A[C?e<CYC=g<EWC<h<FVC:j<HTC8m<IQC7o<KoB5Q=o0O000000000010O00000001O00000001O01O000000001O0001O000001O00000000010O000000001O000001O0001O000000001O0001O000001O00000000010O000000001O000001O0001O000000001O01O00000001O000000010O00000000001O0001O000001O0000000010O0000000001O000001O01O0000O1O1N2O1N3N1O1N2O1N2O1N2O1M4M2N2M3N2M3N2N2M3N3L3N2N2M3N2M3N2M3N2Nlb`3"}, "label": "keyboard on the laptop"}]}
{"id": 213003, "image": "COCO_train2014_000000213003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard on the laptop in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the keyboard on the laptop in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 308, 309, 310, 311, 312, 313, 329, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 380, 381], "bbox": [0.300046875, 0.7692738589211618, 0.62575, 0.9520746887966806], "iscrowd": 0, "area": 10848.761649999997, "rle": {"size": [482, 640], "counts": "\\Xk21Q?0O1O1O100O1O10O01O100O1O100O1O1jNFcC;]<EcC;\\<GbC9^<HaC9_<GaC9^<HaC9^<HaC9_<H`C8_<I`C8_<I`C8`<H_C9a<G_C9a<H]C8d<H[C9e<G[C9e<GZC:f<FZC:g<FXC:h<FXC:h<FXC:h<GWC9i<HVC8j<HVC8j<IUC6l<JTCb0`<_O_Ca0b<^O^Cc0a<^O^Cb0b<_O]Ca0c<_O]Ca0c<A[C?e<CYC=g<EWC<h<FVC:j<HTC8m<IQC7o<KoB5Q=o0O000000000010O00000001O00000001O01O000000001O0001O000001O00000000010O000000001O000001O0001O000000001O0001O000001O00000000010O000000001O000001O0001O000000001O01O00000001O000000010O00000000001O0001O000001O0000000010O0000000001O000001O01O0000O1O1N2O1N3N1O1N2O1N2O1N2O1M4M2N2M3N2M3N2N2M3N3L3N2N2M3N2M3N2M3N2Nlb`3"}, "label": "the keyboard on the laptop in the right hand picture"}]}
{"id": 327694, "image": "COCO_train2014_000000327694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of the man standing behind a child reaching around with a toy\"."}], "task": "refering", "answer_template": "The \"the arm of the man standing behind a child reaching around with a toy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 16, 17, 18, 23, 24, 25, 26, 27, 28, 29, 30, 31, 39, 40, 41, 46, 47, 48, 49, 50, 51, 52, 53, 54, 62, 63, 64, 69, 70, 71, 72, 73, 74, 75, 76, 85, 86, 92, 93, 94, 95, 96, 97, 98, 99, 107, 108, 115, 116, 117, 118, 119, 120, 121, 122, 123, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 339, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0033750000000000004, 0.0045000000000000005, 0.859546875, 0.9842708333333334], "iscrowd": 0, "area": 138918.1024, "rle": {"size": [480, 640], "counts": "Sn0e>;000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000A?ZOf0oNQ1gNY1J600000000001O000000000000000000000000000000001O0000000000000000eKjKlMV4T2jKlMV4S2lKlMT4T2mKkMS4U2mKkMS4T2oKkMQ4U2PLjMP4V2PLjMP4U2RLjMn3V2RLkMm3U2TLjMl3U2VLjMj3V2VLjMj3V2WLiMi3W2WLiMi3V2YLiMg3W2ZLhMf3X2ZLhMf3W2\\LhMd3X2\\LhMd3X2\\LhMd3W2^LhMb3X2^LhMb3X2^LhMb3X2^LhMb3W2_LjM`3V2aLiM_3W2aLiM_3V2bLjM^3V2bLjM^3V2bLjM^3U2cLkM]3U2dLjM\\3V2dLjM\\3U2eLlMZ3T2fLlMZ3T2fLlMZ3T2gLkMY3T2hLlMX3T2hLlMX3T2hLlMX3S2iLmMW3S2iLmMW3S2jLmMU3R2lLnMT3R2lLnMT3Q2mLoMS3Q2mLoMS3Q2nLoMQ3P2PMPNP3P2PMPNP3P2PMPNP3o1QMRNn2n1RMRNn2n1SMQNm2n1TMRNl2n1TMRNl2n1TMSNk2l1VMTNj2l1WMSNi2m1WMSNi2l1XMUNg2k1YMUNg2k1YMUNg2j1ZMVNf2j1ZMVNf2i1[MXNd2h1\\MXNd2g1]MYNc2f1^MZNb2e1_M\\N`2kNXLGY1^1_2jN\\LEU1a1_2hNbLBP1f1^2fNgL@l0k1\\2dNmL\\Oh0P2[2bNRM[Oc0S2[2`NWMYO?W2Z2_N\\MUO;\\2Y2]NaMSO7a2W2ZNhMPO2f2V2XNmMoNMi2V2VNSNlNHn2U2SNYNkNCS3S2PN^NjN@V3R2nMbNjN\\OX3R2kMgNjNXO[3Q2iMkNiNTO_3Q2fMoNhNQOc3o1bMUOhNmNf3n1`MYOhNiNh3n1^M]OgNfNk3m1[MBgNbNo3k1XMGfN_NR4j1WMJeN[NU4k1UMk1k2UNSMm1m2SNRMn1o2QNPMP2P3PNnLR2R3nMmLS2S3mMkLU2V3jMiLV2X3jMgLW2Y3iMeLY2[3gMdLZ2\\3fMcL[2^3dM`L^2`3bM_L_2a3aM^L_2c3aM[La2f3^MYLc2g3]MWLe2i3[MVLf2j3ZMULg2k3YMSLi2n3VMQLk2o3UMPLl2P4TMnKn2R4RMmKo2T4QMjKP3V4QMgKQ3Y4PMeKQ3[4PMcKQ3^4oL`KR3`4nL`KR3`4g30000001O0000000O101O0000000O10001N1000000O101O000O10000O101O000O1000001N10000O10000O10000O2O000O10000O10000O2O000O10000O10000O101O0O10000O10000O10000000000000000O10000000000000000000000000000O10000000000000000000000O10000000000OoKdF_2]9]MhFb2X9]MjFa2W9^MlF`2T9`MmF^2S9cMoF[2Q9dMQGZ2P9fMQGY2o8fMTGX2l8hMUGV2l8jMTGV2k8jMWGT2j8lMVGT2j8lMVGS2k8lMWGS2i8mMWGR2i8oMWGQ2i8nMXGQ2i8oMXGP2h8PNXGo1i8PNXGP2h8PNYGo1f8QN[Gn1f8RNZGn1e8SN[Gl1f8SN\\Gl1c8UN]Gj1d8VN\\Gj1c8VN_Gi1`8XN`Gg1a8YN_Gg1`8YNaGf1`8ZNaGe1^8\\NbGd1]8\\NdGc1]8]NcGc1i7XM_GV1i0`1g7^M^GR1k0`1e7bM]Gn0o0`1a7gM^Gi0R1^1_7lM]Gg0T1]1\\7oM_Gd0V1\\1Z7RN`Gb0V1\\1W7UNbG`0X1Y1T7ZNdG=X1Y1R7\\NeG;Z1Y1o6^NgG9[1W1l6cNiG6[1W1j6dNkG6[1U1i6gNlG3\\1V1f6iNmG2^1T1b6mNPHO^1S1a6POQHL_1T1^6`0cI^O\\6d0dI\\OZ6WOVHH`1i1`5OaJ1]51dJNZ54fJLX55iJKU57kJIS59mJGQ5;oJEo4<RKDl4>TKBj4?WKAg4`0ZK@d4b0\\K^Ob4c0_K]O_4d0bK\\O\\4f0cK[O[4f0fKZOX4g0iKYOU4i0kKWOS4ATJ`Nj1o1P4[O_JdNb1Q2m3VOiJgN[1S2j3POTKlNR1T2h3jN_KPOj0V2e3dNjKTOb0X2\\3dN[LSO9Y2R3kNiLjN6[2g2TOUM_N5]2_2ZO^MXN3_2X2^OgMnM5d2o1BnMcM9k2d1FUNWM>S3X1JQ17jNMW14hNKY16eNJ\\17cNH^18bNG_1:_NFb1;]NDd1<\\NDd1=ZNCf1?YNAg1?YN@h1a0VN_Oj1c0UN]Ok1f0RNYOo1o0hMROW2TOnJFj2U1Y2UOmJGi2T1Y2UOPKFf2U1[2TOPKHd2S1]2TOSKF`2V1\\2UOXKA[2Z1^2TO[K_OW2]1^2SO`K\\OQ2a1_2TOdKWOm1e1_2SOgKVOj1f1_2TOhKVOh1f1a2TOhKUOg1g1a2SOjKUOe1g1a2TOlKTOb1f1e2UOkKTO`1d1h2XOhKTO`1`1k2\\OgKRO^1`1n2]OfKQO]1_1P3_OeKPO\\1^1Q3CcKnN\\1Y1X3H^KmN[1S1^30YKkNZ1m0e37SKjNX1h0m3>kJiNY1a0S4f0fJgNX1;Z4m0`JfNV16b4S1ZJeNU10h4[1TJdNU1Io4c1mIaNU1EV5i1gI`NW9`1kF^NV9a1kF^NU9c1PGWNQ9h1T2O1O2O0O1O1O1O2O0O1O1O101M2O1N2O1N3N1N2N2O1N3N1N2O1N2M4L3L4L4MimY1"}, "label": "the arm of the man standing behind a child reaching around with a toy"}]}
{"id": 327694, "image": "COCO_train2014_000000327694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a father helping his son brush his teeth with a tigger toothbrush\"."}], "task": "refering", "answer_template": "The \"a father helping his son brush his teeth with a tigger toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 16, 17, 18, 23, 24, 25, 26, 27, 28, 29, 30, 31, 39, 40, 41, 46, 47, 48, 49, 50, 51, 52, 53, 54, 62, 63, 64, 69, 70, 71, 72, 73, 74, 75, 76, 85, 86, 92, 93, 94, 95, 96, 97, 98, 99, 107, 108, 115, 116, 117, 118, 119, 120, 121, 122, 123, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 339, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0033750000000000004, 0.0045000000000000005, 0.859546875, 0.9842708333333334], "iscrowd": 0, "area": 138918.1024, "rle": {"size": [480, 640], "counts": "Sn0e>;000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000A?ZOf0oNQ1gNY1J600000000001O000000000000000000000000000000001O0000000000000000eKjKlMV4T2jKlMV4S2lKlMT4T2mKkMS4U2mKkMS4T2oKkMQ4U2PLjMP4V2PLjMP4U2RLjMn3V2RLkMm3U2TLjMl3U2VLjMj3V2VLjMj3V2WLiMi3W2WLiMi3V2YLiMg3W2ZLhMf3X2ZLhMf3W2\\LhMd3X2\\LhMd3X2\\LhMd3W2^LhMb3X2^LhMb3X2^LhMb3X2^LhMb3W2_LjM`3V2aLiM_3W2aLiM_3V2bLjM^3V2bLjM^3V2bLjM^3U2cLkM]3U2dLjM\\3V2dLjM\\3U2eLlMZ3T2fLlMZ3T2fLlMZ3T2gLkMY3T2hLlMX3T2hLlMX3T2hLlMX3S2iLmMW3S2iLmMW3S2jLmMU3R2lLnMT3R2lLnMT3Q2mLoMS3Q2mLoMS3Q2nLoMQ3P2PMPNP3P2PMPNP3P2PMPNP3o1QMRNn2n1RMRNn2n1SMQNm2n1TMRNl2n1TMRNl2n1TMSNk2l1VMTNj2l1WMSNi2m1WMSNi2l1XMUNg2k1YMUNg2k1YMUNg2j1ZMVNf2j1ZMVNf2i1[MXNd2h1\\MXNd2g1]MYNc2f1^MZNb2e1_M\\N`2kNXLGY1^1_2jN\\LEU1a1_2hNbLBP1f1^2fNgL@l0k1\\2dNmL\\Oh0P2[2bNRM[Oc0S2[2`NWMYO?W2Z2_N\\MUO;\\2Y2]NaMSO7a2W2ZNhMPO2f2V2XNmMoNMi2V2VNSNlNHn2U2SNYNkNCS3S2PN^NjN@V3R2nMbNjN\\OX3R2kMgNjNXO[3Q2iMkNiNTO_3Q2fMoNhNQOc3o1bMUOhNmNf3n1`MYOhNiNh3n1^M]OgNfNk3m1[MBgNbNo3k1XMGfN_NR4j1WMJeN[NU4k1UMk1k2UNSMm1m2SNRMn1o2QNPMP2P3PNnLR2R3nMmLS2S3mMkLU2V3jMiLV2X3jMgLW2Y3iMeLY2[3gMdLZ2\\3fMcL[2^3dM`L^2`3bM_L_2a3aM^L_2c3aM[La2f3^MYLc2g3]MWLe2i3[MVLf2j3ZMULg2k3YMSLi2n3VMQLk2o3UMPLl2P4TMnKn2R4RMmKo2T4QMjKP3V4QMgKQ3Y4PMeKQ3[4PMcKQ3^4oL`KR3`4nL`KR3`4g30000001O0000000O101O0000000O10001N1000000O101O000O10000O101O000O1000001N10000O10000O10000O2O000O10000O10000O2O000O10000O10000O101O0O10000O10000O10000000000000000O10000000000000000000000000000O10000000000000000000000O10000000000OoKdF_2]9]MhFb2X9]MjFa2W9^MlF`2T9`MmF^2S9cMoF[2Q9dMQGZ2P9fMQGY2o8fMTGX2l8hMUGV2l8jMTGV2k8jMWGT2j8lMVGT2j8lMVGS2k8lMWGS2i8mMWGR2i8oMWGQ2i8nMXGQ2i8oMXGP2h8PNXGo1i8PNXGP2h8PNYGo1f8QN[Gn1f8RNZGn1e8SN[Gl1f8SN\\Gl1c8UN]Gj1d8VN\\Gj1c8VN_Gi1`8XN`Gg1a8YN_Gg1`8YNaGf1`8ZNaGe1^8\\NbGd1]8\\NdGc1]8]NcGc1i7XM_GV1i0`1g7^M^GR1k0`1e7bM]Gn0o0`1a7gM^Gi0R1^1_7lM]Gg0T1]1\\7oM_Gd0V1\\1Z7RN`Gb0V1\\1W7UNbG`0X1Y1T7ZNdG=X1Y1R7\\NeG;Z1Y1o6^NgG9[1W1l6cNiG6[1W1j6dNkG6[1U1i6gNlG3\\1V1f6iNmG2^1T1b6mNPHO^1S1a6POQHL_1T1^6`0cI^O\\6d0dI\\OZ6WOVHH`1i1`5OaJ1]51dJNZ54fJLX55iJKU57kJIS59mJGQ5;oJEo4<RKDl4>TKBj4?WKAg4`0ZK@d4b0\\K^Ob4c0_K]O_4d0bK\\O\\4f0cK[O[4f0fKZOX4g0iKYOU4i0kKWOS4ATJ`Nj1o1P4[O_JdNb1Q2m3VOiJgN[1S2j3POTKlNR1T2h3jN_KPOj0V2e3dNjKTOb0X2\\3dN[LSO9Y2R3kNiLjN6[2g2TOUM_N5]2_2ZO^MXN3_2X2^OgMnM5d2o1BnMcM9k2d1FUNWM>S3X1JQ17jNMW14hNKY16eNJ\\17cNH^18bNG_1:_NFb1;]NDd1<\\NDd1=ZNCf1?YNAg1?YN@h1a0VN_Oj1c0UN]Ok1f0RNYOo1o0hMROW2TOnJFj2U1Y2UOmJGi2T1Y2UOPKFf2U1[2TOPKHd2S1]2TOSKF`2V1\\2UOXKA[2Z1^2TO[K_OW2]1^2SO`K\\OQ2a1_2TOdKWOm1e1_2SOgKVOj1f1_2TOhKVOh1f1a2TOhKUOg1g1a2SOjKUOe1g1a2TOlKTOb1f1e2UOkKTO`1d1h2XOhKTO`1`1k2\\OgKRO^1`1n2]OfKQO]1_1P3_OeKPO\\1^1Q3CcKnN\\1Y1X3H^KmN[1S1^30YKkNZ1m0e37SKjNX1h0m3>kJiNY1a0S4f0fJgNX1;Z4m0`JfNV16b4S1ZJeNU10h4[1TJdNU1Io4c1mIaNU1EV5i1gI`NW9`1kF^NV9a1kF^NU9c1PGWNQ9h1T2O1O2O0O1O1O1O2O0O1O1O101M2O1N2O1N3N1N2N2O1N3N1N2O1N2M4L3L4L4MimY1"}, "label": "a father helping his son brush his teeth with a tigger toothbrush"}]}
{"id": 327694, "image": "COCO_train2014_000000327694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy getting his teeth brushed with the help of the father\"."}], "task": "refering", "answer_template": "The \"boy getting his teeth brushed with the help of the father\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 54, 55, 56, 57, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 107, 108, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 279, 280, 281, 282, 283, 284, 290, 291, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.165171875, 0.0022500000000000003, 0.9825781250000001, 1.0], "iscrowd": 0, "area": 95425.66700000002, "rle": {"size": [480, 640], "counts": "nTb12k>3eM[2ZNf100001O00000000000000001O00000000000000001O00000000000000001O0000000000oKPLRMP4n2PLRMP4m2QLSMo3m2QLSMo3m2QLTMn3k2TLTMl3l2TLTMl3k2ULUMk3k2ULUMk3k2VLTMj3k2WLUMi3k2WLUMi3j2XLVMh3j2XLWMg3i2ZLVMf3i2[LWMe3i2[LWMe3h2\\LXMd3h2\\LXMd3h2]LWMc3h2^LXMb3h2^LXMb3g2_LZM`3f2aLYM_3g2aLYM_3f2bLZM^3f2bLZM^3e2cL[M]3e2dLZM\\3f2dLZM\\3e2eL[M[3e2eL\\MZ3c2gL]MY3c2hL\\MX3d2hL\\MX3c2iL]MW3c2iL]MW3b2kL]MU3a2mL_MS3a2mL_MS3a2mL`MR3`2nL`MR3`2oL_MQ3a2oL_MQ3`2PM`MP3`2PM`MP3`2PM`MP3`2QM_Mo2a2QM_Mo2a2QM_Mo2a2QM`Mn2`2SM_Mm2a2SM_Mm2a2SM_Mm2a2SM_Mm2a2SM_Mm2a2TM^Ml2b2TM^Ml2b2TM^Ml2b2TM_Mk2a2UM_Mk2a2VM^Mj2b2VM^Mj2b2VM^Mj2a2WM_Mi2a2XM_Mg2a2YM_Mg2a2YM_Mg2a2YM_Mg2WO[L5n0e0f2SO`L7k0e0e2QOeL8f0g0e2oNhL9c0h0e2lNmL:>k0d2hNRM<:l0d2eNWM=5n0d2bN\\M>0P1d2aN^M>MS1d2^NbM=JU1d2]NdM<IW1c2\\NgM:F[1c2ZNjM8D_1a2XNmM8B`1a2WNPN6@c1`2VNRN5^Of1`2TNUN4[Oi1_2RNYN2YOl1^2QN[N1WOo1^2nM_N0TOR2]2mMaN0ROT2\\2jMfNOoNW2[2iMiNMlN[2[2fMlNNiN\\2[2dMPOMfN`2Y2bMSOLdNc2Y2_MWOKaNf2X2^MYOK_Ng2X2\\M]OJ[Nl2W2YMAHYNo2V2WMDIVNP3V2UM<TO_Mg3U2TMY2n2fMPM\\2P3dMoL]2Q3cMmL^2T3bMkL_2V3`MhLb2X3^MfLd2Z3\\MeLd2\\3\\MbLf2_3YM`Lh2`3XM^Li2c3WM\\Lj2d3VMZLl2g3SMWLn2j3RMULo2k3QMSLQ3m3oLRLR3o3mLoKT3R4lLmKU3S4kLkKW3U4iLiKX3X4hLgKY3Z4fLdK\\3\\4dLcK]3]4cLaK^3`4bL_K_3b4`L]Ka3c4_L]K`3d4`L\\K`3d4`L\\K_3e4aL[K_3f4`LZK`3f4`LZK_3g4aLYK_3g4aLYK_3g4aLYK^3i4aLWK_3i4aLWK_3i4aLWK^3j4bLVK^3j4bLVK^3k4aLUK^3l4bLTK^3l4bLTK^3l4bLTK]3m4cLSK]3n4bLRK]3o4cLQK]3o4cLQK]3o4bLRK]3P5bLPK^3P5bLPK^3P5bLPK]3Q5cLoJ]3Q5cLoJ]3R5bLnJ]3S5cLmJ]3S5cLmJ]3S5cLmJ\\3T5dLlJ\\3U5cLkJ]3U5cLkJ\\3V5dLjJ\\3V5dLjJ[3W5eLiJ[3X5dLhJ\\3X5dLhJ[3Y5eLgJ[3Y5eLgJ[3Y5eLgJZ3[5eLeJ[3[5eLeJ[3[5eLeJZ3\\5fLdJZ3\\5fLdJZ3\\5fLdJY3\\5hLdJX3\\5hLdJW3]5iLcJW3]5iLcJW3]5iLcJV3^5jLbJV3^5jLbJV3^5jLbJU3_5kLaJU3_5kLaJU3_5kL`JU3a5kL_JU3a5kL_JU3a5kL_JT3b5lL^JT3b5lL^JT3b5lL^JS3c5mL]JS3b5nL^JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3e5kL[JU3e5kL[JU3e5kL[JU3e5kL[JU3e5kL[JU3e5kL[JU3a5oL_JQ3^5RMbJn2Z5VMfJj2W5YMiJg2U5[MkJe2U5[MkJe2T5\\MlJd2T5\\MlJd2S5]MmJc2S5]MmJc2S5]MmJc2R5^MnJc2Q5]MoJc2P5^MPKb2P5^MPKb2o4_MQKa2o4_MQKa2n4`MRK]2Q5cMoJY2T5hMlJU2W5kMiJQ2Z5PNfJl1^5TNbJi1`5XN`Jd1d5\\N\\Ja1f5`NZJ\\1j5dNUJY1n5hNRJU1Q6kNoIQ1T6POlIm0W6SOiIi0Z6XOfId0^6\\ObIb0_6_OaI`0`6@`I>a6C_I<b6D^I;c6E]I:g6CYI;l6BTI=P7@PI?T7^OlHa0X7\\OhHc0]7YOcHf0a7WO_Hh0e7UO[Hj0i7SOWHl0m7QOSHn0Q8oNoGP1S8oNmGP1T8POlGo0U8QOkGo0U8QOkGI0^NV8i1jGH1_NU8i1jGG3_NS8j1jGF4_NS8k1iGE5`NR8k1iGC8aNo7l1iGB9aNo7m1hGA;aNm7n1hG@<bNl7n1hG_O>aNk7P2gG^O?bNj7P2gG]O`0bNj7Q2fG\\Ob0aNi7S2eGZOd0aNi7U2cGYOf0`Nh7W2bGXOg0_Ni7S1_G_O1n0i0^Nh7U1_G^O3k0g0`Ni7W1^G]O5g0f0cNi7Y1]G\\O6e0f0dNh7[1]G[O8a0e0gNh7]1\\GZO:>d0iNg7_1\\GYO<:c0lNg7a1[GXO>7a0oNg7a1]GWO>4a0QOf7d1\\GVO`00a9j0PFUOb0M_9n0PFTOd0I^9S1oESOf0F\\9W1oEROh0B[9\\1nEQOj0_OY9`1nEPOg;k0ZDTOl;f0UDYOQ<a0PD^OR<`0oC_OR<`0oC_OQ<`0QD_OP<`0QD_OP<`0QD^OP<b0QDZOS<e0nCWOU<i0R1O1O001O00O100O010O100O100O100O100O100O10O0100O100O100O1J6K5J6K5N2N101O1N2O1O1O1N2O1O1N2O1O001N2O1O1N2O1O1N10100000000000O100000000000000O1000000000O1000O100000000000000O1000000000000O10000000O100000O1000001O1O1O001N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1Ooh4"}, "label": "boy getting his teeth brushed with the help of the father"}]}
{"id": 327694, "image": "COCO_train2014_000000327694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with blond hair is getting his teeth brushed\"."}], "task": "refering", "answer_template": "The \"a child with blond hair is getting his teeth brushed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 54, 55, 56, 57, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 107, 108, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 279, 280, 281, 282, 283, 284, 290, 291, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.165171875, 0.0022500000000000003, 0.9825781250000001, 1.0], "iscrowd": 0, "area": 95425.66700000002, "rle": {"size": [480, 640], "counts": "nTb12k>3eM[2ZNf100001O00000000000000001O00000000000000001O00000000000000001O0000000000oKPLRMP4n2PLRMP4m2QLSMo3m2QLSMo3m2QLTMn3k2TLTMl3l2TLTMl3k2ULUMk3k2ULUMk3k2VLTMj3k2WLUMi3k2WLUMi3j2XLVMh3j2XLWMg3i2ZLVMf3i2[LWMe3i2[LWMe3h2\\LXMd3h2\\LXMd3h2]LWMc3h2^LXMb3h2^LXMb3g2_LZM`3f2aLYM_3g2aLYM_3f2bLZM^3f2bLZM^3e2cL[M]3e2dLZM\\3f2dLZM\\3e2eL[M[3e2eL\\MZ3c2gL]MY3c2hL\\MX3d2hL\\MX3c2iL]MW3c2iL]MW3b2kL]MU3a2mL_MS3a2mL_MS3a2mL`MR3`2nL`MR3`2oL_MQ3a2oL_MQ3`2PM`MP3`2PM`MP3`2PM`MP3`2QM_Mo2a2QM_Mo2a2QM_Mo2a2QM`Mn2`2SM_Mm2a2SM_Mm2a2SM_Mm2a2SM_Mm2a2SM_Mm2a2TM^Ml2b2TM^Ml2b2TM^Ml2b2TM_Mk2a2UM_Mk2a2VM^Mj2b2VM^Mj2b2VM^Mj2a2WM_Mi2a2XM_Mg2a2YM_Mg2a2YM_Mg2a2YM_Mg2WO[L5n0e0f2SO`L7k0e0e2QOeL8f0g0e2oNhL9c0h0e2lNmL:>k0d2hNRM<:l0d2eNWM=5n0d2bN\\M>0P1d2aN^M>MS1d2^NbM=JU1d2]NdM<IW1c2\\NgM:F[1c2ZNjM8D_1a2XNmM8B`1a2WNPN6@c1`2VNRN5^Of1`2TNUN4[Oi1_2RNYN2YOl1^2QN[N1WOo1^2nM_N0TOR2]2mMaN0ROT2\\2jMfNOoNW2[2iMiNMlN[2[2fMlNNiN\\2[2dMPOMfN`2Y2bMSOLdNc2Y2_MWOKaNf2X2^MYOK_Ng2X2\\M]OJ[Nl2W2YMAHYNo2V2WMDIVNP3V2UM<TO_Mg3U2TMY2n2fMPM\\2P3dMoL]2Q3cMmL^2T3bMkL_2V3`MhLb2X3^MfLd2Z3\\MeLd2\\3\\MbLf2_3YM`Lh2`3XM^Li2c3WM\\Lj2d3VMZLl2g3SMWLn2j3RMULo2k3QMSLQ3m3oLRLR3o3mLoKT3R4lLmKU3S4kLkKW3U4iLiKX3X4hLgKY3Z4fLdK\\3\\4dLcK]3]4cLaK^3`4bL_K_3b4`L]Ka3c4_L]K`3d4`L\\K`3d4`L\\K_3e4aL[K_3f4`LZK`3f4`LZK_3g4aLYK_3g4aLYK_3g4aLYK^3i4aLWK_3i4aLWK_3i4aLWK^3j4bLVK^3j4bLVK^3k4aLUK^3l4bLTK^3l4bLTK^3l4bLTK]3m4cLSK]3n4bLRK]3o4cLQK]3o4cLQK]3o4bLRK]3P5bLPK^3P5bLPK^3P5bLPK]3Q5cLoJ]3Q5cLoJ]3R5bLnJ]3S5cLmJ]3S5cLmJ]3S5cLmJ\\3T5dLlJ\\3U5cLkJ]3U5cLkJ\\3V5dLjJ\\3V5dLjJ[3W5eLiJ[3X5dLhJ\\3X5dLhJ[3Y5eLgJ[3Y5eLgJ[3Y5eLgJZ3[5eLeJ[3[5eLeJ[3[5eLeJZ3\\5fLdJZ3\\5fLdJZ3\\5fLdJY3\\5hLdJX3\\5hLdJW3]5iLcJW3]5iLcJW3]5iLcJV3^5jLbJV3^5jLbJV3^5jLbJU3_5kLaJU3_5kLaJU3_5kL`JU3a5kL_JU3a5kL_JU3a5kL_JT3b5lL^JT3b5lL^JT3b5lL^JS3c5mL]JS3b5nL^JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JQ3c5oL]JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3b5nL^JR3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3c5mL]JS3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3d5lL\\JT3e5kL[JU3e5kL[JU3e5kL[JU3e5kL[JU3e5kL[JU3e5kL[JU3a5oL_JQ3^5RMbJn2Z5VMfJj2W5YMiJg2U5[MkJe2U5[MkJe2T5\\MlJd2T5\\MlJd2S5]MmJc2S5]MmJc2S5]MmJc2R5^MnJc2Q5]MoJc2P5^MPKb2P5^MPKb2o4_MQKa2o4_MQKa2n4`MRK]2Q5cMoJY2T5hMlJU2W5kMiJQ2Z5PNfJl1^5TNbJi1`5XN`Jd1d5\\N\\Ja1f5`NZJ\\1j5dNUJY1n5hNRJU1Q6kNoIQ1T6POlIm0W6SOiIi0Z6XOfId0^6\\ObIb0_6_OaI`0`6@`I>a6C_I<b6D^I;c6E]I:g6CYI;l6BTI=P7@PI?T7^OlHa0X7\\OhHc0]7YOcHf0a7WO_Hh0e7UO[Hj0i7SOWHl0m7QOSHn0Q8oNoGP1S8oNmGP1T8POlGo0U8QOkGo0U8QOkGI0^NV8i1jGH1_NU8i1jGG3_NS8j1jGF4_NS8k1iGE5`NR8k1iGC8aNo7l1iGB9aNo7m1hGA;aNm7n1hG@<bNl7n1hG_O>aNk7P2gG^O?bNj7P2gG]O`0bNj7Q2fG\\Ob0aNi7S2eGZOd0aNi7U2cGYOf0`Nh7W2bGXOg0_Ni7S1_G_O1n0i0^Nh7U1_G^O3k0g0`Ni7W1^G]O5g0f0cNi7Y1]G\\O6e0f0dNh7[1]G[O8a0e0gNh7]1\\GZO:>d0iNg7_1\\GYO<:c0lNg7a1[GXO>7a0oNg7a1]GWO>4a0QOf7d1\\GVO`00a9j0PFUOb0M_9n0PFTOd0I^9S1oESOf0F\\9W1oEROh0B[9\\1nEQOj0_OY9`1nEPOg;k0ZDTOl;f0UDYOQ<a0PD^OR<`0oC_OR<`0oC_OQ<`0QD_OP<`0QD_OP<`0QD^OP<b0QDZOS<e0nCWOU<i0R1O1O001O00O100O010O100O100O100O100O100O10O0100O100O100O1J6K5J6K5N2N101O1N2O1O1O1N2O1O1N2O1O001N2O1O1N2O1O1N10100000000000O100000000000000O1000000000O1000O100000000000000O1000000000000O10000000O100000O1000001O1O1O001N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1Ooh4"}, "label": "a child with blond hair is getting his teeth brushed"}]}
{"id": 327694, "image": "COCO_train2014_000000327694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and orange toothbrush\"."}], "task": "refering", "answer_template": "The \"a green and orange toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 177, 178, 199, 200, 201, 202, 203, 223, 224, 225, 226, 249], "bbox": [0.644796875, 0.390125, 0.8618593750000001, 0.6267499999999999], "iscrowd": 0, "area": 5177.6055000000015, "rle": {"size": [480, 640], "counts": "ohQ62l>2O100eBLj;4UDNj;2UDOj;2VDOi;1WDOi;1VD0k;OUD2j;NVD2k;MTD4l;LTD5l;JSD7m;HSD:m;ESD;X=1O2O0O1O2O0O2N1O2N1O2N1O2N3M2N2N1O2N1O2N2O1N2N2N2N2N2N3M2N2N3N1O1O1N2O1O00O10O01O1O1000000O1000001N101O000O2O1O2N1N6K2M2O1N10001N100O2O0O101N100O2O0O101N1000000O10000000000O101O00001O00010O00WCSO_;m0`DTOa;l0]DUOc;k0]DUOd;j0[DWOe;i0ZDXOg;g0XDYOi;g0VDZOk;f0TDZOl;f0SDZOo;e0PD\\OP<d0oC\\OS<c0mC\\OT<d0kC]OV<b0iC]OZ<b0eC^O]<a0bC\\Ob<e0n0N2N2N1O2N11O1004L2ERPY1"}, "label": "a green and orange toothbrush"}]}
{"id": 327694, "image": "COCO_train2014_000000327694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tigger toothbrush\"."}], "task": "refering", "answer_template": "The \"tigger toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 177, 178, 199, 200, 201, 202, 203, 223, 224, 225, 226, 249], "bbox": [0.644796875, 0.390125, 0.8618593750000001, 0.6267499999999999], "iscrowd": 0, "area": 5177.6055000000015, "rle": {"size": [480, 640], "counts": "ohQ62l>2O100eBLj;4UDNj;2UDOj;2VDOi;1WDOi;1VD0k;OUD2j;NVD2k;MTD4l;LTD5l;JSD7m;HSD:m;ESD;X=1O2O0O1O2O0O2N1O2N1O2N1O2N3M2N2N1O2N1O2N2O1N2N2N2N2N2N3M2N2N3N1O1O1N2O1O00O10O01O1O1000000O1000001N101O000O2O1O2N1N6K2M2O1N10001N100O2O0O101N100O2O0O101N1000000O10000000000O101O00001O00010O00WCSO_;m0`DTOa;l0]DUOc;k0]DUOd;j0[DWOe;i0ZDXOg;g0XDYOi;g0VDZOk;f0TDZOl;f0SDZOo;e0PD\\OP<d0oC\\OS<c0mC\\OT<d0kC]OV<b0iC]OZ<b0eC^O]<a0bC\\Ob<e0n0N2N2N1O2N11O1004L2ERPY1"}, "label": "tigger toothbrush"}]}
{"id": 180239, "image": "COCO_train2014_000000180239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plant that is under the railing\"."}], "task": "refering", "answer_template": "The \"the plant that is under the railing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 265, 266, 268, 269, 270, 289, 291, 292, 312, 314, 315, 316, 335, 337], "bbox": [0.5367656249999999, 0.6881100478468899, 0.7612812499999999, 0.9924401913875598], "iscrowd": 0, "area": 3650.343549999999, "rle": {"size": [418, 640], "counts": "ag\\42P=0O2O001O1[CLY<6bCN^<;000000O100000001O1`DEW:<gE0n91PF;e9FYFf0\\9ZOdFf0\\9ZOdFf0]9ZOaFg0`8dNfGe0Jg0`8iNaG`0Nh0a8jN_G>0h0a8lN]G=1g0c8lN[G=1h0d8mNYG;3h0e8mNWG;4h0f8mNUG:60^9HZF8i1ZOS7>SG8;i0b8PORG7:k0e8nNPG7:l0g8nNmF5<o0i8jNiF8>n0k8kNdF7a0n0n8BRG>P9@PG`0R9^OnFb0U9ZOkFh0_9mNnE]Oa0f1g9hNZFS1Q:gNVFh0W:TOQF;m]=DnmB7I5MO22oo3GVPLi0WOh0XO9G2N1O2N2N2NDP>TNWD3N1N2O0O2N101VDDc:4dD:f0Ed:d0[E^Od:c0ZE^Of:b0ZE_Of:a0YE@f:`0ZE@f:a0ZE_Of:`0[E@d:`0]EBa:1iDOg03^:InD4d05]:DRE5c06\\:BSE8b05e:K\\E5c:K]E5c:K^E4b:K`E4`:L`E4a:K`E5_:KbE4_:JbE3a:MZ101O010O00001N100OVTn1"}, "label": "the plant that is under the railing"}]}
{"id": 24591, "image": "COCO_train2014_000000024591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black pants is skiing down the hill\"."}], "task": "refering", "answer_template": "The \"a man in black pants is skiing down the hill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 152, 153, 154, 175, 176, 177, 197, 198, 199, 200, 201, 222, 223, 224, 245, 246, 247, 248], "bbox": [0.603515625, 0.29591666666666666, 0.79121875, 0.6369375], "iscrowd": 0, "area": 8072.361150000012, "rle": {"size": [480, 640], "counts": "_Ue53j>4J6O00000M4M2O1O0O2O1OHkAGT>:7101O1O1OOQND]E=a:FQENiN<V<HjDj0U;]ObDe0_;]O]Dd0c;^OZDd0e;_OXDa0i;@UDa0k;@mCg0S<ZOhCi0Y<o00O1O100O100L4N2O2N11^CjMY<W2cCPNY<[2010OjC_Mo;`2nCRN1CV;`3O02M4M6J0020M1N2O1bDeLU;b32O000O1002SEYL]:h3aE[L^:e3`E\\L`:3ZEW35fLa:3[EW33fLc:2[EU2NdN3UOd:3ZES21eMNm00ZOh:1YEP27nN`:SOYER25iNc:TOXEX22bNf:WOXEY20_Nh:YOXEY20\\Ni:[OWEZ20ZNh:]OWEZ22WNh:^OWE[22UNg:AWEY24TNf:CUEZ27PNd:GUEY27nMf:HTEZ26mMf:JSEY28lMf:JSEY28lMe:LSEKNT2:TNg:LPEK2S28TNh:MoDK7o12XNi:NnDJ9o10XNk:MlDK;Q2NUNl:OkDI=R2NTNl:0hDIR1a1[OcNm:2oEX1d:hNcER1Z<O6G0O100O1O1002O1N00100ON3N1O1O10O1O0000O6K5I3M201O2N2N2N2Ng`n1"}, "label": "a man in black pants is skiing down the hill"}]}
{"id": 24591, "image": "COCO_train2014_000000024591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in dark clothes skiing in the snow\"."}], "task": "refering", "answer_template": "The \"a man in dark clothes skiing in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 152, 153, 154, 175, 176, 177, 197, 198, 199, 200, 201, 222, 223, 224, 245, 246, 247, 248], "bbox": [0.603515625, 0.29591666666666666, 0.79121875, 0.6369375], "iscrowd": 0, "area": 8072.361150000012, "rle": {"size": [480, 640], "counts": "_Ue53j>4J6O00000M4M2O1O0O2O1OHkAGT>:7101O1O1OOQND]E=a:FQENiN<V<HjDj0U;]ObDe0_;]O]Dd0c;^OZDd0e;_OXDa0i;@UDa0k;@mCg0S<ZOhCi0Y<o00O1O100O100L4N2O2N11^CjMY<W2cCPNY<[2010OjC_Mo;`2nCRN1CV;`3O02M4M6J0020M1N2O1bDeLU;b32O000O1002SEYL]:h3aE[L^:e3`E\\L`:3ZEW35fLa:3[EW33fLc:2[EU2NdN3UOd:3ZES21eMNm00ZOh:1YEP27nN`:SOYER25iNc:TOXEX22bNf:WOXEY20_Nh:YOXEY20\\Ni:[OWEZ20ZNh:]OWEZ22WNh:^OWE[22UNg:AWEY24TNf:CUEZ27PNd:GUEY27nMf:HTEZ26mMf:JSEY28lMf:JSEY28lMe:LSEKNT2:TNg:LPEK2S28TNh:MoDK7o12XNi:NnDJ9o10XNk:MlDK;Q2NUNl:OkDI=R2NTNl:0hDIR1a1[OcNm:2oEX1d:hNcER1Z<O6G0O100O1O1002O1N00100ON3N1O1O10O1O0000O6K5I3M201O2N2N2N2Ng`n1"}, "label": "a man in dark clothes skiing in the snow"}]}
{"id": 466964, "image": "COCO_train2014_000000466964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple umbrella hanging upside down\"."}], "task": "refering", "answer_template": "The \"a purple umbrella hanging upside down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 74, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 142], "bbox": [0.12990625, 0.12910416666666666, 0.345984375, 0.37375], "iscrowd": 0, "area": 3411.682550000001, "rle": {"size": [480, 640], "counts": "h_W12o>1O0O1O2N1O1O1O2N1O1O1N2O2N1O1O1O2N1O1O0O2N2O1O1O10O10O100O1O001O001O1gBmNc<S1[CoNe<Q1YCPOh<P1UCSOk<n0QCTOQ=k0mBVOT=h0fBoN3<W=d0iB^OX=b0hB]OY=c0gB]OY=c0hB[OY=d0iBZOX=f0iBYOW=g0iBXOX=g0jBWOW=i0jBVOV=j0kBTOV=l0kBROW=l0jBTOV=l0?0000O11OO101O0000kNZOPDf0k;BRD>n;VO^Dj0Q;ChCCf0\\1^;XOPDZ1P<k00YMgEg0Y:[OeEe0]:^O_E`0g:[OYEe0a<0000000000000000000000O1000000000000000000000000O10000000000000000000000000001O00000000000N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1OO0ThT6"}, "label": "a purple umbrella hanging upside down"}]}
{"id": 466964, "image": "COCO_train2014_000000466964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple umbrella at the end of a row of umbrellas\"."}], "task": "refering", "answer_template": "The \"a purple umbrella at the end of a row of umbrellas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 74, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 142], "bbox": [0.12990625, 0.12910416666666666, 0.345984375, 0.37375], "iscrowd": 0, "area": 3411.682550000001, "rle": {"size": [480, 640], "counts": "h_W12o>1O0O1O2N1O1O1O2N1O1O1N2O2N1O1O1O2N1O1O0O2N2O1O1O10O10O100O1O001O001O1gBmNc<S1[CoNe<Q1YCPOh<P1UCSOk<n0QCTOQ=k0mBVOT=h0fBoN3<W=d0iB^OX=b0hB]OY=c0gB]OY=c0hB[OY=d0iBZOX=f0iBYOW=g0iBXOX=g0jBWOW=i0jBVOV=j0kBTOV=l0kBROW=l0jBTOV=l0?0000O11OO101O0000kNZOPDf0k;BRD>n;VO^Dj0Q;ChCCf0\\1^;XOPDZ1P<k00YMgEg0Y:[OeEe0]:^O_E`0g:[OYEe0a<0000000000000000000000O1000000000000000000000000O10000000000000000000000000001O00000000000N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1OO0ThT6"}, "label": "a purple umbrella at the end of a row of umbrellas"}]}
{"id": 466964, "image": "COCO_train2014_000000466964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella with the red border and light blue background\"."}], "task": "refering", "answer_template": "The \"the umbrella with the red border and light blue background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 105, 128, 151, 174, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222], "bbox": [0.42978125, 0.15212499999999998, 0.7668593749999999, 0.5813333333333334], "iscrowd": 0, "area": 7212.4396, "rle": {"size": [480, 640], "counts": "ooR41P?0O100000000O2N1O100O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O101N1O1O1O1O1O1O1O000000000000001O0000000000001O000001O00000001O0000000000001O000000000000001O0000000000010O000L4H8F:@`0@`0@POfD[Oi:a0kE_Oe9=oFCb88RHH^74VILZ60ZJ0a5ASK>n4[OYKe0f4YO]Kg0c4WO_Ki0a4TObKk0^4TOdKl0\\4ROfKn0Z4oNiKQ1V4nNlKQ1U4mNmKS1S4nNlKR1S4POlKP1T4ROjKm0V4TOjKl0V4UOiKk0X4TOhKl0Z4ROfKm0]4QOcKo0_4oNaKQ1a4mN_KS1c4lN\\KS1g4kNYKU1i4iNWKW1Y900O1000001N1000000O1000000O1000000O1000000O1000000O1000000O10000O1O100O1O100O100O1O2O0O100O1O100O1O100O100O1O10O0100O1O100O1O100O100O00100O100O1O100O1O10O0100O1O100O100OjaV2"}, "label": "the umbrella with the red border and light blue background"}]}
{"id": 466964, "image": "COCO_train2014_000000466964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella with a minnie mouse design and pink border hangs upside down in a shop\"."}], "task": "refering", "answer_template": "The \"an umbrella with a minnie mouse design and pink border hangs upside down in a shop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 105, 128, 151, 174, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222], "bbox": [0.42978125, 0.15212499999999998, 0.7668593749999999, 0.5813333333333334], "iscrowd": 0, "area": 7212.4396, "rle": {"size": [480, 640], "counts": "ooR41P?0O100000000O2N1O100O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O101N1O1O1O1O1O1O1O000000000000001O0000000000001O000001O00000001O0000000000001O000000000000001O0000000000010O000L4H8F:@`0@`0@POfD[Oi:a0kE_Oe9=oFCb88RHH^74VILZ60ZJ0a5ASK>n4[OYKe0f4YO]Kg0c4WO_Ki0a4TObKk0^4TOdKl0\\4ROfKn0Z4oNiKQ1V4nNlKQ1U4mNmKS1S4nNlKR1S4POlKP1T4ROjKm0V4TOjKl0V4UOiKk0X4TOhKl0Z4ROfKm0]4QOcKo0_4oNaKQ1a4mN_KS1c4lN\\KS1g4kNYKU1i4iNWKW1Y900O1000001N1000000O1000000O1000000O1000000O1000000O1000000O10000O1O100O1O100O100O1O2O0O100O1O100O1O100O100O1O10O0100O1O100O1O100O100O00100O100O1O100O1O10O0100O1O100O100OjaV2"}, "label": "an umbrella with a minnie mouse design and pink border hangs upside down in a shop"}]}
{"id": 466964, "image": "COCO_train2014_000000466964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella in the middle\"."}], "task": "refering", "answer_template": "The \"the umbrella in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 54, 77, 100, 119, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171], "bbox": [0.192140625, 0.08764583333333334, 0.5578593749999999, 0.43820833333333337], "iscrowd": 0, "area": 10865.410800000003, "rle": {"size": [480, 640], "counts": "lii11o>1O001N2O1O1O0O2O1O1O1O0O2O1O1O1N101O100M]OgAc0[>1O1N2O1O001O1O1O1O0O2O1O1O1O001O4L2NO1N101001O1O001O001O001O1O0O2O001O001O1O001O001O001O001O001O001O001O00000000001O000000000000001O000000000000001O000000000000001O000001O00000001O000000000000000gLY3N2O11O0cMgFSOZ9MeG3[8mNeHS1\\7VNZIj1f6VNZIj1g6TNZIl1m900000000000000000001O000O10000000000000000000000000000000000000001O000000000O10000000000000000O1O1O100O1O2N100O1O1O100O1O1O100O1O1O100O1O1O101N1O1O100O1O1O100O1O1O101N1O2N101N1O2N101N2N101N101N2N101N2N101NlnT4"}, "label": "the umbrella in the middle"}]}
{"id": 466964, "image": "COCO_train2014_000000466964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a orange colour umberrla\"."}], "task": "refering", "answer_template": "The \"a orange colour umberrla\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 54, 77, 100, 119, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171], "bbox": [0.192140625, 0.08764583333333334, 0.5578593749999999, 0.43820833333333337], "iscrowd": 0, "area": 10865.410800000003, "rle": {"size": [480, 640], "counts": "lii11o>1O001N2O1O1O0O2O1O1O1O0O2O1O1O1N101O100M]OgAc0[>1O1N2O1O001O1O1O1O0O2O1O1O1O001O4L2NO1N101001O1O001O001O001O1O0O2O001O001O1O001O001O001O001O001O001O001O00000000001O000000000000001O000000000000001O000000000000001O000001O00000001O000000000000000gLY3N2O11O0cMgFSOZ9MeG3[8mNeHS1\\7VNZIj1f6VNZIj1g6TNZIl1m900000000000000000001O000O10000000000000000000000000000000000000001O000000000O10000000000000000O1O1O100O1O2N100O1O1O100O1O1O100O1O1O100O1O1O101N1O1O100O1O1O100O1O1O101N1O2N101N1O2N101N2N101N101N2N101N2N101NlnT4"}, "label": "a orange colour umberrla"}]}
{"id": 532505, "image": "COCO_train2014_000000532505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing white sitting on a blue sofa\"."}], "task": "refering", "answer_template": "The \"a woman wearing white sitting on a blue sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 238, 239, 240, 242, 243, 265, 266, 288, 289, 290, 312, 313, 314], "bbox": [0.329265625, 0.318625, 0.686265625, 0.8252708333333333], "iscrowd": 0, "area": 15906.243699999995, "rle": {"size": [480, 640], "counts": "SUS3=T>a0^Oa0L5M3M3M2N3L4M3M2iNjMaEY2\\:YNREj1k:X1L4M2N3M3M3L3N3M3M3L3N3M001N2O001O1O10O10000O01000O10000O1000O01000000O10000O10000O101O0O1000000O103L5L3L4M3L5L3L4M3L5L3L4M3L5L3L3N2M4L3N2M3N3L3N2M4M2M3N3L3N2M3N2M3N3L3N2M3N2M3N1N201N1OO1000000O10000O10000O10000O2O000O^OnA8R>GQB7o=HSB7m=HUB7j=IYB5g=J[B5e=J]B5b=L`B2`=MbB2^=MdB2[=NgB1Y=NiB3U=LlB8o<HSC<f=4L4M3L4L4M3L4M3L5K4M3L4L6K6I7I6K6I3NO000010O0001O003N2M3M3M3N2M2N3M3N2M3M3M3M3M3M3M3N0O2N2N2N2N2N2N101N2N2N2N1O00O2O00001O000O2O00001O0O101O001O0:F2O0O00100O00100O0010O01O10O01L4M2M4LTkm2"}, "label": "a woman wearing white sitting on a blue sofa"}]}
{"id": 532505, "image": "COCO_train2014_000000532505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady embarrassing and sitting blue color sofa wearing white color shirt\"."}], "task": "refering", "answer_template": "The \"a lady embarrassing and sitting blue color sofa wearing white color shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 238, 239, 240, 242, 243, 265, 266, 288, 289, 290, 312, 313, 314], "bbox": [0.329265625, 0.318625, 0.686265625, 0.8252708333333333], "iscrowd": 0, "area": 15906.243699999995, "rle": {"size": [480, 640], "counts": "SUS3=T>a0^Oa0L5M3M3M2N3L4M3M2iNjMaEY2\\:YNREj1k:X1L4M2N3M3M3L3N3M3M3L3N3M001N2O001O1O10O10000O01000O10000O1000O01000000O10000O10000O101O0O1000000O103L5L3L4M3L5L3L4M3L5L3L4M3L5L3L3N2M4L3N2M3N3L3N2M4M2M3N3L3N2M3N2M3N3L3N2M3N2M3N1N201N1OO1000000O10000O10000O10000O2O000O^OnA8R>GQB7o=HSB7m=HUB7j=IYB5g=J[B5e=J]B5b=L`B2`=MbB2^=MdB2[=NgB1Y=NiB3U=LlB8o<HSC<f=4L4M3L4L4M3L4M3L5K4M3L4L6K6I7I6K6I3NO000010O0001O003N2M3M3M3N2M2N3M3N2M3M3M3M3M3M3M3N0O2N2N2N2N2N2N101N2N2N2N1O00O2O00001O000O2O00001O0O101O001O0:F2O0O00100O00100O0010O01O10O01L4M2M4LTkm2"}, "label": "a lady embarrassing and sitting blue color sofa wearing white color shirt"}]}
{"id": 532505, "image": "COCO_train2014_000000532505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plaid couch right side\"."}], "task": "refering", "answer_template": "The \"plaid couch right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 204, 205, 206], "bbox": [0.559671875, 0.22795833333333335, 0.9774218749999999, 0.49462500000000004], "iscrowd": 0, "area": 27996.826350000007, "rle": {"size": [480, 640], "counts": "^oW53l>9F:G9G9F:G8H1N3N1O2M2O2J5J6J7J5J7I6O2O0O101O001O001N10001O001O000O2O000000000O1000000000001N10000000000O1000000000000O10O1000000O010000000O1000O10O1000000O10000O100O010O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100000000000000000000000000000000000000000000000O10000000001O000000000000000000001O000000001O00001O0O101O00001O00001O00001O001O00001O00001O0000000000000O10001O000000000000000000000000000000000000000000000O10000010O0000000000000000001O00000000000000004L5K4L4L5K4L3N2M3M3M2N3M3M3M3M3M2N3M2N3M2N3M2N2N3L3C=C>BX\\6"}, "label": "plaid couch right side"}]}
{"id": 532505, "image": "COCO_train2014_000000532505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plaid sofa with two pillows that no one is sitting on\"."}], "task": "refering", "answer_template": "The \"plaid sofa with two pillows that no one is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 204, 205, 206], "bbox": [0.559671875, 0.22795833333333335, 0.9774218749999999, 0.49462500000000004], "iscrowd": 0, "area": 27996.826350000007, "rle": {"size": [480, 640], "counts": "^oW53l>9F:G9G9F:G8H1N3N1O2M2O2J5J6J7J5J7I6O2O0O101O001O001N10001O001O000O2O000000000O1000000000001N10000000000O1000000000000O10O1000000O010000000O1000O10O1000000O10000O100O010O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100000000000000000000000000000000000000000000000O10000000001O000000000000000000001O000000001O00001O0O101O00001O00001O00001O001O00001O00001O0000000000000O10001O000000000000000000000000000000000000000000000O10000010O0000000000000000001O00000000000000004L5K4L4L5K4L3N2M3M3M2N3M3M3M3M3M2N3M2N3M2N3M2N2N3L3C=C>BX\\6"}, "label": "plaid sofa with two pillows that no one is sitting on"}]}
{"id": 532505, "image": "COCO_train2014_000000532505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue couch that women are sitting on\"."}], "task": "refering", "answer_template": "The \"a blue couch that women are sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 187, 188, 189, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 329, 330, 331, 332, 333, 334, 356, 357], "bbox": [0.12607812499999999, 0.45366666666666666, 0.559265625, 0.9105624999999999], "iscrowd": 0, "area": 35750.398949999995, "rle": {"size": [480, 640], "counts": "mXV19a>:F9F:K6J3N2N2N2N2N2N2N2N1N3N1O1nNnMUES2j:RNQEP2n:UNlDl1S;YNhDh1W;]NdDd1[;bN_D_1a;Q1N1O1O1O101N1O1O3N4K4L5K4L4M4K4L4L5K4M4K4L4L5L3L4L5K4L4M4K1O1O2O0O21L5L4K5L3M4K5L4L4K4M4L4M1O2N2O0O2N2N1O2N2N1O2N2N000001O00000000000000000000001O01O0000000000000000000001O00000000000001O00000001O00000000000000000000001O01O001O5K4L2N2N2N2N2N2O0O2N2N2N3M2N2N1O00O1000001O000O1000000000000O10001O00000O1000000000000O101O0000000O10000000000O101N100O100O100O1O100O100O101N100O1O100O100O100O100O2N100O100O100O100O1O10001O001O1O2N1O2N1O1O2N000O2O000000001O000000001O0000O2O0O1O2O0O100O2O0O1O2O0L:F;E:G:E;F9Gd0\\Og0YOh0XOSjS4"}, "label": "a blue couch that women are sitting on"}]}
{"id": 532505, "image": "COCO_train2014_000000532505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue couch with two woman on it\"."}], "task": "refering", "answer_template": "The \"a blue couch with two woman on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 187, 188, 189, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 329, 330, 331, 332, 333, 334, 356, 357], "bbox": [0.12607812499999999, 0.45366666666666666, 0.559265625, 0.9105624999999999], "iscrowd": 0, "area": 35750.398949999995, "rle": {"size": [480, 640], "counts": "mXV19a>:F9F:K6J3N2N2N2N2N2N2N2N1N3N1O1nNnMUES2j:RNQEP2n:UNlDl1S;YNhDh1W;]NdDd1[;bN_D_1a;Q1N1O1O1O101N1O1O3N4K4L5K4L4M4K4L4L5K4M4K4L4L5L3L4L5K4L4M4K1O1O2O0O21L5L4K5L3M4K5L4L4K4M4L4M1O2N2O0O2N2N1O2N2N1O2N2N000001O00000000000000000000001O01O0000000000000000000001O00000000000001O00000001O00000000000000000000001O01O001O5K4L2N2N2N2N2N2O0O2N2N2N3M2N2N1O00O1000001O000O1000000000000O10001O00000O1000000000000O101O0000000O10000000000O101N100O100O100O1O100O100O101N100O1O100O100O100O100O2N100O100O100O100O1O10001O001O1O2N1O2N1O1O2N000O2O000000001O000000001O0000O2O0O1O2O0O100O2O0O1O2O0L:F;E:G:E;F9Gd0\\Og0YOh0XOSjS4"}, "label": "a blue couch with two woman on it"}]}
{"id": 499738, "image": "COCO_train2014_000000499738.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a elephant bigger than the surrounding elephants\"."}], "task": "refering", "answer_template": "The \"a elephant bigger than the surrounding elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 202, 203, 204, 205, 206, 207, 208, 209, 210, 224, 225, 226, 227, 228, 229, 230, 231, 232, 246, 247, 248, 249, 250, 251, 252, 253, 254, 271, 272, 273, 274, 275, 276, 295, 296, 297, 298, 316, 317, 318, 319, 320, 338, 339, 340, 342, 343, 360, 361, 362, 363, 364, 383], "bbox": [0.1779248366013072, 0.39751633986928103, 0.5991013071895425, 0.7804084967320262], "iscrowd": 0, "area": 28481.04390000002, "rle": {"size": [612, 612], "counts": "a^Q22ob04M2M4L3N3L3M4M2K5O2N1O1O1N2O1O2N1O1N2O1O1O2M2O1O1O1O1N3N1O1O1O1N2O2N1O10000O101O0O100O10001N10000O100O2O000O10000O2O0O10000O101O0O100O10000O10000O100O2O000O10000000000000000000010O0000000000000000000000001O00001O001O001O001O001O001O1O001O001O0010O01O1TCPMb9P3YFWMe9j2UF]Mi9c2RFdMl9]2nEjMP:V2kEPNT:Q2fEVNX:j1cE\\N\\:e1^E`Nb:a1XEeNg:\\1SEiNm:X1PEkNP;U1nDlNR;V1jDmNU;T1iDnNV;S1gDoNY;R1eDPOZ;R1bDPO^;Q1`DPO`;Q1]DROb;o0\\DROd;P1XDSOg;n0VDTOj;m0TDUOl;k0QDWOo;o301O001O001O00001O0001O000O100O1VLZD6f;EaD9_;GbD7`;HbD6^;JcD4^;KdD4\\;LfD2Z;NgD0Z;0hDNX;2iDLX;4gDMZ;2eDO[;1cD\\MYOU2T<?bD]MAm1m;e0bD^M0_1^;S1aD_M?P1P;a1`D_Mo0b0a:o1_D`M]14T:\\2^D`Ma12Q:^2]DaMc10Q:]2\\DcMe1Oo9^2[DdMf1No9^2YDeMj1Lm9`2XDcMl1Ml9`2WDdMn1Kk9a2WDcMP2Ki9c2VDbMQ2Ki9c2VDaMS2Kh9c2TDcMT2Jh9d2WD]MS2Ne9f2[DYMQ20d9g2_DTMn15c9h2aDPMm15d9k2WGSMk8n2UGPMl8Q3c31O1O1O1O2N2N2N2N2M4M8H8H5K200O100O100O010O100O100O100O100O16JnAeKb=R4YBSLo=e3kA`L^>X3`AiLj>n2TASMU?e2i@ZMc?]2[@bM^`0h1`_OWNVa0`14L3N3L4L3N3N1N3M2N2N3M2N2O1I8F9G^Pb4"}, "label": "a elephant bigger than the surrounding elephants"}]}
{"id": 499738, "image": "COCO_train2014_000000499738.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"largest elephant standing behind baby elephant\"."}], "task": "refering", "answer_template": "The \"largest elephant standing behind baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 202, 203, 204, 205, 206, 207, 208, 209, 210, 224, 225, 226, 227, 228, 229, 230, 231, 232, 246, 247, 248, 249, 250, 251, 252, 253, 254, 271, 272, 273, 274, 275, 276, 295, 296, 297, 298, 316, 317, 318, 319, 320, 338, 339, 340, 342, 343, 360, 361, 362, 363, 364, 383], "bbox": [0.1779248366013072, 0.39751633986928103, 0.5991013071895425, 0.7804084967320262], "iscrowd": 0, "area": 28481.04390000002, "rle": {"size": [612, 612], "counts": "a^Q22ob04M2M4L3N3L3M4M2K5O2N1O1O1N2O1O2N1O1N2O1O1O2M2O1O1O1O1N3N1O1O1O1N2O2N1O10000O101O0O100O10001N10000O100O2O000O10000O2O0O10000O101O0O100O10000O10000O100O2O000O10000000000000000000010O0000000000000000000000001O00001O001O001O001O001O001O1O001O001O0010O01O1TCPMb9P3YFWMe9j2UF]Mi9c2RFdMl9]2nEjMP:V2kEPNT:Q2fEVNX:j1cE\\N\\:e1^E`Nb:a1XEeNg:\\1SEiNm:X1PEkNP;U1nDlNR;V1jDmNU;T1iDnNV;S1gDoNY;R1eDPOZ;R1bDPO^;Q1`DPO`;Q1]DROb;o0\\DROd;P1XDSOg;n0VDTOj;m0TDUOl;k0QDWOo;o301O001O001O00001O0001O000O100O1VLZD6f;EaD9_;GbD7`;HbD6^;JcD4^;KdD4\\;LfD2Z;NgD0Z;0hDNX;2iDLX;4gDMZ;2eDO[;1cD\\MYOU2T<?bD]MAm1m;e0bD^M0_1^;S1aD_M?P1P;a1`D_Mo0b0a:o1_D`M]14T:\\2^D`Ma12Q:^2]DaMc10Q:]2\\DcMe1Oo9^2[DdMf1No9^2YDeMj1Lm9`2XDcMl1Ml9`2WDdMn1Kk9a2WDcMP2Ki9c2VDbMQ2Ki9c2VDaMS2Kh9c2TDcMT2Jh9d2WD]MS2Ne9f2[DYMQ20d9g2_DTMn15c9h2aDPMm15d9k2WGSMk8n2UGPMl8Q3c31O1O1O1O2N2N2N2N2M4M8H8H5K200O100O100O010O100O100O100O100O16JnAeKb=R4YBSLo=e3kA`L^>X3`AiLj>n2TASMU?e2i@ZMc?]2[@bM^`0h1`_OWNVa0`14L3N3L4L3N3N1N3M2N2N3M2N2O1I8F9G^Pb4"}, "label": "largest elephant standing behind baby elephant"}]}
{"id": 426016, "image": "COCO_train2014_000000426016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange parking meter with grafitti\"."}], "task": "refering", "answer_template": "The \"an orange parking meter with grafitti\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 44, 45, 46, 57, 58, 59, 70, 71, 72, 83, 84, 85, 96, 97, 98, 109, 110, 123], "bbox": [0.39383999999999997, 0.13378, 0.5743466666666667, 0.53032], "iscrowd": 0, "area": 9401.839900000003, "rle": {"size": [500, 375], "counts": "d[X2;X?<D<QBYOZ<T1oB^Oh<o1G9G8H6J7H7J7H8H6J3M4N2O1N2O001O1O0O10001O0000005K8H7J5K00000000O1O1O1O1O001O1O1O4L6I7JdEjJ[:W51O1M3K4]Od0E:K500O1O<Da0@?A<D4L4L4K7mNS1Fan]2"}, "label": "an orange parking meter with grafitti"}]}
{"id": 426016, "image": "COCO_train2014_000000426016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"caramel colored parking meter\"."}], "task": "refering", "answer_template": "The \"caramel colored parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 44, 45, 46, 57, 58, 59, 70, 71, 72, 83, 84, 85, 96, 97, 98, 109, 110, 123], "bbox": [0.39383999999999997, 0.13378, 0.5743466666666667, 0.53032], "iscrowd": 0, "area": 9401.839900000003, "rle": {"size": [500, 375], "counts": "d[X2;X?<D<QBYOZ<T1oB^Oh<o1G9G8H6J7H7J7H8H6J3M4N2O1N2O001O1O0O10001O0000005K8H7J5K00000000O1O1O1O1O001O1O1O4L6I7JdEjJ[:W51O1M3K4]Od0E:K500O1O<Da0@?A<D4L4L4K7mNS1Fan]2"}, "label": "caramel colored parking meter"}]}
{"id": 483363, "image": "COCO_train2014_000000483363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white and red tank\"."}], "task": "refering", "answer_template": "The \"man in white and red tank\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 82, 83, 84, 86, 87, 88, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 127, 128, 129, 130, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 252, 265, 266, 267, 268, 280, 281, 283, 284], "bbox": [0.4658177570093458, 0.193, 0.9620093457943926, 0.831828125], "iscrowd": 0, "area": 34365.66329999999, "rle": {"size": [640, 428], "counts": "Ucl3g0Wc0?A?An0RO4L5K4L4M3M3QNYM`Bj2[=]M`Bf2]=^M^Bf2_=^M]Bd2a=aMZBb2d=bMWBa2f=dMUB_2h=cMVB_2i=bMUBa2h=aMVBb2h=_MUBd2i=^MUBe2h=l1Ne0[O2M3N2N2N2N1O2M3N2N2N2N1O2N2N2N2N2N1O2N2N2O1O1O10O01O100O1O10O0100O1O100O000>CO001O0O2O001O001N1O1O2N1O2N1O`M_EcK_:]4cEbK\\:^4fEbKX:^4jEaKU:`4lE_KS:a4PF^Kn9b4TF]Kk9d4VF\\Kh9d4ZF[Ke9e4]F[Ka9e4aFZK^9g4cFYK[9g4gFXKW9i4kFWKHgNj7S6_HUKDjNl7R6bHSK_OnNm7P6fHRKYOPOP8;kFP5n1dKTOSOS89nFn4l1fKoNUOV87SGk4k1gKiNZOW85XGh4i1iKdN\\OZ83]Gf4f1jK`N_O]81`Ge4c1kK\\NB_8NfGd4_1kKYNEa8MiGc4^1iKUNIc8KnGa4[1kKPNLf8HRH`4X1kKmMOh8FWH_4T1mLd6eNYH_4S1jLd6fN\\H_4Q1jLa6hN_H_4o0hLa6iNcH_4k0gLa6kNfH^4i0fL`6kNiH_4g0eL_6mNlH^4d0dL_6nNoH_4b0aL^6QORI]4`0bL\\6QOWI]4<aL\\6SOYI]4:_L\\6TO]I]47^LZ6VO`I\\46]LY6WOdI\\42\\LY6YOPJR4GaLZ6]O]Jg3XOhL]6BaJa3ROiL_6FfJ]3jNiLc6JiJY3dNhLe6OnJU3\\NhLh63SKHbLX2c1kMh66ZLj1mLnMi68]Lj1iLlMk6:^Lj1fLjMm6;aLi1bLjMn6=bLi1_LhMP7>dLj1[LfMR7`0eLj1XL`MX7e0cLj1Z5VNhJj1V5UNmJk1Q5UNQKk1m4TNVKk1i4UN[Ki1c4VNbKh1\\4XNhKe1W4ZNnKd1o3\\NTLd1i3]NYLc1d3]N_Lb1_3_NcLa1Z3_NiLa1cIhMP9g0`Ma1]IjMQ9f0dM_1YImMR9c0hM`1SIPNR9a0mM_1mHSNT9>RN_1gHUNU9<VNm1h1TNZNk1d1UN^Nk1`1VNbNi1\\1WNgNh1X1XNiNg1V1YNmNf1Q1ZNROd1m0]NTOc1j0]N]O^1a0bNIU17jN2o0LRO<f0CZOf0?XOAQ17nNIV15iNJZ15dNL]13bNMa12]NNf10ZNOi10UN0m1ORN1Q2NnM2T2MlM1W2NiM0X21hMMY24gMJZ27gMGY2;fMC[2>eM_O]2b0cMUOe2l0W81O2N1O1O1O101OO1O1O2N1O1O1O1O2N1O1O1N2O2N2N1O2N1O2N2N3M6Ji_9"}, "label": "man in white and red tank"}]}
{"id": 483363, "image": "COCO_train2014_000000483363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a sleeveless shirt\"."}], "task": "refering", "answer_template": "The \"a man in a sleeveless shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 82, 83, 84, 86, 87, 88, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 127, 128, 129, 130, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 252, 265, 266, 267, 268, 280, 281, 283, 284], "bbox": [0.4658177570093458, 0.193, 0.9620093457943926, 0.831828125], "iscrowd": 0, "area": 34365.66329999999, "rle": {"size": [640, 428], "counts": "Ucl3g0Wc0?A?An0RO4L5K4L4M3M3QNYM`Bj2[=]M`Bf2]=^M^Bf2_=^M]Bd2a=aMZBb2d=bMWBa2f=dMUB_2h=cMVB_2i=bMUBa2h=aMVBb2h=_MUBd2i=^MUBe2h=l1Ne0[O2M3N2N2N2N1O2M3N2N2N2N1O2N2N2N2N2N1O2N2N2O1O1O10O01O100O1O10O0100O1O100O000>CO001O0O2O001O001N1O1O2N1O2N1O`M_EcK_:]4cEbK\\:^4fEbKX:^4jEaKU:`4lE_KS:a4PF^Kn9b4TF]Kk9d4VF\\Kh9d4ZF[Ke9e4]F[Ka9e4aFZK^9g4cFYK[9g4gFXKW9i4kFWKHgNj7S6_HUKDjNl7R6bHSK_OnNm7P6fHRKYOPOP8;kFP5n1dKTOSOS89nFn4l1fKoNUOV87SGk4k1gKiNZOW85XGh4i1iKdN\\OZ83]Gf4f1jK`N_O]81`Ge4c1kK\\NB_8NfGd4_1kKYNEa8MiGc4^1iKUNIc8KnGa4[1kKPNLf8HRH`4X1kKmMOh8FWH_4T1mLd6eNYH_4S1jLd6fN\\H_4Q1jLa6hN_H_4o0hLa6iNcH_4k0gLa6kNfH^4i0fL`6kNiH_4g0eL_6mNlH^4d0dL_6nNoH_4b0aL^6QORI]4`0bL\\6QOWI]4<aL\\6SOYI]4:_L\\6TO]I]47^LZ6VO`I\\46]LY6WOdI\\42\\LY6YOPJR4GaLZ6]O]Jg3XOhL]6BaJa3ROiL_6FfJ]3jNiLc6JiJY3dNhLe6OnJU3\\NhLh63SKHbLX2c1kMh66ZLj1mLnMi68]Lj1iLlMk6:^Lj1fLjMm6;aLi1bLjMn6=bLi1_LhMP7>dLj1[LfMR7`0eLj1XL`MX7e0cLj1Z5VNhJj1V5UNmJk1Q5UNQKk1m4TNVKk1i4UN[Ki1c4VNbKh1\\4XNhKe1W4ZNnKd1o3\\NTLd1i3]NYLc1d3]N_Lb1_3_NcLa1Z3_NiLa1cIhMP9g0`Ma1]IjMQ9f0dM_1YImMR9c0hM`1SIPNR9a0mM_1mHSNT9>RN_1gHUNU9<VNm1h1TNZNk1d1UN^Nk1`1VNbNi1\\1WNgNh1X1XNiNg1V1YNmNf1Q1ZNROd1m0]NTOc1j0]N]O^1a0bNIU17jN2o0LRO<f0CZOf0?XOAQ17nNIV15iNJZ15dNL]13bNMa12]NNf10ZNOi10UN0m1ORN1Q2NnM2T2MlM1W2NiM0X21hMMY24gMJZ27gMGY2;fMC[2>eM_O]2b0cMUOe2l0W81O2N1O1O1O101OO1O1O2N1O1O1O1O2N1O1O1N2O2N2N1O2N1O2N2N3M6Ji_9"}, "label": "a man in a sleeveless shirt"}]}
{"id": 483363, "image": "COCO_train2014_000000483363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy wearing blue tshirt and walking\"."}], "task": "refering", "answer_template": "The \"guy wearing blue tshirt and walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 62, 63, 77, 78, 79, 92, 93, 94, 107, 108, 121, 122, 123, 136, 137, 150, 151, 152, 165, 166, 180, 181], "bbox": [0.01766355140186916, 0.140484375, 0.3050233644859813, 0.568421875], "iscrowd": 0, "area": 10054.650450000001, "rle": {"size": [640, 428], "counts": "_Z55hc06J6J6J5K6K5K3N1N3M2O1N3M2O1N3M2N2N3M2O1N3M2NnNk^O]OTa0b0Q_O[On`0d0W_OVOk`0i0Z_OQOh`0o0T1OVOoNh^OP1Xa0ROg^Om0Ya0VOe^Oi0\\a0YOb^Of0^a0\\Oa^Oc0_a0@_^O?ba0B]^O=ca0FZ^O:fa0HY^O7ha0KV^O4ja0NU^O2ja01T^OOla03R^OMma05R^OKma08Q^OHoa09P^OGoa0U1O1N2J6J6\\Od0O1O001O0^Ob0J7O001O010O001O001TO`LYA_3e>lLRAT3n>o00O100O100O1O2O0O2O0O2O1N1O2O0O2O1N101N2N2O1N101N2O1N2O2M4\\Mo_OS1T`0iNP@U1S`0fNQ@W1R`0fNQ@X1R`0cNR@Z1V`0^Nm_O`1\\`0VNf_Oh1[a0M3N2M3N2M3N2M2N3N2M3N2M3N2M3M2O2M3N4K5Ll^i5"}, "label": "guy wearing blue tshirt and walking"}]}
{"id": 483363, "image": "COCO_train2014_000000483363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"reserve players having their warmup\"."}], "task": "refering", "answer_template": "The \"reserve players having their warmup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 62, 63, 77, 78, 79, 92, 93, 94, 107, 108, 121, 122, 123, 136, 137, 150, 151, 152, 165, 166, 180, 181], "bbox": [0.01766355140186916, 0.140484375, 0.3050233644859813, 0.568421875], "iscrowd": 0, "area": 10054.650450000001, "rle": {"size": [640, 428], "counts": "_Z55hc06J6J6J5K6K5K3N1N3M2O1N3M2O1N3M2N2N3M2O1N3M2NnNk^O]OTa0b0Q_O[On`0d0W_OVOk`0i0Z_OQOh`0o0T1OVOoNh^OP1Xa0ROg^Om0Ya0VOe^Oi0\\a0YOb^Of0^a0\\Oa^Oc0_a0@_^O?ba0B]^O=ca0FZ^O:fa0HY^O7ha0KV^O4ja0NU^O2ja01T^OOla03R^OMma05R^OKma08Q^OHoa09P^OGoa0U1O1N2J6J6\\Od0O1O001O0^Ob0J7O001O010O001O001TO`LYA_3e>lLRAT3n>o00O100O100O1O2O0O2O0O2O1N1O2O0O2O1N101N2N2O1N101N2O1N2O2M4\\Mo_OS1T`0iNP@U1S`0fNQ@W1R`0fNQ@X1R`0cNR@Z1V`0^Nm_O`1\\`0VNf_Oh1[a0M3N2M3N2M3N2M2N3N2M3N2M3N2M3M2O2M3N4K5Ll^i5"}, "label": "reserve players having their warmup"}]}
{"id": 483363, "image": "COCO_train2014_000000483363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an man wearing red at a frisbee match\"."}], "task": "refering", "answer_template": "The \"an man wearing red at a frisbee match\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 65, 66, 67, 80, 81, 82, 94, 95, 96, 97, 108, 109, 110, 111, 112, 123, 124, 125, 126, 138, 139, 140, 141, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 199, 200, 201, 202, 214, 215, 216, 229, 230, 231, 244, 245, 258, 259, 260, 261, 274, 275, 276], "bbox": [0.16801401869158877, 0.134453125, 0.5074065420560747, 0.82659375], "iscrowd": 0, "area": 33180.56799999999, "rle": {"size": [640, 428], "counts": "bZ]12lc03M4K4oNGW^O<da0OS^O5ha06n]OMma0W1J5K6J5L4K6J5K6J5K6K3L3O1O1O10O01O1O1O2O2M3M2N3N2M3M2N3WDmKa7V4]HlKa7V4\\HlKb7V4\\HmKa7V4\\HlKb7W4\\HkKa7X4\\HjKc7X4ZHkKc7W4\\HjKb7Y4[HjKZMSOh9V5kHiK[MWOe9R5mHjK[MYOc9o4QIiKZM^O_9l4TIiKZMA\\9i4WIhK[MA\\9j4WIcLh6^3VIdLh6_3UIcLj6_3SIcLl6_3RIbLm6_3QIcLn6_3oHcLP7_3nHbLQ7`3lHbLS7_3kHcLT7_3iHcLV7_3fHdLY7^3bHfL]7\\3_HgL`7Z3]HiLb7Y3ZHjLe7X3WHkLh7W3THlLk7U3RHnLm7T3oGoLQ8R3kGQMT8Q3hGRMW8o2fGTMY8n2cGUM\\8m2`GVMc7nKPIn6YOWMe7nKPIV;n6mDPIT;n6oDPIR;n6QEPIP;n6SEPIn:o6TEoHm:o6VEoHk:o6XEoHi:o6[EnHf:P7]EnHd:P7_EnHb:Q7W1N2N2N2N2NQN_IcF_6]9hI^FW6a9QJYFm5h9UJWFi5i9YJWFe5i9^JVF`5j9cJUF[5k9hJTFW5k9lJTFR5l9QKSFm4n9UKQFj4n9YKQFh4l9[KSFe4k9^KTFc4i9`KVF`4h9cKWF^4g9dKXF\\4f9fKZF[4c9hK\\FX4b9kK]FU4a9nK^FS4_9PL`FP4_9RL`Fo3S5PJWK2T1R2aNl3o4]JgJ4e1m8_3cH]L_7_3gH]L\\7^3jH^LX7^3nH^LU7\\3RIaLo6[3WIaLl6Z3ZIbLh6Z3^IbLe6Y3bIbL`6Y3gIcL\\6X3jIeLX6V3nIfLV6T3PJhLT6m1WKoMl4e1aKWNc4a1eK[N_4a1eK\\N^4_1gK]N]4_1U7K4L5L4K5L4K5L4K5K5LogS4"}, "label": "an man wearing red at a frisbee match"}]}
{"id": 483363, "image": "COCO_train2014_000000483363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in red shirt with three white x ' s on front\"."}], "task": "refering", "answer_template": "The \"man in red shirt with three white x ' s on front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 65, 66, 67, 80, 81, 82, 94, 95, 96, 97, 108, 109, 110, 111, 112, 123, 124, 125, 126, 138, 139, 140, 141, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 199, 200, 201, 202, 214, 215, 216, 229, 230, 231, 244, 245, 258, 259, 260, 261, 274, 275, 276], "bbox": [0.16801401869158877, 0.134453125, 0.5074065420560747, 0.82659375], "iscrowd": 0, "area": 33180.56799999999, "rle": {"size": [640, 428], "counts": "bZ]12lc03M4K4oNGW^O<da0OS^O5ha06n]OMma0W1J5K6J5L4K6J5K6J5K6K3L3O1O1O10O01O1O1O2O2M3M2N3N2M3M2N3WDmKa7V4]HlKa7V4\\HlKb7V4\\HmKa7V4\\HlKb7W4\\HkKa7X4\\HjKc7X4ZHkKc7W4\\HjKb7Y4[HjKZMSOh9V5kHiK[MWOe9R5mHjK[MYOc9o4QIiKZM^O_9l4TIiKZMA\\9i4WIhK[MA\\9j4WIcLh6^3VIdLh6_3UIcLj6_3SIcLl6_3RIbLm6_3QIcLn6_3oHcLP7_3nHbLQ7`3lHbLS7_3kHcLT7_3iHcLV7_3fHdLY7^3bHfL]7\\3_HgL`7Z3]HiLb7Y3ZHjLe7X3WHkLh7W3THlLk7U3RHnLm7T3oGoLQ8R3kGQMT8Q3hGRMW8o2fGTMY8n2cGUM\\8m2`GVMc7nKPIn6YOWMe7nKPIV;n6mDPIT;n6oDPIR;n6QEPIP;n6SEPIn:o6TEoHm:o6VEoHk:o6XEoHi:o6[EnHf:P7]EnHd:P7_EnHb:Q7W1N2N2N2N2NQN_IcF_6]9hI^FW6a9QJYFm5h9UJWFi5i9YJWFe5i9^JVF`5j9cJUF[5k9hJTFW5k9lJTFR5l9QKSFm4n9UKQFj4n9YKQFh4l9[KSFe4k9^KTFc4i9`KVF`4h9cKWF^4g9dKXF\\4f9fKZF[4c9hK\\FX4b9kK]FU4a9nK^FS4_9PL`FP4_9RL`Fo3S5PJWK2T1R2aNl3o4]JgJ4e1m8_3cH]L_7_3gH]L\\7^3jH^LX7^3nH^LU7\\3RIaLo6[3WIaLl6Z3ZIbLh6Z3^IbLe6Y3bIbL`6Y3gIcL\\6X3jIeLX6V3nIfLV6T3PJhLT6m1WKoMl4e1aKWNc4a1eK[N_4a1eK\\N^4_1gK]N]4_1U7K4L5L4K5L4K5L4K5K5LogS4"}, "label": "man in red shirt with three white x ' s on front"}]}
{"id": 262180, "image": "COCO_train2014_000000262180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bananas on top of a cake\"."}], "task": "refering", "answer_template": "The \"bananas on top of a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131], "bbox": [0.47131249999999997, 0.1000625, 0.728046875, 0.3534583333333333], "iscrowd": 0, "area": 15653.36845, "rle": {"size": [480, 640], "counts": "]e]4<b>7H8I6I5L4E:G:K4M3L5L3M4L3N1O2N2N101N1O2N2N101O0O2O001O1N101O0O2O000O101O0O101O0O1O101N100O1O2O0000001O00001O0000001O0000001O01O0001O0001N10000000001O000000001O00000000001O000000001O00001O000O2O0000001N10001O00001N1000001O0O101O00001O0O101O001N101O0O101N101O0O101O0O2O0O101O1N101O1N2O1N2O1M2N3M3N2M3M4L4M3L4L4M3L4L4M3M4L3K7G9Hd^a2"}, "label": "bananas on top of a cake"}]}
{"id": 262180, "image": "COCO_train2014_000000262180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a caramel doused banana on a chocolate cake\"."}], "task": "refering", "answer_template": "The \"a caramel doused banana on a chocolate cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131], "bbox": [0.47131249999999997, 0.1000625, 0.728046875, 0.3534583333333333], "iscrowd": 0, "area": 15653.36845, "rle": {"size": [480, 640], "counts": "]e]4<b>7H8I6I5L4E:G:K4M3L5L3M4L3N1O2N2N101N1O2N2N101O0O2O001O1N101O0O2O000O101O0O101O0O1O101N100O1O2O0000001O00001O0000001O0000001O01O0001O0001N10000000001O000000001O00000000001O000000001O00001O000O2O0000001N10001O00001N1000001O0O101O00001O0O101O001N101O0O101N101O0O101O0O2O0O101O1N101O1N2O1N2O1M2N3M3N2M3M4L4M3L4L4M3L4L4M3M4L3K7G9Hd^a2"}, "label": "a caramel doused banana on a chocolate cake"}]}
{"id": 417832, "image": "COCO_train2014_000000417832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teddy bear in the middle\"."}], "task": "refering", "answer_template": "The \"the teddy bear in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 215, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 263], "bbox": [0.3350233644859813, 0.470828125, 0.6459345794392523, 0.778109375], "iscrowd": 0, "area": 17171.877549999994, "rle": {"size": [640, 428], "counts": "U^j26ec07J5I8H7cNUO]_OS1_`0TOZ_Oo0d`0UOW_Oo0e`0UOW_On0g`0VOT_Om0j`0VOR_Om0m`0VOn^Ol0Qa0R101N100]NZMlAg2m=`MQBa2m=bMRB_2l=dMRB\\2l=gMRBZ2l=iMSBX2j=WNhAj1W>ZNfAf1X>_NeAb1X>cNfA\\1Y>hNdAY1Z>kNdAU1Y>POdAQ1Z>ROdAo0[>\\2O1O1N3N2N1O2N00001O00000O110O0001O010O01O010O01O01OO2O0O1O2OO01O01O00010O1O2O1O001O1O001O00000000O100000000hM[BUNe=i1`BTNa=h1fBTNZ=i1lBRNW=k1lBRNV=l1mBQNV=l1mBPNV=o1lBnMW=o1lBnMV=Q2mBkMV=R2mBkMU=T2mBiMV=U2lBgMW=X2jBfMY=X2iBeMY=Z2hBdM[=[2eBcM^=[2dBaM_=^2bB^Mc=`2P2N1O2N101N2N1O2O0N3M3M3M3M3M3M2N2J7G8H8M4N100O2N100O6J7J_en2"}, "label": "the teddy bear in the middle"}]}
{"id": 417832, "image": "COCO_train2014_000000417832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"doll which is in the middle\"."}], "task": "refering", "answer_template": "The \"doll which is in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 215, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 263], "bbox": [0.3350233644859813, 0.470828125, 0.6459345794392523, 0.778109375], "iscrowd": 0, "area": 17171.877549999994, "rle": {"size": [640, 428], "counts": "U^j26ec07J5I8H7cNUO]_OS1_`0TOZ_Oo0d`0UOW_Oo0e`0UOW_On0g`0VOT_Om0j`0VOR_Om0m`0VOn^Ol0Qa0R101N100]NZMlAg2m=`MQBa2m=bMRB_2l=dMRB\\2l=gMRBZ2l=iMSBX2j=WNhAj1W>ZNfAf1X>_NeAb1X>cNfA\\1Y>hNdAY1Z>kNdAU1Y>POdAQ1Z>ROdAo0[>\\2O1O1N3N2N1O2N00001O00000O110O0001O010O01O010O01O01OO2O0O1O2OO01O01O00010O1O2O1O001O1O001O00000000O100000000hM[BUNe=i1`BTNa=h1fBTNZ=i1lBRNW=k1lBRNV=l1mBQNV=l1mBPNV=o1lBnMW=o1lBnMV=Q2mBkMV=R2mBkMU=T2mBiMV=U2lBgMW=X2jBfMY=X2iBeMY=Z2hBdM[=[2eBcM^=[2dBaM_=^2bB^Mc=`2P2N1O2N101N2N1O2O0N3M3M3M3M3M3M2N2J7G8H8M4N100O2N100O6J7J_en2"}, "label": "doll which is in the middle"}]}
{"id": 417832, "image": "COCO_train2014_000000417832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan teddy bear with a red hat\"."}], "task": "refering", "answer_template": "The \"tan teddy bear with a red hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 167, 168, 169, 170, 182, 183, 184, 185, 197, 198, 199, 200, 212, 213, 214, 215, 227, 228, 229, 230, 242, 243, 244, 245, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 287, 288, 289, 290, 291], "bbox": [0.13464953271028038, 0.44378124999999996, 0.46282710280373834, 0.8657187499999999], "iscrowd": 0, "area": 23696.5682, "rle": {"size": [640, 428], "counts": "YeT16ec08H7H9H7X_OoNX>X1dAkNV>]1dAhNY>[1bAiN[>[1aAiN\\>[1^AiN`>Y1]AkN`>X1[AlNc>V1ZAmNd>U1XAPOd>T1WAPOg>R1VAROg>P1UATOZ=]NcCb2oNUOX=_NfC^2POUOT=eNgCZ2ROSOR=jNhCU2TOROP=oNiCQ2VOQOl<UOjCm1WOPOm<WOgCm3X<WL\\CS4b<f101N2N1O2N101N2N1O2O0O2N2N101N1O100O1O100O1O100O1O010O1O100O1O100O1O1000000000O010000000001O001O0O101O00001O01O01O010OfL`DTN`;i1iDQNX;k1QEPNn:m1WEQNj:j1ZEVNf:g1^EYNa:d1cE[N^:f0_DYMV1R2Z:c0iDQMQ1[2V:c0PEiLn0c2S:a0hF^OX9`0jF@V9?lF_OV9>mFAS9>nFBS9<oFCQ9iNaDO`2W1P9eNoDHS2b1o8bNTEHn1e1P9_NVEJl1f1o8\\NZEKh1i1o8XN]EMe1k1o8VN_ELc1n1o8TNaEKb1P2n8RNcEL`1R2n8PNeEK^1T2P9nMdEL]1V2P9kMQIU2P7iMQIW2]<O0O101O1O1N3N1O1N2O1O1O2M2O1O1N2O1O2N1N5K4J6J6J6I8I6JiW_4"}, "label": "tan teddy bear with a red hat"}]}
{"id": 417832, "image": "COCO_train2014_000000417832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear has ' smile ' written on its shirt\"."}], "task": "refering", "answer_template": "The \"a bear has ' smile ' written on its shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 167, 168, 169, 170, 182, 183, 184, 185, 197, 198, 199, 200, 212, 213, 214, 215, 227, 228, 229, 230, 242, 243, 244, 245, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 287, 288, 289, 290, 291], "bbox": [0.13464953271028038, 0.44378124999999996, 0.46282710280373834, 0.8657187499999999], "iscrowd": 0, "area": 23696.5682, "rle": {"size": [640, 428], "counts": "YeT16ec08H7H9H7X_OoNX>X1dAkNV>]1dAhNY>[1bAiN[>[1aAiN\\>[1^AiN`>Y1]AkN`>X1[AlNc>V1ZAmNd>U1XAPOd>T1WAPOg>R1VAROg>P1UATOZ=]NcCb2oNUOX=_NfC^2POUOT=eNgCZ2ROSOR=jNhCU2TOROP=oNiCQ2VOQOl<UOjCm1WOPOm<WOgCm3X<WL\\CS4b<f101N2N1O2N101N2N1O2O0O2N2N101N1O100O1O100O1O100O1O010O1O100O1O100O1O1000000000O010000000001O001O0O101O00001O01O01O010OfL`DTN`;i1iDQNX;k1QEPNn:m1WEQNj:j1ZEVNf:g1^EYNa:d1cE[N^:f0_DYMV1R2Z:c0iDQMQ1[2V:c0PEiLn0c2S:a0hF^OX9`0jF@V9?lF_OV9>mFAS9>nFBS9<oFCQ9iNaDO`2W1P9eNoDHS2b1o8bNTEHn1e1P9_NVEJl1f1o8\\NZEKh1i1o8XN]EMe1k1o8VN_ELc1n1o8TNaEKb1P2n8RNcEL`1R2n8PNeEK^1T2P9nMdEL]1V2P9kMQIU2P7iMQIW2]<O0O101O1O1N3N1O1N2O1O1O2M2O1O1N2O1O2N1N5K4J6J6J6I8I6JiW_4"}, "label": "a bear has ' smile ' written on its shirt"}]}
{"id": 417832, "image": "COCO_train2014_000000417832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chocolate brown bear with a red shirt\"."}], "task": "refering", "answer_template": "The \"chocolate brown bear with a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 161, 173, 174, 175, 176, 177, 189, 190, 191, 192, 204, 205, 206, 207, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 296, 297, 298, 307, 308, 309, 312, 313], "bbox": [0.44455607476635517, 0.46057812499999995, 0.9295327102803739, 0.90428125], "iscrowd": 0, "area": 30464.194399999993, "rle": {"size": [640, 428], "counts": "_Wg31Pd03L3N2M4L3S]OJka08j]O3Sb0P1N1N3M3M1N2O1N2O2M2N2O1N3N1N2O1N2000O1O00100O1O00100O1OO1O1O2iMgMdBZ2T=PNiBQ2o<ZNmBg1R=\\NjBg1S=^NiBc1V=`NfBa1Z=cNaB_1^=dN^B]1b=gNYB[1f=gNWB[1h=fNWBZ1i=hNTBZ1k=hNRBY1m=jNPBX1o=iNoAX1Q>jNmAW1S>jNjAX1U>iNiAX1W>jNfAX1Z>iNcAX1]>V2O10OUMaJVG`5i8eJRG[5o8jJlFV5T9kJjFU5W9lJgFT5Z9lJfFR5\\9oJbFQ5_9PK`Fo4a9QK^Fo4c9RKZFo4f9SKWFm4k9SKTFm4m9TKPFm4Q:TKlEm4U:SKPEg5Q;ZJmDe5U;\\JiDd5X;\\JgDd5Z;]JdDc5];^JaDa5a;`J]D`5d;bJYD^5h;cJVD]5k;eJRDo3[=SLbBn3_=SL^Bn3c=SLZBn3f=V11O00100O000000O1O1O100O1O100O1O10000000000O10000000000O01000000000001OcBjIP=f6M2O2M3M2N3N1N3M3N1N2N2O1O2N1O1O000000O10N2N1O2NSMRDkMl;T2WDkMh;JTDD7a0d;IZDA5e0b;F_DA1h0`;DfD^ONm0[;ClD\\OKP1Y;AREZOIT1T;AWEVOHX1R;CWEPOJ\\1o:FXEhNNa1i:H[EbNOe1f:K\\E[N1i1d:M\\ETN5m1_:0_EmM6Q2\\:3_EgM:S2X:7[GGf87^GFc88`GFa88cGE_88dGF]88gGEZ8:hGDY8:kGCV8;nGBT8;nGDX85jGJ[80fG0`8IaG7e8B\\G>j8[OWGe0R>0001N10000000001O000000001O1O2N2N1O2N1M4L3M4L3M4L4L3M4L3MTWb0"}, "label": "chocolate brown bear with a red shirt"}]}
{"id": 417832, "image": "COCO_train2014_000000417832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown teddy bear\"."}], "task": "refering", "answer_template": "The \"brown teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 161, 173, 174, 175, 176, 177, 189, 190, 191, 192, 204, 205, 206, 207, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 296, 297, 298, 307, 308, 309, 312, 313], "bbox": [0.44455607476635517, 0.46057812499999995, 0.9295327102803739, 0.90428125], "iscrowd": 0, "area": 30464.194399999993, "rle": {"size": [640, 428], "counts": "_Wg31Pd03L3N2M4L3S]OJka08j]O3Sb0P1N1N3M3M1N2O1N2O2M2N2O1N3N1N2O1N2000O1O00100O1O00100O1OO1O1O2iMgMdBZ2T=PNiBQ2o<ZNmBg1R=\\NjBg1S=^NiBc1V=`NfBa1Z=cNaB_1^=dN^B]1b=gNYB[1f=gNWB[1h=fNWBZ1i=hNTBZ1k=hNRBY1m=jNPBX1o=iNoAX1Q>jNmAW1S>jNjAX1U>iNiAX1W>jNfAX1Z>iNcAX1]>V2O10OUMaJVG`5i8eJRG[5o8jJlFV5T9kJjFU5W9lJgFT5Z9lJfFR5\\9oJbFQ5_9PK`Fo4a9QK^Fo4c9RKZFo4f9SKWFm4k9SKTFm4m9TKPFm4Q:TKlEm4U:SKPEg5Q;ZJmDe5U;\\JiDd5X;\\JgDd5Z;]JdDc5];^JaDa5a;`J]D`5d;bJYD^5h;cJVD]5k;eJRDo3[=SLbBn3_=SL^Bn3c=SLZBn3f=V11O00100O000000O1O1O100O1O100O1O10000000000O10000000000O01000000000001OcBjIP=f6M2O2M3M2N3N1N3M3N1N2N2O1O2N1O1O000000O10N2N1O2NSMRDkMl;T2WDkMh;JTDD7a0d;IZDA5e0b;F_DA1h0`;DfD^ONm0[;ClD\\OKP1Y;AREZOIT1T;AWEVOHX1R;CWEPOJ\\1o:FXEhNNa1i:H[EbNOe1f:K\\E[N1i1d:M\\ETN5m1_:0_EmM6Q2\\:3_EgM:S2X:7[GGf87^GFc88`GFa88cGE_88dGF]88gGEZ8:hGDY8:kGCV8;nGBT8;nGDX85jGJ[80fG0`8IaG7e8B\\G>j8[OWGe0R>0001N10000000001O000000001O1O2N2N1O2N1M4L3M4L3M4L4L3M4L3MTWb0"}, "label": "brown teddy bear"}]}
{"id": 417834, "image": "COCO_train2014_000000417834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sofa with dark colour pillow\"."}], "task": "refering", "answer_template": "The \"a white sofa with dark colour pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5580625, 0.4172083333333333, 0.9983906250000001, 0.98925], "iscrowd": 0, "area": 63666.63525, "rle": {"size": [480, 640], "counts": "`fW5<^=W1hNX1TOm0L3N2M3O2N1O1O1O1N3N1O1O1O1O2N1N2O1O1O2N1O1O1O0O1000001O00001O2M2O1O1O2N1O1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O2N1N2O1O1O2N1O1O1O2M2O1O1O2N1O1O1O2M2O1O1O1O00000O100000000000000O100000000000000O1000000000000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O1000000O10000O10000O1000000O10000O10000000000000000000000000000000000000000000000000000000000O1O100O100O1O100N2L4L4L400O1O100O1O100O1O100O1O100000000000000O10000001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001OP8"}, "label": "a white sofa with dark colour pillow"}]}
{"id": 417834, "image": "COCO_train2014_000000417834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch with a dark , furry , pillow\"."}], "task": "refering", "answer_template": "The \"a white couch with a dark , furry , pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5580625, 0.4172083333333333, 0.9983906250000001, 0.98925], "iscrowd": 0, "area": 63666.63525, "rle": {"size": [480, 640], "counts": "`fW5<^=W1hNX1TOm0L3N2M3O2N1O1O1O1N3N1O1O1O1O2N1N2O1O1O2N1O1O1O0O1000001O00001O2M2O1O1O2N1O1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O2N1N2O1O1O2N1O1O1O2M2O1O1O2N1O1O1O2M2O1O1O1O00000O100000000000000O100000000000000O1000000000000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O1000000O10000O10000O1000000O10000O10000000000000000000000000000000000000000000000000000000000O1O100O100O1O100N2L4L4L400O1O100O1O100O1O100O1O100000000000000O10000001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001OP8"}, "label": "a white couch with a dark , furry , pillow"}]}
{"id": 573485, "image": "COCO_train2014_000000573485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white color van is running on the road\"."}], "task": "refering", "answer_template": "The \"a white color van is running on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 299], "bbox": [0.0, 0.42716666666666664, 0.316671875, 0.7728333333333333], "iscrowd": 0, "area": 26764.561849999995, "rle": {"size": [480, 640], "counts": "c9e1o;\\1iNX1N2O1N2O1N101N2O1N2O1N101N100O100O1O100O010O100O10000000000000O100O1O1O1O1O1O1O001O1O1O1O100O1O1000000O100000000O1000000O1000000O10000000O01000000O100000000O1000000O1000000O100000000O1000000O1000O10O100000000O1000000O1000000O100000000O1000000O1000000O10O100000O1000000O100000000O1000000O1000000O100000000O1000000O0100000O100000000O1000000O1000000O100000000O1000000O100000O0100000000L4C=D<C=D<C=D<C=D<CPe\\6"}, "label": "a white color van is running on the road"}]}
{"id": 573485, "image": "COCO_train2014_000000573485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white van in road\"."}], "task": "refering", "answer_template": "The \"white van in road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 299], "bbox": [0.0, 0.42716666666666664, 0.316671875, 0.7728333333333333], "iscrowd": 0, "area": 26764.561849999995, "rle": {"size": [480, 640], "counts": "c9e1o;\\1iNX1N2O1N2O1N101N2O1N2O1N101N100O100O1O100O010O100O10000000000000O100O1O1O1O1O1O1O001O1O1O1O100O1O1000000O100000000O1000000O1000000O10000000O01000000O100000000O1000000O1000000O100000000O1000000O1000O10O100000000O1000000O1000000O100000000O1000000O1000000O10O100000O1000000O100000000O1000000O1000000O100000000O1000000O0100000O100000000O1000000O1000000O100000000O1000000O100000O0100000000L4C=D<C=D<C=D<C=D<CPe\\6"}, "label": "white van in road"}]}
{"id": 229422, "image": "COCO_train2014_000000229422.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with the number 24 on him wearing a blue jacket and orange pants\"."}], "task": "refering", "answer_template": "The \"a man with the number 24 on him wearing a blue jacket and orange pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 130, 131, 132, 150, 151, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 242, 243, 244, 245, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312], "bbox": [0.492578125, 0.2289156626506024, 0.791796875, 0.6616523235800343], "iscrowd": 0, "area": 18295.146500000003, "rle": {"size": [581, 640], "counts": "^Rc51Rb05L3L4M4K4M4K2N2O1N2O0O2O1N2N2O0O2O1N2O1N1O2O1N2O1N1iK\\NmGe1T8\\NjGc1X8^NeGc1[8_NcGa1^8`N_G`1c8aN[G_1e8cNXG]1j8dNSG]1n8cNQG]1o8eNnF[1PM\\Nb;:\\GZ1QM^Nc;8ZGZ1TM`N`;7[GY1TMcN`;4ZGY1VMfN_;1YGZ1XMfN^;1YGY1XMiN^;NXGY1ZMlN];KXGY1[MnN[;JXGX1]MPO[;HWGX1]MSO[;EVGY1_MTOY;DVGX1aMVOY;CTGW1bMYOY;@SGW1eMYOX;ARGV1eMZOZ;@oFW1fMZO\\;_OmFV1hMZO\\;AjFV1iMZO^;@gFV1kMZO`;@dFV1lMZO`;AbFV1mMYOc;A_FU1nM[Od;@\\FV1PNZOe;@ZFU1QN[Og;@VFV1RN[Oi;_OSFW1TNYOj;BPFT1VN[Ok;AmEU1WN[Om;@[EZOPOj1I[On;AXE_OkNf1N[OP<@VEDfNb13ZOS<@SEc1kNmNS<@QEc1kNnNU<_OoDb1lNoNV<@mDa1mNoNW<@kDa1mNoNZ<@gDa1oNPO[<_OeDa1POPO\\<_OcD`1QOQO^<@_D_1RORO`<_O]D^1TOROa<@ZD^1TOSOc<_OXD^1TOTOd<_OWD\\1VOTOe<@TD\\1VOUOg<_ORD[1WOVOi<_OoC[1XOVOj<_OmC[1XOWOl<^OkCZ1YOXOn<^OhCZ1ZOXOo<^OfCY1[OYOP=_OcCY1\\OYOR=^OaCY1]OYOS=_O^CW1_OZOU=_O[Ce2f<[MYCV1A[OX=_OVCd2k<\\MTCd2m<\\MRCc2o<^MPCb2Q=^MnBa2T=_MkBa2V=_MiBa2X=_MgB`2[=`MdB`2]=`MbB_2`=bM]B_2d=aM[B_2e=bMZB^2g=o01O1N2O1N2N2N101N2N2O1N2N2N2O0O2N2N100O1O010O1O00100O001O10O01O100O1O2N100O2N1O101N1O2OjATMT=k2mBWMQ=h2oB\\Mo<e2nB_Mo<b2oBaMo<`2oBdMo<]2nBfMP=[2nBiMQ=W2mBlMS=R2kBoMW=P2gBQN[=o1bBRNa=l1]BUNe=j1YBWNi=i1TBWNP>g1oAYNS>k23M5K7H9H8Hg0YO3M3L5L3M3M4L3M3L5L3M3M3M4K4M3M4L3M3L5L3M[W[2"}, "label": "a man with the number 24 on him wearing a blue jacket and orange pants"}]}
{"id": 229422, "image": "COCO_train2014_000000229422.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue jacket and orange pants\"."}], "task": "refering", "answer_template": "The \"a man in a blue jacket and orange pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 130, 131, 132, 150, 151, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 242, 243, 244, 245, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312], "bbox": [0.492578125, 0.2289156626506024, 0.791796875, 0.6616523235800343], "iscrowd": 0, "area": 18295.146500000003, "rle": {"size": [581, 640], "counts": "^Rc51Rb05L3L4M4K4M4K2N2O1N2O0O2O1N2N2O0O2O1N2O1N1O2O1N2O1N1iK\\NmGe1T8\\NjGc1X8^NeGc1[8_NcGa1^8`N_G`1c8aN[G_1e8cNXG]1j8dNSG]1n8cNQG]1o8eNnF[1PM\\Nb;:\\GZ1QM^Nc;8ZGZ1TM`N`;7[GY1TMcN`;4ZGY1VMfN_;1YGZ1XMfN^;1YGY1XMiN^;NXGY1ZMlN];KXGY1[MnN[;JXGX1]MPO[;HWGX1]MSO[;EVGY1_MTOY;DVGX1aMVOY;CTGW1bMYOY;@SGW1eMYOX;ARGV1eMZOZ;@oFW1fMZO\\;_OmFV1hMZO\\;AjFV1iMZO^;@gFV1kMZO`;@dFV1lMZO`;AbFV1mMYOc;A_FU1nM[Od;@\\FV1PNZOe;@ZFU1QN[Og;@VFV1RN[Oi;_OSFW1TNYOj;BPFT1VN[Ok;AmEU1WN[Om;@[EZOPOj1I[On;AXE_OkNf1N[OP<@VEDfNb13ZOS<@SEc1kNmNS<@QEc1kNnNU<_OoDb1lNoNV<@mDa1mNoNW<@kDa1mNoNZ<@gDa1oNPO[<_OeDa1POPO\\<_OcD`1QOQO^<@_D_1RORO`<_O]D^1TOROa<@ZD^1TOSOc<_OXD^1TOTOd<_OWD\\1VOTOe<@TD\\1VOUOg<_ORD[1WOVOi<_OoC[1XOVOj<_OmC[1XOWOl<^OkCZ1YOXOn<^OhCZ1ZOXOo<^OfCY1[OYOP=_OcCY1\\OYOR=^OaCY1]OYOS=_O^CW1_OZOU=_O[Ce2f<[MYCV1A[OX=_OVCd2k<\\MTCd2m<\\MRCc2o<^MPCb2Q=^MnBa2T=_MkBa2V=_MiBa2X=_MgB`2[=`MdB`2]=`MbB_2`=bM]B_2d=aM[B_2e=bMZB^2g=o01O1N2O1N2N2N101N2N2O1N2N2N2O0O2N2N100O1O010O1O00100O001O10O01O100O1O2N100O2N1O101N1O2OjATMT=k2mBWMQ=h2oB\\Mo<e2nB_Mo<b2oBaMo<`2oBdMo<]2nBfMP=[2nBiMQ=W2mBlMS=R2kBoMW=P2gBQN[=o1bBRNa=l1]BUNe=j1YBWNi=i1TBWNP>g1oAYNS>k23M5K7H9H8Hg0YO3M3L5L3M3M4L3M3L5L3M3M3M4K4M3M4L3M3L5L3M[W[2"}, "label": "a man in a blue jacket and orange pants"}]}
{"id": 229422, "image": "COCO_train2014_000000229422.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier with a yellow and black ski outfit ,\"."}], "task": "refering", "answer_template": "The \"a skier with a yellow and black ski outfit ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 187, 188, 189, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 328, 329, 348, 349, 351, 352, 371, 372, 374, 375, 394, 395, 397, 398], "bbox": [0.0874375, 0.3542512908777969, 0.3424375, 0.8643545611015491], "iscrowd": 0, "area": 21990.8402, "rle": {"size": [581, 640], "counts": "`RP13Qb06J6J6J6J5J2O2N1O2N1O1O2N1O2M2O1O2N1O2N1O1O2N1N3N1O1O2N1O2N1O1N3N1O1OM3B>A?00O1OTBgNn:Z1oDROh:m0VE@`:b0\\EI[:<]E0Z:5^E8_9e0[FF[9`0]FMX99aF2V93bF9T9MdF`0R9DhFg0n8_OjFk0n8[OjFi0U9\\OdFf0[9S4O1N1O2O0O2O1N1O2O1O02O0O2O1O0O2O1O0O2O1N101\\MiFXLX9d3WGoKk8n3cGfK]8W4RH]Ko7`4_HSKc7j4kHjJU7S5ZI`Jh6]5R3L;Fh0XO3L3N2M4M2N3L3N3L3N3001N2O1O1N2O1O1N2OYMdAe1[>[NgAd1W>\\NlAc1S>]NoAb1P>]NSBb1k=_NWB`1h=_N[B_1e=aN\\B_1b=bN`B]1_=bNdBa1W=_NkBd1o<\\NTCh1f<XN\\Ck1_<UNbCP2W<PNlCT2n;lMTDW2g;hM\\D\\2];eMeD_2U;`MnDc2m:]MUEg2d:ZM]Ek2]:TMfEo2U:QMmEP3o9PMTFl2n9TMRFk2o9UMRFi2n9WMSFh2n9XM_ESO\\Oc3V;YMTE^OGX3U;[MiDG4l2T;\\M^D2?a2T;cNnD[1R;eNoDZ1R;eNPEY1Q;fNQEX1o:hNREW1o:hNSEV1n:iNSEU1n:kNWEP1j:oN[El0f:SO_Eh0a:WOeEd0\\:UOoEf0R:SOZFg0f9SOeFh0\\9QOoFj0R9oNYGl0g8mNeGn0\\8kNoGP1R8iNYHR1g7hNeHQ1]<K5K5K5K5K5Ki\\^7"}, "label": "a skier with a yellow and black ski outfit ,"}]}
{"id": 229422, "image": "COCO_train2014_000000229422.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"snowboarder in yellow , far left\"."}], "task": "refering", "answer_template": "The \"snowboarder in yellow , far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 187, 188, 189, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 328, 329, 348, 349, 351, 352, 371, 372, 374, 375, 394, 395, 397, 398], "bbox": [0.0874375, 0.3542512908777969, 0.3424375, 0.8643545611015491], "iscrowd": 0, "area": 21990.8402, "rle": {"size": [581, 640], "counts": "`RP13Qb06J6J6J6J5J2O2N1O2N1O1O2N1O2M2O1O2N1O2N1O1O2N1N3N1O1O2N1O2N1O1N3N1O1OM3B>A?00O1OTBgNn:Z1oDROh:m0VE@`:b0\\EI[:<]E0Z:5^E8_9e0[FF[9`0]FMX99aF2V93bF9T9MdF`0R9DhFg0n8_OjFk0n8[OjFi0U9\\OdFf0[9S4O1N1O2O0O2O1N1O2O1O02O0O2O1O0O2O1O0O2O1N101\\MiFXLX9d3WGoKk8n3cGfK]8W4RH]Ko7`4_HSKc7j4kHjJU7S5ZI`Jh6]5R3L;Fh0XO3L3N2M4M2N3L3N3L3N3001N2O1O1N2O1O1N2OYMdAe1[>[NgAd1W>\\NlAc1S>]NoAb1P>]NSBb1k=_NWB`1h=_N[B_1e=aN\\B_1b=bN`B]1_=bNdBa1W=_NkBd1o<\\NTCh1f<XN\\Ck1_<UNbCP2W<PNlCT2n;lMTDW2g;hM\\D\\2];eMeD_2U;`MnDc2m:]MUEg2d:ZM]Ek2]:TMfEo2U:QMmEP3o9PMTFl2n9TMRFk2o9UMRFi2n9WMSFh2n9XM_ESO\\Oc3V;YMTE^OGX3U;[MiDG4l2T;\\M^D2?a2T;cNnD[1R;eNoDZ1R;eNPEY1Q;fNQEX1o:hNREW1o:hNSEV1n:iNSEU1n:kNWEP1j:oN[El0f:SO_Eh0a:WOeEd0\\:UOoEf0R:SOZFg0f9SOeFh0\\9QOoFj0R9oNYGl0g8mNeGn0\\8kNoGP1R8iNYHR1g7hNeHQ1]<K5K5K5K5K5Ki\\^7"}, "label": "snowboarder in yellow , far left"}]}
{"id": 499760, "image": "COCO_train2014_000000499760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with the white beard\"."}], "task": "refering", "answer_template": "The \"man with the white beard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 161, 162, 184, 185, 186, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 254, 255, 257, 258, 277, 278, 280, 281, 301], "bbox": [0.0, 0.3185087719298246, 0.2555, 0.6576140350877193], "iscrowd": 0, "area": 11937.487199999998, "rle": {"size": [570, 640], "counts": "i5o3j=101O0O2O1N101O001OO2J50001N101O3M3M2N3L4M4L4L2N2M4M2N1O1O2N0O2O1O1O1O1N2O02O2N2M4M2N3L4N2N2M4K5L3M1O1N101O1N101N2O0O2N2O0O_MXBQ1h=nN\\Bn0d=RO_Bl0`=TOcBi0^=VOfBf0[=YOhBd0X=\\OlBa0T=_OnB>R=BQC;P=ESC7n<ITC5k<KYC1h<OZCMg<4[CIf<7]CEc<<_CAb<?aC\\Oa<d0aCYO_<h0jCnNW<R1kCkNV<U1lCgNU<Z1mCcNT<\\1V20O010O010O00100O0010O10O010000O100O010O10000O100O010O100O11O0000010O003M4L4L4L1OO1O2N1N2O1O1O1O1O1O2N1O2O1N2N2N2N2N1O2N2N2N2N2O1N2N2N2N1O2N1O1O2N1O2N1O1O2MhUY8"}, "label": "man with the white beard"}]}
{"id": 499760, "image": "COCO_train2014_000000499760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couple sitting on bench . man holding a hat\"."}], "task": "refering", "answer_template": "The \"couple sitting on bench . man holding a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 161, 162, 184, 185, 186, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 254, 255, 257, 258, 277, 278, 280, 281, 301], "bbox": [0.0, 0.3185087719298246, 0.2555, 0.6576140350877193], "iscrowd": 0, "area": 11937.487199999998, "rle": {"size": [570, 640], "counts": "i5o3j=101O0O2O1N101O001OO2J50001N101O3M3M2N3L4M4L4L2N2M4M2N1O1O2N0O2O1O1O1O1N2O02O2N2M4M2N3L4N2N2M4K5L3M1O1N101O1N101N2O0O2N2O0O_MXBQ1h=nN\\Bn0d=RO_Bl0`=TOcBi0^=VOfBf0[=YOhBd0X=\\OlBa0T=_OnB>R=BQC;P=ESC7n<ITC5k<KYC1h<OZCMg<4[CIf<7]CEc<<_CAb<?aC\\Oa<d0aCYO_<h0jCnNW<R1kCkNV<U1lCgNU<Z1mCcNT<\\1V20O010O010O00100O0010O10O010000O100O010O10000O100O010O100O11O0000010O003M4L4L4L1OO1O2N1N2O1O1O1O1O1O2N1O2O1N2N2N2N2N1O2N2N2N2N2O1N2N2N2N1O2N1O1O2N1O2N1O1O2MhUY8"}, "label": "couple sitting on bench . man holding a hat"}]}
{"id": 499760, "image": "COCO_train2014_000000499760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black jacket sitting on a park bench\"."}], "task": "refering", "answer_template": "The \"a woman in a black jacket sitting on a park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 163, 164, 165, 186, 187, 188, 189, 210, 211, 212, 234, 235, 236, 258, 259], "bbox": [0.095046875, 0.29882456140350877, 0.30451562499999996, 0.588701754385965], "iscrowd": 0, "area": 7735.140049999997, "rle": {"size": [570, 640], "counts": "\\UR14]`0>Y@F_?a0^@D]?`0^@E_??\\@Db?>[@Eb?>[@Fb?[1N2N1O00000000000001O00001O00100O1O1O1O2N1O1O101O010001N100O0O10O10O10O001O100O1O010O10O101N100O2O1O1O0O0100O100000001O01O000100O2OO1O100O1O1O010000000O1O1O1O1002M0100O1O1O1O0000O2O001M2O2N2N2N2M3N2N3L3K5L4M9I4K100O1O1O1O0[OR_O>o`0_OU_O?Ta0O1O1O1O1O1O1O1O001O1N1O2N2O]\\g7"}, "label": "a woman in a black jacket sitting on a park bench"}]}
{"id": 499760, "image": "COCO_train2014_000000499760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with coat and sunglasses , sitting on bench by white bag\"."}], "task": "refering", "answer_template": "The \"woman with coat and sunglasses , sitting on bench by white bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 163, 164, 165, 186, 187, 188, 189, 210, 211, 212, 234, 235, 236, 258, 259], "bbox": [0.095046875, 0.29882456140350877, 0.30451562499999996, 0.588701754385965], "iscrowd": 0, "area": 7735.140049999997, "rle": {"size": [570, 640], "counts": "\\UR14]`0>Y@F_?a0^@D]?`0^@E_??\\@Db?>[@Eb?>[@Fb?[1N2N1O00000000000001O00001O00100O1O1O1O2N1O1O101O010001N100O0O10O10O10O001O100O1O010O10O101N100O2O1O1O0O0100O100000001O01O000100O2OO1O100O1O1O010000000O1O1O1O1002M0100O1O1O1O0000O2O001M2O2N2N2N2M3N2N3L3K5L4M9I4K100O1O1O1O0[OR_O>o`0_OU_O?Ta0O1O1O1O1O1O1O1O001O1N1O2N2O]\\g7"}, "label": "woman with coat and sunglasses , sitting on bench by white bag"}]}
{"id": 499760, "image": "COCO_train2014_000000499760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"steel bench in road\"."}], "task": "refering", "answer_template": "The \"steel bench in road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 146, 147, 148, 150, 151, 152, 169, 170, 171, 172, 173, 193, 194, 216, 217, 218], "bbox": [0.349125, 0.2599122807017544, 0.62309375, 0.4992280701754386], "iscrowd": 0, "area": 7259.038100000002, "rle": {"size": [570, 640], "counts": "ikl32ha01N201N2O1N2O0O2M3N2N2M3N2N2N2M3N2N2N2M3N2N2M3N3M3M2M4M3M3M3L3N4ZAeMe=_2VBaMj=e2oA]MQ>V30000O0100000O10BcA^M^>Z2mAcMR>]2SB_Mm=_2h0O1O010O1000O01000O10O1000O101O001NIo@gMQ?V2RAkMm>Q2WAPNg>m1^ASNa>k1aAWN]>j1cAVN\\>k1cAVN[>m1dATNZ>l1fAUNY>j1hAXNV>h1kAXNS>b1[AXNc07P>`1bAUN?;n=`1\\B`NVODU>3nA^1h0kNSOFV>OVB[1c0TOW=@\\BV1`0XOo=c0RB\\OP>b0RB[OQ>b0QB]OQ>a0QB\\OR>b0oA]OS>`0oA^OT>`0mA^OV>`0kA_OW>>e1N2N2M01000O10000O010000O01000O01000O010O10O10O10O10^NJ`A6_>N^A1c>1[AOe>4XALh>6VAIj>:TAFl><RADn>>QAAo>a0T12M3N2N2N1N10O1000O10O1000O10O1000O10002N1N2O00I^_OVOb`0j0`_OTO``0l0b_ORO]`0o07O10000O01000Ob_OSOQ`0m0o_OUOo?k0Q@VOm?k0R@\\Oh?c0Y@]Oi?a0W@_Om?=S@CP`09Q@FT`06k_OKY`01g_OO]`0Mc_O3a`0H`_O7Ra00O11O4K5LdnU4"}, "label": "steel bench in road"}]}
{"id": 499760, "image": "COCO_train2014_000000499760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair behind the dog\"."}], "task": "refering", "answer_template": "The \"chair behind the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 146, 147, 148, 150, 151, 152, 169, 170, 171, 172, 173, 193, 194, 216, 217, 218], "bbox": [0.349125, 0.2599122807017544, 0.62309375, 0.4992280701754386], "iscrowd": 0, "area": 7259.038100000002, "rle": {"size": [570, 640], "counts": "ikl32ha01N201N2O1N2O0O2M3N2N2M3N2N2N2M3N2N2N2M3N2N2M3N3M3M2M4M3M3M3L3N4ZAeMe=_2VBaMj=e2oA]MQ>V30000O0100000O10BcA^M^>Z2mAcMR>]2SB_Mm=_2h0O1O010O1000O01000O10O1000O101O001NIo@gMQ?V2RAkMm>Q2WAPNg>m1^ASNa>k1aAWN]>j1cAVN\\>k1cAVN[>m1dATNZ>l1fAUNY>j1hAXNV>h1kAXNS>b1[AXNc07P>`1bAUN?;n=`1\\B`NVODU>3nA^1h0kNSOFV>OVB[1c0TOW=@\\BV1`0XOo=c0RB\\OP>b0RB[OQ>b0QB]OQ>a0QB\\OR>b0oA]OS>`0oA^OT>`0mA^OV>`0kA_OW>>e1N2N2M01000O10000O010000O01000O01000O010O10O10O10O10^NJ`A6_>N^A1c>1[AOe>4XALh>6VAIj>:TAFl><RADn>>QAAo>a0T12M3N2N2N1N10O1000O10O1000O10O1000O10002N1N2O00I^_OVOb`0j0`_OTO``0l0b_ORO]`0o07O10000O01000Ob_OSOQ`0m0o_OUOo?k0Q@VOm?k0R@\\Oh?c0Y@]Oi?a0W@_Om?=S@CP`09Q@FT`06k_OKY`01g_OO]`0Mc_O3a`0H`_O7Ra00O11O4K5LdnU4"}, "label": "chair behind the dog"}]}
{"id": 131127, "image": "COCO_train2014_000000131127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a man with no shirt and a red headband\"."}], "task": "refering", "answer_template": "The \"the back of a man with no shirt and a red headband\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 115, 138, 139, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 276, 277, 299, 322], "bbox": [0.0, 0.18756440281030445, 0.168171875, 1.0], "iscrowd": 0, "area": 14996.8537, "rle": {"size": [427, 640], "counts": "`2k:a2O001O1O0G:K5L3L5[HlLm4Y3lJnLP5V3kJPMo4T3mJQMn4T3oJoLm4U3PKnLk4W3SKkLh4Z3UKiLf4\\3WKgLd4_3YKcLb4b3[KaL`4d3\\KaL`4d3[K_Lc4f3VK]Li4h3lJ`LR5n5N2N1N3N2N2N1O2M3N2N1O2N2N2N101N2O1N2N101N2N2O1N101N2N2O0O2N2O1N2O0O2N2O1N101N2N2N2M2O2N2M3N1L5L4K5L4O0O2N2N2N2N1O2N2N2N1O2N3M3M2N3N2M2M4M3M2N3M3L3N3M3M3L3N3M3M2M4M3M2N`bm6"}, "label": "the back of a man with no shirt and a red headband"}]}
{"id": 131127, "image": "COCO_train2014_000000131127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s back , and he is wearing a red head band\"."}], "task": "refering", "answer_template": "The \"a man ' s back , and he is wearing a red head band\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 115, 138, 139, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 276, 277, 299, 322], "bbox": [0.0, 0.18756440281030445, 0.168171875, 1.0], "iscrowd": 0, "area": 14996.8537, "rle": {"size": [427, 640], "counts": "`2k:a2O001O1O0G:K5L3L5[HlLm4Y3lJnLP5V3kJPMo4T3mJQMn4T3oJoLm4U3PKnLk4W3SKkLh4Z3UKiLf4\\3WKgLd4_3YKcLb4b3[KaL`4d3\\KaL`4d3[K_Lc4f3VK]Li4h3lJ`LR5n5N2N1N3N2N2N1O2M3N2N1O2N2N2N101N2O1N2N101N2N2O1N101N2N2O0O2N2O1N2O0O2N2O1N101N2N2N2M2O2N2M3N1L5L4K5L4O0O2N2N2N2N1O2N2N2N1O2N3M3M2N3N2M2M4M3M2N3M3L3N3M3M3L3N3M3M2M4M3M2N`bm6"}, "label": "a man ' s back , and he is wearing a red head band"}]}
{"id": 131127, "image": "COCO_train2014_000000131127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding bananas\"."}], "task": "refering", "answer_template": "The \"a man holding bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 120, 121, 123, 124, 125, 126, 127, 128, 142, 143, 144, 147, 148, 149, 150, 151, 165, 166, 167, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.00603125, 0.1348711943793911, 0.911171875, 0.9859016393442623], "iscrowd": 0, "area": 102244.6178, "rle": {"size": [427, 640], "counts": "[R21W=3N2N2M3N2N2VOj0N2O2M2O1N2O1N2N2O1N2O1N2N2O1N2O100O100O1O100O100O100O100O1O100O100O2O0O1O100O100O100O100O1O100O100O100O100O1O100G9O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2K5K5K5VOjJZI[5[6P1I7M3L4L4M3L4M3L4L4O1O1O1O100O1O1O1O100O100O100O100O10000O100O100O10oHnJi5R5TJSKj5m4TJVKk5j4RJ[Kl5e4RJ_Kl5a4QJdKm5\\4QJhKm5X4PJmKn5S4aI`L]6Y5N1O2N2N2N2N2N2N2N2N1O001O0000001O000000001O0000O100O100O100O100O1000000000000000000000000O1O100O100O1O100O100O1O100O100O1O1O100O1O1O1O100O1O1O1O1O1jMVJkLj5h2dJVM]5h2fJVM[5i2gJTM\\5j2fJSM\\5l2fJQM\\5n2fJPM[5o2fJoL\\5P3gJlL[5S3hJiLZ5U3jJgLX5X3^2N2O1O1N2O1N2N2M3N2M3M3M3M3M3M3N22NO1N2O1N2N2N2N2O1N2N2000000O100000000O100000000O1000000O100000000O1000000000000000000000000001O0000000000000000001O000000000000000000001O00000WJmM:S2EoM:Q2EQN:o1DTN;l1DVN;j1DXN;i1BZN=g1A[N>e1A\\N?e1_O]N`0d1]O^Nc0c1[O_Nd0a1[O`Ne0a1XObNg0_1WObNi0_1UObNk0^1SOeNl0\\1ROeNn0\\1POfNo0[1nNgNR1Y1fNPOY1Q1]NXOc1i0SNAl1`0lMGT29jMIV28gMKX26fMKZ26cMM\\23bMO^24^MNa28WMJi2<oLGP3>jLCV3h0]LZOc3R1oKQOP4U501O001O001O00001O001O001O00001O001O001O001O001O1O001O1L4J5K6J6J6O010O100O010O100O00100O100O100O2O0O100O101N100O100O100O00100O010O10O0100O010O010O10O0100O010O1O010O10O01O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O2N1O2N1O2N1OO1000000O1000000O1O1O100O1O1O100O1O1O100O1O1O1O1N2O1N2N1O1O0001O0001O0000000O1N3M2N2N2N2L5[Od0O1000001O0O10000104K4L4L5K4L4LU1kN3M3L5L3M3M4L3M3M4L3M3M4L3M3M4L3M4L5Kh^g0"}, "label": "a man holding bananas"}]}
{"id": 131127, "image": "COCO_train2014_000000131127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy with blue shirt and holding two bananas in each hand\"."}], "task": "refering", "answer_template": "The \"the guy with blue shirt and holding two bananas in each hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 120, 121, 123, 124, 125, 126, 127, 128, 142, 143, 144, 147, 148, 149, 150, 151, 165, 166, 167, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.00603125, 0.1348711943793911, 0.911171875, 0.9859016393442623], "iscrowd": 0, "area": 102244.6178, "rle": {"size": [427, 640], "counts": "[R21W=3N2N2M3N2N2VOj0N2O2M2O1N2O1N2N2O1N2O1N2N2O1N2O100O100O1O100O100O100O100O1O100O100O2O0O1O100O100O100O100O1O100O100O100O100O1O100G9O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2K5K5K5VOjJZI[5[6P1I7M3L4L4M3L4M3L4L4O1O1O1O100O1O1O1O100O100O100O100O10000O100O100O10oHnJi5R5TJSKj5m4TJVKk5j4RJ[Kl5e4RJ_Kl5a4QJdKm5\\4QJhKm5X4PJmKn5S4aI`L]6Y5N1O2N2N2N2N2N2N2N2N1O001O0000001O000000001O0000O100O100O100O100O1000000000000000000000000O1O100O100O1O100O100O1O100O100O1O1O100O1O1O1O100O1O1O1O1O1jMVJkLj5h2dJVM]5h2fJVM[5i2gJTM\\5j2fJSM\\5l2fJQM\\5n2fJPM[5o2fJoL\\5P3gJlL[5S3hJiLZ5U3jJgLX5X3^2N2O1O1N2O1N2N2M3N2M3M3M3M3M3M3N22NO1N2O1N2N2N2N2O1N2N2000000O100000000O100000000O1000000O100000000O1000000000000000000000000001O0000000000000000001O000000000000000000001O00000WJmM:S2EoM:Q2EQN:o1DTN;l1DVN;j1DXN;i1BZN=g1A[N>e1A\\N?e1_O]N`0d1]O^Nc0c1[O_Nd0a1[O`Ne0a1XObNg0_1WObNi0_1UObNk0^1SOeNl0\\1ROeNn0\\1POfNo0[1nNgNR1Y1fNPOY1Q1]NXOc1i0SNAl1`0lMGT29jMIV28gMKX26fMKZ26cMM\\23bMO^24^MNa28WMJi2<oLGP3>jLCV3h0]LZOc3R1oKQOP4U501O001O001O00001O001O001O00001O001O001O001O001O1O001O1L4J5K6J6J6O010O100O010O100O00100O100O100O2O0O100O101N100O100O100O00100O010O10O0100O010O010O10O0100O010O1O010O10O01O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O2N1O2N1O2N1OO1000000O1000000O1O1O100O1O1O100O1O1O100O1O1O1O1N2O1N2N1O1O0001O0001O0000000O1N3M2N2N2N2L5[Od0O1000001O0O10000104K4L4L5K4L4LU1kN3M3L5L3M3M4L3M3M4L3M3M4L3M3M4L3M4L5Kh^g0"}, "label": "the guy with blue shirt and holding two bananas in each hand"}]}
{"id": 131127, "image": "COCO_train2014_000000131127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt behind the man holding bananas\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt behind the man holding bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 269, 270, 271, 272, 293, 294], "bbox": [0.5706562500000001, 0.13580796252927402, 0.8689843750000001, 0.8205152224824356], "iscrowd": 0, "area": 30286.37405, "rle": {"size": [427, 640], "counts": "T^h41X=1O2N3M2N3M2M4M2N3N1O2N101N1O2N101N1O2N101N1O2N101N1O2N101N2N1O2O0O2N1O2O0O2N1O2O0O2N1O2O0O2N2N101N1O2N101N2N2N2O1N2N2N3N1N2N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N101N2M3L4L4L4L4L4ZOf0I7J5K6I7G9G9L4L4L4L4K5M3M3N2N1N3N2M3N2N2M3N2O1N2N2O1N2N101N2N2O1N2N2O1N2O1O1O1O001O1O1O1O1O1O1O1O00O100O100O100O100O1O100O100O100O1O100O100O1PL_KUOb4j0`KSOb4k0aKROa4m0bKnNa4P1bKmN`4Q1cKlN_4S1cKjN_4T1dKiN^4V1dKfN_4X1dKeN^4Y1eKdN]4[1eKbN_4Z1eKbN^4\\1dK`Na4\\1bKaNa4]1aK`Nb4]1aK`Nc4\\1`KaNc4]1_K_Nf4]1]KWNo4g1UKPNR5m1RKjMW5R2nJeMY5Y2jJ_M^5]2gJYMa5e2d2F;C<D=C<D<D=C_nR1"}, "label": "a man in a black shirt behind the man holding bananas"}]}
{"id": 131127, "image": "COCO_train2014_000000131127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mud - covered man looks on from behind at another clutching three bananas\"."}], "task": "refering", "answer_template": "The \"a mud - covered man looks on from behind at another clutching three bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 269, 270, 271, 272, 293, 294], "bbox": [0.5706562500000001, 0.13580796252927402, 0.8689843750000001, 0.8205152224824356], "iscrowd": 0, "area": 30286.37405, "rle": {"size": [427, 640], "counts": "T^h41X=1O2N3M2N3M2M4M2N3N1O2N101N1O2N101N1O2N101N1O2N101N1O2N101N2N1O2O0O2N1O2O0O2N1O2O0O2N1O2O0O2N2N101N1O2N101N2N2N2O1N2N2N3N1N2N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N101N2M3L4L4L4L4L4ZOf0I7J5K6I7G9G9L4L4L4L4K5M3M3N2N1N3N2M3N2N2M3N2O1N2N2O1N2N101N2N2O1N2N2O1N2O1O1O1O001O1O1O1O1O1O1O1O00O100O100O100O100O1O100O100O100O1O100O100O1PL_KUOb4j0`KSOb4k0aKROa4m0bKnNa4P1bKmN`4Q1cKlN_4S1cKjN_4T1dKiN^4V1dKfN_4X1dKeN^4Y1eKdN]4[1eKbN_4Z1eKbN^4\\1dK`Na4\\1bKaNa4]1aK`Nb4]1aK`Nc4\\1`KaNc4]1_K_Nf4]1]KWNo4g1UKPNR5m1RKjMW5R2nJeMY5Y2jJ_M^5]2gJYMa5e2d2F;C<D=C<D<D=C_nR1"}, "label": "a mud - covered man looks on from behind at another clutching three bananas"}]}
{"id": 245817, "image": "COCO_train2014_000000245817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in black shorts\"."}], "task": "refering", "answer_template": "The \"a girl in black shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 17, 18, 19, 20, 21, 30, 31, 32, 33, 34, 35, 43, 44, 45, 46, 47, 48, 55, 56, 57, 58, 59, 60, 66, 67, 68, 69, 70, 71, 72, 73, 78, 79, 80, 81, 82, 83, 84, 85, 86, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 104, 106, 107, 108, 109, 110, 111, 112, 113, 117, 119, 120, 121, 122, 123, 124, 125, 130, 132, 133, 134, 135, 136, 137, 143, 145, 146, 147, 148, 149, 150, 158, 159, 160, 161, 162, 163, 171, 172, 173, 174, 175, 176, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 224, 225, 226, 227, 237, 238, 239, 250, 251, 252, 263, 264, 265, 276, 277, 278, 289, 290], "bbox": [0.0, 0.022625, 0.7290581717451523, 0.9776718750000001], "iscrowd": 0, "area": 78280.38735000003, "rle": {"size": [640, 361], "counts": "R99ea0HY@S1m>WOh@j0W?A^@a0`?JU@7k?3j_ONU`0=`_OD``0m1O101N0100O2N1O2O1N1N3N1O2WOi0M2O2N1N3N2O001O00101N1O2N100O2N1O2N101O00001O0010O100O00100O100O10O01O100O1N2O001O1N2O1O1O0ZAcNi:_1TDKd;6ZCT1]<mNPCm1h<TNUCV2c<jM[C`2^<`M_Cj2Y<WMcCT3V<lLgCY3V<hLgC^3V<bLgCc3V<]LhCh3V<S2M4M2M3N2M4M2M4M4K5L4KT3mLh0WOh0YOa0^O9H9F:G8G:G3L100O1O10O0100O100O10000O10000O1000000O1[Oe0J6J5K6J6M3N2M3N2N1O2N1N2O2N1O2NkMoM^BQ2c=PN\\BP2d=RNYBn1h=TNVBk1k=VNSBj1n=XNPBh1P>YNnAg1S>[NjAe1W>`N`Ad1`>aNWAb1j>cNm@`1T?fNb@^1^?hNX@[1i?]1100O1O100O1O1O100O1O1O1O1O000010O000XNTM]Bl2c=h100000000000002N1O1O1O1O1O1O1O1O2M2O1O1O1O1O1N2O1O2N1O1lLmIZHT6f7PJUHQ6k7SJoGo5Q8UJjGl5W8WJdGj5\\8ZJ_Gg5a8]JZGe5e8_JVGb5j8cJPG^5P9fJjF\\5V9hJeFY5[9kJ`FV5`9nJ[FS5e9QKfE`5Z:S201O01OO1N2N2N2M3N2N3bNnFjGT9`7lGTHV8^7]2L4SLdC1a<LcCOb<NaCNc<O`COc<L`C3c<I`C5c<G`C7c<DaC:a<CbC;a<AcC<`<AbC=a<@bC=a<@bC=a<@aC>a<@bC=a<@bC>`<_ObC`0_<^OdC`0]<^OeCa0\\<]OgC`0[<^OjC>V<APD:Q<DTD8m;EYD7h;G]D4e;J`D2a;KeD0];NhDNY;OmDMT;1QEJP;4TEJn:3UEKn:0UENo:NSE1P;JTE4o:HTE5o:HXE2k:I\\E1X`m1"}, "label": "a girl in black shorts"}]}
{"id": 245817, "image": "COCO_train2014_000000245817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in a white shirt and black shorts kicking a ball\"."}], "task": "refering", "answer_template": "The \"a little girl in a white shirt and black shorts kicking a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 17, 18, 19, 20, 21, 30, 31, 32, 33, 34, 35, 43, 44, 45, 46, 47, 48, 55, 56, 57, 58, 59, 60, 66, 67, 68, 69, 70, 71, 72, 73, 78, 79, 80, 81, 82, 83, 84, 85, 86, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 104, 106, 107, 108, 109, 110, 111, 112, 113, 117, 119, 120, 121, 122, 123, 124, 125, 130, 132, 133, 134, 135, 136, 137, 143, 145, 146, 147, 148, 149, 150, 158, 159, 160, 161, 162, 163, 171, 172, 173, 174, 175, 176, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 224, 225, 226, 227, 237, 238, 239, 250, 251, 252, 263, 264, 265, 276, 277, 278, 289, 290], "bbox": [0.0, 0.022625, 0.7290581717451523, 0.9776718750000001], "iscrowd": 0, "area": 78280.38735000003, "rle": {"size": [640, 361], "counts": "R99ea0HY@S1m>WOh@j0W?A^@a0`?JU@7k?3j_ONU`0=`_OD``0m1O101N0100O2N1O2O1N1N3N1O2WOi0M2O2N1N3N2O001O00101N1O2N100O2N1O2N101O00001O0010O100O00100O100O10O01O100O1N2O001O1N2O1O1O0ZAcNi:_1TDKd;6ZCT1]<mNPCm1h<TNUCV2c<jM[C`2^<`M_Cj2Y<WMcCT3V<lLgCY3V<hLgC^3V<bLgCc3V<]LhCh3V<S2M4M2M3N2M4M2M4M4K5L4KT3mLh0WOh0YOa0^O9H9F:G8G:G3L100O1O10O0100O100O10000O10000O1000000O1[Oe0J6J5K6J6M3N2M3N2N1O2N1N2O2N1O2NkMoM^BQ2c=PN\\BP2d=RNYBn1h=TNVBk1k=VNSBj1n=XNPBh1P>YNnAg1S>[NjAe1W>`N`Ad1`>aNWAb1j>cNm@`1T?fNb@^1^?hNX@[1i?]1100O1O100O1O1O100O1O1O1O1O000010O000XNTM]Bl2c=h100000000000002N1O1O1O1O1O1O1O1O2M2O1O1O1O1O1N2O1O2N1O1lLmIZHT6f7PJUHQ6k7SJoGo5Q8UJjGl5W8WJdGj5\\8ZJ_Gg5a8]JZGe5e8_JVGb5j8cJPG^5P9fJjF\\5V9hJeFY5[9kJ`FV5`9nJ[FS5e9QKfE`5Z:S201O01OO1N2N2N2M3N2N3bNnFjGT9`7lGTHV8^7]2L4SLdC1a<LcCOb<NaCNc<O`COc<L`C3c<I`C5c<G`C7c<DaC:a<CbC;a<AcC<`<AbC=a<@bC=a<@bC=a<@aC>a<@bC=a<@bC>`<_ObC`0_<^OdC`0]<^OeCa0\\<]OgC`0[<^OjC>V<APD:Q<DTD8m;EYD7h;G]D4e;J`D2a;KeD0];NhDNY;OmDMT;1QEJP;4TEJn:3UEKn:0UENo:NSE1P;JTE4o:HTE5o:HXE2k:I\\E1X`m1"}, "label": "a little girl in a white shirt and black shorts kicking a ball"}]}
{"id": 106557, "image": "COCO_train2014_000000106557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl full of fruit , including strawberries and mango\"."}], "task": "refering", "answer_template": "The \"a bowl full of fruit , including strawberries and mango\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 356, 357, 358, 359, 360, 361, 362, 363, 364, 380, 381, 382, 383, 384, 385, 386, 405, 406, 407], "bbox": [0.42251562500000006, 0.35753937007874015, 0.9033906250000001, 0.9556496062992126], "iscrowd": 0, "area": 72129.34620000001, "rle": {"size": [508, 640], "counts": "YXV42f??A?@a0@7H8I7I6I8I7H8J5K6K5J6K4L5J6K3M3N3L3N2N3M2M3N3M2N2N3L3N2N3M2N2M4M2N2N3M2M3N3M2N2M4M2N2N2N2M3O1N2O1N2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N2O1N2N101N100O2N101N1O2O000O2O001N101O0O2O000O2O001N101O0O101O0O2O001N10000O2O000O101N10000O101O000O10001O000000001O0000000O2O0000001O000000001O000000001O000001O01O000000000001O00000O10000000001O0O1000000000001N1000000000000O101O00001O000O2O001O00001O0O101O00001O000O2O001N100O2O0O101N1O101N101N1O101N100O2O1N1O2O1N2O1N1O2O1N2O0O2O1N2N2O0O2O1N2N2O0O2O1N2O0O2N2O1N2O0O2N2O1N1O2N2N2M3N2N2N2M4M2N3L3N3M2N3L3N2N3L3N3M2N3L3N3M2M3N3M2N3L3N3M2M4M2M3N3L3M4M3L7H8F:G9F:G9E;C=D<D<CV^n0"}, "label": "a bowl full of fruit , including strawberries and mango"}]}
{"id": 106557, "image": "COCO_train2014_000000106557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"container with strawberries , peaches and other foods inside\"."}], "task": "refering", "answer_template": "The \"container with strawberries , peaches and other foods inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 356, 357, 358, 359, 360, 361, 362, 363, 364, 380, 381, 382, 383, 384, 385, 386, 405, 406, 407], "bbox": [0.42251562500000006, 0.35753937007874015, 0.9033906250000001, 0.9556496062992126], "iscrowd": 0, "area": 72129.34620000001, "rle": {"size": [508, 640], "counts": "YXV42f??A?@a0@7H8I7I6I8I7H8J5K6K5J6K4L5J6K3M3N3L3N2N3M2M3N3M2N2N3L3N2N3M2N2M4M2N2N3M2M3N3M2N2M4M2N2N2N2M3O1N2O1N2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N2O1N2N101N100O2N101N1O2O000O2O001N101O0O2O000O2O001N101O0O101O0O2O001N10000O2O000O101N10000O101O000O10001O000000001O0000000O2O0000001O000000001O000000001O000001O01O000000000001O00000O10000000001O0O1000000000001N1000000000000O101O00001O000O2O001O00001O0O101O00001O000O2O001N100O2O0O101N1O101N101N1O101N100O2O1N1O2O1N2O1N1O2O1N2O0O2O1N2N2O0O2O1N2N2O0O2O1N2O0O2N2O1N2O0O2N2O1N1O2N2N2M3N2N2N2M4M2N3L3N3M2N3L3N2N3L3N3M2N3L3N3M2M3N3M2N3L3N3M2M4M2M3N3L3M4M3L7H8F:G9F:G9E;C=D<D<CV^n0"}, "label": "container with strawberries , peaches and other foods inside"}]}
{"id": 106557, "image": "COCO_train2014_000000106557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"thermos lid\"."}], "task": "refering", "answer_template": "The \"thermos lid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 125, 126, 127], "bbox": [0.30212500000000003, 0.002244094488188976, 0.618546875, 0.3108070866141732], "iscrowd": 0, "area": 23984.367750000005, "rle": {"size": [508, 640], "counts": "mho21i?;E<E:E6J2N3N2M2N3M3N1N3M3M2O2M3M2N3O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O002N2N2N3M2N2N3M2N2N3M2N2N1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O001O0000001O0000001O0000001O0000O1000000000000O1000000000000O1000000000000O1000000000000O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O1O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1N4M4L4K5L4K5L4L4K5L4L4J6^Ob0B>B>B^Pi3"}, "label": "thermos lid"}]}
{"id": 106557, "image": "COCO_train2014_000000106557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pink bowl\"."}], "task": "refering", "answer_template": "The \"pink bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 125, 126, 127], "bbox": [0.30212500000000003, 0.002244094488188976, 0.618546875, 0.3108070866141732], "iscrowd": 0, "area": 23984.367750000005, "rle": {"size": [508, 640], "counts": "mho21i?;E<E:E6J2N3N2M2N3M3N1N3M3M2O2M3M2N3O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O002N2N2N3M2N2N3M2N2N3M2N2N1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O001O0000001O0000001O0000001O0000O1000000000000O1000000000000O1000000000000O1000000000000O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O1O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1N4M4L4K5L4K5L4L4K5L4L4J6^Ob0B>B>B^Pi3"}, "label": "pink bowl"}]}
{"id": 106557, "image": "COCO_train2014_000000106557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a jar of veggies\"."}], "task": "refering", "answer_template": "The \"a jar of veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 281, 282], "bbox": [0.01175, 0.17687007874015748, 0.46303124999999995, 0.6802362204724409], "iscrowd": 0, "area": 59993.27464999999, "rle": {"size": [508, 640], "counts": "kT48a?9G8H9F:G8H9G8G:G6J6J5K6I7J5K6J6I6K6K4L5L4L3L5L4K4M4L4PFdJX9`5_FhJ_9k5N3N1N3N1O2M2O2O0O1O2O0O2N101N101N1O2O0O1O2O0O2N101N101N101O00001O001O001O001O001O001O00010O001O001O0000001O0000000000001O0000000000000010O000000000001O0000000000001O000000000000001O0001O00000001O000000000000001O0000000000001O000001O00000001O0000000000001O0000000000001OO100000001O0O10000000000O10000000000O10000000000O100000001O0O10000000000O1000000O1O1O1O1O100O1O1O1O1O1O2N1O1O1O1O100O1O1O1O2N1O1O2N1O1O101N1O1O2N1O2N2N3M2N3M2O2M3M2N3M2N3M2N3M2N3M3N1N3M2N3M2N3M3M2N3M2N3M2O2M2N3L4L3L5L3L5L3L5L4K4M4K4M4K4M4K4M4K5L`0@S_Z5"}, "label": "a jar of veggies"}]}
{"id": 106557, "image": "COCO_train2014_000000106557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vessel with a yellow duck like thing on it\"."}], "task": "refering", "answer_template": "The \"the vessel with a yellow duck like thing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 281, 282], "bbox": [0.01175, 0.17687007874015748, 0.46303124999999995, 0.6802362204724409], "iscrowd": 0, "area": 59993.27464999999, "rle": {"size": [508, 640], "counts": "kT48a?9G8H9F:G8H9G8G:G6J6J5K6I7J5K6J6I6K6K4L5L4L3L5L4K4M4L4PFdJX9`5_FhJ_9k5N3N1N3N1O2M2O2O0O1O2O0O2N101N101N1O2O0O1O2O0O2N101N101N101O00001O001O001O001O001O001O00010O001O001O0000001O0000000000001O0000000000000010O000000000001O0000000000001O000000000000001O0001O00000001O000000000000001O0000000000001O000001O00000001O0000000000001O0000000000001OO100000001O0O10000000000O10000000000O10000000000O100000001O0O10000000000O1000000O1O1O1O1O100O1O1O1O1O1O2N1O1O1O1O100O1O1O1O2N1O1O2N1O1O101N1O1O2N1O2N2N3M2N3M2O2M3M2N3M2N3M2N3M2N3M3N1N3M2N3M2N3M3M2N3M2N3M2O2M2N3L4L3L5L3L5L3L5L4K4M4K4M4K4M4K4M4K5L`0@S_Z5"}, "label": "the vessel with a yellow duck like thing on it"}]}
{"id": 409678, "image": "COCO_train2014_000000409678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a tray of pizza on it\"."}], "task": "refering", "answer_template": "The \"a table with a tray of pizza on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 58, 75, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.002984375, 0.19100938967136152, 1.0, 0.9820187793427231], "iscrowd": 0, "area": 136361.43469999998, "rle": {"size": [426, 640], "counts": "_W17d<?A?A?B>A?A?A?A?N2O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100bMcInM^6R2gIhMZ6X2jIcMW6]2nI]MS6b2RJYMo5g2VJTMj5l2ZJoLg5Q3^JiLc5W3aJdL`5\\3eJ^L\\5b3iJYLW5g3mJTLT5k3RKoKo4Q4UKjKl4V4T20000000000O1000000000000O100000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O17I9G1OO1nN_GZMb8a2gGYM[8b2mGZMT8a2THZMn7a2ZH[Mg7d2]HWMe7i2^1000000O1000000000000O1000000000000O1000000000000O11O2N1O2N2N2N2N2N1O2N2N0000000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O10000000000O1000000000000000000000000000000000000000000000000O10gHUMS4k2kKXMT4h2kKYMU4g2iK\\MV4d2hK_MW4a2gKbMX4^2fKdMZ4\\2eKfMZ4Z2dKiM[4V2dKlM\\4T2bKoM]4Q2aKRN^4n1aKTN^4l1`KVN`4j1\\K[Nc4e1XKaNg4_1TKfNl4Z1PKkNo4T1mJROR5n0jJWOU5k0gJWOY5k0dJWO[5k0aJXO^5j0_JWOa5k0[JXOd5j0XJYOg5i0VJXOj5j0RJYOm5h0QJYOo5h0nIZOR6g0lI[OS6f0jI\\OV6f0gI\\OX6e0eI]O[6d0bI_O]6b0aI_O_6g3O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O_E"}, "label": "a table with a tray of pizza on it"}]}
{"id": 524369, "image": "COCO_train2014_000000524369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe on right\"."}], "task": "refering", "answer_template": "The \"giraffe on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 63, 64, 65, 80, 81, 97, 98, 113, 114, 115, 130, 131, 147, 148, 163, 164, 165, 180, 181, 182, 196, 197, 198, 199, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 248, 249, 250, 265, 266, 267, 282, 283, 284, 299, 300, 301, 316, 317, 318, 333, 334, 335, 350, 351, 352, 353, 367, 368, 369, 370], "bbox": [0.5692916666666666, 0.085390625, 0.8509375, 0.9438124999999999], "iscrowd": 0, "area": 28944.291600000004, "rle": {"size": [640, 480], "counts": "PnZ53kc09G9G9G9G9G9dDYNo3P2gKSNW4V2_KmM_4\\2WKgMg4b2oJaMo4g2gJ]MW5l2_JWM_5Q3XJRMf5V3QJmLm5\\3iIgLU6a3bIbL\\6f3[I]Lc6l3SIWLk6Q4lHSLQ7U4fHnKX7Z4^HjK`7_4VHdKh7c4PH`Kn7d4mG_KQ8f4iG]KU8h4eG[KY8i4bGZK\\8k4^GXK`8m4ZGVKd8n4WGUKg8P5SGSKk8Q5oFTKn8Q5lFRKR9S5hFPKV9n7N2N2N2NjMXGXIf8h6hGlHV8T7XH`Hf7`7iHSHU7m7YIgGd6Z8kIYGR6h8a21O1O1O1O1OVMdETL[:o3XF\\Kg9f4mFeJR9^5aGmI^8V6b24M2N3M3M2lLRIUIQ7k6PIQIS7W2QI^OMWNT7i1cIOZOTNV7Y1UJb0fNQNX7j0eJU1UNlMY7;WKh1aMiMZ7MiKY2nLfM\\7kNmL^3hKcMo<]2SC^MQ=a2PC[MT=d2mBXMW=g2jBUMY=k2iBPM[=o2fBmL^=R3cBjL`=W3`BeLd=Z3]BbL]4ZOU4S4`G^LY4GU4j3gGVLR48U4b3mGoKl3g0U4Y3jLnLU3Q3eLXMX3h2cL_MPOjN@k2V2cNPNQOC]2U2jNoMXOEn1V2ROkM_OGa1V2XOjMFIS1V2^OgMNMd0V2FcM5O7V2LbM<1IV23_Mc04ZOV2;]Mi07mNU2b0ZMP19`NV2g0XMX1;QNV2P1UM^1>cMV2V1SMf1?VMW2\\1oLm1d0gLV2P9cMQGV2\\9cMeFU2i9cMXFW2j=L4K4M4K5L4M201O1O0kNf]Og0Zb0XOh]Of0Yb0XOi]Og0Xb0WOk]Og0Ub0YOm]Oe0Tb0YOo]Of0Qb0XOR^Of0na0ZOS^Oe0na0YOU^Oe0la0YOW^Oe0cb0O1N2N1O2N2O1N2N1O2N2O1N2Nam\\1"}, "label": "giraffe on right"}]}
{"id": 524369, "image": "COCO_train2014_000000524369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the right\"."}], "task": "refering", "answer_template": "The \"the giraffe on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 63, 64, 65, 80, 81, 97, 98, 113, 114, 115, 130, 131, 147, 148, 163, 164, 165, 180, 181, 182, 196, 197, 198, 199, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 248, 249, 250, 265, 266, 267, 282, 283, 284, 299, 300, 301, 316, 317, 318, 333, 334, 335, 350, 351, 352, 353, 367, 368, 369, 370], "bbox": [0.5692916666666666, 0.085390625, 0.8509375, 0.9438124999999999], "iscrowd": 0, "area": 28944.291600000004, "rle": {"size": [640, 480], "counts": "PnZ53kc09G9G9G9G9G9dDYNo3P2gKSNW4V2_KmM_4\\2WKgMg4b2oJaMo4g2gJ]MW5l2_JWM_5Q3XJRMf5V3QJmLm5\\3iIgLU6a3bIbL\\6f3[I]Lc6l3SIWLk6Q4lHSLQ7U4fHnKX7Z4^HjK`7_4VHdKh7c4PH`Kn7d4mG_KQ8f4iG]KU8h4eG[KY8i4bGZK\\8k4^GXK`8m4ZGVKd8n4WGUKg8P5SGSKk8Q5oFTKn8Q5lFRKR9S5hFPKV9n7N2N2N2NjMXGXIf8h6hGlHV8T7XH`Hf7`7iHSHU7m7YIgGd6Z8kIYGR6h8a21O1O1O1O1OVMdETL[:o3XF\\Kg9f4mFeJR9^5aGmI^8V6b24M2N3M3M2lLRIUIQ7k6PIQIS7W2QI^OMWNT7i1cIOZOTNV7Y1UJb0fNQNX7j0eJU1UNlMY7;WKh1aMiMZ7MiKY2nLfM\\7kNmL^3hKcMo<]2SC^MQ=a2PC[MT=d2mBXMW=g2jBUMY=k2iBPM[=o2fBmL^=R3cBjL`=W3`BeLd=Z3]BbL]4ZOU4S4`G^LY4GU4j3gGVLR48U4b3mGoKl3g0U4Y3jLnLU3Q3eLXMX3h2cL_MPOjN@k2V2cNPNQOC]2U2jNoMXOEn1V2ROkM_OGa1V2XOjMFIS1V2^OgMNMd0V2FcM5O7V2LbM<1IV23_Mc04ZOV2;]Mi07mNU2b0ZMP19`NV2g0XMX1;QNV2P1UM^1>cMV2V1SMf1?VMW2\\1oLm1d0gLV2P9cMQGV2\\9cMeFU2i9cMXFW2j=L4K4M4K5L4M201O1O0kNf]Og0Zb0XOh]Of0Yb0XOi]Og0Xb0WOk]Og0Ub0YOm]Oe0Tb0YOo]Of0Qb0XOR^Of0na0ZOS^Oe0na0YOU^Oe0la0YOW^Oe0cb0O1N2N1O2N2O1N2N1O2N2O1N2Nam\\1"}, "label": "the giraffe on the right"}]}
{"id": 524369, "image": "COCO_train2014_000000524369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe with his head lower\"."}], "task": "refering", "answer_template": "The \"the giraffe with his head lower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 77, 78, 79, 80, 93, 94, 95, 96, 109, 110, 111, 112, 113, 125, 126, 127, 128, 141, 142, 143, 144, 145, 158, 159, 160, 161, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330, 331, 344, 345, 346, 347, 348, 361, 362, 363, 364, 365], "bbox": [0.21822916666666667, 0.127796875, 0.7145, 0.9372187499999999], "iscrowd": 0, "area": 44877.63035, "rle": {"size": [640, 480], "counts": "RPR29]c0>B=C>F9I7J7I6I7J7I6lCSMk6T3SEiLh2>Q8P3PEkL`2>]8m2mDnLV2?i8j2kD[Mb14`9^5lDlJS;R7M2O1O1O2M2O1O1O1N3N1O1O1N3N[Me2M3L3N3L4M3L3011O1O1`EnIP7T6fHVJY7k5]H_Jb7c5SHgJl7Z5jGPKT8R5bGYK\\8h4SGiKl8X4cFYL\\9h3RFjLl9f6O1O1O1N2O1O1M3N2M3M3M3M4L3M3M3M3M3M3N2M3M3M3O1O1O100O1O`LUHmJk7n4lHaJR7Z5gISJX6h5`JfI_5U6ZKXIf4c6gKUIX4f6ULSIj3h6dLPI[3k6RMnHn2m6T5K5K6K4L4O101N1O1O101NcDnIo8Q6UFWK`9i4cE_LS:P7F;E:F;F9F;E;E:G4K5K6K4K5K5K5L1N001O100OoK]IgJb6Q5`JVJ`5`5dKfI[4Q6hLUIX3c6jMdHU2S7YNiHg1P7bNnH]1o6hNPIW1m6oNQIP1l6UOSIk0i6[OTIe0i6@VI?f6HXI8d6c6M3M3[LgBa0]=^OPC2S=M[CCh<=eCSO^<l0oCeNU<Z1XDVNk;j1bDfMa;Y2T3O0O10001N100O101O0O101N10000O2O0O10001N100OoNkM[@T2c?QNZ@o1d?VNY@i1i?YNT@g1n?ZNP@d1S`0\\Nk_Od1W`0\\Ng_Od1[`0]Nc_Oa1``0_N^_Oa1d`0_NZ_O`1i`0`NU_O`1m`0aNQ_O^1Ra0aNl^O^1Wa0bNg^O^1[a0bNc^O]1`a0dN^^O[1da0?1NF;J5M4K5L3N33L3N3M32MO2N1O2N2N1N3N1O2N2N00K5L3L5K5N21O2N1O2N2M2Ok0RO3ROm0H9O010O0010O01O01OO2O0O2O0O2O0O2O1N5L3L^`e2"}, "label": "the giraffe with his head lower"}]}
{"id": 524369, "image": "COCO_train2014_000000524369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the left of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe to the left of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 77, 78, 79, 80, 93, 94, 95, 96, 109, 110, 111, 112, 113, 125, 126, 127, 128, 141, 142, 143, 144, 145, 158, 159, 160, 161, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330, 331, 344, 345, 346, 347, 348, 361, 362, 363, 364, 365], "bbox": [0.21822916666666667, 0.127796875, 0.7145, 0.9372187499999999], "iscrowd": 0, "area": 44877.63035, "rle": {"size": [640, 480], "counts": "RPR29]c0>B=C>F9I7J7I6I7J7I6lCSMk6T3SEiLh2>Q8P3PEkL`2>]8m2mDnLV2?i8j2kD[Mb14`9^5lDlJS;R7M2O1O1O2M2O1O1O1N3N1O1O1N3N[Me2M3L3N3L4M3L3011O1O1`EnIP7T6fHVJY7k5]H_Jb7c5SHgJl7Z5jGPKT8R5bGYK\\8h4SGiKl8X4cFYL\\9h3RFjLl9f6O1O1O1N2O1O1M3N2M3M3M3M4L3M3M3M3M3M3N2M3M3M3O1O1O100O1O`LUHmJk7n4lHaJR7Z5gISJX6h5`JfI_5U6ZKXIf4c6gKUIX4f6ULSIj3h6dLPI[3k6RMnHn2m6T5K5K6K4L4O101N1O1O101NcDnIo8Q6UFWK`9i4cE_LS:P7F;E:F;F9F;E;E:G4K5K6K4K5K5K5L1N001O100OoK]IgJb6Q5`JVJ`5`5dKfI[4Q6hLUIX3c6jMdHU2S7YNiHg1P7bNnH]1o6hNPIW1m6oNQIP1l6UOSIk0i6[OTIe0i6@VI?f6HXI8d6c6M3M3[LgBa0]=^OPC2S=M[CCh<=eCSO^<l0oCeNU<Z1XDVNk;j1bDfMa;Y2T3O0O10001N100O101O0O101N10000O2O0O10001N100OoNkM[@T2c?QNZ@o1d?VNY@i1i?YNT@g1n?ZNP@d1S`0\\Nk_Od1W`0\\Ng_Od1[`0]Nc_Oa1``0_N^_Oa1d`0_NZ_O`1i`0`NU_O`1m`0aNQ_O^1Ra0aNl^O^1Wa0bNg^O^1[a0bNc^O]1`a0dN^^O[1da0?1NF;J5M4K5L3N33L3N3M32MO2N1O2N2N1N3N1O2N2N00K5L3L5K5N21O2N1O2N2M2Ok0RO3ROm0H9O010O0010O01O01OO2O0O2O0O2O0O2O1N5L3L^`e2"}, "label": "a giraffe to the left of another giraffe"}]}
{"id": 188498, "image": "COCO_train2014_000000188498.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing on some skis and holding ski poles wearing black sunglasses\"."}], "task": "refering", "answer_template": "The \"a man standing on some skis and holding ski poles wearing black sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 97, 98, 99, 112, 113, 114, 115, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 246, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 292, 293, 294, 295, 307, 308, 309, 310, 322, 323, 324, 325, 338, 339, 340], "bbox": [0.4080660377358491, 0.218421875, 0.8513207547169812, 0.99803125], "iscrowd": 0, "area": 52403.9571, "rle": {"size": [640, 424], "counts": "ja\\33gc06I8I6J6I7J6J6I8I6J6J6K5K5K6J5K5K5K5K5K6K4M3M3M3iAoK[<U4bCQLX<S4dCSLV<P4hCULR<n3kCXLo;`2nBkMP1Kl;^2cCWM>`0j8jNSIe3SNaLMU1d8nNWI`3YOgM[7mNXIa3XOgMX7PO[I]3YOiMS7QO`I[3XOiMb6UNRHo0o1W3YOjMc6_OQJk2WOPO_5YNVKj2WORO`5XNTKj2WOSOb5WNRKk2VOSOe5VNPKR3oNmNm5VNoJX3hNgNV6UNmJe7P5_HkJe7Q5`HjJd7S5`HhJd7U5_HgJe7V5]HgJg7W5[HeJi7Z5WHcJm7[5THbJP8\\5RH`JR8_5Y3N2O1N2O1N2N2O1N2O1N2O1N2O1O1N2O1N01O000000000000000000000000000000000000000000000000:F2N2N2N2N2N1O2N2N2N2N2N1O2RLXJ`I\\6d5_JjI_5o5ZK`Ie4Y6gKcIW4W6ULeIj3U6aLgI^3X6hLdIW3\\6nL`IQ3`6TM\\Ij2d6\\MXIc2h6bMTI]2l6hMPIW2o6oMmHP2S7TNjHn1S7WNiHo1\\OlIl6Z4eIP2VOjIS7Z4dIR2POhIZ7[4bI[2\\6iMaIX2\\6mMaIT2]6QN_IQ2]6TN`Im1]6YN_Ih1]6_N_Ib1^6cN_I^1^6hN^IY1^6W6M3L2O2M3N2M3O001O100O2N1O2N1O2N1O100O1O2cJmAk4e>N3M3N1N3N2M3M2O2M3M3N1N3M3N1N3C=L4L9H7H9SOd^OfNda0S1h^OaN`a0W1j0I7J6J6M2N3M3J6IlPW1"}, "label": "a man standing on some skis and holding ski poles wearing black sunglasses"}]}
{"id": 188498, "image": "COCO_train2014_000000188498.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier wearing dark green\"."}], "task": "refering", "answer_template": "The \"the skier wearing dark green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 97, 98, 99, 112, 113, 114, 115, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 246, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 292, 293, 294, 295, 307, 308, 309, 310, 322, 323, 324, 325, 338, 339, 340], "bbox": [0.4080660377358491, 0.218421875, 0.8513207547169812, 0.99803125], "iscrowd": 0, "area": 52403.9571, "rle": {"size": [640, 424], "counts": "ja\\33gc06I8I6J6I7J6J6I8I6J6J6K5K5K6J5K5K5K5K5K6K4M3M3M3iAoK[<U4bCQLX<S4dCSLV<P4hCULR<n3kCXLo;`2nBkMP1Kl;^2cCWM>`0j8jNSIe3SNaLMU1d8nNWI`3YOgM[7mNXIa3XOgMX7PO[I]3YOiMS7QO`I[3XOiMb6UNRHo0o1W3YOjMc6_OQJk2WOPO_5YNVKj2WORO`5XNTKj2WOSOb5WNRKk2VOSOe5VNPKR3oNmNm5VNoJX3hNgNV6UNmJe7P5_HkJe7Q5`HjJd7S5`HhJd7U5_HgJe7V5]HgJg7W5[HeJi7Z5WHcJm7[5THbJP8\\5RH`JR8_5Y3N2O1N2O1N2N2O1N2O1N2O1N2O1O1N2O1N01O000000000000000000000000000000000000000000000000:F2N2N2N2N2N1O2N2N2N2N2N1O2RLXJ`I\\6d5_JjI_5o5ZK`Ie4Y6gKcIW4W6ULeIj3U6aLgI^3X6hLdIW3\\6nL`IQ3`6TM\\Ij2d6\\MXIc2h6bMTI]2l6hMPIW2o6oMmHP2S7TNjHn1S7WNiHo1\\OlIl6Z4eIP2VOjIS7Z4dIR2POhIZ7[4bI[2\\6iMaIX2\\6mMaIT2]6QN_IQ2]6TN`Im1]6YN_Ih1]6_N_Ib1^6cN_I^1^6hN^IY1^6W6M3L2O2M3N2M3O001O100O2N1O2N1O2N1O100O1O2cJmAk4e>N3M3N1N3N2M3M2O2M3M3N1N3M3N1N3C=L4L9H7H9SOd^OfNda0S1h^OaN`a0W1j0I7J6J6M2N3M3J6IlPW1"}, "label": "the skier wearing dark green"}]}
{"id": 106586, "image": "COCO_train2014_000000106586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"helicopter prototype with green parts on it\"."}], "task": "refering", "answer_template": "The \"helicopter prototype with green parts on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 26, 27, 41, 42, 43, 57, 58, 59, 72, 73, 74, 75, 81, 82, 83, 89, 90, 91, 96, 97, 98, 99, 105, 106, 107, 108, 111, 112, 113, 114, 115, 121, 122, 123, 124, 125, 127, 128, 129, 130, 137, 138, 140, 141, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 340, 341, 342, 343, 345, 348, 349, 350, 351, 352, 357, 358, 359, 366, 367, 368, 369, 375, 376], "bbox": [0.041291666666666664, 0.0, 0.8949166666666667, 0.9843125], "iscrowd": 0, "area": 88946.40534999999, "rle": {"size": [640, 480], "counts": "]Q=7fc05n\\OKSb09i]OJka0b0R^OBia0c0S^O@la0c0o]O@Qb0X101N100M3L5J5L3L5WMbMkCa2m;iMmC\\2P<gMlC\\2P<iMkC\\2o;kMlCX2P<mMlCW2o;nMlCW2o;nMmCU2e6`LVM]1RLW2g6]LTM_1RLV2i6]LRM_1SLW2i6[LRMa1RLV2k6[LPMa1SLW2k6ZLnLc1TLU2m6YLmLd1TLV2n6WLkLf1TLU2P7VLjLh1SLU2Q7ULhLi1TLU2S7TLfLj1TLU2T7RLfLk1TLV2U7oKdLo1TLU2V7mKdLQ2TLT2V7lKfLP2SLU2V7lKfLQ2SLT2U7jKjLR2PLV2T7hKYK]OoNe2dNW2S7gKXK@POc2cNW2T7fKXKCPO`2dNW2S7gKYKCnN`2eNX2R7eK[KDnN_2eNX2Q7fK[KCoNa2cNX2Q7eK]KAPOb2bNX2Q7dK]KCoNb2bNX2Q7dK^KAoNi7a5fHaK@oNj7_5gHaK^OQOl7]5fHcK]OPOo7[5eHeK]OmNP8^5bHeK_OkNR8^5`HgK]OjNU8^5_HhK\\OhNW8_5]HiK\\OhNY8]5\\HkK[OgNZ8^5ZHlK\\OfNZ8]5[HlK[OgNZ8]5[HmK[OfN_5iN\\K0g0d6SOmKZOfN`5ROTKHn0b6TOnK[OfN^5[OmJA12:^60nKZOeN^5@iJ@O8;T66oKXOfN_5@hJAHa0>i5:oKZOeN_5AgJd0OX5b0nKXOfN`5AfJh0FW5l0jKYOfN`5AdJa7c0XJXOfNb5AcJ`7c0ZJXOeNb5BbJ^7e0ZJXOfNb5B`J]7f0\\JWOfNd5A^J]7g0[JXOfNe5AZJ_7i0[JWOfNg5k7SK_IVOfNh5j7QK`IXOfNh5h7QKbIVOgNj5e7QKcIVOgNj5e7PKeIUOgNk5d7PKeIUOgNk5c7PKfIVOgNk5a7PKhITOhNl5`7oJhIVOhNk5_7PKiIUOgNl5_7oJkITOdNP6`7mJkITObNS6b7hJmIUO^NV6d7fJmITO]NY6e7cJoITO[NZ6f7bJnIUO\\NY6e7bJoIVO\\NY6c7bJQJTOdNS6Z7jJQJTOgNP6X7kJRJUOeNQ6X7kJRJTOeNT6W7hJUJTOcNU6X7gJTJUOcNU6X7fJVJTOaNX6X7eJVJTObNW6W7eJXJTO`NY6W7cJYJTO_NZ6W7bJZJTO_N[6V7bJ[JSO^N]6U7`J]JTO]N]6U7`J^JRO^N^6S7`JjIlNeN6>_6Q7`JjIoNdN3`0_6Q7_JkIRObN0:H^Og6l6YIiIX1Q1RO_NO<I^Oe6l6[IhIW1S1SO]NM=K^Od6l6[IgIV1U1VO[NJ>M^Ob6l6]IfIU1W1WOXNIa0L]Oc6m6\\IeIU1Y1XOVNHa0N]Ob6l6]I^IHK\\1h1DfNW6X6SIZINLS1m1EdNX6Y6RIYI_1h1ZObNW6[6RIYIn1Z1jNoNX6^6SLnIeMCY6_6PN_IQ2`6PN`IQ2_6PN_IQ2a6oM^IR2b6nM^IS2a6mM^IT2b6lM^IT2b6lM]IV2b6jM^IV2a6kM^IW2a6iM^IX2b6hM]IY2c6c5O001O00001O000000000000000000000O1O1O10O01O1O00cCXJd:g5]EcJX:]5hEkJQ:T5PFnJm9R5TFnJkN\\OY:e5mFoJhN8`9j4hGgLP8Y3lGgInNo2W9Y3jGmLV8S3jGnLV8Q3iGPMV8Q3jGoLV8Q3iGQMU8P3kGPM:hL[6Y6ZIPM8kL]6V6ZIPM8lL[6U6]IoL8nLX6T6_IoL:mLV6T6`IoL:oLS6S6cIoL9PMQ6R6eIoL6cLmN>X7n5eIQM6VMU6h5eISM6VMT6f5fITM7UMU6e5dIVM8UMT6e5cIWM9TMT6d5cIXM9UMU6a5bIZM:UMT6`5bI\\M:TMT6_5bI]M;TMS6^5bI^M<TMR6]5bI`M;TMT6Z5aIbM<TMS6Y5`IdM>SMR6X5`IfM>RMR6W5`IgM?RMR6U5_IiM?UMP6Q5aIjM`0XMl5n4cIkMa0ZMi5j4fIlMb0]Me5f4iImMc0`Mb5a4kIPNb0cM`5\\4nIQNc0fM\\5X4QJRNd0iMX5T4TJTNd0kMU5P4XJTNc0oMT5k3YJVNd0QNQ5h3[JXNd0PNQ5g3[JYNe0QNo4e3\\JZNf0QNn4d3\\J\\Ne0RNn4a3]J]Nf0RNm4`3]J^Ng0SNl4]3^J`Ne0TNm4\\3]J`Ng0TNl4[3]JaNg0VNk4X3^JbNh0VNj4W3^JdNg0WNj4T3_JeNh0WNi4S3_JfNh0XNi4Q3_JhNh0XNi4n2`JiNg0ZNi4l2`JjNh0[Ng4j2aJlNg0[Nh4h2aJmNg0\\Nh4f2aJnNg0^Ng4c2bJPOg0]Ng4b2bJQOg0_Nf4QNkIl3h0Dg0aNg4kMoIl3d0Hf0aNh4iMSJj3>Lg0cNi4dMWJh391g0eNb5W1gI5f0fNc5S1gI7g0fNd5P1eI;f0gNe5l0eI=f0iNf5g0dI`0g0jNd5e0fIa0e0kNf5b0eIc0e0mNe5?fId0f0nNc5>fIe0f0oNd5:fIg0f0POd58fIh0f0ROc55gIj0f0ROa54iIj0f0TO`51jIk0f0VO_5NkIm0f0UO_5MlIm0e0XO]5KnIm0e0ZO\\5HoIo0e0ZO[5FPJP1e0[O[5EoIP1f0\\O\\5BnIS1f0[O\\5AnIT1f0]O]5\\OmIW1f0^O^5YOlIZ1e0^O`5VOkI\\1f0_O`5ROkI^1e0Aa5oNjIa1d0Ad5kNhId1e0AiNcLV6W2\\Je1e0CbNgL]6Q2[Jf1e0C^NkLb6k1[Jg1f07o4PN[Ji1f08o4nM[Jk1e09o4kM\\Jl1f09n4jM\\Jm1f0:o4gM[JP2e0;]ObK_4R2_KQ2e0<\\ObK`4P2`KQ2e0=ZOdK_4n1bKR2d0>gN_K46m4k1dKR2d0?gNbKO4o4k1fKP2f0`0dNgKKOT5j1gKQ2e0`0dNXLm4W1jKQ2e0a0aNbLg4l0SLQ2f0b0]NjLa4d0\\LQ2e0^2f2aKeLQ2e0f2^2YKmLR2d0n2V2PKVMR2e0S3o1lJ[MQ2f0U3l1kJ_MP2d0X3l1fJ`MR2e0Z3i1cJbMS2e0\\3i1_JbMV2d0^3h1[JdMW2d0a3g1WJdMX2f0c3d1TJfMZ2e0d3d1QJhMZ2d0g3d1mIhM]2c0i3e1gIhM`2d0j3e1cIgMc2d0l3g1^IdMg2d0n3i1WIdMj2c0Y4c1iHjMn2d0[4c1cHiMS3c0^4c1]HiMU3d0a4b1YHiMW3e0c4a1THjMX3e0g4`1PHjMY3f0k4_1jGjM\\3f0m4`1eGiM^3h0P5\\1bGlM]3h0T5Z1_GmM^3h0V5Y1\\GnM^3i0Y5W1YGoM_3j0Y5V1XGoM_3k0\\5T1UGQN^3k0`5R1RGRN_3l0a5P1PGSN_3m0c5o0nFSN`3m0e5n0kFUN_3m0i5k0iFWN^3o0j5h0iFXN^3o0k5h0gFXN^3P1m5f0fFYN]3Q1n5f0eFYN]3Q1n5e0fFYN\\3R1o5d0fFYN\\3R1o5b0hF[NY3T1o5>kF]NV3U1P6;mF_NT3U1P6oNaEk0^1POQ3W1Q6lNaEn0\\1oNQ3W1S6kNaEo0[1nNR3W1S6lNaEn0Z1nNR3Y1S6jNbE1Kd0]1WOT3Y1T6jN`E1Nc0Z1XOT3Z1V6hN_E2Od0W1WOU3\\1W6fN_E11d0T1XOV3\\1W6eN`E13d0P1YOV3]1e6dNXEf0l0YOV3^1Z7WOSJh0o5WOPJi0P6YOoIf0Q6[OnIf0S6YOlIg0V6WOkIh0W6WOhIj0Y6UOgIj0[6jNPDNd5X1^6hNmIY1]<O00010O0010O00010O00010O01O01O01O00010O0010O00010O01O01O000100O00100O2N100O2N100O2N1O101N1O01L300100O1O2O0O1O101N1O101N1O100O2N100O2O0O1Ef\\OOfc000O1O1Ojgo0"}, "label": "helicopter prototype with green parts on it"}]}
{"id": 106586, "image": "COCO_train2014_000000106586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an exposed engine with two pairs of rotating helicopter blades\"."}], "task": "refering", "answer_template": "The \"an exposed engine with two pairs of rotating helicopter blades\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 26, 27, 41, 42, 43, 57, 58, 59, 72, 73, 74, 75, 81, 82, 83, 89, 90, 91, 96, 97, 98, 99, 105, 106, 107, 108, 111, 112, 113, 114, 115, 121, 122, 123, 124, 125, 127, 128, 129, 130, 137, 138, 140, 141, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 340, 341, 342, 343, 345, 348, 349, 350, 351, 352, 357, 358, 359, 366, 367, 368, 369, 375, 376], "bbox": [0.041291666666666664, 0.0, 0.8949166666666667, 0.9843125], "iscrowd": 0, "area": 88946.40534999999, "rle": {"size": [640, 480], "counts": "]Q=7fc05n\\OKSb09i]OJka0b0R^OBia0c0S^O@la0c0o]O@Qb0X101N100M3L5J5L3L5WMbMkCa2m;iMmC\\2P<gMlC\\2P<iMkC\\2o;kMlCX2P<mMlCW2o;nMlCW2o;nMmCU2e6`LVM]1RLW2g6]LTM_1RLV2i6]LRM_1SLW2i6[LRMa1RLV2k6[LPMa1SLW2k6ZLnLc1TLU2m6YLmLd1TLV2n6WLkLf1TLU2P7VLjLh1SLU2Q7ULhLi1TLU2S7TLfLj1TLU2T7RLfLk1TLV2U7oKdLo1TLU2V7mKdLQ2TLT2V7lKfLP2SLU2V7lKfLQ2SLT2U7jKjLR2PLV2T7hKYK]OoNe2dNW2S7gKXK@POc2cNW2T7fKXKCPO`2dNW2S7gKYKCnN`2eNX2R7eK[KDnN_2eNX2Q7fK[KCoNa2cNX2Q7eK]KAPOb2bNX2Q7dK]KCoNb2bNX2Q7dK^KAoNi7a5fHaK@oNj7_5gHaK^OQOl7]5fHcK]OPOo7[5eHeK]OmNP8^5bHeK_OkNR8^5`HgK]OjNU8^5_HhK\\OhNW8_5]HiK\\OhNY8]5\\HkK[OgNZ8^5ZHlK\\OfNZ8]5[HlK[OgNZ8]5[HmK[OfN_5iN\\K0g0d6SOmKZOfN`5ROTKHn0b6TOnK[OfN^5[OmJA12:^60nKZOeN^5@iJ@O8;T66oKXOfN_5@hJAHa0>i5:oKZOeN_5AgJd0OX5b0nKXOfN`5AfJh0FW5l0jKYOfN`5AdJa7c0XJXOfNb5AcJ`7c0ZJXOeNb5BbJ^7e0ZJXOfNb5B`J]7f0\\JWOfNd5A^J]7g0[JXOfNe5AZJ_7i0[JWOfNg5k7SK_IVOfNh5j7QK`IXOfNh5h7QKbIVOgNj5e7QKcIVOgNj5e7PKeIUOgNk5d7PKeIUOgNk5c7PKfIVOgNk5a7PKhITOhNl5`7oJhIVOhNk5_7PKiIUOgNl5_7oJkITOdNP6`7mJkITObNS6b7hJmIUO^NV6d7fJmITO]NY6e7cJoITO[NZ6f7bJnIUO\\NY6e7bJoIVO\\NY6c7bJQJTOdNS6Z7jJQJTOgNP6X7kJRJUOeNQ6X7kJRJTOeNT6W7hJUJTOcNU6X7gJTJUOcNU6X7fJVJTOaNX6X7eJVJTObNW6W7eJXJTO`NY6W7cJYJTO_NZ6W7bJZJTO_N[6V7bJ[JSO^N]6U7`J]JTO]N]6U7`J^JRO^N^6S7`JjIlNeN6>_6Q7`JjIoNdN3`0_6Q7_JkIRObN0:H^Og6l6YIiIX1Q1RO_NO<I^Oe6l6[IhIW1S1SO]NM=K^Od6l6[IgIV1U1VO[NJ>M^Ob6l6]IfIU1W1WOXNIa0L]Oc6m6\\IeIU1Y1XOVNHa0N]Ob6l6]I^IHK\\1h1DfNW6X6SIZINLS1m1EdNX6Y6RIYI_1h1ZObNW6[6RIYIn1Z1jNoNX6^6SLnIeMCY6_6PN_IQ2`6PN`IQ2_6PN_IQ2a6oM^IR2b6nM^IS2a6mM^IT2b6lM^IT2b6lM]IV2b6jM^IV2a6kM^IW2a6iM^IX2b6hM]IY2c6c5O001O00001O000000000000000000000O1O1O10O01O1O00cCXJd:g5]EcJX:]5hEkJQ:T5PFnJm9R5TFnJkN\\OY:e5mFoJhN8`9j4hGgLP8Y3lGgInNo2W9Y3jGmLV8S3jGnLV8Q3iGPMV8Q3jGoLV8Q3iGQMU8P3kGPM:hL[6Y6ZIPM8kL]6V6ZIPM8lL[6U6]IoL8nLX6T6_IoL:mLV6T6`IoL:oLS6S6cIoL9PMQ6R6eIoL6cLmN>X7n5eIQM6VMU6h5eISM6VMT6f5fITM7UMU6e5dIVM8UMT6e5cIWM9TMT6d5cIXM9UMU6a5bIZM:UMT6`5bI\\M:TMT6_5bI]M;TMS6^5bI^M<TMR6]5bI`M;TMT6Z5aIbM<TMS6Y5`IdM>SMR6X5`IfM>RMR6W5`IgM?RMR6U5_IiM?UMP6Q5aIjM`0XMl5n4cIkMa0ZMi5j4fIlMb0]Me5f4iImMc0`Mb5a4kIPNb0cM`5\\4nIQNc0fM\\5X4QJRNd0iMX5T4TJTNd0kMU5P4XJTNc0oMT5k3YJVNd0QNQ5h3[JXNd0PNQ5g3[JYNe0QNo4e3\\JZNf0QNn4d3\\J\\Ne0RNn4a3]J]Nf0RNm4`3]J^Ng0SNl4]3^J`Ne0TNm4\\3]J`Ng0TNl4[3]JaNg0VNk4X3^JbNh0VNj4W3^JdNg0WNj4T3_JeNh0WNi4S3_JfNh0XNi4Q3_JhNh0XNi4n2`JiNg0ZNi4l2`JjNh0[Ng4j2aJlNg0[Nh4h2aJmNg0\\Nh4f2aJnNg0^Ng4c2bJPOg0]Ng4b2bJQOg0_Nf4QNkIl3h0Dg0aNg4kMoIl3d0Hf0aNh4iMSJj3>Lg0cNi4dMWJh391g0eNb5W1gI5f0fNc5S1gI7g0fNd5P1eI;f0gNe5l0eI=f0iNf5g0dI`0g0jNd5e0fIa0e0kNf5b0eIc0e0mNe5?fId0f0nNc5>fIe0f0oNd5:fIg0f0POd58fIh0f0ROc55gIj0f0ROa54iIj0f0TO`51jIk0f0VO_5NkIm0f0UO_5MlIm0e0XO]5KnIm0e0ZO\\5HoIo0e0ZO[5FPJP1e0[O[5EoIP1f0\\O\\5BnIS1f0[O\\5AnIT1f0]O]5\\OmIW1f0^O^5YOlIZ1e0^O`5VOkI\\1f0_O`5ROkI^1e0Aa5oNjIa1d0Ad5kNhId1e0AiNcLV6W2\\Je1e0CbNgL]6Q2[Jf1e0C^NkLb6k1[Jg1f07o4PN[Ji1f08o4nM[Jk1e09o4kM\\Jl1f09n4jM\\Jm1f0:o4gM[JP2e0;]ObK_4R2_KQ2e0<\\ObK`4P2`KQ2e0=ZOdK_4n1bKR2d0>gN_K46m4k1dKR2d0?gNbKO4o4k1fKP2f0`0dNgKKOT5j1gKQ2e0`0dNXLm4W1jKQ2e0a0aNbLg4l0SLQ2f0b0]NjLa4d0\\LQ2e0^2f2aKeLQ2e0f2^2YKmLR2d0n2V2PKVMR2e0S3o1lJ[MQ2f0U3l1kJ_MP2d0X3l1fJ`MR2e0Z3i1cJbMS2e0\\3i1_JbMV2d0^3h1[JdMW2d0a3g1WJdMX2f0c3d1TJfMZ2e0d3d1QJhMZ2d0g3d1mIhM]2c0i3e1gIhM`2d0j3e1cIgMc2d0l3g1^IdMg2d0n3i1WIdMj2c0Y4c1iHjMn2d0[4c1cHiMS3c0^4c1]HiMU3d0a4b1YHiMW3e0c4a1THjMX3e0g4`1PHjMY3f0k4_1jGjM\\3f0m4`1eGiM^3h0P5\\1bGlM]3h0T5Z1_GmM^3h0V5Y1\\GnM^3i0Y5W1YGoM_3j0Y5V1XGoM_3k0\\5T1UGQN^3k0`5R1RGRN_3l0a5P1PGSN_3m0c5o0nFSN`3m0e5n0kFUN_3m0i5k0iFWN^3o0j5h0iFXN^3o0k5h0gFXN^3P1m5f0fFYN]3Q1n5f0eFYN]3Q1n5e0fFYN\\3R1o5d0fFYN\\3R1o5b0hF[NY3T1o5>kF]NV3U1P6;mF_NT3U1P6oNaEk0^1POQ3W1Q6lNaEn0\\1oNQ3W1S6kNaEo0[1nNR3W1S6lNaEn0Z1nNR3Y1S6jNbE1Kd0]1WOT3Y1T6jN`E1Nc0Z1XOT3Z1V6hN_E2Od0W1WOU3\\1W6fN_E11d0T1XOV3\\1W6eN`E13d0P1YOV3]1e6dNXEf0l0YOV3^1Z7WOSJh0o5WOPJi0P6YOoIf0Q6[OnIf0S6YOlIg0V6WOkIh0W6WOhIj0Y6UOgIj0[6jNPDNd5X1^6hNmIY1]<O00010O0010O00010O00010O01O01O01O00010O0010O00010O01O01O000100O00100O2N100O2N100O2N1O101N1O01L300100O1O2O0O1O101N1O101N1O100O2N100O2O0O1Ef\\OOfc000O1O1Ojgo0"}, "label": "an exposed engine with two pairs of rotating helicopter blades"}]}
{"id": 344157, "image": "COCO_train2014_000000344157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wedding cake sitting on a table with a tablecloth under it\"."}], "task": "refering", "answer_template": "The \"a wedding cake sitting on a table with a tablecloth under it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.0016875000000000002, 0.494375, 0.614859375, 0.989875], "iscrowd": 0, "area": 60416.12509999999, "rle": {"size": [480, 640], "counts": "Wk0a2_<0O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O2O0O1O1O1O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1O1O100O100O100O100000000000000000000000000001O00000000M3K5L4N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O100O100O100O100O10000O100O100O100O11O000000001O0000001O0001O0001O0000001O00001O00001O00001O001O00001O00001O001O00001O001O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O001O1O001O1O001O000000000000000000000000006J7Ic0]O2N1O1O1O1O001O001O1O1O1O001O1O1O1O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O010O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O1O1O2N2N2N1O2N2N2N2M3N2M3N2M3N2M3M3N2M3N1N3L4\\Od0[OXmb3"}, "label": "a wedding cake sitting on a table with a tablecloth under it"}]}
{"id": 344157, "image": "COCO_train2014_000000344157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that the cake is on\"."}], "task": "refering", "answer_template": "The \"the table that the cake is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.0016875000000000002, 0.494375, 0.614859375, 0.989875], "iscrowd": 0, "area": 60416.12509999999, "rle": {"size": [480, 640], "counts": "Wk0a2_<0O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O2O0O1O1O1O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1O1O100O100O100O100000000000000000000000000001O00000000M3K5L4N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O100O100O100O100O10000O100O100O100O11O000000001O0000001O0001O0001O0000001O00001O00001O00001O001O00001O00001O001O00001O001O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O001O1O001O1O001O000000000000000000000000006J7Ic0]O2N1O1O1O1O001O001O1O1O1O001O1O1O1O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O010O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O1O1O2N2N2N1O2N2N2N2M3N2M3N2M3N2M3M3N2M3N1N3L4\\Od0[OXmb3"}, "label": "the table that the cake is on"}]}
{"id": 344157, "image": "COCO_train2014_000000344157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in dress shirt\"."}], "task": "refering", "answer_template": "The \"a man in dress shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 128, 129, 130, 151, 152, 153, 154, 172, 173, 175, 176, 177, 195, 196, 199, 200, 218, 219, 222, 223, 246], "bbox": [0.5066875, 0.21320833333333333, 0.732921875, 0.6530416666666667], "iscrowd": 0, "area": 8589.165899999996, "rle": {"size": [480, 640], "counts": "bRh43j09k<2hB6R=m0N3M3M4L6J4M2001O1N1O0000ORN`C^1Y=HfLlNQIk0l6[OeI1Y62Z40O1O1O2N1O1O1O1O1O010O10000O101O0O101O010O001O1O1QCWOe;j0XDYOg;j0UDWOk;o0mCTOR<o0iCTOV<n0eCUO[<i1O010000O1O1O1O100O1O2N3M2N3M7I4M2M2N2N2N100000O010O1001O2N1N10O101N1O1O1O100O2N1O1O1O101N4L3M4L4M2M4L3M3M3M3N1N2N2N2N2O0OO2N2N2N3NjC`Mm;^2PDhMo;e200001O100O1O22M3M2O2M2N20O1O1O2N1O2K4L4K9G<D<C<E:E;F7Iok_2"}, "label": "a man in dress shirt"}]}
{"id": 344157, "image": "COCO_train2014_000000344157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the groom\"."}], "task": "refering", "answer_template": "The \"the groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 128, 129, 130, 151, 152, 153, 154, 172, 173, 175, 176, 177, 195, 196, 199, 200, 218, 219, 222, 223, 246], "bbox": [0.5066875, 0.21320833333333333, 0.732921875, 0.6530416666666667], "iscrowd": 0, "area": 8589.165899999996, "rle": {"size": [480, 640], "counts": "bRh43j09k<2hB6R=m0N3M3M4L6J4M2001O1N1O0000ORN`C^1Y=HfLlNQIk0l6[OeI1Y62Z40O1O1O2N1O1O1O1O1O010O10000O101O0O101O010O001O1O1QCWOe;j0XDYOg;j0UDWOk;o0mCTOR<o0iCTOV<n0eCUO[<i1O010000O1O1O1O100O1O2N3M2N3M7I4M2M2N2N2N100000O010O1001O2N1N10O101N1O1O1O100O2N1O1O1O101N4L3M4L4M2M4L3M3M3M3N1N2N2N2N2O0OO2N2N2N3NjC`Mm;^2PDhMo;e200001O100O1O22M3M2O2M2N20O1O1O2N1O2K4L4K9G<D<C<E:E;F7Iok_2"}, "label": "the groom"}]}
{"id": 344157, "image": "COCO_train2014_000000344157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl wearing a wedding dress , and cutting the cake\"."}], "task": "refering", "answer_template": "The \"girl wearing a wedding dress , and cutting the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 149, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 246, 263, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386], "bbox": [0.43537499999999996, 0.2685625, 0.8042812500000001, 0.9985208333333333], "iscrowd": 0, "area": 34794.150249999984, "rle": {"size": [480, 640], "counts": "bRS46i>6I5L2N3M3M2N1O100O1000O01000O010O1O001O1O001O10O01OiK@PJ`0i5JSJ6h52UJNi57UJIk5:QJFo5=nICS6?kI@U6d0gI]OY6e0dI[O\\6i0`IWOa6k0]ITOd6n0YISOf6o0XIQOi6P1VIoNj6R1UInNl6S1RInNm6T1RIkNo6U1PIlNo6U1PIlNP7T1PIlNP7T1oHmNQ7S1nHnNR7R1nHmNR7U1lHlNT7T1kHmNU7T1jHlNV7U1hHlNX7U1fHlNZ7U1eHjN\\7V1cHkN\\7W1bHiN_7X1_HiNa7X1^HgNc7Z1[HgNe7X2[GhMf8X4010O100O100O1O100O100O1O100O100O10dHRKf4m4TKZKl4e4oJbKP5^4iJiKX5U4bJSL_5k3^JXLd5e3ZJ_Lg5_3VJdLk5Z3SJjLn5]ObHj2_1kMQ6WOgHV3h0gMb6POjHk5W7QJlHP6Q82O2M2O1N3M2O1N2O2M2O1N3N1N2N3N1_IfHX5\\7eJXIi4j6RKZIl4h6QKYIP5h6lJZIT5i6hJXIY5i6cJYI]5i6`JXI`5k6[JWIf5j6WJWIi5P81O2N1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O3M2N3M3M2N3M3M2N3M3M3M2N3L4M2N3M3M2N1O2N1O2N1O1O2O0O2N2N4L3M3M3M3M4L3M3M3M3M4L3N2M2N0000003M6J5K5K6J5K5K6J5K6J5K5K4L1O001N2O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1XN]CQ1d<mN_CQ1b<mNaCQ1`<lNdCR1]<lNfCR1[<lNhCQ1Z<mNiCQ1X<lNlCR1U<lNmCf0UOEm=8UBFn=6VBHl=5WBIk=4WBKk=2XBKk=2d0MjTj1"}, "label": "girl wearing a wedding dress , and cutting the cake"}]}
{"id": 344157, "image": "COCO_train2014_000000344157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a cream colored wedding dress cutting cake\"."}], "task": "refering", "answer_template": "The \"woman in a cream colored wedding dress cutting cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 149, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 246, 263, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386], "bbox": [0.43537499999999996, 0.2685625, 0.8042812500000001, 0.9985208333333333], "iscrowd": 0, "area": 34794.150249999984, "rle": {"size": [480, 640], "counts": "bRS46i>6I5L2N3M3M2N1O100O1000O01000O010O1O001O1O001O10O01OiK@PJ`0i5JSJ6h52UJNi57UJIk5:QJFo5=nICS6?kI@U6d0gI]OY6e0dI[O\\6i0`IWOa6k0]ITOd6n0YISOf6o0XIQOi6P1VIoNj6R1UInNl6S1RInNm6T1RIkNo6U1PIlNo6U1PIlNP7T1PIlNP7T1oHmNQ7S1nHnNR7R1nHmNR7U1lHlNT7T1kHmNU7T1jHlNV7U1hHlNX7U1fHlNZ7U1eHjN\\7V1cHkN\\7W1bHiN_7X1_HiNa7X1^HgNc7Z1[HgNe7X2[GhMf8X4010O100O100O1O100O100O1O100O100O10dHRKf4m4TKZKl4e4oJbKP5^4iJiKX5U4bJSL_5k3^JXLd5e3ZJ_Lg5_3VJdLk5Z3SJjLn5]ObHj2_1kMQ6WOgHV3h0gMb6POjHk5W7QJlHP6Q82O2M2O1N3M2O1N2O2M2O1N3N1N2N3N1_IfHX5\\7eJXIi4j6RKZIl4h6QKYIP5h6lJZIT5i6hJXIY5i6cJYI]5i6`JXI`5k6[JWIf5j6WJWIi5P81O2N1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O3M2N3M3M2N3M3M2N3M3M3M2N3L4M2N3M3M2N1O2N1O2N1O1O2O0O2N2N4L3M3M3M3M4L3M3M3M3M4L3N2M2N0000003M6J5K5K6J5K5K6J5K6J5K5K4L1O001N2O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1XN]CQ1d<mN_CQ1b<mNaCQ1`<lNdCR1]<lNfCR1[<lNhCQ1Z<mNiCQ1X<lNlCR1U<lNmCf0UOEm=8UBFn=6VBHl=5WBIk=4WBKk=2XBKk=2d0MjTj1"}, "label": "woman in a cream colored wedding dress cutting cake"}]}
{"id": 385118, "image": "COCO_train2014_000000385118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skateboarding down a ramp\"."}], "task": "refering", "answer_template": "The \"a man skateboarding down a ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 145, 148, 149, 150, 171, 172, 173, 174, 197], "bbox": [0.270859375, 0.20793427230046949, 0.63434375, 0.5657276995305164], "iscrowd": 0, "area": 14680.719300000002, "rle": {"size": [426, 640], "counts": "QSX22Q=:K6I5L3M4L4M100O2N1O10001O001O010O00000001O01O101N10000OO2O00O001O001O1hDVOn9l0oEYOn9h0PF\\On9d0oEAn9`0oEDP:=lEGS:9jEJU:g1O1O0O2O1NmNQFnNn9P1ZFjNf9T1bFgN\\9Y1TGVNn8h1[1N101N2O0O2O1O0O2O1N110000O1000O10O10000000000O1000000000000O10O100000000000O10O100000000000000O1000000g0YO<Dd0\\O1O1O1O1O1O1O1O1O00001O00010O00001O000010O0001O001O00010O00QOdFPN\\9i1kFXNT9b1RG^No8Z1YGeNg8X1\\GiNc8W1^GhNc8W1]GjNb8V1^GjNb8V1_GiNb8W1]GjNb8W1^GhNc8Y1[GgNf8Z1XGgNg8[1XGdNi8]1UGcNk8^1TGcNl8^1RGbNo8_1PG`NQ9a1mF`NS9`1lF`NV9^1kFbNU9]1kFcNV9\\1jFdNW9[1iFfNX9X1iFgNY9W1gFjNY9U1gFkN[9S1fFlN\\9R1dFoN]9o0cFQOo9=RFBS:9mEHX:2hENY:1gE0X:OjE0U:1kEOU:0lE1S:OnE0Q:1oE0P:OQF1o9OQF1n9OTF1k9OUF1k9OUF2i9NYF1g9OYF1g9NZF3d9N\\F2d9N]F2a9N`F2`9N`F2`9MbF3\\9NdF2\\9NdF3[9LfF4Y9MhF2X9MiF4V9LjF4U9MlF3S9LnF5Q9KoF6o8JRG7m8ITG7j8JVG7i8HYG5i8KWG2k8NWGOk82TGKo85RGGP9:QGBR9>nF_OU9a0lF[OV9g0jFVOX9j0d100O2N100O100O101N1000000O1000000O1000000O1K5HTWQ3"}, "label": "a man skateboarding down a ramp"}]}
{"id": 385118, "image": "COCO_train2014_000000385118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a guy in gray pants that is skateboarding\"."}], "task": "refering", "answer_template": "The \"this is a guy in gray pants that is skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 145, 148, 149, 150, 171, 172, 173, 174, 197], "bbox": [0.270859375, 0.20793427230046949, 0.63434375, 0.5657276995305164], "iscrowd": 0, "area": 14680.719300000002, "rle": {"size": [426, 640], "counts": "QSX22Q=:K6I5L3M4L4M100O2N1O10001O001O010O00000001O01O101N10000OO2O00O001O001O1hDVOn9l0oEYOn9h0PF\\On9d0oEAn9`0oEDP:=lEGS:9jEJU:g1O1O0O2O1NmNQFnNn9P1ZFjNf9T1bFgN\\9Y1TGVNn8h1[1N101N2O0O2O1O0O2O1N110000O1000O10O10000000000O1000000000000O10O100000000000O10O100000000000000O1000000g0YO<Dd0\\O1O1O1O1O1O1O1O1O00001O00010O00001O000010O0001O001O00010O00QOdFPN\\9i1kFXNT9b1RG^No8Z1YGeNg8X1\\GiNc8W1^GhNc8W1]GjNb8V1^GjNb8V1_GiNb8W1]GjNb8W1^GhNc8Y1[GgNf8Z1XGgNg8[1XGdNi8]1UGcNk8^1TGcNl8^1RGbNo8_1PG`NQ9a1mF`NS9`1lF`NV9^1kFbNU9]1kFcNV9\\1jFdNW9[1iFfNX9X1iFgNY9W1gFjNY9U1gFkN[9S1fFlN\\9R1dFoN]9o0cFQOo9=RFBS:9mEHX:2hENY:1gE0X:OjE0U:1kEOU:0lE1S:OnE0Q:1oE0P:OQF1o9OQF1n9OTF1k9OUF1k9OUF2i9NYF1g9OYF1g9NZF3d9N\\F2d9N]F2a9N`F2`9N`F2`9MbF3\\9NdF2\\9NdF3[9LfF4Y9MhF2X9MiF4V9LjF4U9MlF3S9LnF5Q9KoF6o8JRG7m8ITG7j8JVG7i8HYG5i8KWG2k8NWGOk82TGKo85RGGP9:QGBR9>nF_OU9a0lF[OV9g0jFVOX9j0d100O2N100O100O101N1000000O1000000O1000000O1K5HTWQ3"}, "label": "this is a guy in gray pants that is skateboarding"}]}
{"id": 254046, "image": "COCO_train2014_000000254046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorbike of the # 99 motor bike\"."}], "task": "refering", "answer_template": "The \"the motorbike of the # 99 motor bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 202, 203, 204, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 263], "bbox": [0.406440281030445, 0.530171875, 0.6764402810304451, 0.7548281250000001], "iscrowd": 0, "area": 7693.653150000005, "rle": {"size": [640, 427], "counts": "gU]3`0Zc09J4K5L5K4K6K4NhNh]Oj0Vb0VOP^Of0oa0[OQ^Oe0ma0]OS^Ob0ma0^OT^Ob0ka0_OU^Oa0la0^OT^Ob0ma0]OS^Oc0na0WOW^Oh0ka0TOY^Ok0ha0ROZ^On0\\b0O1000O01O1O1O1O001N2NK3O302M3N1O2M3N2N2M3N2N2M3N2`NZNd@h1[?YNb@k1\\?VNa@m1^?TNP@0Gn1Y`0RNP@7]Oj1d`0nMo_O_2Q`0bMm_O_2S`0bMl_O^21eMX?Ne@_2JlM`?Ef@_2BUNg?]Of@\\3Y?dLe@^3Z?cLe@]3\\?bLc@`3\\?aLc@_3^?aLa@_3_?aL`@a3`?bL\\@^3d??10O00010O0001N10001N100O1O2N1O1O2M2O1N3N1O1N3N1N3M2M4M2YN^_OMe`02^_OJe`03__OHd`0[OV_O599g`0]OU_O95NQa0Jk^O66CZa07b^O4Wb0Kj]O3Wb0Mk]O1Ub0Om]OOTb00m]ONTb01o]OKTb04l001O1O100O1O1O10O0001O00ale2"}, "label": "the motorbike of the # 99 motor bike"}]}
{"id": 254046, "image": "COCO_train2014_000000254046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dirtbike with the number 99 on it\"."}], "task": "refering", "answer_template": "The \"a dirtbike with the number 99 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 202, 203, 204, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 263], "bbox": [0.406440281030445, 0.530171875, 0.6764402810304451, 0.7548281250000001], "iscrowd": 0, "area": 7693.653150000005, "rle": {"size": [640, 427], "counts": "gU]3`0Zc09J4K5L5K4K6K4NhNh]Oj0Vb0VOP^Of0oa0[OQ^Oe0ma0]OS^Ob0ma0^OT^Ob0ka0_OU^Oa0la0^OT^Ob0ma0]OS^Oc0na0WOW^Oh0ka0TOY^Ok0ha0ROZ^On0\\b0O1000O01O1O1O1O001N2NK3O302M3N1O2M3N2N2M3N2N2M3N2`NZNd@h1[?YNb@k1\\?VNa@m1^?TNP@0Gn1Y`0RNP@7]Oj1d`0nMo_O_2Q`0bMm_O_2S`0bMl_O^21eMX?Ne@_2JlM`?Ef@_2BUNg?]Of@\\3Y?dLe@^3Z?cLe@]3\\?bLc@`3\\?aLc@_3^?aLa@_3_?aL`@a3`?bL\\@^3d??10O00010O0001N10001N100O1O2N1O1O2M2O1N3N1O1N3N1N3M2M4M2YN^_OMe`02^_OJe`03__OHd`0[OV_O599g`0]OU_O95NQa0Jk^O66CZa07b^O4Wb0Kj]O3Wb0Mk]O1Ub0Om]OOTb00m]ONTb01o]OKTb04l001O1O100O1O1O10O0001O00ale2"}, "label": "a dirtbike with the number 99 on it"}]}
{"id": 155751, "image": "COCO_train2014_000000155751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane with the letters ec - kfi on the side of it\"."}], "task": "refering", "answer_template": "The \"an airplane with the letters ec - kfi on the side of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 70, 71, 72, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200], "bbox": [0.02790625, 0.18002906976744185, 0.94396875, 0.6992151162790697], "iscrowd": 0, "area": 39923.4187, "rle": {"size": [344, 640], "counts": "gV61g:3L3N3M1O00001N1000001O00000O2jLDnK<R4HjK8V4LfK4Z4OcK2\\42`KN`46\\KJd4:WKGi4=SKDl4?QKAo4c0mJ]OS5g0iJYOV5l0fJUOY5o0bJRO^5R1^JnNb5U1[JkNe5Y1WJhNh5\\1TJdNl5`1PJ`NP6d1lI\\NT6g1hI[NW6i1eIWN[6m1aISN_6Q2]IoMc6U2YIlMf6X2VIhMj6S30001O00000100O2N1O100O1O2N100O1O1O2O0O1O1O2O0O1O100O2N100O1O101N1O100O1O2O0O1O100O2N100O1O101N1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O101N1O1O1O001O010O001O001O010O001O001O1O010O001O001O010O001O001O1O010O001O001O01O000O100000001O00000000001O0000000000001O00000000001O0000000000001O00000O100000000000001O000000000000000000001O000000000000000000001O000000000O10000000001O000000000000000000001O000000000000000000001O0000000000000O101O00000000000000001O00000000000000001O0000000000000000001O000000000O1000001O00000000000000001O000000000000000000000000000000001O000000000O10000000000000000000000000001O0000000000000000000000000000000000000O101O000000000000000000001O00000000001O00000000001O00000000001O0000000O101O0000001O000000001O000000001O000000001O0000001O000000001O000000000O10O100O100O1O100O100O1O100N2M3N2N2N2M3N2N2N20000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000O1000O100000000000000000000000000000000000000000O100000001O001O001O00001N101O001O001O0O101O001O001O0O2O00001O001O0O2O001O001O1O1N2O1O1O1O1O1O1N101O001O00001N101N1O1O2N1N3N1O1N3N1O3M3L4MYm;"}, "label": "an airplane with the letters ec - kfi on the side of it"}]}
{"id": 155751, "image": "COCO_train2014_000000155751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane with a dark grey tail\"."}], "task": "refering", "answer_template": "The \"an airplane with a dark grey tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 70, 71, 72, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200], "bbox": [0.02790625, 0.18002906976744185, 0.94396875, 0.6992151162790697], "iscrowd": 0, "area": 39923.4187, "rle": {"size": [344, 640], "counts": "gV61g:3L3N3M1O00001N1000001O00000O2jLDnK<R4HjK8V4LfK4Z4OcK2\\42`KN`46\\KJd4:WKGi4=SKDl4?QKAo4c0mJ]OS5g0iJYOV5l0fJUOY5o0bJRO^5R1^JnNb5U1[JkNe5Y1WJhNh5\\1TJdNl5`1PJ`NP6d1lI\\NT6g1hI[NW6i1eIWN[6m1aISN_6Q2]IoMc6U2YIlMf6X2VIhMj6S30001O00000100O2N1O100O1O2N100O1O1O2O0O1O1O2O0O1O100O2N100O1O101N1O100O1O2O0O1O100O2N100O1O101N1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O101N1O1O1O001O010O001O001O010O001O001O1O010O001O001O010O001O001O1O010O001O001O01O000O100000001O00000000001O0000000000001O00000000001O0000000000001O00000O100000000000001O000000000000000000001O000000000000000000001O000000000O10000000001O000000000000000000001O000000000000000000001O0000000000000O101O00000000000000001O00000000000000001O0000000000000000001O000000000O1000001O00000000000000001O000000000000000000000000000000001O000000000O10000000000000000000000000001O0000000000000000000000000000000000000O101O000000000000000000001O00000000001O00000000001O00000000001O0000000O101O0000001O000000001O000000001O000000001O0000001O000000001O000000000O10O100O100O1O100O100O1O100N2M3N2N2N2M3N2N2N20000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000O1000O100000000000000000000000000000000000000000O100000001O001O001O00001N101O001O001O0O101O001O001O0O2O00001O001O0O2O001O001O1O1N2O1O1O1O1O1O1N101O001O00001N101N1O1O2N1N3N1O1N3N1O3M3L4MYm;"}, "label": "an airplane with a dark grey tail"}]}
{"id": 155751, "image": "COCO_train2014_000000155751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plane behind other\"."}], "task": "refering", "answer_template": "The \"plane behind other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140], "bbox": [0.0036249999999999998, 0.3893313953488372, 0.5918593750000001, 0.5780813953488373], "iscrowd": 0, "area": 6151.063299999998, "rle": {"size": [344, 640], "counts": "Wk08U9[10001O00000000000000001O00000000UOoFOQ91oFOQ90QGOo81QGOo80SGOm81SG0l80TG0l8OVG0j80VG0j80WGOi81WGOi80XG0h80YGOg81YGOg81ZGNf82ZGNf81[G0d80]GOc81]GOc81^GNb81_GOa81`GN`82`GN`82`GN`82aGM_82bGN^82cGM]83cGM]83cGN\\81fGNZ82fGNZ82gGMY83gGMY82hGNX8F[GL>=g8CYG=g8CZG<f8DZG<b9O01O01O001O00001O001O000010O01O001O00001O001O1O1O100O1Ob\\?0]c@2N101O001O010O001O1O001O010O001O001O001O010O001O001O001O010O001O001O001O01O0000000000001O0000000000000000000000000000000000O1000000000O100000000000000000000000O10000000000000000000000000000000000O100000O1000000000000000000000000000O10000000000000000000000000000000000000O100000000000000000000000000000000000000000000000001O000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000001O001O1O001O1O001O1N101O1O1O001O1O00kkg2"}, "label": "plane behind other"}]}
{"id": 155751, "image": "COCO_train2014_000000155751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane in the back of the full plane in view\"."}], "task": "refering", "answer_template": "The \"the plane in the back of the full plane in view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140], "bbox": [0.0036249999999999998, 0.3893313953488372, 0.5918593750000001, 0.5780813953488373], "iscrowd": 0, "area": 6151.063299999998, "rle": {"size": [344, 640], "counts": "Wk08U9[10001O00000000000000001O00000000UOoFOQ91oFOQ90QGOo81QGOo80SGOm81SG0l80TG0l8OVG0j80VG0j80WGOi81WGOi80XG0h80YGOg81YGOg81ZGNf82ZGNf81[G0d80]GOc81]GOc81^GNb81_GOa81`GN`82`GN`82`GN`82aGM_82bGN^82cGM]83cGM]83cGN\\81fGNZ82fGNZ82gGMY83gGMY82hGNX8F[GL>=g8CYG=g8CZG<f8DZG<b9O01O01O001O00001O001O000010O01O001O00001O001O1O1O100O1Ob\\?0]c@2N101O001O010O001O1O001O010O001O001O001O010O001O001O001O010O001O001O001O01O0000000000001O0000000000000000000000000000000000O1000000000O100000000000000000000000O10000000000000000000000000000000000O100000O1000000000000000000000000000O10000000000000000000000000000000000000O100000000000000000000000000000000000000000000000001O000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000001O001O1O001O1O001O1N101O1O1O001O1O00kkg2"}, "label": "the plane in the back of the full plane in view"}]}
{"id": 213107, "image": "COCO_train2014_000000213107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in black and orange baseball uniform swinging a bat\"."}], "task": "refering", "answer_template": "The \"a boy in black and orange baseball uniform swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 96, 97, 98, 99, 120, 121, 122, 143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.19190625, 0.2150351288056206, 0.46478125, 0.9858313817330211], "iscrowd": 0, "area": 24791.44855, "rle": {"size": [427, 640], "counts": "P]c13W=5K5K5L2M2N3M2N2fKWOoKMmKl0T8[OkKh0S4]OgKe0X4_OcKc0]4\\ObKf0]4ZObKi0]4VOcKk0\\4UOcKm0YMeNW6>_Lo0WMiNW67aLS1TMkNX62cLU1RMoNX6KeLX1PMSOW6EhL[1mLUOX6@kL\\1jLZOX6YOmL_1hL]OX6TOoLa1fL_OZ6oNoLS3n2mLRMU3l2kLTMX3j2gLUM\\3i2dLWM^3g2bLYMa3c2`L\\Md3a2\\L_Mh3\\2YLdMk3X2VLgMo3S2RLlMS4P2mKPNW4k1jKUN[4e1fKZN^4c1bK]N_4a1bK_N_4`1aK_Na4_1`KaNa4^1_KbNb4\\1_KcNc4\\1]KdNd4Z1]KfNd4Y1\\KfNf4X1[KhNf4V1[KjNf4U1ZKkNg4S1ZKnNf4Q1ZKPOf4n0[KROj5IVJ8j5FWJ;j5CVJ>j5@WJ`0j5_OUJc0l5ZOUJg0k5WOUJj0m5TOSJm0m5QOTJP1l5oNSJS1n5jNSJV1n5iNRJX1n5fNRJ\\1o5bNQJ_1o5_NQJb1P6]NPJd1Q6YNPJh1P6VNPJk1Q6TNoIm1S6oMnIR2P1^Lo2^1PLU2h0fL[3Q1nKX2f0jL]3l0lK[2BYLc0a0R4g0jK_2_O]Lb0?W4b0hKa2_O`L`0>[4>fKe2]OcL5DJh0P58eKh2\\OgL1f0P5HcKk2ZOkLOd0V5CaKo2XOmLNd0[5]O_KQ3XO^N\\5]N\\KV3VO`N_5XN[KW3VObNb5SNYK[3SOeNe5nMWKo5k4mIVKS6m4jIRKX6o4eIRK[6o51N101O1N101N101O1N101O1N101N2O0O2O1N101N2O001iKmHS2U7jMQIR2o6lMWIo1k6nMZIo1g6nM_Im1b6QNdIk1]6RNiIi1X6UNmIh1T6VNQJe1Q6ZNSJb1m5^NVJ^1l5bNWJZ1j5eNYJX1g5hN[JU1g5jN\\JU1c5jN_JU1`5kNbJT1^5kNeJS1[5lNgJT1W5mNkJQ1U5nNmJQ1S5oNoJP1o4POSKo0m4QOUKm0k4ROWKn0g4SOZKl0f4SO]Kk0b4VO_Kj0_4VOdKh0[4YOfKf0_4TOcKl0^4ROdKl0^4QOdKn0^4oNeKo0\\4POeKP1\\4mNfKQ1[4oNgKn0Y4ROiKl0X4TOiKj0W4VOlKg0U4XOmKf0S4[OoKb0R4]OPLa0P4@QL>P4ASL<m3ETL9m3FUL8k3HXL5i3KXL4g3L\\L1e3L_L2a3IfL5[3EmL8S3DSM:P8O2N1O2N1O2N1Ojc^4"}, "label": "a boy in black and orange baseball uniform swinging a bat"}]}
{"id": 213107, "image": "COCO_train2014_000000213107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little boy in a black uniform and orange socks\"."}], "task": "refering", "answer_template": "The \"little boy in a black uniform and orange socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 96, 97, 98, 99, 120, 121, 122, 143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.19190625, 0.2150351288056206, 0.46478125, 0.9858313817330211], "iscrowd": 0, "area": 24791.44855, "rle": {"size": [427, 640], "counts": "P]c13W=5K5K5L2M2N3M2N2fKWOoKMmKl0T8[OkKh0S4]OgKe0X4_OcKc0]4\\ObKf0]4ZObKi0]4VOcKk0\\4UOcKm0YMeNW6>_Lo0WMiNW67aLS1TMkNX62cLU1RMoNX6KeLX1PMSOW6EhL[1mLUOX6@kL\\1jLZOX6YOmL_1hL]OX6TOoLa1fL_OZ6oNoLS3n2mLRMU3l2kLTMX3j2gLUM\\3i2dLWM^3g2bLYMa3c2`L\\Md3a2\\L_Mh3\\2YLdMk3X2VLgMo3S2RLlMS4P2mKPNW4k1jKUN[4e1fKZN^4c1bK]N_4a1bK_N_4`1aK_Na4_1`KaNa4^1_KbNb4\\1_KcNc4\\1]KdNd4Z1]KfNd4Y1\\KfNf4X1[KhNf4V1[KjNf4U1ZKkNg4S1ZKnNf4Q1ZKPOf4n0[KROj5IVJ8j5FWJ;j5CVJ>j5@WJ`0j5_OUJc0l5ZOUJg0k5WOUJj0m5TOSJm0m5QOTJP1l5oNSJS1n5jNSJV1n5iNRJX1n5fNRJ\\1o5bNQJ_1o5_NQJb1P6]NPJd1Q6YNPJh1P6VNPJk1Q6TNoIm1S6oMnIR2P1^Lo2^1PLU2h0fL[3Q1nKX2f0jL]3l0lK[2BYLc0a0R4g0jK_2_O]Lb0?W4b0hKa2_O`L`0>[4>fKe2]OcL5DJh0P58eKh2\\OgL1f0P5HcKk2ZOkLOd0V5CaKo2XOmLNd0[5]O_KQ3XO^N\\5]N\\KV3VO`N_5XN[KW3VObNb5SNYK[3SOeNe5nMWKo5k4mIVKS6m4jIRKX6o4eIRK[6o51N101O1N101N101O1N101O1N101N2O0O2O1N101N2O001iKmHS2U7jMQIR2o6lMWIo1k6nMZIo1g6nM_Im1b6QNdIk1]6RNiIi1X6UNmIh1T6VNQJe1Q6ZNSJb1m5^NVJ^1l5bNWJZ1j5eNYJX1g5hN[JU1g5jN\\JU1c5jN_JU1`5kNbJT1^5kNeJS1[5lNgJT1W5mNkJQ1U5nNmJQ1S5oNoJP1o4POSKo0m4QOUKm0k4ROWKn0g4SOZKl0f4SO]Kk0b4VO_Kj0_4VOdKh0[4YOfKf0_4TOcKl0^4ROdKl0^4QOdKn0^4oNeKo0\\4POeKP1\\4mNfKQ1[4oNgKn0Y4ROiKl0X4TOiKj0W4VOlKg0U4XOmKf0S4[OoKb0R4]OPLa0P4@QL>P4ASL<m3ETL9m3FUL8k3HXL5i3KXL4g3L\\L1e3L_L2a3IfL5[3EmL8S3DSM:P8O2N1O2N1O2N1Ojc^4"}, "label": "little boy in a black uniform and orange socks"}]}
{"id": 82039, "image": "COCO_train2014_000000082039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white snowboard with a red and black pattern on it\"."}], "task": "refering", "answer_template": "The \"a white snowboard with a red and black pattern on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 209, 210, 211, 214, 215, 232, 233], "bbox": [0.6415599999999999, 0.7916533333333333, 1.0, 0.9567466666666666], "iscrowd": 0, "area": 2459.6313500000006, "rle": {"size": [375, 500], "counts": "[ke34b;4L5J5L4K2O10000O2O000O101N100000001O01O0000010O000001O01O000001O01O000001O01O0N2N2N2O2M2N2O1N2N3N1N2N2O1N3MWoP11jPoNg0YO<EO00000000000000001O000001O0000000001O00000000000000010O1O2N1JmE"}, "label": "a white snowboard with a red and black pattern on it"}]}
{"id": 82039, "image": "COCO_train2014_000000082039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the snowboard for a man with a flapped hat\"."}], "task": "refering", "answer_template": "The \"the snowboard for a man with a flapped hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 209, 210, 211, 214, 215, 232, 233], "bbox": [0.6415599999999999, 0.7916533333333333, 1.0, 0.9567466666666666], "iscrowd": 0, "area": 2459.6313500000006, "rle": {"size": [375, 500], "counts": "[ke34b;4L5J5L4K2O10000O2O000O101N100000001O01O0000010O000001O01O000001O01O000001O01O0N2N2N2O2M2N2O1N2N3N1N2N2O1N3MWoP11jPoNg0YO<EO00000000000000001O000001O0000000001O00000000000000010O1O2N1JmE"}, "label": "the snowboard for a man with a flapped hat"}]}
{"id": 376953, "image": "COCO_train2014_000000376953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue chair\"."}], "task": "refering", "answer_template": "The \"a blue chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 138, 139, 140, 141, 142, 149, 150, 153, 154, 161, 162, 164, 165, 166, 173, 174, 176, 177, 178, 198], "bbox": [0.4847447447447447, 0.55148, 0.9044444444444443, 0.9438799999999999], "iscrowd": 0, "area": 6201.5455999999995, "rle": {"size": [500, 333], "counts": "]P_22g=<^CO_<3WC4j<NlB:S=IcB?]=l00O10OE<B>M31N2O000M3J6J6K6I6lDQN]8U2aGnM^8R2^GSNa8j1[G_Nc8W1`GQO^81]EHV2:a8N\\E\\O\\2c0[80VHJn76UHDn7<TH^OP8a0THYOo7g0_3O1O100O1O100O1O100O100O100O100O100O100O100000O100010O000000001O0000000010O00000001O0000hBFV;;eDIY;9cDK[;7aDM];5cDK[;7eDIX;:iDEU;=kDCS;`0mD_OQ;c0oD]Oo:e0REZOm:g0SEYOl:h0UEWOj:j0VEVOh:l0YESOf:o0YEQOf:P1[EoNd:R1\\EnNb:T1^ElNa:V1^EjNa:W1^EjNa:W1_EjN^:Y1aEgN^:Z1bEfN]:[1cEeN\\:\\1dEdN\\:]1cEcN\\:^1dEbN[:_1eEaN[:_1eEaNZ:a1eE_NZ:b1fE^NZ:b1fE]N[:d1cE]N]:c1cE]N]:c1cE]N]:c1dE\\N\\:e1cE[N]:e1cE[N]:e1cE[N]:f1bEZN^:f1bEZN^:f1bEZN^:f1bEZN^:d1dE\\N\\:c1eE]N[:a1]D[Nk04h:_1_DfN`0KQ;]1bDjN:HU;\\1cDoN5DY;[1eDSOOA];[1eDXOJ\\Ob;Z1gD\\O^O_Ol;S1hDMZ;0iDNX;0jD0V;OlDOV;NoDOQ;OSEOm:OWEOj:M\\E0f:K_E4b:GSFGo94l2K6JQ[?"}, "label": "a blue chair"}]}
{"id": 376953, "image": "COCO_train2014_000000376953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark blue chair with a white cloth on one of the arms\"."}], "task": "refering", "answer_template": "The \"a dark blue chair with a white cloth on one of the arms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 138, 139, 140, 141, 142, 149, 150, 153, 154, 161, 162, 164, 165, 166, 173, 174, 176, 177, 178, 198], "bbox": [0.4847447447447447, 0.55148, 0.9044444444444443, 0.9438799999999999], "iscrowd": 0, "area": 6201.5455999999995, "rle": {"size": [500, 333], "counts": "]P_22g=<^CO_<3WC4j<NlB:S=IcB?]=l00O10OE<B>M31N2O000M3J6J6K6I6lDQN]8U2aGnM^8R2^GSNa8j1[G_Nc8W1`GQO^81]EHV2:a8N\\E\\O\\2c0[80VHJn76UHDn7<TH^OP8a0THYOo7g0_3O1O100O1O100O1O100O100O100O100O100O100O100000O100010O000000001O0000000010O00000001O0000hBFV;;eDIY;9cDK[;7aDM];5cDK[;7eDIX;:iDEU;=kDCS;`0mD_OQ;c0oD]Oo:e0REZOm:g0SEYOl:h0UEWOj:j0VEVOh:l0YESOf:o0YEQOf:P1[EoNd:R1\\EnNb:T1^ElNa:V1^EjNa:W1^EjNa:W1_EjN^:Y1aEgN^:Z1bEfN]:[1cEeN\\:\\1dEdN\\:]1cEcN\\:^1dEbN[:_1eEaN[:_1eEaNZ:a1eE_NZ:b1fE^NZ:b1fE]N[:d1cE]N]:c1cE]N]:c1cE]N]:c1dE\\N\\:e1cE[N]:e1cE[N]:e1cE[N]:f1bEZN^:f1bEZN^:f1bEZN^:f1bEZN^:d1dE\\N\\:c1eE]N[:a1]D[Nk04h:_1_DfN`0KQ;]1bDjN:HU;\\1cDoN5DY;[1eDSOOA];[1eDXOJ\\Ob;Z1gD\\O^O_Ol;S1hDMZ;0iDNX;0jD0V;OlDOV;NoDOQ;OSEOm:OWEOj:M\\E0f:K_E4b:GSFGo94l2K6JQ[?"}, "label": "a dark blue chair with a white cloth on one of the arms"}]}
{"id": 376953, "image": "COCO_train2014_000000376953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red couch with two seats\"."}], "task": "refering", "answer_template": "The \"a red couch with two seats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 124, 125, 126, 127, 128, 136, 137, 138, 139, 140, 148, 149, 150, 161, 173], "bbox": [0.34645645645645645, 0.5261399999999999, 0.7032132132132133, 0.81858], "iscrowd": 0, "area": 7047.482449999997, "rle": {"size": [500, 333], "counts": "o^h12X?<[AHg==mALQ>m0M3M3M201O1O10O01O1O1O0010001O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1iCkMm:V2cDYN];h1SDhNk;a20000010O000001O00O2UNaDWOg;h0\\DnNl;R1VDcNS<]1oCYNY<f1k001OM2J7N2N1O2N2N1OM4J7I6J7J6000O1000000O1000000O100001O00001O001O00001O00001O0000O10000O1000000000000001O01O000001O000M3M3M4M2M4L5L5G8H]R`1"}, "label": "a red couch with two seats"}]}
{"id": 376953, "image": "COCO_train2014_000000376953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three red chairs\"."}], "task": "refering", "answer_template": "The \"three red chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 124, 125, 126, 127, 128, 136, 137, 138, 139, 140, 148, 149, 150, 161, 173], "bbox": [0.34645645645645645, 0.5261399999999999, 0.7032132132132133, 0.81858], "iscrowd": 0, "area": 7047.482449999997, "rle": {"size": [500, 333], "counts": "o^h12X?<[AHg==mALQ>m0M3M3M201O1O10O01O1O1O0010001O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1iCkMm:V2cDYN];h1SDhNk;a20000010O000001O00O2UNaDWOg;h0\\DnNl;R1VDcNS<]1oCYNY<f1k001OM2J7N2N1O2N2N1OM4J7I6J7J6000O1000000O1000000O100001O00001O001O00001O00001O0000O10000O1000000000000001O01O000001O000M3M3M4M2M4L5L5G8H]R`1"}, "label": "three red chairs"}]}
{"id": 8320, "image": "COCO_train2014_000000008320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young zebra running\"."}], "task": "refering", "answer_template": "The \"a young zebra running\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 285, 286], "bbox": [0.35529687499999996, 0.3191569086651054, 0.538703125, 0.844824355971897], "iscrowd": 0, "area": 16448.493150000002, "rle": {"size": [427, 640], "counts": "[in21[=5K5K4K6]E^OB5]6=lIC^O<c60lIIWOb0j6FjIMROi0R7YOiI3kNP1Y7lNiI8eNW1_7aNmHA0m0ZO\\1g7VNjHINZ3U7lLjH2HY3\\7dLiH\\4U7dKgH`4Z7_KcHe4]7[K_Hi4`7a01000O0100000ON3M3L3N3L4M2MC>2N1N3N1N3N1N3N1O2M[ITKP5k4_JiK_5U4RJ]Ln5a3bIRM\\6o401O00001O001O001OH8G9H8G9H8G:G8H9K5J6JSIiKP5S4mJTLT5f3jJ`LX5[3dJmL]5m2aJYMa5b2\\JeMd5V2]JmMe5n1[JVNf5d1[J^Nh5]1XJcNk5^1QJcNR6]1jIcNY6^1cIcN`6]1\\IcNh6\\1TIfNn6[1nHeNV7[1fHfN]7Z1_HfNd7[1XHfNk7Z1QHfNS8Z1iGgNY8Z1cGfNa8o22M4M3L3N3L4L5L3M4K4M4L3L5L4L3L5L3M4L4L5L3L5K5L3L5L4K4L5L4K<D;F;Dajj3"}, "label": "a young zebra running"}]}
{"id": 8320, "image": "COCO_train2014_000000008320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby zebra\"."}], "task": "refering", "answer_template": "The \"baby zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 285, 286], "bbox": [0.35529687499999996, 0.3191569086651054, 0.538703125, 0.844824355971897], "iscrowd": 0, "area": 16448.493150000002, "rle": {"size": [427, 640], "counts": "[in21[=5K5K4K6]E^OB5]6=lIC^O<c60lIIWOb0j6FjIMROi0R7YOiI3kNP1Y7lNiI8eNW1_7aNmHA0m0ZO\\1g7VNjHINZ3U7lLjH2HY3\\7dLiH\\4U7dKgH`4Z7_KcHe4]7[K_Hi4`7a01000O0100000ON3M3L3N3L4M2MC>2N1N3N1N3N1N3N1O2M[ITKP5k4_JiK_5U4RJ]Ln5a3bIRM\\6o401O00001O001O001OH8G9H8G9H8G:G8H9K5J6JSIiKP5S4mJTLT5f3jJ`LX5[3dJmL]5m2aJYMa5b2\\JeMd5V2]JmMe5n1[JVNf5d1[J^Nh5]1XJcNk5^1QJcNR6]1jIcNY6^1cIcN`6]1\\IcNh6\\1TIfNn6[1nHeNV7[1fHfN]7Z1_HfNd7[1XHfNk7Z1QHfNS8Z1iGgNY8Z1cGfNa8o22M4M3L3N3L4L5L3M4K4M4L3L5L4L3L5L3M4L4L5L3L5K5L3L5L4K4L5L4K<D;F;Dajj3"}, "label": "baby zebra"}]}
{"id": 8320, "image": "COCO_train2014_000000008320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra showing his back side with his kid\"."}], "task": "refering", "answer_template": "The \"a zebra showing his back side with his kid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 63, 64, 65, 66, 67, 86, 87, 88, 89, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 271, 272, 273, 274, 292, 294, 296, 317, 318, 319, 341, 342], "bbox": [0.706, 0.08522248243559719, 0.9761562500000001, 0.9845667447306791], "iscrowd": 0, "area": 35539.27214999999, "rle": {"size": [427, 640], "counts": "[jl5<j<5O2N1jKMRK6^4=[KF`4b0[KA`4g0YK\\Ob4m0XKUOd4Q1VKSOf4R1gJ@U5f0WJLd5a0gILU6g3L4L4K5M1dIUI[6j5gISKZ6k4mI]JO6^3ElNf5nM]J0OU31jNa5WN^Ja33TN]5bN_JX36SNY5P3hJmLV5W3kJfLm4e3SKYLg4n3ZKoKa4X4`KeK[4b4fK\\KT4l4lKQKn3W5SLgJg3`5ZL]Ji3b5XL[Jj3f5VLXJl3g5ULVJm3j5TLSJo3l5RLRJo3o5QLnIR4Q6oKmIR4S6oKjIT4U6mKiIT4X6lKeIW4Z6jKcIX4]6iKaIY4^6gK`I[4a6eK]I\\4c6dK\\I^4c6cK[I^4e6cKYI_4g6`KWIb4i6_KUIc4j6]KUId4l6n0O0\\LhI5Y6]MiIo0:c1m5SMYJS1Lk1k5mLbJ`07b2W5kLjJ:3k2h6mL\\IS3i6eL[I[3Z801O00010O0O2NXOgL`GY3`8iL^GV3b8lLZGM@U3P8hMUHYO7T3X7`NfH`1W6FeI;]6EaI9b6H\\I7f6T34M5J6K6K5K5K4L3M2N2101N2OTNRH[Nk7c1VH`Ni7]1XHfNf7W1\\HlNc7Q1_HPOa7m0`HTOa7i0`HXOa7d0aH]O`7`0aHA`7;cHE^76eHJ]71fHO\\7LgH4[7GhH8[7CiH<Y7_OjHa0X7ZOkHf0W7UOlHk0V7POmHP1U7kNoHT1AWNd8h1[G[Nd8e1[G]Nd8c1[G^Ne8b1ZG`Ne8`1YGbNg8^1XGdNg8\\1XGfNg8Y1YGhNg8X1YGiNe8X1ZGiNf8W1YGkNf8U1ZGkN_7TOlHQ2DmN\\7J[HY18nNZ71YHP1>POU78WHi0c0QOR7a2mH`MP7c2oH_Mm6d2SI\\Mj6g2UI[Mg6h2XIYMd6j2\\IXM]6n2cIRMX6S3gIoLR6W3mIkLm5Z3RJgLk5\\3TJfLi5\\3VJeLf5^3YJeLV5k3hJWLV5j3iJYLT5i3jJYLS5i3mJYLo4i3PKZLl4i3RKYLl4h3SK[Li4V6L4K6J5K5L5K4M4L3L5L3L5L3M4K4M5K4K5L5K4K6K4K6J8F=D=B=DnQ6"}, "label": "a zebra showing his back side with his kid"}]}
{"id": 8320, "image": "COCO_train2014_000000008320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult zebra\"."}], "task": "refering", "answer_template": "The \"an adult zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 63, 64, 65, 66, 67, 86, 87, 88, 89, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 271, 272, 273, 274, 292, 294, 296, 317, 318, 319, 341, 342], "bbox": [0.706, 0.08522248243559719, 0.9761562500000001, 0.9845667447306791], "iscrowd": 0, "area": 35539.27214999999, "rle": {"size": [427, 640], "counts": "[jl5<j<5O2N1jKMRK6^4=[KF`4b0[KA`4g0YK\\Ob4m0XKUOd4Q1VKSOf4R1gJ@U5f0WJLd5a0gILU6g3L4L4K5M1dIUI[6j5gISKZ6k4mI]JO6^3ElNf5nM]J0OU31jNa5WN^Ja33TN]5bN_JX36SNY5P3hJmLV5W3kJfLm4e3SKYLg4n3ZKoKa4X4`KeK[4b4fK\\KT4l4lKQKn3W5SLgJg3`5ZL]Ji3b5XL[Jj3f5VLXJl3g5ULVJm3j5TLSJo3l5RLRJo3o5QLnIR4Q6oKmIR4S6oKjIT4U6mKiIT4X6lKeIW4Z6jKcIX4]6iKaIY4^6gK`I[4a6eK]I\\4c6dK\\I^4c6cK[I^4e6cKYI_4g6`KWIb4i6_KUIc4j6]KUId4l6n0O0\\LhI5Y6]MiIo0:c1m5SMYJS1Lk1k5mLbJ`07b2W5kLjJ:3k2h6mL\\IS3i6eL[I[3Z801O00010O0O2NXOgL`GY3`8iL^GV3b8lLZGM@U3P8hMUHYO7T3X7`NfH`1W6FeI;]6EaI9b6H\\I7f6T34M5J6K6K5K5K4L3M2N2101N2OTNRH[Nk7c1VH`Ni7]1XHfNf7W1\\HlNc7Q1_HPOa7m0`HTOa7i0`HXOa7d0aH]O`7`0aHA`7;cHE^76eHJ]71fHO\\7LgH4[7GhH8[7CiH<Y7_OjHa0X7ZOkHf0W7UOlHk0V7POmHP1U7kNoHT1AWNd8h1[G[Nd8e1[G]Nd8c1[G^Ne8b1ZG`Ne8`1YGbNg8^1XGdNg8\\1XGfNg8Y1YGhNg8X1YGiNe8X1ZGiNf8W1YGkNf8U1ZGkN_7TOlHQ2DmN\\7J[HY18nNZ71YHP1>POU78WHi0c0QOR7a2mH`MP7c2oH_Mm6d2SI\\Mj6g2UI[Mg6h2XIYMd6j2\\IXM]6n2cIRMX6S3gIoLR6W3mIkLm5Z3RJgLk5\\3TJfLi5\\3VJeLf5^3YJeLV5k3hJWLV5j3iJYLT5i3jJYLS5i3mJYLo4i3PKZLl4i3RKYLl4h3SK[Li4V6L4K6J5K5L5K4M4L3L5L3L5L3M4K4M5K4K5L5K4K6K4K6J8F=D=B=DnQ6"}, "label": "an adult zebra"}]}
{"id": 409732, "image": "COCO_train2014_000000409732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female wearing a blue shirt under her orange shirt and standing closest to the passenger door\"."}], "task": "refering", "answer_template": "The \"female wearing a blue shirt under her orange shirt and standing closest to the passenger door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 146, 167, 168, 169, 189, 190, 191, 192, 193, 212, 213, 214, 215, 237, 238, 260, 261, 283, 284, 306, 307, 329, 330, 352, 353], "bbox": [0.24504687500000003, 0.32583682008368203, 0.39609375, 0.9078661087866108], "iscrowd": 0, "area": 13000.399700000002, "rle": {"size": [478, 640], "counts": "WaY21l>1O100O1O1O1N2@`0_Oa0@`0E;E;F:L4O100O100O100O1O100O100O100O100O100000001O000000000000010O00000gDYM[:j3K5L4L5K4M3L;E_2aM8HO10000000000O2O1O001N1O2N101N1O2N1O2O1N1O1O100[NbHhK_7Q4QIeKP7U4^IbKc6X4lI_KT6[4ZJ\\Kg5]4iJYKX5a4f2J6J6iNW1UOk0N2N2N2K5K6Jf0ZOf0ZOVnc5"}, "label": "female wearing a blue shirt under her orange shirt and standing closest to the passenger door"}]}
{"id": 409732, "image": "COCO_train2014_000000409732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue undershirt\"."}], "task": "refering", "answer_template": "The \"woman in blue undershirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 146, 167, 168, 169, 189, 190, 191, 192, 193, 212, 213, 214, 215, 237, 238, 260, 261, 283, 284, 306, 307, 329, 330, 352, 353], "bbox": [0.24504687500000003, 0.32583682008368203, 0.39609375, 0.9078661087866108], "iscrowd": 0, "area": 13000.399700000002, "rle": {"size": [478, 640], "counts": "WaY21l>1O100O1O1O1N2@`0_Oa0@`0E;E;F:L4O100O100O100O1O100O100O100O100O100000001O000000000000010O00000gDYM[:j3K5L4L5K4M3L;E_2aM8HO10000000000O2O1O001N1O2N101N1O2N1O2O1N1O1O100[NbHhK_7Q4QIeKP7U4^IbKc6X4lI_KT6[4ZJ\\Kg5]4iJYKX5a4f2J6J6iNW1UOk0N2N2N2K5K6Jf0ZOf0ZOVnc5"}, "label": "woman in blue undershirt"}]}
{"id": 409732, "image": "COCO_train2014_000000409732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with blonde hair and black pants holding a tray with orange cups\"."}], "task": "refering", "answer_template": "The \"a woman with blonde hair and black pants holding a tray with orange cups\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 346, 347], "bbox": [0.01178125, 0.3006694560669456, 0.139625, 0.9042887029288703], "iscrowd": 0, "area": 17112.37865, "rle": {"size": [478, 640], "counts": "UR4<\\>?B=B>\\LSOaH[1l6^OaHo0T7]O]HQ1X7VO`HS1X7TOaHQ1[7UO^HQ1_7SOZHR1d7ROUHS1h7T3N2M3M3L4MO1NVOiHRJT7P6TIhIj6[6\\I_Ib6b6j03N101N1O2N3M2N3M3M2N3M3M3K5H8I7H]I_JY4S810O001O1O0010O01O001O0010O0000002N3M2THkIk6W6QInIl6U6PInIn6T6nHPJP7S6kHQJS7Q6iHSJU7P6gHTJV7l6M3K5K6WKRHY2T8`MRHZ2U8^MRH]2R8]MTHU2Z8cMoGR2[8fMnGT2W8gMnGV2U8gMTHP2n7mM`Hd1e7XN\\Ie0j6XO`I:g6B]4MeoP8"}, "label": "a woman with blonde hair and black pants holding a tray with orange cups"}]}
{"id": 409732, "image": "COCO_train2014_000000409732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on the left with grey undershirt and blond hair\"."}], "task": "refering", "answer_template": "The \"woman on the left with grey undershirt and blond hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 346, 347], "bbox": [0.01178125, 0.3006694560669456, 0.139625, 0.9042887029288703], "iscrowd": 0, "area": 17112.37865, "rle": {"size": [478, 640], "counts": "UR4<\\>?B=B>\\LSOaH[1l6^OaHo0T7]O]HQ1X7VO`HS1X7TOaHQ1[7UO^HQ1_7SOZHR1d7ROUHS1h7T3N2M3M3L4MO1NVOiHRJT7P6TIhIj6[6\\I_Ib6b6j03N101N1O2N3M2N3M3M2N3M3M3K5H8I7H]I_JY4S810O001O1O0010O01O001O0010O0000002N3M2THkIk6W6QInIl6U6PInIn6T6nHPJP7S6kHQJS7Q6iHSJU7P6gHTJV7l6M3K5K6WKRHY2T8`MRHZ2U8^MRH]2R8]MTHU2Z8cMoGR2[8fMnGT2W8gMnGV2U8gMTHP2n7mM`Hd1e7XN\\Ie0j6XO`I:g6B]4MeoP8"}, "label": "woman on the left with grey undershirt and blond hair"}]}
{"id": 344196, "image": "COCO_train2014_000000344196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown teddy bear holding a red heart\"."}], "task": "refering", "answer_template": "The \"a brown teddy bear holding a red heart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 146, 147, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215], "bbox": [0.184015625, 0.11654205607476636, 0.41890625000000004, 0.6579439252336449], "iscrowd": 0, "area": 17679.898750000008, "rle": {"size": [428, 640], "counts": "kaa1<o<3N1TKHaL;]3MZL6c33ULNi35TLNj32UL1h31WL1f30YL2e30ZL2d3N[L3d3OZL3d3N\\L2d3OZL2e30ZL1d30\\L0d31[L0c31]LOb33]LMb34^LMa34^LLa35_LKa36^LKa35_LKa37]LJa38^LHb3:\\LFd3;[LFd3<ZLDf3=aJYN6Z1Y5?\\JZN;X1X5cNTIk1T1]N`0U1Y5`0SJ^Ne0R1V5c0PJ_Ni0n0W5d0lIaNk0m0X5e0hIaNl0m0\\5d0eIaNk0k0c5g0]IaNn0f0i5j0UIcNS1=l5S1lHdNX16m5[2TJdMk5]2UJbMl5^2UJaMj5a2VJ]Mj5d2VJ\\Mj5d2WJ[Mi5e2XJYMh5i2XJVMh5j2XJUMi5k2XJUMf5l2[JSMe5m2[JSMe5l2]JSMc5m2^JSM`5n2aJQM_5o2aJQM_5o2aJQM^5P3cJPM\\5P3dJPM\\5o2eJQMZ5o2hJQMW5n2jJRMV5l2lJTMS5l2oJSMQ5l2PKTMP5j2RKVMn4i2TKVMk4j2VKUMk4i2WKWMi4h2YKWMf4i2[KVMf4i2[KWMe4i2\\KVMd4j2\\KUMe4k2[KUMe4k2[KUMe4k2[KSMg4m2YKiLQ5W3oJbLW5_3iJbLV5^3jJbLV5^3jJbLV5^3jJbLV5^3jJbLV5^3jJbL\\N3o5[3eKbLSN<W6S3fKaLiMf0a6i2fK\\MZ4d2fK\\MZ4d2fK\\MZ4d2fK\\MY4e2gK[MY4f2fKZMZ4g2eKYM[4g2eKYM[4h2dKXM[4j2dKVM\\4j2dKVM\\4k2cKUM]4l2bKTM]4n2bKSM]4m2cKSM]4n2bKRM^4o2`KRM`4n2^KTMa4a0hId0c1nNe4>UJ8R1]Oi4;]J1f0Gm47`J0`0Lo45dJM91S51hJLN9Z5KjJKCb0c5BnJIWOm0k5ZOoJImNU1U6QOnJKdN\\1_6gNoJc2R5\\MnJb2V5[MlJc2V5\\MjJb2Y5]MhJ`2[5^MfJa2]5]McJa2b5[M`Jb2d5\\M\\Jc2g5ZM[Jc2j5YMWJe2m5XMUJf2o5VMRJh2R6UMPJh2U6TMlIk2W6RMjIl2[6PMgIm2]8O2M2N2O1N3M2O1N6J6J7G8H9G8G9H9G8GeYk4"}, "label": "a brown teddy bear holding a red heart"}]}
{"id": 344196, "image": "COCO_train2014_000000344196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small ceramic bear holding a heart\"."}], "task": "refering", "answer_template": "The \"a small ceramic bear holding a heart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 146, 147, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215], "bbox": [0.184015625, 0.11654205607476636, 0.41890625000000004, 0.6579439252336449], "iscrowd": 0, "area": 17679.898750000008, "rle": {"size": [428, 640], "counts": "kaa1<o<3N1TKHaL;]3MZL6c33ULNi35TLNj32UL1h31WL1f30YL2e30ZL2d3N[L3d3OZL3d3N\\L2d3OZL2e30ZL1d30\\L0d31[L0c31]LOb33]LMb34^LMa34^LLa35_LKa36^LKa35_LKa37]LJa38^LHb3:\\LFd3;[LFd3<ZLDf3=aJYN6Z1Y5?\\JZN;X1X5cNTIk1T1]N`0U1Y5`0SJ^Ne0R1V5c0PJ_Ni0n0W5d0lIaNk0m0X5e0hIaNl0m0\\5d0eIaNk0k0c5g0]IaNn0f0i5j0UIcNS1=l5S1lHdNX16m5[2TJdMk5]2UJbMl5^2UJaMj5a2VJ]Mj5d2VJ\\Mj5d2WJ[Mi5e2XJYMh5i2XJVMh5j2XJUMi5k2XJUMf5l2[JSMe5m2[JSMe5l2]JSMc5m2^JSM`5n2aJQM_5o2aJQM_5o2aJQM^5P3cJPM\\5P3dJPM\\5o2eJQMZ5o2hJQMW5n2jJRMV5l2lJTMS5l2oJSMQ5l2PKTMP5j2RKVMn4i2TKVMk4j2VKUMk4i2WKWMi4h2YKWMf4i2[KVMf4i2[KWMe4i2\\KVMd4j2\\KUMe4k2[KUMe4k2[KUMe4k2[KSMg4m2YKiLQ5W3oJbLW5_3iJbLV5^3jJbLV5^3jJbLV5^3jJbLV5^3jJbLV5^3jJbL\\N3o5[3eKbLSN<W6S3fKaLiMf0a6i2fK\\MZ4d2fK\\MZ4d2fK\\MZ4d2fK\\MY4e2gK[MY4f2fKZMZ4g2eKYM[4g2eKYM[4h2dKXM[4j2dKVM\\4j2dKVM\\4k2cKUM]4l2bKTM]4n2bKSM]4m2cKSM]4n2bKRM^4o2`KRM`4n2^KTMa4a0hId0c1nNe4>UJ8R1]Oi4;]J1f0Gm47`J0`0Lo45dJM91S51hJLN9Z5KjJKCb0c5BnJIWOm0k5ZOoJImNU1U6QOnJKdN\\1_6gNoJc2R5\\MnJb2V5[MlJc2V5\\MjJb2Y5]MhJ`2[5^MfJa2]5]McJa2b5[M`Jb2d5\\M\\Jc2g5ZM[Jc2j5YMWJe2m5XMUJf2o5VMRJh2R6UMPJh2U6TMlIk2W6RMjIl2[6PMgIm2]8O2M2N2O1N3M2O1N6J6J7G8H9G8G9H9G8GeYk4"}, "label": "a small ceramic bear holding a heart"}]}
{"id": 344196, "image": "COCO_train2014_000000344196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right half of a teddy bear porcelain garden structure , hugging another bear holding a red heart\"."}], "task": "refering", "answer_template": "The \"the right half of a teddy bear porcelain garden structure , hugging another bear holding a red heart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 148, 149, 171, 172, 193, 194, 195, 216], "bbox": [0.35315625, 0.08539719626168224, 0.524484375, 0.613481308411215], "iscrowd": 0, "area": 13371.859149999997, "rle": {"size": [428, 640], "counts": "Zbn26P=:H9J5K6J3M2M4N1N2N3N1N3N2M3M100O100O1O100O00100O100O1O100N2M3M3N2M3M300oG`MV5`2iJfMQ5[2nJkMm4T2SKQNi4o1VKWNe4i1[KVNe4k1ZKVNf4j1YKWNg4i1XKXNh4h1WKYNh4g1XKYNi4g1VKZNj4f1TK\\Nl4d1SK]Nl4d1SK]Nm4c1QK^NP5b1oJ_Ne4POZI`2P2bNd4ROYI]2P2dNd4RO\\IZ2m1gNd4RO_IW2j1iNg4QO_IV2g1lNi4PO^IU2d1POn4kN^IW2]1SOT5gN_IY2U1UO\\5bN_I\\2m0VOd5_N^I^2f0YOk5`1mIeNS6^1eIgNZ6U410O010O01O2O1N2O6I7J1N101N2O1N2N3N1N3N1N2O2M2O2M2O1N2N101N2O0O2O1UMlG;V8CmG8W8ElG7Y8FjG6Z8GhG5]8GgG4]9gNhFV1d9YNeFb1e:H9D;F;ZOYmn3"}, "label": "the right half of a teddy bear porcelain garden structure , hugging another bear holding a red heart"}]}
{"id": 344196, "image": "COCO_train2014_000000344196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear with out the heart\"."}], "task": "refering", "answer_template": "The \"teddy bear with out the heart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 148, 149, 171, 172, 193, 194, 195, 216], "bbox": [0.35315625, 0.08539719626168224, 0.524484375, 0.613481308411215], "iscrowd": 0, "area": 13371.859149999997, "rle": {"size": [428, 640], "counts": "Zbn26P=:H9J5K6J3M2M4N1N2N3N1N3N2M3M100O100O1O100O00100O100O1O100N2M3M3N2M3M300oG`MV5`2iJfMQ5[2nJkMm4T2SKQNi4o1VKWNe4i1[KVNe4k1ZKVNf4j1YKWNg4i1XKXNh4h1WKYNh4g1XKYNi4g1VKZNj4f1TK\\Nl4d1SK]Nl4d1SK]Nm4c1QK^NP5b1oJ_Ne4POZI`2P2bNd4ROYI]2P2dNd4RO\\IZ2m1gNd4RO_IW2j1iNg4QO_IV2g1lNi4PO^IU2d1POn4kN^IW2]1SOT5gN_IY2U1UO\\5bN_I\\2m0VOd5_N^I^2f0YOk5`1mIeNS6^1eIgNZ6U410O010O01O2O1N2O6I7J1N101N2O1N2N3N1N3N1N2O2M2O2M2O1N2N101N2O0O2O1UMlG;V8CmG8W8ElG7Y8FjG6Z8GhG5]8GgG4]9gNhFV1d9YNeFb1e:H9D;F;ZOYmn3"}, "label": "teddy bear with out the heart"}]}
{"id": 581766, "image": "COCO_train2014_000000581766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"urinal on the right\"."}], "task": "refering", "answer_template": "The \"urinal on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 69, 87, 103, 104, 105, 106, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 176, 177], "bbox": [0.74494, 0.20224, 0.9573, 0.7415733333333334], "iscrowd": 0, "area": 11540.7888, "rle": {"size": [375, 500], "counts": "i]X41c;6K6I6J7I6J7I7J5J7I6J7I6J7J4K4N2O1N2N2O1N2O0O2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N101N2N10O0100O010O10OfN[1D<E;M210000O100O1N2M3N2000000O1YLWJ`1i5VNaJj1_5lMlJS2T5bMXK]2h4YMcKe2^4PMmKP3S4fLXLY3T6O0O2O1O1O0O2O1O1N101O1N101O1N2M2O2N2N2M4M2N2N2M4L3L4L4L5K4L4L4L5G8A?@i`7"}, "label": "urinal on the right"}]}
{"id": 581766, "image": "COCO_train2014_000000581766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a urinal on the right\"."}], "task": "refering", "answer_template": "The \"a urinal on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 69, 87, 103, 104, 105, 106, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 176, 177], "bbox": [0.74494, 0.20224, 0.9573, 0.7415733333333334], "iscrowd": 0, "area": 11540.7888, "rle": {"size": [375, 500], "counts": "i]X41c;6K6I6J7I6J7I7J5J7I6J7I6J7J4K4N2O1N2N2O1N2O0O2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N101N2N10O0100O010O10OfN[1D<E;M210000O100O1N2M3N2000000O1YLWJ`1i5VNaJj1_5lMlJS2T5bMXK]2h4YMcKe2^4PMmKP3S4fLXLY3T6O0O2O1O1O0O2O1O1N101O1N101O1N2M2O2N2N2M4M2N2N2M4L3L4L4L5K4L4L4L5G8A?@i`7"}, "label": "a urinal on the right"}]}
{"id": 581766, "image": "COCO_train2014_000000581766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the urinal on the left\"."}], "task": "refering", "answer_template": "The \"the urinal on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 163, 164, 165, 166], "bbox": [0.038939999999999995, 0.39898666666666666, 0.25564, 0.7578933333333334], "iscrowd": 0, "area": 11487.5681, "rle": {"size": [375, 500], "counts": "`U72R;f0_Oa0L3M4L4L4L3L5L4L4L3M3M3M2M4M3M2N3M2N2M3N2N2N2N2N2N2N1O2O0O1O2N1O1O2N101O00001O000000O1O1O1O100O1O1O1O12N5K4L1O001O001O1O000000000000O1000000000000O10001O1O1O1O1O1N2N2O1N2O1N2N2O2M2N2O1N3KcGlL^8P37K5L5K4L5K5K6I6J6I8H<Ef0YOjPX4"}, "label": "the urinal on the left"}]}
{"id": 581766, "image": "COCO_train2014_000000581766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left most urinal\"."}], "task": "refering", "answer_template": "The \"the left most urinal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 163, 164, 165, 166], "bbox": [0.038939999999999995, 0.39898666666666666, 0.25564, 0.7578933333333334], "iscrowd": 0, "area": 11487.5681, "rle": {"size": [375, 500], "counts": "`U72R;f0_Oa0L3M4L4L4L3L5L4L4L3M3M3M2M4M3M2N3M2N2M3N2N2N2N2N2N2N1O2O0O1O2N1O1O2N101O00001O000000O1O1O1O100O1O1O1O12N5K4L1O001O001O1O000000000000O1000000000000O10001O1O1O1O1O1N2N2O1N2O1N2N2O2M2N2O1N3KcGlL^8P37K5L5K4L5K5K6I6J6I8H<Ef0YOjPX4"}, "label": "the left most urinal"}]}
{"id": 581766, "image": "COCO_train2014_000000581766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brightly lit urinal\"."}], "task": "refering", "answer_template": "The \"a brightly lit urinal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 136, 151, 152, 153, 154, 170, 171], "bbox": [0.4045, 0.3887733333333333, 0.59494, 0.7617866666666666], "iscrowd": 0, "area": 9043.586, "rle": {"size": [375, 500], "counts": "YUZ23^;9G9G9K5K5L4K5L4K5K5L4K5L4K5M1O2N2M2O2N1N3N1N3N1O2M3N1O2M2O2M2O1O1N2O1N2O1O1N2O1N2O1O1MYL^H^3a7]LdHd3g71O1O1O1O1O000O0E;1O2O0O2N1O2N101N1O2N1O2M2O2M2N3M3N1N3M3M3M2O2M3M3M3N1N3M3L4L4L3M4L4L4K5L3M4L4L4LcUZ2"}, "label": "a brightly lit urinal"}]}
{"id": 581766, "image": "COCO_train2014_000000581766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"urinal in middle of other urinals\"."}], "task": "refering", "answer_template": "The \"urinal in middle of other urinals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 136, 151, 152, 153, 154, 170, 171], "bbox": [0.4045, 0.3887733333333333, 0.59494, 0.7617866666666666], "iscrowd": 0, "area": 9043.586, "rle": {"size": [375, 500], "counts": "YUZ23^;9G9G9K5K5L4K5L4K5K5L4K5L4K5M1O2N2M2O2N1N3N1N3N1O2M3N1O2M2O2M2O1O1N2O1N2O1O1N2O1N2O1O1MYL^H^3a7]LdHd3g71O1O1O1O1O000O0E;1O2O0O2N1O2N101N1O2N1O2M2O2M2N3M3N1N3M3M3M2O2M3M3M3N1N3M3L4L4L3M4L4L4K5L3M4L4L4LcUZ2"}, "label": "urinal in middle of other urinals"}]}
{"id": 163975, "image": "COCO_train2014_000000163975.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white wash basin\"."}], "task": "refering", "answer_template": "The \"white wash basin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.47778124999999994, 0.7350588235294118, 1.0, 0.9870117647058823], "iscrowd": 0, "area": 27839.554900000006, "rle": {"size": [425, 640], "counts": "Q]o34o<6J6J6J6J6J6J6J6O1O1O1O1O100O1O1O1O100O1O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O11O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O0UOPF"}, "label": "white wash basin"}]}
{"id": 163975, "image": "COCO_train2014_000000163975.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color real sink not the reflected on the mirror\"."}], "task": "refering", "answer_template": "The \"white color real sink not the reflected on the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.47778124999999994, 0.7350588235294118, 1.0, 0.9870117647058823], "iscrowd": 0, "area": 27839.554900000006, "rle": {"size": [425, 640], "counts": "Q]o34o<6J6J6J6J6J6J6J6O1O1O1O1O100O1O1O1O100O1O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O11O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O0UOPF"}, "label": "white color real sink not the reflected on the mirror"}]}
{"id": 131215, "image": "COCO_train2014_000000131215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and blue jeans watching television\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and blue jeans watching television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.54540625, 0.2440566037735849, 0.95003125, 0.984551886792453], "iscrowd": 0, "area": 44295.862099999984, "rle": {"size": [424, 640], "counts": "Ql`41U=3N1O2O1O001O1O001O001O001O001O001O001O001O00001O001O001oC@X;`0dDD\\;<`DH`;9[DKe;5WDOi;h00000L4D<E;D<I7N2N2N2M3N2N2N2N2N2N2M3M3O1N2O1N2O1N2O1N2O100001O001O00001O001O00001O001O001O00\\KTMjNn2S1VMjNl2R1ZMiNi2T1[MhNh2T1]MiNa2X1cMdN\\2\\1iM`NV2c1lMYNS2i1PNTNm1P2UNlMj1V2ZNeMf1\\2]NaMf1]2\\N_Mh1^2[N^Mh1a2ZN\\Mi1a2ZN[Mj1c2XNYMl1d2VNZMm1d2UNXMn1g2SNVMQ2i2PNTMS2k2mMRMW2m2jMoLY2Q3gMmL\\2R3dMkL`2T3aMhLc2W3]MgLe2Y3\\McLh2\\3XMaLl2^3UM_Ln2`3RM^LQ3a3PM]LQ3b3PM\\LS3c3mL\\LU3c3lLZLW3e3iLZLX3f3V3O1O1O000000000000000O100000001O000000000000000000000O11O2N1O2N2N101N2N1O2N2N2N1O2N2N1O2N2N2N1]H`K\\6b4ZIfKf6\\4PImKo6T4gHVLY7l401N101O1O001O1N101O001O1O0O2O1O001O001N2O001O1O001N102N1O2N2N1N3N1O2N1O2N2M2O2N1O2N2M2O2N1O2N1O2M3N1O2N1O2N1N3N2N1N3M2N3M2N3M3M2N3M3M5K5K6J5K6J5K6J5K6J5K5K6J5K6J5K6J5K6J5K5K6J5K6J5K6Jbl<"}, "label": "a man in a white shirt and blue jeans watching television"}]}
{"id": 131215, "image": "COCO_train2014_000000131215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing white t - shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing white t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.54540625, 0.2440566037735849, 0.95003125, 0.984551886792453], "iscrowd": 0, "area": 44295.862099999984, "rle": {"size": [424, 640], "counts": "Ql`41U=3N1O2O1O001O1O001O001O001O001O001O001O001O00001O001O001oC@X;`0dDD\\;<`DH`;9[DKe;5WDOi;h00000L4D<E;D<I7N2N2N2M3N2N2N2N2N2N2M3M3O1N2O1N2O1N2O1N2O100001O001O00001O001O00001O001O001O00\\KTMjNn2S1VMjNl2R1ZMiNi2T1[MhNh2T1]MiNa2X1cMdN\\2\\1iM`NV2c1lMYNS2i1PNTNm1P2UNlMj1V2ZNeMf1\\2]NaMf1]2\\N_Mh1^2[N^Mh1a2ZN\\Mi1a2ZN[Mj1c2XNYMl1d2VNZMm1d2UNXMn1g2SNVMQ2i2PNTMS2k2mMRMW2m2jMoLY2Q3gMmL\\2R3dMkL`2T3aMhLc2W3]MgLe2Y3\\McLh2\\3XMaLl2^3UM_Ln2`3RM^LQ3a3PM]LQ3b3PM\\LS3c3mL\\LU3c3lLZLW3e3iLZLX3f3V3O1O1O000000000000000O100000001O000000000000000000000O11O2N1O2N2N101N2N1O2N2N2N1O2N2N1O2N2N2N1]H`K\\6b4ZIfKf6\\4PImKo6T4gHVLY7l401N101O1O001O1N101O001O1O0O2O1O001O001N2O001O1O001N102N1O2N2N1N3N1O2N1O2N2M2O2N1O2N2M2O2N1O2N1O2M3N1O2N1O2N1N3N2N1N3M2N3M2N3M3M2N3M3M5K5K6J5K6J5K6J5K6J5K5K6J5K6J5K6J5K6J5K5K6J5K6J5K6Jbl<"}, "label": "a man wearing white t - shirt"}]}
{"id": 303247, "image": "COCO_train2014_000000303247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open book without a yellow cover\"."}], "task": "refering", "answer_template": "The \"an open book without a yellow cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.038609375, 0.6647907949790794, 0.64953125, 0.9883891213389121], "iscrowd": 0, "area": 43301.61245, "rle": {"size": [478, 640], "counts": "]S<1l>1O1O2N1O1O1N3N1O1O1O1O2N1O1O1O2N1O1O1N3N1O1O1O1O2N1O1O1O2N1O1O1O1N3N1O1O1O2N1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O2N1O1O1O1O2N1N2O1O2N1O1O1O1O2N1O1O1O2N1O1N2O2N1O1O1O1O2N1O1O1O2N1O1O1N3O0001O00001O00001O00001O000000O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O100000000000000001O00000000000000000000001O00000000000000000000001O000O1000000000000000001O000000000000000000001O00000000000000001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O1O4L5K5K00O1O100O100O1O100O1000000000000000000000000O1000O10000000000000000000000000000000000000000000000000000000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O001O1O1N101O1O1O1O001O1O1O1O001O1O1O1O001O00001O0000001O000K5I7I8H7H8I7I7I8H7I7I7HYVX3"}, "label": "an open book without a yellow cover"}]}
{"id": 303247, "image": "COCO_train2014_000000303247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"open book on the table in front of bananas and the other book\"."}], "task": "refering", "answer_template": "The \"open book on the table in front of bananas and the other book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.038609375, 0.6647907949790794, 0.64953125, 0.9883891213389121], "iscrowd": 0, "area": 43301.61245, "rle": {"size": [478, 640], "counts": "]S<1l>1O1O2N1O1O1N3N1O1O1O1O2N1O1O1O2N1O1O1N3N1O1O1O1O2N1O1O1O2N1O1O1O1N3N1O1O1O2N1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O2N1O1O1O1O2N1N2O1O2N1O1O1O1O2N1O1O1O2N1O1N2O2N1O1O1O1O2N1O1O1O2N1O1O1N3O0001O00001O00001O00001O000000O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O100000000000000001O00000000000000000000001O00000000000000000000001O000O1000000000000000001O000000000000000000001O00000000000000001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O1O4L5K5K00O1O100O100O1O100O1000000000000000000000000O1000O10000000000000000000000000000000000000000000000000000000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O001O1O1N101O1O1O1O001O1O1O1O001O1O1O1O001O00001O0000001O000K5I7I8H7H8I7I7I8H7I7I7HYVX3"}, "label": "open book on the table in front of bananas and the other book"}]}
{"id": 303247, "image": "COCO_train2014_000000303247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow open book\"."}], "task": "refering", "answer_template": "The \"the yellow open book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 320, 341, 342, 343], "bbox": [0.43376562500000004, 0.5841213389121338, 0.94709375, 0.8423640167364016], "iscrowd": 0, "area": 24181.139450000002, "rle": {"size": [478, 640], "counts": "]SR42h>4L4M2M4N2M2O2N1N3N2M4M2N2M3N2N2M3N2M300000000000000000000O1O100O101N1O100O1O100O100O1O2O0O101N1O10001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O001O00001O0O2O00001O001O00001O1O2N1O1O1O1O1O1O1N101O1O1O000000010O0000000001O01O0000000001O01O00000001O00010O001O00001O010O00002N2N2N2N2O1N000000O1O1O1O101N1O1O10000000001O00000000000000001OO10000O1O1O1N2O100O1O100O1O1O101N100O2O0000001O000O2O0000010O01O01O01O00101N1O2N2O1N2N2O1N000001O000010O000001O0000001O0001O0000000001O0001O00000001O0001O0000000001O01O000001O0000010O1O001O1O1O010O1O1O1O1O1O01O0000000002N1O2O1N1O2N2N2N2N2O1N2N2N2N2N2N2N3N1N2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N1O2N2N1O2O1L`_?"}, "label": "the yellow open book"}]}
{"id": 303247, "image": "COCO_train2014_000000303247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open book with a yellow cover\"."}], "task": "refering", "answer_template": "The \"an open book with a yellow cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 320, 341, 342, 343], "bbox": [0.43376562500000004, 0.5841213389121338, 0.94709375, 0.8423640167364016], "iscrowd": 0, "area": 24181.139450000002, "rle": {"size": [478, 640], "counts": "]SR42h>4L4M2M4N2M2O2N1N3N2M4M2N2M3N2N2M3N2M300000000000000000000O1O100O101N1O100O1O100O100O1O2O0O101N1O10001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O001O00001O0O2O00001O001O00001O1O2N1O1O1O1O1O1O1N101O1O1O000000010O0000000001O01O0000000001O01O00000001O00010O001O00001O010O00002N2N2N2N2O1N000000O1O1O1O101N1O1O10000000001O00000000000000001OO10000O1O1O1N2O100O1O100O1O1O101N100O2O0000001O000O2O0000010O01O01O01O00101N1O2N2O1N2N2O1N000001O000010O000001O0000001O0001O0000000001O0001O00000001O0001O0000000001O01O000001O0000010O1O001O1O1O010O1O1O1O1O1O01O0000000002N1O2O1N1O2N2N2N2N2O1N2N2N2N2N2N2N3N1N2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N1O2N2N1O2O1L`_?"}, "label": "an open book with a yellow cover"}]}
{"id": 450707, "image": "COCO_train2014_000000450707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person sitting on bench with hood over face\"."}], "task": "refering", "answer_template": "The \"a person sitting on bench with hood over face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270], "bbox": [0.60878125, 0.2757276995305164, 0.83165625, 0.781338028169014], "iscrowd": 0, "area": 12448.322750000005, "rle": {"size": [426, 640], "counts": "k]R54S=4L5K4L4L5L2M3N2M3N2N2N2N2N3M200O1O1O100O1O1O10O01O010WO[NkEg1U:f00100O3N1O2M2O2N1O2N1O2N2N1O2N2N2M7J6K0O10001N100O100000O1O2M2O1O1O1O1O001O1NhHeLc4[3SIiLY1OgNOn6X3PIPMT1MkNNQ7T3nHXMl0LQOKV7P3kHYOU7_30O1O1O010O1O1O100O0O2O1N2O1_OaHUK`7g4fHUK\\7g4jHVKW7e4j0L4L4TOm0F9H9N1O2N2M3N14M1O001N10000O1O100O6J2OnMUFc0j9WO^Fh0f9mNbFR1m:O0N3M3M4L5K1O2N2N2N1O2N2N2N2N1N3N2M2O1N1O2O1N2O0O2N1Ojd\\1"}, "label": "a person sitting on bench with hood over face"}]}
{"id": 450707, "image": "COCO_train2014_000000450707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person sleeping while sitting on the road bench\"."}], "task": "refering", "answer_template": "The \"a person sleeping while sitting on the road bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270], "bbox": [0.60878125, 0.2757276995305164, 0.83165625, 0.781338028169014], "iscrowd": 0, "area": 12448.322750000005, "rle": {"size": [426, 640], "counts": "k]R54S=4L5K4L4L5L2M3N2M3N2N2N2N2N3M200O1O1O100O1O1O10O01O010WO[NkEg1U:f00100O3N1O2M2O2N1O2N1O2N2N1O2N2N2M7J6K0O10001N100O100000O1O2M2O1O1O1O1O001O1NhHeLc4[3SIiLY1OgNOn6X3PIPMT1MkNNQ7T3nHXMl0LQOKV7P3kHYOU7_30O1O1O010O1O1O100O0O2O1N2O1_OaHUK`7g4fHUK\\7g4jHVKW7e4j0L4L4TOm0F9H9N1O2N2M3N14M1O001N10000O1O100O6J2OnMUFc0j9WO^Fh0f9mNbFR1m:O0N3M3M4L5K1O2N2N2N1O2N2N2N2N1N3N2M2O1N1O2O1N2O0O2N1Ojd\\1"}, "label": "a person sleeping while sitting on the road bench"}]}
{"id": 450707, "image": "COCO_train2014_000000450707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting on bench reading the newspaper\"."}], "task": "refering", "answer_template": "The \"man sitting on bench reading the newspaper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 215, 216, 235, 238, 239, 257, 258, 261, 262], "bbox": [0.21200000000000002, 0.24507042253521127, 0.41051562500000005, 0.7823708920187793], "iscrowd": 0, "area": 14278.2539, "rle": {"size": [426, 640], "counts": "alh18i<9G:F9K5O1L5K4K5hMYNiHk1V7VNeHo1Y7SNcHR2[7oMaHU2^7lM]HY2b7iMYH[2e7gMVH_2h7bMTHb2k7_MPHf2n7\\MnGh2Q8ZMjGi2V8XMgGi2Y8YMfGg2Z8ZMdGg2\\8ZMcGf2\\8]MaGd2_8]M`Gc2`8^M`Ga2_8aMaG^2_8cMaG\\2\\8iMdGU2Y8oMgGP2V8TNjGn0P9VOQG=X9FhFN`96`F@f9d0[FZOe9g0[FYOd9i0[FWOd9j0\\FUOd9l0]FSOb9n0^FQOb9P1^FPOa9Q1`FnN_9S1aFlN_9U1aFkN^9V1cFhN]9Y1cFgN\\9Z1dFfN[9\\1eFbN[9_1eFaNZ9`1fF_NZ9b1fF^NY9c1hF\\NW9e1iFZNW9g1iFYNV9h1kFVNT9l1lFTNo8Q2QGoMj8W2VGgMf8^2ZGbMd8`2\\G_Md8b2]G]Mc8c2]G]Mb8d2^G[Mc8e2]G[Mb8f2_GYMa8g2_GYM`8h2`GXM`8h2aGWM^8k2aGVM^8j2bGVM]8k2dGTM\\8l2dGTM[8m2eGTMZ8l2gGSMX8n2hGRMY8m2gGSMZ8l2fGUMY8k2hGTMY8l2fGTMZ8l2eGUM\\8j2cGWM]8i2aGZM_8e2`G\\M`8d2]G_Md8`2WGeMj8Z2QGlMn8X31O00105J=C5K1005K6J5K6K5J6J6J6J6J5L2M00O100O101N100O100O103L9H7H9TL[Gb2m8kLeGm2d9@`0^Ob0^Oc0]Ob0]Oabl4"}, "label": "man sitting on bench reading the newspaper"}]}
{"id": 450707, "image": "COCO_train2014_000000450707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a dark colored jacket sitting on a bench reading\"."}], "task": "refering", "answer_template": "The \"a man in a dark colored jacket sitting on a bench reading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 215, 216, 235, 238, 239, 257, 258, 261, 262], "bbox": [0.21200000000000002, 0.24507042253521127, 0.41051562500000005, 0.7823708920187793], "iscrowd": 0, "area": 14278.2539, "rle": {"size": [426, 640], "counts": "alh18i<9G:F9K5O1L5K4K5hMYNiHk1V7VNeHo1Y7SNcHR2[7oMaHU2^7lM]HY2b7iMYH[2e7gMVH_2h7bMTHb2k7_MPHf2n7\\MnGh2Q8ZMjGi2V8XMgGi2Y8YMfGg2Z8ZMdGg2\\8ZMcGf2\\8]MaGd2_8]M`Gc2`8^M`Ga2_8aMaG^2_8cMaG\\2\\8iMdGU2Y8oMgGP2V8TNjGn0P9VOQG=X9FhFN`96`F@f9d0[FZOe9g0[FYOd9i0[FWOd9j0\\FUOd9l0]FSOb9n0^FQOb9P1^FPOa9Q1`FnN_9S1aFlN_9U1aFkN^9V1cFhN]9Y1cFgN\\9Z1dFfN[9\\1eFbN[9_1eFaNZ9`1fF_NZ9b1fF^NY9c1hF\\NW9e1iFZNW9g1iFYNV9h1kFVNT9l1lFTNo8Q2QGoMj8W2VGgMf8^2ZGbMd8`2\\G_Md8b2]G]Mc8c2]G]Mb8d2^G[Mc8e2]G[Mb8f2_GYMa8g2_GYM`8h2`GXM`8h2aGWM^8k2aGVM^8j2bGVM]8k2dGTM\\8l2dGTM[8m2eGTMZ8l2gGSMX8n2hGRMY8m2gGSMZ8l2fGUMY8k2hGTMY8l2fGTMZ8l2eGUM\\8j2cGWM]8i2aGZM_8e2`G\\M`8d2]G_Md8`2WGeMj8Z2QGlMn8X31O00105J=C5K1005K6J5K6K5J6J6J6J6J5L2M00O100O101N100O100O103L9H7H9TL[Gb2m8kLeGm2d9@`0^Ob0^Oc0]Ob0]Oabl4"}, "label": "a man in a dark colored jacket sitting on a bench reading"}]}
{"id": 65685, "image": "COCO_train2014_000000065685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass stand in the left side of the image\"."}], "task": "refering", "answer_template": "The \"glass stand in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 143, 144, 145, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283], "bbox": [0.2205625, 0.3624882629107981, 0.32268749999999996, 0.8568779342723004], "iscrowd": 0, "area": 11273.913099999998, "rle": {"size": [426, 640], "counts": "RPk16d<d0ZOc0]Ob0^Ob0]Oc0^Ob0^Ob0]Oc0^Ob0^Ob0_Oa0000000001O000000000000000000O1000000O100000000000000000001O02N2N1O00000001O01O0000000O101O0cM]2D<D<D<D<^Ob0oNQ1\\OW]d5"}, "label": "glass stand in the left side of the image"}]}
{"id": 65685, "image": "COCO_train2014_000000065685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tall rectangular vase with pink flowers in it\"."}], "task": "refering", "answer_template": "The \"tall rectangular vase with pink flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 143, 144, 145, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283], "bbox": [0.2205625, 0.3624882629107981, 0.32268749999999996, 0.8568779342723004], "iscrowd": 0, "area": 11273.913099999998, "rle": {"size": [426, 640], "counts": "RPk16d<d0ZOc0]Ob0^Ob0]Oc0^Ob0^Ob0]Oc0^Ob0^Ob0_Oa0000000001O000000000000000000O1000000O100000000000000000001O02N2N1O00000001O01O0000000O101O0cM]2D<D<D<D<^Ob0oNQ1\\OW]d5"}, "label": "tall rectangular vase with pink flowers in it"}]}
{"id": 106646, "image": "COCO_train2014_000000106646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fast food coffee cup featuring a promotion to win a car\"."}], "task": "refering", "answer_template": "The \"a fast food coffee cup featuring a promotion to win a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284], "bbox": [0.11782812499999999, 0.14572916666666666, 0.515671875, 0.7528124999999999], "iscrowd": 0, "area": 54132.148749999986, "rle": {"size": [480, 640], "counts": "\\XS11a>`0@?M4N1O2N1O2N1O2N1O2N2M2O2N3M4L5K6J6J5K6I6K6J6J5K6J5K6J6J5K6I6K6J6J5K6J5K6J6J5J7J6J5K6J5K6J6J5K6J5J7J6J5K2N001O00001N10001O00001O00001O0O10001O00001O00001N10001O0000001O00001O000000001O00000000000000000000000001O0000000000000001O0000000000000000000000000000000000000000001O00000000000000000001O00000000001O00000O1O1O2N1O1O1N2O1N3M2N2N2O1N2O2O0O1O1O1O100O2N1O1O1O100O2N1O100O1O101N1O100O1O101N1O100O1O101N1O100O1O101N1J6J6J6J6J7I6J6J6H8I7H9H7H8H8I7H8I8G8I7H8I7H8I8G8I7H8K5O1O2O0O1O100O1O2O0O1O1O100O2N107H;EgVa4"}, "label": "a fast food coffee cup featuring a promotion to win a car"}]}
{"id": 106646, "image": "COCO_train2014_000000106646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tim horton ' s coffee cup\"."}], "task": "refering", "answer_template": "The \"a tim horton ' s coffee cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284], "bbox": [0.11782812499999999, 0.14572916666666666, 0.515671875, 0.7528124999999999], "iscrowd": 0, "area": 54132.148749999986, "rle": {"size": [480, 640], "counts": "\\XS11a>`0@?M4N1O2N1O2N1O2N1O2N2M2O2N3M4L5K6J6J5K6I6K6J6J5K6J5K6J6J5K6I6K6J6J5K6J5K6J6J5J7J6J5K6J5K6J6J5K6J5J7J6J5K2N001O00001N10001O00001O00001O0O10001O00001O00001N10001O0000001O00001O000000001O00000000000000000000000001O0000000000000001O0000000000000000000000000000000000000000001O00000000000000000001O00000000001O00000O1O1O2N1O1O1N2O1N3M2N2N2O1N2O2O0O1O1O1O100O2N1O1O1O100O2N1O100O1O101N1O100O1O101N1O100O1O101N1O100O1O101N1J6J6J6J6J7I6J6J6H8I7H9H7H8H8I7H8I8G8I7H8I7H8I8G8I7H8K5O1O2O0O1O100O1O2O0O1O1O100O2N107H;EgVa4"}, "label": "a tim horton ' s coffee cup"}]}
{"id": 106646, "image": "COCO_train2014_000000106646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate glazed covered donut\"."}], "task": "refering", "answer_template": "The \"a chocolate glazed covered donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 341], "bbox": [0.565875, 0.4718541666666667, 0.9966250000000001, 0.8277083333333334], "iscrowd": 0, "area": 37883.965249999994, "rle": {"size": [480, 640], "counts": "PoY51m>5L3L5L3L5L4K4M4L4K4M4K5L4K6K4K5L5J5L4L4K6L3L3N1N3N1O1N3N1O1N3N1O2M2O1N3N1O1N3N1O1N10000O2O000O100O2O000O10001N10000O2O0O10000O2O000O101O0O10000O2O0O10000O2O0000001O0O1000001O00000O2O000000001O0O10001O0000000O2O000000001O0O1000000000000O1000000000001N10000000000O1000000000000O1000000000000O1001O000000001O000000001O00000000001O000001O01O000000001O00000000001O000000001O000000001O000000001O000000001O001O001O001O000O2O001N101O001N101O0O101O001N101O001N101O0O101O001N101O001N10001N101O001N101O001N10001N101O001N102N2M3N2N1O2M3N2L4K5J5L5K5J6J6C=B=D=C=BSe0"}, "label": "a chocolate glazed covered donut"}]}
{"id": 106646, "image": "COCO_train2014_000000106646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan and brown donut with a thick coating of chocolate on top\"."}], "task": "refering", "answer_template": "The \"a tan and brown donut with a thick coating of chocolate on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 341], "bbox": [0.565875, 0.4718541666666667, 0.9966250000000001, 0.8277083333333334], "iscrowd": 0, "area": 37883.965249999994, "rle": {"size": [480, 640], "counts": "PoY51m>5L3L5L3L5L4K4M4L4K4M4K5L4K6K4K5L5J5L4L4K6L3L3N1N3N1O1N3N1O1N3N1O2M2O1N3N1O1N3N1O1N10000O2O000O100O2O000O10001N10000O2O0O10000O2O000O101O0O10000O2O0O10000O2O0000001O0O1000001O00000O2O000000001O0O10001O0000000O2O000000001O0O1000000000000O1000000000001N10000000000O1000000000000O1000000000000O1001O000000001O000000001O00000000001O000001O01O000000001O00000000001O000000001O000000001O000000001O000000001O001O001O001O000O2O001N101O001N101O0O101O001N101O001N101O0O101O001N101O001N10001N101O001N101O001N10001N101O001N102N2M3N2N1O2M3N2L4K5J5L5K5J6J6C=B=D=C=BSe0"}, "label": "a tan and brown donut with a thick coating of chocolate on top"}]}
{"id": 106646, "image": "COCO_train2014_000000106646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red cup behind the other one\"."}], "task": "refering", "answer_template": "The \"the red cup behind the other one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 80, 81, 103, 104, 126, 127, 148, 149, 150, 171, 172, 173, 194, 195, 196], "bbox": [0.291546875, 0.13514583333333335, 0.5651875000000001, 0.5234791666666667], "iscrowd": 0, "area": 8277.421150000002, "rle": {"size": [480, 640], "counts": "Vgg24k>1O1O1O10000000000000000000000000000000000000001O000000000000001O00000000001O0001O00000001O000000000000001O000000000000001O00000000001O000000000O101O0000000000001O00000000001O0000000000001O0000000000000000000000000gF_OS4a0gKEY4;`KL`44ZK2f4ORK8n4HlJ>T5BfJc0[5]O^Jj0b5VOXJP1h5POQJW1o5iNkI]1U6cNdId1\\6\\N^Ij1b6VNWIQ2i6oMPIX2P7hMjH]2W7cMdHb2\\7^M^Hh2b7XMYHm2g7SMTHR3l7nLnGX3R8hLiG]3W8cLgG_3Y8aLfG_3[8aLeG_3[8aLdG`3\\8`LdG_3]8aLbG`3^8aLaG^3`8fL[G[3e8iLWGV3j8nLQGS3o8QMmFo2S9UMhFk2Y9YMcFg2]9]1100O100O100O1XOh0I7J6I7I7I7L4K5@`0J6I7J6J6J7F9H9G8J6O2O0O2ChA0X>MmA0U>KQB2a>NXfR4"}, "label": "the red cup behind the other one"}]}
{"id": 106646, "image": "COCO_train2014_000000106646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee from tim horton ' s on a tray behind another cup of coffee\"."}], "task": "refering", "answer_template": "The \"a cup of coffee from tim horton ' s on a tray behind another cup of coffee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 80, 81, 103, 104, 126, 127, 148, 149, 150, 171, 172, 173, 194, 195, 196], "bbox": [0.291546875, 0.13514583333333335, 0.5651875000000001, 0.5234791666666667], "iscrowd": 0, "area": 8277.421150000002, "rle": {"size": [480, 640], "counts": "Vgg24k>1O1O1O10000000000000000000000000000000000000001O000000000000001O00000000001O0001O00000001O000000000000001O000000000000001O00000000001O000000000O101O0000000000001O00000000001O0000000000001O0000000000000000000000000gF_OS4a0gKEY4;`KL`44ZK2f4ORK8n4HlJ>T5BfJc0[5]O^Jj0b5VOXJP1h5POQJW1o5iNkI]1U6cNdId1\\6\\N^Ij1b6VNWIQ2i6oMPIX2P7hMjH]2W7cMdHb2\\7^M^Hh2b7XMYHm2g7SMTHR3l7nLnGX3R8hLiG]3W8cLgG_3Y8aLfG_3[8aLeG_3[8aLdG`3\\8`LdG_3]8aLbG`3^8aLaG^3`8fL[G[3e8iLWGV3j8nLQGS3o8QMmFo2S9UMhFk2Y9YMcFg2]9]1100O100O100O1XOh0I7J6I7I7I7L4K5@`0J6I7J6J6J7F9H9G8J6O2O0O2ChA0X>MmA0U>KQB2a>NXfR4"}, "label": "a cup of coffee from tim horton ' s on a tray behind another cup of coffee"}]}
{"id": 90277, "image": "COCO_train2014_000000090277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"equipment bag on far right of dugout\"."}], "task": "refering", "answer_template": "The \"equipment bag on far right of dugout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 75, 76, 77, 92, 93, 94, 109, 110, 111, 112, 126, 127, 128, 129, 143, 144, 145, 146, 160, 161, 162, 163, 178, 179, 180], "bbox": [0.4164791666666667, 0.13415625, 0.6436875000000001, 0.447203125], "iscrowd": 0, "area": 13684.264949999999, "rle": {"size": [640, 480], "counts": "`Wm39ec0>B2N3M2N3M2N3M2N2N3M2N3M2N3M2M4M3M3M2N3`M`2mNR1101O0O1000000O101O0O10000O100O1O1O010O1O1O010O1O1O10O01O1O100O001O1O1O1O001O2N2N2N2N1O2N2N2N2N2N2N1O2N2N1O2M3N1O2N1O2N2M2O2N1O2N1O1N2O001N2N2O0O2N2O1N2N102M2N3K4K6K5J5K6ZOe0mNTgZ3"}, "label": "equipment bag on far right of dugout"}]}
{"id": 90277, "image": "COCO_train2014_000000090277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bag in a dugout\"."}], "task": "refering", "answer_template": "The \"a bag in a dugout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 75, 76, 77, 92, 93, 94, 109, 110, 111, 112, 126, 127, 128, 129, 143, 144, 145, 146, 160, 161, 162, 163, 178, 179, 180], "bbox": [0.4164791666666667, 0.13415625, 0.6436875000000001, 0.447203125], "iscrowd": 0, "area": 13684.264949999999, "rle": {"size": [640, 480], "counts": "`Wm39ec0>B2N3M2N3M2N3M2N2N3M2N3M2N3M2M4M3M3M2N3`M`2mNR1101O0O1000000O101O0O10000O100O1O1O010O1O1O010O1O1O10O01O1O100O001O1O1O1O001O2N2N2N2N1O2N2N2N2N2N2N1O2N2N1O2M3N1O2N1O2N2M2O2N1O2N1O1N2O001N2N2O0O2N2O1N2N102M2N3K4K6K5J5K6ZOe0mNTgZ3"}, "label": "a bag in a dugout"}]}
{"id": 90277, "image": "COCO_train2014_000000090277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with blode hair swinging a bat\"."}], "task": "refering", "answer_template": "The \"woman with blode hair swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 180, 181, 182, 183, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 296, 297, 298, 299, 300, 301, 302, 303, 304, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 366, 367, 368, 369, 370, 371, 372, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.3793958333333334, 0.41378125, 0.9301250000000001, 0.98734375], "iscrowd": 0, "area": 61830.12329999999, "rle": {"size": [640, 480], "counts": "]Vb32gc07H7N3O1O001O001O1O001O001O1O0g]OYOZa0h0Z^ODea0<Q^OOna0R1OPOn]O5Pb0LP^O4oa0o0OnNR^O4ma0MS^O3la0MT^O4ka0MU^O3ja0S1MlNY^O3fa0NZ^O2ea0N[^O3da0N\\^O2ca0O]^O1ba0O^^O2aa0O_^O2_a0Na^O3^a0Nb^O2]a0Oc^O1\\a0Od^O2[a0Oe^O1Za0Og^O1Xa00g^O2Va0Ok^O1Ta00k^O1Ta00l^O0Ra01o^OOPa02o^OOPa01Q_O1k`01U_O2g`0OX_O4e`0L\\_O6``0L__O7^`0Ic_O9Z`0Hf_O:V`0Hi_O;T`0Em_O=P`0DP@>l?CU@?h?BX@`0e?A[@b0`?_Oa@c0\\?^Od@d0X?]Oi@e0T?\\Ol@f0m2]MP9m1TDh0h2]MR9m1VDh0e2\\MT9m1WDi0a2\\MW9k1YDk0\\2\\MY9k1[Dk0X2\\M\\9i1]Dn0R2[M`9g1_DP1n1ZMa9g1bDQ1i1ZMd9e1dDS1d1ZMg9b1gDV1^1ZMi9a1jDW1Z1YMk9`1lDY1U1YMn9^1nD[1P1YMP:]1QE\\1k0YMS:[1SE^1f0YMV:Y1UEa1a0WMX:Y1XEb1<WM[:W1ZEd17WM^:U1\\Ef12WM`:T1_Eg1NVMb:S1aEi1IVMe:Q1cEk1DVMg:P1fEl1_OVMj:n0hEn1ZOVMl:m0kEo1VOUMn:l0mE_3R:aLoE_3o9bLRF^3m9bLTF^3k9bLVF^3h9cLYF]3f9cL[F]3b9eL_F[3^9gLcFZ3Z9gLgFY3V9iLkFW3R9kLoFU3n8mLSGS3j8oLWGQ3f8QMYGQ3d8QM\\GP3a8RM_Go2_8RMaGo2\\8SMcGo2[8RMeGo2Z8QMfGP3Y8PMgGQ3X8oLgGS3X8mLhGT3W8lLiGU3V8kLjGV3U8jLjGX3U8hLkGY3T8gLlGZ3S8fLmG[3R8eLmG]3R8cLnG^3Q8bLoG_3P8`LQHa3n7_LQHc3n7]LRHd3m7\\LSHe3l7[LSHg3l7ZLSHg3l7[LRHf3m7\\LQHe3n7]LoGe3P8^LmGc3R8_LlGb3S8`LkGa3U8`LfGd3Z8]L_Gi3a8XLXGn3h8e30000O10000000001O0O1000000000000O1000000000000O10000000000O1000000000000O100000000001O1O1O2N1O1O1O1O1O1O2N1O1O100O1O1O1O2N1O2N2N3M2N3M2N3M2N3M2N3M3M2N`1`N000000001O000000001O0000001O000000001O000000001O000000010O0000001O3M3M3M3M2N3M3_KQDk0R<lMoEY1T:]MnGh1P>UOd0\\O2N3M3M2N3M3M2N[dd0"}, "label": "woman with blode hair swinging a bat"}]}
{"id": 90277, "image": "COCO_train2014_000000090277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl batting\"."}], "task": "refering", "answer_template": "The \"a girl batting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 180, 181, 182, 183, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 296, 297, 298, 299, 300, 301, 302, 303, 304, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 366, 367, 368, 369, 370, 371, 372, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.3793958333333334, 0.41378125, 0.9301250000000001, 0.98734375], "iscrowd": 0, "area": 61830.12329999999, "rle": {"size": [640, 480], "counts": "]Vb32gc07H7N3O1O001O001O1O001O001O1O0g]OYOZa0h0Z^ODea0<Q^OOna0R1OPOn]O5Pb0LP^O4oa0o0OnNR^O4ma0MS^O3la0MT^O4ka0MU^O3ja0S1MlNY^O3fa0NZ^O2ea0N[^O3da0N\\^O2ca0O]^O1ba0O^^O2aa0O_^O2_a0Na^O3^a0Nb^O2]a0Oc^O1\\a0Od^O2[a0Oe^O1Za0Og^O1Xa00g^O2Va0Ok^O1Ta00k^O1Ta00l^O0Ra01o^OOPa02o^OOPa01Q_O1k`01U_O2g`0OX_O4e`0L\\_O6``0L__O7^`0Ic_O9Z`0Hf_O:V`0Hi_O;T`0Em_O=P`0DP@>l?CU@?h?BX@`0e?A[@b0`?_Oa@c0\\?^Od@d0X?]Oi@e0T?\\Ol@f0m2]MP9m1TDh0h2]MR9m1VDh0e2\\MT9m1WDi0a2\\MW9k1YDk0\\2\\MY9k1[Dk0X2\\M\\9i1]Dn0R2[M`9g1_DP1n1ZMa9g1bDQ1i1ZMd9e1dDS1d1ZMg9b1gDV1^1ZMi9a1jDW1Z1YMk9`1lDY1U1YMn9^1nD[1P1YMP:]1QE\\1k0YMS:[1SE^1f0YMV:Y1UEa1a0WMX:Y1XEb1<WM[:W1ZEd17WM^:U1\\Ef12WM`:T1_Eg1NVMb:S1aEi1IVMe:Q1cEk1DVMg:P1fEl1_OVMj:n0hEn1ZOVMl:m0kEo1VOUMn:l0mE_3R:aLoE_3o9bLRF^3m9bLTF^3k9bLVF^3h9cLYF]3f9cL[F]3b9eL_F[3^9gLcFZ3Z9gLgFY3V9iLkFW3R9kLoFU3n8mLSGS3j8oLWGQ3f8QMYGQ3d8QM\\GP3a8RM_Go2_8RMaGo2\\8SMcGo2[8RMeGo2Z8QMfGP3Y8PMgGQ3X8oLgGS3X8mLhGT3W8lLiGU3V8kLjGV3U8jLjGX3U8hLkGY3T8gLlGZ3S8fLmG[3R8eLmG]3R8cLnG^3Q8bLoG_3P8`LQHa3n7_LQHc3n7]LRHd3m7\\LSHe3l7[LSHg3l7ZLSHg3l7[LRHf3m7\\LQHe3n7]LoGe3P8^LmGc3R8_LlGb3S8`LkGa3U8`LfGd3Z8]L_Gi3a8XLXGn3h8e30000O10000000001O0O1000000000000O1000000000000O10000000000O1000000000000O100000000001O1O1O2N1O1O1O1O1O1O2N1O1O100O1O1O1O2N1O2N2N3M2N3M2N3M2N3M2N3M3M2N`1`N000000001O000000001O0000001O000000001O000000001O000000010O0000001O3M3M3M3M2N3M3_KQDk0R<lMoEY1T:]MnGh1P>UOd0\\O2N3M3M2N3M3M2N[dd0"}, "label": "a girl batting"}]}
{"id": 98471, "image": "COCO_train2014_000000098471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white ford superduty truck\"."}], "task": "refering", "answer_template": "The \"the white ford superduty truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 318, 334, 335], "bbox": [0.062359374999999995, 0.23370833333333335, 0.9707812499999999, 0.8426875], "iscrowd": 0, "area": 105685.90440000003, "rle": {"size": [480, 640], "counts": "Pob05b>;POo0F;M2N3N1N3N1N3Nl0SO2O1N2O1L4I8I6K5O1N2O1O2N000O101O00000O2O00000100O2N1O100O2M2M3N2N1N101N2O1N2O0O2O100001N1O1O2N1O1N3N1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1O1N3N1O1O2N1O1O2N1O1O2O0O101O000O2O001O0O101O000O2O00001N10001O0O101O000O2O00001N10001O0O101O00001O00001O001O00001O01O01O00001O00001O00001O00001O00001O00001O00001O00000000O1O100O100O100O1O100O100O100O1O101N100O1000000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O01O01O0000001O00001O0000001O00001O0000001O00001O0eGWJ_7g6N2N3M2N2N2N3M2N2N2N3M2N2N2N2N3N1N2N1O001O010O001O1O0010O01O001O1O010O001O001O10O01O001O010O1O001O0010O010O3N1N3M3M2M4M3M2N3L3N3M3L3N3M3M2M4M3M2M4M00000O10000O1000000O2O000O1000000O1000000O10000O101O000O1000000O1000000O1000000O2O00000O1000000O1J6D<D<G91O0001O01O0000001O0000001O00001O0001O01O0000001O0000001O0000001O01O0001O0000001O0000001O0000010O0000001O0000001O0000001O0001O01O0000001O0000001O0000001O01O0001O0000001O000000O100000001O0O10000000000O10000000000O101O0000000O10000000000001O001O10O01O1O001O1O1O001O0000O1000001N10000O1000000O10000O1000000O2O00000O10000O1N2O1O1N2O1O1N2O2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N3M3M3M3M4L3M2WNQE^OR;>QE@R;>nDAT;>mD@V;@iE>]<O2N2O1J5C_d8"}, "label": "the white ford superduty truck"}]}
{"id": 98471, "image": "COCO_train2014_000000098471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big white truck in the foreground\"."}], "task": "refering", "answer_template": "The \"the big white truck in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 318, 334, 335], "bbox": [0.062359374999999995, 0.23370833333333335, 0.9707812499999999, 0.8426875], "iscrowd": 0, "area": 105685.90440000003, "rle": {"size": [480, 640], "counts": "Pob05b>;POo0F;M2N3N1N3N1N3Nl0SO2O1N2O1L4I8I6K5O1N2O1O2N000O101O00000O2O00000100O2N1O100O2M2M3N2N1N101N2O1N2O0O2O100001N1O1O2N1O1N3N1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1O1N3N1O1O2N1O1O2N1O1O2O0O101O000O2O001O0O101O000O2O00001N10001O0O101O000O2O00001N10001O0O101O00001O00001O001O00001O01O01O00001O00001O00001O00001O00001O00001O00001O00000000O1O100O100O100O1O100O100O100O1O101N100O1000000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O01O01O0000001O00001O0000001O00001O0000001O00001O0eGWJ_7g6N2N3M2N2N2N3M2N2N2N3M2N2N2N2N3N1N2N1O001O010O001O1O0010O01O001O1O010O001O001O10O01O001O010O1O001O0010O010O3N1N3M3M2M4M3M2N3L3N3M3L3N3M3M2M4M3M2M4M00000O10000O1000000O2O000O1000000O1000000O10000O101O000O1000000O1000000O1000000O2O00000O1000000O1J6D<D<G91O0001O01O0000001O0000001O00001O0001O01O0000001O0000001O0000001O01O0001O0000001O0000001O0000010O0000001O0000001O0000001O0001O01O0000001O0000001O0000001O01O0001O0000001O000000O100000001O0O10000000000O10000000000O101O0000000O10000000000001O001O10O01O1O001O1O1O001O0000O1000001N10000O1000000O10000O1000000O2O00000O10000O1N2O1O1N2O1O1N2O2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N3M3M3M3M4L3M2WNQE^OR;>QE@R;>nDAT;>mD@V;@iE>]<O2N2O1J5C_d8"}, "label": "the big white truck in the foreground"}]}
{"id": 467113, "image": "COCO_train2014_000000467113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bundle of unripe bananas being carried by a man\"."}], "task": "refering", "answer_template": "The \"a bundle of unripe bananas being carried by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 188, 189, 190, 191, 200, 202, 203, 204, 205, 206, 219, 220, 221], "bbox": [0.25934426229508195, 0.2743125, 0.7780327868852458, 0.64959375], "iscrowd": 0, "area": 26762.700500000003, "rle": {"size": [640, 427], "counts": "]eU25ec07J7L3N2N3L3N2N3M1N3N2N2N2M201N2N2N2O1N1O2N2O1N4L<D2O0O2N1O2O0YO`Mm_Oa2R`0bMk_O`2^?ZMg@9H^2a?\\Mc@9I]2b?]Mb@9JZ2d?]Ma@=GW2h?\\M`@]3_?cL`@_3_?aL`@a3_?_L_@c3`?^L_@d3`?]L^@d3b?\\L]@e3c?[L\\@f3d?ZL[@g3e?:WOhKjAX4V>hKjAk4b=WK]Bh4d=YK[Bg4e=YK[Bg4d=[K[Bd4f=]KYBc4f=^KZBa4g=`KXB`4g=aKYB^4h=cKWB\\4i=eKVB[4k=fKTBZ4k=gKUBX4l=hKTBW4m=jKRBU4o=kKQBU4P>kKoAT4R>lKnAS4S>nKlAR4U>mKjAS4W>nKgAR4Z>nKeAR4]>mKbAT4^>mK`AS4a>mK^AS4d>mKZAS4g>mKXAT4h>N_AhKc>R4h0K5K6RNm1M3M4L3M3O1O0O2O1O1O1O10O3N1O2N1O2N3L4M3M3M2N2N2M4Me0]AkLQ;V3kDmLU;T3gDoLY;R3dDPM[;R3dDnL\\;T3bDlL];V3PCgLa03^<X3eBQMl0G^<Z3cBRMm0E`<Z3aBSMn0D_<[3`BTMP1A_<]3_BTMQ1_O_<_3]BUMS1\\O`<a4^C`Ka<b4^C^Ka<d4^C\\Ka<e4_C[Ka<f4^CZKb<g4\\CZKd<g4[CYKf<g4YCYKg<h4XCXKh<o4QCQKP=W5fBjJZ=_5]BaJd=m500001O00001O00001O0001O01YO]BoJc=l4bBTK^=h4fBYKY=b4lB_KS=\\4SCdKm<W4WCjKh<T4ZCmKe<Q4]CPLb<n3aCRL_<k3cCVL\\<h3fCQMa;m2aDSM_;m2_DUMa;j2_DWMb;g2^DZMc;e2\\D\\Md;c2\\D^Me;`2[DaMe;_2ZDbMg;\\2YDeMh;Y2XDhMh;X2VDjMk;T2UDmMl;Q2TDPNl;P2RDRNo;l1nCXNS<f1jC^NV<a1hCbNY<i0nB^Mf0m1]<b0UCZM;X2`<<]CVM1`2c<7RDLo;1QD1o;NoC5R<JlC8V<EhC>Z<@gC?[<_OfC`0\\<]OfCb0[<]OgC?]<_OdC>`<_ObC>b<@_C=c<C_C8d<G^C4f<L\\COf<bNlAh0`1a0g<fNmAe0^1`0i<jNmAc0[1?l<lNmAb0W1b0m<jNoAa0V1d0l<iNPBb0U1d0k<iNRBb0T1d0k<hNSBb0U1d0i<hNTBc0T1d0i<gNVBc0S1d0h<gNWBc0S1e0g<fNXBd0S1c0f<hNYBd0R1c0j=VO^Bh0c=QOdBn0Y`0N100O2O000O100O10000O101O0O10=BRPk1"}, "label": "a bundle of unripe bananas being carried by a man"}]}
{"id": 467113, "image": "COCO_train2014_000000467113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a huge bundle of bananas being carried\"."}], "task": "refering", "answer_template": "The \"a huge bundle of bananas being carried\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 188, 189, 190, 191, 200, 202, 203, 204, 205, 206, 219, 220, 221], "bbox": [0.25934426229508195, 0.2743125, 0.7780327868852458, 0.64959375], "iscrowd": 0, "area": 26762.700500000003, "rle": {"size": [640, 427], "counts": "]eU25ec07J7L3N2N3L3N2N3M1N3N2N2N2M201N2N2N2O1N1O2N2O1N4L<D2O0O2N1O2O0YO`Mm_Oa2R`0bMk_O`2^?ZMg@9H^2a?\\Mc@9I]2b?]Mb@9JZ2d?]Ma@=GW2h?\\M`@]3_?cL`@_3_?aL`@a3_?_L_@c3`?^L_@d3`?]L^@d3b?\\L]@e3c?[L\\@f3d?ZL[@g3e?:WOhKjAX4V>hKjAk4b=WK]Bh4d=YK[Bg4e=YK[Bg4d=[K[Bd4f=]KYBc4f=^KZBa4g=`KXB`4g=aKYB^4h=cKWB\\4i=eKVB[4k=fKTBZ4k=gKUBX4l=hKTBW4m=jKRBU4o=kKQBU4P>kKoAT4R>lKnAS4S>nKlAR4U>mKjAS4W>nKgAR4Z>nKeAR4]>mKbAT4^>mK`AS4a>mK^AS4d>mKZAS4g>mKXAT4h>N_AhKc>R4h0K5K6RNm1M3M4L3M3O1O0O2O1O1O1O10O3N1O2N1O2N3L4M3M3M2N2N2M4Me0]AkLQ;V3kDmLU;T3gDoLY;R3dDPM[;R3dDnL\\;T3bDlL];V3PCgLa03^<X3eBQMl0G^<Z3cBRMm0E`<Z3aBSMn0D_<[3`BTMP1A_<]3_BTMQ1_O_<_3]BUMS1\\O`<a4^C`Ka<b4^C^Ka<d4^C\\Ka<e4_C[Ka<f4^CZKb<g4\\CZKd<g4[CYKf<g4YCYKg<h4XCXKh<o4QCQKP=W5fBjJZ=_5]BaJd=m500001O00001O00001O0001O01YO]BoJc=l4bBTK^=h4fBYKY=b4lB_KS=\\4SCdKm<W4WCjKh<T4ZCmKe<Q4]CPLb<n3aCRL_<k3cCVL\\<h3fCQMa;m2aDSM_;m2_DUMa;j2_DWMb;g2^DZMc;e2\\D\\Md;c2\\D^Me;`2[DaMe;_2ZDbMg;\\2YDeMh;Y2XDhMh;X2VDjMk;T2UDmMl;Q2TDPNl;P2RDRNo;l1nCXNS<f1jC^NV<a1hCbNY<i0nB^Mf0m1]<b0UCZM;X2`<<]CVM1`2c<7RDLo;1QD1o;NoC5R<JlC8V<EhC>Z<@gC?[<_OfC`0\\<]OfCb0[<]OgC?]<_OdC>`<_ObC>b<@_C=c<C_C8d<G^C4f<L\\COf<bNlAh0`1a0g<fNmAe0^1`0i<jNmAc0[1?l<lNmAb0W1b0m<jNoAa0V1d0l<iNPBb0U1d0k<iNRBb0T1d0k<hNSBb0U1d0i<hNTBc0T1d0i<gNVBc0S1d0h<gNWBc0S1e0g<fNXBd0S1c0f<hNYBd0R1c0j=VO^Bh0c=QOdBn0Y`0N100O2O000O100O10000O101O0O10=BRPk1"}, "label": "a huge bundle of bananas being carried"}]}
{"id": 188587, "image": "COCO_train2014_000000188587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink and black plane\"."}], "task": "refering", "answer_template": "The \"a pink and black plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 221, 222, 223, 224, 225, 226, 236], "bbox": [0.113171875, 0.258755868544601, 0.87559375, 0.6815727699530517], "iscrowd": 0, "area": 58093.75639999999, "rle": {"size": [426, 640], "counts": "ean01X=101N1O101N101N1O101N101N1O101N101N1O101N101N100XNYOlFi0R9XOmFi0Q9ZOnFg0o8[OPGg0n8[OQGh0k8YOTGk0i8VOVGm0g8SOXGQ1e8POYGT1d8lN\\GW1a8jN]GW1c8iN]GW1c8jN[GX1d8hN\\GX1d8iNZGX1f8hNYGY1h8gNWGY1i8gNVG[1i8eNWG[1i8fNUG[1k8eNUG[1k8eNUG\\1j8dNUG]1k8cNUG]1k8dNTG\\1l8dNTG\\1l8dNTG]1k8cNTG^1l8bNTG^1l8cNSG]1m8cNSG^1l8bNTG^1l8bNTG^1l8bNSG_1m8bNRG^1n8bNRG_1m8aNSG_1m8aNSG_1m8aNRG`1n8aNQG_1o8aNQG`1n8`NRG`1n8`NRG`1n8`NRG`1n8aNPGa1o8_NQGa1o8_NQGa1o8_NQGa1o8_NQGa1o8`NPGa1d0kMn6d0]Hb1b0mMQ7a0]Hb1`0PNS7=]Hc1>RNU7;]Hc1<UNV79]Hc1:VNY77\\Hd1;VNX76]Hd1;VNX76]Hd1:XNX74^He19WNY75]Hd1:XNX74^Hd1:XNX74]He1;XNW73^He1;XNW73^Hf19YNX72^He1:YNX72^He1:ZNW71^Hf1;ZNV71^Hf1;YNW74[Hc1>ZNV75ZHa1`0ZNV77XH_1a0\\NV77WH]1c0\\NV7:SH\\1f0[NV7l2jHTMV7l2jHUMU7k2kHUMU7k2jHWMU7j2jHVMV7j2jHWMU7i2kHWMU7i2kHXMT7i2kHWMU7i2jHYMU7g2kHZMT7f2lHZMT7f2lHZMT7g2kHYMU7g2kHXMV7h2iHXMY7g2gHYMY7g2fHYM[7h2aH[M_7e2_H\\Mb7d2\\H^Md7b2ZH_Mg7b2VH`Mj7`2THaMm7m3100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O2O0O100O100O1002N2N2N2N2N3M2N000000000000000000001O0000000000000000000001O000001O000000000000000000000000^OQHiKo7U4SHkKm7R4WHmKi7Q4YHPLf7m3^HRLb7l3`HTL`7l3aHSL_7m3aHSL_7m3bHRL^7n3bHRL^7n3cHQL]7o3cHQL]7P4cHoK]7Q4cHoK]7Q4dHnK\\7R4dHnK\\7R4eHmK[7S4eHmK[7S4fHmKY7S4gHmKY7S4hHlKY7S4gHmKY7S4hHlKX7T4hHlKX7T4iHkKW7U4iHkKW7U4jHjKV7V4jHjKV7V4jHjKV7V4kHiKU7W4kHiKU7W4lHhKT7X4lHhKT7X4mHhKR7X4nHhKR7X4P11O0000000000000000000000000000000000000001O000001O00000000000000000000000000000000000000000000001O0000000000000001O0000000000000000000000000001O00000000000000000000000000000000000000000001O01O0000000000000000000001O0000001O00001O0000001O01O01O0000001O00001O0000001O00001O0000001O00001O01O0001O00001O0000001O00001O0000001O00001O0000010O00001O0000001O00001O0000001O00001O0000001O00010O0000001O0000?A1O1O1O001O100O1O1O001O1O1O1O1O010O1O1O1O001O1O1bNgEIZ:DYF<g9QOlFn0h:01O1O1O1O1O1O1O100O001O1O1O1O1O1O1O100M2N3M3M3NRQQ1"}, "label": "a pink and black plane"}]}
{"id": 188587, "image": "COCO_train2014_000000188587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pink airplane\"."}], "task": "refering", "answer_template": "The \"pink airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 221, 222, 223, 224, 225, 226, 236], "bbox": [0.113171875, 0.258755868544601, 0.87559375, 0.6815727699530517], "iscrowd": 0, "area": 58093.75639999999, "rle": {"size": [426, 640], "counts": "ean01X=101N1O101N101N1O101N101N1O101N101N1O101N101N100XNYOlFi0R9XOmFi0Q9ZOnFg0o8[OPGg0n8[OQGh0k8YOTGk0i8VOVGm0g8SOXGQ1e8POYGT1d8lN\\GW1a8jN]GW1c8iN]GW1c8jN[GX1d8hN\\GX1d8iNZGX1f8hNYGY1h8gNWGY1i8gNVG[1i8eNWG[1i8fNUG[1k8eNUG[1k8eNUG\\1j8dNUG]1k8cNUG]1k8dNTG\\1l8dNTG\\1l8dNTG]1k8cNTG^1l8bNTG^1l8cNSG]1m8cNSG^1l8bNTG^1l8bNTG^1l8bNSG_1m8bNRG^1n8bNRG_1m8aNSG_1m8aNSG_1m8aNRG`1n8aNQG_1o8aNQG`1n8`NRG`1n8`NRG`1n8`NRG`1n8aNPGa1o8_NQGa1o8_NQGa1o8_NQGa1o8_NQGa1o8`NPGa1d0kMn6d0]Hb1b0mMQ7a0]Hb1`0PNS7=]Hc1>RNU7;]Hc1<UNV79]Hc1:VNY77\\Hd1;VNX76]Hd1;VNX76]Hd1:XNX74^He19WNY75]Hd1:XNX74^Hd1:XNX74]He1;XNW73^He1;XNW73^Hf19YNX72^He1:YNX72^He1:ZNW71^Hf1;ZNV71^Hf1;YNW74[Hc1>ZNV75ZHa1`0ZNV77XH_1a0\\NV77WH]1c0\\NV7:SH\\1f0[NV7l2jHTMV7l2jHUMU7k2kHUMU7k2jHWMU7j2jHVMV7j2jHWMU7i2kHWMU7i2kHXMT7i2kHWMU7i2jHYMU7g2kHZMT7f2lHZMT7f2lHZMT7g2kHYMU7g2kHXMV7h2iHXMY7g2gHYMY7g2fHYM[7h2aH[M_7e2_H\\Mb7d2\\H^Md7b2ZH_Mg7b2VH`Mj7`2THaMm7m3100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O2O0O100O100O1002N2N2N2N2N3M2N000000000000000000001O0000000000000000000001O000001O000000000000000000000000^OQHiKo7U4SHkKm7R4WHmKi7Q4YHPLf7m3^HRLb7l3`HTL`7l3aHSL_7m3aHSL_7m3bHRL^7n3bHRL^7n3cHQL]7o3cHQL]7P4cHoK]7Q4cHoK]7Q4dHnK\\7R4dHnK\\7R4eHmK[7S4eHmK[7S4fHmKY7S4gHmKY7S4hHlKY7S4gHmKY7S4hHlKX7T4hHlKX7T4iHkKW7U4iHkKW7U4jHjKV7V4jHjKV7V4jHjKV7V4kHiKU7W4kHiKU7W4lHhKT7X4lHhKT7X4mHhKR7X4nHhKR7X4P11O0000000000000000000000000000000000000001O000001O00000000000000000000000000000000000000000000001O0000000000000001O0000000000000000000000000001O00000000000000000000000000000000000000000001O01O0000000000000000000001O0000001O00001O0000001O01O01O0000001O00001O0000001O00001O0000001O00001O01O0001O00001O0000001O00001O0000001O00001O0000010O00001O0000001O00001O0000001O00001O0000001O00010O0000001O0000?A1O1O1O001O100O1O1O001O1O1O1O1O010O1O1O1O001O1O1bNgEIZ:DYF<g9QOlFn0h:01O1O1O1O1O1O1O100O001O1O1O1O1O1O1O100M2N3M3M3NRQQ1"}, "label": "pink airplane"}]}
{"id": 188587, "image": "COCO_train2014_000000188587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plane with a blue fin standing close to a pink airplane\"."}], "task": "refering", "answer_template": "The \"a white plane with a blue fin standing close to a pink airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120], "bbox": [0.00596875, 0.09417840375586854, 0.25371875, 0.3699530516431925], "iscrowd": 0, "area": 11432.002100000002, "rle": {"size": [426, 640], "counts": "Rh1\\1n;0000001O00000000001O000O1000001O000000001O00001O00001O00001O00001O001O00001O00001O0O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1N3N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1O1O100O100O100O10O0hN[FmNe9S1\\FlNc9U1^FjNb9V1_FiNa9V1aFiN`9V1aFiN_9W1bFhN^9X1bFhN^9W1dFgN]9Y1dFfN\\9Z1eFeN\\9Y1fFfNZ9Z1gFeNY9[1gFeNY9[1gFeNY9Z1hFfNY9Y1gFgNY9Y1gFfNZ9Z1fFfNZ9Y1gFgNY9Y1hFfNX9Z1hFfNY9Y1gFgNY9X1hFhNX9X1hFhNX9X1hFgNY9Y1gFgNZ9W1hFhNX9X1hFhNX9X1hFhNX9X1iFfNX9Y1iFgNW9Y1jFfNW9Y1iFgNW9X1kFgNU9Y1kFgNU9Y1lFeNU9Z1lFfNU9Y1lFfNT9Z1lFfNT9Y1mFgNS9Y1nFeNS9[1mFeNS9Z1oFeNR9Z1nFfNR9Z1oFeNS9X1nFgNU9W1lFhNV9V1jFjNX9S1jFlNX9R1hFnNZ9P1gFoN[9n0fFQO]9l0dFTO]9i0fFVO\\9f0fFZO\\9b0gF]O[99mFFV9MVG2Q;01O001O0OUTW6"}, "label": "a white plane with a blue fin standing close to a pink airplane"}]}
{"id": 188587, "image": "COCO_train2014_000000188587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane with k - amy on it\"."}], "task": "refering", "answer_template": "The \"the plane with k - amy on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120], "bbox": [0.00596875, 0.09417840375586854, 0.25371875, 0.3699530516431925], "iscrowd": 0, "area": 11432.002100000002, "rle": {"size": [426, 640], "counts": "Rh1\\1n;0000001O00000000001O000O1000001O000000001O00001O00001O00001O00001O001O00001O00001O0O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1N3N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1O1O100O100O100O10O0hN[FmNe9S1\\FlNc9U1^FjNb9V1_FiNa9V1aFiN`9V1aFiN_9W1bFhN^9X1bFhN^9W1dFgN]9Y1dFfN\\9Z1eFeN\\9Y1fFfNZ9Z1gFeNY9[1gFeNY9[1gFeNY9Z1hFfNY9Y1gFgNY9Y1gFfNZ9Z1fFfNZ9Y1gFgNY9Y1hFfNX9Z1hFfNY9Y1gFgNY9X1hFhNX9X1hFhNX9X1hFgNY9Y1gFgNZ9W1hFhNX9X1hFhNX9X1hFhNX9X1iFfNX9Y1iFgNW9Y1jFfNW9Y1iFgNW9X1kFgNU9Y1kFgNU9Y1lFeNU9Z1lFfNU9Y1lFfNT9Z1lFfNT9Y1mFgNS9Y1nFeNS9[1mFeNS9Z1oFeNR9Z1nFfNR9Z1oFeNS9X1nFgNU9W1lFhNV9V1jFjNX9S1jFlNX9R1hFnNZ9P1gFoN[9n0fFQO]9l0dFTO]9i0fFVO\\9f0fFZO\\9b0gF]O[99mFFV9MVG2Q;01O001O0OUTW6"}, "label": "the plane with k - amy on it"}]}
{"id": 188587, "image": "COCO_train2014_000000188587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow and white airplate\"."}], "task": "refering", "answer_template": "The \"yellow and white airplate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 185, 186, 187, 188, 189], "bbox": [0.0, 0.3076291079812207, 0.24690625000000002, 0.5605399061032864], "iscrowd": 0, "area": 8802.107300000001, "rle": {"size": [426, 640], "counts": "S4P3Z:000001O00001O00010O001O00001O00001O00001O01O01O00001O00001O000010O0001O00001O00001O0010O0001O001O00001O001O00010O00001O001O00001]NUEe0k:ZOWEe0j:YOXEg0g:XO\\Ef0d:ZO]Ee0d:ZO\\Ef0d:ZO\\Ef0d:ZO\\Ef0e:YO[Eg0e:YO[Eg0e:YO[Eg0f:XOZEh0f:XOZEh0f:XOZEi0e:WO[Ei0f:VOZEj0f:VOZEj0f:VOZEj0g:UOYEk0g:UOYEl0f:TO[Ek0f:TOZEl0f:TOZEl0f:TOZEl0g:SOYEm0g:SOYEn0f:ROZEn0g:QOZEn0f:ROZEn0f:QO[Eo0e:QO[Eo0f:POZEP1f:POZEQ1e:oN[EQ1f:nN[EQ1e:nN\\ER1d:nN\\ER1e:mN[EGLP1i:YOmEb0R:\\OPFd0Q:XOSFg0m9WOZFd0f9[O^Fb0T;101O00O1M301N2O1O1O2N00001O001O000010O01O00000000001O000000001O00000000001O00000000001O000000010O00\\YX6"}, "label": "yellow and white airplate"}]}
{"id": 188587, "image": "COCO_train2014_000000188587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and orange airplane\"."}], "task": "refering", "answer_template": "The \"a white and orange airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 185, 186, 187, 188, 189], "bbox": [0.0, 0.3076291079812207, 0.24690625000000002, 0.5605399061032864], "iscrowd": 0, "area": 8802.107300000001, "rle": {"size": [426, 640], "counts": "S4P3Z:000001O00001O00010O001O00001O00001O00001O01O01O00001O00001O000010O0001O00001O00001O0010O0001O001O00001O001O00010O00001O001O00001]NUEe0k:ZOWEe0j:YOXEg0g:XO\\Ef0d:ZO]Ee0d:ZO\\Ef0d:ZO\\Ef0d:ZO\\Ef0e:YO[Eg0e:YO[Eg0e:YO[Eg0f:XOZEh0f:XOZEh0f:XOZEi0e:WO[Ei0f:VOZEj0f:VOZEj0f:VOZEj0g:UOYEk0g:UOYEl0f:TO[Ek0f:TOZEl0f:TOZEl0f:TOZEl0g:SOYEm0g:SOYEn0f:ROZEn0g:QOZEn0f:ROZEn0f:QO[Eo0e:QO[Eo0f:POZEP1f:POZEQ1e:oN[EQ1f:nN[EQ1e:nN\\ER1d:nN\\ER1e:mN[EGLP1i:YOmEb0R:\\OPFd0Q:XOSFg0m9WOZFd0f9[O^Fb0T;101O00O1M301N2O1O1O2N00001O001O000010O01O00000000001O000000001O00000000001O00000000001O000000010O00\\YX6"}, "label": "a white and orange airplane"}]}
{"id": 377007, "image": "COCO_train2014_000000377007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white apple computer monitor displaying two web pages\"."}], "task": "refering", "answer_template": "The \"white apple computer monitor displaying two web pages\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247], "bbox": [0.5564843749999999, 0.17711943793911006, 0.9065312499999999, 0.6905854800936768], "iscrowd": 0, "area": 42330.69834999999, "rle": {"size": [427, 640], "counts": "o`d4Z5Q800000000001O00000000000000001O0000000O100000001O00000000000000001O00000000000000001O00000000000O101O00000000000000001O00000000000000001RI^JQ6b5nI_JR6a5lIaJT6_5kIcJT6]5jIeJV6[5iIfJW6Z5gIhJY6X5fIiJZ6W5dIkJ\\6V5bIkJ]6V5aIlJ_6T5_InJa6R5^IoJb6Q5\\IQKd6o4ZISKf6m4YITKg6i5O00000000000000000000001O0000000000000000000O101O0000000000000000000000001O00000000000000000000001O000O100000O1O1N2M3K5\\O[I]Jj6c5?0000000000000000001O00000000000000000O1000000000000000000000000000000000000000000000000000000O2O00000000000000000000000000000000000000000000000000L5iNW1iNX1hNW1iNYoh0"}, "label": "white apple computer monitor displaying two web pages"}]}
{"id": 377007, "image": "COCO_train2014_000000377007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an imac computer beside a computer monitor\"."}], "task": "refering", "answer_template": "The \"an imac computer beside a computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247], "bbox": [0.5564843749999999, 0.17711943793911006, 0.9065312499999999, 0.6905854800936768], "iscrowd": 0, "area": 42330.69834999999, "rle": {"size": [427, 640], "counts": "o`d4Z5Q800000000001O00000000000000001O0000000O100000001O00000000000000001O00000000000000001O00000000000O101O00000000000000001O00000000000000001RI^JQ6b5nI_JR6a5lIaJT6_5kIcJT6]5jIeJV6[5iIfJW6Z5gIhJY6X5fIiJZ6W5dIkJ\\6V5bIkJ]6V5aIlJ_6T5_InJa6R5^IoJb6Q5\\IQKd6o4ZISKf6m4YITKg6i5O00000000000000000000001O0000000000000000000O101O0000000000000000000000001O00000000000000000000001O000O100000O1O1N2M3K5\\O[I]Jj6c5?0000000000000000001O00000000000000000O1000000000000000000000000000000000000000000000000000000O2O00000000000000000000000000000000000000000000000000L5iNW1iNX1hNW1iNYoh0"}, "label": "an imac computer beside a computer monitor"}]}
{"id": 377007, "image": "COCO_train2014_000000377007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smaller computer screen with a blue desktop background\"."}], "task": "refering", "answer_template": "The \"a smaller computer screen with a blue desktop background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.30025, 0.27477751756440283, 0.5453749999999999, 0.5869789227166275], "iscrowd": 0, "area": 20407.327199999996, "rle": {"size": [427, 640], "counts": "nX`2l0`9o2G90000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000O100000000000000000000000000000000000000000M3SOm0SOm0SOm0SOYXi3"}, "label": "a smaller computer screen with a blue desktop background"}]}
{"id": 377007, "image": "COCO_train2014_000000377007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer screen with a blue background and itunes open\"."}], "task": "refering", "answer_template": "The \"a computer screen with a blue background and itunes open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.30025, 0.27477751756440283, 0.5453749999999999, 0.5869789227166275], "iscrowd": 0, "area": 20407.327199999996, "rle": {"size": [427, 640], "counts": "nX`2l0`9o2G90000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000O100000000000000000000000000000000000000000M3SOm0SOm0SOm0SOYXi3"}, "label": "a computer screen with a blue background and itunes open"}]}
{"id": 204979, "image": "COCO_train2014_000000204979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table that has two cameras attached to it\"."}], "task": "refering", "answer_template": "The \"a wooden table that has two cameras attached to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 171, 173, 174, 175, 189, 192, 193, 210], "bbox": [0.4598, 0.5973952095808384, 0.8016, 0.9324550898203593], "iscrowd": 0, "area": 6009.376300000002, "rle": {"size": [334, 500], "counts": "]W[22Y:4L4K4M4L4N2O1O001O1O1O1O001O1O1O10;Df0ZO4L4L4L4L2N0000O100aNVH:j7F^H1d7NeHI[77nH@R7`0VIWOk6i0^InNb6R1]1000000POgFk0^90001O00001O010O00001O001O00001O00001O0010O0001O00001O001O00001O001O01O01O00001O001O00001O00001O0010O0001O001OPF4^9LbF5]9KcF6\\9JdF7[9IeF8Z9HfF9Y9GgF9Y9GgF9Y9GgF9Y9>000O100[OgF5Y9Ke00lGET6<eIK[65^I3a6MXI;g6ERIb0n6^OkHj0T7VOeHR1Z7nN^HZ1b7fN[H^1d7cNZH^1f7bNYH_1g7aNXH`1h7`NWHa1i7_NWHa1i7_NVHb1j7a001O0000000O100000000000J6[Oe0\\Od0[Oe0N20000000000000000000O10O100000000000000000000000000000004K9HmQP1"}, "label": "a wooden table that has two cameras attached to it"}]}
{"id": 204979, "image": "COCO_train2014_000000204979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that is grayish colored that the camera equipment is on\"."}], "task": "refering", "answer_template": "The \"the table that is grayish colored that the camera equipment is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 171, 173, 174, 175, 189, 192, 193, 210], "bbox": [0.4598, 0.5973952095808384, 0.8016, 0.9324550898203593], "iscrowd": 0, "area": 6009.376300000002, "rle": {"size": [334, 500], "counts": "]W[22Y:4L4K4M4L4N2O1O001O1O1O1O001O1O1O10;Df0ZO4L4L4L4L2N0000O100aNVH:j7F^H1d7NeHI[77nH@R7`0VIWOk6i0^InNb6R1]1000000POgFk0^90001O00001O010O00001O001O00001O00001O0010O0001O00001O001O00001O001O01O01O00001O001O00001O00001O0010O0001O001OPF4^9LbF5]9KcF6\\9JdF7[9IeF8Z9HfF9Y9GgF9Y9GgF9Y9GgF9Y9>000O100[OgF5Y9Ke00lGET6<eIK[65^I3a6MXI;g6ERIb0n6^OkHj0T7VOeHR1Z7nN^HZ1b7fN[H^1d7cNZH^1f7bNYH_1g7aNXH`1h7`NWHa1i7_NWHa1i7_NVHb1j7a001O0000000O100000000000J6[Oe0\\Od0[Oe0N20000000000000000000O10O100000000000000000000000000000004K9HmQP1"}, "label": "the table that is grayish colored that the camera equipment is on"}]}
{"id": 204979, "image": "COCO_train2014_000000204979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair directly behind the laptop\"."}], "task": "refering", "answer_template": "The \"the chair directly behind the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 55, 56, 57, 58, 59, 60, 73, 91, 109], "bbox": [0.06756000000000001, 0.2195508982035928, 0.34076, 0.5588622754491018], "iscrowd": 0, "area": 3497.6716499999984, "rle": {"size": [334, 500], "counts": "TW;c1V7e1M3L4M31O0O2N101N1O100O1Lh0VO6G9H8G8H9H8G9H7K600000O1000000O1000000O10000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000001O0000000000001O00000000001O0000000000001O000000001O001O010O001O001O00001O001O001O001O00I7I7I7O1O10000000000010Oja[3"}, "label": "the chair directly behind the laptop"}]}
{"id": 204979, "image": "COCO_train2014_000000204979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the chair on the left that holds the computer\"."}], "task": "refering", "answer_template": "The \"the back of the chair on the left that holds the computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 55, 56, 57, 58, 59, 60, 73, 91, 109], "bbox": [0.06756000000000001, 0.2195508982035928, 0.34076, 0.5588622754491018], "iscrowd": 0, "area": 3497.6716499999984, "rle": {"size": [334, 500], "counts": "TW;c1V7e1M3L4M31O0O2N101N1O100O1Lh0VO6G9H8G8H9H8G9H7K600000O1000000O1000000O10000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000001O0000000000001O00000000001O0000000000001O000000001O001O010O001O001O00001O001O001O001O00I7I7I7O1O10000000000010Oja[3"}, "label": "the back of the chair on the left that holds the computer"}]}
{"id": 467126, "image": "COCO_train2014_000000467126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car\"."}], "task": "refering", "answer_template": "The \"black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 134, 135, 136, 137, 138, 139, 140], "bbox": [0.35218666666666665, 0.47487999999999997, 0.8174933333333333, 0.61674], "iscrowd": 0, "area": 7669.641450000001, "rle": {"size": [500, 375], "counts": "PgP27S??G9H7I3O2N1O2O1N1O2N101N1O2N2O0O2O0O2O001O0O100000000O100000000O2O00001N10001O000010O0001O0000O1O1N2O1N2O1N2001O0000000000000000000000001U\\7SN]eH:F:F:F:F:K5000000001O000001O01O00001O001O1O001O001O1O001O001O001O1O001O001O1O0010O01O1O001O001O1O001O010O10O010O010O001O010O00001O010O001O01N100N2N2N3M2N20001O00101N1O2L4K4M4K\\]Q1"}, "label": "black car"}]}
{"id": 467126, "image": "COCO_train2014_000000467126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black hatchback car\"."}], "task": "refering", "answer_template": "The \"a black hatchback car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 134, 135, 136, 137, 138, 139, 140], "bbox": [0.35218666666666665, 0.47487999999999997, 0.8174933333333333, 0.61674], "iscrowd": 0, "area": 7669.641450000001, "rle": {"size": [500, 375], "counts": "PgP27S??G9H7I3O2N1O2O1N1O2N101N1O2N2O0O2O0O2O001O0O100000000O100000000O2O00001N10001O000010O0001O0000O1O1N2O1N2O1N2001O0000000000000000000000001U\\7SN]eH:F:F:F:F:K5000000001O000001O01O00001O001O1O001O001O1O001O001O001O1O001O001O1O0010O01O1O001O001O1O001O010O10O010O010O001O010O00001O010O001O01N100N2N2N3M2N20001O00101N1O2L4K4M4K\\]Q1"}, "label": "a black hatchback car"}]}
{"id": 278715, "image": "COCO_train2014_000000278715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white car parked on curb\"."}], "task": "refering", "answer_template": "The \"white car parked on curb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 272, 273, 274, 277, 278, 279, 289, 290, 295, 296, 297, 298, 299, 300, 301, 302, 306, 307, 312, 313, 314, 315, 316, 317, 318, 319, 320, 323, 324, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.0029791666666666664, 0.6837500000000001, 0.8490208333333334, 0.919171875], "iscrowd": 0, "area": 27285.721399999995, "rle": {"size": [640, 480], "counts": "UR15kc0]1cN]1cNX1hN0O0100O10O0100O100O010O100O10O1000O1000O100000000O0100eNW@_Ni?`1Y@_Nf?`1]@_Nc?`1_@_Na?_1c@_N]?`1e@_N[?_1h@_NX?`1k@_NU?`1m@_NS?_1QA_No>`1SA_Nm>_1VA`Ni>`1YA_Ng>_1\\A`Nd>^1_A`Nb>_1aA_N_>_1dA`N[>`1fA`NZ>^1hAbNX>\\1jAdNV>[1kAeNU>Y1mAgNR>Y1oAfNR>f0Z@VOf14P>d0_@TOb18o=b0c@RO`1<m=`0h@oN]1a0j=?jBBV=<lBDT=:nBFR=8PCGP=8RCGo<7SCFP=8RCEQ=8RCDR=:PCCR=<PCAS==Y3N2N2N1O2N2N2N`_9o0c_F0000O10001O0000000000000000000000001O00000000000000000000001N1000000000000000000000001O0000000000000000000000001O00cLXObCg0_<ZO`Cf0`<[O_Ce0a<]O]Cc0d<]O[Cc0e<^OZCb0f<@XC`0h<AWC?j<AUC?k<CSC=m<DSC;m<FRC:o<GoB9Q=HnB8S=HnB6R=LPC0P=1SCKn<6TCFl<;WCAj<`0XC\\Oh<f0Y31N101N2N2O1N101N2O1N2O0O2d^OQOc?Q1[@XO]?i0a@AU?a0h@AX??g@AY?a0e@_O\\?`0d@@\\?`0d@@\\?a0b@@_??a@B^?>b@B_?=a@C_?>_@Cb?<^@Db?<^@Dc?;]@Ec?<[@Ee?;[@Ef?:Z@Ff?:X@Hi?8U@Jj?6T@Lm?3R@Nn?2P@0Q`00m_O1S`0Ok_O3U`0Mj_O4W`0Kg_O7Y`0If_O8[`0Hb_O:^`0Fa_O;``0D^_O>b`0B]_O`0b`0A[_Oa0f`0^OY_Oc0g`0[11O001O001O00001O010O001O001O00001O001O0010O0001O001O001O001O000010O01O001O000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1N2O1N2O100O1000000O1000000O1000000O100000000O100000000O100000000O1000000O101O001O1N2O1O1O1O1N2O1O1O1O1N2O1O001O1N2O1O1O1O1N2O1O1O1O1N2O1O0O2N2N2N2N2N2N2N2N2N3M4L5K4LWS]1"}, "label": "white car parked on curb"}]}
{"id": 278715, "image": "COCO_train2014_000000278715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white color car behind a girl catching a disc\"."}], "task": "refering", "answer_template": "The \"a white color car behind a girl catching a disc\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 272, 273, 274, 277, 278, 279, 289, 290, 295, 296, 297, 298, 299, 300, 301, 302, 306, 307, 312, 313, 314, 315, 316, 317, 318, 319, 320, 323, 324, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.0029791666666666664, 0.6837500000000001, 0.8490208333333334, 0.919171875], "iscrowd": 0, "area": 27285.721399999995, "rle": {"size": [640, 480], "counts": "UR15kc0]1cN]1cNX1hN0O0100O10O0100O100O010O100O10O1000O1000O100000000O0100eNW@_Ni?`1Y@_Nf?`1]@_Nc?`1_@_Na?_1c@_N]?`1e@_N[?_1h@_NX?`1k@_NU?`1m@_NS?_1QA_No>`1SA_Nm>_1VA`Ni>`1YA_Ng>_1\\A`Nd>^1_A`Nb>_1aA_N_>_1dA`N[>`1fA`NZ>^1hAbNX>\\1jAdNV>[1kAeNU>Y1mAgNR>Y1oAfNR>f0Z@VOf14P>d0_@TOb18o=b0c@RO`1<m=`0h@oN]1a0j=?jBBV=<lBDT=:nBFR=8PCGP=8RCGo<7SCFP=8RCEQ=8RCDR=:PCCR=<PCAS==Y3N2N2N1O2N2N2N`_9o0c_F0000O10001O0000000000000000000000001O00000000000000000000001N1000000000000000000000001O0000000000000000000000001O00cLXObCg0_<ZO`Cf0`<[O_Ce0a<]O]Cc0d<]O[Cc0e<^OZCb0f<@XC`0h<AWC?j<AUC?k<CSC=m<DSC;m<FRC:o<GoB9Q=HnB8S=HnB6R=LPC0P=1SCKn<6TCFl<;WCAj<`0XC\\Oh<f0Y31N101N2N2O1N101N2O1N2O0O2d^OQOc?Q1[@XO]?i0a@AU?a0h@AX??g@AY?a0e@_O\\?`0d@@\\?`0d@@\\?a0b@@_??a@B^?>b@B_?=a@C_?>_@Cb?<^@Db?<^@Dc?;]@Ec?<[@Ee?;[@Ef?:Z@Ff?:X@Hi?8U@Jj?6T@Lm?3R@Nn?2P@0Q`00m_O1S`0Ok_O3U`0Mj_O4W`0Kg_O7Y`0If_O8[`0Hb_O:^`0Fa_O;``0D^_O>b`0B]_O`0b`0A[_Oa0f`0^OY_Oc0g`0[11O001O001O00001O010O001O001O00001O001O0010O0001O001O001O001O000010O01O001O000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1N2O1N2O100O1000000O1000000O1000000O100000000O100000000O100000000O1000000O101O001O1N2O1O1O1O1N2O1O1O1O1N2O1O001O1N2O1O1O1O1N2O1O1O1O1N2O1O0O2N2N2N2N2N2N2N2N2N3M4L5K4LWS]1"}, "label": "a white color car behind a girl catching a disc"}]}
{"id": 532668, "image": "COCO_train2014_000000532668.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ship with number 154 on it\"."}], "task": "refering", "answer_template": "The \"a ship with number 154 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274], "bbox": [0.08395312499999999, 0.5820140515222483, 0.953546875, 0.7618032786885246], "iscrowd": 0, "area": 27548.411049999995, "rle": {"size": [427, 640], "counts": "Vif01Z=1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O3M6J5K6J2N0000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O100000000O10000000000000000O10000000000000000O10000000000000000O2O000001O0000001O0000001O000000001O0000001O0000001O0000001O00000000000000000000001O00000001O000000000000000000000000000001O0000000000000000000000000000000000001O000000000000O10001N10000O1000000O10000O10000O10000O10000O1000000O1000000O10000O1000000O1000000O1000000O10001O00000000000O1000000000000000000O1000000000000000000O1000000000000000000000000000000001O01O000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O000001O0000000001O000000000000001O00000000000000001O00000000002N5K5K3M00001O00001O001O00001O00010O000000001O00000000000000001O00000000000000001O00000000000000001O0000000000000000000010O0000000000000000000000000001O000000000000000000000000001O00000000000000000000000000000001O0000000000000000000001O0000000000000000001O0000000000000000001N1N2O1N2O1N2O1O1N2O1N3N1N2O1O1N2O1N2O1N2O2N1N2OYW<"}, "label": "a ship with number 154 on it"}]}
{"id": 532668, "image": "COCO_train2014_000000532668.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"excursion ship\"."}], "task": "refering", "answer_template": "The \"excursion ship\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274], "bbox": [0.08395312499999999, 0.5820140515222483, 0.953546875, 0.7618032786885246], "iscrowd": 0, "area": 27548.411049999995, "rle": {"size": [427, 640], "counts": "Vif01Z=1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O3M6J5K6J2N0000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O100000000O10000000000000000O10000000000000000O10000000000000000O2O000001O0000001O0000001O000000001O0000001O0000001O0000001O00000000000000000000001O00000001O000000000000000000000000000001O0000000000000000000000000000000000001O000000000000O10001N10000O1000000O10000O10000O10000O10000O1000000O1000000O10000O1000000O1000000O1000000O10001O00000000000O1000000000000000000O1000000000000000000O1000000000000000000000000000000001O01O000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O000001O0000000001O000000000000001O00000000000000001O00000000002N5K5K3M00001O00001O001O00001O00010O000000001O00000000000000001O00000000000000001O00000000000000001O0000000000000000000010O0000000000000000000000000001O000000000000000000000000001O00000000000000000000000000000001O0000000000000000000001O0000000000000000001O0000000000000000001N1N2O1N2O1N2O1O1N2O1N3N1N2O1O1N2O1N2O1N2O2N1N2OYW<"}, "label": "excursion ship"}]}
{"id": 188606, "image": "COCO_train2014_000000188606.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the forest wearing sunglasses\"."}], "task": "refering", "answer_template": "The \"a man in the forest wearing sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.23809375, 0.19795833333333332, 1.0, 1.0], "iscrowd": 0, "area": 125992.59544999998, "rle": {"size": [480, 640], "counts": "oUX21m>2O100O1O1O100O1O1eN4bCLT<o0ZCRO\\<n1F:F:F:I7I7H8I7L4L4L4M3L4L4L4M3L4N2N2O1N2O1N2O1O1O100YLdJiM]5T2jJgMW5Y2lJdMT5[2PKaMQ5^2RK_Mo4`2TK]Mm4c2YKWMg4h2`KQMa4f2mKTMT4g2SLVMn3i2VLTMj3j2ZLSMg3k2^LQMc3n2`LPM`3o2cLnL^3P3fLmL[3R3hLlLX3S3kLkLU3U3mLhLT3W3nLhLR3W3QMfLP3Z3RMdLn2[3UMcLk2]3VMaLk2^3YM_Lg2a3\\M\\Ld2c3_M[La2e3bMXL^2g3eMWL[2i3gMTLZ2k3iMSLW2l3kMSLU2m3mMQLS2n3PNPLP2o3SNoKm1Q4UNmKk1R4WNmKi1R4ZNlKf1S4\\NlKd1S4^NlKb1S4aNkK_1T4cNkK]1T4eNkK[1U4h4O1O1O1O1N2O1O1O100O1O1O1O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O10000O10000O10000O1O100O1O1O1O100O1O1O10000000000000000000000001O00000000001O0000000000000000000000000000001O001O1O1O001O1O1O001O1O1O001O1O1O000000001O0000000000001O00000000001O00000000001O1O1O1O001O1O1O1O1O001O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O7I4L3M3M3M2N1O001VIgJk3Z5SLkJi3V5TLmJk3V5dKeJRN5[6Y5XK\\Kh4g4iIlJ3h0T6_4^IjLb6i5100000000O10000000000O11O2N1O1O1O1O000000O100O100O10000O100O11O000000000000000000000000000000002N1O1O1O1O0000000000001O000000000000000000000000001O000000000000001O00000000000000O1O1O100O1O100O1O100O1O100O1O1O1O100O1OHoGcIP8U6[HiId7d5U12N100N3N2M3N1N3N2M3M2O2M3L3N3L4M2M4M3N1010O0010O0100O010O00100O101N100O1O100O1O2O0O1O100O1O2O0O1O100O1O101N1O100O1O100O2N100O1O1O100O2N1O100O1O1O101N1O1O100O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1012MO1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1fLaG1a8QO^IL"}, "label": "a man in the forest wearing sunglasses"}]}
{"id": 188606, "image": "COCO_train2014_000000188606.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with sunglasses\"."}], "task": "refering", "answer_template": "The \"man with sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.23809375, 0.19795833333333332, 1.0, 1.0], "iscrowd": 0, "area": 125992.59544999998, "rle": {"size": [480, 640], "counts": "oUX21m>2O100O1O1O100O1O1eN4bCLT<o0ZCRO\\<n1F:F:F:I7I7H8I7L4L4L4M3L4L4L4M3L4N2N2O1N2O1N2O1O1O100YLdJiM]5T2jJgMW5Y2lJdMT5[2PKaMQ5^2RK_Mo4`2TK]Mm4c2YKWMg4h2`KQMa4f2mKTMT4g2SLVMn3i2VLTMj3j2ZLSMg3k2^LQMc3n2`LPM`3o2cLnL^3P3fLmL[3R3hLlLX3S3kLkLU3U3mLhLT3W3nLhLR3W3QMfLP3Z3RMdLn2[3UMcLk2]3VMaLk2^3YM_Lg2a3\\M\\Ld2c3_M[La2e3bMXL^2g3eMWL[2i3gMTLZ2k3iMSLW2l3kMSLU2m3mMQLS2n3PNPLP2o3SNoKm1Q4UNmKk1R4WNmKi1R4ZNlKf1S4\\NlKd1S4^NlKb1S4aNkK_1T4cNkK]1T4eNkK[1U4h4O1O1O1O1N2O1O1O100O1O1O1O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O10000O10000O10000O1O100O1O1O1O100O1O1O10000000000000000000000001O00000000001O0000000000000000000000000000001O001O1O1O001O1O1O001O1O1O001O1O1O000000001O0000000000001O00000000001O00000000001O1O1O1O001O1O1O1O1O001O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O7I4L3M3M3M2N1O001VIgJk3Z5SLkJi3V5TLmJk3V5dKeJRN5[6Y5XK\\Kh4g4iIlJ3h0T6_4^IjLb6i5100000000O10000000000O11O2N1O1O1O1O000000O100O100O10000O100O11O000000000000000000000000000000002N1O1O1O1O0000000000001O000000000000000000000000001O000000000000001O00000000000000O1O1O100O1O100O1O100O1O100O1O1O1O100O1OHoGcIP8U6[HiId7d5U12N100N3N2M3N1N3N2M3M2O2M3L3N3L4M2M4M3N1010O0010O0100O010O00100O101N100O1O100O1O2O0O1O100O1O2O0O1O100O1O101N1O100O1O100O2N100O1O1O100O2N1O100O1O1O101N1O1O100O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1012MO1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1fLaG1a8QO^IL"}, "label": "man with sunglasses"}]}
{"id": 245953, "image": "COCO_train2014_000000245953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog is standing over another dog on a purple leash\"."}], "task": "refering", "answer_template": "The \"a dog is standing over another dog on a purple leash\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 44, 45, 46, 47, 48, 62, 63, 64, 65, 66, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 115, 116, 117, 119, 134, 135, 137], "bbox": [0.42923541247484914, 0.060609375, 0.7076056338028169, 0.35757812499999997], "iscrowd": 0, "area": 15254.967249999998, "rle": {"size": [640, 497], "counts": "TYU41nc0;E2N001O1O1TN6a_OJ_`0:]_OGb`0=Z_ODf`0>W_OCh`0a0T_O_Ol`0e0n^O^ORa0e0g^O^OYa0d101O01O01O01O0O1O2N100100O0l^OlN[?W1Y@SOg?e2M5K4K5L4L4L4L4L1O1O0000001O0000N3L3N2N4L4L4M3N2N2N2N2N2M3N2N1O2N2N2N2N2O1O1O1O1O1O1O1N20O101O00000000001O00000000001O0001O000001Z@_Lk>a3o@TMb>l2XAiMY>W2aAWNW>j3O10O01O1O1O1OSNmAnMR>o1UBmMk=Q2\\BiMe=T2cBfM^=X2hBcMZ=Y2nBaMS=]2SC_Mm<^2XC_Mi<_2[C^Mf<`2^C]Mc<`2bC]M_<a2fC\\MZ<b2jC[MW<c2mCZMU<b2PD[MQ<c2g2N2N2N2N2N2N2N2N3M2N2N2N3M5J6K5K5F:A?A]bj2"}, "label": "a dog is standing over another dog on a purple leash"}]}
{"id": 245953, "image": "COCO_train2014_000000245953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dog laying on the ground behind a plate\"."}], "task": "refering", "answer_template": "The \"a black dog laying on the ground behind a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 234, 235, 236, 241, 242, 243, 244, 245, 246, 247, 253, 254, 260, 261, 262, 263, 271, 272, 280, 281], "bbox": [0.04674044265593561, 0.30846874999999996, 0.7580885311871228, 0.6875], "iscrowd": 0, "area": 42907.53740000001, "rle": {"size": [640, 497], "counts": "^g>3jc0;E;F=B=C4M3L5J5K6J5J6[Of0D;F;E:F:I7N1N2N3N1N2N2O1N2N2O1N2O1N2O2N1O1N2O1O1N2O`NoASMP>n2RBPMl=Q3WBmLh=T3ZBkLd=U3_BiL`=X3bBfL\\=[3gBaLZ=`3hBYL\\=g3gBRL^=n3dBkK`=U4cBdKb=\\4P101O01O010O010O0010O010O01O010O010O01O2O0O1O2O0O101N1O100O2N100O2N100O1O2O0O1O2O0O100O2N100O1O1O100O1O1O1O1O001O1O1O1O1O100O100000000000000000^_OeL_`0[3a_OeL_`0[3`_OgL_`0^3O001O1O001O1O0010O01O1O001O1O00100O010O1O010O00100O00100O010O1O010O1O010O1O010O010O1O010O10O01O10O0100O0010O010O01TOcLSA]3l>m0N1O1O2N1O101N1O1O2N1O2N1O1O21N1O100O1O1000000O2O000000001O001O001O001O001O001N10001N100O1gNRA_Ml>`2WA_Mh>`2[A`Mc>_2_AaM`>^2cAbM[>]2hAbMV>^2mAbMQ>]2RBbMm=]2VBcMh=]2ZBbMe=^2]BbMa=_2_BaM`=`2aB`M]=`2eB_MZ=b2fB_MX=b2iB]MV=d2jB]MT=c2nB]MP=d2QC\\Mm<e2SC\\Ml<d2UC\\Mi<d2YC\\Me<e2[C[Md<f2]CZMa<f2aCZM]<g2cCZM[<g2fCYMX<g2jCYMT<h2lCYMS<g2nCYMP<g2RDZMl;f2UD]Mf;e2YD_Mc;b2]DaM_;_2bDdMY;^2gDdMV;]2jDbMU;`2kD_MU;a2lD^MT;c2kD]MT;e2lDZMT;f2mDYMS;h2mDWMT;i2lDVMT;k2lDTMU;k2lDTMT;_O^C]2]1TNV;\\ObC^2X1UNW;[OeC]2U1WNV;[OiC_1@fNa1?W;YOQDV1]OROZ1?X;XOYDj0[O@T1=Y;VOaDa0XOKo0=X;VOWF=aN<Y;UOXF?_N;Y;TOZFa0]N;j>EVA:k>FUA8m>HSA7m>ISA6o>JQA5P?KPA3Q?MPA2Q?No@1Q?0n@OT?0m@OT?1l@MU?3l@LU?4k@JV?7i@IX?6i@HY?8g@GY?:g@D[?;e@E\\?;d@D\\?=e@@]??f@^OZ?c0h@YOY?g0k@UOV?k0o11O00100O001O10O01O0010O0000010O0000001O00001O00001O01O01O00001O00001O00001O01O01O1O2M3M3M3M3MhdZ2"}, "label": "a black dog laying on the ground behind a plate"}]}
{"id": 245953, "image": "COCO_train2014_000000245953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dog , with a black leash , laying on the ground\"."}], "task": "refering", "answer_template": "The \"a black dog , with a black leash , laying on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 234, 235, 236, 241, 242, 243, 244, 245, 246, 247, 253, 254, 260, 261, 262, 263, 271, 272, 280, 281], "bbox": [0.04674044265593561, 0.30846874999999996, 0.7580885311871228, 0.6875], "iscrowd": 0, "area": 42907.53740000001, "rle": {"size": [640, 497], "counts": "^g>3jc0;E;F=B=C4M3L5J5K6J5J6[Of0D;F;E:F:I7N1N2N3N1N2N2O1N2N2O1N2O1N2O2N1O1N2O1O1N2O`NoASMP>n2RBPMl=Q3WBmLh=T3ZBkLd=U3_BiL`=X3bBfL\\=[3gBaLZ=`3hBYL\\=g3gBRL^=n3dBkK`=U4cBdKb=\\4P101O01O010O010O0010O010O01O010O010O01O2O0O1O2O0O101N1O100O2N100O2N100O1O2O0O1O2O0O100O2N100O1O1O100O1O1O1O1O001O1O1O1O1O100O100000000000000000^_OeL_`0[3a_OeL_`0[3`_OgL_`0^3O001O1O001O1O0010O01O1O001O1O00100O010O1O010O00100O00100O010O1O010O1O010O1O010O010O1O010O10O01O10O0100O0010O010O01TOcLSA]3l>m0N1O1O2N1O101N1O1O2N1O2N1O1O21N1O100O1O1000000O2O000000001O001O001O001O001O001N10001N100O1gNRA_Ml>`2WA_Mh>`2[A`Mc>_2_AaM`>^2cAbM[>]2hAbMV>^2mAbMQ>]2RBbMm=]2VBcMh=]2ZBbMe=^2]BbMa=_2_BaM`=`2aB`M]=`2eB_MZ=b2fB_MX=b2iB]MV=d2jB]MT=c2nB]MP=d2QC\\Mm<e2SC\\Ml<d2UC\\Mi<d2YC\\Me<e2[C[Md<f2]CZMa<f2aCZM]<g2cCZM[<g2fCYMX<g2jCYMT<h2lCYMS<g2nCYMP<g2RDZMl;f2UD]Mf;e2YD_Mc;b2]DaM_;_2bDdMY;^2gDdMV;]2jDbMU;`2kD_MU;a2lD^MT;c2kD]MT;e2lDZMT;f2mDYMS;h2mDWMT;i2lDVMT;k2lDTMU;k2lDTMT;_O^C]2]1TNV;\\ObC^2X1UNW;[OeC]2U1WNV;[OiC_1@fNa1?W;YOQDV1]OROZ1?X;XOYDj0[O@T1=Y;VOaDa0XOKo0=X;VOWF=aN<Y;UOXF?_N;Y;TOZFa0]N;j>EVA:k>FUA8m>HSA7m>ISA6o>JQA5P?KPA3Q?MPA2Q?No@1Q?0n@OT?0m@OT?1l@MU?3l@LU?4k@JV?7i@IX?6i@HY?8g@GY?:g@D[?;e@E\\?;d@D\\?=e@@]??f@^OZ?c0h@YOY?g0k@UOV?k0o11O00100O001O10O01O0010O0000010O0000001O00001O00001O01O01O00001O00001O00001O01O01O1O2M3M3M3M3MhdZ2"}, "label": "a black dog , with a black leash , laying on the ground"}]}
{"id": 319685, "image": "COCO_train2014_000000319685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the polar bear with its mouth closed\"."}], "task": "refering", "answer_template": "The \"the polar bear with its mouth closed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 269, 270, 271, 292, 293, 294, 295, 315, 316, 317, 318], "bbox": [0.563765625, 0.29514583333333333, 0.851578125, 0.8120833333333333], "iscrowd": 0, "area": 23167.75565, "rle": {"size": [480, 640], "counts": "i_Y56h>:F9F7J5K5J6J5K5K5L4K5K4L5H6G:J5N2O2N1O2N1O1O2M2O2N1O2N1O1O2N1O2N1O1O2MYO[EWMb:h2eEVMX:h2oETMP:m2SFQMk9P3YFmLe9U3]FhLb9Y3bFdLZ9`3l06L5K4L5K5K4N3O001O001O001O001O010O10O10010O01O1O100O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N2N1O2N1O3M3M3M2N3N3L5L4K6K<C7J7H01O010O001O001O3M3M2N3M3M3M3M2N5K7I6J6J6J5K5K4L5K4L3M3M3M3M4L3M3M3M3M3M4L3M3M3M3M3M3N2M3M3M3M3M4M2M3M3M3M3M3N2M3M2N3M2N3N1N3M2N3M2N3N2M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2NlT\\1"}, "label": "the polar bear with its mouth closed"}]}
{"id": 319685, "image": "COCO_train2014_000000319685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a polar bear that is wrestling another bear and does not have its mouth open\"."}], "task": "refering", "answer_template": "The \"a polar bear that is wrestling another bear and does not have its mouth open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 269, 270, 271, 292, 293, 294, 295, 315, 316, 317, 318], "bbox": [0.563765625, 0.29514583333333333, 0.851578125, 0.8120833333333333], "iscrowd": 0, "area": 23167.75565, "rle": {"size": [480, 640], "counts": "i_Y56h>:F9F7J5K5J6J5K5K5L4K5K4L5H6G:J5N2O2N1O2N1O1O2M2O2N1O2N1O1O2N1O2N1O1O2MYO[EWMb:h2eEVMX:h2oETMP:m2SFQMk9P3YFmLe9U3]FhLb9Y3bFdLZ9`3l06L5K4L5K5K4N3O001O001O001O001O010O10O10010O01O1O100O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N2N1O2N1O3M3M3M2N3N3L5L4K6K<C7J7H01O010O001O001O3M3M2N3M3M3M3M2N5K7I6J6J6J5K5K4L5K4L3M3M3M3M4L3M3M3M3M3M4L3M3M3M3M3M3N2M3M3M3M3M4M2M3M3M3M3M3N2M3M2N3M2N3N1N3M2N3M2N3N2M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2NlT\\1"}, "label": "a polar bear that is wrestling another bear and does not have its mouth open"}]}
{"id": 319685, "image": "COCO_train2014_000000319685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white polar bear that is fighting and has its mouth open\"."}], "task": "refering", "answer_template": "The \"a white polar bear that is fighting and has its mouth open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334], "bbox": [0.38340625, 0.33172916666666663, 0.697875, 0.8810625], "iscrowd": 0, "area": 27306.2699, "rle": {"size": [480, 640], "counts": "fWc31m>2N2M3N3M2M3N2N2M4L3L4K5L5J5L4L4K5L5H7H8I7I7H9F9E;E;E<E:L4L4L4L5K4L4L4L4L5K4K5K5K5J7J500O1O2N1O1O1O1O2N1O1O100O2N1O100O2O0O100O1O2O0O100O1O2O0O100O101O0O1000000O1000000O1000000O1000000O[OTIXIl6h6WIUIi6j6\\IRId6n6_IoHa6Q7bIlH]6T7hIhHX6X7h000O010O1000O010000O010O1000O0100oKmH4T7L`I@_6`0fI\\OZ6c0iIZOX6f0jIXOV6g0nIVOR6i0QJTOP6l0RJROn5m0UJPOl5o0WJoNi5Q1YJmNg5QOWHO3LR2Q1g5lNZH6OHV2T1c5fN^H<MEW2W1`5_NcHd0I@X2]1Y6POcG]OX2b1W6oNcG[OY2f1U6mNaJS1`5jNbJU1a5hN`JX1a5cNcJ]1^5^NfJb1\\9O1000000O10000O10001O0O2O00001N101O001N10001N101O001N101O000O2O001N101O00001O001O00001O010O00001O00001O0010O01O001O1O002N7J7H8HVSj2"}, "label": "a white polar bear that is fighting and has its mouth open"}]}
{"id": 319685, "image": "COCO_train2014_000000319685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bear\"."}], "task": "refering", "answer_template": "The \"a white bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334], "bbox": [0.38340625, 0.33172916666666663, 0.697875, 0.8810625], "iscrowd": 0, "area": 27306.2699, "rle": {"size": [480, 640], "counts": "fWc31m>2N2M3N3M2M3N2N2M4L3L4K5L5J5L4L4K5L5H7H8I7I7H9F9E;E;E<E:L4L4L4L5K4L4L4L4L5K4K5K5K5J7J500O1O2N1O1O1O1O2N1O1O100O2N1O100O2O0O100O1O2O0O100O1O2O0O100O101O0O1000000O1000000O1000000O1000000O[OTIXIl6h6WIUIi6j6\\IRId6n6_IoHa6Q7bIlH]6T7hIhHX6X7h000O010O1000O010000O010O1000O0100oKmH4T7L`I@_6`0fI\\OZ6c0iIZOX6f0jIXOV6g0nIVOR6i0QJTOP6l0RJROn5m0UJPOl5o0WJoNi5Q1YJmNg5QOWHO3LR2Q1g5lNZH6OHV2T1c5fN^H<MEW2W1`5_NcHd0I@X2]1Y6POcG]OX2b1W6oNcG[OY2f1U6mNaJS1`5jNbJU1a5hN`JX1a5cNcJ]1^5^NfJb1\\9O1000000O10000O10001O0O2O00001N101O001N10001N101O001N101O000O2O001N101O00001O001O00001O010O00001O00001O0010O01O001O1O002N7J7H8HVSj2"}, "label": "a white bear"}]}
{"id": 327881, "image": "COCO_train2014_000000327881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra walking to the right\"."}], "task": "refering", "answer_template": "The \"a zebra walking to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 65, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 190, 191, 193, 194, 208, 209, 210, 211, 212, 226, 227, 228], "bbox": [0.5201399999999999, 0.168, 0.92282, 0.9666666666666667], "iscrowd": 0, "area": 32592.217549999998, "rle": {"size": [375, 500], "counts": "e[o25`;<E;E;D:G4L4K5L4L4ZHkMb5Z2TJnMj5V2mIRNP6R2fIVNW6o1`IXN^6l1XI]Ne6g1RI`Nk6Y3M2N0O10000O100001O001O1O0016I7J6I7J6I7J7H<E;oJ]IX4R7N3N2M100O100O100O10WN^L\\Kc3V4oLfKP3_4oL^KQ3g4mLVKR3P5lLmJT3X5jLeJU3`5jL\\JW3i5hLPJ[3U6Q14M4K4L4_KTIP4[7N101N2N2O0OdMcL`L\\3U3QMhLo2W3TMfLm2Y3WMcLk2[3YM`Li2_3ZM^Lh2`3\\M\\Le2b3`MZLa2e3bMSLe2k3_MQLb2n3bMnK_2Q4eMkK]2S4iMgKX2W4nMdKS2[4f2N1O1N3M2N2N3M2N2N2N3M2N1O1O2NoG`Ln7k3I6K5J7I6K5J8H9H9F9GX1hN1N2M3N2M3N2M3N2N2O1O1DXHhLi7X3]HaLd7_3:00001O00000000O101N10`N\\MmIc2o5cMPJ]2k5iMUJU2f5RNYJn1b5XN]Jg1_5_N`Ja1Z5fNeJY1V5nNjJQ1P5VOoJi0l4^OSKb0k4ATK>l4ERK;l4X3O2O0O2O0O2N2N1L5L3M4L3M4K6K4L4N30O2O0O2ON10001O0O2O00001N101O0O1000O01O0YORHeMP8c0oG:9POi7d0SH3;VOc7g0UHK`0[O\\7i0WHCf0@T7l0[HYOk0Hk6n0]HQOP1Nd6P1^2O1O0O2N3M3M4L4L3M4L4LgV>"}, "label": "a zebra walking to the right"}]}
{"id": 327881, "image": "COCO_train2014_000000327881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on right\"."}], "task": "refering", "answer_template": "The \"zebra on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 65, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 190, 191, 193, 194, 208, 209, 210, 211, 212, 226, 227, 228], "bbox": [0.5201399999999999, 0.168, 0.92282, 0.9666666666666667], "iscrowd": 0, "area": 32592.217549999998, "rle": {"size": [375, 500], "counts": "e[o25`;<E;E;D:G4L4K5L4L4ZHkMb5Z2TJnMj5V2mIRNP6R2fIVNW6o1`IXN^6l1XI]Ne6g1RI`Nk6Y3M2N0O10000O100001O001O1O0016I7J6I7J6I7J7H<E;oJ]IX4R7N3N2M100O100O100O10WN^L\\Kc3V4oLfKP3_4oL^KQ3g4mLVKR3P5lLmJT3X5jLeJU3`5jL\\JW3i5hLPJ[3U6Q14M4K4L4_KTIP4[7N101N2N2O0OdMcL`L\\3U3QMhLo2W3TMfLm2Y3WMcLk2[3YM`Li2_3ZM^Lh2`3\\M\\Le2b3`MZLa2e3bMSLe2k3_MQLb2n3bMnK_2Q4eMkK]2S4iMgKX2W4nMdKS2[4f2N1O1N3M2N2N3M2N2N2N3M2N1O1O2NoG`Ln7k3I6K5J7I6K5J8H9H9F9GX1hN1N2M3N2M3N2M3N2N2O1O1DXHhLi7X3]HaLd7_3:00001O00000000O101N10`N\\MmIc2o5cMPJ]2k5iMUJU2f5RNYJn1b5XN]Jg1_5_N`Ja1Z5fNeJY1V5nNjJQ1P5VOoJi0l4^OSKb0k4ATK>l4ERK;l4X3O2O0O2O0O2N2N1L5L3M4L3M4K6K4L4N30O2O0O2ON10001O0O2O00001N101O0O1000O01O0YORHeMP8c0oG:9POi7d0SH3;VOc7g0UHK`0[O\\7i0WHCf0@T7l0[HYOk0Hk6n0]HQOP1Nd6P1^2O1O0O2N3M3M4L4L3M4L4LgV>"}, "label": "zebra on right"}]}
{"id": 327881, "image": "COCO_train2014_000000327881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left\"."}], "task": "refering", "answer_template": "The \"the zebra on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 41, 42, 43, 44, 58, 59, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 130, 131, 132, 133, 148, 149, 150, 151, 166, 167, 168, 169, 184, 185, 186, 187, 202, 203, 204, 205, 220, 221, 222, 223], "bbox": [0.21634, 0.13074666666666668, 0.45402, 0.9915200000000002], "iscrowd": 0, "area": 23364.43225000002, "rle": {"size": [375, 500], "counts": "YfW1;T;Z1hN<G9F9J2M4M2M4M2N3L5L5K7I9G5K4K4M3M3M3M3kKlKb0W4XOmKg0U4iNVKRNi0U3S4dNZKPNg0[3S4_NYLa1P4TNSLk1T4iMRLV2R4_MVLa2m3TM[Lk2Y6O100O10000OWOjLSIQ2YOQO_7UOTIk1FgNU7CPIg11aNW6_OkI=Hd19^NQ6ObI6L^1e0ZNj5\\3YJbLc5a3_J^L^5e3dJYLY5j3hJULU5o3kJPLQ5R1SJlNO]2n0dMm4U1bJV1c0bMm4V1bJX1b0_M^5g0RJi1_7bMSHO`0_2`800iMmFi1S9SNTGi1m8TNVGl1Z91O1O1O7I1O00O10000N2M3NiMhNTJU1k5XOkIe0V6FaI8^61\\IMc68[IFe6>YI@j6d0PI\\OT7e0hH[OZ7g0aHZOb7YO\\HS2P8[MWHd2_8WMeGT2]8gMgGY2k8001O2N2N2N2WN\\M[Jf2`5fMWJ\\2d5kMXJX2`4]MPKb0<T2\\4dMnJa0c0m1W4mMhJb0m0d1T4ChK9X41bKM]48aKG]4=bKBW1]MT1Y3`MZOZ1aMQ1\\3aMRO\\1gMj0b3fMfNf0ZMKa0g1^6]NVIF`0i1[6aN[I]O`0o1V6dNaJY1`5fNcJW1]5iNeJT1]5kNeJS1\\5mNfJP1[5oNgJn0[5ROfJl0Z5TOhJj0Y5VOiJd0[5[OjJ=Z5CiJ6[5IjJO[5OW3N2O3M3MUmS3"}, "label": "the zebra on the left"}]}
{"id": 327881, "image": "COCO_train2014_000000327881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra to the left of another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra to the left of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 41, 42, 43, 44, 58, 59, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 130, 131, 132, 133, 148, 149, 150, 151, 166, 167, 168, 169, 184, 185, 186, 187, 202, 203, 204, 205, 220, 221, 222, 223], "bbox": [0.21634, 0.13074666666666668, 0.45402, 0.9915200000000002], "iscrowd": 0, "area": 23364.43225000002, "rle": {"size": [375, 500], "counts": "YfW1;T;Z1hN<G9F9J2M4M2M4M2N3L5L5K7I9G5K4K4M3M3M3M3kKlKb0W4XOmKg0U4iNVKRNi0U3S4dNZKPNg0[3S4_NYLa1P4TNSLk1T4iMRLV2R4_MVLa2m3TM[Lk2Y6O100O10000OWOjLSIQ2YOQO_7UOTIk1FgNU7CPIg11aNW6_OkI=Hd19^NQ6ObI6L^1e0ZNj5\\3YJbLc5a3_J^L^5e3dJYLY5j3hJULU5o3kJPLQ5R1SJlNO]2n0dMm4U1bJV1c0bMm4V1bJX1b0_M^5g0RJi1_7bMSHO`0_2`800iMmFi1S9SNTGi1m8TNVGl1Z91O1O1O7I1O00O10000N2M3NiMhNTJU1k5XOkIe0V6FaI8^61\\IMc68[IFe6>YI@j6d0PI\\OT7e0hH[OZ7g0aHZOb7YO\\HS2P8[MWHd2_8WMeGT2]8gMgGY2k8001O2N2N2N2WN\\M[Jf2`5fMWJ\\2d5kMXJX2`4]MPKb0<T2\\4dMnJa0c0m1W4mMhJb0m0d1T4ChK9X41bKM]48aKG]4=bKBW1]MT1Y3`MZOZ1aMQ1\\3aMRO\\1gMj0b3fMfNf0ZMKa0g1^6]NVIF`0i1[6aN[I]O`0o1V6dNaJY1`5fNcJW1]5iNeJT1]5kNeJS1\\5mNfJP1[5oNgJn0[5ROfJl0Z5TOhJj0Y5VOiJd0[5[OjJ=Z5CiJ6[5IjJO[5OW3N2O3M3MUmS3"}, "label": "a zebra to the left of another zebra"}]}
{"id": 164042, "image": "COCO_train2014_000000164042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older man in a brown coat wearing a santa hat and holding a cell phone in his hand\"."}], "task": "refering", "answer_template": "The \"older man in a brown coat wearing a santa hat and holding a cell phone in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.05546875, 0.17079625292740047, 0.566734375, 0.9820140515222482], "iscrowd": 0, "area": 65743.89669999998, "rle": {"size": [427, 640], "counts": "SZ?6Q=7I6I8I7I6J7I7K5K4L5K5K4L5K5K4WGZMg6k2oH\\Mo6h2gH`MW7d2`HcM^7a2YHfMe7]2RHkMl7Y2kGnMS8f3N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2N200O1O100O1O100O1O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O1O1O1O1O100O1O1N2M3M3M3M3M3M3N2M3M3M3M3M3M3N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2N20000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O000000001O00000000001O000VHVMS5j2lJXMS5h2kJZMU5g2iJ[MV5f2hJ\\MW5d2gJ^MY5c2eJ_MZ5b2cJ`M]5a2aJaM^5`2`JbM_5_2^JcMb5^2\\JdMc5]2ZJfMhNZOU6Q3QKfMeN@X6k2QKgMbNC\\6g2oJiM`NF_6b2oJiM^NKa6\\2PKiNn4X1PKjNo4W1oJlNo4U1oJnNo4S1oJPOo4Q1oJQOW1cMY1^3]MROY1bMW1^3]MSOZ1bMV1^3\\MQO_1cMR1_3\\MoNa1fMo0^3\\MnNd1gMm0^3RMVOQ2aMg0f6YOdI=]6ClI3U6LUJJl55]JBc5?fJWO[5h0nJoNS5Q1TKgNl4Y1\\K_Ne4a1bKWN_4h1iKPNX4o1b310O100O0010O010O000010O01O01O01O0010O00010O01OO1O2N1O1O1O1O01O0000000000000000001N1N2O1O1O1O1O1O1O1O2O01O001O001O001O0010OO2O0O2O0O2O0O2O0O2O0O2O0O2O1N3N2M4M2M3N3L3M3N3L3M3L5K4K5L5K4K^Yc3"}, "label": "older man in a brown coat wearing a santa hat and holding a cell phone in his hand"}]}
{"id": 164042, "image": "COCO_train2014_000000164042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy in brown\"."}], "task": "refering", "answer_template": "The \"the guy in brown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.05546875, 0.17079625292740047, 0.566734375, 0.9820140515222482], "iscrowd": 0, "area": 65743.89669999998, "rle": {"size": [427, 640], "counts": "SZ?6Q=7I6I8I7I6J7I7K5K4L5K5K4L5K5K4WGZMg6k2oH\\Mo6h2gH`MW7d2`HcM^7a2YHfMe7]2RHkMl7Y2kGnMS8f3N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2N200O1O100O1O100O1O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O1O1O1O1O100O1O1N2M3M3M3M3M3M3N2M3M3M3M3M3M3N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2N20000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O000000001O00000000001O000VHVMS5j2lJXMS5h2kJZMU5g2iJ[MV5f2hJ\\MW5d2gJ^MY5c2eJ_MZ5b2cJ`M]5a2aJaM^5`2`JbM_5_2^JcMb5^2\\JdMc5]2ZJfMhNZOU6Q3QKfMeN@X6k2QKgMbNC\\6g2oJiM`NF_6b2oJiM^NKa6\\2PKiNn4X1PKjNo4W1oJlNo4U1oJnNo4S1oJPOo4Q1oJQOW1cMY1^3]MROY1bMW1^3]MSOZ1bMV1^3\\MQO_1cMR1_3\\MoNa1fMo0^3\\MnNd1gMm0^3RMVOQ2aMg0f6YOdI=]6ClI3U6LUJJl55]JBc5?fJWO[5h0nJoNS5Q1TKgNl4Y1\\K_Ne4a1bKWN_4h1iKPNX4o1b310O100O0010O010O000010O01O01O01O0010O00010O01OO1O2N1O1O1O1O01O0000000000000000001N1N2O1O1O1O1O1O1O1O2O01O001O001O001O0010OO2O0O2O0O2O0O2O0O2O0O2O0O2O1N3N2M4M2M3N3L3M3N3L3M3L5K4K5L5K4K^Yc3"}, "label": "the guy in brown"}]}
{"id": 164042, "image": "COCO_train2014_000000164042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a stocking cap and red blouse\"."}], "task": "refering", "answer_template": "The \"a woman wearing a stocking cap and red blouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 219, 220, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342], "bbox": [0.5179218750000001, 0.15459016393442623, 0.91346875, 1.0], "iscrowd": 0, "area": 48223.08505000001, "rle": {"size": [427, 640], "counts": "lfZ42X=2M2O2N2N2N1O2N2N2N2M2O2N2N2O10O1000000O11O;EO1O1O100O1O1O10O01O100O001O10O01O1O010O1000O001O00100O001O00100O001O00100O001O00010O00001O00010O00001O01O01O000010O0000000010O00000N2M3N3M2N2M3M3M3M4K4M3UEUNS:n1jEbNf9a1WFkN^9W1_FTOW9i2F9G:F9G2N3M2N2N2N2N2N2M3M3jHSJj6Q6UIPJh6S6VIoIg6T6WInIf6U6YIlIZ3MJc61eII^63hIJY61nILS63PJJQ65RJHo58TJDm5;VJBk5=YJ_Oh5`0_JYOb5g0dJRO]5m0iJlNY5S1iJkNX5U1jJhNW5W1kJgNV5X1mJeNT5Z1nJdNS5\\1oJaNR5^1QK_NP5`1RK^No4b1SK\\Nm4c1UK[Nl4d1WKYNj4f1XKXNi4h1YKVNg4j1[KSNf4l1\\KRNe4n1]KPNc4P2^KnMc4R2_KkMa4U2aKiM`4W2bKgM^4Y2cKeM^4Z2Z30000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O100001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O000O2O00001O00001O0000>B9G8H9G9G8H8H1O1O1O1K5_Oa0@?A`0_Oa0@`0@`0_Oa0@`0@`0_O`0A[1eNa0^Oa0@`hf0"}, "label": "a woman wearing a stocking cap and red blouse"}]}
{"id": 164042, "image": "COCO_train2014_000000164042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women helping other women take a picture with her cell phone\"."}], "task": "refering", "answer_template": "The \"a women helping other women take a picture with her cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 219, 220, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342], "bbox": [0.5179218750000001, 0.15459016393442623, 0.91346875, 1.0], "iscrowd": 0, "area": 48223.08505000001, "rle": {"size": [427, 640], "counts": "lfZ42X=2M2O2N2N2N1O2N2N2N2M2O2N2N2O10O1000000O11O;EO1O1O100O1O1O10O01O100O001O10O01O1O010O1000O001O00100O001O00100O001O00100O001O00010O00001O00010O00001O01O01O000010O0000000010O00000N2M3N3M2N2M3M3M3M4K4M3UEUNS:n1jEbNf9a1WFkN^9W1_FTOW9i2F9G:F9G2N3M2N2N2N2N2N2M3M3jHSJj6Q6UIPJh6S6VIoIg6T6WInIf6U6YIlIZ3MJc61eII^63hIJY61nILS63PJJQ65RJHo58TJDm5;VJBk5=YJ_Oh5`0_JYOb5g0dJRO]5m0iJlNY5S1iJkNX5U1jJhNW5W1kJgNV5X1mJeNT5Z1nJdNS5\\1oJaNR5^1QK_NP5`1RK^No4b1SK\\Nm4c1UK[Nl4d1WKYNj4f1XKXNi4h1YKVNg4j1[KSNf4l1\\KRNe4n1]KPNc4P2^KnMc4R2_KkMa4U2aKiM`4W2bKgM^4Y2cKeM^4Z2Z30000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O100001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O000O2O00001O00001O0000>B9G8H9G9G8H8H1O1O1O1K5_Oa0@?A`0_Oa0@`0@`0_Oa0@`0@`0_O`0A[1eNa0^Oa0@`hf0"}, "label": "a women helping other women take a picture with her cell phone"}]}
{"id": 164043, "image": "COCO_train2014_000000164043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child wearing blue jeans\"."}], "task": "refering", "answer_template": "The \"a child wearing blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 117, 118, 119, 120, 121, 140, 141, 142, 163, 164, 165, 186, 187, 188, 209, 210, 211, 214, 232, 233, 234, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 279, 280, 281, 283, 284, 285, 302, 303, 304, 308, 327, 330, 331, 349, 350, 353, 354, 371, 372, 373, 374, 376, 377], "bbox": [0.10089062499999998, 0.06277083333333333, 0.43721875, 1.0], "iscrowd": 0, "area": 31977.58145, "rle": {"size": [480, 640], "counts": "[gn0:_>:E:F;E;F9I8J6hMgM`G_2\\8eM_G_2^8fM[G`2a8eMYG`2c8dMWGc2d8bMVGd2e8`MUGg2e8_MTGh2g8]MSGi2h8[MRGl2i8YMQGn2i8i1WNiI`J\\6[5fIcJ_6X5cIeJc6V5_IhJf6S5\\IkJi6P5YInJl6m4WIPKn6k4TISKQ7h4QIVKT7e4nHXKd7P1iGY1d0eMd7Q1jGV1f0gMa7S1kGR1h0hM]7W1lGm0k0jMZ7X1mGk0l0kMX7Z1mGg0o0mMT7]1nGc0Q1mMS7_1oG?R1PNP7a1oG;U1RNl6c1QH8W1SNh6e1RH4Z1VNd6f1TH0[1YN`6g1WHL]1\\N\\6h1XHI_1^NY6i1YHEb1aNT6j1]H@d1dNo5l1^H]Oj4lN^Og1iKZOg4UO^O`1lK[Oa4]O_OY1PLZO]4CAS1RLZOX4KCj0VL[OS42Cd0ZLZOQ46C`0]LXOo3=A:aLYOl3b0_O6eLXOj3f0_O2gLXOh3k0^OLkLYOe3o0]OInLWOd3U1[OCSMWO`3[1ZO^OVMWO^3_1YO[OYMVO[3e1YOTO]MWOX3j1XOoN`MVOW3o1VOlNcMUOU3T2UOfNhMUOR3Y2UO`NiMWOR3\\2TO[NjMYOQ3`2TOTNlM[OQ3d2ROoMmM]OQ3h2POiMoM_OQ3k2oNdMQN@P3o2nN^MSNCn2S3mNYMUNCo2W3kNTMVNEo2Z3jNoLWNGo2]3iNjLXNIo2d3dN`L^NLn2n3\\NTLfNNm2Y4UNgKmN1n2b4mM[KUO3n2l4dMPK^O4n2V5\\MdJF6n2`5TMWJO9l2k5mLjI7;l2U6eL^I?=l2_6]LRIg0?l2i6ULgHn0`0m2S7mK]HT1`0o2g7nLZHQ3i7mLWHS3j7lLVHT3l7jLTHV3n7hLRHX3P8fLPHZ3Q8eLoGZ3T8dLlG\\3V8bLjG^3X8`LhG`3Y8_LhG`3Z8]LhGa3[8]LfGa3]8]LdG`3_8_LbGd2lNiMf9A_Gd2mNiMg9A]Gc2POhMg9CZGc2QOgMi9DXGb2QOgMj9FVG`2SOgMX;W2jDgMX;V2lDfMW;W2k0M3MYInMe0n1\\OTNe0i1ZOZN`1j0`NXOg1?YNCl16SNLR2MoM5U2EkM=Y83N2N2M6K4L5K5J5L5K4M4N1N3N2M2O2N1N3N0O100O10O0100O10000O100O100O100O01O001O001O1O001OdEfMR8Z2fGmM\\8R2\\GVNd8j1SGaNl8^1mFjNR9V1nFlNQ9S1oFoNo8Q1QGPOQ9m0oFUOS9g0mF[OT9b0mF^OV9>jFDW99iFIY93gFN[9OeF3]9IcF7`9F`F;c9A^F>e9_O[Fb0f9\\OZFd0g9[OYFf0h9XOXFh0j9VOVFk0j9TOVFl0l9ROUFn0k9QOUFo0m9oNSFR1m9mNSFS1o9kNQFV1P:hNPFX1Q:gNPFY1Q:eNoE[1R:eNmE[1U:eNlEX1V:gNmET1U:mNlEP1V:POkEm0V:SOmEi0U:WOlEf0U:ZOmEc0U:]OlE`0V:_OmE=T:DmE9U:FmE6U:KmE1U:NmEOU:1mEKT:5oEGS:9oECR:>Y2I6JY\\X5"}, "label": "a child wearing blue jeans"}]}
{"id": 164043, "image": "COCO_train2014_000000164043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child on shoulders of man with black hat\"."}], "task": "refering", "answer_template": "The \"child on shoulders of man with black hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 117, 118, 119, 120, 121, 140, 141, 142, 163, 164, 165, 186, 187, 188, 209, 210, 211, 214, 232, 233, 234, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 279, 280, 281, 283, 284, 285, 302, 303, 304, 308, 327, 330, 331, 349, 350, 353, 354, 371, 372, 373, 374, 376, 377], "bbox": [0.10089062499999998, 0.06277083333333333, 0.43721875, 1.0], "iscrowd": 0, "area": 31977.58145, "rle": {"size": [480, 640], "counts": "[gn0:_>:E:F;E;F9I8J6hMgM`G_2\\8eM_G_2^8fM[G`2a8eMYG`2c8dMWGc2d8bMVGd2e8`MUGg2e8_MTGh2g8]MSGi2h8[MRGl2i8YMQGn2i8i1WNiI`J\\6[5fIcJ_6X5cIeJc6V5_IhJf6S5\\IkJi6P5YInJl6m4WIPKn6k4TISKQ7h4QIVKT7e4nHXKd7P1iGY1d0eMd7Q1jGV1f0gMa7S1kGR1h0hM]7W1lGm0k0jMZ7X1mGk0l0kMX7Z1mGg0o0mMT7]1nGc0Q1mMS7_1oG?R1PNP7a1oG;U1RNl6c1QH8W1SNh6e1RH4Z1VNd6f1TH0[1YN`6g1WHL]1\\N\\6h1XHI_1^NY6i1YHEb1aNT6j1]H@d1dNo5l1^H]Oj4lN^Og1iKZOg4UO^O`1lK[Oa4]O_OY1PLZO]4CAS1RLZOX4KCj0VL[OS42Cd0ZLZOQ46C`0]LXOo3=A:aLYOl3b0_O6eLXOj3f0_O2gLXOh3k0^OLkLYOe3o0]OInLWOd3U1[OCSMWO`3[1ZO^OVMWO^3_1YO[OYMVO[3e1YOTO]MWOX3j1XOoN`MVOW3o1VOlNcMUOU3T2UOfNhMUOR3Y2UO`NiMWOR3\\2TO[NjMYOQ3`2TOTNlM[OQ3d2ROoMmM]OQ3h2POiMoM_OQ3k2oNdMQN@P3o2nN^MSNCn2S3mNYMUNCo2W3kNTMVNEo2Z3jNoLWNGo2]3iNjLXNIo2d3dN`L^NLn2n3\\NTLfNNm2Y4UNgKmN1n2b4mM[KUO3n2l4dMPK^O4n2V5\\MdJF6n2`5TMWJO9l2k5mLjI7;l2U6eL^I?=l2_6]LRIg0?l2i6ULgHn0`0m2S7mK]HT1`0o2g7nLZHQ3i7mLWHS3j7lLVHT3l7jLTHV3n7hLRHX3P8fLPHZ3Q8eLoGZ3T8dLlG\\3V8bLjG^3X8`LhG`3Y8_LhG`3Z8]LhGa3[8]LfGa3]8]LdG`3_8_LbGd2lNiMf9A_Gd2mNiMg9A]Gc2POhMg9CZGc2QOgMi9DXGb2QOgMj9FVG`2SOgMX;W2jDgMX;V2lDfMW;W2k0M3MYInMe0n1\\OTNe0i1ZOZN`1j0`NXOg1?YNCl16SNLR2MoM5U2EkM=Y83N2N2M6K4L5K5J5L5K4M4N1N3N2M2O2N1N3N0O100O10O0100O10000O100O100O100O01O001O001O1O001OdEfMR8Z2fGmM\\8R2\\GVNd8j1SGaNl8^1mFjNR9V1nFlNQ9S1oFoNo8Q1QGPOQ9m0oFUOS9g0mF[OT9b0mF^OV9>jFDW99iFIY93gFN[9OeF3]9IcF7`9F`F;c9A^F>e9_O[Fb0f9\\OZFd0g9[OYFf0h9XOXFh0j9VOVFk0j9TOVFl0l9ROUFn0k9QOUFo0m9oNSFR1m9mNSFS1o9kNQFV1P:hNPFX1Q:gNPFY1Q:eNoE[1R:eNmE[1U:eNlEX1V:gNmET1U:mNlEP1V:POkEm0V:SOmEi0U:WOlEf0U:ZOmEc0U:]OlE`0V:_OmE=T:DmE9U:FmE6U:KmE1U:NmEOU:1mEKT:5oEGS:9oECR:>Y2I6JY\\X5"}, "label": "child on shoulders of man with black hat"}]}
{"id": 164043, "image": "COCO_train2014_000000164043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 232, 235, 236, 254, 255, 258, 259, 276, 277, 278, 281, 282, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 374, 375, 376], "bbox": [0.0016875000000000002, 0.3325833333333333, 0.43314062499999995, 0.9820208333333333], "iscrowd": 0, "area": 33079.63495, "rle": {"size": [480, 640], "counts": "gl0l0j<Z1F:F:M3M3M3M3N2N2N2O1N2N2O1N2N2N2O1N2N2O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O2O0O1O1O1O101N1O1O1O100O2O000000001O00000O100000XMdFK]90hF0X9JmF6T9DRG<n8^OXGa0h8ZO]Gg0e8QOaGn0d8iNaGW1b8WNlGi1f:00O010O100O010O100O10O0100O100O010O100O10O0100000000O0100000000O01000000000_JYNiMf1R2dNjM\\1P2oNkMQ1o1WOoMi0j1@TN`0f1GYN9c1L\\N3d1O[N1d11[NOd13[NMe14]NIb1=\\N@c1e0\\NXOc1n0ZNoNf1W1XNfNh1^1UNaNj1e1SNYNl1m1QNlMS2Y2jM]M_2h2^MRMe2T3XMeLl2`3RMYLR3m3n34K4L5L4K5L3M9]IPKc2_5dLlJ\\3\\5VLjJj3Z5QLgJo3\\5lKeJU4_5eKcJ[4`5\\KfJd4^5RKhJn4[5jJiJW5[5_JkJa5X5VJnJj5V5lIPKT6m6100000ObKlI1U6NoIOQ60SJMm52WJKi54[JHf57^JFb59aJE_5:cJF\\59gJFX59kJGS58PKGo48SKHl47WKHh47[KIc46`KI_46dKI[46gKJX45kKIU46nKIQ46RLHn38SLHl38VLFj3:XLDh3;[LDd3<_LAa3?bL^O^3a0fL]OY3b0iL]OY3`0iL@X3<kLCW3:lLDV38mLHT35nLJS33PMLR31QMNP3OQM1Q3KRM4P3IRM7o2nMkNQ2V1_MYOb2c6O1O010O1O10O0100O1O010O1O010O100O00100O00100O1J5J7[Oe0SO8K^BOg=3WBHl=;TB@P>a080N2O1OYOiAf0X>1N101OWOkAg0U>201O010O00001O0010O0001O001O00010O001O0003M3M3M4L3M3M1O001O000OhhY5"}, "label": "a man wearing a blue shirt"}]}
{"id": 164043, "image": "COCO_train2014_000000164043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with the blue shirt and a black hat with a child upon his shoulders\"."}], "task": "refering", "answer_template": "The \"a man with the blue shirt and a black hat with a child upon his shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 232, 235, 236, 254, 255, 258, 259, 276, 277, 278, 281, 282, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 374, 375, 376], "bbox": [0.0016875000000000002, 0.3325833333333333, 0.43314062499999995, 0.9820208333333333], "iscrowd": 0, "area": 33079.63495, "rle": {"size": [480, 640], "counts": "gl0l0j<Z1F:F:M3M3M3M3N2N2N2O1N2N2O1N2N2N2O1N2N2O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O2O0O1O1O1O101N1O1O1O100O2O000000001O00000O100000XMdFK]90hF0X9JmF6T9DRG<n8^OXGa0h8ZO]Gg0e8QOaGn0d8iNaGW1b8WNlGi1f:00O010O100O010O100O10O0100O100O010O100O10O0100000000O0100000000O01000000000_JYNiMf1R2dNjM\\1P2oNkMQ1o1WOoMi0j1@TN`0f1GYN9c1L\\N3d1O[N1d11[NOd13[NMe14]NIb1=\\N@c1e0\\NXOc1n0ZNoNf1W1XNfNh1^1UNaNj1e1SNYNl1m1QNlMS2Y2jM]M_2h2^MRMe2T3XMeLl2`3RMYLR3m3n34K4L5L4K5L3M9]IPKc2_5dLlJ\\3\\5VLjJj3Z5QLgJo3\\5lKeJU4_5eKcJ[4`5\\KfJd4^5RKhJn4[5jJiJW5[5_JkJa5X5VJnJj5V5lIPKT6m6100000ObKlI1U6NoIOQ60SJMm52WJKi54[JHf57^JFb59aJE_5:cJF\\59gJFX59kJGS58PKGo48SKHl47WKHh47[KIc46`KI_46dKI[46gKJX45kKIU46nKIQ46RLHn38SLHl38VLFj3:XLDh3;[LDd3<_LAa3?bL^O^3a0fL]OY3b0iL]OY3`0iL@X3<kLCW3:lLDV38mLHT35nLJS33PMLR31QMNP3OQM1Q3KRM4P3IRM7o2nMkNQ2V1_MYOb2c6O1O010O1O10O0100O1O010O1O010O100O00100O00100O1J5J7[Oe0SO8K^BOg=3WBHl=;TB@P>a080N2O1OYOiAf0X>1N101OWOkAg0U>201O010O00001O0010O0001O001O00010O001O0003M3M3M4L3M3M1O001O000OhhY5"}, "label": "a man with the blue shirt and a black hat with a child upon his shoulders"}]}
{"id": 491727, "image": "COCO_train2014_000000491727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza sitting on a table with a gentlemen eating one of them and making a funny face\"."}], "task": "refering", "answer_template": "The \"a pizza sitting on a table with a gentlemen eating one of them and making a funny face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252], "bbox": [0.637078125, 0.5685277777777777, 0.9985937500000001, 0.8516944444444445], "iscrowd": 0, "area": 12619.067150000003, "rle": {"size": [360, 640], "counts": "Qe_44P;5J7L3O1N3M2O1N2O2M2O1N3N1N2N3N1N2O1N2O0010O100O010O100O10O01O100O100O010O10O0010O01O00001O010O001O001O0010O01O001O00010O001O001O00010O00001O000010O0001O000010O0001O00001O01O01O00001O000010O0001O00001O00001O01O01O00001O00001O01O01O00001O00001O01O00000001O00000001O000001O000000000000001O0000000000001O000001O00000001O000000000000001O0000000000001O0000000000000010O001O00001O0010O01O001O000010O01O001O00001O010O00001O001O00010O001O00001O010O00001O001O00010D;@a0@i3"}, "label": "a pizza sitting on a table with a gentlemen eating one of them and making a funny face"}]}
{"id": 491727, "image": "COCO_train2014_000000491727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza slices on the platter\"."}], "task": "refering", "answer_template": "The \"pizza slices on the platter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252], "bbox": [0.637078125, 0.5685277777777777, 0.9985937500000001, 0.8516944444444445], "iscrowd": 0, "area": 12619.067150000003, "rle": {"size": [360, 640], "counts": "Qe_44P;5J7L3O1N3M2O1N2O2M2O1N3N1N2N3N1N2O1N2O0010O100O010O100O10O01O100O100O010O10O0010O01O00001O010O001O001O0010O01O001O00010O001O001O00010O00001O000010O0001O000010O0001O00001O01O01O00001O000010O0001O00001O00001O01O01O00001O00001O01O01O00001O00001O01O00000001O00000001O000001O000000000000001O0000000000001O000001O00000001O000000000000001O0000000000001O0000000000000010O001O00001O0010O01O001O000010O01O001O00001O010O00001O001O00010O001O00001O010O00001O001O00010D;@a0@i3"}, "label": "pizza slices on the platter"}]}
{"id": 155860, "image": "COCO_train2014_000000155860.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana fully blocked by an apple\"."}], "task": "refering", "answer_template": "The \"a banana fully blocked by an apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 136, 137, 138, 139, 153, 154, 155, 170, 171, 172, 187, 188, 189, 205, 206], "bbox": [0.003, 0.07865625000000001, 0.31460416666666663, 0.5528124999999999], "iscrowd": 0, "area": 21449.1208, "rle": {"size": [640, 480], "counts": "Tm06Rb0i1WNh1WNj1WNi1_Oa0001O0O10001O00000O2O0000001O0O10001O00000O2O0000001N1000001O000O101OO10000O01000O100O10O10O10000O10O10O10000O10O10O10000O0100ZOe0gNZ1I7J5J7I7I6J7I7I6L5M3N1O2N2M2O2N2M2O2N2M2O2N2N1N3N2N1N3N2N1N3N2N1O2M3N1O2M3N1O2N2M2O2N2N1O2N2O0O2N2N101N2N101N2N1O2O0O2N2N101N2N1O2O1N1O2N2O0O2N2O0O2N2N1O2N2N1Ok^]6"}, "label": "a banana fully blocked by an apple"}]}
{"id": 155860, "image": "COCO_train2014_000000155860.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana closest to apple\"."}], "task": "refering", "answer_template": "The \"banana closest to apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 136, 137, 138, 139, 153, 154, 155, 170, 171, 172, 187, 188, 189, 205, 206], "bbox": [0.003, 0.07865625000000001, 0.31460416666666663, 0.5528124999999999], "iscrowd": 0, "area": 21449.1208, "rle": {"size": [640, 480], "counts": "Tm06Rb0i1WNh1WNj1WNi1_Oa0001O0O10001O00000O2O0000001O0O10001O00000O2O0000001N1000001O000O101OO10000O01000O100O10O10O10000O10O10O10000O10O10O10000O0100ZOe0gNZ1I7J5J7I7I6J7I7I6L5M3N1O2N2M2O2N2M2O2N2M2O2N2N1N3N2N1N3N2N1N3N2N1O2M3N1O2M3N1O2N2M2O2N2N1O2N2O0O2N2N101N2N101N2N1O2O0O2N2N101N2N1O2O1N1O2N2O0O2N2O0O2N2N1O2N2N1Ok^]6"}, "label": "banana closest to apple"}]}
{"id": 491733, "image": "COCO_train2014_000000491733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large fishing boat with an american flag\"."}], "task": "refering", "answer_template": "The \"a large fishing boat with an american flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 201, 202, 216, 218, 219, 220, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274], "bbox": [0.40784374999999995, 0.5483129584352078, 0.92053125, 0.766283618581907], "iscrowd": 0, "area": 18086.627100000005, "rle": {"size": [409, 640], "counts": "SaX36X<;F:L41O0000000000001O01O000000000001O00000001O000001O0000000000000010O0000000000000QOVOPFi0P:ZOmEf0S:]O_En0a:f0O000000000O1000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000L4J6J600001O001O001O001O001O001O001O002N2N3M2N1O000000000000000UN^E_1b:`N`E_1`:aNbE]1^:cNdE[1\\:eNfEY1Z:fNiEX1W:hNkEV1U:jNlEU1V:iNlEU1V:hNmEV1V:gNlEX1l:N3M5K1O0000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O1O1O1O1N2N2M3N6J6IaRd0"}, "label": "a large fishing boat with an american flag"}]}
{"id": 491733, "image": "COCO_train2014_000000491733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long boat with people in front and in back\"."}], "task": "refering", "answer_template": "The \"a long boat with people in front and in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 201, 202, 216, 218, 219, 220, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274], "bbox": [0.40784374999999995, 0.5483129584352078, 0.92053125, 0.766283618581907], "iscrowd": 0, "area": 18086.627100000005, "rle": {"size": [409, 640], "counts": "SaX36X<;F:L41O0000000000001O01O000000000001O00000001O000001O0000000000000010O0000000000000QOVOPFi0P:ZOmEf0S:]O_En0a:f0O000000000O1000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000L4J6J600001O001O001O001O001O001O001O002N2N3M2N1O000000000000000UN^E_1b:`N`E_1`:aNbE]1^:cNdE[1\\:eNfEY1Z:fNiEX1W:hNkEV1U:jNlEU1V:iNlEU1V:hNmEV1V:gNlEX1l:N3M5K1O0000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O1O1O1O1N2N2M3N6J6IaRd0"}, "label": "a long boat with people in front and in back"}]}
{"id": 278744, "image": "COCO_train2014_000000278744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant to the right of the blue sign\"."}], "task": "refering", "answer_template": "The \"a plant to the right of the blue sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 132, 133, 134, 155, 156, 157], "bbox": [0.680953125, 0.0, 0.971140625, 0.4665258215962442], "iscrowd": 0, "area": 19988.2695, "rle": {"size": [426, 640], "counts": "\\\\e57o<8L4L0000001O00001O0000001O00000000000000000000001O000000000000001O00000000001O1TE2Z8N\\GU3k5lLlI\\3T6kL]I]3c6lLlH\\3T7]101O0000000000000000001O0000000000000000001O000000000000000000000000000000000000000000000000O10000000000000000O1000000O1N2K5J6J6J6J6K5J6J6J6M3N2M3N2M3M3L4XOaF`Md9`2_FYMe9f2aFPMd9P3;00000000O1000000^NgE3Y:ImE4T:KnE4R:KPF4P:LPF4P:LPF4P:KQF5o9KQF4P:LPF4P:KQF5o9KRF4n9LRF4n9KSF5m9KSF5m9KSF5m9JTF6l9JTF5m9KSF5m9JTF6l9JTF6l9JTFDAI[:b0UFBCLX:b0UF@ENV:b0UF^OG0T:a0VF\\OI3Q:6dEDb01K5o95eECd01I7n95fEAe00I:l94hE_Of01G<k93iE_Og0OG?i93]GLd83]GMc83]GMc82^GNb81_GOa81_GOa80`G0`80`G0`8OaG0`80`G0`8OaG1_8OaG1_8NcG0^80bG0^8OdG0\\80dGO]80dG0\\8OfG0\\8NdG2^8KcG4_8JhG0Y8Ooa7"}, "label": "a plant to the right of the blue sign"}]}
{"id": 278744, "image": "COCO_train2014_000000278744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a potted plant above a tennis racket\"."}], "task": "refering", "answer_template": "The \"a potted plant above a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 132, 133, 134, 155, 156, 157], "bbox": [0.680953125, 0.0, 0.971140625, 0.4665258215962442], "iscrowd": 0, "area": 19988.2695, "rle": {"size": [426, 640], "counts": "\\\\e57o<8L4L0000001O00001O0000001O00000000000000000000001O000000000000001O00000000001O1TE2Z8N\\GU3k5lLlI\\3T6kL]I]3c6lLlH\\3T7]101O0000000000000000001O0000000000000000001O000000000000000000000000000000000000000000000000O10000000000000000O1000000O1N2K5J6J6J6J6K5J6J6J6M3N2M3N2M3M3L4XOaF`Md9`2_FYMe9f2aFPMd9P3;00000000O1000000^NgE3Y:ImE4T:KnE4R:KPF4P:LPF4P:LPF4P:KQF5o9KQF4P:LPF4P:KQF5o9KRF4n9LRF4n9KSF5m9KSF5m9KSF5m9JTF6l9JTF5m9KSF5m9JTF6l9JTF6l9JTFDAI[:b0UFBCLX:b0UF@ENV:b0UF^OG0T:a0VF\\OI3Q:6dEDb01K5o95eECd01I7n95fEAe00I:l94hE_Of01G<k93iE_Og0OG?i93]GLd83]GMc83]GMc82^GNb81_GOa81_GOa80`G0`80`G0`8OaG0`80`G0`8OaG1_8OaG1_8NcG0^80bG0^8OdG0\\80dGO]80dG0\\8OfG0\\8NdG2^8KcG4_8JhG0Y8Ooa7"}, "label": "a potted plant above a tennis racket"}]}
{"id": 155864, "image": "COCO_train2014_000000155864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy in the purple hat that just swung the bat\"."}], "task": "refering", "answer_template": "The \"the guy in the purple hat that just swung the bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 95, 98, 99, 100, 101, 117, 118, 119, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331], "bbox": [0.1244375, 0.2243559718969555, 0.5427500000000001, 0.9883840749414521], "iscrowd": 0, "area": 50632.840299999996, "rle": {"size": [427, 640], "counts": "T`Q1Q1X<3L5L4L3L5L3L5L4M2O2OO100O1O010O100O010O1O100O100O1O100O100O100O1O100O1000000O010000O01000O10O10eF\\Nm6e1PI`Nn6`1oHdNP7[1nHjNo6W1nHmNQ7S1lHROR7n0kHWOR7i0lH[OS7e0jH@S7a0kHCS7=jHGU78iHMT74kHOS71lH1S7OlH3R7MmH6R7JmH8R7HmH:Q7GnH;\\1mMQ3h1aK>X1nMW3d1aK`0Q1QN^3`1`Ka0l0SNd3]1_Kb0f0VNk3Y1]Kd0b0WNR4T1\\Kg0<YNX4Q1[Kh06\\N_4b4YKdKg4`600001O00000000jNZHgLf7X3^HfLa7Z3cHbL]7]3fHaLZ7_3fHaLZ7^3hHaLX7_3iH`LW7_3jHbLU7^3lHaLT7^3nHaLR7_3nHaLR7^3PIaLP7_3QI`Lo6_3RIbLm6^3TIaLl6^3VIaLj6^3WIbLi6^3WIbLi6]3WIdLi6\\3WIdLi6[3WIfLi6Y3WIhLi6V3YIjLg6U3YIlLg6R3[ImLf6R3ZIoLf6o2\\IZL^O4V7a3RJ]Ln5a3UJ^Lk5a3WJ^Li5`3[J^Le5a3S2N2O1N200000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000O1000000O1000000001O00001O001O001O001O00001O001O001O001O001O00001O0lMiLTJX3i5VMkIj2S6dMaI]2]6PNWIP2g6^NmHf1n6`2J6UJeI]4_6YKhIe4^6PKiIn4]6gJjIX5U7N2O2M2N2O1O1O1N2O1O1O1N3N1O1O1N2O1O1O1O1N2O2N1O1N2O1O1O1N2O1O1O2M2O1O1O2N2M3N2N2M3M2N3M3M3M3N2M3M3M2N3M3M3M3M3M3N4K6J6J6J6J6J6J6J6J6K5Jeni3"}, "label": "the guy in the purple hat that just swung the bat"}]}
{"id": 155864, "image": "COCO_train2014_000000155864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black , purple , and grey uniform practicing batting\"."}], "task": "refering", "answer_template": "The \"a man in a black , purple , and grey uniform practicing batting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 95, 98, 99, 100, 101, 117, 118, 119, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331], "bbox": [0.1244375, 0.2243559718969555, 0.5427500000000001, 0.9883840749414521], "iscrowd": 0, "area": 50632.840299999996, "rle": {"size": [427, 640], "counts": "T`Q1Q1X<3L5L4L3L5L3L5L4M2O2OO100O1O010O100O010O1O100O100O1O100O100O100O1O100O1000000O010000O01000O10O10eF\\Nm6e1PI`Nn6`1oHdNP7[1nHjNo6W1nHmNQ7S1lHROR7n0kHWOR7i0lH[OS7e0jH@S7a0kHCS7=jHGU78iHMT74kHOS71lH1S7OlH3R7MmH6R7JmH8R7HmH:Q7GnH;\\1mMQ3h1aK>X1nMW3d1aK`0Q1QN^3`1`Ka0l0SNd3]1_Kb0f0VNk3Y1]Kd0b0WNR4T1\\Kg0<YNX4Q1[Kh06\\N_4b4YKdKg4`600001O00000000jNZHgLf7X3^HfLa7Z3cHbL]7]3fHaLZ7_3fHaLZ7^3hHaLX7_3iH`LW7_3jHbLU7^3lHaLT7^3nHaLR7_3nHaLR7^3PIaLP7_3QI`Lo6_3RIbLm6^3TIaLl6^3VIaLj6^3WIbLi6^3WIbLi6]3WIdLi6\\3WIdLi6[3WIfLi6Y3WIhLi6V3YIjLg6U3YIlLg6R3[ImLf6R3ZIoLf6o2\\IZL^O4V7a3RJ]Ln5a3UJ^Lk5a3WJ^Li5`3[J^Le5a3S2N2O1N200000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000O1000000O1000000001O00001O001O001O001O00001O001O001O001O001O00001O0lMiLTJX3i5VMkIj2S6dMaI]2]6PNWIP2g6^NmHf1n6`2J6UJeI]4_6YKhIe4^6PKiIn4]6gJjIX5U7N2O2M2N2O1O1O1N2O1O1O1N3N1O1O1N2O1O1O1O1N2O2N1O1N2O1O1O1N2O1O1O2M2O1O1O2N2M3N2N2M3M2N3M3M3M3N2M3M3M2N3M3M3M3M3M3N4K6J6J6J6J6J6J6J6J6K5Jeni3"}, "label": "a man in a black , purple , and grey uniform practicing batting"}]}
{"id": 155864, "image": "COCO_train2014_000000155864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player standing behind the player hitting the ball\"."}], "task": "refering", "answer_template": "The \"a baseball player standing behind the player hitting the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 229, 252, 275], "bbox": [0.80678125, 0.34571428571428575, 0.9938437500000001, 0.750023419203747], "iscrowd": 0, "area": 5673.030349999991, "rle": {"size": [427, 640], "counts": "b\\g63P==K6J0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000QOo0eN[1100O1O2N1O1O1O100O2N]GVMf6i2YIXMh6g2gHXMUO1U8f2_H`M[OKW8d2WHjM_OC\\8b2mGTNDZOa8a2cG^NKQOc8\\41O1N2K6J5K5K5J6XOb`1"}, "label": "a baseball player standing behind the player hitting the ball"}]}
{"id": 155864, "image": "COCO_train2014_000000155864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing behind the batting cage , to the very right of the other men\"."}], "task": "refering", "answer_template": "The \"man standing behind the batting cage , to the very right of the other men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 229, 252, 275], "bbox": [0.80678125, 0.34571428571428575, 0.9938437500000001, 0.750023419203747], "iscrowd": 0, "area": 5673.030349999991, "rle": {"size": [427, 640], "counts": "b\\g63P==K6J0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000QOo0eN[1100O1O2N1O1O1O100O2N]GVMf6i2YIXMh6g2gHXMUO1U8f2_H`M[OKW8d2WHjM_OC\\8b2mGTNDZOa8a2cG^NKQOc8\\41O1N2K6J5K5K5J6XOb`1"}, "label": "man standing behind the batting cage , to the very right of the other men"}]}
{"id": 319714, "image": "COCO_train2014_000000319714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing\"."}], "task": "refering", "answer_template": "The \"man standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 122, 123, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 211, 212, 234, 235, 256, 257, 280], "bbox": [0.15826562500000002, 0.25035416666666666, 0.39114062499999996, 0.724125], "iscrowd": 0, "area": 9513.15085, "rle": {"size": [480, 640], "counts": "Pb_15f>;G3M4L5hDXOR8m0iGTOW8P1eGROY8Q1cGROZ8R1cGoN\\8U1_GnN_8U1\\GoNb8V1QGSOm8S1kFROS9Q1hFSOV9P1dFTOZ9Q1_FUO^9m0]FXOa9j0YF]Od9j0PF]Om9l2O100N2O1N2N2O0OO200ZO\\FQLc9k3fFRLY9k3lFTLS9i3RGVLl8h3XGWLh8f3]GYLb8g3`GXL^8h3dGXL[8i3fGVLY8j3hGVLV8k3lGTLS8l3oGSLP8n3PHRLn7\\2iFmN]1dNk7S2jFmM<Q1]:S1SFaNm9a1]FTNd9m1h1N101N2O0O2OJPN]CP2`<8O2O1N202O2M3M3M3L4M3M2O1O1N2O1O0010hN`DoN_;f0nDYOQ;<\\EBd:;bEC^:<dEC[:=gEjNkN?^;g0PFYOP:f0QFZOn9f0SFZOm9e0SF[Om9e0TF[Om9c0TF]Ol9b0UF]Ol9b0UF^Ok9b0TF_Om9H]D;g1Lm9HbD6b12l9GbD7c11l9GbD7b12m9G`D7d11l9HaD6d11k9IaD6e10k9I`D7e10k9I`D7f1Oj9IaD8f1Oi9FcD;e1NV:2jENV:2kEMU:3lEMS:3mEMR:4oEMo93QFOm91SF1j91VFNj92VFOi91WF0h90XF1g90XF0h90XF1h9NXF2h9NYF2f9NZF2f9OZF1e9O[F0f90[FNf91\\FMe93]FIf96\\FFf99a200O101N100O10001N1000000000O2O1O0O100O101OO010O2N1N3O1NYef5"}, "label": "man standing"}]}
{"id": 319714, "image": "COCO_train2014_000000319714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a surfboard , paddling with his dog on the board\"."}], "task": "refering", "answer_template": "The \"a man on a surfboard , paddling with his dog on the board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 122, 123, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 211, 212, 234, 235, 256, 257, 280], "bbox": [0.15826562500000002, 0.25035416666666666, 0.39114062499999996, 0.724125], "iscrowd": 0, "area": 9513.15085, "rle": {"size": [480, 640], "counts": "Pb_15f>;G3M4L5hDXOR8m0iGTOW8P1eGROY8Q1cGROZ8R1cGoN\\8U1_GnN_8U1\\GoNb8V1QGSOm8S1kFROS9Q1hFSOV9P1dFTOZ9Q1_FUO^9m0]FXOa9j0YF]Od9j0PF]Om9l2O100N2O1N2N2O0OO200ZO\\FQLc9k3fFRLY9k3lFTLS9i3RGVLl8h3XGWLh8f3]GYLb8g3`GXL^8h3dGXL[8i3fGVLY8j3hGVLV8k3lGTLS8l3oGSLP8n3PHRLn7\\2iFmN]1dNk7S2jFmM<Q1]:S1SFaNm9a1]FTNd9m1h1N101N2O0O2OJPN]CP2`<8O2O1N202O2M3M3M3L4M3M2O1O1N2O1O0010hN`DoN_;f0nDYOQ;<\\EBd:;bEC^:<dEC[:=gEjNkN?^;g0PFYOP:f0QFZOn9f0SFZOm9e0SF[Om9e0TF[Om9c0TF]Ol9b0UF]Ol9b0UF^Ok9b0TF_Om9H]D;g1Lm9HbD6b12l9GbD7c11l9GbD7b12m9G`D7d11l9HaD6d11k9IaD6e10k9I`D7e10k9I`D7f1Oj9IaD8f1Oi9FcD;e1NV:2jENV:2kEMU:3lEMS:3mEMR:4oEMo93QFOm91SF1j91VFNj92VFOi91WF0h90XF1g90XF0h90XF1h9NXF2h9NYF2f9NZF2f9OZF1e9O[F0f90[FNf91\\FMe93]FIf96\\FFf99a200O101N100O10001N1000000000O2O1O0O100O101OO010O2N1N3O1NYef5"}, "label": "a man on a surfboard , paddling with his dog on the board"}]}
{"id": 16616, "image": "COCO_train2014_000000016616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a back of a white pickup truck in traffic\"."}], "task": "refering", "answer_template": "The \"a back of a white pickup truck in traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119], "bbox": [0.0, 0.03953431372549019, 0.203046875, 0.405906862745098], "iscrowd": 0, "area": 15117.117350000004, "rle": {"size": [408, 640], "counts": "f0c3U9000O100000000O101O1O2N2M2O2N1O2N0O0100O10O10O100O10O10O100O2O000O100O10000O101N10000O1001O0001O00000000000000001O0000000003M2N2O1N3M2N2N2N3M2N10O000001O000001O1O001O010O001O1O001O00100N3N2M3N2M2O2M3M3N2M3N2M3N1N3N1N010O0001O010O0011[MdFk1]9UNfFh1[9WNfFg1[9XNhFf1Y9YNhFe1Z9ZNgFV1h9jNZFd0W:[OjE`0Z:VORFe0S;K6K4K5LdS[6"}, "label": "a back of a white pickup truck in traffic"}]}
{"id": 16616, "image": "COCO_train2014_000000016616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ford pickup truck in traffic\"."}], "task": "refering", "answer_template": "The \"a ford pickup truck in traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119], "bbox": [0.0, 0.03953431372549019, 0.203046875, 0.405906862745098], "iscrowd": 0, "area": 15117.117350000004, "rle": {"size": [408, 640], "counts": "f0c3U9000O100000000O101O1O2N2M2O2N1O2N0O0100O10O10O100O10O10O100O2O000O100O10000O101N10000O1001O0001O00000000000000001O0000000003M2N2O1N3M2N2N2N3M2N10O000001O000001O1O001O010O001O1O001O00100N3N2M3N2M2O2M3M3N2M3N2M3N1N3N1N010O0001O010O0011[MdFk1]9UNfFh1[9WNfFg1[9XNhFf1Y9YNhFe1Z9ZNgFV1h9jNZFd0W:[OjE`0Z:VORFe0S;K6K4K5LdS[6"}, "label": "a ford pickup truck in traffic"}]}
{"id": 65769, "image": "COCO_train2014_000000065769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the persons red shirt behind the boy\"."}], "task": "refering", "answer_template": "The \"the persons red shirt behind the boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 42, 43, 44, 45, 46, 47, 48, 49, 50, 59, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 136, 137, 138, 139, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 210, 211, 212, 225, 226, 240, 241, 255, 256, 270, 271, 285, 300, 315, 330], "bbox": [0.0, 0.00225, 1.0, 0.98046875], "iscrowd": 0, "area": 68788.3822, "rle": {"size": [640, 424], "counts": "12nc04gNa0i]OCc`0k1__OYN[?g2g@]MX>^3jAgLT=T4nBPLR<j4PDZKV;[5kDiJl:Z5VEjJ`:Z5bEjJU:W5oEmJh9T5\\FPKk2mNo1U6ZKRKU2[OX2c5hKVK_1H_2T5VLXKi06h2c4cL[K3d0Q3R4PM^K]OR1Y3b3^MjK]NV1l3Q3kM3l1MYN7]1KgN9P1HTO<c0EA?5COa0H@<d0[O\\Oj0h0lNZOX1j0_NWOe1m0RNTOR2P1dMRO`2R1WMnNn2V1iLkN[3Y1[LiNi3c84L4L4K5L4L4L4L4L4K5L4L4L4L4L4K5L4L4L4L4K5L4L4L4L4L4M3O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1O100oKhFZMX9c2mF[MS9b2RG\\Mn8b2VG\\Mj8a2[G]Me8`2aG^M^8`2fG^MZ8_2kG_MU8_2oG_MQ8^2TH`Ml7]2YHaMg7]2]HaMc7\\2bHbM^7[2hHbMX7\\2lHcMS7Z2RIdMn6Y2WIeMi6Y2[IeMe6X2`IfM`6W2eIgM[6W2iIgMW6V2oIgMQ6V2TJhMl5V2XJiMg5T2^JjMb5S2cJkM]5S2gJkMY5R2lJlMT5Q2QKmMo4Q2VKlMj4Q2[KmMe4Q2_KmMa4P2dKoM[4n1jKPNV4n1RLlMn3Q2bLbM^3[2SMWMm2g2cMkL]2R3Z7M3N2M3M3N2M3M3N2M3M3N2N200O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O10000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1S2mMjM"}, "label": "the persons red shirt behind the boy"}]}
{"id": 65769, "image": "COCO_train2014_000000065769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person wearing red behind little boy\"."}], "task": "refering", "answer_template": "The \"person wearing red behind little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 42, 43, 44, 45, 46, 47, 48, 49, 50, 59, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 136, 137, 138, 139, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 210, 211, 212, 225, 226, 240, 241, 255, 256, 270, 271, 285, 300, 315, 330], "bbox": [0.0, 0.00225, 1.0, 0.98046875], "iscrowd": 0, "area": 68788.3822, "rle": {"size": [640, 424], "counts": "12nc04gNa0i]OCc`0k1__OYN[?g2g@]MX>^3jAgLT=T4nBPLR<j4PDZKV;[5kDiJl:Z5VEjJ`:Z5bEjJU:W5oEmJh9T5\\FPKk2mNo1U6ZKRKU2[OX2c5hKVK_1H_2T5VLXKi06h2c4cL[K3d0Q3R4PM^K]OR1Y3b3^MjK]NV1l3Q3kM3l1MYN7]1KgN9P1HTO<c0EA?5COa0H@<d0[O\\Oj0h0lNZOX1j0_NWOe1m0RNTOR2P1dMRO`2R1WMnNn2V1iLkN[3Y1[LiNi3c84L4L4K5L4L4L4L4L4K5L4L4L4L4L4K5L4L4L4L4K5L4L4L4L4L4M3O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1O100oKhFZMX9c2mF[MS9b2RG\\Mn8b2VG\\Mj8a2[G]Me8`2aG^M^8`2fG^MZ8_2kG_MU8_2oG_MQ8^2TH`Ml7]2YHaMg7]2]HaMc7\\2bHbM^7[2hHbMX7\\2lHcMS7Z2RIdMn6Y2WIeMi6Y2[IeMe6X2`IfM`6W2eIgM[6W2iIgMW6V2oIgMQ6V2TJhMl5V2XJiMg5T2^JjMb5S2cJkM]5S2gJkMY5R2lJlMT5Q2QKmMo4Q2VKlMj4Q2[KmMe4Q2_KmMa4P2dKoM[4n1jKPNV4n1RLlMn3Q2bLbM^3[2SMWMm2g2cMkL]2R3Z7M3N2M3M3N2M3M3N2M3M3N2N200O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O10000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1S2mMjM"}, "label": "person wearing red behind little boy"}]}
{"id": 65769, "image": "COCO_train2014_000000065769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child eating a hot dog\"."}], "task": "refering", "answer_template": "The \"a child eating a hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 36, 37, 38, 39, 40, 41, 42, 43, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 300, 301, 302, 303, 304, 305, 306, 307, 315, 316, 317, 318, 319, 320, 321, 330], "bbox": [0.02375, 0.065171875, 1.0, 0.9617968750000001], "iscrowd": 0, "area": 151971.79524999997, "rle": {"size": [640, 424], "counts": "Uk63ic03L5L4L4K5L4K5L4L4K5L4K4M4L4K5L4K5L4L4K5L4K5L3M4K5L4K5L4L4K5L4K5L4K4M4L4K5L4K5L4L4M3N2N2N1O2N2N2N2N2N2N2N2N2O1N1O2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N20O1000O10000000000O1000O10000000O10000000000`K[HdLe7[3^HdLb7Z3bHdL^7Z3fHdLZ7Z3iHeLW7Y3mHeLS7Y3QIdLP7Z3SIeLm6Y3WIeLi6Y3[IeLe6Y3^IfLb6X3bIfL]6Y3gIdLZ6Z3iIeLW6Y3mIeLS6Y3QJeLo5Y3UJeLk5Y3XJfLh5X3\\JfLd5X3`JeLa5Y3bJfL^5X3fJfLY5Y3kJeLU5Y3nJfLR5X3RKfLn4X3VKeLk4Y3XKfLh4X3\\KfLd4X3`KfL`4X3cKgL]4W3gKgLY4W3kKfLV4X3nKfLQ4Y3RLfLn3X3VLfLj3X3ZLfLf3X3]LgLc3W3aLgL_3W3eLfL\\3X3gLgLY3W3kLgLU3W3oL`LX3^3kLTLa3k3cLfKj3X4ZLYKR4f4e5N1O1O2N1O1O2N1O1O2N1O1O2N100O2N1O1O2N1O1O2N1O2N1O1O2O00001O00001O00010O00001O00001O001O00001O00001O01O100O1O100O1O100O1O010O100O1O100O1O100O1O10O01O100O1O100O1O100O1O10O0100O1O100O1O100O1O100O00100O1O100O1O100O100O1O10O01O100O1O100O1O100O1O10O01O100O1O100O100O1O100O00100O1O100O1O100O1O100O00100O100O1O100O1O101N2N101N1O2O0O2N101N1O2O0O2O0O2N101N2N101N1O2O0O2N101N1O2O0O2N101N2N101N101N1O2O0O2N101N1O2O0O2N2O0O2N101N1O2O0O2O0O2N101N1O2O1N1O2O0O2N101N1O2O0O2N101N1O2O1N101N1O2O0O2N101N1O2O0O2N101N1O2O1N1O2O0O2O0O2N101N1O2Og0XO`1`N`1aN^1aN`1`N`1aN^1aN_C"}, "label": "a child eating a hot dog"}]}
{"id": 65769, "image": "COCO_train2014_000000065769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is a young blond haired boy with a red t - shirt and red cap eating a hotdog\"."}], "task": "refering", "answer_template": "The \"it is a young blond haired boy with a red t - shirt and red cap eating a hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 36, 37, 38, 39, 40, 41, 42, 43, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 300, 301, 302, 303, 304, 305, 306, 307, 315, 316, 317, 318, 319, 320, 321, 330], "bbox": [0.02375, 0.065171875, 1.0, 0.9617968750000001], "iscrowd": 0, "area": 151971.79524999997, "rle": {"size": [640, 424], "counts": "Uk63ic03L5L4L4K5L4K5L4L4K5L4K4M4L4K5L4K5L4L4K5L4K5L3M4K5L4K5L4L4K5L4K5L4K4M4L4K5L4K5L4L4M3N2N2N1O2N2N2N2N2N2N2N2N2O1N1O2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N20O1000O10000000000O1000O10000000O10000000000`K[HdLe7[3^HdLb7Z3bHdL^7Z3fHdLZ7Z3iHeLW7Y3mHeLS7Y3QIdLP7Z3SIeLm6Y3WIeLi6Y3[IeLe6Y3^IfLb6X3bIfL]6Y3gIdLZ6Z3iIeLW6Y3mIeLS6Y3QJeLo5Y3UJeLk5Y3XJfLh5X3\\JfLd5X3`JeLa5Y3bJfL^5X3fJfLY5Y3kJeLU5Y3nJfLR5X3RKfLn4X3VKeLk4Y3XKfLh4X3\\KfLd4X3`KfL`4X3cKgL]4W3gKgLY4W3kKfLV4X3nKfLQ4Y3RLfLn3X3VLfLj3X3ZLfLf3X3]LgLc3W3aLgL_3W3eLfL\\3X3gLgLY3W3kLgLU3W3oL`LX3^3kLTLa3k3cLfKj3X4ZLYKR4f4e5N1O1O2N1O1O2N1O1O2N1O1O2N100O2N1O1O2N1O1O2N1O2N1O1O2O00001O00001O00010O00001O00001O001O00001O00001O01O100O1O100O1O100O1O010O100O1O100O1O100O1O10O01O100O1O100O1O100O1O10O0100O1O100O1O100O1O100O00100O1O100O1O100O100O1O10O01O100O1O100O1O100O1O10O01O100O1O100O100O1O100O00100O1O100O1O100O1O100O00100O100O1O100O1O101N2N101N1O2O0O2N101N1O2O0O2O0O2N101N2N101N1O2O0O2N101N1O2O0O2N101N2N101N101N1O2O0O2N101N1O2O0O2N2O0O2N101N1O2O0O2O0O2N101N1O2O1N1O2O0O2N101N1O2O0O2N101N1O2O1N101N1O2O0O2N101N1O2O0O2N101N1O2O1N1O2O0O2O0O2N101N1O2Og0XO`1`N`1aN^1aN`1`N`1aN^1aN_C"}, "label": "it is a young blond haired boy with a red t - shirt and red cap eating a hotdog"}]}
{"id": 8429, "image": "COCO_train2014_000000008429.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hotdog being held in front of a man in a black shirt\"."}], "task": "refering", "answer_template": "The \"a hotdog being held in front of a man in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221], "bbox": [0.07968, 0.7523733333333333, 0.35698, 0.9950933333333333], "iscrowd": 0, "area": 7972.209350000001, "rle": {"size": [375, 500], "counts": "`n>:W;9J6J4M3L4L4N2N2N2N2M2O2N2O1N2N2O1O1O1O00001O0010O01O001O00001O0010O01O000010O01O001O00010OBaFdN_9Q1i01O0001O01O00001O01O01O000010O02N8YFYNo8W201O0001O01O01O0001O01O01O01O0001O01O01O00010O00001O01O01O00010O001O0001O000O101O00001N101O00001N101O00001N2O1N2L3M4M3L4L4M3L3L5K5L4K5L3O3N3M3M2NWce3"}, "label": "a hotdog being held in front of a man in a black shirt"}]}
{"id": 8429, "image": "COCO_train2014_000000008429.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog in the bottom left picture\"."}], "task": "refering", "answer_template": "The \"a hot dog in the bottom left picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221], "bbox": [0.07968, 0.7523733333333333, 0.35698, 0.9950933333333333], "iscrowd": 0, "area": 7972.209350000001, "rle": {"size": [375, 500], "counts": "`n>:W;9J6J4M3L4L4N2N2N2N2M2O2N2O1N2N2O1O1O1O00001O0010O01O001O00001O0010O01O000010O01O001O00010OBaFdN_9Q1i01O0001O01O00001O01O01O000010O02N8YFYNo8W201O0001O01O01O0001O01O01O01O0001O01O01O00010O00001O01O01O00010O001O0001O000O101O00001N101O00001N101O00001N2O1N2L3M4M3L4L4M3L3L5K5L4K5L3O3N3M3M2NWce3"}, "label": "a hot dog in the bottom left picture"}]}
{"id": 164081, "image": "COCO_train2014_000000164081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tractor\"."}], "task": "refering", "answer_template": "The \"tractor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 85, 86, 87, 88, 89, 105, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136], "bbox": [0.59890625, 0.14138497652582158, 0.92696875, 0.360962441314554], "iscrowd": 0, "area": 11389.003999999995, "rle": {"size": [426, 640], "counts": "j^o41Y=8H6J000000000000001N100000000000000000000O100O1O100O1O1O1000000O1000000000000O1000000000000O1000000O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O100001O01O00000001O0000000000001O0000000000001O0000K5H8I7M3O1O100O10000000000O100000000000001O0000000000000000000O1000000000000000000000000000000000000O1000000000000000000000001O000000000O1000000009GTf1fMV\\N>A?A?B4L01O01O01O001O1O001O1O1O100O001O1O1mNUDe0l;QO`Dl0Q<N3M3M2O2M2N\\]c0"}, "label": "tractor"}]}
{"id": 164081, "image": "COCO_train2014_000000164081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue tractor\"."}], "task": "refering", "answer_template": "The \"a blue tractor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 85, 86, 87, 88, 89, 105, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136], "bbox": [0.59890625, 0.14138497652582158, 0.92696875, 0.360962441314554], "iscrowd": 0, "area": 11389.003999999995, "rle": {"size": [426, 640], "counts": "j^o41Y=8H6J000000000000001N100000000000000000000O100O1O100O1O1O1000000O1000000000000O1000000000000O1000000O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O100001O01O00000001O0000000000001O0000000000001O0000K5H8I7M3O1O100O10000000000O100000000000001O0000000000000000000O1000000000000000000000000000000000000O1000000000000000000000001O000000000O1000000009GTf1fMV\\N>A?A?B4L01O01O01O001O1O001O1O1O100O001O1O1mNUDe0l;QO`Dl0Q<N3M3M2O2M2N\\]c0"}, "label": "a blue tractor"}]}
{"id": 164081, "image": "COCO_train2014_000000164081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a dark blue shirt reaching for the frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a dark blue shirt reaching for the frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 210, 211, 233, 234], "bbox": [0.0385625, 0.40246478873239433, 0.28553125, 0.6878638497652582], "iscrowd": 0, "area": 5860.1140000000005, "rle": {"size": [426, 640], "counts": "ce:7R=1N101N2O1O0O2O1N2OO0010O010O010O010O010O101N1O100O101N10000O101O0O100O2M2M3N3O0O2O0O2O0O2O001O001O000001O000000N2M31O02O1N2O1N101N2O000O2O001O00001O001O1lERNj8o1SGSNm8Q1QFGn0ZOQ9n0SFJf0[OX9i0TF0:]Oa9b0VF6O]Ok9<XFP1h9POYFo0g9QOYFo0g9PO[Fn0f9RO[Fm0e9SO\\Fl0d9TO]Fk0c9UO^Fj0b9VO`Fh0`9WObFg0_9YObFf0^9ZOdFO\\OEo9=fFI@Ii9?nF]O^O3b9b0kG\\OT8f0mGWOR8l0nGSOR8n0oGoNQ8S1PHkNP8V1QHgNP8Z1o12O000O2O000O1010O0UOYD3g;K_D1a;McD2\\;LhD2Y;LiD4Y;GkD7X;DjD=R<O0010O0010O010O01O01O01O010O01O01O01O010O01O01O010O001O00001N101O001O001O00klm5"}, "label": "a man in a dark blue shirt reaching for the frisbee"}]}
{"id": 164081, "image": "COCO_train2014_000000164081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was playing in blue t - shirt\"."}], "task": "refering", "answer_template": "The \"a man was playing in blue t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 210, 211, 233, 234], "bbox": [0.0385625, 0.40246478873239433, 0.28553125, 0.6878638497652582], "iscrowd": 0, "area": 5860.1140000000005, "rle": {"size": [426, 640], "counts": "ce:7R=1N101N2O1O0O2O1N2OO0010O010O010O010O010O101N1O100O101N10000O101O0O100O2M2M3N3O0O2O0O2O0O2O001O001O000001O000000N2M31O02O1N2O1N101N2O000O2O001O00001O001O1lERNj8o1SGSNm8Q1QFGn0ZOQ9n0SFJf0[OX9i0TF0:]Oa9b0VF6O]Ok9<XFP1h9POYFo0g9QOYFo0g9PO[Fn0f9RO[Fm0e9SO\\Fl0d9TO]Fk0c9UO^Fj0b9VO`Fh0`9WObFg0_9YObFf0^9ZOdFO\\OEo9=fFI@Ii9?nF]O^O3b9b0kG\\OT8f0mGWOR8l0nGSOR8n0oGoNQ8S1PHkNP8V1QHgNP8Z1o12O000O2O000O1010O0UOYD3g;K_D1a;McD2\\;LhD2Y;LiD4Y;GkD7X;DjD=R<O0010O0010O010O01O01O01O010O01O01O01O010O01O01O010O001O00001N101O001O001O00klm5"}, "label": "a man was playing in blue t - shirt"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in pink hat staring at horse\"."}], "task": "refering", "answer_template": "The \"girl in pink hat staring at horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 229, 232, 233, 234, 235, 244, 245, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 320, 321, 322, 323, 324, 325, 326, 336, 337, 338, 339, 340, 341], "bbox": [0.2824121779859485, 0.63253125, 0.8046370023419205, 0.9913125], "iscrowd": 0, "area": 30304.0799, "rle": {"size": [640, 427], "counts": "^Q\\21Pd01N2O1O1N2O1O1N2O1O1N2O1O1N3N3M2M4M2N3L3N3M2M3N2N1O2M2O1O2M2O2N1N2O2N1O1O001O1O1O1O1O00100^_OSNn>m1j@lNe>T1VAVOe>k0VA]Of>c0WADe>?TAIg>=PAKl>e20O1O1O1O1O001O1O1OO010O100O100O100O100O010O100O100O100O100@`0O001O1O1O1O1O1O1O1O100O0O2O1N2O1N2O1N2O1O1N2O001O1O1O1O1O1O1N2O1O1O001O1O1O1N2O1O1O100000O100000000000000000000O100000000000000000000O101O00000000000000000O10001O001O001O001O1O0O2O001O001O001O00O1OVLjBl0V=lNVCo0m<gN^CV1g<dN^CX1f<cNdCT1`<hNjCn0Z<mNPDj0T<QOVDf0n;UO]Da0f;[OdD<_;AjD6Y;GoD1T;LTEKQ;2WEEl:9[E_Oh:?^4OO10000O100O100O1000O0O2O1N2N2O1N101N2O1N2O1N2O1N10n\\c1"}, "label": "girl in pink hat staring at horse"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid watching the horse by standing near an old lady\"."}], "task": "refering", "answer_template": "The \"a kid watching the horse by standing near an old lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 229, 232, 233, 234, 235, 244, 245, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 320, 321, 322, 323, 324, 325, 326, 336, 337, 338, 339, 340, 341], "bbox": [0.2824121779859485, 0.63253125, 0.8046370023419205, 0.9913125], "iscrowd": 0, "area": 30304.0799, "rle": {"size": [640, 427], "counts": "^Q\\21Pd01N2O1O1N2O1O1N2O1O1N2O1O1N3N3M2M4M2N3L3N3M2M3N2N1O2M2O1O2M2O2N1N2O2N1O1O001O1O1O1O1O00100^_OSNn>m1j@lNe>T1VAVOe>k0VA]Of>c0WADe>?TAIg>=PAKl>e20O1O1O1O1O001O1O1OO010O100O100O100O100O010O100O100O100O100@`0O001O1O1O1O1O1O1O1O100O0O2O1N2O1N2O1N2O1O1N2O001O1O1O1O1O1O1N2O1O1O001O1O1O1N2O1O1O100000O100000000000000000000O100000000000000000000O101O00000000000000000O10001O001O001O001O1O0O2O001O001O001O00O1OVLjBl0V=lNVCo0m<gN^CV1g<dN^CX1f<cNdCT1`<hNjCn0Z<mNPDj0T<QOVDf0n;UO]Da0f;[OdD<_;AjD6Y;GoD1T;LTEKQ;2WEEl:9[E_Oh:?^4OO10000O100O100O1000O0O2O1N2N2O1N101N2O1N2O1N2O1N10n\\c1"}, "label": "a kid watching the horse by standing near an old lady"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with black hair and a black jacket\"."}], "task": "refering", "answer_template": "The \"a person with black hair and a black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 133, 134, 148, 149, 162, 163, 164, 177, 178, 179, 191, 192, 193, 194, 205, 206, 207, 208, 209, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 266, 267, 268, 269, 281, 282, 283, 284, 297, 298, 299, 312, 313, 314, 326, 327, 328, 329, 341, 342, 343, 344], "bbox": [0.710679156908665, 0.32359375, 1.0, 1.0], "iscrowd": 0, "area": 37305.95619999999, "rle": {"size": [640, 427], "counts": "hhm5Z1Yb0?E;I7H8I7J5N2M4L3N2M4M2iDaLi5b3UJ_Lj5b3UJ_Lk5b3RJ`Lm5a3RJ`Ln5`3QJaLn5a3oIaLP6`3oIaLQ6b3kI_LT6d3hI^LX6d3eI]LZ6g3bIZL]6i3_IYLa6i3\\IXLc6k3ZIVLf6m3UIULj6n3SISLl6P4QIQLo6R4lHPLS7S4jHnKV7T4gHmKX7W4cHkK]7W4\\HnKd7T4UHQLk7Q4nGTLR8o3fGVLY8m3`GXL`8T4nFRLR9Q5fETKZ:Y7O1000000O1000000O1N1M4L4L4L4L4M3L4L4L4L4L4L4M3L4L4L4K5J6J6K5J6O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O11O001O0\\Od0nMS2mM`K"}, "label": "a person with black hair and a black jacket"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with dark hair standing to the right\"."}], "task": "refering", "answer_template": "The \"woman with dark hair standing to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 133, 134, 148, 149, 162, 163, 164, 177, 178, 179, 191, 192, 193, 194, 205, 206, 207, 208, 209, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 266, 267, 268, 269, 281, 282, 283, 284, 297, 298, 299, 312, 313, 314, 326, 327, 328, 329, 341, 342, 343, 344], "bbox": [0.710679156908665, 0.32359375, 1.0, 1.0], "iscrowd": 0, "area": 37305.95619999999, "rle": {"size": [640, 427], "counts": "hhm5Z1Yb0?E;I7H8I7J5N2M4L3N2M4M2iDaLi5b3UJ_Lj5b3UJ_Lk5b3RJ`Lm5a3RJ`Ln5`3QJaLn5a3oIaLP6`3oIaLQ6b3kI_LT6d3hI^LX6d3eI]LZ6g3bIZL]6i3_IYLa6i3\\IXLc6k3ZIVLf6m3UIULj6n3SISLl6P4QIQLo6R4lHPLS7S4jHnKV7T4gHmKX7W4cHkK]7W4\\HnKd7T4UHQLk7Q4nGTLR8o3fGVLY8m3`GXL`8T4nFRLR9Q5fETKZ:Y7O1000000O1000000O1N1M4L4L4L4L4M3L4L4L4L4L4L4M3L4L4L4K5J6J6K5J6O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O11O001O0\\Od0nMS2mM`K"}, "label": "woman with dark hair standing to the right"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in navy blue watching a horse\"."}], "task": "refering", "answer_template": "The \"a woman in navy blue watching a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 120, 135, 136, 150, 151, 165, 166, 180, 181, 195, 196, 197, 210, 211, 212, 213, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 300, 301, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.326578125, 0.3870491803278689, 0.9848125], "iscrowd": 0, "area": 38784.218499999995, "rle": {"size": [640, 427], "counts": "b?T4e7W8WOi02N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O9G:F:F;Eh1XN?A8H001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N6J8H8H8H8H8H9G8H8H8H6J7I6J7I6J6J7I6J7I^TS5"}, "label": "a woman in navy blue watching a horse"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady in dark blue to the left of the horse\"."}], "task": "refering", "answer_template": "The \"the lady in dark blue to the left of the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 120, 135, 136, 150, 151, 165, 166, 180, 181, 195, 196, 197, 210, 211, 212, 213, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 300, 301, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.326578125, 0.3870491803278689, 0.9848125], "iscrowd": 0, "area": 38784.218499999995, "rle": {"size": [640, 427], "counts": "b?T4e7W8WOi02N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O9G:F:F;Eh1XN?A8H001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N6J8H8H8H8H8H9G8H8H8H6J7I6J7I6J6J7I6J7I^TS5"}, "label": "the lady in dark blue to the left of the horse"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elderly woman looking up at the horse\"."}], "task": "refering", "answer_template": "The \"an elderly woman looking up at the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 131, 132, 133, 145, 146, 147, 148, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 189, 190, 191, 192, 204, 205, 219, 220], "bbox": [0.5914754098360656, 0.34304687500000003, 0.8804683840749415, 0.6390156250000001], "iscrowd": 0, "area": 12301.383300000001, "rle": {"size": [640, 427], "counts": "^`n42mc01O2O1N2N2N1O2N1M3M4L3QODX^O>ea0GV^O<ga0HV^O;fa0JV^O8ea0OW^O3ea0W1L3L4M3K6I6M3N1O2M2O2N2N101O1N101O0O2O0O1010O1O101N100O1O1O1O1O1O1O^OSMS@l2j?]MR@a2l?j0N2N101O^Ob0O100O2O0100O1000O100001O1O101N1O1O1O100O1O10000O1000001O0O1000O100O10O0100O100O100O10O10O1000000O1N2O1O1N2O2M4M3M3K4M4K4M2WNP_O?Ta0\\OT_OM`a0Mi^OJ]a00l^OFPVP1"}, "label": "an elderly woman looking up at the horse"}]}
{"id": 319735, "image": "COCO_train2014_000000319735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman smiling at a horse\"."}], "task": "refering", "answer_template": "The \"a woman smiling at a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 131, 132, 133, 145, 146, 147, 148, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 189, 190, 191, 192, 204, 205, 219, 220], "bbox": [0.5914754098360656, 0.34304687500000003, 0.8804683840749415, 0.6390156250000001], "iscrowd": 0, "area": 12301.383300000001, "rle": {"size": [640, 427], "counts": "^`n42mc01O2O1N2N2N1O2N1M3M4L3QODX^O>ea0GV^O<ga0HV^O;fa0JV^O8ea0OW^O3ea0W1L3L4M3K6I6M3N1O2M2O2N2N101O1N101O0O2O0O1010O1O101N100O1O1O1O1O1O1O^OSMS@l2j?]MR@a2l?j0N2N101O^Ob0O100O2O0100O1000O100001O1O101N1O1O1O100O1O10000O1000001O0O1000O100O10O0100O100O100O10O10O1000000O1N2O1O1N2O2M4M3M3K4M4K4M2WNP_O?Ta0\\OT_OM`a0Mi^OJ]a00l^OFPVP1"}, "label": "a woman smiling at a horse"}]}
{"id": 114943, "image": "COCO_train2014_000000114943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a blue t - shirt holding a black bag in the train station\"."}], "task": "refering", "answer_template": "The \"man wearing a blue t - shirt holding a black bag in the train station\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 175, 176, 177, 181, 182, 183, 198, 199, 200, 201, 204, 205, 206, 221, 222, 223, 227, 228, 245, 246, 268, 269, 291, 292, 315, 316], "bbox": [0.617546875, 0.34908878504672897, 0.9730000000000001, 0.8806308411214953], "iscrowd": 0, "area": 13510.449399999998, "rle": {"size": [428, 640], "counts": "iYU51Y=3M2]COi;3PD>d;ETDT1X;b0M3M3M1O1O1O00WOWEPOj:P1aEcNc:[1h0N1O2NUDQO];e1Jl0TO6J3L5L3M5J:hFYLY8b4K6I8I5J4UOgJ\\I]5`6kJXIY5d6mJVIW5f6j0L3O0O10001N100O2O0OoMjInLW6k2TJPMm5j2]JRMb5i2hJRMY5o2kJULfN5`6g3`KRL`4;_Il2W2bLZ4?eIP3e8mL^GW3^8gLcG^3\\8\\LfGi3Y8QLjGR4f85L4KW1jNO01O010O010O10O01O0101N2O1N2XNVHkMk7f1mHoMU7c1bIRN`6X1]J_Ne5[1gJ^NZ5JSHS1g2ROX5HWHm0g2YOT5G\\Hd0h2Ch50bJN`9N1O101N1O1OSYQ12hfnN9H7I7I6Jm0TO1N3N1N2OL5L4J6K6K47I2N3N10[O\\DAb;<bDD^;9fDFZ;7jDE[;6hDG\\;5jDD[;8o0M3M5KO0O5M3M3M3L4M2N3M2N3K4Mg0TOWNhDl1W;5O1N3N1000M3H9N2N2N3L3N3M3I7J6K5L4Khl6"}, "label": "man wearing a blue t - shirt holding a black bag in the train station"}]}
{"id": 114943, "image": "COCO_train2014_000000114943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the young man stands with his luggage outside of the commuter train waiting for his ride\"."}], "task": "refering", "answer_template": "The \"the young man stands with his luggage outside of the commuter train waiting for his ride\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 175, 176, 177, 181, 182, 183, 198, 199, 200, 201, 204, 205, 206, 221, 222, 223, 227, 228, 245, 246, 268, 269, 291, 292, 315, 316], "bbox": [0.617546875, 0.34908878504672897, 0.9730000000000001, 0.8806308411214953], "iscrowd": 0, "area": 13510.449399999998, "rle": {"size": [428, 640], "counts": "iYU51Y=3M2]COi;3PD>d;ETDT1X;b0M3M3M1O1O1O00WOWEPOj:P1aEcNc:[1h0N1O2NUDQO];e1Jl0TO6J3L5L3M5J:hFYLY8b4K6I8I5J4UOgJ\\I]5`6kJXIY5d6mJVIW5f6j0L3O0O10001N100O2O0OoMjInLW6k2TJPMm5j2]JRMb5i2hJRMY5o2kJULfN5`6g3`KRL`4;_Il2W2bLZ4?eIP3e8mL^GW3^8gLcG^3\\8\\LfGi3Y8QLjGR4f85L4KW1jNO01O010O010O10O01O0101N2O1N2XNVHkMk7f1mHoMU7c1bIRN`6X1]J_Ne5[1gJ^NZ5JSHS1g2ROX5HWHm0g2YOT5G\\Hd0h2Ch50bJN`9N1O101N1O1OSYQ12hfnN9H7I7I6Jm0TO1N3N1N2OL5L4J6K6K47I2N3N10[O\\DAb;<bDD^;9fDFZ;7jDE[;6hDG\\;5jDD[;8o0M3M5KO0O5M3M3M3L4M2N3M2N3K4Mg0TOWNhDl1W;5O1N3N1000M3H9N2N2N3L3N3M3I7J6K5L4Khl6"}, "label": "the young man stands with his luggage outside of the commuter train waiting for his ride"}]}
{"id": 303360, "image": "COCO_train2014_000000303360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in pink jacket holding a bow and arrow\"."}], "task": "refering", "answer_template": "The \"woman in pink jacket holding a bow and arrow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 48, 49, 50, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 162, 163, 164, 165], "bbox": [0.016125, 0.194070796460177, 0.35957812499999997, 0.9885840707964603], "iscrowd": 0, "area": 17630.156600000006, "rle": {"size": [226, 640], "counts": "dY21o66J6J4L4M2N3N2N2M3N2N2M101O001N10001O00001O00001O00001O0010O00000N30O00100dKeNP3Z1mLlNP3U1kLROS3m0iLYOV3g0eL@:VOj1Y1hMH8ROP2W1dMM6oNW2S1_M53lN`2n0XM=1hNi2j0PMg01bNP3]3HWLYMg3d2aLXM^3b2c0N3M3M3N2M2O2O1O1O1O1O0O2N2O1N2O1N3N1O1000000O100000000000000001O1O00100O001O1O001O1O1O001O1O001O1O001O100O0ZLQNY2R2cMoM]2T2_MmMa2W2ZMjMf2Z2TMhMl2]2nLdMR3`2hLbMX3b2cL_M]3e2^L\\Mb3R300000000001O000001O0_OeL[M[3[2PMeMo2R2]MkMc2k1lMPNT2h1VNVNj1d1]N[Nc1_1S20001O00001O0O101O000O2O00001O0O10001O00O10O10O100000O1000000000O010000O10000O10000O10O010GIdJYO[5g0`0O2N2N2N3M2K5L5O000010O002bIFQ6g0N1O2O1N2N00010N100O101O0O101OO001O10L3011N2O1N2O1N7J7G4K5LS\\j2"}, "label": "woman in pink jacket holding a bow and arrow"}]}
{"id": 303360, "image": "COCO_train2014_000000303360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman in a pink jacket practicing archery\"."}], "task": "refering", "answer_template": "The \"a young woman in a pink jacket practicing archery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 48, 49, 50, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 162, 163, 164, 165], "bbox": [0.016125, 0.194070796460177, 0.35957812499999997, 0.9885840707964603], "iscrowd": 0, "area": 17630.156600000006, "rle": {"size": [226, 640], "counts": "dY21o66J6J4L4M2N3N2N2M3N2N2M101O001N10001O00001O00001O00001O0010O00000N30O00100dKeNP3Z1mLlNP3U1kLROS3m0iLYOV3g0eL@:VOj1Y1hMH8ROP2W1dMM6oNW2S1_M53lN`2n0XM=1hNi2j0PMg01bNP3]3HWLYMg3d2aLXM^3b2c0N3M3M3N2M2O2O1O1O1O1O0O2N2O1N2O1N3N1O1000000O100000000000000001O1O00100O001O1O001O1O1O001O1O001O1O001O100O0ZLQNY2R2cMoM]2T2_MmMa2W2ZMjMf2Z2TMhMl2]2nLdMR3`2hLbMX3b2cL_M]3e2^L\\Mb3R300000000001O000001O0_OeL[M[3[2PMeMo2R2]MkMc2k1lMPNT2h1VNVNj1d1]N[Nc1_1S20001O00001O0O101O000O2O00001O0O10001O00O10O10O100000O1000000000O010000O10000O10000O10O010GIdJYO[5g0`0O2N2N2N3M2K5L5O000010O002bIFQ6g0N1O2O1N2N00010N100O101O0O101OO001O10L3011N2O1N2O1N7J7G4K5LS\\j2"}, "label": "a young woman in a pink jacket practicing archery"}]}
{"id": 303360, "image": "COCO_train2014_000000303360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman who is in the dark tank - top\"."}], "task": "refering", "answer_template": "The \"the woman who is in the dark tank - top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 59, 64, 65, 66, 82, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 133, 134, 135, 136, 137, 156, 157, 158, 179, 180, 181, 182], "bbox": [0.57865625, 0.19110619469026546, 0.975796875, 1.0], "iscrowd": 0, "area": 15549.444049999998, "rle": {"size": [226, 640], "counts": "hna24400OONX64gI9ODX64hI=[643L1CaI5a6HbI5h6L`=MeB3L4M1O11O0001N>C8H1O0000O102IlIZOX6`0mI^OU6>;L2MaIIX65gINX60hI2Y6LgI5a62N2O1N2N20001O000000000001O000001O000000000O100000001O000000000000000O2O00000000000000000000000000000001O000000000001O0000000000000000001O00000000001O00000O1000001O00000000001O0O100O100O010O1O10O0100O10O01oLRONR1LPO4T1FnN:V1ZOROf0T1QOoNo0Y1cNmN]1T1SM\\Nn0h0o1l0RMcNg0c0W2j0RMlN<<m1oNYNi1_O;V2oMZNf1@;T2SNjMK<h1E:m1PNRN9<]1E:k1aN`NU1E:j1cNaNR1F:g1gNcNn0F;f1iNcNl0G;e1lNcNi0G;e1nNcNg0H;d1POcNe0I;c1ROcNc0K:a1WOaN?N:`1ZO`N=O9`1]O_N:19_1Y1aNgN^1Z1cNeN\\1\\1dNdN[1^1dNbN[1_1eNaN[1_1eNaNZ1`1gN_NY1a1gN_NY1b1fN^NZ1b1fN^NZ1b1fN^NZ1b1gN]NY1c1gN]NY1c1gN]NY1d1fN\\NY1e1gN[NY1e1hNZNX1f1hNZNX1f1hNZNX1g1gNYNY1g1gNYNY1g1hNXNX1h1hNXNY1g1gNYNY1h1fNXN[1g1eNYN\\1f1eNYN[1g1eNYN\\1f1dNZN\\1g1cNYN^1f1bNZN_1e1bNZN^1f1bNZN_1]OjNc0G0`1[OkNe0D0b1WOmNi0A0c1SOoNm0^O0e1QOnNn0^O0e1VOhNj0C0f1POlNP1^O0n1aNlN_1UO0Q2[NnNe1QO0X30iLOW31iLOW31iLOW32hLNX32hLNX32iLLX34hLLX34hLLX35gL\\N6n0T3e0fLXN`0n0j2j0aMQO`2l0eMQO[2m0jMPOW2l0oMQOQ2m0TNPOm1m0XNPOh1n0^NnNb1j0iNSOY1k0jNROW1n0kNoNV1P1`2O0O2O0M4K4OmIAd5>[JGc58\\JKb54]JOb51]J1b5O]J1d5N\\J3d5M[J3f5MXJ4i5LVJ4k5a01O1O001O001N1O2O1N2N2O1N2N2N2O`0_OmT3"}, "label": "the woman who is in the dark tank - top"}]}
{"id": 303360, "image": "COCO_train2014_000000303360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl holding wii remotes reenacting archery\"."}], "task": "refering", "answer_template": "The \"girl holding wii remotes reenacting archery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 59, 64, 65, 66, 82, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 133, 134, 135, 136, 137, 156, 157, 158, 179, 180, 181, 182], "bbox": [0.57865625, 0.19110619469026546, 0.975796875, 1.0], "iscrowd": 0, "area": 15549.444049999998, "rle": {"size": [226, 640], "counts": "hna24400OONX64gI9ODX64hI=[643L1CaI5a6HbI5h6L`=MeB3L4M1O11O0001N>C8H1O0000O102IlIZOX6`0mI^OU6>;L2MaIIX65gINX60hI2Y6LgI5a62N2O1N2N20001O000000000001O000001O000000000O100000001O000000000000000O2O00000000000000000000000000000001O000000000001O0000000000000000001O00000000001O00000O1000001O00000000001O0O100O100O010O1O10O0100O10O01oLRONR1LPO4T1FnN:V1ZOROf0T1QOoNo0Y1cNmN]1T1SM\\Nn0h0o1l0RMcNg0c0W2j0RMlN<<m1oNYNi1_O;V2oMZNf1@;T2SNjMK<h1E:m1PNRN9<]1E:k1aN`NU1E:j1cNaNR1F:g1gNcNn0F;f1iNcNl0G;e1lNcNi0G;e1nNcNg0H;d1POcNe0I;c1ROcNc0K:a1WOaN?N:`1ZO`N=O9`1]O_N:19_1Y1aNgN^1Z1cNeN\\1\\1dNdN[1^1dNbN[1_1eNaN[1_1eNaNZ1`1gN_NY1a1gN_NY1b1fN^NZ1b1fN^NZ1b1fN^NZ1b1gN]NY1c1gN]NY1c1gN]NY1d1fN\\NY1e1gN[NY1e1hNZNX1f1hNZNX1f1hNZNX1g1gNYNY1g1gNYNY1g1hNXNX1h1hNXNY1g1gNYNY1h1fNXN[1g1eNYN\\1f1eNYN[1g1eNYN\\1f1dNZN\\1g1cNYN^1f1bNZN_1e1bNZN^1f1bNZN_1]OjNc0G0`1[OkNe0D0b1WOmNi0A0c1SOoNm0^O0e1QOnNn0^O0e1VOhNj0C0f1POlNP1^O0n1aNlN_1UO0Q2[NnNe1QO0X30iLOW31iLOW31iLOW32hLNX32hLNX32iLLX34hLLX34hLLX35gL\\N6n0T3e0fLXN`0n0j2j0aMQO`2l0eMQO[2m0jMPOW2l0oMQOQ2m0TNPOm1m0XNPOh1n0^NnNb1j0iNSOY1k0jNROW1n0kNoNV1P1`2O0O2O0M4K4OmIAd5>[JGc58\\JKb54]JOb51]J1b5O]J1d5N\\J3d5M[J3f5MXJ4i5LVJ4k5a01O1O001O001N1O2O1N2N2O1N2N2N2O`0_OmT3"}, "label": "girl holding wii remotes reenacting archery"}]}
{"id": 114945, "image": "COCO_train2014_000000114945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman tennis player that is in all pink and has white and yellow tennis shoes\"."}], "task": "refering", "answer_template": "The \"a woman tennis player that is in all pink and has white and yellow tennis shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 43, 44, 45, 60, 61, 62, 76, 77, 78, 92, 93, 94, 95, 109, 110, 111, 112, 125, 126, 127, 128, 129, 142, 143, 144, 145, 158, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 210, 211, 212, 213, 228, 229, 230, 245, 246, 247, 262, 263, 264, 279, 280, 296, 297, 313, 314, 315, 330, 331, 332, 333], "bbox": [0.34756250000000005, 0.0653125, 0.6651666666666666, 0.8675625], "iscrowd": 0, "area": 35358.56425, "rle": {"size": [640, 480], "counts": "QdX3`0Rc0b0J5L5K5J6K4K6K5K4K6K5K3L4M4K4M3M4K4M3M4K4M3L5L3M3L5L3M3L5L2M3N2N2M3N2M3M3iFgJc3[5XLnJb3U5YLoJe3T5ULRKg3R5TLRKj3Q5QLSKm3P5mKUKQ4n4jKVKT4m4fKXKX4k4cKZKZ4i4`K\\K^4f4^K^K`4e4UIQJ<_1\\6d4TIeJGk0S7c4RIZKQO8k7a4PIeMm6^2oHeMo6^2nHdMP7_2lHdMR7_2jHdMT7_2hHdMU7`2gHdMV7^2gHeMW7c7N2O1O1N3N1O1N2O1O1O2M2O1O1O1N2O2NcM`I\\G_6b8fI\\GX6d8lIZGS6d8RJZGm5d8XJZGg5d8^J[G_4mMhKf:N[GX4RNhKa:6ZGP4YNgK[:>ZGi3^NhKU:d0[Gb3cNhKP:k0[G[3hNiKj9Q1]GS3lNjKe9X1]Gl2QOkKj5]OSMQ2PNe2VOkKd5ImLk1XN_2[OkK[56iLd1`NY2_OkKU5b0cL^1hNS2ClKT5g0XL^1QOl1FmKT5m0lK^1YOf1JnKS5R1bK\\1Ab1MnKR5Y1WK[1I\\11nKR5]1mJ]1NW17mKP5b1cJ^14R1<lKo4g1ZJ_19m0a0lKn4k1PJ`1`0g0_7kMjGb1e0c0c7nMaGc1j0`0g7PNWGd1P1=k7h1SHZNm7e1RH\\Nn7e1PH\\N[5WLmK^5fN]Nl4fLWKQO6m5E]Nk4hLTKUO;g5D]Nk4jLUKSO;g5C^Nj4jLYKQO9h5B^Ni4mL\\KmN8i5B]Nh4PM^KjN7j5A^Ng4PMaKiN6j5@^Ng4RMcKfN5j5@_Ne4TMgKcN3k5_O`Nd4UMjK`N2l5^O`Nd4VMmK^NMQ6_O\\Nd4YMPLZND]6CRNf4YMSLYN[O_;`4[FULU:h3nEYLR:d3QF\\Ln9b3TF_Lk9_3XF`Li9\\3ZFeLe9Y3^FgLb9^1XHcNg7\\1[HdNe7Z1\\HgNd7W1]HiNd7T1_HlNa7R1`HoN`7o0aHQO`7m0bHPOa7n0`HQOb7m0_HQOd7m0^HQOd7m0]HQOe7n0\\HQOf7m0\\HPOg7n0ZHQOh7m0YHQOj7m0WHROk7l0WHQOR8f0PHYO\\>N3N1N2O\\[T3"}, "label": "a woman tennis player that is in all pink and has white and yellow tennis shoes"}]}
{"id": 114945, "image": "COCO_train2014_000000114945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female tennis player n pink clothing\"."}], "task": "refering", "answer_template": "The \"a female tennis player n pink clothing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 43, 44, 45, 60, 61, 62, 76, 77, 78, 92, 93, 94, 95, 109, 110, 111, 112, 125, 126, 127, 128, 129, 142, 143, 144, 145, 158, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 210, 211, 212, 213, 228, 229, 230, 245, 246, 247, 262, 263, 264, 279, 280, 296, 297, 313, 314, 315, 330, 331, 332, 333], "bbox": [0.34756250000000005, 0.0653125, 0.6651666666666666, 0.8675625], "iscrowd": 0, "area": 35358.56425, "rle": {"size": [640, 480], "counts": "QdX3`0Rc0b0J5L5K5J6K4K6K5K4K6K5K3L4M4K4M3M4K4M3M4K4M3L5L3M3L5L3M3L5L2M3N2N2M3N2M3M3iFgJc3[5XLnJb3U5YLoJe3T5ULRKg3R5TLRKj3Q5QLSKm3P5mKUKQ4n4jKVKT4m4fKXKX4k4cKZKZ4i4`K\\K^4f4^K^K`4e4UIQJ<_1\\6d4TIeJGk0S7c4RIZKQO8k7a4PIeMm6^2oHeMo6^2nHdMP7_2lHdMR7_2jHdMT7_2hHdMU7`2gHdMV7^2gHeMW7c7N2O1O1N3N1O1N2O1O1O2M2O1O1O1N2O2NcM`I\\G_6b8fI\\GX6d8lIZGS6d8RJZGm5d8XJZGg5d8^J[G_4mMhKf:N[GX4RNhKa:6ZGP4YNgK[:>ZGi3^NhKU:d0[Gb3cNhKP:k0[G[3hNiKj9Q1]GS3lNjKe9X1]Gl2QOkKj5]OSMQ2PNe2VOkKd5ImLk1XN_2[OkK[56iLd1`NY2_OkKU5b0cL^1hNS2ClKT5g0XL^1QOl1FmKT5m0lK^1YOf1JnKS5R1bK\\1Ab1MnKR5Y1WK[1I\\11nKR5]1mJ]1NW17mKP5b1cJ^14R1<lKo4g1ZJ_19m0a0lKn4k1PJ`1`0g0_7kMjGb1e0c0c7nMaGc1j0`0g7PNWGd1P1=k7h1SHZNm7e1RH\\Nn7e1PH\\N[5WLmK^5fN]Nl4fLWKQO6m5E]Nk4hLTKUO;g5D]Nk4jLUKSO;g5C^Nj4jLYKQO9h5B^Ni4mL\\KmN8i5B]Nh4PM^KjN7j5A^Ng4PMaKiN6j5@^Ng4RMcKfN5j5@_Ne4TMgKcN3k5_O`Nd4UMjK`N2l5^O`Nd4VMmK^NMQ6_O\\Nd4YMPLZND]6CRNf4YMSLYN[O_;`4[FULU:h3nEYLR:d3QF\\Ln9b3TF_Lk9_3XF`Li9\\3ZFeLe9Y3^FgLb9^1XHcNg7\\1[HdNe7Z1\\HgNd7W1]HiNd7T1_HlNa7R1`HoN`7o0aHQO`7m0bHPOa7n0`HQOb7m0_HQOd7m0^HQOd7m0]HQOe7n0\\HQOf7m0\\HPOg7n0ZHQOh7m0YHQOj7m0WHROk7l0WHQOR8f0PHYO\\>N3N1N2O\\[T3"}, "label": "a female tennis player n pink clothing"}]}
{"id": 352511, "image": "COCO_train2014_000000352511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with an ornate floral design on the comforter\"."}], "task": "refering", "answer_template": "The \"a bed with an ornate floral design on the comforter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.006583333333333333, 0.48395312500000004, 0.9942291666666667, 1.0], "iscrowd": 0, "area": 130007.03235000002, "rle": {"size": [640, 480], "counts": "bY24bc0;D=C<E;D<E<D;I7O1N2N3M2O1N2N3N1N2N2O2M2N2O1N3M2O1N2N6J8I8G8H9H7H9G8Ij0UO]1cN\\1eN0O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100000O010000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O1000000000O10000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10O10000000000000000000000000000000000000000O10000000000000000000000000000000000000000O100000000000000000000000O100000000000000000O1000000000000000000000000000000001O0000001O00001O00001O0000001O00001O0000001O00001O00001O000O101O00001O0000001O00001O00001O0000001O00001O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N001O1O1O1N2O1O001O1O1O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O0O2O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1N101O1O001O1O001O1O001O1O001O00S3mLeZ1"}, "label": "a bed with an ornate floral design on the comforter"}]}
{"id": 352511, "image": "COCO_train2014_000000352511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left big bed in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the left big bed in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.006583333333333333, 0.48395312500000004, 0.9942291666666667, 1.0], "iscrowd": 0, "area": 130007.03235000002, "rle": {"size": [640, 480], "counts": "bY24bc0;D=C<E;D<E<D;I7O1N2N3M2O1N2N3N1N2N2O2M2N2O1N3M2O1N2N6J8I8G8H9H7H9G8Ij0UO]1cN\\1eN0O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100000O010000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O1000000000O10000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10O10000000000000000000000000000000000000000O10000000000000000000000000000000000000000O100000000000000000000000O100000000000000000O1000000000000000000000000000000001O0000001O00001O00001O0000001O00001O0000001O00001O00001O000O101O00001O0000001O00001O00001O0000001O00001O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N001O1O1O1N2O1O001O1O1O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O0O2O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1N101O1O001O1O001O1O001O1O001O00S3mLeZ1"}, "label": "the left big bed in the right hand picture"}]}
{"id": 155904, "image": "COCO_train2014_000000155904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white bowl full of strawberries\"."}], "task": "refering", "answer_template": "The \"a small white bowl full of strawberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 18, 19, 20, 21, 22, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67, 79, 80, 81], "bbox": [0.2020843091334895, 0.02471875, 0.518688524590164, 0.240453125], "iscrowd": 0, "area": 14711.811599999997, "rle": {"size": [640, 427], "counts": "Wje18gc0g0XO9G3M3M3M3N1N3M3M3M3N1N3M3M3M3N3L4L4L5K4L4M3L4M2N2O1N101N2N101N2O0O2O1N1O2O0O2O1N1O2O1O000000001O0000000000001O000000000000001O00000000O1000000000001O00000O101O000000001O000O10001O000000001O000O10001O1O1O001O1O1N2M3N2N1N3N2N2M3N2N2N1N3N3M3L4K5K5K5L4K5K5K7I6I7H8G9H9GjeP4"}, "label": "a small white bowl full of strawberries"}]}
{"id": 155904, "image": "COCO_train2014_000000155904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl of fresh strawberries at the top of the screen\"."}], "task": "refering", "answer_template": "The \"the bowl of fresh strawberries at the top of the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 18, 19, 20, 21, 22, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67, 79, 80, 81], "bbox": [0.2020843091334895, 0.02471875, 0.518688524590164, 0.240453125], "iscrowd": 0, "area": 14711.811599999997, "rle": {"size": [640, 427], "counts": "Wje18gc0g0XO9G3M3M3M3N1N3M3M3M3N1N3M3M3M3N3L4L4L5K4L4M3L4M2N2O1N101N2N101N2O0O2O1N1O2O0O2O1N1O2O1O000000001O0000000000001O000000000000001O00000000O1000000000001O00000O101O000000001O000O10001O000000001O000O10001O1O1O001O1O1N2M3N2N1N3N2N2M3N2N2N1N3N3M3L4K5K5K5L4K5K5K7I6I7H8G9H9GjeP4"}, "label": "the bowl of fresh strawberries at the top of the screen"}]}
{"id": 155904, "image": "COCO_train2014_000000155904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of healthy food with a spoon and napkin next to it\"."}], "task": "refering", "answer_template": "The \"a bowl of healthy food with a spoon and napkin next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 272, 273, 274, 275, 276, 277, 278, 289, 290, 291], "bbox": [0.11426229508196721, 0.46434375, 0.6855737704918033, 0.8387968750000001], "iscrowd": 0, "area": 46567.99935, "rle": {"size": [640, 427], "counts": "SPo0a0]c0g0XOh0XO7J5J5L5J6J5L5J6K4K5L4K5M3M2M4M3L4M3M2M4M3M3L4M3M2M3N2N2N2O1N2N2O1N2N2O0O2N101N2O0O2N101N1O2O0O2N2O0O2O0O2N101N1O2O1N101O0O2O001N101O1N101O001N10000O10001N1000000O101O0O10000O101O000O10000O2O000000000O2O00000000001O0000000000001O0000O10000000000O1000000000000O100000O1000O10000000000O10001O00000O2O00001O000O2O00001O000O101O00001O000O2O000O101N1O101N100O2O1N2N2O1N101N2O1N2N2O0O2O1N2O1N2O0O2N2O1N2O1N101M3N3M2M3N2N2M4M2N2M3N2N3L3N2N2M3N3M2M3N2N2M4M2N2M3N2L5K4L4L7I7J6I8H7I7H8C=C>B=Cd[c2"}, "label": "a bowl of healthy food with a spoon and napkin next to it"}]}
{"id": 155904, "image": "COCO_train2014_000000155904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of bananas , strawberries , and some other things\"."}], "task": "refering", "answer_template": "The \"a bowl of bananas , strawberries , and some other things\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 272, 273, 274, 275, 276, 277, 278, 289, 290, 291], "bbox": [0.11426229508196721, 0.46434375, 0.6855737704918033, 0.8387968750000001], "iscrowd": 0, "area": 46567.99935, "rle": {"size": [640, 427], "counts": "SPo0a0]c0g0XOh0XO7J5J5L5J6J5L5J6K4K5L4K5M3M2M4M3L4M3M2M4M3M3L4M3M2M3N2N2N2O1N2N2O1N2N2O0O2N101N2O0O2N101N1O2O0O2N2O0O2O0O2N101N1O2O1N101O0O2O001N101O1N101O001N10000O10001N1000000O101O0O10000O101O000O10000O2O000000000O2O00000000001O0000000000001O0000O10000000000O1000000000000O100000O1000O10000000000O10001O00000O2O00001O000O2O00001O000O101O00001O000O2O000O101N1O101N100O2O1N2N2O1N101N2O1N2N2O0O2O1N2O1N2O0O2N2O1N2O1N101M3N3M2M3N2N2M4M2N2M3N2N3L3N2N2M3N3M2M3N2N2M4M2N2M3N2L5K4L4L7I7J6I8H7I7H8C=C>B=Cd[c2"}, "label": "a bowl of bananas , strawberries , and some other things"}]}
{"id": 549127, "image": "COCO_train2014_000000549127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a colorful umbrella shading seafood\"."}], "task": "refering", "answer_template": "The \"a colorful umbrella shading seafood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147], "bbox": [0.0, 0.0, 0.517, 0.4349344978165939], "iscrowd": 0, "area": 57047.279650000004, "rle": {"size": [458, 640], "counts": "b0P2T2SNn7V4O1N1O2O001O1O001O000O2O00001O001N10001O0000001N100000001O0O1000001O1O0O3N001O000O2O001O0O2O001N101O000O2O001O1N101O001N101O000O2O00000O0100000OO2O1N2O0O2O1001O00<D3M1O1O2N00001O0000001O00001O00000000000000001O000000000000001O000000000000000000010O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O0O1N2O1N2O1N2O100O1O2N1O1O2N1O1O1O2N1O11O4M3L3M2OO01O00100N1000O101O0O2O0O101N1O101N101N1O101N100O2N101N100O2N101N101N1O2O1N101N1O101N100O2O0O101N100O2O0O100O2O0O101N100O2O0O101N100O2O0O10001N10001N100O2O000O2O1]NfESO]:h0jESOX:]ObEb0:0U:\\OhE;76T:]OiE686V:AgE199U:DfEL9=V:DmF3\\9JfF1a9LbWZ4"}, "label": "a colorful umbrella shading seafood"}]}
{"id": 549127, "image": "COCO_train2014_000000549127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella covering the food items\"."}], "task": "refering", "answer_template": "The \"umbrella covering the food items\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147], "bbox": [0.0, 0.0, 0.517, 0.4349344978165939], "iscrowd": 0, "area": 57047.279650000004, "rle": {"size": [458, 640], "counts": "b0P2T2SNn7V4O1N1O2O001O1O001O000O2O00001O001N10001O0000001N100000001O0O1000001O1O0O3N001O000O2O001O0O2O001N101O000O2O001O1N101O001N101O000O2O00000O0100000OO2O1N2O0O2O1001O00<D3M1O1O2N00001O0000001O00001O00000000000000001O000000000000001O000000000000000000010O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O0O1N2O1N2O1N2O100O1O2N1O1O2N1O1O1O2N1O11O4M3L3M2OO01O00100N1000O101O0O2O0O101N1O101N101N1O101N100O2N101N100O2N101N101N1O2O1N101N1O101N100O2O0O101N100O2O0O100O2O0O101N100O2O0O101N100O2O0O10001N10001N100O2O000O2O1]NfESO]:h0jESOX:]ObEb0:0U:\\OhE;76T:]OiE686V:AgE199U:DfEL9=V:DmF3\\9JfF1a9LbWZ4"}, "label": "umbrella covering the food items"}]}
{"id": 549127, "image": "COCO_train2014_000000549127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella over lady ' s hand\"."}], "task": "refering", "answer_template": "The \"umbrella over lady ' s hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228], "bbox": [0.5834374999999999, 0.0, 1.0, 0.59353711790393], "iscrowd": 0, "area": 58631.975399999996, "rle": {"size": [458, 640], "counts": "kmV54`=h0E;F:F;M2M3M3M3M3N1N2N1O2O1N2N2N2N1O2N3M3M3L3N3M3M3M3M3M3M3L2O2N2N2N2N1O2M3N3M3M4L4L4L3M4K5L4L4L3M4L4L4K4M2N1O2N2N101O0O2O001O1O0O2O001O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O1O001O001O1O001OO100O10000O10000O100O101O0O100O10000O100O1001O1O1O001O1O1O1O1O1O1O1O1O001O1O1OO100O100O100O100O10000O100O100O100O10000O100O10000O100O10000O10000O100O10000O100O100O10000O100O100O100O10000O100O100O100000000O1000000000000O1N1"}, "label": "umbrella over lady ' s hand"}]}
{"id": 549127, "image": "COCO_train2014_000000549127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"multi colored umbrella on the right side\"."}], "task": "refering", "answer_template": "The \"multi colored umbrella on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228], "bbox": [0.5834374999999999, 0.0, 1.0, 0.59353711790393], "iscrowd": 0, "area": 58631.975399999996, "rle": {"size": [458, 640], "counts": "kmV54`=h0E;F:F;M2M3M3M3M3N1N2N1O2O1N2N2N2N1O2N3M3M3L3N3M3M3M3M3M3M3L2O2N2N2N2N1O2M3N3M3M4L4L4L3M4K5L4L4L3M4L4L4K4M2N1O2N2N101O0O2O001O1O0O2O001O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O1O001O001O1O001OO100O10000O10000O100O101O0O100O10000O100O1001O1O1O001O1O1O1O1O1O1O1O1O001O1O1OO100O100O100O100O10000O100O100O100O10000O100O10000O100O10000O10000O100O10000O100O100O10000O100O100O100O10000O100O100O100000000O1000000000000O1N1"}, "label": "multi colored umbrella on the right side"}]}
{"id": 24842, "image": "COCO_train2014_000000024842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing floral dress preparing food in kitchen\"."}], "task": "refering", "answer_template": "The \"a woman wearing floral dress preparing food in kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 72, 73, 74, 75, 76, 79, 80], "bbox": [0.161078431372549, 0.0, 0.7897058823529411, 0.16800653594771242], "iscrowd": 0, "area": 27642.837700000004, "rle": {"size": [612, 612], "counts": "WWk13nb03M3M3M3M3M4L3M3M3M3M3M3M3M3M4L3M3M3M3N200000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000001O001O1O1O1O002N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O0000000000O101O000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O12N1O2N2N1O2N2N1O2N2N2N101N2N1O2N1O00000000000000000000000003M5\\Nj_O0Z`0Li_O1\\`0Jf_O4^`0He_O5_`0Hb_O6``0Hb_O6``0Hc_O5_`0Ic_O5^`0Je_O3Z`0Nh_O0W`01l_OLS`05o_OIP`08S@El?<V@Bj?>X@@g?a0]11O1O2N1O001O1O1O1O100O1O1O1O001O1O1O1O1O1O100O1O00100000000000000000O10000000000000001O1N2O1O1O1O1O2N1O1O1O1N2O1O1O2M2O1N2O1N2O1N4M3L4M3L4MiQ]2"}, "label": "a woman wearing floral dress preparing food in kitchen"}]}
{"id": 24842, "image": "COCO_train2014_000000024842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"breast area of a woman in floral design shirt\"."}], "task": "refering", "answer_template": "The \"breast area of a woman in floral design shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 72, 73, 74, 75, 76, 79, 80], "bbox": [0.161078431372549, 0.0, 0.7897058823529411, 0.16800653594771242], "iscrowd": 0, "area": 27642.837700000004, "rle": {"size": [612, 612], "counts": "WWk13nb03M3M3M3M3M4L3M3M3M3M3M3M3M3M4L3M3M3M3N200000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000001O001O1O1O1O002N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O0000000000O101O000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O12N1O2N2N1O2N2N1O2N2N2N101N2N1O2N1O00000000000000000000000003M5\\Nj_O0Z`0Li_O1\\`0Jf_O4^`0He_O5_`0Hb_O6``0Hb_O6``0Hc_O5_`0Ic_O5^`0Je_O3Z`0Nh_O0W`01l_OLS`05o_OIP`08S@El?<V@Bj?>X@@g?a0]11O1O2N1O001O1O1O1O100O1O1O1O001O1O1O1O1O1O100O1O00100000000000000000O10000000000000001O1N2O1O1O1O1O2N1O1O1O1N2O1O1O2M2O1N2O1N2O1N4M3L4M3L4MiQ]2"}, "label": "breast area of a woman in floral design shirt"}]}
{"id": 237834, "image": "COCO_train2014_000000237834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mcgriddle with cheese next to a hashbrown\"."}], "task": "refering", "answer_template": "The \"a mcgriddle with cheese next to a hashbrown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 158, 159, 160, 161, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.06727124183006536, 0.30627450980392157, 0.6502287581699346, 0.7188398692810458], "iscrowd": 0, "area": 64461.81610000002, "rle": {"size": [612, 612], "counts": "Vlh04nb0`0A3L2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2M3C=ZOf0O1O100O1O1O1O100O1O1O1O100O1O1JeJbB\\5]=eJcB[5\\=gJbBZ5]=7O1O1N2O100O100O10001N2O0O2O1O1N2O0O2O1O1N101N2O1O1N101N2O1O1N101N2O1O0O2O1N2O1O0O100O10000O100O10000O100O10000O100O1000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000000O10O2O00000000000O100000001O00000O1000000000000O2O0001O001O001O1O001O1O010O001O1O001O1O001O00100O00O100O2O000O101N10000O2O000O101N10000O101N10000O2O0O10001N100O101O0O100O101O0O100O2O000O1O2N1N3N1N3N1O2M2O2M2O2M3N1O2M2O2M2O2M2O2N1IbBfJa=Z550O101N100O100O101N100O10001N01N1YMbBG`=7aBH`=6cBH_=6bBI_=5cBJ_=4cBK]=3eBL]=2dBM^=1dBM]=1eBN]=OeB0]=MeB2^=KdB3_=IcB6`=FbB:_=CdB;_=AcB>`=_OaB`0a=]ObBa0a=[OaBd0b=XOaBf0b=VO`Bi0b=TOaBj0b=SO`Bj0d=RO^Bm0e=oN^Bn0e=oN]BP1f=lN]BR1f=kN[BS1i=iNZBU1h=hNZBV1l?N1N3N1N3N1N3N4K7J5J7Jjao3"}, "label": "a mcgriddle with cheese next to a hashbrown"}]}
{"id": 237834, "image": "COCO_train2014_000000237834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"burger\"."}], "task": "refering", "answer_template": "The \"burger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 158, 159, 160, 161, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.06727124183006536, 0.30627450980392157, 0.6502287581699346, 0.7188398692810458], "iscrowd": 0, "area": 64461.81610000002, "rle": {"size": [612, 612], "counts": "Vlh04nb0`0A3L2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2M3C=ZOf0O1O100O1O1O1O100O1O1O1O100O1O1JeJbB\\5]=eJcB[5\\=gJbBZ5]=7O1O1N2O100O100O10001N2O0O2O1O1N2O0O2O1O1N101N2O1O1N101N2O1O1N101N2O1O0O2O1N2O1O0O100O10000O100O10000O100O10000O100O1000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000000O10O2O00000000000O100000001O00000O1000000000000O2O0001O001O001O1O001O1O010O001O1O001O1O001O00100O00O100O2O000O101N10000O2O000O101N10000O101N10000O2O0O10001N100O101O0O100O101O0O100O2O000O1O2N1N3N1N3N1O2M2O2M2O2M3N1O2M2O2M2O2M2O2N1IbBfJa=Z550O101N100O100O101N100O10001N01N1YMbBG`=7aBH`=6cBH_=6bBI_=5cBJ_=4cBK]=3eBL]=2dBM^=1dBM]=1eBN]=OeB0]=MeB2^=KdB3_=IcB6`=FbB:_=CdB;_=AcB>`=_OaB`0a=]ObBa0a=[OaBd0b=XOaBf0b=VO`Bi0b=TOaBj0b=SO`Bj0d=RO^Bm0e=oN^Bn0e=oN]BP1f=lN]BR1f=kN[BS1i=iNZBU1h=hNZBV1l?N1N3N1N3N1N3N4K7J5J7Jjao3"}, "label": "burger"}]}
{"id": 237834, "image": "COCO_train2014_000000237834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hash - brown next to the sandwich\"."}], "task": "refering", "answer_template": "The \"the hash - brown next to the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350], "bbox": [0.6502287581699346, 0.4408006535947712, 0.9529084967320263, 0.7121078431372548], "iscrowd": 0, "area": 19440.6313, "rle": {"size": [612, 612], "counts": "fV^71Rc02N1O2N1O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2N2N1O2N1O2N2O1N2O0O2N2O1N2O1N101N2O1N2N2O1N2O3L3N3L3N3L4c_OmMb?k200O100O2O0O100O1O101N100O100O101N1O100O100O2O0O100O101N100O10000O2O0O10000O101O0O100O10001N10000O100O2O000O10001N100O10000O2O000O100O101O0O100O10001N10000O101N10000O10001N100O1000001O000O1000001O0000000O101OO1O1O1N2O1N101O1N2O1O1N2M2N3N2aLZA]2i>`M]AZ2f>cM`AW2c>fMdAT2^>jMgAP2\\>mMjAm1X>QNoAh1T>UNRBf1P>WNVBc1m=ZNYB`1j=]N]B\\1j?K5J7I6J6K6I6Jao`0"}, "label": "the hash - brown next to the sandwich"}]}
{"id": 237834, "image": "COCO_train2014_000000237834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hash brown with a bite taken out of it\"."}], "task": "refering", "answer_template": "The \"the hash brown with a bite taken out of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350], "bbox": [0.6502287581699346, 0.4408006535947712, 0.9529084967320263, 0.7121078431372548], "iscrowd": 0, "area": 19440.6313, "rle": {"size": [612, 612], "counts": "fV^71Rc02N1O2N1O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2N2N1O2N1O2N2O1N2O0O2N2O1N2O1N101N2O1N2N2O1N2O3L3N3L3N3L4c_OmMb?k200O100O2O0O100O1O101N100O100O101N1O100O100O2O0O100O101N100O10000O2O0O10000O101O0O100O10001N10000O100O2O000O10001N100O10000O2O000O100O101O0O100O10001N10000O101N10000O10001N100O1000001O000O1000001O0000000O101OO1O1O1N2O1N101O1N2O1O1N2M2N3N2aLZA]2i>`M]AZ2f>cM`AW2c>fMdAT2^>jMgAP2\\>mMjAm1X>QNoAh1T>UNRBf1P>WNVBc1m=ZNYB`1j=]N]B\\1j?K5J7I6J6K6I6Jao`0"}, "label": "the hash brown with a bite taken out of it"}]}
{"id": 237834, "image": "COCO_train2014_000000237834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the food on the wrapper\"."}], "task": "refering", "answer_template": "The \"the food on the wrapper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.09299019607843137, 1.0, 1.0], "iscrowd": 0, "area": 277080.60705, "rle": {"size": [612, 612], "counts": "`4d>_41O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O100O10000O100O10000O100O10000O100O10000O100O10000O10000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000_BiMU:W2jEkMU:T2kEmMU:S2jEoMU:Q2jEPNV:P2jEQNU:o1jERNV:m1jEUNU:k1jEWNU:i1jEXNV:h1iEZNV:e1kE[NU:e1jE]NU:c1jE_NU:a1jE`NV:`1iEbNV:]1jEdNV:\\1jEeNU:[1jEfNV:Z1iEhNV:W1jEkNU:U1jElNV:T1iEnNV:R1jEnNV:R1iEPOV:o0jEROV:n0iETOV:l0iEVOV:j0iEWOW:h0iEZOV:f0jEZOV:f0iE\\OV:d0iE^OV:b0iE_OW:`0hECW:=gEEY:;fEGY:9eEI[:7cEL\\:3dEO[:1cE1]:OaE4^:L`E6`:I`E9_:G_E;a:E]E>b:B]E`0b:@\\Eb0d:]O[Ef0d:ZO[Eg0e:YOYEj0f:VOXEm0g:ROYEo0g:QOWER1h:nNVET1j:lNUEV1j:jNTEX1l:gNSE\\1l:dNRE_1m:aNRE`1n:`NPEc1o:\\NPEf1P;ZNoDh1P;XNnDj1R;VNlDm1S;SNlDo1S;PNlDR2T;nMjDU2U;kMjDV2V;jMhDY2W;fMhD]2W;cMgD_2Y;aMfDa2Y;eM_D]2a;iMWDZ2h;mMPDT2P<a2O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O0000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O100000000O10000000000O10000000000O1000000000000O10000000000O10000000000O11O00n0ROQ3oL"}, "label": "the food on the wrapper"}]}
{"id": 303370, "image": "COCO_train2014_000000303370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple suitcase , being pulled by a woman with blonde hair\"."}], "task": "refering", "answer_template": "The \"a purple suitcase , being pulled by a woman with blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.261625, 0.6908430913348946, 0.465515625, 0.9746135831381734], "iscrowd": 0, "area": 8976.0848, "rle": {"size": [427, 640], "counts": "moU21Y=4L4L3SCIa<a0O2M3N2N1O2N2N2N1N3N2N2N1O2N2N2N2M2dDZNR;h1gDdNR;l1N001O1O0O2O001O1O001O1O0O3N1O1O2N1O1O2M2O1O1O2N1O1O2N1N2O2N1O1O1O000O1000000000000001O1O2M2O1O1O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O1O2N1O1O1N2O2N1O1O1O1O1O2N1O1O2N2N2M4M2N2N2N3M2N2N2N3M2N2N3M2M3N2N3M2N2N2N3M2N2N2NUY^4"}, "label": "a purple suitcase , being pulled by a woman with blonde hair"}]}
{"id": 303370, "image": "COCO_train2014_000000303370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the purple suitcase\"."}], "task": "refering", "answer_template": "The \"the purple suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.261625, 0.6908430913348946, 0.465515625, 0.9746135831381734], "iscrowd": 0, "area": 8976.0848, "rle": {"size": [427, 640], "counts": "moU21Y=4L4L3SCIa<a0O2M3N2N1O2N2N2N1N3N2N2N1O2N2N2N2M2dDZNR;h1gDdNR;l1N001O1O0O2O001O1O001O1O0O3N1O1O2N1O1O2M2O1O1O2N1O1O2N1N2O2N1O1O1O000O1000000000000001O1O2M2O1O1O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O1O2N1O1O1N2O2N1O1O1O1O1O2N1O1O2N2N2M4M2N2N2N3M2N2N2N3M2N2N3M2M3N2N3M2N2N2N3M2N2N2NUY^4"}, "label": "the purple suitcase"}]}
{"id": 24847, "image": "COCO_train2014_000000024847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is traveling to the left of the frame and looking to the right of the frame\"."}], "task": "refering", "answer_template": "The \"a giraffe that is traveling to the left of the frame and looking to the right of the frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 116, 117, 118, 140, 141, 163, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329], "bbox": [0.0455, 0.22977083333333334, 0.3505625, 0.8814583333333333], "iscrowd": 0, "area": 23067.3072, "rle": {"size": [480, 640], "counts": "SV>2m>2O0O2O1N2O1N101O1O1O10O0100O1O010O001lADc=;XBLf=5TB1l=c0N1O2O0J7G8O2M2O2N1O2M2O2N1N3N1O2M2N3L3N3L3N3O001N2O001N1011N4M3L3N0O100O010cNeC2\\<LeC5[<IeC8[<GfC:Z<ChC=Y<AgCa0n1iNo7d0TFc0b1YOU83ZFe0X1EZ8D^Fh0T1MZ8WOeFm0k06\\8hNmFR1c0`0Z8YNWGY17j0^8gM`Gl4X95J6K5Kl0SO=D2N1N3N00M2N3M3M2N3M3M3M3L5L3M3L4L5J5K5L3L4L4M3L4L4M3L4101N2N3M2O1O2N100O2N1O1O1O2N1O101N4L7I5KN2N2M4M2N2N2N2M3N2L4J6K4K6J6K4L51O2M2O2M2O2N1N3N1N2O1O0O100O10000O10001N100O10000O100O101OM2N3L3N3M3M2M4M3M3M4K5J7I6J6J7I6J6J7I6J6J7I6I7JefR6"}, "label": "a giraffe that is traveling to the left of the frame and looking to the right of the frame"}]}
{"id": 24847, "image": "COCO_train2014_000000024847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe leading two other giraffes towards the camera\"."}], "task": "refering", "answer_template": "The \"a giraffe leading two other giraffes towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 116, 117, 118, 140, 141, 163, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329], "bbox": [0.0455, 0.22977083333333334, 0.3505625, 0.8814583333333333], "iscrowd": 0, "area": 23067.3072, "rle": {"size": [480, 640], "counts": "SV>2m>2O0O2O1N2O1N101O1O1O10O0100O1O010O001lADc=;XBLf=5TB1l=c0N1O2O0J7G8O2M2O2N1O2M2O2N1N3N1O2M2N3L3N3L3N3O001N2O001N1011N4M3L3N0O100O010cNeC2\\<LeC5[<IeC8[<GfC:Z<ChC=Y<AgCa0n1iNo7d0TFc0b1YOU83ZFe0X1EZ8D^Fh0T1MZ8WOeFm0k06\\8hNmFR1c0`0Z8YNWGY17j0^8gM`Gl4X95J6K5Kl0SO=D2N1N3N00M2N3M3M2N3M3M3M3L5L3M3L4L5J5K5L3L4L4M3L4L4M3L4101N2N3M2O1O2N100O2N1O1O1O2N1O101N4L7I5KN2N2M4M2N2N2N2M3N2L4J6K4K6J6K4L51O2M2O2M2O2N1N3N1N2O1O0O100O10000O10001N100O10000O100O101OM2N3L3N3M3M2M4M3M3M4K5J7I6J6J7I6J6J7I6J6J7I6I7JefR6"}, "label": "a giraffe leading two other giraffes towards the camera"}]}
{"id": 24847, "image": "COCO_train2014_000000024847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle giraffe in a group of three that is looking away from the camera\"."}], "task": "refering", "answer_template": "The \"the middle giraffe in a group of three that is looking away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 124, 125, 126, 147, 148, 149, 150, 151, 152, 172, 173, 174, 175, 176, 177, 197, 198, 199, 200, 201, 221], "bbox": [0.391015625, 0.23820833333333333, 0.7533749999999999, 0.5348333333333334], "iscrowd": 0, "area": 9388.811750000003, "rle": {"size": [480, 640], "counts": "khe32o>0O1O2O0O1O2N100O2N1XBHg<8RC0l<2lB6Q=LhB<Q=m0J7N2NO003M2N3M2N1N10O10000O11O1O1O1O1N2O5K6K6K2OO000O100000000O101O00000O10000000001N100000000O100000000O100000000O100000001O0O100O100O100O100O100O10000O100O100O10O010O0010O010O010O010O01O10O101N100O2O0O100O2O0O1O2N100O0010O01O00010O001O01O01O00010O001O01O01O0010O000O101O00001O0O2O001O001O001N2O001O001O0O2O001O001N101O1O001O0O2O001O001N101K4J7J6L3010O01O0010O01O0010O01O1O010O0010O10O010O0100O0100O0100O010O01O0O2O001O1O1O1N2N3N1NajY2"}, "label": "the middle giraffe in a group of three that is looking away from the camera"}]}
{"id": 24847, "image": "COCO_train2014_000000024847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is in between 2 other giraffes standing in treesh and bushes\"."}], "task": "refering", "answer_template": "The \"a giraffe that is in between 2 other giraffes standing in treesh and bushes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 124, 125, 126, 147, 148, 149, 150, 151, 152, 172, 173, 174, 175, 176, 177, 197, 198, 199, 200, 201, 221], "bbox": [0.391015625, 0.23820833333333333, 0.7533749999999999, 0.5348333333333334], "iscrowd": 0, "area": 9388.811750000003, "rle": {"size": [480, 640], "counts": "khe32o>0O1O2O0O1O2N100O2N1XBHg<8RC0l<2lB6Q=LhB<Q=m0J7N2NO003M2N3M2N1N10O10000O11O1O1O1O1N2O5K6K6K2OO000O100000000O101O00000O10000000001N100000000O100000000O100000000O100000001O0O100O100O100O100O100O10000O100O100O10O010O0010O010O010O010O01O10O101N100O2O0O100O2O0O1O2N100O0010O01O00010O001O01O01O00010O001O01O01O0010O000O101O00001O0O2O001O001O001N2O001O001O0O2O001O001N101O1O001O0O2O001O001N101K4J7J6L3010O01O0010O01O0010O01O1O010O0010O10O010O0100O0100O0100O010O01O0O2O001O1O1O1N2N3N1NajY2"}, "label": "a giraffe that is in between 2 other giraffes standing in treesh and bushes"}]}
{"id": 24847, "image": "COCO_train2014_000000024847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the giraffe on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 130, 131, 154, 155, 177, 178, 179, 201, 202, 203, 204, 225, 226, 227, 249, 250], "bbox": [0.6539375, 0.1960625, 0.8898906249999999, 0.6567500000000001], "iscrowd": 0, "area": 8353.09975, "rle": {"size": [480, 640], "counts": "``T62n>4M2M3M2O2M2N3K4L5J5A`0N100O1O1O2N1O100O1O1O1F:O2O000O10000NaNZCf0d<YO_Cg0`<XObCi0[<VOhCj0V<UOmCk0Q<UOQDk0m;TOUDm0j;QOYDo0e;mNaDS1];iNiDX1T;eNQE[1m:aNYE_1e:^N_Ec1P<000000001N1O1O100O1O1O1O100O1O1O1O00O2N1N2O1O1O2M2O1O1O1O0O10001O0O10000011N2N2N2O2M2N2O1O1O2N1O2N1O2N1O2N1O00O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O00100O1O1O100O001O1001O00001N101O00001O0O2O001O00001O0O10O1O6I`1aN[`P1"}, "label": "the giraffe on the right in the right hand picture"}]}
{"id": 393493, "image": "COCO_train2014_000000393493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red picnic table\"."}], "task": "refering", "answer_template": "The \"the red picnic table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 164], "bbox": [0.0, 0.376675, 0.36772000000000005, 0.69955], "iscrowd": 0, "area": 13910.0506, "rle": {"size": [400, 500], "counts": "k68g;a0_Oa0_Oa0A?01O0000000000000000001O000000000000000000001O0000000000000000001O00000000000000002N2N1O1O1O1O1O001O1N2O1O0\\GXM`7i2\\HZMd7g2WH]Mi7d2RH`Mn7a2nGbMR8^2jGfMV8[2eGiM[8V3000000000000000000000000000000000J6I7H8H8K5O100O100O10000O100O100O100O10000O100O100O100O10O10O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100bNeEf0\\:XOhEf0X:XOkEg0U:WOoEf0R:XOQFh0n9WOUFh0j9WOXFi0g9UO]Fi0c9VO_Fj0`9UOcFj0\\9UOeFl0Z9TOgFk0Y9TOhFm0X9QOiFP1V9oNkFR1U9mNkFT1X:001O1O1OO5L5K4]O[D2k;FXD:P<N2O1N101O1N2O1N10cck3"}, "label": "the red picnic table"}]}
{"id": 393493, "image": "COCO_train2014_000000393493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red picnic table\"."}], "task": "refering", "answer_template": "The \"the red picnic table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 164], "bbox": [0.0, 0.376675, 0.36772000000000005, 0.69955], "iscrowd": 0, "area": 13910.0506, "rle": {"size": [400, 500], "counts": "k68g;a0_Oa0_Oa0A?01O0000000000000000001O000000000000000000001O0000000000000000001O00000000000000002N2N1O1O1O1O1O001O1N2O1O0\\GXM`7i2\\HZMd7g2WH]Mi7d2RH`Mn7a2nGbMR8^2jGfMV8[2eGiM[8V3000000000000000000000000000000000J6I7H8H8K5O100O100O10000O100O100O100O10000O100O100O100O10O10O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100bNeEf0\\:XOhEf0X:XOkEg0U:WOoEf0R:XOQFh0n9WOUFh0j9WOXFi0g9UO]Fi0c9VO_Fj0`9UOcFj0\\9UOeFl0Z9TOgFk0Y9TOhFm0X9QOiFP1V9oNkFR1U9mNkFT1X:001O1O1OO5L5K4]O[D2k;FXD:P<N2O1N101O1N2O1N10cck3"}, "label": "the red picnic table"}]}
{"id": 393493, "image": "COCO_train2014_000000393493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"park bench behind girl\"."}], "task": "refering", "answer_template": "The \"park bench behind girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 86, 87, 97, 98, 99, 101, 102, 103, 104, 115], "bbox": [0.38672, 0.28190000000000004, 0.8421600000000001, 0.51665], "iscrowd": 0, "area": 8318.6382, "rle": {"size": [400, 500], "counts": "Xn[24\\<1O0O2O1O1_O`0K6ZEROg9P1QFWOn9k0kE[OT:j0bE[O^:V15G9O101NF;100000000003MC^EoN_:Q1gElNV:T1oEjNn9U1YFfNf9Z1g0100000000000000000000000000000000O1000O1000000000000O10000000000O1M2O2M3M3M3M3M3J6J6K5O1O1O1O1O2O0O\\h90eWF2N2N1O2M3N1O2N2N1O2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N1O000000000000000000000000O10O10000000000000000000000000000000O10O10000000000000000000000000000000O010000000000000000000000000000000000O01000000000000000000000000000000000O01004L5K1OPOeEK[:5lEDT:<lE[O]:e0cE[O]:e0cE[O]:e0cE[O]:d0dE\\OP;0PE0[Ro0"}, "label": "park bench behind girl"}]}
{"id": 393493, "image": "COCO_train2014_000000393493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green bench directly behind the girl\"."}], "task": "refering", "answer_template": "The \"the green bench directly behind the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 86, 87, 97, 98, 99, 101, 102, 103, 104, 115], "bbox": [0.38672, 0.28190000000000004, 0.8421600000000001, 0.51665], "iscrowd": 0, "area": 8318.6382, "rle": {"size": [400, 500], "counts": "Xn[24\\<1O0O2O1O1_O`0K6ZEROg9P1QFWOn9k0kE[OT:j0bE[O^:V15G9O101NF;100000000003MC^EoN_:Q1gElNV:T1oEjNn9U1YFfNf9Z1g0100000000000000000000000000000000O1000O1000000000000O10000000000O1M2O2M3M3M3M3M3J6J6K5O1O1O1O1O2O0O\\h90eWF2N2N1O2M3N1O2N2N1O2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N1O000000000000000000000000O10O10000000000000000000000000000000O10O10000000000000000000000000000000O010000000000000000000000000000000000O01000000000000000000000000000000000O01004L5K1OPOeEK[:5lEDT:<lE[O]:e0cE[O]:e0cE[O]:e0cE[O]:d0dE\\OP;0PE0[Ro0"}, "label": "the green bench directly behind the girl"}]}
{"id": 393493, "image": "COCO_train2014_000000393493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl in red pants and shirt\"."}], "task": "refering", "answer_template": "The \"little girl in red pants and shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 81, 82, 99, 100, 101, 117, 118, 119, 136, 154], "bbox": [0.50008, 0.24087499999999998, 0.6415599999999999, 0.604175], "iscrowd": 0, "area": 5055.5967, "rle": {"size": [400, 500], "counts": "bjQ32\\<4L4L2N2M3N1N100O2N2O2M3D<N1WOhNZF[1c9lNVFV1g9POUFQ1h9TOTFn0k9UOQFm0o9k0N101N2O2M3N:E6L3L3M4L3M5aG^Lg7h3oG\\LP8Q4101N3N1N2N3N3L3N3L8I1N1O21oFdLk8a300100M3N1nMVG<l8B_G2d8MgGF\\88oG\\OT8b0VHROn7l0l1O2N2N2N2N2N13N00M3N2M3N]hU2"}, "label": "little girl in red pants and shirt"}]}
{"id": 393493, "image": "COCO_train2014_000000393493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl alone in a park\"."}], "task": "refering", "answer_template": "The \"a girl alone in a park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 81, 82, 99, 100, 101, 117, 118, 119, 136, 154], "bbox": [0.50008, 0.24087499999999998, 0.6415599999999999, 0.604175], "iscrowd": 0, "area": 5055.5967, "rle": {"size": [400, 500], "counts": "bjQ32\\<4L4L2N2M3N1N100O2N2O2M3D<N1WOhNZF[1c9lNVFV1g9POUFQ1h9TOTFn0k9UOQFm0o9k0N101N2O2M3N:E6L3L3M4L3M5aG^Lg7h3oG\\LP8Q4101N3N1N2N3N3L3N3L8I1N1O21oFdLk8a300100M3N1nMVG<l8B_G2d8MgGF\\88oG\\OT8b0VHROn7l0l1O2N2N2N2N2N13N00M3N2M3N]hU2"}, "label": "a girl alone in a park"}]}
{"id": 475415, "image": "COCO_train2014_000000475415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light - colored car in the background behind a chain - link fence\"."}], "task": "refering", "answer_template": "The \"a light - colored car in the background behind a chain - link fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 54, 72, 73, 74, 75, 76, 77, 96, 97, 98], "bbox": [0.01803125, 0.07545667447306792, 0.36515624999999996, 0.30292740046838407], "iscrowd": 0, "area": 10688.418549999997, "rle": {"size": [427, 640], "counts": "lQ52X=3N2N3M2N2N000000000000000000000000000000000001O00@DQD<o;FoC:Q<HmC8S<JkC6U<JkC6T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5S<LmC4S<LmC4S<KnC5R<KnC5R<KnC5R<KnC5R<KnC5R<KnC5R<KnC5R<KnC5Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3o;NQDZ1g:fNYE_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1_:dNaE\\1_:dNaE\\1_:dNaE\\1_:dNaE\\1_:dNaE\\1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1a:bN_E^1a:bN_E^1a:bN_E_1`:aN`E_1`:aN`E`1_:`NaEa1_:^NaEb1_:^NaEc1^:]NbEc1^:]NbEd1]:\\NcEd1]:\\NcEe1]:ZNcEg1\\:YNdEg1\\:YNdEh1[:XNeEh1[:XNeEi1Z:WNfEi1Z:WNfEj1Z:UNfEl1Y:TNgEl1Y:TNgEm1X:SNhEm1X:SNhEn1W:RNiEn1X:QNhEP2W:PNiEQ2V:oMjEQ2V:oMjEQ2V:oMjEQ2V:oMjEQ2W:nMiER2W:nMiER2W:nMiEQ2X:oMhEQ2X:oMhEQ2X:oMhEQ2X:oMhEQ2Y:nMgEQ2Z:oMfEQ2Z:oMfEQ2Z:oMfEQ2Z:oMfEQ2Z:oMfEQ2[:nMeEQ2\\:oMdEP2]:PNcEP2]:PNcEo1^:QNbEn1_:RNaEm1a:RN_Em1b:SN^Em1b:SN^El1e:RN[Em1g:RNYEm1S;O100O1O1000000O10000O10000O10001N101O001N101O0O2O001N101O0O2O001N101O001N101O0O2O3M2M3N2N2M4M2N2M3M3M4L3McVY5"}, "label": "a light - colored car in the background behind a chain - link fence"}]}
{"id": 475415, "image": "COCO_train2014_000000475415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver car in parking lot\"."}], "task": "refering", "answer_template": "The \"silver car in parking lot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 54, 72, 73, 74, 75, 76, 77, 96, 97, 98], "bbox": [0.01803125, 0.07545667447306792, 0.36515624999999996, 0.30292740046838407], "iscrowd": 0, "area": 10688.418549999997, "rle": {"size": [427, 640], "counts": "lQ52X=3N2N3M2N2N000000000000000000000000000000000001O00@DQD<o;FoC:Q<HmC8S<JkC6U<JkC6T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5T<KlC5S<LmC4S<LmC4S<KnC5R<KnC5R<KnC5R<KnC5R<KnC5R<KnC5R<KnC5R<KnC5Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4Q<LoC4P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3P<MPD3o;NQDZ1g:fNYE_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1b:aN^E_1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1a:bN_E^1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1_:dNaE\\1_:dNaE\\1_:dNaE\\1_:dNaE\\1_:dNaE\\1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1`:cN`E]1a:bN_E^1a:bN_E^1a:bN_E_1`:aN`E_1`:aN`E`1_:`NaEa1_:^NaEb1_:^NaEc1^:]NbEc1^:]NbEd1]:\\NcEd1]:\\NcEe1]:ZNcEg1\\:YNdEg1\\:YNdEh1[:XNeEh1[:XNeEi1Z:WNfEi1Z:WNfEj1Z:UNfEl1Y:TNgEl1Y:TNgEm1X:SNhEm1X:SNhEn1W:RNiEn1X:QNhEP2W:PNiEQ2V:oMjEQ2V:oMjEQ2V:oMjEQ2V:oMjEQ2W:nMiER2W:nMiER2W:nMiEQ2X:oMhEQ2X:oMhEQ2X:oMhEQ2X:oMhEQ2Y:nMgEQ2Z:oMfEQ2Z:oMfEQ2Z:oMfEQ2Z:oMfEQ2Z:oMfEQ2[:nMeEQ2\\:oMdEP2]:PNcEP2]:PNcEo1^:QNbEn1_:RNaEm1a:RN_Em1b:SN^Em1b:SN^El1e:RN[Em1g:RNYEm1S;O100O1O1000000O10000O10000O10001N101O001N101O0O2O001N101O0O2O001N101O001N101O0O2O3M2M3N2N2M4M2N2M3M3M4L3McVY5"}, "label": "silver car in parking lot"}]}
{"id": 385305, "image": "COCO_train2014_000000385305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow city bus\"."}], "task": "refering", "answer_template": "The \"a yellow city bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 307], "bbox": [0.06403125, 0.0809020618556701, 0.9659062500000001, 0.934819587628866], "iscrowd": 0, "area": 141800.3444, "rle": {"size": [388, 640], "counts": "Xg?_1b:e1ZN>C4K4M4K4M4K4M4L4K4M4K4M4K4M4K5L3L5L3M4K4M4K5L3O2N1N3N1O2M3N1O2M2O2M2O2N2M2O2N1N3N1O2M2O1O2M2O1O1N3N1O1N2O1N3N1O1N2O2N1N2O1O1N3N1O1N2O2N1N2O1O1N3O0O10001O0000000O2O0000001O00000O101O0000001O000O10001O0000001O0O1000001O0000001N100000001O00000O2O000000001O00000O2O000000001O000O101O000000001O0O10001O000000001N1000001O0000000O2O0000001O00000O101O01O0001O000000001O0000001O0000001O00001O0001O01O00001O0000000000000000000000O10000O101N100O100O10000O100O100O10000O10000O10000O101O0O10000O10000O101N10000O2O000O101O0O10001N10000O2O000O2O0O10001N10000O2O000O101O0O10001N10000O2O0O10001N10000O2O000O101O0O10001O01O00100O001O010O001O010O000001O00010O0000001O0000001N10000O2O0O10001N100O1O2N1O1O1O2M2O1O1O2N1O1N2O2N1O1O1O2N100O1O2N100O101O0O100O2O000O101O0O101O0O10001N10000O2O0O10001N10000O2O000O101O0O10001N100O101O0O10001N10000O2O000O101O0O100O2O000O101O0O101O0O10001N10000O2O0O10001N10000O2O000O101O0O10001N100O101O0O10001N10000O2O000O101O0O100O2O000O101O0O10001N10001N100O101O0O10001N10000O2O000O101O0O100O2O000O101O0O10001N10000O2O00000100O001O100O1O0001O0000001O00001O000O1O2N1N2O1O2M2O1O1O2N100O101N10000O2O0O100O2O0O100O2O0O100O2O0O100O2J5_Oa0_Oa0_Ob0^Oa0_Oa0_ObW8"}, "label": "a yellow city bus"}]}
{"id": 385305, "image": "COCO_train2014_000000385305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bus with the words \" yellow buses \" displayed in white above its door\"."}], "task": "refering", "answer_template": "The \"a yellow bus with the words \" yellow buses \" displayed in white above its door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 307], "bbox": [0.06403125, 0.0809020618556701, 0.9659062500000001, 0.934819587628866], "iscrowd": 0, "area": 141800.3444, "rle": {"size": [388, 640], "counts": "Xg?_1b:e1ZN>C4K4M4K4M4K4M4L4K4M4K4M4K4M4K5L3L5L3M4K4M4K5L3O2N1N3N1O2M3N1O2M2O2M2O2N2M2O2N1N3N1O2M2O1O2M2O1O1N3N1O1N2O1N3N1O1N2O2N1N2O1O1N3N1O1N2O2N1N2O1O1N3O0O10001O0000000O2O0000001O00000O101O0000001O000O10001O0000001O0O1000001O0000001N100000001O00000O2O000000001O00000O2O000000001O000O101O000000001O0O10001O000000001N1000001O0000000O2O0000001O00000O101O01O0001O000000001O0000001O0000001O00001O0001O01O00001O0000000000000000000000O10000O101N100O100O10000O100O100O10000O10000O10000O101O0O10000O10000O101N10000O2O000O101O0O10001N10000O2O000O2O0O10001N10000O2O000O101O0O10001N10000O2O0O10001N10000O2O000O101O0O10001O01O00100O001O010O001O010O000001O00010O0000001O0000001N10000O2O0O10001N100O1O2N1O1O1O2M2O1O1O2N1O1N2O2N1O1O1O2N100O1O2N100O101O0O100O2O000O101O0O101O0O10001N10000O2O0O10001N10000O2O000O101O0O10001N100O101O0O10001N10000O2O000O101O0O100O2O000O101O0O101O0O10001N10000O2O0O10001N10000O2O000O101O0O10001N100O101O0O10001N10000O2O000O101O0O100O2O000O101O0O10001N10001N100O101O0O10001N10000O2O000O101O0O100O2O000O101O0O10001N10000O2O00000100O001O100O1O0001O0000001O00001O000O1O2N1N2O1O2M2O1O1O2N100O101N10000O2O0O100O2O0O100O2O0O100O2O0O100O2J5_Oa0_Oa0_Ob0^Oa0_Oa0_ObW8"}, "label": "a yellow bus with the words \" yellow buses \" displayed in white above its door"}]}
{"id": 16669, "image": "COCO_train2014_000000016669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor turned on\"."}], "task": "refering", "answer_template": "The \"a computer monitor turned on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223], "bbox": [0.284859375, 0.13707259953161594, 0.74815625, 0.6471896955503513], "iscrowd": 0, "area": 58556.26355000001, "rle": {"size": [427, 640], "counts": "UU\\21k<?A?VJXObNW1^1BZMm0f2KRLe0n34jJ;V50PJ?P6^OdIQ1\\6o2000000000O100000000000000000000000001O000O10000000000000000000000000000O100000000000000000001O000000000O10000000000000000000000000000O1000000000001O000000000000000O1000000000000000000000000000000O1000001O000000000000000000000O1000000000000000000000000000000O2O000000000000000000000000000O1000000000000000000000001O00000O10000000000000000000000000000O1000000000000000001O000000000O1000000000000000000000000000000O10000000001O00000000000000000O1000000000000000000000000000000O10001O00000000000000000000000O1000000000000000000000000000001N1000000000000000000000008Hm0SOl0SOn0SOm0SOl0TOm0SOblR2"}, "label": "a computer monitor turned on"}]}
{"id": 16669, "image": "COCO_train2014_000000016669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor with the green pattern on the screen\"."}], "task": "refering", "answer_template": "The \"the monitor with the green pattern on the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223], "bbox": [0.284859375, 0.13707259953161594, 0.74815625, 0.6471896955503513], "iscrowd": 0, "area": 58556.26355000001, "rle": {"size": [427, 640], "counts": "UU\\21k<?A?VJXObNW1^1BZMm0f2KRLe0n34jJ;V50PJ?P6^OdIQ1\\6o2000000000O100000000000000000000000001O000O10000000000000000000000000000O100000000000000000001O000000000O10000000000000000000000000000O1000000000001O000000000000000O1000000000000000000000000000000O1000001O000000000000000000000O1000000000000000000000000000000O2O000000000000000000000000000O1000000000000000000000001O00000O10000000000000000000000000000O1000000000000000001O000000000O1000000000000000000000000000000O10000000001O00000000000000000O1000000000000000000000000000000O10001O00000000000000000000000O1000000000000000000000000000001N1000000000000000000000008Hm0SOl0SOn0SOm0SOl0TOm0SOblR2"}, "label": "the monitor with the green pattern on the screen"}]}
{"id": 16669, "image": "COCO_train2014_000000016669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a monitor with a white screen , text and a picture next to a green monitor\"."}], "task": "refering", "answer_template": "The \"a monitor with a white screen , text and a picture next to a green monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.85759375, 0.2788056206088993, 1.0, 0.6450819672131147], "iscrowd": 0, "area": 12826.9581, "rle": {"size": [427, 640], "counts": "QVU77k<9F:G9G9F:G9G9G9F:G9G9F:G9G9F:J600000000000000000000000001O00000000000000000000000000000000000000000000000O10000000000000001O0000000000000000000000000000000000000000000000000000000YL"}, "label": "a monitor with a white screen , text and a picture next to a green monitor"}]}
{"id": 16669, "image": "COCO_train2014_000000016669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a system was on a table which is showing the movie\"."}], "task": "refering", "answer_template": "The \"a system was on a table which is showing the movie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.85759375, 0.2788056206088993, 1.0, 0.6450819672131147], "iscrowd": 0, "area": 12826.9581, "rle": {"size": [427, 640], "counts": "QVU77k<9F:G9G9F:G9G9G9F:G9G9F:G9G9F:J600000000000000000000000001O00000000000000000000000000000000000000000000000O10000000000000001O0000000000000000000000000000000000000000000000000000000YL"}, "label": "a system was on a table which is showing the movie"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey travel bag packed full , has black wheels on the bottom\"."}], "task": "refering", "answer_template": "The \"grey travel bag packed full , has black wheels on the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355, 356, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.228328125, 0.7686965811965812, 0.5253125, 0.9931837606837607], "iscrowd": 0, "area": 10647.208449999996, "rle": {"size": [468, 640], "counts": "ccS22b>2M2O1O1N2O1N2O1N3N1N2O1N2O1O1N3N0O010O010O010N1O2M2O2M2O2M2O2N2O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N20000O100O100O102M3N3M2M2O0O100O10000O100O100O10000O100O100O100O10000O100O100O101O0O100O100O100O10000O10000000HeDWMZ;b2QE[Mo:]2h0H8J6N2N2N2O0O2N2N2O1N2N2N2O1N1O2N2L4M3L4L41O0001O0001O0001O01O002N3N1N3M2N3N1N3M3M5K4M3L4L4L4L5L3L4L4L2N01O0O1000001O0O2O1O1K5nNR1N2M3N2N2M3N2N2M3N1O2M4M2N2M3N3M2M3NjaZ4"}, "label": "grey travel bag packed full , has black wheels on the bottom"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bag on the table\"."}], "task": "refering", "answer_template": "The \"a bag on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355, 356, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.228328125, 0.7686965811965812, 0.5253125, 0.9931837606837607], "iscrowd": 0, "area": 10647.208449999996, "rle": {"size": [468, 640], "counts": "ccS22b>2M2O1O1N2O1N2O1N3N1N2O1N2O1O1N3N0O010O010O010N1O2M2O2M2O2M2O2N2O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N20000O100O100O102M3N3M2M2O0O100O10000O100O100O10000O100O100O100O10000O100O100O101O0O100O100O100O10000O10000000HeDWMZ;b2QE[Mo:]2h0H8J6N2N2N2O0O2N2N2O1N2N2N2O1N1O2N2L4M3L4L41O0001O0001O0001O01O002N3N1N3M2N3N1N3M3M5K4M3L4L4L4L5L3L4L4L2N01O0O1000001O0O2O1O1K5nNR1N2M3N2N2M3N2N2M3N1O2M4M2N2M3N3M2M3NjaZ4"}, "label": "a bag on the table"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with full of cloths\"."}], "task": "refering", "answer_template": "The \"a bed with full of cloths\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 242, 243, 244, 257, 258, 259, 260, 261, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 326, 327, 328, 329, 349, 350, 351], "bbox": [0.18157812499999998, 0.4275, 0.6609062499999999, 0.9219230769230768], "iscrowd": 0, "area": 27138.218500000003, "rle": {"size": [468, 640], "counts": "e[e1c0l=^1bN4M3L4L5L4K6N1O2N1O2N1O2N2N1O2N1O1O2N1O101N1000001O00001O00001O0000001O00001O0O2O00001O001O001O]LXEZ3h:fLXEZ3h:eLYE[3g:dLZE[3g:dLZE[3g:dLZE[3g:dLZE\\3f:dLZE[3g:dLZE[3g:dLZE\\3o:O1O100O1O100O1O1O1O100O001O100O1O1O1O100O1O1O1O100O1O1O0O2O1O1O1O1O1O1O1O1O1O0O2O1O1O2N2N2N2N2N2N2N2O1O1N2O1O001N10001O0O2O000O2O0O2O000O2O0O101O0O2O0O101O0O100O10O10O100O010O100000O10001O000O101O000O101O00000O2O00001O0O10001O0000001O0O10001O0000001O00001O00000O2O0000001N1000001O0O10000YMZN`He1^7eN[H[1a7oNYHQ1d7YOUHf0h7DRH<k7MoG3m72RHNk75UHJk77UHIk77UHIk77UHHk79UHGk79UHGk79UHGj7:VHEk7;UHEk7;UHEj7<VHDj7<VHCj7>VHBj7=WHCi7=WHBi7?WHAi7?WHAi7?WHAh7?YH@h7`0XH@h7`0YH_Of7b0nFRNn0[1U8c0lFTNn0Y1V8b0kFXNm0W1W8b0kFXNn0V1W8b0iF[No0R1Y8c0gF]Nn0Q1Z8c0fF^NP1o0Z8b0fFaNn0n0\\8a0dFdNo0j0]8c0cFdNP1i0]8Z1bGfN^8[1bGcN^8^1bGbN^8_1bG`N^8a1aG_N^8b1cG\\N^8e1bGZN^8g1aGYN_8g1bGWN_8j1aGUN_8l1`GTN`8l1aGSN_8n1aGPN`8Q2_GoMa8R2_GmMa8T2_GkMa8V2^GiMc8X2]GgMc8Y2^GfMb8[2]GdMd8]2\\GbMd8_2\\G`Md8a2[G_Me8b2[G\\Mf8e2ZGZMf8g2YGYMg8h2XGWMi8j2WGUMi8k2WGUMi8l2VGTMj8m2UGRMl8P3SGoLm8S3QGmLo8U3oFkLR9V3lFiLU9Y3iFgLX9Z3eFgL[9[3bFeL`9\\3]FeLc9]3ZFdLg9]3VFdLj9R42O000O2O001O0O2O001O0O2SMlEQ1[:lNlEl0Y:POnEh0V:UORFb0T:YOSF?R:]OUF;P:@XF8m9DZF4k9H\\F0i9L^FLg90`FHe94`2KknR3"}, "label": "a bed with full of cloths"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dress in the store\"."}], "task": "refering", "answer_template": "The \"dress in the store\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 302, 322, 323], "bbox": [0.0, 0.40782051282051285, 0.2165625, 0.8681196581196581], "iscrowd": 0, "area": 18949.269800000002, "rle": {"size": [468, 640], "counts": "d6R6b80O1000000O010000000000000O0100000000000000O100O1O1O001O1O1O001O10O01O1O001O1O001O1O001O001O1O010O1N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N110O10O100000O010000O1000O3N2N1O2=B>CM2O2M2TOdF^L^9_3oFTLS9j3n0N2M2N3N1N3M2O2M2N3N1N3M3N1N3M2O2M2M4YOSDaNP<Z1WDaNm;Z1ZD`Ni;]1\\D^Nh;^1j0M2N3M2O2M2N3M101N1O2O1N1O2O1N1O2N2O0M4L3N3LcUU7"}, "label": "dress in the store"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a back pack hanging on a wall with the pockets open\"."}], "task": "refering", "answer_template": "The \"a back pack hanging on a wall with the pockets open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 302, 322, 323], "bbox": [0.0, 0.40782051282051285, 0.2165625, 0.8681196581196581], "iscrowd": 0, "area": 18949.269800000002, "rle": {"size": [468, 640], "counts": "d6R6b80O1000000O010000000000000O0100000000000000O100O1O1O001O1O1O001O10O01O1O001O1O001O1O001O001O1O010O1N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N110O10O100000O010000O1000O3N2N1O2=B>CM2O2M2TOdF^L^9_3oFTLS9j3n0N2M2N3N1N3M2O2M2N3N1N3M3N1N3M2O2M2M4YOSDaNP<Z1WDaNm;Z1ZD`Ni;]1\\D^Nh;^1j0M2N3M2O2M2N3M101N1O2O1N1O2O1N1O2N2O0M4L3N3LcUU7"}, "label": "a back pack hanging on a wall with the pockets open"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed and headboard of the bed the man is standing next to\"."}], "task": "refering", "answer_template": "The \"the bed and headboard of the bed the man is standing next to\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 291, 292, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.555875, 0.4457051282051282, 1.0, 0.966068376068376], "iscrowd": 0, "area": 37552.503950000006, "rle": {"size": [468, 640], "counts": "hRS53`>5L3L5K4L4L4M0O2N1O1O101N1O1O2O0O1O1O2O0O1O2O0O100O2O0O101N1O100O2O0O101N1O100O2O0O100O2O0O1O2O0O10001N:G0O2O0O10000O101N100O10000O101N100O10000O101N100O10000O101N100O10000O2O0O100O10000O2O0O100O10`MhDd1X;SNQEm1n:kM\\EU2b;O1O1O1O1O010O000000000001O01O0000000001O00000000001O01O0000000001O0000000000010O0000000000001N10000O10000O100lMcMQH]2o7dMPH\\2P8dMoG]2Q8dMnG\\2R8eMmG[2S8fMkG[2U8fMjGZ2V8hMhGX2m7TNRHl1k6XOSIi0c6B\\I>\\6LbI4\\6ObI2[64bIL\\68bIH\\6<aIE]6?aIA]6c0`I^O^6f0`IZO^6i0`IXO^6l0`ITO^6o0`IRO_6P1`IoNa6S1\\InNd6S1[ImNe6U1XIlNh6U1WIkNi6V1UIkNj6X1SIiNm6X1RIhNn6Y1PIhNP7Z1mHgNS7Z1iHiNW7Z40000O10000000000000000O100000000O10000O10000O100O10000O01000O100O10000O10000000000000000000000000000000000000000000000000000000000000O100000000000000000001O1O1O1O1O1O1RORITJo6\\5iI[JW6V5o1\\Oh0YOg0YOh0XOg0XOi0XOg0YOPH"}, "label": "the bed and headboard of the bed the man is standing next to"}]}
{"id": 278816, "image": "COCO_train2014_000000278816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed that is gentleman is going through his items on . it is to the right of the other bed\"."}], "task": "refering", "answer_template": "The \"the bed that is gentleman is going through his items on . it is to the right of the other bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 291, 292, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.555875, 0.4457051282051282, 1.0, 0.966068376068376], "iscrowd": 0, "area": 37552.503950000006, "rle": {"size": [468, 640], "counts": "hRS53`>5L3L5K4L4L4M0O2N1O1O101N1O1O2O0O1O1O2O0O1O2O0O100O2O0O101N1O100O2O0O101N1O100O2O0O100O2O0O1O2O0O10001N:G0O2O0O10000O101N100O10000O101N100O10000O101N100O10000O101N100O10000O2O0O100O10000O2O0O100O10`MhDd1X;SNQEm1n:kM\\EU2b;O1O1O1O1O010O000000000001O01O0000000001O00000000001O01O0000000001O0000000000010O0000000000001N10000O10000O100lMcMQH]2o7dMPH\\2P8dMoG]2Q8dMnG\\2R8eMmG[2S8fMkG[2U8fMjGZ2V8hMhGX2m7TNRHl1k6XOSIi0c6B\\I>\\6LbI4\\6ObI2[64bIL\\68bIH\\6<aIE]6?aIA]6c0`I^O^6f0`IZO^6i0`IXO^6l0`ITO^6o0`IRO_6P1`IoNa6S1\\InNd6S1[ImNe6U1XIlNh6U1WIkNi6V1UIkNj6X1SIiNm6X1RIhNn6Y1PIhNP7Z1mHgNS7Z1iHiNW7Z40000O10000000000000000O100000000O10000O10000O100O10000O01000O100O10000O10000000000000000000000000000000000000000000000000000000000000O100000000000000000001O1O1O1O1O1O1RORITJo6\\5iI[JW6V5o1\\Oh0YOg0YOh0XOg0XOi0XOg0YOPH"}, "label": "the bed that is gentleman is going through his items on . it is to the right of the other bed"}]}
{"id": 508200, "image": "COCO_train2014_000000508200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vegitable salad\"."}], "task": "refering", "answer_template": "The \"vegitable salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.394, 0.26368544600938965, 1.0, 0.9834037558685447], "iscrowd": 0, "area": 95550.71835000001, "rle": {"size": [426, 640], "counts": "aQY37Q=a0^Ob0^O<E4K5K5L4K5L5L3L4L4M3L4L4M3L4L5L3L5L4K4L5L3M4L3M4L4L3N3L3M4L4L2N2N2N3M2N2N2N2O1N3M2N2N2N2N3M2N2N2N2O1N3M2N2N2N2N3N1N2O1O1N2O1O1N101N101O0O2O1O0O2O0O2O1O0O2O001N2O001N101N2O001N101O1N101N101O1N101O0O2O1O0O2O0O2O1O0O2O001N2O0O2O001N101O1N101O0O2O1N101O001O001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000O10001O0000000001O0000000000000O100000000000001O0000000000000000000000000O2O000000000000000000000000001O000000000O1000000000000000001O000000000000000000001N10001O00001O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N10001N10001N10001N10001N10001N10001N10001N10001N101O1N2O001N2O1O1N1O2N2N2N2N1N3N2N2N2N1O2M3N2N2N1O2N2M3N2N1O2N2N2M3N1O2N2N2N2M2O2N2N2N1O2M3N2N2N1O2N2M3N3M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3M3L4M3MVI"}, "label": "vegitable salad"}]}
{"id": 508200, "image": "COCO_train2014_000000508200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bowl full of bean and tomato salsa\"."}], "task": "refering", "answer_template": "The \"a white bowl full of bean and tomato salsa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.394, 0.26368544600938965, 1.0, 0.9834037558685447], "iscrowd": 0, "area": 95550.71835000001, "rle": {"size": [426, 640], "counts": "aQY37Q=a0^Ob0^O<E4K5K5L4K5L5L3L4L4M3L4L4M3L4L5L3L5L4K4L5L3M4L3M4L4L3N3L3M4L4L2N2N2N3M2N2N2N2O1N3M2N2N2N2N3M2N2N2N2O1N3M2N2N2N2N3N1N2O1O1N2O1O1N101N101O0O2O1O0O2O0O2O1O0O2O001N2O001N101N2O001N101O1N101N101O1N101O0O2O1O0O2O0O2O1O0O2O001N2O0O2O001N101O1N101O0O2O1N101O001O001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000O10001O0000000001O0000000000000O100000000000001O0000000000000000000000000O2O000000000000000000000000001O000000000O1000000000000000001O000000000000000000001N10001O00001O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N10001N10001N10001N10001N10001N10001N10001N10001N101O1N2O001N2O1O1N1O2N2N2N2N1N3N2N2N2N1O2M3N2N2N1O2N2M3N2N1O2N2N2M3N1O2N2N2N2M2O2N2N2N1O2M3N2N2N1O2N2M3N3M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3M3L4M3MVI"}, "label": "a white bowl full of bean and tomato salsa"}]}
{"id": 508200, "image": "COCO_train2014_000000508200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl full of chips that is out of focus in this picture , behind the orange\"."}], "task": "refering", "answer_template": "The \"a bowl full of chips that is out of focus in this picture , behind the orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 26, 27, 28, 29, 30, 31, 32, 33, 34, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170], "bbox": [0.06146875, 0.036032863849765256, 0.5411875, 0.5225117370892018], "iscrowd": 0, "area": 47051.649950000006, "rle": {"size": [426, 640], "counts": "bZ`0c0e<5K6J5K6J5K6J4L4L3M3M4L3N2N3M2N3M2M3N3M2N2N3M2N3L2O2N2N2N2N2N2N2N2N2N1O2N2N2O1N2N2N2N2N2N1O2N2N2N2O1N2N2O0O101N100O2O0O2N100O2O0O101N100O2O0O2O0O1O2O0O101N100O2O0O2O0O101N100O2O000O2O001O000O2O00001O0O101O001O0O101O00000O100000000O101O00000O100000000O100000001O0000000000000000000O10001O0000000000000000000000001O000O010001O000000000000000000001O0000000000000O100000001O00000000000000000000001O000O100O100O100O2O0O100O2O0O101N100N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2N2N3M2N2O2N1O1O1N3N1O2N1O1N3N1O1O2N1N2O2N1O1O2N1O101N100O2N100O2O0O2N101N2N2O0O2O1N2N101N2O1N1O2O1N2O0O2O1O1N101O1N2O0O2O2N2M3N2M3N2N2L4L4L4L4L4L4K6K9G:FiVj3"}, "label": "a bowl full of chips that is out of focus in this picture , behind the orange"}]}
{"id": 508200, "image": "COCO_train2014_000000508200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of wheat thin crackers on a table\"."}], "task": "refering", "answer_template": "The \"a bowl of wheat thin crackers on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 26, 27, 28, 29, 30, 31, 32, 33, 34, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170], "bbox": [0.06146875, 0.036032863849765256, 0.5411875, 0.5225117370892018], "iscrowd": 0, "area": 47051.649950000006, "rle": {"size": [426, 640], "counts": "bZ`0c0e<5K6J5K6J5K6J4L4L3M3M4L3N2N3M2N3M2M3N3M2N2N3M2N3L2O2N2N2N2N2N2N2N2N2N1O2N2N2O1N2N2N2N2N2N1O2N2N2N2O1N2N2O0O101N100O2O0O2N100O2O0O101N100O2O0O2O0O1O2O0O101N100O2O0O2O0O101N100O2O000O2O001O000O2O00001O0O101O001O0O101O00000O100000000O101O00000O100000000O100000001O0000000000000000000O10001O0000000000000000000000001O000O010001O000000000000000000001O0000000000000O100000001O00000000000000000000001O000O100O100O100O2O0O100O2O0O101N100N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2N2N3M2N2O2N1O1O1N3N1O2N1O1N3N1O1O2N1N2O2N1O1O2N1O101N100O2N100O2O0O2N101N2N2O0O2O1N2N101N2O1N1O2O1N2O0O2O1O1N101O1N2O0O2O2N2M3N2M3N2N2L4L4L4L4L4L4K6K9G:FiVj3"}, "label": "a bowl of wheat thin crackers on a table"}]}
{"id": 147760, "image": "COCO_train2014_000000147760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the red with number 10\"."}], "task": "refering", "answer_template": "The \"the man in the red with number 10\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 98, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 265, 266, 281, 282, 283, 284, 285, 286, 288, 289, 304, 305, 306, 307, 308, 309, 312, 313, 327, 328, 329, 330, 331, 332, 335, 336], "bbox": [0.176921875, 0.22920374707259955, 0.641703125, 0.9820140515222482], "iscrowd": 0, "area": 48878.50685, "rle": {"size": [427, 640], "counts": "T\\_13R=<E:E<D;E;F;D;E:G6L4N2O1N3M2N2N2N2N3M2N2O1N3M2N2N2N2N3M2O1N3ZIjKR4[4XKWLg4n3cJdL\\5`3oIRMP6U5O1O1O100O1O1O1O100O2O000O10000O100O10000O10000O100O10000O10000O10000O100O10000O100nNR1N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2N2O100001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000002N2N2N2N2N2N3M6J5K6J6J6J6J6J>B1O1O001O1O1O1O001O1O1O1O001O1O1O1iLkITOX6g0PJTOQ6g0WJUOk5e0]JVOf5e0aJWOa5c0gJXO\\5c0kJYOV5b0SKXOP5e0UKWOm4f0XKVOj4g0[KTOh4i0]KSOd4k0aKPOc4k0dKPO^4l0hKoN[4n0jKnNY4Q1iKjNZ4V1n310O00010O0001O010O00010O00010O00010O00010O0010O00100O1O1O100O1O101N1O100O1O100O1O100O1O1O2O000000O2N100O2N100O1O2O0O1O101O0000001O0000001O00001O000000O1N2O1O0O2O001N101O0O101N101O0O2O000O2O1O2M3N1N3N1N3M2N3L4M2N3L3N2M3N2N2M]`o2"}, "label": "the man in the red with number 10"}]}
{"id": 147760, "image": "COCO_train2014_000000147760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a football player wear red t - shirt numbered as 10\"."}], "task": "refering", "answer_template": "The \"a football player wear red t - shirt numbered as 10\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 98, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 265, 266, 281, 282, 283, 284, 285, 286, 288, 289, 304, 305, 306, 307, 308, 309, 312, 313, 327, 328, 329, 330, 331, 332, 335, 336], "bbox": [0.176921875, 0.22920374707259955, 0.641703125, 0.9820140515222482], "iscrowd": 0, "area": 48878.50685, "rle": {"size": [427, 640], "counts": "T\\_13R=<E:E<D;E;F;D;E:G6L4N2O1N3M2N2N2N2N3M2N2O1N3M2N2N2N2N3M2O1N3ZIjKR4[4XKWLg4n3cJdL\\5`3oIRMP6U5O1O1O100O1O1O1O100O2O000O10000O100O10000O10000O100O10000O10000O10000O100O10000O100nNR1N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2N2O100001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000002N2N2N2N2N2N3M6J5K6J6J6J6J6J>B1O1O001O1O1O1O001O1O1O1O001O1O1O1iLkITOX6g0PJTOQ6g0WJUOk5e0]JVOf5e0aJWOa5c0gJXO\\5c0kJYOV5b0SKXOP5e0UKWOm4f0XKVOj4g0[KTOh4i0]KSOd4k0aKPOc4k0dKPO^4l0hKoN[4n0jKnNY4Q1iKjNZ4V1n310O00010O0001O010O00010O00010O00010O00010O0010O00100O1O1O100O1O101N1O100O1O100O1O100O1O1O2O000000O2N100O2N100O1O2O0O1O101O0000001O0000001O00001O000000O1N2O1O0O2O001N101O0O101N101O0O2O000O2O1O2M3N1N3N1N3M2N3L4M2N3L3N2M3N2N2M]`o2"}, "label": "a football player wear red t - shirt numbered as 10"}]}
{"id": 147760, "image": "COCO_train2014_000000147760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with white t - shirt is trying to get the ball\"."}], "task": "refering", "answer_template": "The \"a person with white t - shirt is trying to get the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.5373281249999999, 0.25098360655737706, 0.9522656250000001, 0.9981733021077284], "iscrowd": 0, "area": 45605.05495, "rle": {"size": [427, 640], "counts": "`j_47S=6K2M3M3N2M3M3M2O1N2N1O2O1N1O2N2O0O2O0O100OTOUD:k;EWD9i;HYD6g;JZD5e;L]D2c;N^D1a;0aDM_;3cDHa;8aDAc;`0d00O10O0100O100O10O0100O100O010O100O10O0100O100O10O0100O1N2O0O2O1nDnNi9T1TFoNj9S1TFoNn6OiJS1VNQOP72cJn0[NSOQ76]Jh0`NTOR7:WJd0dNUOT7=QJ`0iNUOU7b0kI:mNWOW7e0eI5ROYOX7h0`I0VOZOY7m0YIK[O[O[7P1TIF_O\\O\\7T1nHBC]O]7Y1hH[OJ^O]7]1cHVOM@_7b4O001O1O1O1O1O1O1O1N2O1O1O1O1mM[J_Lf5T3nJdLS5o2aKiL`4k2SLmLn3S3ZLdLg3[3aL]L`3b3U300O1O1O10O01O100O1O1O100O1O10000000000000000000000000000000O10000000000000000000O100000000000000000000001O2N2N2M3N2N2N2N2gHSLP5o3kJTLU5n3fJVLY5l3bJWL^5k3]JXLc5j3XJYLh5i3SJZLm5h3nI[LR6g3iI]LV6e3dI_L\\6c3_I`La6b3ZIaLf6a3UIbLk6`3PIdLo6l40000000001O0O1000001O00000000001O000000001O00000000001O1O1O2N1N3N[I`I`6d62N3M2N3L3N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2M4M2N2N3M2N3M2N2N3L3M3M4L3L5L3M5K4L5K4K6K4L5K4L5J5L5K5K4L5K4K6K4L5K4L>Agc<"}, "label": "a person with white t - shirt is trying to get the ball"}]}
{"id": 147760, "image": "COCO_train2014_000000147760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on the right wearing number 13\"."}], "task": "refering", "answer_template": "The \"man on the right wearing number 13\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.5373281249999999, 0.25098360655737706, 0.9522656250000001, 0.9981733021077284], "iscrowd": 0, "area": 45605.05495, "rle": {"size": [427, 640], "counts": "`j_47S=6K2M3M3N2M3M3M2O1N2N1O2O1N1O2N2O0O2O0O100OTOUD:k;EWD9i;HYD6g;JZD5e;L]D2c;N^D1a;0aDM_;3cDHa;8aDAc;`0d00O10O0100O100O10O0100O100O010O100O10O0100O100O10O0100O1N2O0O2O1nDnNi9T1TFoNj9S1TFoNn6OiJS1VNQOP72cJn0[NSOQ76]Jh0`NTOR7:WJd0dNUOT7=QJ`0iNUOU7b0kI:mNWOW7e0eI5ROYOX7h0`I0VOZOY7m0YIK[O[O[7P1TIF_O\\O\\7T1nHBC]O]7Y1hH[OJ^O]7]1cHVOM@_7b4O001O1O1O1O1O1O1O1N2O1O1O1O1mM[J_Lf5T3nJdLS5o2aKiL`4k2SLmLn3S3ZLdLg3[3aL]L`3b3U300O1O1O10O01O100O1O1O100O1O10000000000000000000000000000000O10000000000000000000O100000000000000000000001O2N2N2M3N2N2N2N2gHSLP5o3kJTLU5n3fJVLY5l3bJWL^5k3]JXLc5j3XJYLh5i3SJZLm5h3nI[LR6g3iI]LV6e3dI_L\\6c3_I`La6b3ZIaLf6a3UIbLk6`3PIdLo6l40000000001O0O1000001O00000000001O000000001O00000000001O1O1O2N1N3N[I`I`6d62N3M2N3L3N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2M4M2N2N3M2N3M2N2N3L3M3M4L3L5L3M5K4L5K4K6K4L5K4L5J5L5K5K4L5K4K6K4L5K4L>Agc<"}, "label": "man on the right wearing number 13"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing white color t shirt and holding wine glass\"."}], "task": "refering", "answer_template": "The \"a man wearing white color t shirt and holding wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0016875000000000002, 0.0016875000000000002, 0.6264062499999999, 0.9858958333333334], "iscrowd": 0, "area": 116025.56690000002, "rle": {"size": [480, 640], "counts": "Ql0b1Z:T3bM^2O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1mNS1M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M3M300O1O100O1O100O1O100O100O1O100O1O100O100O100000000000000000000000000000000000000000000O1000^Jf0^KZOb4g0]KYOb4i0]KWOc4k0[KUOe4l0ZKTOf4m0ZKROe4Q1YKoNg4R1XKnNh4S1WKmNi4T1VKlNi4W1UKiNk4X1TKhNl4Y1SKgNm4Z1RKfNm4]1QKcNo4^1PKbNP5_1oJaNQ5`1nJ`NQ5c1mJ]NS5d1lJ\\NT5e1lJZNT5g1kJYNT5j1jJUNW5l1hJTNX5m1gJSNY5V2^JjMa5c2SJ]Mm5o2gIQMY6Z3\\IfLd6d3RI\\Ln6j3lHVLS7Q4gHoKY7T4dHlK\\7m3kHSLU7g3RIXLm6h3TIXLl6h3TIXLl6g3UIYLk6f3VIZLi6g3WIYLi6f3XIZLh6f3XIZLh6e3YI[Lf6e3[I[Le6e3ZI\\Le6d3\\I\\Ld6d3\\I\\Lc6d3^I\\La6e3_IZLa6f3`IZL`6e3`I\\L_6e3aI[L^6e3cI[L]6e3cI[L\\6e3eI[LZ6f3fIZLZ6e3fI\\LY6d3hI\\LW6e3iI[LV6e3kI[LU6e3kI[LT6e3mI[LR6f3mI[LS6d3nI\\LQ6d3PJ\\Lo5e3QJ[Lo5d3RJ\\Lm5e3SJ[Ll5e3TJ\\Ll5c3UJ]Lj5c3WJ]Lh5c3YJ]Lf5c3[J]Lc5d3^J\\L`5e3aJZL^5h3aJYL]5h3dJXLY5j3hJVLV5k3kJULS5l3nJTLP5n3PKRLn4o3RKRLl4P4TKPLi4R4XKnKf4T4ZKlKd4U4]KkKa4W4_KiK_4X4aKiK]4Y4dKfKY4\\4hKdKV4^4jKbKT4_4mKaKQ4a4oK_Ko3b4RL^Kk3e4VLZKh3g4YLXKf3j4ZLVKd3k4]LUKa3m4_LSK_3n4bLRK[3Q5fLnJX3S5iLmJU3U5kLkJS3V5nLjJP3W5QMiJm2W5UMiJi2X5YMgJe2Z5\\MfJc2Z5^MfJ`2[5aMeJ]2[5eMeJZ2[5gMeJW2\\5kMbJT2_5mMaJR2_5oMaJo1`5RN`Jl1`5VN`Jh1a5YN_Jf1a5\\N^Jb1c5_N]J_1d5bN\\J]1d5dN\\JZ1e5gN[JW1f5jNZJT1g5nNXJQ1i5oNWJo0j5ROVJl0k5UOVJi0j5XOVJf0l5ZOTJd0m5^OSJ`0m5ASJ=n5DRJ:P6FQJ7P6JPJ5P6LPJ2Q60oIMS63mILS65mIIT68mIET6<lIBV6?jI_OV6b0lIZOR6j0oISOo5Q1SJlNi5Y1YJcNd5b1]J[N`5k1aJQN\\5T2eJjMX5Z2jJbMW5_2jJ^MW5c2kJZMU5g2mJUMT5l2mJQMT5P3nJmLR5T3oJiLR5X3PKbLS5_3oJZLU5g3lJTLW5m3kJlKZ5T4gJfK]5[4eJ^K`5b4aJYKb5h4`JQKe5o4]JjJg5W5ZJcJk5]5WJ]Jl5d5h10O100O100O100JYJRGh5b8dJ^G\\5a8fJ]G[5c8a00000O10eLbG4^8KkGMU83SHEm7;WHAh7?ZH@f7`0\\H^Od7a0^H^Ob7b0_H]O`7d0aH[O_7d0cH[O]7e0dHnNoNZN]8g2fHnNPOWN\\8j2eHmNSOVNX8m2gHjNf7U1\\HiNe7W1\\HgNe7Y1\\HeNe7Z1]HdNd7\\1]HbNe7\\1]HbNd7^1^H_Nc7a1`HVNf7i1^HSNc7m1`HoMa7Q2bHkM_7T2eHgM^7X2eHdM\\7[2iH`MX7`2kH\\MV7d2c2O1O1M3M4K4M4L4L4L4L4K4M4L4L4L4L3L5L4L4L4L3L5LWh_3"}, "label": "a man wearing white color t shirt and holding wine glass"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding champagne glass\"."}], "task": "refering", "answer_template": "The \"man holding champagne glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0016875000000000002, 0.0016875000000000002, 0.6264062499999999, 0.9858958333333334], "iscrowd": 0, "area": 116025.56690000002, "rle": {"size": [480, 640], "counts": "Ql0b1Z:T3bM^2O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1mNS1M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M3M300O1O100O1O100O1O100O100O1O100O1O100O100O100000000000000000000000000000000000000000000O1000^Jf0^KZOb4g0]KYOb4i0]KWOc4k0[KUOe4l0ZKTOf4m0ZKROe4Q1YKoNg4R1XKnNh4S1WKmNi4T1VKlNi4W1UKiNk4X1TKhNl4Y1SKgNm4Z1RKfNm4]1QKcNo4^1PKbNP5_1oJaNQ5`1nJ`NQ5c1mJ]NS5d1lJ\\NT5e1lJZNT5g1kJYNT5j1jJUNW5l1hJTNX5m1gJSNY5V2^JjMa5c2SJ]Mm5o2gIQMY6Z3\\IfLd6d3RI\\Ln6j3lHVLS7Q4gHoKY7T4dHlK\\7m3kHSLU7g3RIXLm6h3TIXLl6h3TIXLl6g3UIYLk6f3VIZLi6g3WIYLi6f3XIZLh6f3XIZLh6e3YI[Lf6e3[I[Le6e3ZI\\Le6d3\\I\\Ld6d3\\I\\Lc6d3^I\\La6e3_IZLa6f3`IZL`6e3`I\\L_6e3aI[L^6e3cI[L]6e3cI[L\\6e3eI[LZ6f3fIZLZ6e3fI\\LY6d3hI\\LW6e3iI[LV6e3kI[LU6e3kI[LT6e3mI[LR6f3mI[LS6d3nI\\LQ6d3PJ\\Lo5e3QJ[Lo5d3RJ\\Lm5e3SJ[Ll5e3TJ\\Ll5c3UJ]Lj5c3WJ]Lh5c3YJ]Lf5c3[J]Lc5d3^J\\L`5e3aJZL^5h3aJYL]5h3dJXLY5j3hJVLV5k3kJULS5l3nJTLP5n3PKRLn4o3RKRLl4P4TKPLi4R4XKnKf4T4ZKlKd4U4]KkKa4W4_KiK_4X4aKiK]4Y4dKfKY4\\4hKdKV4^4jKbKT4_4mKaKQ4a4oK_Ko3b4RL^Kk3e4VLZKh3g4YLXKf3j4ZLVKd3k4]LUKa3m4_LSK_3n4bLRK[3Q5fLnJX3S5iLmJU3U5kLkJS3V5nLjJP3W5QMiJm2W5UMiJi2X5YMgJe2Z5\\MfJc2Z5^MfJ`2[5aMeJ]2[5eMeJZ2[5gMeJW2\\5kMbJT2_5mMaJR2_5oMaJo1`5RN`Jl1`5VN`Jh1a5YN_Jf1a5\\N^Jb1c5_N]J_1d5bN\\J]1d5dN\\JZ1e5gN[JW1f5jNZJT1g5nNXJQ1i5oNWJo0j5ROVJl0k5UOVJi0j5XOVJf0l5ZOTJd0m5^OSJ`0m5ASJ=n5DRJ:P6FQJ7P6JPJ5P6LPJ2Q60oIMS63mILS65mIIT68mIET6<lIBV6?jI_OV6b0lIZOR6j0oISOo5Q1SJlNi5Y1YJcNd5b1]J[N`5k1aJQN\\5T2eJjMX5Z2jJbMW5_2jJ^MW5c2kJZMU5g2mJUMT5l2mJQMT5P3nJmLR5T3oJiLR5X3PKbLS5_3oJZLU5g3lJTLW5m3kJlKZ5T4gJfK]5[4eJ^K`5b4aJYKb5h4`JQKe5o4]JjJg5W5ZJcJk5]5WJ]Jl5d5h10O100O100O100JYJRGh5b8dJ^G\\5a8fJ]G[5c8a00000O10eLbG4^8KkGMU83SHEm7;WHAh7?ZH@f7`0\\H^Od7a0^H^Ob7b0_H]O`7d0aH[O_7d0cH[O]7e0dHnNoNZN]8g2fHnNPOWN\\8j2eHmNSOVNX8m2gHjNf7U1\\HiNe7W1\\HgNe7Y1\\HeNe7Z1]HdNd7\\1]HbNe7\\1]HbNd7^1^H_Nc7a1`HVNf7i1^HSNc7m1`HoMa7Q2bHkM_7T2eHgM^7X2eHdM\\7[2iH`MX7`2kH\\MV7d2c2O1O1M3M4K4M4L4L4L4L4K4M4L4L4L4L3L5L4L4L4L3L5LWh_3"}, "label": "man holding champagne glass"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt with a glass of drink\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt with a glass of drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0016875000000000002, 0.0016875000000000002, 0.6264062499999999, 0.9858958333333334], "iscrowd": 0, "area": 116025.56690000002, "rle": {"size": [480, 640], "counts": "Ql0b1Z:T3bM^2O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1mNS1M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M3M300O1O100O1O100O1O100O100O1O100O1O100O100O100000000000000000000000000000000000000000000O1000^Jf0^KZOb4g0]KYOb4i0]KWOc4k0[KUOe4l0ZKTOf4m0ZKROe4Q1YKoNg4R1XKnNh4S1WKmNi4T1VKlNi4W1UKiNk4X1TKhNl4Y1SKgNm4Z1RKfNm4]1QKcNo4^1PKbNP5_1oJaNQ5`1nJ`NQ5c1mJ]NS5d1lJ\\NT5e1lJZNT5g1kJYNT5j1jJUNW5l1hJTNX5m1gJSNY5V2^JjMa5c2SJ]Mm5o2gIQMY6Z3\\IfLd6d3RI\\Ln6j3lHVLS7Q4gHoKY7T4dHlK\\7m3kHSLU7g3RIXLm6h3TIXLl6h3TIXLl6g3UIYLk6f3VIZLi6g3WIYLi6f3XIZLh6f3XIZLh6e3YI[Lf6e3[I[Le6e3ZI\\Le6d3\\I\\Ld6d3\\I\\Lc6d3^I\\La6e3_IZLa6f3`IZL`6e3`I\\L_6e3aI[L^6e3cI[L]6e3cI[L\\6e3eI[LZ6f3fIZLZ6e3fI\\LY6d3hI\\LW6e3iI[LV6e3kI[LU6e3kI[LT6e3mI[LR6f3mI[LS6d3nI\\LQ6d3PJ\\Lo5e3QJ[Lo5d3RJ\\Lm5e3SJ[Ll5e3TJ\\Ll5c3UJ]Lj5c3WJ]Lh5c3YJ]Lf5c3[J]Lc5d3^J\\L`5e3aJZL^5h3aJYL]5h3dJXLY5j3hJVLV5k3kJULS5l3nJTLP5n3PKRLn4o3RKRLl4P4TKPLi4R4XKnKf4T4ZKlKd4U4]KkKa4W4_KiK_4X4aKiK]4Y4dKfKY4\\4hKdKV4^4jKbKT4_4mKaKQ4a4oK_Ko3b4RL^Kk3e4VLZKh3g4YLXKf3j4ZLVKd3k4]LUKa3m4_LSK_3n4bLRK[3Q5fLnJX3S5iLmJU3U5kLkJS3V5nLjJP3W5QMiJm2W5UMiJi2X5YMgJe2Z5\\MfJc2Z5^MfJ`2[5aMeJ]2[5eMeJZ2[5gMeJW2\\5kMbJT2_5mMaJR2_5oMaJo1`5RN`Jl1`5VN`Jh1a5YN_Jf1a5\\N^Jb1c5_N]J_1d5bN\\J]1d5dN\\JZ1e5gN[JW1f5jNZJT1g5nNXJQ1i5oNWJo0j5ROVJl0k5UOVJi0j5XOVJf0l5ZOTJd0m5^OSJ`0m5ASJ=n5DRJ:P6FQJ7P6JPJ5P6LPJ2Q60oIMS63mILS65mIIT68mIET6<lIBV6?jI_OV6b0lIZOR6j0oISOo5Q1SJlNi5Y1YJcNd5b1]J[N`5k1aJQN\\5T2eJjMX5Z2jJbMW5_2jJ^MW5c2kJZMU5g2mJUMT5l2mJQMT5P3nJmLR5T3oJiLR5X3PKbLS5_3oJZLU5g3lJTLW5m3kJlKZ5T4gJfK]5[4eJ^K`5b4aJYKb5h4`JQKe5o4]JjJg5W5ZJcJk5]5WJ]Jl5d5h10O100O100O100JYJRGh5b8dJ^G\\5a8fJ]G[5c8a00000O10eLbG4^8KkGMU83SHEm7;WHAh7?ZH@f7`0\\H^Od7a0^H^Ob7b0_H]O`7d0aH[O_7d0cH[O]7e0dHnNoNZN]8g2fHnNPOWN\\8j2eHmNSOVNX8m2gHjNf7U1\\HiNe7W1\\HgNe7Y1\\HeNe7Z1]HdNd7\\1]HbNe7\\1]HbNd7^1^H_Nc7a1`HVNf7i1^HSNc7m1`HoMa7Q2bHkM_7T2eHgM^7X2eHdM\\7[2iH`MX7`2kH\\MV7d2c2O1O1M3M4K4M4L4L4L4L4K4M4L4L4L4L3L5L4L4L4L3L5LWh_3"}, "label": "a man in a white shirt with a glass of drink"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mans arm hanging\"."}], "task": "refering", "answer_template": "The \"mans arm hanging\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229, 251, 252, 274, 275], "bbox": [0.88703125, 0.08181250000000001, 1.0, 0.6828749999999999], "iscrowd": 0, "area": 14872.08199999999, "rle": {"size": [480, 640], "counts": "W]Z84f>6K6I6K5O2^Oa0iBgNo;Q2iCYNR<^2J6K6I7J6I8I6I8I6G9J7J5J7eFiJc8l5M3M3O2M2O1O1O1N2O1O1O2M2O1O0O10000WI\\IV5d6_JhI_5Y6VJRJj5n5kI^JT6b5aIiJ_6m6N101O001N101O001N101O0O2O00001N10001OO0100000O01000M2N3NkN"}, "label": "mans arm hanging"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white shirt and black vest\"."}], "task": "refering", "answer_template": "The \"a man wearing a white shirt and black vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229, 251, 252, 274, 275], "bbox": [0.88703125, 0.08181250000000001, 1.0, 0.6828749999999999], "iscrowd": 0, "area": 14872.08199999999, "rle": {"size": [480, 640], "counts": "W]Z84f>6K6I6K5O2^Oa0iBgNo;Q2iCYNR<^2J6K6I7J6I8I6I8I6G9J7J5J7eFiJc8l5M3M3O2M2O1O1O1N2O1O1O2M2O1O0O10000WI\\IV5d6_JhI_5Y6VJRJj5n5kI^JT6b5aIiJ_6m6N101O001N101O001N101O0O2O00001N10001OO0100000O01000M2N3NkN"}, "label": "a man wearing a white shirt and black vest"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the blue shirt sitting next to the man with the white shirt\"."}], "task": "refering", "answer_template": "The \"the man with the blue shirt sitting next to the man with the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 267, 268, 269, 270, 271, 272, 273, 283, 284, 285, 286, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 335], "bbox": [0.26360937500000003, 0.10808333333333334, 0.9149687500000001, 0.8337083333333334], "iscrowd": 0, "area": 64568.33850000003, "rle": {"size": [480, 640], "counts": "\\`_2e0Z>3M2M3N2N2N2M3N2O1O1N2O001N2O1O0O2O1O1N101O1N2O1O001N2O001N2O001N2O001N2O001O0O2O1O2M3N1O2M3N1O2M01O10O01O1O10O1I7L4L4L4M3L4N101O001O1O1O100O2N1O1O1O1O1O2N1O1O1O1O2N1O1O100O2N1O1O1O1O2N1O1O1O1O2O0O100O1O2O0O100O101N1O100O100O2OO001O0O2N1N3N1N3N2M2N3N1N3N2N1N3N1O2M3M2O2M2N3M3M2N3N2N1O2N1O2N1O2M2N2M4M2N3O01O010O10O01O100O010O1O10O0100O001N2O1O0\\GTMU5m2jJUMS5m2lJUMR5l2mJWMP5i2PKYMm4j2QKXMm4i2RKYMk4i2TKXMh4l2WKVMc4o2\\KRM_4S3`KoL[4U3cKmLX4Y3fKhLU4]3kKdLP4`3oKaLl3d3TL]Lf3h3YLYLb3l3^LUL[3Q4dLPLW3V4hLkKR3Z4mLgKn2^4RMcKg2c4XM^Kc2g4]MZKa2g4^MZK`2h4`MYK^2i4`MXK_2i4aMXK]2i4bMXK]2i4cMXK[2i4dMXK[2i4eMXKY2i4fMXKY2j4fMWKX2j4gMWKX2j4gMWKX2j4gMVKY2k4fMVKY2k4eMWKZ2j4eMVK[2l4cMUK]2k4bMVK]2k4aMVK`2j4_MWK`2j4_MWKa2i4^MWKc2i4[MYKd2i4ZMXKf2h4YMYKg2g4XMZKg2g4WMZKj2g4TMZKk2i4RMXKn2k4nLVKR3l4jLUKV3n4gLSKY3o4WL_Ki3c4oKcKP4`4gKfKZ4\\4^KjKa4Z7100O00100O010O010O1O100O100O00100O100O100000O1000000000001O0O100000000000001N1000000000000001O01O01O0O2O0O2N100O2N101N1O101N1O2O0O1O2O0O2N100O2N101N1O101N1O2O0O1O2O0[Og0N2N2N2N1O2PJcJY2_5bMfJ\\2]5_MgJ_2^5YMgJe2]5TMhJj2]5jLlJT3X5YLYKe3S8N3M2N2N2N2N4L5K5Kh1XN1O00001O001O001O00001O001O010O1O1O010O1O1O010O1O0010O0001O010O000010O0001O00001O00001O00001O00001O00001O00001O001O001O1O1O001O1O001N2M3N1O2N2N1O2N2M3N1O2N2N1Ofoh0"}, "label": "the man with the blue shirt sitting next to the man with the white shirt"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in grey shirt is smiling and sitting with his friend\"."}], "task": "refering", "answer_template": "The \"a man in grey shirt is smiling and sitting with his friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 267, 268, 269, 270, 271, 272, 273, 283, 284, 285, 286, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 335], "bbox": [0.26360937500000003, 0.10808333333333334, 0.9149687500000001, 0.8337083333333334], "iscrowd": 0, "area": 64568.33850000003, "rle": {"size": [480, 640], "counts": "\\`_2e0Z>3M2M3N2N2N2M3N2O1O1N2O001N2O1O0O2O1O1N101O1N2O1O001N2O001N2O001N2O001N2O001O0O2O1O2M3N1O2M3N1O2M01O10O01O1O10O1I7L4L4L4M3L4N101O001O1O1O100O2N1O1O1O1O1O2N1O1O1O1O2N1O1O100O2N1O1O1O1O2N1O1O1O1O2O0O100O1O2O0O100O101N1O100O100O2OO001O0O2N1N3N1N3N2M2N3N1N3N2N1N3N1O2M3M2O2M2N3M3M2N3N2N1O2N1O2N1O2M2N2M4M2N3O01O010O10O01O100O010O1O10O0100O001N2O1O0\\GTMU5m2jJUMS5m2lJUMR5l2mJWMP5i2PKYMm4j2QKXMm4i2RKYMk4i2TKXMh4l2WKVMc4o2\\KRM_4S3`KoL[4U3cKmLX4Y3fKhLU4]3kKdLP4`3oKaLl3d3TL]Lf3h3YLYLb3l3^LUL[3Q4dLPLW3V4hLkKR3Z4mLgKn2^4RMcKg2c4XM^Kc2g4]MZKa2g4^MZK`2h4`MYK^2i4`MXK_2i4aMXK]2i4bMXK]2i4cMXK[2i4dMXK[2i4eMXKY2i4fMXKY2j4fMWKX2j4gMWKX2j4gMWKX2j4gMVKY2k4fMVKY2k4eMWKZ2j4eMVK[2l4cMUK]2k4bMVK]2k4aMVK`2j4_MWK`2j4_MWKa2i4^MWKc2i4[MYKd2i4ZMXKf2h4YMYKg2g4XMZKg2g4WMZKj2g4TMZKk2i4RMXKn2k4nLVKR3l4jLUKV3n4gLSKY3o4WL_Ki3c4oKcKP4`4gKfKZ4\\4^KjKa4Z7100O00100O010O010O1O100O100O00100O100O100000O1000000000001O0O100000000000001N1000000000000001O01O01O0O2O0O2N100O2N101N1O101N1O2O0O1O2O0O2N100O2N101N1O101N1O2O0O1O2O0[Og0N2N2N2N1O2PJcJY2_5bMfJ\\2]5_MgJ_2^5YMgJe2]5TMhJj2]5jLlJT3X5YLYKe3S8N3M2N2N2N2N4L5K5Kh1XN1O00001O001O001O00001O001O010O1O1O010O1O1O010O1O0010O0001O010O000010O0001O00001O00001O00001O00001O00001O00001O001O001O1O1O001O1O001N2M3N1O2N2N1O2N2M3N1O2N2N1Ofoh0"}, "label": "a man in grey shirt is smiling and sitting with his friend"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"placemat that ' s white and blue\"."}], "task": "refering", "answer_template": "The \"placemat that ' s white and blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 321, 344, 381, 382, 383, 384, 385, 386], "bbox": [0.5780937500000001, 0.7370833333333333, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 2793.7727999999997, "rle": {"size": [480, 640], "counts": "fl]52j>4M30000000001N10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000001O0000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000000000000000O1000001O01O5K4L5K4L\\k]11bTbN2M3M2N3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3O0;F9G:F:AfC"}, "label": "placemat that ' s white and blue"}]}
{"id": 82228, "image": "COCO_train2014_000000082228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the utensil right besides the open butter packet\"."}], "task": "refering", "answer_template": "The \"the utensil right besides the open butter packet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 321, 344, 381, 382, 383, 384, 385, 386], "bbox": [0.5780937500000001, 0.7370833333333333, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 2793.7727999999997, "rle": {"size": [480, 640], "counts": "fl]52j>4M30000000001N10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000001O0000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000000000000000O1000001O01O5K4L5K4L\\k]11bTbN2M3M2N3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3O0;F9G:F:AfC"}, "label": "the utensil right besides the open butter packet"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair closest to the donuts\"."}], "task": "refering", "answer_template": "The \"a chair closest to the donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 71, 89, 106, 107, 123, 124, 125, 141, 142, 143, 158, 159, 160, 161, 177, 178, 179, 195, 196, 197, 214, 215, 233], "bbox": [0.81176, 0.11207877461706783, 0.99988, 0.7966301969365426], "iscrowd": 0, "area": 13439.53575, "rle": {"size": [457, 500], "counts": "o]e51V>4M2N2N2M3N3IA_Ba0_=6M3M3M4M2M3M3N3L3M3M3N2M4L3N2M3M3M4M2M3M3N3L3M3N2M3M4L3N2M3M3N3L3M3M3N3L3M3N2M3M4M2M3M3M4M2M3M3N2M4L3M3N2M3M4M2M3M3M4L3M3M3M3M4L3M3M3M3M4]Ob0K5L4L5L3L4L4M3L5K4M3L4L4L5K4L4M3ZN"}, "label": "a chair closest to the donuts"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair nearest to some donuts\"."}], "task": "refering", "answer_template": "The \"a chair nearest to some donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 71, 89, 106, 107, 123, 124, 125, 141, 142, 143, 158, 159, 160, 161, 177, 178, 179, 195, 196, 197, 214, 215, 233], "bbox": [0.81176, 0.11207877461706783, 0.99988, 0.7966301969365426], "iscrowd": 0, "area": 13439.53575, "rle": {"size": [457, 500], "counts": "o]e51V>4M2N2N2M3N3IA_Ba0_=6M3M3M4M2M3M3N3L3M3M3N2M4L3N2M3M3M4M2M3M3N3L3M3N2M3M4L3N2M3M3N3L3M3M3N3L3M3N2M3M4M2M3M3M4M2M3M3N2M4L3M3N2M3M4M2M3M3M4L3M3M3M3M4L3M3M3M3M4]Ob0K5L4L5L3L4L4M3L5K4M3L4L4L5K4L4M3ZN"}, "label": "a chair nearest to some donuts"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chair in the left side of the image\"."}], "task": "refering", "answer_template": "The \"wooden chair in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 180, 181, 182, 198, 199, 200, 216, 217, 218, 219, 234, 235, 236, 237, 252, 253, 254, 255, 270, 271, 272, 273, 274], "bbox": [0.0040999999999999995, 0.5973741794310722, 0.22798, 0.9838949671772429], "iscrowd": 0, "area": 14371.2624, "rle": {"size": [457, 500], "counts": "SU1X1Q=`2`Mi1WN00000000000000000000O10000000000000000000000000000000000000000000000000000000000000001O2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2NUk[5"}, "label": "wooden chair in the left side of the image"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the seat of the wooden chair on the left\"."}], "task": "refering", "answer_template": "The \"the seat of the wooden chair on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 180, 181, 182, 198, 199, 200, 216, 217, 218, 219, 234, 235, 236, 237, 252, 253, 254, 255, 270, 271, 272, 273, 274], "bbox": [0.0040999999999999995, 0.5973741794310722, 0.22798, 0.9838949671772429], "iscrowd": 0, "area": 14371.2624, "rle": {"size": [457, 500], "counts": "SU1X1Q=`2`Mi1WN00000000000000000000O10000000000000000000000000000000000000000000000000000000000000001O2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2NUk[5"}, "label": "the seat of the wooden chair on the left"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a round white plate\"."}], "task": "refering", "answer_template": "The \"a round white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 207, 208, 211, 212, 225, 226, 227, 228, 229, 230], "bbox": [0.4849, 0.6032822757111597, 0.83314, 0.8033916849015317], "iscrowd": 0, "area": 2698.1650999999983, "rle": {"size": [457, 500], "counts": "RY\\3<k=6K5K5J6K5J3N3LDXC\\Oh<a0\\C_Oe<=^CCb<;`CDb<8aCH_<5dCJ^<2gCLY<1kCNV<OkC1V<LlC5S=010000000O100O00100O100O1O1O010O00001O010O001O00001O001O00001O001O001O000010O100O2N1O1O2O0O1O001O1O0010O000001O0001O01O0000001O01O0001O000000000000000000000001O0O1000000000000O100000000O1000000O100O1O1O1O1N2N2N1O2N1100O10000O100O100O100O100O100O100O100O100O10000O1O001O10O01O001O001O000000001N101O001N101O0O2O00_UU1"}, "label": "a round white plate"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plate holding donuts\"."}], "task": "refering", "answer_template": "The \"plate holding donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 207, 208, 211, 212, 225, 226, 227, 228, 229, 230], "bbox": [0.4849, 0.6032822757111597, 0.83314, 0.8033916849015317], "iscrowd": 0, "area": 2698.1650999999983, "rle": {"size": [457, 500], "counts": "RY\\3<k=6K5K5J6K5J3N3LDXC\\Oh<a0\\C_Oe<=^CCb<;`CDb<8aCH_<5dCJ^<2gCLY<1kCNV<OkC1V<LlC5S=010000000O100O00100O100O1O1O010O00001O010O001O00001O001O00001O001O001O000010O100O2N1O1O2O0O1O001O1O0010O000001O0001O01O0000001O01O0001O000000000000000000000001O0O1000000000000O100000000O1000000O100O1O1O1O1N2N2N1O2N1100O10000O100O100O100O100O100O100O100O100O10000O1O001O10O01O001O001O000000001N101O001N101O0O2O00_UU1"}, "label": "plate holding donuts"}]}
{"id": 360759, "image": "COCO_train2014_000000360759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chair in upper right corner of wooden table\"."}], "task": "refering", "answer_template": "The \"wooden chair in upper right corner of wooden table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 85, 86, 87, 88, 102, 103, 104, 105, 120, 121, 122, 123, 139, 140, 141, 158], "bbox": [0.55332, 0.009234135667396061, 0.96318, 0.5563238512035011], "iscrowd": 0, "area": 26668.0406, "rle": {"size": [457, 500], "counts": "\\fk33R>4L4M4K4M3L4L4M3L5L3L4L4M3L4M3L5K4M3L4M3M301O0000000000001O0001O000001O000000001O001O001O001O001O001O001O001O001O001O001O001O001O1O010TGPM]6Q3aIQM^6P3`IRM_6n2`ITM_6m2^IVMa6k2]IWMb6j2\\IXMc6h2\\I[Mb6f2\\I\\Mc6e2[I]Md6d2YI_Mf6b2XI`Mg6`2XIbMg6_2WIcMh6^2VIdMi6]2UIfMi6Z2VIhMi6Y2TIjMk6W2SIkMl6V2RIlMm6X2nHjMQ7[2iHgMV7^2dHeMZ7`2`HbM_7d2YH_Mf7f2TH\\Mk7b4O1O1O1O2N1O1O1O100O1O1O2N1O1O1O1O1O1O2N1O1O1O2N1O2N1O2N2N1O2N1O2N1O2NO1N2M3N2N2N2N2N2N2N3M2M3M3M3L4M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3M3L4M3M3M3M3M3M3M3L4M3M3M3M3M4L3M3M3L4M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3L4I7J6I7IPQ8"}, "label": "wooden chair in upper right corner of wooden table"}]}
{"id": 188727, "image": "COCO_train2014_000000188727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra sitting with his friend in the second place\"."}], "task": "refering", "answer_template": "The \"a zebra sitting with his friend in the second place\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 109, 110, 111, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.44018750000000006, 0.22291666666666668, 0.882046875, 0.7601666666666667], "iscrowd": 0, "area": 24310.338349999998, "rle": {"size": [480, 640], "counts": "Q`T49b>6N2M4M2N2M3N2M4M2N2M3N2N3L3N2N2M2O2N1O2O1N1O2O0O2N1O2O0O2N101N101N101N1O2O0O2O0O1O2O0O2O0O100O1O100O100O2N100O100O10000O100O10001N10000O10000O10O010000O10000O100O100O1O100O00100O1O1O100O100O100O100O100O010O100O100O100O100O100O100O001O1O100O1O1O1O100O1O1O100O1O1O1O10O01O1O1O10O01O1O1O1O001N2O1O1O1O001O1O1O1O001O1O1O1O00eMmEKR:4TFIj96[FGd98bFD]9;iFAV9>oF@o8?VG^Oi8b0[G[Od8d0aGYO^8g0fGWOX8h0mGUOR8k0SHQOl7n0bHfN]7Z1eHfNY7Y1iHgNV7Y1kHgNT7X1oHhNo6X1RIhNm6X1TIhNk6W1WIiNh6W1ZIiNd6V1^IjN`6W1aIiN^6V1eIiNZ6W1gIjNW6U1kIkNS6V1oIiNP6V1RJkNl5T1VJlNi5S1ZJlNd5S1_JlNa5S1aJmN^5R1dJnN[5Q1hJnNW5Q1kJoNS5P1PKPOo4o0TKPOk4o0WKQOh4n0ZKQOe4n0_KQO`4n0bKRO]4m0eKSOZ4l0hKTOV4k0mKUOR4j0PLVOo3i0SLVOm3i0ULWOi3h0ZLXOe3g0^LXOa3g0aLYO^3f0dLZOZ3e0iL[OV3d0lL[OT3d0nL\\OQ3c0QM\\On2c0UM\\Ok2c0WM\\Oh2f0XMYOh2g0YMXOg2i0ZMTOf2m0[MROe2o0[MPOe2Q1[MlNf2U1[MfNi2[1Q5<10O100O2O001N101000O100O010O1000O01O1O1O1N101O1O1O1O1N5L4L4L4M1N1O100O1O2O0O1O100O2O000O10001O0O100N3N1O1N2O1N3JjoR1"}, "label": "a zebra sitting with his friend in the second place"}]}
{"id": 188727, "image": "COCO_train2014_000000188727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffs are relaxing at the zoo\"."}], "task": "refering", "answer_template": "The \"two giraffs are relaxing at the zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 109, 110, 111, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.44018750000000006, 0.22291666666666668, 0.882046875, 0.7601666666666667], "iscrowd": 0, "area": 24310.338349999998, "rle": {"size": [480, 640], "counts": "Q`T49b>6N2M4M2N2M3N2M4M2N2M3N2N3L3N2N2M2O2N1O2O1N1O2O0O2N1O2O0O2N101N101N101N1O2O0O2O0O1O2O0O2O0O100O1O100O100O2N100O100O10000O100O10001N10000O10000O10O010000O10000O100O100O1O100O00100O1O1O100O100O100O100O100O010O100O100O100O100O100O100O001O1O100O1O1O1O100O1O1O100O1O1O1O10O01O1O1O10O01O1O1O1O001N2O1O1O1O001O1O1O1O001O1O1O1O00eMmEKR:4TFIj96[FGd98bFD]9;iFAV9>oF@o8?VG^Oi8b0[G[Od8d0aGYO^8g0fGWOX8h0mGUOR8k0SHQOl7n0bHfN]7Z1eHfNY7Y1iHgNV7Y1kHgNT7X1oHhNo6X1RIhNm6X1TIhNk6W1WIiNh6W1ZIiNd6V1^IjN`6W1aIiN^6V1eIiNZ6W1gIjNW6U1kIkNS6V1oIiNP6V1RJkNl5T1VJlNi5S1ZJlNd5S1_JlNa5S1aJmN^5R1dJnN[5Q1hJnNW5Q1kJoNS5P1PKPOo4o0TKPOk4o0WKQOh4n0ZKQOe4n0_KQO`4n0bKRO]4m0eKSOZ4l0hKTOV4k0mKUOR4j0PLVOo3i0SLVOm3i0ULWOi3h0ZLXOe3g0^LXOa3g0aLYO^3f0dLZOZ3e0iL[OV3d0lL[OT3d0nL\\OQ3c0QM\\On2c0UM\\Ok2c0WM\\Oh2f0XMYOh2g0YMXOg2i0ZMTOf2m0[MROe2o0[MPOe2Q1[MlNf2U1[MfNi2[1Q5<10O100O2O001N101000O100O010O1000O01O1O1O1N101O1O1O1O1N5L4L4L4M1N1O100O1O2O0O1O100O2O000O10001O0O100N3N1O1N2O1N3JjoR1"}, "label": "two giraffs are relaxing at the zoo"}]}
{"id": 188727, "image": "COCO_train2014_000000188727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shorter giraffe laying on grass\"."}], "task": "refering", "answer_template": "The \"a shorter giraffe laying on grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 127, 128, 129, 149, 150, 151, 152, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.14737499999999998, 0.28245833333333337, 0.6478906249999999, 0.7925833333333334], "iscrowd": 0, "area": 27253.383399999995, "rle": {"size": [480, 640], "counts": "a]\\11n>1O1O1O1O1O2N1N2O1O1O1O1O2RO7bBJ]=m0O1O1N2O2N1O1O1O1O1O2M2O1O1O1O1N3N1O1O1NG[C[Nd<f1\\CZNc<g1]CZN`<h1`CXN_<i1bCVN]<k1cCUN[<m1eCSNZ<n1fCSNX<n1hCRNX<n1hCRNX<n1iCQNV<o1kCQNU<o1kCQNU<o1kCRNS<o1mCQNS<o1mCQNS<o1nCPNQ<Q2oCoMQ<Q2oCoMQ<Q2oCPNo;Q2QDoMo;Q2RDnMn;R2RDnMm;R2TDnMl;R2TDnMl;R2TDoMj;R2VDnMj;R2WDmMi;S2WDmMh;T2XDlMh;T2XDjMj;W2UDeMn;\\2;001N10001N10000O2O000O101O0O10000O2O000O101O0O10000O2N1O1O1O1O2N1O1O1O2O0O1O1O1O2N1O1O1O1O1O100O2N1O1O1O1O1O1O100O10001N100O10000O100O10000O101N10000O10000O100O10001N100O10000O100O10000O2O0O10000O100O10000O101N100O100O1O1O1O10]MYFMf92]FMb93`FL_94dFKZ95hFIY96iFIV97mFGR98QGGn89TG`NTO;g9U1XGXNWOc0a9T1ZGQN\\Ok0Y9T1UHjNk7V1VHjNi7U1ZHjNe7V1\\HjNd7U1]HkNb7U1`HjN_7V1bHjN]7V1dHiN]7U1fHjNY7V1hHjNW7V1jHjNU7V1mHiNR7W1oHiNQ7V1PIjNo6U1TIiNl6W1UIiNj6W1WIiNh6W1ZIhNf6W1[IiNd6W1]IiNb6W1`IhN_6X1bIgN^6Y1cIgN]6W1fIhNY6X1hIhNW6X1jIhNU6X1mIgNR6Y1oIgNQ6X1PJgNP6Y1RJfNm5Z1TJfNk5Z1VJfNj5Y1XJfNg5Z1ZJfNe5Z1\\JeNd5[1^JdNb5[1_JeN`5[1aJeN^5[1S401O00001O0000001O0000001O00001O0000001O0000001O00001O1O001O10O01OCcNYC\\1h<fNVCY1k<jNQCV1Q=lNlBS1U=oNiBP1X=;2N1O1O2N1O1O2N1O1O1O11N101N101O0O1C>O001O001O00001O001O001O1O20O01O10O001N1O4L6J5K3N3fNaBk0b=QOaBm0l=M3M2N3M4L6J:G[XY3"}, "label": "a shorter giraffe laying on grass"}]}
{"id": 188727, "image": "COCO_train2014_000000188727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe laying to the left of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe laying to the left of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 127, 128, 129, 149, 150, 151, 152, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.14737499999999998, 0.28245833333333337, 0.6478906249999999, 0.7925833333333334], "iscrowd": 0, "area": 27253.383399999995, "rle": {"size": [480, 640], "counts": "a]\\11n>1O1O1O1O1O2N1N2O1O1O1O1O2RO7bBJ]=m0O1O1N2O2N1O1O1O1O1O2M2O1O1O1O1N3N1O1O1NG[C[Nd<f1\\CZNc<g1]CZN`<h1`CXN_<i1bCVN]<k1cCUN[<m1eCSNZ<n1fCSNX<n1hCRNX<n1hCRNX<n1iCQNV<o1kCQNU<o1kCQNU<o1kCRNS<o1mCQNS<o1mCQNS<o1nCPNQ<Q2oCoMQ<Q2oCoMQ<Q2oCPNo;Q2QDoMo;Q2RDnMn;R2RDnMm;R2TDnMl;R2TDnMl;R2TDoMj;R2VDnMj;R2WDmMi;S2WDmMh;T2XDlMh;T2XDjMj;W2UDeMn;\\2;001N10001N10000O2O000O101O0O10000O2O000O101O0O10000O2N1O1O1O1O2N1O1O1O2O0O1O1O1O2N1O1O1O1O1O100O2N1O1O1O1O1O1O100O10001N100O10000O100O10000O101N10000O10000O100O10001N100O10000O100O10000O2O0O10000O100O10000O101N100O100O1O1O1O10]MYFMf92]FMb93`FL_94dFKZ95hFIY96iFIV97mFGR98QGGn89TG`NTO;g9U1XGXNWOc0a9T1ZGQN\\Ok0Y9T1UHjNk7V1VHjNi7U1ZHjNe7V1\\HjNd7U1]HkNb7U1`HjN_7V1bHjN]7V1dHiN]7U1fHjNY7V1hHjNW7V1jHjNU7V1mHiNR7W1oHiNQ7V1PIjNo6U1TIiNl6W1UIiNj6W1WIiNh6W1ZIhNf6W1[IiNd6W1]IiNb6W1`IhN_6X1bIgN^6Y1cIgN]6W1fIhNY6X1hIhNW6X1jIhNU6X1mIgNR6Y1oIgNQ6X1PJgNP6Y1RJfNm5Z1TJfNk5Z1VJfNj5Y1XJfNg5Z1ZJfNe5Z1\\JeNd5[1^JdNb5[1_JeN`5[1aJeN^5[1S401O00001O0000001O0000001O00001O0000001O0000001O00001O1O001O10O01OCcNYC\\1h<fNVCY1k<jNQCV1Q=lNlBS1U=oNiBP1X=;2N1O1O2N1O1O2N1O1O1O11N101N101O0O1C>O001O001O00001O001O001O1O20O01O10O001N1O4L6J5K3N3fNaBk0b=QOaBm0l=M3M2N3M4L6J:G[XY3"}, "label": "a giraffe laying to the left of another giraffe"}]}
{"id": 385337, "image": "COCO_train2014_000000385337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player in red cap holds ball in one hand and wears mitt on the other\"."}], "task": "refering", "answer_template": "The \"player in red cap holds ball in one hand and wears mitt on the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 240, 241, 242, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311], "bbox": [0.40685937499999997, 0.33894366197183096, 0.592328125, 0.8895070422535211], "iscrowd": 0, "area": 13045.866350000002, "rle": {"size": [426, 640], "counts": "Pn\\32W=2N2N2O1N3M1000000_JKmM6Q2OkM1U23dM0\\24]MOc23WM1i20SM3m2NnL6R3KjL8V3IeL;[3G`L<`3E[L?e3BWLa0i3@RLe0m3[OPLh0P4XOlKl0T4bNiH2P3_1X4VNoH<e2b1i4_NTKd1k4]NQKg1n4ZNoJh1Q5XNlJl1S5TNkJn1U5RNhJQ2X5PNeJh1e5XNXJi1j5WNTJi1o5VNnIk1T6VNiIj1Y6\\N^Ie1d6aNTI_1n6hNhHY1]6dMXIY13S1g6fMTI^1Km0T7eMPIe1Cf0`7fMkH[4Y7eKfHY4\\7iKbHU4a7kK^HS4e7oKXHo3k7XLnGe3T8e04J6K400O2N1O2N1O13N3L4M3jIZKa3j4YLYKf3j4VLXKk3l4mKWKS4\\5XKfJi4l5bJVJ^5k61O001O00001O001UInI\\6S6_ISJ_6n5\\IXJb6Z6N2N2N2N1O1O001@`0^Ob0^Ob0^Ob0^Oa0_Oa0E<N1O2O0O1O2O0O101N2N2N2M3M3L3N3M3M6J5K5K5K6J5K5K5Hk[\\3"}, "label": "player in red cap holds ball in one hand and wears mitt on the other"}]}
{"id": 385337, "image": "COCO_train2014_000000385337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a red number 48 jersey\"."}], "task": "refering", "answer_template": "The \"a man wearing a red number 48 jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 240, 241, 242, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311], "bbox": [0.40685937499999997, 0.33894366197183096, 0.592328125, 0.8895070422535211], "iscrowd": 0, "area": 13045.866350000002, "rle": {"size": [426, 640], "counts": "Pn\\32W=2N2N2O1N3M1000000_JKmM6Q2OkM1U23dM0\\24]MOc23WM1i20SM3m2NnL6R3KjL8V3IeL;[3G`L<`3E[L?e3BWLa0i3@RLe0m3[OPLh0P4XOlKl0T4bNiH2P3_1X4VNoH<e2b1i4_NTKd1k4]NQKg1n4ZNoJh1Q5XNlJl1S5TNkJn1U5RNhJQ2X5PNeJh1e5XNXJi1j5WNTJi1o5VNnIk1T6VNiIj1Y6\\N^Ie1d6aNTI_1n6hNhHY1]6dMXIY13S1g6fMTI^1Km0T7eMPIe1Cf0`7fMkH[4Y7eKfHY4\\7iKbHU4a7kK^HS4e7oKXHo3k7XLnGe3T8e04J6K400O2N1O2N1O13N3L4M3jIZKa3j4YLYKf3j4VLXKk3l4mKWKS4\\5XKfJi4l5bJVJ^5k61O001O00001O001UInI\\6S6_ISJ_6n5\\IXJb6Z6N2N2N2N1O1O001@`0^Ob0^Ob0^Ob0^Oa0_Oa0E<N1O2O0O1O2O0O101N2N2N2M3M3L3N3M3M6J5K5K5K6J5K5K5Hk[\\3"}, "label": "a man wearing a red number 48 jersey"}]}
{"id": 385337, "image": "COCO_train2014_000000385337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player whose number is 55\"."}], "task": "refering", "answer_template": "The \"a baseball player whose number is 55\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 280, 281], "bbox": [0.16946875, 0.3383568075117371, 0.3365, 0.8588262910798122], "iscrowd": 0, "area": 11017.057, "rle": {"size": [426, 640], "counts": "XX]12U=5K6YLDWJa0e5CVJc0e5AVJd0f5@UJe0g5^OUJg0f5^OUJg0g5]OUJf0h5^OTJd0j5_OSJc0k5AQJa0m5CoI?o5EmI=R6EkI=S6HhIF\\NWOj7V1fIBeNVOd7\\1bI\\OPOVO]7`1eHYN4o00SOY7g1[HaN7d0]8h1]G[N8lN`71nGm25WN<PO]7Z4aHkK\\7U4cHoK[7P4cHULZ7l3dHYLY7f3eH_LX7a3gHdLV7[3jHiLR7X3nHjLo6V3PIlLP7S3kHSMU7a4001O0000HPJVIP6i6SJUIm5k6SJUIm5j6TJVIl5i6UJWIk5h6UJYIk5c6YJ^If5[6aJcI`5]6aJaI`5`6e00CcIlI^6i5nIUJR6f5TJZJl5c5WJ\\Jj5b5YJ\\Jh5b5ZJ^Je5a5]J^Jd5_5U1N1O2N2M3N1O3J6I6[Of0[Oe0K5L4K5E;PN]Em0n:QOUEc0V;SOeDN8k0U;QOmDN2P1o;N1O1M2O0O00010O0010O01O010N100O010O1O010O1ON3H7N2N31N2O0O2N2N2N8Ild`5"}, "label": "a baseball player whose number is 55"}]}
{"id": 385337, "image": "COCO_train2014_000000385337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher\"."}], "task": "refering", "answer_template": "The \"a baseball catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 280, 281], "bbox": [0.16946875, 0.3383568075117371, 0.3365, 0.8588262910798122], "iscrowd": 0, "area": 11017.057, "rle": {"size": [426, 640], "counts": "XX]12U=5K6YLDWJa0e5CVJc0e5AVJd0f5@UJe0g5^OUJg0f5^OUJg0g5]OUJf0h5^OTJd0j5_OSJc0k5AQJa0m5CoI?o5EmI=R6EkI=S6HhIF\\NWOj7V1fIBeNVOd7\\1bI\\OPOVO]7`1eHYN4o00SOY7g1[HaN7d0]8h1]G[N8lN`71nGm25WN<PO]7Z4aHkK\\7U4cHoK[7P4cHULZ7l3dHYLY7f3eH_LX7a3gHdLV7[3jHiLR7X3nHjLo6V3PIlLP7S3kHSMU7a4001O0000HPJVIP6i6SJUIm5k6SJUIm5j6TJVIl5i6UJWIk5h6UJYIk5c6YJ^If5[6aJcI`5]6aJaI`5`6e00CcIlI^6i5nIUJR6f5TJZJl5c5WJ\\Jj5b5YJ\\Jh5b5ZJ^Je5a5]J^Jd5_5U1N1O2N2M3N1O3J6I6[Of0[Oe0K5L4K5E;PN]Em0n:QOUEc0V;SOeDN8k0U;QOmDN2P1o;N1O1M2O0O00010O0010O01O010N100O010O1O010O1ON3H7N2N31N2O0O2N2N2N8Ild`5"}, "label": "a baseball catcher"}]}
{"id": 336185, "image": "COCO_train2014_000000336185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother elephant\"."}], "task": "refering", "answer_template": "The \"a mother elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 13, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 174, 175, 186, 187, 188, 189, 190, 192, 193, 194, 195, 197, 198, 208, 209, 210, 211, 212, 215, 216, 217, 218, 220, 221, 231, 232, 233, 234, 235, 239, 240, 241, 242, 243, 244, 254, 255, 257, 258, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 285, 286, 287, 290, 308, 309, 310], "bbox": [0.047984375, 0.026955503512880563, 0.6521874999999999, 0.8876346604215456], "iscrowd": 0, "area": 87901.04144999999, "rle": {"size": [427, 640], "counts": "me=1P=:E;F:M3N3M2N2O1N3M2N2N2oJ`NUNb1i1jNlMX1Q2SOdMo0Z2]OZMe0d2GPM;n21fL2W39]LJa3b0PLCn3b0kKBS4a0gKDV4?fKFW4<dKIY4:cKJ[49_KL_46]KO_44\\K1b41ZK3c4J^K;`4@cKd0[4WOgKo0U4POkKT1S4oNhKU1n3gK^KW3`0V1Q4dK]KY3=W1U4aK\\KZ3:Z1Z4lNaKX1^4iN^K[1a4h3O1O1O1O1O1O100OiNULPIj3k6^LSIa3j6dLUI[3i6jLUIU3h6QMVIn2g6XMWIg2f6_MXIa2d6eMZIZ2d6kMZIT2c6RN[Im1b6YN\\If1a6`N]I_1a6f20001O00000O2O1O1O001O1O1N101002N3L3N3M3M2M4M2N3M3L3N3M2N3M3L3N3M3M2M4M3M2N3L4M2ZLaG\\2b8_M`Ga2c8YM`Gf2c8VM_Gj2d8PM_GP3d8kL_GT3c8gL`GX3d8cL^G]3X90000O2O00000OWLjL_MV3a2UMTMk2k2aMkL^2U3mM`LR2a3RN[Ln1e3UNXLk1g3ZNULf1k3]NRLb1o3aNoK^1Q4fNkKZ1U4iNhKW1W4mNfKS1Z4QOaKP1_4SO^Km0b4VO[Kj0d4ZOYKf0g4^OUKc0j4@SK`0m4CPK=o4HmJ8S5KjJ5V5MgJ4Y5NeJ2Z51dJO\\52cJN]54aJL_55`JK`57^JIa59^JGb5;[JFe5;ZJEf5=XJCg5?XJAh5a0VJ_Oj5b0UJ^Ok5Q4O100000001N1000000O1000000O100000000O1000000O2O00001N1000001O0O10001O0O10001O0O10001O00N101N2O1N101N2O3L9H9G8G9H9F4M2M2O2M2O2M201O0O2O001O0O2O001dMXHSOi7j0YHUOh7i0ZHWOg7f0[HZOe7d0]H\\Od7a0^H^Oc7`0_H@b7=aHB_7<cHD^79dHF]78eHH\\75fHKZ74gHLZ71hHNY70iH0X7MkH2U7GRI9o6_OXI`0i6YO^Ig0c6ROcIn0]6kNjIU1W6cNPJ]1P6\\NXJc1i5UN^Jl1`800001O00010O001O00001O00010O00001O000010O01O00001O00010O00001N1N3M2N2N3M2G9G:E:N2O2aM^LRKc3l4aLQK`3m4eLoJ]3n4gLoJZ3o4jLnJX3m4nLQKR3f4YMWKi2`4aM]K`2[4jMbKW2U4UNgKm1P4]NmKd1k3fNRL\\1d3oNYLR1_3XO^Lj0]3\\O`Le0_3_O^La0`3CmKo0P4W4O2M2N2O2M2N3N1N3M2O1N100O1011N2N3M2N3M3M4L3M4M2M4L3M4L3M4L3M4M3L3M4L3M4L6J;E;F:E<E:G9G9G:F8H8H8G:G8H8H8H8H8HO1O1O100O1O2O2M2N2O1N2N2N3N1iK^NaLd1`2j0oKXOe0SN[OU4^OiMP1n4kNTKo0h8J7I6J6J6Jdml2"}, "label": "a mother elephant"}]}
{"id": 336185, "image": "COCO_train2014_000000336185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large elephant walking with small elephant\"."}], "task": "refering", "answer_template": "The \"large elephant walking with small elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 13, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 174, 175, 186, 187, 188, 189, 190, 192, 193, 194, 195, 197, 198, 208, 209, 210, 211, 212, 215, 216, 217, 218, 220, 221, 231, 232, 233, 234, 235, 239, 240, 241, 242, 243, 244, 254, 255, 257, 258, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 285, 286, 287, 290, 308, 309, 310], "bbox": [0.047984375, 0.026955503512880563, 0.6521874999999999, 0.8876346604215456], "iscrowd": 0, "area": 87901.04144999999, "rle": {"size": [427, 640], "counts": "me=1P=:E;F:M3N3M2N2O1N3M2N2N2oJ`NUNb1i1jNlMX1Q2SOdMo0Z2]OZMe0d2GPM;n21fL2W39]LJa3b0PLCn3b0kKBS4a0gKDV4?fKFW4<dKIY4:cKJ[49_KL_46]KO_44\\K1b41ZK3c4J^K;`4@cKd0[4WOgKo0U4POkKT1S4oNhKU1n3gK^KW3`0V1Q4dK]KY3=W1U4aK\\KZ3:Z1Z4lNaKX1^4iN^K[1a4h3O1O1O1O1O1O100OiNULPIj3k6^LSIa3j6dLUI[3i6jLUIU3h6QMVIn2g6XMWIg2f6_MXIa2d6eMZIZ2d6kMZIT2c6RN[Im1b6YN\\If1a6`N]I_1a6f20001O00000O2O1O1O001O1O1N101002N3L3N3M3M2M4M2N3M3L3N3M2N3M3L3N3M3M2M4M3M2N3L4M2ZLaG\\2b8_M`Ga2c8YM`Gf2c8VM_Gj2d8PM_GP3d8kL_GT3c8gL`GX3d8cL^G]3X90000O2O00000OWLjL_MV3a2UMTMk2k2aMkL^2U3mM`LR2a3RN[Ln1e3UNXLk1g3ZNULf1k3]NRLb1o3aNoK^1Q4fNkKZ1U4iNhKW1W4mNfKS1Z4QOaKP1_4SO^Km0b4VO[Kj0d4ZOYKf0g4^OUKc0j4@SK`0m4CPK=o4HmJ8S5KjJ5V5MgJ4Y5NeJ2Z51dJO\\52cJN]54aJL_55`JK`57^JIa59^JGb5;[JFe5;ZJEf5=XJCg5?XJAh5a0VJ_Oj5b0UJ^Ok5Q4O100000001N1000000O1000000O100000000O1000000O2O00001N1000001O0O10001O0O10001O0O10001O00N101N2O1N101N2O3L9H9G8G9H9F4M2M2O2M2O2M201O0O2O001O0O2O001dMXHSOi7j0YHUOh7i0ZHWOg7f0[HZOe7d0]H\\Od7a0^H^Oc7`0_H@b7=aHB_7<cHD^79dHF]78eHH\\75fHKZ74gHLZ71hHNY70iH0X7MkH2U7GRI9o6_OXI`0i6YO^Ig0c6ROcIn0]6kNjIU1W6cNPJ]1P6\\NXJc1i5UN^Jl1`800001O00010O001O00001O00010O00001O000010O01O00001O00010O00001N1N3M2N2N3M2G9G:E:N2O2aM^LRKc3l4aLQK`3m4eLoJ]3n4gLoJZ3o4jLnJX3m4nLQKR3f4YMWKi2`4aM]K`2[4jMbKW2U4UNgKm1P4]NmKd1k3fNRL\\1d3oNYLR1_3XO^Lj0]3\\O`Le0_3_O^La0`3CmKo0P4W4O2M2N2O2M2N3N1N3M2O1N100O1011N2N3M2N3M3M4L3M4M2M4L3M4L3M4L3M4M3L3M4L3M4L6J;E;F:E<E:G9G9G:F8H8H8G:G8H8H8H8H8HO1O1O100O1O2O2M2N2O1N2N2N3N1iK^NaLd1`2j0oKXOe0SN[OU4^OiMP1n4kNTKo0h8J7I6J6J6Jdml2"}, "label": "large elephant walking with small elephant"}]}
{"id": 336185, "image": "COCO_train2014_000000336185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant , walking alongside a taller elephant , in the brush\"."}], "task": "refering", "answer_template": "The \"a baby elephant , walking alongside a taller elephant , in the brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 315], "bbox": [0.51609375, 0.4592740046838408, 0.7614375000000001, 0.8875175644028104], "iscrowd": 0, "area": 14132.642699999997, "rle": {"size": [427, 640], "counts": "`YZ42X=2M3N1OOhDL[94gFMW92jF0T9OnF3n81oF1o84mFNP97kFKR9;jFF\\8W1_GlN^8Z1]GgNa8^1ZGeNc8`1WGcNg8c1RGaNk8d1nF_NQ9f1iF[NX9l2O1O1O001O0001N100O1O100O2N100O1n0RO1N2O1O1N20dNcGkMd8n1]GPNl8h1UGVNS9b1oF\\NY9]1iF`N_9X1cFgNe9Q1\\FmNl9k0WFROQ:g0PFWOX:?lE_O]:4jEJ]:IjE5^:_OhE?e;M2Egl75oRHe0XOg0WOi0WOj0WOd0[Oc0]Od0N100001O0L5J5L5J5L5K4K6N12O1N101N2N2O1N101N8I8G8I0O10O001^MiGNW80nGMR81THLl71YHLg72^HKc73bHI^76fHHZ77iHFW79nHCS7;QIBo6=UI@l6>YI_Of6a0]I\\Od6c0_IZOa6f0cIVO]6k0eISO[6l0hIQOX6P1kIlNV6S1mIjNS6W1oIfNR6Y1RJdNm5]1UJ`Nl5_1WJ^Ni5b1[JZNf5f1\\JXNc5h1`JUNb5j1Q3N2OXFXNm7f1nG^NU8a1fGaN^8]1]GgNe8Y1_FAd9Q23M3M3M3M3L3N3M3M3K5VOj0WOYbo1"}, "label": "a baby elephant , walking alongside a taller elephant , in the brush"}]}
{"id": 336185, "image": "COCO_train2014_000000336185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young , small elephant walking to the left of a larger adult\"."}], "task": "refering", "answer_template": "The \"a young , small elephant walking to the left of a larger adult\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 315], "bbox": [0.51609375, 0.4592740046838408, 0.7614375000000001, 0.8875175644028104], "iscrowd": 0, "area": 14132.642699999997, "rle": {"size": [427, 640], "counts": "`YZ42X=2M3N1OOhDL[94gFMW92jF0T9OnF3n81oF1o84mFNP97kFKR9;jFF\\8W1_GlN^8Z1]GgNa8^1ZGeNc8`1WGcNg8c1RGaNk8d1nF_NQ9f1iF[NX9l2O1O1O001O0001N100O1O100O2N100O1n0RO1N2O1O1N20dNcGkMd8n1]GPNl8h1UGVNS9b1oF\\NY9]1iF`N_9X1cFgNe9Q1\\FmNl9k0WFROQ:g0PFWOX:?lE_O]:4jEJ]:IjE5^:_OhE?e;M2Egl75oRHe0XOg0WOi0WOj0WOd0[Oc0]Od0N100001O0L5J5L5J5L5K4K6N12O1N101N2N2O1N101N8I8G8I0O10O001^MiGNW80nGMR81THLl71YHLg72^HKc73bHI^76fHHZ77iHFW79nHCS7;QIBo6=UI@l6>YI_Of6a0]I\\Od6c0_IZOa6f0cIVO]6k0eISO[6l0hIQOX6P1kIlNV6S1mIjNS6W1oIfNR6Y1RJdNm5]1UJ`Nl5_1WJ^Ni5b1[JZNf5f1\\JXNc5h1`JUNb5j1Q3N2OXFXNm7f1nG^NU8a1fGaN^8]1]GgNe8Y1_FAd9Q23M3M3M3M3L3N3M3M3K5VOj0WOYbo1"}, "label": "a young , small elephant walking to the left of a larger adult"}]}
{"id": 65855, "image": "COCO_train2014_000000065855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant with long leaves on left\"."}], "task": "refering", "answer_template": "The \"plant with long leaves on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 143, 144, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309], "bbox": [0.195015625, 0.36108235294117647, 0.51971875, 0.9076000000000001], "iscrowd": 0, "area": 24384.707149999995, "rle": {"size": [425, 640], "counts": "YSd11X=1O0O2O1QELD1W74RINA2^71nHO@4c7MjH2WOHUO9l8MeH4YOFVO:m8KbH8YODWO9P9K]H:[OBXO9R9KYH<[OAYO9S9KXH;[OAZO9U9KTH=[O_O\\O9W9KQH>ZO_O]O9Y9JnG?[O^O^O9[9IlGa0YO^O_O9]9HiGb0ZO]O@9_9HfGc0YO]OA8b9HbGd0ZO\\OB9b9H`Ge0ZO[OC8e9H]Ge0XO]OF6g9HYGg0UO^OK4h9FXGi0POAO0k9FTG\\1QOoNl9ERG\\1QOoNo9EnF\\1SOoNQ:EkFn1Y6PNbK3SNm1\\6RN`K1SNl1_6TN]KOSNn1b6TNYKNTNm1e6VNVKMQNP2j6TNTKLmMS2R7RNoJKjMV2Y7PNlJJgMY2_7mMhJKdM\\2e7kMdJ]2^5cM`J]2a5dM]J\\2e5eMWJ]2i5dMUJ\\2m5dMQJ\\2Q6eMlI[2U6fMhI\\2Y6eMdI[2]6fM`I_2^6bM]Id2`6]M\\Ii2b6XMYIn2d6SMYIQ3f6QMTIU3j6nLoHW3DnKP7o0TIY3FkKX7i51O2O0O2N2N1O2N:F5K1O00001O0001O01O00000000001O000000000001O0001ESHXKm7`4b0O100O100O10fGdKg7[4WHhKi7b4jGaKW8h4O100O100O010O100O10000O100O010O100O10000OmJPHW48TLm7k3i01O1O1O1O001O1O1O1O1O1O1O00001O0O1O1O1O2N100O1OcKhGk3W8ULjGl3T8SLoGl3P8SLRHm3m7SLTHm3l7QLWHn3h7RLYHo3e7PL^Ho3a7QL`Ho3_7PLcHP4\\7PLfHo3Y7PLiHQ4V7nKlHQ4S7oKnHQ4Q7nKRIQ4m6oKTIQ4k6nKWIR4h6nKZIR4d6mK^IS4b6lK`IS4_6lKcIT4\\6lKeIT4Z6kKiIU4U6kKlIU4S6kKoIT4P6kKRJU4b7O01OO2N1N3M3M4M2M3M4M2M3M3N3L3M3N2M4YNlE3W:Ke1O1N3N1N2O1N3N1O1N3N1N2O1N3N1O1N2O2M2Ofao3"}, "label": "plant with long leaves on left"}]}
{"id": 65855, "image": "COCO_train2014_000000065855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant in the window\"."}], "task": "refering", "answer_template": "The \"plant in the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 143, 144, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309], "bbox": [0.195015625, 0.36108235294117647, 0.51971875, 0.9076000000000001], "iscrowd": 0, "area": 24384.707149999995, "rle": {"size": [425, 640], "counts": "YSd11X=1O0O2O1QELD1W74RINA2^71nHO@4c7MjH2WOHUO9l8MeH4YOFVO:m8KbH8YODWO9P9K]H:[OBXO9R9KYH<[OAYO9S9KXH;[OAZO9U9KTH=[O_O\\O9W9KQH>ZO_O]O9Y9JnG?[O^O^O9[9IlGa0YO^O_O9]9HiGb0ZO]O@9_9HfGc0YO]OA8b9HbGd0ZO\\OB9b9H`Ge0ZO[OC8e9H]Ge0XO]OF6g9HYGg0UO^OK4h9FXGi0POAO0k9FTG\\1QOoNl9ERG\\1QOoNo9EnF\\1SOoNQ:EkFn1Y6PNbK3SNm1\\6RN`K1SNl1_6TN]KOSNn1b6TNYKNTNm1e6VNVKMQNP2j6TNTKLmMS2R7RNoJKjMV2Y7PNlJJgMY2_7mMhJKdM\\2e7kMdJ]2^5cM`J]2a5dM]J\\2e5eMWJ]2i5dMUJ\\2m5dMQJ\\2Q6eMlI[2U6fMhI\\2Y6eMdI[2]6fM`I_2^6bM]Id2`6]M\\Ii2b6XMYIn2d6SMYIQ3f6QMTIU3j6nLoHW3DnKP7o0TIY3FkKX7i51O2O0O2N2N1O2N:F5K1O00001O0001O01O00000000001O000000000001O0001ESHXKm7`4b0O100O100O10fGdKg7[4WHhKi7b4jGaKW8h4O100O100O010O100O10000O100O010O100O10000OmJPHW48TLm7k3i01O1O1O1O001O1O1O1O1O1O1O00001O0O1O1O1O2N100O1OcKhGk3W8ULjGl3T8SLoGl3P8SLRHm3m7SLTHm3l7QLWHn3h7RLYHo3e7PL^Ho3a7QL`Ho3_7PLcHP4\\7PLfHo3Y7PLiHQ4V7nKlHQ4S7oKnHQ4Q7nKRIQ4m6oKTIQ4k6nKWIR4h6nKZIR4d6mK^IS4b6lK`IS4_6lKcIT4\\6lKeIT4Z6kKiIU4U6kKlIU4S6kKoIT4P6kKRJU4b7O01OO2N1N3M3M4M2M3M4M2M3M3N3L3M3N2M4YNlE3W:Ke1O1N3N1N2O1N3N1O1N3N1N2O1N3N1O1N2O2M2Ofao3"}, "label": "plant in the window"}]}
{"id": 549184, "image": "COCO_train2014_000000549184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beard man sitting along with a man with a cap\"."}], "task": "refering", "answer_template": "The \"a beard man sitting along with a man with a cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 22, 23, 24, 40, 41, 42, 43, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 112, 113, 114, 115, 116], "bbox": [0.16682, 0.0462, 0.47534, 0.49237499999999995], "iscrowd": 0, "area": 16646.769949999994, "rle": {"size": [400, 500], "counts": "ZaP12]<>B4L5K4L4L5L0O1O100O101N1O100O1O100N2O1O1N3N1aEaNk9a1kEiNT:j1O1N2O1N2O1O1N2O2N1N2O1N2O1O1N2kNoMjGR2T8V1M3M4M2O1O100O1O100O1O100O2N10000O10000O01M3M3M2N3O10O10000000000001O001O01O01O01O01O010O1d0[O100O0100O017I`0_O100O0100O010O0100O0O2O001N2O001N101O1N101O0O2O001N2O001N101O1N101O0O2O001N2O001N105K5J7J6J6I6K6J6I7J5K6I7J6J5J7J6J6I6K]UV3"}, "label": "a beard man sitting along with a man with a cap"}]}
{"id": 549184, "image": "COCO_train2014_000000549184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ma in a white shirt is sitting on the couch\"."}], "task": "refering", "answer_template": "The \"a ma in a white shirt is sitting on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 22, 23, 24, 40, 41, 42, 43, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 112, 113, 114, 115, 116], "bbox": [0.16682, 0.0462, 0.47534, 0.49237499999999995], "iscrowd": 0, "area": 16646.769949999994, "rle": {"size": [400, 500], "counts": "ZaP12]<>B4L5K4L4L5L0O1O100O101N1O100O1O100N2O1O1N3N1aEaNk9a1kEiNT:j1O1N2O1N2O1O1N2O2N1N2O1N2O1O1N2kNoMjGR2T8V1M3M4M2O1O100O1O100O1O100O2N10000O10000O01M3M3M2N3O10O10000000000001O001O01O01O01O01O010O1d0[O100O0100O017I`0_O100O0100O010O0100O0O2O001N2O001N101O1N101O0O2O001N2O001N101O1N101O0O2O001N2O001N105K5J7J6J6I6K6J6I7J5K6I7J6J5J7J6J6I6K]UV3"}, "label": "a ma in a white shirt is sitting on the couch"}]}
{"id": 549184, "image": "COCO_train2014_000000549184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black shirt and black cap\"."}], "task": "refering", "answer_template": "The \"a man with a black shirt and black cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 50, 51, 52, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 157, 158], "bbox": [0.55012, 0.08245, 0.9815799999999999, 0.590325], "iscrowd": 0, "area": 17217.680350000006, "rle": {"size": [400, 500], "counts": "[d[32\\<2N2M4M2M3N2M3N3M2M3N2M3N3L3N2N20001O0000001N100000001O000O10001O0000001N100000001O00O001O100O1O100O001O100O1O100O01000O10O1N1O2N1O2N2N1O2N2N1O2N2kESNd9o1[FSNc9o1\\FRNb9o1]FTN`9n1_FSN_9o1_FSN_9n1aFSN]9o1aFSN]9o1bFRN\\9o1cFTNZ9n1eFSNY9n1fFTNX9c2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2N10000000001O00000000001O0001O0O1H8H8H8L4N2O1N2N2O1N2O1ND=ZObLhGP4X840O01O010O010O0100O0100O100O100O11O2O0O1O2N1O101N100O2O0O101N100O2O0O101N100O2mMTG`0l8]OWGb0k8[OXGd0h8ZOZGe0i9VN[Fg1X:O2N1O2M2O2N1O101N101N1O2O0O2O0O2O0O2O0O2O0O2O1N2O1N2O1N2N2O2M5L3L5L4Ka\\3"}, "label": "a man with a black shirt and black cap"}]}
{"id": 549184, "image": "COCO_train2014_000000549184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black t - shirt sitting on couch\"."}], "task": "refering", "answer_template": "The \"man in black t - shirt sitting on couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 50, 51, 52, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 157, 158], "bbox": [0.55012, 0.08245, 0.9815799999999999, 0.590325], "iscrowd": 0, "area": 17217.680350000006, "rle": {"size": [400, 500], "counts": "[d[32\\<2N2M4M2M3N2M3N3M2M3N2M3N3L3N2N20001O0000001N100000001O000O10001O0000001N100000001O00O001O100O1O100O001O100O1O100O01000O10O1N1O2N1O2N2N1O2N2N1O2N2kESNd9o1[FSNc9o1\\FRNb9o1]FTN`9n1_FSN_9o1_FSN_9n1aFSN]9o1aFSN]9o1bFRN\\9o1cFTNZ9n1eFSNY9n1fFTNX9c2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2N10000000001O00000000001O0001O0O1H8H8H8L4N2O1N2N2O1N2O1ND=ZObLhGP4X840O01O010O010O0100O0100O100O100O11O2O0O1O2N1O101N100O2O0O101N100O2O0O101N100O2mMTG`0l8]OWGb0k8[OXGd0h8ZOZGe0i9VN[Fg1X:O2N1O2M2O2N1O101N101N1O2O0O2O0O2O0O2O0O2O0O2O1N2O1N2O1N2N2O2M5L3L5L4Ka\\3"}, "label": "man in black t - shirt sitting on couch"}]}
{"id": 549184, "image": "COCO_train2014_000000549184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wii controller on blue pillow\"."}], "task": "refering", "answer_template": "The \"wii controller on blue pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 152, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245], "bbox": [0.0, 0.559675, 0.8720800000000001, 0.98535], "iscrowd": 0, "area": 48754.39894999999, "rle": {"size": [400, 500], "counts": "R:\\1T;0O2O1O1N2O2N1O1N2O2N0O2O1O1O1N2O1O1N101O1O1N101O1N2O001O0O2O1O0O1000000O10000O101O000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O100000O010000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O0100000O10000O1000000O10000O1000000O10000O1000000O10000O100000O010000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O0100000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O0100000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O01000O1000000O1000000O10000000000O100000000O10000000000O10000O10O1000O10000O10000O1000000O1000000O100000O10O1000000O10O10000O010O100O010O100O010O100O010O1O10O0100O10O0100O00100O10O0100O1O010O100O00100O10O01O100O010O100O00100O10O0100O010O100O100O2O000O101N100O2O0O100O2O0O101N100O2O0O101N100O101N100O101N100O100O2O0O100O2O0O100O101N100O100O2N1O1O101N1O1O1O1O2N100O2N1O2N1O2N2N1N3M2N3M3M7I7IRhh0"}, "label": "wii controller on blue pillow"}]}
{"id": 549184, "image": "COCO_train2014_000000549184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wii remote sitting on the blue chair\"."}], "task": "refering", "answer_template": "The \"wii remote sitting on the blue chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 152, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245], "bbox": [0.0, 0.559675, 0.8720800000000001, 0.98535], "iscrowd": 0, "area": 48754.39894999999, "rle": {"size": [400, 500], "counts": "R:\\1T;0O2O1O1N2O2N1O1N2O2N0O2O1O1O1N2O1O1N101O1O1N101O1N2O001O0O2O1O0O1000000O10000O101O000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O100000O010000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O0100000O10000O1000000O10000O1000000O10000O1000000O10000O100000O010000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O0100000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O0100000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O01000O1000000O1000000O10000000000O100000000O10000000000O10000O10O1000O10000O10000O1000000O1000000O100000O10O1000000O10O10000O010O100O010O100O010O100O010O1O10O0100O10O0100O00100O10O0100O1O010O100O00100O10O01O100O010O100O00100O10O0100O010O100O100O2O000O101N100O2O0O100O2O0O101N100O2O0O101N100O101N100O101N100O100O2O0O100O2O0O100O101N100O100O2N1O1O101N1O1O1O1O2N100O2N1O2N1O2N2N1N3M2N3M3M7I7IRhh0"}, "label": "wii remote sitting on the blue chair"}]}
{"id": 311620, "image": "COCO_train2014_000000311620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car in the center that appears to be white\"."}], "task": "refering", "answer_template": "The \"the car in the center that appears to be white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 174, 175, 176], "bbox": [0.0, 0.231640625, 0.5023868312757201, 0.45465625000000004], "iscrowd": 0, "area": 20053.762599999995, "rle": {"size": [640, 486], "counts": "d41oc0k1UN0000001O01O000000000001O0000000000001O000000000000001O0010O01O00001O001O001O001O00001O001O001O001O01O01O010O010O0010O0010O00102M7I7J6I7J6I6J10O010O00001O00001O010O00001O000010O01O00001O001O01O01O00001O001O01O01O001O000010O0001O001O000010O01O00001O000010O01O00010O01O010O010O0010O01O010O01O00001O01O01O001O00010O00001O0010O0001O00001O0O2O00001O00001O001N10001O00001O0O2N100O2N100O2N101N1O101N1O101N101O00001O001O00001O001O001O00001O001O001O00001O001O0O2O0O2O0O3N4K5L4K5L4K5L4K5L4K5L4Kgof4"}, "label": "the car in the center that appears to be white"}]}
{"id": 311620, "image": "COCO_train2014_000000311620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white quest diagnostics car driving by\"."}], "task": "refering", "answer_template": "The \"white quest diagnostics car driving by\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 174, 175, 176], "bbox": [0.0, 0.231640625, 0.5023868312757201, 0.45465625000000004], "iscrowd": 0, "area": 20053.762599999995, "rle": {"size": [640, 486], "counts": "d41oc0k1UN0000001O01O000000000001O0000000000001O000000000000001O0010O01O00001O001O001O001O00001O001O001O001O01O01O010O010O0010O0010O00102M7I7J6I7J6I6J10O010O00001O00001O010O00001O000010O01O00001O001O01O01O00001O001O01O01O001O000010O0001O001O000010O01O00001O000010O01O00010O01O010O010O0010O01O010O01O00001O01O01O001O00010O00001O0010O0001O00001O0O2O00001O00001O001N10001O00001O0O2N100O2N100O2N101N1O101N1O101N101O00001O001O00001O001O001O00001O001O001O00001O001O0O2O0O2O0O3N4K5L4K5L4K5L4K5L4K5L4Kgof4"}, "label": "white quest diagnostics car driving by"}]}
{"id": 311620, "image": "COCO_train2014_000000311620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suv\"."}], "task": "refering", "answer_template": "The \"a black suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 94, 97, 106, 107, 108, 109, 110, 111, 114, 115, 125, 126, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 184, 185], "bbox": [0.27224279835390947, 0.22246875, 0.9972633744855968, 0.467421875], "iscrowd": 0, "area": 16074.614449999997, "rle": {"size": [640, 486], "counts": "aYc22mc02N1O2N2N1O2N2M2O2N2N1O2N2N1O2N2N2N1O2O1O0O101O0000001O000O2O0000001O00001N10001O0000001O0O101O0000001O000O2O0000001O00001N10001O0000001O0O101O0000001O000O2O0000001O00001N10001O0000001O0OJ7ZOf0M30000O1000O010000O10000O10O10O10000O10000O01000O10000O1000O010000O10000O10O1_X^1UO]haN3L4M3L4M3M3L4M3L4M3M3L4M3V_OdNh>_1RAdNo>_1k@cNV?_1d@cN]?a1\\@bNe?`1U@cNl?`1n_ObNR`0a1h_ObNY`0a1`_ObN``0a1[_OaNf`0\\21O001N101O1O001O001O1O001O0O2O1O001O1O001O1O001N2O1O1O1O1O1O1O1O0011O2N2N1N3N1N3M001O001O00001O01O01O00001O00001O000010O0001O00001O001O00010O00001O00001O00O1O1O2M2O1O1N2O1O1N2O1O1N3N1O1N2O1O1N2O2N2M4N2N1O2N2N2N1O2N2N1O2N2N1O2NIiNg]OU1^b0kN^]OT1fb0mNW]OQ1nb0N4G9[Ol<"}, "label": "a black suv"}]}
{"id": 311620, "image": "COCO_train2014_000000311620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark suv on the far side of the road\"."}], "task": "refering", "answer_template": "The \"dark suv on the far side of the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 94, 97, 106, 107, 108, 109, 110, 111, 114, 115, 125, 126, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 184, 185], "bbox": [0.27224279835390947, 0.22246875, 0.9972633744855968, 0.467421875], "iscrowd": 0, "area": 16074.614449999997, "rle": {"size": [640, 486], "counts": "aYc22mc02N1O2N2N1O2N2M2O2N2N1O2N2N1O2N2N2N1O2O1O0O101O0000001O000O2O0000001O00001N10001O0000001O0O101O0000001O000O2O0000001O00001N10001O0000001O0O101O0000001O000O2O0000001O00001N10001O0000001O0OJ7ZOf0M30000O1000O010000O10000O10O10O10000O10000O01000O10000O1000O010000O10000O10O1_X^1UO]haN3L4M3L4M3M3L4M3L4M3M3L4M3V_OdNh>_1RAdNo>_1k@cNV?_1d@cN]?a1\\@bNe?`1U@cNl?`1n_ObNR`0a1h_ObNY`0a1`_ObN``0a1[_OaNf`0\\21O001N101O1O001O001O1O001O0O2O1O001O1O001O1O001N2O1O1O1O1O1O1O1O0011O2N2N1N3N1N3M001O001O00001O01O01O00001O00001O000010O0001O00001O001O00010O00001O00001O00O1O1O2M2O1O1N2O1O1N2O1O1N3N1O1N2O1O1N2O2N2M4N2N1O2N2N2N1O2N2N1O2N2N1O2NIiNg]OU1^b0kN^]OT1fb0mNW]OQ1nb0N4G9[Ol<"}, "label": "dark suv on the far side of the road"}]}
{"id": 246089, "image": "COCO_train2014_000000246089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of a stove with a pan on it\"."}], "task": "refering", "answer_template": "The \"the top of a stove with a pan on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288, 310], "bbox": [0.185390625, 0.5011249999999999, 0.57978125, 0.7707916666666667], "iscrowd": 0, "area": 22514.130950000006, "rle": {"size": [480, 640], "counts": "iSh14h>5K4N200O100O100O10001N100O100O10000O100O2O0O100O10000O100O101N10000O100O100O100O2O000O100O100O10000O2O0O100O100O10000O101N100O10000O100O100O2E:\\Od0\\Od0B>00000000000001O0000000000000000001O0000000000000000001O00000000000000000010O000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0001O00000000000001O000000000000000000001O0000000000000000001O0000000000000000001O000001O000000000001O0000JQE[Lo:d3TEZLl:e38O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O2N1N2O1O1O1N2O1O1D<C=C=C=Chkm3"}, "label": "the top of a stove with a pan on it"}]}
{"id": 246089, "image": "COCO_train2014_000000246089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front stovetop in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the front stovetop in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288, 310], "bbox": [0.185390625, 0.5011249999999999, 0.57978125, 0.7707916666666667], "iscrowd": 0, "area": 22514.130950000006, "rle": {"size": [480, 640], "counts": "iSh14h>5K4N200O100O100O10001N100O100O10000O100O2O0O100O10000O100O101N10000O100O100O100O2O000O100O100O10000O2O0O100O100O10000O101N100O10000O100O100O2E:\\Od0\\Od0B>00000000000001O0000000000000000001O0000000000000000001O00000000000000000010O000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0001O00000000000001O000000000000000000001O0000000000000000001O0000000000000000001O000001O000000000001O0000JQE[Lo:d3TEZLl:e38O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O2N1N2O1O1O1N2O1O1D<C=C=C=Chkm3"}, "label": "the front stovetop in the right hand picture"}]}
{"id": 246089, "image": "COCO_train2014_000000246089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the oven part of a range stove with a hand towel decorated with a rooster hanging from the handle\"."}], "task": "refering", "answer_template": "The \"the oven part of a range stove with a hand towel decorated with a rooster hanging from the handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.18267187499999998, 0.73325, 0.509375, 0.9838333333333333], "iscrowd": 0, "area": 21963.388500000005, "rle": {"size": [480, 640], "counts": "PVg19g>a0_Oa0_Oa0_Oa0@`0_Oa0_O1O1O000000000001O0000000000000000010O00000000000000001O0000000001O00000001O00000000000000000010O000000000000000001O0000000001O000001O0000000000000000001O0001O00000000000001O00000000000001O0001O00000000000000001O000001O000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O000000000000000VNUlb4"}, "label": "the oven part of a range stove with a hand towel decorated with a rooster hanging from the handle"}]}
{"id": 246089, "image": "COCO_train2014_000000246089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black door of oven\"."}], "task": "refering", "answer_template": "The \"black door of oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.18267187499999998, 0.73325, 0.509375, 0.9838333333333333], "iscrowd": 0, "area": 21963.388500000005, "rle": {"size": [480, 640], "counts": "PVg19g>a0_Oa0_Oa0_Oa0@`0_Oa0_O1O1O000000000001O0000000000000000010O00000000000000001O0000000001O00000001O00000000000000000010O000000000000000001O0000000001O000001O0000000000000000001O0001O00000000000001O00000000000001O0001O00000000000000001O000001O000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O000000000000000VNUlb4"}, "label": "black door of oven"}]}
{"id": 74060, "image": "COCO_train2014_000000074060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult giraffe\"."}], "task": "refering", "answer_template": "The \"the adult giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 192, 193, 208, 209, 210, 211, 212, 216, 217, 231, 232, 234, 235, 239, 240, 254, 255, 257, 258, 259, 278, 280], "bbox": [0.061375, 0.36803278688524593, 0.45989062500000005, 0.8276346604215457], "iscrowd": 0, "area": 18121.022700000005, "rle": {"size": [427, 640], "counts": "ja`02W=4L5K5K5K4K6K2N3M2N2N2M4M2dNgNdF[1\\9nNYFU1g9UOkEn0V:P13H`FdMb8i1kG]NU8k0]H[Ob7f0WH@i7a0QHEo7;fGOZ83TG;n8P21N2O2B_FoLb9h2f0GPOaMdGY2b8hM^GY2_8iM`GY2^8hMbG]2X8eMfGk2i7WMVHm2f7UMWHl2h7VMgGD0V3Y8XMdGE0S3\\8ZMcGD0R3]8]M_GC3P3]8aM]G@4o2`8<QHdLn7o2^1L3M4L3M4L4M2O2M2O2N1O2N1100O100O2O000O100O2O0O100O100O2O0O100O10000O10O01O100O00100O010fERNV9o1cFcNQ9P3nGmKd6V4kHXLT7o40100O1O010O@`0L5J5L5K4_Ob0ZOf0L4N2M2O2N2N2d0\\O6J6J7I6I6K3M4M2M4M2M3N3L3MeMWHWOg7e0SIhNi6V1eI_N[6^1nI[NR6c1WJVNj5f1aJRN_5l1jJmMV5Q2RKjMn4S2V3M4M2N3XOhDXO[;g0kDROX;m0c000O2O001N101N100010O010O01O01O010O010O01O01O010O010O01O10O010O010O0100O0100O100O10000O010O100O10000O10O0100O10000O100O01000001O01O01O01O0010O01O010O0010O01O001[DfNW;Y1_DSO_;_1O1O10<C4LN2N2O1N2N2N3M2N2O1N2N1N2N1N3G9G9J5O1N101N2N2G\\CIf<5801O1O1O00lP`4"}, "label": "the adult giraffe"}]}
{"id": 74060, "image": "COCO_train2014_000000074060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother giraffe lickicking her baby\"."}], "task": "refering", "answer_template": "The \"a mother giraffe lickicking her baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 192, 193, 208, 209, 210, 211, 212, 216, 217, 231, 232, 234, 235, 239, 240, 254, 255, 257, 258, 259, 278, 280], "bbox": [0.061375, 0.36803278688524593, 0.45989062500000005, 0.8276346604215457], "iscrowd": 0, "area": 18121.022700000005, "rle": {"size": [427, 640], "counts": "ja`02W=4L5K5K5K4K6K2N3M2N2N2M4M2dNgNdF[1\\9nNYFU1g9UOkEn0V:P13H`FdMb8i1kG]NU8k0]H[Ob7f0WH@i7a0QHEo7;fGOZ83TG;n8P21N2O2B_FoLb9h2f0GPOaMdGY2b8hM^GY2_8iM`GY2^8hMbG]2X8eMfGk2i7WMVHm2f7UMWHl2h7VMgGD0V3Y8XMdGE0S3\\8ZMcGD0R3]8]M_GC3P3]8aM]G@4o2`8<QHdLn7o2^1L3M4L3M4L4M2O2M2O2N1O2N1100O100O2O000O100O2O0O100O100O2O0O100O10000O10O01O100O00100O010fERNV9o1cFcNQ9P3nGmKd6V4kHXLT7o40100O1O010O@`0L5J5L5K4_Ob0ZOf0L4N2M2O2N2N2d0\\O6J6J7I6I6K3M4M2M4M2M3N3L3MeMWHWOg7e0SIhNi6V1eI_N[6^1nI[NR6c1WJVNj5f1aJRN_5l1jJmMV5Q2RKjMn4S2V3M4M2N3XOhDXO[;g0kDROX;m0c000O2O001N101N100010O010O01O01O010O010O01O01O010O010O01O10O010O010O0100O0100O100O10000O010O100O10000O10O0100O10000O100O01000001O01O01O01O0010O01O010O0010O01O001[DfNW;Y1_DSO_;_1O1O10<C4LN2N2O1N2N2N3M2N2O1N2N1N2N1N3G9G9J5O1N101N2N2G\\CIf<5801O1O1O00lP`4"}, "label": "a mother giraffe lickicking her baby"}]}
{"id": 172370, "image": "COCO_train2014_000000172370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in snow skating behind a lady sitting on the snow\"."}], "task": "refering", "answer_template": "The \"a lady in snow skating behind a lady sitting on the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 105, 126, 127, 128, 148, 149, 150, 151, 172, 173, 174, 196, 197, 219], "bbox": [0.4329375, 0.22702576112412176, 0.586921875, 0.6251288056206089], "iscrowd": 0, "area": 7529.655100000002, "rle": {"size": [427, 640], "counts": "jec32W=3M3M3O0O101O0O101O0O101O00010O01O010O010O00010O010O0010O010O010O00001O0kME[G;b8OVG1i83TGMk86SGJk8:SGFkNAc9l0iFSOLb0FCb9k0hFTOJb0KBa9i0gFXOGT1`9EeF]ODT1d9AeFT1Y9lNeFX1X9iNfF[1X9fNdF^1Z9Z10000O2O000O1000003M6J4L2N1O2N2N1O6K4K01\\KnGn3Q8QLRHn3n7oKVHP4j7nK[Ho3d7nK_HR4]81O1O001M3OO2M3M4L3M3M4L4L3M4L4L4L3L5ZMeEV2_:eMZFb1m:bN[Dl0h;QO[Dk0j;POZDm0i<WOnk]3"}, "label": "a lady in snow skating behind a lady sitting on the snow"}]}
{"id": 172370, "image": "COCO_train2014_000000172370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier who is standing wearing a brown coat\"."}], "task": "refering", "answer_template": "The \"the skier who is standing wearing a brown coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 105, 126, 127, 128, 148, 149, 150, 151, 172, 173, 174, 196, 197, 219], "bbox": [0.4329375, 0.22702576112412176, 0.586921875, 0.6251288056206089], "iscrowd": 0, "area": 7529.655100000002, "rle": {"size": [427, 640], "counts": "jec32W=3M3M3O0O101O0O101O0O101O00010O01O010O010O00010O010O0010O010O010O00001O0kME[G;b8OVG1i83TGMk86SGJk8:SGFkNAc9l0iFSOLb0FCb9k0hFTOJb0KBa9i0gFXOGT1`9EeF]ODT1d9AeFT1Y9lNeFX1X9iNfF[1X9fNdF^1Z9Z10000O2O000O1000003M6J4L2N1O2N2N1O6K4K01\\KnGn3Q8QLRHn3n7oKVHP4j7nK[Ho3d7nK_HR4]81O1O001M3OO2M3M4L3M3M4L4L3M4L4L4L3L5ZMeEV2_:eMZFb1m:bN[Dl0h;QO[Dk0j;POZDm0i<WOnk]3"}, "label": "the skier who is standing wearing a brown coat"}]}
{"id": 254291, "image": "COCO_train2014_000000254291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the left of the other giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe to the left of the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 48, 50, 51, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 164, 165, 166, 167, 168, 169, 170, 184, 185, 189, 190, 191, 192, 193, 194, 208, 209, 215, 216, 217, 231, 232, 254, 255, 256, 278, 279, 301, 302, 303], "bbox": [0.0015, 0.0005620608899297424, 0.4617812499999999, 0.9331381733021077], "iscrowd": 0, "area": 39713.4668, "rle": {"size": [427, 640], "counts": "[=e0g<Y1fNZ1RGWMf4T4]I`La6X5M2O2M2N3M2O2M3M2N3N1N3M2O2M2N3N1O2O1N101O0O2O001O0O2O0010O2O0002O2M2N2N2ON1O1O1kKeJ6[5YOZKc0d4WOhKe0W4ZOoKb0P4^OTL?k3@[L<c3EbL7]3IhL3W3LoL0P30UMLi24]MHb28`MG_29bMG]28eMHY29iMFV29lMGS29nMGQ29PNFo1:TNEk1;VNEi1:YNEg1;[NDc1=^NCa1<aND^1<dNB[1>gNBX1>iNBV1>lN@T1?nNAP1`0QO@n0`0SO_Om0`0VO_Oi0a0XO_Of0a0\\O^Od0b0\\O_Oc0a0^O_Oa0`0A_O>b0C^O<a0F_O9a0G@8`0I_O6a0L_O3a0N_O1`01_OOa01@Ma04]OMb05\\OLd06YOKf0:UOFl0>oNCQ1a0jN_OV1f0eN[O[1h0bNWO_1m0\\NTOc1o0ZNQOg1P1WNQOi1o0VNQOk1P1SNQOm1o0RNQOo1P1oMQOQ2o0nMROR2o0mMPOT2P1kMPOV2Q1hMnNZ2S1dMRN^OYMP3e4aMQNCVMn2j4]MoMISMl2n4ZMnMMQMk2R5VMlM3nLi2n8YMoFi2o8[MmFg2Q9k0N2N2N2O1N2N2N2N2N2O1O1O1O1O1O1O0000001O0000001O000001O10O01O1O100O1O1O010O1O1O100O1O1O010O1O2N2O2M2N2N3M2O2M2N2N3N1N_Oa00001O000O101O0000001O000O101O00UL\\Go2d8PM]GP3c8oL_GP3a8nLbGR3\\8nLfGQ3l8\\LVGc3Z9O10O01O1O100O1O3M3N3L3M3N2M3N3L3N2M3N3L3M100O101N100O100O100O100O1O100O100O100O100O100O100O100O10000O101N100O10000O100O10000O100O10000O100O10000O100O10001N2O1N2O1O1N2O1N3L3N2M3M3N2M3M4M2M3N2M3M3NdU_4"}, "label": "a giraffe to the left of the other giraffe"}]}
{"id": 254291, "image": "COCO_train2014_000000254291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffes leaning over a fence next to a woman in a red shirt and white hat\"."}], "task": "refering", "answer_template": "The \"two giraffes leaning over a fence next to a woman in a red shirt and white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 123, 124, 125, 126, 127, 147, 148, 149, 150, 172, 173, 174, 195, 196, 197, 219, 220, 243], "bbox": [0.097453125, 0.00032786885245901645, 0.61275, 0.6858782201405152], "iscrowd": 0, "area": 29862.786450000007, "rle": {"size": [427, 640], "counts": "Rfj01Z=001O001O001O00001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O1O001O001O1O1O001O1O001O001O1O001O001O1O001O001O1O001O1O1O001O1O1O001O1O1O001O1O1O5K3M3M3M2O1N2N2N2N001O001O001O000000O100O10000HbEbM_:^27O11O1O2N2N2N3M:fERMc9^3M3M2N1O1O1O1O1O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O000000O11O001O1O1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O001O1O001O1N101O1O001O1O0101O0OO2O1N1O2O1N101N2O1OYKeH_3Z7`LhH`3W7_LkHa3T7_LmH`3S7_LoHa3P7_LQIa3n6^LTIb3k6]LVIc3j6\\LWIe3h6ZLZIf3e6YL\\Ih3c6VL`Ij3_6ULcIj3]6ULeIk3Z6TLhIl3W6RLlIn3S6QLoIo3P6PLTJn3k5PLYJ:kNe2k6PM^J6lNh2e6QMbJ3oNi2^6RMgJ1QOT3m5iLYKLPOY3e5iLaKHPO]3]5iLlLT3S3iLSMV3k2hL[MU3d2iLbMU3[6O001O001O10O01O001O0O2O001O0O2O001O0O2O001O0O2N2O1N2N101N2N20O100000000O0100O100O1O010O105J100O0O2O1O001O0O2O0nNeD3\\;JgD5Y;HlD5V;HmD7S;GQE7P;ETE:m:CWE;k:@ZE=l;O0O2N2N1O2O1N1N3L[TW3"}, "label": "two giraffes leaning over a fence next to a woman in a red shirt and white hat"}]}
{"id": 254291, "image": "COCO_train2014_000000254291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffe heads with one of them with its tongue out\"."}], "task": "refering", "answer_template": "The \"two giraffe heads with one of them with its tongue out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 123, 124, 125, 126, 127, 147, 148, 149, 150, 172, 173, 174, 195, 196, 197, 219, 220, 243], "bbox": [0.097453125, 0.00032786885245901645, 0.61275, 0.6858782201405152], "iscrowd": 0, "area": 29862.786450000007, "rle": {"size": [427, 640], "counts": "Rfj01Z=001O001O001O00001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O1O001O001O1O1O001O1O001O001O1O001O001O1O001O001O1O001O1O1O001O1O1O001O1O1O001O1O1O5K3M3M3M2O1N2N2N2N001O001O001O000000O100O10000HbEbM_:^27O11O1O2N2N2N3M:fERMc9^3M3M2N1O1O1O1O1O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O000000O11O001O1O1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O001O1O001O1N101O1O001O1O0101O0OO2O1N1O2O1N101N2O1OYKeH_3Z7`LhH`3W7_LkHa3T7_LmH`3S7_LoHa3P7_LQIa3n6^LTIb3k6]LVIc3j6\\LWIe3h6ZLZIf3e6YL\\Ih3c6VL`Ij3_6ULcIj3]6ULeIk3Z6TLhIl3W6RLlIn3S6QLoIo3P6PLTJn3k5PLYJ:kNe2k6PM^J6lNh2e6QMbJ3oNi2^6RMgJ1QOT3m5iLYKLPOY3e5iLaKHPO]3]5iLlLT3S3iLSMV3k2hL[MU3d2iLbMU3[6O001O001O10O01O001O0O2O001O0O2O001O0O2O001O0O2N2O1N2N101N2N20O100000000O0100O100O1O010O105J100O0O2O1O001O0O2O0nNeD3\\;JgD5Y;HlD5V;HmD7S;GQE7P;ETE:m:CWE;k:@ZE=l;O0O2N2N1O2O1N1N3L[TW3"}, "label": "two giraffe heads with one of them with its tongue out"}]}
{"id": 16725, "image": "COCO_train2014_000000016725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant on a coffee table\"."}], "task": "refering", "answer_template": "The \"a plant on a coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 95, 96, 97, 98, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 153, 166], "bbox": [0.23596, 0.35335999999999995, 0.5662999999999999, 0.7129333333333334], "iscrowd": 0, "area": 11239.4341, "rle": {"size": [375, 500], "counts": "f^[14`;9G8H2N3N1O1O1O1O1O100O1O1O1O001O1H8F:G9F9N300O1000O0100O10000O010O10000O10O10O101O0O2O0O2O001N2O0100O101N10000O20ONnF`MQ9d200O1O1O101^GWMR8j2kGYMT8h2jG[MU8e2iG]MV8d2hG^MW8c2fG^M[8n22N1O1O1O2N1O1O2O0O1O1O2O0O101O0000001O00001O00000O2O0000001O000O2N100O101N2N2O1N2N2O1N2O1N2N2O1O1O100O100O1O10O0100O010O10O010O015K7H7JL4L3M4L3N3L4N1O2M3N1O2N1N3N2N1N3N2N1O2M2O2N2N1N3N1O2NQY_2"}, "label": "a plant on a coffee table"}]}
{"id": 16725, "image": "COCO_train2014_000000016725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pot plant sitting on top of the coffee table\"."}], "task": "refering", "answer_template": "The \"the pot plant sitting on top of the coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 95, 96, 97, 98, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 153, 166], "bbox": [0.23596, 0.35335999999999995, 0.5662999999999999, 0.7129333333333334], "iscrowd": 0, "area": 11239.4341, "rle": {"size": [375, 500], "counts": "f^[14`;9G8H2N3N1O1O1O1O1O100O1O1O1O001O1H8F:G9F9N300O1000O0100O10000O010O10000O10O10O101O0O2O0O2O001N2O0100O101N10000O20ONnF`MQ9d200O1O1O101^GWMR8j2kGYMT8h2jG[MU8e2iG]MV8d2hG^MW8c2fG^M[8n22N1O1O1O2N1O1O2O0O1O1O2O0O101O0000001O00001O00000O2O0000001O000O2N100O101N2N2O1N2N2O1N2O1N2N2O1O1O100O100O1O10O0100O010O10O010O015K7H7JL4L3M4L3N3L4N1O2M3N1O2N1N3N2N1N3N2N1O2M2O2N2N1N3N1O2NQY_2"}, "label": "the pot plant sitting on top of the coffee table"}]}
{"id": 16725, "image": "COCO_train2014_000000016725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch with yellow and red cushions in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the couch with yellow and red cushions in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 97, 98, 111, 112, 113, 114, 115, 116, 129, 130, 131, 132, 147, 148], "bbox": [0.1758, 0.44576, 0.47428000000000003, 0.66104], "iscrowd": 0, "area": 5184.063250000001, "rle": {"size": [375, 500], "counts": "T]P11e;3N2VE0h93SF2k90PF5m9MPF6o9KmE:P:KjE9T:g0N2O0O101N1B^NfFb1X9`0O1O2N1O0O10000O100O10N1O200O1O100O101N100O1N2N2O1N2N2O1N2N2O1N2OQF^Nk9a1UF`Nk9_1TFcNl9\\1UFdNk9\\1TFeNl9Z1TFgNl9X1TFiNl9V1TFkNl9T1UFlNk9S1UFnNl9P1UFPOk9o0UFROk9m0UFSOl9m0TFSOl9l0TFUOl9k0SFVOm9i0TFWOl9i0UFVOk9i0VFVOk9j0UFQOP:o0QFlNS:S1nEnNQ:R1nEoN3Nd9R1ZFQOM1i9n0ZF[Of9e0YF[Oh9d0YF\\Oh9c0WF^Oi9b0WF^Oi9a0XF_Oh9a0WF_Oj9`0WF@i9`0VFAj9?VFYOJOP:g0WFZOJMP:i0UF[OMJn95PF<5EOHl97QF;3GS:NlE91IT:LlE;0IT:LkE5OH26U:MjE5OH34U:NiE70F24W:NgE71H11X:0eE83F01Y:0eE92Fc:O\\E<2Di:;XEEVW:4oSF3M4N1010O10O0100O100O10O1000O1000O010001N101O001N10gTP3"}, "label": "the couch with yellow and red cushions in the right hand picture"}]}
{"id": 16725, "image": "COCO_train2014_000000016725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden couch with a bunch of various pillows on it\"."}], "task": "refering", "answer_template": "The \"wooden couch with a bunch of various pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 97, 98, 111, 112, 113, 114, 115, 116, 129, 130, 131, 132, 147, 148], "bbox": [0.1758, 0.44576, 0.47428000000000003, 0.66104], "iscrowd": 0, "area": 5184.063250000001, "rle": {"size": [375, 500], "counts": "T]P11e;3N2VE0h93SF2k90PF5m9MPF6o9KmE:P:KjE9T:g0N2O0O101N1B^NfFb1X9`0O1O2N1O0O10000O100O10N1O200O1O100O101N100O1N2N2O1N2N2O1N2N2O1N2OQF^Nk9a1UF`Nk9_1TFcNl9\\1UFdNk9\\1TFeNl9Z1TFgNl9X1TFiNl9V1TFkNl9T1UFlNk9S1UFnNl9P1UFPOk9o0UFROk9m0UFSOl9m0TFSOl9l0TFUOl9k0SFVOm9i0TFWOl9i0UFVOk9i0VFVOk9j0UFQOP:o0QFlNS:S1nEnNQ:R1nEoN3Nd9R1ZFQOM1i9n0ZF[Of9e0YF[Oh9d0YF\\Oh9c0WF^Oi9b0WF^Oi9a0XF_Oh9a0WF_Oj9`0WF@i9`0VFAj9?VFYOJOP:g0WFZOJMP:i0UF[OMJn95PF<5EOHl97QF;3GS:NlE91IT:LlE;0IT:LkE5OH26U:MjE5OH34U:NiE70F24W:NgE71H11X:0eE83F01Y:0eE92Fc:O\\E<2Di:;XEEVW:4oSF3M4N1010O10O0100O100O10O1000O1000O010001N101O001N10gTP3"}, "label": "wooden couch with a bunch of various pillows on it"}]}
{"id": 16725, "image": "COCO_train2014_000000016725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tree with fronds near a window\"."}], "task": "refering", "answer_template": "The \"a tree with fronds near a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 86, 87, 88, 89, 105, 106, 107, 123, 124, 125, 141, 142, 143, 161, 178, 179, 197], "bbox": [0.8245399999999999, 0.2702133333333333, 1.0, 0.8473066666666667], "iscrowd": 0, "area": 9413.357850000002, "rle": {"size": [375, 500], "counts": "RPg41e;6J5L4K5K4M3L3N2M3N2M3N2N0O100O1TGjNW7V1gHlNY7T1dHnN\\7S1bHoN^7Q1_HQOa7Q1\\HPOe7P1XHQOi7P1UHQOl7o0SHQOn7o0RHPOn7Q1RHnNo7R1QHmNo7T1PHlNQ8T1oGkNR8U1nGjNS8W1lGiNU8V1jGjNX8U1hGjNZ8U1fGjN\\8V1cGiN_OGi8_1gGkN@Gi8]1gGkNAHi8\\1eGmNAIh8\\1fGjNCJd8_1hGgNEK`8`1jGfNEK_8a1lGcNFMZ8d1nGbNELZ8d1QHbNBKZ8f1SHmNj7V1UHkNk7T1THnNk7S1THmNm7S1QHoNn7R1QHoNn7R1PHPOP8P1oGQOP8P1oGPOQ8R1mGoNS8Q1lGPOS8\\2O2M1WIZLUO8<K[4e3QL[LmNe0<_Og4a3oJ[Lg0Q1\\OUOn4a3mJZLg0b1^4S2jJ\\Lf0b1c4X2YKjMj4T2hJ\\LLb1_5Q2bJ[N_5d1aJ\\N]5g1bJYN\\5j1dJUN\\5S40000000001O0000000001N1000001N6K8Hn2RMPI"}, "label": "a tree with fronds near a window"}]}
{"id": 16725, "image": "COCO_train2014_000000016725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant inside a red vase next to a wooden chair\"."}], "task": "refering", "answer_template": "The \"a plant inside a red vase next to a wooden chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 86, 87, 88, 89, 105, 106, 107, 123, 124, 125, 141, 142, 143, 161, 178, 179, 197], "bbox": [0.8245399999999999, 0.2702133333333333, 1.0, 0.8473066666666667], "iscrowd": 0, "area": 9413.357850000002, "rle": {"size": [375, 500], "counts": "RPg41e;6J5L4K5K4M3L3N2M3N2M3N2N0O100O1TGjNW7V1gHlNY7T1dHnN\\7S1bHoN^7Q1_HQOa7Q1\\HPOe7P1XHQOi7P1UHQOl7o0SHQOn7o0RHPOn7Q1RHnNo7R1QHmNo7T1PHlNQ8T1oGkNR8U1nGjNS8W1lGiNU8V1jGjNX8U1hGjNZ8U1fGjN\\8V1cGiN_OGi8_1gGkN@Gi8]1gGkNAHi8\\1eGmNAIh8\\1fGjNCJd8_1hGgNEK`8`1jGfNEK_8a1lGcNFMZ8d1nGbNELZ8d1QHbNBKZ8f1SHmNj7V1UHkNk7T1THnNk7S1THmNm7S1QHoNn7R1QHoNn7R1PHPOP8P1oGQOP8P1oGPOQ8R1mGoNS8Q1lGPOS8\\2O2M1WIZLUO8<K[4e3QL[LmNe0<_Og4a3oJ[Lg0Q1\\OUOn4a3mJZLg0b1^4S2jJ\\Lf0b1c4X2YKjMj4T2hJ\\LLb1_5Q2bJ[N_5d1aJ\\N]5g1bJYN\\5j1dJUN\\5S40000000001O0000000001N1000001N6K8Hn2RMPI"}, "label": "a plant inside a red vase next to a wooden chair"}]}
{"id": 377174, "image": "COCO_train2014_000000377174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red train in front\"."}], "task": "refering", "answer_template": "The \"the red train in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 434, 435, 436, 450, 451, 452, 453, 454, 473, 474], "bbox": [0.017973856209150325, 0.3280882352941176, 0.8629248366013073, 0.9685457516339869], "iscrowd": 0, "area": 146206.43439999997, "rle": {"size": [612, 612], "counts": "`R75db0;E;E<E:E;E;F:E;E;E;F;D;E;E;F:E;E;E;F;D;E;E;F:E;E;E;F:E<L300000000000001O000O10000000001O00000000001O000000001O0000001O000000001O0000001O00000O101O0000001O000000001O0000001O000000001O0000001O000000001O0000001N100000001O0000001O000000001O00000000000000000000001O000000000000000O100000000000000000000000000000000000000000001O0000000000000000000O10000000000000000000000000000000000000001O00000000000000000000000O1000000000000000000000000000000000001O000000000000000000000000000O100000000000000000000000001O1O001O1O001O1O1O001O1O001O1O001O1O001O1N101O1O001O1O001O1O001O1O001O001O1O001O001O001O1O001O001O1O001O001N101O1O001O001O1O001O001O001O1O001O001OO1000000O1000000O1000000O10O1000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O10O1000O1000000O1000000O1000000O1000000O1000000O101O1M4J6I6K6J6J5K6J5K6I7J5M4N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O9G9G9G9G:F9G9bMgC_Mb<T2cCkMf<g1aCWNh<\\1]CcNm<n0ZCPOo<c0VC\\OS=6TCHU=KPC4Y=^OnB`0\\=ROiBm0`=eNgBY1c?O2N1O2N1O2N101N1O2N1O2N1O2N1O2N101N1O2N1O2N1O000001O0001O00000000000001O0000000000000000001O01O0000000000000001O00000000O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O`fa1"}, "label": "the red train in front"}]}
{"id": 377174, "image": "COCO_train2014_000000377174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white bus train with the number one on the side is sitting next to another train\"."}], "task": "refering", "answer_template": "The \"a red and white bus train with the number one on the side is sitting next to another train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 434, 435, 436, 450, 451, 452, 453, 454, 473, 474], "bbox": [0.017973856209150325, 0.3280882352941176, 0.8629248366013073, 0.9685457516339869], "iscrowd": 0, "area": 146206.43439999997, "rle": {"size": [612, 612], "counts": "`R75db0;E;E<E:E;E;F:E;E;E;F;D;E;E;F:E;E;E;F;D;E;E;F:E;E;E;F:E<L300000000000001O000O10000000001O00000000001O000000001O0000001O000000001O0000001O00000O101O0000001O000000001O0000001O000000001O0000001O000000001O0000001N100000001O0000001O000000001O00000000000000000000001O000000000000000O100000000000000000000000000000000000000000001O0000000000000000000O10000000000000000000000000000000000000001O00000000000000000000000O1000000000000000000000000000000000001O000000000000000000000000000O100000000000000000000000001O1O001O1O001O1O1O001O1O001O1O001O1O001O1N101O1O001O1O001O1O001O1O001O001O1O001O001O001O1O001O001O1O001O001N101O1O001O001O1O001O001O001O1O001O001OO1000000O1000000O1000000O10O1000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O10O1000O1000000O1000000O1000000O1000000O1000000O101O1M4J6I6K6J6J5K6J5K6I7J5M4N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O9G9G9G9G:F9G9bMgC_Mb<T2cCkMf<g1aCWNh<\\1]CcNm<n0ZCPOo<c0VC\\OS=6TCHU=KPC4Y=^OnB`0\\=ROiBm0`=eNgBY1c?O2N1O2N1O2N101N1O2N1O2N1O2N1O2N101N1O2N1O2N1O000001O0001O00000000000001O0000000000000000001O01O0000000000000001O00000000O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O`fa1"}, "label": "a red and white bus train with the number one on the side is sitting next to another train"}]}
{"id": 377174, "image": "COCO_train2014_000000377174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white train car with two large square windows\"."}], "task": "refering", "answer_template": "The \"a red and white train car with two large square windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 390, 391, 392, 393, 394, 395, 414, 415, 416, 417], "bbox": [0.7020261437908496, 0.4373529411764706, 1.0, 0.8369117647058825], "iscrowd": 0, "area": 38038.71729999999, "rle": {"size": [612, 612], "counts": "`XQ86nb06J6J6J7I6J6J6J7H7J6J6J7I6J6XBWMi:o2mDVMR;P3dDTM\\;n2^DWMa;l2XDYMg;j2RD[Mm;h2lC]MS<f2fC^MZ<e2_C`M`<c2YCbMe<b2TCcMk<_2oBeMQ=Y4N2N2N2N2N3M2N2N2N2N2N2M3N00001O000000001O000000001N100000001O000000001O0000000O2O0000001O00000000000000000O2O0000000000000000000000001N1000000000000000000000001O0O1000000000000000000000001O000O100000000000000000001O00000O1000000000000000001O0000000O10000000000000001O000000000O10000000000000001O0000000X1hNk1UNj1VNj1VNm@"}, "label": "a red and white train car with two large square windows"}]}
{"id": 377174, "image": "COCO_train2014_000000377174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and red train behind another train\"."}], "task": "refering", "answer_template": "The \"a white and red train behind another train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 390, 391, 392, 393, 394, 395, 414, 415, 416, 417], "bbox": [0.7020261437908496, 0.4373529411764706, 1.0, 0.8369117647058825], "iscrowd": 0, "area": 38038.71729999999, "rle": {"size": [612, 612], "counts": "`XQ86nb06J6J6J7I6J6J6J7H7J6J6J7I6J6XBWMi:o2mDVMR;P3dDTM\\;n2^DWMa;l2XDYMg;j2RD[Mm;h2lC]MS<f2fC^MZ<e2_C`M`<c2YCbMe<b2TCcMk<_2oBeMQ=Y4N2N2N2N2N3M2N2N2N2N2N2M3N00001O000000001O000000001N100000001O000000001O0000000O2O0000001O00000000000000000O2O0000000000000000000000001N1000000000000000000000001O0O1000000000000000000000001O000O100000000000000000001O00000O1000000000000000001O0000000O10000000000000001O000000000O10000000000000001O0000000X1hNk1UNj1VNj1VNm@"}, "label": "a white and red train behind another train"}]}
{"id": 516439, "image": "COCO_train2014_000000516439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white refrigerator\"."}], "task": "refering", "answer_template": "The \"a white refrigerator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 382, 383, 384, 404, 405, 406, 426, 427, 428, 449, 450], "bbox": [0.39245098039215687, 0.57528125, 0.5922058823529411, 0.9168593749999999], "iscrowd": 0, "area": 16278.137399999998, "rle": {"size": [640, 612], "counts": "P`f4W1g>S4C>N1O2M2O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O1O00001O00001O00001O00001O00001O001N10001OaLRCLn<3^CBb<>jCWOU<i0VDjNl;V1`D]Na;c1kDoMW;Q2TEbMn:^2^EUMc:k2iEgLY:X3X30001O00000000000O10000000000000000O10001O00000000000O10000000000000000O101O00000000000O10000000000XOU_O[Nk`0d1[_OWNe`0h1a_OSN_`0l1h_OmMY`0R2k0O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O]lk4"}, "label": "a white refrigerator"}]}
{"id": 516439, "image": "COCO_train2014_000000516439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a white refrigerator in the kitchen\"."}], "task": "refering", "answer_template": "The \"there is a white refrigerator in the kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 382, 383, 384, 404, 405, 406, 426, 427, 428, 449, 450], "bbox": [0.39245098039215687, 0.57528125, 0.5922058823529411, 0.9168593749999999], "iscrowd": 0, "area": 16278.137399999998, "rle": {"size": [640, 612], "counts": "P`f4W1g>S4C>N1O2M2O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O1O00001O00001O00001O00001O00001O001N10001OaLRCLn<3^CBb<>jCWOU<i0VDjNl;V1`D]Na;c1kDoMW;Q2TEbMn:^2^EUMc:k2iEgLY:X3X30001O00000000000O10000000000000000O10001O00000000000O10000000000000000O101O00000000000O10000000000XOU_O[Nk`0d1[_OWNe`0h1a_OSN_`0l1h_OmMY`0R2k0O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O]lk4"}, "label": "there is a white refrigerator in the kitchen"}]}
{"id": 139612, "image": "COCO_train2014_000000139612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car in front of others\"."}], "task": "refering", "answer_template": "The \"a white car in front of others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.21612499999999998, 0.29891666666666666, 0.927421875, 0.9892500000000001], "iscrowd": 0, "area": 103769.47330000001, "rle": {"size": [480, 640], "counts": "[cQ24j>3L5L3M4L3M4SN]OWEf0g:BQEa0l:HmC4N6U<h1O100O1O1N2N2N2N2N200O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O1M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2N2O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100XOaGnJ`8n4hGnJX8o4mGoJS8P5nGoJS8Q5nGnJR8R5nGmJS8S5mGmJS8R5oGlJR8T5nGlJR8T5oGkJQ8U5oGjJR8V5nGjJR8V5oGhJR8X5nGhJR8W5oGiJQ8W5l00000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O100000000000SO^K\\Ga4d8iKSGW4l8SLkFm3T9P1O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N21O1O2N2N1N3N1O2N2N15K>A2N2N2N3M2N2N2N2N2N2M3N2M3L4M3M3L4M3L6K4G:A>B?G:J5L5K5K5J5G:YOg0_O`0L5K5K7I8H7I7I8H7IUXe0"}, "label": "a white car in front of others"}]}
{"id": 139612, "image": "COCO_train2014_000000139612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white toyota pickup with a cab on the back\"."}], "task": "refering", "answer_template": "The \"a white toyota pickup with a cab on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.21612499999999998, 0.29891666666666666, 0.927421875, 0.9892500000000001], "iscrowd": 0, "area": 103769.47330000001, "rle": {"size": [480, 640], "counts": "[cQ24j>3L5L3M4L3M4SN]OWEf0g:BQEa0l:HmC4N6U<h1O100O1O1N2N2N2N2N200O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O1M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2N2O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100XOaGnJ`8n4hGnJX8o4mGoJS8P5nGoJS8Q5nGnJR8R5nGmJS8S5mGmJS8R5oGlJR8T5nGlJR8T5oGkJQ8U5oGjJR8V5nGjJR8V5oGhJR8X5nGhJR8W5oGiJQ8W5l00000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O100000000000SO^K\\Ga4d8iKSGW4l8SLkFm3T9P1O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N21O1O2N2N1N3N1O2N2N15K>A2N2N2N3M2N2N2N2N2N2M3N2M3L4M3M3L4M3L6K4G:A>B?G:J5L5K5K5J5G:YOg0_O`0L5K5K7I8H7I7I8H7IUXe0"}, "label": "a white toyota pickup with a cab on the back"}]}
{"id": 16735, "image": "COCO_train2014_000000016735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe leaning against fence\"."}], "task": "refering", "answer_template": "The \"giraffe leaning against fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 218, 219, 220, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 292, 313, 314], "bbox": [0.48070312499999995, 0.5081458333333333, 0.7169062499999999, 0.8218333333333334], "iscrowd": 0, "area": 6821.244099999998, "rle": {"size": [480, 640], "counts": "Zd`45i>4L4K3N2N1O2M2O2N11O100O010O100O10O0100O2O0O1O2O000010O010O00010O0010O00010O00010O01O01O01O01O01O00001O00001O0000001O00001O0000001O0000000000000000000000000000001O001O1O1O1O0RBYOc=U1M3M3M3M3M3WEXNT8k1dGZN\\8i1bFUN^O7Q:h1kEoM0;E3`:f1`EUN;g0U:e1kE[NU:f1kEYNU:h1kEVNV:k1iEUNW:l1iESNX:m1hEfMd:[2[EUMV;k2`00O101N101N1]DfM_:Q1VE@9_Oa:Q1cETOLJb:Q1oEjN^O4c:R1\\F^NPO`0e:Q1jF]Nh9b1k110OQE[Nd8d1YG_Nh8`1UGcNk8]1RGfNo8Y1oFhNS9W1kFkNU9V1gFmNZ9R1dFPO\\9P1bFQO`9n0]FUOc9k0[FUOh9j0UFVOn9k0oEUOT:j0jEUOY:k0dEVO_:i0_EWOc:Y1kDfNY;^22M3N3M2M4M2M4M2N3L3H8A`0@?B?@gbd2"}, "label": "giraffe leaning against fence"}]}
{"id": 16735, "image": "COCO_train2014_000000016735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe holds its head high as it walks past another giraffe chewing on a fence\"."}], "task": "refering", "answer_template": "The \"a giraffe holds its head high as it walks past another giraffe chewing on a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 192, 193, 194, 215, 216, 217, 240, 241, 242, 263, 264, 265, 266, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 332, 333, 334, 335, 354, 355, 356, 357, 358], "bbox": [0.371640625, 0.4486875, 0.6709687499999999, 0.933], "iscrowd": 0, "area": 11326.312399999997, "rle": {"size": [480, 640], "counts": "Uj_32i>5L3L5M2O1O2N1O2N1O101N2M3N2M2O2N200000O10O100000O0113N4L3L3aG]N[3c1bL_N^3b1_L_Nc3OcG^1g4eNf3MdG^1b4hNk3HdGa1^4hNT4X1iKjNX4V1eKkN]4U1_KnNb4R1[KoNf4R1TKSOm4l0nJXOT5h0fJ]O[5c0_J_Od5b0VJ]OQ6d0hI[O_6e0[IXOn6h0nHUOY7k0bHWO`7k0ZHVOj7j0eF[Nb0l0l8i0VFhNh0a0U9g0gETOP15\\9e1^F]Ne9c1VF^Nn9c1mE]NW:c1dE\\Na:f1ZEYNj:h1QEWNT;g23M4N2N2N2N1O2N2N2N2N2N13N4L4L4L3bE`LU9c3eF_L[9e3_F]Lb9f3XF\\Li9g3QF[LP:h3jEZLV:X41O1eMbFQO`9k0fFROZ9n0iFlNiNQO_:R2jFiN\\9W1fFeN\\9Z1U201O10XEQN^8n1aGSN`8l1^GVNc8i1[GYNe8h1YGYNh8f1VG\\Nk8c1TG^Nm8a1QGaNP9^1oFcNR9]1kFeNU9[1jFfNW9Y1gFiNZ9V1eFkN[9V1`FiNe9W1QFROQ:m0fE[O[:e0\\ECe:>TEFn::mD@^;`0]DBg;i12O1N2N2N3N1N2N2N4K4M3M4L3M3M4L3010O2N1O^NRCR1m<mN\\Cm0a<SOhCf0W<YOkCh0R<XOoCi0P<VORDk0k;UOWDk0h;TOZDm0c;SO^Dn0a;QOaDP1HmNe:3gEn0[OVOo:KkEm0kN@Z;CQFa1o9_NVF\\1k9cNYFZ1f9fN^FV1b9jNbFP1`9POcFi0b9VObFb0b9^OaF;c9EaF3c9MaFKc95e2O01O1O1O1O100O00100O1000000O10O1000O01000001N101O001N10001O0O2O001N101O00]aR3"}, "label": "a giraffe holds its head high as it walks past another giraffe chewing on a fence"}]}
{"id": 16735, "image": "COCO_train2014_000000016735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe standing near log\"."}], "task": "refering", "answer_template": "The \"giraffe standing near log\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 192, 193, 194, 215, 216, 217, 240, 241, 242, 263, 264, 265, 266, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 332, 333, 334, 335, 354, 355, 356, 357, 358], "bbox": [0.371640625, 0.4486875, 0.6709687499999999, 0.933], "iscrowd": 0, "area": 11326.312399999997, "rle": {"size": [480, 640], "counts": "Uj_32i>5L3L5M2O1O2N1O2N1O101N2M3N2M2O2N200000O10O100000O0113N4L3L3aG]N[3c1bL_N^3b1_L_Nc3OcG^1g4eNf3MdG^1b4hNk3HdGa1^4hNT4X1iKjNX4V1eKkN]4U1_KnNb4R1[KoNf4R1TKSOm4l0nJXOT5h0fJ]O[5c0_J_Od5b0VJ]OQ6d0hI[O_6e0[IXOn6h0nHUOY7k0bHWO`7k0ZHVOj7j0eF[Nb0l0l8i0VFhNh0a0U9g0gETOP15\\9e1^F]Ne9c1VF^Nn9c1mE]NW:c1dE\\Na:f1ZEYNj:h1QEWNT;g23M4N2N2N2N1O2N2N2N2N2N13N4L4L4L3bE`LU9c3eF_L[9e3_F]Lb9f3XF\\Li9g3QF[LP:h3jEZLV:X41O1eMbFQO`9k0fFROZ9n0iFlNiNQO_:R2jFiN\\9W1fFeN\\9Z1U201O10XEQN^8n1aGSN`8l1^GVNc8i1[GYNe8h1YGYNh8f1VG\\Nk8c1TG^Nm8a1QGaNP9^1oFcNR9]1kFeNU9[1jFfNW9Y1gFiNZ9V1eFkN[9V1`FiNe9W1QFROQ:m0fE[O[:e0\\ECe:>TEFn::mD@^;`0]DBg;i12O1N2N2N3N1N2N2N4K4M3M4L3M3M4L3010O2N1O^NRCR1m<mN\\Cm0a<SOhCf0W<YOkCh0R<XOoCi0P<VORDk0k;UOWDk0h;TOZDm0c;SO^Dn0a;QOaDP1HmNe:3gEn0[OVOo:KkEm0kN@Z;CQFa1o9_NVF\\1k9cNYFZ1f9fN^FV1b9jNbFP1`9POcFi0b9VObFb0b9^OaF;c9EaF3c9MaFKc95e2O01O1O1O1O100O00100O1000000O10O1000O01000001N101O001N10001O0O2O001N101O00]aR3"}, "label": "giraffe standing near log"}]}
{"id": 180578, "image": "COCO_train2014_000000180578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair in front of a window\"."}], "task": "refering", "answer_template": "The \"a chair in front of a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 256, 257, 258, 259, 279, 280, 281], "bbox": [0.10546875, 0.4570416666666667, 0.34375, 0.7643333333333333], "iscrowd": 0, "area": 13836.349999999999, "rle": {"size": [480, 640], "counts": "cSP13l>4M3L4M3L4M3L4M3L4M3L4L4M4K6K4K6K5J6K5J5L5J6K5J6K4K6K3L2O0O2O1N2O0O2O1N2O1N101O1O1O001O1O1O0O2O1OO100O100O10000O1000O0100O10000O100O10000O10O010000O100O10000O100O10O10O2O1O0O2O1N2O0PO`E`Ma:_2Q1N101O1N101O1N2O001N2O001N2O1O0O2O1O0O2O1O1N101O1N101O1N101O0O2O1O0O2O1O0O2O1O001N101O1N101PO`B>a=]OdBa0]=[OhBd0Y=WOlBh0g=N2O1O1N2O1O1M3L4L4KVcT6"}, "label": "a chair in front of a window"}]}
{"id": 180578, "image": "COCO_train2014_000000180578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair with 2 bags on it\"."}], "task": "refering", "answer_template": "The \"chair with 2 bags on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 256, 257, 258, 259, 279, 280, 281], "bbox": [0.10546875, 0.4570416666666667, 0.34375, 0.7643333333333333], "iscrowd": 0, "area": 13836.349999999999, "rle": {"size": [480, 640], "counts": "cSP13l>4M3L4M3L4M3L4M3L4M3L4L4M4K6K4K6K5J6K5J5L5J6K5J6K4K6K3L2O0O2O1N2O0O2O1N2O1N101O1O1O001O1O1O0O2O1OO100O100O10000O1000O0100O10000O100O10000O10O010000O100O10000O100O10O10O2O1O0O2O1N2O0PO`E`Ma:_2Q1N101O1N101O1N2O001N2O001N2O1O0O2O1O0O2O1O1N101O1N101O1N101O0O2O1O0O2O1O0O2O1O001N101O1N101PO`B>a=]OdBa0]=[OhBd0Y=WOlBh0g=N2O1O1N2O1O1M3L4L4KVcT6"}, "label": "chair with 2 bags on it"}]}
{"id": 180578, "image": "COCO_train2014_000000180578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown chair situated under a brown desk\"."}], "task": "refering", "answer_template": "The \"a dark brown chair situated under a brown desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 366, 367, 386, 389, 390], "bbox": [0.77528125, 0.5903333333333334, 0.985953125, 0.9858333333333335], "iscrowd": 0, "area": 14372.824900000001, "rle": {"size": [480, 640], "counts": "jkX7^1b=0O101O0O101N101O0O101O0O101O4K>C=B>CO1N1N3N2M2O2N1N3N2N1N3N2M2O2N2M3N1N3N2N1O2000O10O1000O1N1N3M3M3M2N3M3M3N1N3M3M3M2N3M3N2M2N3M3M2N3M3O11N2O1OO10O10O1000O10O1000O10O1000O0100000O0100000OUF`LW8a3dGdL\\8\\3`GhL`8X3[GnLc8R3ZGRMf8n2UGWMk8i2QG\\Mn8c2nFbMQ9_2jFgMU9Y2gFkMY9U2bFPN]9P2`FUN_9f30001N10001O0O2O000O2O00001NN3K5K5J5L5K5J5L5K5J6K4L5J6iNW1YNko3"}, "label": "a dark brown chair situated under a brown desk"}]}
{"id": 180578, "image": "COCO_train2014_000000180578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown office chair pushed under a table\"."}], "task": "refering", "answer_template": "The \"a dark brown office chair pushed under a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 366, 367, 386, 389, 390], "bbox": [0.77528125, 0.5903333333333334, 0.985953125, 0.9858333333333335], "iscrowd": 0, "area": 14372.824900000001, "rle": {"size": [480, 640], "counts": "jkX7^1b=0O101O0O101N101O0O101O0O101O4K>C=B>CO1N1N3N2M2O2N1N3N2N1N3N2M2O2N2M3N1N3N2N1O2000O10O1000O1N1N3M3M3M2N3M3M3N1N3M3M3M2N3M3N2M2N3M3M2N3M3O11N2O1OO10O10O1000O10O1000O10O1000O0100000O0100000OUF`LW8a3dGdL\\8\\3`GhL`8X3[GnLc8R3ZGRMf8n2UGWMk8i2QG\\Mn8c2nFbMQ9_2jFgMU9Y2gFkMY9U2bFPN]9P2`FUN_9f30001N10001O0O2O000O2O00001NN3K5K5J5L5K5J5L5K5J6K4L5J6iNW1YNko3"}, "label": "a dark brown office chair pushed under a table"}]}
{"id": 180578, "image": "COCO_train2014_000000180578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black roller suitcase in the center of the room\"."}], "task": "refering", "answer_template": "The \"the black roller suitcase in the center of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 240, 241, 263, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 379, 380], "bbox": [0.447765625, 0.5348541666666667, 0.5748749999999999, 0.9810208333333333], "iscrowd": 0, "area": 11207.493900000001, "rle": {"size": [480, 640], "counts": "hkV4:e>;]C\\OU:o0[EYOb:S1lDUOR;V1\\DSOb;X2M2N2O2M2`MmLPIV3n6UMfHm2Y7_MZHc2e7jMmGX2S8TN_Go1_8^NSGd1l8X20O1O1O2O0O11O0010O01O001O0010O01O02O5J9H9F9H8G9H6I10O010O00101N2O1N2O2M01O001O00010O0M4K4M4L3N3M2O2M2O2O0O2O0O2O0O2N101N101N101]OmUo3"}, "label": "the black roller suitcase in the center of the room"}]}
{"id": 180578, "image": "COCO_train2014_000000180578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suitcase in the middle of the room\"."}], "task": "refering", "answer_template": "The \"black suitcase in the middle of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 240, 241, 263, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 379, 380], "bbox": [0.447765625, 0.5348541666666667, 0.5748749999999999, 0.9810208333333333], "iscrowd": 0, "area": 11207.493900000001, "rle": {"size": [480, 640], "counts": "hkV4:e>;]C\\OU:o0[EYOb:S1lDUOR;V1\\DSOb;X2M2N2O2M2`MmLPIV3n6UMfHm2Y7_MZHc2e7jMmGX2S8TN_Go1_8^NSGd1l8X20O1O1O2O0O11O0010O01O001O0010O01O02O5J9H9F9H8G9H6I10O010O00101N2O1N2O2M01O001O00010O0M4K4M4L3N3M2O2M2O2O0O2O0O2O0O2N101N101N101]OmUo3"}, "label": "black suitcase in the middle of the room"}]}
{"id": 270696, "image": "COCO_train2014_000000270696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a duct tape suit talking to someone\"."}], "task": "refering", "answer_template": "The \"man in a duct tape suit talking to someone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 65, 66, 67, 68, 69, 80, 81, 82, 83, 84, 96, 97, 98, 99, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 275, 276, 277, 278, 279, 280, 281, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 310, 311, 320, 321, 322, 323, 324, 325, 326, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.21304449648711943, 0.1271875, 0.8671428571428571, 0.98753125], "iscrowd": 0, "area": 92742.0342, "rle": {"size": [640, 427], "counts": "QWi13gc07I7I7I7I7I6O2N2N2N2N2N1O2N2N2N2N2N1O2N2M2N1O1N3N1O1O1N3N1O1N2O2N1N2O2N1N2O1O2M2O1O1N3N1O1N2O2N1N2O108H?B>B>A?B>A]6dI000O100O10000eL]CXOd<f0`CXO`<f0eCVO\\<g0iCWOW<g0nCVOR<h0SDTOn;i0WDUOi;i0\\DSOe;k0`DRO`;l0eDQO[;l0jDQOW;m0nDPOR;o0UEjNl:V1YEeNg:[1_E^Nb:a1eEYN[:g1jESNW:m1kEQNU:n1oEnMR:R2PFlMP:T2RFjMn9U2VFgMk9Y2WFeMi9[2YFbMh9]2\\F`Md9`2^F]Mc9c2_F[Ma9d2cFXM^9h2eFUM[9j2hFSMY9m2jFPMV9P3m3O10000O10000O10000O1000000O10000O1000000O1000000O11O0000001O0000001O0000001O0000001O00002N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N2N2N2N2^BdK\\;^4`DeK_;^4\\DdKd;^4XDdKh;^4UDdKj;^4RDdKn;W60001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O001O001O00001O001O001O001O001O001O0UGnLh0R3XOQMe0P3[OSMa0n2_OTM>m2BVM:l2EVM8k2HWM5j2JZM2h2MZM0g20\\MLf22]MKd25^MHc27aMEa2:aMC`2<dM@^2?dM^O]2a0fM\\O[2d0hMXOZ2f0iMWOY2h0lMROV2l0SNkNo1T1YNcNj1Z1^N^Nd1`1eNWN]1g1kNQNW1m1ROjMQ1S2XOdMj0Z2^O^Md0`2EWM>f2EWM=g2FVM<h2FVM=g2`8N2N2N3L3N2N3M2N2N3M2N2N2N3M2N2N3M2N2J6J7H7I7I8IkYS1"}, "label": "man in a duct tape suit talking to someone"}]}
{"id": 270696, "image": "COCO_train2014_000000270696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the duct tape outfit\"."}], "task": "refering", "answer_template": "The \"the man in the duct tape outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 65, 66, 67, 68, 69, 80, 81, 82, 83, 84, 96, 97, 98, 99, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 275, 276, 277, 278, 279, 280, 281, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 310, 311, 320, 321, 322, 323, 324, 325, 326, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.21304449648711943, 0.1271875, 0.8671428571428571, 0.98753125], "iscrowd": 0, "area": 92742.0342, "rle": {"size": [640, 427], "counts": "QWi13gc07I7I7I7I7I6O2N2N2N2N2N1O2N2N2N2N2N1O2N2M2N1O1N3N1O1O1N3N1O1N2O2N1N2O2N1N2O1O2M2O1O1N3N1O1N2O2N1N2O108H?B>B>A?B>A]6dI000O100O10000eL]CXOd<f0`CXO`<f0eCVO\\<g0iCWOW<g0nCVOR<h0SDTOn;i0WDUOi;i0\\DSOe;k0`DRO`;l0eDQO[;l0jDQOW;m0nDPOR;o0UEjNl:V1YEeNg:[1_E^Nb:a1eEYN[:g1jESNW:m1kEQNU:n1oEnMR:R2PFlMP:T2RFjMn9U2VFgMk9Y2WFeMi9[2YFbMh9]2\\F`Md9`2^F]Mc9c2_F[Ma9d2cFXM^9h2eFUM[9j2hFSMY9m2jFPMV9P3m3O10000O10000O10000O1000000O10000O1000000O1000000O11O0000001O0000001O0000001O0000001O00002N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N2N2N2N2^BdK\\;^4`DeK_;^4\\DdKd;^4XDdKh;^4UDdKj;^4RDdKn;W60001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O001O001O00001O001O001O001O001O001O0UGnLh0R3XOQMe0P3[OSMa0n2_OTM>m2BVM:l2EVM8k2HWM5j2JZM2h2MZM0g20\\MLf22]MKd25^MHc27aMEa2:aMC`2<dM@^2?dM^O]2a0fM\\O[2d0hMXOZ2f0iMWOY2h0lMROV2l0SNkNo1T1YNcNj1Z1^N^Nd1`1eNWN]1g1kNQNW1m1ROjMQ1S2XOdMj0Z2^O^Md0`2EWM>f2EWM=g2FVM<h2FVM=g2`8N2N2N3L3N2N3M2N2N3M2N2N2N3M2N2N3M2N2J6J7H7I7I8IkYS1"}, "label": "the man in the duct tape outfit"}]}
{"id": 270696, "image": "COCO_train2014_000000270696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a blue jacket on holding a beer bottle\"."}], "task": "refering", "answer_template": "The \"a man with a blue jacket on holding a beer bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 30, 31, 45, 46, 60, 61, 75, 76, 90, 91, 105, 106, 120, 121, 135, 136, 150, 151, 165, 166, 180, 181, 182, 195, 196, 197, 210, 211, 212, 225, 226, 227, 240, 241, 242, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 330, 331, 332], "bbox": [0.0, 0.067359375, 0.20074941451990633, 1.0], "iscrowd": 0, "area": 34055.1733, "rle": {"size": [640, 427], "counts": "[1l0Uc0b1TLSN`B_3c6o1ZIBV3a2eLROW1b1gN0UOa0l0Y9001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O1_AnMR;Y2bDPNZ;X2YDQNc;V2QDSNk;U2hCTNT<S2`CVN\\<R2WCWNe<P2nBZNn<m1fB\\NV=[4L4L4L<Db0^Oc0]Ob0^Od0\\O3M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3M3M3M4L3nLXDfMk;T2_DeMd;U2eDeM^;U2lDdMX;T2SEeMP;U2YEeMj:U2`EdMc:V2fEdM^:U2hEhM[:R2jElMY:m1mEnMY:l1mEXMo:b2e3AR1nNged6"}, "label": "a man with a blue jacket on holding a beer bottle"}]}
{"id": 270696, "image": "COCO_train2014_000000270696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding beer\"."}], "task": "refering", "answer_template": "The \"man holding beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 30, 31, 45, 46, 60, 61, 75, 76, 90, 91, 105, 106, 120, 121, 135, 136, 150, 151, 165, 166, 180, 181, 182, 195, 196, 197, 210, 211, 212, 225, 226, 227, 240, 241, 242, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 330, 331, 332], "bbox": [0.0, 0.067359375, 0.20074941451990633, 1.0], "iscrowd": 0, "area": 34055.1733, "rle": {"size": [640, 427], "counts": "[1l0Uc0b1TLSN`B_3c6o1ZIBV3a2eLROW1b1gN0UOa0l0Y9001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O1_AnMR;Y2bDPNZ;X2YDQNc;V2QDSNk;U2hCTNT<S2`CVN\\<R2WCWNe<P2nBZNn<m1fB\\NV=[4L4L4L<Db0^Oc0]Ob0^Od0\\O3M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3M3M3M4L3nLXDfMk;T2_DeMd;U2eDeM^;U2lDdMX;T2SEeMP;U2YEeMj:U2`EdMc:V2fEdM^:U2hEhM[:R2jElMY:m1mEnMY:l1mEXMo:b2e3AR1nNged6"}, "label": "man holding beer"}]}
{"id": 303471, "image": "COCO_train2014_000000303471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the right\"."}], "task": "refering", "answer_template": "The \"the woman on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 89, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 246, 247, 250, 251, 269, 270, 273, 274, 292, 293, 296, 297, 315, 316, 320], "bbox": [0.63125, 0.13245939675174015, 0.9777968749999999, 0.9369605568445477], "iscrowd": 0, "area": 30848.325500000003, "rle": {"size": [431, 640], "counts": "`XZ51n<3\\C1b<N^C4`<M_C4`<L_C6`<J`C7_<J`C6`<J_C8`<I^C9a<G^C;`<G^C;`<:01N2N2N2N1O1O01000O01000O0100001N2O001N2O1O1O001O1O1O1O1O001O1O1O001O010O0TJ\\NNe1O^N0b1OaNO`1ObN1]1NfN0[1NgN1Y1NiN1X1MjN2W1MjN3U1LmN3T1LmN3S1MnN2S1LPO2P1NQO2o0LSO3m0YM`NDd0R3m0XMWOnNNi3k0XMMZNYO]4k0WMa0gMfNR5i0UMn1j2RNUMP2j2QNTMQ2k2oMTMS2k2nMSMU2k2kMTMW2k2jMSMX2l2hMSMY2m2hMQMY2o2gMPM[2P3eMjL`2V3`MfLd2Z3]MaLh2^3YMaLg2_3YM`Lh2`3YM_Lh2`3XM`Lh2`3YM_Lg2a3YM^Lh2b3YM]Le1hNTMk4V1]La1QOUMb4[1\\LZ1\\OWMX4_1[LV1FXMn3c1[Lo01ZMd3h1ZLh0<\\MZ3l1ZLa0g0`MX2MYLS2g09S1cMe1=]Lg1j02_1gMQ1n0aLZ1n0Jj1lM`0[1cLP1R1B_5m1aJQNb5o1]JQNc5o1]JQNc5P2]JoMd5Q2[JoMe5Q2[JoMe5R2ZJmMh5R2XJnMh5S2WJmMi5T2WJkMj5T2VJlMj5U2UJkMl5T2TJmMk5T2TJlMm5T2SJlMn5R2RJoMn5Q2QJoMP6P2PJQNQ6n1nIRNS6n1mIRNS6m1mITNT6k1kIUNV6j1jIWNV6i1iIWNX6i1hIWNY6g1gIZNY6f1fIZN[6e1eI\\N\\6d1bI]N^6f1_IYNb6j1ZIWNg6k1UIVNk6l1RIUNn6n1oHQNS7Q2iHPNY7R42N3L2O1O001N101O001N2O2N3L4M=B6J101NUIYKP5f4QK]Km4b4SK`Kl4^4TKfKi4[4VKgKi4Z4TKiKj4Y4TKjKj4W4TKkKj4W4TKlKj4U4TKmKk4T4SKoKj4S4TKoKk4R4SKPLk4Q4TKRLk4n3SKTLl4[6N2M3M2O2M01O001OO1O1O00100O1O1O1O00O2M2O8H3ROQJWJS6d5oI[JT6`5oI^JV6\\5mITJf6g5d0K6K5J6K4K6J;F5E;E:F;E;D<F:L4M6I:F2O2M3]Oc0M3N2N2M3N2M3N2L5I6KYe5"}, "label": "the woman on the right"}]}
{"id": 303471, "image": "COCO_train2014_000000303471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with the racket in the photo on the right\"."}], "task": "refering", "answer_template": "The \"the girl with the racket in the photo on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 89, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 246, 247, 250, 251, 269, 270, 273, 274, 292, 293, 296, 297, 315, 316, 320], "bbox": [0.63125, 0.13245939675174015, 0.9777968749999999, 0.9369605568445477], "iscrowd": 0, "area": 30848.325500000003, "rle": {"size": [431, 640], "counts": "`XZ51n<3\\C1b<N^C4`<M_C4`<L_C6`<J`C7_<J`C6`<J_C8`<I^C9a<G^C;`<G^C;`<:01N2N2N2N1O1O01000O01000O0100001N2O001N2O1O1O001O1O1O1O1O001O1O1O001O010O0TJ\\NNe1O^N0b1OaNO`1ObN1]1NfN0[1NgN1Y1NiN1X1MjN2W1MjN3U1LmN3T1LmN3S1MnN2S1LPO2P1NQO2o0LSO3m0YM`NDd0R3m0XMWOnNNi3k0XMMZNYO]4k0WMa0gMfNR5i0UMn1j2RNUMP2j2QNTMQ2k2oMTMS2k2nMSMU2k2kMTMW2k2jMSMX2l2hMSMY2m2hMQMY2o2gMPM[2P3eMjL`2V3`MfLd2Z3]MaLh2^3YMaLg2_3YM`Lh2`3YM_Lh2`3XM`Lh2`3YM_Lg2a3YM^Lh2b3YM]Le1hNTMk4V1]La1QOUMb4[1\\LZ1\\OWMX4_1[LV1FXMn3c1[Lo01ZMd3h1ZLh0<\\MZ3l1ZLa0g0`MX2MYLS2g09S1cMe1=]Lg1j02_1gMQ1n0aLZ1n0Jj1lM`0[1cLP1R1B_5m1aJQNb5o1]JQNc5o1]JQNc5P2]JoMd5Q2[JoMe5Q2[JoMe5R2ZJmMh5R2XJnMh5S2WJmMi5T2WJkMj5T2VJlMj5U2UJkMl5T2TJmMk5T2TJlMm5T2SJlMn5R2RJoMn5Q2QJoMP6P2PJQNQ6n1nIRNS6n1mIRNS6m1mITNT6k1kIUNV6j1jIWNV6i1iIWNX6i1hIWNY6g1gIZNY6f1fIZN[6e1eI\\N\\6d1bI]N^6f1_IYNb6j1ZIWNg6k1UIVNk6l1RIUNn6n1oHQNS7Q2iHPNY7R42N3L2O1O001N101O001N2O2N3L4M=B6J101NUIYKP5f4QK]Km4b4SK`Kl4^4TKfKi4[4VKgKi4Z4TKiKj4Y4TKjKj4W4TKkKj4W4TKlKj4U4TKmKk4T4SKoKj4S4TKoKk4R4SKPLk4Q4TKRLk4n3SKTLl4[6N2M3M2O2M01O001OO1O1O00100O1O1O1O00O2M2O8H3ROQJWJS6d5oI[JT6`5oI^JV6\\5mITJf6g5d0K6K5J6K4K6J;F5E;E:F;E;D<F:L4M6I:F2O2M3]Oc0M3N2N2M3N2M3N2L5I6KYe5"}, "label": "the girl with the racket in the photo on the right"}]}
{"id": 303471, "image": "COCO_train2014_000000303471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on the left side with racket in hand\"."}], "task": "refering", "answer_template": "The \"woman on the left side with racket in hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 75, 76, 77, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 239, 240, 258, 259, 261, 262, 263, 281, 282, 285, 286, 304, 305, 308], "bbox": [0.12484375, 0.13436194895591647, 0.48228124999999994, 0.9379350348027842], "iscrowd": 0, "area": 31852.467200000006, "rle": {"size": [431, 640], "counts": "YlQ11]=1N3N1O2N1N3N1O2N1N2O2N1O2N1O2N1O2N2N1O2N2N1O2N1O001O1O1O1O1OO100000O1000O10O1000000O1000001O1N2O1O001O100O1O1O1UJ[NKf11_NMa10cNO^1MfN2Z1LiN4W1KjN4V1LkN3V1LkN3U1MlN2U1MmN1S1OnN0S1OnN1Q1OPO0Q1OPO0P1ORO0o0gMGcM[Oe4n0dM5]MnNn4m0`Md1`2]N[Mh1e2WNVMn1j2SNQMR2n2nMmLX2R3iMlLX2S3iMlLY2S3hMlLX2T3hMlLY2S3hMlLY2R3hMmLY2S3hMlLY2S3gMmLY2S3hMkLZ2S3gMmLY2S3hMlLY2S3gMmLZ2R3gMlLZ2T3fMlL[2S3fMlLZ2U3eMjLY1]N`Mh4X1jLQ1gNfM^4Y1kLk0nNlMW4X1kLe0VORNo3Z1iL?_OWN\\2e2UNnNIZNQ2i2UNgN5[Nf1n2UNbN>]N\\1R3TN\\Nk0^NP1V3UNVNU1`Nf0[3TNoMa1bN:_3TNmMi1aN2c3TNlMo1]NLg3UNlMY5T2fJlMZ5T2eJmM[5T2dJmM[5S2eJmM[5T2dJlM]5S2bJnM_5R2`JnMa5Q2_JPN`5Q2^JPNc5o1]JQNd5o1[JQNf5n1ZJRNg5n1WJSNj5l1VJUNj5k1UJUNl5k1SJUNn5k1QJUNo5l1PJTNQ6l1nITNS6l1kIVNU6j1jIVNW6j1gIWNZ6i1dIXN]6i1_IZNa6f1]I\\Nd6f1WI\\Ni6h1PI[NP7Q410001O0O10001O1N2O1O1O1N2O1O3M5J6K5K4K6J6I6K2N1N3N2OZIeKX4_4_KhK`4[4[KhKc4\\4XKhKf4[4VKgKi4[4SKhKk4Y4SKjKl4V4RKmKm4S4QKPLm4X4jJkKU5\\4bJgK]5a4YJcKd5e4SJ^Kl5T6O2M2O1N2O2N1N2O2N0O10000O01O1O001O1O3M3L3N3M2N3M2nNmIcJW6Y5mIdJV6X5nIdJU6Y5PJbJT6Z5SJ`Jo5]5XJ\\Jl5`5T1K6G9G8H9F9_Ob0_Oa0F9L5L3M4L9G9E<C8I4K6J5L5K5K4M4K4LoR[4"}, "label": "woman on the left side with racket in hand"}]}
{"id": 303471, "image": "COCO_train2014_000000303471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female asian tennis player on left side\"."}], "task": "refering", "answer_template": "The \"female asian tennis player on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 75, 76, 77, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 239, 240, 258, 259, 261, 262, 263, 281, 282, 285, 286, 304, 305, 308], "bbox": [0.12484375, 0.13436194895591647, 0.48228124999999994, 0.9379350348027842], "iscrowd": 0, "area": 31852.467200000006, "rle": {"size": [431, 640], "counts": "YlQ11]=1N3N1O2N1N3N1O2N1N2O2N1O2N1O2N1O2N2N1O2N2N1O2N1O001O1O1O1O1OO100000O1000O10O1000000O1000001O1N2O1O001O100O1O1O1UJ[NKf11_NMa10cNO^1MfN2Z1LiN4W1KjN4V1LkN3V1LkN3U1MlN2U1MmN1S1OnN0S1OnN1Q1OPO0Q1OPO0P1ORO0o0gMGcM[Oe4n0dM5]MnNn4m0`Md1`2]N[Mh1e2WNVMn1j2SNQMR2n2nMmLX2R3iMlLX2S3iMlLY2S3hMlLX2T3hMlLY2S3hMlLY2R3hMmLY2S3hMlLY2S3gMmLY2S3hMkLZ2S3gMmLY2S3hMlLY2S3gMmLZ2R3gMlLZ2T3fMlL[2S3fMlLZ2U3eMjLY1]N`Mh4X1jLQ1gNfM^4Y1kLk0nNlMW4X1kLe0VORNo3Z1iL?_OWN\\2e2UNnNIZNQ2i2UNgN5[Nf1n2UNbN>]N\\1R3TN\\Nk0^NP1V3UNVNU1`Nf0[3TNoMa1bN:_3TNmMi1aN2c3TNlMo1]NLg3UNlMY5T2fJlMZ5T2eJmM[5T2dJmM[5S2eJmM[5T2dJlM]5S2bJnM_5R2`JnMa5Q2_JPN`5Q2^JPNc5o1]JQNd5o1[JQNf5n1ZJRNg5n1WJSNj5l1VJUNj5k1UJUNl5k1SJUNn5k1QJUNo5l1PJTNQ6l1nITNS6l1kIVNU6j1jIVNW6j1gIWNZ6i1dIXN]6i1_IZNa6f1]I\\Nd6f1WI\\Ni6h1PI[NP7Q410001O0O10001O1N2O1O1O1N2O1O3M5J6K5K4K6J6I6K2N1N3N2OZIeKX4_4_KhK`4[4[KhKc4\\4XKhKf4[4VKgKi4[4SKhKk4Y4SKjKl4V4RKmKm4S4QKPLm4X4jJkKU5\\4bJgK]5a4YJcKd5e4SJ^Kl5T6O2M2O1N2O2N1N2O2N0O10000O01O1O001O1O3M3L3N3M2N3M2nNmIcJW6Y5mIdJV6X5nIdJU6Y5PJbJT6Z5SJ`Jo5]5XJ\\Jl5`5T1K6G9G8H9F9_Ob0_Oa0F9L5L3M4L9G9E<C8I4K6J5L5K5K4M4K4LoR[4"}, "label": "female asian tennis player on left side"}]}
{"id": 98672, "image": "COCO_train2014_000000098672.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with mobile phone\"."}], "task": "refering", "answer_template": "The \"a man with mobile phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 117, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 258, 259, 260, 261, 262, 263, 276, 282, 283, 284, 285, 286], "bbox": [0.01196875, 0.3051990632318501, 0.44878125, 0.8276112412177986], "iscrowd": 0, "area": 32440.829300000005, "rle": {"size": [427, 640], "counts": "nd39m<9F;F9F6N2N2N2N2N2N2N2N2N2N2N1O2N101N101N101N101N1O2O0O2O0O2O0O2O0O1O2N1O2M2O1O1O000O101O1N2M6K6J5J7J5J7I6J5K4K6KoNfGZMV8h2RHQMk7R3[HiLa7Z3eH`LX7c3oHWLm6l3YIoKb6U4eIdKW6`4W13M3M2N3M3O1O100O1O1O10O01O100O1O10000O10O1000O10000O100000O100000O1000000000O10O1000000000000O02O1O1O1O1O1O001O1O2M3N1O2N2N1O2N2N2N2N2O0O2N100dLPG\\2P9aMSG`2l8]MWGd2h8YMZGi2e8TM]Gm2`92N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N1O01O000000001O000000001O000000001O0O1000001O0000002N1O1O1O1O1O1O1O2N1O1O1O1N2O1O1O2N1O1O1O1O1O1O2N1nNkFlMV9R2oFjMS9S2QGjMP9T2TGiMm8U2XGgMj8V2ZGgMg8X2W1N3M2N3M2N2100O1O1\\ERNi9n1RFWNm9j1oEYNQ:h1kE\\NU:d1gE_NY:b1cE`N^:V22N1O101N1O2N1O1O2N1O1O2N100O2M2O1O2N1N2O2N1O2M2O1O4L6I7J4L2M3N2N2N2N2N2N2N2N2M3M3Mkkb4"}, "label": "a man with mobile phone"}]}
{"id": 98672, "image": "COCO_train2014_000000098672.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black t - shirt talking in the mobile\"."}], "task": "refering", "answer_template": "The \"a man in black t - shirt talking in the mobile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 117, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 258, 259, 260, 261, 262, 263, 276, 282, 283, 284, 285, 286], "bbox": [0.01196875, 0.3051990632318501, 0.44878125, 0.8276112412177986], "iscrowd": 0, "area": 32440.829300000005, "rle": {"size": [427, 640], "counts": "nd39m<9F;F9F6N2N2N2N2N2N2N2N2N2N2N1O2N101N101N101N101N1O2O0O2O0O2O0O2O0O1O2N1O2M2O1O1O000O101O1N2M6K6J5J7J5J7I6J5K4K6KoNfGZMV8h2RHQMk7R3[HiLa7Z3eH`LX7c3oHWLm6l3YIoKb6U4eIdKW6`4W13M3M2N3M3O1O100O1O1O10O01O100O1O10000O10O1000O10000O100000O100000O1000000000O10O1000000000000O02O1O1O1O1O1O001O1O2M3N1O2N2N1O2N2N2N2N2O0O2N100dLPG\\2P9aMSG`2l8]MWGd2h8YMZGi2e8TM]Gm2`92N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N1O01O000000001O000000001O000000001O0O1000001O0000002N1O1O1O1O1O1O1O2N1O1O1O1N2O1O1O2N1O1O1O1O1O1O2N1nNkFlMV9R2oFjMS9S2QGjMP9T2TGiMm8U2XGgMj8V2ZGgMg8X2W1N3M2N3M2N2100O1O1\\ERNi9n1RFWNm9j1oEYNQ:h1kE\\NU:d1gE_NY:b1cE`N^:V22N1O101N1O2N1O1O2N1O1O2N100O2M2O1O2N1N2O2N1O2M2O1O4L6I7J4L2M3N2N2N2N2N2N2N2N2M3M3Mkkb4"}, "label": "a man in black t - shirt talking in the mobile"}]}
{"id": 336242, "image": "COCO_train2014_000000336242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady wearing the red sweater\"."}], "task": "refering", "answer_template": "The \"the lady wearing the red sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 26, 27, 28, 29, 43, 44, 45, 46, 60, 61, 62, 63, 77, 78, 79, 80, 94, 95, 96, 111, 112, 113, 129, 130, 162, 163, 164, 180, 181, 198], "bbox": [0.5420208333333334, 0.00071875, 0.7617708333333334, 0.486546875], "iscrowd": 0, "area": 17656.5743, "rle": {"size": [640, 480], "counts": "]dR54cc0;D<D<E<C<fNZ1VOj0L5J5L4L4K5L5J5L4K5L5J5gC[JQ:i5mE[Jo9g5oE\\Jn9g5oE]Jm9e5QF^Jm9c5QFaJk9a5SFcJi9^5VFeJg9\\5XFhJd9X5\\FkJa9U5^FPK^9Q5aFSK[9m4eFVKX9j4hFZKT9g4kFZKT9f4lFZKT9f4lFZKT9f4lFZKT9f4lF[KS9d4nF\\KR9d4nF\\KR9d4nF\\KR9d4nF]KQ9c4oF]KQ9c4nF^KR9b4nF_KQ9a4oF_KQ9a4oF_KQ9a4oF_KQ9a4oF`KP9`4PG`KP9`4PG`KP9`4PG`KP9`4PGaKo8_4QGaKo8^4RGbKn8^4RGbKn8^4RGcKn8\\4QGeKo8[4QGeKo8[4\\G[Kc8e4kGmJU8S5YH_Jg7a5ZH^Jf7b5\\H]Jc7c5^H\\Jb7d5`HZJ`7f5aHYJ_7g5cHXJ^7f5cHYJ^7e5eHYJ\\7f5eHYJ]7e5eHZJ[7e5fHZJ\\7d5fHZJ[7e5fHXJ]7g5`3N1kNU1A`0K4L4K6K4L5K4L4L5J5L4L5K4L4L5K4M4M2M3M4L3M3N4K8H7I8H7I8IcUW2"}, "label": "the lady wearing the red sweater"}]}
{"id": 336242, "image": "COCO_train2014_000000336242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in red cardigan behind young girl\"."}], "task": "refering", "answer_template": "The \"woman in red cardigan behind young girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 26, 27, 28, 29, 43, 44, 45, 46, 60, 61, 62, 63, 77, 78, 79, 80, 94, 95, 96, 111, 112, 113, 129, 130, 162, 163, 164, 180, 181, 198], "bbox": [0.5420208333333334, 0.00071875, 0.7617708333333334, 0.486546875], "iscrowd": 0, "area": 17656.5743, "rle": {"size": [640, 480], "counts": "]dR54cc0;D<D<E<C<fNZ1VOj0L5J5L4L4K5L5J5L4K5L5J5gC[JQ:i5mE[Jo9g5oE\\Jn9g5oE]Jm9e5QF^Jm9c5QFaJk9a5SFcJi9^5VFeJg9\\5XFhJd9X5\\FkJa9U5^FPK^9Q5aFSK[9m4eFVKX9j4hFZKT9g4kFZKT9f4lFZKT9f4lFZKT9f4lFZKT9f4lF[KS9d4nF\\KR9d4nF\\KR9d4nF\\KR9d4nF]KQ9c4oF]KQ9c4nF^KR9b4nF_KQ9a4oF_KQ9a4oF_KQ9a4oF_KQ9a4oF`KP9`4PG`KP9`4PG`KP9`4PG`KP9`4PGaKo8_4QGaKo8^4RGbKn8^4RGbKn8^4RGcKn8\\4QGeKo8[4QGeKo8[4\\G[Kc8e4kGmJU8S5YH_Jg7a5ZH^Jf7b5\\H]Jc7c5^H\\Jb7d5`HZJ`7f5aHYJ_7g5cHXJ^7f5cHYJ^7e5eHYJ\\7f5eHYJ]7e5eHZJ[7e5fHZJ\\7d5fHZJ[7e5fHXJ]7g5`3N1kNU1A`0K4L4K6K4L5K4L4L5J5L4L5K4L4L5K4M4M2M3M4L3M3N4K8H7I8H7I8IcUW2"}, "label": "woman in red cardigan behind young girl"}]}
{"id": 336242, "image": "COCO_train2014_000000336242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl is standing in the room with her arms outstretched\"."}], "task": "refering", "answer_template": "The \"a little girl is standing in the room with her arms outstretched\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 91, 92, 93, 108, 109, 110, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 158, 159, 160, 161, 162, 169, 175, 176, 177, 178, 179, 193, 194, 195, 210, 211, 212, 227, 228, 229, 244, 245, 246, 261, 262, 263, 278, 279, 280, 295, 296, 297], "bbox": [0.014979166666666667, 0.2, 1.0, 0.7595468750000001], "iscrowd": 0, "area": 42439.12950000001, "rle": {"size": [640, 480], "counts": "Yd41lc02N3N1N3M2N3N2M2N3M2N3N1N3M2N3N1N3M2010O01000O010O100O010O10O0100O01000O10O0100O010O10O0100O10O01000O0100O100O100O100O100O100O1000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000001O01O0000000010O00000001O0001O01O0000001O001O00001O001O00001O001O0000001O003M4L4L4L4L4L4L4L4L4L4L4L4L5K4L4L4mN_L_Ae3V>gLeA]3P>oLkAU3]=dM^Bb2[=dM`Be2U=bMfBg2P=^2F9G:F:FY2fM5L4M3M3M3L2O00000000000000000O2O000000001O000000001N1000J6E;E;D=[Od0fNZ1G:4K6J5L5J6K4K6J5L5J5K6K4L5K5K4L8H7I7JN3M3M2N3M2N3M2N3M2M3M4mKWFhMl9R2UGPMn8i2SHZLQ8_3\\4J6J7H7J6J7I6I7J7I5J5L4L4L5K4N2M4M2M3N2N2M3N2M3N2N2M3N2N2O10000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000001O0000000000000001O000001O0000000001O000000000001O01O00000000000000001O01O00000000000001O00000001O00000001O000000000000010O00000000000000001OO101O000O1000000O1000000O101O000O1000000O10000O10001O0O1000000O1000000O1000001N1000000O10000OJ7B>A?BZJ"}, "label": "a little girl is standing in the room with her arms outstretched"}]}
{"id": 336242, "image": "COCO_train2014_000000336242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in ivory sweater , with eggplant colored pants and hot pink socks\"."}], "task": "refering", "answer_template": "The \"a girl in ivory sweater , with eggplant colored pants and hot pink socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 91, 92, 93, 108, 109, 110, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 158, 159, 160, 161, 162, 169, 175, 176, 177, 178, 179, 193, 194, 195, 210, 211, 212, 227, 228, 229, 244, 245, 246, 261, 262, 263, 278, 279, 280, 295, 296, 297], "bbox": [0.014979166666666667, 0.2, 1.0, 0.7595468750000001], "iscrowd": 0, "area": 42439.12950000001, "rle": {"size": [640, 480], "counts": "Yd41lc02N3N1N3M2N3N2M2N3M2N3N1N3M2N3N1N3M2010O01000O010O100O010O10O0100O01000O10O0100O010O10O0100O10O01000O0100O100O100O100O100O100O1000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000001O01O0000000010O00000001O0001O01O0000001O001O00001O001O00001O001O0000001O003M4L4L4L4L4L4L4L4L4L4L4L4L5K4L4L4mN_L_Ae3V>gLeA]3P>oLkAU3]=dM^Bb2[=dM`Be2U=bMfBg2P=^2F9G:F:FY2fM5L4M3M3M3L2O00000000000000000O2O000000001O000000001N1000J6E;E;D=[Od0fNZ1G:4K6J5L5J6K4K6J5L5J5K6K4L5K5K4L8H7I7JN3M3M2N3M2N3M2N3M2M3M4mKWFhMl9R2UGPMn8i2SHZLQ8_3\\4J6J7H7J6J7I6I7J7I5J5L4L4L5K4N2M4M2M3N2N2M3N2M3N2N2M3N2N2O10000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000001O0000000000000001O000001O0000000001O000000000001O01O00000000000000001O01O00000000000001O00000001O00000001O000000000000010O00000000000000001OO101O000O1000000O1000000O101O000O1000000O10000O10001O0O1000000O1000000O1000001N1000000O10000OJ7B>A?BZJ"}, "label": "a girl in ivory sweater , with eggplant colored pants and hot pink socks"}]}
{"id": 24948, "image": "COCO_train2014_000000024948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baked goods in a black pot\"."}], "task": "refering", "answer_template": "The \"baked goods in a black pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 214, 215, 216], "bbox": [0.206734375, 0.22054166666666666, 0.519453125, 0.5437916666666667], "iscrowd": 0, "area": 23732.089, "rle": {"size": [480, 640], "counts": "aQn12l>=B>B>B5L4K5K4M4K5L3O2M3M2O2M3M2O2M3M2O2M2O0O1O2O0O1O2O0O1O2O0O1O2O0O101N1O101N1O101N1O101N1O101N100O2N100O2O0O1O2O0O101N1O101N100O2N100O2O0O1O2O00001O000O2O0000000O1000O100000O10000000000O1000O100000O100000000O100000O10000000001O000000000O2O00000000001O000000000O2O000O100O101N1O2O0O2O1N101N101N101N101N1O2O1N101N101N101N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2M3N2M3N2N2M3N2M3N2M3N3M3L4M3L4M4K4M3M3LgV`4"}, "label": "baked goods in a black pot"}]}
{"id": 24948, "image": "COCO_train2014_000000024948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bread in a stockpot\"."}], "task": "refering", "answer_template": "The \"a bread in a stockpot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 214, 215, 216], "bbox": [0.206734375, 0.22054166666666666, 0.519453125, 0.5437916666666667], "iscrowd": 0, "area": 23732.089, "rle": {"size": [480, 640], "counts": "aQn12l>=B>B>B5L4K5K4M4K5L3O2M3M2O2M3M2O2M3M2O2M2O0O1O2O0O1O2O0O1O2O0O1O2O0O101N1O101N1O101N1O101N1O101N100O2N100O2O0O1O2O0O101N1O101N100O2N100O2O0O1O2O00001O000O2O0000000O1000O100000O10000000000O1000O100000O100000000O100000O10000000001O000000000O2O00000000001O000000000O2O000O100O101N1O2O0O2O1N101N101N101N101N1O2O1N101N101N101N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2M3N2M3N2N2M3N2M3N2M3N3M3L4M3L4M4K4M3M3LgV`4"}, "label": "a bread in a stockpot"}]}
{"id": 573815, "image": "COCO_train2014_000000573815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black computer monitor\"."}], "task": "refering", "answer_template": "The \"black computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 30, 31, 43, 44, 45, 46, 47, 48, 49, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 134, 135, 136, 137], "bbox": [0.38832, 0.0764, 0.7263, 0.573025], "iscrowd": 0, "area": 28117.177549999993, "rle": {"size": [400, 500], "counts": "oi[25[<d0\\Oc0]Od0\\Oc0]Od0\\Oc0]Ob0^O5K6J5K3M0000000000000000001O00000000000ZIZKb5f4]J[Kc5e4]J[Kc5e4]J[Kc5e4]J[Kc5f4\\JZKd5f4\\JZKd5f4\\JZKd5f4[J[Ke5e4[J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4[J]Kf5b4ZJ^Kf5c4YJ]Kg5c4YJ^Kf5b4ZJ^Kf5b4ZJ^Kf5b4ZJ^Kf5b4YJ_Kg5a4YJ_Kg5a4YJ_Kg5d4VJ\\Kj5h4RJXKn5k4oIUKQ6n4kISKU6P5hIPKX6d500000O10000000000000000000000O1000000000000000000000000O1000001O000000000000000O1SOfIdKZ6S4oImKQ6S4oImKQ6S4oImKQ6S4oImKQ6R4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJlKP6T4PJlKP6S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5R4RJnKn5R4RJnKn5R4SJmKm5S4SJmKm5S4SJlKn5T4RJlKn5T4RJlKn5T4RJlKn5S4SJmKm5S4SJmKm5S4X10000000000O101O00000000000O100000000000000O1000000000000O100000000000000O100000^_e1"}, "label": "black computer monitor"}]}
{"id": 573815, "image": "COCO_train2014_000000573815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer screen with a cartoon ninja on it with a glass cup in front of it\"."}], "task": "refering", "answer_template": "The \"a computer screen with a cartoon ninja on it with a glass cup in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 30, 31, 43, 44, 45, 46, 47, 48, 49, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 134, 135, 136, 137], "bbox": [0.38832, 0.0764, 0.7263, 0.573025], "iscrowd": 0, "area": 28117.177549999993, "rle": {"size": [400, 500], "counts": "oi[25[<d0\\Oc0]Od0\\Oc0]Od0\\Oc0]Ob0^O5K6J5K3M0000000000000000001O00000000000ZIZKb5f4]J[Kc5e4]J[Kc5e4]J[Kc5e4]J[Kc5f4\\JZKd5f4\\JZKd5f4\\JZKd5f4[J[Ke5e4[J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4\\J\\Kd5d4[J]Kf5b4ZJ^Kf5c4YJ]Kg5c4YJ^Kf5b4ZJ^Kf5b4ZJ^Kf5b4ZJ^Kf5b4YJ_Kg5a4YJ_Kg5a4YJ_Kg5d4VJ\\Kj5h4RJXKn5k4oIUKQ6n4kISKU6P5hIPKX6d500000O10000000000000000000000O1000000000000000000000000O1000001O000000000000000O1SOfIdKZ6S4oImKQ6S4oImKQ6S4oImKQ6S4oImKQ6R4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJlKP6T4PJlKP6S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5S4QJmKo5R4RJnKn5R4RJnKn5R4SJmKm5S4SJmKm5S4SJlKn5T4RJlKn5T4RJlKn5T4RJlKn5S4SJmKm5S4SJmKm5S4X10000000000O101O00000000000O100000000000000O1000000000000O100000000000000O100000^_e1"}, "label": "a computer screen with a cartoon ninja on it with a glass cup in front of it"}]}
{"id": 369019, "image": "COCO_train2014_000000369019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pitcher in white\"."}], "task": "refering", "answer_template": "The \"the pitcher in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 254, 255, 256, 276, 277, 278, 279, 280, 300, 301, 302, 323, 324, 325, 345, 346, 347, 348, 368, 369, 370, 371, 392, 393, 394], "bbox": [0.028234375, 0.55431640625, 0.18429687499999997, 0.9977343750000001], "iscrowd": 0, "area": 12689.738199999998, "rle": {"size": [512, 640], "counts": "R[9:d?5K5L2M3M2O2YCAi9a0oEHo9:jENR:4aE<[:HPEl0l:ZOjDo0R;UOiDP1R;WOfDn0V;Q2M4K4M;F8G1O1O1O1O1O1000000O101OO001N2PO]F`Kf9[4Q1L5M3M3M2N3NVOdLTEY3m:nLmDP3U;k033M7I8G3N2N3M2N2M5L5K4bFoIe8[6iFVJn8^61O001O000010O0001O00`0aI_F_5g9YJ[Fg5o9001O0001bN_1H9POR1iNX1G6I5L3L5K4M4O0O1O100O1O1O1O1N2O1O3M3M3M3M3K5LodT8"}, "label": "the pitcher in white"}]}
{"id": 369019, "image": "COCO_train2014_000000369019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pitcher\"."}], "task": "refering", "answer_template": "The \"the pitcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 254, 255, 256, 276, 277, 278, 279, 280, 300, 301, 302, 323, 324, 325, 345, 346, 347, 348, 368, 369, 370, 371, 392, 393, 394], "bbox": [0.028234375, 0.55431640625, 0.18429687499999997, 0.9977343750000001], "iscrowd": 0, "area": 12689.738199999998, "rle": {"size": [512, 640], "counts": "R[9:d?5K5L2M3M2O2YCAi9a0oEHo9:jENR:4aE<[:HPEl0l:ZOjDo0R;UOiDP1R;WOfDn0V;Q2M4K4M;F8G1O1O1O1O1O1000000O101OO001N2PO]F`Kf9[4Q1L5M3M3M2N3NVOdLTEY3m:nLmDP3U;k033M7I8G3N2N3M2N2M5L5K4bFoIe8[6iFVJn8^61O001O000010O0001O00`0aI_F_5g9YJ[Fg5o9001O0001bN_1H9POR1iNX1G6I5L3L5K4M4O0O1O100O1O1O1O1N2O1O3M3M3M3M3K5LodT8"}, "label": "the pitcher"}]}
{"id": 369019, "image": "COCO_train2014_000000369019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red and blue baseball jersey waiting to hit the baseball\"."}], "task": "refering", "answer_template": "The \"a man in a red and blue baseball jersey waiting to hit the baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 141, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 405, 406, 407, 408, 409, 410, 411, 412, 413], "bbox": [0.129734375, 0.14638671875, 1.0, 0.986484375], "iscrowd": 0, "area": 131727.89500000002, "rle": {"size": [512, 640], "counts": "\\gY1U1d>o0QO:F:G9F;I6M3M4K4M3M3L2O0O101O0O10000O101O0O10000010O001O001O010O001O00100O001O000010O000001O01O010O1O100O1O101N1O1O010O1O010O1O10O01O1O010O100O00100O00100O10O01O10O01O100O00100O010O1O10O01O10O0100O1O010O1O010O1O10O01O10O01O1O10O01O10O01O100O00100O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O001N101O000010O01O00010O0010O0001O010O00010O001O010O00010O001O01O01O010O000010N1O1O2N1O1O1O2N1O101N10000O2O00000O2O000O10000O0100000O10O10O1000O10O1000O010000O10O1000O10O10O1000O10O10000O10000O1000000OWMRKbJn4[4\\L^Kd3a4eLWKZ3i4oLoJQ3P5XMhJh2W5aMaJ_2^5eM_JZ2`5jM^JV2a5lM]JU2b5nM\\JQ2d5RNZJo1d5SN[Jn1c5UN[Jl1b5XN\\Ji1b5YN]Jg1b5\\N\\Je1b5]N\\Je1b5^N[Jd1c5_NZJc1c5`N[Ja1d5bNYJ`1e5_4N2O1N101N2O1N2O0O2O1N2O6I6K6I7J5J7J5J7J6I6K6I6K6I5L00000000000000000000000000000001O000000000000000000000000000000000000000000001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O1O1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N3M6J6J6J6J6J6J6J6J2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M4L4L5K5K5K5cNeC"}, "label": "a man in a red and blue baseball jersey waiting to hit the baseball"}]}
{"id": 369019, "image": "COCO_train2014_000000369019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the batter wearing a red helmet\"."}], "task": "refering", "answer_template": "The \"the batter wearing a red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 141, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 405, 406, 407, 408, 409, 410, 411, 412, 413], "bbox": [0.129734375, 0.14638671875, 1.0, 0.986484375], "iscrowd": 0, "area": 131727.89500000002, "rle": {"size": [512, 640], "counts": "\\gY1U1d>o0QO:F:G9F;I6M3M4K4M3M3L2O0O101O0O10000O101O0O10000010O001O001O010O001O00100O001O000010O000001O01O010O1O100O1O101N1O1O010O1O010O1O10O01O1O010O100O00100O00100O10O01O10O01O100O00100O010O1O10O01O10O0100O1O010O1O010O1O10O01O10O01O1O10O01O10O01O100O00100O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O001N101O000010O01O00010O0010O0001O010O00010O001O010O00010O001O01O01O010O000010N1O1O2N1O1O1O2N1O101N10000O2O00000O2O000O10000O0100000O10O10O1000O10O1000O010000O10O1000O10O10O1000O10O10000O10000O1000000OWMRKbJn4[4\\L^Kd3a4eLWKZ3i4oLoJQ3P5XMhJh2W5aMaJ_2^5eM_JZ2`5jM^JV2a5lM]JU2b5nM\\JQ2d5RNZJo1d5SN[Jn1c5UN[Jl1b5XN\\Ji1b5YN]Jg1b5\\N\\Je1b5]N\\Je1b5^N[Jd1c5_NZJc1c5`N[Ja1d5bNYJ`1e5_4N2O1N101N2O1N2O0O2O1N2O6I6K6I7J5J7J5J7J6I6K6I6K6I5L00000000000000000000000000000001O000000000000000000000000000000000000000000001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O1O1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N3M6J6J6J6J6J6J6J6J2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M4L4L5K5K5K5cNeC"}, "label": "the batter wearing a red helmet"}]}
{"id": 41340, "image": "COCO_train2014_000000041340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman who is sitting down looking back at the camera\"."}], "task": "refering", "answer_template": "The \"the woman who is sitting down looking back at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 174, 175, 176, 195, 196, 197, 198, 218, 219, 265, 266, 268, 269, 288, 289, 291, 292, 315, 316], "bbox": [0.48459375, 0.3124235807860262, 0.761015625, 0.8661353711790393], "iscrowd": 0, "area": 9738.115999999996, "rle": {"size": [458, 640], "counts": "ilZ42R>7M3O1N101O1O1O1N2O1O1O001O0O2O001O001N10001O0000000000000000O1O10O010O0kD\\Oe8e0YG]Of8c0YG@e8a0YGBe8>[GDd8<ZGGd89]GGb89_GH`88`GI^87bGK]85cGL\\83eGMZ84fGMY82gG0X80hG1V8OkG2T8MmG4R8LnG4R8KoG6l5VOiIe0;6f5AhI8c08_5MfIKk09^51`IGS19[55\\IDY17[58WIB_16Z58UIDb13X5;SICf12W5;QIDi11V5<nHEl1OU5=mHEo1NT5>jHFR2LT5>hHGU2KS5?eHHX2IR5`0dHH[2GR5b0`HI^2DS5c0]HJa2CQ5S2oJlMR5S2oJkMS5T2nJjMS5W2nJXMb5g2Q300O010O10O0100O00100O1N3N1O1O1O2M2O2N2M3M3L4M3M31O2NO10a0^O100O100O10O01O1O1O1O1O001O1O1O1O1O001ORFnNe6Q1[ITOa6k0_IZO^6d0bIAb65_IOf6G[I>i6XOXIm0P:5K5K5K5K4L5K5K1O000000O100000000O10001O0O10004K7J6I6K6I3N2M2O1N3N1N3N1O1N3N1N2O1N101N1N3N1NSRT2"}, "label": "the woman who is sitting down looking back at the camera"}]}
{"id": 41340, "image": "COCO_train2014_000000041340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl looking back and smiling\"."}], "task": "refering", "answer_template": "The \"the girl looking back and smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 174, 175, 176, 195, 196, 197, 198, 218, 219, 265, 266, 268, 269, 288, 289, 291, 292, 315, 316], "bbox": [0.48459375, 0.3124235807860262, 0.761015625, 0.8661353711790393], "iscrowd": 0, "area": 9738.115999999996, "rle": {"size": [458, 640], "counts": "ilZ42R>7M3O1N101O1O1O1N2O1O1O001O0O2O001O001N10001O0000000000000000O1O10O010O0kD\\Oe8e0YG]Of8c0YG@e8a0YGBe8>[GDd8<ZGGd89]GGb89_GH`88`GI^87bGK]85cGL\\83eGMZ84fGMY82gG0X80hG1V8OkG2T8MmG4R8LnG4R8KoG6l5VOiIe0;6f5AhI8c08_5MfIKk09^51`IGS19[55\\IDY17[58WIB_16Z58UIDb13X5;SICf12W5;QIDi11V5<nHEl1OU5=mHEo1NT5>jHFR2LT5>hHGU2KS5?eHHX2IR5`0dHH[2GR5b0`HI^2DS5c0]HJa2CQ5S2oJlMR5S2oJkMS5T2nJjMS5W2nJXMb5g2Q300O010O10O0100O00100O1N3N1O1O1O2M2O2N2M3M3L4M3M31O2NO10a0^O100O100O10O01O1O1O1O1O001O1O1O1O1O001ORFnNe6Q1[ITOa6k0_IZO^6d0bIAb65_IOf6G[I>i6XOXIm0P:5K5K5K5K4L5K5K1O000000O100000000O10001O0O10004K7J6I6K6I3N2M2O1N3N1N3N1O1N3N1N2O1N101N1N3N1NSRT2"}, "label": "the girl looking back and smiling"}]}
{"id": 262528, "image": "COCO_train2014_000000262528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player in white and red sliding into a base\"."}], "task": "refering", "answer_template": "The \"a player in white and red sliding into a base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 93, 94, 104, 105, 106, 107, 108, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 137, 138, 139, 140, 147, 148, 149, 150, 151, 161, 162, 163, 173, 174, 184], "bbox": [0.3713836477987421, 0.4621875, 0.9370125786163522, 0.9542916666666666], "iscrowd": 0, "area": 14999.804999999998, "rle": {"size": [480, 318], "counts": "Sfg11n>2M3M3N2N2N1O2N1O2N0001O00001O00001O0000000000000000O1O001O1O1O1O001O1O1O1\\Od0M2N4M2M3PNgNXF[1e9kNWFV1f9oNWFR1i9POTFQ1m9POPFQ1P:QOmEP1T:QOjEn0X:SOfEl0[:VOcEi0_:XO_Eg0b:[O\\Ed0f:]OYE`0i:CTE<n:EPE:Q;HmD7U;JiD5X;4`DL`;j1O2O001N101O10001N100O10O00100O10bEfLT9Z3aFQM`9o2^FSMb9l2]FVMc9j2ZFXMh9f2WF\\MU:X2iEjMX:U2eEnM]:P2aEQNa:n1]ETNe:S31O00O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O100O1O1OgKUF^3j9aLXF_3g9aL[F_3c9aL^F_3a9`LaF`3^9`LdF_3[9aLfF_3Y9aLTGS3k8mLeGd2Z8\\MhGd2V8]MjGc2U8]MmGb2R8^MoGn2d7RM^Hn2`7RMaHn2^7RMdHn2Z7SMgHk2Y7UMhH4dN_1d8]NjHKkNf1[8`NRJ^1n5cNSJ[1m5eNVJW1k5jNVJS1l5mNWJn0l5QOXJi0k5VOYJd0j5[O[J=i5BYJ:i5FYJ6i5JYJ2i5NZJMh53ZJIh57ZJEh5;[J_Oh5a0ZJ[Oh5f0Z41N2N1O001O001O001O001O00O1N2N2N2N2N2N3M4L3M3M3M4LWR9"}, "label": "a player in white and red sliding into a base"}]}
{"id": 262528, "image": "COCO_train2014_000000262528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player running into a base with his left leg up\"."}], "task": "refering", "answer_template": "The \"a baseball player running into a base with his left leg up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 93, 94, 104, 105, 106, 107, 108, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 137, 138, 139, 140, 147, 148, 149, 150, 151, 161, 162, 163, 173, 174, 184], "bbox": [0.3713836477987421, 0.4621875, 0.9370125786163522, 0.9542916666666666], "iscrowd": 0, "area": 14999.804999999998, "rle": {"size": [480, 318], "counts": "Sfg11n>2M3M3N2N2N1O2N1O2N0001O00001O00001O0000000000000000O1O001O1O1O1O001O1O1O1\\Od0M2N4M2M3PNgNXF[1e9kNWFV1f9oNWFR1i9POTFQ1m9POPFQ1P:QOmEP1T:QOjEn0X:SOfEl0[:VOcEi0_:XO_Eg0b:[O\\Ed0f:]OYE`0i:CTE<n:EPE:Q;HmD7U;JiD5X;4`DL`;j1O2O001N101O10001N100O10O00100O10bEfLT9Z3aFQM`9o2^FSMb9l2]FVMc9j2ZFXMh9f2WF\\MU:X2iEjMX:U2eEnM]:P2aEQNa:n1]ETNe:S31O00O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O100O1O1OgKUF^3j9aLXF_3g9aL[F_3c9aL^F_3a9`LaF`3^9`LdF_3[9aLfF_3Y9aLTGS3k8mLeGd2Z8\\MhGd2V8]MjGc2U8]MmGb2R8^MoGn2d7RM^Hn2`7RMaHn2^7RMdHn2Z7SMgHk2Y7UMhH4dN_1d8]NjHKkNf1[8`NRJ^1n5cNSJ[1m5eNVJW1k5jNVJS1l5mNWJn0l5QOXJi0k5VOYJd0j5[O[J=i5BYJ:i5FYJ6i5JYJ2i5NZJMh53ZJIh57ZJEh5;[J_Oh5a0ZJ[Oh5f0Z41N2N1O001O001O001O001O00O1N2N2N2N2N2N3M4L3M3M3M4LWR9"}, "label": "a baseball player running into a base with his left leg up"}]}
{"id": 262528, "image": "COCO_train2014_000000262528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man who is jumping in the image\"."}], "task": "refering", "answer_template": "The \"the man who is jumping in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 25, 26, 28, 29, 36, 37, 38, 39, 45, 46, 47, 48, 49, 55, 56, 57, 58, 59, 66, 68, 69, 70, 79, 80, 81, 82, 90, 91, 92, 93, 101, 102, 103, 104, 105, 112, 113, 115, 123, 124, 134, 135, 145], "bbox": [0.04430817610062893, 0.08435416666666667, 0.7189308176100629, 0.786375], "iscrowd": 0, "area": 18136.627400000005, "rle": {"size": [480, 318], "counts": "^g63i>5K5N1O2O1N1O2O1N1O2O0OO100O2O0O101N100O101O00000001O001O000O2O00001O001O000O2O00001O00001N101O00001O004K8I6SElN>Fm3e1WKJLfNj4g1lJe0]OiMe5h1`Jl0CaMk5i1SJT1JXMP6k1hIZ10PMV6l1\\Ia17dL^6[6ZIgIf6Z70O1O001O1O001O1O001O1O1O1O1O1O1O1O1O00I7N2O1O[NQJ\\Jn5g4TKUKk4a4cK\\K[4b4lK[KS4b4TL[Kk3e4ZLXKd3i4_LTK`3n4bLoJ_3P5Y300O1O010001O1O0O2O1O001O1N1012N3L3N2eMXFYOj9<_FCd93eFL]9JkF6W9_OSG?o8ZOXGe0g8ZO[Gf0d8XO`Gg0^8YOdGf0\\8YOgGf0X8YOjGg0U8WOoGg0P8YORHg0m7XOVHg0i7XOZHf0e7ZO]Hf0b7XObHg0]7XOeHg0[7XOhHg0V7YOlHg0S7XOPIf0P7XOTIg0k6XOWIh0g6XO\\If0d6YO^Ig0a6WOcIh0\\6XOeIP1Q6PORJo0m5QOUJo0i5QOXJP1f5oN]JP1a5QO`JP1^5oNeJQ1Y5oNiJQ1T5oNnJQ1Q5oNQKQ1l4oNWKn0j4ROWKl0i4TOZKi0f4XO\\Ke0e4[O]Kb0c4^O`K?`4BbK;_4DdK9\\4HfK5Z4KiK2X4NjKOV41mKLT44nKIR47QLFo3;TLAm3?UL_Oj3a0YL\\Og3e0Z5O01O010O010O00100O0010O01N1O2N1O2N1O2N100000O01000O01000O100O10000O10000O2O1O1N3N2M2N3L3N3L4M2NdTZ1"}, "label": "the man who is jumping in the image"}]}
{"id": 262528, "image": "COCO_train2014_000000262528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player catching the ball\"."}], "task": "refering", "answer_template": "The \"a baseball player catching the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 25, 26, 28, 29, 36, 37, 38, 39, 45, 46, 47, 48, 49, 55, 56, 57, 58, 59, 66, 68, 69, 70, 79, 80, 81, 82, 90, 91, 92, 93, 101, 102, 103, 104, 105, 112, 113, 115, 123, 124, 134, 135, 145], "bbox": [0.04430817610062893, 0.08435416666666667, 0.7189308176100629, 0.786375], "iscrowd": 0, "area": 18136.627400000005, "rle": {"size": [480, 318], "counts": "^g63i>5K5N1O2O1N1O2O1N1O2O0OO100O2O0O101N100O101O00000001O001O000O2O00001O001O000O2O00001O00001N101O00001O004K8I6SElN>Fm3e1WKJLfNj4g1lJe0]OiMe5h1`Jl0CaMk5i1SJT1JXMP6k1hIZ10PMV6l1\\Ia17dL^6[6ZIgIf6Z70O1O001O1O001O1O001O1O1O1O1O1O1O1O1O00I7N2O1O[NQJ\\Jn5g4TKUKk4a4cK\\K[4b4lK[KS4b4TL[Kk3e4ZLXKd3i4_LTK`3n4bLoJ_3P5Y300O1O010001O1O0O2O1O001O1N1012N3L3N2eMXFYOj9<_FCd93eFL]9JkF6W9_OSG?o8ZOXGe0g8ZO[Gf0d8XO`Gg0^8YOdGf0\\8YOgGf0X8YOjGg0U8WOoGg0P8YORHg0m7XOVHg0i7XOZHf0e7ZO]Hf0b7XObHg0]7XOeHg0[7XOhHg0V7YOlHg0S7XOPIf0P7XOTIg0k6XOWIh0g6XO\\If0d6YO^Ig0a6WOcIh0\\6XOeIP1Q6PORJo0m5QOUJo0i5QOXJP1f5oN]JP1a5QO`JP1^5oNeJQ1Y5oNiJQ1T5oNnJQ1Q5oNQKQ1l4oNWKn0j4ROWKl0i4TOZKi0f4XO\\Ke0e4[O]Kb0c4^O`K?`4BbK;_4DdK9\\4HfK5Z4KiK2X4NjKOV41mKLT44nKIR47QLFo3;TLAm3?UL_Oj3a0YL\\Og3e0Z5O01O010O010O00100O0010O01N1O2N1O2N1O2N100000O01000O01000O100O10000O10000O2O1O1N3N2M2N3L3N3L4M2NdTZ1"}, "label": "a baseball player catching the ball"}]}
{"id": 98689, "image": "COCO_train2014_000000098689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing brown color short on the right side\"."}], "task": "refering", "answer_template": "The \"a man wearing brown color short on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 83, 84, 85, 106, 107, 108, 129, 130, 131, 132, 153, 155, 156, 176], "bbox": [0.610484375, 0.1719431279620853, 0.79621875, 0.47708530805687205], "iscrowd": 0, "area": 5231.8049500000025, "rle": {"size": [422, 640], "counts": "nXQ51T=2nM6eFLW9;dFFY9>eFDW9`0gFBX9>gFCY9>eFDZ9<eFFZ9;[FOe91ZF1e90ZFNh93WFLj94UFMk95SFJn98oEIQ::lEEV:>fEA[:e11N2N2O1N5K2O1N10[FTNP8k1QHVNn7i1RHXNo7g1QHYNo7f1QH[No7d1RH\\Nn7c1RH^Nn7a1RH`Nn7_1SHbNl7]1THdNl7\\1THdNl7Z1UHgNk7T1ZHlNg7o0\\HROd7j0_HWOa7e0WGXO48f8;WGEI5P95UGS1k8nNQGV1o8iNoFY1Q9gNmF[1T9eNhF^1X9bNfF`1Z9`NdFb1]9]NaFi03oN\\99ZFc0i0POn8h1XGRNj8m1[GmMh8P2X1O2N1N3O010O0100000000O10O100O2O0O101N2O1O1N2iNcDb0^;UOTE`0k;O1N01000O100O100O010O10000O10O010O0010O01O1eCBm;=RDEn;=oCCQ<h02N2N1O2N2M2O2N2N1O2N2N1O2N2N1Om\\e1"}, "label": "a man wearing brown color short on the right side"}]}
{"id": 98689, "image": "COCO_train2014_000000098689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shirtless man is playing tennis at a court\"."}], "task": "refering", "answer_template": "The \"a shirtless man is playing tennis at a court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 83, 84, 85, 106, 107, 108, 129, 130, 131, 132, 153, 155, 156, 176], "bbox": [0.610484375, 0.1719431279620853, 0.79621875, 0.47708530805687205], "iscrowd": 0, "area": 5231.8049500000025, "rle": {"size": [422, 640], "counts": "nXQ51T=2nM6eFLW9;dFFY9>eFDW9`0gFBX9>gFCY9>eFDZ9<eFFZ9;[FOe91ZF1e90ZFNh93WFLj94UFMk95SFJn98oEIQ::lEEV:>fEA[:e11N2N2O1N5K2O1N10[FTNP8k1QHVNn7i1RHXNo7g1QHYNo7f1QH[No7d1RH\\Nn7c1RH^Nn7a1RH`Nn7_1SHbNl7]1THdNl7\\1THdNl7Z1UHgNk7T1ZHlNg7o0\\HROd7j0_HWOa7e0WGXO48f8;WGEI5P95UGS1k8nNQGV1o8iNoFY1Q9gNmF[1T9eNhF^1X9bNfF`1Z9`NdFb1]9]NaFi03oN\\99ZFc0i0POn8h1XGRNj8m1[GmMh8P2X1O2N1N3O010O0100000000O10O100O2O0O101N2O1O1N2iNcDb0^;UOTE`0k;O1N01000O100O100O010O10000O10O010O0010O01O1eCBm;=RDEn;=oCCQ<h02N2N1O2N2M2O2N2N1O2N2N1O2N2N1Om\\e1"}, "label": "a shirtless man is playing tennis at a court"}]}
{"id": 196998, "image": "COCO_train2014_000000196998.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with the clock sign\"."}], "task": "refering", "answer_template": "The \"the person with the clock sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 25, 26, 27, 28, 43, 44, 45, 46, 62, 63, 64, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 114, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 137, 138, 153, 154, 155, 156, 173, 174, 191, 192, 209, 210, 227, 228, 245, 246, 262, 263, 278, 279, 280, 281, 296, 297, 298, 299, 314, 315, 316, 317, 332, 333, 334, 335, 349, 350, 351, 352, 368], "bbox": [0.3378915662650603, 0.035953125, 0.7133333333333334, 0.8764062499999999], "iscrowd": 0, "area": 39490.70050000001, "rle": {"size": [640, 498], "counts": "gUY31kc06K4K6K4K6K4K5L5J5L5J5L5J5L3L5L3L4M4M20001O00001O001O00001O00001O00001O001O00001O00001O000fMRNbBo1\\=TNaBm1^=VN_Bl1_=VN`Bj1R9nLXJ[1cLh1P9TMZJW1cLf1n8ZM\\JS1cLd1l8`M^Jo0dLb1g8fMcJj0dLa1`8PNiJb0dL_1Y8[NoJ:eL\\1R8fNVK1fLZ1j7oN^KJeLX1c7ZOeK@fLW1[7ElKWOgLT1T70RLoNgLR1m69[LX2\\3oMdLR2X3RNgLo1Y3PNgLQ2Y3oMfLS2Y3lMhLT2X3lMhLT2Y3jMhLV2X3jMhLV2X3iMiLX2V3hMjLX2V3hMjLX2W3fMjLZ2V3fMjLZ2V3eMkL\\2T3dMlL\\2T3cMmL]2T3bMlL^2T3aMmL`2R3`MnL`2R3_MoLa2Q3_MoLa2R3^MnLb2R3]MoLd2P3\\MPMd2P3[MQMe2o2[MQMe2P3YMQMh2n2XMRMh2n2WMSMi2m2WMSMi2m2WMSMi2n2UMSMl2l2TMTMl2l2SMUMm2k2SMUMm2k2RMVMn2k2QMUMP3j2oLWMQ3i2oLWMQ3i2nLXMR3h2nLXMS3h2lLXMT3h2kLYMU3g2cKaIUOi3X5f2^KjISOd3^5b2[KSJPO]3f5`2VKZJnNX3l5^2WK]JfNX3S6[2XK`J_NX3X6X2ZKlNg4S1ZKlNf4U1ZKjNf4V1[KjNd4V1]KiNd4V1]KiNc4W1^KhNb4X1_KhN`4X1aKgN?mJa1[6ZN^N4ZK`1V6iNVNDgKb1o5YOoMTOULa1i5_OQNnNZLb1a5CUNiN^Lb1Y5H[NdN_Lb1S5M^N_NcLb1l41bN[NfLc1c46gNVNiLb1]4:lNQNkLc1V4?oNlMnLd1o3c0TOhMPMd1h3g0DWMhLP2^3gMTKU2Y5aMXLa2U3mM^Ko1`73kLSNhKg1`74bLZNQL`1^75[L`N[LW1[79TLeNdLP1Y7;kKjNPMh0V7>dKoNYMa0S7`0^KTOcM8Q7c0VKZOlM1n6f0oJ^OVNIl6i0hJB`NCi6j0aJHiN[Og6m0YJNTOROd6P1QJ2_OkNa6S1jI6IeN]6X1aI76^N[6]1VI8e0XNU6b1nH:Q1RNQ6g1eH;^1kMo5l1ZH<l1fMj5Q2PH>Z2^Mg5V2gG`0g2WMc5Y6aJdI`5Y6eJeI[5Y6iJdIY5Y6lJdIT5Y6QKdIP5Z6TKdIm4Z6VKcIk4\\6YKaIg4^6_4N1O1O2N1O1O3M>B>A?A?B<C;E;E<D;E;E<D;E<Ec0\\Oc0]OnQi2"}, "label": "the person with the clock sign"}]}
{"id": 196998, "image": "COCO_train2014_000000196998.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy carried a poster\"."}], "task": "refering", "answer_template": "The \"a boy carried a poster\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 25, 26, 27, 28, 43, 44, 45, 46, 62, 63, 64, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 114, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 137, 138, 153, 154, 155, 156, 173, 174, 191, 192, 209, 210, 227, 228, 245, 246, 262, 263, 278, 279, 280, 281, 296, 297, 298, 299, 314, 315, 316, 317, 332, 333, 334, 335, 349, 350, 351, 352, 368], "bbox": [0.3378915662650603, 0.035953125, 0.7133333333333334, 0.8764062499999999], "iscrowd": 0, "area": 39490.70050000001, "rle": {"size": [640, 498], "counts": "gUY31kc06K4K6K4K6K4K5L5J5L5J5L5J5L3L5L3L4M4M20001O00001O001O00001O00001O00001O001O00001O00001O000fMRNbBo1\\=TNaBm1^=VN_Bl1_=VN`Bj1R9nLXJ[1cLh1P9TMZJW1cLf1n8ZM\\JS1cLd1l8`M^Jo0dLb1g8fMcJj0dLa1`8PNiJb0dL_1Y8[NoJ:eL\\1R8fNVK1fLZ1j7oN^KJeLX1c7ZOeK@fLW1[7ElKWOgLT1T70RLoNgLR1m69[LX2\\3oMdLR2X3RNgLo1Y3PNgLQ2Y3oMfLS2Y3lMhLT2X3lMhLT2Y3jMhLV2X3jMhLV2X3iMiLX2V3hMjLX2V3hMjLX2W3fMjLZ2V3fMjLZ2V3eMkL\\2T3dMlL\\2T3cMmL]2T3bMlL^2T3aMmL`2R3`MnL`2R3_MoLa2Q3_MoLa2R3^MnLb2R3]MoLd2P3\\MPMd2P3[MQMe2o2[MQMe2P3YMQMh2n2XMRMh2n2WMSMi2m2WMSMi2m2WMSMi2n2UMSMl2l2TMTMl2l2SMUMm2k2SMUMm2k2RMVMn2k2QMUMP3j2oLWMQ3i2oLWMQ3i2nLXMR3h2nLXMS3h2lLXMT3h2kLYMU3g2cKaIUOi3X5f2^KjISOd3^5b2[KSJPO]3f5`2VKZJnNX3l5^2WK]JfNX3S6[2XK`J_NX3X6X2ZKlNg4S1ZKlNf4U1ZKjNf4V1[KjNd4V1]KiNd4V1]KiNc4W1^KhNb4X1_KhN`4X1aKgN?mJa1[6ZN^N4ZK`1V6iNVNDgKb1o5YOoMTOULa1i5_OQNnNZLb1a5CUNiN^Lb1Y5H[NdN_Lb1S5M^N_NcLb1l41bN[NfLc1c46gNVNiLb1]4:lNQNkLc1V4?oNlMnLd1o3c0TOhMPMd1h3g0DWMhLP2^3gMTKU2Y5aMXLa2U3mM^Ko1`73kLSNhKg1`74bLZNQL`1^75[L`N[LW1[79TLeNdLP1Y7;kKjNPMh0V7>dKoNYMa0S7`0^KTOcM8Q7c0VKZOlM1n6f0oJ^OVNIl6i0hJB`NCi6j0aJHiN[Og6m0YJNTOROd6P1QJ2_OkNa6S1jI6IeN]6X1aI76^N[6]1VI8e0XNU6b1nH:Q1RNQ6g1eH;^1kMo5l1ZH<l1fMj5Q2PH>Z2^Mg5V2gG`0g2WMc5Y6aJdI`5Y6eJeI[5Y6iJdIY5Y6lJdIT5Y6QKdIP5Z6TKdIm4Z6VKcIk4\\6YKaIg4^6_4N1O1O2N1O1O3M>B>A?A?B<C;E;E<D;E;E<D;E<Ec0\\Oc0]OnQi2"}, "label": "a boy carried a poster"}]}
{"id": 393608, "image": "COCO_train2014_000000393608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked pizza beside a bottle of wine\"."}], "task": "refering", "answer_template": "The \"an uncooked pizza beside a bottle of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 223, 224, 225, 226], "bbox": [0.372953125, 0.3436458333333333, 0.8825312499999999, 0.5770833333333333], "iscrowd": 0, "area": 20447.462299999996, "rle": {"size": [480, 640], "counts": "mW`36i>7I4L1O1O2N1O1WO]OTCd0m<]OPCd0P=^OmBc0T=_OiBa0X=@eBa0[=AbB`0_=A^B`0c=>001O001O1O00100O001O100O001O01O0000001O0000001O0000001N10000O2O00001N10000O2O000O101O000O101O0O10001O0O10001N10000O2O0000001O00001O0000001O0000001O0000001O0000001N100O101O0O10001N10000O01O1O001O1O1O1O00100O1O001O1O1O1O02O0000000O100O1O1O1O1O2N1O1O1O1N2O1O1O1O1O2N1O11O2N1O2N1O00000010O0000000001O0000O101N100O100O2O0000001O01O01O0000001O0001O10O01O00100O0010O01O0001N1O1O1O2M2O1O2N1O1O1010O1O00100O1O1O010O1O1O1O1O000000001O000000000000O1N2O001O1001O1O001O1O00>B000000000000O1N2O2N110O010O010O02O2M2O1N2O1N010O010O01O010O010O010O0010O01O001O010OO100O1O2N1O1O1O101O0O100O1O2N1O100O1O1O1O1O1O1O1O100O1O1O2N101N3M4M3L4L4Lc0^OZmR1"}, "label": "an uncooked pizza beside a bottle of wine"}]}
{"id": 393608, "image": "COCO_train2014_000000393608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked pizza in between a similar pizza nad a bottle of beer\"."}], "task": "refering", "answer_template": "The \"an uncooked pizza in between a similar pizza nad a bottle of beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 223, 224, 225, 226], "bbox": [0.372953125, 0.3436458333333333, 0.8825312499999999, 0.5770833333333333], "iscrowd": 0, "area": 20447.462299999996, "rle": {"size": [480, 640], "counts": "mW`36i>7I4L1O1O2N1O1WO]OTCd0m<]OPCd0P=^OmBc0T=_OiBa0X=@eBa0[=AbB`0_=A^B`0c=>001O001O1O00100O001O100O001O01O0000001O0000001O0000001N10000O2O00001N10000O2O000O101O000O101O0O10001O0O10001N10000O2O0000001O00001O0000001O0000001O0000001O0000001N100O101O0O10001N10000O01O1O001O1O1O1O00100O1O001O1O1O1O02O0000000O100O1O1O1O1O2N1O1O1O1N2O1O1O1O1O2N1O11O2N1O2N1O00000010O0000000001O0000O101N100O100O2O0000001O01O01O0000001O0001O10O01O00100O0010O01O0001N1O1O1O2M2O1O2N1O1O1010O1O00100O1O1O010O1O1O1O1O000000001O000000000000O1N2O001O1001O1O001O1O00>B000000000000O1N2O2N110O010O010O02O2M2O1N2O1N010O010O01O010O010O010O0010O01O001O010OO100O1O2N1O1O1O101O0O100O1O2N1O100O1O1O1O1O1O1O1O100O1O1O2N101N3M4M3L4L4Lc0^OZmR1"}, "label": "an uncooked pizza in between a similar pizza nad a bottle of beer"}]}
{"id": 393608, "image": "COCO_train2014_000000393608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beer of the brand james boag ' s sitting on a counter\"."}], "task": "refering", "answer_template": "The \"a beer of the brand james boag ' s sitting on a counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 29, 30, 31, 52, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 168, 169, 170, 191], "bbox": [0.277828125, 0.0012083333333333332, 0.431328125, 0.480625], "iscrowd": 0, "area": 15034.5474, "rle": {"size": [480, 640], "counts": "Qac27e>:G9F9G:YMdNZGf1`8nNiF[1Q9XOXFR1b9AgEh0T:X2J6J6J6J5K6J6J6J6J5K6J6J6J3M00000001O000001O000000000000000000001O0000O100O1@VHiIl7U6^HaIf7\\6c0K5K5K4M4M3L4M2M4M3O1O0L5K5K3M2N3M2M4VObE_LO6a:Z3jEdLW:\\3jEbLX:\\3jEaLX:_3iE_LY:_3e001N101N10002O1N2N3M4L6J6J6J6I7I7J6I7I7I7I7I7I7EU^Z5"}, "label": "a beer of the brand james boag ' s sitting on a counter"}]}
{"id": 393608, "image": "COCO_train2014_000000393608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of james boag ' s\"."}], "task": "refering", "answer_template": "The \"a bottle of james boag ' s\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 29, 30, 31, 52, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 168, 169, 170, 191], "bbox": [0.277828125, 0.0012083333333333332, 0.431328125, 0.480625], "iscrowd": 0, "area": 15034.5474, "rle": {"size": [480, 640], "counts": "Qac27e>:G9F9G:YMdNZGf1`8nNiF[1Q9XOXFR1b9AgEh0T:X2J6J6J6J5K6J6J6J6J5K6J6J6J3M00000001O000001O000000000000000000001O0000O100O1@VHiIl7U6^HaIf7\\6c0K5K5K4M4M3L4M2M4M3O1O0L5K5K3M2N3M2M4VObE_LO6a:Z3jEdLW:\\3jEbLX:\\3jEaLX:_3iE_LY:_3e001N101N10002O1N2N3M4L6J6J6J6I7I7J6I7I7I7I7I7I7EU^Z5"}, "label": "a bottle of james boag ' s"}]}
{"id": 393608, "image": "COCO_train2014_000000393608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza closest to camera\"."}], "task": "refering", "answer_template": "The \"pizza closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.43820833333333337, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 131166.80735000002, "rle": {"size": [480, 640], "counts": "Y=b1l;b1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O0O2O1N2N2O1N2O1N2O1N2O10000000000O1000000000000000000N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O100O1O1O1O1O100O1O1O1O1O1O13M4L3M1ON2N2O1N2N2N200O1O100O1O100O1O1O100O100O100O100O100O100O100O10000O100O100O100O100O10000O1O1O1O1O1O1O1QOdIYI]6g6gITIZ6l6kInHV6Q7lImHU6S7kIlHV6T7jIkHW6U7iIjHX6U7iIjHX6V7gIjHZ6V7fIiH[6W7eIhH\\6X7dIgH]6X7dIgH]6Y7cIgH]6Y7cIgH]6Y7bIgH_6X7bIhH^6X7bIgH_6Y7aIgH_6Y7aIgH_6Y7aIfH`6Y7aIgH_6Y7`IgHa6Y7_IgHa6Y7_IgHa6Y7_IfHb6Y7_IgHa6Y7_IfHb6Z7^IfHb6Z7]IgHc6Y7]IfHd6Z7\\IfHd6Z7a00000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000O1000000000000000O100000000000]OoHWIQ7i6c0001O001O00003M:F9G000000000000000000000000000000000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O0000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O00000000001OO1O100O1O100O1O100O1O100O1O1O100O1O10000000000000000001O00000000000000000000001O000000000000000000001O000000000000001O0000001O0000001O0000001O0000001O00001N1000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O00000000000000000000000000000000O10000000000000000000000000000000000000000001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O0O2O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1OhE"}, "label": "pizza closest to camera"}]}
{"id": 393608, "image": "COCO_train2014_000000393608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate next to the camera\"."}], "task": "refering", "answer_template": "The \"the plate next to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.43820833333333337, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 131166.80735000002, "rle": {"size": [480, 640], "counts": "Y=b1l;b1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O0O2O1N2N2O1N2O1N2O1N2O10000000000O1000000000000000000N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O100O1O1O1O1O100O1O1O1O1O1O13M4L3M1ON2N2O1N2N2N200O1O100O1O100O1O1O100O100O100O100O100O100O100O10000O100O100O100O100O10000O1O1O1O1O1O1O1QOdIYI]6g6gITIZ6l6kInHV6Q7lImHU6S7kIlHV6T7jIkHW6U7iIjHX6U7iIjHX6V7gIjHZ6V7fIiH[6W7eIhH\\6X7dIgH]6X7dIgH]6Y7cIgH]6Y7cIgH]6Y7bIgH_6X7bIhH^6X7bIgH_6Y7aIgH_6Y7aIgH_6Y7aIfH`6Y7aIgH_6Y7`IgHa6Y7_IgHa6Y7_IgHa6Y7_IfHb6Y7_IgHa6Y7_IfHb6Z7^IfHb6Z7]IgHc6Y7]IfHd6Z7\\IfHd6Z7a00000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000O1000000000000000O100000000000]OoHWIQ7i6c0001O001O00003M:F9G000000000000000000000000000000000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O0000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O00000000001OO1O100O1O100O1O100O1O100O1O1O100O1O10000000000000000001O00000000000000000000001O000000000000000000001O000000000000001O0000001O0000001O0000001O0000001O00001N1000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O00000000000000000000000000000000O10000000000000000000000000000000000000000001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O0O2O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1OhE"}, "label": "the plate next to the camera"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with hite cushion in which no one is there to sit\"."}], "task": "refering", "answer_template": "The \"a chair with hite cushion in which no one is there to sit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 38, 39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 84, 85, 86, 87, 106, 107, 109], "bbox": [0.6365624999999999, 0.0, 0.8989375, 0.31372916666666667], "iscrowd": 0, "area": 12895.2581, "rle": {"size": [480, 640], "counts": "b\\o54i>2N1O1O1O1O1O2N2N2N2]L_OfHb0\\7B]H?e7ETH;o7JiG7Y8M`G3c81VG0l84mFNT97dFJ^9:[FGg9=RFDP:b0hE^OZ:h0]EZOd:_22M3M3N2M3M3N2M3M3M3O11O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O0`EoLQ9Q3jFTMV9m2dFXM\\9h2_F]Ma9d2YFaMg9_2TFfMl9Z2oEkMQ:V2iElMZ:T2bElMb:U34L4M3L4L4L4L5L3L4M300O10000O100O10000O10000O10000O100O10000O1N2N2N2N2N2O1O1O1O1O1O1O1O100O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O9G`0@X^n0"}, "label": "a chair with hite cushion in which no one is there to sit"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a steel chair near a lady and back of the man\"."}], "task": "refering", "answer_template": "The \"a steel chair near a lady and back of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 38, 39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 84, 85, 86, 87, 106, 107, 109], "bbox": [0.6365624999999999, 0.0, 0.8989375, 0.31372916666666667], "iscrowd": 0, "area": 12895.2581, "rle": {"size": [480, 640], "counts": "b\\o54i>2N1O1O1O1O1O2N2N2N2]L_OfHb0\\7B]H?e7ETH;o7JiG7Y8M`G3c81VG0l84mFNT97dFJ^9:[FGg9=RFDP:b0hE^OZ:h0]EZOd:_22M3M3N2M3M3N2M3M3M3O11O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O0`EoLQ9Q3jFTMV9m2dFXM\\9h2_F]Ma9d2YFaMg9_2TFfMl9Z2oEkMQ:V2iElMZ:T2bElMb:U34L4M3L4L4L4L5L3L4M300O10000O100O10000O10000O10000O100O10000O1N2N2N2N2N2O1O1O1O1O1O1O1O100O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O9G`0@X^n0"}, "label": "a steel chair near a lady and back of the man"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mostly full bottle of water with the cap on\"."}], "task": "refering", "answer_template": "The \"a mostly full bottle of water with the cap on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 340, 341, 342, 343, 362, 363, 364, 365, 366, 386, 387, 388], "bbox": [0.77190625, 0.5376458333333333, 0.97921875, 0.9803333333333334], "iscrowd": 0, "area": 17027.030600000013, "rle": {"size": [480, 640], "counts": "moW71m>3M3M3O1K4L5K5K5K5K5K4L5K5K5K4M4K4L4L4L4L5K4L4L4L4L4L5K4L4M4N1N3N1N2O1O2M2O1N2O2M2O1N2O2N1N2O1N3N1N2O1N3N1O1N2O2N100O1O1O100O2N100O1O100O1O100O1O100O1O100O1O1O100O1O2M2N1ON2O1O1O2N1O1N011O3M3M3L4M3M3M3M3M3L5L3M3M3M3M3L4M3M3M3M3M3L4M3M3M4L3M4K4M4L3M4L3L5L3M4kNXCMk<VOWDb0[=G9FVi5"}, "label": "a mostly full bottle of water with the cap on"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman sitting on the chair\"."}], "task": "refering", "answer_template": "The \"the woman sitting on the chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 280, 281, 282], "bbox": [0.104265625, 0.0, 0.4843125, 0.7780208333333333], "iscrowd": 0, "area": 59981.67535, "rle": {"size": [480, 640], "counts": "Wao0>_>f0[Od0[Of0[O?@:G:F:E:G2M3N2M4M2M3N2M4M2N2M3N3M2O1O1O2O0O1O1O2N100O1O2N100O1O2N1O100O2N1RIXJc4i5\\KYJc4g5[K]Jb4e5[K^Jc4c5[KaJb4`5\\KcJb4^5\\KfJb4[5[KhJc4Y5[KkJb4V5\\KnJa4S5]KPKa4R5]KQKa4o4]KTKa4m4[KYKb4h4YK^Ke4d4VKbKh4^4SKhKk4Y4QKmKl4T4hJZLU5j6N3N1N3M2N3N1N3M3M2M2N2N1O1O2N1O1QO`E^Ma:c2cEWM_:i2dERM]:o2gEkLZ:V3iEeLX:\\3lE^LV:b3c01O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1ZGgJd7Z5YHjJe7W5XHlJh7T5THPKk7Q5RHRKm7o4PHTKP8l4mGWKR8j4kGYKT8h4iG[KV8f4gG]KY8c4dG`K[8c5O100O2N100O1O1O100O1O100O1O1O101N1O2O1N2N1O2O1N2N2O1N2N2O0O2N2N2O1N3M4M4K5K5K4M4K5K5J5K6J6I7J5J7J6J6I6K6J6J6K:F:F;E;E7I3M4L3M4L3M3UObBOb=LbB1a=KcB2`=JdB3`=HdB5_=GeB6U>L4MQ_j4"}, "label": "the woman sitting on the chair"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding a fork\"."}], "task": "refering", "answer_template": "The \"the woman holding a fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 280, 281, 282], "bbox": [0.104265625, 0.0, 0.4843125, 0.7780208333333333], "iscrowd": 0, "area": 59981.67535, "rle": {"size": [480, 640], "counts": "Wao0>_>f0[Od0[Of0[O?@:G:F:E:G2M3N2M4M2M3N2M4M2N2M3N3M2O1O1O2O0O1O1O2N100O1O2N100O1O2N1O100O2N1RIXJc4i5\\KYJc4g5[K]Jb4e5[K^Jc4c5[KaJb4`5\\KcJb4^5\\KfJb4[5[KhJc4Y5[KkJb4V5\\KnJa4S5]KPKa4R5]KQKa4o4]KTKa4m4[KYKb4h4YK^Ke4d4VKbKh4^4SKhKk4Y4QKmKl4T4hJZLU5j6N3N1N3M2N3N1N3M3M2M2N2N1O1O2N1O1QO`E^Ma:c2cEWM_:i2dERM]:o2gEkLZ:V3iEeLX:\\3lE^LV:b3c01O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1ZGgJd7Z5YHjJe7W5XHlJh7T5THPKk7Q5RHRKm7o4PHTKP8l4mGWKR8j4kGYKT8h4iG[KV8f4gG]KY8c4dG`K[8c5O100O2N100O1O1O100O1O100O1O1O101N1O2O1N2N1O2O1N2N2O1N2N2O0O2N2N2O1N3M4M4K5K5K4M4K5K5J5K6J6I7J5J7J6J6I6K6J6J6K:F:F;E;E7I3M4L3M4L3M3UObBOb=LbB1a=KcB2`=JdB3`=HdB5_=GeB6U>L4MQ_j4"}, "label": "the woman holding a fork"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the red bracelet to the right of the woman and the plate of food\"."}], "task": "refering", "answer_template": "The \"the man with the red bracelet to the right of the woman and the plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 135, 136, 137], "bbox": [0.8151718750000001, 0.0, 1.0, 0.34047916666666667], "iscrowd": 0, "area": 11248.664400000001, "rle": {"size": [480, 640], "counts": "Xid7:b>7I8I4M2M3N2N2N2O1N2N2O1NM4O00001O1O100O1O1O1O2N100O1O1O1O100O1O1O1O1O10000100O2N1O1O1O1O2N1O1O1O2N1O7I8I4K4L5K;E8H2N1kD]Ml9e2kEdMS:]2iEgMU:[2gEiMX:Y2fEiMX:X2fEjMX:Y2eEjMY:X2eEiMZ:Z2bEhM\\:_3O2N1N2O1O1N2O1O1N2O100O100O100O1000000000O0100000000000000000000000000000000000000000000000000001lLZGmNd9"}, "label": "the man with the red bracelet to the right of the woman and the plate of food"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wit grey color dress in the right side of the image\"."}], "task": "refering", "answer_template": "The \"man wit grey color dress in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 135, 136, 137], "bbox": [0.8151718750000001, 0.0, 1.0, 0.34047916666666667], "iscrowd": 0, "area": 11248.664400000001, "rle": {"size": [480, 640], "counts": "Xid7:b>7I8I4M2M3N2N2N2O1N2N2O1NM4O00001O1O100O1O1O1O2N100O1O1O1O100O1O1O1O1O10000100O2N1O1O1O1O2N1O1O1O2N1O7I8I4K4L5K;E8H2N1kD]Ml9e2kEdMS:]2iEgMU:[2gEiMX:Y2fEiMX:X2fEjMX:Y2eEjMY:X2eEiMZ:Z2bEhM\\:_3O2N1N2O1O1N2O1O1N2O100O100O100O1000000000O0100000000000000000000000000000000000000000000000000001lLZGmNd9"}, "label": "man wit grey color dress in the right side of the image"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the woman is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the woman is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 138, 139, 140, 149, 162, 163, 172, 173, 185, 186, 187, 188, 209, 210, 211, 232, 233, 234, 235, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304], "bbox": [0.016859375, 0.2561875, 0.56334375, 0.8202291666666668], "iscrowd": 0, "area": 11127.40065, "rle": {"size": [480, 640], "counts": "lX56k>9H9F9G7J0O00100O1N2N2M3N2N2N1N3N2N2N2N2M3N2N2N1O2M3N2N2N2M3N2N2N1O2M3N2N2N2M3N2N2N1O2M3N2N2N1O2M3N2N1ORO[EcMb:Z2cEgM[:T2jEnMQ<0TD2V<]OPDc0V=2N2N2O1N1O2O1N2N2N102M4L3N2M4L3M3N2M4L3N2M3M4L3N2M3M4M2M3MQOUDmNg;Q1]DPOa;n0cDROZ;n0hDSOU;n0lDSOR;l0QESOm:n0TEROj:n0XEQOg:P1[EoNd:P1^EoNa:Q1`EPO^:Q1bEPO]:o0cERO\\:o0dEROZ:n0gEROX:o0gESOX:l0iEUOU:l0kETOT:l0lEVOS:IWELf0^O[O=g:8YEMf0^O]O;b:9]ENc0^OA9^:9`E0b0^OC7Z:9cE2?^OG6U:8gE4>]OJ4P::jE5<]OMOo9=iE7;]O0Io9a0iE98\\Om::lD:8\\Ok:8PE<4\\Om:5QE?3\\Ok:4TE`01[Ok:3WEa0N\\Ol:0XEe0L[Ok:N\\Ef0J[Oj:M^Ei0G[Oj:JbEj0E[OS<f0lCZOU<e0lC[OS<f0lCZOU<e0lC[OS<f0lCZOU<e0lCZOT<g0kCZOU<e0lCZOT<g0kCYOV<f0jC[OV<e0jCZOV<f0jC[OV<e0jCZOW<e0iC[OX<e0hC[OW<e0iC[OX<e0hC[OX<d0hC\\OY<d0gC[OZ<d0fC]OZ<c0fC\\O[<c0eC\\O]<d0cCZO^<f0bCYO`<g0m0O100O10O10O10O0100O010O10O010O0100O001N2O0O2O1N101N2N101N2O0O3N2M]jd15_U[N10O00000001O01O0001O0000010O0000000010O000001O0001O0001O00000001O00000000010O00000000001O00000000001O0000000LkbR4"}, "label": "the chair the woman is sitting in"}]}
{"id": 352651, "image": "COCO_train2014_000000352651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the woman is sitting\"."}], "task": "refering", "answer_template": "The \"the chair the woman is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 138, 139, 140, 149, 162, 163, 172, 173, 185, 186, 187, 188, 209, 210, 211, 232, 233, 234, 235, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304], "bbox": [0.016859375, 0.2561875, 0.56334375, 0.8202291666666668], "iscrowd": 0, "area": 11127.40065, "rle": {"size": [480, 640], "counts": "lX56k>9H9F9G7J0O00100O1N2N2M3N2N2N1N3N2N2N2N2M3N2N2N1O2M3N2N2N2M3N2N2N1O2M3N2N2N2M3N2N2N1O2M3N2N2N1O2M3N2N1ORO[EcMb:Z2cEgM[:T2jEnMQ<0TD2V<]OPDc0V=2N2N2O1N1O2O1N2N2N102M4L3N2M4L3M3N2M4L3N2M3M4L3N2M3M4M2M3MQOUDmNg;Q1]DPOa;n0cDROZ;n0hDSOU;n0lDSOR;l0QESOm:n0TEROj:n0XEQOg:P1[EoNd:P1^EoNa:Q1`EPO^:Q1bEPO]:o0cERO\\:o0dEROZ:n0gEROX:o0gESOX:l0iEUOU:l0kETOT:l0lEVOS:IWELf0^O[O=g:8YEMf0^O]O;b:9]ENc0^OA9^:9`E0b0^OC7Z:9cE2?^OG6U:8gE4>]OJ4P::jE5<]OMOo9=iE7;]O0Io9a0iE98\\Om::lD:8\\Ok:8PE<4\\Om:5QE?3\\Ok:4TE`01[Ok:3WEa0N\\Ol:0XEe0L[Ok:N\\Ef0J[Oj:M^Ei0G[Oj:JbEj0E[OS<f0lCZOU<e0lC[OS<f0lCZOU<e0lC[OS<f0lCZOU<e0lCZOT<g0kCZOU<e0lCZOT<g0kCYOV<f0jC[OV<e0jCZOV<f0jC[OV<e0jCZOW<e0iC[OX<e0hC[OW<e0iC[OX<e0hC[OX<d0hC\\OY<d0gC[OZ<d0fC]OZ<c0fC\\O[<c0eC\\O]<d0cCZO^<f0bCYO`<g0m0O100O10O10O10O0100O010O10O010O0100O001N2O0O2O1N101N2N101N2O0O3N2M]jd15_U[N10O00000001O01O0001O0000010O0000000010O000001O0001O0001O00000001O00000000010O00000000001O00000000001O0000000LkbR4"}, "label": "the chair the woman is sitting"}]}
{"id": 336267, "image": "COCO_train2014_000000336267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird in the middle\"."}], "task": "refering", "answer_template": "The \"the bird in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 241, 242, 243], "bbox": [0.396203125, 0.3557831325301205, 0.84840625, 0.7038072289156627], "iscrowd": 0, "area": 20119.591700000004, "rle": {"size": [415, 640], "counts": "eTW3g0n;>H6L4L3M4M2M3J7G8B>N2N2N2N1N3N1O2O1N1O2O1N1O2O0O2N2O0O2O1O001O1O001O001O1O001O1O001O1O001000O1000O010O10O01O1O00100O010O10O10O010O10O10O0100O100O100O0O101O0O100000000000001O000000000000001O0000001O00001O0000001O0O10001N10001N10000O10001N10000O10001N100O101O1N2O001N2O0N3M3N1N3N2M3N2N101N2O1O001N2O00001N101O000O2O0O101O0O101N100F;M2M3N3N100O1O2N100O1O2N1O100O2O0O10001O00000O101O0000000O10000000001O0O2O001O00001N10000000O1000O10000000000O01000000000O100000O100000O1000000000O01000000000000O10001O00000O10000000000000000O100000000000001O0O1000000000001O1O0O2O1O1O001O1O1O2NU`W1"}, "label": "the bird in the middle"}]}
{"id": 336267, "image": "COCO_train2014_000000336267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle of the three birds\"."}], "task": "refering", "answer_template": "The \"middle of the three birds\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 241, 242, 243], "bbox": [0.396203125, 0.3557831325301205, 0.84840625, 0.7038072289156627], "iscrowd": 0, "area": 20119.591700000004, "rle": {"size": [415, 640], "counts": "eTW3g0n;>H6L4L3M4M2M3J7G8B>N2N2N2N1N3N1O2O1N1O2O1N1O2O0O2N2O0O2O1O001O1O001O001O1O001O1O001O1O001000O1000O010O10O01O1O00100O010O10O10O010O10O10O0100O100O100O0O101O0O100000000000001O000000000000001O0000001O00001O0000001O0O10001N10001N10000O10001N10000O10001N100O101O1N2O001N2O0N3M3N1N3N2M3N2N101N2O1O001N2O00001N101O000O2O0O101O0O101N100F;M2M3N3N100O1O2N100O1O2N1O100O2O0O10001O00000O101O0000000O10000000001O0O2O001O00001N10000000O1000O10000000000O01000000000O100000O100000O1000000000O01000000000000O10001O00000O10000000000000000O100000000000001O0O1000000000001O1O0O2O1O1O001O1O1O2NU`W1"}, "label": "middle of the three birds"}]}
{"id": 336267, "image": "COCO_train2014_000000336267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sparrow is sitting along with two others\"."}], "task": "refering", "answer_template": "The \"a sparrow is sitting along with two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 286, 287, 302, 303], "bbox": [0.03625, 0.4279518072289156, 0.51596875, 0.8860240963855421], "iscrowd": 0, "area": 23618.360249999998, "rle": {"size": [415, 640], "counts": "Vn93k<2N1N3N1O1N2O1O1N2O2M2M3M3N2M3M4N2N2O1N2O1N2O1N2O1N5L7I7H9H6I3N2M3N1N3N2N2M3N1O2N2M3N1O2M3N2N1N3N2N001N2O1O1O1O00100O1O1O010O1O1gHmKd5T4ZJmKg5S4WJnKi5R4WJnKi5S4VJmKk5S4TJmKl5T4RJmKn5T4QJlKo5U4PJkKQ6T4oIlKQ6U4mIlKS6U4lIkKT6V4kIjKV6V4iIjKW6W4gIjKY6V4gIjKZ6V4eIjK[6W4dIiK\\6X4bIhK_6Y4`IgKa6Y4^IgKb6Y4^IgKb6Z4\\IfKe6[4ZIeKg6[4XIeKh6\\4WIdKi6\\4WIdKj6\\4TIeKn6Z4QIfKQ7Y4nHgKT7X4kHhKX7V4fHkK\\7T4cHlK_7T4_HkKd7T4[HlKd7W4YHjKg7W4XHiKh7Y4VHgKj7Z4UHfKk7g40000O01O1O1O1O1O10O01O1O1O2O1N2N2N2N2O1N2N2N2O1O1N2O1O1O1N2O1O001N2O1O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N101O1N2O000O10000O2O0O10001N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N3N1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N3N1N2L4O100O1O010O1O1O100O1O010O00010O000010O00010O00010O00010O0001O01O01O01O01O01O01O01O01O00010O00010O00010O00010O000010O00010O00010O0001O01O01EfCOZ<1fC0Y<0gC0Z<OgC0Y<0gCOZ<1gCNZ<1fCOZ<1fCN[<2fCM[<2eCM\\<3dCM\\<3eCK]<4cCL]<4cCK^<57O001O001O001O001O00Rgm3"}, "label": "a sparrow is sitting along with two others"}]}
{"id": 336267, "image": "COCO_train2014_000000336267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bid that is not facing the camera , looking off to the right\"."}], "task": "refering", "answer_template": "The \"a bid that is not facing the camera , looking off to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 64, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 251, 252], "bbox": [0.7850937499999999, 0.09031325301204819, 0.995671875, 0.7118313253012049], "iscrowd": 0, "area": 21008.843100000002, "rle": {"size": [415, 640], "counts": "So[6:=Ga;h0ZDYO_;R1\\DPO^;\\1J6K5K6J5J6I7J7I6B>L4L4L5K4N2M3M3N2M3M3N3L3N2M3M3N2M2N2O1N2N2O1WIhJh5[5SJhJm5Y5nIjJR6X5jIkJV6V5eImJ\\6T5_IPK`6R5[IQKf6d5O0100O1B\\IaJc6b5bIVJ_6k5:1O1O1N2O1O1O1O1O1O2M2O1O1O1O000000O10000O10001N2O0O2003M4L4K5L4K5K5KPJZJl4`5RKdJQ5X5kJmJX5o4eJTK_5g4_J]Kc5`4[JcKd5]4ZJfKe5Y4ZJkKd5U4RJVLm5j3oI\\LP6b3lIeLR6V5N3M2N3M2N2M2O1O1O1N2O1O1O1O1O1O1O1O1O1O3M3M3M3`MRHIQ8FoG[N6i1n7DiH8Z7_OoH=U7[OTIa0o6WOYIf0c9M3M3M4L3M3M3M4L3L4M3M4KVn0"}, "label": "a bid that is not facing the camera , looking off to the right"}]}
{"id": 336267, "image": "COCO_train2014_000000336267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bri facing the right\"."}], "task": "refering", "answer_template": "The \"bri facing the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 64, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 251, 252], "bbox": [0.7850937499999999, 0.09031325301204819, 0.995671875, 0.7118313253012049], "iscrowd": 0, "area": 21008.843100000002, "rle": {"size": [415, 640], "counts": "So[6:=Ga;h0ZDYO_;R1\\DPO^;\\1J6K5K6J5J6I7J7I6B>L4L4L5K4N2M3M3N2M3M3N3L3N2M3M3N2M2N2O1N2N2O1WIhJh5[5SJhJm5Y5nIjJR6X5jIkJV6V5eImJ\\6T5_IPK`6R5[IQKf6d5O0100O1B\\IaJc6b5bIVJ_6k5:1O1O1N2O1O1O1O1O1O2M2O1O1O1O000000O10000O10001N2O0O2003M4L4K5L4K5K5KPJZJl4`5RKdJQ5X5kJmJX5o4eJTK_5g4_J]Kc5`4[JcKd5]4ZJfKe5Y4ZJkKd5U4RJVLm5j3oI\\LP6b3lIeLR6V5N3M2N3M2N2M2O1O1O1N2O1O1O1O1O1O1O1O1O1O3M3M3M3`MRHIQ8FoG[N6i1n7DiH8Z7_OoH=U7[OTIa0o6WOYIf0c9M3M3M4L3M3M3M4L3L4M3M4KVn0"}, "label": "bri facing the right"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in lingerie with a hand on her hip and carrying a pink and black purse\"."}], "task": "refering", "answer_template": "The \"a woman in lingerie with a hand on her hip and carrying a pink and black purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 185, 186, 187, 188, 189, 190, 191, 192, 193, 200, 201, 202, 203, 204, 205, 206, 207, 208, 215, 216, 217, 218, 219, 220, 221, 222, 223, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 307, 308, 309, 310, 311, 312, 313, 314, 321, 322, 323, 324, 325, 326, 327, 328, 329, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.343559718969555, 0.191015625, 0.976768149882904, 0.9887656249999999], "iscrowd": 0, "area": 93690.6298, "rle": {"size": [640, 427], "counts": "WWl24ic07I6J6K6I6J5K6J5K6J6K4K6J5K6J5K6K5K4M4K4L5L3L4L3M3N3L3M3N2M3M3N3L3M3M3N2M4L3N2M3M3N3L3M3]GYJi2j5oL]JP3e5iLaJU3c5bLdJ\\3_5]LgJb3[5VLlJh3W5PLPKn3S5kKSKT4P5cKVK\\4l4_KYK_4j4]KYKa4i4\\KZKd4f4XK^Kg4d4UK_Kk4a4QKcKn4^4oJeKQ5[4kJiKT5Y4hJjKX5V4eJmKZ5T4bJPL^5Q4^JRLa5o3[JULe5k3XJXLk3mLlKm65YLd3WMUL`63]Lc3YMXL[62_Ld3YMYLX6OcLg3WMXLX6MdLi3XMXLT6KhLk3WMXLR6JjLm3VMXLQ6FmLP4VMWLn5EPMR4UMXLk5AUMV4RMWLk5^OWMY4QMXLh5ZO[M]4QMVLe5XO_Ma4nLVLc5UOcMc4mLWLc:g3aEWL_:h3cEWL]:g3fEXLZ:f3iEXLX:h3jEVLV:j3kEULU:l3kESLU:m3lERLT:n3mEQLS:P4mEoKS:Q4nEmKS:S4nElKR:U4nEjKR:V4oEiKQ:X4oEgKQ:Y4PFfKP:Z4QFeKo9\\4QFbKP:^4QFaKo9`4QF_Ko9a4RF^Kn9b4SF]Km9d4SF[Kl9f4UFXKl9h4UFWKk9i4f2O1000000000000000000000000O10000O100O1O100O100O010O100O100O100O100O100O100O100O100O100O1O100O100O10O0100OM4L4K6K4L4L4L4L51N:F9G:F7I01O0000001O001O1O1O001Oc2]M00001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001cCRJf:n5\\ESJa:m5aESJ]:m5eETJX:m5iETJT:m5mESJQ:o5oERJn9P6RFQJk9P6WFoIP1iNj6Z7WHmIj0QOi6T7^HjIe0ZOg6n6eHhI?Bf6h6lHfI9Je6b6SIcI52b6]6ZIaIO:a6V6bI`IHb0`6P6iI]ICk0^6j5QJZI]OS1\\6e5XJXIWO\\1Z6^5`JUIROe1X6X5gJYLS5h3oJ[L_4gK^JR8R1XLY4\\4eKdKU4d4iK\\KQ4m4lKSKn3U5PLjJj3_5TLaJf3g5XLYJb3o5\\LQJ^3X6_LgI\\3a6bL_IW3j6gLVIS3R7kLnHo2Z7oLeHl2d7QM\\Hi2l7UMTHe2T8YMlG`2]8^MbG]2U=J6J<D`0_O`0A?AWg5"}, "label": "a woman in lingerie with a hand on her hip and carrying a pink and black purse"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with the stickers on her butt\"."}], "task": "refering", "answer_template": "The \"the woman with the stickers on her butt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 185, 186, 187, 188, 189, 190, 191, 192, 193, 200, 201, 202, 203, 204, 205, 206, 207, 208, 215, 216, 217, 218, 219, 220, 221, 222, 223, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 307, 308, 309, 310, 311, 312, 313, 314, 321, 322, 323, 324, 325, 326, 327, 328, 329, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.343559718969555, 0.191015625, 0.976768149882904, 0.9887656249999999], "iscrowd": 0, "area": 93690.6298, "rle": {"size": [640, 427], "counts": "WWl24ic07I6J6K6I6J5K6J5K6J6K4K6J5K6J5K6K5K4M4K4L5L3L4L3M3N3L3M3N2M3M3N3L3M3M3N2M4L3N2M3M3N3L3M3]GYJi2j5oL]JP3e5iLaJU3c5bLdJ\\3_5]LgJb3[5VLlJh3W5PLPKn3S5kKSKT4P5cKVK\\4l4_KYK_4j4]KYKa4i4\\KZKd4f4XK^Kg4d4UK_Kk4a4QKcKn4^4oJeKQ5[4kJiKT5Y4hJjKX5V4eJmKZ5T4bJPL^5Q4^JRLa5o3[JULe5k3XJXLk3mLlKm65YLd3WMUL`63]Lc3YMXL[62_Ld3YMYLX6OcLg3WMXLX6MdLi3XMXLT6KhLk3WMXLR6JjLm3VMXLQ6FmLP4VMWLn5EPMR4UMXLk5AUMV4RMWLk5^OWMY4QMXLh5ZO[M]4QMVLe5XO_Ma4nLVLc5UOcMc4mLWLc:g3aEWL_:h3cEWL]:g3fEXLZ:f3iEXLX:h3jEVLV:j3kEULU:l3kESLU:m3lERLT:n3mEQLS:P4mEoKS:Q4nEmKS:S4nElKR:U4nEjKR:V4oEiKQ:X4oEgKQ:Y4PFfKP:Z4QFeKo9\\4QFbKP:^4QFaKo9`4QF_Ko9a4RF^Kn9b4SF]Km9d4SF[Kl9f4UFXKl9h4UFWKk9i4f2O1000000000000000000000000O10000O100O1O100O100O010O100O100O100O100O100O100O100O100O100O1O100O100O10O0100OM4L4K6K4L4L4L4L51N:F9G:F7I01O0000001O001O1O1O001Oc2]M00001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001cCRJf:n5\\ESJa:m5aESJ]:m5eETJX:m5iETJT:m5mESJQ:o5oERJn9P6RFQJk9P6WFoIP1iNj6Z7WHmIj0QOi6T7^HjIe0ZOg6n6eHhI?Bf6h6lHfI9Je6b6SIcI52b6]6ZIaIO:a6V6bI`IHb0`6P6iI]ICk0^6j5QJZI]OS1\\6e5XJXIWO\\1Z6^5`JUIROe1X6X5gJYLS5h3oJ[L_4gK^JR8R1XLY4\\4eKdKU4d4iK\\KQ4m4lKSKn3U5PLjJj3_5TLaJf3g5XLYJb3o5\\LQJ^3X6_LgI\\3a6bL_IW3j6gLVIS3R7kLnHo2Z7oLeHl2d7QM\\Hi2l7UMTHe2T8YMlG`2]8^MbG]2U=J6J<D`0_O`0A?AWg5"}, "label": "the woman with the stickers on her butt"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown haired girl almost completely hidden from view by a girl wearing a purple halter\"."}], "task": "refering", "answer_template": "The \"a brown haired girl almost completely hidden from view by a girl wearing a purple halter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 86, 87, 101, 102, 116, 117, 118, 132, 133, 187, 202, 203, 325, 339, 340], "bbox": [0.46981264637002346, 0.20196874999999997, 0.9036533957845434, 0.9993749999999999], "iscrowd": 0, "area": 3389.0003499999984, "rle": {"size": [640, 427], "counts": "Pom33lc03M3L4M2N3L3N2N2M4M2N2M300001O0O10000O101O0O103M2M4M1O0O101O0O101O005KTdh0FS\\WO2O1N2N2O1N2O1N2O1N2O1O1N2O1O1N2N2O1N2O1N2N2OO2K6J5L4L4H^a>IQ_A3N1O2N3M5K6P^OIj?o0Y_OUOh`0Q21O100O2N1O101N2N1O2O1N2N2N2N3M3M4M2M4L7I7I4L2N1O2N101N2N1O2O1N1O2O1N1O2O1O1O1O1N2O1O1O1O1OdXk0"}, "label": "a brown haired girl almost completely hidden from view by a girl wearing a purple halter"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person standing extreem left\"."}], "task": "refering", "answer_template": "The \"the person standing extreem left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 46, 47, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 153, 155, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 225, 226, 227, 228, 240, 241, 242, 243, 255, 256, 257, 258, 270, 271, 272, 273, 285, 286, 287, 288, 300, 301, 302, 303, 315, 316, 317, 318, 330, 331, 332, 333], "bbox": [0.0033723653395784543, 0.096625, 0.3839812646370024, 0.9842656250000001], "iscrowd": 0, "area": 53450.019400000005, "rle": {"size": [640, 427], "counts": "dV1R1Y8e:hLX3O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1NZD]MS6b2mI`MR6_2mIeMQ6Z2oIhMP6V2PJmMo5R2QJQNm5n1RJUNm5i1TJYNk5f1TJ^Nj5a1VJaNi5^1VJeNi5`1QJcNm5b1lIaNS6d1gI^NX6g1aI]N]6h1]IZNb6k1WIXNh6m1RIVNl6o1mHTNR7Q2hHQNW7T2aGQO]8T1YGROf8S1PGSOo8R1gFUOW9P1_FVO`9P1UFVOj9b5N1O1O2N1O7I6J7I7I1O00O1000C<gNZ1WOh0ZOg0YOg0YOf0F;H7J7I6J7I7I6J7I6J7I7E:01O010O1O001O001O001O10O01O001O1O001O0010O01O1O001N1N3L4L3M4M2M4L4M2M4L3N3L3M4M3L3M4M2MlRT5"}, "label": "the person standing extreem left"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing fishnet stockings with a big tattoo\"."}], "task": "refering", "answer_template": "The \"a woman wearing fishnet stockings with a big tattoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 46, 47, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 153, 155, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 225, 226, 227, 228, 240, 241, 242, 243, 255, 256, 257, 258, 270, 271, 272, 273, 285, 286, 287, 288, 300, 301, 302, 303, 315, 316, 317, 318, 330, 331, 332, 333], "bbox": [0.0033723653395784543, 0.096625, 0.3839812646370024, 0.9842656250000001], "iscrowd": 0, "area": 53450.019400000005, "rle": {"size": [640, 427], "counts": "dV1R1Y8e:hLX3O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1NZD]MS6b2mI`MR6_2mIeMQ6Z2oIhMP6V2PJmMo5R2QJQNm5n1RJUNm5i1TJYNk5f1TJ^Nj5a1VJaNi5^1VJeNi5`1QJcNm5b1lIaNS6d1gI^NX6g1aI]N]6h1]IZNb6k1WIXNh6m1RIVNl6o1mHTNR7Q2hHQNW7T2aGQO]8T1YGROf8S1PGSOo8R1gFUOW9P1_FVO`9P1UFVOj9b5N1O1O2N1O7I6J7I7I1O00O1000C<gNZ1WOh0ZOg0YOg0YOf0F;H7J7I6J7I7I6J7I6J7I7E:01O010O1O001O001O001O10O01O001O1O001O0010O01O1O001N1N3L4L3M4M2M4L4M2M4L3N3L3M4M3L3M4M2MlRT5"}, "label": "a woman wearing fishnet stockings with a big tattoo"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white car in the distance\"."}], "task": "refering", "answer_template": "The \"white car in the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 169, 170], "bbox": [0.12524590163934426, 0.309, 0.5913817330210772, 0.493359375], "iscrowd": 0, "area": 12320.843550000001, "rle": {"size": [640, 427], "counts": "eoQ12mc02O2M2O2N1N3N1N2O2M2O2M2O2M2O2M2O1O2M2O2M2O2M2O1N3N1N3N1N3N1O2M2O1N3N1N3N1N3N1N3N1N2OM3M2N3M2N3M2N3M2N3M2N3M2N34L5J5L4K5L4K5L2M2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N010O10O0100O010000O100O01000O100O1M3M3M3N2M3M3M3M2100000O1000000000000000O10000000000000000000O1O1O0O2O1N2O1O1N2O1O1N2O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2O1O00100O100O10000O100O2O001N2O0O2O1O1N101N2O1O0O2O1N2O0O2N2O1N1O2N2N2N101N2N1O2N2N2N10Sa\\3"}, "label": "white car in the distance"}]}
{"id": 475533, "image": "COCO_train2014_000000475533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car\"."}], "task": "refering", "answer_template": "The \"a white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 169, 170], "bbox": [0.12524590163934426, 0.309, 0.5913817330210772, 0.493359375], "iscrowd": 0, "area": 12320.843550000001, "rle": {"size": [640, 427], "counts": "eoQ12mc02O2M2O2N1N3N1N2O2M2O2M2O2M2O2M2O1O2M2O2M2O2M2O1N3N1N3N1N3N1O2M2O1N3N1N3N1N3N1N3N1N2OM3M2N3M2N3M2N3M2N3M2N3M2N34L5J5L4K5L4K5L2M2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N010O10O0100O010000O100O01000O100O1M3M3M3N2M3M3M3M2100000O1000000000000000O10000000000000000000O1O1O0O2O1N2O1O1N2O1O1N2O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2O1O00100O100O10000O100O2O001N2O0O2O1O1N101N2O1O0O2O1N2O0O2N2O1N1O2N2N2N101N2N1O2N2N2N10Sa\\3"}, "label": "a white car"}]}
{"id": 74127, "image": "COCO_train2014_000000074127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black shirt and green apron holding a baby\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black shirt and green apron holding a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 248, 249, 250, 271, 272, 294], "bbox": [0.77534375, 0.132875, 1.0, 0.7229791666666666], "iscrowd": 0, "area": 20012.856849999993, "rle": {"size": [480, 640], "counts": "ifX72i>6I7J6J6J6J6J6I7`EjNo6^1[HVO_7S1jGAP8h0\\GI^8`0QGMj8X3K5N2N2O0O2N2O1N1O2O1N2N01O000O2O000OYNiJhIW5X6oJbIQ5^6UK\\Ij4e6[KVIe4j6h10O10000O101O1N2O001N2O1O1N2O001N2O1O1N2O001O1N2O1O1N101O1N2O1O1N2O1O2M2O1O1N3N1O1N2O1O2M2O1O3L4M3M3L4M2G:[NjFWM]9e2kFSM\\9g2fFUMd9b2_FZMk9]1cEjNe0Fn9Z1iEdN<NS:W1fEfN:0X:R1bEjNj;T1S1M2N3N1O2N101N1O2N101N1O1O2O0O1O1O101N1002N1O2O0O2N1O2N110O012M4M4K4M0ON3O3M3N1N3M2O3I7J5J7JaJ"}, "label": "a woman wearing a black shirt and green apron holding a baby"}]}
{"id": 74127, "image": "COCO_train2014_000000074127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman and child on right\"."}], "task": "refering", "answer_template": "The \"woman and child on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 248, 249, 250, 271, 272, 294], "bbox": [0.77534375, 0.132875, 1.0, 0.7229791666666666], "iscrowd": 0, "area": 20012.856849999993, "rle": {"size": [480, 640], "counts": "ifX72i>6I7J6J6J6J6J6I7`EjNo6^1[HVO_7S1jGAP8h0\\GI^8`0QGMj8X3K5N2N2O0O2N2O1N1O2O1N2N01O000O2O000OYNiJhIW5X6oJbIQ5^6UK\\Ij4e6[KVIe4j6h10O10000O101O1N2O001N2O1O1N2O001N2O1O1N2O001O1N2O1O1N101O1N2O1O1N2O1O2M2O1O1N3N1O1N2O1O2M2O1O3L4M3M3L4M2G:[NjFWM]9e2kFSM\\9g2fFUMd9b2_FZMk9]1cEjNe0Fn9Z1iEdN<NS:W1fEfN:0X:R1bEjNj;T1S1M2N3N1O2N101N1O2N101N1O1O2O0O1O1O101N1002N1O2O0O2N1O2N110O012M4M4K4M0ON3O3M3N1N3M2O3I7J5J7JaJ"}, "label": "woman and child on right"}]}
{"id": 74127, "image": "COCO_train2014_000000074127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a ponytail and multicolored apron not holding a child\"."}], "task": "refering", "answer_template": "The \"a woman with a ponytail and multicolored apron not holding a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 239, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 333, 334, 355, 356, 357, 378, 379], "bbox": [0.327703125, 0.17341666666666666, 0.545609375, 0.99775], "iscrowd": 0, "area": 27372.58500000001, "rle": {"size": [480, 640], "counts": "keR35f>7K4oAD[=?aBEP9c0UJJgLIo8c0TJHiLIP9d0RJ:l5LoI5o51kI1T66eIKZ67cIK]66aIK^67_IJb66]IJc68ZIJf68VIIj69TIGm6:QIGn6;oHFR7=jHCV7c0cH_O]7R1nGQOR8n3010O0010O0101N1O2O000O2O00001N10001N3O0000000005LO0O1O1O100O1O0010O01O010O001O01O01O010O0010O01O00101RK`FX4c9aKaF_4a9[KdFe4k900010O000YFUMY7l2eHYOX5f0gJ3a4M^K2d4N[K2g4NUK4l4n42O1O2ZOe0L4L5K5K56J:F:oJhFd3Z:1O1O1O1O1O001O1O1O2N2M3N2N2N3M6I6J7I7I7H6K5K4L5J5I8F9F;TMdHdMh7V2g2I6J9G8H9H8G8H9G9F9F;D=DdSX4"}, "label": "a woman with a ponytail and multicolored apron not holding a child"}]}
{"id": 508311, "image": "COCO_train2014_000000508311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird touching its neck with its right feet\"."}], "task": "refering", "answer_template": "The \"a bird touching its neck with its right feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 155, 156, 157, 178, 179, 180, 203], "bbox": [0.54515625, 0.051604166666666666, 1.0, 0.5118333333333334], "iscrowd": 0, "area": 32698.871550000007, "rle": {"size": [480, 640], "counts": "hfS53j>5L4K6J5K5K6J5K5L5J6YCZNk;P2hCXNU<[2N2N2N2O0O01O0010O01O001O001N2O001O0O2N1O2M2O2O001000O1]CfM`<Z2_CgM`<]200O100O01000O100O100O100O100O10000O100O1000O0100O10000O100O100O00100O100O1O100O100O100O1O100O100O1O100O100O1O101N100O1O2O0O3N2M2O1M3K6K4O1O1N2O1O1O2M2O1O1O1O1N2O1O1O1N2O001O1N2O1O2N1O1N3N1O1O2N1O2N100O2N1O101N1O2N1O101N1O1O1O010O001O01O01O001OO2O00000[NjF`MV9a2mF[MT9d2PGXMP9i2SGRMn8n2VGlLl8T3YGeLi8\\3ZG^Li8a3\\GXLf8i3^GPLd8P4V11O010O00001O010O000iFmKa7T4[HRLb7n3ZHYLd7l3RHZLl7j3lG[LS8h3gG]LW8e3cGaL[8b3`GaL`8a3[GbLd8n4O1O1O10`NeGZLZ8Y3]H[Le7T3VIaLk6_3_IVLc6i3Z2O2O000O2O001N101O000O2O001N101O001N2O001N2O001O0O2O1O0O2O1O001N101O1N101O1O0O2O001N101O001N2O001N101O001N101O0O2O001O0O2O001N101O0O2O00000O2O000O101O000O101O0O100N3N1O1O1Oa0_ORM"}, "label": "a bird touching its neck with its right feet"}]}
{"id": 508311, "image": "COCO_train2014_000000508311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird pecking at the water\"."}], "task": "refering", "answer_template": "The \"a bird pecking at the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 155, 156, 157, 178, 179, 180, 203], "bbox": [0.54515625, 0.051604166666666666, 1.0, 0.5118333333333334], "iscrowd": 0, "area": 32698.871550000007, "rle": {"size": [480, 640], "counts": "hfS53j>5L4K6J5K5K6J5K5L5J6YCZNk;P2hCXNU<[2N2N2N2O0O01O0010O01O001O001N2O001O0O2N1O2M2O2O001000O1]CfM`<Z2_CgM`<]200O100O01000O100O100O100O100O10000O100O1000O0100O10000O100O100O00100O100O1O100O100O100O1O100O100O1O100O100O1O101N100O1O2O0O3N2M2O1M3K6K4O1O1N2O1O1O2M2O1O1O1O1N2O1O1O1N2O001O1N2O1O2N1O1N3N1O1O2N1O2N100O2N1O101N1O2N1O101N1O1O1O010O001O01O01O001OO2O00000[NjF`MV9a2mF[MT9d2PGXMP9i2SGRMn8n2VGlLl8T3YGeLi8\\3ZG^Li8a3\\GXLf8i3^GPLd8P4V11O010O00001O010O000iFmKa7T4[HRLb7n3ZHYLd7l3RHZLl7j3lG[LS8h3gG]LW8e3cGaL[8b3`GaL`8a3[GbLd8n4O1O1O10`NeGZLZ8Y3]H[Le7T3VIaLk6_3_IVLc6i3Z2O2O000O2O001N101O000O2O001N101O001N2O001N2O001O0O2O1O0O2O1O001N101O1N101O1O0O2O001N101O001N2O001N101O001N101O0O2O001O0O2O001N101O0O2O00000O2O000O101O000O101O0O100N3N1O1O1Oa0_ORM"}, "label": "a bird pecking at the water"}]}
{"id": 508311, "image": "COCO_train2014_000000508311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird in front\"."}], "task": "refering", "answer_template": "The \"the bird in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 287, 288, 310, 311, 333, 334, 335], "bbox": [0.064234375, 0.18008333333333332, 0.9780468750000001, 0.8781458333333333], "iscrowd": 0, "area": 57260.287149999975, "rle": {"size": [480, 640], "counts": "l[c01n>3N0O0010O01O010O0010O01O0010O01O010O00100O0010O01O0010O010O010O10O0100O010O10O10O01000O01000O010000O01000O01000O01000O10O10O10O10O10O10O10O10O1000O01000O0100O01000O10000000000O100000000000000O1000000000000O1000000000000O100000000000000O10001O0000000O10001O000bCA[:`0bED\\:<bEIZ:8dEMY:4eEOY:1eE4X:MeE8W:IgE<U:EjE>S:DjEa0S:_OkEf0Q:[OmEi0P:YOmEl0o9UOPFn0m9SOQFQ1l9QOQFR1m9oNQFU1l9mNRFU1l9lNRFX1k9kNRFW1m9jNQFY1l9jNQFX1n9iNPFZ1m9iNQFX1n9iNPFZ1m9iNPFY1o9hNoE[1n9hNoEZ1P:m1M2O2M2O2M2O2M2O2M3N1N3N1N3N1N3O0O2N1O2N2O0O2N1O2O0O2N1O10O01O1O10O01O1O00100O001O10O01O1O010001N2O1O1N2O001N2O1O1N2O1O001Na0@e0[O0O01O1O001O1O0010O0001O0000000010O000001O000000010O000000001O01O0001O0000000010O00000001O0000010O000000001O0fHjJj4V5VKjJk4U5TKlJl4T5SKmJm4T5QKnJn4R5QKoJo4Q5PKPKQ5R5kJoJU5V5eJlJZ5X5aJiJ_5[5\\JfJd5_5VJbJj5b5QJ_JP6e5jI]JU6U6XIlIh6V700010O000001O01XNjHdKV7Z4PIbKP7[4VIcKi6[4]IaKc6]4bI`K_6\\4hI`KX6^4nI_KQ6^4UJ^Kl5`4ZJ\\Kf5b4`JZK`5c4fJZK[5c4hJ\\KX5a4kJ_KU5a4lJ^KT5b4lJ^KT5b4mJ]KT5b4lJ^KT5b4mJ]KS5c4mJ\\KT5d4mJ[KS5e4mJ[KT5d4mJ[KS5d4nJ\\KR5d4oJ[KQ5e4oJ[KQ5e4PKZKQ5e4PKZKP5f4PKZKP5f4QKXKP5h4PKXKP5h4QKWKP5h4PKXKP5h4QKWKo4i4QKWKo4i4RKVKn4i4SKWKm4i4TKVKm4i4SKWKm4i4e2000001O000000001O0O10001O001O1O001O001O001O0O2O001O001O001O001O0O2O1O001O001O0O2O001O001O001N101O001O1O001N101O001O001O0O2O001O001O001N101O1O001O001N101O001O001O0O2O001O00001O0O101O00000O2O000O101O0O100O2O000O2O0O10001N100O101O0O10001N100O101O0O101N10000O2O0O10001N100O101O000O2O00000O2O00000O2O3M2N3L1000000O101O000O1000000O101O000O1000000O101N10000O100O10001N100O100O10000O2O0O10000O100O101O0O100O100O10001N100O10000O100O101O0O100O10000000000010O00000000001O00BlA0T>NnA2R>MoA3R>JPB6P>IRB7m=GUB9k=FVB:X>01O0001O01O000G^A4j>N1NWY6"}, "label": "the bird in front"}]}
{"id": 508311, "image": "COCO_train2014_000000508311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird with both feet in the water\"."}], "task": "refering", "answer_template": "The \"bird with both feet in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 287, 288, 310, 311, 333, 334, 335], "bbox": [0.064234375, 0.18008333333333332, 0.9780468750000001, 0.8781458333333333], "iscrowd": 0, "area": 57260.287149999975, "rle": {"size": [480, 640], "counts": "l[c01n>3N0O0010O01O010O0010O01O0010O01O010O00100O0010O01O0010O010O010O10O0100O010O10O10O01000O01000O010000O01000O01000O01000O10O10O10O10O10O10O10O10O1000O01000O0100O01000O10000000000O100000000000000O1000000000000O1000000000000O100000000000000O10001O0000000O10001O000bCA[:`0bED\\:<bEIZ:8dEMY:4eEOY:1eE4X:MeE8W:IgE<U:EjE>S:DjEa0S:_OkEf0Q:[OmEi0P:YOmEl0o9UOPFn0m9SOQFQ1l9QOQFR1m9oNQFU1l9mNRFU1l9lNRFX1k9kNRFW1m9jNQFY1l9jNQFX1n9iNPFZ1m9iNQFX1n9iNPFZ1m9iNPFY1o9hNoE[1n9hNoEZ1P:m1M2O2M2O2M2O2M2O2M3N1N3N1N3N1N3O0O2N1O2N2O0O2N1O2O0O2N1O10O01O1O10O01O1O00100O001O10O01O1O010001N2O1O1N2O001N2O1O1N2O1O001Na0@e0[O0O01O1O001O1O0010O0001O0000000010O000001O000000010O000000001O01O0001O0000000010O00000001O0000010O000000001O0fHjJj4V5VKjJk4U5TKlJl4T5SKmJm4T5QKnJn4R5QKoJo4Q5PKPKQ5R5kJoJU5V5eJlJZ5X5aJiJ_5[5\\JfJd5_5VJbJj5b5QJ_JP6e5jI]JU6U6XIlIh6V700010O000001O01XNjHdKV7Z4PIbKP7[4VIcKi6[4]IaKc6]4bI`K_6\\4hI`KX6^4nI_KQ6^4UJ^Kl5`4ZJ\\Kf5b4`JZK`5c4fJZK[5c4hJ\\KX5a4kJ_KU5a4lJ^KT5b4lJ^KT5b4mJ]KT5b4lJ^KT5b4mJ]KS5c4mJ\\KT5d4mJ[KS5e4mJ[KT5d4mJ[KS5d4nJ\\KR5d4oJ[KQ5e4oJ[KQ5e4PKZKQ5e4PKZKP5f4PKZKP5f4QKXKP5h4PKXKP5h4QKWKP5h4PKXKP5h4QKWKo4i4QKWKo4i4RKVKn4i4SKWKm4i4TKVKm4i4SKWKm4i4e2000001O000000001O0O10001O001O1O001O001O001O0O2O001O001O001O001O0O2O1O001O001O0O2O001O001O001N101O001O1O001N101O001O001O0O2O001O001O001N101O1O001O001N101O001O001O0O2O001O00001O0O101O00000O2O000O101O0O100O2O000O2O0O10001N100O101O0O10001N100O101O0O101N10000O2O0O10001N100O101O000O2O00000O2O00000O2O3M2N3L1000000O101O000O1000000O101O000O1000000O101N10000O100O10001N100O100O10000O2O0O10000O100O101O0O100O100O10001N100O10000O100O101O0O100O10000000000010O00000000001O00BlA0T>NnA2R>MoA3R>JPB6P>IRB7m=GUB9k=FVB:X>01O0001O01O000G^A4j>N1NWY6"}, "label": "bird with both feet in the water"}]}
{"id": 16796, "image": "COCO_train2014_000000016796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a toasted sandwich with a bite taken out of the corner\"."}], "task": "refering", "answer_template": "The \"half of a toasted sandwich with a bite taken out of the corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 399, 400, 401, 402, 403, 404, 405, 422, 423, 424, 425, 426, 427, 428, 446, 447], "bbox": [0.351125, 0.50168, 0.65396875, 0.867984], "iscrowd": 0, "area": 28865.398499999996, "rle": {"size": [625, 640], "counts": "ehY44[c07J6J6J6J6J6I6K6J6J6J6J6I7J6J6J6J6I4M1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O001O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N101O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O100O00100O1O100O1O100O1002N2N1O2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2M3N2N2N2N1O2N2N2N2N2M3L4M3L4M3L4L4M2M4M3L4M3L4M3L4M3L4M3L4L3N3L4M3L4M3L4M3L4J6J6I7I7I6K6I7I7J6Ib^V4"}, "label": "half of a toasted sandwich with a bite taken out of the corner"}]}
{"id": 16796, "image": "COCO_train2014_000000016796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a unfinished piece of bread sandwich\"."}], "task": "refering", "answer_template": "The \"a unfinished piece of bread sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 399, 400, 401, 402, 403, 404, 405, 422, 423, 424, 425, 426, 427, 428, 446, 447], "bbox": [0.351125, 0.50168, 0.65396875, 0.867984], "iscrowd": 0, "area": 28865.398499999996, "rle": {"size": [625, 640], "counts": "ehY44[c07J6J6J6J6J6I6K6J6J6J6J6I7J6J6J6J6I4M1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O001O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N101O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O100O00100O1O100O1O100O1002N2N1O2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2M3N2N2N2N1O2N2N2N2N2M3L4M3L4M3L4L4M2M4M3L4M3L4M3L4M3L4M3L4L3N3L4M3L4M3L4M3L4J6J6I7I7I6K6I7I7J6Ib^V4"}, "label": "a unfinished piece of bread sandwich"}]}
{"id": 16796, "image": "COCO_train2014_000000016796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee in a black mug beside a sandwich and salad\"."}], "task": "refering", "answer_template": "The \"a cup of coffee in a black mug beside a sandwich and salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226], "bbox": [0.621453125, 0.184496, 0.898234375, 0.435824], "iscrowd": 0, "area": 22690.291849999998, "rle": {"size": [625, 640], "counts": "PSc7=kb0d0\\Od0]O?@;H8I7J6K5N2N2N2M3N2N2N2M3N2N1O1N2O001O1N2O001O1N2O001O1O1N101O1O1N101O000O101O000O1000001N100001O000001O00000000001O0000000000010O00000000001O0000000000001O0001O00000001O0000000000000000000001O0O1000001O00001O000O101O00001O00000O2O00001O0000001O0O101O00001O0000001N1O1O2N1N2O1O2M2O1N3N1O1N2O2N1N2O2N1N3M3L4K5K5L3L5K5L4K5K5L3L5L4K5K5L3L7I7J6I7I7Ja`W1"}, "label": "a cup of coffee in a black mug beside a sandwich and salad"}]}
{"id": 16796, "image": "COCO_train2014_000000016796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black cup filled with black coffee\"."}], "task": "refering", "answer_template": "The \"black cup filled with black coffee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226], "bbox": [0.621453125, 0.184496, 0.898234375, 0.435824], "iscrowd": 0, "area": 22690.291849999998, "rle": {"size": [625, 640], "counts": "PSc7=kb0d0\\Od0]O?@;H8I7J6K5N2N2N2M3N2N2N2M3N2N1O1N2O001O1N2O001O1N2O001O1O1N101O1O1N101O000O101O000O1000001N100001O000001O00000000001O0000000000010O00000000001O0000000000001O0001O00000001O0000000000000000000001O0O1000001O00001O000O101O00001O00000O2O00001O0000001O0O101O00001O0000001N1O1O2N1N2O1O2M2O1N3N1O1N2O2N1N2O2N1N3M3L4K5K5L3L5K5L4K5K5L3L5L4K5K5L3L7I7J6I7I7Ja`W1"}, "label": "black cup filled with black coffee"}]}
{"id": 16796, "image": "COCO_train2014_000000016796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left half od the sandwich\"."}], "task": "refering", "answer_template": "The \"the left half od the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 393, 394, 395, 396, 397, 417, 418, 419], "bbox": [0.08301562500000001, 0.45435200000000003, 0.447859375, 0.8570400000000001], "iscrowd": 0, "area": 37587.604649999994, "rle": {"size": [625, 640], "counts": "^fP17ob0c0]Od0A?F:G8G6K3L4M3L4M3L4M3L4L4M3L4M3L5L4K6K4K5L4K5L4K5L4K6K4K5L4K5L1N101N101N101N1O2O0O2O01O3N3L4M2MN3N1N3N1N3M2O1N3N1N3N1N3N101O001O0O101O001O001O0O2O001O00000O0100O100O1O010O1O100O1O010O1O1O1O1O001O1O1N2O1O1O001O1O1N2O1O001O1O1O1N2O001O1O1O1O1N101O1O1O1O1O0O2O1O1O1O1O001N2O1O1O1O1O001N2O1O1O1O001O1N2O1O1O001O1O1N2O1O001O1O1O1N2O001O1O1O1O1N101O1O1O1O1O1N101O1O1O1O1O0O2O1O1O1O1O001N2O1O1O1O001O1N2O2^MY@W1j?eNX@Z1j?bNZ@\\1h?aNZ@^1h?^N[@a1h?[NZ@c1i?YN[@e1g?XN[@g1g?UN\\@j1g?RN[@m1e`0O2N1O1N2O1O2N1O1O3M5J6K5K5K5K5K5K5J3N2N2Ndfg6"}, "label": "the left half od the sandwich"}]}
{"id": 16796, "image": "COCO_train2014_000000016796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half a sandwich that has not been bitten\"."}], "task": "refering", "answer_template": "The \"half a sandwich that has not been bitten\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 393, 394, 395, 396, 397, 417, 418, 419], "bbox": [0.08301562500000001, 0.45435200000000003, 0.447859375, 0.8570400000000001], "iscrowd": 0, "area": 37587.604649999994, "rle": {"size": [625, 640], "counts": "^fP17ob0c0]Od0A?F:G8G6K3L4M3L4M3L4M3L4L4M3L4M3L5L4K6K4K5L4K5L4K5L4K6K4K5L4K5L1N101N101N101N1O2O0O2O01O3N3L4M2MN3N1N3N1N3M2O1N3N1N3N1N3N101O001O0O101O001O001O0O2O001O00000O0100O100O1O010O1O100O1O010O1O1O1O1O001O1O1N2O1O1O001O1O1N2O1O001O1O1O1N2O001O1O1O1O1N101O1O1O1O1O0O2O1O1O1O1O001N2O1O1O1O1O001N2O1O1O1O001O1N2O1O1O001O1O1N2O1O001O1O1O1N2O001O1O1O1O1N101O1O1O1O1O1N101O1O1O1O1O0O2O1O1O1O1O001N2O1O1O1O001O1N2O2^MY@W1j?eNX@Z1j?bNZ@\\1h?aNZ@^1h?^N[@a1h?[NZ@c1i?YN[@e1g?XN[@g1g?UN\\@j1g?RN[@m1e`0O2N1O1N2O1O2N1O1O3M5J6K5K5K5K5K5K5J3N2N2Ndfg6"}, "label": "half a sandwich that has not been bitten"}]}
{"id": 139679, "image": "COCO_train2014_000000139679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white t - shirt who is playing tennis\"."}], "task": "refering", "answer_template": "The \"a man wearing a white t - shirt who is playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 150, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 247, 264, 265, 270, 271, 272, 287, 288, 294, 295, 310, 311, 333, 334, 356, 357, 379], "bbox": [0.365, 0.2869758064516129, 0.8643593749999999, 0.9006451612903227], "iscrowd": 0, "area": 24281.84180000002, "rle": {"size": [496, 640], "counts": "f_a36X?5K7J6J5L5K5K2N1O2N1O2N1O2N1O1N2O1O2N1O000000OcB[NQ=e1nBcNk<]1SCgNk<Y1TCjNj<U1RCROk<l1O1N2O1N10O01000O0100O010O10O2O1O2M3N2M01O1O001O1\\C]MX<e2gC\\MX<e2gC\\MX<e2gC[MY<f2fCZMZ<g2eCYM[<h2dCXM[<j2dCWM[<o20001O1O1O001O1O1O1O001TIiL]1W3bNmL[1R3eNoL[1P3eNQM[1o2eNQM[1Y2`InMU5J\\1W2bIlMQ5N]1W2bIiMQ51]1V2dIfMi1Nm07f3V2dIcM`1;T1Nh3S2eIcMT1h0]1Cj3S2nJeNV1YOl3S2fJlN\\1ROo3R2`JPO_1POQ4Q2\\JSOa1mNS4Q2WJYOb1gNX4Q2QJ_Oa1bN^4P2mIF_1ZNe4R2gIL^1SNk4S2bI1]1mMS5S2\\I7[1gMY5b5aJ_J_5f5ZJ[Jg5i5TJXJl5m5nISJS6R6gIoIX6Z6_IgIa6c701000O2O1N2O1O1N2O1bLPIYNQ7e1TIXNm6e1YIWNg6g1]IVNd6h1`IUL^OQ1S7g2dISL_OQ1n6k2fIoKBS1i6k2jImKBU1d6l2nIkKDT1_6o2RJgKEW1[6n2YKnLj4n2\\KnLf4n2_KoLc4n2aKnLb4n2bKPM`4l2dKRM^4j2fKSM\\4j2hKTMZ4h2jKUMX4i2kKUMV4h2nKVMR4h2RLUMP4i2TLTMl3j2XLSMj3k2_4N1O2N1O2N1N3N101N1O2O0O2O0O2O01O0000000000000001O1O1O1O001O1O1O001O1O1O1O2N100000000O1O1O1O1O2N1N2O1O1O1O1O1O101N01000O10000O01000O10000O010O100O1O010O1O1mNfAn0_>00O1O1O010O1O1O10O01O1O10O01O00100O0010O100O10O01O100O100O00100O10O01O010O0100O10O01O100O100O100O100O1O100O100O100O100O010O100O100O10O0100O11O001N101O001O0O2O1O001N101N101N1O101N100O2O0000N2O1N2O1N2N2O1N3N1N3N1N3M5LgiY1"}, "label": "a man wearing a white t - shirt who is playing tennis"}]}
{"id": 139679, "image": "COCO_train2014_000000139679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white shirt playing tennis\"."}], "task": "refering", "answer_template": "The \"a man in white shirt playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 150, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 247, 264, 265, 270, 271, 272, 287, 288, 294, 295, 310, 311, 333, 334, 356, 357, 379], "bbox": [0.365, 0.2869758064516129, 0.8643593749999999, 0.9006451612903227], "iscrowd": 0, "area": 24281.84180000002, "rle": {"size": [496, 640], "counts": "f_a36X?5K7J6J5L5K5K2N1O2N1O2N1O2N1O1N2O1O2N1O000000OcB[NQ=e1nBcNk<]1SCgNk<Y1TCjNj<U1RCROk<l1O1N2O1N10O01000O0100O010O10O2O1O2M3N2M01O1O001O1\\C]MX<e2gC\\MX<e2gC\\MX<e2gC[MY<f2fCZMZ<g2eCYM[<h2dCXM[<j2dCWM[<o20001O1O1O001O1O1O1O001TIiL]1W3bNmL[1R3eNoL[1P3eNQM[1o2eNQM[1Y2`InMU5J\\1W2bIlMQ5N]1W2bIiMQ51]1V2dIfMi1Nm07f3V2dIcM`1;T1Nh3S2eIcMT1h0]1Cj3S2nJeNV1YOl3S2fJlN\\1ROo3R2`JPO_1POQ4Q2\\JSOa1mNS4Q2WJYOb1gNX4Q2QJ_Oa1bN^4P2mIF_1ZNe4R2gIL^1SNk4S2bI1]1mMS5S2\\I7[1gMY5b5aJ_J_5f5ZJ[Jg5i5TJXJl5m5nISJS6R6gIoIX6Z6_IgIa6c701000O2O1N2O1O1N2O1bLPIYNQ7e1TIXNm6e1YIWNg6g1]IVNd6h1`IUL^OQ1S7g2dISL_OQ1n6k2fIoKBS1i6k2jImKBU1d6l2nIkKDT1_6o2RJgKEW1[6n2YKnLj4n2\\KnLf4n2_KoLc4n2aKnLb4n2bKPM`4l2dKRM^4j2fKSM\\4j2hKTMZ4h2jKUMX4i2kKUMV4h2nKVMR4h2RLUMP4i2TLTMl3j2XLSMj3k2_4N1O2N1O2N1N3N101N1O2O0O2O0O2O01O0000000000000001O1O1O1O001O1O1O001O1O1O1O2N100000000O1O1O1O1O2N1N2O1O1O1O1O1O101N01000O10000O01000O10000O010O100O1O010O1O1mNfAn0_>00O1O1O010O1O1O10O01O1O10O01O00100O0010O100O10O01O100O100O00100O10O01O010O0100O10O01O100O100O100O100O1O100O100O100O100O010O100O100O10O0100O11O001N101O001O0O2O1O001N101N101N1O101N100O2O0000N2O1N2O1N2N2O1N3N1N3N1N3M5LgiY1"}, "label": "a man in white shirt playing tennis"}]}
{"id": 483752, "image": "COCO_train2014_000000483752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lighter colored vehicle behind the darker one\"."}], "task": "refering", "answer_template": "The \"the lighter colored vehicle behind the darker one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 341, 342], "bbox": [0.6434375, 0.6919375, 0.9750937499999999, 0.8471458333333334], "iscrowd": 0, "area": 10734.155950000002, "rle": {"size": [480, 640], "counts": "e_Q6?a>0O1O1O1O2N100O1O1O1O1O2O0O10000O101O2N2N2M2O0000001N10001O00001N10001OO1O010O1O100O00100O0010O010O100O100O100O10000O100O100O100O100O100O101O0O100O100O100O100O100O10000O100O100O100O100O10000000000000000000000000000000000000000O1000000000000000000000000000000000000000001O1O1O1O2N1O001O001O1O001O001O001OO10000000000000000000000000000EbCTN^<l1;0000O2O00001O00001O00001O00001O1N2O001O1O1N2O1O1O1N2O1O1O1N2O1O1O1N4M4L5K4L5J5L2N3M2MoS7"}, "label": "the lighter colored vehicle behind the darker one"}]}
{"id": 483752, "image": "COCO_train2014_000000483752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car drives down the street\"."}], "task": "refering", "answer_template": "The \"a white car drives down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 341, 342], "bbox": [0.6434375, 0.6919375, 0.9750937499999999, 0.8471458333333334], "iscrowd": 0, "area": 10734.155950000002, "rle": {"size": [480, 640], "counts": "e_Q6?a>0O1O1O1O2N100O1O1O1O1O2O0O10000O101O2N2N2M2O0000001N10001O00001N10001OO1O010O1O100O00100O0010O010O100O100O100O10000O100O100O100O100O100O101O0O100O100O100O100O100O10000O100O100O100O100O10000000000000000000000000000000000000000O1000000000000000000000000000000000000000001O1O1O1O2N1O001O001O1O001O001O001OO10000000000000000000000000000EbCTN^<l1;0000O2O00001O00001O00001O00001O1N2O001O1O1N2O1O1O1N2O1O1O1N2O1O1O1N4M4L5K4L5J5L2N3M2MoS7"}, "label": "a white car drives down the street"}]}
{"id": 65962, "image": "COCO_train2014_000000065962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"9424 bus\"."}], "task": "refering", "answer_template": "The \"9424 bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283], "bbox": [0.224609375, 0.30445983379501385, 0.6700625, 0.9353462603878115], "iscrowd": 0, "area": 51138.25925, "rle": {"size": [361, 640], "counts": "Vnb12T;4L3M4L3L5J5L4N3N1O1N1010SOnEc0R:]OPF`0P:@SF>m9BUF<k9CXF:h9G[F6e9J^F2c9MaF0^91eFL[93n00O1Io0YOQ1fN[1dHdL\\4g4bJeKo4^5N2N2N1O1N100000001N10000000000O1000001O0O10000000000O1000001O2M3N2N3M2N3M1N10000000000O10001OO101N1O100O1L4J600000000000000000000O1000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000O10O1000000000000000000000000000000000000000O10000000000000000000000000000000000000O10O1000000000000000000000001O000000001O0000000O2O001O1O1O001O1O1O001O1O001O1O1O001O1O1BgJjJY5V5iJhJX5W5jJfJX5Y5jJeJV5[5lJcJU5[5>O1O0O2O1O1O001O1O0O2O1O13L4M3MO1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1N2O001O1N101O1O1J9Gc0]Ob0_Ob0_O`0AZ1fNTTZ2"}, "label": "9424 bus"}]}
{"id": 65962, "image": "COCO_train2014_000000065962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a city bus , number 42 , driving through an intersection\"."}], "task": "refering", "answer_template": "The \"a city bus , number 42 , driving through an intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283], "bbox": [0.224609375, 0.30445983379501385, 0.6700625, 0.9353462603878115], "iscrowd": 0, "area": 51138.25925, "rle": {"size": [361, 640], "counts": "Vnb12T;4L3M4L3L5J5L4N3N1O1N1010SOnEc0R:]OPF`0P:@SF>m9BUF<k9CXF:h9G[F6e9J^F2c9MaF0^91eFL[93n00O1Io0YOQ1fN[1dHdL\\4g4bJeKo4^5N2N2N1O1N100000001N10000000000O1000001O0O10000000000O1000001O2M3N2N3M2N3M1N10000000000O10001OO101N1O100O1L4J600000000000000000000O1000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000O10O1000000000000000000000000000000000000000O10000000000000000000000000000000000000O10O1000000000000000000000001O000000001O0000000O2O001O1O1O001O1O1O001O1O001O1O1O001O1O1BgJjJY5V5iJhJX5W5jJfJX5Y5jJeJV5[5lJcJU5[5>O1O0O2O1O1O001O1O0O2O1O13L4M3MO1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1N2O001O1N101O1O1J9Gc0]Ob0_Ob0_O`0AZ1fNTTZ2"}, "label": "a city bus , number 42 , driving through an intersection"}]}
{"id": 188845, "image": "COCO_train2014_000000188845.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male playing wii with stars above his head\"."}], "task": "refering", "answer_template": "The \"a male playing wii with stars above his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 39, 40, 49, 50, 51, 52, 59, 60, 61, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 104, 105, 106, 107, 114, 115, 116, 117, 118, 125, 126, 127, 128, 129, 136, 137, 138, 139, 140, 147, 148, 150, 151, 158, 159, 161, 162, 169, 170, 172, 173, 180, 181], "bbox": [0.3976, 0.12979166666666667, 0.9227333333333333, 0.9733541666666667], "iscrowd": 0, "area": 32477.78435000001, "rle": {"size": [480, 300], "counts": "_Vh12h>8H7lAFZ=a0cBFS=`0jBBP=e0mB^OT7B\\NW1^JXOc66gN^1P1lNhN[1o0POhNV1S1RObNW1Z1oNUN_1g1eNlMi1P2ZNcMS2Z2PNXM^2e2eMnLi2o2YMhLP3V3QMgLS3W3oLeLU3Y3mLcLW3[3jLbL[3\\3fLRLl3m3TLmKS4S4mKeK[4Z4jK[K]4e4o2O10\\OXKPGh4n8]KoFb4P9dKiF_4U9f0O2O001O010O10O@a0^Oa0F;M2M4J5MVOcHYJZ7V5nHSJOd0Q7X5dIfJZ6X5lIfJR6X5SJfJk5\\5XJaJg5a5XJ`Je5b5\\J]Jc5d5bJXJ\\5j5hJQJV5Q6oJkIo4V6k12O2N1N2O2M2O1O2M3M3M3M3M4N101O04M3L4L4M4K4M3L5L3L4L4M3LfHTJd5i5XJ\\Jn5\\5PJhJW6P5eISKd6d4ZI^Kn6[4nHiKU7S4iHoKZ7o3UH`KDd0Z8j3kGkKD>e8f3_GoLd8iN\\G]5U9N010O0010O010O01O01O010O01O010O00PIdJ]4\\5WK\\L\\3d3XLnLb3S3_LkLb3T3aLjL^3V3eLgL\\3Y3dLjLX3W3iLoLK_M[1b5POQM[OmMV1T5]OWMSOUNS1e4I]MkN^Nm0V47eMaNeNk0g3b0c0QO_On0k0gNVOX1k0hNVOV1k0iNWOV1i0hNYOW1Z7N2N3M2O1N1O1O01O000000O1O001O1O2N1N2N10O0003M3N2L4M3L4L4Mcc:"}, "label": "a male playing wii with stars above his head"}]}
{"id": 188845, "image": "COCO_train2014_000000188845.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in purple shirt playing wii\"."}], "task": "refering", "answer_template": "The \"man in purple shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 39, 40, 49, 50, 51, 52, 59, 60, 61, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 104, 105, 106, 107, 114, 115, 116, 117, 118, 125, 126, 127, 128, 129, 136, 137, 138, 139, 140, 147, 148, 150, 151, 158, 159, 161, 162, 169, 170, 172, 173, 180, 181], "bbox": [0.3976, 0.12979166666666667, 0.9227333333333333, 0.9733541666666667], "iscrowd": 0, "area": 32477.78435000001, "rle": {"size": [480, 300], "counts": "_Vh12h>8H7lAFZ=a0cBFS=`0jBBP=e0mB^OT7B\\NW1^JXOc66gN^1P1lNhN[1o0POhNV1S1RObNW1Z1oNUN_1g1eNlMi1P2ZNcMS2Z2PNXM^2e2eMnLi2o2YMhLP3V3QMgLS3W3oLeLU3Y3mLcLW3[3jLbL[3\\3fLRLl3m3TLmKS4S4mKeK[4Z4jK[K]4e4o2O10\\OXKPGh4n8]KoFb4P9dKiF_4U9f0O2O001O010O10O@a0^Oa0F;M2M4J5MVOcHYJZ7V5nHSJOd0Q7X5dIfJZ6X5lIfJR6X5SJfJk5\\5XJaJg5a5XJ`Je5b5\\J]Jc5d5bJXJ\\5j5hJQJV5Q6oJkIo4V6k12O2N1N2O2M2O1O2M3M3M3M3M4N101O04M3L4L4M4K4M3L5L3L4L4M3LfHTJd5i5XJ\\Jn5\\5PJhJW6P5eISKd6d4ZI^Kn6[4nHiKU7S4iHoKZ7o3UH`KDd0Z8j3kGkKD>e8f3_GoLd8iN\\G]5U9N010O0010O010O01O01O010O01O010O00PIdJ]4\\5WK\\L\\3d3XLnLb3S3_LkLb3T3aLjL^3V3eLgL\\3Y3dLjLX3W3iLoLK_M[1b5POQM[OmMV1T5]OWMSOUNS1e4I]MkN^Nm0V47eMaNeNk0g3b0c0QO_On0k0gNVOX1k0hNVOV1k0iNWOV1i0hNYOW1Z7N2N3M2O1N1O1O01O000000O1O001O1O2N1N2N10O0003M3N2L4M3L4L4Mcc:"}, "label": "man in purple shirt playing wii"}]}
{"id": 188845, "image": "COCO_train2014_000000188845.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"women holding a white game remote , with pow written in red over it\"."}], "task": "refering", "answer_template": "The \"women holding a white game remote , with pow written in red over it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 45, 46, 47, 56, 57, 58, 59, 66, 67, 68, 69, 70, 78, 79, 80, 81, 88, 89, 90, 91, 92, 99, 100, 101, 102, 103, 110, 111, 112, 113, 114, 121, 122, 123, 124, 133, 134, 135, 143, 144, 145, 146, 154, 155, 156, 157, 165, 166, 167, 168, 176, 177, 178, 179], "bbox": [0.0395, 0.18666666666666665, 0.41769999999999996, 0.9798333333333334], "iscrowd": 0, "area": 26831.3487, "rle": {"size": [480, 300], "counts": "gQ6=c9HTKb0b4HUKa0a4HWK<d4NoJ:k40cJ>X5LTJb0h5HnI=m5LkI8P62gI1T67gILT69hIJS4XM]LS3\\OHU4kMfKc21EW4PN\\Kb2:@X4VNRKa2d0ZOX4]NhJ`2n0TOX4a2fK`MY4b2eK_MZ4m2ZKPLKfNk4e7iKnGW4Q8nKkGR4T8RLjGn3T8VLiGj3V8`1N3N1O1O2N1O1N3M101N101O10O1^OXHmIh7l5cHPJ^7i5kHSJV7k5nHSJR7n5nHPJS7P6oHnIQ7R6QIlIo6T6o00M2N2N2M4M2M3N23iKfGd1]8fLdG;a0h2o7eL^IY3f6^LaI`3b6YLdIf3l8O1O1O1O1O1O1O1O1O001OYOdLkE[3n9mLRFR3f9WMYFj2`9]M`Fb2Z9eMeF[2U9lMkFS2P9RNPGo1i8XNWGg1c8`N\\G`1^8gN`G[1U8POiGQ1m7ZOQHg0f7CXH?d7D[H=b7G\\H:a7J]H8_7L_H<W7HgH`0Q7BmHf0l6Z4kM^GeLe8Y3aG`Lb8^3cG\\L`8b3fGWL]8g3iGSLY8k3lGdK`8\\4S1O001O001[N^EjNb:U1cEgN^:V1hEfNX:Y1lEcNV:[1oEaNQ:_1SF]NT:[1QFaNW:V1nEfNZ:Q1kEkN]:l0iEmNa:k0cElNj:i0cla2"}, "label": "women holding a white game remote , with pow written in red over it"}]}
{"id": 188845, "image": "COCO_train2014_000000188845.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman playing video games\"."}], "task": "refering", "answer_template": "The \"a woman playing video games\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 45, 46, 47, 56, 57, 58, 59, 66, 67, 68, 69, 70, 78, 79, 80, 81, 88, 89, 90, 91, 92, 99, 100, 101, 102, 103, 110, 111, 112, 113, 114, 121, 122, 123, 124, 133, 134, 135, 143, 144, 145, 146, 154, 155, 156, 157, 165, 166, 167, 168, 176, 177, 178, 179], "bbox": [0.0395, 0.18666666666666665, 0.41769999999999996, 0.9798333333333334], "iscrowd": 0, "area": 26831.3487, "rle": {"size": [480, 300], "counts": "gQ6=c9HTKb0b4HUKa0a4HWK<d4NoJ:k40cJ>X5LTJb0h5HnI=m5LkI8P62gI1T67gILT69hIJS4XM]LS3\\OHU4kMfKc21EW4PN\\Kb2:@X4VNRKa2d0ZOX4]NhJ`2n0TOX4a2fK`MY4b2eK_MZ4m2ZKPLKfNk4e7iKnGW4Q8nKkGR4T8RLjGn3T8VLiGj3V8`1N3N1O1O2N1O1N3M101N101O10O1^OXHmIh7l5cHPJ^7i5kHSJV7k5nHSJR7n5nHPJS7P6oHnIQ7R6QIlIo6T6o00M2N2N2M4M2M3N23iKfGd1]8fLdG;a0h2o7eL^IY3f6^LaI`3b6YLdIf3l8O1O1O1O1O1O1O1O1O001OYOdLkE[3n9mLRFR3f9WMYFj2`9]M`Fb2Z9eMeF[2U9lMkFS2P9RNPGo1i8XNWGg1c8`N\\G`1^8gN`G[1U8POiGQ1m7ZOQHg0f7CXH?d7D[H=b7G\\H:a7J]H8_7L_H<W7HgH`0Q7BmHf0l6Z4kM^GeLe8Y3aG`Lb8^3cG\\L`8b3fGWL]8g3iGSLY8k3lGdK`8\\4S1O001O001[N^EjNb:U1cEgN^:V1hEfNX:Y1lEcNV:[1oEaNQ:_1SF]NT:[1QFaNW:V1nEfNZ:Q1kEkN]:l0iEmNa:k0cElNj:i0cla2"}, "label": "a woman playing video games"}]}
{"id": 188847, "image": "COCO_train2014_000000188847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver colored bus is stopping at signal\"."}], "task": "refering", "answer_template": "The \"a silver colored bus is stopping at signal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214], "bbox": [0.0063125, 0.32358333333333333, 0.33496875, 0.7415833333333333], "iscrowd": 0, "area": 26808.655350000005, "rle": {"size": [360, 640], "counts": "Xa1h0a:a1^N\\1eNO1O001O001O1O001O001O1O00O100000001N10000000000O100000000O2O1O001O1O1N101O1O001O1O1O001O1O000000O10000000000O10000000000O1000O1O100O1O00100O000001O0001O100O1O2O0O100O1O100O100O2O0O1O100O1000001O000O1000000000001O0000000O1000001O00000000000O1001O0000000000001O000O1000001O0000000000001O00000000001O00000000001O00000O1000001O00000000001O0000000000001O001O1O001O1O001O001N2O001O1O001O001O1O001O1O1O1O1O1O1O1OT3lL4L3M4L``e4"}, "label": "a silver colored bus is stopping at signal"}]}
{"id": 188847, "image": "COCO_train2014_000000188847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver bus with green stripes on the road behind a blue - green taxi\"."}], "task": "refering", "answer_template": "The \"a silver bus with green stripes on the road behind a blue - green taxi\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214], "bbox": [0.0063125, 0.32358333333333333, 0.33496875, 0.7415833333333333], "iscrowd": 0, "area": 26808.655350000005, "rle": {"size": [360, 640], "counts": "Xa1h0a:a1^N\\1eNO1O001O001O1O001O001O1O00O100000001N10000000000O100000000O2O1O001O1O1N101O1O001O1O1O001O1O000000O10000000000O10000000000O1000O1O100O1O00100O000001O0001O100O1O2O0O100O1O100O100O2O0O1O100O1000001O000O1000000000001O0000000O1000001O00000000000O1001O0000000000001O000O1000001O0000000000001O00000000001O00000000001O00000O1000001O00000000001O0000000000001O001O1O001O1O001O001N2O001O1O001O001O1O001O1O1O1O1O1O1O1OT3lL4L3M4L``e4"}, "label": "a silver bus with green stripes on the road behind a blue - green taxi"}]}
{"id": 139696, "image": "COCO_train2014_000000139696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher crouching in front of the umpire\"."}], "task": "refering", "answer_template": "The \"a catcher crouching in front of the umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 192, 193, 194, 215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 306, 307, 308, 309, 330, 331], "bbox": [0.30989062500000003, 0.4364646464646465, 0.59921875, 0.8225454545454546], "iscrowd": 0, "area": 16049.246199999998, "rle": {"size": [495, 640], "counts": "^QP31X?9F9SA@`>l0AROUBU1f=>K4L4L5K6J5K6L3M4L2N1O2dCTMR<m2mCUMQ<l2nCVMP<l2mCVMR<U3N2M3N101N2N2N2O0O2N2N2O0O2N2N2O0O2N1O2O1N1DeKcE]4\\:dKbE^4\\:eKbE[4]:eKcE]4\\:cKcE_4[:bKdE_4\\:`KcEc4[:9G:L4M2M4L3N1010O1O100O00100O1[OYFbKg9^4[F_Ke9a4]F^Kc9b4^F]Kb9b4`F]K`9c4bFSKKOc9m4f02O1O1O10ZLhES2X:mMlEP2T:oMPFn1o9RNQFo1n9RNRFn1n9QNRFo1n9RNRFm1o9SNPFm1P:UNoEj1R:VNmEj1S:WNlEi1T:XNlEg1U:YNjEh1U:ZNjEc1Y:YNjEe1X:VNnEZ1b:dN_E\\1a:cNaE[1a:bNaE^1_:aNcE^1]:aNdE^1^:QNmDKf0T2^:nMoDLe0T2_:mMnDLe0W2_:lMlDKg0W2`:lMjDKg0Y2`:kMiDKh0Z2a:iMhDKh0[2c:iMfEW2\\:gMdEX2^:gMbEY2`:eMaEZ2j;O100O2O0O1mNkBLU=0PCOQ=0PCNQ=4mBLS=7kBFW=GhBI7=T=IiBA8d0Q=JZC2i<NXCOj<1VCMm<2TCJo<5SCMj<3VC0g<0ZC2d<M]C2c<M^C0e<0\\CMf<3\\1O0O100010O000O100O2O000O100O2O000O1O2N1O10001O1O1O1O1O1O1O001O100O1O1O1OO10000O100O100O10000O101N101N2O2M3N1O2Mi^k3"}, "label": "a catcher crouching in front of the umpire"}]}
{"id": 139696, "image": "COCO_train2014_000000139696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher behind the plate\"."}], "task": "refering", "answer_template": "The \"the catcher behind the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 192, 193, 194, 215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 306, 307, 308, 309, 330, 331], "bbox": [0.30989062500000003, 0.4364646464646465, 0.59921875, 0.8225454545454546], "iscrowd": 0, "area": 16049.246199999998, "rle": {"size": [495, 640], "counts": "^QP31X?9F9SA@`>l0AROUBU1f=>K4L4L5K6J5K6L3M4L2N1O2dCTMR<m2mCUMQ<l2nCVMP<l2mCVMR<U3N2M3N101N2N2N2O0O2N2N2O0O2N2N2O0O2N1O2O1N1DeKcE]4\\:dKbE^4\\:eKbE[4]:eKcE]4\\:cKcE_4[:bKdE_4\\:`KcEc4[:9G:L4M2M4L3N1010O1O100O00100O1[OYFbKg9^4[F_Ke9a4]F^Kc9b4^F]Kb9b4`F]K`9c4bFSKKOc9m4f02O1O1O10ZLhES2X:mMlEP2T:oMPFn1o9RNQFo1n9RNRFn1n9QNRFo1n9RNRFm1o9SNPFm1P:UNoEj1R:VNmEj1S:WNlEi1T:XNlEg1U:YNjEh1U:ZNjEc1Y:YNjEe1X:VNnEZ1b:dN_E\\1a:cNaE[1a:bNaE^1_:aNcE^1]:aNdE^1^:QNmDKf0T2^:nMoDLe0T2_:mMnDLe0W2_:lMlDKg0W2`:lMjDKg0Y2`:kMiDKh0Z2a:iMhDKh0[2c:iMfEW2\\:gMdEX2^:gMbEY2`:eMaEZ2j;O100O2O0O1mNkBLU=0PCOQ=0PCNQ=4mBLS=7kBFW=GhBI7=T=IiBA8d0Q=JZC2i<NXCOj<1VCMm<2TCJo<5SCMj<3VC0g<0ZC2d<M]C2c<M^C0e<0\\CMf<3\\1O0O100010O000O100O2O000O100O2O000O1O2N1O10001O1O1O1O1O1O1O001O100O1O1O1OO10000O100O100O10000O101N101N2O2M3N1O2Mi^k3"}, "label": "the catcher behind the plate"}]}
{"id": 139696, "image": "COCO_train2014_000000139696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 128, 129, 130, 131, 151, 152, 153, 154, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 247, 248, 264, 265, 266, 267, 270, 271, 272, 287, 288, 289, 294, 295], "bbox": [0.499328125, 0.18002020202020202, 0.862484375, 0.7172727272727273], "iscrowd": 0, "area": 17080.3675, "rle": {"size": [495, 640], "counts": "YPk42\\?2N2N1O2N2N2N2N1O2N2N2N1O2N2O1N2O001N2O1O1N101O1N2O1O0O2O1O1N100001N100000001N101O001L4XOg001N101O00O010O1YJHWL7h33oKNP47lKIR4=PKZO^L9b8b0lJVOaL8b8f0iJTOcL7d8h0eJSOfL4e8m0bJPOfL5h8n0^JoNgL5k8o0[JmNgL5o8Q1VJlNhL5R9R1RJkNhL6V9R1oI@Q6c0kI]OV6f0gIZOY6i0cIXO]6k0`ITO`6P1\\IQOc6S1YIlNi6W1TIhNl6\\1PIdNP7`1mH_NS7e1iH[NQ7i2QHWMa7oN]GS4P1lLe7^3WHcLi7_3SHcLn7_3mGcLS8`3hGbLY8`3cGaL]8b3^G`Lc8a3YGaLg8b3UG`Lk8Q501O0100O100O100O10O0100O1O2M2N2O2M2N2O2M2N3N2M3M3N2M4L3N2M3M3M4M2M3M4M2M4L3N2WLRE91c1T;hMTE`00a1e;ZNcDa1_;[NfDb1\\;YNjDd1X;XNmDd1`<01O001N2O001O0O2O001O1N110O001O0010O0001O010O001O0010O00100O1O1O1O1O001O1O1O1O1O1O102M3M3N1NO101N1O1O2N2O1N2N2N2N2O1N2N2N2N2O1N2N1O1O010O1O1O001O100O1O001O100O1O1O1O010O1O1O1O10O01O1O00100O002N2M3L4M2MQWZ1"}, "label": "a baseball player swinging a bat"}]}
{"id": 139696, "image": "COCO_train2014_000000139696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player swinging the bat on the home plate with the umpire and another baseball player\"."}], "task": "refering", "answer_template": "The \"baseball player swinging the bat on the home plate with the umpire and another baseball player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 128, 129, 130, 131, 151, 152, 153, 154, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 247, 248, 264, 265, 266, 267, 270, 271, 272, 287, 288, 289, 294, 295], "bbox": [0.499328125, 0.18002020202020202, 0.862484375, 0.7172727272727273], "iscrowd": 0, "area": 17080.3675, "rle": {"size": [495, 640], "counts": "YPk42\\?2N2N1O2N2N2N2N1O2N2N2N1O2N2O1N2O001N2O1O1N101O1N2O1O0O2O1O1N100001N100000001N101O001L4XOg001N101O00O010O1YJHWL7h33oKNP47lKIR4=PKZO^L9b8b0lJVOaL8b8f0iJTOcL7d8h0eJSOfL4e8m0bJPOfL5h8n0^JoNgL5k8o0[JmNgL5o8Q1VJlNhL5R9R1RJkNhL6V9R1oI@Q6c0kI]OV6f0gIZOY6i0cIXO]6k0`ITO`6P1\\IQOc6S1YIlNi6W1TIhNl6\\1PIdNP7`1mH_NS7e1iH[NQ7i2QHWMa7oN]GS4P1lLe7^3WHcLi7_3SHcLn7_3mGcLS8`3hGbLY8`3cGaL]8b3^G`Lc8a3YGaLg8b3UG`Lk8Q501O0100O100O100O10O0100O1O2M2N2O2M2N2O2M2N3N2M3M3N2M4L3N2M3M3M4M2M3M4M2M4L3N2WLRE91c1T;hMTE`00a1e;ZNcDa1_;[NfDb1\\;YNjDd1X;XNmDd1`<01O001N2O001O0O2O001O1N110O001O0010O0001O010O001O0010O00100O1O1O1O1O001O1O1O1O1O1O102M3M3N1NO101N1O1O2N2O1N2N2N2N2O1N2N2N2N2O1N2N1O1O010O1O1O001O100O1O001O100O1O1O1O010O1O1O1O10O01O1O00100O002N2M3L4M2MQWZ1"}, "label": "baseball player swinging the bat on the home plate with the umpire and another baseball player"}]}
{"id": 139696, "image": "COCO_train2014_000000139696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a refree behind the wicket keeper on the baseball ground\"."}], "task": "refering", "answer_template": "The \"a refree behind the wicket keeper on the baseball ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 141, 142, 143, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 278, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 348], "bbox": [0.048671874999999996, 0.3146060606060606, 0.358046875, 0.8382020202020202], "iscrowd": 0, "area": 21909.3279, "rle": {"size": [495, 640], "counts": "kW?3Z?8G9H8G9H3M3L3N3N1N3N1O2M2O2N1N3N2N1N3N1N3N1O2M2O2N1N3N1oE\\MV7f2gH]MV7f2fH_MW7b2fHaMX7b2dHaM[7`2bHcM]7_2_HdM_7^2_HdM`7^2[HfMd7[2XHiMf7Z2UHjMj7W2RHmMm7U2oGnMo7T2mGoMT8R2gGRNX8P2cGTN]8l1_GXN`8j1[GZNd8g1XG]Nh8d1TG_Nk8b1QGbNo8_1lFeNS9l30O0100O1O100O1^GeIP8[6nGgIQ8Z6mGhIS8X6kGjIT8V6kGlIT8U6jGmIV8S6hGoIW8R6gGoIZ8Q6dGQJ[8b600O100O1O100O100O100O1N101N2O1N2O1N2O1N2O1N2O1N2O1NRL`H0`7LfH3X7JnH5R7FTI9l6B[I<e6_OaI`0_6^OdIa0\\6@dI?]6AcI>]6CdI;\\6FdI9\\6HdI7\\6IeI6[6KeI4[6MfI0\\60dIO\\62dIM\\64dIK\\66dII\\67fIVOiNkMa7P3fIoNROmMX7U3fIhNZOPNQ7X3eIbNCRNh6]3fIYNIYNa6_3fIRNO^N[6R2_HXOW124dNV6P2dHWOS1O9iNQ6m1hHZOn0I?POl5j1lH[Oj0D[8n0oF\\Og0@`8Q1nF\\Oc0]Oe8S1mF_O>WOl8W1jF@;SOQ9Z1iF@h:<]EBd:;aEB`:;dEC]::hECY:9kEGU:6nEIS:4oELR:1PFOP:JWF6j9B]F>[<0O00000001O01O000001O0001O01O000000010O000000001O01O000000000010O0000000001O00O100O2N100O100O1O100O100O1O2O0O1O100O100O1O10V]V6"}, "label": "a refree behind the wicket keeper on the baseball ground"}]}
{"id": 139696, "image": "COCO_train2014_000000139696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire\"."}], "task": "refering", "answer_template": "The \"the umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 141, 142, 143, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 278, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 348], "bbox": [0.048671874999999996, 0.3146060606060606, 0.358046875, 0.8382020202020202], "iscrowd": 0, "area": 21909.3279, "rle": {"size": [495, 640], "counts": "kW?3Z?8G9H8G9H3M3L3N3N1N3N1O2M2O2N1N3N2N1N3N1N3N1O2M2O2N1N3N1oE\\MV7f2gH]MV7f2fH_MW7b2fHaMX7b2dHaM[7`2bHcM]7_2_HdM_7^2_HdM`7^2[HfMd7[2XHiMf7Z2UHjMj7W2RHmMm7U2oGnMo7T2mGoMT8R2gGRNX8P2cGTN]8l1_GXN`8j1[GZNd8g1XG]Nh8d1TG_Nk8b1QGbNo8_1lFeNS9l30O0100O1O100O1^GeIP8[6nGgIQ8Z6mGhIS8X6kGjIT8V6kGlIT8U6jGmIV8S6hGoIW8R6gGoIZ8Q6dGQJ[8b600O100O1O100O100O100O1N101N2O1N2O1N2O1N2O1N2O1N2O1NRL`H0`7LfH3X7JnH5R7FTI9l6B[I<e6_OaI`0_6^OdIa0\\6@dI?]6AcI>]6CdI;\\6FdI9\\6HdI7\\6IeI6[6KeI4[6MfI0\\60dIO\\62dIM\\64dIK\\66dII\\67fIVOiNkMa7P3fIoNROmMX7U3fIhNZOPNQ7X3eIbNCRNh6]3fIYNIYNa6_3fIRNO^N[6R2_HXOW124dNV6P2dHWOS1O9iNQ6m1hHZOn0I?POl5j1lH[Oj0D[8n0oF\\Og0@`8Q1nF\\Oc0]Oe8S1mF_O>WOl8W1jF@;SOQ9Z1iF@h:<]EBd:;aEB`:;dEC]::hECY:9kEGU:6nEIS:4oELR:1PFOP:JWF6j9B]F>[<0O00000001O01O000001O0001O01O000000010O000000001O01O000000000010O0000000001O00O100O2N100O100O1O100O100O1O2O0O1O100O100O1O10V]V6"}, "label": "the umpire"}]}
{"id": 328113, "image": "COCO_train2014_000000328113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheep with black face stands parallel to a tall stone fence\"."}], "task": "refering", "answer_template": "The \"a white sheep with black face stands parallel to a tall stone fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 291, 292, 312], "bbox": [0.511515625, 0.43620386643233744, 0.820296875, 0.6638664323374341], "iscrowd": 0, "area": 13909.4397, "rle": {"size": [569, 640], "counts": "Woe51fa02O0000008a^OHm`0d0J6M3M207H9Go0QO2O00000O1O1O2N100O100001O1O10O0O101O001N2O1N2N3N2L3O2M3NfLdAX3^>2N1O2N_OjA`MT>\\2UBbMh=]2^B`M`=^2dBcM[=Z2iBeMW=V2oBjMQ=o1UCQNj<j1]CVNb<h1`CWN`<j1aCUN^<k1dCSN[<T2b13N1N1O100O1O1O010O000010O00010O0001O000O10000001O10O01O000000001O0O10001O0O100O2O0O1O100O10000O10000000n@SNX>n1dAWNZ>j1aA\\N]>`201O000000M3L3L4M4K4L4K6LbAkM`=U2ZBUNb=j1[BZNf=d1XB_Nh=`1VBdNh=\\1VB]N@IZ>j1RB_NY>b1cA`N_>a1]A]Nh>W23N2O1N2N2N2O0010O001O1O2OO00001O0N201O001O001M3N2N2O1N2L4M201O1O1O1M3N2N3N1N2N2F;O00001O1O1O001O1O1O1O1O1O100O1O100O001O1O1N2O1N3J`do1"}, "label": "a white sheep with black face stands parallel to a tall stone fence"}]}
{"id": 328113, "image": "COCO_train2014_000000328113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheet standing in front of the wall\"."}], "task": "refering", "answer_template": "The \"sheet standing in front of the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 291, 292, 312], "bbox": [0.511515625, 0.43620386643233744, 0.820296875, 0.6638664323374341], "iscrowd": 0, "area": 13909.4397, "rle": {"size": [569, 640], "counts": "Woe51fa02O0000008a^OHm`0d0J6M3M207H9Go0QO2O00000O1O1O2N100O100001O1O10O0O101O001N2O1N2N3N2L3O2M3NfLdAX3^>2N1O2N_OjA`MT>\\2UBbMh=]2^B`M`=^2dBcM[=Z2iBeMW=V2oBjMQ=o1UCQNj<j1]CVNb<h1`CWN`<j1aCUN^<k1dCSN[<T2b13N1N1O100O1O1O010O000010O00010O0001O000O10000001O10O01O000000001O0O10001O0O100O2O0O1O100O10000O10000000n@SNX>n1dAWNZ>j1aA\\N]>`201O000000M3L3L4M4K4L4K6LbAkM`=U2ZBUNb=j1[BZNf=d1XB_Nh=`1VBdNh=\\1VB]N@IZ>j1RB_NY>b1cA`N_>a1]A]Nh>W23N2O1N2N2N2O0010O001O1O2OO00001O0N201O001O001M3N2N2O1N2L4M201O1O1O1M3N2N3N1N2N2F;O00001O1O1O001O1O1O1O1O1O100O1O100O001O1O1N2O1N3J`do1"}, "label": "sheet standing in front of the wall"}]}
{"id": 328113, "image": "COCO_train2014_000000328113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the animal laying down closest to the camera\"."}], "task": "refering", "answer_template": "The \"the animal laying down closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376, 396, 397, 398], "bbox": [0.03640625, 0.617451669595782, 0.36021875000000003, 0.8653075571177504], "iscrowd": 0, "area": 16746.94765, "rle": {"size": [569, 640], "counts": "he=7ba01O1O2N1O1O1O010O1O1O1O1O1O100O1__O^Oj?g0d_OMT`0R1K4L2M1O1O1O1O1O1O1O1O1O000000001O0O2O2N3N4K4M2N3L2O000O1O1O1O1O1O01O01O010O000100O10O00010O0000001O00001O00001O0O10001O00ZM`AS2`>lMaAU2^>iMeAEL]2_>lMgAFJ^2`>jMgAHI_2_>hMjAHG`2`>fMjAJFa2`>cMkALEb2c>]MjA0Cc2T?1O001O1O001O001O1O001O001O00010O0000001O0000001O01O00010O001O01O01O00001O000000000010O000000001O0001O0001O00000000010O000001O000000001O000000001O0000001O0000001O00001O0000001O0000001O001O001N2O001O1O001O001O1O1O1O2M2O1O1O2N1O1M3N3M2M5Lf0YO\\jR7"}, "label": "the animal laying down closest to the camera"}]}
{"id": 328113, "image": "COCO_train2014_000000328113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep laying down with the pink spot\"."}], "task": "refering", "answer_template": "The \"the sheep laying down with the pink spot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376, 396, 397, 398], "bbox": [0.03640625, 0.617451669595782, 0.36021875000000003, 0.8653075571177504], "iscrowd": 0, "area": 16746.94765, "rle": {"size": [569, 640], "counts": "he=7ba01O1O2N1O1O1O010O1O1O1O1O1O100O1__O^Oj?g0d_OMT`0R1K4L2M1O1O1O1O1O1O1O1O1O000000001O0O2O2N3N4K4M2N3L2O000O1O1O1O1O1O01O01O010O000100O10O00010O0000001O00001O00001O0O10001O00ZM`AS2`>lMaAU2^>iMeAEL]2_>lMgAFJ^2`>jMgAHI_2_>hMjAHG`2`>fMjAJFa2`>cMkALEb2c>]MjA0Cc2T?1O001O1O001O001O1O001O001O00010O0000001O0000001O01O00010O001O01O01O00001O000000000010O000000001O0001O0001O00000000010O000001O000000001O000000001O0000001O0000001O00001O0000001O0000001O001O001N2O001O1O001O001O1O1O1O2M2O1O1O2N1O1M3N3M2M5Lf0YO\\jR7"}, "label": "the sheep laying down with the pink spot"}]}
{"id": 328113, "image": "COCO_train2014_000000328113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep sprayed with blue spray paint laying in the field / grass . next to the sheep a smaller sheep with red spray paint on the fur\"."}], "task": "refering", "answer_template": "The \"sheep sprayed with blue spray paint laying in the field / grass . next to the sheep a smaller sheep with red spray paint on the fur\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381, 382, 383, 399, 400, 401, 406], "bbox": [0.257015625, 0.5548154657293497, 0.6939375, 0.8702636203866433], "iscrowd": 0, "area": 26768.264000000006, "rle": {"size": [569, 640], "counts": "``k21ga03M1O1O1N2O1O1QOJY@8d?K[@6d?LZ@5d?NZ@3d?0Y@2f?0X@1f?2X@0g?1V@1i?2T@Ok?3R@Om?P1O0O1O1O2O0E;O1O2M2O1O1O100O100O2N10001N10001N101N101O000010O10O010O011N100O100N3N1O1O1O2N3M3L3N3Ml0TO6J6J4L4L2N001O1O1O001O1O100O100O1O01O001O00001O000O2O001O2N1nK]Cl2e<QM\\Co2e<oL\\CQ3f<lL\\CS3e<kL\\CU3f<hL[CX3f<fL[CY3h<dLYC\\3h<`L[C`3_=000000000000O1O100O1O101N1O1O100O1O1O101N1O1O1O100O1O1O2O0O1O1O100O1O1O100O100O1O101N1O100O1O100O1O2O0O1000000010O0000001O00010O0000001O01O0001O00000001O000001O0001O0000000000010O0000000000010O01O00001O0010O0001O00010O0010O00010O0010O0001O010O000010O0001O010O1O1O1O100O001O1O100O1O001O100O1O010O100O10000O100O2O0O100O10001N3N3L3M3L4L4L3L5J6J6J9Gn^\\3"}, "label": "sheep sprayed with blue spray paint laying in the field / grass . next to the sheep a smaller sheep with red spray paint on the fur"}]}
{"id": 328113, "image": "COCO_train2014_000000328113.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an animal laying down with blue on it ' s fur\"."}], "task": "refering", "answer_template": "The \"an animal laying down with blue on it ' s fur\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381, 382, 383, 399, 400, 401, 406], "bbox": [0.257015625, 0.5548154657293497, 0.6939375, 0.8702636203866433], "iscrowd": 0, "area": 26768.264000000006, "rle": {"size": [569, 640], "counts": "``k21ga03M1O1O1N2O1O1QOJY@8d?K[@6d?LZ@5d?NZ@3d?0Y@2f?0X@1f?2X@0g?1V@1i?2T@Ok?3R@Om?P1O0O1O1O2O0E;O1O2M2O1O1O100O100O2N10001N10001N101N101O000010O10O010O011N100O100N3N1O1O1O2N3M3L3N3Ml0TO6J6J4L4L2N001O1O1O001O1O100O100O1O01O001O00001O000O2O001O2N1nK]Cl2e<QM\\Co2e<oL\\CQ3f<lL\\CS3e<kL\\CU3f<hL[CX3f<fL[CY3h<dLYC\\3h<`L[C`3_=000000000000O1O100O1O101N1O1O100O1O1O101N1O1O1O100O1O1O2O0O1O1O100O1O1O100O100O1O101N1O100O1O100O1O2O0O1000000010O0000001O00010O0000001O01O0001O00000001O000001O0001O0000000000010O0000000000010O01O00001O0010O0001O00010O0010O00010O0010O0001O010O000010O0001O010O1O1O1O100O001O1O100O1O001O100O1O010O100O10000O100O2O0O100O10001N3N3L3M3L4L4L3L5J6J6J9Gn^\\3"}, "label": "an animal laying down with blue on it ' s fur"}]}
{"id": 33204, "image": "COCO_train2014_000000033204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather chair with a gray pillow and a red plaid blanket thrown over the back of it\"."}], "task": "refering", "answer_template": "The \"a brown leather chair with a gray pillow and a red plaid blanket thrown over the back of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 306, 307, 308, 309], "bbox": [0.26809375, 0.6638578680203046, 0.4833125000000001, 0.998020304568528], "iscrowd": 0, "area": 12244.349049999997, "rle": {"size": [394, 640], "counts": "g^R24T<5K4L5K4L4L4L3M4L4L4N3L4L5L4K5K5L3L5L4L4L4M2M4L4L4M3L3M4M3L4L4M1NO101O0000001O0O101G80001O001O1O001N2O1O001O1O001O1O1O001O1O1O0O2O1O001O1O1O001OO100000000000O100000000000O1000000000000000O100000O1000000000000000000000O10O1000000000000000000000155JN3J6J1O001O1O001G8B?A?@?BRRo3"}, "label": "a brown leather chair with a gray pillow and a red plaid blanket thrown over the back of it"}]}
{"id": 33204, "image": "COCO_train2014_000000033204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair with a red blanket on it\"."}], "task": "refering", "answer_template": "The \"a black chair with a red blanket on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 306, 307, 308, 309], "bbox": [0.26809375, 0.6638578680203046, 0.4833125000000001, 0.998020304568528], "iscrowd": 0, "area": 12244.349049999997, "rle": {"size": [394, 640], "counts": "g^R24T<5K4L5K4L4L4L3M4L4L4N3L4L5L4K5K5L3L5L4L4L4M2M4L4L4M3L3M4M3L4L4M1NO101O0000001O0O101G80001O001O1O001N2O1O001O1O001O1O1O001O1O1O0O2O1O001O1O1O001OO100000000000O100000000000O1000000000000000O100000O1000000000000000000000O10O1000000000000000000000155JN3J6J1O001O1O001G8B?A?@?BRRo3"}, "label": "a black chair with a red blanket on it"}]}
{"id": 33204, "image": "COCO_train2014_000000033204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the grey couch facing the windows\"."}], "task": "refering", "answer_template": "The \"the grey couch facing the windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321], "bbox": [0.7533125, 0.6255583756345178, 1.0, 0.9891878172588833], "iscrowd": 0, "area": 17456.86415, "rle": {"size": [394, 640], "counts": "bji52T<4K5K:G4K5L4K5M3N2N2M3N2N2M3N2N2M3N2N2N2M3N2M3N2M3M3N2M3M3N2M3O10000O10000O10000001O001O001O00001O001O00001O001O0000VOj0O1N2N200000000001O00000000000000O100O1O1O100O1O1O100O1O1000000000000001O0000000001O00000001O0000000000000000001O0000O1O1O1O1O1O100O1O1O1O1O1O100001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001_NYI"}, "label": "the grey couch facing the windows"}]}
{"id": 33204, "image": "COCO_train2014_000000033204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sofa sitting across the room from two black chairs\"."}], "task": "refering", "answer_template": "The \"a white sofa sitting across the room from two black chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321], "bbox": [0.7533125, 0.6255583756345178, 1.0, 0.9891878172588833], "iscrowd": 0, "area": 17456.86415, "rle": {"size": [394, 640], "counts": "bji52T<4K5K:G4K5L4K5M3N2N2M3N2N2M3N2N2M3N2N2N2M3N2M3N2M3M3N2M3M3N2M3O10000O10000O10000001O001O001O00001O001O00001O001O0000VOj0O1N2N200000000001O00000000000000O100O1O1O100O1O1O100O1O1000000000000001O0000000001O00000001O0000000000000000001O0000O1O1O1O1O1O100O1O1O1O1O1O100001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001_NYI"}, "label": "a white sofa sitting across the room from two black chairs"}]}
{"id": 303541, "image": "COCO_train2014_000000303541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking to the right who is next to a second giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe looking to the right who is next to a second giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 87, 88, 89, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 154, 155, 156, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360, 361, 380, 382, 383, 384, 403, 405, 406, 407, 426, 427, 428, 429, 430, 449, 450, 451, 452, 453, 472, 473, 474, 476, 499], "bbox": [0.49135937500000004, 0.10787167449139282, 0.9580468750000002, 0.9348356807511735], "iscrowd": 0, "area": 43936.71330000002, "rle": {"size": [639, 640], "counts": "[WT61mc0a0_O`0a_O[Od<V1TBDj=m0m@MQ?b2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2iDUJW8m5SG]KW8e4SGVLe8m3eFiLY9Y3RF[Mm9]6N1O1O2N1O2O0O2O001O0O2O00PLgF[MX9b2SGVMm8g2_GQM_8n2kGjLU8S3VHfLi7W3bHaL^7\\3mH]LQ7a3[IVLe6g3fIQLZ6l3QJjKQ6S4ZJcKh5Z4cJ\\K^5c4lJSKV5j4VKkJl4R5R5M2O2O1N2O1N101N2O1N101N2O1N2N101N2O1N101N_FgK`3X4XLQLh3m3QL]Lm3c3kKfLT4Y3eKQMZ4n2^K[Ma4d2XKfMf4Y2SKPNl4f2VJcMi5U3XITMf6e3YHdLf7T4ZGVLd8c7O2N1O2N1O2N1OhLX3J6J6J6K6I6J6J6J6K5J6J6J6J6K5J6M31O001O001O010O001UDaKU8_4_GYLU8h3^GQMU8o2_GiMU8X2^GaNU8_1_GYOU8h0_G0T80`Gh0T8o5Bb0^O2N1O1O1OlM[IPGd6Z8VJcGi5g7RKUHm4V7mKfHR4i6cLTI[3i6nLSIQ3i6XMSIg2i6bMSI\\2i6mMTIR2h6WNTIh1h6aNTI^1h6jNUIT1h6UOTIj0g6@UI?g6]6M4TMmBiNW=U1PCbNT=\\1RC[NR=d1TCTNP=j1WCmMm<Q2YCfMk<Y2\\C]Mi<b2\\CVMh<h2^CoLg<o2`CgLd<X3R2N2N1O2N2O1N1O2N2N1O2N1O100O2N1O1O1O1O100O2N1O1O1O1O1O101N1O1O1O1O1O1O2O0O1O11O101O0100O1O100O010O1O100O10O01O100O100O00100O100O1O10O011K4K5K61N1O2O1N2N2O0O2N2N2O0O2N2OO01O01O01O0010O00010O00010O01O01O01O00010O002O;D5K100O1O1O2O0O1O100O2N1O1O1O1O2N1O1O1O1N3N1O1O1O1O2N1O_f`0"}, "label": "a giraffe looking to the right who is next to a second giraffe"}]}
{"id": 303541, "image": "COCO_train2014_000000303541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking his right side\"."}], "task": "refering", "answer_template": "The \"a giraffe looking his right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 87, 88, 89, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 154, 155, 156, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360, 361, 380, 382, 383, 384, 403, 405, 406, 407, 426, 427, 428, 429, 430, 449, 450, 451, 452, 453, 472, 473, 474, 476, 499], "bbox": [0.49135937500000004, 0.10787167449139282, 0.9580468750000002, 0.9348356807511735], "iscrowd": 0, "area": 43936.71330000002, "rle": {"size": [639, 640], "counts": "[WT61mc0a0_O`0a_O[Od<V1TBDj=m0m@MQ?b2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2iDUJW8m5SG]KW8e4SGVLe8m3eFiLY9Y3RF[Mm9]6N1O1O2N1O2O0O2O001O0O2O00PLgF[MX9b2SGVMm8g2_GQM_8n2kGjLU8S3VHfLi7W3bHaL^7\\3mH]LQ7a3[IVLe6g3fIQLZ6l3QJjKQ6S4ZJcKh5Z4cJ\\K^5c4lJSKV5j4VKkJl4R5R5M2O2O1N2O1N101N2O1N101N2O1N2N101N2O1N101N_FgK`3X4XLQLh3m3QL]Lm3c3kKfLT4Y3eKQMZ4n2^K[Ma4d2XKfMf4Y2SKPNl4f2VJcMi5U3XITMf6e3YHdLf7T4ZGVLd8c7O2N1O2N1O2N1OhLX3J6J6J6K6I6J6J6J6K5J6J6J6J6K5J6M31O001O001O010O001UDaKU8_4_GYLU8h3^GQMU8o2_GiMU8X2^GaNU8_1_GYOU8h0_G0T80`Gh0T8o5Bb0^O2N1O1O1OlM[IPGd6Z8VJcGi5g7RKUHm4V7mKfHR4i6cLTI[3i6nLSIQ3i6XMSIg2i6bMSI\\2i6mMTIR2h6WNTIh1h6aNTI^1h6jNUIT1h6UOTIj0g6@UI?g6]6M4TMmBiNW=U1PCbNT=\\1RC[NR=d1TCTNP=j1WCmMm<Q2YCfMk<Y2\\C]Mi<b2\\CVMh<h2^CoLg<o2`CgLd<X3R2N2N1O2N2O1N1O2N2N1O2N1O100O2N1O1O1O1O100O2N1O1O1O1O1O101N1O1O1O1O1O1O2O0O1O11O101O0100O1O100O010O1O100O10O01O100O100O00100O100O1O10O011K4K5K61N1O2O1N2N2O0O2N2N2O0O2N2OO01O01O01O0010O00010O00010O01O01O01O00010O002O;D5K100O1O1O2O0O1O100O2N1O1O1O1O2N1O1O1O1N3N1O1O1O1O2N1O_f`0"}, "label": "a giraffe looking his right side"}]}
{"id": 303541, "image": "COCO_train2014_000000303541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult giraffe facing to the left\"."}], "task": "refering", "answer_template": "The \"an adult giraffe facing to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 47, 48, 49, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 118, 119, 120, 121, 141, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 350, 351, 352, 354, 355, 356, 373, 374, 377, 378, 379, 396, 397, 398, 400, 401, 402, 418, 419, 420, 421, 424, 425, 441, 442, 443, 447, 448, 470, 471], "bbox": [0.015703125000000002, 0.08089201877934273, 0.52278125, 0.8876369327073553], "iscrowd": 0, "area": 51619.551750000006, "rle": {"size": [639, 640], "counts": "\\[64cc08O1O1O1O001O1N2O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1QOoNn^OS1Sa0nNj^OS1Va0nNh^OR1Ya0oNe^OR1\\a0oNa^OR1_a0oN_^OR1aa0oN]^OR1ca0POZ^OQ1fa0POX^OQ1ga0QOV^OQ1ja0QOS^OP1ma0QOQ^OP1oa0?010000O10000O100000O1000000000O100000000000O11O2N1102M2O2N1NO2O0O2O1N101O0O2O0O2O1O0O2O0O2O1O001N101O001O1O001O001O1O001O001O001O1O001N101O1O001O001O1O001O001O0\\AoLY<R3dCTMX<m2fCYMU<h2hC^MT<d2hCbMT<_2iChMP5VO0T3mJlMh4WO;o2jJoMd4TOc0n2fJUN_4oNl0m2bJYN[4mNT1l2]J\\NY4kNZ1j2ZJ_NW4iN`1i2WJbNS4gNg1i2RJeNQ4eNm1g2oIhNo3cNS2f2kIlNl3`NZ2e2hIoNh1[MZOV1W5a2dIROd1]MVOZ1b5W2bIWO^1]MSO_1n5n1^I[OY1^MPOb1Z6f1ZIi12jLe6^1WIj1HSMQ7Q:1O100O1N2K5L5K4K5L4L4UOk0C=D=C<C=D<E;J6J6eNZDVJm;l5SDkIS<X6mC_IY<d6hCRI^<Q773`GkHo3X7iKnHV4U7aKRI^4R7YKTIg4n6bIgJ]6\\5XIlJh6W5mHRKR7Q5cHVK^7m4WHZKj7]81N3N1O1O1[OhGaEY8]:jG`EX8]:lGaEU8]:nGaET8\\:oGbER8\\:QHbEP8\\:SHbEn7V8fGaI`0VNl7T8RH[Ik8_6eGUI]8f6QHoHR8k6]HhHf7S7hHbHZ7U5fF[LS<e3aDbKd;^4o100O101O000O1000000O2O0000000O10001O0O10001O0100O10000O10O0100O100N2O1O1N101N2O1O1N2O1N2O001N2O1N2O1O1N2O0O2O1O1N4M3L3N3M3L4M2M4M3N1N3M2N100O1O2O0O1O1O100O1O2N105J8H9HlLYDUN_;c1nD^Nl:[1_EbN`:^1dE_N\\:a1gE\\NX:f1jEWNU:j1nESNQ:n1SFnMm9S2UFjMj9W2YFfMf9[2[FcMf9^2ZF_Mg9b2ZF[Mg9h2WFVMj9m2UFPMm9S3QFiLQ:Z3nEcLU:^3iE`L]:^3bE_Ld:`3ZE]Ll:a3RE]LT;f0iCa0R1eN\\;1ZDY19cNc;\\OkDQ2_OaNR>_1mA^NX>a1gA\\N^>d1`AYNd>h1YAVNl>k1QAQNU?P2g@nM]?T2^@kMg?W2T@gMQ`0Z2k_OdM^`0V2`_OiMk`0m1V_OQNVa0d1k^OYNaa0^1_^O`Nia0Y1W^OfNna0V1S^OhNRb0T1o]OjNVb0S1>M4L4L4K5K5K4L5KTim5"}, "label": "an adult giraffe facing to the left"}]}
{"id": 303541, "image": "COCO_train2014_000000303541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller giraffe\"."}], "task": "refering", "answer_template": "The \"the taller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 47, 48, 49, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 118, 119, 120, 121, 141, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 350, 351, 352, 354, 355, 356, 373, 374, 377, 378, 379, 396, 397, 398, 400, 401, 402, 418, 419, 420, 421, 424, 425, 441, 442, 443, 447, 448, 470, 471], "bbox": [0.015703125000000002, 0.08089201877934273, 0.52278125, 0.8876369327073553], "iscrowd": 0, "area": 51619.551750000006, "rle": {"size": [639, 640], "counts": "\\[64cc08O1O1O1O001O1N2O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1QOoNn^OS1Sa0nNj^OS1Va0nNh^OR1Ya0oNe^OR1\\a0oNa^OR1_a0oN_^OR1aa0oN]^OR1ca0POZ^OQ1fa0POX^OQ1ga0QOV^OQ1ja0QOS^OP1ma0QOQ^OP1oa0?010000O10000O100000O1000000000O100000000000O11O2N1102M2O2N1NO2O0O2O1N101O0O2O0O2O1O0O2O0O2O1O001N101O001O1O001O001O1O001O001O001O1O001N101O1O001O001O1O001O001O0\\AoLY<R3dCTMX<m2fCYMU<h2hC^MT<d2hCbMT<_2iChMP5VO0T3mJlMh4WO;o2jJoMd4TOc0n2fJUN_4oNl0m2bJYN[4mNT1l2]J\\NY4kNZ1j2ZJ_NW4iN`1i2WJbNS4gNg1i2RJeNQ4eNm1g2oIhNo3cNS2f2kIlNl3`NZ2e2hIoNh1[MZOV1W5a2dIROd1]MVOZ1b5W2bIWO^1]MSO_1n5n1^I[OY1^MPOb1Z6f1ZIi12jLe6^1WIj1HSMQ7Q:1O100O1N2K5L5K4K5L4L4UOk0C=D=C<C=D<E;J6J6eNZDVJm;l5SDkIS<X6mC_IY<d6hCRI^<Q773`GkHo3X7iKnHV4U7aKRI^4R7YKTIg4n6bIgJ]6\\5XIlJh6W5mHRKR7Q5cHVK^7m4WHZKj7]81N3N1O1O1[OhGaEY8]:jG`EX8]:lGaEU8]:nGaET8\\:oGbER8\\:QHbEP8\\:SHbEn7V8fGaI`0VNl7T8RH[Ik8_6eGUI]8f6QHoHR8k6]HhHf7S7hHbHZ7U5fF[LS<e3aDbKd;^4o100O101O000O1000000O2O0000000O10001O0O10001O0100O10000O10O0100O100N2O1O1N101N2O1O1N2O1N2O001N2O1N2O1O1N2O0O2O1O1N4M3L3N3M3L4M2M4M3N1N3M2N100O1O2O0O1O1O100O1O2N105J8H9HlLYDUN_;c1nD^Nl:[1_EbN`:^1dE_N\\:a1gE\\NX:f1jEWNU:j1nESNQ:n1SFnMm9S2UFjMj9W2YFfMf9[2[FcMf9^2ZF_Mg9b2ZF[Mg9h2WFVMj9m2UFPMm9S3QFiLQ:Z3nEcLU:^3iE`L]:^3bE_Ld:`3ZE]Ll:a3RE]LT;f0iCa0R1eN\\;1ZDY19cNc;\\OkDQ2_OaNR>_1mA^NX>a1gA\\N^>d1`AYNd>h1YAVNl>k1QAQNU?P2g@nM]?T2^@kMg?W2T@gMQ`0Z2k_OdM^`0V2`_OiMk`0m1V_OQNVa0d1k^OYNaa0^1_^O`Nia0Y1W^OfNna0V1S^OhNRb0T1o]OjNVb0S1>M4L4L4K5K5K4L5KTim5"}, "label": "the taller giraffe"}]}
{"id": 401846, "image": "COCO_train2014_000000401846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man is holding a surfboard\"."}], "task": "refering", "answer_template": "The \"the man is holding a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 265, 285, 308, 309, 330, 331, 332], "bbox": [0.36834375, 0.3997411764705882, 0.5291718750000001, 0.9830823529411763], "iscrowd": 0, "area": 9308.167849999994, "rle": {"size": [425, 640], "counts": "jWR3a0_<=C9G8H8I8eGRNR5V2lJmMl4a1YHROj2^Ol4T1lHZOU2En4c0_IDb1Kn4=iIDV12o4;PJ^Oo0:P58UJZOj0`0o47iLLV34iLMW33gL0W31gL2X3NgL4X3LfL7X3JgL8X3HfL;Y3EeL=Z3DeL=[3MYL3i3b0aK^O_4U1mJlNT5\\1bJdN_5e1VJ]Nk5k1^IcNb6f1fHhN\\7a30O2O001NO2I70O2O000O1I7I7H8H8I7J7MaJdKY2Y4`MQL`2k3[M^Le2_3UMjLk2R3oLXMQ3e2PM]MP3_2SMbMm2[2VMeMk2V2XMkMi2Q2ZMoMg2m1\\MSNe2i1]MXNd2d1_M\\Nb2`1aMaN_2[1cMfN^2V1eMjN\\2TO\\JV1[3FZ2mNeJX1Q3L[2fNlJ[1i2Ja33^LEm39QL_OY4`0gKWOc4i0\\KoNm4P1RKhNX5X1]300O10001O02N2O1N1O2O1N2O1003M2N2N3M2N2N3L3N2N3M2N3M3M3K5L4L4LScl3"}, "label": "the man is holding a surfboard"}]}
{"id": 401846, "image": "COCO_train2014_000000401846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"surfer near the shore\"."}], "task": "refering", "answer_template": "The \"surfer near the shore\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 265, 285, 308, 309, 330, 331, 332], "bbox": [0.36834375, 0.3997411764705882, 0.5291718750000001, 0.9830823529411763], "iscrowd": 0, "area": 9308.167849999994, "rle": {"size": [425, 640], "counts": "jWR3a0_<=C9G8H8I8eGRNR5V2lJmMl4a1YHROj2^Ol4T1lHZOU2En4c0_IDb1Kn4=iIDV12o4;PJ^Oo0:P58UJZOj0`0o47iLLV34iLMW33gL0W31gL2X3NgL4X3LfL7X3JgL8X3HfL;Y3EeL=Z3DeL=[3MYL3i3b0aK^O_4U1mJlNT5\\1bJdN_5e1VJ]Nk5k1^IcNb6f1fHhN\\7a30O2O001NO2I70O2O000O1I7I7H8H8I7J7MaJdKY2Y4`MQL`2k3[M^Le2_3UMjLk2R3oLXMQ3e2PM]MP3_2SMbMm2[2VMeMk2V2XMkMi2Q2ZMoMg2m1\\MSNe2i1]MXNd2d1_M\\Nb2`1aMaN_2[1cMfN^2V1eMjN\\2TO\\JV1[3FZ2mNeJX1Q3L[2fNlJ[1i2Ja33^LEm39QL_OY4`0gKWOc4i0\\KoNm4P1RKhNX5X1]300O10001O02N2O1N1O2O1N2O1003M2N2N3M2N2N3L3N2N3M2N3M3M3K5L4L4LScl3"}, "label": "surfer near the shore"}]}
{"id": 221625, "image": "COCO_train2014_000000221625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player on the left side\"."}], "task": "refering", "answer_template": "The \"the player on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 164, 165, 166, 167, 186, 187, 189, 190, 207, 208, 209, 212, 213, 230, 231, 235, 236], "bbox": [0.02803125, 0.06503125, 0.33071875, 0.9551562499999999], "iscrowd": 0, "area": 22481.041, "rle": {"size": [320, 640], "counts": "Rm52l94M3L4L4M3L2O2M201O0O2O0O2N0001O0000001O0001O01O0000001O00000010O0000000O100PL6kMJT2k0XMUOg2S1RMmNn2U1PMkNo2W1PMiNo2Z1oLgNo2[1PMeNo2]1PMcNo2`1oL`NP3b1oL^NP3e1mL\\NR3f1mLZNR3i1lLWNS3k1lLUNS3m1lLSNS3P2kLQNT3P2kLPNT3S2jLmMU3U2jLlMT3W2jLiMU3Z2iLfMV3]2gLeMW3_2fLaMY3b2eL_MY3d2eL\\MZ3h2cLXM]3j2aLWM^3k2`LUMa3l2]LTMc3o2ZLRMe3P3YLPMh3Q3VLoLj3T3SLmLm3T3PLmLP4U3nKkLS4W3jKiLV4`4001O0O2O00001N101O001N1000_OcJaL\\5^3kJ\\LU5d3PKXLn4h3e0O1O10O01OhN[LXLe3j3aLnK_3S4hLeKY3[4nL\\KS3f4TMQKm2o4Q12O0O101N101N10001N101N10001O0010O0^M`KD`4;bKEk2eM^Nd2jNFf2iM_N_2oNG_2mMbNY2ROIZ2RNbNS2WOJU2VNcNn1\\OKn1ZNeNj1_OKi1^NhNe1BLd1aNkNd1@Kb1cNoNd1^OGa1gNQOd1]OE_1iNUOc1[OC^1lNXOc1YO_O:3=`0XO]OBk0W1IVO`2k0aMUO]2l0eMROZ2n0dKkNQ16Z3P1cKnNR12Y3S1bKmNV1NW3W1aKnNW1JW3[1_KmNZ1HU3]1^KoN\\1CU3`1]KoN]1AU3b1\\KoN]1@V3c1[KPO]1^OV3d1ZKQO^1\\OW3m2fLSMY3P3eLQMY3S3cLnL\\3V3aLjL^3Y3_LiL_3[3^LeLa3_3[LbLd3b3YL^Lg3a3YL`LP2HaNi3_O_Lc17kNZ3B`LV1e0UOl2E_Li0U1^O\\2I`L<e1Fl1M`L1T2M\\12aLOY2JW15bL0Y2IU17cLNZ2JT17cLM[2KR18dLK\\2LQ18dLK\\2LP19dLJ^2Lo09dLI_2Mm09hLF]20k0:kLC[21l0;`0BB`0o43L3N3L3M4L0101N3M3M2O2M3M3N2M3M3N2M3NmlU4"}, "label": "the player on the left side"}]}
{"id": 221625, "image": "COCO_train2014_000000221625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player on the right\"."}], "task": "refering", "answer_template": "The \"the player on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 38, 39, 40, 60, 61, 62, 63, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 223, 224, 225, 242, 246, 247, 248], "bbox": [0.52584375, 0.06965625, 0.8247187500000001, 0.9618125000000001], "iscrowd": 0, "area": 22144.35455, "rle": {"size": [320, 640], "counts": "lbY3a0^92O0O2N1O2O0O2N1O2O0O2N1O10O000O101N1000010O00001O0001O01O00O10001O0^LSORNm0l1\\OmMd0Q2EhM:V20bM0]29\\MFc2c0VM[Oj2n0oLQOP3X1iLfNW3^1gLcNU3c1hL]NV3g1gL[NV3i1hLWNV3m1hLTNU3P2iLPNU3T2iLmMV3U2hLlMW3V2gLjMX3Y2eLiMZ3Y2dLgM\\3[2bLfM\\3]2bLcM^3_2`LbM_3`2_L`Ma3c2\\L^Mb3e2\\L[Md3g2YL[Mf3g2XLZMf3i2XLWMh3k2VLVMi3l2ULTMk3n2SLRMl3Q3RLPMm3R3PLoLP4S3nKmLR4U3lKkLS4X3kKhLU4a400001O001O1O001O00100O001O001OD_J[L`5d3dJYL\\5f3iJWLV5h3nJVLP5k3b0O_NULPMj3R3[LgLe3[3`L_L`3c3cLWL]3k3hLoKX3R4lLiKT3Y4PM`KR3`4SMZKm2g4U12N101N1O1O2O0O1O2O0O1O2O01O01ObM]K_OR3gMZNi2fN_On2kM\\Nc2iNBi2mM]N_2mNCd2PN_N[2QOD]2TNaNV2UOEY2WNbNR2WOFU2[NcNm1\\OGn1^NfNi1_OHj1aNfNf1BHf1dNiNe1@Gd1fNlNf1_OBd1iNmNh1]O^Od1lNoNh1\\O\\Ob1nNSOh1YOYO6=a0=XOTOJl0n02WORO\\OZ1^1lMfNc0>]3P1dKmNo01\\3U1bKlNR1OZ3X1aKkNV1KX3]1^KlNY1FX3`1]KlN\\1BV3b1]KoN^1^OS3d1^KPO^1\\OS3e1]KSO]1YOU3e1]KTO\\1YOU3d1^KVOZ1WOW3d1]KYOY1TOY3d1]KZOX1SOZ3U3cLmL[3V3bLkL]3X3`LiL_3Y3_LhL`3[3]LgLa3\\3\\LeLc3]3\\LcLc3]3]LdLb3]3]LeLj1\\OnMK`0U4GdLZ1N\\N\\Ob0R4HeLm0:gNPOd0R4GeLa0g0POeNg0o3HfL5R1YO[Nj0n3GeL5b13j1GeL4c14h1HfL3a16j1FfL2a1G_M5[41fL1`1IbM3X43fL0a1HdM4V44eLO`1HiM3R46fLM_1JkM2Q46fLL_1JmM3o36eLL_1JQN2k39fLH_1LRN2j3<dLD_1MWN1g3`0bL_O_20o0e0h42N2N3M2NO1O100O001O102M4L3N3L3L4L5JQkR1"}, "label": "the player on the right"}]}
{"id": 221625, "image": "COCO_train2014_000000221625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman hitting a tennis ball to the right of an identical woman\"."}], "task": "refering", "answer_template": "The \"the woman hitting a tennis ball to the right of an identical woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 38, 39, 40, 60, 61, 62, 63, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 223, 224, 225, 242, 246, 247, 248], "bbox": [0.52584375, 0.06965625, 0.8247187500000001, 0.9618125000000001], "iscrowd": 0, "area": 22144.35455, "rle": {"size": [320, 640], "counts": "lbY3a0^92O0O2N1O2O0O2N1O2O0O2N1O10O000O101N1000010O00001O0001O01O00O10001O0^LSORNm0l1\\OmMd0Q2EhM:V20bM0]29\\MFc2c0VM[Oj2n0oLQOP3X1iLfNW3^1gLcNU3c1hL]NV3g1gL[NV3i1hLWNV3m1hLTNU3P2iLPNU3T2iLmMV3U2hLlMW3V2gLjMX3Y2eLiMZ3Y2dLgM\\3[2bLfM\\3]2bLcM^3_2`LbM_3`2_L`Ma3c2\\L^Mb3e2\\L[Md3g2YL[Mf3g2XLZMf3i2XLWMh3k2VLVMi3l2ULTMk3n2SLRMl3Q3RLPMm3R3PLoLP4S3nKmLR4U3lKkLS4X3kKhLU4a400001O001O1O001O00100O001O001OD_J[L`5d3dJYL\\5f3iJWLV5h3nJVLP5k3b0O_NULPMj3R3[LgLe3[3`L_L`3c3cLWL]3k3hLoKX3R4lLiKT3Y4PM`KR3`4SMZKm2g4U12N101N1O1O2O0O1O2O0O1O2O01O01ObM]K_OR3gMZNi2fN_On2kM\\Nc2iNBi2mM]N_2mNCd2PN_N[2QOD]2TNaNV2UOEY2WNbNR2WOFU2[NcNm1\\OGn1^NfNi1_OHj1aNfNf1BHf1dNiNe1@Gd1fNlNf1_OBd1iNmNh1]O^Od1lNoNh1\\O\\Ob1nNSOh1YOYO6=a0=XOTOJl0n02WORO\\OZ1^1lMfNc0>]3P1dKmNo01\\3U1bKlNR1OZ3X1aKkNV1KX3]1^KlNY1FX3`1]KlN\\1BV3b1]KoN^1^OS3d1^KPO^1\\OS3e1]KSO]1YOU3e1]KTO\\1YOU3d1^KVOZ1WOW3d1]KYOY1TOY3d1]KZOX1SOZ3U3cLmL[3V3bLkL]3X3`LiL_3Y3_LhL`3[3]LgLa3\\3\\LeLc3]3\\LcLc3]3]LdLb3]3]LeLj1\\OnMK`0U4GdLZ1N\\N\\Ob0R4HeLm0:gNPOd0R4GeLa0g0POeNg0o3HfL5R1YO[Nj0n3GeL5b13j1GeL4c14h1HfL3a16j1FfL2a1G_M5[41fL1`1IbM3X43fL0a1HdM4V44eLO`1HiM3R46fLM_1JkM2Q46fLL_1JmM3o36eLL_1JQN2k39fLH_1LRN2j3<dLD_1MWN1g3`0bL_O_20o0e0h42N2N3M2NO1O100O001O102M4L3N3L3L4L5JQkR1"}, "label": "the woman hitting a tennis ball to the right of an identical woman"}]}
{"id": 25017, "image": "COCO_train2014_000000025017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue boat\"."}], "task": "refering", "answer_template": "The \"a blue boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225], "bbox": [0.27942187500000004, 0.3237236533957845, 0.925359375, 0.6573067915690866], "iscrowd": 0, "area": 35570.830849999984, "rle": {"size": [427, 640], "counts": "bjZ2`0j<10O0100O1O010O10O0100O1O01O00001O001O0O101O0013L6K4K6J5J3N1N3O1N2O1O1O001O1O1N2O1O001O1O1O1N2O001O1O1O1O1O0O2O1O1O1O1O001N2O1O1O1O1O1O1O1N2O1O1O1O1O1001O2N1O1O0N1O0O2O1N2N2O1N2O1N21O1O1O1O1O1O1O001O1O1O1O1O1O0O10000O100O100O100O1O100O2O0O1N2N2N2N2O1N1O2N2N2001O00000000001O000000001O000000001O01O000010O000001O01O01O00002N100O1O1O1O100O1O1O2N0O001N10000O01000O2O001N2O001N110001N100O10000O101N10000O100000001O0000000000001O0000000000001O00000000000010O00000000001O01O00000001O0001O0001O00001O01O01O00001O1O010O1O17I7H3NN2N2N2M2O2N2N2M3O010OO2O001O001O001O001O1O001O0@[FYMg9d2\\F[Me9b2^F]Mb9b2aF\\M_9b2dF]M\\9c2eF\\M[9d2gFZMY9f2hFYMX9g2iFXMW9h2f001O0001O00000000000000007JO00000000000000000000001O00000000000000000000001O00000000000000000000001O0000000O1N2O1O1O1N2O1O1O1N3N1O1O1O1N2O1O1O1N2O1O1O1N3N1O1O1N2O1O2N1N2O1O1O1N3N1O1O1N2O1O2N1N2O100O101O0O100O101O0O100O10001N100O100O102N;D`jc0"}, "label": "a blue boat"}]}
{"id": 25017, "image": "COCO_train2014_000000025017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue boat with sheep underneath it\"."}], "task": "refering", "answer_template": "The \"a blue boat with sheep underneath it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225], "bbox": [0.27942187500000004, 0.3237236533957845, 0.925359375, 0.6573067915690866], "iscrowd": 0, "area": 35570.830849999984, "rle": {"size": [427, 640], "counts": "bjZ2`0j<10O0100O1O010O10O0100O1O01O00001O001O0O101O0013L6K4K6J5J3N1N3O1N2O1O1O001O1O1N2O1O001O1O1O1N2O001O1O1O1O1O0O2O1O1O1O1O001N2O1O1O1O1O1O1O1N2O1O1O1O1O1001O2N1O1O0N1O0O2O1N2N2O1N2O1N21O1O1O1O1O1O1O001O1O1O1O1O1O0O10000O100O100O100O1O100O2O0O1N2N2N2N2O1N1O2N2N2001O00000000001O000000001O000000001O01O000010O000001O01O01O00002N100O1O1O1O100O1O1O2N0O001N10000O01000O2O001N2O001N110001N100O10000O101N10000O100000001O0000000000001O0000000000001O00000000000010O00000000001O01O00000001O0001O0001O00001O01O01O00001O1O010O1O17I7H3NN2N2N2M2O2N2N2M3O010OO2O001O001O001O001O1O001O0@[FYMg9d2\\F[Me9b2^F]Mb9b2aF\\M_9b2dF]M\\9c2eF\\M[9d2gFZMY9f2hFYMX9g2iFXMW9h2f001O0001O00000000000000007JO00000000000000000000001O00000000000000000000001O00000000000000000000001O0000000O1N2O1O1O1N2O1O1O1N3N1O1O1O1N2O1O1O1N2O1O1O1N3N1O1O1N2O1O2N1N2O1O1O1N3N1O1O1N2O1O2N1N2O100O101O0O100O101O0O100O10001N100O100O102N;D`jc0"}, "label": "a blue boat with sheep underneath it"}]}
{"id": 426427, "image": "COCO_train2014_000000426427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb standing broadside in front of a taller lamb\"."}], "task": "refering", "answer_template": "The \"a lamb standing broadside in front of a taller lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 264, 265], "bbox": [0.311453125, 0.40072599531615927, 0.6771874999999999, 0.8054566744730679], "iscrowd": 0, "area": 22213.18925, "rle": {"size": [427, 640], "counts": "UWc22g12b9>kE>[9o1I7I6K6J6L1O00000000N2N2M3N2M3YNlFUOW9f0j1O01O0010O01O3N1O1O1O2N1O100O2N1O1O1O2N1O3N2M3M4L`0@5K6J5K6J2O0O1O1O1O001O1O100O1O1O1O1O1O001O10O0L4L4M3L4L4L4N2O1N3M2N2O1O100O100O100O100O100000000O100000000000000000000000000000000000000000000003M4L4L4L3M4L4L4L4L010O0000000000000000000000000000001O001O001O0O2O00L4L4LRO[L^Ha3f7cLVHW3P8kLmGo2Z8RMeGk2^8VM`Gj2a8VM^Gj2c8WM[Gi2f8XMXGh2i8YMUGg2k8[MSGd2o8]MnFd2S9^MiFc2X9^MfFc2Z9^McFc2^9e000000O10O100O10O10O1000O010O010O10OO2L4M3M4L3L4N2M3N2M3N2N3L3N2M4M2N2N3M2N2N2N2M3N2M3L4M3M3L4M3N2M3M3J6L4O1N2O1O1O1O1O1O1O1O2N1O3M4M`Tf2"}, "label": "a lamb standing broadside in front of a taller lamb"}]}
{"id": 426427, "image": "COCO_train2014_000000426427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb that is positioned in front of another lamb\"."}], "task": "refering", "answer_template": "The \"a lamb that is positioned in front of another lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 264, 265], "bbox": [0.311453125, 0.40072599531615927, 0.6771874999999999, 0.8054566744730679], "iscrowd": 0, "area": 22213.18925, "rle": {"size": [427, 640], "counts": "UWc22g12b9>kE>[9o1I7I6K6J6L1O00000000N2N2M3N2M3YNlFUOW9f0j1O01O0010O01O3N1O1O1O2N1O100O2N1O1O1O2N1O3N2M3M4L`0@5K6J5K6J2O0O1O1O1O001O1O100O1O1O1O1O1O001O10O0L4L4M3L4L4L4N2O1N3M2N2O1O100O100O100O100O100000000O100000000000000000000000000000000000000000000003M4L4L4L3M4L4L4L4L010O0000000000000000000000000000001O001O001O0O2O00L4L4LRO[L^Ha3f7cLVHW3P8kLmGo2Z8RMeGk2^8VM`Gj2a8VM^Gj2c8WM[Gi2f8XMXGh2i8YMUGg2k8[MSGd2o8]MnFd2S9^MiFc2X9^MfFc2Z9^McFc2^9e000000O10O100O10O10O1000O010O010O10OO2L4M3M4L3L4N2M3N2M3N2N3L3N2M4M2N2N3M2N2N2N2M3N2M3L4M3M3L4M3N2M3M3J6L4O1N2O1O1O1O1O1O1O1O2N1O3M4M`Tf2"}, "label": "a lamb that is positioned in front of another lamb"}]}
{"id": 74176, "image": "COCO_train2014_000000074176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in pink shirt and bikini surfboarding\"."}], "task": "refering", "answer_template": "The \"girl in pink shirt and bikini surfboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 197, 198, 213, 214, 215, 216, 217, 218, 221, 222, 236, 237, 238, 239, 240, 241, 245, 246, 261, 262, 263, 284, 285, 306, 307, 308, 309], "bbox": [0.2690625, 0.1975625, 0.726453125, 0.8231249999999999], "iscrowd": 0, "area": 33374.217150000004, "rle": {"size": [480, 640], "counts": "Wj`22R3Of85WGNe86YGLc88ZGKb89[GJa8:]GH_8<^GG]8>`GE]8>aGD\\8?aGD[8`0cGBZ8`0dGCY8`0dGBY8b0eG@X8c0eG_OY8d0dG_OZ8c0dG^O[8d0bG^O]8d0aG]O^8e0_G\\Oa8f0\\G[Od8g0ZGZOe8h0XGYOh8i0UGYOj8h0UGXOk8i0SGXOm8j0QGVOo8k0oFTOR9o0kFQOV9P1iFnNY9T1dFkN^9V1`FiNb9Y1\\FeNf9]1WFbNk9`1SF_Nn9c1oE\\NS:f1jEYNX:i1fEUN\\:m1aERNa:P2]EnMe:U32M3M20WIPLd2o3ZMTLe2k3[MWLe2h3YMZLh2e3VM^Li2b3UMaLk2]3TMfLk2Z3TMhLl2W3RMkLo2S3PMPMo2P3oLSMQ3m2gL[MX3g2YLfMh3Z2nJPOR5Q1jJROU5P1hJROX5o0dJTO[5m0cJTO^5m0^JVOb5k0[JWOe5j0WJYOi5g0UJZOl5g0PJ\\OP6e0mI]OS6d0iI_OW6a0fIBZ6?bIC_6_400000O100000000O1000000O10000O10000O10000O100O10000O100O10000O10000O100QObI\\I^6`6gI_IY6]6lIaIU6\\6oIcIQ6Y6SJgIm5V6WJiIi5T6[JjIf5T6]JkIc5T6_JkIa5T6f1O1aLWGe0k8[OWGb0j8]OXGa0i8_OYG>h8BYG;i8EYG8h8GZG7g8I[G4f8L[G1g8O[GNf82[GLf83]GJd86^GFd8:^GCc8=`G_Oa8a0aG[Oa8d0bGYO^8h0eGTO\\8l0fGQOZ8P1iGkNX8U1kGhNU8Y1nGcNP8_1SH^Nl7c1XHXNg7k1[HQNe7P2_HhMd7Z2aH\\Mb7e2[22N1O110O0010O01O01O01O0100O10000O100000000000000000000010O010000000O1000001O0O101OcNUDDk;4\\DNc;KdD5Z<fNmCZ1l<2O001N10001O001O00010O001O01O6K2M01O01O010O010O00010O010O0010O0010O0010O010O0001O010O001O010O000010O010O010O00100O2O0O100O1O01N101O001N101O001O0O2N1O2N2N1O2N1O1O001O00001O001O001O1N101O1O1O1O1N101O1N2N2M3Mhga2"}, "label": "girl in pink shirt and bikini surfboarding"}]}
{"id": 74176, "image": "COCO_train2014_000000074176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in red shirt surfing\"."}], "task": "refering", "answer_template": "The \"a woman in red shirt surfing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 197, 198, 213, 214, 215, 216, 217, 218, 221, 222, 236, 237, 238, 239, 240, 241, 245, 246, 261, 262, 263, 284, 285, 306, 307, 308, 309], "bbox": [0.2690625, 0.1975625, 0.726453125, 0.8231249999999999], "iscrowd": 0, "area": 33374.217150000004, "rle": {"size": [480, 640], "counts": "Wj`22R3Of85WGNe86YGLc88ZGKb89[GJa8:]GH_8<^GG]8>`GE]8>aGD\\8?aGD[8`0cGBZ8`0dGCY8`0dGBY8b0eG@X8c0eG_OY8d0dG_OZ8c0dG^O[8d0bG^O]8d0aG]O^8e0_G\\Oa8f0\\G[Od8g0ZGZOe8h0XGYOh8i0UGYOj8h0UGXOk8i0SGXOm8j0QGVOo8k0oFTOR9o0kFQOV9P1iFnNY9T1dFkN^9V1`FiNb9Y1\\FeNf9]1WFbNk9`1SF_Nn9c1oE\\NS:f1jEYNX:i1fEUN\\:m1aERNa:P2]EnMe:U32M3M20WIPLd2o3ZMTLe2k3[MWLe2h3YMZLh2e3VM^Li2b3UMaLk2]3TMfLk2Z3TMhLl2W3RMkLo2S3PMPMo2P3oLSMQ3m2gL[MX3g2YLfMh3Z2nJPOR5Q1jJROU5P1hJROX5o0dJTO[5m0cJTO^5m0^JVOb5k0[JWOe5j0WJYOi5g0UJZOl5g0PJ\\OP6e0mI]OS6d0iI_OW6a0fIBZ6?bIC_6_400000O100000000O1000000O10000O10000O10000O100O10000O100O10000O10000O100QObI\\I^6`6gI_IY6]6lIaIU6\\6oIcIQ6Y6SJgIm5V6WJiIi5T6[JjIf5T6]JkIc5T6_JkIa5T6f1O1aLWGe0k8[OWGb0j8]OXGa0i8_OYG>h8BYG;i8EYG8h8GZG7g8I[G4f8L[G1g8O[GNf82[GLf83]GJd86^GFd8:^GCc8=`G_Oa8a0aG[Oa8d0bGYO^8h0eGTO\\8l0fGQOZ8P1iGkNX8U1kGhNU8Y1nGcNP8_1SH^Nl7c1XHXNg7k1[HQNe7P2_HhMd7Z2aH\\Mb7e2[22N1O110O0010O01O01O01O0100O10000O100000000000000000000010O010000000O1000001O0O101OcNUDDk;4\\DNc;KdD5Z<fNmCZ1l<2O001N10001O001O00010O001O01O6K2M01O01O010O010O00010O010O0010O0010O0010O010O0001O010O001O010O000010O010O010O00100O2O0O100O1O01N101O001N101O001O0O2N1O2N2N1O2N1O1O001O00001O001O001O1N101O1O1O1O1N101O1N2N2M3Mhga2"}, "label": "a woman in red shirt surfing"}]}
{"id": 74179, "image": "COCO_train2014_000000074179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink van\"."}], "task": "refering", "answer_template": "The \"a pink van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 111, 112, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.19496, 0.3549866666666667, 0.5642999999999999, 0.61688], "iscrowd": 0, "area": 13559.131549999998, "rle": {"size": [375, 500], "counts": "jPT12c;2M3N3mD;R:e00O10VFWOf8i0YG^Oa8a0`G_O`8a0_G@a8?_GBa8>_GBa8>^GCb8<_GDa8<^GEb8:_GEb8;]GFd89[GHe87\\GId87[GHg87ZGTO[9l0h000OfFTOk7l0[GNe82ZGOf82YGOf81YG0g80YG0g81WG0i80VG1j8OVG1j8OUG2k8OPG5P9KlF9T9U10006J6J1O00001O0000001O000000000000O10000000000O101O000O1N2M3N2M3O100O1000000000000O11O2N2N1O0000000000000000000000000000001O0000000000000001O000000000000000000000000000000000000000000000001O000000000001O000010O2O1N01O000010O01O001O2N3N2L4M2M4L4M3L4M3L4L3N2M2M2L5K5O1O001O1N2O001O1O1O0O2O1O1O0K6Gnd_2"}, "label": "a pink van"}]}
{"id": 74179, "image": "COCO_train2014_000000074179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pink truck on side of road\"."}], "task": "refering", "answer_template": "The \"pink truck on side of road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 111, 112, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.19496, 0.3549866666666667, 0.5642999999999999, 0.61688], "iscrowd": 0, "area": 13559.131549999998, "rle": {"size": [375, 500], "counts": "jPT12c;2M3N3mD;R:e00O10VFWOf8i0YG^Oa8a0`G_O`8a0_G@a8?_GBa8>_GBa8>^GCb8<_GDa8<^GEb8:_GEb8;]GFd89[GHe87\\GId87[GHg87ZGTO[9l0h000OfFTOk7l0[GNe82ZGOf82YGOf81YG0g80YG0g81WG0i80VG1j8OVG1j8OUG2k8OPG5P9KlF9T9U10006J6J1O00001O0000001O000000000000O10000000000O101O000O1N2M3N2M3O100O1000000000000O11O2N2N1O0000000000000000000000000000001O0000000000000001O000000000000000000000000000000000000000000000001O000000000001O000010O2O1N01O000010O01O001O2N3N2L4M2M4L4M3L4M3L4L3N2M2M2L5K5O1O001O1N2O001O1O1O0O2O1O1O0K6Gnd_2"}, "label": "pink truck on side of road"}]}
{"id": 74179, "image": "COCO_train2014_000000074179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open door on the yellow car\"."}], "task": "refering", "answer_template": "The \"an open door on the yellow car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 136, 137, 138, 139, 140, 141, 155, 156, 157, 158, 159, 175, 176, 177], "bbox": [0.6052799999999999, 0.47461333333333333, 0.8726799999999999, 0.7231733333333333], "iscrowd": 0, "area": 6769.211250000002, "rle": {"size": [375, 500], "counts": "]U_3=Y;2O0O2N2O1N101N2N1XEAX:P1N2O1N2N101N2N010O10O01O10O0010O0O1O2O1O100O10000N2L4L4O1000001O01O00000000001O000000001SN]Fd1c9YN`Fg1h901O000001O01O0000001O0000^FXNU9V2O101N1O1O100O1O000001O00000O2O2N1M3M3M3M3M3M3N2N2O1O1O1O1O1O1O2N1O1O1O1O1O0102M2O000O010O01O001O1N1O2O0O2O1D;01O0O2O001O1N101OeVg0"}, "label": "an open door on the yellow car"}]}
{"id": 74179, "image": "COCO_train2014_000000074179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow colour truck\"."}], "task": "refering", "answer_template": "The \"yellow colour truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 136, 137, 138, 139, 140, 141, 155, 156, 157, 158, 159, 175, 176, 177], "bbox": [0.6052799999999999, 0.47461333333333333, 0.8726799999999999, 0.7231733333333333], "iscrowd": 0, "area": 6769.211250000002, "rle": {"size": [375, 500], "counts": "]U_3=Y;2O0O2N2O1N101N2N1XEAX:P1N2O1N2N101N2N010O10O01O10O0010O0O1O2O1O100O10000N2L4L4O1000001O01O00000000001O000000001SN]Fd1c9YN`Fg1h901O000001O01O0000001O0000^FXNU9V2O101N1O1O100O1O000001O00000O2O2N1M3M3M3M3M3M3N2N2O1O1O1O1O1O1O2N1O1O1O1O1O0102M2O000O010O01O001O1N1O2O0O2O1D;01O0O2O001O1N101OeVg0"}, "label": "yellow colour truck"}]}
{"id": 8649, "image": "COCO_train2014_000000008649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"palm tree and pot that is beside the bench and directly behind the little boy\"."}], "task": "refering", "answer_template": "The \"palm tree and pot that is beside the bench and directly behind the little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 10, 11, 12, 13, 14, 15, 22, 23, 24, 27, 28, 29, 30, 31, 32, 36, 37, 38, 39, 40, 41, 44, 45, 46, 47, 48, 49, 50, 53, 54, 55, 56, 57, 58, 62, 63, 64, 65, 66, 67, 70, 71, 72, 73, 74, 75, 78, 79, 80, 81, 82, 83, 84, 88, 89, 90, 91, 92, 95, 96, 97, 98, 99, 100, 101, 112, 162, 163, 164, 179, 180, 181, 195, 197, 210, 211, 212, 213, 227, 228, 229, 230, 245, 246, 247], "bbox": [0.11985416666666666, 0.0, 1.0, 0.628484375], "iscrowd": 0, "area": 44270.30445000001, "rle": {"size": [640, 480], "counts": "nZT16ic07I7I7I7I5K2N1O1O2N101N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O1O1O2O0000000O101O00000000001O00000O1000001O00000000001PNaM]B_2c=bM[B_2e=cMYB]2g=dMXB\\2h=eMWB[2i=fMVBZ2j=hMTBY2l=gMSBY2m=hMRBX2n=iMQBW2o=kMoAU2Q>lMoAS2Q>nMoAQ2Q>PNnAQ2Q>QNnAn1R>SNnAl1R>UNmAk1S>VNmAi1S>YNlAf1T>[NkAe1U>\\NkAd1T>]NlAb1U>_NiAa1W>`NiA_1W>bNiA]1W>dNiA[1W>gNgAY1Y>hNgAX1X>iNhAV1X>kNgAU1Y>mNfAR1Z>oNfAP1Z>QOeAo0[>ROeAm0[>UOdAk0[>VOdAj0\\>WOdAh0]>XOcAg0]>[OaAe0_>\\OaAc0_>^OaAa0_>A_A?a>B_A=a>D_A;a>F^A:b>H]A7c>J\\A6d>KZA6f>KYA5g>MVA4j>MTA4l>MSA3n>MPA6n>LPA6n>KPA7o>JPA8n>IPA:n>HPA9o>Ho@9Q?Hn@7S?Jk@7U?Y2100O100O2O1O1N2O0OVGaKV2^4YMTLg2j3WMZLi2e3VM^Li2`3XMaLh2^3WMdLi2[3WMgLh2W3XMkLh2T3XMmLg2R3ZMoLf2P3ZMRMe2l2]MTMc2k2]MVMc2h2^MYMb2f2^M\\Ma2c2_M^Ma2`2`MaM_2_2aMbM_2\\2cMeM\\2Z2dMgM\\2`1_GfMV6k0[2W1iGkMl5P1Z2Q1QHkMe5U1Z2\\OgG9?ROa5Z1Y2VOUH26[O\\5^1Y2QOTJ]Od3d1V2nNd0S1\\OkNe0V1[OhNf0Y1ZOeNg0\\1YOWN`G9X9`1YOWNcG3T9g1l8000000O1000O100000000AWNf^Oi1Za0[Nc^Oe1]a0]Na^Oc1^a0aN^^O`1aa0`000O100001N10001O3M3M3M2N3L3N3M2N001O001O1O0O2O1O1O1O1O1O1O1N3N1O1O1O2N1O2N1O2M5Lnh1C[WN`0A7H5L3L5L3L5O002O0\\N[Nk@f1m>`NTAa1d>eN^A[1Z>kNgAV1Q>POQBP1h9ZN]Hj0mMn0c9[NWHm0XNh0_9]NQHQ1aNc0[9`NoGo0hNa0W9cNmGm0mNa0T9eNjGk0UO`0o8gNhGk0ZO?l8iNeGj0@?g8kNdGh0G=c8nNaGg0M<`8PO^Ge05;[8RO\\Ge0::X8TOYGd0a08S8WOXGb0g08o7YOUGa0n07j7ZOUG?S18f7[OSG?Y14c7@PG=_1Od7GhF<e1Hg7N`F;l1Bg75YF<P2ZOj7=RF;V2SOk7e0jE:\\2lNn7l0ZEc0i2\\No7n3RHmKR8S4PHhKS8Y4mGbKW8^4kG]KW8e4iGVK[8j4fGRK^8m4[300000000O10000000000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10001N1000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000O10000000000O100dLZA^1f>bNZA^1f>aN[A_1e>`N\\A`1d>`N\\A`1d>_N]Aa1c>^N^Aa1c>_N]Aa1c>^N^Ab1b>]N_Ac1a>]N_Ac1a>\\N`Ad1`>\\N`Ad1`>[NaAe1_>ZNbAf1g>QNYAo1i>nMXAR2j>kMVAV2l>hMTAX2n>eMSA[2Y`000000000000000000000000000000000000000O10000O1N2N2N2O1N2N2M3N2N2M3N2N2M3N2Nf0YO^M"}, "label": "palm tree and pot that is beside the bench and directly behind the little boy"}]}
{"id": 8649, "image": "COCO_train2014_000000008649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"palm tree and planter shown directly behind the boy\"."}], "task": "refering", "answer_template": "The \"palm tree and planter shown directly behind the boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 10, 11, 12, 13, 14, 15, 22, 23, 24, 27, 28, 29, 30, 31, 32, 36, 37, 38, 39, 40, 41, 44, 45, 46, 47, 48, 49, 50, 53, 54, 55, 56, 57, 58, 62, 63, 64, 65, 66, 67, 70, 71, 72, 73, 74, 75, 78, 79, 80, 81, 82, 83, 84, 88, 89, 90, 91, 92, 95, 96, 97, 98, 99, 100, 101, 112, 162, 163, 164, 179, 180, 181, 195, 197, 210, 211, 212, 213, 227, 228, 229, 230, 245, 246, 247], "bbox": [0.11985416666666666, 0.0, 1.0, 0.628484375], "iscrowd": 0, "area": 44270.30445000001, "rle": {"size": [640, 480], "counts": "nZT16ic07I7I7I7I5K2N1O1O2N101N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O1O1O2O0000000O101O00000000001O00000O1000001O00000000001PNaM]B_2c=bM[B_2e=cMYB]2g=dMXB\\2h=eMWB[2i=fMVBZ2j=hMTBY2l=gMSBY2m=hMRBX2n=iMQBW2o=kMoAU2Q>lMoAS2Q>nMoAQ2Q>PNnAQ2Q>QNnAn1R>SNnAl1R>UNmAk1S>VNmAi1S>YNlAf1T>[NkAe1U>\\NkAd1T>]NlAb1U>_NiAa1W>`NiA_1W>bNiA]1W>dNiA[1W>gNgAY1Y>hNgAX1X>iNhAV1X>kNgAU1Y>mNfAR1Z>oNfAP1Z>QOeAo0[>ROeAm0[>UOdAk0[>VOdAj0\\>WOdAh0]>XOcAg0]>[OaAe0_>\\OaAc0_>^OaAa0_>A_A?a>B_A=a>D_A;a>F^A:b>H]A7c>J\\A6d>KZA6f>KYA5g>MVA4j>MTA4l>MSA3n>MPA6n>LPA6n>KPA7o>JPA8n>IPA:n>HPA9o>Ho@9Q?Hn@7S?Jk@7U?Y2100O100O2O1O1N2O0OVGaKV2^4YMTLg2j3WMZLi2e3VM^Li2`3XMaLh2^3WMdLi2[3WMgLh2W3XMkLh2T3XMmLg2R3ZMoLf2P3ZMRMe2l2]MTMc2k2]MVMc2h2^MYMb2f2^M\\Ma2c2_M^Ma2`2`MaM_2_2aMbM_2\\2cMeM\\2Z2dMgM\\2`1_GfMV6k0[2W1iGkMl5P1Z2Q1QHkMe5U1Z2\\OgG9?ROa5Z1Y2VOUH26[O\\5^1Y2QOTJ]Od3d1V2nNd0S1\\OkNe0V1[OhNf0Y1ZOeNg0\\1YOWN`G9X9`1YOWNcG3T9g1l8000000O1000O100000000AWNf^Oi1Za0[Nc^Oe1]a0]Na^Oc1^a0aN^^O`1aa0`000O100001N10001O3M3M3M2N3L3N3M2N001O001O1O0O2O1O1O1O1O1O1O1N3N1O1O1O2N1O2N1O2M5Lnh1C[WN`0A7H5L3L5L3L5O002O0\\N[Nk@f1m>`NTAa1d>eN^A[1Z>kNgAV1Q>POQBP1h9ZN]Hj0mMn0c9[NWHm0XNh0_9]NQHQ1aNc0[9`NoGo0hNa0W9cNmGm0mNa0T9eNjGk0UO`0o8gNhGk0ZO?l8iNeGj0@?g8kNdGh0G=c8nNaGg0M<`8PO^Ge05;[8RO\\Ge0::X8TOYGd0a08S8WOXGb0g08o7YOUGa0n07j7ZOUG?S18f7[OSG?Y14c7@PG=_1Od7GhF<e1Hg7N`F;l1Bg75YF<P2ZOj7=RF;V2SOk7e0jE:\\2lNn7l0ZEc0i2\\No7n3RHmKR8S4PHhKS8Y4mGbKW8^4kG]KW8e4iGVK[8j4fGRK^8m4[300000000O10000000000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10001N1000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000O10000000000O100dLZA^1f>bNZA^1f>aN[A_1e>`N\\A`1d>`N\\A`1d>_N]Aa1c>^N^Aa1c>_N]Aa1c>^N^Ab1b>]N_Ac1a>]N_Ac1a>\\N`Ad1`>\\N`Ad1`>[NaAe1_>ZNbAf1g>QNYAo1i>nMXAR2j>kMVAV2l>hMTAX2n>eMSA[2Y`000000000000000000000000000000000000000O10000O1N2N2N2O1N2N2M3N2N2M3N2N2M3N2Nf0YO^M"}, "label": "palm tree and planter shown directly behind the boy"}]}
{"id": 197068, "image": "COCO_train2014_000000197068.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mobile on the table\"."}], "task": "refering", "answer_template": "The \"a mobile on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157], "bbox": [0.36462500000000003, 0.0, 0.8293125, 0.4112422360248447], "iscrowd": 0, "area": 39486.391250000015, "rle": {"size": [483, 640], "counts": "R_^35k>3M4L3L5L3M4L3L5L3M4K4M4L3M4K4N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3N1O2O0O2O0O2O001N101N101O0O2O0O2O0O2O001N101O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O0010O01O010O0010O010O01O010O0010O010O01O01O01O01O010O000010O00010O01O00010O0010O0001O01O01O010O000010O00010O01O00010O00010O0010O0001O01O01O010O000010O00010O01O00010O0010OO1N3N1O1O2N1O2N1O1O2N1O1O2N101O00010O00010O0010O00010O00010O01O01O01O01O01O0010O0001O010_N`E]N`:Y1jEgNV:o0UFPOl9f0^FYOb9>gFCX99mFFT96PGIP93UGLl8OXG2g8K]G4c8HaG8`8CeG<[8AhG?X8]OmGc0S8XOQHh0o7UOUHj0l7QOXHo0h7mN]HS1b7jNaHV1g:O00010O001O001O0N3M2N2N3M2N3MTXc1"}, "label": "a mobile on the table"}]}
{"id": 197068, "image": "COCO_train2014_000000197068.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the flip phone\"."}], "task": "refering", "answer_template": "The \"the flip phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157], "bbox": [0.36462500000000003, 0.0, 0.8293125, 0.4112422360248447], "iscrowd": 0, "area": 39486.391250000015, "rle": {"size": [483, 640], "counts": "R_^35k>3M4L3L5L3M4L3L5L3M4K4M4L3M4K4N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3N1O2O0O2O0O2O001N101N101O0O2O0O2O0O2O001N101O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O0010O01O010O0010O010O01O010O0010O010O01O01O01O01O010O000010O00010O01O00010O0010O0001O01O01O010O000010O00010O01O00010O00010O0010O0001O01O01O010O000010O00010O01O00010O0010OO1N3N1O1O2N1O2N1O1O2N1O1O2N101O00010O00010O0010O00010O00010O01O01O01O01O01O0010O0001O010_N`E]N`:Y1jEgNV:o0UFPOl9f0^FYOb9>gFCX99mFFT96PGIP93UGLl8OXG2g8K]G4c8HaG8`8CeG<[8AhG?X8]OmGc0S8XOQHh0o7UOUHj0l7QOXHo0h7mN]HS1b7jNaHV1g:O00010O001O001O0N3M2N2N3M2N3MTXc1"}, "label": "the flip phone"}]}
{"id": 8657, "image": "COCO_train2014_000000008657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a subway wearing a jacket and talking on the phone\"."}], "task": "refering", "answer_template": "The \"a man on a subway wearing a jacket and talking on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.61396875, 0.2741501103752759, 1.0, 0.9842604856512142], "iscrowd": 0, "area": 49928.28164999998, "rle": {"size": [453, 640], "counts": "eT^54Q>8H8I4K1O001O1O1O00100O001O001O0100O1O100O010O100OTN[OYFe0b9A]F>_9H`F8[9N\\23M3M3UNVOVFm0a9\\O\\Fg0XOmNo86\\F>Z1b0UOoNS9n0cG7TOPOV9m0cG6ROQOY9l0bGl0\\2oM[2_1oJc0_2YN\\2_1kJ:a2aN`2`1dJ1d2jNb2f4WLeKe3b4nKiKm3[4iKoKR4V4dKULW4o3_KYL_4g6O1O100O1O100O1O1O100O1O1O10010O000001O0000001O0000001O0000001O00000000001O01O000001O00000000001O000000001O1O001O1O1O1O010O1O1O001O2N2lHVKi4m4RKUKn4n4mJTKS5n4iJUKV5n4eJTK[5o4_JTKa5n4ZJUKf5n4TJUKl5n4nIVKQ6P5fISKZ6R5^IQKb6T5VIoJj6U610000O1000001N10000O1000000O10000001O001O001O001O001O001O00001O010O001O001O001O001O001O1O1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M3M4L3M3C>jNU1kNUF"}, "label": "a man on a subway wearing a jacket and talking on the phone"}]}
{"id": 8657, "image": "COCO_train2014_000000008657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a bus in a black puffy coat talking on his cell phone\"."}], "task": "refering", "answer_template": "The \"a man on a bus in a black puffy coat talking on his cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.61396875, 0.2741501103752759, 1.0, 0.9842604856512142], "iscrowd": 0, "area": 49928.28164999998, "rle": {"size": [453, 640], "counts": "eT^54Q>8H8I4K1O001O1O1O00100O001O001O0100O1O100O010O100OTN[OYFe0b9A]F>_9H`F8[9N\\23M3M3UNVOVFm0a9\\O\\Fg0XOmNo86\\F>Z1b0UOoNS9n0cG7TOPOV9m0cG6ROQOY9l0bGl0\\2oM[2_1oJc0_2YN\\2_1kJ:a2aN`2`1dJ1d2jNb2f4WLeKe3b4nKiKm3[4iKoKR4V4dKULW4o3_KYL_4g6O1O100O1O100O1O1O100O1O1O10010O000001O0000001O0000001O0000001O00000000001O01O000001O00000000001O000000001O1O001O1O1O1O010O1O1O001O2N2lHVKi4m4RKUKn4n4mJTKS5n4iJUKV5n4eJTK[5o4_JTKa5n4ZJUKf5n4TJUKl5n4nIVKQ6P5fISKZ6R5^IQKb6T5VIoJj6U610000O1000001N10000O1000000O10000001O001O001O001O001O001O00001O010O001O001O001O001O001O1O1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M3M4L3M3C>jNU1kNUF"}, "label": "a man on a bus in a black puffy coat talking on his cell phone"}]}
{"id": 8657, "image": "COCO_train2014_000000008657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black hat with a phone to her ear\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black hat with a phone to her ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 313, 314, 336, 337, 359, 360], "bbox": [0.11328125, 0.34487858719646797, 0.682171875, 1.0], "iscrowd": 0, "area": 32307.005350000007, "rle": {"size": [453, 640], "counts": "cdP13o=3N2M3M3N2M3M3N2M3M300O1O100O1O100O100O1O100O100O1O1000000000000000000000000000001O00000[OhN]DX1b;iN^DW1a;jN_DV1`;kN`DU1_;lNaDT1^;mNbDS1];nNcDR1\\;oNdDQ1[;POeDP1Y;ROgDn0X;SOhDm0W;TOiDl0V;UOjDk0U;VOkDj0T;WOlDi0S;XOmDh0R;YOnDg0Q;ZOoDf0P;[OPEe0o:\\OQEd0n:]OREc0m:^OSEb0l:_OTEa0j:AVE?j:AVE?i:BWE>i:BWE>h:CXE=h:CXE=h:CXE=g:DYE<g:DYE=e:D[E<e:D[E<e:D[E<d:E\\E;d:E\\E;c:F]E:a:H_E8`:I`E7_:JaE7]:JcE6[:LeE4Z:MfE3Y:NgE2W:0iE0V:1jEOU:2kENS:4mEMQ:4oELQ:4oELR:3nEMR:3nEMR:3nEMR:3nEMS:2mENS:2mEOR:1nEOR:1nEOS:0mE0S:0mE0S:0mE0S:0mE0S:0mE0T:OlE1T:OlE2l8iN`GU1D_OD@W95YGl0L_O]9L`Ff03^O\\9N`Fd04^O[90`Fb05^OZ92_Fa07\\OZ95^F?8\\OY96^F?9[OX98^F=:[OW9:^F;;ZOW9=\\F:=YOV9?\\F8>YOV9`0ZF8`0XOU9V2kFjMT9W2lFhMU9X2kFhMT9Y2lFgMS9Z2mFfMS9[2lFeMS9\\2mFdMR9]2nFcMR9]2nFcMQ9^2oFbMP9_2PGaMP9_2PGaMo8`2QG`Mn8a2RG_Mn8a2RG_Mm8b2SG^Mm8b2SG]Mn8c2RG]Mm8d2SG\\Mm8e2RG[Mn8e2RG[Mn8e2RG[Mn8e2RG[Mn8e2RG[Mm8f2SGZMj8i2VGWMh8k2XGUMe8n2[GRMc8P3]GPMa8R3_GnL^8U3bGkL\\8W3dGiLY8Z3gGfLW8\\3iGdLV8^3iGbLV8_3jGaLV8_3jGaLU8`3kG`LU8`3kG`LT8a3lG_LT8a3lG_LS8b3mG^LS8b3mG^LR8c3nG]LR8c3nG]LQ8d3oG\\LQ8d3oG\\LP8e3PH[LP8e3PH[Lo7f3QHZLn7g3RHYLn7g3RHYLm7h3SHXLl7i3THWLl7i3THWLk7j3UHVLj7k3VHULj7k3VHULi7l3WHTLh7m3XHSLh7m3XHSLh7m3XHSLh7m3XHSLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLj7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULm7j3SHVLm7j3SHVLn7i3RHWLo7h3QHXLo7h3QHXLP8g3PHYLQ8f3oGZLQ8f3oGZLR8e3nG[LS8d3mG\\LS8d3mG\\LT8c3lG]LU8b3kG_LT8a3lG_LU8`3kG`LU8`3kG`LV8_3jGaLW8^3iGbLW8^3iGbLX8]3hGdLX8[3hGeLX8[3hGeLY8Z3gGfLZ8Y3fGgLZ8Y3fGgL[8X3eGiL[8V3eGjL]8T3cGlL_8R3aGnLb8o2^GQMe8l2[GTMh8i2XGWMk8f2UG[Ml8c2TG]Mo8`2QG`MX9H`Fd18dNT:[1lEeNU:Z1kEfNV:Y1jEhNU:X1kElNR:S1nEROn9m0RFYOi9f0WF\\Oh9c0XF]Oi9c0VF]Ok9b0UF^Ol9a0TF_Ol9a0TF_On9?RFAQ:<oEDU:8kEHX:4iEMY:0gE0\\:MdE3^:KbE5`:I`E7b:G^E9c:\\OgEd0m;0000000001O0000001O0001O01O0000001O0000001O0000001O0000001O003M4KnNUOQEf0m:^OUE>i:GXE5f:O]ELb:6aEE^:>eE^OY:f0iEVOX:j0b101O01O0006K:E0010O001O00001O01O011N4O001O00N2M3Mj0VO3M3M2O2M3M3oCiNn:V24M2M4M3L4L4L3M4M3L3M3M3M4L3N2M3M4L7I8I7H8Hc[i2"}, "label": "a woman wearing a black hat with a phone to her ear"}]}
{"id": 8657, "image": "COCO_train2014_000000008657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black hat , sitting in front of a young man\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black hat , sitting in front of a young man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 313, 314, 336, 337, 359, 360], "bbox": [0.11328125, 0.34487858719646797, 0.682171875, 1.0], "iscrowd": 0, "area": 32307.005350000007, "rle": {"size": [453, 640], "counts": "cdP13o=3N2M3M3N2M3M3N2M3M300O1O100O1O100O100O1O100O100O1O1000000000000000000000000000001O00000[OhN]DX1b;iN^DW1a;jN_DV1`;kN`DU1_;lNaDT1^;mNbDS1];nNcDR1\\;oNdDQ1[;POeDP1Y;ROgDn0X;SOhDm0W;TOiDl0V;UOjDk0U;VOkDj0T;WOlDi0S;XOmDh0R;YOnDg0Q;ZOoDf0P;[OPEe0o:\\OQEd0n:]OREc0m:^OSEb0l:_OTEa0j:AVE?j:AVE?i:BWE>i:BWE>h:CXE=h:CXE=h:CXE=g:DYE<g:DYE=e:D[E<e:D[E<e:D[E<d:E\\E;d:E\\E;c:F]E:a:H_E8`:I`E7_:JaE7]:JcE6[:LeE4Z:MfE3Y:NgE2W:0iE0V:1jEOU:2kENS:4mEMQ:4oELQ:4oELR:3nEMR:3nEMR:3nEMR:3nEMS:2mENS:2mEOR:1nEOR:1nEOS:0mE0S:0mE0S:0mE0S:0mE0S:0mE0T:OlE1T:OlE2l8iN`GU1D_OD@W95YGl0L_O]9L`Ff03^O\\9N`Fd04^O[90`Fb05^OZ92_Fa07\\OZ95^F?8\\OY96^F?9[OX98^F=:[OW9:^F;;ZOW9=\\F:=YOV9?\\F8>YOV9`0ZF8`0XOU9V2kFjMT9W2lFhMU9X2kFhMT9Y2lFgMS9Z2mFfMS9[2lFeMS9\\2mFdMR9]2nFcMR9]2nFcMQ9^2oFbMP9_2PGaMP9_2PGaMo8`2QG`Mn8a2RG_Mn8a2RG_Mm8b2SG^Mm8b2SG]Mn8c2RG]Mm8d2SG\\Mm8e2RG[Mn8e2RG[Mn8e2RG[Mn8e2RG[Mn8e2RG[Mm8f2SGZMj8i2VGWMh8k2XGUMe8n2[GRMc8P3]GPMa8R3_GnL^8U3bGkL\\8W3dGiLY8Z3gGfLW8\\3iGdLV8^3iGbLV8_3jGaLV8_3jGaLU8`3kG`LU8`3kG`LT8a3lG_LT8a3lG_LS8b3mG^LS8b3mG^LR8c3nG]LR8c3nG]LQ8d3oG\\LQ8d3oG\\LP8e3PH[LP8e3PH[Lo7f3QHZLn7g3RHYLn7g3RHYLm7h3SHXLl7i3THWLl7i3THWLk7j3UHVLj7k3VHULj7k3VHULi7l3WHTLh7m3XHSLh7m3XHSLh7m3XHSLh7m3XHSLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLi7l3WHTLj7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLk7l3UHTLl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULl7k3THULm7j3SHVLm7j3SHVLn7i3RHWLo7h3QHXLo7h3QHXLP8g3PHYLQ8f3oGZLQ8f3oGZLR8e3nG[LS8d3mG\\LS8d3mG\\LT8c3lG]LU8b3kG_LT8a3lG_LU8`3kG`LU8`3kG`LV8_3jGaLW8^3iGbLW8^3iGbLX8]3hGdLX8[3hGeLX8[3hGeLY8Z3gGfLZ8Y3fGgLZ8Y3fGgL[8X3eGiL[8V3eGjL]8T3cGlL_8R3aGnLb8o2^GQMe8l2[GTMh8i2XGWMk8f2UG[Ml8c2TG]Mo8`2QG`MX9H`Fd18dNT:[1lEeNU:Z1kEfNV:Y1jEhNU:X1kElNR:S1nEROn9m0RFYOi9f0WF\\Oh9c0XF]Oi9c0VF]Ok9b0UF^Ol9a0TF_Ol9a0TF_On9?RFAQ:<oEDU:8kEHX:4iEMY:0gE0\\:MdE3^:KbE5`:I`E7b:G^E9c:\\OgEd0m;0000000001O0000001O0001O01O0000001O0000001O0000001O0000001O003M4KnNUOQEf0m:^OUE>i:GXE5f:O]ELb:6aEE^:>eE^OY:f0iEVOX:j0b101O01O0006K:E0010O001O00001O01O011N4O001O00N2M3Mj0VO3M3M2O2M3M3oCiNn:V24M2M4M3L4L4L3M4M3L3M3M3M4L3N2M3M4L7I8I7H8Hc[i2"}, "label": "a woman wearing a black hat , sitting in front of a young man"}]}
{"id": 221654, "image": "COCO_train2014_000000221654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with blue pants and yellow helmet\"."}], "task": "refering", "answer_template": "The \"a man with blue pants and yellow helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 164, 165, 166, 187, 188, 210, 211, 212, 233, 234, 235, 257, 258, 280, 281, 282, 303, 304], "bbox": [0.15440625, 0.45607981220657273, 0.293828125, 0.8885211267605634], "iscrowd": 0, "area": 7147.4232, "rle": {"size": [426, 640], "counts": "S^Y17n<9F:F;F9G9G9G9H5M3M2CbMlEa2R:<00O2O001N101N102N1N2O1O2WHaLb5a3ZJbLf5_3WJcLh5_3TJdLl5]3PJfLP6\\3lIfLS6\\3iIgLW6U510O11O1O1]K]H]3d7aLcHY3]7fLjHS3X7kLoHo2R7oLTIl2m6PMVIP3k6iL\\IV3d6dLbI\\3_6^LgI`3R801O0C\\FTMe9i2`FTM`9j2eFSM\\9j2iFSMX9j2e0oMeER1]:kNfER1\\:lNhET1W:iNlEX1R:fNRF[1l9cNWF^1f9bN[F_1d9`N^Fa1_9_NbFb1]9]NgFb1V9^NQG]1n8bNYGW1h8hN_GP1b8POmGa0T8^OmGa0S8]OPHb0Q8[OQHe0P8XOSHf0Z:O001O1O001N2O001O1O001N2O001M3N1O2M3NPfk5"}, "label": "a man with blue pants and yellow helmet"}]}
{"id": 221654, "image": "COCO_train2014_000000221654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier with a yellow helmet is the only skier on the snow covered ground\"."}], "task": "refering", "answer_template": "The \"a skier with a yellow helmet is the only skier on the snow covered ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 164, 165, 166, 187, 188, 210, 211, 212, 233, 234, 235, 257, 258, 280, 281, 282, 303, 304], "bbox": [0.15440625, 0.45607981220657273, 0.293828125, 0.8885211267605634], "iscrowd": 0, "area": 7147.4232, "rle": {"size": [426, 640], "counts": "S^Y17n<9F:F;F9G9G9G9H5M3M2CbMlEa2R:<00O2O001N101N102N1N2O1O2WHaLb5a3ZJbLf5_3WJcLh5_3TJdLl5]3PJfLP6\\3lIfLS6\\3iIgLW6U510O11O1O1]K]H]3d7aLcHY3]7fLjHS3X7kLoHo2R7oLTIl2m6PMVIP3k6iL\\IV3d6dLbI\\3_6^LgI`3R801O0C\\FTMe9i2`FTM`9j2eFSM\\9j2iFSMX9j2e0oMeER1]:kNfER1\\:lNhET1W:iNlEX1R:fNRF[1l9cNWF^1f9bN[F_1d9`N^Fa1_9_NbFb1]9]NgFb1V9^NQG]1n8bNYGW1h8hN_GP1b8POmGa0T8^OmGa0S8]OPHb0Q8[OQHe0P8XOSHf0Z:O001O1O001N2O001O1O001N2O001M3N1O2M3NPfk5"}, "label": "a skier with a yellow helmet is the only skier on the snow covered ground"}]}
{"id": 147930, "image": "COCO_train2014_000000147930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man standing behind the herd of cows\"."}], "task": "refering", "answer_template": "The \"the man standing behind the herd of cows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 307, 308, 309, 310, 330], "bbox": [0.326078125, 0.2711737089201878, 0.5474375, 0.9611267605633803], "iscrowd": 0, "area": 26894.31045, "rle": {"size": [426, 640], "counts": "hUg2`0X<b0L4M3M3N2M3M3N1N3TOeNgE]1Z:gN^E\\1b:f00000QGWN]6i1mH^Ob6b0TI9[6GZIV1T6iNbIS2m5j2LO0O10002M3N3M2M2O000O0100O01000O010O010O10O10O10O01000O0100O01000O00100jNnHSLR7m3RIoKo6P4TImKl6S4XIiKi6V4[IfKf6X4_IdKa6\\4gI\\KZ6c4mIVKT6i4V1O2O1O001O1O010000O10001N100O10000O101O0O10000O101O0O10000iIQKR4P5jKTKV4l4fKYKY4h4bK\\K]4f4`K]K_4d4^K_K`4d4\\K^Kd4c4ZK_Ke4b4XK`Kg4c4UK`Kj4a4SKaKm4`4PKcKn4^6O001N1AdHPK]7g4RIRKn6g4P1O1N2O1N2O1O1N21O001O11N4M2J6K6I6K6I6K6I7J7H7J9F=Di0VO5K4J6J6IPZh3"}, "label": "the man standing behind the herd of cows"}]}
{"id": 147930, "image": "COCO_train2014_000000147930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a turban is walking behind a herd of cows in the street\"."}], "task": "refering", "answer_template": "The \"a man wearing a turban is walking behind a herd of cows in the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 307, 308, 309, 310, 330], "bbox": [0.326078125, 0.2711737089201878, 0.5474375, 0.9611267605633803], "iscrowd": 0, "area": 26894.31045, "rle": {"size": [426, 640], "counts": "hUg2`0X<b0L4M3M3N2M3M3N1N3TOeNgE]1Z:gN^E\\1b:f00000QGWN]6i1mH^Ob6b0TI9[6GZIV1T6iNbIS2m5j2LO0O10002M3N3M2M2O000O0100O01000O010O010O10O10O10O01000O0100O01000O00100jNnHSLR7m3RIoKo6P4TImKl6S4XIiKi6V4[IfKf6X4_IdKa6\\4gI\\KZ6c4mIVKT6i4V1O2O1O001O1O010000O10001N100O10000O101O0O10000O101O0O10000iIQKR4P5jKTKV4l4fKYKY4h4bK\\K]4f4`K]K_4d4^K_K`4d4\\K^Kd4c4ZK_Ke4b4XK`Kg4c4UK`Kj4a4SKaKm4`4PKcKn4^6O001N1AdHPK]7g4RIRKn6g4P1O1N2O1N2O1O1N21O001O11N4M2J6K6I6K6I6K6I7J7H7J9F=Di0VO5K4J6J6IPZh3"}, "label": "a man wearing a turban is walking behind a herd of cows in the street"}]}
{"id": 336350, "image": "COCO_train2014_000000336350.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and red umbrella\"."}], "task": "refering", "answer_template": "The \"a white and red umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151], "bbox": [0.379203125, 0.16933054393305438, 0.65328125, 0.39949790794979073], "iscrowd": 0, "area": 11853.734600000003, "rle": {"size": [478, 640], "counts": "Zba32i>3N2M3N2M3N2M3N2[O]OlBf0S=]OiBd0W=b0O1O1O1O1N2O1bC_N^;b1`DaN^;`1`DbN_;_1_DdN_;^1^DdNa;]1]DfNb;Z1\\DhNc;Y1XDmNg;T1oCUOQ<m1O001O001O001O001O0O2O001O001O001O001O001OO0O2O1O1O1O1O1N2O1O1O1O001O1O100O10000O10O1000O1000000O100O01000O1L4F:C=0000O10O0100000000001O000O10001O0hCXNZ;h1cD\\N\\;d1aD`N_;b1ZDcNe;`1UDdNk;^1oCfNP<T21000O100O100O01O001O010O001N1O2O0O2O0O2N101N101N101N2O0O2O0O2O0O2O0O2YOfChNZ<T1nCgNT<U1l0M4N100O2O0O2N10001O001O00001O00nBWOP<h0oCYOQ<g0mC[OS<i0hCXOY<k0aCWO_<m0[CUOf<n0UCSOk<Q1oBQOR=]10O10S1mN4L3M4L3L5Ln^W3"}, "label": "a white and red umbrella"}]}
{"id": 336350, "image": "COCO_train2014_000000336350.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella with red and white stripes\"."}], "task": "refering", "answer_template": "The \"an umbrella with red and white stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151], "bbox": [0.379203125, 0.16933054393305438, 0.65328125, 0.39949790794979073], "iscrowd": 0, "area": 11853.734600000003, "rle": {"size": [478, 640], "counts": "Zba32i>3N2M3N2M3N2M3N2[O]OlBf0S=]OiBd0W=b0O1O1O1O1N2O1bC_N^;b1`DaN^;`1`DbN_;_1_DdN_;^1^DdNa;]1]DfNb;Z1\\DhNc;Y1XDmNg;T1oCUOQ<m1O001O001O001O001O0O2O001O001O001O001O001OO0O2O1O1O1O1O1N2O1O1O1O001O1O100O10000O10O1000O1000000O100O01000O1L4F:C=0000O10O0100000000001O000O10001O0hCXNZ;h1cD\\N\\;d1aD`N_;b1ZDcNe;`1UDdNk;^1oCfNP<T21000O100O100O01O001O010O001N1O2O0O2O0O2N101N101N101N2O0O2O0O2O0O2O0O2YOfChNZ<T1nCgNT<U1l0M4N100O2O0O2N10001O001O00001O00nBWOP<h0oCYOQ<g0mC[OS<i0hCXOY<k0aCWO_<m0[CUOf<n0UCSOk<Q1oBQOR=]10O10S1mN4L3M4L3L5Ln^W3"}, "label": "an umbrella with red and white stripes"}]}
{"id": 336353, "image": "COCO_train2014_000000336353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog that is closest to the headboard\"."}], "task": "refering", "answer_template": "The \"the dog that is closest to the headboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 102, 103, 104, 119, 120, 121, 122, 123, 137, 138, 139, 140, 141, 157, 158], "bbox": [0.64144, 0.35589333333333334, 0.8483800000000001, 0.6738933333333333], "iscrowd": 0, "area": 6939.266949999998, "rle": {"size": [375, 500], "counts": "jge35`;3M2M3N3M4L5L2M2O1O1O1O1N2O1O1O00000O1OO1110O0I7O1O1N2N30O1O1O1O1O10O10000000O1000XFeNT9V2I5K2N2N1O2N2N1O1O2O1N2N2N2N001O00001O001O0000[OeGoMZ8R2hGkMY8T2gGkMZ8U2gGjMY8V2hGiMX8X2hGfMY8Z2hGeMX8[2iGcMW8_2iG`MV8a2mGfMh7\\2RHhMn7Z2nGiMP8o20100O001O10O018G?A1N2N3L3O2N1O3L5L3M5K6J6K5J7Ij0VO5Kkck0"}, "label": "the dog that is closest to the headboard"}]}
{"id": 336353, "image": "COCO_train2014_000000336353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog lying on the top side of the bed\"."}], "task": "refering", "answer_template": "The \"a brown dog lying on the top side of the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 102, 103, 104, 119, 120, 121, 122, 123, 137, 138, 139, 140, 141, 157, 158], "bbox": [0.64144, 0.35589333333333334, 0.8483800000000001, 0.6738933333333333], "iscrowd": 0, "area": 6939.266949999998, "rle": {"size": [375, 500], "counts": "jge35`;3M2M3N3M4L5L2M2O1O1O1O1N2O1O1O00000O1OO1110O0I7O1O1N2N30O1O1O1O1O10O10000000O1000XFeNT9V2I5K2N2N1O2N2N1O1O2O1N2N2N2N001O00001O001O0000[OeGoMZ8R2hGkMY8T2gGkMZ8U2gGjMY8V2hGiMX8X2hGfMY8Z2hGeMX8[2iGcMW8_2iG`MV8a2mGfMh7\\2RHhMn7Z2nGiMP8o20100O001O10O018G?A1N2N3L3O2N1O3L5L3M5K6J6K5J7Ij0VO5Kkck0"}, "label": "a brown dog lying on the top side of the bed"}]}
{"id": 106978, "image": "COCO_train2014_000000106978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"achair in front of a man in grey hat\"."}], "task": "refering", "answer_template": "The \"achair in front of a man in grey hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 335, 336, 355, 358, 359, 378, 381, 382, 383], "bbox": [0.44851562500000003, 0.5801666666666667, 0.6571562500000001, 0.9990208333333334], "iscrowd": 0, "area": 13422.365500000002, "rle": {"size": [480, 640], "counts": "hoV47Y>`0A?A>bM\\OTFR1k9QOiEY1V:iNfEY1[:hNbEZ1]:hN_EZ1a:iN[EY1e:hNXEY1h:]12M3M400O1\\Od0YOg0ZOf0M3O1O10000000ROiNSDW1S;CmD=P;FPE:P;FoD;Q;EoD;Q;EoD:R;FkD=U;CdDd0\\;`10000000000000000000000000000000000000000O100000O10000000000000000000000000000000000000000000000007Ih0XOh0WOa0@0000000000lMQHWMo7h2U200O100O1000000000000000000000000PNiD7W;HVE^OBYOX;Y1WE\\OB[OW;X1YE\\O@]OV;W1[EZOA^OT;W1]EYO@@S;V1^EYOAAo:R1gE[OZOFl:a0ZFFkN5_:FXG1[Nc0f<UO`CV1P=;F9F;E;E03L;E<D;E;E<D;E;FQVV3"}, "label": "achair in front of a man in grey hat"}]}
{"id": 106978, "image": "COCO_train2014_000000106978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a crem , white and red color chair\"."}], "task": "refering", "answer_template": "The \"a crem , white and red color chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 335, 336, 355, 358, 359, 378, 381, 382, 383], "bbox": [0.44851562500000003, 0.5801666666666667, 0.6571562500000001, 0.9990208333333334], "iscrowd": 0, "area": 13422.365500000002, "rle": {"size": [480, 640], "counts": "hoV47Y>`0A?A>bM\\OTFR1k9QOiEY1V:iNfEY1[:hNbEZ1]:hN_EZ1a:iN[EY1e:hNXEY1h:]12M3M400O1\\Od0YOg0ZOf0M3O1O10000000ROiNSDW1S;CmD=P;FPE:P;FoD;Q;EoD;Q;EoD:R;FkD=U;CdDd0\\;`10000000000000000000000000000000000000000O100000O10000000000000000000000000000000000000000000000007Ih0XOh0WOa0@0000000000lMQHWMo7h2U200O100O1000000000000000000000000PNiD7W;HVE^OBYOX;Y1WE\\OB[OW;X1YE\\O@]OV;W1[EZOA^OT;W1]EYO@@S;V1^EYOAAo:R1gE[OZOFl:a0ZFFkN5_:FXG1[Nc0f<UO`CV1P=;F9F;E;E03L;E<D;E;E<D;E;FQVV3"}, "label": "a crem , white and red color chair"}]}
{"id": 106978, "image": "COCO_train2014_000000106978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the red hat\"."}], "task": "refering", "answer_template": "The \"the man with the red hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 295, 296, 314, 315, 318, 319, 338, 341, 342], "bbox": [0.647109375, 0.3411458333333333, 0.9031874999999999, 0.8401458333333333], "iscrowd": 0, "area": 19279.69374999999, "rle": {"size": [480, 640], "counts": "o[R61]=1PD3m;1nC4n;OmC6P<NkC6R<MjC7S<MhC7U<LgC8V<LdC9Y<JcC;Y<IbC;[<W1M3M3M3L5J5L4K5L4L4K6L3O2M4L3`FgKU8[4eGiKZ8Z4`GkK_8W4[GmKd8W4UGnKi8U4QGoKo8S4kFQLT9o4O0O2N100O2O0O2N1O1N3L3M3M4LYNnGfLo7U3]HhL`7W3gHeLW7]3mH]LT7d3PITLS7l3RImKQ7S4SIeKP7\\4TI\\Ko6e4e11O100O1O1O1O1O100O100O10000O10000O10000O10000O101N10000O10000O10oK`Fb2`9VMiFj2W9SMkFm2U9PMoFo2R9nLPGR3Q9jLSGU3n8hLTGX3l8eLXGZ3i8cLZG\\3g8`L\\G`3e8]L^Gb3P:O2O0O2O001N2O001N101N13N2M3N2NTMgDR2V1]M_8=aFX2k0]Md85hF_2?_Mh8NoFd24`Mn8GTGj2J`MR9BYGP3@`M\\:_2bEcM_:\\2^EfMd:Y2ZEhMh:V2VEmMk:R2REPNP;o1nDRNT;m1iDVNW;i22N3N4K4L4L4M4K4L4M3L5K4M3L4L5K4M3L5K5K5J5L5K5J3N1O2M2O2N1N2O2N1N2O2N1N2O2Ndel0"}, "label": "the man with the red hat"}]}
{"id": 106978, "image": "COCO_train2014_000000106978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green jacket and blue pants and a red hat\"."}], "task": "refering", "answer_template": "The \"a man in a green jacket and blue pants and a red hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 295, 296, 314, 315, 318, 319, 338, 341, 342], "bbox": [0.647109375, 0.3411458333333333, 0.9031874999999999, 0.8401458333333333], "iscrowd": 0, "area": 19279.69374999999, "rle": {"size": [480, 640], "counts": "o[R61]=1PD3m;1nC4n;OmC6P<NkC6R<MjC7S<MhC7U<LgC8V<LdC9Y<JcC;Y<IbC;[<W1M3M3M3L5J5L4K5L4L4K6L3O2M4L3`FgKU8[4eGiKZ8Z4`GkK_8W4[GmKd8W4UGnKi8U4QGoKo8S4kFQLT9o4O0O2N100O2O0O2N1O1N3L3M3M4LYNnGfLo7U3]HhL`7W3gHeLW7]3mH]LT7d3PITLS7l3RImKQ7S4SIeKP7\\4TI\\Ko6e4e11O100O1O1O1O1O100O100O10000O10000O10000O10000O101N10000O10000O10oK`Fb2`9VMiFj2W9SMkFm2U9PMoFo2R9nLPGR3Q9jLSGU3n8hLTGX3l8eLXGZ3i8cLZG\\3g8`L\\G`3e8]L^Gb3P:O2O0O2O001N2O001N101N13N2M3N2NTMgDR2V1]M_8=aFX2k0]Md85hF_2?_Mh8NoFd24`Mn8GTGj2J`MR9BYGP3@`M\\:_2bEcM_:\\2^EfMd:Y2ZEhMh:V2VEmMk:R2REPNP;o1nDRNT;m1iDVNW;i22N3N4K4L4L4M4K4L4M3L5K4M3L4L5K4M3L5K5K5J5L5K5J3N1O2M2O2N1N2O2N1N2O2N1N2O2Ndel0"}, "label": "a man in a green jacket and blue pants and a red hat"}]}
{"id": 25058, "image": "COCO_train2014_000000025058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a motorcycle sitting on the street\"."}], "task": "refering", "answer_template": "The \"there is a motorcycle sitting on the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 220, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 289, 290, 291, 293, 294], "bbox": [0.571109375, 0.4923541666666667, 0.7986093750000001, 0.7470625000000001], "iscrowd": 0, "area": 8457.201000000001, "rle": {"size": [480, 640], "counts": "Wk[53a>OgA31Lk=4RB3NMn=1SB4LNQ>NPBa0?_Oj<NfBe0=Eh<a0SCFk<9RCKm<V1N100O2N1O101N1O1O100O001O010O001O10O0000001O000000001O000000001O00000O1O2N100O1O1O2O0O1N2M3M4L3O1000001OFaNSC`1k<bNTCf1c<[N]Ch1`<YN_C[1MhNc<N_C[10eN`<0`C\\12aN_<3^C]1l<dNSC]1l<cNUC]1j<dNUC]1k<cNUC]1j<cNWC]1h<dNWC]1i<bNXC^1h<bNWC_1i<aNWC_1h<aNYC_1g<aNXC`1h<_NYCa1g<_NXCb1h<^NXCa1i<^NXCa1j<^NUCc1k<]NUCb1m<^NQCc1o<^NPCb1Q=60001N1O100O10O0O2O000100I7M3N2N2_OgMXD\\2g;eMWD\\2j;eMSDX2R<hMkC1134[1Q<fNoCL4\\1n;dNRDN2^1l;aNUD00_1k;`NVDO1`1i;aNWDO0`1\\<`NdC_1IcNW<NPDl1o;UNQD5B\\1W<eNWDND]1S<fNYDLF]1R<fNXDLH]1Q<fNWDMH^1Q<dNXDLI_1P<dNXDKIb1o;]NjDb1Z<19F2O2N1N2N1O10UBoNg=o0WBSOj=R11N101N2O1N3M2N2N2M4L5J6J]Tl1"}, "label": "there is a motorcycle sitting on the street"}]}
{"id": 25058, "image": "COCO_train2014_000000025058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red motorcycle next to the curve\"."}], "task": "refering", "answer_template": "The \"red motorcycle next to the curve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 220, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 289, 290, 291, 293, 294], "bbox": [0.571109375, 0.4923541666666667, 0.7986093750000001, 0.7470625000000001], "iscrowd": 0, "area": 8457.201000000001, "rle": {"size": [480, 640], "counts": "Wk[53a>OgA31Lk=4RB3NMn=1SB4LNQ>NPBa0?_Oj<NfBe0=Eh<a0SCFk<9RCKm<V1N100O2N1O101N1O1O100O001O010O001O10O0000001O000000001O000000001O00000O1O2N100O1O1O2O0O1N2M3M4L3O1000001OFaNSC`1k<bNTCf1c<[N]Ch1`<YN_C[1MhNc<N_C[10eN`<0`C\\12aN_<3^C]1l<dNSC]1l<cNUC]1j<dNUC]1k<cNUC]1j<cNWC]1h<dNWC]1i<bNXC^1h<bNWC_1i<aNWC_1h<aNYC_1g<aNXC`1h<_NYCa1g<_NXCb1h<^NXCa1i<^NXCa1j<^NUCc1k<]NUCb1m<^NQCc1o<^NPCb1Q=60001N1O100O10O0O2O000100I7M3N2N2_OgMXD\\2g;eMWD\\2j;eMSDX2R<hMkC1134[1Q<fNoCL4\\1n;dNRDN2^1l;aNUD00_1k;`NVDO1`1i;aNWDO0`1\\<`NdC_1IcNW<NPDl1o;UNQD5B\\1W<eNWDND]1S<fNYDLF]1R<fNXDLH]1Q<fNWDMH^1Q<dNXDLI_1P<dNXDKIb1o;]NjDb1Z<19F2O2N1N2N1O10UBoNg=o0WBSOj=R11N101N2O1N3M2N2N2M4L5J6J]Tl1"}, "label": "red motorcycle next to the curve"}]}
{"id": 147941, "image": "COCO_train2014_000000147941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in red and white plays tennis\"."}], "task": "refering", "answer_template": "The \"a woman in red and white plays tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 335, 336, 358, 359, 381, 382], "bbox": [0.391890625, 0.2781458333333333, 0.72128125, 0.9853541666666666], "iscrowd": 0, "area": 27046.44564999999, "rle": {"size": [480, 640], "counts": "cke33k>4L5K4L5K3O0O100O1O100O1O100O100O0010O0O2O0O2O001N1O2N1O2N2N2O00100O100O010O100O02O00000O10001O00000000001O000000000O2O0000000oB@b;a0\\DCa;=_DG];9cDI[;7eDJZ;6fDLX;4hDMW;3hDOW;1iD1U;OkD2T;NlD4R;LnD5Q;JPE7o:IQE9m:GSE:l:FTE<j:DVE=i:CWE?g:AYE`0f:@ZEb0d:^O\\Ed0b:\\O^Ee0a:[O_Eg0_:YOaEi0]:WOcEk0[:UOeEo0V:ROjES1Q:mNoEX1l9hNTF\\1h9dNXFa1c9_N]Fe1_9[NaFi1[9WNeF_2e8aM[Ga2b8`M^Gb2`8^M`Ge2]8[McGg2[8YMeGi2Y8WMgGk2W8UMiGm2T8TMlGn2R8UMkGm2S8VMjGm2R8XMjGl2R8WMkGl2R8XM^GW3^8i1M3M3L4M3M3M4L3M3M3M3M3M3M3nIfHX4d7oJkHQ5b80O10000000000O1000000000000O1000O1O1O1O001O100N2N2MlMfJgJY5V5PKdJP5\\5UK`Ji4b5\\KYJc4h5fKnIZ4S6jKiIU4X6nKdIR4]6QL`In3`6VL[Ik3f6XLVIh3k6[LPIf3Q7^LjHb3W7cLbH_3^7P210N2J7I7J6I7N2O1O1N2O1O1O1N2O2N1N3N2M2N3hKPG\\2R9dMnFZ2U9dMmFY2V9eMlFW2W9hMkFT2Y9kMgFR2\\9mMfFP2]9mMfFQ2\\9jMiFU2W9fMnFY2U9aMPG]2S9]MRGa2Q9XMUGf2n8SMWGl2`:N1O1O2N1N3XOnC_NS<^1SD^Nn;^1XD_Nj;\\1\\DaNf;Z1bD`Nd;X1Y1G9GUWc2"}, "label": "a woman in red and white plays tennis"}]}
{"id": 147941, "image": "COCO_train2014_000000147941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with red dress playing tennis\"."}], "task": "refering", "answer_template": "The \"a girl with red dress playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 335, 336, 358, 359, 381, 382], "bbox": [0.391890625, 0.2781458333333333, 0.72128125, 0.9853541666666666], "iscrowd": 0, "area": 27046.44564999999, "rle": {"size": [480, 640], "counts": "cke33k>4L5K4L5K3O0O100O1O100O1O100O100O0010O0O2O0O2O001N1O2N1O2N2N2O00100O100O010O100O02O00000O10001O00000000001O000000000O2O0000000oB@b;a0\\DCa;=_DG];9cDI[;7eDJZ;6fDLX;4hDMW;3hDOW;1iD1U;OkD2T;NlD4R;LnD5Q;JPE7o:IQE9m:GSE:l:FTE<j:DVE=i:CWE?g:AYE`0f:@ZEb0d:^O\\Ed0b:\\O^Ee0a:[O_Eg0_:YOaEi0]:WOcEk0[:UOeEo0V:ROjES1Q:mNoEX1l9hNTF\\1h9dNXFa1c9_N]Fe1_9[NaFi1[9WNeF_2e8aM[Ga2b8`M^Gb2`8^M`Ge2]8[McGg2[8YMeGi2Y8WMgGk2W8UMiGm2T8TMlGn2R8UMkGm2S8VMjGm2R8XMjGl2R8WMkGl2R8XM^GW3^8i1M3M3L4M3M3M4L3M3M3M3M3M3M3nIfHX4d7oJkHQ5b80O10000000000O1000000000000O1000O1O1O1O001O100N2N2MlMfJgJY5V5PKdJP5\\5UK`Ji4b5\\KYJc4h5fKnIZ4S6jKiIU4X6nKdIR4]6QL`In3`6VL[Ik3f6XLVIh3k6[LPIf3Q7^LjHb3W7cLbH_3^7P210N2J7I7J6I7N2O1O1N2O1O1O1N2O2N1N3N2M2N3hKPG\\2R9dMnFZ2U9dMmFY2V9eMlFW2W9hMkFT2Y9kMgFR2\\9mMfFP2]9mMfFQ2\\9jMiFU2W9fMnFY2U9aMPG]2S9]MRGa2Q9XMUGf2n8SMWGl2`:N1O1O2N1N3XOnC_NS<^1SD^Nn;^1XD_Nj;\\1\\DaNf;Z1bD`Nd;X1Y1G9GUWc2"}, "label": "a girl with red dress playing tennis"}]}
{"id": 147941, "image": "COCO_train2014_000000147941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in back\"."}], "task": "refering", "answer_template": "The \"person in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 105, 106, 107, 127, 128, 129, 151, 152, 291, 311, 313, 314, 337], "bbox": [0.541015625, 0.14397916666666666, 0.6809062499999999, 0.8316041666666667], "iscrowd": 0, "area": 5376.061350000001, "rle": {"size": [480, 640], "counts": "YbR51o>0O1O1O2N100O1O1O1O2O0]HO62Gd0E]O9h0CYO:j0EWO8l0HUO6m00lNOX14bNK_1k60O1O1iN]NQEc1n:]NREd1l:^NSEc1l:]NUEb1k:_NTEb1k:cNPE^1o:POdDP1Z;WO`Dj0_;XO`Dh0_;ZO_Dg0_;[O`Dg0^;[O`Df0`;[O\\Dh0d;Z1O10000O1000000O1000O001N2O1O2N1O1O1O1O2N3L3M4TNnCd0U<WOPDf0S<UOSDg0U=M4K4M4L3M2M4M2NoH8oNEo0?QOAl0b0UO\\Oi0g0WOYOf0j0[OTOc0o0]OQO`0R1@mN>V1CiN:Z1FeN9]1HbN7_1I`N6a1L^N2d1N[N1g1b62N2N2N2N2`0UOm0ROnXo2"}, "label": "person in back"}]}
{"id": 147941, "image": "COCO_train2014_000000147941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a purple shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a purple shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 105, 106, 107, 127, 128, 129, 151, 152, 291, 311, 313, 314, 337], "bbox": [0.541015625, 0.14397916666666666, 0.6809062499999999, 0.8316041666666667], "iscrowd": 0, "area": 5376.061350000001, "rle": {"size": [480, 640], "counts": "YbR51o>0O1O1O2N100O1O1O1O2O0]HO62Gd0E]O9h0CYO:j0EWO8l0HUO6m00lNOX14bNK_1k60O1O1iN]NQEc1n:]NREd1l:^NSEc1l:]NUEb1k:_NTEb1k:cNPE^1o:POdDP1Z;WO`Dj0_;XO`Dh0_;ZO_Dg0_;[O`Dg0^;[O`Df0`;[O\\Dh0d;Z1O10000O1000000O1000O001N2O1O2N1O1O1O1O2N3L3M4TNnCd0U<WOPDf0S<UOSDg0U=M4K4M4L3M2M4M2NoH8oNEo0?QOAl0b0UO\\Oi0g0WOYOf0j0[OTOc0o0]OQO`0R1@mN>V1CiN:Z1FeN9]1HbN7_1I`N6a1L^N2d1N[N1g1b62N2N2N2N2`0UOm0ROnXo2"}, "label": "a woman wearing a purple shirt"}]}
{"id": 74215, "image": "COCO_train2014_000000074215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black\"."}], "task": "refering", "answer_template": "The \"a man in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 143, 144, 160, 161, 176, 177, 178, 193, 194, 195, 227, 228, 229, 244, 245, 246, 261, 262, 263, 278, 279, 280, 295, 296, 297, 312, 313, 314, 330, 331], "bbox": [0.3500625, 0.336390625, 0.533125, 0.8359375], "iscrowd": 0, "area": 15502.015899999991, "rle": {"size": [640, 480], "counts": "Y[Y37ic0;E4L5KN2LeN\\Ob_O`0]`0Dd_O8[`0Mh14K401U_O_OU>c0eACZ>=bAH]>9^ALa>5ZA1c>0YA5f>MWA5?WOm;e0dC49>Q;@eDE]NIe1]1S;XOjDCbNC[1l1T;POnDBgN]OQ1[2U;gNSEAMS2j:_NXE_OGZ2n:YNZEg3d:[L[Ee3d:\\L\\Ee3b:]L]Ec3b:_L^Ea3`:aL_E`3_:bL`E^3_:cLaE^3]:dLbE\\3^:eLaE\\3]:fLbEZ3]:hLbEY3Y9VKVFb1a0X3W9aKnEY1k0V3U9hKiET1Q1T3U9kKgES1S1R3T9nKgER1T1P3T9QLfEo0V1P3R9YNmFg1R9ZNnFf1Q9ZNPGg1m8[NSGe1l8[NTGf1l8ZNTGd1m8\\NTGa1o8_NQG5oN]MS:]2nFBHjM]9d2kFgNg0]Nb8l2gFeNm0XN_8R3eFdNF_NLDl9Y3bFcNVO2W:Z1cFj1j8VNVGj1j8UNWGk1i8UNWGl1h8SNYGm1g8QNZGP2f8lM^GT2a8jMbGV2^8fMfGZ2Z8cMiG]2W8`MkGa2U8\\MnGe2R8WMQHi2o7TMTHl2m7PMUHQ3k7lLXHR3j7lLXHR3k7iLYHU3i7hLZHU3i7gLZHX3i7dLZHY3i7oKbEBi2]4h7nKcE@h2`4g7nKdE^Og2a4i7nKeE[Of2c4j7oKhHn3\\7PLfHl3_7PLcHl3b7RL`Hj3d7QL`H9^Ld2^?WMXAk0oNOl?SOmB?W=\\OXC5hQ\\4"}, "label": "a man in black"}]}
{"id": 74215, "image": "COCO_train2014_000000074215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on the skating board with helmet\"."}], "task": "refering", "answer_template": "The \"a man on the skating board with helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 143, 144, 160, 161, 176, 177, 178, 193, 194, 195, 227, 228, 229, 244, 245, 246, 261, 262, 263, 278, 279, 280, 295, 296, 297, 312, 313, 314, 330, 331], "bbox": [0.3500625, 0.336390625, 0.533125, 0.8359375], "iscrowd": 0, "area": 15502.015899999991, "rle": {"size": [640, 480], "counts": "Y[Y37ic0;E4L5KN2LeN\\Ob_O`0]`0Dd_O8[`0Mh14K401U_O_OU>c0eACZ>=bAH]>9^ALa>5ZA1c>0YA5f>MWA5?WOm;e0dC49>Q;@eDE]NIe1]1S;XOjDCbNC[1l1T;POnDBgN]OQ1[2U;gNSEAMS2j:_NXE_OGZ2n:YNZEg3d:[L[Ee3d:\\L\\Ee3b:]L]Ec3b:_L^Ea3`:aL_E`3_:bL`E^3_:cLaE^3]:dLbE\\3^:eLaE\\3]:fLbEZ3]:hLbEY3Y9VKVFb1a0X3W9aKnEY1k0V3U9hKiET1Q1T3U9kKgES1S1R3T9nKgER1T1P3T9QLfEo0V1P3R9YNmFg1R9ZNnFf1Q9ZNPGg1m8[NSGe1l8[NTGf1l8ZNTGd1m8\\NTGa1o8_NQG5oN]MS:]2nFBHjM]9d2kFgNg0]Nb8l2gFeNm0XN_8R3eFdNF_NLDl9Y3bFcNVO2W:Z1cFj1j8VNVGj1j8UNWGk1i8UNWGl1h8SNYGm1g8QNZGP2f8lM^GT2a8jMbGV2^8fMfGZ2Z8cMiG]2W8`MkGa2U8\\MnGe2R8WMQHi2o7TMTHl2m7PMUHQ3k7lLXHR3j7lLXHR3k7iLYHU3i7hLZHU3i7gLZHX3i7dLZHY3i7oKbEBi2]4h7nKcE@h2`4g7nKdE^Og2a4i7nKeE[Of2c4j7oKhHn3\\7PLfHl3_7PLcHl3b7RL`Hj3d7QL`H9^Ld2^?WMXAk0oNOl?SOmB?W=\\OXC5hQ\\4"}, "label": "a man on the skating board with helmet"}]}
{"id": 74215, "image": "COCO_train2014_000000074215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with orange mirrored goggles holding skies\"."}], "task": "refering", "answer_template": "The \"a woman with orange mirrored goggles holding skies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 95, 96, 97, 112, 113, 114, 129, 130, 131, 132, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 178, 179, 180, 181, 182, 183, 185, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 252, 264, 265, 266, 267, 269, 270, 281, 282, 283, 285, 286, 287, 298, 299, 300, 302, 303, 315, 316, 317, 319, 320, 332, 333, 334, 336, 337, 349, 350, 353, 354, 366, 367, 370, 371, 386, 387], "bbox": [0.34497916666666667, 0.185109375, 0.9682083333333333, 1.0], "iscrowd": 0, "area": 55718.91889999997, "rle": {"size": [640, 480], "counts": "bRX3f0Uc06J5M3N3M2O1N2N3M2000001N10001O000O101O00000010O1O1O1O1O1O1O001O1O1O10O01O1O001O1O001O001O00O10O01O100O100O010O10000O01000O100O10OM4M3L4L4M2M4L4M3L4L4K4M4L4L4L4L3O2N2N2M3N2N1QESMk4j3\\IVM_6S3eHfMV7a2WHoMe7Y2gGWNU8P2XG`Nd8h1hFhNV9S6M3N2N2M3N2M2O1N3N1O1N3N1N3N1N2O2N1N2O1N_OjGZEV8c:nG]EP8`:VH_Ei7a:XH_Ed7c:^H]E^7f:cHZEZ7g:iHVEU7m:lHoDT7S;nHjDQ7Y;h02O200O2N101\\O^DgHc;U7dDhH];S7iDjHW;g6\\EUIc:f6fEXIY:a6e101O0O10000101N1O1O2N101N1O1M4F9PNiHPG`7k8UI[FU7`9`IfEj6V:a1L5L4M30O01O100O010O100O10O01O2O0O100O1O1M3L4L4L4M3L4L4L4L4M3K6G8I8G9H8G9H8I7KTLeI\\JX6d5jI^JT6`5nIaJQ6]5QJdJn5[5SJfJ[1XLa2P9VLiJj0fLQ3]8WLnJ9TMa3g5jKQK>U2I`Ma3ZNfLj6HSK:Y4T3mIlL`6MVK5^4LiIb26iMo5K`K0b4HlI]29YN\\5IoJ[OK>g5FlIY2<cNo4KoJ_OM6m5CnIT2>kNf4OmJBNOT6_OoIP2a0RO]43iJG0G[6ZOPJn1d0YOS4U2Y1cLRJi1g0@j3W2^1^LSJd1k0H`3Y2b1ZLUJ_1n0OV3\\2Z2`MeJ6l2]2`2WMhJ>d2^2e2nLlJe0[2`2j2dLQKl0Q2d2o2ZLUKR1i1g2S3QLYKX1a1j2W3hK]K^1Y1m2R6TMkIo2U6QMgIS3Y6mLcIW3]6iL_I[3a6eLZI`3f6aLUIc3j6^LQIg3o6YLmHk3S7ULhHP4X7PLdH^3aM_Kk9S1_HZ3VN[K\\9[1ZHT3kNZKj8c1VHn2\\9RM`Fi2i9WMRFd2Y:[McE`2g:_MTE\\2V;eMeDV2e;iMWDW2m;iMPDU2V<jMfCV2^<jM_CT2g<kMUCT2P=mMlBR2Y=nMbBQ2c=oMZBP2k=PNPBo1V>QNfAo1]>g14M2M4L5_LTAh1P?WNSAb1R?\\NSA]1R?aNQAX1Q?iNTAo0n>QOWAi0i>YO\\A?f>B_A6b>Lh20O2O1N1I8I7I6J7I7I7H72O2N2M3N2M3M3M3N2M3M3M4M4K4L5L3L4L5K2O1N2N2N2K6I6J6J6J6JoQ9"}, "label": "a woman with orange mirrored goggles holding skies"}]}
{"id": 74215, "image": "COCO_train2014_000000074215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in ski googles holding her skis in her hands\"."}], "task": "refering", "answer_template": "The \"a woman in ski googles holding her skis in her hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 95, 96, 97, 112, 113, 114, 129, 130, 131, 132, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 178, 179, 180, 181, 182, 183, 185, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 252, 264, 265, 266, 267, 269, 270, 281, 282, 283, 285, 286, 287, 298, 299, 300, 302, 303, 315, 316, 317, 319, 320, 332, 333, 334, 336, 337, 349, 350, 353, 354, 366, 367, 370, 371, 386, 387], "bbox": [0.34497916666666667, 0.185109375, 0.9682083333333333, 1.0], "iscrowd": 0, "area": 55718.91889999997, "rle": {"size": [640, 480], "counts": "bRX3f0Uc06J5M3N3M2O1N2N3M2000001N10001O000O101O00000010O1O1O1O1O1O1O001O1O1O10O01O1O001O1O001O001O00O10O01O100O100O010O10000O01000O100O10OM4M3L4L4M2M4L4M3L4L4K4M4L4L4L4L3O2N2N2M3N2N1QESMk4j3\\IVM_6S3eHfMV7a2WHoMe7Y2gGWNU8P2XG`Nd8h1hFhNV9S6M3N2N2M3N2M2O1N3N1O1N3N1N3N1N2O2N1N2O1N_OjGZEV8c:nG]EP8`:VH_Ei7a:XH_Ed7c:^H]E^7f:cHZEZ7g:iHVEU7m:lHoDT7S;nHjDQ7Y;h02O200O2N101\\O^DgHc;U7dDhH];S7iDjHW;g6\\EUIc:f6fEXIY:a6e101O0O10000101N1O1O2N101N1O1M4F9PNiHPG`7k8UI[FU7`9`IfEj6V:a1L5L4M30O01O100O010O100O10O01O2O0O100O1O1M3L4L4L4M3L4L4L4L4M3K6G8I8G9H8G9H8I7KTLeI\\JX6d5jI^JT6`5nIaJQ6]5QJdJn5[5SJfJ[1XLa2P9VLiJj0fLQ3]8WLnJ9TMa3g5jKQK>U2I`Ma3ZNfLj6HSK:Y4T3mIlL`6MVK5^4LiIb26iMo5K`K0b4HlI]29YN\\5IoJ[OK>g5FlIY2<cNo4KoJ_OM6m5CnIT2>kNf4OmJBNOT6_OoIP2a0RO]43iJG0G[6ZOPJn1d0YOS4U2Y1cLRJi1g0@j3W2^1^LSJd1k0H`3Y2b1ZLUJ_1n0OV3\\2Z2`MeJ6l2]2`2WMhJ>d2^2e2nLlJe0[2`2j2dLQKl0Q2d2o2ZLUKR1i1g2S3QLYKX1a1j2W3hK]K^1Y1m2R6TMkIo2U6QMgIS3Y6mLcIW3]6iL_I[3a6eLZI`3f6aLUIc3j6^LQIg3o6YLmHk3S7ULhHP4X7PLdH^3aM_Kk9S1_HZ3VN[K\\9[1ZHT3kNZKj8c1VHn2\\9RM`Fi2i9WMRFd2Y:[McE`2g:_MTE\\2V;eMeDV2e;iMWDW2m;iMPDU2V<jMfCV2^<jM_CT2g<kMUCT2P=mMlBR2Y=nMbBQ2c=oMZBP2k=PNPBo1V>QNfAo1]>g14M2M4L5_LTAh1P?WNSAb1R?\\NSA]1R?aNQAX1Q?iNTAo0n>QOWAi0i>YO\\A?f>B_A6b>Lh20O2O1N1I8I7I6J7I7I7H72O2N2M3N2M3M3M3N2M3M3M4M4K4L5L3L4L5K2O1N2N2N2K6I6J6J6J6JoQ9"}, "label": "a woman in ski googles holding her skis in her hands"}]}
{"id": 426478, "image": "COCO_train2014_000000426478.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hotdog to the right of another hotdog and chips\"."}], "task": "refering", "answer_template": "The \"the hotdog to the right of another hotdog and chips\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.688125, 0.2727488151658768, 1.0, 1.0], "iscrowd": 0, "area": 52961.1202, "rle": {"size": [422, 640], "counts": "gbe51R=8I7H8H8H8I7cGdNc4d1QKeNj4c1jJgNP5a1dJiNW5^1]JkN^5W2\\ISN^6]4L3N2M3M3M3N2M3M3M3N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O1000000000000000000000000000000O10000000000000000O100000000000000O10000000000000000O100001O00001O00001O0O101O00001O001O00001O00001O00001O000O2O002N5K5K4L5K5K5K5K9Gb0^OTJ"}, "label": "the hotdog to the right of another hotdog and chips"}]}
{"id": 426478, "image": "COCO_train2014_000000426478.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hotdog with no bun\"."}], "task": "refering", "answer_template": "The \"hotdog with no bun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.688125, 0.2727488151658768, 1.0, 1.0], "iscrowd": 0, "area": 52961.1202, "rle": {"size": [422, 640], "counts": "gbe51R=8I7H8H8H8I7cGdNc4d1QKeNj4c1jJgNP5a1dJiNW5^1]JkN^5W2\\ISN^6]4L3N2M3M3M3N2M3M3M3N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O1000000000000000000000000000000O10000000000000000O100000000000000O10000000000000000O100001O00001O00001O0O101O00001O001O00001O00001O00001O000O2O002N5K5K4L5K5K5K5K9Gb0^OTJ"}, "label": "hotdog with no bun"}]}
{"id": 426478, "image": "COCO_train2014_000000426478.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog with salsa by the green flower on the plate\"."}], "task": "refering", "answer_template": "The \"a hot dog with salsa by the green flower on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 195, 196, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 306, 307], "bbox": [0.157515625, 0.00023696682464454977, 0.613046875, 0.8737677725118483], "iscrowd": 0, "area": 60035.66525000002, "rle": {"size": [422, 640], "counts": "flY18k<;E7I4L5K4L5K4L5K4L5K5K4K6K4L5K4L5K4L4L4L4L4L4L4L4M3M3M3N2M3N2M3M2O1N3N1N2N2O2M2O1N2N3N1N2O1N3N1N3M2O1N3N1N2N3N1N2O2M2N3N1N2O2M2N2O2M2O1N3M2O2M2O1N3N1N3M2O1N3N1N3M2O2M2O1N3M2O2M2O2M2N3N1N3N1N3N2M2N3N1N3N1N3M2O2M3N1N3M2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N1O100O10O0mJcM@\\2b0jMWOV2k0PNmNo1U14gLL[34bLM_34_LLb35\\LKf35XLLh35VLKl36PLKQ46mKKT45jKKW46gKJZ47cKK^45`KL`45^KLc44[KLf45XKLh45UKMk44SKMn43PKNP53nJNR54jJMW54gJMZ53dJN\\53bJN^53`JN`53]JOd51ZJOg53VJNj54SJMm54QJMo55mIMT64iIMW64gILZ66cIK]6_31O0000001O000O101O001O1O1O001O1O1O1N101O1O1O1O001O1O1O001O1N2O1O00100O010O10O0100O010O10O01000O2O000O10000O100O10N2OO001O0O2O0hN^JmJb5e4UKQKm4l4]KlJc4S5i1N1O1O2N1O2N1N2O2N1O1O2N100010O01O2O0O101N1O101N1O2O0O101N1O101N101N1O101N1O4K5L4L4K4M4L4K7Jc0]OlTV3"}, "label": "a hot dog with salsa by the green flower on the plate"}]}
{"id": 426478, "image": "COCO_train2014_000000426478.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a neat hotdog and bun on lunch plate\"."}], "task": "refering", "answer_template": "The \"a neat hotdog and bun on lunch plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 195, 196, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 306, 307], "bbox": [0.157515625, 0.00023696682464454977, 0.613046875, 0.8737677725118483], "iscrowd": 0, "area": 60035.66525000002, "rle": {"size": [422, 640], "counts": "flY18k<;E7I4L5K4L5K4L5K4L5K5K4K6K4L5K4L5K4L4L4L4L4L4L4L4M3M3M3N2M3N2M3M2O1N3N1N2N2O2M2O1N2N3N1N2O1N3N1N3M2O1N3N1N2N3N1N2O2M2N3N1N2O2M2N2O2M2O1N3M2O2M2O1N3N1N3M2O1N3N1N3M2O2M2O1N3M2O2M2O2M2N3N1N3N1N3N2M2N3N1N3N1N3M2O2M3N1N3M2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N1O100O10O0mJcM@\\2b0jMWOV2k0PNmNo1U14gLL[34bLM_34_LLb35\\LKf35XLLh35VLKl36PLKQ46mKKT45jKKW46gKJZ47cKK^45`KL`45^KLc44[KLf45XKLh45UKMk44SKMn43PKNP53nJNR54jJMW54gJMZ53dJN\\53bJN^53`JN`53]JOd51ZJOg53VJNj54SJMm54QJMo55mIMT64iIMW64gILZ66cIK]6_31O0000001O000O101O001O1O1O001O1O1O1N101O1O1O1O001O1O1O001O1N2O1O00100O010O10O0100O010O10O01000O2O000O10000O100O10N2OO001O0O2O0hN^JmJb5e4UKQKm4l4]KlJc4S5i1N1O1O2N1O2N1N2O2N1O1O2N100010O01O2O0O101N1O101N1O2O0O101N1O101N101N1O101N1O4K5L4L4K4M4L4K7Jc0]OlTV3"}, "label": "a neat hotdog and bun on lunch plate"}]}
{"id": 188911, "image": "COCO_train2014_000000188911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the face and the body of the white sheep looking out the slats of the fence\"."}], "task": "refering", "answer_template": "The \"the face and the body of the white sheep looking out the slats of the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.0013124999999999999, 0.3653347280334728, 0.515296875, 0.6624895397489539], "iscrowd": 0, "area": 28073.729000000003, "rle": {"size": [478, 640], "counts": "dg0W1U=b00000000O100000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000O10000000O100000000000000O1000000O1000000O1000000O10000O1000000O10000fNlMjET2V:mMhET2W:nMhER2W:QNgEo1Y:RNeEo1Z:TNdEl1[:VNdEj1\\:WNcEi1\\:ZNaEg1^:[NaEe1_:\\N`Ed1_:_N_Ea1a:`N]Ea1c:`N\\E`1d:`N\\E`1c:aN\\E`1d:`N\\E_1e:bNZE^1f:bNZE^1f:bNYE_1f:bNZE^1f:bNZE^1f:cNYE]1g:cNYE]1g:cNYE]1g:cNZE\\1e:fNZEZ1f:fNZEZ1f:fNZEZ1f:fNZEZ1f:gNYEY1f:hNZEX1f:hNZEX1f:hN[EW1e:jNZEV1f:jNZEV1f:jNZEV1e:kN[EU1e:kN[EU1e:lNZET1f:lNZET1f:lN[ES1e:mN[ES1d:nN\\ER1d:oN[EQ1e:oN[EQ1e:oN[EQ1e:oN[EQ1e:oN[EQ1e:PO[Eo0e:QO[Eo0e:QO[Eo0e:QOYEQ1g:oNVET1j:lNSEW1m:jNPEX1P;[10000000000000000000O1000000000O100O10000O10000O100O10000O100O100O1O100O100O1O100O1O100O100O100000000000000000000001O000O100000000000000000000000000000000000000000000000000000001O000000001O001O000010O0001O001O00001O001O1O1O1J6ZM[El0k:TOYEb0m:]OUE=P;BSE6S;JQ2O00100O0010O01O01N100O100O010O100O100O010O100O100000O2O00001N100O2O000O2O0O2O0O101N10001N10001N101O0O2O0O101O0O2M2O2N1N3N101N1O10X``4"}, "label": "the face and the body of the white sheep looking out the slats of the fence"}]}
{"id": 188911, "image": "COCO_train2014_000000188911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lightest sheep standing at the end next to the brown sheep and whose head you can see\"."}], "task": "refering", "answer_template": "The \"the lightest sheep standing at the end next to the brown sheep and whose head you can see\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.0013124999999999999, 0.3653347280334728, 0.515296875, 0.6624895397489539], "iscrowd": 0, "area": 28073.729000000003, "rle": {"size": [478, 640], "counts": "dg0W1U=b00000000O100000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000O10000000O100000000000000O1000000O1000000O1000000O10000O1000000O10000fNlMjET2V:mMhET2W:nMhER2W:QNgEo1Y:RNeEo1Z:TNdEl1[:VNdEj1\\:WNcEi1\\:ZNaEg1^:[NaEe1_:\\N`Ed1_:_N_Ea1a:`N]Ea1c:`N\\E`1d:`N\\E`1c:aN\\E`1d:`N\\E_1e:bNZE^1f:bNZE^1f:bNYE_1f:bNZE^1f:bNZE^1f:cNYE]1g:cNYE]1g:cNYE]1g:cNZE\\1e:fNZEZ1f:fNZEZ1f:fNZEZ1f:fNZEZ1f:gNYEY1f:hNZEX1f:hNZEX1f:hN[EW1e:jNZEV1f:jNZEV1f:jNZEV1e:kN[EU1e:kN[EU1e:lNZET1f:lNZET1f:lN[ES1e:mN[ES1d:nN\\ER1d:oN[EQ1e:oN[EQ1e:oN[EQ1e:oN[EQ1e:oN[EQ1e:PO[Eo0e:QO[Eo0e:QO[Eo0e:QOYEQ1g:oNVET1j:lNSEW1m:jNPEX1P;[10000000000000000000O1000000000O100O10000O10000O100O10000O100O100O1O100O100O1O100O1O100O100O100000000000000000000001O000O100000000000000000000000000000000000000000000000000000001O000000001O001O000010O0001O001O00001O001O1O1O1J6ZM[El0k:TOYEb0m:]OUE=P;BSE6S;JQ2O00100O0010O01O01N100O100O010O100O100O010O100O100000O2O00001N100O2O000O2O0O2O0O101N10001N10001N101O0O2O0O101O0O2M2O2N1N3N101N1O10X``4"}, "label": "the lightest sheep standing at the end next to the brown sheep and whose head you can see"}]}
{"id": 188911, "image": "COCO_train2014_000000188911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep that doesn ' t have it ' s face showing\"."}], "task": "refering", "answer_template": "The \"the sheep that doesn ' t have it ' s face showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271], "bbox": [0.525875, 0.48326359832635984, 1.0, 0.6517364016736401], "iscrowd": 0, "area": 15222.623799999996, "rle": {"size": [478, 640], "counts": "ecm41l>1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O1001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000000000001N1000000000000000001O0000000000000000001O000000000000000000001O0000000000000000001O0000000000000000001O0O1000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O000000000O2O0000000000001O0000000000001O0000000000001O000000O10000O10000O10000O100O10000O10000O10000O10000O10000O010N2N2N2N2NK5G8H9H8G9G9GSI"}, "label": "the sheep that doesn ' t have it ' s face showing"}]}
{"id": 188911, "image": "COCO_train2014_000000188911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown sheep head\"."}], "task": "refering", "answer_template": "The \"brown sheep head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 151, 152, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 262, 263, 264], "bbox": [0.40499999999999997, 0.36667364016736403, 0.743859375, 0.6566108786610879], "iscrowd": 0, "area": 15288.655499999999, "rle": {"size": [478, 640], "counts": "^Vi32h>8H4L4L4F:F:H800000000000000O100000000000000O1O1XOcNmC_1Q<h0N2N2O100O10000O10000O1kNZMRFf2m9[MRFf2n9[MQFe2n9\\MQFe2o9[MQFe2n9]MPFd2P:\\MPFd2P:]MoEc2P:^MoEc2Q:]MoEc2Q:^MmEc2R:^MnEb2R:_MlEb2T:^MlEb2T:^MkEc2T:_MkEa2U:_MjEb2V:_MhEb2W:_MiEa2W:_MhEb2X:_MfEb2Y:_MgE`2Z:aMdE`2\\:`McEa2\\:`MdE`2\\:aMbE`2^:`MaEa2_:`M`E`2_:aM`E`2`:`M_Ea2a:`M]Ea2b:`M\\Eb2d:_MZEb2f:k0O1000000O10000O01N2O001N2O1N3N1O1N2O1N2O1O1O2M2O1O1O1N2O1O1O2N1N2O1O100O2O0O101N10001N100O2O000O101O0O1O2N1O2N1O1O2N101N10001N101O000O2O001N101O0O2O1O001N2O2K5J5L5I6F;N1O2N000000O10000O1000000O10000O10000O101OO010000O01000000O10000O101O000O10000O100O10000O100O2O000O101N100O1O2N1O1O2N2N4L4MR[\\2"}, "label": "brown sheep head"}]}
{"id": 188911, "image": "COCO_train2014_000000188911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black face sheep stainding with other sheep\"."}], "task": "refering", "answer_template": "The \"a black face sheep stainding with other sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 151, 152, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 262, 263, 264], "bbox": [0.40499999999999997, 0.36667364016736403, 0.743859375, 0.6566108786610879], "iscrowd": 0, "area": 15288.655499999999, "rle": {"size": [478, 640], "counts": "^Vi32h>8H4L4L4F:F:H800000000000000O100000000000000O1O1XOcNmC_1Q<h0N2N2O100O10000O10000O1kNZMRFf2m9[MRFf2n9[MQFe2n9\\MQFe2o9[MQFe2n9]MPFd2P:\\MPFd2P:]MoEc2P:^MoEc2Q:]MoEc2Q:^MmEc2R:^MnEb2R:_MlEb2T:^MlEb2T:^MkEc2T:_MkEa2U:_MjEb2V:_MhEb2W:_MiEa2W:_MhEb2X:_MfEb2Y:_MgE`2Z:aMdE`2\\:`McEa2\\:`MdE`2\\:aMbE`2^:`MaEa2_:`M`E`2_:aM`E`2`:`M_Ea2a:`M]Ea2b:`M\\Eb2d:_MZEb2f:k0O1000000O10000O01N2O001N2O1N3N1O1N2O1N2O1O1O2M2O1O1O1N2O1O1O2N1N2O1O100O2O0O101N10001N100O2O000O101O0O1O2N1O2N1O1O2N101N10001N101O000O2O001N101O0O2O1O001N2O2K5J5L5I6F;N1O2N000000O10000O1000000O10000O10000O101OO010000O01000000O10000O101O000O10000O100O10000O100O2O000O101N100O1O2N1O1O2N2N4L4MR[\\2"}, "label": "a black face sheep stainding with other sheep"}]}
{"id": 41459, "image": "COCO_train2014_000000041459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man riding a bike wearing a rainbow colored hat\"."}], "task": "refering", "answer_template": "The \"man riding a bike wearing a rainbow colored hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 90, 91, 92, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 197, 198, 199, 206, 207, 208, 209, 210, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 277, 278, 279, 294, 295, 296, 311, 312, 313, 328, 329, 330, 345, 346, 347, 362, 363, 364, 365, 379, 380, 381, 382], "bbox": [0.15012526096033402, 0.182015625, 0.7236116910229645, 0.986515625], "iscrowd": 0, "area": 49593.447300000014, "rle": {"size": [640, 479], "counts": "YZ]13]13j`04k^O8n`0Ng^O=Ra0^1I8H7I4L3M3N2M3M3M4L3M3N2M3M3M3M4K4M3M3M3M3M4L3M3M3M3K5I7I8H7I7H8N2O1O1O2O0O1O1O1O100O2N1O1O100O1O101N1O3N2M3M3N2M3M3N2M3N2N2N2N2N2N2N2N2M3N2N2N3mNoGQFS8g9UHWFm7b9YH]Fi7[9_HcFc7V9cHhF`7V9bHeFc7X9e1N2M6K4fKlD_OY;;iDE\\;3gDL^;NdD2`;HbD7d;B^D>f;\\O\\Dc0j;VOXDi0m;QOUDo0o;jNTDU1R<]MWEc2S>O100O11O000000001O0000001O00001O001O00001O000^GoKU1R4eNTLZ1l3bNZL\\1g3^N_La1a3ZNfLd1Z3XNkLg1V3SNQMk1o2QNVMn1k2lM\\MR2d2iMbMV2^2fMhMX2Y2bMmM]2S2_MSNZ1hMYJV4W4YNZ1eM_JR4R4cNU1`MiJm3o3jNP1]MQKj3j3QOm0YMYKf3f3ZOh0TMbKb3b3Bd0PMjK_3]3I`0mLSLZ3>hKY1Y4m1iL\\LW3YOlLY2\\3j1eLdLg4]1mNf1aLlLb4Z1UOb1]LTM_4V1[O]1[L]M[4Q1A^1ULaM[4l0G_1oKeMZ4h0N`1hKhM[4h0Oo0TLYNn3h00`0^LhNb3h026dLRO[3h03KjL]OT3h04_OQMIl2h05TOWM4d2i06iN]M?^2h07^NcMj0W2h0`4XO`Ki0_4WObKh0]4YOdKg0[4YOfKg0Y4XOhKi0V4WOlKi0S4VOoKj0P4UORLk0m3TOTLm0j3SOXLm0g3RO\\Ll0f3QO\\Lo0e3^N`E7n6Z1c3]NaE6n6]1a3]NbE3P7_1_3]NbE1Q7b1^3[NcE0R7d1\\3[NnLe1h:N101O001O1O001O0O2O001O001O001O0O101O00000O2O0O10001N10000O101O0O10000O0100O010O010O10O01O10O0100O010O10O010O01001N2O001N2O001N2O001N2O0O2N2N1O2N2N1O1O00000001O1O001O001O001O1O001O001O001O001O1O00J6G9HjVb2"}, "label": "man riding a bike wearing a rainbow colored hat"}]}
{"id": 41459, "image": "COCO_train2014_000000041459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the multi colored beanie riding the green bike\"."}], "task": "refering", "answer_template": "The \"the man in the multi colored beanie riding the green bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 90, 91, 92, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 197, 198, 199, 206, 207, 208, 209, 210, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 277, 278, 279, 294, 295, 296, 311, 312, 313, 328, 329, 330, 345, 346, 347, 362, 363, 364, 365, 379, 380, 381, 382], "bbox": [0.15012526096033402, 0.182015625, 0.7236116910229645, 0.986515625], "iscrowd": 0, "area": 49593.447300000014, "rle": {"size": [640, 479], "counts": "YZ]13]13j`04k^O8n`0Ng^O=Ra0^1I8H7I4L3M3N2M3M3M4L3M3N2M3M3M3M4K4M3M3M3M3M4L3M3M3M3K5I7I8H7I7H8N2O1O1O2O0O1O1O1O100O2N1O1O100O1O101N1O3N2M3M3N2M3M3N2M3N2N2N2N2N2N2N2N2M3N2N2N3mNoGQFS8g9UHWFm7b9YH]Fi7[9_HcFc7V9cHhF`7V9bHeFc7X9e1N2M6K4fKlD_OY;;iDE\\;3gDL^;NdD2`;HbD7d;B^D>f;\\O\\Dc0j;VOXDi0m;QOUDo0o;jNTDU1R<]MWEc2S>O100O11O000000001O0000001O00001O001O00001O000^GoKU1R4eNTLZ1l3bNZL\\1g3^N_La1a3ZNfLd1Z3XNkLg1V3SNQMk1o2QNVMn1k2lM\\MR2d2iMbMV2^2fMhMX2Y2bMmM]2S2_MSNZ1hMYJV4W4YNZ1eM_JR4R4cNU1`MiJm3o3jNP1]MQKj3j3QOm0YMYKf3f3ZOh0TMbKb3b3Bd0PMjK_3]3I`0mLSLZ3>hKY1Y4m1iL\\LW3YOlLY2\\3j1eLdLg4]1mNf1aLlLb4Z1UOb1]LTM_4V1[O]1[L]M[4Q1A^1ULaM[4l0G_1oKeMZ4h0N`1hKhM[4h0Oo0TLYNn3h00`0^LhNb3h026dLRO[3h03KjL]OT3h04_OQMIl2h05TOWM4d2i06iN]M?^2h07^NcMj0W2h0`4XO`Ki0_4WObKh0]4YOdKg0[4YOfKg0Y4XOhKi0V4WOlKi0S4VOoKj0P4UORLk0m3TOTLm0j3SOXLm0g3RO\\Ll0f3QO\\Lo0e3^N`E7n6Z1c3]NaE6n6]1a3]NbE3P7_1_3]NbE1Q7b1^3[NcE0R7d1\\3[NnLe1h:N101O001O1O001O0O2O001O001O001O0O101O00000O2O0O10001N10000O101O0O10000O0100O010O010O10O01O10O0100O010O10O010O01001N2O001N2O001N2O001N2O0O2N2N1O2N2N1O1O00000001O1O001O001O001O1O001O001O001O001O1O00J6G9HjVb2"}, "label": "the man in the multi colored beanie riding the green bike"}]}
{"id": 41459, "image": "COCO_train2014_000000041459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bike with a green basket attached to the front\"."}], "task": "refering", "answer_template": "The \"a green bike with a green basket attached to the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 294, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 308, 309, 310, 311, 312, 313, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 337, 338, 340, 341, 342, 343, 344, 345, 346, 347, 348, 357, 358, 359, 360, 361, 362, 363, 364, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.014384133611691023, 0.459453125, 0.9586221294363257, 1.0], "iscrowd": 0, "area": 66018.21935, "rle": {"size": [640, 479], "counts": "eo4;`c05K5K5L4L4M3M3M3M3M3M3M3M3N2M3N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O100O1O1O1O100O1O1O1O1O10O01O1O1O100O1O1O1O100O10000O10000O10000O10000O100O10000O10000000000O100000000000000000001N100000000000000000000O1000000000000001O0000001O000000001O0000O1O1O100O1nNeK^B\\4a=fK]B[4b=gK\\BZ4d=gKZBZ4f=gKXBZ4g=iKVBX4k=jKRBV4P>kKlAV4U>kKhAV4Z>kKbAV4d>gKXAZ4P?75K3M3M7H6J5J6G9G9A?H8H<D<D<ROWY<YOThC2M3gDOb22YM3g2LUM:j2FQM`0n2@mLf0R3ZOjLj0U3WOhLm0W3TOdLP1\\3POaLS1_3mN]LX1b3iNZLZ1e3gNWL]1f3fNWLZ1k3hNYHCa0d1[7iNRHFb0^1^7lNPHH`0X1e7POiGK`0U1g7POiGM?R1i7QOgGO?o0j7SOfG1LG_MV1o:ROeG8XM[OQ29UNQ1n:SOcGd0TOYO]No0n:SO]FISOQ1n0UOfNm0m:UOXFKVOR1j0QOmNl0o:UOPFNYOU1=TO^Ob0o:VOgE4^Ob1\\OQOc01n:WOcE7_Oh23gMm:ZO]E;Bb27hMl:ZOmDk0NR2<fMk:]OfDP12k1a0eMi:^OaDW14e1f0_Mj:DXD^16]1b=UMSBc1:V1f=WMlAh1;Q1j=ZOUBd0n=]OPBb0Q>_OnA?S>DkA<U>EjA=T>i2M3O1N101N101N100O10000O10000000O100O10000O100N2O1O1O1O2N1O1O1O1O1O1107H8H5K4L4L1O10O01O001O1O0010N2N1O2N2N2N1O2N2N2N1O1O2N1O1O2N1O2N2NlA[JP>7RBU5b>K4LN2N2N2M2O2K4M30000000001O00001O00001XOYJVCf5j<ZJVCg5i<ZJVCf5j<ZJVCf5j<[JVCe5i<[JWCe5i<\\JUCf5k<YJUCg5k<ZJSCg5m<YJSCh5l<YJRCh5n<\\JlBe5U=_JfBa5[=`02O10001O000000000000000000001O0000000000000000001O001O1O010O1O001O001N2O001O001O1N2O1O1O1O0O2O1O1O1[JgAa5\\>000O10O100002N4L3M3M4L2N00010O01O02O1N2O1N1O2N2N2N2N2N2O1N2N000000O1000001O000O100N2O1N2O2M2OL41O0O2O1O1N3L3M3M4L4M3_LRBTNHY2Y>\\OZBSNB]2X>ZOhBa0\\=ZOiBb0^=UOhBg0^=QOiBj0]=nNiBn0[=lNkBQ1Y=_NWCZ1o<bNUCX1Wb<"}, "label": "a green bike with a green basket attached to the front"}]}
{"id": 41459, "image": "COCO_train2014_000000041459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green bike and cart\"."}], "task": "refering", "answer_template": "The \"green bike and cart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 294, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 308, 309, 310, 311, 312, 313, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 337, 338, 340, 341, 342, 343, 344, 345, 346, 347, 348, 357, 358, 359, 360, 361, 362, 363, 364, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.014384133611691023, 0.459453125, 0.9586221294363257, 1.0], "iscrowd": 0, "area": 66018.21935, "rle": {"size": [640, 479], "counts": "eo4;`c05K5K5L4L4M3M3M3M3M3M3M3M3N2M3N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O100O1O1O1O100O1O1O1O1O10O01O1O1O100O1O1O1O100O10000O10000O10000O10000O100O10000O10000000000O100000000000000000001N100000000000000000000O1000000000000001O0000001O000000001O0000O1O1O100O1nNeK^B\\4a=fK]B[4b=gK\\BZ4d=gKZBZ4f=gKXBZ4g=iKVBX4k=jKRBV4P>kKlAV4U>kKhAV4Z>kKbAV4d>gKXAZ4P?75K3M3M7H6J5J6G9G9A?H8H<D<D<ROWY<YOThC2M3gDOb22YM3g2LUM:j2FQM`0n2@mLf0R3ZOjLj0U3WOhLm0W3TOdLP1\\3POaLS1_3mN]LX1b3iNZLZ1e3gNWL]1f3fNWLZ1k3hNYHCa0d1[7iNRHFb0^1^7lNPHH`0X1e7POiGK`0U1g7POiGM?R1i7QOgGO?o0j7SOfG1LG_MV1o:ROeG8XM[OQ29UNQ1n:SOcGd0TOYO]No0n:SO]FISOQ1n0UOfNm0m:UOXFKVOR1j0QOmNl0o:UOPFNYOU1=TO^Ob0o:VOgE4^Ob1\\OQOc01n:WOcE7_Oh23gMm:ZO]E;Bb27hMl:ZOmDk0NR2<fMk:]OfDP12k1a0eMi:^OaDW14e1f0_Mj:DXD^16]1b=UMSBc1:V1f=WMlAh1;Q1j=ZOUBd0n=]OPBb0Q>_OnA?S>DkA<U>EjA=T>i2M3O1N101N101N100O10000O10000000O100O10000O100N2O1O1O1O2N1O1O1O1O1O1107H8H5K4L4L1O10O01O001O1O0010N2N1O2N2N2N1O2N2N2N1O1O2N1O1O2N1O2N2NlA[JP>7RBU5b>K4LN2N2N2M2O2K4M30000000001O00001O00001XOYJVCf5j<ZJVCg5i<ZJVCf5j<ZJVCf5j<[JVCe5i<[JWCe5i<\\JUCf5k<YJUCg5k<ZJSCg5m<YJSCh5l<YJRCh5n<\\JlBe5U=_JfBa5[=`02O10001O000000000000000000001O0000000000000000001O001O1O010O1O001O001N2O001O001O1N2O1O1O1O0O2O1O1O1[JgAa5\\>000O10O100002N4L3M3M4L2N00010O01O02O1N2O1N1O2N2N2N2N2N2O1N2N000000O1000001O000O100N2O1N2O2M2OL41O0O2O1O1N3L3M3M4L4M3_LRBTNHY2Y>\\OZBSNB]2X>ZOhBa0\\=ZOiBb0^=UOhBg0^=QOiBj0]=nNiBn0[=lNkBQ1Y=_NWCZ1o<bNUCX1Wb<"}, "label": "green bike and cart"}]}
{"id": 303608, "image": "COCO_train2014_000000303608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus in the middle of the crowd\"."}], "task": "refering", "answer_template": "The \"bus in the middle of the crowd\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 157, 158, 159, 160, 161, 162, 164, 165, 166, 170, 172, 173], "bbox": [0.10732732732732733, 0.52062, 0.8950150150150151, 0.8025], "iscrowd": 0, "area": 28485.060750000008, "rle": {"size": [500, 333], "counts": "ema07k>b0H8K5J3L6K5L5J6K5K5L6I;F5M2N3M3N3M4M5J5L6I5L3M000O0100DeD^L\\;^3nD[LS;_3c001N2O100O10000000O1000001RDcLc;^3[DcLe;^3XDdLh;d300001O0GXDfLh;T3_DkLa;U3?00O10000000000001O00002N5KO1M3O1O1O1O1000000O11O00001O2jCjLj;a3O001O1O2N6J5K1O1O1O1O0000000000000000001O1O1O001O00000000000000O100O1N2N2N2N2N2N2M3M3M3N2M3O100O1000000001O001O1O1O1SDdL`;\\3\\DjLb;d3000N2H8O1O1GUDgLm;X38O1O1O1O1O100O1O1001O000000N3L3O100O1O100000000000000001O000000001O0000001O00001O003M1O00001O1O3M5K2N1O0000N2O1O10000O1O1N2O1O1O11O0000001O001O1O001O002N3M4L00000000000000O1O1O1O10000000001O000101N01N10001O0O101N1O1O2L3L5K5Ki0fMbBb0iWa0"}, "label": "bus in the middle of the crowd"}]}
{"id": 303608, "image": "COCO_train2014_000000303608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus\"."}], "task": "refering", "answer_template": "The \"bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 157, 158, 159, 160, 161, 162, 164, 165, 166, 170, 172, 173], "bbox": [0.10732732732732733, 0.52062, 0.8950150150150151, 0.8025], "iscrowd": 0, "area": 28485.060750000008, "rle": {"size": [500, 333], "counts": "ema07k>b0H8K5J3L6K5L5J6K5K5L6I;F5M2N3M3N3M4M5J5L6I5L3M000O0100DeD^L\\;^3nD[LS;_3c001N2O100O10000000O1000001RDcLc;^3[DcLe;^3XDdLh;d300001O0GXDfLh;T3_DkLa;U3?00O10000000000001O00002N5KO1M3O1O1O1O1000000O11O00001O2jCjLj;a3O001O1O2N6J5K1O1O1O1O0000000000000000001O1O1O001O00000000000000O100O1N2N2N2N2N2N2M3M3M3N2M3O100O1000000001O001O1O1O1SDdL`;\\3\\DjLb;d3000N2H8O1O1GUDgLm;X38O1O1O1O1O100O1O1001O000000N3L3O100O1O100000000000000001O000000001O0000001O00001O003M1O00001O1O3M5K2N1O0000N2O1O10000O1O1N2O1O1O11O0000001O001O1O001O002N3M4L00000000000000O1O1O1O10000000001O000101N01N10001O0O101N1O1O2L3L5K5Ki0fMbBb0iWa0"}, "label": "bus"}]}
{"id": 270844, "image": "COCO_train2014_000000270844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue jacket to the right of the man\"."}], "task": "refering", "answer_template": "The \"the blue jacket to the right of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 203, 219, 220, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 303, 304, 305, 321, 322], "bbox": [0.8463541666666666, 0.45493750000000005, 1.0, 0.82521875], "iscrowd": 0, "area": 10283.91835, "rle": {"size": [640, 480], "counts": "dVn72jc05J6J6J5L5J6J6J6K5J6J6J5L5J6J6J6J6J6K5J5N3M3N2M3M3M3N1N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3N101O1O0O2O00000O2O00000O2O00000O2O0L4K5J7J7I8L3M3L4M3MK5G9fNZ1[LbJ"}, "label": "the blue jacket to the right of the man"}]}
{"id": 270844, "image": "COCO_train2014_000000270844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark green coat on a chair beside a smiling man\"."}], "task": "refering", "answer_template": "The \"dark green coat on a chair beside a smiling man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 203, 219, 220, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 303, 304, 305, 321, 322], "bbox": [0.8463541666666666, 0.45493750000000005, 1.0, 0.82521875], "iscrowd": 0, "area": 10283.91835, "rle": {"size": [640, 480], "counts": "dVn72jc05J6J6J5L5J6J6J6K5J6J6J5L5J6J6J6J6J6K5J5N3M3N2M3M3M3N1N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3N101O1O0O2O00000O2O00000O2O00000O2O0L4K5J7J7I8L3M3L4M3MK5G9fNZ1[LbJ"}, "label": "dark green coat on a chair beside a smiling man"}]}
{"id": 270844, "image": "COCO_train2014_000000270844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man smiling at a restaurant\"."}], "task": "refering", "answer_template": "The \"a man smiling at a restaurant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 104, 105, 106, 107, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.005958333333333333, 0.00021875000000000003, 1.0, 0.9867968749999999], "iscrowd": 0, "area": 194870.6347, "rle": {"size": [640, 480], "counts": "_U2Y:g90O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1RKWEd0j:TO`Ej0`:oNiEn0X:kNRFR1n9fN\\FW1e9bNdF[1]9]NnF`1R9YNWGd1j8UN_Gi1a8oMjGm1W8lMRHR2n7fM\\HW2e7bMeHZ2\\7_MmH_2S7YMWId2j6UM`Ih2`6PMjIm2W6nLPJo2Q6QMRJl2n5TMTJi2m5WMUJg2k5YMWJd2j5\\MYJ`2h5`MZJ^2f5bM\\J[2e5eM^JX2b5hM`JU2a5kMaJS2_5mMdJo1]5QNeJl1\\5TNfJj1Z5VNgJh1Z5XNgJg1Y5YNiJd1X5\\NiJb1X5^NiJa1W5_NjJ_1W5aNjJ^1V5bNlJ[1U5eNlJZ1T5fNmJX1T5iNlJU1U5kNlJT1T5lNnJQ1S5oNnJP1R5POoJn0R5ROoJl0R5TOoJk0Q5UOQKh0P5XOQKg0o4YORKe0o4[ORKc0o4]ORKb0n4^OTK?m4ATK>l4BUK<l4DUK;k4EVK9k4GVK7k4IWK5i4KXK3i4MXK2h4NYK0h40YKNh42ZKLf44[KJf46[KIe47b7000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000`DFZ3:cLJ\\36bLM]33`L0`30]L4b3L\\L7c3IZL;e3EXL>h3BVLa0i3@SLd0l3^OoKf0P4]OkKe0U4]OfKg0Y4[ObKi0]4YO_Ki0`4ZO[Kj0d4XOWKl0g4WOUKl0i4WORKl0n4WOlJm0R5VOjJm0U5UOfJo0X5TOcJo0\\5TO`Jo0_5SO\\JQ1b5ROYJR1e5QOUJS1k5oNmIX1Q6lNeI\\1Y6gN_I`1`6bNXId1g6_NPIi1n6ZNjHm1U7UNcHR2[7QN\\HZ2_7iMYH`2d7cMSHg2h7\\MPHm2l7VMkGT3R8nLfG\\3U8gLcGb3Y8aL^Gi3_8YLYGQ4b8RLVGW4f8lKQG_4k8^3L4L5L3L5K4M3L5K4M4K4L4M4K4L5L3L4L5L3L5K4M3L5K4M2M2O1N2N2N1N3N2M3M3M3M2N3N2M3M3M3M2O2M3M3M3M3N1N3M3M3M3M3N1N3M3M3M3N2N12O001O1O1O0O2O1O1O001O1N2O001O1O1O0O2O1O1O001N2O1O001O1O1N1000O1O100O1O100O1O100O2N100O1O1O100O1O100O1O1O1N2N2N2N2N3M2N1O2N2N2N2N1O2N2N2N110e0[O5KO1O001O1O1N101O1O1O1O0O2O1O1O001N2M3N2M2N3N2M3N2M2O2M3M3N1N3N2M3M:Gk0TO]^O"}, "label": "a man smiling at a restaurant"}]}
{"id": 270844, "image": "COCO_train2014_000000270844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man smiling holding a macys box in front of him\"."}], "task": "refering", "answer_template": "The \"a man smiling holding a macys box in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 104, 105, 106, 107, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.005958333333333333, 0.00021875000000000003, 1.0, 0.9867968749999999], "iscrowd": 0, "area": 194870.6347, "rle": {"size": [640, 480], "counts": "_U2Y:g90O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1RKWEd0j:TO`Ej0`:oNiEn0X:kNRFR1n9fN\\FW1e9bNdF[1]9]NnF`1R9YNWGd1j8UN_Gi1a8oMjGm1W8lMRHR2n7fM\\HW2e7bMeHZ2\\7_MmH_2S7YMWId2j6UM`Ih2`6PMjIm2W6nLPJo2Q6QMRJl2n5TMTJi2m5WMUJg2k5YMWJd2j5\\MYJ`2h5`MZJ^2f5bM\\J[2e5eM^JX2b5hM`JU2a5kMaJS2_5mMdJo1]5QNeJl1\\5TNfJj1Z5VNgJh1Z5XNgJg1Y5YNiJd1X5\\NiJb1X5^NiJa1W5_NjJ_1W5aNjJ^1V5bNlJ[1U5eNlJZ1T5fNmJX1T5iNlJU1U5kNlJT1T5lNnJQ1S5oNnJP1R5POoJn0R5ROoJl0R5TOoJk0Q5UOQKh0P5XOQKg0o4YORKe0o4[ORKc0o4]ORKb0n4^OTK?m4ATK>l4BUK<l4DUK;k4EVK9k4GVK7k4IWK5i4KXK3i4MXK2h4NYK0h40YKNh42ZKLf44[KJf46[KIe47b7000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000`DFZ3:cLJ\\36bLM]33`L0`30]L4b3L\\L7c3IZL;e3EXL>h3BVLa0i3@SLd0l3^OoKf0P4]OkKe0U4]OfKg0Y4[ObKi0]4YO_Ki0`4ZO[Kj0d4XOWKl0g4WOUKl0i4WORKl0n4WOlJm0R5VOjJm0U5UOfJo0X5TOcJo0\\5TO`Jo0_5SO\\JQ1b5ROYJR1e5QOUJS1k5oNmIX1Q6lNeI\\1Y6gN_I`1`6bNXId1g6_NPIi1n6ZNjHm1U7UNcHR2[7QN\\HZ2_7iMYH`2d7cMSHg2h7\\MPHm2l7VMkGT3R8nLfG\\3U8gLcGb3Y8aL^Gi3_8YLYGQ4b8RLVGW4f8lKQG_4k8^3L4L5L3L5K4M3L5K4M4K4L4M4K4L5L3L4L5L3L5K4M3L5K4M2M2O1N2N2N1N3N2M3M3M3M2N3N2M3M3M3M2O2M3M3M3M3N1N3M3M3M3M3N1N3M3M3M3N2N12O001O1O1O0O2O1O1O001O1N2O001O1O1O0O2O1O1O001N2O1O001O1O1N1000O1O100O1O100O1O100O2N100O1O1O100O1O100O1O1O1N2N2N2N2N3M2N1O2N2N2N2N1O2N2N2N110e0[O5KO1O001O1O1N101O1O1O1O0O2O1O1O001N2M3N2M2N3N2M3N2M2O2M3M3N1N3N2M3M:Gk0TO]^O"}, "label": "a man smiling holding a macys box in front of him"}]}
{"id": 401917, "image": "COCO_train2014_000000401917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white and black umbrella\"."}], "task": "refering", "answer_template": "The \"the white and black umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 25, 26, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126], "bbox": [0.0909367681498829, 0.0465, 0.8218266978922716, 0.37234375], "iscrowd": 0, "area": 46636.690599999994, "rle": {"size": [640, 427], "counts": "Rah0`0^c05L4K5K5K5K5L3L5L4K5L4L4K3N2M3N2M3N1N3N2N2N2M3N1O2M3N2N2N1N2O001O1N2O1O1O0O2O1O1N2O1O1O0O2O1O1O1N2O1O001N2O1O1N2O1O011N1O2N1O2O0O2N1O2N101N1O2N1O101N1O2N1O2N101N1O2N1O2O0O2N1O2O0O2O0O1O100O100O1O100O100O100O1O100O10O01O100O100O1O100O100O1O100O010O1O01000000O01000O1000000O1000O01000000O10000O0100000O10000O10O1000O10000O1000O10O10000O100000O010000O1000000O10O10O1000000O10000O2O00000O2O000O101O000O10001N1000001N10000O101O000O101O0O10001O0O10000O2O00000O2O000O10001O0O10001N1000001N1O1O1O1O2N1N2O1O2N1O1O1O1O2M2O1O1O2N1O1O1O2N1N2O1O1O2N1O1O1O2M2O1O1O2N1O1O1O1N3N1O1O1O2N1O1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O1O2N1O1N2O2N1O1O1O1OW^_1"}, "label": "the white and black umbrella"}]}
{"id": 401917, "image": "COCO_train2014_000000401917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white checkered umbrella being held up over someone ' s head\"."}], "task": "refering", "answer_template": "The \"a black and white checkered umbrella being held up over someone ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 25, 26, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126], "bbox": [0.0909367681498829, 0.0465, 0.8218266978922716, 0.37234375], "iscrowd": 0, "area": 46636.690599999994, "rle": {"size": [640, 427], "counts": "Rah0`0^c05L4K5K5K5K5L3L5L4K5L4L4K3N2M3N2M3N1N3N2N2N2M3N1O2M3N2N2N1N2O001O1N2O1O1O0O2O1O1N2O1O1O0O2O1O1O1N2O1O001N2O1O1N2O1O011N1O2N1O2O0O2N1O2N101N1O2N1O101N1O2N1O2N101N1O2N1O2O0O2N1O2O0O2O0O1O100O100O1O100O100O100O1O100O10O01O100O100O1O100O100O1O100O010O1O01000000O01000O1000000O1000O01000000O10000O0100000O10000O10O1000O10000O1000O10O10000O100000O010000O1000000O10O10O1000000O10000O2O00000O2O000O101O000O10001N1000001N10000O101O000O101O0O10001O0O10000O2O00000O2O000O10001O0O10001N1000001N1O1O1O1O2N1N2O1O2N1O1O1O1O2M2O1O1O2N1O1O1O2N1N2O1O1O2N1O1O1O2M2O1O1O2N1O1O1O1N3N1O1O1O2N1O1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O1O2N1O1N2O2N1O1O1O1OW^_1"}, "label": "a black and white checkered umbrella being held up over someone ' s head"}]}
{"id": 442878, "image": "COCO_train2014_000000442878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right side sofa is in front of tv\"."}], "task": "refering", "answer_template": "The \"right side sofa is in front of tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233, 251, 269], "bbox": [0.5764600000000001, 0.6480186480186481, 1.0, 0.9970629370629371], "iscrowd": 0, "area": 13584.246550000002, "rle": {"size": [429, 500], "counts": "anh31Z=2N2N2N2O2M2N6J3M2N2N2N200O100000000000000000000O10001O00000000O1O1000000001O000000001O01O000001O000000001O01O0001O00001O001O00010O00001O001O000010O0001O00001N101N100eDgNa:[1VEoNh:Q1QEWOm:f1O1N100O101O01O00001O0001O01O0000001O01O0001O000010O000001O0000010O0000000001O0000000000001O00000000001O0001O00000001O0000000000001O00000000001O0000000001O01O00000000001O0000000000001O00000000001O01O000004L5K5K6J5K5K6J5K5L5J5K1O000001N10000000001O000UOgG"}, "label": "right side sofa is in front of tv"}]}
{"id": 442878, "image": "COCO_train2014_000000442878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch cushion\"."}], "task": "refering", "answer_template": "The \"a brown couch cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233, 251, 269], "bbox": [0.5764600000000001, 0.6480186480186481, 1.0, 0.9970629370629371], "iscrowd": 0, "area": 13584.246550000002, "rle": {"size": [429, 500], "counts": "anh31Z=2N2N2N2O2M2N6J3M2N2N2N200O100000000000000000000O10001O00000000O1O1000000001O000000001O01O000001O000000001O01O0001O00001O001O00010O00001O001O000010O0001O00001N101N100eDgNa:[1VEoNh:Q1QEWOm:f1O1N100O101O01O00001O0001O01O0000001O01O0001O000010O000001O0000010O0000000001O0000000000001O00000000001O0001O00000001O0000000000001O00000000001O0000000001O01O00000000001O0000000000001O00000000001O01O000004L5K5K6J5K5K6J5K5L5J5K1O000001N10000000001O000UOgG"}, "label": "a brown couch cushion"}]}
{"id": 139775, "image": "COCO_train2014_000000139775.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child is on a skateboard\"."}], "task": "refering", "answer_template": "The \"a child is on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 38, 40, 41, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 99, 100, 101, 114, 115, 116, 117, 129, 130, 131, 132, 144, 145, 146, 147, 159, 160, 161, 162, 174, 175, 176, 177, 189, 190, 191, 192, 204, 205, 206, 207, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 252], "bbox": [0.48276995305164316, 0.07571875, 0.8710328638497653, 0.747625], "iscrowd": 0, "area": 34162.22445, "rle": {"size": [640, 426], "counts": "jkP4`0Tc0=N1O2N2N1N3N2N1O2N2N1O2N2N1O2M3N1O2N2N1O2N2N1O2N2M5L4L4L1O1O11O10O01O1O10O01O1O1O10O01O1O1O001O1O1O00100O1O04M3M3M3L3hE\\Mo2f2hLaMX3a2_LeMb3]2ULiMl3Y2kKnMU4T2bKRN_4P2\\GTMS2R1b6l1QGYMW2R1h6f1gFaMZ2o0P7b1\\FfM^2o0V7]1QFmMb2l0QNnM`8i4YI^MRNoMe8b4UIdMPNoMl8]4nHiMQNoMR9W4iHnMPNPNX9R4cHoN^7P1^HUOb7k0cGO^8i50001O01O000001O0001O0N2M3N2N2N2M4M2N2M3N2N1O2M2O_NbFaH_9^7gFYH]9g7hFQH[9o7kFhGX9X8mF`GW9_8oFXGT9g8R1jHTEZ5l:^J`E^5`:[JjEe5S:SJYFl5o;O00N2O0O2N2N2O1N2N1lEPIW7R7gHRIU7P7iHTIT7m6jHWIR7k6lHXIQ7i6mH\\Io6f6oH^In6c6PIaIl6a6RIbIk6`6SIdIi6^6TIgIi6Y6VIjIg6X6WIlIe6h6gH\\IU7j6dH[IY7i6bHZI[7l6_HXI]7n6]HVI_7o6\\HUI`7Q7YHSIe7Q7VHSIf7g9L3M4M3L4L3M4M3L4L3M4L4K4iLfBC^=1TCEP=NbCIb<KPDK[`0G:F9G:F9G:EdnQ1"}, "label": "a child is on a skateboard"}]}
{"id": 139775, "image": "COCO_train2014_000000139775.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child on the skating board with white shirt\"."}], "task": "refering", "answer_template": "The \"the child on the skating board with white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 38, 40, 41, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 99, 100, 101, 114, 115, 116, 117, 129, 130, 131, 132, 144, 145, 146, 147, 159, 160, 161, 162, 174, 175, 176, 177, 189, 190, 191, 192, 204, 205, 206, 207, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 252], "bbox": [0.48276995305164316, 0.07571875, 0.8710328638497653, 0.747625], "iscrowd": 0, "area": 34162.22445, "rle": {"size": [640, 426], "counts": "jkP4`0Tc0=N1O2N2N1N3N2N1O2N2N1O2N2N1O2M3N1O2N2N1O2N2N1O2N2M5L4L4L1O1O11O10O01O1O10O01O1O1O10O01O1O1O001O1O1O00100O1O04M3M3M3L3hE\\Mo2f2hLaMX3a2_LeMb3]2ULiMl3Y2kKnMU4T2bKRN_4P2\\GTMS2R1b6l1QGYMW2R1h6f1gFaMZ2o0P7b1\\FfM^2o0V7]1QFmMb2l0QNnM`8i4YI^MRNoMe8b4UIdMPNoMl8]4nHiMQNoMR9W4iHnMPNPNX9R4cHoN^7P1^HUOb7k0cGO^8i50001O01O000001O0001O0N2M3N2N2N2M4M2N2M3N2N1O2M2O_NbFaH_9^7gFYH]9g7hFQH[9o7kFhGX9X8mF`GW9_8oFXGT9g8R1jHTEZ5l:^J`E^5`:[JjEe5S:SJYFl5o;O00N2O0O2N2N2O1N2N1lEPIW7R7gHRIU7P7iHTIT7m6jHWIR7k6lHXIQ7i6mH\\Io6f6oH^In6c6PIaIl6a6RIbIk6`6SIdIi6^6TIgIi6Y6VIjIg6X6WIlIe6h6gH\\IU7j6dH[IY7i6bHZI[7l6_HXI]7n6]HVI_7o6\\HUI`7Q7YHSIe7Q7VHSIf7g9L3M4M3L4L3M4M3L4L3M4L4K4iLfBC^=1TCEP=NbCIb<KPDK[`0G:F9G:F9G:EdnQ1"}, "label": "the child on the skating board with white shirt"}]}
{"id": 311808, "image": "COCO_train2014_000000311808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in gray playing wii\"."}], "task": "refering", "answer_template": "The \"a man in gray playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 370, 371, 372, 373, 374], "bbox": [0.103125, 0.3914375, 0.46557812500000006, 1.0], "iscrowd": 0, "area": 30622.629500000006, "rle": {"size": [480, 640], "counts": "X[o02l>5J5L5J5I8L3M4K4M4L3M4L3M4K4M4L3M4L3M4K4M4K3N1N2N2O1O100O100O10O10O100O10000O100O10000O100QNhLVHX3h7lLUHU3h7gLaF3f1V3P8iLnGX3j7fLeF3_1X3l7eLeF4]1X3m7eLfF4\\1W3m7fLhF3Y1X3o7eLhF4W1X3P8eLiF3V1Y3Q8dLjF3S1Z3R8dLkF3Q1Z3S8dLlF3P1Y3T8dLmF2n0[3T8dLnF2m0Z3U8dLoF2j0[3W8dLnF1k0[3V8eLPG0h0\\3X8dLPG0h0\\3W8eLRGNf0^3X8eLQGMg0^3X8eLQGMf0_3X8eLQGNf0]3Z8dLPGOe0^3[8dLnFOg0]3[8dLnFOg0]3\\8cLlF1g0]3]8bLlF1g0]3]8cLkFOh0_3]8bLmFLg0b3]8aLkFLi0d3\\8aLjFKj0d3\\8aLiFKk0d3]8aLgFKk0f3_8^LfFLh0i3b8[LhFIc0P4e8XLhFG`0T4h8ULkFE:X4k8SLZGR4f8nKmF_4S9aKmF_4R9bKnF^4R9bKnF^4Q9i000O1O1N2N2N2N2N2O1N2N2O1N2O1N2O1O1NO2L2O1M3G:100O1001O001N101O001O001O001O001O0010O01O0010O01OWO^JVHa5h7dJUH]5i7gJkGLJ\\5Y8mJeGO0S5Z8gNgGPNW8o1nGPNP8R2QHlMl7X2THgMj7\\2WHbMh7`2XHnLGQOP8R4YHiL1nNf7Z4ZHcL:lN\\7b4ZH\\Ld8d3\\GTLk8m3UGmKQ9S4QGkKn8V4UGgKk8Y4WGeKi8\\4YG`Kg8a4ZG^Kf8c4YG]Kf8d4[G\\Kd8e4[G[Ke8e4\\GZKd8g4ZGZKf8f4ZGZKf8f4YG[Kg8f4XGZKh8f4\\GVKd8j4]GUKc8k4]GUKc8j4^GVKb8j4^GVKb8i4_GWKa8i4_GXK`8h4`GWKa8h4aGWK_8i4aGVK`8j4l0O0O10001O001N101O0O2O1O2N101N2N1O3eLYEX2k:aMYE]2k:\\MYEc2m:TMWEl2`;N8H2N2OO01O100O1O1O100O1O100O1O100O1O100O2O0O2O1N101J6H7J7_Oa0J6I6M4L4LjPP5"}, "label": "a man in gray playing wii"}]}
{"id": 311808, "image": "COCO_train2014_000000311808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the white control\"."}], "task": "refering", "answer_template": "The \"the man with the white control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 370, 371, 372, 373, 374], "bbox": [0.103125, 0.3914375, 0.46557812500000006, 1.0], "iscrowd": 0, "area": 30622.629500000006, "rle": {"size": [480, 640], "counts": "X[o02l>5J5L5J5I8L3M4K4M4L3M4L3M4K4M4L3M4L3M4K4M4K3N1N2N2O1O100O100O10O10O100O10000O100O10000O100QNhLVHX3h7lLUHU3h7gLaF3f1V3P8iLnGX3j7fLeF3_1X3l7eLeF4]1X3m7eLfF4\\1W3m7fLhF3Y1X3o7eLhF4W1X3P8eLiF3V1Y3Q8dLjF3S1Z3R8dLkF3Q1Z3S8dLlF3P1Y3T8dLmF2n0[3T8dLnF2m0Z3U8dLoF2j0[3W8dLnF1k0[3V8eLPG0h0\\3X8dLPG0h0\\3W8eLRGNf0^3X8eLQGMg0^3X8eLQGMf0_3X8eLQGNf0]3Z8dLPGOe0^3[8dLnFOg0]3[8dLnFOg0]3\\8cLlF1g0]3]8bLlF1g0]3]8cLkFOh0_3]8bLmFLg0b3]8aLkFLi0d3\\8aLjFKj0d3\\8aLiFKk0d3]8aLgFKk0f3_8^LfFLh0i3b8[LhFIc0P4e8XLhFG`0T4h8ULkFE:X4k8SLZGR4f8nKmF_4S9aKmF_4R9bKnF^4R9bKnF^4Q9i000O1O1N2N2N2N2N2O1N2N2O1N2O1N2O1O1NO2L2O1M3G:100O1001O001N101O001O001O001O001O0010O01O0010O01OWO^JVHa5h7dJUH]5i7gJkGLJ\\5Y8mJeGO0S5Z8gNgGPNW8o1nGPNP8R2QHlMl7X2THgMj7\\2WHbMh7`2XHnLGQOP8R4YHiL1nNf7Z4ZHcL:lN\\7b4ZH\\Ld8d3\\GTLk8m3UGmKQ9S4QGkKn8V4UGgKk8Y4WGeKi8\\4YG`Kg8a4ZG^Kf8c4YG]Kf8d4[G\\Kd8e4[G[Ke8e4\\GZKd8g4ZGZKf8f4ZGZKf8f4YG[Kg8f4XGZKh8f4\\GVKd8j4]GUKc8k4]GUKc8j4^GVKb8j4^GVKb8i4_GWKa8i4_GXK`8h4`GWKa8h4aGWK_8i4aGVK`8j4l0O0O10001O001N101O0O2O1O2N101N2N1O3eLYEX2k:aMYE]2k:\\MYEc2m:TMWEl2`;N8H2N2OO01O100O1O1O100O1O100O1O100O1O100O2O0O2O1N101J6H7J7_Oa0J6I6M4L4LjPP5"}, "label": "the man with the white control"}]}
{"id": 311808, "image": "COCO_train2014_000000311808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting and watching video game\"."}], "task": "refering", "answer_template": "The \"a man sitting and watching video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 202, 203, 204, 205, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.714609375, 0.43820833333333337, 1.0, 0.9865208333333333], "iscrowd": 0, "area": 32281.761199999994, "rle": {"size": [480, 640], "counts": "ief61n>1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2J6K5J6J6J6K5J6lLgM^H4hN[2g8gM[H:`NT2R9hMXHV3e7PMUHS3i7RMQHQ3l7UMnGn2o7XMkGk2R8\\MgGg2V8_MdGd2Z8aM`Gb2]8R2K5L4L4UObIoHc6P7g0O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100000000000000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1YKWHU2j7gMYHY2h7dM[H[2f7aM]H_2e7\\M_Hc2b7ZMaHe2a7VMbHj2_7SMdHl2]7PMfHP3\\7kLhHT3Y7iLiHW3Y7cLlH\\3U7_LPI`3R7YLSIg3n6TLWIk3k6nKZIR4g6hK_IW4e801O1O001O001O001O1O001O001O1O001N101O001O1O001O001O1O6J6J6JT2lM]B"}, "label": "a man sitting and watching video game"}]}
{"id": 311808, "image": "COCO_train2014_000000311808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown t - shirt with dark curly hair\"."}], "task": "refering", "answer_template": "The \"a man in a brown t - shirt with dark curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 202, 203, 204, 205, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.714609375, 0.43820833333333337, 1.0, 0.9865208333333333], "iscrowd": 0, "area": 32281.761199999994, "rle": {"size": [480, 640], "counts": "ief61n>1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2J6K5J6J6J6K5J6lLgM^H4hN[2g8gM[H:`NT2R9hMXHV3e7PMUHS3i7RMQHQ3l7UMnGn2o7XMkGk2R8\\MgGg2V8_MdGd2Z8aM`Gb2]8R2K5L4L4UObIoHc6P7g0O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100000000000000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1YKWHU2j7gMYHY2h7dM[H[2f7aM]H_2e7\\M_Hc2b7ZMaHe2a7VMbHj2_7SMdHl2]7PMfHP3\\7kLhHT3Y7iLiHW3Y7cLlH\\3U7_LPI`3R7YLSIg3n6TLWIk3k6nKZIR4g6hK_IW4e801O1O001O001O001O1O001O001O1O001N101O001O1O001O001O1O6J6J6JT2lM]B"}, "label": "a man in a brown t - shirt with dark curly hair"}]}
{"id": 311808, "image": "COCO_train2014_000000311808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with pink sweater and glasses sitting in the middle\"."}], "task": "refering", "answer_template": "The \"woman with pink sweater and glasses sitting in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 291, 307, 309, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.348875, 0.3910208333333333, 0.67078125, 1.0], "iscrowd": 0, "area": 34439.840500000006, "rle": {"size": [480, 640], "counts": "VlX31m>4L3M3L4M3M3N2O1N2O1VEVO\\7j0aHYO^7h0`HZO`7f0`HZO`7f0_H[Oa7f0]H[Ob7f0]H[Oc7e0]H[Oc7e0\\H\\Od7d0[H]Od7d0[H]Oe7c0ZH^Of7a0[H_Oe7a0ZH@f7?ZHBf7>YHCg7<ZHDg7;XHFh79XHHh78WHIi77WHIi76WHKi75VHLk73THNl72THNl72SHOm71RH0o7OPH2P8NPH2Q8MnG4R8LmG5S8KlG6U8IkG7U8IjG8W8GhG:X8FgG;Z8DfG<Z8DeG=[8CdG>]8AbG`0^8AaG?`8B]G?c8B[G?f8AXG`0h8AWG?j8ATG`0l8ARG`0o8@oFa0Q9@mFa0T9@jF`0V9AhF`0Y9@eFa0[9@cFa0]9@bF`0^9A`F`0`9A^F`0b9A\\F`0d9@\\F`0d9@[Fa0e9_OZFb0e9_OZFb0e9_O[Fa0e9_OZFb0e9_OZFb0f9^OYFc0f9^OZFb0e9_OZFb0f9^OYFc0c9A\\F`0^9FbF:W9MhF4Q93nFNl88SGIk6XNjIV2[OBP6ZOXJZ1G]Ol5FnIT15WOh51fIn0a0ROe5;\\Ij0o0kN`5d3_J]L\\5h3cJYL[5i3dJXL\\5h3dJXL\\5h3cJYL\\5h3bJZL^5f3aJ[L^5f3aJ[L_5e3`J\\L`5d3_J]L`5e3^J\\Lb5l3UJULk5S4kIoKT6_601N10000O10000O100O10000O10000O100O10000O100000000000000001O0000000000001O0000000000001O0000001O001O001O00001O001O001O001O001O001O001O<D=C=C>B=C8H2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N4L7I7I7I7H9F9H8H8H8G9F;E:G9F:FW`R3"}, "label": "woman with pink sweater and glasses sitting in the middle"}]}
{"id": 311808, "image": "COCO_train2014_000000311808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a pink shirt playing nintendo wii\"."}], "task": "refering", "answer_template": "The \"a woman in a pink shirt playing nintendo wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 291, 307, 309, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.348875, 0.3910208333333333, 0.67078125, 1.0], "iscrowd": 0, "area": 34439.840500000006, "rle": {"size": [480, 640], "counts": "VlX31m>4L3M3L4M3M3N2O1N2O1VEVO\\7j0aHYO^7h0`HZO`7f0`HZO`7f0_H[Oa7f0]H[Ob7f0]H[Oc7e0]H[Oc7e0\\H\\Od7d0[H]Od7d0[H]Oe7c0ZH^Of7a0[H_Oe7a0ZH@f7?ZHBf7>YHCg7<ZHDg7;XHFh79XHHh78WHIi77WHIi76WHKi75VHLk73THNl72THNl72SHOm71RH0o7OPH2P8NPH2Q8MnG4R8LmG5S8KlG6U8IkG7U8IjG8W8GhG:X8FgG;Z8DfG<Z8DeG=[8CdG>]8AbG`0^8AaG?`8B]G?c8B[G?f8AXG`0h8AWG?j8ATG`0l8ARG`0o8@oFa0Q9@mFa0T9@jF`0V9AhF`0Y9@eFa0[9@cFa0]9@bF`0^9A`F`0`9A^F`0b9A\\F`0d9@\\F`0d9@[Fa0e9_OZFb0e9_OZFb0e9_O[Fa0e9_OZFb0e9_OZFb0f9^OYFc0f9^OZFb0e9_OZFb0f9^OYFc0c9A\\F`0^9FbF:W9MhF4Q93nFNl88SGIk6XNjIV2[OBP6ZOXJZ1G]Ol5FnIT15WOh51fIn0a0ROe5;\\Ij0o0kN`5d3_J]L\\5h3cJYL[5i3dJXL\\5h3dJXL\\5h3cJYL\\5h3bJZL^5f3aJ[L^5f3aJ[L_5e3`J\\L`5d3_J]L`5e3^J\\Lb5l3UJULk5S4kIoKT6_601N10000O10000O100O10000O10000O100O10000O100000000000000001O0000000000001O0000000000001O0000001O001O001O00001O001O001O001O001O001O001O<D=C=C>B=C8H2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N4L7I7I7I7H9F9H8H8H8G9F;E:G9F:FW`R3"}, "label": "a woman in a pink shirt playing nintendo wii"}]}
{"id": 573961, "image": "COCO_train2014_000000573961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep\"."}], "task": "refering", "answer_template": "The \"sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 285, 286], "bbox": [0.0298125, 0.04418181818181818, 0.8336250000000002, 0.8739220779220779], "iscrowd": 0, "area": 90065.6077, "rle": {"size": [385, 640], "counts": "[\\73i;6K6I7J5J7J5J7J6L3M4L3N3L4L3M3M3N1N2N3M2O1N2N3M2N2O2M2N2N3N1N2N2N3M2O1N3M2O1O2N1O1O1O2O0O1O2N1O1O2O0O1O1O2N1O1O2O0O1O2N1O1O101N1O1O2N1O1O2O0O1O1O2N1O101N1O1O2N1O1O2O0O1O1O2N1O101N1O101N1000001O0O101O1O0O2O001O0O2O001O001N2O001O0O2O001O0O2O1O001N101O001O0O2O1O001N101O001N101O001N2O001O001N101O001N2O001O00000000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000001O0000jM[KPMe4l2fKmLY4P3nKmLR4R3PLmLP4Q3SLnLn3P3TLoLl3P3VLoLj3o2YLPMg3o2ZLQMf3n2\\LQMd3n2^LQMc3l2`LSM`3l2`LUM`3j2aLVM_3h2cLXM]3g2dLYM]3e2cL]M\\3a2fL_MZ3`2gL`MY3_2hLaMX3^2hLcMX3[2jLeMW3Y2jLgMV3X2kLiMT3U2mLlMS3S2nLmMR3R2PMmMP3R2QMnMP3o1RMQNn2n1SMSNl2l1UMTNk2j1XMUNh2j1YMVNg2i1ZMWNg2g1ZMYNf2e1\\M[Nd2d1]M]Nb2b1`M]N`2a1bM_N_2_1bMaN^2_1cM`N]2_1eM`N[2`1eM`N[2`1fM`NY2_1hMaNY2^1hMaNX2^1jMaNV2_1jMaNV2^1lMaNT2_1lMaNT2^1nMbNR2]1nMcNR2\\1PNcNP2]1QNbNo1]1SNbNm1^1UN`Nl1_1VN_Nj1`1YN^Ng1b1[N\\Ne1c1^N[Nb1e1`NYN`1f1dNWN]1h1fNUNZ1k1hNSNX1m1kNPNU1P2mNnMS1R2POkMP1U2SOhMn0W2TOgMl0Y2WOdMi0\\2YO`Mi0`2Z43M2M3N2N2M3N2M3N2N2M3N2O2M2N3N1N3M3N1N3M2O2M2N3N2M2O2M2N3N1N3M3N1N3M5L4L4L4M3M3L4M2M3N1O1N2O1O1N2O1N2O2N1N2O1N2O1O0O100O10O10O1000O0100O01000O10O010000O10000O100O10000O100O10000O10000000000000000O1000000000000000000O10000000000000000O2O00000000000001O0000001O01O000001O0000001O000000001O01O0001O000000001O0000001O01OO10001O1N2O001O0O2O1O001N101O1O0O2O001O0O2O1O001N101L4K4L5K5J7J6J6J6J5K6J6H8B>N2N1YMgGd1[8VNkGg1X8SNmGl1U8mMQHR2Q8hMTHV2m8O1O1O1N2O1O1N3N1O1O1N2O1N2N3L3N3L4M2N3L3M4H7F;H\\SX1"}, "label": "sheep"}]}
{"id": 573961, "image": "COCO_train2014_000000573961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep with lamb\"."}], "task": "refering", "answer_template": "The \"sheep with lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 285, 286], "bbox": [0.0298125, 0.04418181818181818, 0.8336250000000002, 0.8739220779220779], "iscrowd": 0, "area": 90065.6077, "rle": {"size": [385, 640], "counts": "[\\73i;6K6I7J5J7J5J7J6L3M4L3N3L4L3M3M3N1N2N3M2O1N2N3M2N2O2M2N2N3N1N2N2N3M2O1N3M2O1O2N1O1O1O2O0O1O2N1O1O2O0O1O1O2N1O1O2O0O1O2N1O1O101N1O1O2N1O1O2O0O1O1O2N1O101N1O1O2N1O1O2O0O1O1O2N1O101N1O101N1000001O0O101O1O0O2O001O0O2O001O001N2O001O0O2O001O0O2O1O001N101O001O0O2O1O001N101O001N101O001N2O001O001N101O001N2O001O00000000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000001O0000jM[KPMe4l2fKmLY4P3nKmLR4R3PLmLP4Q3SLnLn3P3TLoLl3P3VLoLj3o2YLPMg3o2ZLQMf3n2\\LQMd3n2^LQMc3l2`LSM`3l2`LUM`3j2aLVM_3h2cLXM]3g2dLYM]3e2cL]M\\3a2fL_MZ3`2gL`MY3_2hLaMX3^2hLcMX3[2jLeMW3Y2jLgMV3X2kLiMT3U2mLlMS3S2nLmMR3R2PMmMP3R2QMnMP3o1RMQNn2n1SMSNl2l1UMTNk2j1XMUNh2j1YMVNg2i1ZMWNg2g1ZMYNf2e1\\M[Nd2d1]M]Nb2b1`M]N`2a1bM_N_2_1bMaN^2_1cM`N]2_1eM`N[2`1eM`N[2`1fM`NY2_1hMaNY2^1hMaNX2^1jMaNV2_1jMaNV2^1lMaNT2_1lMaNT2^1nMbNR2]1nMcNR2\\1PNcNP2]1QNbNo1]1SNbNm1^1UN`Nl1_1VN_Nj1`1YN^Ng1b1[N\\Ne1c1^N[Nb1e1`NYN`1f1dNWN]1h1fNUNZ1k1hNSNX1m1kNPNU1P2mNnMS1R2POkMP1U2SOhMn0W2TOgMl0Y2WOdMi0\\2YO`Mi0`2Z43M2M3N2N2M3N2M3N2N2M3N2O2M2N3N1N3M3N1N3M2O2M2N3N2M2O2M2N3N1N3M3N1N3M5L4L4L4M3M3L4M2M3N1O1N2O1O1N2O1N2O2N1N2O1N2O1O0O100O10O10O1000O0100O01000O10O010000O10000O100O10000O100O10000O10000000000000000O1000000000000000000O10000000000000000O2O00000000000001O0000001O01O000001O0000001O000000001O01O0001O000000001O0000001O01OO10001O1N2O001O0O2O1O001N101O1O0O2O001O0O2O1O001N101L4K4L5K5J7J6J6J6J5K6J6H8B>N2N1YMgGd1[8VNkGg1X8SNmGl1U8mMQHR2Q8hMTHV2m8O1O1O1N2O1O1N3N1O1O1N2O1N2N3L3N3L4M2N3L3M4H7F;H\\SX1"}, "label": "sheep with lamb"}]}
{"id": 573961, "image": "COCO_train2014_000000573961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby lamb\"."}], "task": "refering", "answer_template": "The \"a baby lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 198, 199, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 319], "bbox": [0.366109375, 0.5053766233766234, 0.9159218750000001, 0.9419480519480519], "iscrowd": 0, "area": 33963.7784, "rle": {"size": [385, 640], "counts": "XWh2d0\\;2M2O2N2M2O2N2M2O2N2M2O2N1O1N101O0O2O1O0O2O1O1N2O1O1N2O1O1N2O1O1N2O1O2O0O2N101N100O2N101N101N1O2O0O2O0O3M4M2M4M2M4M2N3M2N2N3M2N3M2N3M2N2]HlKU7W4eHmKZ7U4bHmK^7\\4O1O1O1O1O1O0010O010O00010O010O001O001O001O010O001O00001O001O001O2N3M3M3N2M3M3M3M2N2N1O1O1O1O2N1O1O100O2N1O1O1O1O1OO10000O10000O2O000O1000000O10000O10000O2O000O1000000000O100O1O100O100O100O1O2O0O100O1O100O100O100N2O1O1O2N1N2O1O1O1O1N2004L3M3M3M00O2N1O1O1O1O100O1O1O1O100O10QG\\Ma8c2^G_Mc8_2]GbMc8^2[GcMf8\\2YGfMg8Z2XGgMh8X2WGkMh8U2VGmMj8R2VGoMj8Q2TGRNk8m1TGUNl8`201O000010O000001O00001O000O2O00001O000O2O00001O000O2O0000000O100001O00001O0O101O00001O00001O00001O00001O00001O00001O0O2O00001O000O2O00001N10001O000O2O00001N10001O0O101O01O01O001O1O1O1O1O1O2N1O1O1O1O1001O1O1O1O1O00M4M2M3N2N3O1N2O1N2O1N2O1N2O2M2O1N2N2O1N2N2M3N3L3M3N2MYoc0"}, "label": "a baby lamb"}]}
{"id": 573961, "image": "COCO_train2014_000000573961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby white sheep\"."}], "task": "refering", "answer_template": "The \"a baby white sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 198, 199, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 319], "bbox": [0.366109375, 0.5053766233766234, 0.9159218750000001, 0.9419480519480519], "iscrowd": 0, "area": 33963.7784, "rle": {"size": [385, 640], "counts": "XWh2d0\\;2M2O2N2M2O2N2M2O2N2M2O2N1O1N101O0O2O1O0O2O1O1N2O1O1N2O1O1N2O1O1N2O1O2O0O2N101N100O2N101N101N1O2O0O2O0O3M4M2M4M2M4M2N3M2N2N3M2N3M2N3M2N2]HlKU7W4eHmKZ7U4bHmK^7\\4O1O1O1O1O1O0010O010O00010O010O001O001O001O010O001O00001O001O001O2N3M3M3N2M3M3M3M2N2N1O1O1O1O2N1O1O100O2N1O1O1O1O1OO10000O10000O2O000O1000000O10000O10000O2O000O1000000000O100O1O100O100O100O1O2O0O100O1O100O100O100N2O1O1O2N1N2O1O1O1O1N2004L3M3M3M00O2N1O1O1O1O100O1O1O1O100O10QG\\Ma8c2^G_Mc8_2]GbMc8^2[GcMf8\\2YGfMg8Z2XGgMh8X2WGkMh8U2VGmMj8R2VGoMj8Q2TGRNk8m1TGUNl8`201O000010O000001O00001O000O2O00001O000O2O00001O000O2O0000000O100001O00001O0O101O00001O00001O00001O00001O00001O00001O0O2O00001O000O2O00001N10001O000O2O00001N10001O0O101O01O01O001O1O1O1O1O1O2N1O1O1O1O1001O1O1O1O1O00M4M2M3N2N3O1N2O1N2O1N2O1N2O2M2O1N2N2O1N2N2M3N3L3M3N2MYoc0"}, "label": "a baby white sheep"}]}
{"id": 524820, "image": "COCO_train2014_000000524820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white urinal with a blue stopper\"."}], "task": "refering", "answer_template": "The \"a white urinal with a blue stopper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 182, 183, 184, 185, 186, 187, 188, 201, 202, 203, 204, 205], "bbox": [0.0015, 0.0018618618618618619, 0.5148400000000001, 0.9748948948948949], "iscrowd": 0, "area": 66633.35859999999, "rle": {"size": [333, 500], "counts": "j;i0d9R1nNn0QOm0TO1O2M3N2N1O2M3N1O2M3N1O2M3N2N1O2M3N1O2M3N2N1N3N2N1O2M3N2N1N3N2N1O2M3N2N1N3N2N1N3N2N1O2M3N2N1N3N2N1N3N2N2N1N3N2N1N3N2N2N1N3N2N1N3N2N2M2O2N2N1N3N2N1N3N2N2M2O2N2N1N3N2N2M2O2N27G0100O010O010O1O001O001O001O1O001O001O1O001O00001O00001O000O2OO10O10000O1@iFJV96nFGR99QGCP9=SG@m8?a0000O10O10000O10O1000O10000O01000O1000000O01000O100000O10000000000O1000O100000O10000000000O1000000000000O10000O10000O100O100O10000O10O010000O100O10000O100O10000O100O10000O100O10000O100O1O1N2O1O1N2O2N1N2O2N1N2O1N3L3M3M4K4M3M3M4L3L4M4L3J6I7J7I6J6D>mMS2mM\\R_2"}, "label": "a white urinal with a blue stopper"}]}
{"id": 524820, "image": "COCO_train2014_000000524820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a men ' s urinal with a urinal cake next to a trash can and toilet\"."}], "task": "refering", "answer_template": "The \"a men ' s urinal with a urinal cake next to a trash can and toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 182, 183, 184, 185, 186, 187, 188, 201, 202, 203, 204, 205], "bbox": [0.0015, 0.0018618618618618619, 0.5148400000000001, 0.9748948948948949], "iscrowd": 0, "area": 66633.35859999999, "rle": {"size": [333, 500], "counts": "j;i0d9R1nNn0QOm0TO1O2M3N2N1O2M3N1O2M3N1O2M3N2N1O2M3N1O2M3N2N1N3N2N1O2M3N2N1N3N2N1O2M3N2N1N3N2N1N3N2N1O2M3N2N1N3N2N1N3N2N2N1N3N2N1N3N2N2N1N3N2N1N3N2N2M2O2N2N1N3N2N1N3N2N2M2O2N2N1N3N2N2M2O2N27G0100O010O010O1O001O001O001O1O001O001O1O001O00001O00001O000O2OO10O10000O1@iFJV96nFGR99QGCP9=SG@m8?a0000O10O10000O10O1000O10000O01000O1000000O01000O100000O10000000000O1000O100000O10000000000O1000000000000O10000O10000O100O100O10000O10O010000O100O10000O100O10000O100O10000O100O10000O100O1O1N2O1O1N2O2N1N2O2N1N2O1N3L3M3M4K4M3M3M4L3L4M4L3J6I7J7I6J6D>mMS2mM\\R_2"}, "label": "a men ' s urinal with a urinal cake next to a trash can and toilet"}]}
{"id": 328214, "image": "COCO_train2014_000000328214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of salad with tongs\"."}], "task": "refering", "answer_template": "The \"a bowl of salad with tongs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 255, 256, 257, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 323, 324, 325, 340], "bbox": [0.0, 0.6125, 0.20183333333333334, 0.901359375], "iscrowd": 0, "area": 11932.829349999996, "rle": {"size": [640, 480], "counts": "X<h5X>01O001O01O01O0O100O2N1O1O2N100O1O2N1O1O2O0O1O1O2N1O101N1O1O1O2N100O2N1O100O2N1O2N101N1O2N2N101N1O2N1O2O0O2N1O2N101N2N2N1O2N2M2N3M3M3M2N3M3N2M3M3N2M3M3M3N2M4L3N2M3M3N2M3M3N3K4M3L5L5J6K4K6J6K5Jdm^7"}, "label": "a bowl of salad with tongs"}]}
{"id": 328214, "image": "COCO_train2014_000000328214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of salad with tongs in it\"."}], "task": "refering", "answer_template": "The \"a bowl of salad with tongs in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 255, 256, 257, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 323, 324, 325, 340], "bbox": [0.0, 0.6125, 0.20183333333333334, 0.901359375], "iscrowd": 0, "area": 11932.829349999996, "rle": {"size": [640, 480], "counts": "X<h5X>01O001O01O01O0O100O2N1O1O2N100O1O2N1O1O2O0O1O1O2N1O101N1O1O1O2N100O2N1O100O2N1O2N101N1O2N2N101N1O2N1O2O0O2N1O2N101N2N2N1O2N2M2N3M3M3M2N3M3N2M3M3N2M3M3M3N2M4L3N2M3M3N2M3M3N3K4M3L5L5J6K4K6J6K5Jdm^7"}, "label": "a bowl of salad with tongs in it"}]}
{"id": 328214, "image": "COCO_train2014_000000328214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of green with purle leafy greens in it\"."}], "task": "refering", "answer_template": "The \"a bowl of green with purle leafy greens in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 257, 258, 259, 275], "bbox": [0.0, 0.524015625, 0.3266458333333333, 0.709640625], "iscrowd": 0, "area": 10207.512450000004, "rle": {"size": [640, 480], "counts": "a:f1Zb02N001O00001O0O101O00001O0000001O00001O00001O00001O0O101O00001O00001O00001O0010O1O010O10O010O01000O0100O01000O01000O010O10O10O10O01000O01000O010O011O0O10000O101N1O1O100O1O2O0O1O5K4M1N0000O2O00001N1000001N1000001N1000001N10001N100O101N1O100O2N100O2O0O1O101N1O1O1O2N1N3N1O2N1O2N1N3N1O2N1O2M2O2M3M2O2M2N3N1N3M2N3I6IS`Y6"}, "label": "a bowl of green with purle leafy greens in it"}]}
{"id": 328214, "image": "COCO_train2014_000000328214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bowl filled with green and purple salad greens\"."}], "task": "refering", "answer_template": "The \"a white bowl filled with green and purple salad greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 257, 258, 259, 275], "bbox": [0.0, 0.524015625, 0.3266458333333333, 0.709640625], "iscrowd": 0, "area": 10207.512450000004, "rle": {"size": [640, 480], "counts": "a:f1Zb02N001O00001O0O101O00001O0000001O00001O00001O00001O0O101O00001O00001O00001O0010O1O010O10O010O01000O0100O01000O01000O010O10O10O10O01000O01000O010O011O0O10000O101N1O1O100O1O2O0O1O5K4M1N0000O2O00001N1000001N1000001N1000001N10001N100O101N1O100O2N100O2O0O1O101N1O1O1O2N1N3N1O2N1O2N1N3N1O2N1O2M2O2M3M2O2M2N3N1N3M2N3I6IS`Y6"}, "label": "a white bowl filled with green and purple salad greens"}]}
{"id": 541212, "image": "COCO_train2014_000000541212.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella the little boy is holding\"."}], "task": "refering", "answer_template": "The \"the umbrella the little boy is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 145, 146, 147, 148, 169, 170, 171], "bbox": [0.33401562500000004, 0.07568749999999999, 0.5878593750000001, 0.3333125], "iscrowd": 0, "area": 13077.173399999998, "rle": {"size": [640, 640], "counts": "gmU43jc04K4M4L4L3L5L4L4K4M4L4L3L5L4L4K4M4L4K5L3O2N2N1O2N2N2N1O2N2N1O2M3N2N1O2N2N2N1O1O1O1O1O0O2O1O1O1O001O1O1O1O001O1N2O1O001O1O1O1O001O1O1O1O010000000O010000000O01hM`@0`?Oe@M[?3h@JX?6l@FS?:QABP?>TA^Ol>a0UB_Nk=a1VB^Ni=b1YB]Ng=c1ZB[Ng=d1[B[Ne=e1\\BZNd=e1^BZNa=h1^BXNb=i1^BUNc=l1]BQNe=P2[BmMg=U2XBiMh=Y2o11O1O2N1N2O1O1O2N1N2O1O00N2N101O1O100O001O1O100O001O100O001N2M3L4_O`0I8J6K5K5O010O1O100O00100O100O00100O1O10O01O100O1O102M4L3N3L3MTnT5"}, "label": "the umbrella the little boy is holding"}]}
{"id": 541212, "image": "COCO_train2014_000000541212.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella shade over the head of a little boy\"."}], "task": "refering", "answer_template": "The \"an umbrella shade over the head of a little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 145, 146, 147, 148, 169, 170, 171], "bbox": [0.33401562500000004, 0.07568749999999999, 0.5878593750000001, 0.3333125], "iscrowd": 0, "area": 13077.173399999998, "rle": {"size": [640, 640], "counts": "gmU43jc04K4M4L4L3L5L4L4K4M4L4L3L5L4L4K4M4L4K5L3O2N2N1O2N2N2N1O2N2N1O2M3N2N1O2N2N2N1O1O1O1O1O0O2O1O1O1O001O1O1O1O001O1N2O1O001O1O1O1O001O1O1O1O010000000O010000000O01hM`@0`?Oe@M[?3h@JX?6l@FS?:QABP?>TA^Ol>a0UB_Nk=a1VB^Ni=b1YB]Ng=c1ZB[Ng=d1[B[Ne=e1\\BZNd=e1^BZNa=h1^BXNb=i1^BUNc=l1]BQNe=P2[BmMg=U2XBiMh=Y2o11O1O2N1N2O1O1O2N1N2O1O00N2N101O1O100O001O1O100O001O100O001N2M3L4_O`0I8J6K5K5O010O1O100O00100O100O00100O1O10O01O100O1O102M4L3N3L3MTnT5"}, "label": "an umbrella shade over the head of a little boy"}]}
{"id": 426525, "image": "COCO_train2014_000000426525.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle most animal center\"."}], "task": "refering", "answer_template": "The \"the middle most animal center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 101, 116, 117, 120, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270], "bbox": [0.040203125, 0.28471698113207544, 0.8143593749999999, 0.7925707547169811], "iscrowd": 0, "area": 28667.4157, "rle": {"size": [424, 640], "counts": "[l:b0f<1O2N1O2O1N1O2N1O2O02O1O1N2O1N101N2O1N2O1N1O01O0O2O0O101N1O2O0O2O0O010O010O010O10O010O010O0100O0010O010O010O010O010O010OTDXOX;i0fDZOY;e0fD]OY;c0eD@[;`0cDB\\;>bDE];<aDG^;8`DK_;6_DLa;3]D0a;n0N1O1O2N2M3N2N2N2M3N1O1O1N2O1O1O1N3N1O000O100000001N1000000O10001O0O1000000O101O000O10000O101O000O1000000O2O000O1000000O2O00000O1000001O01O00001O0010O0001O001O001O0O101O001N10000O2O00000O10001O0O1000000O2O000O1000001N1000000O10001O0O10000O1000001N10O100O00100O1O010O1O_N`E;a:DdE8[:HiE5V:KnE2R:MRFOn92UFKj95XFJg96ZFJf95[FJe96]FI`9:`FG]9:dFFY9<iFCU9>lFCP9?QGAm8a0TG^Oj8c0WG^Oe8d0\\G\\Ob8e0`GZO]8h0dGYOY8i0gGWOW8j0kGUOT8k0mGVOQ8j0PHVOn7e0ZHZOe7a0aH@]7;iHEU76SIIl62ZIOd6KcI5[6GlI8S6CSJ>k5\\O\\Jd0b5XOeJg0n81O01O01O01O01O01O01O01O01O01N1N2M4L3M3N3L3M3M\\U90gjF1N101N101N101N1O2O0O2O0O2O0_NETF=g9HXF8h9HWF:h9FWF;i9EWF<h9DWF=i9CWF>h9CVF>j9BVF?i9AVF`0j9@VF`0j9@UFb0j9^OUFc0k9^OTFb0l9_ORFb0o9^OPFb0P:^OoEc0Q:^OnEc0Q:^OmEc0S:]OmEc0S:^OkEc0U:^OjEb0W:^OgEd0X:\\OgEe0Y:\\OfEd0Z:]OdEd0\\:]OcEc0]:]ObEd0^:]OaEd0_:\\O_Ee0a:[O_Ee0a:\\O]Ee0c:\\O[Ee0e:\\OZEd0f:\\OYEf0f:[OYEe0h:[OVEf0j:[OUEe0k:[OTEf0l:[OSEf0l:k0001O000001O000001O0001O0000000010O0000000001O000001O0001O0000000010O0000000001O0001O000001O000000001O00000000001O00000000001O000000001O000O1000001O00000000001O000000001O00000000001O0O10000010O00000O100O100O100O101N1O100O1G9G9G9N2O1O1O1N3N1O1N2O1O1O1N2O1O2N1N2O1O1O2M2O1O1O2N100O1O1O2O0O1O101N1O1O100O2N1O100O2No]a1"}, "label": "the middle most animal center"}]}
{"id": 426525, "image": "COCO_train2014_000000426525.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bull in the very center , top of the group of three\"."}], "task": "refering", "answer_template": "The \"the bull in the very center , top of the group of three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 101, 116, 117, 120, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270], "bbox": [0.040203125, 0.28471698113207544, 0.8143593749999999, 0.7925707547169811], "iscrowd": 0, "area": 28667.4157, "rle": {"size": [424, 640], "counts": "[l:b0f<1O2N1O2O1N1O2N1O2O02O1O1N2O1N101N2O1N2O1N1O01O0O2O0O101N1O2O0O2O0O010O010O010O10O010O010O0100O0010O010O010O010O010O010OTDXOX;i0fDZOY;e0fD]OY;c0eD@[;`0cDB\\;>bDE];<aDG^;8`DK_;6_DLa;3]D0a;n0N1O1O2N2M3N2N2N2M3N1O1O1N2O1O1O1N3N1O000O100000001N1000000O10001O0O1000000O101O000O10000O101O000O1000000O2O000O1000000O2O00000O1000001O01O00001O0010O0001O001O001O0O101O001N10000O2O00000O10001O0O1000000O2O000O1000001N1000000O10001O0O10000O1000001N10O100O00100O1O010O1O_N`E;a:DdE8[:HiE5V:KnE2R:MRFOn92UFKj95XFJg96ZFJf95[FJe96]FI`9:`FG]9:dFFY9<iFCU9>lFCP9?QGAm8a0TG^Oj8c0WG^Oe8d0\\G\\Ob8e0`GZO]8h0dGYOY8i0gGWOW8j0kGUOT8k0mGVOQ8j0PHVOn7e0ZHZOe7a0aH@]7;iHEU76SIIl62ZIOd6KcI5[6GlI8S6CSJ>k5\\O\\Jd0b5XOeJg0n81O01O01O01O01O01O01O01O01O01N1N2M4L3M3N3L3M3M\\U90gjF1N101N101N101N1O2O0O2O0O2O0_NETF=g9HXF8h9HWF:h9FWF;i9EWF<h9DWF=i9CWF>h9CVF>j9BVF?i9AVF`0j9@VF`0j9@UFb0j9^OUFc0k9^OTFb0l9_ORFb0o9^OPFb0P:^OoEc0Q:^OnEc0Q:^OmEc0S:]OmEc0S:^OkEc0U:^OjEb0W:^OgEd0X:\\OgEe0Y:\\OfEd0Z:]OdEd0\\:]OcEc0]:]ObEd0^:]OaEd0_:\\O_Ee0a:[O_Ee0a:\\O]Ee0c:\\O[Ee0e:\\OZEd0f:\\OYEf0f:[OYEe0h:[OVEf0j:[OUEe0k:[OTEf0l:[OSEf0l:k0001O000001O000001O0001O0000000010O0000000001O000001O0001O0000000010O0000000001O0001O000001O000000001O00000000001O00000000001O000000001O000O1000001O00000000001O000000001O00000000001O0O10000010O00000O100O100O100O101N1O100O1G9G9G9N2O1O1O1N3N1O1N2O1O1O1N2O1O2N1N2O1O1O2M2O1O1O2N100O1O1O2O0O1O101N1O1O100O2N1O100O2No]a1"}, "label": "the bull in the very center , top of the group of three"}]}
{"id": 426525, "image": "COCO_train2014_000000426525.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yak separate from the rest of the group\"."}], "task": "refering", "answer_template": "The \"a yak separate from the rest of the group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 64, 65, 82, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182], "bbox": [0.5740000000000001, 0.15264150943396226, 0.9309218749999999, 0.4965566037735849], "iscrowd": 0, "area": 15908.45055, "rle": {"size": [424, 640], "counts": "WQh41W=0O2O0O2O0O2O0000001O00003N1N2N3M4M2M2N3N2O1N12M2M2O1N2O1N2N2O1N100O10O0010O0000010O00010O0001O00001_DFR::fE3U:OgE5W:MfE5X:MeE60UOc9g0YF8JYOm9R2mETNR:a2O10001N10000O2O000O10001O000O100000000O10000000000O1000001O0O10010O000010O000001O01O000000001N1000001O0O100O1N3M2N2N2O101N10000OHZMPFf2n9;N1O2N2N1O2N2N2N2O1O002O0OoLfFW2Y9eMPGV2o8hMXGU2f8jM]GU2c8jM^GW2`8hMbGY2\\8gMfGX2Z8gMgGZ2W8fMjGZ2V8eMlG[2R8fMnG[2Q8dMPH]2n7cMTH\\2l7cMUH^2j7bMVH_2h7aMZH^2f7aM[H`2c7aM]H_2c7`M_H_2a7aM_H`2_7`MbH`2^7_MdH`2\\7`MdH`2[7`MfHa2Y7_MhH_2Y7`MhH`2X7`MhH`2X7_MjH`2V7`MjH`2V7_MlH`2T7_MmHa2T7^MlHb2T7]MnHb2R7^MnHb2l8O2O001N101N101N101O000O2O001O01O01O001O1O1O1O1O1O1O101N1O1O1O1O1O1O1N2O1O001O1O1O1O1O1N101O1O1O1O1O1O2M2O2N2N8G4L4L9ESQb0"}, "label": "a yak separate from the rest of the group"}]}
{"id": 426525, "image": "COCO_train2014_000000426525.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bison facing the camera\"."}], "task": "refering", "answer_template": "The \"the bison facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 64, 65, 82, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182], "bbox": [0.5740000000000001, 0.15264150943396226, 0.9309218749999999, 0.4965566037735849], "iscrowd": 0, "area": 15908.45055, "rle": {"size": [424, 640], "counts": "WQh41W=0O2O0O2O0O2O0000001O00003N1N2N3M4M2M2N3N2O1N12M2M2O1N2O1N2N2O1N100O10O0010O0000010O00010O0001O00001_DFR::fE3U:OgE5W:MfE5X:MeE60UOc9g0YF8JYOm9R2mETNR:a2O10001N10000O2O000O10001O000O100000000O10000000000O1000001O0O10010O000010O000001O01O000000001N1000001O0O100O1N3M2N2N2O101N10000OHZMPFf2n9;N1O2N2N1O2N2N2N2O1O002O0OoLfFW2Y9eMPGV2o8hMXGU2f8jM]GU2c8jM^GW2`8hMbGY2\\8gMfGX2Z8gMgGZ2W8fMjGZ2V8eMlG[2R8fMnG[2Q8dMPH]2n7cMTH\\2l7cMUH^2j7bMVH_2h7aMZH^2f7aM[H`2c7aM]H_2c7`M_H_2a7aM_H`2_7`MbH`2^7_MdH`2\\7`MdH`2[7`MfHa2Y7_MhH_2Y7`MhH`2X7`MhH`2X7_MjH`2V7`MjH`2V7_MlH`2T7_MmHa2T7^MlHb2T7]MnHb2R7^MnHb2l8O2O001N101N101N101O000O2O001O01O01O001O1O1O1O1O1O1O101N1O1O1O1O1O1O1N2O1O001O1O1O1O1O1N101O1O1O1O1O1O2M2O2N2N8G4L4L9ESQb0"}, "label": "the bison facing the camera"}]}
{"id": 188956, "image": "COCO_train2014_000000188956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a light blue shirt and dark blue shorts\"."}], "task": "refering", "answer_template": "The \"a little boy in a light blue shirt and dark blue shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 315, 316, 317, 318], "bbox": [0.7045, 0.41122916666666665, 0.8612343750000001, 0.80675], "iscrowd": 0, "area": 10961.050050000005, "rle": {"size": [480, 640], "counts": "Udc6:g>7H100O100O1000O0100O100O100O01000SE[O]7f0`H]O`7c0]H@b7a0ZHCf7=WHFh7;UHHi79THKj77SHLj77SHLj77RHMl75QHNl75QHNm74oG0n72PH1n71oG2n73mG0Q8:cGI[8_3M3M3M2N3M3M3M3M3N20O2O0000000000000O1000000000001N101O001O00001N101O00001O0O2O0O1[K`Fj3a9PLfFm3[9mKlFR4U9gKSGW4m8dKYGZ4i9N2M2TN[E[Og:a0bEXO`:d0hERO]:k0kEgN_:U1iE^N_:_1d1L4N1O2N1O2M3N1O2N1O2M3Na^Y1"}, "label": "a little boy in a light blue shirt and dark blue shorts"}]}
{"id": 188956, "image": "COCO_train2014_000000188956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in int he light blue shirt\"."}], "task": "refering", "answer_template": "The \"the boy in int he light blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 315, 316, 317, 318], "bbox": [0.7045, 0.41122916666666665, 0.8612343750000001, 0.80675], "iscrowd": 0, "area": 10961.050050000005, "rle": {"size": [480, 640], "counts": "Udc6:g>7H100O100O1000O0100O100O100O01000SE[O]7f0`H]O`7c0]H@b7a0ZHCf7=WHFh7;UHHi79THKj77SHLj77SHLj77RHMl75QHNl75QHNm74oG0n72PH1n71oG2n73mG0Q8:cGI[8_3M3M3M2N3M3M3M3M3N20O2O0000000000000O1000000000001N101O001O00001N101O00001O0O2O0O1[K`Fj3a9PLfFm3[9mKlFR4U9gKSGW4m8dKYGZ4i9N2M2TN[E[Og:a0bEXO`:d0hERO]:k0kEgN_:U1iE^N_:_1d1L4N1O2N1O2M3N1O2N1O2M3Na^Y1"}, "label": "the boy in int he light blue shirt"}]}
{"id": 213536, "image": "COCO_train2014_000000213536.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird that has spread wings\"."}], "task": "refering", "answer_template": "The \"a bird that has spread wings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 226, 227, 228, 229, 242, 243], "bbox": [0.0032084309133489463, 0.561375, 0.3570257611241218, 0.739484375], "iscrowd": 0, "area": 8108.935900000003, "rle": {"size": [640, 427], "counts": "_o01mc0201O001O1O001O1O0Z]OIca08\\^OHda09Z^OIea07[^OIea08Z^OIea07Z^OKea06Z^OJfa07Y^OJfa06Y^OKga06X^OKga06X^OJha06W^OLha05W^OKia05V^OMia04V^OLja05U^OLja04U^OMka04T^OMka04T^OMka03T^ONma02R^OOma02R^ONna02Q^O0na01Q^OOoa01P^O1oa00P^O0Pb01n]O1Qb0Oo]O1Qb00m]O2Rb0Ol]O2Tb0Nl]O3Sb0Nk]O4Tb0Ll]O4Tb0m0O00001O001O001O001O001O001O00001O0Y^OnM`a0S2b^ORNVa0n1m^ORNPa0o1Q_OQNm`0o1V_OQNg`0P2[_OoMc`0Q2`_OnM^`0S2c_OmM[`0S2f_OnMY`0Q2h_OPNV`0Q2j_OPNT`0P2m_OQNQ`0P2o_OQNo?o1R@QNm?P2S@oMm?Q2T@mMm?T2S@kMm?U2T@iMm?X2S@gMm?Y2T@fMl?[2__OfMTa0Z2j^OiMUa0W2i^OmMUa0`2O1O1OCl^OoMSa0R2m^OnMRa0R2o^OoMo`0R2R_OmMm`0S2T_OmMk`0S2V_OmMi`0S2X_OmMg`0R2[_OnMLDc`0^2b_OoMICe`0^2c_OoMGCf`0]2f_OPNZ`0P2g_OoMY`0Q2h_OnMX`0R2i_OmMW`0R2k_OnMT`0R2m_OmMS`0S2o_OkMQ`0T2Q@kMo?U2R@jMn?V2S@iMm?W2T@hMm?V2U@iMk?W2W@gMi?Y2o0N101N2N2O0O2N2O0O2N2O0O2N2N2O0O2N2O0O2N2O0O2N2O0O2N2O1N1O4M4L4M3M3M2N3M2N2N2N1O2N2Nk_[5"}, "label": "a bird that has spread wings"}]}
{"id": 213536, "image": "COCO_train2014_000000213536.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bird with wings fluttering on the edge of a landing\"."}], "task": "refering", "answer_template": "The \"a black bird with wings fluttering on the edge of a landing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 226, 227, 228, 229, 242, 243], "bbox": [0.0032084309133489463, 0.561375, 0.3570257611241218, 0.739484375], "iscrowd": 0, "area": 8108.935900000003, "rle": {"size": [640, 427], "counts": "_o01mc0201O001O1O001O1O0Z]OIca08\\^OHda09Z^OIea07[^OIea08Z^OIea07Z^OKea06Z^OJfa07Y^OJfa06Y^OKga06X^OKga06X^OJha06W^OLha05W^OKia05V^OMia04V^OLja05U^OLja04U^OMka04T^OMka04T^OMka03T^ONma02R^OOma02R^ONna02Q^O0na01Q^OOoa01P^O1oa00P^O0Pb01n]O1Qb0Oo]O1Qb00m]O2Rb0Ol]O2Tb0Nl]O3Sb0Nk]O4Tb0Ll]O4Tb0m0O00001O001O001O001O001O001O00001O0Y^OnM`a0S2b^ORNVa0n1m^ORNPa0o1Q_OQNm`0o1V_OQNg`0P2[_OoMc`0Q2`_OnM^`0S2c_OmM[`0S2f_OnMY`0Q2h_OPNV`0Q2j_OPNT`0P2m_OQNQ`0P2o_OQNo?o1R@QNm?P2S@oMm?Q2T@mMm?T2S@kMm?U2T@iMm?X2S@gMm?Y2T@fMl?[2__OfMTa0Z2j^OiMUa0W2i^OmMUa0`2O1O1OCl^OoMSa0R2m^OnMRa0R2o^OoMo`0R2R_OmMm`0S2T_OmMk`0S2V_OmMi`0S2X_OmMg`0R2[_OnMLDc`0^2b_OoMICe`0^2c_OoMGCf`0]2f_OPNZ`0P2g_OoMY`0Q2h_OnMX`0R2i_OmMW`0R2k_OnMT`0R2m_OmMS`0S2o_OkMQ`0T2Q@kMo?U2R@jMn?V2S@iMm?W2T@hMm?V2U@iMk?W2W@gMi?Y2o0N101N2N2O0O2N2O0O2N2O0O2N2N2O0O2N2O0O2N2O0O2N2O0O2N2O1N1O4M4L4M3M3M2N3M2N2N2N1O2N2Nk_[5"}, "label": "a black bird with wings fluttering on the edge of a landing"}]}
{"id": 557602, "image": "COCO_train2014_000000557602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra gazing on taller pile of hay\"."}], "task": "refering", "answer_template": "The \"zebra gazing on taller pile of hay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331], "bbox": [0.09567187499999999, 0.14401273885350319, 0.46515625, 0.6653503184713375], "iscrowd": 0, "area": 57370.982899999995, "rle": {"size": [628, 640], "counts": "YgU1>``0h2K5L4K5K5L4K5K4L5L4K5K5K5L4K4L5Kj0WOb0_OO01O10O0100O00100O010O00100O00100O010^CWIZ<h6fCZIZ<e6eC^IY<b6fCaIY<]6gCeIY<Z6gChIW<X6hCkIW<S6iCoIW<P6hCSJV<m5iCVJV<h5jC[JU<b6O2O1O0O2O1O0O2O1O0O2O1O001N2O001N2O001N2O001N10000O1000O010000000000000O10001O000000000O100000000000000O101O00000000000O1000000000O1O1O101N1O1O2O0O1O2N1O101N1O1O2O0O2N1O1O2N1O1O2N1O1O2N1O2N1O1OoM\\GUIb8l6R21O1N2O1N2O1N2O1N2O1N2O1N2O1N2O0O2O1N2O1N200001O001O00YFSHh7l7RHZHo7e7jGbHV8^7dGhH]8W7[GQIe8o6mF_IT9`6_FmIb9R8001O010O001O1O001N101N1O2O1N1O2O0O2N2O2M4L4M2M4L4M3L4L4M3L4L4M3L4L4K5J5K5J6K5K5J5L5K5K4K6K5K4K6J6G9G8H9G9G8G:G9Goaa6"}, "label": "zebra gazing on taller pile of hay"}]}
{"id": 557602, "image": "COCO_train2014_000000557602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a giraffe eating the brown grass\"."}], "task": "refering", "answer_template": "The \"there is a giraffe eating the brown grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331], "bbox": [0.09567187499999999, 0.14401273885350319, 0.46515625, 0.6653503184713375], "iscrowd": 0, "area": 57370.982899999995, "rle": {"size": [628, 640], "counts": "YgU1>``0h2K5L4K5K5L4K5K4L5L4K5K5K5L4K4L5Kj0WOb0_OO01O10O0100O00100O010O00100O00100O010^CWIZ<h6fCZIZ<e6eC^IY<b6fCaIY<]6gCeIY<Z6gChIW<X6hCkIW<S6iCoIW<P6hCSJV<m5iCVJV<h5jC[JU<b6O2O1O0O2O1O0O2O1O0O2O1O001N2O001N2O001N2O001N10000O1000O010000000000000O10001O000000000O100000000000000O101O00000000000O1000000000O1O1O101N1O1O2O0O1O2N1O101N1O1O2O0O2N1O1O2N1O1O2N1O1O2N1O2N1O1OoM\\GUIb8l6R21O1N2O1N2O1N2O1N2O1N2O1N2O1N2O0O2O1N2O1N200001O001O00YFSHh7l7RHZHo7e7jGbHV8^7dGhH]8W7[GQIe8o6mF_IT9`6_FmIb9R8001O010O001O1O001N101N1O2O1N1O2O0O2N2O2M4L4M2M4L4M3L4L4M3L4L4M3L4L4K5J5K5J6K5K5J5L5K5K4K6K5K4K6J6G9G8H9G9G8G:G9Goaa6"}, "label": "there is a giraffe eating the brown grass"}]}
{"id": 279076, "image": "COCO_train2014_000000279076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing black boots standing next to a dog\"."}], "task": "refering", "answer_template": "The \"a person wearing black boots standing next to a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 60, 61, 62, 63, 64, 65, 66, 67, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 110, 111, 112, 113, 129, 130, 131, 133, 134, 135, 136, 152, 153, 154, 157, 158, 159, 175, 176, 177, 180, 181, 182, 198, 199, 200, 203, 204, 205, 221, 222, 223, 226, 227, 228, 244, 245, 246, 249, 250, 251, 267, 268, 269, 272, 273, 274, 290, 291, 295, 296, 297, 318, 319, 320, 321, 343, 344], "bbox": [0.47781250000000003, 0.001880952380952381, 0.97921875, 0.9456904761904762], "iscrowd": 0, "area": 52143.88945000002, "rle": {"size": [420, 640], "counts": "m`m33n<3M4L3L5L3N3O001O001O001O1O001O001O000O1000001O000000000000001O0000000000001O000000001O0000001O0000001O0000001O0000001O00001O0000001O00001O001O001O001O1O001O001O1O001O001O001O1O001O0fK\\NTMe1]2kNPKAi0g1f3WO^KUOk0n1o2]OSLhNl0T2Z2CgL\\Nn0[2c1I[MoMQ1b2l0OPNbMR1h275dNVMT1o2@;YOiLU1V3jNa0M\\LX1\\3TNg0b0oKY1a8cNcG[1_8bNcG]1^8`NeG_1]8^NfG`1[8]NhGb1Z8[NgGe1[8WNgGi1k9O0000000000001O000000000000001O00000000O1000000O10000O100H8G9G9G9H8G9\\HUMf4T3RKnNk2[1mLb0U1GcNQ2DXN5_7H8O12N1O2N2N2N2N2N2M3N2N2N1O2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M4L3M3Ma1_N4L4L4L4L4L4L4L4L4L4L4L4L4LT1lNS1mN3M2N2N2N1O000001O01O02N3M4L4L3M4M2M4L4L3M4L4M2M4L4L3M4L3M4M3L3M4L4L3M4L4M2M4L4L3M4L3M4M4lM[K_Kk4T4^KjKg4j3bKTLc4`3fK^L_4V3jKhL\\4k2mKSMX4`2SL^MQ4V2XLhMm3l1\\LRNj3a1_L]Nf3<^MBg2kNiNS1f6N2O1N2N2N2N2M3N2M3M4L3N2M3M3M3N_n4"}, "label": "a person wearing black boots standing next to a dog"}]}
{"id": 279076, "image": "COCO_train2014_000000279076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with orange color shoe besides a fur dog\"."}], "task": "refering", "answer_template": "The \"man with orange color shoe besides a fur dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 163, 164, 165, 166, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 307, 328], "bbox": [0.033921875000000004, 0.0045, 0.35245312500000003, 0.9393333333333332], "iscrowd": 0, "area": 42986.6802, "rle": {"size": [420, 640], "counts": "U[94l<5K6J5M4M2M3N3L3N3L3M3N1N2O1N2O1N3N1N2O1N2O1N2N2O1N2O1N3L3M3M3M3M3M3N2M3M3M3M3M4L3M3N2M3M3M3M3M3M3M3M4M2M3M7I;E:E9F6I6K5K5J6K5J6K6L3N2N2N2N2N2_L`H\\1b7aNgHW1\\7fNlHR1V7kNTIl0n6ROYIe0k6XO]I:l6D[IMo60XIBR7<UIVOU7f0SIkNW7R1_2000000000001O0000000000001O0000000000001O00000001O000001O0000000000001O0000000000001O000000000000WM_OoHa0R6?nI@S5`1lJ`NV4_2iKaMR4e2mK[MR4g2mKYMR4j2lKVMR4m2mKSMR4o2nKPMQ4R3nKnLP4U3oKlLo3V3PLjLo3Y3oKgLo3\\3PLdLo3^3PLbLn3a3QL_Ln3c3RL\\Lm3f3RLZLl3i3SLWLl3l3RLTLm3n3RLRLl3Q4SLoKl3S4SLmKl3U4TLjKj3Y4ULgKj3[4ULeKj3\\4VLdKh3^4XLbKg3_4YLaKe3b4ZL^Ke3c4\\L\\Kc3e4\\L\\Kb3f4^LZKa3g4_LYK`3i4^LXK`3j4`LVK_3k4aLUK^3l4aLUK]3n4bLRK]3o4cLQK[3Q5dLPK[3Q5eLoJZ3R5fLnJX3U5fLlJY3U5gLkJX3V5hLkJU3W5jLjJU3X5jLhJU3Y5kLgJS3[5lLfJS3[5mLeJQ3]5oLcJP3_5nLbJQ3_5oLaJo2a5QM_Jn2b5QM_Jn2c5QMUKU2m4kMmKZ1T4eNhL?Z3@aMCb2;U5O1O1N2O1O0O2O1O1N2O1O1N\\^Y5"}, "label": "man with orange color shoe besides a fur dog"}]}
{"id": 279076, "image": "COCO_train2014_000000279076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange rain boots and black leggings\"."}], "task": "refering", "answer_template": "The \"the orange rain boots and black leggings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 163, 164, 165, 166, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 307, 328], "bbox": [0.033921875000000004, 0.0045, 0.35245312500000003, 0.9393333333333332], "iscrowd": 0, "area": 42986.6802, "rle": {"size": [420, 640], "counts": "U[94l<5K6J5M4M2M3N3L3N3L3M3N1N2O1N2O1N3N1N2O1N2O1N2N2O1N2O1N3L3M3M3M3M3M3N2M3M3M3M3M4L3M3N2M3M3M3M3M3M3M3M4M2M3M7I;E:E9F6I6K5K5J6K5J6K6L3N2N2N2N2N2_L`H\\1b7aNgHW1\\7fNlHR1V7kNTIl0n6ROYIe0k6XO]I:l6D[IMo60XIBR7<UIVOU7f0SIkNW7R1_2000000000001O0000000000001O0000000000001O00000001O000001O0000000000001O0000000000001O000000000000WM_OoHa0R6?nI@S5`1lJ`NV4_2iKaMR4e2mK[MR4g2mKYMR4j2lKVMR4m2mKSMR4o2nKPMQ4R3nKnLP4U3oKlLo3V3PLjLo3Y3oKgLo3\\3PLdLo3^3PLbLn3a3QL_Ln3c3RL\\Lm3f3RLZLl3i3SLWLl3l3RLTLm3n3RLRLl3Q4SLoKl3S4SLmKl3U4TLjKj3Y4ULgKj3[4ULeKj3\\4VLdKh3^4XLbKg3_4YLaKe3b4ZL^Ke3c4\\L\\Kc3e4\\L\\Kb3f4^LZKa3g4_LYK`3i4^LXK`3j4`LVK_3k4aLUK^3l4aLUK]3n4bLRK]3o4cLQK[3Q5dLPK[3Q5eLoJZ3R5fLnJX3U5fLlJY3U5gLkJX3V5hLkJU3W5jLjJU3X5jLhJU3Y5kLgJS3[5lLfJS3[5mLeJQ3]5oLcJP3_5nLbJQ3_5oLaJo2a5QM_Jn2b5QM_Jn2c5QMUKU2m4kMmKZ1T4eNhL?Z3@aMCb2;U5O1O1N2O1O0O2O1O1N2O1O1N\\^Y5"}, "label": "the orange rain boots and black leggings"}]}
{"id": 508456, "image": "COCO_train2014_000000508456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on left\"."}], "task": "refering", "answer_template": "The \"zebra on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 100, 101, 102, 118, 119, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 237, 238, 239, 254, 255, 260, 261, 277, 278, 283, 284, 300, 301, 306, 307, 308, 324], "bbox": [0.057296875000000004, 0.17752083333333332, 0.49214062500000005, 0.8719166666666667], "iscrowd": 0, "area": 39098.890100000004, "rle": {"size": [480, 640], "counts": "Uaa0c1\\=;E<D<D;E<D<D;E<D<D8H8H9G8H9G6J5K5K5L4K5L5J5L3L01O001O0101NjNjHdJU7T5UIkJk6k4`IUK^6b4mI]KS6Z4XJdKg5V4aJiK_5V4cJjK[5U4hJjKX5U4kJiKT5W4nJhKR5W4PKgKP5Y4RKfKn4Y4UKdKl4[4VKcKk4\\4WKcKi4[4ZKcKg4\\4\\KbKd4]4^KaKc4^4_K`Kb4_4`K`K`4_4cK^K^4a4Q3N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O11O00001O00001O00001O001O01O01O00001O00001O01O0001O000000001O01O000001O0000000010O00000000000000001N10000000000000001N0100O10000O100O10O010000O100O100O10O10O100O3N5K6I6K6H7J7I6J7I6`FgKn7j5K5QHWI]7W7O1O1O1O1O1ObNcH[K[7c4nHXKQ7d4XIXKg6d4bIXK]6e4lIVKS6j4RJRKm5P5VJlJi5V5ZJeJe5^5oIYJUO5k6d5cIeJDCh6j5VIRK4POe6\\7>3M2N3L3N=C100O1O100O1OgNhIlIW6e5ZJZJf5`5bJ^J]5]5jJbJU5Z5QKeJo4U5YKiJf4S5`KlJ_4o4iKoJV4l4QLSKo3h4WLWKh3d4`LZK_3b4gL]KX3`4nL^KP3b4SM]Kj2d4YM[Kc2g4aMWK[2k4hMTKT2n4PNPKk1S5XNlJc1W5`NgJa1W5cNgJ]1W5X4M2N3M2N3SOPFfLR:X3TFbLo9[3WF_Lk9_3[F[Lh9b3l0N3M2N3M2N3M2N3M2N2N2N1O2N2N2N2O0O2O1O1N2_NeC<[<EjC5W<JPDOQ<0UDIm;6]D@d;?_1O001O1O1O1O4L6J5K6JeWh4"}, "label": "zebra on left"}]}
{"id": 508456, "image": "COCO_train2014_000000508456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall zebra\"."}], "task": "refering", "answer_template": "The \"a tall zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 100, 101, 102, 118, 119, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 237, 238, 239, 254, 255, 260, 261, 277, 278, 283, 284, 300, 301, 306, 307, 308, 324], "bbox": [0.057296875000000004, 0.17752083333333332, 0.49214062500000005, 0.8719166666666667], "iscrowd": 0, "area": 39098.890100000004, "rle": {"size": [480, 640], "counts": "Uaa0c1\\=;E<D<D;E<D<D;E<D<D8H8H9G8H9G6J5K5K5L4K5L5J5L3L01O001O0101NjNjHdJU7T5UIkJk6k4`IUK^6b4mI]KS6Z4XJdKg5V4aJiK_5V4cJjK[5U4hJjKX5U4kJiKT5W4nJhKR5W4PKgKP5Y4RKfKn4Y4UKdKl4[4VKcKk4\\4WKcKi4[4ZKcKg4\\4\\KbKd4]4^KaKc4^4_K`Kb4_4`K`K`4_4cK^K^4a4Q3N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O11O00001O00001O00001O001O01O01O00001O00001O01O0001O000000001O01O000001O0000000010O00000000000000001N10000000000000001N0100O10000O100O10O010000O100O100O10O10O100O3N5K6I6K6H7J7I6J7I6`FgKn7j5K5QHWI]7W7O1O1O1O1O1ObNcH[K[7c4nHXKQ7d4XIXKg6d4bIXK]6e4lIVKS6j4RJRKm5P5VJlJi5V5ZJeJe5^5oIYJUO5k6d5cIeJDCh6j5VIRK4POe6\\7>3M2N3L3N=C100O1O100O1OgNhIlIW6e5ZJZJf5`5bJ^J]5]5jJbJU5Z5QKeJo4U5YKiJf4S5`KlJ_4o4iKoJV4l4QLSKo3h4WLWKh3d4`LZK_3b4gL]KX3`4nL^KP3b4SM]Kj2d4YM[Kc2g4aMWK[2k4hMTKT2n4PNPKk1S5XNlJc1W5`NgJa1W5cNgJ]1W5X4M2N3M2N3SOPFfLR:X3TFbLo9[3WF_Lk9_3[F[Lh9b3l0N3M2N3M2N3M2N3M2N2N2N1O2N2N2N2O0O2O1O1N2_NeC<[<EjC5W<JPDOQ<0UDIm;6]D@d;?_1O001O1O1O1O4L6J5K6JeWh4"}, "label": "a tall zebra"}]}
{"id": 508456, "image": "COCO_train2014_000000508456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in a sunbeam\"."}], "task": "refering", "answer_template": "The \"a zebra in a sunbeam\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 128, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 246, 247, 248, 249, 265, 266, 267, 270, 271, 272, 288, 289, 290, 293, 294, 295, 312, 313, 316, 317, 318, 339], "bbox": [0.458421875, 0.2426875, 0.8460625, 0.8471875000000001], "iscrowd": 0, "area": 32280.109950000002, "rle": {"size": [480, 640], "counts": "koY47g>7J6I6K6I3M101N100O2O0O1N3M2N2N3M2N2O2QOn01O101N1O2O0O1015J5K6K5J3N1N2N2O1N2O1O1O1N2O1O1O3M3L4M4L3M3M3L5L4L7I7I7H8I7I5K6J6I7J6J3M00O01000O1000O10hM]HQMd7i2iHPMW7j2SIRMm6h2]IoLh6l2aIgLh6W3c20O2N2N2O1N2N2N2O0Ob0^Og1ZNN1O1O00000000000001O01O0jNV1gNY1ZOf0000000000001O000000000O10000000001O0000000000000O10001O0000000000000000001N1000000000000000001O00000O10000001O1O100O2N1O1O1O2N1O1O1O101N1O1O1O2N1O1O1O2N100O7I8QGlKb6\\4PIlKn6\\4eHjKZ7^4XHiKg7^4lGjKR8c5O2N100O2N1O2N10N1O1N2N2O2M2N2O1N2N2M4L3M3M3N2M4L3iMXHSMk7Q3P23M3N3L3N3M2O1N3M100O1O100O1O100O1O2O0O11O1O1O1O1O2N2N2N3N[2dMS1mN;E\\Q^1"}, "label": "a zebra in a sunbeam"}]}
{"id": 508456, "image": "COCO_train2014_000000508456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on the right flicking its tail\"."}], "task": "refering", "answer_template": "The \"zebra on the right flicking its tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 128, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 246, 247, 248, 249, 265, 266, 267, 270, 271, 272, 288, 289, 290, 293, 294, 295, 312, 313, 316, 317, 318, 339], "bbox": [0.458421875, 0.2426875, 0.8460625, 0.8471875000000001], "iscrowd": 0, "area": 32280.109950000002, "rle": {"size": [480, 640], "counts": "koY47g>7J6I6K6I3M101N100O2O0O1N3M2N2N3M2N2O2QOn01O101N1O2O0O1015J5K6K5J3N1N2N2O1N2O1O1O1N2O1O1O3M3L4M4L3M3M3L5L4L7I7I7H8I7I5K6J6I7J6J3M00O01000O1000O10hM]HQMd7i2iHPMW7j2SIRMm6h2]IoLh6l2aIgLh6W3c20O2N2N2O1N2N2N2O0Ob0^Og1ZNN1O1O00000000000001O01O0jNV1gNY1ZOf0000000000001O000000000O10000000001O0000000000000O10001O0000000000000000001N1000000000000000001O00000O10000001O1O100O2N1O1O1O2N1O1O1O101N1O1O1O2N1O1O1O2N100O7I8QGlKb6\\4PIlKn6\\4eHjKZ7^4XHiKg7^4lGjKR8c5O2N100O2N1O2N10N1O1N2N2O2M2N2O1N2N2M4L3M3M3N2M4L3iMXHSMk7Q3P23M3N3L3N3M2O1N3M100O1O100O1O100O1O2O0O11O1O1O1O1O2N2N2N3N[2dMS1mN;E\\Q^1"}, "label": "zebra on the right flicking its tail"}]}
{"id": 451119, "image": "COCO_train2014_000000451119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the blue shirt holding the wii controller\"."}], "task": "refering", "answer_template": "The \"the girl in the blue shirt holding the wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 112, 113, 114, 129, 130, 131, 132, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 177, 178, 179, 180, 181, 182, 183, 184, 194, 195, 197, 198, 199, 200, 214, 215, 216, 217, 231, 232, 233, 234, 235, 248, 249, 250, 251, 252, 265, 266, 268, 269, 270, 281, 282, 283, 286, 287, 299, 304], "bbox": [0.4572291666666667, 0.221046875, 0.9083333333333333, 0.75175], "iscrowd": 0, "area": 27956.432249999998, "rle": {"size": [640, 480], "counts": "^UY41nc0;E:F6J1O1O1O1O1O1O1O1O1O1O1O1O1O0O10O1000000O1000000O1000000O101N100O100O100O100O10O0100O10O01O10O0100O10O010YCTO[6k0]I_Oa6b0_I_O_6a0aIA\\6a0dI@Z6a0eIAY6`0gIAW6`0hICV6=jIDT6=kIES6<lIFS6:mIGQ6:nIHP69]HVOlLb0f:8WHFiL4n:7SH5eLEV;8oGc0_LWO`;7mGR2Q8nMjGW2U8jMfG\\2X8eMgG]2W8dMgG_2W8bMgG`2W8bMhG`2V8aMhGb2V8_MhGc2W8^MhGd2V8_MfGd2X8`MiEfN?l3e9bMhEoN7`3o9eMeEQO7]3P:gMeEQO7Y3R:jMbEV4\\:PL_ER4_:h2M3N2N2O10O101]NQFTIo9`6aF\\I_9b6hFYIY9c6nFXIS9f6SGVIm8h6YGSIi8i6^GRIc8l6cGnH_8P7Z2M2O101N1O1O2N1O2O0O1O2N10002M2O1N2010O1O002N3N3L3M3M4M2M4M3M3M3N2M11O4L4L5K2M3N2N2N1O2N1O00O10O1000O0100O100O0WMgBkNY=P1VCfNj<U1ZCkNf<Q1_CoN`<m0dCSO\\<i0hCWOX<e0mCZOT<a0PD_OP<=UDBk;9ZDGh;4\\DKf;0]D0e;L^D3d;H_D8d;C_D<c;@_D`0b;\\ObDc0_;ZOcDf0^;VOfDi0[;TOgDl0Z;POjDo0W;nNkDR1V;kNmDT1T;hNoDX1R;bNTE]1m:\\NYEd1h:UN_Ej1\\>01O001O001O00001O0O2O001O001O001N10000000000O10004L5F;D;E;ElQk0"}, "label": "the girl in the blue shirt holding the wii controller"}]}
{"id": 451119, "image": "COCO_train2014_000000451119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl playing play station\"."}], "task": "refering", "answer_template": "The \"girl playing play station\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 112, 113, 114, 129, 130, 131, 132, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 177, 178, 179, 180, 181, 182, 183, 184, 194, 195, 197, 198, 199, 200, 214, 215, 216, 217, 231, 232, 233, 234, 235, 248, 249, 250, 251, 252, 265, 266, 268, 269, 270, 281, 282, 283, 286, 287, 299, 304], "bbox": [0.4572291666666667, 0.221046875, 0.9083333333333333, 0.75175], "iscrowd": 0, "area": 27956.432249999998, "rle": {"size": [640, 480], "counts": "^UY41nc0;E:F6J1O1O1O1O1O1O1O1O1O1O1O1O1O0O10O1000000O1000000O1000000O101N100O100O100O100O10O0100O10O01O10O0100O10O010YCTO[6k0]I_Oa6b0_I_O_6a0aIA\\6a0dI@Z6a0eIAY6`0gIAW6`0hICV6=jIDT6=kIES6<lIFS6:mIGQ6:nIHP69]HVOlLb0f:8WHFiL4n:7SH5eLEV;8oGc0_LWO`;7mGR2Q8nMjGW2U8jMfG\\2X8eMgG]2W8dMgG_2W8bMgG`2W8bMhG`2V8aMhGb2V8_MhGc2W8^MhGd2V8_MfGd2X8`MiEfN?l3e9bMhEoN7`3o9eMeEQO7]3P:gMeEQO7Y3R:jMbEV4\\:PL_ER4_:h2M3N2N2O10O101]NQFTIo9`6aF\\I_9b6hFYIY9c6nFXIS9f6SGVIm8h6YGSIi8i6^GRIc8l6cGnH_8P7Z2M2O101N1O1O2N1O2O0O1O2N10002M2O1N2010O1O002N3N3L3M3M4M2M4M3M3M3N2M11O4L4L5K2M3N2N2N1O2N1O00O10O1000O0100O100O0WMgBkNY=P1VCfNj<U1ZCkNf<Q1_CoN`<m0dCSO\\<i0hCWOX<e0mCZOT<a0PD_OP<=UDBk;9ZDGh;4\\DKf;0]D0e;L^D3d;H_D8d;C_D<c;@_D`0b;\\ObDc0_;ZOcDf0^;VOfDi0[;TOgDl0Z;POjDo0W;nNkDR1V;kNmDT1T;hNoDX1R;bNTE]1m:\\NYEd1h:UN_Ej1\\>01O001O001O00001O0O2O001O001O001N10000000000O10004L5F;D;E;ElQk0"}, "label": "girl playing play station"}]}
{"id": 451119, "image": "COCO_train2014_000000451119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young male child with a red shirt and blue jeans holding a wii remote\"."}], "task": "refering", "answer_template": "The \"a young male child with a red shirt and blue jeans holding a wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 124, 125, 126, 141, 142, 143, 144, 158, 159, 160, 161, 175, 176, 177, 178, 193, 194, 210, 211, 212, 226, 227, 228, 229, 230, 243, 244, 245, 246, 247, 260, 261, 263], "bbox": [0.2856666666666667, 0.280421875, 0.5453125, 0.68690625], "iscrowd": 0, "area": 15637.615549999999, "rle": {"size": [640, 480], "counts": "dle21mc02O2N2M2O2M3N1N3N2M2O2N2M2O2M3XAQO\\:P1`EUO]:n0^EWO`:k0[EZOc:g0YE]Oe:f0VE_Oh:b0TECi:`0REEl:<PEIm:9oDLo:5mDOQ;3kD2S;OiD6U;LhD7U;KjD7[:c0dE_OP:l0oEUOP:l0oEVOo9l0oETOQ:m0nEQOT:P1kEoNV:R1iEmNX:V3eChL^<X3aCgL`<[3^CcLd<^3[CaLf<Z52O1O1N3N1N2O1N2O1O1N3N2M3L4K5K5K5K5]OkAVKZ>i4fAZKX>e4iA\\KV>c4jA`KT>^4nAcKQ>\\4oAgKo=X4RBiKm=U4TBnKj=Q4WBPLh=n3YBTLf=k3[BWLc=g3^B[La=d3`B^L^=`3cBbL\\=]3eBeLY=Y3iBhLW=V3jBkLU=T3lBnLS=P3nBQMR=m2oBUMP=i2QCXMo<f2RC[Mn<c2SC_Ml<_2UCbMk<F_BZ1g0QOi<CcBZ1e0UOg<^OgB[1c0XOf<YOkB]1a0[Oc<TOPC_1>^Ob<kNXCe18@`<bNaCl11C]<YNjCS2IEj=2_BNa=@QCa0n<]OSCd0m<ZOUCf0k<XOWCi0h<UOYCl0g<RO[Cn0d<PO_CP1a<nN`CS1`<kNbCU1^<iNdCW1\\<gNgCX1Y<fNiCZ1W<dNkC\\1U<bNmC^1S<`NoC_1R<\\NTDb1m;VN[Di1o>O1O1N3N1O1O2N2N2N2N2N2N2N2N2N2WOdlW4"}, "label": "a young male child with a red shirt and blue jeans holding a wii remote"}]}
{"id": 451119, "image": "COCO_train2014_000000451119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy playing the game\"."}], "task": "refering", "answer_template": "The \"a boy playing the game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 124, 125, 126, 141, 142, 143, 144, 158, 159, 160, 161, 175, 176, 177, 178, 193, 194, 210, 211, 212, 226, 227, 228, 229, 230, 243, 244, 245, 246, 247, 260, 261, 263], "bbox": [0.2856666666666667, 0.280421875, 0.5453125, 0.68690625], "iscrowd": 0, "area": 15637.615549999999, "rle": {"size": [640, 480], "counts": "dle21mc02O2N2M2O2M3N1N3N2M2O2N2M2O2M3XAQO\\:P1`EUO]:n0^EWO`:k0[EZOc:g0YE]Oe:f0VE_Oh:b0TECi:`0REEl:<PEIm:9oDLo:5mDOQ;3kD2S;OiD6U;LhD7U;KjD7[:c0dE_OP:l0oEUOP:l0oEVOo9l0oETOQ:m0nEQOT:P1kEoNV:R1iEmNX:V3eChL^<X3aCgL`<[3^CcLd<^3[CaLf<Z52O1O1N3N1N2O1N2O1O1N3N2M3L4K5K5K5K5]OkAVKZ>i4fAZKX>e4iA\\KV>c4jA`KT>^4nAcKQ>\\4oAgKo=X4RBiKm=U4TBnKj=Q4WBPLh=n3YBTLf=k3[BWLc=g3^B[La=d3`B^L^=`3cBbL\\=]3eBeLY=Y3iBhLW=V3jBkLU=T3lBnLS=P3nBQMR=m2oBUMP=i2QCXMo<f2RC[Mn<c2SC_Ml<_2UCbMk<F_BZ1g0QOi<CcBZ1e0UOg<^OgB[1c0XOf<YOkB]1a0[Oc<TOPC_1>^Ob<kNXCe18@`<bNaCl11C]<YNjCS2IEj=2_BNa=@QCa0n<]OSCd0m<ZOUCf0k<XOWCi0h<UOYCl0g<RO[Cn0d<PO_CP1a<nN`CS1`<kNbCU1^<iNdCW1\\<gNgCX1Y<fNiCZ1W<dNkC\\1U<bNmC^1S<`NoC_1R<\\NTDb1m;VN[Di1o>O1O1N3N1O1O2N2N2N2N2N2N2N2N2N2WOdlW4"}, "label": "a boy playing the game"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair in background , front left\"."}], "task": "refering", "answer_template": "The \"chair in background , front left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 27, 28, 29, 30, 44, 45, 46, 47, 61, 62, 63, 64, 77, 78, 79, 80, 95, 96, 97], "bbox": [0.5632291666666667, 0.0, 0.8072291666666668, 0.258703125], "iscrowd": 0, "area": 10551.568999999998, "rle": {"size": [640, 480], "counts": "mkX51kc04L5K4L401O00001O00000010O0001O00001O00010O00001O0000001O01O01O0iL_OoBa0Q=IeB8Y=b0nA^OR>]1SAcNm>W2Y@jMe?Y30`0_O<E00O010O3N1N2O1N3N1N2O0O01O01J5C>0O00100O0010O0100O0010O010i0WO000000O10O10000000O100000O100000O100N2J6J5K6J6J6J6J6J7I6I7I7I8H7H8J7L3M3N2M4L3N2M3M4M4K8H8IaSj1"}, "label": "chair in background , front left"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the chair on the left\"."}], "task": "refering", "answer_template": "The \"the back of the chair on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 27, 28, 29, 30, 44, 45, 46, 47, 61, 62, 63, 64, 77, 78, 79, 80, 95, 96, 97], "bbox": [0.5632291666666667, 0.0, 0.8072291666666668, 0.258703125], "iscrowd": 0, "area": 10551.568999999998, "rle": {"size": [640, 480], "counts": "mkX51kc04L5K4L401O00001O00000010O0001O00001O00010O00001O0000001O01O01O0iL_OoBa0Q=IeB8Y=b0nA^OR>]1SAcNm>W2Y@jMe?Y30`0_O<E00O010O3N1N2O1N3N1N2O0O01O01J5C>0O00100O0010O0100O0010O010i0WO000000O10O10000000O100000O100000O100N2J6J5K6J6J6J6J6J7I6I7I7I8H7H8J7L3M3N2M4L3N2M3M4M4K8H8IaSj1"}, "label": "the back of the chair on the left"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of soup in a white bowl with cashews and a spoon on the side\"."}], "task": "refering", "answer_template": "The \"a bowl of soup in a white bowl with cashews and a spoon on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 148, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 361, 362, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384], "bbox": [0.12883333333333333, 0.35954687500000004, 0.9378333333333333, 0.9820156250000001], "iscrowd": 0, "area": 116879.45309999998, "rle": {"size": [640, 480], "counts": "TUW19dc0?B=C>B>A>C>B=C=B4M4L3M4K5L3M4L3L5L4L3M4K5L3M4L3L5L4L3N3M2N3M3M2N3L3N3M3M2N3M2N3M3M2N3L3N3M3M2N3M2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N101N1O2N1O2O0O2N1O2N101N1O2N1O2N101N1O2N1O2O0O2N1O2N1O2N1N3M2N3L3O2O0O2O0O2N101N101N101N101N1O2O0O2O0O2O0O1O101N100O100O2N100O100O101N1O100O100O2O0O1O100O100O100O1O100O100O100O1O100O100O100O1000O01000000000000O10000000000O1000000000000O1000O10000000O10000000000O1000000000000O100000O100O10O0100O100O10O0100O100O101N100O100O100O100O100O2O0O100O100O100O100O101N100O10000O100O100O101N100O100O100O100O101N2O0O2O1N2O0O2O1N101N2O1N101N2N2N101N2N1O2N2N2N101N2N1O2N2N2N101N2N2N1O2N2O0O2N2N2N1O2N2O0O2N2N2N1O2N2O0O2M3N2M3M3M4M2M4L3N3L3M3M4M2M4L3N3L3M3M4M2M4L3N3L3M4L3M3N3L3M4L3M4L3M3M4L3N3L3M5K4L5K5K5K4M4K5K4L5K5K5K4L5L4K4L5K5K;Ei0WOj0VOcZb0"}, "label": "a bowl of soup in a white bowl with cashews and a spoon on the side"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with the red sauce\"."}], "task": "refering", "answer_template": "The \"the bowl with the red sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 148, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 361, 362, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384], "bbox": [0.12883333333333333, 0.35954687500000004, 0.9378333333333333, 0.9820156250000001], "iscrowd": 0, "area": 116879.45309999998, "rle": {"size": [640, 480], "counts": "TUW19dc0?B=C>B>A>C>B=C=B4M4L3M4K5L3M4L3L5L4L3M4K5L3M4L3L5L4L3N3M2N3M3M2N3L3N3M3M2N3M2N3M3M2N3L3N3M3M2N3M2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N101N1O2N1O2O0O2N1O2N101N1O2N1O2N101N1O2N1O2O0O2N1O2N1O2N1N3M2N3L3O2O0O2O0O2N101N101N101N101N1O2O0O2O0O2O0O1O101N100O100O2N100O100O101N1O100O100O2O0O1O100O100O100O1O100O100O100O1O100O100O100O1000O01000000000000O10000000000O1000000000000O1000O10000000O10000000000O1000000000000O100000O100O10O0100O100O10O0100O100O101N100O100O100O100O100O2O0O100O100O100O100O101N100O10000O100O100O101N100O100O100O100O101N2O0O2O1N2O0O2O1N101N2O1N101N2N2N101N2N1O2N2N2N101N2N1O2N2N2N101N2N2N1O2N2O0O2N2N2N1O2N2O0O2N2N2N1O2N2O0O2M3N2M3M3M4M2M4L3N3L3M3M4M2M4L3N3L3M3M4M2M4L3N3L3M4L3M3N3L3M4L3M4L3M3M4L3N3L3M5K4L5K5K5K4M4K5K4L5K5K5K4L5L4K4L5K5K;Ei0WOj0VOcZb0"}, "label": "the bowl with the red sauce"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the part of the bowl of stew that has the spoon in it\"."}], "task": "refering", "answer_template": "The \"the part of the bowl of stew that has the spoon in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 268, 285, 300, 301, 302, 316, 317, 318, 332, 333, 348, 349, 364, 365, 380, 381], "bbox": [0.3642291666666667, 0.60221875, 0.8232291666666666, 0.9991406250000001], "iscrowd": 0, "area": 4148.4776, "rle": {"size": [640, 480], "counts": "nc^31mc02O1O1O1O1O1O1O1O1O1O001O1O1O1O0000000000000O10000000O10000000000000000000000O100000000000O100000000000O10000000000000000000000O100000O1000000000000000O1000000000000000000000000O10O10000000000000000000O1000000000000000000000O10O1000000000000000000O100000000000000O100000O1001O0O2O10O100O100O10O010000O100O100O10000O100N21OO1O10000O100O2O000O101N10001N10001N100O2OO1O001O010O00O10001N100O100O01O10O010O000N3SNoNj@R1X?RO`@P1b?TOV@o0k?VOl_Ol0V`0XOb_Oj0``0[OS_Ol0o`0P15K5K5L4K5K=D;D<^Ohfd1"}, "label": "the part of the bowl of stew that has the spoon in it"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver spoon in a bowl of red stew\"."}], "task": "refering", "answer_template": "The \"a silver spoon in a bowl of red stew\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 268, 285, 300, 301, 302, 316, 317, 318, 332, 333, 348, 349, 364, 365, 380, 381], "bbox": [0.3642291666666667, 0.60221875, 0.8232291666666666, 0.9991406250000001], "iscrowd": 0, "area": 4148.4776, "rle": {"size": [640, 480], "counts": "nc^31mc02O1O1O1O1O1O1O1O1O1O001O1O1O1O0000000000000O10000000O10000000000000000000000O100000000000O100000000000O10000000000000000000000O100000O1000000000000000O1000000000000000000000000O10O10000000000000000000O1000000000000000000000O10O1000000000000000000O100000000000000O100000O1001O0O2O10O100O100O10O010000O100O100O10000O100N21OO1O10000O100O2O000O101N10001N10001N100O2OO1O001O010O00O10001N100O100O01O10O010O000N3SNoNj@R1X?RO`@P1b?TOV@o0k?VOl_Ol0V`0XOb_Oj0``0[OS_Ol0o`0P15K5K5L4K5K=D;D<^Ohfd1"}, "label": "a silver spoon in a bowl of red stew"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table under white bowl with spoons in it\"."}], "task": "refering", "answer_template": "The \"table under white bowl with spoons in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 308, 323, 324, 325, 326, 340, 341, 342, 343, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380], "bbox": [0.017354166666666667, 0.6171875, 0.3888958333333334, 0.984375], "iscrowd": 0, "area": 15648.069449999994, "rle": {"size": [640, 480], "counts": "]\\51nc0;F;D<E:F;E:F;E;E:G:E;E:F;E;E:F;E;E:F;E;F9F6J00000000000000000001O00000000000000000000001O00000000000000000ZJ]Co3c<kKhCP4Y<iKRDR4n;hK]DS4c;gKhDT4X;fKSEU4m:eK]EW4c:cKhEX4U=K5K5L4K4L4L2N3M2N2N3M2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N1O1O001O001O1O001O010O1O001O001O1O001O001O1O001O001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O0010O0001O00001O001O00001O001O00000000O1O1001O4L[`f5"}, "label": "table under white bowl with spoons in it"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"portion of table defined by the white dishes\"."}], "task": "refering", "answer_template": "The \"portion of table defined by the white dishes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 308, 323, 324, 325, 326, 340, 341, 342, 343, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380], "bbox": [0.017354166666666667, 0.6171875, 0.3888958333333334, 0.984375], "iscrowd": 0, "area": 15648.069449999994, "rle": {"size": [640, 480], "counts": "]\\51nc0;F;D<E:F;E:F;E;E:G:E;E:F;E;E:F;E;E:F;E;F9F6J00000000000000000001O00000000000000000000001O00000000000000000ZJ]Co3c<kKhCP4Y<iKRDR4n;hK]DS4c;gKhDT4X;fKSEU4m:eK]EW4c:cKhEX4U=K5K5L4K4L4L2N3M2N2N3M2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N1O1O001O001O1O001O010O1O001O001O1O001O001O1O001O001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O0010O0001O00001O001O00001O001O00000000O1O1001O4L[`f5"}, "label": "portion of table defined by the white dishes"}]}
{"id": 508467, "image": "COCO_train2014_000000508467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver placemat on right side\"."}], "task": "refering", "answer_template": "The \"silver placemat on right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 160, 166, 167, 168, 169, 184, 185, 186, 201, 202, 203, 219, 220, 237, 254, 270, 271, 287, 288, 304, 305, 320, 321, 322, 336, 337, 338, 339, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.4270833333333333, 0.23176562500000003, 0.9895833333333334, 0.9869843749999999], "iscrowd": 0, "area": 46567.01665, "rle": {"size": [640, 480], "counts": "g[P43kc04L4L4L2N1O2N1O2N1O2N1O2N1cNSOg_On0Z`0SOc_Oo0]`0QOa_OP1``0QO]_OQ1b`0QO[_OP1f`0QOW_OQ1i`0POT_OQ1m`0oNQ_OS1o`0nNn^OS1Sa0nNj^OS1Wa0nNf^OT1Za0mNc^OT1^a0lN`^OV1_a0lN^^OU1ca0lNZ^OV1fa0kNW^OV1ja0`00O1O100O10O0100O100O100O1O100O010O100O100O100000O1]JUNQIl1n6TNRIl1n6TNRIl1m6UNSIk1m6UNSIk1m6UNSIk1l6VNTIj1l6VNTIj1l6VNSIk1m6UNSIl1k6UNUIk1k6UNUIk1k6UNUIk1j6UNWIk1i6UNWIk1i6UNWIk1h6VNXIj1h6VNWIk1i6UNWIl1g6UNYIk1g6UNYIk1g6UNYIk1f6VNZIj1f6VNZIj1f6VNZIj1e6WN[Ii1e6WNZIj1f6VNZIk1d6VN\\Ij1d6VN\\Ij1d6VN\\Ij1c6WN]Ii1c6WN]Ii1c6WN]Ii1b6XN^Ih1b6XN^Ih1b6XN]Ij1a6WN_Ii1a6WN_Ii1a6WN_Ii1a6WN_Ii1`6XN`Ih1`6XN`Ih1`6XN`Ih1_6YNaIh1^6XNaIi1_6WNaIi1^6XNbIh1]6YNcIg1]6YNcIg1\\6ZNdIf1[6[NeIe1[6[NeIe1Z6\\NfIe1X6\\NgIe1Y6[NgIe1X6\\NhId1X6\\NhId1W6]NiId1U6]NkIc1U6]NkIc1T6^NlIc1R6^NnIb1R6^NmIc1R6^NnIb1Q6_NoIb1P6^NPJb1o5_NQJa1n5`NRJ`1n5`NRJa1l5`NTJ`1l5`NTJ`1k5aNTJ`1k5aNUJ`1j5`NVJ`1i5aNWJ_1h5bNXJ^1h5bNXJ_1f5bNZJ^1e5cN[J]1e5cN[J]1d5dN\\J]1b5dN]J]1c5cN]J]1b5dN^J\\1b5dN^J]1`5dN`J\\1_5eNaJ[1_5eNaJ[1^5fNbJ[1\\5fNdJZ1\\5fNcJ[1\\5fNdJZ1[5gNeJZ1Z5fNfJZ1Y5gNgJY1X5hNhJX1X5hNhJY1V5hNjJX1V5hNjJX1U5iNjJY1T5hNlJX1T5hNlJX1S5iNmJW1R5jNnJW1Q5iNoJW1P5jNPKV1o4kNQKU1o4kNQKV1m4kNRKV1m4kNSKU1m4kNSKU1l4lNTKU1k4kNUKV1i4kNWKW1f4jNZKW1e4iN[KY1a4iN_KX1_4iN`KZ1\\4hNdKY1Y4iNgKY1V4hNjKY1S4iNmKY1P4hNPLZ1l3hNTLY1j3hNVLZ1f3hNZLY1d3hN\\LZ1`3hN_LZ1_3gNaL[1[3gNeLZ1X3hNhLZ1U3gNkLZ1R3hNnLZ1o2gNQMZ1l2hNTMZ1i2gNWM[1e2gNZM[1d2fN\\M\\1`2fN`M[1]2gNcM[1Z2fNfM[1W2gNiM[1T2fNlM[1Q2gNoM[1n1fNRN[1k1gNTN\\1i1eNWN\\1f1eN[N]1b1dN^N]1_1eNaN]1[1eNeN]1X1dNhN]1U1eNkN]1R1dNnN]1o0eNPO^1m0cNSO^1j0dNVO^1g0cNYO^1d0dN\\O^1a0cN_O^1>dNB^1:dNF]18dNH^14dNK^13cNM_1OcN1^1McN3_1IcN7^1GcN9_1CcN=^1@dN`0^1]OcNc0^1ZOdNf0]1XOdNg0_1UOcNk0^1SOcNm0_1oNcNQ1^1mNcNS1_1iNcNW1^1dNfN\\1\\1]NiNc1X1VNnNj1S1oMSOP2P1hMVOX2k0aM[O_2g0YM_Og2b0SMCm2?kLGU3:dLL\\36\\L0d31UL5k3MmK9S4HgK=Y4D`Kb0`4@XKf0h4[OQKk0n4T7000000000000000000000000000000O100\\o2"}, "label": "silver placemat on right side"}]}
{"id": 221748, "image": "COCO_train2014_000000221748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in black dress with a bat in his hand\"."}], "task": "refering", "answer_template": "The \"a baseball player in black dress with a bat in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 187, 188, 189, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 278, 279, 280, 281, 301, 302, 303, 324, 325, 326, 346, 347, 348, 349, 369, 370, 371, 372, 373, 393, 394], "bbox": [0.0590625, 0.41759765625, 0.2899375, 0.95451171875], "iscrowd": 0, "area": 19390.961199999998, "rle": {"size": [512, 640], "counts": "c^c07h?1O2N1O2N1O2M2O1O2N1M4M2M4L3O1N3M2O2M2N3]MfN\\F\\1_9jN_FX1\\9lN`FX1]9lN]FY1b9iNXF]1f9dNVF`1h9cNUF`1i9bNUF_1j9bNTFa1i9bNUF_1j9bNUF_1j9cNSF`1j9bNUF_1i9dNTF^1i9eNVF\\1f9iNWFY1f9kNXFV1d9nN[FS1b9QO[FQ1d9PO[F5^OdNX:W1WF4GbNS:Y1TF5LaNo9\\1RF32`Nf9c1UFO7]N^9k1WFI?ZNT9S2[FDc0XNn8Y2[F@j0VNg8^2]F]Oo0SN`8i4eGUKk7X5ZHfJ_7^5fHaJY7^5jH`JV7^5mHaJR7_5oHaJQ7^5QIaJo6]5SIcJl6]5l1O1O0O2O1O0O101O0O1000000OeERLh8c5O1O1O1O1O1O1O1O1O1M3I7I7O100O100O10000O100O100O1000_MQHQLNAP8^4aH[KM1b7b4PJXKP6g4RJXKn5g4TJXKl5f4XJXKh5g4ZJXKf5d4_J[Ka5d4bJZK^5d4eJ\\KZ5c4hJ\\KX5b4jJ^KV5a4lJ^KT5`4oJ_KQ5`4PK`KP5^4SKaKm4^4UKaKk4S1mGT2Z3iLi4R1nGT2Z3jLh4P1PHU2Y3kLh4m0QHV2Z3lLf4l0QHW2c;eM`DX2c;eM_DZ2c;cM_D\\2c;bM^D\\2f;`M]D]2i;]MYDa2m;ZMTDd2e<N2N3M2N2N3M2N3M2N3M3M3J7J5K6I6K6J5K5K5K4LeWS7"}, "label": "a baseball player in black dress with a bat in his hand"}]}
{"id": 221748, "image": "COCO_train2014_000000221748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball umpire stands behind a catcher on a baseball field\"."}], "task": "refering", "answer_template": "The \"a baseball umpire stands behind a catcher on a baseball field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 187, 188, 189, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 278, 279, 280, 281, 301, 302, 303, 324, 325, 326, 346, 347, 348, 349, 369, 370, 371, 372, 373, 393, 394], "bbox": [0.0590625, 0.41759765625, 0.2899375, 0.95451171875], "iscrowd": 0, "area": 19390.961199999998, "rle": {"size": [512, 640], "counts": "c^c07h?1O2N1O2N1O2M2O1O2N1M4M2M4L3O1N3M2O2M2N3]MfN\\F\\1_9jN_FX1\\9lN`FX1]9lN]FY1b9iNXF]1f9dNVF`1h9cNUF`1i9bNUF_1j9bNTFa1i9bNUF_1j9bNUF_1j9cNSF`1j9bNUF_1i9dNTF^1i9eNVF\\1f9iNWFY1f9kNXFV1d9nN[FS1b9QO[FQ1d9PO[F5^OdNX:W1WF4GbNS:Y1TF5LaNo9\\1RF32`Nf9c1UFO7]N^9k1WFI?ZNT9S2[FDc0XNn8Y2[F@j0VNg8^2]F]Oo0SN`8i4eGUKk7X5ZHfJ_7^5fHaJY7^5jH`JV7^5mHaJR7_5oHaJQ7^5QIaJo6]5SIcJl6]5l1O1O0O2O1O0O101O0O1000000OeERLh8c5O1O1O1O1O1O1O1O1O1M3I7I7O100O100O10000O100O100O1000_MQHQLNAP8^4aH[KM1b7b4PJXKP6g4RJXKn5g4TJXKl5f4XJXKh5g4ZJXKf5d4_J[Ka5d4bJZK^5d4eJ\\KZ5c4hJ\\KX5b4jJ^KV5a4lJ^KT5`4oJ_KQ5`4PK`KP5^4SKaKm4^4UKaKk4S1mGT2Z3iLi4R1nGT2Z3jLh4P1PHU2Y3kLh4m0QHV2Z3lLf4l0QHW2c;eM`DX2c;eM_DZ2c;cM_D\\2c;bM^D\\2f;`M]D]2i;]MYDa2m;ZMTDd2e<N2N3M2N2N3M2N3M2N3M3M3J7J5K6I6K6J5K5K5K4LeWS7"}, "label": "a baseball umpire stands behind a catcher on a baseball field"}]}
{"id": 410165, "image": "COCO_train2014_000000410165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden flatbed and tire in front of a jeep\"."}], "task": "refering", "answer_template": "The \"a wooden flatbed and tire in front of a jeep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 279, 280, 299, 322], "bbox": [0.0, 0.5640416666666667, 0.1938125, 0.8539375], "iscrowd": 0, "area": 8256.024400000002, "rle": {"size": [480, 640], "counts": "m8o1P=o1RN0O1O1O0001O100O1O1O100O1O00100O1O1N2N2N2O1N1O2N2fNbDPO`;o0iDgNY;W1PEaNQ;^1WEYNk:e1Y1O1N12O0000001O0000001N1000001O0000O100000O01000000O1000000O0100000O1000000O10O1000O100001O00000O101O000000001O000O10001O001O0010O4M2M4M2M3N3L7J5J7J5J3M001O00001O001O0O100000000000005K8H^aa7"}, "label": "a wooden flatbed and tire in front of a jeep"}]}
{"id": 410165, "image": "COCO_train2014_000000410165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck whell that is in front of the blue car to the right of the horses\"."}], "task": "refering", "answer_template": "The \"the truck whell that is in front of the blue car to the right of the horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 279, 280, 299, 322], "bbox": [0.0, 0.5640416666666667, 0.1938125, 0.8539375], "iscrowd": 0, "area": 8256.024400000002, "rle": {"size": [480, 640], "counts": "m8o1P=o1RN0O1O1O0001O100O1O1O100O1O00100O1O1N2N2N2O1N1O2N2fNbDPO`;o0iDgNY;W1PEaNQ;^1WEYNk:e1Y1O1N12O0000001O0000001N1000001O0000O100000O01000000O1000000O0100000O1000000O10O1000O100001O00000O101O000000001O000O10001O001O0010O4M2M4M2M3N3L7J5J7J5J3M001O00001O001O0O100000000000005K8H^aa7"}, "label": "the truck whell that is in front of the blue car to the right of the horses"}]}
{"id": 410165, "image": "COCO_train2014_000000410165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that is on the left\"."}], "task": "refering", "answer_template": "The \"the horse that is on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 117, 118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 192, 193, 210, 211, 212, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 263, 264, 265, 282, 283, 284, 286, 287, 288, 305, 306, 307, 309, 311, 328, 329, 330, 332, 351, 352], "bbox": [0.10321875, 0.27095833333333336, 0.55, 0.9440833333333334], "iscrowd": 0, "area": 26997.92885, "rle": {"size": [480, 640], "counts": "nRo02j>5K4M4O0O2O1N101N101N101N101N101N2O001N107H;cB=i:n1O0O101O0O101N10001N101N10001O00001O0O101O00001O0000O10O01O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O1O1O1M3N2N2M4M2M3N2M3N2M3N2M3N2M3N2M3m0SO3M3M3M3N2M3M3N2M3M4M4K6J5L4K5L5J8Ho0eIdIU2_6fMcIY2`6bMbI^2b6]M`Ic2b6YM_Ig2e6SM^Il2f6oL[IR3g6iL\\IV3h6eLYI[3k6_LWIb3e2PKd0Z1iLg3Z2ZKk0l0lLk3Q2dKP1>RMn3i1mKR12VMR4d1SLS1H\\MU4]1XLW1@]MX4Y1\\LZ1YO_MZ4U1aL]1RO_M]4R1dL_1lNaM_4n0iLb1eNaMb4j0mLe1_NbMi4`0mLo1VNcMX50eL^2QNcMd5A`Lm2hMdMS6POZL\\3aMeMZ9\\2bFfM_9Z2_FgMb9Y2\\FhMc9k300000000O010000000O1000BhFTKX9i4`0N3M3M3M3M2N3O1UO[E\\Ma:e2aE[MZ:g2iEYMR:j2PFWMj9k2YFUMb9m2aFTM[9m2gFSMX9k2kFUMS9k2oFVMo8i2SGWMk8i2WGXMg8g2[GYMc8g2_GYM`8f2bG[M[8i0fF3Q1TOX8j0kFLo0ZOT8l0RGCl0@S8m0UG]Oj0ER8n0YGVOg0Lo7o0^GoNe01n7o0cGiNa08l7o0gGcN?=j7Q1lG[N<d0h7Q1PHUN:i0g7R1hIiNX6X1lIdNT6\\1QJ^NP6b1TJZNk5f1[JTNf5l1g30O13M5K5K5K5K7H7J5K2N1O2N2N1O2M3N1Oci4N\\VK4L4K5L4L4K5L4L4K5L4L4]O`NaCd1\\<bN]Cb1_<b0M3M3M0011O0O101O0O2O0M3[Of0G9I7I7I7@`0O100000000000000000000000000000000000000000000001O0000U1kN8I5LO000O3N3M2N3La0@:F6J5J5KecV4"}, "label": "the horse that is on the left"}]}
{"id": 410165, "image": "COCO_train2014_000000410165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse which is not visible fully\"."}], "task": "refering", "answer_template": "The \"the horse which is not visible fully\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 117, 118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 192, 193, 210, 211, 212, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 263, 264, 265, 282, 283, 284, 286, 287, 288, 305, 306, 307, 309, 311, 328, 329, 330, 332, 351, 352], "bbox": [0.10321875, 0.27095833333333336, 0.55, 0.9440833333333334], "iscrowd": 0, "area": 26997.92885, "rle": {"size": [480, 640], "counts": "nRo02j>5K4M4O0O2O1N101N101N101N101N101N2O001N107H;cB=i:n1O0O101O0O101N10001N101N10001O00001O0O101O00001O0000O10O01O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O1O1O1M3N2N2M4M2M3N2M3N2M3N2M3N2M3N2M3m0SO3M3M3M3N2M3M3N2M3M4M4K6J5L4K5L5J8Ho0eIdIU2_6fMcIY2`6bMbI^2b6]M`Ic2b6YM_Ig2e6SM^Il2f6oL[IR3g6iL\\IV3h6eLYI[3k6_LWIb3e2PKd0Z1iLg3Z2ZKk0l0lLk3Q2dKP1>RMn3i1mKR12VMR4d1SLS1H\\MU4]1XLW1@]MX4Y1\\LZ1YO_MZ4U1aL]1RO_M]4R1dL_1lNaM_4n0iLb1eNaMb4j0mLe1_NbMi4`0mLo1VNcMX50eL^2QNcMd5A`Lm2hMdMS6POZL\\3aMeMZ9\\2bFfM_9Z2_FgMb9Y2\\FhMc9k300000000O010000000O1000BhFTKX9i4`0N3M3M3M3M2N3O1UO[E\\Ma:e2aE[MZ:g2iEYMR:j2PFWMj9k2YFUMb9m2aFTM[9m2gFSMX9k2kFUMS9k2oFVMo8i2SGWMk8i2WGXMg8g2[GYMc8g2_GYM`8f2bG[M[8i0fF3Q1TOX8j0kFLo0ZOT8l0RGCl0@S8m0UG]Oj0ER8n0YGVOg0Lo7o0^GoNe01n7o0cGiNa08l7o0gGcN?=j7Q1lG[N<d0h7Q1PHUN:i0g7R1hIiNX6X1lIdNT6\\1QJ^NP6b1TJZNk5f1[JTNf5l1g30O13M5K5K5K5K7H7J5K2N1O2N2N1O2M3N1Oci4N\\VK4L4K5L4L4K5L4L4K5L4L4]O`NaCd1\\<bN]Cb1_<b0M3M3M0011O0O101O0O2O0M3[Of0G9I7I7I7@`0O100000000000000000000000000000000000000000000001O0000U1kN8I5LO000O3N3M2N3La0@:F6J5J5KecV4"}, "label": "the horse which is not visible fully"}]}
{"id": 410165, "image": "COCO_train2014_000000410165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue jeep cherokee parked next to the two horses\"."}], "task": "refering", "answer_template": "The \"the blue jeep cherokee parked next to the two horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 212, 233, 234, 235, 256, 257, 258, 259, 260, 261, 262, 266, 267, 281, 282, 283, 284, 285, 289, 305], "bbox": [0.07025, 0.5402708333333333, 0.62034375, 0.7826458333333333], "iscrowd": 0, "area": 6569.382750000001, "rle": {"size": [480, 640], "counts": "b[e01n\\10QRO000O10000O10O10O100O10000O1000O010000O10000O1000O010000O100O10000O01000O10000O11O1O1N101O1O0012N2M3N3M5J5K6J6J6J5K5K00000000001N1O1O1O1O1O1O2O0001O02O1N3M3N2M3M0000O10000000000eN[C:f<FZC:f<F[C9e<G\\C8d<H]C7c<I^C5c<K^C4b<L^C4b<L_C3a<M`C2`<NaC1_<ObCO_<1aCO_<1aCO_<1aCO_<1aCO_<1aCN`<1aCO_<0cCO^<OcC1a<J`C6g<BZC>P=WOQCj0b=6J5K6J6J5L2M1O20O3N1N3M3M3L4L5K5K4L5K4L5K5K4L5LnQ67cmI:F:J5N000O10O1000O01K5I7FX_8GXaG3M2N3M3M3M2N3M3M30O1000O10O1000000O01000O1000O10O1000000O010O1O1N2N2O1N2M`h`1M^W_NGiA<R>FnA:R>FnA;Q>EoA;P>FPB;o=DRB<n=DRB<m=ERB<n=DRB<n=DRB;n=FRB:n=FRB:n=ESB:m=GSB9m=GSB9m=GSB8m=ISB7m=ISB7m=ISB6m=JTB6l=JTB2P>NPBOR>2nAKU>5;0Sca3"}, "label": "the blue jeep cherokee parked next to the two horses"}]}
{"id": 410165, "image": "COCO_train2014_000000410165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue car behind the horse\"."}], "task": "refering", "answer_template": "The \"the blue car behind the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 212, 233, 234, 235, 256, 257, 258, 259, 260, 261, 262, 266, 267, 281, 282, 283, 284, 285, 289, 305], "bbox": [0.07025, 0.5402708333333333, 0.62034375, 0.7826458333333333], "iscrowd": 0, "area": 6569.382750000001, "rle": {"size": [480, 640], "counts": "b[e01n\\10QRO000O10000O10O10O100O10000O1000O010000O10000O1000O010000O100O10000O01000O10000O11O1O1N101O1O0012N2M3N3M5J5K6J6J6J5K5K00000000001N1O1O1O1O1O1O2O0001O02O1N3M3N2M3M0000O10000000000eN[C:f<FZC:f<F[C9e<G\\C8d<H]C7c<I^C5c<K^C4b<L^C4b<L_C3a<M`C2`<NaC1_<ObCO_<1aCO_<1aCO_<1aCO_<1aCO_<1aCN`<1aCO_<0cCO^<OcC1a<J`C6g<BZC>P=WOQCj0b=6J5K6J6J5L2M1O20O3N1N3M3M3L4L5K5K4L5K4L5K5K4L5LnQ67cmI:F:J5N000O10O1000O01K5I7FX_8GXaG3M2N3M3M3M2N3M3M30O1000O10O1000000O01000O1000O10O1000000O010O1O1N2N2O1N2M`h`1M^W_NGiA<R>FnA:R>FnA;Q>EoA;P>FPB;o=DRB<n=DRB<m=ERB<n=DRB<n=DRB;n=FRB:n=FRB:n=ESB:m=GSB9m=GSB9m=GSB8m=ISB7m=ISB7m=ISB6m=JTB6l=JTB2P>NPBOR>2nAKU>5;0Sca3"}, "label": "the blue car behind the horse"}]}
{"id": 123462, "image": "COCO_train2014_000000123462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfer riding the wave\"."}], "task": "refering", "answer_template": "The \"the surfer riding the wave\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 100, 101, 102, 103, 122, 123, 124, 125, 145, 146, 147, 148, 149, 170, 171, 172, 193], "bbox": [0.334296875, 0.25451990632318505, 0.504546875, 0.5562060889929742], "iscrowd": 0, "area": 5010.633449999998, "rle": {"size": [427, 640], "counts": "d\\i2;n<5K4L4L4L4L4L3M2O2N1O2M3N1O1O2N1O1O1O1O1O101N1O1O1N2O100000000O0N3UFnMY8V2eGlMV8Y2cGeMPO4[9Y2mFfMK9LJ\\9W2eFkN[9c2O1000O100mLeFBOV2\\9XNfFAOV2[9YNfFAOU2\\9ZNfF@Ob1<WNP9f0eFA0]1R:SOmE@1`02IS:7jEA1>9_OQ:`0gEB0>P;LSEFM>Q;HVEIJ>R;FUELJ=R;]O^E6_O=n;CSD<l;ETD;l;EUD:k;FVD:i;FWD:i;FXD9g;GZD9f;G[D9d;G]D:a;E`D<_;D`D>_;AbD`0\\;AdDLD=f;FgDMD=d;FiDLD=c;GiDMD<b;FkDND;`;HlDND:_;GoDNB;_;GoDOB:];GREOB9\\;GTEOA:Y;HVENC9V;HXEOC8T;JYEME9P;JnE6P:KQF5l9LVF3i9NXF2f9O[F1d9N_F1_90bF0^9OcF1]9MeF3Y9NhF1W9NlF2R9NPG2o8MSG3l8LVG4h8LYG5f8JZG8f8FZG<e8B\\G?h:0O01OO0001O2N2O2N1N5KooS4"}, "label": "the surfer riding the wave"}]}
{"id": 123462, "image": "COCO_train2014_000000123462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on yellow surf board\"."}], "task": "refering", "answer_template": "The \"man on yellow surf board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 100, 101, 102, 103, 122, 123, 124, 125, 145, 146, 147, 148, 149, 170, 171, 172, 193], "bbox": [0.334296875, 0.25451990632318505, 0.504546875, 0.5562060889929742], "iscrowd": 0, "area": 5010.633449999998, "rle": {"size": [427, 640], "counts": "d\\i2;n<5K4L4L4L4L4L3M2O2N1O2M3N1O1O2N1O1O1O1O1O101N1O1O1N2O100000000O0N3UFnMY8V2eGlMV8Y2cGeMPO4[9Y2mFfMK9LJ\\9W2eFkN[9c2O1000O100mLeFBOV2\\9XNfFAOV2[9YNfFAOU2\\9ZNfF@Ob1<WNP9f0eFA0]1R:SOmE@1`02IS:7jEA1>9_OQ:`0gEB0>P;LSEFM>Q;HVEIJ>R;FUELJ=R;]O^E6_O=n;CSD<l;ETD;l;EUD:k;FVD:i;FWD:i;FXD9g;GZD9f;G[D9d;G]D:a;E`D<_;D`D>_;AbD`0\\;AdDLD=f;FgDMD=d;FiDLD=c;GiDMD<b;FkDND;`;HlDND:_;GoDNB;_;GoDOB:];GREOB9\\;GTEOA:Y;HVENC9V;HXEOC8T;JYEME9P;JnE6P:KQF5l9LVF3i9NXF2f9O[F1d9N_F1_90bF0^9OcF1]9MeF3Y9NhF1W9NlF2R9NPG2o8MSG3l8LVG4h8LYG5f8JZG8f8FZG<e8B\\G?h:0O01OO0001O2N2O2N1N5KooS4"}, "label": "man on yellow surf board"}]}
{"id": 311890, "image": "COCO_train2014_000000311890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase that is sitting on the left side of the table\"."}], "task": "refering", "answer_template": "The \"a vase that is sitting on the left side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 54, 55, 56, 57, 72, 73, 74, 89, 90, 91, 106, 107, 108, 123, 124, 125, 140, 141, 142], "bbox": [0.21125000000000002, 0.11428125, 0.41904166666666665, 0.36103125], "iscrowd": 0, "area": 11154.9554, "rle": {"size": [640, 480], "counts": "cjo13mc02N2M4M2N2N3M2N5J6K5K5K6J5n_O_NZ=g1^B]Na=h1XB[Ng=k1PBYNo=m1hAWNW>o1`AUN_>Q2XARNg>U2PAoMo>W2h@mMW?Y2`@kM_?\\3O1O1N2O1O1O1O0O2O000000000O10000000000000O1000000000O10O1000000000000O1000000000O10O10000000000O100000000000O01jMl@@T?=WA[Oi>a0cAWO]>f0oAQOQ>k0[BmNf=o0cBkN]=Q1iBlNX=Q1mBmNS=P1RCnNn<n0XCoNj<m0[CQOe<k0`CSOa<i0eCUO[<f0lCWOU<d0RDZOn;b0XD[Oj;?]D_Oc;<S4K5LhY^5"}, "label": "a vase that is sitting on the left side of the table"}]}
{"id": 311890, "image": "COCO_train2014_000000311890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red vase on the left\"."}], "task": "refering", "answer_template": "The \"the red vase on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 54, 55, 56, 57, 72, 73, 74, 89, 90, 91, 106, 107, 108, 123, 124, 125, 140, 141, 142], "bbox": [0.21125000000000002, 0.11428125, 0.41904166666666665, 0.36103125], "iscrowd": 0, "area": 11154.9554, "rle": {"size": [640, 480], "counts": "cjo13mc02N2M4M2N2N3M2N5J6K5K5K6J5n_O_NZ=g1^B]Na=h1XB[Ng=k1PBYNo=m1hAWNW>o1`AUN_>Q2XARNg>U2PAoMo>W2h@mMW?Y2`@kM_?\\3O1O1N2O1O1O1O0O2O000000000O10000000000000O1000000000O10O1000000000000O1000000000O10O10000000000O100000000000O01jMl@@T?=WA[Oi>a0cAWO]>f0oAQOQ>k0[BmNf=o0cBkN]=Q1iBlNX=Q1mBmNS=P1RCnNn<n0XCoNj<m0[CQOe<k0`CSOa<i0eCUO[<f0lCWOU<d0RDZOn;b0XD[Oj;?]D_Oc;<S4K5LhY^5"}, "label": "the red vase on the left"}]}
{"id": 246364, "image": "COCO_train2014_000000246364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bull in front is wearing an earing that is red and white\"."}], "task": "refering", "answer_template": "The \"the bull in front is wearing an earing that is red and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 227, 228, 229, 230, 231, 232, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356, 365, 366, 367, 368, 369, 370, 371, 372, 373, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.28862499999999996, 0.531765625, 1.0, 0.988234375], "iscrowd": 0, "area": 67857.42270000002, "rle": {"size": [640, 480], "counts": "V[g2a0Vc0;M3M2N3M3M3N2M2N3M3M3M3M2N100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1\\NWMPBj2l=\\MRBd2j=bMSB_2i=gMSB[2i=kMRBW2k=oMoAU2n=PNmAS2R>PNiAS2W>PNbAS2^>PN]AS2b>PNYAS2g>c1O00100O10O01O100O00100O10O01O100O00100O10O01O100O10O01\\CaJg:_5YEcJf:\\5YEfJf:Y5[EiJc:W5\\ElJb:T5^EmJa:S5^EPK`:o4`ETK^:l4bEUK]:k4bEXK\\:h4dEZKZ:f4eE\\KZ:d4fE^KX:b4gEaKW:_4hEcKW:]4iEeKU:[4jEhKT:X4lEjKR:V4mElKR:T4nEnKP:S4nEnKR:R4nEnKR:R4mEoKS:Q4mEoKS:Q4mEoKT:P4lEPLT:P4kEQLU:o3kEQLU:o3kEQLU:o3kEQLU:o3jERLV:n3jERLV:n3jERLV:n3jERLV:n3iESLW:m3iESLW:m3iESLW:m3iESLW:m3hETLX:l3hETLX:l3hETLX:l3gEULZ:k3eEUL[:k3eEUL[:e4kD[KU;g61000000O10000O1000001N1000000O1000000O1000000O10001O0002N2N2O100O2N1O100O1O1O100O00O2TH^DV7c;iH^DV7c;iH]DW7d;hH]DW7e;gH\\DX7e;gH[DY7f;fH[DY7f;fH[DY7g;eHZDZ7g;eHYD[7h;dHYD[7h;dHYD[7i;cHWD]7j;bHWD]7j;bHWD]7V<01O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O0000O100O10000O100O10000O100O10000O10000O100000000000000O1000000000000O100000000000000O1000000000000O10000000000000000000000O10000000000000000000000000000001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O000010O0001O001O00000]OlBZJU=d4XDPKh;n3eE"}, "label": "the bull in front is wearing an earing that is red and white"}]}
{"id": 246364, "image": "COCO_train2014_000000246364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cow was sitting\"."}], "task": "refering", "answer_template": "The \"a black cow was sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 227, 228, 229, 230, 231, 232, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356, 365, 366, 367, 368, 369, 370, 371, 372, 373, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.28862499999999996, 0.531765625, 1.0, 0.988234375], "iscrowd": 0, "area": 67857.42270000002, "rle": {"size": [640, 480], "counts": "V[g2a0Vc0;M3M2N3M3M3N2M2N3M3M3M3M2N100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1\\NWMPBj2l=\\MRBd2j=bMSB_2i=gMSB[2i=kMRBW2k=oMoAU2n=PNmAS2R>PNiAS2W>PNbAS2^>PN]AS2b>PNYAS2g>c1O00100O10O01O100O00100O10O01O100O00100O10O01O100O10O01\\CaJg:_5YEcJf:\\5YEfJf:Y5[EiJc:W5\\ElJb:T5^EmJa:S5^EPK`:o4`ETK^:l4bEUK]:k4bEXK\\:h4dEZKZ:f4eE\\KZ:d4fE^KX:b4gEaKW:_4hEcKW:]4iEeKU:[4jEhKT:X4lEjKR:V4mElKR:T4nEnKP:S4nEnKR:R4nEnKR:R4mEoKS:Q4mEoKS:Q4mEoKT:P4lEPLT:P4kEQLU:o3kEQLU:o3kEQLU:o3kEQLU:o3jERLV:n3jERLV:n3jERLV:n3jERLV:n3iESLW:m3iESLW:m3iESLW:m3iESLW:m3hETLX:l3hETLX:l3hETLX:l3gEULZ:k3eEUL[:k3eEUL[:e4kD[KU;g61000000O10000O1000001N1000000O1000000O1000000O10001O0002N2N2O100O2N1O100O1O1O100O00O2TH^DV7c;iH^DV7c;iH]DW7d;hH]DW7e;gH\\DX7e;gH[DY7f;fH[DY7f;fH[DY7g;eHZDZ7g;eHYD[7h;dHYD[7h;dHYD[7i;cHWD]7j;bHWD]7j;bHWD]7V<01O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O0000O100O10000O100O10000O100O10000O10000O100000000000000O1000000000000O100000000000000O1000000000000O10000000000000000000000O10000000000000000000000000000001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O000010O0001O001O00000]OlBZJU=d4XDPKh;n3eE"}, "label": "a black cow was sitting"}]}
{"id": 442977, "image": "COCO_train2014_000000442977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard being propelled through the air\"."}], "task": "refering", "answer_template": "The \"a skateboard being propelled through the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246], "bbox": [0.2816458333333333, 0.4449375, 0.803, 0.6449374999999999], "iscrowd": 0, "area": 16189.11655, "rle": {"size": [640, 480], "counts": "]hd21ic07J5J6K6I6M301N100O101N1O101N100O101N1O101N100O2O0O1O101N100O2O0O100O2N100O2O0O101N1O100O2O0O101N1O010O010O10O01O10O0100O010O00100O010O10O01000O010O10O10O10O10O10O10O10O10O10O01000O01000O01000O01000O010O10O10O10O10O10O10O01000O01000O01000O01000O01000O01000O0100O01000O01000O01000O01000O01000O010000O01000O10O10O10000O0100000O0100000O0100000O10O1000O10O1000O1000O10O1000O10O100000O0100000O0100000O001O100O00100O1O010O1O10O01O100O00100O1O00100O1O2O0O2N101N1O2O0O2N101M2N3L3N3L3N3L3N3L3N3L3N3L3N3L3N3LhRk1"}, "label": "a skateboard being propelled through the air"}]}
{"id": 442977, "image": "COCO_train2014_000000442977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboard in the air beneath rider doing a trick\"."}], "task": "refering", "answer_template": "The \"skateboard in the air beneath rider doing a trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246], "bbox": [0.2816458333333333, 0.4449375, 0.803, 0.6449374999999999], "iscrowd": 0, "area": 16189.11655, "rle": {"size": [640, 480], "counts": "]hd21ic07J5J6K6I6M301N100O101N1O101N100O101N1O101N100O2O0O1O101N100O2O0O100O2N100O2O0O101N1O100O2O0O101N1O010O010O10O01O10O0100O010O00100O010O10O01000O010O10O10O10O10O10O10O10O10O10O01000O01000O01000O01000O010O10O10O10O10O10O10O01000O01000O01000O01000O01000O01000O0100O01000O01000O01000O01000O01000O010000O01000O10O10O10000O0100000O0100000O0100000O10O1000O10O1000O1000O10O1000O10O100000O0100000O0100000O001O100O00100O1O010O1O10O01O100O00100O1O00100O1O2O0O2N101N1O2O0O2N101M2N3L3N3L3N3L3N3L3N3L3N3L3N3L3N3LhRk1"}, "label": "skateboard in the air beneath rider doing a trick"}]}
{"id": 442977, "image": "COCO_train2014_000000442977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man jumping in the air , midway through a skateboard trick\"."}], "task": "refering", "answer_template": "The \"a man jumping in the air , midway through a skateboard trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 17, 23, 24, 25, 26, 27, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 177, 178, 179, 180, 194, 195, 196, 211], "bbox": [0.006, 0.006734374999999999, 0.6801458333333332, 0.532578125], "iscrowd": 0, "area": 46643.81489999999, "rle": {"size": [640, 480], "counts": "Tl18hc0<D;E3M00O100O1O100O100O100O100O1O100O100O100O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O100O10000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000XAKe95YF`0T9@iFU1e8kNYGe1Y8[NdGh1\\8XNbGk1]8TNaGo1_8QN_GR2`8nM^GU2a8kMiFLhL\\2_<hMhFOeL\\2b<eMgF3bL[2g<bMeF7`LZ2j<_MeFT3Z9lLdFV3\\9jLbFY3]9gLbF[3]9eLaF]3_9cL_F`3`9`L_Fa3a9_L]Fd3b9\\L\\Fg3c9YL\\Fh3d9XLZFk3e9ULYFn3f9RLYFo3g9QLWFR4h9nKWFS4i9mKVFU4i9kKVFW4i9iKVFX4j9hKUFZ4j9fKUF[4k9eKSF^4l9bKSF`4l9`KSFa4m9_KRFc4m9]KRFe4m9[KRFf4n9[KPFg4o9^KkEc4U:bKeE`4Z:dKaE^4^:gK[E[4e:iKVEY4i:lKQEV4n:nKmDS4S;PLiDR4V;QLfDP4Z;SLbDo3];SL`Do3_;TL]Dm3c;VLYDl3f;VLVDm3i;VLSDm3k;VLQDl3n;T2O1O2N1O1O1O2N1O1O1O2N1O1aEcGY9^8eFeGY9]8eFeGY9\\8eFgGY9Z8eFiGY9X8fFjGX9X8eFkGY9V8eFmGY9T8fFoGW9R8gFQHW9Q8fFRHX9o7gFSHW9n7gFUHW9l7gFWHW9k7gFWHW9j7gFVHZ9k7dFVH\\9k7bFUH_9V90O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100`JdE[1]:eNdEZ1\\:eNfEY1[:gNeEY1[:fNgEX1Z:hNgEW1Y:hNiEV1X:iNjEV1W:iNjEU1Y:hNiEW1Y:gNhEW1[:fNgEY1[:eNfEY1]:dNeE[1]:bNeE\\1^:bNbEg0[OcLU;c2bEi0YOcLX;b2`Ei0YOeLY;_2`Ek0WOeL\\;^2^El0VOfL^;[2^En0TOfLa;Y2]Eo0SOhLb;W2\\EP1ROhLe;U2[ER1POiLg;R2[ES1oNjLi;Q2YET1nNkLk;n1YEV1lNkLn;m1WEW1kNlLP<j1WEX1jNmLS<g1UE[1iNlLV<f1SE\\1gNnLY<b1RE_1eNnL]<`1oD`1eNPM_<Y1QEf1bNnLa<T1SEm1`NkL`<o0WET2_;dMgD[2\\;]MjDb2Y;VMmDi2V;oLPEo2R>M3L4K5L5J<E:E;F;D<E;DRQP3"}, "label": "a man jumping in the air , midway through a skateboard trick"}]}
{"id": 442977, "image": "COCO_train2014_000000442977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy doing a trick on a skateboard\"."}], "task": "refering", "answer_template": "The \"a boy doing a trick on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 17, 23, 24, 25, 26, 27, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 177, 178, 179, 180, 194, 195, 196, 211], "bbox": [0.006, 0.006734374999999999, 0.6801458333333332, 0.532578125], "iscrowd": 0, "area": 46643.81489999999, "rle": {"size": [640, 480], "counts": "Tl18hc0<D;E3M00O100O1O100O100O100O100O1O100O100O100O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O100O10000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000XAKe95YF`0T9@iFU1e8kNYGe1Y8[NdGh1\\8XNbGk1]8TNaGo1_8QN_GR2`8nM^GU2a8kMiFLhL\\2_<hMhFOeL\\2b<eMgF3bL[2g<bMeF7`LZ2j<_MeFT3Z9lLdFV3\\9jLbFY3]9gLbF[3]9eLaF]3_9cL_F`3`9`L_Fa3a9_L]Fd3b9\\L\\Fg3c9YL\\Fh3d9XLZFk3e9ULYFn3f9RLYFo3g9QLWFR4h9nKWFS4i9mKVFU4i9kKVFW4i9iKVFX4j9hKUFZ4j9fKUF[4k9eKSF^4l9bKSF`4l9`KSFa4m9_KRFc4m9]KRFe4m9[KRFf4n9[KPFg4o9^KkEc4U:bKeE`4Z:dKaE^4^:gK[E[4e:iKVEY4i:lKQEV4n:nKmDS4S;PLiDR4V;QLfDP4Z;SLbDo3];SL`Do3_;TL]Dm3c;VLYDl3f;VLVDm3i;VLSDm3k;VLQDl3n;T2O1O2N1O1O1O2N1O1O1O2N1O1aEcGY9^8eFeGY9]8eFeGY9\\8eFgGY9Z8eFiGY9X8fFjGX9X8eFkGY9V8eFmGY9T8fFoGW9R8gFQHW9Q8fFRHX9o7gFSHW9n7gFUHW9l7gFWHW9k7gFWHW9j7gFVHZ9k7dFVH\\9k7bFUH_9V90O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100`JdE[1]:eNdEZ1\\:eNfEY1[:gNeEY1[:fNgEX1Z:hNgEW1Y:hNiEV1X:iNjEV1W:iNjEU1Y:hNiEW1Y:gNhEW1[:fNgEY1[:eNfEY1]:dNeE[1]:bNeE\\1^:bNbEg0[OcLU;c2bEi0YOcLX;b2`Ei0YOeLY;_2`Ek0WOeL\\;^2^El0VOfL^;[2^En0TOfLa;Y2]Eo0SOhLb;W2\\EP1ROhLe;U2[ER1POiLg;R2[ES1oNjLi;Q2YET1nNkLk;n1YEV1lNkLn;m1WEW1kNlLP<j1WEX1jNmLS<g1UE[1iNlLV<f1SE\\1gNnLY<b1RE_1eNnL]<`1oD`1eNPM_<Y1QEf1bNnLa<T1SEm1`NkL`<o0WET2_;dMgD[2\\;]MjDb2Y;VMmDi2V;oLPEo2R>M3L4K5L5J<E:E;F;D<E;DRQP3"}, "label": "a boy doing a trick on a skateboard"}]}
{"id": 402020, "image": "COCO_train2014_000000402020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chicken burger\"."}], "task": "refering", "answer_template": "The \"a chicken burger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.305859375, 0.44957845433255267, 0.9340624999999999, 0.8473302107728338], "iscrowd": 0, "area": 51387.6641, "rle": {"size": [427, 640], "counts": "[Pb22X=3M2N3N1N3M3M2N3M2N3M3M2N3M3M2N3N1N3M3M2N3M2N3M3M2N2N2N1O2N2O1N2N2N2N1O1O2N1O1O2N1O1O101N1O1O2N1O100O2N100O2N100O101N1O101N1O100O2O0O1O2O0O100O2N100O2N100O101N1O101N1O100O2O0O1O2O0O100O2N100O2N100O101N1O101N10000O2O000O2O00000O2O000O2O000O101O000O10000O1000O10O10000O1000000O10000O10000O1000000O10000O1000000O01000O10000000000000000000000001O0000000000000000000O10000000000000000000001O000000000000000000000000000O100000000000001O00000000000000000000000000000000000O10001O0000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000001O000000000O1000001O00001O000O101O0O101O0O10001N10001N10000O2O0O101O0O101O0O10001N10001N10000O2O0O101O0O101O0O10001N10001N10000O2O000O2O0O101O0O10001N10001N10000O2O1O2M2O2N1N3N1N3N1O2mMdER1^:jNfEU1\\:gNhEW1Y:fNkEW1X:dNmEZ1T:cNoE\\1S:`NQF]1Q:_NTF_1m:O2M3N2N1N3N2M2O2N2M3N1O2M3N1O2M3N2N1N3N2N1N3NbVa0"}, "label": "a chicken burger"}]}
{"id": 402020, "image": "COCO_train2014_000000402020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwhich with the inside showing\"."}], "task": "refering", "answer_template": "The \"half of a sandwhich with the inside showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.305859375, 0.44957845433255267, 0.9340624999999999, 0.8473302107728338], "iscrowd": 0, "area": 51387.6641, "rle": {"size": [427, 640], "counts": "[Pb22X=3M2N3N1N3M3M2N3M2N3M3M2N3M3M2N3N1N3M3M2N3M2N3M3M2N2N2N1O2N2O1N2N2N2N1O1O2N1O1O2N1O1O101N1O1O2N1O100O2N100O2N100O101N1O101N1O100O2O0O1O2O0O100O2N100O2N100O101N1O101N1O100O2O0O1O2O0O100O2N100O2N100O101N1O101N10000O2O000O2O00000O2O000O2O000O101O000O10000O1000O10O10000O1000000O10000O10000O1000000O10000O1000000O01000O10000000000000000000000001O0000000000000000000O10000000000000000000001O000000000000000000000000000O100000000000001O00000000000000000000000000000000000O10001O0000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000001O000000000O1000001O00001O000O101O0O101O0O10001N10001N10000O2O0O101O0O101O0O10001N10001N10000O2O0O101O0O101O0O10001N10001N10000O2O000O2O0O101O0O10001N10001N10000O2O1O2M2O2N1N3N1N3N1O2mMdER1^:jNfEU1\\:gNhEW1Y:fNkEW1X:dNmEZ1T:cNoE\\1S:`NQF]1Q:_NTF_1m:O2M3N2N1N3N2M2O2N2M3N1O2M3N1O2M3N2N1N3N2N1N3NbVa0"}, "label": "half of a sandwhich with the inside showing"}]}
{"id": 402020, "image": "COCO_train2014_000000402020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hidden second half burger on the plate\"."}], "task": "refering", "answer_template": "The \"a hidden second half burger on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 190, 191, 192, 193, 212, 213, 214, 215, 235, 236, 237, 258, 259, 260, 261, 283, 284, 285, 286], "bbox": [0.219390625, 0.4005152224824356, 0.6401406249999999, 0.8487353629976581], "iscrowd": 0, "area": 12392.882999999998, "rle": {"size": [427, 640], "counts": "_ej1:o<2M4L3N3N2O0O2O1N1O2O0O1O100O1O100O1O100O1O1O1M3L4M3L4M3L4L4N2N4M2N2N2N3M2M3N2N2N3M2N2N2M3N2N1O2N1O2M3N1O2N1O2N1N\\N^FD`9<cFC[9>gF@X9`0kF_OS9a0PG]Oo8c0SG]Ok8d0WGZOg8g0\\GXOb8h0aGVO^8j0eGTOZ8l0iGSOU8n0mGPOR8P1PHoNo7Q1THmNl7R1WHlNh7T1[HkNd7U1^HiNb7V1aHhN^7X1eHfN[7Y1hHeNW7[1kHdNU7\\1mHcNR7\\1QIbNn6^1UI`Nk6_1XI_Ng6a1\\I]Nd6c1^I[Nb6d1`I\\N^6d1eIZN[6e1hIYNW6h1iIXNW6g1jIYNV6f1kI[NS6e1oIZNQ6e1PJ[No5f1QJZNo5e1RJ[Nn5d1TJ[Nk5e1VJ\\Ni5c1XJ]Ng5d1YJ\\Ng5c1[J\\Ne5c1\\J]Nc5c1^J]Nb5b1`J^N^5c1bJ]N^5b1cJ^N]5a1dJ_N[5a1gJ^NY5a1hJ_NW5a1jJ`NU5_1lJaNT5^1nJaNQ5_1PKaNP5^1QKbNn4^1TKbNk4]1VKcNj4\\1WKdNh4\\1YKdNg4[1[KdNd4\\1]KdNc4[1^KfNa4Y1`KgN_4Y1cKfN]4Y1dKgN[4Y1fKgNZ4X1hKgNX4Y1hKhNV4X1kKhNU4X1kKhNU4W1mKhNS4X1mKhNR4X1oKiNP4W1PLiNP4W1QLhNo3W1RLiNn3W1RLiNm3W1ULhNk3X1ULiNj3V1WLjNi3V1WLjNh3V1ZLiNf3W1ZLiNf3V1[LjNe3V1[LkNd3T1^LkNa3V1_LjNa3U1`LkN`3U1aLjN_3U1bLkN]3V1cLkN\\3T1eLlN[3T1fLkNZ3U1fLkNZ3T1gLlNX3U1hLlNW3S1kLlNU3T1kLlNU3S1lLmNS3T1nLkNR3T1oLlNQ3T1oLmNP3R1QMnNo2R1RMmNm2S1TMmNl2S1TMmNl2R1UMnNk2R1VMnNh2R1YMnNg2R1YMnNg2Q1ZMoNf2Q1j40aHoNc2Q1^MoNb2Q1l4O10O10O1000001N10000O10001O0O10000O101O000O10000O2O00000O10001N1000000O101O0O1000000O2O00000000001O000000001O00000000010O00000000001O000000001O00001O001O001O00001O001O001O001O001O00001O001O00gaP3"}, "label": "a hidden second half burger on the plate"}]}
{"id": 402020, "image": "COCO_train2014_000000402020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hidden half of the sandwich\"."}], "task": "refering", "answer_template": "The \"hidden half of the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 190, 191, 192, 193, 212, 213, 214, 215, 235, 236, 237, 258, 259, 260, 261, 283, 284, 285, 286], "bbox": [0.219390625, 0.4005152224824356, 0.6401406249999999, 0.8487353629976581], "iscrowd": 0, "area": 12392.882999999998, "rle": {"size": [427, 640], "counts": "_ej1:o<2M4L3N3N2O0O2O1N1O2O0O1O100O1O100O1O100O1O1O1M3L4M3L4M3L4L4N2N4M2N2N2N3M2M3N2N2N3M2N2N2M3N2N1O2N1O2M3N1O2N1O2N1N\\N^FD`9<cFC[9>gF@X9`0kF_OS9a0PG]Oo8c0SG]Ok8d0WGZOg8g0\\GXOb8h0aGVO^8j0eGTOZ8l0iGSOU8n0mGPOR8P1PHoNo7Q1THmNl7R1WHlNh7T1[HkNd7U1^HiNb7V1aHhN^7X1eHfN[7Y1hHeNW7[1kHdNU7\\1mHcNR7\\1QIbNn6^1UI`Nk6_1XI_Ng6a1\\I]Nd6c1^I[Nb6d1`I\\N^6d1eIZN[6e1hIYNW6h1iIXNW6g1jIYNV6f1kI[NS6e1oIZNQ6e1PJ[No5f1QJZNo5e1RJ[Nn5d1TJ[Nk5e1VJ\\Ni5c1XJ]Ng5d1YJ\\Ng5c1[J\\Ne5c1\\J]Nc5c1^J]Nb5b1`J^N^5c1bJ]N^5b1cJ^N]5a1dJ_N[5a1gJ^NY5a1hJ_NW5a1jJ`NU5_1lJaNT5^1nJaNQ5_1PKaNP5^1QKbNn4^1TKbNk4]1VKcNj4\\1WKdNh4\\1YKdNg4[1[KdNd4\\1]KdNc4[1^KfNa4Y1`KgN_4Y1cKfN]4Y1dKgN[4Y1fKgNZ4X1hKgNX4Y1hKhNV4X1kKhNU4X1kKhNU4W1mKhNS4X1mKhNR4X1oKiNP4W1PLiNP4W1QLhNo3W1RLiNn3W1RLiNm3W1ULhNk3X1ULiNj3V1WLjNi3V1WLjNh3V1ZLiNf3W1ZLiNf3V1[LjNe3V1[LkNd3T1^LkNa3V1_LjNa3U1`LkN`3U1aLjN_3U1bLkN]3V1cLkN\\3T1eLlN[3T1fLkNZ3U1fLkNZ3T1gLlNX3U1hLlNW3S1kLlNU3T1kLlNU3S1lLmNS3T1nLkNR3T1oLlNQ3T1oLmNP3R1QMnNo2R1RMmNm2S1TMmNl2S1TMmNl2R1UMnNk2R1VMnNh2R1YMnNg2R1YMnNg2Q1ZMoNf2Q1j40aHoNc2Q1^MoNb2Q1l4O10O10O1000001N10000O10001O0O10000O101O000O10000O2O00000O10001N1000000O101O0O1000000O2O00000000001O000000001O00000000010O00000000001O000000001O00001O001O001O00001O001O001O001O001O00001O001O00gaP3"}, "label": "hidden half of the sandwich"}]}
{"id": 25192, "image": "COCO_train2014_000000025192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass right top corner\"."}], "task": "refering", "answer_template": "The \"glass right top corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 113], "bbox": [0.788890625, 0.0, 0.996296875, 0.2543125], "iscrowd": 0, "area": 11235.631050000004, "rle": {"size": [480, 640], "counts": "ci\\74h>4L5K4K5L5K4K6K4L4L5J5L5K4K5L5K4L5J5L4N3O001O00001O001O00001O00001O001O00001O00001O00001O001O00001O01O01O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O000010O0N2L4M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M3N2MQn0"}, "label": "glass right top corner"}]}
{"id": 25192, "image": "COCO_train2014_000000025192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of water on the table next to a plate of meat with salad on top\"."}], "task": "refering", "answer_template": "The \"a glass of water on the table next to a plate of meat with salad on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 113], "bbox": [0.788890625, 0.0, 0.996296875, 0.2543125], "iscrowd": 0, "area": 11235.631050000004, "rle": {"size": [480, 640], "counts": "ci\\74h>4L5K4K5L5K4K6K4L4L5J5L5K4K5L5K4L5J5L4N3O001O00001O001O00001O00001O001O00001O00001O00001O001O00001O01O01O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O000010O0N2L4M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M3N2MQn0"}, "label": "a glass of water on the table next to a plate of meat with salad on top"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white sheep\"."}], "task": "refering", "answer_template": "The \"white sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 259, 260, 261, 262, 279, 280, 284, 285, 302, 303, 307, 308], "bbox": [0.10760937500000001, 0.4172131147540984, 0.58684375, 0.9290398126463701], "iscrowd": 0, "area": 32439.545, "rle": {"size": [427, 640], "counts": "aPm06Q=9G8H9G9G7I7I8I6M3L4M3M4K4M3L6K5K5J5QGcLh7_3SHfLj7^3PHhLm7Z3nGkLP8X3kGmLQ8W3iGnLU8T3fGRMV8S4O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1NeNjHbLW7Z3oHdLQ7X3UIfLk6V3\\IfLe6V3aIhL_6U3fIiLZ6U3jIhLW6U3nIiLR6U3QJjLo5U3TJhLm5W3UJhLk5W3XJgLh5Y3YJeLh5Z3[JdLe5\\3^JaLb5^3bJ^L_5a3V200O100O1O100O100O1O100O1000000000000001O00000000000001O000000000000000000000001O000000000000000001O0000001O01O000001O0000001O000000001O0001O01O000000001O0000001O0000010O00000001N3N2N2N2N2N2M4M5K4L4L4L3L4M3M3M3M3L4M4L4L5K2N1N2O1O1O1O1O1N2O2NO10000O01000]NhHRMY7_2bIUM^6d2UJoLk5Q3ZJkLf5S3^JkLb5T3aJiL`5V3bJiL^5V3eJhL[5W3\\2O2O1O1O0O2O001O0O101O0O2O00001N101O00010O001O001O00001O010O001O001O001O1O001O001O001O001O1O001O001O001O1O001O1O002N2N1O2N2N2N2N1O2N2N2N10100O1O010O1O1O010O1O10O01O1O100O001O101N1N3N1O2N1N3N1O2N1O2M2O2N2N2M3N3M2N2M4M2MdX^3"}, "label": "white sheep"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white sheep\"."}], "task": "refering", "answer_template": "The \"the white sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 259, 260, 261, 262, 279, 280, 284, 285, 302, 303, 307, 308], "bbox": [0.10760937500000001, 0.4172131147540984, 0.58684375, 0.9290398126463701], "iscrowd": 0, "area": 32439.545, "rle": {"size": [427, 640], "counts": "aPm06Q=9G8H9G9G7I7I8I6M3L4M3M4K4M3L6K5K5J5QGcLh7_3SHfLj7^3PHhLm7Z3nGkLP8X3kGmLQ8W3iGnLU8T3fGRMV8S4O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1NeNjHbLW7Z3oHdLQ7X3UIfLk6V3\\IfLe6V3aIhL_6U3fIiLZ6U3jIhLW6U3nIiLR6U3QJjLo5U3TJhLm5W3UJhLk5W3XJgLh5Y3YJeLh5Z3[JdLe5\\3^JaLb5^3bJ^L_5a3V200O100O1O100O100O1O100O1000000000000001O00000000000001O000000000000000000000001O000000000000000001O0000001O01O000001O0000001O000000001O0001O01O000000001O0000001O0000010O00000001N3N2N2N2N2N2M4M5K4L4L4L3L4M3M3M3M3L4M4L4L5K2N1N2O1O1O1O1O1N2O2NO10000O01000]NhHRMY7_2bIUM^6d2UJoLk5Q3ZJkLf5S3^JkLb5T3aJiL`5V3bJiL^5V3eJhL[5W3\\2O2O1O1O0O2O001O0O101O0O2O00001N101O00010O001O001O00001O010O001O001O001O1O001O001O001O001O1O001O001O001O1O001O1O002N2N1O2N2N2N2N1O2N2N2N10100O1O010O1O1O010O1O10O01O1O100O001O101N1N3N1O2N1N3N1O2N1O2M2O2N2N2M3N3M2N2M4M2MdX^3"}, "label": "the white sheep"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with white hair and wearing yellow , white , blue shirt and shot\"."}], "task": "refering", "answer_template": "The \"a boy with white hair and wearing yellow , white , blue shirt and shot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 38, 39, 40, 41, 61, 62, 63, 64, 84, 85, 86, 87, 108, 109, 110, 130, 131, 132, 133, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 291, 292, 293, 294, 315, 316], "bbox": [0.58621875, 0.04494145199063232, 0.835109375, 0.932576112412178], "iscrowd": 0, "area": 30237.482049999988, "rle": {"size": [427, 640], "counts": "dbl41Y=3N2M3M3N2M4M2M3N2M3M3N2M2OO01O100O10O0100O1O10O0100\\DkNQ;V1kDnNT;R1iDROW;n0eDVOZ;k0bDYO^;[1N100O1O100O100O1O1O1N2O001O000[LkMTLW2j3mMSLS2l3QNQLo1n3VNnKk1DgM_1b0jNV2YOXMk1f0hNR2^OZMf1i0iNm1BZMc1n0gNg1H\\M^1Q1gNc1K]M\\1T1fN^10^MX1Y1cNZ16^MT1\\1cNU1;_MQ1_1aNR1>aMo0`1`Nn0c0bMl0c1^Nk0g0bMj0f1[Nh0l0dMg0g1ZNe0P1dMd0k1YN`0U1fM`0m1XN=Y1fM>o1WN:\\1hMRNEW1\\2WO8a1hMoMEX1^2UO4f1iMmMFV1`2RO3k1hMlMFV1b2nN1R2gMiMGV1c2jN2X2dMhMGT1g2gNO^2dMfMHS1g2eN0b2aMfMHR1j2aN3d2[MhMIR1k2^N7e2UMjMKQ1k2[N;g2PMmMJP1n2WN>i2jLoMKP1i6n0\\IQNLo0m6l0XIUNLn0n6k0VIVNMo0P7h0SIYNMo0S7d0QI\\NNP1S7a0oH^NOQ1U7>lHaNOQ1W7<jHbN0R1Y78hHeN1R1Z76eHhN1R1b7N\\HPO3R1j7DTHYO4R1Q8\\OkGB4R1Y8TOcGI5S1a8jN[G34S1j8aNRG;6T1a9kN`FU1`9kN`FU1`9jNbFU1j:000001O000010O000O1O1O1O2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1\\OaM[Fa2Z1^MY69YH[2\\1_MU6=[HU2^1bMQ6?^HQ2_1cMl5`3RJcLo4ETJl3k0cLo4FmIn3P1_LP5JfIn3V1[LR5Y4jJkKT5X4hJkKV5X4fJkKW5Y4dJkKZ5W4cJlK[5W4aJlK]5W4^JmK`5V4\\JmKa5W4[JlKc5P6N2N2N2M3N2N2N8G;F:F:F:F:E<Emo[1"}, "label": "a boy with white hair and wearing yellow , white , blue shirt and shot"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toddler playing with a goat\"."}], "task": "refering", "answer_template": "The \"a toddler playing with a goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 38, 39, 40, 41, 61, 62, 63, 64, 84, 85, 86, 87, 108, 109, 110, 130, 131, 132, 133, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 291, 292, 293, 294, 315, 316], "bbox": [0.58621875, 0.04494145199063232, 0.835109375, 0.932576112412178], "iscrowd": 0, "area": 30237.482049999988, "rle": {"size": [427, 640], "counts": "dbl41Y=3N2M3M3N2M4M2M3N2M3M3N2M2OO01O100O10O0100O1O10O0100\\DkNQ;V1kDnNT;R1iDROW;n0eDVOZ;k0bDYO^;[1N100O1O100O100O1O1O1N2O001O000[LkMTLW2j3mMSLS2l3QNQLo1n3VNnKk1DgM_1b0jNV2YOXMk1f0hNR2^OZMf1i0iNm1BZMc1n0gNg1H\\M^1Q1gNc1K]M\\1T1fN^10^MX1Y1cNZ16^MT1\\1cNU1;_MQ1_1aNR1>aMo0`1`Nn0c0bMl0c1^Nk0g0bMj0f1[Nh0l0dMg0g1ZNe0P1dMd0k1YN`0U1fM`0m1XN=Y1fM>o1WN:\\1hMRNEW1\\2WO8a1hMoMEX1^2UO4f1iMmMFV1`2RO3k1hMlMFV1b2nN1R2gMiMGV1c2jN2X2dMhMGT1g2gNO^2dMfMHS1g2eN0b2aMfMHR1j2aN3d2[MhMIR1k2^N7e2UMjMKQ1k2[N;g2PMmMJP1n2WN>i2jLoMKP1i6n0\\IQNLo0m6l0XIUNLn0n6k0VIVNMo0P7h0SIYNMo0S7d0QI\\NNP1S7a0oH^NOQ1U7>lHaNOQ1W7<jHbN0R1Y78hHeN1R1Z76eHhN1R1b7N\\HPO3R1j7DTHYO4R1Q8\\OkGB4R1Y8TOcGI5S1a8jN[G34S1j8aNRG;6T1a9kN`FU1`9kN`FU1`9jNbFU1j:000001O000010O000O1O1O1O2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1\\OaM[Fa2Z1^MY69YH[2\\1_MU6=[HU2^1bMQ6?^HQ2_1cMl5`3RJcLo4ETJl3k0cLo4FmIn3P1_LP5JfIn3V1[LR5Y4jJkKT5X4hJkKV5X4fJkKW5Y4dJkKZ5W4cJlK[5W4aJlK]5W4^JmK`5V4\\JmKa5W4[JlKc5P6N2N2N2M3N2N2N8G;F:F:F:F:E<Emo[1"}, "label": "a toddler playing with a goat"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black baby sheep\"."}], "task": "refering", "answer_template": "The \"a black baby sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 127, 128, 129, 143, 144, 145, 146, 147, 148, 150, 151, 152, 166, 167, 168, 169, 170, 173, 174, 175, 176, 195, 196, 197, 217, 218, 219, 239, 240, 241, 258, 259, 260, 262, 263, 264, 282, 285, 286, 287], "bbox": [0.241921875, 0.3545433255269321, 0.7034062500000001, 0.8575409836065573], "iscrowd": 0, "area": 13606.40335, "rle": {"size": [427, 640], "counts": "bjP22W=2M3\\FLo56QJMj56UJNf53ZJ0b51^J3\\5OdJ4X5MhJ6T5KmJ8m4JSK9i4HWK<d4E\\K>`4C`Ka0Z4AfK?Y4BgK?W4BiK?U4BkK>T4DkK=T4ClK=T4CmK=Q4DoK<Q4DoK=o3DQL<o3DQL]ObLa0\\72RL^OcL>Z75SL]OeL=W76TL^OfL:V78TL_OgL8S7:VL^OiL6Q7<WL^OiL5o6=XL_OjL3l6`0YL]OnL0i6c0YL^OoLIl6i0UL^OQMBn6Q1QL^Oc4b0]K_Ob4a0^K_Ob4a0^K@`4a0`K@_4`0bK_O^4a0bK@\\4a0dK_O\\4a0dK@[4`0eKAZ4?fKAY4a0fK@Y4`0gKAX4?hKAX4?hK^OZ4c0iKWOZ4i0jKPOY4P1jKjNY4V1m31000000000001O00000O10000000000000O1000000000001O0O100000000000000O1000000000001N1000000000001N10001O001O0O101O00001N100O2O0OoF^O\\5a0cJIT56lJ5j4JVK`0a4@]Kk0Z4TOfKm0Z4SOeKo0Z4POfKQ1Y4POfKQ1Z4nNfKS1Z4mNeKU1Z4jNfKW1Z4iNdKY1\\4gNcKZ1]4eNcK]1\\4cNcK^1]4aNcK`1]4`NbKa1^4^NbKd1]4\\NbKe1^4[NaKf1_4YN`Ki1`4WN_Kk1`4TN`Km1`4SN_Ko0^MTOS7M^Kk0kMQOg63^Kh0WNoNY6:_Kc0P6\\OPJ`0U6@jI<[6CdI:a6F^I6g6IYI6i6IWI<e6B\\Ic0`6\\O`Ii0^95K5J6K5K5K5L4K5K5L1N2N101N2N10SO^ERO`:h0mEROR:h0ZFTOe9e0hFVOV9e0VGUOi8g0aGUO^8j0Q2000010O00001O10^MTOeHk0[7WOdHh0\\7ZOdHd0\\7^OcHb0\\7@cH?]7CcH;]7GbH8^7JaH5a7K_H4b7L]H3f7LYH3j7LVH2l7ORH1o70oGOR83mGKT87jGHW8:gGEY8>fGAZ8a0dG^O]8d0aG[O`8l0YGSOh8U1QGiNP9Y1nFgNR9[1lFdNU9^1jF`NV9c1X12N1O2N2N2N2M3M3M2O2M2N000001O0BSE^Nn:c1TEZNl:f1WEWNj:i1WEUNi:k1;0011N101N1O2O0O2O0O2N102M3M3N3L3N2M3M3M2O1N2N3N1N2N3M2N2O1N3M2N2NcCO^<1O2N101N1O2N1O1O100O1O1O1O1O1O100O1O0100000000O2N1O2N2O1N2N2N2N2O1NhP_2"}, "label": "a black baby sheep"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black sheep being pet\"."}], "task": "refering", "answer_template": "The \"the black sheep being pet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 127, 128, 129, 143, 144, 145, 146, 147, 148, 150, 151, 152, 166, 167, 168, 169, 170, 173, 174, 175, 176, 195, 196, 197, 217, 218, 219, 239, 240, 241, 258, 259, 260, 262, 263, 264, 282, 285, 286, 287], "bbox": [0.241921875, 0.3545433255269321, 0.7034062500000001, 0.8575409836065573], "iscrowd": 0, "area": 13606.40335, "rle": {"size": [427, 640], "counts": "bjP22W=2M3\\FLo56QJMj56UJNf53ZJ0b51^J3\\5OdJ4X5MhJ6T5KmJ8m4JSK9i4HWK<d4E\\K>`4C`Ka0Z4AfK?Y4BgK?W4BiK?U4BkK>T4DkK=T4ClK=T4CmK=Q4DoK<Q4DoK=o3DQL<o3DQL]ObLa0\\72RL^OcL>Z75SL]OeL=W76TL^OfL:V78TL_OgL8S7:VL^OiL6Q7<WL^OiL5o6=XL_OjL3l6`0YL]OnL0i6c0YL^OoLIl6i0UL^OQMBn6Q1QL^Oc4b0]K_Ob4a0^K_Ob4a0^K@`4a0`K@_4`0bK_O^4a0bK@\\4a0dK_O\\4a0dK@[4`0eKAZ4?fKAY4a0fK@Y4`0gKAX4?hKAX4?hK^OZ4c0iKWOZ4i0jKPOY4P1jKjNY4V1m31000000000001O00000O10000000000000O1000000000001O0O100000000000000O1000000000001N1000000000001N10001O001O0O101O00001N100O2O0OoF^O\\5a0cJIT56lJ5j4JVK`0a4@]Kk0Z4TOfKm0Z4SOeKo0Z4POfKQ1Y4POfKQ1Z4nNfKS1Z4mNeKU1Z4jNfKW1Z4iNdKY1\\4gNcKZ1]4eNcK]1\\4cNcK^1]4aNcK`1]4`NbKa1^4^NbKd1]4\\NbKe1^4[NaKf1_4YN`Ki1`4WN_Kk1`4TN`Km1`4SN_Ko0^MTOS7M^Kk0kMQOg63^Kh0WNoNY6:_Kc0P6\\OPJ`0U6@jI<[6CdI:a6F^I6g6IYI6i6IWI<e6B\\Ic0`6\\O`Ii0^95K5J6K5K5K5L4K5K5L1N2N101N2N10SO^ERO`:h0mEROR:h0ZFTOe9e0hFVOV9e0VGUOi8g0aGUO^8j0Q2000010O00001O10^MTOeHk0[7WOdHh0\\7ZOdHd0\\7^OcHb0\\7@cH?]7CcH;]7GbH8^7JaH5a7K_H4b7L]H3f7LYH3j7LVH2l7ORH1o70oGOR83mGKT87jGHW8:gGEY8>fGAZ8a0dG^O]8d0aG[O`8l0YGSOh8U1QGiNP9Y1nFgNR9[1lFdNU9^1jF`NV9c1X12N1O2N2N2N2M3M3M2O2M2N000001O0BSE^Nn:c1TEZNl:f1WEWNj:i1WEUNi:k1;0011N101N1O2O0O2O0O2N102M3M3N3L3N2M3M3M2O1N2N3N1N2N3M2N2O1N3M2N2NcCO^<1O2N101N1O2N1O1O100O1O1O1O1O1O100O1O0100000000O2N1O2N2O1N2N2N2N2O1NhP_2"}, "label": "the black sheep being pet"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in an orange striped shirt\"."}], "task": "refering", "answer_template": "The \"a woman in an orange striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 84, 88, 89, 90, 91, 105, 106, 107, 108, 111, 112, 113, 114, 127, 128, 129, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297], "bbox": [0.539171875, 0.0016861826697892272, 1.0, 0.8075644028103045], "iscrowd": 0, "area": 33869.19169999999, "rle": {"size": [427, 640], "counts": "_P`41Z=0O2N101N100O2N100O2N101N1O101N1O2OO01O100O10O0100O100O010O10O010O010O010O010O010O10O01O010O01O010O1O1O1O001O1O1O100O1O1O1O001O1O1O010O00001O001O0000010O0000000000000010O001O2N1O2O0O2O2M2O2M2O2M2O2M2O2M2O1N3N1N2O1N`]a03\\b^O4M2M2O2M3M2O2M3N2M4M2M4L3N3L3O2N1O2N1O2N1O;E2N3M3M3M2N1O2N1O1O1]G`MY6a2dIiMT6X2iIQNP6P2nIXNk5i1RJ`Ng5a1VJiNb5X1]JoN\\5R1bJVOW5k0hJ\\OQ5f0mJBk4?TKHe49YKIf48YKHg49XKHg49WKHi49VKGj49VKGj49UKIj47VKIj48TKIl47TKIl47SKKl45TKKl45SKLm45RKKn45RKLm44RKMn43RKMn48kJJU5<cJG\\5=^JEb5>YJDg5?TJDk5n3000001O0000001O000000O100O1O1O100O_LjHo0U7lMSJR2l5nMWJP2h5oM\\Jo1d5PN_Jn1`5QNdJm1[5RNiJl1V5TNmJj1R5UNRKi1n4VNUKh1j4WNZKg1e4YN^Ke1a4ZNcKd1]4[NfKc1Y4\\NjKc1U4]NlKc1S4\\NoKd1P4[NRLe1n3ZNSLf1l3YNVLg1i3YNXLg1g3XN[Lh1f3VN[Lj1e3TN]Ll1c3SN^Lm1b3QN`Lo1`3PNaLP2`3mMbLS2^3kMdLU2\\3iMfLW2Z3hMgLX2Y3fMiLZ2X3cMjL]2V3aMlL_2T3_MnLa2R3^MnLc2m60O100O100O100O10000O100O100O100O100O100O100O1B>aN_1bN^1aN^1"}, "label": "a woman in an orange striped shirt"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the orange striped shirt\"."}], "task": "refering", "answer_template": "The \"the woman in the orange striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 84, 88, 89, 90, 91, 105, 106, 107, 108, 111, 112, 113, 114, 127, 128, 129, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297], "bbox": [0.539171875, 0.0016861826697892272, 1.0, 0.8075644028103045], "iscrowd": 0, "area": 33869.19169999999, "rle": {"size": [427, 640], "counts": "_P`41Z=0O2N101N100O2N100O2N101N1O101N1O2OO01O100O10O0100O100O010O10O010O010O010O010O010O10O01O010O01O010O1O1O1O001O1O1O100O1O1O1O001O1O1O010O00001O001O0000010O0000000000000010O001O2N1O2O0O2O2M2O2M2O2M2O2M2O2M2O1N3N1N2O1N`]a03\\b^O4M2M2O2M3M2O2M3N2M4M2M4L3N3L3O2N1O2N1O2N1O;E2N3M3M3M2N1O2N1O1O1]G`MY6a2dIiMT6X2iIQNP6P2nIXNk5i1RJ`Ng5a1VJiNb5X1]JoN\\5R1bJVOW5k0hJ\\OQ5f0mJBk4?TKHe49YKIf48YKHg49XKHg49WKHi49VKGj49VKGj49UKIj47VKIj48TKIl47TKIl47SKKl45TKKl45SKLm45RKKn45RKLm44RKMn43RKMn48kJJU5<cJG\\5=^JEb5>YJDg5?TJDk5n3000001O0000001O000000O100O1O1O100O_LjHo0U7lMSJR2l5nMWJP2h5oM\\Jo1d5PN_Jn1`5QNdJm1[5RNiJl1V5TNmJj1R5UNRKi1n4VNUKh1j4WNZKg1e4YN^Ke1a4ZNcKd1]4[NfKc1Y4\\NjKc1U4]NlKc1S4\\NoKd1P4[NRLe1n3ZNSLf1l3YNVLg1i3YNXLg1g3XN[Lh1f3VN[Lj1e3TN]Ll1c3SN^Lm1b3QN`Lo1`3PNaLP2`3mMbLS2^3kMdLU2\\3iMfLW2Z3hMgLX2Y3fMiLZ2X3cMjL]2V3aMlL_2T3_MnLa2R3^MnLc2m60O100O100O100O10000O100O100O100O100O100O100O1B>aN_1bN^1aN^1"}, "label": "the woman in the orange striped shirt"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in pink sitting in wooden chair\"."}], "task": "refering", "answer_template": "The \"person in pink sitting in wooden chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102], "bbox": [0.23546874999999998, 0.0, 0.48795312499999993, 0.32674473067915694], "iscrowd": 0, "area": 9541.338650000002, "rle": {"size": [427, 640], "counts": "ZQo1d0f<1O2N2N2gN]OgEd0X:EaE;]:J_E7_:JbE6\\:KdE5[:MeE3Y:NgE3W:NjE2U:NkE3S:NmE2R:OoE1n91RF0l91UFOh93XFMf95[FKc96]FK`98_FI_98bFH[9:eFFY9<hFDU9>kFCR9?nFBo8?RGBk8?VGBi8=XGCh8<ZGDe8;\\GFc8:]GGa89`GH_87bGI^86cGK\\84eGMZ82hGNW81jGOV80kG1T8OlG2S8MnG4Q8KPH5P8NmG3R8OlG2S81jGOV83hGNW85gGKX87fGJY8:cGF]8`0]GAb8a0\\G@c8a0]G^Oc8c0\\G^Oc8c0\\G^Oc8b0]G_Ob8b0^G]Ob8d0]GXOg8i0XGPOo8o0RGjNU9W1jFcN\\9^1dFZNc9g1\\FRNk9o1TFkMR:V2mEcMZ:^29001OI7F:F:F:G9L4K5L4L4LiFEf57\\JMb5NbJ5W94L4POAoDb0h:MnD5k:2TEOj:3VEMX:[OhEh00MR:AmEb02Ml9FRF=2Mg9KWF82Mb9O]F41M]94bFN2N[95cFM2M\\96bFM1N]95bFM1N]95bFL2O\\95bFL2O\\94cFM1O\\94cFM1O\\94cFL20[94cFL20[94cFK31Z94cFK31Z94bFL40Z94bFK51Y93cFL41Y93cFL31\\92aFL42[92aFL42[92aFK53d9HWF553o9]OlE`053d:L]E4c:L]E4c:L]E4c:K^E5b:K^E5b:K^E5b:J_E6a:J_E6a:I`E6d:G[E:j:AVE?o:[OREe0j;000ZMZOgHe0Y7BaH>_7IZH7f7OTH0m73PHMP83PHLQ84PHKP84UHHk78YHCh7=\\H_Od7a0`HZOa7f0a21000O10SDZOU;f0kDZOU;f0kDJe:6ZEKf:5ZELe:4[ELe:4ZEMf:2[ENd:3\\EMd:3\\ENc:2\\EOf:OZE1m:HSE8T;AlD?o;101O04J8I6JVfX4"}, "label": "person in pink sitting in wooden chair"}]}
{"id": 336491, "image": "COCO_train2014_000000336491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl sitting in a chair\"."}], "task": "refering", "answer_template": "The \"girl sitting in a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102], "bbox": [0.23546874999999998, 0.0, 0.48795312499999993, 0.32674473067915694], "iscrowd": 0, "area": 9541.338650000002, "rle": {"size": [427, 640], "counts": "ZQo1d0f<1O2N2N2gN]OgEd0X:EaE;]:J_E7_:JbE6\\:KdE5[:MeE3Y:NgE3W:NjE2U:NkE3S:NmE2R:OoE1n91RF0l91UFOh93XFMf95[FKc96]FK`98_FI_98bFH[9:eFFY9<hFDU9>kFCR9?nFBo8?RGBk8?VGBi8=XGCh8<ZGDe8;\\GFc8:]GGa89`GH_87bGI^86cGK\\84eGMZ82hGNW81jGOV80kG1T8OlG2S8MnG4Q8KPH5P8NmG3R8OlG2S81jGOV83hGNW85gGKX87fGJY8:cGF]8`0]GAb8a0\\G@c8a0]G^Oc8c0\\G^Oc8c0\\G^Oc8b0]G_Ob8b0^G]Ob8d0]GXOg8i0XGPOo8o0RGjNU9W1jFcN\\9^1dFZNc9g1\\FRNk9o1TFkMR:V2mEcMZ:^29001OI7F:F:F:G9L4K5L4L4LiFEf57\\JMb5NbJ5W94L4POAoDb0h:MnD5k:2TEOj:3VEMX:[OhEh00MR:AmEb02Ml9FRF=2Mg9KWF82Mb9O]F41M]94bFN2N[95cFM2M\\96bFM1N]95bFM1N]95bFL2O\\95bFL2O\\94cFM1O\\94cFM1O\\94cFL20[94cFL20[94cFK31Z94cFK31Z94bFL40Z94bFK51Y93cFL41Y93cFL31\\92aFL42[92aFL42[92aFK53d9HWF553o9]OlE`053d:L]E4c:L]E4c:L]E4c:K^E5b:K^E5b:K^E5b:J_E6a:J_E6a:I`E6d:G[E:j:AVE?o:[OREe0j;000ZMZOgHe0Y7BaH>_7IZH7f7OTH0m73PHMP83PHLQ84PHKP84UHHk78YHCh7=\\H_Od7a0`HZOa7f0a21000O10SDZOU;f0kDZOU;f0kDJe:6ZEKf:5ZELe:4[ELe:4ZEMf:2[ENd:3\\EMd:3\\ENc:2\\EOf:OZE1m:HSE8T;AlD?o;101O04J8I6JVfX4"}, "label": "girl sitting in a chair"}]}
{"id": 557678, "image": "COCO_train2014_000000557678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor between two other monitors\"."}], "task": "refering", "answer_template": "The \"a computer monitor between two other monitors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 194, 195, 217], "bbox": [0.257140625, 0.0883125, 0.625328125, 0.5428541666666666], "iscrowd": 0, "area": 38115.040349999996, "rle": {"size": [480, 640], "counts": "Q]]2`0_>a0_Ob0^Oa0@a0_O`0@`0@a0_O?A000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000^GSKQ7m4kHWKU7i4gH[KY7e4dH^K\\7b4`HaKa7_4\\HdKd7\\4XHhKh7X4THlKl7T4QHnKP8R4lGRLS8o3jGTLV8l3fGXLZ8h3bG[L_8e3^G^Lb8P5000O1000000O10O1000O1000000O1000000O0100000jN]GnKd8R4fGdKZ8\\4oG[KQ8e4YHQKg7o4U10000000O10000000000000O100000O10000000000000000000000O100000O1000000000000000O100000000000000000O10O10000000000000000000000O10000000O100000000000O1000000000000000000000O01f^`3"}, "label": "a computer monitor between two other monitors"}]}
{"id": 557678, "image": "COCO_train2014_000000557678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black framed monitor with white wires in front of it\"."}], "task": "refering", "answer_template": "The \"black framed monitor with white wires in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 194, 195, 217], "bbox": [0.257140625, 0.0883125, 0.625328125, 0.5428541666666666], "iscrowd": 0, "area": 38115.040349999996, "rle": {"size": [480, 640], "counts": "Q]]2`0_>a0_Ob0^Oa0@a0_O`0@`0@a0_O?A000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000^GSKQ7m4kHWKU7i4gH[KY7e4dH^K\\7b4`HaKa7_4\\HdKd7\\4XHhKh7X4THlKl7T4QHnKP8R4lGRLS8o3jGTLV8l3fGXLZ8h3bG[L_8e3^G^Lb8P5000O1000000O10O1000O1000000O1000000O0100000jN]GnKd8R4fGdKZ8\\4oG[KQ8e4YHQKg7o4U10000000O10000000000000O100000O10000000000000000000000O100000O1000000000000000O100000000000000000O10O10000000000000000000000O10000000O100000000000O1000000000000000000000O01f^`3"}, "label": "black framed monitor with white wires in front of it"}]}
{"id": 557678, "image": "COCO_train2014_000000557678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor reading 0 : 504\"."}], "task": "refering", "answer_template": "The \"a computer monitor reading 0 : 504\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.6565000000000001, 0.22856249999999997, 0.9545468750000001, 0.5038958333333333], "iscrowd": 0, "area": 22469.213699999997, "rle": {"size": [480, 640], "counts": "cRU6n0T;m21O1O1O1000000000000O10000000000000O1000000000000000000000O1000000000000000000000000000O100000O1000000000000000000000000000000000000O1000O10000000000000000000000000000000O100000000000000000O100000000000000000O10000000000000000000000000000000O1000O10000000000000000000000000000000000O10000000O1000000000000000000000000000O1000000000000000000000O10000000000000O1000000000000000000eMma="}, "label": "a computer monitor reading 0 : 504"}]}
{"id": 557678, "image": "COCO_train2014_000000557678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer screen showing the time\"."}], "task": "refering", "answer_template": "The \"a computer screen showing the time\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.6565000000000001, 0.22856249999999997, 0.9545468750000001, 0.5038958333333333], "iscrowd": 0, "area": 22469.213699999997, "rle": {"size": [480, 640], "counts": "cRU6n0T;m21O1O1O1000000000000O10000000000000O1000000000000000000000O1000000000000000000000000000O100000O1000000000000000000000000000000000000O1000O10000000000000000000000000000000O100000000000000000O100000000000000000O10000000000000000000000000000000O1000O10000000000000000000000000000000000O10000000O1000000000000000000000000000O1000000000000000000000O10000000000000O1000000000000000000eMma="}, "label": "a computer screen showing the time"}]}
{"id": 557678, "image": "COCO_train2014_000000557678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a monitor along with keyboard\"."}], "task": "refering", "answer_template": "The \"a monitor along with keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189], "bbox": [0.0038906250000000004, 0.2935, 0.26103125, 0.5194791666666666], "iscrowd": 0, "area": 17409.62525, "rle": {"size": [480, 640], "counts": "hU11^>a0_Oa0_Oa0@`0_Oa0_Oa0K5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000cbm6"}, "label": "a monitor along with keyboard"}]}
{"id": 557678, "image": "COCO_train2014_000000557678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small black computer monitor\"."}], "task": "refering", "answer_template": "The \"the small black computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189], "bbox": [0.0038906250000000004, 0.2935, 0.26103125, 0.5194791666666666], "iscrowd": 0, "area": 17409.62525, "rle": {"size": [480, 640], "counts": "hU11^>a0_Oa0_Oa0@`0_Oa0_Oa0K5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000cbm6"}, "label": "the small black computer monitor"}]}
{"id": 377456, "image": "COCO_train2014_000000377456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male tennis player\"."}], "task": "refering", "answer_template": "The \"male tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 58, 59, 76, 77, 94, 95, 111, 112, 113, 129, 130, 131, 132, 148, 149, 150, 167, 168, 185, 186], "bbox": [0.21056, 0.1972972972972973, 0.3539, 0.9058258258258258], "iscrowd": 0, "area": 7420.0485, "rle": {"size": [333, 500], "counts": "cZR19n9c0^O5J6K6I6J7L3N3M2N3M2N2N3N1N3M2N3M2N3M2N2D=L3N3M2O2M2O2M2O5J5L4K2N20O`JSLa4n3^KTL`4m3_KUL`4j3aKVL^4k3aKWL]4j3cKWL\\4h3eKXLZ4U2[KeN:XOY4P2bKgN5ZOW4l1hKiN2[OV4F]Ka1`0]OM]OT4EaK\\1b0BH_OS4CdKX1f0DDB2ZOW37_LT1h0I@BJB^3ObLo0k0MZOECGe3IcLk0n0OXO>V3gNfLh0P12TO?U3gNiLd0Q15QOa0T3fNkL;Z1=gNc0S3dNnL0c1i0\\Nc0R3dNQMDm1S1QNe0l2hNGc0\\Mg0e2lN1<ZMh0e2kN2<ZMi0d2jN4<WMl0d2gN6=VMl0c2hN8:VMo0a2fN:;TMo0c2eN;;RM9HYOk22>9PM13C_22`09nLFa0MQ23c08jL\\OQ17c13d06aNEk03h04aNGh00m04`NJc0NR14_NLk4OZKO^oX3"}, "label": "male tennis player"}]}
{"id": 377456, "image": "COCO_train2014_000000377456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player\"."}], "task": "refering", "answer_template": "The \"tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 58, 59, 76, 77, 94, 95, 111, 112, 113, 129, 130, 131, 132, 148, 149, 150, 167, 168, 185, 186], "bbox": [0.21056, 0.1972972972972973, 0.3539, 0.9058258258258258], "iscrowd": 0, "area": 7420.0485, "rle": {"size": [333, 500], "counts": "cZR19n9c0^O5J6K6I6J7L3N3M2N3M2N2N3N1N3M2N3M2N3M2N2D=L3N3M2O2M2O2M2O5J5L4K2N20O`JSLa4n3^KTL`4m3_KUL`4j3aKVL^4k3aKWL]4j3cKWL\\4h3eKXLZ4U2[KeN:XOY4P2bKgN5ZOW4l1hKiN2[OV4F]Ka1`0]OM]OT4EaK\\1b0BH_OS4CdKX1f0DDB2ZOW37_LT1h0I@BJB^3ObLo0k0MZOECGe3IcLk0n0OXO>V3gNfLh0P12TO?U3gNiLd0Q15QOa0T3fNkL;Z1=gNc0S3dNnL0c1i0\\Nc0R3dNQMDm1S1QNe0l2hNGc0\\Mg0e2lN1<ZMh0e2kN2<ZMi0d2jN4<WMl0d2gN6=VMl0c2hN8:VMo0a2fN:;TMo0c2eN;;RM9HYOk22>9PM13C_22`09nLFa0MQ23c08jL\\OQ17c13d06aNEk03h04aNGh00m04`NJc0NR14_NLk4OZKO^oX3"}, "label": "tennis player"}]}
{"id": 311933, "image": "COCO_train2014_000000311933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green tree\"."}], "task": "refering", "answer_template": "The \"a green tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 9, 28, 29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 146, 147, 148], "bbox": [0.244453125, 0.03191666666666667, 0.616578125, 0.41102083333333334], "iscrowd": 0, "area": 21960.834450000002, "rle": {"size": [480, 640], "counts": "ddY24j>4M4L3M4L?A1O1N2O1O1O2N1O1N2M3M3M3O100VC]NW<d1gCcNT<]1kCjNn;V1QDoNk;R1TDoNl;Q1RDQOm;P1RDQOm;P1QDROo;n0PDROP<o0nCSOQ<o1N2O1O1O1O0O2O1O1O1O1N20OmDhL^:Y3^EkLa:V3\\ElLd:U3XEoLh:Q3TERMl:b3N2N1N3M2N3N2001N1000000000O1O1O1O1O1N21O1O1O000O001O010O001O010O001O010O00100O2N1O2O0O1OoNnKVGP4j8TLSGl3l8YLPGf3P9_LlF`3T9V101O000102M2N2N3N1N3N1N3N1N2O00O001N101N101O1N10GmEfKR:Z4QFcKn9^4TFaKk9^4YF`Ke9m3TF`LZ:`3jE[LW:d3nEXLQ:i3SFRLn9m3b010O10O01N1O2O0eLmD?2X1T;VNkDa04V1U;VNhDb06V1U;TNfDf08S1U;TNcDi09R1\\;mNeDS1Z;nNfDQ1Z;oNgDP1Y;POhDo0Y;POhDo0X;QOiDP1U;POlDi0_ObNd;e0mDi0B`N\\;j0SEf0Q;^OPEa0m:BTE>l:AUE?k:@VE?k:@VE`0j:@WEa0g:^OZEe0d:YO]Ek0_:TOcEn0Z:QOgER1V:mNkEV1R:jNoEY1m9fNTF]1i9bNXFZOnNR2g:cN]FZOlNU2e:QN\\E4T1EmNV2b:oM^E6T1ElNU2d:nM\\E9Y1i1[9lM^E;W1h1\\9lM]E=W1f1c9[N]Fe1c9[N]Fe1b9\\N]Ff1a9[N^Ff1b9ZN]Fg1b9ZN]Fh1a9YN^Fh1c9WN[Fc1n9\\NQFe1P:ZNoEg1R:XNmEi1T:VNkEa1POYNV;5iEc1SOVNU;6gE3ROQ14dNT;7eE2WOQ13dNR;8cE1[OS12bNP;:bEN@=EF1Nc;1VEKE`0BE2Ma;3TEHMb0]OE4K^;6mE;aNDm;1bE;cNBl;2aE<dNAl;2aE<dNAm;1_E>fN_Ol;2^E`0fN\\On;3\\Ea0P;^OQEa0Q;]OoDc0g<L50O001O1O010O1O001O1O0010O01O1O010O10O01OCjA7P>HPB=l=BUB=k=CVB<k=DUB;l=DTB<l=DUB;l=DUB;k=EVB:k=EUB;l=ETB:m=ETB6Q>IPB4T>JlA6V>HkA7`>O001O2N3Mffb3"}, "label": "a green tree"}]}
{"id": 311933, "image": "COCO_train2014_000000311933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant in the background of a picture\"."}], "task": "refering", "answer_template": "The \"a plant in the background of a picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 9, 28, 29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 146, 147, 148], "bbox": [0.244453125, 0.03191666666666667, 0.616578125, 0.41102083333333334], "iscrowd": 0, "area": 21960.834450000002, "rle": {"size": [480, 640], "counts": "ddY24j>4M4L3M4L?A1O1N2O1O1O2N1O1N2M3M3M3O100VC]NW<d1gCcNT<]1kCjNn;V1QDoNk;R1TDoNl;Q1RDQOm;P1RDQOm;P1QDROo;n0PDROP<o0nCSOQ<o1N2O1O1O1O0O2O1O1O1O1N20OmDhL^:Y3^EkLa:V3\\ElLd:U3XEoLh:Q3TERMl:b3N2N1N3M2N3N2001N1000000000O1O1O1O1O1N21O1O1O000O001O010O001O010O001O010O00100O2N1O2O0O1OoNnKVGP4j8TLSGl3l8YLPGf3P9_LlF`3T9V101O000102M2N2N3N1N3N1N3N1N2O00O001N101N101O1N10GmEfKR:Z4QFcKn9^4TFaKk9^4YF`Ke9m3TF`LZ:`3jE[LW:d3nEXLQ:i3SFRLn9m3b010O10O01N1O2O0eLmD?2X1T;VNkDa04V1U;VNhDb06V1U;TNfDf08S1U;TNcDi09R1\\;mNeDS1Z;nNfDQ1Z;oNgDP1Y;POhDo0Y;POhDo0X;QOiDP1U;POlDi0_ObNd;e0mDi0B`N\\;j0SEf0Q;^OPEa0m:BTE>l:AUE?k:@VE?k:@VE`0j:@WEa0g:^OZEe0d:YO]Ek0_:TOcEn0Z:QOgER1V:mNkEV1R:jNoEY1m9fNTF]1i9bNXFZOnNR2g:cN]FZOlNU2e:QN\\E4T1EmNV2b:oM^E6T1ElNU2d:nM\\E9Y1i1[9lM^E;W1h1\\9lM]E=W1f1c9[N]Fe1c9[N]Fe1b9\\N]Ff1a9[N^Ff1b9ZN]Fg1b9ZN]Fh1a9YN^Fh1c9WN[Fc1n9\\NQFe1P:ZNoEg1R:XNmEi1T:VNkEa1POYNV;5iEc1SOVNU;6gE3ROQ14dNT;7eE2WOQ13dNR;8cE1[OS12bNP;:bEN@=EF1Nc;1VEKE`0BE2Ma;3TEHMb0]OE4K^;6mE;aNDm;1bE;cNBl;2aE<dNAl;2aE<dNAm;1_E>fN_Ol;2^E`0fN\\On;3\\Ea0P;^OQEa0Q;]OoDc0g<L50O001O1O010O1O001O1O0010O01O1O010O10O01OCjA7P>HPB=l=BUB=k=CVB<k=DUB;l=DTB<l=DUB;l=DUB;k=EVB:k=EUB;l=ETB:m=ETB6Q>IPB4T>JlA6V>HkA7`>O001O2N3Mffb3"}, "label": "a plant in the background of a picture"}]}
{"id": 311933, "image": "COCO_train2014_000000311933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman pulling a man ' s tie with her teeth\"."}], "task": "refering", "answer_template": "The \"a woman pulling a man ' s tie with her teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 130, 131, 132, 133, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 342, 343, 344, 366, 367, 389, 390], "bbox": [0.63490625, 0.28341666666666665, 1.0, 0.9888541666666666], "iscrowd": 0, "area": 42148.26, "rle": {"size": [480, 640], "counts": "Ubn51g>>B>B>D:F6K5K5J7J5_Oa0C=L5K4K5L4L5J5M30001O0O10O1O0O2O1O0O2O1O1N101O1N101O1O0O2O1O1O0OdIZLb1e3TNfLl1Y3jMSMT2l2mMVMR2i2mMYMS2g2lM[MR2f2lM\\MT2d2kM^MT2b2jM`MV2`2iMaMX2_2eMcM[2]2dMdM\\2\\2bMfM^2Z2aMgM`2X2_MhMb2X2\\MjMd2V2[MkMe2U2ZJ`LX2\\1`3S2VJhLT2U1g3S2SJnLR2m0m3U2oITMP2f0S4U2kI^Mj1=\\4U2hI`Nl0YO^5W2dIgNh0ROe5W2aIjNi0nNh5h4XJVKj5k4TJUKm5l4RJSKo5n4oIQKT6o4kIPKV6Q5hIoJY6R5fImJ\\6S5cIkJ_6V5_IiJd6W5[IfJh6[5VIdJm6\\5RIbJP7_5oH_JT7a5jH]JY7d5fHZJ]7f5aHYJa7h5^HVJe7j5YHVJh7a61O001N101O001N101O001N3N3M2N3L3N3M2N3M2M4M2N2N2N2N2O1N2O1N1O100O10O0100O1O010O100O1O010O10O00001O010O1O1O00100O1O01000O10O10O100O10O10O1000O010000O01000O10O10O10000O01000O10O10O1000O100000O10O1000000O0100000O1001N2O1O1O1O1O0O2hGWJV7j5bH\\J_7e5[H_Jf7b5VH`Jk7a5PHbJQ8V6001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1OWG"}, "label": "a woman pulling a man ' s tie with her teeth"}]}
{"id": 311933, "image": "COCO_train2014_000000311933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with long blonde hair and a gray dress holding a red tie between her teeth\"."}], "task": "refering", "answer_template": "The \"a woman with long blonde hair and a gray dress holding a red tie between her teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 130, 131, 132, 133, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 342, 343, 344, 366, 367, 389, 390], "bbox": [0.63490625, 0.28341666666666665, 1.0, 0.9888541666666666], "iscrowd": 0, "area": 42148.26, "rle": {"size": [480, 640], "counts": "Ubn51g>>B>B>D:F6K5K5J7J5_Oa0C=L5K4K5L4L5J5M30001O0O10O1O0O2O1O0O2O1O1N101O1N101O1O0O2O1O1O0OdIZLb1e3TNfLl1Y3jMSMT2l2mMVMR2i2mMYMS2g2lM[MR2f2lM\\MT2d2kM^MT2b2jM`MV2`2iMaMX2_2eMcM[2]2dMdM\\2\\2bMfM^2Z2aMgM`2X2_MhMb2X2\\MjMd2V2[MkMe2U2ZJ`LX2\\1`3S2VJhLT2U1g3S2SJnLR2m0m3U2oITMP2f0S4U2kI^Mj1=\\4U2hI`Nl0YO^5W2dIgNh0ROe5W2aIjNi0nNh5h4XJVKj5k4TJUKm5l4RJSKo5n4oIQKT6o4kIPKV6Q5hIoJY6R5fImJ\\6S5cIkJ_6V5_IiJd6W5[IfJh6[5VIdJm6\\5RIbJP7_5oH_JT7a5jH]JY7d5fHZJ]7f5aHYJa7h5^HVJe7j5YHVJh7a61O001N101O001N101O001N3N3M2N3L3N3M2N3M2M4M2N2N2N2N2O1N2O1N1O100O10O0100O1O010O100O1O010O10O00001O010O1O1O00100O1O01000O10O10O100O10O10O1000O010000O01000O10O10O10000O01000O10O10O1000O100000O10O1000000O0100000O1001N2O1O1O1O1O0O2hGWJV7j5bH\\J_7e5[H_Jf7b5VH`Jk7a5PHbJQ8V6001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1OWG"}, "label": "a woman with long blonde hair and a gray dress holding a red tie between her teeth"}]}
{"id": 172669, "image": "COCO_train2014_000000172669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a textured vase holds an array of pink flowers\"."}], "task": "refering", "answer_template": "The \"a textured vase holds an array of pink flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362, 363, 383, 384, 385, 386, 406, 407, 408], "bbox": [0.643765625, 0.5080506822612085, 0.8015625, 0.9733723196881092], "iscrowd": 0, "area": 20650.631200000003, "rle": {"size": [513, 640], "counts": "af^6:c?e0^Oa0_Ob0^Oa0\\N_MgDR3i:PMTEb3Z:`LdEQ4l9QLPFb4^9aK_FP5Q9RKlF_5c8cJ[Gm5V8UJgGk5Z8WJcGj5]8WJaGj5_8XJ^Gi5b8YJ[Gh5e8j00001O0000000O10000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000001O00000000001O00000000001O0000000]I\\Gi5d8TJ`Gk5a8QJcGn5]8oIgGS5V9jJnF_4h9]K]Fn3W:oKmE[3g:cL]Eg2Z<ZOf0\\Od0F:E;F;EhZo1"}, "label": "a textured vase holds an array of pink flowers"}]}
{"id": 172669, "image": "COCO_train2014_000000172669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark blue transparent flower pot with pink colour flowers in it\"."}], "task": "refering", "answer_template": "The \"a dark blue transparent flower pot with pink colour flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362, 363, 383, 384, 385, 386, 406, 407, 408], "bbox": [0.643765625, 0.5080506822612085, 0.8015625, 0.9733723196881092], "iscrowd": 0, "area": 20650.631200000003, "rle": {"size": [513, 640], "counts": "af^6:c?e0^Oa0_Ob0^Oa0\\N_MgDR3i:PMTEb3Z:`LdEQ4l9QLPFb4^9aK_FP5Q9RKlF_5c8cJ[Gm5V8UJgGk5Z8WJcGj5]8WJaGj5_8XJ^Gi5b8YJ[Gh5e8j00001O0000000O10000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000001O00000000001O00000000001O0000000]I\\Gi5d8TJ`Gk5a8QJcGn5]8oIgGS5V9jJnF_4h9]K]Fn3W:oKmE[3g:cL]Eg2Z<ZOf0\\Od0F:E;F;EhZo1"}, "label": "a dark blue transparent flower pot with pink colour flowers in it"}]}
{"id": 172669, "image": "COCO_train2014_000000172669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue vase near two vases with flowers in the m\"."}], "task": "refering", "answer_template": "The \"a blue vase near two vases with flowers in the m\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 257, 258, 279, 280, 281, 282, 301, 302, 303, 304, 305, 323, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 352, 369, 370, 371, 372, 373, 374, 375, 393, 394, 395, 396, 397], "bbox": [0.063046875, 0.5910136452241715, 0.3314375, 0.982027290448343], "iscrowd": 0, "area": 21709.028849999995, "rle": {"size": [513, 640], "counts": "a_d01i?9H8G9G9H8G9I7L5K4L4K5L4L3M2N3M2N2M3N2N2O2N1O2O1N101N2O1N1O2O1N101N2O1N1O2O000O10000O10000O10001N10000O10000O10000O1O2N1N2O1O1O1O1O2N1N2O1O1kN_KiFb4U9bKhF_4V9eKUFo4h9f0N2N200000001O00000O1000001O000000000O101O0000O10001N10000O1000000O1XJ\\FS5d9hJfFS5Z9iJoFQ5T:M2N2M2O2N1N3N2N1O0O2O001N2O001N2O001N101O001N10001N1iLZDS2g;hM`DU2`;gMgDU2[;fMlDW2T;dMSEZ2n:aMYE[2Z<M00000O10000O10000O10000O1000000O10000O10000O103M4F9G=C?]Oooe6"}, "label": "a blue vase near two vases with flowers in the m"}]}
{"id": 320125, "image": "COCO_train2014_000000320125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cell phone with letters and at the bottome\"."}], "task": "refering", "answer_template": "The \"a black cell phone with letters and at the bottome\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355, 356], "bbox": [0.23293750000000002, 0.3952916666666667, 0.49573437499999995, 0.9337083333333334], "iscrowd": 0, "area": 39283.32844999999, "rle": {"size": [480, 640], "counts": "]XV27_>>B>B>B:F:E;F:F:F:F:F:F:F:F:F:E;F:F:F:F:F:F:F:M3O1000000O1000000O1000000O1000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000h0XOS1mNR1nNS1mNR1nNS1mNR1nNjof4"}, "label": "a black cell phone with letters and at the bottome"}]}
{"id": 320125, "image": "COCO_train2014_000000320125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black phone second to the right with the most buttons\"."}], "task": "refering", "answer_template": "The \"the black phone second to the right with the most buttons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355, 356], "bbox": [0.23293750000000002, 0.3952916666666667, 0.49573437499999995, 0.9337083333333334], "iscrowd": 0, "area": 39283.32844999999, "rle": {"size": [480, 640], "counts": "]XV27_>>B>B>B:F:E;F:F:F:F:F:F:F:F:F:E;F:F:F:F:F:F:F:M3O1000000O1000000O1000000O1000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000h0XOS1mNR1nNS1mNR1nNS1mNR1nNjof4"}, "label": "the black phone second to the right with the most buttons"}]}
{"id": 320125, "image": "COCO_train2014_000000320125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver cell phone sitting on top of a laptop on the right most side\"."}], "task": "refering", "answer_template": "The \"a silver cell phone sitting on top of a laptop on the right most side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367], "bbox": [0.7519375, 0.4718125, 1.0, 0.9225833333333334], "iscrowd": 0, "area": 24824.63495, "rle": {"size": [480, 640], "counts": "`fQ72n>5J6K5K5J6K5J6K5K5J6K5K5J6K5K6I6K5J6K5K5J6K5K5J6K5J6K5K5J6K5K5J7J5K5J6K5J6K5K5K3M00000000000000001O00000000000000000000000000O10000000000000000000000000001O1O001O1O001O001O1O001O1O001O001O1O001O1O001N2O001O001O1O001O1O002N3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M3M4L3M3M3M3M4L3L4M3M3M4L3M3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3MaC"}, "label": "a silver cell phone sitting on top of a laptop on the right most side"}]}
{"id": 320125, "image": "COCO_train2014_000000320125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver cell phone next to 3 other cell phones\"."}], "task": "refering", "answer_template": "The \"a silver cell phone next to 3 other cell phones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367], "bbox": [0.7519375, 0.4718125, 1.0, 0.9225833333333334], "iscrowd": 0, "area": 24824.63495, "rle": {"size": [480, 640], "counts": "`fQ72n>5J6K5K5J6K5J6K5K5J6K5K5J6K5K6I6K5J6K5K5J6K5K5J6K5J6K5K5J6K5K5J7J5K5J6K5J6K5K5K3M00000000000000001O00000000000000000000000000O10000000000000000000000000001O1O001O1O001O001O1O001O1O001O001O1O001O1O001N2O001O001O1O001O1O002N3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M3M4L3M3M3M3M4L3L4M3M3M4L3M3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3MaC"}, "label": "a silver cell phone next to 3 other cell phones"}]}
{"id": 320125, "image": "COCO_train2014_000000320125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the phone in the middle of other phones\"."}], "task": "refering", "answer_template": "The \"the phone in the middle of other phones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 361, 362, 363], "bbox": [0.494140625, 0.4101666666666667, 0.794921875, 0.9362083333333333], "iscrowd": 0, "area": 42179.6875, "rle": {"size": [480, 640], "counts": "UZd4>b>a1_Na1_Na1_Na1_NT1lN2N3M1O0000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000O100000001O0000000000000000000000000000000000000000000000001O1O1N2O1O1O1O001O1O1N2O2N3M4L3M3M3M4L3L4M3M4L3M3M3M4L4L7H7J6J6J7I6J6J6J7I6I7J6J7I6J6J6J7I6J6I7J7I6J6J6J7I6J6J6I\\ol1"}, "label": "the phone in the middle of other phones"}]}
{"id": 320125, "image": "COCO_train2014_000000320125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple cellphone\"."}], "task": "refering", "answer_template": "The \"the apple cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 361, 362, 363], "bbox": [0.494140625, 0.4101666666666667, 0.794921875, 0.9362083333333333], "iscrowd": 0, "area": 42179.6875, "rle": {"size": [480, 640], "counts": "UZd4>b>a1_Na1_Na1_Na1_NT1lN2N3M1O0000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000O100000001O0000000000000000000000000000000000000000000000001O1O1N2O1O1O1O001O1O1N2O2N3M4L3M3M3M4L3L4M3M4L3M3M3M4L4L7H7J6J6J7I6J6J6J7I6I7J6J7I6J6J6J7I6J6I7J7I6J6J6J7I6J6J6I\\ol1"}, "label": "the apple cellphone"}]}
{"id": 90751, "image": "COCO_train2014_000000090751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red double decker bus\"."}], "task": "refering", "answer_template": "The \"a red double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 466, 471, 472, 473], "bbox": [0.036765625, 0.04284090909090909, 0.9452031249999999, 0.9192532467532467], "iscrowd": 0, "area": 269897.78390000004, "rle": {"size": [616, 640], "counts": "]j>?`b0b0_O`0_Ob0_O`0_Ob0_O`0_Ob0^Ob0_O`0_Ob0_O`0_Ob0_Oa0^Oa0@a0^O?B8G9H8G9H8G9J7J5L4K5L4K5L4K5K5L5J5L4K5L4K5L4K5L4K6K4K5L4K5L4K5L4K5L5J5L4K5L4K5K5L4O101O000O100000000O10001O0O1000000O1000001N100000000O1000001N1000000O100000001N1000000O1000000O2O00000O1000000000001O00000O10000000001O00000000000000001O0O100000000000001O0000001O1O1O0O2O1O0000000000O1000000000000O10O10O10000O10000O10000O10000O1000O010000O1000000000000000000000O10000000000000000000001O0000000O10000000000000000000000000000O1000001O00000000000000000000000O1000000000000000001O00000000000O10000001O000000000000000000000001O00000000000000000000000000000001O000000000001O00000000000000000000000000000000000000000010O0000000000000000000000000000000000000000001O000O2O00001N10001O000O2O00001O0O101O001O0O2O1O001O0O2O1O001N2O001O1O0O2O1O001O1N101O1O001N2O001O1N110O1O010O10O01O10O0100O010O1O010O10O01O10O0100O0O2M3N1N3N1O2M3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2N101N2O001O1N101O1O0O2O1O001N2O001N2O001O1N101O1O0O2O1O001N2O001N2O001O1N101O001N2O001O1N101O1N101O1O0O2O1O001N101O1O0O2O001N101O001N2O001O0O2O001O0O2O1O001O0100O10O001O1O001O001O0O2O001O1O001O0O2O001O001N2M2N3N1N3M2N3M2N3M3M2O2O001N101O001N2O001N101O001N4Mf0ZOg0YOg0XOh0YOf0ZOd0\\O`0_Oa0@`0@a0^Oa0@`0mMofd0"}, "label": "a red double decker bus"}]}
{"id": 90751, "image": "COCO_train2014_000000090751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red double decker bus\"."}], "task": "refering", "answer_template": "The \"a red double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 466, 471, 472, 473], "bbox": [0.036765625, 0.04284090909090909, 0.9452031249999999, 0.9192532467532467], "iscrowd": 0, "area": 269897.78390000004, "rle": {"size": [616, 640], "counts": "]j>?`b0b0_O`0_Ob0_O`0_Ob0_O`0_Ob0^Ob0_O`0_Ob0_O`0_Ob0_Oa0^Oa0@a0^O?B8G9H8G9H8G9J7J5L4K5L4K5L4K5K5L5J5L4K5L4K5L4K5L4K6K4K5L4K5L4K5L4K5L5J5L4K5L4K5K5L4O101O000O100000000O10001O0O1000000O1000001N100000000O1000001N1000000O100000001N1000000O1000000O2O00000O1000000000001O00000O10000000001O00000000000000001O0O100000000000001O0000001O1O1O0O2O1O0000000000O1000000000000O10O10O10000O10000O10000O10000O1000O010000O1000000000000000000000O10000000000000000000001O0000000O10000000000000000000000000000O1000001O00000000000000000000000O1000000000000000001O00000000000O10000001O000000000000000000000001O00000000000000000000000000000001O000000000001O00000000000000000000000000000000000000000010O0000000000000000000000000000000000000000001O000O2O00001N10001O000O2O00001O0O101O001O0O2O1O001O0O2O1O001N2O001O1O0O2O1O001O1N101O1O001N2O001O1N110O1O010O10O01O10O0100O010O1O010O10O01O10O0100O0O2M3N1N3N1O2M3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2N101N2O001O1N101O1O0O2O1O001N2O001N2O001O1N101O1O0O2O1O001N2O001N2O001O1N101O001N2O001O1N101O1N101O1O0O2O1O001N101O1O0O2O001N101O001N2O001O0O2O001O0O2O1O001O0100O10O001O1O001O001O0O2O001O1O001O0O2O001O001N2M2N3N1N3M2N3M2N3M3M2O2O001N101O001N2O001N101O001N4Mf0ZOg0YOg0XOh0YOf0ZOd0\\O`0_Oa0@`0@a0^Oa0@`0mMofd0"}, "label": "a red double decker bus"}]}
{"id": 172680, "image": "COCO_train2014_000000172680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy pointing at a woman\"."}], "task": "refering", "answer_template": "The \"a boy pointing at a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 250, 251, 252, 260, 261, 267, 268, 269, 270, 271, 277, 278, 279, 280, 284, 285, 286, 287, 288, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 351, 352, 353, 354, 368, 369, 370, 371, 385, 386, 387, 388], "bbox": [0.2816458333333333, 0.584265625, 0.9617916666666667, 0.9887656249999999], "iscrowd": 0, "area": 27636.8333, "rle": {"size": [640, 480], "counts": "^]e21oc0010O001O1O010O001O1O010O001O0^\\O2Rc0f0H8I1N00001O01O000000001O000010O01O001O001O00001O010O001O001O001O001O01O01O001O001O001O0010O01O00001O001O001O010O001O0010O010O0010O010O010O0100O010O010O0010O010O0100O010O010O010O010O0010O0100O010O010O010O010O00100O010O010O010O0010O01O1O010O001O010O001O010O1O0010O01O010O001O010O1O0010O01O1O10O01O1O10O01O1O100O00100O1O00100O1O00100O1O00100O1O00l0UOn0QOl0TO01O0000001O0000001O0000001OlLPLXFP4h9SLUFm3j9WLSFi3m9ZLQFe3n9_LoEa3Q:bLlE^3S:fLjEY3W:jLcDWOHn3d;PM[DXO0h3e;SMSDZO8c3d;WMmC[O?]3d;ZMfC^Of0X3c;YMeCCf0V3d;UMdCKf0Q3f;QMbC3g0m2f;nLbC9g0j2g;kL_Ca0h0e2h;\\NWDe1i;\\NUDe1j;\\NUDe1k;[NSDg1l;[NRDf1n;ZNQDg1n;[NPDf1P<ZNoCg1P<ZNoCg1Q<ZNmCg1R<ZNmCg1S<ZNjCh1U<YNjCh1V<XNiCi1V<YNhCh1X<XNgCi1Y<o200O100000000O1000000O1000000O1000VIkCi5U<VJlCj5T<VJlCj5S<WJmCi5S<VJnCj5R<VJoCi5Q<WJoCi5P<WJQDi5o;WJQDi5o;WJQDi5o;WJRDh5m;XJTDh5l;XJTDh5l;XJTDg5m;XJTDh5k;YJVDf5j;ZJVDd29mNa;^NWDe2[1jM^:AWDe2^2gL[9d0WDd2k>\\MWAc2i>]MWAc2i>jLjAV3V>hLlAW3T>iLmAW3S>hLoAW3Q>hLPBW3Q>hLPBX3o=hLRBX3n=gLSBX3n=gLTBX3l=gLUBY3k=eLWBZ3i=fLXBZ3W?01N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2H7D<D<C=D=Coj:"}, "label": "a boy pointing at a woman"}]}
{"id": 172680, "image": "COCO_train2014_000000172680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"kid pointing at his mom\"."}], "task": "refering", "answer_template": "The \"kid pointing at his mom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 250, 251, 252, 260, 261, 267, 268, 269, 270, 271, 277, 278, 279, 280, 284, 285, 286, 287, 288, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 351, 352, 353, 354, 368, 369, 370, 371, 385, 386, 387, 388], "bbox": [0.2816458333333333, 0.584265625, 0.9617916666666667, 0.9887656249999999], "iscrowd": 0, "area": 27636.8333, "rle": {"size": [640, 480], "counts": "^]e21oc0010O001O1O010O001O1O010O001O0^\\O2Rc0f0H8I1N00001O01O000000001O000010O01O001O001O00001O010O001O001O001O001O01O01O001O001O001O0010O01O00001O001O001O010O001O0010O010O0010O010O010O0100O010O010O0010O010O0100O010O010O010O010O0010O0100O010O010O010O010O00100O010O010O010O0010O01O1O010O001O010O001O010O1O0010O01O010O001O010O1O0010O01O1O10O01O1O10O01O1O100O00100O1O00100O1O00100O1O00100O1O00l0UOn0QOl0TO01O0000001O0000001O0000001OlLPLXFP4h9SLUFm3j9WLSFi3m9ZLQFe3n9_LoEa3Q:bLlE^3S:fLjEY3W:jLcDWOHn3d;PM[DXO0h3e;SMSDZO8c3d;WMmC[O?]3d;ZMfC^Of0X3c;YMeCCf0V3d;UMdCKf0Q3f;QMbC3g0m2f;nLbC9g0j2g;kL_Ca0h0e2h;\\NWDe1i;\\NUDe1j;\\NUDe1k;[NSDg1l;[NRDf1n;ZNQDg1n;[NPDf1P<ZNoCg1P<ZNoCg1Q<ZNmCg1R<ZNmCg1S<ZNjCh1U<YNjCh1V<XNiCi1V<YNhCh1X<XNgCi1Y<o200O100000000O1000000O1000000O1000VIkCi5U<VJlCj5T<VJlCj5S<WJmCi5S<VJnCj5R<VJoCi5Q<WJoCi5P<WJQDi5o;WJQDi5o;WJQDi5o;WJRDh5m;XJTDh5l;XJTDh5l;XJTDg5m;XJTDh5k;YJVDf5j;ZJVDd29mNa;^NWDe2[1jM^:AWDe2^2gL[9d0WDd2k>\\MWAc2i>]MWAc2i>jLjAV3V>hLlAW3T>iLmAW3S>hLoAW3Q>hLPBW3Q>hLPBX3o=hLRBX3n=gLSBX3n=gLTBX3l=gLUBY3k=eLWBZ3i=fLXBZ3W?01N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2H7D<D<C=D=Coj:"}, "label": "kid pointing at his mom"}]}
{"id": 172680, "image": "COCO_train2014_000000172680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing in a kitchen\"."}], "task": "refering", "answer_template": "The \"woman standing in a kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 72, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 136, 137, 138, 139, 153, 154, 155, 156, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 272, 273, 274, 275, 276, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378], "bbox": [0.0, 0.172578125, 0.3012708333333334, 0.986890625], "iscrowd": 0, "area": 63813.97565, "rle": {"size": [640, 480], "counts": "U4b5]>R:nE100O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1000000O10000000000O10000000000O1000000000000000000O10000000000000000000000O100000000000000000000O1001O00001O00001O0000001O00001O00001O00001O00001O00001O000RDZLb7g3]HbLZ7^3eHeLY7\\3eHgLY7Y3fHjLX7V3hHlLV7U3hHnLV7R3iHQMU7P3iHRMV7n2jHTMT7m2dH\\MZ7d2_HdM`7\\2_HgM_7Z2_HiM_7W2`HkM_7V2_HmM_7S2`HPN^7Q2`HQN_7o1`HnNd6R1ZIROd6o0ZIUOP6^1oIeNT5W2jJlMY4o2dKVMi3\\3RLkL]3c3`LbLP3m3kLYLe2U4lL]Ld2P4[MSLV2[4hMhKQ2_4lMeKm1b4PN`Ki1g4TN\\Kf1k4VNXKd1o4XNUK`1R5^NPK\\1W5`NlJY1\\5cNhJV1^5gNeJS1b5iNaJP1g5mN\\J:]6BS7M4L4M2McVa6"}, "label": "woman standing in a kitchen"}]}
{"id": 172680, "image": "COCO_train2014_000000172680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman wearing a grey shirt and a scarf\"."}], "task": "refering", "answer_template": "The \"a blonde woman wearing a grey shirt and a scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 72, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 136, 137, 138, 139, 153, 154, 155, 156, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 272, 273, 274, 275, 276, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378], "bbox": [0.0, 0.172578125, 0.3012708333333334, 0.986890625], "iscrowd": 0, "area": 63813.97565, "rle": {"size": [640, 480], "counts": "U4b5]>R:nE100O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1000000O10000000000O10000000000O1000000000000000000O10000000000000000000000O100000000000000000000O1001O00001O00001O0000001O00001O00001O00001O00001O00001O000RDZLb7g3]HbLZ7^3eHeLY7\\3eHgLY7Y3fHjLX7V3hHlLV7U3hHnLV7R3iHQMU7P3iHRMV7n2jHTMT7m2dH\\MZ7d2_HdM`7\\2_HgM_7Z2_HiM_7W2`HkM_7V2_HmM_7S2`HPN^7Q2`HQN_7o1`HnNd6R1ZIROd6o0ZIUOP6^1oIeNT5W2jJlMY4o2dKVMi3\\3RLkL]3c3`LbLP3m3kLYLe2U4lL]Ld2P4[MSLV2[4hMhKQ2_4lMeKm1b4PN`Ki1g4TN\\Kf1k4VNXKd1o4XNUK`1R5^NPK\\1W5`NlJY1\\5cNhJV1^5gNeJS1b5iNaJP1g5mN\\J:]6BS7M4L4M2McVa6"}, "label": "a blonde woman wearing a grey shirt and a scarf"}]}
{"id": 139914, "image": "COCO_train2014_000000139914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard of a laptop computer\"."}], "task": "refering", "answer_template": "The \"the keyboard of a laptop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 115, 116, 117, 118, 131, 132, 133, 134, 135, 149, 150, 151, 152, 167, 168, 169, 185, 186, 203], "bbox": [0.7453125, 0.222875, 1.0, 0.5045624999999999], "iscrowd": 0, "area": 11387.4111, "rle": {"size": [640, 480], "counts": "[no61nc02N2N2O1N2N2N2O1N2N1O2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N1O2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N1O2N2O0O2N2N1O2O1N2N101O10O01O100O010O10O0100O10O01O10O0100O10O0100O010O100O1O2O1N2O1N2O1N2O1N2O1N2N2OL3ZNT2mMVL"}, "label": "the keyboard of a laptop computer"}]}
{"id": 139914, "image": "COCO_train2014_000000139914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop keyboard sitting on black mat\"."}], "task": "refering", "answer_template": "The \"laptop keyboard sitting on black mat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 115, 116, 117, 118, 131, 132, 133, 134, 135, 149, 150, 151, 152, 167, 168, 169, 185, 186, 203], "bbox": [0.7453125, 0.222875, 1.0, 0.5045624999999999], "iscrowd": 0, "area": 11387.4111, "rle": {"size": [640, 480], "counts": "[no61nc02N2N2O1N2N2N2O1N2N1O2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N1O2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N1O2N2O0O2N2N1O2O1N2N101O10O01O100O010O10O0100O10O01O10O0100O10O0100O010O100O1O2O1N2O1N2O1N2O1N2O1N2N2OL3ZNT2mMVL"}, "label": "laptop keyboard sitting on black mat"}]}
{"id": 656, "image": "COCO_train2014_000000000656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver car in the background reflecting off of the window\"."}], "task": "refering", "answer_template": "The \"the silver car in the background reflecting off of the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78], "bbox": [0.01909375, 0.00163135593220339, 0.4737343750000001, 0.2298728813559322], "iscrowd": 0, "area": 24684.597850000002, "rle": {"size": [472, 640], "counts": "[d52_>6K6I7J6I6K6I7J6I6K6I7J6I7J5J7J6N20O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000O100000000O100000000O100000000O100000000N2O1O1O1N2O1O1O1N2O1O1O1N200O100O100O10000O100O1N2O1O1N2O1O1O1N2O1O1O1N2O1O3L4M3M3M3L4M8H<CkYk4"}, "label": "the silver car in the background reflecting off of the window"}]}
{"id": 656, "image": "COCO_train2014_000000000656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection of a suv\"."}], "task": "refering", "answer_template": "The \"a reflection of a suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78], "bbox": [0.01909375, 0.00163135593220339, 0.4737343750000001, 0.2298728813559322], "iscrowd": 0, "area": 24684.597850000002, "rle": {"size": [472, 640], "counts": "[d52_>6K6I7J6I6K6I7J6I6K6I7J6I7J5J7J6N20O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000O100000000O100000000O100000000O100000000N2O1O1O1N2O1O1O1N2O1O1O1N200O100O100O10000O100O1N2O1O1N2O1O1O1N2O1O1O1N2O1O3L4M3M3M3L4M8H<CkYk4"}, "label": "a reflection of a suv"}]}
{"id": 656, "image": "COCO_train2014_000000000656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car parked behind a motorcycle\"."}], "task": "refering", "answer_template": "The \"a car parked behind a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 63, 64, 66, 67, 68, 91, 113, 114, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.721640625, 0.0021610169491525426, 1.0, 0.4043008474576271], "iscrowd": 0, "area": 15399.651050000004, "rle": {"size": [472, 640], "counts": "UPe63a>4M3L5K4O1O101O0000001O0000001O0000001XBXO^=h0]B]Oc=k00100000O1000000O10000O100000O0O2O1N2N1O2N2N2O0O2N2N2N1O2N2O0O2N2N20O1000000O100000000O1000000000000O1000000000kF\\NV5c1iJbNT5^1jJcNW5]1hJdNX5\\1fJfNZ5Z1dJhN\\5X1bJjN^5V1aJkN_5T1bJlN^5T1aJmN_5S1aJmN_5S1`JnN`5R1`JnN`5R1_JoNa5Q1_JoNa5Q1^JPOb5P1^JPOb5P1^JPOb5P1]JQOc5o0]JQOc5P1[JROd5n0\\JROd5n0[JSOe5m0[JSOe5m0ZJTOf5l0ZJTOf5l0YJUOg5k0YJUOg5k0XJVOh5k0WJUOi5k0VJVOj5j0VJVOj5j0UJWOk5i0UJWOk5i0TJXOl5h0TJXOl5h0SJYOm5g0SJYOm5h0RJXOn5h0QJZOn5f0RJZOn5f0QJ[Oo5e0QJ[Oo5e0PJ\\OP6d0PJ\\OP6d0oI]OQ6c0oI]OQ6d0mI]OS6c0lI^OT6b0jI@V6a0gIAY6`0eIA[6`0bIB^6?`IB`6?]ICc6?ZIBf6?WICi6>TIDl6=RIEm6<PIFP7<mHES7<jHFV7;gHGY7:eHG[7:bHH^7:_HGa7:\\HHd79YHIg78WHIi78THJl78QHIo78nGJR87lGJT88hGJX87eGK[87bGJ^87_GKa86]GLb86ZGLf85WGMi8Q300000000000000000000000000000001O00000<Df3ZLmK"}, "label": "a car parked behind a motorcycle"}]}
{"id": 656, "image": "COCO_train2014_000000000656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"car\"."}], "task": "refering", "answer_template": "The \"car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 63, 64, 66, 67, 68, 91, 113, 114, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.721640625, 0.0021610169491525426, 1.0, 0.4043008474576271], "iscrowd": 0, "area": 15399.651050000004, "rle": {"size": [472, 640], "counts": "UPe63a>4M3L5K4O1O101O0000001O0000001O0000001XBXO^=h0]B]Oc=k00100000O1000000O10000O100000O0O2O1N2N1O2N2N2O0O2N2N2N1O2N2O0O2N2N20O1000000O100000000O1000000000000O1000000000kF\\NV5c1iJbNT5^1jJcNW5]1hJdNX5\\1fJfNZ5Z1dJhN\\5X1bJjN^5V1aJkN_5T1bJlN^5T1aJmN_5S1aJmN_5S1`JnN`5R1`JnN`5R1_JoNa5Q1_JoNa5Q1^JPOb5P1^JPOb5P1^JPOb5P1]JQOc5o0]JQOc5P1[JROd5n0\\JROd5n0[JSOe5m0[JSOe5m0ZJTOf5l0ZJTOf5l0YJUOg5k0YJUOg5k0XJVOh5k0WJUOi5k0VJVOj5j0VJVOj5j0UJWOk5i0UJWOk5i0TJXOl5h0TJXOl5h0SJYOm5g0SJYOm5h0RJXOn5h0QJZOn5f0RJZOn5f0QJ[Oo5e0QJ[Oo5e0PJ\\OP6d0PJ\\OP6d0oI]OQ6c0oI]OQ6d0mI]OS6c0lI^OT6b0jI@V6a0gIAY6`0eIA[6`0bIB^6?`IB`6?]ICc6?ZIBf6?WICi6>TIDl6=RIEm6<PIFP7<mHES7<jHFV7;gHGY7:eHG[7:bHH^7:_HGa7:\\HHd79YHIg78WHIi78THJl78QHIo78nGJR87lGJT88hGJX87eGK[87bGJ^87_GKa86]GLb86ZGLf85WGMi8Q300000000000000000000000000000001O00000<Df3ZLmK"}, "label": "car"}]}
{"id": 336532, "image": "COCO_train2014_000000336532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child is playing base ball\"."}], "task": "refering", "answer_template": "The \"the child is playing base ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 146, 147, 148, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 198, 199, 200, 215, 216, 217, 218, 232, 233, 234, 249, 250, 251, 266, 267, 268, 283, 284, 285, 300, 301, 302, 316, 317, 318, 319, 334, 335], "bbox": [0.59625, 0.339328125, 0.8898958333333333, 0.831453125], "iscrowd": 0, "area": 23708.103650000005, "rle": {"size": [640, 480], "counts": "eob53lc06J7H8I6I8I4K2O2M2O1N3N1O2M2O1O2[EUNS3k1lLVNS3k1lLWNS3l1jLTNV3R2cLoM\\3X2]LjMb3\\2WLeMi3`2QLbMm3a2PL`MP4c2lK_MS4d2lHXM]O4f7h2bHfM]ODP8i2XHUN]OSO[8k2lGdN_ObNd8m2cGRO_ORNn8o2WGAAaMX9i7^FXHb9R8TFoGj9[8lEfGT:S91O00001O001O00001O01O01O00001O001O00001O000010O0001O001O00003Md0\\O6J1O0O100000000O2O0000000O10001O00000O1000001O0O10000mLiD]MW;`2[EQMe:m2lEcLV:Y3Z3N2M3N2M3N2M3N3iL]@k1g?UN`@\\1h?dN^@n0j?RO\\@`0l?@Z@3m?MY@Eo?;X@VOQ`0i0c100000000O101N102M2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N3N1N2N2OQkP1"}, "label": "the child is playing base ball"}]}
{"id": 336532, "image": "COCO_train2014_000000336532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player holding a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player holding a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 146, 147, 148, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 198, 199, 200, 215, 216, 217, 218, 232, 233, 234, 249, 250, 251, 266, 267, 268, 283, 284, 285, 300, 301, 302, 316, 317, 318, 319, 334, 335], "bbox": [0.59625, 0.339328125, 0.8898958333333333, 0.831453125], "iscrowd": 0, "area": 23708.103650000005, "rle": {"size": [640, 480], "counts": "eob53lc06J7H8I6I8I4K2O2M2O1N3N1O2M2O1O2[EUNS3k1lLVNS3k1lLWNS3l1jLTNV3R2cLoM\\3X2]LjMb3\\2WLeMi3`2QLbMm3a2PL`MP4c2lK_MS4d2lHXM]O4f7h2bHfM]ODP8i2XHUN]OSO[8k2lGdN_ObNd8m2cGRO_ORNn8o2WGAAaMX9i7^FXHb9R8TFoGj9[8lEfGT:S91O00001O001O00001O01O01O00001O001O00001O000010O0001O001O00003Md0\\O6J1O0O100000000O2O0000000O10001O00000O1000001O0O10000mLiD]MW;`2[EQMe:m2lEcLV:Y3Z3N2M3N2M3N2M3N3iL]@k1g?UN`@\\1h?dN^@n0j?RO\\@`0l?@Z@3m?MY@Eo?;X@VOQ`0i0c100000000O101N102M2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N3N1N2N2OQkP1"}, "label": "a baseball player holding a bat"}]}
{"id": 25237, "image": "COCO_train2014_000000025237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little boy wearing a dark shirt , brushing his teeth\"."}], "task": "refering", "answer_template": "The \"little boy wearing a dark shirt , brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 138, 139, 140, 141, 155, 156, 157, 158, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 206, 207, 208, 223, 224, 225, 241, 242, 258, 259, 275, 276, 292, 293, 309, 310, 311, 327, 328, 329, 344, 345, 362], "bbox": [0.14852697095435685, 0.30670312499999997, 0.42477178423236517, 0.930859375], "iscrowd": 0, "area": 22190.809600000004, "rle": {"size": [640, 482], "counts": "PW]1h0Y26f=4RB0i=:mAKn=>iAGR>c0eAAU>h0cA]OX>e0fA@U>b0iACQ>`0nADm==RBHi=:UBKf=7YBMa=5^B0]=2aB3Z=OdB6X=KgB9U=GjB>R=ClBa0Q=@nBa0P=@oBb0P=_OnBc0Q=^OmBd0R=\\OnBe0Q=\\OmBf0Q=\\OmBg0Q=U3O1O1O1N2O2N1O1O1N=Dg0YOg0YO9G3L4M4L3M3M3L4M3M3M3M3M3LZNlFeHQ9R7YGPId8g6fG\\IV8[6UHfIh7Q6bHQJZ7g5PI[Jm6\\5]IfJ`6Q5kIPKR6g4XJaK_5V4kJjKS5o3VKQLi4f3bKYL]4^3mKaLS4V3WLjLh3m2bLSM\\3e2nL[MQ3\\2ZMbMg2Z2]MfMb2Y2`MgM`2X2aMhM^2W2eMgM\\2X2eMhMZ2W2hMiMX2U2jMkMU2U2mMiMT2U2nMkMR2T2oMlMP2S2RNmMn1Q2TNnMl1R2VNmMj1Q2XNoMg1Q2ZNoMf1o1\\NPNd1P2^NoMb1o1`NQN`1m1bNSN]1m1dNRN]1l1eNRN\\1n1cNPNa1n1`NoMc1P2]NnMg1Q2YNlMj1S2VNkM\\2f1dMWNn2Y1RMeNa3k0_LROS4>mK_O`47aKFb4:]KDg4:ZKCi4=VKAn4=SK@Q5>oJ@T5`0lJ]OX5a0hJ]O[5c0eJZO_5d0aJZOc5e0]JXOf5g0ZJWOj5g0h601N2O0O2N101N2O0O2O0O2N2O0O2O0O2N2O0O2O0O2Od^]5"}, "label": "little boy wearing a dark shirt , brushing his teeth"}]}
{"id": 25237, "image": "COCO_train2014_000000025237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy brushing his teeth\"."}], "task": "refering", "answer_template": "The \"a small boy brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 138, 139, 140, 141, 155, 156, 157, 158, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 206, 207, 208, 223, 224, 225, 241, 242, 258, 259, 275, 276, 292, 293, 309, 310, 311, 327, 328, 329, 344, 345, 362], "bbox": [0.14852697095435685, 0.30670312499999997, 0.42477178423236517, 0.930859375], "iscrowd": 0, "area": 22190.809600000004, "rle": {"size": [640, 482], "counts": "PW]1h0Y26f=4RB0i=:mAKn=>iAGR>c0eAAU>h0cA]OX>e0fA@U>b0iACQ>`0nADm==RBHi=:UBKf=7YBMa=5^B0]=2aB3Z=OdB6X=KgB9U=GjB>R=ClBa0Q=@nBa0P=@oBb0P=_OnBc0Q=^OmBd0R=\\OnBe0Q=\\OmBf0Q=\\OmBg0Q=U3O1O1O1N2O2N1O1O1N=Dg0YOg0YO9G3L4M4L3M3M3L4M3M3M3M3M3LZNlFeHQ9R7YGPId8g6fG\\IV8[6UHfIh7Q6bHQJZ7g5PI[Jm6\\5]IfJ`6Q5kIPKR6g4XJaK_5V4kJjKS5o3VKQLi4f3bKYL]4^3mKaLS4V3WLjLh3m2bLSM\\3e2nL[MQ3\\2ZMbMg2Z2]MfMb2Y2`MgM`2X2aMhM^2W2eMgM\\2X2eMhMZ2W2hMiMX2U2jMkMU2U2mMiMT2U2nMkMR2T2oMlMP2S2RNmMn1Q2TNnMl1R2VNmMj1Q2XNoMg1Q2ZNoMf1o1\\NPNd1P2^NoMb1o1`NQN`1m1bNSN]1m1dNRN]1l1eNRN\\1n1cNPNa1n1`NoMc1P2]NnMg1Q2YNlMj1S2VNkM\\2f1dMWNn2Y1RMeNa3k0_LROS4>mK_O`47aKFb4:]KDg4:ZKCi4=VKAn4=SK@Q5>oJ@T5`0lJ]OX5a0hJ]O[5c0eJZO_5d0aJZOc5e0]JXOf5g0ZJWOj5g0h601N2O0O2N101N2O0O2O0O2N2O0O2O0O2N2O0O2O0O2Od^]5"}, "label": "a small boy brushing his teeth"}]}
{"id": 25237, "image": "COCO_train2014_000000025237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl brushing her teeth\"."}], "task": "refering", "answer_template": "The \"the girl brushing her teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 149, 150, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 293, 294, 295, 296, 297, 298, 299, 300, 301, 310, 311, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 362, 363, 364, 365, 366, 367, 368, 369, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.2554564315352697, 0.197078125, 0.9832157676348547, 0.9845625], "iscrowd": 0, "area": 107952.33215000002, "rle": {"size": [640, 482], "counts": "[[]23bc0=C<D<D<D<D<E;D<D=C<D<D<D<E;O1N2N3N1N2N2O1N2N2O2M2N2O1N2N2O1N2UEiIj7Z6SHgIm7Z6PHhIo7Z6oGhIo7X6PHjIP8V6nGlIQ8T6kGQJU8o5gGUJX8k5eGYJZ8g5cG]J]8c5_GaJ`8_5\\GfJc8[5YGiJg8V5VGnJi8S5SGQKl8o4PGVKP9j4lFZKS9f4jF^KU9c4gFaKY9^4dFfK[9[4`FjK_9V4^FnKb9R4ZFRLe9n3XFVLg9k3UFYLk9f3QF_Ln9a3oEcLP:^3lEfLT:c6O1O100O1O100O1O100O2O0O1O1O1L4L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4M3jLjC^NZ<b1iCXNZ<h1hCSN[<m1hClM\\<T2fCgM]<Y2fC`M^<_2eC\\M^<d2eCUM_<k2cCPM`<P3cCiLa<W3aCdLb<\\3aC]Lc<c3_CXLd<h3_CQLf<n3\\CmKg<S4\\CgKg<Y4[CaKi<_4\\100000000000000000000O10000000000000000000000000000000000000000O2O00000000000000000000000000000000000000000O100000000000000000000000001O000000000000000O10000000000000000000000000000000000000000O10000000001O01O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1\\KQLUIP4f6dMjG\\2R8XO^Fi0]9P5K5K5J6K5K4L5K5K5J6K5K5K5K4M4O1N2O1O1N2O001N2O1ZKeCe1\\<XNhCf1Y<XNjCe1X<YNjCf1V<XNnCf1S<XNPDe1R<XNRDf1o;XNTDf1m;XNVDe1l;XNXDf1i;XNZDf1f;YN]De1d;XN`De1b;YNaDe1`;YNcDe1^;XNeDf1];XNfDf1\\;VNhDh1Y;VNjDg1X;WNkDg1V;VNnDh1T;UNoDi1R;UNQEh1Q;UNSEi1n:UNUEi1m:TNVEi1l:TNWEk1j:SNYEk1h:RN\\Ek1f:SN]Ek1e:RN^El1c:QNaEm1`:QNcEl1_:RNdEl1^:PNfEn1[:PNhEm1Z:QNhEn1Y:oMiEQ2X:mMiER2S>000000O10000O10000O1N2O1O1O1N2O1O1O2N1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N3N3M2N2N3L3N3M[e4"}, "label": "the girl brushing her teeth"}]}
{"id": 25237, "image": "COCO_train2014_000000025237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl brushing her teeth and behind her is a small boy trying to brush teeth\"."}], "task": "refering", "answer_template": "The \"a young girl brushing her teeth and behind her is a small boy trying to brush teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 149, 150, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 293, 294, 295, 296, 297, 298, 299, 300, 301, 310, 311, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 362, 363, 364, 365, 366, 367, 368, 369, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.2554564315352697, 0.197078125, 0.9832157676348547, 0.9845625], "iscrowd": 0, "area": 107952.33215000002, "rle": {"size": [640, 482], "counts": "[[]23bc0=C<D<D<D<D<E;D<D=C<D<D<D<E;O1N2N3N1N2N2O1N2N2O2M2N2O1N2N2O1N2UEiIj7Z6SHgIm7Z6PHhIo7Z6oGhIo7X6PHjIP8V6nGlIQ8T6kGQJU8o5gGUJX8k5eGYJZ8g5cG]J]8c5_GaJ`8_5\\GfJc8[5YGiJg8V5VGnJi8S5SGQKl8o4PGVKP9j4lFZKS9f4jF^KU9c4gFaKY9^4dFfK[9[4`FjK_9V4^FnKb9R4ZFRLe9n3XFVLg9k3UFYLk9f3QF_Ln9a3oEcLP:^3lEfLT:c6O1O100O1O100O1O100O2O0O1O1O1L4L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4M3jLjC^NZ<b1iCXNZ<h1hCSN[<m1hClM\\<T2fCgM]<Y2fC`M^<_2eC\\M^<d2eCUM_<k2cCPM`<P3cCiLa<W3aCdLb<\\3aC]Lc<c3_CXLd<h3_CQLf<n3\\CmKg<S4\\CgKg<Y4[CaKi<_4\\100000000000000000000O10000000000000000000000000000000000000000O2O00000000000000000000000000000000000000000O100000000000000000000000001O000000000000000O10000000000000000000000000000000000000000O10000000001O01O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1\\KQLUIP4f6dMjG\\2R8XO^Fi0]9P5K5K5J6K5K4L5K5K5J6K5K5K5K4M4O1N2O1O1N2O001N2O1ZKeCe1\\<XNhCf1Y<XNjCe1X<YNjCf1V<XNnCf1S<XNPDe1R<XNRDf1o;XNTDf1m;XNVDe1l;XNXDf1i;XNZDf1f;YN]De1d;XN`De1b;YNaDe1`;YNcDe1^;XNeDf1];XNfDf1\\;VNhDh1Y;VNjDg1X;WNkDg1V;VNnDh1T;UNoDi1R;UNQEh1Q;UNSEi1n:UNUEi1m:TNVEi1l:TNWEk1j:SNYEk1h:RN\\Ek1f:SN]Ek1e:RN^El1c:QNaEm1`:QNcEl1_:RNdEl1^:PNfEn1[:PNhEm1Z:QNhEn1Y:oMiEQ2X:mMiER2S>000000O10000O10000O1N2O1O1O1N2O1O1O2N1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N3N3M2N2N3L3N3M[e4"}, "label": "a young girl brushing her teeth and behind her is a small boy trying to brush teeth"}]}
{"id": 148118, "image": "COCO_train2014_000000148118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird in the background sitting on branches\"."}], "task": "refering", "answer_template": "The \"bird in the background sitting on branches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 83, 84, 85, 86, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153], "bbox": [0.39773437500000003, 0.1579294117647059, 0.7540625000000001, 0.45856470588235293], "iscrowd": 0, "area": 11819.842900000003, "rle": {"size": [425, 640], "counts": "]PZ37Q=1O2N1O1O2N1O001O100O1000O0100O10000O010O10O10O010O010O0100O010O010O1O010O001O010O1O0010O01000O2O0O10001N100O101N10000O2O0O10000O2O0O100O10000O100O101N100O100O10N1O2O0O2O1N1010O01O001O0010O100O100XDZOl:g0PE^Oo:b0lDCS;>jDFU;:hDIW;8fDKZ;R101N100O100O100O100O100O100O1O101N100O1O100O100O100O1O100O100O2N100O100O00100O010O00100O010O010O1O010O010O1O010O010M3K4M4O00100O010O00100O010O010O01O00010O001O010O001O10O01O1O10OO2O1O1O001O1N2O001O1O1O001N2N2G8H9G9M3N1O2M3N2N1N3N4L5K4L5L4L3M4L4L3M\\RQ2"}, "label": "bird in the background sitting on branches"}]}
{"id": 148118, "image": "COCO_train2014_000000148118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown bird standing behind another bird\"."}], "task": "refering", "answer_template": "The \"a dark brown bird standing behind another bird\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 83, 84, 85, 86, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153], "bbox": [0.39773437500000003, 0.1579294117647059, 0.7540625000000001, 0.45856470588235293], "iscrowd": 0, "area": 11819.842900000003, "rle": {"size": [425, 640], "counts": "]PZ37Q=1O2N1O1O2N1O001O100O1000O0100O10000O010O10O10O010O010O0100O010O010O1O010O001O010O1O0010O01000O2O0O10001N100O101N10000O2O0O10000O2O0O100O10000O100O101N100O100O10N1O2O0O2O1N1010O01O001O0010O100O100XDZOl:g0PE^Oo:b0lDCS;>jDFU;:hDIW;8fDKZ;R101N100O100O100O100O100O100O1O101N100O1O100O100O100O1O100O100O2N100O100O00100O010O00100O010O010O1O010O010O1O010O010M3K4M4O00100O010O00100O010O010O01O00010O001O010O001O10O01O1O10OO2O1O1O001O1N2O001O1O1O001N2N2G8H9G9M3N1O2M3N2N1N3N4L5K4L5L4L3M4L4L3M\\RQ2"}, "label": "a dark brown bird standing behind another bird"}]}
{"id": 148118, "image": "COCO_train2014_000000148118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown and white bird in the foreground . the bird is looking towards the left\"."}], "task": "refering", "answer_template": "The \"the brown and white bird in the foreground . the bird is looking towards the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264], "bbox": [0.360859375, 0.4615294117647059, 0.7252812499999999, 0.7654823529411766], "iscrowd": 0, "area": 9109.368950000002, "rle": {"size": [425, 640], "counts": "]`P32V=1O1O1O1O1O1N2O001O1HEaC<_<H\\C9c<7O1O101N1O2N2O0O2N101N1O2N2O0O2N101O001O1N102N100O2N1O2O0O2O001O010O00YN`De1b;1O0O2O001N101N101O00000000000000O10iEVNn8j1nFYNR9h1mFXNT9h1jFYNW9g1gFZNY9g1fFYNZ9g1eFZN\\9f1cFZN]9g1aFZN_9g1`FXNb9h1\\FXNe9i1YFXNg9i1XFVNj9i1UFWNl9j1SFVNm9k1QFUNQ:T2eElM[:`21O2O0O1O2O0O1O2O0O101O0O101O00001O001O001O001N100O2O001N101O001N101O0O2O00001N101N1L5L3M4N1O2M200O2N100O101N100O1O100O100O10O10O10000O10O10O10000O01000O100O01000O1000O0100O1O001O1N2N2K401O1O100O001O1O10O01000000O01000O1000O0100000O010000O10O10O100000O010000O01000O1000O01000000O01000KPC0Q=OPC1o<0RCOn<1500O10VjX2"}, "label": "the brown and white bird in the foreground . the bird is looking towards the left"}]}
{"id": 148118, "image": "COCO_train2014_000000148118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small bird standing on a twig\"."}], "task": "refering", "answer_template": "The \"a small bird standing on a twig\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264], "bbox": [0.360859375, 0.4615294117647059, 0.7252812499999999, 0.7654823529411766], "iscrowd": 0, "area": 9109.368950000002, "rle": {"size": [425, 640], "counts": "]`P32V=1O1O1O1O1O1N2O001O1HEaC<_<H\\C9c<7O1O101N1O2N2O0O2N101N1O2N2O0O2N101O001O1N102N100O2N1O2O0O2O001O010O00YN`De1b;1O0O2O001N101N101O00000000000000O10iEVNn8j1nFYNR9h1mFXNT9h1jFYNW9g1gFZNY9g1fFYNZ9g1eFZN\\9f1cFZN]9g1aFZN_9g1`FXNb9h1\\FXNe9i1YFXNg9i1XFVNj9i1UFWNl9j1SFVNm9k1QFUNQ:T2eElM[:`21O2O0O1O2O0O1O2O0O101O0O101O00001O001O001O001N100O2O001N101O001N101O0O2O00001N101N1L5L3M4N1O2M200O2N100O101N100O1O100O100O10O10O10000O10O10O10000O01000O100O01000O1000O0100O1O001O1N2N2K401O1O100O001O1O10O01000000O01000O1000O0100000O010000O10O10O100000O010000O01000O1000O01000000O01000KPC0Q=OPC1o<0RCOn<1500O10VjX2"}, "label": "a small bird standing on a twig"}]}
{"id": 107156, "image": "COCO_train2014_000000107156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one women wearing black dress is playing video game\"."}], "task": "refering", "answer_template": "The \"there is one women wearing black dress is playing video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 36, 37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 132, 133, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.449515625, 0.0026056338028169015, 0.90075, 0.9870187793427231], "iscrowd": 0, "area": 73266.09195, "rle": {"size": [426, 640], "counts": "`Ph37o<7J4M2O001O1N101O1O0O2J6Gc0QEVOa8S3E;I6K6I7I6K6I7I6K4L2N2O0O2O1N2O0O2O1oJXJ]2j5PMhJP3X5dLUKZ3m4_LYKa3h4YL]Kf3e4SLaKm3_4nKfKQ4\\4hKjKX4d6O100O100O1002N1O1O00O1O1O100O1O1O100O1O1O100O1O1O100N2O1O1O1O1O1PNmFHT95]G\\Od8b0kGPOV8n0YHdNh7Y1V2N2N2M3N2N2M3N2N2M3N2N2N2O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100001O00000000001O00000000001O0000001O00001O0000001O00001O00001O0000001O002N1O1O1O2N1O1O1O2N1gFTOR6m0lIUOS6l0jIWOU6k0hIWOW6j0fIXOZ6i0dIYO[6i0aIZO^6j0]IXOb6n0VIUOi6Q1PIPOP7V1hHmNW7Y1aHjN^7[1XHkNg7]3O001O1O1O001O1O0000001O0000001O0000001O000000001O0000001O0O10001O00001O2N1O2N1O2N2N1O2N1O2VLTNRLm1^3Y4L40O10O0100O0010O000O2O000O2O00003L4M2N3L3N001N2O001O1N101O1N101O1N101O1O0O2O1O0O2O1N101N2O0O2O0O2O0O101N100O2N100O2N1O1N3J5I7I8H7H8I8Ha[j0"}, "label": "there is one women wearing black dress is playing video game"}]}
{"id": 107156, "image": "COCO_train2014_000000107156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a caucasian woman in a black shirt is holding a set of nintendo wii controllers\"."}], "task": "refering", "answer_template": "The \"a caucasian woman in a black shirt is holding a set of nintendo wii controllers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 36, 37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 132, 133, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.449515625, 0.0026056338028169015, 0.90075, 0.9870187793427231], "iscrowd": 0, "area": 73266.09195, "rle": {"size": [426, 640], "counts": "`Ph37o<7J4M2O001O1N101O1O0O2J6Gc0QEVOa8S3E;I6K6I7I6K6I7I6K4L2N2O0O2O1N2O0O2O1oJXJ]2j5PMhJP3X5dLUKZ3m4_LYKa3h4YL]Kf3e4SLaKm3_4nKfKQ4\\4hKjKX4d6O100O100O1002N1O1O00O1O1O100O1O1O100O1O1O100O1O1O100N2O1O1O1O1O1PNmFHT95]G\\Od8b0kGPOV8n0YHdNh7Y1V2N2N2M3N2N2M3N2N2M3N2N2N2O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100001O00000000001O00000000001O0000001O00001O0000001O00001O00001O0000001O002N1O1O1O2N1O1O1O2N1gFTOR6m0lIUOS6l0jIWOU6k0hIWOW6j0fIXOZ6i0dIYO[6i0aIZO^6j0]IXOb6n0VIUOi6Q1PIPOP7V1hHmNW7Y1aHjN^7[1XHkNg7]3O001O1O1O001O1O0000001O0000001O0000001O000000001O0000001O0O10001O00001O2N1O2N1O2N2N1O2N1O2VLTNRLm1^3Y4L40O10O0100O0010O000O2O000O2O00003L4M2N3L3N001N2O001O1N101O1N101O1N101O1O0O2O1O0O2O1N101N2O0O2O0O2O0O101N100O2N100O2N1O1N3J5I7I8H7H8I8Ha[j0"}, "label": "a caucasian woman in a black shirt is holding a set of nintendo wii controllers"}]}
{"id": 246423, "image": "COCO_train2014_000000246423.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black expedition driving through an intersection\"."}], "task": "refering", "answer_template": "The \"black expedition driving through an intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5895312500000001, 0.8277083333333334, 1.0, 0.9853541666666668], "iscrowd": 0, "area": 15133.644450000005, "rle": {"size": [480, 640], "counts": "eUa51k>4K5L4L4K5O100O100000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O1000000O1000001O0O100000000O1000000O1000000O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O100O10000000000000000O1000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O101O00000000000000000000000000000000000000000000000O100001O1O2N1O2N1O1O2N1O2N1O1O2NQC"}, "label": "black expedition driving through an intersection"}]}
{"id": 246423, "image": "COCO_train2014_000000246423.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suv in front\"."}], "task": "refering", "answer_template": "The \"black suv in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5895312500000001, 0.8277083333333334, 1.0, 0.9853541666666668], "iscrowd": 0, "area": 15133.644450000005, "rle": {"size": [480, 640], "counts": "eUa51k>4K5L4L4K5O100O100000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O1000000O1000001O0O100000000O1000000O1000000O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O100O10000000000000000O1000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O101O00000000000000000000000000000000000000000000000O100001O1O2N1O2N1O1O2N1O2N1O1O2NQC"}, "label": "black suv in front"}]}
{"id": 295578, "image": "COCO_train2014_000000295578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher wearing the red equipment\"."}], "task": "refering", "answer_template": "The \"the catcher wearing the red equipment\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 194, 195, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 267, 287, 288, 289, 290, 311, 312, 313], "bbox": [0.407015625, 0.23354166666666665, 0.652203125, 0.8211666666666666], "iscrowd": 0, "area": 23027.210999999996, "rle": {"size": [480, 640], "counts": "PRj31h>>B:M001O1O1O0O2O1O1O4L3M4L2M2O2N1O1O2N1O1N2O1O000bEZNb7e1\\H^Nd7b1YHbNf7^1WHeNi7[1VHgNi7X1WHiNi7W1UHlNj7T1UHnN[6HYH[1[1nNX6OYHR1_1QOU64VHl0d1QOT68UHf0g1TOR68UHe0h1UOP68WHc0i1VOn5:WH`0j1YOn58WH?j1ZOo58UH>l1\\On57TH>m1\\Oo56SH>m1_Oo54RH>n1@o52QH`0o1_Oo53oG`0P2@P6P2oIRNP6n1oISNQ6m1nIUNQ6k1mIWNS6j1kIXNT6h1kIZNS6g1lIZNT6f1jI]NT6d1kI^NT6d1iI]NX6c1fI_NY6c1cI`N\\6a1bI`N^6b1_I`N`6b1]I`Nb6a1[IaNf6`1WIbNh6`1UIbNj6_1TIbNl6`1PIcNo6^1lHgNS7[1eHkN[7^40O01O1O001O010O1O0100O101O0nH_Hc6b7ZI`Hg6`7UIdHj6^7RIdHo6f71O010fHPIb6P7XIWIg6j6SI]Il6c6nHcIo6_6oHdIo6Y7M1N12O001N2O0O2O1N101O1N101N2O0O2M4M2N3nL]HbNe7\\1^HaNe7\\1]HbNe7^NUHg18Ie7_NUHe19Ie7`NTHd19Je7bNTH`1:Kd7dNTH^1:Ld7eNUH[19Oc7fNUHX1;1`7fNWHV1;3_7fNXHT1<4]7gNZHQ1;7[7iN[Hm0=8Y7jN\\Hk0=:X7jN]Hh0>=V7kN]Hd0a0?S7lN^H`0c0c0P7lN`H<d0f0m6nN`H7g0j0j6nNaH4h0m0h6mNcH1j0P1d6nNnHDa0]1b6nN\\KP1e4nN]KP1e4oN]Kn0e4QO\\Km0g4PO\\Km0i4oNXKo0l4mNWKo0o4mNRKQ1a9M2N2N3M2N3M2M4M2N3M3M3MRVX3"}, "label": "the catcher wearing the red equipment"}]}
{"id": 295578, "image": "COCO_train2014_000000295578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher that is about to throw the baseball\"."}], "task": "refering", "answer_template": "The \"a catcher that is about to throw the baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 194, 195, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 267, 287, 288, 289, 290, 311, 312, 313], "bbox": [0.407015625, 0.23354166666666665, 0.652203125, 0.8211666666666666], "iscrowd": 0, "area": 23027.210999999996, "rle": {"size": [480, 640], "counts": "PRj31h>>B:M001O1O1O0O2O1O1O4L3M4L2M2O2N1O1O2N1O1N2O1O000bEZNb7e1\\H^Nd7b1YHbNf7^1WHeNi7[1VHgNi7X1WHiNi7W1UHlNj7T1UHnN[6HYH[1[1nNX6OYHR1_1QOU64VHl0d1QOT68UHf0g1TOR68UHe0h1UOP68WHc0i1VOn5:WH`0j1YOn58WH?j1ZOo58UH>l1\\On57TH>m1\\Oo56SH>m1_Oo54RH>n1@o52QH`0o1_Oo53oG`0P2@P6P2oIRNP6n1oISNQ6m1nIUNQ6k1mIWNS6j1kIXNT6h1kIZNS6g1lIZNT6f1jI]NT6d1kI^NT6d1iI]NX6c1fI_NY6c1cI`N\\6a1bI`N^6b1_I`N`6b1]I`Nb6a1[IaNf6`1WIbNh6`1UIbNj6_1TIbNl6`1PIcNo6^1lHgNS7[1eHkN[7^40O01O1O001O010O1O0100O101O0nH_Hc6b7ZI`Hg6`7UIdHj6^7RIdHo6f71O010fHPIb6P7XIWIg6j6SI]Il6c6nHcIo6_6oHdIo6Y7M1N12O001N2O0O2O1N101O1N101N2O0O2M4M2N3nL]HbNe7\\1^HaNe7\\1]HbNe7^NUHg18Ie7_NUHe19Ie7`NTHd19Je7bNTH`1:Kd7dNTH^1:Ld7eNUH[19Oc7fNUHX1;1`7fNWHV1;3_7fNXHT1<4]7gNZHQ1;7[7iN[Hm0=8Y7jN\\Hk0=:X7jN]Hh0>=V7kN]Hd0a0?S7lN^H`0c0c0P7lN`H<d0f0m6nN`H7g0j0j6nNaH4h0m0h6mNcH1j0P1d6nNnHDa0]1b6nN\\KP1e4nN]KP1e4oN]Kn0e4QO\\Km0g4PO\\Km0i4oNXKo0l4mNWKo0o4mNRKQ1a9M2N2N3M2N3M2M4M2N3M3M3MRVX3"}, "label": "a catcher that is about to throw the baseball"}]}
{"id": 295578, "image": "COCO_train2014_000000295578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rugby player batting wearing blue color t shirt grey color lower\"."}], "task": "refering", "answer_template": "The \"a rugby player batting wearing blue color t shirt grey color lower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 70, 71, 72, 73, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 210, 211, 230, 231, 234, 235, 257, 258, 280, 281], "bbox": [0.0, 0.134125, 0.23739062500000002, 0.7278958333333333], "iscrowd": 0, "area": 20101.53030000001, "rle": {"size": [480, 640], "counts": "g81m>4M2M4M2PBIS=;eBMX=5aB2\\=2bBM]=6cBH]=:cBD\\=`0cB^O\\=e0dBYOZ=l0eBROZ=Q1fBmNY=]10LgBbNX=c10000O100O100O10O01N2N2N2N2]MRNRHQ2m7RNoGo1Q8TNkGn1T8VNgGl1X8WNdGk1[8XNaGj1^8ZN^Gg1a8\\N[Gf1d8]NXGd1g8aNTGa1j8cNRG_1l8eNPG\\1o8hNmFY1R9kNjFV1U9nNgFS1X9QOcFQ1\\9SO`Fn0_9VO]Fj0c9Y2001O00001O01O0101O000000001O0000000000001O0000000O1000001O1O1O1O001O1O1O11O2N3L3O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O100O1O1O1O1O1O001N2O001N104L3QHmIj6W6eHWJY7k6N1O1N3N1O2M2OgJUJ]1j5aN\\J\\1b5dNbJX1_5fNgJU1Z5jNjJR1V5mNoJP1Q5oNTKl0m4ROXKj0h4UO^Kf0c4YOaKc0`4[OfKa0Y4_OmK;T4CTL6m3IYL1l3I\\L1g3JaLOd3LbLNb3MfLL_3OhLJ]30jLKY31nLHW32QMGS34TMFQ35R6JTcT7"}, "label": "a rugby player batting wearing blue color t shirt grey color lower"}]}
{"id": 295578, "image": "COCO_train2014_000000295578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batter for the blue jays at home plate\"."}], "task": "refering", "answer_template": "The \"a batter for the blue jays at home plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 70, 71, 72, 73, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 210, 211, 230, 231, 234, 235, 257, 258, 280, 281], "bbox": [0.0, 0.134125, 0.23739062500000002, 0.7278958333333333], "iscrowd": 0, "area": 20101.53030000001, "rle": {"size": [480, 640], "counts": "g81m>4M2M4M2PBIS=;eBMX=5aB2\\=2bBM]=6cBH]=:cBD\\=`0cB^O\\=e0dBYOZ=l0eBROZ=Q1fBmNY=]10LgBbNX=c10000O100O100O10O01N2N2N2N2]MRNRHQ2m7RNoGo1Q8TNkGn1T8VNgGl1X8WNdGk1[8XNaGj1^8ZN^Gg1a8\\N[Gf1d8]NXGd1g8aNTGa1j8cNRG_1l8eNPG\\1o8hNmFY1R9kNjFV1U9nNgFS1X9QOcFQ1\\9SO`Fn0_9VO]Fj0c9Y2001O00001O01O0101O000000001O0000000000001O0000000O1000001O1O1O1O001O1O1O11O2N3L3O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O100O1O1O1O1O1O001N2O001N104L3QHmIj6W6eHWJY7k6N1O1N3N1O2M2OgJUJ]1j5aN\\J\\1b5dNbJX1_5fNgJU1Z5jNjJR1V5mNoJP1Q5oNTKl0m4ROXKj0h4UO^Kf0c4YOaKc0`4[OfKa0Y4_OmK;T4CTL6m3IYL1l3I\\L1g3JaLOd3LbLNb3MfLL_3OhLJ]30jLKY31nLHW32QMGS34TMFQ35R6JTcT7"}, "label": "a batter for the blue jays at home plate"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large pizza with slice missing and a knife laying where the slice would have been\"."}], "task": "refering", "answer_template": "The \"large pizza with slice missing and a knife laying where the slice would have been\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 383, 397, 398, 399, 400, 401, 402, 403, 404, 421, 422, 423, 424, 425], "bbox": [0.013480392156862746, 0.2584313725490196, 1.0, 0.8853921568627451], "iscrowd": 0, "area": 137151.08574999997, "rle": {"size": [612, 612], "counts": "gT54kb0=C<D=C=C<C>D;G9G8H8H9F9H5K4L4L5J5L4L5K4L5K4K5L5K3M3M3L4M3N2N2M3N2M3N2M3N2M3N2M3N2M3N2N1O2N2M3N1O2N2N2N2N1O2N2N2N1O2N2N1O1O2N1O1O1O2N1O1O1O2M2O1O1O2N1O1O1O1O2O0O1O1O2N100O1O2N1O100O1O1O2N100O1O1O1O101N1O1O1O100O1O2N1O100O1O1O101O0O100O10000O2O000O100O10001N10000O100O10001N10000O100O101O0O10000O100O10000O101O0O100O10000O10000O10000O1000000O1000001N1000000O1000000O10000O1000000O1000000O2O000O10000O0100O1O010O1O001O1O1O010O1O001O1O00100O1O001O1O0O2O1O1O001N2O001O1O001N2O1O1O001N2O1O1L4K4M4L4L4L4L4L4K4N3O1N2N2N2N1O2N2N2O1N2N2N1O2N2N2O1N2N1O2N2N2N2O1N2N1O2N2N2O1O10O01O100O1O1O01O01O2N100O2N100O2N1O101N1O1O2O0O1O2O0O1O2N100O2N1O100O2N100O1O1O100O1O1O2O0O1O100O1O1O100O2O01O1O1O0000O101O0000000O2O000000001N100000001O0O1000001O0O10001O00000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000O101O000O2O000O2O00001N10001N1000001N10001N10001O0O101O0O1010O000001O00001O00001O00010O00001O0000001O00001O01O01O00001O00001O0000001O01O0001O0000000000000000010O0000000000000000001O01O00000000000001O00000010O000001O0000001O00000010O0001O001O001O001O001O00010O001O001O001O00001O0010O01O001O001O10000O100O101N1O1O1N2O1O1O1O1O1O1N2O1O001O1O1O0O2O1O1O1O001O1N2O002N2N2N2N2M3N2N2N2ZOj_OZNX`0_1Z@TNi?c1Q1I7I7I7J7I8G:GQG"}, "label": "large pizza with slice missing and a knife laying where the slice would have been"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with one piece missing sitting on a white plate with a fork and knife\"."}], "task": "refering", "answer_template": "The \"a pizza with one piece missing sitting on a white plate with a fork and knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 383, 397, 398, 399, 400, 401, 402, 403, 404, 421, 422, 423, 424, 425], "bbox": [0.013480392156862746, 0.2584313725490196, 1.0, 0.8853921568627451], "iscrowd": 0, "area": 137151.08574999997, "rle": {"size": [612, 612], "counts": "gT54kb0=C<D=C=C<C>D;G9G8H8H9F9H5K4L4L5J5L4L5K4L5K4K5L5K3M3M3L4M3N2N2M3N2M3N2M3N2M3N2M3N2M3N2N1O2N2M3N1O2N2N2N2N1O2N2N2N1O2N2N1O1O2N1O1O1O2N1O1O1O2M2O1O1O2N1O1O1O1O2O0O1O1O2N100O1O2N1O100O1O1O2N100O1O1O1O101N1O1O1O100O1O2N1O100O1O1O101O0O100O10000O2O000O100O10001N10000O100O10001N10000O100O101O0O10000O100O10000O101O0O100O10000O10000O10000O1000000O1000001N1000000O1000000O10000O1000000O1000000O2O000O10000O0100O1O010O1O001O1O1O010O1O001O1O00100O1O001O1O0O2O1O1O001N2O001O1O001N2O1O1O001N2O1O1L4K4M4L4L4L4L4L4K4N3O1N2N2N2N1O2N2N2O1N2N2N1O2N2N2O1N2N1O2N2N2N2O1N2N1O2N2N2O1O10O01O100O1O1O01O01O2N100O2N100O2N1O101N1O1O2O0O1O2O0O1O2N100O2N1O100O2N100O1O1O100O1O1O2O0O1O100O1O1O100O2O01O1O1O0000O101O0000000O2O000000001N100000001O0O1000001O0O10001O00000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000O101O000O2O000O2O00001N10001N1000001N10001N10001O0O101O0O1010O000001O00001O00001O00010O00001O0000001O00001O01O01O00001O00001O0000001O01O0001O0000000000000000010O0000000000000000001O01O00000000000001O00000010O000001O0000001O00000010O0001O001O001O001O001O00010O001O001O001O00001O0010O01O001O001O10000O100O101N1O1O1N2O1O1O1O1O1O1N2O1O001O1O1O0O2O1O1O1O001O1N2O002N2N2N2N2M3N2N2N2ZOj_OZNX`0_1Z@TNi?c1Q1I7I7I7J7I8G:GQG"}, "label": "a pizza with one piece missing sitting on a white plate with a fork and knife"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork sitting on a pizza near us\"."}], "task": "refering", "answer_template": "The \"a fork sitting on a pizza near us\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 209, 210, 211], "bbox": [0.4720424836601307, 0.2552450980392157, 0.9776307189542482, 0.42303921568627445], "iscrowd": 0, "area": 6696.336899999999, "rle": {"size": [612, 612], "counts": "Sn\\52Sc00O2N2N2N1O2N2N2N2O0O2N2N2N1O2N2N2N2O0O0000000000000001O00000000000000001O00000000000000001O00000001O00000001O00000000000000O100O10000O010O010O010O010O0100O010O0100O010O0100O010O010O010O010O010O10O010O010O10O010O10O010O010O0100O010O10O0100O10O01O1O001O100O001O1O1O001O1O010O1O1O001O1O001O1O10O01O1O10O10000O0100000O10000O10O1000O1000000O10O10O1000000O1000O01000000O1000O10O10000O100000O010000O1000000O0100000O10000O0100000O1000O01000O10O10O10O10O1000O10O1000O01000O010000O010000O010000O01000O10O10O10O1000O10O10O10O10O1000O0100000O01000O010000O01000O0100000O01000O10O10O10O10O1000O01000O10O2O1O1N2O1O1N2O1O1N2O1O1O1N`V8"}, "label": "a fork sitting on a pizza near us"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork laying in a pizza next to a knife\"."}], "task": "refering", "answer_template": "The \"a fork laying in a pizza next to a knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 209, 210, 211], "bbox": [0.4720424836601307, 0.2552450980392157, 0.9776307189542482, 0.42303921568627445], "iscrowd": 0, "area": 6696.336899999999, "rle": {"size": [612, 612], "counts": "Sn\\52Sc00O2N2N2N1O2N2N2N2O0O2N2N2N1O2N2N2N2O0O0000000000000001O00000000000000001O00000000000000001O00000001O00000001O00000000000000O100O10000O010O010O010O010O0100O010O0100O010O0100O010O010O010O010O010O10O010O010O10O010O10O010O010O0100O010O10O0100O10O01O1O001O100O001O1O1O001O1O010O1O1O001O1O001O1O10O01O1O10O10000O0100000O10000O10O1000O1000000O10O10O1000000O1000O01000000O1000O10O10000O100000O010000O1000000O0100000O10000O0100000O1000O01000O10O10O10O10O1000O10O1000O01000O010000O010000O010000O01000O10O10O10O1000O10O10O10O10O1000O0100000O01000O010000O01000O0100000O01000O10O10O10O10O1000O01000O10O2O1O1N2O1O1N2O1O1N2O1O1O1N`V8"}, "label": "a fork laying in a pizza next to a knife"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pizza with only cheese by the knife\"."}], "task": "refering", "answer_template": "The \"a slice of pizza with only cheese by the knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 343, 344, 345, 346, 347, 348, 349, 350, 351, 365, 366, 367, 368, 369, 370, 371, 372, 387, 388, 389, 390, 391, 392, 393, 409, 410, 411, 412, 413, 414, 415, 431, 432, 433, 434, 435, 436], "bbox": [0.5842647058823529, 0.577532679738562, 0.9955065359477124, 0.9123529411764706], "iscrowd": 0, "area": 37712.09575, "rle": {"size": [612, 612], "counts": "R[f65ob06J7J5J6K5J7F9G9H9F9G9I8H7I7I7J7H7I7J7H7I7I8I6I7I7J2M1O100O1O1O010O1O100O1O00100O1O100O10O10O1000000000O10000000000O1000000000000O10O1000000000O10000000000O1000000000000O1000O100000O1000000000000O1000000000000O1000O1000O10000O10000O10000O1000000O10000O10O10O101O0O1000000O10000O101O0O10000O10000O10001O0O10000O10000O10001N1O1O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1N2O2N1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N3N1O1N2K5J6J6J6J6J7J5J6JT^1"}, "label": "a slice of pizza with only cheese by the knife"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cut piece of pizza , still on the plate , near the knife\"."}], "task": "refering", "answer_template": "The \"a cut piece of pizza , still on the plate , near the knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 343, 344, 345, 346, 347, 348, 349, 350, 351, 365, 366, 367, 368, 369, 370, 371, 372, 387, 388, 389, 390, 391, 392, 393, 409, 410, 411, 412, 413, 414, 415, 431, 432, 433, 434, 435, 436], "bbox": [0.5842647058823529, 0.577532679738562, 0.9955065359477124, 0.9123529411764706], "iscrowd": 0, "area": 37712.09575, "rle": {"size": [612, 612], "counts": "R[f65ob06J7J5J6K5J7F9G9H9F9G9I8H7I7I7J7H7I7J7H7I7I8I6I7I7J2M1O100O1O1O010O1O100O1O00100O1O100O10O10O1000000000O10000000000O1000000000000O10O1000000000O10000000000O1000000000000O1000O100000O1000000000000O1000000000000O1000O1000O10000O10000O10000O1000000O10000O10O10O101O0O1000000O10000O101O0O10000O10000O10001O0O10000O10000O10001N1O1O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1N2O2N1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N3N1O1N2K5J6J6J6J6J7J5J6JT^1"}, "label": "a cut piece of pizza , still on the plate , near the knife"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork resting on the full pizza in the back\"."}], "task": "refering", "answer_template": "The \"the fork resting on the full pizza in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 104, 105, 127, 128, 129, 150, 151], "bbox": [0.6129084967320262, 0.0388562091503268, 0.8875490196078433, 0.29151960784313724], "iscrowd": 0, "area": 3045.382899999997, "rle": {"size": [612, 612], "counts": "hTP72Sc0101O1O1EKc]O6Zb0Md]O5Xb0Nh]O3Wb0Mi]O3Vb0Nj]O1Wb00h]O0Xb00h]O0na0KR^O5O0Pb0KP^O60Ooa0LQ^O50Ooa0LQ^O5OOPb0MQ^O4OOPb0MP^O40Ofa0J]^O4M30Nea0M^^O2M20Ofa0N\\^O1N20Oea0O\\^O1N20Nga00Z^O0O10Oha00X^O1000Oha00X^O10O00ha00Y^O1OON1ka0OW^O2O9ja0DX^O3N8ka0EV^O4O6ka0GV^O3N5na0GU^Od0la0\\OS^Oc0na0]OS^Ob0na0^OQ^Oa0Pb0_OQ^O`0oa0AQ^O=Pb0Dm]O>Tb081O1O1O2N1O1O001O001O001O001O010O001O1O1O2N100O1O5K100O0100O010O10O0101O000O10001O0O10001O000000001O0000001O000000010O0000001O000000001O00000001O00O10000000000000000O1000000000O100000O100000000000000O1000000000000000000O10000000O1O1O1O1N101O1N2O1O1N2O1O1N101O1NYRY1"}, "label": "the fork resting on the full pizza in the back"}]}
{"id": 238238, "image": "COCO_train2014_000000238238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fork on the further pizza\"."}], "task": "refering", "answer_template": "The \"fork on the further pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 104, 105, 127, 128, 129, 150, 151], "bbox": [0.6129084967320262, 0.0388562091503268, 0.8875490196078433, 0.29151960784313724], "iscrowd": 0, "area": 3045.382899999997, "rle": {"size": [612, 612], "counts": "hTP72Sc0101O1O1EKc]O6Zb0Md]O5Xb0Nh]O3Wb0Mi]O3Vb0Nj]O1Wb00h]O0Xb00h]O0na0KR^O5O0Pb0KP^O60Ooa0LQ^O50Ooa0LQ^O5OOPb0MQ^O4OOPb0MP^O40Ofa0J]^O4M30Nea0M^^O2M20Ofa0N\\^O1N20Oea0O\\^O1N20Nga00Z^O0O10Oha00X^O1000Oha00X^O10O00ha00Y^O1OON1ka0OW^O2O9ja0DX^O3N8ka0EV^O4O6ka0GV^O3N5na0GU^Od0la0\\OS^Oc0na0]OS^Ob0na0^OQ^Oa0Pb0_OQ^O`0oa0AQ^O=Pb0Dm]O>Tb081O1O1O2N1O1O001O001O001O001O010O001O1O1O2N100O1O5K100O0100O010O10O0101O000O10001O0O10001O000000001O0000001O000000010O0000001O000000001O00000001O00O10000000000000000O1000000000O100000O100000000000000O1000000000000000000O10000000O1O1O1O1N101O1N2O1O1N2O1O1N101O1NYRY1"}, "label": "fork on the further pizza"}]}
{"id": 107176, "image": "COCO_train2014_000000107176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wood chair with a polka dotted cushion\"."}], "task": "refering", "answer_template": "The \"a wood chair with a polka dotted cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 181, 182, 183, 206], "bbox": [0.822015625, 0.19412177985948478, 1.0, 0.5456674473067916], "iscrowd": 0, "area": 13737.157150000003, "rle": {"size": [427, 640], "counts": "T_k6_2[9[16J6J6M210001O000O100000000O2O0000000O1000001O00000000000000001O00000000000O101O000000000000001O00000000000000001O0000000000000O2O000000000000001O00000000000000001O0000001O00001O00001O001O3M9G:F1O00O100000000O100000000O10^1bNS2mMlI"}, "label": "a wood chair with a polka dotted cushion"}]}
{"id": 107176, "image": "COCO_train2014_000000107176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty seat\"."}], "task": "refering", "answer_template": "The \"an empty seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 181, 182, 183, 206], "bbox": [0.822015625, 0.19412177985948478, 1.0, 0.5456674473067916], "iscrowd": 0, "area": 13737.157150000003, "rle": {"size": [427, 640], "counts": "T_k6_2[9[16J6J6M210001O000O100000000O2O0000000O1000001O00000000000000001O00000000000O101O000000000000001O00000000000000001O0000000000000O2O000000000000001O00000000000000001O0000001O00001O00001O001O3M9G:F1O00O100000000O100000000O10^1bNS2mMlI"}, "label": "an empty seat"}]}
{"id": 74409, "image": "COCO_train2014_000000074409.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book shelf\"."}], "task": "refering", "answer_template": "The \"a book shelf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 40, 41, 42, 43, 44, 45, 64], "bbox": [0.5788593750000001, 0.0, 0.99728125, 0.15010025062656643], "iscrowd": 0, "area": 10587.240199999997, "rle": {"size": [399, 640], "counts": "mV`43X;S1F:1O1O1O100O1O1O1O1O1O100O1O1O100O1O1O101N1O1O1O100O1O1O100O1O1O100O1O1O100O1000000000000000000000000O100000O100000000000000000000000001O000000000000001O00001O0000001O00001O0000001O1O001O1O001O1O001O1O001O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O001O00000000000000000000000000000000000000001O0000000000000000O10000000000O1000000000000O1000000000000O10000000000O100000000000000O100000000O1000000O1000000O1000000O100000000O10000000000001O000000000000000000000000000000O1000000O1000000O1000000O100O1I7I7I7H8I7IY17_N7101N2O1N2O2M2O1O1N2Ojg0"}, "label": "a book shelf"}]}
{"id": 434857, "image": "COCO_train2014_000000434857.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white designed chair on the outside dinning table\"."}], "task": "refering", "answer_template": "The \"a black and white designed chair on the outside dinning table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 273, 274, 275, 276, 277, 290, 291, 292], "bbox": [0.06591666666666667, 0.532578125, 0.3295833333333333, 0.7662968750000001], "iscrowd": 0, "area": 14224.658899999997, "rle": {"size": [640, 480], "counts": "mZd08hc09G9G9G9G9G9G9G9G9F:G8H9G9G9G8H000000O1000O10O10000O1000000O1000O10O1000000O1000000O1000O10O1000000O10000O100000O01000000O1000000O100000O01000000O10000O100000O01000000O1000000O100000O01000000O1000000O6K5K4K6K5K4L5J6K4L5K5J5L5K5K4K6K5K4L5J6K4L5K5J6K4LfiX6"}, "label": "a black and white designed chair on the outside dinning table"}]}
{"id": 434857, "image": "COCO_train2014_000000434857.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white floral pattern patio chair\"."}], "task": "refering", "answer_template": "The \"black and white floral pattern patio chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 273, 274, 275, 276, 277, 290, 291, 292], "bbox": [0.06591666666666667, 0.532578125, 0.3295833333333333, 0.7662968750000001], "iscrowd": 0, "area": 14224.658899999997, "rle": {"size": [640, 480], "counts": "mZd08hc09G9G9G9G9G9G9G9G9F:G8H9G9G9G8H000000O1000O10O10000O1000000O1000O10O1000000O1000000O1000O10O1000000O10000O100000O01000000O1000000O100000O01000000O10000O100000O01000000O1000000O100000O01000000O1000000O6K5K4K6K5K4L5J6K4L5K5J5L5K5K4K6K5K4L5J6K4L5K5J6K4LfiX6"}, "label": "black and white floral pattern patio chair"}]}
{"id": 361132, "image": "COCO_train2014_000000361132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe bending his face into the grass\"."}], "task": "refering", "answer_template": "The \"a giraffe bending his face into the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 92, 93, 94, 95, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 230, 231, 232, 241, 242, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281], "bbox": [0.2020985010706638, 0.27206, 0.6832976445396146, 0.93498], "iscrowd": 0, "area": 39106.4537, "rle": {"size": [500, 467], "counts": "SY^11`?3M2M4N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O0O2O00O1O001O001N2O001O1O001O001N2O00O1M4L3N2M3N2M4O01O1O100O1O1O10O01O1O100O1O1O10SOROTCn0`<@_C?h0UOQ:a0VE9g0\\Ol9`0[E4h0@h9a0]E0h0De9a0aEKk0F_9c0dEGo0GW9g0hEBS1HQ9j0iE_OW1Hm8k0jE]O[1Ih8l0kE[O_1Jc8m0kEZOd1J^8Q2dGoMX8R2kGnMR8R2RHmMk7S2XHmMe7T2^HkM_7U2dHkMY7U2kHkMQ7U2RIkMk6a2lHZNV6R2bIQNW6`2YIcMc6o2lHTMP7^3^HfL^7k5SIUHk5o7oIVHn5o7lIUHQ6o7jIUHT6n7gIVHX6l7bIYH\\6a8N2O1N1O100O100O1O100O100O1O100O1O01O010O000010O00010O01O01O01O010O000010O00:G0O100O1O10O01O100O100O1O100O100000000000RObJfG^5W8RK\\Gn4b8aKQG`4k8V1N2M3N2M3N2O100O100O100O100O1O2O0O1I7I7H8I7H8H8I7H8B>B?A?J7N1O1N2O1N2O2M2O1N22N4cH_K_3d4ZL^Kg3f4QL^Ko3f4iK]KX4g4aK[K`4h4YK\\Kj4e4nJ^KW5b4bJaKa5`4WJcKm5^4WIZLm6T64L4M4K<E:EY1hN6I8H7J7H8I6I8Ie0VOl0TOj0VOi0WOTlW2"}, "label": "a giraffe bending his face into the grass"}]}
{"id": 361132, "image": "COCO_train2014_000000361132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating\"."}], "task": "refering", "answer_template": "The \"a giraffe eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 92, 93, 94, 95, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 230, 231, 232, 241, 242, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281], "bbox": [0.2020985010706638, 0.27206, 0.6832976445396146, 0.93498], "iscrowd": 0, "area": 39106.4537, "rle": {"size": [500, 467], "counts": "SY^11`?3M2M4N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O0O2O00O1O001O001N2O001O1O001O001N2O00O1M4L3N2M3N2M4O01O1O100O1O1O10O01O1O100O1O1O10SOROTCn0`<@_C?h0UOQ:a0VE9g0\\Ol9`0[E4h0@h9a0]E0h0De9a0aEKk0F_9c0dEGo0GW9g0hEBS1HQ9j0iE_OW1Hm8k0jE]O[1Ih8l0kE[O_1Jc8m0kEZOd1J^8Q2dGoMX8R2kGnMR8R2RHmMk7S2XHmMe7T2^HkM_7U2dHkMY7U2kHkMQ7U2RIkMk6a2lHZNV6R2bIQNW6`2YIcMc6o2lHTMP7^3^HfL^7k5SIUHk5o7oIVHn5o7lIUHQ6o7jIUHT6n7gIVHX6l7bIYH\\6a8N2O1N1O100O100O1O100O100O1O100O1O01O010O000010O00010O01O01O01O010O000010O00:G0O100O1O10O01O100O100O1O100O100000000000RObJfG^5W8RK\\Gn4b8aKQG`4k8V1N2M3N2M3N2O100O100O100O100O1O2O0O1I7I7H8I7H8H8I7H8B>B?A?J7N1O1N2O1N2O2M2O1N22N4cH_K_3d4ZL^Kg3f4QL^Ko3f4iK]KX4g4aK[K`4h4YK\\Kj4e4nJ^KW5b4bJaKa5`4WJcKm5^4WIZLm6T64L4M4K<E:EY1hN6I8H7J7H8I6I8Ie0VOl0TOj0VOi0WOTlW2"}, "label": "a giraffe eating"}]}
{"id": 377518, "image": "COCO_train2014_000000377518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa facing a television\"."}], "task": "refering", "answer_template": "The \"a sofa facing a television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0014375, 0.6617056074766354, 0.363859375, 0.9864485981308411], "iscrowd": 0, "area": 23298.334149999995, "rle": {"size": [428, 640], "counts": "^f0b2j:a1_N0O100O100O10000O10000O1000000O10000O1001O1O001O1O001O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O[ObL^G\\3c8hLYGW3g8nLTGR3m8RMmFo2S9e0001O001O001O001O001O001O001O001O0010O01O001O00001O001O001O001O00C=1O001O001O001O001O001O001O001O001O001O001O001OO100000000000000000000000000000000O10000000000000000000000001O1O1O1O1O1O1O3M6J7I6J4L1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O00000000O1000000O10000O1000000O10000O1000000O1001O0000001O00001O1O1O2N1O1O1O1O1O1O1O2N1O3M9G8H9G9G`fY5"}, "label": "a sofa facing a television"}]}
{"id": 377518, "image": "COCO_train2014_000000377518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch is placed with other furniture in a living room\"."}], "task": "refering", "answer_template": "The \"a couch is placed with other furniture in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285], "bbox": [0.18487499999999998, 0.5502102803738318, 0.586421875, 0.8569158878504672], "iscrowd": 0, "area": 22610.440749999998, "rle": {"size": [428, 640], "counts": "hba1m0_<001O000O10001O000O10001O000O10001O000O2O1O0O2K5K5K8H7J6I8K4M3M3M4L1O2N1O2N1O2N1O2N1O2M101O1O1O1O1O001O1O1O1O1O1O001O1O1O1OO1000000O10000O1000000O1000000O10000O1000000O10000O10O1000O10000O1000000O10000O1000000O1000000O100000000O100000000O100000000O100000000O10001O000O100000000O100000000O100000000O100000000K4J7O100O100O10O0100O100000000O010000000O1000O10O100000000O10O1000O100000000O010000000O1000O10O10000000b0]O0100O1O0101O00000O2O0000000O2O000000001O0O1000001O00000O2O000000001O0O1000001O00000O101O0000001O`0_Od0]OU_^3"}, "label": "a couch is placed with other furniture in a living room"}]}
{"id": 377518, "image": "COCO_train2014_000000377518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown chair or love seat facing the tv and shelves\"."}], "task": "refering", "answer_template": "The \"brown chair or love seat facing the tv and shelves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.000359375, 0.6559112149532711, 0.369453125, 1.0], "iscrowd": 0, "area": 24371.40705, "rle": {"size": [428, 640], "counts": "j<`0X;d1\\Nd1@`0O1O100O100O1O100O100O1O100O100001O1O001O001O1O001O1O001O001O1O001O001O1O001O2N1O1O1O1O2N1O1O1O1O1O2N@`0001O001O001O001O00001O001O001O001O001O0010O01O001O000000O1M3L4M3L41O001O1O001O001O1O001O1O001O001O1O001O001O1O0000O10000O100O10000O10000O11O001O001O001O001O001O001O1O1O1O1O2N1O1O1O2N7I2N1O1O2N1O1O2N1O1O2N1O2N2N2N2N3M2N2N2N001O001O001O001O001O001O001O001O0010O01O001O001O00O10000O1000000O10000O1000000O10000O11O1O1O001O1O1O001O1O1O001O1O1O001O1O1O006J7I6J6J7I6J6JZ^X5"}, "label": "brown chair or love seat facing the tv and shelves"}]}
{"id": 377518, "image": "COCO_train2014_000000377518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beige love seat sitting across a television set\"."}], "task": "refering", "answer_template": "The \"a beige love seat sitting across a television set\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.000359375, 0.6559112149532711, 0.369453125, 1.0], "iscrowd": 0, "area": 24371.40705, "rle": {"size": [428, 640], "counts": "j<`0X;d1\\Nd1@`0O1O100O100O1O100O100O1O100O100001O1O001O001O1O001O1O001O001O1O001O001O1O001O2N1O1O1O1O2N1O1O1O1O1O2N@`0001O001O001O001O00001O001O001O001O001O0010O01O001O000000O1M3L4M3L41O001O1O001O001O1O001O1O001O001O1O001O001O1O0000O10000O100O10000O10000O11O001O001O001O001O001O001O1O1O1O1O2N1O1O1O2N7I2N1O1O2N1O1O2N1O1O2N1O2N2N2N2N3M2N2N2N001O001O001O001O001O001O001O001O0010O01O001O001O00O10000O1000000O10000O1000000O10000O11O1O1O001O1O1O001O1O1O001O1O1O001O1O1O006J7I6J6J7I6J6JZ^X5"}, "label": "a beige love seat sitting across a television set"}]}
{"id": 557746, "image": "COCO_train2014_000000557746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pitza in front of a man\"."}], "task": "refering", "answer_template": "The \"pitza in front of a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 352, 353, 354, 355], "bbox": [0.2545, 0.6752083333333334, 0.57978125, 0.9111666666666667], "iscrowd": 0, "area": 16105.18475, "rle": {"size": [480, 640], "counts": "Yh\\21n>2N2N101N2N1O2N2N1O2N2O0O2N2N1O2N2N101O1N2O0O2O\\1dN1N100O2O000O101N10001N100O101O0O100O2O000O2O0O10001N100O2O000O101N10001N100O101O00001O0000001O0000001O000000000O10O100000000O100000000O1000000O10000000O0100000000O100000000O100000000O100000O10O100000000O2O00000O2O0000001N100000001N1000001O0O10001O000O101O00000O2O0000001N1000001N101O0N3K4M4K4L5L3L5K4M4K4L5N101O0O2O001O0O2O001O001N101O001O0O2O001O2M4M2N2N2N3L3N2N2N]gm3"}, "label": "pitza in front of a man"}]}
{"id": 557746, "image": "COCO_train2014_000000557746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in front of the man\"."}], "task": "refering", "answer_template": "The \"the pizza in front of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 352, 353, 354, 355], "bbox": [0.2545, 0.6752083333333334, 0.57978125, 0.9111666666666667], "iscrowd": 0, "area": 16105.18475, "rle": {"size": [480, 640], "counts": "Yh\\21n>2N2N101N2N1O2N2N1O2N2O0O2N2N1O2N2N101O1N2O0O2O\\1dN1N100O2O000O101N10001N100O101O0O100O2O000O2O0O10001N100O2O000O101N10001N100O101O00001O0000001O0000001O000000000O10O100000000O100000000O1000000O10000000O0100000000O100000000O100000000O100000O10O100000000O2O00000O2O0000001N100000001N1000001O0O10001O000O101O00000O2O0000001N1000001N101O0N3K4M4K4L5L3L5K4M4K4L5N101O0O2O001O0O2O001O001N101O001O0O2O001O2M4M2N2N2N3L3N2N2N]gm3"}, "label": "the pizza in front of the man"}]}
{"id": 557746, "image": "COCO_train2014_000000557746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza in front of a woman wearing a brown and blue shirt\"."}], "task": "refering", "answer_template": "The \"a pizza in front of a woman wearing a brown and blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341], "bbox": [0.5780937500000001, 0.6468125, 0.9016875000000001, 0.8580625000000001], "iscrowd": 0, "area": 15746.537400000001, "rle": {"size": [480, 640], "counts": "Ti]54c>:F:I6O2O1O1N101N2O001N2O1N101O1N101N2O1O0O2O1N101O1N2O0O2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O001N101O001O0O2O001O001O0O2O001O001N101O000000000O1000000000000000000O101O000000000000000O1000000000000000000O11O00010O00001O00001O00001O00001O00001O00001O00001N10001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0O101N100O2O0O101N100O2O0O101N100O2O0O101N100O2O0O1O2O1N2O2M2K5L4L4K5L4L5J5L4L4L4KPVm0"}, "label": "a pizza in front of a woman wearing a brown and blue shirt"}]}
{"id": 557746, "image": "COCO_train2014_000000557746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza in front of the woman on the table\"."}], "task": "refering", "answer_template": "The \"a pizza in front of the woman on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341], "bbox": [0.5780937500000001, 0.6468125, 0.9016875000000001, 0.8580625000000001], "iscrowd": 0, "area": 15746.537400000001, "rle": {"size": [480, 640], "counts": "Ti]54c>:F:I6O2O1O1N101N2O001N2O1N101O1N101N2O1O0O2O1N101O1N2O0O2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O001N101O001O0O2O001O001O0O2O001O001N101O000000000O1000000000000000000O101O000000000000000O1000000000000000000O11O00010O00001O00001O00001O00001O00001O00001O00001N10001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0O101N100O2O0O101N100O2O0O101N100O2O0O101N100O2O0O1O2O1N2O2M2K5L4L4K5L4L5J5L4L4L4KPVm0"}, "label": "a pizza in front of the woman on the table"}]}
{"id": 557746, "image": "COCO_train2014_000000557746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table under green cloth\"."}], "task": "refering", "answer_template": "The \"table under green cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 345, 346, 347, 368, 369, 370], "bbox": [0.003359375, 0.6815, 0.359859375, 0.9864375], "iscrowd": 0, "area": 15997.550199999998, "rle": {"size": [480, 640], "counts": "c\\15Y>b0^Ob0]Oc0^Ob0^Ob0^Ob0N201O00000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000OcNoDmNQ;S1XEdNh:[1aE]N_:b1kEUNU:k1SFmMm9R2]FeMc9[2a1O010O100O100O10O01O100O100O100O010O100O100O1O10O0100O100O100O010O1O100O100O10O0100O100O100O1O010O100O100O10O0100O100O1O100O010O100O100O100O010O1O100O100O010O100O100O1O10O0100O100O100O010O100O1O100O10O0100O100O100O10O01O100O100O10O0100O100O1O100O010O100O100O100O010O1O100O100O010O100O100N2L4L3N3Lmko5"}, "label": "table under green cloth"}]}
{"id": 557746, "image": "COCO_train2014_000000557746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the triangle of exposed table\"."}], "task": "refering", "answer_template": "The \"the triangle of exposed table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 345, 346, 347, 368, 369, 370], "bbox": [0.003359375, 0.6815, 0.359859375, 0.9864375], "iscrowd": 0, "area": 15997.550199999998, "rle": {"size": [480, 640], "counts": "c\\15Y>b0^Ob0]Oc0^Ob0^Ob0^Ob0N201O00000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000OcNoDmNQ;S1XEdNh:[1aE]N_:b1kEUNU:k1SFmMm9R2]FeMc9[2a1O010O100O100O10O01O100O100O100O010O100O100O1O10O0100O100O100O010O1O100O100O10O0100O100O100O1O010O100O100O10O0100O100O1O100O010O100O100O100O010O1O100O100O010O100O100O1O10O0100O100O100O010O100O1O100O10O0100O100O100O10O01O100O100O10O0100O100O1O100O010O100O100O100O010O1O100O100O010O100O100N2L4L3N3Lmko5"}, "label": "the triangle of exposed table"}]}
{"id": 565938, "image": "COCO_train2014_000000565938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two men touching the horses\"."}], "task": "refering", "answer_template": "The \"two men touching the horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 204, 225, 226, 227, 248, 249, 250, 271, 272, 273, 295, 296, 297, 318, 319, 320, 341, 342, 343, 364], "bbox": [0.779859375, 0.48314814814814816, 0.930015625, 0.8943827160493828], "iscrowd": 0, "area": 11065.8848, "rle": {"size": [486, 640], "counts": "ma]71U?2O1O1N2O1N2O1O1N2O1N2O1O1N2N2K5J6J6K5J6K6J?D5J6K5J3M3N2M3M3N8G;F;D;E;F:E=Cg0ZOO1000000000000O1000000000000O2O0000000000000O10000O2O1O1O1N2O1O1N2O1O1VLaFk1`9PNgFl1[9oMlFn1V9kMSGo1Q9kMVGP2o8iMXGR2l8gM\\GT2h8fM_GU2e8eMbGV2c8bMeGY2h:K5K6J7I7I6J7I8H7I7I7I^od0"}, "label": "two men touching the horses"}]}
{"id": 565938, "image": "COCO_train2014_000000565938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 people in front of horses\"."}], "task": "refering", "answer_template": "The \"2 people in front of horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 204, 225, 226, 227, 248, 249, 250, 271, 272, 273, 295, 296, 297, 318, 319, 320, 341, 342, 343, 364], "bbox": [0.779859375, 0.48314814814814816, 0.930015625, 0.8943827160493828], "iscrowd": 0, "area": 11065.8848, "rle": {"size": [486, 640], "counts": "ma]71U?2O1O1N2O1N2O1O1N2O1N2O1O1N2N2K5J6J6K5J6K6J?D5J6K5J3M3N2M3M3N8G;F;D;E;F:E=Cg0ZOO1000000000000O1000000000000O2O0000000000000O10000O2O1O1O1N2O1O1N2O1O1VLaFk1`9PNgFl1[9oMlFn1V9kMSGo1Q9kMVGP2o8iMXGR2l8gM\\GT2h8fM_GU2e8eMbGV2c8bMeGY2h:K5K6J7I7I6J7I8H7I7I7I^od0"}, "label": "2 people in front of horses"}]}
{"id": 492219, "image": "COCO_train2014_000000492219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in red shirt\"."}], "task": "refering", "answer_template": "The \"man in red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 136, 137, 138, 139, 151, 152, 153, 154, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 255, 256, 258, 259, 270, 271, 274, 275, 285, 286, 289, 290, 291, 300, 301, 315, 330], "bbox": [0.0067289719626168224, 0.20224999999999999, 0.6350934579439252, 0.9977499999999999], "iscrowd": 0, "area": 41642.496349999994, "rle": {"size": [640, 428], "counts": "o[2P4m?3L4M3M3L4MoNi@_MT?c2QAYMk>j2ZARMc>P3bAhL_>Z3gA^LX>e3o02N2M3N2WOh0aM`2K5L4L4K5L4L4K5L4L4K5L4L4K5L4QOhFSG\\9k8eFQG`9m8bFoFa9P9aFlFa9T9`FhFc9X9_FdFb9\\9aF`Fa9`9>000O10000O3N3M3L3I8I6I8H7J7H8I6L5K4M4K4L5L3L5L4K4L5L3L5L3L5K5L3L50O3N2M3N2M3N2\\NkIYEW6d:e1M3N2M3M3N1N3M3N2M3M3N2M3M3M3N3L5K4M4K5K5LSLaF\\M[9d2iF]MT9`2PGbMl8]2XGcMe8[2`GfM\\8Y2hGhMT8W2oGjMo7U2THlMh7U2ZHmMa7T2aHPNX7Q2iHUNo6l1SIXNg6h1[I]N]6d1eI`NT6a1mIaNo5`1SJ]Nm5d1UJYNl5LiHSM^1m2i5MnHUM]1i2e50SIUM[1g2b51XIWMZ1c2^53^IXMX1`2[54bI[MV1]2X55hIZMV1\\2R58lIYMX1[2k49SJXMW1[2f4:XJWMX1[2`4;]JWMW1[2]4:bJWMU1\\2Y4:gJWMT1\\2U4;lJTMT1^2P4;QKTMS1^2l3;WKSMR1^2h3;[KTMQ1^2d3;aKSMQ1]2^3=N\\O2d06TOJl0=nNCP1f0hN[OW1l0cNTO\\1P1`NQO^1R1aNnN^1S1bNmN4fHG_83mNBaHE9a0\\87jNBdHA7g0]85hNAiH]O4m0]83gNAZI;R83dN@[I=R82dN_OZI`0T80cN]OZIc0U8OaN\\O[Id0W8N_N\\O[If0X8M]N[O[Ii0Z8J\\N[O[Ij0\\8JYNZO\\Il0\\8JYNXO[In0^8HYO8j0GUO9k0HTO8m0GSO8o0HoN9R1FnN:S1FlN9V1FjN:W1DiN=X1AiN>Y1AgN?Z1_OgNa0Z1]OgNc0Z1\\OeNd0]1ZOdNf0]1XOdNh0\\1XOdNg0^1WObNj0_1TObNl0_1SOaNl0Z:0000O10000O100O100O10000O010O10O010O0100O010O10O0100O01000O010O10O010O0100O010O010O10O01000O010O10001N100O2O001N10001N101O000O2O00001N101O000O2O001OO2N100O2O0O2N100O2O0O2N101N1L4H9FkZQ3"}, "label": "man in red shirt"}]}
{"id": 492219, "image": "COCO_train2014_000000492219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a red shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 136, 137, 138, 139, 151, 152, 153, 154, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 255, 256, 258, 259, 270, 271, 274, 275, 285, 286, 289, 290, 291, 300, 301, 315, 330], "bbox": [0.0067289719626168224, 0.20224999999999999, 0.6350934579439252, 0.9977499999999999], "iscrowd": 0, "area": 41642.496349999994, "rle": {"size": [640, 428], "counts": "o[2P4m?3L4M3M3L4MoNi@_MT?c2QAYMk>j2ZARMc>P3bAhL_>Z3gA^LX>e3o02N2M3N2WOh0aM`2K5L4L4K5L4L4K5L4L4K5L4L4K5L4QOhFSG\\9k8eFQG`9m8bFoFa9P9aFlFa9T9`FhFc9X9_FdFb9\\9aF`Fa9`9>000O10000O3N3M3L3I8I6I8H7J7H8I6L5K4M4K4L5L3L5L4K4L5L3L5L3L5K5L3L50O3N2M3N2M3N2\\NkIYEW6d:e1M3N2M3M3N1N3M3N2M3M3N2M3M3M3N3L5K4M4K5K5LSLaF\\M[9d2iF]MT9`2PGbMl8]2XGcMe8[2`GfM\\8Y2hGhMT8W2oGjMo7U2THlMh7U2ZHmMa7T2aHPNX7Q2iHUNo6l1SIXNg6h1[I]N]6d1eI`NT6a1mIaNo5`1SJ]Nm5d1UJYNl5LiHSM^1m2i5MnHUM]1i2e50SIUM[1g2b51XIWMZ1c2^53^IXMX1`2[54bI[MV1]2X55hIZMV1\\2R58lIYMX1[2k49SJXMW1[2f4:XJWMX1[2`4;]JWMW1[2]4:bJWMU1\\2Y4:gJWMT1\\2U4;lJTMT1^2P4;QKTMS1^2l3;WKSMR1^2h3;[KTMQ1^2d3;aKSMQ1]2^3=N\\O2d06TOJl0=nNCP1f0hN[OW1l0cNTO\\1P1`NQO^1R1aNnN^1S1bNmN4fHG_83mNBaHE9a0\\87jNBdHA7g0]85hNAiH]O4m0]83gNAZI;R83dN@[I=R82dN_OZI`0T80cN]OZIc0U8OaN\\O[Id0W8N_N\\O[If0X8M]N[O[Ii0Z8J\\N[O[Ij0\\8JYNZO\\Il0\\8JYNXO[In0^8HYO8j0GUO9k0HTO8m0GSO8o0HoN9R1FnN:S1FlN9V1FjN:W1DiN=X1AiN>Y1AgN?Z1_OgNa0Z1]OgNc0Z1\\OeNd0]1ZOdNf0]1XOdNh0\\1XOdNg0^1WObNj0_1TObNl0_1SOaNl0Z:0000O10000O100O100O10000O010O10O010O0100O010O10O0100O01000O010O10O010O0100O010O010O10O01000O010O10001N100O2O001N10001N101O000O2O00001N101O000O2O001OO2N100O2O0O2N100O2O0O2N101N1L4H9FkZQ3"}, "label": "a man wearing a red shirt"}]}
{"id": 492219, "image": "COCO_train2014_000000492219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in jeans and a blue shirt playing wii\"."}], "task": "refering", "answer_template": "The \"a man in jeans and a blue shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 99, 100, 101, 114, 115, 116, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 202, 203, 204, 205, 218, 219, 220, 233, 234, 235, 248, 249, 250, 263, 264, 265, 277, 278, 279, 280, 281, 292, 293, 294, 295, 296, 307, 308, 309, 323, 324], "bbox": [0.508177570093458, 0.24167187499999998, 0.9470560747663552, 0.9338749999999999], "iscrowd": 0, "area": 36154.94929999999, "rle": {"size": [640, 428], "counts": "lhX4?b=CYHa0\\77RHNc7`0UHEb7f0XH]Ob7m0VHXOb7R1WHTOa7V1XHoNa7Y1SHPOg7Z1lGoNo7Y1dGQOW8W1]GRO^8V1VGSOd8U1RGSOi8U1mFROn8g1XF_Nc9P2mEVNm9R6K5K5K5L4L4L4M3L4M3L301N2N2O1N2N2O0O1O2N10mMfG`HY8^7nG^HQ8`7VH\\Hj7a7\\H\\Hc7b7dHXGUOa0U8T8mHUGROf0Q8R8bIlG]6P8iIoGV6m7QJQHo5j7WJUHi5f7^JYH`5c7fJ[H[5`7lJ]HU5^7QK`Ho4^5eHTKc2\\Oh4e5RMYJn2l5PMQJQ3S6lLkIn12lIW6T4dIn1=hIU6V4]Io1g0eIP6U;WJgDm5a1_IU7j0VGm5i0RJl77XGS8d8QHXGT8d8oGYGX8_8g1M3M3M3M2N001000000000000000hIVH:j7EZH8e7I]H5c7J`H4`7LbH2^7NfHNZ72nHFR7:UI^Ol6b0\\IVOd6j0dIoN[6P1mIiNS6W1UJaNk5_1]JYNc5g1dJSN[5m1mJkMS5U2TKdMl4[2ZKfJQMV2f7S3VLdLj3\\3`L[L_3e3WMeKi2[4eMXK[2g4fMXKZ2g4iMXKV2h4kMWKV2h4kMXKT2h4mMWKS2i4oMVKP2j4QNUKP2j4QNVKn1i4TNVKl1j4TNVKm1i4TNWKk1i4VNVKj1j4VNWKj1h4WNWKj1X1RHd1U6SMj1T1XHg1o5TMj1HUHi0:X2h5VMi1E\\Hi07Y2d5XM^2<oG[2d5XM_2:oG]2c5XM`21WGJg0l2b5XMa20]GFb0Q3a5XMa2OYHh2W5XM`2O\\Hg2T5YMa2N^Hh2R5YMR4f2oKVMU4i2kKUMX4j2`6O<D001O001O0SN_^O[1Ub0N2N2N2M3N2M010O0O2N2N101N1O2N10O010000000O100000JU]OXOjb0i0W]OVOib0j0X]OUOgb0l0[]OROeb0n051O0100^OS]O2mb0MU]O2kb0MW]O2ib0MX]OFJ;nb0NX]OIJ8mb0OZ]OJI7lb0M^]O8ab0G`]O:_b0Ec]O;\\b0De]O=Zb0Bg]O?mb00O010O1K4N2M23L4_Oi\\O2]c0Jj\\ON[g>"}, "label": "a man in jeans and a blue shirt playing wii"}]}
{"id": 492219, "image": "COCO_train2014_000000492219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue shirt and glasses holding a wii controller\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue shirt and glasses holding a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 99, 100, 101, 114, 115, 116, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 202, 203, 204, 205, 218, 219, 220, 233, 234, 235, 248, 249, 250, 263, 264, 265, 277, 278, 279, 280, 281, 292, 293, 294, 295, 296, 307, 308, 309, 323, 324], "bbox": [0.508177570093458, 0.24167187499999998, 0.9470560747663552, 0.9338749999999999], "iscrowd": 0, "area": 36154.94929999999, "rle": {"size": [640, 428], "counts": "lhX4?b=CYHa0\\77RHNc7`0UHEb7f0XH]Ob7m0VHXOb7R1WHTOa7V1XHoNa7Y1SHPOg7Z1lGoNo7Y1dGQOW8W1]GRO^8V1VGSOd8U1RGSOi8U1mFROn8g1XF_Nc9P2mEVNm9R6K5K5K5L4L4L4M3L4M3L301N2N2O1N2N2O0O1O2N10mMfG`HY8^7nG^HQ8`7VH\\Hj7a7\\H\\Hc7b7dHXGUOa0U8T8mHUGROf0Q8R8bIlG]6P8iIoGV6m7QJQHo5j7WJUHi5f7^JYH`5c7fJ[H[5`7lJ]HU5^7QK`Ho4^5eHTKc2\\Oh4e5RMYJn2l5PMQJQ3S6lLkIn12lIW6T4dIn1=hIU6V4]Io1g0eIP6U;WJgDm5a1_IU7j0VGm5i0RJl77XGS8d8QHXGT8d8oGYGX8_8g1M3M3M3M2N001000000000000000hIVH:j7EZH8e7I]H5c7J`H4`7LbH2^7NfHNZ72nHFR7:UI^Ol6b0\\IVOd6j0dIoN[6P1mIiNS6W1UJaNk5_1]JYNc5g1dJSN[5m1mJkMS5U2TKdMl4[2ZKfJQMV2f7S3VLdLj3\\3`L[L_3e3WMeKi2[4eMXK[2g4fMXKZ2g4iMXKV2h4kMWKV2h4kMXKT2h4mMWKS2i4oMVKP2j4QNUKP2j4QNVKn1i4TNVKl1j4TNVKm1i4TNWKk1i4VNVKj1j4VNWKj1h4WNWKj1X1RHd1U6SMj1T1XHg1o5TMj1HUHi0:X2h5VMi1E\\Hi07Y2d5XM^2<oG[2d5XM_2:oG]2c5XM`21WGJg0l2b5XMa20]GFb0Q3a5XMa2OYHh2W5XM`2O\\Hg2T5YMa2N^Hh2R5YMR4f2oKVMU4i2kKUMX4j2`6O<D001O001O0SN_^O[1Ub0N2N2N2M3N2M010O0O2N2N101N1O2N10O010000000O100000JU]OXOjb0i0W]OVOib0j0X]OUOgb0l0[]OROeb0n051O0100^OS]O2mb0MU]O2kb0MW]O2ib0MX]OFJ;nb0NX]OIJ8mb0OZ]OJI7lb0M^]O8ab0G`]O:_b0Ec]O;\\b0De]O=Zb0Bg]O?mb00O010O1K4N2M23L4_Oi\\O2]c0Jj\\ON[g>"}, "label": "a man wearing a blue shirt and glasses holding a wii controller"}]}
{"id": 524991, "image": "COCO_train2014_000000524991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that is running near the rock\"."}], "task": "refering", "answer_template": "The \"a zebra that is running near the rock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 292, 293, 294, 315, 316, 337, 338, 360, 361], "bbox": [0.67753125, 0.2900596421471173, 0.91115625, 0.8581510934393638], "iscrowd": 0, "area": 17579.668749999997, "rle": {"size": [503, 640], "counts": "Vce61d?3N2N2N1O2N2N2L4M2M3M3N2M3M3N2M3M3N2M2N100O1O100000000O01000[MbNSG]1i7jNaFM`1Z1P8lN]FN]1R1]8QOTF2X1i0j8WOlE4R1f0U9XOgE5j0i0`9UObE6e0j0j9RO_E7=m0U:nN\\E86o0_:lNXE8OR1j:hNUEb2l:`MRE_2o:R12OnNZLVFe3j9aLQF^3P:S1A_JkF`5T9dJjF[5V9hJhFW5W9mJgFR5X9QKfFo4[9SKcFl4`9TK^Fk4e9UKYFj4i9`03M3M2010O2O001N101O0M4L3M4J5K8J8J00000000gNWGdKj8Z4ZGcKf8[4^GcKb8[4cGbK]8]4_1N2O1N2O1N2O2N102N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N2N1O1N101O1O001O001O1O001O001O001O001O001O1O001O001O001M2N3M2hMfCQ1]<[NYD`1i;ZN`D`1Q=J6J6J7H:G9G`gk0"}, "label": "a zebra that is running near the rock"}]}
{"id": 524991, "image": "COCO_train2014_000000524991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra not eating grass\"."}], "task": "refering", "answer_template": "The \"a zebra not eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 292, 293, 294, 315, 316, 337, 338, 360, 361], "bbox": [0.67753125, 0.2900596421471173, 0.91115625, 0.8581510934393638], "iscrowd": 0, "area": 17579.668749999997, "rle": {"size": [503, 640], "counts": "Vce61d?3N2N2N1O2N2N2L4M2M3M3N2M3M3N2M3M3N2M2N100O1O100000000O01000[MbNSG]1i7jNaFM`1Z1P8lN]FN]1R1]8QOTF2X1i0j8WOlE4R1f0U9XOgE5j0i0`9UObE6e0j0j9RO_E7=m0U:nN\\E86o0_:lNXE8OR1j:hNUEb2l:`MRE_2o:R12OnNZLVFe3j9aLQF^3P:S1A_JkF`5T9dJjF[5V9hJhFW5W9mJgFR5X9QKfFo4[9SKcFl4`9TK^Fk4e9UKYFj4i9`03M3M2010O2O001N101O0M4L3M4J5K8J8J00000000gNWGdKj8Z4ZGcKf8[4^GcKb8[4cGbK]8]4_1N2O1N2O1N2O2N102N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N2N1O1N101O1O001O001O1O001O001O001O001O001O1O001O001O001M2N3M2hMfCQ1]<[NYD`1i;ZN`D`1Q=J6J6J7H:G9G`gk0"}, "label": "a zebra not eating grass"}]}
{"id": 524991, "image": "COCO_train2014_000000524991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing behind two others picking at the ground\"."}], "task": "refering", "answer_template": "The \"a zebra standing behind two others picking at the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 108, 109, 110, 111, 130, 131, 132, 133, 134, 152, 153, 154, 155, 157, 175, 176, 177, 198, 199, 200], "bbox": [0.64390625, 0.20693836978131214, 0.8732031250000001, 0.49343936381709746], "iscrowd": 0, "area": 8298.232299999998, "rle": {"size": [503, 640], "counts": "]cZ67l>d0\\Oe0[Oe0L3O2N2N2O0O2N2N1O2ON1O100O1O1O2O0O1O1O100O2N1O1O100O1O2N1A?00O10001N1000PClN^;S1UD\\Oj;d0gCLX<j101N1O1O0001N10000O100O2O000N2@`0A?H6XNWBg1j=21O1O00100O001O1O1O001O10O^Oc01O001N2O1O001N2O000O01O1O1O001O10O01O1O1O001O1O10O101O0O2O00001N101O0O2O001O0O2O00000O10O10O0100O2O1O1NdBjN[<T1bCoN_<o0`CSOa<l0\\CWOe<i0WCZOj<f0RC]Oo<c0mB@T=`0hBCY==dBE]=;_BHb=V11M3J7H7J6I7J6I7I7J6I\\eW1"}, "label": "a zebra standing behind two others picking at the ground"}]}
{"id": 524991, "image": "COCO_train2014_000000524991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with its head down eating grass to the left of two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra with its head down eating grass to the left of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 328, 329], "bbox": [0.2321875, 0.30968190854870775, 0.470125, 0.8359840954274353], "iscrowd": 0, "area": 22098.815700000014, "rle": {"size": [503, 640], "counts": "b_Y25^?6K5J6N2O1O001O1jAYOX=h0dB\\O\\=e0`B^O`=b0]BBb=>[BEe=;WBIi=7TB9^=T1Mg0YO2N3M2M3N1O6JO1OcM]MfGc2T8iMgGV2T8TNhGk1W8k2O1O1O1N2O1O1O1O0O101O00000O10000O10ROVGPKj8j4cGoJ\\8W5fGbJZ8c5hGUJY8P6c05L2M3N2N2N1N2O1O1O1O1OO1M3`N_1O2M3M3N2M3N2O1O1N2O1O1N3N1O2M3N1N3J6H710O010O1O010O1O0100UFgLX7[3eHfLZ7]3cHdL]7^3`HcL`7_3\\HcLd7_3YHbLg7b3RHaLn7c3kG`LU8c3eG_L\\8e3\\G^Le8e3UG^Lk8f3mF]LT9f3eF]L\\9n41O100O1O1L4L4L5K4L4hMlEiNX:W1T2O1O100O2O0O1O1R1nN3M3M3M4M;D3MZ1fNO1O100O2N2N3M3N2O1L4kMlEbNW:W1RFdNQ:V1WFeNl9T1]FgNe9S1dFgN`9S1hFhN[9Q1cGnMg8k1g2CV[V5"}, "label": "a zebra with its head down eating grass to the left of two other zebras"}]}
{"id": 524991, "image": "COCO_train2014_000000524991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left\"."}], "task": "refering", "answer_template": "The \"the zebra on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 328, 329], "bbox": [0.2321875, 0.30968190854870775, 0.470125, 0.8359840954274353], "iscrowd": 0, "area": 22098.815700000014, "rle": {"size": [503, 640], "counts": "b_Y25^?6K5J6N2O1O001O1jAYOX=h0dB\\O\\=e0`B^O`=b0]BBb=>[BEe=;WBIi=7TB9^=T1Mg0YO2N3M2M3N1O6JO1OcM]MfGc2T8iMgGV2T8TNhGk1W8k2O1O1O1N2O1O1O1O0O101O00000O10000O10ROVGPKj8j4cGoJ\\8W5fGbJZ8c5hGUJY8P6c05L2M3N2N2N1N2O1O1O1O1OO1M3`N_1O2M3M3N2M3N2O1O1N2O1O1N3N1O2M3N1N3J6H710O010O1O010O1O0100UFgLX7[3eHfLZ7]3cHdL]7^3`HcL`7_3\\HcLd7_3YHbLg7b3RHaLn7c3kG`LU8c3eG_L\\8e3\\G^Le8e3UG^Lk8f3mF]LT9f3eF]L\\9n41O100O1O1L4L4L5K4L4hMlEiNX:W1T2O1O100O2O0O1O1R1nN3M3M3M4M;D3MZ1fNO1O100O2N2N3M3N2O1L4kMlEbNW:W1RFdNQ:V1WFeNl9T1]FgNe9S1dFgN`9S1hFhN[9Q1cGnMg8k1g2CV[V5"}, "label": "the zebra on the left"}]}
{"id": 221889, "image": "COCO_train2014_000000221889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant with its back to the camera on the left '\"."}], "task": "refering", "answer_template": "The \"the elephant with its back to the camera on the left '\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 235, 236], "bbox": [0.17840625000000002, 0.22410798122065728, 0.347796875, 0.7060798122065727], "iscrowd": 0, "area": 12964.943450000004, "rle": {"size": [426, 640], "counts": "nd_11Z;0cF1Z94cFMZ97dFKY98eFIX9<eFEX9?fFCW9`0gFAX9b0dF@\\9a0bF@]9b0aF@^9b0_F_Oa9b0^F^Ob9c0]F^Ob9e0[F[Od9h0ZFXOf9j0WFXOh9j0VFVOj9l0QFXOn9k0lEXOS:k0hEXOX:j0dEYO[:k101YFZMh8f2jFoMn8^3I3M1O0O2O0O2O000O2J5J7H7I7O2N1O2N2N101N1O2N1O2N100O1O1O1O100O1O1O100O1000000O10000O01N2N2N1O2M3N1N3M3N2M2N3N2M2N3TMWH6j7@dH:^7[OPIb0o6ZOYIb0h6ZO`Ib0`6ZOgIb0Z6ZOnIb0Q6[OVJa0k5ZO^Ja0c5[OdJb0\\5[OkJ`0U5@oJ=Q5BRK;P5CRK;o4DTK7o4HTK4n4KUK0n4OUKLn43UKIm46VKFl48XKFh49U4O1O3M3Mai]5"}, "label": "the elephant with its back to the camera on the left '"}]}
{"id": 221889, "image": "COCO_train2014_000000221889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant nearly fully behind the lead elephant\"."}], "task": "refering", "answer_template": "The \"an elephant nearly fully behind the lead elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 235, 236], "bbox": [0.17840625000000002, 0.22410798122065728, 0.347796875, 0.7060798122065727], "iscrowd": 0, "area": 12964.943450000004, "rle": {"size": [426, 640], "counts": "nd_11Z;0cF1Z94cFMZ97dFKY98eFIX9<eFEX9?fFCW9`0gFAX9b0dF@\\9a0bF@]9b0aF@^9b0_F_Oa9b0^F^Ob9c0]F^Ob9e0[F[Od9h0ZFXOf9j0WFXOh9j0VFVOj9l0QFXOn9k0lEXOS:k0hEXOX:j0dEYO[:k101YFZMh8f2jFoMn8^3I3M1O0O2O0O2O000O2J5J7H7I7O2N1O2N2N101N1O2N1O2N100O1O1O1O100O1O1O100O1000000O10000O01N2N2N1O2M3N1N3M3N2M2N3N2M2N3TMWH6j7@dH:^7[OPIb0o6ZOYIb0h6ZO`Ib0`6ZOgIb0Z6ZOnIb0Q6[OVJa0k5ZO^Ja0c5[OdJb0\\5[OkJ`0U5@oJ=Q5BRK;P5CRK;o4DTK7o4HTK4n4KUK0n4OUKLn43UKIm46VKFl48XKFh49U4O1O3M3Mai]5"}, "label": "an elephant nearly fully behind the lead elephant"}]}
{"id": 221889, "image": "COCO_train2014_000000221889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant walking to the side with only one eye showing\"."}], "task": "refering", "answer_template": "The \"an elephant walking to the side with only one eye showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 201, 202, 204, 205, 221, 222, 223, 224, 227], "bbox": [0.610484375, 0.1997887323943662, 0.9624843750000001, 0.6610093896713615], "iscrowd": 0, "area": 30549.262599999995, "rle": {"size": [426, 640], "counts": "\\jR5c1e;m0TO2N001O00eMkEc1T:aN\\Fk0d9XOaFa0^9CfF5\\O[OU9c0cGLSOFX9c0iG_OjN5[9?YHGf7=QHIn7:jGLU8T1fFROY9`2O2[NYLdIh3\\6ZL_Ii3`6ZL\\Ih3d6[LVIh3j6ZLRIi3l6YLPIj3P7XLlHl3R7ULlHm3R7ULlHn3R7SLlHo3S7SLjHn3U7TLiHn3V7VLdHl3\\7Q1000000O100000001O0000000O101O00\\NiHUMW7`2`ITM`6h2TJhLl5V3YJgLg5W3\\JiLb5V3aJiL_5U3cJkL]5T3eJjL\\5X3cJfL^5[3bJbL`5^3aJ`L`5`3`J^Lb5b3_J\\Lb5d3Q210000000000000000O10001O000000000O1000000000000gH[Lm4e3oJ_LQ5a3jJdLV5\\3fJhLZ5X3bJlL^5U3]JoLc5Q3YJSMg5m2UJWMk5i2QJ[Mo5e2mI_MS6a2iIcMX6\\2cIiM]6W2_ImMa6T2ZIPNf6T2RIPNn6S2kHQNU7R40O1N2M3N2N2M3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2N2M3N3L3N2N2M3N2M3N2M3N2N201O01O00000000O1O1O1O1O2N1O1O1O1O1001O2N1O2N1O2O1N4L4L3M4L<Df0ZOf0ZO3M0000001O00000000O100000001O0PMcH3]7MfHO[70hHNX72jHZOLXNZ7]2nHQO3^No6a2PIbNa0jN_6d2SIQNP1WOm5h2UIbMl8]2Q10XGdMb6\\2]IfMb6[2\\IgMc6Z2[IiMc6X2\\IhMd6Y2ZIgMg6Z2WIgMi6Z2VIfMj6^2oHeMQ7`2fHcM[7c2[H`Mf7e2QH^MQ8f2fG]M[8i30O1O1O1O2N3M4L3M4L3N3H7H9F9gMUFi0T:QO\\F<l9_OdFNe9MkF_O]9<m1J5Kak9"}, "label": "an elephant walking to the side with only one eye showing"}]}
{"id": 221889, "image": "COCO_train2014_000000221889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on the left not directly facing camera\"."}], "task": "refering", "answer_template": "The \"elephant on the left not directly facing camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 201, 202, 204, 205, 221, 222, 223, 224, 227], "bbox": [0.610484375, 0.1997887323943662, 0.9624843750000001, 0.6610093896713615], "iscrowd": 0, "area": 30549.262599999995, "rle": {"size": [426, 640], "counts": "\\jR5c1e;m0TO2N001O00eMkEc1T:aN\\Fk0d9XOaFa0^9CfF5\\O[OU9c0cGLSOFX9c0iG_OjN5[9?YHGf7=QHIn7:jGLU8T1fFROY9`2O2[NYLdIh3\\6ZL_Ii3`6ZL\\Ih3d6[LVIh3j6ZLRIi3l6YLPIj3P7XLlHl3R7ULlHm3R7ULlHn3R7SLlHo3S7SLjHn3U7TLiHn3V7VLdHl3\\7Q1000000O100000001O0000000O101O00\\NiHUMW7`2`ITM`6h2TJhLl5V3YJgLg5W3\\JiLb5V3aJiL_5U3cJkL]5T3eJjL\\5X3cJfL^5[3bJbL`5^3aJ`L`5`3`J^Lb5b3_J\\Lb5d3Q210000000000000000O10001O000000000O1000000000000gH[Lm4e3oJ_LQ5a3jJdLV5\\3fJhLZ5X3bJlL^5U3]JoLc5Q3YJSMg5m2UJWMk5i2QJ[Mo5e2mI_MS6a2iIcMX6\\2cIiM]6W2_ImMa6T2ZIPNf6T2RIPNn6S2kHQNU7R40O1N2M3N2N2M3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2N2M3N3L3N2N2M3N2M3N2M3N2N201O01O00000000O1O1O1O1O2N1O1O1O1O1001O2N1O2N1O2O1N4L4L3M4L<Df0ZOf0ZO3M0000001O00000000O100000001O0PMcH3]7MfHO[70hHNX72jHZOLXNZ7]2nHQO3^No6a2PIbNa0jN_6d2SIQNP1WOm5h2UIbMl8]2Q10XGdMb6\\2]IfMb6[2\\IgMc6Z2[IiMc6X2\\IhMd6Y2ZIgMg6Z2WIgMi6Z2VIfMj6^2oHeMQ7`2fHcM[7c2[H`Mf7e2QH^MQ8f2fG]M[8i30O1O1O1O2N3M4L3M4L3N3H7H9F9gMUFi0T:QO\\F<l9_OdFNe9MkF_O]9<m1J5Kak9"}, "label": "elephant on the left not directly facing camera"}]}
{"id": 221889, "image": "COCO_train2014_000000221889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the front walking\"."}], "task": "refering", "answer_template": "The \"the elephant in the front walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 313, 335], "bbox": [0.299828125, 0.15652582159624415, 0.670125, 0.9403051643192488], "iscrowd": 0, "area": 52532.48385000001, "rle": {"size": [426, 640], "counts": "XQ`26P=8I6I8H8H6J6J6J6J7K4L4L4PGgMi6]2SIgMi6]2RIhMj6\\2QIjMj6Z2RIjMj6Z2QIlMj6X2RIlMj6X2QImMk6W2QInMl6T2oHQNn6U2kHnMS7Z2cHjM\\7S4N2O1N2N2O1N2N2O1N3M2dJcIi3`6kKfIW4^6]KlIb4W6UKoIk4T6lJRJS5T7O00001O0O2O00001N2O2M2bKiGf3Y8RLnGm3U8kKRHT4m7kKUHT4a8OYOmKYHR4`8O2N1M3L4M3M301N10000O100O2O000O100O110O00000000O2O00000O1000O10O0PNfLRJZ3o4kLbI9Y1l2T5RNfJo1Y5WN`Jj1_5]N[Jd1d5cNUJ^1i5jNPJW1o5oNjIS1T6UOeIl0Z6R3O0100]KQJj1P6QNUJn1l5PNWJo1i5nMZJQ2g5mM\\JR2d5kM_JT2b5jMaJU2^5iMeJV2]5gMfJX2^5aMeJ^2m5oLTJR3R6eLQJ[3U6]LmIc3m7000000000000000000000000000000000dNWMWHi2b7aM[H_2_7iM_HW2\\7QNaHo1^7TN`Hl1^7XN`Hi1^7R2O1O2NSKgHk3W7RLQIk3n6RLYIk3f6nKeIo3Z6cKXJY4`7aLSG[2o8aMTG]2o8_MSGa2n8[MVGd2l8XMVGg2l8VMWGi2k8RMYGl2i8QMYGo2b9O0O2O0O2O000001O1O001O001O1O001O00100O001O001O1O000000O10000OeJQM=o2_OjMLV21bNZO]1c0[OjNd0S14WNMf1k0dMUOY2d1RM]Nj2m1nLTNn2Q2PMQNl2S2RMnMk2V2SMlMg2Z2XMfMd2`2YMbMa2d2]M]M_2h2_MZM\\2k2bMVM]2l2aMVM]2l2bMTM]2o2`MSM^2o2`MTM]2n2aMTM^2m2`MUM^2m2aMTM]2n6N2N3M2M3N2N2M3N2N3L3N2N2M3L4K5L4K5L4N2O1N2O1N2O1N2O2M5L4KSeg2"}, "label": "the elephant in the front walking"}]}
{"id": 221889, "image": "COCO_train2014_000000221889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant infront of camera\"."}], "task": "refering", "answer_template": "The \"elephant infront of camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 313, 335], "bbox": [0.299828125, 0.15652582159624415, 0.670125, 0.9403051643192488], "iscrowd": 0, "area": 52532.48385000001, "rle": {"size": [426, 640], "counts": "XQ`26P=8I6I8H8H6J6J6J6J7K4L4L4PGgMi6]2SIgMi6]2RIhMj6\\2QIjMj6Z2RIjMj6Z2QIlMj6X2RIlMj6X2QImMk6W2QInMl6T2oHQNn6U2kHnMS7Z2cHjM\\7S4N2O1N2N2O1N2N2O1N3M2dJcIi3`6kKfIW4^6]KlIb4W6UKoIk4T6lJRJS5T7O00001O0O2O00001N2O2M2bKiGf3Y8RLnGm3U8kKRHT4m7kKUHT4a8OYOmKYHR4`8O2N1M3L4M3M301N10000O100O2O000O100O110O00000000O2O00000O1000O10O0PNfLRJZ3o4kLbI9Y1l2T5RNfJo1Y5WN`Jj1_5]N[Jd1d5cNUJ^1i5jNPJW1o5oNjIS1T6UOeIl0Z6R3O0100]KQJj1P6QNUJn1l5PNWJo1i5nMZJQ2g5mM\\JR2d5kM_JT2b5jMaJU2^5iMeJV2]5gMfJX2^5aMeJ^2m5oLTJR3R6eLQJ[3U6]LmIc3m7000000000000000000000000000000000dNWMWHi2b7aM[H_2_7iM_HW2\\7QNaHo1^7TN`Hl1^7XN`Hi1^7R2O1O2NSKgHk3W7RLQIk3n6RLYIk3f6nKeIo3Z6cKXJY4`7aLSG[2o8aMTG]2o8_MSGa2n8[MVGd2l8XMVGg2l8VMWGi2k8RMYGl2i8QMYGo2b9O0O2O0O2O000001O1O001O001O1O001O00100O001O001O1O000000O10000OeJQM=o2_OjMLV21bNZO]1c0[OjNd0S14WNMf1k0dMUOY2d1RM]Nj2m1nLTNn2Q2PMQNl2S2RMnMk2V2SMlMg2Z2XMfMd2`2YMbMa2d2]M]M_2h2_MZM\\2k2bMVM]2l2aMVM]2l2bMTM]2o2`MSM^2o2`MTM]2n2aMTM^2m2`MUM^2m2aMTM]2n6N2N3M2M3N2N2M3N2N3L3N2N2M3L4K5L4K5L4N2O1N2O1N2O1N2O2M5L4KSeg2"}, "label": "elephant infront of camera"}]}
{"id": 254660, "image": "COCO_train2014_000000254660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black chair\"."}], "task": "refering", "answer_template": "The \"black chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 18, 19, 20, 35, 36, 37, 52, 53, 54, 68, 69, 70, 71, 85, 86, 87, 88, 102, 103, 104, 119, 120, 121], "bbox": [0.000397489539748954, 0.0, 0.187092050209205, 0.356015625], "iscrowd": 0, "area": 13646.327850000001, "rle": {"size": [640, 478], "counts": "a64Sa02TA:l>]2000O10O1000O1000000O1000000O100000000O10000001O001O1O001XLTMZFl2f9ZMTFg2k9`MnE`2R:gMgEZ2X:lMbEU2]:RN\\En1d:YNUEh1j:^NPEb1P;eNiD\\1V;kNcDV1\\;PO^DP1b;WOWDj0h;]OQDc0o;CkC>T<IeC7[<0^C1a<5YCLf<;SCEm<h3O0010O01O1O001O0000O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1[NcBfL^=Z3eBbL\\=^3gB^LZ=b3jBYLW=f3mBVLT=j3oBRLR=n3QCnKP=Q4TCkKm<U4WCeKk<[4XCaKi<^4[C^Kf<b4]CZKd<f4j1@aSc7"}, "label": "black chair"}]}
{"id": 254660, "image": "COCO_train2014_000000254660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair with a leather seat\"."}], "task": "refering", "answer_template": "The \"a black chair with a leather seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 18, 19, 20, 35, 36, 37, 52, 53, 54, 68, 69, 70, 71, 85, 86, 87, 88, 102, 103, 104, 119, 120, 121], "bbox": [0.000397489539748954, 0.0, 0.187092050209205, 0.356015625], "iscrowd": 0, "area": 13646.327850000001, "rle": {"size": [640, 478], "counts": "a64Sa02TA:l>]2000O10O1000O1000000O1000000O100000000O10000001O001O1O001XLTMZFl2f9ZMTFg2k9`MnE`2R:gMgEZ2X:lMbEU2]:RN\\En1d:YNUEh1j:^NPEb1P;eNiD\\1V;kNcDV1\\;PO^DP1b;WOWDj0h;]OQDc0o;CkC>T<IeC7[<0^C1a<5YCLf<;SCEm<h3O0010O01O1O001O0000O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1[NcBfL^=Z3eBbL\\=^3gB^LZ=b3jBYLW=f3mBVLT=j3oBRLR=n3QCnKP=Q4TCkKm<U4WCeKk<[4XCaKi<^4[C^Kf<b4]CZKd<f4j1@aSc7"}, "label": "a black chair with a leather seat"}]}
{"id": 459465, "image": "COCO_train2014_000000459465.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and black sheep looking this way\"."}], "task": "refering", "answer_template": "The \"a white and black sheep looking this way\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 205, 206, 207, 208, 209, 210, 211, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291], "bbox": [0.0, 0.27742187500000004, 0.5965116279069766, 0.769890625], "iscrowd": 0, "area": 34448.9742, "rle": {"size": [640, 473], "counts": "[=g0Xc0\\1eN00O1000000O1000O1000O1000000O10O100000O1000000O10O100000O1000000O10O100000O100000hMSN]Bl1c=UN]Bk1b=VN]Bk1b=VN^Bj1a=WN_Bi1`=XN`Bg1`=ZN`Bf1_=[NaBe1^=\\NbBd1]=]NcBb1]=_NbBb1]=_NcBa1\\=_NeBa1Z=`NfB`1Y=aNgB^1Y=cNgB]1X=dNhB\\1W=eNiB[1V=fNjBY1V=hNiBY1V=hNjBX1U=iNkBW1T=jNlBV1S=kNmBT1S=mNmBS1R=nNnBR1Q=oNoBQ1P=POPCo0P=ROoBo0P=ROPCn0o<SOQCm0n<TORCl0m<UOSCj0m<WOSCi0l<XOTCh0k<YOUCg0j<ZOVCe0j<\\OUCe0j<[OWCe0h<\\OXCd0g<]OYCc0f<^OZCa0f<@ZC`0e<A[C?d<B\\C>c<C\\C=d<D\\C<c<E]C;b<F^C:a<G_C9`<H`C7`<J`C6_<KaC5^<LbC4]<MbC3^<NbC2]<OcC1k:YNcDg1b00j:YNdDh1b0Oi:ZNdDh1c0Mi:\\NcDh1d0Lh:]NcDh1e0Kg:^NcDh1f0Jf:^NdDh1f0Jf:_NcDh1g0Ie:`NcDh1h0Hd:aNcDh1i0Gc:aNdDi1i0Fb:bNdDi1j0Dc:cNbDj1k0Cb:dNbDj1l0Ba:dNcDk1l0A`:eNcDk1l0@a:fNbDk1m0_O`:gNbDk1n0^O_:hNbDk1o0]O^:hNcDl1o0\\O]:iNcDl1P1ZO]:kNbDl1Q1YO\\:lNbDl1R1XO[:lNcDm1R1WOZ:mNcDm1R1VO[:nNbDm1S1UOZ:oNbDm1T1TOZ:nNbDo1T1SOY:oNbDo1U1ROX:PObDo1V1POX:RO`DP2X1nNW:SO`Do1Z1nNU:SO`DQ2Z1mNU:SO_DR2\\1jNU:UO]DS2^1hNT:VO]DS2_1gNS:VO]DU2`1eNR:WO\\DV2b1cNR:WO[DW2c1aNR:YOYDX2e1_NQ:ZOXDY2g1]NP:ZOYDZ2f1]NQ:YOWD\\2h1ZNQ:[OUD]2j1XNP:\\OTD^2l1VNP:[OTD`2l1UNo9\\OSDa2n1SNn9n2RFQMo9o2QFQMn9P3RFPMm9Q3RFPMn9P3RFoLn9R3RFnLm9S3SFmLl9T3TFlLl9T3TFlLk9U3UFjLk9V3VFjLj9V3VFjLi9W3VFjLi9W3WFhLj9X3VFhLi9Y3WFgLh9Z3XFfLg9[3YFeLg9[3YFdLg9]3YFcLf9^3YFcLg9]3YFcLf9^3ZFaLf9`3ZF`Lf9`3ZF`Le9a3[F_Ld9b3\\F^Ld9b3\\F]Le9c3[F]Le9c3ZF^Lf9b3ZF^Lf9b3ZF]Lf9d3ZF\\Lf9d3ZF\\Lf9d3ZF\\Lf9c3[F]Le9c3[F\\Le9e3[F[Le9e3ZF\\Lf9d3ZF\\Lf9d3ZF[Lg9e3YF[Lg9e3YF[Lf9f3ZFZLf9f3ZFZLf9f3ZFYLg9g3YFYLg9g3XFZLh9f3XFZLg9g3YFXLh9h3XFXLh9h3XFXLh9h3XFXLh9h3XFXLh9h3XFWLh9j3WFWLi9i3WFWLi9i3WFWLi9i3WFVLj9j3VFVLj9i3WFWLh9j3XFVLh9j3XFVLh9j3XFULi9k3VFVLj9c2^DVMh13n9g2^DRMd1OV:P3ZDlL`1K_:Z3UDfL\\1Hh:b3PDaLX1EP;k3lC[LU>f3oAULR>k3RBPLn=P4Q110O00010O001O00001O00001O00001O001O1O001O1O001O1O1O001O1O102M4L4L3M4L3oMT_Oh0Pa0TOW_Od0n`0XOX_O<Pa0CV_O0Qa0OV_ODRa0:T_OZOSa0f0T1O2N1O2O0O1O2N101N1O1O2O0O2N100O2N1O2OmSg3"}, "label": "a white and black sheep looking this way"}]}
{"id": 459465, "image": "COCO_train2014_000000459465.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep\"."}], "task": "refering", "answer_template": "The \"a sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 205, 206, 207, 208, 209, 210, 211, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291], "bbox": [0.0, 0.27742187500000004, 0.5965116279069766, 0.769890625], "iscrowd": 0, "area": 34448.9742, "rle": {"size": [640, 473], "counts": "[=g0Xc0\\1eN00O1000000O1000O1000O1000000O10O100000O1000000O10O100000O1000000O10O100000O100000hMSN]Bl1c=UN]Bk1b=VN]Bk1b=VN^Bj1a=WN_Bi1`=XN`Bg1`=ZN`Bf1_=[NaBe1^=\\NbBd1]=]NcBb1]=_NbBb1]=_NcBa1\\=_NeBa1Z=`NfB`1Y=aNgB^1Y=cNgB]1X=dNhB\\1W=eNiB[1V=fNjBY1V=hNiBY1V=hNjBX1U=iNkBW1T=jNlBV1S=kNmBT1S=mNmBS1R=nNnBR1Q=oNoBQ1P=POPCo0P=ROoBo0P=ROPCn0o<SOQCm0n<TORCl0m<UOSCj0m<WOSCi0l<XOTCh0k<YOUCg0j<ZOVCe0j<\\OUCe0j<[OWCe0h<\\OXCd0g<]OYCc0f<^OZCa0f<@ZC`0e<A[C?d<B\\C>c<C\\C=d<D\\C<c<E]C;b<F^C:a<G_C9`<H`C7`<J`C6_<KaC5^<LbC4]<MbC3^<NbC2]<OcC1k:YNcDg1b00j:YNdDh1b0Oi:ZNdDh1c0Mi:\\NcDh1d0Lh:]NcDh1e0Kg:^NcDh1f0Jf:^NdDh1f0Jf:_NcDh1g0Ie:`NcDh1h0Hd:aNcDh1i0Gc:aNdDi1i0Fb:bNdDi1j0Dc:cNbDj1k0Cb:dNbDj1l0Ba:dNcDk1l0A`:eNcDk1l0@a:fNbDk1m0_O`:gNbDk1n0^O_:hNbDk1o0]O^:hNcDl1o0\\O]:iNcDl1P1ZO]:kNbDl1Q1YO\\:lNbDl1R1XO[:lNcDm1R1WOZ:mNcDm1R1VO[:nNbDm1S1UOZ:oNbDm1T1TOZ:nNbDo1T1SOY:oNbDo1U1ROX:PObDo1V1POX:RO`DP2X1nNW:SO`Do1Z1nNU:SO`DQ2Z1mNU:SO_DR2\\1jNU:UO]DS2^1hNT:VO]DS2_1gNS:VO]DU2`1eNR:WO\\DV2b1cNR:WO[DW2c1aNR:YOYDX2e1_NQ:ZOXDY2g1]NP:ZOYDZ2f1]NQ:YOWD\\2h1ZNQ:[OUD]2j1XNP:\\OTD^2l1VNP:[OTD`2l1UNo9\\OSDa2n1SNn9n2RFQMo9o2QFQMn9P3RFPMm9Q3RFPMn9P3RFoLn9R3RFnLm9S3SFmLl9T3TFlLl9T3TFlLk9U3UFjLk9V3VFjLj9V3VFjLi9W3VFjLi9W3WFhLj9X3VFhLi9Y3WFgLh9Z3XFfLg9[3YFeLg9[3YFdLg9]3YFcLf9^3YFcLg9]3YFcLf9^3ZFaLf9`3ZF`Lf9`3ZF`Le9a3[F_Ld9b3\\F^Ld9b3\\F]Le9c3[F]Le9c3ZF^Lf9b3ZF^Lf9b3ZF]Lf9d3ZF\\Lf9d3ZF\\Lf9d3ZF\\Lf9c3[F]Le9c3[F\\Le9e3[F[Le9e3ZF\\Lf9d3ZF\\Lf9d3ZF[Lg9e3YF[Lg9e3YF[Lf9f3ZFZLf9f3ZFZLf9f3ZFYLg9g3YFYLg9g3XFZLh9f3XFZLg9g3YFXLh9h3XFXLh9h3XFXLh9h3XFXLh9h3XFXLh9h3XFWLh9j3WFWLi9i3WFWLi9i3WFWLi9i3WFVLj9j3VFVLj9i3WFWLh9j3XFVLh9j3XFVLh9j3XFULi9k3VFVLj9c2^DVMh13n9g2^DRMd1OV:P3ZDlL`1K_:Z3UDfL\\1Hh:b3PDaLX1EP;k3lC[LU>f3oAULR>k3RBPLn=P4Q110O00010O001O00001O00001O00001O001O1O001O1O001O1O1O001O1O102M4L4L3M4L3oMT_Oh0Pa0TOW_Od0n`0XOX_O<Pa0CV_O0Qa0OV_ODRa0:T_OZOSa0f0T1O2N1O2O0O1O2N101N1O1O2O0O2N100O2N1O2OmSg3"}, "label": "a sheep"}]}
{"id": 49866, "image": "COCO_train2014_000000049866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a messanger bag is wear by the mand\"."}], "task": "refering", "answer_template": "The \"a messanger bag is wear by the mand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 109, 110, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205], "bbox": [0.7638906249999999, 0.24944289693593313, 0.9403593749999999, 0.6539275766016713], "iscrowd": 0, "area": 9127.241100000003, "rle": {"size": [359, 640], "counts": "bd[5<Z:b0^Oa0^Ob0G9O100O1O1O2YNf100O10000O100003NQ1nNc0]O5K1OO100000001O000000000000000N3M2N2N2O100000001O000001O0000000001O00000000000000010O000001O00001O00010O001O00001O1O100O1O00001O00000001O0001O0000000000010OkNoMTIo1n6o06K5J<D`0A?@`0A5J7I6K6I6J7JaU="}, "label": "a messanger bag is wear by the mand"}]}
{"id": 49866, "image": "COCO_train2014_000000049866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black colored bag worn by a man in the local train\"."}], "task": "refering", "answer_template": "The \"a black colored bag worn by a man in the local train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 109, 110, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205], "bbox": [0.7638906249999999, 0.24944289693593313, 0.9403593749999999, 0.6539275766016713], "iscrowd": 0, "area": 9127.241100000003, "rle": {"size": [359, 640], "counts": "bd[5<Z:b0^Oa0^Ob0G9O100O1O1O2YNf100O10000O100003NQ1nNc0]O5K1OO100000001O000000000000000N3M2N2N2O100000001O000001O0000000001O00000000000000010O000001O00001O00010O001O00001O1O100O1O00001O00000001O0001O0000000000010OkNoMTIo1n6o06K5J<D`0A?@`0A5J7I6K6I6J7JaU="}, "label": "a black colored bag worn by a man in the local train"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass goblet in front of the other is clear and white\"."}], "task": "refering", "answer_template": "The \"the glass goblet in front of the other is clear and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 210, 225, 240, 255, 270, 285, 286, 300, 301, 302, 303, 315, 316, 317, 318, 330, 331, 332], "bbox": [0.0, 0.5519999999999999, 0.2426229508196721, 1.0], "iscrowd": 0, "area": 6758.09765, "rle": {"size": [640, 427], "counts": "R>g1m3]N[7e4bHbK\\7]4`HdKd7Z4YHgKk7W4QHkKS8S4jGnK[8P4aGQLc8m3ZGTLj8j3SGWLQ9h3kFYL[9Z1cD_On1YOe9m0eDIc1[On9b0fD2Y1]OW:6hD=m0^O`:KkDf0b0@i:@lDo08B[==bBDa=<ZBFi=9TBHo=7nAJU>6gAK[>5bALa>3\\ANg>1VA0m>0n@2U?Mh@4[?Kb@6a?I\\@8g?HU@9n?k13M3M2N3M1O2N1O2N1O2N1O2N1O2N1O2N1O2M2N3N1N101N2N2O1N1O2O1N2N100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1N3L4M3L3N3L4MgmY6"}, "label": "the glass goblet in front of the other is clear and white"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass of water in table\"."}], "task": "refering", "answer_template": "The \"glass of water in table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 210, 225, 240, 255, 270, 285, 286, 300, 301, 302, 303, 315, 316, 317, 318, 330, 331, 332], "bbox": [0.0, 0.5519999999999999, 0.2426229508196721, 1.0], "iscrowd": 0, "area": 6758.09765, "rle": {"size": [640, 427], "counts": "R>g1m3]N[7e4bHbK\\7]4`HdKd7Z4YHgKk7W4QHkKS8S4jGnK[8P4aGQLc8m3ZGTLj8j3SGWLQ9h3kFYL[9Z1cD_On1YOe9m0eDIc1[On9b0fD2Y1]OW:6hD=m0^O`:KkDf0b0@i:@lDo08B[==bBDa=<ZBFi=9TBHo=7nAJU>6gAK[>5bALa>3\\ANg>1VA0m>0n@2U?Mh@4[?Kb@6a?I\\@8g?HU@9n?k13M3M2N3M1O2N1O2N1O2N1O2N1O2N1O2N1O2M2N3N1N101N2N2O1N1O2O1N2N100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1N3L4M3L3N3L4MgmY6"}, "label": "glass of water in table"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of chocolate cake\"."}], "task": "refering", "answer_template": "The \"slice of chocolate cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246, 247, 257, 258, 259, 260, 261, 272, 273, 274, 275], "bbox": [0.14594847775175643, 0.6185625, 0.5059718969555035, 0.811984375], "iscrowd": 0, "area": 13251.313899999997, "rle": {"size": [640, 427], "counts": "oUW11lc0:F:E<F9F:F5L4L4M2N3N2M3N1N3N2M3G8L5N2N2N001N2O1O001O1O1N101O1O00000O1000O100000O10O1000000O010000000O10O1000O10000000O01000000O10O100000O1000O10O100000000O1000000O100000000O10005K0O100000000O01O1O1O001O1O11O0O01O100O00100000O101O00000O102N1O2M3N1O2N2ZOk^O]NVa0]1R_O_NQa0Z1V_OdNk`0V1T1H7J7M2N10001N103M3L2O002M6K3M0N2N3O1O2O2M10Pd00QjR4"}, "label": "slice of chocolate cake"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chocolate cake in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the chocolate cake in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246, 247, 257, 258, 259, 260, 261, 272, 273, 274, 275], "bbox": [0.14594847775175643, 0.6185625, 0.5059718969555035, 0.811984375], "iscrowd": 0, "area": 13251.313899999997, "rle": {"size": [640, 427], "counts": "oUW11lc0:F:E<F9F:F5L4L4M2N3N2M3N1N3N2M3G8L5N2N2N001N2O1O001O1O1N101O1O00000O1000O100000O10O1000000O010000000O10O1000O10000000O01000000O10O100000O1000O10O100000000O1000000O100000000O10005K0O100000000O01O1O1O001O1O11O0O01O100O00100000O101O00000O102N1O2M3N1O2N2ZOk^O]NVa0]1R_O_NQa0Z1V_OdNk`0V1T1H7J7M2N10001N103M3L2O002M6K3M0N2N3O1O2O2M10Pd00QjR4"}, "label": "the chocolate cake in the right hand picture"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cheese cake with coco graham cracker crust , fresh whipped cream with caramel drizzle\"."}], "task": "refering", "answer_template": "The \"cheese cake with coco graham cracker crust , fresh whipped cream with caramel drizzle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 265, 266, 267, 268, 269, 280, 281, 282], "bbox": [0.6688992974238877, 0.6172187499999999, 1.0, 0.8065625000000001], "iscrowd": 0, "area": 10489.406500000003, "rle": {"size": [640, 427], "counts": "cUc56hc09G<E2N2N2M3O1N2O1O1O1O2M2O2N:F3L4M4K5L4L5J6K4K2N2O0O1O010O1O010O1O010O001O1O010O1O001O10O0100O010O10O0100O010O10O10O10O1000O01000O01000O010001N2O001N2O002M4M3M2M3N2N1N2O1O1N101O0O2O000O100O100O100O100O2O1N2O1O1N2O2M3N2M2O0O2O0O101O0O010O10O010O010O01000O02O2M2O2M3N2M3N1N3N1N3N:FgA"}, "label": "cheese cake with coco graham cracker crust , fresh whipped cream with caramel drizzle"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of white cheese cake on the table\"."}], "task": "refering", "answer_template": "The \"a slice of white cheese cake on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 265, 266, 267, 268, 269, 280, 281, 282], "bbox": [0.6688992974238877, 0.6172187499999999, 1.0, 0.8065625000000001], "iscrowd": 0, "area": 10489.406500000003, "rle": {"size": [640, 427], "counts": "cUc56hc09G<E2N2N2M3O1N2O1O1O1O2M2O2N:F3L4M4K5L4L5J6K4K2N2O0O1O010O1O010O1O010O001O1O010O1O001O10O0100O010O10O0100O010O10O10O10O1000O01000O01000O010001N2O001N2O002M4M3M2M3N2N1N2O1O1N101O0O2O000O100O100O100O100O2O1N2O1O1N2O2M3N2M2O0O2O0O101O0O010O10O010O010O01000O02O2M2O2M3N2M3N1N3N1N3N:FgA"}, "label": "a slice of white cheese cake on the table"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown marble table\"."}], "task": "refering", "answer_template": "The \"the brown marble table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 274, 275, 282, 283, 284, 288, 289, 290, 291, 293, 294, 295, 296, 297, 298, 299, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.2551288056206089, 0.7678906249999999, 0.9994847775175644, 0.992609375], "iscrowd": 0, "area": 31828.915650000003, "rle": {"size": [640, 427], "counts": "edT2U3k`00O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O1000mLd_Ob2[`0]Mg_Oc2Y`0\\Mi_Oc2W`0\\Mk_Oc2U`0\\Ml_Od2S`0\\Mo_Oc2Q`0[MR@d2n?[MS@e2l?[MV@d2j?[MX@d2h?[MY@e2g?YM\\@f2c?ZM_@e2a?ZM`@f2`?YMb@f2^?YMd@f2[?[Me@e2[?ZMg@e2[`0O001O1O001O1O001O1O001O000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O1000000O1000000O1000000O1000000O10001O0O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O1b0^Od0\\Od0\\Od0\\Od0\\Od0\\Od0\\O[\\O"}, "label": "the brown marble table"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stained table with plates of pie on it\"."}], "task": "refering", "answer_template": "The \"a stained table with plates of pie on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 274, 275, 282, 283, 284, 288, 289, 290, 291, 293, 294, 295, 296, 297, 298, 299, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.2551288056206089, 0.7678906249999999, 0.9994847775175644, 0.992609375], "iscrowd": 0, "area": 31828.915650000003, "rle": {"size": [640, 427], "counts": "edT2U3k`00O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O1000mLd_Ob2[`0]Mg_Oc2Y`0\\Mi_Oc2W`0\\Mk_Oc2U`0\\Ml_Od2S`0\\Mo_Oc2Q`0[MR@d2n?[MS@e2l?[MV@d2j?[MX@d2h?[MY@e2g?YM\\@f2c?ZM_@e2a?ZM`@f2`?YMb@f2^?YMd@f2[?[Me@e2[?ZMg@e2[`0O001O1O001O1O001O1O001O000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O1000000O1000000O1000000O1000000O10001O0O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O1b0^Od0\\Od0\\Od0\\Od0\\Od0\\Od0\\O[\\O"}, "label": "a stained table with plates of pie on it"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a fork and wearing a napkin\"."}], "task": "refering", "answer_template": "The \"a man holding a fork and wearing a napkin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 37, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 159, 160, 161, 162, 174, 175, 176, 177, 189, 190, 191, 192, 200, 201, 202, 203, 204, 205, 206, 207, 215, 216, 217, 218, 219, 220], "bbox": [0.3806088992974239, 0.058421875, 1.0, 0.638203125], "iscrowd": 0, "area": 47649.312150000005, "rle": {"size": [640, 427], "counts": "[XV36ec07J4L4L4L4K5O100001O00000000000001O0001O0000000000001O0000000000001O00000000YGXOVNh0i1^ORNc0k1DQN;n1KmM5R21iMOU29eMGZ2?aMA]2b0bM^O]2d0cM[O\\2g0cMYO[2j0dMVO[2l0dMTOZ2P1dMPO[2R1dMnNZ2U1eMkNZ2W1fMhNY2Z1fMfNX2]1gMcNX2_1gMaNW2b1hM^NW2d1hM\\NV2g1jMXNU2j1jMVNU2l1kMTNR2o1mMQNR2Q2nMnMP2U2oMkMP2W2PNhMn1[2RNdMm1^2RNbMm1`2SN_Mk1e2SN[Ml1g2TNXMj1k2UNUMk1l2UNSMk1n2TNRMk1P3UNoLk1R3UNmLk1T3TNlLl1U3TNjLl1W3SNiLm1X3SNgLl1[3SNeLm1\\3SNcLm1^3RNbLn1_3RN`Ln1a3RN_Lm1b3RN^Lm1d3SN[Lm1f3RNZLn1g3RNXLn1j3PNVLP2k3PNTLo1n3YJUL:M]5o3WJfLLZOm5Q4UJXMYOjNb6o3SJjMiMWOT8P3QJIo58oIIP69nIHR69lIHT69jIHV69hIHX69fIHY6:fIFZ6;dIF\\6:cIG]6:aIG_69`II_68_II`68_IIa68]IIc68[IIe67ZIJf67XIJg67XIJh67VIJj66UIKk66SIKm65RILn65QIKn67PIJQ75nHLR75lHLU74iHMW76eHK\\77`HJ`7:[HGf7;VHFj7=RHCP8?lGAU8b0gG]O\\8e0`G[Oa8h0[GXOg8o51N3N1O2M2O1O2M2O2N1N3N1O2N102N1N2O2N1O2N1N3N1O1O2N1O2M2O2N1O1O2M2O2N0000000O1000001O2M4M2N2N3M2N2M3N3M2mGcDd53bK\\;gNdD`59eKU;hNdD]5?gKQ;hNbD[5e0jKk:hNcDW5k0mKd:iNaDV5S1nKh:m3aEoK_:n3gEoKX:Q4lEkKU:T4PFhKP:W4TFfKk9[4XFbKh9]4]F_Kc9`4aF\\K_9e4eFWK[9h4iFUKW9j4o210O1O010O1O010O1O010O1O100O1O100O1O101N1O2O0O2N2N101N101N2O0O2O1N101N101N2O0O10000O2O0O100O101N100O100O101N100O1N2\\Oe0[Od0]Oc0\\Od0\\O]L"}, "label": "a man holding a fork and wearing a napkin"}]}
{"id": 197323, "image": "COCO_train2014_000000197323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is eating desert at the table\"."}], "task": "refering", "answer_template": "The \"a man is eating desert at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 37, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 159, 160, 161, 162, 174, 175, 176, 177, 189, 190, 191, 192, 200, 201, 202, 203, 204, 205, 206, 207, 215, 216, 217, 218, 219, 220], "bbox": [0.3806088992974239, 0.058421875, 1.0, 0.638203125], "iscrowd": 0, "area": 47649.312150000005, "rle": {"size": [640, 427], "counts": "[XV36ec07J4L4L4L4K5O100001O00000000000001O0001O0000000000001O0000000000001O00000000YGXOVNh0i1^ORNc0k1DQN;n1KmM5R21iMOU29eMGZ2?aMA]2b0bM^O]2d0cM[O\\2g0cMYO[2j0dMVO[2l0dMTOZ2P1dMPO[2R1dMnNZ2U1eMkNZ2W1fMhNY2Z1fMfNX2]1gMcNX2_1gMaNW2b1hM^NW2d1hM\\NV2g1jMXNU2j1jMVNU2l1kMTNR2o1mMQNR2Q2nMnMP2U2oMkMP2W2PNhMn1[2RNdMm1^2RNbMm1`2SN_Mk1e2SN[Ml1g2TNXMj1k2UNUMk1l2UNSMk1n2TNRMk1P3UNoLk1R3UNmLk1T3TNlLl1U3TNjLl1W3SNiLm1X3SNgLl1[3SNeLm1\\3SNcLm1^3RNbLn1_3RN`Ln1a3RN_Lm1b3RN^Lm1d3SN[Lm1f3RNZLn1g3RNXLn1j3PNVLP2k3PNTLo1n3YJUL:M]5o3WJfLLZOm5Q4UJXMYOjNb6o3SJjMiMWOT8P3QJIo58oIIP69nIHR69lIHT69jIHV69hIHX69fIHY6:fIFZ6;dIF\\6:cIG]6:aIG_69`II_68_II`68_IIa68]IIc68[IIe67ZIJf67XIJg67XIJh67VIJj66UIKk66SIKm65RILn65QIKn67PIJQ75nHLR75lHLU74iHMW76eHK\\77`HJ`7:[HGf7;VHFj7=RHCP8?lGAU8b0gG]O\\8e0`G[Oa8h0[GXOg8o51N3N1O2M2O1O2M2O2N1N3N1O2N102N1N2O2N1O2N1N3N1O1O2N1O2M2O2N1O1O2M2O2N0000000O1000001O2M4M2N2N3M2N2M3N3M2mGcDd53bK\\;gNdD`59eKU;hNdD]5?gKQ;hNbD[5e0jKk:hNcDW5k0mKd:iNaDV5S1nKh:m3aEoK_:n3gEoKX:Q4lEkKU:T4PFhKP:W4TFfKk9[4XFbKh9]4]F_Kc9`4aF\\K_9e4eFWK[9h4iFUKW9j4o210O1O010O1O010O1O010O1O100O1O100O1O101N1O2O0O2N2N101N101N2O0O2O1N101N101N2O0O10000O2O0O100O101N100O100O101N100O1N2\\Oe0[Od0]Oc0\\Od0\\O]L"}, "label": "a man is eating desert at the table"}]}
{"id": 443084, "image": "COCO_train2014_000000443084.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with red hair sitting at a table\"."}], "task": "refering", "answer_template": "The \"a woman with red hair sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 285, 286, 287], "bbox": [0.42534375, 0.37387892376681614, 0.693734375, 0.770269058295964], "iscrowd": 0, "area": 15350.082399999998, "rle": {"size": [446, 640], "counts": "gif31k=2M2O2N2N2N2M3N2N2M3L4L4L4M3L4L3M400O100O100O10O01O100O100O100O010O100O1000000O010000000O1000O1000O100O1O1N2O001O1N2]NPNWGQ2h8RNUGo1k8SNRGn1n8TNoFm1P9UNmFl1T9VNiFk1T9ZNiFg1U9]NhFd1U9aNgFa1V9dNgF]1V9gNgF[1V9jNfFX1Z9jNbFX1^9jN^FW1c9b100O100000000H710O100O100O1N2O001O1H8N2N2O1O1N101001O001O001O001N101O001O001O002N1N2O1O1O1O1O1O1O1O001N101O001O001O1O001O1N2O1O1O1aM^F`0b9\\OfF`0[9]OjF`0W9^OmF?T9^OQG?P9^OTG?n8_OTG`0m8]OVGb0l8ZOWGe0l8SO[Gk0h8lN_GS1d8cNdG\\1`:O1O1O2M2O1N2M3N2N3M2N2N2M3N2N2N2N2N2M4M2N2NXRe2"}, "label": "a woman with red hair sitting at a table"}]}
{"id": 443084, "image": "COCO_train2014_000000443084.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman getting her hair cut\"."}], "task": "refering", "answer_template": "The \"a woman getting her hair cut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 285, 286, 287], "bbox": [0.42534375, 0.37387892376681614, 0.693734375, 0.770269058295964], "iscrowd": 0, "area": 15350.082399999998, "rle": {"size": [446, 640], "counts": "gif31k=2M2O2N2N2N2M3N2N2M3L4L4L4M3L4L3M400O100O100O10O01O100O100O100O010O100O1000000O010000000O1000O1000O100O1O1N2O001O1N2]NPNWGQ2h8RNUGo1k8SNRGn1n8TNoFm1P9UNmFl1T9VNiFk1T9ZNiFg1U9]NhFd1U9aNgFa1V9dNgF]1V9gNgF[1V9jNfFX1Z9jNbFX1^9jN^FW1c9b100O100000000H710O100O100O1N2O001O1H8N2N2O1O1N101001O001O001O001N101O001O001O002N1N2O1O1O1O1O1O1O1O001N101O001O001O1O001O1N2O1O1O1aM^F`0b9\\OfF`0[9]OjF`0W9^OmF?T9^OQG?P9^OTG?n8_OTG`0m8]OVGb0l8ZOWGe0l8SO[Gk0h8lN_GS1d8cNdG\\1`:O1O1O2M2O1N2M3N2N3M2N2N2M3N2N2N2N2N2M4M2N2NXRe2"}, "label": "a woman getting her hair cut"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife on top of the pie in the back of picture\"."}], "task": "refering", "answer_template": "The \"the knife on top of the pie in the back of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 40, 41, 42, 65, 66], "bbox": [0.638234375, 0.0, 0.90928125, 0.18966057441253265], "iscrowd": 0, "area": 1373.8314500000006, "rle": {"size": [383, 640], "counts": "Woh41n;1O001O1O1O001O001O010O001O001O001O1O001O000001O01O01O00010O001O0010O01O000010O01O010O010O01O01O010O010O0010O0010O0010O010O01O01O010O010O01O01O010O01O010O01O01O010O010O01O01O010O010O010O01O01O010O010O01O01O010O010O01O010O01O010O01O010O010O01O01O010O010O010O00010O010O010O0010O0010O010O010O0010O0010O010O010O010O100O100O1O10N101N101N101N101N10SPf0"}, "label": "the knife on top of the pie in the back of picture"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife on top of the pie\"."}], "task": "refering", "answer_template": "The \"the knife on top of the pie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 40, 41, 42, 65, 66], "bbox": [0.638234375, 0.0, 0.90928125, 0.18966057441253265], "iscrowd": 0, "area": 1373.8314500000006, "rle": {"size": [383, 640], "counts": "Woh41n;1O001O1O1O001O001O010O001O001O001O1O001O000001O01O01O00010O001O0010O01O000010O01O010O010O01O01O010O010O0010O0010O0010O010O01O01O010O010O01O01O010O01O010O01O01O010O010O01O01O010O010O010O01O01O010O010O01O01O010O010O01O010O01O010O01O010O010O01O01O010O010O010O00010O010O010O0010O0010O010O010O0010O0010O010O010O010O100O100O1O10N101N101N101N101N10SPf0"}, "label": "the knife on top of the pie"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spoon in the snacks which is exactly under the fork\"."}], "task": "refering", "answer_template": "The \"a spoon in the snacks which is exactly under the fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 165, 166, 187, 188, 189, 209, 210, 211, 231, 232, 233], "bbox": [0.069765625, 0.4602872062663185, 0.269765625, 0.7737597911227154], "iscrowd": 0, "area": 4459.422050000001, "rle": {"size": [383, 640], "counts": "gSa04f;5K5N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O00000000001O00000000001O000000001O00000000001N10000000001O000000001O00000000001O00000000001O00000000001O000000001O00000000001O000000000100O001O001O010O001O00100O001O001O010O001O001O10O01O001O0010O01O0]Od0M3M2N3Mnk^5"}, "label": "a spoon in the snacks which is exactly under the fork"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of metal tongs submerged underneath the food\"."}], "task": "refering", "answer_template": "The \"part of metal tongs submerged underneath the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 165, 166, 187, 188, 189, 209, 210, 211, 231, 232, 233], "bbox": [0.069765625, 0.4602872062663185, 0.269765625, 0.7737597911227154], "iscrowd": 0, "area": 4459.422050000001, "rle": {"size": [383, 640], "counts": "gSa04f;5K5N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O00000000001O00000000001O000000001O00000000001N10000000001O000000001O00000000001O00000000001O00000000001O000000001O00000000001O000000000100O001O001O010O001O00100O001O001O010O001O001O10O01O001O0010O01O0]Od0M3M2N3Mnk^5"}, "label": "part of metal tongs submerged underneath the food"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large bowl of seafood and vegetables\"."}], "task": "refering", "answer_template": "The \"a large bowl of seafood and vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.188265625, 0.1438120104438642, 0.937328125, 0.9887728459530026], "iscrowd": 0, "area": 119842.23185000003, "rle": {"size": [383, 640], "counts": "nQ]13b;g0YOg0XOi0]Ob0B?A3M2N2N2O1N2N3M2N2N2N2N2N2N2O2M2N2N2N2N2N2N2N2O2M2N2N2N2N2N2N2O2M2N2O1N2N2N2O1N2N3N1N2N2O1N2N2N2O2M2N2O1N2N2O1N2N2N3N0O2N2O0O2N2N101N2N2O0O2N2O0O2N2N101N2N2O0O2N2N101N2N101O1O1O0O2O1O001O1O0O2O1O001O1O1N101O1O001O1N101O1O1O001N2O001O00000O2O0000001O0O10001O00000O2O00001O000O101O0000001N1000001O000O2O0000001O0O10001O00000O2O0000001O0O101O0000001N1000001O000O101O00001N1000001O000O101O0000001N10000000000O10000000000O10000000000O10000000000O100000O1000O10000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O00000000000O2O0000000O10001O0O1000000O101O000O1000001N100000000O101O000O1000000O2O00000O1000001N1000001O0O2O001O0O2O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O001N101N1O2O0O101N1O2O0O2N101N1O2O0O2N101N1O2O0O2N100O2O0O2N101N1O2O0O2N101N1O2O0O2N100O2O0O2N101N1O2N1N3N2M2N3N2M2N3N1N3M3N1N3N1N3L4K4M4L3M4K5L3M4K4M4L4K4M4L4K4M4L3M4K5L3M5J6K5K4K6K5K5L4L3N3L4L4M3L3M4L4M3L3M4L4M3L4L3N3L4L4L4MWi>"}, "label": "a large bowl of seafood and vegetables"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange bowl with a shrimp dish in it\"."}], "task": "refering", "answer_template": "The \"an orange bowl with a shrimp dish in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.188265625, 0.1438120104438642, 0.937328125, 0.9887728459530026], "iscrowd": 0, "area": 119842.23185000003, "rle": {"size": [383, 640], "counts": "nQ]13b;g0YOg0XOi0]Ob0B?A3M2N2N2O1N2N3M2N2N2N2N2N2N2O2M2N2N2N2N2N2N2N2O2M2N2N2N2N2N2N2O2M2N2O1N2N2N2O1N2N3N1N2N2O1N2N2N2O2M2N2O1N2N2O1N2N2N3N0O2N2O0O2N2N101N2N2O0O2N2O0O2N2N101N2N2O0O2N2N101N2N101O1O1O0O2O1O001O1O0O2O1O001O1O1N101O1O001O1N101O1O1O001N2O001O00000O2O0000001O0O10001O00000O2O00001O000O101O0000001N1000001O000O2O0000001O0O10001O00000O2O0000001O0O101O0000001N1000001O000O101O00001N1000001O000O101O0000001N10000000000O10000000000O10000000000O10000000000O100000O1000O10000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O00000000000O2O0000000O10001O0O1000000O101O000O1000001N100000000O101O000O1000000O2O00000O1000001N1000001O0O2O001O0O2O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O001N101N1O2O0O101N1O2O0O2N101N1O2O0O2N101N1O2O0O2N100O2O0O2N101N1O2O0O2N101N1O2O0O2N100O2O0O2N101N1O2N1N3N2M2N3N2M2N3N1N3M3N1N3N1N3L4K4M4L3M4K5L3M4K4M4L4K4M4L4K4M4L3M4K5L3M5J6K5K4K6K5K5L4L3N3L4L4M3L3M4L4M3L3M4L4M3L4L3N3L4L4L4MWi>"}, "label": "an orange bowl with a shrimp dish in it"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver tongs on left side of bowl\"."}], "task": "refering", "answer_template": "The \"silver tongs on left side of bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 97, 98, 99, 100, 119, 120, 121, 122, 123, 141, 142, 143, 144, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 254, 255], "bbox": [0.00621875, 0.2543342036553525, 0.39509374999999997, 0.8100783289817232], "iscrowd": 0, "area": 16383.328200000002, "rle": {"size": [383, 640], "counts": "gh13h;5J6J5K6J6M2O2N2N1O2O1N1O2N2N1O2O1N1O2N2N1O2O0O2N2N1O2O1N1O2N1O1O100O1O1O1O2N1O1O2N101N1O1O2N1O1O1O1O1O1O10O00000000000O10000O1O2O0O100O1O1000000000001O000000000001O0000010O0000001O000000001O01O00000001O00000000jNkFDT9<lFDU9;lFDT9<mFCS9=mFCS9=nFBR9>nFCQ9<QGCo8=QGCo8=QGCo8=RGBn8>RGBo8=QGDn8<RGDn8<SGCm8=SGCm8=SGCm8=TGCk8=UGCk8=UGCk8=VGBj8=WGCi8=WGCj8<VGEi8;XGDh8<XGDh8<XGDh8<YGCg8=YGCg8=YGDf8<[GCe8=[GCe8=[GCe8=[GCf8<[GDd8<\\GDd8;]GFb8:_GEa8;_GF`8:`GF`8:aGE_8;aGF^8:bGF^8:bGG^88cGG]89cGG]89cGH\\88eGG[89eGHZ88fGHZ88gGGY88hGIW87iGIW87iGJV86kGIV86jGKU85kGKU85lGJT86lGSO_O6d8g0mGROG0\\8n0nGoNL0U8Q1oGiN44m7T1nHjNS7V1mHjNS7U1oHiNQ7X1oHhNQ7X1k1O0100O01000O10000O10O0101O0O10000O10000O2O0O10000O1000001N100000000O2O000000000O2O0000000O11O0000001O0000001O0000001O00000O2O0O1O2O0O3M4L3K6J6K4K5K6J5JTd`4"}, "label": "silver tongs on left side of bowl"}]}
{"id": 434894, "image": "COCO_train2014_000000434894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of serving tongs in a bowl of food\"."}], "task": "refering", "answer_template": "The \"a pair of serving tongs in a bowl of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 97, 98, 99, 100, 119, 120, 121, 122, 123, 141, 142, 143, 144, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 254, 255], "bbox": [0.00621875, 0.2543342036553525, 0.39509374999999997, 0.8100783289817232], "iscrowd": 0, "area": 16383.328200000002, "rle": {"size": [383, 640], "counts": "gh13h;5J6J5K6J6M2O2N2N1O2O1N1O2N2N1O2O1N1O2N2N1O2O0O2N2N1O2O1N1O2N1O1O100O1O1O1O2N1O1O2N101N1O1O2N1O1O1O1O1O1O10O00000000000O10000O1O2O0O100O1O1000000000001O000000000001O0000010O0000001O000000001O01O00000001O00000000jNkFDT9<lFDU9;lFDT9<mFCS9=mFCS9=nFBR9>nFCQ9<QGCo8=QGCo8=QGCo8=RGBn8>RGBo8=QGDn8<RGDn8<SGCm8=SGCm8=SGCm8=TGCk8=UGCk8=UGCk8=VGBj8=WGCi8=WGCj8<VGEi8;XGDh8<XGDh8<XGDh8<YGCg8=YGCg8=YGDf8<[GCe8=[GCe8=[GCe8=[GCf8<[GDd8<\\GDd8;]GFb8:_GEa8;_GF`8:`GF`8:aGE_8;aGF^8:bGF^8:bGG^88cGG]89cGG]89cGH\\88eGG[89eGHZ88fGHZ88gGGY88hGIW87iGIW87iGJV86kGIV86jGKU85kGKU85lGJT86lGSO_O6d8g0mGROG0\\8n0nGoNL0U8Q1oGiN44m7T1nHjNS7V1mHjNS7U1oHiNQ7X1oHhNQ7X1k1O0100O01000O10000O10O0101O0O10000O10000O2O0O10000O1000001N100000000O2O000000000O2O0000000O11O0000001O0000001O0000001O00000O2O0O1O2O0O3M4L3K6J6K4K5K6J5JTd`4"}, "label": "a pair of serving tongs in a bowl of food"}]}
{"id": 74461, "image": "COCO_train2014_000000074461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a very long and green park bench\"."}], "task": "refering", "answer_template": "The \"this is a very long and green park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 225, 226, 227, 228, 229, 230, 231, 232, 233, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 270, 271, 272, 273, 274, 275, 276, 277, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.0, 0.501125, 0.6129976580796253, 0.9887656249999999], "iscrowd": 0, "area": 59743.92989999999, "rle": {"size": [640, 427], "counts": "U=[2ea0V4iK10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10001N100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O101N100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O2O0O100O1O100O1O100O100O1O100O100O1O100OO1L5nMTElJo:S5TEhJP;V5TEdJo:[5TE`Jo:_5TE[JQ;c5SEWJP;h5SESJP;l5SEoIQ;P6REjIQ;W6PEcIU;]6lD^IW;c6R1100O100O1O3N2M3M3N2M3N2M3M3N2M3N:HN2N1O2N2N2N4M2M4L2cM\\CZMd<d2iCQMW<n2UDRLYOCb<Y4^DjKUOK]<Z4`DjKUOJ[<Z4cDkKSOJZ<[4dDjKTOJW<\\4fDiKTOJV<]4gDhKUOIT<_4hDgKVOHR<a4iDgKUOGR<g3^ClL\\1EUOGQ<f3aCmLZ1EUOGP<e3eCmLW1FUOGo;d3hCnLU1FTOHo;d3jClLT1GTOIm;d3mCjLS1Hi;^3UCiLg>W3[AgLh>h1d@aN02e0Dn>a1_@hNO2f0CS?Z1[@POM2f0CV?W1X@ROO2e0CX?U1U@UOO2e0C[?R1R@XOO2f0B\\?=l_OO3N13f0A^?8m_O4OO13Xa0Bo^O9J013`b0Kl0O1N2O1O1NaZW3"}, "label": "this is a very long and green park bench"}]}
{"id": 74461, "image": "COCO_train2014_000000074461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"row of benches\"."}], "task": "refering", "answer_template": "The \"row of benches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 225, 226, 227, 228, 229, 230, 231, 232, 233, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 270, 271, 272, 273, 274, 275, 276, 277, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.0, 0.501125, 0.6129976580796253, 0.9887656249999999], "iscrowd": 0, "area": 59743.92989999999, "rle": {"size": [640, 427], "counts": "U=[2ea0V4iK10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10001N100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O101N100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O2O0O100O1O100O1O100O100O1O100O100O1O100OO1L5nMTElJo:S5TEhJP;V5TEdJo:[5TE`Jo:_5TE[JQ;c5SEWJP;h5SESJP;l5SEoIQ;P6REjIQ;W6PEcIU;]6lD^IW;c6R1100O100O1O3N2M3M3N2M3N2M3M3N2M3N:HN2N1O2N2N2N4M2M4L2cM\\CZMd<d2iCQMW<n2UDRLYOCb<Y4^DjKUOK]<Z4`DjKUOJ[<Z4cDkKSOJZ<[4dDjKTOJW<\\4fDiKTOJV<]4gDhKUOIT<_4hDgKVOHR<a4iDgKUOGR<g3^ClL\\1EUOGQ<f3aCmLZ1EUOGP<e3eCmLW1FUOGo;d3hCnLU1FTOHo;d3jClLT1GTOIm;d3mCjLS1Hi;^3UCiLg>W3[AgLh>h1d@aN02e0Dn>a1_@hNO2f0CS?Z1[@POM2f0CV?W1X@ROO2e0CX?U1U@UOO2e0C[?R1R@XOO2f0B\\?=l_OO3N13f0A^?8m_O4OO13Xa0Bo^O9J013`b0Kl0O1N2O1O1NaZW3"}, "label": "row of benches"}]}
{"id": 735, "image": "COCO_train2014_000000000735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white table cloth under the table with the blue top\"."}], "task": "refering", "answer_template": "The \"the white table cloth under the table with the blue top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 204, 225, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.670296875, 0.6143490304709142, 0.9965937500000001, 0.9865373961218836], "iscrowd": 0, "area": 12017.701200000001, "rle": {"size": [361, 640], "counts": "g^g41T;5O0001O0001O000000001O000001O01O000000001O00001O0001O00000000000010O000001O00001O0000001O0000001O0001O01O0000010O0000010O0001O00001O01O01O00001O00001O00010O1O1O1O1O1O010O1O1O1O1O100O2N1O00010O00001O0_F^O\\8a0eGDU8=lGGn79RHMh73WH4c7M[H:`7E`Ha0Z7@dHg0V7YO[H\\1_7eN]Hd1^7[NbHh1[7YNcHi1\\7WNdHi1\\7XNbHi1_7VNaHj1_7WN_Hj1a7VN_Hj1a7VN^Hk1b7VN]Hj1c7VN\\Hk1e7TN[Hl1e7UNYHl1g7TNYHl1g7k000001O000000000000000000000000000000000000000000000000001O001O001O001O000000001O00000000001O000000001O0000000000001O000000000000001O000000000000001O000000000000001O000000c>"}, "label": "the white table cloth under the table with the blue top"}]}
{"id": 735, "image": "COCO_train2014_000000000735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white table cloth\"."}], "task": "refering", "answer_template": "The \"a white table cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 204, 225, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.670296875, 0.6143490304709142, 0.9965937500000001, 0.9865373961218836], "iscrowd": 0, "area": 12017.701200000001, "rle": {"size": [361, 640], "counts": "g^g41T;5O0001O0001O000000001O000001O01O000000001O00001O0001O00000000000010O000001O00001O0000001O0000001O0001O01O0000010O0000010O0001O00001O01O01O00001O00001O00010O1O1O1O1O1O010O1O1O1O1O100O2N1O00010O00001O0_F^O\\8a0eGDU8=lGGn79RHMh73WH4c7M[H:`7E`Ha0Z7@dHg0V7YO[H\\1_7eN]Hd1^7[NbHh1[7YNcHi1\\7WNdHi1\\7XNbHi1_7VNaHj1_7WN_Hj1a7VN_Hj1a7VN^Hk1b7VN]Hj1c7VN\\Hk1e7TN[Hl1e7UNYHl1g7TNYHl1g7k000001O000000000000000000000000000000000000000000000000001O001O001O001O000000001O00000000001O000000001O0000000000001O000000000000001O000000000000001O000000000000001O000000c>"}, "label": "a white table cloth"}]}
{"id": 735, "image": "COCO_train2014_000000000735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a really big cake with candle on it\"."}], "task": "refering", "answer_template": "The \"a really big cake with candle on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247], "bbox": [0.48435937500000004, 0.5848476454293629, 0.8123906250000001, 0.857617728531856], "iscrowd": 0, "area": 11346.827599999997, "rle": {"size": [361, 640], "counts": "Qb]32T;3N2M3N3L3N2M4L3K5I8O00001N10001O000O2O00000O2O00001O0O101O00001N10001O000O2O00001O0O10001O000O2O00001O0O101O00001N1000O1N1O2N2001O0O100000000O2O0000000O10001O000XOeNaG\\1^8h000`0@5LO00000000000O2O1O1O001O1O1N101O1O001O000O1000O10000O10O010000O10000O01000O10000O2O000O101O0O100O2O000O101O0O101O0O10001N10000O2O000O101O0O100O2O000O101O0O10001N10000O2O000O101O0O100O2O000O101O0O101O0O10001N10000O2O000O101O0O101N2O0EdEI]:LnE2f:O1O1N2O1Oa\\Z1"}, "label": "a really big cake with candle on it"}]}
{"id": 735, "image": "COCO_train2014_000000000735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cake with a candle\"."}], "task": "refering", "answer_template": "The \"the cake with a candle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247], "bbox": [0.48435937500000004, 0.5848476454293629, 0.8123906250000001, 0.857617728531856], "iscrowd": 0, "area": 11346.827599999997, "rle": {"size": [361, 640], "counts": "Qb]32T;3N2M3N3L3N2M4L3K5I8O00001N10001O000O2O00000O2O00001O0O101O00001N10001O000O2O00001O0O10001O000O2O00001O0O101O00001N1000O1N1O2N2001O0O100000000O2O0000000O10001O000XOeNaG\\1^8h000`0@5LO00000000000O2O1O1O001O1O1N101O1O001O000O1000O10000O10O010000O10000O01000O10000O2O000O101O0O100O2O000O101O0O101O0O10001N10000O2O000O101O0O100O2O000O101O0O10001N10000O2O000O101O0O100O2O000O101O0O101O0O10001N10000O2O000O101O0O101N2O0EdEI]:LnE2f:O1O1N2O1Oa\\Z1"}, "label": "the cake with a candle"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green chair\"."}], "task": "refering", "answer_template": "The \"green chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 167, 190, 213, 214, 215, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 357, 358, 359], "bbox": [0.27154687499999997, 0.3135833333333334, 0.62715625, 0.9256458333333335], "iscrowd": 0, "area": 20944.0657, "rle": {"size": [480, 640], "counts": "kia2<b>3M3]Mc0_E@`:b2M2N3Mg1YN2N101N2N2O0O10QMhF5X9HRG2n8JWG4i8I\\G7b8FcG:\\8BhG>W8_OnGb0P8mNbHS1\\7jNiHV1V7fNoH[1o6bNTI_1k6_NXIa1f6^N\\Id1b6YNbIg1]6WNeIj1Z6TNhIn1V6PNmIP2R6oMoIR2P6mMRJT2l5jMWJV2h5iMYJX2f5gM\\JZ2b5dMaJ\\2^5cMdJ]2[5bMfJ`2X5^MkJb2T5]MnJc2Q5\\MPKf2n4XMUKh2j4WMWKj2h4VMXKZ2_LQNo;o1a001O00000000001O00000001O01O01O00001O01O01O01O10O100O2O1N6J0001O0001OO2O0O3N7H2O1N100O2O000000000000O100O1_DnMS:S2_ElM^O7o:m1cEmM[O8Q;k1eEnMWO9S;j1fEnMTO9W;i1eE]N[:c1dE^N\\:b1dE^N[:c1eE\\N\\:d1dE\\N\\:d1dE\\NW:i1hEXNW:i1jEVNV:j1jEUNV:l1jETNV:k1kEUNU:k1kETNV:l1jERNX:n1hEQNY:o1fEPN\\:o1eERNZ:n1eEVNX:j1hEXNV:h1iEZNU:f1lE[NS:e1lE]NS:c1mE^NR:b1mE`NR:`1mEaNS:^1nEcNQ:]1nEeNQ:[1nEgNQ:Y1nEiNQ:W1oEiNP:W1PFkNo9U1PFlNP:T1oEnNP:R1PFnNP:R1oEPOP:o0PFROP:n0PFSOo9m0PFTOP:k0PFWOo9h0RFXOn9h0QFZOo9d0QF]Oo9c0QF^Oo9`0QFBn9>RFBo9<QFFo99PFHR:5oELR:2mE0T:MlE4V:JjE7V:HiE:Y:BhE?\\:\\OdEd0a:WO_Ej0d:QO\\EQ1h:jNXEW1i:gNWEZ1k:bNVE^1m:_NREc1o:[NQEe1R;WNnDk1R;TNmDm1U;QNjDP2W;nMhDU2Y;hMfDZ2[;dMdD^2^;_McDa2^;^MbDc2l;10001O0O10000O2N100O1O101O01O0001O0001O01O01O01O01O0010O01O0010O01O00100O001O0lMiC_1X<^NkC`1W<^NkCa1U<]NmCb1U<[NnCd1S<ZNnCe1T<ZNmCd1T<[NnCa1V<]NkCa1X<^NiC^1[<`NfC]1Q=M2M4QOcB6`=GgB1^=Lje_3"}, "label": "green chair"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green lawn chair\"."}], "task": "refering", "answer_template": "The \"a green lawn chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 167, 190, 213, 214, 215, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 357, 358, 359], "bbox": [0.27154687499999997, 0.3135833333333334, 0.62715625, 0.9256458333333335], "iscrowd": 0, "area": 20944.0657, "rle": {"size": [480, 640], "counts": "kia2<b>3M3]Mc0_E@`:b2M2N3Mg1YN2N101N2N2O0O10QMhF5X9HRG2n8JWG4i8I\\G7b8FcG:\\8BhG>W8_OnGb0P8mNbHS1\\7jNiHV1V7fNoH[1o6bNTI_1k6_NXIa1f6^N\\Id1b6YNbIg1]6WNeIj1Z6TNhIn1V6PNmIP2R6oMoIR2P6mMRJT2l5jMWJV2h5iMYJX2f5gM\\JZ2b5dMaJ\\2^5cMdJ]2[5bMfJ`2X5^MkJb2T5]MnJc2Q5\\MPKf2n4XMUKh2j4WMWKj2h4VMXKZ2_LQNo;o1a001O00000000001O00000001O01O01O00001O01O01O01O10O100O2O1N6J0001O0001OO2O0O3N7H2O1N100O2O000000000000O100O1_DnMS:S2_ElM^O7o:m1cEmM[O8Q;k1eEnMWO9S;j1fEnMTO9W;i1eE]N[:c1dE^N\\:b1dE^N[:c1eE\\N\\:d1dE\\N\\:d1dE\\NW:i1hEXNW:i1jEVNV:j1jEUNV:l1jETNV:k1kEUNU:k1kETNV:l1jERNX:n1hEQNY:o1fEPN\\:o1eERNZ:n1eEVNX:j1hEXNV:h1iEZNU:f1lE[NS:e1lE]NS:c1mE^NR:b1mE`NR:`1mEaNS:^1nEcNQ:]1nEeNQ:[1nEgNQ:Y1nEiNQ:W1oEiNP:W1PFkNo9U1PFlNP:T1oEnNP:R1PFnNP:R1oEPOP:o0PFROP:n0PFSOo9m0PFTOP:k0PFWOo9h0RFXOn9h0QFZOo9d0QF]Oo9c0QF^Oo9`0QFBn9>RFBo9<QFFo99PFHR:5oELR:2mE0T:MlE4V:JjE7V:HiE:Y:BhE?\\:\\OdEd0a:WO_Ej0d:QO\\EQ1h:jNXEW1i:gNWEZ1k:bNVE^1m:_NREc1o:[NQEe1R;WNnDk1R;TNmDm1U;QNjDP2W;nMhDU2Y;hMfDZ2[;dMdD^2^;_McDa2^;^MbDc2l;10001O0O10000O2N100O1O101O01O0001O0001O01O01O01O01O0010O01O0010O01O00100O001O0lMiC_1X<^NkC`1W<^NkCa1U<]NmCb1U<[NnCd1S<ZNnCe1T<ZNmCd1T<[NnCa1V<]NkCa1X<^NiC^1[<`NfC]1Q=M2M4QOcB6`=GgB1^=Lje_3"}, "label": "a green lawn chair"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the outdoor chair the woman is sitting on\"."}], "task": "refering", "answer_template": "The \"the outdoor chair the woman is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 253, 254, 276, 277, 300, 323, 324, 347, 352, 353, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0032187500000000003, 0.40358333333333335, 0.37821875, 0.9872708333333333], "iscrowd": 0, "area": 8290.155850000003, "rle": {"size": [480, 640], "counts": "]V16c>:F;G9L3M4L4L3M4L4L3M4L4L2N3M2N3F9M4L3N3L3M4M2M4M2M4M3L3N3LeNREhNk:W1ZEiNc:T1cEkNZ:T1lEkNQ:R1UFmNh9R1]FnN`9o0jFlNS9k0eGjNW8W1lGhNR8W1RHiNj7W1YHiNd7X1_HhN]7X1gHgNV7Y1mHhNP7V1UIjNg6U1^IjN_6U1eIlNW6S1nIlNo5R1WJnNf5P1^JQO^5n0gJQOV5o0nJQOn4o0VKQOe4P1^KPO_4P1eKPOW4P1mKPOo3Q1SLoNj3Q1ZLoNb3Q1bLROV3o0mLVOj2k0ZMUOa2l0cMSOY2m0lMQOQ2n0TNROh1m0^NRO^1n0gNPOW1n0nNROn0m0WOROg0m0]OROc0i0CVO=f0GZO9b0L^O3>1BO:6EJ6;JE0b0N_OKh05XODP1;k6O2O0O2O0O101N1O1N2O00001O0000001O000O101O0000001O00O1000O10O1000000O100000000O1000000000000O100000000000000O10000000001O0O1000000000000O1000000000000O100000000O1000000O1000000O10000O2O00000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O101O0O1000000O10000O1000O10O1000O10O1000O01000O0100000O2O000O1O2N1N2O2N2N3K4IUTj5"}, "label": "the outdoor chair the woman is sitting on"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the woman is sitting in in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the chair the woman is sitting in in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 253, 254, 276, 277, 300, 323, 324, 347, 352, 353, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0032187500000000003, 0.40358333333333335, 0.37821875, 0.9872708333333333], "iscrowd": 0, "area": 8290.155850000003, "rle": {"size": [480, 640], "counts": "]V16c>:F;G9L3M4L4L3M4L4L3M4L4L2N3M2N3F9M4L3N3L3M4M2M4M2M4M3L3N3LeNREhNk:W1ZEiNc:T1cEkNZ:T1lEkNQ:R1UFmNh9R1]FnN`9o0jFlNS9k0eGjNW8W1lGhNR8W1RHiNj7W1YHiNd7X1_HhN]7X1gHgNV7Y1mHhNP7V1UIjNg6U1^IjN_6U1eIlNW6S1nIlNo5R1WJnNf5P1^JQO^5n0gJQOV5o0nJQOn4o0VKQOe4P1^KPO_4P1eKPOW4P1mKPOo3Q1SLoNj3Q1ZLoNb3Q1bLROV3o0mLVOj2k0ZMUOa2l0cMSOY2m0lMQOQ2n0TNROh1m0^NRO^1n0gNPOW1n0nNROn0m0WOROg0m0]OROc0i0CVO=f0GZO9b0L^O3>1BO:6EJ6;JE0b0N_OKh05XODP1;k6O2O0O2O0O101N1O1N2O00001O0000001O000O101O0000001O00O1000O10O1000000O100000000O1000000000000O100000000000000O10000000001O0O1000000000000O1000000000000O100000000O1000000O1000000O10000O2O00000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O101O0O1000000O10000O1000O10O1000O10O1000O01000O0100000O2O000O1O2N1N2O2N2N3K4IUTj5"}, "label": "the chair the woman is sitting in in the right hand picture"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"nobody is sitting in the red chair\"."}], "task": "refering", "answer_template": "The \"nobody is sitting in the red chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.788703125, 0.5591458333333333, 1.0, 0.98925], "iscrowd": 0, "area": 23229.555150000015, "rle": {"size": [480, 640], "counts": "TR]7V3W;c0\\Od0]Oc0\\Od0F:001O0O100I7YOg0YOg0A?3M3M3M3M3M4L3M3M3M3M3dF^KT8e4bGdK\\8_4[GjKc8X4TGQLi8U5N3M1O000000000000000000000000001O01O000000000O1O1O1O1O1O1O1O1O1O100O1lNT1M3M4M2O1001O00000000001O00000000001O0000003Nk0TO000000010O000001O00001O0000001O0000001O0000001O00001O0000001O0000001N1000001O00001O00000O2OZ1fNc2\\MWC"}, "label": "nobody is sitting in the red chair"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red camp chair\"."}], "task": "refering", "answer_template": "The \"red camp chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.788703125, 0.5591458333333333, 1.0, 0.98925], "iscrowd": 0, "area": 23229.555150000015, "rle": {"size": [480, 640], "counts": "TR]7V3W;c0\\Od0]Oc0\\Od0F:001O0O100I7YOg0YOg0A?3M3M3M3M3M4L3M3M3M3M3dF^KT8e4bGdK\\8_4[GjKc8X4TGQLi8U5N3M1O000000000000000000000000001O01O000000000O1O1O1O1O1O1O1O1O1O100O1lNT1M3M4M2O1001O00000000001O00000000001O0000003Nk0TO000000010O000001O00001O0000001O0000001O0000001O00001O0000001O0000001N1000001O00001O00000O2OZ1fNc2\\MWC"}, "label": "red camp chair"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman on an outside lawn chair next to a man on a laptop computer\"."}], "task": "refering", "answer_template": "The \"a young woman on an outside lawn chair next to a man on a laptop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 365, 366, 370, 371, 372, 373, 374, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.0558125, 0.3813958333333333, 0.9436093750000001, 0.988375], "iscrowd": 0, "area": 55617.332050000005, "rle": {"size": [480, 640], "counts": "^Va0<`>b0^O<D7I7I7I7J6J6K5J6J6J6J6K5J6J6J6K5J3M5K5K5K5K4L5L4K5K4L5K5K5K5J6J7J6I7I8H7J5J6L3N3L4M3M2N3M3M1O1O1O1O1O001O1O2N1O1O1O1O100000000000000000000000000000000000O100O10000O100O100O100000000000000O10001O000000000000000000001O0O10000O10000O10001N2O2N1N3N1N3N2N1N3N1O2M2O2N1N3jJnHZ2S7`MSI_2o6ZMWId2k6VM[Ii2g6PM`In2b6kLdIS3^6fLiIY3Y6^LoIa3R6WLWJf3l5QL\\Jn3e5kKbJT4`5cKhJ[4Q8N1N2O1O1N3N1O2M2O2N2M3N1O2M3N2M2O2N2M3N2N2M4M2N2M4M2M4MN2I6J7K5000O101O000O101O000O101O000O101O000001O01OlDSMU:n2iESMW:m2hETMX:m2fETM[:k2dEVM\\:j2bEXM^:i2`EXM`:h2^EZMc:e2\\E\\Md:e2YE]Mg:c2XE^Mh:b2VE`Mk:`2SEaMm:_2QEcMo:^2oDcMQ;]2mDeMU;Y2jDhMY;;gD_1\\;ZNhDf1\\;QNjDn1Y;jMlDV2W;cMnD\\2k;01O1O001O001O001O001O001O001O001O0000001O000000001O00000000001O000000001O000000001O00000000001O000000001O001O1O001N2O1O001O1O001O1O001O1O001O1O001O1O001O1O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O000000001O000000001O000000001O0000001O000000001O000000001O0000000010O000001O000000001O000000001O000000001O0000001O000000001O00000000000000000000000000000000000000000001O0001O0000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000001O0000001O00010O0000001O0000001O00001O0000001O0000001O0000001O0000O10000O100O100O100O100O10000O100O100O100O100O100M3N2M3M3N2M3M3M3M3M3M3M3L4O100001O00000000001O000000001O00000000001O00000000001O00000000001O00000oNgC[OY<<Z1GW_`0"}, "label": "a young woman on an outside lawn chair next to a man on a laptop computer"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black tank top sitting in a blue folding chair holding a can\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black tank top sitting in a blue folding chair holding a can\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 365, 366, 370, 371, 372, 373, 374, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.0558125, 0.3813958333333333, 0.9436093750000001, 0.988375], "iscrowd": 0, "area": 55617.332050000005, "rle": {"size": [480, 640], "counts": "^Va0<`>b0^O<D7I7I7I7J6J6K5J6J6J6J6K5J6J6J6K5J3M5K5K5K5K4L5L4K5K4L5K5K5K5J6J7J6I7I8H7J5J6L3N3L4M3M2N3M3M1O1O1O1O1O001O1O2N1O1O1O1O100000000000000000000000000000000000O100O10000O100O100O100000000000000O10001O000000000000000000001O0O10000O10000O10001N2O2N1N3N1N3N2N1N3N1O2M2O2N1N3jJnHZ2S7`MSI_2o6ZMWId2k6VM[Ii2g6PM`In2b6kLdIS3^6fLiIY3Y6^LoIa3R6WLWJf3l5QL\\Jn3e5kKbJT4`5cKhJ[4Q8N1N2O1O1N3N1O2M2O2N2M3N1O2M3N2M2O2N2M3N2N2M4M2N2M4M2M4MN2I6J7K5000O101O000O101O000O101O000O101O000001O01OlDSMU:n2iESMW:m2hETMX:m2fETM[:k2dEVM\\:j2bEXM^:i2`EXM`:h2^EZMc:e2\\E\\Md:e2YE]Mg:c2XE^Mh:b2VE`Mk:`2SEaMm:_2QEcMo:^2oDcMQ;]2mDeMU;Y2jDhMY;;gD_1\\;ZNhDf1\\;QNjDn1Y;jMlDV2W;cMnD\\2k;01O1O001O001O001O001O001O001O001O0000001O000000001O00000000001O000000001O000000001O00000000001O000000001O001O1O001N2O1O001O1O001O1O001O1O001O1O001O1O001O1O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O000000001O000000001O000000001O0000001O000000001O000000001O0000000010O000001O000000001O000000001O000000001O0000001O000000001O00000000000000000000000000000000000000000001O0001O0000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000001O0000001O00010O0000001O0000001O00001O0000001O0000001O0000001O0000O10000O100O100O100O100O10000O100O100O100O100O100M3N2M3M3N2M3M3M3M3M3M3M3L4O100001O00000000001O000000001O00000000001O00000000001O00000000001O00000oNgC[OY<<Z1GW_`0"}, "label": "a woman wearing a black tank top sitting in a blue folding chair holding a can"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man working on his computer\"."}], "task": "refering", "answer_template": "The \"a man working on his computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 245, 246, 247, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 313, 314, 315, 316, 336, 337, 338, 339, 359, 360, 361, 362], "bbox": [0.293265625, 0.11910416666666666, 0.7870781250000001, 0.9438125], "iscrowd": 0, "area": 49068.34925, "rle": {"size": [480, 640], "counts": "QZh2e0U><H7J7I7I5J5L4L4L4L4K6K1O1O1N2O1O1O1O1O1mERMS8o2gGWMX8k2aG\\M]8e2]GaMb8`2XGgMf8Z2TGmMk8S2TGnMk8S2UGnMi8S2WGmMh8T2YGlMe8U2[GkMd8V2\\GkMb8V2^GkM`8W2_GiMa8W2_GjM_8X2`GhM]7iNhHa3JgM[7oNeH[30fMY7VOaHV36dMW7\\O]HQ3<dMT7@\\Hn2?bMS7D[Hl2a0aMQ7H[Hg2d0aMo6LZHe2f0`Mn6j3RIVLn6k3QIVLn6j3SIVLl6k3SIULm6l3RIULm6k3SIULl6m3SITLl6l3TITLk6n3TISLj6o3UIRLj6n3WIQLh6Q4WIPLh6P4XIPLg6R4XIoKf6R4ZInKe6T4ZImKc6V4\\IkKa6W4_IiK_6Z4aIfK\\6\\4dIdKY6^4hIcKS6`4nIaKn5a4SJ_Ki5d4XJ]Kd5e4]J[K`5g4aJZK]5f4eJYKZ5f4hJ[KW5d4jJ\\KV5c4kJ^KS5c4mJ]KS5c4mJ]KS5b4nJ_KP5b4PK^KP5b4QK]Ko4c4QK^Km4c4SK]Km4c4SK^Kg4VN_J\\6j0^Kg4WN]J\\6l0]Kg4WN]J\\6l0^Kg4VN\\J\\6m0^Kg4e4ZKZKf4f4ZK[Ke4e4[K[Ke4e4[K\\Kd4d4\\K\\Kd4d4\\K\\Kd4d4\\K]Kc4b4^K^Kc4a4^K^Kb4g4YKZKf4k4UKUKk4P5PKPKP5Q5oJoJQ5Q5oJnJR5R5nJnJR5R5nJnJR5R5oJlJR5T5nJlJS5S5mJmJS5S5mJlJT5S5mJmJS5S5nJlJR5T5nJkJS5U5mJkJS5U5mJkJS5U5nJiJS5W5mJiJT5U5mJkJT5T5mJjJU5U5kJkJV5T5jJlJV5T5jJkJX5T5iJkJX5T5hJlJY5R5hJmJZ5R5gJmJZ5R5fJnJ[5Q5eJnJ]5Q5cJoJ^5P5cJoJ^5P5bJoJ`5o4aJQKa5m4_JSKb5l4_JRKd5l4\\JTKe5a0cIY2h0VMg52PJf2:VMh51SJf25YMi5NVJf22\\Mj5IYJi2M]Ml5G\\Ji2I_MV6[OUJT3E`Mc6nNlI`3AbMY8[2hGdMZ8Z2fGeM]8Y2cGgM_8W2aGiMa8U2`GiMc8U2]GkMd8T2\\GlMe8S2[GlMg8S2YGmMi8Q2WGoMj8P2VGoMl8P2TGPNm8o1TGPNm8o1SGRNn8l1RGTNo8k1QGUNP9j1PGWNP9h1PGXNQ9g1oFYNS9e1mF\\NS9c1mF]NT9b1mF]NU9a1kF`NW9]1iFcNY9Z1hFfNZ9X1fFiN[9T1fFlN\\9Q1eFoN[93`EEV19X90eEGS19W9OhEGS19T9NlEHQ1;Q9LPFHP1<o8KSFHo0=m8LTFGo0>k8LVFEP1?i8MWFCQ1`0h8MWFBR1b0e8LZFBQ1b0e8LZFBR1a0c8N[FAR1b0b8M\\F@S1c0`8N^F^OR1d0`8N^F^OR1e0`8L^F_OR1e0`8L^F^OS1f0`8K]F_OS1g0_8I_F@R1g0_8I_F@S1f0_8I^F@T1h0]8H_F@T1h0]8H_F@T1h0^8G^FAT1i0]8F_F@U1j0]8D_FBT1j0]8D_FBT1k0\\8BaFCT1j0\\8BcF@R1n0[8AjF[Ok0U1Z8_OiHa0X7^OhHb0X7]OiHd0W7ZOjHf0W7YOiHg0W7XOiHj0W7UOiHk0W7TOjHl0W7ROiHP1V7POjHP1W7nNjHS1V7jNkHW1U7fNnHZ1S7aNQI`1V:100O2nD_Nc8a1ZFCe9=WFGh9;RFJn96nEOQ:1kE3T:OgE5Y:KbE;]:E_E?`:C[Ea0e:n1O0O10001N100O2O000O2O000O2OO10O0100O010O010O10ON3L4M2M4M3L3N3L4M2M4M2M4M3L3N]E[MQ9a2`FRN`9k1PFhNo9U301O0O10000O2O0O10001N10000O101O0O100010O00001O01O01L3J6K6I6J6K6I8I6I7I7J6I7I7J6I7I7J6I7I7J6I7I7J6I7I^^o1"}, "label": "a man working on his computer"}]}
{"id": 377570, "image": "COCO_train2014_000000377570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on laptop\"."}], "task": "refering", "answer_template": "The \"man on laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 245, 246, 247, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 313, 314, 315, 316, 336, 337, 338, 339, 359, 360, 361, 362], "bbox": [0.293265625, 0.11910416666666666, 0.7870781250000001, 0.9438125], "iscrowd": 0, "area": 49068.34925, "rle": {"size": [480, 640], "counts": "QZh2e0U><H7J7I7I5J5L4L4L4L4K6K1O1O1N2O1O1O1O1O1mERMS8o2gGWMX8k2aG\\M]8e2]GaMb8`2XGgMf8Z2TGmMk8S2TGnMk8S2UGnMi8S2WGmMh8T2YGlMe8U2[GkMd8V2\\GkMb8V2^GkM`8W2_GiMa8W2_GjM_8X2`GhM]7iNhHa3JgM[7oNeH[30fMY7VOaHV36dMW7\\O]HQ3<dMT7@\\Hn2?bMS7D[Hl2a0aMQ7H[Hg2d0aMo6LZHe2f0`Mn6j3RIVLn6k3QIVLn6j3SIVLl6k3SIULm6l3RIULm6k3SIULl6m3SITLl6l3TITLk6n3TISLj6o3UIRLj6n3WIQLh6Q4WIPLh6P4XIPLg6R4XIoKf6R4ZInKe6T4ZImKc6V4\\IkKa6W4_IiK_6Z4aIfK\\6\\4dIdKY6^4hIcKS6`4nIaKn5a4SJ_Ki5d4XJ]Kd5e4]J[K`5g4aJZK]5f4eJYKZ5f4hJ[KW5d4jJ\\KV5c4kJ^KS5c4mJ]KS5c4mJ]KS5b4nJ_KP5b4PK^KP5b4QK]Ko4c4QK^Km4c4SK]Km4c4SK^Kg4VN_J\\6j0^Kg4WN]J\\6l0]Kg4WN]J\\6l0^Kg4VN\\J\\6m0^Kg4e4ZKZKf4f4ZK[Ke4e4[K[Ke4e4[K\\Kd4d4\\K\\Kd4d4\\K\\Kd4d4\\K]Kc4b4^K^Kc4a4^K^Kb4g4YKZKf4k4UKUKk4P5PKPKP5Q5oJoJQ5Q5oJnJR5R5nJnJR5R5nJnJR5R5oJlJR5T5nJlJS5S5mJmJS5S5mJlJT5S5mJmJS5S5nJlJR5T5nJkJS5U5mJkJS5U5mJkJS5U5nJiJS5W5mJiJT5U5mJkJT5T5mJjJU5U5kJkJV5T5jJlJV5T5jJkJX5T5iJkJX5T5hJlJY5R5hJmJZ5R5gJmJZ5R5fJnJ[5Q5eJnJ]5Q5cJoJ^5P5cJoJ^5P5bJoJ`5o4aJQKa5m4_JSKb5l4_JRKd5l4\\JTKe5a0cIY2h0VMg52PJf2:VMh51SJf25YMi5NVJf22\\Mj5IYJi2M]Ml5G\\Ji2I_MV6[OUJT3E`Mc6nNlI`3AbMY8[2hGdMZ8Z2fGeM]8Y2cGgM_8W2aGiMa8U2`GiMc8U2]GkMd8T2\\GlMe8S2[GlMg8S2YGmMi8Q2WGoMj8P2VGoMl8P2TGPNm8o1TGPNm8o1SGRNn8l1RGTNo8k1QGUNP9j1PGWNP9h1PGXNQ9g1oFYNS9e1mF\\NS9c1mF]NT9b1mF]NU9a1kF`NW9]1iFcNY9Z1hFfNZ9X1fFiN[9T1fFlN\\9Q1eFoN[93`EEV19X90eEGS19W9OhEGS19T9NlEHQ1;Q9LPFHP1<o8KSFHo0=m8LTFGo0>k8LVFEP1?i8MWFCQ1`0h8MWFBR1b0e8LZFBQ1b0e8LZFBR1a0c8N[FAR1b0b8M\\F@S1c0`8N^F^OR1d0`8N^F^OR1e0`8L^F_OR1e0`8L^F^OS1f0`8K]F_OS1g0_8I_F@R1g0_8I_F@S1f0_8I^F@T1h0]8H_F@T1h0]8H_F@T1h0^8G^FAT1i0]8F_F@U1j0]8D_FBT1j0]8D_FBT1k0\\8BaFCT1j0\\8BcF@R1n0[8AjF[Ok0U1Z8_OiHa0X7^OhHb0X7]OiHd0W7ZOjHf0W7YOiHg0W7XOiHj0W7UOiHk0W7TOjHl0W7ROiHP1V7POjHP1W7nNjHS1V7jNkHW1U7fNnHZ1S7aNQI`1V:100O2nD_Nc8a1ZFCe9=WFGh9;RFJn96nEOQ:1kE3T:OgE5Y:KbE;]:E_E?`:C[Ea0e:n1O0O10001N100O2O000O2O000O2OO10O0100O010O010O10ON3L4M2M4M3L3N3L4M2M4M2M4M3L3N]E[MQ9a2`FRN`9k1PFhNo9U301O0O10000O2O0O10001N10000O101O0O100010O00001O01O01L3J6K6I6J6K6I8I6I7I7J6I7I7J6I7I7J6I7I7J6I7I7J6I7I^^o1"}, "label": "man on laptop"}]}
{"id": 533218, "image": "COCO_train2014_000000533218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partially made bed with pillows on it\"."}], "task": "refering", "answer_template": "The \"a partially made bed with pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 264, 265, 280, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 329, 330, 331, 332, 333, 334, 353, 354, 355, 356], "bbox": [0.20340625, 0.5797353497164461, 0.5325468750000001, 0.8571455576559547], "iscrowd": 0, "area": 14614.851200000005, "rle": {"size": [529, 640], "counts": "maS21]`03L4bNM[B6d=<gAHX>`0^ACb>V11gNWAh0j>VOXAj0i>SOYAl0h>QO[An0f>PO\\Ao0f>mN`Ao0V?K00000000O10000000001O00000000000000000000000000001O000000000001O0000000000000001O00000000000O1O1O1O1O1O1O1O101N1O1O1O1O1O100O1O1O100O1O100O2N1O100O1O100O1O1O100O1O1O100O2N1N2O1N2O1N2O1N2O1N2O1N2O1N2O2M2N2O1N2O1N2O1N2N2O10000000001O000000000000000000000000001O0000001O0000001O0000001O00002N1O2N2N1O2N2N3M3M2N3M3M2NN3M2N2N2N3M2N2N2M4M2N2N2N3M2N2N2N3M2N2N3M2N2M3N2N1O2N1O2NcBiMV<c2jC`MU<d2eC^M[<]3O0013L5RMUCl1h=bN^1L4M;E;EeSj4"}, "label": "a partially made bed with pillows on it"}]}
{"id": 533218, "image": "COCO_train2014_000000533218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a turned down bed with pillow\"."}], "task": "refering", "answer_template": "The \"a turned down bed with pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 264, 265, 280, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 329, 330, 331, 332, 333, 334, 353, 354, 355, 356], "bbox": [0.20340625, 0.5797353497164461, 0.5325468750000001, 0.8571455576559547], "iscrowd": 0, "area": 14614.851200000005, "rle": {"size": [529, 640], "counts": "maS21]`03L4bNM[B6d=<gAHX>`0^ACb>V11gNWAh0j>VOXAj0i>SOYAl0h>QO[An0f>PO\\Ao0f>mN`Ao0V?K00000000O10000000001O00000000000000000000000000001O000000000001O0000000000000001O00000000000O1O1O1O1O1O1O1O101N1O1O1O1O1O100O1O1O100O1O100O2N1O100O1O100O1O1O100O1O1O100O2N1N2O1N2O1N2O1N2O1N2O1N2O1N2O2M2N2O1N2O1N2O1N2N2O10000000001O000000000000000000000000001O0000001O0000001O0000001O00002N1O2N2N1O2N2N3M3M2N3M3M2NN3M2N2N2N3M2N2N2M4M2N2N2N3M2N2N2N3M2N2N3M2N2M3N2N1O2N1O2NcBiMV<c2jC`MU<d2eC^M[<]3O0013L5RMUCl1h=bN^1L4M;E;EeSj4"}, "label": "a turned down bed with pillow"}]}
{"id": 41700, "image": "COCO_train2014_000000041700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in coveralls\"."}], "task": "refering", "answer_template": "The \"woman in coveralls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 110, 111, 112, 132, 133, 134, 135, 156, 157, 158, 179, 180, 181, 202, 203, 204, 227], "bbox": [0.775375, 0.30112727272727274, 0.9047499999999999, 0.9146181818181819], "iscrowd": 0, "area": 5768.316200000001, "rle": {"size": [275, 640], "counts": "a[U42_88H3N2M3NO000O2O0O2O01O100O010O1K50O10O10O1000TK]Oh1d0UNAi1`0TNCk1>RNEm1=nMHo1;mMHR2:XLZO?`0X38aK2l0Nb3X1XLlNf3X1ULjNh3\\1TLeNi3a1RLbN^3mN]Lg21]Nb3R2YLPNf3R3[LlKd3S4`LlK_3R4eLjK\\3V4`0O2N10000100010O10O01000OTLnKX3g41N1O100O103L`N\\KUO5cNo3V2SL^On3`0XL[Oh3ROPNi0P2WOUNd0l1[OXNa0h1_O\\N=d1C`N9a1FcN6]1JfN3\\1KgN1[1NhNNZ11gNLW18jNFo0b0QO]OP1b0RO[OP1e0POZOQ1f0POXOQ1g0^30001N4LjX`0"}, "label": "woman in coveralls"}]}
{"id": 41700, "image": "COCO_train2014_000000041700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing overalls\"."}], "task": "refering", "answer_template": "The \"a person wearing overalls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 110, 111, 112, 132, 133, 134, 135, 156, 157, 158, 179, 180, 181, 202, 203, 204, 227], "bbox": [0.775375, 0.30112727272727274, 0.9047499999999999, 0.9146181818181819], "iscrowd": 0, "area": 5768.316200000001, "rle": {"size": [275, 640], "counts": "a[U42_88H3N2M3NO000O2O0O2O01O100O010O1K50O10O10O1000TK]Oh1d0UNAi1`0TNCk1>RNEm1=nMHo1;mMHR2:XLZO?`0X38aK2l0Nb3X1XLlNf3X1ULjNh3\\1TLeNi3a1RLbN^3mN]Lg21]Nb3R2YLPNf3R3[LlKd3S4`LlK_3R4eLjK\\3V4`0O2N10000100010O10O01000OTLnKX3g41N1O100O103L`N\\KUO5cNo3V2SL^On3`0XL[Oh3ROPNi0P2WOUNd0l1[OXNa0h1_O\\N=d1C`N9a1FcN6]1JfN3\\1KgN1[1NhNNZ11gNLW18jNFo0b0QO]OP1b0RO[OP1e0POZOQ1f0POXOQ1g0^30001N4LjX`0"}, "label": "a person wearing overalls"}]}
{"id": 41700, "image": "COCO_train2014_000000041700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding the cow wearing all white\"."}], "task": "refering", "answer_template": "The \"the woman holding the cow wearing all white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 101, 102, 124, 125, 126, 147, 148, 170, 171, 193, 194], "bbox": [0.397, 0.3447272727272727, 0.508515625, 0.8480727272727272], "iscrowd": 0, "area": 4048.9543999999973, "rle": {"size": [275, 640], "counts": "V[T25W88N2N2NK60O00000010O0010O0oG1`7N^H<]1[O[4]1YKiNc4]1WKhNe4Y1ZKiNc4Y1[KjNb4X1]KjN`4Y1^KhNm3HaKb1`0iNl3J_K_1d0kNi3j1ULXNh3j1VLWNj3j1PL[NP4g1fKaNZ4l2O100000UOgK\\MY4d2iKZMV4g2lKWMT4i2nKUMR4j2QLTMo3l2SLRMm3m2VLQMj3n2[LnLe3Q3]LnLc3R3^LmLb3R3`LlLa3KTL4E^2g0aMY4]2iKaMY4]2hKbMZ4]2n0O1dNSJ0o5LWJ0k5K]J0e51\\JJS6KoIMZ60hII_65P10000000O10O100O100O100O1O1O100O1OU^d2"}, "label": "the woman holding the cow wearing all white"}]}
{"id": 41700, "image": "COCO_train2014_000000041700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman dressed in white leading a black and white cow by a rope down the street\"."}], "task": "refering", "answer_template": "The \"a woman dressed in white leading a black and white cow by a rope down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 101, 102, 124, 125, 126, 147, 148, 170, 171, 193, 194], "bbox": [0.397, 0.3447272727272727, 0.508515625, 0.8480727272727272], "iscrowd": 0, "area": 4048.9543999999973, "rle": {"size": [275, 640], "counts": "V[T25W88N2N2NK60O00000010O0010O0oG1`7N^H<]1[O[4]1YKiNc4]1WKhNe4Y1ZKiNc4Y1[KjNb4X1]KjN`4Y1^KhNm3HaKb1`0iNl3J_K_1d0kNi3j1ULXNh3j1VLWNj3j1PL[NP4g1fKaNZ4l2O100000UOgK\\MY4d2iKZMV4g2lKWMT4i2nKUMR4j2QLTMo3l2SLRMm3m2VLQMj3n2[LnLe3Q3]LnLc3R3^LmLb3R3`LlLa3KTL4E^2g0aMY4]2iKaMY4]2hKbMZ4]2n0O1dNSJ0o5LWJ0k5K]J0e51\\JJS6KoIMZ60hII_65P10000000O10O100O100O100O1O1O100O1OU^d2"}, "label": "a woman dressed in white leading a black and white cow by a rope down the street"}]}
{"id": 344805, "image": "COCO_train2014_000000344805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on a laptop sitting on a couch\"."}], "task": "refering", "answer_template": "The \"a woman on a laptop sitting on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 223, 224, 225, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 311, 312, 335], "bbox": [0.525765625, 0.477125, 0.83128125, 0.8387291666666667], "iscrowd": 0, "area": 11362.986799999988, "rle": {"size": [480, 640], "counts": "dkm41n>8G2O1O1O1O1N100O1O100OO1I7MN7L3O1L4N3PCZOd;g0YD^Oe;b0YDAe;a0VDEh;;SDKl;5lC4S<MhC9U<HhC<V<Y1N101N101N0001O001O1O1O1nNeC]O]<b0eC\\O\\<c0fC[O[<e0fCYO[<f0hCWOX<i0lCmN[<S1iCeN[<Z1f0O1O010O1001O000O10000000000O100000001O000O100000000O101O00000O1000000O100O10000O1000000O1000000000000O100000000000000O100000O100O10O010O001O0O2N1O\\OoB\\OQ=a0`CnNb<P1g0002O0O2O1N2N2N101O1N2N2M2O2gB`No<c1QCeNe<]1[CgN_<[1aCeN]<Q2M]OfCdNn;]OYDR2IaNm;_OWDR2L_Nk;`2O1O2M1000O101N20000_OVDlMj;T2VDkMk;U2UDkMk;U2UDkMk;U2UDkMk;U2UDkMk;U2UDkMk;U2UDlMj;T2VDlMj;T2WDkMi;U2WDkMi;U2WDkMj;T2VDlMj;T2WDlMh;T2XDlMh;T2XDlMh;T2XDlMj;R2WDnMi;Q2WDPNj;n1VDRNm;l1SDTNn;j1SDUNS<e1mC\\Nf<P1[CoNi<m0WCTOj<k0VCTOl<_12O1N2N2N2O1N5K5K3N2M4L4L5K6I9GhYb1"}, "label": "a woman on a laptop sitting on a couch"}]}
{"id": 344805, "image": "COCO_train2014_000000344805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on couch\"."}], "task": "refering", "answer_template": "The \"woman on couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 223, 224, 225, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 311, 312, 335], "bbox": [0.525765625, 0.477125, 0.83128125, 0.8387291666666667], "iscrowd": 0, "area": 11362.986799999988, "rle": {"size": [480, 640], "counts": "dkm41n>8G2O1O1O1O1N100O1O100OO1I7MN7L3O1L4N3PCZOd;g0YD^Oe;b0YDAe;a0VDEh;;SDKl;5lC4S<MhC9U<HhC<V<Y1N101N101N0001O001O1O1O1nNeC]O]<b0eC\\O\\<c0fC[O[<e0fCYO[<f0hCWOX<i0lCmN[<S1iCeN[<Z1f0O1O010O1001O000O10000000000O100000001O000O100000000O101O00000O1000000O100O10000O1000000O1000000000000O100000000000000O100000O100O10O010O001O0O2N1O\\OoB\\OQ=a0`CnNb<P1g0002O0O2O1N2N2N101O1N2N2M2O2gB`No<c1QCeNe<]1[CgN_<[1aCeN]<Q2M]OfCdNn;]OYDR2IaNm;_OWDR2L_Nk;`2O1O2M1000O101N20000_OVDlMj;T2VDkMk;U2UDkMk;U2UDkMk;U2UDkMk;U2UDkMk;U2UDkMk;U2UDlMj;T2VDlMj;T2WDkMi;U2WDkMi;U2WDkMj;T2VDlMj;T2WDlMh;T2XDlMh;T2XDlMh;T2XDlMj;R2WDnMi;Q2WDPNj;n1VDRNm;l1SDTNn;j1SDUNS<e1mC\\Nf<P1[CoNi<m0WCTOj<k0VCTOl<_12O1N2N2N2O1N5K5K3N2M4L4L5K6I9GhYb1"}, "label": "woman on couch"}]}
{"id": 344805, "image": "COCO_train2014_000000344805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on couch holding laptop\"."}], "task": "refering", "answer_template": "The \"man on couch holding laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 196, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 286, 287], "bbox": [0.43020312499999996, 0.4451041666666667, 0.7204843750000001, 0.7243333333333334], "iscrowd": 0, "area": 8049.651549999999, "rle": {"size": [480, 640], "counts": "[fQ42m>2N1O2O1N100O1O1O4M4L0gNCoC>n;HnC8P<?[CAc<b0\\C_Oa<d0^C]O_<f0`CZO_<h0`CXO`<j0^CUOc<c10100000O10000012M1O0001O0000001N2O1M2L5M3O1N2O2LOImBdNT=]152O0O2O1O003N1N3N1N3N1N3N1N2O2O0O2O0O1000000001O000O101O100O2\\OWCnNi<P1YCPOg<o0[CoNf<o0\\CQOe<m0]CQOd<m0^CROc<m0^CPOf<m0\\CQOf<n0[CoNh<P1YCnNj<o0e0O2M2O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2M2OA2=2M102N01N2O0002M7I2O1N2N1O03N3M]OAhB<X=DkB9T=InB4R=LQC1o<0RCNm<3o01O2M201N101[ODcB=[=GcB:\\=GcB9]=HbB9]=GcB9X=3aBNY=9eBGW=T1N1O1N3O0O2O00001O0O2O001O002N4L3M4L4L4LO2M2O2O1O1\\NhC=Y<AiC?X<^OkC`0W<]OlCb0V<ZOmCe0V<UOnCj0U=O001O1L3O2N2M201N4L7I_dc2"}, "label": "man on couch holding laptop"}]}
{"id": 344805, "image": "COCO_train2014_000000344805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing sunglasses staring at a computer screen\"."}], "task": "refering", "answer_template": "The \"a man wearing sunglasses staring at a computer screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 196, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 286, 287], "bbox": [0.43020312499999996, 0.4451041666666667, 0.7204843750000001, 0.7243333333333334], "iscrowd": 0, "area": 8049.651549999999, "rle": {"size": [480, 640], "counts": "[fQ42m>2N1O2O1N100O1O1O4M4L0gNCoC>n;HnC8P<?[CAc<b0\\C_Oa<d0^C]O_<f0`CZO_<h0`CXO`<j0^CUOc<c10100000O10000012M1O0001O0000001N2O1M2L5M3O1N2O2LOImBdNT=]152O0O2O1O003N1N3N1N3N1N3N1N2O2O0O2O0O1000000001O000O101O100O2\\OWCnNi<P1YCPOg<o0[CoNf<o0\\CQOe<m0]CQOd<m0^CROc<m0^CPOf<m0\\CQOf<n0[CoNh<P1YCnNj<o0e0O2M2O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2M2OA2=2M102N01N2O0002M7I2O1N2N1O03N3M]OAhB<X=DkB9T=InB4R=LQC1o<0RCNm<3o01O2M201N101[ODcB=[=GcB:\\=GcB9]=HbB9]=GcB9X=3aBNY=9eBGW=T1N1O1N3O0O2O00001O0O2O001O002N4L3M4L4L4LO2M2O2O1O1\\NhC=Y<AiC?X<^OkC`0W<]OlCb0V<ZOmCe0V<UOnCj0U=O001O1L3O2N2M201N4L7I_dc2"}, "label": "a man wearing sunglasses staring at a computer screen"}]}
{"id": 180968, "image": "COCO_train2014_000000180968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man has big rubber boots on\"."}], "task": "refering", "answer_template": "The \"the man has big rubber boots on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 162, 163, 185, 186, 187, 208, 209, 210, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 324, 325, 347, 348, 349], "bbox": [0.06340625, 0.38426356589147287, 0.19929687499999998, 0.8606782945736434], "iscrowd": 0, "area": 11153.593149999999, "rle": {"size": [516, 640], "counts": "knd06j?5dMh0fC^OW<f0cC^O[<e0`C@]<d0]CA`<c0ZCAd<c0RCFk<T3eN8H8H8I7I7I6J7D<B>B=D7O000001N10000O100O100O100ZN[GlKf8T4^GhKc8W4bGdKa8Y4cGbKb8Z4QHSK^8^4d101O001O001O001O0O2g1XN10O100kJnFV3S9eLVGW3j8iLYGT3g8kLjGd2W8[MoG`2R8^MUH[2P8`MVH[2m7aMXH[2l7`MZHZ2k7`M[H\\2g7aMbHV2_7iMgIQ1[6mNiIo0X6POmIk0T6SOQJi0Q6UOSJg0n5WOUJg0e:O1O2N1O1O1N2O1O1N2O0O2O001N101O2M3N_RR8"}, "label": "the man has big rubber boots on"}]}
{"id": 180968, "image": "COCO_train2014_000000180968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man leading elephant\"."}], "task": "refering", "answer_template": "The \"man leading elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 162, 163, 185, 186, 187, 208, 209, 210, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 324, 325, 347, 348, 349], "bbox": [0.06340625, 0.38426356589147287, 0.19929687499999998, 0.8606782945736434], "iscrowd": 0, "area": 11153.593149999999, "rle": {"size": [516, 640], "counts": "knd06j?5dMh0fC^OW<f0cC^O[<e0`C@]<d0]CA`<c0ZCAd<c0RCFk<T3eN8H8H8I7I7I6J7D<B>B=D7O000001N10000O100O100O100ZN[GlKf8T4^GhKc8W4bGdKa8Y4cGbKb8Z4QHSK^8^4d101O001O001O001O0O2g1XN10O100kJnFV3S9eLVGW3j8iLYGT3g8kLjGd2W8[MoG`2R8^MUH[2P8`MVH[2m7aMXH[2l7`MZHZ2k7`M[H\\2g7aMbHV2_7iMgIQ1[6mNiIo0X6POmIk0T6SOQJi0Q6UOSJg0n5WOUJg0e:O1O2N1O1O1N2O1O1N2O0O2O001N101O2M3N_RR8"}, "label": "man leading elephant"}]}
{"id": 426728, "image": "COCO_train2014_000000426728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large black motorcycle beneath a man in a helmet\"."}], "task": "refering", "answer_template": "The \"a large black motorcycle beneath a man in a helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 334, 335, 336], "bbox": [0.39159375, 0.3287058823529411, 0.729578125, 0.9811764705882353], "iscrowd": 0, "area": 41297.840800000005, "rle": {"size": [425, 640], "counts": "f^X3P1Q<:H8I7I6O1O1O1O1O001O1N2O1O002N2cMiMdIY2U6PNgIR2V6RNgIP2W6SNeIP2Y6TNcIm1[6UNdIm1Z6TNeIm1Z6TNeIn1X6TNgIn1W6SNhIn1W6SNhIo1U6SNiIo1V6RNiIP2U6RNiIo1V6VNeIl1X6ZNcIh1[6]N`Id1_6_N^Ia1a6aNdIY1\\6hNmIn0S6SOnIl0Q6TOnIm0Q6UOmIl0S6UOkIl0U6UOiIm0U6UOiIl0W6TOhIm0X6TOfIm0Y6UOeIm0Y6UOfIk0Y6WOfIi0Y6XOfIi0Y6YOfIh0X6ZOgIf0Y6[OfIe0W6_OgIb0V6AjI`0R6DmI<o5IPJ8l5LSJ5k5MSJ5k5LUJ5j5LUJ6i5KVJ6i5KUJ7i5KVJ6i5g3N1N3N2N2N1N101O0O1000000O10001N1000000O0100O010O10O01O100O010O1O010O100O0102N2M4M2N3M2M4M2N1O2N1O2N1O1O10O01O00001O001O00000001O0000000000000000001N1O1O2O0O1O2N1O1K6J5K5L5L3M4L3M3M4N1O101N101O2M3N1N3N2M2O2N1N3N1N2O0O10001N1cLmHd0S7[OQIa0P7^OSI`0n6_OUI=l6BVI=j6CVI=j6BWI=j6CVI=k6AVI>k6AWI>i6BWI=j6BWI>i6BWI=j6BWI>i6AYI=h6CXI<i6CXI;k6DUI:m6ETI:m6DVI9l6FUI8m6FUI9l6FXI5j6J\\IOf6O`ILa63dIG_67gIB[6<kI_OV6`0oIZOS6d0TJTOo5k0VJnNo5o0VJjNo5R1XJfNm5W1T3M4H8J6I7J5J7IkcW2"}, "label": "a large black motorcycle beneath a man in a helmet"}]}
{"id": 426728, "image": "COCO_train2014_000000426728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large black motorcycle being driven by a man in a helmet\"."}], "task": "refering", "answer_template": "The \"a large black motorcycle being driven by a man in a helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 334, 335, 336], "bbox": [0.39159375, 0.3287058823529411, 0.729578125, 0.9811764705882353], "iscrowd": 0, "area": 41297.840800000005, "rle": {"size": [425, 640], "counts": "f^X3P1Q<:H8I7I6O1O1O1O1O001O1N2O1O002N2cMiMdIY2U6PNgIR2V6RNgIP2W6SNeIP2Y6TNcIm1[6UNdIm1Z6TNeIm1Z6TNeIn1X6TNgIn1W6SNhIn1W6SNhIo1U6SNiIo1V6RNiIP2U6RNiIo1V6VNeIl1X6ZNcIh1[6]N`Id1_6_N^Ia1a6aNdIY1\\6hNmIn0S6SOnIl0Q6TOnIm0Q6UOmIl0S6UOkIl0U6UOiIm0U6UOiIl0W6TOhIm0X6TOfIm0Y6UOeIm0Y6UOfIk0Y6WOfIi0Y6XOfIi0Y6YOfIh0X6ZOgIf0Y6[OfIe0W6_OgIb0V6AjI`0R6DmI<o5IPJ8l5LSJ5k5MSJ5k5LUJ5j5LUJ6i5KVJ6i5KUJ7i5KVJ6i5g3N1N3N2N2N1N101O0O1000000O10001N1000000O0100O010O10O01O100O010O1O010O100O0102N2M4M2N3M2M4M2N1O2N1O2N1O1O10O01O00001O001O00000001O0000000000000000001N1O1O2O0O1O2N1O1K6J5K5L5L3M4L3M3M4N1O101N101O2M3N1N3N2M2O2N1N3N1N2O0O10001N1cLmHd0S7[OQIa0P7^OSI`0n6_OUI=l6BVI=j6CVI=j6BWI=j6CVI=k6AVI>k6AWI>i6BWI=j6BWI>i6BWI=j6BWI>i6AYI=h6CXI<i6CXI;k6DUI:m6ETI:m6DVI9l6FUI8m6FUI9l6FXI5j6J\\IOf6O`ILa63dIG_67gIB[6<kI_OV6`0oIZOS6d0TJTOo5k0VJnNo5o0VJjNo5R1XJfNm5W1T3M4H8J6I7J5J7IkcW2"}, "label": "a large black motorcycle being driven by a man in a helmet"}]}
{"id": 426728, "image": "COCO_train2014_000000426728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"santa claus riding motorcycle\"."}], "task": "refering", "answer_template": "The \"santa claus riding motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216], "bbox": [0.1480625, 0.1943294117647059, 0.40675, 0.6690117647058824], "iscrowd": 0, "area": 24354.1399, "rle": {"size": [425, 640], "counts": "RdW19m<>B=D=B>B=D<C3M3N2M3M3QOcMPG_2P9n0O000O10000000000O10000000001N10000eN[10000O1000000O01N2O1O001O100000O1000000O1000O10O1000000O1000000O1000000O100000000O01002N8G7J00000O1000000O1000000001O2K4M3L4M3L4M3L5N1O1N2O1O1O1O2N1O1O1O1O1O1O2N1O1O1O0000001O000000001O000O3N3N2O1O1O1O00N2OO001N101O000O2O001N101O001lLSGi1n8PNZGo1g8iMaGV2_8cMhG\\2Z8[MoGd2\\9O1O1N2O5K5I7G9G9G9G9H8G9Gf_m4"}, "label": "santa claus riding motorcycle"}]}
{"id": 426728, "image": "COCO_train2014_000000426728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and black motorcycle with a santa riding it\"."}], "task": "refering", "answer_template": "The \"a red and black motorcycle with a santa riding it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216], "bbox": [0.1480625, 0.1943294117647059, 0.40675, 0.6690117647058824], "iscrowd": 0, "area": 24354.1399, "rle": {"size": [425, 640], "counts": "RdW19m<>B=D=B>B=D<C3M3N2M3M3QOcMPG_2P9n0O000O10000000000O10000000001N10000eN[10000O1000000O01N2O1O001O100000O1000000O1000O10O1000000O1000000O1000000O100000000O01002N8G7J00000O1000000O1000000001O2K4M3L4M3L4M3L5N1O1N2O1O1O1O2N1O1O1O1O1O1O2N1O1O1O0000001O000000001O000O3N3N2O1O1O1O00N2OO001N101O000O2O001N101O001lLSGi1n8PNZGo1g8iMaGV2_8cMhG\\2Z8[MoGd2\\9O1O1N2O5K5I7G9G9G9G9H8G9Gf_m4"}, "label": "a red and black motorcycle with a santa riding it"}]}
{"id": 492268, "image": "COCO_train2014_000000492268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"inside of canoe\"."}], "task": "refering", "answer_template": "The \"inside of canoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 330, 331, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 376, 377], "bbox": [0.312078125, 0.5125208333333333, 1.0, 0.9733750000000001], "iscrowd": 0, "area": 34377.5433, "rle": {"size": [480, 640], "counts": "a_Q31o>0om00QRO000o>0QA00000O10O1000O1000O10O100000O01000000O10O1000O1000O10O100000O01^Mb2O100O10O0100O1000O0100O100O010O10000O10O0100O100O01000O100O10O010000O10O0100O100O01000O100O10O0100O1000O0100O100O010O10000O10O010000O10O1000000000O0100000000O010000000O10O10000000O1000O1000O100000O10O1000000000O0100000000O010000000O10O10000000O1000O01O1O1O100O001O1O1O001O00010O001O001O00001O010O00001O001O010O01O10O0100O0010O0100O010000O010O10O10O10O01000O010000O010O10O10O10O10O100O01000O01000O010O10O10O1000O0100O01000O010O10O10O1000O0100O01000O01000O10O01000O01000O0100O01000O10O10O10O01000O0100O10O10O10O10O10O01000O010000O010O10O10O10O10O10O010000O01000O010O10O10O1000O0100O01000O010O10O10O1000O0100O01000O01000O10O01000O01000O0100O01000O10O10O10O01000O0100O01000O01000O10O01000O01000O0100O01000O10O10O10O01000O01000O0100O01000O10O01000O01000O0100O01000O10O10O10O01000O010g0YOQ1nNcF"}, "label": "inside of canoe"}]}
{"id": 492268, "image": "COCO_train2014_000000492268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pears\"."}], "task": "refering", "answer_template": "The \"pears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 330, 331, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 376, 377], "bbox": [0.312078125, 0.5125208333333333, 1.0, 0.9733750000000001], "iscrowd": 0, "area": 34377.5433, "rle": {"size": [480, 640], "counts": "a_Q31o>0om00QRO000o>0QA00000O10O1000O1000O10O100000O01000000O10O1000O1000O10O100000O01^Mb2O100O10O0100O1000O0100O100O010O10000O10O0100O100O01000O100O10O010000O10O0100O100O01000O100O10O0100O1000O0100O100O010O10000O10O010000O10O1000000000O0100000000O010000000O10O10000000O1000O1000O100000O10O1000000000O0100000000O010000000O10O10000000O1000O01O1O1O100O001O1O1O001O00010O001O001O00001O010O00001O001O010O01O10O0100O0010O0100O010000O010O10O10O10O01000O010000O010O10O10O10O10O100O01000O01000O010O10O10O1000O0100O01000O010O10O10O1000O0100O01000O01000O10O01000O01000O0100O01000O10O10O10O01000O0100O10O10O10O10O10O01000O010000O010O10O10O10O10O10O010000O01000O010O10O10O1000O0100O01000O010O10O10O1000O0100O01000O01000O10O01000O01000O0100O01000O10O10O10O01000O0100O01000O01000O10O01000O01000O0100O01000O10O10O10O01000O01000O0100O01000O10O01000O01000O0100O01000O10O10O10O01000O010g0YOQ1nNcF"}, "label": "pears"}]}
{"id": 492268, "image": "COCO_train2014_000000492268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat full of pitchers and fruit\"."}], "task": "refering", "answer_template": "The \"a boat full of pitchers and fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [274, 275, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 355, 356, 357, 358, 359, 360, 361, 362, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.338765625, 0.6629166666666666, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 16547.867399999996, "rle": {"size": [480, 640], "counts": "cUV33m>1N10001O0O10000O1000000O1000000O1000000O1000000O1000000O1000000O2O00000O1000000O1000000O1000000O1000000O10000O1000000O10001O0O1000000O1000000O1000000O1000000O1000000O1000000O1000001N1000O10O100000O010000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O10O1000O10O1000O1000O10O100000O010000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O0100000O01000O001O1O0O2O1O0O2O1O1N101O10O01000O01000O0100O01000O01000O010O10O10O100O01000O01000O010O10O10O10O10O10O01000O0100O01000O10O10O10O01000O0100O01000O01000O010O10O10O10O10O100O01000O010O10O10O10O10O10O01000O01000O0100O01000O10O01000O01000O0100O01000O010O10O10O10O10O10O010000O01000O010O10O10O10O01000O01000O0100O01000O0100O010O010O10O010O0100O010O10O010O0100O0010O0100O010O10O010O0100O010O10O010O01O10O010O0100O010O10O010O0100O010O10O01O010O10O010O0100O010O10O010O0100O010O00100O0UOnF"}, "label": "a boat full of pitchers and fruit"}]}
{"id": 492268, "image": "COCO_train2014_000000492268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat under the greenery\"."}], "task": "refering", "answer_template": "The \"boat under the greenery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [274, 275, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 355, 356, 357, 358, 359, 360, 361, 362, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.338765625, 0.6629166666666666, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 16547.867399999996, "rle": {"size": [480, 640], "counts": "cUV33m>1N10001O0O10000O1000000O1000000O1000000O1000000O1000000O1000000O2O00000O1000000O1000000O1000000O1000000O10000O1000000O10001O0O1000000O1000000O1000000O1000000O1000000O1000000O1000001N1000O10O100000O010000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O10O1000O10O1000O1000O10O100000O010000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O0100000O01000O001O1O0O2O1O0O2O1O1N101O10O01000O01000O0100O01000O01000O010O10O10O100O01000O01000O010O10O10O10O10O10O01000O0100O01000O10O10O10O01000O0100O01000O01000O010O10O10O10O10O100O01000O010O10O10O10O10O10O01000O01000O0100O01000O10O01000O01000O0100O01000O010O10O10O10O10O10O010000O01000O010O10O10O10O01000O01000O0100O01000O0100O010O010O10O010O0100O010O10O010O0100O0010O0100O010O10O010O0100O010O10O010O01O10O010O0100O010O10O010O0100O010O10O01O010O10O010O0100O010O10O010O0100O010O00100O0UOnF"}, "label": "boat under the greenery"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vinagarette dressing in a plastic container\"."}], "task": "refering", "answer_template": "The \"vinagarette dressing in a plastic container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 122], "bbox": [0.207765625, 0.0529375, 0.4324375, 0.30066666666666664], "iscrowd": 0, "area": 13427.594050000002, "rle": {"size": [480, 640], "counts": "Z]n13j>8H7I7I8H7I8H7I4N3N1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N3M2N101N1O2O0O1O2N101N10001N101O001O000O2O001O00001N101O001O00001N101O00001O0O1001O0000000000000000001O000000000000000O101O0000000000000O10001O0O101O001N2O001O0O2O001O1N101O001N101O1O0O2O001O1N1O2M2O2N2M4M2N2M4M2N2N3L3N2N3L3N2M4I6J7I6J6JPSZ5"}, "label": "vinagarette dressing in a plastic container"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plastic cup of red salad dressing\"."}], "task": "refering", "answer_template": "The \"plastic cup of red salad dressing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 122], "bbox": [0.207765625, 0.0529375, 0.4324375, 0.30066666666666664], "iscrowd": 0, "area": 13427.594050000002, "rle": {"size": [480, 640], "counts": "Z]n13j>8H7I7I8H7I8H7I4N3N1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N3M2N101N1O2O0O1O2N101N10001N101O001O000O2O001O00001N101O001O00001N101O00001O0O1001O0000000000000000001O000000000000000O101O0000000000000O10001O0O101O001N2O001O0O2O001O1N101O001N101O1O0O2O001O1N1O2M2O2N2M4M2N2M4M2N2N3L3N2N3L3N2M4I6J7I6J6JPSZ5"}, "label": "plastic cup of red salad dressing"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white sauce in plastic bowl\"."}], "task": "refering", "answer_template": "The \"white sauce in plastic bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 382, 383, 384, 385, 386, 387], "bbox": [0.519828125, 0.59, 0.907125, 1.0], "iscrowd": 0, "area": 36814.31295000001, "rle": {"size": [480, 640], "counts": "Q^l47g>8H7I8H7I7I3M3M4L3M4L3O1N3M2N3M2N2N3M2O1N2O1N2O2M2O1N2O1O2N1O2O0O2N2N1O2N1O2O1N1O2N1O2N2N1O2N1N3N2N1O2M2O2M3N1N3N1N3N2M2M4L4L3DSK^FS5_9801N2O1O0O2O1N2O1O0O100O10000O100O10000O10000000000O100000000000000O1000000000000O1000O10000000O1000000000000O1000000000000O100000000000000001O000000000000000000000000000000000000001O0000000000000000001O0000000O2O00000O2O000O10001N10001O0O2O000O2O001N101O000N3M2O2M2N2O2M2M4M2N2N3M2M3N3M2N3L3N2N3M2M3N3M2N3L3N4L3M3M3L4M3N2N2M3N2N2N2M3N2N3L3N2N2N2M4K5K4L5K5K5K5K5Ke0[OmYk0"}, "label": "white sauce in plastic bowl"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small container of potatoe salad\"."}], "task": "refering", "answer_template": "The \"a small container of potatoe salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 382, 383, 384, 385, 386, 387], "bbox": [0.519828125, 0.59, 0.907125, 1.0], "iscrowd": 0, "area": 36814.31295000001, "rle": {"size": [480, 640], "counts": "Q^l47g>8H7I8H7I7I3M3M4L3M4L3O1N3M2N3M2N2N3M2O1N2O1N2O2M2O1N2O1O2N1O2O0O2N2N1O2N1O2O1N1O2N1O2N2N1O2N1N3N2N1O2M2O2M3N1N3N1N3N2M2M4L4L3DSK^FS5_9801N2O1O0O2O1N2O1O0O100O10000O100O10000O10000000000O100000000000000O1000000000000O1000O10000000O1000000000000O1000000000000O100000000000000001O000000000000000000000000000000000000001O0000000000000000001O0000000O2O00000O2O000O10001N10001O0O2O000O2O001N101O000N3M2O2M2N2O2M2M4M2N2N3M2M3N3M2N3L3N2N3M2M3N3M2N3L3N4L3M3M3L4M3N2N2M3N2N2N2M3N2N3L3N2N2N2M4K5K4L5K5K5K5K5Ke0[OmYk0"}, "label": "a small container of potatoe salad"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup with some pink sause\"."}], "task": "refering", "answer_template": "The \"a cup with some pink sause\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 121, 122, 123], "bbox": [0.20978124999999997, 0.04645833333333334, 0.428296875, 0.31064583333333334], "iscrowd": 0, "area": 12883.166200000001, "rle": {"size": [480, 640], "counts": "cln11l>5K5K5K4L5K5K5K4L5K5K4L5K5K5N1N3N2M3N1O2M3N1N3N2N2M2O2M010O2O1O0O2O1N2O001N2O0O2O1N2O001N2O0O2O1O1N101N2O0O2O1O0O100O1001N1000000000000O10001O0000000O10000000001O0O1000000000000O101O0000001O1N2O1O1O1O1O001N2O1O1O1N2N2N1O2N2N2N2N2N2M3N2N1O2N2N2N2N2N3M3M4K4M3M4L3M3L5J5K6J5K5K6J5K[_[5"}, "label": "a cup with some pink sause"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of red dressing\"."}], "task": "refering", "answer_template": "The \"a cup of red dressing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 121, 122, 123], "bbox": [0.20978124999999997, 0.04645833333333334, 0.428296875, 0.31064583333333334], "iscrowd": 0, "area": 12883.166200000001, "rle": {"size": [480, 640], "counts": "cln11l>5K5K5K4L5K5K5K4L5K5K4L5K5K5N1N3N2M3N1O2M3N1N3N2N2M2O2M010O2O1O0O2O1N2O001N2O0O2O1N2O001N2O0O2O1O1N101N2O0O2O1O0O100O1001N1000000000000O10001O0000000O10000000001O0O1000000000000O101O0000001O1N2O1O1O1O1O001N2O1O1O1N2N2N1O2N2N2N2N2N2M3N2N1O2N2N2N2N2N3M3M4K4M3M4L3M3L5J5K6J5K5K6J5K[_[5"}, "label": "a cup of red dressing"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white ranch dressing in a cup\"."}], "task": "refering", "answer_template": "The \"the white ranch dressing in a cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.5253125, 0.5952083333333333, 0.9079062499999999, 0.9997083333333333], "iscrowd": 0, "area": 38026.836449999995, "rle": {"size": [480, 640], "counts": "_jm47g>j0WOg0XO4M2M4M2M3N3L3N3L3N3L3N3M2N3N1N2O2M2O2M2O2M2O2M2O1N3N1N3N1N3N1N3N1N3N1N3N1N3N2N1O2M2O2N1O2M2O2N1O2M2O2N1N3N2N1O2M2O2N1O2M2O1O000O1000000O1000000O10000O10000O100O10000O10000O1000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000001O00000O101O000000001O0O1000001O00000O2O001O001O00001N1N3M2N3M2N3M2N3M2N2N3L3N3M3M3M3M3M3M2N3M3M3M3M3M3L4M3M3M3L4M3M3M3L5L3M3L4M3L4M4K4L4M0O106I8I7H8H8I7HPZk0"}, "label": "the white ranch dressing in a cup"}]}
{"id": 361197, "image": "COCO_train2014_000000361197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white dressing in a plastic container\"."}], "task": "refering", "answer_template": "The \"white dressing in a plastic container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.5253125, 0.5952083333333333, 0.9079062499999999, 0.9997083333333333], "iscrowd": 0, "area": 38026.836449999995, "rle": {"size": [480, 640], "counts": "_jm47g>j0WOg0XO4M2M4M2M3N3L3N3L3N3L3N3M2N3N1N2O2M2O2M2O2M2O2M2O1N3N1N3N1N3N1N3N1N3N1N3N1N3N2N1O2M2O2N1O2M2O2N1O2M2O2N1N3N2N1O2M2O2N1O2M2O1O000O1000000O1000000O10000O10000O100O10000O10000O1000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000001O00000O101O000000001O0O1000001O00000O2O001O001O00001N1N3M2N3M2N3M2N3M2N2N3L3N3M3M3M3M3M3M2N3M3M3M3M3M3L4M3M3M3L4M3M3M3L5L3M3L4M3L4M4K4L4M0O106I8I7H8H8I7HPZk0"}, "label": "white dressing in a plastic container"}]}
{"id": 33527, "image": "COCO_train2014_000000033527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mostly black kitten\"."}], "task": "refering", "answer_template": "The \"a mostly black kitten\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 163, 164], "bbox": [0.02148, 0.1527027027027027, 0.38242000000000004, 0.8000000000000002], "iscrowd": 0, "area": 26372.781849999992, "rle": {"size": [333, 500], "counts": "Qg37S:<D<oF^Ob7j1K6J5K5L4J7J8H9G7H5L5K4L4M3N2M4L3N2M3N2M4L3N2M2O2O0O2N101N1O2O1N1O2O0O2O0O2N101N1O2O0O2N2O0O2O0O2N10O000O2O000N3M2N3N1010O100000O10O10000000O0100000000O010000000O010000000O10O1N2O1O001O1UN^KiMd4U2^KjMc4T2`KjMa4T2bKjM_4U2cKiM^4U2eKiM\\4W2eKgM\\4Y2eKeM\\4Z2fKdM[4\\2fKbM[4^2fK`M[4`2eK_M\\4a2eK]M\\4b2fK\\M[4d2fK[MZ4e2gKZMY4f2hKYMW4h2jKWMV4i2jKWMV4h2lKWMT4i2mKVMS4j2nKUMR4k2oKTMQ4l2PLSMP4l2QLTMo3m2QLRMo3n2RLQMn3P3QLPMo3P3RLoLn3R3RLmLn3T3QLlLo3T3[LbLe3_3d11O1O101N100O2O0O101N100O2O0O1O101N100O2O0O10PLmId3S6[LoIe3Q6YLQJP2L_OR6^NUJS2K^O\\6a0fI]O[6b0gI]OX6c0jI\\OV6b0mI\\OT6c0nI\\OR6c0oI]OP6c0RJ\\On5c0TJ[On5b0UJ]Ok5b0WJ]Oj5a0WJ^Oj5a0XJ^Oi5`0YJ_Oh5>YJAj5=WJAn5;SJCV65jIJ^8N2N1O2N1O2N2N:FfYT3"}, "label": "a mostly black kitten"}]}
{"id": 33527, "image": "COCO_train2014_000000033527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"almost all black cat\"."}], "task": "refering", "answer_template": "The \"almost all black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 163, 164], "bbox": [0.02148, 0.1527027027027027, 0.38242000000000004, 0.8000000000000002], "iscrowd": 0, "area": 26372.781849999992, "rle": {"size": [333, 500], "counts": "Qg37S:<D<oF^Ob7j1K6J5K5L4J7J8H9G7H5L5K4L4M3N2M4L3N2M3N2M4L3N2M2O2O0O2N101N1O2O1N1O2O0O2O0O2N101N1O2O0O2N2O0O2O0O2N10O000O2O000N3M2N3N1010O100000O10O10000000O0100000000O010000000O010000000O10O1N2O1O001O1UN^KiMd4U2^KjMc4T2`KjMa4T2bKjM_4U2cKiM^4U2eKiM\\4W2eKgM\\4Y2eKeM\\4Z2fKdM[4\\2fKbM[4^2fK`M[4`2eK_M\\4a2eK]M\\4b2fK\\M[4d2fK[MZ4e2gKZMY4f2hKYMW4h2jKWMV4i2jKWMV4h2lKWMT4i2mKVMS4j2nKUMR4k2oKTMQ4l2PLSMP4l2QLTMo3m2QLRMo3n2RLQMn3P3QLPMo3P3RLoLn3R3RLmLn3T3QLlLo3T3[LbLe3_3d11O1O101N100O2O0O101N100O2O0O1O101N100O2O0O10PLmId3S6[LoIe3Q6YLQJP2L_OR6^NUJS2K^O\\6a0fI]O[6b0gI]OX6c0jI\\OV6b0mI\\OT6c0nI\\OR6c0oI]OP6c0RJ\\On5c0TJ[On5b0UJ]Ok5b0WJ]Oj5a0WJ^Oj5a0XJ^Oi5`0YJ_Oh5>YJAj5=WJAn5;SJCV65jIJ^8N2N1O2N1O2N2N:FfYT3"}, "label": "almost all black cat"}]}
{"id": 33527, "image": "COCO_train2014_000000033527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat with other cat sitting with other cat\"."}], "task": "refering", "answer_template": "The \"a black and white cat with other cat sitting with other cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 42, 43, 44, 45, 46, 51, 52, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142], "bbox": [0.33226, 0.14831831831831832, 0.9862799999999999, 0.6719219219219219], "iscrowd": 0, "area": 38250.339900000006, "rle": {"size": [333, 500], "counts": "cQf12W:5O2N1N2O2N1O1O1N2O1O1O1N2O1O1O1Nl0UO2M4M3L3M4M3L3N3L4M3L3M4M2M100O2O0O101O00001TJlL]4T3`KPM^4Q3_KSM`4n2aJSM:3S5k2bJRN]5o1aJSN^5n1`JTN_5m1`JSN_5^31N1O101O01O100O1O101N00000000000O1000000000000O10O10000000001O00001O0O2O001N100O2O001N101O0O101O0O2O0O2O000O2O001N2O101N1O2N1O2N1O2N2N1O2N1O10O000001O0000001O0000000010O0001O1O001O001O1O001O010O1O001O001O00000000000001O0O10000000000O1000000000000O10001O00000O1000000000000O10000000000O1000001O00000O10000000000O10000000000O1000001O00000O10000000000O10000000000000000O1000O1O1O100O1N2O[I]La6b3`IaL^6^3aIfL]6X3dIlLY6S3gIPMW6o2hIUMV6j2jIYMT6e2mI_MP6`2PJaMP6^2QJbMn5_2RJaMn5_2QJbMo5]2RJcMn5]2RJcMn5]2QJdMo5\\2QJeMn5[2RJeMn5[2QJfMo5Z2QJfMn5[2RJeMn5]30001O0000O1O1O001O1O1O100O1O1O1O001O1O1O1O100O1ON2O1N2N2N2N2002N100O1O1O1O1O1O5K6J6J6K5J4L2N2N1O2M2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N3M2IkU2"}, "label": "a black and white cat with other cat sitting with other cat"}]}
{"id": 33527, "image": "COCO_train2014_000000033527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat watching on the remote on the sofa\"."}], "task": "refering", "answer_template": "The \"cat watching on the remote on the sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 42, 43, 44, 45, 46, 51, 52, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142], "bbox": [0.33226, 0.14831831831831832, 0.9862799999999999, 0.6719219219219219], "iscrowd": 0, "area": 38250.339900000006, "rle": {"size": [333, 500], "counts": "cQf12W:5O2N1N2O2N1O1O1N2O1O1O1N2O1O1O1Nl0UO2M4M3L3M4M3L3N3L4M3L3M4M2M100O2O0O101O00001TJlL]4T3`KPM^4Q3_KSM`4n2aJSM:3S5k2bJRN]5o1aJSN^5n1`JTN_5m1`JSN_5^31N1O101O01O100O1O101N00000000000O1000000000000O10O10000000001O00001O0O2O001N100O2O001N101O0O101O0O2O0O2O000O2O001N2O101N1O2N1O2N1O2N2N1O2N1O10O000001O0000001O0000000010O0001O1O001O001O1O001O010O1O001O001O00000000000001O0O10000000000O1000000000000O10001O00000O1000000000000O10000000000O1000001O00000O10000000000O10000000000O1000001O00000O10000000000O10000000000000000O1000O1O1O100O1N2O[I]La6b3`IaL^6^3aIfL]6X3dIlLY6S3gIPMW6o2hIUMV6j2jIYMT6e2mI_MP6`2PJaMP6^2QJbMn5_2RJaMn5_2QJbMo5]2RJcMn5]2RJcMn5]2QJdMo5\\2QJeMn5[2RJeMn5[2QJfMo5Z2QJfMn5[2RJeMn5]30001O0000O1O1O001O1O1O100O1O1O1O001O1O1O1O100O1ON2O1N2N2N2N2002N100O1O1O1O1O1O5K6J6J6K5J4L2N2N1O2M2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N3M2IkU2"}, "label": "cat watching on the remote on the sofa"}]}
{"id": 221949, "image": "COCO_train2014_000000221949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red and tan purse with the hearts\"."}], "task": "refering", "answer_template": "The \"the red and tan purse with the hearts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314], "bbox": [0.454953125, 0.6480841121495327, 0.6897968750000001, 0.9030373831775701], "iscrowd": 0, "area": 11847.829599999995, "rle": {"size": [428, 640], "counts": "mni37f<c0I8G8I5K4M3L4M3H8G9M3N1N3N2M3N2N2M3N2O1N2O1O0O2O1N2O0O100O11O0000001O000000001O0000001O000000001O0000001O000000001O001O0O2O001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O001O001O1O001N1O2O0O2O0O2N2O0O2O1M3N1N3N2M2O2M3N2N1N3N2M2O3L3N2M3N2M3JSkb2"}, "label": "the red and tan purse with the hearts"}]}
{"id": 221949, "image": "COCO_train2014_000000221949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bag with hearts\"."}], "task": "refering", "answer_template": "The \"a bag with hearts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314], "bbox": [0.454953125, 0.6480841121495327, 0.6897968750000001, 0.9030373831775701], "iscrowd": 0, "area": 11847.829599999995, "rle": {"size": [428, 640], "counts": "mni37f<c0I8G8I5K4M3L4M3H8G9M3N1N3N2M3N2N2M3N2O1N2O1O0O2O1N2O0O100O11O0000001O000000001O0000001O000000001O0000001O000000001O001O0O2O001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O001O001O1O001N1O2O0O2O0O2N2O0O2O1M3N1N3N2M2O2M3N2N1N3N2M2O3L3N2M3N2M3JSkb2"}, "label": "a bag with hearts"}]}
{"id": 221949, "image": "COCO_train2014_000000221949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bag\"."}], "task": "refering", "answer_template": "The \"brown bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.6737968750000001, 0.7402570093457943, 0.9792500000000001, 0.9358644859813083], "iscrowd": 0, "area": 7396.767600000005, "rle": {"size": [428, 640], "counts": "b^d52Z=5K4L4K5L5K4L`0@100O1O1O100O1O100O1O1O100O1O100O1O0010O01O010O00010O001O010O010O0010O01O010O0010O01O01O01O010O0001O001O01O01O001O01O01O001O0010O0001O001O01O01O001O0010O01O1O1O1O010O1O1O001O100O001O1O1O10O01O1O1O001O1O0O10001O0000001N100000001O0O1000001O00000O2O000000001O0O1000001O000O10001O0000001O0000001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O1O001O001O1O010O1O001O001O]=0Sd4"}, "label": "brown bag"}]}
{"id": 221949, "image": "COCO_train2014_000000221949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue purse next to other purse and suit ties\"."}], "task": "refering", "answer_template": "The \"a blue purse next to other purse and suit ties\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327], "bbox": [0.0936875, 0.7511915887850467, 0.371953125, 0.9981308411214953], "iscrowd": 0, "area": 7014.761149999998, "rle": {"size": [428, 640], "counts": "_^i01hj00cB0mB4f<901O0010O01O001O1O001O010O001O001O001O001O001O001O00001O00001O00001O001O0O101N100O2O0O1O2O0O101N2O0O2O1N2O1N101N2O1N10O010O0010O010O01O010O010O0010O0010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O00010O010O0010O010O01O010O010O010O0010O010O01O010O010O0010O010O00010O010O01O010O010O0010O010O01O010O001O010O1O100O1O103L3M4L4M2L5K4L5KSfW5"}, "label": "a blue purse next to other purse and suit ties"}]}
{"id": 221949, "image": "COCO_train2014_000000221949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bag with pink trim under a black case\"."}], "task": "refering", "answer_template": "The \"a blue bag with pink trim under a black case\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327], "bbox": [0.0936875, 0.7511915887850467, 0.371953125, 0.9981308411214953], "iscrowd": 0, "area": 7014.761149999998, "rle": {"size": [428, 640], "counts": "_^i01hj00cB0mB4f<901O0010O01O001O1O001O010O001O001O001O001O001O001O00001O00001O00001O001O0O101N100O2O0O1O2O0O101N2O0O2O1N2O1N101N2O1N10O010O0010O010O01O010O010O0010O0010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O00010O010O0010O010O01O010O010O010O0010O010O01O010O010O0010O010O00010O010O01O010O010O0010O010O01O010O001O010O1O100O1O103L3M4L4M2L5K4L5KSfW5"}, "label": "a blue bag with pink trim under a black case"}]}
{"id": 271106, "image": "COCO_train2014_000000271106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plants in the cistern of the toilet\"."}], "task": "refering", "answer_template": "The \"the plants in the cistern of the toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 73, 74, 75, 90, 91, 92, 93, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 156, 157, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 225, 226, 227, 228], "bbox": [0.207125, 0.157203125, 0.8320416666666667, 0.60121875], "iscrowd": 0, "area": 37497.1904, "rle": {"size": [640, 480], "counts": "bSn12cc0?D;M3L10001N10000O10O1O001O000000O101O00000O202O0O1O1Z^OjN``0X1]_O_Om?b0Q@5Y?Le@j0f>WOWAW1]>kN`AV1a>jN]AX1b>iN\\AX1d>iNZAY1e>hNYAZ1g>gNUA[1k>fNl@c1S?e110ZO[KoAe4P>g0N2N101[N\\JkDe5T;`JgDb5W;`JgDb5W;^JiDc5V;]JjDe5T;[JlDf5R;\\JQDF>P6_;ZJSD40d5\\;gJdDE0e5Q;PKoD[O0g5P;mJPE[O1i5P;kJnD\\O3j5n:iJoD]O3k5o:fJnD_O3m5n:fJlD]O6o5m:gJiDYO;Q6m:VKSEl4l:SKTEn4m:PKSER5l:R20O010O010O001O00001O001O001O001O00000000000O100O10000O100O010O100O100O100O10O0100O100O10000O2O1N2O1N2O1N2ZHnC\\7S<dHoCX7S<gHoCW7R<hHQDT7Q<lHQDQ7P<nHRDo6P<PISDm6n;SIUDh6m;WIXDd6i;[I]D_6c;bI\\D]6f;bIYD_6d<O10O0100O010O10O10O01000O010O10O10O1O0N3N3M4K5L4L4L5oJeAU4_>fKdAX4`>bKdA\\4`>_KcA_4P?N2L4M3L4M3L4M3L4TO^@VMf?f2l0L41O001O1O1O1O001O1O1O1O1O001OfNT_OAk`03b_OK_`04c_OK]`04e_OJ\\`06e_OIZ`07i_OFX`09j_OFU`0;l_OCT`0=n_OBR`0=P@AP`0?R@ZOT`0f0m_OROY`0n0\\1010O01O01O01O010O000010O00010O00010O0010O00010O010O1O010O10O0100O00100O010O1O010O10O01O100O010O10O01000O0100O01000O0100O010O1O010O010O10O10O10O10000O010O10000O1000000000O1000O101O00001O00001Cn\\OJRc06U]OCkb0==O00000N]_b1"}, "label": "the plants in the cistern of the toilet"}]}
{"id": 271106, "image": "COCO_train2014_000000271106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plant with the red and green leaves\"."}], "task": "refering", "answer_template": "The \"the plant with the red and green leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 73, 74, 75, 90, 91, 92, 93, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 156, 157, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 225, 226, 227, 228], "bbox": [0.207125, 0.157203125, 0.8320416666666667, 0.60121875], "iscrowd": 0, "area": 37497.1904, "rle": {"size": [640, 480], "counts": "bSn12cc0?D;M3L10001N10000O10O1O001O000000O101O00000O202O0O1O1Z^OjN``0X1]_O_Om?b0Q@5Y?Le@j0f>WOWAW1]>kN`AV1a>jN]AX1b>iN\\AX1d>iNZAY1e>hNYAZ1g>gNUA[1k>fNl@c1S?e110ZO[KoAe4P>g0N2N101[N\\JkDe5T;`JgDb5W;`JgDb5W;^JiDc5V;]JjDe5T;[JlDf5R;\\JQDF>P6_;ZJSD40d5\\;gJdDE0e5Q;PKoD[O0g5P;mJPE[O1i5P;kJnD\\O3j5n:iJoD]O3k5o:fJnD_O3m5n:fJlD]O6o5m:gJiDYO;Q6m:VKSEl4l:SKTEn4m:PKSER5l:R20O010O010O001O00001O001O001O001O00000000000O100O10000O100O010O100O100O100O10O0100O100O10000O2O1N2O1N2O1N2ZHnC\\7S<dHoCX7S<gHoCW7R<hHQDT7Q<lHQDQ7P<nHRDo6P<PISDm6n;SIUDh6m;WIXDd6i;[I]D_6c;bI\\D]6f;bIYD_6d<O10O0100O010O10O10O01000O010O10O10O1O0N3N3M4K5L4L4L5oJeAU4_>fKdAX4`>bKdA\\4`>_KcA_4P?N2L4M3L4M3L4M3L4TO^@VMf?f2l0L41O001O1O1O1O001O1O1O1O1O001OfNT_OAk`03b_OK_`04c_OK]`04e_OJ\\`06e_OIZ`07i_OFX`09j_OFU`0;l_OCT`0=n_OBR`0=P@AP`0?R@ZOT`0f0m_OROY`0n0\\1010O01O01O01O010O000010O00010O00010O0010O00010O010O1O010O10O0100O00100O010O1O010O10O01O100O010O10O01000O0100O01000O0100O010O1O010O010O10O10O10O10000O010O10000O1000000000O1000O101O00001O00001Cn\\OJRc06U]OCkb0==O00000N]_b1"}, "label": "the plant with the red and green leaves"}]}
{"id": 271106, "image": "COCO_train2014_000000271106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant growing out of a toilet bowl\"."}], "task": "refering", "answer_template": "The \"a plant growing out of a toilet bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 228, 229, 230, 231, 232, 233, 234, 236, 237, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 347, 348, 349, 364], "bbox": [0.40206250000000004, 0.46473437500000003, 0.9905208333333333, 0.9233750000000001], "iscrowd": 0, "area": 39721.21910000001, "rle": {"size": [640, 480], "counts": "^Pi32kc04K6K4M3N2O1N2O2M2O1N2O1N2O;D>C5J7J5J3`BdMX9]2eFdM[9^2aFeM_9\\2]FfMc9\\2YFgMg9Z2UFhMk9Z2RFhMn9Y2nEjMQ:X2kEjMV:V2gEmMX:U2eElM\\:U2`EnM_:T2]EnMd:S2XEPNg:R2UEQNk:P2REQNn:Q2nDRNR;o1jDSNV;o1fDTNZ;m1bDUN^;m1_DUNa;l1[DWNd;k1XDWNi;j1SDYNl;i1PDYNQ<h1lCZNS<h1iCZNX<g1dC\\N[<f1aC]N_<d1^C]Nb<e1ZC^Nf<c1VC_Nj<c1RC`Nn<`1oBbNQ=`1lBbNT=_1hBdNW=R410O010O010O10O01M2N3M2O21O0O2O1N2O1N2O001N\\O`IkC_6R<jIiCT6X<QJdCn5_<TJ]Cj5g<f02N3L4M3L4MN1O2N2M2O2N1O2N1O2N2000O01000000O1O1O1O001N2O1O1O1O1N101O1O1N2O1O1O10O10O100000000O01000000000O1000O1000O10000000O0100000000O1000O100000O100000O10O11O001O001O0O2O001O1O001O0O2O001O001O001O0O2O001O001O001N1\\IgCa5Z<_J`Df4`;ZKoDH^Nm3d<[LoDGgNc3\\<dLnDHPOZ3U<lLlDHZOQ3l;VMkDHBU1dN5Q=lNjDILj0_N;o<POgDJ4`0\\Na0k<TOfDI>7VNg0h<WOeDK<6\\Nd0d<[OeDI;7bN`0^<@QG1fL:Z<EoF0lL7V<ImFOSM2R<NkF0XMMm;4iFOP;0PEOR;1mDNT;3jDNW;1iDNY;2fDM\\;2cDO];2bDM`;3_DLc;3\\DNd;3[DLg;3YDLi;4UDMl;3SDLn;4RDKP<5nCLS<4lCKV<4jCKW<6gCKZ<4fCK\\<5g310O100O01000O010O1000O0100O01000O10O01000O010000O010O10O10O100O01000O10O01000O0100O10O10O10O10O10Pd00P\\O0Pd00l[3"}, "label": "a plant growing out of a toilet bowl"}]}
{"id": 271106, "image": "COCO_train2014_000000271106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower planted in a toilet bowl\"."}], "task": "refering", "answer_template": "The \"a flower planted in a toilet bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 228, 229, 230, 231, 232, 233, 234, 236, 237, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 347, 348, 349, 364], "bbox": [0.40206250000000004, 0.46473437500000003, 0.9905208333333333, 0.9233750000000001], "iscrowd": 0, "area": 39721.21910000001, "rle": {"size": [640, 480], "counts": "^Pi32kc04K6K4M3N2O1N2O2M2O1N2O1N2O;D>C5J7J5J3`BdMX9]2eFdM[9^2aFeM_9\\2]FfMc9\\2YFgMg9Z2UFhMk9Z2RFhMn9Y2nEjMQ:X2kEjMV:V2gEmMX:U2eElM\\:U2`EnM_:T2]EnMd:S2XEPNg:R2UEQNk:P2REQNn:Q2nDRNR;o1jDSNV;o1fDTNZ;m1bDUN^;m1_DUNa;l1[DWNd;k1XDWNi;j1SDYNl;i1PDYNQ<h1lCZNS<h1iCZNX<g1dC\\N[<f1aC]N_<d1^C]Nb<e1ZC^Nf<c1VC_Nj<c1RC`Nn<`1oBbNQ=`1lBbNT=_1hBdNW=R410O010O010O10O01M2N3M2O21O0O2O1N2O1N2O001N\\O`IkC_6R<jIiCT6X<QJdCn5_<TJ]Cj5g<f02N3L4M3L4MN1O2N2M2O2N1O2N1O2N2000O01000000O1O1O1O001N2O1O1O1O1N101O1O1N2O1O1O10O10O100000000O01000000000O1000O1000O10000000O0100000000O1000O100000O100000O10O11O001O001O0O2O001O1O001O0O2O001O001O001O0O2O001O001O001N1\\IgCa5Z<_J`Df4`;ZKoDH^Nm3d<[LoDGgNc3\\<dLnDHPOZ3U<lLlDHZOQ3l;VMkDHBU1dN5Q=lNjDILj0_N;o<POgDJ4`0\\Na0k<TOfDI>7VNg0h<WOeDK<6\\Nd0d<[OeDI;7bN`0^<@QG1fL:Z<EoF0lL7V<ImFOSM2R<NkF0XMMm;4iFOP;0PEOR;1mDNT;3jDNW;1iDNY;2fDM\\;2cDO];2bDM`;3_DLc;3\\DNd;3[DLg;3YDLi;4UDMl;3SDLn;4RDKP<5nCLS<4lCKV<4jCKW<6gCKZ<4fCK\\<5g310O100O01000O010O1000O0100O01000O10O01000O010000O010O10O10O100O01000O10O01000O0100O10O10O10O10O10Pd00P\\O0Pd00l[3"}, "label": "a flower planted in a toilet bowl"}]}
{"id": 410373, "image": "COCO_train2014_000000410373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse in between 2 other horses\"."}], "task": "refering", "answer_template": "The \"a horse in between 2 other horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126], "bbox": [0.3288125, 0.46049107142857143, 0.53421875, 0.7423214285714286], "iscrowd": 0, "area": 4163.2529, "rle": {"size": [224, 640], "counts": "QS^11n61O1O1O2N1O1G:@?N2NM3101N110O010O0100O001O001O10YJRO[5n0eJSO[5l0dJVO\\5k0aJXO]5j0`JXO`5T1O10001ON2N2M3N2M3N2N2M3O1001O000000001O000000001O0001O1O100O1O001O100L4L3M4L4K4Nd6N^C2R50hJ4X5NcJ5\\5O]J5c50TJ4l5a0N2O1_JoNQ5`1O1000O0M2M3N4N2000000000000000000000mJeNa4S1PKRO>Lb4n0TKVO8Nd4k0VKVO50e4i0WKWO31f4MoJg08[O12m4b0SK\\OO3n4`0TK]OM3P5`0TKLk44VKKk44VKHn48SKCQ5<RK^OR5a0g00O01001O2N2O001JVORJP1m50O010O0O1N2O14L3M2F9F;I_RQ2"}, "label": "a horse in between 2 other horses"}]}
{"id": 410373, "image": "COCO_train2014_000000410373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with a white saddle\"."}], "task": "refering", "answer_template": "The \"a horse with a white saddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126], "bbox": [0.3288125, 0.46049107142857143, 0.53421875, 0.7423214285714286], "iscrowd": 0, "area": 4163.2529, "rle": {"size": [224, 640], "counts": "QS^11n61O1O1O2N1O1G:@?N2NM3101N110O010O0100O001O001O10YJRO[5n0eJSO[5l0dJVO\\5k0aJXO]5j0`JXO`5T1O10001ON2N2M3N2M3N2N2M3O1001O000000001O000000001O0001O1O100O1O001O100L4L3M4L4K4Nd6N^C2R50hJ4X5NcJ5\\5O]J5c50TJ4l5a0N2O1_JoNQ5`1O1000O0M2M3N4N2000000000000000000000mJeNa4S1PKRO>Lb4n0TKVO8Nd4k0VKVO50e4i0WKWO31f4MoJg08[O12m4b0SK\\OO3n4`0TK]OM3P5`0TKLk44VKKk44VKHn48SKCQ5<RK^OR5a0g00O01001O2N2O001JVORJP1m50O010O0O1N2O14L3M2F9F;I_RQ2"}, "label": "a horse with a white saddle"}]}
{"id": 410373, "image": "COCO_train2014_000000410373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse of the man in red\"."}], "task": "refering", "answer_template": "The \"the horse of the man in red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 104, 105, 106, 107, 108, 127, 128, 129, 130], "bbox": [0.52734375, 0.47098214285714285, 0.7390625, 0.7433035714285714], "iscrowd": 0, "area": 3768.25, "rle": {"size": [224, 640], "counts": "QSZ22l61O2N2M2N3N1N3M3N1O001O1N1N2oI]Og5c0YJCc5;^JG`57aJK_52cJO\\50fJ1Y5MiJ1X5NiJO[5NhJOZ5NXJN:1a5>YJCf5c0UJ]Oj5m00O]JnNY5S1dJPO\\5P1aJSO^5X100O100O100000H8N2N2M3L4M301O000O2O0010O1O10O01JXORJh0o55000O100O1O001O100O1O002N2N3M2N4L4MdIJP62kI3W6KfI8[6GcI;]64N2QO]O^Kf0b4\\OZKf0f4\\OUKg0l4\\OmJf0T5_OdJb0^5>2N3N010O010O010O10O0100O1001O001O00G[JWOe5g0^JYOa5f0`JZO_5?jJAU5=nJBR5<PKDP5:SKFl48WKGi47YKJf44]KIe48\\KEe4<n02O1O1O1O1O0O2O001L40O2O1O1N2O001M2N2N3M3M3MS]T1"}, "label": "the horse of the man in red"}]}
{"id": 410373, "image": "COCO_train2014_000000410373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black horse that is leading the pack of other horses\"."}], "task": "refering", "answer_template": "The \"a black horse that is leading the pack of other horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 104, 105, 106, 107, 108, 127, 128, 129, 130], "bbox": [0.52734375, 0.47098214285714285, 0.7390625, 0.7433035714285714], "iscrowd": 0, "area": 3768.25, "rle": {"size": [224, 640], "counts": "QSZ22l61O2N2M2N3N1N3M3N1O001O1N1N2oI]Og5c0YJCc5;^JG`57aJK_52cJO\\50fJ1Y5MiJ1X5NiJO[5NhJOZ5NXJN:1a5>YJCf5c0UJ]Oj5m00O]JnNY5S1dJPO\\5P1aJSO^5X100O100O100000H8N2N2M3L4M301O000O2O0010O1O10O01JXORJh0o55000O100O1O001O100O1O002N2N3M2N4L4MdIJP62kI3W6KfI8[6GcI;]64N2QO]O^Kf0b4\\OZKf0f4\\OUKg0l4\\OmJf0T5_OdJb0^5>2N3N010O010O010O10O0100O1001O001O00G[JWOe5g0^JYOa5f0`JZO_5?jJAU5=nJBR5<PKDP5:SKFl48WKGi47YKJf44]KIe48\\KEe4<n02O1O1O1O1O0O2O001L40O2O1O1N2O001M2N2N3M3M3MS]T1"}, "label": "a black horse that is leading the pack of other horses"}]}
{"id": 197383, "image": "COCO_train2014_000000197383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red taxi standing next to a yellow sign\"."}], "task": "refering", "answer_template": "The \"a red taxi standing next to a yellow sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278], "bbox": [0.0016250000000000001, 0.5273749999999999, 0.3904583333333333, 0.73259375], "iscrowd": 0, "area": 18838.26620000001, "rle": {"size": [640, 480], "counts": "bo0o2Pa00O01000O01O100O1001O1O1O1O100O1O1O1O10000O100O1000000000000O10000000000000000O10000000001O0000000O100000000000000000000000001O000000000O1O100O100O100O100000000O101O0000000O10000000000000000001O0000000000001O0000000000101N2N2N05K100O1O1O0010O0001O00010O00001O00O2O0000000O101N2M2O2L4I7M3M3N2M201O1O1N2O1O100O001O1O1O2N1O1O2N1O0010O1O0O2O1O001O0000PN^^Oc1aa0]N_^Oc1ba0\\N_^Oc1ba0\\N_^Od1aa0[N`^Oc1fa0XN\\^Of1Pb0O1O1N2O3M7hN]]Ob0mYg5"}, "label": "a red taxi standing next to a yellow sign"}]}
{"id": 197383, "image": "COCO_train2014_000000197383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red san francisco taxi cab climbs a small incline\"."}], "task": "refering", "answer_template": "The \"a red san francisco taxi cab climbs a small incline\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278], "bbox": [0.0016250000000000001, 0.5273749999999999, 0.3904583333333333, 0.73259375], "iscrowd": 0, "area": 18838.26620000001, "rle": {"size": [640, 480], "counts": "bo0o2Pa00O01000O01O100O1001O1O1O1O100O1O1O1O10000O100O1000000000000O10000000000000000O10000000001O0000000O100000000000000000000000001O000000000O1O100O100O100O100000000O101O0000000O10000000000000000001O0000000000001O0000000000101N2N2N05K100O1O1O0010O0001O00010O00001O00O2O0000000O101N2M2O2L4I7M3M3N2M201O1O1N2O1O100O001O1O1O2N1O1O2N1O0010O1O0O2O1O001O0000PN^^Oc1aa0]N_^Oc1ba0\\N_^Oc1ba0\\N_^Od1aa0[N`^Oc1fa0XN\\^Of1Pb0O1O1N2O3M7hN]]Ob0mYg5"}, "label": "a red san francisco taxi cab climbs a small incline"}]}
{"id": 172813, "image": "COCO_train2014_000000172813.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a curved sofa with many pillows\"."}], "task": "refering", "answer_template": "The \"a curved sofa with many pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 272, 285, 286, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.25166107382550335, 0.7804932735426009, 0.9630369127516779, 1.0], "iscrowd": 0, "area": 22296.790650000006, "rle": {"size": [446, 596], "counts": "[hQ21l=2O1N1O1O1O1O100O1O1O1O1O2N1O100O1O1O1O1000000000000000000O10000000000O10000000000O10000O1O1O1O1O1O1O1O1000000O1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O101O00000O1000000O100000000O1000000O100000000O10000000000O10000000000000000O10000000000000000O100000000002N4L4L4L3M00O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O2O0O100O1O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100OQNaEa0n;0O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O10000O101c0\\OO1O1O1O1N200O100O100O1O100O100O1N2N2N2O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O100O1O100O1O100OM3B?B=B>C>A>Ci\\18ScN:F:F:F:F:G9F:F:L4M31O1O001O1O001O1O001O1O000000000000000000001O0000000000000000000000000000000000O1000000O1O100O100O^Ob0[Of0[Od0\\O[X9"}, "label": "a curved sofa with many pillows"}]}
{"id": 172813, "image": "COCO_train2014_000000172813.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"all five pillows in the foreground\"."}], "task": "refering", "answer_template": "The \"all five pillows in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 272, 285, 286, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.25166107382550335, 0.7804932735426009, 0.9630369127516779, 1.0], "iscrowd": 0, "area": 22296.790650000006, "rle": {"size": [446, 596], "counts": "[hQ21l=2O1N1O1O1O1O100O1O1O1O1O2N1O100O1O1O1O1000000000000000000O10000000000O10000000000O10000O1O1O1O1O1O1O1O1000000O1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O101O00000O1000000O100000000O1000000O100000000O10000000000O10000000000000000O10000000000000000O100000000002N4L4L4L3M00O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O2O0O100O1O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100OQNaEa0n;0O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O10000O101c0\\OO1O1O1O1N200O100O100O1O100O100O1N2N2N2O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O100O1O100O1O100OM3B?B=B>C>A>Ci\\18ScN:F:F:F:F:G9F:F:L4M31O1O001O1O001O1O001O1O000000000000000000001O0000000000000000000000000000000000O1000000O1O100O100O^Ob0[Of0[Od0\\O[X9"}, "label": "all five pillows in the foreground"}]}
{"id": 172813, "image": "COCO_train2014_000000172813.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown sofa at the left most side\"."}], "task": "refering", "answer_template": "The \"the brown sofa at the left most side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [252, 273, 274, 275, 276, 277, 294, 295, 296, 297, 298, 299, 300, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327], "bbox": [0.0, 0.8022197309417041, 1.0, 1.0], "iscrowd": 0, "area": 8214.070049999993, "rle": {"size": [446, 596], "counts": "W;9d=5L5K4L4L3M1O010O010O01O010O010O010O1O10O010O0010O01O01O01O0010O01O0010O0001O010O001O010O000010O01O010O0010O0010O010O010O010O01O01O001O001O0O101N101N101N101O0O101O0O2O0O2O001N10001N101O0O2O000O101O00001O000O101O00001O0000001O01O01O00010O0000010O000100O1O00100O00100O1O1O100O1O1O1O001O1O001O001O001LdCdN\\<[1eCeN\\<Z1dCfN\\<Y1eCgN\\<W1eCiN[<V1fCjN[<T1fClNZ<T1fClN[<R1fCnNZ<Q1gCoNZ<o0gCQOY<o0gCQOY<n0hCROY<l0hCTOX<l0hCTOX<k0iCUOX<i0iCWOW<i0jCVOV<i0kCWOV<g0kCYOU<e0mC[OS<d0nC\\OS<c0mC]OS<c0mC]OS<b0nC^OS<a0mC_OS<a0mC_OS<a0mC@S<?mCAS<?mCAS<?mCAS<?mCAS<?mCAS<?mCAT<>lCBT<>lCBT<>lCBT<>lCBT<>lCBU<=kCCU<=kCCU<=kCCU<=kCCU<=kCCV<<jCDV<=iCCW<=iCCW<8nCHR<6PDJQ<5oCKQ<4PDLP<4PDLP<3QDMP<1QDOo;0RD0n<00000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000][40cdK000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000YV<0giC0000000000000000000000000000000000000000000000000000000Qmd00oR[O000000000000000YV<"}, "label": "the brown sofa at the left most side"}]}
{"id": 172813, "image": "COCO_train2014_000000172813.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the light colored couch\"."}], "task": "refering", "answer_template": "The \"the light colored couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [252, 273, 274, 275, 276, 277, 294, 295, 296, 297, 298, 299, 300, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327], "bbox": [0.0, 0.8022197309417041, 1.0, 1.0], "iscrowd": 0, "area": 8214.070049999993, "rle": {"size": [446, 596], "counts": "W;9d=5L5K4L4L3M1O010O010O01O010O010O010O1O10O010O0010O01O01O01O0010O01O0010O0001O010O001O010O000010O01O010O0010O0010O010O010O010O01O01O001O001O0O101N101N101N101O0O101O0O2O0O2O001N10001N101O0O2O000O101O00001O000O101O00001O0000001O01O01O00010O0000010O000100O1O00100O00100O1O1O100O1O1O1O001O1O001O001O001LdCdN\\<[1eCeN\\<Z1dCfN\\<Y1eCgN\\<W1eCiN[<V1fCjN[<T1fClNZ<T1fClN[<R1fCnNZ<Q1gCoNZ<o0gCQOY<o0gCQOY<n0hCROY<l0hCTOX<l0hCTOX<k0iCUOX<i0iCWOW<i0jCVOV<i0kCWOV<g0kCYOU<e0mC[OS<d0nC\\OS<c0mC]OS<c0mC]OS<b0nC^OS<a0mC_OS<a0mC_OS<a0mC@S<?mCAS<?mCAS<?mCAS<?mCAS<?mCAS<?mCAT<>lCBT<>lCBT<>lCBT<>lCBT<>lCBU<=kCCU<=kCCU<=kCCU<=kCCU<=kCCV<<jCDV<=iCCW<=iCCW<8nCHR<6PDJQ<5oCKQ<4PDLP<4PDLP<3QDMP<1QDOo;0RD0n<00000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000][40cdK000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000YV<0giC0000000000000000000000000000000000000000000000000000000Qmd00oR[O000000000000000YV<"}, "label": "the light colored couch"}]}
{"id": 492302, "image": "COCO_train2014_000000492302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the out of focus tall plant on the right\"."}], "task": "refering", "answer_template": "The \"the out of focus tall plant on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 37, 38, 39, 40, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390, 410, 411, 412, 413, 433, 434, 435, 436, 456, 457, 458, 459], "bbox": [0.617109375, 0.0045, 1.0, 0.8716250000000001], "iscrowd": 0, "area": 65620.60740000001, "rle": {"size": [640, 640], "counts": "Zag72nc01O2N1O1O2N1O2N1O1O2N1O2N1O1O1O1O001O1O1O100O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O00100O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1V_OlMf?U2W@nMh?S2V@oMi?Q2V@QNi?P2T@SNk?n1S@TNl?m1R@UNm?l1P@WNo?j1o_OXNP`0i1n_OYNQ`0g1m_O\\NR`0e1l_O]NS`0d1k_O^NT`0c1i_O`NV`0a1h_OaNW`0e200O1O1O100kCVLS8j3kGXLV8g3hG[LX8e3fG]L[8b3dG_L\\8a3bGaL_8^3_GdLa8\\3^GeLb8[3\\GfLf8Y3XGiLh8W3VGkLk8T3TGmLl8S3RGoLo8P3oFRMQ9n2mFTMT9k2kFVMV9i2hFYMZ9e2dF]M]9b2aF`Ma9^2^FcMd9[2ZFgMg9X2WFjMk9T2SFnMo9P2PFQNROjNR9U3iGSNSOkNS9S3gGTNTOkNT9T3dGSNVOlNU9R3cGSNWOmNU9Q3aGTNXOnNV9P3^GTNZOnNW9P3]GSNZOPOX9n2PF`Lc0e12POZ9m2`ElLS1Y11QOZ9_3dFbM1PO[9^3cFcM0SO[9Y3dFfMOUO[9T3eFiMNXOY9P3gFjMNZOY9k2iFlML^OX9e2kFoML_OV9c2lFPNLBU9]2nFSNKET9W2QGTNJIS9S2QGTNLNo8n1TGUNM0m8k1TGUNO5j8e1WGVNN9h8b1XGUN0>e8^1XGUN2a0d8\\1WGRN5g0a8Y1VGQN9i0c1oL_4Y4bIoM<n0X1QM`1i0jN[3nMnM?Q1o0UMi1h1eMV2QOkMb0W1d0WMT2a9WM\\I;[M]2X9XMbI0^Mg2P9YMfIEaMR3h8YMbKf2^4YMcKg2\\4YMfKf2Z4YMhKf2W4ZMjKf2V4ZMkKe2T4[MmKe2S4ZMoKe2P4[MRLd2n3[MSLe2l3[MVLd2l3ZMULe2m3XMTLh2m3VMULi2m3TMTLl2n3QMTLn2`:O001O0000001O000000000000001O00000000000000001O00000000000000001O0000000000000000iEZMmJ1W6d2iNaMnJ3P6\\2oNfMQK4h5V2TOkMTK3b5R2WOQNVK0^5o1ZOUNXK0X5k1@WNWK2T5g1FYNUK4o4c1L[NUK4l4`1O\\NWK2j4b1O]NWK0j4c10]NWKNi4e10]NWKNi4e10^NWKLi4f11^NVKKi4g11^NWKJh4h11_NVKIi4h12_NTKIk4g11aNSKIk4f12aNRKJl4e13aNPKJm4e13bNnJJo4d13bNmJKP5c14bNkJKQ5c14cNiJKS5b14cNiJLR5a16cNfJMU5_15eNeJMU5^16fNcJMW5]17eNaJ0W5[18fN`J0W5Z19gN^J0Y5Y1:fN]J2X5X1;gN[J2Z5V1<iNYJ2[5T1<jNXJ3\\5S1=jNUJ5]5n0"}, "label": "the out of focus tall plant on the right"}]}
{"id": 492302, "image": "COCO_train2014_000000492302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green plant in a pot on the floor\"."}], "task": "refering", "answer_template": "The \"the green plant in a pot on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 37, 38, 39, 40, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390, 410, 411, 412, 413, 433, 434, 435, 436, 456, 457, 458, 459], "bbox": [0.617109375, 0.0045, 1.0, 0.8716250000000001], "iscrowd": 0, "area": 65620.60740000001, "rle": {"size": [640, 640], "counts": "Zag72nc01O2N1O1O2N1O2N1O1O2N1O2N1O1O1O1O001O1O1O100O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O00100O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1V_OlMf?U2W@nMh?S2V@oMi?Q2V@QNi?P2T@SNk?n1S@TNl?m1R@UNm?l1P@WNo?j1o_OXNP`0i1n_OYNQ`0g1m_O\\NR`0e1l_O]NS`0d1k_O^NT`0c1i_O`NV`0a1h_OaNW`0e200O1O1O100kCVLS8j3kGXLV8g3hG[LX8e3fG]L[8b3dG_L\\8a3bGaL_8^3_GdLa8\\3^GeLb8[3\\GfLf8Y3XGiLh8W3VGkLk8T3TGmLl8S3RGoLo8P3oFRMQ9n2mFTMT9k2kFVMV9i2hFYMZ9e2dF]M]9b2aF`Ma9^2^FcMd9[2ZFgMg9X2WFjMk9T2SFnMo9P2PFQNROjNR9U3iGSNSOkNS9S3gGTNTOkNT9T3dGSNVOlNU9R3cGSNWOmNU9Q3aGTNXOnNV9P3^GTNZOnNW9P3]GSNZOPOX9n2PF`Lc0e12POZ9m2`ElLS1Y11QOZ9_3dFbM1PO[9^3cFcM0SO[9Y3dFfMOUO[9T3eFiMNXOY9P3gFjMNZOY9k2iFlML^OX9e2kFoML_OV9c2lFPNLBU9]2nFSNKET9W2QGTNJIS9S2QGTNLNo8n1TGUNM0m8k1TGUNO5j8e1WGVNN9h8b1XGUN0>e8^1XGUN2a0d8\\1WGRN5g0a8Y1VGQN9i0c1oL_4Y4bIoM<n0X1QM`1i0jN[3nMnM?Q1o0UMi1h1eMV2QOkMb0W1d0WMT2a9WM\\I;[M]2X9XMbI0^Mg2P9YMfIEaMR3h8YMbKf2^4YMcKg2\\4YMfKf2Z4YMhKf2W4ZMjKf2V4ZMkKe2T4[MmKe2S4ZMoKe2P4[MRLd2n3[MSLe2l3[MVLd2l3ZMULe2m3XMTLh2m3VMULi2m3TMTLl2n3QMTLn2`:O001O0000001O000000000000001O00000000000000001O00000000000000001O0000000000000000iEZMmJ1W6d2iNaMnJ3P6\\2oNfMQK4h5V2TOkMTK3b5R2WOQNVK0^5o1ZOUNXK0X5k1@WNWK2T5g1FYNUK4o4c1L[NUK4l4`1O\\NWK2j4b1O]NWK0j4c10]NWKNi4e10]NWKNi4e10^NWKLi4f11^NVKKi4g11^NWKJh4h11_NVKIi4h12_NTKIk4g11aNSKIk4f12aNRKJl4e13aNPKJm4e13bNnJJo4d13bNmJKP5c14bNkJKQ5c14cNiJKS5b14cNiJLR5a16cNfJMU5_15eNeJMU5^16fNcJMW5]17eNaJ0W5[18fN`J0W5Z19gN^J0Y5Y1:fN]J2X5X1;gN[J2Z5V1<iNYJ2[5T1<jNXJ3\\5S1=jNUJ5]5n0"}, "label": "the green plant in a pot on the floor"}]}
{"id": 492302, "image": "COCO_train2014_000000492302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green plant in a green vase sitting on a wood table\"."}], "task": "refering", "answer_template": "The \"a green plant in a green vase sitting on a wood table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 54, 55, 56, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 262, 263, 264], "bbox": [0.2690625, 0.076234375, 0.6255625, 0.4910312500000001], "iscrowd": 0, "area": 31552.084050000023, "rle": {"size": [640, 640], "counts": "nl\\32nc00O100O101N100O100O2O0O100O2O0O100O101N100O100O10000O100O1fNAS_O?l`0Jk^O7Ua01c^OO\\a03b^OOB@`a0a0n^OO@Bba0`0l^O:Sa0Hl^O8a0YOW??W@9`0_OT?9[@8>BV?7[@89G\\?2Z@77K^?OY@69L^?NY@68M_?NX@58Na?LW@68Na?LX@57Oa?MX@450c?LY@52Of?KX@71Of?KY@7ONh?KZ@7MNj?KY@8JNm?JZ@8HNn?J[@9DNQ`0JY@a1h?^NT@f1S?TOh@Q1S?P2M3L4L4M3jBPK\\;P5cDSKZ;m4eDXKX;h4gD[KW;e4hD_KU;b4iDaKU;_4jDeKS;[4lDhKR;X4mDlKo:T4PEoKo:Q4PEPLP;P4nDRLR;n3lDTLT;m3hDVLX;k3eDWL[;j3bDXL];j3^DZLb;g3UDaLk;k5O001O1O1N101O00000001O3N2M2N1O1OO1N2N2O1N2N2N2O1N2N200000000001O00000000100O4L3M<D7I2M3N2N2N1O2N0000001O0000001O0000001O00000J7M2N2N2N3M2N2N2N3N1N101N2N100O2O1N2N2M3M3F:G:O01O1O1O1O1O00O2O0O100YLPAb06Jk>Do@a09Hi>FPA`0:Hf>GQA`0=Ed>IQA??F`>JRA?a0D^>LSA>b0D[>MTA=f0BW>0UA<k0]OR>6SA<S1VOk==TA;Y1POd=d0TA:a1jN\\=l0TA8n`0HR_O6Qa0IP_O5Qa0Ko^O4Ra0Lo^O1Sa0Om^O0Ta00m^ONUa00l^O0Ta00m^OOSa01m^OOSa01n^OMTa02l^ONTa02m^OMSa02n^OMSa03n^OLRa04n^OLSa03n^OKSa05o^OIQa07R_OFn`0:U_OBl`0>\\1O0000001O001O1O1O2M2O1N3M2O1N3N1N]le4"}, "label": "a green plant in a green vase sitting on a wood table"}]}
{"id": 492302, "image": "COCO_train2014_000000492302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small plant in a planter set on a wooden table\"."}], "task": "refering", "answer_template": "The \"a small plant in a planter set on a wooden table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 54, 55, 56, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 262, 263, 264], "bbox": [0.2690625, 0.076234375, 0.6255625, 0.4910312500000001], "iscrowd": 0, "area": 31552.084050000023, "rle": {"size": [640, 640], "counts": "nl\\32nc00O100O101N100O100O2O0O100O2O0O100O101N100O100O10000O100O1fNAS_O?l`0Jk^O7Ua01c^OO\\a03b^OOB@`a0a0n^OO@Bba0`0l^O:Sa0Hl^O8a0YOW??W@9`0_OT?9[@8>BV?7[@89G\\?2Z@77K^?OY@69L^?NY@68M_?NX@58Na?LW@68Na?LX@57Oa?MX@450c?LY@52Of?KX@71Of?KY@7ONh?KZ@7MNj?KY@8JNm?JZ@8HNn?J[@9DNQ`0JY@a1h?^NT@f1S?TOh@Q1S?P2M3L4L4M3jBPK\\;P5cDSKZ;m4eDXKX;h4gD[KW;e4hD_KU;b4iDaKU;_4jDeKS;[4lDhKR;X4mDlKo:T4PEoKo:Q4PEPLP;P4nDRLR;n3lDTLT;m3hDVLX;k3eDWL[;j3bDXL];j3^DZLb;g3UDaLk;k5O001O1O1N101O00000001O3N2M2N1O1OO1N2N2O1N2N2N2O1N2N200000000001O00000000100O4L3M<D7I2M3N2N2N1O2N0000001O0000001O0000001O00000J7M2N2N2N3M2N2N2N3N1N101N2N100O2O1N2N2M3M3F:G:O01O1O1O1O1O00O2O0O100YLPAb06Jk>Do@a09Hi>FPA`0:Hf>GQA`0=Ed>IQA??F`>JRA?a0D^>LSA>b0D[>MTA=f0BW>0UA<k0]OR>6SA<S1VOk==TA;Y1POd=d0TA:a1jN\\=l0TA8n`0HR_O6Qa0IP_O5Qa0Ko^O4Ra0Lo^O1Sa0Om^O0Ta00m^ONUa00l^O0Ta00m^OOSa01m^OOSa01n^OMTa02l^ONTa02m^OMSa02n^OMSa03n^OLRa04n^OLSa03n^OKSa05o^OIQa07R_OFn`0:U_OBl`0>\\1O0000001O001O1O1O2M2O1N3M2O1N3N1N]le4"}, "label": "a small plant in a planter set on a wooden table"}]}
{"id": 99086, "image": "COCO_train2014_000000099086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young zebra running in between two other zebras\"."}], "task": "refering", "answer_template": "The \"a young zebra running in between two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 144, 145, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 238, 240, 257, 258, 260, 261, 263, 281, 283, 286, 287], "bbox": [0.175359375, 0.3408541666666667, 0.57940625, 0.7414166666666666], "iscrowd": 0, "area": 19725.107650000013, "rle": {"size": [480, 640], "counts": "^jd11n>2M3N2N2N1O1O1O0kMDgE=W:GfE:Y:IcE8\\:JcE6\\:LaE6_:J_E8`:J^E7b:J[E8d:I[E8e:IXE9h:HVE9j:HSE9n:GQE9P;HmD:T;FjD:W;FgD<Y;EYDf0i;[OWDa0l;_OTD>P<BPD;R<FmC9T<GmC7T<IlC6U<JkC6U<JjC6X<IgC7Z<IeC8[<HdC9\\<FeC:[<FdC<Z<EeC<Z<EeC<[<DdC=V1mNb9f0YE=S1RO`9`0^E?o0VO`9;aE`0m0ZO^96eE`0l0^O\\92`Eg0T1[OY9`1gFeNU9Z1kFjNT9U1TFFi9;RFKl95TFLl94SFMl93TFNl92SFNm92SFlN0El9`1SFfNV;Z1iDdNY;\\1gDbN[;_1`DdNb;[1[DgNf;W23N1O1N3O1O001N2O001O1O1O010O1O1O010O1O001O1O100O1O1O1O1O1O1O100O1O1001_EUNk7k1SHWNm7i1QHYNo7g1nG\\NR8e1kG]NU8c1iG_NW8a1gGaNY8`1eGaN[8_1cGcN^8\\1aGeN_8[1`GdNc8]1ZG`Nj8b1SG[NQ9f1mFWNX9i1fFTN^9l1aFSNb9m1\\FRNf9n1YFQNj9n1TFRNn9o1PFPNR:P2lEPNW:o1hEPNZ:P2dEPN_:o1`EPNb:P2]EoMf:Q2WEoMk:Q2TEnMn:R2PEnMS;Q2kDoMW;Q2gDoM[;Q2bDPNa;o1]DRNd;n1ZDSNg;m1WDTNk;k1SDUNo;]21O2M2O1O1O2M2O1O2N1O1O2N1O101O00000O1000O10O10O10000O101O0O2O1O0O2O1L3N3M2O1O006J7J4K4M2N2M;F5K4K5L4K6KoMTFROh9l0iFjNQ9V1TGiNi8U1]GgNc8X1bGdN]8[1hGcNW8[1nGbNR8]1RHaNl7^1YH_Ng7`1]H^Nb7`1bH^N]7b1fH\\NZ7Z1RIeNl6Y1YIeNf6\\1\\IcNc6]1_IaN`6_1cI`N\\6`1fI^NY6b1jI\\NU6e1PJWNo5h1d31O01OM3N2M22O1O1N2O1O1N3N1001O00010O000010O00010O0010O01O10O01O100O2N1O2O3L7I5M001N100O10000N2N2N2M4M2N2N2N3Knkm3"}, "label": "a young zebra running in between two other zebras"}]}
{"id": 99086, "image": "COCO_train2014_000000099086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the middle of the picture\"."}], "task": "refering", "answer_template": "The \"the zebra in the middle of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 144, 145, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 238, 240, 257, 258, 260, 261, 263, 281, 283, 286, 287], "bbox": [0.175359375, 0.3408541666666667, 0.57940625, 0.7414166666666666], "iscrowd": 0, "area": 19725.107650000013, "rle": {"size": [480, 640], "counts": "^jd11n>2M3N2N2N1O1O1O0kMDgE=W:GfE:Y:IcE8\\:JcE6\\:LaE6_:J_E8`:J^E7b:J[E8d:I[E8e:IXE9h:HVE9j:HSE9n:GQE9P;HmD:T;FjD:W;FgD<Y;EYDf0i;[OWDa0l;_OTD>P<BPD;R<FmC9T<GmC7T<IlC6U<JkC6U<JjC6X<IgC7Z<IeC8[<HdC9\\<FeC:[<FdC<Z<EeC<Z<EeC<[<DdC=V1mNb9f0YE=S1RO`9`0^E?o0VO`9;aE`0m0ZO^96eE`0l0^O\\92`Eg0T1[OY9`1gFeNU9Z1kFjNT9U1TFFi9;RFKl95TFLl94SFMl93TFNl92SFNm92SFlN0El9`1SFfNV;Z1iDdNY;\\1gDbN[;_1`DdNb;[1[DgNf;W23N1O1N3O1O001N2O001O1O1O010O1O1O010O1O001O1O100O1O1O1O1O1O1O100O1O1001_EUNk7k1SHWNm7i1QHYNo7g1nG\\NR8e1kG]NU8c1iG_NW8a1gGaNY8`1eGaN[8_1cGcN^8\\1aGeN_8[1`GdNc8]1ZG`Nj8b1SG[NQ9f1mFWNX9i1fFTN^9l1aFSNb9m1\\FRNf9n1YFQNj9n1TFRNn9o1PFPNR:P2lEPNW:o1hEPNZ:P2dEPN_:o1`EPNb:P2]EoMf:Q2WEoMk:Q2TEnMn:R2PEnMS;Q2kDoMW;Q2gDoM[;Q2bDPNa;o1]DRNd;n1ZDSNg;m1WDTNk;k1SDUNo;]21O2M2O1O1O2M2O1O2N1O1O2N1O101O00000O1000O10O10O10000O101O0O2O1O0O2O1L3N3M2O1O006J7J4K4M2N2M;F5K4K5L4K6KoMTFROh9l0iFjNQ9V1TGiNi8U1]GgNc8X1bGdN]8[1hGcNW8[1nGbNR8]1RHaNl7^1YH_Ng7`1]H^Nb7`1bH^N]7b1fH\\NZ7Z1RIeNl6Y1YIeNf6\\1\\IcNc6]1_IaN`6_1cI`N\\6`1fI^NY6b1jI\\NU6e1PJWNo5h1d31O01OM3N2M22O1O1N2O1O1N3N1001O00010O000010O00010O0010O01O10O01O100O2N1O2O3L7I5M001N100O10000N2N2N2M4M2N2N2N3Knkm3"}, "label": "the zebra in the middle of the picture"}]}
{"id": 99086, "image": "COCO_train2014_000000099086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back half of the zebra\"."}], "task": "refering", "answer_template": "The \"the back half of the zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 272, 273, 275, 293, 295, 296, 298], "bbox": [0.72465625, 0.38627083333333334, 1.0, 0.757875], "iscrowd": 0, "area": 16264.259549999997, "rle": {"size": [480, 640], "counts": "eii61l>3GO`A3_>7O2O0O1hBAo;a0nCCP<=nCHn;9QDKk;7RDMl;3RD2j;0TD4h;MVD7g;JWD;f;EYD?c;A\\Dc0a;]O_De0`;ZO_Di0_;XO_Dk0_;UOaDm0^;RO`DS1];mNZD^1d;bN^D^1`;cN`Di0NmNa;9bDh03lNY;<eDf09jNP;?iDd0>kNe:a0PEa0b0jN\\:e0SE?Q<@QD>o;BSD<n;CTD;m;DUD:k;FXD7i;HYD6h;IZD7d;I^D7a;I`D7_;L`D3^;2_DN`;7\\DIb;=ZDCe;a0XD_Og;f0QD^Om;k1O1O1001N2O001O001O001O001OTD`MY;`2fDcMY;\\2fDhMW;X2hDlMV;S2iDQNT;n1lDVNR;i1mD[NQ;h1iD\\NU;h1fD\\NX;h2M3N2M3N1N3N2M3N1N3N2MQObE_M]:`2fE`MX:_2jEbMS:^2oEbMP:]2RFdMk9\\2WFeMg9Z2[FgMb9Y2aFfM^9Z2cFgMZ9Y2hFhMV9W2lFiMR9W2PGjMn8U2UGjMi8V2YGiMg8V2[GiMd8X2]GgMd8W2^GiMa8V2bGhM^8W2dGhM\\8X2eGgM[8X2gGeM\\8Y2fGeM[8[2fGbM\\8]2gG`MZ8_2iG]MY8b2jG[MX8d2jGYMW8f2lGVMV8i2nGSMS8m2P2O2N1O100O1O20O1O1O1O100O1O1O1O01O0001O0001O0000000010O0000000001O0001O0001O0000000010O00000001O000000010O000000001O0001O0gEZMX8f2bG`M_8_2[GgMe8Y2UGmMk8S2dF^N\\9b1RFoNo9Q3o1bMWH"}, "label": "the back half of the zebra"}]}
{"id": 99086, "image": "COCO_train2014_000000099086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra leading other one in the forest\"."}], "task": "refering", "answer_template": "The \"a zebra leading other one in the forest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 272, 273, 275, 293, 295, 296, 298], "bbox": [0.72465625, 0.38627083333333334, 1.0, 0.757875], "iscrowd": 0, "area": 16264.259549999997, "rle": {"size": [480, 640], "counts": "eii61l>3GO`A3_>7O2O0O1hBAo;a0nCCP<=nCHn;9QDKk;7RDMl;3RD2j;0TD4h;MVD7g;JWD;f;EYD?c;A\\Dc0a;]O_De0`;ZO_Di0_;XO_Dk0_;UOaDm0^;RO`DS1];mNZD^1d;bN^D^1`;cN`Di0NmNa;9bDh03lNY;<eDf09jNP;?iDd0>kNe:a0PEa0b0jN\\:e0SE?Q<@QD>o;BSD<n;CTD;m;DUD:k;FXD7i;HYD6h;IZD7d;I^D7a;I`D7_;L`D3^;2_DN`;7\\DIb;=ZDCe;a0XD_Og;f0QD^Om;k1O1O1001N2O001O001O001O001OTD`MY;`2fDcMY;\\2fDhMW;X2hDlMV;S2iDQNT;n1lDVNR;i1mD[NQ;h1iD\\NU;h1fD\\NX;h2M3N2M3N1N3N2M3N1N3N2MQObE_M]:`2fE`MX:_2jEbMS:^2oEbMP:]2RFdMk9\\2WFeMg9Z2[FgMb9Y2aFfM^9Z2cFgMZ9Y2hFhMV9W2lFiMR9W2PGjMn8U2UGjMi8V2YGiMg8V2[GiMd8X2]GgMd8W2^GiMa8V2bGhM^8W2dGhM\\8X2eGgM[8X2gGeM\\8Y2fGeM[8[2fGbM\\8]2gG`MZ8_2iG]MY8b2jG[MX8d2jGYMW8f2lGVMV8i2nGSMS8m2P2O2N1O100O1O20O1O1O1O100O1O1O1O01O0001O0001O0000000010O0000000001O0001O0001O0000000010O00000001O000000010O000000001O0001O0gEZMX8f2bG`M_8_2[GgMe8Y2UGmMk8S2dF^N\\9b1RFoNo9Q3o1bMWH"}, "label": "a zebra leading other one in the forest"}]}
{"id": 49933, "image": "COCO_train2014_000000049933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white playing tennis\"."}], "task": "refering", "answer_template": "The \"a woman in white playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.449796875, 0.07192037470725995, 1.0, 1.0], "iscrowd": 0, "area": 84677.73174999999, "rle": {"size": [427, 640], "counts": "nUh34V=6K2M1O2N1O2N101N101O0O2O0O10000O10O01000O0100O10O11N2O2QD\\OR;f0`DJ];T1N3L3ZGlMiM6_4LYNU2XOcN[2ZOYNY2UOdN_2UOWN^2QOeNd2oNWNb2mNhNh2gNXNg2jNgNj2eNYNi2gNgNm2bNYNl2dNgNo2aNYNk2dNiNP3`NVNl2eNiNS3]NTNn2dNjNW3XNRNQ3cNlNY3RNSNS3dNnNW3nMTNV3cNPOW3iMUNY3cNQOU3fMWN[3bNSOU3aMXN]3bNVOS3]MZN_3bNWOQ3ZM\\N`3bN^Ok2RMbNa3bNFj1ULjNc0h0c3eNLb1oKPOb0h0d3eN4[1hKVO`0i0d3fN;e0RLKOi0e3gN?7WL8Ei0f3gNm1?\\Ji0g3hNo1>ZJj0g3hNQ2;YJm0f3gNR2;YJm0f3hNR29YJo0V:POkEP1V:nNkEQ1V:oNiER1X;0000O1000000O1000000O1000000O1000000O11O00000000000000001O00001O00001O001O001O001O0O2O1O001O001O_O[D\\Oe;e0]DWOd;i0^DTOc;l0`DQO`;o0bDnN_;S1<00004L6J5K5K1N1000O10000O10000O01000O100O100001O1N101O1O001O1O001O1O0O2O1O2iFmMR7U2iHPNU7S2eGiM87Q8S2dGkM56U8Q2cGTNK0`8o1aGRO\\8Q1aGSO\\8o0bGSO\\8o0aGUO[8o0aGTO]8n2O0O2O001N1000WKiG_4U8aKlG_4T8_KnGa4R8]KPHd4n7[KTHe4l7ZKUHg4j7WKXHi4U8001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O0000000000001O00001O000UHjJa7V5^HlJa7U5^HkJb7U5]HmJb7S5]HoJb7^5O001O1O1O001O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O"}, "label": "a woman in white playing tennis"}]}
{"id": 49933, "image": "COCO_train2014_000000049933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the female in white ready to paly tennis\"."}], "task": "refering", "answer_template": "The \"the female in white ready to paly tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.449796875, 0.07192037470725995, 1.0, 1.0], "iscrowd": 0, "area": 84677.73174999999, "rle": {"size": [427, 640], "counts": "nUh34V=6K2M1O2N1O2N101N101O0O2O0O10000O10O01000O0100O10O11N2O2QD\\OR;f0`DJ];T1N3L3ZGlMiM6_4LYNU2XOcN[2ZOYNY2UOdN_2UOWN^2QOeNd2oNWNb2mNhNh2gNXNg2jNgNj2eNYNi2gNgNm2bNYNl2dNgNo2aNYNk2dNiNP3`NVNl2eNiNS3]NTNn2dNjNW3XNRNQ3cNlNY3RNSNS3dNnNW3nMTNV3cNPOW3iMUNY3cNQOU3fMWN[3bNSOU3aMXN]3bNVOS3]MZN_3bNWOQ3ZM\\N`3bN^Ok2RMbNa3bNFj1ULjNc0h0c3eNLb1oKPOb0h0d3eN4[1hKVO`0i0d3fN;e0RLKOi0e3gN?7WL8Ei0f3gNm1?\\Ji0g3hNo1>ZJj0g3hNQ2;YJm0f3gNR2;YJm0f3hNR29YJo0V:POkEP1V:nNkEQ1V:oNiER1X;0000O1000000O1000000O1000000O1000000O11O00000000000000001O00001O00001O001O001O001O0O2O1O001O001O_O[D\\Oe;e0]DWOd;i0^DTOc;l0`DQO`;o0bDnN_;S1<00004L6J5K5K1N1000O10000O10000O01000O100O100001O1N101O1O001O1O001O1O0O2O1O2iFmMR7U2iHPNU7S2eGiM87Q8S2dGkM56U8Q2cGTNK0`8o1aGRO\\8Q1aGSO\\8o0bGSO\\8o0aGUO[8o0aGTO]8n2O0O2O001N1000WKiG_4U8aKlG_4T8_KnGa4R8]KPHd4n7[KTHe4l7ZKUHg4j7WKXHi4U8001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O0000000000001O00001O000UHjJa7V5^HlJa7U5^HkJb7U5]HmJb7S5]HoJb7^5O001O1O1O001O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O"}, "label": "the female in white ready to paly tennis"}]}
{"id": 287507, "image": "COCO_train2014_000000287507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle piece of the chicken roll up\"."}], "task": "refering", "answer_template": "The \"the middle piece of the chicken roll up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 186, 187, 188, 189, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 289, 290, 291, 292], "bbox": [0.27214622641509434, 0.5130625, 0.6180896226415095, 0.86640625], "iscrowd": 0, "area": 22304.935849999994, "rle": {"size": [640, 424], "counts": "]\\X22jc0<C6K5K5J7J5K5J6L4K6K4K5K5L3L4L4M3M4L3M3N2M4L3N2M3M3N3L3M3N2M4L3N2M3M3N3L3N2N2N3M2N2N1O2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O0000001O0002N1O100O1O2N100O1O1O2O0O1O2N101M2N3M3M2O2M2N3M2N3M2M4J6I6J7J5J7L3M4L4K4M4L3M4L3M4L4L3M4L3L5L3M4K5K4K6K:E<Df0[OUlT3"}, "label": "the middle piece of the chicken roll up"}]}
{"id": 287507, "image": "COCO_train2014_000000287507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chicken nearest the broccoli\"."}], "task": "refering", "answer_template": "The \"the chicken nearest the broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 170, 171, 172, 173, 184, 185, 186, 187, 199, 200, 201, 214, 215, 229, 230, 244], "bbox": [0.2632311320754717, 0.44296875, 0.5547877358490566, 0.703921875], "iscrowd": 0, "area": 8531.072300000002, "rle": {"size": [640, 424], "counts": "i[V2n0Pc0;D;F;E9G3L4M2N3L4M2N3N2M3N1O2N2M2O2N1O1N101O001N101O0ON4J5L4K5K5L4N3N100O100O101O0O100O101N100O100O100O101N100O10000000001O000O10001O001O001O001O001O1O001O001N101O001O001O001O100O010O100O10O10O100O10O0100O100O1N101N2N2N2N2N2N2N2N2M3M3M3M3M3M6JTke3"}, "label": "the chicken nearest the broccoli"}]}
{"id": 287507, "image": "COCO_train2014_000000287507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of the wrap that is next to broccoli\"."}], "task": "refering", "answer_template": "The \"a slice of the wrap that is next to broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 170, 171, 172, 173, 184, 185, 186, 187, 199, 200, 201, 214, 215, 229, 230, 244], "bbox": [0.2632311320754717, 0.44296875, 0.5547877358490566, 0.703921875], "iscrowd": 0, "area": 8531.072300000002, "rle": {"size": [640, 424], "counts": "i[V2n0Pc0;D;F;E9G3L4M2N3L4M2N3N2M3N1O2N2M2O2N1O1N101O001N101O0ON4J5L4K5K5L4N3N100O100O101O0O100O101N100O100O100O101N100O10000000001O000O10001O001O001O001O001O1O001O001N101O001O001O001O100O010O100O10O10O100O10O0100O100O1N101N2N2N2N2N2N2N2N2M3M3M3M3M3M6JTke3"}, "label": "a slice of the wrap that is next to broccoli"}]}
{"id": 140053, "image": "COCO_train2014_000000140053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"big elaphant in the right\"."}], "task": "refering", "answer_template": "The \"big elaphant in the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 28, 29, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 173, 174], "bbox": [0.54298, 0.003950177935943061, 1.0, 0.938505338078292], "iscrowd": 0, "area": 36173.411499999995, "rle": {"size": [281, 500], "counts": "adZ22g82N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N1O2N1O2N1O2N2N1O2QNRNoLo1n2UNPMm1m2VNQMk1l2YNRMj1i2ZNUMh1h2[NVMg1f2\\NYMg1b2]N\\Me1`2_N^Mc1^2eN\\M^1_2mNXMU1d2VOSMP1f2ROYMU1^2lNcM[1T2gNlM`1k1bNUNd1c1^N]Nm1V1TNkN^3AdL?k50O1000000O1000000O1000000O1000000O1000000O1000000O100000000O100000000O10000000000O10000000000O100000000O1000000]KUOX1k0gNZOU1f0kN_Oo0c0POAdN_O>P1n0FQNNk0<T1K]M=Z1HY1l1aNTN_1Q2[NPNe1V2UNjMj1[2QNfMo1^2mMbMS2S2oLUMj0i0W2Q2SMVMb0j0Z2P2XMWM:i0^2o1\\MXM2j0b2n1_MXMKk0f2l1dMYMBk0i2l1iMYMZOl0m2j1mMZMSOl0P3i1QNbNQ2[OUMi0o0Hn1ZO]Mf0i0Ml1WOfMd0b02i1UOoMb0;7h1ROWN?1`0j1kN`NO5V1^5100O10000O100O100O10000O100O1000000O1000000O1000000O10000O1000000O10000000000000000000000XOjNmIV1l5o0H8I7I7H8I7I7H8I7H8I7I7H8I7I7fN[JXOm5g0_JgNh5X1m0O1O1O100O1O1O1O1N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N0"}, "label": "big elaphant in the right"}]}
{"id": 140053, "image": "COCO_train2014_000000140053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant to the far left\"."}], "task": "refering", "answer_template": "The \"elephant to the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 24, 25, 36, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 72, 73, 75, 76, 77, 78, 90, 91, 93, 94, 95, 96, 108, 109, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 144, 145, 147, 148], "bbox": [0.0, 0.004626334519572954, 0.436, 0.8611387900355872], "iscrowd": 0, "area": 35057.90305, "rle": {"size": [281, 500], "counts": "\\6o0m5n1QNo1RNn1\\Od0O001O1O1O00O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O100O1]MgNdM[1`1KSN8_1?VNCk1e0jM]OV2m0]MVOd2R1QMPOo2]3001O001O00001O000000O100O100O100O100O1O100O100O100O100OVNnLoMR3n1RMQNm2m1VMSNi2k1[MTNc2j1aMVN^2h1fMWNY2g1jMYNT2f1PNYNo1d1UN\\Nj1b1YN^Ne1a1_N^N`1`1cN`N\\1^1hNaNV1]1nNcNQ1[1SOdNl0Z1WOfNh0X1[OhNc0Km04l301O001O001O00001O001O001O00001O001O001O001O00N2N2N2N2N200O1O100O1O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1nJEe1<XN\\12eNKV5M3O1001O0102N2M3N1N1000O0100O01000O010O10O10O10O01O1O01O000001O000O1Ld[]2"}, "label": "elephant to the far left"}]}
{"id": 140053, "image": "COCO_train2014_000000140053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with the large tusks\"."}], "task": "refering", "answer_template": "The \"elephant with the large tusks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 24, 25, 36, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 72, 73, 75, 76, 77, 78, 90, 91, 93, 94, 95, 96, 108, 109, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 144, 145, 147, 148], "bbox": [0.0, 0.004626334519572954, 0.436, 0.8611387900355872], "iscrowd": 0, "area": 35057.90305, "rle": {"size": [281, 500], "counts": "\\6o0m5n1QNo1RNn1\\Od0O001O1O1O00O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O100O1]MgNdM[1`1KSN8_1?VNCk1e0jM]OV2m0]MVOd2R1QMPOo2]3001O001O00001O000000O100O100O100O100O1O100O100O100O100OVNnLoMR3n1RMQNm2m1VMSNi2k1[MTNc2j1aMVN^2h1fMWNY2g1jMYNT2f1PNYNo1d1UN\\Nj1b1YN^Ne1a1_N^N`1`1cN`N\\1^1hNaNV1]1nNcNQ1[1SOdNl0Z1WOfNh0X1[OhNc0Km04l301O001O001O00001O001O001O00001O001O001O001O00N2N2N2N2N200O1O100O1O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1nJEe1<XN\\12eNKV5M3O1001O0102N2M3N1N1000O0100O01000O010O10O10O10O01O1O01O000001O000O1Ld[]2"}, "label": "elephant with the large tusks"}]}
{"id": 140053, "image": "COCO_train2014_000000140053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest elephant in the image\"."}], "task": "refering", "answer_template": "The \"the smallest elephant in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 136, 151, 152, 153, 154, 171, 172], "bbox": [0.37594, 0.26907473309608543, 0.5963999999999999, 0.9266903914590746], "iscrowd": 0, "area": 14771.680699999997, "rle": {"size": [281, 500], "counts": "Wgc1;j7h0D;J7I6J7I6J6J7I7I7K4M4K5L4K6K6I`0A2M2O2M2O1N3N1N2OO00100O10O01O100O1O001001O1O1O001O1O1O1O001O001O1O001O001O001O1O2N0000]NfLhMZ3S2lLkMT3Q2RMmMn2T2SMjMm2W2UMgMj2Z2WMdMi2]2YM`Mg2b2YM]Mf2d2\\MYMe2g2\\MVMe2k2]MQMe2o2\\MnLf2R3]MhLe2Z3`MZLf2g3V11N2O1O1N>[K\\LZ3]45L4K5KYLfKX3m42N1O100O1O2N3M2N2N3M2N2N3L3M3M4XNg1E<D<J6I8I6G:D;E`Yg1"}, "label": "the smallest elephant in the image"}]}
{"id": 140053, "image": "COCO_train2014_000000140053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant between two adult elephants\"."}], "task": "refering", "answer_template": "The \"baby elephant between two adult elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 136, 151, 152, 153, 154, 171, 172], "bbox": [0.37594, 0.26907473309608543, 0.5963999999999999, 0.9266903914590746], "iscrowd": 0, "area": 14771.680699999997, "rle": {"size": [281, 500], "counts": "Wgc1;j7h0D;J7I6J7I6J6J7I7I7K4M4K5L4K6K6I`0A2M2O2M2O1N3N1N2OO00100O10O01O100O1O001001O1O1O001O1O1O1O001O001O1O001O001O001O1O2N0000]NfLhMZ3S2lLkMT3Q2RMmMn2T2SMjMm2W2UMgMj2Z2WMdMi2]2YM`Mg2b2YM]Mf2d2\\MYMe2g2\\MVMe2k2]MQMe2o2\\MnLf2R3]MhLe2Z3`MZLf2g3V11N2O1O1N>[K\\LZ3]45L4K5KYLfKX3m42N1O100O1O2N3M2N2N3M2N2N3L3M3M4XNg1E<D<J6I8I6G:D;E`Yg1"}, "label": "baby elephant between two adult elephants"}]}
{"id": 797, "image": "COCO_train2014_000000000797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on left\"."}], "task": "refering", "answer_template": "The \"elephant on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 81, 82, 83, 84, 85, 86, 92, 93, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 231, 233, 234, 235, 236, 237], "bbox": [0.0, 0.2326388888888889, 0.752703125, 0.8315277777777779], "iscrowd": 0, "area": 48404.744149999984, "rle": {"size": [360, 640], "counts": "i3S4T7O2M3N2M33gHVLn6R401VInK\\6R4]IULc6Y4O001O1O001N2O001O001O001O2O0O10M2N3M2O2O0O1O2O05L7H10O0H9M21O100O010O100O0010J_IdK`6\\4cIbK]6]4eIaK[6^49O001O000O27HO2N1O2N1O2N1O1101N010OJWImKj6Q47O2M2O1O2N1O1N21O2N2N10O0000000001O00O100001O:F1O1O002N2N1O010O000000M2N1N3M23N2M2O2M3N2M2O2M3N1N3N2N101N3M2O1N2O2M2OO0O1O2M2O2N100O2O0O2O000O2O001O01O3N2M3N2M3N2M2O2N1N01N2O001O1O001O10O10000O10000O10001N10001N=D6J3M2M1000O010000O0100O0O2N2N2N2N200000O101O1O1O2N1O1O1O1O0O2O001O001O00SOZKWKg4d4bKWK_4e4hKWKX4e4PLWKQ4d4XLVKj3f4^1K5L4L40000000OH9O1O1N2O1O1O1O1O001N2O1O1N2O1N2O1O1N2O2O000O101O0O10001N10001N100O2O0O101N100O2N100O2O000O2O00001O0O101O00001O0O101VO`GjN`8V1aGiN_8W1aGiN`8W1`GgNa8Y1`GfN`8[1`GcNb8\\1_GcNa8^1^GaNc8_1^G`Nc8`1]G_Nc8a1^G]Nc8d1]G[Nd8d1]GZNd8g1\\GXNd8h1?0000O10O010000O10O01000O0010O01O1O001O001OO2N100O1O1O100O2N0O2L4L4L3M4L4M3O001O1N2O001O1N2O001N2O3M4L4K4M4Lh:N[E1N3N1N2O1O2M2O1N2O2M100O010O1000O0100O100O01000O10O10O1000O010000O10O10O1000O010000O0100000O01000O10O1000O10O1000O10O10O100000O1000000O010000000O100000000O1000O10000000O1000000000001O000O2O00001O00001O000O101O00001O000010O0001O0000010O001O001O010O001N1Kd^g1"}, "label": "elephant on left"}]}
{"id": 797, "image": "COCO_train2014_000000000797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the left\"."}], "task": "refering", "answer_template": "The \"the elephant in the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 81, 82, 83, 84, 85, 86, 92, 93, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 231, 233, 234, 235, 236, 237], "bbox": [0.0, 0.2326388888888889, 0.752703125, 0.8315277777777779], "iscrowd": 0, "area": 48404.744149999984, "rle": {"size": [360, 640], "counts": "i3S4T7O2M3N2M33gHVLn6R401VInK\\6R4]IULc6Y4O001O1O001N2O001O001O001O2O0O10M2N3M2O2O0O1O2O05L7H10O0H9M21O100O010O100O0010J_IdK`6\\4cIbK]6]4eIaK[6^49O001O000O27HO2N1O2N1O2N1O1101N010OJWImKj6Q47O2M2O1O2N1O1N21O2N2N10O0000000001O00O100001O:F1O1O002N2N1O010O000000M2N1N3M23N2M2O2M3N2M2O2M3N1N3N2N101N3M2O1N2O2M2OO0O1O2M2O2N100O2O0O2O000O2O001O01O3N2M3N2M3N2M2O2N1N01N2O001O1O001O10O10000O10000O10001N10001N=D6J3M2M1000O010000O0100O0O2N2N2N2N200000O101O1O1O2N1O1O1O1O0O2O001O001O00SOZKWKg4d4bKWK_4e4hKWKX4e4PLWKQ4d4XLVKj3f4^1K5L4L40000000OH9O1O1N2O1O1O1O1O001N2O1O1N2O1N2O1O1N2O2O000O101O0O10001N10001N100O2O0O101N100O2N100O2O000O2O00001O0O101O00001O0O101VO`GjN`8V1aGiN_8W1aGiN`8W1`GgNa8Y1`GfN`8[1`GcNb8\\1_GcNa8^1^GaNc8_1^G`Nc8`1]G_Nc8a1^G]Nc8d1]G[Nd8d1]GZNd8g1\\GXNd8h1?0000O10O010000O10O01000O0010O01O1O001O001OO2N100O1O1O100O2N0O2L4L4L3M4L4M3O001O1N2O001O1N2O001N2O3M4L4K4M4Lh:N[E1N3N1N2O1O2M2O1N2O2M100O010O1000O0100O100O01000O10O10O1000O010000O10O10O1000O010000O0100000O01000O10O1000O10O1000O10O10O100000O1000000O010000000O100000000O1000O10000000O1000000000001O000O2O00001O00001O000O101O00001O000010O0001O0000010O001O001O010O001N1Kd^g1"}, "label": "the elephant in the left"}]}
{"id": 797, "image": "COCO_train2014_000000000797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"romance of african elephsnt\"."}], "task": "refering", "answer_template": "The \"romance of african elephsnt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298], "bbox": [0.5133125, 0.28775, 1.0, 0.967111111111111], "iscrowd": 0, "area": 49669.1678, "rle": {"size": [360, 640], "counts": "Skc38i:;F9G:E:G:H7L5J5L5J5L5K4K6K3L2O2M3N1TOdLjI_3U6dLgI^3W6fLeI[3[6hLaIZ3`6gL\\I[3d6fLYI[3h6c0001N10000O101O0nMTIaNM<P7n0RJQOm5j0YJVOg5d0`JZO`5b0eJ^OZ5c0gJ\\OY5c0hJ\\OX5e0iJZOV5g0kJXOT5h0mJWOR5k0oJTOP5m0oJSOQ5m0PKROP5m0RKROm4o0TKPOl4P1UKoNk4Q1VKoNi4P1YKoNf4R1[KmNe4S1[KmNe4R1]KmNc4S1^KlNa4U1`KjN`4U1bKkN]4U1dKjN\\4V1cI\\Nf1=f4W1cI`Ne19h4W1aIcNg15h4X1^IgNj10g4Z1]IiNl1Lg4n1XKRNh4o1WKRNh4P2VKPNi4Q2WKoMi4R2VKnMj4R2VKnMj4R2VKnMi4T2WKkMi4U2WKlMh4T2XKlMh4U2WKkMh4V2XKjMh4V2XKjMh4W2WKiMi4W2WKiMh4X2YKgMg4Z2XKgMg4Y2YKgMg4Z2XKfMg4\\2XKdMh4]2WKcMi4^2VKbMj4_2UKaMj4a2VK^Mj4c2UK]Mk4d2TK\\Mk4f2TKYMm4h2RKXMn4h2RKXMn4i2QKVMo4k2QKUMo4k2RKTMn4m2QKSMo4m2RKQMn4Q3RKnLn4R3SKmLm4S3TKlLl4U3TKiLl4X3UKgLk4Z3UKeLk4[3a10O2O001O00001O00001]HhLT7h3N2N2N3M2N2N3M0000O10O10000000O100000000001O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N3gKdJ`2_5[McJe2`5VMcJi2`5RMbJn2a5lLcJS3`5hLcJW3`5dLbJ]3`5^LcJa3a5YLbJf3a5ULaJk3b5PLaJo3]6O0010O01000O0100O010O10O10N10000000O10O10000000000O10000000000O10000000O1000000000000O1000000O10000O10000O10001O0O1001O001O001O001O001O000001O010O001O1O001YJQLS4o3jKULU4l3gKWLY4i3gKXLX4i3fKXLZ4h3fKYLZ4g3eKYL[4g3eKYL[4h3dKYL[4g3dKZL\\4g3cKZL\\4f3dKZL\\4g3cKZL\\4f3dKZL\\4g3cKZL\\4f3cK[L]4f3`K]L_4c3_K_Lb4a3[KbLd4^3ZKdLf4]3WKfLh4Z3UKiLk4d3fJ_LY5h40001O001O010O001O001O001O00001O00001N10001O00001O001N10O10000O1000O010000O10000O01000O10000O1m1RN`H"}, "label": "romance of african elephsnt"}]}
{"id": 541472, "image": "COCO_train2014_000000541472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple with a union jack sticker\"."}], "task": "refering", "answer_template": "The \"an apple with a union jack sticker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.43059375, 0.4740654205607477, 0.8059687499999999, 0.9858411214953271], "iscrowd": 0, "area": 34080.188700000006, "rle": {"size": [428, 640], "counts": "Sec39n<9F8N2N1O2N000N3N1\\DPOj:R1VEPOg:Q1YEROb:P1]ETO_:m0aEUO[:n0dEUOX:l0hEWOS:k0lEXOQ:i0oEZOl9h0TF[Oh9g0WF\\Od9f0\\F\\Oa9e0^F_O]9c0cF@Y9a0gF[NAo0e9i0iFVNDR1a9i0jFTNGT1\\9j0mFQNHV1X9k0WGVOg8k0YGVOd8l0[GVOb8m0]GTOa8m0_GTO^8n0bGSO\\8n0dGSOY8o0gGROW8P1hGQOU8Q1kGoNS8S1lGoNR8R1nGnNP8T1PHmNn7U1QHkNm7W1SHjNk7W1UHiNi7Y1WHhNg7Y1YHgNe7\\1ZHeNc7]1\\HdNc7]1]HdNa7]1_HcNa7]1_HdNa7[1_HeNa7[1_HfN`7Z1`HfN`7Z1`HgN_7Y1`HhN`7X1`HiN_7W1aHiN_7W1aHjN_7U1aHkNl6RNYIS3KlNj6SNZIQ3LlNj6TNYIP3MmNi6TNYIn2OnNg6UNZIm2NnNi6VNXIl2OnNi6VNYIk2NoNi6VNYIk2NoNh6WNZIj2NoNh6WNZIj2NnNi6XNYIj2NnNh6YNZIi2NnNh6YNZIi2NnNh6YN[Ih2LPOh6YN\\Ig2LoNi6ZN[Ig2LoNi6ZN[Ig2LoNi6ZN[Ig2LoNi6ZN[Ig2LoNh6[N\\If2LnNi6\\N\\Ie2KoNi6\\N\\Ie2JPOj6[N\\Ie2JPOj6[N\\Ie2JPOi6\\N]Id2JoNj6]N]Ic2IPOj6]N]Ic2IPOj6]N]Ic2IPOj6]N]Ic2IPOi6^N_Ia2HPOj6_N^Ia2GQOk6]N_Ib2FQOk6]N_Ib2FQOk6]N_Ib2FQOj6^N`Ia2FPOk6_N_Ia2FPOk6_N_I`2GQOj6_N_I`2GQOi6`N`I_2FROj6_N`I_2FROj6_N`I_2FROj6_N`I_2FQOj6aN`I^2FQOj6aN`I^2FQOj6aN`I^2FQOi6bNaI]2FQOi6bNaI]2FQOi6bNaI]2EROj6aNaI]2EQOj6cNaI\\2EQOj6cNaI\\2EQOj6cNaI[2FROi6cNaI[2FROi6cNaI[2FROi6cNaI[2FQOj6dN`I[2FQOj6dN`IZ2GROi6dN`IZ2GROi6dN`IZ2GROi6dN`IZ2GQOj6eN_IZ2GQOj6eN_IY2HROi6eN_IY2HROi6eN_IY2HROi6eN^IZ2IQOi6eN^IZ2IPOj6fN]IY2JQOi6fN]IY2JQOi6fN]IY2JQOi6fN]IY2JQOi6fN]IY2JPOj6gN\\IY2JPOk6fN[IZ2JPOk6fNZI[2KoNk6fNZI[2KoNk6fNZIZ2LoNk6gNYIZ2LoNk6gNYIZ2LoNk6gNYIZ2KPOl6fNXI[2LoNl6fNXI[2LnNm6gNWI[2LnNm6gNWI[2LnNn6eNVI^2LmNn6eNVI^2LmN2YN[6<gI^2LlN3[NY6;hI]2MmN2\\NX6:iI]2MmN2\\NX6:hI^2NlN1^NY67hI_2NlN1_NX66iI_2NkN2aNV65jI_2NkN2bNU64kI_2NkN1cNV63jI`2OjN1cNV63jI`2OjN1cNV63jI`2OiN2dNV62iIa2OiN1eNW61iIa2OiN1eNW61iI`20jN0eNW61iI`20iN1fNV61iI`20iN0gNW60iI`20iN0gNX6OhIa20iN0gNX6OhIa20iNOhNY6NhIa20hN0iNX6NhIa2OiN1hNX6NgIb20hN1hNY6MfIb21iNOiNZ6LfIb21iNOiNZ6LfIb21hN0jNZ6JfId20hN0jNZ6JfId20hNOkN\\6HeId21iNNlN[6FgIe20iNNlN[6FgIe2OiN0lN[6EfIf2OiNOmN\\6CgIg2NiNOmN\\6CgIg2NiNOmN]6AgIh2MkNOlN]6AgIh2MkNNmN^6@gIh2MjNOnNQ7X2PIjNOnNQ7X2oHkN0mNP7X2QIkNNnNQ7W2QIkNNnNQ7W2QIkNNnNQ7W2QIjNOoNP7W2PIkNOoNQ7V2PIkNOoNQ7U2QIkNOPOP7U2QIkNOPOP7U2PIlNOPOQ7T2PIkN0QOP7T2PIkN0POQ7T2PIlNOoNR7U2nHlN1nNR7V2mHlN1mNS7W2lHlN1lNT7X2kHkN1mNU7X2iHlN2kNV7X2iHmN1iNW7[2hHkN2iNW7\\2gHkN2gNY7^2dHlN3dN[7`2bHkN4cN\\7b2`HkN4aN^7c2_HlN3_Na7d2\\HlNX8T1gGmNZ8R1fGnN[8Q1eGoN[8Q1eGnN]8P1cGQO^8n0bGRO^8n0bGQO`8n0`GROa8l0_GUOa8k0_GTOc8k0]GUOe8i0[GWOf8h0YGYOi8d0XG\\Oj8b0VG^Ok8a0UG^On8`0QGAQ9<PGDQ9;oFES99mFGU96kFJW94jFKY91hFO[9NfF1_9JbF4e9E\\F9l9AUF=R:[OPFc0X:SOmEk0Z:lNiER1Z;I<D>BY`c1"}, "label": "an apple with a union jack sticker"}]}
{"id": 541472, "image": "COCO_train2014_000000541472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple with red white and blue sticker on the side\"."}], "task": "refering", "answer_template": "The \"apple with red white and blue sticker on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.43059375, 0.4740654205607477, 0.8059687499999999, 0.9858411214953271], "iscrowd": 0, "area": 34080.188700000006, "rle": {"size": [428, 640], "counts": "Sec39n<9F8N2N1O2N000N3N1\\DPOj:R1VEPOg:Q1YEROb:P1]ETO_:m0aEUO[:n0dEUOX:l0hEWOS:k0lEXOQ:i0oEZOl9h0TF[Oh9g0WF\\Od9f0\\F\\Oa9e0^F_O]9c0cF@Y9a0gF[NAo0e9i0iFVNDR1a9i0jFTNGT1\\9j0mFQNHV1X9k0WGVOg8k0YGVOd8l0[GVOb8m0]GTOa8m0_GTO^8n0bGSO\\8n0dGSOY8o0gGROW8P1hGQOU8Q1kGoNS8S1lGoNR8R1nGnNP8T1PHmNn7U1QHkNm7W1SHjNk7W1UHiNi7Y1WHhNg7Y1YHgNe7\\1ZHeNc7]1\\HdNc7]1]HdNa7]1_HcNa7]1_HdNa7[1_HeNa7[1_HfN`7Z1`HfN`7Z1`HgN_7Y1`HhN`7X1`HiN_7W1aHiN_7W1aHjN_7U1aHkNl6RNYIS3KlNj6SNZIQ3LlNj6TNYIP3MmNi6TNYIn2OnNg6UNZIm2NnNi6VNXIl2OnNi6VNYIk2NoNi6VNYIk2NoNh6WNZIj2NoNh6WNZIj2NnNi6XNYIj2NnNh6YNZIi2NnNh6YNZIi2NnNh6YN[Ih2LPOh6YN\\Ig2LoNi6ZN[Ig2LoNi6ZN[Ig2LoNi6ZN[Ig2LoNi6ZN[Ig2LoNh6[N\\If2LnNi6\\N\\Ie2KoNi6\\N\\Ie2JPOj6[N\\Ie2JPOj6[N\\Ie2JPOi6\\N]Id2JoNj6]N]Ic2IPOj6]N]Ic2IPOj6]N]Ic2IPOj6]N]Ic2IPOi6^N_Ia2HPOj6_N^Ia2GQOk6]N_Ib2FQOk6]N_Ib2FQOk6]N_Ib2FQOj6^N`Ia2FPOk6_N_Ia2FPOk6_N_I`2GQOj6_N_I`2GQOi6`N`I_2FROj6_N`I_2FROj6_N`I_2FROj6_N`I_2FQOj6aN`I^2FQOj6aN`I^2FQOj6aN`I^2FQOi6bNaI]2FQOi6bNaI]2FQOi6bNaI]2EROj6aNaI]2EQOj6cNaI\\2EQOj6cNaI\\2EQOj6cNaI[2FROi6cNaI[2FROi6cNaI[2FROi6cNaI[2FQOj6dN`I[2FQOj6dN`IZ2GROi6dN`IZ2GROi6dN`IZ2GROi6dN`IZ2GQOj6eN_IZ2GQOj6eN_IY2HROi6eN_IY2HROi6eN_IY2HROi6eN^IZ2IQOi6eN^IZ2IPOj6fN]IY2JQOi6fN]IY2JQOi6fN]IY2JQOi6fN]IY2JQOi6fN]IY2JPOj6gN\\IY2JPOk6fN[IZ2JPOk6fNZI[2KoNk6fNZI[2KoNk6fNZIZ2LoNk6gNYIZ2LoNk6gNYIZ2LoNk6gNYIZ2KPOl6fNXI[2LoNl6fNXI[2LnNm6gNWI[2LnNm6gNWI[2LnNn6eNVI^2LmNn6eNVI^2LmN2YN[6<gI^2LlN3[NY6;hI]2MmN2\\NX6:iI]2MmN2\\NX6:hI^2NlN1^NY67hI_2NlN1_NX66iI_2NkN2aNV65jI_2NkN2bNU64kI_2NkN1cNV63jI`2OjN1cNV63jI`2OjN1cNV63jI`2OiN2dNV62iIa2OiN1eNW61iIa2OiN1eNW61iI`20jN0eNW61iI`20iN1fNV61iI`20iN0gNW60iI`20iN0gNX6OhIa20iN0gNX6OhIa20iNOhNY6NhIa20hN0iNX6NhIa2OiN1hNX6NgIb20hN1hNY6MfIb21iNOiNZ6LfIb21iNOiNZ6LfIb21hN0jNZ6JfId20hN0jNZ6JfId20hNOkN\\6HeId21iNNlN[6FgIe20iNNlN[6FgIe2OiN0lN[6EfIf2OiNOmN\\6CgIg2NiNOmN\\6CgIg2NiNOmN]6AgIh2MkNOlN]6AgIh2MkNNmN^6@gIh2MjNOnNQ7X2PIjNOnNQ7X2oHkN0mNP7X2QIkNNnNQ7W2QIkNNnNQ7W2QIkNNnNQ7W2QIjNOoNP7W2PIkNOoNQ7V2PIkNOoNQ7U2QIkNOPOP7U2QIkNOPOP7U2PIlNOPOQ7T2PIkN0QOP7T2PIkN0POQ7T2PIlNOoNR7U2nHlN1nNR7V2mHlN1mNS7W2lHlN1lNT7X2kHkN1mNU7X2iHlN2kNV7X2iHmN1iNW7[2hHkN2iNW7\\2gHkN2gNY7^2dHlN3dN[7`2bHkN4cN\\7b2`HkN4aN^7c2_HlN3_Na7d2\\HlNX8T1gGmNZ8R1fGnN[8Q1eGoN[8Q1eGnN]8P1cGQO^8n0bGRO^8n0bGQO`8n0`GROa8l0_GUOa8k0_GTOc8k0]GUOe8i0[GWOf8h0YGYOi8d0XG\\Oj8b0VG^Ok8a0UG^On8`0QGAQ9<PGDQ9;oFES99mFGU96kFJW94jFKY91hFO[9NfF1_9JbF4e9E\\F9l9AUF=R:[OPFc0X:SOmEk0Z:lNiER1Z;I<D>BY`c1"}, "label": "apple with red white and blue sticker on the side"}]}
{"id": 541472, "image": "COCO_train2014_000000541472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red apple at the bottom\"."}], "task": "refering", "answer_template": "The \"a red apple at the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.118015625, 0.7105841121495327, 0.5239843750000001, 0.9858878504672897], "iscrowd": 0, "area": 15976.415099999997, "rle": {"size": [428, 640], "counts": "hSP14U=4ZC1i;3XD7Y;NfD4U;OlD1o:4PEMk:7VEIf:;YEFd:<]EDa:>^EC`:>aEC[:`0dEAZ:`0gE@W:b0hE_OV:b0kE^OR:e0mE\\OQ:e0PF[Oo9f0PF[Oo9e0RF[Om9f0RF\\Ol9d0UF\\Ok9d0TF]Ok9c0VF\\Oj9e0UF[Ok9e0VFZOj9g0UFYOl9f0UFYOk9h0TFXOl9h0UFWOk9j0TFVOl9j0UFUOk9l0TFTOm9k0TFTOl9m0SFSOm9m0TFROl9o0SFQOm9o0TFPOm9P1RFPOn9P1SFoNm9R1RFnNn9R1SFmNm9T1RFlNo9S1RFlNn9U1QFkNo9U1RFjNn9W1QFiNo9W1RFhNo9W1QFiNo9W1RFhNn9Y1QFgNo9Y1RFfNn9Z1RFfNo9Y1RFfNn9Z1RFfNn9[1QFeNo9[1QFeNo9[1QFeNP:Z1QFeNo9\\1PFdNP:\\1PFdNP:\\1PFdNP:\\1QFcNo9^1PFbNP:^1PFbNP:^1PFbNP:^1PFbNP:^1QFaNo9`1PF`No9a1QF_No9a1QF_No9a1QF_No9b1QF]No9c1QF]No9c1QF]Nn9d1RF\\Nn9d1RF\\No9d1QF[No9e1QF[No9e1QF[No9e1QF[NP:e1oE[NQ:e1PFZNP:f1PFZNP:f1PFZNQ:e1oE[NQ:f1nEZNR:f1oEYNQ:g1oEYNR:f1nEZNR:g1mEYNS:g1mEYNS:g1nEXNR:h1nEXNS:g1mEYNS:h1lEXNT:h1lEXNT:h1mEWNT:h1lEXNT:i1kEWNU:i1kEWNU:i1kEWNV:h1kEWNU:i1kEWNU:j1jEVNV:j1jEVNV:j1jEVNW:i1jEVNV:k1iEUNW:k1iEUNW:k1iEUNX:j1iEUNW:k1iEUNW:l1hETNX:l1hETNX:l1hETNX:l1iESNW:n1hERNX:n1hERNY:m1gESNY:m1gESNY:m1hERNX:n1hERNX:n1hERNX:o1gEQNY:o1gEQNY:o1hEPNY:o1gEQNY:o1gEQNY:o1gEQNY:o1gEQNY:P2a0000001O00000001O000001O000001O00000001O000001O00000]OPNjEP2U:QNlEn1S:SNmEm1S:SNmEn1Q:SNoEm1Q:TNnEl1Q:UNoEk1Q:UNoEk1P:VNQFi1n9XNRFh1n9XNRFh1m9YNSFh1l9XNTFh1k9YNUFg1j9ZNVFf1j9ZNVFf1i9\\NWFc1i9]NWFc1h9^NXFc1g9]NYFc1f9^NZFb1e9_N[Fa1e9_N[Fa1d9`N]F_1c9aN]F_1b9bN^F_1`9bN`F^1`9cN_F]1`9dN`F\\1`9dN`F\\1`9dN`F\\1`9dNaF[1^9fNbFZ1^9fNbF[1]9eNcF[1]9eNcF[1]9eNcF[1\\9gNcFY1]9gNdFX1\\9hNdFX1\\9hNdFY1[9YNTF1a0f1Z9VN[F2;h1Z9QNbF54j1n9TNRFl1P:RNQFm1Q:QNoEo1g:0000000002N3M3M3M3M3M3M3M3M3M3M3M3M3M2N1O001O001O1O010O001O1O001O1O001O001O1O001O001O1O001O001O1O001O1O001O00^Ro3"}, "label": "a red apple at the bottom"}]}
{"id": 541472, "image": "COCO_train2014_000000541472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple on the bottom on the left\"."}], "task": "refering", "answer_template": "The \"the apple on the bottom on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.118015625, 0.7105841121495327, 0.5239843750000001, 0.9858878504672897], "iscrowd": 0, "area": 15976.415099999997, "rle": {"size": [428, 640], "counts": "hSP14U=4ZC1i;3XD7Y;NfD4U;OlD1o:4PEMk:7VEIf:;YEFd:<]EDa:>^EC`:>aEC[:`0dEAZ:`0gE@W:b0hE_OV:b0kE^OR:e0mE\\OQ:e0PF[Oo9f0PF[Oo9e0RF[Om9f0RF\\Ol9d0UF\\Ok9d0TF]Ok9c0VF\\Oj9e0UF[Ok9e0VFZOj9g0UFYOl9f0UFYOk9h0TFXOl9h0UFWOk9j0TFVOl9j0UFUOk9l0TFTOm9k0TFTOl9m0SFSOm9m0TFROl9o0SFQOm9o0TFPOm9P1RFPOn9P1SFoNm9R1RFnNn9R1SFmNm9T1RFlNo9S1RFlNn9U1QFkNo9U1RFjNn9W1QFiNo9W1RFhNo9W1QFiNo9W1RFhNn9Y1QFgNo9Y1RFfNn9Z1RFfNo9Y1RFfNn9Z1RFfNn9[1QFeNo9[1QFeNo9[1QFeNP:Z1QFeNo9\\1PFdNP:\\1PFdNP:\\1PFdNP:\\1QFcNo9^1PFbNP:^1PFbNP:^1PFbNP:^1PFbNP:^1QFaNo9`1PF`No9a1QF_No9a1QF_No9a1QF_No9b1QF]No9c1QF]No9c1QF]Nn9d1RF\\Nn9d1RF\\No9d1QF[No9e1QF[No9e1QF[No9e1QF[NP:e1oE[NQ:e1PFZNP:f1PFZNP:f1PFZNQ:e1oE[NQ:f1nEZNR:f1oEYNQ:g1oEYNR:f1nEZNR:g1mEYNS:g1mEYNS:g1nEXNR:h1nEXNS:g1mEYNS:h1lEXNT:h1lEXNT:h1mEWNT:h1lEXNT:i1kEWNU:i1kEWNU:i1kEWNV:h1kEWNU:i1kEWNU:j1jEVNV:j1jEVNV:j1jEVNW:i1jEVNV:k1iEUNW:k1iEUNW:k1iEUNX:j1iEUNW:k1iEUNW:l1hETNX:l1hETNX:l1hETNX:l1iESNW:n1hERNX:n1hERNY:m1gESNY:m1gESNY:m1hERNX:n1hERNX:n1hERNX:o1gEQNY:o1gEQNY:o1hEPNY:o1gEQNY:o1gEQNY:o1gEQNY:o1gEQNY:P2a0000001O00000001O000001O000001O00000001O000001O00000]OPNjEP2U:QNlEn1S:SNmEm1S:SNmEn1Q:SNoEm1Q:TNnEl1Q:UNoEk1Q:UNoEk1P:VNQFi1n9XNRFh1n9XNRFh1m9YNSFh1l9XNTFh1k9YNUFg1j9ZNVFf1j9ZNVFf1i9\\NWFc1i9]NWFc1h9^NXFc1g9]NYFc1f9^NZFb1e9_N[Fa1e9_N[Fa1d9`N]F_1c9aN]F_1b9bN^F_1`9bN`F^1`9cN_F]1`9dN`F\\1`9dN`F\\1`9dN`F\\1`9dNaF[1^9fNbFZ1^9fNbF[1]9eNcF[1]9eNcF[1]9eNcF[1\\9gNcFY1]9gNdFX1\\9hNdFX1\\9hNdFY1[9YNTF1a0f1Z9VN[F2;h1Z9QNbF54j1n9TNRFl1P:RNQFm1Q:QNoEo1g:0000000002N3M3M3M3M3M3M3M3M3M3M3M3M3M2N1O001O001O1O010O001O1O001O1O001O001O1O001O001O1O001O001O1O001O1O001O00^Ro3"}, "label": "the apple on the bottom on the left"}]}
{"id": 320292, "image": "COCO_train2014_000000320292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red and yellow coloured bike\"."}], "task": "refering", "answer_template": "The \"red and yellow coloured bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 66, 67, 68, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.8222968749999999, 0.0828125, 1.0, 0.4153958333333333], "iscrowd": 0, "area": 10852.9133, "rle": {"size": [480, 640], "counts": "Uff71m>2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1hBBm;?PDDo;=nCFR<:kCIT<8iCKW<5gCNW<3fC0Y<1dC2\\<NaC5^<L_C8`<H]C;b<F[C=d<DYC?g<m0O2O000000001O0000001O00000SOhMREX2m:VNfDk1X;XNfDh1Z;YNeDg1Z;\\NdDd1Z;_NeDb1Y;_NgDa1X;`NhD`1Q;gNoDY1Q;hNnDX1R;hNmDZ1Q;gNoDY1Q;gNoDY1Q;gNoDY1Q;gNoDY1P;hNPEY1o:gNQEY1o:gNoD[1Q;[1N101O001O01O01O010O001O010O01O010O10O010L3L5N1N3N1O1O2N1O2N1O1O10O0jKfEh3Z:WLgEh3Y:YLhEf3X:YLiEf3W:ZLjEf3V:ZLkEe3T:[LlEe3U:ZLlEf3S:ZLnEf3Q:[LoEe3P:[LQFe3n9[LRFg3l9YLUFg3j9XLXFh3h9WLYFj3f9TL^Fj3b9UL`Fk3_9SLeFk3]9RLeFm3Q8"}, "label": "red and yellow coloured bike"}]}
{"id": 320292, "image": "COCO_train2014_000000320292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red , yellow and chrome fixtures on the front of a motor bike\"."}], "task": "refering", "answer_template": "The \"red , yellow and chrome fixtures on the front of a motor bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 66, 67, 68, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.8222968749999999, 0.0828125, 1.0, 0.4153958333333333], "iscrowd": 0, "area": 10852.9133, "rle": {"size": [480, 640], "counts": "Uff71m>2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1hBBm;?PDDo;=nCFR<:kCIT<8iCKW<5gCNW<3fC0Y<1dC2\\<NaC5^<L_C8`<H]C;b<F[C=d<DYC?g<m0O2O000000001O0000001O00000SOhMREX2m:VNfDk1X;XNfDh1Z;YNeDg1Z;\\NdDd1Z;_NeDb1Y;_NgDa1X;`NhD`1Q;gNoDY1Q;hNnDX1R;hNmDZ1Q;gNoDY1Q;gNoDY1Q;gNoDY1Q;gNoDY1P;hNPEY1o:gNQEY1o:gNoD[1Q;[1N101O001O01O01O010O001O010O01O010O10O010L3L5N1N3N1O1O2N1O2N1O1O10O0jKfEh3Z:WLgEh3Y:YLhEf3X:YLiEf3W:ZLjEf3V:ZLkEe3T:[LlEe3U:ZLlEf3S:ZLnEf3Q:[LoEe3P:[LQFe3n9[LRFg3l9YLUFg3j9XLXFh3h9WLYFj3f9TL^Fj3b9UL`Fk3_9SLeFk3]9RLeFm3Q8"}, "label": "red , yellow and chrome fixtures on the front of a motor bike"}]}
{"id": 320292, "image": "COCO_train2014_000000320292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle with a yellow card on its seat\"."}], "task": "refering", "answer_template": "The \"a motorcycle with a yellow card on its seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 293, 294, 295, 296, 297, 300, 301, 302, 303, 304, 317, 318, 319, 320], "bbox": [0.0, 0.0, 0.987640625, 0.8247291666666666], "iscrowd": 0, "area": 139437.21360000002, "rle": {"size": [480, 640], "counts": "b9d0V=W1hNY1ROm0N2N2M3N3M2N2M3N2N3M2M3N2N2M4M2N2N2M3N3M2M3N2N2N3L3N2O11O100N3NO001O1O001O1O001O1O001O001N2O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O0O2O1O001O1O001O001OO10000000000O100000000O2O00001O0O10001O000O2O00001O0O101O00000O2O00001O0O101O00001N1000001O000O2O00001O0O101O00000O2O00001O0O101O00001N10F:L4K5K4M4L400000O101O00000N2L4L4L5L3L4L4O100O2O0O100O100O1O2O0O100O100O2O01O0000001O01O000001O0001O01O0000001O01O000001O0001O01O0000001O01O000001O0000010O0000001O01O0001O000000010O0000001O01O0000000000001O00O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O10O01O1O100O1O1O1O100O1O1O1O1eN[1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O10000O10000O1000O012N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2N1O0N101O00001N101O00001N101O001O0O101O001O0O101OSNXHcLh7^3TIfKk6[4lN[KeKg4Z4`K_Ka4`4fKYK[4f4mKRKT4l4TLmJm3R5ZLfJh3Y5`L_Ja3`5fLkIi3T6^LUIS4j6R2N2O1O1O1O1O1N2003M4L4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L4L4L3M4L4L2N1O00001O001O00001O001O00001O1O1O2N1O1O1O1O2N1O1O1O1O2NaI\\Kc2c4mLnKS3Q4]L`Lc3_3PLoLQ4o2aK`M_4_2TKoMl4P2fJ_NZ5`1XJoNh5P1hIAY6c4000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O000O1O101N000010O0000001O0000001O000000001O0000001O0001O01O0000001O0000001O0000001O0000001O0001O01O0000001O0000001O0000001O0000001O1O2N2O1N1K6K5K4K6K5J6K4K6K5K5J5L5J6K4K6K5K5J5L5J6K4K6K5K5J5L5J6K5K4K6K5J5LX`3"}, "label": "a motorcycle with a yellow card on its seat"}]}
{"id": 320292, "image": "COCO_train2014_000000320292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small black motorcycle on display\"."}], "task": "refering", "answer_template": "The \"a small black motorcycle on display\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 293, 294, 295, 296, 297, 300, 301, 302, 303, 304, 317, 318, 319, 320], "bbox": [0.0, 0.0, 0.987640625, 0.8247291666666666], "iscrowd": 0, "area": 139437.21360000002, "rle": {"size": [480, 640], "counts": "b9d0V=W1hNY1ROm0N2N2M3N3M2N2M3N2N3M2M3N2N2M4M2N2N2M3N3M2M3N2N2N3L3N2O11O100N3NO001O1O001O1O001O1O001O001N2O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O0O2O1O001O1O001O001OO10000000000O100000000O2O00001O0O10001O000O2O00001O0O101O00000O2O00001O0O101O00001N1000001O000O2O00001O0O101O00000O2O00001O0O101O00001N10F:L4K5K4M4L400000O101O00000N2L4L4L5L3L4L4O100O2O0O100O100O1O2O0O100O100O2O01O0000001O01O000001O0001O01O0000001O01O000001O0001O01O0000001O01O000001O0000010O0000001O01O0001O000000010O0000001O01O0000000000001O00O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O10O01O1O100O1O1O1O100O1O1O1O1eN[1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O10000O10000O1000O012N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2N1O0N101O00001N101O00001N101O001O0O101O001O0O101OSNXHcLh7^3TIfKk6[4lN[KeKg4Z4`K_Ka4`4fKYK[4f4mKRKT4l4TLmJm3R5ZLfJh3Y5`L_Ja3`5fLkIi3T6^LUIS4j6R2N2O1O1O1O1O1N2003M4L4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L4L4L3M4L4L2N1O00001O001O00001O001O00001O1O1O2N1O1O1O1O2N1O1O1O1O2NaI\\Kc2c4mLnKS3Q4]L`Lc3_3PLoLQ4o2aK`M_4_2TKoMl4P2fJ_NZ5`1XJoNh5P1hIAY6c4000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O000O1O101N000010O0000001O0000001O000000001O0000001O0001O01O0000001O0000001O0000001O0000001O0001O01O0000001O0000001O0000001O0000001O1O2N2O1N1K6K5K4K6K5J6K4K6K5K5J5L5J6K4K6K5K5J5L5J6K4K6K5K5J5L5J6K5K4K6K5J5LX`3"}, "label": "a small black motorcycle on display"}]}
{"id": 574248, "image": "COCO_train2014_000000574248.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long blue board held diagonally behind people on the right\"."}], "task": "refering", "answer_template": "The \"the long blue board held diagonally behind people on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 202, 268, 290, 291, 313, 314], "bbox": [0.625296875, 0.4232083333333333, 0.8438125, 0.801], "iscrowd": 0, "area": 1963.7109499999988, "rle": {"size": [480, 640], "counts": "ekk59f>1O1N2O1N3N1O1N2O1N2O1O1N2O1N2O1O1N01O010O1O001O10O01O001O1O010O001O1O010O001O1O010O3M4L4L4L4M3L^[g01adXO4M2N2M3N3L3N2M3N0ON30O100O010O100O100O100O010O10000001O1O1N2O1O1O1O1O1O1O1N2K5Jke^1"}, "label": "the long blue board held diagonally behind people on the right"}]}
{"id": 82729, "image": "COCO_train2014_000000082729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pot boiling water with green bell peppers in man ' s kitchen\"."}], "task": "refering", "answer_template": "The \"pot boiling water with green bell peppers in man ' s kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319], "bbox": [0.5336666666666667, 0.62665625, 0.8238541666666668, 0.8082499999999999], "iscrowd": 0, "area": 13966.739099999995, "rle": {"size": [640, 480], "counts": "Y]P51nc04o]O3g?2h_O=U`0G[_Oh0b`0]Om^OR1o`0R1M3M3M3L4O1O1O00001O000000001O0O10001O000000001O0000001O0000000O2O0000001O000000001O00000O10000000001O0000000000000O1000000000000000000000000O1000000000000001O00001O0000001O00001O00001O0O101O0000001O00001N10001O00001N100O101N1O101N1O2O0O2N2M2N3M3N2K4J7H8H7J7H8I7H7A`0Jlfd1"}, "label": "pot boiling water with green bell peppers in man ' s kitchen"}]}
{"id": 82729, "image": "COCO_train2014_000000082729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pan with food cooking on the gas\"."}], "task": "refering", "answer_template": "The \"a pan with food cooking on the gas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319], "bbox": [0.5336666666666667, 0.62665625, 0.8238541666666668, 0.8082499999999999], "iscrowd": 0, "area": 13966.739099999995, "rle": {"size": [640, 480], "counts": "Y]P51nc04o]O3g?2h_O=U`0G[_Oh0b`0]Om^OR1o`0R1M3M3M3L4O1O1O00001O000000001O0O10001O000000001O0000001O0000000O2O0000001O000000001O00000O10000000001O0000000000000O1000000000000000000000000O1000000000000001O00001O0000001O00001O00001O0O101O0000001O00001N10001O00001N100O101N1O101N1O2O0O2N2M2N3M3N2K4J7H8H7J7H8I7H7A`0Jlfd1"}, "label": "a pan with food cooking on the gas"}]}
{"id": 336683, "image": "COCO_train2014_000000336683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smiling lady\"."}], "task": "refering", "answer_template": "The \"smiling lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 46, 47, 48, 64, 65, 66, 67, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.50224, 0.12584, 0.9185399999999999, 0.9865066666666666], "iscrowd": 0, "area": 40732.91405, "rle": {"size": [375, 500], "counts": "fTl24Y;:E<M200O100SOVOlFl0S9WOjFi0U9[OgFf0Y9]OdFc0[9AaF`0_9D]F<b9HZF9m3QOf1k1UNZNk1f1PN_No1b1mMbNS2^1hMgNW2Z1dMkN\\2V1_MnN`2S1[MROe2n0YMTOf2m0WMVOi2j0WMVOi2j0WMVOh2k0XMUOh2k0XMUOg2l0YMTOg2l0YMTOf2n0YMROc0hNhNV2e0RO<QOmNm1g0RO5[OPOd1k0QOOCTO\\1m0QOH2ROn0V1POF=kNc0_1PODb0iN?c1oNBf0iN;e1oNAi0gN9h1nN_Om0gN5j1nN]OQ1gN2k1mN]OT1eN0n1lN[Oi0WNTN=X2Q2kNYOm0QO:f1iNXOo0PO9h1hNVOS1nN7l1fNUOU1mN6n1eNTOX1jN5R2cNTOY1hN5S2cNUOZ1eN4V2bNTO\\1cN5X2_NUO^1`N4[2^NTO`1^N4]2]NUO`1\\N4_2\\NUOb1XN4c2ZNTOe4l0[KTOe4k0\\KTOe4l0[KTOe4l0[KTOe4l0[KSOf4l0[KTOf4k0ZKTOj4i0VKWOm4f0SKZOQ5b0oJ]OU5?lJAX5;hJE[58eJH^55bJKb50_J0d5M\\J4g5HYJ8j5DWJ<m5@SJ`0Q6\\OoId0U6WOlIi0X6SOhIm0\\6nNeIR1_6jNaIV1c6fN]IZ1g6aNZI_1j6]NVIc1_800000000001O000000001O2N2N2N2N2N2N2N2N2N2N1O2N2N2N4L4L5K5K5K4L5K5K5K9G6J2N1O1O2N1O1O1O2N2N3M2N1O1O0000001O000000001O0000001O0000001O000000001O000000001O000000001O00000000001O000000001O2N3M2N3aN_JbLd5Z3`JcLc5Y3`JeLb5X3aJfLb5V3aJhLb5S3cJjL`5R3cJlL_5Q3dJmL_5o2dJoL_5m2eJPM^5l2eJRM^5i2fJUM]5g2TKiLo4R3n1M4K5L4K5L4K5K5L4K8I7H9H7H8I8G8Ihg>"}, "label": "smiling lady"}]}
{"id": 336683, "image": "COCO_train2014_000000336683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a grey shirt holding a cake with a man\"."}], "task": "refering", "answer_template": "The \"a woman with a grey shirt holding a cake with a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 46, 47, 48, 64, 65, 66, 67, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.50224, 0.12584, 0.9185399999999999, 0.9865066666666666], "iscrowd": 0, "area": 40732.91405, "rle": {"size": [375, 500], "counts": "fTl24Y;:E<M200O100SOVOlFl0S9WOjFi0U9[OgFf0Y9]OdFc0[9AaF`0_9D]F<b9HZF9m3QOf1k1UNZNk1f1PN_No1b1mMbNS2^1hMgNW2Z1dMkN\\2V1_MnN`2S1[MROe2n0YMTOf2m0WMVOi2j0WMVOi2j0WMVOh2k0XMUOh2k0XMUOg2l0YMTOg2l0YMTOf2n0YMROc0hNhNV2e0RO<QOmNm1g0RO5[OPOd1k0QOOCTO\\1m0QOH2ROn0V1POF=kNc0_1PODb0iN?c1oNBf0iN;e1oNAi0gN9h1nN_Om0gN5j1nN]OQ1gN2k1mN]OT1eN0n1lN[Oi0WNTN=X2Q2kNYOm0QO:f1iNXOo0PO9h1hNVOS1nN7l1fNUOU1mN6n1eNTOX1jN5R2cNTOY1hN5S2cNUOZ1eN4V2bNTO\\1cN5X2_NUO^1`N4[2^NTO`1^N4]2]NUO`1\\N4_2\\NUOb1XN4c2ZNTOe4l0[KTOe4k0\\KTOe4l0[KTOe4l0[KTOe4l0[KSOf4l0[KTOf4k0ZKTOj4i0VKWOm4f0SKZOQ5b0oJ]OU5?lJAX5;hJE[58eJH^55bJKb50_J0d5M\\J4g5HYJ8j5DWJ<m5@SJ`0Q6\\OoId0U6WOlIi0X6SOhIm0\\6nNeIR1_6jNaIV1c6fN]IZ1g6aNZI_1j6]NVIc1_800000000001O000000001O2N2N2N2N2N2N2N2N2N2N1O2N2N2N4L4L5K5K5K4L5K5K5K9G6J2N1O1O2N1O1O1O2N2N3M2N1O1O0000001O000000001O0000001O0000001O000000001O000000001O000000001O00000000001O000000001O2N3M2N3aN_JbLd5Z3`JcLc5Y3`JeLb5X3aJfLb5V3aJhLb5S3cJjL`5R3cJlL_5Q3dJmL_5o2dJoL_5m2eJPM^5l2eJRM^5i2fJUM]5g2TKiLo4R3n1M4K5L4K5L4K5K5L4K8I7H9H7H8I8G8Ihg>"}, "label": "a woman with a grey shirt holding a cake with a man"}]}
{"id": 336688, "image": "COCO_train2014_000000336688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sofa\"."}], "task": "refering", "answer_template": "The \"the sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 147, 148, 149, 150, 151, 152, 153, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 168, 169, 170, 171, 172, 173, 174, 175, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 190, 191, 192, 193, 194, 195, 196, 197, 201, 202, 203, 204, 205, 206, 207, 208, 209, 213, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293], "bbox": [0.1287581699346405, 0.14591503267973857, 1.0, 0.6051470588235295], "iscrowd": 0, "area": 90393.87465000003, "rle": {"size": [612, 612], "counts": "V\\_16lb09F9G9H8G:G8G9H9F9H8G9G:G8G9H8G:G8G9G:G8G9H8H9I6J6I8I6J6J6J7I5K00000000000000000000000000O100000O100000000000000000000000O10000000000000000000O10000000O1000000000000000000000000000000O10O10000000000000000000000000O100000000000000000O100000000000O10000000000000000000000000000O10O1000000000000000000000000000O1000000000000000O100000000000O10000000000000000000000000000O10O1000RNTEdKl:Z4^E^Kb:`4hEXKX:f4QFSKo9k4[FmJe9Q5eFgJ[9W5oFaJQ9]5\\2N2N2N2N2N2N2O1N2N200O1000O10000000000000O100000000000000000000O10000000O1000000000O1000000000000000000K5J6J6J6O1O100O01000dNVA]Nk>]1]AaNc>Z1cAdN^>W1iAgNW>S1QBkNo=Q1VBmNk=R1XBlNh=S1[BgNi=Y1YB`Nl=_1VB[No=e1SBSNS>l1c1O100O100O1O100O10O01O100O100O1O100O100O10000O100O100O10000O100O01000O100O100O10000O100O100O10000O11O00001O001N1k@eNa<\\1]CfNb<Z1]CiNa<X1]CjNb<V1]ClNb<U1\\CmNc<S1\\CPOb<P1]CROb<o0\\CSOc<m0\\CUOc<l0[CWOc<i0\\CYOc<h0[CZOd<f0[C\\Od<e0ZC^Od<b0[C@d<`0[CBd<`0YCBf<b0UCAi<d0QC^On<f0mB\\OR=i0gBZOX=j0cBWO\\=o0^BROb=R1YBoNg=V1SBkNm=Y1nAhNR>]1hAdNX>T30000000000000000000000000O10000000O100000000000000000000000000000000000000000000O010000000000000000000000000000000000000000000O1000000000O1000000000000000000000000000000000O10000000000000000000O100000000000000000000000O100000000000000000000000000000O10000000000000O1000000000000000000000000000000000000000O1000O10000000000000000000000000000000000000000000I6UNl1TNSO"}, "label": "the sofa"}]}
{"id": 336688, "image": "COCO_train2014_000000336688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black couch with pillows on it\"."}], "task": "refering", "answer_template": "The \"the black couch with pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 147, 148, 149, 150, 151, 152, 153, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 168, 169, 170, 171, 172, 173, 174, 175, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 190, 191, 192, 193, 194, 195, 196, 197, 201, 202, 203, 204, 205, 206, 207, 208, 209, 213, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293], "bbox": [0.1287581699346405, 0.14591503267973857, 1.0, 0.6051470588235295], "iscrowd": 0, "area": 90393.87465000003, "rle": {"size": [612, 612], "counts": "V\\_16lb09F9G9H8G:G8G9H9F9H8G9G:G8G9H8G:G8G9G:G8G9H8H9I6J6I8I6J6J6J7I5K00000000000000000000000000O100000O100000000000000000000000O10000000000000000000O10000000O1000000000000000000000000000000O10O10000000000000000000000000O100000000000000000O100000000000O10000000000000000000000000000O10O1000000000000000000000000000O1000000000000000O100000000000O10000000000000000000000000000O10O1000RNTEdKl:Z4^E^Kb:`4hEXKX:f4QFSKo9k4[FmJe9Q5eFgJ[9W5oFaJQ9]5\\2N2N2N2N2N2N2O1N2N200O1000O10000000000000O100000000000000000000O10000000O1000000000O1000000000000000000K5J6J6J6O1O100O01000dNVA]Nk>]1]AaNc>Z1cAdN^>W1iAgNW>S1QBkNo=Q1VBmNk=R1XBlNh=S1[BgNi=Y1YB`Nl=_1VB[No=e1SBSNS>l1c1O100O100O1O100O10O01O100O100O1O100O100O10000O100O100O10000O100O01000O100O100O10000O100O100O10000O11O00001O001N1k@eNa<\\1]CfNb<Z1]CiNa<X1]CjNb<V1]ClNb<U1\\CmNc<S1\\CPOb<P1]CROb<o0\\CSOc<m0\\CUOc<l0[CWOc<i0\\CYOc<h0[CZOd<f0[C\\Od<e0ZC^Od<b0[C@d<`0[CBd<`0YCBf<b0UCAi<d0QC^On<f0mB\\OR=i0gBZOX=j0cBWO\\=o0^BROb=R1YBoNg=V1SBkNm=Y1nAhNR>]1hAdNX>T30000000000000000000000000O10000000O100000000000000000000000000000000000000000000O010000000000000000000000000000000000000000000O1000000000O1000000000000000000000000000000000O10000000000000000000O100000000000000000000000O100000000000000000000000000000O10000000000000O1000000000000000000000000000000000000000O1000O10000000000000000000000000000000000000000000I6UNl1TNSO"}, "label": "the black couch with pillows on it"}]}
{"id": 336688, "image": "COCO_train2014_000000336688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black sofa with a little boy sitting on it\"."}], "task": "refering", "answer_template": "The \"a black sofa with a little boy sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 344, 345, 346, 347, 348, 349, 350, 366, 367, 368, 369, 370], "bbox": [0.6488235294117647, 0.4134803921568628, 0.944330065359477, 0.7560947712418301], "iscrowd": 0, "area": 24735.351650000004, "rle": {"size": [612, 612], "counts": "ke]71Rc03M2N3M3L3N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3L3N3L4M1N2O0O2N2O1lKgM[G[2d8hMYGY2g8hMVGZ2j8gMTGY2m8hMPGZ2P9hMlFZ2T9gMjFZ2V9gMgF[2Y9fMeF[2[9fMbF\\2^9fM_FZ2b9gM[F[2e9fMQDc4o;^KkCg4U<[KcCk4]<VK\\CP5d<h00O10O1000000000O1000000000000O1000000000O10O1000000000000O101O1O1O001O1N2O1O001O1O1O1N101O1O1O1O001N2O1O001O1O1O1N101O1O1O1O001N2O1O1O001O1O1N2O001O1O1O001N2O1O1O001O1O1N2O001O1O1O1O0O2O1O2N1O1O1O1N2O1O2N1O1O1O1N2O1O2N1O1O1O1N2O2N1O1O1O1O1N2O2N1O1O1O1O1N2O2N`0@a0_Ob0^Ob0^Oa0^Ojlc0"}, "label": "a black sofa with a little boy sitting on it"}]}
{"id": 336688, "image": "COCO_train2014_000000336688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black ottoman being sat on by a little kid\"."}], "task": "refering", "answer_template": "The \"a black ottoman being sat on by a little kid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 344, 345, 346, 347, 348, 349, 350, 366, 367, 368, 369, 370], "bbox": [0.6488235294117647, 0.4134803921568628, 0.944330065359477, 0.7560947712418301], "iscrowd": 0, "area": 24735.351650000004, "rle": {"size": [612, 612], "counts": "ke]71Rc03M2N3M3L3N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3L3N3L4M1N2O0O2N2O1lKgM[G[2d8hMYGY2g8hMVGZ2j8gMTGY2m8hMPGZ2P9hMlFZ2T9gMjFZ2V9gMgF[2Y9fMeF[2[9fMbF\\2^9fM_FZ2b9gM[F[2e9fMQDc4o;^KkCg4U<[KcCk4]<VK\\CP5d<h00O10O1000000000O1000000000000O1000000000O10O1000000000000O101O1O1O001O1N2O1O001O1O1O1N101O1O1O1O001N2O1O001O1O1O1N101O1O1O1O001N2O1O1O001O1O1N2O001O1O1O001N2O1O1O001O1O1N2O001O1O1O1O0O2O1O2N1O1O1O1N2O1O2N1O1O1O1N2O1O2N1O1O1O1N2O2N1O1O1O1O1N2O2N1O1O1O1O1N2O2N`0@a0_Ob0^Ob0^Oa0^Ojlc0"}, "label": "a black ottoman being sat on by a little kid"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"huge bagage on back of a lady\"."}], "task": "refering", "answer_template": "The \"huge bagage on back of a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 132, 133, 134, 155, 156, 157, 158, 178, 179, 180, 202, 203, 204, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274], "bbox": [0.737625, 0.2712756264236902, 0.9663906250000001, 0.7288610478359909], "iscrowd": 0, "area": 11569.624549999999, "rle": {"size": [439, 640], "counts": "``Z63c=6J5K4L2N2N2N2O1N2N2N2N2N2O1O1N2O1N2O1O1N101O1N2O1O1N2O0O2O2N1N2O2N1N3N1O2N1O2N1O2N1O1O2N101N1O2N3mF\\Mm6h2iH^MW7e2_HbMa7a2VHfMi7]2nGiMR8Z2eGlM[8X2ZGPNe8a30001O00001O00001O000WO`K\\H`4^7mKZHS4g7n0001N101O0O2O0O3N2XLQHc1R8ZNSHb1n7[NYH`1i7mLoGS1=k1U8QNPHk1R8RNSHi1Q8SNSHj1P8RNSHk1T8mMoGR2m9M3N1O2M2O2M3N1N2O1O2M2:FO1O2N1O1O1O100O1O001O1K5K4K6N11O000000002N4M3L4L3M001O001000O010O1L3ZOg0N2N1O2M2O2N2M2O2M3K4K6KeU9"}, "label": "huge bagage on back of a lady"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue backpack being wore by a woman\"."}], "task": "refering", "answer_template": "The \"a blue backpack being wore by a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 132, 133, 134, 155, 156, 157, 158, 178, 179, 180, 202, 203, 204, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274], "bbox": [0.737625, 0.2712756264236902, 0.9663906250000001, 0.7288610478359909], "iscrowd": 0, "area": 11569.624549999999, "rle": {"size": [439, 640], "counts": "``Z63c=6J5K4L2N2N2N2O1N2N2N2N2N2O1O1N2O1N2O1O1N101O1N2O1O1N2O0O2O2N1N2O2N1N3N1O2N1O2N1O2N1O1O2N101N1O2N3mF\\Mm6h2iH^MW7e2_HbMa7a2VHfMi7]2nGiMR8Z2eGlM[8X2ZGPNe8a30001O00001O00001O000WO`K\\H`4^7mKZHS4g7n0001N101O0O2O0O3N2XLQHc1R8ZNSHb1n7[NYH`1i7mLoGS1=k1U8QNPHk1R8RNSHi1Q8SNSHj1P8RNSHk1T8mMoGR2m9M3N1O2M2O2M3N1N2O1O2M2:FO1O2N1O1O1O100O1O001O1K5K4K6N11O000000002N4M3L4L3M001O001000O010O1L3ZOg0N2N1O2M2O2N2M2O2M3K4K6KeU9"}, "label": "a blue backpack being wore by a woman"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog with gear strapped to him\"."}], "task": "refering", "answer_template": "The \"a dog with gear strapped to him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.43531250000000005, 0.8078132118451025, 0.774671875, 1.0], "iscrowd": 0, "area": 14237.353100000004, "rle": {"size": [439, 640], "counts": "]og36a=1N2O1N101N101O1N100O100O1oB@j<a0UCBi<>VCEf<=YCDe<h0O1N2O1O1O1O3N1N3N1N2O1N2O1N2O1N2O1O2N1O2N101N1O100O3N3L3M10O000000000001O000001O00000000000000000000001O000000001O000000001O000000001O000000001O00000000000000000000000000000000000000000000000000O10000000000O1000000O100000000O1000000O10000O1000000O10000O10000O100O100O100O10000000000000000000000000000001O0000000000001O001O001O1O001O001O001O1O001O001O001O1O001O1O1O2N3M3UN]D\\1Q<N2N2N3M2N2N2N2N3M2N3M4L7I5KPZm1"}, "label": "a dog with gear strapped to him"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog with an attached bag standing before three people carrying camping equipment\"."}], "task": "refering", "answer_template": "The \"dog with an attached bag standing before three people carrying camping equipment\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.43531250000000005, 0.8078132118451025, 0.774671875, 1.0], "iscrowd": 0, "area": 14237.353100000004, "rle": {"size": [439, 640], "counts": "]og36a=1N2O1N101N101O1N100O100O1oB@j<a0UCBi<>VCEf<=YCDe<h0O1N2O1O1O1O3N1N3N1N2O1N2O1N2O1N2O1O2N1O2N101N1O100O3N3L3M10O000000000001O000001O00000000000000000000001O000000001O000000001O000000001O000000001O00000000000000000000000000000000000000000000000000O10000000000O1000000O100000000O1000000O10000O1000000O10000O10000O100O100O100O10000000000000000000000000000001O0000000000001O001O001O1O001O001O001O1O001O001O001O1O001O1O1O2N3M3UN]D\\1Q<N2N2N3M2N2N2N2N3M2N3M4L7I5KPZm1"}, "label": "dog with an attached bag standing before three people carrying camping equipment"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the solid blue shirt\"."}], "task": "refering", "answer_template": "The \"the man in the solid blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 307, 308], "bbox": [0.24925000000000003, 0.5185421412300683, 0.45549999999999996, 0.8583371298405466], "iscrowd": 0, "area": 12631.9109, "rle": {"size": [439, 640], "counts": "RmT25]=6K5K4L5L3L4L5M2O1N2N`0@8H9H0O1O2N2M3N2N2M3N2O1O1O001O1O10O0O100O2O1O010000O10000O10O10O0100O00100O100O10000O1000000OMe0SOa0J3M4L3M4L3M3N2N01O0O2O001O1O001N2O1O001O0O101O0O1O1O2N100O1O2N1O100O10000XLXFa3h9\\L\\Fb3e9\\L^Fc3k901O1O1N2O2N1O1O1N2O2N1O5J6K2N3M2N2M4M2L4M3M4K4M2M4ROPD0S<MoC1T<LoC0U<NlC0W<Mk0N3Eild4"}, "label": "the man in the solid blue shirt"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt on the left\"."}], "task": "refering", "answer_template": "The \"man in blue shirt on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 307, 308], "bbox": [0.24925000000000003, 0.5185421412300683, 0.45549999999999996, 0.8583371298405466], "iscrowd": 0, "area": 12631.9109, "rle": {"size": [439, 640], "counts": "RmT25]=6K5K4L5L3L4L5M2O1N2N`0@8H9H0O1O2N2M3N2N2M3N2O1O1O001O1O10O0O100O2O1O010000O10000O10O10O0100O00100O100O10000O1000000OMe0SOa0J3M4L3M4L3M3N2N01O0O2O001O1O001N2O1O001O0O101O0O1O1O2N100O1O2N1O100O10000XLXFa3h9\\L\\Fb3e9\\L^Fc3k901O1O1N2O2N1O1O1N2O2N1O5J6K2N3M2N2M4M2L4M3M4K4M2M4ROPD0S<MoC1T<LoC0U<NlC0W<Mk0N3Eild4"}, "label": "man in blue shirt on the left"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl on right\"."}], "task": "refering", "answer_template": "The \"girl on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 316, 317, 339, 340, 362, 363], "bbox": [0.636609375, 0.33708428246013666, 0.8246718749999999, 0.98876993166287], "iscrowd": 0, "area": 17975.605199999998, "rle": {"size": [439, 640], "counts": "`d^5f0o<=C5K4N3L3N1N2O1N3N1O1N2O1O001O001RHSNP4n1jK^Nn3c1nKeNn3\\1lKjNS4V1iKnNW4S1cKSO\\4m0_KXOa4i0XK^Oh4a0SKDm4=mJIS56gJ0Z50`J5`5LPImMk0]2U6EmHTNe0]2_6_OiHYN`0d1FYNS7j1dH`N6c18PNP7m1_HQ1e0oLn6R5VIlJk6Q5P1M2M4L4M2M4L3N32M5L5J5L5J5L5J2OO0010O010O0010O0017H8I2M2O1N2O1=DO0N4M3L4M3M0O2O1N2O0O2O1N2O2M4M3L3N3L3N3M2M3NO001N101O1O1O2N1O2M2O2eKVI53dNh6T1XI2:eN`6U1XI2`0eNY6V1ZI0f0eNR6V1[I0Q1`Ng5\\1[IO_1XNW5f1\\INo8NTGLS90PGMU9NmF1V9KmF4U9HmF7V9EmF:V9@mF?V9]OmFb0W9XOkFg0Z9SOiFh0_9ROcFh0g9RO\\Fh0];ISg_1"}, "label": "girl on right"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girll with white shirt and tan shorts\"."}], "task": "refering", "answer_template": "The \"girll with white shirt and tan shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 316, 317, 339, 340, 362, 363], "bbox": [0.636609375, 0.33708428246013666, 0.8246718749999999, 0.98876993166287], "iscrowd": 0, "area": 17975.605199999998, "rle": {"size": [439, 640], "counts": "`d^5f0o<=C5K4N3L3N1N2O1N3N1O1N2O1O001O001RHSNP4n1jK^Nn3c1nKeNn3\\1lKjNS4V1iKnNW4S1cKSO\\4m0_KXOa4i0XK^Oh4a0SKDm4=mJIS56gJ0Z50`J5`5LPImMk0]2U6EmHTNe0]2_6_OiHYN`0d1FYNS7j1dH`N6c18PNP7m1_HQ1e0oLn6R5VIlJk6Q5P1M2M4L4M2M4L3N32M5L5J5L5J5L5J2OO0010O010O0010O0017H8I2M2O1N2O1=DO0N4M3L4M3M0O2O1N2O0O2O1N2O2M4M3L3N3L3N3M2M3NO001N101O1O1O2N1O2M2O2eKVI53dNh6T1XI2:eN`6U1XI2`0eNY6V1ZI0f0eNR6V1[I0Q1`Ng5\\1[IO_1XNW5f1\\INo8NTGLS90PGMU9NmF1V9KmF4U9HmF7V9EmF:V9@mF?V9]OmFb0W9XOkFg0Z9SOiFh0_9ROcFh0g9RO\\Fh0];ISg_1"}, "label": "girll with white shirt and tan shorts"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"backpack of man in dark blue shirt\"."}], "task": "refering", "answer_template": "The \"backpack of man in dark blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 193, 194, 212, 213, 214], "bbox": [0.24978125, 0.38034168564920273, 0.4635625, 0.6234851936218678], "iscrowd": 0, "area": 7433.240599999999, "rle": {"size": [439, 640], "counts": "`iT27^=4L2`CJ];8^Da0k:BTE`0j:ATEa0j:@VEb0h:_OWEb0h:EQE<n:X1N1O2N1O1O010O1O01O0O2O0O101N2O0O2O1N2O1O001O1O100O1000O10O0100O10O00010O001L3M4M210O1O010O1O100O001O100O1O001O1N2M3M3O010O1O1O10O01O10000000000000O10000000000000001O0000001O100O1O4\\DaNh:c1SE`Nk:i1kDXNU;j1iDTNY;R24N10000O10000O2O1O1O1N9A9H7O2N2O001O1M2O2M3N1N3M4K3M3N1O2N2N1N3NhXc4"}, "label": "backpack of man in dark blue shirt"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the backpack of the mustached man\"."}], "task": "refering", "answer_template": "The \"the backpack of the mustached man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 193, 194, 212, 213, 214], "bbox": [0.24978125, 0.38034168564920273, 0.4635625, 0.6234851936218678], "iscrowd": 0, "area": 7433.240599999999, "rle": {"size": [439, 640], "counts": "`iT27^=4L2`CJ];8^Da0k:BTE`0j:ATEa0j:@VEb0h:_OWEb0h:EQE<n:X1N1O2N1O1O010O1O01O0O2O0O101N2O0O2O1N2O1O001O1O100O1000O10O0100O10O00010O001L3M4M210O1O010O1O100O001O100O1O001O1N2M3M3O010O1O1O10O01O10000000000000O10000000000000001O0000001O100O1O4\\DaNh:c1SE`Nk:i1kDXNU;j1iDTNY;R24N10000O10000O2O1O1O1N9A9H7O2N2O001O1M2O2M3N1N3M4K3M3N1O2N2N1N3NhXc4"}, "label": "the backpack of the mustached man"}]}
{"id": 320308, "image": "COCO_train2014_000000320308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lower part of a dog , its head is not visible\"."}], "task": "refering", "answer_template": "The \"the lower part of a dog , its head is not visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [282, 283, 304, 305, 306, 307, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.09862499999999999, 0.793986332574032, 0.35621875000000003, 1.0], "iscrowd": 0, "area": 8320.55225, "rle": {"size": [439, 640], "counts": "Y]k01e=4K6K4M<D000000000000000000O101O001O1O001O001O001O1O001O001O1O1O1O1O1O0O10001O00000000001O00O10000O100O10000O100N1O2N2N2N2N2L4M3M3N3M2O1O1O1N2O1O1O1N2O2N1O1N2O1O1O1N2O1O1O2M2O1O1O1N200O10000O100O100O100O100O10000O100O100O1O1O1O1O100O1O1O1O1O01000000000000O10000O100O100O100O02O0000001O0000001O100O01O1O1O001O3N2M3M4L2N3WO_DVOd;d0dDWO^;c0lDUOj;4Q1JgW`5"}, "label": "the lower part of a dog , its head is not visible"}]}
{"id": 164663, "image": "COCO_train2014_000000164663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the right\"."}], "task": "refering", "answer_template": "The \"the man on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 135, 156, 157, 158, 179, 180, 181, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 249, 250, 251], "bbox": [0.7618125, 0.43131250000000004, 0.9635468750000001, 0.98865625], "iscrowd": 0, "area": 8300.201850000003, "rle": {"size": [320, 640], "counts": "Yih48d96E9hFZOU9j0M0nFWOo8k0RGUOl8l0UGTOh8n0YGQOd8Q13[OYG7e8IbG2[8OnGHQ89i010O01O100001O001O001O001O0100000O01POC^H>_7F]H>_7F]H=a7F]H;a7H\\H;b7G\\H:c7G\\H:c7H[H:c7H[H9f7FXH<h7EVH=i7CUH?k7BSH?m7f04K6N11bIjNP4W1hKPOY4P1aKUO`4k0[KYOf4h0TK\\Om4f0lJ^OU5d0dJ@\\5a0VJLk5T2001O1OO1O1O1N2O1O1O1O100O1O1O1N2M3N2N2N2M3VO^KTLo0@^2\\4fLoK=ICOY3Y4gLlK=OAOZ3U4iLkK<4_OOZ3P4mLjK;j0g2]3mLgK<Q1d2Z3oLaK=Z1b2V3cMoL[2R3aMVMZ2k2_M`M\\2d2ZMeMc2m4O1O1O2N3M1O2N;E001O00001O001O2N1O2N1O2N1O2N1O2N3M4L3M4L4L4L3M3M3M2N3M3M3M2N1O2NUV7"}, "label": "the man on the right"}]}
{"id": 164663, "image": "COCO_train2014_000000164663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the purple shirt who caught the frisbee\"."}], "task": "refering", "answer_template": "The \"the person in the purple shirt who caught the frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 135, 156, 157, 158, 179, 180, 181, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 249, 250, 251], "bbox": [0.7618125, 0.43131250000000004, 0.9635468750000001, 0.98865625], "iscrowd": 0, "area": 8300.201850000003, "rle": {"size": [320, 640], "counts": "Yih48d96E9hFZOU9j0M0nFWOo8k0RGUOl8l0UGTOh8n0YGQOd8Q13[OYG7e8IbG2[8OnGHQ89i010O01O100001O001O001O001O0100000O01POC^H>_7F]H>_7F]H=a7F]H;a7H\\H;b7G\\H:c7G\\H:c7H[H:c7H[H9f7FXH<h7EVH=i7CUH?k7BSH?m7f04K6N11bIjNP4W1hKPOY4P1aKUO`4k0[KYOf4h0TK\\Om4f0lJ^OU5d0dJ@\\5a0VJLk5T2001O1OO1O1O1N2O1O1O1O100O1O1O1N2M3N2N2N2M3VO^KTLo0@^2\\4fLoK=ICOY3Y4gLlK=OAOZ3U4iLkK<4_OOZ3P4mLjK;j0g2]3mLgK<Q1d2Z3oLaK=Z1b2V3cMoL[2R3aMVMZ2k2_M`M\\2d2ZMeMc2m4O1O1O2N3M1O2N;E001O00001O001O2N1O2N1O2N1O2N1O2N3M4L3M4L4L4L3M3M3M2N3M3M3M2N1O2NUV7"}, "label": "the person in the purple shirt who caught the frisbee"}]}
{"id": 164663, "image": "COCO_train2014_000000164663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in dark blue shirt getting ready to leap\"."}], "task": "refering", "answer_template": "The \"man in dark blue shirt getting ready to leap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 234], "bbox": [0.002359375, 0.7074374999999999, 0.195296875, 0.9898125], "iscrowd": 0, "area": 6319.789, "rle": {"size": [320, 640], "counts": "_l0=b98H0000000001O000001O000hG\\OU7d0kH]OT7c0kH_OU7`0kHAT7?lHBT7=kHEU7:kHGU78kHIU77jHJV75iHMW72iHOW71hH0X71fH0Z70dH2\\7ObH2^7O`H2`7N_H3a7N\\H3e7MZH4f7MXH4h7MVH4j7LTH6m7JQH7o7JoG7Q8InG8S8HjG:V8GiG9W8h0001O00001O0000001O00001O00001O001OBfNVHZ1i7?O1O1O1000000O010D<O100O100O100O100O100O1O100O11O00001O00001O001O0N2O2M2O3N2N2N2N5K1O1O101N100dNQHe0P8RO[Hl0^8N3N2M100O1O100O100O100O100L4I7HndP5"}, "label": "man in dark blue shirt getting ready to leap"}]}
{"id": 164663, "image": "COCO_train2014_000000164663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in purple on the ground preparing to jump\"."}], "task": "refering", "answer_template": "The \"man in purple on the ground preparing to jump\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 234], "bbox": [0.002359375, 0.7074374999999999, 0.195296875, 0.9898125], "iscrowd": 0, "area": 6319.789, "rle": {"size": [320, 640], "counts": "_l0=b98H0000000001O000001O000hG\\OU7d0kH]OT7c0kH_OU7`0kHAT7?lHBT7=kHEU7:kHGU78kHIU77jHJV75iHMW72iHOW71hH0X71fH0Z70dH2\\7ObH2^7O`H2`7N_H3a7N\\H3e7MZH4f7MXH4h7MVH4j7LTH6m7JQH7o7JoG7Q8InG8S8HjG:V8GiG9W8h0001O00001O0000001O00001O00001O001OBfNVHZ1i7?O1O1O1000000O010D<O100O100O100O100O100O1O100O11O00001O00001O001O0N2O2M2O3N2N2N2N5K1O1O101N100dNQHe0P8RO[Hl0^8N3N2M100O1O100O100O100O100L4I7HndP5"}, "label": "man in purple on the ground preparing to jump"}]}
{"id": 164663, "image": "COCO_train2014_000000164663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the purple shirt in the second frame\"."}], "task": "refering", "answer_template": "The \"the man in the purple shirt in the second frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 125, 126, 148, 149, 171, 172, 173, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242], "bbox": [0.39775, 0.43975, 0.5516875, 0.9858437500000001], "iscrowd": 0, "area": 9779.39615, "rle": {"size": [320, 640], "counts": "ho_23k92N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2gLeNkM]1S2fN`K1\\1[1R3fN_K7X1U1X3fN]K;U1Q1\\3gN\\K=R1o0`3gNZK?R1l0b3gNZKa0P1j0f3fNVKf0n0g0l3dNSKi0m0e0P42kK1U42fK0Z42aK1^4i20000000000001O1O2N1O2N1O2N1O1OO100O1^O[KiKe4T4`KjK`4T4cKjK^4V4cKiK]4W4dKhK\\4Y4dKeK]4[4dKdK\\4\\4cKdK^4\\4bKdK^4[4bKfK^4Z4aKfK`4[4`KdK`4_4]KaKc4a4\\K]Ke4c4\\K\\K:Mg3h4nKZK;4b3b4TLYK8<_3\\4XLXK8a0\\3X4jLoKQ3Q4nLULm2l3PMZLl2f3SM_Li2b3TMcLi2^3UMgLg2Y3WMkLg2V3WMlLh2T3VMnLj2S3TMoLk2T5O001O1O1XMnH`2R7`MSI[2n6dMXIV2i6iM\\IR2^7K5K6J5K5K8H7I8H7I8HX\\i2"}, "label": "the man in the purple shirt in the second frame"}]}
{"id": 164663, "image": "COCO_train2014_000000164663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy is playing with a frisbee\"."}], "task": "refering", "answer_template": "The \"a boy is playing with a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 125, 126, 148, 149, 171, 172, 173, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242], "bbox": [0.39775, 0.43975, 0.5516875, 0.9858437500000001], "iscrowd": 0, "area": 9779.39615, "rle": {"size": [320, 640], "counts": "ho_23k92N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2gLeNkM]1S2fN`K1\\1[1R3fN_K7X1U1X3fN]K;U1Q1\\3gN\\K=R1o0`3gNZK?R1l0b3gNZKa0P1j0f3fNVKf0n0g0l3dNSKi0m0e0P42kK1U42fK0Z42aK1^4i20000000000001O1O2N1O2N1O2N1O1OO100O1^O[KiKe4T4`KjK`4T4cKjK^4V4cKiK]4W4dKhK\\4Y4dKeK]4[4dKdK\\4\\4cKdK^4\\4bKdK^4[4bKfK^4Z4aKfK`4[4`KdK`4_4]KaKc4a4\\K]Ke4c4\\K\\K:Mg3h4nKZK;4b3b4TLYK8<_3\\4XLXK8a0\\3X4jLoKQ3Q4nLULm2l3PMZLl2f3SM_Li2b3TMcLi2^3UMgLg2Y3WMkLg2V3WMlLh2T3VMnLj2S3TMoLk2T5O001O1O1XMnH`2R7`MSI[2n6dMXIV2i6iM\\IR2^7K5K6J5K5K8H7I8H7I8HX\\i2"}, "label": "a boy is playing with a frisbee"}]}
{"id": 9018, "image": "COCO_train2014_000000009018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white truck with dent in the door\"."}], "task": "refering", "answer_template": "The \"white truck with dent in the door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 138, 139, 140, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235], "bbox": [0.00375, 0.3657772621809745, 0.269296875, 0.7336194895591647], "iscrowd": 0, "area": 18816.9475, "rle": {"size": [431, 640], "counts": "[Q1\\3h8\\1K4000001O0000001O0001O000O1000000O10000O101O000O1O1M3N2M3M3O10001O000000000000000000010O0000000000000000000O1O2N1O1O1O1O100O1O1O1O4L3M4L4M2N3M3M2N3M3M2N3M3N1N3M3M2N2N00000000000000000000000001O005K5K5K5K5K1O00000000000000000000000000000000000000000000000000N2L4N20000000000O10000000000000000000000O10000000000nNcERO]:m0jEmNV:Q1SFhNm9W1ZFcNf9[1bF_N^9`1Y1N2O1O2M5K5F;D;FchT6"}, "label": "white truck with dent in the door"}]}
{"id": 9018, "image": "COCO_train2014_000000009018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck in front of a yellow truck\"."}], "task": "refering", "answer_template": "The \"a white truck in front of a yellow truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 138, 139, 140, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235], "bbox": [0.00375, 0.3657772621809745, 0.269296875, 0.7336194895591647], "iscrowd": 0, "area": 18816.9475, "rle": {"size": [431, 640], "counts": "[Q1\\3h8\\1K4000001O0000001O0001O000O1000000O10000O101O000O1O1M3N2M3M3O10001O000000000000000000010O0000000000000000000O1O2N1O1O1O1O100O1O1O1O4L3M4L4M2N3M3M2N3M3M2N3M3N1N3M3M2N2N00000000000000000000000001O005K5K5K5K5K1O00000000000000000000000000000000000000000000000000N2L4N20000000000O10000000000000000000000O10000000000nNcERO]:m0jEmNV:Q1SFhNm9W1ZFcNf9[1bF_N^9`1Y1N2O1O2M5K5F;D;FchT6"}, "label": "a white truck in front of a yellow truck"}]}
{"id": 9018, "image": "COCO_train2014_000000009018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"old yellow and white truck parked behind other truck\"."}], "task": "refering", "answer_template": "The \"old yellow and white truck parked behind other truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 236, 238, 239, 240, 242, 243], "bbox": [0.252734375, 0.3893271461716938, 0.62046875, 0.7151740139211137], "iscrowd": 0, "area": 23355.41265, "rle": {"size": [431, 640], "counts": "Y^T26V=9G9G9G9G4M2M3J6WOi0M3L4M2NO100003L4M3N2O1O1O1O1O1O1O100O1O1O100O100000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000001O3L4M3M4L1O000000000000O1M3N2N2N2M2O23M3M4L3M3M1O0000001O0000001O000O10001O0000001O0000N2N2N2M3N2N2M2O2N200000000000000000001N2O1O1O1O1O1O1O1O1O1O1O1O1O2N2N3M2N2N2N3M2N2M3N2N3M2N1O00001O000002N3M3M3M3M1O1O000000000000001O00000001O000001O000000000000N2G9G9I7M3N2O100O100O100O1O100O100O1O100O100O00102M9H7H8I8G8HRQV3"}, "label": "old yellow and white truck parked behind other truck"}]}
{"id": 9018, "image": "COCO_train2014_000000009018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck that is yellow on the top and white on the bottom half\"."}], "task": "refering", "answer_template": "The \"a truck that is yellow on the top and white on the bottom half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 236, 238, 239, 240, 242, 243], "bbox": [0.252734375, 0.3893271461716938, 0.62046875, 0.7151740139211137], "iscrowd": 0, "area": 23355.41265, "rle": {"size": [431, 640], "counts": "Y^T26V=9G9G9G9G4M2M3J6WOi0M3L4M2NO100003L4M3N2O1O1O1O1O1O1O100O1O1O100O100000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000001O3L4M3M4L1O000000000000O1M3N2N2N2M2O23M3M4L3M3M1O0000001O0000001O000O10001O0000001O0000N2N2N2M3N2N2M2O2N200000000000000000001N2O1O1O1O1O1O1O1O1O1O1O1O1O2N2N3M2N2N2N3M2N2M3N2N3M2N1O00001O000002N3M3M3M3M1O1O000000000000001O00000001O000001O000000000000N2G9G9I7M3N2O100O100O100O1O100O100O1O100O100O00102M9H7H8I8G8HRQV3"}, "label": "a truck that is yellow on the top and white on the bottom half"}]}
{"id": 9018, "image": "COCO_train2014_000000009018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a gray shirt\"."}], "task": "refering", "answer_template": "The \"a man in a gray shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 133, 154, 155, 156, 177, 178, 179, 199, 200, 202, 203, 222, 226, 245, 246, 249, 272], "bbox": [0.65225, 0.26675174013921116, 0.8716875, 0.7476334106728538], "iscrowd": 0, "area": 8653.506299999997, "rle": {"size": [431, 640], "counts": "gW`51^=0O1O1O1O1O100O1O1O1ZO3iCNR<j0J6J6K5J5K6M3O1N2N2N2O1N2N2NdNdEI[:HYF4e9LbFO\\92jFIT98RGCf7N`G`0o0^OZ78fG;[:HcE9[:IdE8Y:KfE6X:MfE4X:NgE3V:1gE1W:1hE0W:2fE0Z:0bE4^:M]E7c:IXE<i:U1O001O100O0^O]MYFd2f9`MVFa2i9cMTF]2l9fMPF[2o9c0O1O001O100O1O1O1O010O11O2N1O1N3N1O1O2N3M5K4K2O1O100O1O1O1O100O2N1O100001O0001O0001O0000PNSE]1m:aNXE]1j:]N]E_1];M3QOTD4o;DVD<n;ZOXDg0X<001O3M3M3M2N3iDjNn9Y1jElNV:V1bEPO_:R1XESOi:P1oDUOQ;e1000O5L4L4K6K4L6Ja0^Oa0@0O1O1O1O1O1O1O1O1O2M2O1OmcR1"}, "label": "a man in a gray shirt"}]}
{"id": 9018, "image": "COCO_train2014_000000009018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two men on a horse\"."}], "task": "refering", "answer_template": "The \"two men on a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 133, 154, 155, 156, 177, 178, 179, 199, 200, 202, 203, 222, 226, 245, 246, 249, 272], "bbox": [0.65225, 0.26675174013921116, 0.8716875, 0.7476334106728538], "iscrowd": 0, "area": 8653.506299999997, "rle": {"size": [431, 640], "counts": "gW`51^=0O1O1O1O1O100O1O1O1ZO3iCNR<j0J6J6K5J5K6M3O1N2N2N2O1N2N2NdNdEI[:HYF4e9LbFO\\92jFIT98RGCf7N`G`0o0^OZ78fG;[:HcE9[:IdE8Y:KfE6X:MfE4X:NgE3V:1gE1W:1hE0W:2fE0Z:0bE4^:M]E7c:IXE<i:U1O001O100O0^O]MYFd2f9`MVFa2i9cMTF]2l9fMPF[2o9c0O1O001O100O1O1O1O010O11O2N1O1N3N1O1O2N3M5K4K2O1O100O1O1O1O100O2N1O100001O0001O0001O0000PNSE]1m:aNXE]1j:]N]E_1];M3QOTD4o;DVD<n;ZOXDg0X<001O3M3M3M2N3iDjNn9Y1jElNV:V1bEPO_:R1XESOi:P1oDUOQ;e1000O5L4L4K6K4L6Ja0^Oa0@0O1O1O1O1O1O1O1O1O2M2O1OmcR1"}, "label": "two men on a horse"}]}
{"id": 271166, "image": "COCO_train2014_000000271166.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange snowboard\"."}], "task": "refering", "answer_template": "The \"the orange snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 98, 99, 121, 122, 123, 144, 145, 146, 167, 168, 169, 170, 191, 192, 193, 214, 215, 216, 217, 238, 239, 240, 261, 262, 263, 264, 284, 285, 286, 287], "bbox": [0.29235937500000003, 0.2445667447306792, 0.490265625, 0.8490632318501171], "iscrowd": 0, "area": 14415.368049999997, "rle": {"size": [427, 640], "counts": "kT^29\\<l0UOk0VOj0_Oa0J4L3M2N3M2N3M2N3M2N3N2O001O001O001O001O001O1O001O001O0O2O001O1O0O2O001O0O2O001O1O0O2O001O001N101O1O1O1O10O0N2O2N1N2O1O2M2O1O2M2O1O2M2O1O2M2O1O2M2O1N3M2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3N1N2N2M4M2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2L4K5L4K5K5K5KdcW4"}, "label": "the orange snowboard"}]}
{"id": 271166, "image": "COCO_train2014_000000271166.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange and red snowboard\"."}], "task": "refering", "answer_template": "The \"an orange and red snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 98, 99, 121, 122, 123, 144, 145, 146, 167, 168, 169, 170, 191, 192, 193, 214, 215, 216, 217, 238, 239, 240, 261, 262, 263, 264, 284, 285, 286, 287], "bbox": [0.29235937500000003, 0.2445667447306792, 0.490265625, 0.8490632318501171], "iscrowd": 0, "area": 14415.368049999997, "rle": {"size": [427, 640], "counts": "kT^29\\<l0UOk0VOj0_Oa0J4L3M2N3M2N3M2N3M2N3N2O001O001O001O001O001O1O001O001O0O2O001O1O0O2O001O0O2O001O1O0O2O001O001N101O1O1O1O10O0N2O2N1N2O1O2M2O1O2M2O1O2M2O1O2M2O1O2M2O1N3M2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3N1N2N2M4M2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2L4K5L4K5K5K5KdcW4"}, "label": "an orange and red snowboard"}]}
{"id": 115519, "image": "COCO_train2014_000000115519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with his hand on his chin talking to family\"."}], "task": "refering", "answer_template": "The \"a man with his hand on his chin talking to family\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 316, 317, 318, 338, 339, 340, 341, 361, 362, 384, 385], "bbox": [0.691015625, 0.3438125, 0.8932656250000001, 0.9887708333333334], "iscrowd": 0, "area": 21630.743150000002, "rle": {"size": [480, 640], "counts": "Z^_61l>3N3M3M3M3L3N3M3M3M3QHUOd1n0TNZOj1i0lM_OS2d0cME[2>\\MI\\1mNlJ]1_3Md1nNdJX1ZOeNP4V1o1WO^JQ1]OiNT4n0P2AUJZ1i3VOP2HnIV1P4ROQ2V2mMkMQ2Y2mMfMQ2^2mMcMQ2`2mM`MR2c2lM]MS2f2kM[MS2i2kMVMT2m2jMTMV2m2hMSMY2n2eMSM[2n2dMlKhLh0e5^3`MkKnLd0b5b3^MjKTM`0_5g3[MjKZM:\\5m3YMiK^M7Y5R4VMhKdM2W5`4jL_KSNMR5S5\\LPKfNIo4d5nKcJWOEk4V6_KVJd5X6mIiIS6e6^I\\Ib6`7100000000O1000O1000O100000000O100000000O10O100000O1000001O2M2O1O1O2N1N2O1O1O2N1N6KP1PO1O1O1N2O1O1O1O1N2O1O1O1O1N2O2N1O1O1N2O1O1O1O1N2O1O1K5A?A?A?A?A?TOl0M3L5L3L4M3L4M3L4M3L4I7H8Inco0"}, "label": "a man with his hand on his chin talking to family"}]}
{"id": 115519, "image": "COCO_train2014_000000115519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting at a table across from another man and woman\"."}], "task": "refering", "answer_template": "The \"a man sitting at a table across from another man and woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 316, 317, 318, 338, 339, 340, 341, 361, 362, 384, 385], "bbox": [0.691015625, 0.3438125, 0.8932656250000001, 0.9887708333333334], "iscrowd": 0, "area": 21630.743150000002, "rle": {"size": [480, 640], "counts": "Z^_61l>3N3M3M3M3L3N3M3M3M3QHUOd1n0TNZOj1i0lM_OS2d0cME[2>\\MI\\1mNlJ]1_3Md1nNdJX1ZOeNP4V1o1WO^JQ1]OiNT4n0P2AUJZ1i3VOP2HnIV1P4ROQ2V2mMkMQ2Y2mMfMQ2^2mMcMQ2`2mM`MR2c2lM]MS2f2kM[MS2i2kMVMT2m2jMTMV2m2hMSMY2n2eMSM[2n2dMlKhLh0e5^3`MkKnLd0b5b3^MjKTM`0_5g3[MjKZM:\\5m3YMiK^M7Y5R4VMhKdM2W5`4jL_KSNMR5S5\\LPKfNIo4d5nKcJWOEk4V6_KVJd5X6mIiIS6e6^I\\Ib6`7100000000O1000O1000O100000000O100000000O10O100000O1000001O2M2O1O1O2N1N2O1O1O2N1N6KP1PO1O1O1N2O1O1O1O1N2O1O1O1O1N2O2N1O1O1N2O1O1O1O1N2O1O1K5A?A?A?A?A?TOl0M3L5L3L4M3L4M3L4M3L4I7H8Inco0"}, "label": "a man sitting at a table across from another man and woman"}]}
{"id": 541505, "image": "COCO_train2014_000000541505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mustard - colored jar on the right is the middle - sized jar\"."}], "task": "refering", "answer_template": "The \"the mustard - colored jar on the right is the middle - sized jar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 248, 249, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335], "bbox": [0.49995833333333334, 0.60853125, 0.7290833333333333, 0.8708281249999998], "iscrowd": 0, "area": 16404.87985, "rle": {"size": [640, 480], "counts": "l`f45ba00^@9T>Fo@7CG:f0l>AQA8CH8d0R?Am@9BI7a0X?Bi@;AI6>_?Cd@^2Z?gMa@[2^?T1N2N2O1N2N2D<J60000O10000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000001O:F:F2N1O1O1aL`@3O1N1d?Gc@4N10Ob?He@4L21Kd?Kd@3L4OHf?Nc@2K50Dh?1a@30OLFi?3`@45IW`0Oi_O3hb0Mdi`2"}, "label": "the mustard - colored jar on the right is the middle - sized jar"}]}
{"id": 541505, "image": "COCO_train2014_000000541505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller of the three vases with flowers coming out of it\"."}], "task": "refering", "answer_template": "The \"the taller of the three vases with flowers coming out of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 227, 228, 229, 230, 244, 245, 246, 261, 262, 263, 279, 280, 297, 314, 331], "bbox": [0.3655833333333333, 0.544734375, 0.5958125, 0.854453125], "iscrowd": 0, "area": 8181.0845, "rle": {"size": [640, 480], "counts": "j[^36hc06I6\\]O_Oaa0g0X^O]Ofa0g0U^O]Oha0g0S^O]Oka0e0P^O^Ooa0e0m]O]ORb0Y101N1O100O2N100O1I7J7O00000010O000010O0001O00010O001OL5O0000000O2O2N3M3M3M3M2N2O1N1[@hLM6\\=T3cBnLH6`=m2dBVMC5d=g2eB\\M^O4i=c2cB`N\\=c1_B`N`=d1YB_Nh=l3000O1000O`MUBlNk=S1YBkNg=3WB^N5\\1e=5ZBZN5_1a=5_BWN41Ak0l=l0cBTN30E=T>^1XBPN31G:Q>c1ZBmM21J8m=i1[CjMoN6i=o1T3N2O1O1N2[O\\^OmNea0S1e0O0O10000O10001O2M2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O2N2N2N1O2N2N2N2NPlh3"}, "label": "the taller of the three vases with flowers coming out of it"}]}
{"id": 541505, "image": "COCO_train2014_000000541505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"jar being used as a flower vase\"."}], "task": "refering", "answer_template": "The \"jar being used as a flower vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 227, 228, 229, 230, 244, 245, 246, 261, 262, 263, 279, 280, 297, 314, 331], "bbox": [0.3655833333333333, 0.544734375, 0.5958125, 0.854453125], "iscrowd": 0, "area": 8181.0845, "rle": {"size": [640, 480], "counts": "j[^36hc06I6\\]O_Oaa0g0X^O]Ofa0g0U^O]Oha0g0S^O]Oka0e0P^O^Ooa0e0m]O]ORb0Y101N1O100O2N100O1I7J7O00000010O000010O0001O00010O001OL5O0000000O2O2N3M3M3M3M2N2O1N1[@hLM6\\=T3cBnLH6`=m2dBVMC5d=g2eB\\M^O4i=c2cB`N\\=c1_B`N`=d1YB_Nh=l3000O1000O`MUBlNk=S1YBkNg=3WB^N5\\1e=5ZBZN5_1a=5_BWN41Ak0l=l0cBTN30E=T>^1XBPN31G:Q>c1ZBmM21J8m=i1[CjMoN6i=o1T3N2O1O1N2[O\\^OmNea0S1e0O0O10000O10001O2M2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O2N2N2N1O2N2N2N2NPlh3"}, "label": "jar being used as a flower vase"}]}
{"id": 148292, "image": "COCO_train2014_000000148292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a group of cut crots with herbal seasoning\"."}], "task": "refering", "answer_template": "The \"a group of cut crots with herbal seasoning\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.338859375, 0.6051877934272301, 0.7036562500000001, 0.9844131455399061], "iscrowd": 0, "area": 23940.2807, "rle": {"size": [426, 640], "counts": "lcj22V=4L4L4L4L4L3M4M3L4L4L4M2N3N2N2M3N1O2M3N1O0O2O0O101O0O101O0O101O0O101N10001N10001N10001N100O2O000O2O000O1N3N1N2O2N0O2O1N2O001N2O1O0O2O1N101O1N2O010O1O100O1O2O010O001O001O1O001O0010O01O00001O0000000000010O01O1O1O1O100O1O1O1O1O00100O1O1O00000001OO1O2N1O1O1O1O1O1O1O1O1O1O1O100O2N1O1O1O100000XFXMn8h2QGYMo8h2oFYMQ9g2nFZMR9f2mF[MQ9g2nFZMP9h2nFZMP9h2oFYMo8i2PGXMn8k2PGVMn8l2QGUMm8m2RGTMl8n2RGTMm8m2RGTMl8n2SGSMk8P3SGQMk8Q3TGPMj8R3UGoLi8S3UGoLi8j3N2O1N2O1O10000O10000O10000O100O10000O10000O10000O10YLfG^2Z8bMfG^2Y8cMhG\\2X8dMhG\\2X8cMjG\\2U8eMkG[2U8eMlGZ2T8eMmG[2R8fMoGY2Q8gMoGY2Q8fMQHY2P8fMPHY2Q8gMPHX2Q8fMPHZ2Q8eMoG[2R8dMnG\\2R8cMPH\\2Q8cMoG]2R8bMnG^2R8aMoG_2R8`MnG`2S8_MnG`2R8_MoGa2R8^MnGb2T8\\MlGd2_8QMaGo2j8eLXGZ3\\9SNTF<o9^OUFa0P:XOSFg0R:ROQFm0S:mNQFQ1T:hNoEW1V:bNmE\\1R;O1O2M2O1O1O2M2O1J6@Pf^2"}, "label": "a group of cut crots with herbal seasoning"}]}
{"id": 148292, "image": "COCO_train2014_000000148292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slices of orange carrots surround by meat and yellow and green vegetables\"."}], "task": "refering", "answer_template": "The \"slices of orange carrots surround by meat and yellow and green vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.338859375, 0.6051877934272301, 0.7036562500000001, 0.9844131455399061], "iscrowd": 0, "area": 23940.2807, "rle": {"size": [426, 640], "counts": "lcj22V=4L4L4L4L4L3M4M3L4L4L4M2N3N2N2M3N1O2M3N1O0O2O0O101O0O101O0O101O0O101N10001N10001N10001N100O2O000O2O000O1N3N1N2O2N0O2O1N2O001N2O1O0O2O1N101O1N2O010O1O100O1O2O010O001O001O1O001O0010O01O00001O0000000000010O01O1O1O1O100O1O1O1O1O00100O1O1O00000001OO1O2N1O1O1O1O1O1O1O1O1O1O1O100O2N1O1O1O100000XFXMn8h2QGYMo8h2oFYMQ9g2nFZMR9f2mF[MQ9g2nFZMP9h2nFZMP9h2oFYMo8i2PGXMn8k2PGVMn8l2QGUMm8m2RGTMl8n2RGTMm8m2RGTMl8n2SGSMk8P3SGQMk8Q3TGPMj8R3UGoLi8S3UGoLi8j3N2O1N2O1O10000O10000O10000O100O10000O10000O10000O10YLfG^2Z8bMfG^2Y8cMhG\\2X8dMhG\\2X8cMjG\\2U8eMkG[2U8eMlGZ2T8eMmG[2R8fMoGY2Q8gMoGY2Q8fMQHY2P8fMPHY2Q8gMPHX2Q8fMPHZ2Q8eMoG[2R8dMnG\\2R8cMPH\\2Q8cMoG]2R8bMnG^2R8aMoG_2R8`MnG`2S8_MnG`2R8_MoGa2R8^MnGb2T8\\MlGd2_8QMaGo2j8eLXGZ3\\9SNTF<o9^OUFa0P:XOSFg0R:ROQFm0S:mNQFQ1T:hNoEW1V:bNmE\\1R;O1O2M2O1O1O2M2O1J6@Pf^2"}, "label": "slices of orange carrots surround by meat and yellow and green vegetables"}]}
{"id": 9029, "image": "COCO_train2014_000000009029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red train on the tracks\"."}], "task": "refering", "answer_template": "The \"a red train on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 253, 254, 255, 276, 277, 278], "bbox": [0.0015, 0.25169411764705885, 0.19996874999999997, 0.8719058823529411], "iscrowd": 0, "area": 25006.0686, "rle": {"size": [425, 640], "counts": "d`0T4U9S4mK0001N10000O10000O2O000O10000O101O0O100O10000O2O000O10000O101O0O10000O101O0O10000O100O2O000O10000O101O0O10000O10001N10000O10000O2O000O100O10001L3M3L4M3M3M3M4K4POTI`Ko6`4_IoJd6P5k0O100O2N100O1O10000010O0000001O0001O01OO100000001O0iMeI\\M_6_2X2K5K5K5K6J5K5K5K5K5K5K6J5K5K5K5KX\\d6"}, "label": "a red train on the tracks"}]}
{"id": 9029, "image": "COCO_train2014_000000009029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red train on the other side of a green one with no passengers\"."}], "task": "refering", "answer_template": "The \"red train on the other side of a green one with no passengers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 253, 254, 255, 276, 277, 278], "bbox": [0.0015, 0.25169411764705885, 0.19996874999999997, 0.8719058823529411], "iscrowd": 0, "area": 25006.0686, "rle": {"size": [425, 640], "counts": "d`0T4U9S4mK0001N10000O10000O2O000O10000O101O0O100O10000O2O000O10000O101O0O10000O101O0O10000O100O2O000O10000O101O0O10000O10001N10000O10000O2O000O100O10001L3M3L4M3M3M3M4K4POTI`Ko6`4_IoJd6P5k0O100O2N100O1O10000010O0000001O0001O01OO100000001O0iMeI\\M_6_2X2K5K5K5K6J5K5K5K5K5K5K6J5K5K5K5KX\\d6"}, "label": "red train on the other side of a green one with no passengers"}]}
{"id": 9029, "image": "COCO_train2014_000000009029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the entire train that is closest to the people\"."}], "task": "refering", "answer_template": "The \"the entire train that is closest to the people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 220, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 303, 304, 305], "bbox": [0.13878125, 0.23077647058823528, 0.7998593749999999, 0.8936941176470589], "iscrowd": 0, "area": 58001.20180000001, "rle": {"size": [425, 640], "counts": "cWU12U=4M3L4M3L4M3L4L4M3L4M3L4M3L4L4M3L4M3L4L4M3L4M3_MkMhIY2R6^NZIe1`6QOlHS1m6D_H`0[7j2K5K5K5K5K5K5K5K5K5K5K5L4O1O1N2O1N2O000O010O10000O100O10000O010O10000O100O10000O10O010000O100O10000O10O010000O100O10000O100O01000O100O10O1O0010O01O1O001O10O01O001O10O01O001O1O101N1O1O1O1O1O2O0O1O1O1O1O2N100O1O1O2N1O1O100O1O2N1O1O1O101N1O1O1O1O1O2O0O1^OfITJ[6l5eISJ\\6m5dIRJ]6n5dIPJ^6o5bIPJ_6o5cInI_6R6aImI`6S6aIkI`6U6`IiIc6V6^IhIc6X6700O10001O0000000O2O00000000001N10000000001O0O1000001O00000O10001O0000000O101O000000001N1N2O1O1N2O2M2O1O1N2O2M2O1O2M2O1N3N1O1N3N1WNaGfNa8X1cGeN_8Y1cGdN_8Z1eGcN\\8[1hGbNY8]1iG`NZ8]1jG`NW8_1kG^NW8`1mG]NU8a1nG\\NS8b1k1O2M2N2O2M2M3N3L3M3M4M2N2O1O2O0O1O2N100O2N1O101N11O100O001O100O1O100O00100O1O100O1eFfNe6Z1YIhNh6X1UIiNl6X1QIjNo6X1lHjNV7^1[G\\N8:]8_3100O2O0O101N100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O101N1TNcEl0]:TOoE?S:@TF9l9G[F1f9OaFJ_96hFBZ9<nF]OR9c0j101O00001O000O20O1O101N6J6J5L5J6J6K4K6J6K0OO101N1O1]NjDR1X;mNlDo0T;QOb110O0000001O01O10O01O01O00001O00001O0000001O00001O00001O0O101O00001O00001O001O001O00001O001O0000Z=0Ynd1"}, "label": "the entire train that is closest to the people"}]}
{"id": 9029, "image": "COCO_train2014_000000009029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green train with a yellow front\"."}], "task": "refering", "answer_template": "The \"a green train with a yellow front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 220, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 303, 304, 305], "bbox": [0.13878125, 0.23077647058823528, 0.7998593749999999, 0.8936941176470589], "iscrowd": 0, "area": 58001.20180000001, "rle": {"size": [425, 640], "counts": "cWU12U=4M3L4M3L4M3L4L4M3L4M3L4M3L4L4M3L4M3L4L4M3L4M3_MkMhIY2R6^NZIe1`6QOlHS1m6D_H`0[7j2K5K5K5K5K5K5K5K5K5K5K5L4O1O1N2O1N2O000O010O10000O100O10000O010O10000O100O10000O10O010000O100O10000O10O010000O100O10000O100O01000O100O10O1O0010O01O1O001O10O01O001O10O01O001O1O101N1O1O1O1O1O2O0O1O1O1O1O2N100O1O1O2N1O1O100O1O2N1O1O1O101N1O1O1O1O1O2O0O1^OfITJ[6l5eISJ\\6m5dIRJ]6n5dIPJ^6o5bIPJ_6o5cInI_6R6aImI`6S6aIkI`6U6`IiIc6V6^IhIc6X6700O10001O0000000O2O00000000001N10000000001O0O1000001O00000O10001O0000000O101O000000001N1N2O1O1N2O2M2O1O1N2O2M2O1O2M2O1N3N1O1N3N1WNaGfNa8X1cGeN_8Y1cGdN_8Z1eGcN\\8[1hGbNY8]1iG`NZ8]1jG`NW8_1kG^NW8`1mG]NU8a1nG\\NS8b1k1O2M2N2O2M2M3N3L3M3M4M2N2O1O2O0O1O2N100O2N1O101N11O100O001O100O1O100O00100O1O100O1eFfNe6Z1YIhNh6X1UIiNl6X1QIjNo6X1lHjNV7^1[G\\N8:]8_3100O2O0O101N100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O101N1TNcEl0]:TOoE?S:@TF9l9G[F1f9OaFJ_96hFBZ9<nF]OR9c0j101O00001O000O20O1O101N6J6J5L5J6J6K4K6J6K0OO101N1O1]NjDR1X;mNlDo0T;QOb110O0000001O01O10O01O01O00001O00001O0000001O00001O00001O0O101O00001O00001O001O001O00001O001O0000Z=0Ynd1"}, "label": "a green train with a yellow front"}]}
{"id": 484171, "image": "COCO_train2014_000000484171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dining knife\"."}], "task": "refering", "answer_template": "The \"the dining knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 159, 160, 181, 182, 204, 205, 227, 228, 250, 251, 273, 274, 296, 297, 319, 320, 342, 343, 365, 366, 388, 389], "bbox": [0.86965625, 0.2426875, 0.964046875, 0.9887708333333334], "iscrowd": 0, "area": 9625.579100000012, "rle": {"size": [480, 640], "counts": "iaU82h>6I7J5K6J6I7J6J6J6I7J6J6J6I7J6J6J6J6J6PM_M\\If2o4ROhJT1T5SOdJR1Y5TO^JR1^5UOZJP1b5WOUJo0h5V4L4M3L4L4M3L4L4M3L4L4M_KbKcN]4_1`41O2N1O1O2N1O1100O2N101N2N1M4XOg0YOh0XOh0WOh0YOh0XOh0WOTf:"}, "label": "the dining knife"}]}
{"id": 484171, "image": "COCO_train2014_000000484171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver knife next to the plate of food\"."}], "task": "refering", "answer_template": "The \"the silver knife next to the plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 159, 160, 181, 182, 204, 205, 227, 228, 250, 251, 273, 274, 296, 297, 319, 320, 342, 343, 365, 366, 388, 389], "bbox": [0.86965625, 0.2426875, 0.964046875, 0.9887708333333334], "iscrowd": 0, "area": 9625.579100000012, "rle": {"size": [480, 640], "counts": "iaU82h>6I7J5K6J6I7J6J6J6I7J6J6J6I7J6J6J6J6J6PM_M\\If2o4ROhJT1T5SOdJR1Y5TO^JR1^5UOZJP1b5WOUJo0h5V4L4M3L4L4M3L4L4M3L4L4M_KbKcN]4_1`41O2N1O1O2N1O1100O2N101N2N1M4XOg0YOh0XOh0WOh0YOh0XOh0WOTf:"}, "label": "the silver knife next to the plate of food"}]}
{"id": 426829, "image": "COCO_train2014_000000426829.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bubba gump novelty fortune magazine\"."}], "task": "refering", "answer_template": "The \"bubba gump novelty fortune magazine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.0458125, 0.09287337662337662, 0.381953125, 0.49749999999999994], "iscrowd": 0, "area": 44754.588599999995, "rle": {"size": [616, 640], "counts": "fSb09ob08H8H7I8H8H8H7I8H8G9H7I8H8H8H7I8H8H8H7I8H8H8H7I8H8G9H7I8H8H2N0000000000000000O10O100000000000000000O1000000000O1000000000O100000000000000000O10O100000000000000000000O1000O1000000000000000O100000000000O10000000O1000000000000000000O010000000000000000000O10000000O100000000000O100000000000000000O10O100000000000000000000O1000O1000000000000000O100000000000O10000000O1000000000000000000O01000000003M3M3M4nIbER3d=F:E;F9G:F:F8H9G9G8H9G9G8GXn]7"}, "label": "bubba gump novelty fortune magazine"}]}
{"id": 426829, "image": "COCO_train2014_000000426829.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"magazine next to feather\"."}], "task": "refering", "answer_template": "The \"magazine next to feather\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.0458125, 0.09287337662337662, 0.381953125, 0.49749999999999994], "iscrowd": 0, "area": 44754.588599999995, "rle": {"size": [616, 640], "counts": "fSb09ob08H8H7I8H8H8H7I8H8G9H7I8H8H8H7I8H8H8H7I8H8H8H7I8H8G9H7I8H8H2N0000000000000000O10O100000000000000000O1000000000O1000000000O100000000000000000O10O100000000000000000000O1000O1000000000000000O100000000000O10000000O1000000000000000000O010000000000000000000O10000000O100000000000O100000000000000000O10O100000000000000000000O1000O1000000000000000O100000000000O10000000O1000000000000000000O01000000003M3M3M4nIbER3d=F:E;F9G:F:F8H9G9G8H9G9G8GXn]7"}, "label": "magazine next to feather"}]}
{"id": 426829, "image": "COCO_train2014_000000426829.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow book\"."}], "task": "refering", "answer_template": "The \"a yellow book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314], "bbox": [0.5274375, 0.4201948051948052, 0.7875, 0.6272727272727272], "iscrowd": 0, "area": 12949.53245, "rle": {"size": [616, 640], "counts": "ob[61Wc0001O1O1O3M2M4M3M1O001O001O00001O000O2O001O1O1O1O2N1O1O1O1N3N1O1g^OnNZ`0S1__OVO^`0l0]_OZO``0g0^_O^O^`0c0`_OA]`0`0a_OD\\`0=b_OHZ`0:c_OJZ`07d_ONX`0d1L4M4L3M4L3M3M1O00O1000O0100O10000O10000O100O10O10O100O10000O10000O10O010000O100O10000O1000O0100O10000O10000O100O10O10O100O10000O10000O10O010000O100O10000O10000O010O10000O10000O100O10O10O100O10000O101O3L4M4K4M3M4K4M3L5L3M3L5L3M3L5L3L4M4L3L4M3M4K4M3L^_a2"}, "label": "a yellow book"}]}
{"id": 426829, "image": "COCO_train2014_000000426829.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"curious george yellow book\"."}], "task": "refering", "answer_template": "The \"curious george yellow book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314], "bbox": [0.5274375, 0.4201948051948052, 0.7875, 0.6272727272727272], "iscrowd": 0, "area": 12949.53245, "rle": {"size": [616, 640], "counts": "ob[61Wc0001O1O1O3M2M4M3M1O001O001O00001O000O2O001O1O1O1O2N1O1O1O1N3N1O1g^OnNZ`0S1__OVO^`0l0]_OZO``0g0^_O^O^`0c0`_OA]`0`0a_OD\\`0=b_OHZ`0:c_OJZ`07d_ONX`0d1L4M4L3M4L3M3M1O00O1000O0100O10000O10000O100O10O10O100O10000O10000O10O010000O100O10000O1000O0100O10000O10000O100O10O10O100O10000O10000O10O010000O100O10000O10000O010O10000O10000O100O10O10O100O10000O101O3L4M4K4M3M4K4M3L5L3M3L5L3M3L5L3L4M4L3L4M3M4K4M3L^_a2"}, "label": "curious george yellow book"}]}
{"id": 451407, "image": "COCO_train2014_000000451407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of cheesy pizza on a white plate with a fork and knife on the right side\"."}], "task": "refering", "answer_template": "The \"a slice of cheesy pizza on a white plate with a fork and knife on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.18696875, 0.47640845070422533, 0.617765625, 0.7977464788732394], "iscrowd": 0, "area": 22899.84859999999, "rle": {"size": [426, 640], "counts": "VWb14T=4K4L5K4M3L5K4M4K4L4L4M2M4L3M4M2M4L3N3N1N3N1N3N1N2O1N101N2O1N2O1N2O1O010O1O1O1O100O001O100O1O1O10O01O1O1O100O1O1O010O1O1O100O1O001O100O1O1O100O001O1000000O10O1001O0000000O101O0000000000001N1000000001O100O001O1O1O10O01O1O1O001O100O1O001O1O100O001O1O1O1O010O1O1O1O001O100O1O001O1O100O001O1O1O1O010O1O1O1O001O1O1O001O00001O00001O00001O001O00010O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O000000001O0000010O0000001O000000001O000O1O2N1O1O1N2O2N1O1O1O2N1O2N2N2N1N3N2N2N2N2O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O10O0001O00001O001O0O1H]dU3"}, "label": "a slice of cheesy pizza on a white plate with a fork and knife on the right side"}]}
{"id": 451407, "image": "COCO_train2014_000000451407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of pizza on a plate with fork and knife\"."}], "task": "refering", "answer_template": "The \"slice of pizza on a plate with fork and knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.18696875, 0.47640845070422533, 0.617765625, 0.7977464788732394], "iscrowd": 0, "area": 22899.84859999999, "rle": {"size": [426, 640], "counts": "VWb14T=4K4L5K4M3L5K4M4K4L4L4M2M4L3M4M2M4L3N3N1N3N1N3N1N2O1N101N2O1N2O1N2O1O010O1O1O1O100O001O100O1O1O10O01O1O1O100O1O1O010O1O1O100O1O001O100O1O1O100O001O1000000O10O1001O0000000O101O0000000000001N1000000001O100O001O1O1O10O01O1O1O001O100O1O001O1O100O001O1O1O1O010O1O1O1O001O100O1O001O1O100O001O1O1O1O010O1O1O1O001O1O1O001O00001O00001O00001O001O00010O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O000000001O0000010O0000001O000000001O000O1O2N1O1O1N2O2N1O1O1O2N1O2N2N2N1N3N2N2N2N2O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O10O0001O00001O001O0O1H]dU3"}, "label": "slice of pizza on a plate with fork and knife"}]}
{"id": 287567, "image": "COCO_train2014_000000287567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and blak dog in the front of another dog\"."}], "task": "refering", "answer_template": "The \"a white and blak dog in the front of another dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [320, 321, 322, 342, 343, 344, 345, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 403, 404, 405, 406, 407, 408, 409, 410, 411, 425, 426, 427, 428, 429, 430, 431, 432, 433, 447, 448, 449, 450, 451, 452, 453, 454, 455, 474, 475], "bbox": [0.32965231788079474, 0.62828125, 0.7213907284768213, 0.941421875], "iscrowd": 0, "area": 28212.24514999999, "rle": {"size": [640, 604], "counts": "Yll34lc0>B4L3M4L3M4L3N3L2N00000000O1000000001O00010O01O10FfNl]OZ1Rb0iNn]OV1Qb0lNo]OT1na0oNR^OP1ma0ROR^On0la0UOT^Oj0ka0XOU^Oh0ha0[OW^Oe0ha0^OW^Oa0ga0BY^O>ea0DZ^O<aa0kN\\^Ol038`a0lN]^On035^a01a^OO^a03b^OL\\a07d^OH[a0:e^OEYa0>f^OBXa0a0h^O^OWa0e0h^OZOVa0i1O1N2O1N2O1O1O1O100O1O1O100O2N1O2O0O2N1O101N1O2N101N1O2N100O2N1O2O0O2N00100O1O1O100O100000O1000000000000000000O10000000000000O1000O100000000000000000O0100O010O01000O01J7J6J5L5N2N2O1N2N2N1O2N2O1N2O1N2N2O0O2O2M2O1N3M2O2M2O2M2O1N00100O100O1000TO\\B\\Ke=a4_B]Ka=a4cB]K]=a4hB\\KX=b4lB\\KT=b4QC[Ko<d4TCYKm<e4XCXKh<f4\\CXKe<e4`CXK`<f4^1O1N2N2O101O000000000000001O0000001O1O1O20O1O100O2N100O1O100O10000O10O0N3N2M2N3N2M2O2M2O2M3M2O2mNm_OTNV`0e1Q@YNP`0`1X@]Nk?\\1\\@bNe?W1c@fN`?S1g@jN\\?n0l@POU?j0RASOQ?e0VAZOl>?[A^Og>;aABb>7g2J^aX3"}, "label": "a white and blak dog in the front of another dog"}]}
{"id": 287567, "image": "COCO_train2014_000000287567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mountain dog who is sitting down in the snow in front of the other dog and person\"."}], "task": "refering", "answer_template": "The \"the mountain dog who is sitting down in the snow in front of the other dog and person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [320, 321, 322, 342, 343, 344, 345, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 403, 404, 405, 406, 407, 408, 409, 410, 411, 425, 426, 427, 428, 429, 430, 431, 432, 433, 447, 448, 449, 450, 451, 452, 453, 454, 455, 474, 475], "bbox": [0.32965231788079474, 0.62828125, 0.7213907284768213, 0.941421875], "iscrowd": 0, "area": 28212.24514999999, "rle": {"size": [640, 604], "counts": "Yll34lc0>B4L3M4L3M4L3N3L2N00000000O1000000001O00010O01O10FfNl]OZ1Rb0iNn]OV1Qb0lNo]OT1na0oNR^OP1ma0ROR^On0la0UOT^Oj0ka0XOU^Oh0ha0[OW^Oe0ha0^OW^Oa0ga0BY^O>ea0DZ^O<aa0kN\\^Ol038`a0lN]^On035^a01a^OO^a03b^OL\\a07d^OH[a0:e^OEYa0>f^OBXa0a0h^O^OWa0e0h^OZOVa0i1O1N2O1N2O1O1O1O100O1O1O100O2N1O2O0O2N1O101N1O2N101N1O2N100O2N1O2O0O2N00100O1O1O100O100000O1000000000000000000O10000000000000O1000O100000000000000000O0100O010O01000O01J7J6J5L5N2N2O1N2N2N1O2N2O1N2O1N2N2O0O2O2M2O1N3M2O2M2O2M2O1N00100O100O1000TO\\B\\Ke=a4_B]Ka=a4cB]K]=a4hB\\KX=b4lB\\KT=b4QC[Ko<d4TCYKm<e4XCXKh<f4\\CXKe<e4`CXK`<f4^1O1N2N2O101O000000000000001O0000001O1O1O20O1O100O2N100O1O100O10000O10O0N3N2M2N3N2M2O2M2O2M3M2O2mNm_OTNV`0e1Q@YNP`0`1X@]Nk?\\1\\@bNe?W1c@fN`?S1g@jN\\?n0l@POU?j0RASOQ?e0VAZOl>?[A^Og>;aABb>7g2J^aX3"}, "label": "the mountain dog who is sitting down in the snow in front of the other dog and person"}]}
{"id": 287567, "image": "COCO_train2014_000000287567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dog standing between a human and another dog\"."}], "task": "refering", "answer_template": "The \"a black dog standing between a human and another dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 254, 271, 272, 273, 274, 275, 276, 277, 293, 294, 295, 296, 297, 298, 299, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 360, 361], "bbox": [0.3429139072847682, 0.49378124999999995, 0.6394205298013246, 0.7248906249999999], "iscrowd": 0, "area": 16643.97425000001, "rle": {"size": [640, 604], "counts": "bfQ4>ac06J4K3N2N3L3N2c]OTOca0n0W^O^Oaa0d0Y^OEaa0a1L4L4M2N2M3g_OSM[?o2b@ZMV?g2g@cMP?`2m@eMo>]2m@eMS?]2j@dMV?^2f@cMZ?`2c@aM]?E`@X3`?gLa@Y3^?hLb@X3^?gLc@Y3]?gLc@X3]?hLd@X3\\?hLd@X3\\?gLe@Y3[?gLd@Z3[?fLf@Y3[?gLe@Y3[?fLf@Z3Y?gLg@Y3Y?fLh@Z3X?fLg@[3Y?dLh@[3Y?dLh@\\3X?dLh@\\3X?cLh@]3Y?bLh@[3[?cLg@Y3]?dLf@Z3\\?dLe@[3]?cLe@\\3\\?bLf@]3o?N101O1N2O1N101001O1O001ON1N3N2M3O11O001O0O2O00001O010O01O01O010O01O01O010O010O00010O010O010O1O1O1O1O000O10000O100O10000O10000O10000O01O0O2O1N1O2O1N2O1M3L4L40O01O100O100O1O100O100O00100O101N1O100O2O0O1O101N1O100O2O0O1O101O0O2O001O001O001O001O0L5H7I8K5O2N3M2N2M3GTmW4"}, "label": "a black dog standing between a human and another dog"}]}
{"id": 287567, "image": "COCO_train2014_000000287567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog with its mouth closed\"."}], "task": "refering", "answer_template": "The \"the dog with its mouth closed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 254, 271, 272, 273, 274, 275, 276, 277, 293, 294, 295, 296, 297, 298, 299, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 360, 361], "bbox": [0.3429139072847682, 0.49378124999999995, 0.6394205298013246, 0.7248906249999999], "iscrowd": 0, "area": 16643.97425000001, "rle": {"size": [640, 604], "counts": "bfQ4>ac06J4K3N2N3L3N2c]OTOca0n0W^O^Oaa0d0Y^OEaa0a1L4L4M2N2M3g_OSM[?o2b@ZMV?g2g@cMP?`2m@eMo>]2m@eMS?]2j@dMV?^2f@cMZ?`2c@aM]?E`@X3`?gLa@Y3^?hLb@X3^?gLc@Y3]?gLc@X3]?hLd@X3\\?hLd@X3\\?gLe@Y3[?gLd@Z3[?fLf@Y3[?gLe@Y3[?fLf@Z3Y?gLg@Y3Y?fLh@Z3X?fLg@[3Y?dLh@[3Y?dLh@\\3X?dLh@\\3X?cLh@]3Y?bLh@[3[?cLg@Y3]?dLf@Z3\\?dLe@[3]?cLe@\\3\\?bLf@]3o?N101O1N2O1N101001O1O001ON1N3N2M3O11O001O0O2O00001O010O01O01O010O01O01O010O010O00010O010O010O1O1O1O1O000O10000O100O10000O10000O10000O01O0O2O1N1O2O1N2O1M3L4L40O01O100O100O1O100O100O00100O101N1O100O2O0O1O101N1O100O2O0O1O101O0O2O001O001O001O001O0L5H7I8K5O2N3M2N2M3GTmW4"}, "label": "the dog with its mouth closed"}]}
{"id": 279377, "image": "COCO_train2014_000000279377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the darker dog on the right\"."}], "task": "refering", "answer_template": "The \"the darker dog on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 247, 266, 267], "bbox": [0.540828125, 0.2592723004694836, 0.7890937499999999, 0.7673474178403755], "iscrowd": 0, "area": 19826.473749999997, "rle": {"size": [426, 640], "counts": "PS`42W=6I6J7I7J5J7I7K4K6K1O1N2O1O0O2O1O1N101O1N2VHlM\\4U2`KnM`4R2]KRNa4P2[KTNd4m1WKWNi4j1SKZNl4f1kJdNT5]1_JoNa5R1RJ[Om5e0fIIY66ZI8f6HQIb0n6^OPIe0o6ZOPIh0P7XOnHi0S7WOkHk0U7UOiHm0W7ROhHP1X7POfHQ1[7oNbHT1^7\\2000O1000000OO2O1N2O0O2O1N20O01O10`NQIbLQ7[3UIaLl6]3ZI]Li6_3^I\\Lc6b3cIZL]6f3gIVLZ6i3jITLU6l3mISLR6m3PJQLQ6n3QJQLn5o3TJPLk5P4VJoKk5P4WJoKh5Q4ZJnKe5R4]JlKd5S4]JmKb5T4i1101N1O101N101O00001O001O1O1O1O1O100O1O100OO101O00001O00001O0N2M4L3M3M4L3N2M4N10001O00001O01O;F9F7I3N2M3M3N2M3M[LkGV2U8gMnGX2R8eMRH[2o7`MVH^2R8WMRHh2V8mLoGR3X8cLlG\\3S9N101N1O2O0N3L3M4L4L3TO^ElNf:R1`EfNd:W1cEaN`:]1i0M3N2M2N3M3M2O2Bkng1"}, "label": "the darker dog on the right"}]}
{"id": 279377, "image": "COCO_train2014_000000279377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hairy dog with his friend\"."}], "task": "refering", "answer_template": "The \"a hairy dog with his friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 247, 266, 267], "bbox": [0.540828125, 0.2592723004694836, 0.7890937499999999, 0.7673474178403755], "iscrowd": 0, "area": 19826.473749999997, "rle": {"size": [426, 640], "counts": "PS`42W=6I6J7I7J5J7I7K4K6K1O1N2O1O0O2O1O1N101O1N2VHlM\\4U2`KnM`4R2]KRNa4P2[KTNd4m1WKWNi4j1SKZNl4f1kJdNT5]1_JoNa5R1RJ[Om5e0fIIY66ZI8f6HQIb0n6^OPIe0o6ZOPIh0P7XOnHi0S7WOkHk0U7UOiHm0W7ROhHP1X7POfHQ1[7oNbHT1^7\\2000O1000000OO2O1N2O0O2O1N20O01O10`NQIbLQ7[3UIaLl6]3ZI]Li6_3^I\\Lc6b3cIZL]6f3gIVLZ6i3jITLU6l3mISLR6m3PJQLQ6n3QJQLn5o3TJPLk5P4VJoKk5P4WJoKh5Q4ZJnKe5R4]JlKd5S4]JmKb5T4i1101N1O101N101O00001O001O1O1O1O1O100O1O100OO101O00001O00001O0N2M4L3M3M4L3N2M4N10001O00001O01O;F9F7I3N2M3M3N2M3M[LkGV2U8gMnGX2R8eMRH[2o7`MVH^2R8WMRHh2V8mLoGR3X8cLlG\\3S9N101N1O2O0N3L3M4L4L3TO^ElNf:R1`EfNd:W1cEaN`:]1i0M3N2M2N3M3M2O2Bkng1"}, "label": "a hairy dog with his friend"}]}
{"id": 279377, "image": "COCO_train2014_000000279377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog on the left\"."}], "task": "refering", "answer_template": "The \"the dog on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239], "bbox": [0.299875, 0.1696948356807512, 0.5526718749999999, 0.6767605633802817], "iscrowd": 0, "area": 17165.771599999996, "rle": {"size": [426, 640], "counts": "To_23V=5K5K6J5L2M2N1O2N1O1O2N1O2O0Oo0QO;E<aHVMh3T3jKQMT4U3cKPM\\4S3]KRMb4Q3WKSMh4P3UKQMk4R3QKoLn4U3nJlLR5V3lJjLT5X3iJiLV5[3fJfLZ5\\3cJeL]5]3oJULP5o3gJWLY5k3^J\\La5g3VJ`Lj5c3mIcLR6`3eIgL[6[3]IkLb6o40O001O001O001B=L5K4L5L4K6J5K5L4K5K6J5L4N210O010O00100O0010O01O0010O0QIlKi4S4UKoKk4R4QKQLP5n3mJULT5j3iJYLY5e3eJ^L[5b3aJaL`5^3]JeLe5]3TJfLm5\\3mIgLT6\\3fIfL\\6[3^IiLb6Z3WIiLj6Y3PIjLR7X3gHkLZ7b42N2O1N2N2N2N1_Ob0eNZ1A`0M3N1N3M1O2N2O1N101O1N2O001N2O0O2O1O1N101O1N101N2O1O0O2O1O0O10001O0000001O000O101O0000001O0000001O000O20O100O01O014K4L4L5L3L4L8I7H7I8Hohf3"}, "label": "the dog on the left"}]}
{"id": 279377, "image": "COCO_train2014_000000279377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog ot the left of another dog\"."}], "task": "refering", "answer_template": "The \"a dog ot the left of another dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239], "bbox": [0.299875, 0.1696948356807512, 0.5526718749999999, 0.6767605633802817], "iscrowd": 0, "area": 17165.771599999996, "rle": {"size": [426, 640], "counts": "To_23V=5K5K6J5L2M2N1O2N1O1O2N1O2O0Oo0QO;E<aHVMh3T3jKQMT4U3cKPM\\4S3]KRMb4Q3WKSMh4P3UKQMk4R3QKoLn4U3nJlLR5V3lJjLT5X3iJiLV5[3fJfLZ5\\3cJeL]5]3oJULP5o3gJWLY5k3^J\\La5g3VJ`Lj5c3mIcLR6`3eIgL[6[3]IkLb6o40O001O001O001B=L5K4L5L4K6J5K5L4K5K6J5L4N210O010O00100O0010O01O0010O0QIlKi4S4UKoKk4R4QKQLP5n3mJULT5j3iJYLY5e3eJ^L[5b3aJaL`5^3]JeLe5]3TJfLm5\\3mIgLT6\\3fIfL\\6[3^IiLb6Z3WIiLj6Y3PIjLR7X3gHkLZ7b42N2O1N2N2N2N1_Ob0eNZ1A`0M3N1N3M1O2N2O1N101O1N2O001N2O0O2O1O1N101O1N101N2O1O0O2O1O0O10001O0000001O000O101O0000001O0000001O000O20O100O01O014K4L4L5L3L4L8I7H7I8Hohf3"}, "label": "a dog ot the left of another dog"}]}
{"id": 295759, "image": "COCO_train2014_000000295759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dragonfly kite being flown\"."}], "task": "refering", "answer_template": "The \"dragonfly kite being flown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 12, 13, 14, 15, 16, 17, 18, 19, 20, 33, 37, 38, 39, 40, 43, 56, 60, 61, 66, 79, 84, 89, 101, 107, 112, 124, 130, 134, 135, 157, 179, 180, 202, 225, 247, 248], "bbox": [0.391515625, 8.547008547008547e-05, 0.905375, 0.6543803418803419], "iscrowd": 0, "area": 9536.174599999997, "rle": {"size": [468, 640], "counts": "ckb32a>1N3N2N2M2O0000O010O1O10O0O2O000O2O0O2O000O2O00M2O2M201O1O0O2O1O00]MLhF4\\9N^F3f9NSF4R:MgE4_:KRE>T;]16Jc1]N\\k`0JmT_O2N2N1O2N0000001O000000001O0000001O00001O1O001O1O001O1O1O1O001O1O1O1O1O1O001O1O1OO100000000001O0000001O00001O001O00001O00001O1O001O1O1O1O001O1O001O001O1O001O1O001O8H:F:F:F6J6J6J6J9G:F:F9G:F9GM3M3M3M3VM\\F?i9\\OfF3a9FPGGY92WG[OS9d0_2000000O1000000O1O100O1O1O1O1O2O0O10000O100O10000O100000000000000000000000000001O000000O100O10hJTOhLk0X3XOfLh0X3\\OfLc0Z3]OgLc0X3\\OjLc0V3]OkLc0S3^OnLa0R3^OPMb0o2^ORMa0n2^OTMb0j2_OWM`0i2@XM`0g2@ZM?f2A[M?c2B^M=b2B`M>_2BbM>]2BdM>Z2CgM=X2CiM=V2CkM=S2DnM=o1DRN<l1DVN<h1EYN<e1C]N=a1BbN>\\1AgN`0V1^OnNb0Q1\\OROd0l0[OWOf0=BF>6CM=OE3;IG99GD<<CC?<BB`0>VOIm07QOIQ16mNKU15hNK[14cNM_13^NNd11ZNOi11TN0n1OoM3S2MgM7[2HdM5a2K^M2f2MZM0j20TMMQ32oLJV36hLG]38cLG_39_LGc38\\LIe37YLIi36VLKk35SLLn33QLMQ43mKNT41kK0V40hK0Z40eK0\\4OdK1]4ObK0`4O`K2`4N_K2b4M_K2b4N^K1c4N^K1c4O\\K1e4N\\K2d4N\\K1e4NRKKjK6U9OjJ5mKLZ9OdJ<PLD]9OaJ0\\L6T9J]J0dL3P9LZJ1jL1m8NnG?4BUO0i8OgGn00SODNf82aGc0ROFf0J5Ic84\\G>EKG0_:GRF88=n9[OfE6h0Nm9LWE7g<IVC6n<JnB7U=IgB8\\=H`B8e=GWB9m=08I6M4N10Plk0"}, "label": "dragonfly kite being flown"}]}
{"id": 295759, "image": "COCO_train2014_000000295759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a multi - colored kite in the sky with woman who has her arms outstretched towards it\"."}], "task": "refering", "answer_template": "The \"a multi - colored kite in the sky with woman who has her arms outstretched towards it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 12, 13, 14, 15, 16, 17, 18, 19, 20, 33, 37, 38, 39, 40, 43, 56, 60, 61, 66, 79, 84, 89, 101, 107, 112, 124, 130, 134, 135, 157, 179, 180, 202, 225, 247, 248], "bbox": [0.391515625, 8.547008547008547e-05, 0.905375, 0.6543803418803419], "iscrowd": 0, "area": 9536.174599999997, "rle": {"size": [468, 640], "counts": "ckb32a>1N3N2N2M2O0000O010O1O10O0O2O000O2O0O2O000O2O00M2O2M201O1O0O2O1O00]MLhF4\\9N^F3f9NSF4R:MgE4_:KRE>T;]16Jc1]N\\k`0JmT_O2N2N1O2N0000001O000000001O0000001O00001O1O001O1O001O1O1O1O001O1O1O1O1O1O001O1O1OO100000000001O0000001O00001O001O00001O00001O1O001O1O1O1O001O1O001O001O1O001O1O001O8H:F:F:F6J6J6J6J9G:F:F9G:F9GM3M3M3M3VM\\F?i9\\OfF3a9FPGGY92WG[OS9d0_2000000O1000000O1O100O1O1O1O1O2O0O10000O100O10000O100000000000000000000000000001O000000O100O10hJTOhLk0X3XOfLh0X3\\OfLc0Z3]OgLc0X3\\OjLc0V3]OkLc0S3^OnLa0R3^OPMb0o2^ORMa0n2^OTMb0j2_OWM`0i2@XM`0g2@ZM?f2A[M?c2B^M=b2B`M>_2BbM>]2BdM>Z2CgM=X2CiM=V2CkM=S2DnM=o1DRN<l1DVN<h1EYN<e1C]N=a1BbN>\\1AgN`0V1^OnNb0Q1\\OROd0l0[OWOf0=BF>6CM=OE3;IG99GD<<CC?<BB`0>VOIm07QOIQ16mNKU15hNK[14cNM_13^NNd11ZNOi11TN0n1OoM3S2MgM7[2HdM5a2K^M2f2MZM0j20TMMQ32oLJV36hLG]38cLG_39_LGc38\\LIe37YLIi36VLKk35SLLn33QLMQ43mKNT41kK0V40hK0Z40eK0\\4OdK1]4ObK0`4O`K2`4N_K2b4M_K2b4N^K1c4N^K1c4O\\K1e4N\\K2d4N\\K1e4NRKKjK6U9OjJ5mKLZ9OdJ<PLD]9OaJ0\\L6T9J]J0dL3P9LZJ1jL1m8NnG?4BUO0i8OgGn00SODNf82aGc0ROFf0J5Ic84\\G>EKG0_:GRF88=n9[OfE6h0Nm9LWE7g<IVC6n<JnB7U=IgB8\\=H`B8e=GWB9m=08I6M4N10Plk0"}, "label": "a multi - colored kite in the sky with woman who has her arms outstretched towards it"}]}
{"id": 271185, "image": "COCO_train2014_000000271185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lamb peeking out shyly\"."}], "task": "refering", "answer_template": "The \"lamb peeking out shyly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 174, 175, 176, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265], "bbox": [0.5393125000000001, 0.30752336448598133, 0.74353125, 0.7462383177570093], "iscrowd": 0, "area": 9462.01345, "rle": {"size": [428, 640], "counts": "V``42V=4L5L3O1TOEfD=W;GfD:Z:CkE6H9[:FhE4K7[:JeE2N6Z:k0dEVOZ:m0dETO\\:m0bESO^:n0aERO_:P1_EPOa:R1]EkNf:j13LTNeMPIW2U7mMaHT2d7QNTHn1P8WNiGf1\\8_N\\G_1i8fNPGV1V9oNaFa0S:CdEb0\\:V12N1O1O1O100O1O1OWOcMeF\\2\\9kM]FT2d9lM\\FU2c9lM\\FU2d9jM\\FX2b9hM^FY2a9gM_F[2_9eMaF\\2^9eMaF]2]9cMcF_2OVMo8;RGo2k8QMVGR3f8nLZGV3b8kL]GP4i7oKWHT4f7lKZHU4e7lKZHU4e7kK[HU4f7jK[HV4d7jK\\HV4d7jK\\HW4c7iK^HV4b7jK^HV4b7jK_HV4`7kK_HU4a7lK_HU3GVMi7G_Hl21[M`7JWHA0W2^1]NZ6KUHG1P2a1]NY6`2lI\\MT6c2QJYMo5f2XJTMg5l2_JPM^5R3]20000O2N1O2N2N2N2M3N2N2M3N2N2M4M2N2N3N1N3M2N3M2N2M4M2N3L3N2M3M3G810O10000O1000000O10000O010O1O100O1O1O100O001N2O1O1O1N2N2L3M^]T2"}, "label": "lamb peeking out shyly"}]}
{"id": 271185, "image": "COCO_train2014_000000271185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black and white lamb\"."}], "task": "refering", "answer_template": "The \"the black and white lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 174, 175, 176, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265], "bbox": [0.5393125000000001, 0.30752336448598133, 0.74353125, 0.7462383177570093], "iscrowd": 0, "area": 9462.01345, "rle": {"size": [428, 640], "counts": "V``42V=4L5L3O1TOEfD=W;GfD:Z:CkE6H9[:FhE4K7[:JeE2N6Z:k0dEVOZ:m0dETO\\:m0bESO^:n0aERO_:P1_EPOa:R1]EkNf:j13LTNeMPIW2U7mMaHT2d7QNTHn1P8WNiGf1\\8_N\\G_1i8fNPGV1V9oNaFa0S:CdEb0\\:V12N1O1O1O100O1O1OWOcMeF\\2\\9kM]FT2d9lM\\FU2c9lM\\FU2d9jM\\FX2b9hM^FY2a9gM_F[2_9eMaF\\2^9eMaF]2]9cMcF_2OVMo8;RGo2k8QMVGR3f8nLZGV3b8kL]GP4i7oKWHT4f7lKZHU4e7lKZHU4e7kK[HU4f7jK[HV4d7jK\\HV4d7jK\\HW4c7iK^HV4b7jK^HV4b7jK_HV4`7kK_HU4a7lK_HU3GVMi7G_Hl21[M`7JWHA0W2^1]NZ6KUHG1P2a1]NY6`2lI\\MT6c2QJYMo5f2XJTMg5l2_JPM^5R3]20000O2N1O2N2N2N2M3N2N2M3N2N2M4M2N2N3N1N3M2N3M2N2M4M2N3L3N2M3M3G810O10000O1000000O10000O010O1O100O1O1O100O001N2O1O1O1N2N2L3M^]T2"}, "label": "the black and white lamb"}]}
{"id": 271185, "image": "COCO_train2014_000000271185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooly sheep with its face in the green grass standing beside a two tone sheep\"."}], "task": "refering", "answer_template": "The \"wooly sheep with its face in the green grass standing beside a two tone sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 191, 192, 193, 194, 195, 196, 212, 214, 215, 216, 217, 218, 219, 235, 237, 238, 239, 240, 241, 242, 258, 261, 262, 263, 264, 265, 285, 286, 287, 308, 309], "bbox": [0.21140625000000002, 0.13764018691588784, 0.5767031250000001, 0.9311915887850466], "iscrowd": 0, "area": 48945.20245000002, "rle": {"size": [428, 640], "counts": "_ch12X;OaFa0U9D`Fl0W9WO^F[1W9jN^Ff1X9^N^FR2Y9RN\\F_2Y9S1F;E:GW1hN5K5L4K4L3N1N3M2OeNiIdKT6X4TJfKj5W4^JfK`5W4gJgKX5U4PKhKn4U4YKiKf4S4bKjK]4T4hKjKW4V4mKgKR4X4RLfKn3Y4VLdKi3\\4ZLbKe3^4`L^K`3W4nLfKQ3Z4SMcKl2_4o2101N1O101N101O0O101N101O0O2O000O2O001N10001N101N10001N101O001N10001O2N1O2N2N1O2N2M3N1O2N2N1O2YJnIP4T6[K`Jd4a5iJSKU5c6O1N2O1O1O0000000001O00000ZNjJVKW5i4RKnJn4R5[KeJe4[5dK\\J\\4d5hKXJY4g5iKWJW4i5lKTJT4o5lKnIT4S6nKjIS4W6mKgIS4Z6oKcIQ4^6^12O0O2N1O101N1O1O2O1N2N2N101N2PJRJ^4P6YKZJe4T72N2N3M2N3O02O1N2O1N2O1N2N2O2M2RLjGc2W8oLUHQ3S91O001O00001O001O001O001O0O2N1O1O2N1O2N1O2N2N1O2N2N2N2N1O2M3M3L4M3M2M4M3M3M3L3N3N2M3N2M2O2M3N2M3N2N1N3N2M3N2M2O2M3N2M3N2N3L3K6I6J7I6J7M2O2M2\\OnGlKT8P4oGnKS8o3oGQLS8k3PHSLR8j3QHULQ8g3QHXLQ8f2cGkM?^OP8d2QIZMR7c2oH]MR7`2QI^MR7_2PI`MQ7^2PIaMT7[2nHdMW7U2lHiMZ7Q2U2K6J6F9F;F:F:FfRa3"}, "label": "wooly sheep with its face in the green grass standing beside a two tone sheep"}]}
{"id": 271185, "image": "COCO_train2014_000000271185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult sheep with white wool\"."}], "task": "refering", "answer_template": "The \"an adult sheep with white wool\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 191, 192, 193, 194, 195, 196, 212, 214, 215, 216, 217, 218, 219, 235, 237, 238, 239, 240, 241, 242, 258, 261, 262, 263, 264, 265, 285, 286, 287, 308, 309], "bbox": [0.21140625000000002, 0.13764018691588784, 0.5767031250000001, 0.9311915887850466], "iscrowd": 0, "area": 48945.20245000002, "rle": {"size": [428, 640], "counts": "_ch12X;OaFa0U9D`Fl0W9WO^F[1W9jN^Ff1X9^N^FR2Y9RN\\F_2Y9S1F;E:GW1hN5K5L4K4L3N1N3M2OeNiIdKT6X4TJfKj5W4^JfK`5W4gJgKX5U4PKhKn4U4YKiKf4S4bKjK]4T4hKjKW4V4mKgKR4X4RLfKn3Y4VLdKi3\\4ZLbKe3^4`L^K`3W4nLfKQ3Z4SMcKl2_4o2101N1O101N101O0O101N101O0O2O000O2O001N10001N101N10001N101O001N10001O2N1O2N2N1O2N2M3N1O2N2N1O2YJnIP4T6[K`Jd4a5iJSKU5c6O1N2O1O1O0000000001O00000ZNjJVKW5i4RKnJn4R5[KeJe4[5dK\\J\\4d5hKXJY4g5iKWJW4i5lKTJT4o5lKnIT4S6nKjIS4W6mKgIS4Z6oKcIQ4^6^12O0O2N1O101N1O1O2O1N2N2N101N2PJRJ^4P6YKZJe4T72N2N3M2N3O02O1N2O1N2O1N2N2O2M2RLjGc2W8oLUHQ3S91O001O00001O001O001O001O0O2N1O1O2N1O2N1O2N2N1O2N2N2N2N1O2M3M3L4M3M2M4M3M3M3L3N3N2M3N2M2O2M3N2M3N2N1N3N2M3N2M2O2M3N2M3N2N3L3K6I6J7I6J7M2O2M2\\OnGlKT8P4oGnKS8o3oGQLS8k3PHSLR8j3QHULQ8g3QHXLQ8f2cGkM?^OP8d2QIZMR7c2oH]MR7`2QI^MR7_2PI`MQ7^2PIaMT7[2nHdMW7U2lHiMZ7Q2U2K6J6F9F;F:F:FfRa3"}, "label": "an adult sheep with white wool"}]}
{"id": 17236, "image": "COCO_train2014_000000017236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in grey shirt hugging another man\"."}], "task": "refering", "answer_template": "The \"man in grey shirt hugging another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 105, 106, 107, 129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340], "bbox": [0.3858125, 0.16180522565320665, 0.7997187499999999, 1.0], "iscrowd": 0, "area": 39836.34299999998, "rle": {"size": [421, 640], "counts": "UkU3:j<4L3M4L2N2O2M2N2N2O1O001O001O001O010OO1N2O1O1O1N2000000000000000000O1001O00000000001O000000001O0O2O0O2O001N101O0O2O001N101O0O2O001N101O0O2O0O10000O1000O010000O01000O10O10O100O10O10O100O01000O10O010000O10O0100N2O1O001O1O1O1O1N101O1O1O1nJmMQOT2l0RNPOo1m0WNoNi1n0^NmNd1P1aNmN`1P1fNlN[1Q1kNkNV1Q1QOkNP1T1SOiNm0dNiMI]1`1k0^NQN2U1^1k0XNXN9m0_1l0oM`Na0e0_1l0hMgNi0=_1l0cMlNm08`1l0`MPOo05`1l0\\MTOT10_1n0WMWOZ1J_1U1nLUOc1F^1Z1fLUOl1A\\1`1`LSOS2^OW1R4iNmKR1Y4nNgKl0_4TOaKf0e4ZOZKa0l4_OTKgNcN>_6k0nJdNfN;b6P1iJfNdN4i6V1bJgNdN1m6X1^JhNeNOn6Y1\\JiNfNLP7[1YJjNgNIR7]1WJjNgNGT7_1TJkNhNEU7`1RJlNiNBW7a1QJmNhN@Y7c1nInNiN]O[7e1lInNiN\\O\\7f1jIoNjNYO^7h1hIoNjNWO`7j1eIPOjNVOb7j1dIQOiNSOe7k1bISOiNQOf7l1aISOiNPOg7m1`ITOhNmNj7o1^ITOhNlNk7P2]ITOm6m0QIUOn6k0RIUOn6k0RIUOn6k0RIVOm6j0SIVOm6j0SIPNhNn0T8R1UImMiNQ1R8S1TIjMlNS1P8S1\\IbN^NBU8l1^IaN_NAS8n1^IbN`N^OR8P2^IbNaN^OP8P2_IbNaN^OP8P2`IbN`N]OP8Q2`IbN`N^Oo7Q2aIaN`N[OQ8T2_IaNaNYOQ8V2_I`Ni6`1WIaNi6_1WI`Ni6a1WI_Nh6b1XI]Nh6d1XI[Nh6f1XIXNi6i1WIVNi6k1WISNj6n1iHjMlN6[8Q2eHnMPON\\8U2`HQNUOG\\8Y2[HTNZO_O]8^2UHXN^OVO_8c2aGiNf8m2N1O2N1O1O2N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O2N3M3M3M3M3M3M3M3L3L8G>C<D=E;E;D<E:F7H6K6J5J6K5K5J6K6J5J6K5J6K8H:E;FXXd1"}, "label": "man in grey shirt hugging another man"}]}
{"id": 17236, "image": "COCO_train2014_000000017236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in long grey shirt and dark cap\"."}], "task": "refering", "answer_template": "The \"man in long grey shirt and dark cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 105, 106, 107, 129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340], "bbox": [0.3858125, 0.16180522565320665, 0.7997187499999999, 1.0], "iscrowd": 0, "area": 39836.34299999998, "rle": {"size": [421, 640], "counts": "UkU3:j<4L3M4L2N2O2M2N2N2O1O001O001O001O010OO1N2O1O1O1N2000000000000000000O1001O00000000001O000000001O0O2O0O2O001N101O0O2O001N101O0O2O001N101O0O2O0O10000O1000O010000O01000O10O10O100O10O10O100O01000O10O010000O10O0100N2O1O001O1O1O1O1N101O1O1O1nJmMQOT2l0RNPOo1m0WNoNi1n0^NmNd1P1aNmN`1P1fNlN[1Q1kNkNV1Q1QOkNP1T1SOiNm0dNiMI]1`1k0^NQN2U1^1k0XNXN9m0_1l0oM`Na0e0_1l0hMgNi0=_1l0cMlNm08`1l0`MPOo05`1l0\\MTOT10_1n0WMWOZ1J_1U1nLUOc1F^1Z1fLUOl1A\\1`1`LSOS2^OW1R4iNmKR1Y4nNgKl0_4TOaKf0e4ZOZKa0l4_OTKgNcN>_6k0nJdNfN;b6P1iJfNdN4i6V1bJgNdN1m6X1^JhNeNOn6Y1\\JiNfNLP7[1YJjNgNIR7]1WJjNgNGT7_1TJkNhNEU7`1RJlNiNBW7a1QJmNhN@Y7c1nInNiN]O[7e1lInNiN\\O\\7f1jIoNjNYO^7h1hIoNjNWO`7j1eIPOjNVOb7j1dIQOiNSOe7k1bISOiNQOf7l1aISOiNPOg7m1`ITOhNmNj7o1^ITOhNlNk7P2]ITOm6m0QIUOn6k0RIUOn6k0RIUOn6k0RIVOm6j0SIVOm6j0SIPNhNn0T8R1UImMiNQ1R8S1TIjMlNS1P8S1\\IbN^NBU8l1^IaN_NAS8n1^IbN`N^OR8P2^IbNaN^OP8P2_IbNaN^OP8P2`IbN`N]OP8Q2`IbN`N^Oo7Q2aIaN`N[OQ8T2_IaNaNYOQ8V2_I`Ni6`1WIaNi6_1WI`Ni6a1WI_Nh6b1XI]Nh6d1XI[Nh6f1XIXNi6i1WIVNi6k1WISNj6n1iHjMlN6[8Q2eHnMPON\\8U2`HQNUOG\\8Y2[HTNZO_O]8^2UHXN^OVO_8c2aGiNf8m2N1O2N1O1O2N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O2N3M3M3M3M3M3M3M3L3L8G>C<D=E;E;D<E:F7H6K6J5J6K5K5J6K6J5J6K5J6K8H:E;FXXd1"}, "label": "man in long grey shirt and dark cap"}]}
{"id": 17236, "image": "COCO_train2014_000000017236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone or something covered in black obstructed by three white paper plates\"."}], "task": "refering", "answer_template": "The \"someone or something covered in black obstructed by three white paper plates\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 185, 186, 187, 233, 234, 235, 256, 257, 258, 259, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.058734375000000005, 0.5277434679334917, 0.35534375, 1.0], "iscrowd": 0, "area": 14861.226700000001, "rle": {"size": [421, 640], "counts": "PQ`03P=2N2O100000000O100iJHaM8^2IbM7\\2KcM6\\2KdM5\\2KcM6\\2KdM5[2McM4\\2MdM3[2NeM2[2NdM3[2NeM2[2NdM3[2NeM2[2NdM3[2NeM2[2OcM2\\2OcM2\\2OcM2]2NcM2\\2OcM2]2NcM2\\2OcM2]2NbM3]2ObM1^2OaM2^20`M1`20_M0`21_M0a21^MOa2ERIM[4?c2@ZINQ4c0e2]OaIKj3h0d2]OWLJaNj0X5[OXLM]Ni0[5YOXL0\\Ng0[5YOZL2XNf0^5XOYL4XNd0_5WOZL7TNc0a5WOZL8SNb0c5VOZL:QN`0d5VO[L;PN`0e5UO[L;oMa0f5SO[L=oM`0f5SO[L<oMb0f5QO[L>oMa0f5QOZL?oMa0g5POYL?PNb0g5oNXL`0oMc0i5mNVLb0PNb0j5lNSLd0SNa0k5jNoKh0UN?l5iNhJ0[Oj01>m5gNgJ4UOj06<n5fNfJ9oNh0<:f7QOhGg0b09g7SO]Gj0j05j7?UHBk7=UHDl7;SHFn78RHIn77QHJP85oGLQ83nGOS81kG0V8OiG2W8OgG2Z8MeG4\\8KcG6]8KaG6`8I^G9c8F[G<e8EYG<g8HQG<o8m10000001O00000000000000001O000000000000O10000000000O100000000O1O1O1O1O1O1O11O3M2N3M2N3M3M3M3M3M3M3M3M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O001O001O001O001O1O1O1O1O1O1O1O1O001_NZEa0g:SOhEi0^;L4L3M4L4L4L_\\Y5"}, "label": "someone or something covered in black obstructed by three white paper plates"}]}
{"id": 17236, "image": "COCO_train2014_000000017236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a seated person in blue with paper plates taped on him\"."}], "task": "refering", "answer_template": "The \"a seated person in blue with paper plates taped on him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 185, 186, 187, 233, 234, 235, 256, 257, 258, 259, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.058734375000000005, 0.5277434679334917, 0.35534375, 1.0], "iscrowd": 0, "area": 14861.226700000001, "rle": {"size": [421, 640], "counts": "PQ`03P=2N2O100000000O100iJHaM8^2IbM7\\2KcM6\\2KdM5\\2KcM6\\2KdM5[2McM4\\2MdM3[2NeM2[2NdM3[2NeM2[2NdM3[2NeM2[2NdM3[2NeM2[2OcM2\\2OcM2\\2OcM2]2NcM2\\2OcM2]2NcM2\\2OcM2]2NbM3]2ObM1^2OaM2^20`M1`20_M0`21_M0a21^MOa2ERIM[4?c2@ZINQ4c0e2]OaIKj3h0d2]OWLJaNj0X5[OXLM]Ni0[5YOXL0\\Ng0[5YOZL2XNf0^5XOYL4XNd0_5WOZL7TNc0a5WOZL8SNb0c5VOZL:QN`0d5VO[L;PN`0e5UO[L;oMa0f5SO[L=oM`0f5SO[L<oMb0f5QO[L>oMa0f5QOZL?oMa0g5POYL?PNb0g5oNXL`0oMc0i5mNVLb0PNb0j5lNSLd0SNa0k5jNoKh0UN?l5iNhJ0[Oj01>m5gNgJ4UOj06<n5fNfJ9oNh0<:f7QOhGg0b09g7SO]Gj0j05j7?UHBk7=UHDl7;SHFn78RHIn77QHJP85oGLQ83nGOS81kG0V8OiG2W8OgG2Z8MeG4\\8KcG6]8KaG6`8I^G9c8F[G<e8EYG<g8HQG<o8m10000001O00000000000000001O000000000000O10000000000O100000000O1O1O1O1O1O1O11O3M2N3M2N3M3M3M3M3M3M3M3M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O001O001O001O001O1O1O1O1O1O1O1O1O001_NZEa0g:SOhEi0^;L4L3M4L4L4L_\\Y5"}, "label": "a seated person in blue with paper plates taped on him"}]}
{"id": 17236, "image": "COCO_train2014_000000017236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two people hugging\"."}], "task": "refering", "answer_template": "The \"two people hugging\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 155, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 262, 263, 285, 286, 287, 288, 307, 308, 309, 310, 311, 330, 331, 332, 333, 334], "bbox": [0.368734375, 0.2457482185273159, 0.774328125, 0.9878859857482185], "iscrowd": 0, "area": 27343.853099999993, "rle": {"size": [421, 640], "counts": "c]Q32o<5K4K5L4L4L4L4L4L4L4L5K4M3L4L4nMSNhHQ2l6[NPIi1`6gN^I[1Y6mNhIS1R6SOnIm0l5YOSJi0g5[OZJe0`5A_J`0^5CbJ=[5EfJ;X5GgJ:V5IjJ7S5LlJ5Q5MPK3n4ORK2j43SKNj46UKJg4<VKEg4?XKAe4c0ZK]Oc4h0ZKYOc4k0\\KUOb4o0[KROc4Q1\\KoNb4T1]KmN`4W1]KjNa4X1_KhN_4Z1aKfN]4\\1bKeN\\4]1dKcN[4_1cKbN\\4_1dKaN\\4_1dKaN[4`1dKaN\\4_1dKbNZ4_1fKaNY4`1fKaNZ4_1gK`NX4a1hK_NW4b1jK]NV4d1iK\\NV4e1kKZNT4g1lKYNT4g1mKYNQ4h1PLWNP4i1PLWNo3j1RLUNm3l1SLTNk3n1VLQNi3Q2VLoMh3S2YLlMf3U2ZLkM^3]2cLbMX3b2jL]MT3e2lL[MQ3g2QMXMm2j2SMVMj2l2XMSMf2o2ZMQMd2P3^MoLa2R3_MnL_2S3cMlL\\2U3eMjLZ2V3gMjLX2W3iMhLV2X3kMhLT2Y3mMfLR2Z3oMfLP2[3QNdLn1\\3SNdLm1\\3SNdLl1\\3UNdLk1\\3UNdLk1[3VNeLi1\\3VNeLj1Z3WNfLi1Z3WNfLh1[3XNeLh1Z3YNfLg1Z3YNfLf1Z3[NeLf1[3YNfLg1Y3ZNgLf1Y3ZNgLf1X3[NhLe1X3[NhLe1W3\\NiLd1W3\\NiLd1V3]NjLb1W3]NjLd1T3]N`KYN8Z3X4]N^K\\N9W3X4^N\\K`N:S3Y4]NZKeN;n2Z4^NYKhN;j2\\4^NVKmN<e2]4^NUKRO<a2]4^NTKUO>]2]4_NSKXO>Y2]4aNRK[O?U2G`Lg117P2iM_O`0P2HaLe148m1iMB`0l1IcLb16:j1hMFa0h1HdLc189d3ZN@IfL`1:;b3[N^OIhL^1;<b3[N[OJiL^1=:b3\\NYOJjL_1<9d3\\NVOKkL_1=8e3\\NSOLmL_1=5g3^NPOLnL`1<4h3_NA]1hL2j3_N_O]1jL1Q80PHNQ82QHKP85QHI]7hNoG^1f0H^7kNhG^1l0E^7nNcG]1P1C`7QO]G\\1T1Ab7TOVG[1[1^Ob7XOPGX1a1^Ob7d0bHZOa7a0dH]O^7>fH@]7;gHCZ7:kHCV7:nHDR7;QICP7=RIAn6?SI_On6a0TIZOn6g0h20O100O100O100O2N100O100O100O1O10000O10001N10000O100O10000OoNXDg0h;XOYDh0h;VOYDj0g;UOZDl0f;RO[Dn0e;QO\\Do0n;O00000ImC]OR<c0oC\\OR<c0PD[OP<f0PDYOP<g070001O01O0002N5K2N2L4M3L4L`l`02aS_O2N2M3N2N2N2N1N01O100O1O01ePk1"}, "label": "two people hugging"}]}
{"id": 17236, "image": "COCO_train2014_000000017236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy in white shirt\"."}], "task": "refering", "answer_template": "The \"guy in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 155, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 262, 263, 285, 286, 287, 288, 307, 308, 309, 310, 311, 330, 331, 332, 333, 334], "bbox": [0.368734375, 0.2457482185273159, 0.774328125, 0.9878859857482185], "iscrowd": 0, "area": 27343.853099999993, "rle": {"size": [421, 640], "counts": "c]Q32o<5K4K5L4L4L4L4L4L4L4L5K4M3L4L4nMSNhHQ2l6[NPIi1`6gN^I[1Y6mNhIS1R6SOnIm0l5YOSJi0g5[OZJe0`5A_J`0^5CbJ=[5EfJ;X5GgJ:V5IjJ7S5LlJ5Q5MPK3n4ORK2j43SKNj46UKJg4<VKEg4?XKAe4c0ZK]Oc4h0ZKYOc4k0\\KUOb4o0[KROc4Q1\\KoNb4T1]KmN`4W1]KjNa4X1_KhN_4Z1aKfN]4\\1bKeN\\4]1dKcN[4_1cKbN\\4_1dKaN\\4_1dKaN[4`1dKaN\\4_1dKbNZ4_1fKaNY4`1fKaNZ4_1gK`NX4a1hK_NW4b1jK]NV4d1iK\\NV4e1kKZNT4g1lKYNT4g1mKYNQ4h1PLWNP4i1PLWNo3j1RLUNm3l1SLTNk3n1VLQNi3Q2VLoMh3S2YLlMf3U2ZLkM^3]2cLbMX3b2jL]MT3e2lL[MQ3g2QMXMm2j2SMVMj2l2XMSMf2o2ZMQMd2P3^MoLa2R3_MnL_2S3cMlL\\2U3eMjLZ2V3gMjLX2W3iMhLV2X3kMhLT2Y3mMfLR2Z3oMfLP2[3QNdLn1\\3SNdLm1\\3SNdLl1\\3UNdLk1\\3UNdLk1[3VNeLi1\\3VNeLj1Z3WNfLi1Z3WNfLh1[3XNeLh1Z3YNfLg1Z3YNfLf1Z3[NeLf1[3YNfLg1Y3ZNgLf1Y3ZNgLf1X3[NhLe1X3[NhLe1W3\\NiLd1W3\\NiLd1V3]NjLb1W3]NjLd1T3]N`KYN8Z3X4]N^K\\N9W3X4^N\\K`N:S3Y4]NZKeN;n2Z4^NYKhN;j2\\4^NVKmN<e2]4^NUKRO<a2]4^NTKUO>]2]4_NSKXO>Y2]4aNRK[O?U2G`Lg117P2iM_O`0P2HaLe148m1iMB`0l1IcLb16:j1hMFa0h1HdLc189d3ZN@IfL`1:;b3[N^OIhL^1;<b3[N[OJiL^1=:b3\\NYOJjL_1<9d3\\NVOKkL_1=8e3\\NSOLmL_1=5g3^NPOLnL`1<4h3_NA]1hL2j3_N_O]1jL1Q80PHNQ82QHKP85QHI]7hNoG^1f0H^7kNhG^1l0E^7nNcG]1P1C`7QO]G\\1T1Ab7TOVG[1[1^Ob7XOPGX1a1^Ob7d0bHZOa7a0dH]O^7>fH@]7;gHCZ7:kHCV7:nHDR7;QICP7=RIAn6?SI_On6a0TIZOn6g0h20O100O100O100O2N100O100O100O1O10000O10001N10000O100O10000OoNXDg0h;XOYDh0h;VOYDj0g;UOZDl0f;RO[Dn0e;QO\\Do0n;O00000ImC]OR<c0oC\\OR<c0PD[OP<f0PDYOP<g070001O01O0002N5K2N2L4M3L4L`l`02aS_O2N2M3N2N2N2N1N01O100O1O01ePk1"}, "label": "guy in white shirt"}]}
{"id": 394065, "image": "COCO_train2014_000000394065.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big black and yellow truck\"."}], "task": "refering", "answer_template": "The \"the big black and yellow truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 348, 349, 350, 351], "bbox": [0.0, 0.3141875, 0.9763437500000001, 0.9290625], "iscrowd": 0, "area": 102009.0179, "rle": {"size": [480, 640], "counts": "\\;k0_<f1kNV1N101O0O2O001N101N101O0O101N101DmKiES4U:nKkET4Q:oKPFP4m9SLTFm3h9VLYFi3d9ZL\\Fg3`9\\L`Fd3_9]LaFd3]9\\LdFd3[9]LeFb3[9_LeFa3Z9`LfF`3Z9`LeFa3Z9`LfF`3Y9aLgF^3Y9cLgF]3X9dLhF\\3W9eLiF[3W9eLiF[3V9fLjFY3V9hLjFX3U9iLkFW3T9jLlFV3T9jLlFU3T9lLlFT3S9mLmFS3R9nLnFR3R9nLnFR3R9nLnFQ3T9nLlFR3T9nLlFQ3U9oLkFQ3U9oLkFP3V9PMjFP3V9PMjFo2W9QMiFn2X9RMhFn2X9RMhFm2Y9SMgFm2Z9RMfFm2[9SMeFm2[9SMeFl2\\9TMdFl2\\9TMdFk2]9UMcFk2]9UMcFj2^9VMbFi2_9WMaFi2_9WMaFh2`9XM`Fh2`9XM`Fg2a9YM_Fg2a9YM_Fh2`9WMaFP3X9PMhFW3Q9iLoFX3P9hLPGZ3n8fLRG[3m8eLSG]3k8cLUG^3i8cLWG_3g8aLYG`3f8`LZGb3d8^L\\Gc3c8]L]Ge3a8[L_Gf3`8ZL`Gh3^8XLbGi3]8WLcGk3[8ULeGl3Z8TLfGn3X8RLhGo3W8QLiGQ4U8oKkGR4T8nKlGS4R8nKnGT4P8lKPHU4o7kKQHW4m7iKSHX4l7hKTHZ4j7fKVH[4i7eKWH\\4h7dKXH\\4h7dKXH]4g7cKYH]4g7cKYH]4g7cKYH^4f7bKZH^4f7bKZH^4f7bKZH_4d7bK\\H^4d7bK\\H_4c7aKfHV4Z7iKhIV3X6bLQK^2n4ZM\\Ld1d3TNeMl0Z2mNnN2`n1<UYNP1POo0POP1_MdLhF0K\\4]8_LcG_4_7aKaH^4`7bK`H^4`7bK`H^4`7bK`H]4a7cK_H]4a7cK_H]4a7cK_H]4a7cK_H\\4b7dK^H\\4b7dK^H\\4b7dK^H\\4b7dK^H[4c7eK]H[4c7eK]H[4c7eK]H[4c7eK]HZ4d7fK\\HZ4d7fK\\HZ4d7fK\\HZ4d7fK\\HY4e7hKZHX4f7hKZHX4f7hKZHW4g7iKYHW4h7hKXHX4h7hKXHX4h7hKXHW4i7iKWHW4i7iKWHW4i7iKWHW4i7iKWHV4j7jKVHV4j7jKVHV4j7jKVHV4j7jKVHU4k7kKUHU4k7kKUHU4k7kKUHU4k7kKUHT4l7lKTHT4l7lKTHT4l7lKTHS4m7mKSHS4m7mKSHS4m7mKSHS4m7mKSHR4n7nKRHQ4o7oKQHo3Q8QLoGn3R8RLnGm3S8SLmGl3T8TLlGk3R8XLnGg3P8\\LPHc3o7_LQH`3m7cLSH[3l7hLTHW3j7lLVHS3i7oLWHP3g7SMYHl2e7WM[Hh2d7ZM\\He2b7^M^Hb2_7aMaH^2^7dMbH[2\\7hMdHX2Y7kMgHT2W7oMiHQ2U7QNkHn1U7SNkHl1U7UNkHk1U7UNkHj1e6kKVI\\25h1f6nKRI[28g1f6QLoHX2;f1g6TLkHW2>d1g6XLiHT2`0d1g6kNYIT1h6lNXIS1i6mNWIS1i6mNWIR1j6nNVIQ1j6POVIP1j6POVIo0k6QOUIo0k6QOUIn0l6ROTIm0m6SOSIm0m6SOSIl0m6UOSIk0m6UOSIk0m6UOSIj0n6VORIj0n6VORIi0o6XOPIh0o6YOQIg0o6YOQIf0P7ZOPIf0P7ZOPIf0P7ZOPIe0Q7[OoHe0P7\\OPIc0Q7]OoHc0Q7]OoHc0Q7]OoHb0R7^OnHb0R7^OnHb0R7^OnHa0S7_OmHa0S7_OmH`0T7@lH`0T7@lH`0U7_OkHa0W7]OiHc0Y7[OgHe0[7YOeHg0]7WOcHi0_7UOaHj0_7WOaHi0_7WOaHi0_7WOaHi0_7WOaHi0_7WOaHi0^7XObHh0^7XObHh0^7XObHg0_7YOaHg0_7YOaHg0^7ZObHf0^7ZObHf0^7ZObHf0^7ZObHf0^7ZObHf0^7ZObHe0^7\\ObHd0^7\\ObHd0^7\\ObHd0^7\\ObHd0^7\\ObHd0]7]OcHc0]7]OcHb0^7^ObHb0^7^ObHb0^7^ObHb0^7^ObHb0]7_OcHa0]7_OcHa0]7_OcHa0]7_OcH`0^7@bH`0]7AcH?]7AcH?]7AcH?]7AcH?]7AcH?]7AcH?\\7BdH=]7CmH3S7MoIQOQ6m0TKPNl4Z1h4ZOf0ZOf0ZOhl1JTTN2N:SG2T2j0`Ke0a48VIT1j6E]H;c7E]H;c7E\\H<d7D\\H;e7E[H;e7E[H;e7E[H;e7E[H;e7E[H;e7E[H:f7FZH:f7FZH:f7FZH:f7FZH:f7FZH9g7GYH9g7GYH9g7GYH9g7GYH9g7GYH9g7GYH8h7HXH8h7HXH8h7HXH8h7HXH8h7HXH7i7IWH7i7IWH7i7IWH7i7IWH7i7IWH7i7IWH6j7JVH6j7JVH6j7JVH6i7KVH6j7JVH5k7KUH5k7KUH5k7KUH5k7KUH5k7KUH5k7KUH4l7LTH4l7LTH4l7LTH4l7LTH4l7LTH3m7MSH3m7MSH3m7MSH3m7MSH3m7MSH3m7MSH2n7NRH2n7NRH2n7NRH2n7NRH2n7NRH2o7MQH2P8NPH2Q8MoG3Q8MoG3R8WM`G[2>>R8VMaG\\2==T8UM`G^2<=T8TMaG_2;=U8RMaGa2:=U8QMbGb29=]8CbG>^8BbG=_8CaG=_8CaG=_8CaG=_8CaG=_8CaG<`8D`G<`8D`G<`8D`G<`8D`G<_8EaG;_8EaG:`8F`G:`8F`G:`8F`G:`8F`G:`8F`G9a8G_G9a8G_G9a8G_G9a8G_G9a8G_G9Y8OgG0Z80fG0Z80fG0Z80fG0Z80fG0Z80fGOZ82fGNZ82fGNZ82fGNZ82fGNZ82eGOZ82fGM[83eGM[83eGM[82fGNZ82fGNZ82fGMZ84fGLZ84fGLZ84fGLZ84fGLZ84fGLZ84fGKZ86fGJZ86fGJZ86fGJZ86fGJZ86fGI[87eGIZ88fGHZ88fGHZ88fGHZ88fGH[87eGH\\88dGH\\88dGH]87cGI]87cGI]87cGH_8aMSG^2>1_8]MWGb2:1g8OYG1g8OYG1g8OXG2g8OYG0h80XG0h80XG0h80jH^NV7b1T3J6E;E;D<E;E;DnFEW40gK=R9<dM_OlEn0k9[OhER1n9XOfET1Q:UObEW1U:SOeES1R:VOnEj0h9@XF`0S9`MmFY207S95mFKS95mFKS95nFJR96nFIS97mFIP9:PGFm8=SGCl8>TGBl8>TGBl8>TGBm8=SGCm8=SGCm8=SGBn8>RGBn8>RGBn8>RGBn8>RGBn8>RGBn8>RGBo8=QGCo8=QGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGAQ9?oFAQ9?nFBR9>nFBR9>nFBR9>nFBS9=mFCT9<lFDV9:jFFZ96fFI]95cFK]95cFK]95cFK\\96dFJ\\96dFJ\\96dFJ\\96dFJ[97eFH\\98dFH\\98dFH\\98dFH[99eFG[99eFG[99eFG[99eFGZ9:fFE[9;eFE[9;eFE[9;eFEZ9<fFDZ9<fFDZ9<fFDZ9<fFD[9;eFHX98hFMT92lF4n8LRG9j8FVG=h8BXG>h8BXG>i8AWG?i8AWG?j8@VG`0j8@VG`0k8_OUGa0l8^OTGb0l8^OTGb0m8]OSGc0m8]OSGc0n8\\ORGd0n8\\ORGd0o8ZORGf0n8ZORGf0n8ZORGf0o8YOQGe0Q9[OoFc0T9\\OlFc0U9]OkFa0X9^OhF`0Z9@fF?\\9@cFa0]9_OcF`0_9_OaFa0_9_OaF`0`9@`F`0a9_O_Fa0a9_O_F`0c9_O]Fa0c9_O]F`0e9_O[Fh0^9XObFi0^9VObFj0^9VObFk0^9TObFl0^9TObFl0_9SOaFn0^9RObFn0^9RObFo0^9PObFP1^9PObFQ1^9nNbFR1^9nNbFP1`9PO`Fn0c9QO]Fm0e9SO[Fk0h9TOXF=W:CiEHl:8TESOb;h0bDcNS<T1Q1G:F9G9GgQ17hmNc0]Ob0_Ob0]Oc0^Oa0H91O2M2O1O2M2O1O2M2O1O2M2O1O1N2O2M2O1O1N2Of0ZO4K4M2N3L3N3M2M4M2N3L4M2N5J7J6J5Jeg6"}, "label": "the big black and yellow truck"}]}
{"id": 394065, "image": "COCO_train2014_000000394065.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and yellow signal truck with black tires , and red rims\"."}], "task": "refering", "answer_template": "The \"a black and yellow signal truck with black tires , and red rims\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 348, 349, 350, 351], "bbox": [0.0, 0.3141875, 0.9763437500000001, 0.9290625], "iscrowd": 0, "area": 102009.0179, "rle": {"size": [480, 640], "counts": "\\;k0_<f1kNV1N101O0O2O001N101N101O0O101N101DmKiES4U:nKkET4Q:oKPFP4m9SLTFm3h9VLYFi3d9ZL\\Fg3`9\\L`Fd3_9]LaFd3]9\\LdFd3[9]LeFb3[9_LeFa3Z9`LfF`3Z9`LeFa3Z9`LfF`3Y9aLgF^3Y9cLgF]3X9dLhF\\3W9eLiF[3W9eLiF[3V9fLjFY3V9hLjFX3U9iLkFW3T9jLlFV3T9jLlFU3T9lLlFT3S9mLmFS3R9nLnFR3R9nLnFR3R9nLnFQ3T9nLlFR3T9nLlFQ3U9oLkFQ3U9oLkFP3V9PMjFP3V9PMjFo2W9QMiFn2X9RMhFn2X9RMhFm2Y9SMgFm2Z9RMfFm2[9SMeFm2[9SMeFl2\\9TMdFl2\\9TMdFk2]9UMcFk2]9UMcFj2^9VMbFi2_9WMaFi2_9WMaFh2`9XM`Fh2`9XM`Fg2a9YM_Fg2a9YM_Fh2`9WMaFP3X9PMhFW3Q9iLoFX3P9hLPGZ3n8fLRG[3m8eLSG]3k8cLUG^3i8cLWG_3g8aLYG`3f8`LZGb3d8^L\\Gc3c8]L]Ge3a8[L_Gf3`8ZL`Gh3^8XLbGi3]8WLcGk3[8ULeGl3Z8TLfGn3X8RLhGo3W8QLiGQ4U8oKkGR4T8nKlGS4R8nKnGT4P8lKPHU4o7kKQHW4m7iKSHX4l7hKTHZ4j7fKVH[4i7eKWH\\4h7dKXH\\4h7dKXH]4g7cKYH]4g7cKYH]4g7cKYH^4f7bKZH^4f7bKZH^4f7bKZH_4d7bK\\H^4d7bK\\H_4c7aKfHV4Z7iKhIV3X6bLQK^2n4ZM\\Ld1d3TNeMl0Z2mNnN2`n1<UYNP1POo0POP1_MdLhF0K\\4]8_LcG_4_7aKaH^4`7bK`H^4`7bK`H^4`7bK`H]4a7cK_H]4a7cK_H]4a7cK_H]4a7cK_H\\4b7dK^H\\4b7dK^H\\4b7dK^H\\4b7dK^H[4c7eK]H[4c7eK]H[4c7eK]H[4c7eK]HZ4d7fK\\HZ4d7fK\\HZ4d7fK\\HZ4d7fK\\HY4e7hKZHX4f7hKZHX4f7hKZHW4g7iKYHW4h7hKXHX4h7hKXHX4h7hKXHW4i7iKWHW4i7iKWHW4i7iKWHW4i7iKWHV4j7jKVHV4j7jKVHV4j7jKVHV4j7jKVHU4k7kKUHU4k7kKUHU4k7kKUHU4k7kKUHT4l7lKTHT4l7lKTHT4l7lKTHS4m7mKSHS4m7mKSHS4m7mKSHS4m7mKSHR4n7nKRHQ4o7oKQHo3Q8QLoGn3R8RLnGm3S8SLmGl3T8TLlGk3R8XLnGg3P8\\LPHc3o7_LQH`3m7cLSH[3l7hLTHW3j7lLVHS3i7oLWHP3g7SMYHl2e7WM[Hh2d7ZM\\He2b7^M^Hb2_7aMaH^2^7dMbH[2\\7hMdHX2Y7kMgHT2W7oMiHQ2U7QNkHn1U7SNkHl1U7UNkHk1U7UNkHj1e6kKVI\\25h1f6nKRI[28g1f6QLoHX2;f1g6TLkHW2>d1g6XLiHT2`0d1g6kNYIT1h6lNXIS1i6mNWIS1i6mNWIR1j6nNVIQ1j6POVIP1j6POVIo0k6QOUIo0k6QOUIn0l6ROTIm0m6SOSIm0m6SOSIl0m6UOSIk0m6UOSIk0m6UOSIj0n6VORIj0n6VORIi0o6XOPIh0o6YOQIg0o6YOQIf0P7ZOPIf0P7ZOPIf0P7ZOPIe0Q7[OoHe0P7\\OPIc0Q7]OoHc0Q7]OoHc0Q7]OoHb0R7^OnHb0R7^OnHb0R7^OnHa0S7_OmHa0S7_OmH`0T7@lH`0T7@lH`0U7_OkHa0W7]OiHc0Y7[OgHe0[7YOeHg0]7WOcHi0_7UOaHj0_7WOaHi0_7WOaHi0_7WOaHi0_7WOaHi0_7WOaHi0^7XObHh0^7XObHh0^7XObHg0_7YOaHg0_7YOaHg0^7ZObHf0^7ZObHf0^7ZObHf0^7ZObHf0^7ZObHf0^7ZObHe0^7\\ObHd0^7\\ObHd0^7\\ObHd0^7\\ObHd0^7\\ObHd0]7]OcHc0]7]OcHb0^7^ObHb0^7^ObHb0^7^ObHb0^7^ObHb0]7_OcHa0]7_OcHa0]7_OcHa0]7_OcH`0^7@bH`0]7AcH?]7AcH?]7AcH?]7AcH?]7AcH?]7AcH?\\7BdH=]7CmH3S7MoIQOQ6m0TKPNl4Z1h4ZOf0ZOf0ZOhl1JTTN2N:SG2T2j0`Ke0a48VIT1j6E]H;c7E]H;c7E\\H<d7D\\H;e7E[H;e7E[H;e7E[H;e7E[H;e7E[H;e7E[H:f7FZH:f7FZH:f7FZH:f7FZH:f7FZH9g7GYH9g7GYH9g7GYH9g7GYH9g7GYH9g7GYH8h7HXH8h7HXH8h7HXH8h7HXH8h7HXH7i7IWH7i7IWH7i7IWH7i7IWH7i7IWH7i7IWH6j7JVH6j7JVH6j7JVH6i7KVH6j7JVH5k7KUH5k7KUH5k7KUH5k7KUH5k7KUH5k7KUH4l7LTH4l7LTH4l7LTH4l7LTH4l7LTH3m7MSH3m7MSH3m7MSH3m7MSH3m7MSH3m7MSH2n7NRH2n7NRH2n7NRH2n7NRH2n7NRH2o7MQH2P8NPH2Q8MoG3Q8MoG3R8WM`G[2>>R8VMaG\\2==T8UM`G^2<=T8TMaG_2;=U8RMaGa2:=U8QMbGb29=]8CbG>^8BbG=_8CaG=_8CaG=_8CaG=_8CaG=_8CaG<`8D`G<`8D`G<`8D`G<`8D`G<_8EaG;_8EaG:`8F`G:`8F`G:`8F`G:`8F`G:`8F`G9a8G_G9a8G_G9a8G_G9a8G_G9a8G_G9Y8OgG0Z80fG0Z80fG0Z80fG0Z80fG0Z80fGOZ82fGNZ82fGNZ82fGNZ82fGNZ82eGOZ82fGM[83eGM[83eGM[82fGNZ82fGNZ82fGMZ84fGLZ84fGLZ84fGLZ84fGLZ84fGLZ84fGKZ86fGJZ86fGJZ86fGJZ86fGJZ86fGI[87eGIZ88fGHZ88fGHZ88fGHZ88fGH[87eGH\\88dGH\\88dGH]87cGI]87cGI]87cGH_8aMSG^2>1_8]MWGb2:1g8OYG1g8OYG1g8OXG2g8OYG0h80XG0h80XG0h80jH^NV7b1T3J6E;E;D<E;E;DnFEW40gK=R9<dM_OlEn0k9[OhER1n9XOfET1Q:UObEW1U:SOeES1R:VOnEj0h9@XF`0S9`MmFY207S95mFKS95mFKS95nFJR96nFIS97mFIP9:PGFm8=SGCl8>TGBl8>TGBl8>TGBm8=SGCm8=SGCm8=SGBn8>RGBn8>RGBn8>RGBn8>RGBn8>RGBn8>RGBo8=QGCo8=QGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGBP9>PGAQ9?oFAQ9?nFBR9>nFBR9>nFBR9>nFBS9=mFCT9<lFDV9:jFFZ96fFI]95cFK]95cFK]95cFK\\96dFJ\\96dFJ\\96dFJ\\96dFJ[97eFH\\98dFH\\98dFH\\98dFH[99eFG[99eFG[99eFG[99eFGZ9:fFE[9;eFE[9;eFE[9;eFEZ9<fFDZ9<fFDZ9<fFDZ9<fFD[9;eFHX98hFMT92lF4n8LRG9j8FVG=h8BXG>h8BXG>i8AWG?i8AWG?j8@VG`0j8@VG`0k8_OUGa0l8^OTGb0l8^OTGb0m8]OSGc0m8]OSGc0n8\\ORGd0n8\\ORGd0o8ZORGf0n8ZORGf0n8ZORGf0o8YOQGe0Q9[OoFc0T9\\OlFc0U9]OkFa0X9^OhF`0Z9@fF?\\9@cFa0]9_OcF`0_9_OaFa0_9_OaF`0`9@`F`0a9_O_Fa0a9_O_F`0c9_O]Fa0c9_O]F`0e9_O[Fh0^9XObFi0^9VObFj0^9VObFk0^9TObFl0^9TObFl0_9SOaFn0^9RObFn0^9RObFo0^9PObFP1^9PObFQ1^9nNbFR1^9nNbFP1`9PO`Fn0c9QO]Fm0e9SO[Fk0h9TOXF=W:CiEHl:8TESOb;h0bDcNS<T1Q1G:F9G9GgQ17hmNc0]Ob0_Ob0]Oc0^Oa0H91O2M2O1O2M2O1O2M2O1O2M2O1O1N2O2M2O1O1N2Of0ZO4K4M2N3L3N3M2M4M2N3L4M2N5J7J6J5Jeg6"}, "label": "a black and yellow signal truck with black tires , and red rims"}]}
{"id": 500561, "image": "COCO_train2014_000000500561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand taking apart a hotdog with tomato and pickle\"."}], "task": "refering", "answer_template": "The \"a hand taking apart a hotdog with tomato and pickle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 150, 161, 162, 163, 164, 167, 168, 169, 185, 186, 202, 203, 219, 220, 235, 236, 237, 251, 252, 253, 254, 268, 269, 270, 271, 287, 288, 304, 305], "bbox": [0.5055416666666667, 0.289515625, 1.0, 0.766625], "iscrowd": 0, "area": 19344.1931, "rle": {"size": [640, 480], "counts": "cSh4?_c02M3N2N1O2N2M3N101O1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O2N100O100O2N1O1O2N1O1O2N1O1O101O00001O0000001O00001O00001O0000001O001O001O001O1O0010O010O010O10O01N1O2M2O2M2G:M2M4M2O2O0O2N1O2N101N1O2N1O2N1O2N1O2N2O0O2N1O2N1O2N1O2N101N101N101N101N101NbBOo61oH2Q7LnH6R7ImH:S7EkH=V7_OjHd0W=2N2N2N2O1N2N3M2N2O1N2N2N2N200O1O1O1O1O1O100O1O1O1O100O1lKVNRFj1m9^NkEc1U:^NiEc1W:^NhEb1X:^NgEc1Y:^NfEb1[:^NcEc1^:]N`Ed1`:]N_Ec1b:]N\\Ed1d:\\N\\Ed1e:\\NYEe1g:\\NWEe1j:[NUEe1k:\\NSEe1n:[NQEe1P;[NnDf1R;[NlDf1U;ZNjDf1V;ZNiDg1X;YNgDg1Y;ZNeDg1\\;YNbDh1^;YN`Dh1a;XN]Di1d;WNZDj1f;WNXDj1i;VNUDk1k;VNSDk1n;VNoCl1P<UNnCn1Q<SNlCP2R<RNkCQ2T<oMjCT2]N_Ml<>dDR3Z;oLcDU3\\;lLaDW3];kL_DY3`;gL[D_3c;cLWDc3h;^LRDh3l;ZLnCk3R<VLhCn3Y<SLaCQ4_<WLUCm3l<c1010O001O001O10O01O001O0010O01O1O001O010O001O001O01d1eK]H"}, "label": "a hand taking apart a hotdog with tomato and pickle"}]}
{"id": 500561, "image": "COCO_train2014_000000500561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver watch\"."}], "task": "refering", "answer_template": "The \"the silver watch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 150, 151, 152, 169], "bbox": [0.19518749999999999, 0.00225, 1.0, 0.42117187500000003], "iscrowd": 0, "area": 45855.87475000002, "rle": {"size": [640, 480], "counts": "Qhj13mc03M2N3M3M2N3M3M3M2N3M2N1O001O001O00001O001O001O001O0Q^OPOn`0Q1j^OWOUa0i0d^O_O[a0b0]^OFba0]1O1O2N1O1O1O1O1O1O1O000000000000001O0000000000001O00000000001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O0000000000000000000000001O000000000000000000000000001O1O001O001O11N101O0O101O0O101N10001N101O0O1000O0010O0100O010O10O01O10O010O0100O010O1O010O10O010O0100O010O1O010O010O10O0100O00100O010O010O10O01O10O01N2O0O2O0O2N2O0O2O1O001000O100O1000O010000O10000O010O10000O1000O0100O10000O10O10O10000O100O01000O1000O010O10O010O01O10O0100O010O10O01O010O10O0100O00100O010O010O10O01O10O010O0100O010O1O010O010O010O10O01O010O010O10O01O010O010O010O10O01O010O010O10O010O01O010O010O10O01O010O010O10O010O01O010O010O10O01P3oLYI"}, "label": "the silver watch"}]}
{"id": 500561, "image": "COCO_train2014_000000500561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one hand and arm wearing a watch , and a portion of another hand holding food\"."}], "task": "refering", "answer_template": "The \"one hand and arm wearing a watch , and a portion of another hand holding food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 150, 151, 152, 169], "bbox": [0.19518749999999999, 0.00225, 1.0, 0.42117187500000003], "iscrowd": 0, "area": 45855.87475000002, "rle": {"size": [640, 480], "counts": "Qhj13mc03M2N3M3M2N3M3M3M2N3M2N1O001O001O00001O001O001O001O0Q^OPOn`0Q1j^OWOUa0i0d^O_O[a0b0]^OFba0]1O1O2N1O1O1O1O1O1O1O000000000000001O0000000000001O00000000001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O0000000000000000000000001O000000000000000000000000001O1O001O001O11N101O0O101O0O101N10001N101O0O1000O0010O0100O010O10O01O10O010O0100O010O1O010O10O010O0100O010O1O010O010O10O0100O00100O010O010O10O01O10O01N2O0O2O0O2N2O0O2O1O001000O100O1000O010000O10000O010O10000O1000O0100O10000O10O10O10000O100O01000O1000O010O10O010O01O10O0100O010O10O01O010O10O0100O00100O010O010O10O01O10O010O0100O010O1O010O010O010O10O01O010O010O10O01O010O010O010O10O01O010O010O10O010O01O010O010O10O01O010O010O10O010O01O010O010O10O01P3oLYI"}, "label": "one hand and arm wearing a watch , and a portion of another hand holding food"}]}
{"id": 287575, "image": "COCO_train2014_000000287575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small green vase on the left with a flower in it\"."}], "task": "refering", "answer_template": "The \"small green vase on the left with a flower in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 128, 129, 130, 131, 132, 146, 147, 148, 149, 164, 165, 166, 183, 184], "bbox": [0.1146, 0.50336, 0.37416, 0.8202133333333332], "iscrowd": 0, "area": 7786.539149999998, "rle": {"size": [375, 500], "counts": "QTe01b;9_O`0I5M2N2N1O2N2N2N1O2N101N1O2N1O2O1N1O1O101N2N2N2N2O1N2N1O2O1N101O1N101N2O0UGaM^8a2^GaMa8a2[GcMd8^2YGdMf8g20000O100O100O1O1N2O1O1O1O1O1O1M3L4M300000000000000000VObH[M^7a2kH[MT7c2PI[MP7c2S1N2M3M3L4M3O2N10000O10000O100O10001N4M2N2M4L4L2N2N2N1O2N2N2N1O2N2N1O2N1O2N1O101O0O101N101N100O2O0O2O0O2O1O1O0O3N2Nm\\b3"}, "label": "small green vase on the left with a flower in it"}]}
{"id": 287575, "image": "COCO_train2014_000000287575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase that does not have a flower in it\"."}], "task": "refering", "answer_template": "The \"the vase that does not have a flower in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 208], "bbox": [0.24956, 0.4956, 0.6254, 0.9001066666666666], "iscrowd": 0, "area": 18857.597499999996, "rle": {"size": [375, 500], "counts": "aQ^1e0k:;G8K4M4L2M3N2M2O2N2N2N2O0O2O0O2N100O2O0O101N1O101N100O101N100O2O0O101N100O2O0O10001N100O2O0O101N1O1O101N1O1O2N100N2O1O1N2O2M2O1O1N2O1N2N3N1N2N2O1N2O1N2N3H7O100O100O1O2O0000000000010O000000000001O01OO100000001O000000002N1O1N2O2N1O3M2N2N3M2N3M1O1O1N2O1O1O1O1O1O1O1O1O1O1O0O2O001O001O001O001O001O0O2O0O1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N2N2N2N2N2N2N3M2N2M3L4L4L4L4L4L4L>BQWT2"}, "label": "the vase that does not have a flower in it"}]}
{"id": 287575, "image": "COCO_train2014_000000287575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a squat green vase with no flowers in it\"."}], "task": "refering", "answer_template": "The \"a squat green vase with no flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 208], "bbox": [0.24956, 0.4956, 0.6254, 0.9001066666666666], "iscrowd": 0, "area": 18857.597499999996, "rle": {"size": [375, 500], "counts": "aQ^1e0k:;G8K4M4L2M3N2M2O2N2N2N2O0O2O0O2N100O2O0O101N1O101N100O101N100O2O0O101N100O2O0O10001N100O2O0O101N1O1O101N1O1O2N100N2O1O1N2O2M2O1O1N2O1N2N3N1N2N2O1N2O1N2N3H7O100O100O1O2O0000000000010O000000000001O01OO100000001O000000002N1O1N2O2N1O3M2N2N3M2N3M1O1O1N2O1O1O1O1O1O1O1O1O1O1O0O2O001O001O001O001O001O0O2O0O1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N2N2N2N2N2N2N3M2N2M3L4L4L4L4L4L4L>BQWT2"}, "label": "a squat green vase with no flowers in it"}]}
{"id": 287575, "image": "COCO_train2014_000000287575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green vase to the right of two other vaces\"."}], "task": "refering", "answer_template": "The \"a green vase to the right of two other vaces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213], "bbox": [0.63446, 0.55648, 0.9833200000000001, 0.8580266666666666], "iscrowd": 0, "area": 14129.342200000005, "rle": {"size": [375, 500], "counts": "S[d32c;7H7J7H7J5J4M3N2N2N2N2N2N2N2N2O1N2N2O1O1O1N2O1O001O0O2O0O2O0O2N101N101N101N101N101N101N101N101N2O1N101N2O0O100O100O10000O100O1O1N2O1O1N2O1001O001O001O1O001O1O0000001O0000000000001O0000000000001O000000000000O2O00000001O001O1O1O001O1O00001O000O2M200O2O0O100O2O0O101O0O101N101N101N101N101N2O0O2O0O2O1N1O2N1O2N2M2O2N1O2N2N2N2N2N2N2N2N2N2N2N3M3M4L3M5K>BSe2"}, "label": "a green vase to the right of two other vaces"}]}
{"id": 287575, "image": "COCO_train2014_000000287575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vase which was at the extreme right in the picture\"."}], "task": "refering", "answer_template": "The \"vase which was at the extreme right in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213], "bbox": [0.63446, 0.55648, 0.9833200000000001, 0.8580266666666666], "iscrowd": 0, "area": 14129.342200000005, "rle": {"size": [375, 500], "counts": "S[d32c;7H7J7H7J5J4M3N2N2N2N2N2N2N2N2O1N2N2O1O1O1N2O1O001O0O2O0O2O0O2N101N101N101N101N101N101N101N101N2O1N101N2O0O100O100O10000O100O1O1N2O1O1N2O1001O001O001O1O001O1O0000001O0000000000001O0000000000001O000000000000O2O00000001O001O1O1O001O1O00001O000O2M200O2O0O100O2O0O101O0O101N101N101N101N101N2O0O2O0O2O1N1O2N1O2N2M2O2N1O2N2N2N2N2N2N2N2N2N2N2N3M3M4L3M5K>BSe2"}, "label": "vase which was at the extreme right in the picture"}]}
{"id": 181084, "image": "COCO_train2014_000000181084.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white police van\"."}], "task": "refering", "answer_template": "The \"a white police van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 231, 235, 236], "bbox": [0.002140625, 0.23529109589041095, 0.302734375, 0.4894178082191781], "iscrowd": 0, "area": 24806.470999999998, "rle": {"size": [584, 640], "counts": "df0d3d>001O001O1O001O1O001O1O001O001O1O001N2O001O001O1O1O1O1O1O1O1O1O1O001O1O1O1O00000000O100000000000000O100000000O010000000O1000000O100000000O100000000O1000000O1000000O100O10000O100O100000000000000000000000000000000000000001O000000000000000O10000000000001O000O10001O000000001O000O10001O0000000O1100O001O1O100O1O010O100O010O1000O011O2M2O2M3M2N3M3M2N3N2M2N3M3L3N3M2N2N2N1N3M2N3M2O2M2N3M2N3MnTn7"}, "label": "a white police van"}]}
{"id": 181084, "image": "COCO_train2014_000000181084.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white police van to the left of the crowd\"."}], "task": "refering", "answer_template": "The \"white police van to the left of the crowd\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 231, 235, 236], "bbox": [0.002140625, 0.23529109589041095, 0.302734375, 0.4894178082191781], "iscrowd": 0, "area": 24806.470999999998, "rle": {"size": [584, 640], "counts": "df0d3d>001O001O1O001O1O001O1O001O001O1O001N2O001O001O1O1O1O1O1O1O1O1O1O001O1O1O1O00000000O100000000000000O100000000O010000000O1000000O100000000O100000000O1000000O1000000O100O10000O100O100000000000000000000000000000000000000001O000000000000000O10000000000001O000O10001O000000001O000O10001O0000000O1100O001O1O100O1O010O100O010O1000O011O2M2O2M3M2N3M3M2N3N2M2N3M3L3N3M2N2N2N1N3M2N3M2O2M2N3M2N3MnTn7"}, "label": "white police van to the left of the crowd"}]}
{"id": 475999, "image": "COCO_train2014_000000475999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giant stack of brownies , a dozen layers high\"."}], "task": "refering", "answer_template": "The \"a giant stack of brownies , a dozen layers high\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269], "bbox": [0.2084375, 0.16104166666666667, 0.746109375, 0.7038958333333334], "iscrowd": 0, "area": 64613.71965000001, "rle": {"size": [480, 640], "counts": "S`n11n>4L5K5L3L5K5K5K2N2N2N2N2N2N2N2N2O0O2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N1O2O1N2N2O1N2N2O1N2N101N2N2O2M3M3N1N3M3N2M3M2O2M3M3N2M3M2O2M3M3N2M2N3N2M3M3N1N3M2O0O10001N1000000O2O00000O2O00000O2O00000O2O0000000O2O00000O2O00000O2O00000O101O000O101O000O101O000O101O000O10001O0000001O0000001O0000001O0O1000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001OVOj00000001O000O2O0000001O00001O00000010O00100O1O1O10O01O1[L\\HMd71`HLa73bHJ_74eHI[77iHFW78mHET79PIDQ7;RICn6;VIBj6>YI_Oh6?\\I^ORO\\MX7U3iI]OjNbM^7o2lIEV6:mIBU6<oI@R6`0QJ\\OQ6b0SJZOo5d0VJWOl5h0WJTOk5j0YJROi5m0ZJoNg5P1]JlNe5S1^JiNd5U1`JXN\\M_OV8X2aJUN\\M@U8Y2cJSN[MAT8[2dJPNZMCS8\\2gJmMYMEQ8\\2jJkMWMGW8W2eJnMS6j1RJRNV6f1i3H8I8G8I7H9H7H\\h[2"}, "label": "a giant stack of brownies , a dozen layers high"}]}
{"id": 475999, "image": "COCO_train2014_000000475999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chocolate cakes in a shop\"."}], "task": "refering", "answer_template": "The \"chocolate cakes in a shop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269], "bbox": [0.2084375, 0.16104166666666667, 0.746109375, 0.7038958333333334], "iscrowd": 0, "area": 64613.71965000001, "rle": {"size": [480, 640], "counts": "S`n11n>4L5K5L3L5K5K5K2N2N2N2N2N2N2N2N2O0O2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N1O2O1N2N2O1N2N2O1N2N101N2N2O2M3M3N1N3M3N2M3M2O2M3M3N2M3M2O2M3M3N2M2N3N2M3M3N1N3M2O0O10001N1000000O2O00000O2O00000O2O00000O2O0000000O2O00000O2O00000O2O00000O101O000O101O000O101O000O101O000O10001O0000001O0000001O0000001O0O1000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001OVOj00000001O000O2O0000001O00001O00000010O00100O1O1O10O01O1[L\\HMd71`HLa73bHJ_74eHI[77iHFW78mHET79PIDQ7;RICn6;VIBj6>YI_Oh6?\\I^ORO\\MX7U3iI]OjNbM^7o2lIEV6:mIBU6<oI@R6`0QJ\\OQ6b0SJZOo5d0VJWOl5h0WJTOk5j0YJROi5m0ZJoNg5P1]JlNe5S1^JiNd5U1`JXN\\M_OV8X2aJUN\\M@U8Y2cJSN[MAT8[2dJPNZMCS8\\2gJmMYMEQ8\\2jJkMWMGW8W2eJnMS6j1RJRNV6f1i3H8I8G8I7H9H7H\\h[2"}, "label": "chocolate cakes in a shop"}]}
{"id": 475999, "image": "COCO_train2014_000000475999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stack of brownies at a buffet\"."}], "task": "refering", "answer_template": "The \"a stack of brownies at a buffet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269], "bbox": [0.2084375, 0.16104166666666667, 0.746109375, 0.7038958333333334], "iscrowd": 0, "area": 64613.71965000001, "rle": {"size": [480, 640], "counts": "S`n11n>4L5K5L3L5K5K5K2N2N2N2N2N2N2N2N2O0O2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2N2O1N2N1O2O1N2N2O1N2N2O1N2N101N2N2O2M3M3N1N3M3N2M3M2O2M3M3N2M3M2O2M3M3N2M2N3N2M3M3N1N3M2O0O10001N1000000O2O00000O2O00000O2O00000O2O0000000O2O00000O2O00000O2O00000O101O000O101O000O101O000O101O000O10001O0000001O0000001O0000001O0O1000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001OVOj00000001O000O2O0000001O00001O00000010O00100O1O1O10O01O1[L\\HMd71`HLa73bHJ_74eHI[77iHFW78mHET79PIDQ7;RICn6;VIBj6>YI_Oh6?\\I^ORO\\MX7U3iI]OjNbM^7o2lIEV6:mIBU6<oI@R6`0QJ\\OQ6b0SJZOo5d0VJWOl5h0WJTOk5j0YJROi5m0ZJoNg5P1]JlNe5S1^JiNd5U1`JXN\\M_OV8X2aJUN\\M@U8Y2cJSN[MAT8[2dJPNZMCS8\\2gJmMYMEQ8\\2jJkMWMGW8W2eJnMS6j1RJRNV6f1i3H8I8G8I7H9H7H\\h[2"}, "label": "a stack of brownies at a buffet"}]}
{"id": 525152, "image": "COCO_train2014_000000525152.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra that ' s center , facing to the left\"."}], "task": "refering", "answer_template": "The \"the zebra that ' s center , facing to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 254, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 282, 283, 292, 293, 294, 297, 298, 307, 308, 309, 311, 312, 313, 321, 322, 323, 326, 327, 328, 336, 337, 338, 341, 342, 343], "bbox": [0.12154929577464789, 0.46896875, 0.9959389671361502, 0.9813281249999999], "iscrowd": 0, "area": 54447.11094999998, "rle": {"size": [640, 426], "counts": "PlP13lc03N2M4L3M4L3M4M2M4L2N1O2N2O0O2N2N1N3M3M2O2M3N1O00001O001O00001O001O001O000010O01O010O010O0010O01O010O0V^OPNea0Q2Z^OPNea0P2[^OPNda0R2[^OoMca0R2\\^OPNba0X2N1O2N2N1O2N2N1O2N2O00100O101N100O100O101N100O100O101N101N101O010O1O001O10O01O001O1O010O1O001O10O01O1O1O010O1O1O1O00100O1O1O1O100O1O1O1O3N1N3N2M2O2N1N2O1eEaKQ5a4kJbKU5_4fJdK[5\\4bJgK]5[4^JhKc5Y4YJjKf5X4VJjKk5X4PJjKQ6W4kIjKV6Y4eIiK\\6X4`IjK`6Y4[IiKf6X4mHSLT7o3aHYL_7i3[H[Lf7g3SH]Lm7f3lG]LV8f3bG^L_8P4nFTLR9Z4\\FjKe9X7001O1O1N2O1M3L4L4L4L4K5A?@`0H8H8H8H8G96RCcIk;o6bCUI\\<i6cC[I\\<c6cCbI[<\\6dChI[<V6eCmIZ<Q6eCTJY<j5fC`JS<^5lCXKa;f4_D^K_;b60O1000000O1O001O001O1O001O0O2N2N1O2N1O2VOTD\\Io;]6]D\\Id;^6iDZIY;_6W1J7I7oNP1L4L4L5I6G9H8K5000001O0000000O1000001O0000000001O000000001O0000001O0000001O000000O1O1O1O1O1O1O100O1O1O0010000001O0000001O0O10001O0000001O0[C]Lm8c3kFfLS9\\3dFmL[9S3aFRM^9n2aFTM^9l2aFVM^9k2_FXM`9j2]FXMb9i2[FZMd9g2ZF[Me9f2XF]Mg9d2WF^Mg9e2VF\\Mj9f2RF\\Mn9g2nE[MQ:h2jEZMV:h2gEYMY:j2`E[M_:h2UEaMk:b2aDoM_;T2WDSNh;T5O00001O00001O00000O2O00001O000000O1000O01000000O1N2K6J5K6K5J5K6K5J6M2M4M3M2M4M3L4M2N3L4eNhBjK[=R4[1M4L4M2M4L4L4L3N3J6I6J7K5J6K4L5H8I7M3M3M3M3M3M3M3M3M3L4G9Ffk0"}, "label": "the zebra that ' s center , facing to the left"}]}
{"id": 525152, "image": "COCO_train2014_000000525152.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra close to the camera and in front of giraffe\"."}], "task": "refering", "answer_template": "The \"the zebra close to the camera and in front of giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 254, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 282, 283, 292, 293, 294, 297, 298, 307, 308, 309, 311, 312, 313, 321, 322, 323, 326, 327, 328, 336, 337, 338, 341, 342, 343], "bbox": [0.12154929577464789, 0.46896875, 0.9959389671361502, 0.9813281249999999], "iscrowd": 0, "area": 54447.11094999998, "rle": {"size": [640, 426], "counts": "PlP13lc03N2M4L3M4L3M4M2M4L2N1O2N2O0O2N2N1N3M3M2O2M3N1O00001O001O00001O001O001O000010O01O010O010O0010O01O010O0V^OPNea0Q2Z^OPNea0P2[^OPNda0R2[^OoMca0R2\\^OPNba0X2N1O2N2N1O2N2N1O2N2O00100O101N100O100O101N100O100O101N101N101O010O1O001O10O01O001O1O010O1O001O10O01O1O1O010O1O1O1O00100O1O1O1O100O1O1O1O3N1N3N2M2O2N1N2O1eEaKQ5a4kJbKU5_4fJdK[5\\4bJgK]5[4^JhKc5Y4YJjKf5X4VJjKk5X4PJjKQ6W4kIjKV6Y4eIiK\\6X4`IjK`6Y4[IiKf6X4mHSLT7o3aHYL_7i3[H[Lf7g3SH]Lm7f3lG]LV8f3bG^L_8P4nFTLR9Z4\\FjKe9X7001O1O1N2O1M3L4L4L4L4K5A?@`0H8H8H8H8G96RCcIk;o6bCUI\\<i6cC[I\\<c6cCbI[<\\6dChI[<V6eCmIZ<Q6eCTJY<j5fC`JS<^5lCXKa;f4_D^K_;b60O1000000O1O001O001O1O001O0O2N2N1O2N1O2VOTD\\Io;]6]D\\Id;^6iDZIY;_6W1J7I7oNP1L4L4L5I6G9H8K5000001O0000000O1000001O0000000001O000000001O0000001O0000001O000000O1O1O1O1O1O1O100O1O1O0010000001O0000001O0O10001O0000001O0[C]Lm8c3kFfLS9\\3dFmL[9S3aFRM^9n2aFTM^9l2aFVM^9k2_FXM`9j2]FXMb9i2[FZMd9g2ZF[Me9f2XF]Mg9d2WF^Mg9e2VF\\Mj9f2RF\\Mn9g2nE[MQ:h2jEZMV:h2gEYMY:j2`E[M_:h2UEaMk:b2aDoM_;T2WDSNh;T5O00001O00001O00000O2O00001O000000O1000O01000000O1N2K6J5K6K5J5K6K5J6M2M4M3M2M4M3L4M2N3L4eNhBjK[=R4[1M4L4M2M4L4L4L3N3J6I6J7K5J6K4L5H8I7M3M3M3M3M3M3M3M3M3L4G9Ffk0"}, "label": "the zebra close to the camera and in front of giraffe"}]}
{"id": 426849, "image": "COCO_train2014_000000426849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and black cat laying down\"."}], "task": "refering", "answer_template": "The \"white and black cat laying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 25, 26, 27, 28, 29, 30, 31, 32, 42, 43, 44, 45, 46, 47, 48, 49, 58, 59, 60, 61, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220], "bbox": [0.3835355648535565, 0.0012968749999999999, 1.0, 0.5612031249999999], "iscrowd": 0, "area": 71274.94720000001, "rle": {"size": [640, 478], "counts": "Pbb32]c0d0C=K5K5K7I`0@:F2N2N2N3M2N2N2N3M2N2N2N3M2O1N2N3M2N2N2N3M2N2M3M3N2M3M4M2M3M3N2M3MnNn@\\MP?c2VAZMg>f2^AYM^>f2iAWMU>h2QBUMl=k2ZBRMc=m2dBPMY=Q3lBlLQ=V3RChLk<[3ZC`Lc<b3bC[LZ<g3kCULR<m3n12N2N3M2POP11O1OoBbKb:]4]EfKb:Y4]EiKc:W4ZEmKe:R4ZEQLd:o3[ETLd:k3ZEXLf:g3YE\\Lf:c3YE`Lf:`3WEdLh:[3XEgLg:X3XEjLh:U3WEnLh:Q3XEQMg:n2XEUMg:j2XEXMh:g2XEZMh:e2WE]Mi:b2VE`Mj:_2UEcMk:\\2UEeMk:Z2TEhMl:Z2PEhMP;Y2nDhMR;Z2jDhMV;Z2fDhMZ;Y2dDhM\\;Z2`DhM`;Z2\\DhMd;Y2ZDhMf;Z2VDhMj;W50000000000000000000000000000000O10000O100O10000O100O10000O100O1001O001O001O001O001O0^FXHl6i7bHiH]7W7WHVIh7b9O1N101O0000001O00001O0000001O00001O0000001O00001O000000O10000000000000000O1000000000000000000O100000000000001O1N2O1O1O1O001O1O1O001O001N101O1O001O001O0O2O001N101O1N101N101O0O2O001N2O001N2O1O1O1O1O1O1O2M2O1O1O1O1O2N1O2N2N2N2N1O2O1N2N101N2N2N2O1N4M3M3M3L4M3M3mMPDoKT<j3PDULa<X3aCgLj<l2YCTMk<e2YCZMj<a2XC_Mk<[2XCeMk<V2XChMm<Q2VCoMm<l1UCTN^=V1eBiN`=Q1bBmNd=l0^BSOh=f0[BXOk=b0VB^Oa>EfA9^>]OiAb0[>TOmAk0V>lNPBS1T>cNTB[1Y`001N2N101N1O2N1I8G8I8H]F"}, "label": "white and black cat laying down"}]}
{"id": 426849, "image": "COCO_train2014_000000426849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat laying with its head almost touching a black keyboard\"."}], "task": "refering", "answer_template": "The \"a cat laying with its head almost touching a black keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 25, 26, 27, 28, 29, 30, 31, 32, 42, 43, 44, 45, 46, 47, 48, 49, 58, 59, 60, 61, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220], "bbox": [0.3835355648535565, 0.0012968749999999999, 1.0, 0.5612031249999999], "iscrowd": 0, "area": 71274.94720000001, "rle": {"size": [640, 478], "counts": "Pbb32]c0d0C=K5K5K7I`0@:F2N2N2N3M2N2N2N3M2N2N2N3M2O1N2N3M2N2N2N3M2N2M3M3N2M3M4M2M3M3N2M3MnNn@\\MP?c2VAZMg>f2^AYM^>f2iAWMU>h2QBUMl=k2ZBRMc=m2dBPMY=Q3lBlLQ=V3RChLk<[3ZC`Lc<b3bC[LZ<g3kCULR<m3n12N2N3M2POP11O1OoBbKb:]4]EfKb:Y4]EiKc:W4ZEmKe:R4ZEQLd:o3[ETLd:k3ZEXLf:g3YE\\Lf:c3YE`Lf:`3WEdLh:[3XEgLg:X3XEjLh:U3WEnLh:Q3XEQMg:n2XEUMg:j2XEXMh:g2XEZMh:e2WE]Mi:b2VE`Mj:_2UEcMk:\\2UEeMk:Z2TEhMl:Z2PEhMP;Y2nDhMR;Z2jDhMV;Z2fDhMZ;Y2dDhM\\;Z2`DhM`;Z2\\DhMd;Y2ZDhMf;Z2VDhMj;W50000000000000000000000000000000O10000O100O10000O100O10000O100O1001O001O001O001O001O0^FXHl6i7bHiH]7W7WHVIh7b9O1N101O0000001O00001O0000001O00001O0000001O00001O000000O10000000000000000O1000000000000000000O100000000000001O1N2O1O1O1O001O1O1O001O001N101O1O001O001O0O2O001N101O1N101N101O0O2O001N2O001N2O1O1O1O1O1O1O2M2O1O1O1O1O2N1O2N2N2N2N1O2O1N2N101N2N2N2O1N4M3M3M3L4M3M3mMPDoKT<j3PDULa<X3aCgLj<l2YCTMk<e2YCZMj<a2XC_Mk<[2XCeMk<V2XChMm<Q2VCoMm<l1UCTN^=V1eBiN`=Q1bBmNd=l0^BSOh=f0[BXOk=b0VB^Oa>EfA9^>]OiAb0[>TOmAk0V>lNPBS1T>cNTB[1Y`001N2N101N1O2N1I8G8I8H]F"}, "label": "a cat laying with its head almost touching a black keyboard"}]}
{"id": 426849, "image": "COCO_train2014_000000426849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cat sitting on someone ' s left arm\"."}], "task": "refering", "answer_template": "The \"an orange cat sitting on someone ' s left arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 312, 313, 314, 324, 325, 326, 330, 331, 332, 347, 348, 349, 350, 365, 366], "bbox": [0.0, 0.35203125, 0.6507322175732217, 0.9606406249999999], "iscrowd": 0, "area": 73279.45435, "rle": {"size": [640, 478], "counts": "^?`0_b0R1nNR1nNQ1oMcL^B_4[<\\LoBf4j;[2E<O1O0O2O1O001N2O001O0O2O1O001N2O001O1O0O2O1O001N2O001O0O2O1O001N2O001O1O0O2O1O001N2O001O0O10O100000O0100000O010000O0100000O01000O1000O10O1000O0100000O0101O000O10000O1000000O10000O2O00000O10000O100O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O100O1O100O100O1O2O0O100O1O100O100O1O100O101N1O100O1001O2N3M2N3M2N2N3N1N2N2N2N2N3M2N2N2N2N2N3M2N2O1N2N2N2N3M2N2N2N2N2NbKhEWOU:j0nETOP:m0RFROm9n0UFQOi9o0ZFPOd9Q1_FmN_9T1cFlNZ9U1hFjNV9W1lFhNR9X1RGfNm8Z1UGeNi8\\1YGcNe8^1^GaN`8_1bG`N\\8`1gG_NX8`1kG_NS8`1RH^Nm7a1VH^Nh7a1\\H]Nd7a1`H\\N`7d1dHYN\\7e1hHXNX7g1lHWNT7g1QIVNn6j1UISNk6l1XISNh6k1[ISNe6l1^ISNb6l1`IRN`6m1cIRN]6k1fITNZ6j1jIUNV6h1nIWNQ6f1TJXNm5e1WJZNh5d1\\JZNe5b1`J]N_5a1eJ]N[5a1iJ^NW5^1nJaNQ5]1SKaNn4\\1VKcNi4[1[KcNf4Y1_KfN`4X1dKfN]4W1gKhNX4U1mKiNT4T1PLkNo3ROQGa0T5<j3SO\\G5n4f0g3UOeGGi4S1a3VOPH[Oc4]1^3XOXHoN^4h1Y3YO_Nf0a1ZO`Nd0`1\\ObNc0^1]ObNc0]1]OeNa0\\1_OeN`0Z1@hN?X1AiN>V1BlN<T1ElN;T1DmN<R1EnN:S1EnN;Q1FoN:Q1EPO:P1GoN:Q1EPO[OPH:o8<QOYOQH4U9c0jNXORHM[9k0dNVOTHMY9m0cNUOUHMX9n0dNTOUHMW9P1cNSOWHLW9P1cNSOWHLV9R1cNPOYHMU9R1cNPOYHMT9T1cNnNZHMT9T1cNnNZHMS9V1cNlN[HMS9V1bNlN]HMQ9X1bNjN^HMQ9X1bNjN^HMP9Z1bNhNT2X1mMfNU2Z1kMeNU2[1kMeNV2[1W8010O010O010O10O01O010O010O010O010O01N1_Ob0^OgSX3"}, "label": "an orange cat sitting on someone ' s left arm"}]}
{"id": 426849, "image": "COCO_train2014_000000426849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange tabby kitty helping type a email\"."}], "task": "refering", "answer_template": "The \"orange tabby kitty helping type a email\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 312, 313, 314, 324, 325, 326, 330, 331, 332, 347, 348, 349, 350, 365, 366], "bbox": [0.0, 0.35203125, 0.6507322175732217, 0.9606406249999999], "iscrowd": 0, "area": 73279.45435, "rle": {"size": [640, 478], "counts": "^?`0_b0R1nNR1nNQ1oMcL^B_4[<\\LoBf4j;[2E<O1O0O2O1O001N2O001O0O2O1O001N2O001O1O0O2O1O001N2O001O0O2O1O001N2O001O1O0O2O1O001N2O001O0O10O100000O0100000O010000O0100000O01000O1000O10O1000O0100000O0101O000O10000O1000000O10000O2O00000O10000O100O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O100O1O100O100O1O2O0O100O1O100O100O1O100O101N1O100O1001O2N3M2N3M2N2N3N1N2N2N2N2N3M2N2N2N2N2N3M2N2O1N2N2N2N3M2N2N2N2N2NbKhEWOU:j0nETOP:m0RFROm9n0UFQOi9o0ZFPOd9Q1_FmN_9T1cFlNZ9U1hFjNV9W1lFhNR9X1RGfNm8Z1UGeNi8\\1YGcNe8^1^GaN`8_1bG`N\\8`1gG_NX8`1kG_NS8`1RH^Nm7a1VH^Nh7a1\\H]Nd7a1`H\\N`7d1dHYN\\7e1hHXNX7g1lHWNT7g1QIVNn6j1UISNk6l1XISNh6k1[ISNe6l1^ISNb6l1`IRN`6m1cIRN]6k1fITNZ6j1jIUNV6h1nIWNQ6f1TJXNm5e1WJZNh5d1\\JZNe5b1`J]N_5a1eJ]N[5a1iJ^NW5^1nJaNQ5]1SKaNn4\\1VKcNi4[1[KcNf4Y1_KfN`4X1dKfN]4W1gKhNX4U1mKiNT4T1PLkNo3ROQGa0T5<j3SO\\G5n4f0g3UOeGGi4S1a3VOPH[Oc4]1^3XOXHoN^4h1Y3YO_Nf0a1ZO`Nd0`1\\ObNc0^1]ObNc0]1]OeNa0\\1_OeN`0Z1@hN?X1AiN>V1BlN<T1ElN;T1DmN<R1EnN:S1EnN;Q1FoN:Q1EPO:P1GoN:Q1EPO[OPH:o8<QOYOQH4U9c0jNXORHM[9k0dNVOTHMY9m0cNUOUHMX9n0dNTOUHMW9P1cNSOWHLW9P1cNSOWHLV9R1cNPOYHMU9R1cNPOYHMT9T1cNnNZHMT9T1cNnNZHMS9V1cNlN[HMS9V1bNlN]HMQ9X1bNjN^HMQ9X1bNjN^HMP9Z1bNhNT2X1mMfNU2Z1kMeNU2[1kMeNV2[1W8010O010O010O10O01O010O010O010O010O01N1_Ob0^OgSX3"}, "label": "orange tabby kitty helping type a email"}]}
{"id": 9057, "image": "COCO_train2014_000000009057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady watching the water\"."}], "task": "refering", "answer_template": "The \"a lady watching the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 255, 256, 257, 278, 279, 280, 300, 301, 302, 303, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.026968750000000003, 0.6269583333333333, 0.2528125, 0.9842708333333333], "iscrowd": 0, "area": 14480.43625, "rle": {"size": [480, 640], "counts": "[]82k>3M4K5L4K4M4K5L4K4M4L3L4M4K4M4K4M3L5L3O1O2N1O1O100O1O001O1O1O00100O1O00100O00100O100O100O1O100O100O1cNcMVF^2h9eMWF[2g9hMWFY2g9jMWFX2g9jMVFX2i9lMRFV2m9oMmES2R:^1O100O1O1O1O1O100O1O1O1O10000O100O10000O100O10000O1001O1O001O1O1O1O10O01O1O1O1O5XLoEW2U:aMPF^2U:YMPFf2U:PMRFn2R:jLSFU3m:O1O1O1O1O100O1O1O1O2N1O1O1O1O1O1O1O1O1O101N3M3M4L3M4L3M3M4L3M3M4L3M4J5K5L5L3L6K6I6Kaco6"}, "label": "a lady watching the water"}]}
{"id": 9057, "image": "COCO_train2014_000000009057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women see the sea water in back position\"."}], "task": "refering", "answer_template": "The \"a women see the sea water in back position\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 255, 256, 257, 278, 279, 280, 300, 301, 302, 303, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.026968750000000003, 0.6269583333333333, 0.2528125, 0.9842708333333333], "iscrowd": 0, "area": 14480.43625, "rle": {"size": [480, 640], "counts": "[]82k>3M4K5L4K4M4K5L4K4M4L3L4M4K4M4K4M3L5L3O1O2N1O1O100O1O001O1O1O00100O1O00100O00100O100O100O1O100O100O1cNcMVF^2h9eMWF[2g9hMWFY2g9jMWFX2g9jMVFX2i9lMRFV2m9oMmES2R:^1O100O1O1O1O1O100O1O1O1O10000O100O10000O100O10000O1001O1O001O1O1O1O10O01O1O1O1O5XLoEW2U:aMPF^2U:YMPFf2U:PMRFn2R:jLSFU3m:O1O1O1O1O100O1O1O1O2N1O1O1O1O1O1O1O1O1O101N3M3M4L3M4L3M3M4L3M3M4L3M4J5K5L5L3L6K6I6Kaco6"}, "label": "a women see the sea water in back position"}]}
{"id": 9057, "image": "COCO_train2014_000000009057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt watching the jet flying\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt watching the jet flying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 246, 247, 248, 269, 270, 291, 292, 293, 294, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 382, 383, 384, 385, 386], "bbox": [0.623421875, 0.5449166666666667, 0.86146875, 0.9857291666666667], "iscrowd": 0, "area": 20183.144599999996, "rle": {"size": [480, 640], "counts": "g_k52_=OjC6R<OgC7U<NeC7W<MdC8Y<LaC9[<L_C9]<L]C9`<U1L4L4L4M3L4L4L4M3L4L4G91OO100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O1YObKlF^4P9hKmFY4o8mKnFT4n8QLQGo3l8VLQGk3l8V1N2M3M3M3N2M3N200O10000O1000O010000O10000O10000O10000O10000O10000000ZKdGh2\\8WMgGg2Z8VMjGh2W8VMlGh2V8TMnGj2S8TMPHj2Q8SMQHm2P8QMRHn2o7oLTHP3m7nLUHQ3l7fL]HY3e7aL`H^3a7^LcHa3^7[LeHe3\\7VLiHi3X7SLlHl3U7PLoHo3R9O1O1O001O1O1O1O1O1O001O1O1]OiD[MX;d2c0N2O001O1N2O1O1O1O0O2O1O1O1O1N3N4L3M4L4K4M4J6H8H7H9H8H7IUZY1"}, "label": "a man in an orange shirt watching the jet flying"}]}
{"id": 9057, "image": "COCO_train2014_000000009057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man was seeing water\"."}], "task": "refering", "answer_template": "The \"man was seeing water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 246, 247, 248, 269, 270, 291, 292, 293, 294, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 382, 383, 384, 385, 386], "bbox": [0.623421875, 0.5449166666666667, 0.86146875, 0.9857291666666667], "iscrowd": 0, "area": 20183.144599999996, "rle": {"size": [480, 640], "counts": "g_k52_=OjC6R<OgC7U<NeC7W<MdC8Y<LaC9[<L_C9]<L]C9`<U1L4L4L4M3L4L4L4M3L4L4G91OO100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O1YObKlF^4P9hKmFY4o8mKnFT4n8QLQGo3l8VLQGk3l8V1N2M3M3M3N2M3N200O10000O1000O010000O10000O10000O10000O10000O10000000ZKdGh2\\8WMgGg2Z8VMjGh2W8VMlGh2V8TMnGj2S8TMPHj2Q8SMQHm2P8QMRHn2o7oLTHP3m7nLUHQ3l7fL]HY3e7aL`H^3a7^LcHa3^7[LeHe3\\7VLiHi3X7SLlHl3U7PLoHo3R9O1O1O001O1O1O1O1O1O001O1O1]OiD[MX;d2c0N2O001O1N2O1O1O1O0O2O1O1O1O1N3N4L3M4L4K4M4J6H8H7H9H8H7IUZY1"}, "label": "man was seeing water"}]}
{"id": 90985, "image": "COCO_train2014_000000090985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange in the 5 o ' clock position\"."}], "task": "refering", "answer_template": "The \"the orange in the 5 o ' clock position\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.487140625, 0.5448148148148148, 0.916984375, 0.9882098765432098], "iscrowd": 0, "area": 48801.381700000005, "rle": {"size": [486, 640], "counts": "R_d45m>6I7I7I7J6I7I7J6I7I7I7J7K5L4K5L3L5K5L4K5L3L5K5L4K5L3L5K5L4K2O1N2N2O1N2O1N2O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O1O100O100O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000001O000000001O00000000001O000000001O000000001N100000001O000000001O000000001O00001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1N2M3N3L3N3L4M3L5L3L4M3L5L3L4M3L5K4M3L4M4K4M3L4MWih0"}, "label": "the orange in the 5 o ' clock position"}]}
{"id": 90985, "image": "COCO_train2014_000000090985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange that does not have any light reflections on it\"."}], "task": "refering", "answer_template": "The \"the orange that does not have any light reflections on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.487140625, 0.5448148148148148, 0.916984375, 0.9882098765432098], "iscrowd": 0, "area": 48801.381700000005, "rle": {"size": [486, 640], "counts": "R_d45m>6I7I7I7J6I7I7J6I7I7I7J7K5L4K5L3L5K5L4K5L3L5K5L4K5L3L5K5L4K2O1N2N2O1N2O1N2O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O1O100O100O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000001O000000001O00000000001O000000001O000000001N100000001O000000001O000000001O00001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1N2M3N3L3N3L4M3L5L3L4M3L5L3L4M3L5K4M3L4M4K4M3L4MWih0"}, "label": "the orange that does not have any light reflections on it"}]}
{"id": 90985, "image": "COCO_train2014_000000090985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"6 oranges with a small tomato perched on top\"."}], "task": "refering", "answer_template": "The \"6 oranges with a small tomato perched on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 30, 31, 32, 33, 34, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.012531249999999999, 0.05306584362139918, 0.8793749999999999, 0.9846707818930042], "iscrowd": 0, "area": 156166.93769999995, "rle": {"size": [486, 640], "counts": "Uo34o>7I7I8H7I7I8G8I7I8H8H8H7I8I7H6K3M3_GgLR5Z3mJUNd3m1YLWNc3l1[LWNb3k1[LYNb3i1\\LZN`3i1]L[N`3g1]L]N`3e1^L_N]3c1aLaN\\3a1bLbN[3`1bLdN[3^1cLeNZ3]1cLgNY3\\1eLgNX3[1eLjNW3X1gLkNV3W1gLmNV3T1iLoNT3S1iLQOT3Q1jLROR3Q1kLSOR3o0kLVOR3k0lLXOQ3j0lLZOQ3h0mL[OP3f0nL^Oo2d0oL_On2c0oLAo2`0oLDm2>PMFm2<QMGl2;QMIl29RMJk28RMLl24PM1n21nL1R3R600001O0000000000000O10000000000000000000000O10000000001O0000O1000O10000000O1000000000000O0100000000000O1000000O01000O100O10000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000001O001O00N1O2O1N2N2N2O1N2N2N2O1N1O2O1N2N2O1O1O1O1O1N2O1O001O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1N2O1O1O1O0010000000000000000000O1000000000001O000000000O100000000000000000000O1000000000001O000000000O10000000000000000000000O10000000000N2N2N2N2O1N2N2N2N2`KiMoKY2Q4iMlKX2T4jMhKX2Y4jMbKX2^4jM^KX2c4iMYKY2g4iMVKX2j4jMRKX2o4iMmJY2S5iMiJY2W5jMdJX2]5iM_JY2a5iM\\JX2d5jMXJX2h5jMTJX2l5jMPJX2Q6iMkIY2U6iMgIY2Y6jMcIW2]6^NlHd1T7fN`H\\1`7mNUHU1l7TOhGn0X8[O]Gg0c8n201O1O2O0O2N1O2N2N1O2O0O2N1O2N1O2O0O2N2N1O2N1O2O0O2N1O2N2N101N1O2N1O2N1O2O0O2N8G<E<oInFU5c9N01N100O2O0O100O2O0O101N100K6A>N2N2M4O0O101N10001N1000010O01O10O01O010O1O0010O01O010O1O010O0010O10O0100O00100O010O1O010O1O010O1O010O10O0100O01000O010O10O10000lNhGYKY8g4jGVKU8k4mGSKS8m4PHoJQ8Q5RHlJn7S5UHkJk7U5XHgJi7X5ZHdJg7]5Q100O100O100O10000O100O101O0O100O100O1[OaFcK_9Z4kF_KU9_4j0O100O1O101N1O100O1O100O1O100O1O100O1O101N100O100O1O100O100O1O100O100O2N100O100O1O100O2O0O1O3N1N2O1N2N2O1N2O1N3M2O1N2O1N2O1N2N2O1N3N1N2N2O1N2O1N2N2O2M2O1N2N2O2M2O1N3M3N2M2O2L5K9F:G6J6Id]T1"}, "label": "6 oranges with a small tomato perched on top"}]}
{"id": 90985, "image": "COCO_train2014_000000090985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small tomato in the middle on top of a group on oranges in a basket\"."}], "task": "refering", "answer_template": "The \"a small tomato in the middle on top of a group on oranges in a basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 30, 31, 32, 33, 34, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.012531249999999999, 0.05306584362139918, 0.8793749999999999, 0.9846707818930042], "iscrowd": 0, "area": 156166.93769999995, "rle": {"size": [486, 640], "counts": "Uo34o>7I7I8H7I7I8G8I7I8H8H8H7I8I7H6K3M3_GgLR5Z3mJUNd3m1YLWNc3l1[LWNb3k1[LYNb3i1\\LZN`3i1]L[N`3g1]L]N`3e1^L_N]3c1aLaN\\3a1bLbN[3`1bLdN[3^1cLeNZ3]1cLgNY3\\1eLgNX3[1eLjNW3X1gLkNV3W1gLmNV3T1iLoNT3S1iLQOT3Q1jLROR3Q1kLSOR3o0kLVOR3k0lLXOQ3j0lLZOQ3h0mL[OP3f0nL^Oo2d0oL_On2c0oLAo2`0oLDm2>PMFm2<QMGl2;QMIl29RMJk28RMLl24PM1n21nL1R3R600001O0000000000000O10000000000000000000000O10000000001O0000O1000O10000000O1000000000000O0100000000000O1000000O01000O100O10000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000001O001O00N1O2O1N2N2N2O1N2N2N2O1N1O2O1N2N2O1O1O1O1O1N2O1O001O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1N2O1O1O1O0010000000000000000000O1000000000001O000000000O100000000000000000000O1000000000001O000000000O10000000000000000000000O10000000000N2N2N2N2O1N2N2N2N2`KiMoKY2Q4iMlKX2T4jMhKX2Y4jMbKX2^4jM^KX2c4iMYKY2g4iMVKX2j4jMRKX2o4iMmJY2S5iMiJY2W5jMdJX2]5iM_JY2a5iM\\JX2d5jMXJX2h5jMTJX2l5jMPJX2Q6iMkIY2U6iMgIY2Y6jMcIW2]6^NlHd1T7fN`H\\1`7mNUHU1l7TOhGn0X8[O]Gg0c8n201O1O2O0O2N1O2N2N1O2O0O2N1O2N1O2O0O2N2N1O2N1O2O0O2N1O2N2N101N1O2N1O2N1O2O0O2N8G<E<oInFU5c9N01N100O2O0O100O2O0O101N100K6A>N2N2M4O0O101N10001N1000010O01O10O01O010O1O0010O01O010O1O010O0010O10O0100O00100O010O1O010O1O010O1O010O10O0100O01000O010O10O10000lNhGYKY8g4jGVKU8k4mGSKS8m4PHoJQ8Q5RHlJn7S5UHkJk7U5XHgJi7X5ZHdJg7]5Q100O100O100O10000O100O101O0O100O100O1[OaFcK_9Z4kF_KU9_4j0O100O1O101N1O100O1O100O1O100O1O100O1O101N100O100O1O100O100O1O100O100O2N100O100O1O100O2O0O1O3N1N2O1N2N2O1N2O1N3M2O1N2O1N2O1N2N2O1N3N1N2N2O1N2O1N2N2O2M2O1N2N2O2M2O1N3M3N2M2O2L5K9F:G6J6Id]T1"}, "label": "a small tomato in the middle on top of a group on oranges in a basket"}]}
{"id": 115564, "image": "COCO_train2014_000000115564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"five car train to right of photo\"."}], "task": "refering", "answer_template": "The \"five car train to right of photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225], "bbox": [0.0375625, 0.3752803738317757, 0.793484375, 0.6584345794392523], "iscrowd": 0, "area": 27624.930899999992, "rle": {"size": [428, 640], "counts": "gV:6P=7N1O100O1O100O2N100O1O100O1O2O0O1O1O1O1O101N1000000000001O000001O001O010O1O001O001O010O00001O000010O00000000001O001O001O001O001mCROf;n0YDSOg;n0WDSOi;m0UDUOk;l0SDUOm;k0RDWOl;U10000O1000001N10000O10001N1001O000001O000000000010O000000000001O000001O0001O00000000001O01O00000001O00000001O01O0000000000001O01O00000001O00000000010O00000000001O000001O000001O000000000010O0000000001O00000001O01O0000000000001O01O00000001O00000000010O00000000001O0001O00000001O00000000010O00000000001O000001O0001O0000000000001O0001O000001O00000000001O000001O0001O0000000000001O0001O000001O00000000001O000001O0001O0000000000001O000001O0001O00000000001O00000001O01O0000000000001O000001O0001O00000000001O00000001O01O0000000000001O000001O0001O00000000001O0001O000001O00000000000010O0000000001O000000000010O000000000001O00000001O01O00000000001O000001O0001O0000000000001O0001O000001O00000000001O01O00000001O00000000000010O0000000001O00000000010O00000000001O000000010O000000001O01O00010O001O00010O001O00012Md0\\Oe0[O?A010O001O1O010M2N3L4M2MVmf1"}, "label": "five car train to right of photo"}]}
{"id": 115564, "image": "COCO_train2014_000000115564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bunch of colorful train cars\"."}], "task": "refering", "answer_template": "The \"a bunch of colorful train cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225], "bbox": [0.0375625, 0.3752803738317757, 0.793484375, 0.6584345794392523], "iscrowd": 0, "area": 27624.930899999992, "rle": {"size": [428, 640], "counts": "gV:6P=7N1O100O1O100O2N100O1O100O1O2O0O1O1O1O1O101N1000000000001O000001O001O010O1O001O001O010O00001O000010O00000000001O001O001O001O001mCROf;n0YDSOg;n0WDSOi;m0UDUOk;l0SDUOm;k0RDWOl;U10000O1000001N10000O10001N1001O000001O000000000010O000000000001O000001O0001O00000000001O01O00000001O00000001O01O0000000000001O01O00000001O00000000010O00000000001O000001O000001O000000000010O0000000001O00000001O01O0000000000001O01O00000001O00000000010O00000000001O0001O00000001O00000000010O00000000001O000001O0001O0000000000001O0001O000001O00000000001O000001O0001O0000000000001O0001O000001O00000000001O000001O0001O0000000000001O000001O0001O00000000001O00000001O01O0000000000001O000001O0001O00000000001O00000001O01O0000000000001O000001O0001O00000000001O0001O000001O00000000000010O0000000001O000000000010O000000000001O00000001O01O00000000001O000001O0001O0000000000001O0001O000001O00000000001O01O00000001O00000000000010O0000000001O00000000010O00000000001O000000010O000000001O01O00010O001O00010O001O00012Md0\\Oe0[O?A010O001O1O010M2N3L4M2MVmf1"}, "label": "a bunch of colorful train cars"}]}
{"id": 377709, "image": "COCO_train2014_000000377709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus with its lights on\"."}], "task": "refering", "answer_template": "The \"the bus with its lights on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 280, 281, 282, 283, 284, 285, 296], "bbox": [0.047546875, 0.2396, 0.9130312499999998, 0.895075], "iscrowd": 0, "area": 119348.75185000002, "rle": {"size": [400, 640], "counts": "lQ<2n;`0_Ob0^Oa0@`0_Ob0^Oa0@`0_Oa0_Ob0_O`0D<01O0O101O000O2O00001N1000000O10000O1000000O10O1000O1000000O1000000O1000000O0100H8G8H9G9G9K410O1000O10O1000O0100000O01000O102N1O1N3N1O2N4K>C=C<D4K1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O100000000O100000000O100000000O100000000O1000001O0O1000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O10lNoJiJQ5S5aK_J_4]5SLUJm3g5_1N3L4L4L4M2N3001O000O10001O000O1000001O0O1000001O0O1000001O8G;F:F:F:F3L100000000O10O100000O1000O1000O10000000O0100000000O010000000O1000O1000O100000O10O100000000O010000000O10O100000O1000O1000O10000000O0100000000O010000000O100000000O10000000000000001O000000000001O00000000000000000001O0000000001O000000000000000000000001O00000001O00000000000000000000000001O0001O0000000000000000000000000000010O00000000000000000000000000000010O00000000000000000000000000000001O01O0000000000000000000000000001O000001O000000000000000000000001O00000001O0000000001O00000000001O0000000000010O00000000001O0000000000001O00000001O0001O00009Gc0]Ob0^Ob0^Oc0]Ob0^Oc0]Ob0^Ob0^Oa0_O`0@`0@?Bdae0"}, "label": "the bus with its lights on"}]}
{"id": 377709, "image": "COCO_train2014_000000377709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bus parked\"."}], "task": "refering", "answer_template": "The \"white bus parked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 280, 281, 282, 283, 284, 285, 296], "bbox": [0.047546875, 0.2396, 0.9130312499999998, 0.895075], "iscrowd": 0, "area": 119348.75185000002, "rle": {"size": [400, 640], "counts": "lQ<2n;`0_Ob0^Oa0@`0_Ob0^Oa0@`0_Oa0_Ob0_O`0D<01O0O101O000O2O00001N1000000O10000O1000000O10O1000O1000000O1000000O1000000O0100H8G8H9G9G9K410O1000O10O1000O0100000O01000O102N1O1N3N1O2N4K>C=C<D4K1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O100000000O100000000O100000000O100000000O1000001O0O1000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O10lNoJiJQ5S5aK_J_4]5SLUJm3g5_1N3L4L4L4M2N3001O000O10001O000O1000001O0O1000001O0O1000001O8G;F:F:F:F3L100000000O10O100000O1000O1000O10000000O0100000000O010000000O1000O1000O100000O10O100000000O010000000O10O100000O1000O1000O10000000O0100000000O010000000O100000000O10000000000000001O000000000001O00000000000000000001O0000000001O000000000000000000000001O00000001O00000000000000000000000001O0001O0000000000000000000000000000010O00000000000000000000000000000010O00000000000000000000000000000001O01O0000000000000000000000000001O000001O000000000000000000000001O00000001O0000000001O00000000001O0000000000010O00000000001O0000000000001O00000001O0001O00009Gc0]Ob0^Ob0^Oc0]Ob0^Oc0]Ob0^Ob0^Oa0_O`0@`0@?Bdae0"}, "label": "white bus parked"}]}
{"id": 500594, "image": "COCO_train2014_000000500594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating from a branch a woman is holding\"."}], "task": "refering", "answer_template": "The \"a giraffe eating from a branch a woman is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 260, 261, 262, 263, 278, 279, 280, 281, 285, 301, 302, 303, 304, 324, 325, 326, 327], "bbox": [0.08545312499999999, 0.32583138173302106, 0.46328125, 0.9865105386416863], "iscrowd": 0, "area": 30884.565249999996, "rle": {"size": [427, 640], "counts": "[Yg0\\1i;=C9C:F:F:N2M3N2N2M3N2N2M3N2M3N2N2J6M3N2O1N2O1N2O1N2O1N2N3N1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1OVOaIjJ^6S5mIfJS6V5XJcJh5Y5bJaJ^5T5`InJj7l4\\HQKh7i4`0J6J5J7J6J6J5M4L4L4L4L4L4N21O001O1O1N2N1N3M3L4L4]OkLjFY3S9b0L4K5L3M4M2M3N3M201N10001N1O2N1O1O2M20010O010O01O01O010O01O010O01O03M7J6I7I7I5L3L4L4L4L2OO01O1O010O001O0010O010000O100O2O0O100O2O0O2O1N2O1N2O7H9G9H2M2O2M2O1O000O2O000O2O00001N101O1N101O1O0O2N2M2O2M3M101BUDYOm;f0UDWOm;h0TDWOl;i0UDUOm;j0TDUOm;j0UDSOm;k0:01O1O001O1O0M4K5L3M4N1N3McT_4"}, "label": "a giraffe eating from a branch a woman is holding"}]}
{"id": 500594, "image": "COCO_train2014_000000500594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe taking a branch from the woman\"."}], "task": "refering", "answer_template": "The \"the giraffe taking a branch from the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 260, 261, 262, 263, 278, 279, 280, 281, 285, 301, 302, 303, 304, 324, 325, 326, 327], "bbox": [0.08545312499999999, 0.32583138173302106, 0.46328125, 0.9865105386416863], "iscrowd": 0, "area": 30884.565249999996, "rle": {"size": [427, 640], "counts": "[Yg0\\1i;=C9C:F:F:N2M3N2N2M3N2N2M3N2M3N2N2J6M3N2O1N2O1N2O1N2O1N2N3N1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1OVOaIjJ^6S5mIfJS6V5XJcJh5Y5bJaJ^5T5`InJj7l4\\HQKh7i4`0J6J5J7J6J6J5M4L4L4L4L4L4N21O001O1O1N2N1N3M3L4L4]OkLjFY3S9b0L4K5L3M4M2M3N3M201N10001N1O2N1O1O2M20010O010O01O01O010O01O010O01O03M7J6I7I7I5L3L4L4L4L2OO01O1O010O001O0010O010000O100O2O0O100O2O0O2O1N2O1N2O7H9G9H2M2O2M2O1O000O2O000O2O00001N101O1N101O1O0O2N2M2O2M3M101BUDYOm;f0UDWOm;h0TDWOl;i0UDUOm;j0TDUOm;j0UDSOm;k0:01O1O001O1O0M4K5L3M4N1N3McT_4"}, "label": "the giraffe taking a branch from the woman"}]}
{"id": 500594, "image": "COCO_train2014_000000500594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a taller giraffe\"."}], "task": "refering", "answer_template": "The \"a taller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 54, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 187, 188], "bbox": [0.1469375, 0.1039344262295082, 0.47078125000000004, 0.5938173302107729], "iscrowd": 0, "area": 16251.666100000006, "rle": {"size": [427, 640], "counts": "jXX11X=1O2N1O1O1O1O1O1O2N1N2O1O1O1O1VLAhJ`0Y5CbJ?_5D[J=f5FUJ;l5HoI9R6HjI:W6GfI:[6GbI:_6G^I;c6EZI<g6EVI<k6ERI<o6EnH<S7EjH<W7EeH=\\7DaH=a7D[H>e7CWH?j7BSH?n7BnG`0S8AjG`0X8_OfGb0[8^ObGd0`8[O^Gf0c8ZOZGi0f8WOXGj0j8UOSGm0n8TOoFm0S9ROkFo0V9QOgFQ1[9nNcFS1^9mN_FV1a9jN]FW1e9hNYFY1h9gNUF[1m9eNPF\\1P:eNmE]1T:l000010O01O1O10O01O100O00100O00O2O0O1ObNmLgHR3[7SM_Hl2c7XMXHh2i7\\MRHm1[ORNe87iGf1MhM\\8h0`G`1W9fNbFY1`9mNYFR1i9V12O1N2N2O3L4L3N3N2N2N1O2N2N1O2N2N1O2N2NN2M2M4M3O1O1N2O1O4L3M3M4L4N2N1O2N2N2N2N1N4O3MO1M3M1O01O00001O000000000000O01001O2N2N1O2N2M2O2N2N2N1O2N2M2O2N2N2N1O1O2M2O1O1O1O2N1O1N2O100O2N1O100O1O2N1O100O1O2N100O1O1O1O2O0O1O1O1N3M2N2N2N2N3M2N2N2N2N3N1M3J6J6ImV]4"}, "label": "a taller giraffe"}]}
{"id": 500594, "image": "COCO_train2014_000000500594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller of the two giraffes eating leaves\"."}], "task": "refering", "answer_template": "The \"the taller of the two giraffes eating leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 54, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 187, 188], "bbox": [0.1469375, 0.1039344262295082, 0.47078125000000004, 0.5938173302107729], "iscrowd": 0, "area": 16251.666100000006, "rle": {"size": [427, 640], "counts": "jXX11X=1O2N1O1O1O1O1O1O2N1N2O1O1O1O1VLAhJ`0Y5CbJ?_5D[J=f5FUJ;l5HoI9R6HjI:W6GfI:[6GbI:_6G^I;c6EZI<g6EVI<k6ERI<o6EnH<S7EjH<W7EeH=\\7DaH=a7D[H>e7CWH?j7BSH?n7BnG`0S8AjG`0X8_OfGb0[8^ObGd0`8[O^Gf0c8ZOZGi0f8WOXGj0j8UOSGm0n8TOoFm0S9ROkFo0V9QOgFQ1[9nNcFS1^9mN_FV1a9jN]FW1e9hNYFY1h9gNUF[1m9eNPF\\1P:eNmE]1T:l000010O01O1O10O01O100O00100O00O2O0O1ObNmLgHR3[7SM_Hl2c7XMXHh2i7\\MRHm1[ORNe87iGf1MhM\\8h0`G`1W9fNbFY1`9mNYFR1i9V12O1N2N2O3L4L3N3N2N2N1O2N2N1O2N2N1O2N2NN2M2M4M3O1O1N2O1O4L3M3M4L4N2N1O2N2N2N2N1N4O3MO1M3M1O01O00001O000000000000O01001O2N2N1O2N2M2O2N2N2N1O2N2M2O2N2N2N1O1O2M2O1O1O1O2N1O1N2O100O2N1O100O1O2N1O100O1O2N100O1O1O1O2O0O1O1O1N3M2N2N2N2N3M2N2N2N2N3N1M3J6J6ImV]4"}, "label": "the taller of the two giraffes eating leaves"}]}
{"id": 525171, "image": "COCO_train2014_000000525171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a black and white dress playing tennis\"."}], "task": "refering", "answer_template": "The \"a girl in a black and white dress playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 213, 214, 215, 235, 236, 238, 239, 257, 258, 262, 280, 281, 285], "bbox": [0.19042187500000002, 0.24897196261682245, 0.43937500000000007, 0.8374766355140187], "iscrowd": 0, "area": 12341.429049999999, "rle": {"size": [428, 640], "counts": "gZc12W=4M4L3L4M2N2O1N2O1O1O100O1O1O1N10O100OOO1O2M2O11PMYO\\Ih0d6ZOXIh0h6ZOUIg0j6\\OTIe0k6]ORId0o6]OoHc0R7FdH;\\7GaH:`7F_H9b7H^H7b7I^H7b7J]H5d7K]H4c7M\\H3d7M]H1c7N_H2a7LbH3]7LeH3]7JfH5[7JeH6\\7JdH5]7KZH=g7DnGe0S8[OeGl0\\8TO`Go0a8QO\\GQ1o7^N]Ga0`0T1S8\\N]G`0<W1W8ZN]G?8Z1[8XN\\G?6[1]8XN]G=2^1_8YN^G9O`1U8gNlGIKc1X8fNmGGHe1[8fNlGEEh1^8eNmGCAk1a8eNlGP2U8QNiGo1W8ZN]Gi1d8]12O1O2M2O1N2001O1O1N2O1O1O2M2O1N2O1N2O1N200000000000000000000000000001O12OmKcGU3\\8gLhGZ3W8bLmG_3R8^LPHc3Q8XLSHh3o7RLUHn3e81N1000\\MkFQ1T9iNWGR1h8hNeGR1[8hNPHT1Q8cN[HX1g7_NdH]1b7VNhHh1a9N2N2N2N2N3M3M2N10O01O1O01O000O1O1O2O2M2ZDgNT;h11O3M5K6J6J5K5K2N3M2N2N2N2N2N2N1O1O2N1O1M3Mkfe4"}, "label": "a girl in a black and white dress playing tennis"}]}
{"id": 525171, "image": "COCO_train2014_000000525171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing white and black getting ready to swing at the tennis ball\"."}], "task": "refering", "answer_template": "The \"a girl wearing white and black getting ready to swing at the tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 213, 214, 215, 235, 236, 238, 239, 257, 258, 262, 280, 281, 285], "bbox": [0.19042187500000002, 0.24897196261682245, 0.43937500000000007, 0.8374766355140187], "iscrowd": 0, "area": 12341.429049999999, "rle": {"size": [428, 640], "counts": "gZc12W=4M4L3L4M2N2O1N2O1O1O100O1O1O1N10O100OOO1O2M2O11PMYO\\Ih0d6ZOXIh0h6ZOUIg0j6\\OTIe0k6]ORId0o6]OoHc0R7FdH;\\7GaH:`7F_H9b7H^H7b7I^H7b7J]H5d7K]H4c7M\\H3d7M]H1c7N_H2a7LbH3]7LeH3]7JfH5[7JeH6\\7JdH5]7KZH=g7DnGe0S8[OeGl0\\8TO`Go0a8QO\\GQ1o7^N]Ga0`0T1S8\\N]G`0<W1W8ZN]G?8Z1[8XN\\G?6[1]8XN]G=2^1_8YN^G9O`1U8gNlGIKc1X8fNmGGHe1[8fNlGEEh1^8eNmGCAk1a8eNlGP2U8QNiGo1W8ZN]Gi1d8]12O1O2M2O1N2001O1O1N2O1O1O2M2O1N2O1N2O1N200000000000000000000000000001O12OmKcGU3\\8gLhGZ3W8bLmG_3R8^LPHc3Q8XLSHh3o7RLUHn3e81N1000\\MkFQ1T9iNWGR1h8hNeGR1[8hNPHT1Q8cN[HX1g7_NdH]1b7VNhHh1a9N2N2N2N2N3M3M2N10O01O1O01O000O1O1O2O2M2ZDgNT;h11O3M5K6J6J5K5K2N3M2N2N2N2N2N2N1O1O2N1O1M3Mkfe4"}, "label": "a girl wearing white and black getting ready to swing at the tennis ball"}]}
{"id": 287608, "image": "COCO_train2014_000000287608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller of the two teddybears\"."}], "task": "refering", "answer_template": "The \"the smaller of the two teddybears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 70, 71, 72, 73, 74, 75, 82, 83, 84, 85, 86, 87, 88, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 136, 137, 138, 139, 150, 151, 152, 153, 154, 163, 164, 165, 166], "bbox": [0.30058666666666667, 0.233, 0.87216, 0.7147], "iscrowd": 0, "area": 27329.883149999998, "rle": {"size": [500, 375], "counts": "i[g12a?2M2O2M3M2O2M3SC@c:a0YECe:`0VEDi:>REFm:<nDHQ;9kDKT;7gDNW;4eDO[;1eDOZ;3dDN[;4cDM\\;4dDL[;6cDL\\;5bDL];5bDL];6bDJ];8aDI_;7`DJ_;8`DH_;:_DH_;9`DH_;:_DGa;:^DFa;<]DEb;<]DEb;=]DDb;=\\DCd;>[DCd;?ZDAf;a0YD_Og;b0WD^Oi;c0VD^Oi;d0VD[Oj;g0TDZOk;g0TDYOm;W2O000001O000000001O0000000O2O000000001O00001O000010^DaLn:`3PEdLm:]3PEhLn:X3oDmLn:T3oDQMn:P3oDUMo:k2oDYMn:i3M3N2M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3N2Mg0ZO000000000001O0000001N101O1O1O001O1O1O1O2N1O1O1hIXGb5i8YJ\\Gf5f8TJ`Gi5b8RJdGl5S9O2N1O2N2N2N2M3oJnE\\4T:\\KQFe4^:O001O1N2O1N2O0011O:F:F7I2M200O2O001N[N[F]Me9b2_F\\M_9d2cF[M]9d2dF]MZ9b2hF^MX9a2iF_MV9a2kF_MU9`2lF`MS9_2PG`MP90\\FQ1e0oNn81^Fn0e0QOm80_Fm0f0SOj81aFj0f0UOi80cFi0e0WOg81dFg0f0XOf80fFe0g0ZOc81XG26Mb80aGIO7_81fGAN>\\80kG\\OKd0Z80PHTOIl0W8OUHoNGQ1S81ZHhNEW1Q80nI0R60nI0R6OoI1Q6OoI1P6ORJ0n50RJ0n5OSJ1m5NTJ2k5NVJ1k5NVJ1l5MVJ2j5MWJ2j5NVJ1k5NUJ2m5LSJ4n5KSJ5n5IRJ7o5HQJ8Q6FPJ9R6FmI;T6ClI=V6AkI>X6_OhIa0Z6\\OhId0Z6YOfIg0\\6WOdIi0_6SOcIl0`6QO`IP1a6mN`IS1a:O00002N5K5K5J6K4L5KjSg0"}, "label": "the smaller of the two teddybears"}]}
{"id": 287608, "image": "COCO_train2014_000000287608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small stuffed bear in front of a larger one\"."}], "task": "refering", "answer_template": "The \"a small stuffed bear in front of a larger one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 70, 71, 72, 73, 74, 75, 82, 83, 84, 85, 86, 87, 88, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 136, 137, 138, 139, 150, 151, 152, 153, 154, 163, 164, 165, 166], "bbox": [0.30058666666666667, 0.233, 0.87216, 0.7147], "iscrowd": 0, "area": 27329.883149999998, "rle": {"size": [500, 375], "counts": "i[g12a?2M2O2M3M2O2M3SC@c:a0YECe:`0VEDi:>REFm:<nDHQ;9kDKT;7gDNW;4eDO[;1eDOZ;3dDN[;4cDM\\;4dDL[;6cDL\\;5bDL];5bDL];6bDJ];8aDI_;7`DJ_;8`DH_;:_DH_;9`DH_;:_DGa;:^DFa;<]DEb;<]DEb;=]DDb;=\\DCd;>[DCd;?ZDAf;a0YD_Og;b0WD^Oi;c0VD^Oi;d0VD[Oj;g0TDZOk;g0TDYOm;W2O000001O000000001O0000000O2O000000001O00001O000010^DaLn:`3PEdLm:]3PEhLn:X3oDmLn:T3oDQMn:P3oDUMo:k2oDYMn:i3M3N2M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3N2Mg0ZO000000000001O0000001N101O1O1O001O1O1O1O2N1O1O1hIXGb5i8YJ\\Gf5f8TJ`Gi5b8RJdGl5S9O2N1O2N2N2N2M3oJnE\\4T:\\KQFe4^:O001O1N2O1N2O0011O:F:F7I2M200O2O001N[N[F]Me9b2_F\\M_9d2cF[M]9d2dF]MZ9b2hF^MX9a2iF_MV9a2kF_MU9`2lF`MS9_2PG`MP90\\FQ1e0oNn81^Fn0e0QOm80_Fm0f0SOj81aFj0f0UOi80cFi0e0WOg81dFg0f0XOf80fFe0g0ZOc81XG26Mb80aGIO7_81fGAN>\\80kG\\OKd0Z80PHTOIl0W8OUHoNGQ1S81ZHhNEW1Q80nI0R60nI0R6OoI1Q6OoI1P6ORJ0n50RJ0n5OSJ1m5NTJ2k5NVJ1k5NVJ1l5MVJ2j5MWJ2j5NVJ1k5NUJ2m5LSJ4n5KSJ5n5IRJ7o5HQJ8Q6FPJ9R6FmI;T6ClI=V6AkI>X6_OhIa0Z6\\OhId0Z6YOfIg0\\6WOdIi0_6SOcIl0`6QO`IP1a6mN`IS1a:O00002N5K5K5J6K4L5KjSg0"}, "label": "a small stuffed bear in front of a larger one"}]}
{"id": 287608, "image": "COCO_train2014_000000287608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown big teady bear side of small one\"."}], "task": "refering", "answer_template": "The \"a brown big teady bear side of small one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 27, 28, 29, 30, 31, 32, 39, 40, 41, 42, 43, 44, 45, 46, 52, 53, 54, 55, 56, 57, 58, 66, 67, 68, 69, 70, 80, 81, 82, 94, 95, 106, 107, 108, 109, 119, 120, 121, 122, 134, 135, 136, 145, 146, 147, 148, 149, 150, 161, 162], "bbox": [0.060186666666666666, 0.07392, 0.5568, 0.6924399999999999], "iscrowd": 0, "area": 26398.3613, "rle": {"size": [500, 375], "counts": "^Z;h0l>3L3N2M3N2M4L3N2M3N2M4M2M3M3N2M2N2N2N2N2N2N2O1N2N2N2N2O1O10O01O100O1O100O1O100O1O100O1O1O1N1O2N2O1fGiLT4Y3hKjLW4X3fKjLY4W3dKmLZ4T3cKoL]4Q3hIhLJ8\\6R3_IPM5N[6S3ZIVM;GY6T3[IVM<FX6V3YIVM`0DV6V3YIXM`0BV6W3YIXMb0AS6Y3XIXMe0_OR6Z3XIYMf0\\OQ6\\3WIZMh0[OP6[3WI[Mj0YOn5^3UI\\Ml0WOm5^3VI\\Mn0UOk5`3VI\\Mo0UOj5`3TI^MR1QOi5b3TI^MS1QOg5b3TI_MV1nNe5e3SI_MW1lNf5e3QIaMZ1jNd5e3QIcMZ1hNd5g3PIcM\\1eNd5h3nHfM]1bNe5h3mHgM_1aNc5h3lHjM`1^Nc5j3kHjMb1[Nc5k3jHlMb1YNd5Q6[JoIe5R6ZJnIe5T6ZJlIf5U6YJkIg5U6YJlIf5U6YJkIg5V6XJjIg5X6XJhIh5h3mHZNZ1nMi5g3oH[NW1nMj5f3QI\\NT1nMj5e3TI^NQ1mMk5d3VI^No0nMk5c3XI_Nl0oMk5a3[I`Ni0oMk5a3]IaNg0nMl5`3_I[NEhMo0=m5_3aIUNIPNh0<n5^3bIQNMVNb0;o5]3dIkM1^N;:o5]3gIeM4dN6:o5\\3iIaM6jN1:o5Y3lIbM5lNO9P6X3nIaM4oNM8P6X3QJ_M3ROK7Q6W3RJ_M3TOI6R6V3TJ^M2WOG5S6U3VJ]M1YOF5R6U3XJ\\M1[OD4S6T3ZJ\\MO]OC3T6S3\\J[MN@A3T6Q3_JZMMC_O2S6R3bJXMME]O1S6R3eJVMLHZO1T6R3gJSMLJWO3T6Q3jJQMLLTO3V6o2lJPMKOQO3W6o2nJnLJ1oN3Y6m2oJnLJ3lN4Y6l2RKkLJ6iN4[6j2TKjLI9fN4]6h2VKiLH;dN5]6h2WKgLI=aN5_6f2YKfLH`0^N5`6f2[KcLI`1l4l1\\KdLH`1k4l1_KbLGc1i4k1aK`LFf1h4j1cK_LEh1h4i1ZMXNf2g1ZMZNe2f1\\MZNd2f1[M[Nd2e1\\M\\Nd2d1[M]Nd2c1\\M^Nd2b1[M`Nc2`1]MaNc2^1]McNb2^1]McNc2\\1]MeNd2Z1]MeNc2Z1^MfNc2X1_MgNb2X1_MgNa2X1`MhNa2W1`MhNa2V1`MkN_2T1cMkN^2T1cMkN^2S1cMmN]2S1dMlN]2R1dMnN]2P1eMoN[2Q1fMnN[2P1fMPO[2o0fMPOZ2o0hMROW2m0iMTOW2j0kMVOT2i0mMXOS2g0nMYOR2e0PN[OP2c0QN^On1a0TN^Om1`0VN^Ok1`0XN^Oh1b0ZN\\Og1b0\\N[Og1b0\\N\\Of1a0^N\\Oe1`0^N^Od1?_N_Od1=_NAc1<`NBc1:`NCc1;_NCd19_NEc18`NEd17`ND\\2AgM:m8L4L4L4L4Ldg`2"}, "label": "a brown big teady bear side of small one"}]}
{"id": 287608, "image": "COCO_train2014_000000287608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger of two small teddy bears on a canon camera\"."}], "task": "refering", "answer_template": "The \"the larger of two small teddy bears on a canon camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 27, 28, 29, 30, 31, 32, 39, 40, 41, 42, 43, 44, 45, 46, 52, 53, 54, 55, 56, 57, 58, 66, 67, 68, 69, 70, 80, 81, 82, 94, 95, 106, 107, 108, 109, 119, 120, 121, 122, 134, 135, 136, 145, 146, 147, 148, 149, 150, 161, 162], "bbox": [0.060186666666666666, 0.07392, 0.5568, 0.6924399999999999], "iscrowd": 0, "area": 26398.3613, "rle": {"size": [500, 375], "counts": "^Z;h0l>3L3N2M3N2M4L3N2M3N2M4M2M3M3N2M2N2N2N2N2N2N2O1N2N2N2N2O1O10O01O100O1O100O1O100O1O100O1O1O1N1O2N2O1fGiLT4Y3hKjLW4X3fKjLY4W3dKmLZ4T3cKoL]4Q3hIhLJ8\\6R3_IPM5N[6S3ZIVM;GY6T3[IVM<FX6V3YIVM`0DV6V3YIXM`0BV6W3YIXMb0AS6Y3XIXMe0_OR6Z3XIYMf0\\OQ6\\3WIZMh0[OP6[3WI[Mj0YOn5^3UI\\Ml0WOm5^3VI\\Mn0UOk5`3VI\\Mo0UOj5`3TI^MR1QOi5b3TI^MS1QOg5b3TI_MV1nNe5e3SI_MW1lNf5e3QIaMZ1jNd5e3QIcMZ1hNd5g3PIcM\\1eNd5h3nHfM]1bNe5h3mHgM_1aNc5h3lHjM`1^Nc5j3kHjMb1[Nc5k3jHlMb1YNd5Q6[JoIe5R6ZJnIe5T6ZJlIf5U6YJkIg5U6YJlIf5U6YJkIg5V6XJjIg5X6XJhIh5h3mHZNZ1nMi5g3oH[NW1nMj5f3QI\\NT1nMj5e3TI^NQ1mMk5d3VI^No0nMk5c3XI_Nl0oMk5a3[I`Ni0oMk5a3]IaNg0nMl5`3_I[NEhMo0=m5_3aIUNIPNh0<n5^3bIQNMVNb0;o5]3dIkM1^N;:o5]3gIeM4dN6:o5\\3iIaM6jN1:o5Y3lIbM5lNO9P6X3nIaM4oNM8P6X3QJ_M3ROK7Q6W3RJ_M3TOI6R6V3TJ^M2WOG5S6U3VJ]M1YOF5R6U3XJ\\M1[OD4S6T3ZJ\\MO]OC3T6S3\\J[MN@A3T6Q3_JZMMC_O2S6R3bJXMME]O1S6R3eJVMLHZO1T6R3gJSMLJWO3T6Q3jJQMLLTO3V6o2lJPMKOQO3W6o2nJnLJ1oN3Y6m2oJnLJ3lN4Y6l2RKkLJ6iN4[6j2TKjLI9fN4]6h2VKiLH;dN5]6h2WKgLI=aN5_6f2YKfLH`0^N5`6f2[KcLI`1l4l1\\KdLH`1k4l1_KbLGc1i4k1aK`LFf1h4j1cK_LEh1h4i1ZMXNf2g1ZMZNe2f1\\MZNd2f1[M[Nd2e1\\M\\Nd2d1[M]Nd2c1\\M^Nd2b1[M`Nc2`1]MaNc2^1]McNb2^1]McNc2\\1]MeNd2Z1]MeNc2Z1^MfNc2X1_MgNb2X1_MgNa2X1`MhNa2W1`MhNa2V1`MkN_2T1cMkN^2T1cMkN^2S1cMmN]2S1dMlN]2R1dMnN]2P1eMoN[2Q1fMnN[2P1fMPO[2o0fMPOZ2o0hMROW2m0iMTOW2j0kMVOT2i0mMXOS2g0nMYOR2e0PN[OP2c0QN^On1a0TN^Om1`0VN^Ok1`0XN^Oh1b0ZN\\Og1b0\\N[Og1b0\\N\\Of1a0^N\\Oe1`0^N^Od1?_N_Od1=_NAc1<`NBc1:`NCc1;_NCd19_NEc18`NEd17`ND\\2AgM:m8L4L4L4L4Ldg`2"}, "label": "the larger of two small teddy bears on a canon camera"}]}
{"id": 500603, "image": "COCO_train2014_000000500603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bat of the man on the left\"."}], "task": "refering", "answer_template": "The \"the bat of the man on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 237, 238, 259, 282, 304, 326, 327, 348, 349, 371, 372], "bbox": [0.142984375, 0.5393320235756385, 0.36459375000000005, 0.9056188605108054], "iscrowd": 0, "area": 2649.3610499999995, "rle": {"size": [509, 640], "counts": "`U^11j?3N1O1N3N1N3N1O1N3N1N2O2N1N2O2N1N3N1N2O2N0O10O10O010O0100O010O10O010O0100O010O010O10O0100O010O010O10O0100O011N2O1N2O0O2O1N2O1N101N3N2M4M2M4L^^54TaJ:J61O010O1O10O01O100O1O011N2N2N2O1N1O2O1N2N2O0O2NP_25i`M30O10O100O10000O10O010000O100O10O10O100001O1O001N101M[QZ6"}, "label": "the bat of the man on the left"}]}
{"id": 500603, "image": "COCO_train2014_000000500603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball bat of the man on the left\"."}], "task": "refering", "answer_template": "The \"the baseball bat of the man on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 237, 238, 259, 282, 304, 326, 327, 348, 349, 371, 372], "bbox": [0.142984375, 0.5393320235756385, 0.36459375000000005, 0.9056188605108054], "iscrowd": 0, "area": 2649.3610499999995, "rle": {"size": [509, 640], "counts": "`U^11j?3N1O1N3N1N3N1O1N3N1N2O2N1N2O2N1N3N1N2O2N0O10O10O010O0100O010O10O010O0100O010O010O10O0100O010O010O10O0100O011N2O1N2O0O2O1N2O1N101N3N2M4M2M4L^^54TaJ:J61O010O1O10O01O100O1O011N2N2N2O1N1O2O1N2N2O0O2NP_25i`M30O10O100O10000O10O010000O100O10O10O100001O1O001N101M[QZ6"}, "label": "the baseball bat of the man on the left"}]}
{"id": 500603, "image": "COCO_train2014_000000500603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player kneeling down on his right knee\"."}], "task": "refering", "answer_template": "The \"a baseball player kneeling down on his right knee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 75, 76, 77, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 309, 310, 329, 330, 332, 333, 355, 356, 357], "bbox": [0.18270312500000002, 0.13062868369351668, 0.526625, 0.8828880157170923], "iscrowd": 0, "area": 42924.513499999994, "rle": {"size": [509, 640], "counts": "T\\j17a?=D;E4K5L5J5L4L4M4K4M3M3M4K4M3N2O2N1O1N2O2N1O1O1O2N3M3kFiL_5Z3]JiLa5Z3ZJjLd5Y3XJjLf5Y3UJkLi5X3RJmLk5Y3nIjLP6a3bIdL\\6h3hHhLX7d3bGWM^8S501O001O001O010O002N3M2O1N1O2O1N2NO2LfN_H_J_7_5QIUJg5n6c14M2M34M3L4L4M3L4L5K5L4L4K5L4L5K5K5K2N2N2N1O2M3N1O2N2N1O2N1O1O1O2N1O1O1O1O1O0000000000000001O0000001O00000O2N1O3M4L3M4XGYJS7j5jHWJU7m5gHUJW7P6dHQJ[7Z7oLTInLm6R3^IbLd6]3hIWLX6i3mIQLU6n3nInKS6R4oIkKS6T4PJgKR6Z4oI`KU6a4nIXKU6i4XJeJm5\\5Y21O1O1O2O0O1O2N2O1N2N3M2N2O1O1O00001O00001O000O200O10O10O10O0100OiKTG]1l8^NXGc1h8YN[Gi1c8UN`Gl1_8QNcGQ2[8nMfGT2Y8iMjGX2U8eMmG]2S8`MnGb2Q8[MRHf2m7WMUHk2j7RMYHQ3d7mL_HU3^7hLeHZ3Y7cLjH_3T7^LoHc3P7[LRIf3m6WLVIi3m6QLVIo3\\91O100O4M4K31N4M4J5I6J7I7I7I6J7I8Hc0]O4L3M2N3M2N3M2N3M2M3N3M2N3M3M3M3L3N3M3MQVf4"}, "label": "a baseball player kneeling down on his right knee"}]}
{"id": 500603, "image": "COCO_train2014_000000500603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting in the left side of the image\"."}], "task": "refering", "answer_template": "The \"man sitting in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 75, 76, 77, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 309, 310, 329, 330, 332, 333, 355, 356, 357], "bbox": [0.18270312500000002, 0.13062868369351668, 0.526625, 0.8828880157170923], "iscrowd": 0, "area": 42924.513499999994, "rle": {"size": [509, 640], "counts": "T\\j17a?=D;E4K5L5J5L4L4M4K4M3M3M4K4M3N2O2N1O1N2O2N1O1O1O2N3M3kFiL_5Z3]JiLa5Z3ZJjLd5Y3XJjLf5Y3UJkLi5X3RJmLk5Y3nIjLP6a3bIdL\\6h3hHhLX7d3bGWM^8S501O001O001O010O002N3M2O1N1O2O1N2NO2LfN_H_J_7_5QIUJg5n6c14M2M34M3L4L4M3L4L5K5L4L4K5L4L5K5K5K2N2N2N1O2M3N1O2N2N1O2N1O1O1O2N1O1O1O1O1O0000000000000001O0000001O00000O2N1O3M4L3M4XGYJS7j5jHWJU7m5gHUJW7P6dHQJ[7Z7oLTInLm6R3^IbLd6]3hIWLX6i3mIQLU6n3nInKS6R4oIkKS6T4PJgKR6Z4oI`KU6a4nIXKU6i4XJeJm5\\5Y21O1O1O2O0O1O2N2O1N2N3M2N2O1O1O00001O00001O000O200O10O10O10O0100OiKTG]1l8^NXGc1h8YN[Gi1c8UN`Gl1_8QNcGQ2[8nMfGT2Y8iMjGX2U8eMmG]2S8`MnGb2Q8[MRHf2m7WMUHk2j7RMYHQ3d7mL_HU3^7hLeHZ3Y7cLjH_3T7^LoHc3P7[LRIf3m6WLVIi3m6QLVIo3\\91O100O4M4K31N4M4J5I6J7I7I7I6J7I8Hc0]O4L3M2N3M2N3M2N3M2M3N3M2N3M3M3M3L3N3M3MQVf4"}, "label": "man sitting in the left side of the image"}]}
{"id": 500603, "image": "COCO_train2014_000000500603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bat held in the mans left hand\"."}], "task": "refering", "answer_template": "The \"the bat held in the mans left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 245, 246, 269, 270, 293, 294, 317, 318, 341, 342, 343, 365], "bbox": [0.6480937499999999, 0.536679764243615, 0.9227812499999999, 0.8436738703339882], "iscrowd": 0, "area": 3772.249700000009, "rle": {"size": [509, 640], "counts": "ha^61l?1N2O1N2N2O1N20001O0O10000000000000000O10000000000000000O10000000000000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O10000000000000000O101O00000000000000000000000O0100000000000000000000000O100000000000000000000000000O100000O1000000O1O1O1O1O1O001O1O1O1N2O1O1O1N2O1N2N2O1N1O2O1NW^h0"}, "label": "the bat held in the mans left hand"}]}
{"id": 500603, "image": "COCO_train2014_000000500603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bat held by the man on the right\"."}], "task": "refering", "answer_template": "The \"the bat held by the man on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 245, 246, 269, 270, 293, 294, 317, 318, 341, 342, 343, 365], "bbox": [0.6480937499999999, 0.536679764243615, 0.9227812499999999, 0.8436738703339882], "iscrowd": 0, "area": 3772.249700000009, "rle": {"size": [509, 640], "counts": "ha^61l?1N2O1N2N2O1N20001O0O10000000000000000O10000000000000000O10000000000000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O10000000000000000O101O00000000000000000000000O0100000000000000000000000O100000000000000000000000000O100000O1000000O1O1O1O1O1O001O1O1O1N2O1O1O1N2O1N2N2O1N1O2O1NW^h0"}, "label": "the bat held by the man on the right"}]}
{"id": 516990, "image": "COCO_train2014_000000516990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt\"."}], "task": "refering", "answer_template": "The \"man in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 278], "bbox": [0.0, 0.006744730679156909, 0.21140625000000002, 0.8157377049180328], "iscrowd": 0, "area": 27803.612250000002, "rle": {"size": [427, 640], "counts": "3Z1Q<ObDd7o1[HRNe7n1ZHSNe7n1ZHSNf7n1XHTNg7m1WHUNg7n1VHTNi7m1UHUNi7n1THTNk7n1RHTNm7m1^GhN`8S3O1N2O001N2O1O1N2O1O1O1O1O1N0100000000000O10000000O10O10O1000000O01000O1000000O2O000O2O000O2O00001N10001N10001O000O2O001O00001O0O100000O100O010O10000O10O0100H8VOj0WOi0VOi0XOi0VOj0B>L4M3M3L3N3L5L3L5L3L5L3L6M2N3M3M2M4M3M2N3M3L3N3M3M2M3N2N2N2N2M3N2N2N2N2M3N3MZ_b6"}, "label": "man in white shirt"}]}
{"id": 516990, "image": "COCO_train2014_000000516990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the white shirt half off camera\"."}], "task": "refering", "answer_template": "The \"the man with the white shirt half off camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 278], "bbox": [0.0, 0.006744730679156909, 0.21140625000000002, 0.8157377049180328], "iscrowd": 0, "area": 27803.612250000002, "rle": {"size": [427, 640], "counts": "3Z1Q<ObDd7o1[HRNe7n1ZHSNe7n1ZHSNf7n1XHTNg7m1WHUNg7n1VHTNi7m1UHUNi7n1THTNk7n1RHTNm7m1^GhN`8S3O1N2O001N2O1O1N2O1O1O1O1O1N0100000000000O10000000O10O10O1000000O01000O1000000O2O000O2O000O2O00001N10001N10001O000O2O001O00001O0O100000O100O010O10000O10O0100H8VOj0WOi0VOi0XOi0VOj0B>L4M3M3L3N3L5L3L5L3L5L3L6M2N3M3M2M4M3M2N3M3L3N3M3M2M3N2N2N2N2M3N2N2N2N2M3N3MZ_b6"}, "label": "the man with the white shirt half off camera"}]}
{"id": 516990, "image": "COCO_train2014_000000516990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chef looking down intently\"."}], "task": "refering", "answer_template": "The \"chef looking down intently\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 34, 35, 36, 37, 56, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.290671875, 0.033302107728337235, 0.8358593750000001, 0.8549648711943795], "iscrowd": 0, "area": 63949.75285, "rle": {"size": [427, 640], "counts": "Uj]2g0d<1N1O2O1N1O2O0O2N2N101N1O2O1N1O2O0O2N2O001N101O1N101O001N2O001N101O0O2O000O10O10O100000O010000O10O0100O1O1O001O100O001O001O1O100O1O1O001O1O100O1O001O1O100O1O001jF^Mb7c2]H^Mb7c2]H_Ma7b2_H^M_7c2aH^M^7c2aH_M\\7c2cH^M\\7c2cH_M[7b2dH_MZ7b2eH`MZ7a2dHbMZ7_2eHbMY7`2eHcMY7^2eHdMY7^2gHbMX7_2gHcMW7^2hHcMV7`2hHaMW7`2iHaMT7a2kH`MT7a2kHaMS7a2kH`MS7b2mH^MQ7d2nH^Mo6d2PI]Mo6e2oH]Mn6e2RI[Ml6g2SIZMk6i2SIYMj6i2UIXMi6j2WIWMe6l2ZIUMd6n2ZISMd6Q3YIQMQ4hNoMZ4oMnLm3POnMV4SNlLi3UOPNR4UNjLf3[OQNo3VNgLd3BQNj3ZNeLb3EPNi3\\NcLb3IlM^5R2eJjM]5T2gJhM[5V2jJeMX5Z2kJaMX5]2k2O2O1N2N2N2O1O10O01O100O1O100O00100O1O1O100O1O010O10000O100O01000O100O10000O010O10000O100O01000O1001O000O10001O0000000O2O000000001O0000001N101O1O001O001O001O001O001O1O001O001O001O0QFkLg9W3YFkLd9U3\\FmLb9T3]FnLa9R3_FPM_9Q3`FRM]9o2bFSM\\9m2dFUMZ9l2eFVMY9j2gFSN\\8n1cGcNl7]1THgNh7a3M4L3M3M4L3M4L2N3M3M2N3M1O2N1O2N1O2N1O2N1N2O1O1O1O1O1O1O1O2N1O00001O001O0O101O001O00001O001O00001O001O00001O002N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O3M2N2N2N3M2N2N3M2N4L4L4L4L3M2N2M3N2M3N2Mc0^O4K5L5J5K5L4K4L4L3L4L5K4M3Lj^[1"}, "label": "chef looking down intently"}]}
{"id": 516990, "image": "COCO_train2014_000000516990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt and a black and white striped apron stirring something in a metal container\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt and a black and white striped apron stirring something in a metal container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 34, 35, 36, 37, 56, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.290671875, 0.033302107728337235, 0.8358593750000001, 0.8549648711943795], "iscrowd": 0, "area": 63949.75285, "rle": {"size": [427, 640], "counts": "Uj]2g0d<1N1O2O1N1O2O0O2N2N101N1O2O1N1O2O0O2N2O001N101O1N101O001N2O001N101O0O2O000O10O10O100000O010000O10O0100O1O1O001O100O001O001O1O100O1O1O001O1O100O1O001O1O100O1O001jF^Mb7c2]H^Mb7c2]H_Ma7b2_H^M_7c2aH^M^7c2aH_M\\7c2cH^M\\7c2cH_M[7b2dH_MZ7b2eH`MZ7a2dHbMZ7_2eHbMY7`2eHcMY7^2eHdMY7^2gHbMX7_2gHcMW7^2hHcMV7`2hHaMW7`2iHaMT7a2kH`MT7a2kHaMS7a2kH`MS7b2mH^MQ7d2nH^Mo6d2PI]Mo6e2oH]Mn6e2RI[Ml6g2SIZMk6i2SIYMj6i2UIXMi6j2WIWMe6l2ZIUMd6n2ZISMd6Q3YIQMQ4hNoMZ4oMnLm3POnMV4SNlLi3UOPNR4UNjLf3[OQNo3VNgLd3BQNj3ZNeLb3EPNi3\\NcLb3IlM^5R2eJjM]5T2gJhM[5V2jJeMX5Z2kJaMX5]2k2O2O1N2N2N2O1O10O01O100O1O100O00100O1O1O100O1O010O10000O100O01000O100O10000O010O10000O100O01000O1001O000O10001O0000000O2O000000001O0000001N101O1O001O001O001O001O001O1O001O001O001O0QFkLg9W3YFkLd9U3\\FmLb9T3]FnLa9R3_FPM_9Q3`FRM]9o2bFSM\\9m2dFUMZ9l2eFVMY9j2gFSN\\8n1cGcNl7]1THgNh7a3M4L3M3M4L3M4L2N3M3M2N3M1O2N1O2N1O2N1O2N1N2O1O1O1O1O1O1O1O2N1O00001O001O0O101O001O00001O001O00001O001O00001O002N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O3M2N2N2N3M2N2N3M2N4L4L4L4L3M2N2M3N2M3N2Mc0^O4K5L5J5K5L4K4L4L3L4L5K4M3Lj^[1"}, "label": "a man wearing a black shirt and a black and white striped apron stirring something in a metal container"}]}
{"id": 197503, "image": "COCO_train2014_000000197503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young kid closest to projector\"."}], "task": "refering", "answer_template": "The \"young kid closest to projector\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.24128124999999997, 0.41429166666666667, 0.48534375, 0.8596041666666667], "iscrowd": 0, "area": 17305.9517, "rle": {"size": [480, 640], "counts": "]^X21i>7I7O2M2O1O2N1O1O1O1O1O1O2N1O2N101N2N1hBKc;]OmC9;?f;VOQD<5c0h;oNVDV2f;hM]DZ2`;eMaD^2\\;`MgDb2V;[MnDg2d;2O?@3N8G5L3N3K2_NTLiGm3R8[LjGf3Q8aLlG`3o7hLnGX3m7nLQHS3j7SMUHl2g7ZMWHg2h7[MWHe2h7^MmGk2R8WMkGk2T8WMjGj2V8VMhGl2W8VMfGl2Z8m100O1000000O100000000000O1000000O1000000O1000001N10001O0FeGnI]8m5=L5J5I8N101N101O0O2O001_L[Fd1f9WN^Fj1c9QNaFn1a9PN`FP2a9mMaFS2`9jMbFU2`9hMaFY2b9bM`F^2b9^M`Fa2f9XM\\Fh2n:00O2O0O2O1O1N2O1O1N2O2M2O1O1N2O1O1N2O0O2@cC[N^<^1jC`NX<]1kC`NX<^1kC_NX<^1jC`NX<^1g0N1O001O1O1O1O1O1O1O00001O00001O00001N1O100N3M2M3N3M2O1O2M3M3M4LYmi4"}, "label": "young kid closest to projector"}]}
{"id": 197503, "image": "COCO_train2014_000000197503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little boy with the orange shirt\"."}], "task": "refering", "answer_template": "The \"the little boy with the orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.24128124999999997, 0.41429166666666667, 0.48534375, 0.8596041666666667], "iscrowd": 0, "area": 17305.9517, "rle": {"size": [480, 640], "counts": "]^X21i>7I7O2M2O1O2N1O1O1O1O1O1O2N1O2N101N2N1hBKc;]OmC9;?f;VOQD<5c0h;oNVDV2f;hM]DZ2`;eMaD^2\\;`MgDb2V;[MnDg2d;2O?@3N8G5L3N3K2_NTLiGm3R8[LjGf3Q8aLlG`3o7hLnGX3m7nLQHS3j7SMUHl2g7ZMWHg2h7[MWHe2h7^MmGk2R8WMkGk2T8WMjGj2V8VMhGl2W8VMfGl2Z8m100O1000000O100000000000O1000000O1000000O1000001N10001O0FeGnI]8m5=L5J5I8N101N101O0O2O001_L[Fd1f9WN^Fj1c9QNaFn1a9PN`FP2a9mMaFS2`9jMbFU2`9hMaFY2b9bM`F^2b9^M`Fa2f9XM\\Fh2n:00O2O0O2O1O1N2O1O1N2O2M2O1O1N2O1O1N2O0O2@cC[N^<^1jC`NX<]1kC`NX<^1kC_NX<^1jC`NX<^1g0N1O001O1O1O1O1O1O1O00001O00001O00001N1O100N3M2M3N3M2O1O2M3M3M4LYmi4"}, "label": "the little boy with the orange shirt"}]}
{"id": 197503, "image": "COCO_train2014_000000197503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark colored boy standing while holding a remote control\"."}], "task": "refering", "answer_template": "The \"a dark colored boy standing while holding a remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 314, 337, 338], "bbox": [0.4848125, 0.12685416666666666, 0.748234375, 0.8463958333333333], "iscrowd": 0, "area": 32382.448400000016, "rle": {"size": [480, 640], "counts": "_`a47g0OP=Y1D6K6K5L4M2M4L4N1O2N2N2N2N2N2M2O2N2N2bFoLj6R3kH]MP7d2mH_MR7c2kH`MR7b2kHbMR7`2kHcMS7_2eHiMZ7X2\\HRNb7P2TH[Ni7g1mGcNS8]1iGgNV8Z1gGiN_6iMjJ]3eNnN^6jMiJY3fNPO`6kMgJU3fNSOa6mMeJP3hNVOa6PNcJe5[5aJ`J`5_5eJ]J[5a5kJZJU5e5RKVJl4k5UKTJj4l5YKSJd4n5_KPJ`4Q6bKnI\\4S6fKlIY4T6hKkIY4T6iKkIW4T6kKjIV4U6kKkIU4U6lKjIT4U6mKjIT4U6nKjIR4V6nKhIU4V6[2O100O1000000000O2O00000000001O000000lI]Gk5P9L1O01O01O00001O00010O0000001O03M3M4L4L4L5L4K8H:F4L3M2O1N;E010O010O010O01O010O010O010O010O010O100O100O10001O2M4M4L:E2O00fKnHe0R7WOUIf0j6VO^Ig0b6UOeIZO5oMW6f2hIPO:WNP6g2kIfN`0`Nf5h2WJUN=oN^5j2dJYMa0Im4l2\\MRMe2m2h4N2N2N1O3N4K4L4L4L4M2M3M2N3M3M3N2M3M4L4L4L4M3L4K5K6I6KTX[2"}, "label": "a dark colored boy standing while holding a remote control"}]}
{"id": 197503, "image": "COCO_train2014_000000197503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the boy on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 314, 337, 338], "bbox": [0.4848125, 0.12685416666666666, 0.748234375, 0.8463958333333333], "iscrowd": 0, "area": 32382.448400000016, "rle": {"size": [480, 640], "counts": "_`a47g0OP=Y1D6K6K5L4M2M4L4N1O2N2N2N2N2N2M2O2N2N2bFoLj6R3kH]MP7d2mH_MR7c2kH`MR7b2kHbMR7`2kHcMS7_2eHiMZ7X2\\HRNb7P2TH[Ni7g1mGcNS8]1iGgNV8Z1gGiN_6iMjJ]3eNnN^6jMiJY3fNPO`6kMgJU3fNSOa6mMeJP3hNVOa6PNcJe5[5aJ`J`5_5eJ]J[5a5kJZJU5e5RKVJl4k5UKTJj4l5YKSJd4n5_KPJ`4Q6bKnI\\4S6fKlIY4T6hKkIY4T6iKkIW4T6kKjIV4U6kKkIU4U6lKjIT4U6mKjIT4U6nKjIR4V6nKhIU4V6[2O100O1000000000O2O00000000001O000000lI]Gk5P9L1O01O01O00001O00010O0000001O03M3M4L4L4L5L4K8H:F4L3M2O1N;E010O010O010O01O010O010O010O010O010O100O100O10001O2M4M4L:E2O00fKnHe0R7WOUIf0j6VO^Ig0b6UOeIZO5oMW6f2hIPO:WNP6g2kIfN`0`Nf5h2WJUN=oN^5j2dJYMa0Im4l2\\MRMe2m2h4N2N2N1O3N4K4L4L4L4M2M3M2N3M3M3N2M3M4L4L4L4M3L4K5K6I6KTX[2"}, "label": "the boy on the right in the right hand picture"}]}
{"id": 467840, "image": "COCO_train2014_000000467840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bag of the man in the gray coat\"."}], "task": "refering", "answer_template": "The \"bag of the man in the gray coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 257, 272, 273, 274, 275, 289, 290, 291, 292, 307, 308, 309, 324, 325, 326, 327, 342, 343, 344, 359, 360], "bbox": [0.0002708333333333333, 0.6758437500000001, 0.27877083333333336, 0.9594531249999999], "iscrowd": 0, "area": 10325.963249999997, "rle": {"size": [640, 480], "counts": "`>6Zc0`0B?O1000000000O10000000O1000O100000000000O0100000000000000O0100000000000O1000O10000000O1002N2N3M2M3N3M2H9G8H8I8G8H9G8I7H9J5M3M4L3M4L3M3M3M2N2N2N2N2N3M2N2001O100O0M4L4M2M4L3M4L3N3L3M3M4M2M4L3M4L3N3L3M4L3M3N3L3M4L3M4M2M4L3M3N3L3M4L3M4M2M4L3M4L3N2M4L3M4L3N3L3M4L3N3L3MUfg6"}, "label": "bag of the man in the gray coat"}]}
{"id": 467840, "image": "COCO_train2014_000000467840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black messenger bag\"."}], "task": "refering", "answer_template": "The \"a black messenger bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 257, 272, 273, 274, 275, 289, 290, 291, 292, 307, 308, 309, 324, 325, 326, 327, 342, 343, 344, 359, 360], "bbox": [0.0002708333333333333, 0.6758437500000001, 0.27877083333333336, 0.9594531249999999], "iscrowd": 0, "area": 10325.963249999997, "rle": {"size": [640, 480], "counts": "`>6Zc0`0B?O1000000000O10000000O1000O100000000000O0100000000000000O0100000000000O1000O10000000O1002N2N3M2M3N3M2H9G8H8I8G8H9G8I7H9J5M3M4L3M4L3M3M3M2N2N2N2N2N3M2N2001O100O0M4L4M2M4L3M4L3N3L3M3M4M2M4L3M4L3N3L3M4L3M3N3L3M4L3M4M2M4L3M3N3L3M4L3M4M2M4L3M4L3N2M4L3M4L3N3L3M4L3N3L3MUfg6"}, "label": "a black messenger bag"}]}
{"id": 467840, "image": "COCO_train2014_000000467840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue adidas bag\"."}], "task": "refering", "answer_template": "The \"a blue adidas bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 233, 249, 250, 266, 267, 283, 284, 300, 301, 317, 318, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 367, 368, 369, 370, 371, 372, 373, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6097083333333334, 0.5165, 1.0, 1.0], "iscrowd": 0, "area": 19697.2333, "rle": {"size": [640, 480], "counts": "`fg5=_c08I7I8H4K6K4L5K5K4L5K4L5K4L3L5K4L5N1O2N1O2O0O1O2N1O2N1O2J5G9G9F:G9G9F:G9G9G9F:G9G9F:G9G9F:G9G9F:GnKmEiNo9c0jF\\OV9LbG4^8UOZHj0g7]NRIb1n6fMjIZ2V6nLcJQ3]5XL[Kg3`:O001O1O001O1O002N1O1O2N1O1O2N1O1O1O2N1O1O001O0000000000001O00000000001O0O100000001O000000001O0000001O0000001O0000001O0000001O00001O00001O00001O00001O00001O0000001O00000O100000000000000000000000001O0000001O000000001O0000001O0000001O1O001O1O1O001O1O1O8H=C<D=@?Ba\\O"}, "label": "a blue adidas bag"}]}
{"id": 467840, "image": "COCO_train2014_000000467840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adidas bag\"."}], "task": "refering", "answer_template": "The \"the adidas bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 233, 249, 250, 266, 267, 283, 284, 300, 301, 317, 318, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 367, 368, 369, 370, 371, 372, 373, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6097083333333334, 0.5165, 1.0, 1.0], "iscrowd": 0, "area": 19697.2333, "rle": {"size": [640, 480], "counts": "`fg5=_c08I7I8H4K6K4L5K5K4L5K4L5K4L3L5K4L5N1O2N1O2O0O1O2N1O2N1O2J5G9G9F:G9G9F:G9G9G9F:G9G9F:G9G9F:G9G9F:GnKmEiNo9c0jF\\OV9LbG4^8UOZHj0g7]NRIb1n6fMjIZ2V6nLcJQ3]5XL[Kg3`:O001O1O001O1O002N1O1O2N1O1O2N1O1O1O2N1O1O001O0000000000001O00000000001O0O100000001O000000001O0000001O0000001O0000001O0000001O00001O00001O00001O00001O00001O0000001O00000O100000000000000000000000001O0000001O000000001O0000001O0000001O1O001O1O1O001O1O1O8H=C<D=@?Ba\\O"}, "label": "the adidas bag"}]}
{"id": 426888, "image": "COCO_train2014_000000426888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana to the left of some other bananas\"."}], "task": "refering", "answer_template": "The \"a banana to the left of some other bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 98, 114, 115, 116, 132, 133, 134, 135, 151, 152, 153, 170, 171, 188, 189, 207], "bbox": [0.3436, 0.36013333333333336, 0.5569400000000001, 0.8859733333333334], "iscrowd": 0, "area": 8816.139199999996, "rle": {"size": [375, 500], "counts": "jTo1=o:`0J7H8I4K3N3M2M4M2M3N3N1N3M2O2M2N2O2M2N3N1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N100O1O2O0O101N1O101N100O101N10001O0O101O0O2O1O0O2O11N2O0O2O001N2O1O1N2O1N2O5K8G4M2N3L3N3L3N2N3L3gM]GX1d9M3N1N1O100O1O1O100O1O10O01M3L4M3L4M3L4L4Mfo`2"}, "label": "a banana to the left of some other bananas"}]}
{"id": 426888, "image": "COCO_train2014_000000426888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far left banana in the bowl\"."}], "task": "refering", "answer_template": "The \"the far left banana in the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 98, 114, 115, 116, 132, 133, 134, 135, 151, 152, 153, 170, 171, 188, 189, 207], "bbox": [0.3436, 0.36013333333333336, 0.5569400000000001, 0.8859733333333334], "iscrowd": 0, "area": 8816.139199999996, "rle": {"size": [375, 500], "counts": "jTo1=o:`0J7H8I4K3N3M2M4M2M3N3N1N3M2O2M2N2O2M2N3N1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N100O1O2O0O101N1O101N100O101N10001O0O101O0O2O1O0O2O11N2O0O2O001N2O1O1N2O1N2O5K8G4M2N3L3N3L3N2N3L3gM]GX1d9M3N1N1O100O1O1O100O1O10O01M3L4M3L4M3L4L4Mfo`2"}, "label": "the far left banana in the bowl"}]}
{"id": 426888, "image": "COCO_train2014_000000426888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana is lower than the other two\"."}], "task": "refering", "answer_template": "The \"the banana is lower than the other two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 119, 120, 121, 137, 138, 139, 140, 155, 156, 157, 173, 174, 190, 191], "bbox": [0.60338, 0.43981333333333333, 0.79214, 0.8353066666666666], "iscrowd": 0, "area": 7950.61705, "rle": {"size": [375, 500], "counts": "cl^38j9IaG`0W81WG8c86mF2n8^1L4K5L4N2N2N2O1N2N2N2N2N2M3N1O0O2O0O2O1N101N2H54M2N3L3N3M3M3O100O1O100O10O01O100O100O100O1O10O0100O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2M3N2N2N2N4L4L3L5L4K4M4K4M4K5L3LPlU1"}, "label": "the banana is lower than the other two"}]}
{"id": 426888, "image": "COCO_train2014_000000426888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana in between two other bananas , wearing a pair of eyeglasses\"."}], "task": "refering", "answer_template": "The \"a banana in between two other bananas , wearing a pair of eyeglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 81, 82, 83, 99, 100, 101, 117, 118, 119, 135, 136, 137, 153, 154, 171, 172, 189, 190, 208], "bbox": [0.50648, 0.24234666666666665, 0.6653, 0.86352], "iscrowd": 0, "area": 10722.454600000001, "rle": {"size": [375, 500], "counts": "jhl2Z1[:6K5J7I6J6J7I6J6J7J5J6J7I6J6L5J5K5L>Ac0]Od0]Ob0]O>B100O1O10O10O10000000000000N2[Oe0@?K6J6J6J6J6J6M32OJ5L5J6K4L5L4M2M4M3L3N3L4M2O2M3M2O2M3M2N3N2M3M3N3L3L4L4L4L5J9H8H8HQQm1"}, "label": "a banana in between two other bananas , wearing a pair of eyeglasses"}]}
{"id": 426888, "image": "COCO_train2014_000000426888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow banana in the middle of a bunch\"."}], "task": "refering", "answer_template": "The \"a yellow banana in the middle of a bunch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 81, 82, 83, 99, 100, 101, 117, 118, 119, 135, 136, 137, 153, 154, 171, 172, 189, 190, 208], "bbox": [0.50648, 0.24234666666666665, 0.6653, 0.86352], "iscrowd": 0, "area": 10722.454600000001, "rle": {"size": [375, 500], "counts": "jhl2Z1[:6K5J7I6J6J7I6J6J7J5J6J7I6J6L5J5K5L>Ac0]Od0]Ob0]O>B100O1O10O10O10000000000000N2[Oe0@?K6J6J6J6J6J6M32OJ5L5J6K4L5L4M2M4M3L3N3L4M2O2M3M2O2M3M2N3N2M3M3N3L3L4L4L4L5J9H8H8HQQm1"}, "label": "a yellow banana in the middle of a bunch"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing yellow dress\"."}], "task": "refering", "answer_template": "The \"a women wearing yellow dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317, 318, 338, 339, 340, 341], "bbox": [0.707484375, 0.5662060889929743, 0.8746562499999999, 0.9890398126463701], "iscrowd": 0, "area": 10765.262200000001, "rle": {"size": [427, 640], "counts": "mVm56R=8I8G8I8G8I7I7H8I7aNoMWGY2\\8QN`GV2O_MU7c0gHR3Y7PMcHR3]7oL`HS3`7oL\\HS3c7PMYHR3g7PMVHQ3j7X10000O1000000O100000000O1000000O100000000O1000000001O001O001O000PLZH[2g7bM^H[2b7dMbHY2_7dMfHY2Z7fMiHW2X7hMkHV2V7gMnHW2R7eMTIY2m6cMXI[2h6aM^I]2b6`McI^2^6ZMkId2U6QMXJm2U8N2N2N1O2N2N2N2N2N2N3M7I8H8H8H6J1O2N2N2N2N2N1O2N2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1N2O1O1Oh[Q1"}, "label": "a women wearing yellow dress"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with yellow dress\"."}], "task": "refering", "answer_template": "The \"a lady with yellow dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317, 318, 338, 339, 340, 341], "bbox": [0.707484375, 0.5662060889929743, 0.8746562499999999, 0.9890398126463701], "iscrowd": 0, "area": 10765.262200000001, "rle": {"size": [427, 640], "counts": "mVm56R=8I8G8I8G8I7I7H8I7aNoMWGY2\\8QN`GV2O_MU7c0gHR3Y7PMcHR3]7oL`HS3`7oL\\HS3c7PMYHR3g7PMVHQ3j7X10000O1000000O100000000O1000000O100000000O1000000001O001O001O000PLZH[2g7bM^H[2b7dMbHY2_7dMfHY2Z7fMiHW2X7hMkHV2V7gMnHW2R7eMTIY2m6cMXI[2h6aM^I]2b6`McI^2^6ZMkId2U6QMXJm2U8N2N2N1O2N2N2N2N2N2N3M7I8H8H8H6J1O2N2N2N2N2N1O2N2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1N2O1O1Oh[Q1"}, "label": "a lady with yellow dress"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white umbrella\"."}], "task": "refering", "answer_template": "The \"a black and white umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 75, 76, 77, 78, 92, 93, 94, 98, 99, 115, 116, 117, 139], "bbox": [0.0, 0.0, 0.456734375, 0.42313817330210773], "iscrowd": 0, "area": 26700.52179999999, "rle": {"size": [427, 640], "counts": "S3e1d;3N2M3N1N3N2M3N2M2O2M3N2M2O2M3O1N101N2O1N2N101N2O1N3N2M3N2M1O100O010O10O01O10O0100O010O1O001O10O01O1O00100O001OO101O000O10010O01O001O00100O00HgF`LX9]3<N2M3N2M4M2N2N2J6D<00000001N10000000000000001O000001O001O0100O100O100O100O100O100O100O100O100O10O0100O1000000001O00000O2O00001O00001O00000O101O000000001O00000O2O00001O001O001N10001O001O00001O0O2O001O00001jF`M_7`2ZHfMg7[2eG\\MH<c8X2dG`MD9h8X2cGcM@6n8W2aGXN_8h1aGWN`8i1_GWNb8i1\\GYNd8h1WG[Nj8e1RG^No8b1mFbNS9j21O100O1O010O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O100O1O100O1O100O010O1O100O100O100O1O100O100O1O100O010O100O1O100O100O1O101N100O2O0O1O2O0O101N1O101N100O2O0O2O0O101N1O101N100O2O0O101N100O2O0XNcEa0^:[OlE?d;K6K4K`Qa4"}, "label": "a black and white umbrella"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white umbrella\"."}], "task": "refering", "answer_template": "The \"a black and white umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 75, 76, 77, 78, 92, 93, 94, 98, 99, 115, 116, 117, 139], "bbox": [0.0, 0.0, 0.456734375, 0.42313817330210773], "iscrowd": 0, "area": 26700.52179999999, "rle": {"size": [427, 640], "counts": "S3e1d;3N2M3N1N3N2M3N2M2O2M3N2M2O2M3O1N101N2O1N2N101N2O1N3N2M3N2M1O100O010O10O01O10O0100O010O1O001O10O01O1O00100O001OO101O000O10010O01O001O00100O00HgF`LX9]3<N2M3N2M4M2N2N2J6D<00000001N10000000000000001O000001O001O0100O100O100O100O100O100O100O100O100O10O0100O1000000001O00000O2O00001O00001O00000O101O000000001O00000O2O00001O001O001N10001O001O00001O0O2O001O00001jF`M_7`2ZHfMg7[2eG\\MH<c8X2dG`MD9h8X2cGcM@6n8W2aGXN_8h1aGWN`8i1_GWNb8i1\\GYNd8h1WG[Nj8e1RG^No8b1mFbNS9j21O100O1O010O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O100O1O100O1O100O010O1O100O100O100O1O100O100O1O100O010O100O1O100O100O1O101N100O2O0O1O2O0O101N1O101N100O2O0O2O0O101N1O101N100O2O0O101N100O2O0XNcEa0^:[OlE?d;K6K4K`Qa4"}, "label": "a black and white umbrella"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of a white pop open tent covering at a market on the sidewalk\"."}], "task": "refering", "answer_template": "The \"the top of a white pop open tent covering at a market on the sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 260, 261, 262, 263, 264, 265, 268, 269], "bbox": [0.295359375, 0.5578688524590164, 0.754140625, 0.7915690866510539], "iscrowd": 0, "area": 11560.3609, "rle": {"size": [427, 640], "counts": "aa_21Y=2O0O1O1O100O1O1O100O1O1O100O1O2N2O001N2O1O001O1N101O1O000O100000000O010000000O1000000O10O10O010O1O010O1O010O10O01O10O01O10O0100O1O100O100O1O010O100O1O100O100O1O010O1O010O001O1O001O1O0O2O1O1O1O1O1O1O1O1O0O2O1O1O100O1O1O001O100O1O1O1O1O010O1O1O1O1O100O1O001O1O100O1O1O10O2O00001O0000001O0O10001O00000O2O001N101O001N2O001N2O001N101O1N101O1N101O0O2O1O0O2O1O001N101O1N101O1O0O2O001N2O001O0O101O0O20O01O00010O001O01O01O010O001O01O01O010O000010O01O0010O00010O01O00010O001O010O00010O001O000010O01O001O000000010O00000000001O01O000000O2O00000000001N10000000001O000O1000001O00000O2O00001N1N3N8G^UQ2"}, "label": "the top of a white pop open tent covering at a market on the sidewalk"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white tent top\"."}], "task": "refering", "answer_template": "The \"the white tent top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 260, 261, 262, 263, 264, 265, 268, 269], "bbox": [0.295359375, 0.5578688524590164, 0.754140625, 0.7915690866510539], "iscrowd": 0, "area": 11560.3609, "rle": {"size": [427, 640], "counts": "aa_21Y=2O0O1O1O100O1O1O100O1O1O100O1O2N2O001N2O1O001O1N101O1O000O100000000O010000000O1000000O10O10O010O1O010O1O010O10O01O10O01O10O0100O1O100O100O1O010O100O1O100O100O1O010O1O010O001O1O001O1O0O2O1O1O1O1O1O1O1O1O0O2O1O1O100O1O1O001O100O1O1O1O1O010O1O1O1O1O100O1O001O1O100O1O1O10O2O00001O0000001O0O10001O00000O2O001N101O001N2O001N2O001N101O1N101O1N101O0O2O1O0O2O1O001N101O1N101O1O0O2O001N2O001O0O101O0O20O01O00010O001O01O01O010O001O01O01O010O000010O01O0010O00010O01O00010O001O010O00010O001O000010O01O001O000000010O00000000001O01O000000O2O00000000001N10000000001O000O1000001O00000O2O00001N1N3N8G^UQ2"}, "label": "the white tent top"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing behind a man that is holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a man standing behind a man that is holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324], "bbox": [0.001171875, 0.401288056206089, 0.09601562500000001, 1.0], "iscrowd": 0, "area": 12092.415250000002, "rle": {"size": [427, 640], "counts": "ec0;m<?mCd4\\5`KaJb4[5cKcJ^4Z5fKcJ\\4Z5hKdJY4Y5lKcJV4[5mKcJT4[5oKbJS4\\5QLaJP4\\5TLaJn3]5Q2O11O00000000000000001O000000001O0000001O01O02hIaJ`4a5YKjJa4W5XKUKb4m4VK^Ke4b4VKfKg4Z4UKlKi4T4UKQLh4o3UKVLi4j3TK\\Li4d3UK`Li4`3TKfLi4Z3TKkLj4U3TKPMj4j5N3M3M2N3M3M2K7G`0@a0_Og0SNYEMZ<E:GVRa7"}, "label": "a man standing behind a man that is holding an umbrella"}]}
{"id": 99211, "image": "COCO_train2014_000000099211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black shirt and with a hat is standing behind a man with the umbrella\"."}], "task": "refering", "answer_template": "The \"a man in black shirt and with a hat is standing behind a man with the umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324], "bbox": [0.001171875, 0.401288056206089, 0.09601562500000001, 1.0], "iscrowd": 0, "area": 12092.415250000002, "rle": {"size": [427, 640], "counts": "ec0;m<?mCd4\\5`KaJb4[5cKcJ^4Z5fKcJ\\4Z5hKdJY4Y5lKcJV4[5mKcJT4[5oKbJS4\\5QLaJP4\\5TLaJn3]5Q2O11O00000000000000001O000000001O0000001O01O02hIaJ`4a5YKjJa4W5XKUKb4m4VK^Ke4b4VKfKg4Z4UKlKi4T4UKQLh4o3UKVLi4j3TK\\Li4d3UK`Li4`3TKfLi4Z3TKkLj4U3TKPMj4j5N3M3M2N3M3M2K7G`0@a0_Og0SNYEMZ<E:GVRa7"}, "label": "a man in black shirt and with a hat is standing behind a man with the umbrella"}]}
{"id": 66443, "image": "COCO_train2014_000000066443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sad looking man in a train\"."}], "task": "refering", "answer_template": "The \"a sad looking man in a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219], "bbox": [0.45692187500000003, 0.326056338028169, 0.67078125, 0.6637089201877935], "iscrowd": 0, "area": 9900.476999999999, "rle": {"size": [426, 640], "counts": "^fi32P=b0A8J5L4M3L4M3M3M3M3M3M3M3N2M3M3M2O2M2N3M2O2N1OO01001N100O10000O10ZOjMXFV2_9VN]Fj1]9_N`Fb1]9V1O2M2000000001O000000001O001O1O001O001O1O1O1N2O1O1O1O1O1O1O1O1O2N1O2N1O101O001O1O1O1O1O1O1O1O3M6J6cM_Ee1X;K1O00001O001O001O1O1O001O1O1O1O1O1O1O2N1O1O1O1O1O001O001O001O001O1O001O001O001O001O1O1O001O1O1O2N3Mkag2"}, "label": "a sad looking man in a train"}]}
{"id": 66443, "image": "COCO_train2014_000000066443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a beige shirt leaning out the window of a train\"."}], "task": "refering", "answer_template": "The \"a man with a beige shirt leaning out the window of a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219], "bbox": [0.45692187500000003, 0.326056338028169, 0.67078125, 0.6637089201877935], "iscrowd": 0, "area": 9900.476999999999, "rle": {"size": [426, 640], "counts": "^fi32P=b0A8J5L4M3L4M3M3M3M3M3M3M3N2M3M3M2O2M2N3M2O2N1OO01001N100O10000O10ZOjMXFV2_9VN]Fj1]9_N`Fb1]9V1O2M2000000001O000000001O001O1O001O001O1O1O1N2O1O1O1O1O1O1O1O1O2N1O2N1O101O001O1O1O1O1O1O1O1O3M6J6cM_Ee1X;K1O00001O001O001O1O1O001O1O1O1O1O1O1O2N1O1O1O1O1O001O001O001O001O1O001O001O001O001O1O1O001O1O1O2N3Mkag2"}, "label": "a man with a beige shirt leaning out the window of a train"}]}
{"id": 164751, "image": "COCO_train2014_000000164751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an external keyboard in front of a laptop\"."}], "task": "refering", "answer_template": "The \"an external keyboard in front of a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 306, 307, 308, 309, 310], "bbox": [0.024520833333333332, 0.645421875, 0.6935, 0.7974218750000001], "iscrowd": 0, "area": 24950.3609, "rle": {"size": [640, 480], "counts": "lo72jc04L4M3L4L4L4M3L4L4M3L4L4L4M2M4L4M3L4L4L4M3L4L4N200000000O10O1000000000000000000000000000O0100000000000000000000000000000O01000000000000000000000000000000O0100000000000000000000000000000O01000000000000000000000000000O10N2000000000000000000000000000O2O00000000000000000000000000000O10000000000000000000000000000O2O00000000000000000000000000000O1000000000000000000000000000001N1000000000000000000000000000000O10000000000000000000000000001N1000000000000000000000000000000O10000000000000000000000000001O0O1000000000000000000000000000000O100000000000000000000000001O0O1000000000000000000000000000000O100000003M3M4L4L4L3M4L4L4L3M4L4K4M4L4L4L3M4L4Lg\\k2"}, "label": "an external keyboard in front of a laptop"}]}
{"id": 164751, "image": "COCO_train2014_000000164751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"i see a picture of a keyboard\"."}], "task": "refering", "answer_template": "The \"i see a picture of a keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 306, 307, 308, 309, 310], "bbox": [0.024520833333333332, 0.645421875, 0.6935, 0.7974218750000001], "iscrowd": 0, "area": 24950.3609, "rle": {"size": [640, 480], "counts": "lo72jc04L4M3L4L4L4M3L4L4M3L4L4L4M2M4L4M3L4L4L4M3L4L4N200000000O10O1000000000000000000000000000O0100000000000000000000000000000O01000000000000000000000000000000O0100000000000000000000000000000O01000000000000000000000000000O10N2000000000000000000000000000O2O00000000000000000000000000000O10000000000000000000000000000O2O00000000000000000000000000000O1000000000000000000000000000001N1000000000000000000000000000000O10000000000000000000000000001N1000000000000000000000000000000O10000000000000000000000000001O0O1000000000000000000000000000000O100000000000000000000000001O0O1000000000000000000000000000000O100000003M3M4L4L4L3M4L4L4L3M4L4K4M4L4L4L3M4L4Lg\\k2"}, "label": "i see a picture of a keyboard"}]}
{"id": 320403, "image": "COCO_train2014_000000320403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe in front looking down\"."}], "task": "refering", "answer_template": "The \"giraffe in front looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 235, 236, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 288, 289, 290, 292, 293, 304, 305, 306, 308, 309, 321, 322, 324, 325, 337, 338, 340, 341, 353, 354, 356, 357], "bbox": [0.026541019955654103, 0.49521875, 0.8029268292682927, 0.9973906249999999], "iscrowd": 0, "area": 35453.08394999999, "rle": {"size": [640, 451], "counts": "eP82lc05K3L3N2N2N3M2N2M3N2_O`000001O0N2L5KH8\\Oe0ZOg0N3O10lN__OjNa`0S1h_OeNZ`0X1W1Me_OjNU>_1bAgNX>h1XA_Nc>o1m@XNn>R2g@TNo1XO`:m2UCWNh1XOQ;X4^DSL`;V4mCVLP<S4_CWL`<f5N2N2M3N2N2N1N3NcMPDmLn;R3YDjLe;T3dDgLY;X3PEcLn:[3[E_Ld:_3dE]LZ:a3oEYLQ:d3XFVLg9i3aFQL_9l3iFoKV9o3SGlKl8Q4\\GjKc8T4bGjK^8S4gGkKY8R4lGeKZ8Z4g3M2N3M3M2O2N101N101N2O0O2O0O2O0O2O1N101N101N2O0O2O0O101N12O1N100O1O100O100O10O0100O1O100O100O100O00100O100O1N2M300O10O0100O100O10:ET1eClJd7T6RGZJi8h6lEfIo9Z8L0O100O100O010O100O10O0100O100O10O0100O10[NUFUIj9b6hFVIX9a6\\GSId8d6iGYIW8]6VH`Ii7Y6bHdI^7V6kHgIT7T6UIiIk6P6f3J6J6J6M3M4M2M3M3M3M3M3M3M3M3M3M3M3M4L3E;N2N2M3N2\\Nd_OB^`0=f_O]O]`0c0f_OXO\\`0g0g_OUO[`0k0h_OPOZ`0o0j_OkNY`0U1j_OfNY`0X1j_OdNX`0\\1k_O_NW`0`1U1O100O100O100O10000O2O00000000000O10000000000O2O01O0000000000000000001O0001O00000000000001O0001O001O00001O00010O00001O00001O001O01O01O01O01O01O01O01O01O010O00010O00010O00010O00010O0010O0010O00010O00010O00010O0010O00010O00010O00100O1O010O100O001D<O1O001O1O1O0012N1O00I601O1O1O001O1002M3N2N2NO001O_NV^Om0ia0ROd^Oc0Za0]Oj^Oa0Ua0^Oo^O?Qa0@R_O?m`0@W_O>h`0BZ_O=e`0B__O;Ub0O1N3K4LXWg1"}, "label": "giraffe in front looking down"}]}
{"id": 320403, "image": "COCO_train2014_000000320403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a crouching giraffe\"."}], "task": "refering", "answer_template": "The \"a crouching giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 235, 236, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 288, 289, 290, 292, 293, 304, 305, 306, 308, 309, 321, 322, 324, 325, 337, 338, 340, 341, 353, 354, 356, 357], "bbox": [0.026541019955654103, 0.49521875, 0.8029268292682927, 0.9973906249999999], "iscrowd": 0, "area": 35453.08394999999, "rle": {"size": [640, 451], "counts": "eP82lc05K3L3N2N2N3M2N2M3N2_O`000001O0N2L5KH8\\Oe0ZOg0N3O10lN__OjNa`0S1h_OeNZ`0X1W1Me_OjNU>_1bAgNX>h1XA_Nc>o1m@XNn>R2g@TNo1XO`:m2UCWNh1XOQ;X4^DSL`;V4mCVLP<S4_CWL`<f5N2N2M3N2N2N1N3NcMPDmLn;R3YDjLe;T3dDgLY;X3PEcLn:[3[E_Ld:_3dE]LZ:a3oEYLQ:d3XFVLg9i3aFQL_9l3iFoKV9o3SGlKl8Q4\\GjKc8T4bGjK^8S4gGkKY8R4lGeKZ8Z4g3M2N3M3M2O2N101N101N2O0O2O0O2O0O2O1N101N101N2O0O2O0O101N12O1N100O1O100O100O10O0100O1O100O100O100O00100O100O1N2M300O10O0100O100O10:ET1eClJd7T6RGZJi8h6lEfIo9Z8L0O100O100O010O100O10O0100O100O10O0100O10[NUFUIj9b6hFVIX9a6\\GSId8d6iGYIW8]6VH`Ii7Y6bHdI^7V6kHgIT7T6UIiIk6P6f3J6J6J6M3M4M2M3M3M3M3M3M3M3M3M3M3M3M4L3E;N2N2M3N2\\Nd_OB^`0=f_O]O]`0c0f_OXO\\`0g0g_OUO[`0k0h_OPOZ`0o0j_OkNY`0U1j_OfNY`0X1j_OdNX`0\\1k_O_NW`0`1U1O100O100O100O10000O2O00000000000O10000000000O2O01O0000000000000000001O0001O00000000000001O0001O001O00001O00010O00001O00001O001O01O01O01O01O01O01O01O01O010O00010O00010O00010O00010O0010O0010O00010O00010O00010O0010O00010O00010O00100O1O010O100O001D<O1O001O1O1O0012N1O00I601O1O1O001O1002M3N2N2NO001O_NV^Om0ia0ROd^Oc0Za0]Oj^Oa0Ua0^Oo^O?Qa0@R_O?m`0@W_O>h`0BZ_O=e`0B__O;Ub0O1N3K4LXWg1"}, "label": "a crouching giraffe"}]}
{"id": 320403, "image": "COCO_train2014_000000320403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe reaching to eat some leaves off of a tree\"."}], "task": "refering", "answer_template": "The \"a giraffe reaching to eat some leaves off of a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 45, 46, 61, 62, 76, 77, 78, 93, 108, 109, 124, 125, 140, 141, 155, 156, 157, 170, 171, 172, 173, 185, 186, 187, 188, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 231, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 300, 311, 312, 314, 315, 327, 328, 330, 331, 343, 344, 345, 346, 347, 348, 359, 360, 361, 362, 364], "bbox": [0.4481818181818182, 0.061406249999999996, 0.9276940133037693, 0.9913593749999998], "iscrowd": 0, "area": 34183.226049999954, "rle": {"size": [640, 451], "counts": "fkn31mc04L3XOIc]O9Zb0Lb]O6\\b0N`]O4^b00]]O3ab01[]O0db0g0N2M3N2N2N2K5XOd0c^OeMSa0c221ZKbNhEBnMP2Z<lNZEc1X;gN_D>Z<L]C3e<5RCIR=>fB@]=^33M3H8NjClJe9S5TFWKj9h4nEbKQ:^4jEhKV:V4fEoK[:o3aEWL_:h3]E\\Ld:\\3^EjLb:k2cE[M]:l2TE\\Mm:m2]D_Mc;_501O00001O01jNPE[Io:]6[EaIe:W6eEgI[:Q6nEoIQ:i5WFWJi9d5]F\\Jc9_5a2M3L4L4L5K4L4MhN\\KZC`4j<`KWC[4l<gKSCX4m<jKRCT4P=lKPCS4P=oKPCn3Q=TLnBk3S=ULmBi3T=YLkBf3U=[LkBc3W=]LjB`3W=bLhB]3X=dLhBZ3Z=fLfBY3m2YLY6?jFU3n2^LV6>lFS3m2aLW6<lFQ3W2dLSM0j9<kFo2W2mLmLIP:<lFl2X2bMk6BnFj2V2fMl6AmFg2X2iMk6@mFe2X2mMj6_OoF`2V2UNk6[OoF]2R2^NP7VOmFZ2o1fNS7QOoFU2j1QOW7jNoFR2f1[O[7dNnFo1`1Fa7\\NPGj15hM\\O\\2[9\\MnEf0V1g14^1g8QMTGb10a1k8nLUGe1Gb1S9jLWGg1]Oc1\\9hLVGj1SOW1MSLi9m0WGP6[9QJfFQ5U:RKjEh4[:YKeE`4a:aK`EX4e:jKZEo3l:RLTEh3Q;YLoD`3W;bLiDY3Y;iLgDS3\\;nLdDo2^;SMbDh2a;YM_Dc2d;^M\\DX1\\OdM\\<V1WDn0KbMQ<a1UDd0[=]OeB=`=E_B7d=J\\B2g=0YBKj=6VBEm==TB^Om=f0SBVOn=l0SBROl=P1UBnNj=U1UBjNk=W1VBgNi=[1XBcNh=_1XB_Ng=c1ZB[NW1VOh9V2\\EcNg0A3Hm6T2[HaN`0NLEW7m1^H^NOl1iNRNh8e1`H^NG\\3g7WNcH\\N@c3m7QNdH[NkNY4BnJe8_2oHZNfN]4DkJf8_2QIXNbNa4ClJi8^2QIUN`Nj4^8\\MiHiMfNn4`8SMZG^M_1^6V7SL^G\\M^1a6T7QLaG[M^1c6P7RL[Ib1QNQNc8=^I[1eNiMk7n0`IR1GTMg6l1bIj0=jLo5^2eI>g0RMb5c2gILW1_Mo4i2jI^Ob1gMa4o2lITOk1kMV4^3fIaN[2oMm3b3iIXNb2TNc3f3lIPNe:S2[EgMi:[2fCULU1Z1V;e2dCWLR1o0[;m2bC[Ln0e0_;S3bCeLc05j;Z3bCeLb0Mm;a3_CfLb0Fo;f3^ChLa0^OR<l3\\CjL?XOU<P4ZClL`0POW<j5fCSJ[<Q6bCnIV:HbG_6UNgIW:NcG_6SNcIV:3eG]6SN`IV:6dGh6]8XI`G^3YN\\OW:WM\\GZ3jNUOn9`MVGX3XOmNf9lMnFT3GgN`9UNeFR34aN[9]N^FS3=UN\\9dNXFX3c0_Mf9UOgE^3h0WMi8TO\\E532m0`3R1mLc8AVE271j0`3c1_LW8<[EDg0d3n<jLYBBe0f3T=iLVBBa0c3`=lLmAAb0Z3k=WMaA@b0U3S>gMkAW2Y>jMeAS2^>PN]Ao1g>`14M3M2N3M^OkK`AR4a>SLZAj3k>XLPAe3V?<6J6J6J7I7_Oa0N2M4L5L3eMZ_OW1k`0fNW_OV1m`0gNU_Oo0Wa0mNl^OQ1Rb000010000WO`]O7Zb0Gj]O8Ub0Eo]O;Qb0BQ^O=Rb0^OR^Oa0gb000O100O10Z`d0"}, "label": "a giraffe reaching to eat some leaves off of a tree"}]}
{"id": 320403, "image": "COCO_train2014_000000320403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe stretching it ' s head to try and reach some leaves\"."}], "task": "refering", "answer_template": "The \"the giraffe stretching it ' s head to try and reach some leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 45, 46, 61, 62, 76, 77, 78, 93, 108, 109, 124, 125, 140, 141, 155, 156, 157, 170, 171, 172, 173, 185, 186, 187, 188, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 231, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 300, 311, 312, 314, 315, 327, 328, 330, 331, 343, 344, 345, 346, 347, 348, 359, 360, 361, 362, 364], "bbox": [0.4481818181818182, 0.061406249999999996, 0.9276940133037693, 0.9913593749999998], "iscrowd": 0, "area": 34183.226049999954, "rle": {"size": [640, 451], "counts": "fkn31mc04L3XOIc]O9Zb0Lb]O6\\b0N`]O4^b00]]O3ab01[]O0db0g0N2M3N2N2N2K5XOd0c^OeMSa0c221ZKbNhEBnMP2Z<lNZEc1X;gN_D>Z<L]C3e<5RCIR=>fB@]=^33M3H8NjClJe9S5TFWKj9h4nEbKQ:^4jEhKV:V4fEoK[:o3aEWL_:h3]E\\Ld:\\3^EjLb:k2cE[M]:l2TE\\Mm:m2]D_Mc;_501O00001O01jNPE[Io:]6[EaIe:W6eEgI[:Q6nEoIQ:i5WFWJi9d5]F\\Jc9_5a2M3L4L4L5K4L4MhN\\KZC`4j<`KWC[4l<gKSCX4m<jKRCT4P=lKPCS4P=oKPCn3Q=TLnBk3S=ULmBi3T=YLkBf3U=[LkBc3W=]LjB`3W=bLhB]3X=dLhBZ3Z=fLfBY3m2YLY6?jFU3n2^LV6>lFS3m2aLW6<lFQ3W2dLSM0j9<kFo2W2mLmLIP:<lFl2X2bMk6BnFj2V2fMl6AmFg2X2iMk6@mFe2X2mMj6_OoF`2V2UNk6[OoF]2R2^NP7VOmFZ2o1fNS7QOoFU2j1QOW7jNoFR2f1[O[7dNnFo1`1Fa7\\NPGj15hM\\O\\2[9\\MnEf0V1g14^1g8QMTGb10a1k8nLUGe1Gb1S9jLWGg1]Oc1\\9hLVGj1SOW1MSLi9m0WGP6[9QJfFQ5U:RKjEh4[:YKeE`4a:aK`EX4e:jKZEo3l:RLTEh3Q;YLoD`3W;bLiDY3Y;iLgDS3\\;nLdDo2^;SMbDh2a;YM_Dc2d;^M\\DX1\\OdM\\<V1WDn0KbMQ<a1UDd0[=]OeB=`=E_B7d=J\\B2g=0YBKj=6VBEm==TB^Om=f0SBVOn=l0SBROl=P1UBnNj=U1UBjNk=W1VBgNi=[1XBcNh=_1XB_Ng=c1ZB[NW1VOh9V2\\EcNg0A3Hm6T2[HaN`0NLEW7m1^H^NOl1iNRNh8e1`H^NG\\3g7WNcH\\N@c3m7QNdH[NkNY4BnJe8_2oHZNfN]4DkJf8_2QIXNbNa4ClJi8^2QIUN`Nj4^8\\MiHiMfNn4`8SMZG^M_1^6V7SL^G\\M^1a6T7QLaG[M^1c6P7RL[Ib1QNQNc8=^I[1eNiMk7n0`IR1GTMg6l1bIj0=jLo5^2eI>g0RMb5c2gILW1_Mo4i2jI^Ob1gMa4o2lITOk1kMV4^3fIaN[2oMm3b3iIXNb2TNc3f3lIPNe:S2[EgMi:[2fCULU1Z1V;e2dCWLR1o0[;m2bC[Ln0e0_;S3bCeLc05j;Z3bCeLb0Mm;a3_CfLb0Fo;f3^ChLa0^OR<l3\\CjL?XOU<P4ZClL`0POW<j5fCSJ[<Q6bCnIV:HbG_6UNgIW:NcG_6SNcIV:3eG]6SN`IV:6dGh6]8XI`G^3YN\\OW:WM\\GZ3jNUOn9`MVGX3XOmNf9lMnFT3GgN`9UNeFR34aN[9]N^FS3=UN\\9dNXFX3c0_Mf9UOgE^3h0WMi8TO\\E532m0`3R1mLc8AVE271j0`3c1_LW8<[EDg0d3n<jLYBBe0f3T=iLVBBa0c3`=lLmAAb0Z3k=WMaA@b0U3S>gMkAW2Y>jMeAS2^>PN]Ao1g>`14M3M2N3M^OkK`AR4a>SLZAj3k>XLPAe3V?<6J6J6J7I7_Oa0N2M4L5L3eMZ_OW1k`0fNW_OV1m`0gNU_Oo0Wa0mNl^OQ1Rb000010000WO`]O7Zb0Gj]O8Ub0Eo]O;Qb0BQ^O=Rb0^OR^Oa0gb000O100O10Z`d0"}, "label": "the giraffe stretching it ' s head to try and reach some leaves"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car between two other cars\"."}], "task": "refering", "answer_template": "The \"a white car between two other cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 33, 34, 52, 53, 54, 56, 57, 58, 59, 60, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107], "bbox": [0.29246875, 0.11212048192771085, 0.6644375, 0.29038554216867474], "iscrowd": 0, "area": 5559.792200000002, "rle": {"size": [415, 640], "counts": "oW\\22m<1N2O1O1N2O1N2O1N101N101N100O2O001N10001O001O0O10000000000O10000000O10O10000000O100O10O0100O100O100AkC3V<KoC2Q<LQD4o;JTD4\\<00O2O1N10nh=5jVB3N3CERD=n;EoC=P<901O00001O00001O00001O01H7L400000000O10000000000000001O00001O0100O0dCMm;j0H8H7J7H8IO0O2O001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001N101O[Z9NgeF01O00001O00001O0000001O00001O00001O000O3N1O1O1O2N1O1O2N2N4L3GWQg2"}, "label": "a white car between two other cars"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car parked next to a red vehicle\"."}], "task": "refering", "answer_template": "The \"a white car parked next to a red vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 33, 34, 52, 53, 54, 56, 57, 58, 59, 60, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107], "bbox": [0.29246875, 0.11212048192771085, 0.6644375, 0.29038554216867474], "iscrowd": 0, "area": 5559.792200000002, "rle": {"size": [415, 640], "counts": "oW\\22m<1N2O1O1N2O1N2O1N101N101N100O2O001N10001O001O0O10000000000O10000000O10O10000000O100O10O0100O100O100AkC3V<KoC2Q<LQD4o;JTD4\\<00O2O1N10nh=5jVB3N3CERD=n;EoC=P<901O00001O00001O00001O01H7L400000000O10000000000000001O00001O0100O0dCMm;j0H8H7J7H8IO0O2O001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001N101O[Z9NgeF01O00001O00001O0000001O00001O00001O000O3N1O1O1O2N1O1O2N2N4L3GWQg2"}, "label": "a white car parked next to a red vehicle"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white station wagon with dark windows\"."}], "task": "refering", "answer_template": "The \"a white station wagon with dark windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99], "bbox": [0.002921875, 0.09214457831325301, 0.35700000000000004, 0.30785542168674696], "iscrowd": 0, "area": 12279.4911, "rle": {"size": [415, 640], "counts": "bk0`1^;e0[O2O0O100O1O100O1O100O1O101N10000000000000000000001O0000000O10000000000000001O00000000000000000000001O0000000000000000O10OO2XOh0_Oa0O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1000000O1000000O10000000O10004L8H9G8H9G8H9G1O001O001O00001O001O00001O000000001O0000000O10000000000000O10001O1O001O001O0_NeE>\\:AgE<Z:BjE:W:FlE7U:HmE6S:IQF4P:KSFLT:2PFFU::nE^OX:`0T101O1N101N101N2O001N101N101N2O001O010001N100O10001N`1aNO010O1O001O0010O0001O000001O0001O000001O01O000001O01O0WOjDIV;6PEEP;:WE@i:?]E]Oc:a0cEZO]:e0o0O1O1O2M2O1O1O1O1O2N1O1O1O1O2N1N2O1OilV5"}, "label": "a white station wagon with dark windows"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the parked silver car closest to the raised fist and shoulder of the man in a brown tee shirt\"."}], "task": "refering", "answer_template": "The \"the parked silver car closest to the raised fist and shoulder of the man in a brown tee shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99], "bbox": [0.002921875, 0.09214457831325301, 0.35700000000000004, 0.30785542168674696], "iscrowd": 0, "area": 12279.4911, "rle": {"size": [415, 640], "counts": "bk0`1^;e0[O2O0O100O1O100O1O100O1O101N10000000000000000000001O0000000O10000000000000001O00000000000000000000001O0000000000000000O10OO2XOh0_Oa0O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1000000O1000000O10000000O10004L8H9G8H9G8H9G1O001O001O00001O001O00001O000000001O0000000O10000000000000O10001O1O001O001O0_NeE>\\:AgE<Z:BjE:W:FlE7U:HmE6S:IQF4P:KSFLT:2PFFU::nE^OX:`0T101O1N101N101N2O001N101N101N2O001O010001N100O10001N`1aNO010O1O001O0010O0001O000001O0001O000001O01O000001O01O0WOjDIV;6PEEP;:WE@i:?]E]Oc:a0cEZO]:e0o0O1O1O2M2O1O1O1O1O2N1O1O1O1O2N1N2O1OilV5"}, "label": "the parked silver car closest to the raised fist and shoulder of the man in a brown tee shirt"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown shirt playing frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a brown shirt playing frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 31, 32, 33, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 191, 192, 193, 214, 215, 236, 237, 238, 259, 260, 282, 283, 305, 306], "bbox": [0.241890625, 0.044939759036144576, 0.47065625000000005, 0.9258313253012047], "iscrowd": 0, "area": 21248.339249999994, "rle": {"size": [415, 640], "counts": "YPo11n<0O2N101N101N1O101N1O3N5J2O0O1mGF`3;oK6Q4J^Kh0`4YO_Kj0^4WOcKi0Z4YOfKh0X4YOiKh0S4ZOnKf0o3\\OQLe0l3]OULc0h3^OYLd0c3^O^Lb0_3_ObL`0]3AeL=[3DdL<[3EdL;\\3GbL:]3GbL:]3DeL<[3BgL?X3^OkLIfLH^6=oLIQM]OQ6g0PMKl33VLLk34VLJk37TLIl38TLHk39ULGhMBZ4j0lMEeM0S4=XNC`M8S47XKnN`2d0QN<S45ZKRO]2=QNb0\\3jN]LW1GWOY27QNh0_3hN[LU1I[OW2X2a1^MYLR1M_OS2T2c1_MYLP1ODP2n1e1YOYLPOl1i1k1l0PNUOP2o0jMSOU2R1fMPOY2T1bMmN]2c1RM_Nn2U2]LlMb3j5O0OdN`L[I`3b6UMjHm2S7`1M4M2N3M2M4M2N3M3M2M4M2N3M2M4M2N3M3L3NZOnIeJQ6X5TJeJm5Y5WJeJh5Z5]JbJd5[5aJcJ_5[5eJbJ\\5\\5hJaJY5`5iJ]JW5c5R11O001O1O001N101N2O0O2O001N2O0O2O0O2O1N101N101N2O1N3N1N3N2M2H9^NPIgL]7T3iHjLW7S3mHkLU7Q3PIlLQ72iHf1;VNm63mH_1:\\Nk63mH]1=^Nf65nHY1a0_Nk6OdH_1e0aNm6J_Ha1i0bNP7EYHd1m0dNf7X1_HfNa7W1cHgN_7U1fHiN[7S1jHjNX7R1lHlNU7Q1_2L4L4K4M4L4L3M4LkYY4"}, "label": "a man in a brown shirt playing frisbee"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red car\"."}], "task": "refering", "answer_template": "The \"a red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88], "bbox": [0.496578125, 0.08783132530120483, 0.8735937500000001, 0.2700240963855422], "iscrowd": 0, "area": 7472.529, "rle": {"size": [415, 640], "counts": "cmP42l<2O001N10001N101O0O2O001O001N101O001O0O2O001O000O2O001O001O0O2O001O001O0O2O001O001O0O101O001OO1O0L5O1O1O1000O100000O100000O100001O000000000001O2N2N9G2N1O001O000000001O0000001O000000001O000010O0001O002N2N3M3M2N3M3M2N5K1O01O0000001O0aN^Ea0b:^OeE<[:C_FDb9;a10001O0000001O00001O00001O001O100O010O10O103M>XDPOc:i1NO010O1O0010O01O1O001O00001O001O001O00001O001N10001O001O00001O00001O001O000010O01O2N>Bc0]ObX7KagH3N5\\CIX<b0M4L;E1O000000O2O0000001O0000001O0000001O00001O0O3N6J2M1O2N2N2IhgP1"}, "label": "a red car"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red car in background\"."}], "task": "refering", "answer_template": "The \"red car in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88], "bbox": [0.496578125, 0.08783132530120483, 0.8735937500000001, 0.2700240963855422], "iscrowd": 0, "area": 7472.529, "rle": {"size": [415, 640], "counts": "cmP42l<2O001N10001N101O0O2O001O001N101O001O0O2O001O000O2O001O001O0O2O001O001O0O2O001O001O0O101O001OO1O0L5O1O1O1000O100000O100000O100001O000000000001O2N2N9G2N1O001O000000001O0000001O000000001O000010O0001O002N2N3M3M2N3M3M2N5K1O01O0000001O0aN^Ea0b:^OeE<[:C_FDb9;a10001O0000001O00001O00001O001O100O010O10O103M>XDPOc:i1NO010O1O0010O01O1O001O00001O001O001O00001O001N10001O001O00001O00001O001O000010O01O2N>Bc0]ObX7KagH3N5\\CIX<b0M4L;E1O000000O2O0000001O0000001O0000001O00001O0O3N6J2M1O2N2N2IhgP1"}, "label": "red car in background"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man jumping to catch frisbee\"."}], "task": "refering", "answer_template": "The \"man jumping to catch frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 132, 133, 134, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244], "bbox": [0.46823437500000004, 0.3046265060240964, 0.9094374999999999, 0.6869879518072289], "iscrowd": 0, "area": 18661.541100000013, "rle": {"size": [415, 640], "counts": "gji336N^<;N1O2N2O1N2N101O1O001O00000001O0001O0lNZOoEf0P:\\OoEd0P:_OnEa0Q:BmE>R:DmE<R:GlE9T:HkE8T:IlE7T:JlE5S:LmE4S:MlE3T:MlE3T:MlE3T:NkE2U:NkE2T:0kE0U:0kE0U:1jEOV:1jEOV:0kE0U:0kE0T:1lEOT:0mE0S:0mE0S:0mE0S:0mEOS:2mENS:2mENS:3lENR:3nEMR:4nELQ:4oELP:6oEKP:5PFKP:5PFLn96QFJo96RFJm97TFGl99PFKQ:4kE0U:1gE3Y:LfE5Z:L_ESO1Q1`:L_E:b:E^E;b:Q1000000O101O000O100000000O101O000O10000000000000O10O10000000O10000000000O100000000O10000000000O1000O100O01O010O00100O1O100O1O100O001O1O1OO2O1O1N2000001O000O10O10O010O001O001O001O010O001O1O001O1O001M3M2O2O010000O01000O10O0100O10O0100O10O01O100O01K4K6N101O0010O0000001O001O001O001O001O1O1000000O100kNoE\\OQ:c0QF\\Oo9d0QF[OP:d0RF[Oo9c0RF]On9b0TF]Ol9c0UF\\Ok9c0VF]Oj9c0WF\\Oj9b0XF^Og9b0YF^Oh9a0YF_Of9`0\\F_Oc9b0]F^Oc9b0^F^Oa9b0_F^O`9DTF6`05\\9]OoEN95d0a0Q9\\OVFK93b0f0n8\\OlFI9k0i8[OTHe0k7[OWHd0h7\\OYHe0f7ZO\\He0c7[O_Hd0`7]OaHc0^7^ObHa0]7@dH?[7BeH>[7BfH>X7ChH=W7DjH;V7EjH<T7EmH:T7ElH;T7ElH<S7DnH;R7EnH;R7EnH<Q7DPI;P7EPI;Q7DlGN@>c8DhG4D8d8CeG;F1e8DeG<E1e8CgG=B0g8CgG?AMi8CgG`0_ONi8AiGc0YOOn8^OmGh1T8WNlGi1T8VNjGn1V8oMdGY2\\8fMdG[2^93N0011O0O2O6JN10hNaE2_:KjENW:0RFIo93ZFGf9D[E>W<@oC:b<I_Zg0"}, "label": "man jumping to catch frisbee"}]}
{"id": 369557, "image": "COCO_train2014_000000369557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man that is laid out in the air\"."}], "task": "refering", "answer_template": "The \"the man that is laid out in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 132, 133, 134, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244], "bbox": [0.46823437500000004, 0.3046265060240964, 0.9094374999999999, 0.6869879518072289], "iscrowd": 0, "area": 18661.541100000013, "rle": {"size": [415, 640], "counts": "gji336N^<;N1O2N2O1N2N101O1O001O00000001O0001O0lNZOoEf0P:\\OoEd0P:_OnEa0Q:BmE>R:DmE<R:GlE9T:HkE8T:IlE7T:JlE5S:LmE4S:MlE3T:MlE3T:MlE3T:NkE2U:NkE2T:0kE0U:0kE0U:1jEOV:1jEOV:0kE0U:0kE0T:1lEOT:0mE0S:0mE0S:0mE0S:0mEOS:2mENS:2mENS:3lENR:3nEMR:4nELQ:4oELP:6oEKP:5PFKP:5PFLn96QFJo96RFJm97TFGl99PFKQ:4kE0U:1gE3Y:LfE5Z:L_ESO1Q1`:L_E:b:E^E;b:Q1000000O101O000O100000000O101O000O10000000000000O10O10000000O10000000000O100000000O10000000000O1000O100O01O010O00100O1O100O1O100O001O1O1OO2O1O1N2000001O000O10O10O010O001O001O001O010O001O1O001O1O001M3M2O2O010000O01000O10O0100O10O0100O10O01O100O01K4K6N101O0010O0000001O001O001O001O001O1O1000000O100kNoE\\OQ:c0QF\\Oo9d0QF[OP:d0RF[Oo9c0RF]On9b0TF]Ol9c0UF\\Ok9c0VF]Oj9c0WF\\Oj9b0XF^Og9b0YF^Oh9a0YF_Of9`0\\F_Oc9b0]F^Oc9b0^F^Oa9b0_F^O`9DTF6`05\\9]OoEN95d0a0Q9\\OVFK93b0f0n8\\OlFI9k0i8[OTHe0k7[OWHd0h7\\OYHe0f7ZO\\He0c7[O_Hd0`7]OaHc0^7^ObHa0]7@dH?[7BeH>[7BfH>X7ChH=W7DjH;V7EjH<T7EmH:T7ElH;T7ElH<S7DnH;R7EnH;R7EnH<Q7DPI;P7EPI;Q7DlGN@>c8DhG4D8d8CeG;F1e8DeG<E1e8CgG=B0g8CgG?AMi8CgG`0_ONi8AiGc0YOOn8^OmGh1T8WNlGi1T8VNjGn1V8oMdGY2\\8fMdG[2^93N0011O0O2O6JN10hNaE2_:KjENW:0RFIo93ZFGf9D[E>W<@oC:b<I_Zg0"}, "label": "the man that is laid out in the air"}]}
{"id": 426903, "image": "COCO_train2014_000000426903.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling , blonde - haired girl sitting on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a smiling , blonde - haired girl sitting on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 97, 98, 99, 112, 113, 114, 127, 128, 129, 141, 142, 143, 144, 156, 157, 158, 159, 172, 173, 174, 189], "bbox": [0.45341232227488154, 0.238765625, 0.6663270142180094, 0.554640625], "iscrowd": 0, "area": 10715.14, "rle": {"size": [640, 422], "counts": "idg31ic07J7H7I7I7J7K4K5K5\\O[Ng^Ok1Wa0XN_^OP2_a0:O1N2O2M2O1N2O1N2O1N2O1N3TOPM[@Z3b?d0M3M3M3M3M3M3M4L2M3N2N110O100O100O010O100O10O010000000O100000001N100000001N3N1O1O1`NiAUMX>h2fBnLi<Q3QCWMn<h2kBaMU=]2dBlM[=S2^BVNa=i1XB`Ng=^1SBkNl=T1mAVOQ>i0eAC[>eNjAn2[>eLPBX3Y>VLRBf3V?M3M4K4dMc_Ok0e`0iNg_OP1i`0\\Nc_Od0^b0J6J6J6Jljg2"}, "label": "a smiling , blonde - haired girl sitting on a motorcycle"}]}
{"id": 426903, "image": "COCO_train2014_000000426903.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with straight blonde hair sitting on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a girl with straight blonde hair sitting on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 97, 98, 99, 112, 113, 114, 127, 128, 129, 141, 142, 143, 144, 156, 157, 158, 159, 172, 173, 174, 189], "bbox": [0.45341232227488154, 0.238765625, 0.6663270142180094, 0.554640625], "iscrowd": 0, "area": 10715.14, "rle": {"size": [640, 422], "counts": "idg31ic07J7H7I7I7J7K4K5K5\\O[Ng^Ok1Wa0XN_^OP2_a0:O1N2O2M2O1N2O1N2O1N2O1N3TOPM[@Z3b?d0M3M3M3M3M3M3M4L2M3N2N110O100O100O010O100O10O010000000O100000001N100000001N3N1O1O1`NiAUMX>h2fBnLi<Q3QCWMn<h2kBaMU=]2dBlM[=S2^BVNa=i1XB`Ng=^1SBkNl=T1mAVOQ>i0eAC[>eNjAn2[>eLPBX3Y>VLRBf3V?M3M4K4dMc_Ok0e`0iNg_OP1i`0\\Nc_Od0^b0J6J6J6Jljg2"}, "label": "a girl with straight blonde hair sitting on a motorcycle"}]}
{"id": 451482, "image": "COCO_train2014_000000451482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book on top of a pile of books\"."}], "task": "refering", "answer_template": "The \"a book on top of a pile of books\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202], "bbox": [0.086640625, 0.08860416666666666, 0.8213437499999999, 0.5058958333333333], "iscrowd": 0, "area": 69563.97045000001, "rle": {"size": [480, 640], "counts": "QPj03o=n0QOo0ROn0TOl0N2N2F:O1O1O2N100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1L4L5K4L40000001O0000001O00001O0000001O00001O0000001O0000001O000000010O000000001O0000001O000000001O000000001O0000001O000000001O000001O01O000000001O0000001O000000001O000000001O0000001O000000001O000001O01O0000001O000000001O000000001O0000001O000000001O000000001O0001O01O000000001O000000001O00000000001O000000001O00000000001O000001O01O00000000001O000000001O00000000001O000000001O00000000001O0001O0001O00000000001O000000001O00000000001O000000001O00000000001O01O000001O00000000001O000000001O00000000001O000000001O000000000001O00000001O000000000000000000000000000000000000000000000000000001O00000000000000000000000O1000000O1000000O100000000O1000000O12N2O1N2N1O2N1O1O001O001O001O001O001O001O001O001O001O001O001O001O0010O01O001O001O1O001O001O1O001O00001O00001O00001O00001O001O00001O00010O00001O00001O001O00001O00001O00001O00001O001O00001O00001O000010O01O00001O9Gd0\\Od0\\Oc0[Of0UOoVe1"}, "label": "a book on top of a pile of books"}]}
{"id": 451482, "image": "COCO_train2014_000000451482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the book titled the poisonwood bible\"."}], "task": "refering", "answer_template": "The \"the book titled the poisonwood bible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202], "bbox": [0.086640625, 0.08860416666666666, 0.8213437499999999, 0.5058958333333333], "iscrowd": 0, "area": 69563.97045000001, "rle": {"size": [480, 640], "counts": "QPj03o=n0QOo0ROn0TOl0N2N2F:O1O1O2N100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1L4L5K4L40000001O0000001O00001O0000001O00001O0000001O0000001O000000010O000000001O0000001O000000001O000000001O0000001O000000001O000001O01O000000001O0000001O000000001O000000001O0000001O000000001O000001O01O0000001O000000001O000000001O0000001O000000001O000000001O0001O01O000000001O000000001O00000000001O000000001O00000000001O000001O01O00000000001O000000001O00000000001O000000001O00000000001O0001O0001O00000000001O000000001O00000000001O000000001O00000000001O01O000001O00000000001O000000001O00000000001O000000001O000000000001O00000001O000000000000000000000000000000000000000000000000000001O00000000000000000000000O1000000O1000000O100000000O1000000O12N2O1N2N1O2N1O1O001O001O001O001O001O001O001O001O001O001O001O001O0010O01O001O001O1O001O001O1O001O00001O00001O00001O00001O001O00001O00010O00001O00001O001O00001O00001O00001O00001O001O00001O00001O000010O01O00001O9Gd0\\Od0\\Oc0[Of0UOoVe1"}, "label": "the book titled the poisonwood bible"}]}
{"id": 451482, "image": "COCO_train2014_000000451482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dare to be square book\"."}], "task": "refering", "answer_template": "The \"dare to be square book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.082578125, 0.7438125, 0.808984375, 0.8763958333333333], "iscrowd": 0, "area": 18542.715450000003, "rle": {"size": [480, 640], "counts": "WVi01o>001O1O1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O0000000000000000O100000000O010000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000000000O100000000000000000000000000000000000000000000000000O10000001O0000000000000000001O000000000000000000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O001O00001O00001O001O00001O001OO1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O1O100O1O10000O0100000000000000000000000000000001O000000000000001O00000000000000001O000000000000001O00000000001O000000001O00000000001O000000001O00000000001O000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1OlYi1"}, "label": "dare to be square book"}]}
{"id": 451482, "image": "COCO_train2014_000000451482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white book underneath two other books\"."}], "task": "refering", "answer_template": "The \"a green and white book underneath two other books\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.082578125, 0.7438125, 0.808984375, 0.8763958333333333], "iscrowd": 0, "area": 18542.715450000003, "rle": {"size": [480, 640], "counts": "WVi01o>001O1O1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O0000000000000000O100000000O010000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000000000O100000000000000000000000000000000000000000000000000O10000001O0000000000000000001O000000000000000000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O001O00001O00001O001O00001O001OO1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O1O100O1O10000O0100000000000000000000000000000001O000000000000001O00000000000000001O000000000000001O00000000001O000000001O00000000001O000000001O00000000001O000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1OlYi1"}, "label": "a green and white book underneath two other books"}]}
{"id": 451482, "image": "COCO_train2014_000000451482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the book called embroidery stitches is good\"."}], "task": "refering", "answer_template": "The \"the book called embroidery stitches is good\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.0, 0.4561875, 1.0, 0.7438125], "iscrowd": 0, "area": 70210.07460000002, "rle": {"size": [480, 640], "counts": "`7a0_>P1oNj0WO9G8G9H0O101O0O101N100O2O000O101N100O2O000O2O0O100O2O000O2O0O101N10000O2O0O101N10001N100O101O0O1010O1O010O001O10O01O010O1O010O001O0000000000000000000000000000000000001O0000000000000000000000000000000000000O2O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000O100000001O00000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000O100000000000001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000000000000O1000000000000000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000LiL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;jL^DV3b;jL^DV3b;jL^DV3b;jL^DV3b;jL^DV3b;jL^DV3b;kL]DU3c;kL]DU3c;kL]DU3c;kL^DT3b;lL^DT3b;lL^DT3b;lL^DT3b;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;5000000O10000O10000O10000O10000O1000000O1000O010O10O10O10O0101N10001N101N100O2O000O2O0O101N101O0O101N100O2O0O2O000O2O0O101N10001N101N100O2O000O2O0O101N101O0O101N100O2O001N100O101N10001N100O101N10000O2O0O101N10000O2O0O100O2O0O101N1N2N2M4M2N2NhG"}, "label": "the book called embroidery stitches is good"}]}
{"id": 451482, "image": "COCO_train2014_000000451482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and blue book that reads : embroidery stitches\"."}], "task": "refering", "answer_template": "The \"white and blue book that reads : embroidery stitches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.0, 0.4561875, 1.0, 0.7438125], "iscrowd": 0, "area": 70210.07460000002, "rle": {"size": [480, 640], "counts": "`7a0_>P1oNj0WO9G8G9H0O101O0O101N100O2O000O101N100O2O000O2O0O100O2O000O2O0O101N10000O2O0O101N10001N100O101O0O1010O1O010O001O10O01O010O1O010O001O0000000000000000000000000000000000001O0000000000000000000000000000000000000O2O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000O100000001O00000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000O100000000000001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000000000000O1000000000000000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000LiL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;iL_DW3a;jL^DV3b;jL^DV3b;jL^DV3b;jL^DV3b;jL^DV3b;jL^DV3b;kL]DU3c;kL]DU3c;kL]DU3c;kL^DT3b;lL^DT3b;lL^DT3b;lL^DT3b;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;mL]DS3c;5000000O10000O10000O10000O10000O1000000O1000O010O10O10O10O0101N10001N101N100O2O000O2O0O101N101O0O101N100O2O0O2O000O2O0O101N10001N101N100O2O000O2O0O101N101O0O101N100O2O001N100O101N10001N100O101N10000O2O0O101N10000O2O0O100O2O0O101N1N2N2M4M2N2NhG"}, "label": "white and blue book that reads : embroidery stitches"}]}
{"id": 394139, "image": "COCO_train2014_000000394139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suv parked on side\"."}], "task": "refering", "answer_template": "The \"black suv parked on side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 294, 295, 297, 298], "bbox": [0.712171875, 0.7146153846153847, 0.9888281250000001, 0.9572944297082229], "iscrowd": 0, "area": 12080.096199999996, "rle": {"size": [377, 640], "counts": "UVX57[;:G9H8M4K4M3L4M2N0O10000O10000O10000O01M3M2M40O10000O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O10000000000]FQN\\9W2O1O2N1O2N1O2N1O2N1O1O00O100O10000O100O10000O100O10000O100O1O001O1O1O1O1O1000000000000001O0000000000000000000000001O000000000000000000000000001O00000000000000000O10001O1O2N101N101N1O2O0O01O001O00001O001O002N3M2N2M4K4L4K6K4L4L5H7G9G:F9GoX2"}, "label": "black suv parked on side"}]}
{"id": 394139, "image": "COCO_train2014_000000394139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car with a yellow license plate\"."}], "task": "refering", "answer_template": "The \"a blue car with a yellow license plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 294, 295, 297, 298], "bbox": [0.712171875, 0.7146153846153847, 0.9888281250000001, 0.9572944297082229], "iscrowd": 0, "area": 12080.096199999996, "rle": {"size": [377, 640], "counts": "UVX57[;:G9H8M4K4M3L4M2N0O10000O10000O10000O01M3M2M40O10000O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O10000000000]FQN\\9W2O1O2N1O2N1O2N1O2N1O1O00O100O10000O100O10000O100O10000O100O1O001O1O1O1O1O1000000000000001O0000000000000000000000001O000000000000000000000000001O00000000000000000O10001O1O2N101N101N1O2O0O01O001O00001O001O002N3M2N2M4K4L4K6K4L4L5H7G9G:F9GoX2"}, "label": "a blue car with a yellow license plate"}]}
{"id": 172957, "image": "COCO_train2014_000000172957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man who is holding a baby and using a keyboard\"."}], "task": "refering", "answer_template": "The \"a man who is holding a baby and using a keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 66, 67, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 90, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 113, 120, 121, 122, 123, 124, 125, 126, 127, 136, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237], "bbox": [0.11368750000000001, 0.0022535211267605635, 0.9528125, 0.7258450704225352], "iscrowd": 0, "area": 65464.15585, "rle": {"size": [426, 640], "counts": "Udn0?j<1N2O2N1N3N1O2M2O1O2M2O0O01000O10000O010000000001N10000V1kN0O2N1O2O0O2N1O101N1O0001O01O0001O01O0001ROeElN[:X1i04L4M4K4L4M3L4L5K4M2M1\\MVMgJk2X5\\MaJe2]5cM]J^2`5kMXJV2f5RNSJo1k5ZNmIh1P6`NjI`1T6iNdIX1Z6PO_IR1_6VOYIk0e6]OUIc0i6Q3M2N2N3MSO[IYKd6a4fI\\KX6b4oI[Ko5e4VJXKh5i4ZJVKd5k4_JSK\\5Q5fJnJR5X5RKeJg4`5cKYJU4l5^15J6K5K7I8H8H9N10000000006J6J6J6J6J6J6J6J6J6J1O1O00000000000000000000001O00000000000000SNbLlI`3T6RMVIR3j6_MaHe2`7b100000000001O0000001O1O1O1O001O1O1O1O1O00O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O2O0O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O1N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1O100O100O100O100O100O100O1O100O100O100O100O1000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O10000001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1a0BM2O2M3M2N3N2M2N3M3N1O2O10O0100O10O010O001O001O0010O01O001O2N100O2N1O1O2K4I7I8H7ImZ<"}, "label": "a man who is holding a baby and using a keyboard"}]}
{"id": 172957, "image": "COCO_train2014_000000172957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a baby using a laptop\"."}], "task": "refering", "answer_template": "The \"a man holding a baby using a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 66, 67, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 90, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 113, 120, 121, 122, 123, 124, 125, 126, 127, 136, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237], "bbox": [0.11368750000000001, 0.0022535211267605635, 0.9528125, 0.7258450704225352], "iscrowd": 0, "area": 65464.15585, "rle": {"size": [426, 640], "counts": "Udn0?j<1N2O2N1N3N1O2M2O1O2M2O0O01000O10000O010000000001N10000V1kN0O2N1O2O0O2N1O101N1O0001O01O0001O01O0001ROeElN[:X1i04L4M4K4L4M3L4L5K4M2M1\\MVMgJk2X5\\MaJe2]5cM]J^2`5kMXJV2f5RNSJo1k5ZNmIh1P6`NjI`1T6iNdIX1Z6PO_IR1_6VOYIk0e6]OUIc0i6Q3M2N2N3MSO[IYKd6a4fI\\KX6b4oI[Ko5e4VJXKh5i4ZJVKd5k4_JSK\\5Q5fJnJR5X5RKeJg4`5cKYJU4l5^15J6K5K7I8H8H9N10000000006J6J6J6J6J6J6J6J6J6J1O1O00000000000000000000001O00000000000000SNbLlI`3T6RMVIR3j6_MaHe2`7b100000000001O0000001O1O1O1O001O1O1O1O1O00O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O2O0O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O1N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1O100O100O100O100O100O100O1O100O100O100O100O1000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O10000001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1a0BM2O2M3M2N3N2M2N3M3N1O2O10O0100O10O010O001O001O0010O01O001O2N100O2N1O1O2K4I7I8H7ImZ<"}, "label": "a man holding a baby using a laptop"}]}
{"id": 574368, "image": "COCO_train2014_000000574368.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a baseball uniform with a brace on his left ankle swinging a bat\"."}], "task": "refering", "answer_template": "The \"a man in a baseball uniform with a brace on his left ankle swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 290, 291, 292, 308, 309, 313, 314, 315, 316, 331, 337, 338], "bbox": [0.386125, 0.042388758782201406, 0.786359375, 0.9547306791569087], "iscrowd": 0, "area": 37002.04209999999, "rle": {"size": [427, 640], "counts": "eZW31Y=5J5L4L3M3YD\\Ob:f0XEDc:>WEKc:8VE0g:1SE4m:NmD6S;Q101O000000000O10000000M3M3L4M3M3M3M3L4M2010O1O010O010O0010O0010O01O01hIlN8T1@\\O9c0_OM94^O;<D]Ok0;VO\\OV1`0jN]O\\1a0cN^O`1`0aN^Oc1?]N@g1>XNAk1=VNAn1<RNCR2;mMDV2cLaLP3Y1<Z2VLoLX3g0`0[3[OeLe0^3XObLh0a3VO_Lj0c3SO]Ln0e3oN[LQ1h3mNXLS1j3jNVLV1m3gNSLZ1o3dNPL\\1R4bNoK^1R4`NnKa1S4^NlKb1V4\\NkKd1V4ZNjKf1X4YNgKh1Z4VNgKi1[4UNeKl1\\4SNcKm1_4QNbKo1_4oMaKR2`4mM_KR2d4lM\\Ko1k4oMVKk1Q5SNoJh1X5WNgJg1]5WNdJh1^5VNbJk1_5SNaJm1a5QN`Jo1a5oM_JQ2i5hMVJY2P6`MPJ`2X6XMiIh2]6nLfIR3d6`L`Ia3U80O0010O01O1O10O01O100O1O100O00100O100O1O10O01O100O2N101N1O101N1O2O0O1O2J5K52O1N3N2M2O2N_HiKn5V4oIlKf15P1m3VMQLh1;n0b3XMTLh1c0k0X3\\MULg1l0h0n2`MVLf1U1e0d2dMXLd1^1b0Y2iMYLc1g1?o1mMZLb1n1>g1oM[La1R2`0Q2_MmKo1W2a0l1^MnKn1Z2d0g1]MoKV11fN^2f2b1\\MoKX1X3\\1h0ZMQLZ1Z3[1e0WMSL_1Y3Z1d0TMTLb1[3Y1`0SMWLc1[3Z1o0fNTOY1j0gNXOY1f0gN]OX1a0hNAX1>hNDW1:iNHW16iNMV11jN1V1MjN6U1IjN:b1WO^Nm0j1hNVN[1h1cNXNa1e1_NZNd1d1[N\\Ni1`1WN`Nm1\\1TNcNP2Z1oMfNU2V1lMiNX2S1hMnNZ2o0gMPO[2n0eMRO\\2m0dMSO^2k0cMTO^2k0bMUO`2i0aMVOa2h0_MYOa2f0`MYOb2e0^M[Oc2d0]M\\Oe2b0\\M]Oe2b0[M^Og2`0ZM_Og2a0XM@i2>XMAj2=VMCk2<UMCn2;SMDn2;RMEP39PMFR39nLGT37lLIU36kLJX33hLL\\31dLO_3NbL1b3K^L4f3IZL7j3EVL;k3DUL<k3DUL;m3DSL<n3CSL<n3CRL<o3DQL<P4CQL<P4BQL>o3BRL<P4CPL=Q4BoK>R4AoK=R4CnK=S4BnK=R4CnK=R4CoK<R4CnK<S4DnK;R4EnK;R4DPL;P4EPL;P4DQL<o3CRL=n3BSL>n3@SL`0o3^OQLa0R4\\OoKd0S4YOnKg0[80000000000O1000001O0000000O2O0000003M4L7IZPi1"}, "label": "a man in a baseball uniform with a brace on his left ankle swinging a bat"}]}
{"id": 574368, "image": "COCO_train2014_000000574368.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man swinging a bat\"."}], "task": "refering", "answer_template": "The \"a man swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 290, 291, 292, 308, 309, 313, 314, 315, 316, 331, 337, 338], "bbox": [0.386125, 0.042388758782201406, 0.786359375, 0.9547306791569087], "iscrowd": 0, "area": 37002.04209999999, "rle": {"size": [427, 640], "counts": "eZW31Y=5J5L4L3M3YD\\Ob:f0XEDc:>WEKc:8VE0g:1SE4m:NmD6S;Q101O000000000O10000000M3M3L4M3M3M3M3L4M2010O1O010O010O0010O0010O01O01hIlN8T1@\\O9c0_OM94^O;<D]Ok0;VO\\OV1`0jN]O\\1a0cN^O`1`0aN^Oc1?]N@g1>XNAk1=VNAn1<RNCR2;mMDV2cLaLP3Y1<Z2VLoLX3g0`0[3[OeLe0^3XObLh0a3VO_Lj0c3SO]Ln0e3oN[LQ1h3mNXLS1j3jNVLV1m3gNSLZ1o3dNPL\\1R4bNoK^1R4`NnKa1S4^NlKb1V4\\NkKd1V4ZNjKf1X4YNgKh1Z4VNgKi1[4UNeKl1\\4SNcKm1_4QNbKo1_4oMaKR2`4mM_KR2d4lM\\Ko1k4oMVKk1Q5SNoJh1X5WNgJg1]5WNdJh1^5VNbJk1_5SNaJm1a5QN`Jo1a5oM_JQ2i5hMVJY2P6`MPJ`2X6XMiIh2]6nLfIR3d6`L`Ia3U80O0010O01O1O10O01O100O1O100O00100O100O1O10O01O100O2N101N1O101N1O2O0O1O2J5K52O1N3N2M2O2N_HiKn5V4oIlKf15P1m3VMQLh1;n0b3XMTLh1c0k0X3\\MULg1l0h0n2`MVLf1U1e0d2dMXLd1^1b0Y2iMYLc1g1?o1mMZLb1n1>g1oM[La1R2`0Q2_MmKo1W2a0l1^MnKn1Z2d0g1]MoKV11fN^2f2b1\\MoKX1X3\\1h0ZMQLZ1Z3[1e0WMSL_1Y3Z1d0TMTLb1[3Y1`0SMWLc1[3Z1o0fNTOY1j0gNXOY1f0gN]OX1a0hNAX1>hNDW1:iNHW16iNMV11jN1V1MjN6U1IjN:b1WO^Nm0j1hNVN[1h1cNXNa1e1_NZNd1d1[N\\Ni1`1WN`Nm1\\1TNcNP2Z1oMfNU2V1lMiNX2S1hMnNZ2o0gMPO[2n0eMRO\\2m0dMSO^2k0cMTO^2k0bMUO`2i0aMVOa2h0_MYOa2f0`MYOb2e0^M[Oc2d0]M\\Oe2b0\\M]Oe2b0[M^Og2`0ZM_Og2a0XM@i2>XMAj2=VMCk2<UMCn2;SMDn2;RMEP39PMFR39nLGT37lLIU36kLJX33hLL\\31dLO_3NbL1b3K^L4f3IZL7j3EVL;k3DUL<k3DUL;m3DSL<n3CSL<n3CRL<o3DQL<P4CQL<P4BQL>o3BRL<P4CPL=Q4BoK>R4AoK=R4CnK=S4BnK=R4CnK=R4CoK<R4CnK<S4DnK;R4EnK;R4DPL;P4EPL;P4DQL<o3CRL=n3BSL>n3@SL`0o3^OQLa0R4\\OoKd0S4YOnKg0[80000000000O1000001O0000000O2O0000003M4L7IZPi1"}, "label": "a man swinging a bat"}]}
{"id": 574368, "image": "COCO_train2014_000000574368.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher with an open mitt\"."}], "task": "refering", "answer_template": "The \"a baseball catcher with an open mitt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 147, 148, 163, 164, 165, 166, 167, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0075, 0.34156908665105384, 0.46628125000000004, 0.9842622950819672], "iscrowd": 0, "area": 38698.645449999996, "rle": {"size": [427, 640], "counts": "j\\28Q=;D<D<E;D7I6K5J6J6K5J5L5J5J6J6K5J5K6J6K5K5M3M2M2O1N2O2N1N2O1O2M2O1O1N2O2N1N2O1N3N1O1O1O2N1O1O1O101N1O1O1O1O101N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O100O2O000O100O100O100O100O100O100O100O100O100O1O2O0OZOlIWJT6i5mIVJR6j5PJUJo5k5RJUJn5j5TJUJk5k5WJTJh5l5ZJSJf5m5[JRJd5n5^JQJl4CYK\\6LQJj4HVKV62QJg4NRKQ69PJd44nJl5`0oIb48kJi5d0nI`4h6bKWI]4i6dKWI[4i6gKVIY4i6iKVIV4j6lKTIT4l6nKSIQ4n6PLQIP4n6RLQIm3o6TLQIk3o6X101O00010O001O1O1\\MdIXN]6g1dIXN]6h1cIWN^6h1cIXN]6g1dIXN\\6i1dIVN]6i1dIVN]6i1dIVN]6i1dIVN]6i1dIVN]6i1dIVN]6h1eIXN[6g1fIXNZ6h1gIWNZ6h1gIWNZ6c0YI[N>Q1Z6d0XIZN?Q1Z6e0WIYN`0Q1Z6f0VIXNa0Q1Z6f0VIYN`0Q1Z6f0VIYN`0P1Z6h0VIXN`0o0[6i0UIXNa0m0[6j0UIYN`0l0\\6k0TIYN`0k0]6l0SIYN`0j0^6m0RIZN?i0_6l0SI[N>h0b6k0PI]N>g0d6k0nH^N>f0f6k0lH_N>e0i6i0jHbN=d0k6i0hHcN=c0m6i0fHdN>a0o6i0cHgN>?P7h0cHiN>=Q7i0aHjN?;S7i0^HlN`06W7m0YHmNa02Z7o0VHoNa0M_7R1PHQOb0Hc7U1lGSOb0Df7X1hGUOb0^Ol7Z1cGWOc0ZOo7]1_GYOc0UOS8`1ZG\\Od0POW8b1UG^On9`0SF_Oo9?RFAP:<QFDR:9nEGW:3iEM]:GjE9[:VOQFj0X;00O10O1000O100000O10O1000000O0100000O1000O10O1000000O0100000O1000O0O2O1N2N101N2N2N101N2O1N101O1O1O0O2O1O001O1O1L3L5L4000O1000000O10001O0O100000000O1000000O100000000O1000001N101O1O001N101O1O001N101O001L4K4M4K4M4_Ol]^4"}, "label": "a baseball catcher with an open mitt"}]}
{"id": 574368, "image": "COCO_train2014_000000574368.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher wearing a red vest with red shin guards\"."}], "task": "refering", "answer_template": "The \"a baseball catcher wearing a red vest with red shin guards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 147, 148, 163, 164, 165, 166, 167, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0075, 0.34156908665105384, 0.46628125000000004, 0.9842622950819672], "iscrowd": 0, "area": 38698.645449999996, "rle": {"size": [427, 640], "counts": "j\\28Q=;D<D<E;D7I6K5J6J6K5J5L5J5J6J6K5J5K6J6K5K5M3M2M2O1N2O2N1N2O1O2M2O1O1N2O2N1N2O1N3N1O1O1O2N1O1O1O101N1O1O1O1O101N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O100O2O000O100O100O100O100O100O100O100O100O100O1O2O0OZOlIWJT6i5mIVJR6j5PJUJo5k5RJUJn5j5TJUJk5k5WJTJh5l5ZJSJf5m5[JRJd5n5^JQJl4CYK\\6LQJj4HVKV62QJg4NRKQ69PJd44nJl5`0oIb48kJi5d0nI`4h6bKWI]4i6dKWI[4i6gKVIY4i6iKVIV4j6lKTIT4l6nKSIQ4n6PLQIP4n6RLQIm3o6TLQIk3o6X101O00010O001O1O1\\MdIXN]6g1dIXN]6h1cIWN^6h1cIXN]6g1dIXN\\6i1dIVN]6i1dIVN]6i1dIVN]6i1dIVN]6i1dIVN]6i1dIVN]6h1eIXN[6g1fIXNZ6h1gIWNZ6h1gIWNZ6c0YI[N>Q1Z6d0XIZN?Q1Z6e0WIYN`0Q1Z6f0VIXNa0Q1Z6f0VIYN`0Q1Z6f0VIYN`0P1Z6h0VIXN`0o0[6i0UIXNa0m0[6j0UIYN`0l0\\6k0TIYN`0k0]6l0SIYN`0j0^6m0RIZN?i0_6l0SI[N>h0b6k0PI]N>g0d6k0nH^N>f0f6k0lH_N>e0i6i0jHbN=d0k6i0hHcN=c0m6i0fHdN>a0o6i0cHgN>?P7h0cHiN>=Q7i0aHjN?;S7i0^HlN`06W7m0YHmNa02Z7o0VHoNa0M_7R1PHQOb0Hc7U1lGSOb0Df7X1hGUOb0^Ol7Z1cGWOc0ZOo7]1_GYOc0UOS8`1ZG\\Od0POW8b1UG^On9`0SF_Oo9?RFAP:<QFDR:9nEGW:3iEM]:GjE9[:VOQFj0X;00O10O1000O100000O10O1000000O0100000O1000O10O1000000O0100000O1000O0O2O1N2N101N2N2N101N2O1N101O1O1O0O2O1O001O1O1L3L5L4000O1000000O10001O0O100000000O1000000O100000000O1000001N101O1O001N101O1O001N101O001L4K4M4K4M4_Ol]^4"}, "label": "a baseball catcher wearing a red vest with red shin guards"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray chair that is not being sat in\"."}], "task": "refering", "answer_template": "The \"a gray chair that is not being sat in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366], "bbox": [0.656046875, 0.6502401746724891, 0.984953125, 0.999235807860262], "iscrowd": 0, "area": 27749.420249999992, "rle": {"size": [458, 640], "counts": "gTl5;o=9F:G:F9G9G9F;F9G9G:F9F:G5K2N2N1N3N1O2N2N1N3N1O2N2O001O00100OO1O1O1O1O1O1O1O1O2N000000000000O1000O1000O1000000O1000000O100000000O1000000O100000000O1000000O10000000000O1000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O00000000001O000000001O000000001O000000001O00000000001O000000001O000000001OI700000000O100000000O10M3J6K5J5K6K5J6J6K4K6J7J6I8J5L4K5L4L4L4L4L5K4L4K5L4L4L4L4L5K4LRU4"}, "label": "a gray chair that is not being sat in"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the grey color single seater sofa infront of the boy with laptop\"."}], "task": "refering", "answer_template": "The \"the grey color single seater sofa infront of the boy with laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366], "bbox": [0.656046875, 0.6502401746724891, 0.984953125, 0.999235807860262], "iscrowd": 0, "area": 27749.420249999992, "rle": {"size": [458, 640], "counts": "gTl5;o=9F:G:F9G9G9F;F9G9G:F9F:G5K2N2N1N3N1O2N2N1N3N1O2N2O001O00100OO1O1O1O1O1O1O1O1O2N000000000000O1000O1000O1000000O1000000O100000000O1000000O100000000O1000000O10000000000O1000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O00000000001O000000001O000000001O000000001O00000000001O000000001O000000001OI700000000O100000000O10M3J6K5J5K6K5J6J6K4K6J7J6I8J5L4K5L4L4L4L4L5K4L4K5L4L4L4L4L5K4LRU4"}, "label": "the grey color single seater sofa infront of the boy with laptop"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beige couch\"."}], "task": "refering", "answer_template": "The \"a beige couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.0016093750000000001, 0.465174672489083, 0.225140625, 0.9191048034934499], "iscrowd": 0, "area": 20990.887649999997, "rle": {"size": [458, 640], "counts": "Ye0n2\\;n2QM1O2O0O2O0O100O1O100O100O100O100O11O1O2N1O1O1O2N1O1O1O2N1O1O2N1O2N2N2N2N2O@gJoGW5Q8PKjGn4V8ZKeGc4[8f0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000ZLcG_1]8`NfG^1Z8aNiG]1W8aNlG_1S8`NPH_1o7`N\\HX1a7iNmHj0R7UO]I>b6AlI2T6M[JEe5:jJXOV5g0YKkNg4T1gK_NZ4_1[4N1N3N2O1N1O2N2N2O0O2N0010O01O0010O01O0010O01O0010O01O0010O01O0K6D;EZcm6"}, "label": "a beige couch"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan couch on the left that has a pillow with circles on it\"."}], "task": "refering", "answer_template": "The \"tan couch on the left that has a pillow with circles on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.0016093750000000001, 0.465174672489083, 0.225140625, 0.9191048034934499], "iscrowd": 0, "area": 20990.887649999997, "rle": {"size": [458, 640], "counts": "Ye0n2\\;n2QM1O2O0O2O0O100O1O100O100O100O100O11O1O2N1O1O1O2N1O1O1O2N1O1O2N1O2N2N2N2N2O@gJoGW5Q8PKjGn4V8ZKeGc4[8f0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000ZLcG_1]8`NfG^1Z8aNiG]1W8aNlG_1S8`NPH_1o7`N\\HX1a7iNmHj0R7UO]I>b6AlI2T6M[JEe5:jJXOV5g0YKkNg4T1gK_NZ4_1[4N1N3N2O1N1O2N2N2O0O2N0010O01O0010O01O0010O01O0010O01O0010O01O0K6D;EZcm6"}, "label": "tan couch on the left that has a pillow with circles on it"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray armchair is being sat on by a boy with a laptop\"."}], "task": "refering", "answer_template": "The \"a gray armchair is being sat on by a boy with a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.15362499999999998, 0.6190829694323144, 0.464171875, 1.0], "iscrowd": 0, "area": 26404.303900000006, "rle": {"size": [458, 640], "counts": "UT\\12n=;N3M2N2N3Me0[On0ROh0YO;D<D=C<E;E:F6J000000000O100000000000000000000001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O001O00001O00001O00001O00001O00001O00001O00001O0000001O00000000000O1000000000000000001O0000000000000000000000000000001O0O1000000000000XLbFd2^9RMPGj2P9TMTGj2l8TMVGl2j8RMXGn2h8QMYGo2g8oL[GQ3e8lL_GS3a8iLcGW3h9000000000001O0000000000001O0000010O0000001O0000001O000000001O0000001O01O0004L3M4L[1eN001O1O1O1O1O1O1O0J8G9F:FVPi4"}, "label": "a gray armchair is being sat on by a boy with a laptop"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey chair with a boy sitting in it\"."}], "task": "refering", "answer_template": "The \"a grey chair with a boy sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.15362499999999998, 0.6190829694323144, 0.464171875, 1.0], "iscrowd": 0, "area": 26404.303900000006, "rle": {"size": [458, 640], "counts": "UT\\12n=;N3M2N2N3Me0[On0ROh0YO;D<D=C<E;E:F6J000000000O100000000000000000000001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O001O00001O00001O00001O00001O00001O00001O00001O0000001O00000000000O1000000000000000001O0000000000000000000000000000001O0O1000000000000XLbFd2^9RMPGj2P9TMTGj2l8TMVGl2j8RMXGn2h8QMYGo2g8oL[GQ3e8lL_GS3a8iLcGW3h9000000000001O0000000000001O0000010O0000001O0000001O000000001O0000001O01O0004L3M4L[1eN001O1O1O1O1O1O1O0J8G9F:FVPi4"}, "label": "a grey chair with a boy sitting in it"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray chair with a boy sitting on it\"."}], "task": "refering", "answer_template": "The \"a gray chair with a boy sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.156109375, 0.6171615720524019, 0.46214062499999997, 0.9865065502183408], "iscrowd": 0, "area": 25711.489199999996, "rle": {"size": [458, 640], "counts": "aP]17R>>B>A?B>B>B>B?A>A?B>B>B5L0000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O0000000^LYFa2h9XM^Fh2b9RMeFm2[9lLlFT3T9fLRGZ3P:O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O4L9G9G9G9G9G9G4L00000000000000000000000WOY_i4"}, "label": "a gray chair with a boy sitting on it"}]}
{"id": 238502, "image": "COCO_train2014_000000238502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gray chair with boy sitting in it\"."}], "task": "refering", "answer_template": "The \"gray chair with boy sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.156109375, 0.6171615720524019, 0.46214062499999997, 0.9865065502183408], "iscrowd": 0, "area": 25711.489199999996, "rle": {"size": [458, 640], "counts": "aP]17R>>B>A?B>B>B>B?A>A?B>B>B5L0000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O0000000^LYFa2h9XM^Fh2b9RMeFm2[9lLlFT3T9fLRGZ3P:O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O4L9G9G9G9G9G9G4L00000000000000000000000WOY_i4"}, "label": "gray chair with boy sitting in it"}]}
{"id": 394151, "image": "COCO_train2014_000000394151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark green suitcase\"."}], "task": "refering", "answer_template": "The \"a dark green suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334], "bbox": [0.302953125, 0.5127634660421545, 0.53996875, 0.9965807962529273], "iscrowd": 0, "area": 25614.5929, "rle": {"size": [427, 640], "counts": "jWa2\\1m;5L2M3M2N3M2O2ZNVOgFS1j8W2[O9L4N2N2N2O100O1000000O1000000O10O1000O1000000O1000000O1000000O1000000O1mNoJlIQ5T6S1000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000f0ZO4L2N1O1O1O2N1O1O1O001O0000001O00001O00001O00001O0000001O00001O001O001O1O001O001O1O001O1O001O001O1O001O0eM`2ZOj0VOi0WO[Zj3"}, "label": "a dark green suitcase"}]}
{"id": 394151, "image": "COCO_train2014_000000394151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green suitcase by the back of a car\"."}], "task": "refering", "answer_template": "The \"a green suitcase by the back of a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334], "bbox": [0.302953125, 0.5127634660421545, 0.53996875, 0.9965807962529273], "iscrowd": 0, "area": 25614.5929, "rle": {"size": [427, 640], "counts": "jWa2\\1m;5L2M3M2N3M2O2ZNVOgFS1j8W2[O9L4N2N2N2O100O1000000O1000000O10O1000O1000000O1000000O1000000O1000000O1mNoJlIQ5T6S1000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000f0ZO4L2N1O1O1O2N1O1O1O001O0000001O00001O00001O00001O0000001O00001O001O001O1O001O001O1O001O1O001O001O1O001O0eM`2ZOj0VOi0WO[Zj3"}, "label": "a green suitcase by the back of a car"}]}
{"id": 189353, "image": "COCO_train2014_000000189353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foot almost on the skateboard in the foreground\"."}], "task": "refering", "answer_template": "The \"the foot almost on the skateboard in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 117, 118, 119, 140, 141, 142, 163, 164], "bbox": [0.0, 0.0026995305164319245, 0.2854375, 0.4717840375586855], "iscrowd": 0, "area": 18315.04075, "rle": {"size": [426, 640], "counts": "=7g<<01O1O1O001O1O1O001O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O2N1O2N2N1O2N1O2N2N3M2hHfM\\3\\2aLeM_3]2]LeMc3]2ZLdMf3^2VLdMj3^2SLcMm3_2oKcMQ4_2lKbMT4`2hKbMX4`2RI^Mk19S5[2mHQNX1Jk5\\3jIjLV6[3`IjL`6P501O000000000000000002N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1N3N1O1O1O1N100O2N1O101N1O1O2N1O1N3N1O2M2O1O2M2O1O2N1O2N1O1O2O0O2N1N2N3M2N3M2N2N3N1N2N3M2N3M2N2N3M2O1N3M2N3N1O1N3N1O2N1N2O2N1N3N1O1N3N1N3N1N3M2O3L3M3M3M3M3M3M3M3L4M4L3M3M3Mk^n5"}, "label": "the foot almost on the skateboard in the foreground"}]}
{"id": 189353, "image": "COCO_train2014_000000189353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black shoe with a white sock lifting off a skateboard at the heel\"."}], "task": "refering", "answer_template": "The \"a black shoe with a white sock lifting off a skateboard at the heel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 117, 118, 119, 140, 141, 142, 163, 164], "bbox": [0.0, 0.0026995305164319245, 0.2854375, 0.4717840375586855], "iscrowd": 0, "area": 18315.04075, "rle": {"size": [426, 640], "counts": "=7g<<01O1O1O001O1O1O001O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O2N1O2N2N1O2N1O2N2N3M2hHfM\\3\\2aLeM_3]2]LeMc3]2ZLdMf3^2VLdMj3^2SLcMm3_2oKcMQ4_2lKbMT4`2hKbMX4`2RI^Mk19S5[2mHQNX1Jk5\\3jIjLV6[3`IjL`6P501O000000000000000002N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1N3N1O1O1O1N100O2N1O101N1O1O2N1O1N3N1O2M2O1O2M2O1O2N1O2N1O1O2O0O2N1N2N3M2N3M2N2N3N1N2N3M2N3M2N2N3M2O1N3M2N3N1O1N3N1O2N1N2O2N1N3N1O1N3N1N3N1N3M2O3L3M3M3M3M3M3M3M3L4M4L3M3M3Mk^n5"}, "label": "a black shoe with a white sock lifting off a skateboard at the heel"}]}
{"id": 189353, "image": "COCO_train2014_000000189353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pushing off with left foot\"."}], "task": "refering", "answer_template": "The \"pushing off with left foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 38, 39, 40, 42, 43, 44, 58, 59, 60, 61, 62, 63, 66, 67, 79, 80, 81, 82, 83, 84, 85, 89, 90, 102, 103, 104, 105, 112, 113, 126, 127, 135, 136, 137, 149, 150, 158, 159, 160], "bbox": [0.46959375000000003, 0.0, 1.0, 0.4681220657276995], "iscrowd": 0, "area": 23748.223249999995, "rle": {"size": [426, 640], "counts": "RZm3h0a<h0YO3L4L4M3L3N3L4L4M3L4M2M4M3M3M3M2N1O1O001O1O001O1O001O1O001O1O001OO1K5L4K5K5M2M4M2N3M3M2N3M2N3M2N3L4M2N3M2N3O001000O010O010O10O10O010O10O010O10O10O010O10O010O10O010O10O10O010O01000O010O0100O0100O0100O010O0100O0100O0100O0100O010O01000O010O010O10O10O010O010O10O1000M2eN_NlFd1T9dN`F_1a9iNTFZ1l9mNjEV1V:n010000O100O10000O100O10000O10000O100O10000O10O10O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2N2N2N2N2N2N2N2N2N2N2N2K5G9H8H8O10000001O00000000001O1O3M2N2N3M2N3M2N2N3M2N3M1O001O0000001O00001O00001O00001O0000000000O1000O10001O7I7I6J7I7I6J7QH_Mg4h2oJ_Mm4g2jJ`MR5g2dJ`MX5f2_JaM]5e2[J]Me5i2RJZMn5l2iIWMW6o2aISM_6P3\\IRMd6Q3WIQMi6h400000000000000000000000000000O1000000000000000004L3M4L3mMcHSN`7Z1RIdNR7i0_ITOe69mIEW6G[J7h5WOiJg0Z5fNWKY1^8O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O001O001O001O0O2O001O1O001O008H9GRJ"}, "label": "pushing off with left foot"}]}
{"id": 189353, "image": "COCO_train2014_000000189353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with with shorts\"."}], "task": "refering", "answer_template": "The \"a person with with shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 38, 39, 40, 42, 43, 44, 58, 59, 60, 61, 62, 63, 66, 67, 79, 80, 81, 82, 83, 84, 85, 89, 90, 102, 103, 104, 105, 112, 113, 126, 127, 135, 136, 137, 149, 150, 158, 159, 160], "bbox": [0.46959375000000003, 0.0, 1.0, 0.4681220657276995], "iscrowd": 0, "area": 23748.223249999995, "rle": {"size": [426, 640], "counts": "RZm3h0a<h0YO3L4L4M3L3N3L4L4M3L4M2M4M3M3M3M2N1O1O001O1O001O1O001O1O001O1O001OO1K5L4K5K5M2M4M2N3M3M2N3M2N3M2N3L4M2N3M2N3O001000O010O010O10O10O010O10O010O10O10O010O10O010O10O010O10O10O010O01000O010O0100O0100O0100O010O0100O0100O0100O0100O010O01000O010O010O10O10O010O010O10O1000M2eN_NlFd1T9dN`F_1a9iNTFZ1l9mNjEV1V:n010000O100O10000O100O10000O10000O100O10000O10O10O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2N2N2N2N2N2N2N2N2N2N2N2K5G9H8H8O10000001O00000000001O1O3M2N2N3M2N3M2N2N3M2N3M1O001O0000001O00001O00001O00001O0000000000O1000O10001O7I7I6J7I7I6J7QH_Mg4h2oJ_Mm4g2jJ`MR5g2dJ`MX5f2_JaM]5e2[J]Me5i2RJZMn5l2iIWMW6o2aISM_6P3\\IRMd6Q3WIQMi6h400000000000000000000000000000O1000000000000000004L3M4L3mMcHSN`7Z1RIdNR7i0_ITOe69mIEW6G[J7h5WOiJg0Z5fNWKY1^8O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O001O001O001O0O2O001O1O001O008H9GRJ"}, "label": "a person with with shorts"}]}
{"id": 418736, "image": "COCO_train2014_000000418736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green leaves against wall\"."}], "task": "refering", "answer_template": "The \"green leaves against wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 71, 72, 93, 94, 95, 96, 97, 116, 117, 119, 120], "bbox": [0.068984375, 0.12983568075117372, 0.262109375, 0.3966666666666667], "iscrowd": 0, "area": 4939.450900000001, "rle": {"size": [426, 640], "counts": "Y]b01X=5kNKmD:n:FRE>i:DVEa0d:A[Ed0_:^O_Eh0\\:YOcEl0W:VOhEo0R:SOmEo0P:ROPFn0P:SOnEn0Q:TOnEk0R:WOmEg0T:[OkEc0W:^OgEa0Z:@fE?Z:EcE9^:L^E2d:2XELi:8SEGn:Z10O2N1O1O110O010O1O010O100O100O10O01O100O100O1O10O0100O1O1O1O1I_DeN`;Z1bDeN_;X1dDgN];W1eDhN\\;V1fDiN[;T1hDkNY;S1`0O1N2O1O1O1O1N2O1O1O10000000000001O000000000000001O0YEWOm8h0SGZOl8d0VG]Oi8a0XGAg8=[GDd8:]GHb86`GK_83bGO]8OeG2Z8LgG6W8IjG9U8EmG<R8BoGa0o7]OSHd0X:2O1N4L4M3L4LH9I62O1N]OdC:[<DhC=W<@lCa0Q<^ORDc0l;[OWDf0f;ZO\\Dg0a;YOaDh0];XOdDi0Y;WOiDj0k;1O00N6I7J5K6JlVT6"}, "label": "green leaves against wall"}]}
{"id": 418736, "image": "COCO_train2014_000000418736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"indoor tree behind the man\"."}], "task": "refering", "answer_template": "The \"indoor tree behind the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 71, 72, 93, 94, 95, 96, 97, 116, 117, 119, 120], "bbox": [0.068984375, 0.12983568075117372, 0.262109375, 0.3966666666666667], "iscrowd": 0, "area": 4939.450900000001, "rle": {"size": [426, 640], "counts": "Y]b01X=5kNKmD:n:FRE>i:DVEa0d:A[Ed0_:^O_Eh0\\:YOcEl0W:VOhEo0R:SOmEo0P:ROPFn0P:SOnEn0Q:TOnEk0R:WOmEg0T:[OkEc0W:^OgEa0Z:@fE?Z:EcE9^:L^E2d:2XELi:8SEGn:Z10O2N1O1O110O010O1O010O100O100O10O01O100O100O1O10O0100O1O1O1O1I_DeN`;Z1bDeN_;X1dDgN];W1eDhN\\;V1fDiN[;T1hDkNY;S1`0O1N2O1O1O1O1N2O1O1O10000000000001O000000000000001O0YEWOm8h0SGZOl8d0VG]Oi8a0XGAg8=[GDd8:]GHb86`GK_83bGO]8OeG2Z8LgG6W8IjG9U8EmG<R8BoGa0o7]OSHd0X:2O1N4L4M3L4LH9I62O1N]OdC:[<DhC=W<@lCa0Q<^ORDc0l;[OWDf0f;ZO\\Dg0a;YOaDh0];XOdDi0Y;WOiDj0k;1O00N6I7J5K6JlVT6"}, "label": "indoor tree behind the man"}]}
{"id": 418736, "image": "COCO_train2014_000000418736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black leather coat\"."}], "task": "refering", "answer_template": "The \"a man in a black leather coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327], "bbox": [0.07179687500000001, 0.17528169014084508, 0.26625, 0.9820187793427231], "iscrowd": 0, "area": 29981.653000000002, "rle": {"size": [426, 640], "counts": "]Zc0a0Z<Q1oNQ1XOh0]Oc0]O;D:G9G8H2N1N3N1O1O1O1N2O2N1O1O1N2O2N1O1O1N2O1O2N1O1O1O1O2M2O1O1O1O1O8He0[Oe0[Oe0\\Od0[O;E100O100O1O100O1O100O1M3N2M3M3M2O2M3N2O100O1O1O100O1O100O1O100000O101O0000001O0000001O0\\GZMd6f2\\I]Ma6d2]I`M`6`2_IdM^6\\2bIgM[6Y2dIkMY6V2eInMW6S2iIPNT6P2kITNR6l1mIXNP6i1nI[No5f1oI]No5d1oI`Nn5a1PJcNm5^1hIoNU6R1iIROT6o0jIUOS6l0kIXOR6i0lI[OQ6f0mI^OP6c0nIAo5`0mIFP6;mIJf5a0VJET5o0hJVOW4j1fK[Na3\\2[LmM\\3X6H8G9G9H8G9C>B=B>C_XS6"}, "label": "a man in a black leather coat"}]}
{"id": 418736, "image": "COCO_train2014_000000418736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black leather coat standing in the livingroom\"."}], "task": "refering", "answer_template": "The \"a man in a black leather coat standing in the livingroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327], "bbox": [0.07179687500000001, 0.17528169014084508, 0.26625, 0.9820187793427231], "iscrowd": 0, "area": 29981.653000000002, "rle": {"size": [426, 640], "counts": "]Zc0a0Z<Q1oNQ1XOh0]Oc0]O;D:G9G8H2N1N3N1O1O1O1N2O2N1O1O1N2O2N1O1O1N2O1O2N1O1O1O1O2M2O1O1O1O1O8He0[Oe0[Oe0\\Od0[O;E100O100O1O100O1O100O1M3N2M3M3M2O2M3N2O100O1O1O100O1O100O1O100000O101O0000001O0000001O0\\GZMd6f2\\I]Ma6d2]I`M`6`2_IdM^6\\2bIgM[6Y2dIkMY6V2eInMW6S2iIPNT6P2kITNR6l1mIXNP6i1nI[No5f1oI]No5d1oI`Nn5a1PJcNm5^1hIoNU6R1iIROT6o0jIUOS6l0kIXOR6i0lI[OQ6f0mI^OP6c0nIAo5`0mIFP6;mIJf5a0VJET5o0hJVOW4j1fK[Na3\\2[LmM\\3X6H8G9G9H8G9C>B=B>C_XS6"}, "label": "a man in a black leather coat standing in the livingroom"}]}
{"id": 418736, "image": "COCO_train2014_000000418736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair closest to the lady\"."}], "task": "refering", "answer_template": "The \"a chair closest to the lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 199, 200, 201, 202, 221, 222, 223, 224, 244, 245, 246, 268, 269], "bbox": [0.61828125, 0.47305164319248827, 0.80190625, 0.7596948356807511], "iscrowd": 0, "area": 7916.104150000004, "rle": {"size": [426, 640], "counts": "gPU56n<7I8L3M^ObC=\\<AfCa0W<]OlCe0Z<3N2N1O2M3N2N1O2N101N101O001O001O001N101O001N1O2O0O1O2J5N3M2N3M2M3M4K4K6K4L5L3I7J7M2N3N101O00000O100000000000001OO2L3L4J6L4M4K4M3N2O2N1O1N2O1O2N1O100O2N1O1O1O1O2O0O100O1O101N1O100O1O101N1O100O1O1O2M2N2M3M3L5K4H8J6M4L3J6JWdd1"}, "label": "a chair closest to the lady"}]}
{"id": 418736, "image": "COCO_train2014_000000418736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair sits behind a coffee table next to a woman in white\"."}], "task": "refering", "answer_template": "The \"a chair sits behind a coffee table next to a woman in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 199, 200, 201, 202, 221, 222, 223, 224, 244, 245, 246, 268, 269], "bbox": [0.61828125, 0.47305164319248827, 0.80190625, 0.7596948356807511], "iscrowd": 0, "area": 7916.104150000004, "rle": {"size": [426, 640], "counts": "gPU56n<7I8L3M^ObC=\\<AfCa0W<]OlCe0Z<3N2N1O2M3N2N1O2N101N101O001O001O001N101O001N1O2O0O1O2J5N3M2N3M2M3M4K4K6K4L5L3I7J7M2N3N101O00000O100000000000001OO2L3L4J6L4M4K4M3N2O2N1O1N2O1O2N1O100O2N1O1O1O1O2O0O100O1O101N1O100O1O101N1O100O1O1O2M2N2M3M3L5K4H8J6M4L3J6JWdd1"}, "label": "a chair sits behind a coffee table next to a woman in white"}]}
{"id": 230321, "image": "COCO_train2014_000000230321.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"matress pink and yellow in color and on the blue spread\"."}], "task": "refering", "answer_template": "The \"matress pink and yellow in color and on the blue spread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.21404687500000003, 0.6005625, 0.9572968750000002, 0.9848333333333332], "iscrowd": 0, "area": 56090.9295, "rle": {"size": [480, 640], "counts": "geP21n>1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100001O0000000000001O0000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O100O100O1O100O100O100O2N100O100O1O100O100O100O1O1001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001N1N2YLgEf2[:XMgEf2Z:YMgEf2Z:YMgEf2[:XMgEf2Z:YMgEf2Z:YMhEd2[:ZMfEe2[:ZMgEd2Z:[MgEd2[:ZMfEe2[:ZMgEd2Z:[MgEc2\\:[MfEc2[:\\MfEc2[:\\MfEc2\\:[MfEc2[:\\MfEc2[:\\MgEa2\\:]MeEb2\\:]MeEb2\\:]MfEa2\\:]MeEb2^;N1N3N1O1O2N1O2N1N2O2N1O2N1O1N3N1O2N1O2N1N3N4L4L4L4L3L5L;E;EYX<"}, "label": "matress pink and yellow in color and on the blue spread"}]}
{"id": 230321, "image": "COCO_train2014_000000230321.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cot with blue bedspray and pink color blanket on it\"."}], "task": "refering", "answer_template": "The \"a cot with blue bedspray and pink color blanket on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.21404687500000003, 0.6005625, 0.9572968750000002, 0.9848333333333332], "iscrowd": 0, "area": 56090.9295, "rle": {"size": [480, 640], "counts": "geP21n>1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100001O0000000000001O0000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O100O100O1O100O100O100O2N100O100O1O100O100O100O1O1001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001N1N2YLgEf2[:XMgEf2Z:YMgEf2Z:YMgEf2[:XMgEf2Z:YMgEf2Z:YMhEd2[:ZMfEe2[:ZMgEd2Z:[MgEd2[:ZMfEe2[:ZMgEd2Z:[MgEc2\\:[MfEc2[:\\MfEc2[:\\MfEc2\\:[MfEc2[:\\MfEc2[:\\MgEa2\\:]MeEb2\\:]MeEb2\\:]MfEa2\\:]MeEb2^;N1N3N1O1O2N1O2N1N2O2N1O2N1O1N3N1O2N1O2N1N3N4L4L4L4L3L5L;E;EYX<"}, "label": "a cot with blue bedspray and pink color blanket on it"}]}
{"id": 230321, "image": "COCO_train2014_000000230321.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bed on the extreme end of the room\"."}], "task": "refering", "answer_template": "The \"bed on the extreme end of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 253, 254], "bbox": [0.005078125, 0.3876875, 0.4029375, 0.7195208333333334], "iscrowd": 0, "area": 18082.879599999997, "rle": {"size": [480, 640], "counts": "Yd1`3_;01O1O1O001O1O1O1O001O1O1O001O1O1N2O001O1O1O001O1O1O1O001O1O1O1O00100O10O010000O10O10O1000O010000O10O10O10000O01000O10O10O10000O01000O10O010000O10O10O10000O01000O10O10O10000O01000O10O10O10000O01000O10O010000O10O10O10000O01000O10O10O10000O01000O10O10O10000O01000O100O01000O10O10O10000O01000O10O10O10000O01000O1000O010000O01000O100O01000O10O10O1000O00100O00100O00100O00100O0101O0O2O000O2O000O2O0O2O000O2O001N10001N101O0O101O0O2O000O2O001N10001N101O0O101N101O0O101O0O2O000O2O001N10001N10001N101O1N4M3M4K4M3M3L5Lfjb5"}, "label": "bed on the extreme end of the room"}]}
{"id": 230321, "image": "COCO_train2014_000000230321.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a floor mattress in the middle of the room with a blue flower blanket and a brown flower blanket\"."}], "task": "refering", "answer_template": "The \"a floor mattress in the middle of the room with a blue flower blanket and a brown flower blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 369, 370, 371, 372], "bbox": [0.0016875000000000002, 0.478, 0.5925468749999999, 0.9881458333333334], "iscrowd": 0, "area": 49178.38749999999, "rle": {"size": [480, 640], "counts": "mi01l>5L3M3L4M4M2N2O0O1O010O00100O001O10O04L3M3RCoNh;T1RDPOn;Q1mCSOR<P1iCTOU<o0eCUOZ<m0aCWO^<f1O2O0mCaMd;a2WDlM_;U2^DVNY;R3F2N2N101N1O2`E^L_9c3`F^L_9d3_F]La9c3^F^Lb9c3\\F^Lc9c3\\F^Ld9c3YF_Lg9b3WF_Lh9]40000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10O010000O10000O10000O10000O10000O1000O010O1O010O1O010O10O01O10O01O10O01O10O01O10O0100O00100O00100O00100O001O1O010O001O1O001O10O01O1O010000O1000O011O0000000O1000001O000O1O1O1O1O2N001N101O1O001O010O010O10O10O010O010O10O10O010O010O10O01O010O001O00100O001O0010O01O1O010O0100O01000O010O10O01000O0100O010O10O10O10O010O01000O0100O010O10O1O010O001O010O1O010O1O2K4M4K4M4L3L5L4K4M4L301N101O0O2O0O2O0O100O01O010O0010O01O010O1O010O0010O01O010O0010O100O2O000O2O0O2O0O2O0O101N101N100O2O0O2O001N100O2O0O2O0O2O0O101N101N101O0O101N101N101N100O2O0O2O0O2O0O101O0O2O0O2O0O101N1oNcB?^=^OeB`0\\=]OgBc0Z=YOjBe0W=XOmBg0S=VOQCh0e=O0O2O1O1NXRj3"}, "label": "a floor mattress in the middle of the room with a blue flower blanket and a brown flower blanket"}]}
{"id": 230321, "image": "COCO_train2014_000000230321.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mattress with a blue top blanket that is between two other mattresses\"."}], "task": "refering", "answer_template": "The \"a mattress with a blue top blanket that is between two other mattresses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 369, 370, 371, 372], "bbox": [0.0016875000000000002, 0.478, 0.5925468749999999, 0.9881458333333334], "iscrowd": 0, "area": 49178.38749999999, "rle": {"size": [480, 640], "counts": "mi01l>5L3M3L4M4M2N2O0O1O010O00100O001O10O04L3M3RCoNh;T1RDPOn;Q1mCSOR<P1iCTOU<o0eCUOZ<m0aCWO^<f1O2O0mCaMd;a2WDlM_;U2^DVNY;R3F2N2N101N1O2`E^L_9c3`F^L_9d3_F]La9c3^F^Lb9c3\\F^Lc9c3\\F^Ld9c3YF_Lg9b3WF_Lh9]40000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10O010000O10000O10000O10000O10000O1000O010O1O010O1O010O10O01O10O01O10O01O10O01O10O0100O00100O00100O00100O001O1O010O001O1O001O10O01O1O010000O1000O011O0000000O1000001O000O1O1O1O1O2N001N101O1O001O010O010O10O10O010O010O10O10O010O010O10O01O010O001O00100O001O0010O01O1O010O0100O01000O010O10O01000O0100O010O10O10O10O010O01000O0100O010O10O1O010O001O010O1O010O1O2K4M4K4M4L3L5L4K4M4L301N101O0O2O0O2O0O100O01O010O0010O01O010O1O010O0010O01O010O0010O100O2O000O2O0O2O0O2O0O101N101N100O2O0O2O001N100O2O0O2O0O2O0O101N101N101O0O101N101N101N100O2O0O2O0O2O0O101O0O2O0O2O0O101N1oNcB?^=^OeB`0\\=]OgBc0Z=YOjBe0W=XOmBg0S=VOQCh0e=O0O2O1O1NXRj3"}, "label": "a mattress with a blue top blanket that is between two other mattresses"}]}
{"id": 140209, "image": "COCO_train2014_000000140209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person standing behind batter\"."}], "task": "refering", "answer_template": "The \"person standing behind batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 29, 30, 32, 33, 50, 51, 68, 86, 104, 122], "bbox": [0.6530409356725146, 0.0, 0.8471150097465887, 0.311109375], "iscrowd": 0, "area": 5923.341350000002, "rle": {"size": [640, 513], "counts": "P\\a6Y1gb0O100O100O100O2O0O100O100O100O100O100O100O101N100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O101N1000000001O00001O00001^^OG[?9a@=m>Co@T1^>mN\\Ak1R>TNjAc2c=^MXBW3W=iLeBk3k<ULPCU4k<lKPCW4Q=iKkBZ4V=hKdB\\4\\=hK[B\\4f=gKSB]4m=j01O100O1O100O2O0O1O100O1O1TMaCWN_<U1`3E;E;D<E;E;DVh`1"}, "label": "person standing behind batter"}]}
{"id": 140209, "image": "COCO_train2014_000000140209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the legs of a player in the back\"."}], "task": "refering", "answer_template": "The \"the legs of a player in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 29, 30, 32, 33, 50, 51, 68, 86, 104, 122], "bbox": [0.6530409356725146, 0.0, 0.8471150097465887, 0.311109375], "iscrowd": 0, "area": 5923.341350000002, "rle": {"size": [640, 513], "counts": "P\\a6Y1gb0O100O100O100O2O0O100O100O100O100O100O100O101N100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O101N1000000001O00001O00001^^OG[?9a@=m>Co@T1^>mN\\Ak1R>TNjAc2c=^MXBW3W=iLeBk3k<ULPCU4k<lKPCW4Q=iKkBZ4V=hKdB\\4\\=hK[B\\4f=gKSB]4m=j01O100O1O100O2O0O1O100O1O1TMaCWN_<U1`3E;E;D<E;E;DVh`1"}, "label": "the legs of a player in the back"}]}
{"id": 312247, "image": "COCO_train2014_000000312247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden dining table chair\"."}], "task": "refering", "answer_template": "The \"a wooden dining table chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 274, 275], "bbox": [0.8192656250000001, 0.2568958333333333, 1.0, 0.714625], "iscrowd": 0, "area": 17268.020699999994, "rle": {"size": [480, 640], "counts": "ome71f>9G9H8SNWOZES1_:SOYEU1b:POUEY1g:lNoD^1o:cNiDe1W;\\NeDg1Z;ZNcDj1[;XNaDk1_;UN_Dm1a;i00N2K6J5JM7B`0@`0]NbLTGn3m8R101O00000000000010O000000000001O01O0000000001O000001O000001O00000001O0001O0000000000010O0000000000001TG_JW8a5`GkJ]8P6L1O00001O0000001O000010O000001O00E;A?G9H8G9G9G9G9J6H8G9H8G9G9H8G9I7K6I6J6J6J6K5L4LQL"}, "label": "a wooden dining table chair"}]}
{"id": 312247, "image": "COCO_train2014_000000312247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wood chair on the right next to another just like it\"."}], "task": "refering", "answer_template": "The \"wood chair on the right next to another just like it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 274, 275], "bbox": [0.8192656250000001, 0.2568958333333333, 1.0, 0.714625], "iscrowd": 0, "area": 17268.020699999994, "rle": {"size": [480, 640], "counts": "ome71f>9G9H8SNWOZES1_:SOYEU1b:POUEY1g:lNoD^1o:cNiDe1W;\\NeDg1Z;ZNcDj1[;XNaDk1_;UN_Dm1a;i00N2K6J5JM7B`0@`0]NbLTGn3m8R101O00000000000010O000000000001O01O0000000001O000001O000001O00000001O0001O0000000000010O0000000000001TG_JW8a5`GkJ]8P6L1O00001O0000001O000010O000001O00E;A?G9H8G9G9G9G9J6H8G9H8G9G9H8G9I7K6I6J6J6J6K5L4LQL"}, "label": "wood chair on the right next to another just like it"}]}
{"id": 312247, "image": "COCO_train2014_000000312247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man using a laptop\"."}], "task": "refering", "answer_template": "The \"a man using a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 31, 32, 33, 34, 35, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 307, 308, 309, 330], "bbox": [0.117453125, 0.0270625, 0.634234375, 0.8950625], "iscrowd": 0, "area": 62788.03089999999, "rle": {"size": [480, 640], "counts": "S]S12k>7I7I7I7I7I7I7I7I7I7I5K4L3N3L3M4L4L3M4L3M4M2M4M2N3M2N3M3M2N3M2N3L3N2N3M2N2N2N3M2N2M4M2N2N2N3M2N2N2N3N1O1N3N1O1N2O2N1O1N2O2N1N2O2N1N2O1O2M2O1O1O2M2O101N1O100O2O0O100O2N100O2O0O1O101N100O101N1O101N100O1O2O0O100O2O0O1O2O0O100O2O0O10001O000O2O00000O2O00000O2O0000001N100O2O000O101N10000O2O0O101N10000SL_GZ1c8cN^G^1b8_N`Ga1h9VMYFk2m9mLTFT3m9iLTFW3o9eLRF\\3o9aLRF_3P:^LQFc3P:YLRFg3Q:ULQFi3R:TLoEk3c:O100O1O1O1O100O1O1O100O1O10000001O00001O00001O00001O00001O00001UIXL]2h3aM`LY2`3eMgLW2Y3eMPMU2P3iMWMR2i2lM_Mn1b2oMeMl1[2RNmMh1T2VNRNe1n1]NUN^1k1cNZNW1g1iN]NR1d1oN^Nm0c1SO_Nj0b1WO^Ng0b1ZO_Nd0b1]O^Na0c1_O^N?c1A^N=b1E^N9c1G^N7c1J^N3c1M^N1c1O^NOb10aNN`11bNM_12cNL^12eNL[14gNJZ14iNJX15kNHV17lNGU17nNGR19POEQ1:QODP1:SODn0;TOCm0<UOCj0<YOBh0=ZOAg0>[O@f0>]O@c0`0@]Oa0b0A\\O`0b0B]O`0c0_O\\Oc0c0]O\\Oe0c0[O\\Og0c0YO\\Oh0d0XO[Oj0d0VO[Ol0d0UOZOm0e0SOZOo0e0QOZOQ1c0QO\\OQ1=UOBm08XOGj03[OLf03[OLg02[OLg02ZOMh01YONi00XOOj0OWO0k0NVO1l0MUO2m0JVO5k0GYO8i0B]O<e0A]O>e0A[O>g0AYO>i0AWO>k0AUO?l0@TO?n0@RO?o0ARO=P1BPO>Q1AoN?R1@nN`0S1_OmNa0T1^OlNb0U1]OkNc0V1\\OjNd0`1ROaNm0i700000002N2L5L3L4M2M201O001O001O00001O001O001O00001O010O00001O001O010O00001O001O0010O01O001O0N3L3N3M2O2N1N3Nm\\]3"}, "label": "a man using a laptop"}]}
{"id": 312247, "image": "COCO_train2014_000000312247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing yellow line shirt & blue jeans using laptop\"."}], "task": "refering", "answer_template": "The \"a man wearing yellow line shirt & blue jeans using laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 31, 32, 33, 34, 35, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 307, 308, 309, 330], "bbox": [0.117453125, 0.0270625, 0.634234375, 0.8950625], "iscrowd": 0, "area": 62788.03089999999, "rle": {"size": [480, 640], "counts": "S]S12k>7I7I7I7I7I7I7I7I7I7I5K4L3N3L3M4L4L3M4L3M4M2M4M2N3M2N3M3M2N3M2N3L3N2N3M2N2N2N3M2N2M4M2N2N2N3M2N2N2N3N1O1N3N1O1N2O2N1O1N2O2N1N2O2N1N2O1O2M2O1O1O2M2O101N1O100O2O0O100O2N100O2O0O1O101N100O101N1O101N100O1O2O0O100O2O0O1O2O0O100O2O0O10001O000O2O00000O2O00000O2O0000001N100O2O000O101N10000O2O0O101N10000SL_GZ1c8cN^G^1b8_N`Ga1h9VMYFk2m9mLTFT3m9iLTFW3o9eLRF\\3o9aLRF_3P:^LQFc3P:YLRFg3Q:ULQFi3R:TLoEk3c:O100O1O1O1O100O1O1O100O1O10000001O00001O00001O00001O00001O00001UIXL]2h3aM`LY2`3eMgLW2Y3eMPMU2P3iMWMR2i2lM_Mn1b2oMeMl1[2RNmMh1T2VNRNe1n1]NUN^1k1cNZNW1g1iN]NR1d1oN^Nm0c1SO_Nj0b1WO^Ng0b1ZO_Nd0b1]O^Na0c1_O^N?c1A^N=b1E^N9c1G^N7c1J^N3c1M^N1c1O^NOb10aNN`11bNM_12cNL^12eNL[14gNJZ14iNJX15kNHV17lNGU17nNGR19POEQ1:QODP1:SODn0;TOCm0<UOCj0<YOBh0=ZOAg0>[O@f0>]O@c0`0@]Oa0b0A\\O`0b0B]O`0c0_O\\Oc0c0]O\\Oe0c0[O\\Og0c0YO\\Oh0d0XO[Oj0d0VO[Ol0d0UOZOm0e0SOZOo0e0QOZOQ1c0QO\\OQ1=UOBm08XOGj03[OLf03[OLg02[OLg02ZOMh01YONi00XOOj0OWO0k0NVO1l0MUO2m0JVO5k0GYO8i0B]O<e0A]O>e0A[O>g0AYO>i0AWO>k0AUO?l0@TO?n0@RO?o0ARO=P1BPO>Q1AoN?R1@nN`0S1_OmNa0T1^OlNb0U1]OkNc0V1\\OjNd0`1ROaNm0i700000002N2L5L3L4M2M201O001O001O00001O001O001O00001O010O00001O001O010O00001O001O0010O01O001O0N3L3N3M2O2N1N3Nm\\]3"}, "label": "a man wearing yellow line shirt & blue jeans using laptop"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right computer in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the right computer in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270], "bbox": [0.5060156250000001, 0.3142857142857143, 0.790234375, 0.7714285714285714], "iscrowd": 0, "area": 28701.671599999998, "rle": {"size": [427, 640], "counts": "hZW4>_<d0[Oe0\\Oc0]Od0[Oe0\\Od0\\Od0H7K6J6J6J5K6J4MO000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O1000001O0]MgHTOY7<YIBg6MlI1T6^O_J`0a5oNSKn0m4bNeK\\1[4SNXLk1f7N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2Nkag1"}, "label": "the right computer in the right hand picture"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the computer on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270], "bbox": [0.5060156250000001, 0.3142857142857143, 0.790234375, 0.7714285714285714], "iscrowd": 0, "area": 28701.671599999998, "rle": {"size": [427, 640], "counts": "hZW4>_<d0[Oe0\\Oc0]Od0[Oe0\\Od0\\Od0H7K6J6J6J5K6J4MO000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O1000001O0]MgHTOY7<YIBg6MlI1T6^O_J`0a5oNSKn0m4bNeK\\1[4SNXLk1f7N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2Nkag1"}, "label": "the computer on the right in the right hand picture"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color laptop near by a small white color laptop\"."}], "task": "refering", "answer_template": "The \"a black color laptop near by a small white color laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308], "bbox": [0.11234375, 0.19201405152224824, 0.5392812499999999, 0.9407259953161592], "iscrowd": 0, "area": 65337.800049999976, "rle": {"size": [427, 640], "counts": "bSn04W=:F9G9G:F9G9G9F;F9G9G:`KoLRNZ3P1]MnNl24jMK_2WOWNh0S2YNdNe1e1]MQOb2X1_L_O`3V5000000000O100000000O1000000O100000O10O100000000O100000000O1000000O100000000O100000000O1000O1000O1000000O100000000O100000000O100000000O1000000O1000O1000O100000000O100000000O100000000O100000000000000O10O1000000000O1000000000000O100000000000000O1000000000000O010000000000000O1000000000000O1000000000000O100000000000O10O1000000000000O100000000000000O1000000000000O10000000O1000O100000000000000O100000b1^NW2iMU1kN2M3N2N2N2N2N3M2N2M3N2N2N2N2N2N2M3N2N3M2N2N2N2M3N2N2N2N2N2N2N3L3N2N2N2N2N2N2M3N2N2N3M2N2N2N2M3N2N2N2N2L4K5Laej3"}, "label": "a black color laptop near by a small white color laptop"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black laptop with light blue screen\"."}], "task": "refering", "answer_template": "The \"black laptop with light blue screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308], "bbox": [0.11234375, 0.19201405152224824, 0.5392812499999999, 0.9407259953161592], "iscrowd": 0, "area": 65337.800049999976, "rle": {"size": [427, 640], "counts": "bSn04W=:F9G9G:F9G9G9F;F9G9G:`KoLRNZ3P1]MnNl24jMK_2WOWNh0S2YNdNe1e1]MQOb2X1_L_O`3V5000000000O100000000O1000000O100000O10O100000000O100000000O1000000O100000000O100000000O1000O1000O1000000O100000000O100000000O100000000O1000000O1000O1000O100000000O100000000O100000000O100000000000000O10O1000000000O1000000000000O100000000000000O1000000000000O010000000000000O1000000000000O1000000000000O100000000000O10O1000000000000O100000000000000O1000000000000O10000000O1000O100000000000000O100000b1^NW2iMU1kN2M3N2N2N2N2N3M2N2M3N2N2N2N2N2N2M3N2N3M2N2N2N2M3N2N2N2N2N2N2N3L3N2N2N2N2N2N2M3N2N2N3M2N2N2N2M3N2N2N2N2L4K5Laej3"}, "label": "black laptop with light blue screen"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a grey color laptop beside black colored laptop\"."}], "task": "refering", "answer_template": "The \"there is a grey color laptop beside black colored laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271], "bbox": [0.50753125, 0.5915222482435597, 0.79340625, 0.7739344262295083], "iscrowd": 0, "area": 12054.4516, "rle": {"size": [427, 640], "counts": "dhW44W=6J6J6J6J6J6J6J6J6J7I6J6J00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000001O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2NUgf1"}, "label": "there is a grey color laptop beside black colored laptop"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard of the black laptop\"."}], "task": "refering", "answer_template": "The \"the keyboard of the black laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305], "bbox": [0.12471874999999999, 0.5732084309133489, 0.5289375, 0.9065339578454333], "iscrowd": 0, "area": 26791.788150000004, "rle": {"size": [427, 640], "counts": "]gQ16m<8H8G8I8H8H8H8G9H8H7I8G9H8H8J6000O100000000000O10O1000000000000O10000000O1000O1000000000000O1000O10000000O1000000000000O10O1000000000O100000000000O01000000000000O1000000000O1000O1000000000000O100000O100000O1000000000000O10O1000000000O1000000000000O0100000000000O1000000000O10O1000000000000O10000000O100000O1000000000000O1000O10000000O1000000000000O0100000000000O100000000000O01000000000000O10000000O1000O1000000000000O3N2N2N3M2N2N2N2M3N2N3M2N2N2N2M3N2N3M2N2N2N2M3N3M2N2N2N2N2M3N3M2N2N2N2N2M3N3M2N2N2N2N2M4M2N2N2N2N2NaUm3"}, "label": "the keyboard of the black laptop"}]}
{"id": 295864, "image": "COCO_train2014_000000295864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard of a large laptop next to a smaller laptop\"."}], "task": "refering", "answer_template": "The \"the keyboard of a large laptop next to a smaller laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305], "bbox": [0.12471874999999999, 0.5732084309133489, 0.5289375, 0.9065339578454333], "iscrowd": 0, "area": 26791.788150000004, "rle": {"size": [427, 640], "counts": "]gQ16m<8H8G8I8H8H8H8G9H8H7I8G9H8H8J6000O100000000000O10O1000000000000O10000000O1000O1000000000000O1000O10000000O1000000000000O10O1000000000O100000000000O01000000000000O1000000000O1000O1000000000000O100000O100000O1000000000000O10O1000000000O1000000000000O0100000000000O1000000000O10O1000000000000O10000000O100000O1000000000000O1000O10000000O1000000000000O0100000000000O100000000000O01000000000000O10000000O1000O1000000000000O3N2N2N3M2N2N2N2M3N2N3M2N2N2N2M3N2N3M2N2N2N2M3N3M2N2N2N2N2M3N3M2N2N2N2N2M3N3M2N2N2N2N2M4M2N2N2N2N2NaUm3"}, "label": "the keyboard of a large laptop next to a smaller laptop"}]}
{"id": 394172, "image": "COCO_train2014_000000394172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the salad with a fork in it\"."}], "task": "refering", "answer_template": "The \"the salad with a fork in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 183, 184, 185, 186, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 203, 204, 205, 206], "bbox": [0.04896, 0.4308, 0.9001866666666667, 0.87146], "iscrowd": 0, "area": 51650.933549999994, "rle": {"size": [500, 375], "counts": "fT91[?;F:E;F:E<E:F:L4L4L4L4L5K4L4L4K5L4L4L2N1O001O001O1O001O1O001O0O2O1O001O001O1O002N3L4M3M3M3M3M3M3M3M1O1O1N101O001O1O001O001N2O001O001O1O0O2O001O1O1O1O001N2O1O1O010O10000O10O10O10000O01000O10000O010O10000O01000O1000O010000OH9G9G801O1O1O1O1O1O1O1O100001N101O001O001O1O001O00O1N2N2N101O1O1O1O1O1O1O1O1O1O1O001O1O001O1O001>B1N2O001N20O01O100O001O10O01O1O010O1O00100O1O0010O01O10O01O00100O001O10O01O00100O001O010O1O0010O01O1O010O1O010O00100O010O1O010O010O1O010O10O00001O0010O01O00001O001O000010O01O00001O001O001O01O01O1O001O001O1O010O001O1O001O001O1O001O0O2O1O001O001O1N101O001O1O001N2O1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O1O2N1O1O4H7I8G9G9G8H9F:G8H9GQha0"}, "label": "the salad with a fork in it"}]}
{"id": 394172, "image": "COCO_train2014_000000394172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate closest to the pepper\"."}], "task": "refering", "answer_template": "The \"a plate closest to the pepper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 183, 184, 185, 186, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 203, 204, 205, 206], "bbox": [0.04896, 0.4308, 0.9001866666666667, 0.87146], "iscrowd": 0, "area": 51650.933549999994, "rle": {"size": [500, 375], "counts": "fT91[?;F:E;F:E<E:F:L4L4L4L4L5K4L4L4K5L4L4L2N1O001O001O1O001O1O001O0O2O1O001O001O1O002N3L4M3M3M3M3M3M3M3M1O1O1N101O001O1O001O001N2O001O001O1O0O2O001O1O1O1O001N2O1O1O010O10000O10O10O10000O01000O10000O010O10000O01000O1000O010000OH9G9G801O1O1O1O1O1O1O1O100001N101O001O001O1O001O00O1N2N2N101O1O1O1O1O1O1O1O1O1O1O001O1O001O1O001>B1N2O001N20O01O100O001O10O01O1O010O1O00100O1O0010O01O10O01O00100O001O10O01O00100O001O010O1O0010O01O1O010O1O010O00100O010O1O010O010O1O010O10O00001O0010O01O00001O001O000010O01O00001O001O001O01O01O1O001O001O1O010O001O1O001O001O1O001O0O2O1O001O001O1N101O001O1O001N2O1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O1O2N1O1O4H7I8G9G9G8H9F:G8H9GQha0"}, "label": "a plate closest to the pepper"}]}
{"id": 394172, "image": "COCO_train2014_000000394172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork on the plate with the salad\"."}], "task": "refering", "answer_template": "The \"the fork on the plate with the salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220], "bbox": [0.32722666666666667, 0.77428, 1.0, 0.94786], "iscrowd": 0, "area": 4507.71405, "rle": {"size": [500, 375], "counts": "c^l11b?200O1O010O1HLo@4o>MQA3P?Mo@4j>LTA123h>MWAO14h>NVAO23g>NWA012a>L`A2NO14a>K_A2O012a>M_A1OO14a>K_A2O012a>M_A0O202a>L`A1O003b>L^A10102a>M_A00004`>L`A1O003b>L_A0O113`>L`A1OO14`>M`AOO113`>M_A00O14`>M_A00004a>L_A0O023`>M_A1OO14`>M`AOO023`>M_A1N023`>M_A00O15_>M`AOOO34_>M_A00O15_>M`AOOO34^>N`AOO023`>N_AOOO25_>M`AOO023`>N^AO1O15_>M`AOOO34^>O_AN0024_>M`AOOO25_>M`AOO023_>O`AM0015_>M`AOO023_>O_AN1O15^>NaAM0104`>M`AOO024^>NaAM0015_>M`AN0113_>O`AM0015^>NaAM0104`>M`AN0114^>NaAM0104_>NaAM0=`>EaANO<`>G`AM0<`>GaALO=a>G_A?a>A^A`0b>91O001O000010O0001O001O0O2N100O2O0O2N101N100O2O001O1N200O1O1O1O00000001O01O00000000000010O0000000001O0001O00000001O0001O0000010O0000000010O000001O0000010O0000000010O000001O0000010O0000000010O000001O0000010O00001O0001O01O00001O0001O01O00001O01O0001O00000010O0001O0000010O00001O0010O0001O00010O000010O0001O0010O0001O00010O00001O010O000010O0001O00010O00001O010O00001O01O01O00010O00010nA"}, "label": "the fork on the plate with the salad"}]}
{"id": 394172, "image": "COCO_train2014_000000394172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver fork sitting on plate bottom of screen\"."}], "task": "refering", "answer_template": "The \"silver fork sitting on plate bottom of screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220], "bbox": [0.32722666666666667, 0.77428, 1.0, 0.94786], "iscrowd": 0, "area": 4507.71405, "rle": {"size": [500, 375], "counts": "c^l11b?200O1O010O1HLo@4o>MQA3P?Mo@4j>LTA123h>MWAO14h>NVAO23g>NWA012a>L`A2NO14a>K_A2O012a>M_A1OO14a>K_A2O012a>M_A0O202a>L`A1O003b>L^A10102a>M_A00004`>L`A1O003b>L_A0O113`>L`A1OO14`>M`AOO113`>M_A00O14`>M_A00004a>L_A0O023`>M_A1OO14`>M`AOO023`>M_A1N023`>M_A00O15_>M`AOOO34_>M_A00O15_>M`AOOO34^>N`AOO023`>N_AOOO25_>M`AOO023`>N^AO1O15_>M`AOOO34^>O_AN0024_>M`AOOO25_>M`AOO023_>O`AM0015_>M`AOO023_>O_AN1O15^>NaAM0104`>M`AOO024^>NaAM0015_>M`AN0113_>O`AM0015^>NaAM0104`>M`AN0114^>NaAM0104_>NaAM0=`>EaANO<`>G`AM0<`>GaALO=a>G_A?a>A^A`0b>91O001O000010O0001O001O0O2N100O2O0O2N101N100O2O001O1N200O1O1O1O00000001O01O00000000000010O0000000001O0001O00000001O0001O0000010O0000000010O000001O0000010O0000000010O000001O0000010O0000000010O000001O0000010O00001O0001O01O00001O0001O01O00001O01O0001O00000010O0001O0000010O00001O0010O0001O00010O000010O0001O0010O0001O00010O00001O010O000010O0001O00010O00001O010O00001O01O01O00010O00010nA"}, "label": "silver fork sitting on plate bottom of screen"}]}
{"id": 279485, "image": "COCO_train2014_000000279485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a blue hat talking on her cell phone\"."}], "task": "refering", "answer_template": "The \"a woman with a blue hat talking on her cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 210, 211, 233, 234, 256, 257, 278, 279, 280, 301, 302, 303], "bbox": [0.09425, 0.14573770491803278, 0.24532812499999998, 0.8789227166276347], "iscrowd": 0, "area": 19359.330099999996, "rle": {"size": [427, 640], "counts": "\\Ui07n<7I7K5J6J5UJnNYOW1e0jNXOZ1g0gNVO\\1aLeNR20\\1]1ZLnNV2F]1^1ULXOZ2[O_1^1PLB\\2QOb1_1nKH\\2jNd1_1nKK[2gN_1f1TLF[2eN\\1j1WLD[2cNY1n1ZLB[2aNV1R2]L_O\\2`NS1U2_L^O\\2^NR1W2`L^O]2[NP1Z2aL^O]2YNo0\\2bL]O^2XNo0\\2aL_O^2VNP1\\2`LA]2UNR1[2_LCY2WNW1W2^LEU2YN\\1S2[LJR2XNa1P2YLNP2VNg1R5gL^LY3j6O100000O100O001O1O1O12M3N2M2O1N11O1O1O1N2O1O1O1O0O2O1O1O1O1N2O1O1O1O1N2O1O001O2I6A?A?A?B>A?A`0D;J7I7I7I6J7K5L3M4L4mM]GWOg8e0`GUOe8e0bGUOh8`0Z2F6K4L5L4K4M6I`XY6"}, "label": "a woman with a blue hat talking on her cell phone"}]}
{"id": 279485, "image": "COCO_train2014_000000279485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a bright blue hat wearing black winter clothes\"."}], "task": "refering", "answer_template": "The \"a woman in a bright blue hat wearing black winter clothes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 210, 211, 233, 234, 256, 257, 278, 279, 280, 301, 302, 303], "bbox": [0.09425, 0.14573770491803278, 0.24532812499999998, 0.8789227166276347], "iscrowd": 0, "area": 19359.330099999996, "rle": {"size": [427, 640], "counts": "\\Ui07n<7I7K5J6J5UJnNYOW1e0jNXOZ1g0gNVO\\1aLeNR20\\1]1ZLnNV2F]1^1ULXOZ2[O_1^1PLB\\2QOb1_1nKH\\2jNd1_1nKK[2gN_1f1TLF[2eN\\1j1WLD[2cNY1n1ZLB[2aNV1R2]L_O\\2`NS1U2_L^O\\2^NR1W2`L^O]2[NP1Z2aL^O]2YNo0\\2bL]O^2XNo0\\2aL_O^2VNP1\\2`LA]2UNR1[2_LCY2WNW1W2^LEU2YN\\1S2[LJR2XNa1P2YLNP2VNg1R5gL^LY3j6O100000O100O001O1O1O12M3N2M2O1N11O1O1O1N2O1O1O1O0O2O1O1O1O1N2O1O1O1O1N2O1O001O2I6A?A?A?B>A?A`0D;J7I7I7I6J7K5L3M4L4mM]GWOg8e0`GUOe8e0bGUOh8`0Z2F6K4L5L4K4M6I`XY6"}, "label": "a woman in a bright blue hat wearing black winter clothes"}]}
{"id": 279485, "image": "COCO_train2014_000000279485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady in the middle\"."}], "task": "refering", "answer_template": "The \"lady in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 100, 101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 262, 263, 285, 286, 308, 309], "bbox": [0.35353124999999996, 0.2020608899297424, 0.560734375, 0.8930210772833724], "iscrowd": 0, "area": 20127.753950000002, "rle": {"size": [427, 640], "counts": "o]n24T==C>B=C:F=C3M3M3M3M3M3M3M3M3ZOf000O2O7H9H8G8I7H5L4H8D<D:F8G9L4O1O100O1OH9G80001N3O01O1O10O01YKlIU2X6cMPJY2W6YMSJe2S6lLUJT3R6]LWJb3P6PLWJQ4a70000001O00001O0000009G2N2O2M2N2N1O1O0QMSK`Mn4_2YKZMh4e2_KTMa4l2fKmL[4R3oKdLR4[3WL\\Li3d3`LSLa3l3hLkKY3T4o201O1O001O4L3M1O2N1O1O1O1O1O1O2N1O1_NVG_Nk8_1YG^Nh8a1\\G[Ne8d1_GXNb8f1bGWN`8g1dGoMc8o1bGfMf8Y2]G]Mk8a2Q1O1O1N3N1N2O1O2M4M2^NhDl0\\;QOfDm0];oNfDo0\\;iNlDT1g;O2M3N2N1N3N2M3N2M2O2N2M3N1N3Klnd3"}, "label": "lady in the middle"}]}
{"id": 279485, "image": "COCO_train2014_000000279485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a red scarf and a black bag\"."}], "task": "refering", "answer_template": "The \"a person with a red scarf and a black bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 100, 101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 262, 263, 285, 286, 308, 309], "bbox": [0.35353124999999996, 0.2020608899297424, 0.560734375, 0.8930210772833724], "iscrowd": 0, "area": 20127.753950000002, "rle": {"size": [427, 640], "counts": "o]n24T==C>B=C:F=C3M3M3M3M3M3M3M3M3ZOf000O2O7H9H8G8I7H5L4H8D<D:F8G9L4O1O100O1OH9G80001N3O01O1O10O01YKlIU2X6cMPJY2W6YMSJe2S6lLUJT3R6]LWJb3P6PLWJQ4a70000001O00001O0000009G2N2O2M2N2N1O1O0QMSK`Mn4_2YKZMh4e2_KTMa4l2fKmL[4R3oKdLR4[3WL\\Li3d3`LSLa3l3hLkKY3T4o201O1O001O4L3M1O2N1O1O1O1O1O1O2N1O1_NVG_Nk8_1YG^Nh8a1\\G[Ne8d1_GXNb8f1bGWN`8g1dGoMc8o1bGfMf8Y2]G]Mk8a2Q1O1O1N3N1N2O1O2M4M2^NhDl0\\;QOfDm0];oNfDo0\\;iNlDT1g;O2M3N2N1N3N2M3N2M2O2N2M3N1N3Klnd3"}, "label": "a person with a red scarf and a black bag"}]}
{"id": 279485, "image": "COCO_train2014_000000279485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a black jacket and blue jeans\"."}], "task": "refering", "answer_template": "The \"man wearing a black jacket and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 134, 135, 136, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 249, 250, 251, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320], "bbox": [0.8208125000000001, 0.270304449648712, 0.9774375000000001, 0.8910772833723654], "iscrowd": 0, "area": 16823.806399999994, "rle": {"size": [427, 640], "counts": "\\Sk6<m<_2aMM4L3M3O13M3M3M3M3M3gHPMS4S3jKQMQ4S3lKQMo3S3nKQMh3X3ULlLW3g3eL^LW3e3fL^LX3h3aL\\L\\3k3\\LXLa3o3XLSLg3S4RLoKm3X4kKjKS4h6O1O1N2O1O1N200010O001O001O001O001O2N1O2oMQ2B=N20000000000000000000000O1000000001O2N2N3M2N2N2N2O1N2TIZJo5g5bIeJ_6S62M2O1TM\\IoNf6=oI_OS6>QJ@Q6<TJAm5<XJ@k5=XJAi5<\\JAe5<`J@c5=`JAa5<dJ@_5<fJ@\\5=hJ@[5=iJ_OY5>lJ^OW5>mJ^OV5`0nJ\\OU5`0PK\\OS5`0RK\\OP5b0SK[OP5a0UK[Om4b0dKmN`4o0Q4M2Ml0TOlc5"}, "label": "man wearing a black jacket and blue jeans"}]}
{"id": 279485, "image": "COCO_train2014_000000279485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black jacket\"."}], "task": "refering", "answer_template": "The \"a man in a black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 134, 135, 136, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 249, 250, 251, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320], "bbox": [0.8208125000000001, 0.270304449648712, 0.9774375000000001, 0.8910772833723654], "iscrowd": 0, "area": 16823.806399999994, "rle": {"size": [427, 640], "counts": "\\Sk6<m<_2aMM4L3M3O13M3M3M3M3M3gHPMS4S3jKQMQ4S3lKQMo3S3nKQMh3X3ULlLW3g3eL^LW3e3fL^LX3h3aL\\L\\3k3\\LXLa3o3XLSLg3S4RLoKm3X4kKjKS4h6O1O1N2O1O1N200010O001O001O001O001O2N1O2oMQ2B=N20000000000000000000000O1000000001O2N2N3M2N2N2N2O1N2TIZJo5g5bIeJ_6S62M2O1TM\\IoNf6=oI_OS6>QJ@Q6<TJAm5<XJ@k5=XJAi5<\\JAe5<`J@c5=`JAa5<dJ@_5<fJ@\\5=hJ@[5=iJ_OY5>lJ^OW5>mJ^OV5`0nJ\\OU5`0PK\\OS5`0RK\\OP5b0SK[OP5a0UK[Om4b0dKmN`4o0Q4M2Ml0TOlc5"}, "label": "a man in a black jacket"}]}
{"id": 25533, "image": "COCO_train2014_000000025533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a fanny - pack is skiing\"."}], "task": "refering", "answer_template": "The \"a man wearing a fanny - pack is skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 190, 191, 192, 193, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 276, 277, 278, 293, 294, 295, 309, 310, 311, 312, 326, 327, 343, 344], "bbox": [0.19939583333333333, 0.42973437499999995, 0.4345, 0.8880624999999999], "iscrowd": 0, "area": 15723.949700000012, "rle": {"size": [640, 480], "counts": "oZl14jc07J4L2M3N1K6I7M2N3N1O00001dAjNhL?`;g0cGUOVLh0T<4`Gn0]8RO^GV1_8jN]G[1a8fNWDkNd02?c2e:]NYDTO6<`0Y2P;UN[D\\OHf0?Q2[;j0oCEl;Z4K5N2O100OO1O2N1O10BLa0L5J6L3N4K4M4K5L4L3L5L4K5L4L4K5L4K5L4L4K7J5LaBoJV<X5`BYK1N_=P6O1O112O2ObIdBT6b=4L1N2O1N10N1O101N1O1O2N1O1N3A_B\\Jb=b5?UKVBl1HbNU>CXBb1O_Nm=OVB\\1<VNa=`0TBS1k0nMT=P1TB_O\\O4i`0=m_OWOA7e`0d0]@YOg?f0Z@WOh?i0X@TOl?k0U@QOo?o0a101O06K6I6KO0O2O0O2Oa]OgNVb0X1h]OlNXb0_10N1O2N2N2N2N2N2M3L9H:F:FSPY5"}, "label": "a man wearing a fanny - pack is skiing"}]}
{"id": 25533, "image": "COCO_train2014_000000025533.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skaters moving around in the snow\"."}], "task": "refering", "answer_template": "The \"skaters moving around in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 190, 191, 192, 193, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 276, 277, 278, 293, 294, 295, 309, 310, 311, 312, 326, 327, 343, 344], "bbox": [0.19939583333333333, 0.42973437499999995, 0.4345, 0.8880624999999999], "iscrowd": 0, "area": 15723.949700000012, "rle": {"size": [640, 480], "counts": "oZl14jc07J4L2M3N1K6I7M2N3N1O00001dAjNhL?`;g0cGUOVLh0T<4`Gn0]8RO^GV1_8jN]G[1a8fNWDkNd02?c2e:]NYDTO6<`0Y2P;UN[D\\OHf0?Q2[;j0oCEl;Z4K5N2O100OO1O2N1O10BLa0L5J6L3N4K4M4K5L4L3L5L4K5L4L4K5L4K5L4L4K7J5LaBoJV<X5`BYK1N_=P6O1O112O2ObIdBT6b=4L1N2O1N10N1O101N1O1O2N1O1N3A_B\\Jb=b5?UKVBl1HbNU>CXBb1O_Nm=OVB\\1<VNa=`0TBS1k0nMT=P1TB_O\\O4i`0=m_OWOA7e`0d0]@YOg?f0Z@WOh?i0X@TOl?k0U@QOo?o0a101O06K6I6KO0O2O0O2Oa]OgNVb0X1h]OlNXb0_10N1O2N2N2N2N2N2M3L9H:F:FSPY5"}, "label": "skaters moving around in the snow"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing the white shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 397, 398, 399, 401, 402, 403, 404, 405, 406, 420, 421, 424, 425, 426, 427, 428, 429, 443, 444, 447, 448, 449, 450, 451, 452, 466, 467, 471], "bbox": [0.16312500000000002, 0.19707086614173228, 0.68828125, 0.886267716535433], "iscrowd": 0, "area": 88323.21504999997, "rle": {"size": [635, 640], "counts": "j_Q27bc05J6K6J5J6K6K4K5L3M3M3L4M3M3M3M4K4M3M3M3L4M2N2N3L3N001N2O1O1N2O0O2O1O1N2WNi1M3N1N3M3M3M3M3L4G9I7J6I7J6J6I7J6I7H8H8N3M4L3M4L5K6J5K6J6J6J6Jo1QN3M4L3M3M4M2N2N2O2M101N1O100O1O100O1O100O101N1O100O1O100O100O2N1000000O10nMiI`FW6]9SJ\\Fm5a9\\JYFd5e9cJWF\\5h9fJWFY5i9jJUFV5k9kJTFU5k9nJSFR5l9QKRFo4m9SKRFm4n9XKmEg4S:gKaEX4^:ULVEk3i:e2O0O1O2N100O2O0000001N10001O000O2O00001O0O101O00001N10001O000O2O0000001N2O1O1O1O1N20O01O1O1O100O1O1O1O010O1O1O10O01O1O1O01000001O0O2\\FkF_8Y:01O000O2OO1000O010001N2O001000O2O0000:E3NN2N2O0O2O3L6K6I6K6I6J6I7J5J5L5J5L5K4K6K4L2M101O0O2O001N101O1N101N101N101N101N101N2O001O001O001O1O001O001O1O001O002N2O1N2N2N2N101N2N2N2N2O1N2N2N3M2N2N3L3N2N2N3M2N2N3M2M3N2N3M2M3N3M2N2M3N3M2N2M3N3M2N2M4M2N2N2M4M2N2N3L3N2N2N3L3N2N2N3L3N2N3gNQ@`NQ`0\\1T@`No?\\1U@bNm?Z1X@cNk?X1Y@fNi?W1`1L5K4L4G9F;Fdnj3"}, "label": "a woman wearing the white shirt"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting at the table of a cafe having a drink\"."}], "task": "refering", "answer_template": "The \"a woman sitting at the table of a cafe having a drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 397, 398, 399, 401, 402, 403, 404, 405, 406, 420, 421, 424, 425, 426, 427, 428, 429, 443, 444, 447, 448, 449, 450, 451, 452, 466, 467, 471], "bbox": [0.16312500000000002, 0.19707086614173228, 0.68828125, 0.886267716535433], "iscrowd": 0, "area": 88323.21504999997, "rle": {"size": [635, 640], "counts": "j_Q27bc05J6K6J5J6K6K4K5L3M3M3L4M3M3M3M4K4M3M3M3L4M2N2N3L3N001N2O1O1N2O0O2O1O1N2WNi1M3N1N3M3M3M3M3L4G9I7J6I7J6J6I7J6I7H8H8N3M4L3M4L5K6J5K6J6J6J6Jo1QN3M4L3M3M4M2N2N2O2M101N1O100O1O100O1O100O101N1O100O1O100O100O2N1000000O10nMiI`FW6]9SJ\\Fm5a9\\JYFd5e9cJWF\\5h9fJWFY5i9jJUFV5k9kJTFU5k9nJSFR5l9QKRFo4m9SKRFm4n9XKmEg4S:gKaEX4^:ULVEk3i:e2O0O1O2N100O2O0000001N10001O000O2O00001O0O101O00001N10001O000O2O0000001N2O1O1O1O1N20O01O1O1O100O1O1O1O010O1O1O10O01O1O1O01000001O0O2\\FkF_8Y:01O000O2OO1000O010001N2O001000O2O0000:E3NN2N2O0O2O3L6K6I6K6I6J6I7J5J5L5J5L5K4K6K4L2M101O0O2O001N101O1N101N101N101N101N101N2O001O001O001O1O001O001O1O001O002N2O1N2N2N2N101N2N2N2N2O1N2N2N3M2N2N3L3N2N2N3M2N2N3M2M3N2N3M2M3N3M2N2M3N3M2N2M3N3M2N2M4M2N2N2M4M2N2N3L3N2N2N3L3N2N2N3L3N2N3gNQ@`NQ`0\\1T@`No?\\1U@bNm?Z1X@cNk?X1Y@fNi?W1`1L5K4L4G9F;Fdnj3"}, "label": "a woman sitting at the table of a cafe having a drink"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting in a wooden chair with her back turned\"."}], "task": "refering", "answer_template": "The \"a woman sitting in a wooden chair with her back turned\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 337, 338, 339, 340, 341, 342, 361, 362, 363], "bbox": [0.6309843749999999, 0.31236220472440945, 0.87846875, 0.6966299212598426], "iscrowd": 0, "area": 19861.7523, "rle": {"size": [635, 640], "counts": "Zkj72hc01N2N3N1N2N3N1N2O2M2N3N1N2O2M2Nb0_O;D4K6I7J5J7J6J6I6K6J5J6K5K5K5J6K5K5J6K5K5J6K5N200O100O100O01000O10O0O2N2`KXBj2h=kLhBo2Z=fLUCT3l<aLeCY3n:hLcFR3X9]MdF^2V9PNgFj1U9cNhFW1T9UOjFf0V9]OlF=T9HmF2T91nFJQ9:QG@P9c0RGWOn8m0TGjNo8Z1SGZNT9i1nFkMX9Y2jF[M]9i2h34L4M4K6K5J7J5J6J6K6I6K5J6J7J5J6K5J7J5J6JO2O001O2M3N2N2N2M4M2N2N2M3N2N2N2M4M2N2N2N2M3N2N2N3L3N1O2N1N3N1O2N1N3N1O2N2N1N3N1O2N1N3N1O2N2M2O2N1O2M4M2N2N2N2UN]_O9f`0_Of_O;\\`0]Oo_O>S`0[OX@?k?XO`@b0ba0K6J4L2M4M3MXo_1"}, "label": "a woman sitting in a wooden chair with her back turned"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with yellow sweater\"."}], "task": "refering", "answer_template": "The \"woman with yellow sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 337, 338, 339, 340, 341, 342, 361, 362, 363], "bbox": [0.6309843749999999, 0.31236220472440945, 0.87846875, 0.6966299212598426], "iscrowd": 0, "area": 19861.7523, "rle": {"size": [635, 640], "counts": "Zkj72hc01N2N3N1N2N3N1N2O2M2N3N1N2O2M2Nb0_O;D4K6I7J5J7J6J6I6K6J5J6K5K5K5J6K5K5J6K5K5J6K5N200O100O100O01000O10O0O2N2`KXBj2h=kLhBo2Z=fLUCT3l<aLeCY3n:hLcFR3X9]MdF^2V9PNgFj1U9cNhFW1T9UOjFf0V9]OlF=T9HmF2T91nFJQ9:QG@P9c0RGWOn8m0TGjNo8Z1SGZNT9i1nFkMX9Y2jF[M]9i2h34L4M4K6K5J7J5J6J6K6I6K5J6J7J5J6K5J7J5J6JO2O001O2M3N2N2N2M4M2N2N2M3N2N2N2M4M2N2N2N2M3N2N2N3L3N1O2N1N3N1O2N1N3N1O2N2N1N3N1O2N1N3N1O2N2M2O2N1O2M4M2N2N2N2UN]_O9f`0_Of_O;\\`0]Oo_O>S`0[OX@?k?XO`@b0ba0K6J4L2M4M3MXo_1"}, "label": "woman with yellow sweater"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pole with an umbrella\"."}], "task": "refering", "answer_template": "The \"the pole with an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 130, 131, 153, 154, 177, 200, 223, 224, 246, 247, 269, 270, 292, 293, 315, 316, 339, 362, 385, 386, 408, 409, 431, 432, 454, 455, 477, 478], "bbox": [0.490546875, 0.0026614173228346454, 0.8545624999999999, 0.8885669291338583], "iscrowd": 0, "area": 33018.71365000001, "rle": {"size": [635, 640], "counts": "ZjR64]c0:F:G9F:G9F:F:G9F:G9G90000000000000000000000000001O00000000000000000000001O000000000000000000001O0000000000000000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O000000000000000000000000000000O1000000000000000O01O1O1O1O1O1O1O100001O001O001O001O001O001O002N5K6J5K5K6J5K5K6J5K5K5K6J5K5K6J6J8H7I8H7I8HkKYD=`;CgD>Q;BVE>c:AeE`0S:@TF`0e9@bFa0V9^ORGb0g8^O`Gb0Y8^OnGc0j7\\O^Hd0[7\\OlHe0l6[O[Ie0^6[OiIf0o5YOYJg0`5YOgJh0Q5WOWKi0b4VOfKk0R4UOULk0d3TOdLm0T3ROTMn0e2RObMo0V2PORNP1g1oNbNQ1V1oNQOQ1h0nN@S18mNOS1JlN>U1ZOjNn0V1kNjN\\1W1\\NhNl1X1mMgN[2Z1]MfNj2Z1oLeNY3\\1_LcNi3]1PLcNW4^1aKaNg4_1RK`NV5a1bJ_Ne5a1TJ^NT6c1dI\\Nd6d1UI\\NR7e1fHZNb7a1\\H_Nk7Z1UHeNS8S1nGlNZ8m0fGSOa8f0_GYOg8a0ZG_Ok8<UGCQ98oFHW91jFN\\9MdF2b9I^F7h9BYF=m9^OSFb0U:VOkEi0^:nNcER1f:eNZEZ1o:^NQEa1X;VNiDj1_;nMaDQ2i;fMWDZ2Q<]MPDb2e>00O1O102M4M3L4M2M4L4M3L3N3L4M2M4L4M3L3N3L4M3L3M4M3L3N3L4MYci1"}, "label": "the pole with an umbrella"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pole with an umbrella\"."}], "task": "refering", "answer_template": "The \"the pole with an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 130, 131, 153, 154, 177, 200, 223, 224, 246, 247, 269, 270, 292, 293, 315, 316, 339, 362, 385, 386, 408, 409, 431, 432, 454, 455, 477, 478], "bbox": [0.490546875, 0.0026614173228346454, 0.8545624999999999, 0.8885669291338583], "iscrowd": 0, "area": 33018.71365000001, "rle": {"size": [635, 640], "counts": "ZjR64]c0:F:G9F:G9F:F:G9F:G9G90000000000000000000000000001O00000000000000000000001O000000000000000000001O0000000000000000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O000000000000000000000000000000O1000000000000000O01O1O1O1O1O1O1O100001O001O001O001O001O001O002N5K6J5K5K6J5K5K6J5K5K5K6J5K5K6J6J8H7I8H7I8HkKYD=`;CgD>Q;BVE>c:AeE`0S:@TF`0e9@bFa0V9^ORGb0g8^O`Gb0Y8^OnGc0j7\\O^Hd0[7\\OlHe0l6[O[Ie0^6[OiIf0o5YOYJg0`5YOgJh0Q5WOWKi0b4VOfKk0R4UOULk0d3TOdLm0T3ROTMn0e2RObMo0V2PORNP1g1oNbNQ1V1oNQOQ1h0nN@S18mNOS1JlN>U1ZOjNn0V1kNjN\\1W1\\NhNl1X1mMgN[2Z1]MfNj2Z1oLeNY3\\1_LcNi3]1PLcNW4^1aKaNg4_1RK`NV5a1bJ_Ne5a1TJ^NT6c1dI\\Nd6d1UI\\NR7e1fHZNb7a1\\H_Nk7Z1UHeNS8S1nGlNZ8m0fGSOa8f0_GYOg8a0ZG_Ok8<UGCQ98oFHW91jFN\\9MdF2b9I^F7h9BYF=m9^OSFb0U:VOkEi0^:nNcER1f:eNZEZ1o:^NQEa1X;VNiDj1_;nMaDQ2i;fMWDZ2Q<]MPDb2e>00O1O102M4M3L4M2M4L4M3L3N3L4M2M4L4M3L3N3L4M3L3M4M3L3N3L4MYci1"}, "label": "the pole with an umbrella"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair of the woman who is behind the woman reaching for her glass\"."}], "task": "refering", "answer_template": "The \"the chair of the woman who is behind the woman reaching for her glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387, 388, 406, 407, 408, 409, 410, 411, 429, 430, 431, 432, 433, 434, 452, 453, 455, 456, 457, 478, 479, 480], "bbox": [0.609453125, 0.49822047244094486, 0.905125, 0.9044409448818896], "iscrowd": 0, "area": 28678.63705000001, "rle": {"size": [635, 640], "counts": "nUb72ic04L3M4K4M3M2N2N1O2M3N2N2N2N2M3N2N2N2N2N2M3N3M2N3M2N3L3N3M2N3M2N3L3N3M2N3M2N3M2M6K5`AmLn;X3jCkLV<Z3aCjL_<[3YChLg<]3PCgLo<_3hBeLX=`3`BbLa=c3VBaLj=h40000O1000000000O010000000000O10000000O1000O1006J6J7I6J7I6J7H7J6J7I6J7I6J7I6J7I6JbLbMeEW2\\:lMhEm1X:VNkEd1T:cNmEV1S:POmEj0S:\\OnE=R:InE1R:5oEDP:b0QFXOo9n0QFlNo9Z1RF_Nn9g1RFRNo9T2h37J6J6J6J6I6K6J6J6J6I7J6J6J6J6I6K6J6J6J6I7J3M3M3L4M4L3M3M3L4M4L3M3M3L4M3M00000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000kHnDa5R;_JoE]4S:cKnFX3Y9fLhGU2^8hMdHR1b7mNmI_OZ6>d6M2Ga`U1"}, "label": "the chair of the woman who is behind the woman reaching for her glass"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a chair with a woman sitting in it\"."}], "task": "refering", "answer_template": "The \"the back of a chair with a woman sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387, 388, 406, 407, 408, 409, 410, 411, 429, 430, 431, 432, 433, 434, 452, 453, 455, 456, 457, 478, 479, 480], "bbox": [0.609453125, 0.49822047244094486, 0.905125, 0.9044409448818896], "iscrowd": 0, "area": 28678.63705000001, "rle": {"size": [635, 640], "counts": "nUb72ic04L3M4K4M3M2N2N1O2M3N2N2N2N2M3N2N2N2N2N2M3N3M2N3M2N3L3N3M2N3M2N3L3N3M2N3M2N3M2M6K5`AmLn;X3jCkLV<Z3aCjL_<[3YChLg<]3PCgLo<_3hBeLX=`3`BbLa=c3VBaLj=h40000O1000000000O010000000000O10000000O1000O1006J6J7I6J7I6J7H7J6J7I6J7I6J7I6J7I6JbLbMeEW2\\:lMhEm1X:VNkEd1T:cNmEV1S:POmEj0S:\\OnE=R:InE1R:5oEDP:b0QFXOo9n0QFlNo9Z1RF_Nn9g1RFRNo9T2h37J6J6J6J6I6K6J6J6J6I7J6J6J6J6I6K6J6J6J6I7J3M3M3L4M4L3M3M3L4M4L3M3M3L4M3M00000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000kHnDa5R;_JoE]4S:cKnFX3Y9fLhGU2^8hMdHR1b7mNmI_OZ6>d6M2Ga`U1"}, "label": "the back of a chair with a woman sitting in it"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tabletop with food and drinks in front of a sitting woman\"."}], "task": "refering", "answer_template": "The \"a tabletop with food and drinks in front of a sitting woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [371, 372, 373, 394, 395, 396, 397, 398, 399, 400, 401, 417, 418, 419, 420, 421, 422, 423, 424, 428, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 449, 450, 451, 452, 453, 454, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528], "bbox": [0.0268125, 0.7083307086614173, 1.0, 0.9898582677165355], "iscrowd": 0, "area": 58260.92135000001, "rle": {"size": [635, 640], "counts": "lT;1fc04L4M3L4M3L4L4M3L4L4M3L4M3L4L4M3L4M3L4L4M3O100O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O11O00001O00001O00001O00lMZMSCf2l<gMiBX2W=UN\\Bk1d=aNPB_1o=nNfAQ1Z>[OZAe0f>X2O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000001O00001O001O00001O00001O001O00001O3M6J7I6J7I6K5J7I6J7I6J6J2N2N1O2N2N2N2N1O2N2N2N2N2N1O1O0000001O000000001O000000000000kMPN_BP2a=SN\\Bm1c=XNYBh1g=\\NUBd1j=aNRB_1n=dNoA\\1Q>hNkAX1T>mNhAS1X>POeAP1Z>UObAk0^>YO^Ag0a>]O\\Ac0d>AXA?g>FUA:k>IRA7n>Mn@3Q?2k@NU?5h@KW?]200O100O100O10000O100O100O100O10000O100O100O10000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000UNo@lNQ?d0^A]Ob>3nAMR>B_B>a=ROnBo0S=`N]C`1^?100O100O100O10000O100O10000O100O10000O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O1O100O1O100O1O100O100O1O100O100O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O101O0O100O10000O10000O100O11O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O000010O0001O001O00001O00001O00001O00001O001O00001J5L4K6J5L4K_]O"}, "label": "a tabletop with food and drinks in front of a sitting woman"}]}
{"id": 467905, "image": "COCO_train2014_000000467905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table in the drink\"."}], "task": "refering", "answer_template": "The \"table in the drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [371, 372, 373, 394, 395, 396, 397, 398, 399, 400, 401, 417, 418, 419, 420, 421, 422, 423, 424, 428, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 449, 450, 451, 452, 453, 454, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528], "bbox": [0.0268125, 0.7083307086614173, 1.0, 0.9898582677165355], "iscrowd": 0, "area": 58260.92135000001, "rle": {"size": [635, 640], "counts": "lT;1fc04L4M3L4M3L4L4M3L4L4M3L4M3L4L4M3L4M3L4L4M3O100O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O11O00001O00001O00001O00lMZMSCf2l<gMiBX2W=UN\\Bk1d=aNPB_1o=nNfAQ1Z>[OZAe0f>X2O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000001O00001O001O00001O00001O001O00001O3M6J7I6J7I6K5J7I6J7I6J6J2N2N1O2N2N2N2N1O2N2N2N2N2N1O1O0000001O000000001O000000000000kMPN_BP2a=SN\\Bm1c=XNYBh1g=\\NUBd1j=aNRB_1n=dNoA\\1Q>hNkAX1T>mNhAS1X>POeAP1Z>UObAk0^>YO^Ag0a>]O\\Ac0d>AXA?g>FUA:k>IRA7n>Mn@3Q?2k@NU?5h@KW?]200O100O100O10000O100O100O100O10000O100O100O10000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000UNo@lNQ?d0^A]Ob>3nAMR>B_B>a=ROnBo0S=`N]C`1^?100O100O100O10000O100O10000O100O10000O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O1O100O1O100O1O100O100O1O100O100O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O101O0O100O10000O10000O100O11O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O000010O0001O001O00001O00001O00001O00001O001O00001J5L4K6J5L4K_]O"}, "label": "table in the drink"}]}
{"id": 558018, "image": "COCO_train2014_000000558018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white baseball bat , held by a person\"."}], "task": "refering", "answer_template": "The \"a white baseball bat , held by a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 163, 186, 187, 209, 210, 232, 233, 256, 279, 280, 326], "bbox": [0.0818125, 0.4378454332552693, 0.2168125, 0.9766510538641685], "iscrowd": 0, "area": 3519.3068000000007, "rle": {"size": [427, 640], "counts": "ZYf04V=5L4K5L3M4K5L4K5L4L4K4M4K5L4L4K5L3L5L400O10000O01000O10O001O1O001O1O001OTEcMg:a2101N1O1O2N1O1TOcEfN^:U1eElN]:o0eERO\\:i0gEXOZ:c0iE^OY:=iEDX:7kEJV:2lEOV:KmE4V:GmE8V:DlE:c;00O2N100O10000O10000O2O0O10001N10b12YN5K5N11O0O2N1O2N101N1O2N1OW``6"}, "label": "a white baseball bat , held by a person"}]}
{"id": 558018, "image": "COCO_train2014_000000558018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bat\"."}], "task": "refering", "answer_template": "The \"the white bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 163, 186, 187, 209, 210, 232, 233, 256, 279, 280, 326], "bbox": [0.0818125, 0.4378454332552693, 0.2168125, 0.9766510538641685], "iscrowd": 0, "area": 3519.3068000000007, "rle": {"size": [427, 640], "counts": "ZYf04V=5L4K5L3M4K5L4K5L4L4K4M4K5L4L4K5L3L5L400O10000O01000O10O001O1O001O1O001OTEcMg:a2101N1O1O2N1O1TOcEfN^:U1eElN]:o0eERO\\:i0gEXOZ:c0iE^OY:=iEDX:7kEJV:2lEOV:KmE4V:GmE8V:DlE:c;00O2N100O10000O10000O2O0O10001N10b12YN5K5N11O0O2N1O2N101N1O2N1OW``6"}, "label": "the white bat"}]}
{"id": 558018, "image": "COCO_train2014_000000558018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"darker baseball bat\"."}], "task": "refering", "answer_template": "The \"darker baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 28, 29, 30, 52, 53, 75, 76, 99, 122, 123, 145, 146, 192, 238, 239, 261, 262], "bbox": [0.234421875, 0.006299765807962529, 0.424890625, 0.7716393442622951], "iscrowd": 0, "area": 5732.19835, "rle": {"size": [427, 640], "counts": "ean11Z=2N3M3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M3M3M3M40O1O2N1O2N2O0O2N1O2N101N2N2N2N2N2O2M2N2N2N010O00001O00010O000L4L5L3L4M4K4L4M3L5LUEVOX9f0gF]O[9?eFB_9:`FGc95^FJf92[FNh9OXF1a;00O10O01000VEN[81lFNTO5m9MoF0QO4P:LnF3oN3S:HoF8kN0k;4PDMo;h00000O1001O0000001O0000004K5L8H1O0O2O000O2O1N2N2N3LcTi4"}, "label": "darker baseball bat"}]}
{"id": 558018, "image": "COCO_train2014_000000558018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two hands holding a purple bat\"."}], "task": "refering", "answer_template": "The \"two hands holding a purple bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 28, 29, 30, 52, 53, 75, 76, 99, 122, 123, 145, 146, 192, 238, 239, 261, 262], "bbox": [0.234421875, 0.006299765807962529, 0.424890625, 0.7716393442622951], "iscrowd": 0, "area": 5732.19835, "rle": {"size": [427, 640], "counts": "ean11Z=2N3M3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M3M3M3M40O1O2N1O2N2O0O2N1O2N101N2N2N2N2N2O2M2N2N2N010O00001O00010O000L4L5L3L4M4K4L4M3L5LUEVOX9f0gF]O[9?eFB_9:`FGc95^FJf92[FNh9OXF1a;00O10O01000VEN[81lFNTO5m9MoF0QO4P:LnF3oN3S:HoF8kN0k;4PDMo;h00000O1001O0000001O0000004K5L8H1O0O2O000O2O1N2N2N3LcTi4"}, "label": "two hands holding a purple bat"}]}
{"id": 320454, "image": "COCO_train2014_000000320454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog is lying down\"."}], "task": "refering", "answer_template": "The \"a brown dog is lying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105], "bbox": [0.3988125, 0.1496487119437939, 0.692265625, 0.3242388758782201], "iscrowd": 0, "area": 8100.321250000002, "rle": {"size": [427, 640], "counts": "S^Z31V=5L3L4M4K4O100O1O2N1O100O1O1O100O1O1O1O100O1O1O1O10000O100000000O100000001N10001O2N2M3N2N2N2N1N3N1O1O1O1O001O1O1O000000YOiDYOW;h0kDVOU;j0mDTOS;m0nDPOS;P1oDnNQ;R1PEmNP;S1REkNn:U1TEhNm:X1UEfNk:Z1WEdNi:\\1c00O2O00000000000000001N100000000000001OO10O1000O1000000O0100000O10000000O1000000000000O10000000000000000000000000000000000001O000O10000000000000000000000000001N100010O0001O00001O00001O00001O00001O01OO2F9F;N1O2N1O1O2N1O2N1O1O2N1O2N1Od_b2"}, "label": "a brown dog is lying down"}]}
{"id": 320454, "image": "COCO_train2014_000000320454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog layiing down next to a big chair\"."}], "task": "refering", "answer_template": "The \"a brown dog layiing down next to a big chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105], "bbox": [0.3988125, 0.1496487119437939, 0.692265625, 0.3242388758782201], "iscrowd": 0, "area": 8100.321250000002, "rle": {"size": [427, 640], "counts": "S^Z31V=5L3L4M4K4O100O1O2N1O100O1O1O100O1O1O1O100O1O1O1O10000O100000000O100000001N10001O2N2M3N2N2N2N1N3N1O1O1O1O001O1O1O000000YOiDYOW;h0kDVOU;j0mDTOS;m0nDPOS;P1oDnNQ;R1PEmNP;S1REkNn:U1TEhNm:X1UEfNk:Z1WEdNi:\\1c00O2O00000000000000001N100000000000001OO10O1000O1000000O0100000O10000000O1000000000000O10000000000000000000000000000000000001O000O10000000000000000000000000001N100010O0001O00001O00001O00001O00001O01OO2F9F;N1O2N1O1O2N1O2N1O1O2N1O2N1Od_b2"}, "label": "a brown dog layiing down next to a big chair"}]}
{"id": 320454, "image": "COCO_train2014_000000320454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white colored pillow which has some mokey cartoons on it\"."}], "task": "refering", "answer_template": "The \"a white colored pillow which has some mokey cartoons on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 68, 84, 85, 89, 90, 91, 107, 108, 110, 111, 112, 113, 114, 130, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.66025, 0.14466042154566747, 1.0, 0.8329742388758782], "iscrowd": 0, "area": 36034.7026, "rle": {"size": [427, 640], "counts": "S``5b0]<f0A9I3M3M3N2O001O1O001O1O001O001N10001O001O000010O3N1ZOhDVO[;g0iDUOi4C`1V1kITOd4Ha1R1lIUOc4Jc1n0kIWOa4Me1j0lIXO^4OW27\\II]42S3LmL6S3GoL:P82M2O1O1N3N1O1N3N2N1O2O1N2N1O2N2N2N1O2N2O1N1O2N2N2N1O2N2N2N1N3N2M3M2O2M3M3M2O2M3M3N1N3M3N2M2N3M3N2M2N3L4J5K6I6K6J5J6dNYK\\I`5b6e001N100O2N101N1O2O0O1O2O0O2O0O1O2O0O2N101N1O101N1O2O0O101N1O2O0O2N100O2N101N100O2N10000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O100O1O1O001N2O1O1O001O1O1O1N101O1O1O001O1O1O10O01O100O1O00100O1O10O01O1O1O1O001OO1K5K5K4M4K5KXN"}, "label": "a white colored pillow which has some mokey cartoons on it"}]}
{"id": 320454, "image": "COCO_train2014_000000320454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bean bag couch with sock monkies on it\"."}], "task": "refering", "answer_template": "The \"a white bean bag couch with sock monkies on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 68, 84, 85, 89, 90, 91, 107, 108, 110, 111, 112, 113, 114, 130, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.66025, 0.14466042154566747, 1.0, 0.8329742388758782], "iscrowd": 0, "area": 36034.7026, "rle": {"size": [427, 640], "counts": "S``5b0]<f0A9I3M3M3N2O001O1O001O1O001O001N10001O001O000010O3N1ZOhDVO[;g0iDUOi4C`1V1kITOd4Ha1R1lIUOc4Jc1n0kIWOa4Me1j0lIXO^4OW27\\II]42S3LmL6S3GoL:P82M2O1O1N3N1O1N3N2N1O2O1N2N1O2N2N2N1O2N2O1N1O2N2N2N1O2N2N2N1N3N2M3M2O2M3M3M2O2M3M3N1N3M3N2M2N3M3N2M2N3L4J5K6I6K6J5J6dNYK\\I`5b6e001N100O2N101N1O2O0O1O2O0O2O0O1O2O0O2N101N1O101N1O2O0O101N1O2O0O2N100O2N101N100O2N10000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O100O1O1O001N2O1O1O001O1O1O1N101O1O1O001O1O1O10O01O100O1O00100O1O10O01O1O1O1O001OO1K5K5K4M4K5KXN"}, "label": "a white bean bag couch with sock monkies on it"}]}
{"id": 320454, "image": "COCO_train2014_000000320454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog\"."}], "task": "refering", "answer_template": "The \"a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149], "bbox": [0.17690625, 0.04718969555035128, 0.5179375, 0.4671194379391101], "iscrowd": 0, "area": 18926.415549999998, "rle": {"size": [427, 640], "counts": "_U_15U=8H4L3`CFi;P1K40N1O2N101N1O2N2N1O2O1N1102N2M3N001N3N4L3L3N1O1O2N100O100O1O0O2PEZNX:h1`EfNY:\\2I1O100O1O010O1O1O10O01O10OkGoLQ6R3mIQMQ6P3mIRMS6m2lIUMS6l2lIUMS6l2kIVMT6j2kIXMU6h2iIZMV6i2fIYMZ6k2aIVM_6n2QI^Mo6o2[HXMg7V402N00O2N100M3M3M3nNPHXLL2V8a3UHZLI0T8a3]HYLC2Q8_3]1O14L4L3oFUL\\8o3\\GXLa8\\4M2N100O2N1O101N1O1O2O0O001O010O00010O000001O00001O1O001O1O001O1O001O1O002N2N2N_G_KZ8_4bGeK`8c41N2N2N2N2N2N2N2N1O2N1N3N1O2]MfFP1a9bNgF^1_9UNhFj1_9hMjFT2X:M3O1010O000001O000100O010O10000O10O010O2N1O2N1O2N1O2L3N6I9H8G3M3N3N1O1O1O101O0O10O1O001O00001O000000010O00A]C6c<J^C5b<K^C5b<L^C3b<M_C2a<N_C2a<N`C1a<N`C1`<0_COb<1_CLd<3]CKd<5\\CJe<7701O010O1O001O001N2O0O2OceP4"}, "label": "a dog"}]}
{"id": 320454, "image": "COCO_train2014_000000320454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black dog sitting next to a laptop and staring at the screen\"."}], "task": "refering", "answer_template": "The \"black dog sitting next to a laptop and staring at the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149], "bbox": [0.17690625, 0.04718969555035128, 0.5179375, 0.4671194379391101], "iscrowd": 0, "area": 18926.415549999998, "rle": {"size": [427, 640], "counts": "_U_15U=8H4L3`CFi;P1K40N1O2N101N1O2N2N1O2O1N1102N2M3N001N3N4L3L3N1O1O2N100O100O1O0O2PEZNX:h1`EfNY:\\2I1O100O1O010O1O1O10O01O10OkGoLQ6R3mIQMQ6P3mIRMS6m2lIUMS6l2lIUMS6l2kIVMT6j2kIXMU6h2iIZMV6i2fIYMZ6k2aIVM_6n2QI^Mo6o2[HXMg7V402N00O2N100M3M3M3nNPHXLL2V8a3UHZLI0T8a3]HYLC2Q8_3]1O14L4L3oFUL\\8o3\\GXLa8\\4M2N100O2N1O101N1O1O2O0O001O010O00010O000001O00001O1O001O1O001O1O001O1O002N2N2N_G_KZ8_4bGeK`8c41N2N2N2N2N2N2N2N1O2N1N3N1O2]MfFP1a9bNgF^1_9UNhFj1_9hMjFT2X:M3O1010O000001O000100O010O10000O10O010O2N1O2N1O2N1O2L3N6I9H8G3M3N3N1O1O1O101O0O10O1O001O00001O000000010O00A]C6c<J^C5b<K^C5b<L^C3b<M_C2a<N_C2a<N`C1a<N`C1`<0_COb<1_CLd<3]CKd<5\\CJe<7701O010O1O001O001N2O0O2OceP4"}, "label": "black dog sitting next to a laptop and staring at the screen"}]}
{"id": 394185, "image": "COCO_train2014_000000394185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with curly hair helping an elderly woman play a video game\"."}], "task": "refering", "answer_template": "The \"a woman with curly hair helping an elderly woman play a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 236, 237, 238, 239, 240, 241, 242, 243, 244, 255, 256, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 283, 284, 285, 286, 287, 288, 289, 290, 301, 302, 303, 307, 308, 309, 310, 311, 312, 324, 325, 326, 332, 333, 334], "bbox": [0.02328125, 0.20927058823529412, 0.65975, 0.9982117647058824], "iscrowd": 0, "area": 66907.31704999998, "rle": {"size": [425, 640], "counts": "k^63T=4L4L4L4M3L5K4L4L4L4L4M4K4L4L4L4L4M4K4L;G2NVNdEh0[:UOjEj0U:SOoEm0P:POUFo0k9oNXFP1g9nN]FP1c9oN`FP1_9nNbFT1]9kNcFW1\\9gNeF[1[9cNfF^1[9^NfFd1^:3M2N2O1N2N2N2SFiMd8Z2PGTNl8o1oFWNl88hF?:\\Ok85lF`06@j80PGb04Bg8MUGc02De8IZGd0NHc8E_Ge0L;P8POTHa3U7_LkHh3m6YLTIg3k6YLUIh3i6YLWIh3h6WLYIj3f6VL[Ij3d6VL\\Ik3d6TL\\Im3c6SL]In3c6QL]IP4b6PL_IP4`6oKaIR4_6mKaIT4^6lKbIU4]6kKdIU4\\6jKdIW4[6iKeIX4[6fKfI[4Y6eKhI[4W6eKiIh0Bo1e6XMiIh0DP2d6VMhIi0GP2a6UMjIi0GR2_6TMjIi0IS2^6RMiIj0LS2[6RMiIj0NT2Y6QMjIi0OV2X6oLiIj02V2U6nLjIk03W2W6hLgIQ13W2R7gMQIX2n6WMYHIk0P3l6VM[HGk0S3j6UM]HDm0V3f6UM`H_On0\\3b6TMjIk2V6TMWJ`2i5_MYJ`2g5_M\\J_2d5`M^J_2c5_M_J`2c5]M`Ja2b5\\M`Jc2U8N1O:F001O1O001O1O0aNPEe0Q;lN`EQ1`;L3M3M3M4L3M3M3M3M4L3M3M3MnQ1:emN3N3O1N2N101N2O1N1lCZOe;g0YD[Og;e0VD_Oh;b0VD_Oj;b0TD@k;o0001O101N2M4M2N2M3N1N10000O100O10000000O101O001N2O1O1O1N2O1O1cNnNaFS1AnNk84cGo0]OSO2Io79aHl0ZOVO4Hn78cHk0VO[O6DP88bHj0TO]O9Co78cHi0POA<@P87cHg1\\OSNQ87bHg1[ORNS89aHd1\\OTNS81gGCi0X2\\OUNe7E[H:m0k1SOWNd7GYH8P1j1QOZNd7HXH4S1j1PO\\Nc7JWH0V1i1PO_Na7:_IW1oNaN`78bIW1mNcN_77eIV1jNeN_76gIW1gNeN_76jIV2T6kMlIV2R6kMnIV2o5lMQJU2k5nMVIiN?Z3X6oMXIkN=W3W6QN[IlN;U3U6RN`IlN8S3U6SNbInN6P3T6UNeImN6o2R6VNgInN5n2o5WNkImN5l2n5YNlImN5j2m5[NnIlN4j2l5\\NnIlN6g2k5BSJ>l5DTJ;l5FTJ:j5ITJ7k5KUJ4j5OTJ2k5OUJ1i51VJOi54VJLi5k3N1O2N101N2N1O2O1N1O2N1O2N2O0O2N2N1O2N101N2N2N1O2N2O1O1O1O1O1N101O1O1O1O1O1O1O001N2O1O1O1O1O0O2O0O1O100O1O100O1O100O1O10001O0O1000000O1000000O1002N2O1O1O2OO0O2N1N3N1O1O2M2O2N1O1N01000O1000001N10001N01O2N2O2M2N2N2O1M4L3N2M3N2M4N1N2N2O1N1O2O0O2O0O2N2O001O0O2O001N2O001N2O1O1N2O1O1O1N2O1O1N2O1O2M3N3M2M3N3aJ`Jo2b5XKYJV1?Z3[5\\KfJi05f3X5]KgJj06d3Q6XLTJd3n5YLWJc3l5YLZJb3h5YL_Jd3c5ULfJf3b7L4L4L3L5L4L4L3M4L4L<D8H4K4M3M4L3M2N1O1O1O1O2M2O1N2N2O1N3N1N2O1N2N3N1N2O1N3N4K4LTUj2"}, "label": "a woman with curly hair helping an elderly woman play a video game"}]}
{"id": 394185, "image": "COCO_train2014_000000394185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with curly hair who is helping the older woman to play the video game\"."}], "task": "refering", "answer_template": "The \"a woman with curly hair who is helping the older woman to play the video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 236, 237, 238, 239, 240, 241, 242, 243, 244, 255, 256, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 283, 284, 285, 286, 287, 288, 289, 290, 301, 302, 303, 307, 308, 309, 310, 311, 312, 324, 325, 326, 332, 333, 334], "bbox": [0.02328125, 0.20927058823529412, 0.65975, 0.9982117647058824], "iscrowd": 0, "area": 66907.31704999998, "rle": {"size": [425, 640], "counts": "k^63T=4L4L4L4M3L5K4L4L4L4L4M4K4L4L4L4L4M4K4L;G2NVNdEh0[:UOjEj0U:SOoEm0P:POUFo0k9oNXFP1g9nN]FP1c9oN`FP1_9nNbFT1]9kNcFW1\\9gNeF[1[9cNfF^1[9^NfFd1^:3M2N2O1N2N2N2SFiMd8Z2PGTNl8o1oFWNl88hF?:\\Ok85lF`06@j80PGb04Bg8MUGc02De8IZGd0NHc8E_Ge0L;P8POTHa3U7_LkHh3m6YLTIg3k6YLUIh3i6YLWIh3h6WLYIj3f6VL[Ij3d6VL\\Ik3d6TL\\Im3c6SL]In3c6QL]IP4b6PL_IP4`6oKaIR4_6mKaIT4^6lKbIU4]6kKdIU4\\6jKdIW4[6iKeIX4[6fKfI[4Y6eKhI[4W6eKiIh0Bo1e6XMiIh0DP2d6VMhIi0GP2a6UMjIi0GR2_6TMjIi0IS2^6RMiIj0LS2[6RMiIj0NT2Y6QMjIi0OV2X6oLiIj02V2U6nLjIk03W2W6hLgIQ13W2R7gMQIX2n6WMYHIk0P3l6VM[HGk0S3j6UM]HDm0V3f6UM`H_On0\\3b6TMjIk2V6TMWJ`2i5_MYJ`2g5_M\\J_2d5`M^J_2c5_M_J`2c5]M`Ja2b5\\M`Jc2U8N1O:F001O1O001O1O0aNPEe0Q;lN`EQ1`;L3M3M3M4L3M3M3M3M4L3M3M3MnQ1:emN3N3O1N2N101N2O1N1lCZOe;g0YD[Og;e0VD_Oh;b0VD_Oj;b0TD@k;o0001O101N2M4M2N2M3N1N10000O100O10000000O101O001N2O1O1O1N2O1O1cNnNaFS1AnNk84cGo0]OSO2Io79aHl0ZOVO4Hn78cHk0VO[O6DP88bHj0TO]O9Co78cHi0POA<@P87cHg1\\OSNQ87bHg1[ORNS89aHd1\\OTNS81gGCi0X2\\OUNe7E[H:m0k1SOWNd7GYH8P1j1QOZNd7HXH4S1j1PO\\Nc7JWH0V1i1PO_Na7:_IW1oNaN`78bIW1mNcN_77eIV1jNeN_76gIW1gNeN_76jIV2T6kMlIV2R6kMnIV2o5lMQJU2k5nMVIiN?Z3X6oMXIkN=W3W6QN[IlN;U3U6RN`IlN8S3U6SNbInN6P3T6UNeImN6o2R6VNgInN5n2o5WNkImN5l2n5YNlImN5j2m5[NnIlN4j2l5\\NnIlN6g2k5BSJ>l5DTJ;l5FTJ:j5ITJ7k5KUJ4j5OTJ2k5OUJ1i51VJOi54VJLi5k3N1O2N101N2N1O2O1N1O2N1O2N2O0O2N2N1O2N101N2N2N1O2N2O1O1O1O1O1N101O1O1O1O1O1O1O001N2O1O1O1O1O0O2O0O1O100O1O100O1O100O1O10001O0O1000000O1000000O1002N2O1O1O2OO0O2N1N3N1O1O2M2O2N1O1N01000O1000001N10001N01O2N2O2M2N2N2O1M4L3N2M3N2M4N1N2N2O1N1O2O0O2O0O2N2O001O0O2O001N2O001N2O1O1N2O1O1O1N2O1O1N2O1O2M3N3M2M3N3aJ`Jo2b5XKYJV1?Z3[5\\KfJi05f3X5]KgJj06d3Q6XLTJd3n5YLWJc3l5YLZJb3h5YL_Jd3c5ULfJf3b7L4L4L3L5L4L4L3M4L4L<D8H4K4M3M4L3M2N1O1O1O1O2M2O1N2N2O1N3N1N2O1N2N3N1N2O1N3N4K4LTUj2"}, "label": "a woman with curly hair who is helping the older woman to play the video game"}]}
{"id": 451529, "image": "COCO_train2014_000000451529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red coach that is to the left of the mirror on the wall\"."}], "task": "refering", "answer_template": "The \"the red coach that is to the left of the mirror on the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183], "bbox": [0.0, 0.5590390390390391, 0.26422, 0.8758858858858859], "iscrowd": 0, "area": 10019.311399999999, "rle": {"size": [333, 500], "counts": "]6X2T8`0A0O10000O100O10000O100O10O10O100O10000O010O10000O100O01000O100O10000O100O10000O100O10000O100001O001N2O1O2N2N2N1O2N1N101O00001O000O3N1O1O1O2N2M3N2N3M00000O100000000000000O10000000000000000001O1O3M0000O100N2M3O1N2O1N2O1N2O1N2O1N2O1N2N2N2N2N2N2O1O1N2O1O1O2N2K4I8Gk^g3"}, "label": "the red coach that is to the left of the mirror on the wall"}]}
{"id": 451529, "image": "COCO_train2014_000000451529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dining chair on the left side of the table\"."}], "task": "refering", "answer_template": "The \"a dining chair on the left side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 172, 173, 191, 192, 193, 209, 210, 211, 212], "bbox": [0.5973999999999999, 0.6809609609609609, 0.80892, 0.9921321321321321], "iscrowd": 0, "area": 4360.627349999996, "rle": {"size": [333, 500], "counts": "b^Q3U1Z92M2N3M2N2N2N10O0001O2N2N2N2O1O14J5L5J5L5J10N1N3N1O1O2N1O1O2M1000000000000O1000000000000O2O00000000000O100000000000000O10001O0000000O12N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1N2O2N1O1Okln0"}, "label": "a dining chair on the left side of the table"}]}
{"id": 25548, "image": "COCO_train2014_000000025548.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl about to throw a baseball\"."}], "task": "refering", "answer_template": "The \"little girl about to throw a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 21, 22, 23, 24, 38, 39, 40, 41, 55, 56, 57, 58, 72, 73, 74, 75, 88, 89, 90, 91, 92, 93, 104, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 246, 257, 258, 259, 262, 263, 274, 275, 276, 280, 281, 290, 291, 292, 297, 298, 307, 308, 309, 314, 315, 324, 325, 341, 342, 358, 359], "bbox": [0.06402083333333333, 0.03934375, 0.5609375, 0.9493281249999999], "iscrowd": 0, "area": 59010.69034999999, "rle": {"size": [640, 480], "counts": "Umc0d0Zc0e0\\O3L3M3N3M2M3N3L3N2N3L3M3M4K4M2C]M[_Oe2b`0cMW_O_2f`0>M4M3M3M2N3L4M2N3M3M3M2M4M3M2N3M3L3N3M3MjNcAoLZ>d2XBYMg=d2aBVMa=f2gBRM\\=k2kBoL[50f0o2U1VMhNi2W1\\MfNf2X1]MfNd2Y1_MdNa2]1`MaNa2^1aM`N`2_1bM_N_2a1bM]N_2b1cM\\N]2d1eMZN\\2f1fMWNZ2i1hMUNX2k1jMSNU2n1mMcMoNmIT3a8nM`MSOjIn2g8QN]MVOgIi2l8SN[MY2g2hMWMX2i2jMUMU2l2mMRMS2o2nMoLQ2R3QNZLnNhJQ3n8SNQLUOoJg2Q9VNnKTOPKf2S9WNjKX2WKXLP9c1fKT2e4UNYGVNi2e3o5\\NnF\\Ng2X3\\6eNaFbNQ1HGQ3g8T2oFkMR9`62O0O101N1O2O0O2O0O2N2O0O2N2N1O2N2N1O2N1O2N2N1O2MaNnFVHP9P7YHbHf7X7lH^HR7b7QI]Hm6c7WIZHh6f7[IYHc6g7aIVH^6j7eIUHY6k7jITHT6l7PJQHo5o7TJPHk5P8WJoGg5R8\\JkGc5W8Q31O2O0O2N1O101N1O2N101N2N2N1O2O1N2N2N2O1N20000001O001O001O0YGWEW8i:^GeE^8R;M2N2N2N1O2O0O2O0O0010O0010O00010O010O00010O01O0O2O0O1O2O0O20O1001O1O1oFfG[6Y8aIjG`6V8\\ImGf6R8VIQHk6o7QITHP7l7lHWHV7h7jF`HS1JT8a8gGbG[8^8`GeGa8[8[GhGf8Y8RGmGn8U8hFSHX9[91O001O00001O000eEhFm9j9M3L4L2N1O\\H\\Fm4b9oJdFPO6P4T9lLlF[Nj0h4X8jLSGeM_1_5]7iL^IV3`6hLhIT3\\6cLgI_3]6YLfId3b6SLaIk3g6mK[IP4o6fKTIX4T7`KoH]4Y7[KhHd4`7SKbHl4Z;O1O1bLSAg1n>UN]Ac1d>YNgA_1Z>]NQBoN\\Of1n>mNcBo0Y`0N3M2N2N2N2N2N2N4L5I6J6J7Ib\\S4"}, "label": "little girl about to throw a baseball"}]}
{"id": 25548, "image": "COCO_train2014_000000025548.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with baseball glove\"."}], "task": "refering", "answer_template": "The \"boy with baseball glove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 21, 22, 23, 24, 38, 39, 40, 41, 55, 56, 57, 58, 72, 73, 74, 75, 88, 89, 90, 91, 92, 93, 104, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 246, 257, 258, 259, 262, 263, 274, 275, 276, 280, 281, 290, 291, 292, 297, 298, 307, 308, 309, 314, 315, 324, 325, 341, 342, 358, 359], "bbox": [0.06402083333333333, 0.03934375, 0.5609375, 0.9493281249999999], "iscrowd": 0, "area": 59010.69034999999, "rle": {"size": [640, 480], "counts": "Umc0d0Zc0e0\\O3L3M3N3M2M3N3L3N2N3L3M3M4K4M2C]M[_Oe2b`0cMW_O_2f`0>M4M3M3M2N3L4M2N3M3M3M2M4M3M2N3M3L3N3M3MjNcAoLZ>d2XBYMg=d2aBVMa=f2gBRM\\=k2kBoL[50f0o2U1VMhNi2W1\\MfNf2X1]MfNd2Y1_MdNa2]1`MaNa2^1aM`N`2_1bM_N_2a1bM]N_2b1cM\\N]2d1eMZN\\2f1fMWNZ2i1hMUNX2k1jMSNU2n1mMcMoNmIT3a8nM`MSOjIn2g8QN]MVOgIi2l8SN[MY2g2hMWMX2i2jMUMU2l2mMRMS2o2nMoLQ2R3QNZLnNhJQ3n8SNQLUOoJg2Q9VNnKTOPKf2S9WNjKX2WKXLP9c1fKT2e4UNYGVNi2e3o5\\NnF\\Ng2X3\\6eNaFbNQ1HGQ3g8T2oFkMR9`62O0O101N1O2O0O2O0O2N2O0O2N2N1O2N2N1O2N1O2N2N1O2MaNnFVHP9P7YHbHf7X7lH^HR7b7QI]Hm6c7WIZHh6f7[IYHc6g7aIVH^6j7eIUHY6k7jITHT6l7PJQHo5o7TJPHk5P8WJoGg5R8\\JkGc5W8Q31O2O0O2N1O101N1O2N101N2N2N1O2O1N2N2N2O1N20000001O001O001O0YGWEW8i:^GeE^8R;M2N2N2N1O2O0O2O0O0010O0010O00010O010O00010O01O0O2O0O1O2O0O20O1001O1O1oFfG[6Y8aIjG`6V8\\ImGf6R8VIQHk6o7QITHP7l7lHWHV7h7jF`HS1JT8a8gGbG[8^8`GeGa8[8[GhGf8Y8RGmGn8U8hFSHX9[91O001O00001O000eEhFm9j9M3L4L2N1O\\H\\Fm4b9oJdFPO6P4T9lLlF[Nj0h4X8jLSGeM_1_5]7iL^IV3`6hLhIT3\\6cLgI_3]6YLfId3b6SLaIk3g6mK[IP4o6fKTIX4T7`KoH]4Y7[KhHd4`7SKbHl4Z;O1O1bLSAg1n>UN]Ac1d>YNgA_1Z>]NQBoN\\Of1n>mNcBo0Y`0N3M2N2N2N2N2N2N4L5I6J6J7Ib\\S4"}, "label": "boy with baseball glove"}]}
{"id": 500686, "image": "COCO_train2014_000000500686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black woman in long blue dress\"."}], "task": "refering", "answer_template": "The \"black woman in long blue dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 30, 31, 32, 47, 48, 49, 50, 65, 66, 67, 68, 69, 83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 209, 210, 212, 213, 227, 228, 245, 246], "bbox": [0.59894, 0.016013513513513513, 0.88078, 0.8568018018018019], "iscrowd": 0, "area": 33848.69155, "rle": {"size": [444, 500], "counts": "_iQ41\\20m88gF2S95bF4Y94[F6_92UF7f90oE:GaNi9]1TFh0d9l1J7I7I7I7I7I7I6K6UKZJ\\1h5`N[J_1k5XNYJg1Q6mMRJQ2Y6bMjI^2`6UMdIj2c6mL`IR3a6jLcIU3_6gLcIX3`6dLbI\\3_6aLcI_3_6\\LcIe3^6XLdIh3^6TLdIk3^6RLcIo3_6mKcIS4^6jKdIV4^6fKdIZ4]6cKdI]4n7000O10O1000O1000cNjFSNW9j1oFSNP9l1UGQNk8l1[GQNe8m1`GPN_8o1gGlMZ8Q2nGjMP8W2THfMk7Z2XHdMo60lG^2\\1[M^6b0SHT2_9XN_Fi1V9dNfF]1X9e10O10O0100O010O1000O10000O1000000O1000000000001O001O001O001gH_KU5a4gJcKY5^4bJgK]5Y4bJhK^5Y4`JiK_5X4_JiKa5W4^JkKb5U4\\JlKd5T4\\JkKf5V4WJjKj5X4SJgKR6W4lIjKV6V4gIkK\\6T4aImKb6R4YIQLj6m3SIULP7i3lHZLW7d3fH^L]7`3_HcLd7Z3ZHhLi7V3THlLm7T3PHoLn7S3oGoLQ8_4N1O2N2O5J3M2O0O2N1]KdIe1^6XNeIg1\\6WNfIg1[6WNhIg1Z6WNiIf1X6YNjIf1W6YNjIe1X6YNjIe1W6[NjIc1Y6ZNjId1X6ZNiId1Y6ZNiId1Z6YNhIe1[6WLmIn0J_2V8QMmGR1RODi9VOWGn0YO_On9_OkFj0_OEj9^OhFe0GJk:N^EOe:HcE6_:CiE:o;M3M2NXhi0"}, "label": "black woman in long blue dress"}]}
{"id": 500686, "image": "COCO_train2014_000000500686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a blue dress\"."}], "task": "refering", "answer_template": "The \"a lady in a blue dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 30, 31, 32, 47, 48, 49, 50, 65, 66, 67, 68, 69, 83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 209, 210, 212, 213, 227, 228, 245, 246], "bbox": [0.59894, 0.016013513513513513, 0.88078, 0.8568018018018019], "iscrowd": 0, "area": 33848.69155, "rle": {"size": [444, 500], "counts": "_iQ41\\20m88gF2S95bF4Y94[F6_92UF7f90oE:GaNi9]1TFh0d9l1J7I7I7I7I7I7I6K6UKZJ\\1h5`N[J_1k5XNYJg1Q6mMRJQ2Y6bMjI^2`6UMdIj2c6mL`IR3a6jLcIU3_6gLcIX3`6dLbI\\3_6aLcI_3_6\\LcIe3^6XLdIh3^6TLdIk3^6RLcIo3_6mKcIS4^6jKdIV4^6fKdIZ4]6cKdI]4n7000O10O1000O1000cNjFSNW9j1oFSNP9l1UGQNk8l1[GQNe8m1`GPN_8o1gGlMZ8Q2nGjMP8W2THfMk7Z2XHdMo60lG^2\\1[M^6b0SHT2_9XN_Fi1V9dNfF]1X9e10O10O0100O010O1000O10000O1000000O1000000000001O001O001O001gH_KU5a4gJcKY5^4bJgK]5Y4bJhK^5Y4`JiK_5X4_JiKa5W4^JkKb5U4\\JlKd5T4\\JkKf5V4WJjKj5X4SJgKR6W4lIjKV6V4gIkK\\6T4aImKb6R4YIQLj6m3SIULP7i3lHZLW7d3fH^L]7`3_HcLd7Z3ZHhLi7V3THlLm7T3PHoLn7S3oGoLQ8_4N1O2N2O5J3M2O0O2N1]KdIe1^6XNeIg1\\6WNfIg1[6WNhIg1Z6WNiIf1X6YNjIf1W6YNjIe1X6YNjIe1W6[NjIc1Y6ZNjId1X6ZNiId1Y6ZNiId1Z6YNhIe1[6WLmIn0J_2V8QMmGR1RODi9VOWGn0YO_On9_OkFj0_OEj9^OhFe0GJk:N^EOe:HcE6_:CiE:o;M3M2NXhi0"}, "label": "a lady in a blue dress"}]}
{"id": 500686, "image": "COCO_train2014_000000500686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old woman in a white hat\"."}], "task": "refering", "answer_template": "The \"an old woman in a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 38, 39, 40, 56, 57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 145, 146, 147, 148, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 218, 236, 237], "bbox": [0.0, 0.08988738738738739, 0.32127999999999995, 0.8606756756756756], "iscrowd": 0, "area": 24373.359199999995, "rle": {"size": [444, 500], "counts": "R:2e=6J5L5L300O2N101N100O101N100O1O2O0O100O101N100O1O101N100O100O1SNhNTGZ1j8hNUGY1i8iNUGX1j8kNTGV1j8lNTGV1k8lNSGU1k8mNSGU1k8mNTGT1j8oNSGS1k8oNTGR1j8QOSGP1l8RORGP1l8ROQGQ1m8ROPGP1n8ROPGP1n8UOlFn0R9YOgFi0a6JeICCf0h6N\\IBFa0n65SIe0:bMc6P5PJ^Jo5b5U1OeI`J^4_5_KgJ^4X5bKlJ\\4S5dKPKo2@bL`5>TKk2@eL[5`0XKi2_OfLX5a0\\Kg2\\OgLW5?dKg2WOiLT5<lKi2QOiLS59TLl2jNiLQ58\\Lm2cNkLP53bLS3_NhLP8X3QHgLn7Z3RHeLn7\\3SHbLm7_3TH`Lk7`3WH^Li7c3WH]Lh7d3YHZLg7OfGU3d0kLe70hGU3d0iLe70iGW3\\9iLeFW3Z9hLgFY3X9gLiFY3V9fLlFY3T9gLmFY3R9fLPGZ3S8dLRH2LZ3n7TMiGA:\\3i7fMXHZ2d7hM^HX2^7kMcHS2[7PNfHn1X7WNgHg1W7]NhHb1V7bNjH\\1T7hNlHV1R7POlHn0R7ZOjHd0U7BhH>U7HgH9V7MgH3W7Z3M3N2M3N2M3N2M3N3L3N2M3N2M3O10000O10000O101O1O1N101O1O1N101O1N2O001O1N101O1N2Nf0YO5L5J5K6K4L5K4M2M2gNoEdNS:W1TFfNm9U1ZFgNh9S1`FiNb9R1eFkN\\9o0lFmNV9n0QGoNP9k0WGROk8i0\\GSOf8g0bGVO_8e0_2K5J6KPZc4"}, "label": "an old woman in a white hat"}]}
{"id": 500686, "image": "COCO_train2014_000000500686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white woman sitting on a bench wearing a white hat\"."}], "task": "refering", "answer_template": "The \"a white woman sitting on a bench wearing a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 38, 39, 40, 56, 57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 145, 146, 147, 148, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 218, 236, 237], "bbox": [0.0, 0.08988738738738739, 0.32127999999999995, 0.8606756756756756], "iscrowd": 0, "area": 24373.359199999995, "rle": {"size": [444, 500], "counts": "R:2e=6J5L5L300O2N101N100O101N100O1O2O0O100O101N100O1O101N100O100O1SNhNTGZ1j8hNUGY1i8iNUGX1j8kNTGV1j8lNTGV1k8lNSGU1k8mNSGU1k8mNTGT1j8oNSGS1k8oNTGR1j8QOSGP1l8RORGP1l8ROQGQ1m8ROPGP1n8ROPGP1n8UOlFn0R9YOgFi0a6JeICCf0h6N\\IBFa0n65SIe0:bMc6P5PJ^Jo5b5U1OeI`J^4_5_KgJ^4X5bKlJ\\4S5dKPKo2@bL`5>TKk2@eL[5`0XKi2_OfLX5a0\\Kg2\\OgLW5?dKg2WOiLT5<lKi2QOiLS59TLl2jNiLQ58\\Lm2cNkLP53bLS3_NhLP8X3QHgLn7Z3RHeLn7\\3SHbLm7_3TH`Lk7`3WH^Li7c3WH]Lh7d3YHZLg7OfGU3d0kLe70hGU3d0iLe70iGW3\\9iLeFW3Z9hLgFY3X9gLiFY3V9fLlFY3T9gLmFY3R9fLPGZ3S8dLRH2LZ3n7TMiGA:\\3i7fMXHZ2d7hM^HX2^7kMcHS2[7PNfHn1X7WNgHg1W7]NhHb1V7bNjH\\1T7hNlHV1R7POlHn0R7ZOjHd0U7BhH>U7HgH9V7MgH3W7Z3M3N2M3N2M3N2M3N3L3N2M3N2M3O10000O10000O101O1O1N101O1O1N101O1N2O001O1N101O1N2Nf0YO5L5J5K6K4L5K4M2M2gNoEdNS:W1TFfNm9U1ZFgNh9S1`FiNb9R1eFkN\\9o0lFmNV9n0QGoNP9k0WGROk8i0\\GSOf8g0bGVO_8e0_2K5J6KPZc4"}, "label": "a white woman sitting on a bench wearing a white hat"}]}
{"id": 279503, "image": "COCO_train2014_000000279503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"translucent bucket of green sludge\"."}], "task": "refering", "answer_template": "The \"translucent bucket of green sludge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 129, 130, 131, 132, 133, 148, 149, 150, 151, 166, 167, 168], "bbox": [0.15046, 0.31408, 0.4124, 0.7259200000000001], "iscrowd": 0, "area": 14063.102050000003, "rle": {"size": [375, 500], "counts": "Wdk04^;=C9J3N1O2M3N2N2M2O2N2M3N2N2M4M3M2N3L4M2O2M3N1N3N2M3M2O2M2O2M3N1N3N2M2O2M3N1N3M2O2N2M2O2N2N1N3N2N1O2N1N3N2N1O2M2O001O001N10001O0O2O001O000O2O0000000001O01O0001O000[MVI=k6@ZI=f6BZI?f6_O[Ib0e6\\O\\Ie0e6YO[Ig0f6WO[Ij0e6UOZIm0f6QO[IP1f6mN[IS1f6lNZIU1f6iN[IX1e6fN\\I[1e6cN[I]1f6aNZIa1f6]N[Id1e6[N[If1f6WN[Ii1f6VNZIk1f6SN[Im1f6QN[IP2f6nMZIS2f6kMZIV2g6hMZIY2f6fMZIZ2h6cMYI^2g6`MZIa2f6^MZIb2h6[MYIf2g6YMYIg2i6VMWIl2j6QMWIP3j6nLVIR3l6kLUIV3e700001N1101N101N`0Aa0^Ob0_O9F5L4K5LX^[3"}, "label": "translucent bucket of green sludge"}]}
{"id": 279503, "image": "COCO_train2014_000000279503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bottle with green chattini\"."}], "task": "refering", "answer_template": "The \"white bottle with green chattini\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 129, 130, 131, 132, 133, 148, 149, 150, 151, 166, 167, 168], "bbox": [0.15046, 0.31408, 0.4124, 0.7259200000000001], "iscrowd": 0, "area": 14063.102050000003, "rle": {"size": [375, 500], "counts": "Wdk04^;=C9J3N1O2M3N2N2M2O2N2M3N2N2M4M3M2N3L4M2O2M3N1N3N2M3M2O2M2O2M3N1N3N2M2O2M3N1N3M2O2N2M2O2N2N1N3N2N1O2N1N3N2N1O2M2O001O001N10001O0O2O001O000O2O0000000001O01O0001O000[MVI=k6@ZI=f6BZI?f6_O[Ib0e6\\O\\Ie0e6YO[Ig0f6WO[Ij0e6UOZIm0f6QO[IP1f6mN[IS1f6lNZIU1f6iN[IX1e6fN\\I[1e6cN[I]1f6aNZIa1f6]N[Id1e6[N[If1f6WN[Ii1f6VNZIk1f6SN[Im1f6QN[IP2f6nMZIS2f6kMZIV2g6hMZIY2f6fMZIZ2h6cMYI^2g6`MZIa2f6^MZIb2h6[MYIf2g6YMYIg2i6VMWIl2j6QMWIP3j6nLVIR3l6kLUIV3e700001N1101N101N`0Aa0^Ob0_O9F5L4K5LX^[3"}, "label": "white bottle with green chattini"}]}
{"id": 977, "image": "COCO_train2014_000000000977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is an orange tabby cat lying on its side\"."}], "task": "refering", "answer_template": "The \"this is an orange tabby cat lying on its side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 266, 267, 268, 290], "bbox": [0.31121875, 0.27956639566395663, 0.9435781249999999, 0.9655826558265581], "iscrowd": 0, "area": 36786.59189999999, "rle": {"size": [369, 640], "counts": "lmW21`;0O2O001N101N2O001N101N2O0O2O001N101N2O001N101N2O001N101N2O0O2O001N101O1O001N101O1O0000000O100000O100000O100000000000000O10O1000000000O1000000000000O10O1000000000O10000000000000O01000000000000O010000O100O10O010000O100O10O010000OWGlNW7T1hHPOU7o0kHSOT7m0lHSOT7l0lHVOR7k0nHUOR7k0mHWOR7h0oHXOQ7h0nHZOQ7e0oH\\OQ7d0oH]OP7b0PI_Oo6b0QI_On6a0QI@o6?RIBm6>RICn6<RIFm6:SIFm69SIIk68UIHk68TIJk65VIKj65UILk63UIOj61VIOj61TI1k6OUI2k6NTI3l6LTI5l6KSI6m6IRI:m6FRI;n6EQI<o6CQI>n6CQI>o6APIa0P7_OoHb0Q7@lHb0S7@jHa0V7@hHa0X7AdHa0[7BbH?]7CaH>]7F`H;^7H`H9_7J]H9`7K]H6a7M]H4a70\\H1c7o1N3M3M3M3N2M4L3M3M3N2M3N2N2N2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2M4M2N3L3M4L3M2N1O1O1O1O1O1O1O1O2N1O1O02O1N2O1O1N101N2O1N2O1O1N2O1N101N2O1N2O1O1N2O001O100O10O0100O10O10O100O10O0100O1000OeJlJZ4U5aKPK]4R5]KUKa4l4ZKYKd4h4WK^Kg4d4TKaKk4c5N2N3N1N2N2N2O1N2N3@?_Oa0G9N2N2N3M2O1N2000000O2O000000000O101O000000000O2O00000000000O2O000000000O2O00000000000O1000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000000O101O0000001O1O001O1O1O1O1O001O1O1O1N2O001O1O1O1O001O1O1L4L4L3M4L4L4L4L4M2M4L4L4Lci<"}, "label": "this is an orange tabby cat lying on its side"}]}
{"id": 977, "image": "COCO_train2014_000000000977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mature cat stretching on a leather couch\"."}], "task": "refering", "answer_template": "The \"a mature cat stretching on a leather couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 266, 267, 268, 290], "bbox": [0.31121875, 0.27956639566395663, 0.9435781249999999, 0.9655826558265581], "iscrowd": 0, "area": 36786.59189999999, "rle": {"size": [369, 640], "counts": "lmW21`;0O2O001N101N2O001N101N2O0O2O001N101N2O001N101N2O001N101N2O0O2O001N101O1O001N101O1O0000000O100000O100000O100000000000000O10O1000000000O1000000000000O10O1000000000O10000000000000O01000000000000O010000O100O10O010000O100O10O010000OWGlNW7T1hHPOU7o0kHSOT7m0lHSOT7l0lHVOR7k0nHUOR7k0mHWOR7h0oHXOQ7h0nHZOQ7e0oH\\OQ7d0oH]OP7b0PI_Oo6b0QI_On6a0QI@o6?RIBm6>RICn6<RIFm6:SIFm69SIIk68UIHk68TIJk65VIKj65UILk63UIOj61VIOj61TI1k6OUI2k6NTI3l6LTI5l6KSI6m6IRI:m6FRI;n6EQI<o6CQI>n6CQI>o6APIa0P7_OoHb0Q7@lHb0S7@jHa0V7@hHa0X7AdHa0[7BbH?]7CaH>]7F`H;^7H`H9_7J]H9`7K]H6a7M]H4a70\\H1c7o1N3M3M3M3N2M4L3M3M3N2M3N2N2N2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2M4M2N3L3M4L3M2N1O1O1O1O1O1O1O1O2N1O1O02O1N2O1O1N101N2O1N2O1O1N2O1N101N2O1N2O1O1N2O001O100O10O0100O10O10O100O10O0100O1000OeJlJZ4U5aKPK]4R5]KUKa4l4ZKYKd4h4WK^Kg4d4TKaKk4c5N2N3N1N2N2N2O1N2N3@?_Oa0G9N2N2N3M2O1N2000000O2O000000000O101O000000000O2O00000000000O2O000000000O2O00000000000O1000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000000O101O0000001O1O001O1O1O1O1O001O1O1O1N2O001O1O1O1O001O1O1L4L4L3M4L4L4L4L4M2M4L4L4Lci<"}, "label": "a mature cat stretching on a leather couch"}]}
{"id": 977, "image": "COCO_train2014_000000000977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white semi - long haired cat sleeping on a couch\"."}], "task": "refering", "answer_template": "The \"a black and white semi - long haired cat sleeping on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145], "bbox": [0.01984375, 0.3161246612466125, 0.34221874999999996, 0.5333333333333333], "iscrowd": 0, "area": 12886.595550000002, "rle": {"size": [369, 640], "counts": "Pk4>S;1O1O010O001O10O01O0010O`NBmG>S8FiG:X8IdG8[8MaG2_88VGIj8:SGFm8>nFCS9?jFAV9b0fF_OZ9\\1001O000001O01O000000001O0000000010O000O100O100O100O100O1O1000000000000000000O100000000000000000000000000000000000000000001O000000000000001O00000000000000001O000000O10000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O000000000001O0010O01O001O0010O010O010O01O10O0100O1O1N101N2N1O2N1O2N1N3N1O2N1N3N1O2N3L3N2N1O2M3N2N2M2O2MTbg4"}, "label": "a black and white semi - long haired cat sleeping on a couch"}]}
{"id": 977, "image": "COCO_train2014_000000000977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and black cat\"."}], "task": "refering", "answer_template": "The \"white and black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145], "bbox": [0.01984375, 0.3161246612466125, 0.34221874999999996, 0.5333333333333333], "iscrowd": 0, "area": 12886.595550000002, "rle": {"size": [369, 640], "counts": "Pk4>S;1O1O010O001O10O01O0010O`NBmG>S8FiG:X8IdG8[8MaG2_88VGIj8:SGFm8>nFCS9?jFAV9b0fF_OZ9\\1001O000001O01O000000001O0000000010O000O100O100O100O100O1O1000000000000000000O100000000000000000000000000000000000000000001O000000000000001O00000000000000001O000000O10000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O000000000001O0010O01O001O0010O010O010O01O10O0100O1O1N101N2N1O2N1O2N1N3N1O2N1N3N1O2N3L3N2N1O2M3N2N2M2O2MTbg4"}, "label": "white and black cat"}]}
{"id": 977, "image": "COCO_train2014_000000000977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and black cat laying on orange cat\"."}], "task": "refering", "answer_template": "The \"white and black cat laying on orange cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172], "bbox": [0.318734375, 0.3303252032520325, 0.61284375, 0.5707859078590786], "iscrowd": 0, "area": 10363.70035, "rle": {"size": [369, 640], "counts": "YfY26X;6M2M4L10O010O01O010O010O010O10O010O0102M3N1N3N1N2O2M2O2M100O1O100O10O0100O100O10O100000000O10O10000000O2O00001O0001O100SOgN`G[1^8oNXGQ1h8TOTGk0l8UOTGk0l8UOTGk0l8VOTGi0k8YOTGg0l8\\OQGd0o8@nF>S9P100000000O1O001O100O1O1O1O1O010O1O1O1O100000O10000000O100000O10000000O1000O1000000000000O1000000O100000000O1000000O10000000000O101O00000O1000001O000O100000001N100O1O1O1O2N1O1N2N2N2M3N3M2M3N1O2N2N2M2O2N2N2O001O1O001O1O1O00100O1O001N2N3MoWi2"}, "label": "white and black cat laying on orange cat"}]}
{"id": 484307, "image": "COCO_train2014_000000484307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right shoe is on the right of the banana\"."}], "task": "refering", "answer_template": "The \"the right shoe is on the right of the banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 105, 106, 107, 123, 124, 125, 141, 142, 143, 158, 159, 160, 161, 176, 177, 178, 179, 194, 195, 196, 197, 213, 215], "bbox": [0.7835599999999999, 0.30349333333333334, 1.0, 0.8650133333333333], "iscrowd": 0, "area": 15147.442499999997, "rle": {"size": [375, 500], "counts": "_j_43c;4K4L5K4M3L5K4L5K4M3L5K3M3M3N3L3M3M3M3N2M4L3M2N2O1N2N3M2N2O1N2N2O1N2O2M2O1O1N2\\OeLkH\\3n6mLnHU3j6UMSIl2f6^MUId2f6X1O2O1O0O2O1O0O2O1O1N101O1O0O2O1O1N2O1O0O2O1O1O1N2OO101O0001O100O1O1O100O00100O1O1O100O1O2N2O0O2N101N2O0010O01O010O1O010O001O010YL"}, "label": "the right shoe is on the right of the banana"}]}
{"id": 484307, "image": "COCO_train2014_000000484307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis shoe on a right foot\"."}], "task": "refering", "answer_template": "The \"a tennis shoe on a right foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 105, 106, 107, 123, 124, 125, 141, 142, 143, 158, 159, 160, 161, 176, 177, 178, 179, 194, 195, 196, 197, 213, 215], "bbox": [0.7835599999999999, 0.30349333333333334, 1.0, 0.8650133333333333], "iscrowd": 0, "area": 15147.442499999997, "rle": {"size": [375, 500], "counts": "_j_43c;4K4L5K4M3L5K4L5K4M3L5K3M3M3N3L3M3M3M3N2M4L3M2N2O1N2N3M2N2O1N2N2O1N2O2M2O1O1N2\\OeLkH\\3n6mLnHU3j6UMSIl2f6^MUId2f6X1O2O1O0O2O1O0O2O1O1N101O1O0O2O1O1N2O1O0O2O1O1O1N2OO101O0001O100O1O1O100O00100O1O1O100O1O2N2O0O2N101N2O0010O01O010O1O010O001O010YL"}, "label": "a tennis shoe on a right foot"}]}
{"id": 484307, "image": "COCO_train2014_000000484307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a thumb holding the banana\"."}], "task": "refering", "answer_template": "The \"a thumb holding the banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 216, 217, 222, 223, 224], "bbox": [0.0016799999999999999, 0.58072, 0.46918, 0.9865600000000001], "iscrowd": 0, "area": 14418.520999999997, "rle": {"size": [375, 500], "counts": "Sc0V4`701O10O0100O0010O0100O010O1O010O010O10O01O10O010O0100O00100O010O00100O010O10O01O010O10O0100O00100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O100O2N100O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O1_MSGX2m8gMTGW2n8hMTGU2n8jMSGT2o8kMSGR2o8mMRGQ2P9nMQGP2Q9oMQGn1Q9RNoFl1_9N3M2O2M2N2N3M2N2N3M2N4L5L4K5K4L5K5Kim<M\\RC0O100O100O100O100O100O100O100O010O100O100O100O100O100O100O100O10000001O0000001O000O2O00001O0000001O00001O00001O001O1O1O1O1O1O1OQfP3"}, "label": "a thumb holding the banana"}]}
{"id": 484307, "image": "COCO_train2014_000000484307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fingers holding on to a banana\"."}], "task": "refering", "answer_template": "The \"fingers holding on to a banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 216, 217, 222, 223, 224], "bbox": [0.0016799999999999999, 0.58072, 0.46918, 0.9865600000000001], "iscrowd": 0, "area": 14418.520999999997, "rle": {"size": [375, 500], "counts": "Sc0V4`701O10O0100O0010O0100O010O1O010O010O10O01O10O010O0100O00100O010O00100O010O10O01O010O10O0100O00100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O100O2N100O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O1_MSGX2m8gMTGW2n8hMTGU2n8jMSGT2o8kMSGR2o8mMRGQ2P9nMQGP2Q9oMQGn1Q9RNoFl1_9N3M2O2M2N2N3M2N2N3M2N4L5L4K5K4L5K5Kim<M\\RC0O100O100O100O100O100O100O100O010O100O100O100O100O100O100O100O10000001O0000001O000O2O00001O0000001O00001O00001O001O1O1O1O1O1O1OQfP3"}, "label": "fingers holding on to a banana"}]}
{"id": 484307, "image": "COCO_train2014_000000484307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left shoe of a person holding a banana\"."}], "task": "refering", "answer_template": "The \"the left shoe of a person holding a banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 18, 19, 20, 21, 36, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 126, 127, 128, 129], "bbox": [0.0, 0.00224, 0.20394, 0.6157333333333334], "iscrowd": 0, "area": 18669.56205, "rle": {"size": [375, 500], "counts": "f6a0S:S1nNQ1gMZ2N2N2M2O2N2M3L3M4L4L3N3L4L4L3M4L4O1O001O1O1O0O2O1O1O0010000O01000000O01000O100000O010000000O10O100000O1000O1000001O0O101O00001O0O101O1O1O1N2O1N101N2N2O1N2O1N101N2O1RMWJKk52[JJh51^JJg50_JKf5O`JMd5MbJNc5LcJ0a5JeJ1`5IfJ3^5FiJ5\\5EjJ7\\5AjJ:`5ZOhJ`0a5VOgJd0\\8I7J6IVba4"}, "label": "the left shoe of a person holding a banana"}]}
{"id": 484307, "image": "COCO_train2014_000000484307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left shoe\"."}], "task": "refering", "answer_template": "The \"left shoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 18, 19, 20, 21, 36, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 126, 127, 128, 129], "bbox": [0.0, 0.00224, 0.20394, 0.6157333333333334], "iscrowd": 0, "area": 18669.56205, "rle": {"size": [375, 500], "counts": "f6a0S:S1nNQ1gMZ2N2N2M2O2N2M3L3M4L4L3N3L4L4L3M4L4O1O001O1O1O0O2O1O1O0010000O01000000O01000O100000O010000000O10O100000O1000O1000001O0O101O00001O0O101O1O1O1N2O1N101N2N2O1N2O1N101N2O1RMWJKk52[JJh51^JJg50_JKf5O`JMd5MbJNc5LcJ0a5JeJ1`5IfJ3^5FiJ5\\5EjJ7\\5AjJ:`5ZOhJ`0a5VOgJd0\\8I7J6IVba4"}, "label": "left shoe"}]}
{"id": 574420, "image": "COCO_train2014_000000574420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in dark gray shirt\"."}], "task": "refering", "answer_template": "The \"the man in dark gray shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 156, 157, 158, 159, 160, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 218, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.48390625, 0.13032710280373833, 0.9753281250000001, 1.0], "iscrowd": 0, "area": 46066.98445000001, "rle": {"size": [428, 640], "counts": "heR41[=2N2N3M2N2N3M2N2N2N001O001O001O00001O001O01O010O10O01O10O01O001O1O001O1O001O1O010O001O1O001O00O10000O10000O100O10001O00001O001O001O1O001O0O2O001O001O1O00001O00001O00000O2O0000001O00001O0000001O00000O2O00001O0000001O00001O0000000O100000000000000000000000000O00100O1O100O100O1O100O1O100O100O1O100O00100O010O1O0010O01O10O01O10O01O010O1O010O1O015J8H8I3L2N3M2O1N3VLRMQMP3m2SMQMn2m2TMQMn2m2UMPMn2l2VMRMk2l2XMQMj2m2YMUJBZ2X3\\3mMaLT2Y3SNdLo1U3YNhLj1P3_NmLb1m2eNPM]1k2iNRMY1k2lNRMV1k2lNSM`1`2bN_Mg1X2[NgMg1U2[NjMh1R2[NlMg1R2ZNmMi1o1YNQNi1k1ZNSNi1j1XNUNk1g1XNXNi1e1YNZNj1c1WN\\Nl1a1VN^Nl1_1UN`Nn1]1TNaNo1\\1RNcNP2[1RNdNP2Y1QNfNU2U1kMjN_2l0cMSO]2l0dMSO]2l0eMRO\\2m0fMQO[2n0gMPOZ2o0hMoNY2P1hMoNY2P1iMnNX2R1iMlNX2S1jMkNW2T1kMjNV2U1lMiNU2W1lMfNV2Z1kMdNV2\\1kMbNV2]1lMaNU2_1lM_NU2a1kM^NV2b1kM\\NV2d1kMZNV2e1lMYNU2g1lMVNV2j1f401N1000000O1000000001O001O00001O001O00001O001O001O00001O001O0O2O1O1O1O001O1O1O001O1O1O001N2O1O001O001O001O001O001WG_Me6a2VIfMh6[2RIlMk6U2PIRNn6o1kHXNT7h1gH_NW7b1cHeN[7[1`HlN\\7W1]HQO_7Q1\\HVO`7m0ZHZOb7h0YH_O_7]3G8WJdKW2f4]MkKU2]4_MULS2S4bM^LP2j7A?ZOf0ZOU\\6"}, "label": "the man in dark gray shirt"}]}
{"id": 574420, "image": "COCO_train2014_000000574420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a man with thin rimmed glasses and a black scarf\"."}], "task": "refering", "answer_template": "The \"this is a man with thin rimmed glasses and a black scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 156, 157, 158, 159, 160, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 218, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.48390625, 0.13032710280373833, 0.9753281250000001, 1.0], "iscrowd": 0, "area": 46066.98445000001, "rle": {"size": [428, 640], "counts": "heR41[=2N2N3M2N2N3M2N2N2N001O001O001O00001O001O01O010O10O01O10O01O001O1O001O1O001O1O010O001O1O001O00O10000O10000O100O10001O00001O001O001O1O001O0O2O001O001O1O00001O00001O00000O2O0000001O00001O0000001O00000O2O00001O0000001O00001O0000000O100000000000000000000000000O00100O1O100O100O1O100O1O100O100O1O100O00100O010O1O0010O01O10O01O10O01O010O1O010O1O015J8H8I3L2N3M2O1N3VLRMQMP3m2SMQMn2m2TMQMn2m2UMPMn2l2VMRMk2l2XMQMj2m2YMUJBZ2X3\\3mMaLT2Y3SNdLo1U3YNhLj1P3_NmLb1m2eNPM]1k2iNRMY1k2lNRMV1k2lNSM`1`2bN_Mg1X2[NgMg1U2[NjMh1R2[NlMg1R2ZNmMi1o1YNQNi1k1ZNSNi1j1XNUNk1g1XNXNi1e1YNZNj1c1WN\\Nl1a1VN^Nl1_1UN`Nn1]1TNaNo1\\1RNcNP2[1RNdNP2Y1QNfNU2U1kMjN_2l0cMSO]2l0dMSO]2l0eMRO\\2m0fMQO[2n0gMPOZ2o0hMoNY2P1hMoNY2P1iMnNX2R1iMlNX2S1jMkNW2T1kMjNV2U1lMiNU2W1lMfNV2Z1kMdNV2\\1kMbNV2]1lMaNU2_1lM_NU2a1kM^NV2b1kM\\NV2d1kMZNV2e1lMYNU2g1lMVNV2j1f401N1000000O1000000001O001O00001O001O00001O001O001O00001O001O0O2O1O1O1O001O1O1O001O1O1O001N2O1O001O001O001O001O001WG_Me6a2VIfMh6[2RIlMk6U2PIRNn6o1kHXNT7h1gH_NW7b1cHeN[7[1`HlN\\7W1]HQO_7Q1\\HVO`7m0ZHZOb7h0YH_O_7]3G8WJdKW2f4]MkKU2]4_MULS2S4bM^LP2j7A?ZOf0ZOU\\6"}, "label": "this is a man with thin rimmed glasses and a black scarf"}]}
{"id": 574420, "image": "COCO_train2014_000000574420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the beard who is holding a cat\"."}], "task": "refering", "answer_template": "The \"the man with the beard who is holding a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236], "bbox": [0.023984375, 0.1255607476635514, 0.349375, 0.7399065420560748], "iscrowd": 0, "area": 31199.524199999996, "rle": {"size": [428, 640], "counts": "h_64V=S1nN9F9G7J0O2N2NQOlDIT;8nDFP;<SE@l:a0WE]Oh:d0[EYOc:i0`ESO_:n0dEPO[:Q1hEkNW:W1lEfNR:[1QFcNm9_1o01N2N101N2N101N2N1O2O1O1a0^OR1oNTc4Lb\\K=D=B=C>D;_EfMT:X2kEjM\\1Nm6X2fGkM[1Oo6U2fGnMX1OR7S2eGPNV1OT7Q2fGQNb0IC6U8R2cGQNa0O@0\\8U2]GmMe04\\OMa8b2oGUNP8m1mGUNS8k1jGXNV8i1gGYNY8h1cG[N]8e1aG\\N_8Y300000000O100000OVN[KoJe4o4ULXJl3f5WLYJi3e5ZLZJf3d5]L[Jc3c5`L\\J`3a5dL]J]3a5fL^JZ3`5iL_JV3`5mL_JS3_5PM`JP3^5SM`Jn2^5UMaJk2\\5YMcJg2[5\\MdJd2Z5_MeJa2Y5bMeJ_2Y5k2N101000J^JbHb5^75100000O10O100000000O010000000O1000O1000O100000001N<E0000000O10001O0001O1O001O1O001O100O001O1O001O1O1O00100O1O001O1O001O1gJ[Ji2f5QM_Jo2a5lLeJT3[5fLjJY3X5aLnJ^3R5]LSKb3o4XLWKg3j4SL[Kl3T7O0O2O001N101N101N1N3J5K6J5nNWG\\Mo8W2Y1E<C<L5N1O2N1O2N1O1O2N16Ko1PN01O0O2O1O1N101O1O1N2O001N2A?[Oe0[Od0\\Oe0[Onf]5"}, "label": "the man with the beard who is holding a cat"}]}
{"id": 574420, "image": "COCO_train2014_000000574420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses at a candle lit table holding a cat\"."}], "task": "refering", "answer_template": "The \"a man with glasses at a candle lit table holding a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236], "bbox": [0.023984375, 0.1255607476635514, 0.349375, 0.7399065420560748], "iscrowd": 0, "area": 31199.524199999996, "rle": {"size": [428, 640], "counts": "h_64V=S1nN9F9G7J0O2N2NQOlDIT;8nDFP;<SE@l:a0WE]Oh:d0[EYOc:i0`ESO_:n0dEPO[:Q1hEkNW:W1lEfNR:[1QFcNm9_1o01N2N101N2N101N2N1O2O1O1a0^OR1oNTc4Lb\\K=D=B=C>D;_EfMT:X2kEjM\\1Nm6X2fGkM[1Oo6U2fGnMX1OR7S2eGPNV1OT7Q2fGQNb0IC6U8R2cGQNa0O@0\\8U2]GmMe04\\OMa8b2oGUNP8m1mGUNS8k1jGXNV8i1gGYNY8h1cG[N]8e1aG\\N_8Y300000000O100000OVN[KoJe4o4ULXJl3f5WLYJi3e5ZLZJf3d5]L[Jc3c5`L\\J`3a5dL]J]3a5fL^JZ3`5iL_JV3`5mL_JS3_5PM`JP3^5SM`Jn2^5UMaJk2\\5YMcJg2[5\\MdJd2Z5_MeJa2Y5bMeJ_2Y5k2N101000J^JbHb5^75100000O10O100000000O010000000O1000O1000O100000001N<E0000000O10001O0001O1O001O1O001O100O001O1O001O1O1O00100O1O001O1O001O1gJ[Ji2f5QM_Jo2a5lLeJT3[5fLjJY3X5aLnJ^3R5]LSKb3o4XLWKg3j4SL[Kl3T7O0O2O001N101N101N1N3J5K6J5nNWG\\Mo8W2Y1E<C<L5N1O2N1O2N1O1O2N16Ko1PN01O0O2O1O1N101O1O1N2O001N2A?[Oe0[Od0\\Oe0[Onf]5"}, "label": "a man with glasses at a candle lit table holding a cat"}]}
{"id": 50134, "image": "COCO_train2014_000000050134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch in a living room\"."}], "task": "refering", "answer_template": "The \"a brown couch in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 337, 338, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6112968750000001, 0.6666666666666666, 1.0, 1.0], "iscrowd": 0, "area": 26472.408399999997, "rle": {"size": [480, 640], "counts": "ngg51j>5L4L4L4K5L4L4L4K5L4L4O1O1O100O1O100O1O100O1O100O1O100O1O100O100O100O1000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O000000000000001O00000000O1O1O1O1N2O1I7G9G9F:G9G9F;H7N2N2N2M3N2N2N2N2M3N2N2O100O1O100O1000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000VOj0\\Nd1[NbG"}, "label": "a brown couch in a living room"}]}
{"id": 50134, "image": "COCO_train2014_000000050134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair\"."}], "task": "refering", "answer_template": "The \"chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 337, 338, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6112968750000001, 0.6666666666666666, 1.0, 1.0], "iscrowd": 0, "area": 26472.408399999997, "rle": {"size": [480, 640], "counts": "ngg51j>5L4L4L4K5L4L4L4K5L4L4O1O1O100O1O100O1O100O1O100O1O100O1O100O100O100O1000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O000000000000001O00000000O1O1O1O1N2O1I7G9G9F:G9G9F;H7N2N2N2M3N2N2N2N2M3N2N2O100O1O100O1000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000VOj0\\Nd1[NbG"}, "label": "chair"}]}
{"id": 50134, "image": "COCO_train2014_000000050134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown sofa next to the lamp , across the room from the television\"."}], "task": "refering", "answer_template": "The \"the brown sofa next to the lamp , across the room from the television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 279, 280, 302], "bbox": [0.13415625, 0.48814583333333333, 0.3863125, 0.8114166666666667], "iscrowd": 0, "area": 12233.0406, "rle": {"size": [480, 640], "counts": "eaX1a0_>g0YOg0XOh0YOh0XOg0XO9HH8I7J501N2O0O2O1N101N2O0O2O1N20O10O10O1000O010000O10O1000002M3N2N2N2M3N1O2N2N2M1000000O100000000O1000000O1000000O100000000O1000000O100000000O10000O101O000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O2M2N3M3L4M3N2O1O1N2O1O1O001O000000001O000O10001O000000001O00000O101O00000000001K4J6I7I7Igng5"}, "label": "the brown sofa next to the lamp , across the room from the television"}]}
{"id": 50134, "image": "COCO_train2014_000000050134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch sits empty behind the coffee table\"."}], "task": "refering", "answer_template": "The \"the couch sits empty behind the coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 279, 280, 302], "bbox": [0.13415625, 0.48814583333333333, 0.3863125, 0.8114166666666667], "iscrowd": 0, "area": 12233.0406, "rle": {"size": [480, 640], "counts": "eaX1a0_>g0YOg0XOh0YOh0XOg0XO9HH8I7J501N2O0O2O1N101N2O0O2O1N20O10O10O1000O010000O10O1000002M3N2N2N2M3N1O2N2N2M1000000O100000000O1000000O1000000O100000000O1000000O100000000O10000O101O000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O2M2N3M3L4M3N2O1O1N2O1O1O001O000000001O000O10001O000000001O00000O101O00000000001K4J6I7I7Igng5"}, "label": "the couch sits empty behind the coffee table"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow jacket\"."}], "task": "refering", "answer_template": "The \"man in yellow jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 133, 134, 135, 136, 156, 157, 158, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 288, 289, 290, 291, 310, 311, 312, 313, 332, 333, 334, 335, 354, 355, 356, 357, 376, 377, 378, 379, 400], "bbox": [0.03812091503267974, 0.24782679738562088, 0.2708823529411764, 0.8334477124183005], "iscrowd": 0, "area": 30963.146500000003, "rle": {"size": [612, 612], "counts": "aP>2Pc0;D:G3L5L3M4L4M2M4L4L3N3L2N3N2M2O2N2O1N1O2N2O0O2N2N2ZNcMbB^2Y=hMfBY2T=nMiBS2S=RNkBo1T=SNjBo1U=SNhBo1T4VMj3m0QHn1R4XMk3l0QHn1R4WMl3n0oGm1R4XMm3m0PHm1o3XMP4m0oGm1f1QMC8g6k0nGo1a1UMG2i6l0nGo1\\1XMJ0j6k0nGa2g0jL>Lk6k0nGe2`0jLc0Hn6k0nGd4Q1cJQ7j0lGf4o0bJT7j0lGe4k0dJZ7h0iGf4f0hJ`7c0iGh4`0iJh7`0gGh4;mJm7=fGh48nJS8;cGi43RKY8l810O010O100O010O1000000O10O10000000000O100O1M3M3M3N2N2N2O1N2N2N3iMVGPJl8Q6WGiIk8X6WGdIj8]6XG`Ii8`6YG\\Ih8d6[GXIf8j6[GSIe8U7VGgHl8c7lFYHf0BS7a8nGjGm00l6Y8WHcGl06m6Z8WH\\Gl0;n6[8WHVGh0b0Q7Y8YIiGg6X8UIkGl6U8QImGP7S8lHPHV7o7gHSHZ7m7bHUH`7k7\\HXHe7h7WH[Hj7n7jGTHX8[91O1hFnG[7T8cHnG\\7S8aHoG^7T8_HnG_7T8_HnG_7V8\\HmGc7W8XHkGf7[8THgGj7_8oGcGP8j9N2O1N3M3M3eJ\\EQ2f:hMeES2]:fMlET2X:hMnEQ2W:lMnE^1eNkLb;e1VFS1_NnL_;n1_Ff0P:XORFf0P:WOSFg0n9XOUFe0m9YOUFe0l9ZOVFd0l9ZOWFb0k9\\OXFb0j9\\OYFa0h9^O[F?g9^O\\F>j9^OYF>l9]OZF<k9_OeF2a9EoFMY9JVGHQ9Oe]Z8"}, "label": "man in yellow jacket"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow jacket\"."}], "task": "refering", "answer_template": "The \"man in yellow jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 133, 134, 135, 136, 156, 157, 158, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 288, 289, 290, 291, 310, 311, 312, 313, 332, 333, 334, 335, 354, 355, 356, 357, 376, 377, 378, 379, 400], "bbox": [0.03812091503267974, 0.24782679738562088, 0.2708823529411764, 0.8334477124183005], "iscrowd": 0, "area": 30963.146500000003, "rle": {"size": [612, 612], "counts": "aP>2Pc0;D:G3L5L3M4L4M2M4L4L3N3L2N3N2M2O2N2O1N1O2N2O0O2N2N2ZNcMbB^2Y=hMfBY2T=nMiBS2S=RNkBo1T=SNjBo1U=SNhBo1T4VMj3m0QHn1R4XMk3l0QHn1R4WMl3n0oGm1R4XMm3m0PHm1o3XMP4m0oGm1f1QMC8g6k0nGo1a1UMG2i6l0nGo1\\1XMJ0j6k0nGa2g0jL>Lk6k0nGe2`0jLc0Hn6k0nGd4Q1cJQ7j0lGf4o0bJT7j0lGe4k0dJZ7h0iGf4f0hJ`7c0iGh4`0iJh7`0gGh4;mJm7=fGh48nJS8;cGi43RKY8l810O010O100O010O1000000O10O10000000000O100O1M3M3M3N2N2N2O1N2N2N3iMVGPJl8Q6WGiIk8X6WGdIj8]6XG`Ii8`6YG\\Ih8d6[GXIf8j6[GSIe8U7VGgHl8c7lFYHf0BS7a8nGjGm00l6Y8WHcGl06m6Z8WH\\Gl0;n6[8WHVGh0b0Q7Y8YIiGg6X8UIkGl6U8QImGP7S8lHPHV7o7gHSHZ7m7bHUH`7k7\\HXHe7h7WH[Hj7n7jGTHX8[91O1hFnG[7T8cHnG\\7S8aHoG^7T8_HnG_7T8_HnG_7V8\\HmGc7W8XHkGf7[8THgGj7_8oGcGP8j9N2O1N3M3M3eJ\\EQ2f:hMeES2]:fMlET2X:hMnEQ2W:lMnE^1eNkLb;e1VFS1_NnL_;n1_Ff0P:XORFf0P:WOSFg0n9XOUFe0m9YOUFe0l9ZOVFd0l9ZOWFb0k9\\OXFb0j9\\OYFa0h9^O[F?g9^O\\F>j9^OYF>l9]OZF<k9_OeF2a9EoFMY9JVGHQ9Oe]Z8"}, "label": "man in yellow jacket"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skiis being worn by a man with an unzipped jacket\"."}], "task": "refering", "answer_template": "The \"skiis being worn by a man with an unzipped jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [344, 346, 366, 368, 369, 389, 390, 391, 392, 411, 412, 414, 415, 434, 437, 438], "bbox": [0.6454084967320262, 0.6964052287581699, 0.9229901960784314, 0.8858169934640523], "iscrowd": 0, "area": 2585.3485500000006, "rle": {"size": [612, 612], "counts": "h_\\73Rc03L2M3N1O2M2O2N1N3N2O1N1ON3L3N5K5K6Ji0O\\O1O001N101O0O2O000O1000000O01000O1N2N2N1N3N2N2N2M3N1O2M3M3000O1000O100UNnNbAR1[>POfAo0X>ROiAo0U>ROlAn0Q>TOPBj0P>WOQBg0n=ZOSBe0m=\\OTBb0l=_OUB?j=CWB;i=FXB8h=IXB6h=KYB3g=NZB0f=1[BMf=4YBKh=5YBIg=8ZBFg=:ZBDg=<YBCg=>ZB@g=`0[B]Of=c0^2000010Da]O0_b0Oc]O0]b0Oe]O0[b0Oh]OOXb00j]OOVb00k]O0Ub0Nm]O2Sb0Mn]O3Rb0Lo]O4ab01O000000000000000000001O0000000001O0000010O000000001O0001O01O0000001O00O101N100O100O100O100O2O0O100O1O100M3J600001O0000000O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N3NVUl0"}, "label": "skiis being worn by a man with an unzipped jacket"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue outfit standing in the snow\"."}], "task": "refering", "answer_template": "The \"a man in blue outfit standing in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 164, 165, 166, 185, 186, 187, 188, 189, 206, 207, 208, 209, 210, 211, 228, 229, 230, 231, 232, 233, 250, 251, 252, 253, 254, 255, 273, 274, 275, 276, 277, 295, 296, 297, 298, 317, 318, 319, 320, 321, 338, 339, 340, 342, 343, 360, 361, 364, 365, 382, 383], "bbox": [0.3920751633986928, 0.26217320261437904, 0.6198529411764706, 0.8067156862745097], "iscrowd": 0, "area": 26324.272349999992, "rle": {"size": [612, 612], "counts": "gf_44nb09jBGR7a0VH\\OWK`0R<h0ZHIa7?SHKh7=kGLP86jG0T83dG4Z8N`G9^8I[G=c8FXG>f8DYG=f8EYG<d8GZG:d8H\\G7c8L\\G4c8N[G2d80UG6j8NRG4m8NQG2n81mF2S90cF6_9LWF:k9H[En0g:AjD=Y;EdD9^;d32N2O1N3N1N2O0O2N1O2N1N3N1O2N1N3N1O1O1NYOdIlD[6U;SJ]Dl5d;h01O100O1O2N3N2M2N3M3N2M3M3N2M3M3M2O2M3M3O13M3M3M4L3M2M3O1O101N1O1O100O1009H6I0000000001O0O100O1O101N1002N00O1O2M2O1O1N2O1O2M2O1OO01000]MnCSNQ<l1SDRNn;k1WDQNk;l1ZDRNf;k1_DRNa;m1cDPN^;m1gDPNZ;m1kDPNV;m1nDRNR;k1RESNo:j1UETNl:i1YETNg:j1]ESNe:c1fEZN\\:_1lE^NV:^1oE^NT:\\1SF]NT:\\1VFXNV:^1P4F;D;F:F:DPoZ4"}, "label": "a man in blue outfit standing in the snow"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue ski jacket and black helmet and black gloves on a snowy mountain\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue ski jacket and black helmet and black gloves on a snowy mountain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 164, 165, 166, 185, 186, 187, 188, 189, 206, 207, 208, 209, 210, 211, 228, 229, 230, 231, 232, 233, 250, 251, 252, 253, 254, 255, 273, 274, 275, 276, 277, 295, 296, 297, 298, 317, 318, 319, 320, 321, 338, 339, 340, 342, 343, 360, 361, 364, 365, 382, 383], "bbox": [0.3920751633986928, 0.26217320261437904, 0.6198529411764706, 0.8067156862745097], "iscrowd": 0, "area": 26324.272349999992, "rle": {"size": [612, 612], "counts": "gf_44nb09jBGR7a0VH\\OWK`0R<h0ZHIa7?SHKh7=kGLP86jG0T83dG4Z8N`G9^8I[G=c8FXG>f8DYG=f8EYG<d8GZG:d8H\\G7c8L\\G4c8N[G2d80UG6j8NRG4m8NQG2n81mF2S90cF6_9LWF:k9H[En0g:AjD=Y;EdD9^;d32N2O1N3N1N2O0O2N1O2N1N3N1O2N1N3N1O1O1NYOdIlD[6U;SJ]Dl5d;h01O100O1O2N3N2M2N3M3N2M3M3N2M3M3M2O2M3M3O13M3M3M4L3M2M3O1O101N1O1O100O1009H6I0000000001O0O100O1O101N1002N00O1O2M2O1O1N2O1O2M2O1OO01000]MnCSNQ<l1SDRNn;k1WDQNk;l1ZDRNf;k1_DRNa;m1cDPN^;m1gDPNZ;m1kDPNV;m1nDRNR;k1RESNo:j1UETNl:i1YETNg:j1]ESNe:c1fEZN\\:_1lE^NV:^1oE^NT:\\1SF]NT:\\1VFXNV:^1P4F;D;F:F:DPoZ4"}, "label": "a man wearing a blue ski jacket and black helmet and black gloves on a snowy mountain"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy wearing blue setter and black pants and a cooling glass\"."}], "task": "refering", "answer_template": "The \"a guy wearing blue setter and black pants and a cooling glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 148, 149, 150, 169, 170, 171, 172, 191, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 301, 302, 303, 304, 323, 324, 325, 326, 345, 346, 347, 348, 367, 369, 370, 389, 392], "bbox": [0.6869444444444445, 0.25, 0.8806372549019609, 0.8153104575163399], "iscrowd": 0, "area": 24204.467199999985, "rle": {"size": [612, 612], "counts": "gXk72ob0a0@a0^BQOn7a1_GfN[8l1RG\\Nh8l1oF[Nj8n1lFYNn8o1iFYNQ9o1fFWNT9Q2cFWNV9R2aFUNY9R2_FTN[9P2bFRN\\9Q2aFQN^9R2]FQNa9R2\\FPNc9R2YFQNf9R2VFPNh9S2TFPNk9Q2SFQNk9Q2RFmLR;T3lDkLV;V3hDjLX;W3fDiL\\;X3bDgL`;Z3]DfLd;\\3ZDcLh;^3UDbLl;_3SDaLn;_3PDaLR<_3mCaLS<_3mC`LT<a3jC`LV<a3iC^LW<W50]NhC[LY<f3gCYLY<h3gCVLZ<[3oC]KHV1Y<_3VD_Lk;b3UD]Lk;d3UDZLl;g3TDXLl;i3TDULm;m3SDQLn;o3RDoKP<Q4PDnKP<S4PDlKQ<T4oCjKS<W4mCgKS<Z4gCjK[<^50O2O1N101O11N5L4K5L4L4M3N2M3N2M3M3N2UFbHj7a7gGmHW8V7YGXIg8j6jFdIT9X8O1N2N2O1N2N2O05L4L4L3M2N3M2N1N2O001O2L4K4M2N1O1N3N1O2M2O2VOWEUIj:`6eE[I]:j2QGVNQOiNP:^2WJ]Mk5`2[J[Mf5d2^5O6JW1iN6J4L3M3M3L4M3M3HbZ[1"}, "label": "a guy wearing blue setter and black pants and a cooling glass"}]}
{"id": 328663, "image": "COCO_train2014_000000328663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid wearing sun glasses in the snow\"."}], "task": "refering", "answer_template": "The \"a kid wearing sun glasses in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 148, 149, 150, 169, 170, 171, 172, 191, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 301, 302, 303, 304, 323, 324, 325, 326, 345, 346, 347, 348, 367, 369, 370, 389, 392], "bbox": [0.6869444444444445, 0.25, 0.8806372549019609, 0.8153104575163399], "iscrowd": 0, "area": 24204.467199999985, "rle": {"size": [612, 612], "counts": "gXk72ob0a0@a0^BQOn7a1_GfN[8l1RG\\Nh8l1oF[Nj8n1lFYNn8o1iFYNQ9o1fFWNT9Q2cFWNV9R2aFUNY9R2_FTN[9P2bFRN\\9Q2aFQN^9R2]FQNa9R2\\FPNc9R2YFQNf9R2VFPNh9S2TFPNk9Q2SFQNk9Q2RFmLR;T3lDkLV;V3hDjLX;W3fDiL\\;X3bDgL`;Z3]DfLd;\\3ZDcLh;^3UDbLl;_3SDaLn;_3PDaLR<_3mCaLS<_3mC`LT<a3jC`LV<a3iC^LW<W50]NhC[LY<f3gCYLY<h3gCVLZ<[3oC]KHV1Y<_3VD_Lk;b3UD]Lk;d3UDZLl;g3TDXLl;i3TDULm;m3SDQLn;o3RDoKP<Q4PDnKP<S4PDlKQ<T4oCjKS<W4mCgKS<Z4gCjK[<^50O2O1N101O11N5L4K5L4L4M3N2M3N2M3M3N2UFbHj7a7gGmHW8V7YGXIg8j6jFdIT9X8O1N2N2O1N2N2O05L4L4L3M2N3M2N1N2O001O2L4K4M2N1O1N3N1O2M2O2VOWEUIj:`6eE[I]:j2QGVNQOiNP:^2WJ]Mk5`2[J[Mf5d2^5O6JW1iN6J4L3M3M3L4M3M3HbZ[1"}, "label": "a kid wearing sun glasses in the snow"}]}
{"id": 304088, "image": "COCO_train2014_000000304088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing baseball wearing red helmet outside the green field\"."}], "task": "refering", "answer_template": "The \"a man playing baseball wearing red helmet outside the green field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 99, 100, 101, 102, 123, 124, 125, 146, 147, 169, 170, 192, 193, 215, 216, 217, 238, 239], "bbox": [0.323203125, 0.3119375, 0.46284375000000005, 0.9763437500000001], "iscrowd": 0, "area": 8256.957449999996, "rle": {"size": [320, 640], "counts": "aiP23j95L3O1O1N2O1O1O001O1O1O1O1O2N100^HF\\5;^JLa55XJ2f5OTJ8`1ZOZ1?oL>c1VO]1>iLc0f1ROa1c2[N`Md1b2XN`Mi1a2TNaMk1`2RNbMo1_2mMdMR2]2kMeMV2[2gMhMX2Y2eMiM\\2X2\\MoMc2R2VMTNj2m1RMVNo2j1nLXNR3j1jLXNV3i1gLYNY3j1bLXN^3k1]LWNc3l1YLUNh3X3VL\\Kj3b4[L[Ke3c4_L[Ka3c4bL\\K^3b4fL\\K[3a4hL^KV3b4nL\\Ki2k4ZMTKd2l4_MSKa2l4aMQK`2n4cMPK^2n4dMQKk2`4WM^Kj2`4\\1O11lJaKS42oKT22BF^NX4^2eLfNQOmNZ4Z2oLcNeNTO\\4U2YMnNf2o0aMmN_2P1iMkNW2R1oMlNP2Q1UNmNk1P1[NlNf1P1`NmNa1n0fNPOZ1Q1fNnN[19bKHU3MZ19cKJT3KY1:eKJS3KY18fKNR3HX19hKNP3JX16jK0n2IX16kK1m2IZ13kK3k2J[12kK4j2I]10kK6h2J_1MkK8g2J`1KkK;d2Jb1JkK;d2Id1IjK=b2Ja25`MJa26^MJc25^MJc25]MJg23ZMLh22XMNk2OVM0l2NUM1m][3"}, "label": "a man playing baseball wearing red helmet outside the green field"}]}
{"id": 304088, "image": "COCO_train2014_000000304088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man getting ready to swing a baseball bat\"."}], "task": "refering", "answer_template": "The \"a man getting ready to swing a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 99, 100, 101, 102, 123, 124, 125, 146, 147, 169, 170, 192, 193, 215, 216, 217, 238, 239], "bbox": [0.323203125, 0.3119375, 0.46284375000000005, 0.9763437500000001], "iscrowd": 0, "area": 8256.957449999996, "rle": {"size": [320, 640], "counts": "aiP23j95L3O1O1N2O1O1O001O1O1O1O1O2N100^HF\\5;^JLa55XJ2f5OTJ8`1ZOZ1?oL>c1VO]1>iLc0f1ROa1c2[N`Md1b2XN`Mi1a2TNaMk1`2RNbMo1_2mMdMR2]2kMeMV2[2gMhMX2Y2eMiM\\2X2\\MoMc2R2VMTNj2m1RMVNo2j1nLXNR3j1jLXNV3i1gLYNY3j1bLXN^3k1]LWNc3l1YLUNh3X3VL\\Kj3b4[L[Ke3c4_L[Ka3c4bL\\K^3b4fL\\K[3a4hL^KV3b4nL\\Ki2k4ZMTKd2l4_MSKa2l4aMQK`2n4cMPK^2n4dMQKk2`4WM^Kj2`4\\1O11lJaKS42oKT22BF^NX4^2eLfNQOmNZ4Z2oLcNeNTO\\4U2YMnNf2o0aMmN_2P1iMkNW2R1oMlNP2Q1UNmNk1P1[NlNf1P1`NmNa1n0fNPOZ1Q1fNnN[19bKHU3MZ19cKJT3KY1:eKJS3KY18fKNR3HX19hKNP3JX16jK0n2IX16kK1m2IZ13kK3k2J[12kK4j2I]10kK6h2J_1MkK8g2J`1KkK;d2Jb1JkK;d2Id1IjK=b2Ja25`MJa26^MJc25^MJc25]MJg23ZMLh22XMNk2OVM0l2NUM1m][3"}, "label": "a man getting ready to swing a baseball bat"}]}
{"id": 197591, "image": "COCO_train2014_000000197591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in pink on skiis\"."}], "task": "refering", "answer_template": "The \"a woman in pink on skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 98, 99, 120, 121, 122, 123, 142, 143, 144, 145, 146, 166, 167, 168, 189, 190, 191, 213, 214, 236, 237, 259, 260], "bbox": [0.20720312500000002, 0.2566355140186916, 0.379625, 0.8007710280373832], "iscrowd": 0, "area": 11208.579000000002, "rle": {"size": [428, 640], "counts": "Qig12X=2N2O1O1N2O2N1O1O1O1N2O1O1O1O1O1O1O2M2O1O1O1O1M0O11011O0bDoNb:Q1\\E[OZ:f0bEFU::iE;d9EPFl0k9^1K6K5K5F:I6I7I7J6I7I7J6L4L4L4Lc0]O>C0O1O101N10000000001O000000000011N2N2N1O1O2O0O1kLkIlN\\6R1RJZNU6e1YJaMT6^2_2N1O1O2N1O2M2N2N3nN[FXNg9Z1]1O2N8H6K1N1O101N1O2N1O2N1O101N1O2N1O2N1N2M4L3JYhU5"}, "label": "a woman in pink on skiis"}]}
{"id": 197591, "image": "COCO_train2014_000000197591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing pink jacket\"."}], "task": "refering", "answer_template": "The \"woman wearing pink jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 98, 99, 120, 121, 122, 123, 142, 143, 144, 145, 146, 166, 167, 168, 189, 190, 191, 213, 214, 236, 237, 259, 260], "bbox": [0.20720312500000002, 0.2566355140186916, 0.379625, 0.8007710280373832], "iscrowd": 0, "area": 11208.579000000002, "rle": {"size": [428, 640], "counts": "Qig12X=2N2O1O1N2O2N1O1O1O1N2O1O1O1O1O1O1O2M2O1O1O1O1M0O11011O0bDoNb:Q1\\E[OZ:f0bEFU::iE;d9EPFl0k9^1K6K5K5F:I6I7I7J6I7I7J6L4L4L4Lc0]O>C0O1O101N10000000001O000000000011N2N2N1O1O2O0O1kLkIlN\\6R1RJZNU6e1YJaMT6^2_2N1O1O2N1O2M2N2N3nN[FXNg9Z1]1O2N8H6K1N1O101N1O2N1O2N1O101N1O2N1O2N1N2M4L3JYhU5"}, "label": "woman wearing pink jacket"}]}
{"id": 304092, "image": "COCO_train2014_000000304092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a medium sized elephant standing between a large elephant and some baby elephants\"."}], "task": "refering", "answer_template": "The \"a medium sized elephant standing between a large elephant and some baby elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173], "bbox": [0.216640625, 0.30264583333333334, 0.5454375, 0.4567083333333334], "iscrowd": 0, "area": 10616.769149999996, "rle": {"size": [480, 640], "counts": "`ZQ23k>4L4L3N0O1O2N1O101O1O100O100O100O1O100O010O100O3N2N1O2M3N2]BnNT=_101O1N2O00000O100000000000000000O100000000010O00000001O00001O00001O001O00001O00001O000000001O0000001O000O10001O0O100O10000O100O1001O0000001O0000001O0000001N100000O100000O100O100O10O0100O100O101N2O1O1O001O1kBZNm<m100000O100000O1000N2O10000000O100000001O1O2N1O1N2O1O00O100O100O10000O100O10O10O100O10001N100O2O00001O01O010O0010O1O1O1O2N1O2N2N2N1O2N2N2N2N3K6K4K5L5J5L8GlVX4"}, "label": "a medium sized elephant standing between a large elephant and some baby elephants"}]}
{"id": 304092, "image": "COCO_train2014_000000304092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the largest elephant\"."}], "task": "refering", "answer_template": "The \"the largest elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 177, 178, 200, 201, 223], "bbox": [0.315171875, 0.218125, 0.7550625, 0.5417291666666667], "iscrowd": 0, "area": 15581.400399999997, "rle": {"size": [480, 640], "counts": "cjn21n>1O2N1O2O0O2N1O2N1O1O2N1O2N1O2O0O1O2N1O2N1O2O0O2N001O10O01O1O10O01O1O00100O001O100O001O10O10000O100000O010000O1000000O10000O10000O1O0100000000000000001O0000001O00001O0000001O00000000000000000000000001O0O10001O0000001O0000001O000000001O000O101O0000001O00001O1O1O1O1O001O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1N2O1N2O1O1N2O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1OO2O000O10000O100O10000O100O2O000O10000O100O10000O100O1000000010O000000001O000001O100O1O001O100O1O100O100O1O100O1O100O1O010O1O100O100O1O1002N4M3M3N2M3M3N2M3N2M3MO2M2O2M2N2N3N1N2N3M2O2M2N2N3N1N3H7H8H9H7H9G8I^\\Y2"}, "label": "the largest elephant"}]}
{"id": 304092, "image": "COCO_train2014_000000304092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant behind three other elephants\"."}], "task": "refering", "answer_template": "The \"an elephant behind three other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 177, 178, 200, 201, 223], "bbox": [0.315171875, 0.218125, 0.7550625, 0.5417291666666667], "iscrowd": 0, "area": 15581.400399999997, "rle": {"size": [480, 640], "counts": "cjn21n>1O2N1O2O0O2N1O2N1O1O2N1O2N1O2O0O1O2N1O2N1O2O0O2N001O10O01O1O10O01O1O00100O001O100O001O10O10000O100000O010000O1000000O10000O10000O1O0100000000000000001O0000001O00001O0000001O00000000000000000000000001O0O10001O0000001O0000001O000000001O000O101O0000001O00001O1O1O1O1O001O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1N2O1N2O1O1N2O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1OO2O000O10000O100O10000O100O2O000O10000O100O10000O100O1000000010O000000001O000001O100O1O001O100O1O100O100O1O100O1O100O1O010O1O100O100O1O1002N4M3M3N2M3M3N2M3N2M3MO2M2O2M2N2N3N1N2N3M2O2M2N2N3N1N3H7H8H9H7H9G8I^\\Y2"}, "label": "an elephant behind three other elephants"}]}
{"id": 99293, "image": "COCO_train2014_000000099293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with curly hair drinks a glass of milk and has a donut\"."}], "task": "refering", "answer_template": "The \"a boy with curly hair drinks a glass of milk and has a donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284], "bbox": [0.11756249999999999, 0.24944444444444444, 0.42598437499999997, 0.9842777777777779], "iscrowd": 0, "area": 35353.1928, "rle": {"size": [360, 640], "counts": "Qfj05a:c0^Ob0J5L5L4L3L5L4K4M4K5L3M4K5L3L5L4K4M4L4K4M4K5L3L5M3N2N1O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1[O_JbKb5\\4aJbK`5[4cJdK^5Z4eJdK\\5Y4hJeKY5Y4jJeKW5Y4kJgKU5V4oJhKR5V4QKhKP5U4TKiKm4U4VKiKk4T4V1N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N10100O1O100O100O1O100O100O1O100O100O1O100O100O100000000001O000000000000001O0000000000001O000000001O001O0L5C<C>C<D=B=D=C<C=H9O010O0010O01O010O0010O00010O01O010O0010O01O01ObMUI?l6_OYI=g6A_I<b6@eI;^6AgI;[6AkI<W6@nI<T6@TJ:o5A[J8f5EcJ3`5HjJ0X5LQKNQ5NUKMm4OYKNi4N\\KNg4M^K0Q8MhhP4"}, "label": "a boy with curly hair drinks a glass of milk and has a donut"}]}
{"id": 99293, "image": "COCO_train2014_000000099293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy drinking milk\"."}], "task": "refering", "answer_template": "The \"a boy drinking milk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284], "bbox": [0.11756249999999999, 0.24944444444444444, 0.42598437499999997, 0.9842777777777779], "iscrowd": 0, "area": 35353.1928, "rle": {"size": [360, 640], "counts": "Qfj05a:c0^Ob0J5L5L4L3L5L4K4M4K5L3M4K5L3L5L4K4M4L4K4M4K5L3L5M3N2N1O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1[O_JbKb5\\4aJbK`5[4cJdK^5Z4eJdK\\5Y4hJeKY5Y4jJeKW5Y4kJgKU5V4oJhKR5V4QKhKP5U4TKiKm4U4VKiKk4T4V1N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N10100O1O100O100O1O100O100O1O100O100O1O100O100O100000000001O000000000000001O0000000000001O000000001O001O0L5C<C>C<D=B=D=C<C=H9O010O0010O01O010O0010O00010O01O010O0010O01O01ObMUI?l6_OYI=g6A_I<b6@eI;^6AgI;[6AkI<W6@nI<T6@TJ:o5A[J8f5EcJ3`5HjJ0X5LQKNQ5NUKMm4OYKNi4N\\KNg4M^K0Q8MhhP4"}, "label": "a boy drinking milk"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a potted plant next to a bench with a lady on it\"."}], "task": "refering", "answer_template": "The \"a potted plant next to a bench with a lady on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 38, 39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 109, 110, 111, 131, 132, 133, 154, 155, 156, 157, 177, 178, 179, 180, 201, 202, 225, 248, 271], "bbox": [0.665546875, 0.0013411764705882352, 0.9380312499999999, 0.7748941176470587], "iscrowd": 0, "area": 20975.1947, "rle": {"size": [425, 640], "counts": "ak`54T=8I8G9H7\\DnNb:[1REhNo:`1fDcNY;f110O01000O010O010O010O10OfFUNW7l1hHXNU7g1jH^NR7c1jHdNS7[1kHlNQ7T1lHROQ7n0lHXOP7h0PI\\Om6d0SI@i6?WIEf6;ZIHc67]IM_64aI0[60dI4Y6KhI8U6HjI9U6GlI:S6FlI;T6DlI>S6BmI>R6CnI>Q6APJ?P6APJ?P6@PJb0n5_ORJa0n5^OSJc0l5]OTJc0l5]OTJd0k5[OUJf0j5[OVJf0i5YOXJg0h5YOXJh0c4aMkJf1c0i0b4cMhJe1f0h0b4eMfJc1g0j0b4eMeJ`1l0i0^4jMdJ]1P1h0[4mMbJ[1U1h0X4oMaJY1Y1k0Q4nMdJV1]1m0m3oMdJW1^1h0n3SNaJ[1]1a0R4VN_J_1[19V4ZN]Jb1Y13Z4]N[Jf1W1K^4aNXJj1S1Ge4aNVJn1k0Eo4a1hJcNX5b1_JcN`5c1VJaNj5_400000000000000000001O0hKkH\\2U7UM_If2a6lLSJn2Y3dLbM:^Om2f2SMlM0Dg2V2cMVNFJa2f1SN`N\\OK`2\\1^NhNROL`2[1_NiNQOKa2]1]NhNROKa2^1\\NgNSOPO@^OQ3k2\\NgNSOkNFAl2n2[NfNROgNMCf2Q3ZNeNSObN4F_20PM\\2Z1\\OSO^N:GZ2GZMj2n0ZOTOYNa0Jh2e2cMVOUOWNe0Mb2g2dMTOVOSNk0O[2l2cMROWOoMQ11T2OZLh2Y1XOYOkMW14l11\\Lh2X1WOX1nMT13]Li2W1TO[1nMP15_Li2V1TO]1lMm07aLQ2L_NX1\\1S5TNjIQ2M^NU1X1W5ZNhIP2N\\NS1S1]5bNbIP2OZNR1l0c5jN]IP2OZNP1f0i5POXIQ21WNn0a0o5WOSIQ22VNk0;V6^OnHR22TNj04\\6FiHR23RNh0Ob6MdHR24QNe0Ii64_HR25PNe8NWGS25mMc80YGS26lM`80[GT27kM]81\\GT29iM[82]GV28hMZ81_GW29gMW81aGX2:eMU82bGY2;dMQ83dG[2<aMo73fG\\2[9cMgF=ZO^1o9TNhF>YO^1o9SNiF?XO_1n9RNjF>ZO_1g:`NYE`1g:`NYEa1f:_N[E`1e:_N\\Ea1e:^N\\ET1KlNi:0\\ET1KmNh:O]ET1LlNg:O_ET1KlNf:0_ET1LkNf:0^EU1LlNe:O`EU1KkNe:OaEV1KjNd:0aEV1LiNd:0aEV1KkNc:ObEV1LjNb:OdEV1KjNa:0dEV1LiNa:0cEW1LjN`:OeEV1LjN_:OfEW1LiN^:0fEW1MhN^:0fEW1LjN]:OgEX1LhN]:OhEY1LgN\\:0iEX1LgN\\:0hEY1LhN[:OjEX1LhNZ:OkEY1LgNY:0lEX1LhNm:X1SEhNn:W1c0O001O010O1O001L4L3L5L4K4L5L4KYa`0"}, "label": "a potted plant next to a bench with a lady on it"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pot holding a large green plant next to a bench with a couple sitting on it\"."}], "task": "refering", "answer_template": "The \"a pot holding a large green plant next to a bench with a couple sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 38, 39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 109, 110, 111, 131, 132, 133, 154, 155, 156, 157, 177, 178, 179, 180, 201, 202, 225, 248, 271], "bbox": [0.665546875, 0.0013411764705882352, 0.9380312499999999, 0.7748941176470587], "iscrowd": 0, "area": 20975.1947, "rle": {"size": [425, 640], "counts": "ak`54T=8I8G9H7\\DnNb:[1REhNo:`1fDcNY;f110O01000O010O010O010O10OfFUNW7l1hHXNU7g1jH^NR7c1jHdNS7[1kHlNQ7T1lHROQ7n0lHXOP7h0PI\\Om6d0SI@i6?WIEf6;ZIHc67]IM_64aI0[60dI4Y6KhI8U6HjI9U6GlI:S6FlI;T6DlI>S6BmI>R6CnI>Q6APJ?P6APJ?P6@PJb0n5_ORJa0n5^OSJc0l5]OTJc0l5]OTJd0k5[OUJf0j5[OVJf0i5YOXJg0h5YOXJh0c4aMkJf1c0i0b4cMhJe1f0h0b4eMfJc1g0j0b4eMeJ`1l0i0^4jMdJ]1P1h0[4mMbJ[1U1h0X4oMaJY1Y1k0Q4nMdJV1]1m0m3oMdJW1^1h0n3SNaJ[1]1a0R4VN_J_1[19V4ZN]Jb1Y13Z4]N[Jf1W1K^4aNXJj1S1Ge4aNVJn1k0Eo4a1hJcNX5b1_JcN`5c1VJaNj5_400000000000000000001O0hKkH\\2U7UM_If2a6lLSJn2Y3dLbM:^Om2f2SMlM0Dg2V2cMVNFJa2f1SN`N\\OK`2\\1^NhNROL`2[1_NiNQOKa2]1]NhNROKa2^1\\NgNSOPO@^OQ3k2\\NgNSOkNFAl2n2[NfNROgNMCf2Q3ZNeNSObN4F_20PM\\2Z1\\OSO^N:GZ2GZMj2n0ZOTOYNa0Jh2e2cMVOUOWNe0Mb2g2dMTOVOSNk0O[2l2cMROWOoMQ11T2OZLh2Y1XOYOkMW14l11\\Lh2X1WOX1nMT13]Li2W1TO[1nMP15_Li2V1TO]1lMm07aLQ2L_NX1\\1S5TNjIQ2M^NU1X1W5ZNhIP2N\\NS1S1]5bNbIP2OZNR1l0c5jN]IP2OZNP1f0i5POXIQ21WNn0a0o5WOSIQ22VNk0;V6^OnHR22TNj04\\6FiHR23RNh0Ob6MdHR24QNe0Ii64_HR25PNe8NWGS25mMc80YGS26lM`80[GT27kM]81\\GT29iM[82]GV28hMZ81_GW29gMW81aGX2:eMU82bGY2;dMQ83dG[2<aMo73fG\\2[9cMgF=ZO^1o9TNhF>YO^1o9SNiF?XO_1n9RNjF>ZO_1g:`NYE`1g:`NYEa1f:_N[E`1e:_N\\Ea1e:^N\\ET1KlNi:0\\ET1KmNh:O]ET1LlNg:O_ET1KlNf:0_ET1LkNf:0^EU1LlNe:O`EU1KkNe:OaEV1KjNd:0aEV1LiNd:0aEV1KkNc:ObEV1LjNb:OdEV1KjNa:0dEV1LiNa:0cEW1LjN`:OeEV1LjN_:OfEW1LiN^:0fEW1MhN^:0fEW1LjN]:OgEX1LhN]:OhEY1LgN\\:0iEX1LgN\\:0hEY1LhN[:OjEX1LhNZ:OkEY1LgNY:0lEX1LhNm:X1SEhNn:W1c0O001O010O1O001L4L3L5L4K4L5L4KYa`0"}, "label": "a pot holding a large green plant next to a bench with a couple sitting on it"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the groom\"."}], "task": "refering", "answer_template": "The \"the groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 147, 148, 149, 150, 151, 166, 170, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 242, 243, 244, 245, 265, 266, 267, 268, 269, 270, 288, 289, 291, 292, 311, 312, 334, 335, 336], "bbox": [0.214890625, 0.07889411764705882, 0.7610625000000001, 0.9867764705882353], "iscrowd": 0, "area": 36763.921299999995, "rle": {"size": [425, 640], "counts": "g^i14Q=5K4L5K4L4M4O0O2O0O2O0O101O0O2O0O2O0O101O0O2O0O2O0O01L4L3O2N1O2N1O2N2M2O20O100O010O10000O100O100O10O10O100O100O10000O100O010O10000O100O100O10O10O100O100O10000O100O010O1N2O1O1O0O2O001O0O2O001O0O2O00001O0O2O0100O100O010O1002N3M3M2aFZNW7i1VHjNg7X1THnNi7U1RHoNk7U1oGQOn7Q1nGSOP8P1jGVOS8l0iGXOU8k0fGZOW8n2N3M2N3M1N10000000001O0000000O101O000000001O000O100[OfJ[IZ5Z6Q1E:N2N2N2M3N3M2N2N2N2N3J50000000001N10000000001O000000001O0000001O00001O0000001O00001O00001OXJeJb3\\5[LfJe3[5XLgJi3X5ULjJn3T5oKnJT4P5iKRKZ4l4cKVK`4h4]KZKe4e4XK]Kk4a4RKaKQ5]4lJdKX5Z4`JmKc5T4PJULS6o3\\I[Lf6S53M3M3M3M5K5K6J5KeKaJQ1Z5mNjJV1P5hNUK`1c4[NaK[2m3^MWLW3X3`LlL_3Z3YLjLe3[3SLjLk3\\3lKhLT4\\3eKhLY4^3aJcM^5V5N1O2N2N1O2N1O1O1O1VLWHR2j7nMZHm1g7SN]Hh1d7XN`Hc1a7]NcH^1^7bNfHY1[7gNiHT1X7lNlHo0U7QOoHj0R7VORIe0o6[OUI]OmNJo7i0XIROSO1f7m0[IgNYO8]7Q1^I\\N^O`0U7T1eJhN\\5X1gJdNZ5\\1hJaNY5_1jJ]NW5c1X300000000000000000000000000000000000000O100000000000000000000000000000000O1O1O1O001O001O001O001O00001O001O001O001O00001O001O0O2N1O2N100O2N1O1O1O100O1O2N1O100O1O2N2N1O2O1N2N1O2N2O1N2N1OmUo1"}, "label": "the groom"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couple sitting on a colorful bench in dress up clothing\"."}], "task": "refering", "answer_template": "The \"a couple sitting on a colorful bench in dress up clothing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 147, 148, 149, 150, 151, 166, 170, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 242, 243, 244, 245, 265, 266, 267, 268, 269, 270, 288, 289, 291, 292, 311, 312, 334, 335, 336], "bbox": [0.214890625, 0.07889411764705882, 0.7610625000000001, 0.9867764705882353], "iscrowd": 0, "area": 36763.921299999995, "rle": {"size": [425, 640], "counts": "g^i14Q=5K4L5K4L4M4O0O2O0O2O0O101O0O2O0O2O0O101O0O2O0O2O0O01L4L3O2N1O2N1O2N2M2O20O100O010O10000O100O100O10O10O100O100O10000O100O010O10000O100O100O10O10O100O100O10000O100O010O1N2O1O1O0O2O001O0O2O001O0O2O00001O0O2O0100O100O010O1002N3M3M2aFZNW7i1VHjNg7X1THnNi7U1RHoNk7U1oGQOn7Q1nGSOP8P1jGVOS8l0iGXOU8k0fGZOW8n2N3M2N3M1N10000000001O0000000O101O000000001O000O100[OfJ[IZ5Z6Q1E:N2N2N2M3N3M2N2N2N2N3J50000000001N10000000001O000000001O0000001O00001O0000001O00001O00001OXJeJb3\\5[LfJe3[5XLgJi3X5ULjJn3T5oKnJT4P5iKRKZ4l4cKVK`4h4]KZKe4e4XK]Kk4a4RKaKQ5]4lJdKX5Z4`JmKc5T4PJULS6o3\\I[Lf6S53M3M3M3M5K5K6J5KeKaJQ1Z5mNjJV1P5hNUK`1c4[NaK[2m3^MWLW3X3`LlL_3Z3YLjLe3[3SLjLk3\\3lKhLT4\\3eKhLY4^3aJcM^5V5N1O2N2N1O2N1O1O1O1VLWHR2j7nMZHm1g7SN]Hh1d7XN`Hc1a7]NcH^1^7bNfHY1[7gNiHT1X7lNlHo0U7QOoHj0R7VORIe0o6[OUI]OmNJo7i0XIROSO1f7m0[IgNYO8]7Q1^I\\N^O`0U7T1eJhN\\5X1gJdNZ5\\1hJaNY5_1jJ]NW5c1X300000000000000000000000000000000000000O100000000000000000000000000000000O1O1O1O001O001O001O001O00001O001O001O001O00001O001O0O2N1O2N100O2N1O1O1O100O1O2N1O100O1O2N2N1O2O1N2N1O2N2O1N2N1OmUo1"}, "label": "a couple sitting on a colorful bench in dress up clothing"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green plant to the left of a bench\"."}], "task": "refering", "answer_template": "The \"a green plant to the left of a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 162, 163, 164, 184, 185, 186, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 257, 277, 278, 279], "bbox": [0.0015, 0.3825882352941176, 0.207421875, 0.872470588235294], "iscrowd": 0, "area": 10533.236949999999, "rle": {"size": [425, 640], "counts": "bb08P=a0@a0^Oa0@`0_Oa0@a0^Oa0@`0_Oa0@2M2O0011N2O1N2O2nKcGR3_8jLdGV3]8gLeGX3]8dLgG[3Z8aLiG^3Y8^LjGa3X8\\LjGd3X8WLkGh3l801N100O100O101N10O001N101O0000[N^FDa9<`FC`9<aFE]9<dFC\\9=dFD[9;gFDY9<gFEX9:iFFW9:jFFU9:kFFU99mFGQ9:oFFQ99QGGn89RGGn89RGHm87UGaNJa0Q9n0UGaNO<l8S1VG`NO=i8S1YG`NO<h8T1ZG_NO=f8S1\\G`NO>c8R1_G_NO`0`82oF1b0]O0`0^81QG1b0_OO`0\\80SG1b0_O0a0Y8NWG0a0A0a0V8NZG0`0A1b0S8L^G0>B1d0Q8IaG0>C1d0o7HcG1=C2e0l7GfGO=E2f0i7EiG0;F4e0g7DkG1:F5f0c7CPH08G6d0c7DPH17H6a0d7FoG17H7?d7GPH15I8=d7HPH24I9;d7IPH33I9:e7IQH31J:8e7KPH31J;6d7MQH30J<4d7NRH3ML>1d7ORH4LL?Od70SH4JM`0Md72RH4JMa0Kd73RH5INa0He74SH5GOb0Fe75SH6FOc0Dd77TH6EOd0Bd79TH5D0d0Ae79TH6C0c9IkF7B0c9HmF6A2b9GnF7_O3c9FnF7_O3b9FQG6]O4b9ERG7\\O5a9CUG7ZO6a9BVG8YO6a9BVG8YO6a9AXG8WO7a9@YG8WO8`9_OZG9VO8_9_O]G8TO9_9_O]G8TO9R;GmD:T;EmD:S;FnD9R;GoD9Q;EQE:o:FQE:o:FRE9o:FRE9n:GSE8n:GSE8m:HTE7l:HVE7k:HVE7j:IVE7j:IWE7i:HXE7h:IYE6g:JZE5n;O1O1O001O1OSab6"}, "label": "a green plant to the left of a bench"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in white dress sitting on bench\"."}], "task": "refering", "answer_template": "The \"woman in white dress sitting on bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 177, 197, 198, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 269, 270, 290, 291, 292, 293, 314, 315, 316, 338, 339], "bbox": [0.56534375, 0.09797647058823529, 0.7911875, 0.9763529411764705], "iscrowd": 0, "area": 22887.286249999997, "rle": {"size": [425, 640], "counts": "P\\f46R=8H8H8H8G9RNWOmF_1\\8cNcGV2b7nM\\Hh2m6ZMQIi2l6ZMRIh2k6[MSIg2j6[MUIg2i6[MUIg2i6[MTIh2j6YMUIi2h6YMWIi2g6YMXIh2f6YMYIi2e6YMZIh2c6ZM\\Ih2b6YM^Ih2`6ZM^Ih2`6YM`Ih2^6ZM`Ih2^6YMbIh2]6XMbIj2]6WMbIj2]6WMaIk2^6VMaIl2]6Q2N2O1O1O1O1O2N1N2O1O1O2N1O1O1N2O1O2N1O1O1O1dL]HQ1d7lN_HU1^7jNfHU1Y7jNiHV1V7hNnHW1Q7gNSIX1l6gNVIY1h6fN\\IY1c6eNYIWN_OU3V7bNXI^NDo2S7bNUIeNIh2Q7aNSIlNLe2n6^NRITOO_2Q1kLb4a1ZJZO2]2m0jLh4]1VJB4X2i0mLm4^2YJh0e0lLR5]2YJj0>mLY5Y2XJm0:lL_5W2WJP13mLf5R2WJT1NlLl5P2VJl1i5TNWJl1j5TNVJl1j5SNVJn1j5RNVJo1h5QNXJP2g5QNYJo1f5QNZJP2e5QN[Jo1d5kLZJC2d3T1]KQ3[1kKD0d3d0RL[3e0TLCMg35fLc30\\LCLd5b3hJdLCJj5]3bJlLBHn5Y3_JQMCER6V3ZJWMCDU6R3WJ]MC@Y6P3SJbMC_O\\6l2QJfMC]O_6j2mIkMC\\Ob6f2jIPNDYOc6f2hISNDXOd6d2gIVNEUOe6d2eIYNETOf6b2dI\\NFQOg6c2bI]NFQOh6b2`I_NHnNh6d2^I`NImNi6c2\\IbNKjNj6e2XIcNMiNk6c2WIfNNfNl6d2TIhNOeNl6c2UIhN0dNl6d2RIjN1cNm6c2PImN1`NP7b2nHPO1_NQ7a2lHRO2]NS7`2jHVO0\\NU7_2jHVO0[NW7_2gHXO1ZNX7]2fH[O1WN[7^2bH^O1SN_7^2]HC2oMb7_2WHG6jMd7l4[HSKg7n4WHQKk7o4THPKn7]3WHkMJgNP8Y3_HoMc7l1fHnM\\7l1mHoMU7T1lIfNV6W1lIhNV6U1lIjNW6Q1mImNU6P1mIPOU6l0nIROT6j0oIUOT6g0nIXOT6d0PJZOS6c0nI\\OT6a0oI]OT6`0mI_OU6>mIAV6<lIAW6=jIAY6<jIAZ6<[3N3L3N3LW[g1"}, "label": "woman in white dress sitting on bench"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white strapless dress\"."}], "task": "refering", "answer_template": "The \"a woman in a white strapless dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 177, 197, 198, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 269, 270, 290, 291, 292, 293, 314, 315, 316, 338, 339], "bbox": [0.56534375, 0.09797647058823529, 0.7911875, 0.9763529411764705], "iscrowd": 0, "area": 22887.286249999997, "rle": {"size": [425, 640], "counts": "P\\f46R=8H8H8H8G9RNWOmF_1\\8cNcGV2b7nM\\Hh2m6ZMQIi2l6ZMRIh2k6[MSIg2j6[MUIg2i6[MUIg2i6[MTIh2j6YMUIi2h6YMWIi2g6YMXIh2f6YMYIi2e6YMZIh2c6ZM\\Ih2b6YM^Ih2`6ZM^Ih2`6YM`Ih2^6ZM`Ih2^6YMbIh2]6XMbIj2]6WMbIj2]6WMaIk2^6VMaIl2]6Q2N2O1O1O1O1O2N1N2O1O1O2N1O1O1N2O1O2N1O1O1O1dL]HQ1d7lN_HU1^7jNfHU1Y7jNiHV1V7hNnHW1Q7gNSIX1l6gNVIY1h6fN\\IY1c6eNYIWN_OU3V7bNXI^NDo2S7bNUIeNIh2Q7aNSIlNLe2n6^NRITOO_2Q1kLb4a1ZJZO2]2m0jLh4]1VJB4X2i0mLm4^2YJh0e0lLR5]2YJj0>mLY5Y2XJm0:lL_5W2WJP13mLf5R2WJT1NlLl5P2VJl1i5TNWJl1j5TNVJl1j5SNVJn1j5RNVJo1h5QNXJP2g5QNYJo1f5QNZJP2e5QN[Jo1d5kLZJC2d3T1]KQ3[1kKD0d3d0RL[3e0TLCMg35fLc30\\LCLd5b3hJdLCJj5]3bJlLBHn5Y3_JQMCER6V3ZJWMCDU6R3WJ]MC@Y6P3SJbMC_O\\6l2QJfMC]O_6j2mIkMC\\Ob6f2jIPNDYOc6f2hISNDXOd6d2gIVNEUOe6d2eIYNETOf6b2dI\\NFQOg6c2bI]NFQOh6b2`I_NHnNh6d2^I`NImNi6c2\\IbNKjNj6e2XIcNMiNk6c2WIfNNfNl6d2TIhNOeNl6c2UIhN0dNl6d2RIjN1cNm6c2PImN1`NP7b2nHPO1_NQ7a2lHRO2]NS7`2jHVO0\\NU7_2jHVO0[NW7_2gHXO1ZNX7]2fH[O1WN[7^2bH^O1SN_7^2]HC2oMb7_2WHG6jMd7l4[HSKg7n4WHQKk7o4THPKn7]3WHkMJgNP8Y3_HoMc7l1fHnM\\7l1mHoMU7T1lIfNV6W1lIhNV6U1lIjNW6Q1mImNU6P1mIPOU6l0nIROT6j0oIUOT6g0nIXOT6d0PJZOS6c0nI\\OT6a0oI]OT6`0mI_OU6>mIAV6<lIAW6=jIAY6<jIAZ6<[3N3L3N3LW[g1"}, "label": "a woman in a white strapless dress"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bush in the rightmost pot\"."}], "task": "refering", "answer_template": "The \"a bush in the rightmost pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275], "bbox": [0.843140625, 0.42247058823529415, 1.0, 0.7528], "iscrowd": 0, "area": 11758.60015, "rle": {"size": [425, 640], "counts": "jZP78P=3N3L3N1N2_Oa0N2N2O1N2N3M6J5K5K5L?@i0WO5K4M4L001N1000001O000O101O0000001N10000O10000O100O10000O100O2O000O100O10000O1000001O000O10000000000000001O000000000000000000000000000000000000000000000000O10000000\\J"}, "label": "a bush in the rightmost pot"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant on right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"plant on right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275], "bbox": [0.843140625, 0.42247058823529415, 1.0, 0.7528], "iscrowd": 0, "area": 11758.60015, "rle": {"size": [425, 640], "counts": "jZP78P=3N3L3N1N2_Oa0N2N2O1N2N3M6J5K5K5L?@i0WO5K4M4L001N1000001O000O101O0000001N10000O10000O100O10000O100O2O000O100O10000O1000001O000O10000000000000001O000000000000000000000000000000000000000000000000O10000000\\J"}, "label": "plant on right in the right hand picture"}]}
{"id": 287718, "image": "COCO_train2014_000000287718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tree behind a man on a bench\"."}], "task": "refering", "answer_template": "The \"a tree behind a man on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 29, 30, 31, 32, 33, 34, 35, 52, 53, 54, 55, 56, 58, 75, 76, 77, 78, 79, 98, 99, 100], "bbox": [0.24317187499999998, 0.004588235294117647, 0.579890625, 0.32148235294117644], "iscrowd": 0, "area": 16685.9074, "rle": {"size": [425, 640], "counts": "ngP24U=3M3M3M2N3M3M2N3M2N2N2N2gEROS8P1hGTOW8n0dGVO[8l0`GXO_8j0[G[Od8l0RGXOm8P1hFSOX9T1^FPOa9X1TFlNk9Z2O1O1O1O2N1O2N1O1O2N1O2N1O1O1O1O1O1O1O1O000000001O00000000001O00O1000000000000000000000000O100O1O1O100O1O1O1O1N3M2N2O1N2N2O1O1N2O1N2O1O1N2O1O1000000O1000000000000000000000000O100000000000000000000000000000000O1O100O100[NeF[O\\9d0oFQOR9n0e1O1O1O1O1O1O100O1O1O1O1O100O1O1O10000001O000000001O0000001O0000001O000000001O0000001O002N2N2N1O2N2N2N2N1O6J:F001O1O001O0000N2M3N2N2M3M3K5K5K5L4M3M3M3M3M3M3N2M3M3M3Med_3"}, "label": "a tree behind a man on a bench"}]}
{"id": 345062, "image": "COCO_train2014_000000345062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"image didnt load\"."}], "task": "refering", "answer_template": "The \"image didnt load\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 324, 325, 326], "bbox": [0.5325816993464052, 0.35504901960784313, 0.9932516339869281, 0.6494444444444444], "iscrowd": 0, "area": 39377.78444999999, "rle": {"size": [612, 612], "counts": "QTS6=Tb0g0J6J5L5J6J6K5J5L2M2N3N2M3N2M3M3N2M2O2N2O1N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N101N101N100O2O0O2O001N101O001N101O001N101O001N101O0O2O001O0O2O00001N101O001N101O0O2O00000O2O00000O100O2N1O1O1O100O2N11O000010O000001O00010O00001O00O2OgJWBW5i=200000001O0000000O101O0000000000001O0000000O100000000000001O00000000000001O0000000010O000001O000000001O00000010O00000001O0000001O000000001N1000001N1000000O2O001N101O001N10001O0O2O001N101O000O2O001O0O2O001N10001O0O2O001N101O0O101N1O2O0O2O0O1O2O0O2O1N1O2O1N2O1N2N101N2O1N2N2O1N101N2N2O1N2N1O2M3N2N2M3N2M2O2N2M4M2N3L3N2M4M2N3L3N2M4M2N2M4M[S2"}, "label": "image didnt load"}]}
{"id": 345062, "image": "COCO_train2014_000000345062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"image\"."}], "task": "refering", "answer_template": "The \"image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 324, 325, 326], "bbox": [0.5325816993464052, 0.35504901960784313, 0.9932516339869281, 0.6494444444444444], "iscrowd": 0, "area": 39377.78444999999, "rle": {"size": [612, 612], "counts": "QTS6=Tb0g0J6J5L5J6J6K5J5L2M2N3N2M3N2M3M3N2M2O2N2O1N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N101N101N100O2O0O2O001N101O001N101O001N101O001N101O0O2O001O0O2O00001N101O001N101O0O2O00000O2O00000O100O2N1O1O1O100O2N11O000010O000001O00010O00001O00O2OgJWBW5i=200000001O0000000O101O0000000000001O0000000O100000000000001O00000000000001O0000000010O000001O000000001O00000010O00000001O0000001O000000001N1000001N1000000O2O001N101O001N10001O0O2O001N101O000O2O001O0O2O001N10001O0O2O001N101O0O101N1O2O0O2O0O1O2O0O2O1N1O2O1N2O1N2N101N2O1N2N2O1N101N2N2O1N2N1O2M3N2N2M3N2M2O2N2M4M2N3L3N2M4M2N3L3N2M4M2N2M4M[S2"}, "label": "image"}]}
{"id": 345062, "image": "COCO_train2014_000000345062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leftmost pizza\"."}], "task": "refering", "answer_template": "The \"the leftmost pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 184, 200, 201, 202, 203, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.017973856209150325, 0.39550653594771246, 0.5258496732026143, 0.7280882352941177], "iscrowd": 0, "area": 48945.021199999996, "rle": {"size": [612, 612], "counts": "fl68kb0`0_O`0@7I3M4L3M3N3L3M3M4L3M3M3N3L3M3N3M2N2N2O2M2N2N3M2N101N1O2N1O2O0O2N2N1O2O0O2N1O2N1O2O0O2N1O2N101N2N1O2N101N101N101O0O2O001N101O1N101O0O2O001N101O0O2O001N101O1N101O0O2O001N101O0O10000O10001N10000O10000O2O000O10000O101O0O10000O10001N10000O10000O101O00000000000O2O0000000000001O0000000000001O00000O100000001O0000000000001O0000000000001N010000O100O10000O10000O10001N100O10000O10000O2O000O100O10000O10001N10000O100O10000O2O000O10000O100O101O0O10000O10000O101N10000O10000O100O2O000O10000O2O001N100O2O001N101O0O2O001N100O2O001N101O0O2O001N100O2O001N101N1N3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N<Ch0YOf0ZO[o\\5"}, "label": "the leftmost pizza"}]}
{"id": 345062, "image": "COCO_train2014_000000345062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in the left\"."}], "task": "refering", "answer_template": "The \"the pizza in the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 184, 200, 201, 202, 203, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.017973856209150325, 0.39550653594771246, 0.5258496732026143, 0.7280882352941177], "iscrowd": 0, "area": 48945.021199999996, "rle": {"size": [612, 612], "counts": "fl68kb0`0_O`0@7I3M4L3M3N3L3M3M4L3M3M3N3L3M3N3M2N2N2O2M2N2N3M2N101N1O2N1O2O0O2N2N1O2O0O2N1O2N1O2O0O2N1O2N101N2N1O2N101N101N101O0O2O001N101O1N101O0O2O001N101O0O2O001N101O1N101O0O2O001N101O0O10000O10001N10000O10000O2O000O10000O101O0O10000O10001N10000O10000O101O00000000000O2O0000000000001O0000000000001O00000O100000001O0000000000001O0000000000001N010000O100O10000O10000O10001N100O10000O10000O2O000O100O10000O10001N10000O100O10000O2O000O10000O100O101O0O10000O10000O101N10000O10000O100O2O000O10000O2O001N100O2O001N101O0O2O001N100O2O001N101O0O2O001N100O2O001N101N1N3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N<Ch0YOf0ZO[o\\5"}, "label": "the pizza in the left"}]}
{"id": 279530, "image": "COCO_train2014_000000279530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"darker dog on the left\"."}], "task": "refering", "answer_template": "The \"darker dog on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 216, 217, 218, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 273, 274, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 297, 298, 299, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 351, 352, 353, 354, 370, 371, 373, 392, 411], "bbox": [0.37944123314065514, 0.443859375, 0.7545086705202313, 0.9344374999999999], "iscrowd": 0, "area": 29297.59149999999, "rle": {"size": [640, 519], "counts": "dak32jc07J4L3N3fLCnB?k<IRC9h<MUC6f<ORC8l<KoB:n<JkB<R=GiB>T=EeBb0X=A`Bf0^=]OZBk0d=WOTBQ1j=QOSBR1k=QOQBQ1n=ROoAP1o=TOlAo0R>TOkAo0S>TOiAP1T>ROiAR1T>POiAS1T>QOhAS1U>oNhAT1V>^2L3N3M2N3N2M2N3M2N2OO0O1oNcB_K]=`4fB^KZ=a4iB\\KY=b4jB\\KW=b4lB[KV=c4mB[KT=c4oB[KR=c4QCZKQ=d4RCZKo<d4YCUKh<i4X1O2004L5K5L8G=C=C;F53M6JK4MN3N1L4M2N2N3N1N00000001O000000001O000001O0mMbB`LH>e=Q3iB^L10V=b3kB]LW>c3kA\\LT>d3mA]LQ>c3QB\\Ln=d3SB\\Ll=d3VB[Li=e3YB[Ld=f3]BZLb=e3aBZL]=f3eBZLZ=f3hBXLW=h3kBWLT=i3oBULQ=j3RCTLm<m3TCRLk<n3WCQLh<o3d10100O2N2M2N3N12O0O2O1N101O0O2O0O2N1O2N1O1O1m@mKV>V4fAnKX>S4eAQLY>P4dATLZ>m3cAWL[>j3aA\\L\\>e3aA`Lb1QOU;d5bDeJ[;^5\\DjJa;k600000O1002N3M10O01O1O010cC^HZ<d7001O1O1O1N1TMjCnMW<Q2TDnLAoN\\<R4]DcL\\<\\3oCVLU<i3UDjKo;U4o1O1K4N4L3L4M3M2N3M3L4M3M3M3O1O1O001O1O1O1O1N2^MV_Ol1aa0N2M3M2O2M3M2N3N2M4I6K6I<E:EYb_2"}, "label": "darker dog on the left"}]}
{"id": 279530, "image": "COCO_train2014_000000279530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog next to another that has more black on it ' s face\"."}], "task": "refering", "answer_template": "The \"a dog next to another that has more black on it ' s face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 216, 217, 218, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 273, 274, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 297, 298, 299, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 351, 352, 353, 354, 370, 371, 373, 392, 411], "bbox": [0.37944123314065514, 0.443859375, 0.7545086705202313, 0.9344374999999999], "iscrowd": 0, "area": 29297.59149999999, "rle": {"size": [640, 519], "counts": "dak32jc07J4L3N3fLCnB?k<IRC9h<MUC6f<ORC8l<KoB:n<JkB<R=GiB>T=EeBb0X=A`Bf0^=]OZBk0d=WOTBQ1j=QOSBR1k=QOQBQ1n=ROoAP1o=TOlAo0R>TOkAo0S>TOiAP1T>ROiAR1T>POiAS1T>QOhAS1U>oNhAT1V>^2L3N3M2N3N2M2N3M2N2OO0O1oNcB_K]=`4fB^KZ=a4iB\\KY=b4jB\\KW=b4lB[KV=c4mB[KT=c4oB[KR=c4QCZKQ=d4RCZKo<d4YCUKh<i4X1O2004L5K5L8G=C=C;F53M6JK4MN3N1L4M2N2N3N1N00000001O000000001O000001O0mMbB`LH>e=Q3iB^L10V=b3kB]LW>c3kA\\LT>d3mA]LQ>c3QB\\Ln=d3SB\\Ll=d3VB[Li=e3YB[Ld=f3]BZLb=e3aBZL]=f3eBZLZ=f3hBXLW=h3kBWLT=i3oBULQ=j3RCTLm<m3TCRLk<n3WCQLh<o3d10100O2N2M2N3N12O0O2O1N101O0O2O0O2N1O2N1O1O1m@mKV>V4fAnKX>S4eAQLY>P4dATLZ>m3cAWL[>j3aA\\L\\>e3aA`Lb1QOU;d5bDeJ[;^5\\DjJa;k600000O1002N3M10O01O1O010cC^HZ<d7001O1O1O1N1TMjCnMW<Q2TDnLAoN\\<R4]DcL\\<\\3oCVLU<i3UDjKo;U4o1O1K4N4L3L4M3M2N3M3L4M3M3M3O1O1O001O1O1O1O1N2^MV_Ol1aa0N2M3M2O2M3M2N3N2M4I6K6I<E:EYb_2"}, "label": "a dog next to another that has more black on it ' s face"}]}
{"id": 279530, "image": "COCO_train2014_000000279530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"husky with lighter coloring\"."}], "task": "refering", "answer_template": "The \"husky with lighter coloring\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 222, 223, 224, 225, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 318, 319, 320, 321, 322, 336, 337, 338, 339, 340, 356, 357, 358, 359, 375, 376, 377, 395, 396, 415], "bbox": [0.6352793834296724, 0.44470312500000003, 0.9702119460500963, 0.91875], "iscrowd": 0, "area": 25121.431049999996, "rle": {"size": [640, 519], "counts": "lb^61oc01N1O101N1O1O2O0O1O2O0O1O2N100O2N1O101N1O2N2N2N2N1O2N2N2N2N2N2N2N2O0O2N2N2N2M3N2UAUN^;m1`DUN^;m1`DUN];n1bDTN[;o1bDSN\\;P2aDRN\\;R2bDPN[;S2bDPN[;S2cDoMY;U2WB`Mn1?g;c2VD`Mh;c2UD_Mh;d2UD_Mh;d2UD`Mg;d2VD^Mg;e2VD^Mg;e2VD_Mf;e2WD]Mf;f2WD]Mf;f2WD^Me;m2QDUMl;f3YC]Ld<b5iCgHc;\\7SDkHl;f7N10O01O0010O100O100O100O101N100O100O100O2O0O100O100O10SDUIe:j6PEcIP;g7O1O02O00000O10000O1000001N1000000]HiD[6X;bIkD^6U;`ImD`6S;^InDb6S;[IPEe6f;cH\\D\\7T<01N2O0]NhC^KY<X4TDdKm;[4WDaKi;^4\\D]Kf;b4]D[Kd;c4aDYK`;f4cDWK^;g4gDTK[;k4hDRKZ;k4kDQKV;m4nDPKS;o4P2N2N2N2O1N3M2VOk@fLU?Z3QAaLn>`3WAZLj>e3h0001OO1N2O1N2N2O3L5K6K4K5K5M3M3N2M4M2N2M3N3L5L4K5L4L4K5L3L5L4K=D=Com8"}, "label": "husky with lighter coloring"}]}
{"id": 287723, "image": "COCO_train2014_000000287723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unoccupied chair\"."}], "task": "refering", "answer_template": "The \"an unoccupied chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 143, 156, 157, 158, 169, 170, 171, 172, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225], "bbox": [0.002986666666666667, 0.54922, 0.3497866666666667, 0.9842000000000001], "iscrowd": 0, "area": 14159.4178, "rle": {"size": [500, 375], "counts": "dh0d4n:2M3M3N2M33M4L4L5K4L5J5L5K4L5K4L4L5K4LQETM[9g2kESNY:b35K4L4L5K2N00O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O100000O010000O1000000O100003M2N3M3M3M2N3M3M2N3M3M2N3M3M3M2N3M7I`0@`0@`0@`0@`0@b[54PcJZ1jNW1O00000O1000000O11O5K6K4L4L5K4_Ob0POlgf3"}, "label": "an unoccupied chair"}]}
{"id": 287723, "image": "COCO_train2014_000000287723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chair with no one sitting in it\"."}], "task": "refering", "answer_template": "The \"wooden chair with no one sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 143, 156, 157, 158, 169, 170, 171, 172, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225], "bbox": [0.002986666666666667, 0.54922, 0.3497866666666667, 0.9842000000000001], "iscrowd": 0, "area": 14159.4178, "rle": {"size": [500, 375], "counts": "dh0d4n:2M3M3N2M33M4L4L5K4L5J5L5K4L5K4L4L5K4LQETM[9g2kESNY:b35K4L4L5K2N00O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O100000O010000O1000000O100003M2N3M3M3M2N3M3M2N3M3M2N3M3M3M2N3M7I`0@`0@`0@`0@`0@b[54PcJZ1jNW1O00000O1000000O11O5K6K4L4L5K4_Ob0POlgf3"}, "label": "wooden chair with no one sitting in it"}]}
{"id": 287723, "image": "COCO_train2014_000000287723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the women on the chair is drinking\"."}], "task": "refering", "answer_template": "The \"the women on the chair is drinking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 113, 114, 115, 116, 126, 127, 128, 129, 137, 139, 140, 141, 142, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 179, 180, 181, 192, 193, 194, 202, 203, 204, 205, 206, 207, 214, 215, 216, 217, 218, 219, 220, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.43693333333333334, 0.41752, 0.9945333333333333, 0.9996], "iscrowd": 0, "area": 28464.336549999996, "rle": {"size": [500, 375], "counts": "Qb`22_?3M3M3M3M3M3M3M3N2M3M3M3M3M300O100O100O10000O100O100O100O100O100O10000O10000O1000000O1000000O1000000O10000O10000O1eKYNiJg1n4cNQK]1n4eNQK[1n4gNPKZ1o4hNPKX1o4jNPKV1o4jNRKV1m4hNVKX1i4gNYKY1a900000000000O100]LTN]Im1c6ZNUIg1k6_NoHa1Q7cNjH^1V7bNjH^1V7bNiH_1W7aNiH_1W7bNgH_1Y7aNfH`1Z7`NfH`1[7_NdHb1Z7`NfH`1Y7bNeH_1Y7cNgH]1X7dNgH]1W7eNiH[1V7gNhHZ1V7hNjHX1U7iNkHW1T7jNkHW1V7hNjHX1R7iM_Fn0_2Y1o6lMdFi0\\2]1l6mMhFf0\\2]1S5iMQJ4_Ng0]2\\1P5nMRIK92ZOi0[2\\1n4RNPIK;M\\Ok0Z2\\1l4VNoHJ>F_On0X2\\1j4[NkHJk0`0`1[1h4_NiH2e03j1]1e4@_ITOl1\\1c4C`IQOm1\\1a4F`IoNn1\\1a4G`ImNo1\\1`4I`IkNP2\\1_4K_IjNQ2\\1^4N_IfNS2\\1]40^IeNU2[1\\43]IbNV2\\1\\44\\I`NY2\\1Z47[I]NZ2]1Z48ZI\\N[2]1Z4^1eKcNZ4^1dKdN[4^1cKcN]4^1aKcN_4_1^KbNb4_1\\KbNd4_1ZKbNf4`1WKaNh4a1VK`Nj4a1TK`Nl4g1kJ[NU5o1_JSNa5T2VJnMj5T2RJnMm5V2nIlMR6W2iIkMW6W2dIlM\\6W2_IkMa6d50000O10N2L4M3M3M2N3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3L4M3MXIeI]4X6bKlI^4Q6aKSJ_4j5`KYJa4f5\\K^Jd4b5XKaJi4^5UKeJk4[5QKhJP5X5lJlJV5Q5gJRK[5m4aJWK`5h4\\J[Kf5d4VJ_Km5^4PJcKT6\\4hIeK\\6Z4`IhKd6U4YIlKk6S4QInKS7f0XHR1e0UNW7d0XHW1a0RNKgNm6j1kH]1=oMKmNn6a1oHc17mMMPOS9R3QGkMMUOR9P3QGhMMZOS9n2PGeMN_OR9l2PGcMNCR9j2PG`MNHS9h2nF^MOLT9@\\Fo2a0bM0O]9]2dFaMO3`9[2aF_MO7c9Y2^F]M0:e9X2[F[M0=i9W2VFZM1`0l9U2SFYM0c0P:R2fFPN]9o1aFPNd9n1]FoMf9P2ZFnMi9Q2XFlMk9S2UFiMP:T2SFgM];`0PE\\OTo0"}, "label": "the women on the chair is drinking"}]}
{"id": 287723, "image": "COCO_train2014_000000287723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady with the cup\"."}], "task": "refering", "answer_template": "The \"lady with the cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 113, 114, 115, 116, 126, 127, 128, 129, 137, 139, 140, 141, 142, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 179, 180, 181, 192, 193, 194, 202, 203, 204, 205, 206, 207, 214, 215, 216, 217, 218, 219, 220, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.43693333333333334, 0.41752, 0.9945333333333333, 0.9996], "iscrowd": 0, "area": 28464.336549999996, "rle": {"size": [500, 375], "counts": "Qb`22_?3M3M3M3M3M3M3M3N2M3M3M3M3M300O100O100O10000O100O100O100O100O100O10000O10000O1000000O1000000O1000000O10000O10000O1eKYNiJg1n4cNQK]1n4eNQK[1n4gNPKZ1o4hNPKX1o4jNPKV1o4jNRKV1m4hNVKX1i4gNYKY1a900000000000O100]LTN]Im1c6ZNUIg1k6_NoHa1Q7cNjH^1V7bNjH^1V7bNiH_1W7aNiH_1W7bNgH_1Y7aNfH`1Z7`NfH`1[7_NdHb1Z7`NfH`1Y7bNeH_1Y7cNgH]1X7dNgH]1W7eNiH[1V7gNhHZ1V7hNjHX1U7iNkHW1T7jNkHW1V7hNjHX1R7iM_Fn0_2Y1o6lMdFi0\\2]1l6mMhFf0\\2]1S5iMQJ4_Ng0]2\\1P5nMRIK92ZOi0[2\\1n4RNPIK;M\\Ok0Z2\\1l4VNoHJ>F_On0X2\\1j4[NkHJk0`0`1[1h4_NiH2e03j1]1e4@_ITOl1\\1c4C`IQOm1\\1a4F`IoNn1\\1a4G`ImNo1\\1`4I`IkNP2\\1_4K_IjNQ2\\1^4N_IfNS2\\1]40^IeNU2[1\\43]IbNV2\\1\\44\\I`NY2\\1Z47[I]NZ2]1Z48ZI\\N[2]1Z4^1eKcNZ4^1dKdN[4^1cKcN]4^1aKcN_4_1^KbNb4_1\\KbNd4_1ZKbNf4`1WKaNh4a1VK`Nj4a1TK`Nl4g1kJ[NU5o1_JSNa5T2VJnMj5T2RJnMm5V2nIlMR6W2iIkMW6W2dIlM\\6W2_IkMa6d50000O10N2L4M3M3M2N3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3L4M3MXIeI]4X6bKlI^4Q6aKSJ_4j5`KYJa4f5\\K^Jd4b5XKaJi4^5UKeJk4[5QKhJP5X5lJlJV5Q5gJRK[5m4aJWK`5h4\\J[Kf5d4VJ_Km5^4PJcKT6\\4hIeK\\6Z4`IhKd6U4YIlKk6S4QInKS7f0XHR1e0UNW7d0XHW1a0RNKgNm6j1kH]1=oMKmNn6a1oHc17mMMPOS9R3QGkMMUOR9P3QGhMMZOS9n2PGeMN_OR9l2PGcMNCR9j2PG`MNHS9h2nF^MOLT9@\\Fo2a0bM0O]9]2dFaMO3`9[2aF_MO7c9Y2^F]M0:e9X2[F[M0=i9W2VFZM1`0l9U2SFYM0c0P:R2fFPN]9o1aFPNd9n1]FoMf9P2ZFnMi9Q2XFlMk9S2UFiMP:T2SFgM];`0PE\\OTo0"}, "label": "lady with the cup"}]}
{"id": 287723, "image": "COCO_train2014_000000287723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man working on his laptop while sitting at the table\"."}], "task": "refering", "answer_template": "The \"a man working on his laptop while sitting at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 84, 85, 96, 97, 98, 108, 109, 110, 111, 121, 122], "bbox": [0.3477066666666666, 0.32195999999999997, 0.6199733333333333, 0.55728], "iscrowd": 0, "area": 5959.028300000002, "rle": {"size": [500, 375], "counts": "Qgo11a?4K5L8G8H7J3L3N3L;E2O2M3N2M3M2O2O100000O01000000O10O100000O10O1J6G8N3O1O1HSBiNl=W1900O10O1ZOSOeBm0Y=XOcBi0\\=g002M2J6O101N100O3N1O0O100O1000000O10000000O10001O0000001O00001^NPC>P=ASC=n<AUC=l<BVC<k<BXC<m<_OTC`0P>000001O001O1O001O001O2N3M2N3M3M3\\OWA7U?N2N2NgbU2"}, "label": "a man working on his laptop while sitting at the table"}]}
{"id": 287723, "image": "COCO_train2014_000000287723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 84, 85, 96, 97, 98, 108, 109, 110, 111, 121, 122], "bbox": [0.3477066666666666, 0.32195999999999997, 0.6199733333333333, 0.55728], "iscrowd": 0, "area": 5959.028300000002, "rle": {"size": [500, 375], "counts": "Qgo11a?4K5L8G8H7J3L3N3L;E2O2M3N2M3M2O2O100000O01000000O10O100000O10O1J6G8N3O1O1HSBiNl=W1900O10O1ZOSOeBm0Y=XOcBi0\\=g002M2J6O101N100O3N1O0O100O1000000O10000000O10001O0000001O00001^NPC>P=ASC=n<AUC=l<BVC<k<BXC<m<_OTC`0P>000001O001O1O001O001O2N3M2N3M3M3\\OWA7U?N2N2NgbU2"}, "label": "a man wearing glasses"}]}
{"id": 426988, "image": "COCO_train2014_000000426988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch facing the windows that is brown\"."}], "task": "refering", "answer_template": "The \"the couch facing the windows that is brown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 326, 327, 328], "bbox": [0.047703125, 0.5718544600938967, 0.320671875, 0.9450000000000001], "iscrowd": 0, "area": 19404.701499999996, "rle": {"size": [426, 640], "counts": "QW=8k<:F:F:G9K3M2N2M3O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10000000002N3M2O2N1O2N1N3N3M3M3M3L4M0000000O1O1O1N2O000O100O10000O10000010O000000001O000000000000000000O2O0000000O100000000001O00000001O0000000001O001O001O001O001O0010OO7H<D2O1N2O2O01O0001O000O100O2O0O2O1O0O2O1N4M6I8I7H5L2jN]EAf::V1N3ZOf0HYed5"}, "label": "the couch facing the windows that is brown"}]}
{"id": 426988, "image": "COCO_train2014_000000426988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch on the left side of the fireplace\"."}], "task": "refering", "answer_template": "The \"the couch on the left side of the fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 326, 327, 328], "bbox": [0.047703125, 0.5718544600938967, 0.320671875, 0.9450000000000001], "iscrowd": 0, "area": 19404.701499999996, "rle": {"size": [426, 640], "counts": "QW=8k<:F:F:G9K3M2N2M3O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10000000002N3M2O2N1O2N1N3N3M3M3M3L4M0000000O1O1O1N2O000O100O10000O10000010O000000001O000000000000000000O2O0000000O100000000001O00000001O0000000001O001O001O001O001O0010OO7H<D2O1N2O2O01O0001O000O100O2O0O2O1O0O2O1N4M6I8I7H5L2jN]EAf::V1N3ZOf0HYed5"}, "label": "the couch on the left side of the fireplace"}]}
{"id": 426988, "image": "COCO_train2014_000000426988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a three seat couch\"."}], "task": "refering", "answer_template": "The \"a three seat couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 294, 310, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338], "bbox": [0.45234375, 0.6473004694835681, 0.957640625, 1.0], "iscrowd": 0, "area": 22382.33409999999, "rle": {"size": [426, 640], "counts": "gnh39P=8H7I3N1O2M3N1O2M3N1O2N2N2M2O2N2N2N2N1O2M3N2N2N1O1O001N101O001O001O001O1O0O2O001O001O0O2O001O0O2O0N3N2M2M4DTMVFn2j9:O001O001O1O001O1O010O1O1O00100O001O10O01O1O00100O00100O0010O01O10O01O010O01O00001O010O00001O00010O001O000010O00000000000001N1000000000000000001N1000000O100O10000O10O0100000000000O10O100aNVF^Oj9a0XF^Oh9`0\\F^Od9a0^F^Oa9b0aF]O_9b0cF]O]9b0eF\\O\\9d0fFZOY9g0hFXOX9h0iFWOW9h0lFVOT9j0mFUOS9k0oFSOP9n0RGoNo8P1SGoNm8Q1UGmNk8S1WGkNh8T1[GkNe8T1^GjNb8U1bGhN]8W1hGfNX8Y1kGdNU8[1PHbNP8]1TH`Nl7^1Q2N1N3N2N110O10O1000O01000000O01000000O10O100000O100000000O0101N101N1O2N1O2M2O2O0O2N1O2N1O2N1O2N101O000O10000O100000O001O1J6L4K6O0O2O0O1O2O0O2O0O1O2O0O2O0O101O00000O101O000000010O00000000000O1000000O101O00000O10O100O1O010O1O101N2N2O1N1O2O1OO1O1O100O1N2O1O1O1N2O1N2O1O1N2O2M3N2O1N2O1N2O1N2O1N2O2M2O1Ne[;"}, "label": "a three seat couch"}]}
{"id": 426988, "image": "COCO_train2014_000000426988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark brown couch in front of a table covered in books\"."}], "task": "refering", "answer_template": "The \"dark brown couch in front of a table covered in books\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 294, 310, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338], "bbox": [0.45234375, 0.6473004694835681, 0.957640625, 1.0], "iscrowd": 0, "area": 22382.33409999999, "rle": {"size": [426, 640], "counts": "gnh39P=8H7I3N1O2M3N1O2M3N1O2N2N2M2O2N2N2N2N1O2M3N2N2N1O1O001N101O001O001O001O1O0O2O001O001O0O2O001O0O2O0N3N2M2M4DTMVFn2j9:O001O001O1O001O1O010O1O1O00100O001O10O01O1O00100O00100O0010O01O10O01O010O01O00001O010O00001O00010O001O000010O00000000000001N1000000000000000001N1000000O100O10000O10O0100000000000O10O100aNVF^Oj9a0XF^Oh9`0\\F^Od9a0^F^Oa9b0aF]O_9b0cF]O]9b0eF\\O\\9d0fFZOY9g0hFXOX9h0iFWOW9h0lFVOT9j0mFUOS9k0oFSOP9n0RGoNo8P1SGoNm8Q1UGmNk8S1WGkNh8T1[GkNe8T1^GjNb8U1bGhN]8W1hGfNX8Y1kGdNU8[1PHbNP8]1TH`Nl7^1Q2N1N3N2N110O10O1000O01000000O01000000O10O100000O100000000O0101N101N1O2N1O2M2O2O0O2N1O2N1O2N1O2N101O000O10000O100000O001O1J6L4K6O0O2O0O1O2O0O2O0O1O2O0O2O0O101O00000O101O000000010O00000000000O1000000O101O00000O10O100O1O010O1O101N2N2O1N1O2O1OO1O1O100O1N2O1O1O1N2O1N2O1O1N2O2M3N2O1N2O1N2O1N2O1N2O2M2O1Ne[;"}, "label": "dark brown couch in front of a table covered in books"}]}
{"id": 517101, "image": "COCO_train2014_000000517101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big knife on the counter\"."}], "task": "refering", "answer_template": "The \"a big knife on the counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 322, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 363], "bbox": [0.34269311064718166, 0.765796875, 1.0, 0.923828125], "iscrowd": 0, "area": 16068.594200000009, "rle": {"size": [640, 479], "counts": "^RW31kc04L4L3M4M3O001O1O1O001O1N101O1O1O001O1O001O1O1O001O1N101O1O1O010O10O10O100O010O10O10O100O01000O10O01000O0100O10O01000O0100O10O10O10O0100O10O10O10O010000O010O10O10O100O010O10O10O100O01000O10O0100O01000O10O01000O0100O10O10O10O0100O10O10O10O010000O010O10O0100000O10000O1000000O0100000O11O00000000000O1000000000001O0000000O10000000000000001O00000O10000000000000001O000O100000000000000000001N10000000000000000000001N100000000000000000000O101O00000000000000000O10O10000000000000000O10O10000000000000O100000O1000000000O1000000000O10000000O10000000000000O10O10000000001O00000O10000000000000001O0O10000000000000000O2O000000000000000O1000c@"}, "label": "a big knife on the counter"}]}
{"id": 517101, "image": "COCO_train2014_000000517101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"metal pizza slicer\"."}], "task": "refering", "answer_template": "The \"metal pizza slicer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 322, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 363], "bbox": [0.34269311064718166, 0.765796875, 1.0, 0.923828125], "iscrowd": 0, "area": 16068.594200000009, "rle": {"size": [640, 479], "counts": "^RW31kc04L4L3M4M3O001O1O1O001O1N101O1O1O001O1O001O1O1O001O1N101O1O1O010O10O10O100O010O10O10O100O01000O10O01000O0100O10O01000O0100O10O10O10O0100O10O10O10O010000O010O10O10O100O010O10O10O100O01000O10O0100O01000O10O01000O0100O10O10O10O0100O10O10O10O010000O010O10O0100000O10000O1000000O0100000O11O00000000000O1000000000001O0000000O10000000000000001O00000O10000000000000001O000O100000000000000000001N10000000000000000000001N100000000000000000000O101O00000000000000000O10O10000000000000000O10O10000000000000O100000O1000000000O1000000000O10000000O10000000000000O10O10000000001O00000O10000000000000001O0O10000000000000000O2O000000000000000O1000c@"}, "label": "metal pizza slicer"}]}
{"id": 91123, "image": "COCO_train2014_000000091123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in black\"."}], "task": "refering", "answer_template": "The \"a skier in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 102, 103, 104, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 194, 195, 196, 217, 219], "bbox": [0.43587499999999996, 0.22435416666666666, 0.608796875, 0.5872499999999999], "iscrowd": 0, "area": 7805.185100000002, "rle": {"size": [480, 640], "counts": "`QS43i>4L4`KFlF4T2:o6EhF8P28W7CdF;n16]7BaF>j14d7A]Fa0g13k7_OZFd0X1=\\8QOZFf0i0g0m8dNWFh0i0f0o8dNVFg0j0f0o83oFOP92nF0R9OnF2P90nF2R9NlF5S9JlF7T9JkF5W9JgF8Z9EgF;\\9AeF?_9^O_F?i9^OhEl0`:b19E;C=J6M3M100O2N1O1O2N100O2N1O1>C4K3M3N2M3M2O0O2N2O0O2N101N1OPOREPNm:l1XETNf:j1]EVNb:h1bEXN]:d1hE[NW:b1nE^NQ:^1SFcNk9Y1[FeNf9W1_FhNa9T1dFkN[9R1jFmNV9n0PGPOP9LWH2i7H^H7b7H`H7_7JcH3^7McH2]7NdH0\\70gHNY72hHMX73j310O10O100O10O1000001N10001O0O10001O000O10O1O10O001N1N201O001O1O2N10a_e3"}, "label": "a skier in black"}]}
{"id": 91123, "image": "COCO_train2014_000000091123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboarder that is upright , wearing black , and has snow on his rear\"."}], "task": "refering", "answer_template": "The \"a snowboarder that is upright , wearing black , and has snow on his rear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 102, 103, 104, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 194, 195, 196, 217, 219], "bbox": [0.43587499999999996, 0.22435416666666666, 0.608796875, 0.5872499999999999], "iscrowd": 0, "area": 7805.185100000002, "rle": {"size": [480, 640], "counts": "`QS43i>4L4`KFlF4T2:o6EhF8P28W7CdF;n16]7BaF>j14d7A]Fa0g13k7_OZFd0X1=\\8QOZFf0i0g0m8dNWFh0i0f0o8dNVFg0j0f0o83oFOP92nF0R9OnF2P90nF2R9NlF5S9JlF7T9JkF5W9JgF8Z9EgF;\\9AeF?_9^O_F?i9^OhEl0`:b19E;C=J6M3M100O2N1O1O2N100O2N1O1>C4K3M3N2M3M2O0O2N2O0O2N101N1OPOREPNm:l1XETNf:j1]EVNb:h1bEXN]:d1hE[NW:b1nE^NQ:^1SFcNk9Y1[FeNf9W1_FhNa9T1dFkN[9R1jFmNV9n0PGPOP9LWH2i7H^H7b7H`H7_7JcH3^7McH2]7NdH0\\70gHNY72hHMX73j310O10O100O10O1000001N10001O0O10001O000O10O1O10O001N1N201O001O1O2N10a_e3"}, "label": "a snowboarder that is upright , wearing black , and has snow on his rear"}]}
{"id": 246777, "image": "COCO_train2014_000000246777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby with eyes open , looking into the camera\"."}], "task": "refering", "answer_template": "The \"a baby with eyes open , looking into the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 21, 22, 23, 24, 25, 38, 39, 40, 41, 42, 43, 44, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221], "bbox": [0.0033799999999999998, 0.0, 0.45608, 0.9977600000000001], "iscrowd": 0, "area": 59056.29645, "rle": {"size": [375, 500], "counts": "ZQ1i1o7o1cN]1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O1QOXJoKh5a3mJYLT5^3ZK]Lf4W3iKdLY4n2YLkLj3Q3`2M3N2M3N2N2N2O1O1O1O1O1O10000O10000O0100O0100O001O10O01O001O1O0O2O000J6XOh0G8G:I71O1O1N2O1O1O1000000O12N7I6J7I1O00000000O1000000000000O100000004L4L4L4K2O2N2N1O2N000O010000O100O1000O010000O100O10O0100O1O100O10O1000001O000bNMXG3h8^1001O0000001O00fLPNmLP2Q3TNmLl1R3WNlLi1R3[NmLd1Q3aNlL_1R3eNlL[1S3hNkLX1T3kNkLT1U3nNiLR1V3ROgLn0X3UOfLk0Z3WOcLj0\\3YOaLh0^3[O`Le0`3]O]Ld0b3_O[Lb0d3AZL?f3CWL?g3CWL>h3DVL>g3EXL;f3HWL:g3IWL9f3JWL8g3KVL7h3KWL7f3LWL6g3MVL5h3NVL4g30UL2i31TL2i32TL0i33TL0i33TL0i3Q4M3M3L4K6J5K5L4K5K5K6J6K5K5K5K5K5K5K7J6I8H9Gc0]Om`S3"}, "label": "a baby with eyes open , looking into the camera"}]}
{"id": 246777, "image": "COCO_train2014_000000246777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby girl in green shirt next to another baby\"."}], "task": "refering", "answer_template": "The \"baby girl in green shirt next to another baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 21, 22, 23, 24, 25, 38, 39, 40, 41, 42, 43, 44, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221], "bbox": [0.0033799999999999998, 0.0, 0.45608, 0.9977600000000001], "iscrowd": 0, "area": 59056.29645, "rle": {"size": [375, 500], "counts": "ZQ1i1o7o1cN]1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O1QOXJoKh5a3mJYLT5^3ZK]Lf4W3iKdLY4n2YLkLj3Q3`2M3N2M3N2N2N2O1O1O1O1O1O10000O10000O0100O0100O001O10O01O001O1O0O2O000J6XOh0G8G:I71O1O1N2O1O1O1000000O12N7I6J7I1O00000000O1000000000000O100000004L4L4L4K2O2N2N1O2N000O010000O100O1000O010000O100O10O0100O1O100O10O1000001O000bNMXG3h8^1001O0000001O00fLPNmLP2Q3TNmLl1R3WNlLi1R3[NmLd1Q3aNlL_1R3eNlL[1S3hNkLX1T3kNkLT1U3nNiLR1V3ROgLn0X3UOfLk0Z3WOcLj0\\3YOaLh0^3[O`Le0`3]O]Ld0b3_O[Lb0d3AZL?f3CWL?g3CWL>h3DVL>g3EXL;f3HWL:g3IWL9f3JWL8g3KVL7h3KWL7f3LWL6g3MVL5h3NVL4g30UL2i31TL2i32TL0i33TL0i33TL0i3Q4M3M3L4K6J5K5L4K5K5K6J6K5K5K5K5K5K5K7J6I8H9Gc0]Om`S3"}, "label": "baby girl in green shirt next to another baby"}]}
{"id": 246777, "image": "COCO_train2014_000000246777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small kid sleeping on the bed with some toy in its hand\"."}], "task": "refering", "answer_template": "The \"a small kid sleeping on the bed with some toy in its hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 64, 65, 66, 67, 68, 69, 70, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.42568, 0.10135999999999999, 1.0, 0.9842399999999999], "iscrowd": 0, "area": 71139.84389999999, "rle": {"size": [375, 500], "counts": "jY^25^;8H7I8H8H7I8H8G8I8H8H8H7I8G8I4L4N2O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2M3M3M3L4M3M3I7D<E;D<D<E;L4N2O1O1N2O1O1N2O1O1N2O1O1O1O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O1O2N1O100O1O1O1O100O1O1O1O1O100O1O1O1O1000000O10000O1000000O10000O1000000O10000O1000000000000000000O1000000000000000000000000001O001O00001O00001O00001O00001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O1O1O2N1O1O100O2N1O1O1O1O1O2N1O1O1O1QJeM`2]2\\MfMc2[2[MfMe2\\2YMeMf2]2WMeMh2\\2VMeMj2]2TMdMk2]2SMeMl2]2QMdMo2]2oLeMP3_2lLcMR3b2hLaMV3d2dL_MZ3f2`L]M^3i2\\LYMb3m2WLVMg3Z3hKiLV4X3iKiLV4X3hKkLV4V3iKlLU4U3iKmLV4S3jKoLT4R3jKQMT4o2lKRMS4o2lKSMR4m2nKUMP4l2oKUMP4k2PLWMn3j2RLWMl3i2TLXMk3h2ULZMi3g2VL[Mh3e2YL[Mf3e2[L\\Mc3d2^L]M`3c2bL\\M]3e2cL\\M[3d2fL]MX3c2jL\\MU3d2mL\\MQ3e2VMTMi2l2XMUMf2k2\\MUMb2k2`MTM_2m2aMTMY2P3fMSMm0X4ZNbL;1"}, "label": "a small kid sleeping on the bed with some toy in its hand"}]}
{"id": 246777, "image": "COCO_train2014_000000246777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby on the right with its eyes closed\"."}], "task": "refering", "answer_template": "The \"the baby on the right with its eyes closed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 64, 65, 66, 67, 68, 69, 70, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.42568, 0.10135999999999999, 1.0, 0.9842399999999999], "iscrowd": 0, "area": 71139.84389999999, "rle": {"size": [375, 500], "counts": "jY^25^;8H7I8H8H7I8H8G8I8H8H8H7I8G8I4L4N2O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2M3M3M3L4M3M3I7D<E;D<D<E;L4N2O1O1N2O1O1N2O1O1N2O1O1O1O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O1O2N1O100O1O1O1O100O1O1O1O1O100O1O1O1O1000000O10000O1000000O10000O1000000O10000O1000000000000000000O1000000000000000000000000001O001O00001O00001O00001O00001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O1O1O2N1O1O100O2N1O1O1O1O1O2N1O1O1O1QJeM`2]2\\MfMc2[2[MfMe2\\2YMeMf2]2WMeMh2\\2VMeMj2]2TMdMk2]2SMeMl2]2QMdMo2]2oLeMP3_2lLcMR3b2hLaMV3d2dL_MZ3f2`L]M^3i2\\LYMb3m2WLVMg3Z3hKiLV4X3iKiLV4X3hKkLV4V3iKlLU4U3iKmLV4S3jKoLT4R3jKQMT4o2lKRMS4o2lKSMR4m2nKUMP4l2oKUMP4k2PLWMn3j2RLWMl3i2TLXMk3h2ULZMi3g2VL[Mh3e2YL[Mf3e2[L\\Mc3d2^L]M`3c2bL\\M]3e2cL\\M[3d2fL]MX3c2jL\\MU3d2mL\\MQ3e2VMTMi2l2XMUMf2k2\\MUMb2k2`MTM_2m2aMTMY2P3fMSMm0X4ZNbL;1"}, "label": "the baby on the right with its eyes closed"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"food in a blue container inside a container with other food\"."}], "task": "refering", "answer_template": "The \"food in a blue container inside a container with other food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 330, 331], "bbox": [0.0, 0.27714285714285714, 0.57678125, 0.9152834467120181], "iscrowd": 0, "area": 64571.6017, "rle": {"size": [441, 640], "counts": "X7n1i;3N2M2O2M3N2M2O2M3N1N3N2M3N1N3M3N1N3N2M3N1N3N2M2O2M3N1N3N1N2O2M2O1N3M2O1N3N1N2O2M2O1N3N1N2O2M2O1N3N1N2N3N1N2O2M2O1O2O0O101N1O101N1O2O0O1O2O0O1O2O0O1O2O0O101N1O101N1O101N1O101N1O101N1O1010O00010O00010O0001O01O01O01O01O01O01O01O01O01O01O01O01O01O01O0010O00010O00010O00010O00010O00010O0001O01O01O01O01O01O01O010O00010O0010O01O01O01O0010O01O01O01O010O00010O0010O01O01O01O0010O01O01O01O010O00010O0010O01O01O01O010O001O01O01O010O0010O00010O01O01O01O010O001O01O01O010O0010O00010O01O01O01O010O0010O0001O010O0010O00010O01O01O01O010O0010O000O1O2M2O1N3N1N2N3N1N2O2M2O1N3M2O1N3N1N2O2M2O1N2N3N1N2O2M2O1N3M2O1N3N1N2O2M2N2O2M2O1N2O2M2N2O2M2O1N3N1N2O2M2N2O2M2O1N3N1N2N3N1N2N2N3M2N2N3M2N2N3L3N2N3M2N2N3M2N2M4M2N2N2N3M2N2N3L3N2N3M3M3M3M3M3L4M3M3M3M3M3M3Mo_d3"}, "label": "food in a blue container inside a container with other food"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue plate holding fresh fruit and vegetables\"."}], "task": "refering", "answer_template": "The \"a blue plate holding fresh fruit and vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 330, 331], "bbox": [0.0, 0.27714285714285714, 0.57678125, 0.9152834467120181], "iscrowd": 0, "area": 64571.6017, "rle": {"size": [441, 640], "counts": "X7n1i;3N2M2O2M3N2M2O2M3N1N3N2M3N1N3M3N1N3N2M3N1N3N2M2O2M3N1N3N1N2O2M2O1N3M2O1N3N1N2O2M2O1N3N1N2O2M2O1N3N1N2N3N1N2O2M2O1O2O0O101N1O101N1O2O0O1O2O0O1O2O0O1O2O0O101N1O101N1O101N1O101N1O101N1O1010O00010O00010O0001O01O01O01O01O01O01O01O01O01O01O01O01O01O01O0010O00010O00010O00010O00010O00010O0001O01O01O01O01O01O01O010O00010O0010O01O01O01O0010O01O01O01O010O00010O0010O01O01O01O0010O01O01O01O010O00010O0010O01O01O01O010O001O01O01O010O0010O00010O01O01O01O010O001O01O01O010O0010O00010O01O01O01O010O0010O0001O010O0010O00010O01O01O01O010O0010O000O1O2M2O1N3N1N2N3N1N2O2M2O1N3M2O1N3N1N2O2M2O1N2N3N1N2O2M2O1N3M2O1N3N1N2O2M2N2O2M2O1N2O2M2N2O2M2O1N3N1N2O2M2N2O2M2O1N3N1N2N3N1N2N2N3M2N2N3M2N2N3L3N2N3M2N2N3M2N2M4M2N2N2N3M2N2N3L3N2N3M3M3M3M3M3L4M3M3M3M3M3M3Mo_d3"}, "label": "a blue plate holding fresh fruit and vegetables"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"trail mix in a tray next to fruits and vegetables\"."}], "task": "refering", "answer_template": "The \"trail mix in a tray next to fruits and vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 332, 333, 334, 335, 336, 337, 338, 339, 340, 355, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.45290625, 0.5015873015873016, 0.843359375, 1.0], "iscrowd": 0, "area": 41378.11455, "rle": {"size": [441, 640], "counts": "^Xm3=Z=a0_O5K3M3M4L3M4K5L4L4L5K5K4L5K5K4L2M3N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2M3N2N2N2N3M2N2N2N2N2M3N2N2O1N2N2N2N2N2N2N2O1N2N2N200O100O1O100O100O1O100O100O1000000O1000000O100000000O1000000O100000000000000O1000000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0O1M4M2M3N3L3M3N3L3N2M4L3N2M4L3N2M4M2M4L3N2M4M2M3M4M2M3M4M2M3N3L3M3N3L4M2M4L4M3L4L3N3L4M4K5K5L3L5L5J6J7J5JUiZ1"}, "label": "trail mix in a tray next to fruits and vegetables"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tray on which pebbles are stored\"."}], "task": "refering", "answer_template": "The \"a tray on which pebbles are stored\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 332, 333, 334, 335, 336, 337, 338, 339, 340, 355, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.45290625, 0.5015873015873016, 0.843359375, 1.0], "iscrowd": 0, "area": 41378.11455, "rle": {"size": [441, 640], "counts": "^Xm3=Z=a0_O5K3M3M4L3M4K5L4L4L5K5K4L5K5K4L2M3N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2M3N2N2N2N3M2N2N2N2N2M3N2N2O1N2N2N2N2N2N2N2O1N2N2N200O100O1O100O100O1O100O100O1000000O1000000O100000000O1000000O100000000000000O1000000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0O1M4M2M3N3L3M3N3L3N2M4L3N2M4L3N2M4M2M4L3N2M4M2M3M4M2M3M4M2M3N3L3M3N3L4M2M4L4M3L4L3N3L4M4K5K5L3L5L5J6J7J5JUiZ1"}, "label": "a tray on which pebbles are stored"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little red container holding peanut butter\"."}], "task": "refering", "answer_template": "The \"a little red container holding peanut butter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 225, 226], "bbox": [0.6449374999999999, 0.3546485260770975, 0.867140625, 0.5827437641723356], "iscrowd": 0, "area": 9341.330250000005, "rle": {"size": [441, 640], "counts": "PRb52e=5K4L4M3L4L5K4L3N2M2N3M2O1N3M2N3M2O1N3M2N2O2M2N3M2NjMbDU2a;1O00001O001O00001O001O00001O00001O001O00001O00001O001N10001O010O00010O00010O0001O010O00010O00010O000010O00010O01O00010O00010O00010O000010O01O01O01O01O01O00010O00010O001O0O101O1O0O2O001O0O2L4L3N3L3M4M2M4L3M4M3L3M4L3M4M2M4L4L3M4M2M^]T1"}, "label": "a little red container holding peanut butter"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"penut butter\"."}], "task": "refering", "answer_template": "The \"penut butter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 225, 226], "bbox": [0.6449374999999999, 0.3546485260770975, 0.867140625, 0.5827437641723356], "iscrowd": 0, "area": 9341.330250000005, "rle": {"size": [441, 640], "counts": "PRb52e=5K4L4M3L4L5K4L3N2M2N3M2O1N3M2N3M2O1N3M2N2O2M2N3M2NjMbDU2a;1O00001O001O00001O001O00001O00001O001O00001O00001O001N10001O010O00010O00010O0001O010O00010O00010O000010O00010O01O00010O00010O00010O000010O01O01O01O01O01O00010O00010O001O0O101O1O0O2O001O0O2L4L3N3L3M4M2M4L3M4M3L3M4L3M4M2M4L4L3M4M2M^]T1"}, "label": "penut butter"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fruit salad in cute rose container\"."}], "task": "refering", "answer_template": "The \"fruit salad in cute rose container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 172, 173, 174, 197], "bbox": [0.148984375, 0.0033786848072562357, 0.69371875, 0.5123809523809524], "iscrowd": 0, "area": 45256.442500000005, "rle": {"size": [441, 640], "counts": "Z^Y17a=>B1N2O2N1O1N3N1O1N2O2N1N2O2N1O1N3N1O1N3N1O1N3N1O1O2M2O1O2M2O1O1O2M2O1O2M2O1O2M2O1O2N1N2O2N1N2O2N1O1N2O2N1N2O2N1N2O2N1O1N3N1O1N3N1O1O2M2O1O2M2O1O1N3N1O1O2M2O1O2N10001O00001O00001O00001O0000001O00001O00001O00001O000O2O000010O0001O01O0001O01O01O00010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O01O00010O0000010O00001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O010O1O1O101N2N2N2O1N2N2N01N10000O2O000O101N10000TOjGTLX8i3PHQLP8m3YHlKg7R4aHgK`7W4S1M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2O2M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M3M2N3M3M2N3M3M3M4M3LR[d2"}, "label": "fruit salad in cute rose container"}]}
{"id": 91130, "image": "COCO_train2014_000000091130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the box with the smiley face food in it\"."}], "task": "refering", "answer_template": "The \"the box with the smiley face food in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 172, 173, 174, 197], "bbox": [0.148984375, 0.0033786848072562357, 0.69371875, 0.5123809523809524], "iscrowd": 0, "area": 45256.442500000005, "rle": {"size": [441, 640], "counts": "Z^Y17a=>B1N2O2N1O1N3N1O1N2O2N1N2O2N1O1N3N1O1N3N1O1N3N1O1O2M2O1O2M2O1O1O2M2O1O2M2O1O2M2O1O2N1N2O2N1N2O2N1O1N2O2N1N2O2N1N2O2N1O1N3N1O1N3N1O1O2M2O1O2M2O1O1N3N1O1O2M2O1O2N10001O00001O00001O00001O0000001O00001O00001O00001O000O2O000010O0001O01O0001O01O01O00010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O01O00010O0000010O00001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O010O1O1O101N2N2N2O1N2N2N01N10000O2O000O101N10000TOjGTLX8i3PHQLP8m3YHlKg7R4aHgK`7W4S1M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2O2M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M3M2N3M3M2N3M3M3M4M3LR[d2"}, "label": "the box with the smiley face food in it"}]}
{"id": 476155, "image": "COCO_train2014_000000476155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left ski attatched to a mans foot\"."}], "task": "refering", "answer_template": "The \"the left ski attatched to a mans foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [309, 310, 312, 313, 314, 315, 316, 317, 318, 319, 327, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.2230625, 0.792578125, 0.9585416666666666, 0.87021875], "iscrowd": 0, "area": 6624.442200000019, "rle": {"size": [640, 480], "counts": "`\\S21mc03N1N2N2N3M2N2N2N2O2M2O11O01O0001O0001O0001O0001OM4H7J6Jl`k0OY_TO3M3N2M2N2O2M2N2O10000000000000000010O0000000000000000000001O01O0000000000000000000001O0001O000000000000000001O000001O000000000000000001O00000001O00000000000001O0000000001O000000000001O000000000001O000000000001O00000000000001O00000001O0000000000000000000000001O00000000000000000000001O00000000000000000001O01O0000000000000000000000001O00000000000000000000001O0000000000000000000000001O000001O0000000000000001O00000000000000000000001O0000000000000000000000001O00000000000000000001O01O000000000000O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O106IZo;"}, "label": "the left ski attatched to a mans foot"}]}
{"id": 476155, "image": "COCO_train2014_000000476155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left ski of a skier\"."}], "task": "refering", "answer_template": "The \"the left ski of a skier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [309, 310, 312, 313, 314, 315, 316, 317, 318, 319, 327, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.2230625, 0.792578125, 0.9585416666666666, 0.87021875], "iscrowd": 0, "area": 6624.442200000019, "rle": {"size": [640, 480], "counts": "`\\S21mc03N1N2N2N3M2N2N2N2O2M2O11O01O0001O0001O0001O0001OM4H7J6Jl`k0OY_TO3M3N2M2N2O2M2N2O10000000000000000010O0000000000000000000001O01O0000000000000000000001O0001O000000000000000001O000001O000000000000000001O00000001O00000000000001O0000000001O000000000001O000000000001O000000000001O00000000000001O00000001O0000000000000000000000001O00000000000000000000001O00000000000000000001O01O0000000000000000000000001O00000000000000000000001O0000000000000000000000001O000001O0000000000000001O00000000000000000000001O0000000000000000000000001O00000000000000000001O01O000000000000O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O106IZo;"}, "label": "the left ski of a skier"}]}
{"id": 476155, "image": "COCO_train2014_000000476155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person ' s right ski\"."}], "task": "refering", "answer_template": "The \"the person ' s right ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [350, 351, 352, 353, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 374], "bbox": [0.0069166666666666664, 0.877921875, 0.8242499999999999, 0.9610312499999999], "iscrowd": 0, "area": 5510.984050000001, "rle": {"size": [640, 480], "counts": "US32lc02N2O1N2N2O1000O10000000000000000000O100000000000O10000000000000O1000000000000000O1000000000O1000000000000000000000O100000O10000000000000000000000000O0100000000000000000000000000O1000O100000000000000000000000O10000000O100000000000000000O10000000000000O100000000000O10000000000000000000O100000O100000000000000000000000O1000O100000000000000000000000000O010000000000000000000000000O100000O1000000000000000000000O1000000000O1000000000000000O1000000000000000O1000000000O1000000000000000000000O100000O10000000000000000000000000O0100000000000000000000000000O10O100000000000000000000000O10000000O1000O100000000O100000000O10000000O10O100000000O100000000O100000000O01000000000O100000000O100000000O10O10000000O100000000O100000000O1000O10000O00100O1O10O01O100O10O01O100O1O010O0010O01O010O005L8GSnc1"}, "label": "the person ' s right ski"}]}
{"id": 197629, "image": "COCO_train2014_000000197629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an onlooker with a camera shoots a trio of flying kites\"."}], "task": "refering", "answer_template": "The \"an onlooker with a camera shoots a trio of flying kites\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 196, 197, 198, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 240, 241, 242, 243, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 330, 331, 332], "bbox": [0.006721311475409836, 0.53609375, 0.30245901639344264, 0.9912500000000002], "iscrowd": 0, "area": 22404.403849999995, "rle": {"size": [640, 427], "counts": "`\\2W3d<U4O1000O10O100000O010000OO2O1N2N1O2O1N1O2N2N2O04M>B5I2L4L4M3M3N2M3M3N2M3N200O1O101N1O100O1O100O1O100O100O1O100O101O0O1000000000000000YHSE[6m:bIVE^6j:^I[Ea6e:\\I_Ec6a:\\I`Ed6`:ZIcEZ6g:fI[Em5Q;RJPEb5\\;\\JgDW5e;hJ\\Db2^OhNb<eNRD^2F[Ng<SOeC\\20nMj<AXC]28`Mo<OkB[2P?aMRAY2V?bMj_O9d0Q2k?QNT@f1T`0\\Nk_OT1d`0nNZ_Ob0Va0_Oj^O9]a0Ib^O6^a0L`^O5^a0Na^O1^a02a^OM_a05_^OLaa05^^OJba08]^OGda0:Z^OGea0;Z^ODga0=X^OCga0?W^OAja0?V^O@ja0b0U^O^Ola0Z18H9H6I1O2O1N101N2N101N1O2O1N1O2O1N1N3M2M4M3M2M4M3Mfji5"}, "label": "an onlooker with a camera shoots a trio of flying kites"}]}
{"id": 197629, "image": "COCO_train2014_000000197629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black shirt and jeans photographing kites\"."}], "task": "refering", "answer_template": "The \"man in black shirt and jeans photographing kites\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 196, 197, 198, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 240, 241, 242, 243, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 330, 331, 332], "bbox": [0.006721311475409836, 0.53609375, 0.30245901639344264, 0.9912500000000002], "iscrowd": 0, "area": 22404.403849999995, "rle": {"size": [640, 427], "counts": "`\\2W3d<U4O1000O10O100000O010000OO2O1N2N1O2O1N1O2N2N2O04M>B5I2L4L4M3M3N2M3M3N2M3N200O1O101N1O100O1O100O1O100O100O1O100O101O0O1000000000000000YHSE[6m:bIVE^6j:^I[Ea6e:\\I_Ec6a:\\I`Ed6`:ZIcEZ6g:fI[Em5Q;RJPEb5\\;\\JgDW5e;hJ\\Db2^OhNb<eNRD^2F[Ng<SOeC\\20nMj<AXC]28`Mo<OkB[2P?aMRAY2V?bMj_O9d0Q2k?QNT@f1T`0\\Nk_OT1d`0nNZ_Ob0Va0_Oj^O9]a0Ib^O6^a0L`^O5^a0Na^O1^a02a^OM_a05_^OLaa05^^OJba08]^OGda0:Z^OGea0;Z^ODga0=X^OCga0?W^OAja0?V^O@ja0b0U^O^Ola0Z18H9H6I1O2O1N101N2N101N1O2O1N1O2O1N1N3M2M4M3M2M4M3Mfji5"}, "label": "man in black shirt and jeans photographing kites"}]}
{"id": 361469, "image": "COCO_train2014_000000361469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man eating donut\"."}], "task": "refering", "answer_template": "The \"man eating donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 82, 83, 84, 95, 96, 97, 107, 108, 109, 110, 119, 120, 121, 122, 123, 132, 133, 134, 135, 136, 144, 145, 146, 147, 148, 149, 150, 151, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227], "bbox": [0.09029333333333334, 0.30982, 0.6952533333333334, 0.9825], "iscrowd": 0, "area": 41024.51, "rle": {"size": [500, 375], "counts": "no`07W?:F;F9F;E:G:E:F;F9F;E:F;F9F:F;F9H9I4L1O1O1O1O1O1O2N1O2N2N2M3N2N2O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2M3M3L4L4L4L4L4N2O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1001O00001O00001O00001O00001O0gHPKn3P5QLVKj3k4ULZKf3h4WL^Kd3d4YLbKc3`4ZLeKa3]4\\LiK_3Z4^LlKQOhN`2^5\\NPLcNTOk2n4`NSLUNZO[3f4\\NhMd1Z2ZNfMf1^2UNcMk1b2oM_MQ2e2kM[MU2i2fMXMc0c4YO]K1]5JdJJl52TJNP6MQJ2i:00O100O100O100OO2K4K6J5K6J5N30O100O00100O10O0100O10O0100O10O0100O1O001O1O1O00100O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O2N8H9G9G^Zg1"}, "label": "man eating donut"}]}
{"id": 361469, "image": "COCO_train2014_000000361469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a doughnut\"."}], "task": "refering", "answer_template": "The \"a man holding a doughnut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 82, 83, 84, 95, 96, 97, 107, 108, 109, 110, 119, 120, 121, 122, 123, 132, 133, 134, 135, 136, 144, 145, 146, 147, 148, 149, 150, 151, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227], "bbox": [0.09029333333333334, 0.30982, 0.6952533333333334, 0.9825], "iscrowd": 0, "area": 41024.51, "rle": {"size": [500, 375], "counts": "no`07W?:F;F9F;E:G:E:F;F9F;E:F;F9F:F;F9H9I4L1O1O1O1O1O1O2N1O2N2N2M3N2N2O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2M3M3L4L4L4L4L4N2O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1001O00001O00001O00001O00001O0gHPKn3P5QLVKj3k4ULZKf3h4WL^Kd3d4YLbKc3`4ZLeKa3]4\\LiK_3Z4^LlKQOhN`2^5\\NPLcNTOk2n4`NSLUNZO[3f4\\NhMd1Z2ZNfMf1^2UNcMk1b2oM_MQ2e2kM[MU2i2fMXMc0c4YO]K1]5JdJJl52TJNP6MQJ2i:00O100O100O100OO2K4K6J5K6J5N30O100O00100O10O0100O10O0100O10O0100O1O001O1O1O00100O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O2N8H9G9G^Zg1"}, "label": "a man holding a doughnut"}]}
{"id": 173056, "image": "COCO_train2014_000000173056.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown goat with white spot on head\"."}], "task": "refering", "answer_template": "The \"brown goat with white spot on head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292], "bbox": [0.5628437500000001, 0.39737991266375544, 0.839453125, 0.779410480349345], "iscrowd": 0, "area": 18801.903800000004, "rle": {"size": [458, 640], "counts": "RWQ51U>4M210O100O1O010O1O100O010O1O100O010O10001O3M2N2N3L3N3M3L4L4L4L4M2M4L4L4L4]FnMg6V2UIkMk6X2QIjMm6Y2PIhMP7Y2nHiMQ7Y2lHhMT7Z2hHiMW7X2gHiMZ7X2cHjM\\7W2bHjM^7X2_HjM`7X2]HiMc7X2[HjMd7X2VHlMj7U2QHPNn7R2lGRNU8o1eGVNZ8k1RGhNn8[3O0101M2O0O2O0O1O100O10lFVKi8h4UG[Kk8e4RG]Ko8o40000000O2O0000000O2O00001N101O000N3N1N2O1N2N2O1N2O1N2D;3N1O2N2N1O001O00000O1000000O10000O100O10000O10000O10O1M2O2N2N1O2M3N1O2N1O2N2N3M3M2O2dNZEPOi:n0ZEnNh:Q1[EkNh:S1[EiNh:V1ZEfNh:\\1XE`Nk:b1TEZNo:h1QETNP;n1f02N2M2O2N1N5L3M3L5L3XEYMf9f3O01O10O01O10O01O10O01mLaFY1`9[NnFc1Q9RN]Gk1d:M3N2M3M3^Ohi]1"}, "label": "brown goat with white spot on head"}]}
{"id": 173056, "image": "COCO_train2014_000000173056.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle sized goat\"."}], "task": "refering", "answer_template": "The \"the middle sized goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292], "bbox": [0.5628437500000001, 0.39737991266375544, 0.839453125, 0.779410480349345], "iscrowd": 0, "area": 18801.903800000004, "rle": {"size": [458, 640], "counts": "RWQ51U>4M210O100O1O010O1O100O010O1O100O010O10001O3M2N2N3L3N3M3L4L4L4L4M2M4L4L4L4]FnMg6V2UIkMk6X2QIjMm6Y2PIhMP7Y2nHiMQ7Y2lHhMT7Z2hHiMW7X2gHiMZ7X2cHjM\\7W2bHjM^7X2_HjM`7X2]HiMc7X2[HjMd7X2VHlMj7U2QHPNn7R2lGRNU8o1eGVNZ8k1RGhNn8[3O0101M2O0O2O0O1O100O10lFVKi8h4UG[Kk8e4RG]Ko8o40000000O2O0000000O2O00001N101O000N3N1N2O1N2N2O1N2O1N2D;3N1O2N2N1O001O00000O1000000O10000O100O10000O10000O10O1M2O2N2N1O2M3N1O2N1O2N2N3M3M2O2dNZEPOi:n0ZEnNh:Q1[EkNh:S1[EiNh:V1ZEfNh:\\1XE`Nk:b1TEZNo:h1QETNP;n1f02N2M2O2N1N5L3M3L5L3XEYMf9f3O01O10O01O10O01O10O01mLaFY1`9[NnFc1Q9RN]Gk1d:M3N2M3M3^Ohi]1"}, "label": "the middle sized goat"}]}
{"id": 173056, "image": "COCO_train2014_000000173056.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall grey sheep stands beside a black baby sheep\"."}], "task": "refering", "answer_template": "The \"a tall grey sheep stands beside a black baby sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 171, 172, 190, 191, 194, 212, 213, 214, 215, 217, 236, 237], "bbox": [0.18467187499999999, 0.13334061135371178, 0.5355624999999999, 0.6494541484716156], "iscrowd": 0, "area": 28916.418599999997, "rle": {"size": [458, 640], "counts": "jZe11X>2O1N2O1N2O0O100O10000O100O1000000O0100000O101O2N2M4M4L3M4K5L3M4Le0ZO2O1O1O1N101O1O1N3N2N3QGdMj5^2RJoMb5U2ZJkMg5X2fIVN\\6l1VI_Nj6d1hHhNX7[1]HnNa7U1oGYOQ8i0kGZOT8_3N1O2N1O2M2O1O00001N101O00001O01O1000O10001O001O00001O00001N101O00001O00001O001O000O01O1O1O1N2O1O1nNbHVK`7b4kHZKU7_4VI]Kk6\\4_I_Kc6Y4hIcKX6V4SJfKn5R4X2A`0O11N100O2O000O2O0O2O001N2O0O2O001N1000000O101O0000000O10001O00001N10001O0000O01000O100O10O0100O100ObGnKf6R4WITLf6k3YIYLd6k3WIXLh6l3RIWLm6k3lHZLT7i3dH]L[7e3^H`Lb7b3oGkLQ8h4O000O10001O0000N2N2O0O2M3J6J6N3N1N3M3M3M2D=L4M2N3L4M3L3N3M3\\Od0M2N3N2M2O2M3M3N1N3M3M2M4M3I7G8F;G9L3M4L4L4K4M7I7HngT4"}, "label": "a tall grey sheep stands beside a black baby sheep"}]}
{"id": 173056, "image": "COCO_train2014_000000173056.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother sheep looking at the camera\"."}], "task": "refering", "answer_template": "The \"a mother sheep looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 171, 172, 190, 191, 194, 212, 213, 214, 215, 217, 236, 237], "bbox": [0.18467187499999999, 0.13334061135371178, 0.5355624999999999, 0.6494541484716156], "iscrowd": 0, "area": 28916.418599999997, "rle": {"size": [458, 640], "counts": "jZe11X>2O1N2O1N2O0O100O10000O100O1000000O0100000O101O2N2M4M4L3M4K5L3M4Le0ZO2O1O1O1N101O1O1N3N2N3QGdMj5^2RJoMb5U2ZJkMg5X2fIVN\\6l1VI_Nj6d1hHhNX7[1]HnNa7U1oGYOQ8i0kGZOT8_3N1O2N1O2M2O1O00001N101O00001O01O1000O10001O001O00001O00001N101O00001O00001O001O000O01O1O1O1N2O1O1nNbHVK`7b4kHZKU7_4VI]Kk6\\4_I_Kc6Y4hIcKX6V4SJfKn5R4X2A`0O11N100O2O000O2O0O2O001N2O0O2O001N1000000O101O0000000O10001O00001N10001O0000O01000O100O10O0100O100ObGnKf6R4WITLf6k3YIYLd6k3WIXLh6l3RIWLm6k3lHZLT7i3dH]L[7e3^H`Lb7b3oGkLQ8h4O000O10001O0000N2N2O0O2M3J6J6N3N1N3M3M3M2D=L4M2N3L4M3L3N3M3\\Od0M2N3N2M2O2M3M3N1N3M3M2M4M3I7G8F;G9L3M4L4L4K4M7I7HngT4"}, "label": "a mother sheep looking at the camera"}]}
{"id": 173056, "image": "COCO_train2014_000000173056.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little black lamb\"."}], "task": "refering", "answer_template": "The \"the little black lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 308, 309, 311, 331], "bbox": [0.42056250000000006, 0.48314410480349346, 0.565390625, 0.9040829694323144], "iscrowd": 0, "area": 9076.200099999996, "rle": {"size": [458, 640], "counts": "hfh38n=3M4L3M4L4L3M4L3M4J5K4L4L4L4L4K5J6J6J6J6K5J6000O001O1N101N2O001N2O0O2O0011O2N100O2N1O102M3M3M3N2M4L5L3O1N2N3M2O1N2N2O2M2N1O100O2N100O1O1O101N1O100O1O0001N1QNfFnNZ9Q1oFfNR9X1TGdNP9V1VGfNn8T1WGhNn8R1XGjNm8P1XGkNn8n0XGnNn8j0WGSOo8e0WGVOP9b0VGZOP9>UG_OP9;UG@R98g2HjQl3"}, "label": "the little black lamb"}]}
{"id": 173056, "image": "COCO_train2014_000000173056.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is a black goat\"."}], "task": "refering", "answer_template": "The \"it is a black goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 308, 309, 311, 331], "bbox": [0.42056250000000006, 0.48314410480349346, 0.565390625, 0.9040829694323144], "iscrowd": 0, "area": 9076.200099999996, "rle": {"size": [458, 640], "counts": "hfh38n=3M4L3M4L4L3M4L3M4J5K4L4L4L4L4K5J6J6J6J6K5J6000O001O1N101N2O001N2O0O2O0011O2N100O2N1O102M3M3M3N2M4L5L3O1N2N3M2O1N2N2O2M2N1O100O2N100O1O1O101N1O100O1O0001N1QNfFnNZ9Q1oFfNR9X1TGdNP9V1VGfNn8T1WGhNn8R1XGjNm8P1XGkNn8n0XGnNn8j0WGSOo8e0WGVOP9b0VGZOP9>UG_OP9;UG@R98g2HjQl3"}, "label": "it is a black goat"}]}
{"id": 9218, "image": "COCO_train2014_000000009218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the shadows riding an elephant\"."}], "task": "refering", "answer_template": "The \"the person in the shadows riding an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 28, 29, 30, 51, 52, 53, 54, 74, 75], "bbox": [0.16078125000000001, 0.0011241217798594848, 0.36665625, 0.25789227166276346], "iscrowd": 0, "area": 5446.044999999999, "rle": {"size": [427, 640], "counts": "gnZ11Z=2N1O2M3N2N1O000O1000O1000O1000O1000O10O1000000O0100000O10OFRC9o<1O10O01O100O00100000O2O0000bEIh77XH4]7KcH6]7JcH6]7JcH6]7JcH6]7IdH6]7JbH7^7IbH7^7IbH7^7HcH8]7HcH8]7HcH8]7HmG7YN2j9G]Gh0hN@l9GYGo0iNZOn9GVGe1j8[NTGg1l8YNQGj1o8UNoFn1Q9UNjFm1V9S1100O1M3M3M3M3N2M3M3M3M3M3M3M3M3M4O000001O000000000000001O000000000000001O00001O1O1O1O1O1O1O1O1O1O1OlNUEIj:_OPF`0_;N2O1N2N2M3N2N2N3M2M3N2N2N2NUWY5"}, "label": "the person in the shadows riding an elephant"}]}
{"id": 9218, "image": "COCO_train2014_000000009218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with chains attached in the sunlight\"."}], "task": "refering", "answer_template": "The \"an elephant with chains attached in the sunlight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 199, 200, 202, 203, 217, 218, 219, 222, 223], "bbox": [0.442296875, 0.17079625292740047, 0.8590937499999999, 0.662927400468384], "iscrowd": 0, "area": 34944.943799999994, "rle": {"size": [427, 640], "counts": "\\Xf31Y=2N2N1N3eM9oFIf8S1gFoNo8[1oFfNh8]1\\GfN_8X1fGjNX8Q1mGPOR8k0SHWOk7e0XH^Of7>^HD`78dHIZ74jHNT7NPI4n6HUI:j6BZI`0d6CYI?e6CYI?e6DWI>h6DVI>h6EUI=i6EUI=i6ETI=k6FRI<l6FRI<k6HRI9m6IQI9m6JoH9o6IoH8P7n0jGSOU8R3O1O100O100O100O1O100O101N100O1O100O100O100O1O100O100OhN[HlLd7S3bHiL^7U3iHfLV7Y3PIcLP7[3VIaLi6^3]I^Lc6`3cI\\L\\6c3jIYLU6f3QJVLo5h3j1O2N1O2O0O1O2N1O2O0100O1O100O1O100O10O01O100O1O100O1O100O1O100O10O01O100O1O100O1O100O10000000O11O4L4L4L4L2N001O1O001O1N101O1O001O1O001O001O1O001O1O1O2N3M2N3M2N2N3M2N2N3M2N2N3MO2O001O0O2O001O000O2O001O0O2O001O000O2O001O0O2O0K6cN\\10010O001O010O0010O0001O010O001O010O00010O001O010O1O2jMaF=`9ZOfFg0]9nNkFR1W9eNoF\\1T9YNSGj1m8mMZGU2l93M4M2M4L3N2M4L3M4M2K5K6J3L4M3M3M3M4L3M3L4O11O2N1O1O2N1O1O2N1O2N1O1O2N101N2N1O2N2N2N2kN\\F_Nf9\\1cF^N_9\\1kF^NW9]1RG]No8^1ZG]Nh8^1aG\\Na8^1j1J6K5UOi\\U1"}, "label": "an elephant with chains attached in the sunlight"}]}
{"id": 9218, "image": "COCO_train2014_000000009218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant in the sun\"."}], "task": "refering", "answer_template": "The \"elephant in the sun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 199, 200, 202, 203, 217, 218, 219, 222, 223], "bbox": [0.442296875, 0.17079625292740047, 0.8590937499999999, 0.662927400468384], "iscrowd": 0, "area": 34944.943799999994, "rle": {"size": [427, 640], "counts": "\\Xf31Y=2N2N1N3eM9oFIf8S1gFoNo8[1oFfNh8]1\\GfN_8X1fGjNX8Q1mGPOR8k0SHWOk7e0XH^Of7>^HD`78dHIZ74jHNT7NPI4n6HUI:j6BZI`0d6CYI?e6CYI?e6DWI>h6DVI>h6EUI=i6EUI=i6ETI=k6FRI<l6FRI<k6HRI9m6IQI9m6JoH9o6IoH8P7n0jGSOU8R3O1O100O100O100O1O100O101N100O1O100O100O100O1O100O100OhN[HlLd7S3bHiL^7U3iHfLV7Y3PIcLP7[3VIaLi6^3]I^Lc6`3cI\\L\\6c3jIYLU6f3QJVLo5h3j1O2N1O2O0O1O2N1O2O0100O1O100O1O100O10O01O100O1O100O1O100O1O100O10O01O100O1O100O1O100O10000000O11O4L4L4L4L2N001O1O001O1N101O1O001O1O001O001O1O001O1O1O2N3M2N3M2N2N3M2N2N3M2N2N3MO2O001O0O2O001O000O2O001O0O2O001O000O2O001O0O2O0K6cN\\10010O001O010O0010O0001O010O001O010O00010O001O010O1O2jMaF=`9ZOfFg0]9nNkFR1W9eNoF\\1T9YNSGj1m8mMZGU2l93M4M2M4L3N2M4L3M4M2K5K6J3L4M3M3M3M4L3M3L4O11O2N1O1O2N1O1O2N1O2N1O1O2N101N2N1O2N2N2N2kN\\F_Nf9\\1cF^N_9\\1kF^NW9]1RG]No8^1ZG]Nh8^1aG\\Na8^1j1J6K5UOi\\U1"}, "label": "elephant in the sun"}]}
{"id": 9218, "image": "COCO_train2014_000000009218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant in deep shade\"."}], "task": "refering", "answer_template": "The \"elephant in deep shade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 188, 189, 190, 191, 192, 207, 208, 209, 212, 213, 214, 215, 230, 231, 232, 236, 237, 238, 239, 253, 254, 255, 256], "bbox": [0.0, 0.004473067915690867, 0.40090624999999996, 0.7981967213114755], "iscrowd": 0, "area": 64715.51895, "rle": {"size": [427, 640], "counts": "j7^2o5n4aM_2O100O1O1O100O1O1O10000O1000001O000O101O00001O0O2O001O001N2O001O001O001O1O00000000001O0000000000000000001O000000000000000000O100000000000000000000000000QMkMhJU2P5aNbJ_1V5WO\\Ji0]5MTJ3d5b0oI^Oi5k0VJUOc5T1[JlN_5[1_JfN`5\\1_JdN_5`1_J`N`5b1_J^N`5d1_J]N^5g1`JYN_5i1`JWN_5k1`JUN^5o1`JQN`5P2_JPNb5P2]JPNd5k4001O1O001O1O001O1O4L3M4L6J9G5K1O1O1O2N1O1O1O1O2N1O1O10001N10000O101O0O10001N1000000O2O000O10001N10000O2O0O100O2O0O1O2O0O101N100O1O2O0O101N1O101N100O1O2O0O101N1O01O00001O00001N1000001O00001O00001O00001O00001O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N4L3M3M4L3M3M4PM`ImNc6k0lIkNW6n0WJhNm5P1aJfNb5S1lJcNW5V1WK`Nm4W1_KbNd4W1bKgNa4R1dKmN_4l0gKRO[4g0kKXOW4b0oK\\OT4=RLAP49ULFm33ZLKi3N\\L1f3I`L5]8N1O2N1OY]o4"}, "label": "elephant in deep shade"}]}
{"id": 9218, "image": "COCO_train2014_000000009218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant in the shade helping another elephant roll a log\"."}], "task": "refering", "answer_template": "The \"an elephant in the shade helping another elephant roll a log\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 188, 189, 190, 191, 192, 207, 208, 209, 212, 213, 214, 215, 230, 231, 232, 236, 237, 238, 239, 253, 254, 255, 256], "bbox": [0.0, 0.004473067915690867, 0.40090624999999996, 0.7981967213114755], "iscrowd": 0, "area": 64715.51895, "rle": {"size": [427, 640], "counts": "j7^2o5n4aM_2O100O1O1O100O1O1O10000O1000001O000O101O00001O0O2O001O001N2O001O001O001O1O00000000001O0000000000000000001O000000000000000000O100000000000000000000000000QMkMhJU2P5aNbJ_1V5WO\\Ji0]5MTJ3d5b0oI^Oi5k0VJUOc5T1[JlN_5[1_JfN`5\\1_JdN_5`1_J`N`5b1_J^N`5d1_J]N^5g1`JYN_5i1`JWN_5k1`JUN^5o1`JQN`5P2_JPNb5P2]JPNd5k4001O1O001O1O001O1O4L3M4L6J9G5K1O1O1O2N1O1O1O1O2N1O1O10001N10000O101O0O10001N1000000O2O000O10001N10000O2O0O100O2O0O1O2O0O101N100O1O2O0O101N1O101N100O1O2O0O101N1O01O00001O00001N1000001O00001O00001O00001O00001O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N4L3M3M4L3M3M4PM`ImNc6k0lIkNW6n0WJhNm5P1aJfNb5S1lJcNW5V1WK`Nm4W1_KbNd4W1bKgNa4R1dKmN_4l0gKRO[4g0kKXOW4b0oK\\OT4=RLAP49ULFm33ZLKi3N\\L1f3I`L5]8N1O2N1OY]o4"}, "label": "an elephant in the shade helping another elephant roll a log"}]}
{"id": 295940, "image": "COCO_train2014_000000295940.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby in a green shirt holds a paper plate\"."}], "task": "refering", "answer_template": "The \"a baby in a green shirt holds a paper plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 98, 99, 100, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 193, 194, 261, 283, 284, 285, 307, 308, 330, 331], "bbox": [0.2485, 0.19530516431924883, 0.47467187499999997, 1.0], "iscrowd": 0, "area": 14395.654650000004, "rle": {"size": [426, 640], "counts": "V[R22T=5J7I6L5L3L4L5L3L5L3L3M2O2M200001O00UOaNmE^1R:fNkE[1T:gNkEX1U:kNhEU1P:WOmEh0l9BQF>h9NTF2g9j110O1O1O001O1O1O1O1O2N2N2O1N2O001N101O0O100O01O10O01000O1nJmKm0R4SOQLl0g3UKWLn35k0c3\\O_Lc0a3]OaLa0_3^OdL`0\\3_OgL?X3BjL<V3CmL<R3CQM;o2ERM:m2FVM8j2GYM7g2H\\M6d2J^M5a2JbM4^2KeM4[2KfM4Z2KiM4V2KmM4R2LRN0o1NlNYOS1g0YOnNf0S1YOnNf0S1YOmNh0S1WOnNh0R1XOoNg0R1XOnNh0S1WOnNi0R1VOoN[OPM[OQ4Z1oNYOQM^OQ4X1oNXOQM@Q4W1oNVORMDo3U1oNUOUMEl3W1TOlNRMMk3V1YOjMoL00o0i3V1^O^MQM:JS1g3T15fMVLV1g3R14eMXLY1i3l00gMZL]1l3f0c0ZOC?=AJ86H110O3MM34LM43KM54JL74HM=NB2a0L^O5a0M]O3d0NZO3e0OWO4i0MQO8n05bN0]1o5O10000O2OYDaNb;d10O100012M3M3M3MO2M2N2N2N2N3M2N2N2M3N3M2M2N3M3M3M3M3M3M3M]b[4"}, "label": "a baby in a green shirt holds a paper plate"}]}
{"id": 295940, "image": "COCO_train2014_000000295940.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling baby in a green shirt being held by a woman\"."}], "task": "refering", "answer_template": "The \"a smiling baby in a green shirt being held by a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 98, 99, 100, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 193, 194, 261, 283, 284, 285, 307, 308, 330, 331], "bbox": [0.2485, 0.19530516431924883, 0.47467187499999997, 1.0], "iscrowd": 0, "area": 14395.654650000004, "rle": {"size": [426, 640], "counts": "V[R22T=5J7I6L5L3L4L5L3L5L3L3M2O2M200001O00UOaNmE^1R:fNkE[1T:gNkEX1U:kNhEU1P:WOmEh0l9BQF>h9NTF2g9j110O1O1O001O1O1O1O1O2N2N2O1N2O001N101O0O100O01O10O01000O1nJmKm0R4SOQLl0g3UKWLn35k0c3\\O_Lc0a3]OaLa0_3^OdL`0\\3_OgL?X3BjL<V3CmL<R3CQM;o2ERM:m2FVM8j2GYM7g2H\\M6d2J^M5a2JbM4^2KeM4[2KfM4Z2KiM4V2KmM4R2LRN0o1NlNYOS1g0YOnNf0S1YOnNf0S1YOmNh0S1WOnNh0R1XOoNg0R1XOnNh0S1WOnNi0R1VOoN[OPM[OQ4Z1oNYOQM^OQ4X1oNXOQM@Q4W1oNVORMDo3U1oNUOUMEl3W1TOlNRMMk3V1YOjMoL00o0i3V1^O^MQM:JS1g3T15fMVLV1g3R14eMXLY1i3l00gMZL]1l3f0c0ZOC?=AJ86H110O3MM34LM43KM54JL74HM=NB2a0L^O5a0M]O3d0NZO3e0OWO4i0MQO8n05bN0]1o5O10000O2OYDaNb;d10O100012M3M3M3MO2M2N2N2N2N3M2N2N2M3N3M2M2N3M3M3M3M3M3M3M]b[4"}, "label": "a smiling baby in a green shirt being held by a woman"}]}
{"id": 295940, "image": "COCO_train2014_000000295940.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman and a small child\"."}], "task": "refering", "answer_template": "The \"a woman and a small child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 25, 26, 27, 28, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330], "bbox": [0.051203125, 0.05053990610328639, 0.43156249999999996, 0.9868075117370891], "iscrowd": 0, "area": 50138.8851, "rle": {"size": [426, 640], "counts": "on=b0Q<i0VOi0K6L3M4L4L3M4L3M4L3M4L3M4M3L3M4L3M4L3M4L3M4L4L3M4L3M4L3]NVLbIn3W6f1J7J6I6J7I6K6I6J7K5M2N3M2N3M3M4L3M4L4L3M4L3M4L4L3M4L3N3L3N3M3M2^NVE?n:[OWEc0k:YOXEf0k:UOXEi0l:QOYEm0i:oNZEP1a;N3N000O1000000O10000O10000O100000000000000001O00000000001O000PHUOa3k0[LYOe3g0ZLZOf3f0ZLZOf3f0YL[Og3f0XLZOh3f0XLZOh3f0WL[Oi3e0WL[Oi3e0WL[Oi3e0WL[Oi3e0VL]Oi3d0SL_Om3a0QLAo3?oKCQ4=mKES4<jKFV4;gKGY4:cKI]48`KJ`47]KKc46ZKLf45WKMi44TKOk42QK1WNjNa5W1TL2WNlNc5S1SL5VNkNf5Q1QL8TNlNi5m0PL:TNlNk5k0nK=RNlNo5h0kKa0QNlNR6d0jKc0QNlNT6b0hKf0oMmNW6>hKm1eNdLi4b1^Ll1dNhLl4^1^Ll1`NlLP5[1ZKcN0X3APMS5X1VKeN3R4d4[MTKfN8Q4b4\\MQKfN;Q4b4WN]Kk1a4UN^Kn1`4SN^KQ2_4_3N2N2N2N2N001O001O1O001O001O1O001O001O1O001O001O0000001O000000001O0000001O0000000iMkF9V9EmF9S9FPG8P9HRG6n8IVG4j8LXG2i8LZG2f8N\\G0d8O`GN`82bGL_82dGL\\83gGKY85jGHV87mGGS89oGER89QHEo7;THAm7>VH]Om7c0UHWOo7h0UHQOP8m0SHmNQ8S1QHhNR8W1P210O02O0O2O001N101N2O0O2O0O2O001N100O010O01O00100O0010O01O0POQ1O0O2N3N3L3N3L3MkTg4"}, "label": "a woman and a small child"}]}
{"id": 295940, "image": "COCO_train2014_000000295940.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady holding a child\"."}], "task": "refering", "answer_template": "The \"a lady holding a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 25, 26, 27, 28, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330], "bbox": [0.051203125, 0.05053990610328639, 0.43156249999999996, 0.9868075117370891], "iscrowd": 0, "area": 50138.8851, "rle": {"size": [426, 640], "counts": "on=b0Q<i0VOi0K6L3M4L4L3M4L3M4L3M4L3M4M3L3M4L3M4L3M4L3M4L4L3M4L3M4L3]NVLbIn3W6f1J7J6I6J7I6K6I6J7K5M2N3M2N3M3M4L3M4L4L3M4L3M4L4L3M4L3N3L3N3M3M2^NVE?n:[OWEc0k:YOXEf0k:UOXEi0l:QOYEm0i:oNZEP1a;N3N000O1000000O10000O10000O100000000000000001O00000000001O000PHUOa3k0[LYOe3g0ZLZOf3f0ZLZOf3f0YL[Og3f0XLZOh3f0XLZOh3f0WL[Oi3e0WL[Oi3e0WL[Oi3e0WL[Oi3e0VL]Oi3d0SL_Om3a0QLAo3?oKCQ4=mKES4<jKFV4;gKGY4:cKI]48`KJ`47]KKc46ZKLf45WKMi44TKOk42QK1WNjNa5W1TL2WNlNc5S1SL5VNkNf5Q1QL8TNlNi5m0PL:TNlNk5k0nK=RNlNo5h0kKa0QNlNR6d0jKc0QNlNT6b0hKf0oMmNW6>hKm1eNdLi4b1^Ll1dNhLl4^1^Ll1`NlLP5[1ZKcN0X3APMS5X1VKeN3R4d4[MTKfN8Q4b4\\MQKfN;Q4b4WN]Kk1a4UN^Kn1`4SN^KQ2_4_3N2N2N2N2N001O001O1O001O001O1O001O001O1O001O001O0000001O000000001O0000001O0000000iMkF9V9EmF9S9FPG8P9HRG6n8IVG4j8LXG2i8LZG2f8N\\G0d8O`GN`82bGL_82dGL\\83gGKY85jGHV87mGGS89oGER89QHEo7;THAm7>VH]Om7c0UHWOo7h0UHQOP8m0SHmNQ8S1QHhNR8W1P210O02O0O2O001N101N2O0O2O0O2O001N100O010O01O00100O0010O01O0POQ1O0O2N3N3L3N3L3MkTg4"}, "label": "a lady holding a child"}]}
{"id": 9226, "image": "COCO_train2014_000000009226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train with 931 on the front\"."}], "task": "refering", "answer_template": "The \"a train with 931 on the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246], "bbox": [0.298453125, 0.2898823529411765, 0.744640625, 0.703364705882353], "iscrowd": 0, "area": 32709.518899999995, "rle": {"size": [425, 640], "counts": "e^_21W=2N2M2O2N2M2O2N2N2O001N2O1O2M3N2N1N2O0O10001N10001N10001N2O001N2O001N101O0O101O0O2O001N101O0O2O000O2O001O001O0O2O001O00001O001N2O001O1O001O1O0O2O001O1O001O1O0O2O001O00001O000O2O001O00001O000O2O001N10001N100O2O001N101O0O2O000O2O001N101O0O101N101O0O2O001N101O1N2O1O0O2O1N2O1O1N101O1N2O1O1N101N2O1N2O1N2OLTGkKk8W4UGiKj8X4WGgKg8[4YGeKf8\\4ZGdKd8^4]GaKb8_441N2O0O2O1N2O0O2O1N101N2O1N101N101N100O2O0O10001N100O2O000000000010O0000000000000000000O1000000000000000001N1000000000000000000O1000000001O00000010O000001O000000001O0001O010O001O010O00001O001O001O001O0O101O001O001O001O1O2M3N1N3N2N1N3N2N1N3N2N1N3M3L3N3nNQGdMS9Q2bGbMb:2o^S2"}, "label": "a train with 931 on the front"}]}
{"id": 9226, "image": "COCO_train2014_000000009226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train on the track , number 931\"."}], "task": "refering", "answer_template": "The \"a train on the track , number 931\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246], "bbox": [0.298453125, 0.2898823529411765, 0.744640625, 0.703364705882353], "iscrowd": 0, "area": 32709.518899999995, "rle": {"size": [425, 640], "counts": "e^_21W=2N2M2O2N2M2O2N2N2O001N2O1O2M3N2N1N2O0O10001N10001N10001N2O001N2O001N101O0O101O0O2O001N101O0O2O000O2O001O001O0O2O001O00001O001N2O001O1O001O1O0O2O001O1O001O1O0O2O001O00001O000O2O001O00001O000O2O001N10001N100O2O001N101O0O2O000O2O001N101O0O101N101O0O2O001N101O1N2O1O0O2O1N2O1O1N101O1N2O1O1N101N2O1N2O1N2OLTGkKk8W4UGiKj8X4WGgKg8[4YGeKf8\\4ZGdKd8^4]GaKb8_441N2O0O2O1N2O0O2O1N101N2O1N101N101N100O2O0O10001N100O2O000000000010O0000000000000000000O1000000000000000001N1000000000000000000O1000000001O00000010O000001O000000001O0001O010O001O010O00001O001O001O001O0O101O001O001O001O1O2M3N1N3N2N1N3N2N1N3N2N1N3M3L3N3nNQGdMS9Q2bGbMb:2o^S2"}, "label": "a train on the track , number 931"}]}
{"id": 541706, "image": "COCO_train2014_000000541706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccoli on the right\"."}], "task": "refering", "answer_template": "The \"the broccoli on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 138, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194], "bbox": [0.48944, 0.55656, 0.7992999999999999, 0.84296], "iscrowd": 0, "area": 12209.878249999998, "rle": {"size": [375, 500], "counts": "Xni23c;3M4L3M3M3M3M4L3M3N2jERO^9Q1WFZOf9_1L4L4L4L4M3M3L3N1O2N1O2N1N3N1O2N1O2N1O0O10O1000000O0100000O1000O01000000O1000000O1000000O1000000000000000000000000000000000000000000000001O0O1000001O0000001O000000001O000000001O00001N2O1O001O1O001N2O1O001O1O001N2O001O3M2M3N3M2N3M2M3N3M1O1O0O2O001O001N101O001O0O2O0O1O2M2O2M2O3L8I7H7JnfT1"}, "label": "the broccoli on the right"}]}
{"id": 541706, "image": "COCO_train2014_000000541706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli in some paper below a sign\"."}], "task": "refering", "answer_template": "The \"broccoli in some paper below a sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 138, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194], "bbox": [0.48944, 0.55656, 0.7992999999999999, 0.84296], "iscrowd": 0, "area": 12209.878249999998, "rle": {"size": [375, 500], "counts": "Xni23c;3M4L3M3M3M3M4L3M3N2jERO^9Q1WFZOf9_1L4L4L4L4M3M3L3N1O2N1O2N1N3N1O2N1O2N1O0O10O1000000O0100000O1000O01000000O1000000O1000000O1000000000000000000000000000000000000000000000001O0O1000001O0000001O000000001O000000001O00001N2O1O001O1O001N2O1O001O1O001N2O001O3M2M3N3M2N3M2M3N3M1O1O0O2O001O001N101O001O0O2O0O1O2M2O2M2O3L8I7H7JnfT1"}, "label": "broccoli in some paper below a sign"}]}
{"id": 173073, "image": "COCO_train2014_000000173073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black cow\"."}], "task": "refering", "answer_template": "The \"black cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 235, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 331, 334, 335, 345, 346, 347], "bbox": [0.092, 0.23165624999999998, 0.8628958333333332, 0.895703125], "iscrowd": 0, "area": 69927.78035000002, "rle": {"size": [640, 480], "counts": "gdk0=cc03M2N2N2N2N2N1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O100O100O100O2O0O100O100O01O0010O01O001O00W_ObNQ?^1k@gNS?Y1j@lNU?S1h@SOU?m0i@VOW?i0f@ZOZ?g0b@\\O_?c0_@_Oa?a0[@Cf?<V@Hj?8S@Kn?4n_O0S`00i_O3X`0Ld_O:[`0Eb_O?^`0@^_Of0b`0XOZ_On0f`0V110M2N3M2M4M;Ef0YO3N1O1N2O1O100O2O1N2O1N2O1N2O1N101N2O1N2O1O1N2O1N2O001N2O1N2O1O1N2O1N101O1mHmI6U6DPJ<P6@TJ`0m5ZOXJg0h5YMjI_Mc0X5d5XMmIWMd0a5_5]KfI_O;YOf0k5Z5[KhI_O<mNl0Y6Q5YKiI@=`NT1g6f4WKmI@W2Y5m3UKnIVN:=j1X6o3SKnJdNR1Z6Q4PKPKdNo0]6P4nJTKcNl0_6Q4kJUKfNj0_6R4iJUKhNi0_6S4gJUKjNh0_6S4fJVKkNg0_6T4dJVKmNf0_6U4aJWKPOd0_6U4`JXKQOc0`6U4]JYKSOb0`6V4[JYKUOa0`6W4XJZKXO?`6W4WJ[KYO=a6Y4TJ[K[O;c6Z4PJ\\K]O:e6Y4lI^K_O8g6Y4hIaKA5i6Z4dIbKC3l6Z4_IdKE1n6Z4\\IfKFOP7[4VIiKKKQ7[4PImKOGS7\\4iHQL4BU7\\4dHUL7^OW7]4^HXL;ZOZ7]4WH\\L?VO\\7]7cHaH1QNl6]9SI]H4XNk6Y9QIYH7aNj6T9oHVH:hNh6Q9nHRH=oNg6m8lHnG`0XOf6h8jHkGc0_Od6e8iHfGg0Gb6P1gH_50jIe0If6l0fH`5OjIc0Mj6f0eHc5NjIa00l6b0fHc5MkI?2o6>gHd5KlI<5S7:gHd5JmI:7W75gHf5HmI9:Y72fHg5HmI7<Z70iHf5GmI3`0[7OlHc5FnI1b0\\7MoHb5DoIOd0\\7LSI`5BoINg0[7LUI^5BoIKk0[7IZI\\5@PJIm0\\7H\\IZ5_OQJGo0\\7G`IX5]ORJDR1_7DaIW5\\ORJCU1`7AbIX5[ORJAW1a7@dIV5ZOSJ_OY1c7]OfIV5XOSLQ7hNhIT5WOSLR7hNiIT5UOTLR7hNjIS5TOULQ7hNlIS5SOTLR7hNmIS5QOULR7hNnIR5POVLQ7hNQJQ5nNVLR7iNPJR5mNULS7hNRJR5kNULT7hNSJR5iNVLT7hNSJR5jNULS7hNUJR5hNSLV7jNTJR5fNRLX7lNRJR5fNPLY7nNSJQ5dNnK\\7QOQJP5cNmK^7ROPJQ5bNkK`7SOPJQ5`NiKc7VOnIP5_NhKe7WOmIQ5^NfKg7XOmIQ5\\NdKj7[OkIP5[NcKl7\\OjIQ5ZNaKo7\\OiIR5XN_KR8^OhIR5VN^KU8^OfIT5UN\\KW8@eIS5TNZK[8AcIT5RNYK]8BbIU5QNWK`8BaIV5oMUKc8D`IV5mMTKf8D^IX5lMRKh8F]IW5kMPKl8GZIY5kMmJm8IZIY5iMlJP9IYIZ5gMjJS9KWI[5fMhJV9KVI\\5dMgJ\\9IQI_5cMeJg<[5YCcJj<\\5VCcJl<\\5TCdJl<]5SCcJn<\\5RCeJn<Z5RCfJn<Z5RCfJo<Y5QCgJo<Y5QCgJP=X5PChJP=X5PChJP=X5PCiJP=V5PCjJP=V5PCjJQ=U5oBkJQ=U5oBkJQ=U5PCjJQ=U5oBkJQ=U5oBlJP=T5PClJQ=S5oBmJQ=S5oBmJQ=S5oBmJR=R5nBnJR=R5nBoJR=P5nBPKR=P5nBPKR=P5nBPKS=o4mBQKS=o4mBQKS=o4mBQKT=n4lBRKT=n4lBRKT=n4lBRKU=m4kBSKU=m4kBSKV=l4jBTKV=l4jBTKV=m4jBRKW=m4iBSKW=m4iBSKW=m4iBSKX=l4hBTKX=l4hBTKX=l4hBTKY=k4gBUKY=k4gBUKZ=j4fBVKZ=j4fBVKZ=j4fBVK[=i4eBWK[=i4eBWK[=i4eBWK\\=h4dBXK\\=h4dBXK\\=h4dBXK]=g4cBYK]=g4cBYK^=f4bBYK_=g4aBYK_=g4bBXK_=g4aBYK_=g4aBYK_=g4aBYK_=g4aBYK`=f4`BZK`=f4`BZK`=f4`BZK`=f4`BZKa=e4_B[Ka=f4^BZKb=f4^BZKb=f4^BZKc=f4\\BbK\\=^4dBiKU=W4kBPLn<Q4QCVLi<i3WC^Lb<c3]CdL\\<\\3dCkLU<U3kCSMm;n2RDQMP<n2QDQMo;o2QDPMP<Q3oCnLR<R3nCmLT<R3lCmLU<T3jCkLW<U3iCjLX<W3gChL[<W3eChL\\<X3dCiL[<X3dChL\\<X3dCiL[<W3eCjL[<V3dCkL[<U3eClLZ<T3fCmLY<T3fClL[<S3eCWKI[1a<^3fCSKO\\1\\<a3eCoJ5[1Y<e3TDWLo;i3SDRLP<n3RDfKY<X4jC\\K`<d4^100O10000O10001N10000O100O1O100O2O0O1O100O1O100O101N1O101N2N2O1N3N1N2N2O2M2N2O1N4M2L5L6I6K5J7J4K3N2N2M3XOi0B=E;BclX1"}, "label": "black cow"}]}
{"id": 173073, "image": "COCO_train2014_000000173073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black cow behind the fence\"."}], "task": "refering", "answer_template": "The \"the black cow behind the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 235, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 331, 334, 335, 345, 346, 347], "bbox": [0.092, 0.23165624999999998, 0.8628958333333332, 0.895703125], "iscrowd": 0, "area": 69927.78035000002, "rle": {"size": [640, 480], "counts": "gdk0=cc03M2N2N2N2N2N1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O100O100O100O2O0O100O100O01O0010O01O001O00W_ObNQ?^1k@gNS?Y1j@lNU?S1h@SOU?m0i@VOW?i0f@ZOZ?g0b@\\O_?c0_@_Oa?a0[@Cf?<V@Hj?8S@Kn?4n_O0S`00i_O3X`0Ld_O:[`0Eb_O?^`0@^_Of0b`0XOZ_On0f`0V110M2N3M2M4M;Ef0YO3N1O1N2O1O100O2O1N2O1N2O1N2O1N101N2O1N2O1O1N2O1N2O001N2O1N2O1O1N2O1N101O1mHmI6U6DPJ<P6@TJ`0m5ZOXJg0h5YMjI_Mc0X5d5XMmIWMd0a5_5]KfI_O;YOf0k5Z5[KhI_O<mNl0Y6Q5YKiI@=`NT1g6f4WKmI@W2Y5m3UKnIVN:=j1X6o3SKnJdNR1Z6Q4PKPKdNo0]6P4nJTKcNl0_6Q4kJUKfNj0_6R4iJUKhNi0_6S4gJUKjNh0_6S4fJVKkNg0_6T4dJVKmNf0_6U4aJWKPOd0_6U4`JXKQOc0`6U4]JYKSOb0`6V4[JYKUOa0`6W4XJZKXO?`6W4WJ[KYO=a6Y4TJ[K[O;c6Z4PJ\\K]O:e6Y4lI^K_O8g6Y4hIaKA5i6Z4dIbKC3l6Z4_IdKE1n6Z4\\IfKFOP7[4VIiKKKQ7[4PImKOGS7\\4iHQL4BU7\\4dHUL7^OW7]4^HXL;ZOZ7]4WH\\L?VO\\7]7cHaH1QNl6]9SI]H4XNk6Y9QIYH7aNj6T9oHVH:hNh6Q9nHRH=oNg6m8lHnG`0XOf6h8jHkGc0_Od6e8iHfGg0Gb6P1gH_50jIe0If6l0fH`5OjIc0Mj6f0eHc5NjIa00l6b0fHc5MkI?2o6>gHd5KlI<5S7:gHd5JmI:7W75gHf5HmI9:Y72fHg5HmI7<Z70iHf5GmI3`0[7OlHc5FnI1b0\\7MoHb5DoIOd0\\7LSI`5BoINg0[7LUI^5BoIKk0[7IZI\\5@PJIm0\\7H\\IZ5_OQJGo0\\7G`IX5]ORJDR1_7DaIW5\\ORJCU1`7AbIX5[ORJAW1a7@dIV5ZOSJ_OY1c7]OfIV5XOSLQ7hNhIT5WOSLR7hNiIT5UOTLR7hNjIS5TOULQ7hNlIS5SOTLR7hNmIS5QOULR7hNnIR5POVLQ7hNQJQ5nNVLR7iNPJR5mNULS7hNRJR5kNULT7hNSJR5iNVLT7hNSJR5jNULS7hNUJR5hNSLV7jNTJR5fNRLX7lNRJR5fNPLY7nNSJQ5dNnK\\7QOQJP5cNmK^7ROPJQ5bNkK`7SOPJQ5`NiKc7VOnIP5_NhKe7WOmIQ5^NfKg7XOmIQ5\\NdKj7[OkIP5[NcKl7\\OjIQ5ZNaKo7\\OiIR5XN_KR8^OhIR5VN^KU8^OfIT5UN\\KW8@eIS5TNZK[8AcIT5RNYK]8BbIU5QNWK`8BaIV5oMUKc8D`IV5mMTKf8D^IX5lMRKh8F]IW5kMPKl8GZIY5kMmJm8IZIY5iMlJP9IYIZ5gMjJS9KWI[5fMhJV9KVI\\5dMgJ\\9IQI_5cMeJg<[5YCcJj<\\5VCcJl<\\5TCdJl<]5SCcJn<\\5RCeJn<Z5RCfJn<Z5RCfJo<Y5QCgJo<Y5QCgJP=X5PChJP=X5PChJP=X5PCiJP=V5PCjJP=V5PCjJQ=U5oBkJQ=U5oBkJQ=U5PCjJQ=U5oBkJQ=U5oBlJP=T5PClJQ=S5oBmJQ=S5oBmJQ=S5oBmJR=R5nBnJR=R5nBoJR=P5nBPKR=P5nBPKR=P5nBPKS=o4mBQKS=o4mBQKS=o4mBQKT=n4lBRKT=n4lBRKT=n4lBRKU=m4kBSKU=m4kBSKV=l4jBTKV=l4jBTKV=m4jBRKW=m4iBSKW=m4iBSKW=m4iBSKX=l4hBTKX=l4hBTKX=l4hBTKY=k4gBUKY=k4gBUKZ=j4fBVKZ=j4fBVKZ=j4fBVK[=i4eBWK[=i4eBWK[=i4eBWK\\=h4dBXK\\=h4dBXK\\=h4dBXK]=g4cBYK]=g4cBYK^=f4bBYK_=g4aBYK_=g4bBXK_=g4aBYK_=g4aBYK_=g4aBYK_=g4aBYK`=f4`BZK`=f4`BZK`=f4`BZK`=f4`BZKa=e4_B[Ka=f4^BZKb=f4^BZKb=f4^BZKc=f4\\BbK\\=^4dBiKU=W4kBPLn<Q4QCVLi<i3WC^Lb<c3]CdL\\<\\3dCkLU<U3kCSMm;n2RDQMP<n2QDQMo;o2QDPMP<Q3oCnLR<R3nCmLT<R3lCmLU<T3jCkLW<U3iCjLX<W3gChL[<W3eChL\\<X3dCiL[<X3dChL\\<X3dCiL[<W3eCjL[<V3dCkL[<U3eClLZ<T3fCmLY<T3fClL[<S3eCWKI[1a<^3fCSKO\\1\\<a3eCoJ5[1Y<e3TDWLo;i3SDRLP<n3RDfKY<X4jC\\K`<d4^100O10000O10001N10000O100O1O100O2O0O1O100O1O100O101N1O101N2N2O1N3N1N2N2O2M2N2O1N4M2L5L6I6K5J7J4K3N2N2M3XOi0B=E;BclX1"}, "label": "the black cow behind the fence"}]}
{"id": 517144, "image": "COCO_train2014_000000517144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the male tennis player on the right is starting to run\"."}], "task": "refering", "answer_template": "The \"the male tennis player on the right is starting to run\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 101, 102, 103, 119, 120, 121, 122, 136, 137, 138, 139, 140, 141, 156, 157, 158, 159, 174, 175, 176, 177, 178, 196, 197], "bbox": [0.5581799999999999, 0.3786581469648562, 0.9722, 0.9881469648562301], "iscrowd": 0, "area": 10122.956749999998, "rle": {"size": [313, 500], "counts": "Tie22f92N1O100000001N100000001O000O1000001O00000O10000000O1000O010000O10000O01000O10O1O10O01nM9RJGm5=oICQ6a0lI_OS6e0iI\\OV6f0hIZOX6g0hIYOW6i0gIXOX6i0hIVOX6l0fIUOZ6l0dIUO\\6k0dIUO\\6l0bITO`6k0`IUO`6l0^IUOb6k0^IUO?Bm4Y1bJUO`0Jh4Q1gJVO`0Kh4P1gJUO`0Mi4m0fJWO`0Oh4l0cJWOd00h4j0`JYOg0Oi4i0ZJ\\Ol0Nh4b1WKaNh4_1WKcNi4\\1VKgNh4Y1VKjNi4U1WKmNi4R1UKROi4n0UKUOk4j0SKZOk4e0TKK^45aKL`43^KOb41[K2g4KVK9l4EPK?R5S23M2N2N2N2N1O1N2O1O1O1O1fNZJ_Nh5_1]J\\Nd5b1_J\\Nb5b1`J]Na5a1aJ^N`5`1cJ^N^5a1cJ^N^5`1dJ_N]5_1eJ`N]5_1dJ_N]5`1dJ_N\\5a1eJ^N\\5a1eJ^N\\5b1dJ\\N^5c1dJ[N]5d1dJZN^5f1bJYN_5f1cJXN^5g1eJUN]5k1X110O1000000000001O0001O2N1O2N2O1N1O0000000001O001O3M4L3M4L3L4L3M4L3N3L:F9G1O00010O0000001O01O01O00000010O00011N101O1N101O1N101O1N2O0O3N1O0O01000O10O10O100O01N2O0O2O1N3M2O1N3N1N2O2M2N2N3M2Oeo3"}, "label": "the male tennis player on the right is starting to run"}]}
{"id": 517144, "image": "COCO_train2014_000000517144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male tennis player dressed in white watching his partner hit the ball\"."}], "task": "refering", "answer_template": "The \"a male tennis player dressed in white watching his partner hit the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 101, 102, 103, 119, 120, 121, 122, 136, 137, 138, 139, 140, 141, 156, 157, 158, 159, 174, 175, 176, 177, 178, 196, 197], "bbox": [0.5581799999999999, 0.3786581469648562, 0.9722, 0.9881469648562301], "iscrowd": 0, "area": 10122.956749999998, "rle": {"size": [313, 500], "counts": "Tie22f92N1O100000001N100000001O000O1000001O00000O10000000O1000O010000O10000O01000O10O1O10O01nM9RJGm5=oICQ6a0lI_OS6e0iI\\OV6f0hIZOX6g0hIYOW6i0gIXOX6i0hIVOX6l0fIUOZ6l0dIUO\\6k0dIUO\\6l0bITO`6k0`IUO`6l0^IUOb6k0^IUO?Bm4Y1bJUO`0Jh4Q1gJVO`0Kh4P1gJUO`0Mi4m0fJWO`0Oh4l0cJWOd00h4j0`JYOg0Oi4i0ZJ\\Ol0Nh4b1WKaNh4_1WKcNi4\\1VKgNh4Y1VKjNi4U1WKmNi4R1UKROi4n0UKUOk4j0SKZOk4e0TKK^45aKL`43^KOb41[K2g4KVK9l4EPK?R5S23M2N2N2N2N1O1N2O1O1O1O1fNZJ_Nh5_1]J\\Nd5b1_J\\Nb5b1`J]Na5a1aJ^N`5`1cJ^N^5a1cJ^N^5`1dJ_N]5_1eJ`N]5_1dJ_N]5`1dJ_N\\5a1eJ^N\\5a1eJ^N\\5b1dJ\\N^5c1dJ[N]5d1dJZN^5f1bJYN_5f1cJXN^5g1eJUN]5k1X110O1000000000001O0001O2N1O2N2O1N1O0000000001O001O3M4L3M4L3L4L3M4L3N3L:F9G1O00010O0000001O01O01O00000010O00011N101O1N101O1N101O1N2O0O3N1O0O01000O10O10O100O01N2O0O2O1N3M2O1N3N1N2O2M2N2N3M2Oeo3"}, "label": "a male tennis player dressed in white watching his partner hit the ball"}]}
{"id": 238618, "image": "COCO_train2014_000000238618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow ' s ear with a circular tag\"."}], "task": "refering", "answer_template": "The \"a cow ' s ear with a circular tag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 45, 46, 47, 48, 49, 50, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 90, 91, 92, 93, 105, 106, 107, 120, 121, 122, 123, 124, 135, 136, 137, 138, 139, 140, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 197, 198, 199, 200, 214, 215], "bbox": [0.0, 0.12045312500000001, 0.3695560747663551, 0.6383281249999999], "iscrowd": 0, "area": 36465.75424999998, "rle": {"size": [640, 428], "counts": "i2e1[b0Y3gLZ2fM01O0O100O10001N100O10000O101N100O10000O2O0O10O10O100O0d0]O5K3M3L4M3M4L3L3N1O001O1N101O1O0O2O1O001N2O001O1O0O2O1O000O1000000O1000000O100000000O101O00`J]Ed1a:\\NQFS1o9mNWFm0i9ROXFo0g9POZFP1e9PO\\FQ1c9nN^FR1b9mN_FT1`9lN`FT1`9kNaFU1^9kNcFV1\\9iNeFW1[9hNfFY1Y9fNhFZ1W9fNjF[1U9eNkF[1U9dNlF\\1T9cNmF^1S9`NoF_1Q9`NPGa1P9^NPGb1P9]NQGd1n8\\NRGd1o8ZNRGf1n8YNSGh1l8XNTGh1m8VNTGk1k8UNUGk1l8SNUGn1j8QNWGo1i8QNWGP2i8nMYGQ2g8nMZGS2f8lMZGT2f8kM\\GU2c8kM]GV2c8hM^GX2b8gM`GY2_8gMaGY2`8eMaG\\2^8dMcG[2^8cMcG^2\\8aMeG_2[8aMeG`2[8^MgGa2Y8]MiGd2V8[MkGe2V8XMmGh2R8WMoGi2R8UMoGl2P8RMSHm2m7RMTHo2l7nLVHR3j7mLXHS3h7kLYHV3f7hL]HW3c7hL`HW3`7fLcHY3]7fLeHZ3Z7eLiHY3X7dLkH\\3U7bLmH]3T7`LPI_3o6`LSI_3n6_LTIa3l6\\LWIc3j6ZLZIe3f6WL`If3a6WLdIg3\\6ULjIh3V6ULQJh3o5ULYJe3h5WL`Je3`5YLgJb3Y5[LnJ`3S5^LTK]3l4aLZKZ3l:K4K7_Oc0QOo0SOT^X5"}, "label": "a cow ' s ear with a circular tag"}]}
{"id": 238618, "image": "COCO_train2014_000000238618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an all black cow\"."}], "task": "refering", "answer_template": "The \"an all black cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 45, 46, 47, 48, 49, 50, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 90, 91, 92, 93, 105, 106, 107, 120, 121, 122, 123, 124, 135, 136, 137, 138, 139, 140, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 197, 198, 199, 200, 214, 215], "bbox": [0.0, 0.12045312500000001, 0.3695560747663551, 0.6383281249999999], "iscrowd": 0, "area": 36465.75424999998, "rle": {"size": [640, 428], "counts": "i2e1[b0Y3gLZ2fM01O0O100O10001N100O10000O101N100O10000O2O0O10O10O100O0d0]O5K3M3L4M3M4L3L3N1O001O1N101O1O0O2O1O001N2O001O1O0O2O1O000O1000000O1000000O100000000O101O00`J]Ed1a:\\NQFS1o9mNWFm0i9ROXFo0g9POZFP1e9PO\\FQ1c9nN^FR1b9mN_FT1`9lN`FT1`9kNaFU1^9kNcFV1\\9iNeFW1[9hNfFY1Y9fNhFZ1W9fNjF[1U9eNkF[1U9dNlF\\1T9cNmF^1S9`NoF_1Q9`NPGa1P9^NPGb1P9]NQGd1n8\\NRGd1o8ZNRGf1n8YNSGh1l8XNTGh1m8VNTGk1k8UNUGk1l8SNUGn1j8QNWGo1i8QNWGP2i8nMYGQ2g8nMZGS2f8lMZGT2f8kM\\GU2c8kM]GV2c8hM^GX2b8gM`GY2_8gMaGY2`8eMaG\\2^8dMcG[2^8cMcG^2\\8aMeG_2[8aMeG`2[8^MgGa2Y8]MiGd2V8[MkGe2V8XMmGh2R8WMoGi2R8UMoGl2P8RMSHm2m7RMTHo2l7nLVHR3j7mLXHS3h7kLYHV3f7hL]HW3c7hL`HW3`7fLcHY3]7fLeHZ3Z7eLiHY3X7dLkH\\3U7bLmH]3T7`LPI_3o6`LSI_3n6_LTIa3l6\\LWIc3j6ZLZIe3f6WL`If3a6WLdIg3\\6ULjIh3V6ULQJh3o5ULYJe3h5WL`Je3`5YLgJb3Y5[LnJ`3S5^LTK]3l4aLZKZ3l:K4K7_Oc0QOo0SOT^X5"}, "label": "an all black cow"}]}
{"id": 238618, "image": "COCO_train2014_000000238618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cow\"."}], "task": "refering", "answer_template": "The \"a black and white cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 304, 305, 306, 307, 308, 309, 311, 312, 313, 314, 321, 322, 327, 328, 329, 344], "bbox": [0.1780841121495327, 0.1016875, 1.0, 0.9690937500000001], "iscrowd": 0, "area": 142399.8842, "rle": {"size": [640, 428], "counts": "Tf_1<bc0d0]O3M1O1O1O1O1O2N1O1O1O1N2O1O2N100O100O1O100O2N100O1O100O101N1O100O1O100O2N100O100O1bFkMP1V2mNQNm0Q2QOUNj0l1SOZNh0i1UO]Nf0d1XOaNd0a1XOfNb0]1[OhNa0Z1\\OlN?V1]OPO>S1_OSO<n0AYO:i0C\\O8g0E_O6c0FC6?GG3;JI49IJ58GL67FL96DM:5BO;5@N?7[OLb08YOLd08XOJe0:WOJf0;TOIh0`0POBm0f0kN^OQ1k0fNYOV1P1bNRO\\1U1\\NoNb1X1UNlNh1\\1PNfNo1`1iMdNU2c1bMaN[2g1ZM^Ne2h1]LROa3U1_KF^4o7O2M3N1O2M3N1N3N1O2M3N1N3N2N1N3N2N101O1N101O1O001O1O001O1O0O2O00O10000O1000000O01000O10000O1000000O01000O10000O1000000O01000O10000O100000O010000O10000O100000O010000O1000000O1000O010000O1000000O100O001O1O1O1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O1O2N1O1N2O2N1O1O1O2N1O2M3N2N3M2N2N2M3G9G9G9G:H7J6J6K5J6J6K6I6J6K5J6J5L4K6h2WMd0\\O3N0O1O100O1O100O1O100O1O1O100O1O100O1O1N2O1O1O1N2O1O1O1N101O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2O1O001N2O1O001N2O1O1O0O2O1O1O001N2O1O001N2O1O1O0O2O1O1O0O2O1O1O0O2O1O1O1N101O1O0O1000O10O1000O1000O10O1000O1000O107Ia2_Ma2^Ma2`Ma2_M`2`Ma2^MS^O"}, "label": "a black and white cow"}]}
{"id": 238618, "image": "COCO_train2014_000000238618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow has a black and white face\"."}], "task": "refering", "answer_template": "The \"the cow has a black and white face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 304, 305, 306, 307, 308, 309, 311, 312, 313, 314, 321, 322, 327, 328, 329, 344], "bbox": [0.1780841121495327, 0.1016875, 1.0, 0.9690937500000001], "iscrowd": 0, "area": 142399.8842, "rle": {"size": [640, 428], "counts": "Tf_1<bc0d0]O3M1O1O1O1O1O2N1O1O1O1N2O1O2N100O100O1O100O2N100O1O100O101N1O100O1O100O2N100O100O1bFkMP1V2mNQNm0Q2QOUNj0l1SOZNh0i1UO]Nf0d1XOaNd0a1XOfNb0]1[OhNa0Z1\\OlN?V1]OPO>S1_OSO<n0AYO:i0C\\O8g0E_O6c0FC6?GG3;JI49IJ58GL67FL96DM:5BO;5@N?7[OLb08YOLd08XOJe0:WOJf0;TOIh0`0POBm0f0kN^OQ1k0fNYOV1P1bNRO\\1U1\\NoNb1X1UNlNh1\\1PNfNo1`1iMdNU2c1bMaN[2g1ZM^Ne2h1]LROa3U1_KF^4o7O2M3N1O2M3N1N3N1O2M3N1N3N2N1N3N2N101O1N101O1O001O1O001O1O0O2O00O10000O1000000O01000O10000O1000000O01000O10000O1000000O01000O10000O100000O010000O10000O100000O010000O1000000O1000O010000O1000000O100O001O1O1O1O1N2O2N1O1O1O2N1O1N2O2N1O1O1O1O2N1O1N2O2N1O1O1O2N1O2M3N2N3M2N2N2M3G9G9G9G:H7J6J6K5J6J6K6I6J6K5J6J5L4K6h2WMd0\\O3N0O1O100O1O100O1O100O1O1O100O1O100O1O1N2O1O1O1N2O1O1O1N101O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2O1O001N2O1O001N2O1O1O0O2O1O1O001N2O1O001N2O1O1O0O2O1O1O0O2O1O1O0O2O1O1O1N101O1O0O1000O10O1000O1000O10O1000O1000O107Ia2_Ma2^Ma2`Ma2_M`2`Ma2^MS^O"}, "label": "the cow has a black and white face"}]}
{"id": 345114, "image": "COCO_train2014_000000345114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suitcase in the reflection\"."}], "task": "refering", "answer_template": "The \"the suitcase in the reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 159, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 235, 249, 250, 251, 252], "bbox": [0.38333333333333336, 0.33490625, 0.8794791666666666, 0.63890625], "iscrowd": 0, "area": 28488.902599999998, "rle": {"size": [640, 480], "counts": "gVc33oc04K8Il0TOk0TOm0TOh0WOa0@N2O0O2N2N101N1O2N2N101N000000010O00000001O01O0001O000000010O0000000010O00000001O0001O01O000000010O000000001O01O000001O0001O01O000000010O000000001O01O000001O0001O01O0000000010O00000001O01O0001O000001O01O0000000010O00000001O01O0001O000000010O0000000010O00000001O0001O01O000000010O0000000010O00000001O0001O01O000000010O000000001O01O000001O0001O01O0000000010O00000001O01O000001O0001O01O0000000010O00000001O01N1M3N2M3K5K5J7J5K5K5K5J6K6J5M3N2M3N2M4M2M3N2K7I8G8I8H7IRoS1"}, "label": "the suitcase in the reflection"}]}
{"id": 197663, "image": "COCO_train2014_000000197663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding wine glass\"."}], "task": "refering", "answer_template": "The \"woman holding wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 216, 217, 230, 231, 232, 233, 234, 235, 236, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.12122065727699531, 0.45565625, 1.0], "iscrowd": 0, "area": 66435.41720000001, "rle": {"size": [426, 640], "counts": "o;W1d9_2hNX1O1O1O1O1N2O1O1\\MZKcLg4]3hKRLZ4m3iKPLX4P4iKnKX4R4jKjKX4U4jKeK[4[4gK_K]4a4dKYKa4f4bKTKb4l4bKnJ`4R5fKgJ[4X5l100O10000O10000O1mN_H]La7[3PI[LQ7]3b1H8H8J6N2N2O1N2N2N2O1N2N2N101N2O100O100O10000O100O100O100O100O100O10000O100O100O100O10O010000O100O10000O1000000O10000000000000000000000000000000000O100000000000000000000000001O00000001O1cJVNXOj1f0YNYOh1e0ZNZOf1d0]N[Od1c0^N\\Ob1b0`N^Oa1`0aN_O_1`0cN_O^1>eNA\\1>eNA[1>fNB[1=fNBZ1=hNBY1<iNDV1<kNCV1;kNEV1:kNEU1:mNET19nNFR1:oNER19POFP1:POFQ18QOGP18QOGo08SOGn07TOHl07VOHk05YOIh05[OIf05]OIc06@Ha06AJ?4DJ=3GK94JJ74LJ54NJ34OK231KO35KL37KJ39KH3;KE4>JC4`0JA3c0K_O2d0L_OOe0O^OLf02]OHh07YOEk09XOBl0<WO_Om0?VO[Oo0c0TOXOP1f0ROUOS1i0POROT1l0oNnNW1o0mNjNY1S1lNfNX1X1mN`NY1]1lN\\NY1a1lNWNY1g1lNRNY1k1lNmMZ1P2kNiMY1U2W5G8I8G2N2N2N1OO1O1O1O1N2O1O1O1O1O1O1O1O2N1cN]1K5N2N2N2N2N2N2N101O010O00010O010O010O00010O010O0010N101O001O000O2O001O001O0O101O001O000O2O0O10000O10000O10000O2O0O10000O2O2M2O2M3M2O2M2N3N1N3M2O2M6J8I8C=A>B?AT_`4"}, "label": "woman holding wine glass"}]}
{"id": 197663, "image": "COCO_train2014_000000197663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding an almost empty wine glass and wearing sunglasses\"."}], "task": "refering", "answer_template": "The \"the woman holding an almost empty wine glass and wearing sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 216, 217, 230, 231, 232, 233, 234, 235, 236, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.12122065727699531, 0.45565625, 1.0], "iscrowd": 0, "area": 66435.41720000001, "rle": {"size": [426, 640], "counts": "o;W1d9_2hNX1O1O1O1O1N2O1O1\\MZKcLg4]3hKRLZ4m3iKPLX4P4iKnKX4R4jKjKX4U4jKeK[4[4gK_K]4a4dKYKa4f4bKTKb4l4bKnJ`4R5fKgJ[4X5l100O10000O10000O1mN_H]La7[3PI[LQ7]3b1H8H8J6N2N2O1N2N2N2O1N2N2N101N2O100O100O10000O100O100O100O100O100O10000O100O100O100O10O010000O100O10000O1000000O10000000000000000000000000000000000O100000000000000000000000001O00000001O1cJVNXOj1f0YNYOh1e0ZNZOf1d0]N[Od1c0^N\\Ob1b0`N^Oa1`0aN_O_1`0cN_O^1>eNA\\1>eNA[1>fNB[1=fNBZ1=hNBY1<iNDV1<kNCV1;kNEV1:kNEU1:mNET19nNFR1:oNER19POFP1:POFQ18QOGP18QOGo08SOGn07TOHl07VOHk05YOIh05[OIf05]OIc06@Ha06AJ?4DJ=3GK94JJ74LJ54NJ34OK231KO35KL37KJ39KH3;KE4>JC4`0JA3c0K_O2d0L_OOe0O^OLf02]OHh07YOEk09XOBl0<WO_Om0?VO[Oo0c0TOXOP1f0ROUOS1i0POROT1l0oNnNW1o0mNjNY1S1lNfNX1X1mN`NY1]1lN\\NY1a1lNWNY1g1lNRNY1k1lNmMZ1P2kNiMY1U2W5G8I8G2N2N2N1OO1O1O1O1N2O1O1O1O1O1O1O1O2N1cN]1K5N2N2N2N2N2N2N101O010O00010O010O010O00010O010O0010N101O001O000O2O001O001O0O101O001O000O2O0O10000O10000O10000O2O0O10000O2O2M2O2M3M2O2M2N3N1N3M2O2M6J8I8C=A>B?AT_`4"}, "label": "the woman holding an almost empty wine glass and wearing sunglasses"}]}
{"id": 197663, "image": "COCO_train2014_000000197663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass with little vodka in it , is holding by a beautiful lady\"."}], "task": "refering", "answer_template": "The \"a wine glass with little vodka in it , is holding by a beautiful lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 263, 264, 286], "bbox": [0.324578125, 0.4195305164319249, 0.504078125, 0.8150469483568076], "iscrowd": 0, "area": 5968.403999999997, "rle": {"size": [426, 640], "counts": "Tgf23V=6K4L5J5L5J6K4L5J5L5J5L5K5J3N1O1N2O1N101O1N2O1O1N101N2O1O1N100O1000000O10000O10000O10000O100001O1O2N101O00001M2O2M2N3M2O2M2N3M2O2EnDYNT;e1PFZNP9c1PG_NP9_1QGcNo8Z1QGhNo8V1RGiNQ9T1oFkNU9R1kFnNW9P1jFnNZ9o0fFQO\\9m0eFRO^9k0bFTOb9h0oGZOe6=\\IFi60YI3m92O1N2O1N2O1N2O001O01O100O100O1O100O100O1O100O100O1O100O100O1O100O1M[oS4"}, "label": "a wine glass with little vodka in it , is holding by a beautiful lady"}]}
{"id": 197663, "image": "COCO_train2014_000000197663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine class with wine holding a lady\"."}], "task": "refering", "answer_template": "The \"wine class with wine holding a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 263, 264, 286], "bbox": [0.324578125, 0.4195305164319249, 0.504078125, 0.8150469483568076], "iscrowd": 0, "area": 5968.403999999997, "rle": {"size": [426, 640], "counts": "Tgf23V=6K4L5J5L5J6K4L5J5L5J5L5K5J3N1O1N2O1N101O1N2O1O1N101N2O1O1N100O1000000O10000O10000O10000O100001O1O2N101O00001M2O2M2N3M2O2M2N3M2O2EnDYNT;e1PFZNP9c1PG_NP9_1QGcNo8Z1QGhNo8V1RGiNQ9T1oFkNU9R1kFnNW9P1jFnNZ9o0fFQO\\9m0eFRO^9k0bFTOb9h0oGZOe6=\\IFi60YI3m92O1N2O1N2O1N2O001O01O100O100O1O100O100O1O100O100O1O100O100O1O100O1M[oS4"}, "label": "wine class with wine holding a lady"}]}
{"id": 197663, "image": "COCO_train2014_000000197663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man toasting\"."}], "task": "refering", "answer_template": "The \"man toasting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.3754375, 0.06896713615023474, 0.913921875, 0.9858215962441315], "iscrowd": 0, "area": 82204.48580000001, "rle": {"size": [426, 640], "counts": "kST33U=:F7I1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1000001O000000001O000000001jHcN^2]1ZMlNd2U1TMTOk2l0nL\\OQ3d0gLEX3=^LLa39RL0m35hK2W44_K1<hMc1\\2hM1d0gM`1]2dM0k0fM^1_2^M0S1eM[1]6cNgIZ1Z6eNjIW1W6hNlIU1U6jNoIR1R6mNRJo0o5POTJn0l5QOXJl0h5RO\\Jk0e5TO_Ji0a5VObJh0^5WOfJe0[5ZOhJd0X5[OlJa0U5]OPK`0P5_OSK?m4@WK<j4CYK;g4D]K8c4H`K6`4IdK4\\4JiK2X4MlK0T4OoKOQ40SLLn33VLJj36YLFh39\\LDd3;_LCa3<cL@^3`0eL]O[3b0iLZOX3e0lLXOT3h0nLVOR3i0RMSOo2l0UMPOl2P1WMlNj2S1ZMiNg2V1g4O100O1O100O10000000000000000000000000000000000000000000000001O000000001O0000000O101O00000000001O00000000001O0000002N2N3M2N2N3M2N3M2N2N3M2N3M2N2N2N1O001O1O001O1O001O3M4L3M4L3M4gFlLT8W3eGnLZ8V3^GPM`8S3YGRMf8k3N1O2N1O2N1O2N1O2N1N3N2N000000001O00000000001O001O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O000000001O0000000000000000001O0000000000000000001O000O100000000000001O00001O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N2N2N2M3N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N3M4L4L3M4L4L4L4L3M4L4L4L4L3M4K5L4L4L4L4L4L4L:Fa0_Oa0_Oa_f0"}, "label": "man toasting"}]}
{"id": 197663, "image": "COCO_train2014_000000197663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a glass\"."}], "task": "refering", "answer_template": "The \"a man holding a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.3754375, 0.06896713615023474, 0.913921875, 0.9858215962441315], "iscrowd": 0, "area": 82204.48580000001, "rle": {"size": [426, 640], "counts": "kST33U=:F7I1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1000001O000000001O000000001jHcN^2]1ZMlNd2U1TMTOk2l0nL\\OQ3d0gLEX3=^LLa39RL0m35hK2W44_K1<hMc1\\2hM1d0gM`1]2dM0k0fM^1_2^M0S1eM[1]6cNgIZ1Z6eNjIW1W6hNlIU1U6jNoIR1R6mNRJo0o5POTJn0l5QOXJl0h5RO\\Jk0e5TO_Ji0a5VObJh0^5WOfJe0[5ZOhJd0X5[OlJa0U5]OPK`0P5_OSK?m4@WK<j4CYK;g4D]K8c4H`K6`4IdK4\\4JiK2X4MlK0T4OoKOQ40SLLn33VLJj36YLFh39\\LDd3;_LCa3<cL@^3`0eL]O[3b0iLZOX3e0lLXOT3h0nLVOR3i0RMSOo2l0UMPOl2P1WMlNj2S1ZMiNg2V1g4O100O1O100O10000000000000000000000000000000000000000000000001O000000001O0000000O101O00000000001O00000000001O0000002N2N3M2N2N3M2N3M2N2N3M2N3M2N2N2N1O001O1O001O1O001O3M4L3M4L3M4gFlLT8W3eGnLZ8V3^GPM`8S3YGRMf8k3N1O2N1O2N1O2N1O2N1N3N2N000000001O00000000001O001O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O000000001O0000000000000000001O0000000000000000001O000O100000000000001O00001O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N2N2N2M3N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N3M4L4L3M4L4L4L4L3M4L4L4L4L3M4K5L4L4L4L4L4L4L:Fa0_Oa0_Oa_f0"}, "label": "a man holding a glass"}]}
{"id": 140320, "image": "COCO_train2014_000000140320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch with red , beige and black pillows\"."}], "task": "refering", "answer_template": "The \"a couch with red , beige and black pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 139, 140, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 232, 233], "bbox": [0.7370599999999999, 0.514, 0.9989199999999999, 0.986], "iscrowd": 0, "area": 13543.672549999994, "rle": {"size": [375, 500], "counts": "l\\W49n9JUGa0f8FmFd0o8CeFg0V9o001O1O001N2O001O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1OiMZG^1d8`N_Ga1`8\\NbGf1]8WNfGj1Y8SNjGn1U8oMnGR2Q8kMRHV2m7gMWHY2c800100O100O100O100O1O100O10O0100O100O1O100O100O100O1O001O1O100O1O1O1O1O100O1O1O1O001O100O1O1O1O1O1O1_NeLeJ[3Y5gLfJZ3X5gLhJZ3V5hLiJX3m4SMRKn2l4TMSKm2l4SMUKm2j4TMTKn2l4RMRKo2n4RMPKn2Q5RMoJn2Q5SMmJn2T5RMjJo2W5QMgJP3Z5oLeJR3\\5nLbJS3_5lLaJT3`5lL^JT3c5lL\\JU3d5lLXJW3i5hLUJZ3k5gLQJ\\3o5dLoI^3Q6cLkI`3V6_LhIc3X6^LeIe3Z6[LcIh3]6i01O0000001O0000001M2N2N2L5YOf0ZOf0gLdHa1[="}, "label": "a couch with red , beige and black pillows"}]}
{"id": 140320, "image": "COCO_train2014_000000140320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey couch with pillows sits in a living room\"."}], "task": "refering", "answer_template": "The \"a grey couch with pillows sits in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 139, 140, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 232, 233], "bbox": [0.7370599999999999, 0.514, 0.9989199999999999, 0.986], "iscrowd": 0, "area": 13543.672549999994, "rle": {"size": [375, 500], "counts": "l\\W49n9JUGa0f8FmFd0o8CeFg0V9o001O1O001N2O001O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1OiMZG^1d8`N_Ga1`8\\NbGf1]8WNfGj1Y8SNjGn1U8oMnGR2Q8kMRHV2m7gMWHY2c800100O100O100O100O1O100O10O0100O100O1O100O100O100O1O001O1O100O1O1O1O1O100O1O1O1O001O100O1O1O1O1O1O1_NeLeJ[3Y5gLfJZ3X5gLhJZ3V5hLiJX3m4SMRKn2l4TMSKm2l4SMUKm2j4TMTKn2l4RMRKo2n4RMPKn2Q5RMoJn2Q5SMmJn2T5RMjJo2W5QMgJP3Z5oLeJR3\\5nLbJS3_5lLaJT3`5lL^JT3c5lL\\JU3d5lLXJW3i5hLUJZ3k5gLQJ\\3o5dLoI^3Q6cLkI`3V6_LhIc3X6^LeIe3Z6[LcIh3]6i01O0000001O0000001M2N2N2L5YOf0ZOf0gLdHa1[="}, "label": "a grey couch with pillows sits in a living room"}]}
{"id": 140320, "image": "COCO_train2014_000000140320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch on the bottom left corner\"."}], "task": "refering", "answer_template": "The \"the couch on the bottom left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 186, 198, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.0014, 0.7622933333333334, 0.38198, 0.9994666666666667], "iscrowd": 0, "area": 12951.231950000005, "rle": {"size": [375, 500], "counts": "Re0[2a9K5K5K5K5K5K5K5K1O2N1O2N1O1O2N1OPNYORIe0n6]OQIb0o6_OQI`0o6BQI<o6EQI:o6HQI6o6KQI4o6MPI3P7MPI3P7NoH2Q7OoH0Q70oH0Q71nHOR71nHOR72mHNS73lHMT73lHMT74kHLU74kHLU75jHKV76iHJW76iHJW77hHIX77hHIX78gHHY79fHG[78eHH[79dHG\\79dHG\\7:cHF]7:cHF]7;bHE^7<aHD_7<aHD_7=`HC`7>_HBa7`0]H@c7a0\\H_Od7b0[H^Oe7d0YH\\Og7e0XH[Oh7f0XHYOh7i0VHWOj7j0UHVOk7l0SHTOm7m0RHSOn7n0QHROo7P1oGPOQ8Q1nGoNR8R1mGnNS8T1kGlNU8U1jGkNV8W1hGiNX8X1gGhNY8Y1fGgNZ8[1dGeN\\8\\1cGdN]8]1bGcN^8_1`GaN`8`1_G`Na8b1]G^Nc8c1\\G]Nd8e1ZG[Nf8g1XGYNh8i1VGWNj8j1UGVNk8l1SGTNm8n1QGRNo8[200000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O2N1O2N1O2N^1bN]j`3"}, "label": "the couch on the bottom left corner"}]}
{"id": 140320, "image": "COCO_train2014_000000140320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"that is a comfortable chair in a living room with other furniture\"."}], "task": "refering", "answer_template": "The \"that is a comfortable chair in a living room with other furniture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 186, 198, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.0014, 0.7622933333333334, 0.38198, 0.9994666666666667], "iscrowd": 0, "area": 12951.231950000005, "rle": {"size": [375, 500], "counts": "Re0[2a9K5K5K5K5K5K5K5K1O2N1O2N1O1O2N1OPNYORIe0n6]OQIb0o6_OQI`0o6BQI<o6EQI:o6HQI6o6KQI4o6MPI3P7MPI3P7NoH2Q7OoH0Q70oH0Q71nHOR71nHOR72mHNS73lHMT73lHMT74kHLU74kHLU75jHKV76iHJW76iHJW77hHIX77hHIX78gHHY79fHG[78eHH[79dHG\\79dHG\\7:cHF]7:cHF]7;bHE^7<aHD_7<aHD_7=`HC`7>_HBa7`0]H@c7a0\\H_Od7b0[H^Oe7d0YH\\Og7e0XH[Oh7f0XHYOh7i0VHWOj7j0UHVOk7l0SHTOm7m0RHSOn7n0QHROo7P1oGPOQ8Q1nGoNR8R1mGnNS8T1kGlNU8U1jGkNV8W1hGiNX8X1gGhNY8Y1fGgNZ8[1dGeN\\8\\1cGdN]8]1bGcN^8_1`GaN`8`1_G`Na8b1]G^Nc8c1\\G]Nd8e1ZG[Nf8g1XGYNh8i1VGWNj8j1UGVNk8l1SGTNm8n1QGRNo8[200000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O2N1O2N1O2N^1bN]j`3"}, "label": "that is a comfortable chair in a living room with other furniture"}]}
{"id": 574497, "image": "COCO_train2014_000000574497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboard with a spiral design\"."}], "task": "refering", "answer_template": "The \"a snowboard with a spiral design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 219, 220, 221, 243, 244, 245, 268, 269, 270, 292, 293, 294, 316, 317], "bbox": [0.540109375, 0.5904066985645933, 0.80759375, 0.9108373205741627], "iscrowd": 0, "area": 4715.449099999998, "rle": {"size": [418, 640], "counts": "`_]45l<4M1N2O0O2O1N101N2O001O1O001O1O001O001O10O01O1O1O00100O10O10O1000O1000O10O100000O0100000000O0100000O10O1000O100000O01000000O0100000O1000O1000O1000O10O1000000O0100000O10O1000O10O1000O01000000O10O10O10000O0100000O1000O010000O10O1000O10000O0100000O10O10O10000O10O1000O1000O010000O1000000O01000O10000O1000O010O10O01O010O010O1O2O0N3N1N2O2M2N5K8EYka1"}, "label": "a snowboard with a spiral design"}]}
{"id": 574497, "image": "COCO_train2014_000000574497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow and red snowboard\"."}], "task": "refering", "answer_template": "The \"a yellow and red snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 219, 220, 221, 243, 244, 245, 268, 269, 270, 292, 293, 294, 316, 317], "bbox": [0.540109375, 0.5904066985645933, 0.80759375, 0.9108373205741627], "iscrowd": 0, "area": 4715.449099999998, "rle": {"size": [418, 640], "counts": "`_]45l<4M1N2O0O2O1N101N2O001O1O001O1O001O001O10O01O1O1O00100O10O10O1000O1000O10O100000O0100000000O0100000O10O1000O100000O01000000O0100000O1000O1000O1000O10O1000000O0100000O10O1000O10O1000O01000000O10O10O10000O0100000O1000O010000O10O1000O10000O0100000O10O10O10000O10O1000O1000O010000O1000000O01000O10000O1000O010O10O01O010O010O1O2O0N3N1N2O2M2N5K8EYka1"}, "label": "a yellow and red snowboard"}]}
{"id": 66593, "image": "COCO_train2014_000000066593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue umbrella held by woman in striped shirt\"."}], "task": "refering", "answer_template": "The \"blue umbrella held by woman in striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 101, 102, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 125, 127, 128, 129, 130, 131, 158, 173], "bbox": [0.26824355971896957, 0.235203125, 0.888056206088993, 0.513625], "iscrowd": 0, "area": 14179.457099999996, "rle": {"size": [640, 427], "counts": "_RX21nc02N2N1O2N1O2N2N101N2N1O2N2N1O2N1O2N2O0O2N2O0O2N2N101N1O2O1N1O2O1N1O2O1N1O2O0O2N2N101N2N101N2N10O0O2N2N2N1O2O1O1O010O1O1O001O1O001O1O1O010001O0O1000000O2O00000O101O000O10001O0O1000001N101O1N101O001N101O0O2O1O001N101O001N10000O1000000000O10000000000000O1000000^@dM`=\\2_BnMX=R2hBWN`N]On=\\2dC^NSNWOZ>[2hC`NX<`1lC\\NT<d1QDVNP<j1TDQNm;o1XDkMi;U2[DfMf;Z2P300000001O0000001O0000001N1000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001N10001O001O001O001O001O001O001N101O0O2O001N101O001N2O001N101O0O2O001N101O001N101O0O2O1O1N101O1N2O001O1N2O001N2O1O0O2O1O1N101O1O1N3N1O1N2O`im0"}, "label": "blue umbrella held by woman in striped shirt"}]}
{"id": 66593, "image": "COCO_train2014_000000066593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light blue umbrella\"."}], "task": "refering", "answer_template": "The \"a light blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 101, 102, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 125, 127, 128, 129, 130, 131, 158, 173], "bbox": [0.26824355971896957, 0.235203125, 0.888056206088993, 0.513625], "iscrowd": 0, "area": 14179.457099999996, "rle": {"size": [640, 427], "counts": "_RX21nc02N2N1O2N1O2N2N101N2N1O2N2N1O2N1O2N2O0O2N2O0O2N2N101N1O2O1N1O2O1N1O2O1N1O2O0O2N2N101N2N101N2N10O0O2N2N2N1O2O1O1O010O1O1O001O1O001O1O1O010001O0O1000000O2O00000O101O000O10001O0O1000001N101O1N101O001N101O0O2O1O001N101O001N10000O1000000000O10000000000000O1000000^@dM`=\\2_BnMX=R2hBWN`N]On=\\2dC^NSNWOZ>[2hC`NX<`1lC\\NT<d1QDVNP<j1TDQNm;o1XDkMi;U2[DfMf;Z2P300000001O0000001O0000001N1000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001N10001O001O001O001O001O001O001N101O0O2O001N101O001N2O001N101O0O2O001N101O001N101O0O2O1O1N101O1N2O001O1N2O001N2O1O0O2O1O1N101O1O1N3N1O1N2O`im0"}, "label": "a light blue umbrella"}]}
{"id": 66593, "image": "COCO_train2014_000000066593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red umbrella on the left\"."}], "task": "refering", "answer_template": "The \"the red umbrella on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 90, 91, 92, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 152, 153], "bbox": [0.0, 0.253375, 0.39159250585480093, 0.4448125], "iscrowd": 0, "area": 12942.23605, "rle": {"size": [640, 427], "counts": "X7V1d`0V210O000001O00010O00001O0000010O00001O00010O0000001O00010O00001O00010O0000001O01O01O00001O0001O01O00001O01O01O0000001O01O01O001O0010O01O001O0010O01O001O1O0010O01O001O0010O01O001O1O0010OO2O001O0O2O001O0O2O001O1O0O2O001O0O2O001O0O2O001O1O0O2O1O2N1N2O1O1O1N3N1O1O1O1N3N1O1O1N2O1O2N1O1N2O1O1O2M2O1O1O1N3N1O1O1O1N2O2N1O1N2O1O2N1N2O1O1O1OfWR5"}, "label": "the red umbrella on the left"}]}
{"id": 66593, "image": "COCO_train2014_000000066593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the maroon umbrella\"."}], "task": "refering", "answer_template": "The \"the maroon umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 90, 91, 92, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 152, 153], "bbox": [0.0, 0.253375, 0.39159250585480093, 0.4448125], "iscrowd": 0, "area": 12942.23605, "rle": {"size": [640, 427], "counts": "X7V1d`0V210O000001O00010O00001O0000010O00001O00010O0000001O00010O00001O00010O0000001O01O01O00001O0001O01O00001O01O01O0000001O01O01O001O0010O01O001O0010O01O001O1O0010O01O001O0010O01O001O1O0010OO2O001O0O2O001O0O2O001O1O0O2O001O0O2O001O0O2O001O1O0O2O1O2N1N2O1O1O1N3N1O1O1O1N3N1O1O1N2O1O2N1O1N2O1O1O2M2O1O1O1N3N1O1O1O1N2O2N1O1N2O1O2N1N2O1O1O1OfWR5"}, "label": "the maroon umbrella"}]}
{"id": 566308, "image": "COCO_train2014_000000566308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder mid - air\"."}], "task": "refering", "answer_template": "The \"a skateboarder mid - air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 127, 128, 129, 141, 142, 143, 156, 157, 158, 172, 173, 174, 175, 188, 189, 190], "bbox": [0.4013443396226415, 0.304390625, 0.708372641509434, 0.570515625], "iscrowd": 0, "area": 7403.039950000004, "rle": {"size": [640, 424], "counts": "``Z31kc05M3O2N1O1O1O1O1O1O2N1O10010O01O010O1OcM_OZA`0f>AYA`0f>AYA?g>CWA=i>DUA=k>DTA<l>DTA;m>FQA;o>EQA;o>EQA;P?En@;T?Dl@<T?Dk@=V?CS_OI]1d0`?Cl^O2b1:b?2[@Od?3[@Md?6Z@Je?8Y@Hg?;W@Eh?o0e_OQOZ`0V1__OkN_`0]1[_ObNf`0e1S_O[Nm`0\\20000000O101O07I8H9G8I5LO0001O001O0006KN1O2N00010O001O000010O0001O00fLX@\\2i?cMZ@k0>WOW?N^@b0b0_OQ?N`@:f0Gj>0b@0j01d>Ne@Jl07`>Ng@Dn0>Z>Nk@]Oo0e0W>MfB2[=MeB3[=MfB2[=MfB2[=MeB3\\=LeB3[=MeB3\\=LeB3\\=LeB3[=MeB3\\=McB3^=Dl_OOf2=_=Cm_ONd2?_=Cn_OMc2`0`=Ao_ONa2a0a=@o_ON`2b0b=_Oo_ON_2b0d=^OfBb0\\`000000001O0000000000m\\OCcb0m00000000O2O0000001K4L4L4M3L5K4LfU]2"}, "label": "a skateboarder mid - air"}]}
{"id": 566308, "image": "COCO_train2014_000000566308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder jumping in midair\"."}], "task": "refering", "answer_template": "The \"a skateboarder jumping in midair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 127, 128, 129, 141, 142, 143, 156, 157, 158, 172, 173, 174, 175, 188, 189, 190], "bbox": [0.4013443396226415, 0.304390625, 0.708372641509434, 0.570515625], "iscrowd": 0, "area": 7403.039950000004, "rle": {"size": [640, 424], "counts": "``Z31kc05M3O2N1O1O1O1O1O1O2N1O10010O01O010O1OcM_OZA`0f>AYA`0f>AYA?g>CWA=i>DUA=k>DTA<l>DTA;m>FQA;o>EQA;o>EQA;P?En@;T?Dl@<T?Dk@=V?CS_OI]1d0`?Cl^O2b1:b?2[@Od?3[@Md?6Z@Je?8Y@Hg?;W@Eh?o0e_OQOZ`0V1__OkN_`0]1[_ObNf`0e1S_O[Nm`0\\20000000O101O07I8H9G8I5LO0001O001O0006KN1O2N00010O001O000010O0001O00fLX@\\2i?cMZ@k0>WOW?N^@b0b0_OQ?N`@:f0Gj>0b@0j01d>Ne@Jl07`>Ng@Dn0>Z>Nk@]Oo0e0W>MfB2[=MeB3[=MfB2[=MfB2[=MeB3\\=LeB3[=MeB3\\=LeB3\\=LeB3[=MeB3\\=McB3^=Dl_OOf2=_=Cm_ONd2?_=Cn_OMc2`0`=Ao_ONa2a0a=@o_ON`2b0b=_Oo_ON_2b0d=^OfBb0\\`000000001O0000000000m\\OCcb0m00000000O2O0000001K4L4L4M3L5K4LfU]2"}, "label": "a skateboarder jumping in midair"}]}
{"id": 140328, "image": "COCO_train2014_000000140328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white mug with cream and coffee in it\"."}], "task": "refering", "answer_template": "The \"a white mug with cream and coffee in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 142, 143, 144, 145], "bbox": [0.11173437500000001, 0.09874999999999999, 0.389390625, 0.4125208333333334], "iscrowd": 0, "area": 18076.30925, "rle": {"size": [480, 640], "counts": "SkQ14i>4M3L5K4M3L5K4L4M3L5N1N2O2M2O1N3N1N2O1N3N1N2N3N1N3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M2N3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N1N11O0O100000000O100000001O000000000000000000001O000000000000000001O01O000000000000000000001O00000O2N101N1O2N1O2N101N1O2N1O2N101N1O2N1O2N101N1O2N1O2N2N101N1O2N1O3M2O1N3M2N3M2N2O2L3M3N2M2N2O1N2N2O1N2N2O1N2N101N2N2N2O2M2N3M2N3M2O2M2N3M2N3M2I^Vg5"}, "label": "a white mug with cream and coffee in it"}]}
{"id": 140328, "image": "COCO_train2014_000000140328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white tea cup with half cup tea with saucer in the dining table\"."}], "task": "refering", "answer_template": "The \"white tea cup with half cup tea with saucer in the dining table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 142, 143, 144, 145], "bbox": [0.11173437500000001, 0.09874999999999999, 0.389390625, 0.4125208333333334], "iscrowd": 0, "area": 18076.30925, "rle": {"size": [480, 640], "counts": "SkQ14i>4M3L5K4M3L5K4L4M3L5N1N2O2M2O1N3N1N2O1N3N1N2N3N1N3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M2N3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N1N11O0O100000000O100000001O000000000000000000001O000000000000000001O01O000000000000000000001O00000O2N101N1O2N1O2N101N1O2N1O2N101N1O2N1O2N101N1O2N1O2N2N101N1O2N1O3M2O1N3M2N3M2N2O2L3M3N2M2N2O1N2N2O1N2N2O1N2N101N2N2N2O2M2N3M2N3M2O2M2N3M2N3M2I^Vg5"}, "label": "white tea cup with half cup tea with saucer in the dining table"}]}
{"id": 214059, "image": "COCO_train2014_000000214059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white commuter bus parked at an airport\"."}], "task": "refering", "answer_template": "The \"a white commuter bus parked at an airport\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.700171875, 0.3389461358313817, 1.0, 0.983887587822014], "iscrowd": 0, "area": 46036.49435, "rle": {"size": [427, 640], "counts": "mPk57m<a0F5J2O1N2O0O01O00100O1O001OnF7Y4lNgHl0Q3<lLhNk6:RIb0X3`0hLdNi6Z1aL;Y3FhK^1R4bNQKa2j4^MYJd3b5V2J6L4M2O1O2N1N2O1O1O1O2O000K5000001O0001O000001O0001O0000000000O10001O00000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000001O00000O100000001O00000000000000000000001O00000000000000000000001O000000000000001O000000000000001O00000000000O101O0000000000001O00000000001O00000000001O0000000000001O0000000000O10]1cNh4WK[E"}, "label": "a white commuter bus parked at an airport"}]}
{"id": 214059, "image": "COCO_train2014_000000214059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rb - 4 are the numbers on this white bus that is going down the road\"."}], "task": "refering", "answer_template": "The \"rb - 4 are the numbers on this white bus that is going down the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.700171875, 0.3389461358313817, 1.0, 0.983887587822014], "iscrowd": 0, "area": 46036.49435, "rle": {"size": [427, 640], "counts": "mPk57m<a0F5J2O1N2O0O01O00100O1O001OnF7Y4lNgHl0Q3<lLhNk6:RIb0X3`0hLdNi6Z1aL;Y3FhK^1R4bNQKa2j4^MYJd3b5V2J6L4M2O1O2N1N2O1O1O1O2O000K5000001O0001O000001O0001O0000000000O10001O00000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000001O00000O100000001O00000000000000000000001O00000000000000000000001O000000000000001O000000000000001O00000000000O101O0000000000001O00000000001O00000000001O0000000000001O0000000000O10]1cNh4WK[E"}, "label": "rb - 4 are the numbers on this white bus that is going down the road"}]}
{"id": 427051, "image": "COCO_train2014_000000427051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the waiting area to the right of the fron of the bus , with clear glass and green trim\"."}], "task": "refering", "answer_template": "The \"the waiting area to the right of the fron of the bus , with clear glass and green trim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.79378125, 0.4123958333333333, 0.999203125, 0.7042499999999999], "iscrowd": 0, "area": 15465.462950000003, "rle": {"size": [480, 640], "counts": "aZ^77i>?A?A?A?A6J3M3L4M3M3M2N000000O100001O0O10000000000000001O2N1O1N101O00001O001O1O2N2N5J5L4L4L3M0000000000001O0O10000000000000000000000O2O000000000000000000000O1000001O0000000000000O1000000000001O000000000O10000000000000001O00000O100000000000000000001N10000000000j8"}, "label": "the waiting area to the right of the fron of the bus , with clear glass and green trim"}]}
{"id": 427051, "image": "COCO_train2014_000000427051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and yellow bus headed to stop dill hall circular\"."}], "task": "refering", "answer_template": "The \"a red and yellow bus headed to stop dill hall circular\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313], "bbox": [0.12471874999999999, 0.3012708333333334, 0.8106718749999999, 0.7978958333333334], "iscrowd": 0, "area": 85083.58065000002, "rle": {"size": [480, 640], "counts": "UiU1j0Q=U1lNU1]Ob0O2O001O1O0O2O001O001N2O001O001N101O1O001N101O1O0O2O001O2N2M3N1O2N001N100000001N10000000O010O1O10O01O1OJ70O10000001N1000001O0O1000001O000O101O0000001N1000001O0O10001O00000O101O00000O2O0000001O0O10001O000O10001O00000O2O0000001N1000001O000O10001O00000O2O0000001N1000001O000O10001O000O101O0000001N1000001O0O1000001O000O101O00000O2O0000001O0O3N1O2N2N1O2M3N1O2N1O0O10000000000O100000000O100000O1O1O001O1O1O1O001O1O1O1000O100000000O2O000000000O10000000000O100000000O10000000000O1000001O0O10000000000O100000000O10000000000O100000000O01000000000O1000O100000000000000O100000000000001N100000000000000O1000000000001O0O10000000000001O001O1O001O10O01O001O1O001O000000O100O1O101N100O100O100O100O1000000000001N10000000000000000000000000001O00000000000O10000000000000001O0000000000000O100000000O101O00000O10000000000O100000000O2O0000000O10000000000O100000<D<C>C=C<D=C=C<C>C;E2N2N2M3N2N2`N\\_h1"}, "label": "a red and yellow bus headed to stop dill hall circular"}]}
{"id": 427051, "image": "COCO_train2014_000000427051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and tan bus that has the number 5\"."}], "task": "refering", "answer_template": "The \"a red and tan bus that has the number 5\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313], "bbox": [0.12471874999999999, 0.3012708333333334, 0.8106718749999999, 0.7978958333333334], "iscrowd": 0, "area": 85083.58065000002, "rle": {"size": [480, 640], "counts": "UiU1j0Q=U1lNU1]Ob0O2O001O1O0O2O001O001N2O001O001N101O1O001N101O1O0O2O001O2N2M3N1O2N001N100000001N10000000O010O1O10O01O1OJ70O10000001N1000001O0O1000001O000O101O0000001N1000001O0O10001O00000O101O00000O2O0000001O0O10001O000O10001O00000O2O0000001N1000001O000O10001O00000O2O0000001N1000001O000O10001O000O101O0000001N1000001O0O1000001O000O101O00000O2O0000001O0O3N1O2N2N1O2M3N1O2N1O0O10000000000O100000000O100000O1O1O001O1O1O1O001O1O1O1000O100000000O2O000000000O10000000000O100000000O10000000000O1000001O0O10000000000O100000000O10000000000O100000000O01000000000O1000O100000000000000O100000000000001N100000000000000O1000000000001O0O10000000000001O001O1O001O10O01O001O1O001O000000O100O1O101N100O100O100O100O1000000000001N10000000000000000000000000001O00000000000O10000000000000001O0000000000000O100000000O101O00000O10000000000O100000000O2O0000000O10000000000O100000<D<C>C=C<D=C=C<C>C;E2N2N2M3N2N2`N\\_h1"}, "label": "a red and tan bus that has the number 5"}]}
{"id": 123949, "image": "COCO_train2014_000000123949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing purple talking on a cell phone\"."}], "task": "refering", "answer_template": "The \"a woman wearing purple talking on a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.569390625, 0.20729437229437228, 0.9927812500000002, 0.9915800865800865], "iscrowd": 0, "area": 68027.7272, "rle": {"size": [462, 640], "counts": "S`T51\\>:F9G:E:G:F9G:F9F;F9G9G:F5K2O2N2N2N1N3N2N2N1O2N2N1O2N2N2N1N3N2N2N1O001O1O1O1O1O1N2O1O1O1O1O1O1O1Ob0^O00000HRJPHn5o7SJQHm5m7VJRHj5m7WJSHi5k7YJUHg5j7ZJVHf5h7\\JXHd5g7^JWHc5g7_JYHa5f7`JZH`5e7aJ[H_5d7cJ[H]5m6SJdIa0_O\\5m6XJ^I=E[5l6]JZI9JZ5k6cJUI5OX5k6_KUIa4j6`KVI`4i6aKWI_4h6d1O1POdHnJ\\7P5R1M3N2M3N2N2M3N2O1O100O1O1O1O1O100O1O100O10000O100O10000M3F:J6N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N101O1O1O1N200O100O1O100O100O100O100000000000000000000001O0000000000001O001O0000000000O100000000000000000000000000001O1O1O1O1O1O001O1O1O1O1O1O1O1K5L4L3L5L4L4K5L4L4L4K5L4L4K4M4L4L4K5cLnIYNV6e1lIWNX6h1hIWNZ6g1hI\\NU6c1kIaNR6]1PJgNk5Y1UJkNh5S1YJQOd5m0^JWO^5h0bJ\\O[5b0gJAV5>jJGR57PKLm43SK1j4LYK8c4F^K>_4_OcKe0Y4ZOiKi0T4TOnKQ1n3mNTLS1T4bNnK^1_4TNcKk1j4fMXKZ2T5YMnJf2[5oLgJQ3_5hLbJX3c5aL`J^3f5[L[Je3j5ULXJj3n5nKTJR4Q6hKQJW4U6aKmI_4X6[KjId4P80002N2N2N2N2N2N6J7I8G8I7I7I7I7I8H7I7I7I7I8H7I7I7I7I7ITj1"}, "label": "a woman wearing purple talking on a cell phone"}]}
{"id": 123949, "image": "COCO_train2014_000000123949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of a woman in purple on her phone\"."}], "task": "refering", "answer_template": "The \"a picture of a woman in purple on her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.569390625, 0.20729437229437228, 0.9927812500000002, 0.9915800865800865], "iscrowd": 0, "area": 68027.7272, "rle": {"size": [462, 640], "counts": "S`T51\\>:F9G:E:G:F9G:F9F;F9G9G:F5K2O2N2N2N1N3N2N2N1O2N2N1O2N2N2N1N3N2N2N1O001O1O1O1O1O1N2O1O1O1O1O1O1O1Ob0^O00000HRJPHn5o7SJQHm5m7VJRHj5m7WJSHi5k7YJUHg5j7ZJVHf5h7\\JXHd5g7^JWHc5g7_JYHa5f7`JZH`5e7aJ[H_5d7cJ[H]5m6SJdIa0_O\\5m6XJ^I=E[5l6]JZI9JZ5k6cJUI5OX5k6_KUIa4j6`KVI`4i6aKWI_4h6d1O1POdHnJ\\7P5R1M3N2M3N2N2M3N2O1O100O1O1O1O1O100O1O100O10000O100O10000M3F:J6N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N101O1O1O1N200O100O1O100O100O100O100000000000000000000001O0000000000001O001O0000000000O100000000000000000000000000001O1O1O1O1O1O001O1O1O1O1O1O1O1K5L4L3L5L4L4K5L4L4L4K5L4L4K4M4L4L4K5cLnIYNV6e1lIWNX6h1hIWNZ6g1hI\\NU6c1kIaNR6]1PJgNk5Y1UJkNh5S1YJQOd5m0^JWO^5h0bJ\\O[5b0gJAV5>jJGR57PKLm43SK1j4LYK8c4F^K>_4_OcKe0Y4ZOiKi0T4TOnKQ1n3mNTLS1T4bNnK^1_4TNcKk1j4fMXKZ2T5YMnJf2[5oLgJQ3_5hLbJX3c5aL`J^3f5[L[Je3j5ULXJj3n5nKTJR4Q6hKQJW4U6aKmI_4X6[KjId4P80002N2N2N2N2N2N6J7I8G8I7I7I7I7I8H7I7I7I7I8H7I7I7I7I7ITj1"}, "label": "a picture of a woman in purple on her phone"}]}
{"id": 17451, "image": "COCO_train2014_000000017451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a parking meter labeled number 32\"."}], "task": "refering", "answer_template": "The \"a parking meter labeled number 32\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 146, 147, 148, 149, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287], "bbox": [0.37146875, 0.23104377104377105, 0.53421875, 0.5995622895622895], "iscrowd": 0, "area": 14031.503399999994, "rle": {"size": [594, 640], "counts": "mVZ43]b05K5L3L2O2M3M2O2M3M2O2O1N101O1N101O1N101N2O001N101N2O001O1O0\\CYO^7g0]G=c8EXF^1h9dNRE`2n:a2100000000O10001O000O100000001O0O2O1O001O1N101O1O001N2O1O1O2N1N3N1O2N1O2M2O2N1O2N2M5L3M4L3M4L6I6K5K1O1O1N2O2N1O1O1N2O1O1O1O1N2O1O1O1O1N2O1Odd\\5"}, "label": "a parking meter labeled number 32"}]}
{"id": 17451, "image": "COCO_train2014_000000017451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the parking meter labeled 32\"."}], "task": "refering", "answer_template": "The \"the parking meter labeled 32\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 146, 147, 148, 149, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287], "bbox": [0.37146875, 0.23104377104377105, 0.53421875, 0.5995622895622895], "iscrowd": 0, "area": 14031.503399999994, "rle": {"size": [594, 640], "counts": "mVZ43]b05K5L3L2O2M3M2O2M3M2O2O1N101O1N101O1N101N2O001N101N2O001O1O0\\CYO^7g0]G=c8EXF^1h9dNRE`2n:a2100000000O10001O000O100000001O0O2O1O001O1N101O1O001N2O1O1O2N1N3N1O2N1O2M2O2N1O2N2M5L3M4L3M4L6I6K5K1O1O1N2O2N1O1O1N2O1O1O1O1N2O1O1O1O1N2O1Odd\\5"}, "label": "the parking meter labeled 32"}]}
{"id": 17451, "image": "COCO_train2014_000000017451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old parking meter\"."}], "task": "refering", "answer_template": "The \"an old parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 348, 349, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379, 396, 397, 398, 399, 400, 401, 419, 420, 421, 422, 423, 424], "bbox": [0.08759375, 0.10942760942760943, 0.538109375, 0.895942760942761], "iscrowd": 0, "area": 86864.74334999998, "rle": {"size": [594, 640], "counts": "ldP1h0ja0_1`Na1`N`1`Nl0TO9G9G4L1O2N1N3N1O2N1O2N1O2N1N3N1O2I6K6J5I8B=B?F9N3L3N3L3M4M2M3M4M2M3M3N2M2N3N2M2N3N2M2N3N2M3M2O2O1N101O1N101O1N2O0O2O1O0O2O1N101O1N2O0O2O1O0O2O1O0O3N2M3N2N2M3N1N3N2N2O1N2N101N2O1N2N2O1N101N2O1N2N2O0O2O2M2N3N1N3NO000010O01O00001O001O00001O0010O0001O001O00001O001O002O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N001O001O00100O001O001O001O001O1O0010O01O001O001O1O001N1O2N101N1VJaH`0a7dNZI[1h6SOiHl0X7[LaKd3g9O001O1O1O10O01O1O1O1O001O1O1O1O00100O1O1O1O001O1O1O1O00100O1O1O001O1O1O1N1M4L4L4L4L3M4L4L4L4L4L3M4L4L4L4L3M4L4L4L4M2M4L4L4L4L3M4L4L4L4L4L3M4L4L4Lm[[5"}, "label": "an old parking meter"}]}
{"id": 17451, "image": "COCO_train2014_000000017451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a two meter standing\"."}], "task": "refering", "answer_template": "The \"a two meter standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 348, 349, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379, 396, 397, 398, 399, 400, 401, 419, 420, 421, 422, 423, 424], "bbox": [0.08759375, 0.10942760942760943, 0.538109375, 0.895942760942761], "iscrowd": 0, "area": 86864.74334999998, "rle": {"size": [594, 640], "counts": "ldP1h0ja0_1`Na1`N`1`Nl0TO9G9G4L1O2N1N3N1O2N1O2N1O2N1N3N1O2I6K6J5I8B=B?F9N3L3N3L3M4M2M3M4M2M3M3N2M2N3N2M2N3N2M2N3N2M3M2O2O1N101O1N101O1N2O0O2O1O0O2O1N101O1N2O0O2O1O0O2O1O0O3N2M3N2N2M3N1N3N2N2O1N2N101N2O1N2N2O1N101N2O1N2N2O0O2O2M2N3N1N3NO000010O01O00001O001O00001O0010O0001O001O00001O001O002O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N001O001O00100O001O001O001O001O1O0010O01O001O001O1O001N1O2N101N1VJaH`0a7dNZI[1h6SOiHl0X7[LaKd3g9O001O1O1O10O01O1O1O1O001O1O1O1O00100O1O1O1O001O1O1O1O00100O1O1O001O1O1O1N1M4L4L4L4L3M4L4L4L4L4L3M4L4L4L4L3M4L4L4L4M2M4L4L4L4L3M4L4L4L4L4L3M4L4L4Lm[[5"}, "label": "a two meter standing"}]}
{"id": 427060, "image": "COCO_train2014_000000427060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black woman in blue jeans\"."}], "task": "refering", "answer_template": "The \"black woman in blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 61, 62, 78, 79, 80, 97, 98, 99, 115, 116, 117, 133, 134, 151, 152, 170], "bbox": [0.33254, 0.20717717717717715, 0.52212, 0.7881681681681681], "iscrowd": 0, "area": 6606.3538, "rle": {"size": [333, 500], "counts": "l]f11\\:001O00001O00000O2O00000000000000006J001N100000001N100O1O100O1L4N2N2O1N2SHUOW6l0_IAd5BkIn0=Fe5^OlIo0<Gc5FfIe0d0Ja5CkId0b0M]5CoIb0a00\\5_ORJd0?1[5]OTJd0>4Y5\\OVJb0?5X5[OWJd0=4Y5o0bJUO[5n0bJUO\\5Q3M2N3M2N3L4M2N3]KbJQ4_5mKcJR4]5mKdJS4\\5lKeJU4Z5jKgJV4h3hK\\MOoNZ4c3kK[MITO]4_3nKiMS4U2QLhMo3SOcK]2`0^NP4oNcKb2`0\\NU4hN_Kj2`0[Na4c1bK[N_4c1eKZN[4e1hKYNY4f1iKXNX4f1lKWNU4g1nKWNR4h1RLUNo3i1TLTNQ4g1RLWNQ4f1RLVNS4e1PLYNS4c1PLZNU4a1oK[NU4a1nK[NX4`1^2L5K4L5K4L5K4L5K4L4L5Kma]2"}, "label": "black woman in blue jeans"}]}
{"id": 427060, "image": "COCO_train2014_000000427060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black vest and sun glasses riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black vest and sun glasses riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 61, 62, 78, 79, 80, 97, 98, 99, 115, 116, 117, 133, 134, 151, 152, 170], "bbox": [0.33254, 0.20717717717717715, 0.52212, 0.7881681681681681], "iscrowd": 0, "area": 6606.3538, "rle": {"size": [333, 500], "counts": "l]f11\\:001O00001O00000O2O00000000000000006J001N100000001N100O1O100O1L4N2N2O1N2SHUOW6l0_IAd5BkIn0=Fe5^OlIo0<Gc5FfIe0d0Ja5CkId0b0M]5CoIb0a00\\5_ORJd0?1[5]OTJd0>4Y5\\OVJb0?5X5[OWJd0=4Y5o0bJUO[5n0bJUO\\5Q3M2N3M2N3L4M2N3]KbJQ4_5mKcJR4]5mKdJS4\\5lKeJU4Z5jKgJV4h3hK\\MOoNZ4c3kK[MITO]4_3nKiMS4U2QLhMo3SOcK]2`0^NP4oNcKb2`0\\NU4hN_Kj2`0[Na4c1bK[N_4c1eKZN[4e1hKYNY4f1iKXNX4f1lKWNU4g1nKWNR4h1RLUNo3i1TLTNQ4g1RLWNQ4f1RLVNS4e1PLYNS4c1PLZNU4a1oK[NU4a1nK[NX4`1^2L5K4L5K4L5K4L5K4L4L5Kma]2"}, "label": "a woman wearing a black vest and sun glasses riding a motorcycle"}]}
{"id": 459835, "image": "COCO_train2014_000000459835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red colour speed boat flying in the air\"."}], "task": "refering", "answer_template": "The \"a red colour speed boat flying in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 82, 83, 84, 85, 86, 106, 107, 108, 109], "bbox": [0.540734375, 0.1753125, 0.774078125, 0.44690624999999995], "iscrowd": 0, "area": 6157.46545, "rle": {"size": [320, 640], "counts": "PV\\33m92N2O1N2O1N3N1N2O1N2O1N2O0O010O1000O0100O10O0100O1HWOUGk0k8UOSGn0m83O001O100O001O1CoNiGR1V8SOeGm0Z8>O1O1N101O1O1N1010000O01000O10O0100O01000000O100000O01000000O100000O01000000O100000O0^O\\NiHd1V7`NgH_1X7fNeHX1\\7lNaHS1_7f010000O01000O100O10O03M3M3M2N4L5K4L5L4L4L3M4L001O001O1O001O001O001O0010O01O001O001O10O01O001O00000001O0O10O101O2N2N2N6I3N1O2N1OoU]1"}, "label": "a red colour speed boat flying in the air"}]}
{"id": 459835, "image": "COCO_train2014_000000459835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red boat with two men riding it\"."}], "task": "refering", "answer_template": "The \"a red boat with two men riding it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 82, 83, 84, 85, 86, 106, 107, 108, 109], "bbox": [0.540734375, 0.1753125, 0.774078125, 0.44690624999999995], "iscrowd": 0, "area": 6157.46545, "rle": {"size": [320, 640], "counts": "PV\\33m92N2O1N2O1N3N1N2O1N2O1N2O0O010O1000O0100O10O0100O1HWOUGk0k8UOSGn0m83O001O100O001O1CoNiGR1V8SOeGm0Z8>O1O1N101O1O1N1010000O01000O10O0100O01000000O100000O01000000O100000O01000000O100000O0^O\\NiHd1V7`NgH_1X7fNeHX1\\7lNaHS1_7f010000O01000O100O10O03M3M3M2N4L5K4L5L4L4L3M4L001O001O1O001O001O001O0010O01O001O001O10O01O001O00000001O0O10O101O2N2N2N6I3N1O2N1OoU]1"}, "label": "a red boat with two men riding it"}]}
{"id": 459835, "image": "COCO_train2014_000000459835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue , white , and red air - boat that is ahead of two other air - boats\"."}], "task": "refering", "answer_template": "The \"a blue , white , and red air - boat that is ahead of two other air - boats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 96, 97, 117, 118, 119, 120, 142], "bbox": [0.015734375000000002, 0.2653125, 0.22584374999999998, 0.5686875], "iscrowd": 0, "area": 5130.7819500000005, "rle": {"size": [320, 640], "counts": "eV32n94M3M1N3M3M2O2M4L3N2M1O1O010O1O010O100O010O1O10O0100O010O100O01000O10O0100O10O0100O100O00100O10O0100O1O010O100O1O010O100O00100O10O0100O100O010O1000O0100O10O00100O10O01O010O010O0010O010O01O010O010O0010O010O01O01O001O000O2OH7I8O010O001O010O1O010O1000000O011O001N2O001M2LQbj4"}, "label": "a blue , white , and red air - boat that is ahead of two other air - boats"}]}
{"id": 459835, "image": "COCO_train2014_000000459835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue jetski is in the lead of the others\"."}], "task": "refering", "answer_template": "The \"the blue jetski is in the lead of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 96, 97, 117, 118, 119, 120, 142], "bbox": [0.015734375000000002, 0.2653125, 0.22584374999999998, 0.5686875], "iscrowd": 0, "area": 5130.7819500000005, "rle": {"size": [320, 640], "counts": "eV32n94M3M1N3M3M2O2M4L3N2M1O1O010O1O010O100O010O1O10O0100O010O100O01000O10O0100O10O0100O100O00100O10O0100O1O010O100O1O010O100O00100O10O0100O100O010O1000O0100O10O00100O10O01O010O010O0010O010O01O010O010O0010O010O01O01O001O000O2OH7I8O010O001O010O1O010O1000000O011O001N2O001M2LQbj4"}, "label": "the blue jetski is in the lead of the others"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy sitting against the wall , reading\"."}], "task": "refering", "answer_template": "The \"the boy sitting against the wall , reading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 196], "bbox": [0.474609375, 0.15183333333333332, 0.860171875, 0.47743749999999996], "iscrowd": 0, "area": 15930.01975, "rle": {"size": [480, 640], "counts": "jf^41n>100O1O101N100O1000001O000000010O00001O00001O000010O000001O0000001N100O10000O2O0O100O101N100O10000O10O0100O10O010000O10O0100O10O_OYOjB0Af0d=\\OkBe0T=]OiBe0V=]OhBc0X=_OfBb0Y=@dBb0[=_OdBb0[=@cB`0]=BaB?^=b0O1O1O001N2O1N101O1N2O0O2O1N2O001N2O1N2O0O2O1N2O1N101N2N2O1N2O1N2000000O10000000O1000O11O1O001O1O001O1O001O1O001N2O001O1O001OK5L4L4L40000001O2M3N1O2N1O1O001O1O1O001O1O1O00UOmMiDS2V;PNhDP2X;RNfDn1Z;TNdDl1[;XNbDg1_;[N_De1`;^N^Db1b;aN[D_1e;n0O10000O10000O010BmLnDT3Q;>O1N2O1O1O1N2O1O1N101O1O1000O1O001N2O1O1N101O1N2O1N3N1O2M3N2N1N3N2M2O2N2M2O3M2M3N3M2M3PNeCY1]<cNfC\\1P=00O100N2O1N2N2N2N2N2O1N1O2N3M4L4M4K4L4L4L5K4M3LUbY1"}, "label": "the boy sitting against the wall , reading"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy sitting on a floor reading something\"."}], "task": "refering", "answer_template": "The \"a boy sitting on a floor reading something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 196], "bbox": [0.474609375, 0.15183333333333332, 0.860171875, 0.47743749999999996], "iscrowd": 0, "area": 15930.01975, "rle": {"size": [480, 640], "counts": "jf^41n>100O1O101N100O1000001O000000010O00001O00001O000010O000001O0000001N100O10000O2O0O100O101N100O10000O10O0100O10O010000O10O0100O10O_OYOjB0Af0d=\\OkBe0T=]OiBe0V=]OhBc0X=_OfBb0Y=@dBb0[=_OdBb0[=@cB`0]=BaB?^=b0O1O1O001N2O1N101O1N2O0O2O1N2O001N2O1N2O0O2O1N2O1N101N2N2O1N2O1N2000000O10000000O1000O11O1O001O1O001O1O001O1O001N2O001O1O001OK5L4L4L40000001O2M3N1O2N1O1O001O1O1O001O1O1O00UOmMiDS2V;PNhDP2X;RNfDn1Z;TNdDl1[;XNbDg1_;[N_De1`;^N^Db1b;aN[D_1e;n0O10000O10000O010BmLnDT3Q;>O1N2O1O1O1N2O1O1N101O1O1000O1O001N2O1O1N101O1N2O1N3N1O2M3N2N1N3N2M2O2N2M2O3M2M3N3M2M3PNeCY1]<cNfC\\1P=00O100N2O1N2N2N2N2N2O1N1O2N3M4L4M4K4L4L4L5K4M3LUbY1"}, "label": "a boy sitting on a floor reading something"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a green sweater facing the window\"."}], "task": "refering", "answer_template": "The \"a person in a green sweater facing the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.348875, 0.46066666666666667, 0.67246875, 1.0], "iscrowd": 0, "area": 35574.4723, "rle": {"size": [480, 640], "counts": "ZmX31n>4L3M3M3M3M4L3M3M3M3M4L3eBjNg<Y1QCoNk<e1O2M2O2N1O2M3N1O2N1O2M2O2N1O2N1L5VOi0K6L3N3M2M4M2N3L3N1O1N2O1N2O1O1N2O1N2O1O1N2O1N2nNSKPHn4m7T1L4L4L4L4L4L4N2N2N2N2N2N2M3N2N2N2N2N2O100O10000O100O10000O100O10000O100000000000000000000000000000000000000001O00001O001O00001N101O00001O001O001O00001O001O00001O001O005K5K5K4L5K4L3M3M3M3M3M3M3jJVGP4m8eK[G[4h8ZKaGe4b8PKfGP5U900000001O00000000001O000000001O000000001O2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2kLUEW11`Nl:6eE63CZ:5_GIc85^GJe83\\GLf81]GMe81\\GNe81]GMe81\\GLg83[GHj85R3O1O2N1OYRR3"}, "label": "a person in a green sweater facing the window"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with longhair and has her back towards the camera\"."}], "task": "refering", "answer_template": "The \"the girl with longhair and has her back towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.348875, 0.46066666666666667, 0.67246875, 1.0], "iscrowd": 0, "area": 35574.4723, "rle": {"size": [480, 640], "counts": "ZmX31n>4L3M3M3M3M4L3M3M3M3M4L3eBjNg<Y1QCoNk<e1O2M2O2N1O2M3N1O2N1O2M2O2N1O2N1L5VOi0K6L3N3M2M4M2N3L3N1O1N2O1N2O1O1N2O1N2O1O1N2O1N2nNSKPHn4m7T1L4L4L4L4L4L4N2N2N2N2N2N2M3N2N2N2N2N2O100O10000O100O10000O100O10000O100000000000000000000000000000000000000001O00001O001O00001N101O00001O001O001O00001O001O00001O001O005K5K5K4L5K4L3M3M3M3M3M3M3jJVGP4m8eK[G[4h8ZKaGe4b8PKfGP5U900000001O00000000001O000000001O000000001O2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2kLUEW11`Nl:6eE63CZ:5_GIc85^GJe83\\GLf81]GMe81\\GNe81]GMe81\\GLg83[GHj85R3O1O2N1OYRR3"}, "label": "the girl with longhair and has her back towards the camera"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women seated cross - legged and taking a picture with a cell phone\"."}], "task": "refering", "answer_template": "The \"a women seated cross - legged and taking a picture with a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 138, 139, 140, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281], "bbox": [0.00042187500000000005, 0.243875, 0.28853125, 0.7637916666666666], "iscrowd": 0, "area": 26885.381549999987, "rle": {"size": [480, 640], "counts": "d48]>`0J6YCWOl:o0kDVOS;P1fDROY;T1`DQO];S1]DUO];n0_DXO\\;k0_DXO`;k0[DWOd;S201^DoLR;c3L4L4L3L5L:bF]Kl7o4gGSKW8Z5\\GiJb8l5N2N2N1O2O001O1O4L3M4L3M3M1O1O1O1O1O1O1O100O1O1O1O00001O000010O00010O010O001O010O00010O2N2O1N2N2N2^InGc0O`4T8kJnGc0Ob4U8hJoGc0Nd4U8fJoG3?W5c7dJPH2?Y5n7fJRHZ5o7dJSH\\5i8O1O00100O1O010O1O10O01O1O10O01O100O00100O1O00100O1O001L4L4M2UMfEn0]:POdEP1]:nNcES1^:kNcEU1]:jNdEV1]:hNcEZ1\\:eNeE[1\\:cNeE]1[:bNeE_1\\:_NeEa1\\:]NeEd1Z:[NfEf1Z:YNgEg1X:YNiEg1W:YNhEh1W:XNjEh1U:XNlEh1S:XNmEi1S:VNnEj1Q:VNPFj1o9VNQFk1o9TNRFk1n9UNSFk1l9UNTFk1l9VNTFj1l9UNUFj1k9VNUFk1j9UNWFj1i9VNXFi1h9WNXFj1g9WNYFh1g9XNZFh1e9XN[Fh1e9YN[Ff1d9[N]Fc1c9^N]Fb1c9^N^Fa1b9XNjEUOd0a2c9[NjEROc0c2d9ZNkEPOc0d2d9\\NbFc1_9\\NaFc1`9VNdE\\Om0]2`9WNeEZOk0_2`9WNgEWOi0b2`9WNgFh1Z9XNfFg1[9XNeFi1\\9VNeFh1\\9WNeFh1\\9XNdFg1\\9ZNeFV1nN[N\\:?gFV1kN\\N^:?gFT1kN]N^:`0hFQ1jN_N_:`0gFQ1hNaNb:=fFQ1hNbNe::eFn0mNeN_:OYE5[1V1POdN]:NYE9Z1a0eND=Ac::\\F=gNG<@a:4PE1^1a0gNIf;DeE`0gNKe;CgE?fNM\\=NhB1Z=IkB6W=ElB<l=O1O1O1O1O1O001O1O1O2O2MPPe6"}, "label": "a women seated cross - legged and taking a picture with a cell phone"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl wearing black holding a cell phone while sitting on the ground studying along with 3 others\"."}], "task": "refering", "answer_template": "The \"a young girl wearing black holding a cell phone while sitting on the ground studying along with 3 others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 138, 139, 140, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281], "bbox": [0.00042187500000000005, 0.243875, 0.28853125, 0.7637916666666666], "iscrowd": 0, "area": 26885.381549999987, "rle": {"size": [480, 640], "counts": "d48]>`0J6YCWOl:o0kDVOS;P1fDROY;T1`DQO];S1]DUO];n0_DXO\\;k0_DXO`;k0[DWOd;S201^DoLR;c3L4L4L3L5L:bF]Kl7o4gGSKW8Z5\\GiJb8l5N2N2N1O2O001O1O4L3M4L3M3M1O1O1O1O1O1O1O100O1O1O1O00001O000010O00010O010O001O010O00010O2N2O1N2N2N2^InGc0O`4T8kJnGc0Ob4U8hJoGc0Nd4U8fJoG3?W5c7dJPH2?Y5n7fJRHZ5o7dJSH\\5i8O1O00100O1O010O1O10O01O1O10O01O100O00100O1O00100O1O001L4L4M2UMfEn0]:POdEP1]:nNcES1^:kNcEU1]:jNdEV1]:hNcEZ1\\:eNeE[1\\:cNeE]1[:bNeE_1\\:_NeEa1\\:]NeEd1Z:[NfEf1Z:YNgEg1X:YNiEg1W:YNhEh1W:XNjEh1U:XNlEh1S:XNmEi1S:VNnEj1Q:VNPFj1o9VNQFk1o9TNRFk1n9UNSFk1l9UNTFk1l9VNTFj1l9UNUFj1k9VNUFk1j9UNWFj1i9VNXFi1h9WNXFj1g9WNYFh1g9XNZFh1e9XN[Fh1e9YN[Ff1d9[N]Fc1c9^N]Fb1c9^N^Fa1b9XNjEUOd0a2c9[NjEROc0c2d9ZNkEPOc0d2d9\\NbFc1_9\\NaFc1`9VNdE\\Om0]2`9WNeEZOk0_2`9WNgEWOi0b2`9WNgFh1Z9XNfFg1[9XNeFi1\\9VNeFh1\\9WNeFh1\\9XNdFg1\\9ZNeFV1nN[N\\:?gFV1kN\\N^:?gFT1kN]N^:`0hFQ1jN_N_:`0gFQ1hNaNb:=fFQ1hNbNe::eFn0mNeN_:OYE5[1V1POdN]:NYE9Z1a0eND=Ac::\\F=gNG<@a:4PE1^1a0gNIf;DeE`0gNKe;CgE?fNM\\=NhB1Z=IkB6W=ElB<l=O1O1O1O1O1O001O1O1O2O2MPPe6"}, "label": "a young girl wearing black holding a cell phone while sitting on the ground studying along with 3 others"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the teal jacket\"."}], "task": "refering", "answer_template": "The \"the girl in the teal jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 156, 157, 158, 159, 160, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319], "bbox": [0.6956562500000001, 0.32410416666666664, 1.0, 0.8056458333333333], "iscrowd": 0, "area": 28360.61814999999, "rle": {"size": [480, 640], "counts": "Sl`61l>4K5K5L40O0100O1O10O0100O010O1O10O0100O10O01O1N2O1O101QOSOgCn0X<ROiCo0T<ROlCo0S<QOmCP1Q<ROoCo0o;QOQDP1m;QOTDP1j;POVDR1g;oNYDS1e;nN[DS1b;nN^DT1`;lNaDU1\\;lNdDV1Z;jNfDX12bNQ:8mEW1MeNU:5mEX1JfNW:6mEU1HiNY:4mEU1FjN[:5lES1EkN^:4lER1BnN_:4lEP1CmN`:5lEo0AnNa:7kEk1S:WNkEi1T:[NjEc1W:_NgEa1X:cNfE\\1Z:fNdEZ1[:jNbEV1^:lNaER1_:h100O100O010O1O10O0100O1O10O0100O1O10O0100`N]KbHd4^7gKVHZ4i7kKSHT4m7nKeG]OLf4^8oKcG]OOd4]8RL`G]O1b4^8SL^G]O3a4^8aLaG^3_8cL`G^3`8cL_G]3`8dL_G]3`8b1O001O1O010O001O1O001O00100O001O1O001O010000O100O10000000000000001O001O001O001O001O00001O001O001N101O003M3M2N3M1OO100O100O100O100O100O100001O1O001O1O1N2O2N1O2N1O2N1O2N1O2N3M2N2M3M4L3L4M3M3M3L3N3M3M3L4M3M3M3L4MkH"}, "label": "the girl in the teal jacket"}]}
{"id": 17468, "image": "COCO_train2014_000000017468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in white reads a paper\"."}], "task": "refering", "answer_template": "The \"a girl in white reads a paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 156, 157, 158, 159, 160, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319], "bbox": [0.6956562500000001, 0.32410416666666664, 1.0, 0.8056458333333333], "iscrowd": 0, "area": 28360.61814999999, "rle": {"size": [480, 640], "counts": "Sl`61l>4K5K5L40O0100O1O10O0100O010O1O10O0100O10O01O1N2O1O101QOSOgCn0X<ROiCo0T<ROlCo0S<QOmCP1Q<ROoCo0o;QOQDP1m;QOTDP1j;POVDR1g;oNYDS1e;nN[DS1b;nN^DT1`;lNaDU1\\;lNdDV1Z;jNfDX12bNQ:8mEW1MeNU:5mEX1JfNW:6mEU1HiNY:4mEU1FjN[:5lES1EkN^:4lER1BnN_:4lEP1CmN`:5lEo0AnNa:7kEk1S:WNkEi1T:[NjEc1W:_NgEa1X:cNfE\\1Z:fNdEZ1[:jNbEV1^:lNaER1_:h100O100O010O1O10O0100O1O10O0100O1O10O0100`N]KbHd4^7gKVHZ4i7kKSHT4m7nKeG]OLf4^8oKcG]OOd4]8RL`G]O1b4^8SL^G]O3a4^8aLaG^3_8cL`G^3`8cL_G]3`8dL_G]3`8b1O001O1O010O001O1O001O00100O001O1O001O010000O100O10000000000000001O001O001O001O001O00001O001O001N101O003M3M2N3M1OO100O100O100O100O100O100001O1O001O1O1N2O2N1O2N1O2N1O2N1O2N3M2N2M3M4L3L4M3M3M3L3N3M3M3L4M3M3M3L4MkH"}, "label": "a girl in white reads a paper"}]}
{"id": 443455, "image": "COCO_train2014_000000443455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child beside an adult\"."}], "task": "refering", "answer_template": "The \"a child beside an adult\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 338, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 402, 403, 404, 421, 422, 423, 424, 425, 426, 443, 444, 445, 446, 447, 448, 449, 464, 465, 466, 467, 468, 469, 470, 471], "bbox": [0.07486928104575163, 0.5389705882352942, 0.44483660130718955, 1.0], "iscrowd": 0, "area": 41554.75805000001, "rle": {"size": [612, 612], "counts": "Ylk0c0[b0<I4L4L4L4L4K5K5K4L5J6K5K5K4L4L4L4K5N2N2O0O2M3N2N5K5J6K2N3M2M4VDaKh8b4SGcKj8a4oFeKn8^4lFhKQ9\\4hFkKT9X4fFnKX9U4bFXKS:k4gE]KT:g4eEaKV:b4dEeKX:_4cEgKX:[4dElKW:W4eEmKX:U4dEnK[:T4aEoK^:S4_EoKa:S4ZEPLe:R4WEQLh:V6O100O10K4O2N101N1O2N101O001OYEYHS:f7cEeH\\:R80000O100O10000O10000O10000O10000O100O10000O10000000000001O00000000001O00000000001O1O2N1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N1O1O2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2TJlBd5[=O1O1\\MbB@_=;fBD[=7jBHX=3mBKU=OPC0Q=KTC4n<EXC:j<YOcCe0^<POmCo0U<kNPDT1S<eNSDY1[<UNjCj1i>O2N1O1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N001O1O001O1O001O1O001O1O001O1O001M4I7I7KRdZ6"}, "label": "a child beside an adult"}]}
{"id": 443455, "image": "COCO_train2014_000000443455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child\"."}], "task": "refering", "answer_template": "The \"a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 338, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 402, 403, 404, 421, 422, 423, 424, 425, 426, 443, 444, 445, 446, 447, 448, 449, 464, 465, 466, 467, 468, 469, 470, 471], "bbox": [0.07486928104575163, 0.5389705882352942, 0.44483660130718955, 1.0], "iscrowd": 0, "area": 41554.75805000001, "rle": {"size": [612, 612], "counts": "Ylk0c0[b0<I4L4L4L4L4K5K5K4L5J6K5K5K4L4L4L4K5N2N2O0O2M3N2N5K5J6K2N3M2M4VDaKh8b4SGcKj8a4oFeKn8^4lFhKQ9\\4hFkKT9X4fFnKX9U4bFXKS:k4gE]KT:g4eEaKV:b4dEeKX:_4cEgKX:[4dElKW:W4eEmKX:U4dEnK[:T4aEoK^:S4_EoKa:S4ZEPLe:R4WEQLh:V6O100O10K4O2N101N1O2N101O001OYEYHS:f7cEeH\\:R80000O100O10000O10000O10000O10000O100O10000O10000000000001O00000000001O00000000001O1O2N1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N1O1O2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2TJlBd5[=O1O1\\MbB@_=;fBD[=7jBHX=3mBKU=OPC0Q=KTC4n<EXC:j<YOcCe0^<POmCo0U<kNPDT1S<eNSDY1[<UNjCj1i>O2N1O1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N001O1O001O1O001O1O001O1O001O1O001M4I7I7KRdZ6"}, "label": "a child"}]}
{"id": 443455, "image": "COCO_train2014_000000443455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"old woman holding cake\"."}], "task": "refering", "answer_template": "The \"old woman holding cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 30, 31, 32, 33, 34, 35, 36, 52, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 404, 405, 406, 407, 408, 414, 415, 426, 427, 428, 429, 436, 437, 449, 450, 451, 458, 459, 471, 472, 473, 480, 481, 482], "bbox": [0.3149183006535948, 0.0, 0.9331045751633986, 0.9993954248366013], "iscrowd": 0, "area": 126378.6849, "rle": {"size": [612, 612], "counts": "kcc3d0\\b0j0UO=C<E;D<G9H8I5K4K6K4L4K6K4L4K5L4L3L4M2N3L4M3M2M4M3M3L3N3nDTJf8o5SGWJj8l5PGZJn8i5kF]JS9f5fF_JX9e5aFaJ]9a5]FeJS3oNX2_6^JhJU3De1g5oJjJX3J`1_5RKmJY3Hd1^7VNgHi1[7RNjHk1X7QNlHn1V7oMlHRNRNj2T9QOlHRNTNj2S9POlHRNUNm2Q9nNlHQNXNo2m8nNmHPNYNP3m8mNlHoM\\NR3j8lNlHnM_NT3g8kNlHnM`NU3f8kNmHkMbNX3c8jNUJU1m5gNWJV1l5gNWJW1k5fNWJY1j5fNXJW1j5hNXJV1i5iNYJU1h5jNYJT1i5kNYJS1h5kN[JS1f5lN\\JQ1f5nN[JQ1f5nN\\JP1e5oN]Jn0e5QO]Jm0c5SO_Jk0b5SO`Jj0c5UO_Jg0d5XO^Jd0e5[O]Ja0f5^O[J?h5@ZJ<i5CYJ9j5EVJ:m5ESJ:o5EQJ;P6DoI<S6CmI<Q7gNoHY1Q7gNoHX1R7gNnHZ1R7fNnHY1S7gNnHX1R7hNPIU1Q7kNPIT1P7lNRIQ1o6oNSIo0m6QOUIl0l6TOVIj0j6VOXIg0i6YOYId0h6\\OZIb0f6^O\\I?e6A]I=c6C^I;c6E_I4f6LY6000000000000000000000000000000000000000O1O1O1O100O1O1lNJY_O7d`06Q_OKl`0c0h^O^OUa0Z1M3M3M3N3M2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2N2O1O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000lE[Lh4f3WK`Ld4`3[KfL_4[3`KkL[4U3dKRMV4n2iKXMQ4i2nK\\Mn3d2QL]Mn3d2RL]Mm3c2RL^Mn3d2oK]MQ4e2lK]MS4e2jK\\MV4f2gK\\MY4d2dK^M\\4d2`K^M`4d2]K^Mb4d2ZK^Mf4d2WK^Mh4d2TK^Ml4c2RK_Mm4c2oJ_MP5d2mJ]MS5e2iJ^MV5d2kH[K6R2o6e2gH_K4o1T7d2dHbK4k1X7d2`HhK2f1^7d2\\HkK2c1a7d2YHnK1`1f7d4PHcKP8_4]GRLc8W7100O10O10O10O0O2M2O2N2N1O2M2O2N1O2N2M2O2N12O2M2O1O1N2O1N3N1N2OA>_NPEnJP;o4VEnJk:o4YEoJg:n4^EPKb:n4cEoJ]:P5fEnJ[:Q5gEmJY:R5kEkJU:U5mEeJX:Y5lEaJW:^5lE]JW:c5kEWJZ:g5b11O0010O00010O0001O010O0010O01O1002NhCjJc:T5PEXKT;e4iD^KY;`4cDdK`;Y4]DiKg;S4WDPLk;n3RDULQ<h3lC[LT<Z510000O1000O10O10000O100000000001O0O23M5J5L5J5L5K5J5L5J5L5J5L_DoJR9l4fFZK`9a4YFeKn9V4jEoK]:l3\\EZLk:a3nDdLY;W3_DoLg;l2RDZMU<a2cCeMc<V2VCPNQ=k1gB[N_=b37I5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K4L4L4L4L4L4L4L4L4L4L4K5L4L3M3M3M3M3M4L3M3M3M2N1O1O2N1O1OSmg0"}, "label": "old woman holding cake"}]}
{"id": 443455, "image": "COCO_train2014_000000443455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"old lady holding the cake\"."}], "task": "refering", "answer_template": "The \"old lady holding the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 30, 31, 32, 33, 34, 35, 36, 52, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 404, 405, 406, 407, 408, 414, 415, 426, 427, 428, 429, 436, 437, 449, 450, 451, 458, 459, 471, 472, 473, 480, 481, 482], "bbox": [0.3149183006535948, 0.0, 0.9331045751633986, 0.9993954248366013], "iscrowd": 0, "area": 126378.6849, "rle": {"size": [612, 612], "counts": "kcc3d0\\b0j0UO=C<E;D<G9H8I5K4K6K4L4K6K4L4K5L4L3L4M2N3L4M3M2M4M3M3L3N3nDTJf8o5SGWJj8l5PGZJn8i5kF]JS9f5fF_JX9e5aFaJ]9a5]FeJS3oNX2_6^JhJU3De1g5oJjJX3J`1_5RKmJY3Hd1^7VNgHi1[7RNjHk1X7QNlHn1V7oMlHRNRNj2T9QOlHRNTNj2S9POlHRNUNm2Q9nNlHQNXNo2m8nNmHPNYNP3m8mNlHoM\\NR3j8lNlHnM_NT3g8kNlHnM`NU3f8kNmHkMbNX3c8jNUJU1m5gNWJV1l5gNWJW1k5fNWJY1j5fNXJW1j5hNXJV1i5iNYJU1h5jNYJT1i5kNYJS1h5kN[JS1f5lN\\JQ1f5nN[JQ1f5nN\\JP1e5oN]Jn0e5QO]Jm0c5SO_Jk0b5SO`Jj0c5UO_Jg0d5XO^Jd0e5[O]Ja0f5^O[J?h5@ZJ<i5CYJ9j5EVJ:m5ESJ:o5EQJ;P6DoI<S6CmI<Q7gNoHY1Q7gNoHX1R7gNnHZ1R7fNnHY1S7gNnHX1R7hNPIU1Q7kNPIT1P7lNRIQ1o6oNSIo0m6QOUIl0l6TOVIj0j6VOXIg0i6YOYId0h6\\OZIb0f6^O\\I?e6A]I=c6C^I;c6E_I4f6LY6000000000000000000000000000000000000000O1O1O1O100O1O1lNJY_O7d`06Q_OKl`0c0h^O^OUa0Z1M3M3M3N3M2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2N2O1O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000lE[Lh4f3WK`Ld4`3[KfL_4[3`KkL[4U3dKRMV4n2iKXMQ4i2nK\\Mn3d2QL]Mn3d2RL]Mm3c2RL^Mn3d2oK]MQ4e2lK]MS4e2jK\\MV4f2gK\\MY4d2dK^M\\4d2`K^M`4d2]K^Mb4d2ZK^Mf4d2WK^Mh4d2TK^Ml4c2RK_Mm4c2oJ_MP5d2mJ]MS5e2iJ^MV5d2kH[K6R2o6e2gH_K4o1T7d2dHbK4k1X7d2`HhK2f1^7d2\\HkK2c1a7d2YHnK1`1f7d4PHcKP8_4]GRLc8W7100O10O10O10O0O2M2O2N2N1O2M2O2N1O2N2M2O2N12O2M2O1O1N2O1N3N1N2OA>_NPEnJP;o4VEnJk:o4YEoJg:n4^EPKb:n4cEoJ]:P5fEnJ[:Q5gEmJY:R5kEkJU:U5mEeJX:Y5lEaJW:^5lE]JW:c5kEWJZ:g5b11O0010O00010O0001O010O0010O01O1002NhCjJc:T5PEXKT;e4iD^KY;`4cDdK`;Y4]DiKg;S4WDPLk;n3RDULQ<h3lC[LT<Z510000O1000O10O10000O100000000001O0O23M5J5L5J5L5K5J5L5J5L5J5L_DoJR9l4fFZK`9a4YFeKn9V4jEoK]:l3\\EZLk:a3nDdLY;W3_DoLg;l2RDZMU<a2cCeMc<V2VCPNQ=k1gB[N_=b37I5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K4L4L4L4L4L4L4L4L4L4L4K5L4L3M3M3M3M3M4L3M3M3M2N1O1O2N1O1OSmg0"}, "label": "old lady holding the cake"}]}
{"id": 410687, "image": "COCO_train2014_000000410687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a white refrigerator in a kitchen next to oven and stovetop\"."}], "task": "refering", "answer_template": "The \"the side of a white refrigerator in a kitchen next to oven and stovetop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 162, 175, 176, 177, 178, 190, 191, 192, 193, 205, 206, 207, 208, 222, 223, 237, 238, 252, 253, 267, 268, 282, 297], "bbox": [0.6803747072599531, 0.46254687499999997, 0.8959250585480094, 0.840078125], "iscrowd": 0, "area": 9420.855749999999, "rle": {"size": [640, 427], "counts": "^Vf5<Yc0<C<D=O1O1O1O1OP1QOO0000000000000000001O0000000000000000001O000000000000000000001O0000000000000000001O01O00001O00001O00010O0WC_MS8a2^GmMc8S2oF[NQ9f1_FiNa9W1PFXOP:h0bEF^:;RE4n:LcDb0^;^OTDP1m;POcC_1]<R300000O100000000@a0]Ob0_Oa0_Oa0^Ob0_Oa0^Ob0_Oa0^Ob0_Oa0^Ob0_Ob0]OnVk0"}, "label": "the side of a white refrigerator in a kitchen next to oven and stovetop"}]}
{"id": 132165, "image": "COCO_train2014_000000132165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and brown sleeping dog\"."}], "task": "refering", "answer_template": "The \"a white and brown sleeping dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 219, 220, 235, 236], "bbox": [0.147234375, 0.21885416666666666, 0.7808906250000001, 0.6028125000000001], "iscrowd": 0, "area": 47641.2902, "rle": {"size": [480, 640], "counts": "cY\\11k>6J6J6J5N3L4M1O2M3N1N3N1O2M3N1N2O0O2N2O1N2_NXN_D1R1i1[:fN^E\\1^:mN[EU1b:VOUEk0i:_OiDh0U;c1O1N101O1O1O1N2O1N2O1N2N2O1N2O1N2N3N1N3M2O2M2O1N2N2O1N100O1O100000000O100000000O100000000000O10000O100O10000O2N100O2N1O1O1O2N0000000000000O2O1O1O1O2K4K5K6L3M3L4M4M2O100O1O1O100O10000001O0000001O0000000O101O1N2O1N7J2N1N2O001N101O1N10O10O100000O1000O01O1O1O100O001O1O100000O2O00000000001N01000000000O10O10000000O100000000000000000001O0O100000000000001O0000000000000000001O000000000000001N1000001O000000001O0000001O000000001O0000001O0000001O0000001O1O1O001O1O1O001O1O1O001O001O0010O01O00010O00000001N100O101O0fNmEeMU:[2oE_MS:`2SFZMn9f2VFTMl9l2XFnLj9Q3[FjLf9V3R10O010000O10O10O10000O101O0O10001N11O0001O0000001O01O01O001O00001O0010O01O001O001O0010O01O001O001O001O001O0010O0001O001O001O001O02O0O01O100O1O1O1O1O1O1O001O1O1O1O1O2N2N1O2N1O2M2O2N1O2N2N1N3N2N1O2O1N2N101N2N101N2N1O2N2N1O5L0O0O2M4M3L5L3L5Ic]Q2"}, "label": "a white and brown sleeping dog"}]}
{"id": 132165, "image": "COCO_train2014_000000132165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog sleeping in front o the other sleeping dog\"."}], "task": "refering", "answer_template": "The \"the dog sleeping in front o the other sleeping dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 219, 220, 235, 236], "bbox": [0.147234375, 0.21885416666666666, 0.7808906250000001, 0.6028125000000001], "iscrowd": 0, "area": 47641.2902, "rle": {"size": [480, 640], "counts": "cY\\11k>6J6J6J5N3L4M1O2M3N1N3N1O2M3N1N2O0O2N2O1N2_NXN_D1R1i1[:fN^E\\1^:mN[EU1b:VOUEk0i:_OiDh0U;c1O1N101O1O1O1N2O1N2O1N2N2O1N2O1N2N3N1N3M2O2M2O1N2N2O1N100O1O100000000O100000000O100000000000O10000O100O10000O2N100O2N1O1O1O2N0000000000000O2O1O1O1O2K4K5K6L3M3L4M4M2O100O1O1O100O10000001O0000001O0000000O101O1N2O1N7J2N1N2O001N101O1N10O10O100000O1000O01O1O1O100O001O1O100000O2O00000000001N01000000000O10O10000000O100000000000000000001O0O100000000000001O0000000000000000001O000000000000001N1000001O000000001O0000001O000000001O0000001O0000001O0000001O1O1O001O1O1O001O1O1O001O001O0010O01O00010O00000001N100O101O0fNmEeMU:[2oE_MS:`2SFZMn9f2VFTMl9l2XFnLj9Q3[FjLf9V3R10O010000O10O10O10000O101O0O10001N11O0001O0000001O01O01O001O00001O0010O01O001O001O0010O01O001O001O001O001O0010O0001O001O001O001O02O0O01O100O1O1O1O1O1O1O001O1O1O1O1O2N2N1O2N1O2M2O2N1O2N2N1N3N2N1O2O1N2N101N2N101N2N1O2N2N1O5L0O0O2M4M3L5L3L5Ic]Q2"}, "label": "the dog sleeping in front o the other sleeping dog"}]}
{"id": 336966, "image": "COCO_train2014_000000336966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second plane in the row from the front\"."}], "task": "refering", "answer_template": "The \"the second plane in the row from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 164, 165, 166, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201], "bbox": [0.1114375, 0.5056318681318681, 0.7749843750000001, 0.6768131868131868], "iscrowd": 0, "area": 5868.637599999999, "rle": {"size": [364, 640], "counts": "`_j01[;000000001O0000001O0000000010O00000001O00001O2N1O1O1N2O1O2N1O1O1O1O1O2N100O1O1O2N1O1O100O1O2N100O1O1O101N1O100000001O00000000001O000000000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0000000000000000000000000O100000000000001O0000000000000000001O0000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O10000000000000000000000001O000010O0001O00001O00001O001O2N2N_nb1"}, "label": "the second plane in the row from the front"}]}
{"id": 336966, "image": "COCO_train2014_000000336966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second white plane in group\"."}], "task": "refering", "answer_template": "The \"second white plane in group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 164, 165, 166, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201], "bbox": [0.1114375, 0.5056318681318681, 0.7749843750000001, 0.6768131868131868], "iscrowd": 0, "area": 5868.637599999999, "rle": {"size": [364, 640], "counts": "`_j01[;000000001O0000001O0000000010O00000001O00001O2N1O1O1N2O1O2N1O1O1O1O1O2N100O1O1O2N1O1O100O1O2N100O1O1O101N1O100000001O00000000001O000000000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0000000000000000000000000O100000000000001O0000000000000000001O0000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O10000000000000000000000001O000010O0001O00001O00001O001O2N2N_nb1"}, "label": "second white plane in group"}]}
{"id": 336966, "image": "COCO_train2014_000000336966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first of four flights which is clearly visible\"."}], "task": "refering", "answer_template": "The \"the first of four flights which is clearly visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 161, 162, 163, 164, 165, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.0025781249999999997, 0.5154945054945055, 0.944046875, 0.8465659340659341], "iscrowd": 0, "area": 32927.902449999994, "rle": {"size": [364, 640], "counts": "Pm01[;0O1M3L4M3000000000000000001O00000000000000000000000000000000000000001O0O10000000000000000000000000000000001O2N2N3M2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3N1N2N2N3N1N2N2N3N1N2N3OO000O1O2N100O1O2N100O1O2N100O1O2N100O1O2N100O1O2N100O1O2N100O1O2N100O1O2N100O1O2N100O1O2N1O0000000000000000000000O1000000000001O0000000000000000000000000000000000000000000000000000000O2O000000000000000000000000000000000000000000000000000000001O000000000O10000000000000000000000000000000000000000000001O000000000000000000000O10000000000000000000000000000000001O000000000000000000000000000000000O1000000000000000000000001O000000000000000000000000000000000000000000000O10000000001O000000000000000000000000000000000000000000000000000000001N100000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000O10O100000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001N10000000000000001O0000000O1000001O000000000000001O0O100000000000001O00000000000O101O001O001O001O1O001O1O0O2O1O001O1O001O001O1O001O1N101O1O001O1O001O001O1O001O1N101O1O001O1O001O0O2N2M2O2N2M2O2M3N1Od`<"}, "label": "the first of four flights which is clearly visible"}]}
{"id": 336966, "image": "COCO_train2014_000000336966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the 3rd farthest plane in the row of 4\"."}], "task": "refering", "answer_template": "The \"the 3rd farthest plane in the row of 4\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.20960937500000001, 0.5011263736263736, 0.7349062500000001, 0.6404395604395604], "iscrowd": 0, "area": 5046.026000000003, "rle": {"size": [364, 640], "counts": "Qca11[;0000000000000000001O00000000000000002N2N2L4M2N3L4M3O1N2N2N2N2N2O1N2N2N2N2N2O1N1O01O1O1O1N2O1O1O1O1O1O1N2O1O1O100O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OO100O100O100O10000000001O0000000000000000000000000000000000000000000000000000O1O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O0O101O0O101OPbl1"}, "label": "the 3rd farthest plane in the row of 4"}]}
{"id": 336966, "image": "COCO_train2014_000000336966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"airplane 3rd in the row\"."}], "task": "refering", "answer_template": "The \"airplane 3rd in the row\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.20960937500000001, 0.5011263736263736, 0.7349062500000001, 0.6404395604395604], "iscrowd": 0, "area": 5046.026000000003, "rle": {"size": [364, 640], "counts": "Qca11[;0000000000000000001O00000000000000002N2N2L4M2N3L4M3O1N2N2N2N2N2O1N2N2N2N2N2O1N1O01O1O1O1N2O1O1O1O1O1O1N2O1O1O100O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OO100O100O100O10000000001O0000000000000000000000000000000000000000000000000000O1O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O0O101O0O101OPbl1"}, "label": "airplane 3rd in the row"}]}
{"id": 369735, "image": "COCO_train2014_000000369735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a thick sandwich which appears to be made with bananas and bacon on whole wheat bread sits on a wooden cutting board\"."}], "task": "refering", "answer_template": "The \"a thick sandwich which appears to be made with bananas and bacon on whole wheat bread sits on a wooden cutting board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.30562091503267974, 0.6179738562091504, 0.784264705882353], "iscrowd": 0, "area": 88197.96245000002, "rle": {"size": [612, 612], "counts": "X=Y1[?`2_Mb2UOj0M4O0O2O00001O001O001N101O00001O001N101O00001O001N101O001O00000O100000000000000O1000000000000O1000000000000O100000000000001N1000000000000O10000000000000000000000000O10O1000000000O100000000000000O1000000000O10O100000000000000000000000000001O000O100000000000000000001O0000000000000O1000000000001O00000000000000000O100000000000000000000000000O10000000000000000000000000O0100000O1000000O1000000O1000O01000000O1000000O0100000O1000O10O100000O010000O10O10O10000O10O10O100O10O10O10000O10O10O10000O010O10000O01000O10000O01000O1000O0100O10000O010O100O1O010O1O100O010O1O100O10O01O100O1O010O100O1O10O01O1O1O1O001O1O1O1O1O001O1O1O1O0012N1O00O010O100O12N2M1000bKSDd1n;[NUDc1k;\\NYD`1g;`N\\D^1d;aN`D[1a;dNbDZ1^;eNeDY1[;fNiDV1X;iNkDU1T;kNoDR1R;mNREP1n:POTEm0m:ROWEk0i:UOYEi0f:XO\\Ee0e:[O]Ec0c:\\O^Ec0c:]O]Ec0c:]O]Eb0c:^O^Eb0b:^O^Eb0b:^O^ElMeNT2n;O^EjMiNU2i;1^EgMmNV2f;3QFMn93SFLn94SFKm95SFKm94TFKm95TFJk97UFHl97VFHj98VFGk99UFGk98WFGi99WFFi9;XFYOS:g0mEiNc:V1^EZNR;f1oDiMa;W2g2H7ZOf0YOg0E;O2M2O1O1O1O2N[e[4"}, "label": "a thick sandwich which appears to be made with bananas and bacon on whole wheat bread sits on a wooden cutting board"}]}
{"id": 369735, "image": "COCO_train2014_000000369735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich which is roughly triangular in shape\"."}], "task": "refering", "answer_template": "The \"a sandwich which is roughly triangular in shape\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.30562091503267974, 0.6179738562091504, 0.784264705882353], "iscrowd": 0, "area": 88197.96245000002, "rle": {"size": [612, 612], "counts": "X=Y1[?`2_Mb2UOj0M4O0O2O00001O001O001N101O00001O001N101O00001O001N101O001O00000O100000000000000O1000000000000O1000000000000O100000000000001N1000000000000O10000000000000000000000000O10O1000000000O100000000000000O1000000000O10O100000000000000000000000000001O000O100000000000000000001O0000000000000O1000000000001O00000000000000000O100000000000000000000000000O10000000000000000000000000O0100000O1000000O1000000O1000O01000000O1000000O0100000O1000O10O100000O010000O10O10O10000O10O10O100O10O10O10000O10O10O10000O010O10000O01000O10000O01000O1000O0100O10000O010O100O1O010O1O100O010O1O100O10O01O100O1O010O100O1O10O01O1O1O1O001O1O1O1O1O001O1O1O1O0012N1O00O010O100O12N2M1000bKSDd1n;[NUDc1k;\\NYD`1g;`N\\D^1d;aN`D[1a;dNbDZ1^;eNeDY1[;fNiDV1X;iNkDU1T;kNoDR1R;mNREP1n:POTEm0m:ROWEk0i:UOYEi0f:XO\\Ee0e:[O]Ec0c:\\O^Ec0c:]O]Ec0c:]O]Eb0c:^O^Eb0b:^O^Eb0b:^O^ElMeNT2n;O^EjMiNU2i;1^EgMmNV2f;3QFMn93SFLn94SFKm95SFKm94TFKm95TFJk97UFHl97VFHj98VFGk99UFGk98WFGi99WFFi9;XFYOS:g0mEiNc:V1^EZNR;f1oDiMa;W2g2H7ZOf0YOg0E;O2M2O1O1O1O2N[e[4"}, "label": "a sandwich which is roughly triangular in shape"}]}
{"id": 369735, "image": "COCO_train2014_000000369735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwhich behind other sandwhich\"."}], "task": "refering", "answer_template": "The \"sandwhich behind other sandwhich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305], "bbox": [0.4803921568627451, 0.2213562091503268, 0.9420751633986927, 0.6057679738562092], "iscrowd": 0, "area": 49556.81575, "rle": {"size": [612, 612], "counts": "Tl_5<Yb0b0L3M3M3M4L3M3M2N001O1O1O001O1O1000O100000O010000000O01000000O10O1\\A\\Nh;d1WD_Ng;a1XDbNf;]1ZDgNc;Y1\\DjNb;V1]DnN_;S1_DQO_;n0aDVO\\;j0cDYO[;g0dD\\OZ;d0eD@W;a0hDBV;=jDGS;9mDHR;8mDJR;6nDJR;5nDLQ;5oDKQ;5oDLP;4oDMQ;2PENP;2oDOQ;1oDOP;2oD0lMaNf<^1^Ea0a:^O`Eb0`:^O_Ed0`:\\O`Ed0`:\\O_Ef0`:YOaEg0_:YO`Ei0_:WOaEj0^:VOaEk0_:UOaEl0^:SOcEm0^:ROaEP1^:PObEP1^:POaER1^:mNcEU1[:lNcEZ1X:fNZDm2a;SM]DT3^;mL_DW3_;iL_DZ3`;gL]D[3d;dLZD_3e;aLYDa3g;`LVDc3i;]LUDf3j;[LSDg3m;YLQDj3n;f10000001O000000000000000000001O0000000000000000001O0000000000000000000000001O000000000000000000000O10001O000000000000000000000000001O0000000000000001O0000000001O0000000000000000001O000001O0001O0000000000001O00000000001O0001O00000001O00000000001O00000000001O000001O000001O00000000001O0000000000001O0001O000001O000O1O100O1O2O0O1O100O1O2O0O1O1O100O2L3N2M4M3VMjBBZ=;gBB]=;fBB]=<dBB_=;cBBa=<`BBc=<_BAe=;]BBg=<ZBBm=7VBER>5oAIX>0jAM]>NdAMf>M\\AOl>JVA3R?GQA4W?Dn@7Y?^OPA=X?WOQAc0m`0H7InWe0"}, "label": "sandwhich behind other sandwhich"}]}
{"id": 369735, "image": "COCO_train2014_000000369735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"end view of sandwich with banana on wheat grain bread\"."}], "task": "refering", "answer_template": "The \"end view of sandwich with banana on wheat grain bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305], "bbox": [0.4803921568627451, 0.2213562091503268, 0.9420751633986927, 0.6057679738562092], "iscrowd": 0, "area": 49556.81575, "rle": {"size": [612, 612], "counts": "Tl_5<Yb0b0L3M3M3M4L3M3M2N001O1O1O001O1O1000O100000O010000000O01000000O10O1\\A\\Nh;d1WD_Ng;a1XDbNf;]1ZDgNc;Y1\\DjNb;V1]DnN_;S1_DQO_;n0aDVO\\;j0cDYO[;g0dD\\OZ;d0eD@W;a0hDBV;=jDGS;9mDHR;8mDJR;6nDJR;5nDLQ;5oDKQ;5oDLP;4oDMQ;2PENP;2oDOQ;1oDOP;2oD0lMaNf<^1^Ea0a:^O`Eb0`:^O_Ed0`:\\O`Ed0`:\\O_Ef0`:YOaEg0_:YO`Ei0_:WOaEj0^:VOaEk0_:UOaEl0^:SOcEm0^:ROaEP1^:PObEP1^:POaER1^:mNcEU1[:lNcEZ1X:fNZDm2a;SM]DT3^;mL_DW3_;iL_DZ3`;gL]D[3d;dLZD_3e;aLYDa3g;`LVDc3i;]LUDf3j;[LSDg3m;YLQDj3n;f10000001O000000000000000000001O0000000000000000001O0000000000000000000000001O000000000000000000000O10001O000000000000000000000000001O0000000000000001O0000000001O0000000000000000001O000001O0001O0000000000001O00000000001O0001O00000001O00000000001O00000000001O000001O000001O00000000001O0000000000001O0001O000001O000O1O100O1O2O0O1O100O1O2O0O1O1O100O2L3N2M4M3VMjBBZ=;gBB]=;fBB]=<dBB_=;cBBa=<`BBc=<_BAe=;]BBg=<ZBBm=7VBER>5oAIX>0jAM]>NdAMf>M\\AOl>JVA3R?GQA4W?Dn@7Y?^OPA=X?WOQAc0m`0H7InWe0"}, "label": "end view of sandwich with banana on wheat grain bread"}]}
{"id": 369735, "image": "COCO_train2014_000000369735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cut up banana on the first half of sandwich on the left\"."}], "task": "refering", "answer_template": "The \"the cut up banana on the first half of sandwich on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.004477124183006536, 0.4643790849673202, 0.5629901960784314, 0.6147549019607843], "iscrowd": 0, "area": 15379.268449999998, "rle": {"size": [612, 612], "counts": "PS2]1ha0O0000000000001O000000000000000001O000000000000000001O000000000000000000000000000001O0001O000000001O000000001O01O000001O0000000010O00000001O000001O01O000000001O0001O0001O000000001O01O00000000000000001O0001O0000000001O00000000000000001O01O00000000000001O000000000000000010O0000000000000001O0000000000000010O0000000000000001O00000000000001O01O000000000000000000000000001O01O0000000000000000000000000001O0001O000000000000000000000001O00000000000000000O10000000001O000O100000000000000O100000001O0000000O100000000000000O10001O000000000O1000000O100O100O1O101N100O100O100O1O100OO2O0010O01O0010O0000001O01O000001O0001O01O00000010O000001O0001O01O00000003N2M4L4L3N3L3M4L3N3L3M4L3N3L3M4L4M2MXko4"}, "label": "the cut up banana on the first half of sandwich on the left"}]}
{"id": 140360, "image": "COCO_train2014_000000140360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bowl of cat food that is shaped like a cat\"."}], "task": "refering", "answer_template": "The \"a white bowl of cat food that is shaped like a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337], "bbox": [0.471015625, 0.7528, 0.7023125, 1.0], "iscrowd": 0, "area": 11553.530350000006, "rle": {"size": [425, 640], "counts": "gYm32T==C5K5L4M2N3M3M3L4M2N3M1O1O2N1O1O2N1N3N1O1O2N1O2N1O2N1O2M2O1O2N1O2N1O1O1O1O1N2O1O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O1O1O1N102N1O1O1O2M2O1O1O1O2N1N2O1O2N1O1O2N3L4M2ROo0K4Lil^2"}, "label": "a white bowl of cat food that is shaped like a cat"}]}
{"id": 140360, "image": "COCO_train2014_000000140360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl of food\"."}], "task": "refering", "answer_template": "The \"the bowl of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337], "bbox": [0.471015625, 0.7528, 0.7023125, 1.0], "iscrowd": 0, "area": 11553.530350000006, "rle": {"size": [425, 640], "counts": "gYm32T==C5K5L4M2N3M3M3L4M2N3M1O1O2N1O1O2N1N3N1O1O2N1O2N1O2N1O2M2O1O2N1O2N1O1O1O1O1N2O1O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O1O1O1N102N1O1O1O2M2O1O1O1O2N1N2O1O2N1O1O2N3L4M2ROo0K4Lil^2"}, "label": "the bowl of food"}]}
{"id": 435272, "image": "COCO_train2014_000000435272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown cow in the forefront of the picture\"."}], "task": "refering", "answer_template": "The \"the brown cow in the forefront of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.390265625, 0.06667447306791568, 0.998625, 0.9892505854800936], "iscrowd": 0, "area": 104329.85949999999, "rle": {"size": [427, 640], "counts": "k_X35R=7I6L5J6J6G8I7M2N3M2N2N3M2N2N5K2O0O2SKkMaNV2^1RN[No1c1SN[Nn1d1TN[Nl1d1VNZNl1e1TNZNm1e1UNYNm1f1TNYNl1f1VNXNl1f1VNXNk1h1VNVNl1h1VNVNk1i1WNUNj1k1WNSNk1k1WNTNi1l1XNRNi1m1YNPNj1o1VNoMl1Q2UNlMn1R2TNkMn1U2SNhMo1W2SNfMP2Y2QNcMR2]2oM`MS2_2oM]MU2b2kM[MX2d2jMXMY2h2hMTM[2k2gMQM]2n2cMXL]NPOW4g4_M[KP4e4QLWKR4i4oKRKU4o4kKmJX4S5iKiJZ4W5n1O000001O0000010O001N1O2N1O2N1O1O2N101O0O2O1O1O001O1O001O1O000000O100O100O10000O100O10000O010O10000O1O1O1O100O1O001O1O1O001O1O0001O001O001O001O001O0101O001O0O101O001O00001N101O000O10O1O1O001O2N100O12N2N1O2N1O1O1O1O001O100O01000fNiG_MV8_2mG`MT8]2oGcMP8[2SHeMm7Y2THhMk7V2WHkMi7R2YHnMg7P2[HQNe7l1]HUNb7j1_HVNb7g1`HZNa7b1aH^Nb7]1`HdNb7X1_HhNb7W1^HjNb7T1_HlNb7R1_HnNc7P1]HQOc7m0^HSOc7l0]HTOd7i0^HWOc7f0_HZOc7a0`H_Ob7=`HDb77`HIb73`HMc7N_H2c7K^H5c7H_H8b7E`H<`7C`H=a7A`H?a7_O`Ha0b7\\O_Hd0c7YO^Hg0d7WO\\Hi0f7TO[Hl0g7QOZHo0i7mNXHT1i7iNXHW1j7eNXH[1l90000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N1O2N1O1O1O1O1O1O1O001O1O1O1O1O2N1O1O1O1O1O1O001O001O1O001O1O001O001O00001O001O00001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O0O2O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000000000Y9"}, "label": "the brown cow in the forefront of the picture"}]}
{"id": 435272, "image": "COCO_train2014_000000435272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby cow in front of picture\"."}], "task": "refering", "answer_template": "The \"baby cow in front of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.390265625, 0.06667447306791568, 0.998625, 0.9892505854800936], "iscrowd": 0, "area": 104329.85949999999, "rle": {"size": [427, 640], "counts": "k_X35R=7I6L5J6J6G8I7M2N3M2N2N3M2N2N5K2O0O2SKkMaNV2^1RN[No1c1SN[Nn1d1TN[Nl1d1VNZNl1e1TNZNm1e1UNYNm1f1TNYNl1f1VNXNl1f1VNXNk1h1VNVNl1h1VNVNk1i1WNUNj1k1WNSNk1k1WNTNi1l1XNRNi1m1YNPNj1o1VNoMl1Q2UNlMn1R2TNkMn1U2SNhMo1W2SNfMP2Y2QNcMR2]2oM`MS2_2oM]MU2b2kM[MX2d2jMXMY2h2hMTM[2k2gMQM]2n2cMXL]NPOW4g4_M[KP4e4QLWKR4i4oKRKU4o4kKmJX4S5iKiJZ4W5n1O000001O0000010O001N1O2N1O2N1O1O2N101O0O2O1O1O001O1O001O1O000000O100O100O10000O100O10000O010O10000O1O1O1O100O1O001O1O1O001O1O0001O001O001O001O001O0101O001O0O101O001O00001N101O000O10O1O1O001O2N100O12N2N1O2N1O1O1O1O001O100O01000fNiG_MV8_2mG`MT8]2oGcMP8[2SHeMm7Y2THhMk7V2WHkMi7R2YHnMg7P2[HQNe7l1]HUNb7j1_HVNb7g1`HZNa7b1aH^Nb7]1`HdNb7X1_HhNb7W1^HjNb7T1_HlNb7R1_HnNc7P1]HQOc7m0^HSOc7l0]HTOd7i0^HWOc7f0_HZOc7a0`H_Ob7=`HDb77`HIb73`HMc7N_H2c7K^H5c7H_H8b7E`H<`7C`H=a7A`H?a7_O`Ha0b7\\O_Hd0c7YO^Hg0d7WO\\Hi0f7TO[Hl0g7QOZHo0i7mNXHT1i7iNXHW1j7eNXH[1l90000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N1O2N1O1O1O1O1O1O1O001O1O1O1O1O2N1O1O1O1O1O1O001O001O1O001O1O001O001O00001O001O00001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O0O2O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000000000Y9"}, "label": "baby cow in front of picture"}]}
{"id": 435272, "image": "COCO_train2014_000000435272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow looking at the camera that is sitting behind another cow\"."}], "task": "refering", "answer_template": "The \"a cow looking at the camera that is sitting behind another cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288], "bbox": [0.11368750000000001, 0.04395784543325527, 0.695609375, 0.8376580796252927], "iscrowd": 0, "area": 55044.522450000004, "rle": {"size": [427, 640], "counts": "T_n09R==C2N2N1N3N1O1O2N1O2N1N2O2N1O2N1O1O2M2O2N1O2N1O1O2M3O0O2O0O2O1N1O2O1N101N101N2O1N4M2M3N3L3M4M2M3kGmLm5V3kInLU6U3dIPM\\6R3]ISMb6Q3VITMj6Q3lHSMT7S3bHRM^7R3YHSMf7Z4O010O1O0N3N2M3N1N3N2M2O2M3O1O001N2O001O0O101O000fKhIa1Y6nMXJP2i5dMcJ\\2]5aMfJ_2[5^MhJa2X5\\MkJc2W5ZMkJf2U5WMoJh2Q5VMQKj2P5RMSKm2n4QMTKo2l4nLXKQ3i4lLYKT3g4iL\\KW3g4dL\\KZ3i4_LZKa3j4YLXKg3W7O0O10001O000O2O00001N1000001O0O101O000O2O00O100O2O0O101N100O2O0O2O0O101N101N11O2O1N1O2O1N101N1O2O1N1O2O1N101N1O01O10O01O1O001O10O01O001O10O01O1O001O10O01O0jL^LcLc3W3eLfL\\3T3kLjLU3Q3SMmLm2m2ZMQMg2l2^MQMb2n2aMPM_2o2dMoL[2Q3iMmLU2S3nMkLR2T3RNiLm1X3UNfLj1Z3WNfLi1Y3YNfLf1Z3[NgLc1Y3_NeLb1Z3_NfL`1Z3bNeL]1[3dNeL\\1[3eNdLZ1\\3gNdLX1gNfLe3T2bMV1gNjLe3Q2dMU1cNoLg3n1eMR1aNUMh3j1gMQ1]NYMk3g1hMP1YN^Mm3d1iMm0ZNaMk3c1jMm0[NaMi3d1kMk0\\NbMh3c1lMk0\\NcMf3d1mMh0]NfMd3c1oMg0]NgMb3d1PNe0^NhM`3d1QNe0_NhM^3e1RNb0aNjM[3e1TNa0aNkMY3f1UN?aNlMY3f1VN>aNmMV3g1YN;bNoMR3i1ZN9dNoMo2j1]N7dNPNl2l1_N4dNRNi2m1cN0eNTNe2o1eNMfNVN^2R2lNHfNWNV2Y2ROAhNXNm1^2[OZOhNYNd1f2CPOiN]N[1j2LiNiN^No0U37]NjN`N`0`3f0PNjNaN3k3R1dMlNcNCW4`1VMmNh4S1XKmNi4R1WKnNi4R1VKnNl4Q1UKnNk4R1UKnNl4P1UKPOk4P1VKoNk4P1UKPOk4P1UKPOj4Q1WKnNi4R1WKnNi4Q1XKnNh4S1YKlNg4T1YKlNf4U1ZKkNf4U1[KjNd4W1\\KiNd4V1]KjNb4W1_KiN`4W1`KjN_4V1aKjN^4W1cKiN\\4W1dKjNZ4V1gKjNY4V1hKjNV4W1jKjNU4V1kKjNU4V1lKjNR4W1nKiNR4V1oKkNo3V1RLjNm3V1SLjNm3V1SLkNl3U1ULkNj3U1VLkNj3T1WLmNh3S1YLmNf3S1ZLmNe3T1[LlNe3T1\\LkNd3U1\\LjNe3U1]LjNc3V1]LjNc3V1^LhNc3X1^LgNb3Y1`LeN`3[1bLcN^3\\1dLbN\\3_1X400000000O100000O1000O10000000000O01O1O1O1O1O001O1O1O1O100O01000001O000O10000000001N10000000000O10001O000000001O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N1N3N1O2N1O2N1O2N1O2N2N1O2N1O2Neo`2"}, "label": "a cow looking at the camera that is sitting behind another cow"}]}
{"id": 435272, "image": "COCO_train2014_000000435272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"animals next to eachother with the one on the left looking at the camera\"."}], "task": "refering", "answer_template": "The \"animals next to eachother with the one on the left looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288], "bbox": [0.11368750000000001, 0.04395784543325527, 0.695609375, 0.8376580796252927], "iscrowd": 0, "area": 55044.522450000004, "rle": {"size": [427, 640], "counts": "T_n09R==C2N2N1N3N1O1O2N1O2N1N2O2N1O2N1O1O2M2O2N1O2N1O1O2M3O0O2O0O2O1N1O2O1N101N101N2O1N4M2M3N3L3M4M2M3kGmLm5V3kInLU6U3dIPM\\6R3]ISMb6Q3VITMj6Q3lHSMT7S3bHRM^7R3YHSMf7Z4O010O1O0N3N2M3N1N3N2M2O2M3O1O001N2O001O0O101O000fKhIa1Y6nMXJP2i5dMcJ\\2]5aMfJ_2[5^MhJa2X5\\MkJc2W5ZMkJf2U5WMoJh2Q5VMQKj2P5RMSKm2n4QMTKo2l4nLXKQ3i4lLYKT3g4iL\\KW3g4dL\\KZ3i4_LZKa3j4YLXKg3W7O0O10001O000O2O00001N1000001O0O101O000O2O00O100O2O0O101N100O2O0O2O0O101N101N11O2O1N1O2O1N101N1O2O1N1O2O1N101N1O01O10O01O1O001O10O01O001O10O01O1O001O10O01O0jL^LcLc3W3eLfL\\3T3kLjLU3Q3SMmLm2m2ZMQMg2l2^MQMb2n2aMPM_2o2dMoL[2Q3iMmLU2S3nMkLR2T3RNiLm1X3UNfLj1Z3WNfLi1Y3YNfLf1Z3[NgLc1Y3_NeLb1Z3_NfL`1Z3bNeL]1[3dNeL\\1[3eNdLZ1\\3gNdLX1gNfLe3T2bMV1gNjLe3Q2dMU1cNoLg3n1eMR1aNUMh3j1gMQ1]NYMk3g1hMP1YN^Mm3d1iMm0ZNaMk3c1jMm0[NaMi3d1kMk0\\NbMh3c1lMk0\\NcMf3d1mMh0]NfMd3c1oMg0]NgMb3d1PNe0^NhM`3d1QNe0_NhM^3e1RNb0aNjM[3e1TNa0aNkMY3f1UN?aNlMY3f1VN>aNmMV3g1YN;bNoMR3i1ZN9dNoMo2j1]N7dNPNl2l1_N4dNRNi2m1cN0eNTNe2o1eNMfNVN^2R2lNHfNWNV2Y2ROAhNXNm1^2[OZOhNYNd1f2CPOiN]N[1j2LiNiN^No0U37]NjN`N`0`3f0PNjNaN3k3R1dMlNcNCW4`1VMmNh4S1XKmNi4R1WKnNi4R1VKnNl4Q1UKnNk4R1UKnNl4P1UKPOk4P1VKoNk4P1UKPOk4P1UKPOj4Q1WKnNi4R1WKnNi4Q1XKnNh4S1YKlNg4T1YKlNf4U1ZKkNf4U1[KjNd4W1\\KiNd4V1]KjNb4W1_KiN`4W1`KjN_4V1aKjN^4W1cKiN\\4W1dKjNZ4V1gKjNY4V1hKjNV4W1jKjNU4V1kKjNU4V1lKjNR4W1nKiNR4V1oKkNo3V1RLjNm3V1SLjNm3V1SLkNl3U1ULkNj3U1VLkNj3T1WLmNh3S1YLmNf3S1ZLmNe3T1[LlNe3T1\\LkNd3U1\\LjNe3U1]LjNc3V1]LjNc3V1^LhNc3X1^LgNb3Y1`LeN`3[1bLcN^3\\1dLbN\\3_1X400000000O100000O1000O10000000000O01O1O1O1O1O001O1O1O1O100O01000001O000O10000000001N10000000000O10001O000000001O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N1N3N1O2N1O2N1O2N1O2N2N1O2N1O2Neo`2"}, "label": "animals next to eachother with the one on the left looking at the camera"}]}
{"id": 435272, "image": "COCO_train2014_000000435272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an out of focus body of a cow behind two other cows\"."}], "task": "refering", "answer_template": "The \"an out of focus body of a cow behind two other cows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212], "bbox": [0.002984375, 0.00468384074941452, 0.26568749999999997, 0.6243793911007026], "iscrowd": 0, "area": 28525.65865, "rle": {"size": [427, 640], "counts": "nj0S8X5000000000O10O10000000O10000000000O10000000000O10000000000O10000000000O1000000001O000000001O000000001O000000001O0000001O000000001kHiJT6W5lIkJR6U5mIoJP6Q5PJQKn5o4QJTKm5m4RJVKk5j4TJYKj5g4VJ[Kh5e4XJ^Kd5d4ZJ_Kd5a4\\JaKb5_4]JdKa5\\4_JeK`5[4_JhK_5Y4`JhK_5X4aJjK]5V4bJlK]5T4cJnK[5R4dJPL[5Q4dJQLZ5Q4cJRL[5P4cJQL\\5Q4bJQL\\5Q4aJQL^5Q4`JQL^5Q4_JQL`5Q4^JQL`5n5O2N1O2N1O2N1O2N1O2N2N1O2N1O3M3M4L4L4L3M4L4L3M4L4L3M4L3M4XNPFMS:MPF2U:GoE8T:APF?S:[OQFd0S:TORFk0Q:oNRFP1R:jNRFU1S;O001N101O1O001N101O1O0O2O1O001O0O2O1O001O0O2O1O001N2O001O001N2O001O2M2O2N2N2N1NfgS6"}, "label": "an out of focus body of a cow behind two other cows"}]}
{"id": 435272, "image": "COCO_train2014_000000435272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow whose only body is seen not mouth\"."}], "task": "refering", "answer_template": "The \"the cow whose only body is seen not mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212], "bbox": [0.002984375, 0.00468384074941452, 0.26568749999999997, 0.6243793911007026], "iscrowd": 0, "area": 28525.65865, "rle": {"size": [427, 640], "counts": "nj0S8X5000000000O10O10000000O10000000000O10000000000O10000000000O10000000000O1000000001O000000001O000000001O000000001O0000001O000000001kHiJT6W5lIkJR6U5mIoJP6Q5PJQKn5o4QJTKm5m4RJVKk5j4TJYKj5g4VJ[Kh5e4XJ^Kd5d4ZJ_Kd5a4\\JaKb5_4]JdKa5\\4_JeK`5[4_JhK_5Y4`JhK_5X4aJjK]5V4bJlK]5T4cJnK[5R4dJPL[5Q4dJQLZ5Q4cJRL[5P4cJQL\\5Q4bJQL\\5Q4aJQL^5Q4`JQL^5Q4_JQL`5Q4^JQL`5n5O2N1O2N1O2N1O2N1O2N2N1O2N1O3M3M4L4L4L3M4L4L3M4L4L3M4L3M4XNPFMS:MPF2U:GoE8T:APF?S:[OQFd0S:TORFk0Q:oNRFP1R:jNRFU1S;O001N101O1O001N101O1O0O2O1O001O0O2O1O001O0O2O1O001N2O001O001N2O001O2M2O2N2N2N1NfgS6"}, "label": "the cow whose only body is seen not mouth"}]}
{"id": 377926, "image": "COCO_train2014_000000377926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a black shirt eating a sausage with both hands\"."}], "task": "refering", "answer_template": "The \"a person in a black shirt eating a sausage with both hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 134], "bbox": [0.575609375, 0.0, 1.0, 0.40225], "iscrowd": 0, "area": 28642.37075, "rle": {"size": [360, 640], "counts": "]^Q42R;4K5L4K5L4L4L5L4M2M4L3N3L4L3M4M3L3M4M3L4L4L4M4M2O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O001O1O1O001O1O1O001O1O0000000000O10000000O2O001O1O1O1O1O1O1O001O1O0000000000000000000000000000SOmHeMS7Z2oHeMQ7[2PIdMP7\\2QIcMo6]2RIbMn6^2RIbMn6^2RIbMn6^2RIbMn6^2RIcMm6]2SIcMm6]2SIcMm6]2SIcMm6]2SIcMm6]2SIWMLMQ7l2SIVMMNP7l2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMOLn6n2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2RIWMNMP7l2RIWMNMP7l2RIVMONo6l2RIVMONo6l2RIVMNOP7k2RIVMNOP7k2QIWMONP7l2PIVM0NP7l2PIVM0MQ7n1oHdN0A0MQ7m1RIcNMB2LP7o1RIbNLC2KQ7o1]IXN_ON4KP7n1^IYN^ON4JQ7o1]IYN^ON5IP7o1^IZN]ON5GR7P2]I[N\\OM6FS7R2[I[N\\OMb7g1SI\\NZONc7e1TI]NYONc7d1UI^NXOMd7d1UI_NWOLe7d1UI`NVOLe7b1WIbNVOId7d1WIcNWOFc7f1WIdNXOCb7i1VIdNR7\\1nHdNR7\\1nHdNR7\\1nHdNR7\\1nHdNR7\\1nHdNQ7]1oHcNQ7]1PIaNQ7_1PI_NROJh7g1WI]NoN3f7`1bIfNZ6[1^IPO^6P1bISO[6m0dIWOY6i0fIYOY6f0hI[OW6e0iI]OU6b0lI_OS6a0mI@R6?oICo5=RJCm5=SJDl5;UJFj5:VJGi58ZJFf5:\\JEc5:`JD`5<aJD^5;eJDZ5<TJ]NDU1X6=SJfN@m0\\6<TJiN@j0\\6<UJlN^Oh0\\6<VJmN^Oe0]6>UJmN_Od0\\6`0SJnN@a0^6a0RJnNA`0]6b0RJoNA=^6d0QJoNB<]6f0PJnNC;^6g0oIoNC9^6h0PJnNC8^6j0oIoNC6^6l0nInND5_6m0mInNE4^6o0mImND3`6P1lImND3`6Q1kImND2a6Q1lIlNC3a6R1kIlNB2d6R1kIkNA3d6Q1lImN_O2e6Q1lImN_O1f6R1lImN\\O2h6Q1lImN[O3i6Q1lIlNYO3l6S1iIjNZO4m6S1hIiNZO5n6T1gIgNYO5Q7U1eIfNYO6R7V1dI]O]6c0cI]O]6d0bI[O_6e0`I[Oa6f0^IYOc6h0[IYOe6h0YIXOh6k0UITOl6o0PIQOQ7S1jHnNV7R1jHnNV7S1hHnNX7R1hHmNY7a2000O1O1O1O1O1O1000000000000000O10000000O100O100O100O100O10000O100O10000O10000O10000O10000O100O100O1O100O100O100O1"}, "label": "a person in a black shirt eating a sausage with both hands"}]}
{"id": 377926, "image": "COCO_train2014_000000377926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding a piece of pepperoni pizza\"."}], "task": "refering", "answer_template": "The \"man holding a piece of pepperoni pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 134], "bbox": [0.575609375, 0.0, 1.0, 0.40225], "iscrowd": 0, "area": 28642.37075, "rle": {"size": [360, 640], "counts": "]^Q42R;4K5L4K5L4L4L5L4M2M4L3N3L4L3M4M3L3M4M3L4L4L4M4M2O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O001O1O1O001O1O1O001O1O0000000000O10000000O2O001O1O1O1O1O1O1O001O1O0000000000000000000000000000SOmHeMS7Z2oHeMQ7[2PIdMP7\\2QIcMo6]2RIbMn6^2RIbMn6^2RIbMn6^2RIbMn6^2RIcMm6]2SIcMm6]2SIcMm6]2SIcMm6]2SIcMm6]2SIWMLMQ7l2SIVMMNP7l2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMOLn6n2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2SIVMNMo6m2RIWMNMP7l2RIWMNMP7l2RIVMONo6l2RIVMONo6l2RIVMNOP7k2RIVMNOP7k2QIWMONP7l2PIVM0NP7l2PIVM0MQ7n1oHdN0A0MQ7m1RIcNMB2LP7o1RIbNLC2KQ7o1]IXN_ON4KP7n1^IYN^ON4JQ7o1]IYN^ON5IP7o1^IZN]ON5GR7P2]I[N\\OM6FS7R2[I[N\\OMb7g1SI\\NZONc7e1TI]NYONc7d1UI^NXOMd7d1UI_NWOLe7d1UI`NVOLe7b1WIbNVOId7d1WIcNWOFc7f1WIdNXOCb7i1VIdNR7\\1nHdNR7\\1nHdNR7\\1nHdNR7\\1nHdNR7\\1nHdNQ7]1oHcNQ7]1PIaNQ7_1PI_NROJh7g1WI]NoN3f7`1bIfNZ6[1^IPO^6P1bISO[6m0dIWOY6i0fIYOY6f0hI[OW6e0iI]OU6b0lI_OS6a0mI@R6?oICo5=RJCm5=SJDl5;UJFj5:VJGi58ZJFf5:\\JEc5:`JD`5<aJD^5;eJDZ5<TJ]NDU1X6=SJfN@m0\\6<TJiN@j0\\6<UJlN^Oh0\\6<VJmN^Oe0]6>UJmN_Od0\\6`0SJnN@a0^6a0RJnNA`0]6b0RJoNA=^6d0QJoNB<]6f0PJnNC;^6g0oIoNC9^6h0PJnNC8^6j0oIoNC6^6l0nInND5_6m0mInNE4^6o0mImND3`6P1lImND3`6Q1kImND2a6Q1lIlNC3a6R1kIlNB2d6R1kIkNA3d6Q1lImN_O2e6Q1lImN_O1f6R1lImN\\O2h6Q1lImN[O3i6Q1lIlNYO3l6S1iIjNZO4m6S1hIiNZO5n6T1gIgNYO5Q7U1eIfNYO6R7V1dI]O]6c0cI]O]6d0bI[O_6e0`I[Oa6f0^IYOc6h0[IYOe6h0YIXOh6k0UITOl6o0PIQOQ7S1jHnNV7R1jHnNV7S1hHnNX7R1hHmNY7a2000O1O1O1O1O1O1000000000000000O10000000O100O100O100O100O10000O100O10000O10000O10000O10000O100O100O1O100O100O100O1"}, "label": "man holding a piece of pepperoni pizza"}]}
{"id": 377926, "image": "COCO_train2014_000000377926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 links of meat on a plate\"."}], "task": "refering", "answer_template": "The \"2 links of meat on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212], "bbox": [0.096421875, 0.4805277777777778, 0.29075, 0.7532500000000001], "iscrowd": 0, "area": 8599.3166, "rle": {"size": [360, 640], "counts": "bPf0;l:6J6I4M2N3L3N2N2M3N1O2N1O2N1O2N2N101N1O2N1O2N101N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N100O101O000O10O100O10O10O100O10O010000O10O010000O100O100O10000O100O10000O100O100O1O010O1O100O1O1O100O1O100O1O100O1O1N2O1N2O2M2O1N2O2M2O1N2N3N1UO\\F0f9M^F0d9N`FOb9NbFNa9NcFOb9LbF1\\:MP]o4"}, "label": "2 links of meat on a plate"}]}
{"id": 377926, "image": "COCO_train2014_000000377926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of bread holds slaw , hot dog and sauce\"."}], "task": "refering", "answer_template": "The \"a piece of bread holds slaw , hot dog and sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287], "bbox": [0.275296875, 0.22133333333333335, 0.800734375, 0.9813333333333334], "iscrowd": 0, "area": 53410.3427, "rle": {"size": [360, 640], "counts": "_]n12U;2L4M3M3M3M3M3L4M3M3M3N3N1N2O1N2O1N2O1N2O1O1N3N1N2O1N2O1N2O1O1N2O1N3N1N2O1N2O1N2O1O1N2O1N2O2M2O1N2O1YOSMQIn2l6YMnHi2P7\\MlHe2Q7bMiH`2U7k0M6J9H7H9H8G8I8G9G:G8H9G8I8H1N10000O100O10001N10000O10000O100O10000O2O000O11O00000000000010O000000000001O000O100O1O101N100O1O100O1O2O0O1O100O1O2O0O1O100O1O101N1O100O10O1O2N1N2O2N1N3N1O1N3N1O0O101O0O2O000O2O000O2O01O100O1O100O1O100O1O100O100O1O2O0O1O100O1O100O1O100O1O100O100O1O2O0O1O100O1O100O1O100O1O2O0O2O1N2N2O0O2N2O1N2N101N2N2O1N1O3N1N2O1N2N3N1N2N2O1N2N3N1N2N2O1N3M2O1N2O1N3M2O1N2N2O2M2N2O1N2N3N1N2N2O1N3N1N2N2O1N1O2O1N1O2O1N1O2O1N1O2O1N2O0O2N2O0O2N2O0O2N2O0O2N2O0O2N2O1N1O2O1N1O2O1N1O2O1N1O2N2O1N1O2N2O0O2N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2M3M3M3M3M3M3M3M3M3Mji\\1"}, "label": "a piece of bread holds slaw , hot dog and sauce"}]}
{"id": 377926, "image": "COCO_train2014_000000377926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slightly burnt sausage on a piece of wheat bread\"."}], "task": "refering", "answer_template": "The \"a slightly burnt sausage on a piece of wheat bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287], "bbox": [0.275296875, 0.22133333333333335, 0.800734375, 0.9813333333333334], "iscrowd": 0, "area": 53410.3427, "rle": {"size": [360, 640], "counts": "_]n12U;2L4M3M3M3M3M3L4M3M3M3N3N1N2O1N2O1N2O1N2O1O1N3N1N2O1N2O1N2O1O1N2O1N3N1N2O1N2O1N2O1O1N2O1N2O2M2O1N2O1YOSMQIn2l6YMnHi2P7\\MlHe2Q7bMiH`2U7k0M6J9H7H9H8G8I8G9G:G8H9G8I8H1N10000O100O10001N10000O10000O100O10000O2O000O11O00000000000010O000000000001O000O100O1O101N100O1O100O1O2O0O1O100O1O2O0O1O100O1O101N1O100O10O1O2N1N2O2N1N3N1O1N3N1O0O101O0O2O000O2O000O2O01O100O1O100O1O100O1O100O100O1O2O0O1O100O1O100O1O100O1O100O100O1O2O0O1O100O1O100O1O100O1O2O0O2O1N2N2O0O2N2O1N2N101N2N2O1N1O3N1N2O1N2N3N1N2N2O1N2N3N1N2N2O1N3M2O1N2O1N3M2O1N2N2O2M2N2O1N2N3N1N2N2O1N3N1N2N2O1N1O2O1N1O2O1N1O2O1N1O2O1N2O0O2N2O0O2N2O0O2N2O0O2N2O0O2N2O1N1O2O1N1O2O1N1O2O1N1O2N2O1N1O2N2O0O2N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2M3M3M3M3M3M3M3M3M3Mji\\1"}, "label": "a slightly burnt sausage on a piece of wheat bread"}]}
{"id": 410708, "image": "COCO_train2014_000000410708.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup with red flower printed placed near the donuts\"."}], "task": "refering", "answer_template": "The \"a white cup with red flower printed placed near the donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113], "bbox": [0.785390625, 0.0, 1.0, 0.2786458333333333], "iscrowd": 0, "area": 14457.8371, "rle": {"size": [480, 640], "counts": "Tj[7;d><D=C5K101N2N1O2N2N1O2O1N1O2N2N1O2O1N1O2N2O002N1O2N2ZCQNV<_2N3M2N3M2N2N2N0O2O0000001O000O2O0O1O2N1O1O2M2O1O1O2N1N2O2O00001O00001O0000001O00001O00001O0000000000000000000000000000000000000000000000O01000O100O100O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1O1O100O1O1O100O1O100O1O100N2F:Ga1^NRO"}, "label": "a white cup with red flower printed placed near the donuts"}]}
{"id": 410708, "image": "COCO_train2014_000000410708.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ceramic coffee much with blue lines on bottom and a red flower on the side\"."}], "task": "refering", "answer_template": "The \"a ceramic coffee much with blue lines on bottom and a red flower on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113], "bbox": [0.785390625, 0.0, 1.0, 0.2786458333333333], "iscrowd": 0, "area": 14457.8371, "rle": {"size": [480, 640], "counts": "Tj[7;d><D=C5K101N2N1O2N2N1O2O1N1O2N2N1O2O1N1O2N2O002N1O2N2ZCQNV<_2N3M2N3M2N2N2N0O2O0000001O000O2O0O1O2N1O1O2M2O1O1O2N1N2O2O00001O00001O0000001O00001O00001O0000000000000000000000000000000000000000000000O01000O100O100O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1O1O100O1O1O100O1O100O1O100N2F:Ga1^NRO"}, "label": "a ceramic coffee much with blue lines on bottom and a red flower on the side"}]}
{"id": 410708, "image": "COCO_train2014_000000410708.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mug is empty\"."}], "task": "refering", "answer_template": "The \"the mug is empty\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362], "bbox": [0.56965625, 0.5125, 0.913484375, 0.92825], "iscrowd": 0, "area": 29775.478299999995, "rle": {"size": [480, 640], "counts": "X\\[5<d>=C=@`0_Oa0_O=C8I7K5K5K6J5K5K5K5L4K5K6J3M2N1O2N1O2O0O2O0O2O0O2O0O2O0O2O0O101N100O2O001N100O2O0O101O001N10001O0O2O00001O0O101O001N10001O0O101O001O0O101O00001O001OO2O0O100O10000O100O10000O2O0O10000O100O10000O100O2O000O100O100O101O0O2N101N1O2N101N1O2N101N1O2N101N1O2N101N1O2N101N1O2O0O2N1O2N1O2M2O2M3N2M4M2M3N2M4M2M3N2M4M2M3N2N3L3N2O1N3N1O1O1O2N1N2O2N1O1O1O2M2O1O1O2N1N2O1O0000000O2O000000000O2O0O1O1O100O2N1O1O100O2N2N2N2O1N2N2N2N2O3K4K5J7J5J6K`^i0"}, "label": "the mug is empty"}]}
{"id": 410708, "image": "COCO_train2014_000000410708.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty red coffee cup with a white inside\"."}], "task": "refering", "answer_template": "The \"empty red coffee cup with a white inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362], "bbox": [0.56965625, 0.5125, 0.913484375, 0.92825], "iscrowd": 0, "area": 29775.478299999995, "rle": {"size": [480, 640], "counts": "X\\[5<d>=C=@`0_Oa0_O=C8I7K5K5K6J5K5K5K5L4K5K6J3M2N1O2N1O2O0O2O0O2O0O2O0O2O0O2O0O101N100O2O001N100O2O0O101O001N10001O0O2O00001O0O101O001N10001O0O101O001O0O101O00001O001OO2O0O100O10000O100O10000O2O0O10000O100O10000O100O2O000O100O100O101O0O2N101N1O2N101N1O2N101N1O2N101N1O2N101N1O2N101N1O2O0O2N1O2N1O2M2O2M3N2M4M2M3N2M4M2M3N2M4M2M3N2N3L3N2O1N3N1O1O1O2N1N2O2N1O1O1O2M2O1O1O2N1N2O1O0000000O2O000000000O2O0O1O1O100O2N1O1O100O2N2N2N2O1N2N2N2N2O3K4K5J7J5J6K`^i0"}, "label": "empty red coffee cup with a white inside"}]}
{"id": 132183, "image": "COCO_train2014_000000132183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a llama sticking his head out of a fence\"."}], "task": "refering", "answer_template": "The \"a llama sticking his head out of a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 289, 290, 291, 292, 311, 312, 313, 314], "bbox": [0.3565, 0.001288056206088993, 0.7588125, 0.9347540983606557], "iscrowd": 0, "area": 63756.940350000004, "rle": {"size": [427, 640], "counts": "gVo21Y=3M3M4L3M3M4L3RMYOSIk0e6_OUId0d6FVI=d6LVI8b62YI0`6:ZII_6`0\\ID\\6f0^I]O[6m0_IVO_6o0[IUOb6P1XISOf6R1TIROi6S1QIPOm6U1mHnNQ7W1iHmNT7W1gHlNW7f3N3M2N2N3M2N3M2N2N3M2N2N3M2N2N3M2N3M2N4L3M4L3M4L4L3M4L3M4L3M4L4L3M4L3M4L4L3M4L3M2N2N2N3M2N2N2N1O0000jLRMoKn2Q4UMmKj2S4YMjKg2U4]MhKc2X4aMdK_2\\4gM_KY2`4nMYKR2f4UNTKk1l4[NnJe1R5aNiJ^1V5iNdJW1\\5oN^JQ1b5QO\\Jo0e5ROZJm0f5UOXJk0i5VOSJl0m5VOnIm0R6UOiIn0X6SOcIP1]6RO]IR1d6oNWIT1i6nNRIU1n6d21O001O00001O001O00001O001O00001O3M6J5K5K5K5K5K5K6J5K5K5K5K5K5K5K6aNQFVOW:;QFDW:NPF1X:@oE`0Y:QOoEn0X;O1O1O001O1O1O1O0001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M6J5K5K5K5J7J5K5K5J7I6J6J6J7I6J6J6J6K2M4L3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3N2M4L3M3M4L6J5K5K5K5J6H8G9H9F9H8G9G9H8Gdmo1"}, "label": "a llama sticking his head out of a fence"}]}
{"id": 132183, "image": "COCO_train2014_000000132183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown sheep sticking it ' s head through the fence to eat grass\"."}], "task": "refering", "answer_template": "The \"a large brown sheep sticking it ' s head through the fence to eat grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 289, 290, 291, 292, 311, 312, 313, 314], "bbox": [0.3565, 0.001288056206088993, 0.7588125, 0.9347540983606557], "iscrowd": 0, "area": 63756.940350000004, "rle": {"size": [427, 640], "counts": "gVo21Y=3M3M4L3M3M4L3RMYOSIk0e6_OUId0d6FVI=d6LVI8b62YI0`6:ZII_6`0\\ID\\6f0^I]O[6m0_IVO_6o0[IUOb6P1XISOf6R1TIROi6S1QIPOm6U1mHnNQ7W1iHmNT7W1gHlNW7f3N3M2N2N3M2N3M2N2N3M2N2N3M2N2N3M2N3M2N4L3M4L3M4L4L3M4L3M4L3M4L4L3M4L3M4L4L3M4L3M2N2N2N3M2N2N2N1O0000jLRMoKn2Q4UMmKj2S4YMjKg2U4]MhKc2X4aMdK_2\\4gM_KY2`4nMYKR2f4UNTKk1l4[NnJe1R5aNiJ^1V5iNdJW1\\5oN^JQ1b5QO\\Jo0e5ROZJm0f5UOXJk0i5VOSJl0m5VOnIm0R6UOiIn0X6SOcIP1]6RO]IR1d6oNWIT1i6nNRIU1n6d21O001O00001O001O00001O001O00001O3M6J5K5K5K5K5K5K6J5K5K5K5K5K5K5K6aNQFVOW:;QFDW:NPF1X:@oE`0Y:QOoEn0X;O1O1O001O1O1O1O0001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M6J5K5K5K5J7J5K5K5J7I6J6J6J7I6J6J6J6K2M4L3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3N2M4L3M3M4L6J5K5K5K5J6H8G9H9F9H8G9G9H8Gdmo1"}, "label": "a large brown sheep sticking it ' s head through the fence to eat grass"}]}
{"id": 132183, "image": "COCO_train2014_000000132183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a standing sheep to the left of a sheep with its head in a fence\"."}], "task": "refering", "answer_template": "The \"a standing sheep to the left of a sheep with its head in a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 211, 212, 230, 231, 232, 235, 254, 255], "bbox": [0.0, 0.0032318501170960183, 0.38679687500000004, 0.787704918032787], "iscrowd": 0, "area": 53251.91365, "rle": {"size": [427, 640], "counts": "1h0c<_1aN_1aN`1`N_1aN8H4L4L4L4L4L4L4L4L4L3M2N5L4K5K6J5K5K5K5K5K6J1O1O001O1O1O001O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O00O1O1O1N2O1O1O1O1N2O1O1lLgLULZ3j3QMkKP3U4ZM`Kg2_4S300O100O1O100O100O1O100O100O1O100O11O00001O00001O001O00001O2N2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2N2N3M1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O\\N^K`Ja4`5aK_J^4a5cK^J]4b5eK]JZ4d5fK[JZ4e5hKZJW4f5kKXJU4h5lKWJT4i5nKVJQ4j5PLUJP4j5SLUJl3k5ULTJk3k5XLSJh3m5YLSJf3l5]LRJc3n5_LQJ`3n5bLQJ^3o5dLoI\\3P6fLPJY3P6iLnIW3Q6kLnIU3R6mLkIT3T6nLjIS3V6oLgIR3X6QMfIo2Z6RMdIo2[6TMbIm2^6UM`Ik2_6XM^Ii2a6ZM\\Ig2c6]MZIc2d6S2O1O1N2O1N2O1N2O1N5L5A`0_O`0J6L4K5K6K4K5K5L4K6K4K5K5L4K6J5L4K5K5L5JUMmNdIn0]6WObIe0^6@aI;`6K^I0d64\\IGd6>[I]Of6i0XISOh6R1WIiNj6\\1UI_Nl6f1TIUNl6Q2RIjMP7Z2oHaMR7d2mHXMS7m2lHnLU7X3\\15K5K5K6K4K5K2NO1O100O1O2N1O1O1O1O1O101N5K8H7I8H8H9G8H8H8H8H8H8H8He0[OR[S5"}, "label": "a standing sheep to the left of a sheep with its head in a fence"}]}
{"id": 132183, "image": "COCO_train2014_000000132183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep to the right\"."}], "task": "refering", "answer_template": "The \"the sheep to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 113, 114, 136, 137, 158, 159, 160, 181, 182], "bbox": [0.8063125, 0.0, 1.0, 0.5027868852459016], "iscrowd": 0, "area": 14022.2085, "rle": {"size": [427, 640], "counts": "YYg67P=:G3L4O1O2M2O1O2M1M2O1N2O1O2M2O0O01O0OoNDZE;f:R1100012M3M3M2O200001N2O1M3N2M3N2G9N2N2O1N2O1N2O12N2N3M2N2N2N2N2N2N2N2N1O1O2N1O2N1O1O2aHaMn3`2kKiMR4X2hKoMV4S2gKQNV4S2eKPNY4V2`KmM^4X2\\KjMc4\\2VKfMi4`2oJbMQ5b2gJbMY5a2iIWL3\\1T6`2dIYL1[1[6_2^I\\L1Y1`6m400000001OM3N2N2N1N3F:H8K5J6K5J6M3N2N2N2N2N2D<1O001O001O0bHfKk5[4oInKm5R4PJTLm5m3QJULn5k3PJWLP6j3mIXLS6h3kI[LT6f3gI^LY6b3dIaL\\6_3`IfL_6Z3^IhLc6X3[IiLf6X3VIjLk6V3SIkLn6U3oHlLS7T3kHmLV7c41O1OP3oLSM"}, "label": "the sheep to the right"}]}
{"id": 287833, "image": "COCO_train2014_000000287833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a plaid skirt walking out of a subway\"."}], "task": "refering", "answer_template": "The \"a woman in a plaid skirt walking out of a subway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 119, 120, 142, 143, 165, 166, 188, 189, 211, 212, 234, 235, 257, 258, 281, 304, 326, 327], "bbox": [0.173484375, 0.31901639344262295, 0.2604375, 0.9796955503512882], "iscrowd": 0, "area": 9265.009799999998, "rle": {"size": [427, 640], "counts": "k]^1j0]3XOd5U1nInNP6V1jImNU6V1_IUO_6n0lHFR7<aH1]71YH;c7GRHf0i7]OmGn0P8TOeGY1W8P2M4L:F;F1O1O1\\KQJj1P6SNRJm1n5QNTJo1m5nMUJR2l5lMUJT2k5jMWJV2i5hMYJY2e5fM\\J[2d5cM^J^2a5`MaJa2^5\\MeJe2Z5XMiJi2V5TMmJm2Q5QMRKP3m4lLWKU3h4hLZKZ3X5PLjJS4j5SKYJo4P72M3N3M2M3N3M2L4M4]OWJ^Il5l1[JQ2NnKj5j1VK]1VOcLh5i1SLg0_4ROdLN_3M\\M[Og2>nMQOU2h0Z5J7I6J7Il_U6"}, "label": "a woman in a plaid skirt walking out of a subway"}]}
{"id": 287833, "image": "COCO_train2014_000000287833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a red checkered dress who is exiting a train\"."}], "task": "refering", "answer_template": "The \"a woman with a red checkered dress who is exiting a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 119, 120, 142, 143, 165, 166, 188, 189, 211, 212, 234, 235, 257, 258, 281, 304, 326, 327], "bbox": [0.173484375, 0.31901639344262295, 0.2604375, 0.9796955503512882], "iscrowd": 0, "area": 9265.009799999998, "rle": {"size": [427, 640], "counts": "k]^1j0]3XOd5U1nInNP6V1jImNU6V1_IUO_6n0lHFR7<aH1]71YH;c7GRHf0i7]OmGn0P8TOeGY1W8P2M4L:F;F1O1O1\\KQJj1P6SNRJm1n5QNTJo1m5nMUJR2l5lMUJT2k5jMWJV2i5hMYJY2e5fM\\J[2d5cM^J^2a5`MaJa2^5\\MeJe2Z5XMiJi2V5TMmJm2Q5QMRKP3m4lLWKU3h4hLZKZ3X5PLjJS4j5SKYJo4P72M3N3M2M3N3M2L4M4]OWJ^Il5l1[JQ2NnKj5j1VK]1VOcLh5i1SLg0_4ROdLN_3M\\M[Og2>nMQOU2h0Z5J7I6J7Il_U6"}, "label": "a woman with a red checkered dress who is exiting a train"}]}
{"id": 42081, "image": "COCO_train2014_000000042081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut with brown frosting and two bites taken\"."}], "task": "refering", "answer_template": "The \"a donut with brown frosting and two bites taken\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239], "bbox": [0.159328125, 0.293326359832636, 0.436171875, 0.6346652719665272], "iscrowd": 0, "area": 21360.136999999995, "rle": {"size": [478, 640], "counts": "`i_1?^>g0YO4L4K5L4L4L4L4L4L4L4K5L4L4L4L4L4L4K4M2N1O2N1O2N1O2M2O2N2O001N101O0O2O001N101O1N101O0O2O001N101O1N101O0O2O001N100O101O0O10000O10000O10001N10000O1001O00001O0010O0001O00001O001O00010O00001O2N2N2N2N2N2O1N2N2N00000000000000000O10001N10000O1000000O10001N2O001N101O1N101O0O2O001N2O3M6I6K001N101O0O101MUDQMk;n24O0O1ORDUMl;i2501N1O1O2O0O2N1O101N3M3M3N2M3M3M3N2M3M2N3N2Mh0XOjXX5"}, "label": "a donut with brown frosting and two bites taken"}]}
{"id": 42081, "image": "COCO_train2014_000000042081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partially eaten chocolate frosted donut\"."}], "task": "refering", "answer_template": "The \"a partially eaten chocolate frosted donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239], "bbox": [0.159328125, 0.293326359832636, 0.436171875, 0.6346652719665272], "iscrowd": 0, "area": 21360.136999999995, "rle": {"size": [478, 640], "counts": "`i_1?^>g0YO4L4K5L4L4L4L4L4L4L4K5L4L4L4L4L4L4K4M2N1O2N1O2N1O2M2O2N2O001N101O0O2O001N101O1N101O0O2O001N101O1N101O0O2O001N100O101O0O10000O10000O10001N10000O1001O00001O0010O0001O00001O001O00010O00001O2N2N2N2N2N2O1N2N2N00000000000000000O10001N10000O1000000O10001N2O001N101O1N101O0O2O001N2O3M6I6K001N101O0O101MUDQMk;n24O0O1ORDUMl;i2501N1O1O2O0O2N1O101N3M3M3N2M3M3M3N2M3M2N3N2Mh0XOjXX5"}, "label": "a partially eaten chocolate frosted donut"}]}
{"id": 42081, "image": "COCO_train2014_000000042081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glazed jelly donut with white frosting and sprinkles on it\"."}], "task": "refering", "answer_template": "The \"a glazed jelly donut with white frosting and sprinkles on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 313, 314], "bbox": [0.489375, 0.3935564853556485, 0.8302968749999999, 0.7730753138075314], "iscrowd": 0, "area": 30153.969850000005, "rle": {"size": [478, 640], "counts": "e\\b43h>6I7I8I6I6J6K4K6J5L5K4L5J5L4L5K4N3M2N3M2N3M2N2M4M2N3M2N3M2N2N2M2O2N2N2N1O2N2N2N1N3O1O1O1O001O1O1O001O000000001O01O00001O01O01O00001O010O00O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N100O1O100O1O100O1O100O1O100O1O100O10000001O0000000000000000001O0000000000000000001O00000000000O100O101O0O100O2O001N101N2O001N101O1N100O101O0O100O10001N100O10001N100O102N1N2O1O1N2O1N2O1O1N101N2N2N2N2N2N2M2O2N2N2N2N2N2N1O2N2N2N2N2N3M4K4K5L4K5L4K5L4K6K4K5L4L4K:G=Bbcb1"}, "label": "a glazed jelly donut with white frosting and sprinkles on it"}]}
{"id": 42081, "image": "COCO_train2014_000000042081.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white frosted donut with sprinkles\"."}], "task": "refering", "answer_template": "The \"a white frosted donut with sprinkles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 313, 314], "bbox": [0.489375, 0.3935564853556485, 0.8302968749999999, 0.7730753138075314], "iscrowd": 0, "area": 30153.969850000005, "rle": {"size": [478, 640], "counts": "e\\b43h>6I7I8I6I6J6K4K6J5L5K4L5J5L4L5K4N3M2N3M2N3M2N2M4M2N3M2N3M2N2N2M2O2N2N2N1O2N2N2N1N3O1O1O1O001O1O1O001O000000001O01O00001O01O01O00001O010O00O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N100O1O100O1O100O1O100O1O100O1O100O10000001O0000000000000000001O0000000000000000001O00000000000O100O101O0O100O2O001N101N2O001N101O1N100O101O0O100O10001N100O10001N100O102N1N2O1O1N2O1N2O1O1N101N2N2N2N2N2N2M2O2N2N2N2N2N2N1O2N2N2N2N2N3M4K4K5L4K5L4K5L4K6K4K5L4L4K:G=Bbcb1"}, "label": "a white frosted donut with sprinkles"}]}
{"id": 271464, "image": "COCO_train2014_000000271464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black charter bus\"."}], "task": "refering", "answer_template": "The \"black charter bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 57, 58, 59, 60, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339], "bbox": [0.23595312499999999, 0.10562500000000001, 0.9235937499999999, 0.8651666666666666], "iscrowd": 0, "area": 120005.33919999999, "rle": {"size": [480, 640], "counts": "TnV2b0]>V1kNU1jNk0VO0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N1O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O101N101N101N101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O2O0O2O0O1O2O0O2O0O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101O001O0010O0001O001O0010O01O001O0010O01O00001O0010O01O001O000001O0O1000000O2O0000000O10001O000O100000001N1000000O10001O000O1000001O0O1000000O101O00000O1000001O0O100000001N1000000O10001O000O100000001N100000000O2O00000O1000001O0O100000000O2O0000000O101O000O100000001N100000000O101O000O1000001O0O100000000O2O0000000O10001O0O100000000O2O0000000O101O3M4L4K6K4L4L5J5L4L5K4L4K6K4L4L5K4K5L5K4L4L5J5L5K4L4K6K4L4L5K4K5L5K4L4K6I6I7J7I6I7J7H7J6fNlgf0"}, "label": "black charter bus"}]}
{"id": 271464, "image": "COCO_train2014_000000271464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bus\"."}], "task": "refering", "answer_template": "The \"a black bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 57, 58, 59, 60, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339], "bbox": [0.23595312499999999, 0.10562500000000001, 0.9235937499999999, 0.8651666666666666], "iscrowd": 0, "area": 120005.33919999999, "rle": {"size": [480, 640], "counts": "TnV2b0]>V1kNU1jNk0VO0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N1O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O101N101N101N101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O2O0O2O0O1O2O0O2O0O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101O001O0010O0001O001O0010O01O001O0010O01O00001O0010O01O001O000001O0O1000000O2O0000000O10001O000O100000001N1000000O10001O000O1000001O0O1000000O101O00000O1000001O0O100000001N1000000O10001O000O100000001N100000000O2O00000O1000001O0O100000000O2O0000000O101O000O100000001N100000000O101O000O1000001O0O100000000O2O0000000O10001O0O100000000O2O0000000O101O3M4L4K6K4L4L5J5L4L5K4L4K6K4L4L5K4K5L5K4L4L5J5L5K4L4K6K4L4L5K4K5L5K4L4K6I6I7J7I6I7J7H7J6fNlgf0"}, "label": "a black bus"}]}
{"id": 263275, "image": "COCO_train2014_000000263275.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man skating in the front of the image\"."}], "task": "refering", "answer_template": "The \"man skating in the front of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 37, 38, 39, 48, 49, 50, 59, 60, 61, 70, 71, 72, 73], "bbox": [0.20245847176079734, 0.042159999999999996, 0.7512624584717608, 0.37754], "iscrowd": 0, "area": 9302.40015, "rle": {"size": [500, 301], "counts": "]jm04`?001O010O010O1FLl@:U?3O0100O01O1O00O2O01O01O01O00O100O100O010000000000000000000000000000000000000000000O1001O1O001O001O1PBK_<5\\C2a<0\\C4a<N]C4b<L[C9b<IZC;V1ROb9c0UE?P1WOj9;nDg0R1SOo97mDh0n0WOS:2nDi0i0\\OW:e1eE_NY:b1cEcN[:^1bEeN\\:]1^EhNb:]1WEeNi:S300O10O1000000BREVLo:e3UE[Lk:c3WE]Li:`3ZE`Lf:^3\\EaLe:]3]EcLc:[3`EdL`:Y3cEfL^:W3eEiL[:U3hEjLX:T3kEjLV:R3oElLR:R3PFmLQ:T3nEjLT:W3T11fChLV<[3O2O0O2O0OeNSDUOh;i0^D[O[;b0iDLg:2]E>Q:ASF>k9AXF>f9A^F=b9BaF>]9AgF=X9AlF>T9@oF>R9APG=Q9CPG<P9EPG:P9FRG8n8IRG6n8JSG5n8KRG4n8LTG2m8SO`ENd1n0l8SOaENd1n0l8SOaEOd1m0k8ROcE0c1m0k8ROcE1g1f0g8YOcE0R2;\\8CcE3b2Hl73cE5e2Cm7LYEL8d0a2DR8HfEb0X2El8:TGFm89RGGo89PGHQ97oFHR98nFHS98lFGV98jFGW99jFEX9:hFEY9;iFAY9?n2000O0001O1O1O1O100O10000O10000001O001O2NLk@HW?74Ke@N]?1kaT1"}, "label": "man skating in the front of the image"}]}
{"id": 263275, "image": "COCO_train2014_000000263275.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy in white shirt skateboarding down rail\"."}], "task": "refering", "answer_template": "The \"guy in white shirt skateboarding down rail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 37, 38, 39, 48, 49, 50, 59, 60, 61, 70, 71, 72, 73], "bbox": [0.20245847176079734, 0.042159999999999996, 0.7512624584717608, 0.37754], "iscrowd": 0, "area": 9302.40015, "rle": {"size": [500, 301], "counts": "]jm04`?001O010O010O1FLl@:U?3O0100O01O1O00O2O01O01O01O00O100O100O010000000000000000000000000000000000000000000O1001O1O001O001O1PBK_<5\\C2a<0\\C4a<N]C4b<L[C9b<IZC;V1ROb9c0UE?P1WOj9;nDg0R1SOo97mDh0n0WOS:2nDi0i0\\OW:e1eE_NY:b1cEcN[:^1bEeN\\:]1^EhNb:]1WEeNi:S300O10O1000000BREVLo:e3UE[Lk:c3WE]Li:`3ZE`Lf:^3\\EaLe:]3]EcLc:[3`EdL`:Y3cEfL^:W3eEiL[:U3hEjLX:T3kEjLV:R3oElLR:R3PFmLQ:T3nEjLT:W3T11fChLV<[3O2O0O2O0OeNSDUOh;i0^D[O[;b0iDLg:2]E>Q:ASF>k9AXF>f9A^F=b9BaF>]9AgF=X9AlF>T9@oF>R9APG=Q9CPG<P9EPG:P9FRG8n8IRG6n8JSG5n8KRG4n8LTG2m8SO`ENd1n0l8SOaENd1n0l8SOaEOd1m0k8ROcE0c1m0k8ROcE1g1f0g8YOcE0R2;\\8CcE3b2Hl73cE5e2Cm7LYEL8d0a2DR8HfEb0X2El8:TGFm89RGGo89PGHQ97oFHR98nFHS98lFGV98jFGW99jFEX9:hFEY9;iFAY9?n2000O0001O1O1O1O100O10000O10000001O001O2NLk@HW?74Ke@N]?1kaT1"}, "label": "guy in white shirt skateboarding down rail"}]}
{"id": 476269, "image": "COCO_train2014_000000476269.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle rider\"."}], "task": "refering", "answer_template": "The \"a motorcycle rider\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 151, 152, 153, 173, 174, 175, 176, 198, 199, 221, 222, 244, 245, 268], "bbox": [0.55171875, 0.3518266978922716, 0.6905625000000001, 0.7562295081967212], "iscrowd": 0, "area": 4903.372400000001, "rle": {"size": [427, 640], "counts": "`]c41X=3M31NN101N100O2O0O2O0O101N100O2O0O2O0O101O000O2O001O000O2O1O1M2N3N2M3M3L4L4L4M4M2O1N3N1N2O2O000_EmMj9T2UFmMj9S2VFnMi9R2VFPNj9o1VFRNi9o1VFRNi9n1VF`N]9`1cFfNW9Z1iFlNQ9U1mFROm8n0SGXOh8g0XG^Oc8c0[GBb8=^GG^89bGJ\\86bGO[80eG4[8WN`Gk05R1[8iMgGU1GgNO_2T9iNjFkN2`2P9lNoF]O1j0m8IPG]O6i0j8IPG\\O;g0f8h0]GVOd8g0aGUOa8i0bGUO_8h0iGQOX8n0QHjNP8S1WHhNj7e0kFRO_15g7h0SGiN\\1;b7k0QIQOQ7n0SInNn6Q1UIkNn6S1h2M3L5L3L5G8H`[b2"}, "label": "a motorcycle rider"}]}
{"id": 476269, "image": "COCO_train2014_000000476269.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red coat wearing a helmet on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man in a red coat wearing a helmet on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 151, 152, 153, 173, 174, 175, 176, 198, 199, 221, 222, 244, 245, 268], "bbox": [0.55171875, 0.3518266978922716, 0.6905625000000001, 0.7562295081967212], "iscrowd": 0, "area": 4903.372400000001, "rle": {"size": [427, 640], "counts": "`]c41X=3M31NN101N100O2O0O2O0O101N100O2O0O2O0O101O000O2O001O000O2O1O1M2N3N2M3M3L4L4L4M4M2O1N3N1N2O2O000_EmMj9T2UFmMj9S2VFnMi9R2VFPNj9o1VFRNi9o1VFRNi9n1VF`N]9`1cFfNW9Z1iFlNQ9U1mFROm8n0SGXOh8g0XG^Oc8c0[GBb8=^GG^89bGJ\\86bGO[80eG4[8WN`Gk05R1[8iMgGU1GgNO_2T9iNjFkN2`2P9lNoF]O1j0m8IPG]O6i0j8IPG\\O;g0f8h0]GVOd8g0aGUOa8i0bGUO_8h0iGQOX8n0QHjNP8S1WHhNj7e0kFRO_15g7h0SGiN\\1;b7k0QIQOQ7n0SInNn6Q1UIkNn6S1h2M3L5L3L5G8H`[b2"}, "label": "a man in a red coat wearing a helmet on a motorcycle"}]}
{"id": 263278, "image": "COCO_train2014_000000263278.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long white surfboard with light and blue stripes in the middle and a small child at the end of it\"."}], "task": "refering", "answer_template": "The \"a long white surfboard with light and blue stripes in the middle and a small child at the end of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286], "bbox": [0.310359375, 0.4268028846153846, 0.4801875, 0.845721153846154], "iscrowd": 0, "area": 13315.053499999998, "rle": {"size": [416, 640], "counts": "jTa2:\\<;F9F;F:K5K5K5K5K5K5K5J6K5K5K5K5K5M3L4L4M3L4M3L3N2M4M2M3N2M3N2O100O2O000OSKVHf4U8M3M3M3N1N2N2N00000001O0O100000000FfKnGZ4S8hKjGY4U8jKgGW4Y881O4L000001N1000O1O010ON3L3L5K4M4K42O2M4L3N2M3N2M3N2M3N2M3M3N2M4M4J5K5L5J5K5L4K6J5L4K6J5L4K=C=D=B=CioV4"}, "label": "a long white surfboard with light and blue stripes in the middle and a small child at the end of it"}]}
{"id": 263278, "image": "COCO_train2014_000000263278.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"long white surfboard with a little boy on it\"."}], "task": "refering", "answer_template": "The \"long white surfboard with a little boy on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286], "bbox": [0.310359375, 0.4268028846153846, 0.4801875, 0.845721153846154], "iscrowd": 0, "area": 13315.053499999998, "rle": {"size": [416, 640], "counts": "jTa2:\\<;F9F;F:K5K5K5K5K5K5K5J6K5K5K5K5K5M3L4L4M3L4M3L3N2M4M2M3N2M3N2O100O2O000OSKVHf4U8M3M3M3N1N2N2N00000001O0O100000000FfKnGZ4S8hKjGY4U8jKgGW4Y881O4L000001N1000O1O010ON3L3L5K4M4K42O2M4L3N2M3N2M3N2M3N2M3M3N2M4M4J5K5L5J5K5L4K6J5L4K6J5L4K=C=D=B=CioV4"}, "label": "long white surfboard with a little boy on it"}]}
{"id": 509039, "image": "COCO_train2014_000000509039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork on a plate\"."}], "task": "refering", "answer_template": "The \"a fork on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 291, 292], "bbox": [0.201375, 0.5972594142259414, 0.8555781249999999, 0.7279916317991633], "iscrowd": 0, "area": 5538.550549999998, "rle": {"size": [478, 640], "counts": "Zal12i>5L4L4M10O100000O100000O100000O100000O100000O1000O1000000000O10O1000000000O10O1000000000O10O1000000000O01000000000000O0100000000000O01000000000O010000000000O010000000O1000O100000O1000O1000O10000000O1000000000000O10000000O10O1000000000000O10000000000O1000O10000000O1000000000000O10000000000O1000000000000O1000000000000O10000001O0000000000000000000000000000000000000000000000000000000001O0001O0000000000000000000001O0000000000000000000000001O0000000000000001O0000000010O0001O00001O001O00001O001O01O01O00001O001O01O01O0010O0001O01O01O010O000010O01O01O01O00010O001O01O01O010O00010O01O01O01O01O010O01O01O01O01O01O0100OO101N10001N100O2O0O2O0O2O1O1N2O1N_S51alJ1N101ZAL`>4`AM^>4bAM]>2cAO\\>2dAO[>1dA1Z>OgA0Z>0fA0MLZ>3iA1NLX>4jA0MNX>1kA2MNV>1mA0M1U>NoA1L1T>OoA0N01Ln=3TB1L12Lm=2TB1N01Ok=1VB0M12Oj=OXB0M112h=NYB0N012h=MZB0M113g=MZB0N013g=L[B0M113h=LZB0M113g=L[B1MO14h=LZB0M113g=MZB0NO14h=LZB0M113g=M[BOM023g=MZB0N003i=MYB0M023g=MZB0N003i=MYB0M023g=M[BOM103i=MYB0M023g=M[BOM103i=MYB0NO14h=MYB0M013i=MZBOM103i=MYB0M013i=MZBOM103i=MYB0NO13h=NZBOM103i=MZBOM013i=MYB0M103i=MZBOM013i=NXBOO0O4i=MZBOM013i=NYBNN1O4i=MZBOM013i=NYBOM004j=LYB0NO04i=NYBOM004j=LZBOM004i=NYBOM004j=LZBOM004j=MXB0M013i=NYBNN004j=MYBOM004i=NYBNN004k=LXB0M004k=LWB1MO14m=JVB2M004m=JVB2MO05U>LkA004U>LkAO05U>LkA003V>MjAO05V>KkA0O4]>McA3]>LcA4]>McA3]>LcA4^>LbA4c>010O0101N2O1NfS[1"}, "label": "a fork on a plate"}]}
{"id": 509039, "image": "COCO_train2014_000000509039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone with a black t - shirt with a small star sittring in front of their plate of food\"."}], "task": "refering", "answer_template": "The \"someone with a black t - shirt with a small star sittring in front of their plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 84], "bbox": [0.186296875, 0.0023221757322175736, 0.686421875, 0.21119246861924687], "iscrowd": 0, "area": 20227.9805, "rle": {"size": [478, 640], "counts": "cag13k>8H8H8H5K2N2N3M2fBhNj<Z1QCiNo<Z1kBhNV=`10000000000000000000001O0000000000001O0000000000001O000000000000000000001O0000000000000000000000001O001O001O001O001O001O1O0000001O000000000000000000000000000000000000001O0000000000000000000000O10000000000O100000000O10000000000O10000000000O10000000000O100000000O10000000000O1001O000000000000001O0001O000000000001O000000000000001O00000000000000001O000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O0000001O001O001O00001O001O001O00001O001O004L4L4L4LZORkm2"}, "label": "someone with a black t - shirt with a small star sittring in front of their plate of food"}]}
{"id": 509039, "image": "COCO_train2014_000000509039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with tomatoes and no bites taken out\"."}], "task": "refering", "answer_template": "The \"a sandwich with tomatoes and no bites taken out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 149], "bbox": [0.29479687499999996, 0.13433054393305438, 0.64640625, 0.3628033472803347], "iscrowd": 0, "area": 16145.77825, "rle": {"size": [478, 640], "counts": "VZh29d>1O2N1O2N1O101N1O2N1O2N1O1O2O0O2O0O2O0O101N1O2O0O101N101N101N101N101N2O1N2O1N2O1O1O0O2O1O1O1O1N2O1O1O001N2O1O1O1O1O1N101O1O1O1N101O1O1O1N101O1O100O001O1O1O1O001O100O001O1O1O1O00010O0000001O00000010O000000010O000010O0000010O0001O01O01O000001O00000000O101O0O100O100O10000O100O100O10000O10000000000000001O0O10000O1000000O1001O1O1O1O100O1O1OKQD\\Mo;c2QD^Mn;a2TD^Ml;c2TD[Mm;e2SDZMo;e2RDZMn;f250001O01O0001O000N2O2M2ZNcCi0_<WOaCh0`<XO`Cf0c<YO]Cf0d<YO]Ce0e<[O\\Cc0e<]O[Ca0h<^OXCb0h<]OYCc0g<]OYCb0h<^OXCb0i<]OXCb0h<^OXCb0h<]OYCb0h<^OXCb0i<]OWCIK=n<JXCHJ>o<HXCIK?m<GXCJL?k<GZCHLb0j<EZCIMb0i<D[CHNd0f<C]CIMe0f<A^CHNh0c<_O`CHMj0b<^OaCGOk0`<\\OcCHNl0^<\\ORDb0Q<]OPDa0R<^OoC`0R<@nC>U<@mC>U<AlC<V<DjC;X<Da1O2OhkY3"}, "label": "a sandwich with tomatoes and no bites taken out"}]}
{"id": 509039, "image": "COCO_train2014_000000509039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the part of the sandwich closest to the person eating it\"."}], "task": "refering", "answer_template": "The \"the part of the sandwich closest to the person eating it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 149], "bbox": [0.29479687499999996, 0.13433054393305438, 0.64640625, 0.3628033472803347], "iscrowd": 0, "area": 16145.77825, "rle": {"size": [478, 640], "counts": "VZh29d>1O2N1O2N1O101N1O2N1O2N1O1O2O0O2O0O2O0O101N1O2O0O101N101N101N101N101N2O1N2O1N2O1O1O0O2O1O1O1O1N2O1O1O001N2O1O1O1O1O1N101O1O1O1N101O1O1O1N101O1O100O001O1O1O1O001O100O001O1O1O1O00010O0000001O00000010O000000010O000010O0000010O0001O01O01O000001O00000000O101O0O100O100O10000O100O100O10000O10000000000000001O0O10000O1000000O1001O1O1O1O100O1O1OKQD\\Mo;c2QD^Mn;a2TD^Ml;c2TD[Mm;e2SDZMo;e2RDZMn;f250001O01O0001O000N2O2M2ZNcCi0_<WOaCh0`<XO`Cf0c<YO]Cf0d<YO]Ce0e<[O\\Cc0e<]O[Ca0h<^OXCb0h<]OYCc0g<]OYCb0h<^OXCb0i<]OXCb0h<^OXCb0h<]OYCb0h<^OXCb0i<]OWCIK=n<JXCHJ>o<HXCIK?m<GXCJL?k<GZCHLb0j<EZCIMb0i<D[CHNd0f<C]CIMe0f<A^CHNh0c<_O`CHMj0b<^OaCGOk0`<\\OcCHNl0^<\\ORDb0Q<]OPDa0R<^OoC`0R<@nC>U<@mC>U<AlC<V<DjC;X<Da1O2OhkY3"}, "label": "the part of the sandwich closest to the person eating it"}]}
{"id": 443505, "image": "COCO_train2014_000000443505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"catcher behind person batting\"."}], "task": "refering", "answer_template": "The \"catcher behind person batting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 277, 278, 279, 300, 301, 302, 324], "bbox": [0.0, 0.3626463700234192, 0.354421875, 0.949391100702576], "iscrowd": 0, "area": 21713.908650000005, "rle": {"size": [427, 640], "counts": "c8`0i<=D6J3N2N3M2N2M2N2N1N3M3M2N3M2N3M2O2O0O2N2N101N1O100O100O1SGhMf6Y2XIiMg6X2WIjMh6W2VIkMi6V2VIkMi6V2VIkMi6V2VIkMi6V2VIkMj6U2UIlMj6U2hHaMoN;X8U2hHYNW7h1hHYNW7h1gHYNY7h1fHZNX7g1fH[NY7f1fH\\NY7d1fH]NY7d1fH^NX7c1gH^NX7c1hH]NX7c1gH_NW7b1gH`NX7a1gH`NX7a1gH`NY7`1fHbNX7_1gHbNX7_1gHbNX7_1gHbNX7_1gHbNY7^1eHdNZ7]1eHdNZ7]1eHdNZ7]1dHeN[7\\1dHeN[7\\1cHfN]7Z1bHgN]7Z1aHiN]7X1bHiN]7c3O1O1O2O00000O101O0000000O100OXOlH[KS7e4oHZKQ7e4RIYKn6f4UIXKQ6BZJU5IWKh5NZJj42SKc59XJc4e6`KYI_4f6eKUI\\1Li0m6nMUIY1Oh0l6QNSIW12h0i6;XIDh6<YIBh6?XIoNKVNn6j2WIPOORNk6m2WIPO[7o0fHQOZ7n0fHRO[7n0eHQO\\7n0eHQO\\7n0fHQOZ7m0hHQOZ7m0hHQOY7n0iHoNZ7o0hHoNY7P1iHoNX7P1iHoNW7Q1jHnNW7Q1jHnNW7R1hHnNY7S1cHoN^7Q1bHnN_7S1_HmNb7S1^HlNc7T1]HkNd7V1\\HhNe7X1\\HeNf7[1ZHdNg7]1YH`Ni7`1WH_Nk7a1UH\\Nm7c1UH[Nl7d1VHZNk7e1WHYNk7e1WH[Nh7d1ZHlMUOIa8Z2\\HlMVODb8^2ZHlMXO_Ob8d2XHlMR8R2QHkMQ8S2SHgMQ8W2c1XNTEn0n:oNTEQ1m:mNTES1m:kNTEU1n:iNREW1P;gNPEZ1P;eNPEc0OFR;EPEe0OER;EoDe00FR;DnDc04Ho:DmDb05Ko:AmDa07Nm:@mD>93T;JoD6R;FPE;Q;@SE`0l;0O1000O010O1O10O0100O01O01O01O010O0010O01O01O01O010O010O01O010O010O010O0010O0SO@QE?n:GgD`0V;EfD<X;IdD8\\;k001O0000N110000001O001O0100000000000000001O001N1O2N2N1M4M4L3J6G;J8HfQ\\5"}, "label": "catcher behind person batting"}]}
{"id": 443505, "image": "COCO_train2014_000000443505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"batter\"."}], "task": "refering", "answer_template": "The \"batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 291, 292, 309, 310, 314, 315], "bbox": [0.43629687500000003, 0.15281030444964872, 0.74515625, 0.9213583138173302], "iscrowd": 0, "area": 24913.862999999998, "rle": {"size": [427, 640], "counts": "Vfd35T==B4M4M2O2M3N1N3N000O10000O010O1000O0100O0100N101O1O001O001O000O20O0010O01O010O0010O010O`L]OZJc0d5DVJ=h5JSJ5k51QJ0m56oIIP6<mIDQ6b0jI_OS6g0jIXOU6l0iITOU4ROeKn13POV4YO`Kk19lNU4_O\\Ki1<hNW4DXKg1`0eNU4MSKa1g0aNV45eJd1R1XNX4<\\J`1[1SNX4e0SJ]1c1oMY4j3dKVL\\4l3cKTL\\4n3bKRL^4Q4_KoKa4S4]KnKb4T4]KkKc4W4[KjKd4Y4YKgKg4[4XKeKg4]4WKdKh4^4WKaKi4a4VK_Ki4e4SK[Kl4i4RKWKm4n4nJSKP5S5lJlJS5Z5gJhJW5_5bJbJ]5g6N1O2M3L3L5K4N3L4O0010O01O01000O0100O1000O0101N100O2O000O100M3M3N2ROn0K5K5L400O100VNbIgL^6S3iIkLY6Q3kInLU6o2oIPMR6g0^IId0@m5g0`IEg0Di5e0cICf0Hg5e0dI@h0Kd5d0fI]Oi0O`5d0hIZOk02\\5c0lIWOj06[5b0lISOn0;V5a0mIoNR1`0P5a0oIiNV1f0l4?QJeNX1l0g4>RJaN\\1Q1b4=SJ[Nb1X1[4=dLDa36`LJd32\\LOh3KYL5l3EUL<P4^OPLb0T4YOmK^1[3bNfL_1Y3aNgLa1RN`Mh3n0WNd1lMaMm3j0XNf1gMbMR4g0XNi1`McMX4d0YNj1[MdM]4a0YNn1TMdMc4=ZNT2jLbMm49ZNo2f1QM[Nn2f1PM\\NP3d1nL^NQ3c1nL^NQ3c1mL_NS3a1kLaNT3a1jL`NU3b1hL`NW3k500O1O101N102SNVF3T:WO`Fb0\\;H5K2O2M2O1N2N2O2M100O100O1O100O100O101N1O1O1N2O1O1N2O1N2OlbS2"}, "label": "batter"}]}
{"id": 443505, "image": "COCO_train2014_000000443505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batter in a purple jersey taking a swing\"."}], "task": "refering", "answer_template": "The \"a batter in a purple jersey taking a swing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 291, 292, 309, 310, 314, 315], "bbox": [0.43629687500000003, 0.15281030444964872, 0.74515625, 0.9213583138173302], "iscrowd": 0, "area": 24913.862999999998, "rle": {"size": [427, 640], "counts": "Vfd35T==B4M4M2O2M3N1N3N000O10000O010O1000O0100O0100N101O1O001O001O000O20O0010O01O010O0010O010O`L]OZJc0d5DVJ=h5JSJ5k51QJ0m56oIIP6<mIDQ6b0jI_OS6g0jIXOU6l0iITOU4ROeKn13POV4YO`Kk19lNU4_O\\Ki1<hNW4DXKg1`0eNU4MSKa1g0aNV45eJd1R1XNX4<\\J`1[1SNX4e0SJ]1c1oMY4j3dKVL\\4l3cKTL\\4n3bKRL^4Q4_KoKa4S4]KnKb4T4]KkKc4W4[KjKd4Y4YKgKg4[4XKeKg4]4WKdKh4^4WKaKi4a4VK_Ki4e4SK[Kl4i4RKWKm4n4nJSKP5S5lJlJS5Z5gJhJW5_5bJbJ]5g6N1O2M3L3L5K4N3L4O0010O01O01000O0100O1000O0101N100O2O000O100M3M3N2ROn0K5K5L400O100VNbIgL^6S3iIkLY6Q3kInLU6o2oIPMR6g0^IId0@m5g0`IEg0Di5e0cICf0Hg5e0dI@h0Kd5d0fI]Oi0O`5d0hIZOk02\\5c0lIWOj06[5b0lISOn0;V5a0mIoNR1`0P5a0oIiNV1f0l4?QJeNX1l0g4>RJaN\\1Q1b4=SJ[Nb1X1[4=dLDa36`LJd32\\LOh3KYL5l3EUL<P4^OPLb0T4YOmK^1[3bNfL_1Y3aNgLa1RN`Mh3n0WNd1lMaMm3j0XNf1gMbMR4g0XNi1`McMX4d0YNj1[MdM]4a0YNn1TMdMc4=ZNT2jLbMm49ZNo2f1QM[Nn2f1PM\\NP3d1nL^NQ3c1nL^NQ3c1mL_NS3a1kLaNT3a1jL`NU3b1hL`NW3k500O1O101N102SNVF3T:WO`Fb0\\;H5K2O2M2O1N2N2O2M100O100O1O100O100O101N1O1O1N2O1O1N2O1N2OlbS2"}, "label": "a batter in a purple jersey taking a swing"}]}
{"id": 418935, "image": "COCO_train2014_000000418935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with curly hair\"."}], "task": "refering", "answer_template": "The \"woman with curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0, 0.5619483568075117, 0.22337500000000002, 0.987605633802817], "iscrowd": 0, "area": 17764.642900000006, "rle": {"size": [426, 640], "counts": "T:=m<j0QNlNSFn1Q9VNnFc2U8bMiGg2l7]MRHd2j7aMUH_2i7dMUH]2k7dMSH]2l7eMSH[2l7gMRHZ2m7iMPHX2o7iMQHW2n7jMQHW2o7iMQHW2n7jMQHW2n7kMPHV2n7lMRHT2k7oMTHR2h7RNWHo1g7SNYHm1f7TNYHm1g7SNYHm1g7TNWHm1h7TNWHm1i7VNTHj1l7j1O100OfJSHY5m721ONfJTHZ5m7100O10000O10000O100O10000O100O10000O10000O100O10000O11O001O001O001O001O001O001O001O001kLTHj0l7VOTHj0m7UOSHk0m7UOTHj0m7UOSHk0m7UOSHk0n7TOSHk0m7VORHj0o7UOQHk0o7UORHj0o7UOQHk0P8TOPHl0Q8SOPHl0Q8SOoGm0R8ROnGm0T8ROmGj0W8SOkGi0Z8UOgGh0]8VOeGf0_8XObGe0c8XO^Ge0f8YO\\Gc0h8[OYGf0g8XOZGj0e8SO^Go0a8nN`GU1_8hNbG[1]8bNeG_1[8^NjGa1U8\\NPHc1o7ZNWHc1h7YN_Hd1b7XNeHc1]7XNjHd1X7WNPId1R7XNTId1[9N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N1O2N1OV[^6"}, "label": "woman with curly hair"}]}
{"id": 418935, "image": "COCO_train2014_000000418935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with curly hair\"."}], "task": "refering", "answer_template": "The \"a woman with curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0, 0.5619483568075117, 0.22337500000000002, 0.987605633802817], "iscrowd": 0, "area": 17764.642900000006, "rle": {"size": [426, 640], "counts": "T:=m<j0QNlNSFn1Q9VNnFc2U8bMiGg2l7]MRHd2j7aMUH_2i7dMUH]2k7dMSH]2l7eMSH[2l7gMRHZ2m7iMPHX2o7iMQHW2n7jMQHW2o7iMQHW2n7jMQHW2n7kMPHV2n7lMRHT2k7oMTHR2h7RNWHo1g7SNYHm1f7TNYHm1g7SNYHm1g7TNWHm1h7TNWHm1i7VNTHj1l7j1O100OfJSHY5m721ONfJTHZ5m7100O10000O10000O100O10000O100O10000O10000O100O10000O11O001O001O001O001O001O001O001O001kLTHj0l7VOTHj0m7UOSHk0m7UOTHj0m7UOSHk0m7UOSHk0n7TOSHk0m7VORHj0o7UOQHk0o7UORHj0o7UOQHk0P8TOPHl0Q8SOPHl0Q8SOoGm0R8ROnGm0T8ROmGj0W8SOkGi0Z8UOgGh0]8VOeGf0_8XObGe0c8XO^Ge0f8YO\\Gc0h8[OYGf0g8XOZGj0e8SO^Go0a8nN`GU1_8hNbG[1]8bNeG_1[8^NjGa1U8\\NPHc1o7ZNWHc1h7YN_Hd1b7XNeHc1]7XNjHd1X7WNPId1R7XNTId1[9N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N1O2N1OV[^6"}, "label": "a woman with curly hair"}]}
{"id": 181367, "image": "COCO_train2014_000000181367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bicycle\"."}], "task": "refering", "answer_template": "The \"a blue bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.038875, 0.4875833333333333, 0.4130625, 0.8871458333333333], "iscrowd": 0, "area": 23104.88464999999, "rle": {"size": [480, 640], "counts": "kQ<:Z>>G:K4L5L3L5K4L5K4L5K4L2N3M2N3O1N101N101N1O2O1N101N101N2N101N101N100O1O100O100O1O10000O100O10000O100O10000O10000000O10O01O1O100O1O100O1O1O10N1D<1O100O001O100O1O10O01O1O100O1O00100O1O1O10OlM\\Dm0e;PO^Do0c;nN`DQ1a;kNcDU1];hNfDW1[;fNhDY1X;dNlD[1U;bNnD]1S;`NPE`1X<O1OoMXNlFg1U9\\NhFc1Y9`NdF_1]9dN`F[1`9gN_FX1b9iN]FV1d9jN\\FV1d9kN[FT1f9mNYFR1h9nNXFQ1i9POUFP1k9QOUFn0l9SOSFl0n9UOQFj0P:VOPFj0R:UOmEj0U:UOkEj0W:VOhEi0[:UOeEj0]:UOcEj0_:VO`Ei0c:UO]Ej0e:UO[Ej0g:VOXEj0j:TOVEk0l:TOTEj0o:VOPEi0S;UOmDj0U;UOkDj0W;VOhDi0[;UOeDi0];WOcDh04ZOg9OTFf01Bj9GTFg0NJk9_OWFf0^O?X:lNYFc0^Of0X:fNZFc0\\Om0W:`N]Fb0ZOT1V:[N^Fa0ZOY1W:UN_Fa0XO`1V:oMaF`0XOg1T:jMcF>WOl1U:fMcF>WOo1V:bMcF>WOR2U:aMcF<XOT2U:`MbF<YOV2U:]MbF=XOX2U:\\MaF<ZOY2U:\\M`F;ZO[2U:[M_F;[O\\2U:ZM_F:[O]2V:ZM^F8\\O`2U:YM]F8]Oa2U:WM^F8\\Ob2V:WM\\F_3c9bL\\F_3c9eLXF\\3h9gLUFZ3i9kLSFU3m9oLnES3P:P1O001N2O001N2ORKYFc4g9\\K_F`4_9_KiF\\4V9dKnFX4R9hKRGU4l8lKXGP4h8PLbGe3_8[LbGd3]8]LcGc3]8]LcGc3]8]LcGc3\\8^LdGb3\\8^LdGb3\\8^LdGa3\\8`LdG`3\\8`LeG_3[8aLeG_3\\8^LfGb3\\8[LeGe3k900O10001O001O001N2O001O1O001O1N101O1O001O1N1O2N2N1O2M3N1O2N2N1N3N1O2N2N2M5K5I6J7I7I7I6J7J6I6J7IQm_5"}, "label": "a blue bicycle"}]}
{"id": 181367, "image": "COCO_train2014_000000181367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ladies ' bike is leaning on its kickstand\"."}], "task": "refering", "answer_template": "The \"a ladies ' bike is leaning on its kickstand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.038875, 0.4875833333333333, 0.4130625, 0.8871458333333333], "iscrowd": 0, "area": 23104.88464999999, "rle": {"size": [480, 640], "counts": "kQ<:Z>>G:K4L5L3L5K4L5K4L5K4L2N3M2N3O1N101N101N1O2O1N101N101N2N101N101N100O1O100O100O1O10000O100O10000O100O10000O10000000O10O01O1O100O1O100O1O1O10N1D<1O100O001O100O1O10O01O1O100O1O00100O1O1O10OlM\\Dm0e;PO^Do0c;nN`DQ1a;kNcDU1];hNfDW1[;fNhDY1X;dNlD[1U;bNnD]1S;`NPE`1X<O1OoMXNlFg1U9\\NhFc1Y9`NdF_1]9dN`F[1`9gN_FX1b9iN]FV1d9jN\\FV1d9kN[FT1f9mNYFR1h9nNXFQ1i9POUFP1k9QOUFn0l9SOSFl0n9UOQFj0P:VOPFj0R:UOmEj0U:UOkEj0W:VOhEi0[:UOeEj0]:UOcEj0_:VO`Ei0c:UO]Ej0e:UO[Ej0g:VOXEj0j:TOVEk0l:TOTEj0o:VOPEi0S;UOmDj0U;UOkDj0W;VOhDi0[;UOeDi0];WOcDh04ZOg9OTFf01Bj9GTFg0NJk9_OWFf0^O?X:lNYFc0^Of0X:fNZFc0\\Om0W:`N]Fb0ZOT1V:[N^Fa0ZOY1W:UN_Fa0XO`1V:oMaF`0XOg1T:jMcF>WOl1U:fMcF>WOo1V:bMcF>WOR2U:aMcF<XOT2U:`MbF<YOV2U:]MbF=XOX2U:\\MaF<ZOY2U:\\M`F;ZO[2U:[M_F;[O\\2U:ZM_F:[O]2V:ZM^F8\\O`2U:YM]F8]Oa2U:WM^F8\\Ob2V:WM\\F_3c9bL\\F_3c9eLXF\\3h9gLUFZ3i9kLSFU3m9oLnES3P:P1O001N2O001N2ORKYFc4g9\\K_F`4_9_KiF\\4V9dKnFX4R9hKRGU4l8lKXGP4h8PLbGe3_8[LbGd3]8]LcGc3]8]LcGc3]8]LcGc3\\8^LdGb3\\8^LdGb3\\8^LdGa3\\8`LdG`3\\8`LeG_3[8aLeG_3\\8^LfGb3\\8[LeGe3k900O10001O001O001N2O001O1O001O1N101O1O001O1N1O2N2N1O2M3N1O2N2N1N3N1O2N2N2M5K5I6J7I7I7I6J7J6I6J7IQm_5"}, "label": "a ladies ' bike is leaning on its kickstand"}]}
{"id": 181367, "image": "COCO_train2014_000000181367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue and black shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a blue and black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 140, 141, 142, 163, 164, 165, 186, 187, 188, 189, 209, 210, 211, 232, 234, 256, 257, 279, 302, 303], "bbox": [0.09775, 0.32583333333333336, 0.237640625, 0.8179791666666667], "iscrowd": 0, "area": 8264.771649999999, "rle": {"size": [480, 640], "counts": "Thm07a>8I8G8RCXOa;k0XD_Od;b0TDHj;8oC2m;OkC;R<\\1M0010O0000010O0ObDWMh:i2[EZM_:f2bE^MY:a2iEcMb9GPFe2?iMZ9HRF_2f0lMS9JRFY2m0QNk8b2VGcMd8]2]GQOT7P1mHXOh6j0XIUOh6l0XIROh6P1XIoNh6R1YIkNg6W1ZIfNg6[1ZIcNf6^1\\I^Ng6a1ZI]Ng6c1[IYNg6g1[IUNg6k1ZIRNh6n1ZIlMj6T2XIgMk6Y2VIbMn6^2TI\\MP7d2QIWMT7h2nHlL\\7T3fH`Ld7`3n1000eG`LV5`3gJcLY5]3fJeLY5g0jGd1l2gMY5c0nGe1h2jMY5>SHg1f2jMW5<VHi1f2iMV59VHn1f2gMW56VHR2e2gMW51XHW2NjMk1MR6LXH\\2JSNh1Cn6\\1jF]O?Mc1YOT7\\1lF\\O<7_1POY7\\1nF\\O8:h::RE[O4;l:8SE\\OO;Q;8RE]OJ:W;8QE]OF;];5oD_OB;c;4mD?X;^OjD?Y;_OiD=[;AgD;^;BgD8\\;EiD5[;IjD0Z;NjDMX;1nDHT;8PEBS;=i1N2O1O0O2O1O1N2O001N2O1OeoT7"}, "label": "a woman in a blue and black shirt"}]}
{"id": 181367, "image": "COCO_train2014_000000181367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person standing next to bicycle\"."}], "task": "refering", "answer_template": "The \"person standing next to bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 140, 141, 142, 163, 164, 165, 186, 187, 188, 189, 209, 210, 211, 232, 234, 256, 257, 279, 302, 303], "bbox": [0.09775, 0.32583333333333336, 0.237640625, 0.8179791666666667], "iscrowd": 0, "area": 8264.771649999999, "rle": {"size": [480, 640], "counts": "Thm07a>8I8G8RCXOa;k0XD_Od;b0TDHj;8oC2m;OkC;R<\\1M0010O0000010O0ObDWMh:i2[EZM_:f2bE^MY:a2iEcMb9GPFe2?iMZ9HRF_2f0lMS9JRFY2m0QNk8b2VGcMd8]2]GQOT7P1mHXOh6j0XIUOh6l0XIROh6P1XIoNh6R1YIkNg6W1ZIfNg6[1ZIcNf6^1\\I^Ng6a1ZI]Ng6c1[IYNg6g1[IUNg6k1ZIRNh6n1ZIlMj6T2XIgMk6Y2VIbMn6^2TI\\MP7d2QIWMT7h2nHlL\\7T3fH`Ld7`3n1000eG`LV5`3gJcLY5]3fJeLY5g0jGd1l2gMY5c0nGe1h2jMY5>SHg1f2jMW5<VHi1f2iMV59VHn1f2gMW56VHR2e2gMW51XHW2NjMk1MR6LXH\\2JSNh1Cn6\\1jF]O?Mc1YOT7\\1lF\\O<7_1POY7\\1nF\\O8:h::RE[O4;l:8SE\\OO;Q;8RE]OJ:W;8QE]OF;];5oD_OB;c;4mD?X;^OjD?Y;_OiD=[;AgD;^;BgD8\\;EiD5[;IjD0Z;NjDMX;1nDHT;8PEBS;=i1N2O1O0O2O1O1N2O001N2O1OeoT7"}, "label": "person standing next to bicycle"}]}
{"id": 148602, "image": "COCO_train2014_000000148602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"military flight number 854 in yellow handle\"."}], "task": "refering", "answer_template": "The \"military flight number 854 in yellow handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 149, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 247, 248, 264, 265], "bbox": [0.10071875, 0.38967136150234744, 0.9977656250000001, 0.755868544600939], "iscrowd": 0, "area": 31780.317150000017, "rle": {"size": [426, 640], "counts": "jXk03V=1N3N2M3N2O1O1N101O0O100000001N100000001O0O100000001O0O100000000O10000000000O2O0000000O100001O000000000001O00000000000000000001O000001O0000000000000000000000000010O0000000000000000000000000001O000001O00000000000000000001O00000000000001O000000000001O000000000000000000000000001O01O000001gCTOR<R100000000000000000000000000YDnNV;R1fDROZ;n0cDVO\\;j0cDWO];i0bDYO^;f0bDZO^;g0`D[O_;e0aD\\O^;d0aD]O_;c0aD^O^;b0bD^O^;b0aD@^;`0bD@^;V10001O0000000000001O000001O0001O0000000000001O0000000000001O00000000010O00000000000000000000O10000000000O1000001O000O10000000000O1000O101O2N2N1O2N1O2N2N1O2N1B>I6K4L5L3N3N2M2N3N1O2O13L5L3M\\FbL`9Y3_FlLa9Q3]FSMd9i2ZF]Mg9Q33L5L3L5C<H9K4M4K3M010O1O100O1O01N101O001oDRNe:o1YEVNc:k1\\EYN_:j1_EZN]:g1aE\\N]:Y2O1O1N2O1O1O1O1OO0N3M3F:1O1N2O1O001O1O100O1O010001O001O1N101O1O1O0O2O1O1O001N2O001O1O0O2O001O001N10oNjMoFU2R9mMkFS2U9QNgFo1Z9SNdFl1]9UNaFj1a9WN^Fh1c9ZNZFf1f9\\NYFc1h9^NVFb1k9_NRFa1P:aNlE`1T:bNhE`1Y:h01O1N2O001O000O2O00001O0OXO^MjFb2S9cMkF]2R9hMmFW2Q9mMmFS2S9PNjFo1W9SNhFl1X9VNfFj1Y9YNeFg1[9\\NaFe1_9]N^Fd1b9P12M3M3M30000000000000000000000O101O1O1O001O1O1O001O1O1O001O1O1O001O10OWNdEf0Z:XOiEh0U:XOmEi0Q:VOQFj0n9TOTFm0k9dNmEL9b1i9aNQFI77Gn0P:ROVFC5:HS1k9POhFEEj1T9aNiGg1o7YNQHg1o7YNQHg1o7YNPHi1P8VNPHj1P8VNPHj1P8VNPHj1P8VNPHj1P8UNQHk1R8QNoGj0eN1a9oNkGP1hNMm:3WEIi:7ZEFf::^EBb:>bE^O]:b0hEZOX:f0lEVOT:j0oESOQ:m0nETOR:l0lEWOS:i0lEYOS:g0lE[OS:e0lE]OS:b0mE@R:`0mEAS:?lEBT:>jEDV:<iEEV:<iEEW:;hEEY:;jEBV:>lE@T:?oE_OQ:a0QF]Oo9c0TFZOl9f0VFROP:n0V100000O0100000000000000O10000000O10000000O10000000000000O0100000000000000O10000000O100000O1000000000000000O0100000000000000O10000000O1kCQOo;U100000O10000000000O10O10000000O1000000000000O10O10000000O1000000000000O01000000000O104L6J5K6J6Jgb0"}, "label": "military flight number 854 in yellow handle"}]}
{"id": 148602, "image": "COCO_train2014_000000148602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two airplanes in a field with yellow wings and red and white striped tails . one in numbered 001 and the other 854\"."}], "task": "refering", "answer_template": "The \"two airplanes in a field with yellow wings and red and white striped tails . one in numbered 001 and the other 854\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 149, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 247, 248, 264, 265], "bbox": [0.10071875, 0.38967136150234744, 0.9977656250000001, 0.755868544600939], "iscrowd": 0, "area": 31780.317150000017, "rle": {"size": [426, 640], "counts": "jXk03V=1N3N2M3N2O1O1N101O0O100000001N100000001O0O100000001O0O100000000O10000000000O2O0000000O100001O000000000001O00000000000000000001O000001O0000000000000000000000000010O0000000000000000000000000001O000001O00000000000000000001O00000000000001O000000000001O000000000000000000000000001O01O000001gCTOR<R100000000000000000000000000YDnNV;R1fDROZ;n0cDVO\\;j0cDWO];i0bDYO^;f0bDZO^;g0`D[O_;e0aD\\O^;d0aD]O_;c0aD^O^;b0bD^O^;b0aD@^;`0bD@^;V10001O0000000000001O000001O0001O0000000000001O0000000000001O00000000010O00000000000000000000O10000000000O1000001O000O10000000000O1000O101O2N2N1O2N1O2N2N1O2N1B>I6K4L5L3N3N2M2N3N1O2O13L5L3M\\FbL`9Y3_FlLa9Q3]FSMd9i2ZF]Mg9Q33L5L3L5C<H9K4M4K3M010O1O100O1O01N101O001oDRNe:o1YEVNc:k1\\EYN_:j1_EZN]:g1aE\\N]:Y2O1O1N2O1O1O1O1OO0N3M3F:1O1N2O1O001O1O100O1O010001O001O1N101O1O1O0O2O1O1O001N2O001O1O0O2O001O001N10oNjMoFU2R9mMkFS2U9QNgFo1Z9SNdFl1]9UNaFj1a9WN^Fh1c9ZNZFf1f9\\NYFc1h9^NVFb1k9_NRFa1P:aNlE`1T:bNhE`1Y:h01O1N2O001O000O2O00001O0OXO^MjFb2S9cMkF]2R9hMmFW2Q9mMmFS2S9PNjFo1W9SNhFl1X9VNfFj1Y9YNeFg1[9\\NaFe1_9]N^Fd1b9P12M3M3M30000000000000000000000O101O1O1O001O1O1O001O1O1O001O1O1O001O10OWNdEf0Z:XOiEh0U:XOmEi0Q:VOQFj0n9TOTFm0k9dNmEL9b1i9aNQFI77Gn0P:ROVFC5:HS1k9POhFEEj1T9aNiGg1o7YNQHg1o7YNQHg1o7YNPHi1P8VNPHj1P8VNPHj1P8VNPHj1P8VNPHj1P8UNQHk1R8QNoGj0eN1a9oNkGP1hNMm:3WEIi:7ZEFf::^EBb:>bE^O]:b0hEZOX:f0lEVOT:j0oESOQ:m0nETOR:l0lEWOS:i0lEYOS:g0lE[OS:e0lE]OS:b0mE@R:`0mEAS:?lEBT:>jEDV:<iEEV:<iEEW:;hEEY:;jEBV:>lE@T:?oE_OQ:a0QF]Oo9c0TFZOl9f0VFROP:n0V100000O0100000000000000O10000000O10000000O10000000000000O0100000000000000O10000000O100000O1000000000000000O0100000000000000O10000000O1kCQOo;U100000O10000000000O10O10000000O1000000000000O10O10000000O1000000000000O01000000000O104L6J5K6J6Jgb0"}, "label": "two airplanes in a field with yellow wings and red and white striped tails . one in numbered 001 and the other 854"}]}
{"id": 148602, "image": "COCO_train2014_000000148602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane facing the red and white sign\"."}], "task": "refering", "answer_template": "The \"the plane facing the red and white sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195], "bbox": [0.0006562499999999999, 0.2952112676056338, 0.6310156250000001, 0.5513380281690141], "iscrowd": 0, "area": 16056.722399999991, "rle": {"size": [426, 640], "counts": "n52X=a0_O2N1O00001O0010O00000001O000000000000001O000000000000001O0000000000001O0001O00000000000001O000000000000000000000000001O00000000000001O000000000001O000000000000000000000000001O0000000000000001O00000000000000000000000000000000000000000000000000000000001O000000000000oMXO`Gh0^8[OaGe0]8^ObGb0^8_OaGa0`8^O`Gb0a8^O^Gb0b8_O]Ga0d8^O\\Gb0d8_O[Ga0f8_OYGa0g8@XG`0i8_OWGa0i8@VG`0k8@TG`0m8_OSGa0m8@RG`0o8@PG`0P9AoF?R9@nF`0R9AmF?T9AkF?V9@jF`0V9AjF>W9AiF?W9BhF>Y9AcEIf0g0f9AaENe0a0k9@^E2f0>m9_O[E4i0=l90TF0m9NTF2l9NTF2m9MSF3m9LkEZOCj0c:KiE]OCh0e:JgE@Cf0f:IgEBCe0g:HeEECc0h:HdEFDb0i:FbELC>k:F^E2E8n:c0SE]On:a0SE_Om:`0TE@l:`0TE@l:?VE@i:`0XEAg:?YEAf:a0YE_Of:c0YE]Og:d0XE\\O^:YOcE\\1O[O\\:]OcEY10ZO\\:_ObEY11XO\\:U1cEkN]:Q200000000001OfEVMT:j2lEVMT:j2kEWMU:i2kEWMV:h2iEYMW:g2iEYM`:_2^EbMb:^2^EbMb:^2^EbMc:]2]EcMc:b2001O00010OO100001O00000000000000001O001O2N2N3M2O1N00000000000000000000001O2N2N0000000000012UNjD\\1X;`NkD_1W;]NmDa1_;O1O1O1O1O1O1O1O1O4L0000O100000001N100E;D<K5L4O1O1001O001O1O3M3M3M3M3M3M4L8IO0000000O@TDDl;<TDDl;;UDEk;;UDDl;<TDDl;<TDDl;<TDDk;<VDDj;<VDDj;<VDDj;<VDDj;<VDDj;<VDCk;<VDDj;<VDDi;=WDCi;=WDCi;=WDYOS<b0=K5L4K4O2N2N`Z30beL7I7I7H5L00000jCVOl;j0TDXOj;h0VDZOh;f0XD\\Of;c0[D^Od;b0\\D^Od;b0\\D^Od;b0\\D^Oc;c0]D]Oc;c0^D[Oc;e0]D[Oc;e0]D[Oc;d0^D\\Ob;d0^D[Oc;e0^DZOb;f0^DZOb;f0^DZOb;f0^DYOb;h0_DWOa;i0_DWOa;h0aDWO_;i0a0000000000000O0100000000000O1O1O1O001G9N2N2N2N2N20000000O1000000000000000000O10000000000000LYPR3"}, "label": "the plane facing the red and white sign"}]}
{"id": 148602, "image": "COCO_train2014_000000148602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a propeller airplane with the numbers 001 on the tail\"."}], "task": "refering", "answer_template": "The \"a propeller airplane with the numbers 001 on the tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195], "bbox": [0.0006562499999999999, 0.2952112676056338, 0.6310156250000001, 0.5513380281690141], "iscrowd": 0, "area": 16056.722399999991, "rle": {"size": [426, 640], "counts": "n52X=a0_O2N1O00001O0010O00000001O000000000000001O000000000000001O0000000000001O0001O00000000000001O000000000000000000000000001O00000000000001O000000000001O000000000000000000000000001O0000000000000001O00000000000000000000000000000000000000000000000000000000001O000000000000oMXO`Gh0^8[OaGe0]8^ObGb0^8_OaGa0`8^O`Gb0a8^O^Gb0b8_O]Ga0d8^O\\Gb0d8_O[Ga0f8_OYGa0g8@XG`0i8_OWGa0i8@VG`0k8@TG`0m8_OSGa0m8@RG`0o8@PG`0P9AoF?R9@nF`0R9AmF?T9AkF?V9@jF`0V9AjF>W9AiF?W9BhF>Y9AcEIf0g0f9AaENe0a0k9@^E2f0>m9_O[E4i0=l90TF0m9NTF2l9NTF2m9MSF3m9LkEZOCj0c:KiE]OCh0e:JgE@Cf0f:IgEBCe0g:HeEECc0h:HdEFDb0i:FbELC>k:F^E2E8n:c0SE]On:a0SE_Om:`0TE@l:`0TE@l:?VE@i:`0XEAg:?YEAf:a0YE_Of:c0YE]Og:d0XE\\O^:YOcE\\1O[O\\:]OcEY10ZO\\:_ObEY11XO\\:U1cEkN]:Q200000000001OfEVMT:j2lEVMT:j2kEWMU:i2kEWMV:h2iEYMW:g2iEYM`:_2^EbMb:^2^EbMb:^2^EbMc:]2]EcMc:b2001O00010OO100001O00000000000000001O001O2N2N3M2O1N00000000000000000000001O2N2N0000000000012UNjD\\1X;`NkD_1W;]NmDa1_;O1O1O1O1O1O1O1O1O4L0000O100000001N100E;D<K5L4O1O1001O001O1O3M3M3M3M3M3M4L8IO0000000O@TDDl;<TDDl;;UDEk;;UDDl;<TDDl;<TDDl;<TDDk;<VDDj;<VDDj;<VDDj;<VDDj;<VDDj;<VDCk;<VDDj;<VDDi;=WDCi;=WDCi;=WDYOS<b0=K5L4K4O2N2N`Z30beL7I7I7H5L00000jCVOl;j0TDXOj;h0VDZOh;f0XD\\Of;c0[D^Od;b0\\D^Od;b0\\D^Od;b0\\D^Oc;c0]D]Oc;c0^D[Oc;e0]D[Oc;e0]D[Oc;d0^D\\Ob;d0^D[Oc;e0^DZOb;f0^DZOb;f0^DZOb;f0^DYOb;h0_DWOa;i0_DWOa;h0aDWO_;i0a0000000000000O0100000000000O1O1O1O001G9N2N2N2N2N20000000O1000000000000000000O10000000000000LYPR3"}, "label": "a propeller airplane with the numbers 001 on the tail"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in front of boy sitting\"."}], "task": "refering", "answer_template": "The \"person in front of boy sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 40, 41, 42, 43, 44, 45, 58, 59, 75, 76, 77, 93, 110], "bbox": [0.3828333333333333, 0.001234375, 0.6996666666666668, 0.28093749999999995], "iscrowd": 0, "area": 13234.5435, "rle": {"size": [640, 480], "counts": "QPc3?ac0d0\\O3M4L3M4L3M4L3M3M4L3M4L1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N3M3M3M2N3M3M3M?Am0SO4L00O1O100O100O1O100O100O1O100O1M3N2M3M3M3N2M3gMg@H\\?8j@_OY?`0n@XOT?h0RAoNQ?P1VAgNm>X1ZA_Ni>a1^AUNe>j1bAmMa>S2f1O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O10000000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000;EW1jN\\ni2"}, "label": "person in front of boy sitting"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in a grey coat and dark pants\"."}], "task": "refering", "answer_template": "The \"person in a grey coat and dark pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 40, 41, 42, 43, 44, 45, 58, 59, 75, 76, 77, 93, 110], "bbox": [0.3828333333333333, 0.001234375, 0.6996666666666668, 0.28093749999999995], "iscrowd": 0, "area": 13234.5435, "rle": {"size": [640, 480], "counts": "QPc3?ac0d0\\O3M4L3M4L3M4L3M3M4L3M4L1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N3M3M3M2N3M3M3M?Am0SO4L00O1O100O100O1O100O100O1O100O1M3N2M3M3M3N2M3gMg@H\\?8j@_OY?`0n@XOT?h0RAoNQ?P1VAgNm>X1ZA_Ni>a1^AUNe>j1bAmMa>S2f1O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O10000000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000;EW1jN\\ni2"}, "label": "person in a grey coat and dark pants"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue suitcase that a man is sitting on\"."}], "task": "refering", "answer_template": "The \"a blue suitcase that a man is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 230, 231, 232, 233, 234, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 316, 317, 318, 319, 333, 334, 335, 336], "bbox": [0.48485416666666664, 0.56490625, 0.8221458333333334, 0.8534062500000001], "iscrowd": 0, "area": 19560.185599999986, "rle": {"size": [640, 480], "counts": "loa4>]c07H7K4O2N1O1001O001O1O02O000O2O000O2O4L5J2O1O0O2O1O0O2O001N2O1O2M4M2N3L4M2N3L3N3M3L3N3M2M3N2M3M4K4M3M3M3M3M3M3M3M3M3L4M00001O0000001O000O2O0000001O00001O000O101O00001O00001O0O10001O00001O0000001N10001O0000001O00001N10001O0000001O0O10000O1000000O10000O1000O001O1O10O01O1O001O1O001O1N101N2N1O2O0O1O2N101N1O2N10b0]Oe0[Od0\\Oe0\\Od0[Od0\\Oe0[Ocdd1"}, "label": "a blue suitcase that a man is sitting on"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue suitcase sitting on the floor\"."}], "task": "refering", "answer_template": "The \"a blue suitcase sitting on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 230, 231, 232, 233, 234, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 316, 317, 318, 319, 333, 334, 335, 336], "bbox": [0.48485416666666664, 0.56490625, 0.8221458333333334, 0.8534062500000001], "iscrowd": 0, "area": 19560.185599999986, "rle": {"size": [640, 480], "counts": "loa4>]c07H7K4O2N1O1001O001O1O02O000O2O000O2O4L5J2O1O0O2O1O0O2O001N2O1O2M4M2N3L4M2N3L3N3M3L3N3M2M3N2M3M4K4M3M3M3M3M3M3M3M3M3L4M00001O0000001O000O2O0000001O00001O000O101O00001O00001O0O10001O00001O0000001N10001O0000001O00001N10001O0000001O0O10000O1000000O10000O1000O001O1O10O01O1O001O1O001O1N101N2N1O2O0O1O2N101N1O2N10b0]Oe0[Od0\\Oe0\\Od0[Od0\\Oe0[Ocdd1"}, "label": "a blue suitcase sitting on the floor"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person next to boy\"."}], "task": "refering", "answer_template": "The \"person next to boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 31, 32, 33, 47, 48, 49, 50, 64, 65, 66, 67, 81, 82, 83, 84, 98, 99, 100, 101, 115, 116, 117, 118, 133, 134, 135, 151, 152, 168, 169, 185, 186, 201, 202, 203, 219], "bbox": [0.7981874999999999, 0.0025, 1.0, 0.5308437500000001], "iscrowd": 0, "area": 23419.619650000008, "rle": {"size": [640, 480], "counts": "Qa_75Zc0d0\\Nd1I6K6I7I6J7I7I6J7J6E:E<D<D;I7M1O1O1O1N2N2N2N2N2O10000000000000000O1kEhIb6X6]ImI_6S6aIQJ[6o5dIWJW6j5hHQJUN:o8e5kHSJTN9Q9e5iHSJUN9R9d5hHUJUN7S9e5fHUJVN7T9e5eHUJUN7V9d5dHWJTN6X9d5bHWJUN6Y9c5aHYJTN5[9c5_HYJVN4[9d5]HZJVN3]9c5\\H[JVN3^9c5[H[JUN3_9d5ZH\\Kf7d4YH\\Kh7e4VH\\Kj7d4VH[Kk7f4SH[Km7e4RH[Ko7e4oG]KQ8c4nG]KS8c4lG]KU8c4jG^KV8b4iG^KX8b4fG`KZ8`4cGbK^8^4\\GhKd8X4VGmKk8d70O10O0100O100O1O1O1O1O1O100O2N1O1O1O1O101NNZFlEe9U:[FkEd9V:\\FjEc9W:^FhEa9Y:_FgE`9[:`FeE^9\\:bFdE^9\\:cN]1"}, "label": "person next to boy"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a trench coat and brown suede shoes\"."}], "task": "refering", "answer_template": "The \"a man wearing a trench coat and brown suede shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 31, 32, 33, 47, 48, 49, 50, 64, 65, 66, 67, 81, 82, 83, 84, 98, 99, 100, 101, 115, 116, 117, 118, 133, 134, 135, 151, 152, 168, 169, 185, 186, 201, 202, 203, 219], "bbox": [0.7981874999999999, 0.0025, 1.0, 0.5308437500000001], "iscrowd": 0, "area": 23419.619650000008, "rle": {"size": [640, 480], "counts": "Qa_75Zc0d0\\Nd1I6K6I7I6J7I7I6J7J6E:E<D<D;I7M1O1O1O1N2N2N2N2N2O10000000000000000O1kEhIb6X6]ImI_6S6aIQJ[6o5dIWJW6j5hHQJUN:o8e5kHSJTN9Q9e5iHSJUN9R9d5hHUJUN7S9e5fHUJVN7T9e5eHUJUN7V9d5dHWJTN6X9d5bHWJUN6Y9c5aHYJTN5[9c5_HYJVN4[9d5]HZJVN3]9c5\\H[JVN3^9c5[H[JUN3_9d5ZH\\Kf7d4YH\\Kh7e4VH\\Kj7d4VH[Kk7f4SH[Km7e4RH[Ko7e4oG]KQ8c4nG]KS8c4lG]KU8c4jG^KV8b4iG^KX8b4fG`KZ8`4cGbK^8^4\\GhKd8X4VGmKk8d70O10O0100O100O1O1O1O1O1O100O2N1O1O1O1O101NNZFlEe9U:[FkEd9V:\\FjEc9W:^FhEa9Y:_FgE`9[:`FeE^9\\:bFdE^9\\:cN]1"}, "label": "a man wearing a trench coat and brown suede shoes"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the white shirt\"."}], "task": "refering", "answer_template": "The \"the boy in the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 178, 179, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 229, 230, 231, 232, 233, 234, 235, 252], "bbox": [0.46029166666666665, 0.076546875, 0.8790208333333334, 0.666671875], "iscrowd": 0, "area": 47910.4038, "rle": {"size": [640, 480], "counts": "mYZ4:dc0f0ZO8H6J6K6J5L5J5\\@oMQ=U2bBVN\\=o1WB\\Ne=i1nAbNP>c1cAhNY>]1ZAoNc>Y3AhJQB\\5k=eJUB]5h=dJWB`5e=aJ[Bb5`=`J_Bc5^=_JaBb5]=_JbBd5Q=TJQC9Ne5o<WJmB53g5m<cJSC_5j<bJUCa5d:nIWGa0UNc5a:SJSG;[Ne5_:WJoF5bNe5^:\\JiF0hNg5\\:eJaFDSOg5\\:]KhEmNKh5[:]LeEc3[:]LeEc3Z:^LeEd3Y:]LgEc3Y:R3O101N100O1O2O0O100O2N100O101N100O2N1O100O2O000O101N100O101O0O101N10000O2O0001O01O0000010O000001O0O10001O0O10001O0O10000O2O00000O2O00000O2O00000O101O000O101iHTFX4l9eKXFZ4h9cK[F\\4f9aK]F_4d9]K_Fc4a9ZKcFd4_9XKdFh4b9oJaFQ5f9`JcF_5d9nIhFQ6`9[ImFe6X;O000O101O000O2O00000O2O000O2O00001N1000001N10001O0O10001O0O101O000O2O000O101O000O2O002N3L4M32NXMnAEl=6VB4c=fN[BRO5b2Y=YNbBSO7m2Q=mMiBRO9V2JQNl<e0QCRO;U2NWNc<?UCQO=V21UNb<a0PCRO?U24SNc<d0jBQOa0V27PNc<f0eBQOd0V29oMc<g0aBQOe0V2<nMc<h0\\BQOh0V2>mMc<i0XBQOi0V2b0kMb<j0TCX1e>aN_A]1f>\\N^Aa1h>YN[Ad1k>UNYAh1a`0M3M5L5J6J5K6J6J5K6J5K\\nS1"}, "label": "the boy in the white shirt"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy sitting on a luggage\"."}], "task": "refering", "answer_template": "The \"a boy sitting on a luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 178, 179, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 229, 230, 231, 232, 233, 234, 235, 252], "bbox": [0.46029166666666665, 0.076546875, 0.8790208333333334, 0.666671875], "iscrowd": 0, "area": 47910.4038, "rle": {"size": [640, 480], "counts": "mYZ4:dc0f0ZO8H6J6K6J5L5J5\\@oMQ=U2bBVN\\=o1WB\\Ne=i1nAbNP>c1cAhNY>]1ZAoNc>Y3AhJQB\\5k=eJUB]5h=dJWB`5e=aJ[Bb5`=`J_Bc5^=_JaBb5]=_JbBd5Q=TJQC9Ne5o<WJmB53g5m<cJSC_5j<bJUCa5d:nIWGa0UNc5a:SJSG;[Ne5_:WJoF5bNe5^:\\JiF0hNg5\\:eJaFDSOg5\\:]KhEmNKh5[:]LeEc3[:]LeEc3Z:^LeEd3Y:]LgEc3Y:R3O101N100O1O2O0O100O2N100O101N100O2N1O100O2O000O101N100O101O0O101N10000O2O0001O01O0000010O000001O0O10001O0O10001O0O10000O2O00000O2O00000O2O00000O101O000O101iHTFX4l9eKXFZ4h9cK[F\\4f9aK]F_4d9]K_Fc4a9ZKcFd4_9XKdFh4b9oJaFQ5f9`JcF_5d9nIhFQ6`9[ImFe6X;O000O101O000O2O00000O2O000O2O00001N1000001N10001O0O10001O0O101O000O2O000O101O000O2O002N3L4M32NXMnAEl=6VB4c=fN[BRO5b2Y=YNbBSO7m2Q=mMiBRO9V2JQNl<e0QCRO;U2NWNc<?UCQO=V21UNb<a0PCRO?U24SNc<d0jBQOa0V27PNc<f0eBQOd0V29oMc<g0aBQOe0V2<nMc<h0\\BQOh0V2>mMc<i0XBQOi0V2b0kMb<j0TCX1e>aN_A]1f>\\N^Aa1h>YN[Ad1k>UNYAh1a`0M3M5L5J6J5K6J6J5K6J5K\\nS1"}, "label": "a boy sitting on a luggage"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black wheeled luggage with silver handle\"."}], "task": "refering", "answer_template": "The \"black wheeled luggage with silver handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 280, 281, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 326, 327, 328, 329, 330, 331, 332, 333, 343, 344, 345, 346, 347, 348, 349, 362, 363, 364, 365], "bbox": [0.15581250000000002, 0.62246875, 0.60525, 0.9572968749999999], "iscrowd": 0, "area": 33594.460299999984, "rle": {"size": [640, 480], "counts": "mZ_1n0Ub0P1L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L3M2N1O1O2N1O2N1O1O2N1O2N1O1O2N1N3N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O101O001O001O00001O001O00001O001O00001O001O00001O010O00010O010O00010O0010O0010O0010O00010O010O0001O010O00001O010O001O00010O001O00010O001O01O01O001O01O01O001O01O01O001O01O01O001O01O01O001O01O01O0010O0001O0001NN2O1O001O1O1001N2O001O1O1O001N2O2N2N2N3M2N2M3N3M2N2N2N2M4M5K4L5K4L5K5J5L5K5K4L5K5K4K6K4L5K5K4L5K5J5L5K5K4L5K4K6K5K4L5K5K4LZWf3"}, "label": "black wheeled luggage with silver handle"}]}
{"id": 205963, "image": "COCO_train2014_000000205963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black travel bag near its owner sitting in another bag\"."}], "task": "refering", "answer_template": "The \"a black travel bag near its owner sitting in another bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 280, 281, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 326, 327, 328, 329, 330, 331, 332, 333, 343, 344, 345, 346, 347, 348, 349, 362, 363, 364, 365], "bbox": [0.15581250000000002, 0.62246875, 0.60525, 0.9572968749999999], "iscrowd": 0, "area": 33594.460299999984, "rle": {"size": [640, 480], "counts": "mZ_1n0Ub0P1L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L3M2N1O1O2N1O2N1O1O2N1O2N1O1O2N1N3N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O101O001O001O00001O001O00001O001O00001O001O00001O010O00010O010O00010O0010O0010O0010O00010O010O0001O010O00001O010O001O00010O001O00010O001O01O01O001O01O01O001O01O01O001O01O01O001O01O01O001O01O01O0010O0001O0001NN2O1O001O1O1001N2O001O1O1O001N2O2N2N2N3M2N2M3N3M2N2N2N2M4M5K4L5K4L5K5J5L5K5K4L5K5K4K6K4L5K5K4L5K5J5L5K5K4L5K4K6K5K4L5K5K4LZWf3"}, "label": "a black travel bag near its owner sitting in another bag"}]}
{"id": 402575, "image": "COCO_train2014_000000402575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older child taking bath\"."}], "task": "refering", "answer_template": "The \"older child taking bath\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 315, 316], "bbox": [0.53478125, 0.1842687747035573, 0.8226093750000001, 0.7325889328063241], "iscrowd": 0, "area": 34196.433800000006, "rle": {"size": [506, 640], "counts": "aSY5g1R>3N2M3nESNS6o1kIRNS6R2iIPNV6R2hIPNU6S2iInMV6T2hInMU6U2hImMV6V2hIkMW6W2gIkMV6X2gIjMX6X2fIiMX6Z2fIhMX6Z2eIhMY6[2eIgMY6[2eIfMY6]2eIdMY6_2dIdMZ6^2dIcMZ6`2dIbMZ6`2cIbM[6a2cI`M\\6`2dIbMY6`2eIbMZ6_2eIcMX6_2gIbMW6`2hIaMW6_2hIdMU6^2jIcMU6^2jIdMS6^2kIdMT6\\2kIfMS6\\2iIiMU6X2hIkMV6W2fImMX6T2eIQNY6P2cITN[6n1aIWN]6j1_IZN_6g1^I]Na6d1[IaNb6a1[IbNd6_1XIdNh6\\1UIgNk6Z1QIiNo6X1mHlNR7U1jHnNV7R1gHQOY7P1cHSO]7e40000000000000001O000000000000000000001O00000000000000000000001O000000001O0000000011N1O2N1O2O1N1O2N101N1O2N1O2O1N1O2N101N1O2kJiGc2X8[MjGd2X8YMjGg2V8XMkGg2W8VMjGj2X8SMjGl2W8RMkGn2W8nLkGQ3Y8jLiGU3[8gLeGY3^8cLdG\\3`8_LbGa3`8[LbGc3c8XL^Gh3e8TL]Gk3g8PL[Go3h8nKYGQ4k8jKWGU4l8gKUGX4P9cKRG\\4Q9`KQG_4V:O1O001N2O1O001O1O1N2O001O1O1O0O2O1O1O1O001N2O1O1O001O1N2O001O1O1O1O0O2O1O1O1L3K6J6J6J:F;F:E:F;E;E;EaQh1"}, "label": "older child taking bath"}]}
{"id": 402575, "image": "COCO_train2014_000000402575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older child facing his left\"."}], "task": "refering", "answer_template": "The \"an older child facing his left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 315, 316], "bbox": [0.53478125, 0.1842687747035573, 0.8226093750000001, 0.7325889328063241], "iscrowd": 0, "area": 34196.433800000006, "rle": {"size": [506, 640], "counts": "aSY5g1R>3N2M3nESNS6o1kIRNS6R2iIPNV6R2hIPNU6S2iInMV6T2hInMU6U2hImMV6V2hIkMW6W2gIkMV6X2gIjMX6X2fIiMX6Z2fIhMX6Z2eIhMY6[2eIgMY6[2eIfMY6]2eIdMY6_2dIdMZ6^2dIcMZ6`2dIbMZ6`2cIbM[6a2cI`M\\6`2dIbMY6`2eIbMZ6_2eIcMX6_2gIbMW6`2hIaMW6_2hIdMU6^2jIcMU6^2jIdMS6^2kIdMT6\\2kIfMS6\\2iIiMU6X2hIkMV6W2fImMX6T2eIQNY6P2cITN[6n1aIWN]6j1_IZN_6g1^I]Na6d1[IaNb6a1[IbNd6_1XIdNh6\\1UIgNk6Z1QIiNo6X1mHlNR7U1jHnNV7R1gHQOY7P1cHSO]7e40000000000000001O000000000000000000001O00000000000000000000001O000000001O0000000011N1O2N1O2O1N1O2N101N1O2N1O2O1N1O2N101N1O2kJiGc2X8[MjGd2X8YMjGg2V8XMkGg2W8VMjGj2X8SMjGl2W8RMkGn2W8nLkGQ3Y8jLiGU3[8gLeGY3^8cLdG\\3`8_LbGa3`8[LbGc3c8XL^Gh3e8TL]Gk3g8PL[Go3h8nKYGQ4k8jKWGU4l8gKUGX4P9cKRG\\4Q9`KQG_4V:O1O001N2O1O001O1O1N2O001O1O1O0O2O1O1O1O001N2O1O1O001O1N2O001O1O1O1O0O2O1O1O1L3K6J6J6J:F;F:E:F;E;E;EaQh1"}, "label": "an older child facing his left"}]}
{"id": 402575, "image": "COCO_train2014_000000402575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller child on left side of bathtub\"."}], "task": "refering", "answer_template": "The \"smaller child on left side of bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 258], "bbox": [0.043828125, 0.28810276679841895, 0.25828125, 0.6189130434782608], "iscrowd": 0, "area": 14551.733500000002, "rle": {"size": [506, 640], "counts": "QP>9_?d0]Oc0]O8H4L4L3L5L4L3TDcMV:a2eEaMZ:c2aE_M^:d2^E^Ma:c2]E_Ma:c2]E`Ma:b2\\E`Mc:a2[EaMe:_2YEcMg:]2WEeMi:\\2SEgMm:Y2QEiMo:W2oDkMQ;V2mDlMR;T2mDmMS;S2kDoMT;R2kDoMU;R2iDoMW;Q2gDQNY;o1fDRNZ;n1dDUN[;U300000000000001O000O100000001O001O1O001N101O00001O000000001N10000000001O0001O1O101N1O1O1O2O0O1O2N1O1O101N1O1O1O2O0O1O2N1O2O1N1O2N2N2O0_MQDU1Q<eNTDZ1n;SNdDl1f<O1O100O1O2N4L3N3L3M4L2N101N2N2N2N1O2O1N2N000001O0001O0001O000000000013L<Dj]Z7"}, "label": "smaller child on left side of bathtub"}]}
{"id": 402575, "image": "COCO_train2014_000000402575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"younger of the two boys in the tub\"."}], "task": "refering", "answer_template": "The \"younger of the two boys in the tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 258], "bbox": [0.043828125, 0.28810276679841895, 0.25828125, 0.6189130434782608], "iscrowd": 0, "area": 14551.733500000002, "rle": {"size": [506, 640], "counts": "QP>9_?d0]Oc0]O8H4L4L3L5L4L3TDcMV:a2eEaMZ:c2aE_M^:d2^E^Ma:c2]E_Ma:c2]E`Ma:b2\\E`Mc:a2[EaMe:_2YEcMg:]2WEeMi:\\2SEgMm:Y2QEiMo:W2oDkMQ;V2mDlMR;T2mDmMS;S2kDoMT;R2kDoMU;R2iDoMW;Q2gDQNY;o1fDRNZ;n1dDUN[;U300000000000001O000O100000001O001O1O001N101O00001O000000001N10000000001O0001O1O101N1O1O1O2O0O1O2N1O1O101N1O1O1O2O0O1O2N1O2O1N1O2N2N2O0_MQDU1Q<eNTDZ1n;SNdDl1f<O1O100O1O2N4L3N3L3M4L2N101N2N2N2N1O2O1N2N000001O0001O0001O000000000013L<Dj]Z7"}, "label": "younger of the two boys in the tub"}]}
{"id": 230545, "image": "COCO_train2014_000000230545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase with the green leaves\"."}], "task": "refering", "answer_template": "The \"the vase with the green leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 211, 212], "bbox": [0.58426, 0.3685538461538462, 0.9815799999999999, 0.9483076923076923], "iscrowd": 0, "area": 24847.186800000003, "rle": {"size": [325, 500], "counts": "\\jl22n96K6M3N1N3N1N3M2O2M3N1N3M5L5J7J5J6K5J6J6K5J6K5J6J6K2N2N1N2O2N1O1O2N1N2O2N2N=C>A3N001O1O001N2O001O1O1N101O1O001O1N101O1O001N2O001O000O2O000000000O2O00000000001O01O00000001O01O000001O0001O000001O0001O000001O01O00000O100O2O0O100O100O2O000O100O101N100O100O101N100O100O101N100O1O1M3L5L3M3L4M3L4M4M2N2O1N2N3M2N2O1N2N2N3M2O2M2N3M3M2O2M2N3M3M2L5L4K4L5K5K4M4K5K4L5N2M2O2N1O2M3N1O2N2M2O2N2M2O2N2N1N3N1O2NdP3"}, "label": "the vase with the green leaves"}]}
{"id": 230545, "image": "COCO_train2014_000000230545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vase on right\"."}], "task": "refering", "answer_template": "The \"vase on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 211, 212], "bbox": [0.58426, 0.3685538461538462, 0.9815799999999999, 0.9483076923076923], "iscrowd": 0, "area": 24847.186800000003, "rle": {"size": [325, 500], "counts": "\\jl22n96K6M3N1N3N1N3M2O2M3N1N3M5L5J7J5J6K5J6J6K5J6K5J6J6K2N2N1N2O2N1O1O2N1N2O2N2N=C>A3N001O1O001N2O001O1O1N101O1O001O1N101O1O001N2O001O000O2O000000000O2O00000000001O01O00000001O01O000001O0001O000001O0001O000001O01O00000O100O2O0O100O100O2O000O100O101N100O100O101N100O100O101N100O1O1M3L5L3M3L4M3L4M4M2N2O1N2N3M2N2O1N2N2N3M2O2M2N3M3M2O2M2N3M3M2L5L4K4L5K5K4M4K5K4L5N2M2O2N1O2M3N1O2N2M2O2N2M2O2N2N1N3N1O2NdP3"}, "label": "vase on right"}]}
{"id": 230545, "image": "COCO_train2014_000000230545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass jar\"."}], "task": "refering", "answer_template": "The \"a glass jar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 147, 148, 149], "bbox": [0.06426000000000001, 0.11686153846153845, 0.4236, 0.7235999999999999], "iscrowd": 0, "area": 27033.237999999998, "rle": {"size": [325, 500], "counts": "[X:8j9m0SO7J4K5L4M3L4M3M3L4M3L4M3M3L4M3M3M3N2M3M3N2M3N2M3M3N2M3N2M3M101@mKkJU4Q5PLmJQ4P5TLlJn3Q5VLlJl3Q5f0O1O1O1O1N2O1O1O001O1O1N2O1O1O1O1O1O1O1N2O001O1O1O001N2O001O1O001O1O000O100000000000000000000O100000000000000000000O100000000000000001O000O10000000001O0O10000O100O101O0O10000O100O101O0O100O1O1O1O1O2N1O1O1O1O3QKYK[4j4bKXK\\4k4aKWK\\4m4aKTK]4P5_KSK^4\\5N1N3M3M2N3M3N2M2N3M3M2N3N2M3M2N3M3M2O2M3M2N3M3M3N4K5K6J5K5F:D=D;E;EhYk2"}, "label": "a glass jar"}]}
{"id": 230545, "image": "COCO_train2014_000000230545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase which is white in colour\"."}], "task": "refering", "answer_template": "The \"a vase which is white in colour\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 147, 148, 149], "bbox": [0.06426000000000001, 0.11686153846153845, 0.4236, 0.7235999999999999], "iscrowd": 0, "area": 27033.237999999998, "rle": {"size": [325, 500], "counts": "[X:8j9m0SO7J4K5L4M3L4M3M3L4M3L4M3M3L4M3M3M3N2M3M3N2M3N2M3M3N2M3N2M3M101@mKkJU4Q5PLmJQ4P5TLlJn3Q5VLlJl3Q5f0O1O1O1O1N2O1O1O001O1O1N2O1O1O1O1O1O1O1N2O001O1O1O001N2O001O1O001O1O000O100000000000000000000O100000000000000000000O100000000000000001O000O10000000001O0O10000O100O101O0O10000O100O101O0O100O1O1O1O1O2N1O1O1O1O3QKYK[4j4bKXK\\4k4aKWK\\4m4aKTK]4P5_KSK^4\\5N1N3M3M2N3M3N2M2N3M3M2N3N2M3M2N3M3M2O2M3M2N3M3M3N4K5K6J5K5F:D=D;E;EhYk2"}, "label": "a vase which is white in colour"}]}
{"id": 173202, "image": "COCO_train2014_000000173202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wet elephant blocking the view of two other elephants\"."}], "task": "refering", "answer_template": "The \"a wet elephant blocking the view of two other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 267, 268, 269, 270, 272, 287, 288, 291, 292, 293, 311], "bbox": [0.3195, 0.24289583333333334, 0.8609843749999999, 0.7765416666666667], "iscrowd": 0, "area": 59766.74555000002, "rle": {"size": [480, 640], "counts": "bXP36i>5J6VBGd<>QCKk<:kBOQ=S1M4G8E;F:M3M3M3L4M3M4L3M3L3N2N1O2M201N2O001O001O1O001O001O1O001O0O1000O10000O1000000O10000O01001O0O10000O2O000O10001O0O10001N100001O2O1N1O2N1O2bFnKk7R4nG[Lk7g3nG^LR8d3gGaLY8`3`GfL`8\\3YGiLh8g40000O101O0000000O10000000001N3N1O2N2N2N2M3N2N1O2N2N2M3N2N001O0O2O001O1O001N101O001O1O0O2O001O001O0O`NfH]KY7b4lH[KS7e4QIXKn6g4VIXKh6g4]IUKc6j4bIRK]6n4hImJY6R5lIiJU6S5SJhJn5T5ZJgJg5U5aJfJ_5Z5fJaJ[5_5S2O100000O1000001O000O2O00001O000O2O0000001O0O101O00001N10001O00001N1000001O00001O000010O0001O0001O01O000001O0001O0001O000000001O000000001O000000001O000000001O00nG[Ji6e5UIaJg6_5WIfJg6Y5VInJf6R5XIPKh6P5UISKk6m4SITKn6l4oHWKR7h4lHZKT7f4iH\\KX7d4fH^KZ7c4cH_K]7a4`HaKb7^4\\HdKd7l50O10001N10000O101O000O10001N10000O101O0O10000O2O000O2O001O0O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1N3N1N3N1N3N1N2O2M2O2M2O2M2O1O2M2O2M3N2M3N2M3N3L3N2M3M3M4L3M3L4M3M4L3L4M3M3M4K4M3M3M3L5L^EhMa8U2]GoMd8n1YGWNh8f1VG_Nj8]1TGhNn8T1PGPOQ9[1^FkNb9Y31O1L5J5K5Lc0\\Og0ZOe0ZOf0ZOf0[OW^Y1"}, "label": "a wet elephant blocking the view of two other elephants"}]}
{"id": 173202, "image": "COCO_train2014_000000173202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female elephant standing in the water with two other elephants standing next to it\"."}], "task": "refering", "answer_template": "The \"a female elephant standing in the water with two other elephants standing next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 267, 268, 269, 270, 272, 287, 288, 291, 292, 293, 311], "bbox": [0.3195, 0.24289583333333334, 0.8609843749999999, 0.7765416666666667], "iscrowd": 0, "area": 59766.74555000002, "rle": {"size": [480, 640], "counts": "bXP36i>5J6VBGd<>QCKk<:kBOQ=S1M4G8E;F:M3M3M3L4M3M4L3M3L3N2N1O2M201N2O001O001O1O001O001O1O001O0O1000O10000O1000000O10000O01001O0O10000O2O000O10001O0O10001N100001O2O1N1O2N1O2bFnKk7R4nG[Lk7g3nG^LR8d3gGaLY8`3`GfL`8\\3YGiLh8g40000O101O0000000O10000000001N3N1O2N2N2N2M3N2N1O2N2N2M3N2N001O0O2O001O1O001N101O001O1O0O2O001O001O0O`NfH]KY7b4lH[KS7e4QIXKn6g4VIXKh6g4]IUKc6j4bIRK]6n4hImJY6R5lIiJU6S5SJhJn5T5ZJgJg5U5aJfJ_5Z5fJaJ[5_5S2O100000O1000001O000O2O00001O000O2O0000001O0O101O00001N10001O00001N1000001O00001O000010O0001O0001O01O000001O0001O0001O000000001O000000001O000000001O000000001O00nG[Ji6e5UIaJg6_5WIfJg6Y5VInJf6R5XIPKh6P5UISKk6m4SITKn6l4oHWKR7h4lHZKT7f4iH\\KX7d4fH^KZ7c4cH_K]7a4`HaKb7^4\\HdKd7l50O10001N10000O101O000O10001N10000O101O0O10000O2O000O2O001O0O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1N3N1N3N1N3N1N2O2M2O2M2O2M2O1O2M2O2M3N2M3N2M3N3L3N2M3M3M4L3M3L4M3M4L3L4M3M3M4K4M3M3M3L5L^EhMa8U2]GoMd8n1YGWNh8f1VG_Nj8]1TGhNn8T1PGPOQ9[1^FkNb9Y31O1L5J5K5Lc0\\Og0ZOe0ZOf0ZOf0[OW^Y1"}, "label": "a female elephant standing in the water with two other elephants standing next to it"}]}
{"id": 173202, "image": "COCO_train2014_000000173202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant between two others\"."}], "task": "refering", "answer_template": "The \"an elephant between two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 132, 133, 134, 157], "bbox": [0.42134375, 0.2038125, 0.8528125, 0.4105416666666667], "iscrowd": 0, "area": 7002.0396, "rle": {"size": [480, 640], "counts": "Tfn31n>1O1N2O1O1O1N2O1N2O1N2N2N2O1N2O1O1O1O1O10O01O1O10000000000001O00000O1000000000000000O10000000000000O1000000000001O0O2O1O0O2O1O0O2O1O0O2O00000O101O0O1000000O1000000000000O100000000000000O1000000000000O1000000000000O1000O100O100O100O100O1O100O100O100O1000000O101O001N10001O001N10001O0O101O00001O0000001O000000001O0000001O0000001O000000001O000000001O000001O01O01O00001O01O01O000010O0001O00010O0000010O00001O01O01O000010O010O0010O01O010O1O010O0010O01O10O0100O00100O1O010O10O01O10O01O100O010O0100O100O010O10000O010O100O10O2001O000000001O01O0000N4M3M3M4L4L[l[1"}, "label": "an elephant between two others"}]}
{"id": 83093, "image": "COCO_train2014_000000083093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a red shirt playing wii\"."}], "task": "refering", "answer_template": "The \"a man with a red shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 67, 68, 69, 70, 85, 86, 87, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213], "bbox": [0.52374, 0.18840361445783133, 0.9131800000000001, 0.9884036144578314], "iscrowd": 0, "area": 27364.831149999987, "rle": {"size": [332, 500], "counts": "kUe27o97O2O1N2N2N1O2N100O1O001O1O010O001O001O010O00001O5K?A8H5L4K4L5K5K4L000001O00000000000000001O000000000000001O0000O10000[M]MaMc2[2jM\\MV2`2QN]Mo1`2TN`Ml1`2UN_Mk1b2TN^Ml1b2UN]Mk1c2UN]Mk1d2TN\\Ml1d2UN[Mk1f2TNZMl1f2UNYMk1g2UNYMk1h2TNXMl1h2UNVMl1k2SNUMm1k2SNUMm1l2SNSMm1n2RNRMn1o2QNQMo1P3QNoLo1R3QNmLo1T3SNiLm1X3UNeLk1\\3XN`Lh1`3[N]Le1d3]NYLc1h3_NULa1l3bNPL^1Q4dNlK\\1U4dNjK\\1W4dNgK]1Z4cNeK]1\\4cNcK]1^4dN`K\\1a4dN^K\\1d4cN[K]1g4bNXK^1j4aNUK_1l4aNSK_1o4`NPK`1R5]NoJc1S5[NmJe1T5ZNlJf1V5WNkJi1W5UNiJk1Y5SNgJm1V5UNhJn1U5UNgJo1W5SNfJP2W5RNgJQ2V5RNgJQ2V5RNfJR2W5PNgJS2V5PNgJS2W5b1\\NQKhMR5W2QKdMR5l1iJXM8h0R5n1iJWM;d0o4T2hJVMa0=i4[2hJVMh05c4c2hJVMh02c4f2hJUMi00b4i2i1N2N2O1N2N2N2N2O1O100O010O100O010O10O0100O10O0100O10O10000O2O001N100O2O0O2O0O101N1O2O0O1YJ]Me3d2YL^Me3c2ZL`Md3a2ZLbMc3_2\\LdMb3\\2]LgM`3[2]LhMb3X2]LkM`3W2^LlM`3T2_LoM^3S2_LQN_3o1_LTN_3m1_LWN_3j1]L[N`3f1^L^N`3f1ZL^Nc3f1XL]Ng3g1SL]Nj3h1PL\\Nn3h1lK\\NQ4i1iK]NS4R4iNfISN_6[1ZJZNj5S1oJbNV5l0_2^Oc0^O^h="}, "label": "a man with a red shirt playing wii"}]}
{"id": 83093, "image": "COCO_train2014_000000083093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young man in orange shirt on the right\"."}], "task": "refering", "answer_template": "The \"young man in orange shirt on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 67, 68, 69, 70, 85, 86, 87, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213], "bbox": [0.52374, 0.18840361445783133, 0.9131800000000001, 0.9884036144578314], "iscrowd": 0, "area": 27364.831149999987, "rle": {"size": [332, 500], "counts": "kUe27o97O2O1N2N2N1O2N100O1O001O1O010O001O001O010O00001O5K?A8H5L4K4L5K5K4L000001O00000000000000001O000000000000001O0000O10000[M]MaMc2[2jM\\MV2`2QN]Mo1`2TN`Ml1`2UN_Mk1b2TN^Ml1b2UN]Mk1c2UN]Mk1d2TN\\Ml1d2UN[Mk1f2TNZMl1f2UNYMk1g2UNYMk1h2TNXMl1h2UNVMl1k2SNUMm1k2SNUMm1l2SNSMm1n2RNRMn1o2QNQMo1P3QNoLo1R3QNmLo1T3SNiLm1X3UNeLk1\\3XN`Lh1`3[N]Le1d3]NYLc1h3_NULa1l3bNPL^1Q4dNlK\\1U4dNjK\\1W4dNgK]1Z4cNeK]1\\4cNcK]1^4dN`K\\1a4dN^K\\1d4cN[K]1g4bNXK^1j4aNUK_1l4aNSK_1o4`NPK`1R5]NoJc1S5[NmJe1T5ZNlJf1V5WNkJi1W5UNiJk1Y5SNgJm1V5UNhJn1U5UNgJo1W5SNfJP2W5RNgJQ2V5RNgJQ2V5RNfJR2W5PNgJS2V5PNgJS2W5b1\\NQKhMR5W2QKdMR5l1iJXM8h0R5n1iJWM;d0o4T2hJVMa0=i4[2hJVMh05c4c2hJVMh02c4f2hJUMi00b4i2i1N2N2O1N2N2N2N2O1O100O010O100O010O10O0100O10O0100O10O10000O2O001N100O2O0O2O0O101N1O2O0O1YJ]Me3d2YL^Me3c2ZL`Md3a2ZLbMc3_2\\LdMb3\\2]LgM`3[2]LhMb3X2]LkM`3W2^LlM`3T2_LoM^3S2_LQN_3o1_LTN_3m1_LWN_3j1]L[N`3f1^L^N`3f1ZL^Nc3f1XL]Ng3g1SL]Nj3h1PL\\Nn3h1lK\\NQ4i1iK]NS4R4iNfISN_6[1ZJZNj5S1oJbNV5l0_2^Oc0^O^h="}, "label": "young man in orange shirt on the right"}]}
{"id": 83093, "image": "COCO_train2014_000000083093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in brown sweatshirt playing wii\"."}], "task": "refering", "answer_template": "The \"a boy in brown sweatshirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 60, 61, 62, 78, 79, 80, 96, 97, 98, 99, 113, 114, 115, 116, 117, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 202, 203, 205, 206], "bbox": [0.2196, 0.22265060240963855, 0.53258, 0.9882831325301206], "iscrowd": 0, "area": 18521.08385, "rle": {"size": [332, 500], "counts": "QoS18S:1O1O1O001O1O1O1O10O1000O1000O10O10000O01000000O10O1K5K5J6K5K5XMhNVL^1i3eNPLa1n3bNjKd1U4_NdKf1Z4]N_Ki1`4ZN[Ki1d4YNZKg1g4YNXKh1h4XNWKh1i4YNVKh1j4WNVKi1k4WNTKj1l4VNSKk1m4UNRKk1o4UNPKl1P5TNoJl1R5TNmJm1S5SNlJa1POWNT68kJ`1VOTNP6<iJ^1\\OTNk5>hJ]1ARNh5a0fJ[1o5eNPJZ1R6fNnIX1T6hNkIV1X6jNhIT1Z6lNeIR1^6nNbIP1`6PO_IR1`6nN`IS1_6mN`IU1_6kNaIV1^6jNaIX1^6hNbIY1]6gNbIZ1^6fNbI[1\\6fNcI\\1g5TNQJc08Z1f3YNTM0kN`0:Y1b3_NnL0TO9<Y1`3cNfL1^O3;[1_3fNbLND1;[1_3jN]LLHO;]1^3lN[LJKM<^1\\3oNYLGOL;`1[3QOVLF3I<a1Z36ZLXN;d1Y37ZLUN=e1W3:YLRN>e1Y3[1aLfN]3a1\\LaNb3V4O1N3M2O1O2N2N2N2N2LgHQM[7m24O1N2O0O2O0O]NZMaKf2^4[McKc2]4^McKb2\\4`MbK`2^4aMaK`2^4aMaK_2_4bM`K_2_4cM_K]2a4dM^K]2a4dM^K\\2a4gM]KZ2b4gM]KY2b4iM]KX2b4jM\\KV2K[Mc3a0aLW2@aMo39aLY2VOeMW45aLV3^3kLaLV3]3lLbLT3]3oLaLR3^3oLaLR3]3PMbLQ3]3QMaLo2^3SMaLn2^3SMaLn2]3UMaLl2^3UMaLk2^3WMaLj2]3YMaLh2^3YMaLg2^3[MaLf2^3\\M`Le2^3]MaLd2^3]MaLc2^3`M`La2_3`M`La2^3aMaL`2]3bMbL^2^3cMaL^2]3dMbL^2\\3cMcL_2Z3cMeL_2Y3bMfL`2W3bMhL`2U3_2M2M3N2M4L;F8G2L5]OjGgN\\8g0aG]OP9>`f[2"}, "label": "a boy in brown sweatshirt playing wii"}]}
{"id": 83093, "image": "COCO_train2014_000000083093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a brown sweater playing a video game\"."}], "task": "refering", "answer_template": "The \"a boy in a brown sweater playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 60, 61, 62, 78, 79, 80, 96, 97, 98, 99, 113, 114, 115, 116, 117, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 202, 203, 205, 206], "bbox": [0.2196, 0.22265060240963855, 0.53258, 0.9882831325301206], "iscrowd": 0, "area": 18521.08385, "rle": {"size": [332, 500], "counts": "QoS18S:1O1O1O001O1O1O1O10O1000O1000O10O10000O01000000O10O1K5K5J6K5K5XMhNVL^1i3eNPLa1n3bNjKd1U4_NdKf1Z4]N_Ki1`4ZN[Ki1d4YNZKg1g4YNXKh1h4XNWKh1i4YNVKh1j4WNVKi1k4WNTKj1l4VNSKk1m4UNRKk1o4UNPKl1P5TNoJl1R5TNmJm1S5SNlJa1POWNT68kJ`1VOTNP6<iJ^1\\OTNk5>hJ]1ARNh5a0fJ[1o5eNPJZ1R6fNnIX1T6hNkIV1X6jNhIT1Z6lNeIR1^6nNbIP1`6PO_IR1`6nN`IS1_6mN`IU1_6kNaIV1^6jNaIX1^6hNbIY1]6gNbIZ1^6fNbI[1\\6fNcI\\1g5TNQJc08Z1f3YNTM0kN`0:Y1b3_NnL0TO9<Y1`3cNfL1^O3;[1_3fNbLND1;[1_3jN]LLHO;]1^3lN[LJKM<^1\\3oNYLGOL;`1[3QOVLF3I<a1Z36ZLXN;d1Y37ZLUN=e1W3:YLRN>e1Y3[1aLfN]3a1\\LaNb3V4O1N3M2O1O2N2N2N2N2LgHQM[7m24O1N2O0O2O0O]NZMaKf2^4[McKc2]4^McKb2\\4`MbK`2^4aMaK`2^4aMaK_2_4bM`K_2_4cM_K]2a4dM^K]2a4dM^K\\2a4gM]KZ2b4gM]KY2b4iM]KX2b4jM\\KV2K[Mc3a0aLW2@aMo39aLY2VOeMW45aLV3^3kLaLV3]3lLbLT3]3oLaLR3^3oLaLR3]3PMbLQ3]3QMaLo2^3SMaLn2^3SMaLn2]3UMaLl2^3UMaLk2^3WMaLj2]3YMaLh2^3YMaLg2^3[MaLf2^3\\M`Le2^3]MaLd2^3]MaLc2^3`M`La2_3`M`La2^3aMaL`2]3bMbL^2^3cMaL^2]3dMbL^2\\3cMcL_2Z3cMeL_2Y3bMfL`2W3bMhL`2U3_2M2M3N2M4L;F8G2L5]OjGgN\\8g0aG]OP9>`f[2"}, "label": "a boy in a brown sweater playing a video game"}]}
{"id": 328855, "image": "COCO_train2014_000000328855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown patterned lunch bag and the pink strap\"."}], "task": "refering", "answer_template": "The \"the brown patterned lunch bag and the pink strap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241], "bbox": [0.293921875, 0.49040302267002517, 0.54121875, 0.7818136020151133], "iscrowd": 0, "area": 10239.948950000002, "rle": {"size": [397, 640], "counts": "RUY22U<8I7L4M3N1O0O2O0O2O1O0O2O0000001O000O2O0000001O00001O1O001O1O1O1O1N2O1O00001O00000000000O2O000000000O10001O00000O1000001O000O1O1N2N2N2O1N2N2N2O1N2N2N3N1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N200O1O0000001O00001O01O01O00000010O1O100O1O100O1O101O000000000000000000001O001O1O1O1O1O1O001O1O1O1aNcFJ`92fFF`98eF^Ob9a0dFSOc9k0P1O2N2O1N2N2O1L4K4L5Kgha3"}, "label": "the brown patterned lunch bag and the pink strap"}]}
{"id": 328855, "image": "COCO_train2014_000000328855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink patterned bag in the middle of a school lunch table\"."}], "task": "refering", "answer_template": "The \"a pink patterned bag in the middle of a school lunch table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241], "bbox": [0.293921875, 0.49040302267002517, 0.54121875, 0.7818136020151133], "iscrowd": 0, "area": 10239.948950000002, "rle": {"size": [397, 640], "counts": "RUY22U<8I7L4M3N1O0O2O0O2O1O0O2O0000001O000O2O0000001O00001O1O001O1O1O1O1N2O1O00001O00000000000O2O000000000O10001O00000O1000001O000O1O1N2N2N2O1N2N2N2O1N2N2N3N1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N200O1O0000001O00001O01O01O00000010O1O100O1O100O1O101O000000000000000000001O001O1O1O1O1O1O001O1O1O1aNcFJ`92fFF`98eF^Ob9a0dFSOc9k0P1O2N2O1N2N2O1L4K4L5Kgha3"}, "label": "a pink patterned bag in the middle of a school lunch table"}]}
{"id": 328855, "image": "COCO_train2014_000000328855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white table that kids are using to eat\"."}], "task": "refering", "answer_template": "The \"white table that kids are using to eat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.150546875, 0.32808564231738035, 0.8280156249999999, 0.9775314861460956], "iscrowd": 0, "area": 67841.50605, "rle": {"size": [397, 640], "counts": "[oU15T<4L4K5L4L4L4K5L4L4K5L4L4K5L4L4L4K5L4L4K5L4L4K5L4L4L4K5L4L4K5L4L4K5L4L4L4K5L4J6K5J6J6K5J6J6K5J6J6K5J6J6K5N20000000000000000O10000000000000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000001O1O1O1O1O1O00100O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O00cXZ1"}, "label": "white table that kids are using to eat"}]}
{"id": 328855, "image": "COCO_train2014_000000328855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large table near people in a room\"."}], "task": "refering", "answer_template": "The \"a large table near people in a room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.150546875, 0.32808564231738035, 0.8280156249999999, 0.9775314861460956], "iscrowd": 0, "area": 67841.50605, "rle": {"size": [397, 640], "counts": "[oU15T<4L4K5L4L4L4K5L4L4K5L4L4K5L4L4L4K5L4L4K5L4L4K5L4L4L4K5L4L4K5L4L4K5L4L4L4K5L4J6K5J6J6K5J6J6K5J6J6K5J6J6K5N20000000000000000O10000000000000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000001O1O1O1O1O1O00100O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O00cXZ1"}, "label": "a large table near people in a room"}]}
{"id": 91288, "image": "COCO_train2014_000000091288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey van that is parked\"."}], "task": "refering", "answer_template": "The \"a grey van that is parked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 160, 161, 162, 163, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 197, 198, 199, 200, 201, 211, 212, 213, 226, 227, 228], "bbox": [0.11175644028103045, 0.469546875, 0.9929274004683841, 0.698640625], "iscrowd": 0, "area": 10093.685249999999, "rle": {"size": [640, 427], "counts": "V]n03lc04L5K2N2QOCX^O`0da0HV^O9ha03l]OORb05j]OMUb06h]OKVb0Q1O2M101O1N101N2O0O100O100O100O010O10000O100O100O10_O^^OcNba0Q1l^OnNTa0n0R_OPOm`0m0X_OSOg`0i0^_OVOa`0i0b_OVO^`0h0e_OWO[`0g0h_OXOW`0g0l_OXOT`0f0P@XOo?g0f1N2O001O1O010O1O001O001000O1000O0100O1O1O001O1O1O10O01O1O1O1O1O001O1O1O1O001O100O1O001O1O1O1O1O001O1O100O001O1O1O1O001O1O100O1O001O1O1000O01000000O10O10O10000O01000O1000O10000000000O100000000O1000000O1O1Eh]OnNXb0f0T^OYOla0=_^OCaa02j^OMWa03h^OMYa03g^OMYa03g^OLZa04f^OLZa04f^OK[a05e^OK[a05e^OJ]a05c^OJ^a06b^OJ^a06V100001O000000000000000000000000000000000000000000000001O000000a\\OOob0`01O001O1O1O1O1O1O10HCm\\O=Sc0Hh\\O8Xc0800000O1000000000000000000000O10000000000000O100000000000000000000000000000000000000O1000000000O1000000000000000000000000004K>CeS8:mkG4N2000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O10000000000000001O000000000000000000000000001O00001O1O001O00001O000O1000001O00001O001O00010O2N2O1N3Noa1"}, "label": "a grey van that is parked"}]}
{"id": 91288, "image": "COCO_train2014_000000091288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark van\"."}], "task": "refering", "answer_template": "The \"the dark van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 160, 161, 162, 163, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 197, 198, 199, 200, 201, 211, 212, 213, 226, 227, 228], "bbox": [0.11175644028103045, 0.469546875, 0.9929274004683841, 0.698640625], "iscrowd": 0, "area": 10093.685249999999, "rle": {"size": [640, 427], "counts": "V]n03lc04L5K2N2QOCX^O`0da0HV^O9ha03l]OORb05j]OMUb06h]OKVb0Q1O2M101O1N101N2O0O100O100O100O010O10000O100O100O10_O^^OcNba0Q1l^OnNTa0n0R_OPOm`0m0X_OSOg`0i0^_OVOa`0i0b_OVO^`0h0e_OWO[`0g0h_OXOW`0g0l_OXOT`0f0P@XOo?g0f1N2O001O1O010O1O001O001000O1000O0100O1O1O001O1O1O10O01O1O1O1O1O001O1O1O1O001O100O1O001O1O1O1O1O001O1O100O001O1O1O1O001O1O100O1O001O1O1000O01000000O10O10O10000O01000O1000O10000000000O100000000O1000000O1O1Eh]OnNXb0f0T^OYOla0=_^OCaa02j^OMWa03h^OMYa03g^OMYa03g^OLZa04f^OLZa04f^OK[a05e^OK[a05e^OJ]a05c^OJ^a06b^OJ^a06V100001O000000000000000000000000000000000000000000000001O000000a\\OOob0`01O001O1O1O1O1O1O10HCm\\O=Sc0Hh\\O8Xc0800000O1000000000000000000000O10000000000000O100000000000000000000000000000000000000O1000000000O1000000000000000000000000004K>CeS8:mkG4N2000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O10000000000000001O000000000000000000000000001O00001O1O001O00001O000O1000001O00001O001O00010O2N2O1N3Noa1"}, "label": "the dark van"}]}
{"id": 91288, "image": "COCO_train2014_000000091288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car\"."}], "task": "refering", "answer_template": "The \"a white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 214, 215, 216, 217, 220, 221, 222, 223, 224, 228, 229, 230, 231, 232, 235, 237, 238, 239, 243, 244, 245, 246, 247, 250, 252, 253, 254, 261, 262], "bbox": [0.232248243559719, 0.52371875, 1.0, 0.7505468750000001], "iscrowd": 0, "area": 18003.097549999988, "rle": {"size": [640, 427], "counts": "mmn12nc01N2O1N2O1O1N2O1M4L3M6K2M2N3M2O2N1N3N1O2M2O1O1N2O1O1N3N1O1N2O1O1O101O0O10000O1000001N10000O10000O101O000O10000O101O0O10000O1000000O10000O2O001O001O0O2O001O001N101O0000000O1000001MYMk^Og2Va01000000000VNn^Of0Pa0[OQ_Oe0o`0[OQ_Oe0Pa0ZOP_Of0Pa0ZOP_Of0Pa0ZOP_Of0o`0[OQ_Od0Pa0\\OP_Od0o`0\\OR_Od0n`0\\OR_Od0n`0\\OR_Od0n`0\\OQ_Oe0o`0[OQ_Oe0o`0[OQ_Od0Ra0YOo^Og0da0fNo^Og0Za0PO^dS10]]lN6I100000000000001O00000000000000000000000e^ONh>2k@;U?E]@i0c?WOY@m0g?SOY@m0g?SOU_O2`0k0[`0Hd_O8\\`0Ic_O7]`0Jb_O6^`0Ka_O5^`0La_O5_`0L`_O4``0M__OBRa0?W11O1O00000000000000000o_7N__HVOS^Oi0^a0Fb^O:^a0Fb^O:^a0Fb^O;]a0Ec^O;]a0Ec^O;]a0Ec^O;]a0Ec^O;]a0Ec^O<\\a0Dd^O<\\a0Dd^O<\\a0Dd^O<\\a0Dd^O=[a0Ce^Of0Ra0ZOn^Of0JFn?DX@f0KEm?DY@g0JEm?DY@g0JEm?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DX@f0MFk?DX@f0MFk?DX@f0MFk?DX@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEm?DV@g0MEm?DV@g0MEm?DV@g0MEm?DV@g0MEm?DV@g0MEm?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NEm?EU@f0NEm?EU@f0NEm?EU@f0NEm?EU@f0NEm?EU@f0NEm?EU@e0OFl?EU@e0OFl?EU@e0OFl?ET@f00El?ET@f00El?ET@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?Dg@3\\O9m?D[A_OhNm0m?DkA;U>EkA;U>EkA;U>EkA;U>EkA;U>EkA:V>]ORBd0n=RObBh0n2"}, "label": "a white car"}]}
{"id": 91288, "image": "COCO_train2014_000000091288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car is white\"."}], "task": "refering", "answer_template": "The \"the car is white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 214, 215, 216, 217, 220, 221, 222, 223, 224, 228, 229, 230, 231, 232, 235, 237, 238, 239, 243, 244, 245, 246, 247, 250, 252, 253, 254, 261, 262], "bbox": [0.232248243559719, 0.52371875, 1.0, 0.7505468750000001], "iscrowd": 0, "area": 18003.097549999988, "rle": {"size": [640, 427], "counts": "mmn12nc01N2O1N2O1O1N2O1M4L3M6K2M2N3M2O2N1N3N1O2M2O1O1N2O1O1N3N1O1N2O1O1O101O0O10000O1000001N10000O10000O101O000O10000O101O0O10000O1000000O10000O2O001O001O0O2O001O001N101O0000000O1000001MYMk^Og2Va01000000000VNn^Of0Pa0[OQ_Oe0o`0[OQ_Oe0Pa0ZOP_Of0Pa0ZOP_Of0Pa0ZOP_Of0o`0[OQ_Od0Pa0\\OP_Od0o`0\\OR_Od0n`0\\OR_Od0n`0\\OR_Od0n`0\\OQ_Oe0o`0[OQ_Oe0o`0[OQ_Od0Ra0YOo^Og0da0fNo^Og0Za0PO^dS10]]lN6I100000000000001O00000000000000000000000e^ONh>2k@;U?E]@i0c?WOY@m0g?SOY@m0g?SOU_O2`0k0[`0Hd_O8\\`0Ic_O7]`0Jb_O6^`0Ka_O5^`0La_O5_`0L`_O4``0M__OBRa0?W11O1O00000000000000000o_7N__HVOS^Oi0^a0Fb^O:^a0Fb^O:^a0Fb^O;]a0Ec^O;]a0Ec^O;]a0Ec^O;]a0Ec^O;]a0Ec^O<\\a0Dd^O<\\a0Dd^O<\\a0Dd^O<\\a0Dd^O=[a0Ce^Of0Ra0ZOn^Of0JFn?DX@f0KEm?DY@g0JEm?DY@g0JEm?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DY@f0KFl?DX@f0MFk?DX@f0MFk?DX@f0MFk?DX@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEl?EW@f0MEm?DV@g0MEm?DV@g0MEm?DV@g0MEm?DV@g0MEm?DV@g0MEm?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NFl?DV@f0NEm?EU@f0NEm?EU@f0NEm?EU@f0NEm?EU@f0NEm?EU@f0NEm?EU@e0OFl?EU@e0OFl?EU@e0OFl?ET@f00El?ET@f00El?ET@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?DT@f0OFm?Dg@3\\O9m?D[A_OhNm0m?DkA;U>EkA;U>EkA;U>EkA;U>EkA;U>EkA:V>]ORBd0n=RObBh0n2"}, "label": "the car is white"}]}
{"id": 91288, "image": "COCO_train2014_000000091288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red car\"."}], "task": "refering", "answer_template": "The \"the red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 192, 198, 199, 200, 201, 202, 213, 214, 215, 216, 228, 229], "bbox": [0.19569086651053866, 0.500109375, 1.0, 0.695640625], "iscrowd": 0, "area": 8003.236899999999, "rle": {"size": [640, 427], "counts": "lld1d0Xc05J6K5J6M4M2M3N2N1N2O2N1N2O101O0000001N10001O0000O0O2N2N2N2N2N1O2N2N2N2N2N2N1N3N2N2N2O1O01000O10000O01000O10000O10O1000O10000O01000O10000O0100000O1000O010000O100000O0100000000O010000000O10O1000O100000O10O10000000O010O100O00100O100O00100O10ZOo\\O=Pc0CR]O;nb0ES]O;mb0ET]O:kb0FV]O:ib0FY]O7ib0IW]O4kb0LW]O0lb0OV]ONlb02T]OKob04?O100O1O100O1O10S`23h_M2000000O10000000000O1@`0O1N2O1O1N2O1N2O1O100000000000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O01000O10000O100O10000O10000O10000O10000O100O10000O10000O10X`7En_H5L400000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000009GgE"}, "label": "the red car"}]}
{"id": 91288, "image": "COCO_train2014_000000091288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red vehicle between two other vehicles\"."}], "task": "refering", "answer_template": "The \"a red vehicle between two other vehicles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 192, 198, 199, 200, 201, 202, 213, 214, 215, 216, 228, 229], "bbox": [0.19569086651053866, 0.500109375, 1.0, 0.695640625], "iscrowd": 0, "area": 8003.236899999999, "rle": {"size": [640, 427], "counts": "lld1d0Xc05J6K5J6M4M2M3N2N1N2O2N1N2O101O0000001N10001O0000O0O2N2N2N2N2N1O2N2N2N2N2N2N1N3N2N2N2O1O01000O10000O01000O10000O10O1000O10000O01000O10000O0100000O1000O010000O100000O0100000000O010000000O10O1000O100000O10O10000000O010O100O00100O100O00100O10ZOo\\O=Pc0CR]O;nb0ES]O;mb0ET]O:kb0FV]O:ib0FY]O7ib0IW]O4kb0LW]O0lb0OV]ONlb02T]OKob04?O100O1O100O1O10S`23h_M2000000O10000000000O1@`0O1N2O1O1N2O1N2O1O100000000000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O01000O10000O100O10000O10000O10000O10000O100O10000O10000O10X`7En_H5L400000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000009GgE"}, "label": "a red vehicle between two other vehicles"}]}
{"id": 124055, "image": "COCO_train2014_000000124055.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep laying in front of another sheep\"."}], "task": "refering", "answer_template": "The \"a sheep laying in front of another sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 197, 198, 199, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 351, 352, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 374, 375, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 397, 403, 404, 406, 407, 408, 409, 410, 432, 433, 456, 457, 479, 480, 481, 503, 504], "bbox": [0.26309375, 0.358984375, 0.9591875, 0.9424375000000002], "iscrowd": 0, "area": 84011.80680000006, "rle": {"size": [640, 640], "counts": "jRZ35gc05L4L4L3M4L3M3\\NKZ_O9a`06R_OLk`0g1K4L4L4L4M2M4M3M3M1N3N1O2N1O2M200010O01O01O000O101TOW@`Mj?U2SAYMm>h2U10O1O1000000O100000001O1O1O0O2O1O1O001O1O0O2O1O1N101N2N2N2N3N1N2N2N2O1N2N2N2N3M2N3M2N3M2N3N2N101N2O1N101N2O0O2O1N2O0O2O1O1N101N2O0O2O1O1N101N2O1O0O3N1N2O1O1N101N2O001O1O1N101O1O001O1N101O001O1O001O1O001O0O2O1O001O001O1O0O2O1O001O1O1O001O1O1O0O2O1O1O1O001O1O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O00001O00001O0000001O0001N10000000000O1000000O1O1ROoClIR<T6PDiIQ<X6oCfIS<Y6oCdIR<\\6PDaIR<_6oC^IR<c6TDUIm;k6e00O10O0100O010O1O010O10O010O01O10O01O10O01O10O01O10O0010O00010O0001O01O01O00010O00010O001O01O01O00010O00010O001O00001O001O01O01O001O0000O1O1O1O1N3O0O100O100O100O100O100000000O1000000O100000000O1000000O1O101N2O1N2N2O1N2O1N2O1N2O1O1N2O1O1O2N1O1O1O1O1O1N2O1O1O100O2N2O0O2N2O^NnDUJQ;k5PEVJn:j5TEVJk:h5XEXJf:h5[EXJe:f5^EZJ`:f5aE[J]:e5eE[JZ:c5hE]JW:c5kE]JT:a5nE`JP:`5RF_Jn9`5SFaJk9^5XFbJg9]5ZFgJa9X5bFkJZ9T5gFlJX9S5kFlJT9T5nFkJR9T5oFmJo8R5TGmJl8R5UGnJj8Q5YGnJf8R5[GnJd8R5^GmJ`8R5dGmJ[8R5gGnJX8P5lGPKQ8o4RHQKn7l4VHSKi7k4ZHUKf7h4^HXKa7<YFW2Z2]M[74fFZ2R2cMW7DXG\\2m1PNi6gN_F6W1a2P2cNY6eNeFL[1[2W2TOh5dNcHj1Q2C[5bNgHe1S2HV5bNjH`1U2MR5aNkHP1d2?a4_NnHh0k2h0X4]NoHg0m2l0h6PO[IQ1e6lN]IT1c6jN_IV1b6gN`IY1d<01O0000000O1O1O1N3N1O1O1O1N2O1O1O001N101O1O001O1N101O1N1O2O0O2N2O0O2N2O0O2N1O;Ffe?"}, "label": "a sheep laying in front of another sheep"}]}
{"id": 124055, "image": "COCO_train2014_000000124055.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheep is propped on the ground next to a sheep lying down\"."}], "task": "refering", "answer_template": "The \"a white sheep is propped on the ground next to a sheep lying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 348], "bbox": [0.002140625, 0.21803124999999998, 0.399140625, 0.664375], "iscrowd": 0, "area": 46207.81884999999, "rle": {"size": [640, 640], "counts": "Qm01jc0?B=B?B>A>C8G5L5M2N3M3M2O2M2N3M3N1N3M2N3N1N3M3M2O2M2N3N2M2O2N1O2N1O2N2N1N3N1O2N2N2N1O2N2N2N1O2N2N2N2O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1N2O1O001O2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N0O1000O1000000O100000WOjHhDV7X;jHhDU7X;mHgDS7Y;mHgDS7Y;nHfDR7Z;oHeDP7\\;PIdDP7[;RIdDn6\\;RIdDn6\\;SIcDm6];SIcDP6^OcJP<^OaDP6HXJg;I`Do53nI\\;4`DP6Z<QJeCo5[<QJeCo5[<RJdCo5[<RJdCn5[<SJeCn5Z<SJeCm5[<SJeCn5Z<SJeCm5[<TJdCm5Z<TJfCl5Z<UJeCl5Z<TJfCl5Z<UJeCk5[<UJeCl5Z<UJeCk5Z<WJeCj5Z<VJfCj5Z<WJeCj5Z<VJfCj5Z<WJeCj5Y<XJfCh5Z<XJfCi5Y<XJfCh5Z<XJeCi5[<o0O0O2O001O0O2O0O2N100O2O0O2O0O2O0O2N101N100O2O0O2N101N101N10N2M2M4L4M3L4L4M2M4L4M3L4L3N3L4L4M3L3M4M3L4L4N1O2O1O1N2O1O2M2O2N2M2O2M3N2N1N3N2N1N3N2N1N3N1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1fL`@R2b?mMfAk0[>SOkAh0V>WOlAf0W>WOlAf0V>YOkAe0W>YOlAd0V>ZOmAc0U>\\OlAb0V>\\OmAa0U>^OmA>V>@lA>V>AlA<V>BmA;U>DmA9U>EmA9o`0M3M2N_\\`7"}, "label": "a white sheep is propped on the ground next to a sheep lying down"}]}
{"id": 156827, "image": "COCO_train2014_000000156827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one of three zebras is grazing on the left\"."}], "task": "refering", "answer_template": "The \"one of three zebras is grazing on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 50, 51, 52, 53, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282], "bbox": [0.11068750000000001, 0.11685446009389672, 0.37843750000000004, 0.8404460093896713], "iscrowd": 0, "area": 28883.38014999999, "rle": {"size": [426, 640], "counts": "iim01X=1O2N1O1O2N1O1O2M2O1O2N1O1O2N1O2N1O100O01O010O0010O01O0010O0001O010O00TO@PE?d::oDGe:d1I7M5K5L4K5K6K4K5M3O2M2N2N2O2M2[NPLmIR4o5jLWIX3e6RMTIP3h6ZMoHh2m6m1L5K4M3M3N3M2N3M3N3L4L3M4L3M4L3M4M2M3M2N3M2O2M2O2M2O2M2O2M2O1N3N1N3M2N1O2N1N2O2N1O2O001O01O01O001O0010O0001OM3L5K4K5L4L4L5lNS12N1O1O2N10aMmJgLQ5Y3VLaKi3`4XL`Kf3c4YL\\Kf3h4XLXKg3k4WLUKg3n4XLQKg3S5WLmJg3V5YLhJf3\\5XLdJh3]5WLcJi3_5UL`Jm3a5QL_Jo3c5oK\\JR4f5lKXJW4i5gKUJ[4l5cKRJ`4n5_KQJd4o5[KoIg4Q6XKmIk4T6SKkIP5T6oJjIT5W6jJhIX5X6hJeI\\5[6bJdI`5\\6i02M4M3M5J:G[KVJg1i5RNjJa1X5WN\\K[1f4]NmKU1U4dN^Lo0l3aNfLR1e3^NlLW1^3YNTM[1b8kN2O1O0O2O1N2OaWU5"}, "label": "one of three zebras is grazing on the left"}]}
{"id": 156827, "image": "COCO_train2014_000000156827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra with head turned sideways\"."}], "task": "refering", "answer_template": "The \"zebra with head turned sideways\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 50, 51, 52, 53, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282], "bbox": [0.11068750000000001, 0.11685446009389672, 0.37843750000000004, 0.8404460093896713], "iscrowd": 0, "area": 28883.38014999999, "rle": {"size": [426, 640], "counts": "iim01X=1O2N1O1O2N1O1O2M2O1O2N1O1O2N1O2N1O100O01O010O0010O01O0010O0001O010O00TO@PE?d::oDGe:d1I7M5K5L4K5K6K4K5M3O2M2N2N2O2M2[NPLmIR4o5jLWIX3e6RMTIP3h6ZMoHh2m6m1L5K4M3M3N3M2N3M3N3L4L3M4L3M4L3M4M2M3M2N3M2O2M2O2M2O2M2O2M2O1N3N1N3M2N1O2N1N2O2N1O2O001O01O01O001O0010O0001OM3L5K4K5L4L4L5lNS12N1O1O2N10aMmJgLQ5Y3VLaKi3`4XL`Kf3c4YL\\Kf3h4XLXKg3k4WLUKg3n4XLQKg3S5WLmJg3V5YLhJf3\\5XLdJh3]5WLcJi3_5UL`Jm3a5QL_Jo3c5oK\\JR4f5lKXJW4i5gKUJ[4l5cKRJ`4n5_KQJd4o5[KoIg4Q6XKmIk4T6SKkIP5T6oJjIT5W6jJhIX5X6hJeI\\5[6bJdI`5\\6i02M4M3M5J:G[KVJg1i5RNjJa1X5WN\\K[1f4]NmKU1U4dN^Lo0l3aNfLR1e3^NlLW1^3YNTM[1b8kN2O1O0O2O1N2OaWU5"}, "label": "zebra with head turned sideways"}]}
{"id": 156827, "image": "COCO_train2014_000000156827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a larger zebra facing the right\"."}], "task": "refering", "answer_template": "The \"a larger zebra facing the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 32, 33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 78, 79, 82, 83, 84, 85, 86, 87, 101, 102, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 147, 152, 153, 154, 155, 156, 157, 170, 175, 176, 177, 178, 179, 180, 181, 193, 198, 199, 200, 201, 202, 203, 216, 217, 221, 223, 224, 225, 226, 244, 246, 247, 248, 249, 267, 268, 270, 271, 272], "bbox": [0.394890625, 0.03370892018779342, 0.9079375000000001, 0.7775352112676057], "iscrowd": 0, "area": 43104.89225000001, "rle": {"size": [426, 640], "counts": "VZY34T=4L5K5K5L4K4^FROb6S1ZI]OU6h0gIHh5=SJ3]52_J>P5GlJi0c4\\OYKo0[4VOaKl0\\4YO`Ki0]4\\O_Ke0_4@bK=[4GkK1S44SLDk3>\\LZOc3g0fLQOX3o0QMjNm2W1\\M`Nc2`1fMYNX2h1QNoMn1Q2[NhMd1X2eN`MY1`2POXMo0h2i4O1N2O1N2O1N2O2M2N2O101O100O100O10000O100O100O10000O100O10000O100O100O10000O101N100O2O000O2O0O101O0O100O2O0O101O0O101N100010O1O010O1O010O1O010O1O010O00100O00100O00100O010O1O010O10O01O10O01O10O0100O00100O010O1O010O1O010O101Nn1RN100O103L8H8QJ^Kg2i4hLdKT3d4[LiKgNUOU4k5RL]K]OPOa4Y8O001O0O2O001O001O00O100\\N`L^I`3^6fL`IZ3\\6lLbIS3[6SMcIm2Y6YMeIg2R5fKcKh1YOb2m4[LXKY1I\\2g4WOWKi0b4@\\K`0]4Z4M31O1O2M2O1O2N100O2O0O3N2N3L3N2M3N3M2M3N2N2M4M2M3N2N3L3N1N3N1O1N2O2M2O1O4K5L3L5L4L3L4L1N3N1O1N3N1O2M3N2N2M3N2N2M3N1O0O100001O1O1O2N100O1O2N1O1O1O2N1O100O1O2N100O1O2N100O1O2N100O1O2O0O1O2N3N1N3M2N3M2M4M2N3M2M4M2N3M2`NQI`LS7[3ZIZLi6`3fITL\\6h3RJlKQ6o3h1L5K4L5J5L5K4L5K4LY1QNm0I2N2N1O2N2001O0000001O000O101O000000001O0001O100O1O00100N2M3M2N3M3MiZh0"}, "label": "a larger zebra facing the right"}]}
{"id": 156827, "image": "COCO_train2014_000000156827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra facing sideways behind another zebra\"."}], "task": "refering", "answer_template": "The \"zebra facing sideways behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 32, 33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 78, 79, 82, 83, 84, 85, 86, 87, 101, 102, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 147, 152, 153, 154, 155, 156, 157, 170, 175, 176, 177, 178, 179, 180, 181, 193, 198, 199, 200, 201, 202, 203, 216, 217, 221, 223, 224, 225, 226, 244, 246, 247, 248, 249, 267, 268, 270, 271, 272], "bbox": [0.394890625, 0.03370892018779342, 0.9079375000000001, 0.7775352112676057], "iscrowd": 0, "area": 43104.89225000001, "rle": {"size": [426, 640], "counts": "VZY34T=4L5K5K5L4K4^FROb6S1ZI]OU6h0gIHh5=SJ3]52_J>P5GlJi0c4\\OYKo0[4VOaKl0\\4YO`Ki0]4\\O_Ke0_4@bK=[4GkK1S44SLDk3>\\LZOc3g0fLQOX3o0QMjNm2W1\\M`Nc2`1fMYNX2h1QNoMn1Q2[NhMd1X2eN`MY1`2POXMo0h2i4O1N2O1N2O1N2O2M2N2O101O100O100O10000O100O100O10000O100O10000O100O100O10000O101N100O2O000O2O0O101O0O100O2O0O101O0O101N100010O1O010O1O010O1O010O1O010O00100O00100O00100O010O1O010O10O01O10O01O10O0100O00100O010O1O010O1O010O101Nn1RN100O103L8H8QJ^Kg2i4hLdKT3d4[LiKgNUOU4k5RL]K]OPOa4Y8O001O0O2O001O001O00O100\\N`L^I`3^6fL`IZ3\\6lLbIS3[6SMcIm2Y6YMeIg2R5fKcKh1YOb2m4[LXKY1I\\2g4WOWKi0b4@\\K`0]4Z4M31O1O2M2O1O2N100O2O0O3N2N3L3N2M3N3M2M3N2N2M4M2M3N2N3L3N1N3N1O1N2O2M2O1O4K5L3L5L4L3L4L1N3N1O1N3N1O2M3N2N2M3N2N2M3N1O0O100001O1O1O2N100O1O2N1O1O1O2N1O100O1O2N100O1O2N100O1O2N100O1O2O0O1O2N3N1N3M2N3M2M4M2N3M2M4M2N3M2`NQI`LS7[3ZIZLi6`3fITL\\6h3RJlKQ6o3h1L5K4L5J5L5K4L5K4LY1QNm0I2N2N1O2N2001O0000001O000O101O000000001O0001O100O1O00100N2M3M2N3M3MiZh0"}, "label": "zebra facing sideways behind another zebra"}]}
{"id": 156827, "image": "COCO_train2014_000000156827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra eating grass with his tail raised\"."}], "task": "refering", "answer_template": "The \"a zebra eating grass with his tail raised\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 312, 313], "bbox": [0.371609375, 0.16368544600938967, 0.616375, 0.9394600938967137], "iscrowd": 0, "area": 31352.509900000005, "rle": {"size": [426, 640], "counts": "fTS31X=2N1O2N2N1O2N2O1N1\\GCZ4>fKCX4>hKDV4<iKFV49kKHS49mKHR48mKKQ45oKLP43PLOn32RLOm31SL1k3OTL3k3MUL4j3KWL6g3KXL8f3HZL8f3HYL9g3FZL:f3FZL:e3GZL:f3FTI@W2j0e4FQIMP2=o4EnH<h1O[5EjHi0a1Bd5FhHU1Z1UOn5FeHa1S1iNY6FaHk1o0_N`6E_HU2j0WNf6E]H[2i0oMk6FYHb2h0hMo6FVHi2f0bMT7T3hHkLY7Y3cHmLV7X3eHUMn6P3nH[Me6k2VIbM\\6S5B=C>Bl0TO2N2N2N2N2N2N3M200O100O101N100O100O100O0010O0100O010O10O0100O01N1O2O0K6E:G9G:E:G:F9G:E:H91N3N1N2O1N3N1N2WKmHY3T7[LWId3l6oK_IQ4b6cKiI^4b7O01O10O01O1O10O01O1O010O1O010000O10000O0100000O100bKPHa3l8001O1RM^LRLb3g3hLULZ3d3PMXLQ3`3YM]Lg2\\3dM`L]2X3mMdLU2T3VNhLj1Q3`NlLa1l2jNPMW1i2ROTMn0e2]OVMe0b2FZM;^2O_M2Y29cMGV2c0gM^OQ2m0jMUOn1U1oMkNj1`1RNaNf1i1WN`NY1k1bN^No0l1nN[Nd0P2XOXN9R2DcTV3"}, "label": "a zebra eating grass with his tail raised"}]}
{"id": 156827, "image": "COCO_train2014_000000156827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra bending down eating grass in between two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra bending down eating grass in between two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 312, 313], "bbox": [0.371609375, 0.16368544600938967, 0.616375, 0.9394600938967137], "iscrowd": 0, "area": 31352.509900000005, "rle": {"size": [426, 640], "counts": "fTS31X=2N1O2N2N1O2N2O1N1\\GCZ4>fKCX4>hKDV4<iKFV49kKHS49mKHR48mKKQ45oKLP43PLOn32RLOm31SL1k3OTL3k3MUL4j3KWL6g3KXL8f3HZL8f3HYL9g3FZL:f3FZL:e3GZL:f3FTI@W2j0e4FQIMP2=o4EnH<h1O[5EjHi0a1Bd5FhHU1Z1UOn5FeHa1S1iNY6FaHk1o0_N`6E_HU2j0WNf6E]H[2i0oMk6FYHb2h0hMo6FVHi2f0bMT7T3hHkLY7Y3cHmLV7X3eHUMn6P3nH[Me6k2VIbM\\6S5B=C>Bl0TO2N2N2N2N2N2N3M200O100O101N100O100O100O0010O0100O010O10O0100O01N1O2O0K6E:G9G:E:G:F9G:E:H91N3N1N2O1N3N1N2WKmHY3T7[LWId3l6oK_IQ4b6cKiI^4b7O01O10O01O1O10O01O1O010O1O010000O10000O0100000O100bKPHa3l8001O1RM^LRLb3g3hLULZ3d3PMXLQ3`3YM]Lg2\\3dM`L]2X3mMdLU2T3VNhLj1Q3`NlLa1l2jNPMW1i2ROTMn0e2]OVMe0b2FZM;^2O_M2Y29cMGV2c0gM^OQ2m0jMUOn1U1oMkNj1`1RNaNf1i1WN`NY1k1bN^No0l1nN[Nd0P2XOXN9R2DcTV3"}, "label": "a zebra bending down eating grass in between two other zebras"}]}
{"id": 17566, "image": "COCO_train2014_000000017566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting in the front with check shirt\"."}], "task": "refering", "answer_template": "The \"man sitting in the front with check shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 227, 228, 242, 243, 257, 258, 259, 273, 274, 275, 288, 289, 290, 303, 304, 305, 319, 320, 334, 335, 336, 337], "bbox": [0.14672131147540984, 0.6262187499999999, 0.5226229508196721, 1.0], "iscrowd": 0, "area": 11981.137599999996, "rle": {"size": [640, 427], "counts": "VjW14jc03M4L4l\\OF\\b0>^]OG_b0<]]OI]b0P1K6J5M4L3M4L3M4L3L4M3M3M3M3M3M3M2N3O100O2N2O0O2N2O1N2N2O1N3MiMP@a0n?[OW@f0f?WO^@k0`?QOe@P1]?iNg@X1`?\\Nc@f1]?VNf@k1[?dMSA_2V`02N2M3N2N2N2N2M3N2N1O1N2O0O2O1N2O1O1N5L4K4Mb0]O10O10O010O10O1O010O001O1O1O10O01O1O1O001O100O4L4L4L4L4L4L5K8I7H5K4L4Lc1]N0000000000000001O00000001O00000000000000000000000000000000000000000000000000O100001N1000001O00001O000O101O1O2N2M3N5I]ln3"}, "label": "man sitting in the front with check shirt"}]}
{"id": 17566, "image": "COCO_train2014_000000017566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the audience member wearing a white and red checked shirt\"."}], "task": "refering", "answer_template": "The \"the audience member wearing a white and red checked shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 227, 228, 242, 243, 257, 258, 259, 273, 274, 275, 288, 289, 290, 303, 304, 305, 319, 320, 334, 335, 336, 337], "bbox": [0.14672131147540984, 0.6262187499999999, 0.5226229508196721, 1.0], "iscrowd": 0, "area": 11981.137599999996, "rle": {"size": [640, 427], "counts": "VjW14jc03M4L4l\\OF\\b0>^]OG_b0<]]OI]b0P1K6J5M4L3M4L3M4L3L4M3M3M3M3M3M3M2N3O100O2N2O0O2N2O1N2N2O1N3MiMP@a0n?[OW@f0f?WO^@k0`?QOe@P1]?iNg@X1`?\\Nc@f1]?VNf@k1[?dMSA_2V`02N2M3N2N2N2N2M3N2N1O1N2O0O2O1N2O1O1N5L4K4Mb0]O10O10O010O10O1O010O001O1O1O10O01O1O1O001O100O4L4L4L4L4L4L5K8I7H5K4L4Lc1]N0000000000000001O00000001O00000000000000000000000000000000000000000000000000O100001N1000001O00001O000O101O1O2N2M3N5I]ln3"}, "label": "the audience member wearing a white and red checked shirt"}]}
{"id": 17566, "image": "COCO_train2014_000000017566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man holding a skateboard\"."}], "task": "refering", "answer_template": "The \"the man holding a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 82, 83, 96, 97, 98, 110, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 144, 145, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 279, 292, 293, 294, 307, 308, 309, 322, 323, 324], "bbox": [0.29370023419203745, 0.1891875, 0.7595316159250586, 0.945953125], "iscrowd": 0, "area": 46854.69424999999, "rle": {"size": [640, 427], "counts": "\\\\^2c0Yc04L4M3L4L4M4K4L4N2N2O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1O1O1N2O1O1O1N2O100O1O100O1O1O100O101O1O1O0O2O4L8Ha0_Of0YOV1kN8H8H8H8G9H7H7J5J7J6I7J6I7J6IT1mN3L5L4K5L4K5JZGmFg6c7oGjH^1Fd6S7_HlHP14b6^1[G^3d3UKU5R1]Gi3^3VKY5e0`GT4X3XK[5:cG^4R3XK`5NeGj4k2XKi6i4WIWKi6h4XIXKh6h4WIYKi6g4WIYKi6f4XIZKh6f4XIZKmLFn8Q5UJYKkLIo8m4WJ[KhLJP9k4XJ\\KeLLR9h4YJ\\KdLNR9f4ZJ]KbLNT9d4[J^K_L1T9b4]J^K]L2U9`4^J^K[L5V9\\4`JeKSL0]9[4`J^L`5b3`J^L`5b3`J^L`5a3aJ_L^5b3bJ]L_5c3aJ]L_5d3`J\\L_5e3aJ[L]5h3aJYL^5h3bJXL\\5k3cJUL\\5l3dJTL[5n3cJSL[5o3eJQLZ5Q4eJoKY5S4fJmKZ5U4eJkK[5U4eJkK\\5U4bJkKa5S4_JmKb5S4]JmKd5R4\\JnKe5R4YJnKi5Q4WJoKj5R4TJnKm5R4QJnKQ6S4mImKT6T4jIlKW6T4gImKZ6T4dIkK_6T4`IlKa6U4]IjKe6`4oH`KR7c5kG\\JV8e5iGZJX8g5fGYJ[8i5cGWJ]8j5aGVJ`8k5_GTJb8m5]GRJd8o5ZGQJg8P6XGoIi8R6VG[IoNBn9S7RGXIb9i6]FVId9k6[FUIe9k6ZFUIg9l6XFSIh9o6WFQIi9o6VFQIk9P7TFPIl9Q7RFoHo9Q7QFoHo9R7PFmHK5T7KWIS7InH09`60dIj6KmH4>k58RJ_6MlH8d0T5`0cJQ61jH;l0\\4f0VKf52hH=T1Y3Y1ULl44hH`0\\1Z1d2oMZ36fHc0i17k2mNg28fHf0U2TOQ3LV29dHi0g`0VOY_Om0f`0ROZ_OP1d`0PO\\_OT1a`0jN`_O[1[`0eNe_O`1X`0]Ni_Og1W`0UNh_OP2W`0lMj_OW2V`0fMj_O]2k`03M3L6K5Kn0SOO2N1O1O1O1O2N1O1O1O2N2N4K5L3Mn0RO7H8I7I8H3M3L5L8H8FjPP2"}, "label": "the man holding a skateboard"}]}
{"id": 17566, "image": "COCO_train2014_000000017566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a shirt and jeans holding a skate board in his hand\"."}], "task": "refering", "answer_template": "The \"a person wearing a shirt and jeans holding a skate board in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 82, 83, 96, 97, 98, 110, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 144, 145, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 279, 292, 293, 294, 307, 308, 309, 322, 323, 324], "bbox": [0.29370023419203745, 0.1891875, 0.7595316159250586, 0.945953125], "iscrowd": 0, "area": 46854.69424999999, "rle": {"size": [640, 427], "counts": "\\\\^2c0Yc04L4M3L4L4M4K4L4N2N2O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1O1O1N2O1O1O1N2O100O1O100O1O1O100O101O1O1O0O2O4L8Ha0_Of0YOV1kN8H8H8H8G9H7H7J5J7J6I7J6I7J6IT1mN3L5L4K5L4K5JZGmFg6c7oGjH^1Fd6S7_HlHP14b6^1[G^3d3UKU5R1]Gi3^3VKY5e0`GT4X3XK[5:cG^4R3XK`5NeGj4k2XKi6i4WIWKi6h4XIXKh6h4WIYKi6g4WIYKi6f4XIZKh6f4XIZKmLFn8Q5UJYKkLIo8m4WJ[KhLJP9k4XJ\\KeLLR9h4YJ\\KdLNR9f4ZJ]KbLNT9d4[J^K_L1T9b4]J^K]L2U9`4^J^K[L5V9\\4`JeKSL0]9[4`J^L`5b3`J^L`5b3`J^L`5a3aJ_L^5b3bJ]L_5c3aJ]L_5d3`J\\L_5e3aJ[L]5h3aJYL^5h3bJXL\\5k3cJUL\\5l3dJTL[5n3cJSL[5o3eJQLZ5Q4eJoKY5S4fJmKZ5U4eJkK[5U4eJkK\\5U4bJkKa5S4_JmKb5S4]JmKd5R4\\JnKe5R4YJnKi5Q4WJoKj5R4TJnKm5R4QJnKQ6S4mImKT6T4jIlKW6T4gImKZ6T4dIkK_6T4`IlKa6U4]IjKe6`4oH`KR7c5kG\\JV8e5iGZJX8g5fGYJ[8i5cGWJ]8j5aGVJ`8k5_GTJb8m5]GRJd8o5ZGQJg8P6XGoIi8R6VG[IoNBn9S7RGXIb9i6]FVId9k6[FUIe9k6ZFUIg9l6XFSIh9o6WFQIi9o6VFQIk9P7TFPIl9Q7RFoHo9Q7QFoHo9R7PFmHK5T7KWIS7InH09`60dIj6KmH4>k58RJ_6MlH8d0T5`0cJQ61jH;l0\\4f0VKf52hH=T1Y3Y1ULl44hH`0\\1Z1d2oMZ36fHc0i17k2mNg28fHf0U2TOQ3LV29dHi0g`0VOY_Om0f`0ROZ_OP1d`0PO\\_OT1a`0jN`_O[1[`0eNe_O`1X`0]Ni_Og1W`0UNh_OP2W`0lMj_OW2V`0fMj_O]2k`03M3L6K5Kn0SOO2N1O1O1O1O2N1O1O1O2N2N4K5L3Mn0RO7H8I7I8H3M3L5L8H8FjPP2"}, "label": "a person wearing a shirt and jeans holding a skate board in his hand"}]}
{"id": 230559, "image": "COCO_train2014_000000230559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.132078125, 0.4187353629976581, 0.40604687500000003, 0.988688524590164], "iscrowd": 0, "area": 28176.117150000005, "rle": {"size": [427, 640], "counts": "miS1:n<;F;E:E<E5K3L3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2N2N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N200O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1QOhJjIY5T6jJkIV5S6lJlIU5S6mJlIS5R6PKlIQ5S6QKkIP5S6SKkIn4T6TKjIm4T6VKjIk4U6VKjIk4T6XKjIi4U6YKiIh4U6[KiIf4V6\\KhIe4V6^KhIc4W6^KhIc4V6`KhIa4W6Z1N2N2I7J6000000000000000000001O0000000000000000000000000000001O00000WKWJP2i5nM[JP2e5nM`Jo1`5oMeJn1[5PNiJm1X5QNmJl1S5RNRKk1n4RNWKl1i4RN\\Kk1d4SNaKj1_4TNeKj1[4TNjKi1V4UNoKh1Q4VNSLh1n3UNWLh1i3VN\\Lg1d3WN`Lg1`3WNeLf1[3XNjLe1V3YNiNJW14ZO]Of0a0BYO>e0KTO5j03POMn0;lNcXn4"}, "label": "a man in a red shirt"}]}
{"id": 230559, "image": "COCO_train2014_000000230559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.132078125, 0.4187353629976581, 0.40604687500000003, 0.988688524590164], "iscrowd": 0, "area": 28176.117150000005, "rle": {"size": [427, 640], "counts": "miS1:n<;F;E:E<E5K3L3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2N2N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N200O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1QOhJjIY5T6jJkIV5S6lJlIU5S6mJlIS5R6PKlIQ5S6QKkIP5S6SKkIn4T6TKjIm4T6VKjIk4U6VKjIk4T6XKjIi4U6YKiIh4U6[KiIf4V6\\KhIe4V6^KhIc4W6^KhIc4V6`KhIa4W6Z1N2N2I7J6000000000000000000001O0000000000000000000000000000001O00000WKWJP2i5nM[JP2e5nM`Jo1`5oMeJn1[5PNiJm1X5QNmJl1S5RNRKk1n4RNWKl1i4RN\\Kk1d4SNaKj1_4TNeKj1[4TNjKi1V4UNoKh1Q4VNSLh1n3UNWLh1i3VN\\Lg1d3WN`Lg1`3WNeLf1[3XNjLe1V3YNiNJW14ZO]Of0a0BYO>e0KTO5j03POMn0;lNcXn4"}, "label": "a man in a red shirt"}]}
{"id": 230559, "image": "COCO_train2014_000000230559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black man in a green shirt\"."}], "task": "refering", "answer_template": "The \"a black man in a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 199, 200, 222, 223, 224, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316, 337, 338, 339], "bbox": [0.6581875, 0.4933255269320843, 0.7721875, 0.9865105386416863], "iscrowd": 0, "area": 11249.881399999998, "rle": {"size": [427, 640], "counts": "Xm_58Y<R1E9G4cLZNoJj1Q5XNiJm1U5UNgJo1X5RNcJS2[5oMXJ^2g5cMQJd2n5_MnId2Q6aMiIa2U6dMeI_2cNYMR7=UJP3j5UMQJl2o5UMnIn2P6SMnIo2Q6RMjIS3U6nLdIY3\\6gL\\Ia3c6`LWIf3h6[LWIf3h6\\LVIe3i6\\LVIe3j6[LUIf3j6^1O1O1O100O100001O001O001O1O001O001O0fJQIX4P7gKQIX4o6gKSIX4n6gKSIY4l6XKPIO5h4l6WKQIO5i4j6VKTIO4j4i6TKVI02l4Z7SKgHl4Z7RKhHm4Y7QKiHn4X7QKiHn4k7O1O1O2N2N2N2N2N2N2N2N2N2cLRGX2Q9cMmG\\1[8_NQHR1W8hNUHj0S8QOYH`0o7ZO^H7hjl1"}, "label": "a black man in a green shirt"}]}
{"id": 230559, "image": "COCO_train2014_000000230559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a local with green shirt\"."}], "task": "refering", "answer_template": "The \"a local with green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 199, 200, 222, 223, 224, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316, 337, 338, 339], "bbox": [0.6581875, 0.4933255269320843, 0.7721875, 0.9865105386416863], "iscrowd": 0, "area": 11249.881399999998, "rle": {"size": [427, 640], "counts": "Xm_58Y<R1E9G4cLZNoJj1Q5XNiJm1U5UNgJo1X5RNcJS2[5oMXJ^2g5cMQJd2n5_MnId2Q6aMiIa2U6dMeI_2cNYMR7=UJP3j5UMQJl2o5UMnIn2P6SMnIo2Q6RMjIS3U6nLdIY3\\6gL\\Ia3c6`LWIf3h6[LWIf3h6\\LVIe3i6\\LVIe3j6[LUIf3j6^1O1O1O100O100001O001O001O1O001O001O0fJQIX4P7gKQIX4o6gKSIX4n6gKSIY4l6XKPIO5h4l6WKQIO5i4j6VKTIO4j4i6TKVI02l4Z7SKgHl4Z7RKhHm4Y7QKiHn4X7QKiHn4k7O1O1O2N2N2N2N2N2N2N2N2N2cLRGX2Q9cMmG\\1[8_NQHR1W8hNUHj0S8QOYH`0o7ZO^H7hjl1"}, "label": "a local with green shirt"}]}
{"id": 58531, "image": "COCO_train2014_000000058531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing on a surf board , riding a wave\"."}], "task": "refering", "answer_template": "The \"a man standing on a surf board , riding a wave\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 102, 103, 104, 105, 106, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195], "bbox": [0.46375, 0.22228329809725159, 0.65834375, 0.5019450317124736], "iscrowd": 0, "area": 5222.750150000001, "rle": {"size": [473, 640], "counts": "jYY41h>1O1O3[AK`>700O101O00001N10000O01OaD3S8MlG4U8KjG6W8IiG7W8IhG8Y8FgG;Y8EfG;\\8DcG=^8BaG?_8B_G?b8A\\G`0e8_OYGc0g8^OSGg0n8YOnFj0R9VOoFi0R9WOmFi0T9WOkFh0V9XOjFh0W9XOiFg0W9ZOhFf0Y9ZOfFf0[9YOeFg0[9ZOeFe0\\9[OcFe0]9\\ObFd0_9\\OaFc0`9]O_Fb0b9_O^F`0c9@]F>d9C[F<g9DYF:i9FVF:j9FWF9k9FUF:k9GSF:m9GRF9o9GPF9P:HnE9S:GlE9T:HkE8V:HhE9X:IfE7Z:JeE6\\:JbE7^:JaE6`:J_E6a:K]E6c:K\\E5e:KZE5f:LXE5i:KVE5j:8iDIW;<bDE_;`0[D@e;f0UDZOk;l0nCUOQ<j10000O100001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1O1O1M2N3N2M3M3M32002N2N3M2N2NdBmNR=Z1=D5K3L5L3M3M3L2O1O1N2O1MTnT3"}, "label": "a man standing on a surf board , riding a wave"}]}
{"id": 58531, "image": "COCO_train2014_000000058531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfer standing on a surf board\"."}], "task": "refering", "answer_template": "The \"a surfer standing on a surf board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 102, 103, 104, 105, 106, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195], "bbox": [0.46375, 0.22228329809725159, 0.65834375, 0.5019450317124736], "iscrowd": 0, "area": 5222.750150000001, "rle": {"size": [473, 640], "counts": "jYY41h>1O1O3[AK`>700O101O00001N10000O01OaD3S8MlG4U8KjG6W8IiG7W8IhG8Y8FgG;Y8EfG;\\8DcG=^8BaG?_8B_G?b8A\\G`0e8_OYGc0g8^OSGg0n8YOnFj0R9VOoFi0R9WOmFi0T9WOkFh0V9XOjFh0W9XOiFg0W9ZOhFf0Y9ZOfFf0[9YOeFg0[9ZOeFe0\\9[OcFe0]9\\ObFd0_9\\OaFc0`9]O_Fb0b9_O^F`0c9@]F>d9C[F<g9DYF:i9FVF:j9FWF9k9FUF:k9GSF:m9GRF9o9GPF9P:HnE9S:GlE9T:HkE8V:HhE9X:IfE7Z:JeE6\\:JbE7^:JaE6`:J_E6a:K]E6c:K\\E5e:KZE5f:LXE5i:KVE5j:8iDIW;<bDE_;`0[D@e;f0UDZOk;l0nCUOQ<j10000O100001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1O1O1M2N3N2M3M3M32002N2N3M2N2NdBmNR=Z1=D5K3L5L3M3M3L2O1O1N2O1MTnT3"}, "label": "a surfer standing on a surf board"}]}
{"id": 58531, "image": "COCO_train2014_000000058531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"surfer laying down\"."}], "task": "refering", "answer_template": "The \"surfer laying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 234, 235, 236, 258, 259, 260, 282, 283, 285, 286, 309], "bbox": [0.20604687500000002, 0.533784355179704, 0.45621875, 0.7750528541226216], "iscrowd": 0, "area": 3231.79885, "rle": {"size": [473, 640], "counts": "oWm16a>8I3M3AA`B`0\\=F`B=]=F`B<^=a0M3M4M3M3N2M7I101O001O1O0O11O010J5M3N4gNSCc0n<ZOTCf0o<SOTCo0[=1N2O1N10000O1O10O01O10SOgB9X=GjB7V=IlB6T=HoB6Q=JQC5o<JVC1k<MWC3i<LWC4k<IVC8j<GWC8k<EVC<l<@UC`0b=1O010O001O01O01O01O00001O001O1O1O1O2N2M3N2N1O2NiZd02Te[O5L1N100O1O1O2N1O1O101N10000010O1O10O01000O10O1O001O1O001O1O1O1O1O1O2N2N2N101N^\\P5"}, "label": "surfer laying down"}]}
{"id": 58531, "image": "COCO_train2014_000000058531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man lying down on a surfboard\"."}], "task": "refering", "answer_template": "The \"a man lying down on a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 234, 235, 236, 258, 259, 260, 282, 283, 285, 286, 309], "bbox": [0.20604687500000002, 0.533784355179704, 0.45621875, 0.7750528541226216], "iscrowd": 0, "area": 3231.79885, "rle": {"size": [473, 640], "counts": "oWm16a>8I3M3AA`B`0\\=F`B=]=F`B<^=a0M3M4M3M3N2M7I101O001O1O0O11O010J5M3N4gNSCc0n<ZOTCf0o<SOTCo0[=1N2O1N10000O1O10O01O10SOgB9X=GjB7V=IlB6T=HoB6Q=JQC5o<JVC1k<MWC3i<LWC4k<IVC8j<GWC8k<EVC<l<@UC`0b=1O010O001O01O01O01O00001O001O1O1O1O2N2M3N2N1O2NiZd02Te[O5L1N100O1O1O2N1O1O101N10000010O1O10O01000O10O1O001O1O001O1O1O1O1O1O2N2N2N101N^\\P5"}, "label": "a man lying down on a surfboard"}]}
{"id": 124069, "image": "COCO_train2014_000000124069.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car parked at a transportation terminal\"."}], "task": "refering", "answer_template": "The \"a black car parked at a transportation terminal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 130, 131, 132, 149, 150, 151, 153, 154, 155, 172, 173, 178, 179, 180, 202, 203, 204, 226, 227, 249, 250], "bbox": [0.48353124999999997, 0.30352083333333335, 0.8929375, 0.6517708333333334], "iscrowd": 0, "area": 8379.427100000003, "rle": {"size": [480, 640], "counts": "ZPa42m>b0]O5L1N3N2M3M3N2O1N2O1O0O2O1O1N2O1O1N2O1O1N100O01N2O002M2O1O2M2O101O1O2M3N2N1N3N2N1O1N2O1O1N101O1O1N2O001N2O1O1O0O2O^fj0JeYUO5L4K5K4L5L4N2N2N1O2N1O1O00001O001O00001O00001O000010O01O00001O00001O01O010O01O01O01O01O01O010O0002O0O2N101N1O2O0O2O0O2N101N1O2O1N1O2O0O2O0O2N101N1O2O0O2N10000000010O000001O0000001O001O2N2N2N2N3L3N2M3N2M3M4M2M3N2M3M3N2M4M2M3M3N2M3N3L3M3N2M01N10001O0O2O00001N101O000O2N1N3M2M3N3M2N3Om0RO4L4J6EhRP1"}, "label": "a black car parked at a transportation terminal"}]}
{"id": 124069, "image": "COCO_train2014_000000124069.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car behind two men\"."}], "task": "refering", "answer_template": "The \"a car behind two men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 130, 131, 132, 149, 150, 151, 153, 154, 155, 172, 173, 178, 179, 180, 202, 203, 204, 226, 227, 249, 250], "bbox": [0.48353124999999997, 0.30352083333333335, 0.8929375, 0.6517708333333334], "iscrowd": 0, "area": 8379.427100000003, "rle": {"size": [480, 640], "counts": "ZPa42m>b0]O5L1N3N2M3M3N2O1N2O1O0O2O1O1N2O1O1N2O1O1N100O01N2O002M2O1O2M2O101O1O2M3N2N1N3N2N1O1N2O1O1N101O1O1N2O001N2O1O1O0O2O^fj0JeYUO5L4K5K4L5L4N2N2N1O2N1O1O00001O001O00001O00001O000010O01O00001O00001O01O010O01O01O01O01O01O010O0002O0O2N101N1O2O0O2O0O2N101N1O2O1N1O2O0O2O0O2N101N1O2O0O2N10000000010O000001O0000001O001O2N2N2N2N3L3N2M3N2M3M4M2M3N2M3M3N2M4M2M3M3N2M3N3L3M3N2M01N10001O0O2O00001N101O000O2N1N3M2M3N3M2N3Om0RO4L4J6EhRP1"}, "label": "a car behind two men"}]}
{"id": 500904, "image": "COCO_train2014_000000500904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver spoon with a little bit of oatmeal on it\"."}], "task": "refering", "answer_template": "The \"a silver spoon with a little bit of oatmeal on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [386, 390, 407, 408, 409, 410, 411, 412, 413, 431, 432, 433, 434, 435, 455, 456, 457, 478, 479, 480], "bbox": [0.699796875, 0.725375, 1.0, 0.90740625], "iscrowd": 0, "area": 9304.991949999998, "rle": {"size": [640, 640], "counts": "\\Si81oc00O100O100O100O10000O100O100O100O100O2O0O10001N10001N10001N10001N10001O0O2O1O0O2O1O1N2O1O1N2O6J6Ic0^O001O0O101O0O2O000O2O001O001O01On^OTNP`0Q3M00O1000000O2O00000O1000000O2O00000O1000000O2O00000O1000000O2O001O0O101N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2M2O2N1O1O2N1O2N1O2N1O1O1O1O001O1N1N3M2N3M3M2N3M2N3L3O200O0010O0100O0010O0100O010O0010O0100O0010O0100O010O0010O0100O0010O0100O010O00100O010O006K7HUA"}, "label": "a silver spoon with a little bit of oatmeal on it"}]}
{"id": 500904, "image": "COCO_train2014_000000500904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spoon with food on it\"."}], "task": "refering", "answer_template": "The \"a spoon with food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [386, 390, 407, 408, 409, 410, 411, 412, 413, 431, 432, 433, 434, 435, 455, 456, 457, 478, 479, 480], "bbox": [0.699796875, 0.725375, 1.0, 0.90740625], "iscrowd": 0, "area": 9304.991949999998, "rle": {"size": [640, 640], "counts": "\\Si81oc00O100O100O100O10000O100O100O100O100O2O0O10001N10001N10001N10001N10001O0O2O1O0O2O1O1N2O1O1N2O6J6Ic0^O001O0O101O0O2O000O2O001O001O01On^OTNP`0Q3M00O1000000O2O00000O1000000O2O00000O1000000O2O00000O1000000O2O001O0O101N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2M2O2N1O1O2N1O2N1O2N1O1O1O1O001O1N1N3M2N3M3M2N3M2N3L3O200O0010O0100O0010O0100O010O0010O0100O0010O0100O010O0010O0100O0010O0100O010O00100O010O006K7HUA"}, "label": "a spoon with food on it"}]}
{"id": 459951, "image": "COCO_train2014_000000459951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl playing catcher\"."}], "task": "refering", "answer_template": "The \"a girl playing catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 120, 121, 122, 123, 145, 146, 167, 168, 169, 190, 191, 192, 213, 214, 215], "bbox": [0.236359375, 0.03434272300469484, 0.46451562500000004, 0.6618779342723005], "iscrowd": 0, "area": 17400.52885, "rle": {"size": [426, 640], "counts": "onn11V=5K5J6[Oc0K6K4L5L2N3L3O1O1O2N2M4M2M_OVEaNNJk:e1[E]Nk:a1>01O0O3N1O2N2O02fDmN\\:U1_EQO^:V1XEPOf:n1K4M3L4L3N2M2YIQMQ3Q3aKnLRO5\\5o2_KZMeNKl5m2RKeNl4]1QKeNn4^1nJdNR5_1hJdNW5`1dJbN\\5a1_JaN`5d1ZJ^Nf5c4O10000]ORIjJo6R5TImJm6P5TIRKl6l4VITKj6j4WIWKi6g4aIPK`6n4cIPK^6o4dImJ_6R5m000O2O000O100O1000O1JSKoGm4Q86001O0000000001O1O1VOoJYIR5e6oJ[IR5?hJQ57`JR5<iJS56bJQ58kJT56dJP55lJU57eJm45mJT58gJl43mJT59iJk41mJT5:kJl4MkJV5<lJm4IhJX5?mJn4ChJa2KJd0gMRO3[6S2]JF?RNjN5Z6P2dJgNROJ^1_Om4l1PKRNUO=S1Eh4i1nK_N\\OGf4f1SL`NYOJd4c1YL_NTOMe4`1_N_Na1d1]NYNg1f1o401O1O1N3N2N1NO1O10O010O01O010O0010O0O101O1O1KUDjNl;V15N1000000010O0001O001O0N3N1O2N2N1N3N4M2N2O2Ejc^4"}, "label": "a girl playing catcher"}]}
{"id": 459951, "image": "COCO_train2014_000000459951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher waiting for the ball\"."}], "task": "refering", "answer_template": "The \"the catcher waiting for the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 120, 121, 122, 123, 145, 146, 167, 168, 169, 190, 191, 192, 213, 214, 215], "bbox": [0.236359375, 0.03434272300469484, 0.46451562500000004, 0.6618779342723005], "iscrowd": 0, "area": 17400.52885, "rle": {"size": [426, 640], "counts": "onn11V=5K5J6[Oc0K6K4L5L2N3L3O1O1O2N2M4M2M_OVEaNNJk:e1[E]Nk:a1>01O0O3N1O2N2O02fDmN\\:U1_EQO^:V1XEPOf:n1K4M3L4L3N2M2YIQMQ3Q3aKnLRO5\\5o2_KZMeNKl5m2RKeNl4]1QKeNn4^1nJdNR5_1hJdNW5`1dJbN\\5a1_JaN`5d1ZJ^Nf5c4O10000]ORIjJo6R5TImJm6P5TIRKl6l4VITKj6j4WIWKi6g4aIPK`6n4cIPK^6o4dImJ_6R5m000O2O000O100O1000O1JSKoGm4Q86001O0000000001O1O1VOoJYIR5e6oJ[IR5?hJQ57`JR5<iJS56bJQ58kJT56dJP55lJU57eJm45mJT58gJl43mJT59iJk41mJT5:kJl4MkJV5<lJm4IhJX5?mJn4ChJa2KJd0gMRO3[6S2]JF?RNjN5Z6P2dJgNROJ^1_Om4l1PKRNUO=S1Eh4i1nK_N\\OGf4f1SL`NYOJd4c1YL_NTOMe4`1_N_Na1d1]NYNg1f1o401O1O1N3N2N1NO1O10O010O01O010O0010O0O101O1O1KUDjNl;V15N1000000010O0001O001O0N3N1O2N2N1N3N4M2N2O2Ejc^4"}, "label": "the catcher waiting for the ball"}]}
{"id": 459951, "image": "COCO_train2014_000000459951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl fielding a softball\"."}], "task": "refering", "answer_template": "The \"a little girl fielding a softball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 197, 198, 199, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 336, 338], "bbox": [0.48409375, 0.4727230046948357, 0.72959375, 0.9688262910798122], "iscrowd": 0, "area": 17384.638049999998, "rle": {"size": [426, 640], "counts": "_[Q4:h<8O1N2N2O0O2N2O1N2N2O1N2N2O1N2N2O1N2N2N2O1NO1O20O1O10O01O10O01O100O00100O1O010O0010OWN`NjG`1U8oN]GR1b8[ORGd0n8IeF8Y9l101O0O101O0\\Oe0N101O1O2N2M3N2N2N2M4N10j0VO2N1O1O2N1O1O2N1O1O2O0O101O00001O00O1O0O2N2O1N2O1N2O1N102M3M3N2M2O2M3N2aNmHbLV7V3[I_Lf6Z3h1I8I7I6M41N3N2N2M2O2N2M3N1N3N2N2M2O2N2M3N2M2O3N2M4L3M3M4M2M000001N2O1O1N2O001N2O1O1RLTHZ2d9K4K6J5J7J5K6I9H9F;F9F`bW2"}, "label": "a little girl fielding a softball"}]}
{"id": 459951, "image": "COCO_train2014_000000459951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pitcher bending down\"."}], "task": "refering", "answer_template": "The \"the pitcher bending down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 197, 198, 199, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 336, 338], "bbox": [0.48409375, 0.4727230046948357, 0.72959375, 0.9688262910798122], "iscrowd": 0, "area": 17384.638049999998, "rle": {"size": [426, 640], "counts": "_[Q4:h<8O1N2N2O0O2N2O1N2N2O1N2N2O1N2N2O1N2N2N2O1NO1O20O1O10O01O10O01O100O00100O1O010O0010OWN`NjG`1U8oN]GR1b8[ORGd0n8IeF8Y9l101O0O101O0\\Oe0N101O1O2N2M3N2N2N2M4N10j0VO2N1O1O2N1O1O2N1O1O2O0O101O00001O00O1O0O2N2O1N2O1N2O1N102M3M3N2M2O2M3N2aNmHbLV7V3[I_Lf6Z3h1I8I7I6M41N3N2N2M2O2N2M3N1N3N2N2M2O2N2M3N2M2O3N2M4L3M3M4M2M000001N2O1O1N2O001N2O1O1RLTHZ2d9K4K6J5J7J5K6I9H9F;F9F`bW2"}, "label": "the pitcher bending down"}]}
{"id": 246959, "image": "COCO_train2014_000000246959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with number 10 jersey\"."}], "task": "refering", "answer_template": "The \"girl with number 10 jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 159, 160, 161, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 202, 203, 204, 205, 206, 207, 219, 220, 221, 233, 234, 235, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 277, 278, 279, 282, 283], "bbox": [0.515480093676815, 0.39323437499999997, 0.9005620608899297, 0.8235468749999999], "iscrowd": 0, "area": 17673.6628, "rle": {"size": [640, 427], "counts": "[kY41lc04b@Mk:7SEJl:7SEKk:6TEKk:6TEKk:6TEKk:5TEMj:4VEMi:3WENh:2XEOf:1[EOc:3\\EOa:3_EM^:6bEJ\\:8dEHY::gEFW:=iECU:?kEAR:b0nE^OP:d0PF\\On9e0QF]On9d0QF]On9d0PF_On9b0QF_Om9<YEf0e:WOeD^OiM]1a=SOgDEdMZ1c=oNjDL^MW1f=lNmDP2Q;nMQES2n:lMSEV2k:kMSEW2k:jMUEX2i:hMVEZ2i:gMVEZ2h:gMXEY2h:gMXEZ2f:hMXEZ2f:gMZEZ2e:fM[E[2c:fM]Eh1`MjMR=?]Ee1iMgMh<e0_Eb1PNaMd<m0\\E_1[;bNeD\\1];dNcDZ1`;fN_DX1d;gN\\D8eM@R>8UCAVOg0JPOU>h0bBM[O;i>IeA6]O0o>KRAj0K[OT?U2h@kMY?]3001000O01000O100oN`K`Bd4]=bK\\Bb4a=R1L3N2M4M2M4M2M3N3L3N2M4M2N001O0000001O0000001O0001SNXC`Lh<_3fCULZ<i3TDjKn;R4]DeKd;Y4]DhKd;U4]DoK`;o3aDQL`;m3aDTL`;i3aDWL`;f3bD[L^;b3eD]L];^3fDcL\\;X3fDhL];OjBe14lMg1`0_;KgBg1<dM`1k0_;GfBg1^2bNo:DdBh1^2dNP;AcBk1]2eNQ;^OcBk1^2fNn:@eBi1]2gNm:AgBg1\\2hNl:@kBg1Y2jNk:^OnBf1X2lNj:\\OQCg1U2mNl:XORCj1S2mNl:UOUCm1P2nNe;Q1[DoNe;P1]DoNd;P1]DPOb;P1_DoNb;P1_DoNa;P1aDPO_;o0aDQO_;n0cDRO];m0dDRO];l0eDTO\\;i0fDVO\\;h0eDXO\\;e0fDZO\\;c0fD]O\\;`0eD_O];>dDC^;:cDE`;7bDIa;2aDNa;NaD1b;L_D5c;F_D;b;A_Do0T;mNnDV1R;fNnDZ1m>O001O001O001O0O2O0O10000O2N1L4L4L4N3N2N2M3N2L3Kfhi0"}, "label": "girl with number 10 jersey"}]}
{"id": 246959, "image": "COCO_train2014_000000246959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing the number ten on a yellow shirt with black , blue , and white shorts\"."}], "task": "refering", "answer_template": "The \"a woman wearing the number ten on a yellow shirt with black , blue , and white shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 159, 160, 161, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 202, 203, 204, 205, 206, 207, 219, 220, 221, 233, 234, 235, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 277, 278, 279, 282, 283], "bbox": [0.515480093676815, 0.39323437499999997, 0.9005620608899297, 0.8235468749999999], "iscrowd": 0, "area": 17673.6628, "rle": {"size": [640, 427], "counts": "[kY41lc04b@Mk:7SEJl:7SEKk:6TEKk:6TEKk:6TEKk:5TEMj:4VEMi:3WENh:2XEOf:1[EOc:3\\EOa:3_EM^:6bEJ\\:8dEHY::gEFW:=iECU:?kEAR:b0nE^OP:d0PF\\On9e0QF]On9d0QF]On9d0PF_On9b0QF_Om9<YEf0e:WOeD^OiM]1a=SOgDEdMZ1c=oNjDL^MW1f=lNmDP2Q;nMQES2n:lMSEV2k:kMSEW2k:jMUEX2i:hMVEZ2i:gMVEZ2h:gMXEY2h:gMXEZ2f:hMXEZ2f:gMZEZ2e:fM[E[2c:fM]Eh1`MjMR=?]Ee1iMgMh<e0_Eb1PNaMd<m0\\E_1[;bNeD\\1];dNcDZ1`;fN_DX1d;gN\\D8eM@R>8UCAVOg0JPOU>h0bBM[O;i>IeA6]O0o>KRAj0K[OT?U2h@kMY?]3001000O01000O100oN`K`Bd4]=bK\\Bb4a=R1L3N2M4M2M4M2M3N3L3N2M4M2N001O0000001O0000001O0001SNXC`Lh<_3fCULZ<i3TDjKn;R4]DeKd;Y4]DhKd;U4]DoK`;o3aDQL`;m3aDTL`;i3aDWL`;f3bD[L^;b3eD]L];^3fDcL\\;X3fDhL];OjBe14lMg1`0_;KgBg1<dM`1k0_;GfBg1^2bNo:DdBh1^2dNP;AcBk1]2eNQ;^OcBk1^2fNn:@eBi1]2gNm:AgBg1\\2hNl:@kBg1Y2jNk:^OnBf1X2lNj:\\OQCg1U2mNl:XORCj1S2mNl:UOUCm1P2nNe;Q1[DoNe;P1]DoNd;P1]DPOb;P1_DoNb;P1_DoNa;P1aDPO_;o0aDQO_;n0cDRO];m0dDRO];l0eDTO\\;i0fDVO\\;h0eDXO\\;e0fDZO\\;c0fD]O\\;`0eD_O];>dDC^;:cDE`;7bDIa;2aDNa;NaD1b;L_D5c;F_D;b;A_Do0T;mNnDV1R;fNnDZ1m>O001O001O001O0O2O0O10000O2N1L4L4L4N3N2N2M3N2L3Kfhi0"}, "label": "a woman wearing the number ten on a yellow shirt with black , blue , and white shorts"}]}
{"id": 246959, "image": "COCO_train2014_000000246959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy running in yellow\"."}], "task": "refering", "answer_template": "The \"the boy running in yellow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 147, 148, 161, 162, 163, 164, 177, 178, 179, 192, 193, 194, 207, 208, 209, 221, 222, 223, 224, 236, 237, 238, 239, 251, 252, 253, 266, 267, 281], "bbox": [0.739344262295082, 0.36345312500000004, 1.0, 0.814125], "iscrowd": 0, "area": 16360.018849999997, "rle": {"size": [640, 427], "counts": "joU63jc03O2N2N2N1O2N2N2M2_NIX_O9S`0;l_OG[?R1c@POY?T1e@mNZ?U1d@mNY?V1e@lNY?V1j@gNU<UOlEV2RNXN[<D`EU2YNPNY<NZEU2`NfMY<7TEU2R<nMkCT2T<nMiCT2U<oMgCT2Z;bMTC<_1T2o:nM`C1^1R2P;PN`C1]1Q2P;QNbC1Y1Q2R;QNcC1X1P2S;QNdC1V1P2S;RNeC1U1o1S;TNfCOS1P2U;SNfC0R1n1V;UNgCOP1n1W;UNgC0o0m1W;VNiCOm0m1W;WNjC0k0k1Z;VNjC1h0l1];M`D5_;M^D4b;N[D2f;0WDOk;3QDLR<6kCIW<9fCF\\<<aCDa<l31O1O1O1O1O2N1O2M3N1O2N1O2N2N1O2N2N1O2N1O3M3M3M4L3M4L3Mc0]O1O2O0O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1J6E<D;D<I7100O2N1O100O2N10000?B=BR1nNnD"}, "label": "the boy running in yellow"}]}
{"id": 246959, "image": "COCO_train2014_000000246959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow and black uniform playing soccer\"."}], "task": "refering", "answer_template": "The \"man in yellow and black uniform playing soccer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 147, 148, 161, 162, 163, 164, 177, 178, 179, 192, 193, 194, 207, 208, 209, 221, 222, 223, 224, 236, 237, 238, 239, 251, 252, 253, 266, 267, 281], "bbox": [0.739344262295082, 0.36345312500000004, 1.0, 0.814125], "iscrowd": 0, "area": 16360.018849999997, "rle": {"size": [640, 427], "counts": "joU63jc03O2N2N2N1O2N2N2M2_NIX_O9S`0;l_OG[?R1c@POY?T1e@mNZ?U1d@mNY?V1e@lNY?V1j@gNU<UOlEV2RNXN[<D`EU2YNPNY<NZEU2`NfMY<7TEU2R<nMkCT2T<nMiCT2U<oMgCT2Z;bMTC<_1T2o:nM`C1^1R2P;PN`C1]1Q2P;QNbC1Y1Q2R;QNcC1X1P2S;QNdC1V1P2S;RNeC1U1o1S;TNfCOS1P2U;SNfC0R1n1V;UNgCOP1n1W;UNgC0o0m1W;VNiCOm0m1W;WNjC0k0k1Z;VNjC1h0l1];M`D5_;M^D4b;N[D2f;0WDOk;3QDLR<6kCIW<9fCF\\<<aCDa<l31O1O1O1O1O2N1O2M3N1O2N1O2N2N1O2N2N1O2N1O3M3M3M4L3M4L3Mc0]O1O2O0O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1J6E<D;D<I7100O2N1O100O2N10000?B=BR1nNnD"}, "label": "man in yellow and black uniform playing soccer"}]}
{"id": 246959, "image": "COCO_train2014_000000246959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green shirt kicking a red ball\"."}], "task": "refering", "answer_template": "The \"a man in a green shirt kicking a red ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 125, 126, 140, 141, 142, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 231, 232, 233, 246, 247, 248, 262, 263, 277, 278], "bbox": [0.23576112412177985, 0.338640625, 0.6231147540983606, 0.8038124999999999], "iscrowd": 0, "area": 20194.864100000003, "rle": {"size": [640, 427], "counts": "i^o1j0Uc03L3N2N2N2M3N2N2M3N2N3M2M3NN2J6K4K6L4K4O2O1O1O0010M2M3N2M3M3N2M31O100O2N1O2N1O2N1O1O2O06Ke0ZOa0@>A>B5L2M4M3ZOe0K6J6J5K6K3N2M3N3L3N2M3N2M4M2M3N2M4M2O100O2O0O100O101N100O101N100O101N14M6UJ]Bm4f=aJeBb5l=3L3M3M4M2M4L3M4M2M3M4L3N0O1O2N1O2O0O2NlMXCQMg<P3\\CnLc<Q3aClL^<U3dCjLZ<W3iCfLW<Y3lCeLS<\\3PDbLn;^3VD_Lj;a3XD^Lf;c3]DZLb;f3_D[L`;e3aDYL_;g3cDVL_;j3bDSL_;4iCT2i0eM_;MVDZ2<gM`;OVDX2;fM`;2YDU28fM`;5[DR27fM`;8\\Do15gM_;:_Dm12gMa;;`Dl10fMa;?bDh1NfMa;b0dDf1LeMb;d0fDd1IeMb;h0gDa1IcMa;l0jD]1HcM`;o0kD\\1i<dNZCY1f<fN]CX1c<gN`CW1`<hNdCU1\\<jNgCT1Y<lNjCQ1U<oNnCm0T<ROPDf0U<YOnCa0V<^OnC:W<FkC3Z<LiCN[<1iCG\\<8gCB]<=c301O1N2N2M3N2N2N2N2N1NdZT3"}, "label": "a man in a green shirt kicking a red ball"}]}
{"id": 246959, "image": "COCO_train2014_000000246959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in a green shirt playing soccer\"."}], "task": "refering", "answer_template": "The \"a young man in a green shirt playing soccer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 125, 126, 140, 141, 142, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 231, 232, 233, 246, 247, 248, 262, 263, 277, 278], "bbox": [0.23576112412177985, 0.338640625, 0.6231147540983606, 0.8038124999999999], "iscrowd": 0, "area": 20194.864100000003, "rle": {"size": [640, 427], "counts": "i^o1j0Uc03L3N2N2N2M3N2N2M3N2N3M2M3NN2J6K4K6L4K4O2O1O1O0010M2M3N2M3M3N2M31O100O2N1O2N1O2N1O1O2O06Ke0ZOa0@>A>B5L2M4M3ZOe0K6J6J5K6K3N2M3N3L3N2M3N2M4M2M3N2M4M2O100O2O0O100O101N100O101N100O101N14M6UJ]Bm4f=aJeBb5l=3L3M3M4M2M4L3M4M2M3M4L3N0O1O2N1O2O0O2NlMXCQMg<P3\\CnLc<Q3aClL^<U3dCjLZ<W3iCfLW<Y3lCeLS<\\3PDbLn;^3VD_Lj;a3XD^Lf;c3]DZLb;f3_D[L`;e3aDYL_;g3cDVL_;j3bDSL_;4iCT2i0eM_;MVDZ2<gM`;OVDX2;fM`;2YDU28fM`;5[DR27fM`;8\\Do15gM_;:_Dm12gMa;;`Dl10fMa;?bDh1NfMa;b0dDf1LeMb;d0fDd1IeMb;h0gDa1IcMa;l0jD]1HcM`;o0kD\\1i<dNZCY1f<fN]CX1c<gN`CW1`<hNdCU1\\<jNgCT1Y<lNjCQ1U<oNnCm0T<ROPDf0U<YOnCa0V<^OnC:W<FkC3Z<LiCN[<1iCG\\<8gCB]<=c301O1N2N2M3N2N2N2N2N1NdZT3"}, "label": "a young man in a green shirt playing soccer"}]}
{"id": 66737, "image": "COCO_train2014_000000066737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the single stem of broccoli in the middle toward the bottom\"."}], "task": "refering", "answer_template": "The \"the single stem of broccoli in the middle toward the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 217, 218, 231, 232, 233, 234, 235, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 286, 287, 288, 289, 290, 291, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381], "bbox": [0.070421875, 0.5647916666666667, 0.688375, 1.0], "iscrowd": 0, "area": 42528.61109999998, "rle": {"size": [480, 640], "counts": "e\\e02l>3M3M4L3M3M3M3M4L3M3]Oc00000O2O000O1O1O2O0O1O1O100O1O0010O01O1O0010O01O00100O001O010O001O10O01O0010O01O1O010O001O010O1O0010O010O0100O010O010O010O10O0100O1O001O1O001O1O00100O001O1O1O001O1O001O10O01O1O001O1O1O001O1O010O1O001001N101O1O0O2O001N101O0O2O001N101O1O001O001O001N101O000000O10000O100O10O10O100O10000O10O001O10O01O1O010O001O10O01O001O10O01O00100O001O1O010O001O10O01O00100O001O00100O001O010O1O05L4L4K5L4K5L4L2M01O001O10O01O001O1O001O1O001O001O1O001O001OcL_M\\J`2`5fM^JZ2^5lM`JT2]5QN`JP2`5RN^Jn1b5TN\\Jl1d5VNYJk1g5WNVHUOV1d2d6XNTHWOV1a2f6YNSHXOT1`2i6YNQHZOT1]2k6ZNoG\\OT1Z2m6[NnG]OR1Y2P7[NlG_OR1V2R7\\NkG@Q1T2T7]NiGBQ1Q2V7^NgGDP1o1Y7^NfGEo0m1[7_NdGFP1k1\\7`NcGFo0k1^7`NaGHo0h1`7aN_GIP1f1a7aN_GKm0e1d7I[H7e7JZH6f7LWH6h7KWH5i7LVH4j7NTH2l7ORH2n70PH0Q80nG0R81lG0T82jGNV83iGMW84gGMY85eGK[86dGJ\\88aGI`87_GIa88^GHb8[300000000001O00001O1O001O001O1O001O001O1O001O001O1O010O001O001O001O001O001O00001O001O001O001O001N1O2M2O2N1N2O2N1N3N1O2M3N1O2M2O2N2M2O2N2M2O2O1O001O1N101O1O001N2O001O1O001N2O1O1O1O1N2O1O1O1O001N2O1O1O1O1N2O1O1O1O1N2O1N2N3M3M2N3M3M2O2M3M2N3M3M2N6J=D?@o0QOhkl2"}, "label": "the single stem of broccoli in the middle toward the bottom"}]}
{"id": 66737, "image": "COCO_train2014_000000066737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoli with a long stem\"."}], "task": "refering", "answer_template": "The \"a piece of broccoli with a long stem\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 217, 218, 231, 232, 233, 234, 235, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 286, 287, 288, 289, 290, 291, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 356, 357, 358, 359, 360, 376, 377, 378, 379, 380, 381], "bbox": [0.070421875, 0.5647916666666667, 0.688375, 1.0], "iscrowd": 0, "area": 42528.61109999998, "rle": {"size": [480, 640], "counts": "e\\e02l>3M3M4L3M3M3M3M4L3M3]Oc00000O2O000O1O1O2O0O1O1O100O1O0010O01O1O0010O01O00100O001O010O001O10O01O0010O01O1O010O001O010O1O0010O010O0100O010O010O010O10O0100O1O001O1O001O1O00100O001O1O1O001O1O001O10O01O1O001O1O1O001O1O010O1O001001N101O1O0O2O001N101O0O2O001N101O1O001O001O001N101O000000O10000O100O10O10O100O10000O10O001O10O01O1O010O001O10O01O001O10O01O00100O001O1O010O001O10O01O00100O001O00100O001O010O1O05L4L4K5L4K5L4L2M01O001O10O01O001O1O001O1O001O001O1O001O001OcL_M\\J`2`5fM^JZ2^5lM`JT2]5QN`JP2`5RN^Jn1b5TN\\Jl1d5VNYJk1g5WNVHUOV1d2d6XNTHWOV1a2f6YNSHXOT1`2i6YNQHZOT1]2k6ZNoG\\OT1Z2m6[NnG]OR1Y2P7[NlG_OR1V2R7\\NkG@Q1T2T7]NiGBQ1Q2V7^NgGDP1o1Y7^NfGEo0m1[7_NdGFP1k1\\7`NcGFo0k1^7`NaGHo0h1`7aN_GIP1f1a7aN_GKm0e1d7I[H7e7JZH6f7LWH6h7KWH5i7LVH4j7NTH2l7ORH2n70PH0Q80nG0R81lG0T82jGNV83iGMW84gGMY85eGK[86dGJ\\88aGI`87_GIa88^GHb8[300000000001O00001O1O001O001O1O001O001O1O001O001O1O010O001O001O001O001O001O00001O001O001O001O001N1O2M2O2N1N2O2N1N3N1O2M3N1O2M2O2N2M2O2N2M2O2O1O001O1N101O1O001N2O001O1O001N2O1O1O1O1N2O1O1O1O001N2O1O1O1O1N2O1O1O1O1N2O1N2N3M3M2N3M3M2O2M3M2N3M3M2N6J=D?@o0QOhkl2"}, "label": "a piece of broccoli with a long stem"}]}
{"id": 66737, "image": "COCO_train2014_000000066737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cooked broccoli in a mix of veggies\"."}], "task": "refering", "answer_template": "The \"cooked broccoli in a mix of veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203], "bbox": [0.573109375, 0.08018750000000001, 1.0, 0.5377291666666667], "iscrowd": 0, "area": 27234.55345, "rle": {"size": [480, 640], "counts": "^V\\51n>1O100O1O2N100O10001O0O1000001N1000000O2O00000O101O000O10001O0O1000001N10000O101O0O10000O10000O10000O10O10O10000O10000O1000O010001N2O001N2O1O0O2O1O1N101O1N2O1O001N2O1O0O2O1O1N101O1N2O001N2O1O0bCRNi;P2VDRNh;o1WDSNg;n1XDTNe;m1[DUNc;l1[DVNd;k1[DWNb;k1]DWNa;i1_DYN^;i1`DZN^;g1`D\\N];f1aD]N];c1bD_N\\;c1bD`N[;e2N2M3N2M2N10O0100O010O00100O010O00100O010O10O01O010LPLWEQ4i:oKWEP4i:4O10O01O010O10O01O10O01O010O10O01O010O1O010O10O01O01000O010O10O010O10O10O0100O0100O10000O100O01000O10000O100O1000O0O2O1O001O1O0O2O1O1O001N2O001O1O1O001O1N101O1O1O001O1N10100001N101O0O101O001N10001N101ON2O001O1N101O1O1O0O200O00100O10O0100O00100O10O0100O00100O10O0100O010iN"}, "label": "cooked broccoli in a mix of veggies"}]}
{"id": 66737, "image": "COCO_train2014_000000066737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccoli to the left of the purple stem\"."}], "task": "refering", "answer_template": "The \"the broccoli to the left of the purple stem\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 234, 253, 254, 255, 276, 277, 278, 300], "bbox": [0.0, 0.32808333333333334, 0.21742187500000001, 0.7775208333333333], "iscrowd": 0, "area": 19418.99225, "rle": {"size": [480, 640], "counts": "R6S1b=Q2XNg1XNX1hN101N2O1N12O001O1N101O1O0O2O1O001O1O10O01000O01000O010bGSJi7m5VHTJi7l5VHVJk7i5THXJl7h5RHZJn7e5RH\\Jo7c5oG_JQ8`5nGbJR8^5mGdJS8[5kGgJU8X5kGiJU8W5iGkJX8T5gGmJY8R5fGPKZ8g3cGjL1A]8_3iGnLID^8Y3oGQMBH_8Q3VHRNk7h1\\HVNd7d1cHZN^7f1cHYN^7f1dHXN\\7h1eHWN[7i1fHVN[7i1fHVNZ7j1hHTNX7l1iHRNY7m1hHRNX7n1jHPNV7P2kHoMU7P2mHoMT7P2nHnMR7R2oHkMS7U2nHhMT7X2mHfMT7Z2nHbMT7^2QI\\MP7d2XIRMj6n2]IjLd6V3c2000001O000000000000000O100000000000000000000001O1O001O1O001O001O1O001O001O00001O001O00001O000011N2N3N1N2N2O2M2N2O1N3L3M3N2M3N3L3D=@?D=E:G_dZ7"}, "label": "the broccoli to the left of the purple stem"}]}
{"id": 246990, "image": "COCO_train2014_000000246990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in black outfit skateboarding in an indoor park\"."}], "task": "refering", "answer_template": "The \"boy in black outfit skateboarding in an indoor park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 55, 56, 57, 58, 70, 71, 72, 73, 86, 87, 88, 101, 102, 103, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 189, 190, 191, 192, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 232, 233, 242, 243, 244, 247, 248, 262, 263], "bbox": [0.13473067915690867, 0.12134375, 0.9498126463700235, 0.7617968749999999], "iscrowd": 0, "area": 45367.68990000001, "rle": {"size": [640, 427], "counts": "jeT1g0Yc00O100O100O100O100O1O100O100O2O0O100O100O1O100O100O100O100O100O1O2O0O1O100O1O100O1O1f]O`NTb0`1k]O`NUb0a1j]O`NUb0a1k]O^NUb0c1j]O^NVb0b1i]O^NWb0f1O010O00100O0010O01O0010O000000O0O2O1O1O1O1O1O001O1O1N2O1O1O001O1N2O1O1O1O0O2O1O1O1O11O1O010O10O0100O10O0100O10O0100O010O100O010O100O0O2O1N2NnMiM^BW2b=jM^BT2c=mM]BQ2d=QNZBn1g=SNYBl1g=VNXBh1j=XNVBf1k=\\NSBd1m=]NSBa1n=`NRB^1o=dNPB[1P>fNoAY1R>iNmAU1T>lNlAR1U>oNjAQ1W>POhAn0Y>SOgAm0X>UOgAm0V>TOiAn0U>TOjAm0T>TOlAm0R>TOnAm0P>UOnAn0P>ROPBo0n=SOQBn0m=SORBo0m=QORBQ1m=oNSBR1k=nNUBT1j=kNWBV1h=iNXBY1f=gNZB[1e=dN\\B]1l0hMV:j0nD`1h0iMY:g0oDb1e0iM\\:d0PEd1`0kM_:`0REg1;mMb:;SEj17nMf:7TEl10RNk:3TEn1KSNQ;NUER2BUNY;ITEU2]OWN];DVEY2VOWNd;@VE\\2nNYNl;ZOVEa2gNYNR<WOVEd2`NZNZ<QOWE_4i:aKVEa4h:_KYEa4g:_KXEb4h:]KXEd4g:]KYEc4g:\\KYEe4f:\\KYEf4f:YK[Eg4e:YKZEh4e:XK\\Eh4d:XK[Ei4e:VK[El4c:UK]Ek4c:TK]Em4c:SK\\En4c:RK^En4b:RK]Eo4c:PK]ER5a:oJ^EP5d:PK[Eo4f:RKYEn4h:QKXEc3gNmLS<@UEc3mNhLo;ESEb3SOdLk;KRE`3WO`Li;0oD`3]O[Le;5mD_3CWLb;:jD_3HRL_;?iDS2iN_NT1kN[;c0gDT2mNZNg=B[CU2POUNg=FXCU2UOQNc=KWCU2XOlMc=OTCV2[OhMb=1SCW2_OdM_=5QCX2B_M^=:oBX2E[M]==mBY2IUM\\=b0kBX2LSMZ=e0iBY2V>hMiAY2W>fMiAZ2X>fMgA[2Y>eMfA\\2Y>eMfA\\2Z>dMeA\\2[>eMdA\\2\\>dMcA]2\\>cMdA]2]>cMbA^2]>cMbA^2^>bMaA_2^>bMaA^2_>cMaA]2_>bMaA_2^>bMaA^2`>bM_A_2`>bM_A_2a>aM^A`2a>aM^A_2c>aM[Aa2d>_M\\Ab2d>^M[Ac2d>^M[Ab2e>_MZAb2f>^MYAc2f>^MXAc2i>]MVAd2i>\\MWAe2i>[MVAf2i>[MVAe2f;[LWGP1RMf2e;^LVGm0SMg2e;aLUGh0UMh2d;dLTGd0XMi2b;gLSG^4l8fKQG[4m8jKPGV4n8nKoFS4o8QLnFo3Q9ULmFk3R9XLkFi3U9WLjFj3V9VLiFk3V9VLgFl3Z9ULaFo3_9QL\\FT4c9mKYFW4g9jKSFZ4n9fKnE^4Q:cKkE`4V:`KgEc4Y:^KdEc4\\:cK]E_4c:fKVE[4k:jKnDX4Q;mKiDT4X;Z20O00100O010O10O0100O010O10O0100N2O001O1N101O1O0O2O1O001N2O001O1N2O1N2N3M2N2N2N3M2O1N2UJ_CV4c<hK`CT4c<kK`CQ4c<lK`CQ4c<lK`CP4d<nK^Co3d<oK`Cl3d<QL^Cm3e<PL^Cm3d<RL]Ck3g<RL\\Ck3g<RL[Ck3i<RLYCl3i<SLYCj3j<SLXCj3l<SLWCj3k<ULVCh3n<ULTCi3o<TLTCi3o<TLSCi3P=VLRCf3f>M3M4L3M4M3L3M4L4L3M4L4M3L5K5K6J5K5K5K5D<ROWn<"}, "label": "boy in black outfit skateboarding in an indoor park"}]}
{"id": 246990, "image": "COCO_train2014_000000246990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboarder has a white helmet on\"."}], "task": "refering", "answer_template": "The \"the skateboarder has a white helmet on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 55, 56, 57, 58, 70, 71, 72, 73, 86, 87, 88, 101, 102, 103, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 189, 190, 191, 192, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 232, 233, 242, 243, 244, 247, 248, 262, 263], "bbox": [0.13473067915690867, 0.12134375, 0.9498126463700235, 0.7617968749999999], "iscrowd": 0, "area": 45367.68990000001, "rle": {"size": [640, 427], "counts": "jeT1g0Yc00O100O100O100O100O1O100O100O2O0O100O100O1O100O100O100O100O100O1O2O0O1O100O1O100O1O1f]O`NTb0`1k]O`NUb0a1j]O`NUb0a1k]O^NUb0c1j]O^NVb0b1i]O^NWb0f1O010O00100O0010O01O0010O000000O0O2O1O1O1O1O1O001O1O1N2O1O1O001O1N2O1O1O1O0O2O1O1O1O11O1O010O10O0100O10O0100O10O0100O010O100O010O100O0O2O1N2NnMiM^BW2b=jM^BT2c=mM]BQ2d=QNZBn1g=SNYBl1g=VNXBh1j=XNVBf1k=\\NSBd1m=]NSBa1n=`NRB^1o=dNPB[1P>fNoAY1R>iNmAU1T>lNlAR1U>oNjAQ1W>POhAn0Y>SOgAm0X>UOgAm0V>TOiAn0U>TOjAm0T>TOlAm0R>TOnAm0P>UOnAn0P>ROPBo0n=SOQBn0m=SORBo0m=QORBQ1m=oNSBR1k=nNUBT1j=kNWBV1h=iNXBY1f=gNZB[1e=dN\\B]1l0hMV:j0nD`1h0iMY:g0oDb1e0iM\\:d0PEd1`0kM_:`0REg1;mMb:;SEj17nMf:7TEl10RNk:3TEn1KSNQ;NUER2BUNY;ITEU2]OWN];DVEY2VOWNd;@VE\\2nNYNl;ZOVEa2gNYNR<WOVEd2`NZNZ<QOWE_4i:aKVEa4h:_KYEa4g:_KXEb4h:]KXEd4g:]KYEc4g:\\KYEe4f:\\KYEf4f:YK[Eg4e:YKZEh4e:XK\\Eh4d:XK[Ei4e:VK[El4c:UK]Ek4c:TK]Em4c:SK\\En4c:RK^En4b:RK]Eo4c:PK]ER5a:oJ^EP5d:PK[Eo4f:RKYEn4h:QKXEc3gNmLS<@UEc3mNhLo;ESEb3SOdLk;KRE`3WO`Li;0oD`3]O[Le;5mD_3CWLb;:jD_3HRL_;?iDS2iN_NT1kN[;c0gDT2mNZNg=B[CU2POUNg=FXCU2UOQNc=KWCU2XOlMc=OTCV2[OhMb=1SCW2_OdM_=5QCX2B_M^=:oBX2E[M]==mBY2IUM\\=b0kBX2LSMZ=e0iBY2V>hMiAY2W>fMiAZ2X>fMgA[2Y>eMfA\\2Y>eMfA\\2Z>dMeA\\2[>eMdA\\2\\>dMcA]2\\>cMdA]2]>cMbA^2]>cMbA^2^>bMaA_2^>bMaA^2_>cMaA]2_>bMaA_2^>bMaA^2`>bM_A_2`>bM_A_2a>aM^A`2a>aM^A_2c>aM[Aa2d>_M\\Ab2d>^M[Ac2d>^M[Ab2e>_MZAb2f>^MYAc2f>^MXAc2i>]MVAd2i>\\MWAe2i>[MVAf2i>[MVAe2f;[LWGP1RMf2e;^LVGm0SMg2e;aLUGh0UMh2d;dLTGd0XMi2b;gLSG^4l8fKQG[4m8jKPGV4n8nKoFS4o8QLnFo3Q9ULmFk3R9XLkFi3U9WLjFj3V9VLiFk3V9VLgFl3Z9ULaFo3_9QL\\FT4c9mKYFW4g9jKSFZ4n9fKnE^4Q:cKkE`4V:`KgEc4Y:^KdEc4\\:cK]E_4c:fKVE[4k:jKnDX4Q;mKiDT4X;Z20O00100O010O10O0100O010O10O0100N2O001O1N101O1O0O2O1O001N2O001O1N2O1N2N3M2N2N2N3M2O1N2UJ_CV4c<hK`CT4c<kK`CQ4c<lK`CQ4c<lK`CP4d<nK^Co3d<oK`Cl3d<QL^Cm3e<PL^Cm3d<RL]Ck3g<RL\\Ck3g<RL[Ck3i<RLYCl3i<SLYCj3j<SLXCj3l<SLWCj3k<ULVCh3n<ULTCi3o<TLTCi3o<TLSCi3P=VLRCf3f>M3M4L3M4M3L3M4L4L3M4L4M3L5K5K6J5K5K5K5D<ROWn<"}, "label": "the skateboarder has a white helmet on"}]}
{"id": 246990, "image": "COCO_train2014_000000246990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboard being ridden by a skater in a indoor park\"."}], "task": "refering", "answer_template": "The \"skateboard being ridden by a skater in a indoor park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 276, 277], "bbox": [0.045503512880562064, 0.70525, 0.5941217798594848, 0.8070937499999999], "iscrowd": 0, "area": 3105.5186, "rle": {"size": [640, 427], "counts": "SZ<4mc02M3M010O1O010O001O10O01O010O00100O01O010O0010O01O1O010O1O1O010O1O0010O000010O00d\\O4db0LY]O7gb0JS]O<lb0<1O00000000001O000000000O1K5K5K5M4L3O10001N100000001O000000000000000000000001O000O1O1H8000000001O00000001O0001O000001O0000010O0000001O0001O01O0001O0001O0001O000000001O0000000010O00000001O00000000000001O0000000000000000000000001O00000000000001O04L5K5K2O0O8H001O010O001O0010O000O2N1[OU]O5lb0JW]O2jb0NX]OOjb00X]OJlb05a0O1O001N101O000O2O001O010O001O0001O00000O101O00000O10O10000000O10O100000O100000O1000O01N2O1NoX\\3"}, "label": "skateboard being ridden by a skater in a indoor park"}]}
{"id": 246990, "image": "COCO_train2014_000000246990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black skateboard with white wheels\"."}], "task": "refering", "answer_template": "The \"black skateboard with white wheels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 276, 277], "bbox": [0.045503512880562064, 0.70525, 0.5941217798594848, 0.8070937499999999], "iscrowd": 0, "area": 3105.5186, "rle": {"size": [640, 427], "counts": "SZ<4mc02M3M010O1O010O001O10O01O010O00100O01O010O0010O01O1O010O1O1O010O1O0010O000010O00d\\O4db0LY]O7gb0JS]O<lb0<1O00000000001O000000000O1K5K5K5M4L3O10001N100000001O000000000000000000000001O000O1O1H8000000001O00000001O0001O000001O0000010O0000001O0001O01O0001O0001O0001O000000001O0000000010O00000001O00000000000001O0000000000000000000000001O00000000000001O04L5K5K2O0O8H001O010O001O0010O000O2N1[OU]O5lb0JW]O2jb0NX]OOjb00X]OJlb05a0O1O001N101O000O2O001O010O001O0001O00000O101O00000O10O10000000O10O100000O100000O1000O01N2O1NoX\\3"}, "label": "black skateboard with white wheels"}]}
{"id": 394447, "image": "COCO_train2014_000000394447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"partial view on a woman ' s head\"."}], "task": "refering", "answer_template": "The \"partial view on a woman ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 52, 53, 65, 130, 131, 143], "bbox": [0.0002666666666666667, 0.17228, 0.14613333333333334, 0.62018], "iscrowd": 0, "area": 2228.2706, "rle": {"size": [500, 375], "counts": "]46]?>eD0X6NjI2U6MmI3R6LoI5Q6IQJ7n5ISJ7l5IUJ7k5HVJ7j5JVJ6j5IWJ7h5IXJ8h5GYJ9g5FZJ:]5OcJ1Y52hJNU54lJLQ56PKJm49RKHl49UKGh4;YKEd4<^KD`4<bKC]4=eKCY4>gKCX4=iKCV4>jKBT4?mKAR4?oKAP4?QLAm3a0SL_Ol3a0TL@l3?ULAk3>VLBi3>WLCi3<WLEi3:WLFi3:XLFh39_LAa3>mLUOR3k0l5O1O1O2N1O1O1O1O3M3M4L3M4K4LmTl4"}, "label": "partial view on a woman ' s head"}]}
{"id": 394447, "image": "COCO_train2014_000000394447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a red shirt\"."}], "task": "refering", "answer_template": "The \"a woman with a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 52, 53, 65, 130, 131, 143], "bbox": [0.0002666666666666667, 0.17228, 0.14613333333333334, 0.62018], "iscrowd": 0, "area": 2228.2706, "rle": {"size": [500, 375], "counts": "]46]?>eD0X6NjI2U6MmI3R6LoI5Q6IQJ7n5ISJ7l5IUJ7k5HVJ7j5JVJ6j5IWJ7h5IXJ8h5GYJ9g5FZJ:]5OcJ1Y52hJNU54lJLQ56PKJm49RKHl49UKGh4;YKEd4<^KD`4<bKC]4=eKCY4>gKCX4=iKCV4>jKBT4?mKAR4?oKAP4?QLAm3a0SL_Ol3a0TL@l3?ULAk3>VLBi3>WLCi3<WLEi3:WLFi3:XLFh39_LAa3>mLUOR3k0l5O1O1O2N1O1O1O1O3M3M4L3M4K4LmTl4"}, "label": "a woman with a red shirt"}]}
{"id": 394447, "image": "COCO_train2014_000000394447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a jacket with a white hood\"."}], "task": "refering", "answer_template": "The \"a man wearing a jacket with a white hood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 14, 15, 16, 27, 28, 29, 30, 40, 41, 42, 43, 44, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 91, 92, 93, 94, 95, 96, 97, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 170, 171, 172, 173, 183, 184, 185, 186, 197, 198, 199, 210, 211, 212, 223, 224], "bbox": [0.002986666666666667, 0.037520000000000005, 0.50336, 0.9835799999999999], "iscrowd": 0, "area": 54903.750850000004, "rle": {"size": [500, 375], "counts": "]d0<X?X1gNX1hNT1lN2O0O2N101N1O2N100O2N101N1O2O0O2N1O10O0000000000iF_K\\7a4cHeKX7[4fHlKU7T4jHRLQ7n3nHXLm6h3RI_Lh6a3WIeLd6[3[IoL\\6Q3cIZMS6f2lIbMm5^2RJeMl5[2SJhMk5X2SJlMk5T2lIXNQ6h1eIdNY6\\1^IoN`6Q1WIZOP5aLSLZ8]3WH`Lk7]3ZH_Li7]3[H`Lg7]3]H`Le7]3`H_Lb7^3bH_L`7^3eH^L]7_3gH^L\\7^3hH_LZ7^3kH^LW7_3mH^LZ7Z3Q3@`0]NjBHf=D^B;S>QOQBo0Z>00O10000O10000O10000O1000000000000000001O00000000000001O000000000000001O1O001O001O1O0010O01O1O002N3M4K5I6J7H7jBnMa<Y2XCkMd<e2K5L4L5K4K5L4L4K5L5K4K5L4L4K6J5K5K5J6K6J5K5J6K5K6J5K5J6K5K6J5J6K5K7I6J7H7J6J7I6I8I6J5K5J6K5K5K4L5J6K5K4L5J6K5K5K5K7H8I7I8H7H8I7I7I7I8G8I7I7I7Hddj2"}, "label": "a man wearing a jacket with a white hood"}]}
{"id": 394447, "image": "COCO_train2014_000000394447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue and white windbreaker turned away\"."}], "task": "refering", "answer_template": "The \"man in blue and white windbreaker turned away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 14, 15, 16, 27, 28, 29, 30, 40, 41, 42, 43, 44, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 91, 92, 93, 94, 95, 96, 97, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 170, 171, 172, 173, 183, 184, 185, 186, 197, 198, 199, 210, 211, 212, 223, 224], "bbox": [0.002986666666666667, 0.037520000000000005, 0.50336, 0.9835799999999999], "iscrowd": 0, "area": 54903.750850000004, "rle": {"size": [500, 375], "counts": "]d0<X?X1gNX1hNT1lN2O0O2N101N1O2N100O2N101N1O2O0O2N1O10O0000000000iF_K\\7a4cHeKX7[4fHlKU7T4jHRLQ7n3nHXLm6h3RI_Lh6a3WIeLd6[3[IoL\\6Q3cIZMS6f2lIbMm5^2RJeMl5[2SJhMk5X2SJlMk5T2lIXNQ6h1eIdNY6\\1^IoN`6Q1WIZOP5aLSLZ8]3WH`Lk7]3ZH_Li7]3[H`Lg7]3]H`Le7]3`H_Lb7^3bH_L`7^3eH^L]7_3gH^L\\7^3hH_LZ7^3kH^LW7_3mH^LZ7Z3Q3@`0]NjBHf=D^B;S>QOQBo0Z>00O10000O10000O10000O1000000000000000001O00000000000001O000000000000001O1O001O001O1O0010O01O1O002N3M4K5I6J7H7jBnMa<Y2XCkMd<e2K5L4L5K4K5L4L4K5L5K4K5L4L4K6J5K5K5J6K6J5K5J6K5K6J5K5J6K5K6J5J6K5K7I6J7H7J6J7I6I8I6J5K5J6K5K5K4L5J6K5K4L5J6K5K5K5K7H8I7I8H7H8I7I7I7I8G8I7I7I7Hddj2"}, "label": "man in blue and white windbreaker turned away"}]}
{"id": 394447, "image": "COCO_train2014_000000394447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses holding a hot dog\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses holding a hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 20, 21, 22, 23, 24, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 59, 60, 61, 62, 63, 72, 73, 74, 75, 76, 84, 85, 86, 87, 88, 89, 90, 96, 97, 98, 99, 100, 101, 102, 103, 109, 110, 111, 112, 113, 114, 115, 116, 122, 123, 124, 125, 126, 127, 128, 129, 134, 135, 136, 137, 138, 139, 140, 141, 142, 147, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 167, 168, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.25767999999999996, 0.0, 1.0, 0.98426], "iscrowd": 0, "area": 95987.0618, "rle": {"size": [500, 375], "counts": "gg_11a?4M2N2N2N2N2N2N2N2N2M3N2N2N2N3M2N2N2N2N2N2M3iNW1M3N2M3M3N2M3M4L3N2M3M3N2M3M3M3N2M3M3]FiKf7Y4nGRLP8Q4cG[L[8i3WGcLg8`3lFlLS9k4N2N2O1N2N2L4M3L4L4L4M3L4L4M3L4L4M3M3N2M3M3N2M3N2M300000000000000O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1[M^FWOc9c0eF[O[9>oF^OR98ZGEg80fGM[8HRH5o7YNaFa0n1R1b7]NgI`1Z6`NhI]1Y6cNiIY1Y6gNjIP1\\6POfIh0`6XOcI?c6A`I:b6F`I7a6IbI3_6McI1]6OfIM[63gIKY65jIGW69lIDT6<b400000000000000000000O2O00000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O100000ZD0h70WH>\\7BdHd0V7\\OiHf0V7ZOjHg0U7YOjHi0U7WOkHj0T7VOkHl0T7TOlHm0S7SOlHo0S7QOmHP1R7QOlHQ1S7QOkHo0U7SOgHP1X7ROcHR1\\7PO_HT1`7nN[HV1d7lNWHX1h7jNSHZ1l7hNoG\\1P84\\F1c90YF4f9MVF7i9ITF;k9FQF=o9DmE`0R:@kEd0T:_2O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001cM]2XKkN"}, "label": "a man wearing glasses holding a hot dog"}]}
{"id": 394447, "image": "COCO_train2014_000000394447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glass with camera and burger in his hand\"."}], "task": "refering", "answer_template": "The \"a man wearing glass with camera and burger in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 20, 21, 22, 23, 24, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 59, 60, 61, 62, 63, 72, 73, 74, 75, 76, 84, 85, 86, 87, 88, 89, 90, 96, 97, 98, 99, 100, 101, 102, 103, 109, 110, 111, 112, 113, 114, 115, 116, 122, 123, 124, 125, 126, 127, 128, 129, 134, 135, 136, 137, 138, 139, 140, 141, 142, 147, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 167, 168, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.25767999999999996, 0.0, 1.0, 0.98426], "iscrowd": 0, "area": 95987.0618, "rle": {"size": [500, 375], "counts": "gg_11a?4M2N2N2N2N2N2N2N2N2M3N2N2N2N3M2N2N2N2N2N2M3iNW1M3N2M3M3N2M3M4L3N2M3M3N2M3M3M3N2M3M3]FiKf7Y4nGRLP8Q4cG[L[8i3WGcLg8`3lFlLS9k4N2N2O1N2N2L4M3L4L4L4M3L4L4M3L4L4M3M3N2M3M3N2M3N2M300000000000000O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1[M^FWOc9c0eF[O[9>oF^OR98ZGEg80fGM[8HRH5o7YNaFa0n1R1b7]NgI`1Z6`NhI]1Y6cNiIY1Y6gNjIP1\\6POfIh0`6XOcI?c6A`I:b6F`I7a6IbI3_6McI1]6OfIM[63gIKY65jIGW69lIDT6<b400000000000000000000O2O00000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O100000ZD0h70WH>\\7BdHd0V7\\OiHf0V7ZOjHg0U7YOjHi0U7WOkHj0T7VOkHl0T7TOlHm0S7SOlHo0S7QOmHP1R7QOlHQ1S7QOkHo0U7SOgHP1X7ROcHR1\\7PO_HT1`7nN[HV1d7lNWHX1h7jNSHZ1l7hNoG\\1P84\\F1c90YF4f9MVF7i9ITF;k9FQF=o9DmE`0R:@kEd0T:_2O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001cM]2XKkN"}, "label": "a man wearing glass with camera and burger in his hand"}]}
{"id": 189646, "image": "COCO_train2014_000000189646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a grilled cheese sandwich sitting on top of its other half\"."}], "task": "refering", "answer_template": "The \"half of a grilled cheese sandwich sitting on top of its other half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 171, 172, 173], "bbox": [0.0471875, 0.12583333333333332, 0.589890625, 0.4179791666666667], "iscrowd": 0, "area": 36509.106100000005, "rle": {"size": [480, 640], "counts": "eV>1m>3M3M2N3M3M3M2O2O1N2N101N2N2O0O2N2O8G8I2M3M2O2M2N3N1N2N2O1N2N101N2N100O1O2O0O100O1O100O1O100O1O100O100O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O1000O0100O100O10000O100O10000O100O100O10000O100O10000O100O100O10O10O100O10000O100O100O1000O0100O100O100001O000000000O10000000000000000000000010O0000000000000000000000001O000001O0000000000000001O000000000000000001O000001O01O00001O00001O00001O01O01O00001O01O01O001O000010O01O001O001O01N10000000001O0000000000O1000O1001O1O1O1O1O1O1O1O001O1O001O001O001O1O001O001O001O1O001O001O1O001O0000000000000000000000000000000O1000000O10001O1N3N1O1O1N2O2N1O1N3N4L4L4K5L4L4L3L4M4L3L4M3M4L3L4M4J5\\OWej3"}, "label": "half of a grilled cheese sandwich sitting on top of its other half"}]}
{"id": 189646, "image": "COCO_train2014_000000189646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich half on top of other sandwich half\"."}], "task": "refering", "answer_template": "The \"sandwich half on top of other sandwich half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 171, 172, 173], "bbox": [0.0471875, 0.12583333333333332, 0.589890625, 0.4179791666666667], "iscrowd": 0, "area": 36509.106100000005, "rle": {"size": [480, 640], "counts": "eV>1m>3M3M2N3M3M3M2O2O1N2N101N2N2O0O2N2O8G8I2M3M2O2M2N3N1N2N2O1N2N101N2N100O1O2O0O100O1O100O1O100O1O100O100O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O1000O0100O100O10000O100O10000O100O100O10000O100O10000O100O100O10O10O100O10000O100O100O1000O0100O100O100001O000000000O10000000000000000000000010O0000000000000000000000001O000001O0000000000000001O000000000000000001O000001O01O00001O00001O00001O01O01O00001O01O01O001O000010O01O001O001O01N10000000001O0000000000O1000O1001O1O1O1O1O1O1O1O001O1O001O001O001O1O001O001O001O1O001O001O1O001O0000000000000000000000000000000O1000000O10001O1N3N1O1O1N2O2N1O1N3N4L4L4K5L4L4L3L4M4L3L4M3M4L3L4M4J5\\OWej3"}, "label": "sandwich half on top of other sandwich half"}]}
{"id": 189646, "image": "COCO_train2014_000000189646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"roasted brown color bread with in the yellow color plate\"."}], "task": "refering", "answer_template": "The \"roasted brown color bread with in the yellow color plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242], "bbox": [0.089125, 0.4079583333333333, 0.5767968750000001, 0.6209583333333334], "iscrowd": 0, "area": 26773.472349999996, "rle": {"size": [480, 640], "counts": "anj02m>3N1N3N1N3N2M2O2N1N3N2M2O2M3N1N3N1N3N2M2O2N1N3N2M2O2M3N1N2O0O101N100O10000O100O100O100O100O100O10000O100O100O100O101N100O100O10000O100O100O100O100O100O10000O100O100O101N100O100O100O10000O100000000000000000001O0000000000000000000000000000000000001O000000000000000O100000000000000000001O000000000000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O00000000000000000000000001O0O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000001O0O100000000000000000000O1000000000000000000O10000O100O10000O100O10000O10000O2O0O10000O100O10000O100nNR1F:E;F:FSkn3"}, "label": "roasted brown color bread with in the yellow color plate"}]}
{"id": 328917, "image": "COCO_train2014_000000328917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little boy with overalls on and a red shirt\"."}], "task": "refering", "answer_template": "The \"the little boy with overalls on and a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 125, 126, 127, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265], "bbox": [0.378359375, 0.26258333333333334, 0.5700625, 0.6773958333333334], "iscrowd": 0, "area": 16268.445049999998, "rle": {"size": [480, 640], "counts": "jfa32l>9H6I`0@6K3L4M4K3N1N3N1N2O1N2N3M2O1N2N2O10000O100O10000O100O100POcM]E]2Z:nMcES2Z:RNdEm1Z:XNcEi1Z:^NbEb1\\:bNaE_1b9fMbFo0I]1b9kM_Fi3]9aLZFb3b9P1M3O1N2O1N2N2O1N2O1N2O1000O0100O100O100O10000O100O010O100O10000000000O1000001O000O10000001O010O001O0WMVG\\Ok8c0YGYOg8g0^GTOc8j0cGRO\\8n0gGoNZ8P1hGnNX8Q1kGmNV8R1lGlNT8T1mGkNT8S1oGkNQ8U1PHkNP8T1RHjNn7U1THjNm7U1UHiNk7V1WHiNj7V1XHhNi7W1XHgNk7V1XHhNj7V1WHiNk7U1WHhNl7e0jFVN\\1T1k7b0jH[OY7`0kH_OW7<nHBS79SIDP77UIFm66XIFk66YIFn62XIIo6AeI:d:J6JhcP4"}, "label": "the little boy with overalls on and a red shirt"}]}
{"id": 328917, "image": "COCO_train2014_000000328917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in overalls petting a goat\"."}], "task": "refering", "answer_template": "The \"a little boy in overalls petting a goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 125, 126, 127, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265], "bbox": [0.378359375, 0.26258333333333334, 0.5700625, 0.6773958333333334], "iscrowd": 0, "area": 16268.445049999998, "rle": {"size": [480, 640], "counts": "jfa32l>9H6I`0@6K3L4M4K3N1N3N1N2O1N2N3M2O1N2N2O10000O100O10000O100O100POcM]E]2Z:nMcES2Z:RNdEm1Z:XNcEi1Z:^NbEb1\\:bNaE_1b9fMbFo0I]1b9kM_Fi3]9aLZFb3b9P1M3O1N2O1N2N2O1N2O1N2O1000O0100O100O100O10000O100O010O100O10000000000O1000001O000O10000001O010O001O0WMVG\\Ok8c0YGYOg8g0^GTOc8j0cGRO\\8n0gGoNZ8P1hGnNX8Q1kGmNV8R1lGlNT8T1mGkNT8S1oGkNQ8U1PHkNP8T1RHjNn7U1THjNm7U1UHiNk7V1WHiNj7V1XHhNi7W1XHgNk7V1XHhNj7V1WHiNk7U1WHhNl7e0jFVN\\1T1k7b0jH[OY7`0kH_OW7<nHBS79SIDP77UIFm66XIFk66YIFn62XIIo6AeI:d:J6JhcP4"}, "label": "a little boy in overalls petting a goat"}]}
{"id": 328917, "image": "COCO_train2014_000000328917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman petting a lamb\"."}], "task": "refering", "answer_template": "The \"a woman petting a lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.6809062499999999, 0.0022500000000000003, 0.9977499999999999, 0.9865208333333333], "iscrowd": 0, "area": 58465.08239999995, "rle": {"size": [480, 640], "counts": "fn\\6c0X>k0UO`0@4L4L5K4L4L4L4L4L4L4L4L4L5K4M3N2O1O1O1N2O1O2M2O1O1N2O1O1O2O100000000000000000000000000000000000000000000000000kJaKI_46bKJ^46bKJ^45cKK]4[OcKWL1^4\\4[OdKVL1^4[4\\OfKSL0a4Z4[OhKRLOc4Y4[OjKPLMe4Y4ZOlKoKLg4X4ZOmKmKLi4X4XOoKmKIk4X4XOPLkKIm4W4WOSLiKGP5V4WOTLhKFQ5V4WOaLi0_3VObLj0^3VObLj0^3UOcLk0]3UOcLk0]3TOdLl0\\3TOcLm0]3SOcLm0]3ROdLn0\\3ROcLo0]3POdLTNQO=[4e0eKYMo0b1YO;S4i0SLlLb0m1@9k3m0_MiNmNIVO^N^4o2`MgNTO@UOjNW4n2aMeN[OWOTOVOP4m2aMdNCmNSOBi3l2bMbNJdNRONc3j2bMaNV1eNX1i2cM_NX1hNU1h2cM^N[1jNR1g2dM\\N^1lNn0g2eMZN`1oNk0f2eMYNc1QOh0e2fMWNe1TOe0d2gMUNg1WOb0c2hMSNi1ZO?b2hMRNl1\\O<a2iMPNn1_O9`2jMnMP2B6_2jMmMS2E2]2lMkMU2HO\\2mMiMW2KL[2nMgMY2NIZ2nMfM\\20FY2oMdM^23CX2PNbM`26@W2PNaMc28^OU2U3kMkLT2V3lMjLS2W3mMiLR2X3nMhLQ2Y3oMgLP2Z3PNfLo1[3QNeLn1\\3RNdLn1\\3RNdLm1]3SNcLm1]3SNcLm1]3SNcLm1]3SNcLl1^3TNbLl1^3TNbLl1^3TNbLl1^3TNbLk1_3UNaLk1_3UNaLk1_3UNbLj1^3VNbLi1_3WNaLi1_3WNaLi1_3WNaLh1`3XN`Lh1`3XN`Lh1`3XN`Lh1`3XN`Lg1a3YN_Lg1a3YN_Lf1b3ZN^Lf1b3ZN^Lf1b3ZN^Le1c3[N]Le1c3[N]Ld1d3\\N\\Ld1d3\\N\\Lc1e3]N[Lc1e3]N[Lc1e3]N\\La1e3_N[La1e3_N[L`1f3`NZL`1f3`NZL_1g3aNYL_1g3aNYL^1h3bNXL^1h3bNXL^1h3bNXL]1i3cNXL\\1h3dNYLZ1h3fNYLY1g3gNZLW1g3iNZLV1f3jNZLU1g3kNZLT1f3lN[LS1e3mN\\LQ1e3oN\\LP1d3PO]Ln0d3RO]Lm0c3SO^Lk0c3UO^Lj0b3VO_Li0a3WO`Lg0b3XO_Lg0a3YO_Lf0b3ZO_Le0a3[O`Lc0a3]O`Lb0`3^OaL`0`3@aL?_3AbL=]3EdL:U3MlL1o25RMJh2<YMBb2d0^M\\O\\2j0eMTOY2o0hMPOV2R1kMlNU2U1lMjNR2X1oMfNP2\\1QNcNm1_1TN_Nk1c1VN\\Ni1e1XNYNg1i1f52N2N2NP?"}, "label": "a woman petting a lamb"}]}
{"id": 328917, "image": "COCO_train2014_000000328917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the dotted and striped shirt is petting the sheep\"."}], "task": "refering", "answer_template": "The \"the woman in the dotted and striped shirt is petting the sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.6809062499999999, 0.0022500000000000003, 0.9977499999999999, 0.9865208333333333], "iscrowd": 0, "area": 58465.08239999995, "rle": {"size": [480, 640], "counts": "fn\\6c0X>k0UO`0@4L4L5K4L4L4L4L4L4L4L4L4L5K4M3N2O1O1O1N2O1O2M2O1O1N2O1O1O2O100000000000000000000000000000000000000000000000000kJaKI_46bKJ^46bKJ^45cKK]4[OcKWL1^4\\4[OdKVL1^4[4\\OfKSL0a4Z4[OhKRLOc4Y4[OjKPLMe4Y4ZOlKoKLg4X4ZOmKmKLi4X4XOoKmKIk4X4XOPLkKIm4W4WOSLiKGP5V4WOTLhKFQ5V4WOaLi0_3VObLj0^3VObLj0^3UOcLk0]3UOcLk0]3TOdLl0\\3TOcLm0]3SOcLm0]3ROdLn0\\3ROcLo0]3POdLTNQO=[4e0eKYMo0b1YO;S4i0SLlLb0m1@9k3m0_MiNmNIVO^N^4o2`MgNTO@UOjNW4n2aMeN[OWOTOVOP4m2aMdNCmNSOBi3l2bMbNJdNRONc3j2bMaNV1eNX1i2cM_NX1hNU1h2cM^N[1jNR1g2dM\\N^1lNn0g2eMZN`1oNk0f2eMYNc1QOh0e2fMWNe1TOe0d2gMUNg1WOb0c2hMSNi1ZO?b2hMRNl1\\O<a2iMPNn1_O9`2jMnMP2B6_2jMmMS2E2]2lMkMU2HO\\2mMiMW2KL[2nMgMY2NIZ2nMfM\\20FY2oMdM^23CX2PNbM`26@W2PNaMc28^OU2U3kMkLT2V3lMjLS2W3mMiLR2X3nMhLQ2Y3oMgLP2Z3PNfLo1[3QNeLn1\\3RNdLn1\\3RNdLm1]3SNcLm1]3SNcLm1]3SNcLm1]3SNcLl1^3TNbLl1^3TNbLl1^3TNbLl1^3TNbLk1_3UNaLk1_3UNaLk1_3UNbLj1^3VNbLi1_3WNaLi1_3WNaLi1_3WNaLh1`3XN`Lh1`3XN`Lh1`3XN`Lh1`3XN`Lg1a3YN_Lg1a3YN_Lf1b3ZN^Lf1b3ZN^Lf1b3ZN^Le1c3[N]Le1c3[N]Ld1d3\\N\\Ld1d3\\N\\Lc1e3]N[Lc1e3]N[Lc1e3]N\\La1e3_N[La1e3_N[L`1f3`NZL`1f3`NZL_1g3aNYL_1g3aNYL^1h3bNXL^1h3bNXL^1h3bNXL]1i3cNXL\\1h3dNYLZ1h3fNYLY1g3gNZLW1g3iNZLV1f3jNZLU1g3kNZLT1f3lN[LS1e3mN\\LQ1e3oN\\LP1d3PO]Ln0d3RO]Lm0c3SO^Lk0c3UO^Lj0b3VO_Li0a3WO`Lg0b3XO_Lg0a3YO_Lf0b3ZO_Le0a3[O`Lc0a3]O`Lb0`3^OaL`0`3@aL?_3AbL=]3EdL:U3MlL1o25RMJh2<YMBb2d0^M\\O\\2j0eMTOY2o0hMPOV2R1kMlNU2U1lMjNR2X1oMfNP2\\1QNcNm1_1TN_Nk1c1VN\\Ni1e1XNYNg1i1f52N2N2NP?"}, "label": "the woman in the dotted and striped shirt is petting the sheep"}]}
{"id": 328917, "image": "COCO_train2014_000000328917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 259, 260, 261], "bbox": [0.0016718750000000002, 0.00020833333333333335, 0.498328125, 0.669125], "iscrowd": 0, "area": 73717.18930000001, "rle": {"size": [480, 640], "counts": "Yb03j>8I6I7J7H7I7J7H7J7H7I7J7H7J6I8H7J6I8I6I7I8I6I7J7H7I7J7H7J6I:F>C3L4L4M3L4M3L4L4M3N2O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O8H<D01O000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O01O00000O1N2M3N2M3N2M3N2M3L4K5L4K5N21O001O1O001O1O001O2N2N2N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O1O001O00000000000000001O000000000000000001O0001O00000000000000000000001O00PLoJQNQ5n1QKQNo4o1SKoMm4Q2UKmMk4R2WKmMi4S2YKkMg4T2\\KjMd4V2]KiMc4V2`KhM`4X2bKfM^4Y2dKgM[4Y2gKeMY4Z2iKeMW4[2kKcMU4\\2nKbMR4^2dLlL\\3S3^MTLb2l3WN[Ki1e4PObJP1]5Z400O100O100O100O1O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O2N1O1O1N3N1O1O2N1O2N2O05L4L4K5L4L4L1O1O1O001OROo0L3N2N2N2M4M2N2N2N2M4M2N2N2N3M2N2O1O1N3N1O1N2O1O2M2O1O1N2O2M2O1O1N3N1O1kNoCXOP>M3M3LoZf4"}, "label": "a man in a blue shirt"}]}
{"id": 328917, "image": "COCO_train2014_000000328917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaning down\"."}], "task": "refering", "answer_template": "The \"a man leaning down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 259, 260, 261], "bbox": [0.0016718750000000002, 0.00020833333333333335, 0.498328125, 0.669125], "iscrowd": 0, "area": 73717.18930000001, "rle": {"size": [480, 640], "counts": "Yb03j>8I6I7J7H7I7J7H7J7H7I7J7H7J6I8H7J6I8I6I7I8I6I7J7H7I7J7H7J6I:F>C3L4L4M3L4M3L4L4M3N2O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O8H<D01O000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O01O00000O1N2M3N2M3N2M3N2M3L4K5L4K5N21O001O1O001O1O001O2N2N2N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O1O001O00000000000000001O000000000000000001O0001O00000000000000000000001O00PLoJQNQ5n1QKQNo4o1SKoMm4Q2UKmMk4R2WKmMi4S2YKkMg4T2\\KjMd4V2]KiMc4V2`KhM`4X2bKfM^4Y2dKgM[4Y2gKeMY4Z2iKeMW4[2kKcMU4\\2nKbMR4^2dLlL\\3S3^MTLb2l3WN[Ki1e4PObJP1]5Z400O100O100O100O1O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O2N1O1O1N3N1O1O2N1O2N2O05L4L4K5L4L4L1O1O1O001OROo0L3N2N2N2M4M2N2N2N2M4M2N2N2N3M2N2O1O1N3N1O1N2O1O2M2O1O1N2O2M2O1O1N3N1O1kNoCXOP>M3M3LoZf4"}, "label": "a man leaning down"}]}
{"id": 165077, "image": "COCO_train2014_000000165077.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with red sweater and brown hair\"."}], "task": "refering", "answer_template": "The \"a boy with red sweater and brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 26, 27, 40, 41, 55, 69, 83, 96, 97, 110, 111, 124, 125, 138, 139, 152, 153], "bbox": [0.85955, 0.0022333333333333333, 1.0, 0.9887666666666667], "iscrowd": 0, "area": 11252.208100000002, "rle": {"size": [300, 400], "counts": "fiT3f0d84L3M4M2M3M4\\NHnI;j5LTJ8c5O\\J3\\54bJ0U57iJLn4<QKGc4c0]K^OW4m0iKUOo3Q1QLPOk3S1ULoNg3S1YLnNc3V1\\LlN`3V1`LkN]3W1cLkNX3X1hLiNU3Y1kLiNQ3Y1oLhNm2[1SMfNj2\\1VMfNf2\\1ZMeNb2^1^MdN^2^1bMdN[2]1eMgNU2[1kMiNP2X1PNlNj1V1VNnNe1S1[NQO_1Q1aNSOY1o0gNUOT1l0lNYOm0i0SO[Oh0f0XO^Ob0d0^O@<b0DB1e0O\\4000000000000000000000001O0^3bLaL"}, "label": "a boy with red sweater and brown hair"}]}
{"id": 165077, "image": "COCO_train2014_000000165077.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with wavy hair and a red shirt watching giraffes\"."}], "task": "refering", "answer_template": "The \"a person with wavy hair and a red shirt watching giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 26, 27, 40, 41, 55, 69, 83, 96, 97, 110, 111, 124, 125, 138, 139, 152, 153], "bbox": [0.85955, 0.0022333333333333333, 1.0, 0.9887666666666667], "iscrowd": 0, "area": 11252.208100000002, "rle": {"size": [300, 400], "counts": "fiT3f0d84L3M4M2M3M4\\NHnI;j5LTJ8c5O\\J3\\54bJ0U57iJLn4<QKGc4c0]K^OW4m0iKUOo3Q1QLPOk3S1ULoNg3S1YLnNc3V1\\LlN`3V1`LkN]3W1cLkNX3X1hLiNU3Y1kLiNQ3Y1oLhNm2[1SMfNj2\\1VMfNf2\\1ZMeNb2^1^MdN^2^1bMdN[2]1eMgNU2[1kMiNP2X1PNlNj1V1VNnNe1S1[NQO_1Q1aNSOY1o0gNUOT1l0lNYOm0i0SO[Oh0f0XO^Ob0d0^O@<b0DB1e0O\\4000000000000000000000001O0^3bLaL"}, "label": "a person with wavy hair and a red shirt watching giraffes"}]}
{"id": 165077, "image": "COCO_train2014_000000165077.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a red colour dress seeing their animals\"."}], "task": "refering", "answer_template": "The \"a boy wearing a red colour dress seeing their animals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 28, 29, 30, 31, 42, 43, 44, 45, 56, 57, 58, 59, 70, 71, 72, 73, 84, 85, 86, 87, 98, 99, 100, 101, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146], "bbox": [0.0, 0.16196666666666668, 0.49097499999999994, 0.9858999999999999], "iscrowd": 0, "area": 28705.693199999994, "rle": {"size": [300, 400], "counts": "c1W5Y1nK?_6000000000000001O001O00001O000000O100O100O100O10000O100000000000000O10000001O001O001O001O001O00O10000O10000O1000000O10000O100001O1O1O1O2N1O1O1O00001O001O00001O0nHRNf6o1XIRNh6o1UISNk6X2O001O001O001O001O001O002N3M2N3M4L5K4ZNiLoL[3k2gLUM]3e2fLZM\\3b2gL]M\\3^2fLbM]3_1bLaM5o0[3Z1kL^MLX1^3R1aMmNe2j0^MVOi2`0[M_Oj29YMGi24[MKf2WOWNi0o1mNUNS1S5O000000001O0000001O000000001O000000001O0000001OO1000000000000O100000000000000000000000000000000000000000000000000000000001O1O1O001O1O1O001O1O001O2N4L4L3M4L4L3M\\_k1"}, "label": "a boy wearing a red colour dress seeing their animals"}]}
{"id": 165077, "image": "COCO_train2014_000000165077.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy standing toward the back side in the left most side of the image\"."}], "task": "refering", "answer_template": "The \"boy standing toward the back side in the left most side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 28, 29, 30, 31, 42, 43, 44, 45, 56, 57, 58, 59, 70, 71, 72, 73, 84, 85, 86, 87, 98, 99, 100, 101, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146], "bbox": [0.0, 0.16196666666666668, 0.49097499999999994, 0.9858999999999999], "iscrowd": 0, "area": 28705.693199999994, "rle": {"size": [300, 400], "counts": "c1W5Y1nK?_6000000000000001O001O00001O000000O100O100O100O10000O100000000000000O10000001O001O001O001O001O00O10000O10000O1000000O10000O100001O1O1O1O2N1O1O1O00001O001O00001O0nHRNf6o1XIRNh6o1UISNk6X2O001O001O001O001O001O002N3M2N3M4L5K4ZNiLoL[3k2gLUM]3e2fLZM\\3b2gL]M\\3^2fLbM]3_1bLaM5o0[3Z1kL^MLX1^3R1aMmNe2j0^MVOi2`0[M_Oj29YMGi24[MKf2WOWNi0o1mNUNS1S5O000000001O0000001O000000001O000000001O0000001OO1000000000000O100000000000000000000000000000000000000000000000000000000001O1O1O001O1O1O001O1O001O2N4L4L3M4L4L3M\\_k1"}, "label": "boy standing toward the back side in the left most side of the image"}]}
{"id": 165077, "image": "COCO_train2014_000000165077.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe standing closest to the people\"."}], "task": "refering", "answer_template": "The \"the giraffe standing closest to the people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 47, 48, 49, 50, 60, 61, 62, 64, 74, 75, 76, 88, 89, 91, 102, 103], "bbox": [0.31420000000000003, 0.25, 0.6486500000000001, 0.6689333333333334], "iscrowd": 0, "area": 6387.8906, "rle": {"size": [300, 400], "counts": "]QU1:o8:F7I8H7I:G9H8H5L2M3M3M3M3M2NTObIlN]6n0oImNQ6o0XJmNf5T1cJcN\\5^1X12N1O1Oa0@1N3M2M4M2N2M3N2N3L6K4L3NO10OjN]JeNc5W1kJ_NT5b1RKXNn4g1^KnMb4R2\\100O100O1000000O0100000002N1O1O1O1O`NSIb0l6^OXI`0f6@^I=a6CTJIk57WJHh57ZJIe57]JHb58_JGa59aJF^5:cJF\\59fJGY59hJFX5:jJFT5:mJIo47RKMi42YK3a4M`K6\\4JfK8V4HkK9S4GoK8P4GTL7k3IXL4h3L`LL`34bLJ^36dLG]38gLB\\3>i20O1000O10000000000000000000000000000000000000JA_G`0`87O0O2001OO1M2O2N2O1O11N5L2N3M3L6K3L1O2N2N1MaVY1"}, "label": "the giraffe standing closest to the people"}]}
{"id": 165077, "image": "COCO_train2014_000000165077.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe bending its neck and front knees toward a group of visitors\"."}], "task": "refering", "answer_template": "The \"a giraffe bending its neck and front knees toward a group of visitors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 47, 48, 49, 50, 60, 61, 62, 64, 74, 75, 76, 88, 89, 91, 102, 103], "bbox": [0.31420000000000003, 0.25, 0.6486500000000001, 0.6689333333333334], "iscrowd": 0, "area": 6387.8906, "rle": {"size": [300, 400], "counts": "]QU1:o8:F7I8H7I:G9H8H5L2M3M3M3M3M2NTObIlN]6n0oImNQ6o0XJmNf5T1cJcN\\5^1X12N1O1Oa0@1N3M2M4M2N2M3N2N3L6K4L3NO10OjN]JeNc5W1kJ_NT5b1RKXNn4g1^KnMb4R2\\100O100O1000000O0100000002N1O1O1O1O`NSIb0l6^OXI`0f6@^I=a6CTJIk57WJHh57ZJIe57]JHb58_JGa59aJF^5:cJF\\59fJGY59hJFX5:jJFT5:mJIo47RKMi42YK3a4M`K6\\4JfK8V4HkK9S4GoK8P4GTL7k3IXL4h3L`LL`34bLJ^36dLG]38gLB\\3>i20O1000O10000000000000000000000000000000000000JA_G`0`87O0O2001OO1M2O2N2O1O11N5L2N3M3L6K3L1O2N2N1MaVY1"}, "label": "a giraffe bending its neck and front knees toward a group of visitors"}]}
{"id": 361685, "image": "COCO_train2014_000000361685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall urinal\"."}], "task": "refering", "answer_template": "The \"a tall urinal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375, 394, 395, 396, 397, 418, 419], "bbox": [0.08784375, 0.40428125000000004, 0.353609375, 0.789421875], "iscrowd": 0, "area": 34040.524900000004, "rle": {"size": [640, 640], "counts": "QYS1;ec0T1kNU1lNj0VO4K6K4L4K5L4L4K5L4L4K5L5K4K5L4L2M2O2N1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1O1N2O1O1N10000O2O000O10000O10000O101O0O10000O10000O10001N10000O10000O10000O2O00000001O00001N10001N100O2O0O101O0O101N10001N100O2O0O101O0O101N100O2O000O2O0O101N10001N1N2M4M2N2N3M2N2N3M2M5L4L4L4L4L3M4L4K5L4L4L4L4L3M4K5L4L4L4L4L4L4L3L5L4fNU_R8"}, "label": "a tall urinal"}]}
{"id": 361685, "image": "COCO_train2014_000000361685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one urinal lower than the other for young kids to use\"."}], "task": "refering", "answer_template": "The \"there is one urinal lower than the other for young kids to use\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 272, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389, 406, 407, 408, 409, 410, 411, 412, 430, 431, 432, 433, 434, 453, 454, 455, 456, 457, 478, 479], "bbox": [0.6851875, 0.509875, 0.929171875, 0.87878125], "iscrowd": 0, "area": 31015.539599999993, "rle": {"size": [640, 640], "counts": "mgb8j2Ra07I7I7I7I7K5K5K5K6J5K5K3M4L4L4L3M4L4N1O2M3N2N1N3N2N2N1N2O1O1O1N2O1O1O1N2O1O1O0O2O1O1O1O1O1O1O1O1O1O1O001O1O1N2O001O1O001O001O00000000000000000O1000000000000000000000000000000O1001O0000001O0000001O0000001O00001O0000001O00001O1N2N1O2N2M2O2N2N2N1O2M3N2N1O2N2M2O2M3M3L3N3L4M3L3N3L4M2M4M3L5L4K6J5L4K5L5J9Da0ZOf0[Oe0ZOf0[Od0\\Okfk0"}, "label": "there is one urinal lower than the other for young kids to use"}]}
{"id": 361685, "image": "COCO_train2014_000000361685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"urinal placed lower on wall\"."}], "task": "refering", "answer_template": "The \"urinal placed lower on wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 272, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389, 406, 407, 408, 409, 410, 411, 412, 430, 431, 432, 433, 434, 453, 454, 455, 456, 457, 478, 479], "bbox": [0.6851875, 0.509875, 0.929171875, 0.87878125], "iscrowd": 0, "area": 31015.539599999993, "rle": {"size": [640, 640], "counts": "mgb8j2Ra07I7I7I7I7K5K5K5K6J5K5K3M4L4L4L3M4L4N1O2M3N2N1N3N2N2N1N2O1O1O1N2O1O1O1N2O1O1O0O2O1O1O1O1O1O1O1O1O1O1O001O1O1N2O001O1O001O001O00000000000000000O1000000000000000000000000000000O1001O0000001O0000001O0000001O00001O0000001O00001O1N2N1O2N2M2O2N2N2N1O2M3N2N1O2N2M2O2M3M3L3N3L4M3L3N3L4M2M4M3L5L4K6J5L4K5L5J9Da0ZOf0[Oe0ZOf0[Od0\\Okfk0"}, "label": "urinal placed lower on wall"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby wearing a bib with giraffes on and sitting in a pink chair\"."}], "task": "refering", "answer_template": "The \"a baby wearing a bib with giraffes on and sitting in a pink chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 127, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 242, 243, 244, 245], "bbox": [0.50846875, 0.25, 0.70225, 0.5457169117647059], "iscrowd": 0, "area": 14132.049949999999, "rle": {"size": [544, 640], "counts": "em\\52\\`0d0I8J5J6K6I4L4M3N2M4L3M3M3M3L4N3N100000000O10000QO^M`Cb2]<eM_C[2^<kM^CV2_<PN^CP2_<VN[Cm1b<YNXCj1e<Z1M3M3N2M3M3N20000O10000O1000000O10000O10000O101O000O10000O100001O000010O0001O00001O001O00001O010O001O1O1O1O001O1O1O1O010O1O7I9`LWCW2Q=ZMZCb2d=M2O1O1O001O1N101O001O000O10000000001N100O1O1N2O1N2kNdAE^>OTBIm=KeBMeeU3"}, "label": "a baby wearing a bib with giraffes on and sitting in a pink chair"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child sitting at a table with a plate of food\"."}], "task": "refering", "answer_template": "The \"a young child sitting at a table with a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 127, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 242, 243, 244, 245], "bbox": [0.50846875, 0.25, 0.70225, 0.5457169117647059], "iscrowd": 0, "area": 14132.049949999999, "rle": {"size": [544, 640], "counts": "em\\52\\`0d0I8J5J6K6I4L4M3N2M4L3M3M3M3L4N3N100000000O10000QO^M`Cb2]<eM_C[2^<kM^CV2_<PN^CP2_<VN[Cm1b<YNXCj1e<Z1M3M3N2M3M3N20000O10000O1000000O10000O10000O101O000O10000O100001O000010O0001O00001O001O00001O010O001O1O1O1O001O1O1O1O010O1O7I9`LWCW2Q=ZMZCb2d=M2O1O1O001O1N101O001O000O10000000001N100O1O1N2O1N2kNdAE^>OTBIm=KeBMeeU3"}, "label": "a young child sitting at a table with a plate of food"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bumbo seat with a bear in it near a toddler and another person sitting at a red table\"."}], "task": "refering", "answer_template": "The \"a blue bumbo seat with a bear in it near a toddler and another person sitting at a red table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 207, 208, 209, 210, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351], "bbox": [0.005124999999999999, 0.4647426470588235, 0.38240625, 0.8378308823529411], "iscrowd": 0, "area": 34462.85990000001, "rle": {"size": [544, 640], "counts": "ko17R`0h0XOi0WOh0XOh0XOi0K4N2M4M2N2N2M2O1O1O2M2O1O1N2N2M4L3N2M3N2M3N3N100O10000O2O02N2N3M3M2O2M2N00000010O0000000001O01O00000001O01O000001O0001O000001O0001O000001O0001O000001O0001O0001O000001O0001O000001O0001O000001O0001O000001O01O1O001O010O1O001O001000O101O0O10O0O2O1O001N101O1O001N2O001O001N2O001O0O2O1O00O10O1000lN_KiEa4W:dKdE[4\\:kK^EV4b:oKYEQ4f:ULUEk3k:WLSEi3l:XLTEg3m:XLSEi3l:XLTEh3l:WLUEi3j:TLZEl3f:RL\\Em3d:TL\\El3d:SL\\En3c:SL]Em3c:RL^Em3b:TL^El3b:TL^El3a:VL]Ek3c:UL]Ek3b:WL]Eh3d:XL\\Eh3d:YL[Eg3f:YLYEg3g:YLXEh3h:YLWEf3k:ZLTEd3n:\\LRE^3T;cLkDX3[;hLcDT3b;lL^DQ3e;oL[DP3g;nLZDR3f;nLZDQ3g;oLYDP3h;oLXDQ3j;nLVDQ3k;oLUDQ3k;oLTDQ3n;mLRDS3o;mLQDR3P<nLoCR3S<mLmCR3T<mLlCT3T<mLkCR3W<mLhCS3Y<mLfCS3[<mLeCR3]<nLaCR3a<mL_CS3b<lL]CT3e<kLZCd1Y>a01N3N1O3L6K2M10O01000O0100O010O10000O2O1O001O1O001O1O001N2O003M7I8H7I8H7I8H7HX`a6"}, "label": "a blue bumbo seat with a bear in it near a toddler and another person sitting at a red table"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue child ' s seat\"."}], "task": "refering", "answer_template": "The \"a blue child ' s seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 207, 208, 209, 210, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351], "bbox": [0.005124999999999999, 0.4647426470588235, 0.38240625, 0.8378308823529411], "iscrowd": 0, "area": 34462.85990000001, "rle": {"size": [544, 640], "counts": "ko17R`0h0XOi0WOh0XOh0XOi0K4N2M4M2N2N2M2O1O1O2M2O1O1N2N2M4L3N2M3N2M3N3N100O10000O2O02N2N3M3M2O2M2N00000010O0000000001O01O00000001O01O000001O0001O000001O0001O000001O0001O000001O0001O0001O000001O0001O000001O0001O000001O0001O000001O01O1O001O010O1O001O001000O101O0O10O0O2O1O001N101O1O001N2O001O001N2O001O0O2O1O00O10O1000lN_KiEa4W:dKdE[4\\:kK^EV4b:oKYEQ4f:ULUEk3k:WLSEi3l:XLTEg3m:XLSEi3l:XLTEh3l:WLUEi3j:TLZEl3f:RL\\Em3d:TL\\El3d:SL\\En3c:SL]Em3c:RL^Em3b:TL^El3b:TL^El3a:VL]Ek3c:UL]Ek3b:WL]Eh3d:XL\\Eh3d:YL[Eg3f:YLYEg3g:YLXEh3h:YLWEf3k:ZLTEd3n:\\LRE^3T;cLkDX3[;hLcDT3b;lL^DQ3e;oL[DP3g;nLZDR3f;nLZDQ3g;oLYDP3h;oLXDQ3j;nLVDQ3k;oLUDQ3k;oLTDQ3n;mLRDS3o;mLQDR3P<nLoCR3S<mLmCR3T<mLlCT3T<mLkCR3W<mLhCS3Y<mLfCS3[<mLeCR3]<nLaCR3a<mL_CS3b<lL]CT3e<kLZCd1Y>a01N3N1O3L6K2M10O01000O0100O010O10000O2O1O001O1O001O1O001N2O003M7I8H7I8H7I8H7HX`a6"}, "label": "a blue child ' s seat"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hands that are eating the almost empty plate\"."}], "task": "refering", "answer_template": "The \"the hands that are eating the almost empty plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 320, 321, 343, 344, 365, 366, 367, 388, 389, 390, 411, 412, 413, 435, 436], "bbox": [0.59196875, 0.5286029411764706, 1.0, 0.9594117647058825], "iscrowd": 0, "area": 18839.703649999992, "rle": {"size": [544, 640], "counts": "PdY6k0U`01O1O2O0O1O1O101N1O1O100O1O1O1O100O1O1O100O000001O0O10000O2O00000O2O000O1000000O10000O02O0000001O00001O000010O001O100O1O100O1O1O100O0010O01O0010O01O010O000001O0000001O0000001O0000001O00000010O00000001O0000001O0000001O0000001O0000001O01O0001O0000001O000000001O00000O2O0000001O0000001O0O10001`@aN]?_1c@aN]?_1c@aN]?_1c@aN]?a1000000O100000001O0000000O10000000000000001O0O1000000000000000000O101O000000000001O000000001O0000000000RDWOf7h0oGEo7;eG3Y8M\\G`0b8@YGg0e8YOZGj0e8UOYGo0e8QOZGR1d8nN[GU1c8kN[GX1d8hN[GY1f8fNXG]1g8cNXG^1h8bNWG_1i8aNUGa1k8_NTGb1l8^NPGf1Q9YNiFm1W9RNcFV2\\9jM^F\\2b9dM\\F^2c9cM[F_2e9aMZF`2e9`M[Fa2e9_MYFc2g9]MXFd2g9]MXFe2g9ZMYFg2f9ZMYFg2g9YMXFh2g9YMXFh2h9WMXFj2h9VMWFk2h9VMWFl2h9TMWFm2h9TMWFm2i9RMXFn2h9RMXFn2g9SMYFm2g9SMZFk2f9UM[Fk2e9UM[Fk2e9UM[Fk2d9VM\\Fi2e9WM[Fi2d9WM]Fi2c9WM]Fi2c9WM]Fh2c9YM]Fg2c9XM^Fh2a9YM_Fg2a9YM_Ff2a9[M^Ff2b9ZM^Ff2b9YM_Fg2`9ZM`FSOG[2j9bN_FQO7o1Y9QO`FoNb0_1U9AZFnNa0m0k9_OaF\\OD`0`:]OlFIeN5m1"}, "label": "the hands that are eating the almost empty plate"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arms of the person to the right of the child\"."}], "task": "refering", "answer_template": "The \"the arms of the person to the right of the child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 320, 321, 343, 344, 365, 366, 367, 388, 389, 390, 411, 412, 413, 435, 436], "bbox": [0.59196875, 0.5286029411764706, 1.0, 0.9594117647058825], "iscrowd": 0, "area": 18839.703649999992, "rle": {"size": [544, 640], "counts": "PdY6k0U`01O1O2O0O1O1O101N1O1O100O1O1O1O100O1O1O100O000001O0O10000O2O00000O2O000O1000000O10000O02O0000001O00001O000010O001O100O1O100O1O1O100O0010O01O0010O01O010O000001O0000001O0000001O0000001O00000010O00000001O0000001O0000001O0000001O0000001O01O0001O0000001O000000001O00000O2O0000001O0000001O0O10001`@aN]?_1c@aN]?_1c@aN]?_1c@aN]?a1000000O100000001O0000000O10000000000000001O0O1000000000000000000O101O000000000001O000000001O0000000000RDWOf7h0oGEo7;eG3Y8M\\G`0b8@YGg0e8YOZGj0e8UOYGo0e8QOZGR1d8nN[GU1c8kN[GX1d8hN[GY1f8fNXG]1g8cNXG^1h8bNWG_1i8aNUGa1k8_NTGb1l8^NPGf1Q9YNiFm1W9RNcFV2\\9jM^F\\2b9dM\\F^2c9cM[F_2e9aMZF`2e9`M[Fa2e9_MYFc2g9]MXFd2g9]MXFe2g9ZMYFg2f9ZMYFg2g9YMXFh2g9YMXFh2h9WMXFj2h9VMWFk2h9VMWFl2h9TMWFm2h9TMWFm2i9RMXFn2h9RMXFn2g9SMYFm2g9SMZFk2f9UM[Fk2e9UM[Fk2e9UM[Fk2d9VM\\Fi2e9WM[Fi2d9WM]Fi2c9WM]Fi2c9WM]Fh2c9YM]Fg2c9XM^Fh2a9YM_Fg2a9YM_Ff2a9[M^Ff2b9ZM^Ff2b9YM_Fg2`9ZM`FSOG[2j9bN_FQO7o1Y9QO`FoNb0_1U9AZFnNa0m0k9_OaF\\OD`0`:]OlFIeN5m1"}, "label": "the arms of the person to the right of the child"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty wooden chair to the side of a child\"."}], "task": "refering", "answer_template": "The \"an empty wooden chair to the side of a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 119, 120, 138, 139, 140, 141, 142, 143, 165, 166, 189, 212], "bbox": [0.023234374999999998, 0.12707720588235294, 0.24484375000000003, 0.49299632352941175], "iscrowd": 0, "area": 10332.298050000003, "rle": {"size": [544, 640], "counts": "eT89[`0;D=C=kNcNoAT2Q>oMlAR2S>QNjAP2T>SNiAo1V>>E<N1N20000012nLUB<0k1n=fMYB8Kk1^>nMaAN3i1g>WN_A^1X?D:J6O10O10O1000O00100O001O001O10O01O0VAUO`=l0`BTO_=m0aBSO[=P1fBnNW=W1hBhNU=]1kBcNi<2VB^1R1`Nh<h1XCXNh<h1WCYNh<g1YCZNf<f1ZCZNf<f1ZCZNe<f1[C\\Nd<d1\\C\\Nc<e1]C[Nb<e1_C\\N_<e1aC]N\\<d1cC_NZ<a1gC`NX<`1hC`NW<`1jC`NU<a1jC`NU<`1lC`NS<a1lC`NT<_1mCaNR<`1nCaNP<`1oC`NR<_1oCaNQ<_1oCaNQ<_1nCbNQ<_1oCaNQ<_1oCaNQ<_1nCaNS<_1mCbNR<^1mCcNS<]1mCdNR<\\1nCdNS<[1mCeNS<[1nCdNR<\\1nCdNR<\\1nCdNR<\\1nCdNR<]1mCcNS<]1mCcNS<^1lCbNU<^1jCbNV<_1iCbNW<]1iCdNV<]1jCcNV<]1iCcNX<]1hCbNY<]1gCcNY<^1gCaNZ<^1fCbN[<]1fCbN[<]1fCbN[<]1eCdN[<[1fCeNZ<Z1fChNY<W1hCiNW<W1iCjNW<U1jCkNV<U1iClNV<U1iClNW<S1iCnNZ<o0eCROb<f0_CYOe<d0ZC]Og<b0XC^Ok<?UCBl<=SCCP=:QCFQ=8oBGS=7nBIT=5kBKV=4kBLU=3lBLT=5lBKS=5mBKT=5lBKS=5nBJS=6lBKS=5nBJS=6mBJ`<[OkBk0e0JT<[OQC15Q1`0BZ<\\OPC61[18SOf<]OPC5Ol1MbNR=CmB04f2j<kMQCa2h<T1G3M2O1XD_KP;OdDU48nKU;KdDY43nKe;ORDi30_Lo;CnC14_4n;`KoC04_4n;`KXD_4a<D<C=Do0QOj1RNX[P8"}, "label": "an empty wooden chair to the side of a child"}]}
{"id": 181475, "image": "COCO_train2014_000000181475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of a wooden chair behind a teddy bear\"."}], "task": "refering", "answer_template": "The \"the top of a wooden chair behind a teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 119, 120, 138, 139, 140, 141, 142, 143, 165, 166, 189, 212], "bbox": [0.023234374999999998, 0.12707720588235294, 0.24484375000000003, 0.49299632352941175], "iscrowd": 0, "area": 10332.298050000003, "rle": {"size": [544, 640], "counts": "eT89[`0;D=C=kNcNoAT2Q>oMlAR2S>QNjAP2T>SNiAo1V>>E<N1N20000012nLUB<0k1n=fMYB8Kk1^>nMaAN3i1g>WN_A^1X?D:J6O10O10O1000O00100O001O001O10O01O0VAUO`=l0`BTO_=m0aBSO[=P1fBnNW=W1hBhNU=]1kBcNi<2VB^1R1`Nh<h1XCXNh<h1WCYNh<g1YCZNf<f1ZCZNf<f1ZCZNe<f1[C\\Nd<d1\\C\\Nc<e1]C[Nb<e1_C\\N_<e1aC]N\\<d1cC_NZ<a1gC`NX<`1hC`NW<`1jC`NU<a1jC`NU<`1lC`NS<a1lC`NT<_1mCaNR<`1nCaNP<`1oC`NR<_1oCaNQ<_1oCaNQ<_1nCbNQ<_1oCaNQ<_1oCaNQ<_1nCaNS<_1mCbNR<^1mCcNS<]1mCdNR<\\1nCdNS<[1mCeNS<[1nCdNR<\\1nCdNR<\\1nCdNR<\\1nCdNR<]1mCcNS<]1mCcNS<^1lCbNU<^1jCbNV<_1iCbNW<]1iCdNV<]1jCcNV<]1iCcNX<]1hCbNY<]1gCcNY<^1gCaNZ<^1fCbN[<]1fCbN[<]1fCbN[<]1eCdN[<[1fCeNZ<Z1fChNY<W1hCiNW<W1iCjNW<U1jCkNV<U1iClNV<U1iClNW<S1iCnNZ<o0eCROb<f0_CYOe<d0ZC]Og<b0XC^Ok<?UCBl<=SCCP=:QCFQ=8oBGS=7nBIT=5kBKV=4kBLU=3lBLT=5lBKS=5mBKT=5lBKS=5nBJS=6lBKS=5nBJS=6mBJ`<[OkBk0e0JT<[OQC15Q1`0BZ<\\OPC61[18SOf<]OPC5Ol1MbNR=CmB04f2j<kMQCa2h<T1G3M2O1XD_KP;OdDU48nKU;KdDY43nKe;ORDi30_Lo;CnC14_4n;`KoC04_4n;`KXD_4a<D<C=Do0QOj1RNX[P8"}, "label": "the top of a wooden chair behind a teddy bear"}]}
{"id": 255203, "image": "COCO_train2014_000000255203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the body and extended arm of the person in black and white seated behind cupcakes\"."}], "task": "refering", "answer_template": "The \"the body and extended arm of the person in black and white seated behind cupcakes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 55, 58, 59, 60, 69, 70, 71, 72, 73, 74, 82, 83, 92, 93, 94, 95, 96, 97, 115, 116, 117, 135, 138, 139, 140, 158, 159, 161, 162, 181, 182, 184, 185, 204, 205, 207, 227, 228], "bbox": [0.0, 0.0, 0.95778125, 0.6778537735849057], "iscrowd": 0, "area": 40361.154449999995, "rle": {"size": [424, 640], "counts": "U4T4Q5R4100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000^OiJRIX5l6c0M3N2N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2M3M3N2M3M3M3M3N2M3M210O100O100O100O100O100O100O100O1O100O100O1000000O1000000000000O10000000000O10000000000O100000000H8H8G9H8G9N200001O000000001O0000001OR1nN000000000000000000J6D<C=I7N2N2N2N2N2N2N2O1N2N2N2O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100mNiD4X;LR100O100O10000O100008H9G8H8H1O1O00001O00001O0000001O00001O000O2O00001O00001O00001O00001O00001O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O00M3N2N2M3N2N2M3N2N2N2O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100001O001O10O10000O10O10O100O1000O010000O1000O010000O1000O0100O10000O01000O10000O01000O100O10O10O10000O10O10O10000O10O010000O1G9H70100O1O100O01ZDLW:3cE3^:MZE:g:ESEa0n:_OkDg0W;WOcDo0^;QO[DU1f;61O1O1O1O1O1O1M3L4M3M4L3M3M3L4M3M3M3Mb]T24YbkM5WOG]D>a;DYDa0e;BUDd0h;_ORDf0l;>N1N3aF_NQ7c1gHbN[7`1]HdNd7`1SHeNm7c1eGbN]8d1WG`Nj8h1iF\\NX9i22N1O101N1O1O2N3M3N2M4L3M3M3M3M3N2L5K4L4L4L4L4L5K4M3N2M3N2M3N1N3Nmn:"}, "label": "the body and extended arm of the person in black and white seated behind cupcakes"}]}
{"id": 255203, "image": "COCO_train2014_000000255203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in brown and tan sweater holding cupcakes\"."}], "task": "refering", "answer_template": "The \"woman in brown and tan sweater holding cupcakes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 55, 58, 59, 60, 69, 70, 71, 72, 73, 74, 82, 83, 92, 93, 94, 95, 96, 97, 115, 116, 117, 135, 138, 139, 140, 158, 159, 161, 162, 181, 182, 184, 185, 204, 205, 207, 227, 228], "bbox": [0.0, 0.0, 0.95778125, 0.6778537735849057], "iscrowd": 0, "area": 40361.154449999995, "rle": {"size": [424, 640], "counts": "U4T4Q5R4100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000^OiJRIX5l6c0M3N2N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2M3M3N2M3M3M3M3N2M3M210O100O100O100O100O100O100O100O1O100O100O1000000O1000000000000O10000000000O10000000000O100000000H8H8G9H8G9N200001O000000001O0000001OR1nN000000000000000000J6D<C=I7N2N2N2N2N2N2N2O1N2N2N2O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100mNiD4X;LR100O100O10000O100008H9G8H8H1O1O00001O00001O0000001O00001O000O2O00001O00001O00001O00001O00001O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O00M3N2N2M3N2N2M3N2N2N2O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100001O001O10O10000O10O10O100O1000O010000O1000O010000O1000O0100O10000O01000O10000O01000O100O10O10O10000O10O10O10000O10O010000O1G9H70100O1O100O01ZDLW:3cE3^:MZE:g:ESEa0n:_OkDg0W;WOcDo0^;QO[DU1f;61O1O1O1O1O1O1M3L4M3M4L3M3M3L4M3M3M3Mb]T24YbkM5WOG]D>a;DYDa0e;BUDd0h;_ORDf0l;>N1N3aF_NQ7c1gHbN[7`1]HdNd7`1SHeNm7c1eGbN]8d1WG`Nj8h1iF\\NX9i22N1O101N1O1O2N3M3N2M4L3M3M3M3M3N2L5K4L4L4L4L4L5K4M3N2M3N2M3N1N3Nmn:"}, "label": "woman in brown and tan sweater holding cupcakes"}]}
{"id": 255203, "image": "COCO_train2014_000000255203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red sleeve arm\"."}], "task": "refering", "answer_template": "The \"red sleeve arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67], "bbox": [0.5297031249999999, 0.0011320754716981133, 1.0, 0.1970754716981132], "iscrowd": 0, "area": 16887.71865, "rle": {"size": [424, 640], "counts": "[\\\\46f<=I7O001O100O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O00100O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O0000O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O10000000000000000001O0000000000000000001O0001O000000000001O0000000000000000001O0000000000000000100O00100O010O00100O010O1O010O010O1O010O1O010O010O0001O001N10001O0O10000O1000000O10001N1000000O10000O1000000O2O00000O10000O1000000O101O0O1000000O10000O1000001O0O10000000000O100000000O0100O10000O01000O1000O0100O1000O0100O1O100O001O1O1O1O001O1O1O1O10O100000000O100000000O100O1000O010000O100O10000O10000O100O10000O1M2J7I7"}, "label": "red sleeve arm"}]}
{"id": 255203, "image": "COCO_train2014_000000255203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red sleeve of a person\"."}], "task": "refering", "answer_template": "The \"the red sleeve of a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67], "bbox": [0.5297031249999999, 0.0011320754716981133, 1.0, 0.1970754716981132], "iscrowd": 0, "area": 16887.71865, "rle": {"size": [424, 640], "counts": "[\\\\46f<=I7O001O100O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O00100O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O0000O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O10000000000000000001O0000000000000000001O0001O000000000001O0000000000000000001O0000000000000000100O00100O010O00100O010O1O010O010O1O010O1O010O010O0001O001N10001O0O10000O1000000O10001N1000000O10000O1000000O2O00000O10000O1000000O101O0O1000000O10000O1000001O0O10000000000O100000000O0100O10000O01000O1000O0100O1000O0100O1O100O001O1O1O1O001O1O1O1O10O100000000O100000000O100O1000O010000O100O10000O10000O100O10000O1M2J7I7"}, "label": "the red sleeve of a person"}]}
{"id": 378090, "image": "COCO_train2014_000000378090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair on the right side of the clock\"."}], "task": "refering", "answer_template": "The \"a chair on the right side of the clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 128, 129, 130, 131, 139, 140, 141, 142, 150, 151, 152, 153, 164], "bbox": [0.6609121621621622, 0.6387612612612613, 1.0, 0.9217117117117117], "iscrowd": 0, "area": 6429.9886, "rle": {"size": [444, 296], "counts": "jXe21j=101O1O1O6K6I8Hk0UO`0@5K2NN2J6F;O000000001O0001O010O010O010O00001O010O001O001O001N101O00001O001N101N1O2O001O00001O010O010O010O010O0010O000mM\\Do1c;oM_DR2`;mMaDT2c;1O1M3N2N2N2N2L4J6K4O<D3N3L3M3N0ON3fNRFcNQ:\\1ZFWNh9i1dFiM_9V2T1O001O1O0gNZ1K5K5KYG"}, "label": "a chair on the right side of the clock"}]}
{"id": 378090, "image": "COCO_train2014_000000378090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden armchair with striped cushion to the right of the table clock\"."}], "task": "refering", "answer_template": "The \"wooden armchair with striped cushion to the right of the table clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 128, 129, 130, 131, 139, 140, 141, 142, 150, 151, 152, 153, 164], "bbox": [0.6609121621621622, 0.6387612612612613, 1.0, 0.9217117117117117], "iscrowd": 0, "area": 6429.9886, "rle": {"size": [444, 296], "counts": "jXe21j=101O1O1O6K6I8Hk0UO`0@5K2NN2J6F;O000000001O0001O010O010O010O00001O010O001O001O001N101O00001O001N101N1O2O001O00001O010O010O010O010O0010O000mM\\Do1c;oM_DR2`;mMaDT2c;1O1M3N2N2N2N2L4J6K4O<D3N3L3M3N0ON3fNRFcNQ:\\1ZFWNh9i1dFiM_9V2T1O001O1O0gNZ1K5K5KYG"}, "label": "wooden armchair with striped cushion to the right of the table clock"}]}
{"id": 378090, "image": "COCO_train2014_000000378090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden couch with pillows on it on the left of a big living room with other furnitures in it\"."}], "task": "refering", "answer_template": "The \"wooden couch with pillows on it on the left of a big living room with other furnitures in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 99, 100, 101, 102, 111, 112, 113], "bbox": [0.05, 0.5040540540540541, 0.35172297297297295, 0.6794594594594595], "iscrowd": 0, "area": 3773.8437000000004, "rle": {"size": [444, 296], "counts": "kg62j=0O1Oo0RO0O2O0O1OO20O1O100O10003L4M1N10000O101N10000O0100O0100O10O10O100O01000O01000O010O10001N100O2O001N<E7MJ6I7VO[DSOh;l0\\DmNg;R1b0O1O100001Oh0WO10000000[OWDVOj;i0^DPOb;n0g0O1O1N2O1O1N2O1O1N2O1OSC]Oc<a0ZCDf<i00O1VOj0N2NfPc2"}, "label": "wooden couch with pillows on it on the left of a big living room with other furnitures in it"}]}
{"id": 378090, "image": "COCO_train2014_000000378090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch to the left of the fireplace\"."}], "task": "refering", "answer_template": "The \"the couch to the left of the fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 99, 100, 101, 102, 111, 112, 113], "bbox": [0.05, 0.5040540540540541, 0.35172297297297295, 0.6794594594594595], "iscrowd": 0, "area": 3773.8437000000004, "rle": {"size": [444, 296], "counts": "kg62j=0O1Oo0RO0O2O0O1OO20O1O100O10003L4M1N10000O101N10000O0100O0100O10O10O100O01000O01000O010O10001N100O2O001N<E7MJ6I7VO[DSOh;l0\\DmNg;R1b0O1O100001Oh0WO10000000[OWDVOj;i0^DPOb;n0g0O1O1N2O1O1N2O1O1N2O1OSC]Oc<a0ZCDf<i00O1VOj0N2NfPc2"}, "label": "the couch to the left of the fireplace"}]}
{"id": 378090, "image": "COCO_train2014_000000378090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair to the left of a clock\"."}], "task": "refering", "answer_template": "The \"a wooden chair to the left of a clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 144, 145, 147, 155], "bbox": [0.003243243243243243, 0.6236486486486486, 0.4049662162162162, 0.9130630630630631], "iscrowd": 0, "area": 8902.411849999997, "rle": {"size": [444, 296], "counts": "af0P1m<j0VOj0UO00D<L40000O100O100O1001O1O1O1O001O1O1O1O1O1O1O1O1O001O100001Y1fN01N1O1O00000O01000000O10O1K4G:F:M3O010O1O10O010000O010O10O10O101O2M4M3L3N3M3L4M3L2OO10O10O10O100O01000O10O01000O010000O010O1000O0100O01000O10O109G8G1000O10O1000nMjD^1V;bN_Eh0b:XOTF2S:GR2ObP\\2"}, "label": "a wooden chair to the left of a clock"}]}
{"id": 378090, "image": "COCO_train2014_000000378090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the bottom left\"."}], "task": "refering", "answer_template": "The \"the chair on the bottom left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 144, 145, 147, 155], "bbox": [0.003243243243243243, 0.6236486486486486, 0.4049662162162162, 0.9130630630630631], "iscrowd": 0, "area": 8902.411849999997, "rle": {"size": [444, 296], "counts": "af0P1m<j0VOj0UO00D<L40000O100O100O1001O1O1O1O001O1O1O1O1O1O1O1O1O001O100001Y1fN01N1O1O00000O01000000O10O1K4G:F:M3O010O1O10O010000O010O10O10O101O2M4M3L3N3M3L4M3L2OO10O10O10O100O01000O10O01000O010000O010O1000O0100O01000O10O109G8G1000O10O1000nMjD^1V;bN_Eh0b:XOTF2S:GR2ObP\\2"}, "label": "the chair on the bottom left"}]}
{"id": 238828, "image": "COCO_train2014_000000238828.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the beach chair the man is sitting on\"."}], "task": "refering", "answer_template": "The \"the beach chair the man is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 93, 94, 95, 96, 117, 118, 119, 120, 141, 142, 143, 165, 166, 167, 189, 190, 191, 213, 214, 215, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 295, 296, 297, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 313, 314, 315, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 346, 347, 348, 349, 350, 351, 352, 353, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 369, 370, 371, 372, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.040453125, 0.17527083333333332, 0.9994375, 0.9887708333333334], "iscrowd": 0, "area": 67185.46734999999, "rle": {"size": [480, 640], "counts": "RY<1n>2O1N2O1O1N2O1N2O1N2O1N2O1N2O1kLZOmGh0o7[OQHf0k7]OTHe0h7^OXHc0d7@\\Ha0`7C^H?^7DbH=Z7FeH<W7GiH:S7IlH9P7JPI7l6LSI6j6LVI5f6NYI4c6O^I1^62cINY65hIKT68lIIP6:QJFk5=VJCf5`0[J@a5c0`J]O\\5f0eJZOW5i0jJWOR5l0oJTOm4o0TKQOh4R1YKnNc4U1^KkN_4W1bKiNZ4Z1gKfNU4]1lKcNP4`1QL`Nk3c1VL]Nf3f1[LZNa3i1`LWN\\3l1eLTNW3o1jLQNR3R2oLnMm2U2TMkMh2X2YMhMe2Y2\\MgMc2Y2^MhM`2X2aMhM^2X2bMiM^2V2cMjM\\2V2eMjMZ2V2fMkMZ2T2fMmMZ2R2fMoM[2o1dMSN\\2l1dMUN\\2j1dMWN[2i1eMXNY2i1fMYNX2h1iMXNT2j1mMVNQ2k1PNUNn1l1SNTNk1m1VNSNh1o1XNQNf1P2[NPNc1Q2^NoM`1R2aNnM]1S2dNmMZ1T2hNkMV1T2mNlMQ1T2QOlMm0S2VOmMh0R2[OnMc0Q2@oM>Q2DoM:P2IPN5o1NQN0n13RNKn18QNFn1=RNAm1b0SN\\Ol1g0TNWOl1k0TNSOk1P1UNnNj1U1VNiNi1Z1WNdNi1^1WN`Nh1c1XN[Ng1h1YNUNh1n1WNPNh1S2XNkMg1X2YNfMg1\\2YNbMh1_2XN_Mi1b2WN\\Mk1d2UNZMl1g2TNWMm1j2SNTMn1m2RNQMP2P3oMnLR2S3nMkLS2V3mMhLU2X3kMfLV2[3jMcLW2^3iM`LY2`3gM^LZ2c3fM[L[2f3eMXL]2h3cMVL^2k3bMSL_2o3`MoKb2Q4^MmKc2T4]MjKd2W4\\MgKe2Z4[MdKg2\\4YMbKh2_4XM_Ki2b4WM\\Kk2d4UMYKm2h4SMVKn2k4RMSKP3n4oLPKR3Q5nLmJS3T5mLjJU3V5kLhJV3Y5jLeJW3^5hL_JY3e5dLYJ]3k5`LSJa3P6]LnId3S6\\LkIe3V6[LhIf3Y6ZLeIf3]6ZLaIg3`6YL^Ih3c6XL[Ii3f6WLXIj3i6VLUIk3l6ULRIl3o6TLoHl3S7TLkHm3V7SLhHn3Y7RLeHo3\\7QLbHP4_7PL_HQ4b7oK\\HQ4f7oKXHR4i7nKUHS4m7lKPHV4Q8jKmGW4T8Z11O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1N3N1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2N2M3N2N2N2N2M3N3M2N2N2M3N2N2N2N2M3K5L4K5L4L6I6K6M3O1N2O1N10O0O2O1N101O0O2O1O0O2O0O2O001N2O0010O100O10000O2O0O10000O100O100O2O000O100O100O10001N100O10N2N1O2N1O2N00000O101O0000000000000000000000000000000000000001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O00100O1O1O1O001O1O1O1O1O001OO1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O10000O10000O2O000O100O01000O10000O100O01000O10000O100O10O10000000000000000O10O10000000000000O100000000000000O01000000000000000O100000000000O1000O100000000000000O1000000000O100000O10000000000000000O1000O100000000000O100000000000000O10O10000000000000O10000000000000OO2L4M3L4M3L4M3L4K5J6K5K5J6O11O2M2O1O1O1O2N1O1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O00000000000000000O1001O00001O0000001O00001O0000001O00001O00010O01O1O010O10O10000O01000O10000O10000O01000O10000O3N2N2M4M2N3L3N3M2L5C<D=CmB"}, "label": "the beach chair the man is sitting on"}]}
{"id": 238828, "image": "COCO_train2014_000000238828.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair the man wearing a black shirt is sitting on\"."}], "task": "refering", "answer_template": "The \"chair the man wearing a black shirt is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 93, 94, 95, 96, 117, 118, 119, 120, 141, 142, 143, 165, 166, 167, 189, 190, 191, 213, 214, 215, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 295, 296, 297, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 313, 314, 315, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 346, 347, 348, 349, 350, 351, 352, 353, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 369, 370, 371, 372, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.040453125, 0.17527083333333332, 0.9994375, 0.9887708333333334], "iscrowd": 0, "area": 67185.46734999999, "rle": {"size": [480, 640], "counts": "RY<1n>2O1N2O1O1N2O1N2O1N2O1N2O1N2O1kLZOmGh0o7[OQHf0k7]OTHe0h7^OXHc0d7@\\Ha0`7C^H?^7DbH=Z7FeH<W7GiH:S7IlH9P7JPI7l6LSI6j6LVI5f6NYI4c6O^I1^62cINY65hIKT68lIIP6:QJFk5=VJCf5`0[J@a5c0`J]O\\5f0eJZOW5i0jJWOR5l0oJTOm4o0TKQOh4R1YKnNc4U1^KkN_4W1bKiNZ4Z1gKfNU4]1lKcNP4`1QL`Nk3c1VL]Nf3f1[LZNa3i1`LWN\\3l1eLTNW3o1jLQNR3R2oLnMm2U2TMkMh2X2YMhMe2Y2\\MgMc2Y2^MhM`2X2aMhM^2X2bMiM^2V2cMjM\\2V2eMjMZ2V2fMkMZ2T2fMmMZ2R2fMoM[2o1dMSN\\2l1dMUN\\2j1dMWN[2i1eMXNY2i1fMYNX2h1iMXNT2j1mMVNQ2k1PNUNn1l1SNTNk1m1VNSNh1o1XNQNf1P2[NPNc1Q2^NoM`1R2aNnM]1S2dNmMZ1T2hNkMV1T2mNlMQ1T2QOlMm0S2VOmMh0R2[OnMc0Q2@oM>Q2DoM:P2IPN5o1NQN0n13RNKn18QNFn1=RNAm1b0SN\\Ol1g0TNWOl1k0TNSOk1P1UNnNj1U1VNiNi1Z1WNdNi1^1WN`Nh1c1XN[Ng1h1YNUNh1n1WNPNh1S2XNkMg1X2YNfMg1\\2YNbMh1_2XN_Mi1b2WN\\Mk1d2UNZMl1g2TNWMm1j2SNTMn1m2RNQMP2P3oMnLR2S3nMkLS2V3mMhLU2X3kMfLV2[3jMcLW2^3iM`LY2`3gM^LZ2c3fM[L[2f3eMXL]2h3cMVL^2k3bMSL_2o3`MoKb2Q4^MmKc2T4]MjKd2W4\\MgKe2Z4[MdKg2\\4YMbKh2_4XM_Ki2b4WM\\Kk2d4UMYKm2h4SMVKn2k4RMSKP3n4oLPKR3Q5nLmJS3T5mLjJU3V5kLhJV3Y5jLeJW3^5hL_JY3e5dLYJ]3k5`LSJa3P6]LnId3S6\\LkIe3V6[LhIf3Y6ZLeIf3]6ZLaIg3`6YL^Ih3c6XL[Ii3f6WLXIj3i6VLUIk3l6ULRIl3o6TLoHl3S7TLkHm3V7SLhHn3Y7RLeHo3\\7QLbHP4_7PL_HQ4b7oK\\HQ4f7oKXHR4i7nKUHS4m7lKPHV4Q8jKmGW4T8Z11O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1N3N1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2N2M3N2N2N2N2M3N3M2N2N2M3N2N2N2N2M3K5L4K5L4L6I6K6M3O1N2O1N10O0O2O1N101O0O2O1O0O2O0O2O001N2O0010O100O10000O2O0O10000O100O100O2O000O100O100O10001N100O10N2N1O2N1O2N00000O101O0000000000000000000000000000000000000001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O00100O1O1O1O001O1O1O1O1O001OO1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O10000O10000O2O000O100O01000O10000O100O01000O10000O100O10O10000000000000000O10O10000000000000O100000000000000O01000000000000000O100000000000O1000O100000000000000O1000000000O100000O10000000000000000O1000O100000000000O100000000000000O10O10000000000000O10000000000000OO2L4M3L4M3L4M3L4K5J6K5K5J6O11O2M2O1O1O1O2N1O1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O00000000000000000O1001O00001O0000001O00001O0000001O00001O00010O01O1O010O10O10000O01000O10000O10000O01000O10000O3N2N2M4M2N3L3N3M2L5C<D=CmB"}, "label": "chair the man wearing a black shirt is sitting on"}]}
{"id": 238828, "image": "COCO_train2014_000000238828.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man reclining in a beach chair\"."}], "task": "refering", "answer_template": "The \"a man reclining in a beach chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364], "bbox": [0.22804687499999998, 0.14077083333333332, 1.0, 0.8952708333333333], "iscrowd": 0, "area": 61315.93979999998, "rle": {"size": [480, 640], "counts": "PbT28Z>a0C>J5K5K5K5K6J5K5K5K5K6J5K5L4L4M3L2N2N2O1N3M2N2O1N2N2O1N2N2O1O2N100O1O100O1O100O2O0O1O100O1O100O1O101N1O100O100O1O100O2N1O1O1O1O1N2O1O1O1N3N1O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1O1O1N2O1O1O001O2N1O2N1O2N2O0O2O0O2O0O2O0O2O1N100O1O10O0100O100O100O100O102M2O1N2O1N2O2M2N2O1N2O1N3N1O1O100O101O0O10oJZIf1e6VN^Il1b6PNaIQ2_6jMeIX2Y6dMkI]2U6_MmIc2R6YMRJh2n5TMUJm2k5oLXJR3g5^K]If0m0n3f5VKeIf0h0S4d5RKlIe0c0X4c5mJRJa0a0a4a6UKfIj4]6lJkIS5R8N101O1O0O2O1O1N101O1N101O1O1N101O1N101O1N2O001000O1000000O10000O1000000O1000000O1000000O1000000O10000O1000000O1000O1N1O2N2N2N1O2N2O1N1O2N2N1O2N2N2N1O2N2N2O0O2N2N2N1O2N2NeEgLR9W3nFjLR9U3nFlLS9R3mFoLT9o2lFRMT9n2kFSMV9k2jFVMW9h2iFYMX9e2hF\\MX9c2hF^MY9`2gFaMZ9^2eFcM\\9\\2cFeM]9Z2cFgM^9X2`FjMa9U2^FlMc9S2\\FnMd9R2[FoMf9P2YFPNi9o1VFRNk9m1TFTNl9k1UFUNl9j1TFVNm9i1SFWNn9h1RFXNn9h1RFXNo9g1QFYNP:f1oE[NR:d1nE\\NR:d1nE\\NS:b1nE^NS:a1mE_NS:a1mE_NT:`1lE`NU:_1kEaNV:^1jEbNV:^1iEcNX:\\1hEdNY:Z1hEfNY:Y1gEgNY:Y1gEgNZ:X1fEhN[:W1eEiN\\:V1dEjN\\:V1dEjN]:U1cEkN^:S1bEnN_:Q1aEoN_:Q1aEoN`:P1`EPOa:o0_EQOb:n0^EROb:n0^EROc:m0]ESOc:l0^ETO`:n0_ESO_:o0aEQO]:Q1cEoN[:S1eEmNZ:T1fElNX:V1hEjNW:W1iEiNW:V1jEjNU:W1kEiNU:W1kEiNT:X1kEiNU:W1kEiNT:X1lEhNT:X1lEhNS:Y1mEgNS:Y1mEgNR:Y1oEgNQ:Y1oEgNQ:Y1oEgNP:Z1oEgNQ:Y1oEgNP:Z1PFfNZ:P1fEPOZ:P1fEPOZ:o0gEQOY:o0gEPOZ:P1fEPOZ:P1fEPOZ:P1eEQO[:o0eEQO[:o0eEQO[:o0eEQO[:n0fEROZ:n0fEROZ:n0fEROZ:n0fEROZ:n0fERO[:m0eESO[:m0dETO\\:l0dETO\\:k0eEUO[:k0eEUO[:k0eEUO[:k0eEUOY:m0gESOX:n0hEROV:P1jEPOU:Q1jEPOU:Q1kEoNS:R1nEnNQ:S1oEmNo9U1QFkNn9V1RFjNm9W1SFiNl9X1TFhNn9V1RFjNo9U1PFlNQ:R1PFnNR:P1nEPOS:o0mEQOU:m0kESOV:l0jETOW:k0iEUOY:i0gEWOZ:h0fEXO[:f0fEZO\\:d0cE]O^:b0bE^O`:`0`E@a:?_EAb:>^EBd:<\\EDe:;[EEg:8ZEHg:7YEIh:6WEKj:4VELj:4VELj:4VELj:4VELj:4VEKk:4VEKk:5UEJl:6TEIl:8TEFn::QEFP;:PEEQ;;oDDR;<nDCS;=mDAU;>lDAU;?kD@V;`0jD_OW;a0iD^OX;b0gD^OZ;b0fD\\O\\;d0dD[O];e0cDZO^;e0cD\\O\\;d0dD]O[;c0eD^OY;c0gD]OY;c0gD^OX;b0gD@X;`0hDAW;?iDBV;=kDDT;<lDES;;mDFR;:nDGQ;9oDHP;8PEHP;8PEIo:7PEKo:4REMm:3SENl:2TEOj:2VEOi:1WEOi:1WEOi:1WEOi:1WEOi:0WE1i:OWE0j:1UEOk:1UEOk:2TENl:2TENl:3SELn:4SEKm:6UEGk:9WEEi:<XEBh:>[E_Od:c0]EZOd:f0^EXOb:i0`ETO`:l0bERO^:n0dEPO\\:Q1fEkN[:U1gEiNY:X1hEfNX:Z1g11O001O001O001O0O2O001O00001O001O001O001O001O3M3N2M4M2M3N2M4M2M3N2MN3M2N3M2N3I7G8HmE"}, "label": "a man reclining in a beach chair"}]}
{"id": 238828, "image": "COCO_train2014_000000238828.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on the beach reading his tablet\"."}], "task": "refering", "answer_template": "The \"a man on the beach reading his tablet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364], "bbox": [0.22804687499999998, 0.14077083333333332, 1.0, 0.8952708333333333], "iscrowd": 0, "area": 61315.93979999998, "rle": {"size": [480, 640], "counts": "PbT28Z>a0C>J5K5K5K5K6J5K5K5K5K6J5K5L4L4M3L2N2N2O1N3M2N2O1N2N2O1N2N2O1O2N100O1O100O1O100O2O0O1O100O1O100O1O101N1O100O100O1O100O2N1O1O1O1O1N2O1O1O1N3N1O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1O1O1N2O1O1O001O2N1O2N1O2N2O0O2O0O2O0O2O0O2O1N100O1O10O0100O100O100O100O102M2O1N2O1N2O2M2N2O1N2O1N3N1O1O100O101O0O10oJZIf1e6VN^Il1b6PNaIQ2_6jMeIX2Y6dMkI]2U6_MmIc2R6YMRJh2n5TMUJm2k5oLXJR3g5^K]If0m0n3f5VKeIf0h0S4d5RKlIe0c0X4c5mJRJa0a0a4a6UKfIj4]6lJkIS5R8N101O1O0O2O1O1N101O1N101O1O1N101O1N101O1N2O001000O1000000O10000O1000000O1000000O1000000O1000000O10000O1000000O1000O1N1O2N2N2N1O2N2O1N1O2N2N1O2N2N2N1O2N2N2O0O2N2N2N1O2N2NeEgLR9W3nFjLR9U3nFlLS9R3mFoLT9o2lFRMT9n2kFSMV9k2jFVMW9h2iFYMX9e2hF\\MX9c2hF^MY9`2gFaMZ9^2eFcM\\9\\2cFeM]9Z2cFgM^9X2`FjMa9U2^FlMc9S2\\FnMd9R2[FoMf9P2YFPNi9o1VFRNk9m1TFTNl9k1UFUNl9j1TFVNm9i1SFWNn9h1RFXNn9h1RFXNo9g1QFYNP:f1oE[NR:d1nE\\NR:d1nE\\NS:b1nE^NS:a1mE_NS:a1mE_NT:`1lE`NU:_1kEaNV:^1jEbNV:^1iEcNX:\\1hEdNY:Z1hEfNY:Y1gEgNY:Y1gEgNZ:X1fEhN[:W1eEiN\\:V1dEjN\\:V1dEjN]:U1cEkN^:S1bEnN_:Q1aEoN_:Q1aEoN`:P1`EPOa:o0_EQOb:n0^EROb:n0^EROc:m0]ESOc:l0^ETO`:n0_ESO_:o0aEQO]:Q1cEoN[:S1eEmNZ:T1fElNX:V1hEjNW:W1iEiNW:V1jEjNU:W1kEiNU:W1kEiNT:X1kEiNU:W1kEiNT:X1lEhNT:X1lEhNS:Y1mEgNS:Y1mEgNR:Y1oEgNQ:Y1oEgNQ:Y1oEgNP:Z1oEgNQ:Y1oEgNP:Z1PFfNZ:P1fEPOZ:P1fEPOZ:o0gEQOY:o0gEPOZ:P1fEPOZ:P1fEPOZ:P1eEQO[:o0eEQO[:o0eEQO[:o0eEQO[:n0fEROZ:n0fEROZ:n0fEROZ:n0fEROZ:n0fERO[:m0eESO[:m0dETO\\:l0dETO\\:k0eEUO[:k0eEUO[:k0eEUO[:k0eEUOY:m0gESOX:n0hEROV:P1jEPOU:Q1jEPOU:Q1kEoNS:R1nEnNQ:S1oEmNo9U1QFkNn9V1RFjNm9W1SFiNl9X1TFhNn9V1RFjNo9U1PFlNQ:R1PFnNR:P1nEPOS:o0mEQOU:m0kESOV:l0jETOW:k0iEUOY:i0gEWOZ:h0fEXO[:f0fEZO\\:d0cE]O^:b0bE^O`:`0`E@a:?_EAb:>^EBd:<\\EDe:;[EEg:8ZEHg:7YEIh:6WEKj:4VELj:4VELj:4VELj:4VELj:4VEKk:4VEKk:5UEJl:6TEIl:8TEFn::QEFP;:PEEQ;;oDDR;<nDCS;=mDAU;>lDAU;?kD@V;`0jD_OW;a0iD^OX;b0gD^OZ;b0fD\\O\\;d0dD[O];e0cDZO^;e0cD\\O\\;d0dD]O[;c0eD^OY;c0gD]OY;c0gD^OX;b0gD@X;`0hDAW;?iDBV;=kDDT;<lDES;;mDFR;:nDGQ;9oDHP;8PEHP;8PEIo:7PEKo:4REMm:3SENl:2TEOj:2VEOi:1WEOi:1WEOi:1WEOi:1WEOi:0WE1i:OWE0j:1UEOk:1UEOk:2TENl:2TENl:3SELn:4SEKm:6UEGk:9WEEi:<XEBh:>[E_Od:c0]EZOd:f0^EXOb:i0`ETO`:l0bERO^:n0dEPO\\:Q1fEkN[:U1gEiNY:X1hEfNX:Z1g11O001O001O001O0O2O001O00001O001O001O001O001O3M3N2M4M2M3N2M4M2M3N2MN3M2N3M2N3I7G8HmE"}, "label": "a man on the beach reading his tablet"}]}
{"id": 156914, "image": "COCO_train2014_000000156914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow table\"."}], "task": "refering", "answer_template": "The \"the yellow table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 14, 15, 16, 17, 22, 23, 24, 25, 26, 38, 39, 40, 41, 42, 44, 45, 46, 47, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 87, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 135, 136, 137, 159, 160, 183], "bbox": [0.002984375, 0.004694835680751174, 1.0, 0.4879107981220657], "iscrowd": 0, "area": 27362.8309, "rle": {"size": [426, 640], "counts": "`l0[2b9]1000O2O001N10001N101O0O101O0O2O000O2O000O2O001N10001N101O0O101O0O2O000O2O001N10001N101O0O10000O10000O10000O10000O01000O10000O10000O10000O01000O10000O10000O10000O01000O10000O10000O10000O10000O10000O10000O10000O100O8I8H9F8I1N101O0O2O1N101N101O0O2O0O2O1O0O2O0O2O001N101N2O001N101N101O0O100O01000O10O010000O010O10000O010O1000O0100O10O010000O010O10000O010O1000O0100O10O10O10cj00g_f20eeXM1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N101O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O00001O001O002N2N2N1O2N2N2N2N1O2N2N2N23L7I6J1O0O10000O1000000O10000O10000O10000O10000O10000O10000O10000O0100O001O1O1O001O1O00100O001O1O001O1O001O100O1O1O1O1O1O1O1O100O001O1O1O1O1O1O1O100O1O1O1O1O001O1O1O100O1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O1O100O1O1O1O0N3N2M3M3M3N2M3M3M3M3N2M3M3M2O2M3M3M3N2M3MmNmGTMP8m2SHRMk7n2YHoLe7R3^HmL_7T3eHjLX7W3kHhLS7X3QIfLn6Y3UIfLQ7R3SIlLT7k2oHTMW7d2mHZMY7^2jH`M]7X2gHfM_7R2dHmMb7k1bHSNd7e1_HZNg7^1]H`Ni7X1ZHgNl7Q1XHmNn7k0UHSOR8OhHO[:OaI"}, "label": "the yellow table"}]}
{"id": 156914, "image": "COCO_train2014_000000156914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wood lunch table is beautifully set and people are enjoying sald and cheese quiche for the meal\"."}], "task": "refering", "answer_template": "The \"the wood lunch table is beautifully set and people are enjoying sald and cheese quiche for the meal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 14, 15, 16, 17, 22, 23, 24, 25, 26, 38, 39, 40, 41, 42, 44, 45, 46, 47, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 87, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 135, 136, 137, 159, 160, 183], "bbox": [0.002984375, 0.004694835680751174, 1.0, 0.4879107981220657], "iscrowd": 0, "area": 27362.8309, "rle": {"size": [426, 640], "counts": "`l0[2b9]1000O2O001N10001N101O0O101O0O2O000O2O000O2O001N10001N101O0O101O0O2O000O2O001N10001N101O0O10000O10000O10000O10000O01000O10000O10000O10000O01000O10000O10000O10000O01000O10000O10000O10000O10000O10000O10000O10000O100O8I8H9F8I1N101O0O2O1N101N101O0O2O0O2O1O0O2O0O2O001N101N2O001N101N101O0O100O01000O10O010000O010O10000O010O1000O0100O10O010000O010O10000O010O1000O0100O10O10O10cj00g_f20eeXM1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N101O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O00001O001O002N2N2N1O2N2N2N2N1O2N2N2N23L7I6J1O0O10000O1000000O10000O10000O10000O10000O10000O10000O10000O0100O001O1O1O001O1O00100O001O1O001O1O001O100O1O1O1O1O1O1O1O100O001O1O1O1O1O1O1O100O1O1O1O1O001O1O1O100O1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O1O100O1O1O1O0N3N2M3M3M3N2M3M3M3M3N2M3M3M2O2M3M3M3N2M3MmNmGTMP8m2SHRMk7n2YHoLe7R3^HmL_7T3eHjLX7W3kHhLS7X3QIfLn6Y3UIfLQ7R3SIlLT7k2oHTMW7d2mHZMY7^2jH`M]7X2gHfM_7R2dHmMb7k1bHSNd7e1_HZNg7^1]H`Ni7X1ZHgNl7Q1XHmNn7k0UHSOR8OhHO[:OaI"}, "label": "the wood lunch table is beautifully set and people are enjoying sald and cheese quiche for the meal"}]}
{"id": 156914, "image": "COCO_train2014_000000156914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"patterned placemat under the plate of food\"."}], "task": "refering", "answer_template": "The \"patterned placemat under the plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 92, 93, 94, 95, 115, 116, 117, 135, 136, 139, 159, 160, 161, 162, 180, 181, 182, 183, 184, 185, 203, 204, 206, 207, 208, 226, 227, 228, 229, 230, 231, 232, 249, 250, 251, 252, 253, 254, 255, 271, 272, 273, 274, 275, 276, 277, 278, 279, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.20178403755868543, 0.995453125, 0.9887793427230046], "iscrowd": 0, "area": 45333.24210000002, "rle": {"size": [426, 640], "counts": "X<m0c:j1WNi1_LbK\\LX6_3lIPLf6l3]IULc6f3aI[L_6`3eIaL[6[3hIfLX6U3lIlLT6S3mImLS6S3mIlLT6T3lIlLT6S3mImLS6S3mImLS6R3nInLR6R3nInLR6R3nInLR6Q3oInLR6R3nInLR6R3nInLR6Q3oIoLQ6Q3oIoLQ6Q3oIoLQ6P3PJPMP6P3PJoLQ6Q3oIoLQ6P3PJPMP6P3PJPMP6o2QJQMo5o2QJQMo5o2QJQMo5n2RJQMo5o2QJQMo5o2QJQMo5n2RJRMn5n2RJRMn5n2RJRMn5m2SJRMn5n2RJRMn5m2SJSMm5m2SJSMm5m2RJTMn5k2SJUMm5k2SJUMm5k2SJTMn5k2SJUM7SOS4h3fKUM2\\OT4_3jKUMMFT4T3PLVMG0T4j2ULVMB9U4a2YLVM]Oc0U4V2_LVMXOn0T4l1dLVMSOW1U4b1iLWMlNb1V4W1nL5m2KRM:j2EWM`0d2@\\Md0`2\\O`Mh0\\2WOeMk0Y2UOgMm0W2SOiMn0V2QOkMQ1S2oNmMS1Q2mNoMT1P2kNQNW1m1iNSNY1k1fNUN]1i1cNVN_1i1aNWNa1g1^NYNe1e1[NZNg1e1YN[Ni1c1VN]Nm1a1SN^NP2`1PN`NQ2_1nMaNU2]1kMbNX2\\1hMdNY2[1fMdN^2Z1bMeNa2Y1^MgNe2W1[MhNg2W1YMhNj2V1UMjNn2T1RMjNQ3U1oLjNT3T1kLlNX3R1hLmN[3Q1eLnN]3Q1bLoN`3P1`LnNc3Q1\\LoNf3P1ZLoNh3P1XLoNi3Q1VLnNm3Q1SLnNo3Q1QLnNQ4Q1nKnNU4Q1kKnNW4Q1iKnNY4Q1fKnN]4Q1cKnN_4Q1aKmNb4R1]KnNe4Q1[KnNg4Q1XKnNk4Q1UKnNm4Q1SKnNo4Q1PKnNR5R1nJmNT5R1lJlNW5S1hJmNZ5R1fJmN\\5R1dJnN]5Q1bJoN`5P1`JPOa5o0_JPOc5o0\\JROe5m0[JROg5m0XJSOj5l0VJTOk5k0UJTOm5k0RJVOo5i0QJVOQ6i0oIVOR6j0mIWOT6h0lIWOV6h0jIXOW6g0hIYOZ6f0fIZO[6e0eIZO]6e0bI[O`6d0`I\\Oa6c0^I]Oc6c0]I]Oc6c0]I\\Oe6c0ZI]Og6c0YI]Og6c0YI\\Oi6c0VI^Oj6b0VI]Ok6c0UI]Ol6b0SI^On6b0m201O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000`Vc20`i\\M000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3K5K5J6bNTKYJQ5g5YKjIl4W6]KYIi4g6m01O001O001O001O001O001O00]MkIRNS6P2nInMR6S2oIlMP6U2QJiMo5X2RJgMm5Z2TJdMl5]2UJbMj5_2WJ_Mi5b2XJ]Mg5d2ZJZMf5f2\\JYMc5h2^JVMb5k2_JTM`5m2aJQM_5P3aJPM^5Q3cJmL]5T3dJQMU5P3lJYMi4h2WKbM^4_2cKiMS4X2nKhMP4Y2QLeMo3\\2QLdMn3]2SLaMm3`2TL_Mk3a2WL]Mi3d2XL[Mg3f2YLZMf3g2[LWMe3j2\\LUMc3l2]LSMc3n2]LRMb3o2^LPMb3Q3]LPMb3Q3^LnLb3S3^LmLa3T3_LlL`3U3_LjLb3W3^LfLd3[3\\LbLf3_3ZL_Lg3a3YL]Li3d3WLZLj3g3VLVLl3k3TLSLm3n3RLPLP4Q4PLmKQ4T4oKiKS4X4mKeKU4\\4jKcKW4^4kK]KW4d4QLRKP4o4WL\\JV4e5g11O1O001O1O1O1O1O1O1O1O1O1O1O1Og0YO]2cM^2bMPl0"}, "label": "patterned placemat under the plate of food"}]}
{"id": 156914, "image": "COCO_train2014_000000156914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a placemat with flowers on it\"."}], "task": "refering", "answer_template": "The \"a placemat with flowers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 92, 93, 94, 95, 115, 116, 117, 135, 136, 139, 159, 160, 161, 162, 180, 181, 182, 183, 184, 185, 203, 204, 206, 207, 208, 226, 227, 228, 229, 230, 231, 232, 249, 250, 251, 252, 253, 254, 255, 271, 272, 273, 274, 275, 276, 277, 278, 279, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.20178403755868543, 0.995453125, 0.9887793427230046], "iscrowd": 0, "area": 45333.24210000002, "rle": {"size": [426, 640], "counts": "X<m0c:j1WNi1_LbK\\LX6_3lIPLf6l3]IULc6f3aI[L_6`3eIaL[6[3hIfLX6U3lIlLT6S3mImLS6S3mIlLT6T3lIlLT6S3mImLS6S3mImLS6R3nInLR6R3nInLR6R3nInLR6Q3oInLR6R3nInLR6R3nInLR6Q3oIoLQ6Q3oIoLQ6Q3oIoLQ6P3PJPMP6P3PJoLQ6Q3oIoLQ6P3PJPMP6P3PJPMP6o2QJQMo5o2QJQMo5o2QJQMo5n2RJQMo5o2QJQMo5o2QJQMo5n2RJRMn5n2RJRMn5n2RJRMn5m2SJRMn5n2RJRMn5m2SJSMm5m2SJSMm5m2RJTMn5k2SJUMm5k2SJUMm5k2SJTMn5k2SJUM7SOS4h3fKUM2\\OT4_3jKUMMFT4T3PLVMG0T4j2ULVMB9U4a2YLVM]Oc0U4V2_LVMXOn0T4l1dLVMSOW1U4b1iLWMlNb1V4W1nL5m2KRM:j2EWM`0d2@\\Md0`2\\O`Mh0\\2WOeMk0Y2UOgMm0W2SOiMn0V2QOkMQ1S2oNmMS1Q2mNoMT1P2kNQNW1m1iNSNY1k1fNUN]1i1cNVN_1i1aNWNa1g1^NYNe1e1[NZNg1e1YN[Ni1c1VN]Nm1a1SN^NP2`1PN`NQ2_1nMaNU2]1kMbNX2\\1hMdNY2[1fMdN^2Z1bMeNa2Y1^MgNe2W1[MhNg2W1YMhNj2V1UMjNn2T1RMjNQ3U1oLjNT3T1kLlNX3R1hLmN[3Q1eLnN]3Q1bLoN`3P1`LnNc3Q1\\LoNf3P1ZLoNh3P1XLoNi3Q1VLnNm3Q1SLnNo3Q1QLnNQ4Q1nKnNU4Q1kKnNW4Q1iKnNY4Q1fKnN]4Q1cKnN_4Q1aKmNb4R1]KnNe4Q1[KnNg4Q1XKnNk4Q1UKnNm4Q1SKnNo4Q1PKnNR5R1nJmNT5R1lJlNW5S1hJmNZ5R1fJmN\\5R1dJnN]5Q1bJoN`5P1`JPOa5o0_JPOc5o0\\JROe5m0[JROg5m0XJSOj5l0VJTOk5k0UJTOm5k0RJVOo5i0QJVOQ6i0oIVOR6j0mIWOT6h0lIWOV6h0jIXOW6g0hIYOZ6f0fIZO[6e0eIZO]6e0bI[O`6d0`I\\Oa6c0^I]Oc6c0]I]Oc6c0]I\\Oe6c0ZI]Og6c0YI]Og6c0YI\\Oi6c0VI^Oj6b0VI]Ok6c0UI]Ol6b0SI^On6b0m201O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000`Vc20`i\\M000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3K5K5J6bNTKYJQ5g5YKjIl4W6]KYIi4g6m01O001O001O001O001O001O00]MkIRNS6P2nInMR6S2oIlMP6U2QJiMo5X2RJgMm5Z2TJdMl5]2UJbMj5_2WJ_Mi5b2XJ]Mg5d2ZJZMf5f2\\JYMc5h2^JVMb5k2_JTM`5m2aJQM_5P3aJPM^5Q3cJmL]5T3dJQMU5P3lJYMi4h2WKbM^4_2cKiMS4X2nKhMP4Y2QLeMo3\\2QLdMn3]2SLaMm3`2TL_Mk3a2WL]Mi3d2XL[Mg3f2YLZMf3g2[LWMe3j2\\LUMc3l2]LSMc3n2]LRMb3o2^LPMb3Q3]LPMb3Q3^LnLb3S3^LmLa3T3_LlL`3U3_LjLb3W3^LfLd3[3\\LbLf3_3ZL_Lg3a3YL]Li3d3WLZLj3g3VLVLl3k3TLSLm3n3RLPLP4Q4PLmKQ4T4oKiKS4X4mKeKU4\\4jKcKW4^4kK]KW4d4QLRKP4o4WL\\JV4e5g11O1O001O1O1O1O1O1O1O1O1O1O1O1Og0YO]2cM^2bMPl0"}, "label": "a placemat with flowers on it"}]}
{"id": 500982, "image": "COCO_train2014_000000500982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing a scarf\"."}], "task": "refering", "answer_template": "The \"the woman wearing a scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 130, 131, 132, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 183, 184, 185, 186, 187, 188, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223], "bbox": [0.15878, 0.46058666666666664, 0.45608000000000004, 0.9876], "iscrowd": 0, "area": 16914.50195, "rle": {"size": [375, 500], "counts": "bXm0i0l:2M3M3M3M3M3M3M3M3N2_Oa0_Oa0^Ob0G9O1O1O100O1O1O1O1O1O100O1O1O1O1OkN[HZNd7f1_HXNa7g1aHXN^7h1dHWN[7h1hHWNV7j1lHUNR7l1PISNn6n1TIQNj6P2XIoMf6R2\\ImMa6U2aIjM]6W2eIhMY6Y2iIfMb5\\ORJo2>dM^5BoIk2e0bMZ5HlIg2l0`MV5MjIe2Q1]MT5W3lJiLR5Z3mJfLQ5\\3oJdLo4^3QKbLn4`3QK`Lm4b3SK^Lk4e3TK[Lk4f3UKZLi4i3VKWLj4i3VKWLi4k3VKULj4k3VKULj4l3TKULk4l3UKTLk4m3TKSLk4n3UKRLk4o3TKQLk4P4UKPLk4P4UKPLk4Q4TKoKk4R4UKmKl4T4SKlKl4U4TKkKl4U4TKkKk4U4VKkKj4U4VKkKj4T4WKlKi4T4WKlKi4S4WKnKi4R4WKnKi4Q4XKoKh4Q4WKPLi4o3XKQLh4o3XKQLh4n3YKRLg4n3XKSLh4l3YKTLg4k3ZKULf4k3ZKULf4j3ZKWLf4i3ZKWLf4h3[KXLe4h3ZKYLf4f3[KZLe4f3[KZLe4e3\\K[Ld4e3[K\\Le4c3\\K]Ld4c3\\K]Le4b3[K^Le4a3[K`Le4`3[K`Le4`3[K`Lf4_3YKbLi4[3XKeLi4Z3WKfLj4Y3VKgLl4W3SKjLn4LcJQ2`0SNn4HfJT2<TNP5DhJV27WNR5_OkJY23`MXO=k5GmJ[2<mMi4CPK]29oMh4ARK_27oMo5P2RJoMo5P2RJoMP6n1RJQNo5n1RJQNo5n1RJQNo5n1RJQNP6l1RJSNo5l1RJSNo5l1RJSNP6k1QJTNP6j1RJUNo5j1RJUNo5j1RJUNP6i1QJVNP6i1Q2M2O2N2N2M2O2N2N2M3N1O2N2M3N1O2N2M3N2N1O3M9FZZS3"}, "label": "the woman wearing a scarf"}]}
{"id": 500982, "image": "COCO_train2014_000000500982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady with grey sweater\"."}], "task": "refering", "answer_template": "The \"the lady with grey sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 130, 131, 132, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 183, 184, 185, 186, 187, 188, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223], "bbox": [0.15878, 0.46058666666666664, 0.45608000000000004, 0.9876], "iscrowd": 0, "area": 16914.50195, "rle": {"size": [375, 500], "counts": "bXm0i0l:2M3M3M3M3M3M3M3M3N2_Oa0_Oa0^Ob0G9O1O1O100O1O1O1O1O1O100O1O1O1O1OkN[HZNd7f1_HXNa7g1aHXN^7h1dHWN[7h1hHWNV7j1lHUNR7l1PISNn6n1TIQNj6P2XIoMf6R2\\ImMa6U2aIjM]6W2eIhMY6Y2iIfMb5\\ORJo2>dM^5BoIk2e0bMZ5HlIg2l0`MV5MjIe2Q1]MT5W3lJiLR5Z3mJfLQ5\\3oJdLo4^3QKbLn4`3QK`Lm4b3SK^Lk4e3TK[Lk4f3UKZLi4i3VKWLj4i3VKWLi4k3VKULj4k3VKULj4l3TKULk4l3UKTLk4m3TKSLk4n3UKRLk4o3TKQLk4P4UKPLk4P4UKPLk4Q4TKoKk4R4UKmKl4T4SKlKl4U4TKkKl4U4TKkKk4U4VKkKj4U4VKkKj4T4WKlKi4T4WKlKi4S4WKnKi4R4WKnKi4Q4XKoKh4Q4WKPLi4o3XKQLh4o3XKQLh4n3YKRLg4n3XKSLh4l3YKTLg4k3ZKULf4k3ZKULf4j3ZKWLf4i3ZKWLf4h3[KXLe4h3ZKYLf4f3[KZLe4f3[KZLe4e3\\K[Ld4e3[K\\Le4c3\\K]Ld4c3\\K]Le4b3[K^Le4a3[K`Le4`3[K`Le4`3[K`Lf4_3YKbLi4[3XKeLi4Z3WKfLj4Y3VKgLl4W3SKjLn4LcJQ2`0SNn4HfJT2<TNP5DhJV27WNR5_OkJY23`MXO=k5GmJ[2<mMi4CPK]29oMh4ARK_27oMo5P2RJoMo5P2RJoMP6n1RJQNo5n1RJQNo5n1RJQNo5n1RJQNP6l1RJSNo5l1RJSNo5l1RJSNP6k1QJTNP6j1RJUNo5j1RJUNo5j1RJUNP6i1QJVNP6i1Q2M2O2N2N2M2O2N2N2M3N1O2N2M3N1O2N2M3N2N1O3M9FZZS3"}, "label": "the lady with grey sweater"}]}
{"id": 500982, "image": "COCO_train2014_000000500982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting with green shirt\"."}], "task": "refering", "answer_template": "The \"man sitting with green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 99, 100, 116, 117, 118, 119, 134, 135, 136, 137, 138, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 189, 190, 191, 192, 207, 208, 209, 226, 227], "bbox": [0.45, 0.34381333333333336, 0.7028, 0.9842666666666667], "iscrowd": 0, "area": 15942.2744, "rle": {"size": [375, 500], "counts": "cdb23a;3L5K5L4K4M4K5K4M5J6J6K5J6J5M3N2N101N2O0O2O1N1O2O0O2O1N101N2fNVMdIk2Z6YMdIh2Z6ZMdIh2Z6ZMeIg2Y6[MeIg2X6]MeIg2W6\\MgIh2T6]MgIg2U6^MgIf2S6aMgIc2U6bMgIb2U6bMfIb2V6d1K4M4K1O2N2O0O2N101O1O001O001O001O1O001O0010O10000O100XKcJT3^5hLeJY3[5bLiJ^3X5^LkJc3U5YLnJh3Q5TLSKl3P5nKSKS4M\\KY4;nKY4BdKb4MoK`4XOiKl4@PL`5S51O001O1O1OO1M3M3M3M3N3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M4L4L4eNZG_Oj8=ZG_Oj8;[GBi87^GFe83aGJc80bGNb8LcG0a8KcG3`8GfG6]8EgG8o9M2N3LUYf1"}, "label": "man sitting with green shirt"}]}
{"id": 500982, "image": "COCO_train2014_000000500982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man sitting on a wooden chair with a green t - shirt on and khaki pants\"."}], "task": "refering", "answer_template": "The \"an older man sitting on a wooden chair with a green t - shirt on and khaki pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 99, 100, 116, 117, 118, 119, 134, 135, 136, 137, 138, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 189, 190, 191, 192, 207, 208, 209, 226, 227], "bbox": [0.45, 0.34381333333333336, 0.7028, 0.9842666666666667], "iscrowd": 0, "area": 15942.2744, "rle": {"size": [375, 500], "counts": "cdb23a;3L5K5L4K4M4K5K4M5J6J6K5J6J5M3N2N101N2O0O2O1N1O2O0O2O1N101N2fNVMdIk2Z6YMdIh2Z6ZMdIh2Z6ZMeIg2Y6[MeIg2X6]MeIg2W6\\MgIh2T6]MgIg2U6^MgIf2S6aMgIc2U6bMgIb2U6bMfIb2V6d1K4M4K1O2N2O0O2N101O1O001O001O001O1O001O0010O10000O100XKcJT3^5hLeJY3[5bLiJ^3X5^LkJc3U5YLnJh3Q5TLSKl3P5nKSKS4M\\KY4;nKY4BdKb4MoK`4XOiKl4@PL`5S51O001O1O1OO1M3M3M3M3N3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M4L4L4eNZG_Oj8=ZG_Oj8;[GBi87^GFe83aGJc80bGNb8LcG0a8KcG3`8GfG6]8EgG8o9M2N3LUYf1"}, "label": "an older man sitting on a wooden chair with a green t - shirt on and khaki pants"}]}
{"id": 107767, "image": "COCO_train2014_000000107767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bag in the right side of the pink bag\"."}], "task": "refering", "answer_template": "The \"bag in the right side of the pink bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.44000000000000006, 0.7444166666666666, 0.7138125000000001, 0.9346666666666666], "iscrowd": 0, "area": 10666.302699999998, "rle": {"size": [480, 640], "counts": "cbT46j>6I7I8I6I7I7J1N10000O10000O2O00000O10000O101OO10O0100O010O100O010O10O1000000O01000O1000000O10000O1000000O1000O010000O1000000O10000O2O0O10000O100O10000O100O100O10000A?001O0O2O001O00001O001O1O0O2O1O000000O10000O10O010000O100O10000O010O100O10O10O100O10O100000001N101O001O0O2O00001O0O2O001O001O0O101O001O001O001N101O1O1O2N1O2N1O2M2O2N1O2N1O1O2N1N3M2B?@Q]e2"}, "label": "bag in the right side of the pink bag"}]}
{"id": 107767, "image": "COCO_train2014_000000107767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and black bag with silver strap , next to pink one\"."}], "task": "refering", "answer_template": "The \"blue and black bag with silver strap , next to pink one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.44000000000000006, 0.7444166666666666, 0.7138125000000001, 0.9346666666666666], "iscrowd": 0, "area": 10666.302699999998, "rle": {"size": [480, 640], "counts": "cbT46j>6I7I8I6I7I7J1N10000O10000O2O00000O10000O101OO10O0100O010O100O010O10O1000000O01000O1000000O10000O1000000O1000O010000O1000000O10000O2O0O10000O100O10000O100O100O10000A?001O0O2O001O00001O001O1O0O2O1O000000O10000O10O010000O100O10000O010O100O10O10O100O10O100000001N101O001O0O2O00001O0O2O001O001O0O101O001O001O001N101O1O1O2N1O2N1O2M2O2N1O2N1O1O2N1N3M2B?@Q]e2"}, "label": "blue and black bag with silver strap , next to pink one"}]}
{"id": 214265, "image": "COCO_train2014_000000214265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white man stands in the background while obama fixes a tie\"."}], "task": "refering", "answer_template": "The \"a white man stands in the background while obama fixes a tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 196, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 311, 312, 313, 314], "bbox": [0.46867187499999996, 0.38646484375, 0.726734375, 0.77314453125], "iscrowd": 0, "area": 18239.30865, "rle": {"size": [512, 640], "counts": "V[f42k?3L5L3M3L5L3O1N3O0O1O2O0O1O2O00001O000O2O000O2O000O2O0O101O0O101O0O101O0O2O0O2O000O2O0O2O0O2O0O2O0O101N101O0O2O0O101N1VMjNjFW1S9WOaFi0\\97gEJV:<fEDW:c0dE^OZ:f0dE[OZ:g0eEYO[:h0dEYO[:h0dEXO[:j0dEWO[:j0cEWO]:j0bEWO\\:W30001N101O0O1000001N10000O10001N10000O101O0O10000O101O0O2O1OO100O1O100O00100O1O100O100O1O1M3M3M3M2N3M3M3M3M3M3N2N2M3N4K4M4L3M3M4M2M4M2M4L3N3L3N3L3N3L3M4M2M3N3L3M4M2M4M3L4M3L4I7H8I7J7H<DRXg2"}, "label": "a white man stands in the background while obama fixes a tie"}]}
{"id": 214265, "image": "COCO_train2014_000000214265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a striped shirt with a red and white polka dot tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a striped shirt with a red and white polka dot tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 196, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 311, 312, 313, 314], "bbox": [0.46867187499999996, 0.38646484375, 0.726734375, 0.77314453125], "iscrowd": 0, "area": 18239.30865, "rle": {"size": [512, 640], "counts": "V[f42k?3L5L3M3L5L3O1N3O0O1O2O0O1O2O00001O000O2O000O2O000O2O0O101O0O101O0O101O0O2O0O2O000O2O0O2O0O2O0O2O0O101N101O0O2O0O101N1VMjNjFW1S9WOaFi0\\97gEJV:<fEDW:c0dE^OZ:f0dE[OZ:g0eEYO[:h0dEYO[:h0dEXO[:j0dEWO[:j0cEWO]:j0bEWO\\:W30001N101O0O1000001N10000O10001N10000O101O0O10000O101O0O2O1OO100O1O100O00100O1O100O100O1O1M3M3M3M2N3M3M3M3M3M3N2N2M3N4K4M4L3M3M4M2M4M2M4L3N3L3N3L3N3L3M4M2M3N3L3M4M2M4M3L4M3L4I7H8I7J7H<DRXg2"}, "label": "a man wearing a striped shirt with a red and white polka dot tie"}]}
{"id": 214265, "image": "COCO_train2014_000000214265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man having his tie fixed by the president of the united states\"."}], "task": "refering", "answer_template": "The \"the man having his tie fixed by the president of the united states\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400], "bbox": [0.0017968749999999999, 0.121015625, 0.44564062499999996, 0.9867968749999999], "iscrowd": 0, "area": 83283.24314999998, "rle": {"size": [512, 640], "counts": "\\m05c?>B>B>]OcNnAl1i==G9H7H9G8H8H8H8H5K3M3M3M3M3M2N3M3M3M3M3M3M3M3M3M3M3N2O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N3N1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2O1O1N2O2M2O1N2O1N2O1]M[GiMg8T2`GgMa8V2eGfM\\8X2jGbMX8[2oG`MR8]2TH^Mn7`2XH[Mi7b2]HYMe7d2bHWM_7g2fHTM\\7i2kHRMV7k2QIoLQ7o2TImLm6P3ZIjLh6S3^IiLd6T3bIgL_6V3W3M3N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O100000000000000000000000000000000000000000000000000000000000000O1000000000001O000000000000000000000000000000[HmM]1S2bNQN[1P2dNTNX1l1gNXNV1h1jN[NS1f1lN]NQ1c1nN_NQ1b1nN^NR1b1mN`NR1a1mN_NS1a1lNaNS1_1mNaNS1`1kNbNT1^1lNbNT1_1jNcNU1]1kNcNU1^1iNdNV1\\1jNdNV1\\1jNeNU1\\1iNeNW1[1iNeNW1\\1gNfNX1Z1hNfNlKQOh3Z2:gNjKWOg3R2?gNhK]Oe3l1b0iNfKAd3g1e0hNeKHa3`1i0jNcKL`3[1l0iNbK1_3V1n0kNZK;d3k0Q1jNSKf0i3`0T1kNiJR1o33W1b1fN_NY1d1cN^N[1g1aNZN^1i1_NXN_1l1^NTNb1n1\\NSNb1o1]NRNb1o1\\NSNb1n1^NSNa1n1^NSN`1o1_NRN`1o1_NQN`1Q2^NQN`1Q2_NPN_1R2`NoM^1S2aNnM^1S2aNnM]1T2bNmM\\1U2bNmM\\1U2cNkM]1V2bNkM\\1W2cNjM[1X2dNiMc0jLSM_5X2iMc0oLnLZ5^2hMc0TMiLU5c2hMb0Q3]OoLb0S3]OnLb0S3]OnLa0T3]OnLD_LeNd6<^IN`3o0`LeNd69bI1[3o0`LfNe65fI4V3P1`LfNe61kI7Q3P1aLfNe6NnI:n2Q1`LfNe6JSJ=P3j0ZLmNd6IWJ>R3e0TLROe6FZJb0T3>oKYOd6B_Je0V38hK_OY86YL1`KHY85bLHVK1T?N2O1N2O1N2N2OQZa5"}, "label": "the man having his tie fixed by the president of the united states"}]}
{"id": 214265, "image": "COCO_train2014_000000214265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man fixes another man ' s tie\"."}], "task": "refering", "answer_template": "The \"a man fixes another man ' s tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 205, 215, 223, 224, 225, 226, 227, 228, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413], "bbox": [0.315671875, 0.11052734375, 1.0, 1.0], "iscrowd": 0, "area": 105186.84455000002, "rle": {"size": [512, 640], "counts": "VYU3:d?3O001O1O1O1O001O1O1000O0100O1D^OYAb0f>BWA=i>FTA;^1\\Ok;;fB99K9Ee<o0QCE0@l<l0RCHN@l<i0UCLJ_On<f0WC`0f<AWCd0f<]OXCg0e<[1N10000000001N11O00001O01O01O00001O00010O00001O01O100O00100O001\\EdLc8\\3YGhLf8Y3VGlLh8T3UGPMj8Q3RGSMn8m2nFXMP9P3eFTM\\9Q3ZFTMi9n2mEVMW:l2_EYMd:P44L4M3L4M2M100O1O1c0]ON2N2N2N2OO01O001O001O001O001O001O001O010O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O00000000000000O10000000000O1000000O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2N2fIZKX1h4fNZKX1h4gNZKV1h4hN[KV1f4iN]KS1e4lN_Ko0c4oNaKm0a4ROfKf0\\4XOoK=S4B^LLd33`LHb36dLD^3;fLA[3=jL^OX3a0mLYOU3f0oLUOS3j0oLSOS3l0oLQOS3m0oLQOS3n0oLoNS3P1oLmNS3P1PMoNQ3n0SMoNo2n0TMPOn2a0bM\\O`2b0bM\\O`2c0bM[O_2c0dMZO^2d0eMYO]2e0fMXO\\2f0fMYO[2f0gMWO[2g0hMVOZ2h0iMUOY2i0jMUOW2j0kMSOW2k0lMROV2m0lMPOV2o0lMnNV2P1mMmNU2R1mMkNU2T1mMiNU2U1nMhNT2W1mMgNU2X1mMeNU2Y1nMdNT2[1nMbNT2]1nM`NT2^1nM`NT2`1mM]NU2b1mM[NU2d1mMYNU2f1mMWNU2i1RNmMQ2R2o5O100O1O1O100O1O100O1O1O100000000000000000000000000000000000000O100000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000000001O00000000001O1aFXNd4i1RK`Nn4a1oJbNP5^1oJcNQ5^1lJeNS5\\1kJeNU5\\1iJfNV5[1gJhNX5Y1fJiNY5X1dJlNZ5T1eJnNZ5S1dJoN[5R1cJPO\\5Q1bJRO\\5o0bJSO]5n0aJTO^5l0`JYO]5h0aJ\\O\\5e0bJ^O\\5c0bJA[5`0cJDZ5>cJFZ5<cJHY5;cJJZ58bJK]57_JL`56\\JNb54ZJOe53WJ1g52TJ1k51QJ2n50mI5Q6MkI6T6LhI8V6JfI9Y6IcI:\\6HVIf0h6\\OfHU1Y7k3N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N4L3M4L4L3M4L4L4L3M4L4L3M4L4L4L3MkD"}, "label": "a man fixes another man ' s tie"}]}
{"id": 214265, "image": "COCO_train2014_000000214265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"barrack obama adjusting a person ' s tie\"."}], "task": "refering", "answer_template": "The \"barrack obama adjusting a person ' s tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 205, 215, 223, 224, 225, 226, 227, 228, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413], "bbox": [0.315671875, 0.11052734375, 1.0, 1.0], "iscrowd": 0, "area": 105186.84455000002, "rle": {"size": [512, 640], "counts": "VYU3:d?3O001O1O1O1O001O1O1000O0100O1D^OYAb0f>BWA=i>FTA;^1\\Ok;;fB99K9Ee<o0QCE0@l<l0RCHN@l<i0UCLJ_On<f0WC`0f<AWCd0f<]OXCg0e<[1N10000000001N11O00001O01O01O00001O00010O00001O01O100O00100O001\\EdLc8\\3YGhLf8Y3VGlLh8T3UGPMj8Q3RGSMn8m2nFXMP9P3eFTM\\9Q3ZFTMi9n2mEVMW:l2_EYMd:P44L4M3L4M2M100O1O1c0]ON2N2N2N2OO01O001O001O001O001O001O001O010O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O00000000000000O10000000000O1000000O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2N2fIZKX1h4fNZKX1h4gNZKV1h4hN[KV1f4iN]KS1e4lN_Ko0c4oNaKm0a4ROfKf0\\4XOoK=S4B^LLd33`LHb36dLD^3;fLA[3=jL^OX3a0mLYOU3f0oLUOS3j0oLSOS3l0oLQOS3m0oLQOS3n0oLoNS3P1oLmNS3P1PMoNQ3n0SMoNo2n0TMPOn2a0bM\\O`2b0bM\\O`2c0bM[O_2c0dMZO^2d0eMYO]2e0fMXO\\2f0fMYO[2f0gMWO[2g0hMVOZ2h0iMUOY2i0jMUOW2j0kMSOW2k0lMROV2m0lMPOV2o0lMnNV2P1mMmNU2R1mMkNU2T1mMiNU2U1nMhNT2W1mMgNU2X1mMeNU2Y1nMdNT2[1nMbNT2]1nM`NT2^1nM`NT2`1mM]NU2b1mM[NU2d1mMYNU2f1mMWNU2i1RNmMQ2R2o5O100O1O1O100O1O100O1O1O100000000000000000000000000000000000000O100000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000000001O00000000001O1aFXNd4i1RK`Nn4a1oJbNP5^1oJcNQ5^1lJeNS5\\1kJeNU5\\1iJfNV5[1gJhNX5Y1fJiNY5X1dJlNZ5T1eJnNZ5S1dJoN[5R1cJPO\\5Q1bJRO\\5o0bJSO]5n0aJTO^5l0`JYO]5h0aJ\\O\\5e0bJ^O\\5c0bJA[5`0cJDZ5>cJFZ5<cJHY5;cJJZ58bJK]57_JL`56\\JNb54ZJOe53WJ1g52TJ1k51QJ2n50mI5Q6MkI6T6LhI8V6JfI9Y6IcI:\\6HVIf0h6\\OfHU1Y7k3N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N4L3M4L4L3M4L4L4L3M4L4L3M4L4L4L3MkD"}, "label": "barrack obama adjusting a person ' s tie"}]}
{"id": 263420, "image": "COCO_train2014_000000263420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird that has orange spots on its wing\"."}], "task": "refering", "answer_template": "The \"the bird that has orange spots on its wing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.35484375, 0.2516041666666667, 0.961296875, 0.98925], "iscrowd": 0, "area": 69659.30059999999, "rle": {"size": [480, 640], "counts": "ZbZ32l>:F9H6]BWOd<k0YCYOd<j0XC[Od<g0YC^Oc<d0YCAd<a0YCDc<>ZCGc<]1M4L4L2N2N2N2N2N2N2N3N1N2N2N2N2N2N2N3M2N2N2O1N2N2N2O2N0O2O00001N1000A\\EbLd:^3_E`L_:a3dE\\L\\:e3fEXLZ:h3?0O2O00001N10001O0O101O000O2O00000O1000000O1000000O1000O10O1000001N3N2N1O2M2O2N1O2M2O2N1O2M2O1O2N1O1O1O1O1O2N1O1O1O2N6J6J6J6J6J6J6J6J6K5J`2`M2N2N2N2N2N2N3M6J3M2N2N2N2N2N2N1O2N000000000000000001O00000000000000001O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M2N3M3M4L5K5K2N3M2N2N2M3N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1N3N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O2N1O2N2N1O2N1O2N1O2N2NYX;"}, "label": "the bird that has orange spots on its wing"}]}
{"id": 263420, "image": "COCO_train2014_000000263420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green and red bird that is looking down\"."}], "task": "refering", "answer_template": "The \"the green and red bird that is looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.35484375, 0.2516041666666667, 0.961296875, 0.98925], "iscrowd": 0, "area": 69659.30059999999, "rle": {"size": [480, 640], "counts": "ZbZ32l>:F9H6]BWOd<k0YCYOd<j0XC[Od<g0YC^Oc<d0YCAd<a0YCDc<>ZCGc<]1M4L4L2N2N2N2N2N2N2N3N1N2N2N2N2N2N2N3M2N2N2O1N2N2N2O2N0O2O00001N1000A\\EbLd:^3_E`L_:a3dE\\L\\:e3fEXLZ:h3?0O2O00001N10001O0O101O000O2O00000O1000000O1000000O1000O10O1000001N3N2N1O2M2O2N1O2M2O2N1O2M2O1O2N1O1O1O1O1O2N1O1O1O2N6J6J6J6J6J6J6J6J6K5J`2`M2N2N2N2N2N2N3M6J3M2N2N2N2N2N2N1O2N000000000000000001O00000000000000001O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M2N3M3M4L5K5K2N3M2N2N2M3N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1N3N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O2N1O2N2N1O2N1O2N1O2N2NYX;"}, "label": "the green and red bird that is looking down"}]}
{"id": 263420, "image": "COCO_train2014_000000263420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"love bird with his back to the camera\"."}], "task": "refering", "answer_template": "The \"love bird with his back to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.08351562500000001, 0.252875, 0.55246875, 0.9873541666666666], "iscrowd": 0, "area": 70321.06964999998, "rle": {"size": [480, 640], "counts": "l`i05g>6J5K5K5^NEYD`0e;BWDb0g;ASDe0k;\\ORDg0l;[ORDg0l;\\OQDf0n;[OoCh0o;[OnCg0Q<ZOmCi0P<YOmCj0Q<T1O1N2N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2fHTL^3m3]LcLV3_3eLRMm2P3nL`Mf2b2TMnM_2a1dJhLh2X2V2Q1YKaL\\2n2o1`0bN1P10lN5R1LjN8V1HfN<Y1EcN?\\1B`Nb0`1^O\\Nf0c1[OYNi0f1WOWNm0i1SOSNQ1l1PORNR1n1nNQNS1o1mNPNT1P2lNoMU1Q2kNnMV1R2jNmMW1S2iNlMX1T2gNmMY1R2hNmMY1S2gNlMZ1T2fNkM[1U2eNjM\\1V2dNiM]1W2dNgM]1Y2gNbMZ1^2jN]MW1b2nNYMS1g2QOTMP1l2UOnLl0R3XOiLi0W3[OdLf0\\3^O_Lc0a3AZL`0f3@ZL`0e3BYL?g3BWL?i3AVL`0j3ATL`0l3ARL`0n3@QLa0o3X500O100000000000000O11O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O001O001O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1hGkIc7W6ZHmIc7T6ZHPJd7Q6ZHRJd7P6YHTJd7m5YHWJe7j5YHYJe7d6M2N2N3M2N2N2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O00001O0000001O0000DfISHY6m7iIQHW6P8kImGU6T8lIkGS6U8PJhGP6Y8;1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O1O001O1O1O001O1O1O001O1O1O00001O00001O00001O00001O00001O0000001O5K9G:F9G7I4L4L4L4L4L3M4L4ZMPET1X=VO2N2N1O2N2N2M3J5J7I7IPdU4"}, "label": "love bird with his back to the camera"}]}
{"id": 263420, "image": "COCO_train2014_000000263420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird on the left\"."}], "task": "refering", "answer_template": "The \"the bird on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.08351562500000001, 0.252875, 0.55246875, 0.9873541666666666], "iscrowd": 0, "area": 70321.06964999998, "rle": {"size": [480, 640], "counts": "l`i05g>6J5K5K5^NEYD`0e;BWDb0g;ASDe0k;\\ORDg0l;[ORDg0l;\\OQDf0n;[OoCh0o;[OnCg0Q<ZOmCi0P<YOmCj0Q<T1O1N2N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2fHTL^3m3]LcLV3_3eLRMm2P3nL`Mf2b2TMnM_2a1dJhLh2X2V2Q1YKaL\\2n2o1`0bN1P10lN5R1LjN8V1HfN<Y1EcN?\\1B`Nb0`1^O\\Nf0c1[OYNi0f1WOWNm0i1SOSNQ1l1PORNR1n1nNQNS1o1mNPNT1P2lNoMU1Q2kNnMV1R2jNmMW1S2iNlMX1T2gNmMY1R2hNmMY1S2gNlMZ1T2fNkM[1U2eNjM\\1V2dNiM]1W2dNgM]1Y2gNbMZ1^2jN]MW1b2nNYMS1g2QOTMP1l2UOnLl0R3XOiLi0W3[OdLf0\\3^O_Lc0a3AZL`0f3@ZL`0e3BYL?g3BWL?i3AVL`0j3ATL`0l3ARL`0n3@QLa0o3X500O100000000000000O11O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O001O001O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1hGkIc7W6ZHmIc7T6ZHPJd7Q6ZHRJd7P6YHTJd7m5YHWJe7j5YHYJe7d6M2N2N3M2N2N2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O00001O0000001O0000DfISHY6m7iIQHW6P8kImGU6T8lIkGS6U8PJhGP6Y8;1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O1O001O1O1O001O1O1O001O1O1O00001O00001O00001O00001O00001O0000001O5K9G:F9G7I4L4L4L4L4L3M4L4ZMPET1X=VO2N2N1O2N2N2M3J5J7I7IPdU4"}, "label": "the bird on the left"}]}
{"id": 296191, "image": "COCO_train2014_000000296191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe further from the fence\"."}], "task": "refering", "answer_template": "The \"the giraffe further from the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 104, 124, 125, 126, 146, 147, 148, 149, 169, 170, 190, 191, 213, 214, 215, 216, 237, 238, 239, 260, 261, 262, 283, 284, 285, 306, 307, 308, 329, 330, 331], "bbox": [0.26232812499999997, 0.21876168224299064, 0.531875, 0.9846028037383177], "iscrowd": 0, "area": 12074.227200000003, "rle": {"size": [428, 640], "counts": "W_V21Y=2N101N2N2N2O0O20000001N1000000O1010O0000000001O000001O01O000000001O0001O0001O1O1O1O1O12M5QDc0f9AUFf2e7^MVHc2j7`MRH`2o7_MPHc2P8\\MPHe2P8ZMPHf2P8YMQHh2n7XMRHi2n7VMQHk2o7UMQHl2n7TMRHm2m7SMSHm2m7SMRHo2n7PMRHh2V8XMjGm1Q9SNoFR1l9nNTFi0V:VOiEj0X:UOiEj0X:VOhEj0X:VOhEi0k7jNkI=YNi0k7mNlI8ZNj0j7QOjI5\\Ni0j7TOiI3]Nh0i7XOiI0^Nh0h7[OhIL`Ni0h7]OhII`Ni0g7AhIFaNR1]7[OPJCcNR2\\6]NPK@eNS3Y5aMPL\\OfN]3P5YMZLYOfN^3W3\\LhMo0:WOgN^3W3^LdMQ1<ROjN_3V3`LaMQ1>POjN`3W3aL^MQ1`0nNkN`3W3cLZMS1c0iNlNa3W3eLWMS1e0fNnNb3V3gLTMS1g0dNoNb3V3iLPMU1i0`NPOc3W3jLmLU1k0^NQOc3W3jLlLX1k0YNSOe3V3kLiLZ1m0VNTOe3V3kLhL\\1m0TNTO`3]3QM_L_1n0oMWOZ3c3YMULa1P1lMXOR3k3aMjKe1S1gMXOk2S4jM_Kh1T1cMZOd2Z4QNVKi1o1OS3YNjJi1U2GY3`N_Ji1Z2_O_3V1dLcNd3Z1]LbNi3[1YL`Nm3]1UL^NQ4^1RL]NV3lNPKe2l1[NX3UOhJ]2Q2ZN\\3^O_JT2X2YN^3HVJl1^2WNa32mId1d2VNb3<eI\\1j2TNf3k2_LnLf3P3\\3N2M3N3M5K4K6M3L4L3N0O1O100O1O100O11O1M2M4L4K4N3O1N101O1O000O19Ge0ZO10O10000O00100OD<J7L3M3M3L5N1O101N101N2O2M2O1N3N1N3N1N2OlXm3"}, "label": "the giraffe further from the fence"}]}
{"id": 296191, "image": "COCO_train2014_000000296191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is smaller than the other one\"."}], "task": "refering", "answer_template": "The \"a giraffe that is smaller than the other one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 104, 124, 125, 126, 146, 147, 148, 149, 169, 170, 190, 191, 213, 214, 215, 216, 237, 238, 239, 260, 261, 262, 283, 284, 285, 306, 307, 308, 329, 330, 331], "bbox": [0.26232812499999997, 0.21876168224299064, 0.531875, 0.9846028037383177], "iscrowd": 0, "area": 12074.227200000003, "rle": {"size": [428, 640], "counts": "W_V21Y=2N101N2N2N2O0O20000001N1000000O1010O0000000001O000001O01O000000001O0001O0001O1O1O1O1O12M5QDc0f9AUFf2e7^MVHc2j7`MRH`2o7_MPHc2P8\\MPHe2P8ZMPHf2P8YMQHh2n7XMRHi2n7VMQHk2o7UMQHl2n7TMRHm2m7SMSHm2m7SMRHo2n7PMRHh2V8XMjGm1Q9SNoFR1l9nNTFi0V:VOiEj0X:UOiEj0X:VOhEj0X:VOhEi0k7jNkI=YNi0k7mNlI8ZNj0j7QOjI5\\Ni0j7TOiI3]Nh0i7XOiI0^Nh0h7[OhIL`Ni0h7]OhII`Ni0g7AhIFaNR1]7[OPJCcNR2\\6]NPK@eNS3Y5aMPL\\OfN]3P5YMZLYOfN^3W3\\LhMo0:WOgN^3W3^LdMQ1<ROjN_3V3`LaMQ1>POjN`3W3aL^MQ1`0nNkN`3W3cLZMS1c0iNlNa3W3eLWMS1e0fNnNb3V3gLTMS1g0dNoNb3V3iLPMU1i0`NPOc3W3jLmLU1k0^NQOc3W3jLlLX1k0YNSOe3V3kLiLZ1m0VNTOe3V3kLhL\\1m0TNTO`3]3QM_L_1n0oMWOZ3c3YMULa1P1lMXOR3k3aMjKe1S1gMXOk2S4jM_Kh1T1cMZOd2Z4QNVKi1o1OS3YNjJi1U2GY3`N_Ji1Z2_O_3V1dLcNd3Z1]LbNi3[1YL`Nm3]1UL^NQ4^1RL]NV3lNPKe2l1[NX3UOhJ]2Q2ZN\\3^O_JT2X2YN^3HVJl1^2WNa32mId1d2VNb3<eI\\1j2TNf3k2_LnLf3P3\\3N2M3N3M5K4K6M3L4L3N0O1O100O1O100O11O1M2M4L4K4N3O1N101O1O000O19Ge0ZO10O10000O00100OD<J7L3M3M3L5N1O101N101N2O2M2O1N3N1N3N1N2OlXm3"}, "label": "a giraffe that is smaller than the other one"}]}
{"id": 296191, "image": "COCO_train2014_000000296191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe whose head appears larger\"."}], "task": "refering", "answer_template": "The \"the giraffe whose head appears larger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 214, 215, 216], "bbox": [0.0015, 0.21796728971962617, 0.42379687499999996, 0.6426869158878504], "iscrowd": 0, "area": 28352.974349999993, "rle": {"size": [428, 640], "counts": "jd08d<`0@`0_Oa0@`0@`0@?010000001O00000O2O0000001O00000O2O0000001O00000O10O10000O10000O01000O1000000O10O10O10000O10000O01000O10000O1000O010000000000000O100000000000000O100000000000000O10000O1000000O10000O100000hNW110000O1000000O10O101O00010O01O010O00100O010O0010O00001O00001N10001O001O00010O01O01O01O010O0010O01O01O01O010O001O00010O001O001O01O01O001O00100O1O001O10O01O1O1O00100O001O1O1O010O2N2N1O2N2O1N2N1O2N2O1N1O2N2O001O0000000O2O000000001O000O100O1O2N100O1O1O2O0O1O1O101N1O1O100O1O2N100O1O2N101N1O2N1O2N2N101N1O2N1O2N1O2N1O2N1O2N2N1O2O0O2N4L5K4L4L4Gaoi4"}, "label": "the giraffe whose head appears larger"}]}
{"id": 296191, "image": "COCO_train2014_000000296191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraff in front of the woman eating leaves\"."}], "task": "refering", "answer_template": "The \"the giraff in front of the woman eating leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 214, 215, 216], "bbox": [0.0015, 0.21796728971962617, 0.42379687499999996, 0.6426869158878504], "iscrowd": 0, "area": 28352.974349999993, "rle": {"size": [428, 640], "counts": "jd08d<`0@`0_Oa0@`0@`0@?010000001O00000O2O0000001O00000O2O0000001O00000O10O10000O10000O01000O1000000O10O10O10000O10000O01000O10000O1000O010000000000000O100000000000000O100000000000000O10000O1000000O10000O100000hNW110000O1000000O10O101O00010O01O010O00100O010O0010O00001O00001N10001O001O00010O01O01O01O010O0010O01O01O01O010O001O00010O001O001O01O01O001O00100O1O001O10O01O1O1O00100O001O1O1O010O2N2N1O2N2O1N2N1O2N2O1N1O2N2O001O0000000O2O000000001O000O100O1O2N100O1O1O2O0O1O1O101N1O1O100O1O2N100O1O2N101N1O2N1O2N2N101N1O2N1O2N1O2N1O2N1O2N2N1O2O0O2N4L5K4L4L4Gaoi4"}, "label": "the giraff in front of the woman eating leaves"}]}
{"id": 255233, "image": "COCO_train2014_000000255233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow above - ground tram making its way through the city\"."}], "task": "refering", "answer_template": "The \"a yellow above - ground tram making its way through the city\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 267, 268, 269, 270, 271, 291, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.591875, 0.6862264150943396, 1.0, 0.9998113207547169], "iscrowd": 0, "area": 12383.082100000003, "rle": {"size": [424, 640], "counts": "oVm4;i<401O010O1O001O1O00100O001O00100O00100O00100O0010O0100O00100O03N9G9FWW1d0ShN3O001O001O1O001O10O01O010O1O010O00100O0010O01O1O010O00100O0010O01O10O01O010O1O0010O01O10O01O010O1O010O00100O001O10O01O010O1O010O00100O0010O01O1O010O00100O0010O01O10O01O010O1O0010O01O10O01O010O1O010O010O1O010O00100O010O1O010O0010O010O00010O01O010O010O00010O010O0010O01O01O010O01O010O00010O010O0010O010O00010O01O010O01O01O010O0010O010O00010O00001O001O00001O010O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O0000"}, "label": "a yellow above - ground tram making its way through the city"}]}
{"id": 255233, "image": "COCO_train2014_000000255233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a subway cart that has windows\"."}], "task": "refering", "answer_template": "The \"a subway cart that has windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 267, 268, 269, 270, 271, 291, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.591875, 0.6862264150943396, 1.0, 0.9998113207547169], "iscrowd": 0, "area": 12383.082100000003, "rle": {"size": [424, 640], "counts": "oVm4;i<401O010O1O001O1O00100O001O00100O00100O00100O0010O0100O00100O03N9G9FWW1d0ShN3O001O001O1O001O10O01O010O1O010O00100O0010O01O1O010O00100O0010O01O10O01O010O1O0010O01O10O01O010O1O010O00100O001O10O01O010O1O010O00100O0010O01O1O010O00100O0010O01O10O01O010O1O0010O01O10O01O010O1O010O010O1O010O00100O010O1O010O0010O010O00010O01O010O010O00010O010O0010O01O01O010O01O010O00010O010O0010O010O00010O01O010O01O01O010O0010O010O00010O00001O001O00001O010O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O0000"}, "label": "a subway cart that has windows"}]}
{"id": 337156, "image": "COCO_train2014_000000337156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with its mouth open\"."}], "task": "refering", "answer_template": "The \"a zebra with its mouth open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 189, 190, 194, 212, 217, 235], "bbox": [0.22349999999999998, 0.25, 0.59140625, 0.7827247191011235], "iscrowd": 0, "area": 20577.362999999998, "rle": {"size": [356, 640], "counts": "^ka12P;4L5_GMi58PJ3g51TJ8d5MVJ=`5H[Jb0\\5C^Jf0Z5@_Jj0X5]O`Jh0\\5@[Jc0c5DUJ?j5HlI<T6Z20O100O010\\Od0K5N2N2N2N2N2N2N1O2N2N2N2N2N2O1N2N2N2N101N2N2N2N2O1N2N2O1O001O100O1O1O2N11O100O1O100O1O010O1O100O10O00010O000010O0001O01O01O0010O0001O00010O00001O00001O0000000001O000001O00000000000000010O00000000000000001O0001O000000000001O0000000000000h1XN10O01O1O0O2N1O2N1O1O1O1O100O1O1O1O1O1O1O1OcNjInMU6m1\\JhMd5Q2oJcMQ5W2P2L5O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O1O1O001O1O1O001N2NHVNZGi1e8[NYGc1g8aNVG^1i8gNUGW1k8a0O02N2O2M2N23M3M3N2M3M4L3M3NO0O10001N2O1O1N101N1M1M4L3_NXGd0l8\\OYG<l8AZG6l8HYG0m8LYGKm82U1Mnoj2"}, "label": "a zebra with its mouth open"}]}
{"id": 337156, "image": "COCO_train2014_000000337156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in a grassy field with a look on its face that makes it seem like it ' s smiling for the camera\"."}], "task": "refering", "answer_template": "The \"a zebra in a grassy field with a look on its face that makes it seem like it ' s smiling for the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 189, 190, 194, 212, 217, 235], "bbox": [0.22349999999999998, 0.25, 0.59140625, 0.7827247191011235], "iscrowd": 0, "area": 20577.362999999998, "rle": {"size": [356, 640], "counts": "^ka12P;4L5_GMi58PJ3g51TJ8d5MVJ=`5H[Jb0\\5C^Jf0Z5@_Jj0X5]O`Jh0\\5@[Jc0c5DUJ?j5HlI<T6Z20O100O010\\Od0K5N2N2N2N2N2N2N1O2N2N2N2N2N2O1N2N2N2N101N2N2N2N2O1N2N2O1O001O100O1O1O2N11O100O1O100O1O010O1O100O10O00010O000010O0001O01O01O0010O0001O00010O00001O00001O0000000001O000001O00000000000000010O00000000000000001O0001O000000000001O0000000000000h1XN10O01O1O0O2N1O2N1O1O1O1O100O1O1O1O1O1O1O1OcNjInMU6m1\\JhMd5Q2oJcMQ5W2P2L5O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O1O1O001O1O1O001N2NHVNZGi1e8[NYGc1g8aNVG^1i8gNUGW1k8a0O02N2O2M2N23M3M3N2M3M4L3M3NO0O10001N2O1O1N101N1M1M4L3_NXGd0l8\\OYG<l8AZG6l8HYG0m8LYGKm82U1Mnoj2"}, "label": "a zebra in a grassy field with a look on its face that makes it seem like it ' s smiling for the camera"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blonde wearing a tanktop\"."}], "task": "refering", "answer_template": "The \"the blonde wearing a tanktop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 354, 373, 374, 375, 376, 377], "bbox": [0.238171875, 0.5292708333333334, 0.40709375000000003, 0.9819791666666667], "iscrowd": 0, "area": 18023.625, "rle": {"size": [480, 640], "counts": "UdW2`1[=<E:F;E:E;F6K3N2O1N2N2O1N2N2O1N2O1N2O1N2M3L4K5L4K5L4K5L4K5L4K5L4K5M3O1O1O1O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O1O100O10000O10000O10000O10000001O0000001O0000001O1O001O0QMgGWOZ8e0jGZOW8b0nG\\OR8a0RH^Oo7>UHAk7<ZHBg7:]HEd78`HF`77dHH]74hHIZ74iHKX72lHLU71QIKP72WIIj64\\IHd66bIF_67gIEZ68kIEV68PJDQ69TJDm58ZJDg58_JEb56fJE\\57jJ@^5:jJ[Oi55^JAeaa5"}, "label": "the blonde wearing a tanktop"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black tank top with sunglasses\"."}], "task": "refering", "answer_template": "The \"a woman in a black tank top with sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 354, 373, 374, 375, 376, 377], "bbox": [0.238171875, 0.5292708333333334, 0.40709375000000003, 0.9819791666666667], "iscrowd": 0, "area": 18023.625, "rle": {"size": [480, 640], "counts": "UdW2`1[=<E:F;E:E;F6K3N2O1N2N2O1N2N2O1N2O1N2O1N2M3L4K5L4K5L4K5L4K5L4K5L4K5M3O1O1O1O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O1O100O10000O10000O10000O10000001O0000001O0000001O1O001O0QMgGWOZ8e0jGZOW8b0nG\\OR8a0RH^Oo7>UHAk7<ZHBg7:]HEd78`HF`77dHH]74hHIZ74iHKX72lHLU71QIKP72WIIj64\\IHd66bIF_67gIEZ68kIEV68PJDQ69TJDm58ZJDg58_JEb56fJE\\57jJ@^5:jJ[Oi55^JAeaa5"}, "label": "a woman in a black tank top with sunglasses"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with black hair wearing black shirt standing with three other peoples\"."}], "task": "refering", "answer_template": "The \"a girl with black hair wearing black shirt standing with three other peoples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 241, 262, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 377, 378, 379, 380, 381], "bbox": [0.39775, 0.5842708333333333, 0.603375, 0.9887708333333334], "iscrowd": 0, "area": 17402.914950000002, "rle": {"size": [480, 640], "counts": "kkg3<X>>A>C>I6N2M4M2cEaNU7b1WG1h82RG2l81oF3P90jF4T9OgF5X9McF7[9h2O1N3O0O1000000O10000O1000000O10000O1000000O10000O11O00000010O000001O0000001O0000001O0000001O0000001O002N2N3M2N2N2N2O1N2N3M2N2N2N6J6J5K6J5K;E<D5K1O001O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O001O4L4L4L5K4L4L4L5K4L4L5K4L4L4L5K4M3L5KXcf3"}, "label": "a girl with black hair wearing black shirt standing with three other peoples"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"asian lady in front\"."}], "task": "refering", "answer_template": "The \"asian lady in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 241, 262, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 377, 378, 379, 380, 381], "bbox": [0.39775, 0.5842708333333333, 0.603375, 0.9887708333333334], "iscrowd": 0, "area": 17402.914950000002, "rle": {"size": [480, 640], "counts": "kkg3<X>>A>C>I6N2M4M2cEaNU7b1WG1h82RG2l81oF3P90jF4T9OgF5X9McF7[9h2O1N3O0O1000000O10000O1000000O10000O1000000O10000O11O00000010O000001O0000001O0000001O0000001O0000001O002N2N3M2N2N2N2O1N2N3M2N2N2N6J6J5K6J5K;E<D5K1O001O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O001O4L4L4L5K4L4L4L5K4L4L5K4L4L4L5K4M3L5KXcf3"}, "label": "asian lady in front"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blonde woman standing behind the woman with black hair\"."}], "task": "refering", "answer_template": "The \"the blonde woman standing behind the woman with black hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 218, 219, 220, 241, 242, 243, 244, 264, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 335, 336, 337, 358, 359, 360, 361, 381, 382, 383], "bbox": [0.500921875, 0.5028125, 0.70171875, 0.9947291666666668], "iscrowd": 0, "area": 17257.245800000004, "rle": {"size": [480, 640], "counts": "]gf4k0o=:I6I7J7K4L4M3M4K4M3M4L3L4M3\\D[Mg:_3O2O001O0O2O000O2O001O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N21O0001O011N4L3M3N2M4L3M3N2M3M4M2M3M3M4M2M3M3M3N2MO1O1O2N3M3M4L4L3M4L3M4L3M4L4L3M3M1O001O001O1O001O1O001O001O1O001O1O001O001O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M4L3M6J8H7E<B>B>A>C`Si2"}, "label": "the blonde woman standing behind the woman with black hair"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest girl\"."}], "task": "refering", "answer_template": "The \"tallest girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 218, 219, 220, 241, 242, 243, 244, 264, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 335, 336, 337, 358, 359, 360, 361, 381, 382, 383], "bbox": [0.500921875, 0.5028125, 0.70171875, 0.9947291666666668], "iscrowd": 0, "area": 17257.245800000004, "rle": {"size": [480, 640], "counts": "]gf4k0o=:I6I7J7K4L4M3M4K4M3M4L3L4M3\\D[Mg:_3O2O001O0O2O000O2O001O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N21O0001O011N4L3M3N2M4L3M3N2M3M4M2M3M3M4M2M3M3M3N2MO1O1O2N3M3M4L4L3M4L3M4L3M4L4L3M3M1O001O001O1O001O1O001O001O1O001O1O001O001O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M4L3M6J8H7E<B>B>A>C`Si2"}, "label": "tallest girl"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man behind a group of three ladies by a stop sign\"."}], "task": "refering", "answer_template": "The \"a man behind a group of three ladies by a stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 191, 192, 193, 215, 216, 235, 236, 239, 258, 259, 280, 281, 282, 303, 304, 326, 327, 349, 350, 372], "bbox": [0.170609375, 0.4425625, 0.43918749999999995, 1.0], "iscrowd": 0, "area": 8142.8165999999965, "rle": {"size": [480, 640], "counts": "mac13c>:F:F:F9G:F:L4N2N2M3N2N1N3N2N2M3N2N2M3^OXMjDi2U;[MgDf2X;_MbDa2_;?0O100O100O100O010O100O100O100O1O1K5K5L4N2N2N1O2N2N2N2O1N3M4M3N1O2N2N2N2N1O2N2N2M2M4H]BPOf=m0\\BPOh=l08M2N3L4M3M3L3N3M3LR]?a0`b@1O100O1O100O1O1lAYOo=g0PBZOo=g0PBZOo=g0PBZOo=l000O1O101N1O100O1O100O1O101N2N3O0O2N1O2N101N1O2N101N1O2N1O2O0O`0@O100O2N2N3N2M4L3M3N2M4L3M3N2@fBUO]=b0mBZOW=<l0G`QX5"}, "label": "a man behind a group of three ladies by a stop sign"}]}
{"id": 58633, "image": "COCO_train2014_000000058633.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man behind of the women\"."}], "task": "refering", "answer_template": "The \"man behind of the women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 191, 192, 193, 215, 216, 235, 236, 239, 258, 259, 280, 281, 282, 303, 304, 326, 327, 349, 350, 372], "bbox": [0.170609375, 0.4425625, 0.43918749999999995, 1.0], "iscrowd": 0, "area": 8142.8165999999965, "rle": {"size": [480, 640], "counts": "mac13c>:F:F:F9G:F:L4N2N2M3N2N1N3N2N2M3N2N2M3^OXMjDi2U;[MgDf2X;_MbDa2_;?0O100O100O100O010O100O100O100O1O1K5K5L4N2N2N1O2N2N2N2O1N3M4M3N1O2N2N2N2N1O2N2N2M2M4H]BPOf=m0\\BPOh=l08M2N3L4M3M3L3N3M3LR]?a0`b@1O100O1O100O1O1lAYOo=g0PBZOo=g0PBZOo=g0PBZOo=l000O1O101N1O100O1O100O1O101N2N3O0O2N1O2N101N1O2N101N1O2N1O2O0O`0@O100O2N2N3N2M4L3M3N2M4L3M3N2@fBUO]=b0mBZOW=<l0G`QX5"}, "label": "man behind of the women"}]}
{"id": 369931, "image": "COCO_train2014_000000369931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one women wearing green dress holding cake near her mouth\"."}], "task": "refering", "answer_template": "The \"there is one women wearing green dress holding cake near her mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.48707812500000003, 0.3617916666666667, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 66150.9767, "rle": {"size": [480, 640], "counts": "afb41n>1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1N3N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O100O1O2N100O1O1O100O1O1O10O01O1O1O100O1O1O100O1O1O100O1O1O100[Ne1ZOf0H8H8I7H8I700000000001O000000000000000000001O000000000000000000001O0000000000000000001O000000000000000000001O000000000000000000N2L4L4L4M3L4L4L4L4M3L4L4L4M3M3N2N2M3N2N2N2N2M3N2N2N2N2M3N2N2N2N2O1N2N2O1N2N2N2O1N2O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O1000000001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O1O2N1O2N1O1O2N1O2N101N1O1O2N1O2N1O1O2N1O2N1O6J:F:F9G:B>eN[1eN[1eN[1eNSG"}, "label": "there is one women wearing green dress holding cake near her mouth"}]}
{"id": 369931, "image": "COCO_train2014_000000369931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark haired woman wearing green shirt eating a piece of cake\"."}], "task": "refering", "answer_template": "The \"a dark haired woman wearing green shirt eating a piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.48707812500000003, 0.3617916666666667, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 66150.9767, "rle": {"size": [480, 640], "counts": "afb41n>1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1N3N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O100O1O2N100O1O1O100O1O1O10O01O1O1O100O1O1O100O1O1O100O1O1O100[Ne1ZOf0H8H8I7H8I700000000001O000000000000000000001O000000000000000000001O0000000000000000001O000000000000000000001O000000000000000000N2L4L4L4M3L4L4L4L4M3L4L4L4M3M3N2N2M3N2N2N2N2M3N2N2N2N2M3N2N2N2N2O1N2N2O1N2N2N2O1N2O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O1000000001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O1O2N1O2N1O1O2N1O2N101N1O1O2N1O2N1O1O2N1O2N1O6J:F:F9G:B>eN[1eN[1eN[1eNSG"}, "label": "a dark haired woman wearing green shirt eating a piece of cake"}]}
{"id": 369931, "image": "COCO_train2014_000000369931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding the glass\"."}], "task": "refering", "answer_template": "The \"the woman holding the glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 190, 191, 192, 193, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0050625, 0.40675, 0.510671875, 0.9865208333333334], "iscrowd": 0, "area": 59379.162449999996, "rle": {"size": [480, 640], "counts": "hh1>a>]1cNY1gN1O1N2O1O1O1O1O001O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O10O01O1O1000000O100000000O100000000O10000000000001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001OO1O1O1O1N2O1O1O1O1O1O1N2O1O1YO\\JQHe5m7bJmG_5P8jJiGW5U8k0N2M3N2N2M3N2N2M201N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O11O001O001O001O001O001O001O001O001O1O001N101O0^KXIl0i6ROXIn0h6ROXIn0i6POXIP1h6oNYIQ1h6mNYIS1g6lNZIT1g6jNZIV1f6iN[IW1f6gN[IY1f6eN[I[1e6dN\\I\\1e6bN\\I^1d6aN]I_1d6_N]Ia1c6^N^Ib1c6\\N^Id1d6YN]Ig1f6UN[Ik1h6QNYIo1j6mMWIc1DWLX7R2UIg1^7VNbHi1b7SN_Hm1d7mM_HR2d7gMaHY2b7^MdHa2b7TMdHl2V8ULoGk3c9O2O0O10000O100O100O1O1O2M3N2N2N2M3N3M2N2N3M5K6J5L2M3M2N3M3M3M3M3M3M3M3M3N5J5K6J5K5K6J5K6JWZb4"}, "label": "the woman holding the glass"}]}
{"id": 369931, "image": "COCO_train2014_000000369931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding liquor glass in his right hand\"."}], "task": "refering", "answer_template": "The \"a woman holding liquor glass in his right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 190, 191, 192, 193, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0050625, 0.40675, 0.510671875, 0.9865208333333334], "iscrowd": 0, "area": 59379.162449999996, "rle": {"size": [480, 640], "counts": "hh1>a>]1cNY1gN1O1N2O1O1O1O1O001O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O10O01O1O1000000O100000000O100000000O10000000000001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001OO1O1O1O1N2O1O1O1O1O1O1N2O1O1YO\\JQHe5m7bJmG_5P8jJiGW5U8k0N2M3N2N2M3N2N2M201N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O11O001O001O001O001O001O001O001O001O1O001N101O0^KXIl0i6ROXIn0h6ROXIn0i6POXIP1h6oNYIQ1h6mNYIS1g6lNZIT1g6jNZIV1f6iN[IW1f6gN[IY1f6eN[I[1e6dN\\I\\1e6bN\\I^1d6aN]I_1d6_N]Ia1c6^N^Ib1c6\\N^Id1d6YN]Ig1f6UN[Ik1h6QNYIo1j6mMWIc1DWLX7R2UIg1^7VNbHi1b7SN_Hm1d7mM_HR2d7gMaHY2b7^MdHa2b7TMdHl2V8ULoGk3c9O2O0O10000O100O100O1O1O2M3N2N2N2M3N3M2N2N3M5K6J5L2M3M2N3M3M3M3M3M3M3M3M3N5J5K6J5K5K6J5K6JWZb4"}, "label": "a woman holding liquor glass in his right hand"}]}
{"id": 484620, "image": "COCO_train2014_000000484620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a red bus\"."}], "task": "refering", "answer_template": "The \"the front of a red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 54, 55, 72, 73, 90, 91, 108, 109, 126, 127, 144, 145], "bbox": [0.0015, 0.17751497005988023, 0.09306, 0.7213473053892215], "iscrowd": 0, "area": 7142.52015, "rle": {"size": [334, 500], "counts": "T?k2h4k21O001O001O001O00001N101O001O001O001O00001O001O001O001O001O001O0O101O0G:E:G:E:G:F9G:]Ob0lNU1jNkfc4"}, "label": "the front of a red bus"}]}
{"id": 484620, "image": "COCO_train2014_000000484620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of a bus behind another bus\"."}], "task": "refering", "answer_template": "The \"part of a bus behind another bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 54, 55, 72, 73, 90, 91, 108, 109, 126, 127, 144, 145], "bbox": [0.0015, 0.17751497005988023, 0.09306, 0.7213473053892215], "iscrowd": 0, "area": 7142.52015, "rle": {"size": [334, 500], "counts": "T?k2h4k21O001O001O001O00001N101O001O001O001O00001O001O001O001O001O001O0O101O0G:E:G:E:G:F9G:]Ob0lNU1jNkfc4"}, "label": "part of a bus behind another bus"}]}
{"id": 435471, "image": "COCO_train2014_000000435471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sliding on the snow with a lady along with him\"."}], "task": "refering", "answer_template": "The \"a man sliding on the snow with a lady along with him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 51, 52, 53, 61, 62, 63, 64, 71, 72, 73, 74, 81, 82, 83, 92, 93, 102, 103, 112, 113, 122, 123], "bbox": [0.11006968641114982, 0.3136342592592593, 0.4394425087108014, 0.8149305555555556], "iscrowd": 0, "area": 11227.223100000003, "rle": {"size": [432, 287], "counts": "Tf==o<5M3M3N2M2N2O2M2O1O1O2N1O001O1O2N2N2eDhN\\:Y2L3N2M4M2M2jF]M\\7e2aH`M\\7W3lGlLQ8\\4fGSKm7Q5jGUKT8U5N1THhJS72jHX51jJo6d5PI^Jm6d5QI_Jk6V6N100O2O000O1000O10O0100N2^Ob0O1ZNlHmLW7W3QISL[7P4Q14M2M;F:F2NO24L18IL5J6J01O000010O01O1O1N2O1lL_H6c7FUIUOSOgNj7R2TIUNjNc0<ROi7[1mGSO\\1GhNg0_8?\\GROm1>i6>[GQOP2?f6?[GPOR2=g6a0XGPOT2;h6d0UGmNV2;i6g0cImNi6Q1l2O2O0ROfCd0d<0OO2N2N3L4M3GUgS2"}, "label": "a man sliding on the snow with a lady along with him"}]}
{"id": 435471, "image": "COCO_train2014_000000435471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a ski jacket and pants that is cross country skiing\"."}], "task": "refering", "answer_template": "The \"a man wearing a ski jacket and pants that is cross country skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 51, 52, 53, 61, 62, 63, 64, 71, 72, 73, 74, 81, 82, 83, 92, 93, 102, 103, 112, 113, 122, 123], "bbox": [0.11006968641114982, 0.3136342592592593, 0.4394425087108014, 0.8149305555555556], "iscrowd": 0, "area": 11227.223100000003, "rle": {"size": [432, 287], "counts": "Tf==o<5M3M3N2M2N2O2M2O1O1O2N1O001O1O2N2N2eDhN\\:Y2L3N2M4M2M2jF]M\\7e2aH`M\\7W3lGlLQ8\\4fGSKm7Q5jGUKT8U5N1THhJS72jHX51jJo6d5PI^Jm6d5QI_Jk6V6N100O2O000O1000O10O0100N2^Ob0O1ZNlHmLW7W3QISL[7P4Q14M2M;F:F2NO24L18IL5J6J01O000010O01O1O1N2O1lL_H6c7FUIUOSOgNj7R2TIUNjNc0<ROi7[1mGSO\\1GhNg0_8?\\GROm1>i6>[GQOP2?f6?[GPOR2=g6a0XGPOT2;h6d0UGmNV2;i6g0cImNi6Q1l2O2O0ROfCd0d<0OO2N2N3L4M3GUgS2"}, "label": "a man wearing a ski jacket and pants that is cross country skiing"}]}
{"id": 435471, "image": "COCO_train2014_000000435471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on the right\"."}], "task": "refering", "answer_template": "The \"person on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 66, 67, 68, 76, 77, 78, 85, 86, 87, 88, 96, 97, 98, 106, 107, 108, 116, 117, 118, 126, 127], "bbox": [0.5740418118466899, 0.3573842592592592, 0.9020905923344947, 0.8585185185185185], "iscrowd": 0, "area": 11159.410799999998, "rle": {"size": [432, 287], "counts": "PkU23\\=2N101N1O2O0N3M2N3L3N2N2M3]FBR6?jIHR6:iIMS65gI2U60gI6V6J_Ic0_6]OUIQ1i6POjH_1S7aNbHl1\\7UN]HT2`7lM\\HZ2b7gMXHa2e7_MVHh2h7YMSHn2j7]1O2N100O2OO0010ONJcJ_H[5_7iJaHT5^7PKaHj4a7[K_H^4e7a06JPIdJY6X5dIRKY6i4hI^KV6[4mIiKR6Q4nIVLQ6c3PJdLo5U3RJRMm5m2mI[MT6c2cIgM]6f4001O0000001O00001O00001O1O1_OnHaJT7[5XIZJk6?kH^4^8L3M4L4J5K6J5[Of0fMPFo0T:POPFi0S:VOQFe0R:[OPF?S:BPFL`:3nEUO]:j0V1O1O1O1N2O1O1O1N2N1O2M3K5KWc;"}, "label": "person on the right"}]}
{"id": 435471, "image": "COCO_train2014_000000435471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier in all black\"."}], "task": "refering", "answer_template": "The \"the skier in all black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 66, 67, 68, 76, 77, 78, 85, 86, 87, 88, 96, 97, 98, 106, 107, 108, 116, 117, 118, 126, 127], "bbox": [0.5740418118466899, 0.3573842592592592, 0.9020905923344947, 0.8585185185185185], "iscrowd": 0, "area": 11159.410799999998, "rle": {"size": [432, 287], "counts": "PkU23\\=2N101N1O2O0N3M2N3L3N2N2M3]FBR6?jIHR6:iIMS65gI2U60gI6V6J_Ic0_6]OUIQ1i6POjH_1S7aNbHl1\\7UN]HT2`7lM\\HZ2b7gMXHa2e7_MVHh2h7YMSHn2j7]1O2N100O2OO0010ONJcJ_H[5_7iJaHT5^7PKaHj4a7[K_H^4e7a06JPIdJY6X5dIRKY6i4hI^KV6[4mIiKR6Q4nIVLQ6c3PJdLo5U3RJRMm5m2mI[MT6c2cIgM]6f4001O0000001O00001O00001O1O1_OnHaJT7[5XIZJk6?kH^4^8L3M4L4J5K6J5[Of0fMPFo0T:POPFi0S:VOQFe0R:[OPF?S:BPFL`:3nEUO]:j0V1O1O1O1N2O1O1O1N2N1O2M3K5KWc;"}, "label": "the skier in all black"}]}
{"id": 320785, "image": "COCO_train2014_000000320785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the obscured baseball player\"."}], "task": "refering", "answer_template": "The \"the obscured baseball player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 151, 152, 153, 166, 167, 168, 169, 182, 183, 184, 197, 198, 199, 212, 213, 214, 215, 228, 229, 230, 243], "bbox": [0.05870892018779343, 0.40353125, 0.3835446009389671, 0.70109375], "iscrowd": 0, "area": 8623.209500000003, "rle": {"size": [640, 426], "counts": "Sm?3lc03M3M3L4M2O1O00001O002n\\O[O\\b01`]Of01\\OZb05`]Ob03[O\\b0Y100O2N1O10011N2N101O1N12OO0100O010O01O01O2O1N3M2N8H6I6I4L5N1N2O001O1N1O\\]Og0c`0XOU_OT1h`0jNn^Of1m`0[NP_Ok1k`0h0N1O1O1N2O0n_OnLU?R3i@PMV?R3g@oLZ?Q3d@PM]?P3a@QM_?P3_@RM`?Q3\\@PMe?T3U@mLk?e3O0QNVLhCi3Z<XLdCh3]<YLaCg3`<ZL^Cf3c<[L[Ce3f<\\LXCd1fNLS>cNSC\\1SOLk=kNoBV1\\OLg=mNlBT1DKa=QOjBQ1KK\\=TOhBn01LX=UOgBl05NV=TOfBj090S=TOdBl0lNI<5W>TOaBP2ZOjNV>UO`BR2ZOhNX>TO^BU2[OeNY>SO]BX2\\OcNa?^1`@_Na?b1`@\\N_?e1c@XN^?i1c@UN]?l1d@QN]?P2d@mM]?T2d@jM\\?W2V11O001O1O1O1O1O1O01000003XNW_O2k`0TOm_Ol0da001O01O000001O0O1O1O101N1001O102M3M00O10000O100O0O3M3L5L3L5KmnS5"}, "label": "the obscured baseball player"}]}
{"id": 320785, "image": "COCO_train2014_000000320785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in backround\"."}], "task": "refering", "answer_template": "The \"person in backround\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 151, 152, 153, 166, 167, 168, 169, 182, 183, 184, 197, 198, 199, 212, 213, 214, 215, 228, 229, 230, 243], "bbox": [0.05870892018779343, 0.40353125, 0.3835446009389671, 0.70109375], "iscrowd": 0, "area": 8623.209500000003, "rle": {"size": [640, 426], "counts": "Sm?3lc03M3M3L4M2O1O00001O002n\\O[O\\b01`]Of01\\OZb05`]Ob03[O\\b0Y100O2N1O10011N2N101O1N12OO0100O010O01O01O2O1N3M2N8H6I6I4L5N1N2O001O1N1O\\]Og0c`0XOU_OT1h`0jNn^Of1m`0[NP_Ok1k`0h0N1O1O1N2O0n_OnLU?R3i@PMV?R3g@oLZ?Q3d@PM]?P3a@QM_?P3_@RM`?Q3\\@PMe?T3U@mLk?e3O0QNVLhCi3Z<XLdCh3]<YLaCg3`<ZL^Cf3c<[L[Ce3f<\\LXCd1fNLS>cNSC\\1SOLk=kNoBV1\\OLg=mNlBT1DKa=QOjBQ1KK\\=TOhBn01LX=UOgBl05NV=TOfBj090S=TOdBl0lNI<5W>TOaBP2ZOjNV>UO`BR2ZOhNX>TO^BU2[OeNY>SO]BX2\\OcNa?^1`@_Na?b1`@\\N_?e1c@XN^?i1c@UN]?l1d@QN]?P2d@mM]?T2d@jM\\?W2V11O001O1O1O1O1O1O01000003XNW_O2k`0TOm_Ol0da001O01O000001O0O1O1O101N1001O102M3M00O10000O100O0O3M3L5L3L5KmnS5"}, "label": "person in backround"}]}
{"id": 320785, "image": "COCO_train2014_000000320785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy pitcher\"."}], "task": "refering", "answer_template": "The \"boy pitcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 106, 107, 115, 116, 117, 121, 122, 126, 127, 128, 129, 130, 131, 132, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 167, 168, 169, 170, 171, 172, 173, 174, 176, 177, 178, 187, 188, 189, 190, 202, 203, 204, 205, 206, 217, 218, 219, 232, 233, 247, 248, 262, 263, 277, 278, 292, 293], "bbox": [0.008215962441314555, 0.260609375, 0.8972535211267606, 0.8536718750000001], "iscrowd": 0, "area": 36019.20470000003, "rle": {"size": [640, 426], "counts": "`e29fc01O2N1O1O2O0O2O00001O01O001O01M2O10001O1O1O1N2O1O1O1O100O00000010OO100W]OTO^b0l0b]OWO[b0j0d]OXOZb0h0f]O[OWb0f0h]O[OWb0e0i]O]OUb0d0i]O^OUb0c0k]O_OSb0X1O2N1O2N1N2O2N1]^OPNTa0`2N3M2N3M2N3M3M3M4L4L7I8H5K2N1O2N2O0O2N]Lc@e2[?[Mj@a2V?]MPA`2o>]MXA_2i>RMiAj2h?N10O02O2M4M7H7J2N1N2N3M2N2N3M2N2N3M2N2N3M2N2N2N000000O1000000000000O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O1O100O100O1O100O100O1O100O100O1O100O1O2O0O100O1O100O1O100O100O100O100O100000000O2O0000001O00001O00000O2O0O101O2M2O2N1N5L3^DbM\\5b2\\JcMb5b2VJcMi5`2PJfMn5^2ZI_L`LW1U:^2PIeLdLQ1\\:_2hHeLfLP1c:`2]HWNc7o1THUNk7Q2jGTNV8Q2PGeNo8a1hFcNW9c1`FaN_9Q60OO1000000O100000000O100000000OQOiEWHW:h7TFnGl9R8UFmGk9R8WFmGi9R8YFmGg9R8[FmGd9T8^FjGb9k7dEUHa;i7cDUH];i7gDTHZ;j7jDTHV;\\7^DoHR<n6\\DfHf;V7d0L4M3L3M3C>E:F;E:_Ob0mNR1101N1101N10O001O001N101O00XOn_OcMQ`0U2W@kMh?n1a@RN^?h1i@WNV?j1k@VNT?i1m@WNR?j1o@VNP?i1RAVNm>k1SAVNl>i1VAVNj>j1WAVNi>h1YAWNg>i1YAXNg>f1[AYNf>f1[AZNd>e1^AZNc>e1]A\\Nb>c1`A\\N`>c1bA]NR>Ah@R2V1]NP>Gf@k1\\1^Nk=Q2VBnMg=T2[BlMb=U2_BkM_=V2cBjM[=V2fBjMY=V2iBjMU=V2mBiMR=W2oBjMo<V2SCiMl<W2UCjMi<V2YCiMg<V2[CjMc<V2^CjMb<U2`CkM_<T2bCmM\\<S2fClMZ<S2hCmMW<R2jCoMT<Q2nCoMQ<P2PDQNn;P2SDoMm;R2SDnMk;T2TDmMk;T2UDlMi;V2VDkMi;V2VDjMk;V2TDkMk;V2TDkMl;U2TDkMk;V2TDkMk;V2TDjMm;V2RDjMn;W2QDhMP<Y2PDfMQ<T3TClLl<T3TClLl<T3TCkLn<b1iAZOZ1SOm<b1kAZOY1SOl<a1nA[OW1SOk<a1oA\\OW1ROk<_1oA@W1POj<_1oABY1mNi<^1PBD[1kNf<_1oAG_1fNc<`1PBJb1aN^<d1QBKP`02R@No?NT@2n?GW@9k?@Z@?i?\\O[@c0g?mNZ_O0P1R1i?eN__O8h0S1W`0kNi_OT1Y`0kNg_OU1[`0jNd_OU1^`0jNc_OU1_`0jN`_OV1a`0lN\\_OS1j`0gNW_OY1ia0N100O10001N100O2O000O2O000O2O0O101O0O2O0N3M3L4M3M3M3M3MnVk0"}, "label": "boy pitcher"}]}
{"id": 320785, "image": "COCO_train2014_000000320785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy pitching a baseball\"."}], "task": "refering", "answer_template": "The \"the boy pitching a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 106, 107, 115, 116, 117, 121, 122, 126, 127, 128, 129, 130, 131, 132, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 167, 168, 169, 170, 171, 172, 173, 174, 176, 177, 178, 187, 188, 189, 190, 202, 203, 204, 205, 206, 217, 218, 219, 232, 233, 247, 248, 262, 263, 277, 278, 292, 293], "bbox": [0.008215962441314555, 0.260609375, 0.8972535211267606, 0.8536718750000001], "iscrowd": 0, "area": 36019.20470000003, "rle": {"size": [640, 426], "counts": "`e29fc01O2N1O1O2O0O2O00001O01O001O01M2O10001O1O1O1N2O1O1O1O100O00000010OO100W]OTO^b0l0b]OWO[b0j0d]OXOZb0h0f]O[OWb0f0h]O[OWb0e0i]O]OUb0d0i]O^OUb0c0k]O_OSb0X1O2N1O2N1N2O2N1]^OPNTa0`2N3M2N3M2N3M3M3M4L4L7I8H5K2N1O2N2O0O2N]Lc@e2[?[Mj@a2V?]MPA`2o>]MXA_2i>RMiAj2h?N10O02O2M4M7H7J2N1N2N3M2N2N3M2N2N3M2N2N3M2N2N2N000000O1000000000000O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O1O100O100O1O100O100O1O100O100O1O100O1O2O0O100O1O100O1O100O100O100O100O100000000O2O0000001O00001O00000O2O0O101O2M2O2N1N5L3^DbM\\5b2\\JcMb5b2VJcMi5`2PJfMn5^2ZI_L`LW1U:^2PIeLdLQ1\\:_2hHeLfLP1c:`2]HWNc7o1THUNk7Q2jGTNV8Q2PGeNo8a1hFcNW9c1`FaN_9Q60OO1000000O100000000O100000000OQOiEWHW:h7TFnGl9R8UFmGk9R8WFmGi9R8YFmGg9R8[FmGd9T8^FjGb9k7dEUHa;i7cDUH];i7gDTHZ;j7jDTHV;\\7^DoHR<n6\\DfHf;V7d0L4M3L3M3C>E:F;E:_Ob0mNR1101N1101N10O001O001N101O00XOn_OcMQ`0U2W@kMh?n1a@RN^?h1i@WNV?j1k@VNT?i1m@WNR?j1o@VNP?i1RAVNm>k1SAVNl>i1VAVNj>j1WAVNi>h1YAWNg>i1YAXNg>f1[AYNf>f1[AZNd>e1^AZNc>e1]A\\Nb>c1`A\\N`>c1bA]NR>Ah@R2V1]NP>Gf@k1\\1^Nk=Q2VBnMg=T2[BlMb=U2_BkM_=V2cBjM[=V2fBjMY=V2iBjMU=V2mBiMR=W2oBjMo<V2SCiMl<W2UCjMi<V2YCiMg<V2[CjMc<V2^CjMb<U2`CkM_<T2bCmM\\<S2fClMZ<S2hCmMW<R2jCoMT<Q2nCoMQ<P2PDQNn;P2SDoMm;R2SDnMk;T2TDmMk;T2UDlMi;V2VDkMi;V2VDjMk;V2TDkMk;V2TDkMl;U2TDkMk;V2TDkMk;V2TDjMm;V2RDjMn;W2QDhMP<Y2PDfMQ<T3TClLl<T3TClLl<T3TCkLn<b1iAZOZ1SOm<b1kAZOY1SOl<a1nA[OW1SOk<a1oA\\OW1ROk<_1oA@W1POj<_1oABY1mNi<^1PBD[1kNf<_1oAG_1fNc<`1PBJb1aN^<d1QBKP`02R@No?NT@2n?GW@9k?@Z@?i?\\O[@c0g?mNZ_O0P1R1i?eN__O8h0S1W`0kNi_OT1Y`0kNg_OU1[`0jNd_OU1^`0jNc_OU1_`0jN`_OV1a`0lN\\_OS1j`0gNW_OY1ia0N100O10001N100O2O000O2O000O2O0O101O0O2O0N3M3L4M3M3M3M3MnVk0"}, "label": "the boy pitching a baseball"}]}
{"id": 124178, "image": "COCO_train2014_000000124178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag carried by the man in the green shirt\"."}], "task": "refering", "answer_template": "The \"the bag carried by the man in the green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 77, 78, 91, 92, 105, 106, 107, 120, 121, 122, 136], "bbox": [0.03892271662763466, 0.187015625, 0.3192271662763466, 0.415578125], "iscrowd": 0, "area": 3530.0964499999995, "rle": {"size": [640, 427], "counts": "gk:4gc05K6J5K5L4L4N2N3N1N2O1N2N3N1N2O1O1O2O2M2O2M3N1N1O2O0O2N1O2N1O00N2N3N1N2N2N2N2N100N3L3M3L5L3M3L5N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1100O101N10001N101N100O2O0010O1O100O1O010O100O1O10O0100O1O100O010O1O100O1O10O01O1O100O1O001O100O1O1O1O010O1O1OY\\f5"}, "label": "the bag carried by the man in the green shirt"}]}
{"id": 525589, "image": "COCO_train2014_000000525589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the left\"."}], "task": "refering", "answer_template": "The \"the man on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208], "bbox": [0.0016875000000000002, 0.0, 0.12276562499999999, 0.5515625], "iscrowd": 0, "area": 12579.45085, "rle": {"size": [480, 640], "counts": "ec0[3P7e401O00000000001O000000000010O000000000001O00000000001O00000000001O0000lG^Ij7b6VHhIe5FlKb6_NRJP5H`LV6`N\\J\\4HTMl5`NfJg3IjM`5_NQKS3I^NV5_N[K^2KROj4`NeKj1KFZ7FPI:Z7QORIn0`:1O01O00ZIgNEY1;nN^OR1b0VOWOj0h0]OQOc0o0DjN:X1NaNOa15[NIg17YNFj1;VNBl1?SN_Oo1a0QN\\OR2e0nMXOT2h0lMVOV2k0iMSO[2k0eMRO`2m0`MPOc2o0]MoNg2P1XMmNm2R1SMkNQ3S1oLkNU3T1jLjNZ3S1hLiN]3P1]5J7J6I7I[]W8"}, "label": "the man on the left"}]}
{"id": 525589, "image": "COCO_train2014_000000525589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someones legs wearing olive green pants and black shoes standing to the left of an animal\"."}], "task": "refering", "answer_template": "The \"someones legs wearing olive green pants and black shoes standing to the left of an animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208], "bbox": [0.0016875000000000002, 0.0, 0.12276562499999999, 0.5515625], "iscrowd": 0, "area": 12579.45085, "rle": {"size": [480, 640], "counts": "ec0[3P7e401O00000000001O000000000010O000000000001O00000000001O00000000001O0000lG^Ij7b6VHhIe5FlKb6_NRJP5H`LV6`N\\J\\4HTMl5`NfJg3IjM`5_NQKS3I^NV5_N[K^2KROj4`NeKj1KFZ7FPI:Z7QORIn0`:1O01O00ZIgNEY1;nN^OR1b0VOWOj0h0]OQOc0o0DjN:X1NaNOa15[NIg17YNFj1;VNBl1?SN_Oo1a0QN\\OR2e0nMXOT2h0lMVOV2k0iMSO[2k0eMRO`2m0`MPOc2o0]MoNg2P1XMmNm2R1SMkNQ3S1oLkNU3T1jLjNZ3S1hLiN]3P1]5J7J6I7I[]W8"}, "label": "someones legs wearing olive green pants and black shoes standing to the left of an animal"}]}
{"id": 304406, "image": "COCO_train2014_000000304406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an image of a black , tortoiseshell cat in the mirror\"."}], "task": "refering", "answer_template": "The \"an image of a black , tortoiseshell cat in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 76, 96, 97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 146, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238], "bbox": [0.136234375, 0.20108606557377048, 0.364546875, 0.6244672131147542], "iscrowd": 0, "area": 18466.33265, "rle": {"size": [488, 640], "counts": "WhY11S?4L4N2N2O1N2O1O1N2O1O1M3J6I7I7J6O1O1N2O1N2N2N2M3M3O1O1O1N2O100N2M3N2O1O1O1O1M3H9I6D<O1N2J6iMoLWGZO=i3Z8SMQG\\O?c3_8QNQGQ2m8RNQGo1l8X200O1001O2N1O001O1O1O1O1O1O2N1O2N1N2O2N1O1O1O1O1O1O0000O100000000000000000000000000000000O100O100O100000000O1O1O1O1O1O1O1O1O1O1N2O1N2O1O1ObLRGd0m8]OTGb0k8WOUGbM1V3i8ZOVG`M2T3g8]OXG`M0c06\\1a8b0ZG_MO9a0d1U8f0[G]M8L>n1Q8<WGmM2NV1f1P9kNmE\\OI044X1b1n9ZNgD034Z1_1l9]NgD134_1W1g9dNhD124`1U1f9fNiD114`1T1f9gNiD214`1Q1f9iNjD3O3b1o0e9kNjD4O1h1j0_9QOkD5NNT2`0S9]OmD5KMX2>Q9_OnD5JK]2<k8DPE5GJ`2;j8EQE4[26e8FQE1\\27d8Ig30O010O1ohQ6"}, "label": "an image of a black , tortoiseshell cat in the mirror"}]}
{"id": 304406, "image": "COCO_train2014_000000304406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cats reflection\"."}], "task": "refering", "answer_template": "The \"the cats reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 76, 96, 97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 146, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238], "bbox": [0.136234375, 0.20108606557377048, 0.364546875, 0.6244672131147542], "iscrowd": 0, "area": 18466.33265, "rle": {"size": [488, 640], "counts": "WhY11S?4L4N2N2O1N2O1O1N2O1O1M3J6I7I7J6O1O1N2O1N2N2N2M3M3O1O1O1N2O100N2M3N2O1O1O1O1M3H9I6D<O1N2J6iMoLWGZO=i3Z8SMQG\\O?c3_8QNQGQ2m8RNQGo1l8X200O1001O2N1O001O1O1O1O1O1O2N1O2N1N2O2N1O1O1O1O1O1O0000O100000000000000000000000000000000O100O100O100000000O1O1O1O1O1O1O1O1O1O1N2O1N2O1O1ObLRGd0m8]OTGb0k8WOUGbM1V3i8ZOVG`M2T3g8]OXG`M0c06\\1a8b0ZG_MO9a0d1U8f0[G]M8L>n1Q8<WGmM2NV1f1P9kNmE\\OI044X1b1n9ZNgD034Z1_1l9]NgD134_1W1g9dNhD124`1U1f9fNiD114`1T1f9gNiD214`1Q1f9iNjD3O3b1o0e9kNjD4O1h1j0_9QOkD5NNT2`0S9]OmD5KMX2>Q9_OnD5JK]2<k8DPE5GJ`2;j8EQE4[26e8FQE1\\27d8Ig30O010O1ohQ6"}, "label": "the cats reflection"}]}
{"id": 304406, "image": "COCO_train2014_000000304406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat\"."}], "task": "refering", "answer_template": "The \"a black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 88, 89, 106, 107, 108, 110, 111, 112, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6427968749999999, 0.20860655737704917, 1.0, 0.9698975409836066], "iscrowd": 0, "area": 61657.96425000001, "rle": {"size": [488, 640], "counts": "o^T68k>7iCCZ9e0ZFF_9b0UFJg99nE3o90hE9U:JcE>Z:F\\Eb0b:BUEf0h:^OnDk0o:j1M3M2N3M1O1O2N1O101O000O2jKWKjMi4k1aKVN_4_1kKaNW4S1SLnNS4a0WL@P4NZL3m3[O]Le0S91O1O001O1O1O001O1O001O1O1O001O1O001O1O0010O01O1O001O1O1O2N1O1O1O2N1O0000001O00000000001O0000000000001O00000000001O00000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2O1N2N2N2N2N2O10000O100O10000O100O10000O100O100O1O1O100O1O1N2N2M3O1100O1O001O1O1O1O1O001O1O3M4L4L4L3M4L4Lm0SO1O2N2N2N1O2N2N=C=C;E4L3M4L3M4L6J8H7I7I4L2N2N2N3M2N2N2N2N2N3M2N2N1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1J6G9H9E:E_1bNhD"}, "label": "a black cat"}]}
{"id": 304406, "image": "COCO_train2014_000000304406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat staring at its own reflection\"."}], "task": "refering", "answer_template": "The \"a cat staring at its own reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 88, 89, 106, 107, 108, 110, 111, 112, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6427968749999999, 0.20860655737704917, 1.0, 0.9698975409836066], "iscrowd": 0, "area": 61657.96425000001, "rle": {"size": [488, 640], "counts": "o^T68k>7iCCZ9e0ZFF_9b0UFJg99nE3o90hE9U:JcE>Z:F\\Eb0b:BUEf0h:^OnDk0o:j1M3M2N3M1O1O2N1O101O000O2jKWKjMi4k1aKVN_4_1kKaNW4S1SLnNS4a0WL@P4NZL3m3[O]Le0S91O1O001O1O1O001O1O001O1O1O001O1O001O1O0010O01O1O001O1O1O2N1O1O1O2N1O0000001O00000000001O0000000000001O00000000001O00000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2O1N2N2N2N2N2O10000O100O10000O100O10000O100O100O1O1O100O1O1N2N2M3O1100O1O001O1O1O1O1O001O1O3M4L4L4L3M4L4Lm0SO1O2N2N2N1O2N2N=C=C;E4L3M4L3M4L6J8H7I7I4L2N2N2N3M2N2N2N2N2N3M2N2N1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1J6G9H9E:E_1bNhD"}, "label": "a cat staring at its own reflection"}]}
{"id": 525591, "image": "COCO_train2014_000000525591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mountain goat with a very shaggy coat\"."}], "task": "refering", "answer_template": "The \"a mountain goat with a very shaggy coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 283, 284, 285, 286, 306, 307, 308, 309], "bbox": [0.28907812499999996, 0.519625, 0.53640625, 0.8108541666666667], "iscrowd": 0, "area": 11877.996500000001, "rle": {"size": [480, 640], "counts": "RPg23l>5K6J6J5J7J6J5K6J5K6J3M4L4L5J6K5Ko0QO00O100[OmD^MS;`2UEYMl:e2\\EUMc:j2j02N2N2N2N2O1O1O100O1O100O1O2O1N1O2O1NVOeDQN[;l1nDmMR;Q2UEkMk:R2WEoMh:o1[EPNe:o1\\EoMe:P2^EmMd:Q2_ElMb:T2_EjMb:V2`EhM`:X2T10000000000000000010O0000000000000000000000000000000000O1000001`0_O;E1O1O1O1O1O1O1O2N1O1NN2N3TOk0O100O2O01O0000010O0000100O001O100O100O10O0100O100O100O1O10O002N2N2N1O2N2TO^CWOd<g0_CVOc<h0`CUOa<j0bCTO_<i0eCQO`<m0k0N2N1O2N2N2N2N1N3N2M3N2M3N3L3N2ManZ4"}, "label": "a mountain goat with a very shaggy coat"}]}
{"id": 525591, "image": "COCO_train2014_000000525591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woolly animal looking at the camera\"."}], "task": "refering", "answer_template": "The \"a woolly animal looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 283, 284, 285, 286, 306, 307, 308, 309], "bbox": [0.28907812499999996, 0.519625, 0.53640625, 0.8108541666666667], "iscrowd": 0, "area": 11877.996500000001, "rle": {"size": [480, 640], "counts": "RPg23l>5K6J6J5J7J6J5K6J5K6J3M4L4L5J6K5Ko0QO00O100[OmD^MS;`2UEYMl:e2\\EUMc:j2j02N2N2N2N2O1O1O100O1O100O1O2O1N1O2O1NVOeDQN[;l1nDmMR;Q2UEkMk:R2WEoMh:o1[EPNe:o1\\EoMe:P2^EmMd:Q2_ElMb:T2_EjMb:V2`EhM`:X2T10000000000000000010O0000000000000000000000000000000000O1000001`0_O;E1O1O1O1O1O1O1O2N1O1NN2N3TOk0O100O2O01O0000010O0000100O001O100O100O10O0100O100O100O1O10O002N2N2N1O2N2TO^CWOd<g0_CVOc<h0`CUOa<j0bCTO_<i0eCQO`<m0k0N2N1O2N2N2N2N1N3N2M3N2M3N3L3N2ManZ4"}, "label": "a woolly animal looking at the camera"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man ' s skiis\"."}], "task": "refering", "answer_template": "The \"the man ' s skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 290, 291, 292, 293, 311, 312, 313, 314], "bbox": [0.512734375, 0.7592271662763466, 0.79778125, 0.9342857142857143], "iscrowd": 0, "area": 1620.007199999999, "rle": {"size": [427, 640], "counts": "WbY41Y=2O1O0O1000O01000O01000O01000O01000O010000O01000O010000O0100O01000O010O10O01000O0100O0_CKn;5QDLo;4QDMm;3SDNm;2RDOm;1TDNm;2RDOn;1QD0n;0SD0m;0RD0n;1RDOn;0RD1m;0SDOn;0RD1n;ORD0n;1QD0o;ORD0n;1QD0o;0QDOP<0PD1o;0QDOP<1oC0P<0QDOP<1oC0Q<OPD0P<1oC0Q<0oCOQ<1oC0Q<0oCOQ<2nCOR<0oCOR<1mC0R<1nCNS<1mC0S<0mCOT<0lC1S<0mCOT<1kC0U<OlC0U<0jC1U<0kCOV<0jC1V<OjC0V<1iC0W<OjC0W<0hC1X<NiC1W<0hC1X<OhC0Y<OgC2Y<NgC1Y<OgC2Y<NgC2Y<NfC2Z<NgC2Y<NfC2[<MfC3Z<MeC3[<NeC2[<MfC3Z<MeC3\\<LeC4Z<MeC3\\<MdC3[<MeC3\\<MdC3[<NdC3\\<LeC3\\<McC4\\<LeC3\\<McC4\\<MdC2]<McC4\\<MdC3\\<McC3]<MdC3\\<McC3]<MdC3i<010O10O010O10O010000O010O100O01000O10O0100O10O10OCLeC4Z<MfC2[<NdC3\\<MdC3[<NeC1\\<OdC1\\<OcC1]<0cC0]<0bC0_<0aC0^<1aC0_<0aCO`<1_C0`<0aCO`<1_C0a<O`C0a<0_C0`<1_C0a<O`C0a<0^C1a<O`C0a<0^C1b<N_C1a<0^C1b<N_C2a<N^C2b<O^C1b<N_C1b<O]C2b<N_C1b<O]C2b<N_C2`<O_C1d<L]C4l<05Lh^e1"}, "label": "the man ' s skiis"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skiis that the man has on\"."}], "task": "refering", "answer_template": "The \"the skiis that the man has on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 290, 291, 292, 293, 311, 312, 313, 314], "bbox": [0.512734375, 0.7592271662763466, 0.79778125, 0.9342857142857143], "iscrowd": 0, "area": 1620.007199999999, "rle": {"size": [427, 640], "counts": "WbY41Y=2O1O0O1000O01000O01000O01000O01000O010000O01000O010000O0100O01000O010O10O01000O0100O0_CKn;5QDLo;4QDMm;3SDNm;2RDOm;1TDNm;2RDOn;1QD0n;0SD0m;0RD0n;1RDOn;0RD1m;0SDOn;0RD1n;ORD0n;1QD0o;ORD0n;1QD0o;0QDOP<0PD1o;0QDOP<1oC0P<0QDOP<1oC0Q<OPD0P<1oC0Q<0oCOQ<1oC0Q<0oCOQ<2nCOR<0oCOR<1mC0R<1nCNS<1mC0S<0mCOT<0lC1S<0mCOT<1kC0U<OlC0U<0jC1U<0kCOV<0jC1V<OjC0V<1iC0W<OjC0W<0hC1X<NiC1W<0hC1X<OhC0Y<OgC2Y<NgC1Y<OgC2Y<NgC2Y<NfC2Z<NgC2Y<NfC2[<MfC3Z<MeC3[<NeC2[<MfC3Z<MeC3\\<LeC4Z<MeC3\\<MdC3[<MeC3\\<MdC3[<NdC3\\<LeC3\\<McC4\\<LeC3\\<McC4\\<MdC2]<McC4\\<MdC3\\<McC3]<MdC3\\<McC3]<MdC3i<010O10O010O10O010000O010O100O01000O10O0100O10O10OCLeC4Z<MfC2[<NdC3\\<MdC3[<NeC1\\<OdC1\\<OcC1]<0cC0]<0bC0_<0aC0^<1aC0_<0aCO`<1_C0`<0aCO`<1_C0a<O`C0a<0_C0`<1_C0a<O`C0a<0^C1a<O`C0a<0^C1b<N_C1a<0^C1b<N_C2a<N^C2b<O^C1b<N_C1b<O]C2b<N_C1b<O]C2b<N_C2`<O_C1d<L]C4l<05Lh^e1"}, "label": "the skiis that the man has on"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing all black\"."}], "task": "refering", "answer_template": "The \"the man wearing all black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 83, 84, 85, 105, 106, 107, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199, 200, 221, 222, 223, 244, 245, 246, 267, 268, 269, 270, 290, 291, 292, 293], "bbox": [0.577078125, 0.17585480093676817, 0.7625, 0.8563700234192037], "iscrowd": 0, "area": 14932.157199999994, "rle": {"size": [427, 640], "counts": "d]j44V=4M3L4L4M3L4L3N2M4L3N2M3N3L3M3^EbNS9`1cFiN[9[1ZFmNe9V1PFSOP:T2O2N100O1O3K4K6K5M3L4M2M4L4M3L4PIoKa4S4nJeLi4^3jJVMm4l2kJ_MP5c2iJdMU5^2cJjM[5U5O1N2N2N2L4K3M2N2N2000bNWK[Ji4d5bKRJ^4n5lKhIT4W6oKgIQ4X6RLfIn3Y6ULeIk3Z6XLdIf3]6]LaIb3_6aL_I^3a6eL]I\\3a6gL]IZ3a6iL]IY3`6iL_IX3_6jL`IW3^6jLbIW3[6lLdIV3Y6lLfIU3Y1SKROe0W3U1^LT3W1XKmNb0\\3T1_LS3V1[LY2d0`LS3T1[LGbNj1m1[MW3R1\\LEgNn1i6[NaJEmNn1`6[NeJCROR2[6TNdJJVOP2\\6lM`J4WOP2]6cM^J<ZOo1S8RNQHm1MkMa67fIm1TO`NU7CjIY3S6hLoIW3P6hLTJW3j5iLXJW3f5iL]JU3b5jLbJU3\\5kLfJT3Y5kLkJT3S5lLPKR3o4nLSKP3m4oLWKm2i4TMYKj2g4VM\\Kg2d4XM`Ke2`4[MbKc2g4SM_Kh2g7K5K5K5iNREGS;E]E:l;N1O1O1O2N1O1O1O00001O00000L4M4K4MhQo1"}, "label": "the man wearing all black"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy wearing a black jacket looking at the camera\"."}], "task": "refering", "answer_template": "The \"a guy wearing a black jacket looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 83, 84, 85, 105, 106, 107, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199, 200, 221, 222, 223, 244, 245, 246, 267, 268, 269, 270, 290, 291, 292, 293], "bbox": [0.577078125, 0.17585480093676817, 0.7625, 0.8563700234192037], "iscrowd": 0, "area": 14932.157199999994, "rle": {"size": [427, 640], "counts": "d]j44V=4M3L4L4M3L4L3N2M4L3N2M3N3L3M3^EbNS9`1cFiN[9[1ZFmNe9V1PFSOP:T2O2N100O1O3K4K6K5M3L4M2M4L4M3L4PIoKa4S4nJeLi4^3jJVMm4l2kJ_MP5c2iJdMU5^2cJjM[5U5O1N2N2N2L4K3M2N2N2000bNWK[Ji4d5bKRJ^4n5lKhIT4W6oKgIQ4X6RLfIn3Y6ULeIk3Z6XLdIf3]6]LaIb3_6aL_I^3a6eL]I\\3a6gL]IZ3a6iL]IY3`6iL_IX3_6jL`IW3^6jLbIW3[6lLdIV3Y6lLfIU3Y1SKROe0W3U1^LT3W1XKmNb0\\3T1_LS3V1[LY2d0`LS3T1[LGbNj1m1[MW3R1\\LEgNn1i6[NaJEmNn1`6[NeJCROR2[6TNdJJVOP2\\6lM`J4WOP2]6cM^J<ZOo1S8RNQHm1MkMa67fIm1TO`NU7CjIY3S6hLoIW3P6hLTJW3j5iLXJW3f5iL]JU3b5jLbJU3\\5kLfJT3Y5kLkJT3S5lLPKR3o4nLSKP3m4oLWKm2i4TMYKj2g4VM\\Kg2d4XM`Ke2`4[MbKc2g4SM_Kh2g7K5K5K5iNREGS;E]E:l;N1O1O1O2N1O1O1O00001O00000L4M4K4MhQo1"}, "label": "a guy wearing a black jacket looking at the camera"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier in the red jacket\"."}], "task": "refering", "answer_template": "The \"the skier in the red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 110, 111, 112, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319, 339, 340, 341, 342], "bbox": [0.74065625, 0.24718969555035128, 0.9445625, 0.9887587822014051], "iscrowd": 0, "area": 22905.82119999999, "rle": {"size": [427, 640], "counts": "^kU64P=9G8K5O100O1O1O100O2N100O1O1O100O11O1N2O1O1O1N10dHZOR2f0iMHm19nM0n1OnM5R2LiM8W2IdM;]2D_M`0a2A\\ITOg3_1k2@\\IUOb3_1Q3JhL9V3MaIH_1`0c2XNhLL6a32_No2ZNcLG:k3GXNU3mMTL`0`0B=T4]OQNa3`NcL]O=i6n2mIcL\\O<i6P3mIbL\\O;i6R3mIaL\\O:i6S3nIaLZO9k6U3nI_LYO6n6Z3kI^LYO1R7_3hI^Lo6a3^1O1N2000000000N2M3M3M3M3N2M3K5]Oc0B>C=N2N2N2O1NcJkHh4m7\\LmGn1T8lMQHU2o7dMWH^2i7[M\\Hg2c7RMcHQ3[7iLjHY3U7bLoH`3Q7ZLSIi3k6QLZIQ4P82M2N2N1O1O1O001O1O1O001O1O1O1O1O1O1O1K5G9G9gMZKbKP5R4QLoJY4e4]2D<D<D<M3O1N2N2O2M2O1N3N1N2O1N3N1N5L6I6K6I7G9E;F:E;F:E;Fm[>"}, "label": "the skier in the red jacket"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in red looking back\"."}], "task": "refering", "answer_template": "The \"a woman in red looking back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 110, 111, 112, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319, 339, 340, 341, 342], "bbox": [0.74065625, 0.24718969555035128, 0.9445625, 0.9887587822014051], "iscrowd": 0, "area": 22905.82119999999, "rle": {"size": [427, 640], "counts": "^kU64P=9G8K5O100O1O1O100O2N100O1O1O100O11O1N2O1O1O1N10dHZOR2f0iMHm19nM0n1OnM5R2LiM8W2IdM;]2D_M`0a2A\\ITOg3_1k2@\\IUOb3_1Q3JhL9V3MaIH_1`0c2XNhLL6a32_No2ZNcLG:k3GXNU3mMTL`0`0B=T4]OQNa3`NcL]O=i6n2mIcL\\O<i6P3mIbL\\O;i6R3mIaL\\O:i6S3nIaLZO9k6U3nI_LYO6n6Z3kI^LYO1R7_3hI^Lo6a3^1O1N2000000000N2M3M3M3M3N2M3K5]Oc0B>C=N2N2N2O1NcJkHh4m7\\LmGn1T8lMQHU2o7dMWH^2i7[M\\Hg2c7RMcHQ3[7iLjHY3U7bLoH`3Q7ZLSIi3k6QLZIQ4P82M2N2N1O1O1O001O1O1O001O1O1O1O1O1O1O1K5G9G9gMZKbKP5R4QLoJY4e4]2D<D<D<M3O1N2N2O2M2O1N3N1N2O1N3N1N5L6I6K6I7G9E;F:E;F:E;Fm[>"}, "label": "a woman in red looking back"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person standing in the back not looking at the camera\"."}], "task": "refering", "answer_template": "The \"the person standing in the back not looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327], "bbox": [0.07346875, 0.2741686182669789, 0.253375, 0.9887587822014051], "iscrowd": 0, "area": 22077.134599999994, "rle": {"size": [427, 640], "counts": "Qhc03X=6I7ZCGn;k000O1000O010000O101O1N2O1O1N2O3M7I7H8I7I5K3UFcMb8a2XGlM]8W2]GWNX8m1bG`NS8c1hGdNS8_1gGgNV8Z3M3M3M3M3M2K6K5J6K5J6K5L4L4Mf0YO`1`N1O1O100O1O1O1O1O100O100O10000O1000000O10000O1000000O10000000000000000001O0O2O1O1O1O1O001O1bGnKd02[5Q4oIoKd07W5m3QJoKe0<S5g3VJnKe0c0n4b3YJmKg0h0j4]3\\JnKg0m0f4X3_JmKi0S1a4R3cJnKi0W1^4n2eJmKj0^1Y4[3dKnLT4U3iKTMn3n2oK^Me3f2WLeM_3]2_LmMW3U2fLWNn2m1oL]Ng2e1VMfN`2\\6F7H8I7H8C=B>BnnV6"}, "label": "the person standing in the back not looking at the camera"}]}
{"id": 271641, "image": "COCO_train2014_000000271641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier that is not looking at the camera\"."}], "task": "refering", "answer_template": "The \"a skier that is not looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327], "bbox": [0.07346875, 0.2741686182669789, 0.253375, 0.9887587822014051], "iscrowd": 0, "area": 22077.134599999994, "rle": {"size": [427, 640], "counts": "Qhc03X=6I7ZCGn;k000O1000O010000O101O1N2O1O1N2O3M7I7H8I7I5K3UFcMb8a2XGlM]8W2]GWNX8m1bG`NS8c1hGdNS8_1gGgNV8Z3M3M3M3M3M2K6K5J6K5J6K5L4L4Mf0YO`1`N1O1O100O1O1O1O1O100O100O10000O1000000O10000O1000000O10000000000000000001O0O2O1O1O1O1O001O1bGnKd02[5Q4oIoKd07W5m3QJoKe0<S5g3VJnKe0c0n4b3YJmKg0h0j4]3\\JnKg0m0f4X3_JmKi0S1a4R3cJnKi0W1^4n2eJmKj0^1Y4[3dKnLT4U3iKTMn3n2oK^Me3f2WLeM_3]2_LmMW3U2fLWNn2m1oL]Ng2e1VMfN`2\\6F7H8I7H8C=B>BnnV6"}, "label": "a skier that is not looking at the camera"}]}
{"id": 394527, "image": "COCO_train2014_000000394527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red double decker bus next to the stoplight\"."}], "task": "refering", "answer_template": "The \"red double decker bus next to the stoplight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 122, 123, 124, 125, 140, 141, 142, 143, 160, 161], "bbox": [0.8005599999999999, 0.3887733333333333, 1.0, 0.6472], "iscrowd": 0, "area": 7453.507650000001, "rle": {"size": [375, 500], "counts": "leb49o:k0UOe0D600j0WOh0WO`R1I^lNY1M3000000000000O100000001O3M2N3M2N3M1N2O001O001O00001O001N101O006J5K4L1O00001OJXG]Mg8d26001O000000001O001N2O001O1O00O1O10000001N1000000000000000000O101O00000000000000000O1000001O00000^K"}, "label": "red double decker bus next to the stoplight"}]}
{"id": 394527, "image": "COCO_train2014_000000394527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red color double decker bus is taking turn towards road\"."}], "task": "refering", "answer_template": "The \"a red color double decker bus is taking turn towards road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 122, 123, 124, 125, 140, 141, 142, 143, 160, 161], "bbox": [0.8005599999999999, 0.3887733333333333, 1.0, 0.6472], "iscrowd": 0, "area": 7453.507650000001, "rle": {"size": [375, 500], "counts": "leb49o:k0UOe0D600j0WOh0WO`R1I^lNY1M3000000000000O100000001O3M2N3M2N3M1N2O001O001O00001O001N101O006J5K4L1O00001OJXG]Mg8d26001O000000001O001N2O001O1O00O1O10000001N1000000000000000000O101O00000000000000000O1000001O00000^K"}, "label": "a red color double decker bus is taking turn towards road"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white table in front of a little girl\"."}], "task": "refering", "answer_template": "The \"a white table in front of a little girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 226, 227, 228, 237, 240, 241, 242, 243, 244, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 287, 288, 289, 290, 291, 292, 295, 296, 297, 298, 299, 304, 305, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.6365624999999999, 0.9985613207547169, 1.0], "iscrowd": 0, "area": 31955.46265000001, "rle": {"size": [640, 424], "counts": "d?9Wc0?I800n@SOk:O\\Al0g3XOm:NZAj0i3XOm:OZAh0i3YOm:0YAf0k3ZOl:0ZAe0j3[Ol:0ZAe0j3\\Ok:O[Ae0j3\\Ok:O\\Ad0i3]Ok:O\\Ad0i3]Ok:O]Ad0g3]Ol:O]Ad0g3]Ol:0]Ab0g3^Ol:0]Ab0g3^Ol:0^Ab0e3^Om:0^Ab0f3]Ol:1_Aa0e3^Ol:1_Aa0e3^Ol:1_Ab0d3]Om:1`Aa0c3^Om:2_A`0d3^Om:2`A?b3@n:1`A`0a3_Oo:1aA?_3AP;0aA?_3AP;0bA>]3CQ;ObA?[3CS;NbA?[3CS;NcA>Y3ET;NbA=Z3ET;NcA=W3FV;McA=V3GW;LdA<U3HW;LdA=S3HX;LfA>o2FZ;MgA`0k2D\\;NjA`0g2B^;OkAb0c2@`;1lAb0`2^Oc;1nAa0^2^Oc;2oA`0]2_Ob;3RB=\\2@a;4SB<[2A`;5WB9W2Ci:CPCb035T2Fc:0mB7<3S2G\\:Z1bCnNR2HR:e1kCdNQ2HU:1WCj0c0]OP2IV:b1jCeNP2IV:b1kCdNn1KW:a1kCeNm1JX:a1kCeNl1KY:`1kCeNl1KY:`1lCdNj1MZ:NWCi0d0\\Oj1N\\:LVCj0c0^Oj1L]:LVCj0c0^Oi1M^:]1hCgNj1L^:]1gChNj1L_:\\1fCjNi1L`:Z1fCkNj1K`:IWCh0?Di1Lb:GVCi0?Di1Lb:GVCj0>Dh1Ld:FVCj0>Dg1Me:W1dClNg1Me:W1dClNf1Nf:V1dCmNe1Mg:EVCh0>Fd1Ni:CUCi0?Ec1Oi:CUCi0?Eb10j:BUCi0?Ea11k:AUCj0>Ea10l:AUCj0`0C^13m:AUCh0b0B\\15m:AUCh0e0_OX19n:@UCh0h0]OS1<Q;^OTCi0k0ZOP1?Q;^OTCi0l0YOn0a0R;]OTCi0l0YOn0a0R;]OTCi0m0YOk0b0T;\\OUCi0k0YOl0b0T;]OTCh0l0YOk0c0U;\\OTCh0m0XOi0e0W;ZOSCi0m0YOh0d0X;ZOSCi0m0YOg0e0Y;YOSCi0m0YOg0e0Y;YOSCi0m0YOf0f0Z;XOTCh0k0[Of0f0[;WOTCi0j0[Of0e0\\;WOTCi0j0[Oe0f0^;VORCi0k0[Oe0f0^;VORCi0j0\\Oe0f0_;UORCi0j0]Oc0f0a;TORCi0j0]Oc0f0a;TOSCh0h0_Oc0f0b;SOSCh0h0_Oc0f0b;SOSCh0g0Ab0e0d;gNRCN1V1e0Ad0d0e;cNXCNJZ1e0Ac0e0f;_N_CNB]1e0Bc0e0g;[NdCO]O_1d0Cd0d0`<HlBEb0d0d<EiBHc0c0f<CfBKc0c0j<_ObBOc0c0m<\\O`B2b0b0P=[O\\B4c0b0S=XOYB7d0a0U=VOWB9c0b0X=SOTB=b0a0\\=POQB`0c0`0^=nNoAb0b0a0b=jNkAf0c0`0d=hNhAj0b0`0g=eNfAk0c0`0i=cNcAn0c0`0j=bNcAn0b0a0k=aNcAn0b0a0k=aNdAn0?b0m=`NdAn0?b0m=`NdAn0>c0n=_NdAn0=d0o=^NdAo0<c0Q>^NbAo0<d0R>]NbAo0<d0R>]NbAo0;e0S>\\NcAo09e0T>\\NcAo08f0U>[NcAo07g0V>ZNcAo07g0V>ZNcAP15g0X>YNcAP15g0X>ZNbAo05h0Y>YNcAn03j0[>^N[Ah0:j0[>^N[Ai08j0]>]N[Ai08j0]>^NZAh08k0^>]NZAh07kN6T1Y>YOZAi06iN^1OR=?ZAi05iNd`0>X_Oh04hNf`0`0V_Oh03hNh`0a0T_Oh03fNj`0b0S_Oh02eNn`0b0P_Oi01eNPa0b0o^Oi01dNQa0c0n^Oi00cNTa0d0l^Oj0OaNVa0e0l^On0Va0SOi^Ol0Xa0TOh^Ok0Ya0UOg^Oi0[a0WOe^Oh0\\a0XOd^Og0]a0YOc^Oe0`a0ZOa^Od0`a0]O_^Ob0ba0^O^^O`0da0@\\^O?ea0A[^O?ea0A[^O?ea0A[^O>fa0BZ^O>fa0CZ^O;ga0UO]^OMLn0ga0oNd^O2Eo0\\b0POd]Oo0]b0QOc]Oo0^b0POb]OP1_b0oNa]OP1`b0QO_]Oo0bb0PO_]On0cb0QO]]Oo0cb0QO]]Oo0db0PO\\]Oo0fb0POZ]OP1fb0QOY]Oo0hb0POX]Oo0jb0POV]OP1kb0oNV]Oo0nb0000000000001O000000000000001O000000000000001OO10000000000000000000KQ]OYOnb0h05O100O10000O10000O10O0100000000O10000000O100000000000000000000O100000000000000000000O101O0000000O1O1O100O2N1O1O10000O100O2O000O10000O100O101O0O100O10000O101O0O100O10000O101O0O100O10000O1O101N1O1O1O1O100O1O2N1O100O1O1O1O1O2O0O1O1N2O1O1O1N3N1O1N2N2N2N2O2M2LTMR_On2n`020000000001O01O0000000001O0001O0000000001O0001O00000001O000001O00000001O000001O00000001O000001O000001O00000010O0001O0000001O01O01O0000001O00019F:FUNh7"}, "label": "a white table in front of a little girl"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that the girl is sitting at\"."}], "task": "refering", "answer_template": "The \"the table that the girl is sitting at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 226, 227, 228, 237, 240, 241, 242, 243, 244, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 287, 288, 289, 290, 291, 292, 295, 296, 297, 298, 299, 304, 305, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.6365624999999999, 0.9985613207547169, 1.0], "iscrowd": 0, "area": 31955.46265000001, "rle": {"size": [640, 424], "counts": "d?9Wc0?I800n@SOk:O\\Al0g3XOm:NZAj0i3XOm:OZAh0i3YOm:0YAf0k3ZOl:0ZAe0j3[Ol:0ZAe0j3\\Ok:O[Ae0j3\\Ok:O\\Ad0i3]Ok:O\\Ad0i3]Ok:O]Ad0g3]Ol:O]Ad0g3]Ol:0]Ab0g3^Ol:0]Ab0g3^Ol:0^Ab0e3^Om:0^Ab0f3]Ol:1_Aa0e3^Ol:1_Aa0e3^Ol:1_Ab0d3]Om:1`Aa0c3^Om:2_A`0d3^Om:2`A?b3@n:1`A`0a3_Oo:1aA?_3AP;0aA?_3AP;0bA>]3CQ;ObA?[3CS;NbA?[3CS;NcA>Y3ET;NbA=Z3ET;NcA=W3FV;McA=V3GW;LdA<U3HW;LdA=S3HX;LfA>o2FZ;MgA`0k2D\\;NjA`0g2B^;OkAb0c2@`;1lAb0`2^Oc;1nAa0^2^Oc;2oA`0]2_Ob;3RB=\\2@a;4SB<[2A`;5WB9W2Ci:CPCb035T2Fc:0mB7<3S2G\\:Z1bCnNR2HR:e1kCdNQ2HU:1WCj0c0]OP2IV:b1jCeNP2IV:b1kCdNn1KW:a1kCeNm1JX:a1kCeNl1KY:`1kCeNl1KY:`1lCdNj1MZ:NWCi0d0\\Oj1N\\:LVCj0c0^Oj1L]:LVCj0c0^Oi1M^:]1hCgNj1L^:]1gChNj1L_:\\1fCjNi1L`:Z1fCkNj1K`:IWCh0?Di1Lb:GVCi0?Di1Lb:GVCj0>Dh1Ld:FVCj0>Dg1Me:W1dClNg1Me:W1dClNf1Nf:V1dCmNe1Mg:EVCh0>Fd1Ni:CUCi0?Ec1Oi:CUCi0?Eb10j:BUCi0?Ea11k:AUCj0>Ea10l:AUCj0`0C^13m:AUCh0b0B\\15m:AUCh0e0_OX19n:@UCh0h0]OS1<Q;^OTCi0k0ZOP1?Q;^OTCi0l0YOn0a0R;]OTCi0l0YOn0a0R;]OTCi0m0YOk0b0T;\\OUCi0k0YOl0b0T;]OTCh0l0YOk0c0U;\\OTCh0m0XOi0e0W;ZOSCi0m0YOh0d0X;ZOSCi0m0YOg0e0Y;YOSCi0m0YOg0e0Y;YOSCi0m0YOf0f0Z;XOTCh0k0[Of0f0[;WOTCi0j0[Of0e0\\;WOTCi0j0[Oe0f0^;VORCi0k0[Oe0f0^;VORCi0j0\\Oe0f0_;UORCi0j0]Oc0f0a;TORCi0j0]Oc0f0a;TOSCh0h0_Oc0f0b;SOSCh0h0_Oc0f0b;SOSCh0g0Ab0e0d;gNRCN1V1e0Ad0d0e;cNXCNJZ1e0Ac0e0f;_N_CNB]1e0Bc0e0g;[NdCO]O_1d0Cd0d0`<HlBEb0d0d<EiBHc0c0f<CfBKc0c0j<_ObBOc0c0m<\\O`B2b0b0P=[O\\B4c0b0S=XOYB7d0a0U=VOWB9c0b0X=SOTB=b0a0\\=POQB`0c0`0^=nNoAb0b0a0b=jNkAf0c0`0d=hNhAj0b0`0g=eNfAk0c0`0i=cNcAn0c0`0j=bNcAn0b0a0k=aNcAn0b0a0k=aNdAn0?b0m=`NdAn0?b0m=`NdAn0>c0n=_NdAn0=d0o=^NdAo0<c0Q>^NbAo0<d0R>]NbAo0<d0R>]NbAo0;e0S>\\NcAo09e0T>\\NcAo08f0U>[NcAo07g0V>ZNcAo07g0V>ZNcAP15g0X>YNcAP15g0X>ZNbAo05h0Y>YNcAn03j0[>^N[Ah0:j0[>^N[Ai08j0]>]N[Ai08j0]>^NZAh08k0^>]NZAh07kN6T1Y>YOZAi06iN^1OR=?ZAi05iNd`0>X_Oh04hNf`0`0V_Oh03hNh`0a0T_Oh03fNj`0b0S_Oh02eNn`0b0P_Oi01eNPa0b0o^Oi01dNQa0c0n^Oi00cNTa0d0l^Oj0OaNVa0e0l^On0Va0SOi^Ol0Xa0TOh^Ok0Ya0UOg^Oi0[a0WOe^Oh0\\a0XOd^Og0]a0YOc^Oe0`a0ZOa^Od0`a0]O_^Ob0ba0^O^^O`0da0@\\^O?ea0A[^O?ea0A[^O?ea0A[^O>fa0BZ^O>fa0CZ^O;ga0UO]^OMLn0ga0oNd^O2Eo0\\b0POd]Oo0]b0QOc]Oo0^b0POb]OP1_b0oNa]OP1`b0QO_]Oo0bb0PO_]On0cb0QO]]Oo0cb0QO]]Oo0db0PO\\]Oo0fb0POZ]OP1fb0QOY]Oo0hb0POX]Oo0jb0POV]OP1kb0oNV]Oo0nb0000000000001O000000000000001O000000000000001OO10000000000000000000KQ]OYOnb0h05O100O10000O10000O10O0100000000O10000000O100000000000000000000O100000000000000000000O101O0000000O1O1O100O2N1O1O10000O100O2O000O10000O100O101O0O100O10000O101O0O100O10000O101O0O100O10000O1O101N1O1O1O1O100O1O2N1O100O1O1O1O1O2O0O1O1N2O1O1O1N3N1O1N2N2N2N2O2M2LTMR_On2n`020000000001O01O0000000001O0001O0000000001O0001O00000001O000001O00000001O000001O00000001O000001O000001O00000010O0001O0000001O01O01O0000001O00019F:FUNh7"}, "label": "the table that the girl is sitting at"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl with her tongue sticking out , holding up a fork full of noodles\"."}], "task": "refering", "answer_template": "The \"a young girl with her tongue sticking out , holding up a fork full of noodles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 35, 36, 37, 50, 51, 52, 53, 54, 55, 56, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 252, 253, 254], "bbox": [0.21660377358490568, 0.0765, 0.9476179245283018, 0.715515625], "iscrowd": 0, "area": 81145.10289999998, "rle": {"size": [640, 424], "counts": "eei1m2Ra05J6K5J7J5J6K5J6K5J6J6K2M3N2M3N2M4M2M3N2M3hC\\JQ:f5jE^JT:e5gE_JX:d5bE`J]:b5]EcJb:_5YEeJf:]5UEgJj:[5QEiJm:Z5mDkJR;W5iDmJV;U5eDnJ[;U5_DoJ`;S5ZDRKe;e6O1O1O1O1O1O1O1O1O1N2O1O1O1O100O1O010O100O1O100O100O1O100N2L4L4L4M3L4ROiFoF[9P9k00O01O100O100O1O100O100O1O100O1O100001O001N101O00001O001O00001O001O001O101N1O100O2O0O100O1O2O0O100O2N100O3N8G10O00001O001O001O00001O001O001O00000000O100O10000O100O100O100O100O100000000O100000000O10000000000O100000000O100001O100O2N1O100O1O2N100O1O1O1O2O0O1O1O101N1O1O100O001O010O001O010O001O010O001O010O001O0010O002N2O1N3M2N2O2M2N2N2O2M2nIiEW2Y:cMmE[2U:_MQF`2Q:ZMSFe2o9UMWFi2k9QMZFn2i9kL]FT3d9fLaFY3a9bLdF\\3^9^LhF`3[9YLjFg3W9SLoFk3S9PLQGo3Q9kKUGS4n8fKWGZ4j8`K\\G^4f8]K_Ga4d8XKaGg4a8SKeGl4\\8nJiGQ5e;N1O2O1N1O2N101N1O2N2O2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M6KS1lNQZ="}, "label": "a young girl with her tongue sticking out , holding up a fork full of noodles"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl holding up a fork\"."}], "task": "refering", "answer_template": "The \"a little girl holding up a fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 35, 36, 37, 50, 51, 52, 53, 54, 55, 56, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 252, 253, 254], "bbox": [0.21660377358490568, 0.0765, 0.9476179245283018, 0.715515625], "iscrowd": 0, "area": 81145.10289999998, "rle": {"size": [640, 424], "counts": "eei1m2Ra05J6K5J7J5J6K5J6K5J6J6K2M3N2M3N2M4M2M3N2M3hC\\JQ:f5jE^JT:e5gE_JX:d5bE`J]:b5]EcJb:_5YEeJf:]5UEgJj:[5QEiJm:Z5mDkJR;W5iDmJV;U5eDnJ[;U5_DoJ`;S5ZDRKe;e6O1O1O1O1O1O1O1O1O1N2O1O1O1O100O1O010O100O1O100O100O1O100N2L4L4L4M3L4ROiFoF[9P9k00O01O100O100O1O100O100O1O100O1O100001O001N101O00001O001O00001O001O001O101N1O100O2O0O100O1O2O0O100O2N100O3N8G10O00001O001O001O00001O001O001O00000000O100O10000O100O100O100O100O100000000O100000000O10000000000O100000000O100001O100O2N1O100O1O2N100O1O1O1O2O0O1O1O101N1O1O100O001O010O001O010O001O010O001O010O001O0010O002N2O1N3M2N2O2M2N2N2O2M2nIiEW2Y:cMmE[2U:_MQF`2Q:ZMSFe2o9UMWFi2k9QMZFn2i9kL]FT3d9fLaFY3a9bLdF\\3^9^LhF`3[9YLjFg3W9SLoFk3S9PLQGo3Q9kKUGS4n8fKWGZ4j8`K\\G^4f8]K_Ga4d8XKaGg4a8SKeGl4\\8nJiGQ5e;N1O2O1N1O2N101N1O2N2O2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M3N2M2N3M6KS1lNQZ="}, "label": "a little girl holding up a fork"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black chair that the little girl is sitting in\"."}], "task": "refering", "answer_template": "The \"black chair that the little girl is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 177, 178, 179, 193, 194, 208, 209, 224, 239, 254], "bbox": [0.7801650943396227, 0.42403124999999997, 0.9904716981132077, 0.7049375], "iscrowd": 0, "area": 5418.985449999998, "rle": {"size": [640, 424], "counts": "`h_61oc0010O1O100O101N1O101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O1O2O0O2N100O2O0O2N100O2N100O2O0O2N100O2N101N1O101N100O2N101N6J6K5J6K5J6J6K5J6J6K5J6K5J6J6K5J6J10O00001O001O00001O00001O00`V2"}, "label": "black chair that the little girl is sitting in"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the girl eating pasta is on\"."}], "task": "refering", "answer_template": "The \"the chair the girl eating pasta is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 177, 178, 179, 193, 194, 208, 209, 224, 239, 254], "bbox": [0.7801650943396227, 0.42403124999999997, 0.9904716981132077, 0.7049375], "iscrowd": 0, "area": 5418.985449999998, "rle": {"size": [640, 424], "counts": "`h_61oc0010O1O100O101N1O101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O1O2O0O2N100O2O0O2N100O2N100O2O0O2N100O2N101N1O101N100O2N101N6J6K5J6K5J6J6K5J6J6K5J6K5J6J6K5J6J10O00001O001O00001O00001O00`V2"}, "label": "the chair the girl eating pasta is on"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table in the background\"."}], "task": "refering", "answer_template": "The \"the table in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 121, 122, 136, 137, 150, 151, 152, 165, 166, 180, 181], "bbox": [0.0, 0.207625, 0.27752358490566037, 0.5663750000000001], "iscrowd": 0, "area": 10612.750600000001, "rle": {"size": [640, 424], "counts": "Y4h1Wb02N1O1O2O001O00001O00001O00001O000010O01O00001]BUNo8k1`FgN_9Y1PFYOP:g0_EJ`:6oD<P;D_Dn0`;SOoC^1P<bN_Co1a<RNmB`2R=X2O100O1O1O1O00000001O00000O10002N1J6G:G8G:G8G9H9ZMUAKT?3SA@V?>PATOY?j0n@iNZ?U1m@]N]?`1j@SN]?m1[1O1O100O1O1O100O1O1O100O1O00100N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1O1N101N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O00Wgo5"}, "label": "the table in the background"}]}
{"id": 410916, "image": "COCO_train2014_000000410916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table behind a girl to the left sitting eating her lunch\"."}], "task": "refering", "answer_template": "The \"table behind a girl to the left sitting eating her lunch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 121, 122, 136, 137, 150, 151, 152, 165, 166, 180, 181], "bbox": [0.0, 0.207625, 0.27752358490566037, 0.5663750000000001], "iscrowd": 0, "area": 10612.750600000001, "rle": {"size": [640, 424], "counts": "Y4h1Wb02N1O1O2O001O00001O00001O00001O000010O01O00001]BUNo8k1`FgN_9Y1PFYOP:g0_EJ`:6oD<P;D_Dn0`;SOoC^1P<bN_Co1a<RNmB`2R=X2O100O1O1O1O00000001O00000O10002N1J6G:G8G:G8G9H9ZMUAKT?3SA@V?>PATOY?j0n@iNZ?U1m@]N]?`1j@SN]?m1[1O1O100O1O1O100O1O1O100O1O00100N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1O1N101N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O00Wgo5"}, "label": "table behind a girl to the left sitting eating her lunch"}]}
{"id": 345388, "image": "COCO_train2014_000000345388.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall suitcase behind a dog\"."}], "task": "refering", "answer_template": "The \"a tall suitcase behind a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200], "bbox": [0.34040625, 0.04392523364485981, 0.7907500000000001, 0.5461682242990654], "iscrowd": 0, "area": 39453.51619999999, "rle": {"size": [428, 640], "counts": "[Vk22V=4L4L4A>N3O1O1O1O001O1O1O10O01O1O010O1O010O1O010000O10000001O1O1O00000O10000000000000000000001O0000001O00001O00001O00001N10001O00001O00001O00002N1O1O1O1O1O1O1O1O1O00000O101O001QGPN`6Q2^IRN`6o1]IUNa6l1]IWNa6j1]IYNa6h1]I\\N`6e1^I_N_6c1^IaN_6`1_IdN^6]1`IhN\\6Y1aIlN\\6U1bIoN[6R1cIROZ6o0dIVOX6k0fIXOX6i0eIYO[6h0cIYO]6h0aIZO^6g0`IZO`6g0]I\\Ob6e0\\I\\Od6e0YI]Of6e0VI_Oi6b0TI@l6a0PICo6>oHCQ7?mHBR7`0kHBT7`0iHBV7`0gHBV7\\3L4M3M3M3L4N2O0000000000000000000000O101O00000000000000000000000000000000000000001O0000000000000000000000000000000000000O2O001O3M3M4L1OO10000O100O100O11O0000001O00001O0000001O00001O0000001O0000001O00001N1000001O000000000000000000000000000000000000001O000000000O1N2O1O1O1N2O1O1N2O1N2M3L4L4M4K4M4K5K4M4K4M4K5K4K6H7J7I6J7I7I8H8H8H8H8G9H8H8H8HZng1"}, "label": "a tall suitcase behind a dog"}]}
{"id": 345388, "image": "COCO_train2014_000000345388.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown suitcase directly behind the dog\"."}], "task": "refering", "answer_template": "The \"brown suitcase directly behind the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200], "bbox": [0.34040625, 0.04392523364485981, 0.7907500000000001, 0.5461682242990654], "iscrowd": 0, "area": 39453.51619999999, "rle": {"size": [428, 640], "counts": "[Vk22V=4L4L4A>N3O1O1O1O001O1O1O10O01O1O010O1O010O1O010000O10000001O1O1O00000O10000000000000000000001O0000001O00001O00001O00001N10001O00001O00001O00002N1O1O1O1O1O1O1O1O1O00000O101O001QGPN`6Q2^IRN`6o1]IUNa6l1]IWNa6j1]IYNa6h1]I\\N`6e1^I_N_6c1^IaN_6`1_IdN^6]1`IhN\\6Y1aIlN\\6U1bIoN[6R1cIROZ6o0dIVOX6k0fIXOX6i0eIYO[6h0cIYO]6h0aIZO^6g0`IZO`6g0]I\\Ob6e0\\I\\Od6e0YI]Of6e0VI_Oi6b0TI@l6a0PICo6>oHCQ7?mHBR7`0kHBT7`0iHBV7`0gHBV7\\3L4M3M3M3L4N2O0000000000000000000000O101O00000000000000000000000000000000000000001O0000000000000000000000000000000000000O2O001O3M3M4L1OO10000O100O100O11O0000001O00001O0000001O00001O0000001O0000001O00001N1000001O000000000000000000000000000000000000001O000000000O1N2O1O1O1N2O1O1N2O1N2M3L4L4M4K4M4K5K4M4K4M4K5K4K6H7J7I6J7I7I8H8H8H8H8G9H8H8H8HZng1"}, "label": "brown suitcase directly behind the dog"}]}
{"id": 345388, "image": "COCO_train2014_000000345388.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown suitcase next to a dog\"."}], "task": "refering", "answer_template": "The \"brown suitcase next to a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313], "bbox": [0.47896875000000005, 0.550677570093458, 0.8735937500000001, 0.9267990654205607], "iscrowd": 0, "area": 27394.934649999996, "rle": {"size": [428, 640], "counts": "]eP4`0_<a0^Ob0_Oa0_Oa0C=C9H8G8K600000000O1000000000000000000000000000O00100O1O100O1O100O1O100O1O10O01O100O1O100O1O1000000O1000O100000O100000000O100000000O01000000000O100000000O1000O1000O10000000DbLjF^3U9iLeFV3Z9?O1O1O10O10O100000O01000000O10O1000O1000O10O10000000O010000000O10O100000O10000000000O100000001N100000000O100000001O0O101O001O0O2O001O001N101O001O1O3L4M3M2N000O10O100000O100000O1000O100000000O010000000O100000O10000000O1000O100000000000O10O1000000000000O0100M3N2M3N2N2M2O2M3N2N2M3N2N1N3N2M3N2N3L3N2M3N2N4K5K5K5K5K5K5J6JWcQ1"}, "label": "brown suitcase next to a dog"}]}
{"id": 345388, "image": "COCO_train2014_000000345388.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown briefcase\"."}], "task": "refering", "answer_template": "The \"brown briefcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313], "bbox": [0.47896875000000005, 0.550677570093458, 0.8735937500000001, 0.9267990654205607], "iscrowd": 0, "area": 27394.934649999996, "rle": {"size": [428, 640], "counts": "]eP4`0_<a0^Ob0_Oa0_Oa0C=C9H8G8K600000000O1000000000000000000000000000O00100O1O100O1O100O1O100O1O10O01O100O1O100O1O1000000O1000O100000O100000000O100000000O01000000000O100000000O1000O1000O10000000DbLjF^3U9iLeFV3Z9?O1O1O10O10O100000O01000000O10O1000O1000O10O10000000O010000000O10O100000O10000000000O100000001N100000000O100000001O0O101O001O0O2O001O001N101O001O1O3L4M3M2N000O10O100000O100000O1000O100000000O010000000O100000O10000000O1000O100000000000O10O1000000000000O0100M3N2M3N2N2M2O2M3N2N2M3N2N1N3N2M3N2N3L3N2M3N2N4K5K5K5K5K5K5J6JWcQ1"}, "label": "brown briefcase"}]}
{"id": 345388, "image": "COCO_train2014_000000345388.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first suitcase of the picture\"."}], "task": "refering", "answer_template": "The \"the first suitcase of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5019375, 0.7617990654205608, 1.0, 0.9887616822429907], "iscrowd": 0, "area": 19559.386400000003, "rle": {"size": [428, 640], "counts": "moV42X=2O1N2O1N2N101N2N2O1N2O1N2N2O1N2N101N2O1N2N200O1O2N10NmCkNS<U1mCkNR<V1nCjNR<V1nCjNR<V1nCjNR<V1nCjNR<V1nCjNQ<W1oCiNQ<W1oCiNQ<W1oCiNQ<V1oCkNQ<U1oCkNP<V1PDjNP<V1PDjNP<V1PDjNP<V1PDjNP<V1PDjNo;W1QDiNo;W1QDiNo;W1QDiNo;V1RDjNn;V1QDkNn;V1RDjNn;V1RDjNn;V1RDjNn;V1RDjNn;V1RDjNn;V1RDjNm;W1SDiNm;W1SDiNm;W1SDiNm;V1TDjNl;V1SDkNl;V1TDjNl;V1TDjNl;V1TDjNl;V1TDjNl;V1TDjNk;W1UDiNk;W1UDiNk;W1UDiNk;V1VDjNj;V1VDjNi;W1VDjNj;V1VDjNj;V1VDjNj;V1VDjNj;V1VDjNi;W1WDiNi;W1WDiNi;W1WDiNi;V1XDjNh;V1XDjNh;V1XDjNg;W1XDjNh;V1XDjNh;V1XDjNh;V1XDjNh;V1XDjNg;W1YDiNg;W1YDiNg;V1ZDjNf;V1ZDjNf;V1ZDjNe;W1[DiNe;W1[DiNe;W1ZDjNf;V1ZDjNf;V1ZDjNe;W1[DiNe;W1[DiNe;V1\\DkNc;U1]DkNc;U1]DkNb;V1^DjNb;V1^DjNb;V1^DjNb;V1]DkNc;U1]DkNb;V1^DjNb;V1^DjNb;V1^DjNb;U1_DkNa;U1_DkNa;U1_DkN`;V1`DjN`;V1`DjN`;V1`DjN`;V1_DkNa;U1_DkN`;V1`DjN`;V1`DjN`;U1aDkN_;U1aDkN_;U1aDkN^;V1bDjN^;V1bDjN^;V1bDjN^;V1bDjN^;V1aDkN^;V1bDjN^;V1bDjN^;U1cDkN];U1cDkN];U1cDkN\\;V1dDjN\\;V1dDjN\\;V1dDjN\\;V1dDjN\\;V1dDjN\\;V1cDkN\\;V1dDjN\\;U1eDkN[;U1eDkN[;U1eDkN[;U1eDkNZ;V1fDjNZ;V1fDjNZ;V1fDjNZ;V1fDjNZ;V1fDjNY;W1gDiNY;W1fDjNZ;U1gDkNY;U1gDkNY;U1gDkNX;V1hDjNX;V1hDjNX;V1hDjNX;V1hDjNX;V1hDjNW;W1iDiNW;W1iDiNW;V1iDkNW;U1iDkNW;U1iDkNW;U1iDkNV;V1jDjNV;V1jDjNV;V1jDjNV;V1jDjNV;V1jDjNU;W1kDiNU;V1lDjNT;V1kDkNU;U1kDkNU;U1kDkNT;V1lDjNT;V1lDjNT;V1lDjNT;V1lDjNT;V1lDjNS;W1mDiNS;V1nDjNR;V1nDjNR;V1mDkNS;U1mDkNR;V1nDjNR;V1nDjNR;V1nDjNR;V1nDjNR;V1nDjNQ;W1oDiNQ;W1oDiNQ;V1PEjNP;V1PEjNP;V1oDkNQ;U1oDkNP;V1PEjNP;V1PEjNP;V1PEjNP;V1PEjNP;V1PEjNo:W1QEiNo:V1REjNn:V1REjNn:V1REjNn:V1REjNm:W1REjNn:V1REjNn:V1REjNn:V1REjNn:V1REjNm:W1SEiNm:V1TEjNl:V1TEjNl:V1TEjNl:V1TEjNk:W1UEiNk:W1TEjNl:V1TEjNl:V1TEjNl:V1TEjNl:V1TEjNk:V1VEjNj:V1VEjNj:V1VEjNj:V1VEjNj:V1VEjNi:W1WEiNi:W1VEjNj:V1VEjNj:V1VEjNj:V1VEjNi:W1WEiNi:V1XEjNh:V1XEjNh:V1XEjNh:V1XEjNg:W1YEiNg:W1YEiNg:W1XEjNh:V1XEjNh:V1XEjNg:W1YEiNg:V1ZEjNf:V1ZEjNf:V1ZEjNf:V1ZEjNe:W1[EiNe:W1[EiNe:W1[EiNe:W1[EiNe:W1ZEjNf:V1ZEjNe:V1\\EjNd:V1\\EjNd:V1\\EjNd:V1\\EjNd:V1\\EjNc:W1]EiNc:W1]EiNc:W1]EiNc:W1]EiNc:W1\\EjNc:V1^EjNb:V1^EjNb:V1^EjNb:V1^EjNb:V1^EjNa:W1_EiNa:W1_EiNa:W1_EiNa:W1_EiNa:W1_EiN`:X1_EiNa:V1`EjN`:V1`EjN`:V1`EjN`:V1`EjN`:V1`EjN_:W1aEiN_:W1aEiN_:W1aEiN_:W1aEiN_:W1aEiN^:W1bEjN^:V1bEjN^:V1bEjN^:V1bEjN^:V1bEjN]:W1cEhN^:X1bEhN^:X1bEgN_:Y1aEfN`:Z1`EeN`:[1aEdN`:\\1_EdNb:\\1^EcNc:]1]EbNd:^1\\EaNd:`1\\E_Ne:a1[E^Nf:b1ZE]Ng:c1YE\\Nh:d1XE\\Nh:c1YE\\Ng:e1YEZNh:f1XEYNi:g1VEYNk:g1UEXNl:h1TEWNl:j1TEUNm:k1SETNn:l1RESNo:m1QERNP;n1PEQNP;o1QEQNo:l1TESNm:h1XEWNY<4kCL`0"}, "label": "the first suitcase of the picture"}]}
{"id": 345388, "image": "COCO_train2014_000000345388.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a suitcase in the very bottom right - hand corner\"."}], "task": "refering", "answer_template": "The \"a suitcase in the very bottom right - hand corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5019375, 0.7617990654205608, 1.0, 0.9887616822429907], "iscrowd": 0, "area": 19559.386400000003, "rle": {"size": [428, 640], "counts": "moV42X=2O1N2O1N2N101N2N2O1N2O1N2N2O1N2N101N2O1N2N200O1O2N10NmCkNS<U1mCkNR<V1nCjNR<V1nCjNR<V1nCjNR<V1nCjNR<V1nCjNQ<W1oCiNQ<W1oCiNQ<W1oCiNQ<V1oCkNQ<U1oCkNP<V1PDjNP<V1PDjNP<V1PDjNP<V1PDjNP<V1PDjNo;W1QDiNo;W1QDiNo;W1QDiNo;V1RDjNn;V1QDkNn;V1RDjNn;V1RDjNn;V1RDjNn;V1RDjNn;V1RDjNn;V1RDjNm;W1SDiNm;W1SDiNm;W1SDiNm;V1TDjNl;V1SDkNl;V1TDjNl;V1TDjNl;V1TDjNl;V1TDjNl;V1TDjNk;W1UDiNk;W1UDiNk;W1UDiNk;V1VDjNj;V1VDjNi;W1VDjNj;V1VDjNj;V1VDjNj;V1VDjNj;V1VDjNi;W1WDiNi;W1WDiNi;W1WDiNi;V1XDjNh;V1XDjNh;V1XDjNg;W1XDjNh;V1XDjNh;V1XDjNh;V1XDjNh;V1XDjNg;W1YDiNg;W1YDiNg;V1ZDjNf;V1ZDjNf;V1ZDjNe;W1[DiNe;W1[DiNe;W1ZDjNf;V1ZDjNf;V1ZDjNe;W1[DiNe;W1[DiNe;V1\\DkNc;U1]DkNc;U1]DkNb;V1^DjNb;V1^DjNb;V1^DjNb;V1]DkNc;U1]DkNb;V1^DjNb;V1^DjNb;V1^DjNb;U1_DkNa;U1_DkNa;U1_DkN`;V1`DjN`;V1`DjN`;V1`DjN`;V1_DkNa;U1_DkN`;V1`DjN`;V1`DjN`;U1aDkN_;U1aDkN_;U1aDkN^;V1bDjN^;V1bDjN^;V1bDjN^;V1bDjN^;V1aDkN^;V1bDjN^;V1bDjN^;U1cDkN];U1cDkN];U1cDkN\\;V1dDjN\\;V1dDjN\\;V1dDjN\\;V1dDjN\\;V1dDjN\\;V1cDkN\\;V1dDjN\\;U1eDkN[;U1eDkN[;U1eDkN[;U1eDkNZ;V1fDjNZ;V1fDjNZ;V1fDjNZ;V1fDjNZ;V1fDjNY;W1gDiNY;W1fDjNZ;U1gDkNY;U1gDkNY;U1gDkNX;V1hDjNX;V1hDjNX;V1hDjNX;V1hDjNX;V1hDjNW;W1iDiNW;W1iDiNW;V1iDkNW;U1iDkNW;U1iDkNW;U1iDkNV;V1jDjNV;V1jDjNV;V1jDjNV;V1jDjNV;V1jDjNU;W1kDiNU;V1lDjNT;V1kDkNU;U1kDkNU;U1kDkNT;V1lDjNT;V1lDjNT;V1lDjNT;V1lDjNT;V1lDjNS;W1mDiNS;V1nDjNR;V1nDjNR;V1mDkNS;U1mDkNR;V1nDjNR;V1nDjNR;V1nDjNR;V1nDjNR;V1nDjNQ;W1oDiNQ;W1oDiNQ;V1PEjNP;V1PEjNP;V1oDkNQ;U1oDkNP;V1PEjNP;V1PEjNP;V1PEjNP;V1PEjNP;V1PEjNo:W1QEiNo:V1REjNn:V1REjNn:V1REjNn:V1REjNm:W1REjNn:V1REjNn:V1REjNn:V1REjNn:V1REjNm:W1SEiNm:V1TEjNl:V1TEjNl:V1TEjNl:V1TEjNk:W1UEiNk:W1TEjNl:V1TEjNl:V1TEjNl:V1TEjNl:V1TEjNk:V1VEjNj:V1VEjNj:V1VEjNj:V1VEjNj:V1VEjNi:W1WEiNi:W1VEjNj:V1VEjNj:V1VEjNj:V1VEjNi:W1WEiNi:V1XEjNh:V1XEjNh:V1XEjNh:V1XEjNg:W1YEiNg:W1YEiNg:W1XEjNh:V1XEjNh:V1XEjNg:W1YEiNg:V1ZEjNf:V1ZEjNf:V1ZEjNf:V1ZEjNe:W1[EiNe:W1[EiNe:W1[EiNe:W1[EiNe:W1ZEjNf:V1ZEjNe:V1\\EjNd:V1\\EjNd:V1\\EjNd:V1\\EjNd:V1\\EjNc:W1]EiNc:W1]EiNc:W1]EiNc:W1]EiNc:W1\\EjNc:V1^EjNb:V1^EjNb:V1^EjNb:V1^EjNb:V1^EjNa:W1_EiNa:W1_EiNa:W1_EiNa:W1_EiNa:W1_EiN`:X1_EiNa:V1`EjN`:V1`EjN`:V1`EjN`:V1`EjN`:V1`EjN_:W1aEiN_:W1aEiN_:W1aEiN_:W1aEiN_:W1aEiN^:W1bEjN^:V1bEjN^:V1bEjN^:V1bEjN^:V1bEjN]:W1cEhN^:X1bEhN^:X1bEgN_:Y1aEfN`:Z1`EeN`:[1aEdN`:\\1_EdNb:\\1^EcNc:]1]EbNd:^1\\EaNd:`1\\E_Ne:a1[E^Nf:b1ZE]Ng:c1YE\\Nh:d1XE\\Nh:c1YE\\Ng:e1YEZNh:f1XEYNi:g1VEYNk:g1UEXNl:h1TEWNl:j1TEUNm:k1SETNn:l1RESNo:m1QERNP;n1PEQNP;o1QEQNo:l1TESNm:h1XEWNY<4kCL`0"}, "label": "a suitcase in the very bottom right - hand corner"}]}
{"id": 427308, "image": "COCO_train2014_000000427308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver cell phone\"."}], "task": "refering", "answer_template": "The \"a silver cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 135, 157, 158, 180, 181, 203, 204, 271, 272, 294, 295, 317, 318, 341], "bbox": [0.8081406250000001, 0.2482916666666667, 0.9143281250000002, 0.8595208333333333], "iscrowd": 0, "area": 3763.460049999994, "rle": {"size": [480, 640], "counts": "Zeb74k>b0_O`0@`0@3M3M3M3M3M3L4M3M3M3MOVM]MYIc2e6_M[I`2c6dM\\I\\2a6hM^IW2`6lM_IU2^6nMeIn1Y6UNjIg1T6\\NnIa1P6bNSJZ1l5hNWJT1h5nN[Jm0e5UO^Jg0b5ZO`Jc0_5_OdJ=[5EhJ6X5KlJ1S51PKKP56RKFn4<UK@j4b0YKSOm4o0VKdNR5^1W41N2O1O10O1N1O2M3E;E:E<EdNc0SC[Of=P1O1O1O1O1O00100O100O15K5K5K5J6K5K5KVdi0"}, "label": "a silver cell phone"}]}
{"id": 427308, "image": "COCO_train2014_000000427308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and grey cell phone on the right side of a purse\"."}], "task": "refering", "answer_template": "The \"blue and grey cell phone on the right side of a purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 135, 157, 158, 180, 181, 203, 204, 271, 272, 294, 295, 317, 318, 341], "bbox": [0.8081406250000001, 0.2482916666666667, 0.9143281250000002, 0.8595208333333333], "iscrowd": 0, "area": 3763.460049999994, "rle": {"size": [480, 640], "counts": "Zeb74k>b0_O`0@`0@3M3M3M3M3M3L4M3M3M3MOVM]MYIc2e6_M[I`2c6dM\\I\\2a6hM^IW2`6lM_IU2^6nMeIn1Y6UNjIg1T6\\NnIa1P6bNSJZ1l5hNWJT1h5nN[Jm0e5UO^Jg0b5ZO`Jc0_5_OdJ=[5EhJ6X5KlJ1S51PKKP56RKFn4<UK@j4b0YKSOm4o0VKdNR5^1W41N2O1O10O1N1O2M3E;E:E<EdNc0SC[Of=P1O1O1O1O1O00100O100O15K5K5K5J6K5K5KVdi0"}, "label": "blue and grey cell phone on the right side of a purse"}]}
{"id": 427310, "image": "COCO_train2014_000000427310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep in front with a leg raised into the air\"."}], "task": "refering", "answer_template": "The \"sheep in front with a leg raised into the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.18156250000000002, 0.6728997867803838, 0.50946875, 0.8770149253731343], "iscrowd": 0, "area": 9063.663199999997, "rle": {"size": [469, 640], "counts": "R^e11d>2M4M3M2M4M3M3M2N3N1OO10O1000O10O10O1000O10O1000O0100O10O10O100O0102N4K4M4K7J2M2O2N1N3N1N10O010O10O10O010O10001N101N2N1O2O1N1O2N1O2N01N1000000O101O000O10001O0O1000000O100O1O100O1O1O1O1O10O01O1O1O1O100O1O1O1O1O10O10000000000000000000000O2O0000000000000000001O1N101O1O001O1O001O1O001O0O2O1O001O1O001O1O001O1O1N101O1O1OZOjNmCV1S<jNnCU1R<lNnCS1R<mNoCR1R<nNmCR1S<nNnCQ1R<POnCn0S<ROlCo0T<ROkCn0U<ROkCm0V<TOiCl0X<SOhCm0X<TOfCl0[<TOeCl0[<UOdCk0\\<UOdCj0]<WOaCj0`<UO`Ck0`<VO_Ci0b<WO^Ci0b<WO]Ci0d<XO[Ch0e<XO[Ch0e<XO[Cg0f<YOZCg0f<WO\\Ci0d<UO^Cj0c<UO^Ck0V=0O010000O100O10000O1O002N2N2N3M2N2N2M3N3M2N2NY^_4"}, "label": "sheep in front with a leg raised into the air"}]}
{"id": 427310, "image": "COCO_train2014_000000427310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep being shorn in foreground\"."}], "task": "refering", "answer_template": "The \"sheep being shorn in foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.18156250000000002, 0.6728997867803838, 0.50946875, 0.8770149253731343], "iscrowd": 0, "area": 9063.663199999997, "rle": {"size": [469, 640], "counts": "R^e11d>2M4M3M2M4M3M3M2N3N1OO10O1000O10O10O1000O10O1000O0100O10O10O100O0102N4K4M4K7J2M2O2N1N3N1N10O010O10O10O010O10001N101N2N1O2O1N1O2N1O2N01N1000000O101O000O10001O0O1000000O100O1O100O1O1O1O1O10O01O1O1O1O100O1O1O1O1O10O10000000000000000000000O2O0000000000000000001O1N101O1O001O1O001O1O001O0O2O1O001O1O001O1O001O1O1N101O1O1OZOjNmCV1S<jNnCU1R<lNnCS1R<mNoCR1R<nNmCR1S<nNnCQ1R<POnCn0S<ROlCo0T<ROkCn0U<ROkCm0V<TOiCl0X<SOhCm0X<TOfCl0[<TOeCl0[<UOdCk0\\<UOdCj0]<WOaCj0`<UO`Ck0`<VO_Ci0b<WO^Ci0b<WO]Ci0d<XO[Ch0e<XO[Ch0e<XO[Cg0f<YOZCg0f<WO\\Ci0d<UO^Cj0c<UO^Ck0V=0O010000O100O10000O1O002N2N2N3M2N2N2M3N3M2N2NY^_4"}, "label": "sheep being shorn in foreground"}]}
{"id": 468276, "image": "COCO_train2014_000000468276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the paper plate\"."}], "task": "refering", "answer_template": "The \"the paper plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 200, 201, 202, 224, 225, 226, 235, 236, 248, 249, 259, 260, 261, 266, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 354, 355, 357, 358, 359, 360, 361, 362, 363, 378, 380, 381, 382, 383, 384], "bbox": [0.24881250000000002, 0.41754166666666664, 0.8817499999999999, 0.9842083333333332], "iscrowd": 0, "area": 35707.85265000001, "rle": {"size": [480, 640], "counts": "YjZ21n>5K4L4M3M3M3M4M2M3M3M3N2N2N3L3N2N2N2N2M4M2N2N2N1N10001O001N10001O000O2O001O000O2O001O000O2O001O000O2O001O00O0O2N2O1N2O0O100O2N100O2O0O100O2O0O1O101N100O2O0O1O101N100O101N1O101N100OPD`Nb:`1]EaNd:]1\\EeNc:Z1]EgNc:Y1\\EiNc:V1\\ElNe:S1ZEnNf:Q1ZEQOe:n0[ESOf:k0ZEWOe:i0ZEXOf:g0YE\\Of:c0ZE^Og:`0YEBf:>YECg:<YEEg::YEHg:6XELh:4WENh:1XE0i:NWE4h:KXE6h:JVE8j:GVE;i:DWE=h:CXE?g:AXE`0h:_OXEc0g:\\OXEf0h:YOXEi0f:XOYEi0g:VOYEk0g:TOYEn0f:QOZEP1f:POXES1f:mNZET1f:kNZEW1e:hN[EY1e:gNZE[1e:dN[E]1e:bNZE`1e:`N[Eb1d:^N[Ec1e:\\N[Ee1e:ZN[Ee1g:ZNYEd1i:]NUEb1n:]NREa1Q;^NoD`1T;_NlD_1W;aNhD]1Z;cNfD\\1\\;cNeDZ1^;eNbDY1a;gN_DV1d;iN]DT1f;kNZDS1h;mNYDQ1i;oNWDn0l;QOUDl0n;SORDk0Q<TOQDh0Q<YOPDc0S<\\OPD?S<@oC=S<BPD9S<GnC5T<KnC1U<N_1O100OQS7R2kjH3M3M3N2N3M2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O100O100O10000O100O100O11O1O00100O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001OjJhLeNV3]1jLbNV3^1kLbNS3`1mL`NR3`1oL_NP3c1PM]Nn2d1SM[Nm2f1SMYNl2h1UMWNk2j1UMTNk2m1VMRNj2o1VMPNi2R2WMmMi2S2XMkMh2W2XMhMh2X2YMgMf2[2[MbMf2^2[MaMd2a2\\M^Md2b2]M]Mb2e2^MYMc2g2^MXMa2j2_MTMa2n2_MQMa2o2`MPM_2R3aMlL`2T3aMkL^2W3bMhL^2X3cMfL]2\\3cMcL]2]3dMbL[2`3eM^L\\2b3eM]LZ2e3fMYL[2h3fMVLY2k3hMTLX2m3hMQLX2P4iMoKV2S4jMkKW2U4jMiKV2Y4jMeKW2[4jMcKV2_4jM_KW2a4jM]KV2e4jMYKW2g4jMWKV2k4jMSKW2n4iMPKW2Q5jMlJX2U5hMiJX2X5iMfJX2[5hMcJX2^5jM_JW2b5iM\\JW2e5iMZJW2h5iMVJX2j5iMTJW2n5iMPJX2Q6gMnIY2S6hMkIY2V6gMhIY2Y6hMeIY2\\6fMbI[2_6fM_I[2b6eM\\I[2e6fMYI[2h6dMVI]2k6dMQI_2P7aMkHb2W7^MeHd2\\7\\M`Hh2a7XMZHk2g7VMUHm2l7SMoGP3R8PMkGS3V8mLeGV3\\8kL`GX3a8hLZG[3g8eLUG_3l8aLPG`3R9aLiFa3Z9P13M2N3M2N3M2N3M2N3M2N3M4L4L4L4L5K4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L5K4LkYS1"}, "label": "the paper plate"}]}
{"id": 468276, "image": "COCO_train2014_000000468276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a salad , sandwich with bacon on a paper plate\"."}], "task": "refering", "answer_template": "The \"a salad , sandwich with bacon on a paper plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 200, 201, 202, 224, 225, 226, 235, 236, 248, 249, 259, 260, 261, 266, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 354, 355, 357, 358, 359, 360, 361, 362, 363, 378, 380, 381, 382, 383, 384], "bbox": [0.24881250000000002, 0.41754166666666664, 0.8817499999999999, 0.9842083333333332], "iscrowd": 0, "area": 35707.85265000001, "rle": {"size": [480, 640], "counts": "YjZ21n>5K4L4M3M3M3M4M2M3M3M3N2N2N3L3N2N2N2N2M4M2N2N2N1N10001O001N10001O000O2O001O000O2O001O000O2O001O000O2O001O00O0O2N2O1N2O0O100O2N100O2O0O100O2O0O1O101N100O2O0O1O101N100O101N1O101N100OPD`Nb:`1]EaNd:]1\\EeNc:Z1]EgNc:Y1\\EiNc:V1\\ElNe:S1ZEnNf:Q1ZEQOe:n0[ESOf:k0ZEWOe:i0ZEXOf:g0YE\\Of:c0ZE^Og:`0YEBf:>YECg:<YEEg::YEHg:6XELh:4WENh:1XE0i:NWE4h:KXE6h:JVE8j:GVE;i:DWE=h:CXE?g:AXE`0h:_OXEc0g:\\OXEf0h:YOXEi0f:XOYEi0g:VOYEk0g:TOYEn0f:QOZEP1f:POXES1f:mNZET1f:kNZEW1e:hN[EY1e:gNZE[1e:dN[E]1e:bNZE`1e:`N[Eb1d:^N[Ec1e:\\N[Ee1e:ZN[Ee1g:ZNYEd1i:]NUEb1n:]NREa1Q;^NoD`1T;_NlD_1W;aNhD]1Z;cNfD\\1\\;cNeDZ1^;eNbDY1a;gN_DV1d;iN]DT1f;kNZDS1h;mNYDQ1i;oNWDn0l;QOUDl0n;SORDk0Q<TOQDh0Q<YOPDc0S<\\OPD?S<@oC=S<BPD9S<GnC5T<KnC1U<N_1O100OQS7R2kjH3M3M3N2N3M2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O100O100O10000O100O100O11O1O00100O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001OjJhLeNV3]1jLbNV3^1kLbNS3`1mL`NR3`1oL_NP3c1PM]Nn2d1SM[Nm2f1SMYNl2h1UMWNk2j1UMTNk2m1VMRNj2o1VMPNi2R2WMmMi2S2XMkMh2W2XMhMh2X2YMgMf2[2[MbMf2^2[MaMd2a2\\M^Md2b2]M]Mb2e2^MYMc2g2^MXMa2j2_MTMa2n2_MQMa2o2`MPM_2R3aMlL`2T3aMkL^2W3bMhL^2X3cMfL]2\\3cMcL]2]3dMbL[2`3eM^L\\2b3eM]LZ2e3fMYL[2h3fMVLY2k3hMTLX2m3hMQLX2P4iMoKV2S4jMkKW2U4jMiKV2Y4jMeKW2[4jMcKV2_4jM_KW2a4jM]KV2e4jMYKW2g4jMWKV2k4jMSKW2n4iMPKW2Q5jMlJX2U5hMiJX2X5iMfJX2[5hMcJX2^5jM_JW2b5iM\\JW2e5iMZJW2h5iMVJX2j5iMTJW2n5iMPJX2Q6gMnIY2S6hMkIY2V6gMhIY2Y6hMeIY2\\6fMbI[2_6fM_I[2b6eM\\I[2e6fMYI[2h6dMVI]2k6dMQI_2P7aMkHb2W7^MeHd2\\7\\M`Hh2a7XMZHk2g7VMUHm2l7SMoGP3R8PMkGS3V8mLeGV3\\8kL`GX3a8hLZG[3g8eLUG_3l8aLPG`3R9aLiFa3Z9P13M2N3M2N3M2N3M2N3M2N3M4L4L4L4L5K4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L5K4LkYS1"}, "label": "a salad , sandwich with bacon on a paper plate"}]}
{"id": 468276, "image": "COCO_train2014_000000468276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beer bottle\"."}], "task": "refering", "answer_template": "The \"a beer bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 134, 135, 136, 157, 158, 159], "bbox": [0.83296875, 0.0, 1.0, 0.36947916666666664], "iscrowd": 0, "area": 11400.050349999998, "rle": {"size": [480, 640], "counts": "UPj71k>4L5K4M4K4L4L5K4L5L3L4L5K4L5L3L4L5K4L5L3L4L5K4L5L3L4L5K4L4M3L4L4L4M3N2O0O2O1N2O1O1N2O0O2O1N2O1O1N2O0O2O1N2O1N2O1O0O2O10M4L3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3N2N2M3N2N2M3N1"}, "label": "a beer bottle"}]}
{"id": 468276, "image": "COCO_train2014_000000468276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of beer sitting on a table next to a plate of food\"."}], "task": "refering", "answer_template": "The \"a bottle of beer sitting on a table next to a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 134, 135, 136, 157, 158, 159], "bbox": [0.83296875, 0.0, 1.0, 0.36947916666666664], "iscrowd": 0, "area": 11400.050349999998, "rle": {"size": [480, 640], "counts": "UPj71k>4L5K4M4K4L4L5K4L5L3L4L5K4L5L3L4L5K4L5L3L4L5K4L5L3L4L5K4L4M3L4L4L4M3N2O0O2O1N2O1O1N2O0O2O1N2O1O1N2O0O2O1N2O1N2O1O0O2O10M4L3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3N2N2M3N2N2M3N1"}, "label": "a bottle of beer sitting on a table next to a plate of food"}]}
{"id": 58677, "image": "COCO_train2014_000000058677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red plastic bottle of curry ketchup\"."}], "task": "refering", "answer_template": "The \"a red plastic bottle of curry ketchup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 22, 23, 24, 25, 26, 27, 28, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 158, 159, 160, 161, 162, 175, 176, 177, 178, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 259, 260, 261, 262, 277, 278], "bbox": [0.26206250000000003, 0.0016250000000000001, 0.6623541666666667, 0.727328125], "iscrowd": 0, "area": 54258.1214, "rle": {"size": [640, 480], "counts": "^S_2U2ia03M3M3M2N3M3M3M2M4H8H8H7H9H8H7I8G9H8H7I8H8G9H7I8H8G9E:eN\\1eN[1POo0K6J6J6K4K6J6J6K4K6J6J5L3L5K4L4M4N10001O0000001O00001O000010O000001O00001O01O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O1K5G9G9G9G9G9G9I7M3L4L4M3L4L4M3L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M4K5K6K4K5K6K4K6J5L4K6J5L4K6J5L5J5K5L5J5K5L5J5K6K4K5K6K4K5K6K4K6J5L4K6J5L4K6J5L5J\\VU3"}, "label": "a red plastic bottle of curry ketchup"}]}
{"id": 58677, "image": "COCO_train2014_000000058677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of curry ketchup\"."}], "task": "refering", "answer_template": "The \"a bottle of curry ketchup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 22, 23, 24, 25, 26, 27, 28, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 158, 159, 160, 161, 162, 175, 176, 177, 178, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 259, 260, 261, 262, 277, 278], "bbox": [0.26206250000000003, 0.0016250000000000001, 0.6623541666666667, 0.727328125], "iscrowd": 0, "area": 54258.1214, "rle": {"size": [640, 480], "counts": "^S_2U2ia03M3M3M2N3M3M3M2M4H8H8H7H9H8H7I8G9H8H7I8H8G9H7I8H8G9E:eN\\1eN[1POo0K6J6J6K4K6J6J6K4K6J6J5L3L5K4L4M4N10001O0000001O00001O000010O000001O00001O01O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O1K5G9G9G9G9G9G9I7M3L4L4M3L4L4M3L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M4K5K6K4K5K6K4K6J5L4K6J5L4K6J5L5J5K5L5J5K5L5J5K6K4K5K6K4K5K6K4K6J5L4K6J5L4K6J5L5J\\VU3"}, "label": "a bottle of curry ketchup"}]}
{"id": 58677, "image": "COCO_train2014_000000058677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass bottle with a reddish orange sauce and a red and yellow label\"."}], "task": "refering", "answer_template": "The \"a glass bottle with a reddish orange sauce and a red and yellow label\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 17, 18, 34, 35, 51, 52, 68, 69, 85, 86, 87, 102, 103, 104, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 188, 189, 205], "bbox": [0.0, 0.012562499999999999, 0.14958333333333332, 0.5387500000000001], "iscrowd": 0, "area": 18459.843750000004, "rle": {"size": [640, 480], "counts": "P1j4U?f1ZN9G8G8I4L4L4K6F9H8J7L3M3M?A00O100000000O100000000O100[1eN000000O10004K7J5K4L1N2O2N1N3N1O2cFkEk8V:TGVF_8m9_G_FU8R;D=B=D<D4L1N2O1O1N2O1O1O1N2O1N2N2M3N2N4K5L4L4K5[Lcnn7"}, "label": "a glass bottle with a reddish orange sauce and a red and yellow label"}]}
{"id": 58677, "image": "COCO_train2014_000000058677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall bottle of hot sauce with a pepper on the wrapper\"."}], "task": "refering", "answer_template": "The \"a tall bottle of hot sauce with a pepper on the wrapper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 17, 18, 34, 35, 51, 52, 68, 69, 85, 86, 87, 102, 103, 104, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 188, 189, 205], "bbox": [0.0, 0.012562499999999999, 0.14958333333333332, 0.5387500000000001], "iscrowd": 0, "area": 18459.843750000004, "rle": {"size": [640, 480], "counts": "P1j4U?f1ZN9G8G8I4L4L4K6F9H8J7L3M3M?A00O100000000O100000000O100[1eN000000O10004K7J5K4L1N2O2N1N3N1O2cFkEk8V:TGVF_8m9_G_FU8R;D=B=D<D4L1N2O1O1N2O1O1O1N2O1N2N2M3N2N4K5L4L4K5[Lcnn7"}, "label": "a tall bottle of hot sauce with a pepper on the wrapper"}]}
{"id": 566584, "image": "COCO_train2014_000000566584.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing orange t - shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing orange t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 71, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.008796875, 0.14816901408450703, 0.440234375, 0.9867840375586855], "iscrowd": 0, "area": 69012.4998, "rle": {"size": [426, 640], "counts": "kj2m0X<8H7@`0^Ob0L4L4L3M3L5J5J6K5M3L4L4M3L4A?N2M3N2M3M3N2M3M3N2M3M3N2M3M3N1N2N2N2N2N2N2M3N2N2M300O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100gLTI3m6KXI1i6N[IOe6O_INb61_IOa6OaI0`6OaIiMC^1m6h0bIeMI_1e6j0dIdMj0a0c5j1dIaMP1b0\\5k1gI^MS1e0V5l1kKPNV4o1nKnMR4P2RLmMo3S2TLjMl3U2VLiMk3W2VLhMj3W2XLhMh3X2YLfMh3Y2ZLfMf3Z2[LdMf3\\2d3O100O100O100O100O100O100O100O10000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O<D>XGZLP7]4THlKh7o4L2N1O0000001O000000001O0000001O001O001O001O1O001O1O001O1O001O001N2O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O002N2N2N2N2N3M2N2N2N2N3M2N4L4L4L3M4L4L4Lc0RNbGjMl8f0j2\\O<E:EUdd4"}, "label": "a man wearing orange t - shirt"}]}
{"id": 566584, "image": "COCO_train2014_000000566584.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange colored shirt with a pair of glasses hanging from the shirt\"."}], "task": "refering", "answer_template": "The \"a man in an orange colored shirt with a pair of glasses hanging from the shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 71, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.008796875, 0.14816901408450703, 0.440234375, 0.9867840375586855], "iscrowd": 0, "area": 69012.4998, "rle": {"size": [426, 640], "counts": "kj2m0X<8H7@`0^Ob0L4L4L3M3L5J5J6K5M3L4L4M3L4A?N2M3N2M3M3N2M3M3N2M3M3N2M3M3N1N2N2N2N2N2N2M3N2N2M300O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100gLTI3m6KXI1i6N[IOe6O_INb61_IOa6OaI0`6OaIiMC^1m6h0bIeMI_1e6j0dIdMj0a0c5j1dIaMP1b0\\5k1gI^MS1e0V5l1kKPNV4o1nKnMR4P2RLmMo3S2TLjMl3U2VLiMk3W2VLhMj3W2XLhMh3X2YLfMh3Y2ZLfMf3Z2[LdMf3\\2d3O100O100O100O100O100O100O100O10000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O<D>XGZLP7]4THlKh7o4L2N1O0000001O000000001O0000001O001O001O001O1O001O1O001O1O001O001N2O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O002N2N2N2N2N3M2N2N2N2N3M2N4L4L4L3M4L4L4Lc0RNbGjMl8f0j2\\O<E:EUdd4"}, "label": "a man in an orange colored shirt with a pair of glasses hanging from the shirt"}]}
{"id": 566584, "image": "COCO_train2014_000000566584.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green vest\"."}], "task": "refering", "answer_template": "The \"a woman in a green vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.538484375, 0.2532394366197183, 1.0, 0.9813380281690142], "iscrowd": 0, "area": 60183.856750000006, "rle": {"size": [426, 640], "counts": "Vi_4;f<;E<D;E<D;E<D<D;E<D;E<D;I8L3L5L4L3L5L3M4K4M4K5L3M4K4M4L3L5L5K7H9H8H7H9H8G9H7I2M9H1O1N2O1O1N2O1O100000000000000001O00000000000000000000001O00000000000000000000001O0000000000000000000000001O0000002N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N:F1O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001On0RO2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N2N2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O3M3M4L4L4L4L4L3M4L4L4L4L3M4L4LPC"}, "label": "a woman in a green vest"}]}
{"id": 566584, "image": "COCO_train2014_000000566584.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mature woman with blonde hair looking into her phone\"."}], "task": "refering", "answer_template": "The \"mature woman with blonde hair looking into her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.538484375, 0.2532394366197183, 1.0, 0.9813380281690142], "iscrowd": 0, "area": 60183.856750000006, "rle": {"size": [426, 640], "counts": "Vi_4;f<;E<D;E<D;E<D<D;E<D;E<D;I8L3L5L4L3L5L3M4K4M4K5L3M4K4M4L3L5L5K7H9H8H7H9H8G9H7I2M9H1O1N2O1O1N2O1O100000000000000001O00000000000000000000001O00000000000000000000001O0000000000000000000000001O0000002N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N:F1O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001On0RO2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N2N2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O3M3M4L4L4L4L4L3M4L4L4L4L3M4L4LPC"}, "label": "mature woman with blonde hair looking into her phone"}]}
{"id": 42297, "image": "COCO_train2014_000000042297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child reaching for cake\"."}], "task": "refering", "answer_template": "The \"child reaching for cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 33, 48, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 219, 222], "bbox": [0.122078125, 0.006745283018867924, 0.6892968749999999, 0.6786556603773585], "iscrowd": 0, "area": 58077.58424999999, "rle": {"size": [424, 640], "counts": "bZP1V1Q<2O2M3M2N3N2M2N3M2O2L4M2M4M3L3N3L3N3L4M2M4M3L4M3L4N2N2N3M2N2N2N2N2N2N3M2SJQL]2Q4`MUL[2m3aMYL[2i3bM]LY2e3cMbLX2`3eMeLW2^3dMiLW2Y3fMmLU2U3gMQMU2Q3hMQMW2Q3fMmL]2U3_MjLd2X3YMfLj2[3SMdLP3]3nLcLS3^3jLcLW3^3gLcLY3^3eLcL[3^3bLcL_3_3^LbLb3_3[LbLf3_3XLbLh3_3ULbLl3_3RLbLn3_3PLbLP4_3mKbLT4`3iKaLW4`3fKbLZ4_3dKbL\\4_3bKbL^4_3_KcLa4^3]KcLb4_3[KcLe4^3YKcLg4_3UKcLk4m50O10000000000O100000000O100O100O100O100O1000000000000000000O100000000O100000000O100000000000000O1000000000000000000O1000000000000000000O1001O0000001O000000001O000000001O0000001O0000001O00000100O1O100O1O100O1O1O100O1O100N2N2N2N2N2M3O1O1N2O1O1O1N2O1O1N2M2N3M3N2M4L5K5K5K5L4M3M3M3MdHdJo6Y5mHkJV7R5fHRK]7k4_HYKd7T53M3M3M3M3M3M3M3M3M3M3M3M3M3M3M:F4L00001O00010O000010O0001O0001O01_OhFPMX9o2a001O01OO2N1OWFUMW9i2eF]M\\9a2`FeM_9S3O010O1O011N2N2O1N3M01K4L5M3cMaFo0`9fNiFZ1Y9bNiF^1Y9`NgF`1[9^NeFb1]9\\NcFd1_9ZNbFf1_9XNcFf1_9YNbFe1_9ZNcFd1_9ZNcFe1^9YNdFf1\\9YNfFf1[9YNfFe1\\9YNeFg1[9XNfFh1]:O010000O10000O1000000O10000O1N2O11lNS101O001O1O0O2O1O0o0RO3L01O1O010O001O1O001O1O0010O01O1O010O1O0011N2O1O1N2ON1O2O1O001O1O001O001O1O010O1O010O00105J6J6K4HWUb2"}, "label": "child reaching for cake"}]}
{"id": 42297, "image": "COCO_train2014_000000042297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black haired toddler reaching for cake\"."}], "task": "refering", "answer_template": "The \"a black haired toddler reaching for cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 33, 48, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 219, 222], "bbox": [0.122078125, 0.006745283018867924, 0.6892968749999999, 0.6786556603773585], "iscrowd": 0, "area": 58077.58424999999, "rle": {"size": [424, 640], "counts": "bZP1V1Q<2O2M3M2N3N2M2N3M2O2L4M2M4M3L3N3L3N3L4M2M4M3L4M3L4N2N2N3M2N2N2N2N2N2N3M2SJQL]2Q4`MUL[2m3aMYL[2i3bM]LY2e3cMbLX2`3eMeLW2^3dMiLW2Y3fMmLU2U3gMQMU2Q3hMQMW2Q3fMmL]2U3_MjLd2X3YMfLj2[3SMdLP3]3nLcLS3^3jLcLW3^3gLcLY3^3eLcL[3^3bLcL_3_3^LbLb3_3[LbLf3_3XLbLh3_3ULbLl3_3RLbLn3_3PLbLP4_3mKbLT4`3iKaLW4`3fKbLZ4_3dKbL\\4_3bKbL^4_3_KcLa4^3]KcLb4_3[KcLe4^3YKcLg4_3UKcLk4m50O10000000000O100000000O100O100O100O100O1000000000000000000O100000000O100000000O100000000000000O1000000000000000000O1000000000000000000O1001O0000001O000000001O000000001O0000001O0000001O00000100O1O100O1O100O1O1O100O1O100N2N2N2N2N2M3O1O1N2O1O1O1N2O1O1N2M2N3M3N2M4L5K5K5K5L4M3M3M3MdHdJo6Y5mHkJV7R5fHRK]7k4_HYKd7T53M3M3M3M3M3M3M3M3M3M3M3M3M3M3M:F4L00001O00010O000010O0001O0001O01_OhFPMX9o2a001O01OO2N1OWFUMW9i2eF]M\\9a2`FeM_9S3O010O1O011N2N2O1N3M01K4L5M3cMaFo0`9fNiFZ1Y9bNiF^1Y9`NgF`1[9^NeFb1]9\\NcFd1_9ZNbFf1_9XNcFf1_9YNbFe1_9ZNcFd1_9ZNcFe1^9YNdFf1\\9YNfFf1[9YNfFe1\\9YNeFg1[9XNfFh1]:O010000O10000O1000000O10000O1N2O11lNS101O001O1O0O2O1O0o0RO3L01O1O010O001O1O001O1O0010O01O1O010O1O0011N2O1O1N2ON1O2O1O001O1O001O001O1O010O1O010O00105J6J6K4HWUb2"}, "label": "a black haired toddler reaching for cake"}]}
{"id": 42297, "image": "COCO_train2014_000000042297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand that is about to grab the child ' s arm\"."}], "task": "refering", "answer_template": "The \"the hand that is about to grab the child ' s arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 250, 251, 252, 275], "bbox": [0.549359375, 0.006745283018867924, 1.0, 0.7438207547169812], "iscrowd": 0, "area": 46650.15004999999, "rle": {"size": [424, 640], "counts": "fma42T=3M2N3M2N2N3N1N3M2O1J6K50O0100O010O010O10O10O10O0100O01O001O0O2O1O0010O010O01O010O010O10O01O010O0100O010O100O10000O010O100O100O100O01000O100O100O100O10O010000O100O10000O100O1000O0100O10000O100O10000O100O01000O100O10000SFbNo7a1hGhNV8Y1bGPO[8S1^GTO_8n0\\GVOc8m0WGVOj8l0PGVOQ9l0jFWOV9l0dFWO\\9k0_FXOa9V22O00000O101O0O10000O2O0O10000O101N100O10001NPHlLo5T3oIoLo5Q3PJRMm5P3QJSMm5m2RJWMj5k2TJXMj5i2TJZMj5f2UJ]Mh5d2WJ_Mh3\\O\\LT3LdMg3X4YLkKg3S4XLPLh3m3XLWLg3f3YL]Lg3`3YLcLg3Z3YLjLf3S3ZLPMf3o2XLTMh3k2VLXMj3g2TL]Mk3c2RL`Mn3_2PLdMP4[2nKiMQ4W2lKlMT4S2jKPNV4o1hKTNX4l1eKXNZ4g1dK\\N\\4c1bK`N^4_1`KeN_4[1^KhNb4W1\\KlNd4S1ZKQOe4o0XKTOh4k0VKXOj4g0TK]Ok4c0RKAm4>QKEo4[4O1O001O1O001O1O1O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000O1O1O100O1O1O1O1[Oe0L4N2O1O1O1N2ON"}, "label": "the hand that is about to grab the child ' s arm"}]}
{"id": 42297, "image": "COCO_train2014_000000042297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand reaching out\"."}], "task": "refering", "answer_template": "The \"the hand reaching out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 250, 251, 252, 275], "bbox": [0.549359375, 0.006745283018867924, 1.0, 0.7438207547169812], "iscrowd": 0, "area": 46650.15004999999, "rle": {"size": [424, 640], "counts": "fma42T=3M2N3M2N2N3N1N3M2O1J6K50O0100O010O010O10O10O10O0100O01O001O0O2O1O0010O010O01O010O010O10O01O010O0100O010O100O10000O010O100O100O100O01000O100O100O100O10O010000O100O10000O100O1000O0100O10000O100O10000O100O01000O100O10000SFbNo7a1hGhNV8Y1bGPO[8S1^GTO_8n0\\GVOc8m0WGVOj8l0PGVOQ9l0jFWOV9l0dFWO\\9k0_FXOa9V22O00000O101O0O10000O2O0O10000O101N100O10001NPHlLo5T3oIoLo5Q3PJRMm5P3QJSMm5m2RJWMj5k2TJXMj5i2TJZMj5f2UJ]Mh5d2WJ_Mh3\\O\\LT3LdMg3X4YLkKg3S4XLPLh3m3XLWLg3f3YL]Lg3`3YLcLg3Z3YLjLf3S3ZLPMf3o2XLTMh3k2VLXMj3g2TL]Mk3c2RL`Mn3_2PLdMP4[2nKiMQ4W2lKlMT4S2jKPNV4o1hKTNX4l1eKXNZ4g1dK\\N\\4c1bK`N^4_1`KeN_4[1^KhNb4W1\\KlNd4S1ZKQOe4o0XKTOh4k0VKXOj4g0TK]Ok4c0RKAm4>QKEo4[4O1O001O1O001O1O1O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000O1O1O100O1O1O1O1[Oe0L4N2O1O1O1N2ON"}, "label": "the hand reaching out"}]}
{"id": 533827, "image": "COCO_train2014_000000533827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver car beside tiny truck\"."}], "task": "refering", "answer_template": "The \"silver car beside tiny truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 301, 302], "bbox": [0.000359375, 0.23457556935817805, 0.1683125, 0.8014285714285715], "iscrowd": 0, "area": 18541.411500000006, "rle": {"size": [483, 640], "counts": "b3m7U711O1N101O1O001O1O100O101N1O1O2N1N2O2M2O1N3N1O2N2N2N2N2N1O3M3M2N3L4M3M3M4L3M4M2M2O0O011N4K3N0O00O2NYK`Fo3`9QLcFk3]9VLeFh3[9XLhFd3Y9\\LjFc3T9]LnFi3j8WLWGk3f8UL\\Gm3`8SLaGo3\\8QLdGP4[8oKfGR4Y8nKhGQ4Y8nKgGS4X8lKiGU4V8kKkGU4T8jKmGV4S8jKmGV4S8iKnGW4S8hKnGW4R8iKnGW4R8hKPHW4Q8gKPHY4P8fKQHZ4P8eKQHZ4P8dKQH\\4T8ZKQHf4W9O1O001O1O001O1O1N2O1N2O1O1N2O1J7H7H9H8H9I8J9F9H:F:E5L3J6J7H9G8Gcdj7"}, "label": "silver car beside tiny truck"}]}
{"id": 533827, "image": "COCO_train2014_000000533827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car\"."}], "task": "refering", "answer_template": "The \"a white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 301, 302], "bbox": [0.000359375, 0.23457556935817805, 0.1683125, 0.8014285714285715], "iscrowd": 0, "area": 18541.411500000006, "rle": {"size": [483, 640], "counts": "b3m7U711O1N101O1O001O1O100O101N1O1O2N1N2O2M2O1N3N1O2N2N2N2N2N1O3M3M2N3L4M3M3M4L3M4M2M2O0O011N4K3N0O00O2NYK`Fo3`9QLcFk3]9VLeFh3[9XLhFd3Y9\\LjFc3T9]LnFi3j8WLWGk3f8UL\\Gm3`8SLaGo3\\8QLdGP4[8oKfGR4Y8nKhGQ4Y8nKgGS4X8lKiGU4V8kKkGU4T8jKmGV4S8jKmGV4S8iKnGW4S8hKnGW4R8iKnGW4R8hKPHW4Q8gKPHY4P8fKQHZ4P8eKQHZ4P8dKQH\\4T8ZKQHf4W9O1O001O1O001O1O1N2O1N2O1O1N2O1J7H7H9H8H9I8J9F9H:F:E5L3J6J7H9G8Gcdj7"}, "label": "a white car"}]}
{"id": 533827, "image": "COCO_train2014_000000533827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small yellow and white service vehicle\"."}], "task": "refering", "answer_template": "The \"a small yellow and white service vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 341, 350, 351, 352, 356, 357, 358], "bbox": [0.248453125, 0.15223602484472049, 0.8555624999999999, 0.9368737060041408], "iscrowd": 0, "area": 106717.95835000002, "rle": {"size": [483, 640], "counts": "T\\[2g0n=b0^Ob0G9G:G8G8G7H8H9I6K5J7J5J5L4K6F9L4L5J5G:G8H8G:G8G9H9F9H8H9F9ZN]GdLk8[3XGZLP9f3RGQLU9o3nFgKZ9W4iF_K_9b4`0001O1O1O001O1O1O001O100M2M4M3L4L301000O1000000O1000000O1000000O10O1000O100000000000000O1000000000000000000O1000O10000000000000O100000000000000000000000000000O1000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O100002N3M2O2M3M2N3M2N1O2N1O1O2N1O1O1O1O001O1O1N2M2M4M3L3M3N2M3M3N1N3N2N2000O10O100000O100000000O1000000L5I7J5J7I8I8K5M2M4M3M2N3L4M3M3L101O001N10001N10001O0O1000000O10000O10001O0O10000O1000000O10000O2O00000O1000000O10000O101O01O1O1O2N1O2N1O1O2N1O2O0O1O2N1O2N1O1O1O000000010O00000000001O0000000001O01O0000000000J6J6I7I7J7O000001O0000000000000001O01O00dLjIkMV6h1WJXNi5Z1eJfN[5m0RKSOn4`0`K_O`43nKMR4F[L:e3XOiLh0W3kNVMV1i2]NdMc1\\2oMRNQ2n1bM_N^2a1TMmNl2f6O000000000001O01O00000000000000010O000000000000001O01O00000000000001M2M8H>]Ob0QOP1oNgR[1"}, "label": "a small yellow and white service vehicle"}]}
{"id": 533827, "image": "COCO_train2014_000000533827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little car on the road\"."}], "task": "refering", "answer_template": "The \"a little car on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 341, 350, 351, 352, 356, 357, 358], "bbox": [0.248453125, 0.15223602484472049, 0.8555624999999999, 0.9368737060041408], "iscrowd": 0, "area": 106717.95835000002, "rle": {"size": [483, 640], "counts": "T\\[2g0n=b0^Ob0G9G:G8G8G7H8H9I6K5J7J5J5L4K6F9L4L5J5G:G8H8G:G8G9H9F9H8H9F9ZN]GdLk8[3XGZLP9f3RGQLU9o3nFgKZ9W4iF_K_9b4`0001O1O1O001O1O1O001O100M2M4M3L4L301000O1000000O1000000O1000000O10O1000O100000000000000O1000000000000000000O1000O10000000000000O100000000000000000000000000000O1000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O100002N3M2O2M3M2N3M2N1O2N1O1O2N1O1O1O1O001O1O1N2M2M4M3L3M3N2M3M3N1N3N2N2000O10O100000O100000000O1000000L5I7J5J7I8I8K5M2M4M3M2N3L4M3M3L101O001N10001N10001O0O1000000O10000O10001O0O10000O1000000O10000O2O00000O1000000O10000O101O01O1O1O2N1O2N1O1O2N1O2O0O1O2N1O2N1O1O1O000000010O00000000001O0000000001O01O0000000000J6J6I7I7J7O000001O0000000000000001O01O00dLjIkMV6h1WJXNi5Z1eJfN[5m0RKSOn4`0`K_O`43nKMR4F[L:e3XOiLh0W3kNVMV1i2]NdMc1\\2oMRNQ2n1bM_N^2a1TMmNl2f6O000000000001O01O00000000000000010O000000000000001O01O00000000000001M2M8H>]Ob0QOP1oNgR[1"}, "label": "a little car on the road"}]}
{"id": 353607, "image": "COCO_train2014_000000353607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red double - decker bus\"."}], "task": "refering", "answer_template": "The \"a red double - decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 61, 62, 63, 64, 65, 66, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388], "bbox": [0.40954687500000003, 0.1039375, 0.991015625, 0.9848333333333332], "iscrowd": 0, "area": 118488.1737, "rle": {"size": [480, 640], "counts": "bTk34`><C=D<C=D<D<M3O1O1O1O1N2O1O1O2N2N3M2M4M3M2N3M2N3L4M2N3_MVLgIl3W6VLgIm3W6TLgIn3W6TL^HL9T4V7QLSHi04^3b7kKmGc10j2k7b2J7H8H7J2M2O0O2N100O2N101N100O2N101N1O101N101N1O101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O1O2O0O2O0O1O2O0O2O0O1O2O0O2N100O2O0O2N100O2N101N100O2N101N1O101N101N1O101N1O2O0O101N1O101N1O101N100O2N100O2N100O2O0O1O2O0O101N1O101N1O101N100O2N100O2N101N100O2N100O2N100O2O0O1O2O0O1O2O0O101N1O101N1O101N100O2N100O2O0O1O2O0O1O2O0O101N1O101N1O101N100O2N100O2O000010O00010O0O10001O000O2O00000O2O0000001N10001O000O101O00001N1000001O0O10001O000O2O00000O2O00001O0O10001O000O101O00001N1000001O0O101O00000O2O00000O2O00001O0O10001O000O2O0000001N10001O000O101O00000O2O00001N1000001O0O101O00000O2O0000001N10004L6J5J7J5K6I6K6I6J7J5J7J5J7J5J7I6K5J7J5J7J5J7I6K6I6K6I6K6I6J7J5J7J7H<D<E<C<E;D<E;D<D=D;D<E;D<EQ]2"}, "label": "a red double - decker bus"}]}
{"id": 353607, "image": "COCO_train2014_000000353607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beautiful red double decker bus on road\"."}], "task": "refering", "answer_template": "The \"a beautiful red double decker bus on road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 61, 62, 63, 64, 65, 66, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388], "bbox": [0.40954687500000003, 0.1039375, 0.991015625, 0.9848333333333332], "iscrowd": 0, "area": 118488.1737, "rle": {"size": [480, 640], "counts": "bTk34`><C=D<C=D<D<M3O1O1O1O1N2O1O1O2N2N3M2M4M3M2N3M2N3L4M2N3_MVLgIl3W6VLgIm3W6TLgIn3W6TL^HL9T4V7QLSHi04^3b7kKmGc10j2k7b2J7H8H7J2M2O0O2N100O2N101N100O2N101N1O101N101N1O101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O1O2O0O2O0O1O2O0O2O0O1O2O0O2N100O2O0O2N100O2N101N100O2N101N1O101N101N1O101N1O2O0O101N1O101N1O101N100O2N100O2N100O2O0O1O2O0O101N1O101N1O101N100O2N100O2N101N100O2N100O2N100O2O0O1O2O0O1O2O0O101N1O101N1O101N100O2N100O2O0O1O2O0O1O2O0O101N1O101N1O101N100O2N100O2O000010O00010O0O10001O000O2O00000O2O0000001N10001O000O101O00001N1000001O0O10001O000O2O00000O2O00001O0O10001O000O101O00001N1000001O0O101O00000O2O00000O2O00001O0O10001O000O2O0000001N10001O000O101O00000O2O00001N1000001O0O101O00000O2O0000001N10004L6J5J7J5K6I6K6I6J7J5J7J5J7J5J7I6K5J7J5J7J5J7I6K6I6K6I6K6I6J7J5J7J7H<D<E<C<E;D<E;D<D=D;D<E;D<EQ]2"}, "label": "a beautiful red double decker bus on road"}]}
{"id": 353607, "image": "COCO_train2014_000000353607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bus parked next to a red bus\"."}], "task": "refering", "answer_template": "The \"a green bus parked next to a red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 48, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.02190625, 0.1016875, 0.46179687499999994, 0.9870833333333333], "iscrowd": 0, "area": 99959.38130000002, "rle": {"size": [480, 640], "counts": "PP7j0P=V1M3M3M3M3M3N2M3M3M3M3M3M2N3M3H8\\Od0]Oc0\\Od0\\Od0\\Od0]Oc0H8I7I7I7I7I7I7J6I7I7I7I7I7I7J6I7K5M3L4L4M3L4L4M3L4M3L4L4M3L4M3L4L4M2M4M3M3N2M3N2000000000000000000001O000000000000000000000000001O00000000000000000000000O101O000000000000000000000000001O000000000000000000000000001O0000000000000000000000000O101O000000000000000000000000001O0000000000000000000000001O00001O00001O0000001O00001O0O101O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O0O101O0000001O00001O00001O00001O000eK^MXLb2]3jMbLW2Q3VNnLj1f2cNYM]1[2oNeMR1o1ZOPNf0e1FZN:e1HZN9e1HZN8e1I[N7d1K[N9a1H^N<]1FbN>Y1DfN`0V1@jNd0Q1^OnNf0m0\\OROh0i0ZOVOj0f0WOYOm0b0TO^OQ1<QOCS19nNFV15lNJX11iNO[1MfN2^1IiN1[1JlNOY1MnNLV1OQOKS10TOJP11WOGP14WO_OS1<TOYOU1b0ROSOW1c0VOQOe8D<EejP5"}, "label": "a green bus parked next to a red bus"}]}
{"id": 353607, "image": "COCO_train2014_000000353607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark green double decker tour bus parked next to another bus\"."}], "task": "refering", "answer_template": "The \"a dark green double decker tour bus parked next to another bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 48, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.02190625, 0.1016875, 0.46179687499999994, 0.9870833333333333], "iscrowd": 0, "area": 99959.38130000002, "rle": {"size": [480, 640], "counts": "PP7j0P=V1M3M3M3M3M3N2M3M3M3M3M3M2N3M3H8\\Od0]Oc0\\Od0\\Od0\\Od0]Oc0H8I7I7I7I7I7I7J6I7I7I7I7I7I7J6I7K5M3L4L4M3L4L4M3L4M3L4L4M3L4M3L4L4M2M4M3M3N2M3N2000000000000000000001O000000000000000000000000001O00000000000000000000000O101O000000000000000000000000001O000000000000000000000000001O0000000000000000000000000O101O000000000000000000000000001O0000000000000000000000001O00001O00001O0000001O00001O0O101O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O0O101O0000001O00001O00001O00001O000eK^MXLb2]3jMbLW2Q3VNnLj1f2cNYM]1[2oNeMR1o1ZOPNf0e1FZN:e1HZN9e1HZN8e1I[N7d1K[N9a1H^N<]1FbN>Y1DfN`0V1@jNd0Q1^OnNf0m0\\OROh0i0ZOVOj0f0WOYOm0b0TO^OQ1<QOCS19nNFV15lNJX11iNO[1MfN2^1IiN1[1JlNOY1MnNLV1OQOKS10TOJP11WOGP14WO_OS1<TOYOU1b0ROSOW1c0VOQOe8D<EejP5"}, "label": "a dark green double decker tour bus parked next to another bus"}]}
{"id": 247114, "image": "COCO_train2014_000000247114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding the sandwiches\"."}], "task": "refering", "answer_template": "The \"the woman holding the sandwiches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 66, 67, 68, 69, 81, 82, 83, 84, 96, 97, 98, 99, 100, 111, 112, 113, 114, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 154, 155, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 229, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 274, 275, 276, 277, 278, 279, 280, 281, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 320, 321, 322, 323, 324, 325, 326, 327, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.27557377049180326, 0.161703125, 0.8233489461358313, 0.9868125000000001], "iscrowd": 0, "area": 78404.19194999998, "rle": {"size": [640, 427], "counts": "aRZ2:`c0;F;D;F;D;F;D6L11O2N2N101N1O2Nk@Mb:4_E1Y:0hE7P:JPF<h9EZFa0^9@bFo0n8QOSG^1\\8cNfGa1T8`NlGd1n7]NTHf1g7[NZHi1_7YNbHj1Y7WNiHk1Q7WNPIl1j6WNWIl1c6UN^In1^6RNcIQ2X6PNjIP2T6QNmIo1P6SNQJm1m5TNUJl1g5VNZJj1d5WN^Jh1_5ZNbJf1\\5\\NeJb1Y5aNgJ_1W5cNiJ]1T5fNlJY1S5iNnJV1o4mNRKR1l4POUKn0j4TOWKj0h4XOZKd0e4_OfK4Z4NgKOY4;aKA_4k0XKROg4Q1[KkNe4W1\\KfNd4\\1]KaNc4`1]K_Nb4d1]K[Nc4g1\\KXNd4j1[KUNe4m1[KQNd4Q2\\KnMd4T2[KkMe4W2ZKhMQ2SMoLX5m0fMP2ZMoLQ5o0fMn1bMnLj4Q1eMm1iMnLd4Q1eMm1oMmL_4S1cMn1SNlL[4S1dMo1UNlLY4R1cMo1ZNlLU4Q1cMQ2\\NkLT4Q1aMR2`NhLR4S1`MR2cNhLP4S1^MT2fNdLP4U1[MU2iNbLP4U1YMX2kN^Lo3o0`Mb2dN[LP4b0kMR3YNXLo36VNa3W5oKWKP4]:O1N2O1O1O1O1O1O1N2O1O1O1O100O100O100O10000O100O100O100O100O100O100O1001O0000000000000000000000000000001O00000000000000001O1O001O1O1O1O1O001O1O1O1O1O001O1O3^CTLj8o3lFXLT9k3cF[L]9h3YF`Lf9d3PFbLP:b3gEcLY:a3`EbL`:e3VE_Li:h3nDZLR;m3dDVL\\;Q4[DRLd;T4SDoKm;R60001O001O0XK`HPM`7n2dHPM]7n2fHPMZ7o2jHnLW7o2nHnLR7Q3QImLo6Q3VIlLk6R3XIlLh61RHTOY1j0e60VHSOW1l0d6NYHROV1o0a6M]HQOT1Q1`6K`HPOS1T1]6JdHPOP1U1\\6HiHPOm0W1[6FlHQOj0X1Z6GnHnNj0Z1Y6FPInNi0Z1W6HRIkNh0]1V6GUIiNg0_1U6GVIaNm0g1m5GYI`Mj1i2m4G[IPMW2X3_4FVM:j2FWM9j2FWM9i2FXM:h2FYM9h2EZM:f2FZM:g2D[M;e2E[M;e2E\\M:e2D]M;d2D\\M<f2A\\M>f2@ZM`0i2\\OYMc0i2[OXMd0j2YOWMg0l2UOVMj0l2SOVMl0l2QOUMo0o:OS1mN4L5K4L4L4L5K4L4L5K4L4L4L5K4LYX^1"}, "label": "the woman holding the sandwiches"}]}
{"id": 247114, "image": "COCO_train2014_000000247114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding sandwiches\"."}], "task": "refering", "answer_template": "The \"a woman holding sandwiches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 66, 67, 68, 69, 81, 82, 83, 84, 96, 97, 98, 99, 100, 111, 112, 113, 114, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 154, 155, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 229, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 274, 275, 276, 277, 278, 279, 280, 281, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 320, 321, 322, 323, 324, 325, 326, 327, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.27557377049180326, 0.161703125, 0.8233489461358313, 0.9868125000000001], "iscrowd": 0, "area": 78404.19194999998, "rle": {"size": [640, 427], "counts": "aRZ2:`c0;F;D;F;D;F;D6L11O2N2N101N1O2Nk@Mb:4_E1Y:0hE7P:JPF<h9EZFa0^9@bFo0n8QOSG^1\\8cNfGa1T8`NlGd1n7]NTHf1g7[NZHi1_7YNbHj1Y7WNiHk1Q7WNPIl1j6WNWIl1c6UN^In1^6RNcIQ2X6PNjIP2T6QNmIo1P6SNQJm1m5TNUJl1g5VNZJj1d5WN^Jh1_5ZNbJf1\\5\\NeJb1Y5aNgJ_1W5cNiJ]1T5fNlJY1S5iNnJV1o4mNRKR1l4POUKn0j4TOWKj0h4XOZKd0e4_OfK4Z4NgKOY4;aKA_4k0XKROg4Q1[KkNe4W1\\KfNd4\\1]KaNc4`1]K_Nb4d1]K[Nc4g1\\KXNd4j1[KUNe4m1[KQNd4Q2\\KnMd4T2[KkMe4W2ZKhMQ2SMoLX5m0fMP2ZMoLQ5o0fMn1bMnLj4Q1eMm1iMnLd4Q1eMm1oMmL_4S1cMn1SNlL[4S1dMo1UNlLY4R1cMo1ZNlLU4Q1cMQ2\\NkLT4Q1aMR2`NhLR4S1`MR2cNhLP4S1^MT2fNdLP4U1[MU2iNbLP4U1YMX2kN^Lo3o0`Mb2dN[LP4b0kMR3YNXLo36VNa3W5oKWKP4]:O1N2O1O1O1O1O1O1N2O1O1O1O100O100O100O10000O100O100O100O100O100O100O1001O0000000000000000000000000000001O00000000000000001O1O001O1O1O1O1O001O1O1O1O1O001O1O3^CTLj8o3lFXLT9k3cF[L]9h3YF`Lf9d3PFbLP:b3gEcLY:a3`EbL`:e3VE_Li:h3nDZLR;m3dDVL\\;Q4[DRLd;T4SDoKm;R60001O001O0XK`HPM`7n2dHPM]7n2fHPMZ7o2jHnLW7o2nHnLR7Q3QImLo6Q3VIlLk6R3XIlLh61RHTOY1j0e60VHSOW1l0d6NYHROV1o0a6M]HQOT1Q1`6K`HPOS1T1]6JdHPOP1U1\\6HiHPOm0W1[6FlHQOj0X1Z6GnHnNj0Z1Y6FPInNi0Z1W6HRIkNh0]1V6GUIiNg0_1U6GVIaNm0g1m5GYI`Mj1i2m4G[IPMW2X3_4FVM:j2FWM9j2FWM9i2FXM:h2FYM9h2EZM:f2FZM:g2D[M;e2E[M;e2E\\M:e2D]M;d2D\\M<f2A\\M>f2@ZM`0i2\\OYMc0i2[OXMd0j2YOWMg0l2UOVMj0l2SOVMl0l2QOUMo0o:OS1mN4L5K4L4L4L5K4L4L5K4L4L4L5K4LYX^1"}, "label": "a woman holding sandwiches"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that the plate is on\"."}], "task": "refering", "answer_template": "The \"the table that the plate is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 18, 19, 20, 21, 36, 38, 54, 55, 72, 162, 180, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233], "bbox": [0.0, 0.004506666666666667, 0.99776, 0.9887733333333334], "iscrowd": 0, "area": 15300.505399999996, "rle": {"size": [375, 500], "counts": "^8m1e1WOWOG`4U2oJYNa0i0[3l0TL[Na0l0Z3g0UL]Na0j0^3g0QL`Na0g0`3h0oKaNa0e0d3h0kKcNa0d0g3g0hKeNa0b0j3h0eKfNb0`0l3h0bKiNa0=P4i0_KjNa0<R4i0]KkNb09U4j0YKmNb08W4j0WKnNb06Z4k0TKPOa04^4j0QKROb01`4l0nJSOb0Oc4m0kJTOb0Nf4l0hJWOb0Ji4n0eJXOb0Ik4n0cJYOb0Go4n0_J[Ob0FQ5n0]J\\Oc0CS5P1ZJ^Ob0AW5o0WJ@b0_OZ5P1TJAb0^O\\5P1RJBc0[O_5Q1nIDc0ZOa5Q1lIFb0WOe5R1iIGc0UOg5R1fIIc0SOj5S1cIJn75QHLP83PHNP81PHOQ80oG0R8OnG1S8NmG3R8MnG3S8LmG4T8KlG5U8JkG6V8IjG8U8HkG8V8GjG9W8FiG:X8EhG;Y8DgG=X8ChG=Y8BgG>Z8AfG?[8@eG`0\\8_OdGb0\\8]OdGc0\\8]OdGc0]8\\OcGd0^8[ObGf0^8YObGg0_8XOaGh0_8XOaGh0`8WO`Gi0a8VO_G[1Q8dNoG[1S8dNmG\\1S8dNmG[1U8dNkG\\1V8cNjG\\1X8cNhG\\1Z8cNfG[1\\8eNdGZ1^8eNbGY1`8gN`GW1c8hN]GV1e8jN[GU1f8kNZG:5\\Od89WG;b9E^F:c9F\\F;e9D[F<e9D[F<f9CZF=f9CZF=f9CZF=g9BYF>g9BYF>g9BYF=i9BWF>i9BWF>j9AVF?j9AVF?j9AVF?k9@UF`0k9@UF`0k9@UF`0l9_OTF`0m9@SF`0m9@SF`0n9_ORFa0n9_ORFa0o9^OQFb0o9^OQFb0o9^OQFb0P:]OPFb0Q:^OoEb0Q:^OoE`0T:_OlE?V:AjE=Y:BgE<[:DeE:]:FcE9_:FaE8a:H_E6c:J]E4f:KZE3h:MXE1k:NUE0m:0SENo:2<O00001O00001O00ki[20UVdM000O1000000O10000O10000O10000O10000O10000O10000O10000O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O1000000001O00000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000O1M3L4M3M3L4M3N2001O00001O001O00001O001mNl2"}, "label": "the table that the plate is on"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table holding a plate with a sandwich on it\"."}], "task": "refering", "answer_template": "The \"a wooden table holding a plate with a sandwich on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 70, 71], "bbox": [0.65898, 0.006373333333333333, 1.0, 0.2737866666666667], "iscrowd": 0, "area": 11613.39625, "rle": {"size": [375, 500], "counts": "Ykh32e;2N2N2N2UEIR:9iEJW:8dEL[:5bEN]:4^E0a:d0O001O1O1O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O000000001O000000001O00000000001O000000001O00000000001O0000002N1O1O2N1O1O1O2N1O1O2N1O1O1O2NO100000O1000O10000000000O10000000000O10000000000OO"}, "label": "a wooden table holding a plate with a sandwich on it"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table\"."}], "task": "refering", "answer_template": "The \"the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 70, 71], "bbox": [0.65898, 0.006373333333333333, 1.0, 0.2737866666666667], "iscrowd": 0, "area": 11613.39625, "rle": {"size": [375, 500], "counts": "Ykh32e;2N2N2N2UEIR:9iEJW:8dEL[:5bEN]:4^E0a:d0O001O1O1O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O000000001O000000001O00000000001O000000001O00000000001O0000002N1O1O2N1O1O1O2N1O1O2N1O1O1O2NO100000O1000O10000000000O10000000000O10000000000OO"}, "label": "the table"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich with egg inside it\"."}], "task": "refering", "answer_template": "The \"sandwich with egg inside it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 116, 130], "bbox": [0.05718, 0.019280000000000002, 0.70292, 0.5506666666666666], "iscrowd": 0, "area": 43706.127049999996, "rle": {"size": [375, 500], "counts": "af:2d;2N2M3N3M2N2O1O1N2O1O2YEHk9T1N000000O1000O1000O100000000O100000O10O100002N2N1O2N2N1N1N2L4UOPNkGU2S8h0O1O1O1O1O2N1O1O1O100O1PHlL]7U3XHVMg7^3O1O1O1N2O1O=C0000001N1000000000O010O100O1O10O01O11O2N1O2N2M5L6J7I6J5KO0100000O100000O01000000O10O1000O100000O10O1000000O10O1000O100000O01000000O1000O1000O100000O01000000O10O1000O1000000O010000000O10O1000O1000000O0100000O1000O10O10000000O01000000O1000O10O1000000O010000000O1000000O1000000O10O1000O1000000O100000000O1000000O1000000O1000000O1000000O10000000O0100O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O101N1O100O2N100O2N100O1O2O0O1O101N1O100O2N100O2_NoGYOR8g0RHUOn7\\OnGh07Hm7^OnGi08@P8FjGi09XOS8NeGj0;oNW85`Gk0=fNY8>\\Gk0Z9TOgFl0Y9SOiFl0X9ROjFm0V9ROkFn0U9ROlFn0T9POnFo0R9POoFP1Q9oNQGP1P9nNRGQ1n8nNSGR1m8mNYGn0l9K5J7H7JP`f1"}, "label": "sandwich with egg inside it"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich fast in back of another sandwich ass\"."}], "task": "refering", "answer_template": "The \"the sandwich fast in back of another sandwich ass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 116, 130], "bbox": [0.05718, 0.019280000000000002, 0.70292, 0.5506666666666666], "iscrowd": 0, "area": 43706.127049999996, "rle": {"size": [375, 500], "counts": "af:2d;2N2M3N3M2N2O1O1N2O1O2YEHk9T1N000000O1000O1000O100000000O100000O10O100002N2N1O2N2N1N1N2L4UOPNkGU2S8h0O1O1O1O1O2N1O1O1O100O1PHlL]7U3XHVMg7^3O1O1O1N2O1O=C0000001N1000000000O010O100O1O10O01O11O2N1O2N2M5L6J7I6J5KO0100000O100000O01000000O10O1000O100000O10O1000000O10O1000O100000O01000000O1000O1000O100000O01000000O10O1000O1000000O010000000O10O1000O1000000O0100000O1000O10O10000000O01000000O1000O10O1000000O010000000O1000000O1000000O10O1000O1000000O100000000O1000000O1000000O1000000O1000000O10000000O0100O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O101N1O100O2N100O2N100O1O2O0O1O101N1O100O2N100O2_NoGYOR8g0RHUOn7\\OnGh07Hm7^OnGi08@P8FjGi09XOS8NeGj0;oNW85`Gk0=fNY8>\\Gk0Z9TOgFl0Y9SOiFl0X9ROjFm0V9ROkFn0U9ROlFn0T9POnFo0R9POoFP1Q9oNQGP1P9nNRGQ1n8nNSGR1m8mNYGn0l9K5J7H7JP`f1"}, "label": "the sandwich fast in back of another sandwich ass"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partially eaten sandwich half to the right of an uneaten sandwich half\"."}], "task": "refering", "answer_template": "The \"a partially eaten sandwich half to the right of an uneaten sandwich half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.44662, 0.18877333333333335, 1.0, 0.8898933333333334], "iscrowd": 0, "area": 54693.671200000004, "rle": {"size": [375, 500], "counts": "eXb22c;3N1N2N2O2M2N2O2M2N2O2M2N2N2O2M2N2O2lFlNa7V1]HlNa7V1\\HmNc7U1ZHmNd7U1YHnNe7T1YHnNf7S1XHoNf7T1VHoNh7S1VHoNi7R1THQOj7R1SHQOj7Q1SHROk7P1SHROl7P1QHROm7P1PHSOn7o0PHSOo7n0nGUOP8Z2N2O1N2N2O1N2N3N1N2N2O1N2N2O1O1O100O1O1O:F6K0O1O1O2N100O1O1O100O2O0O100O100O2O0O100O100O2O0O100O100O2O0O100O100O101N100O100[OVJ\\Kk5d4WJZKj5d4ZJXKg5h4[JVKe5i4_JSKb5l4aJRK_5n4e0O2O0O1O100O100O2O0O1O100O101N100O1O100O2O0O1O100N2N3N1N2O1N2N2O1N2N2O1N2N2O1N2O10000000000000001O000000000000000000000000000000000000000000000010O0000000000000000001O00000000001O000000001O00010O001O00001O00001O001O0000001O00001O00001O0001O01O00001O00001O0000001ON2O1O1O1O1O1O1O1O1NmHaLR6^3gIjLY6U3^IUMc6i2TI`Mm6e30000000O10N2N2M2N3N2M3M2O2M3N21N101O00001N101O001N101O0iL"}, "label": "a partially eaten sandwich half to the right of an uneaten sandwich half"}]}
{"id": 296267, "image": "COCO_train2014_000000296267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a blt sandwich on a white plate\"."}], "task": "refering", "answer_template": "The \"half of a blt sandwich on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.44662, 0.18877333333333335, 1.0, 0.8898933333333334], "iscrowd": 0, "area": 54693.671200000004, "rle": {"size": [375, 500], "counts": "eXb22c;3N1N2N2O2M2N2O2M2N2O2M2N2N2O2M2N2O2lFlNa7V1]HlNa7V1\\HmNc7U1ZHmNd7U1YHnNe7T1YHnNf7S1XHoNf7T1VHoNh7S1VHoNi7R1THQOj7R1SHQOj7Q1SHROk7P1SHROl7P1QHROm7P1PHSOn7o0PHSOo7n0nGUOP8Z2N2O1N2N2O1N2N3N1N2N2O1N2N2O1O1O100O1O1O:F6K0O1O1O2N100O1O1O100O2O0O100O100O2O0O100O100O2O0O100O100O2O0O100O100O101N100O100[OVJ\\Kk5d4WJZKj5d4ZJXKg5h4[JVKe5i4_JSKb5l4aJRK_5n4e0O2O0O1O100O100O2O0O1O100O101N100O1O100O2O0O1O100N2N3N1N2O1N2N2O1N2N2O1N2N2O1N2O10000000000000001O000000000000000000000000000000000000000000000010O0000000000000000001O00000000001O000000001O00010O001O00001O00001O001O0000001O00001O00001O0001O01O00001O00001O0000001ON2O1O1O1O1O1O1O1O1NmHaLR6^3gIjLY6U3^IUMc6i2TI`Mm6e30000000O10N2N2M2N3N2M3M2O2M3N21N101O00001N101O001N101O0iL"}, "label": "half of a blt sandwich on a white plate"}]}
{"id": 443725, "image": "COCO_train2014_000000443725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with white blankets next to a couch with pillows\"."}], "task": "refering", "answer_template": "The \"a bed with white blankets next to a couch with pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.035390624999999995, 0.38164583333333335, 0.9016874999999999, 0.9906458333333333], "iscrowd": 0, "area": 99221.2184, "rle": {"size": [480, 640], "counts": "gS;8d>:F:F:F:H8J6I7I7J5J3N2M3N2M3M3N2M3N2M3M2O2M3N2N2M3N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2M2O2N2N2N1O0O1000001O0O10000000000O100000000O100000001O0O100000000O10000000000O100000001N100000000O10000000000O100000000O101O0000000O100000000O10000000000O100000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000O100O01000O100O10O10O100O10O010000O010O10000O010O100O10O10O100O10O10O100O10O010000O10O010000O010O100O10O10O100O10O10O100O100O01000O100O01000O100ON3L3L5L4K4M4K4L5L4K4M4K4M4K4M4M30O2O0O101O0O2O0O2O0O101O0O2O0O2O001N100O2O0O2O00O0O2M2N3N2M2O21N10001N101N101O0O101N101O0O101O0O2O0O2O000O2O0O2O000O2O0O2O001N100O2O001N100O2O001N101O0O101N101O0O101N101O0O2O0O10001N100O10001N100O10001N10000O2O0O10000O2O0O10000O2O0O10000O2O0O10001N10000O101N10000O101N10000O101N10000O101N10000O2O0O10000O2O000O100_MPF1R:JSF5m9GWF9i9B]F<d9^ObFb0^9XOiFf0Y9SOnFl0R9nNTGR1l8iNZGU1g8eN_G[1b8^NeG`1\\8ZNkGe1U8UNQHk1o7PNWHn1^:01N2O001N101N2O001N101N2J5J7J5J7J6J5J7JUWm0"}, "label": "a bed with white blankets next to a couch with pillows"}]}
{"id": 443725, "image": "COCO_train2014_000000443725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a twin bed close to the window\"."}], "task": "refering", "answer_template": "The \"a twin bed close to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.035390624999999995, 0.38164583333333335, 0.9016874999999999, 0.9906458333333333], "iscrowd": 0, "area": 99221.2184, "rle": {"size": [480, 640], "counts": "gS;8d>:F:F:F:H8J6I7I7J5J3N2M3N2M3M3N2M3N2M3M2O2M3N2N2M3N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2M2O2N2N2N1O0O1000001O0O10000000000O100000000O100000001O0O100000000O10000000000O100000001N100000000O10000000000O100000000O101O0000000O100000000O10000000000O100000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000O100O01000O100O10O10O100O10O010000O010O10000O010O100O10O10O100O10O10O100O10O010000O10O010000O010O100O10O10O100O10O10O100O100O01000O100O01000O100ON3L3L5L4K4M4K4L5L4K4M4K4M4K4M4M30O2O0O101O0O2O0O2O0O101O0O2O0O2O001N100O2O0O2O00O0O2M2N3N2M2O21N10001N101N101O0O101N101O0O101O0O2O0O2O000O2O0O2O000O2O0O2O001N100O2O001N100O2O001N101O0O101N101O0O101N101O0O2O0O10001N100O10001N100O10001N10000O2O0O10000O2O0O10000O2O0O10000O2O0O10001N10000O101N10000O101N10000O101N10000O101N10000O2O0O10000O2O000O100_MPF1R:JSF5m9GWF9i9B]F<d9^ObFb0^9XOiFf0Y9SOnFl0R9nNTGR1l8iNZGU1g8eN_G[1b8^NeG`1\\8ZNkGe1U8UNQHk1o7PNWHn1^:01N2O001N101N2O001N101N2J5J7J5J7J6J5J7JUWm0"}, "label": "a twin bed close to the window"}]}
{"id": 230735, "image": "COCO_train2014_000000230735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child facing a mirror brushing his teeth\"."}], "task": "refering", "answer_template": "The \"a child facing a mirror brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 69, 70, 71, 72, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 121, 122, 123, 124, 134, 135, 136, 137, 147, 148, 149, 150, 160, 161, 162, 163, 173, 174, 175, 176, 186, 187, 188, 199, 200, 201], "bbox": [0.26965333333333336, 0.26966, 0.6651733333333333, 0.8921400000000002], "iscrowd": 0, "area": 26884.732450000007, "rle": {"size": [500, 375], "counts": "a`a17o>?O2N1O2N1O2N1O2N1O2M2O2N2N1O2N1SGmNW3[1QLVOn3X1TKYOk4U1WJ[Oi5Y5O1O1O001O1N1M4M3M3M2N3M3L4M2N3M3M2M4M3N2N9F2OL3M4L4M2M4L3M4O1N1O2O1N10kISHf4k7WK[Hi4b7SKeHm4Y7nJnHR5o6jJXIV5f6eJaI[5\\6aJkI^5S6^JTJb5o70001O0000O100O10O10O100O100O100O10000O100O101L3L4L4M3L4M3L4M4K4M3L4L5L5J5L5J5K6@?A`0@`0@?A`0@?D=M4Kb0^O3N2M3M2O2M3M3N2M3M3N1N3M3N2M3M3N2M3M3N1N3M3M3N2M3M3M3N3L3M4M2M4L3M4M2MVZm1"}, "label": "a child facing a mirror brushing his teeth"}]}
{"id": 230735, "image": "COCO_train2014_000000230735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the picture , not the image\"."}], "task": "refering", "answer_template": "The \"the boy in the picture , not the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 69, 70, 71, 72, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 121, 122, 123, 124, 134, 135, 136, 137, 147, 148, 149, 150, 160, 161, 162, 163, 173, 174, 175, 176, 186, 187, 188, 199, 200, 201], "bbox": [0.26965333333333336, 0.26966, 0.6651733333333333, 0.8921400000000002], "iscrowd": 0, "area": 26884.732450000007, "rle": {"size": [500, 375], "counts": "a`a17o>?O2N1O2N1O2N1O2N1O2M2O2N2N1O2N1SGmNW3[1QLVOn3X1TKYOk4U1WJ[Oi5Y5O1O1O001O1N1M4M3M3M2N3M3L4M2N3M3M2M4M3N2N9F2OL3M4L4M2M4L3M4O1N1O2O1N10kISHf4k7WK[Hi4b7SKeHm4Y7nJnHR5o6jJXIV5f6eJaI[5\\6aJkI^5S6^JTJb5o70001O0000O100O10O10O100O100O100O10000O100O101L3L4L4M3L4M3L4M4K4M3L4L5L5J5L5J5K6@?A`0@`0@?A`0@?D=M4Kb0^O3N2M3M2O2M3M3N2M3M3N1N3M3N2M3M3N2M3M3N1N3M3M3N2M3M3M3N3L3M4M2M4L3M4M2MVZm1"}, "label": "the boy in the picture , not the image"}]}
{"id": 165199, "image": "COCO_train2014_000000165199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in baseball uniform with dark hair and eyes standing to the right of another , shorter baseball player\"."}], "task": "refering", "answer_template": "The \"a man in baseball uniform with dark hair and eyes standing to the right of another , shorter baseball player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 370, 371, 372, 373, 374, 375, 376, 377, 379, 389, 390, 391, 392, 393, 394, 395, 408, 409, 410, 411, 412, 413, 414, 415, 428, 429, 430, 431, 432, 433, 434, 435], "bbox": [0.4521509433962264, 0.130265625, 0.9963584905660376, 0.9890156250000001], "iscrowd": 0, "area": 114734.4385, "rle": {"size": [640, 530], "counts": "iXf4d0[c0f0ZOf0YOg0ZO?@8I7H8I8H7H8I8G8I7H8I8H7H9H8G:G8G:G8H8G:G8G9H9F9H9G8G9H8G9H7H9H8H8G8I8G9H6I2O1O1N2O1N2O1N2O1O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2F;]Ob0^Ob0J6N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O10000000000O10000000000O10000000000O10000000000O2O00000000000000000000000000000000000000000000000000000000000000000000000000000002N3M2O2gAmL_;U3^DlLb;V3ZDlLf;W3VDkLi;W3SDkLm;X3oCiLQ<Y3kCiLU<Y3hCiLW<Z3dChL\\<Z3aCgL_<\\3\\CfLd<]3XCeLg<`3RCbLn<c3lB^LT=f3gB[LY=j3`BYL_=l3[BULe=P4TBRLl=R5O0000001O0000001O0000001SNSJeEm5W:YJgEg5T:`JjE`5Q:hJlEY5o9mJoES5o9QKoEo4P:TKnEm4o9XKnEh4P:\\KnEd4P:`KnE`4P:cKoE^4n9eKQF[4m9gKSFY4k9jKUFU4i9mKWFT4f9oKYFQ4e9RLZFn3d9TL\\Fl3b9WL]Fj3a9WL_Fi3_9ZL`Ff3^9\\LbFd3^9]LaFd3^9\\LbFd3]9^LbFb3^9^LbFc3]9^LbFb3]9_LcFa3]9`LbF`3^9`LbFb3[9`LdFb3Z9^LfFd3W9^LhFd3V9\\LjFf3T9ZLlFh3Q9YLnFj3P9VLoFm3o8SLQGo3l8RLSGQ4k8PLTGR4j8nKUGU4h8lKWGX4f8hKZGZ4d8fK[G]4b8dK]G_4a8aK_Ga4_8`K_Gc4^8^KbGd4\\8\\KcGg4[8YKdGj4Y8WKgGk4W8UKhGn4V8RKiGQ5T8QKkGQ5S8oJlGT5R8lJmGW5P8jJPHY5m7gJRH\\5l7dJTH^5k7bJSHa5l7[3O1O1O1O1O100O1O3M2N3M2O2M2N2N3M2N3M2O2M2hKjBa1Y=PN^Ce1d<mMTDg1o;kMgDj1[;hM]Em1e:dMRFQ2[>@?YOh0XObh0"}, "label": "a man in baseball uniform with dark hair and eyes standing to the right of another , shorter baseball player"}]}
{"id": 165199, "image": "COCO_train2014_000000165199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller of two men in an old picture wearing baseball uniforms\"."}], "task": "refering", "answer_template": "The \"the taller of two men in an old picture wearing baseball uniforms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 370, 371, 372, 373, 374, 375, 376, 377, 379, 389, 390, 391, 392, 393, 394, 395, 408, 409, 410, 411, 412, 413, 414, 415, 428, 429, 430, 431, 432, 433, 434, 435], "bbox": [0.4521509433962264, 0.130265625, 0.9963584905660376, 0.9890156250000001], "iscrowd": 0, "area": 114734.4385, "rle": {"size": [640, 530], "counts": "iXf4d0[c0f0ZOf0YOg0ZO?@8I7H8I8H7H8I8G8I7H8I8H7H9H8G:G8G:G8H8G:G8G9H9F9H9G8G9H8G9H7H9H8H8G8I8G9H6I2O1O1N2O1N2O1N2O1O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2F;]Ob0^Ob0J6N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O10000000000O10000000000O10000000000O10000000000O2O00000000000000000000000000000000000000000000000000000000000000000000000000000002N3M2O2gAmL_;U3^DlLb;V3ZDlLf;W3VDkLi;W3SDkLm;X3oCiLQ<Y3kCiLU<Y3hCiLW<Z3dChL\\<Z3aCgL_<\\3\\CfLd<]3XCeLg<`3RCbLn<c3lB^LT=f3gB[LY=j3`BYL_=l3[BULe=P4TBRLl=R5O0000001O0000001O0000001SNSJeEm5W:YJgEg5T:`JjE`5Q:hJlEY5o9mJoES5o9QKoEo4P:TKnEm4o9XKnEh4P:\\KnEd4P:`KnE`4P:cKoE^4n9eKQF[4m9gKSFY4k9jKUFU4i9mKWFT4f9oKYFQ4e9RLZFn3d9TL\\Fl3b9WL]Fj3a9WL_Fi3_9ZL`Ff3^9\\LbFd3^9]LaFd3^9\\LbFd3]9^LbFb3^9^LbFc3]9^LbFb3]9_LcFa3]9`LbF`3^9`LbFb3[9`LdFb3Z9^LfFd3W9^LhFd3V9\\LjFf3T9ZLlFh3Q9YLnFj3P9VLoFm3o8SLQGo3l8RLSGQ4k8PLTGR4j8nKUGU4h8lKWGX4f8hKZGZ4d8fK[G]4b8dK]G_4a8aK_Ga4_8`K_Gc4^8^KbGd4\\8\\KcGg4[8YKdGj4Y8WKgGk4W8UKhGn4V8RKiGQ5T8QKkGQ5S8oJlGT5R8lJmGW5P8jJPHY5m7gJRH\\5l7dJTH^5k7bJSHa5l7[3O1O1O1O1O100O1O3M2N3M2O2M2N2N3M2N3M2O2M2hKjBa1Y=PN^Ce1d<mMTDg1o;kMgDj1[;hM]Em1e:dMRFQ2[>@?YOh0XObh0"}, "label": "the taller of two men in an old picture wearing baseball uniforms"}]}
{"id": 165199, "image": "COCO_train2014_000000165199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with very long hair and a beard is standing in his baseball uniform next to a man with shorter long hair and a bigger beard\"."}], "task": "refering", "answer_template": "The \"a man with very long hair and a beard is standing in his baseball uniform next to a man with shorter long hair and a bigger beard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 78, 79, 80, 81, 97, 98, 99, 100, 115, 116, 117, 118, 119, 134, 135, 136, 137, 138, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 252, 253, 254, 255, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 327, 328, 329, 330, 331, 342, 343, 344, 345, 346, 347, 348, 349, 350, 361, 362, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384, 385, 386, 387, 388, 399, 400, 401, 402, 403, 404, 405, 406, 407, 418, 419, 420, 421, 422, 423, 424, 425], "bbox": [0.00269811320754717, 0.170296875, 0.473811320754717, 0.9841875], "iscrowd": 0, "area": 101807.44184999999, "rle": {"size": [640, 530], "counts": "Pm07gc0Q2nMR2oMk1UN7H8I7jE\\Id6k6cHkIX7]6nGZJo7l5WGkJe8Z8N3M3M3M3M2N3M3M3M3M2N2N1N2O1O1N2M3M3L4M3L4M3M3L4M3L4M3L4M3M3L4L4M3L4M3L4M3L4M3L4O1N2O1O1O1N3N1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1N3N1000000000000000000000000000000000000000000O10000000000000000000000001O001O0010O01O1O1O1O2N1O1O2N1O1O2N1O2N2N1O2N1O2N2N1O2QCaKY:`4XEnKh:T4iDZLV;h3[DeLe;l5O1O001O001O1O001O001O1O0010O01O1O001O1O001O001O1O0O2N1O2N2O0O2N1O2N2N101N2N3M2N2N2O2M2N2N2N3M2O2M3M2N3M3N2M3N2N2M3N2N2M3N3M4K4M3M4K4M4K4M4L3L4M4L3L5L3M4K4M4L3RJTDa3P<YLQEi2S;QMoEo1T:kMoFT1V9fNmG9V8BlH]OY7<Z6K5J7JQP^5"}, "label": "a man with very long hair and a beard is standing in his baseball uniform next to a man with shorter long hair and a bigger beard"}]}
{"id": 165199, "image": "COCO_train2014_000000165199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the long hair on the left\"."}], "task": "refering", "answer_template": "The \"the man with the long hair on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 78, 79, 80, 81, 97, 98, 99, 100, 115, 116, 117, 118, 119, 134, 135, 136, 137, 138, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 252, 253, 254, 255, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 327, 328, 329, 330, 331, 342, 343, 344, 345, 346, 347, 348, 349, 350, 361, 362, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384, 385, 386, 387, 388, 399, 400, 401, 402, 403, 404, 405, 406, 407, 418, 419, 420, 421, 422, 423, 424, 425], "bbox": [0.00269811320754717, 0.170296875, 0.473811320754717, 0.9841875], "iscrowd": 0, "area": 101807.44184999999, "rle": {"size": [640, 530], "counts": "Pm07gc0Q2nMR2oMk1UN7H8I7jE\\Id6k6cHkIX7]6nGZJo7l5WGkJe8Z8N3M3M3M3M2N3M3M3M3M2N2N1N2O1O1N2M3M3L4M3L4M3M3L4M3L4M3L4M3M3L4L4M3L4M3L4M3L4M3L4O1N2O1O1O1N3N1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1N3N1000000000000000000000000000000000000000000O10000000000000000000000001O001O0010O01O1O1O1O2N1O1O2N1O1O2N1O2N2N1O2N1O2N2N1O2QCaKY:`4XEnKh:T4iDZLV;h3[DeLe;l5O1O001O001O1O001O001O1O0010O01O1O001O1O001O001O1O0O2N1O2N2O0O2N1O2N2N101N2N3M2N2N2O2M2N2N2N3M2O2M3M2N3M3N2M3N2N2M3N2N2M3N3M4K4M3M4K4M4K4M4L3L4M4L3L5L3M4K4M4L3RJTDa3P<YLQEi2S;QMoEo1T:kMoFT1V9fNmG9V8BlH]OY7<Z6K5J7JQP^5"}, "label": "the man with the long hair on the left"}]}
{"id": 9557, "image": "COCO_train2014_000000009557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a black wheelchair with a boy wearing a grey and red shirt sitting in it\"."}], "task": "refering", "answer_template": "The \"the back of a black wheelchair with a boy wearing a grey and red shirt sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 154, 155], "bbox": [0.41618, 0.2628, 0.65858, 0.6922666666666667], "iscrowd": 0, "area": 11926.8441, "rle": {"size": [375, 500], "counts": "d[\\29\\;;G9F:F:G9]OYNcFi1W9?I7N3M2N2N2N2O0O1O1N101O001N2O1L4J6K5J5C:MO0mNlHVMo7k2?11OO1O100O1O1O010O001O1O010O001O001O0010O01O001O001O001O00001O2N3M3M1O1O010O001O001ON1O2VOQHiMR8Y2PHaMR8`2QHZMP8h2RHRMQ8P3:1O1N2O2N2N4K4M3M3L5K3M3L4M3M3M3N2NVMRIm0l6jN^IV1b6YNoIg1R6RNTJn1n5oMSJQ2m5nMSJR2n5nMRJR2n5oMRJP2n5oMSJP2o5mMTJP2o5mMTJP2o5nMRJP2Q6nMQJo1R6nMQJo1R6oMoIP2S6nMoIo1o7M3L4L4K6K7G9G9Ga]n1"}, "label": "the back of a black wheelchair with a boy wearing a grey and red shirt sitting in it"}]}
{"id": 9557, "image": "COCO_train2014_000000009557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wheelchair\"."}], "task": "refering", "answer_template": "The \"the wheelchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 154, 155], "bbox": [0.41618, 0.2628, 0.65858, 0.6922666666666667], "iscrowd": 0, "area": 11926.8441, "rle": {"size": [375, 500], "counts": "d[\\29\\;;G9F:F:G9]OYNcFi1W9?I7N3M2N2N2N2O0O1O1N101O001N2O1L4J6K5J5C:MO0mNlHVMo7k2?11OO1O100O1O1O010O001O1O010O001O001O0010O01O001O001O001O00001O2N3M3M1O1O010O001O001ON1O2VOQHiMR8Y2PHaMR8`2QHZMP8h2RHRMQ8P3:1O1N2O2N2N4K4M3M3L5K3M3L4M3M3M3N2NVMRIm0l6jN^IV1b6YNoIg1R6RNTJn1n5oMSJQ2m5nMSJR2n5nMRJR2n5oMRJP2n5oMSJP2o5mMTJP2o5mMTJP2o5nMRJP2Q6nMQJo1R6nMQJo1R6oMoIP2S6nMoIo1o7M3L4L4K6K7G9G9Ga]n1"}, "label": "the wheelchair"}]}
{"id": 140630, "image": "COCO_train2014_000000140630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver suv on the road\"."}], "task": "refering", "answer_template": "The \"a silver suv on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.0, 0.7847500000000001, 0.6367708333333333, 0.988796875], "iscrowd": 0, "area": 32729.002850000008, "rle": {"size": [640, 480], "counts": "Ta0a2_a0000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100O1O1O101N1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1000000000000000000000000000000O100000001O0000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O2O0000000000000000000000000000000000N2N2O1N2O1N2N200000000O01000000000O1000002N3L4M3M3M3M3M3L5L3M3M3M3M3L4M3M4L3M3L4M3M3M3M3M3L5L1O00000O1000000M3K5J6K5Joi\\3"}, "label": "a silver suv on the road"}]}
{"id": 140630, "image": "COCO_train2014_000000140630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey vehicle is near a stop sign in the street\"."}], "task": "refering", "answer_template": "The \"a grey vehicle is near a stop sign in the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.0, 0.7847500000000001, 0.6367708333333333, 0.988796875], "iscrowd": 0, "area": 32729.002850000008, "rle": {"size": [640, 480], "counts": "Ta0a2_a0000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100O1O1O101N1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1000000000000000000000000000000O100000001O0000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O2O0000000000000000000000000000000000N2N2O1N2O1N2N200000000O01000000000O1000002N3L4M3M3M3M3M3L5L3M3M3M3M3L4M3M4L3M3L4M3M3M3M3M3L5L1O00000O1000000M3K5J6K5Joi\\3"}, "label": "a grey vehicle is near a stop sign in the street"}]}
{"id": 148824, "image": "COCO_train2014_000000148824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy carring a suitcase\"."}], "task": "refering", "answer_template": "The \"the boy carring a suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 41, 42, 43, 53, 54, 55, 64, 65, 66, 67, 76, 77, 78, 79, 80, 88, 89, 90, 91, 92, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 124, 125, 126, 127, 136, 137, 138, 139, 148, 149, 150, 151, 152, 160, 161, 162, 163, 164, 172, 173, 174, 175, 184, 185, 186, 195, 196, 197], "bbox": [0.3052710843373494, 0.15990000000000001, 0.6953614457831325, 0.9234199999999999], "iscrowd": 0, "area": 33018.544, "rle": {"size": [500, 332], "counts": "UXb11b?1O010O1O1O1O1N2O1N2THj0XOXO`0h1fNYNX1c2mM_Mo1U3_MmL^2_3WMcLe2i3QMYLl2R4hLRLT3[4_LiK^3b4VLbKg3j4lKZKQ4Q5cKSK[4W5[KjJc4^5WKcJg4a5VK`Jg4e5VK\\Jg4h5WKYJg4k5UKWJh4n5UKSJj4P6SKQJm4P6QKQJn4R6oJoIP5T6mJlIT5U6jJlIU5W6hJjIW5Y6fJgI[5[6bJfI]5\\6aJeI_5]6^JcIb5X8kNUF`Lk9^3[F_Le9_3`F]La9a3dF[L\\9d3iFVLZ9i3jFQLX9P4kFjKX9V4S1O010O010O010O10O010O010O010O010O10000O100O100O100O100O100O2O000O2O0O2O0O2O0O101M2M4K4M4K4L5L3M4M2OS1lN:G8G4K4K6J5L5L4L3M4L3L5N2O1N2N2aMXGnMj8m1_GmMc8n1eGmM]8n1lGlMW8m1SHmMP8l1ZHnMi7k1`HPNb7i1iHQNZ7h1oHSNT7e1WIUNl6d1]IWNf6b1dIXN_6`1P4Ie0ZOg0ZOPSa1"}, "label": "the boy carring a suitcase"}]}
{"id": 148824, "image": "COCO_train2014_000000148824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male wearing suitcase on head and sign around neck\"."}], "task": "refering", "answer_template": "The \"male wearing suitcase on head and sign around neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 41, 42, 43, 53, 54, 55, 64, 65, 66, 67, 76, 77, 78, 79, 80, 88, 89, 90, 91, 92, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 124, 125, 126, 127, 136, 137, 138, 139, 148, 149, 150, 151, 152, 160, 161, 162, 163, 164, 172, 173, 174, 175, 184, 185, 186, 195, 196, 197], "bbox": [0.3052710843373494, 0.15990000000000001, 0.6953614457831325, 0.9234199999999999], "iscrowd": 0, "area": 33018.544, "rle": {"size": [500, 332], "counts": "UXb11b?1O010O1O1O1O1N2O1N2THj0XOXO`0h1fNYNX1c2mM_Mo1U3_MmL^2_3WMcLe2i3QMYLl2R4hLRLT3[4_LiK^3b4VLbKg3j4lKZKQ4Q5cKSK[4W5[KjJc4^5WKcJg4a5VK`Jg4e5VK\\Jg4h5WKYJg4k5UKWJh4n5UKSJj4P6SKQJm4P6QKQJn4R6oJoIP5T6mJlIT5U6jJlIU5W6hJjIW5Y6fJgI[5[6bJfI]5\\6aJeI_5]6^JcIb5X8kNUF`Lk9^3[F_Le9_3`F]La9a3dF[L\\9d3iFVLZ9i3jFQLX9P4kFjKX9V4S1O010O010O010O10O010O010O010O010O10000O100O100O100O100O100O2O000O2O0O2O0O2O0O101M2M4K4M4K4L5L3M4M2OS1lN:G8G4K4K6J5L5L4L3M4L3L5N2O1N2N2aMXGnMj8m1_GmMc8n1eGmM]8n1lGlMW8m1SHmMP8l1ZHnMi7k1`HPNb7i1iHQNZ7h1oHSNT7e1WIUNl6d1]IWNf6b1dIXN_6`1P4Ie0ZOg0ZOPSa1"}, "label": "male wearing suitcase on head and sign around neck"}]}
{"id": 410969, "image": "COCO_train2014_000000410969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck covered in the snow furthest to the right\"."}], "task": "refering", "answer_template": "The \"the truck covered in the snow furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338], "bbox": [0.47757812499999996, 0.4438333333333333, 0.998875, 0.8563958333333334], "iscrowd": 0, "area": 47582.08875, "rle": {"size": [480, 640], "counts": "ci_47f>8I7H5K2O2M3M3N2M3M3N2M3M3N2M3N2M3M3N2M2N3N3L7I6K6I6J6L2O0O1O2N1O100O1O2N100O1O2N1O100O1O2N1000O00100O1O00100O1O010O1000O010000O10O10O1000O010000O0100000O01000O10O10O1000O010000O0101O3M3M4K4M00000O2O00000O2O00000O2O00000O10O1000O100000O01000000N2O001O1O1O1O1N101O1O1O1O1O001N2O1O1O1O001O1N2O100O010000000O10O100000O1000O1O0N3L4M3M30O100000000O100000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000O1N2O1O1N2O1O1O1O10000000000000000000O10000000000000000000000N2N1N3000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000b1^Ni6"}, "label": "the truck covered in the snow furthest to the right"}]}
{"id": 410969, "image": "COCO_train2014_000000410969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old truck covered in snow except for the grill and door\"."}], "task": "refering", "answer_template": "The \"an old truck covered in snow except for the grill and door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338], "bbox": [0.47757812499999996, 0.4438333333333333, 0.998875, 0.8563958333333334], "iscrowd": 0, "area": 47582.08875, "rle": {"size": [480, 640], "counts": "ci_47f>8I7H5K2O2M3M3N2M3M3N2M3M3N2M3N2M3M3N2M2N3N3L7I6K6I6J6L2O0O1O2N1O100O1O2N100O1O2N1O100O1O2N1000O00100O1O00100O1O010O1000O010000O10O10O1000O010000O0100000O01000O10O10O1000O010000O0101O3M3M4K4M00000O2O00000O2O00000O2O00000O10O1000O100000O01000000N2O001O1O1O1O1N101O1O1O1O1O001N2O1O1O1O001O1N2O100O010000000O10O100000O1000O1O0N3L4M3M30O100000000O100000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000O1N2O1O1N2O1O1O1O10000000000000000000O10000000000000000000000N2N1N3000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000b1^Ni6"}, "label": "an old truck covered in snow except for the grill and door"}]}
{"id": 410969, "image": "COCO_train2014_000000410969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old truck behind a tree\"."}], "task": "refering", "answer_template": "The \"an old truck behind a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 242, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308], "bbox": [0.17696875, 0.44906250000000003, 0.701125, 0.7906458333333333], "iscrowd": 0, "area": 24717.481299999992, "rle": {"size": [480, 640], "counts": "^Ye13m>:F1O010O001O001O001O001M2I8G8kNnNUDZ1j;XO]Co0b<i0N101N1O2O0O2N101N2N3N2M3N2M3M3N3L3M100O2O000O10000O10000O1000000O10000O101OO010O10O0100O10O0100O010O10O0100O010O100O010O10O01000O10000O101O000O100O101N10000O101N100O100O2O0O100O10001N100O100O2O0O100O10000O10O0100O100O10O0100O10000O100O2O0O100O100O2O000O100O101N100O100O2O0O1000000O2O00000O01N2N2N2N1O2N2N2cN]1XOg0E<D<F:I6J7I7H8Il]`0?Xa_Oi0K6J5K6J5K5K6JO1O1O1O2N1O1O1N2O1O1O1O1O1O1O1N3O00000O4M5K6J5J7J00000O010000000O1000000O1000000O1000000O100000000O1000000O1000000O100000O1N1O2DTOcBm0]=WO]Bk0c=90O01000000O0100000O10O100000O10O1000O1000O1000001O0O1000000O1O2N1O1O1O1O1O1O1O2N1O1O1O1O`0@bYi2"}, "label": "an old truck behind a tree"}]}
{"id": 410969, "image": "COCO_train2014_000000410969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck behind the tree\"."}], "task": "refering", "answer_template": "The \"truck behind the tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 242, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308], "bbox": [0.17696875, 0.44906250000000003, 0.701125, 0.7906458333333333], "iscrowd": 0, "area": 24717.481299999992, "rle": {"size": [480, 640], "counts": "^Ye13m>:F1O010O001O001O001O001M2I8G8kNnNUDZ1j;XO]Co0b<i0N101N1O2O0O2N101N2N3N2M3N2M3M3N3L3M100O2O000O10000O10000O1000000O10000O101OO010O10O0100O10O0100O010O10O0100O010O100O010O10O01000O10000O101O000O100O101N10000O101N100O100O2O0O100O10001N100O100O2O0O100O10000O10O0100O100O10O0100O10000O100O2O0O100O100O2O000O100O101N100O100O2O0O1000000O2O00000O01N2N2N2N1O2N2N2cN]1XOg0E<D<F:I6J7I7H8Il]`0?Xa_Oi0K6J5K6J5K5K6JO1O1O1O2N1O1O1N2O1O1O1O1O1O1O1N3O00000O4M5K6J5J7J00000O010000000O1000000O1000000O1000000O100000000O1000000O1000000O100000O1N1O2DTOcBm0]=WO]Bk0c=90O01000000O0100000O10O100000O10O1000O1000O1000001O0O1000000O1O2N1O1O1O1O1O1O1O2N1O1O1O1O`0@bYi2"}, "label": "truck behind the tree"}]}
{"id": 345436, "image": "COCO_train2014_000000345436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man and women hugging under an umbrella\"."}], "task": "refering", "answer_template": "The \"a man and women hugging under an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 139, 155, 156, 157, 174, 175, 176, 192, 193, 210, 211], "bbox": [0.64896, 0.5735608308605341, 0.82594, 0.987507418397626], "iscrowd": 0, "area": 4570.974799999999, "rle": {"size": [337, 500], "counts": "dU[33]:2N2N2M3N2N7I9G2M3N2N2ZGUOc7m0[HWO`7l0^HWO_7k0^HYO_7i0^HZO`7h0]H\\O_7f0_H^O^7d0_H@^7k1N3L5L4L4L5L2O1O1O1O1O1O1O1O1O10O0O1XN^I_Oc6?_I@b6?_I@b6`0^I_Od6B^IBOk0f6WOgIKDm0i6QOhI0@n0k6mNhI4^On0m6hNiI:ZOm0S8QOPHm0Q8ROPHm0Q8ROPHm0P8SOTHi0l7WOZHc0f7]OgH6Y7JUIHk68g1O100000000000000O100000001O0O11O0O101O000O2O001O0O101O000014K3N1N4Lb[l0"}, "label": "a man and women hugging under an umbrella"}]}
{"id": 345436, "image": "COCO_train2014_000000345436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man i white shirt and white pants hugging a woman\"."}], "task": "refering", "answer_template": "The \"a man i white shirt and white pants hugging a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 139, 155, 156, 157, 174, 175, 176, 192, 193, 210, 211], "bbox": [0.64896, 0.5735608308605341, 0.82594, 0.987507418397626], "iscrowd": 0, "area": 4570.974799999999, "rle": {"size": [337, 500], "counts": "dU[33]:2N2N2M3N2N7I9G2M3N2N2ZGUOc7m0[HWO`7l0^HWO_7k0^HYO_7i0^HZO`7h0]H\\O_7f0_H^O^7d0_H@^7k1N3L5L4L4L5L2O1O1O1O1O1O1O1O1O10O0O1XN^I_Oc6?_I@b6?_I@b6`0^I_Od6B^IBOk0f6WOgIKDm0i6QOhI0@n0k6mNhI4^On0m6hNiI:ZOm0S8QOPHm0Q8ROPHm0Q8ROPHm0P8SOTHi0l7WOZHc0f7]OgH6Y7JUIHk68g1O100000000000000O100000001O0O11O0O101O000O2O001O0O101O000014K3N1N4Lb[l0"}, "label": "a man i white shirt and white pants hugging a woman"}]}
{"id": 427362, "image": "COCO_train2014_000000427362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl is holding a vote sign\"."}], "task": "refering", "answer_template": "The \"a girl is holding a vote sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290, 312, 313, 314, 335, 336, 337, 359], "bbox": [0.55140625, 0.44945833333333335, 0.6703906249999999, 0.8927499999999999], "iscrowd": 0, "area": 9443.128200000001, "rle": {"size": [480, 640], "counts": "\\gU52i>5L4K6K4L4L5N1N2O1O2N1O0gFmNf4c0oFFZ4Jd49YGNR4La45_G0g0Jk13m53aG1c04g1IT62cG1>>d1@Z60eG39g0a1WO]73nFP1^1nNc73kFZ1Z1dNk72fFe1LcN9HT91cFZ3HfLd91`FP4_9QL]FS4c9e0O1O1O100O1O1O1J6K3OO00001D?_Oa0_O`0G:8G9H7I8G9UGeJa7c5UH_Jj7j5lGXJT8o5bGUJ]8W6O2O00001O001O0001O000oLeG\\O\\8c0nGTOR8l0VHkNk72dGaNk0T1e7bNbGj03[Oo0P1a7cNcGn04ZOP1m0]7cNeGR14YOS1i0m7KUGXOU1c0k71VGWOX17m7>QGWOg;f0\\1L6J7J5J`dR3"}, "label": "a girl is holding a vote sign"}]}
{"id": 427362, "image": "COCO_train2014_000000427362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady hold a poster\"."}], "task": "refering", "answer_template": "The \"lady hold a poster\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290, 312, 313, 314, 335, 336, 337, 359], "bbox": [0.55140625, 0.44945833333333335, 0.6703906249999999, 0.8927499999999999], "iscrowd": 0, "area": 9443.128200000001, "rle": {"size": [480, 640], "counts": "\\gU52i>5L4K6K4L4L5N1N2O1O2N1O0gFmNf4c0oFFZ4Jd49YGNR4La45_G0g0Jk13m53aG1c04g1IT62cG1>>d1@Z60eG39g0a1WO]73nFP1^1nNc73kFZ1Z1dNk72fFe1LcN9HT91cFZ3HfLd91`FP4_9QL]FS4c9e0O1O1O100O1O1O1J6K3OO00001D?_Oa0_O`0G:8G9H7I8G9UGeJa7c5UH_Jj7j5lGXJT8o5bGUJ]8W6O2O00001O001O0001O000oLeG\\O\\8c0nGTOR8l0VHkNk72dGaNk0T1e7bNbGj03[Oo0P1a7cNcGn04ZOP1m0]7cNeGR14YOS1i0m7KUGXOU1c0k71VGWOX17m7>QGWOg;f0\\1L6J7J5J`dR3"}, "label": "lady hold a poster"}]}
{"id": 476520, "image": "COCO_train2014_000000476520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing to the right in a grassy plain\"."}], "task": "refering", "answer_template": "The \"a zebra standing to the right in a grassy plain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 84, 105, 106, 107, 108, 129, 130], "bbox": [0.5774843749999999, 0.25972972972972974, 0.705703125, 0.6652895752895753], "iscrowd": 0, "area": 4213.8170500000015, "rle": {"size": [259, 640], "counts": "odm22Q83N1N0XH1\\70bHb0n6]OQIe0o68O10OO1O0_ISOl5o0oIYOm5b1L3M3N2M5L2M4M2M23`MgJo1g54L1O100O03TO^JQOc5m0aJnN`5R1cJkN\\5W1hJcNX5^1[JcNP6k1N3L5K01O00OTOWJ^Ok5b0[JkNR6U1=1000O010000O1000O10O10000O4M4L2N2N1O1nI\\Nd5Q20ROWJAk5;YJDg5;\\JCe5=[JBe5?[JRONNh5R1ZJlNV6T1jIjNW6W1;3M9G8IN2N1TOPJCR6<QJ^OR6c0QJTOc6=d0N1O218Go\\_1"}, "label": "a zebra standing to the right in a grassy plain"}]}
{"id": 476520, "image": "COCO_train2014_000000476520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 84, 105, 106, 107, 108, 129, 130], "bbox": [0.5774843749999999, 0.25972972972972974, 0.705703125, 0.6652895752895753], "iscrowd": 0, "area": 4213.8170500000015, "rle": {"size": [259, 640], "counts": "odm22Q83N1N0XH1\\70bHb0n6]OQIe0o68O10OO1O0_ISOl5o0oIYOm5b1L3M3N2M5L2M4M2M23`MgJo1g54L1O100O03TO^JQOc5m0aJnN`5R1cJkN\\5W1hJcNX5^1[JcNP6k1N3L5K01O00OTOWJ^Ok5b0[JkNR6U1=1000O010000O1000O10O10000O4M4L2N2N1O1nI\\Nd5Q20ROWJAk5;YJDg5;\\JCe5=[JBe5?[JRONNh5R1ZJlNV6T1jIjNW6W1;3M9G8IN2N1TOPJCR6<QJ^OR6c0QJTOc6=d0N1O218Go\\_1"}, "label": "the zebra on the right"}]}
{"id": 476520, "image": "COCO_train2014_000000476520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra which is standing on the left side\"."}], "task": "refering", "answer_template": "The \"a zebra which is standing on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 77, 98, 99, 100, 101, 122, 123, 145], "bbox": [0.28684375, 0.2296911196911197, 0.401953125, 0.7240540540540541], "iscrowd": 0, "area": 4989.333650000001, "rle": {"size": [259, 640], "counts": "Wc^1<l1Fi3b0PLAP4b0jKAV4b0bKD]4R1hJSOX5l102N4L5L3M3M1O00001O0000:E;D000OgNTL^Nk3ChKd1?cNh3JnKa1>^Nf3OUL\\1;]Na35VL^1l4`NVK_1j4`NXK`1f5O03M5K4L9G3M0O100000000O10VOdJmN\\5U1kJbNU5_1g01O2M201NM3M3N2O2O0a0_O3M10O001O2L3L4J6H8I7K5L4L4M3Na0@3M3M3MkgP3"}, "label": "a zebra which is standing on the left side"}]}
{"id": 476520, "image": "COCO_train2014_000000476520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra that is too the left of the rightmost zebra\"."}], "task": "refering", "answer_template": "The \"zebra that is too the left of the rightmost zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 77, 98, 99, 100, 101, 122, 123, 145], "bbox": [0.28684375, 0.2296911196911197, 0.401953125, 0.7240540540540541], "iscrowd": 0, "area": 4989.333650000001, "rle": {"size": [259, 640], "counts": "Wc^1<l1Fi3b0PLAP4b0jKAV4b0bKD]4R1hJSOX5l102N4L5L3M3M1O00001O0000:E;D000OgNTL^Nk3ChKd1?cNh3JnKa1>^Nf3OUL\\1;]Na35VL^1l4`NVK_1j4`NXK`1f5O03M5K4L9G3M0O100000000O10VOdJmN\\5U1kJbNU5_1g01O2M201NM3M3N2O2O0a0_O3M10O001O2L3L4J6H8I7K5L4L4M3Na0@3M3M3MkgP3"}, "label": "zebra that is too the left of the rightmost zebra"}]}
{"id": 460139, "image": "COCO_train2014_000000460139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old dirty scooter parked next to a metal fence\"."}], "task": "refering", "answer_template": "The \"an old dirty scooter parked next to a metal fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.517875, 0.5070779220779221, 0.891125, 0.8044805194805195], "iscrowd": 0, "area": 16942.001299999996, "rle": {"size": [462, 640], "counts": "^ee41[>2N3M3L4[NKmD7h:4VEO_:;_EH`:8]EKb:6\\EMc:3[E0c:1[E1e:OYE4e:MZE5e:KYE7f:JYE8g:GXE;l:@REb0T;YOjDi0Z;ROeDP1_;kN_DW1e;eNZD_1c;aN[Dc1c;]N\\Dd1d;\\N\\Dd1d;\\N\\De1c;[N]De1d;ZN\\Dg1c;YN]Dg1c;YN]Dg1b;ZN^Dg1a;YN_Dg1a;YN_Dh1_;ZN`Df1`;ZN`Dg1_;YNaDg1_;YNaDg1_;YNaDh1^;XNbDh1^;XNbDi1];WNcDi1];WNcDi1];WNcDj1\\;VNdDj1\\;VNdDk1[;UNeDk1[;UNeDk1[;UNeDl1Z;UNeDk1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Y;WNgDi1Y;WNgDi1X;XNhDh1X;XNhDh1X;XNhDh1W;YNiDg1W;YNiDg1W;YNiDg1V;ZNjDf1V;ZNjDf1V;ZNjDf1S;]NmDc1o:aNQE_1k:eNUE[1g:iNYEW1f:jNZEV1f:jNZEV1f:jNZEV1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1i:gNWEY1i:gNWEY1i:gNWEY1i:gNWEBIV1P;XOWE@KW1o:YOVE^OMY1m:YOVE\\OO[1k:YOVEZO1]1i:YOVEXO3_1g:YOVEWO4`1g:XOUEVO6b1e:XOfEh0Z:XOgEg0Y:YOgEg0Z:XOfEh0Z:XOfEh0Z:XOfEh0Z:XOfEh0[:WOeEi0[:WOeEi0[:WOeEi0[:WOeEh0\\:VOgEi0Z:SOiEm0W:POlEP1T:mNoES1Q:jNRFV1o9fNTFZ1`;000001O00000000000O1000001O00\\NlNnET1Q:lNPFT1n9nNRFR1l9oNUFQ1j9POVFP1h9QOYFo0f9ROZFn0d9TO\\Fl0b9UO`Fj0]9YObFh0\\9YOeFf0Z9\\OeFe0X9]OhFd0V9^OjFa0U9@kF`0T9BkF?R9CoF<P9FoF:R9FlF;V9EfF=[9EaF=_91PF1Q:4gENZ:2dEN^:2bEM_:3aELa:3^ELd:4\\EKe:5[EJf:6ZEHh:8XEGi::VEEl::TEDn:<RECo:=QEBP;>PE@R;`0mD@T;`0lD_O?jNR:f1_EAM[Oc:0\\E`04j0`:dNaE?On0_:aNfE?KP1_:_NkE>FT1^:\\NPF>BV1_:YNTF>]OZ1^:VNYF>YO\\1_:SN\\F?UO_1c;_N]Dn0MSOh;M[Dd0=\\OX;O\\D8l0Ej:1ZDM]1OZ:2Z2N2N2N1Of]10gaNc0]Od0\\Oc0N210O01O00001O001O0O101O001O000O2O001N101O0O101O0O2hN\\Cd0Z=L4M2N3L4K4KeWo0"}, "label": "an old dirty scooter parked next to a metal fence"}]}
{"id": 460139, "image": "COCO_train2014_000000460139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old moped leaning on a railing\"."}], "task": "refering", "answer_template": "The \"an old moped leaning on a railing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.517875, 0.5070779220779221, 0.891125, 0.8044805194805195], "iscrowd": 0, "area": 16942.001299999996, "rle": {"size": [462, 640], "counts": "^ee41[>2N3M3L4[NKmD7h:4VEO_:;_EH`:8]EKb:6\\EMc:3[E0c:1[E1e:OYE4e:MZE5e:KYE7f:JYE8g:GXE;l:@REb0T;YOjDi0Z;ROeDP1_;kN_DW1e;eNZD_1c;aN[Dc1c;]N\\Dd1d;\\N\\Dd1d;\\N\\De1c;[N]De1d;ZN\\Dg1c;YN]Dg1c;YN]Dg1b;ZN^Dg1a;YN_Dg1a;YN_Dh1_;ZN`Df1`;ZN`Dg1_;YNaDg1_;YNaDg1_;YNaDh1^;XNbDh1^;XNbDi1];WNcDi1];WNcDi1];WNcDj1\\;VNdDj1\\;VNdDk1[;UNeDk1[;UNeDk1[;UNeDl1Z;UNeDk1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Z;VNfDj1Y;WNgDi1Y;WNgDi1X;XNhDh1X;XNhDh1X;XNhDh1W;YNiDg1W;YNiDg1W;YNiDg1V;ZNjDf1V;ZNjDf1V;ZNjDf1S;]NmDc1o:aNQE_1k:eNUE[1g:iNYEW1f:jNZEV1f:jNZEV1f:jNZEV1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1g:iNYEW1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1h:hNXEX1i:gNWEY1i:gNWEY1i:gNWEY1i:gNWEBIV1P;XOWE@KW1o:YOVE^OMY1m:YOVE\\OO[1k:YOVEZO1]1i:YOVEXO3_1g:YOVEWO4`1g:XOUEVO6b1e:XOfEh0Z:XOgEg0Y:YOgEg0Z:XOfEh0Z:XOfEh0Z:XOfEh0Z:XOfEh0[:WOeEi0[:WOeEi0[:WOeEi0[:WOeEh0\\:VOgEi0Z:SOiEm0W:POlEP1T:mNoES1Q:jNRFV1o9fNTFZ1`;000001O00000000000O1000001O00\\NlNnET1Q:lNPFT1n9nNRFR1l9oNUFQ1j9POVFP1h9QOYFo0f9ROZFn0d9TO\\Fl0b9UO`Fj0]9YObFh0\\9YOeFf0Z9\\OeFe0X9]OhFd0V9^OjFa0U9@kF`0T9BkF?R9CoF<P9FoF:R9FlF;V9EfF=[9EaF=_91PF1Q:4gENZ:2dEN^:2bEM_:3aELa:3^ELd:4\\EKe:5[EJf:6ZEHh:8XEGi::VEEl::TEDn:<RECo:=QEBP;>PE@R;`0mD@T;`0lD_O?jNR:f1_EAM[Oc:0\\E`04j0`:dNaE?On0_:aNfE?KP1_:_NkE>FT1^:\\NPF>BV1_:YNTF>]OZ1^:VNYF>YO\\1_:SN\\F?UO_1c;_N]Dn0MSOh;M[Dd0=\\OX;O\\D8l0Ej:1ZDM]1OZ:2Z2N2N2N1Of]10gaNc0]Od0\\Oc0N210O01O00001O001O0O101O001O000O2O001N101O0O101O0O2hN\\Cd0Z=L4M2N3L4K4KeWo0"}, "label": "an old moped leaning on a railing"}]}
{"id": 460139, "image": "COCO_train2014_000000460139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motor bike with registration number rj14sa774\"."}], "task": "refering", "answer_template": "The \"a motor bike with registration number rj14sa774\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 210, 211, 212, 213, 234, 235, 236, 237, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 348, 349, 350, 351], "bbox": [0.134640625, 0.5438095238095239, 0.337421875, 0.9865151515151515], "iscrowd": 0, "area": 17821.76325, "rle": {"size": [462, 640], "counts": "hTW12W>6J6K6I6N3[CTOd;m0UDAV9B^HP1TNOR9ROiHQ1nMT1W8lMjIR1hM[1Z8eMlIf3o5[LPJg3m5[LSJe3l5\\LSJe3m5\\LQJe3o5[LPJf3P6ZLPJe3Q6\\LnId3R6\\LmIe3S6[LmIe3R6]LmIc3S6]LmIc3S6]LmIc3S6^LkIc3U6]LkIc3U6]LkIc3U6^LiIc3W6]LiIc3W6]LiIc3W6^LhIb3X6^LgIc3Y6]LgIc3Y6]LgIb3Z6^LeIc3[6^LdIb3\\6^LdIb3\\6^LcIc3]6]LcIc3^6\\LbId3^6]L`Id3`6\\L`Id3a6[L_Ie3a6[L^If3b6ZL^If3c6ZL\\If3d6ZLZIh3f6XLXIj3i6VLSIm3m6SLQIo3o6RLnHo3T7QLhHR4X7oKeHS4[7nKbHT4_7lK[HY4e7hKSH_4m7bKnGb4S8]KiGg4W8h0001O00001O00001O0O101OO1O1O1O1O1O1O100O100001O1I7M2O2M3N1N3N2M3M2O2M3N2M2O2N2N101N2N2O0O2M3N2jLUFf1l9YNXFc1j9[NYFa1j9]NXFa1j9]NXFa1k9\\NVFc1o9WNTFf1S:TNnEk1X:oMjEo1[:lMgEQ2`:iMaEV2];M3N1N4J5K6J5K6J5K6K4K6J5K5K6Jgnn5"}, "label": "a motor bike with registration number rj14sa774"}]}
{"id": 460139, "image": "COCO_train2014_000000460139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle with rj14 on the plate\"."}], "task": "refering", "answer_template": "The \"a black motorcycle with rj14 on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 210, 211, 212, 213, 234, 235, 236, 237, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 348, 349, 350, 351], "bbox": [0.134640625, 0.5438095238095239, 0.337421875, 0.9865151515151515], "iscrowd": 0, "area": 17821.76325, "rle": {"size": [462, 640], "counts": "hTW12W>6J6K6I6N3[CTOd;m0UDAV9B^HP1TNOR9ROiHQ1nMT1W8lMjIR1hM[1Z8eMlIf3o5[LPJg3m5[LSJe3l5\\LSJe3m5\\LQJe3o5[LPJf3P6ZLPJe3Q6\\LnId3R6\\LmIe3S6[LmIe3R6]LmIc3S6]LmIc3S6]LmIc3S6^LkIc3U6]LkIc3U6]LkIc3U6^LiIc3W6]LiIc3W6]LiIc3W6^LhIb3X6^LgIc3Y6]LgIc3Y6]LgIb3Z6^LeIc3[6^LdIb3\\6^LdIb3\\6^LcIc3]6]LcIc3^6\\LbId3^6]L`Id3`6\\L`Id3a6[L_Ie3a6[L^If3b6ZL^If3c6ZL\\If3d6ZLZIh3f6XLXIj3i6VLSIm3m6SLQIo3o6RLnHo3T7QLhHR4X7oKeHS4[7nKbHT4_7lK[HY4e7hKSH_4m7bKnGb4S8]KiGg4W8h0001O00001O00001O0O101OO1O1O1O1O1O1O100O100001O1I7M2O2M3N1N3N2M3M2O2M3N2M2O2N2N101N2N2O0O2M3N2jLUFf1l9YNXFc1j9[NYFa1j9]NXFa1j9]NXFa1k9\\NVFc1o9WNTFf1S:TNnEk1X:oMjEo1[:lMgEQ2`:iMaEV2];M3N1N4J5K6J5K6J5K6K4K6J5K5K6Jgnn5"}, "label": "a black motorcycle with rj14 on the plate"}]}
{"id": 451953, "image": "COCO_train2014_000000451953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange in the middle\"."}], "task": "refering", "answer_template": "The \"the orange in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278], "bbox": [0.0, 0.5586885245901639, 0.17540625, 0.8544496487119438], "iscrowd": 0, "area": 10338.4407, "rle": {"size": [427, 640], "counts": "a77S=a0@?A?A`0@?A`0@?A8G100000000000001O0000000000000000000000O100O1O1O100O1O100O1O1O100O1O1O100O1O100O100O2O0O2O0O101N101O0O101N101N100O2O0O101N101N100O2O0O101N101N1O101N1O2N2N2O0O2N2N1O2O1N2N2N2N2N2N2N2N2N3M2N3M2N3M3M2N3M2N4L7Fblk6"}, "label": "the orange in the middle"}]}
{"id": 42355, "image": "COCO_train2014_000000042355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor that is turned off\"."}], "task": "refering", "answer_template": "The \"a computer monitor that is turned off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 44, 45, 46, 47, 60, 61, 62, 63, 76, 77, 78, 79], "bbox": [0.7511111111111111, 0.10278106508875741, 0.9906666666666667, 0.3805621301775148], "iscrowd": 0, "area": 8187.30535, "rle": {"size": [338, 450], "counts": "de_34W:7H9H7H8I7H8I8G8I7I70001O000000001O000O10001O00000000001O0000000O2O000000000000001O000000000O100000001O0000000000000001O01O0001O0000001O000001O01O0000001O00000010O00000001O0000001O0001OK6I6K5J6K5K6I6K5J6K6J5J6K5K6IoX1"}, "label": "a computer monitor that is turned off"}]}
{"id": 42355, "image": "COCO_train2014_000000042355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a right side monitor\"."}], "task": "refering", "answer_template": "The \"a right side monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 44, 45, 46, 47, 60, 61, 62, 63, 76, 77, 78, 79], "bbox": [0.7511111111111111, 0.10278106508875741, 0.9906666666666667, 0.3805621301775148], "iscrowd": 0, "area": 8187.30535, "rle": {"size": [338, 450], "counts": "de_34W:7H9H7H8I7H8I8G8I7I70001O000000001O000O10001O00000000001O0000000O2O000000000000001O000000000O100000001O0000000000000001O01O0001O0000001O000001O01O0000001O00000010O00000001O0000001O0001OK6I6K5J6K5K6I6K5J6K6J5J6K5K6IoX1"}, "label": "a right side monitor"}]}
{"id": 42355, "image": "COCO_train2014_000000042355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer monitor in the corner\"."}], "task": "refering", "answer_template": "The \"the computer monitor in the corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 19, 20, 21, 22, 23, 35, 36, 37, 38, 39, 52, 53, 54], "bbox": [0.20206666666666667, 0.04636094674556213, 0.44282222222222223, 0.3285207100591716], "iscrowd": 0, "area": 8396.760600000001, "rle": {"size": [338, 450], "counts": "YRn03_:a0^O>C1O1O2N1O1O1O1O1O2N0000001I:D;M2N1O1O1O1O1O1O1O0O2O1O1O1O9G001O0000001O000000O10000000O0100000000O100000000O100000000O100000O10O100000000O100000000O100000000O1000O1000O100000000O1000000O100000000O100000000O10g0YOoib2"}, "label": "the computer monitor in the corner"}]}
{"id": 42355, "image": "COCO_train2014_000000042355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer monitor on the desk\"."}], "task": "refering", "answer_template": "The \"a white computer monitor on the desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 19, 20, 21, 22, 23, 35, 36, 37, 38, 39, 52, 53, 54], "bbox": [0.20206666666666667, 0.04636094674556213, 0.44282222222222223, 0.3285207100591716], "iscrowd": 0, "area": 8396.760600000001, "rle": {"size": [338, 450], "counts": "YRn03_:a0^O>C1O1O2N1O1O1O1O1O2N0000001I:D;M2N1O1O1O1O1O1O1O0O2O1O1O1O9G001O0000001O000000O10000000O0100000000O100000000O100000000O100000O10O100000000O100000000O100000000O1000O1000O100000000O1000000O100000000O100000000O10g0YOoib2"}, "label": "a white computer monitor on the desk"}]}
{"id": 206196, "image": "COCO_train2014_000000206196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sky chefs truck behind all the others\"."}], "task": "refering", "answer_template": "The \"a sky chefs truck behind all the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.676875, 0.6417291666666666, 0.994890625, 0.8752916666666667], "iscrowd": 0, "area": 16119.607199999999, "rle": {"size": [480, 640], "counts": "VZ[61k>5K5L4M3M3M3M3N2M2N3N1O2O001N101O1N3N1O1O1O2N1O1O1O1O2O000O101N100O1N2N1O1O101N1O1O2N1N2N2N3M2N2N3M2N2N20000001O00000000010O01O001O1O010O1O001O1O001O00001O00000000001O01O000000010O0000000010O00000000010O000000010ON2O1N2N2O2M2N2O1N2N30O1O1O1O100O1O1O1O1O100O001O1O100O1O1O1O1O100O1O1O00O2O000O100O2O00000000001O0O1000001O000000001O00000000001O000000001O00000000001N100O100O2O0O100O1O2N1O1O1O1O1N3N1N2O3L4L4lN_S1"}, "label": "a sky chefs truck behind all the others"}]}
{"id": 206196, "image": "COCO_train2014_000000206196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sky chefs van on the right\"."}], "task": "refering", "answer_template": "The \"the sky chefs van on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.676875, 0.6417291666666666, 0.994890625, 0.8752916666666667], "iscrowd": 0, "area": 16119.607199999999, "rle": {"size": [480, 640], "counts": "VZ[61k>5K5L4M3M3M3M3N2M2N3N1O2O001N101O1N3N1O1O1O2N1O1O1O1O2O000O101N100O1N2N1O1O101N1O1O2N1N2N2N3M2N2N3M2N2N20000001O00000000010O01O001O1O010O1O001O1O001O00001O00000000001O01O000000010O0000000010O00000000010O000000010ON2O1N2N2O2M2N2O1N2N30O1O1O1O100O1O1O1O1O100O001O1O100O1O1O1O1O100O1O1O00O2O000O100O2O00000000001O0O1000001O000000001O00000000001O000000001O00000000001N100O100O2O0O100O1O2N1O1O1O1O1N3N1N2O3L4L4lN_S1"}, "label": "the sky chefs van on the right"}]}
{"id": 517494, "image": "COCO_train2014_000000517494.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male zebra , bigger body , looking off the screen standing by the edge of the river\"."}], "task": "refering", "answer_template": "The \"male zebra , bigger body , looking off the screen standing by the edge of the river\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242], "bbox": [0.369390625, 0.5076044568245125, 0.587421875, 0.8455153203342617], "iscrowd": 0, "area": 7486.71805, "rle": {"size": [359, 640], "counts": "`nb21U;1N3N2M3N2N2M3N2M2O2M3N1N100O2ASOcFm0^9VO]Fk0d9YOWFg0j9;0O1O1O1O10O01O1O1001O1O1O1O100O010O100O100O1O100O2O2M2O1N3N0O2O0O2N101N2O0O2O1N3N2M4M2M3N2N3M2kG[Mc7g2XH[Mj7f2PH]MQ8k22O1kNmGTOS8n0QHlNn7V1WHcNj7_1[HZNe7g1`HSN`7o1k0;Ea0_ObNlGJS85THFj7;\\H_Od7a0aHZO_7f0gHSOZ7m0kHlNW7T1oHeNR7[1]10O0100000000000000O10O100000000nG^N_6b1_I`Na6_1^IcNb6_1ZIcNf6a1TIaNl6b1nHaNR7b1iH`NW7c1dH_N\\7d1_H^Na7e1ZH]Nf7f1UH\\Nk71PHn00SOo7MUHh1k7UNXHk1h7RN[Hn1e7oM_Hm1d7QN^Hl1e7QN_Hj1e7SN^Hj1e7SN^Hi1c8K4M3L3N2\\ObF]Ob9b0a000O10000a0@<C1O05L5J5L5J[Zl2"}, "label": "male zebra , bigger body , looking off the screen standing by the edge of the river"}]}
{"id": 517494, "image": "COCO_train2014_000000517494.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra ahead of the other zebra\"."}], "task": "refering", "answer_template": "The \"a zebra ahead of the other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242], "bbox": [0.369390625, 0.5076044568245125, 0.587421875, 0.8455153203342617], "iscrowd": 0, "area": 7486.71805, "rle": {"size": [359, 640], "counts": "`nb21U;1N3N2M3N2N2M3N2M2O2M3N1N100O2ASOcFm0^9VO]Fk0d9YOWFg0j9;0O1O1O1O10O01O1O1001O1O1O1O100O010O100O100O1O100O2O2M2O1N3N0O2O0O2N101N2O0O2O1N3N2M4M2M3N2N3M2kG[Mc7g2XH[Mj7f2PH]MQ8k22O1kNmGTOS8n0QHlNn7V1WHcNj7_1[HZNe7g1`HSN`7o1k0;Ea0_ObNlGJS85THFj7;\\H_Od7a0aHZO_7f0gHSOZ7m0kHlNW7T1oHeNR7[1]10O0100000000000000O10O100000000nG^N_6b1_I`Na6_1^IcNb6_1ZIcNf6a1TIaNl6b1nHaNR7b1iH`NW7c1dH_N\\7d1_H^Na7e1ZH]Nf7f1UH\\Nk71PHn00SOo7MUHh1k7UNXHk1h7RN[Hn1e7oM_Hm1d7QN^Hl1e7QN_Hj1e7SN^Hj1e7SN^Hi1c8K4M3L3N2\\ObF]Ob9b0a000O10000a0@<C1O05L5J5L5J[Zl2"}, "label": "a zebra ahead of the other zebra"}]}
{"id": 517494, "image": "COCO_train2014_000000517494.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 221, 222, 223, 224, 244, 246, 247], "bbox": [0.548375, 0.5015598885793872, 0.773859375, 0.8476323119777159], "iscrowd": 0, "area": 7892.753450000002, "rle": {"size": [359, 640], "counts": "aXk38l:5K3N3L3M3BXO[Fj0d9ZOXFh0g9XOXFh0j9WOVFi0j9WOUFi0l9WORFj0o9VOPFj0j9<00100O100O1O100O1O100O2O01O100O101N101N10000O10O12M2O2N1N2O1N2O1O1N3N3L3M3N6iGlMm6`2eHbM\\7a2\\HcMd7P31O1O6J1O00001L3M3oN]H]Nd7[1hH_NZ7\\1\\1O000000001O0000000000000000O10000000000O100000000O01000O01000O10O10O10O10O10O00100O1O10O2O2N2N2N2nG[N^6g1_IZNa6h1\\IYNd6i1XIYNh6i1TIYNm6h1oHZNQ7i1jHYNV7j1bH[N^7h1XH_Ni7_2000M3N3N1K5I8I7I7D;2O1O1aGXN[7j1\\H\\Ng7a21O2N2N2N2RNeGl0^8lNQHDO3Slb1"}, "label": "the zebra on the right"}]}
{"id": 230777, "image": "COCO_train2014_000000230777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man doing skateboard trick\"."}], "task": "refering", "answer_template": "The \"man doing skateboard trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 127, 128, 129], "bbox": [0.33271875, 0.0, 0.673125, 0.3635362997658079], "iscrowd": 0, "area": 14716.143000000004, "rle": {"size": [427, 640], "counts": "Whi21Z=0O010O100O010O100O010O100O00100O10O0100O10O01O100O100O00100O1O10O0100O1O100O00100O100O1O010O1O100O10O01O100O1O010O100O100O010O1O100O1O1O1O1N2O1O100O1O1O1O1O010O1O1O1O10O01O1O100O1O00100O10001N10000QOeNmE\\1n9jNQFV1m9mN`EB<b1S:nN^EC>_1S:UOlEk0T:UOkEl0U:UOjEl0U:UOjEk0U:WOiEj0W:VOhEl0i0ZN[8k0iFn0j0YN[8k0iFm0k0ZN[8j0hFn0k0YN]8i2aGYM^8h2`GZM^8h2`GZM_8g2_GZMa8g2]G[Mb8f2\\G\\Mc8g300O001O1O100O1O010O10000000O101N2N1O:F2N1O010O00O10000O10001O000000001O000000001N1000001O0000000000000010O001O001O001O0O2N1O2N2N2M3N2N2M3N3L4M3L4M3L4L2M4M2N2M4M2O101O001O0000100O2N2N1Nm0SO6J:E=D\\Qg2"}, "label": "man doing skateboard trick"}]}
{"id": 230777, "image": "COCO_train2014_000000230777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white t - shirt and dark pants jumping with his skateboard\"."}], "task": "refering", "answer_template": "The \"a man in a white t - shirt and dark pants jumping with his skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 127, 128, 129], "bbox": [0.33271875, 0.0, 0.673125, 0.3635362997658079], "iscrowd": 0, "area": 14716.143000000004, "rle": {"size": [427, 640], "counts": "Whi21Z=0O010O100O010O100O010O100O00100O10O0100O10O01O100O100O00100O1O10O0100O1O100O00100O100O1O010O1O100O10O01O100O1O010O100O100O010O1O100O1O1O1O1N2O1O100O1O1O1O1O010O1O1O1O10O01O1O100O1O00100O10001N10000QOeNmE\\1n9jNQFV1m9mN`EB<b1S:nN^EC>_1S:UOlEk0T:UOkEl0U:UOjEl0U:UOjEk0U:WOiEj0W:VOhEl0i0ZN[8k0iFn0j0YN[8k0iFm0k0ZN[8j0hFn0k0YN]8i2aGYM^8h2`GZM^8h2`GZM_8g2_GZMa8g2]G[Mb8f2\\G\\Mc8g300O001O1O100O1O010O10000000O101N2N1O:F2N1O010O00O10000O10001O000000001O000000001N1000001O0000000000000010O001O001O001O0O2N1O2N2N2M3N2N2M3N3L4M3L4M3L4L2M4M2N2M4M2O101O001O0000100O2N2N1Nm0SO6J:E=D\\Qg2"}, "label": "a man in a white t - shirt and dark pants jumping with his skateboard"}]}
{"id": 58753, "image": "COCO_train2014_000000058753.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color tv on the wall\"."}], "task": "refering", "answer_template": "The \"white color tv on the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.7051875, 0.4311666666666667, 0.977921875, 0.6935], "iscrowd": 0, "area": 20353.817750000002, "rle": {"size": [480, 640], "counts": "hfc6:i;m2F:000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000O10000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000O100000000000000000001O00000000000000000000000000001O00000000]Oc0\\Od0\\Od0\\Od0[Oe0\\OU\\6"}, "label": "white color tv on the wall"}]}
{"id": 58753, "image": "COCO_train2014_000000058753.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3d wallpaper inside room\"."}], "task": "refering", "answer_template": "The \"3d wallpaper inside room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.7051875, 0.4311666666666667, 0.977921875, 0.6935], "iscrowd": 0, "area": 20353.817750000002, "rle": {"size": [480, 640], "counts": "hfc6:i;m2F:000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000O10000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000O100000000000000000001O00000000000000000000000000001O00000000]Oc0\\Od0\\Od0\\Od0[Oe0\\OU\\6"}, "label": "3d wallpaper inside room"}]}
{"id": 460164, "image": "COCO_train2014_000000460164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich in the person ' s left hand\"."}], "task": "refering", "answer_template": "The \"a sandwich in the person ' s left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 204, 205, 206], "bbox": [0.25172, 0.43141689373297004, 0.6915399999999999, 0.8792370572207084], "iscrowd": 0, "area": 22213.1102, "rle": {"size": [367, 500], "counts": "h[]1g0h:8G:G8G3N1N3N2M3N1N3N1O1N101N101N101N101N2O001N101N2O7H4M1N3N2M3N2N2M2O2M2O1N3N1N2O2N1O2M2O1O2N00O01000000O1000000O100000O10O1000000O0100000O1000O1000O100000O01000000O10O1000O10O100000O010000O10O1000O10O10O100000O01000000O01000O10O1000O1000O010000O10O10O1000O010000O010SMjHa1V7^NRI[1m6eNWIW1j6hNXIW1h6hNYIX1f6iN[IU1f6jN\\IU1d6jN^IU1b6jN`IT1`6lNbIS1^6nNbIQ1^6oNcIP1\\6QOdIo0\\6QOeIn0[6ROeIn0Z6SOfIm0Z6SOgIl0Y6TOgIl0Y6TOgIl0X6UOiIj0W6VOiIj0W6UOjIk0U6VOlIi0T6WOlIi0T6WOlIi0T6WOmIh0R6XOoIi0P6WOPJi0P6WOPJi0P6WOPJi0o5WOSJh0m5XOSJh0m5XOSJg0m5ZOSJe0n5ZOSJf0m5ZOSJe0n5[ORJd0n5]ORJc0n5]OSJa0n5_ORJ`0o5@QJ?o5BQJ>o5BQJ=P6BQJ=P6CPJ=o5DQJ;P6EPJ:Q6FoI9R6GnI8R6InI6S6JmI6S6JmI5T6JmI5S6LmI3T6MlI2U6NkI1V6OjI0V61jIOV61jINX61hINY61gIOZ61eIO\\61dIN]62bIN_62aIM`63_IMb63^ILc63]IMe62ZINg62YIMi62VIOj61WIMk62UILn63SIIP77QIDT7;lHXOb7h0b101O001O107H8H8HVPg1"}, "label": "a sandwich in the person ' s left hand"}]}
{"id": 460164, "image": "COCO_train2014_000000460164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich that is facing horizontal in the man ' s left hand\"."}], "task": "refering", "answer_template": "The \"the sandwich that is facing horizontal in the man ' s left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 204, 205, 206], "bbox": [0.25172, 0.43141689373297004, 0.6915399999999999, 0.8792370572207084], "iscrowd": 0, "area": 22213.1102, "rle": {"size": [367, 500], "counts": "h[]1g0h:8G:G8G3N1N3N2M3N1N3N1O1N101N101N101N101N2O001N101N2O7H4M1N3N2M3N2N2M2O2M2O1N3N1N2O2N1O2M2O1O2N00O01000000O1000000O100000O10O1000000O0100000O1000O1000O100000O01000000O10O1000O10O100000O010000O10O1000O10O10O100000O01000000O01000O10O1000O1000O010000O10O10O1000O010000O010SMjHa1V7^NRI[1m6eNWIW1j6hNXIW1h6hNYIX1f6iN[IU1f6jN\\IU1d6jN^IU1b6jN`IT1`6lNbIS1^6nNbIQ1^6oNcIP1\\6QOdIo0\\6QOeIn0[6ROeIn0Z6SOfIm0Z6SOgIl0Y6TOgIl0Y6TOgIl0X6UOiIj0W6VOiIj0W6UOjIk0U6VOlIi0T6WOlIi0T6WOlIi0T6WOmIh0R6XOoIi0P6WOPJi0P6WOPJi0P6WOPJi0o5WOSJh0m5XOSJh0m5XOSJg0m5ZOSJe0n5ZOSJf0m5ZOSJe0n5[ORJd0n5]ORJc0n5]OSJa0n5_ORJ`0o5@QJ?o5BQJ>o5BQJ=P6BQJ=P6CPJ=o5DQJ;P6EPJ:Q6FoI9R6GnI8R6InI6S6JmI6S6JmI5T6JmI5S6LmI3T6MlI2U6NkI1V6OjI0V61jIOV61jINX61hINY61gIOZ61eIO\\61dIN]62bIN_62aIM`63_IMb63^ILc63]IMe62ZINg62YIMi62VIOj61WIMk62UILn63SIIP77QIDT7;lHXOb7h0b101O001O107H8H8HVPg1"}, "label": "the sandwich that is facing horizontal in the man ' s left hand"}]}
{"id": 460164, "image": "COCO_train2014_000000460164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding two rib sandwiches\"."}], "task": "refering", "answer_template": "The \"a man holding two rib sandwiches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 82, 83, 84, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.20288, 0.0, 1.0, 0.979100817438692], "iscrowd": 0, "area": 85285.08505, "rle": {"size": [367, 500], "counts": "adT12[;101N2N2O001O1O1O1O001N2O1O1O001O1O1O1O001N2O1O001O1O1O001O1O1O00100O1O0010000O010O10000O2O000O10001O00000000008H0O10O10000000000000O2O1O1_GdNm6]1QIcNP7]1PIcNP7^1oHbNQ7_1mHbNS7_1lHaNT7`1kH`NU7a1iH`NV7c1hH]NX7f1eHZN[7j1`HVNa7m1\\HSNd7Q2XHoMh7S2UHnMk7S2THmMl7S2SHnMm7S2RHmMd5IUL\\2UNlMe5KTLZ2VNkMd5OTLW2VNjMf52RLT2XNjMf52RLU2VNjMg52SLU2UNiMh52SLV2SNhMj53SLV2RNfMl54RLb2o3_MPL`2P4aMoK_2R4aMnK^2R4cMnK\\2S4dMmK[2T4eMlKZ2T4gMlKX2U4hMkKW2U4jMkKU2V4kMiKU2X4lMgKS2Y4nMgKQ2Z4oMfKP2[4PNdKP2]4PNcKo1]4RNcKm1^4SNaKm1`4SN`Kl1a4TN_Kk1b4UN^Kj1b4XN\\Kh1e4XN[Kg1f4YNZKf1g4ZNYKf1f4[NYKe1h4[NWKe1j4[NVKd1k4\\NTKd1m4\\NSKc1m4^NRKb1o4_NPK`1Q5`NnJ`1S5`NlJ`1U5`NkJ_1U5bNkJ]1V5cNjJ\\1W5dNiJ[1X5eNhJZ1Y5fNgJY1Z5gNfJX1Z5iNfJV1[5jNeJU1\\5kNdJT1]5lNdJR1]5nNcJR1]5nNcJQ1]5POcJP1]5POcJo0^5SO`Jm0`5XO[Jg0f5_OSJb0m5^OSJb0m5^OSJa0o5^OQJb0o5^OQJa0P6_OoIb0Q6^OoIa0R6_OnIa0S6^OmIb0R6_OmIb0S6^OmIb0S6^OmIb0S6^OmIb0k2TNZOZ1kMb0k2VNWOY1mMb0k2XNVOV1oMb0k2ZNTOT1QNb0j2]NSOQ1SNb0j2_NPOP1UNb0j2aNoNm0WNb0j2cNmNk0YNb0i2fNlNh0[Nb0i2gNjNh0]Na0h2iNjNf0]Nb0i2hNjNf0]Nb0h2jNiNe0_Na0h2jNiNe0_Na0g2lNhNd0`Nb0g2kNhNc0aNb0f2lNhNc0bNa0f2mNgNb0cNa0f2mNfNc0dN`0f2nNeNb0dNa0g2nNcNb0fN`0g2nNcNb0fN`0f2PObNa0hN?f2POaNb0hN?g2oNaNb0hN?g2oN`Nc0iN>g2PO_Nb0jN>g2PO^Nc0kN=g2PO]Nd0kN=g2PO^Nc0kN=g2PO]Nd0lN<]1nMoNR1h0d0lN<Y1RNSOn0g0e0lN<V1UNWOj0f0f0mN;R1YN[Of0f0f0mN;n0\\N@c0d0g0nN:j0`NC`0e0f0nN:f0dNG=c0f0oN:c0gNK9b0g0PO9?kNO5b0g0PO9;oN31a0h0QO87SO7Ma0h0QO83WO:Ja0h0QO80ZO>F`0i0RO7L]Oc0C>j0SO6H_Oi0A;k0TO5D@P1@7l0TO5ABU1^O4l0VO4]OFY1ZO3m0WOW2d0mL4m0XOV2d0mL3n0YOU2d0mL3n0XOW2d0kL3o0YOV2d0kL2P1ZOV2c0jL2Q1[OU2c0jL1R1[OU2c0jL1R1\\OU2b0iL1S1\\OU2c0iLOS1\\OV2e0gLNT1\\OW2e0eLMV1\\OW2g0cLLW1\\OW2h0bLKX1[OX2i0aLKX1[OX2j0`LJY1[OX2k0_LIZ1\\OW2k0_LH[1]OW2j0^LH\\1^OV2j0_LF\\1@U2i0`LF\\1AT2i0`LE]1BS2i0`LD^1CR2i0`LC_1DQ2i0`LB`1EP2i0`LAa1Fo1i0`L@b1Gn1h0aL@b1HP2e0^L@e1KP2b0[L@h1NP2?YL@i11Q2<VL@l14Q29SL@o17Q26CJ`02ANb0O^O1e0L[O4h0IXO7P6000000000000O1000000000000000000000000000000000000000000000000000lJIPO7o0JQO6o0JQO6P1IPO7P1IPO7P1IPO7P1JoN6R1InN7R1InN7R1InN7R1InN7S1IlN7T1IlN7T1IlN7V6000oJIjN7V1IjN7W600000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000\\KIPN7P2IPN7P2IPN7d60]KInM7R2InM7R2InM7R2JmM6S2JmM6S2JmM6T2IlM7S2JmM6S2JmM6S2JmM6S2InM7R2InM7R2InM7Q2JoM6Q2JoM6Q2JoM6Q2JoM6Q2IPN7o1JQN6o1JQN6o1JQN6o1JQN6o1JQN6o1JQN6n1JSN6m1JSN6m1JSN6m1JSN6m1JSN6l1KTN5l1JUN6k1JUN6k1JUN6k1JUN6k1JUN6j1KVN5j1JWN6i1JWN6i1JWN6i1JWN6h1KXN5h1KXN5h1JYN6g1JYN8e1H[Nk0R1UOnNX2DiM<b5YL^Jg35"}, "label": "a man holding two rib sandwiches"}]}
{"id": 460164, "image": "COCO_train2014_000000460164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 82, 83, 84, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.20288, 0.0, 1.0, 0.979100817438692], "iscrowd": 0, "area": 85285.08505, "rle": {"size": [367, 500], "counts": "adT12[;101N2N2O001O1O1O1O001N2O1O1O001O1O1O1O001N2O1O001O1O1O001O1O1O00100O1O0010000O010O10000O2O000O10001O00000000008H0O10O10000000000000O2O1O1_GdNm6]1QIcNP7]1PIcNP7^1oHbNQ7_1mHbNS7_1lHaNT7`1kH`NU7a1iH`NV7c1hH]NX7f1eHZN[7j1`HVNa7m1\\HSNd7Q2XHoMh7S2UHnMk7S2THmMl7S2SHnMm7S2RHmMd5IUL\\2UNlMe5KTLZ2VNkMd5OTLW2VNjMf52RLT2XNjMf52RLU2VNjMg52SLU2UNiMh52SLV2SNhMj53SLV2RNfMl54RLb2o3_MPL`2P4aMoK_2R4aMnK^2R4cMnK\\2S4dMmK[2T4eMlKZ2T4gMlKX2U4hMkKW2U4jMkKU2V4kMiKU2X4lMgKS2Y4nMgKQ2Z4oMfKP2[4PNdKP2]4PNcKo1]4RNcKm1^4SNaKm1`4SN`Kl1a4TN_Kk1b4UN^Kj1b4XN\\Kh1e4XN[Kg1f4YNZKf1g4ZNYKf1f4[NYKe1h4[NWKe1j4[NVKd1k4\\NTKd1m4\\NSKc1m4^NRKb1o4_NPK`1Q5`NnJ`1S5`NlJ`1U5`NkJ_1U5bNkJ]1V5cNjJ\\1W5dNiJ[1X5eNhJZ1Y5fNgJY1Z5gNfJX1Z5iNfJV1[5jNeJU1\\5kNdJT1]5lNdJR1]5nNcJR1]5nNcJQ1]5POcJP1]5POcJo0^5SO`Jm0`5XO[Jg0f5_OSJb0m5^OSJb0m5^OSJa0o5^OQJb0o5^OQJa0P6_OoIb0Q6^OoIa0R6_OnIa0S6^OmIb0R6_OmIb0S6^OmIb0S6^OmIb0S6^OmIb0k2TNZOZ1kMb0k2VNWOY1mMb0k2XNVOV1oMb0k2ZNTOT1QNb0j2]NSOQ1SNb0j2_NPOP1UNb0j2aNoNm0WNb0j2cNmNk0YNb0i2fNlNh0[Nb0i2gNjNh0]Na0h2iNjNf0]Nb0i2hNjNf0]Nb0h2jNiNe0_Na0h2jNiNe0_Na0g2lNhNd0`Nb0g2kNhNc0aNb0f2lNhNc0bNa0f2mNgNb0cNa0f2mNfNc0dN`0f2nNeNb0dNa0g2nNcNb0fN`0g2nNcNb0fN`0f2PObNa0hN?f2POaNb0hN?g2oNaNb0hN?g2oN`Nc0iN>g2PO_Nb0jN>g2PO^Nc0kN=g2PO]Nd0kN=g2PO^Nc0kN=g2PO]Nd0lN<]1nMoNR1h0d0lN<Y1RNSOn0g0e0lN<V1UNWOj0f0f0mN;R1YN[Of0f0f0mN;n0\\N@c0d0g0nN:j0`NC`0e0f0nN:f0dNG=c0f0oN:c0gNK9b0g0PO9?kNO5b0g0PO9;oN31a0h0QO87SO7Ma0h0QO83WO:Ja0h0QO80ZO>F`0i0RO7L]Oc0C>j0SO6H_Oi0A;k0TO5D@P1@7l0TO5ABU1^O4l0VO4]OFY1ZO3m0WOW2d0mL4m0XOV2d0mL3n0YOU2d0mL3n0XOW2d0kL3o0YOV2d0kL2P1ZOV2c0jL2Q1[OU2c0jL1R1[OU2c0jL1R1\\OU2b0iL1S1\\OU2c0iLOS1\\OV2e0gLNT1\\OW2e0eLMV1\\OW2g0cLLW1\\OW2h0bLKX1[OX2i0aLKX1[OX2j0`LJY1[OX2k0_LIZ1\\OW2k0_LH[1]OW2j0^LH\\1^OV2j0_LF\\1@U2i0`LF\\1AT2i0`LE]1BS2i0`LD^1CR2i0`LC_1DQ2i0`LB`1EP2i0`LAa1Fo1i0`L@b1Gn1h0aL@b1HP2e0^L@e1KP2b0[L@h1NP2?YL@i11Q2<VL@l14Q29SL@o17Q26CJ`02ANb0O^O1e0L[O4h0IXO7P6000000000000O1000000000000000000000000000000000000000000000000000lJIPO7o0JQO6o0JQO6P1IPO7P1IPO7P1IPO7P1JoN6R1InN7R1InN7R1InN7R1InN7S1IlN7T1IlN7T1IlN7V6000oJIjN7V1IjN7W600000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000\\KIPN7P2IPN7P2IPN7d60]KInM7R2InM7R2InM7R2JmM6S2JmM6S2JmM6T2IlM7S2JmM6S2JmM6S2JmM6S2InM7R2InM7R2InM7Q2JoM6Q2JoM6Q2JoM6Q2JoM6Q2IPN7o1JQN6o1JQN6o1JQN6o1JQN6o1JQN6o1JQN6n1JSN6m1JSN6m1JSN6m1JSN6m1JSN6l1KTN5l1JUN6k1JUN6k1JUN6k1JUN6k1JUN6j1KVN5j1JWN6i1JWN6i1JWN6i1JWN6h1KXN5h1KXN5h1JYN6g1JYN8e1H[Nk0R1UOnNX2DiM<b5YL^Jg35"}, "label": "a man wearing a black shirt"}]}
{"id": 279946, "image": "COCO_train2014_000000279946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black container divided into two half containing rice and gravy\"."}], "task": "refering", "answer_template": "The \"black container divided into two half containing rice and gravy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [315, 316, 317, 318, 319, 320, 321, 322, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 427, 428, 429, 430, 431, 432], "bbox": [0.25086601307189543, 0.6730555555555556, 0.7131372549019609, 0.8746241830065359], "iscrowd": 0, "area": 30084.241899999997, "rle": {"size": [612, 612], "counts": "]`l24lb04M4K5L3M4M3N2M2N3M3N2M2N3M3M2O2M3M3M2O2M3M2N4L6K5J6J6J5K4M1N200O10000O10001N100O10000O10000O100O10000O10000O100O1000000000000000001O00000000000000000000000000000001O0000000000000000000001O0000000000000000000000000000000000000000000000000000001O000000000000000001O000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O000000000000000000000001O000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O0000000000000000000000000001O000000000N2M3N3L3M4M2M4M2M4O0O2N1O2O0O2N2N101N1O2N101N1O2N3Nh0WOh0XOTgX3"}, "label": "black container divided into two half containing rice and gravy"}]}
{"id": 279946, "image": "COCO_train2014_000000279946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black plastic compartment dish with rice in the left side and lentils in the right side\"."}], "task": "refering", "answer_template": "The \"a black plastic compartment dish with rice in the left side and lentils in the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [315, 316, 317, 318, 319, 320, 321, 322, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 427, 428, 429, 430, 431, 432], "bbox": [0.25086601307189543, 0.6730555555555556, 0.7131372549019609, 0.8746241830065359], "iscrowd": 0, "area": 30084.241899999997, "rle": {"size": [612, 612], "counts": "]`l24lb04M4K5L3M4M3N2M2N3M3N2M2N3M3M2O2M3M3M2O2M3M2N4L6K5J6J6J5K4M1N200O10000O10001N100O10000O10000O100O10000O10000O100O1000000000000000001O00000000000000000000000000000001O0000000000000000000001O0000000000000000000000000000000000000000000000000000001O000000000000000001O000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O000000000000000000000001O000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O0000000000000000000000000001O000000000N2M3N3L3M4M2M4M2M4O0O2N1O2O0O2N2N101N1O2N101N1O2N3Nh0WOh0XOTgX3"}, "label": "a black plastic compartment dish with rice in the left side and lentils in the right side"}]}
{"id": 271760, "image": "COCO_train2014_000000271760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing closest to the window\"."}], "task": "refering", "answer_template": "The \"a zebra standing closest to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 280, 281, 282, 283, 303, 304, 305, 306, 307, 308, 309, 330, 331, 332, 353, 354], "bbox": [0.20079687499999999, 0.6864583333333333, 0.44340624999999995, 0.9079583333333333], "iscrowd": 0, "area": 6192.7241, "rle": {"size": [480, 640], "counts": "cjl1<b>4L3M4M2N1N2O001O00001O001O1O001O001N2O0O2N2O0O2G90O1000000O1000000O14L4L4MO1O1O001O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O100O1O001O1O100O001O1O00100O001O10O0100O010O1O010O010O01O00001O001O010O001O00001O002N4M3L4L4L3M4L4L1O00000000000000000000001O000000000000000000000000001O000000K5J7H8I8GY`V5"}, "label": "a zebra standing closest to the window"}]}
{"id": 271760, "image": "COCO_train2014_000000271760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second zebra in field\"."}], "task": "refering", "answer_template": "The \"second zebra in field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [284, 285, 286, 307, 308, 309, 310, 311, 312, 313, 333, 334, 335, 336, 356, 358, 359, 381], "bbox": [0.356609375, 0.7190208333333333, 0.622109375, 0.9990833333333333], "iscrowd": 0, "area": 6389.051999999999, "rle": {"size": [480, 640], "counts": "df[33k>3M2N2BHRB:m=InA:Q>HlA:R>;00O1O1O1000000001O0001O0001O001O001O001O001O001O001O0010O01O00001O001O001O001O001O0010O010O10O[P?Ngo@b0UBYOY<g1000001O0000A?000000O100000O1000O10000000000O1000000000O0100000000O1000000000O101O1O1O1O1O2N2N1O2M3N101N2N101XDjMb:X2ZEjMg:W2VEjMj:X2REjMn:X2nDjMS;V2kDkMU;P3001O01OO2O0O1O2N2J6J6K5K6J5K5H8G:E:Bb0FkQa3"}, "label": "second zebra in field"}]}
{"id": 402833, "image": "COCO_train2014_000000402833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319, 340, 341, 342, 364, 365, 387, 388], "bbox": [0.78278125, 0.5773333333333334, 0.9042656250000001, 1.0], "iscrowd": 0, "area": 8588.746499999996, "rle": {"size": [480, 640], "counts": "oV[75g>7J3LNgAFR>?26]O[OeBj0Y=[O_Bj0_=<M5L4LGPCiNl<V1ZCiNa<W1dCiNX<V1mChNS<U1TDhNl;U1Q1L4M3L4O12PCYOc;j0VDF^;<\\D1[;2^D;Y;H_Df0X;j1F:G;E<D;E4K1VOoJdGR5X8VKbGl4Z8\\KaGe4_8^K]Gc4c8m00O1000O10000000000001O0O101O000fJYG[4h8cKZG\\4f8bK]G]4d8`K^G`4c8]K`GS4P9iKSGn3W9mKnFP3R;K6N101N101N101N2O0O2N1N3M2XNj1N1N:G;DYXl0"}, "label": "a man wearing blue shirt"}]}
{"id": 402833, "image": "COCO_train2014_000000402833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy is walking with blue color shirt near the elephant\"."}], "task": "refering", "answer_template": "The \"a boy is walking with blue color shirt near the elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319, 340, 341, 342, 364, 365, 387, 388], "bbox": [0.78278125, 0.5773333333333334, 0.9042656250000001, 1.0], "iscrowd": 0, "area": 8588.746499999996, "rle": {"size": [480, 640], "counts": "oV[75g>7J3LNgAFR>?26]O[OeBj0Y=[O_Bj0_=<M5L4LGPCiNl<V1ZCiNa<W1dCiNX<V1mChNS<U1TDhNl;U1Q1L4M3L4O12PCYOc;j0VDF^;<\\D1[;2^D;Y;H_Df0X;j1F:G;E<D;E4K1VOoJdGR5X8VKbGl4Z8\\KaGe4_8^K]Gc4c8m00O1000O10000000000001O0O101O000fJYG[4h8cKZG\\4f8bK]G]4d8`K^G`4c8]K`GS4P9iKSGn3W9mKnFP3R;K6N101N101N101N2O0O2N1N3M2XNj1N1N:G;DYXl0"}, "label": "a boy is walking with blue color shirt near the elephant"}]}
{"id": 566681, "image": "COCO_train2014_000000566681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman in a white shirt and long black skirt\"."}], "task": "refering", "answer_template": "The \"a blonde woman in a white shirt and long black skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 137, 138, 139, 140, 154, 155, 156, 157, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 346, 358, 359, 360, 361, 362, 363, 375, 376, 377, 378, 379, 380], "bbox": [0.08408333333333333, 0.326578125, 0.4774791666666667, 0.9977499999999999], "iscrowd": 0, "area": 46391.104449999984, "rle": {"size": [640, 480], "counts": "eZi03Xc0k0UOl0SOm0ROn0E;H8G9G9G6J3[HoJ7T5TO`Kj0c4bNPL\\1S4oMaLP2b3[MQMc2R3iLaMU3b2YLoMe3T2nKXNP4k1cKaN\\4b1VKjNi4X1jJTOU5n0^J^Oa5d0QJIo58dI4[6OVI>i6]600O1O100O1O100O1O100O1O1O100O1O100001O000mB_Ig<b6WC_Ii<a6WC_Ii<a6VCaIi<`6UCaIk<_6UCaIk<_6TCcIk<]6TCdIl<]6SCcIm<]6RCeIm<g60001O001O00001O00001O00001O00001O00001O00001O00001O002N1eElHh7V7RHnHn7T7lGPIT8R7fGRIZ8P7`GTI`8m6ZGXIf8j6SG[Im8g6lF^IT9d6eFaI[9a6^FdIb9]6XFhIh9Z6QFkIo9S8000000000000bGTFh6l9XIVFf6j9ZIXFd6h9[I[Fc6e9]I]Fa6c9_I`F^6a9`IbF^6^9bIdF\\6\\9dIfFZ6[9eIhFX6X9hIjFV6V9jIlFT6U9kImFS6S9mIoFQ6R9nIQGo5o8QJSGm5m8SJUGk5l8TJVGj5j8VJYGg5g8YJ[Ge5f8ZJ\\Gd5d8\\J^Gb5b8^J`G`5a8_JbG^5^8bJdG\\5\\8dJfGZ5[8eJgGY5Y8gJjGV5W8iJkGU5U8kJmGS5S8mJoGd3_9[LWG]1[:cNb5O001O00001O001O001O001O001O00001O001O001O001O00001O001O0010O01O000010O01O001O0010O01N10001O001N101O001N1000o`l4"}, "label": "a blonde woman in a white shirt and long black skirt"}]}
{"id": 566681, "image": "COCO_train2014_000000566681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one small girl wearing white top is touching the elephant\"."}], "task": "refering", "answer_template": "The \"there is one small girl wearing white top is touching the elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 137, 138, 139, 140, 154, 155, 156, 157, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 346, 358, 359, 360, 361, 362, 363, 375, 376, 377, 378, 379, 380], "bbox": [0.08408333333333333, 0.326578125, 0.4774791666666667, 0.9977499999999999], "iscrowd": 0, "area": 46391.104449999984, "rle": {"size": [640, 480], "counts": "eZi03Xc0k0UOl0SOm0ROn0E;H8G9G9G6J3[HoJ7T5TO`Kj0c4bNPL\\1S4oMaLP2b3[MQMc2R3iLaMU3b2YLoMe3T2nKXNP4k1cKaN\\4b1VKjNi4X1jJTOU5n0^J^Oa5d0QJIo58dI4[6OVI>i6]600O1O100O1O100O1O100O1O1O100O1O100001O000mB_Ig<b6WC_Ii<a6WC_Ii<a6VCaIi<`6UCaIk<_6UCaIk<_6TCcIk<]6TCdIl<]6SCcIm<]6RCeIm<g60001O001O00001O00001O00001O00001O00001O00001O00001O002N1eElHh7V7RHnHn7T7lGPIT8R7fGRIZ8P7`GTI`8m6ZGXIf8j6SG[Im8g6lF^IT9d6eFaI[9a6^FdIb9]6XFhIh9Z6QFkIo9S8000000000000bGTFh6l9XIVFf6j9ZIXFd6h9[I[Fc6e9]I]Fa6c9_I`F^6a9`IbF^6^9bIdF\\6\\9dIfFZ6[9eIhFX6X9hIjFV6V9jIlFT6U9kImFS6S9mIoFQ6R9nIQGo5o8QJSGm5m8SJUGk5l8TJVGj5j8VJYGg5g8YJ[Ge5f8ZJ\\Gd5d8\\J^Gb5b8^J`G`5a8_JbG^5^8bJdG\\5\\8dJfGZ5[8eJgGY5Y8gJjGV5W8iJkGU5U8kJmGS5S8mJoGd3_9[LWG]1[:cNb5O001O00001O001O001O001O001O00001O001O001O001O00001O001O0010O01O000010O01O001O0010O01N10001O001N101O001N1000o`l4"}, "label": "there is one small girl wearing white top is touching the elephant"}]}
{"id": 50591, "image": "COCO_train2014_000000050591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black armchair to the right of the viewer\"."}], "task": "refering", "answer_template": "The \"the black armchair to the right of the viewer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338], "bbox": [0.56503125, 0.6693442622950819, 0.8444062500000001, 0.9599531615925058], "iscrowd": 0, "area": 16291.742949999996, "rle": {"size": [427, 640], "counts": "lWg45Q=7J6L4L4L4M3N2O2M3N2M4M2N2M6K?@2O0O100O101N100O100O100O100O100O2O0O100O100O100O10001N100O100O1O1O1O1O1O1O2N1O1O1O1O1O00O10000000O01000O10O0100O1000O0100O10O10O100O01000O100O010O1000O0100O1000O0100O10O10O1000O10O100000O10O1000O1000O10O100000O10O1000000O100000000O1000000O1000000O100000000O10001O0O2O001O1N101O001O1N101O001N101O1O001N101O001N2O001O0O2O1ZOe0ROi]Y1"}, "label": "the black armchair to the right of the viewer"}]}
{"id": 50591, "image": "COCO_train2014_000000050591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black chair on the right\"."}], "task": "refering", "answer_template": "The \"the black chair on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338], "bbox": [0.56503125, 0.6693442622950819, 0.8444062500000001, 0.9599531615925058], "iscrowd": 0, "area": 16291.742949999996, "rle": {"size": [427, 640], "counts": "lWg45Q=7J6L4L4L4M3N2O2M3N2M4M2N2M6K?@2O0O100O101N100O100O100O100O100O2O0O100O100O100O10001N100O100O1O1O1O1O1O1O2N1O1O1O1O1O00O10000000O01000O10O0100O1000O0100O10O10O100O01000O100O010O1000O0100O1000O0100O10O10O1000O10O100000O10O1000O1000O10O100000O10O1000000O100000000O1000000O1000000O100000000O10001O0O2O001O1N101O001O1N101O001N101O1O001N101O001N2O001O0O2O1ZOe0ROi]Y1"}, "label": "the black chair on the right"}]}
{"id": 50591, "image": "COCO_train2014_000000050591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black chair on the left with a blue pillow on it\"."}], "task": "refering", "answer_template": "The \"black chair on the left with a blue pillow on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308], "bbox": [0.184421875, 0.6375644028103045, 0.4526093749999999, 0.9311709601873537], "iscrowd": 0, "area": 14318.195150000001, "rle": {"size": [427, 640], "counts": "j_a1?l<5J5L5K5K5K5J5LQ1oN0000001N100000000O2O000000001N100000001N100000001O0O1000001O0O100000001O0O11O00010O00000010O000001O01O01O0000010O00001O0001O01O0000010O00001O01O0001O00010O0000001O01O0001O00010O00000010O0001O0001O01O0000010O00001O00O1O00O2O0000001N1000001O0O10001O000O101O00000O2O0000001N1000001O0O10001O0O10001O000O101O00002M3N1UOREXOP;a0cEQO_:i0U1I6K6IUSb4"}, "label": "black chair on the left with a blue pillow on it"}]}
{"id": 50591, "image": "COCO_train2014_000000050591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair to the left of the table with pink flowers on it\"."}], "task": "refering", "answer_template": "The \"chair to the left of the table with pink flowers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308], "bbox": [0.184421875, 0.6375644028103045, 0.4526093749999999, 0.9311709601873537], "iscrowd": 0, "area": 14318.195150000001, "rle": {"size": [427, 640], "counts": "j_a1?l<5J5L5K5K5K5J5LQ1oN0000001N100000000O2O000000001N100000001N100000001O0O1000001O0O100000001O0O11O00010O00000010O000001O01O01O0000010O00001O0001O01O0000010O00001O01O0001O00010O0000001O01O0001O00010O00000010O0001O0001O01O0000010O00001O00O1O00O2O0000001N1000001O0O10001O000O101O00000O2O0000001N1000001O0O10001O0O10001O000O101O00002M3N1UOREXOP;a0cEQO_:i0U1I6K6IUSb4"}, "label": "chair to the left of the table with pink flowers on it"}]}
{"id": 542112, "image": "COCO_train2014_000000542112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant was eatting grass\"."}], "task": "refering", "answer_template": "The \"elephant was eatting grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 13, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 355, 356, 357, 358], "bbox": [0.015171875000000001, 0.017979166666666668, 0.8662968750000001, 0.9236041666666667], "iscrowd": 0, "area": 156990.03980000006, "rle": {"size": [480, 640], "counts": "bj45i>8H7I6I6K5K4L5K5K4L5K4L4L5K4L5K4L5K5K5K4L5K5K4L5K5K4L3M3N2M2N3M3M2N3M3M2N3M3M3L5M2O1N2N2N3N1N2N2N101N1O2O0O1O2N101N1O101N1O2O0O101N101N1O101N2O1N101N2N2O0O2O1N2O0O2O1N2N101N2O001N2O0O2O1O1N101N2O1O0O2O1N2O1O0O2O1N2O0[JSHV4o7eKTH\\4m7_KWH`4k7[KXHf4h7VK]Hh4e7SK`Hl4a7oJdHo4^7lJgHS5Z7hJkHV5a8O1N2O1N2O1N101N2O1N2O1N2O1O1O1N2O00O1O10O0100O1O100O010O1O100O1O010O100O1O01000001O001N101O001O0O2O001O001O001O001O013L3M4L4L4L3N3L2NO100O101N100O100O2N100O100O101N100O100O2N100O100O2O0O100O101N1O100O101O01O001O01O01O001O010O001O0010O0001O0010O01O001O010OO2O001O0O2O001O0O2O001O1N101O001N101O0O2O001O1N101O001N10XLQF_2n9aMUF]2j9dMXF[2f9fM\\FX2b9jM_FU2`9kMcFS2\\9nMfFP2Y9QNiFn1T9TNmFk1f8fLTG_19i1`8fNbGX1\\8jNeGU1X8mNkGQ1R8ROPHm0m7UOUHi0h7ZOYHe0d7]O^Hb0`7@aH?\\7DeH<Y7EhH:X7EjH:V7FkH8V7HjH8V7HkH7U7HmH6T7JmH5S7KnH4R7KoH4R7LoH3Q7MPI1Q7OPI0P7ORI0n60RIOo61RINn62SILn63TILl64UIKl64TIKm64TILl64TILm63SILn63SIMn61SINo60RI0n60RI0o6NRI1P7MQI3o6MQI2Q7LPI4Q7JPI6P7JPI5R7IoH7Q7HPI8Q7GoH8S7FnH:R7EoH;R7DnH<S7BnH>R7AoH>S7@nH`0S7_OmHa0T7]OmHc0S7\\OnHd0S7ZOnHf0S7VOPIj0Q7SOQIl0Q7PORIP1o6lNTIT1l6iNWIW1j6fNXIZ1i6bNYI_1k6ZNXIe1m6TNVIl1o90000000000O100000O1O10O01O100O1O100O1O101N1O100O2N100O1O101N1O100O2N100O1O101N1O101N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N1O2N1O2N2M2O2M3M2N3M3M2O2M3M3M4L3N2M4L3M2N2N101N2N1O2N101O010O010O10O0100O2O0O101N10000O2O0O101N10000O2O000O101O0O101N10000O2O000O2O000O101O0O2O001N101N101O0O2O001N101O0O2O001N101O0O2O0O2N101N1O2O0O2O1N4L3N2M3M3N3K4I7I7H8I8L3N3M3L3N3M3K4M4K5K4M4K4L5D=D;E<C=E:F;E:E<E;E:F;EbTX1"}, "label": "elephant was eatting grass"}]}
{"id": 542112, "image": "COCO_train2014_000000542112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large elephant\"."}], "task": "refering", "answer_template": "The \"large elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 13, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 355, 356, 357, 358], "bbox": [0.015171875000000001, 0.017979166666666668, 0.8662968750000001, 0.9236041666666667], "iscrowd": 0, "area": 156990.03980000006, "rle": {"size": [480, 640], "counts": "bj45i>8H7I6I6K5K4L5K5K4L5K4L4L5K4L5K4L5K5K5K4L5K5K4L5K5K4L3M3N2M2N3M3M2N3M3M2N3M3M3L5M2O1N2N2N3N1N2N2N101N1O2O0O1O2N101N1O101N1O2O0O101N101N1O101N2O1N101N2N2O0O2O1N2O0O2O1N2N101N2O001N2O0O2O1O1N101N2O1O0O2O1N2O1O0O2O1N2O0[JSHV4o7eKTH\\4m7_KWH`4k7[KXHf4h7VK]Hh4e7SK`Hl4a7oJdHo4^7lJgHS5Z7hJkHV5a8O1N2O1N2O1N101N2O1N2O1N2O1O1O1N2O00O1O10O0100O1O100O010O1O100O1O010O100O1O01000001O001N101O001O0O2O001O001O001O001O013L3M4L4L4L3N3L2NO100O101N100O100O2N100O100O101N100O100O2N100O100O2O0O100O101N1O100O101O01O001O01O01O001O010O001O0010O0001O0010O01O001O010OO2O001O0O2O001O0O2O001O1N101O001N101O0O2O001O1N101O001N10XLQF_2n9aMUF]2j9dMXF[2f9fM\\FX2b9jM_FU2`9kMcFS2\\9nMfFP2Y9QNiFn1T9TNmFk1f8fLTG_19i1`8fNbGX1\\8jNeGU1X8mNkGQ1R8ROPHm0m7UOUHi0h7ZOYHe0d7]O^Hb0`7@aH?\\7DeH<Y7EhH:X7EjH:V7FkH8V7HjH8V7HkH7U7HmH6T7JmH5S7KnH4R7KoH4R7LoH3Q7MPI1Q7OPI0P7ORI0n60RIOo61RINn62SILn63TILl64UIKl64TIKm64TILl64TILm63SILn63SIMn61SINo60RI0n60RI0o6NRI1P7MQI3o6MQI2Q7LPI4Q7JPI6P7JPI5R7IoH7Q7HPI8Q7GoH8S7FnH:R7EoH;R7DnH<S7BnH>R7AoH>S7@nH`0S7_OmHa0T7]OmHc0S7\\OnHd0S7ZOnHf0S7VOPIj0Q7SOQIl0Q7PORIP1o6lNTIT1l6iNWIW1j6fNXIZ1i6bNYI_1k6ZNXIe1m6TNVIl1o90000000000O100000O1O10O01O100O1O100O1O101N1O100O2N100O1O101N1O100O2N100O1O101N1O101N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N1O2N1O2N2M2O2M3M2N3M3M2O2M3M3M4L3N2M4L3M2N2N101N2N1O2N101O010O010O10O0100O2O0O101N10000O2O0O101N10000O2O000O101O0O101N10000O2O000O2O000O101O0O2O001N101N101O0O2O001N101O0O2O001N101O0O2O0O2N101N1O2O0O2O1N4L3N2M3M3N3K4I7I7H8I8L3N3M3L3N3M3K4M4K5K4M4K4L5D=D;E<C=E:F;E:E<E;E:F;EbTX1"}, "label": "large elephant"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle being held by a man in a blue shirt with a yellow collar and light shorts\"."}], "task": "refering", "answer_template": "The \"a bicycle being held by a man in a blue shirt with a yellow collar and light shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240, 241, 263, 264, 286, 287], "bbox": [0.370171875, 0.37254166666666666, 0.5278906250000001, 0.7630833333333332], "iscrowd": 0, "area": 9742.244700000001, "rle": {"size": [480, 640], "counts": "PY_35k>1N10000O101O000000001O0mBJ\\;6\\D5POJm;1kD`0TO@Q<0iDe0RO]OT<OgDU1AaNX;:UEX1AdNU;4XE[1BaNV;4VE]1C`NW;3VE^1A`NY;4TEn1l:UNQET2f:nMXES2\\:cMVE=;Q2^:dMUE=;P2_:^N_Ea1a:aN]E`1c:bNZE_1e:cNYE^1f:dNXE\\1h:^10O1N3N2N3M4K4M2N1M38H?A2N2M3N2N2N2N2N2M3N2N2N2NZKXFV4f9jKaFQ4\\9PLkFk3S9ULoFj3P9VLQGk3m8TLUGl3j8TLWGm3g8SLZGm3e8SL\\Gn3b8RL_Go3_8QLbGn3^8QLfGl3Z8TLjGh3V8XLnGd3R8\\LRH`3n7`LTH^3l7bLVH[3k7dLXHZ3h7gLXHX3h7hLZHV3f7kLZHT3f7lL\\HR3d7oL\\HP3d7PM^Hn2c7QM_Hm2b7RM`Hl2a7SMaHj2`7VMbHh2_7WMcHg2]7YMeHb2^7^MeH]2]7cMeHY2]7gMeHT2^7lMdHP2^7PNdHl1_7RNdHi1_7WNcHe1_7[NdH_1_7aNcH[1_7eNcHW1_7iN]4O0000000MR\\]4"}, "label": "a bicycle being held by a man in a blue shirt with a yellow collar and light shorts"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike belonging to man in black jacket\"."}], "task": "refering", "answer_template": "The \"bike belonging to man in black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 154, 155, 156, 157, 158, 178, 179, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 315, 316], "bbox": [0.6866875, 0.32599999999999996, 0.8867499999999999, 0.8086041666666667], "iscrowd": 0, "area": 9572.640049999998, "rle": {"size": [480, 640], "counts": "Tn]61n>1O2O1N2N2N1O2N2O1N2mDAd7>\\HDb7=]HDb7<]HFb7;]HGb78]HJb76^HKa76]HLb74^HMa73^HOb70]H2b7O\\H3c7M\\Hb0V7^OiHf0T7ZOkHi0T7WOjHl0T7TOkHP1R7POmHS1Q7mNnHV1nM[OW7_O`Jb1mM\\O`7RO_Ji1gM\\Oj7kN[Jg3e5YLPJR4P6oKkIU4V6jKfIZ4Z6fKbI^4^6cK]Ia4c6_K[Ib4g6]KXId4h6\\KWIe4i6[KVIf4j6ZKUIg4k6YKTIh4m6XKQIi4o6WKQIi4o6WKPIj4P7VKPIj4Q7UKoHi4S7WKlHg4W7YKhHd4\\7\\KbHc4a7^K]H_4h7`KWH]4m7cKQH\\4R8dKjG\\4Z8dKaG^4b8bKZG^4k8g04M3L4M4K4L4L4K5K6J5K5K5L5J5K5I7hM`Di0g;TO^Dd0h;ZO\\D=k;AXD7o;GUD0R<MRDLT<2PDEW<9lC_O[<?T1N2O20O0000000000000000001O000000000001O0001O0000000000000001O01O0001O000O1O2M2O1N2O1N3N1N2OaQR1"}, "label": "bike belonging to man in black jacket"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike ridden by a man wearing all black and holding his hand out\"."}], "task": "refering", "answer_template": "The \"a bike ridden by a man wearing all black and holding his hand out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 154, 155, 156, 157, 158, 178, 179, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 315, 316], "bbox": [0.6866875, 0.32599999999999996, 0.8867499999999999, 0.8086041666666667], "iscrowd": 0, "area": 9572.640049999998, "rle": {"size": [480, 640], "counts": "Tn]61n>1O2O1N2N2N1O2N2O1N2mDAd7>\\HDb7=]HDb7<]HFb7;]HGb78]HJb76^HKa76]HLb74^HMa73^HOb70]H2b7O\\H3c7M\\Hb0V7^OiHf0T7ZOkHi0T7WOjHl0T7TOkHP1R7POmHS1Q7mNnHV1nM[OW7_O`Jb1mM\\O`7RO_Ji1gM\\Oj7kN[Jg3e5YLPJR4P6oKkIU4V6jKfIZ4Z6fKbI^4^6cK]Ia4c6_K[Ib4g6]KXId4h6\\KWIe4i6[KVIf4j6ZKUIg4k6YKTIh4m6XKQIi4o6WKQIi4o6WKPIj4P7VKPIj4Q7UKoHi4S7WKlHg4W7YKhHd4\\7\\KbHc4a7^K]H_4h7`KWH]4m7cKQH\\4R8dKjG\\4Z8dKaG^4b8bKZG^4k8g04M3L4M4K4L4L4K5K6J5K5K5L5J5K5I7hM`Di0g;TO^Dd0h;ZO\\D=k;AXD7o;GUD0R<MRDLT<2PDEW<9lC_O[<?T1N2O20O0000000000000000001O000000000001O0001O0000000000000001O01O0001O000O1O2M2O1N2O1N3N1N2OaQR1"}, "label": "a bike ridden by a man wearing all black and holding his hand out"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt and shorts leaning against a bike\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt and shorts leaning against a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 282, 283], "bbox": [0.22415625, 0.10562500000000001, 0.42134375, 0.7191041666666667], "iscrowd": 0, "area": 18740.686349999996, "rle": {"size": [480, 640], "counts": "RVS21h>;D<E;E;G9J6I7mFnMm4X2oJjMm4\\2iHcMd03_6`2cHeMk0N_6c2\\HfMS1I_6e2THjMZ1D`6f2mGmMa1_O`6o3]IULa6l3]IVLa6l3\\IWLb6k3\\IWLb6k3\\IWLb6m3YIVLe6m3XIULg6n3TIULj6n3SIULj6o3QITLm6o3PISLo6P4mHRLQ7R4jHQLU7Q4hHQLW7R4dHQLZ7R4cHQL[7k5N2O0O1oN_HcJ`7W5mHcJS7Z5UI`Jl6`5YI[Jg6e5^IUJb6l5Y1O1000O10000000000fNSJaIl5_6WJ_Ii5`6ZJ^If5a6]J]Ic5b6`J\\I`5c6cJZI^5e6eJYI[5f6hJXIX5g6kJWIU5h6nJVIQ5j6RKTIn4k6UKSIk4l6YKQIg4n6\\KPId4o6_KoHa4P7bKnH^4Q7eKmH\\4Q7gKmHZ4Q7iKmHY4P7m1N2N2N3M2N2N2N2N3kIaGd5a8XJdGe5^8VJgG9Fl4Y9mJnF`4S:]Oc0]Oc0\\Od0\\Od0[Oe0N20000001O0000000010O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1N2O1O1N2N2N1N3N2N2M5L3M3M^g]5"}, "label": "a man in a blue shirt and shorts leaning against a bike"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person to the left of the bicycle that he is touching\"."}], "task": "refering", "answer_template": "The \"the person to the left of the bicycle that he is touching\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 282, 283], "bbox": [0.22415625, 0.10562500000000001, 0.42134375, 0.7191041666666667], "iscrowd": 0, "area": 18740.686349999996, "rle": {"size": [480, 640], "counts": "RVS21h>;D<E;E;G9J6I7mFnMm4X2oJjMm4\\2iHcMd03_6`2cHeMk0N_6c2\\HfMS1I_6e2THjMZ1D`6f2mGmMa1_O`6o3]IULa6l3]IVLa6l3\\IWLb6k3\\IWLb6k3\\IWLb6m3YIVLe6m3XIULg6n3TIULj6n3SIULj6o3QITLm6o3PISLo6P4mHRLQ7R4jHQLU7Q4hHQLW7R4dHQLZ7R4cHQL[7k5N2O0O1oN_HcJ`7W5mHcJS7Z5UI`Jl6`5YI[Jg6e5^IUJb6l5Y1O1000O10000000000fNSJaIl5_6WJ_Ii5`6ZJ^If5a6]J]Ic5b6`J\\I`5c6cJZI^5e6eJYI[5f6hJXIX5g6kJWIU5h6nJVIQ5j6RKTIn4k6UKSIk4l6YKQIg4n6\\KPId4o6_KoHa4P7bKnH^4Q7eKmH\\4Q7gKmHZ4Q7iKmHY4P7m1N2N2N3M2N2N2N2N3kIaGd5a8XJdGe5^8VJgG9Fl4Y9mJnF`4S:]Oc0]Oc0\\Od0\\Od0[Oe0N20000001O0000000010O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1N2O1O1N2N2N1N3N2N2M5L3M3M^g]5"}, "label": "the person to the left of the bicycle that he is touching"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black jacket\"."}], "task": "refering", "answer_template": "The \"man in black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 41, 42, 43, 64, 65, 66, 67, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 179, 180, 181, 202, 203, 204, 225, 226, 227, 248, 249, 250, 271, 272, 273, 295, 296], "bbox": [0.62453125, 0.018479166666666665, 0.939015625, 0.7367916666666667], "iscrowd": 0, "area": 23941.288250000012, "rle": {"size": [480, 640], "counts": "[ck53j>5N2N5K1O1O001O1O00000FASB?m=DPB<P>:00000O2O001O1O1O1O1N2O1O1O0000O01001O001N101O000O2O001O000O2O001O0O1000000O1000000O1000000O1000000O1000000O1000000O10O1000O01000O01000O010000O01000O01UCPOf;o0RDDAB`;j0jD<U;EjD<V;CjD>V;BiD?W;AhD`0W;@hDb0X;^OgDc0Y;]OfDe0X;[OhDf0P4\\Ne2n0ZIg0o3]Nf2m0ZIg0m3^Nj2j0YIj0j3^Nl2i0YIk0g3_No2i0WIj0g3_NR3i0TIj0f3_NV3j0QIi0f3_NY3i0oHj0d3`N\\3o2eI]LT2f0W4R3[I^LR2k0b4m2RI^Lm1R1Q5R3hJRMZ5o2`JTMa5m2ZJUMi5l2QJWMP6k2iIYMY6e51O1O2N1O2N1O2N1O2MUO\\I_Ia6a6lIUIR6k6P10QOcI^I\\6d6n01N2002N2N1O2N2N`JWIm2[8]KSHX4R8`KVH\\4n7TKaHi4o8L1N101O1N101O0O2O1O0O2O001N2O00cNlKnGS4P8TLjGn3T8YLfGh3W8_LdGb3i6hKkIm0XO]3j6lKaIS1AR3m6QLTIY1Kh2n6WLhH^14]2j6iNRIX1Z6R4@?G:N3M2mK]Ff2f9TM_Fj2d9PMaFn2b9mLbFQ3b9jLaFT3d9eL`FY3e9aL^F^3c:O1N2O1O1N3N2N1N3N2N2M3N2N2M3N2N2M4M2N2M4M4BoUb0"}, "label": "man in black jacket"}]}
{"id": 427435, "image": "COCO_train2014_000000427435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black raincoat and helmet , standing to the right of his bicycle\"."}], "task": "refering", "answer_template": "The \"a man in a black raincoat and helmet , standing to the right of his bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 41, 42, 43, 64, 65, 66, 67, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 179, 180, 181, 202, 203, 204, 225, 226, 227, 248, 249, 250, 271, 272, 273, 295, 296], "bbox": [0.62453125, 0.018479166666666665, 0.939015625, 0.7367916666666667], "iscrowd": 0, "area": 23941.288250000012, "rle": {"size": [480, 640], "counts": "[ck53j>5N2N5K1O1O001O1O00000FASB?m=DPB<P>:00000O2O001O1O1O1O1N2O1O1O0000O01001O001N101O000O2O001O000O2O001O0O1000000O1000000O1000000O1000000O1000000O1000000O10O1000O01000O01000O010000O01000O01UCPOf;o0RDDAB`;j0jD<U;EjD<V;CjD>V;BiD?W;AhD`0W;@hDb0X;^OgDc0Y;]OfDe0X;[OhDf0P4\\Ne2n0ZIg0o3]Nf2m0ZIg0m3^Nj2j0YIj0j3^Nl2i0YIk0g3_No2i0WIj0g3_NR3i0TIj0f3_NV3j0QIi0f3_NY3i0oHj0d3`N\\3o2eI]LT2f0W4R3[I^LR2k0b4m2RI^Lm1R1Q5R3hJRMZ5o2`JTMa5m2ZJUMi5l2QJWMP6k2iIYMY6e51O1O2N1O2N1O2N1O2MUO\\I_Ia6a6lIUIR6k6P10QOcI^I\\6d6n01N2002N2N1O2N2N`JWIm2[8]KSHX4R8`KVH\\4n7TKaHi4o8L1N101O1N101O0O2O1O0O2O001N2O00cNlKnGS4P8TLjGn3T8YLfGh3W8_LdGb3i6hKkIm0XO]3j6lKaIS1AR3m6QLTIY1Kh2n6WLhH^14]2j6iNRIX1Z6R4@?G:N3M2mK]Ff2f9TM_Fj2d9PMaFn2b9mLbFQ3b9jLaFT3d9eL`FY3e9aL^F^3c:O1N2O1O1N3N2N1N3N2N2M3N2N2M3N2N2M4M2N2M4M4BoUb0"}, "label": "a man in a black raincoat and helmet , standing to the right of his bicycle"}]}
{"id": 173484, "image": "COCO_train2014_000000173484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog on right wearing green bow tie and hat\"."}], "task": "refering", "answer_template": "The \"dog on right wearing green bow tie and hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 250, 251, 271, 272, 273, 274, 294], "bbox": [0.697265625, 0.09395238095238095, 0.985703125, 0.8076666666666666], "iscrowd": 0, "area": 37223.340750000025, "rle": {"size": [420, 640], "counts": "kVg57]<a0J5K6J6I7J5K6J6J5_Oa0^Ob0WNSMiH^3V7Y1N1O2N2N101N1O2M3K4L^OgHcKT7b4VISKf6Q5dIfJW6_5=:O100O1O1O1O001O1O1O100O1O001O1O4L4L4L4M3L4L4L4M3L4M3L4M3L3N2N2M3N2M3N2M3N2M3N2N2M3N0O100O10cMeKQL[4o3lKjKS4X4RLcKm3]4ZL\\Ke3f4`LTK`3l4fLnJY3S5nLfJR3[5SM_Jl2b5VM\\Jj2e5UMZJl2f5UMYJj2i5VMVJj2j5WMTJi2n5WMQJi2o5XMPJg2R6YMlIh2T6YMkIf2W6ZMhIf2X6[MfIf2Y6\\MfIc2[6^MdIb2[6`McI`2^6cM_I]2a6fM\\IY2d6kMXIV2h6R2010O10O010O10O10O010001O0gJUIX4l6fKWIY4i6eKZIZ4f6dK]I[4c6cK`I\\4`6cKbI\\4^6bKeI]4\\6`KhI]4Y6aKjI^4V6`KmI_4S6`KoI_4Q6_KRJ`4n5_KTJ`4l5^KWJa4i5^KXJb4h5\\KZJc4h5[KYJe4T7O00000000000O20O001O001O1O010O001O1L3K6J5I8I6J7I7I6I8J5M4L4M2M4L3M4L3ZKdHa3`7^LcH]3`7bLdHW3_7iLeHP3_7nLfHk2]7UMfHe2^7ZMfH_2U9J6K4L5K5L4K5K4L5I7F:G8G:F:F:GQa3"}, "label": "dog on right wearing green bow tie and hat"}]}
{"id": 173484, "image": "COCO_train2014_000000173484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pug to the right of another pug looking down\"."}], "task": "refering", "answer_template": "The \"a pug to the right of another pug looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 250, 251, 271, 272, 273, 274, 294], "bbox": [0.697265625, 0.09395238095238095, 0.985703125, 0.8076666666666666], "iscrowd": 0, "area": 37223.340750000025, "rle": {"size": [420, 640], "counts": "kVg57]<a0J5K6J6I7J5K6J6J5_Oa0^Ob0WNSMiH^3V7Y1N1O2N2N101N1O2M3K4L^OgHcKT7b4VISKf6Q5dIfJW6_5=:O100O1O1O1O001O1O1O100O1O001O1O4L4L4L4M3L4L4L4M3L4M3L4M3L3N2N2M3N2M3N2M3N2M3N2N2M3N0O100O10cMeKQL[4o3lKjKS4X4RLcKm3]4ZL\\Ke3f4`LTK`3l4fLnJY3S5nLfJR3[5SM_Jl2b5VM\\Jj2e5UMZJl2f5UMYJj2i5VMVJj2j5WMTJi2n5WMQJi2o5XMPJg2R6YMlIh2T6YMkIf2W6ZMhIf2X6[MfIf2Y6\\MfIc2[6^MdIb2[6`McI`2^6cM_I]2a6fM\\IY2d6kMXIV2h6R2010O10O010O10O10O010001O0gJUIX4l6fKWIY4i6eKZIZ4f6dK]I[4c6cK`I\\4`6cKbI\\4^6bKeI]4\\6`KhI]4Y6aKjI^4V6`KmI_4S6`KoI_4Q6_KRJ`4n5_KTJ`4l5^KWJa4i5^KXJb4h5\\KZJc4h5[KYJe4T7O00000000000O20O001O001O1O010O001O1L3K6J5I8I6J7I7I6I8J5M4L4M2M4L3M4L3ZKdHa3`7^LcH]3`7bLdHW3_7iLeHP3_7nLfHk2]7UMfHe2^7ZMfH_2U9J6K4L5K5L4K5K4L5I7F:G8G:F:F:GQa3"}, "label": "a pug to the right of another pug looking down"}]}
{"id": 312748, "image": "COCO_train2014_000000312748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and blue old bus parked in a parking lot\"."}], "task": "refering", "answer_template": "The \"a red and blue old bus parked in a parking lot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 178, 179, 196, 197, 214, 215, 232, 233, 250, 251, 268, 269], "bbox": [0.865931863727455, 0.28887999999999997, 1.0, 0.81976], "iscrowd": 0, "area": 12857.183699999996, "rle": {"size": [500, 499], "counts": "cRc63a?;E;E;E;E;E;E;E;E;E;E:F;D<E;E;E;E;E;bGiIi6W6UIlIj6S6VIoIi6P6VIRJj6m5UIVJj6f5XI]Jg6`5[IaJe6]5[IfJd6X5\\IjJd6S5^IoJa6o4_ISKa6n4^ISKa6m4_ISKa6n4^ISKa6m4_ISKa6m4_ISKa6n4^ISKa6m4_ISKa6n4]ISKc6m4]ITKb6m4]ISKb6n4^ISKa6m4_ISKa6n4^IRKb6n4^ISKa6n4^IRKb6n4^IRKb6n4]ITKb6m4]ISKc6m4]ITKb6m4]ISKc6m4]ISKc6n4[ITKd6l4[IUKe6k4YIWKg6j4WIXKh6h4WIYKi6h4UIYKk6g4TI[Kk6e4TI\\Kl6i4nHYKQ7e60001O000O2O001_K"}, "label": "a red and blue old bus parked in a parking lot"}]}
{"id": 312748, "image": "COCO_train2014_000000312748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bus with red trim around it ' s windows\"."}], "task": "refering", "answer_template": "The \"a blue bus with red trim around it ' s windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 178, 179, 196, 197, 214, 215, 232, 233, 250, 251, 268, 269], "bbox": [0.865931863727455, 0.28887999999999997, 1.0, 0.81976], "iscrowd": 0, "area": 12857.183699999996, "rle": {"size": [500, 499], "counts": "cRc63a?;E;E;E;E;E;E;E;E;E;E:F;D<E;E;E;E;E;bGiIi6W6UIlIj6S6VIoIi6P6VIRJj6m5UIVJj6f5XI]Jg6`5[IaJe6]5[IfJd6X5\\IjJd6S5^IoJa6o4_ISKa6n4^ISKa6m4_ISKa6n4^ISKa6m4_ISKa6m4_ISKa6n4^ISKa6m4_ISKa6n4]ISKc6m4]ITKb6m4]ISKb6n4^ISKa6m4_ISKa6n4^IRKb6n4^ISKa6n4^IRKb6n4^IRKb6n4]ITKb6m4]ISKc6m4]ITKb6m4]ISKc6m4]ISKc6n4[ITKd6l4[IUKe6k4YIWKg6j4WIXKh6h4WIYKi6h4UIYKk6g4TI[Kk6e4TI\\Kl6i4nHYKQ7e60001O000O2O001_K"}, "label": "a blue bus with red trim around it ' s windows"}]}
{"id": 312748, "image": "COCO_train2014_000000312748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an olive green vehicle\"."}], "task": "refering", "answer_template": "The \"an olive green vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 41, 42, 43, 44, 45, 46, 47, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 294, 295, 296, 297, 298, 299, 300, 313, 314, 315, 316], "bbox": [0.04953907815631262, 0.07415999999999999, 0.9907414829659319, 0.97752], "iscrowd": 0, "area": 138946.69475000005, "rle": {"size": [500, 499], "counts": "S_<7T?;E:E;F:F:F;D;F:F:F;I6N2N2N3M2N2N2N2N3M2N2N2N3M2O1N2N2N3M2N2N3M2N2N3M2N3M2N2N3N1N3M2N2N3M2N3M2N2N3M2N3M2N2N3N1N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2O1N3M2N3M2N2N3M2N3M2N2N3M2N3M2O1N3M2N3M6J8H7I7I7I6J1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O101N100O100O100O100O100O2O1N2O1O2M2O1N3N1N2O2M2O1N2O2M2O1N3N1N2O1N3N1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O1N3N1N2O2M100O100O100O100O100O100O100O10000O100O100O1001O00001O001O00001O001O01O01O001O00001O001O00001N101O0O101O0O2O0O101O0O2O000O2O001N100O2O001N10001N101O0O101N101O0O101O0O2O000O2O0O2O000O2O001N10001N101N10001N101O0O101O0O2O0O101O0O2O000O2O001N100O2O001N10001N101O0O101N101O0O101O0O2O000O2O0O2O000O2O001N10001N101N10001N1O2N1O1O2N1O1O2N1O2N1O1N3N1O2N1O1O2N1O2N1O1O2N1O2M2O1O2N1N3M2N2M4M2N3L3N2N3L3N3M2M3N3M201N10001N101O0O101N101O0O101O0O2O000O2O0O2O8H<C=D;E<C=D<D;D=D<C=D<D;D=D<D<C=D;D=D<D5KN1O2N2N2N2M3N2N2N2N2N21O001N10001O00001O00001O001O00001N10001O00001O001O00001O000O2O001O0M3J7IaS2"}, "label": "an olive green vehicle"}]}
{"id": 312748, "image": "COCO_train2014_000000312748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray - green bus\"."}], "task": "refering", "answer_template": "The \"the gray - green bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 41, 42, 43, 44, 45, 46, 47, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 294, 295, 296, 297, 298, 299, 300, 313, 314, 315, 316], "bbox": [0.04953907815631262, 0.07415999999999999, 0.9907414829659319, 0.97752], "iscrowd": 0, "area": 138946.69475000005, "rle": {"size": [500, 499], "counts": "S_<7T?;E:E;F:F:F;D;F:F:F;I6N2N2N3M2N2N2N2N3M2N2N2N3M2O1N2N2N3M2N2N3M2N2N3M2N3M2N2N3N1N3M2N2N3M2N3M2N2N3M2N3M2N2N3N1N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2O1N3M2N3M2N2N3M2N3M2N2N3M2N3M2O1N3M2N3M6J8H7I7I7I6J1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O101N100O100O100O100O100O2O1N2O1O2M2O1N3N1N2O2M2O1N2O2M2O1N3N1N2O1N3N1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O1N3N1N2O2M100O100O100O100O100O100O100O10000O100O100O1001O00001O001O00001O001O01O01O001O00001O001O00001N101O0O101O0O2O0O101O0O2O000O2O001N100O2O001N10001N101O0O101N101O0O101O0O2O000O2O0O2O000O2O001N10001N101N10001N101O0O101O0O2O0O101O0O2O000O2O001N100O2O001N10001N101O0O101N101O0O101O0O2O000O2O0O2O000O2O001N10001N101N10001N1O2N1O1O2N1O1O2N1O2N1O1N3N1O2N1O1O2N1O2N1O1O2N1O2M2O1O2N1N3M2N2M4M2N3L3N2N3L3N3M2M3N3M201N10001N101O0O101N101O0O101O0O2O000O2O0O2O8H<C=D;E<C=D<D;D=D<C=D<D;D=D<D<C=D;D=D<D5KN1O2N2N2N2M3N2N2N2N2N21O001N10001O00001O00001O001O00001N10001O00001O001O00001O000O2O001O0M3J7IaS2"}, "label": "the gray - green bus"}]}
{"id": 452014, "image": "COCO_train2014_000000452014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red apple under 2 bananas with only red showing\"."}], "task": "refering", "answer_template": "The \"a red apple under 2 bananas with only red showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 278, 279, 280, 296, 297, 298, 299, 300, 301, 302, 303, 317, 318, 319, 320, 321, 322, 323, 340, 341, 342, 343], "bbox": [0.4462745098039216, 0.5375, 0.7889869281045752, 0.7131862745098039], "iscrowd": 0, "area": 11435.142250000003, "rle": {"size": [612, 612], "counts": "caS52lb07I7J6M2O2N2N2M3N2N2M3N2N2M3N2O1O2N1O2O1N1O2N2N1O2N2N1O2N0000000001O0O100000000000000O100000001O0O1000000O0100000O100000O01000000O1000O01000000O1000O10O1000000O10O1000O1000000O0100000O10000O10O1000O1000000O0100000O100000O0100000O010O1000O0100O01000O010O0100O010O01000O010O010O10O010O10O010O10O010O10O10O010O0100O0100O0100O010O010O10O10O010O10O010O10O10O010O0100O0100O0100O010O010O3N3M2M3N3L3N3L3N3L3N3M2M3N3L3N3L3N3M2M^g\\2"}, "label": "a red apple under 2 bananas with only red showing"}]}
{"id": 452014, "image": "COCO_train2014_000000452014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fruit between the bananas that is red\"."}], "task": "refering", "answer_template": "The \"the fruit between the bananas that is red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 278, 279, 280, 296, 297, 298, 299, 300, 301, 302, 303, 317, 318, 319, 320, 321, 322, 323, 340, 341, 342, 343], "bbox": [0.4462745098039216, 0.5375, 0.7889869281045752, 0.7131862745098039], "iscrowd": 0, "area": 11435.142250000003, "rle": {"size": [612, 612], "counts": "caS52lb07I7J6M2O2N2N2M3N2N2M3N2N2M3N2O1O2N1O2O1N1O2N2N1O2N2N1O2N0000000001O0O100000000000000O100000001O0O1000000O0100000O100000O01000000O1000O01000000O1000O10O1000000O10O1000O1000000O0100000O10000O10O1000O1000000O0100000O100000O0100000O010O1000O0100O01000O010O0100O010O01000O010O010O10O010O10O010O10O010O10O10O010O0100O0100O0100O010O010O10O10O010O10O010O10O10O010O0100O0100O0100O010O010O3N3M2M3N3L3N3L3N3L3N3M2M3N3L3N3L3N3M2M^g\\2"}, "label": "the fruit between the bananas that is red"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blond dog with a black snout and black fringed ears lieing with its tail near white chair\"."}], "task": "refering", "answer_template": "The \"a blond dog with a black snout and black fringed ears lieing with its tail near white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 330, 331, 335], "bbox": [0.32596875000000003, 0.6311737089201878, 0.635640625, 0.9832394366197184], "iscrowd": 0, "area": 12926.880299999997, "rle": {"size": [426, 640], "counts": "mWg21Y=1N1O1O1XCOV<2gC4a1Hg85hE3_1Lf83iE2`1Me81kE3^10d8NnE2\\12f8LmE2^12e8LmE2^12e8LmE2]13e8LmE2^12e8KnE3]12e8KnE3]12e8KmE3^13e8JmE3^13d8JoE2^14c8JoE2^14c8JnE2_14d8InE2_14c8KnE1_13d8LlE1`14d8JmE2_13d8LmE0_14kNFU96`F0a13jNLQ90eF1_14jNMQ9NfF0_15iN0P9LgF0`13iN2P9KgFO`14iN4o8IgFOa15hN4o8IhFNa15gN4Q9IfFOa14iN3Q9JeFOa14hN4R9IeFO`14iN4R9JdFO`14jN2S9KcFO`13jN3T9KbF0^13lN1U9L`F1_11lN2V9L_F1^12lN1W9M_F0^12lN0X9N^F0]12mN0Y9N]F0]11nN0Y9O\\F1[10PO1X9O]F0[1OPO2Y90[FO[1OQO2Z9?eG_OPO3[9>dG_OQO3[9?cG^ORO3\\9?bG]ORO5\\9>aG^OSO3\\9`0aG\\OSO4\\9a0`G\\OTO3[9c0`GYOUO4\\9c0^GZOUO4\\9c0^GYOWO3[9e0^GXOVO3\\9f0]GWOWO4[9f0^GVOWO3\\9g0\\GVOXO3\\9h0[GVOYO2\\9h0[GUOYO3]9h0YGVOYO2^9i0XGUOZO3^9h0XGUOYO3_9i0WGTOZO3`9h0WGUOXO4a9g0VGUOYO4a9h0UGUOYO3c9h0TGUOYO3b9h0UGUOYO3c9h0TGUOXO4c9h0TGTOZO3b9j0TGSOYO4b9j0TGROZO4c9i0SGTOZO3b9j0SGTOZO2c9k0PGVO\\O0d9j0ZFL2YOd9X2[FhMf9X2YFiMg9V2ZFjMf9V2YFjMg9m2000O01000O10000O10000O1000000000001O001O00001O001O001O1O001O001O1O0000002N1O2N1O1O1O2N1O1O1O2N2N2N2N1O2N3N1O2O0N2O1N2O0O2N101N2010N1O2N000000O1000000O0100O10O001O001O001O001O1N2K5L4K5K5M3N2O1N2N2N2O2M2M2L4M3O2N100O1O100O101N1OabP3"}, "label": "a blond dog with a black snout and black fringed ears lieing with its tail near white chair"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan dog laying down in the middle of the floor to the right of the basket\"."}], "task": "refering", "answer_template": "The \"tan dog laying down in the middle of the floor to the right of the basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 330, 331, 335], "bbox": [0.32596875000000003, 0.6311737089201878, 0.635640625, 0.9832394366197184], "iscrowd": 0, "area": 12926.880299999997, "rle": {"size": [426, 640], "counts": "mWg21Y=1N1O1O1XCOV<2gC4a1Hg85hE3_1Lf83iE2`1Me81kE3^10d8NnE2\\12f8LmE2^12e8LmE2^12e8LmE2]13e8LmE2^12e8KnE3]12e8KnE3]12e8KmE3^13e8JmE3^13d8JoE2^14c8JoE2^14c8JnE2_14d8InE2_14c8KnE1_13d8LlE1`14d8JmE2_13d8LmE0_14kNFU96`F0a13jNLQ90eF1_14jNMQ9NfF0_15iN0P9LgF0`13iN2P9KgFO`14iN4o8IgFOa15hN4o8IhFNa15gN4Q9IfFOa14iN3Q9JeFOa14hN4R9IeFO`14iN4R9JdFO`14jN2S9KcFO`13jN3T9KbF0^13lN1U9L`F1_11lN2V9L_F1^12lN1W9M_F0^12lN0X9N^F0]12mN0Y9N]F0]11nN0Y9O\\F1[10PO1X9O]F0[1OPO2Y90[FO[1OQO2Z9?eG_OPO3[9>dG_OQO3[9?cG^ORO3\\9?bG]ORO5\\9>aG^OSO3\\9`0aG\\OSO4\\9a0`G\\OTO3[9c0`GYOUO4\\9c0^GZOUO4\\9c0^GYOWO3[9e0^GXOVO3\\9f0]GWOWO4[9f0^GVOWO3\\9g0\\GVOXO3\\9h0[GVOYO2\\9h0[GUOYO3]9h0YGVOYO2^9i0XGUOZO3^9h0XGUOYO3_9i0WGTOZO3`9h0WGUOXO4a9g0VGUOYO4a9h0UGUOYO3c9h0TGUOYO3b9h0UGUOYO3c9h0TGUOXO4c9h0TGTOZO3b9j0TGSOYO4b9j0TGROZO4c9i0SGTOZO3b9j0SGTOZO2c9k0PGVO\\O0d9j0ZFL2YOd9X2[FhMf9X2YFiMg9V2ZFjMf9V2YFjMg9m2000O01000O10000O10000O1000000000001O001O00001O001O001O1O001O001O1O0000002N1O2N1O1O1O2N1O1O1O2N2N2N2N1O2N3N1O2O0N2O1N2O0O2N101N2010N1O2N000000O1000000O0100O10O001O001O001O001O1N2K5L4K5K5M3N2O1N2N2N2O2M2M2L4M3O2N100O1O100O101N1OabP3"}, "label": "tan dog laying down in the middle of the floor to the right of the basket"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dining table\"."}], "task": "refering", "answer_template": "The \"dining table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 201, 202, 203, 223, 224], "bbox": [0.53203125, 0.2278638497652582, 0.8688125, 0.6344600938967135], "iscrowd": 0, "area": 13569.095949999995, "rle": {"size": [426, 640], "counts": "jn]48R=1QDJh:7WELf:5YEMe:4ZELf:5YELf:5ZEJf:7YEIg:8XEIg:8XEHg:9ZEFc:>\\ECa:`0^E@a:b0_E]Oa:d0^E]Ob:c0]E]Oc:d0\\E\\Od:e0\\E[Oc:f0\\EZOd:g0[EYOf:f0ZE[Oe:f0[EYOe:h0ZEXOh:g0WEZOk:d0UE[On:c0QE]OQ;b0nD_OR;`0nD@S;`0lD@T;`0lDAS;`0kDAV;>jDBV;>iDDW;:jDFV;:iDGX;7iDJV;5jDLV;4jDLW;2iD0V;OjD2W;MiD3W;LiD6B[Oc;=kD9C[O`;<lD:D[O_;:mD=C[O^;7oD?C\\O\\;5PE`0E[OZ;4QEc0D[OY;1SEe0D\\OW;NUEg0D\\OV;MUEW1i:hNWEk0E[OT;HWEn0E[Oh:IZEM:_1\\:kNdEV1\\:jNcEX1\\:hNdEX1\\:hNdEX1\\:hNcEZ1\\:fNdEZ1\\:fNdE\\1[:cNdEd1V:\\NjEh1R:XNnEV1c:kN\\En0l:ROTEj0P;VOPEb0W;_OhD:`;F`D2h;NXDJP<6PDGS<;jCFV<g0003MREVOmN2R:d0kF@SOLV:`0aFJYOFZ:<WF4_O@]:9_EI3a01^O\\:f0dEL0^O]:e0cEM0^O]:e0cEM0^O]:e0cEM0^O]:e0cEM0^O]:e0cEM0^O]:e0cEMO_O_:c0bENO@^:b0bEO0_O^:b0bEO0_O^:b0bEO0_O^:b0bEO0_O_:a0aE00_O_:`0bE1O_O_:`0bE1O_O_:`0bE1O@^:?cE1O@_:>bE2O@_:>bE2O@_:>bE2O@_:>bE2O@_:>bE2O@_:>bE2O@`:=aE3OA_:<bE3OA_:<bE3OA_:;cE4NA_:;cE4NA`::bE5NA`::bE5NB^::dE4MD^:8eE4ME\\:8gE3ME\\:8gE3MFZ:8iE2MGY:7jE2MHX:6kE2MIW:5lE2MJV:4mE2MJV:3nE3LKU:2nE4MKT:1oE4MLS:0PF4MMS:NPF5MNR:MQF5MOQ:LRF5MKU:0nE5M]Oc:>`E5Q;KoD5R2ZO\\6a0bG4m1@b6<aG0k1Ie67`GLi12h62_GHh1:j6M_GFd1c0m6G_GBb1l0P7B^G^O`1U1S7]OWIh0j6QOWIT1j6lNPIY1Q7gNiH^1X7bNcHc1]7]N]Hh1d7XNVHm1k7SNoGR2R8oMgGV2Z8jM`G[2a8eMXGb2h8^MRGg2o8i01O1EcFhL^9X3eFdL\\9\\3gFaLY9_3900O2O00001N1O1N3N1O2M20001O00001O00001O001O00001O000N3N1N2O2N1O2O00001N100010O1O100O10O01O100O00100O02O5J7J5J6K5J6K6I6K5K5JfgR1"}, "label": "dining table"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a striped blue , white , and green table cloth\"."}], "task": "refering", "answer_template": "The \"a table with a striped blue , white , and green table cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 201, 202, 203, 223, 224], "bbox": [0.53203125, 0.2278638497652582, 0.8688125, 0.6344600938967135], "iscrowd": 0, "area": 13569.095949999995, "rle": {"size": [426, 640], "counts": "jn]48R=1QDJh:7WELf:5YEMe:4ZELf:5YELf:5ZEJf:7YEIg:8XEIg:8XEHg:9ZEFc:>\\ECa:`0^E@a:b0_E]Oa:d0^E]Ob:c0]E]Oc:d0\\E\\Od:e0\\E[Oc:f0\\EZOd:g0[EYOf:f0ZE[Oe:f0[EYOe:h0ZEXOh:g0WEZOk:d0UE[On:c0QE]OQ;b0nD_OR;`0nD@S;`0lD@T;`0lDAS;`0kDAV;>jDBV;>iDDW;:jDFV;:iDGX;7iDJV;5jDLV;4jDLW;2iD0V;OjD2W;MiD3W;LiD6B[Oc;=kD9C[O`;<lD:D[O_;:mD=C[O^;7oD?C\\O\\;5PE`0E[OZ;4QEc0D[OY;1SEe0D\\OW;NUEg0D\\OV;MUEW1i:hNWEk0E[OT;HWEn0E[Oh:IZEM:_1\\:kNdEV1\\:jNcEX1\\:hNdEX1\\:hNdEX1\\:hNcEZ1\\:fNdEZ1\\:fNdE\\1[:cNdEd1V:\\NjEh1R:XNnEV1c:kN\\En0l:ROTEj0P;VOPEb0W;_OhD:`;F`D2h;NXDJP<6PDGS<;jCFV<g0003MREVOmN2R:d0kF@SOLV:`0aFJYOFZ:<WF4_O@]:9_EI3a01^O\\:f0dEL0^O]:e0cEM0^O]:e0cEM0^O]:e0cEM0^O]:e0cEM0^O]:e0cEM0^O]:e0cEMO_O_:c0bENO@^:b0bEO0_O^:b0bEO0_O^:b0bEO0_O^:b0bEO0_O_:a0aE00_O_:`0bE1O_O_:`0bE1O_O_:`0bE1O@^:?cE1O@_:>bE2O@_:>bE2O@_:>bE2O@_:>bE2O@_:>bE2O@_:>bE2O@`:=aE3OA_:<bE3OA_:<bE3OA_:;cE4NA_:;cE4NA`::bE5NA`::bE5NB^::dE4MD^:8eE4ME\\:8gE3ME\\:8gE3MFZ:8iE2MGY:7jE2MHX:6kE2MIW:5lE2MJV:4mE2MJV:3nE3LKU:2nE4MKT:1oE4MLS:0PF4MMS:NPF5MNR:MQF5MOQ:LRF5MKU:0nE5M]Oc:>`E5Q;KoD5R2ZO\\6a0bG4m1@b6<aG0k1Ie67`GLi12h62_GHh1:j6M_GFd1c0m6G_GBb1l0P7B^G^O`1U1S7]OWIh0j6QOWIT1j6lNPIY1Q7gNiH^1X7bNcHc1]7]N]Hh1d7XNVHm1k7SNoGR2R8oMgGV2Z8jM`G[2a8eMXGb2h8^MRGg2o8i01O1EcFhL^9X3eFdL\\9\\3gFaLY9_3900O2O00001N1O1N3N1O2M20001O00001O00001O001O00001O000N3N1N2O2N1O2O00001N100010O1O100O10O01O100O00100O02O5J7J5J6K5J6K6I6K5K5JfgR1"}, "label": "a table with a striped blue , white , and green table cloth"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty dining chair\"."}], "task": "refering", "answer_template": "The \"empty dining chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 220, 221, 222, 223, 244, 246], "bbox": [0.592484375, 0.3273474178403756, 0.76240625, 0.7389671361502347], "iscrowd": 0, "area": 11128.061949999996, "rle": {"size": [426, 640], "counts": "lmm44j<g0ZO;D=TO>=2M101O1O1O1O3M4L4K4M9G8H5K00000O1000ML1M4K4M3L5K43M3N2M4L7J5N200000001O2N1mLoEk2X:1O00000000000000000bFSM_8m2]GWMc8i2XG\\Mh8d2TGaMk8_2PGfMP9Z2lFkMS9k2lF`LT9a3oF[LQ9f3RGWLm8j392N1O1O1O1O1O1O1O2N1O1O1O1O1aGhKj7Y4kGQLU8c4O00001O00I7J6K5J6J6J6K5J6J6J6K5J6J6K5J6J6J6K1N1O6J5K6I7I7I6J:GWSo1"}, "label": "empty dining chair"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is any empty chair at the table\"."}], "task": "refering", "answer_template": "The \"there is any empty chair at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 220, 221, 222, 223, 244, 246], "bbox": [0.592484375, 0.3273474178403756, 0.76240625, 0.7389671361502347], "iscrowd": 0, "area": 11128.061949999996, "rle": {"size": [426, 640], "counts": "lmm44j<g0ZO;D=TO>=2M101O1O1O1O3M4L4K4M9G8H5K00000O1000ML1M4K4M3L5K43M3N2M4L7J5N200000001O2N1mLoEk2X:1O00000000000000000bFSM_8m2]GWMc8i2XG\\Mh8d2TGaMk8_2PGfMP9Z2lFkMS9k2lF`LT9a3oF[LQ9f3RGWLm8j392N1O1O1O1O1O1O1O2N1O1O1O1O1aGhKj7Y4kGQLU8c4O00001O00I7J6K5J6J6J6K5J6J6J6K5J6J6K5J6J6J6K1N1O6J5K6I7I7I6J:GWSo1"}, "label": "there is any empty chair at the table"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair with a brown haired woman wearing a red shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"a white chair with a brown haired woman wearing a red shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 218], "bbox": [0.432109375, 0.2594835680751174, 0.5797656250000001, 0.6267840375586854], "iscrowd": 0, "area": 4748.758249999998, "rle": {"size": [426, 640], "counts": "Q[c35U=5K6J5K6J5K7K7M3J5HW2iM10O01O01N1\\NWGfNj8[1c1OoEkMm8V2RGkMm8U2SGkMn8U2QGkMo8V2oFjMR9V2nFjMS9V2lFjMk84ZFR2k0jMk8`2TG`Ml8a2RG`Mn8`2RG_Mo8b2PG^MP9b2PG^MV9]2iFcMI?b8g1eGmNb8m0iFG^9o17I6J7I7I7nNgEVNo:Y1[EgNl:h0gDYO=NP;=lDE4NP;=lDD40P;<kDE5OP;<kDE5OP;;lDF4OP;;lDF4NQ;<kDF3OR;;jDF5OP;<kDE5OP;<kDE5OP;;lDF4OP;;lDF4NQ;<kDE40Q;;jDF5OQ;;jDF5OQ;;fDJ9KQ;Z1bDfN^;Z1aDfN`;Z1`DfN_;\\1`DdN`;\\1`DdN`;:]D1f0Em::aEG_:9aEF`::`EF`::`EF`::`EF`:9`EH`:8`EH`:8`EG`::`EF`::`EF`:9`EH`:8`EH_:9aEG_:9aEF`::`EF`::`EF_::aEG_:9aEGPO0V;9mENR:2WEf0g:o0000000O11POP1A>E<K5MSg_3"}, "label": "a white chair with a brown haired woman wearing a red shirt and blue jeans"}]}
{"id": 296385, "image": "COCO_train2014_000000296385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair that the woman with the purple shirt is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair that the woman with the purple shirt is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 218], "bbox": [0.432109375, 0.2594835680751174, 0.5797656250000001, 0.6267840375586854], "iscrowd": 0, "area": 4748.758249999998, "rle": {"size": [426, 640], "counts": "Q[c35U=5K6J5K6J5K7K7M3J5HW2iM10O01O01N1\\NWGfNj8[1c1OoEkMm8V2RGkMm8U2SGkMn8U2QGkMo8V2oFjMR9V2nFjMS9V2lFjMk84ZFR2k0jMk8`2TG`Ml8a2RG`Mn8`2RG_Mo8b2PG^MP9b2PG^MV9]2iFcMI?b8g1eGmNb8m0iFG^9o17I6J7I7I7nNgEVNo:Y1[EgNl:h0gDYO=NP;=lDE4NP;=lDD40P;<kDE5OP;<kDE5OP;;lDF4OP;;lDF4NQ;<kDF3OR;;jDF5OP;<kDE5OP;<kDE5OP;;lDF4OP;;lDF4NQ;<kDE40Q;;jDF5OQ;;jDF5OQ;;fDJ9KQ;Z1bDfN^;Z1aDfN`;Z1`DfN_;\\1`DdN`;\\1`DdN`;:]D1f0Em::aEG_:9aEF`::`EF`::`EF`::`EF`:9`EH`:8`EH`:8`EG`::`EF`::`EF`:9`EH`:8`EH_:9aEG_:9aEF`::`EF`::`EF_::aEG_:9aEGPO0V;9mENR:2WEf0g:o0000000O11POP1A>E<K5MSg_3"}, "label": "the chair that the woman with the purple shirt is sitting in"}]}
{"id": 460228, "image": "COCO_train2014_000000460228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of bread to the right of the cup of sauce\"."}], "task": "refering", "answer_template": "The \"the piece of bread to the right of the cup of sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 318], "bbox": [0.46323437500000003, 0.2426987447698745, 1.0, 0.7752719665271967], "iscrowd": 0, "area": 63711.4687, "rle": {"size": [478, 640], "counts": "d[Z4`0X>7I8H7I8H8H7K6M2N3M3L3N3M2N3M3M2N2M2O1O2N101O0O101N101N100O2O0O101N101N10001N101N100O2O0O101N101N10001O0O101O001O00001O0O2O00001O00001N101O00001O001O0O101O00001O001N10001O00001O001N10001O001O001O1O001O1O0O2O1O1O001O1O001O1O001O1O0O2O1O001O1O1O001O1O001O1O0O2O1O001O1O001O1O1O001O1N101O1O1O001O1O1O001O1O1O0O2O1O1O001O1O1O001O1O1O001O100O001O1O1O001O1O1O001O1O1O001O100O001O1O1O001O1O1O001O1O1O1O10O01O1O1O1O001O1O1O1O1O001O1O1O100O1O001O1O1O1O1O1O1O1O1O1O2N1OO2O000000000O1000000000000O10001O0000000O1000000000001N10000000000O10000000001O0O1000000000000O10001O00000O1000000000000O101O000000001N10000000001N100000001O000O10001O00000O1O2N1N2O1O1O1N3N1O1O1N2O2N1N2O1O1O2M2O1O1O4K7J5H9D;D=DWK"}, "label": "the piece of bread to the right of the cup of sauce"}]}
{"id": 460228, "image": "COCO_train2014_000000460228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baked bun that is next to the cup\"."}], "task": "refering", "answer_template": "The \"a baked bun that is next to the cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 318], "bbox": [0.46323437500000003, 0.2426987447698745, 1.0, 0.7752719665271967], "iscrowd": 0, "area": 63711.4687, "rle": {"size": [478, 640], "counts": "d[Z4`0X>7I8H7I8H8H7K6M2N3M3L3N3M2N3M3M2N2M2O1O2N101O0O101N101N100O2O0O101N101N10001N101N100O2O0O101N101N10001O0O101O001O00001O0O2O00001O00001N101O00001O001O0O101O00001O001N10001O00001O001N10001O001O001O1O001O1O0O2O1O1O001O1O001O1O001O1O0O2O1O001O1O1O001O1O001O1O0O2O1O001O1O001O1O1O001O1N101O1O1O001O1O1O001O1O1O0O2O1O1O001O1O1O001O1O1O001O100O001O1O1O001O1O1O001O1O1O001O100O001O1O1O001O1O1O001O1O1O1O10O01O1O1O1O001O1O1O1O1O001O1O1O100O1O001O1O1O1O1O1O1O1O1O1O2N1OO2O000000000O1000000000000O10001O0000000O1000000000001N10000000000O10000000001O0O1000000000000O10001O00000O1000000000000O101O000000001N10000000001N100000001O000O10001O00000O1O2N1N2O1O1O1N3N1O1O1N2O2N1N2O1O1O2M2O1O1O4K7J5H9D;D=DWK"}, "label": "a baked bun that is next to the cup"}]}
{"id": 157125, "image": "COCO_train2014_000000157125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in straw hat on the phone\"."}], "task": "refering", "answer_template": "The \"man in straw hat on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 241, 242, 243, 244, 265, 266, 267, 286, 287, 288, 289, 290, 308, 309, 330, 331, 353, 376], "bbox": [0.375828125, 0.27133333333333337, 0.677796875, 0.9978125], "iscrowd": 0, "area": 15989.947750000003, "rle": {"size": [480, 640], "counts": "e]a39\\>9G:F8I7I21L6J6N310O2O2M3N2N101O1N2O1O01001O0O2O00001O010O010O00010O010O0010O010O01O100O100O1O01000O1000O0100O100O010O100O010O100O1O010O1O1O10O01O1O10O01O1O1000O1000O1000O100000ORJO]J6oNKb61[Jh0bNXOQ72TJY1aNeNY75kIa1kNZNY76bHKBP2b0nM[77`H3ROS2Q1dM]76_Ho23kL^76gGI4Y3f0hL_76eGM3W3h0fL`77bG11U3l0cLa77`G50S3m0bLb77_G70R3n0`Lb78]G:0P3Q1^L`7U1]G]2S1_L^7n4bHRK]7n4cHSK[7n4fHRKY7o4fHRKX7o4hHRKW7o4iHPKV7Q5jHPKU7Q5kHoJS7R5mHoJR7R5mHnJR7S5oHmJP7S5PInJn6S5SImJl6S5TImJk6T5VIlJh6U5XIlJg6T5ZIlJf6S5ZImJh6Q5YIoJg6P5YIQKg6n4ZIRKf6P5WIPKk6P5SIQKm6S5oHmJQ7X5iHiJW7Y5gHgJZ7Y5dHgJ]7Z5`HhJ`7Y5\\HjJd7X5VHlJk7T5QHoJo7Q60O101O00001O0000001O0000001O0000001O0O1N2oJcG^3_8`LeG[3]8dLfGX3\\8gLhGU3Y8jLkGQ3X8mLkGo2W8PMmGl2Z8jLkGS3R:N3M3K4L5[NdE_Nb:\\1jEWN[:h1`1M3N2M2O2M2O2N2M2O2N2N1N3N2N2N2N1O2N2K5K5L3LR[P3"}, "label": "man in straw hat on the phone"}]}
{"id": 157125, "image": "COCO_train2014_000000157125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the hat on\"."}], "task": "refering", "answer_template": "The \"the man with the hat on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 241, 242, 243, 244, 265, 266, 267, 286, 287, 288, 289, 290, 308, 309, 330, 331, 353, 376], "bbox": [0.375828125, 0.27133333333333337, 0.677796875, 0.9978125], "iscrowd": 0, "area": 15989.947750000003, "rle": {"size": [480, 640], "counts": "e]a39\\>9G:F8I7I21L6J6N310O2O2M3N2N101O1N2O1O01001O0O2O00001O010O010O00010O010O0010O010O01O100O100O1O01000O1000O0100O100O010O100O010O100O1O010O1O1O10O01O1O10O01O1O1000O1000O1000O100000ORJO]J6oNKb61[Jh0bNXOQ72TJY1aNeNY75kIa1kNZNY76bHKBP2b0nM[77`H3ROS2Q1dM]76_Ho23kL^76gGI4Y3f0hL_76eGM3W3h0fL`77bG11U3l0cLa77`G50S3m0bLb77_G70R3n0`Lb78]G:0P3Q1^L`7U1]G]2S1_L^7n4bHRK]7n4cHSK[7n4fHRKY7o4fHRKX7o4hHRKW7o4iHPKV7Q5jHPKU7Q5kHoJS7R5mHoJR7R5mHnJR7S5oHmJP7S5PInJn6S5SImJl6S5TImJk6T5VIlJh6U5XIlJg6T5ZIlJf6S5ZImJh6Q5YIoJg6P5YIQKg6n4ZIRKf6P5WIPKk6P5SIQKm6S5oHmJQ7X5iHiJW7Y5gHgJZ7Y5dHgJ]7Z5`HhJ`7Y5\\HjJd7X5VHlJk7T5QHoJo7Q60O101O00001O0000001O0000001O0000001O0O1N2oJcG^3_8`LeG[3]8dLfGX3\\8gLhGU3Y8jLkGQ3X8mLkGo2W8PMmGl2Z8jLkGS3R:N3M3K4L5[NdE_Nb:\\1jEWN[:h1`1M3N2M2O2M2O2N2M2O2N2N1N3N2N2N2N1O2N2K5K5L3LR[P3"}, "label": "the man with the hat on"}]}
{"id": 157125, "image": "COCO_train2014_000000157125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in white shirt with a red stripe on the shoulder\"."}], "task": "refering", "answer_template": "The \"person in white shirt with a red stripe on the shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 175, 176, 177, 178, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 376, 377, 378], "bbox": [0.37753125, 0.2719166666666667, 0.78034375, 0.9865208333333333], "iscrowd": 0, "area": 30992.502050000003, "rle": {"size": [480, 640], "counts": "bla35f>5J7I6K5J6K5J6J6K5O100O1O1O1O1O1O100O1O1O1O1O1O100O2N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O_NRD3l;NkD[OT;e0c1100O1O10O01O100O1O010O1O100O00100O1O10O0100O1O100O00100O1O10O01O100000O100000000O100000000O10O10000000O1000000000000001O000000000000000O10001O000000000000000O10000O2O0O10000O100O100O10001N100O10000O100O1]KnMXLR2c3TN]JNWOo1W6XNaJ8jNa1a5[ObKD_NQ1k5_OeK0PNb0X6BfK`2V4dMhK^2T4gMjKZ2S4iMlKX2P4mMmKU2S4kMlKV2T4kMjKV2V4kMgKW2X4jMgKX2X4iMfKX2Z4hMdKZ2\\4gMbKZ2^4fMaK[2_4fM^K\\2b4dM]K]2c4dM[K]2e4cMYK_2g4bMWK_2i4aMVK`2j4aMSKb2l4_MRKb2n4^MQKc2o4^MnJd2R5\\MmJe2S5\\MkJe2U5\\MhJf2X5[MfJf2Z5[MdJf2\\5\\M`Jf2`5[M^Jf2b5[M\\Jf2d5\\MXJg2g5ZMWJg2i5ZMUJg2k5W30000000000001O01O01O001O001O001O00001dImIj3R6RLRJn3o5nKTJR4l5jKXJU4j5gKYJW4i5eK[JY4g5dK]JY4f5bK^J[4e5cK]J[4f5aK]J]4e5aK]J]4f5_K]J^4f5`K\\J^4f5_K]J_4f5^K\\J`4f5]K]J`4g5]K[Ja4g5]K\\J`4g5\\K\\Ja4g5]K[Ja4g5\\K\\Jb4g5[K[Jc4g5ZK\\Jc4h5ZKZJd4h5YK[Je4h5XK[Je4T8J6K5K4L5J6K4L5J6K4L5K5J5L5K5F:A>C>A?BijQ2"}, "label": "person in white shirt with a red stripe on the shoulder"}]}
{"id": 157125, "image": "COCO_train2014_000000157125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teenager wearing a white shirt with a red and blue stripe\"."}], "task": "refering", "answer_template": "The \"teenager wearing a white shirt with a red and blue stripe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 175, 176, 177, 178, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 376, 377, 378], "bbox": [0.37753125, 0.2719166666666667, 0.78034375, 0.9865208333333333], "iscrowd": 0, "area": 30992.502050000003, "rle": {"size": [480, 640], "counts": "bla35f>5J7I6K5J6K5J6J6K5O100O1O1O1O1O1O100O1O1O1O1O1O100O2N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O_NRD3l;NkD[OT;e0c1100O1O10O01O100O1O010O1O100O00100O1O10O0100O1O100O00100O1O10O01O100000O100000000O100000000O10O10000000O1000000000000001O000000000000000O10001O000000000000000O10000O2O0O10000O100O100O10001N100O10000O100O1]KnMXLR2c3TN]JNWOo1W6XNaJ8jNa1a5[ObKD_NQ1k5_OeK0PNb0X6BfK`2V4dMhK^2T4gMjKZ2S4iMlKX2P4mMmKU2S4kMlKV2T4kMjKV2V4kMgKW2X4jMgKX2X4iMfKX2Z4hMdKZ2\\4gMbKZ2^4fMaK[2_4fM^K\\2b4dM]K]2c4dM[K]2e4cMYK_2g4bMWK_2i4aMVK`2j4aMSKb2l4_MRKb2n4^MQKc2o4^MnJd2R5\\MmJe2S5\\MkJe2U5\\MhJf2X5[MfJf2Z5[MdJf2\\5\\M`Jf2`5[M^Jf2b5[M\\Jf2d5\\MXJg2g5ZMWJg2i5ZMUJg2k5W30000000000001O01O01O001O001O001O00001dImIj3R6RLRJn3o5nKTJR4l5jKXJU4j5gKYJW4i5eK[JY4g5dK]JY4f5bK^J[4e5cK]J[4f5aK]J]4e5aK]J]4f5_K]J^4f5`K\\J^4f5_K]J_4f5^K\\J`4f5]K]J`4g5]K[Ja4g5]K\\J`4g5\\K\\Ja4g5]K[Ja4g5\\K\\Jb4g5[K[Jc4g5ZK\\Jc4h5ZKZJd4h5YK[Je4h5XK[Je4T8J6K5K4L5J6K4L5J6K4L5K5J5L5K5F:A>C>A?BijQ2"}, "label": "teenager wearing a white shirt with a red and blue stripe"}]}
{"id": 157125, "image": "COCO_train2014_000000157125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two shirtless men talk on their cell phones between two individuals with shirts on a dock\"."}], "task": "refering", "answer_template": "The \"two shirtless men talk on their cell phones between two individuals with shirts on a dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 148, 149, 150, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 239, 240, 241, 243, 263, 264, 265, 266, 285, 286, 287, 288, 289, 307, 308, 330], "bbox": [0.37496874999999996, 0.3403541666666667, 0.589578125, 0.9110416666666666], "iscrowd": 0, "area": 12744.060249999997, "rle": {"size": [480, 640], "counts": "Wl`36h><D<D<DE<L30000O2O1O2N1O2M2O2N1O2O0O2O001O0O2O00oKKYI3i6f0^HYOb7V1oGjNR8Z1jGdNW8a1cGaN]8c1^G^Na8f1[G[Nd8j1WGVNj8n1QGSNn8R2mFoMS9S2jFnMU9U2gFlMZ9V2bFlM\\9X2`FjMY8GaG_25lMX8K]G[29kMY8OZGW2;lMX84XGP2?nMW88TGl1c0nMW8<QGg1f0nMX8]3gGeLX8\\3gGeLX8\\3gGeLY8Z3gGfLY8[3fGfLZ8Z3eGgLZ8Y3fGhLZ8X3dGiL\\8X3bGjL^8U3aGmL^8T3`GnL`8Q3_GPMa8Q3]GQMc8o2ZGTMe8l2ZGVMf8j2XGXMg8i2WGYMi8f2VG[Mj8e2UG]Mk8a2TGbMk8^2TGdMl8X4ORNUGiMk8V4O100O11O001O1O001O0O2O1O001N2O001O0O2O1O001ZMkF^OW9;XG\\Oi8=fGZO]8=THXOo7?`HWOc7a0cH_O`78gHF\\72jHMZ7KkH4X7DnH;l:O00100O1O1O10O10O0100O100O010O100O10O0100O1000O0100O100O01H:L5K4M4L4L3M4Lonj3"}, "label": "two shirtless men talk on their cell phones between two individuals with shirts on a dock"}]}
{"id": 157125, "image": "COCO_train2014_000000157125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with tattoos on his wrist\"."}], "task": "refering", "answer_template": "The \"a man with tattoos on his wrist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 148, 149, 150, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 239, 240, 241, 243, 263, 264, 265, 266, 285, 286, 287, 288, 289, 307, 308, 330], "bbox": [0.37496874999999996, 0.3403541666666667, 0.589578125, 0.9110416666666666], "iscrowd": 0, "area": 12744.060249999997, "rle": {"size": [480, 640], "counts": "Wl`36h><D<D<DE<L30000O2O1O2N1O2M2O2N1O2O0O2O001O0O2O00oKKYI3i6f0^HYOb7V1oGjNR8Z1jGdNW8a1cGaN]8c1^G^Na8f1[G[Nd8j1WGVNj8n1QGSNn8R2mFoMS9S2jFnMU9U2gFlMZ9V2bFlM\\9X2`FjMY8GaG_25lMX8K]G[29kMY8OZGW2;lMX84XGP2?nMW88TGl1c0nMW8<QGg1f0nMX8]3gGeLX8\\3gGeLX8\\3gGeLY8Z3gGfLY8[3fGfLZ8Z3eGgLZ8Y3fGhLZ8X3dGiL\\8X3bGjL^8U3aGmL^8T3`GnL`8Q3_GPMa8Q3]GQMc8o2ZGTMe8l2ZGVMf8j2XGXMg8i2WGYMi8f2VG[Mj8e2UG]Mk8a2TGbMk8^2TGdMl8X4ORNUGiMk8V4O100O11O001O1O001O0O2O1O001N2O001O0O2O1O001ZMkF^OW9;XG\\Oi8=fGZO]8=THXOo7?`HWOc7a0cH_O`78gHF\\72jHMZ7KkH4X7DnH;l:O00100O1O1O10O10O0100O100O010O100O10O0100O1000O0100O100O01H:L5K4M4L4L3M4Lonj3"}, "label": "a man with tattoos on his wrist"}]}
{"id": 148937, "image": "COCO_train2014_000000148937.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in costume\"."}], "task": "refering", "answer_template": "The \"woman in costume\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 98, 105, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 226, 227, 228, 229, 230, 231, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 271, 272, 273, 274, 275, 276, 286, 287, 288, 289, 290, 291, 301, 302, 303, 304, 305, 306], "bbox": [0.04714285714285714, 0.1483125, 0.5624824355971897, 0.8943749999999999], "iscrowd": 0, "area": 75108.99775000001, "rle": {"size": [640, 427], "counts": "[f<3fc0;D;F:F:E;J6L5K4K5L4L4L4K4M1O2N1O1N2O2N1O1N3N1\\EjL]4W3ZK[M\\4f2\\KkMY4X2bHoL2[1Q7g1eHlM[Oo0f7V1fHjNfNa0Z8f0hHg2n6[MhHY3k6iLmHU4\\6lK_Io4j5RKSJ`5]5bJ_J`5`5bJ]J`5c5aJZJb5e5`JWJc5h5^JUJe5i5]JVJd5i5]JVJd5i5^JVJb5i5_JWJa5h5`JXJ`5g5aJXJ`5g5bJXJ^5h5bJXJ^5g5cJYJ]5f5dJYJ]5g5^4O100O1O100O1O100O1O100O1O100O1O100O1O100O1O10000000000O101O00000000000O1000000000000O100000000000000O1000000000000001O000000000000001O000000000000001O01O000000000001O000000000000001O000000000000001O000000000000001O00000000001O1O1O001\\MgL_DZ3f:^3M3L3N3L4L4M3ZJ`Fg0c9XOaFc0b9[ObF?b9^OcF=`9AdF9_9FeF5^9IfFNa9OdFEf99^F[Ol9c0WFROR:m0RFgNX:V1mE^N]:`1TFgMV:W2]FnLl9Q3gFTLd9h3]3M3M2N3L4M2N3M3L4M2N3M3L4M2N4L9F;F8H4L3L5L4L3M4K5L3M4L4K4M4Laed3"}, "label": "woman in costume"}]}
{"id": 148937, "image": "COCO_train2014_000000148937.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a prehistoric costume\"."}], "task": "refering", "answer_template": "The \"a woman wearing a prehistoric costume\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 98, 105, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 226, 227, 228, 229, 230, 231, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 271, 272, 273, 274, 275, 276, 286, 287, 288, 289, 290, 291, 301, 302, 303, 304, 305, 306], "bbox": [0.04714285714285714, 0.1483125, 0.5624824355971897, 0.8943749999999999], "iscrowd": 0, "area": 75108.99775000001, "rle": {"size": [640, 427], "counts": "[f<3fc0;D;F:F:E;J6L5K4K5L4L4L4K4M1O2N1O1N2O2N1O1N3N1\\EjL]4W3ZK[M\\4f2\\KkMY4X2bHoL2[1Q7g1eHlM[Oo0f7V1fHjNfNa0Z8f0hHg2n6[MhHY3k6iLmHU4\\6lK_Io4j5RKSJ`5]5bJ_J`5`5bJ]J`5c5aJZJb5e5`JWJc5h5^JUJe5i5]JVJd5i5]JVJd5i5^JVJb5i5_JWJa5h5`JXJ`5g5aJXJ`5g5bJXJ^5h5bJXJ^5g5cJYJ]5f5dJYJ]5g5^4O100O1O100O1O100O1O100O1O100O1O100O1O100O1O10000000000O101O00000000000O1000000000000O100000000000000O1000000000000001O000000000000001O000000000000001O01O000000000001O000000000000001O000000000000001O000000000000001O00000000001O1O1O001\\MgL_DZ3f:^3M3L3N3L4L4M3ZJ`Fg0c9XOaFc0b9[ObF?b9^OcF=`9AdF9_9FeF5^9IfFNa9OdFEf99^F[Ol9c0WFROR:m0RFgNX:V1mE^N]:`1TFgMV:W2]FnLl9Q3gFTLd9h3]3M3M2N3L4M2N3M3L4M2N3M3L4M2N4L9F;F8H4L3L5L4L3M4K5L3M4L4K4M4Laed3"}, "label": "a woman wearing a prehistoric costume"}]}
{"id": 99788, "image": "COCO_train2014_000000099788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry picture of someones side boob that is wearing a tan shirt\"."}], "task": "refering", "answer_template": "The \"a blurry picture of someones side boob that is wearing a tan shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 90, 91, 108, 109, 110, 111, 126, 127, 128, 129, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202], "bbox": [0.00146, 0.34796923076923075, 0.28253999999999996, 0.9898769230769232], "iscrowd": 0, "area": 18922.188299999998, "rle": {"size": [325, 500], "counts": "f=a6f3N2N2N1O2N2N2N2N2N1O2N2N2N1O2N2N2N2N1N3N2N001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O1O1O1O1O2N1O1O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N5K5K5K5K5K4L5K5K5K5K5K5K5KWda3"}, "label": "a blurry picture of someones side boob that is wearing a tan shirt"}]}
{"id": 99788, "image": "COCO_train2014_000000099788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person out of focus\"."}], "task": "refering", "answer_template": "The \"a person out of focus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 90, 91, 108, 109, 110, 111, 126, 127, 128, 129, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202], "bbox": [0.00146, 0.34796923076923075, 0.28253999999999996, 0.9898769230769232], "iscrowd": 0, "area": 18922.188299999998, "rle": {"size": [325, 500], "counts": "f=a6f3N2N2N1O2N2N2N2N2N1O2N2N2N1O2N2N2N2N1N3N2N001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O1O1O1O1O2N1O1O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N5K5K5K5K5K4L5K5K5K5K5K5K5KWda3"}, "label": "a person out of focus"}]}
{"id": 99788, "image": "COCO_train2014_000000099788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blurry arm of a person near a grill\"."}], "task": "refering", "answer_template": "The \"blurry arm of a person near a grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 125, 142, 143, 160, 161, 177, 178, 179, 195, 196, 197, 213, 214, 215], "bbox": [0.82236, 0.4427076923076923, 1.0, 0.9887692307692308], "iscrowd": 0, "area": 8068.754750000002, "rle": {"size": [325, 500], "counts": "[RS42Q:2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2`K"}, "label": "blurry arm of a person near a grill"}]}
{"id": 99788, "image": "COCO_train2014_000000099788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurred object of a pinkish color\"."}], "task": "refering", "answer_template": "The \"a blurred object of a pinkish color\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 125, 142, 143, 160, 161, 177, 178, 179, 195, 196, 197, 213, 214, 215], "bbox": [0.82236, 0.4427076923076923, 1.0, 0.9887692307692308], "iscrowd": 0, "area": 8068.754750000002, "rle": {"size": [325, 500], "counts": "[RS42Q:2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2M3M3M3M3N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2`K"}, "label": "a blurred object of a pinkish color"}]}
{"id": 370124, "image": "COCO_train2014_000000370124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and red train engine with the numbers 41096\"."}], "task": "refering", "answer_template": "The \"a black and red train engine with the numbers 41096\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 30, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 353, 354, 355, 356, 357, 358, 360], "bbox": [0.1485625, 0.036229166666666666, 0.8168906250000001, 0.9227916666666667], "iscrowd": 0, "area": 122336.51549999992, "rle": {"size": [480, 640], "counts": "gh\\15j>:F9H9F9G:F7J1N3M2N3M2O1L5K8G:G5K5K5J4M4K5J5M4N2N2M2O2N2N2N1N3N2N1O2O1N2N101N2N1O2O1N2O001N2O1O0O2O1O0O2N2N2N2N2N2O1N2N1O2N2N2N2N2N2N2N2O1N1O2N2N2N2N2N2L4K5J6K4K6J6K5J6K5J6M3M3M3M2M4M3M3M3M3M3L4M2N3M3M3M3L4M3M2O2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N2N1O2N2M3N2NmDlL]:R3^EUMb:i2XE^Mi:_2REhMo:T3010O1O1O00100O1O00100O001O100O001O1O010O000O1N2O1N2O1O2M2O1N2I7L4H9H7J6O1O1N3N1O1O10000O2O000O2O001N10001N01O1O2M2K5I7J6L4M3N2N2M3N26J5K00001O00001O01O01O0fMbCR2^<mMdCR2\\<mMfCR2[<lMfCT2Z<kMhCT2X<kMiCU2b<O001O001O0010O00010O01O0010O0010O0010O001O00001O00001O00001O001O0010O01O001O001O001O001O001O001O001O001N101O001O001O001N1N3M2L5K4L5N2O00100O00100O010O1O10O01O10O01O1N101O1O1O1O1O1O2N1O1N2N3M3M3L4M3L4M3L4M2K4L3M40O2O1N5L3L4M4K4MdLbKYLZ4h3fK\\LV4c3kK`LQ4`3PLbLn3^3RLbLm3^3TLaLm3_3SLaLl3_3ULaLk3^3VL\\K3\\Ng3Y6VLZKa0oMZ3f6VLZKo4f4RKXKP5h4PKWKP5i4QKVKP5i4QKVKo4k4QKSKQ5l4PKQKR5n4PKPKR5o4oJQKP5P1dJk0<UNQ5P1eJg0;ZNo4o0iJd08^No4n0lJ`04dNo4m0TK6NmNo4m0ZKNGWOm4k0cKFB_Ok4k0hK@^OFh4k0lKVMoNg1>9f4j0PLPMUOe17a0c4k0nLXN`Nn0a4i0PMWN`NR1^4h0RMUNaNU1[4f0TMTNbN^1Q4?]MRNcNf1i38dMPNeNl1b34jMoMeNm1a34jMUMiM=m0[2^34lMQMmM>j0]2]34lMQMoM:j0a2Z34nMQMoM8j0c2Y34nMQMPN6j0e2W35oMPMQN4j0g2V35oMPMRN1l0i2R37PNoLVNLi0n2Q36QNPMBi2]27QNPMBi2]27QNPMCh2\\28QNPMCh2\\28QNPMDg2[28RNPMDh2\\26PNRMEg2]25nMTMEg2`22kMWMFf2a20jMZMEf2c2NhM\\MEf2f2KeM_MFn1PNaNf4P2dMaMFn1UN\\Nc4S2bMcMGl1UN]Nf4P2_MgMFl1VN\\Ng4o1]MiMGj1VN^Ni4l1EFcK]Nj4k1DGbK^Nl4i1BHcK_Nn4e1AKbK_No4d1_OLcK`NP5b1]ONdK_NU5\\1XO5cK_N[5S1UO=aK_Na5l0POe0`K^Ng5c0lNn0^K_Nl5:iNW1\\K^NY6J^Nh1YK^Nf;a1[D_Ne;a1\\D^Nd;b1\\D^Nd;b1]D]Nc;c1]D\\Nd;d1]DYNf;f1ZDYNg;f1ZDYNg;g1i0000001O0O1O1O1O2N1O1N2N4M5J5K5J6I8H7Ia`f1"}, "label": "a black and red train engine with the numbers 41096"}]}
{"id": 370124, "image": "COCO_train2014_000000370124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and red train with the numbers 41 096 in front\"."}], "task": "refering", "answer_template": "The \"a black and red train with the numbers 41 096 in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 30, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 353, 354, 355, 356, 357, 358, 360], "bbox": [0.1485625, 0.036229166666666666, 0.8168906250000001, 0.9227916666666667], "iscrowd": 0, "area": 122336.51549999992, "rle": {"size": [480, 640], "counts": "gh\\15j>:F9H9F9G:F7J1N3M2N3M2O1L5K8G:G5K5K5J4M4K5J5M4N2N2M2O2N2N2N1N3N2N1O2O1N2N101N2N1O2O1N2O001N2O1O0O2O1O0O2N2N2N2N2N2O1N2N1O2N2N2N2N2N2N2N2O1N1O2N2N2N2N2N2L4K5J6K4K6J6K5J6K5J6M3M3M3M2M4M3M3M3M3M3L4M2N3M3M3M3L4M3M2O2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N2N1O2N2M3N2NmDlL]:R3^EUMb:i2XE^Mi:_2REhMo:T3010O1O1O00100O1O00100O001O100O001O1O010O000O1N2O1N2O1O2M2O1N2I7L4H9H7J6O1O1N3N1O1O10000O2O000O2O001N10001N01O1O2M2K5I7J6L4M3N2N2M3N26J5K00001O00001O01O01O0fMbCR2^<mMdCR2\\<mMfCR2[<lMfCT2Z<kMhCT2X<kMiCU2b<O001O001O0010O00010O01O0010O0010O0010O001O00001O00001O00001O001O0010O01O001O001O001O001O001O001O001O001N101O001O001O001N1N3M2L5K4L5N2O00100O00100O010O1O10O01O10O01O1N101O1O1O1O1O1O2N1O1N2N3M3M3L4M3L4M3L4M2K4L3M40O2O1N5L3L4M4K4MdLbKYLZ4h3fK\\LV4c3kK`LQ4`3PLbLn3^3RLbLm3^3TLaLm3_3SLaLl3_3ULaLk3^3VL\\K3\\Ng3Y6VLZKa0oMZ3f6VLZKo4f4RKXKP5h4PKWKP5i4QKVKP5i4QKVKo4k4QKSKQ5l4PKQKR5n4PKPKR5o4oJQKP5P1dJk0<UNQ5P1eJg0;ZNo4o0iJd08^No4n0lJ`04dNo4m0TK6NmNo4m0ZKNGWOm4k0cKFB_Ok4k0hK@^OFh4k0lKVMoNg1>9f4j0PLPMUOe17a0c4k0nLXN`Nn0a4i0PMWN`NR1^4h0RMUNaNU1[4f0TMTNbN^1Q4?]MRNcNf1i38dMPNeNl1b34jMoMeNm1a34jMUMiM=m0[2^34lMQMmM>j0]2]34lMQMoM:j0a2Z34nMQMoM8j0c2Y34nMQMPN6j0e2W35oMPMQN4j0g2V35oMPMRN1l0i2R37PNoLVNLi0n2Q36QNPMBi2]27QNPMBi2]27QNPMCh2\\28QNPMCh2\\28QNPMDg2[28RNPMDh2\\26PNRMEg2]25nMTMEg2`22kMWMFf2a20jMZMEf2c2NhM\\MEf2f2KeM_MFn1PNaNf4P2dMaMFn1UN\\Nc4S2bMcMGl1UN]Nf4P2_MgMFl1VN\\Ng4o1]MiMGj1VN^Ni4l1EFcK]Nj4k1DGbK^Nl4i1BHcK_Nn4e1AKbK_No4d1_OLcK`NP5b1]ONdK_NU5\\1XO5cK_N[5S1UO=aK_Na5l0POe0`K^Ng5c0lNn0^K_Nl5:iNW1\\K^NY6J^Nh1YK^Nf;a1[D_Ne;a1\\D^Nd;b1\\D^Nd;b1]D]Nc;c1]D\\Nd;d1]DYNf;f1ZDYNg;f1ZDYNg;g1i0000001O0O1O1O1O2N1O1N2N4M5J5K5J6I8H7Ia`f1"}, "label": "a black and red train with the numbers 41 096 in front"}]}
{"id": 370124, "image": "COCO_train2014_000000370124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train to the right of the black locmotive , three men , two standing , one sitting\"."}], "task": "refering", "answer_template": "The \"the train to the right of the black locmotive , three men , two standing , one sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 252, 272, 275, 295, 296, 298], "bbox": [0.692703125, 0.11197916666666667, 1.0, 0.7636666666666667], "iscrowd": 0, "area": 34855.56544999998, "rle": {"size": [480, 640], "counts": "Yn_61Z;0ZH0e72YHOg73VHOh73VHNj73THNl73SHNk74SHMm75PHLo76oGLP85nGLR86jGLU8;dGG[8>^GDa8b0XG@h8e0QG^On8h0kFYOT9m0eFUO[9m0aFVO^9k0_FWO`9k0^FVOb9l0ZFWOd9k0YFWOg9j0mEAS:>mEDQ:=oECQ:<oEEP:<oEEQ::oEGQ:9oEGP::oEGQ::mEFS:<kEEU:<jEDV:<iEEV:=hEDX:=fEDZ:=eECZ:>eEB\\:?bEB]:`0aEA_:`0^EBb:\\2101O001O1O001O1O001O00\\NPLZHn3h7^LlGa3U8jL_GV3a8WMSGh2n8cMfF\\2\\9c1001O001N2O001O001O11N101N2O0iFaJm8`5QGcJl8_5SGdJj8\\5UGgJi8Z5UGiJh8j5N2M3N2N2M3N2M3N2N2M3N2M4M3M3L5L3M3L4M3L4M3MVOVIaIf6_6\\IbIb6]6`IcI^6]6dIcI[6\\6gIeIW6Z6kIfIS6Z6oIfIP6Y6RJhIk5Y6VJgIi5X6YJhIf5W6\\JjIa5V6aJhI`5W6bJhI^5W6dJhI[5X6gJgIY5Y6hJfIW5Y6lJfIT5Y6nJfIR5Y6RKdIm4[6YKaIg4^6_K]I`4b6fKZIZ4e6kKVIV4i6P2O2O001O1O001O1O001O0VOQHdJP8Y5m0M2O2M2O2O1O001O10O01O0101O1N3N1N2O1O1N3N1N2O1O1N3N1O1N3N3L5L4L3L5L3L2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1VJ^HS4d7jK_HU4b7gKaHX4a7eKbHZ4_7cKcHj4Q7RKSIY6b5dIaJg7T4UHoKW9f2[FhMP;]2]1bN^1cNfE"}, "label": "the train to the right of the black locmotive , three men , two standing , one sitting"}]}
{"id": 370124, "image": "COCO_train2014_000000370124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver train car with man poking head out through open doors\"."}], "task": "refering", "answer_template": "The \"silver train car with man poking head out through open doors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 252, 272, 275, 295, 296, 298], "bbox": [0.692703125, 0.11197916666666667, 1.0, 0.7636666666666667], "iscrowd": 0, "area": 34855.56544999998, "rle": {"size": [480, 640], "counts": "Yn_61Z;0ZH0e72YHOg73VHOh73VHNj73THNl73SHNk74SHMm75PHLo76oGLP85nGLR86jGLU8;dGG[8>^GDa8b0XG@h8e0QG^On8h0kFYOT9m0eFUO[9m0aFVO^9k0_FWO`9k0^FVOb9l0ZFWOd9k0YFWOg9j0mEAS:>mEDQ:=oECQ:<oEEP:<oEEQ::oEGQ:9oEGP::oEGQ::mEFS:<kEEU:<jEDV:<iEEV:=hEDX:=fEDZ:=eECZ:>eEB\\:?bEB]:`0aEA_:`0^EBb:\\2101O001O1O001O1O001O00\\NPLZHn3h7^LlGa3U8jL_GV3a8WMSGh2n8cMfF\\2\\9c1001O001N2O001O001O11N101N2O0iFaJm8`5QGcJl8_5SGdJj8\\5UGgJi8Z5UGiJh8j5N2M3N2N2M3N2M3N2N2M3N2M4M3M3L5L3M3L4M3L4M3MVOVIaIf6_6\\IbIb6]6`IcI^6]6dIcI[6\\6gIeIW6Z6kIfIS6Z6oIfIP6Y6RJhIk5Y6VJgIi5X6YJhIf5W6\\JjIa5V6aJhI`5W6bJhI^5W6dJhI[5X6gJgIY5Y6hJfIW5Y6lJfIT5Y6nJfIR5Y6RKdIm4[6YKaIg4^6_K]I`4b6fKZIZ4e6kKVIV4i6P2O2O001O1O001O1O001O0VOQHdJP8Y5m0M2O2M2O2O1O001O10O01O0101O1N3N1N2O1O1N3N1N2O1O1N3N1O1N3N3L5L4L3L5L3L2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1VJ^HS4d7jK_HU4b7gKaHX4a7eKbHZ4_7cKcHj4Q7RKSIY6b5dIaJg7T4UHoKW9f2[FhMP;]2]1bN^1cNfE"}, "label": "silver train car with man poking head out through open doors"}]}
{"id": 83407, "image": "COCO_train2014_000000083407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue stop sign\"."}], "task": "refering", "answer_template": "The \"a blue stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 183, 184, 185, 186, 187, 188, 206, 207, 208, 209, 210, 229, 230, 231], "bbox": [0.3415686274509804, 0.1375, 0.5900163398692811, 0.48663398692810456], "iscrowd": 0, "area": 25386.327499999996, "rle": {"size": [612, 612], "counts": "\\Qm39jb0c0]Ok0UOl0TO`0_O4M3M4L3M3M2N2N3M2N2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2O1O01O00001O0000001O0000010O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0000001O00010O0000001O1O1O2N1O2N1O1N3L3N3M2M4M2M4M2M4M2N2M4M2M4M2N3L3N3L3N3M2M3N3L3N3M2M4M2M4M2N3L3N2M4M2M4M9FSke4"}, "label": "a blue stop sign"}]}
{"id": 83407, "image": "COCO_train2014_000000083407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue stop sign\"."}], "task": "refering", "answer_template": "The \"a blue stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 183, 184, 185, 186, 187, 188, 206, 207, 208, 209, 210, 229, 230, 231], "bbox": [0.3415686274509804, 0.1375, 0.5900163398692811, 0.48663398692810456], "iscrowd": 0, "area": 25386.327499999996, "rle": {"size": [612, 612], "counts": "\\Qm39jb0c0]Ok0UOl0TO`0_O4M3M4L3M3M2N2N3M2N2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2O1O01O00001O0000001O0000010O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0000001O00010O0000001O1O1O2N1O2N1O1N3L3N3M2M4M2M4M2M4M2N2M4M2M4M2N3L3N3L3N3M2M3N3L3N3M2M4M2M4M2N3L3N2M4M2M4M9FSke4"}, "label": "a blue stop sign"}]}
{"id": 312785, "image": "COCO_train2014_000000312785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table covered in newspaper\"."}], "task": "refering", "answer_template": "The \"a table covered in newspaper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 104, 105, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.411234375, 0.2045, 1.0, 0.9977499999999999], "iscrowd": 0, "area": 106668.70915, "rle": {"size": [480, 640], "counts": "mgk32g>7H8I7I7I7I7I7I7I7H8I7J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6O1O10000O10000O1000000O10000O1000000O1000000O10000O1000000O10000WMlHfMT7Y2mHfMT7Y2mHgMS7X2nHhMR7W2PIgMQ7X2PIhMP7V2RIjMn6U2SIkMm6T2TIkMm6T2TIlMl6S2VIlMj6S2WImMi6R2XImMi6R2XInMh6P2ZIPNf6o1[IPNf6o1\\IPNd6o1]IQNc6n1^IRNb6m1_IRNb6m1_ISNa6l1`ITN`6k1bIRN`6l1bIRN`6m1aIQNa6n1`IPNb6o1_IoMc6P2^InMd6R2]IkMe6T2\\IiMg6W2YIgMi6X2XIfMj6Z2VIdMl6[2UIcMm6]2TI`Mn6_2SI_Mo6b2PI\\MR7d2nHZMT7g2nHTMT7l2QImLQ7T3RIfLP7Z3UI_Lm6b3WIWLk6i3YIQLi6P4S21O001O001O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000N2N2O1N2O1N200O10000O10000O100O10000O10000O10000O100O10000O10000O10000O10000O100O10000O100H8F:M3000000O1000000O1000000O1000000O1000000O0101O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O3M2N3M3M2N3M2N2N1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O2N3M2N2N2N3M2N2N3M2N2M3N3M2M3N2M4M2N2M4M2N2M3N3L3N2N2M4M2M3N3M2J6jNV1jNW1iNV1jNeG"}, "label": "a table covered in newspaper"}]}
{"id": 312785, "image": "COCO_train2014_000000312785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table covered in newspaper for crafts\"."}], "task": "refering", "answer_template": "The \"a table covered in newspaper for crafts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 104, 105, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.411234375, 0.2045, 1.0, 0.9977499999999999], "iscrowd": 0, "area": 106668.70915, "rle": {"size": [480, 640], "counts": "mgk32g>7H8I7I7I7I7I7I7I7H8I7J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6O1O10000O10000O1000000O10000O1000000O1000000O10000O1000000O10000WMlHfMT7Y2mHfMT7Y2mHgMS7X2nHhMR7W2PIgMQ7X2PIhMP7V2RIjMn6U2SIkMm6T2TIkMm6T2TIlMl6S2VIlMj6S2WImMi6R2XImMi6R2XInMh6P2ZIPNf6o1[IPNf6o1\\IPNd6o1]IQNc6n1^IRNb6m1_IRNb6m1_ISNa6l1`ITN`6k1bIRN`6l1bIRN`6m1aIQNa6n1`IPNb6o1_IoMc6P2^InMd6R2]IkMe6T2\\IiMg6W2YIgMi6X2XIfMj6Z2VIdMl6[2UIcMm6]2TI`Mn6_2SI_Mo6b2PI\\MR7d2nHZMT7g2nHTMT7l2QImLQ7T3RIfLP7Z3UI_Lm6b3WIWLk6i3YIQLi6P4S21O001O001O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000N2N2O1N2O1N200O10000O10000O100O10000O10000O10000O100O10000O10000O10000O10000O100O10000O100H8F:M3000000O1000000O1000000O1000000O1000000O0101O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O3M2N3M3M2N3M2N2N1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O2N3M2N2N2N3M2N2N3M2N2M3N3M2M3N2M4M2N2M4M2N2M3N3L3N2N2M4M2M3N3M2J6jNV1jNW1iNV1jNeG"}, "label": "a table covered in newspaper for crafts"}]}
{"id": 58836, "image": "COCO_train2014_000000058836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a suitcase with papers on it\"."}], "task": "refering", "answer_template": "The \"a suitcase with papers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [398, 399, 400, 401, 402, 403, 421, 422, 423, 424, 425, 426, 427, 443, 444, 445, 446, 448, 449, 450, 467, 468, 469, 470, 471, 472, 473, 491, 492, 493, 494, 495, 496, 514, 515], "bbox": [0.30012500000000003, 0.747, 0.6046250000000001, 0.9644999999999999], "iscrowd": 0, "area": 16694.301299999992, "rle": {"size": [640, 640], "counts": "l`h34ec08I8H7H9H7H9H7I8N1O2N1O2N2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1N3N1O2N2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N0O100000O10000000000000O010006J<D=oLZ_On1^a0N1N3N1O2N1O2N1O2M2O2N1O2N1O4L5J5L4L2N1O1O1O0O1000000000000O100000oLTOnBl0P=YOmBg0Q=]OmBc0S=@jB?W=DfB<Z=FdB:\\=I`B8`=K]B5c=M[B3e=0WB1i=2TBMm=7oAIQ>:lAFT>=hADX>i0[AWOe>k0YAUOg>l0XATOh>n0UAROl>o0SAQOm>Q1QAoNo>R1PAnNP?T1n@lNR?U1l@lNT?V1j@jNU?Y1h@hNX?Y1g@fNZ?\\1c@eN]?]1`@dN`?]1^@dNb?^1[@cNe?_1Y@aNg?`1W@aNi?a1T@_Nm?c1P@^NP`0i2000000000O2O2N1O1O1O1O1N2O2N1O1O1O1O1N2O2N1O1O1O1O1N3N1O1O1O1O1O1N3N1O1O1O1O1O1N3N1O1O1O1O1O2M2O1O1O1O1O1O2N1N2O1O1O1Om0SOWbm4"}, "label": "a suitcase with papers on it"}]}
{"id": 58836, "image": "COCO_train2014_000000058836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown suitcase underneath a woman ' s bare foot\"."}], "task": "refering", "answer_template": "The \"a brown suitcase underneath a woman ' s bare foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [398, 399, 400, 401, 402, 403, 421, 422, 423, 424, 425, 426, 427, 443, 444, 445, 446, 448, 449, 450, 467, 468, 469, 470, 471, 472, 473, 491, 492, 493, 494, 495, 496, 514, 515], "bbox": [0.30012500000000003, 0.747, 0.6046250000000001, 0.9644999999999999], "iscrowd": 0, "area": 16694.301299999992, "rle": {"size": [640, 640], "counts": "l`h34ec08I8H7H9H7H9H7I8N1O2N1O2N2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1N3N1O2N2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N0O100000O10000000000000O010006J<D=oLZ_On1^a0N1N3N1O2N1O2N1O2M2O2N1O2N1O4L5J5L4L2N1O1O1O0O1000000000000O100000oLTOnBl0P=YOmBg0Q=]OmBc0S=@jB?W=DfB<Z=FdB:\\=I`B8`=K]B5c=M[B3e=0WB1i=2TBMm=7oAIQ>:lAFT>=hADX>i0[AWOe>k0YAUOg>l0XATOh>n0UAROl>o0SAQOm>Q1QAoNo>R1PAnNP?T1n@lNR?U1l@lNT?V1j@jNU?Y1h@hNX?Y1g@fNZ?\\1c@eN]?]1`@dN`?]1^@dNb?^1[@cNe?_1Y@aNg?`1W@aNi?a1T@_Nm?c1P@^NP`0i2000000000O2O2N1O1O1O1O1N2O2N1O1O1O1O1N2O2N1O1O1O1O1N3N1O1O1O1O1O1N3N1O1O1O1O1O1N3N1O1O1O1O1O2M2O1O1O1O1O1O2N1N2O1O1O1Om0SOWbm4"}, "label": "a brown suitcase underneath a woman ' s bare foot"}]}
{"id": 58836, "image": "COCO_train2014_000000058836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a suitcase with hat on it\"."}], "task": "refering", "answer_template": "The \"a suitcase with hat on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [365, 366, 367, 388, 389, 390, 412, 413, 434, 435, 436, 453, 454, 455, 456, 457, 458, 459, 476, 477, 478, 479, 480, 481, 482, 499, 500, 501, 502, 503, 504, 505, 523, 524, 525, 526], "bbox": [0.727265625, 0.657140625, 1.0, 0.984421875], "iscrowd": 0, "area": 20164.9655, "rle": {"size": [640, 640], "counts": "gTS93mc0h0YOg0XOi0WO8H01O00000010O000001O0001O0001O0000010O0000001O01O0001O000001O01O00000010O000001O0001O01O00000010O000000001O000O100O101O0O10000O101N10000O010O10O10O01000O010O10O10O010O10O10O01000O010O01000O001N2N1O2TMVM`Dk2a;XMZDj2f;YMUDi2j;[MQDf2P<^MjCd2U<`MfCa2[<bM`C`2`<cM[C_2d<eMWC\\2j<gMQC[2n<iMmBY2S=jMhBW2Y=lMbBV2]=nM^BS2c=PNXBR2g=RNTBP2l=SNoAn1R>UNiAm1V>n110O10O10O10O010O10O10O010O10000O2O0O2O000O2O001N101N101O0O101N101O0O2O000O2O0O2O001N100O2O001N101O0O2O0O\\B"}, "label": "a suitcase with hat on it"}]}
{"id": 58836, "image": "COCO_train2014_000000058836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown piece of luggage with a woman ' s hat on top\"."}], "task": "refering", "answer_template": "The \"a brown piece of luggage with a woman ' s hat on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [365, 366, 367, 388, 389, 390, 412, 413, 434, 435, 436, 453, 454, 455, 456, 457, 458, 459, 476, 477, 478, 479, 480, 481, 482, 499, 500, 501, 502, 503, 504, 505, 523, 524, 525, 526], "bbox": [0.727265625, 0.657140625, 1.0, 0.984421875], "iscrowd": 0, "area": 20164.9655, "rle": {"size": [640, 640], "counts": "gTS93mc0h0YOg0XOi0WO8H01O00000010O000001O0001O0001O0000010O0000001O01O0001O000001O01O00000010O000001O0001O01O00000010O000000001O000O100O101O0O10000O101N10000O010O10O10O01000O010O10O10O010O10O10O01000O010O01000O001N2N1O2TMVM`Dk2a;XMZDj2f;YMUDi2j;[MQDf2P<^MjCd2U<`MfCa2[<bM`C`2`<cM[C_2d<eMWC\\2j<gMQC[2n<iMmBY2S=jMhBW2Y=lMbBV2]=nM^BS2c=PNXBR2g=RNTBP2l=SNoAn1R>UNiAm1V>n110O10O10O10O010O10O10O010O10000O2O0O2O000O2O001N101N101O0O101N101O0O2O000O2O0O2O001N100O2O001N101O0O2O0O\\B"}, "label": "a brown piece of luggage with a woman ' s hat on top"}]}
{"id": 542173, "image": "COCO_train2014_000000542173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back - end of an ostrich\"."}], "task": "refering", "answer_template": "The \"the back - end of an ostrich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 197, 198, 220, 221, 243], "bbox": [0.4915, 0.25395833333333334, 0.7176875, 0.6394583333333334], "iscrowd": 0, "area": 12886.30035, "rle": {"size": [480, 640], "counts": "mjc45i>4K5K5L2M4M3L3M4M3L3N3L4L3N3L4M2M4M3L4M201O1N101O1O0O2O1O001N2O001O1N101O1O0O2O001O0O101O0O101O0O101O000O2O001N10000O10000O1\\EeMc8[2XGlMf8T2VGRNg8P2SGVNl8j1oF]No8c1mFcNQ9^1iFiNT9a31H8I6J7I7H7O2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O2O0O1O1O1O1O1O1O1O2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3L3N3L3N3L3N3L3N3M2M4M3L5L4K5L4K5L]ed2"}, "label": "the back - end of an ostrich"}]}
{"id": 542173, "image": "COCO_train2014_000000542173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird standing on top of rocks next to a zebra\"."}], "task": "refering", "answer_template": "The \"a bird standing on top of rocks next to a zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 197, 198, 220, 221, 243], "bbox": [0.4915, 0.25395833333333334, 0.7176875, 0.6394583333333334], "iscrowd": 0, "area": 12886.30035, "rle": {"size": [480, 640], "counts": "mjc45i>4K5K5L2M4M3L3M4M3L3N3L4L3N3L4M2M4M3L4M201O1N101O1O0O2O1O001N2O001O1N101O1O0O2O001O0O101O0O101O0O101O000O2O001N10000O10000O1\\EeMc8[2XGlMf8T2VGRNg8P2SGVNl8j1oF]No8c1mFcNQ9^1iFiNT9a31H8I6J7I7H7O2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O2O0O1O1O1O1O1O1O1O2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3L3N3L3N3L3N3L3N3M2M4M3L5L4K5L4K5L]ed2"}, "label": "a bird standing on top of rocks next to a zebra"}]}
{"id": 542173, "image": "COCO_train2014_000000542173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an otrich opening up it ' s feathers on a rock\"."}], "task": "refering", "answer_template": "The \"an otrich opening up it ' s feathers on a rock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 110, 111, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 223, 224, 225, 226, 248], "bbox": [0.633875, 0.21989583333333332, 0.9, 0.5962291666666667], "iscrowd": 0, "area": 14549.13379999999, "rle": {"size": [480, 640], "counts": "gan51n>1N2O1O1O1O1N2O1O2N1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1RCfNU<[1hChNW<Z1fChNY<Y1eCiNZ<X1dCjNZ<Y1dChNY<[1fCfNW<^1hCbNU<a1jC`NS<c1lC^NR<e1lC]NP<f1oC[Nn;i1PDXNn;j1RDVNk;a2M300O1O1O100O1O1O1O2O0O1O1O00O2O0O10001N101N2N1O2N2O1000O100000000O100000O10O10000E;E;D<C=0000000000000O1000000001O1O2WF[KS9f4gFaKW9`4cFgK[9Q5N2N000000000000eNiFmLW9R3nFjLS9T3RGhLo8V3UGfLm8X3XGdLj8Y3[GcLf8[3cG]L^8a3kGWLW8f3c1N2N3M2M3QM`D[2c;`M`D_2b;^MaD`2a;\\MbDc2o;M2O2N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1N3M2O2M2O2M2N3N1N3N1N3M6K9F;Fajm0"}, "label": "an otrich opening up it ' s feathers on a rock"}]}
{"id": 542173, "image": "COCO_train2014_000000542173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ostrich cleaning it ' s wing\"."}], "task": "refering", "answer_template": "The \"an ostrich cleaning it ' s wing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 110, 111, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 223, 224, 225, 226, 248], "bbox": [0.633875, 0.21989583333333332, 0.9, 0.5962291666666667], "iscrowd": 0, "area": 14549.13379999999, "rle": {"size": [480, 640], "counts": "gan51n>1N2O1O1O1O1N2O1O2N1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1RCfNU<[1hChNW<Z1fChNY<Y1eCiNZ<X1dCjNZ<Y1dChNY<[1fCfNW<^1hCbNU<a1jC`NS<c1lC^NR<e1lC]NP<f1oC[Nn;i1PDXNn;j1RDVNk;a2M300O1O1O100O1O1O1O2O0O1O1O00O2O0O10001N101N2N1O2N2O1000O100000000O100000O10O10000E;E;D<C=0000000000000O1000000001O1O2WF[KS9f4gFaKW9`4cFgK[9Q5N2N000000000000eNiFmLW9R3nFjLS9T3RGhLo8V3UGfLm8X3XGdLj8Y3[GcLf8[3cG]L^8a3kGWLW8f3c1N2N3M2M3QM`D[2c;`M`D_2b;^MaD`2a;\\MbDc2o;M2O2N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1N3M2O2M2O2M2N3N1N3N1N3M6K9F;Fajm0"}, "label": "an ostrich cleaning it ' s wing"}]}
{"id": 42463, "image": "COCO_train2014_000000042463.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a gray tee shirt sitting at a table\"."}], "task": "refering", "answer_template": "The \"a man wearing a gray tee shirt sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 45, 46, 47, 48, 63, 64, 65, 66, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 139, 140, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194], "bbox": [0.43386, 0.13602666666666666, 0.8458600000000001, 0.8019466666666667], "iscrowd": 0, "area": 32121.3409, "rle": {"size": [375, 500], "counts": "cf_24^;7I8UELh9S1K5M3N3M4K4M3M3L4M3M3M3L5M200O1O1O1O100O2N1O1O100O1O1O1O2O0O1O1O1O100O1O2N100O1O1O1O101N1O1O1O100bMhL\\LY3b3eMaK\\2]4SNUKo1h4a2N2N2N2O100O1O2N1O1O100O1O1O1O2N100O1O100000001N10000000000O2O000000000O10001O0001O001O000010O01O00001O0000100O1O1O1O2N1O1O100O1O1O1O1O2VKVKc2k4[MVKe2k4XMXKh2h4UMZKk2g4SMZKm2g4PM[KP3f4nL\\KQ3e4lL]KT3d4iL^KW3d4fL]K[3f5_K]J`4]6000001O0000010O00001O00000010O000001O00000010O000001O001O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O2N4L3N2M4]OiG]M[8`2lGWMZ8e2?M4M2M4L3M4M2M4L3L5L3L5SOiE1[:GQF2S:EZF2[Xl0"}, "label": "a man wearing a gray tee shirt sitting at a table"}]}
{"id": 42463, "image": "COCO_train2014_000000042463.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man with red hair and glasses\"."}], "task": "refering", "answer_template": "The \"a young man with red hair and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 45, 46, 47, 48, 63, 64, 65, 66, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 139, 140, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194], "bbox": [0.43386, 0.13602666666666666, 0.8458600000000001, 0.8019466666666667], "iscrowd": 0, "area": 32121.3409, "rle": {"size": [375, 500], "counts": "cf_24^;7I8UELh9S1K5M3N3M4K4M3M3L4M3M3M3L5M200O1O1O1O100O2N1O1O100O1O1O1O2O0O1O1O1O100O1O2N100O1O1O1O101N1O1O1O100bMhL\\LY3b3eMaK\\2]4SNUKo1h4a2N2N2N2O100O1O2N1O1O100O1O1O1O2N100O1O100000001N10000000000O2O000000000O10001O0001O001O000010O01O00001O0000100O1O1O1O2N1O1O100O1O1O1O1O2VKVKc2k4[MVKe2k4XMXKh2h4UMZKk2g4SMZKm2g4PM[KP3f4nL\\KQ3e4lL]KT3d4iL^KW3d4fL]K[3f5_K]J`4]6000001O0000010O00001O00000010O000001O00000010O000001O001O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O2N4L3N2M4]OiG]M[8`2lGWMZ8e2?M4M2M4L3M4M2M4L3L5L3L5SOiE1[:GQF2S:EZF2[Xl0"}, "label": "a young man with red hair and glasses"}]}
{"id": 42463, "image": "COCO_train2014_000000042463.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a striped sleeve\"."}], "task": "refering", "answer_template": "The \"a person with a striped sleeve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 106, 107, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197, 214, 215], "bbox": [0.8408, 0.31165333333333334, 1.0, 0.8946133333333334], "iscrowd": 0, "area": 10415.3044, "rle": {"size": [375, 500], "counts": "URj41d;5J6K5J6K5J6K5J6K5J6K5J6K5J6K3L3N3L3N2M4M2N3M2N3M2M3N3M2N3M2N2N3L3N3M2N2N3M2N3L3N3M2N2N3M2N3L3N2N3M2N3M2M3N3M2NYNUJgMh5Z2ZJeMd5[2_JeM^5\\2cJdMZ5]2iJbMU5_2mJaMP5`2QK`Mm4`2VK`Mg4a2[K^Mc4b2_K^Ma4a2aK_M^4_2eK`M[4_2fKbMY4]2iKbMV4]2mKbMS4]2nKdMQ4Z2RLeMn3Z2SLgMl3W2WLhMi3W2YLiMe3W2\\LiMd3U2_LjMa3U2QM[Mn2c2kMfL`N"}, "label": "a person with a striped sleeve"}]}
{"id": 42463, "image": "COCO_train2014_000000042463.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of a man wearing a black and white shirt\"."}], "task": "refering", "answer_template": "The \"the arm of a man wearing a black and white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 106, 107, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197, 214, 215], "bbox": [0.8408, 0.31165333333333334, 1.0, 0.8946133333333334], "iscrowd": 0, "area": 10415.3044, "rle": {"size": [375, 500], "counts": "URj41d;5J6K5J6K5J6K5J6K5J6K5J6K5J6K3L3N3L3N2M4M2N3M2N3M2M3N3M2N3M2N2N3L3N3M2N2N3M2N3L3N3M2N2N3M2N3L3N2N3M2N3M2M3N3M2NYNUJgMh5Z2ZJeMd5[2_JeM^5\\2cJdMZ5]2iJbMU5_2mJaMP5`2QK`Mm4`2VK`Mg4a2[K^Mc4b2_K^Ma4a2aK_M^4_2eK`M[4_2fKbMY4]2iKbMV4]2mKbMS4]2nKdMQ4Z2RLeMn3Z2SLgMl3W2WLhMi3W2YLiMe3W2\\LiMd3U2_LjMa3U2QM[Mn2c2kMfL`N"}, "label": "the arm of a man wearing a black and white shirt"}]}
{"id": 247265, "image": "COCO_train2014_000000247265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant on the left\"."}], "task": "refering", "answer_template": "The \"the elephant on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280], "bbox": [0.0048125, 0.17901408450704226, 0.303828125, 0.8395305164319249], "iscrowd": 0, "area": 37392.8729, "rle": {"size": [426, 640], "counts": "U]15Q=:G:F:F:F5J7J5K6J5K5M4L3M4L3M3N3L3M3M4L3M4L4L4L4L4L3M4L4L4L4L4M5J;E<mH^Jh5l5lIaJh5k5kI^Jn5`6N2N2O1N2O1N2N101N101N2N101N100O1O101N100O1O100O101N1O100O100O1O1O100O1O1O1O100O10O100000001O000000000O101O00000O100000O010O100O010O1O10O0100O100O010O1000000O1000nN^K\\Ib4_6gK]IX4`6oK]IQ4_6WL]Ii3^6`L^I`3^6gL_IX3]6QM_Io2]6XM`Ih2[6V2L4K5L4M3M3N2M3M3M3N2L4L4L4L4L`H^K`6^4^IkK_6Q4aIWL[6g3gI\\LV6c3kI_LS6`3nI`LR6_3oIaLR6]3oIdLP6\\3PJdLQ6Z3PJfLP6Y3PJiLP6U3QJkLo5T3RJlLo5S3RJmLn5Q3SJoLm5Q3SJoLn5P3RJQMm5P3RJPMo5o2QJPMQ6T3eIQM[6V3VIRMk6f4001O1N102N2N1N3M3N1N3N2M3M2O2M3N1M6VLTHl1R8dM^HV2g7cMaHX2d7`MdHZ2[9J6J6J6J6J5D=[O][i5"}, "label": "the elephant on the left"}]}
{"id": 247265, "image": "COCO_train2014_000000247265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left most elephant\"."}], "task": "refering", "answer_template": "The \"the left most elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280], "bbox": [0.0048125, 0.17901408450704226, 0.303828125, 0.8395305164319249], "iscrowd": 0, "area": 37392.8729, "rle": {"size": [426, 640], "counts": "U]15Q=:G:F:F:F5J7J5K6J5K5M4L3M4L3M3N3L3M3M4L3M4L4L4L4L4L3M4L4L4L4L4M5J;E<mH^Jh5l5lIaJh5k5kI^Jn5`6N2N2O1N2O1N2N101N101N2N101N100O1O101N100O1O100O101N1O100O100O1O1O100O1O1O1O100O10O100000001O000000000O101O00000O100000O010O100O010O1O10O0100O100O010O1000000O1000nN^K\\Ib4_6gK]IX4`6oK]IQ4_6WL]Ii3^6`L^I`3^6gL_IX3]6QM_Io2]6XM`Ih2[6V2L4K5L4M3M3N2M3M3M3N2L4L4L4L4L`H^K`6^4^IkK_6Q4aIWL[6g3gI\\LV6c3kI_LS6`3nI`LR6_3oIaLR6]3oIdLP6\\3PJdLQ6Z3PJfLP6Y3PJiLP6U3QJkLo5T3RJlLo5S3RJmLn5Q3SJoLm5Q3SJoLn5P3RJQMm5P3RJPMo5o2QJPMQ6T3eIQM[6V3VIRMk6f4001O1N102N2N1N3M3N1N3N2M3M2O2M3N1M6VLTHl1R8dM^HV2g7cMaHX2d7`MdHZ2[9J6J6J6J6J5D=[O][i5"}, "label": "the left most elephant"}]}
{"id": 247265, "image": "COCO_train2014_000000247265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with all four legs visable\"."}], "task": "refering", "answer_template": "The \"elephant with all four legs visable\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 244, 245, 246, 249, 250, 251, 267, 268, 269, 272, 273, 274, 275, 290, 291, 292, 295, 296, 297], "bbox": [0.428328125, 0.1592018779342723, 0.9909687500000001, 0.858755868544601], "iscrowd": 0, "area": 68064.14344999999, "rle": {"size": [426, 640], "counts": "\\Wb33n<=D<D<D<D<C=D<D<D;E9F;M2N3M2N3M3M2N3M2N3M3MmNfGYMX8f2kG[MR8d2QH\\Mm7c2VH\\Mi7c2ZH]Md7b2_H^M_7a2dH^M[7a2hH_MV7`2mH_MR7`2QI_Mn6`2TI`Mk6_2VIcMh6\\2ZIdMe6Z2^IfMa6W2cIjM[6S2iImMV6Q2mIoMQ6o1RJRNl5l1XJUNf5h1^JXN`5f1dJZNZ5d1jJ\\NU5a1oJ`Nn4^1WKaNh4^1ZK`Nf4a1\\KZNf4f1]KVNe4j1\\KSNe4n1]KnMd4R2SIWNS1Cl5`3TJ^Ll5d3UJXLl5j3TJTLm5m3TJPLl5R4d12O0O1O100O1O10000010O1O001O1O001O10O010000O01000O01000O010O10000O101O0O101O00001N1000001N10001O0O1010O1O1O001O1O1O001O1O1O0O2O1O2Nb0^Ob0^Ob0`JPJb2a6cLjI\\3h6hKbIX4h701O0000001N10001N10000O2O00001N101O0O101O0O2O001O0O2O000O2O001N100000O010000O100000O001O1N101N2O0O2O1N2I6G:F:F9H9F:F:G9F:F:I710O01O001O001O01O01O001O001O010O001O00001O010O001O001O00010O001O001O001ON3O0000000000010O0000000001O01O0000000001O01O00000001O000[KRJo1m5mMXJR2h5jM\\JV2d5fM`JZ2`5bMdJ^2]5]MgJc2Y5YMkJg2U5UMoJk2Q5QMSKP3l4mLWKS3i4iL[KW3f4eL]K[3e4_L_Ka3c4ZLaKf3_4ULeKk3]4oKgKR4Z4iKiKX4W4cKmK]4U4^KnKc4\\61O001O001O10O01O1O001O001O1O001N2O001O1N101O001N2O001O1N101O1O0O2O001O1O0O2O1O001O1N101O001O1N101O1O0O2O1O001N2\\MPJlMQ6Q2UJkMl5R2[JhMh5U2^JfMc5W2dJdM]5Y2iJcMX5Z2oJ`MS5]2TK^Mm4`2XK\\Mi4a2^KXMe4d2cKVM`4e2hKUMZ4h2lKSMV4i2RLQMP4l2WLnLk3n2[3M3L4J7I6I7J6I7J6J6I7J7H7J6I7JX[2"}, "label": "elephant with all four legs visable"}]}
{"id": 247265, "image": "COCO_train2014_000000247265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult elephant facing towards the left and eating with it ' s trunk\"."}], "task": "refering", "answer_template": "The \"an adult elephant facing towards the left and eating with it ' s trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 244, 245, 246, 249, 250, 251, 267, 268, 269, 272, 273, 274, 275, 290, 291, 292, 295, 296, 297], "bbox": [0.428328125, 0.1592018779342723, 0.9909687500000001, 0.858755868544601], "iscrowd": 0, "area": 68064.14344999999, "rle": {"size": [426, 640], "counts": "\\Wb33n<=D<D<D<D<C=D<D<D;E9F;M2N3M2N3M3M2N3M2N3M3MmNfGYMX8f2kG[MR8d2QH\\Mm7c2VH\\Mi7c2ZH]Md7b2_H^M_7a2dH^M[7a2hH_MV7`2mH_MR7`2QI_Mn6`2TI`Mk6_2VIcMh6\\2ZIdMe6Z2^IfMa6W2cIjM[6S2iImMV6Q2mIoMQ6o1RJRNl5l1XJUNf5h1^JXN`5f1dJZNZ5d1jJ\\NU5a1oJ`Nn4^1WKaNh4^1ZK`Nf4a1\\KZNf4f1]KVNe4j1\\KSNe4n1]KnMd4R2SIWNS1Cl5`3TJ^Ll5d3UJXLl5j3TJTLm5m3TJPLl5R4d12O0O1O100O1O10000010O1O001O1O001O10O010000O01000O01000O010O10000O101O0O101O00001N1000001N10001O0O1010O1O1O001O1O1O001O1O1O0O2O1O2Nb0^Ob0^Ob0`JPJb2a6cLjI\\3h6hKbIX4h701O0000001N10001N10000O2O00001N101O0O101O0O2O001O0O2O000O2O001N100000O010000O100000O001O1N101N2O0O2O1N2I6G:F:F9H9F:F:G9F:F:I710O01O001O001O01O01O001O001O010O001O00001O010O001O001O00010O001O001O001ON3O0000000000010O0000000001O01O0000000001O01O00000001O000[KRJo1m5mMXJR2h5jM\\JV2d5fM`JZ2`5bMdJ^2]5]MgJc2Y5YMkJg2U5UMoJk2Q5QMSKP3l4mLWKS3i4iL[KW3f4eL]K[3e4_L_Ka3c4ZLaKf3_4ULeKk3]4oKgKR4Z4iKiKX4W4cKmK]4U4^KnKc4\\61O001O001O10O01O1O001O001O1O001N2O001O1N101O001N2O001O1N101O1O0O2O001O1O0O2O1O001O1N101O001O1N101O1O0O2O1O001N2\\MPJlMQ6Q2UJkMl5R2[JhMh5U2^JfMc5W2dJdM]5Y2iJcMX5Z2oJ`MS5]2TK^Mm4`2XK\\Mi4a2^KXMe4d2cKVM`4e2hKUMZ4h2lKSMV4i2RLQMP4l2WLnLk3n2[3M3L4J7I6I7J6I7J6J6I7J7H7J6I7JX[2"}, "label": "an adult elephant facing towards the left and eating with it ' s trunk"}]}
{"id": 247265, "image": "COCO_train2014_000000247265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant digging\"."}], "task": "refering", "answer_template": "The \"a baby elephant digging\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 168, 169, 170, 171, 172, 173, 191, 192, 193, 195, 196, 214, 215, 216, 218, 219, 237, 238, 239, 241, 242, 261, 262, 265], "bbox": [0.315984375, 0.45429577464788734, 0.56684375, 0.78556338028169], "iscrowd": 0, "area": 9442.241199999997, "rle": {"size": [426, 640], "counts": "TZd2>i<2N1O2N2M2\\Oe0B>I601O10O0100O0010O001O1O101O001O010O001O001O0010O01O001O1O1000O10OZFYNk7g1SH[Nm7e1PH^No7c1oG_NQ8a1lGcNS8\\1kGgNU8Y1iGiNW8W1fGlNY8U1eGmN[8R1cGQO]8o0aGTO^8l0_GWO`8j0]GYOc8f0\\G\\Od8e0XG^Oh8f0PG^OP9f0iF^OU9g0cF]O]9X2000N1O2N2N2N2N1N3N2N2N2N1O2N2N2N2\\Nd1G9H8G9Gmi8J_VG4L4L3M4K4M4L4L3L5L4L3L5K5K4L5M3M2N3N2M4a0_O;D0102M2N2cDdNf:Q2N2N2O1N2N2O1N2N2N2O1N2N2N1O2N2N2N2NB[FZMd9`2eF^M[9_2jF_MV9^2m0L4M3L4M5H9D;D=E;H7J7Hiob3"}, "label": "a baby elephant digging"}]}
{"id": 247265, "image": "COCO_train2014_000000247265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small elephant stands between two larger elephants\"."}], "task": "refering", "answer_template": "The \"a small elephant stands between two larger elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 168, 169, 170, 171, 172, 173, 191, 192, 193, 195, 196, 214, 215, 216, 218, 219, 237, 238, 239, 241, 242, 261, 262, 265], "bbox": [0.315984375, 0.45429577464788734, 0.56684375, 0.78556338028169], "iscrowd": 0, "area": 9442.241199999997, "rle": {"size": [426, 640], "counts": "TZd2>i<2N1O2N2M2\\Oe0B>I601O10O0100O0010O001O1O101O001O010O001O001O0010O01O001O1O1000O10OZFYNk7g1SH[Nm7e1PH^No7c1oG_NQ8a1lGcNS8\\1kGgNU8Y1iGiNW8W1fGlNY8U1eGmN[8R1cGQO]8o0aGTO^8l0_GWO`8j0]GYOc8f0\\G\\Od8e0XG^Oh8f0PG^OP9f0iF^OU9g0cF]O]9X2000N1O2N2N2N2N1N3N2N2N2N1O2N2N2N2\\Nd1G9H8G9Gmi8J_VG4L4L3M4K4M4L4L3L5L4L3L5K5K4L5M3M2N3N2M4a0_O;D0102M2N2cDdNf:Q2N2N2O1N2N2O1N2N2N2O1N2N2N1O2N2N2N2NB[FZMd9`2eF^M[9_2jF_MV9^2m0L4M3L4M5H9D;D=E;H7J7Hiob3"}, "label": "a small elephant stands between two larger elephants"}]}
{"id": 189924, "image": "COCO_train2014_000000189924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person stand behind obama\"."}], "task": "refering", "answer_template": "The \"a person stand behind obama\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 73, 74, 86, 87, 88, 99, 100, 101, 102, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 138, 139, 140, 141, 153, 154], "bbox": [0.63008356545961, 0.22069767441860466, 0.9294150417827297, 0.69276955602537], "iscrowd": 0, "area": 11708.767249999999, "rle": {"size": [473, 359], "counts": "`eX31e>3N2M4M2M3N3L3L4L5K4L4\\KkN`KZ1`4POQKT1P5]OYJh0h5ZOoIm0Q6UOeIR1\\6QOZIU1g6mNPIZ1P7hNoGSO[O\\2g8cN^Ga2c8k10O1O100O1O100O1O100O1O100ZLgFj1\\9nMiFR2\\9eMiF[2[9]MhFc2]9WMeFh2^9SMdFn2\\9PMfFo2\\9mLgFR3Z9lLgFU3Y9hLjFW3Y9eLhF[3[:1O0001O01O00001O01O0001O01O010O001O010O0010O0001O010O0010O01O0010O01O01O010O100O010O10O10O10O10O100O01000O010000O10OK6C=C=C=DoZ;"}, "label": "a person stand behind obama"}]}
{"id": 189924, "image": "COCO_train2014_000000189924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing behind the president\"."}], "task": "refering", "answer_template": "The \"a man standing behind the president\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 73, 74, 86, 87, 88, 99, 100, 101, 102, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 138, 139, 140, 141, 153, 154], "bbox": [0.63008356545961, 0.22069767441860466, 0.9294150417827297, 0.69276955602537], "iscrowd": 0, "area": 11708.767249999999, "rle": {"size": [473, 359], "counts": "`eX31e>3N2M4M2M3N3L3L4L5K4L4\\KkN`KZ1`4POQKT1P5]OYJh0h5ZOoIm0Q6UOeIR1\\6QOZIU1g6mNPIZ1P7hNoGSO[O\\2g8cN^Ga2c8k10O1O100O1O100O1O100O1O100ZLgFj1\\9nMiFR2\\9eMiF[2[9]MhFc2]9WMeFh2^9SMdFn2\\9PMfFo2\\9mLgFR3Z9lLgFU3Y9hLjFW3Y9eLhF[3[:1O0001O01O00001O01O0001O01O010O001O010O0010O0001O010O0010O01O0010O01O01O010O100O010O10O10O10O10O100O01000O010000O10OK6C=C=C=DoZ;"}, "label": "a man standing behind the president"}]}
{"id": 189924, "image": "COCO_train2014_000000189924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"our preident obama\"."}], "task": "refering", "answer_template": "The \"our preident obama\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 41, 42, 43, 44, 45, 46, 47, 54, 55, 56, 57, 58, 59, 60, 67, 68, 69, 70, 71, 72, 73, 79, 80, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 97, 98, 99, 106, 107, 108, 109, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 131, 132, 133, 134, 135, 136, 137, 138, 139, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.06824512534818941, 0.11374207188160676, 0.925849582172702, 0.9020295983086681], "iscrowd": 0, "area": 76040.29025, "rle": {"size": [473, 359], "counts": "mj;]3\\;<D000001O0001O0O10000O10000O10000O10000O100bLULnKk3o3YLoKh3n3\\LQLd3k3bLSL^3k3eLSL\\3m3fLQLZ3n3hLQLX3o3jLnKW3Q4kLnKU3R4lLmKT3R4oLkKR3U4oLjKQ3V4QMgKP3X4TMeKl2[4XM`Ki2_4[M]Kf2c4^MXKc2g4bMTK_2l4eMoJ\\2Q5gMkJZ2T5hMjJY2S5kMkJV2k4UNSKl1c4`NZKa1\\4jNbKW1T4TOjKm0m3\\OQLf0e3DZL=c3F\\L;b3F^L;`3G_L:_3H`L9^3IaL8]3JbL7\\3JdL7Z3KeL6Y3LfL5X3NfL3X30VLa0h3Y500O100O2O0O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O10000000000O1000000000000000000O100000001O000000000O1000000000000000000O1000000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O0010O01O001O001O001O001O001O001O001O001O0dKQLnMR4j1TLUNP4c1TL]NP4Z1ULfNP4Q1ULnNo3i0VLWOo3`0UL@o37WLHm30WL0n3FWL:m3^OXLa0m3UOXLk0l3kNZLU1j3_N_L`1f3TNaLl1c3iMdLW2`3]MhLc2]3QMkLn2Y3gLnLY3Y701O00001O00001O00001O00001O001O001O1O001O10O01O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O010O1O001O[T<"}, "label": "our preident obama"}]}
{"id": 189924, "image": "COCO_train2014_000000189924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"president barack obama\"."}], "task": "refering", "answer_template": "The \"president barack obama\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 41, 42, 43, 44, 45, 46, 47, 54, 55, 56, 57, 58, 59, 60, 67, 68, 69, 70, 71, 72, 73, 79, 80, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 97, 98, 99, 106, 107, 108, 109, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 131, 132, 133, 134, 135, 136, 137, 138, 139, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.06824512534818941, 0.11374207188160676, 0.925849582172702, 0.9020295983086681], "iscrowd": 0, "area": 76040.29025, "rle": {"size": [473, 359], "counts": "mj;]3\\;<D000001O0001O0O10000O10000O10000O10000O100bLULnKk3o3YLoKh3n3\\LQLd3k3bLSL^3k3eLSL\\3m3fLQLZ3n3hLQLX3o3jLnKW3Q4kLnKU3R4lLmKT3R4oLkKR3U4oLjKQ3V4QMgKP3X4TMeKl2[4XM`Ki2_4[M]Kf2c4^MXKc2g4bMTK_2l4eMoJ\\2Q5gMkJZ2T5hMjJY2S5kMkJV2k4UNSKl1c4`NZKa1\\4jNbKW1T4TOjKm0m3\\OQLf0e3DZL=c3F\\L;b3F^L;`3G_L:_3H`L9^3IaL8]3JbL7\\3JdL7Z3KeL6Y3LfL5X3NfL3X30VLa0h3Y500O100O2O0O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O10000000000O1000000000000000000O100000001O000000000O1000000000000000000O1000000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O0010O01O001O001O001O001O001O001O001O001O0dKQLnMR4j1TLUNP4c1TL]NP4Z1ULfNP4Q1ULnNo3i0VLWOo3`0UL@o37WLHm30WL0n3FWL:m3^OXLa0m3UOXLk0l3kNZLU1j3_N_L`1f3TNaLl1c3iMdLW2`3]MhLc2]3QMkLn2Y3gLnLY3Y701O00001O00001O00001O00001O001O001O1O001O10O01O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O010O1O001O[T<"}, "label": "president barack obama"}]}
{"id": 132585, "image": "COCO_train2014_000000132585.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the breadstick on the top of the others\"."}], "task": "refering", "answer_template": "The \"the breadstick on the top of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335], "bbox": [0.226546875, 0.17386792452830188, 0.79465625, 0.9873584905660377], "iscrowd": 0, "area": 73347.2335, "rle": {"size": [424, 640], "counts": "fTl14T=6I6J7I7I3M3M3N2M4L3M3M3M3M4M2M3M3M4L3M3M3N2M4L3N2N2N2N3M2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N2M4M2N2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N2M4M2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2M4M2N2N2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2M4M2N00000000000000O10000000000O1000O100000O1000000000000O10000000000O100000001O0O10001O0000001N1000001O000OO2N2N2N2N2N1011O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O2M2O2N1O2N1O2M2O2N1O1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1O2M2O2N1N3N1O2M3N2N3L3N3M3L3N3M2M4M2N3L3N3M2M4M2N3L4M2N3L3N3M2N3L3N3M1N3N1O2M3N1O2M3N1O2M3N1O2M2O2N2M2O2N2M2O2N2M2O2N1N3N2N1O2M3N1O2M3N1O2M2O2N1N2O1O2M2O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1O2M2O1O1N3N1O1N2O2N1N2O2N1N2O1O2M2O1O1N3N1O1N2O2N2M3N2M3M3M3M3M3M3M3Mjle1"}, "label": "the breadstick on the top of the others"}]}
{"id": 132585, "image": "COCO_train2014_000000132585.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the very top piece of food\"."}], "task": "refering", "answer_template": "The \"the very top piece of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335], "bbox": [0.226546875, 0.17386792452830188, 0.79465625, 0.9873584905660377], "iscrowd": 0, "area": 73347.2335, "rle": {"size": [424, 640], "counts": "fTl14T=6I6J7I7I3M3M3N2M4L3M3M3M3M4M2M3M3M4L3M3M3N2M4L3N2N2N2N3M2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N2M4M2N2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N2M4M2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2M4M2N2N2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2M4M2N00000000000000O10000000000O1000O100000O1000000000000O10000000000O100000001O0O10001O0000001N1000001O000OO2N2N2N2N2N1011O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O2M2O2N1O2N1O2M2O2N1O1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1O2M2O2N1N3N1O2M3N2N3L3N3M3L3N3M2M4M2N3L3N3M2M4M2N3L4M2N3L3N3M2N3L3N3M1N3N1O2M3N1O2M3N1O2M3N1O2M2O2N2M2O2N2M2O2N2M2O2N1N3N2N1O2M3N1O2M3N1O2M2O2N1N2O1O2M2O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1O2M2O1O1N3N1O1N2O2N1N2O2N1N2O1O2M2O1O1N3N1O1N2O2N2M3N2M3M3M3M3M3M3M3Mjle1"}, "label": "the very top piece of food"}]}
{"id": 132585, "image": "COCO_train2014_000000132585.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza on the bottom closest to the camera\"."}], "task": "refering", "answer_template": "The \"the pizza on the bottom closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.044625, 0.6065094339622642, 0.42907812500000003, 0.9821698113207549], "iscrowd": 0, "area": 25447.235850000005, "rle": {"size": [424, 640], "counts": "f\\<1W=0O101M2L4L4L5L3L4L4L5L3L4L4M4K4L4L4M4K4L4O10001N100O101O0O10001N100O101O0O100O2O000O101N10000O101O0O100O2O000O101N10000O2O000O1N3L3N2N2N2M4M2N2N2M3N2N2N2M3N2N2N2M3N2N2N2N2O1000000O10000O10000O1000000O10000O10000O1000000O1001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O002N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2NbTg4"}, "label": "the pizza on the bottom closest to the camera"}]}
{"id": 132585, "image": "COCO_train2014_000000132585.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cheesey breadstick on the bottom closest to the camera\"."}], "task": "refering", "answer_template": "The \"the cheesey breadstick on the bottom closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.044625, 0.6065094339622642, 0.42907812500000003, 0.9821698113207549], "iscrowd": 0, "area": 25447.235850000005, "rle": {"size": [424, 640], "counts": "f\\<1W=0O101M2L4L4L5L3L4L4L5L3L4L4M4K4L4L4M4K4L4O10001N100O101O0O10001N100O101O0O100O2O000O101N10000O101O0O100O2O000O101N10000O2O000O1N3L3N2N2N2M4M2N2N2M3N2N2N2M3N2N2N2M3N2N2N2N2O1000000O10000O10000O1000000O10000O10000O1000000O1001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O002N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2NbTg4"}, "label": "the cheesey breadstick on the bottom closest to the camera"}]}
{"id": 558570, "image": "COCO_train2014_000000558570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis ball\"."}], "task": "refering", "answer_template": "The \"tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 179], "bbox": [0.7076718750000001, 0.30133177570093456, 1.0, 0.4808878504672897], "iscrowd": 0, "area": 8855.667950000003, "rle": {"size": [428, 640], "counts": "b`m52W=5K5L4M3L4M2N1O1O10000O2O000O10001N10000O10001O000O100000000O0100O100O010O100O10O010bf1XOnYN:G7K5J6J6L40001N10000O100O10000O2N100O000001O00001O0001O010O1O100O100O100O10000O100O100O10000O100O100O100O100000000O10000000000000000O10000000000000000O10000000000000000000000000000000000000O100000000O1000000O1000000O1L4L4L4L4L4O1O1N2O1O1O1O1O1O1O100O1O101N1O1O100O1O1M3N2M4M9GhK"}, "label": "tennis ball"}]}
{"id": 558570, "image": "COCO_train2014_000000558570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow tennis ball in the air\"."}], "task": "refering", "answer_template": "The \"the yellow tennis ball in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 179], "bbox": [0.7076718750000001, 0.30133177570093456, 1.0, 0.4808878504672897], "iscrowd": 0, "area": 8855.667950000003, "rle": {"size": [428, 640], "counts": "b`m52W=5K5L4M3L4M2N1O1O10000O2O000O10001N10000O10001O000O100000000O0100O100O010O100O10O010bf1XOnYN:G7K5J6J6L40001N10000O100O10000O2N100O000001O00001O0001O010O1O100O100O100O10000O100O100O10000O100O100O100O100000000O10000000000000000O10000000000000000O10000000000000000000000000000000000000O100000000O1000000O1000000O1L4L4L4L4L4O1O1N2O1O1O1O1O1O1O100O1O101N1O1O100O1O1M3N2M4M9GhK"}, "label": "the yellow tennis ball in the air"}]}
{"id": 558570, "image": "COCO_train2014_000000558570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue and white tennis outfit\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue and white tennis outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 191, 192, 193, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 261, 262, 284, 285, 286, 308, 309], "bbox": [0.224625, 0.254088785046729, 0.4614062499999999, 0.8901869158878505], "iscrowd": 0, "area": 14258.138049999996, "rle": {"size": [428, 640], "counts": "``l13V=3L5L3M4[Od0I701O001O01O01O001O01O01O0000N3M2N2N2K5K5N101N2N2N2O1N101O100O10000O10000000000001N1000000000000000000000000gNDaE=[:JbE6[:ObE2Z:4dELX::fEF\\8MUHc0\\OAk7e0aH0A[Oh7Q1]HKHTOk7V1THNLmNP8Z1kG11eNS8d3iG]L\\7[O[IQ5d6QKZIP5e6SKYIm4f6UKXIl4g6VKXIj4g6YKWIf4i6\\KUIe4i6^KVIb4i6aKTI`4k6bKSI_4m6bKRI^4m6eKPI\\4P7eKnHX4V7iKiHS4[7k05K4L4L4L33N3L4L3N3L4L4M3M3dJkI[3X6_LlIa3W6XLlIh3W6SLlIl3W6nKmIR4U6hKnIX4`6nJmIQ5W70100OfJ[Km1e4fMTLo1l3oMULQ2k3nMWLQ2i3nMYLR2f3lM]LS2d3kM^LU2a3jMaLU2`3iMbLV2_3gMcLY2_3dMcL[2^3cMdL\\2]3aMfL;aN]Od5SOVLR1fNHT5VOWLd0UO3c4ZOYL5D?R4]OZLH2h0c3A\\LC7i0\\3F]L^O=h0V3J^L[Oa0h0P3N`LWOe0h0k22`LTOi0g0f26LH3:ME3;ND1=0B0?0@Oa02^ONc02\\OMe03ZONf03XONi02UOOk02SO0m00QO1o00oN2P1OnN4P1MnN5P1MoN5o0U6N3N2M3MeQ`4"}, "label": "a woman wearing a blue and white tennis outfit"}]}
{"id": 558570, "image": "COCO_train2014_000000558570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman playing tennis on a court\"."}], "task": "refering", "answer_template": "The \"woman playing tennis on a court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 191, 192, 193, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 261, 262, 284, 285, 286, 308, 309], "bbox": [0.224625, 0.254088785046729, 0.4614062499999999, 0.8901869158878505], "iscrowd": 0, "area": 14258.138049999996, "rle": {"size": [428, 640], "counts": "``l13V=3L5L3M4[Od0I701O001O01O01O001O01O01O0000N3M2N2N2K5K5N101N2N2N2O1N101O100O10000O10000000000001N1000000000000000000000000gNDaE=[:JbE6[:ObE2Z:4dELX::fEF\\8MUHc0\\OAk7e0aH0A[Oh7Q1]HKHTOk7V1THNLmNP8Z1kG11eNS8d3iG]L\\7[O[IQ5d6QKZIP5e6SKYIm4f6UKXIl4g6VKXIj4g6YKWIf4i6\\KUIe4i6^KVIb4i6aKTI`4k6bKSI_4m6bKRI^4m6eKPI\\4P7eKnHX4V7iKiHS4[7k05K4L4L4L33N3L4L3N3L4L4M3M3dJkI[3X6_LlIa3W6XLlIh3W6SLlIl3W6nKmIR4U6hKnIX4`6nJmIQ5W70100OfJ[Km1e4fMTLo1l3oMULQ2k3nMWLQ2i3nMYLR2f3lM]LS2d3kM^LU2a3jMaLU2`3iMbLV2_3gMcLY2_3dMcL[2^3cMdL\\2]3aMfL;aN]Od5SOVLR1fNHT5VOWLd0UO3c4ZOYL5D?R4]OZLH2h0c3A\\LC7i0\\3F]L^O=h0V3J^L[Oa0h0P3N`LWOe0h0k22`LTOi0g0f26LH3:ME3;ND1=0B0?0@Oa02^ONc02\\OMe03ZONf03XONi02UOOk02SO0m00QO1o00oN2P1OnN4P1MnN5P1MoN5o0U6N3N2M3MeQ`4"}, "label": "woman playing tennis on a court"}]}
{"id": 345578, "image": "COCO_train2014_000000345578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"groom in pink tie next to woman\"."}], "task": "refering", "answer_template": "The \"groom in pink tie next to woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 141, 142, 143, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 347, 348, 349, 350, 351, 371, 372, 373, 374], "bbox": [0.091375, 0.23731249999999998, 0.323609375, 0.9885833333333333], "iscrowd": 0, "area": 36298.573650000006, "rle": {"size": [480, 640], "counts": "c_k01i>7I8I6I8H7I8H7J;D>B>B>B>C<C=C=D=G8H8H8H8H9G8I7H8H9G8H8H8H2N1O1O100O1O1O1O1O1O1O1O1O100O1O1O2N1O1O1O1O1[Ne100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1001O010O001O001O00001O001O001O001O001O00001O001O0SGQLi6P4SIULk6l3QIXLn6m3iHYLU7k3dHZLZ7j3_H[L_7j3PHcLo7X5O001O1O00100O001O1G8H9G8G:G9G8H=Bc0@?D=D;D=C<E<C<D=D;C>A<E4L5J5N2O2M2N3K4L5K4LfdZ6"}, "label": "groom in pink tie next to woman"}]}
{"id": 345578, "image": "COCO_train2014_000000345578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with coat and suit beside a lady\"."}], "task": "refering", "answer_template": "The \"a man with coat and suit beside a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 141, 142, 143, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 347, 348, 349, 350, 351, 371, 372, 373, 374], "bbox": [0.091375, 0.23731249999999998, 0.323609375, 0.9885833333333333], "iscrowd": 0, "area": 36298.573650000006, "rle": {"size": [480, 640], "counts": "c_k01i>7I8I6I8H7I8H7J;D>B>B>B>C<C=C=D=G8H8H8H8H9G8I7H8H9G8H8H8H2N1O1O100O1O1O1O1O1O1O1O1O100O1O1O2N1O1O1O1O1[Ne100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1001O010O001O001O00001O001O001O001O001O00001O001O0SGQLi6P4SIULk6l3QIXLn6m3iHYLU7k3dHZLZ7j3_H[L_7j3PHcLo7X5O001O1O00100O001O1G8H9G8G:G9G8H=Bc0@?D=D;D=C<E<C<D=D;C>A<E4L5J5N2O2M2N3K4L5K4LfdZ6"}, "label": "a man with coat and suit beside a lady"}]}
{"id": 345578, "image": "COCO_train2014_000000345578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the white wedding dress\"."}], "task": "refering", "answer_template": "The \"the woman in the white wedding dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 352, 353], "bbox": [0.301828125, 0.2739375, 0.57184375, 0.9496041666666667], "iscrowd": 0, "area": 30709.737699999994, "rle": {"size": [480, 640], "counts": "`gj2;b>6J5K4K5M2M4L4M2M3M4L3M3M4M2N2N3M2oGQNo2Q2RLoNk3S1oKQOQ4Q1jKROV4Q1cKSO^4n0]KVOc4k0XKXOi4i0SKYOn4h0kJ^OU5c0eJA[5a0PJ2P60^Ib0a6_OTIk0m6VOjHQ1V7\\3010O001O00100O000001O0000000O2N101N1O101A>J7J5K5M4M2N3N1O1010O010O10O010O010O010O0100O010O011N10000O101N100O101N100O10001O0O101O1O0O2O1O1O1N2O001O1N2O1O2N1O2N2N2N3M3M3M3M3M3M8Hi0VO4VOaGmJa8S4h1YOe00O01O1O001O1N1O2O0O2N2N1O2N2N2N2M3N2N3M2M3N2N2N2N2M3N2N2N2M4M2N2N3M2M3N3M3M2N3L3N3M2M4M2M3N3L3N3LfUP4"}, "label": "the woman in the white wedding dress"}]}
{"id": 345578, "image": "COCO_train2014_000000345578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on the right screen\"."}], "task": "refering", "answer_template": "The \"a woman on the right screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 352, 353], "bbox": [0.301828125, 0.2739375, 0.57184375, 0.9496041666666667], "iscrowd": 0, "area": 30709.737699999994, "rle": {"size": [480, 640], "counts": "`gj2;b>6J5K4K5M2M4L4M2M3M4L3M3M4M2N2N3M2oGQNo2Q2RLoNk3S1oKQOQ4Q1jKROV4Q1cKSO^4n0]KVOc4k0XKXOi4i0SKYOn4h0kJ^OU5c0eJA[5a0PJ2P60^Ib0a6_OTIk0m6VOjHQ1V7\\3010O001O00100O000001O0000000O2N101N1O101A>J7J5K5M4M2N3N1O1010O010O10O010O010O010O0100O010O011N10000O101N100O101N100O10001O0O101O1O0O2O1O1O1N2O001O1N2O1O2N1O2N2N2N3M3M3M3M3M3M8Hi0VO4VOaGmJa8S4h1YOe00O01O1O001O1N1O2O0O2N2N1O2N2N2N2M3N2N3M2M3N2N2N2N2M3N2N2N2M4M2N2N3M2M3N3M3M2N3L3N3M2M4M2M3N3L3N3LfUP4"}, "label": "a woman on the right screen"}]}
{"id": 214507, "image": "COCO_train2014_000000214507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that is bending over and situated to the left of a rhinoceros\"."}], "task": "refering", "answer_template": "The \"a zebra that is bending over and situated to the left of a rhinoceros\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237], "bbox": [0.190015625, 0.2622291666666667, 0.44057812500000004, 0.6411458333333333], "iscrowd": 0, "area": 16539.095599999997, "rle": {"size": [480, 640], "counts": "V]i14i>4N2O1O1O1O1O2N2nBB^;a0PDOo;b1O1O1O1O1O2N1O1O1M3L4K5L3L4M3L4M3L4N2N0O100000000O10O100000O10000000L4K5KXEkLR:o2TFSMk9e2\\F\\Mc9^2cFdM[9W2iFkMV9T2kFmMT9R2mFPNQ9o1PGRNo8R2mFPNR9T2hFnMW9V2dFlM[9X2`FkM^9Z2\\FhMd9[2VFhMi9h3O01O0000001O02N2N3N1N2O3M3M3M4L2M10000000SNmFlMS9l1ZGPNf8f1iGUNW8a1WHZNj7e1i2000O100000000O100000O1000O100000000O01000000000O010000001N2O2eDmMe9U2gE]NY:U30000O1000O1N2M3M3M2L5K5K5K5L4KbEoLW9n2iFVMV9l2fFWMY9j2cFZM]9g2]F]Mc9e2VF`Mj9a2PFdMP:^2iEgMW:Z2cEkM^:[3000O101O0K5E=C=F:K4L5K5K5L3L5K5^Nb1AheW5"}, "label": "a zebra that is bending over and situated to the left of a rhinoceros"}]}
{"id": 214507, "image": "COCO_train2014_000000214507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing behind a wire fence eating grass next to a rhinoceros\"."}], "task": "refering", "answer_template": "The \"a zebra standing behind a wire fence eating grass next to a rhinoceros\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237], "bbox": [0.190015625, 0.2622291666666667, 0.44057812500000004, 0.6411458333333333], "iscrowd": 0, "area": 16539.095599999997, "rle": {"size": [480, 640], "counts": "V]i14i>4N2O1O1O1O1O2N2nBB^;a0PDOo;b1O1O1O1O1O2N1O1O1M3L4K5L3L4M3L4M3L4N2N0O100000000O10O100000O10000000L4K5KXEkLR:o2TFSMk9e2\\F\\Mc9^2cFdM[9W2iFkMV9T2kFmMT9R2mFPNQ9o1PGRNo8R2mFPNR9T2hFnMW9V2dFlM[9X2`FkM^9Z2\\FhMd9[2VFhMi9h3O01O0000001O02N2N3N1N2O3M3M3M4L2M10000000SNmFlMS9l1ZGPNf8f1iGUNW8a1WHZNj7e1i2000O100000000O100000O1000O100000000O01000000000O010000001N2O2eDmMe9U2gE]NY:U30000O1000O1N2M3M3M2L5K5K5K5L4KbEoLW9n2iFVMV9l2fFWMY9j2cFZM]9g2]F]Mc9e2VF`Mj9a2PFdMP:^2iEgMW:Z2cEkM^:[3000O101O0K5E=C=F:K4L5K5K5L3L5K5^Nb1AheW5"}, "label": "a zebra standing behind a wire fence eating grass next to a rhinoceros"}]}
{"id": 34285, "image": "COCO_train2014_000000034285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with its head up\"."}], "task": "refering", "answer_template": "The \"a horse with its head up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 237, 238, 239, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330], "bbox": [0.1856875, 0.602887323943662, 0.43851562499999996, 0.9891549295774648], "iscrowd": 0, "area": 14581.953600000006, "rle": {"size": [426, 640], "counts": "^ka17n<:E;F9G:F9H6M3N2M3N2M3M3Na0^O100O100O100O1O1000000000000000O10000000003M3M4L9Gc0]O8H9Gn0RO4L3M3MYY37TfL`0H8I7H8I6J3M4L3M4M2O2M2N3N2M:F2O1N2N2M3M3M3L4O1N2O1O1O1N2O1O1O1N2O1O1N2O0001O01O2N2N2N1O1O100XOiK\\HX4d7iKZHX4f7iKXHX4h7iKVHX4i7jKUHW4k7jKSHW4m7jKQHW4n7kKQHU4o7lKoGU4Q8lKmGU4S8mKjGT4W8?1O001O1O1O1O1O001O1O1O000M2L5SN]GnNh8o0]GjNh8R1_GgNe8V1aGbNd8Z1cG^Nb8_1dG[N_8b1fGXN^8e1hGUNZ8m1a12N1O1O1M3N2M3N2O1O1O100O00iM[Ei1f:TN]Ek1c:RNbEl1^:RNeEm1m:\\NjDR1X;lNlDP1U;nNPEo0P;POSEm0o:POVEl0k:SOYEe0l:YOYE<R<D[Qe4"}, "label": "a horse with its head up"}]}
{"id": 34285, "image": "COCO_train2014_000000034285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse on the left of the group of horses\"."}], "task": "refering", "answer_template": "The \"horse on the left of the group of horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 237, 238, 239, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330], "bbox": [0.1856875, 0.602887323943662, 0.43851562499999996, 0.9891549295774648], "iscrowd": 0, "area": 14581.953600000006, "rle": {"size": [426, 640], "counts": "^ka17n<:E;F9G:F9H6M3N2M3N2M3M3Na0^O100O100O100O1O1000000000000000O10000000003M3M4L9Gc0]O8H9Gn0RO4L3M3MYY37TfL`0H8I7H8I6J3M4L3M4M2O2M2N3N2M:F2O1N2N2M3M3M3L4O1N2O1O1O1N2O1O1O1N2O1O1N2O0001O01O2N2N2N1O1O100XOiK\\HX4d7iKZHX4f7iKXHX4h7iKVHX4i7jKUHW4k7jKSHW4m7jKQHW4n7kKQHU4o7lKoGU4Q8lKmGU4S8mKjGT4W8?1O001O1O1O1O1O001O1O1O000M2L5SN]GnNh8o0]GjNh8R1_GgNe8V1aGbNd8Z1cG^Nb8_1dG[N_8b1fGXN^8e1hGUNZ8m1a12N1O1O1M3N2M3N2O1O1O100O00iM[Ei1f:TN]Ek1c:RNbEl1^:RNeEm1m:\\NjDR1X;lNlDP1U;nNPEo0P;POSEm0o:POVEl0k:SOYEe0l:YOYE<R<D[Qe4"}, "label": "horse on the left of the group of horses"}]}
{"id": 34285, "image": "COCO_train2014_000000034285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse is wearing a flower wreath on it ' s head\"."}], "task": "refering", "answer_template": "The \"the horse is wearing a flower wreath on it ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 266, 267, 268, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338, 339], "bbox": [0.580984375, 0.6584037558685446, 0.745953125, 0.9881455399061034], "iscrowd": 0, "area": 9704.040400000002, "rle": {"size": [426, 640], "counts": "QRk44S=7I6J6K5kE[O^7j0VH@h7j0gG]OX8P1UGXOj8T1dFRO[9`2O1N101N101N101N100OVLfFf3Z9ZLhFd3W9]LlF`3T9`LmF_3R9bLnF^3R9bLnF^3Q9cLoF]3Q9cLoF]3Q9cLoF]3P9dLPG\\3P9dLPG\\3o8eLQG[3o8eLQG[3n8fLRGZ3n8fLRGZ3m8fLTGZ3l8dLVG\\3i8dLXG\\3h8bLZG^3e8aL\\G`3d8_L\\Gb3c8]L^Gd3b8[L^Gf3a8YLWG3Ne3l8dLPG^3m8a000001O00001O00001O0RMPGb1Q9]NPGb1P9]NRGb1n8^NTG_1n8_NTG`1l8`NUG_1k8`NWG_1j8`NXG^1i8`NYG_1i8_NXG`1j8]NXGb1k8[NWGc1k8ZNWGe1l8XNUGg1m8VNUGi1m8UNTGj1o8RNSGm1o8QNRGn1T9kMnFT2U:O1O001O1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M5K4L5K4K6Ko\\S2"}, "label": "the horse is wearing a flower wreath on it ' s head"}]}
{"id": 34285, "image": "COCO_train2014_000000034285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front most palomino horse with strip and flowers\"."}], "task": "refering", "answer_template": "The \"front most palomino horse with strip and flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 266, 267, 268, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338, 339], "bbox": [0.580984375, 0.6584037558685446, 0.745953125, 0.9881455399061034], "iscrowd": 0, "area": 9704.040400000002, "rle": {"size": [426, 640], "counts": "QRk44S=7I6J6K5kE[O^7j0VH@h7j0gG]OX8P1UGXOj8T1dFRO[9`2O1N101N101N101N100OVLfFf3Z9ZLhFd3W9]LlF`3T9`LmF_3R9bLnF^3R9bLnF^3Q9cLoF]3Q9cLoF]3Q9cLoF]3P9dLPG\\3P9dLPG\\3o8eLQG[3o8eLQG[3n8fLRGZ3n8fLRGZ3m8fLTGZ3l8dLVG\\3i8dLXG\\3h8bLZG^3e8aL\\G`3d8_L\\Gb3c8]L^Gd3b8[L^Gf3a8YLWG3Ne3l8dLPG^3m8a000001O00001O00001O0RMPGb1Q9]NPGb1P9]NRGb1n8^NTG_1n8_NTG`1l8`NUG_1k8`NWG_1j8`NXG^1i8`NYG_1i8_NXG`1j8]NXGb1k8[NWGc1k8ZNWGe1l8XNUGg1m8VNUGi1m8UNTGj1o8RNSGm1o8QNRGn1T9kMnFT2U:O1O001O1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M5K4L5K4K6Ko\\S2"}, "label": "front most palomino horse with strip and flowers"}]}
{"id": 574957, "image": "COCO_train2014_000000574957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a blue shirt\"."}], "task": "refering", "answer_template": "The \"a boy wearing a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.3909375, 0.1993125, 0.986578125, 1.0], "iscrowd": 0, "area": 81017.30904999998, "rle": {"size": [480, 640], "counts": "h`e33m>6I1O2N1O1O100O1O1O1O1O10000000010O0000000000000001O0000000001O0001O0001O0000001O00010O0000001O00000010O000001O00001O01O0001^D]Of8b0ZG^Od8d0\\G\\Oc8f0[G[Oc8h0\\GXOb8k0]GTOc8m0\\GTOb8o0]GQOb8Q1mFjN_N5b:T1nFgN_N6LJ[:[1XGfNaN5KKZ:]1YGdNaN3MMW:]1[GcN`N4NLU:`1[GaNbN3NLU:a1ZG`NcN3NMT:a1YG`NeN2NMT:b1XG_NeN2ONT:b1WG^NfN2OOS:b1VG^NhN1OOS:b1VG^NgN20NS:c1TG^NiN10OQ:d1UG\\NjN1O0R:d1SG\\NkN10OR:e1RG[NlN100Q:e1RGZNmN100Q:g1oFYNoN110P:h1nFWNQO110P:i1lFWNSO001Q:j1jFUNUO002P:k1hFTNWO010Q:n1eFRNYO010P:Q2dFoM[O010P:S2aFnM]O02OP:U2_FlM_O010Q:V2\\FkMBO10Q:W2[FkMAO3OQ:Y2YFiMCO3OP:\\2XFfMEO3Ni9g2]F\\MF04Mh9i2]FZMG03N2E^9T3eFYMH03N1F^9U3eFXMG05M1F]9X3dFUMIO6NOG^9W3dFUMIO6NNH^9X3dFSMJO5OOG]9Z3dF`M0^OOI\\9Z3eF_M0^ONJ]9Z3dF^M1^ONJ\\9\\3dF\\M2^OMK\\9\\3eF[M2^OLL\\9]3eFYM1@NJ\\9^3dFXM1ANJ\\9^3eFWMOM[9n2eFfM0kN[9Z5O1O1O100O1O1O100O1O1O001O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1aNmHQKR7k4UIRKl6j4ZISKg6h4`IUKa6^3WIXL`07Y6`3WKXLj4g3ZKULg4j3\\KSLe4l3_KQLa4n3bKoK_4P4eKlK\\4S4Y3O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O1000000O1000000O100000000O0100000O1000000O1000000O1000000O10000000000O100000000000000O100000000000000O10000000000000000001O001O001O000O2O001O00001O001O001O00001O001O00001O001O001O00002N1O1O2dH[LY3f3aL`L^3b3[LcLe3^3SLjLl3X3kKPMT4Q3dKVM\\4l2[K\\Md4e2YK^Mf4d2UK_Mk4b2RKaMm4`2PKcMo4_2mJcMS5^2jJeMU5]2fJgMY5Z2dJhM\\5Z2`JiM_5[2[JhMd5^2SJeMm5a2kIbMT6d2dI^M\\6h2\\I[Mc6b5O001O1O001N2O001O1O1O2N2N2N2N2N2N2N2N2eL`HROb77VIFl66YIGi66\\IFf66_IGd65`IHb64dIH^65fIH\\64iIIY63lIJV63oIIS63RJJP63UJIP60VJLP6MVJNP6JTJ4\\:01O1O001O1O001OVh3"}, "label": "a boy wearing a blue shirt"}]}
{"id": 574957, "image": "COCO_train2014_000000574957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child wearing blue shirt sitting on ground\"."}], "task": "refering", "answer_template": "The \"child wearing blue shirt sitting on ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.3909375, 0.1993125, 0.986578125, 1.0], "iscrowd": 0, "area": 81017.30904999998, "rle": {"size": [480, 640], "counts": "h`e33m>6I1O2N1O1O100O1O1O1O1O10000000010O0000000000000001O0000000001O0001O0001O0000001O00010O0000001O00000010O000001O00001O01O0001^D]Of8b0ZG^Od8d0\\G\\Oc8f0[G[Oc8h0\\GXOb8k0]GTOc8m0\\GTOb8o0]GQOb8Q1mFjN_N5b:T1nFgN_N6LJ[:[1XGfNaN5KKZ:]1YGdNaN3MMW:]1[GcN`N4NLU:`1[GaNbN3NLU:a1ZG`NcN3NMT:a1YG`NeN2NMT:b1XG_NeN2ONT:b1WG^NfN2OOS:b1VG^NhN1OOS:b1VG^NgN20NS:c1TG^NiN10OQ:d1UG\\NjN1O0R:d1SG\\NkN10OR:e1RG[NlN100Q:e1RGZNmN100Q:g1oFYNoN110P:h1nFWNQO110P:i1lFWNSO001Q:j1jFUNUO002P:k1hFTNWO010Q:n1eFRNYO010P:Q2dFoM[O010P:S2aFnM]O02OP:U2_FlM_O010Q:V2\\FkMBO10Q:W2[FkMAO3OQ:Y2YFiMCO3OP:\\2XFfMEO3Ni9g2]F\\MF04Mh9i2]FZMG03N2E^9T3eFYMH03N1F^9U3eFXMG05M1F]9X3dFUMIO6NOG^9W3dFUMIO6NNH^9X3dFSMJO5OOG]9Z3dF`M0^OOI\\9Z3eF_M0^ONJ]9Z3dF^M1^ONJ\\9\\3dF\\M2^OMK\\9\\3eF[M2^OLL\\9]3eFYM1@NJ\\9^3dFXM1ANJ\\9^3eFWMOM[9n2eFfM0kN[9Z5O1O1O100O1O1O100O1O1O001O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1aNmHQKR7k4UIRKl6j4ZISKg6h4`IUKa6^3WIXL`07Y6`3WKXLj4g3ZKULg4j3\\KSLe4l3_KQLa4n3bKoK_4P4eKlK\\4S4Y3O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O1000000O1000000O100000000O0100000O1000000O1000000O1000000O10000000000O100000000000000O100000000000000O10000000000000000001O001O001O000O2O001O00001O001O001O00001O001O00001O001O001O00002N1O1O2dH[LY3f3aL`L^3b3[LcLe3^3SLjLl3X3kKPMT4Q3dKVM\\4l2[K\\Md4e2YK^Mf4d2UK_Mk4b2RKaMm4`2PKcMo4_2mJcMS5^2jJeMU5]2fJgMY5Z2dJhM\\5Z2`JiM_5[2[JhMd5^2SJeMm5a2kIbMT6d2dI^M\\6h2\\I[Mc6b5O001O1O001N2O001O1O1O2N2N2N2N2N2N2N2N2eL`HROb77VIFl66YIGi66\\IFf66_IGd65`IHb64dIH^65fIH\\64iIIY63lIJV63oIIS63RJJP63UJIP60VJLP6MVJNP6JTJ4\\:01O1O001O1O001OVh3"}, "label": "child wearing blue shirt sitting on ground"}]}
{"id": 370162, "image": "COCO_train2014_000000370162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl full of green lettuce salad\"."}], "task": "refering", "answer_template": "The \"a bowl full of green lettuce salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.27246875, 0.826975, 0.761234375, 1.0], "iscrowd": 0, "area": 14388.942900000004, "rle": {"size": [400, 640], "counts": "hZT2d0i;3N3L4L3N3L4N@lDFT;:jDZO1:U;;jD_O13U;=kDB0OU;?kDEOIW;a0kDHODU;e0kDHX;7iDIW;7iDIW;6iDKW;4jDLV;4jDLU;4lDLT;4kDMU;2lDNT;1mDOS;1mDOR;1nD0R;0mD1S;NmD3S;_OmD7O;T;_OXEb0g:AVE`0j:ASEa0l:ARE`0m:CPE>P;CnD>Q;DmD=S;`0000000000001O000000000000000000000000000O10000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000O10000000000000000000000000001N100000000000000000000000000O10000000000000000000000000000O100000000000000N2M3M3M3N20000O10000000000000000000000000000000000000000000000O100000000000000000000000000000000001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1NS\\k1"}, "label": "a bowl full of green lettuce salad"}]}
{"id": 370162, "image": "COCO_train2014_000000370162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table and big bowl on green stuff on it\"."}], "task": "refering", "answer_template": "The \"a table and big bowl on green stuff on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.27246875, 0.826975, 0.761234375, 1.0], "iscrowd": 0, "area": 14388.942900000004, "rle": {"size": [400, 640], "counts": "hZT2d0i;3N3L4L3N3L4N@lDFT;:jDZO1:U;;jD_O13U;=kDB0OU;?kDEOIW;a0kDHODU;e0kDHX;7iDIW;7iDIW;6iDKW;4jDLV;4jDLU;4lDLT;4kDMU;2lDNT;1mDOS;1mDOR;1nD0R;0mD1S;NmD3S;_OmD7O;T;_OXEb0g:AVE`0j:ASEa0l:ARE`0m:CPE>P;CnD>Q;DmD=S;`0000000000001O000000000000000000000000000O10000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000O10000000000000000000000000001N100000000000000000000000000O10000000000000000000000000000O100000000000000N2M3M3M3N20000O10000000000000000000000000000000000000000000000O100000000000000000000000000000000001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1NS\\k1"}, "label": "a table and big bowl on green stuff on it"}]}
{"id": 157172, "image": "COCO_train2014_000000157172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a black seat\"."}], "task": "refering", "answer_template": "The \"a chair with a black seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 229, 251, 252, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.7943125, 0.599367681498829, 1.0, 1.0], "iscrowd": 0, "area": 10692.932249999994, "rle": {"size": [427, 640], "counts": "odd65U=4bNK^E8\\:3]ENZ:a0^EBY:m1N01N1N3N2hNnD5T;LQEHT;:REWOV;j0e01O1O1O1O1O1O1O1O1O1O1O1O1O0O2O001mDdNV:\\1^EPOa:m1000000000000000L4I7I7M3N2O1O101O0000000001O0000000000000000000001O000000000;E00001O011N9HOM4J5G9G9I7LF9L45L4E;D;M4M3N2O1N20O01O100O1O010O1O0jGfK]7\\4XHmKg7l400O101O000O10O1L4D<_Oa0YOYGcLn8Y3d0K5K5N2M3M3K5J7I6K5L4K5M3M4ITH"}, "label": "a chair with a black seat"}]}
{"id": 157172, "image": "COCO_train2014_000000157172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and brown chair\"."}], "task": "refering", "answer_template": "The \"a black and brown chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 229, 251, 252, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.7943125, 0.599367681498829, 1.0, 1.0], "iscrowd": 0, "area": 10692.932249999994, "rle": {"size": [427, 640], "counts": "odd65U=4bNK^E8\\:3]ENZ:a0^EBY:m1N01N1N3N2hNnD5T;LQEHT;:REWOV;j0e01O1O1O1O1O1O1O1O1O1O1O1O1O0O2O001mDdNV:\\1^EPOa:m1000000000000000L4I7I7M3N2O1O101O0000000001O0000000000000000000001O000000000;E00001O011N9HOM4J5G9G9I7LF9L45L4E;D;M4M3N2O1N20O01O100O1O010O1O0jGfK]7\\4XHmKg7l400O101O000O10O1L4D<_Oa0YOYGcLn8Y3d0K5K5N2M3M3K5J7I6K5L4K5M3M4ITH"}, "label": "a black and brown chair"}]}
{"id": 452088, "image": "COCO_train2014_000000452088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small airplane that is metal\"."}], "task": "refering", "answer_template": "The \"a small airplane that is metal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 100, 101, 102, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134], "bbox": [0.5475294117647058, 0.213484375, 1.0, 0.378046875], "iscrowd": 0, "area": 8844.4017, "rle": {"size": [640, 425], "counts": "^ka44fc06J6J6J6J6O100000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000gNQOa_Oo0[`0XOb_Oh0Z`0\\Of_Od0Z`0]Oe_Oc0[`0^Od_Ob0\\`0^Od_Ob0\\`0^Od_Ob0\\`0^OP_OF1l0o`0^Oj^O4O>Wa0^Oe^O`1\\a0b0000000000000001O0000H8I7M3N2N21O1O1O1O001O00000000000O100000002N1O2N1O2M2O2N1O1O2N1O1O2M2O1O1OX1hN1O0000000O2O000000001O00000O100000000000000O10O10000000000000O100000O10000000O1000000000O100000O100000000000O1000O100000000000000O01000L4G9FPJ"}, "label": "a small airplane that is metal"}]}
{"id": 452088, "image": "COCO_train2014_000000452088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small silver colored plane in front of a larger plane\"."}], "task": "refering", "answer_template": "The \"a small silver colored plane in front of a larger plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 100, 101, 102, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134], "bbox": [0.5475294117647058, 0.213484375, 1.0, 0.378046875], "iscrowd": 0, "area": 8844.4017, "rle": {"size": [640, 425], "counts": "^ka44fc06J6J6J6J6O100000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000gNQOa_Oo0[`0XOb_Oh0Z`0\\Of_Od0Z`0]Oe_Oc0[`0^Od_Ob0\\`0^Od_Ob0\\`0^Od_Ob0\\`0^OP_OF1l0o`0^Oj^O4O>Wa0^Oe^O`1\\a0b0000000000000001O0000H8I7M3N2N21O1O1O1O001O00000000000O100000002N1O2N1O2M2O2N1O1O2N1O1O2M2O1O1OX1hN1O0000000O2O000000001O00000O100000000000000O10O10000000000000O100000O10000000O1000000000O100000O100000000000O1000O100000000000000O01000L4G9FPJ"}, "label": "a small silver colored plane in front of a larger plane"}]}
{"id": 452088, "image": "COCO_train2014_000000452088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white airplane behind a red and white airplane\"."}], "task": "refering", "answer_template": "The \"a black and white airplane behind a red and white airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 35, 36, 46, 47, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 76, 77, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 98, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 302, 317], "bbox": [0.0, 0.1016875, 1.0, 0.9443750000000002], "iscrowd": 0, "area": 120386.24325000003, "rle": {"size": [640, 425], "counts": "R>R1d<QOgGk2Y8[NaFe1_9@\\Ea0c:a0[D_Oe;a0[D_Oe;a0[D@d;`0\\D@d;a0[D_Oe;a0[D@d;`0\\D@d;`0\\D@d;`0\\DAd;>\\DBd;>\\DBd;>\\DCc;>\\DBd;>\\DF`;:`D0V;0gD7U;IgD;Y;EdD>\\;B`Db0`;^O\\Df0d;[OWDi0i;WOSDm0m;d3O000001O00000001O000000000000001O00000001O00WJi53M2O2MVJhBo4U=iJ^Cn4`<jJSDn4b=N01000O100O010O10000O010O12L5H8H9H7H8I8UMmIYG[6`8PJTGX6e8SJPGT6i8WJkFR6n8UJjFR6o8RJlFV6m8oImFZ6k8jIQG[6k8hIPGP6Y9TJcFb5g9bJTFV5U:Q29F;I6O1000001O001O1O001N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O001O1000000000000000000000gLnHdIR7[6oHeIQ7Z6PIfIP7Y6RIfIn6Y6SIgIm6X6TIhIl6W6UIiIk6V6VIjIj6T6XIlIh6S6YIlIh6S6YImIg6R6ZInIf6Q6[IoIe6P6\\IPJd6o5]IQJc6n5_IQJa6n5`IRJ`6m5aISJ_6l5bITJ^6k5cIUJ]6j5dIVJ\\6i5eIWJ[6h5fIXJZ6g5gIYJY6f5hIZJX6d5jI\\JV6c5lI\\JT6e5kI[JV6d5jI\\JY6a5gI_J[6_5eIaJ^6]5aIcJa6[5_IdJe6Y5[IgJg6W5YIiJj6U5UIkJm6S5SImJP7P5PIPKR7o4nHPKT7n4lHRKW7k4iHUKY7i4gHWK\\7g4cHYK]7g4cHYK]7g4cHYK^7f4bHZK^7g4aHYK_7g4aHYK`7f4`HZK`7f4`HZK`7g4`HXKa7g4_HYKa7g4_HYKa7h4^HXKc7g4]HYKc7g4]HXKd7h4\\HXKd7i4[HWKf7h4ZHXKf7h4ZHXKf7h4ZHXKg7h4XHXKh7h4YHWKg7i4YHWKh7h4XHXKh7h4XHXKj7f4VHZKl7d4TH\\Kn7b4RH^KP8_4QHaKQ8]4oGcKT8Z4lGfKV8X4jGhKX8V4iGiKY8U4gGkK[8S4eGmK]8NoFo0e0SO^8DUGY1=SO`8YO\\Gd14ROc8POaGn1LROe8fNgGX2DROg8\\NmGb2\\OROi8RNSHk2UOSO_;m0aDSO_;m0aDSO^;n0cDQO];o0cDQO];o0cDQO];n0dDRO\\;n0dDRO\\;n0dDRO\\;n0dDRO\\;n0dDRO\\;n0dDRO\\;m0eDSO[;m0eDSO[;m0eDSO[;m0fDROZ;n0fDROY;o0gDQOY;o0gDQOY;n0hDQOY;o0gDQOY;o0gDQOY;o0gDQOY;o0gDQOY;o0gDQOY;n0hDROX;n0hDROX;n0iDQOW;o0iDQOW;o0iDQOW;o0iDQOV;o0kDQOU;o0kDQOU;o0kDQOU;o0kDQOU;o0kDQOU;o0kDQOU;n0lDROT;n0mDQOS;o0mDQOS;o0mDQOS;o0mDPOT;P1lDPOT;o0mDQOR;P1nDPOR;P1nDPOR;P1nDPOR;P1nDPOR;P1nDPOR;P1nDPOR;o0PEPOP;P1PEPOP;P1PEPOP;P1PEPOP;P1PEPOP;P1PEPOP;o0QEQOo:o0QEQOn:P1REPOn:P1REPOn:P1REPOn:P1REPOn:o0TEPOl:P1TEPOl:P1TEoNm:Q1SEoNm:Q1SEoNm:Q1SEoNm:P1TEPOl:P1TEPOl:P1TEPOl:P1TEPOk:Q1UEoNk:Q1VEnNj:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:P1XEPOh:P1XEPOh:P1XEPOh:P1XEPOh:P1YEoNf:R1ZEnNf:Q1[EnNf:R1ZEnNf:R1ZEnNf:R1ZEnNf:R1ZEnNf:R1ZEnNf:Q1[EoNe:Q1[EoNe:Q1[EoNd:R1\\EnNd:R1]EmNc:S1]EmNc:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNa:R1`EnN`:R1`EnN`:R1`EnN`:R1aEmN_:S1aElN`:T1`ElN`:S1aEmN_:S1aEmN_:S1aEmN^:T1bElN^:T1bElN^:T1bElN^:S1cEmN]:S1cEmN]:S1dElN\\:T1dElN\\:T1dElN\\:T1dElN[:T1fElNZ:T1fElNZ:T1fElNZ:T1fElNZ:T1fElNZ:T1fElNZ:S1gEmNY:S1gEmNY:S1hEkNY:U1gEkNX:V1hEjNX:V1hEjNX:V1hEjNX:U1iEkNW:U1iEkNW:U1iEkNW:U1iEkNW:U1iEkNW:U1iEkNV:U1kEkNU:U1lEjNT:V1lEjNT:V1lEjNT:V1lEjNT:V1lEjNT:U1mEkNS:U1mEkNS:U1mEkNR:V1nEjNR:V1nEjNR:V1nEjNR:U1PFiNQ:W1oEiNQ:W1oEiNQ:W1oEiNQ:W1oEiNQ:W1oEiNP:W1QFiNo9W1QFiNo9W1QFiNo9W1QFiNo9W1QFiNo9W1QFiNo9W1RFhNn9W1SFiNm9W1SFiNm9W1SFiNl9X1TFhNl9X1TFhNl9X1TFhNl9W1UFiNk9W1UFiNk9W1UFiNk9W1UFiNk9W1UFhNl9X1UFgNj9Y1WFgNi9Y1WFgNi9Y1WFgNi9Y1WFgNi9Y1WFgNl:6TETOS=ZOnBGX4"}, "label": "a black and white airplane behind a red and white airplane"}]}
{"id": 452088, "image": "COCO_train2014_000000452088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"larger plane on grass\"."}], "task": "refering", "answer_template": "The \"larger plane on grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 35, 36, 46, 47, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 76, 77, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 98, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 302, 317], "bbox": [0.0, 0.1016875, 1.0, 0.9443750000000002], "iscrowd": 0, "area": 120386.24325000003, "rle": {"size": [640, 425], "counts": "R>R1d<QOgGk2Y8[NaFe1_9@\\Ea0c:a0[D_Oe;a0[D_Oe;a0[D@d;`0\\D@d;a0[D_Oe;a0[D@d;`0\\D@d;`0\\D@d;`0\\DAd;>\\DBd;>\\DBd;>\\DCc;>\\DBd;>\\DF`;:`D0V;0gD7U;IgD;Y;EdD>\\;B`Db0`;^O\\Df0d;[OWDi0i;WOSDm0m;d3O000001O00000001O000000000000001O00000001O00WJi53M2O2MVJhBo4U=iJ^Cn4`<jJSDn4b=N01000O100O010O10000O010O12L5H8H9H7H8I8UMmIYG[6`8PJTGX6e8SJPGT6i8WJkFR6n8UJjFR6o8RJlFV6m8oImFZ6k8jIQG[6k8hIPGP6Y9TJcFb5g9bJTFV5U:Q29F;I6O1000001O001O1O001N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O001O1000000000000000000000gLnHdIR7[6oHeIQ7Z6PIfIP7Y6RIfIn6Y6SIgIm6X6TIhIl6W6UIiIk6V6VIjIj6T6XIlIh6S6YIlIh6S6YImIg6R6ZInIf6Q6[IoIe6P6\\IPJd6o5]IQJc6n5_IQJa6n5`IRJ`6m5aISJ_6l5bITJ^6k5cIUJ]6j5dIVJ\\6i5eIWJ[6h5fIXJZ6g5gIYJY6f5hIZJX6d5jI\\JV6c5lI\\JT6e5kI[JV6d5jI\\JY6a5gI_J[6_5eIaJ^6]5aIcJa6[5_IdJe6Y5[IgJg6W5YIiJj6U5UIkJm6S5SImJP7P5PIPKR7o4nHPKT7n4lHRKW7k4iHUKY7i4gHWK\\7g4cHYK]7g4cHYK]7g4cHYK^7f4bHZK^7g4aHYK_7g4aHYK`7f4`HZK`7f4`HZK`7g4`HXKa7g4_HYKa7g4_HYKa7h4^HXKc7g4]HYKc7g4]HXKd7h4\\HXKd7i4[HWKf7h4ZHXKf7h4ZHXKf7h4ZHXKg7h4XHXKh7h4YHWKg7i4YHWKh7h4XHXKh7h4XHXKj7f4VHZKl7d4TH\\Kn7b4RH^KP8_4QHaKQ8]4oGcKT8Z4lGfKV8X4jGhKX8V4iGiKY8U4gGkK[8S4eGmK]8NoFo0e0SO^8DUGY1=SO`8YO\\Gd14ROc8POaGn1LROe8fNgGX2DROg8\\NmGb2\\OROi8RNSHk2UOSO_;m0aDSO_;m0aDSO^;n0cDQO];o0cDQO];o0cDQO];n0dDRO\\;n0dDRO\\;n0dDRO\\;n0dDRO\\;n0dDRO\\;n0dDRO\\;m0eDSO[;m0eDSO[;m0eDSO[;m0fDROZ;n0fDROY;o0gDQOY;o0gDQOY;n0hDQOY;o0gDQOY;o0gDQOY;o0gDQOY;o0gDQOY;o0gDQOY;n0hDROX;n0hDROX;n0iDQOW;o0iDQOW;o0iDQOW;o0iDQOV;o0kDQOU;o0kDQOU;o0kDQOU;o0kDQOU;o0kDQOU;o0kDQOU;n0lDROT;n0mDQOS;o0mDQOS;o0mDQOS;o0mDPOT;P1lDPOT;o0mDQOR;P1nDPOR;P1nDPOR;P1nDPOR;P1nDPOR;P1nDPOR;P1nDPOR;o0PEPOP;P1PEPOP;P1PEPOP;P1PEPOP;P1PEPOP;P1PEPOP;o0QEQOo:o0QEQOn:P1REPOn:P1REPOn:P1REPOn:P1REPOn:o0TEPOl:P1TEPOl:P1TEoNm:Q1SEoNm:Q1SEoNm:Q1SEoNm:P1TEPOl:P1TEPOl:P1TEPOl:P1TEPOk:Q1UEoNk:Q1VEnNj:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:Q1WEoNi:P1XEPOh:P1XEPOh:P1XEPOh:P1XEPOh:P1YEoNf:R1ZEnNf:Q1[EnNf:R1ZEnNf:R1ZEnNf:R1ZEnNf:R1ZEnNf:R1ZEnNf:Q1[EoNe:Q1[EoNe:Q1[EoNd:R1\\EnNd:R1]EmNc:S1]EmNc:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNb:R1^EnNa:R1`EnN`:R1`EnN`:R1`EnN`:R1aEmN_:S1aElN`:T1`ElN`:S1aEmN_:S1aEmN_:S1aEmN^:T1bElN^:T1bElN^:T1bElN^:S1cEmN]:S1cEmN]:S1dElN\\:T1dElN\\:T1dElN\\:T1dElN[:T1fElNZ:T1fElNZ:T1fElNZ:T1fElNZ:T1fElNZ:T1fElNZ:S1gEmNY:S1gEmNY:S1hEkNY:U1gEkNX:V1hEjNX:V1hEjNX:V1hEjNX:U1iEkNW:U1iEkNW:U1iEkNW:U1iEkNW:U1iEkNW:U1iEkNV:U1kEkNU:U1lEjNT:V1lEjNT:V1lEjNT:V1lEjNT:V1lEjNT:U1mEkNS:U1mEkNS:U1mEkNR:V1nEjNR:V1nEjNR:V1nEjNR:U1PFiNQ:W1oEiNQ:W1oEiNQ:W1oEiNQ:W1oEiNQ:W1oEiNP:W1QFiNo9W1QFiNo9W1QFiNo9W1QFiNo9W1QFiNo9W1QFiNo9W1RFhNn9W1SFiNm9W1SFiNm9W1SFiNl9X1TFhNl9X1TFhNl9X1TFhNl9W1UFiNk9W1UFiNk9W1UFiNk9W1UFiNk9W1UFhNl9X1UFgNj9Y1WFgNi9Y1WFgNi9Y1WFgNi9Y1WFgNi9Y1WFgNl:6TETOS=ZOnBGX4"}, "label": "larger plane on grass"}]}
{"id": 26105, "image": "COCO_train2014_000000026105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white t - shirt is screen printed with lecanto tennis and is being worn by a boy\"."}], "task": "refering", "answer_template": "The \"a white t - shirt is screen printed with lecanto tennis and is being worn by a boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.382328125, 0.36180327868852463, 0.71815625, 0.9842622950819673], "iscrowd": 0, "area": 35681.96975, "rle": {"size": [427, 640], "counts": "X`V34T=4L4oCKg:9TENf:6UE1e:3UE5e:OVE8d:LVE<d:Z1J6J5K3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3O1O100O10000O100O100O100O100O100O100O100O100O100O100O10000O01000O10000O10000O10000O10000O10000O10000O1YNkJXKU5b4ZKUKf4d4iKSKX4g4TLSKl3g4]LVKc3e4dLYK\\3g4fLVK[3i4iLTKW3k4lLSKT3l4oLoJT3Q5oLfJW3Y5^2O1O1O100O1O1O1O100O1O1O100000000000000000000000000000000000000000O101O001O1O001O001O001O001O001O001UJYKP3g4mL\\KS3e4hL`KW3`4gLcKX3^4fLcKZ3]4eLeKZ3\\4cLgK\\3Y4cLhK]3Y4`LjK_3[4[LfKe3_4TLdKk3a4oKaKP4d4iK^KW4l6O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O2N2N3L3N3M3M2lNfFSN]9j1oFjMS9T2U1L3N3M2M4M2N3L4M2M4M2N3L3N3M4K7J6J5J7IiS[2"}, "label": "a white t - shirt is screen printed with lecanto tennis and is being worn by a boy"}]}
{"id": 26105, "image": "COCO_train2014_000000026105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a white and green shirt playing frisbee with another boy\"."}], "task": "refering", "answer_template": "The \"a boy in a white and green shirt playing frisbee with another boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.382328125, 0.36180327868852463, 0.71815625, 0.9842622950819673], "iscrowd": 0, "area": 35681.96975, "rle": {"size": [427, 640], "counts": "X`V34T=4L4oCKg:9TENf:6UE1e:3UE5e:OVE8d:LVE<d:Z1J6J5K3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3O1O100O10000O100O100O100O100O100O100O100O100O100O100O10000O01000O10000O10000O10000O10000O10000O10000O1YNkJXKU5b4ZKUKf4d4iKSKX4g4TLSKl3g4]LVKc3e4dLYK\\3g4fLVK[3i4iLTKW3k4lLSKT3l4oLoJT3Q5oLfJW3Y5^2O1O1O100O1O1O1O100O1O1O100000000000000000000000000000000000000000O101O001O1O001O001O001O001O001O001UJYKP3g4mL\\KS3e4hL`KW3`4gLcKX3^4fLcKZ3]4eLeKZ3\\4cLgK\\3Y4cLhK]3Y4`LjK_3[4[LfKe3_4TLdKk3a4oKaKP4d4iK^KW4l6O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O2N2N3L3N3M3M2lNfFSN]9j1oFjMS9T2U1L3N3M2M4M2N3L4M2M4M2N3L3N3M4K7J6J5J7IiS[2"}, "label": "a boy in a white and green shirt playing frisbee with another boy"}]}
{"id": 9723, "image": "COCO_train2014_000000009723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the waffle at the bottom of the plate\"."}], "task": "refering", "answer_template": "The \"the waffle at the bottom of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 366, 367], "bbox": [0.22522916666666667, 0.63175, 0.7897916666666667, 0.9358124999999999], "iscrowd": 0, "area": 30140.425900000002, "rle": {"size": [640, 480], "counts": "nnS21mc03M3M2M4M3M2N3L4M2N3M3L3O2N2O0O2O1N2N101N2N101N2N101N2N1O2O1N1O2N2O0O2N2N101O1N101O1N101O1N2O001O1N101O10O0100O01000O10O0100N101O1O001O1O1O001O1N101O1O1O001O1O001O1O1N101O1O1O001O1O1O001N2O1O1O001O1O1O001O1O1N101O1O1O001O1O1O1O0O2O1O1O001O1O1O0O2N2O1N2N101N101N1O101N1O2O0O2N101N1O101N1O2O0O2N101O000010O01O001O010O001O01O01O0010O01O0010O0001O001N1O2N1O2O0O1O2N1O2O0O2O001O00001O001OO1O1O1N101O1O1N2O1O1O1O1N2O1O001N2O1O1O1O1N2O1O1O1N101O1O1N2O1O1E;D<E;I7I7H7J7M4L3N2N2N2N2N2N2M3N3M2N2N2N2N2M3N2N2N2N3M2N2M3N2N2N2N2N2N2M4M2N2N2N2N2N2M3N2N3MYgn1"}, "label": "the waffle at the bottom of the plate"}]}
{"id": 9723, "image": "COCO_train2014_000000009723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grilled cheese closest to the bottom\"."}], "task": "refering", "answer_template": "The \"grilled cheese closest to the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 366, 367], "bbox": [0.22522916666666667, 0.63175, 0.7897916666666667, 0.9358124999999999], "iscrowd": 0, "area": 30140.425900000002, "rle": {"size": [640, 480], "counts": "nnS21mc03M3M2M4M3M2N3L4M2N3M3L3O2N2O0O2O1N2N101N2N101N2N101N2N1O2O1N1O2N2O0O2N2N101O1N101O1N101O1N2O001O1N101O10O0100O01000O10O0100N101O1O001O1O1O001O1N101O1O1O001O1O001O1O1N101O1O1O001O1O1O001N2O1O1O001O1O1O001O1O1N101O1O1O001O1O1O1O0O2O1O1O001O1O1O0O2N2O1N2N101N101N1O101N1O2O0O2N101N1O101N1O2O0O2N101O000010O01O001O010O001O01O01O0010O01O0010O0001O001N1O2N1O2O0O1O2N1O2O0O2O001O00001O001OO1O1O1N101O1O1N2O1O1O1O1N2O1O001N2O1O1O1O1N2O1O1O1N101O1O1N2O1O1E;D<E;I7I7H7J7M4L3N2N2N2N2N2N2M3N3M2N2N2N2N2M3N2N2N2N3M2N2M3N2N2N2N2N2N2M4M2N2N2N2N2N2M3N2N3MYgn1"}, "label": "grilled cheese closest to the bottom"}]}
{"id": 214523, "image": "COCO_train2014_000000214523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green truck is on the grass\"."}], "task": "refering", "answer_template": "The \"a green truck is on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 327, 328, 329, 332, 333], "bbox": [0.0055625, 0.2710416666666667, 0.55321875, 0.8774791666666666], "iscrowd": 0, "area": 64177.60369999999, "rle": {"size": [480, 640], "counts": "QW26g>8H7I7J6I7I7K6J5L2M0010O00100O010O100O100O010O100O1O100O17I1O0000001O0000001O000000001O0000001O0000001O00000O1O2N1O1O1O2O0O10001N1000000O2O00O010O1O001O1O010O001O00010O001O01O01000O10000000001O0000000O10000000001dNmMiES2S8bMbHj0[Od1R8dMTHW1IU1R8fMmG\\11n0Q8gMnG[10o0Q8hMnGZ1NP1S8hMmGY1YOf1i8SMmGc4R8_KmGa4R8aKmG_4R8cKjGa4T8aKcGg4\\8o0O1O1N2O1N2N2N2N3M2N2N2BVIdHl6\\7TIbHn6^7RIaHo6^7SI_Ho6a7801O1O2N2N3M5K4M3<A1OO1O2N1O100O000SO[H_Je7^5`H`J`7]5eHbJY7]5kHaJU7^5nH`JR7_5QI_Jo6`5TI^Jl6a5WI]Ji6b5ZI]Jd6c5_I[Ja6e5aIYJ_6g5cIWJ]6i5]1001N10000000000O10000000000O10001O00000O10000000000O100000001O0O100000000H8G9K5000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000000000000002N2N2N3M2N2M4M2N2N3M2N2N2N36IN2OO0100O0010O010O010O0100O00010O01O0010O0001O0K6K401O001O00001O001O010O0000O100O1O:G1N2O2M2O1N2N2O1N2J6G9F:O7PKoFb3Z9kKUGP4R:K:F2N100gNPEaNQ;Z1XEaNj:Y1^EcNc:\\OjDk0`1Hg9]O^Ga0c8_O^G?d8@]G?c;I7HQgU4"}, "label": "a green truck is on the grass"}]}
{"id": 214523, "image": "COCO_train2014_000000214523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green truck\"."}], "task": "refering", "answer_template": "The \"the green truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 327, 328, 329, 332, 333], "bbox": [0.0055625, 0.2710416666666667, 0.55321875, 0.8774791666666666], "iscrowd": 0, "area": 64177.60369999999, "rle": {"size": [480, 640], "counts": "QW26g>8H7I7J6I7I7K6J5L2M0010O00100O010O100O100O010O100O1O100O17I1O0000001O0000001O000000001O0000001O0000001O00000O1O2N1O1O1O2O0O10001N1000000O2O00O010O1O001O1O010O001O00010O001O01O01000O10000000001O0000000O10000000001dNmMiES2S8bMbHj0[Od1R8dMTHW1IU1R8fMmG\\11n0Q8gMnG[10o0Q8hMnGZ1NP1S8hMmGY1YOf1i8SMmGc4R8_KmGa4R8aKmG_4R8cKjGa4T8aKcGg4\\8o0O1O1N2O1N2N2N2N3M2N2N2BVIdHl6\\7TIbHn6^7RIaHo6^7SI_Ho6a7801O1O2N2N3M5K4M3<A1OO1O2N1O100O000SO[H_Je7^5`H`J`7]5eHbJY7]5kHaJU7^5nH`JR7_5QI_Jo6`5TI^Jl6a5WI]Ji6b5ZI]Jd6c5_I[Ja6e5aIYJ_6g5cIWJ]6i5]1001N10000000000O10000000000O10001O00000O10000000000O100000001O0O100000000H8G9K5000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000000000000002N2N2N3M2N2M4M2N2N3M2N2N2N36IN2OO0100O0010O010O010O0100O00010O01O0010O0001O0K6K401O001O00001O001O010O0000O100O1O:G1N2O2M2O1N2N2O1N2J6G9F:O7PKoFb3Z9kKUGP4R:K:F2N100gNPEaNQ;Z1XEaNj:Y1^EcNc:\\OjDk0`1Hg9]O^Ga0c8_O^G?d8@]G?c;I7HQgU4"}, "label": "the green truck"}]}
{"id": 214523, "image": "COCO_train2014_000000214523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue semi truck parked to the right of a green semi truck ,\"."}], "task": "refering", "answer_template": "The \"a blue semi truck parked to the right of a green semi truck ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 151, 155, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 337, 338], "bbox": [0.541375, 0.32579166666666665, 0.9532187499999999, 0.8426875], "iscrowd": 0, "area": 40082.8255, "rle": {"size": [480, 640], "counts": "g`R51a>?ZOe0O2O2N3M4L3M4L3M0O101O000000000000001N10000\\KTNTLl1l3XNPLh1P4\\NlKd1T4aNSKJjLf1S8cNQJe0lMh0T8dNPIb1lNJV8eN^Hn1\\O]OW8gNUHQ2BZOZ8fNmGU2EYO_8dNdGX2IXOe8Q4100O2N1O1O1O3M2N4M3^JZGa4e9NM3N2O1O1O1O100000O10000000000000000O1000000006J:F;E0N2O1N2O1O10001O0O2O001O1O001N2O010O001O1O0000000001O00O1000000000000O100004M0O1O000000O1M300O100O01000O010O10OYOiIfHX6X7QJaHn5^7ZJ\\Hf5a7l02M3N2N3M:Gb0]O9G3O000000001O000O101O0000001O0000001O000O2O000;F2MO10O1O1O10000O100000000O100000000001N2O001O4L1O1O001O001O001O1O2N1O2aJdFV5m9Eh0YO001O001O001O1O001O001O001O1O001N101O001O1O001O001O001O1O01001N8H8H7I4LO1O1O1O1O1O001000O1M4M2M4dNfCM]<3dCK^<4cCJb<2^CNd<0\\COg<OYC1Q=EoB:[=]OeBc0^=ZObBf0a=XO^Bg0e=VO\\Bj0m=O10000O105Ja0@cf="}, "label": "a blue semi truck parked to the right of a green semi truck ,"}]}
{"id": 214523, "image": "COCO_train2014_000000214523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large blue truck\"."}], "task": "refering", "answer_template": "The \"the large blue truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 151, 155, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 337, 338], "bbox": [0.541375, 0.32579166666666665, 0.9532187499999999, 0.8426875], "iscrowd": 0, "area": 40082.8255, "rle": {"size": [480, 640], "counts": "g`R51a>?ZOe0O2O2N3M4L3M4L3M0O101O000000000000001N10000\\KTNTLl1l3XNPLh1P4\\NlKd1T4aNSKJjLf1S8cNQJe0lMh0T8dNPIb1lNJV8eN^Hn1\\O]OW8gNUHQ2BZOZ8fNmGU2EYO_8dNdGX2IXOe8Q4100O2N1O1O1O3M2N4M3^JZGa4e9NM3N2O1O1O1O100000O10000000000000000O1000000006J:F;E0N2O1N2O1O10001O0O2O001O1O001N2O010O001O1O0000000001O00O1000000000000O100004M0O1O000000O1M300O100O01000O010O10OYOiIfHX6X7QJaHn5^7ZJ\\Hf5a7l02M3N2N3M:Gb0]O9G3O000000001O000O101O0000001O0000001O000O2O000;F2MO10O1O1O10000O100000000O100000000001N2O001O4L1O1O001O001O001O1O2N1O2aJdFV5m9Eh0YO001O001O001O1O001O001O001O1O001N101O001O1O001O001O001O1O01001N8H8H7I4LO1O1O1O1O1O001000O1M4M2M4dNfCM]<3dCK^<4cCJb<2^CNd<0\\COg<OYC1Q=EoB:[=]OeBc0^=ZObBf0a=XO^Bg0e=VO\\Bj0m=O10000O105Ja0@cf="}, "label": "the large blue truck"}]}
{"id": 255495, "image": "COCO_train2014_000000255495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman riding a bike behind the lady wearing a red shirt\"."}], "task": "refering", "answer_template": "The \"a woman riding a bike behind the lady wearing a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 148, 171, 172, 195, 196, 211, 212, 219, 220, 234, 235, 257, 258], "bbox": [0.18837500000000001, 0.12909426987060998, 0.6025, 0.6236044362292051], "iscrowd": 0, "area": 7533.108299999999, "rle": {"size": [541, 640], "counts": "lVP29^`0;E;G8K6J6J3M2N3L3N2N3M0O1N100O101N15K8H\\_`1WN^b_N3L4M3L4N2M2O1N2O2M2O1N2O2M2O1N2N3M2N2O1N1O1O100O1O1O000O1000000aA^Nd=c1]B]N`=e1aB[N\\=g1dBZNZ=f1hBZNU=h1lBWNR=k1oBUNn<m1RCTNk<m1WCSNf<o1[CQNc<o1^CRN_<P2bCPN\\<P2fCPNW<R2iCoMh;_2YDaMf;^2\\DbMc;^2^DbMb;\\2_DeM`;[2aDeM^;[2cDeM];Z2cDgM\\;Y2eDgMZ;Y2gDfMZ;Y2fDhMY;X2hDhMW;X2jDhMV;W2kDiMT;V2mDkMR;U2oDkMQ;T2PEkMP;U2PElMo:T2RElMn:S2SEmMl:S2TEnMk:R2VEnMj:Q2WEnMi:R2WEoMi:P2XEPNi:m1YESNg:l1ZETNf:j1[EWNe:h1\\EXNd:f1]E[Nc:c1_E]Na:a1`E`N`:^1bEbN^:]1bEdN^:Z1dEfN\\:X1fEgN[:W1fEjNZ:T1hElNX:S1hEnNX:P1jEPOV:k0nEVOR:f0RFZOQ:?RFBR:6RFJR:NQF3R:FRF:Y=0O100O10O01O10O01O00001O0010O01O0o_O\\Oi?e0V@\\Oi?d0X@]Of?c0Z@^Oe?c0Z@^Of?a0[@_Oa`02N100O1O2N100O1O2O0O1O100O2N1O1O10O01O001O001O001O010O001OO1O2N101N2N102M5K5GhmU4"}, "label": "a woman riding a bike behind the lady wearing a red shirt"}]}
{"id": 255495, "image": "COCO_train2014_000000255495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the handbar of a cycle on which a girl is sitting with purple colored tank top\"."}], "task": "refering", "answer_template": "The \"the handbar of a cycle on which a girl is sitting with purple colored tank top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 313, 314, 315, 316, 408, 409, 410, 411, 412, 430, 431, 432, 433, 434, 435, 436], "bbox": [0.594984375, 0.5667652495378928, 0.981359375, 0.9895194085027728], "iscrowd": 0, "area": 13095.596450000003, "rle": {"size": [541, 640], "counts": "_dY64g`02N3N2M3M3N1N3N2M3N1O2M3N2N2N1O2M3N2N1O2M3N2M3N1O2M3N2M2O2N2M3N2M2O2N2M3O000001O001O00001O00001O00001O001O00001O00001O00001O00001O001O000O2O0O101N100O2O001N100O2O0O101O0O10bEXNh5f1VJ]Nj5c1SJ`Nn5^1PJeNP6[1mIhNS6W1mIjNT6U1kIlNU6T1jImNV6R1jIoNW6P1hIQOX6n0hISOX6k0iIVOX6h0iIXOW6f0jI[OV6d0jI]OW6`0jIAV6>jICW6:jIGV67lIIT66lIKU62lIOT60mI0S6NnI3S6KnI5R6IoI8Q6GoI:R6EnI;R6DnI<T6CkI>U6AlI?T6AkI`0V6_OjIa0V6^OjIc0V6]OiId0X6ZOiIf0W6ZOhIg0X6YOhIg0Y6WOgIj0W6XOiIh0S6[OmIf0o5^OQJb0l5ATJ?h5DYJ<c5H]J8_5KaJ6]5LcJ4^5KbJ5^5JcJ6^5IaJ8_5GbJ9^5GbJ9_5FaJ:_5EbJ;_5D`J=`5BaJ>_5BaJ>`5A`J?`5@aJ`0`5_O_Jb0a5]O`Jc0`5]O`Jc0a5\\O_Jd0a5[O`Je0a5ZO^Jg0b5XO_Jh0a5XO_Jh0b5VO_Jj0a5VO_Jj0b5TO^Jm0b5SO^Jm0b5RO_Jn0b5QO^JP1a5oN`JQ1a5nN_JR1a5mN_JT1a5lN_JT1b5jN_JV1a5jN_JV1b5hN_JX1a5hN^JY1b5fN_JZ1b5eN^J[1b5dN_J\\1b5cN^J]1b5bN_J^1_:100000000O100000001O0O100000000O10000001O001O1O001O1O1O001O1O001O10O01O1O001O1O001O1O001O1O001O1O001O1O001O1O001N2O001N2O001N2O0O2O1O0O2O1O0O2O1N101O1Nej5"}, "label": "the handbar of a cycle on which a girl is sitting with purple colored tank top"}]}
{"id": 255495, "image": "COCO_train2014_000000255495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady in black skirt on bike\"."}], "task": "refering", "answer_template": "The \"lady in black skirt on bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 267, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 383, 399, 400, 401, 402, 403, 404, 405, 406, 426, 427, 428], "bbox": [0.18615625, 0.0, 0.70475, 0.9887615526802217], "iscrowd": 0, "area": 92472.17045, "rle": {"size": [541, 640], "counts": "bXo14\\`0>B>C<H9J6K4N3N2M2N3N2M3M2N3E;B=VKk4M3N1N3N2M3mMbGdK_8S4mGhKV8m3WHoKj7h3cHRL`7c3mHXLU7a3UI[Ll6e3WIVLl6i3WISLj6m3ZImKi6Q4[IkKf6U4[IhKg6X4YIgKi6W4XIgKj6Y4VIeKl6[4SIeKo6Z4QIeKP7Z4QIeKP7[4oHeKS7Z4mHeKT7Z4mHeKU7Z4kHeKV7[4iHeKX7[4hHdKZ7Z4gHeKZ7[4eHdK]7\\4cHcK_7[4b201O0O101I6I7I8H7I7O2O001O001O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O00VERNf6m1ZITNf6k1YIXNf6g1YI[Ng6d1YI^Ne6b1ZI`Nf6_1YIdNf6[1ZIfNf6Y1YIjNf6U1YInNf6Q1ZIPOe6P1ZISOe6l0ZIVOf6i0ZIXOf6g0YI\\Of6c0YI_Og6`0YIAg6>XIEf6;YIGg68YIJf65YIMg62XI0h6OXI3g6LXI6g6JXI8h6GXI;g6EWI=]O[N[4W1WLa0[OZN^4T1WLc0XO\\Na4P1VLf0WO]Nb4l0ULk0UO\\Ne4i0ULm0TO\\Ng4f0SLQ1SO\\Nj4b0QLV1RO[Nl4>QLY1PO\\No4:oK^1oNZNR57mKb1nNcNk4KVLd1mNlNd4_O]Li1lNTO]4ROeLm1nNZOT4hNmLP2oN^On3aNQMT2QO[Om3aNQMV2ROYOk3bNQMX2TOUOk3cNoL[2UOSOk3bNnL^2WOoNj3dNnL_2XOmNi3dNmLb2ZOiNi3eNkLe2\\OeNh3gNkLf2]OcNg3gNjLi2_O_Ng3hNhLl2A\\Ne3iNiLm2AZNf3iNgLP3CWNe3iNfLS3ESNd3kNeLU3GoMd3lNdLW3HmMc3lNbL[3KhMb3nN`L^3MeMb3mN_L`30bMa3nN\\Ld33^M_3oN\\Lf34[M`3oNYLj37WM_3oNXLm38TM_3POVLP4;oL_3QOSLT4>kL^3QORLW4?hL^3ROPLZ4b0dL]3ROoK]4c0aL^3ROlKa4f0]L\\3SOlKc4g0ZL]3m5cLSJ\\3n5dLQJ[3Q6dLoI\\3R6dLnI[3S6dLmI[3U6eLkIZ3V6eLjI[3W6eLiIY3Y6gLfIY3Z6gLfIY3[6gLeIW3]6hLcIX3^6hLbIW3_6iL`IV3b6iL_IV3b6jL]IV3d6iL]IU3e6kLZIU3g6jLYIV3h6jLXIT3j6lLUIT3l6kLTIU3m6kLRIT3P7kLQIT3P7lLoHT3R7lLmHS3U7lLlHS3U7mLjHR3X7mLhHR3Z7nLfHQ3[7nLeHQ3\\7PMcHP3^7oLbHP3`7PM`Ho2_OhKe6X1lIo2^OkKg6V1jIo2]OmKj6S1jIn2[OQLl6P1iIo2YOSLo6n0gIn2YOVLQ7k0gIn2VOYLT7i0eIm2WO[LU7g0dIm2VO^LW7e0bIm2UO`LZ7b0bIl2SOdL\\7`0`Il2ROfL_7=_Il2QOiLa7;^Ik2oNlLd78]Ik2oNnLe77[Ik2nNPMh74[Ij2lNTMj72YIj2jNWMn7NXIj2hN[MQ8KWIi2eN_MU8GVIi2bNdMX8DUIh2`NgM\\8@TIh2^NkM_8]OSIf2\\NPNb8YORIg2YNSNf8VOPIa4Q7^KPIa4Q7_KnHa4S7^KmHb4T7^KlHa4U7^KkHb4V7^KiHb4X7]KhHc4Y7]KgHb4Z7]KfHc4[7]KdHc4]7\\KdHc4]7]KbHc4^7]KcHb4]7_KbHa4_7^KbHa4^7`KaH`4_7`KbH_4^7bKaH^4`7aKaH^4_7cK`H]4`7cKaH\\4_7eK`H[4a7dK_H\\4a7eK_HZ4a7fK_HZ4a7gK_HX4b7gK^HY4b7hK^HW4b7iK^HW4b7jK^HU4c7jK]HV4c7jK^HU4b7lK]HT4c7lK^HS4c7mK\\HS4d7mK]HR4d7mK\\HS4d7nK\\HQ4e7nK[HQ4f7PLZHn3g7RLYHl3j7TLVHj3k7nKbF_Oc1a4n7lKcFB`1`4n7jKgFFZ1^4R8hKiFHV1^4R8gKlFKQ1\\4V8dKfHZ4o9N3N1N3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N2M2M4I7J6J6I7J6I7J6I7J5K^^S3"}, "label": "lady in black skirt on bike"}]}
{"id": 255495, "image": "COCO_train2014_000000255495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a skirt on and a helmet sitting on her bike looking directly at the camera\"."}], "task": "refering", "answer_template": "The \"a woman with a skirt on and a helmet sitting on her bike looking directly at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 267, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 383, 399, 400, 401, 402, 403, 404, 405, 406, 426, 427, 428], "bbox": [0.18615625, 0.0, 0.70475, 0.9887615526802217], "iscrowd": 0, "area": 92472.17045, "rle": {"size": [541, 640], "counts": "bXo14\\`0>B>C<H9J6K4N3N2M2N3N2M3M2N3E;B=VKk4M3N1N3N2M3mMbGdK_8S4mGhKV8m3WHoKj7h3cHRL`7c3mHXLU7a3UI[Ll6e3WIVLl6i3WISLj6m3ZImKi6Q4[IkKf6U4[IhKg6X4YIgKi6W4XIgKj6Y4VIeKl6[4SIeKo6Z4QIeKP7Z4QIeKP7[4oHeKS7Z4mHeKT7Z4mHeKU7Z4kHeKV7[4iHeKX7[4hHdKZ7Z4gHeKZ7[4eHdK]7\\4cHcK_7[4b201O0O101I6I7I8H7I7O2O001O001O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O00VERNf6m1ZITNf6k1YIXNf6g1YI[Ng6d1YI^Ne6b1ZI`Nf6_1YIdNf6[1ZIfNf6Y1YIjNf6U1YInNf6Q1ZIPOe6P1ZISOe6l0ZIVOf6i0ZIXOf6g0YI\\Of6c0YI_Og6`0YIAg6>XIEf6;YIGg68YIJf65YIMg62XI0h6OXI3g6LXI6g6JXI8h6GXI;g6EWI=]O[N[4W1WLa0[OZN^4T1WLc0XO\\Na4P1VLf0WO]Nb4l0ULk0UO\\Ne4i0ULm0TO\\Ng4f0SLQ1SO\\Nj4b0QLV1RO[Nl4>QLY1PO\\No4:oK^1oNZNR57mKb1nNcNk4KVLd1mNlNd4_O]Li1lNTO]4ROeLm1nNZOT4hNmLP2oN^On3aNQMT2QO[Om3aNQMV2ROYOk3bNQMX2TOUOk3cNoL[2UOSOk3bNnL^2WOoNj3dNnL_2XOmNi3dNmLb2ZOiNi3eNkLe2\\OeNh3gNkLf2]OcNg3gNjLi2_O_Ng3hNhLl2A\\Ne3iNiLm2AZNf3iNgLP3CWNe3iNfLS3ESNd3kNeLU3GoMd3lNdLW3HmMc3lNbL[3KhMb3nN`L^3MeMb3mN_L`30bMa3nN\\Ld33^M_3oN\\Lf34[M`3oNYLj37WM_3oNXLm38TM_3POVLP4;oL_3QOSLT4>kL^3QORLW4?hL^3ROPLZ4b0dL]3ROoK]4c0aL^3ROlKa4f0]L\\3SOlKc4g0ZL]3m5cLSJ\\3n5dLQJ[3Q6dLoI\\3R6dLnI[3S6dLmI[3U6eLkIZ3V6eLjI[3W6eLiIY3Y6gLfIY3Z6gLfIY3[6gLeIW3]6hLcIX3^6hLbIW3_6iL`IV3b6iL_IV3b6jL]IV3d6iL]IU3e6kLZIU3g6jLYIV3h6jLXIT3j6lLUIT3l6kLTIU3m6kLRIT3P7kLQIT3P7lLoHT3R7lLmHS3U7lLlHS3U7mLjHR3X7mLhHR3Z7nLfHQ3[7nLeHQ3\\7PMcHP3^7oLbHP3`7PM`Ho2_OhKe6X1lIo2^OkKg6V1jIo2]OmKj6S1jIn2[OQLl6P1iIo2YOSLo6n0gIn2YOVLQ7k0gIn2VOYLT7i0eIm2WO[LU7g0dIm2VO^LW7e0bIm2UO`LZ7b0bIl2SOdL\\7`0`Il2ROfL_7=_Il2QOiLa7;^Ik2oNlLd78]Ik2oNnLe77[Ik2nNPMh74[Ij2lNTMj72YIj2jNWMn7NXIj2hN[MQ8KWIi2eN_MU8GVIi2bNdMX8DUIh2`NgM\\8@TIh2^NkM_8]OSIf2\\NPNb8YORIg2YNSNf8VOPIa4Q7^KPIa4Q7_KnHa4S7^KmHb4T7^KlHa4U7^KkHb4V7^KiHb4X7]KhHc4Y7]KgHb4Z7]KfHc4[7]KdHc4]7\\KdHc4]7]KbHc4^7]KcHb4]7_KbHa4_7^KbHa4^7`KaH`4_7`KbH_4^7bKaH^4`7aKaH^4_7cK`H]4`7cKaH\\4_7eK`H[4a7dK_H\\4a7eK_HZ4a7fK_HZ4a7gK_HX4b7gK^HY4b7hK^HW4b7iK^HW4b7jK^HU4c7jK]HV4c7jK^HU4b7lK]HT4c7lK^HS4c7mK\\HS4d7mK]HR4d7mK\\HS4d7nK\\HQ4e7nK[HQ4f7PLZHn3g7RLYHl3j7TLVHj3k7nKbF_Oc1a4n7lKcFB`1`4n7jKgFFZ1^4R8hKiFHV1^4R8gKlFKQ1\\4V8dKfHZ4o9N3N1N3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N2M2M4I7J6J6I7J6I7J6I7J5K^^S3"}, "label": "a woman with a skirt on and a helmet sitting on her bike looking directly at the camera"}]}
{"id": 255495, "image": "COCO_train2014_000000255495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle with a basket on its front\"."}], "task": "refering", "answer_template": "The \"the bicycle with a basket on its front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 198, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 264, 265, 266, 287, 288, 289, 290, 291, 313, 314, 315, 316, 338, 339, 362, 363, 385, 386], "bbox": [0.493265625, 0.4164695009242144, 0.791609375, 0.8870609981515711], "iscrowd": 0, "area": 9760.052199999996, "rle": {"size": [541, 640], "counts": "PYW52k`00O1O2N1O2N1O2O0SMFVE<g:FXE;g:FXE<f:EYE<e:FZE<d:E\\E;c:F\\E;c:F\\E<a:F^E;a:F^E<`:E_E<_:F`E<^:EbE;_:D`E>_:B`E?a:@^Eb0a:]O_Ed0a:\\O^Ee0c:ZO]Eg0b:YO]Eh0c:XO\\Ej0d:UO[El0e:TOZEn0f:QOYEP1g:POgDc1X;]NjDa1W;^NkDa1T;_NnD_1S;nN`DQ1`;TO]Dk0b;UO`Di0a;VOaDi0^;WOdDg0];XOdDh0[;YOfDe0`;UObDj0P<cNRD[1j=NkNfNiBY1U=hNlBW1T=jNmBU1Q=lNQCR1o<oNRCo0m<ROUCm0i<UOWCj0i<VOYCi0e<XO]Cf0b<\\O_Cb0a<^OaCa0]<AdC=\\<CeC=Y<EhC9W<HkC6U<JmC5Q<MPD1o;0RD0m;1TDMk;4WDKg;7YDHg;8YDIe;8[DHe;8[DHd;9\\DHb;9^DGb;9^DH`;:_DF`;;`DF_;:aDF^;;bDF\\;;dDE\\;;dDFZ;;fDEZ;<eDEY;<gDDX;=hDDW;<iDDV;=jDDT;>kDBU;>kDBT;`0kDAT;`0kD@T;a0lD@R;b0lD_OT;b0kD_OS;b0mD^OR;d0mD]OR;c0nD]OQ;e0nD\\OP;f0oDZOQ;f0oD[OP;f0nD[OR;e0nD\\OR;d0mD\\OT;d0kD\\OU;d0kD]OU;c0jD]OV;d0jD\\OV;c0kD\\OV;d0jD\\OU;d0lD[OU;d0lD\\OT;d0lD[OT;e0mD[OS;d0mD\\OS;d0nD\\OR;c0oD\\OR;d0nD\\OQ;d0PE[OQ;d0PE\\Oo:d0RE[Oo:e0QEZOP;e0QE[On:e0RE[Oo:d0RE\\Om:e0SEZOn:e0SE[Om:d0SE\\Om:d0TE\\Ol:c0UE\\Ol:d0TE\\Ok:d0VE[Ok:d0UE]Oj:c0WE\\Oj:c0WE]Oi:b0XE]Oh:b0ZE^Of:a0ZE_Of:a0\\E]Oe:b0^E\\Ob:c0`E[Og:=\\EBj:7YEFo:2SEMS;LPE1W;HkD7o=N3N1Nfe36oYL9F:F;L3M3N3L3N0O2O0O2N2O0O2O0O2N101N6K6I7JN10001O0O2O2N3L4K5E;F:E;F:E;ESkU2"}, "label": "the bicycle with a basket on its front"}]}
{"id": 255495, "image": "COCO_train2014_000000255495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front of bicycle with wire basket attached to handlebars\"."}], "task": "refering", "answer_template": "The \"front of bicycle with wire basket attached to handlebars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 198, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 264, 265, 266, 287, 288, 289, 290, 291, 313, 314, 315, 316, 338, 339, 362, 363, 385, 386], "bbox": [0.493265625, 0.4164695009242144, 0.791609375, 0.8870609981515711], "iscrowd": 0, "area": 9760.052199999996, "rle": {"size": [541, 640], "counts": "PYW52k`00O1O2N1O2N1O2O0SMFVE<g:FXE;g:FXE<f:EYE<e:FZE<d:E\\E;c:F\\E;c:F\\E<a:F^E;a:F^E<`:E_E<_:F`E<^:EbE;_:D`E>_:B`E?a:@^Eb0a:]O_Ed0a:\\O^Ee0c:ZO]Eg0b:YO]Eh0c:XO\\Ej0d:UO[El0e:TOZEn0f:QOYEP1g:POgDc1X;]NjDa1W;^NkDa1T;_NnD_1S;nN`DQ1`;TO]Dk0b;UO`Di0a;VOaDi0^;WOdDg0];XOdDh0[;YOfDe0`;UObDj0P<cNRD[1j=NkNfNiBY1U=hNlBW1T=jNmBU1Q=lNQCR1o<oNRCo0m<ROUCm0i<UOWCj0i<VOYCi0e<XO]Cf0b<\\O_Cb0a<^OaCa0]<AdC=\\<CeC=Y<EhC9W<HkC6U<JmC5Q<MPD1o;0RD0m;1TDMk;4WDKg;7YDHg;8YDIe;8[DHe;8[DHd;9\\DHb;9^DGb;9^DH`;:_DF`;;`DF_;:aDF^;;bDF\\;;dDE\\;;dDFZ;;fDEZ;<eDEY;<gDDX;=hDDW;<iDDV;=jDDT;>kDBU;>kDBT;`0kDAT;`0kD@T;a0lD@R;b0lD_OT;b0kD_OS;b0mD^OR;d0mD]OR;c0nD]OQ;e0nD\\OP;f0oDZOQ;f0oD[OP;f0nD[OR;e0nD\\OR;d0mD\\OT;d0kD\\OU;d0kD]OU;c0jD]OV;d0jD\\OV;c0kD\\OV;d0jD\\OU;d0lD[OU;d0lD\\OT;d0lD[OT;e0mD[OS;d0mD\\OS;d0nD\\OR;c0oD\\OR;d0nD\\OQ;d0PE[OQ;d0PE\\Oo:d0RE[Oo:e0QEZOP;e0QE[On:e0RE[Oo:d0RE\\Om:e0SEZOn:e0SE[Om:d0SE\\Om:d0TE\\Ol:c0UE\\Ol:d0TE\\Ok:d0VE[Ok:d0UE]Oj:c0WE\\Oj:c0WE]Oi:b0XE]Oh:b0ZE^Of:a0ZE_Of:a0\\E]Oe:b0^E\\Ob:c0`E[Og:=\\EBj:7YEFo:2SEMS;LPE1W;HkD7o=N3N1Nfe36oYL9F:F;L3M3N3L3N0O2O0O2N2O0O2O0O2N101N6K6I7JN10001O0O2O2N3L4K5E;F:E;F:E;ESkU2"}, "label": "front of bicycle with wire basket attached to handlebars"}]}
{"id": 574983, "image": "COCO_train2014_000000574983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in front\"."}], "task": "refering", "answer_template": "The \"the zebra in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 63, 64, 65, 66, 77, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 166, 167, 169, 170, 171, 184, 185, 187, 188, 189, 203, 204, 205, 206, 207, 223, 224], "bbox": [0.22, 0.18805405405405404, 0.6966600000000001, 0.9876216216216215], "iscrowd": 0, "area": 31231.18755, "rle": {"size": [370, 500], "counts": "dmW12^;6J6J5K6J6J5K6J6J5K5K5K4L5L3M4L3M4L3M4M2M4L3M5K>B>B?B=B>B9G8H3M100O001O1000O101O0O10000O10001UNgJ[MY5b2mJZMT5c2RKZMn4c2YKXMh4f2]KTMf4j2_KQMc4m2cKoL^4n2gKoLY4o2lKnLT4P3RLlLn3R3WLkLi3S3]LhLd3V3`LhL`3V3cLiL^3T3eLkL[3U3fLjLZ3V3gLiLY3X3`2000001O000001O000001O000001O0001O00000001O00000000000PKmLa1S3WNVMh1k2nM^MQ2c2gMfMX2Z2_MoMa2Q2WMXNh2h1PM`NP3Q501O0O100000kN^KYKc4d4bKZK^4b4iK[KW4b4oKZKQ4d4ULYKk3d4ZLZKf3d4_LXKb3g4mLkJS3U5e1O1O0H9@`0_Oa0LCZLVIc3m6`LPI^3Q7gLkHV3X7nLdHP3]7d0O2[JRLa3Q4oKbLl3`3eKSMT4Z3RKZMh4m4J6K1N3M2O2M010O010O01iNQLkJn3Q5YLlJh3o4_LoJ`3m4gLPKZ3k4mLSKR3i4UMSKm2g4[MTKg2h4`MRKc2j4dMQK_2i4W2O2M2N3M2O2M2N3M2N3N1N3N1N2O2M2O2M2O2N0O2O1N101N2O001N2O0O2O1N101O0O21O2M2O101O0O101O1O1O1O1N2O1O2N1O1O1N2O2N2N2N2N2O1001O000O1N2M3N2N2NO001O1O001OSMaHi1^7VNjHd1U7\\NSI\\1m6cN[IW1d6iNcIQ1\\6nNmIj0U6TORJe0P6XOXJ4W6JQJB]6;_2N2M3M3N2M3Nhjf1"}, "label": "the zebra in front"}]}
{"id": 574983, "image": "COCO_train2014_000000574983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra without a boner\"."}], "task": "refering", "answer_template": "The \"the zebra without a boner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 63, 64, 65, 66, 77, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 166, 167, 169, 170, 171, 184, 185, 187, 188, 189, 203, 204, 205, 206, 207, 223, 224], "bbox": [0.22, 0.18805405405405404, 0.6966600000000001, 0.9876216216216215], "iscrowd": 0, "area": 31231.18755, "rle": {"size": [370, 500], "counts": "dmW12^;6J6J5K6J6J5K6J6J5K5K5K4L5L3M4L3M4L3M4M2M4L3M5K>B>B?B=B>B9G8H3M100O001O1000O101O0O10000O10001UNgJ[MY5b2mJZMT5c2RKZMn4c2YKXMh4f2]KTMf4j2_KQMc4m2cKoL^4n2gKoLY4o2lKnLT4P3RLlLn3R3WLkLi3S3]LhLd3V3`LhL`3V3cLiL^3T3eLkL[3U3fLjLZ3V3gLiLY3X3`2000001O000001O000001O000001O0001O00000001O00000000000PKmLa1S3WNVMh1k2nM^MQ2c2gMfMX2Z2_MoMa2Q2WMXNh2h1PM`NP3Q501O0O100000kN^KYKc4d4bKZK^4b4iK[KW4b4oKZKQ4d4ULYKk3d4ZLZKf3d4_LXKb3g4mLkJS3U5e1O1O0H9@`0_Oa0LCZLVIc3m6`LPI^3Q7gLkHV3X7nLdHP3]7d0O2[JRLa3Q4oKbLl3`3eKSMT4Z3RKZMh4m4J6K1N3M2O2M010O010O01iNQLkJn3Q5YLlJh3o4_LoJ`3m4gLPKZ3k4mLSKR3i4UMSKm2g4[MTKg2h4`MRKc2j4dMQK_2i4W2O2M2N3M2O2M2N3M2N3N1N3N1N2O2M2O2M2O2N0O2O1N101N2O001N2O0O2O1N101O0O21O2M2O101O0O101O1O1O1O1N2O1O2N1O1O1N2O2N2N2N2N2O1001O000O1N2M3N2N2NO001O1O001OSMaHi1^7VNjHd1U7\\NSI\\1m6cN[IW1d6iNcIQ1\\6nNmIj0U6TORJe0P6XOXJ4W6JQJB]6;_2N2M3M3N2M3Nhjf1"}, "label": "the zebra without a boner"}]}
{"id": 574983, "image": "COCO_train2014_000000574983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra behind the zebra facing the camera\"."}], "task": "refering", "answer_template": "The \"the zebra behind the zebra facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 77, 78, 79, 80, 84, 102, 103, 118, 119, 120, 121, 135, 136, 137, 138, 139, 155, 156, 157, 173, 174, 175, 191, 192, 209], "bbox": [0.28716, 0.24500000000000002, 0.7491, 0.8951081081081081], "iscrowd": 0, "area": 10507.025949999996, "rle": {"size": [370, 500], "counts": "_Td12^;3M3N2M3ZEFQ:<jEHT:;gEIX:i0N2O1N201O0O100000000O2O00000O10001O000O1000001O0001O001O0010O01O001O0010O01O001O0010O01O001O001O001O001O001O001O1O001O001O0O2O001O001O001O001O1O001O001O1O0AYE3h:K^E0b:NeEM\\:1Uk=0\\UB2N3M2N3M2N3M2N3L3N2N3M2N2N1O2N1O1O2N1O2N1O2O0101N2N2O1N2O1N2O1N2N2O1N2Q1oN5K0000O1J6J6I7MbHeNk4[1RKfNP5Z1mJhNS5X1kJiNW5W1fJjN\\5U1aJmN`5T1]JmNe5R1YJoNi5Q1TJQOl5o0QJSOn5P1oIQOQ6P1mIQOS6Q1jIPOU6T1fInNZ6U1bImN]6U1_IcNiNLg7e1RIhN\\O@b7k1dHnNOTO]7W3?O2NaMUM[Li2h3[MRLd2Q4`MjK^2Y4fMbKX2a4lMYKS2i4RNRKl1Q5XNjJf1Y5W23M2N3M3M3M3M2aNlI^MW6^2a1M3L40O9H8H8H9G6I1000000ObM`Hl0_7QOhHk0X7ROoHj0Q7SOUIi0l6TO[Ic0i6ZO_I6o6EXI2S7ETI3V7DPI5Z7BmH6^9Hdm\\1"}, "label": "the zebra behind the zebra facing the camera"}]}
{"id": 574983, "image": "COCO_train2014_000000574983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing behind another zebra\"."}], "task": "refering", "answer_template": "The \"zebra standing behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 77, 78, 79, 80, 84, 102, 103, 118, 119, 120, 121, 135, 136, 137, 138, 139, 155, 156, 157, 173, 174, 175, 191, 192, 209], "bbox": [0.28716, 0.24500000000000002, 0.7491, 0.8951081081081081], "iscrowd": 0, "area": 10507.025949999996, "rle": {"size": [370, 500], "counts": "_Td12^;3M3N2M3ZEFQ:<jEHT:;gEIX:i0N2O1N201O0O100000000O2O00000O10001O000O1000001O0001O001O0010O01O001O0010O01O001O0010O01O001O001O001O001O001O001O1O001O001O0O2O001O001O001O001O1O001O001O1O0AYE3h:K^E0b:NeEM\\:1Uk=0\\UB2N3M2N3M2N3M2N3L3N2N3M2N2N1O2N1O1O2N1O2N1O2O0101N2N2O1N2O1N2O1N2N2O1N2Q1oN5K0000O1J6J6I7MbHeNk4[1RKfNP5Z1mJhNS5X1kJiNW5W1fJjN\\5U1aJmN`5T1]JmNe5R1YJoNi5Q1TJQOl5o0QJSOn5P1oIQOQ6P1mIQOS6Q1jIPOU6T1fInNZ6U1bImN]6U1_IcNiNLg7e1RIhN\\O@b7k1dHnNOTO]7W3?O2NaMUM[Li2h3[MRLd2Q4`MjK^2Y4fMbKX2a4lMYKS2i4RNRKl1Q5XNjJf1Y5W23M2N3M3M3M3M2aNlI^MW6^2a1M3L40O9H8H8H9G6I1000000ObM`Hl0_7QOhHk0X7ROoHj0Q7SOUIi0l6TO[Ic0i6ZO_I6o6EXI2S7ETI3V7DPI5Z7BmH6^9Hdm\\1"}, "label": "zebra standing behind another zebra"}]}
{"id": 476681, "image": "COCO_train2014_000000476681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red tee shirt rollerblading\"."}], "task": "refering", "answer_template": "The \"a man in a red tee shirt rollerblading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 292], "bbox": [0.6116406249999999, 0.19350467289719625, 0.7946718749999999, 0.8728504672897196], "iscrowd": 0, "area": 15735.52965, "rle": {"size": [428, 640], "counts": "_hS52Y=c0]O5K1O1N3N1M3M3M3M3M3J6I7J6I8H6J7]O\\M\\Ff2a9_M[Fc2c9`MZFa2e9cMWF_2g9`02N1A?1NG92N3M2N2M3O0101N1O2N1O\\N^McH^2\\7oM[He1m7hNkGW1Q8WOeGk0Y8CZG>e8U2N2O2M2N2O1N2O0nIZK`3b4^KfL_4V3aKPM^4k2cK[M[4`2eKfMZ4W2dKoM[4n1cKVN^4f1aK]Na4`1^KbNd4\\1YKfNj4W1TKlNn4_1bJcN`5c42N2M3N1N101N2O0O0_OMd0K62J6B>E;M3M3L4M3L4N2O1M3N2O1N2O1N2O1O11O0010O0100O01hKRHQ3m7jLXHV3i7dL]H\\3b7aLbH^3_7^LeHa3]7[LfH_3f8J7J6J6J6L3L5H8[Oe0I7J9D;B?Gecf1"}, "label": "a man in a red tee shirt rollerblading"}]}
{"id": 476681, "image": "COCO_train2014_000000476681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in red t - shirt does skating on the road\"."}], "task": "refering", "answer_template": "The \"a man in red t - shirt does skating on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 292], "bbox": [0.6116406249999999, 0.19350467289719625, 0.7946718749999999, 0.8728504672897196], "iscrowd": 0, "area": 15735.52965, "rle": {"size": [428, 640], "counts": "_hS52Y=c0]O5K1O1N3N1M3M3M3M3M3J6I7J6I8H6J7]O\\M\\Ff2a9_M[Fc2c9`MZFa2e9cMWF_2g9`02N1A?1NG92N3M2N2M3O0101N1O2N1O\\N^McH^2\\7oM[He1m7hNkGW1Q8WOeGk0Y8CZG>e8U2N2O2M2N2O1N2O0nIZK`3b4^KfL_4V3aKPM^4k2cK[M[4`2eKfMZ4W2dKoM[4n1cKVN^4f1aK]Na4`1^KbNd4\\1YKfNj4W1TKlNn4_1bJcN`5c42N2M3N1N101N2O0O0_OMd0K62J6B>E;M3M3L4M3L4N2O1M3N2O1N2O1N2O1O11O0010O0100O01hKRHQ3m7jLXHV3i7dL]H\\3b7aLbH^3_7^LeHa3]7[LfH_3f8J7J6J6J6L3L5H8[Oe0I7J9D;B?Gecf1"}, "label": "a man in red t - shirt does skating on the road"}]}
{"id": 329230, "image": "COCO_train2014_000000329230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfboard stuck in the sand\"."}], "task": "refering", "answer_template": "The \"a surfboard stuck in the sand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 149, 150, 151, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219], "bbox": [0.3132579185520362, 0.35575, 0.7373076923076923, 0.60784375], "iscrowd": 0, "area": 7710.590700000001, "rle": {"size": [640, 442], "counts": "[Sg23mc02N100h\\OKdb06Y]OOcb02Z]O4cb0L[]O6db0LY]O6gb0JW]O7ib0KS]O8mb0=100L4J6I7NAJQ]O;nb0;101N20O100O110O01O0010O000O100O2O0O100O2O000O1O1@a0N1O1O1O010O100O13M2N3M3M2N3M1O00O100O100O100O10000O100L4K5J6L40000002N2N2N2N2N2N2N2N0O100000O100000000O10O100000O1000000000O10000000000000000O0100000000000000000O10000000000000000O10O10000000000000O1000000000000000000O10O100000O1O1O1O1O1O1O1N3N1O1O2N1O1O2N1O1O1O2N1O1N3N1O2N2N2N2N2N2N3MPhX2"}, "label": "a surfboard stuck in the sand"}]}
{"id": 329230, "image": "COCO_train2014_000000329230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfboard in the sand in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the surfboard in the sand in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 149, 150, 151, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219], "bbox": [0.3132579185520362, 0.35575, 0.7373076923076923, 0.60784375], "iscrowd": 0, "area": 7710.590700000001, "rle": {"size": [640, 442], "counts": "[Sg23mc02N100h\\OKdb06Y]OOcb02Z]O4cb0L[]O6db0LY]O6gb0JW]O7ib0KS]O8mb0=100L4J6I7NAJQ]O;nb0;101N20O100O110O01O0010O000O100O2O0O100O2O000O1O1@a0N1O1O1O010O100O13M2N3M3M2N3M1O00O100O100O100O10000O100L4K5J6L40000002N2N2N2N2N2N2N2N0O100000O100000000O10O100000O1000000000O10000000000000000O0100000000000000000O10000000000000000O10O10000000000000O1000000000000000000O10O100000O1O1O1O1O1O1O1N3N1O1O2N1O1O2N1O1O1O2N1O1N3N1O2N2N2N2N2N2N3MPhX2"}, "label": "the surfboard in the sand in the right hand picture"}]}
{"id": 566798, "image": "COCO_train2014_000000566798.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with red necklaces and a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman with red necklaces and a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 360, 361, 362, 363, 364], "bbox": [0.40632812500000004, 0.24141666666666667, 0.8752812500000001, 0.9276458333333333], "iscrowd": 0, "area": 59454.80664999998, "rle": {"size": [480, 640], "counts": "jWj31n>2N2N2M2O2N2N2M2O2N2N2N1N3N2N2N1N3N2N2N1N3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M2N2N2N2N2M3N3M2N2N2O1N2O1N3N1O1N2O1N2O1N3N1O1N2O1N2O1O2M200O1O1O100O2N1O100O1O1O101N1O100O1O1O101N1O1O100O1O101N1O10000O100O2O000O100O10000O100O10000O10O0100M3M3M3M3M3cN]1J6J6J6L4O1N101N2O1N2O1N2O2M2O1N2O1O1N3N2M3N2M3N2M3N2M3N2M3O1O1O1O1O10O01O1O0000000010O00000001O0001O000001O000001O01O000000000010O00000001O0001O000001O000001O010O1O001O0010O01O1O001O001O1N101O001O001O1O0O2O001O2N1O1O1N2kI[Jm2g5oL[JQ3f5lL]JS3d5iL_JW3b5fL`JY3b5dLaJ[3a5`LbJ`3_5]LdJa3^5\\LdJd3b5SLaJm3h5hK[JV4P6]KSJc4W6nJnIR5[6aJjI]5m70001N10001O1O1O010O1O100O00100O1O010O1O100O00100O1O10O01O1O1N101O1N2O1N1iNcFkL^9W1fHbN[7Z1PI]NR7^1[3L3M4K5L4K4M4J6J6I7JTWU1"}, "label": "a woman with red necklaces and a white shirt"}]}
{"id": 566798, "image": "COCO_train2014_000000566798.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white with a sly look\"."}], "task": "refering", "answer_template": "The \"a woman in white with a sly look\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 360, 361, 362, 363, 364], "bbox": [0.40632812500000004, 0.24141666666666667, 0.8752812500000001, 0.9276458333333333], "iscrowd": 0, "area": 59454.80664999998, "rle": {"size": [480, 640], "counts": "jWj31n>2N2N2M2O2N2N2M2O2N2N2N1N3N2N2N1N3N2N2N1N3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M2N2N2N2N2M3N3M2N2N2O1N2O1N3N1O1N2O1N2O1N3N1O1N2O1N2O1O2M200O1O1O100O2N1O100O1O1O101N1O100O1O1O101N1O1O100O1O101N1O10000O100O2O000O100O10000O100O10000O10O0100M3M3M3M3M3cN]1J6J6J6L4O1N101N2O1N2O1N2O2M2O1N2O1O1N3N2M3N2M3N2M3N2M3N2M3O1O1O1O1O10O01O1O0000000010O00000001O0001O000001O000001O01O000000000010O00000001O0001O000001O000001O010O1O001O0010O01O1O001O001O1N101O001O001O1O0O2O001O2N1O1O1N2kI[Jm2g5oL[JQ3f5lL]JS3d5iL_JW3b5fL`JY3b5dLaJ[3a5`LbJ`3_5]LdJa3^5\\LdJd3b5SLaJm3h5hK[JV4P6]KSJc4W6nJnIR5[6aJjI]5m70001N10001O1O1O010O1O100O00100O1O010O1O100O00100O1O10O01O1O1N101O1N2O1N1iNcFkL^9W1fHbN[7Z1PI]NR7^1[3L3M4K5L4K4M4J6J6I7JTWU1"}, "label": "a woman in white with a sly look"}]}
{"id": 132626, "image": "COCO_train2014_000000132626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycles side view mirror in front of the woman with the green jacket\"."}], "task": "refering", "answer_template": "The \"the motorcycles side view mirror in front of the woman with the green jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 120, 121, 122, 123, 145, 146, 167, 168, 169, 192, 193, 215, 216], "bbox": [0.24431250000000002, 0.29370023419203745, 0.403765625, 0.6828103044496487], "iscrowd": 0, "area": 4185.142850000002, "rle": {"size": [427, 640], "counts": "`cQ29j<8N2O100O1O1O1O1O101N1000000000000000000O100000000^EXOa8h0^G[O`8e0_G_O^8a0aGB]8>bGC^8=aGC`8=_GCa8>_GAb8?]GAd8?[GAf8?YGAh8?WGAj8?UGAl8`0RG@o8`0PG@Q9`0oF^OS9b0lF^OU9c0iF]OW9d0hF\\OZ9d0dF\\O]9e0aF[Oa9d0^F\\Od9d0ZF\\Oh9d0WF[Oj9e0UF[Om9e0QF[OQ:f0lEZOV:f0hEZOY:g0eEXO^:i0_EWOc:i0[EWOg:i0XEVOi:k0UEUOm:n0nDROT;P1hDPOY;S1cDmN_;^11O1O001O1N2O001O1O1O001O1O1O001004M3N2N10O01OkNSEn1X:3LN3M2N3M2N3J5J7I6J7H7J6J6J6J6JhPo4"}, "label": "the motorcycles side view mirror in front of the woman with the green jacket"}]}
{"id": 132626, "image": "COCO_train2014_000000132626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the windshield of the motorcycle in the front of the line\"."}], "task": "refering", "answer_template": "The \"the windshield of the motorcycle in the front of the line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.22742187500000002, 0.1800234192037471, 0.70634375, 0.9881498829039813], "iscrowd": 0, "area": 58544.79095000001, "rle": {"size": [427, 640], "counts": "XXm12Z=1N2N2N3N1N2N2N2O0O1O1O100O1O1O100O1O1O10O0000001O01O000001O0000001O0000001O0000001OO2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1N2O1O1O1O1O1O1O1O1O1N2M3N3M2M3N2M3N2M3N2N2M3N2M3N2bJYM5j2CjMOX2I\\NIf1OjNFY15oNFT16RODQ19VOAl0<ZO^Oi0?]O\\Oe0b0@YOb0d0DXO=e0JUO8h0NSO4m0OoN2P12kN0U13fNOY15cNL]18nLgKX1S4i1:nLdKT1T4n1;mLbKQ1T4Q2>nL^Kl0V4V2?mL\\Ki0V4Z2a0lLZKe0W4^2k0\\MWOd2l0XMUOg2o0TMROm2Q1nLQOQ3e50000O100O100O100O10000O100O100O100O10000O10O0100O10000O100O100O100O10000O100O100000000001O000O100000001O0000000000001O0000000000001O00000000000O2O00000000001O0000000000001O0000000000001O00000O1000001O00000000001O002N2N1O2N2N1O2N2N1O2N1O2N2N1N3N2bJ^Lo0c3hNhLU1Z3`NQM^1Q3WN[Mf1f2PNfMm1\\2hMmMX2U2`MQN`2P2\\MRNe2P2WMRNi2P2RMSNn2n1oLTNQ3n1jLUNV3m1eLVN[3k1`LYN[O^MZ3[4VMYNYOfM_3S4SMZN^OeM]3R4PM]NDbMZ3S4mL^NIaMX3R4mL^NLaMU3S4mL\\NNcMS3S4nLYN0eMP3S4oLXN1hMm2R4QMTN4kMh2T4RMQN6mMf2R4UMPN6oMc2Q4WMPN6QNa2o3YMPN7RN^2m3]MPN5UN\\2k3_MQN5UNZ2j3bMPN4XNX2h3dMPN5YNU2f3hMQN3ZNS2d3kMRN2\\NQ2a3oMRN1^Nn1`3QNRN1`Nl1]3TNTN0`Nj1[3XNWNK`Nk1X3[NdN_OUNT2W3^Na0`1^OaNd0]1\\OdNe0Z1ZOgNh0W1XOiNj0U1UOmNl0Q1UOnNm0P1SOPOo0n0QOSOS1h0mNXOZ1`0gN@_1:aNFf13ZNMl1MUN2R2GnM9Y2@gM`0_2ZOaMg0a7000002N1O1N2K6J5K5KoW^2"}, "label": "the windshield of the motorcycle in the front of the line"}]}
{"id": 132626, "image": "COCO_train2014_000000132626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red and silver motorcycle is leading the pack down the street\"."}], "task": "refering", "answer_template": "The \"the red and silver motorcycle is leading the pack down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.22742187500000002, 0.1800234192037471, 0.70634375, 0.9881498829039813], "iscrowd": 0, "area": 58544.79095000001, "rle": {"size": [427, 640], "counts": "XXm12Z=1N2N2N3N1N2N2N2O0O1O1O100O1O1O100O1O1O10O0000001O01O000001O0000001O0000001O0000001OO2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1N2O1O1O1O1O1O1O1O1O1N2M3N3M2M3N2M3N2M3N2N2M3N2M3N2bJYM5j2CjMOX2I\\NIf1OjNFY15oNFT16RODQ19VOAl0<ZO^Oi0?]O\\Oe0b0@YOb0d0DXO=e0JUO8h0NSO4m0OoN2P12kN0U13fNOY15cNL]18nLgKX1S4i1:nLdKT1T4n1;mLbKQ1T4Q2>nL^Kl0V4V2?mL\\Ki0V4Z2a0lLZKe0W4^2k0\\MWOd2l0XMUOg2o0TMROm2Q1nLQOQ3e50000O100O100O100O10000O100O100O100O10000O10O0100O10000O100O100O100O10000O100O100000000001O000O100000001O0000000000001O0000000000001O00000000000O2O00000000001O0000000000001O0000000000001O00000O1000001O00000000001O002N2N1O2N2N1O2N2N1O2N1O2N2N1N3N2bJ^Lo0c3hNhLU1Z3`NQM^1Q3WN[Mf1f2PNfMm1\\2hMmMX2U2`MQN`2P2\\MRNe2P2WMRNi2P2RMSNn2n1oLTNQ3n1jLUNV3m1eLVN[3k1`LYN[O^MZ3[4VMYNYOfM_3S4SMZN^OeM]3R4PM]NDbMZ3S4mL^NIaMX3R4mL^NLaMU3S4mL\\NNcMS3S4nLYN0eMP3S4oLXN1hMm2R4QMTN4kMh2T4RMQN6mMf2R4UMPN6oMc2Q4WMPN6QNa2o3YMPN7RN^2m3]MPN5UN\\2k3_MQN5UNZ2j3bMPN4XNX2h3dMPN5YNU2f3hMQN3ZNS2d3kMRN2\\NQ2a3oMRN1^Nn1`3QNRN1`Nl1]3TNTN0`Nj1[3XNWNK`Nk1X3[NdN_OUNT2W3^Na0`1^OaNd0]1\\OdNe0Z1ZOgNh0W1XOiNj0U1UOmNl0Q1UOnNm0P1SOPOo0n0QOSOS1h0mNXOZ1`0gN@_1:aNFf13ZNMl1MUN2R2GnM9Y2@gM`0_2ZOaMg0a7000002N1O1N2K6J5K5KoW^2"}, "label": "the red and silver motorcycle is leading the pack down the street"}]}
{"id": 534037, "image": "COCO_train2014_000000534037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich in the plate\"."}], "task": "refering", "answer_template": "The \"sandwich in the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313], "bbox": [0.006744730679156909, 0.49214062500000005, 1.0, 0.87865625], "iscrowd": 0, "area": 88707.21825000002, "rle": {"size": [640, 427], "counts": "WZ23lc04L4M3L4L4L4L3M4L3M3M3M3N2M3M2N3M3M1O2N1O2N1O2O0O2N1O2N1O2N2N1I8F9hMRMVCW3g<QMjBX3T=PM^BY3^=QMSBW3k=QMfAY3W>Y1N3O001O001O001O0O2O1O001O001O001O0O2O001O001O001O0O2O001O1O001O001O0O2O0000001O0000000O101O000000001O0O100000001O0000000O2O000000001O00000O10001O000000001O0O100000001O00000O101O000000001O000O1000001O0000000010O0000000001O000000010O000000001O00000001O01O000000001O0001O000001O000000001O01O000001O000000000010O000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000O100000000000001O000000000000000000000000000000001O00000000000000000000000000000000001O0000000000000000000000000O1000001O0000000000000000000000000001O00001O00001O0000001O00001N10001O0000001O00001O000O2O0000001O00001O0000001O0O101O00001O0000001O00001O0O101O0000001O00001O0000001N10001O00001O0000001O000O2O00001O0000001O001O2N1O2M3N2N1O2YOg0jMU2jMlF"}, "label": "sandwich in the plate"}]}
{"id": 534037, "image": "COCO_train2014_000000534037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich half behind another sandwich\"."}], "task": "refering", "answer_template": "The \"a sandwich half behind another sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 135, 136, 137, 138, 139, 140, 141, 142, 143, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 180, 181, 195, 196, 210, 211, 225, 226, 240], "bbox": [0.0033723653395784543, 0.368546875, 0.6433255269320843, 0.707859375], "iscrowd": 0, "area": 22991.95515, "rle": {"size": [640, 427], "counts": "Pl0>bc0]4cKY1fN10O1O10O01O100O00100O1O10O01O100O00100O1O10O01O100O001O100O00100O1O10OL5J6I7J5K6I7J6J6I6I8H8H8G8O2O1000O010000O10O10O1000O010000O10O10O1000O010000O10O10O1000O010000O01000O10000O2O000O101O0O10001N10000O2O000O101O0O10001N10000O2O000O2O000O101O0000001O01O01O0001O01O01O01O0001O01O00010O0000010O0000010O000010O000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001N1000001O00001O000O10001O00000000001N100000001O0000000O2O00000000001O000O10001O000000001O0O1O1O1O1O2M2O1O1N2O2L3M3M3L4M4L3MYgn2"}, "label": "a sandwich half behind another sandwich"}]}
{"id": 534037, "image": "COCO_train2014_000000534037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich in back\"."}], "task": "refering", "answer_template": "The \"sandwich in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 135, 136, 137, 138, 139, 140, 141, 142, 143, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 180, 181, 195, 196, 210, 211, 225, 226, 240], "bbox": [0.0033723653395784543, 0.368546875, 0.6433255269320843, 0.707859375], "iscrowd": 0, "area": 22991.95515, "rle": {"size": [640, 427], "counts": "Pl0>bc0]4cKY1fN10O1O10O01O100O00100O1O10O01O100O00100O1O10O01O100O001O100O00100O1O10OL5J6I7J5K6I7J6J6I6I8H8H8G8O2O1000O010000O10O10O1000O010000O10O10O1000O010000O10O10O1000O010000O01000O10000O2O000O101O0O10001N10000O2O000O101O0O10001N10000O2O000O2O000O101O0000001O01O01O0001O01O01O01O0001O01O00010O0000010O0000010O000010O000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001N1000001O00001O000O10001O00000000001N100000001O0000000O2O00000000001O000O10001O000000001O0O1O1O1O1O2M2O1O1N2O2L3M3M3L4M4L3MYgn2"}, "label": "sandwich in back"}]}
{"id": 296474, "image": "COCO_train2014_000000296474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue train\"."}], "task": "refering", "answer_template": "The \"a blue train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.557078125, 0.17160416666666667, 0.996296875, 0.9920833333333333], "iscrowd": 0, "area": 59162.76039999999, "rle": {"size": [480, 640], "counts": "caW55j>7I1O1O00100ON2D<1O1ONUA1m>0O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O101N100O10000O100O100O100O10001N100O100O100O10000O100O2O0O101N10000O2O0O101N100O1dFSOe4n0nJMc43RKg0^4ZOVK[1a4eNTKk1f4WNnJY2m4gMRKZ2m4gMSKY2m4hMQKY2n4hMRKX2n4hMRKX2n4hMRKX2m4jMRKV2n4jMRKV2m4QNmJo1S5WNfJj1Y5\\NbJd1^5bN\\J^1c5hNXJX1h5hNXJX1g5hNZJX1f5hNYJZ1f5fNZJZ1e5gN[JZ1d5fN\\J[1b5eN_J\\1`5eN_J\\1_5fN`J[1_5fN_J\\1_5fN`JZ1`5gN_JZ1`5fN`J[1^5gNaJZ1^5gNaJZ1]5hNbJY1]5hNaJY1^5iN_JZ1`5fNYJb1e5`NRJi1m5XNkIP2T6_3N2O0O2O1N2O1N2O002M2O1N3N1N2O2M2O1O2M2O1N3N1N2O1N101O1N101N101N2O0O2O001N101N2O0O2O0O2O000O100O10O0100O10000O100O010O100O10000O10O0100O100O10000O100O101N10000O101N100O100O2O000O100O2O0O100O10001N100O100O2O0O10000O2O0O100O101N10000O101N100O100O2O000O100O2O0O10000O2O0O100O101N100O1]k0"}, "label": "a blue train"}]}
{"id": 296474, "image": "COCO_train2014_000000296474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a two tone blue passenger train car\"."}], "task": "refering", "answer_template": "The \"a two tone blue passenger train car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.557078125, 0.17160416666666667, 0.996296875, 0.9920833333333333], "iscrowd": 0, "area": 59162.76039999999, "rle": {"size": [480, 640], "counts": "caW55j>7I1O1O00100ON2D<1O1ONUA1m>0O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O101N100O10000O100O100O100O10001N100O100O100O10000O100O2O0O101N10000O2O0O101N100O1dFSOe4n0nJMc43RKg0^4ZOVK[1a4eNTKk1f4WNnJY2m4gMRKZ2m4gMSKY2m4hMQKY2n4hMRKX2n4hMRKX2n4hMRKX2m4jMRKV2n4jMRKV2m4QNmJo1S5WNfJj1Y5\\NbJd1^5bN\\J^1c5hNXJX1h5hNXJX1g5hNZJX1f5hNYJZ1f5fNZJZ1e5gN[JZ1d5fN\\J[1b5eN_J\\1`5eN_J\\1_5fN`J[1_5fN_J\\1_5fN`JZ1`5gN_JZ1`5fN`J[1^5gNaJZ1^5gNaJZ1]5hNbJY1]5hNaJY1^5iN_JZ1`5fNYJb1e5`NRJi1m5XNkIP2T6_3N2O0O2O1N2O1N2O002M2O1N3N1N2O2M2O1O2M2O1N3N1N2O1N101O1N101N101N2O0O2O001N101N2O0O2O0O2O000O100O10O0100O10000O100O010O100O10000O10O0100O100O10000O100O101N10000O101N100O100O2O000O100O2O0O100O10001N100O100O2O0O10000O2O0O100O101N10000O101N100O100O2O000O100O2O0O10000O2O0O100O101N100O1]k0"}, "label": "a two tone blue passenger train car"}]}
{"id": 83487, "image": "COCO_train2014_000000083487.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown horse closer to the foreground\"."}], "task": "refering", "answer_template": "The \"a large brown horse closer to the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139], "bbox": [0.47952, 0.37888000000000005, 0.77604, 0.5840266666666667], "iscrowd": 0, "area": 6849.5522, "rle": {"size": [375, 500], "counts": "WSh24g:7aE2\\:1dEOY:4fEMW:6iEJV:7jEIU:8kEHT:9lEGS::mEFR:;nEER:;nEEQ:<oEDP:=PFCo9>XF[Oh9e0ZFWOg9j0PFUOY:l0gEUOW:k0jEWOT:i0lEYOQ:h0oE^Oj9c0VF]Oi9d0WF\\Oi9d0WF\\Oh9e0XF[Og9f0YFZOf9g0ZFYOf9g0ZFYOf9e0`FWO`9h0h00O1000000000002bESOn9n0QFYOh9g0XFYOi9e0XF[Oh9e0XF[Oh9e0XF[Oh9e0XF[Oh9e0XF[Oh9e0XF[Oi9d0WF\\Oi9d0WF\\Oi9d0WF\\Oi9d0WF\\Oj9c0VF]Oj9c0VF]Oj9c0VF]Ok9b0UF^Ok9b0UF^Ok9b0UF^Ok9b0UF^Ok9b0UF^Ol9a0TF_Ol9`0UF@k9`0UF@k9`0UF@k9`0UF@k9`0UF@k9`0UF@k9`0UF@l9?TFAl9?TFAl9?TFAl9?TFAl9?SFBm9>SFBm9>SFBm9>SFBm9>SFBm9>SFBm9>SFBm9>SFBm9=TFCl9=TFCm9<SFDm9<SFDl9=TFCl9=TFCl9=TFCl9=TFCl9=TFBm9>SFBl9?TFAl9?TFAl9?TFAk9`0UF@b9C^Fm00@a9E^Fk01@a9F\\Fk03_Ob9FZFk04_Ob9GXFk06^Oc9h0]FXOd9g0\\FYOd9g0\\FYOc9h0]FXOc9h0]FXOc9h0]FXOc9h0]FXOd9g0\\FYOd9g0\\FYOd9f0]FZOb9g0^FYOa9h0_FXOa9h0_FXO`9i0`FWO`9i0`FWO_9j0aFVO^9k0bFUO^9k0bFUO]9l0cFTO]9l0cFTO]9l0cFUO[9l0dFUO\\9k0dFUO[9l0eFTO[9l0eFTOZ9m0fFSOZ9m0fFSOZ9m0fFSOZ9m0fFSOZ9m0fFSOZ9m0fFRO[9n0eFRO\\9m0dFSO]9l0cFTO]9l0cFSO]9n0cFRO]9n0cFRO\\9o0dFQOZ9Q1fFnNY9T1gFlNX9U1hFkNX9U1hFkNW9V1iFjN\\9Q1dFjNIKf9Y1aFhNh9W1;1N2O10K5M2kNQFc0oUY1"}, "label": "a large brown horse closer to the foreground"}]}
{"id": 83487, "image": "COCO_train2014_000000083487.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown horse with ropes\"."}], "task": "refering", "answer_template": "The \"a dark brown horse with ropes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139], "bbox": [0.47952, 0.37888000000000005, 0.77604, 0.5840266666666667], "iscrowd": 0, "area": 6849.5522, "rle": {"size": [375, 500], "counts": "WSh24g:7aE2\\:1dEOY:4fEMW:6iEJV:7jEIU:8kEHT:9lEGS::mEFR:;nEER:;nEEQ:<oEDP:=PFCo9>XF[Oh9e0ZFWOg9j0PFUOY:l0gEUOW:k0jEWOT:i0lEYOQ:h0oE^Oj9c0VF]Oi9d0WF\\Oi9d0WF\\Oh9e0XF[Og9f0YFZOf9g0ZFYOf9g0ZFYOf9e0`FWO`9h0h00O1000000000002bESOn9n0QFYOh9g0XFYOi9e0XF[Oh9e0XF[Oh9e0XF[Oh9e0XF[Oh9e0XF[Oh9e0XF[Oi9d0WF\\Oi9d0WF\\Oi9d0WF\\Oi9d0WF\\Oj9c0VF]Oj9c0VF]Oj9c0VF]Ok9b0UF^Ok9b0UF^Ok9b0UF^Ok9b0UF^Ok9b0UF^Ol9a0TF_Ol9`0UF@k9`0UF@k9`0UF@k9`0UF@k9`0UF@k9`0UF@k9`0UF@l9?TFAl9?TFAl9?TFAl9?TFAl9?SFBm9>SFBm9>SFBm9>SFBm9>SFBm9>SFBm9>SFBm9>SFBm9=TFCl9=TFCm9<SFDm9<SFDl9=TFCl9=TFCl9=TFCl9=TFCl9=TFBm9>SFBl9?TFAl9?TFAl9?TFAk9`0UF@b9C^Fm00@a9E^Fk01@a9F\\Fk03_Ob9FZFk04_Ob9GXFk06^Oc9h0]FXOd9g0\\FYOd9g0\\FYOc9h0]FXOc9h0]FXOc9h0]FXOc9h0]FXOd9g0\\FYOd9g0\\FYOd9f0]FZOb9g0^FYOa9h0_FXOa9h0_FXO`9i0`FWO`9i0`FWO_9j0aFVO^9k0bFUO^9k0bFUO]9l0cFTO]9l0cFTO]9l0cFUO[9l0dFUO\\9k0dFUO[9l0eFTO[9l0eFTOZ9m0fFSOZ9m0fFSOZ9m0fFSOZ9m0fFSOZ9m0fFSOZ9m0fFRO[9n0eFRO\\9m0dFSO]9l0cFTO]9l0cFSO]9n0cFRO]9n0cFRO\\9o0dFQOZ9Q1fFnNY9T1gFlNX9U1hFkNX9U1hFkNW9V1iFjN\\9Q1dFjNIKf9Y1aFhNh9W1;1N2O10K5M2kNQFc0oUY1"}, "label": "a dark brown horse with ropes"}]}
{"id": 42529, "image": "COCO_train2014_000000042529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a burgundy rolls - royce sits at a traffic signal\"."}], "task": "refering", "answer_template": "The \"a burgundy rolls - royce sits at a traffic signal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 407, 408, 425], "bbox": [0.24506250000000002, 0.6757874762808349, 0.782890625, 0.9572485768500949], "iscrowd": 0, "area": 36560.766500000005, "rle": {"size": [527, 640], "counts": "bWa2e0h?5K5E9C<G:M3M2O2N2N101N2O0O2O1O1N3N1N2O1O1O1O2N000O101O000000001O0O1000001O0000000O10000000000000000O10000000000000000000M2N3000000000000001O0O1000000000000000001O0O1000000000000O10001O00000O100000000O1000000O100O1O1O1O100O001O1O100O1O1O1O100O1O1O00100O1O1O100O1O1O100O1O00100O1O100O1O100O2O1O1O2N1N2O1O1O1O2M2O1O00000O100000000O1000O10O100000000000000000O100O100O100O100O100O1O1N2O1N2N2O0O2O1O10000O1000000000000O10000000000O1000001O000O10000000000O10000000000O1000000000000O100000001O0O10000000000O100000001O0O100000001O0O100000001O000O1000001O000O101O0000001N1000001O000O10001O001O0O101O001O000011N101N2O00O1O1O1O100O1O2N1O1O2N1O2N1O2N1O2N1N3N2N0O1O2N1O2O0O2J5J7J6N2N2N101N2N2N1M4lNj@h0d?C\\TW2"}, "label": "a burgundy rolls - royce sits at a traffic signal"}]}
{"id": 42529, "image": "COCO_train2014_000000042529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red color car on the road\"."}], "task": "refering", "answer_template": "The \"a red color car on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 407, 408, 425], "bbox": [0.24506250000000002, 0.6757874762808349, 0.782890625, 0.9572485768500949], "iscrowd": 0, "area": 36560.766500000005, "rle": {"size": [527, 640], "counts": "bWa2e0h?5K5E9C<G:M3M2O2N2N101N2O0O2O1O1N3N1N2O1O1O1O2N000O101O000000001O0O1000001O0000000O10000000000000000O10000000000000000000M2N3000000000000001O0O1000000000000000001O0O1000000000000O10001O00000O100000000O1000000O100O1O1O1O100O001O1O100O1O1O1O100O1O1O00100O1O1O100O1O1O100O1O00100O1O100O1O100O2O1O1O2N1N2O1O1O1O2M2O1O00000O100000000O1000O10O100000000000000000O100O100O100O100O100O1O1N2O1N2N2O0O2O1O10000O1000000000000O10000000000O1000001O000O10000000000O10000000000O1000000000000O100000001O0O10000000000O100000001O0O100000001O0O100000001O000O1000001O000O101O0000001N1000001O000O10001O001O0O101O001O000011N101N2O00O1O1O1O100O1O2N1O1O2N1O2N1O2N1O2N1N3N2N0O1O2N1O2O0O2J5J7J6N2N2N101N2N2N1M4lNj@h0d?C\\TW2"}, "label": "a red color car on the road"}]}
{"id": 214563, "image": "COCO_train2014_000000214563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair in front of a pink sony laptop\"."}], "task": "refering", "answer_template": "The \"a chair in front of a pink sony laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 227, 228, 250, 251], "bbox": [0.676484375, 0.2323125, 0.994390625, 0.6299166666666667], "iscrowd": 0, "area": 17143.1827, "rle": {"size": [480, 640], "counts": "eS\\61m>2O1N3M2O1N3M2O1N2N3N1N2N3N1N2N3N1N2N3N1N2N2O2M2N2O2M2N2O2M2N2O2M2N2O1N3M2O1N3M2O1N3M2O1N2N3N1N2N3N1N2N3N1N2N3N1N2N2L5O00001O000010O3M7J0O001O010O00010O0010O01O010O000010O01O010O0010O01O00010O0010O01O010O00010O001O010O0010O00010O01O010O001OTEnLj9S3QFRMn9n2mEWMT:h2hE]MW:d2dEaM\\:^2`EfM`:Z2\\EjMe:W31O2N10001O0100O010O010XEnK`:Z4O010O010O010O010O0010O010O0N3L3L5K4L4J7J5J7I6J7\\OnCRNX<m1?M2O1N3M2O2M2O2M2N3N1N2O2M2O2M2N3N1N3N1N3M2O1N010O00011N2O1N2N2O1N3N1N2Ogf1"}, "label": "a chair in front of a pink sony laptop"}]}
{"id": 214563, "image": "COCO_train2014_000000214563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair behind pink computer in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the chair behind pink computer in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 227, 228, 250, 251], "bbox": [0.676484375, 0.2323125, 0.994390625, 0.6299166666666667], "iscrowd": 0, "area": 17143.1827, "rle": {"size": [480, 640], "counts": "eS\\61m>2O1N3M2O1N3M2O1N2N3N1N2N3N1N2N3N1N2N3N1N2N2O2M2N2O2M2N2O2M2N2O2M2N2O1N3M2O1N3M2O1N3M2O1N2N3N1N2N3N1N2N3N1N2N3N1N2N2L5O00001O000010O3M7J0O001O010O00010O0010O01O010O000010O01O010O0010O01O00010O0010O01O010O00010O001O010O0010O00010O01O010O001OTEnLj9S3QFRMn9n2mEWMT:h2hE]MW:d2dEaM\\:^2`EfM`:Z2\\EjMe:W31O2N10001O0100O010O010XEnK`:Z4O010O010O010O010O0010O010O0N3L3L5K4L4J7J5J7I6J7\\OnCRNX<m1?M2O1N3M2O2M2O2M2N3N1N2O2M2O2M2N3N1N3N1N3M2O1N010O00011N2O1N2N2O1N3N1N2Ogf1"}, "label": "the chair behind pink computer in the right hand picture"}]}
{"id": 157221, "image": "COCO_train2014_000000157221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown elephant stands in the dirt\"."}], "task": "refering", "answer_template": "The \"a brown elephant stands in the dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 217, 218, 235, 236, 237, 238, 240, 241], "bbox": [0.20012500000000003, 0.2755373831775701, 0.53853125, 0.729018691588785], "iscrowd": 0, "area": 29755.6873, "rle": {"size": [428, 640], "counts": "jge11Y=2O2N2]M1eG1Q8:mGHh7c0UH_O`7m0^HUOT7Y1jHhNU7Z1hHiNV7Y1hHhNX7Y1eHhN\\7Y1bHfNa7Z1]HfNd7[1YHeNi7\\1UHdNm7\\1PHdNR8]1lGbNV8_1gGbNZ8_1dG`N_8`1_G`Nb8a1[G_Ng8b1WG]Nl8c1QG^NP9b1oF]NS9d1kF\\Nj06T6]1mH`NNV1U7;fHP1Z7PO`HV1a7Z20000001O00000N2M3L4M3L4M3O11O2UHnJX7T5aHQK_7P5ZHVKf7X50000000000000000000000000000000nN`HXL`7d3fHZLZ7e3iHYLW7f3mHWLS7h3PIULQ7j3RIRLP7l3UIPLl6o3WInKj6Q4YIkKi6T4X1O1O1O1O1O1O1O1O1O1O1O1O[HWLi5h3VJZLj5e3nHhL5Dm6b3bH\\M<SOR7`3\\HgM=kNW7n4eHSK[7P5`HSK_7]5000000000001O0000000000001O00000000dN`HlL`7Q3PIbLP7[3aIXL^6d3mIULS6h3TJTLl5h3\\JTLd5i3Q2L4M3M3001O001O00001O00001O001O00001O001O00001O00001O001O00O1000000O1O1O1002N2oGXLZ6i3eIYLY6i3dIYL[6i3bIZL\\6g3bI\\L\\6f3aI\\L^6f3_I[La6_4dHbK\\7b4_H_Ka7V500000000000N2O2O1O1O010O2N1O101N1O100O2N1O101N1O1O101N1O1O2N1O1O1O2]N`GVNa8f1oGmMR8o1^HdMd7W2kH]MV7_2n1L4L5K4L4L5N13M3M3N1NO1O2J5ROo0POkTk3"}, "label": "a brown elephant stands in the dirt"}]}
{"id": 157221, "image": "COCO_train2014_000000157221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown elephant with very pretty tusks , in a green pasture\"."}], "task": "refering", "answer_template": "The \"a large brown elephant with very pretty tusks , in a green pasture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 217, 218, 235, 236, 237, 238, 240, 241], "bbox": [0.20012500000000003, 0.2755373831775701, 0.53853125, 0.729018691588785], "iscrowd": 0, "area": 29755.6873, "rle": {"size": [428, 640], "counts": "jge11Y=2O2N2]M1eG1Q8:mGHh7c0UH_O`7m0^HUOT7Y1jHhNU7Z1hHiNV7Y1hHhNX7Y1eHhN\\7Y1bHfNa7Z1]HfNd7[1YHeNi7\\1UHdNm7\\1PHdNR8]1lGbNV8_1gGbNZ8_1dG`N_8`1_G`Nb8a1[G_Ng8b1WG]Nl8c1QG^NP9b1oF]NS9d1kF\\Nj06T6]1mH`NNV1U7;fHP1Z7PO`HV1a7Z20000001O00000N2M3L4M3L4M3O11O2UHnJX7T5aHQK_7P5ZHVKf7X50000000000000000000000000000000nN`HXL`7d3fHZLZ7e3iHYLW7f3mHWLS7h3PIULQ7j3RIRLP7l3UIPLl6o3WInKj6Q4YIkKi6T4X1O1O1O1O1O1O1O1O1O1O1O1O[HWLi5h3VJZLj5e3nHhL5Dm6b3bH\\M<SOR7`3\\HgM=kNW7n4eHSK[7P5`HSK_7]5000000000001O0000000000001O00000000dN`HlL`7Q3PIbLP7[3aIXL^6d3mIULS6h3TJTLl5h3\\JTLd5i3Q2L4M3M3001O001O00001O00001O001O00001O001O00001O00001O001O00O1000000O1O1O1002N2oGXLZ6i3eIYLY6i3dIYL[6i3bIZL\\6g3bI\\L\\6f3aI\\L^6f3_I[La6_4dHbK\\7b4_H_Ka7V500000000000N2O2O1O1O010O2N1O101N1O100O2N1O101N1O1O101N1O1O2N1O1O1O2]N`GVNa8f1oGmMR8o1^HdMd7W2kH]MV7_2n1L4L5K4L4L5N13M3M3N1NO1O2J5ROo0POkTk3"}, "label": "a large brown elephant with very pretty tusks , in a green pasture"}]}
{"id": 468518, "image": "COCO_train2014_000000468518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy standing and talking with a lady\"."}], "task": "refering", "answer_template": "The \"young boy standing and talking with a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 328, 329, 351, 352], "bbox": [0.23157812500000002, 0.393304535637149, 0.40973437500000004, 0.9232829373650109], "iscrowd": 0, "area": 16386.8069, "rle": {"size": [463, 640], "counts": "bVS2Y1S=9H4QO[NmDj1n:\\NlDi1P;\\NkDh1Q;]NiDh1S;]NhDg1S;n0M3M3N2N1N3N2M3N2M3N2M4M2M:G:F6I7POjJ[H\\5V7PKfHT5Q7TKmHn4n6XKnHj4o6[KmHh4o6^KlHe4P7_1OO10O01O1O010O001O10O01O10O01O00100O00100O002N2O0O2N2]OaHSJa7b5QIVJP7g5m0N3M2N3M3M2N3jKQGj2o8RMUGn2l8lLYGT3g8gL^GY3c8_LeG_3\\8_LfGa3`8WLbGi3d8PL^Go3a901O001O0O2N101N1M4K4N3M2N3M2N3L3M4M3N3M3N3L3N3M2M4M3M2M4N1N3N2K4J7I7I6J7I7H7JPdZ5"}, "label": "young boy standing and talking with a lady"}]}
{"id": 468518, "image": "COCO_train2014_000000468518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young main in a striped shirt\"."}], "task": "refering", "answer_template": "The \"a young main in a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 328, 329, 351, 352], "bbox": [0.23157812500000002, 0.393304535637149, 0.40973437500000004, 0.9232829373650109], "iscrowd": 0, "area": 16386.8069, "rle": {"size": [463, 640], "counts": "bVS2Y1S=9H4QO[NmDj1n:\\NlDi1P;\\NkDh1Q;]NiDh1S;]NhDg1S;n0M3M3N2N1N3N2M3N2M3N2M4M2M:G:F6I7POjJ[H\\5V7PKfHT5Q7TKmHn4n6XKnHj4o6[KmHh4o6^KlHe4P7_1OO10O01O1O010O001O10O01O10O01O00100O00100O002N2O0O2N2]OaHSJa7b5QIVJP7g5m0N3M2N3M3M2N3jKQGj2o8RMUGn2l8lLYGT3g8gL^GY3c8_LeG_3\\8_LfGa3`8WLbGi3d8PL^Go3a901O001O0O2N101N1M4K4N3M2N3M2N3L3M4M3N3M3N3L3N3M2M4M3M2M4N1N3N2K4J7I7I6J7I7H7JPdZ5"}, "label": "a young main in a striped shirt"}]}
{"id": 468518, "image": "COCO_train2014_000000468518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the pink and white shirt\"."}], "task": "refering", "answer_template": "The \"the girl in the pink and white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 197, 198, 220, 221, 222, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 311, 312, 313, 314], "bbox": [0.5327968750000001, 0.44129589632829375, 0.73321875, 0.8219222462203023], "iscrowd": 0, "area": 11458.813000000004, "rle": {"size": [463, 640], "counts": "Y`j45S>9G8H9J6J6J6J6J8H8H8H4N3M3M2N3M3M2M4M2O2N2O0O2N1O2N2VOaLbF_3l8ZMlFh2R9ZMlFg2S9\\MjFf2T9\\MjFe2U9]MjFd2U9]MiFd2V9[100O100O10000O100O100O1000000001O001O001O001O001O0O2O002O1N2N2N2N2N2N2O0O2N2N1O2N3M4M2\\MQFb0Q:mNjFg0X9XOQG>Q9ARG;P9DSG8o8GTG5n8JUG3l8LWG0k8OXGMj82YGJi85ZGGh88[GDg8;\\GAe8?bGZO_8d0h2O1N2O0O2O0O2O0O100O2O0O101N10O010O10O01O010O1O010O01O0O101N1O2O0O1O2O0OhP]2"}, "label": "the girl in the pink and white shirt"}]}
{"id": 468518, "image": "COCO_train2014_000000468518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a kitchen\"."}], "task": "refering", "answer_template": "The \"a girl in a kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 197, 198, 220, 221, 222, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 311, 312, 313, 314], "bbox": [0.5327968750000001, 0.44129589632829375, 0.73321875, 0.8219222462203023], "iscrowd": 0, "area": 11458.813000000004, "rle": {"size": [463, 640], "counts": "Y`j45S>9G8H9J6J6J6J6J8H8H8H4N3M3M2N3M3M2M4M2O2N2O0O2N1O2N2VOaLbF_3l8ZMlFh2R9ZMlFg2S9\\MjFf2T9\\MjFe2U9]MjFd2U9]MiFd2V9[100O100O10000O100O100O1000000001O001O001O001O001O0O2O002O1N2N2N2N2N2N2O0O2N2N1O2N3M4M2\\MQFb0Q:mNjFg0X9XOQG>Q9ARG;P9DSG8o8GTG5n8JUG3l8LWG0k8OXGMj82YGJi85ZGGh88[GDg8;\\GAe8?bGZO_8d0h2O1N2O0O2O0O2O0O100O2O0O101N10O010O10O01O010O1O010O01O0O101N1O2O0O1O2O0OhP]2"}, "label": "a girl in a kitchen"}]}
{"id": 468518, "image": "COCO_train2014_000000468518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right door of a white refrigerator\"."}], "task": "refering", "answer_template": "The \"the right door of a white refrigerator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 294, 295, 296, 317, 318, 319, 340, 341], "bbox": [0.7266874999999999, 0.3591792656587473, 0.889265625, 0.8468250539956804], "iscrowd": 0, "area": 18098.582700000006, "rle": {"size": [463, 640], "counts": "X]b6Q2^<k2UMM3M3N2M3N2M3M3N2M3N24L5K5K5K5K1O0000000000000O1000000000000000000000000000001O1O001O1O1O001\\HoJh5R5VJPKi5P5VJQKj5P5TJQKl5Q5QJPKo5Q5oIPKP6R5nIoJR6S5kIoJT6R5jIoJV6R5hIoJX6S5eInJ[6S5cInJ]6S5aInJ_6T5^ImJb6T5\\ImJd6T5ZImJf6T5XImJh6T5VImJj6T5TInJk6S5SInJm6S5QInJo6S5nHoJR7R5lHoJT7R5jHoJV7Q5iHoJX7R5fHoJZ7R5dHnJ]7S5aHnJ_7P6100O100O10O01O1O1N3M2O1N2O2M2N2O2M2O1N2N3N1N2ZNbGZM_8c2dG[M_8b2cG\\M_8n1\\2hNX1hNVoo0"}, "label": "the right door of a white refrigerator"}]}
{"id": 468518, "image": "COCO_train2014_000000468518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right fridge\"."}], "task": "refering", "answer_template": "The \"the right fridge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 294, 295, 296, 317, 318, 319, 340, 341], "bbox": [0.7266874999999999, 0.3591792656587473, 0.889265625, 0.8468250539956804], "iscrowd": 0, "area": 18098.582700000006, "rle": {"size": [463, 640], "counts": "X]b6Q2^<k2UMM3M3N2M3N2M3M3N2M3N24L5K5K5K5K1O0000000000000O1000000000000000000000000000001O1O001O1O1O001\\HoJh5R5VJPKi5P5VJQKj5P5TJQKl5Q5QJPKo5Q5oIPKP6R5nIoJR6S5kIoJT6R5jIoJV6R5hIoJX6S5eInJ[6S5cInJ]6S5aInJ_6T5^ImJb6T5\\ImJd6T5ZImJf6T5XImJh6T5VImJj6T5TInJk6S5SInJm6S5QInJo6S5nHoJR7R5lHoJT7R5jHoJV7Q5iHoJX7R5fHoJZ7R5dHnJ]7S5aHnJ_7P6100O100O10O01O1O1N3M2O1N2O2M2N2O2M2O1N2N3N1N2ZNbGZM_8c2dG[M_8b2cG\\M_8n1\\2hNX1hNVoo0"}, "label": "the right fridge"}]}
{"id": 206377, "image": "COCO_train2014_000000206377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller younger baby bear looking off to the side\"."}], "task": "refering", "answer_template": "The \"smaller younger baby bear looking off to the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302], "bbox": [0.3830561797752809, 0.546703125, 0.9426067415730338, 0.8113906249999999], "iscrowd": 0, "area": 29399.4291, "rle": {"size": [640, 445], "counts": "_eZ33Zc0g0H8G9H8H7I7I6J6M3L5K4M2M3N2M3N2N2M3N1O2M2N3N1N3N101N1O101N101O001O0O101O00000O2O00001O0001O01O00001O000000001O01O000001O000000001O1O1O2M2O1O1O1O000000000O0100000000O1000O10000O010O100000000000O1000001YL_@m2a?oLc@Q3]?kLg@U3Z?fLj@Y3X?bLl@^3U?]Lo@c3g?00O10000O10O1000O1000000000000001O000O1000001O000000001N10000000001O00000O10001O00000000000O1000001O000001O000000000010O00000001O01O000000010O00000001O01O01O00001O010O001O01O01O010O00010O0010O01O0010O0100O0100O0100O0100O0100O100O101N1O101N101N100O2N101N100O2N1O2L3M3M5K4L5K4I8E:F;E:F^j?"}, "label": "smaller younger baby bear looking off to the side"}]}
{"id": 206377, "image": "COCO_train2014_000000206377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown bear laying down , and looking to his left\"."}], "task": "refering", "answer_template": "The \"the brown bear laying down , and looking to his left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302], "bbox": [0.3830561797752809, 0.546703125, 0.9426067415730338, 0.8113906249999999], "iscrowd": 0, "area": 29399.4291, "rle": {"size": [640, 445], "counts": "_eZ33Zc0g0H8G9H8H7I7I6J6M3L5K4M2M3N2M3N2N2M3N1O2M2N3N1N3N101N1O101N101O001O0O101O00000O2O00001O0001O01O00001O000000001O01O000001O000000001O1O1O2M2O1O1O1O000000000O0100000000O1000O10000O010O100000000000O1000001YL_@m2a?oLc@Q3]?kLg@U3Z?fLj@Y3X?bLl@^3U?]Lo@c3g?00O10000O10O1000O1000000000000001O000O1000001O000000001N10000000001O00000O10001O00000000000O1000001O000001O000000000010O00000001O01O000000010O00000001O01O01O00001O010O001O01O01O010O00010O0010O01O0010O0100O0100O0100O0100O0100O100O101N1O101N101N100O2N101N100O2N1O2L3M3M5K4L5K4I8E:F;E:F^j?"}, "label": "the brown bear laying down , and looking to his left"}]}
{"id": 206377, "image": "COCO_train2014_000000206377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest bear , showing her protective side\"."}], "task": "refering", "answer_template": "The \"the biggest bear , showing her protective side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 193, 194, 195, 209, 210, 211, 224, 225, 226, 227, 240, 241, 242], "bbox": [0.012921348314606741, 0.3550625, 0.4298426966292135, 0.7056250000000001], "iscrowd": 0, "area": 18244.8354, "rle": {"size": [640, 445], "counts": "kT43ic05L5K4M4M2N3N1N3M001O1O001O1O001O1O10ON3M3M2N3M3M2N3M3M2N3M3M2N3`LVNfDn1Y;`NWDb1j;_NRDd1m;^NoCe1P<]NlCe1T<]NhCf1X<[NQC[2n<gM`Bj2a=WMVBQ3i=PMnAZ3P>\\1O101N1O1O1O2O0O100O2O00003M5J5LO1N2N2M2O2M3N1N2O0O2O0O2O001N2O0O2O0O2O0O2O0O2O1O0O2O0O2O0O2O0O2L3L5L4L4K5L4K5L4L4K5L4L4K5L4L4L4OO010O02O1N3N1N2O1N2O1N2O2M2K5J6K5L4M3L5K4O100O10001N101O0O101O0O2O01O2O1N3N1N2N100O2NO2O001N101O001N101O0O2O001N2O001N101O0O2O001N101O1O2M3M3L4L5L3L4L4M3Ld_n4"}, "label": "the biggest bear , showing her protective side"}]}
{"id": 206377, "image": "COCO_train2014_000000206377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest black bear behind two smaller black bears\"."}], "task": "refering", "answer_template": "The \"the biggest black bear behind two smaller black bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 193, 194, 195, 209, 210, 211, 224, 225, 226, 227, 240, 241, 242], "bbox": [0.012921348314606741, 0.3550625, 0.4298426966292135, 0.7056250000000001], "iscrowd": 0, "area": 18244.8354, "rle": {"size": [640, 445], "counts": "kT43ic05L5K4M4M2N3N1N3M001O1O001O1O001O1O10ON3M3M2N3M3M2N3M3M2N3M3M2N3`LVNfDn1Y;`NWDb1j;_NRDd1m;^NoCe1P<]NlCe1T<]NhCf1X<[NQC[2n<gM`Bj2a=WMVBQ3i=PMnAZ3P>\\1O101N1O1O1O2O0O100O2O00003M5J5LO1N2N2M2O2M3N1N2O0O2O0O2O001N2O0O2O0O2O0O2O0O2O1O0O2O0O2O0O2O0O2L3L5L4L4K5L4K5L4L4K5L4L4K5L4L4L4OO010O02O1N3N1N2O1N2O1N2O2M2K5J6K5L4M3L5K4O100O10001N101O0O101O0O2O01O2O1N3N1N2N100O2NO2O001N101O001N101O0O2O001N2O001N101O0O2O001N101O1O2M3M3L4L5L3L4L4M3Ld_n4"}, "label": "the biggest black bear behind two smaller black bears"}]}
{"id": 50736, "image": "COCO_train2014_000000050736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller zebra looking into the distance\"."}], "task": "refering", "answer_template": "The \"the smaller zebra looking into the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 279, 280, 281, 282, 303, 304], "bbox": [0.146765625, 0.523411214953271, 0.419203125, 0.9219158878504674], "iscrowd": 0, "area": 13230.97745, "rle": {"size": [428, 640], "counts": "PbW13Y=5J7I6J6J6K5J7I6J6K6I6J6J7I6K5J7I6J6K6IM3M3L5K5D<D<E;00O1O11O3M3M3M3_EiMg9Q3M3M2N3M3M2N3N2M2N3NO01OmNPGkMo8m1]GoMd8k1dGRN\\8j1kGRNV8j1RHQNP8j1WHRNj7n1YHoMg7Q2k1O0000001O00000000000000000000000000001O0000O100000<D7I3M4K5L1N3L4M3L3N3L4MO000O2O001O1N2O1O2M2O1N3N1N2O1N3N1N2O2M2O1O2N1O101N1O1O2N1O2O0O2N1O2N1O2N101N101O001O1N101O1O1O001N2001N2O1O1O1O10O01O1O1O100O1O1O1O100O1OVNmD\\1R;cNQEZ1o:fNTEX1l:fNXEX1j:eNYEY1j:cNZEZ1^;N2N2N2N3M4L5JPWk4"}, "label": "the smaller zebra looking into the distance"}]}
{"id": 50736, "image": "COCO_train2014_000000050736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that is on the back of other zebra\"."}], "task": "refering", "answer_template": "The \"a zebra that is on the back of other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 279, 280, 281, 282, 303, 304], "bbox": [0.146765625, 0.523411214953271, 0.419203125, 0.9219158878504674], "iscrowd": 0, "area": 13230.97745, "rle": {"size": [428, 640], "counts": "PbW13Y=5J7I6J6J6K5J7I6J6K6I6J6J7I6K5J7I6J6K6IM3M3L5K5D<D<E;00O1O11O3M3M3M3_EiMg9Q3M3M2N3M3M2N3N2M2N3NO01OmNPGkMo8m1]GoMd8k1dGRN\\8j1kGRNV8j1RHQNP8j1WHRNj7n1YHoMg7Q2k1O0000001O00000000000000000000000000001O0000O100000<D7I3M4K5L1N3L4M3L3N3L4MO000O2O001O1N2O1O2M2O1N3N1N2O1N3N1N2O2M2O1O2N1O101N1O1O2N1O2O0O2N1O2N1O2N101N101O001O1N101O1O1O001N2001N2O1O1O1O10O01O1O1O100O1O1O1O100O1OVNmD\\1R;cNQEZ1o:fNTEX1l:fNXEX1j:eNYEY1j:cNZEZ1^;N2N2N2N3M4L5JPWk4"}, "label": "a zebra that is on the back of other zebra"}]}
{"id": 50736, "image": "COCO_train2014_000000050736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult zebra standing in a meadow\"."}], "task": "refering", "answer_template": "The \"an adult zebra standing in a meadow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 287, 288, 290, 291, 310, 311, 313, 314], "bbox": [0.48121875000000003, 0.4437616822429907, 0.80175, 0.9201168224299066], "iscrowd": 0, "area": 18727.4055, "rle": {"size": [428, 640], "counts": "ooP46Q=<D<H7H6K6J6kD`Nm0L`7j1UGoNh0_OR8P3\\GXMc8i3O1O2O0O1O100O10TOZGPMg8m2bGmL^8S3gGgL[8X3kGbLU8^3m01N5L5K4K8I8H3M0O10000O10000OnNaGYM_8e2iGUMW8i2PHRMP8l2XHnLh7P3aHiL_7U3]1O1N2N2N3L3N2M3N3L3N2N3L3N2001O01O01O00000010O0001O01O010O001O00O100000000O100000000O10000000000O01000O100O10O0100O100O10O01`EhMl9X2mETNm9k2^FcLk8a3oFXMY8l2cGVM\\8m3O2O0N2M3N3M2M3O2N10^O`HXK_7c4jH[KU7^4o0N3M2O1N3M2M4J5N2O2N1N3N1N3N1O1O2N1O2N1O2N1N3N1G:O001O0O2N1O2N1O2N1O2N1O2O001OO10O01000O2O2M2O1N2O2M2O110O010O00100O10O0100O10M2OdN`Dn0^;SOdDm0Y;UOhDk0V;VOkDj0S;VOoDj0P;UOREk0l:SOXEl0g:SO\\Em0c:POaEP1_:mNhEn0X:mNQFP1Y;L4K5L2M3M2N2M4M4Kijd1"}, "label": "an adult zebra standing in a meadow"}]}
{"id": 50736, "image": "COCO_train2014_000000050736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in front of a smaller zebra\"."}], "task": "refering", "answer_template": "The \"a zebra in front of a smaller zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 287, 288, 290, 291, 310, 311, 313, 314], "bbox": [0.48121875000000003, 0.4437616822429907, 0.80175, 0.9201168224299066], "iscrowd": 0, "area": 18727.4055, "rle": {"size": [428, 640], "counts": "ooP46Q=<D<H7H6K6J6kD`Nm0L`7j1UGoNh0_OR8P3\\GXMc8i3O1O2O0O1O100O10TOZGPMg8m2bGmL^8S3gGgL[8X3kGbLU8^3m01N5L5K4K8I8H3M0O10000O10000OnNaGYM_8e2iGUMW8i2PHRMP8l2XHnLh7P3aHiL_7U3]1O1N2N2N3L3N2M3N3L3N2N3L3N2001O01O01O00000010O0001O01O010O001O00O100000000O100000000O10000000000O01000O100O10O0100O100O10O01`EhMl9X2mETNm9k2^FcLk8a3oFXMY8l2cGVM\\8m3O2O0N2M3N3M2M3O2N10^O`HXK_7c4jH[KU7^4o0N3M2O1N3M2M4J5N2O2N1N3N1N3N1O1O2N1O2N1O2N1N3N1G:O001O0O2N1O2N1O2N1O2N1O2O001OO10O01000O2O2M2O1N2O2M2O110O010O00100O10O0100O10M2OdN`Dn0^;SOdDm0Y;UOhDk0V;VOkDj0S;VOoDj0P;UOREk0l:SOXEl0g:SO\\Em0c:POaEP1_:mNhEn0X:mNQFP1Y;L4K5L2M3M2N2M4M4Kijd1"}, "label": "a zebra in front of a smaller zebra"}]}
{"id": 157239, "image": "COCO_train2014_000000157239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman taking backhand swing in the shot to the left\"."}], "task": "refering", "answer_template": "The \"woman taking backhand swing in the shot to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 26, 27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 96, 97, 98, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 168, 169, 170, 187, 188, 192, 193, 194, 210, 211, 216, 217, 233, 234, 240], "bbox": [0.14851562499999998, 0.059750000000000004, 0.46309374999999997, 0.9568749999999999], "iscrowd": 0, "area": 15979.41099999999, "rle": {"size": [320, 640], "counts": "chm02m94L5J4fNEjH>R7ElH=^22?BRM?Y28`NRO49RO?U2>_NUON1]O?P2d0^NQO0NB>k1l0\\NjN42A;k1o0UNRNZOd0U13]O:l1S1TNmM\\Oe0U14\\O9l1W1SNgM@e0T17ZO9k1Z1SNbMCd0U19WO:j1^1RN[MHd0T1<VO8h1l1nM`MT1?RO8h1c2UOYMoN7h1b2YOlNd0V1[OmNa0U1_OlN=W1BlN9V1HlN3W1MjNOY11iNKY15iNGfN^O8k0S1DeNJ1b0\\1AbN6WNZOV1o0S2^O`Nl0XOGZ2UObNY1oNDa2kNdNf1gNAm4e0oJ[OP5k0jJWOS5o0iJROU5Q1iJoNV5T1gJnNW5U1fJlNX5X1eJjNY5Y1dJiNZ5Z1bJhN\\5P3N2N5K=B3N2N2N2M3N2N2N2M2O2N101O001O1O1O1O1O1O2N000[LjI`3V6_LlI`3T6_LmIa3S6_LnI`3Z6O001N2O1O001O3M4M3L4M4L4L4L4L4L2N000000001O0PN\\Hb1R801O000[NmG\\1\\800eN_GV1`8iNcGV1]8iNdGW1c801O001N101O001O001O001O1N101O001O001N10001O000O101O00001O0O10001O00O0100000O1000O10O100O1000O010000O100001O0O101O0000001O000O2O00001O000010O1O100O1O2O0O2N2O1M3M3N2M2N010O0001O00001O1N2O1N6K4K3M3K4L5K\\W[3"}, "label": "woman taking backhand swing in the shot to the left"}]}
{"id": 157239, "image": "COCO_train2014_000000157239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on the left playing tennis\"."}], "task": "refering", "answer_template": "The \"a woman on the left playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 26, 27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 96, 97, 98, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 168, 169, 170, 187, 188, 192, 193, 194, 210, 211, 216, 217, 233, 234, 240], "bbox": [0.14851562499999998, 0.059750000000000004, 0.46309374999999997, 0.9568749999999999], "iscrowd": 0, "area": 15979.41099999999, "rle": {"size": [320, 640], "counts": "chm02m94L5J4fNEjH>R7ElH=^22?BRM?Y28`NRO49RO?U2>_NUON1]O?P2d0^NQO0NB>k1l0\\NjN42A;k1o0UNRNZOd0U13]O:l1S1TNmM\\Oe0U14\\O9l1W1SNgM@e0T17ZO9k1Z1SNbMCd0U19WO:j1^1RN[MHd0T1<VO8h1l1nM`MT1?RO8h1c2UOYMoN7h1b2YOlNd0V1[OmNa0U1_OlN=W1BlN9V1HlN3W1MjNOY11iNKY15iNGfN^O8k0S1DeNJ1b0\\1AbN6WNZOV1o0S2^O`Nl0XOGZ2UObNY1oNDa2kNdNf1gNAm4e0oJ[OP5k0jJWOS5o0iJROU5Q1iJoNV5T1gJnNW5U1fJlNX5X1eJjNY5Y1dJiNZ5Z1bJhN\\5P3N2N5K=B3N2N2N2M3N2N2N2M2O2N101O001O1O1O1O1O1O2N000[LjI`3V6_LlI`3T6_LmIa3S6_LnI`3Z6O001N2O1O001O3M4M3L4M4L4L4L4L4L2N000000001O0PN\\Hb1R801O000[NmG\\1\\800eN_GV1`8iNcGV1]8iNdGW1c801O001N101O001O001O001O1N101O001O001N10001O000O101O00001O0O10001O00O0100000O1000O10O100O1000O010000O100001O0O101O0000001O000O2O00001O000010O1O100O1O2O0O2N2O1M3M3N2M2N010O0001O00001O1N2O1N6K4K3M3K4L5K\\W[3"}, "label": "a woman on the left playing tennis"}]}
{"id": 157242, "image": "COCO_train2014_000000157242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hotdog on the left\"."}], "task": "refering", "answer_template": "The \"the hotdog on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 114, 115, 116, 117, 132, 133, 134, 135], "bbox": [0.30995999999999996, 0.2765066666666667, 0.55782, 0.5994933333333333], "iscrowd": 0, "area": 10262.598250000003, "rle": {"size": [375, 500], "counts": "Wlh12d;4M3L5K4M3L5L3L4L5L3L4M4L3M3M3M4L3M3M4L3N2N3M2M3N3M2N2M4MO10000OM4K4K6J4N4O0000001O00001O0000010O0000010OO1O1O2O0O1O1O1O100N2O1O1000O1N2O101O0000010O000001O01O000001O000O101O00001O3M2M4M2N3M3M2N3M3L3N3M3M2N3M3M2M4M3M2N3L4L3M4L4L3M4L3M4L4L3L5L4L3Mhg`2"}, "label": "the hotdog on the left"}]}
{"id": 157242, "image": "COCO_train2014_000000157242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hot dog made of icing on the left side\"."}], "task": "refering", "answer_template": "The \"the hot dog made of icing on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 114, 115, 116, 117, 132, 133, 134, 135], "bbox": [0.30995999999999996, 0.2765066666666667, 0.55782, 0.5994933333333333], "iscrowd": 0, "area": 10262.598250000003, "rle": {"size": [375, 500], "counts": "Wlh12d;4M3L5K4M3L5L3L4L5L3L4M4L3M3M3M4L3M3M4L3N2N3M2M3N3M2N2M4MO10000OM4K4K6J4N4O0000001O00001O0000010O0000010OO1O1O2O0O1O1O1O100N2O1O1000O1N2O101O0000010O000001O01O000001O000O101O00001O3M2M4M2N3M3M2N3M3L3N3M3M2N3M3M2M4M3M2N3L4L3M4L4L3M4L3M4L4L3L5L4L3Mhg`2"}, "label": "the hot dog made of icing on the left side"}]}
{"id": 247368, "image": "COCO_train2014_000000247368.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one of two zebras , standing behind a goose\"."}], "task": "refering", "answer_template": "The \"one of two zebras , standing behind a goose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 132, 133, 135, 136, 149, 150, 151, 155, 156, 158, 159, 172, 173, 174, 178, 179, 181, 182, 195, 196, 201, 202, 204, 205], "bbox": [0.467203125, 0.019316770186335402, 0.98071875, 0.7283229813664597], "iscrowd": 0, "area": 40389.080699999955, "rle": {"size": [322, 640], "counts": "lSn26j9:E9G4M4L3M4L3N1N2N2N2L4M3L5K3M4L3L5LhNX11`I\\OlNE\\3o0jM3iM\\OZ4b0jMn0T2SOjMQ1T2[O_Mg0_2^O[Mf0c2^OWMe0lNcM]3l1cMd0mNbM_3n1^MP1`2TO[Mo0d2UOWMm0h2VOSMm0m2VOnLl0Q3XOjLj0U3YOfLkNM8]3e2QMoLn2P3`MdL_2[3oMYLQ2d3XNWLf1h3`NTL_1k3fNRLZ1k3jNTLU1l3mNSLR1m3PORLP1m3RORLm0n3UOQLk0n3WOPLj0S4TOlKl0Z4oNeKP1^4oNaKQ1a4W22N2M3N2N3M3L4M3M4L3L4M3M3M4K3N3M2N2M4M2N[N`LgLa3U3QM]Lo2_3XM]Li2_3]M^Lc2`3cM]L]2`3iM\\LX2b3mM[LS2b3SNZLn1c3WNZLi1d3\\NYLe1d3`NXLb1e3_2N2M3M2N3N2M3M3N2O11O1N101O001O1O010O0010O010O01O10O010O010O01O010O10O01O000000010O00000000001O000001O01O00000000001O0001O000000000O100000001O000O1000000000000O0100000000000O100000O100000O100000O100000O010000O10001Nb0_Ob0[KiK[2i4jLbKS3h5O2N2aLdIS3g6O1O0O2O0O101N1000000000000000000000000000dNcMSK^2^4SN_Km1Z4]NcKc1T4hNjKX1Q4oNmKR1n3TOPLl0l3ZOQLg0l3^ORLb0k3CSL>i3EWL;g3HXL8e3K[L5b3O]L2`30`L0^33aLM]35cLK[3`3M2N2O2O00001O000O1010O0001O02N2N101N101O010O0010O010O01TJPMc4U3SKQMm4S3iJRMW5m3010O03N00000O0100O100O010O1007H4M000000O0O2O1N3L3N3M3M2N3L4M3M2M4M3L4L3N2M4kN^IbNf6W1nIWNW6b1TJVNP6e1Y1jNlG8W8GnG3U8MnGMU83nG_O]8a0b00O1000O1000O0100N2O1O001N2O1O1N101OVe3"}, "label": "one of two zebras , standing behind a goose"}]}
{"id": 247368, "image": "COCO_train2014_000000247368.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra eating grass with a goose in front of it\"."}], "task": "refering", "answer_template": "The \"zebra eating grass with a goose in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 132, 133, 135, 136, 149, 150, 151, 155, 156, 158, 159, 172, 173, 174, 178, 179, 181, 182, 195, 196, 201, 202, 204, 205], "bbox": [0.467203125, 0.019316770186335402, 0.98071875, 0.7283229813664597], "iscrowd": 0, "area": 40389.080699999955, "rle": {"size": [322, 640], "counts": "lSn26j9:E9G4M4L3M4L3N1N2N2N2L4M3L5K3M4L3L5LhNX11`I\\OlNE\\3o0jM3iM\\OZ4b0jMn0T2SOjMQ1T2[O_Mg0_2^O[Mf0c2^OWMe0lNcM]3l1cMd0mNbM_3n1^MP1`2TO[Mo0d2UOWMm0h2VOSMm0m2VOnLl0Q3XOjLj0U3YOfLkNM8]3e2QMoLn2P3`MdL_2[3oMYLQ2d3XNWLf1h3`NTL_1k3fNRLZ1k3jNTLU1l3mNSLR1m3PORLP1m3RORLm0n3UOQLk0n3WOPLj0S4TOlKl0Z4oNeKP1^4oNaKQ1a4W22N2M3N2N3M3L4M3M4L3L4M3M3M4K3N3M2N2M4M2N[N`LgLa3U3QM]Lo2_3XM]Li2_3]M^Lc2`3cM]L]2`3iM\\LX2b3mM[LS2b3SNZLn1c3WNZLi1d3\\NYLe1d3`NXLb1e3_2N2M3M2N3N2M3M3N2O11O1N101O001O1O010O0010O010O01O10O010O010O01O010O10O01O000000010O00000000001O000001O01O00000000001O0001O000000000O100000001O000O1000000000000O0100000000000O100000O100000O100000O100000O010000O10001Nb0_Ob0[KiK[2i4jLbKS3h5O2N2aLdIS3g6O1O0O2O0O101N1000000000000000000000000000dNcMSK^2^4SN_Km1Z4]NcKc1T4hNjKX1Q4oNmKR1n3TOPLl0l3ZOQLg0l3^ORLb0k3CSL>i3EWL;g3HXL8e3K[L5b3O]L2`30`L0^33aLM]35cLK[3`3M2N2O2O00001O000O1010O0001O02N2N101N101O010O0010O010O01TJPMc4U3SKQMm4S3iJRMW5m3010O03N00000O0100O100O010O1007H4M000000O0O2O1N3L3N3M3M2N3L4M3M2M4M3L4L3N2M4kN^IbNf6W1nIWNW6b1TJVNP6e1Y1jNlG8W8GnG3U8MnGMU83nG_O]8a0b00O1000O1000O0100N2O1O001N2O1O1N101OVe3"}, "label": "zebra eating grass with a goose in front of it"}]}
{"id": 575049, "image": "COCO_train2014_000000575049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling young woman with long hair and glasses\"."}], "task": "refering", "answer_template": "The \"a smiling young woman with long hair and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.0, 0.04089583333333333, 0.584578125, 0.9894999999999999], "iscrowd": 0, "area": 95880.15699999999, "rle": {"size": [480, 640], "counts": "o5\\4c:]4cK1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M4M2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2O1O1N2O1N2O1N2O1N200O10000O10000O10000O10000O10000O1000000000000O1000000000000000000000000000000O10001O0000000000000000000000000O10000000000000000000000000000001O001O1O001O1O1O001O1O001O1O001O1SGjNn3V1RLmNk3T1SLoNk3R1TLQOcMHg3W1eNUO\\MKm3Q1fNWOUMOS4k0fNYOQM2W4f0gN[OjL7\\4>jN^ObL;b48jNA\\L>h42kNCULb0n4KlNFoKf0S5EmNHhKj0Y5_OmNKcKl0^5YOoNU2o0lMoNW2o0jMPOX2n0hMnN_2o0bMlNd2R1]MhNl2T1TMhNV3R1kLhN`3R1aLiNi3Q1XLjNR4P1nKkN]4o0dKnNd4l0]KSOi4g0WKXOP5b0QK\\OV5>UKWOQ5c0jKaN]4Y1VLSNQ4g1SLUNQ4g1RLUNP4j1SLSNn3l1ULPNm3o1VLmMl3R2WLkMj3T2YLhMi3W2[LdMf3\\2_41O00000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O01O000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O00000@TClNn<R1YCgNi<V1c0N1O2M3N2M3N2N2M3N2M3N2N2KXXl3"}, "label": "a smiling young woman with long hair and glasses"}]}
{"id": 575049, "image": "COCO_train2014_000000575049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing glasses and a blue sweater\"."}], "task": "refering", "answer_template": "The \"a woman wearing glasses and a blue sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.0, 0.04089583333333333, 0.584578125, 0.9894999999999999], "iscrowd": 0, "area": 95880.15699999999, "rle": {"size": [480, 640], "counts": "o5\\4c:]4cK1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M4M2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2O1O1N2O1N2O1N2O1N200O10000O10000O10000O10000O10000O1000000000000O1000000000000000000000000000000O10001O0000000000000000000000000O10000000000000000000000000000001O001O1O001O1O1O001O1O001O1O001O1SGjNn3V1RLmNk3T1SLoNk3R1TLQOcMHg3W1eNUO\\MKm3Q1fNWOUMOS4k0fNYOQM2W4f0gN[OjL7\\4>jN^ObL;b48jNA\\L>h42kNCULb0n4KlNFoKf0S5EmNHhKj0Y5_OmNKcKl0^5YOoNU2o0lMoNW2o0jMPOX2n0hMnN_2o0bMlNd2R1]MhNl2T1TMhNV3R1kLhN`3R1aLiNi3Q1XLjNR4P1nKkN]4o0dKnNd4l0]KSOi4g0WKXOP5b0QK\\OV5>UKWOQ5c0jKaN]4Y1VLSNQ4g1SLUNQ4g1RLUNP4j1SLSNn3l1ULPNm3o1VLmMl3R2WLkMj3T2YLhMi3W2[LdMf3\\2_41O00000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O01O000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O00000@TClNn<R1YCgNi<V1c0N1O2M3N2M3N2N2M3N2M3N2N2KXXl3"}, "label": "a woman wearing glasses and a blue sweater"}]}
{"id": 575049, "image": "COCO_train2014_000000575049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with black t - shirt holding a remote and smilling\"."}], "task": "refering", "answer_template": "The \"a boy with black t - shirt holding a remote and smilling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.304375, 0.11345833333333334, 0.91815625, 0.990125], "iscrowd": 0, "area": 77111.91475000004, "rle": {"size": [480, 640], "counts": "`ek29^><D=B=D=C=C<D=D;L5L3L5K4M4K5L3L5L3L5L3L5K4M4K4M4K5L3L5K4M4K4M2M101N100O1O100O101N100O100O1O2O0O100O100O1O2O0O100O100O2O0O1O100O100O2O0O100O10001O0000000000001O000O1000001L3K5K5K5K5J6K6J5K5K5J6K5M4M2N2M3N2N2M3N3M2N2M3O1O1O2N1O1O1O1N2O1O2N1O1O100O101N100O10000O2O0O100O10001N10000000001O000O1000001O0000000O2O001O1O1O1O1N2O010O1O1O100eMhDl0Y;ROlDj0U;TOmDk0T;SOmDm0T;QOmDP1S;nNnDR1S;lNnDT1S;jNnDV1T;gNnDX1U;aNoD`1S;ZNPEf1S;TNPEl1R<00001O0UHPNf2P2YMTNd2l1[MXNb2i1]M[N_2e1`M_N]2a1bMcN[2]1eMfNY2Y1fMlNV2U1hMoNU2Q1kMROR2n0mMVOP2j0oMZOn1f0RN^Oj1c0SNBj1>TNGi19VNKg16VNOg12VN4f1LYN8d1IYN=d1B[Nb0b1_O[Ng0a1YO]Nl0`1UO^NP1^1QO_NU1]1kN`N[1]1fN_Na1]1_N`Ng1]1ZN`Nl1\\1UN`NQ2]1oMbNV2Z1kMeNX2X1hMiN[2S1fMlN]2Q1cMoNa2m0`MROc2k0^MTOf2h0ZMXOi2e0XMZOl2b0TM^Oo2?RM_NdMfN\\5h2QM^NjMfNX5i2nL^NQNeNQ5m2oLZNXNdNi4R3oLWN_NcNb4V3PMSNeNcNZ4[3QMoMlNbNS4_3RMkMe4U2\\KgMg4Y2YKeMi4[2XKaMj4`2VK]Mm4c2TKZMn4f2SKVMP5j2PKSMS5m2nJPMS5Q3mJlLV5T3kJhLX5X3hJfLZ5Z3gJbL\\5^3eJ^L]5c3P3000000000O2O0000dMgDQ1Y;nNjDP1U;QOlDn0T;ROnDl0R;SOPEl0P;TOREj0n:VOSEi0l:XOVEf0j:YOXEf0h:ZOZEd0f:\\O[Ec0e:\\O\\Ed0c:]O]Ec0c:]O\\Ed0d:\\O\\Ed0d:[O]Ee0c:[O]Ee0b:\\O^Ed0b:\\O^Ed0b:[O^Ef0b:ZO^Ef0b:ZO^Ef0a:ZO`Ef0`:ZO`Ef0`:ZO`Ef0`:ZO_Eg0a:XO`Eh0_:YOaEg0_:YOaEg0_:XObEh0^:XObEh0^:XOaEi0^:XObEh0^:WOcEi0]:WObEj0^:VObEj0^:VObEj0]:VOcEk0]:UOcEk0]:UOcEk0]:VObEj0^:VOaEk0_:VO`Ej0_:WOaEi0_:XO_Ei0a:WO_Ei0a:XO^Eh0b:XO^Eh0a:ZO]Eg0c:YO^Ef0b:[O]Ee0c:[O]Ee0c:\\O\\Ed0c:]O]Ec0c:^O\\Eb0d:^O]Ea0c:@\\E`0d:@\\E`0c:A]E?c:B\\E>d:B\\E>d:B]E=c:C]E=b:E]E;c:E]E;c:E]E;c:E]E;c:F]E9b:H_E7a:I`E6`:JaE5_:LbE2^:NcE1\\:0eEO[:1fENZ:3fELZ:4hEJX:6iEIW:7jEHV:9jEFV::lEDT:<mECS:=QF[OT:d0PFVOR:j0RFPOP:o0UFWNa:i1[100000000O2O000000000O100000001O0O10000000000O1jNfBf0[=YOfBf0Z=ZOgBe0Y=[OgBe0Y=[OhBd0X=\\OiBc0W=]OjBa0X=^OhBb0X=^OiBa0X=^OiBa0X=]OjBa0X=^OhBb0X=^OiB`0Y=^OiB`0Y=_OhB?Z=_OhB?Y=AgB>o=N100O2N1O1O1O1O100O2N1O1O1O1OYPh0"}, "label": "a boy with black t - shirt holding a remote and smilling"}]}
{"id": 575049, "image": "COCO_train2014_000000575049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fully visible man in a black t - shirt and holding a wii controller\"."}], "task": "refering", "answer_template": "The \"fully visible man in a black t - shirt and holding a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.304375, 0.11345833333333334, 0.91815625, 0.990125], "iscrowd": 0, "area": 77111.91475000004, "rle": {"size": [480, 640], "counts": "`ek29^><D=B=D=C=C<D=D;L5L3L5K4M4K5L3L5L3L5L3L5K4M4K4M4K5L3L5K4M4K4M2M101N100O1O100O101N100O100O1O2O0O100O100O1O2O0O100O100O2O0O1O100O100O2O0O100O10001O0000000000001O000O1000001L3K5K5K5K5J6K6J5K5K5J6K5M4M2N2M3N2N2M3N3M2N2M3O1O1O2N1O1O1O1N2O1O2N1O1O100O101N100O10000O2O0O100O10001N10000000001O000O1000001O0000000O2O001O1O1O1O1N2O010O1O1O100eMhDl0Y;ROlDj0U;TOmDk0T;SOmDm0T;QOmDP1S;nNnDR1S;lNnDT1S;jNnDV1T;gNnDX1U;aNoD`1S;ZNPEf1S;TNPEl1R<00001O0UHPNf2P2YMTNd2l1[MXNb2i1]M[N_2e1`M_N]2a1bMcN[2]1eMfNY2Y1fMlNV2U1hMoNU2Q1kMROR2n0mMVOP2j0oMZOn1f0RN^Oj1c0SNBj1>TNGi19VNKg16VNOg12VN4f1LYN8d1IYN=d1B[Nb0b1_O[Ng0a1YO]Nl0`1UO^NP1^1QO_NU1]1kN`N[1]1fN_Na1]1_N`Ng1]1ZN`Nl1\\1UN`NQ2]1oMbNV2Z1kMeNX2X1hMiN[2S1fMlN]2Q1cMoNa2m0`MROc2k0^MTOf2h0ZMXOi2e0XMZOl2b0TM^Oo2?RM_NdMfN\\5h2QM^NjMfNX5i2nL^NQNeNQ5m2oLZNXNdNi4R3oLWN_NcNb4V3PMSNeNcNZ4[3QMoMlNbNS4_3RMkMe4U2\\KgMg4Y2YKeMi4[2XKaMj4`2VK]Mm4c2TKZMn4f2SKVMP5j2PKSMS5m2nJPMS5Q3mJlLV5T3kJhLX5X3hJfLZ5Z3gJbL\\5^3eJ^L]5c3P3000000000O2O0000dMgDQ1Y;nNjDP1U;QOlDn0T;ROnDl0R;SOPEl0P;TOREj0n:VOSEi0l:XOVEf0j:YOXEf0h:ZOZEd0f:\\O[Ec0e:\\O\\Ed0c:]O]Ec0c:]O\\Ed0d:\\O\\Ed0d:[O]Ee0c:[O]Ee0b:\\O^Ed0b:\\O^Ed0b:[O^Ef0b:ZO^Ef0b:ZO^Ef0a:ZO`Ef0`:ZO`Ef0`:ZO`Ef0`:ZO_Eg0a:XO`Eh0_:YOaEg0_:YOaEg0_:XObEh0^:XObEh0^:XOaEi0^:XObEh0^:WOcEi0]:WObEj0^:VObEj0^:VObEj0]:VOcEk0]:UOcEk0]:UOcEk0]:VObEj0^:VOaEk0_:VO`Ej0_:WOaEi0_:XO_Ei0a:WO_Ei0a:XO^Eh0b:XO^Eh0a:ZO]Eg0c:YO^Ef0b:[O]Ee0c:[O]Ee0c:\\O\\Ed0c:]O]Ec0c:^O\\Eb0d:^O]Ea0c:@\\E`0d:@\\E`0c:A]E?c:B\\E>d:B\\E>d:B]E=c:C]E=b:E]E;c:E]E;c:E]E;c:E]E;c:F]E9b:H_E7a:I`E6`:JaE5_:LbE2^:NcE1\\:0eEO[:1fENZ:3fELZ:4hEJX:6iEIW:7jEHV:9jEFV::lEDT:<mECS:=QF[OT:d0PFVOR:j0RFPOP:o0UFWNa:i1[100000000O2O000000000O100000001O0O10000000000O1jNfBf0[=YOfBf0Z=ZOgBe0Y=[OgBe0Y=[OhBd0X=\\OiBc0W=]OjBa0X=^OhBb0X=^OiBa0X=^OiBa0X=]OjBa0X=^OhBb0X=^OiB`0Y=^OiB`0Y=_OhB?Z=_OhB?Y=AgB>o=N100O2N1O1O1O1O100O2N1O1O1O1OYPh0"}, "label": "fully visible man in a black t - shirt and holding a wii controller"}]}
{"id": 575049, "image": "COCO_train2014_000000575049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"th guy oni the far right\"."}], "task": "refering", "answer_template": "The \"th guy oni the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 205, 206, 228, 229, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.857625, 0.426, 1.0, 0.9820625], "iscrowd": 0, "area": 14423.010349999999, "rle": {"size": [480, 640], "counts": "bgQ83m>4K5L4L5J5L4K5L4K5L4L4K5oMXNbFi1]9XNaFi1^9XNbFh1]9ZNaFg1_9YNaFg1^9[NaFe1^9]N`Fd1_9]NaFc1_9^N_Fc1`9^N`Fb1_9`N_Fa1a9_N_Fa1`9aN_F^1a9cN^F^1b9cN]F]1b9eN\\F\\1c9eN]F[1c9fN[F^1a9cN_Fb1[9`NdFf1V9ZNiFj1S9XNlFh1S9ZNkFg1T9ZNlFf1T9[NjFf1U9[NkFe1T9]NkFc1U9]NjFd1a6eMgJi0gNb1`6VNYJ9VOb1`6eNkIJDa1`6nNcICK`1a6oNbIBL`1a6POaIBL_1b6POaICJ_1c6QObIBG_1f6QOaICF]1h6QOaIEC\\1k6QO`IFBZ1m6RO_IG@Y1o6RO`In2`6SM_Im2`6UM^Il2b6UMXIP3h6PMRIV3m6_200O100O100O100O100O100O100O10000O100O1000000O1000O01000000a0fJlIe0b;bN^1aN`B"}, "label": "th guy oni the far right"}]}
{"id": 575049, "image": "COCO_train2014_000000575049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black shirt on wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man with a black shirt on wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 205, 206, 228, 229, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.857625, 0.426, 1.0, 0.9820625], "iscrowd": 0, "area": 14423.010349999999, "rle": {"size": [480, 640], "counts": "bgQ83m>4K5L4L5J5L4K5L4K5L4L4K5oMXNbFi1]9XNaFi1^9XNbFh1]9ZNaFg1_9YNaFg1^9[NaFe1^9]N`Fd1_9]NaFc1_9^N_Fc1`9^N`Fb1_9`N_Fa1a9_N_Fa1`9aN_F^1a9cN^F^1b9cN]F]1b9eN\\F\\1c9eN]F[1c9fN[F^1a9cN_Fb1[9`NdFf1V9ZNiFj1S9XNlFh1S9ZNkFg1T9ZNlFf1T9[NjFf1U9[NkFe1T9]NkFc1U9]NjFd1a6eMgJi0gNb1`6VNYJ9VOb1`6eNkIJDa1`6nNcICK`1a6oNbIBL`1a6POaIBL_1b6POaICJ_1c6QObIBG_1f6QOaICF]1h6QOaIEC\\1k6QO`IFBZ1m6RO_IG@Y1o6RO`In2`6SM_Im2`6UM^Il2b6UMXIP3h6PMRIV3m6_200O100O100O100O100O100O100O10000O100O1000000O1000O01000000a0fJlIe0b;bN^1aN`B"}, "label": "a man with a black shirt on wearing glasses"}]}
{"id": 190026, "image": "COCO_train2014_000000190026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bike at the back of the men\"."}], "task": "refering", "answer_template": "The \"a black bike at the back of the men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227], "bbox": [0.839328125, 0.17977083333333335, 0.9960625000000001, 0.5483125], "iscrowd": 0, "area": 14353.298150000006, "rle": {"size": [480, 640], "counts": "cjk72U10b<5XC4`<0\\C8[<N_C:Z<KaC:Z<K`C;[<EdC`0W<@jCe0Q<ZOoCl0l;TOTDP1h;oNXDW1b;iN^D\\1^;dNbDa1Y;^NgDh1S;XNmDm1o:RNRES2i:mMVEY2e:fM\\E^2_:bMaEd2Z:\\MeEj2U:VMlEn2o9RMQFP3l9RMTFn2k9UMRFl2m9ZMmEh2Q:T1N2O1O2N1O1N2O1O2N1N2O1O2N1000O2N1O1O1N2O2N1O1N2O1O2N1N2O1O1O2N100001O01O01O0000010O00001O01O0001O00010O00001O0001O01O000010O000001O01O01O0000010O000mY1"}, "label": "a black bike at the back of the men"}]}
{"id": 190026, "image": "COCO_train2014_000000190026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back half of a motorcycle behind the man\"."}], "task": "refering", "answer_template": "The \"the back half of a motorcycle behind the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227], "bbox": [0.839328125, 0.17977083333333335, 0.9960625000000001, 0.5483125], "iscrowd": 0, "area": 14353.298150000006, "rle": {"size": [480, 640], "counts": "cjk72U10b<5XC4`<0\\C8[<N_C:Z<KaC:Z<K`C;[<EdC`0W<@jCe0Q<ZOoCl0l;TOTDP1h;oNXDW1b;iN^D\\1^;dNbDa1Y;^NgDh1S;XNmDm1o:RNRES2i:mMVEY2e:fM\\E^2_:bMaEd2Z:\\MeEj2U:VMlEn2o9RMQFP3l9RMTFn2k9UMRFl2m9ZMmEh2Q:T1N2O1O2N1O1N2O1O2N1N2O1O2N1000O2N1O1O1N2O2N1O1N2O1O2N1N2O1O1O2N100001O01O01O0000010O00001O01O0001O00010O00001O0001O01O000010O000001O01O01O0000010O000mY1"}, "label": "the back half of a motorcycle behind the man"}]}
{"id": 190026, "image": "COCO_train2014_000000190026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"customized motorcycle with fat rear tire\"."}], "task": "refering", "answer_template": "The \"customized motorcycle with fat rear tire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 31, 32, 33, 52, 53, 54, 73, 74, 75, 76, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338], "bbox": [0.12142187499999998, 0.0476875, 0.7941093749999999, 0.8723124999999999], "iscrowd": 0, "area": 80267.48924999998, "rle": {"size": [480, 640], "counts": "jgT1f0V>7I5K6N1N2O1O1O1N2O2]NmNUET1i:nNVES1h:oNVES1h:oNWER1g:POXEQ1f:QOYEP1e:ROYEQ1e:POZEQ1d:ROZEo0e:ROZEo0d:RO[EP1d:QO[EP1c:QO]Eo0c:QO]EP1a:RO^Eo0a:QO_EP1_:ROaEm0_:SOaEn0]:TObEm0]:SOcEn0[:SOeEn0Z:ROfEn0Y:ROhEo0W:QOiEP1U:POlEQ1S:oNmEQ1T:nNlES1S:lNnEU1Q:kNoEV1P:iNQFX1o9gNQFY1o9gNQFZ1n9eNTF[1k9eNUF[1l9gNQFZ1n9jNnEV1R:nNjES1V:QOeEo0\\:TO`Em0_:k10O010O01O0M4M2M4M2N3N1O1O1O2N1O100O101O0001O01O00000O100VHjJi5W5P2OoFQK9Ii6V5nHQK8Ji6V5nHQK9Ii6V5mHQK:Ji6U5lHRK:Ij6V5kHRK;Hj6V5jHSK;Hj6U5kHSK<Hi6U5jHTK<Hi6U5jHTK<Hj6S5kHUK;Gj6W3]H_N=dN;Gj6V3aH\\N:gN;Hj6T3dH[N6jN<Gi6S3iHZN1mN<Gj6Q3lHXNNQO;Fk6Q3oHVNJSO=Fi6Q3SISNGWO<Fj6o2VIRNCZO=Ei6n2[IQN_O\\O<Fi6m2_InM\\O@;Ek6l2aImMXOB=Ei6l2eIjMUOF<Ej6i2iIjMPOI<Ej6i2lIgMmNL=Di6k2oI_MnN1:Ej6k2PJYMoN87Di6m2RJRMPO>4Di6m2UJkLROd0OEj6m2mJ_MYNDi6n2nJ]MYNFi6n2lJ_MXNCl6o2kJ`MXNAl6Q3kJ_MWNAn6P3jJbMVN^Oo6R3iJbMVN]OP7R3hJdMVNZOS7S3lHiLU1l0kNXOS7j3nIRMlNUOV7k3kIRMmNTOW7m3iIQMoNROW7P4gIbMZ6`2bIbM]6a2`I`M_6[NkHc3c0SNb6YNmHe3=TNe6XNPId38UNh6VNRIf33UNj6UNVIg3KVNn6TNXIh3FVNQ7QNkIl5S6TJoIk5P6VJPJk5o5TJSJl5k5TJWJk5i5UJWJl5g5TJ[Jl5c5TJ_Jl5`5TJ`Jl5_5TJcJl5[5TJgJl5X5TJhJl5W5TJkJl5T5SJnJm5P5TJPKl5o4TJSKl5l4SJVKm5h4TJXKm5f4SJ\\Kl5d4SJ^Km5`4TJ`Km5_4RJcKm5\\4SJfKm5X4SJiKn5V4RJkKm5T4SJnKm5P4SJQLm5o3SJRLm5l3SJVLl5j3SJXLm5f3TJ[Lk5d3UJ]Ll5b3SJ`Lm5^3TJcLk5]3TJeLl5Y3TJiLj5W3WJjLh5V3WJlLg5T3YJnLe5R3\\JoLc5Q3\\JPMc5P3]JRMa5o2_JRM`5m2`JUM^5k2bJWM\\5j2dJWM[5h2eJZMY5f2gJ\\MW5e2iJ\\MV5c2jJ^MU5c2jJ_MU5`2lJaMR5_2nJcMP5^2oJdMP5[2QKfMo4X2QKjMo4U2PKlMQ5R2PKoMQ5o1nJSNR5k1nJWNR5g1oJZNR5d1mJ]NT5a1lJaNT5]1mJdNT5Z1kJhNU5V1kJlNU5dL`Jj3;BW5^LbJo37EW5WLgJR42IW5PLkJW4NJi55VJMi53VJNk51UJ0j50UJ2k5MTJ5k5KUJ7j5HUJ:j5FUJ<i5EWJ<h5DWJ>h5BWJ`0g5BXJ?g5BWJ`0g5BWJ`0h5AWJ`0h5^4M2O1O1N2O1O1N2O000O10O010000O11O2N1O1O1O1O2N1N2O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1N2O2N1O1O1O1O2M2O1O1O1O2N1N2O1O1N3N1O000O101O0O1000000O101O0O1000000O2O000O1000000O2O000O100000O010000O10000O1000000O10000000O100000000000000000000000000O10000000000000000000000000000O101O00001O00001O0000001O000O2O00001O0000001N1O1O2M2O1O2N2N1O2N2N2N2N1N3N2N2N1O2N2N2N2N1O2N2M3N1O2N2N2N2N1O2N2N2N1N3N2N2N2N1O2L4K5L3L6K5J6K4fMTE:R;AaELd:OoE]OV:`0\\FmNi9n0X2L3L5L4K4M4K5L3Ledm1"}, "label": "customized motorcycle with fat rear tire"}]}
{"id": 190026, "image": "COCO_train2014_000000190026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and black motorcycle with two red tail lights\"."}], "task": "refering", "answer_template": "The \"a brown and black motorcycle with two red tail lights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 31, 32, 33, 52, 53, 54, 73, 74, 75, 76, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338], "bbox": [0.12142187499999998, 0.0476875, 0.7941093749999999, 0.8723124999999999], "iscrowd": 0, "area": 80267.48924999998, "rle": {"size": [480, 640], "counts": "jgT1f0V>7I5K6N1N2O1O1O1N2O2]NmNUET1i:nNVES1h:oNVES1h:oNWER1g:POXEQ1f:QOYEP1e:ROYEQ1e:POZEQ1d:ROZEo0e:ROZEo0d:RO[EP1d:QO[EP1c:QO]Eo0c:QO]EP1a:RO^Eo0a:QO_EP1_:ROaEm0_:SOaEn0]:TObEm0]:SOcEn0[:SOeEn0Z:ROfEn0Y:ROhEo0W:QOiEP1U:POlEQ1S:oNmEQ1T:nNlES1S:lNnEU1Q:kNoEV1P:iNQFX1o9gNQFY1o9gNQFZ1n9eNTF[1k9eNUF[1l9gNQFZ1n9jNnEV1R:nNjES1V:QOeEo0\\:TO`Em0_:k10O010O01O0M4M2M4M2N3N1O1O1O2N1O100O101O0001O01O00000O100VHjJi5W5P2OoFQK9Ii6V5nHQK8Ji6V5nHQK9Ii6V5mHQK:Ji6U5lHRK:Ij6V5kHRK;Hj6V5jHSK;Hj6U5kHSK<Hi6U5jHTK<Hi6U5jHTK<Hj6S5kHUK;Gj6W3]H_N=dN;Gj6V3aH\\N:gN;Hj6T3dH[N6jN<Gi6S3iHZN1mN<Gj6Q3lHXNNQO;Fk6Q3oHVNJSO=Fi6Q3SISNGWO<Fj6o2VIRNCZO=Ei6n2[IQN_O\\O<Fi6m2_InM\\O@;Ek6l2aImMXOB=Ei6l2eIjMUOF<Ej6i2iIjMPOI<Ej6i2lIgMmNL=Di6k2oI_MnN1:Ej6k2PJYMoN87Di6m2RJRMPO>4Di6m2UJkLROd0OEj6m2mJ_MYNDi6n2nJ]MYNFi6n2lJ_MXNCl6o2kJ`MXNAl6Q3kJ_MWNAn6P3jJbMVN^Oo6R3iJbMVN]OP7R3hJdMVNZOS7S3lHiLU1l0kNXOS7j3nIRMlNUOV7k3kIRMmNTOW7m3iIQMoNROW7P4gIbMZ6`2bIbM]6a2`I`M_6[NkHc3c0SNb6YNmHe3=TNe6XNPId38UNh6VNRIf33UNj6UNVIg3KVNn6TNXIh3FVNQ7QNkIl5S6TJoIk5P6VJPJk5o5TJSJl5k5TJWJk5i5UJWJl5g5TJ[Jl5c5TJ_Jl5`5TJ`Jl5_5TJcJl5[5TJgJl5X5TJhJl5W5TJkJl5T5SJnJm5P5TJPKl5o4TJSKl5l4SJVKm5h4TJXKm5f4SJ\\Kl5d4SJ^Km5`4TJ`Km5_4RJcKm5\\4SJfKm5X4SJiKn5V4RJkKm5T4SJnKm5P4SJQLm5o3SJRLm5l3SJVLl5j3SJXLm5f3TJ[Lk5d3UJ]Ll5b3SJ`Lm5^3TJcLk5]3TJeLl5Y3TJiLj5W3WJjLh5V3WJlLg5T3YJnLe5R3\\JoLc5Q3\\JPMc5P3]JRMa5o2_JRM`5m2`JUM^5k2bJWM\\5j2dJWM[5h2eJZMY5f2gJ\\MW5e2iJ\\MV5c2jJ^MU5c2jJ_MU5`2lJaMR5_2nJcMP5^2oJdMP5[2QKfMo4X2QKjMo4U2PKlMQ5R2PKoMQ5o1nJSNR5k1nJWNR5g1oJZNR5d1mJ]NT5a1lJaNT5]1mJdNT5Z1kJhNU5V1kJlNU5dL`Jj3;BW5^LbJo37EW5WLgJR42IW5PLkJW4NJi55VJMi53VJNk51UJ0j50UJ2k5MTJ5k5KUJ7j5HUJ:j5FUJ<i5EWJ<h5DWJ>h5BWJ`0g5BXJ?g5BWJ`0g5BWJ`0h5AWJ`0h5^4M2O1O1N2O1O1N2O000O10O010000O11O2N1O1O1O1O2N1N2O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1N2O2N1O1O1O1O2M2O1O1O1O2N1N2O1O1N3N1O000O101O0O1000000O101O0O1000000O2O000O1000000O2O000O100000O010000O10000O1000000O10000000O100000000000000000000000000O10000000000000000000000000000O101O00001O00001O0000001O000O2O00001O0000001N1O1O2M2O1O2N2N1O2N2N2N2N1N3N2N2N1O2N2N2N2N1O2N2M3N1O2N2N2N2N1O2N2N2N1N3N2N2N2N1O2L4K5L3L6K5J6K4fMTE:R;AaELd:OoE]OV:`0\\FmNi9n0X2L3L5L4K4M4K5L3Ledm1"}, "label": "a brown and black motorcycle with two red tail lights"}]}
{"id": 493131, "image": "COCO_train2014_000000493131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the snowboard the boy in the green hat is riding\"."}], "task": "refering", "answer_template": "The \"the snowboard the boy in the green hat is riding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 263, 264, 265, 266, 284, 285, 286, 287, 307, 308, 309], "bbox": [0.34178125000000004, 0.6439387308533916, 0.5901875000000001, 0.8496936542669583], "iscrowd": 0, "area": 3312.391449999999, "rle": {"size": [457, 640], "counts": "ZSR31W>2N100O2N100O2O0O1O2O0O1O2O0O2N100O2O0O2N100O1O2O0O1O2O0O1O101N1O100O2O0O100001O001O1O001O001O001O001N2O0O2N101N1O2O0O2N2N010O1O100O1O100O1O1O10O01OROHmC8R<InC6R<LmC3S<NlC2T<0kC0T<1lCNU<3jCLV<5iCLV<6iCIW<8hCHX<:gCE[<;cCF^<9bCF_<;_CEc<:YCJh<6TCLm<4RCLP=4lBNV=b00O2N100O2N1O101N1O100O1O1O100O1O1O100O1O100O1OTOGiC9X<GhC9W<HiC7W<JiC5X<LgC4X<MhC2X<OgC1Z<OfC0Z<2eCNZ<7bCH^<=^CBc<b0PCGP=h01O100O1O010O100O1O100O2N101N1O101N1O101N1O2O0O1O2O0O4L5J5LQdd3"}, "label": "the snowboard the boy in the green hat is riding"}]}
{"id": 493131, "image": "COCO_train2014_000000493131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red snowboard is holding a young boy on a small hill\"."}], "task": "refering", "answer_template": "The \"a red snowboard is holding a young boy on a small hill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 263, 264, 265, 266, 284, 285, 286, 287, 307, 308, 309], "bbox": [0.34178125000000004, 0.6439387308533916, 0.5901875000000001, 0.8496936542669583], "iscrowd": 0, "area": 3312.391449999999, "rle": {"size": [457, 640], "counts": "ZSR31W>2N100O2N100O2O0O1O2O0O1O2O0O2N100O2O0O2N100O1O2O0O1O2O0O1O101N1O100O2O0O100001O001O1O001O001O001O001N2O0O2N101N1O2O0O2N2N010O1O100O1O100O1O1O10O01OROHmC8R<InC6R<LmC3S<NlC2T<0kC0T<1lCNU<3jCLV<5iCLV<6iCIW<8hCHX<:gCE[<;cCF^<9bCF_<;_CEc<:YCJh<6TCLm<4RCLP=4lBNV=b00O2N100O2N1O101N1O100O1O1O100O1O1O100O1O100O1OTOGiC9X<GhC9W<HiC7W<JiC5X<LgC4X<MhC2X<OgC1Z<OfC0Z<2eCNZ<7bCH^<=^CBc<b0PCGP=h01O100O1O010O100O1O100O2N101N1O101N1O101N1O2O0O1O2O0O4L5J5LQdd3"}, "label": "a red snowboard is holding a young boy on a small hill"}]}
{"id": 230996, "image": "COCO_train2014_000000230996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus\"."}], "task": "refering", "answer_template": "The \"a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183], "bbox": [0.631546875, 0.0, 1.0, 0.5204694835680751], "iscrowd": 0, "area": 43132.90735000003, "rle": {"size": [426, 640], "counts": "dWX52W=4L4L3M2N1O2N2N2N1O2N2N1O1O1O2N1N2ZKTOlLn0T3^O_Lc0a3HRL:n31eK2Z49bIbN?W1o5b1mI_NR6f1iI[NW6h1eIYN[6k1`IVN`6m1\\ITNd6o1XISNg6Q2TIPNl6S2PInMP7V400000000000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000000000001O000O2O00001O00001O00001O001O001O00001O001O00001O001O001O5K4L1O1O1O001O001O001O1O001O001O001O1O001O0000O100000000000000O100000O10000000O100000000000000O100000000000000O100000000000000O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O100O100O1O100O100O1O11O1N2O2N1O1O1O2N1O1O00000000001O0000000000001O0000O1O1O1O1O1O1O1"}, "label": "a bus"}]}
{"id": 230996, "image": "COCO_train2014_000000230996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a trolley bus stopped at a crosswalk\"."}], "task": "refering", "answer_template": "The \"a trolley bus stopped at a crosswalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183], "bbox": [0.631546875, 0.0, 1.0, 0.5204694835680751], "iscrowd": 0, "area": 43132.90735000003, "rle": {"size": [426, 640], "counts": "dWX52W=4L4L3M2N1O2N2N2N1O2N2N1O1O1O2N1N2ZKTOlLn0T3^O_Lc0a3HRL:n31eK2Z49bIbN?W1o5b1mI_NR6f1iI[NW6h1eIYN[6k1`IVN`6m1\\ITNd6o1XISNg6Q2TIPNl6S2PInMP7V400000000000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000000000001O000O2O00001O00001O00001O001O001O00001O001O00001O001O001O5K4L1O1O1O001O001O001O1O001O001O001O1O001O0000O100000000000000O100000O10000000O100000000000000O100000000000000O100000000000000O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O100O100O1O100O100O1O11O1N2O2N1O1O1O2N1O1O00000000001O0000000000001O0000O1O1O1O1O1O1O1"}, "label": "a trolley bus stopped at a crosswalk"}]}
{"id": 34389, "image": "COCO_train2014_000000034389.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown sheep laying in hay next to a cat\"."}], "task": "refering", "answer_template": "The \"a brown sheep laying in hay next to a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 80, 81, 82, 83, 88, 89, 95, 96, 97, 98, 99, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 202, 203, 204, 205, 206, 207, 208, 209, 217, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237], "bbox": [0.2997658079625293, 0.18709375, 1.0, 0.675265625], "iscrowd": 0, "area": 63921.40925, "rle": {"size": [640, 427], "counts": "jV`22nc01N3M3M2O2M3M3N1Y]OAja0b0b]O1\\b0k0N2O1N1O2O1N2N2N2N1N3M3M3N2M2bNeMUA]2j>hMPAZ2o>jMl@W2S?[101O000O2O00001N101O001N2O1O10000O10000000O01000002N2M3N2N2N3M2M3N3M2N2N2M4M2N3M;E;E9G6J5J7J5K1O1O2N1O1N2O1O1O2N1O1O1N2ZD_Hk:b7oDeHo:]7jDkHS;V7gDQIW;n7N2N2N2N3^EUGj9`900000000001O00000000001O00000000001O003M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3L4M1O001N2O1O1N2O001O1N2O1O1O0O2O1O2M4M2N3M3L4M2N3M3L4M3M1OO2O0O101N100O101N100O101N100O2O0O100O2O0O100O2O0O100O2O0O100O100O100O1O1O001O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1O100O1O1O1O10O01O1O010O1O001O10O01O1O00100O001O10O01O001O010O010O0100O010O010O10O0100O010O10O0100O010O10O01000O0100O0100O0100VK`DT1_;kNdDS1];kNeDU1[;iNgDV1Y;iNiDW1W;gNkDY1U;gNkDX1U;hNlDX1T;gNnDX1R;hNnDW1R;iNoDW1Q;iNoDW1Q;hNPEW1P;iNQEW1o:iNQEV1o:jNREV1n:jNREV1n:iNSEV1m:jNTE?l;VOVD]OT>_OSBSOf9"}, "label": "a brown sheep laying in hay next to a cat"}]}
{"id": 34389, "image": "COCO_train2014_000000034389.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black goat next to a white cat in the grass\"."}], "task": "refering", "answer_template": "The \"black goat next to a white cat in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 80, 81, 82, 83, 88, 89, 95, 96, 97, 98, 99, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 202, 203, 204, 205, 206, 207, 208, 209, 217, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237], "bbox": [0.2997658079625293, 0.18709375, 1.0, 0.675265625], "iscrowd": 0, "area": 63921.40925, "rle": {"size": [640, 427], "counts": "jV`22nc01N3M3M2O2M3M3N1Y]OAja0b0b]O1\\b0k0N2O1N1O2O1N2N2N2N1N3M3M3N2M2bNeMUA]2j>hMPAZ2o>jMl@W2S?[101O000O2O00001N101O001N2O1O10000O10000000O01000002N2M3N2N2N3M2M3N3M2N2N2M4M2N3M;E;E9G6J5J7J5K1O1O2N1O1N2O1O1O2N1O1O1N2ZD_Hk:b7oDeHo:]7jDkHS;V7gDQIW;n7N2N2N2N3^EUGj9`900000000001O00000000001O00000000001O003M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3L4M1O001N2O1O1N2O001O1N2O1O1O0O2O1O2M4M2N3M3L4M2N3M3L4M3M1OO2O0O101N100O101N100O101N100O2O0O100O2O0O100O2O0O100O2O0O100O100O100O1O1O001O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1O100O1O1O1O10O01O1O010O1O001O10O01O1O00100O001O10O01O001O010O010O0100O010O010O10O0100O010O10O0100O010O10O01000O0100O0100O0100VK`DT1_;kNdDS1];kNeDU1[;iNgDV1Y;iNiDW1W;gNkDY1U;gNkDX1U;hNlDX1T;gNnDX1R;hNnDW1R;iNoDW1Q;iNoDW1Q;hNPEW1P;iNQEW1o:iNQEV1o:jNREV1n:jNREV1n:iNSEV1m:jNTE?l;VOVD]OT>_OSBSOf9"}, "label": "black goat next to a white cat in the grass"}]}
{"id": 34389, "image": "COCO_train2014_000000034389.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an animal in the background with only the leg showing\"."}], "task": "refering", "answer_template": "The \"an animal in the background with only the leg showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 24, 25, 26, 27, 28, 29, 39, 40, 54, 55, 69, 70, 84, 85, 100], "bbox": [0.4886182669789227, 0.0025468749999999997, 0.9968618266978923, 0.276375], "iscrowd": 0, "area": 7710.237550000002, "rle": {"size": [640, 427], "counts": "RXS41oc0001O1O1O001O1O001O1O001O000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000009G:T@QOS<Y1bCgN_<`1YCaNg<b1VC]Nk<e1RC\\Nn<f1oB[NQ=g1lBYNU=j1gBWNY=k1aBYN_=i1ZB\\Nf=g1RB^Nn=d1jAbNV>`1cAfN\\>_301O001O00001O001O001O001O4L7I8H8HO1cM`AYOa>a0^BdNd=V1XCSNi<g1kCiMW<Q2o2L4M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M300O100O100O10000O100O10000O100O100O100O100O100O100O100O13M:F7I2N1O2N1O2N1O1O2N1O1O00000000000000O100000000ROc]O9]b0Em]O3Sb0MV^OJja06P100O1000000000000O10000001O2N3M1O2N2N2N1O2N2N1O2N2N2N1O2N2N00000000000000000000YOT]O:lb0F\\]O2db0Nd]OJ\\b06g0O100000nc0"}, "label": "an animal in the background with only the leg showing"}]}
{"id": 34389, "image": "COCO_train2014_000000034389.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white hoof of an animal\"."}], "task": "refering", "answer_template": "The \"a white hoof of an animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 24, 25, 26, 27, 28, 29, 39, 40, 54, 55, 69, 70, 84, 85, 100], "bbox": [0.4886182669789227, 0.0025468749999999997, 0.9968618266978923, 0.276375], "iscrowd": 0, "area": 7710.237550000002, "rle": {"size": [640, 427], "counts": "RXS41oc0001O1O1O001O1O001O1O001O000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000009G:T@QOS<Y1bCgN_<`1YCaNg<b1VC]Nk<e1RC\\Nn<f1oB[NQ=g1lBYNU=j1gBWNY=k1aBYN_=i1ZB\\Nf=g1RB^Nn=d1jAbNV>`1cAfN\\>_301O001O00001O001O001O001O4L7I8H8HO1cM`AYOa>a0^BdNd=V1XCSNi<g1kCiMW<Q2o2L4M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M300O100O100O10000O100O10000O100O100O100O100O100O100O100O13M:F7I2N1O2N1O2N1O1O2N1O1O00000000000000O100000000ROc]O9]b0Em]O3Sb0MV^OJja06P100O1000000000000O10000001O2N3M1O2N2N2N1O2N2N1O2N2N2N1O2N2N00000000000000000000YOT]O:lb0F\\]O2db0Nd]OJ\\b06g0O100000nc0"}, "label": "a white hoof of an animal"}]}
{"id": 362071, "image": "COCO_train2014_000000362071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe facing away\"."}], "task": "refering", "answer_template": "The \"the giraffe facing away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 102, 103, 126, 149, 150, 172, 173, 174, 195, 196, 197, 219, 220, 221, 243, 244], "bbox": [0.4150625, 0.11762500000000001, 0.6234375, 0.62425], "iscrowd": 0, "area": 7364.795599999999, "rle": {"size": [480, 640], "counts": "]hl31o>2N101N1O2N1O1O2O0O2N1O2N1O2N3N2M4aBTOo;5fCj0M^O[<o1N2N2N1O2O10O0M4L4L4L3N3L4L2N000000O103M3M3M3M3O1ON3M2I8G800100O10O012N3M2N3]OaB^Oe>LYBh0h;TO\\DP1j<3M3M3M3M4L3M3M3M3K6F9G3L3N2N3M2N3M2M4M2002O2M2O1N2N3N1N2O1N2N3N1N2O1N3M2O1N2O2M2N2O1N2O2M1O2O1N2O0O2N2O1N101N2N101N2O1N101N2N101N2O2M4L5L4K5L3L5Kge`3"}, "label": "the giraffe facing away"}]}
{"id": 362071, "image": "COCO_train2014_000000362071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe sticking its head out of the bushes to see further into the forest\"."}], "task": "refering", "answer_template": "The \"a giraffe sticking its head out of the bushes to see further into the forest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 102, 103, 126, 149, 150, 172, 173, 174, 195, 196, 197, 219, 220, 221, 243, 244], "bbox": [0.4150625, 0.11762500000000001, 0.6234375, 0.62425], "iscrowd": 0, "area": 7364.795599999999, "rle": {"size": [480, 640], "counts": "]hl31o>2N101N1O2N1O1O2O0O2N1O2N1O2N3N2M4aBTOo;5fCj0M^O[<o1N2N2N1O2O10O0M4L4L4L3N3L4L2N000000O103M3M3M3M3O1ON3M2I8G800100O10O012N3M2N3]OaB^Oe>LYBh0h;TO\\DP1j<3M3M3M3M4L3M3M3M3K6F9G3L3N2N3M2N3M2M4M2002O2M2O1N2N3N1N2O1N2N3N1N2O1N3M2O1N2O2M2N2O1N2O2M1O2O1N2O0O2N2O1N101N2N101N2O1N101N2N101N2O2M4L5L4K5L3L5Kge`3"}, "label": "a giraffe sticking its head out of the bushes to see further into the forest"}]}
{"id": 255576, "image": "COCO_train2014_000000255576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is train on the railway tracks\"."}], "task": "refering", "answer_template": "The \"there is train on the railway tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.17881249999999999, 0.23311475409836066, 0.8475156250000001, 0.9065339578454333], "iscrowd": 0, "area": 97571.14249999997, "rle": {"size": [427, 640], "counts": "Qi_11W==DV1iN9H7H7I5L4K4M4L3N3M3N1N3M2N2O1N2N1O2O1N2N2N2O1N1O2N2O1N1O2N100O2N1O1O2N1O101N1O1O2N1O1O2M2O1O2N1O1O1O2N1O1O2N1O1N2O2N1O1O2M2O1O1O2M2O1O1O2M2O1O2O0O100O10000O101N10000O100O101O0O100O10000O2O0O10000O100O2O000O1000000O2O000O10001O0O2O001O0O2O001O001O1O001N101O001N1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O101O00000O1000000O1000000O1000000O2O00000O1000001N101O000O2O00001N10001O000O2O00001N100000O00100O1O1O010O1O100O1O0101O1O2N1O1N2O1O1O1O0O10001O0000001O01O01O00000010O000001O01O0001O0001OO2O0O100O1O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O10000000000000000000000000000000000001O0000001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001O001N1O2M2O1N3N1N3N1N3O0100O10O0100O100O010OO2N1O101N1O2N1O2N101N1O2N3M3M4L3N3L3M4M3L3M4POZ1bNYgX1"}, "label": "there is train on the railway tracks"}]}
{"id": 255576, "image": "COCO_train2014_000000255576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black train with yellow on front\"."}], "task": "refering", "answer_template": "The \"black train with yellow on front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.17881249999999999, 0.23311475409836066, 0.8475156250000001, 0.9065339578454333], "iscrowd": 0, "area": 97571.14249999997, "rle": {"size": [427, 640], "counts": "Qi_11W==DV1iN9H7H7I5L4K4M4L3N3M3N1N3M2N2O1N2N1O2O1N2N2N2O1N1O2N2O1N1O2N100O2N1O1O2N1O101N1O1O2N1O1O2M2O1O2N1O1O1O2N1O1O2N1O1N2O2N1O1O2M2O1O1O2M2O1O1O2M2O1O2O0O100O10000O101N10000O100O101O0O100O10000O2O0O10000O100O2O000O1000000O2O000O10001O0O2O001O0O2O001O001O1O001N101O001N1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O101O00000O1000000O1000000O1000000O2O00000O1000001N101O000O2O00001N10001O000O2O00001N100000O00100O1O1O010O1O100O1O0101O1O2N1O1N2O1O1O1O0O10001O0000001O01O01O00000010O000001O01O0001O0001OO2O0O100O1O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O10000000000000000000000000000000000001O0000001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001O001N1O2M2O1N3N1N3N1N3O0100O10O0100O100O010OO2N1O101N1O2N1O2N101N1O2N3M3M4L3N3L3M4M3L3M4POZ1bNYgX1"}, "label": "black train with yellow on front"}]}
{"id": 108123, "image": "COCO_train2014_000000108123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a base ball player playing\"."}], "task": "refering", "answer_template": "The \"a base ball player playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 338, 339, 340, 341, 342, 357, 358, 363, 364, 365, 379, 380, 381, 387, 388, 402, 403, 404], "bbox": [0.47928125, 0.162578125, 0.91759375, 0.97328125], "iscrowd": 0, "area": 47605.24075000001, "rle": {"size": [512, 640], "counts": "\\oi41n?1O2O0O2N1O2O0O1O2N1O100O2XGCj1>TNDk1=SNEm1;RNFm1;QNHm19QNIn18PNJo17oMLP24nMNQ23mMOR22mM0Q21mM1R20lM2T2NjM5T2LjM6U2KiM7V2JhM9W2GiM9S2KmM5P2NoM3n10RN0j14UNMh15YNLb18]NI`1:`NF]1=bNDZ1`0fN@W1c0iN]OS1g0lNZOQ1i0oNWOl0n0SOTOf0R1ZOnN?Y1@hN:^1FbN3e1L\\NMk13UNGP2:PNAU2>lM^OX2b0iMXO\\2g0eMUO_2k0aMQOc2n0^MnNf2R1ZMjNj2V1VMfNn2Y1SMcNQ3]1oL_NU3`1mL[NW3e1iLWN[3h1fLUN]3k1cLSN_3l1bLRN_3o1aLnMb3R2^LlMd3S2]LjMf3V2[LgMg3X2ZLeMi3[2WLcMk3\\2VLbMl3^2TL_Mm3b2TL\\Ml3f2TLWMm3k2SLSMl3o2VLmLk3U3ULiLk3X3VLeLQ3W4oLgKo2\\4RMbKl2b4TM[Kk2i4UMUKi2n4XMoJg2U5ZMhJd2[5]MbJd2`5\\M^Jc2d5^MZJa2i5`MSJa2n5`MQJ^2R6cMlI\\2U6fMiIX2Z6hMdIX2]6jMaIT2b6lM]IR2e6PNYIm1lNeKm7^2VIk1n6WNoHh1T7XNkHf1X7[NeHd1]7]NbHa1a7`N\\H_1f7bNYH\\1j7eNSHZ1o7hNoGV1T8jNjGU1X8mNeGR1^8nNaGP1a8RO\\Gm0g8SOXGk0j8WOSGh0P9XOoFf0S9\\OjFc0Y9]OfFa0\\9[3G6I8I7H7J1N2O1N2O1O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1N2N2N1O2N101N1O1O2N1O2N101N1O1N3N1O2N1O2N1O1O2N1O2N1O2N1O2N1OeMjJoIT5n5QKQJn4k5XKUJf4h5`KWJ^4e5hKZJW4c5nK\\JQ4c5SL[Jl3d5VL]Jh3a5\\L^Jc3a5`L^J_3a5cL_J\\3_5hLaJW3]5lLbJT3\\5nLdJS3X5QMgJo2W5TMiJk2U5WMkJi2S5ZMlJf2R5\\MnJd2P5_MPK`2n4cMQK]2l4fMTKZ2j4iMUKW2i4lMWKS2g4oMYKR2d4QN[Ko1P4PKnJS3R1m1n3iNRLV1l3mNSLS1k3oNULQ1i3ROWLm0g3VOXLj0f3XOZLh0d3[O[Le0c3^O]La0a3A_L`0^3CaL=]3EcL=Y3FgL;U3HjL:S3GmL:Q3HnL:o2HQM9k2IUM9h2IWM9f2IYM9d2H]M8`2K_M7^2JbM8[2JdM8Y2IhM:R2ImM<m1ESN`0g1AYNa0d1A\\Na0_1AaNa0\\1@dNb0Y1@gNb0U1_OkNc0Q1_OoNd0m0^OROd0:NG3L84K_O`0`0BTOh0l0V74L3M4M4K7I7I6J7I01O01O000010O0001O00010O00001O00010O00000001N1O100O100O3M6J7I8H7H8I6J6J6I8IbQj0"}, "label": "a base ball player playing"}]}
{"id": 108123, "image": "COCO_train2014_000000108123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player who just has hit a baseball with his bat\"."}], "task": "refering", "answer_template": "The \"a baseball player who just has hit a baseball with his bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 338, 339, 340, 341, 342, 357, 358, 363, 364, 365, 379, 380, 381, 387, 388, 402, 403, 404], "bbox": [0.47928125, 0.162578125, 0.91759375, 0.97328125], "iscrowd": 0, "area": 47605.24075000001, "rle": {"size": [512, 640], "counts": "\\oi41n?1O2O0O2N1O2O0O1O2N1O100O2XGCj1>TNDk1=SNEm1;RNFm1;QNHm19QNIn18PNJo17oMLP24nMNQ23mMOR22mM0Q21mM1R20lM2T2NjM5T2LjM6U2KiM7V2JhM9W2GiM9S2KmM5P2NoM3n10RN0j14UNMh15YNLb18]NI`1:`NF]1=bNDZ1`0fN@W1c0iN]OS1g0lNZOQ1i0oNWOl0n0SOTOf0R1ZOnN?Y1@hN:^1FbN3e1L\\NMk13UNGP2:PNAU2>lM^OX2b0iMXO\\2g0eMUO_2k0aMQOc2n0^MnNf2R1ZMjNj2V1VMfNn2Y1SMcNQ3]1oL_NU3`1mL[NW3e1iLWN[3h1fLUN]3k1cLSN_3l1bLRN_3o1aLnMb3R2^LlMd3S2]LjMf3V2[LgMg3X2ZLeMi3[2WLcMk3\\2VLbMl3^2TL_Mm3b2TL\\Ml3f2TLWMm3k2SLSMl3o2VLmLk3U3ULiLk3X3VLeLQ3W4oLgKo2\\4RMbKl2b4TM[Kk2i4UMUKi2n4XMoJg2U5ZMhJd2[5]MbJd2`5\\M^Jc2d5^MZJa2i5`MSJa2n5`MQJ^2R6cMlI\\2U6fMiIX2Z6hMdIX2]6jMaIT2b6lM]IR2e6PNYIm1lNeKm7^2VIk1n6WNoHh1T7XNkHf1X7[NeHd1]7]NbHa1a7`N\\H_1f7bNYH\\1j7eNSHZ1o7hNoGV1T8jNjGU1X8mNeGR1^8nNaGP1a8RO\\Gm0g8SOXGk0j8WOSGh0P9XOoFf0S9\\OjFc0Y9]OfFa0\\9[3G6I8I7H7J1N2O1N2O1O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1N2N2N1O2N101N1O1O2N1O2N101N1O1N3N1O2N1O2N1O1O2N1O2N1O2N1O2N1OeMjJoIT5n5QKQJn4k5XKUJf4h5`KWJ^4e5hKZJW4c5nK\\JQ4c5SL[Jl3d5VL]Jh3a5\\L^Jc3a5`L^J_3a5cL_J\\3_5hLaJW3]5lLbJT3\\5nLdJS3X5QMgJo2W5TMiJk2U5WMkJi2S5ZMlJf2R5\\MnJd2P5_MPK`2n4cMQK]2l4fMTKZ2j4iMUKW2i4lMWKS2g4oMYKR2d4QN[Ko1P4PKnJS3R1m1n3iNRLV1l3mNSLS1k3oNULQ1i3ROWLm0g3VOXLj0f3XOZLh0d3[O[Le0c3^O]La0a3A_L`0^3CaL=]3EcL=Y3FgL;U3HjL:S3GmL:Q3HnL:o2HQM9k2IUM9h2IWM9f2IYM9d2H]M8`2K_M7^2JbM8[2JdM8Y2IhM:R2ImM<m1ESN`0g1AYNa0d1A\\Na0_1AaNa0\\1@dNb0Y1@gNb0U1_OkNc0Q1_OoNd0m0^OROd0:NG3L84K_O`0`0BTOh0l0V74L3M4M4K7I7I6J7I01O01O000010O0001O00010O00001O00010O00000001N1O100O100O3M6J7I8H7H8I6J6J6I8IbQj0"}, "label": "a baseball player who just has hit a baseball with his bat"}]}
{"id": 108123, "image": "COCO_train2014_000000108123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"georgia player with his hands on his hips\"."}], "task": "refering", "answer_template": "The \"georgia player with his hands on his hips\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 53, 54, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 168, 169, 170, 191, 192, 193, 194, 214, 215, 216, 217, 238, 239, 240], "bbox": [0.27740624999999997, 0.0941796875, 0.48321875, 0.6087109374999999], "iscrowd": 0, "area": 17006.5324, "rle": {"size": [512, 640], "counts": "VTi29`?:J6L3M4L4L4L3N3L4L3M4L4L3M4L3M2N3M3M2N3M3M2N3M3M2N3M3M3M2N3M3L5L5K3M1QOTLhEm3GPLQ:?kEb34SLl9j4SFZKi9g4WF]Kd9d4\\FcK\\9^4dFiKT9X4kFPLn8o3SGYLd8h3\\G_L\\8b3dGeLT8\\3lGlLk7U3THSMd7n2\\HSMb7n2^HRMb7m2_HTM`7l2`HTM`7l2_HUMb7j2^HWMb7h2^HXMc7g2]HYMc7g2]HZMc7d2]H]Md7b2\\H^Me7a2[H_Mf7`2ZHaMe7_2[HaMf7^2YHcMh7\\2XHeMh7Z2XHfMh7Y2YHeKJZ1n7P3XHiK6i0e7Z3THQLd05c7_3iG_LP1Ca7d3_GlLX:T3hEoLV:P3jESMT:k2lEYMQ:g2oE\\Mo9c2QF`Ml9`2TFcMj9\\2VFgMg9Y2XFkMf9T2ZFoMd9o1]FTN`9l1`FWN^9h1bF[N[9e1dF_NZ9`1fFcNX9\\1hFgNU9Y1kFjNS9T1nFlNR9T1mFmNT9R1lFnNU9Q1kFoNU9Q1kFoNV9l0nFTOR9>[GDe8OgG1Y8NhG2Y8KiG5X8IhG8X8GiG9X8ElG9S8ETH6m7HZH2g7L_HOa70fHJ[73mHGS78SICn6;YI@g6>_I]Oa6a0gIYOZ6e0lITOV6k0QJ]Nb6a1P4M3N2N3M2N2M3N3M2N2M3N2N3M2M3N2N3M2M3N2Nl[U5"}, "label": "georgia player with his hands on his hips"}]}
{"id": 108123, "image": "COCO_train2014_000000108123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing wearing a baseball uniform with the word georgia on it\"."}], "task": "refering", "answer_template": "The \"a man standing wearing a baseball uniform with the word georgia on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 53, 54, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 168, 169, 170, 191, 192, 193, 194, 214, 215, 216, 217, 238, 239, 240], "bbox": [0.27740624999999997, 0.0941796875, 0.48321875, 0.6087109374999999], "iscrowd": 0, "area": 17006.5324, "rle": {"size": [512, 640], "counts": "VTi29`?:J6L3M4L4L4L3N3L4L3M4L4L3M4L3M2N3M3M2N3M3M2N3M3M2N3M3M3M2N3M3L5L5K3M1QOTLhEm3GPLQ:?kEb34SLl9j4SFZKi9g4WF]Kd9d4\\FcK\\9^4dFiKT9X4kFPLn8o3SGYLd8h3\\G_L\\8b3dGeLT8\\3lGlLk7U3THSMd7n2\\HSMb7n2^HRMb7m2_HTM`7l2`HTM`7l2_HUMb7j2^HWMb7h2^HXMc7g2]HYMc7g2]HZMc7d2]H]Md7b2\\H^Me7a2[H_Mf7`2ZHaMe7_2[HaMf7^2YHcMh7\\2XHeMh7Z2XHfMh7Y2YHeKJZ1n7P3XHiK6i0e7Z3THQLd05c7_3iG_LP1Ca7d3_GlLX:T3hEoLV:P3jESMT:k2lEYMQ:g2oE\\Mo9c2QF`Ml9`2TFcMj9\\2VFgMg9Y2XFkMf9T2ZFoMd9o1]FTN`9l1`FWN^9h1bF[N[9e1dF_NZ9`1fFcNX9\\1hFgNU9Y1kFjNS9T1nFlNR9T1mFmNT9R1lFnNU9Q1kFoNU9Q1kFoNV9l0nFTOR9>[GDe8OgG1Y8NhG2Y8KiG5X8IhG8X8GiG9X8ElG9S8ETH6m7HZH2g7L_HOa70fHJ[73mHGS78SICn6;YI@g6>_I]Oa6a0gIYOZ6e0lITOV6k0QJ]Nb6a1P4M3N2N3M2N2M3N3M2N2M3N2N3M2M3N2N3M2M3N2Nl[U5"}, "label": "a man standing wearing a baseball uniform with the word georgia on it"}]}
{"id": 280156, "image": "COCO_train2014_000000280156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an arm propped on the window ledge inside a car\"."}], "task": "refering", "answer_template": "The \"an arm propped on the window ledge inside a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [330, 331, 352, 353, 354, 355, 374, 375, 376, 377, 378, 379, 396, 397, 398, 399, 400, 401, 402, 420, 421, 422, 423, 424, 425, 426, 445, 446, 447, 448, 468, 469, 470, 471], "bbox": [0.0, 0.7146078431372549, 0.44102941176470595, 1.0], "iscrowd": 0, "area": 18099.56254999999, "rle": {"size": [612, 612], "counts": "g=^2f`00O1O11O000001O00000000001O000001O0001O00000000001O0000010O000010O000010O00010O00010O0001O010O00010O0010O0001O010O0010O0001O010O0010O00010O01O00010O001O0010O0001O010O001O0010O0001O010O001O010O001O00010O1O0010O01O1O00100O001O10O01O010O10O0100O010O010O10O0100O010O010O10O0100O010O10O010O0100O010O10O010O0100O010O10O010O0100O010000O010O10O10O10O10O100O01000O01001N3N2N3L3N3]_OhMR`0e21O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O002N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O1O1O1O2N1O1O4Lhi[6"}, "label": "an arm propped on the window ledge inside a car"}]}
{"id": 280156, "image": "COCO_train2014_000000280156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vehicle driver ' s arm resting on the windowsill as he drives\"."}], "task": "refering", "answer_template": "The \"a vehicle driver ' s arm resting on the windowsill as he drives\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [330, 331, 352, 353, 354, 355, 374, 375, 376, 377, 378, 379, 396, 397, 398, 399, 400, 401, 402, 420, 421, 422, 423, 424, 425, 426, 445, 446, 447, 448, 468, 469, 470, 471], "bbox": [0.0, 0.7146078431372549, 0.44102941176470595, 1.0], "iscrowd": 0, "area": 18099.56254999999, "rle": {"size": [612, 612], "counts": "g=^2f`00O1O11O000001O00000000001O000001O0001O00000000001O0000010O000010O000010O00010O00010O0001O010O00010O0010O0001O010O0010O0001O010O0010O00010O01O00010O001O0010O0001O010O001O0010O0001O010O001O010O001O00010O1O0010O01O1O00100O001O10O01O010O10O0100O010O010O10O0100O010O010O10O0100O010O10O010O0100O010O10O010O0100O010O10O010O0100O010000O010O10O10O10O10O100O01000O01001N3N2N3L3N3]_OhMR`0e21O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O002N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O1O1O1O2N1O1O4Lhi[6"}, "label": "a vehicle driver ' s arm resting on the windowsill as he drives"}]}
{"id": 280156, "image": "COCO_train2014_000000280156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a helmet and riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man wearing a helmet and riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 181, 182, 183, 203, 204, 205, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 339, 340, 361, 362, 383], "bbox": [0.21573529411764705, 0.3591830065359477, 0.49663398692810456, 0.7861437908496732], "iscrowd": 0, "area": 15675.605600000004, "rle": {"size": [612, 612], "counts": "UV_23mb06J6J7J5J6J7R_OkNe?Y1W@jNg?Y1U@iNj?Z1Q@jNm?Z1n_OiNP`0Z1k_OjNS`0Y1i_OiNV`0Z1f_OiNX`0S2N2N1nNVMiAl2U>WMhAk2U>YMhAi2U>[MhAf2V>^MgAd2V>`MgAb2V>bMgA`2V>dMgA]2W>gMfA[2X>hMeAZ2X>Z1O2O1O1O001O1O1O001O1N2O001O1O00001O000O2O00001O000010QNaBWN`=e1dBZN]=a1hB_NX=]1lBbNU=3VB@i0<R=3VB]Ol0`0n<1XB[On0c0l<OYB\\Ol0d0l<NZB^Oi0e0m<LZB_Oj0d0m<K[BAg0d0P=H[BDd0e0Q=F\\BDd0e0R=C\\BIa0d0U=TOhB83d0U=iNSCb0He0g>\\OYAc0h>]OWAd0i>[OXAd0i>\\OVAd0j>]OUAd0k>[OVAd0k>\\OTAd0m>\\OSAc0m>]OSAc0n>]ORAb0o>]OQAc0o>^OPAc0P?]OPAb0Q?]Oo@c0R?]Om@c0S?^Ol@b0U?]Ok@c0V?]Oi@c0W?^Oh@b0Y?]Og@d0Y?\\Of@d0[?\\Od@d0\\?\\Od@d0]?\\Ob@d0_?\\O`@d0`?\\O`@d0a?\\O^@d0c?\\O\\@e0c?[O]@e0c?\\O\\@d0e?[O[@e0e?\\OZ@d0g?\\OX@d0h?\\OX@d0h?]OW@c0j?]OU@d0j?\\OV@d0k?\\OT@d0l?]OS@c0m?]OS@c0n?]OR@b0n?_OQ@a0P`0^OP@b0P`0]OR@c0m?]OS@c0n?\\OR@d0n?\\OS@c0n?\\OR@d0n?\\OR@d0n?\\OS@g0j?XOV@l0f?TOZ@P1c?oN^@S1_?mNa@V1\\?jNd@Y1Z?fNg@\\1V?dNj@_1T?`Nl@c1Q?]NPAe1m>ZNTAi1j>VNVAm1g>SNZAo1d>PN\\AS2a>mM_AW2^>hMdAY2Y>gMiAZ2T>fMnA[2P>dMRB]2k=cMWB^2g=aM[B`2b=`M`Ba2]=_MeBb2Y=]MiBd2T=\\MnBc2R=\\MPCa2Q=^MRC_2P=`MQCV1eNCZ>WOSCn0iNJU>XOTCf0mN1Q>XOTC?QO8l=YOUC7UO`0g=XOVC0ZOf0a=ZOWCH^Om0\\=[OaDe0`;ZO`De0a;[O_Dd0c;WOYANU3i0c;YOaDg0`;XO`Dg0a;YO_Df0b;ZO^Df0c;YO^De0c;\\O\\Dc0f;\\OZDc0g;]OYDc0T?O2N102M4L4L5LUdg5"}, "label": "a man wearing a helmet and riding a motorcycle"}]}
{"id": 280156, "image": "COCO_train2014_000000280156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"biker wearing jeans , a white t - shirt and black helmet with a dog in a bag on his lap\"."}], "task": "refering", "answer_template": "The \"biker wearing jeans , a white t - shirt and black helmet with a dog in a bag on his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 181, 182, 183, 203, 204, 205, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 339, 340, 361, 362, 383], "bbox": [0.21573529411764705, 0.3591830065359477, 0.49663398692810456, 0.7861437908496732], "iscrowd": 0, "area": 15675.605600000004, "rle": {"size": [612, 612], "counts": "UV_23mb06J6J7J5J6J7R_OkNe?Y1W@jNg?Y1U@iNj?Z1Q@jNm?Z1n_OiNP`0Z1k_OjNS`0Y1i_OiNV`0Z1f_OiNX`0S2N2N1nNVMiAl2U>WMhAk2U>YMhAi2U>[MhAf2V>^MgAd2V>`MgAb2V>bMgA`2V>dMgA]2W>gMfA[2X>hMeAZ2X>Z1O2O1O1O001O1O1O001O1N2O001O1O00001O000O2O00001O000010QNaBWN`=e1dBZN]=a1hB_NX=]1lBbNU=3VB@i0<R=3VB]Ol0`0n<1XB[On0c0l<OYB\\Ol0d0l<NZB^Oi0e0m<LZB_Oj0d0m<K[BAg0d0P=H[BDd0e0Q=F\\BDd0e0R=C\\BIa0d0U=TOhB83d0U=iNSCb0He0g>\\OYAc0h>]OWAd0i>[OXAd0i>\\OVAd0j>]OUAd0k>[OVAd0k>\\OTAd0m>\\OSAc0m>]OSAc0n>]ORAb0o>]OQAc0o>^OPAc0P?]OPAb0Q?]Oo@c0R?]Om@c0S?^Ol@b0U?]Ok@c0V?]Oi@c0W?^Oh@b0Y?]Og@d0Y?\\Of@d0[?\\Od@d0\\?\\Od@d0]?\\Ob@d0_?\\O`@d0`?\\O`@d0a?\\O^@d0c?\\O\\@e0c?[O]@e0c?\\O\\@d0e?[O[@e0e?\\OZ@d0g?\\OX@d0h?\\OX@d0h?]OW@c0j?]OU@d0j?\\OV@d0k?\\OT@d0l?]OS@c0m?]OS@c0n?]OR@b0n?_OQ@a0P`0^OP@b0P`0]OR@c0m?]OS@c0n?\\OR@d0n?\\OS@c0n?\\OR@d0n?\\OR@d0n?\\OS@g0j?XOV@l0f?TOZ@P1c?oN^@S1_?mNa@V1\\?jNd@Y1Z?fNg@\\1V?dNj@_1T?`Nl@c1Q?]NPAe1m>ZNTAi1j>VNVAm1g>SNZAo1d>PN\\AS2a>mM_AW2^>hMdAY2Y>gMiAZ2T>fMnA[2P>dMRB]2k=cMWB^2g=aM[B`2b=`M`Ba2]=_MeBb2Y=]MiBd2T=\\MnBc2R=\\MPCa2Q=^MRC_2P=`MQCV1eNCZ>WOSCn0iNJU>XOTCf0mN1Q>XOTC?QO8l=YOUC7UO`0g=XOVC0ZOf0a=ZOWCH^Om0\\=[OaDe0`;ZO`De0a;[O_Dd0c;WOYANU3i0c;YOaDg0`;XO`Dg0a;YO_Df0b;ZO^Df0c;YO^De0c;\\O\\Dc0f;\\OZDc0g;]OYDc0T?O2N102M4L4L5LUdg5"}, "label": "biker wearing jeans , a white t - shirt and black helmet with a dog in a bag on his lap"}]}
{"id": 198240, "image": "COCO_train2014_000000198240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the kid who eats very messy\"."}], "task": "refering", "answer_template": "The \"the kid who eats very messy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 266, 267], "bbox": [0.38071875, 0.14497641509433962, 0.733359375, 0.7759198113207547], "iscrowd": 0, "area": 35008.258899999986, "rle": {"size": [424, 640], "counts": "iXU34R=4L4L3M4L4L4L3M4L4L4L3M4L4L4J4M3L4L4L4K5L4N2N2N2N2N2N2N3M2N2O1N3M2N2O2M2L5bN]LPIh3m6PMXHW3d7X1N3L3M2O2M3N2O001O1O1O001O1O1O1O010O1O1O1O1O001O1O1O1O010O1O1O100O01000O3N1O1N3N1O2N1N2O2N1N2O000O1000001N10000O101O0O1000001N10000O110OO2O00001O0O2O001O00001N101O00QJQIf5Y7O100O1O100O10O0100O1O100O\\KZJg1f5WN[Jj1d5TN^Jm1b5QN`Jo1_5oMcJS2[5lMfJU2Z5hMiJX2V5gMkJZ2U5cMnJ]2Q5bMPK^2Q5_MQKb2n4]MTKc2k4[MWKe2j4VMZKk2e4QM`Kn2a4mLcKT3\\4hLhKW3Z4dLkK[3W4`LlK`3V4ZLoKd3T4VLPLj3R4PLRLo3Q4kKTLT4n3gKULX4n3bKVL^4l3\\KYLc4V6O101N1O1M4L3N3L3M301O0O2O001O001N101O001N101O002M2O2N2N1N3N1O2M2O2M2N3N2M2N3N2M3M2O2M3M2N3M3N2M2N3M3L4M2N3M4L3M3M4L3L4M4L3M3MgaV2"}, "label": "the kid who eats very messy"}]}
{"id": 198240, "image": "COCO_train2014_000000198240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby\"."}], "task": "refering", "answer_template": "The \"baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 266, 267], "bbox": [0.38071875, 0.14497641509433962, 0.733359375, 0.7759198113207547], "iscrowd": 0, "area": 35008.258899999986, "rle": {"size": [424, 640], "counts": "iXU34R=4L4L3M4L4L4L3M4L4L4L3M4L4L4J4M3L4L4L4K5L4N2N2N2N2N2N2N3M2N2O1N3M2N2O2M2L5bN]LPIh3m6PMXHW3d7X1N3L3M2O2M3N2O001O1O1O001O1O1O1O010O1O1O1O1O001O1O1O1O010O1O1O100O01000O3N1O1N3N1O2N1N2O2N1N2O000O1000001N10000O101O0O1000001N10000O110OO2O00001O0O2O001O00001N101O00QJQIf5Y7O100O1O100O10O0100O1O100O\\KZJg1f5WN[Jj1d5TN^Jm1b5QN`Jo1_5oMcJS2[5lMfJU2Z5hMiJX2V5gMkJZ2U5cMnJ]2Q5bMPK^2Q5_MQKb2n4]MTKc2k4[MWKe2j4VMZKk2e4QM`Kn2a4mLcKT3\\4hLhKW3Z4dLkK[3W4`LlK`3V4ZLoKd3T4VLPLj3R4PLRLo3Q4kKTLT4n3gKULX4n3bKVL^4l3\\KYLc4V6O101N1O1M4L3N3L3M301O0O2O001O001N101O001N101O002M2O2N2N1N3N1O2M2O2M2N3N2M2N3N2M3M2O2M3M2N3M3N2M2N3M3L4M2N3M4L3M3M4L3L4M4L3M3MgaV2"}, "label": "baby"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first sheep in the background facing away in the image\"."}], "task": "refering", "answer_template": "The \"the first sheep in the background facing away in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157], "bbox": [0.65153125, 0.22661214953271028, 0.899796875, 0.4454906542056075], "iscrowd": 0, "area": 10298.46335, "rle": {"size": [428, 640], "counts": "d^^52V=4L4L4M3L5K4M3M3M3M4L3M3M3M3O2N1O1O1O101N1O1O1O1O1O10O01O1O1O10O01O1O100O1O00100[EjMS:W2iEmMV:T2eEQNZ:o1bEWN]:\\20000000001O00000000001O000000O1000000000000000000000000000001O0000000000000000000O100O100O2O000O100O101N100O2O0O2O0O2O00001N101O001N101O000O2O00000O2O00000O10001O0O10000O101O1N5L5K4K5L4L5J5L4L1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1NYPk0"}, "label": "the first sheep in the background facing away in the image"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheeps butt\"."}], "task": "refering", "answer_template": "The \"a sheeps butt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157], "bbox": [0.65153125, 0.22661214953271028, 0.899796875, 0.4454906542056075], "iscrowd": 0, "area": 10298.46335, "rle": {"size": [428, 640], "counts": "d^^52V=4L4L4M3L5K4M3M3M3M4L3M3M3M3O2N1O1O1O101N1O1O1O1O1O10O01O1O1O10O01O1O100O1O00100[EjMS:W2iEmMV:T2eEQNZ:o1bEWN]:\\20000000001O00000000001O000000O1000000000000000000000000000001O0000000000000000000O100O100O2O000O100O101N100O2O0O2O0O2O00001N101O001N101O000O2O00000O2O00000O10001O0O10000O101O1N5L5K4K5L4L5J5L4L1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1NYPk0"}, "label": "a sheeps butt"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep in the bottom right looking at the camera\"."}], "task": "refering", "answer_template": "The \"the sheep in the bottom right looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.43656249999999996, 0.34186915887850466, 0.9974374999999999, 0.9889953271028038], "iscrowd": 0, "area": 74230.4954, "rle": {"size": [428, 640], "counts": "hVe35V=5K3N1N2O001N2O1O001O1O1O001O1O1O001O1O000001O0001O000001O01O000001O01O000001O0002N5K5K5L4K6J5K5L4K5K5K5L4K5K5K5L4K5K`0RGUL[7T4WHSLg7T4nGRLP8d3VHbLh7`3QHgLl7^4O2N2N2N2M3N2N2N2N1O2M3N2N2N2N2N2M3N1O2N2N2N2M3N2N2N2N1O2M3O1N2N2N2N10O0000001O000000001O000000001O000000001O000000001O000000001O00O100O1O100O1O1O100O1O1O100O1O1O1000000000000O100000000000000O10000000000000000O10000000000001eH]Jk6c5RI`Jn6a5nHbJR7_5jHdJV7i50000000000000000000000000001O0000000000000000000000000000000000000000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O10000O100O10000O010O100O100O100O100O100O100O100000000O1000000O1000000O100000000Uf0"}, "label": "the sheep in the bottom right looking at the camera"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closets sheep in the front\"."}], "task": "refering", "answer_template": "The \"the closets sheep in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.43656249999999996, 0.34186915887850466, 0.9974374999999999, 0.9889953271028038], "iscrowd": 0, "area": 74230.4954, "rle": {"size": [428, 640], "counts": "hVe35V=5K3N1N2O001N2O1O001O1O1O001O1O1O001O1O000001O0001O000001O01O000001O01O000001O0002N5K5K5L4K6J5K5L4K5K5K5L4K5K5K5L4K5K`0RGUL[7T4WHSLg7T4nGRLP8d3VHbLh7`3QHgLl7^4O2N2N2N2M3N2N2N2N1O2M3N2N2N2N2N2M3N1O2N2N2N2M3N2N2N2N1O2M3O1N2N2N2N10O0000001O000000001O000000001O000000001O000000001O000000001O00O100O1O100O1O1O100O1O1O100O1O1O1000000000000O100000000000000O10000000000000000O10000000000001eH]Jk6c5RI`Jn6a5nHbJR7_5jHdJV7i50000000000000000000000000001O0000000000000000000000000000000000000000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O10000O100O10000O010O100O100O100O100O100O100O100000000O1000000O1000000O100000000Uf0"}, "label": "the closets sheep in the front"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheared sheep behind three other sheep\"."}], "task": "refering", "answer_template": "The \"a sheared sheep behind three other sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137], "bbox": [0.694640625, 0.12689252336448598, 0.9923437500000001, 0.38495327102803734], "iscrowd": 0, "area": 9483.890899999999, "rle": {"size": [428, 640], "counts": "TSj57l<9G80000O10001O0O10000O101O0O10001O1N2O1O001000000000O100000O10O10000000000O10O100000001O00001N10001O0000001O000O2O00001O00001O0O101O00001O00001O0O101O00001O00000O2O01O00000001O000001O00000001O000001O00000001O2N4L3N2M3M3M4L3M3M2O0O2N1O1O2N1N3N1O2M2O1O2M201N100O2O00N101N101O1N101N101O1N101O0O2O1N101O6I8I7H7J7I4K3N1N2O2N1N2O1N3N1O1N3N0O01O10O01O1OZOf010O000001O01O0001O01O2O1N3M2N[o1"}, "label": "a sheared sheep behind three other sheep"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fatter sheep on left\"."}], "task": "refering", "answer_template": "The \"fatter sheep on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332], "bbox": [0.14750000000000002, 0.2866355140186916, 0.526984375, 0.9877570093457944], "iscrowd": 0, "area": 55468.34129999999, "rle": {"size": [428, 640], "counts": "[kW11[=1hG5m2LbLi0Y3XOUL_1e3aNkKU2o3lMeKe2U4\\MfKj2X4WMbKn2]4TM^KQ3a4oL[KV3d4kLVK[3i4fLRK^3n4cLmJb3R5_LlJc3S5]LkJe3U5\\LiJf3V5ZLhJi3W5XLgJi3Y5WLeJl3Z5ULdJm3[5SLcJo3]5QLbJQ4]5PL`JS4_5mK^JV4b5kKZJY4e5gKWJ^4h5cKTJ`4l5`KQJd4n5]KnIg4Q6ZKlIi4S6XKiIl4V6UKgIo4W6RKeIR5Z6oJcIU5[6kJcIX5\\6m0O2N1O2N1O1O2N1O2N1O2N3M2N2M3N2N1N2O1N2O1O1N3M2N2O1N2N2N2O1N1O2N100O1O1O100O1O10000O1000000O101O000O1000000O1000002M3N3M1O0O1O1O1O1O100000000000000001N1O1O100O1O1O1L4K6K400000001O00000000001O000000001O0001O000001O0001O1O100O2N1O1O5K10O0000001O00001O0O2O00001O00001O00001O00001O000O2O001O00001O00001O0gHdKb5]4\\JeKc5\\4\\JdKd5\\4\\JeKc5\\4[JfKd5[4[JfKd5[4ZJfKf5Z4ZJgKe5Z4ZJgKe5Z4YJhKf5Y4YJgKg5Z4XJgKg5Y4XJiKg5X4XJhKi5X4VJiKi5X4VJhKj5X4VJhKj5Y4UJhKi5Z4VJfKh5]4XJbKf5a4YJ_Ke5c4[J]Kc5f4\\J[Ka5h4^JXK`5k4_JUK^5n4cJQK[5R5dJnJZ5U5eJlJW5W5iJiJU5Z5jJfJT5]5kJcJS5_5nJ`JP5c5oJ]JU1]Ol1X6oL[JP1Ii1n5VMZJl02f1g5]MWJi0<b1_5eMUJe0f0^1X5mMQJ`0P1]1Q5SNoI<X1\\1k4WNmI9_1Z1h4\\NiI5h1Y1a4bNgI1P2W1\\4gNdINX2U1V4nNaIHb2T1P4SOSMh0o2WOUMc0m2]OXM<k2D[M4f2LbMJa25[510O1O0010OM4MR_n3"}, "label": "fatter sheep on left"}]}
{"id": 337509, "image": "COCO_train2014_000000337509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep to the far left\"."}], "task": "refering", "answer_template": "The \"sheep to the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332], "bbox": [0.14750000000000002, 0.2866355140186916, 0.526984375, 0.9877570093457944], "iscrowd": 0, "area": 55468.34129999999, "rle": {"size": [428, 640], "counts": "[kW11[=1hG5m2LbLi0Y3XOUL_1e3aNkKU2o3lMeKe2U4\\MfKj2X4WMbKn2]4TM^KQ3a4oL[KV3d4kLVK[3i4fLRK^3n4cLmJb3R5_LlJc3S5]LkJe3U5\\LiJf3V5ZLhJi3W5XLgJi3Y5WLeJl3Z5ULdJm3[5SLcJo3]5QLbJQ4]5PL`JS4_5mK^JV4b5kKZJY4e5gKWJ^4h5cKTJ`4l5`KQJd4n5]KnIg4Q6ZKlIi4S6XKiIl4V6UKgIo4W6RKeIR5Z6oJcIU5[6kJcIX5\\6m0O2N1O2N1O1O2N1O2N1O2N3M2N2M3N2N1N2O1N2O1O1N3M2N2O1N2N2N2O1N1O2N100O1O1O100O1O10000O1000000O101O000O1000000O1000002M3N3M1O0O1O1O1O1O100000000000000001N1O1O100O1O1O1L4K6K400000001O00000000001O000000001O0001O000001O0001O1O100O2N1O1O5K10O0000001O00001O0O2O00001O00001O00001O00001O000O2O001O00001O00001O0gHdKb5]4\\JeKc5\\4\\JdKd5\\4\\JeKc5\\4[JfKd5[4[JfKd5[4ZJfKf5Z4ZJgKe5Z4ZJgKe5Z4YJhKf5Y4YJgKg5Z4XJgKg5Y4XJiKg5X4XJhKi5X4VJiKi5X4VJhKj5X4VJhKj5Y4UJhKi5Z4VJfKh5]4XJbKf5a4YJ_Ke5c4[J]Kc5f4\\J[Ka5h4^JXK`5k4_JUK^5n4cJQK[5R5dJnJZ5U5eJlJW5W5iJiJU5Z5jJfJT5]5kJcJS5_5nJ`JP5c5oJ]JU1]Ol1X6oL[JP1Ii1n5VMZJl02f1g5]MWJi0<b1_5eMUJe0f0^1X5mMQJ`0P1]1Q5SNoI<X1\\1k4WNmI9_1Z1h4\\NiI5h1Y1a4bNgI1P2W1\\4gNdINX2U1V4nNaIHb2T1P4SOSMh0o2WOUMc0m2]OXM<k2D[M4f2LbMJa25[510O1O0010OM4MR_n3"}, "label": "sheep to the far left"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow cushioned chair\"."}], "task": "refering", "answer_template": "The \"a yellow cushioned chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.77696875, 0.08314583333333332, 0.9960625000000001, 0.7101041666666668], "iscrowd": 0, "area": 32569.38969999999, "rle": {"size": [480, 640], "counts": "TWY71n>1O1N2O1cL]3I7N2O1O100O100O100O010O100N2N2M3M3M3M3M3M3L4M3VGWKZ7`6F:F:F;E5K4L5K4L4L3N2O1O00100O1O1O100O1O1O0010O0001O01O0K5K5L4010O1O1O100O1O1O100O1O1O1O100O1O1O1000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000b0^OV1jNU1kNU1kNU1kNU1kNU1kNWT1"}, "label": "a yellow cushioned chair"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow glider chair\"."}], "task": "refering", "answer_template": "The \"a yellow glider chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.77696875, 0.08314583333333332, 0.9960625000000001, 0.7101041666666668], "iscrowd": 0, "area": 32569.38969999999, "rle": {"size": [480, 640], "counts": "TWY71n>1O1N2O1cL]3I7N2O1O100O100O100O010O100N2N2M3M3M3M3M3M3L4M3VGWKZ7`6F:F:F;E5K4L5K4L4L3N2O1O00100O1O1O100O1O1O0010O0001O01O0K5K5L4010O1O1O100O1O1O100O1O1O1O100O1O1O1000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000b0^OV1jNU1kNU1kNU1kNU1kNU1kNWT1"}, "label": "a yellow glider chair"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a naked child\"."}], "task": "refering", "answer_template": "The \"a naked child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 244, 264, 265, 266, 267, 269, 288, 289, 290, 291, 292, 293, 314, 315, 316], "bbox": [0.46484375, 0.37616666666666665, 0.7707656249999999, 0.7990833333333334], "iscrowd": 0, "area": 12237.889099999997, "rle": {"size": [480, 640], "counts": "Yo[43k>2N3M3L4N2O10O1000O1000O0100]M9lEGn9?RFAi9e0UF[Of9k0ZFUOd9n0[FROb9Q1^FoN_9U1_FlN_9V1aFiN]9[1bFeN[9_1dFaNZ9a1eF`NZ9b1eF^NZ9c1fF]NX9f1fFZNZ9g1fFYNX9j1gFVNX9k1hFUNW9m1gFTNW9o1hFPNX9Q2hFoMW9S2hFmMV9U2iFlMV9V2iFjMV9W2jFhMU9[2jFeMU9\\2jFeMU9]2jFcMU9_2jFaMU9`2jF`MV9b2iF^MV9c2jF]MU9e2jF[MV9e2jF[MU9g2kFXMT9h2nFVMR9k2oFTMP9l2RGSMm8n2TGQMk8o2WGoLi8Q3YGnLf8S3[GkLe8U3]GiLc8W3_GhL`8Y3_GgLa8Y3_GhL`8X3`GhLa8X3^GhLb8X3^GiLa8W3_GiLa8X3_GgLb8X3^GiLa8W3_GiLa8X3^GhLc8W3TGdLTO6g9W3TGRMl8n2UGQMl8n2UGoLm8R3RGmLo8S3RGjLQ9U3oFjLR9W3nFhLR9X3nFhLS9W3nFhLR9Y3nFeLT9Z3lFfLT9Z3lFfLU9Z3iFgLX9Y3fFgL[9[3bFfL_9\\3]FeLc9]3ZFdLg9]3WFcLi9`3mEgLT:Z3cEnL\\:l30O6K4L3L4M2M4M4L5L4aL[EX2h;O001O2N2N2O0O2_OUChNm<T1WCjNk<Q1YCnNi<n0ZCQOh<j0]CUOd<g0_CXOb<e0aCZOa<a0cC^O^<?fC_O\\<=gCBZ<:jCEX<7kCIU<3`m9OjSF8H8H9G5K6J5K6J5M32O1L4M2N2M3N2M3N2N2M3N2N2M3NO0100O010O10O01O010O10O010O0100O010O1O0010O01O1O100O1O2N100OVbT2"}, "label": "a naked child"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child in her mothers lap\"."}], "task": "refering", "answer_template": "The \"a child in her mothers lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 244, 264, 265, 266, 267, 269, 288, 289, 290, 291, 292, 293, 314, 315, 316], "bbox": [0.46484375, 0.37616666666666665, 0.7707656249999999, 0.7990833333333334], "iscrowd": 0, "area": 12237.889099999997, "rle": {"size": [480, 640], "counts": "Yo[43k>2N3M3L4N2O10O1000O1000O0100]M9lEGn9?RFAi9e0UF[Of9k0ZFUOd9n0[FROb9Q1^FoN_9U1_FlN_9V1aFiN]9[1bFeN[9_1dFaNZ9a1eF`NZ9b1eF^NZ9c1fF]NX9f1fFZNZ9g1fFYNX9j1gFVNX9k1hFUNW9m1gFTNW9o1hFPNX9Q2hFoMW9S2hFmMV9U2iFlMV9V2iFjMV9W2jFhMU9[2jFeMU9\\2jFeMU9]2jFcMU9_2jFaMU9`2jF`MV9b2iF^MV9c2jF]MU9e2jF[MV9e2jF[MU9g2kFXMT9h2nFVMR9k2oFTMP9l2RGSMm8n2TGQMk8o2WGoLi8Q3YGnLf8S3[GkLe8U3]GiLc8W3_GhL`8Y3_GgLa8Y3_GhL`8X3`GhLa8X3^GhLb8X3^GiLa8W3_GiLa8X3_GgLb8X3^GiLa8W3_GiLa8X3^GhLc8W3TGdLTO6g9W3TGRMl8n2UGQMl8n2UGoLm8R3RGmLo8S3RGjLQ9U3oFjLR9W3nFhLR9X3nFhLS9W3nFhLR9Y3nFeLT9Z3lFfLT9Z3lFfLU9Z3iFgLX9Y3fFgL[9[3bFfL_9\\3]FeLc9]3ZFdLg9]3WFcLi9`3mEgLT:Z3cEnL\\:l30O6K4L3L4M2M4M4L5L4aL[EX2h;O001O2N2N2O0O2_OUChNm<T1WCjNk<Q1YCnNi<n0ZCQOh<j0]CUOd<g0_CXOb<e0aCZOa<a0cC^O^<?fC_O\\<=gCBZ<:jCEX<7kCIU<3`m9OjSF8H8H9G5K6J5K6J5M32O1L4M2N2M3N2M3N2N2M3N2N2M3NO0100O010O10O01O010O10O010O0100O010O1O0010O01O1O100O1O2N100OVbT2"}, "label": "a child in her mothers lap"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark haired lady clipping a little girl ' s toenails\"."}], "task": "refering", "answer_template": "The \"a dark haired lady clipping a little girl ' s toenails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 175, 176, 193, 194, 195, 198, 199, 216, 217, 218, 219, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 316, 317, 318, 319], "bbox": [0.34381249999999997, 0.15731250000000002, 0.913484375, 0.8], "iscrowd": 0, "area": 46659.36085, "rle": {"size": [480, 640], "counts": "T^W35b><E:F;E;M2N3M2N3L4M2N3M3M2N3M2M4M3M2N3M00000O1000000000000O1000000000000O10000[NaMiF`2Q9\\NXFd1c9o1K5J6K5K5K5J6K5K5K5J6K5K5M3O1O100O1O1O1O100O2N1O100O1O1O100O1O1O1O10000O^MeHaM[7Z2kHeMU7U2RIjMm6U2UIkMk6U2VIjMj6W2VIhMi6Y2WIgMi6Z2WIeMi6[2WIeMh6]2XIcMg6^2YIaMg6_2YIaMf6a2ZI^Mf6c2ZI\\Me6f2ZIZMf6g2ZIXMf6ETJLUO?f6FWJITO`0e6FYJJRO?e6F\\JJnN`0e6F_JJlN?e6FbJJiN?e6FdJJgN`0d6FhJIdN`0d6FjJJbN?d6GkJJ`N?d6GoJI]N?d6GQKI[N`0d6FTKIXN`0c6GWKIVN?c6GZKIRN`0d6F\\KJPN?c6G`KImM`0b6GbKHlMa0a6GfKGiMa0a6HgKGhM`0a6IiKFeMa0a6KjKDeM`0a6LkKCdMa0a6LmKBbMa0`6NoKAaM`0`6OQL@^Ma0a6ORL@]M`0`62TL]O\\M`0_64VL\\OZM`0]67[LYOWM?[6;_LVOVM>X6?dLSORM>W6b0hLQOPM<U6g0lLmNnL;S6k0PMkNlL9Q6o0SMjNkL6o5S1WMgNjL5l5W1ZMfNiL2j5\\1]McNgL1i5_1`MbNfLOf5b1eM_NeLNe5d1fM^NfLMd5f1eM^NgLKc5h1gM\\NgLKb5i1gM\\NhLJa5j1hM\\NgLI`5m1hMZNhLI`5m1iMZNfLIa5m1iMZNgLH_5o1jMYNgLH_5o1kMoMcLTO2n0`5P2jMnMfLRO0P1_5Q2lMmMgLoNOR1^5R2lMmMjLmNKT1_5R2mMmMmL1U5T2mMkMnL0V5U2lMlMkL1Y5S2mMkMfL6\\5P2nMkMcL7_5o1nMiMcL8^5P2oMhMcL8^5P2PNgMbL9^5P2PNfMcL:\\5R2PNdMeL9[5S2QNbMeL:[5U2oMaMfL:Z5W2PN]MgL<Y5X2oM\\MhL<Y5Y2oMYMiL>W5[2oMWMkL=V5]2PNSMkL`0U5^2TNmLgLe0T5`2YNeLdLk0S5a2\\N`LaLn0T5c2_NYL^LT1S5d2oN[MQ1f2nNZMS1g2kNYMU1i2jNVMV1l2hNTMY1m2fNRMZ1P3eNoL[1S3dNlL]1U3bNiL_1Z3^NfLb1\\3]NcLd1_3ZN`Lf1f3UNYLk1m3PNRLQ2S4cMRKmLj0`5Z4[MoJUMf0`5a4TMkJ]Mb0`5h4lLhJdM>a5P5cLeJmM9`5W5]LbJSN6a5]5UL_JZN3a5c5nK]JbNN`5j5hKZJhNKa5P6`KWJPOG`5Y2XJV1Q1mLWOCa5Y2XJ[1j0kL]O_Oc5Y2XJb1a0hLEZOd5[2XJf18hLLUOg5Z2WJn1OeL4oNi5]2UJS2HcLa6W1iI\\3W6cLkI[3V6cLlI]3U6aLlI_3T6_LoI`3R6^LoIb3Q6\\LQJd3o5[LRJe3o5XLTJg3l5XLUJf3n5WLTJh3m5WLTJh3m5VLVJg3m5WLTJh3m5oK\\Jo3g5gKfJT4[5cKRKX4o4_K^K[4e4[KhK`4a7J6K5K5J6K5J6F;E:F:F:F:F:J6M3M00N2N2N1O2M3N2N2N2N0O10O1000O100000000O010000000O1000000O010hCUN[;k1bDXN^;h1^D\\Nb;d1ZD`Ne;a1WDcNi;]1SDgNm;Y1oCkNQ<T1lCPOT<n1O010001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N1O00O1000000O1000000O1000000O1000000O1000000O1000000O10000000000001O001O00001N1O2N1N2O2N1N3K4M3L5K4M4K]mi0"}, "label": "a dark haired lady clipping a little girl ' s toenails"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman grooming a small girl\"."}], "task": "refering", "answer_template": "The \"a woman grooming a small girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 175, 176, 193, 194, 195, 198, 199, 216, 217, 218, 219, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 316, 317, 318, 319], "bbox": [0.34381249999999997, 0.15731250000000002, 0.913484375, 0.8], "iscrowd": 0, "area": 46659.36085, "rle": {"size": [480, 640], "counts": "T^W35b><E:F;E;M2N3M2N3L4M2N3M3M2N3M2M4M3M2N3M00000O1000000000000O1000000000000O10000[NaMiF`2Q9\\NXFd1c9o1K5J6K5K5K5J6K5K5K5J6K5K5M3O1O100O1O1O1O100O2N1O100O1O1O100O1O1O1O10000O^MeHaM[7Z2kHeMU7U2RIjMm6U2UIkMk6U2VIjMj6W2VIhMi6Y2WIgMi6Z2WIeMi6[2WIeMh6]2XIcMg6^2YIaMg6_2YIaMf6a2ZI^Mf6c2ZI\\Me6f2ZIZMf6g2ZIXMf6ETJLUO?f6FWJITO`0e6FYJJRO?e6F\\JJnN`0e6F_JJlN?e6FbJJiN?e6FdJJgN`0d6FhJIdN`0d6FjJJbN?d6GkJJ`N?d6GoJI]N?d6GQKI[N`0d6FTKIXN`0c6GWKIVN?c6GZKIRN`0d6F\\KJPN?c6G`KImM`0b6GbKHlMa0a6GfKGiMa0a6HgKGhM`0a6IiKFeMa0a6KjKDeM`0a6LkKCdMa0a6LmKBbMa0`6NoKAaM`0`6OQL@^Ma0a6ORL@]M`0`62TL]O\\M`0_64VL\\OZM`0]67[LYOWM?[6;_LVOVM>X6?dLSORM>W6b0hLQOPM<U6g0lLmNnL;S6k0PMkNlL9Q6o0SMjNkL6o5S1WMgNjL5l5W1ZMfNiL2j5\\1]McNgL1i5_1`MbNfLOf5b1eM_NeLNe5d1fM^NfLMd5f1eM^NgLKc5h1gM\\NgLKb5i1gM\\NhLJa5j1hM\\NgLI`5m1hMZNhLI`5m1iMZNfLIa5m1iMZNgLH_5o1jMYNgLH_5o1kMoMcLTO2n0`5P2jMnMfLRO0P1_5Q2lMmMgLoNOR1^5R2lMmMjLmNKT1_5R2mMmMmL1U5T2mMkMnL0V5U2lMlMkL1Y5S2mMkMfL6\\5P2nMkMcL7_5o1nMiMcL8^5P2oMhMcL8^5P2PNgMbL9^5P2PNfMcL:\\5R2PNdMeL9[5S2QNbMeL:[5U2oMaMfL:Z5W2PN]MgL<Y5X2oM\\MhL<Y5Y2oMYMiL>W5[2oMWMkL=V5]2PNSMkL`0U5^2TNmLgLe0T5`2YNeLdLk0S5a2\\N`LaLn0T5c2_NYL^LT1S5d2oN[MQ1f2nNZMS1g2kNYMU1i2jNVMV1l2hNTMY1m2fNRMZ1P3eNoL[1S3dNlL]1U3bNiL_1Z3^NfLb1\\3]NcLd1_3ZN`Lf1f3UNYLk1m3PNRLQ2S4cMRKmLj0`5Z4[MoJUMf0`5a4TMkJ]Mb0`5h4lLhJdM>a5P5cLeJmM9`5W5]LbJSN6a5]5UL_JZN3a5c5nK]JbNN`5j5hKZJhNKa5P6`KWJPOG`5Y2XJV1Q1mLWOCa5Y2XJ[1j0kL]O_Oc5Y2XJb1a0hLEZOd5[2XJf18hLLUOg5Z2WJn1OeL4oNi5]2UJS2HcLa6W1iI\\3W6cLkI[3V6cLlI]3U6aLlI_3T6_LoI`3R6^LoIb3Q6\\LQJd3o5[LRJe3o5XLTJg3l5XLUJf3n5WLTJh3m5WLTJh3m5VLVJg3m5WLTJh3m5oK\\Jo3g5gKfJT4[5cKRKX4o4_K^K[4e4[KhK`4a7J6K5K5J6K5J6F;E:F:F:F:F:J6M3M00N2N2N1O2M3N2N2N2N0O10O1000O100000000O010000000O1000000O010hCUN[;k1bDXN^;h1^D\\Nb;d1ZD`Ne;a1WDcNi;]1SDgNm;Y1oCkNQ<T1lCPOT<n1O010001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N1O00O1000000O1000000O1000000O1000000O1000000O1000000O10000000000001O001O00001N1O2N1N2O2N1N3K4M3L5K4M4K]mi0"}, "label": "a woman grooming a small girl"}]}
{"id": 427628, "image": "COCO_train2014_000000427628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leg of a wooden chair near a pink and white bathrobe\"."}], "task": "refering", "answer_template": "The \"the leg of a wooden chair near a pink and white bathrobe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 210, 211, 233, 234, 256, 257, 258, 279, 280, 281, 304, 305, 328], "bbox": [0.136984375, 0.48814583333333333, 0.2923125, 0.8422916666666667], "iscrowd": 0, "area": 4828.0098, "rle": {"size": [480, 640], "counts": "\\_Y13n>:Gi0WOi0VOi0XO<CN3N2N1O2N1O2N2N1N3N2N1O2N1O2N2N1N3N1O2N2N1O2N1O2M3N1O2N1O2N2O010O10O01O010O10O010O0100O0010O0100O010O010O1O010O10O010O0100O0010O0100O010O010O1O010O1O0O2M2O2M3N1N3M2O2M3N1N3N1N3N2M2O2M2Ognc6"}, "label": "the leg of a wooden chair near a pink and white bathrobe"}]}
{"id": 124532, "image": "COCO_train2014_000000124532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"aeroplane\"."}], "task": "refering", "answer_template": "The \"aeroplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 110, 111, 112, 125, 132, 133, 134, 135, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 213, 214, 215, 216, 217, 218, 219], "bbox": [0.014984375, 0.21284037558685445, 0.975953125, 0.6655399061032864], "iscrowd": 0, "area": 39809.8435, "rle": {"size": [426, 640], "counts": "Y\\41X=2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2O0O2N2N101N101N101N2N101N101N2O0O2O0O2O0O4M6I3N0O2UEoM[:Q2aESN^:n1aESN_:m1`ESN`:n1`ERN`:P2]ERNb:o1\\ESNc:\\2N000000YO^EgNa:Y1bEdN^:\\1eEaN[:^1hE`NX:`1iE_NV:b1kEZNX:f1iEVNZ:j1gERN\\:n1`000000000000000000000000O100000000000000000O10000000000000O1000000000000000000000000000000O100000O10000000000000000000000000O1000000000000000000000000000O10O100000000000000000000000000000000O101O001O001O001O001O00O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000O1O1O1O1O1O10000000000000000001O2N2N2N1O2N2N2N2N1O2N2N000000000000O10000000000000000000N2N2O1N2N101N2N2O1N2O1N2N2O1N2N2O_MUFe1i9YN[Fg1c9WNaFh1`9UNcFk1]9QNgFo1Y9nMjFR2U9kMoFU2Q9gMSGY2m8dMVG\\2l900000000000000O1000000001O1O001O1O001O1O00001O00001O00O100O10000O1000000C[EUNe:k1[EUNe:j1\\EVNd:i1]EWNc:_1[E^N33a:_1^E[N26`:_1_EZN17`:`1`EVN1:_:`1iE_NW:a1iE_NW:a1iE_NW:a1jE^NV:b1jE[NY:f1fEVN^:j1a00000001O0000000O101O00000000001O00000000001O0000000000LkDTNT;l15000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000O100000000O100000000O1000000O100000000O100000O001O1O1O001O1O1O001O1O1O001O1O1O0O2O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O00WOZFPNe9o1^FPNa9n1bFRN]9m1eFRN[9l1hFTNW9j1lFVNS9i1PGUNo8j1TGVNk8i1WGWNh8g1[GYNd8f1^GYNb8e1bGZN]8d1fG\\NY8c1iG]NV8a1mG^NS8a1oG_NP8_1TH`Nl7^1VHbNj7[1YHdNh7Y1[HgNe7W1]HiNc7T1aHkN_7S1cHlN^7Q1eHhNb7V1`HjN`7S1cHmN]7P1fHPOZ7n0hHROX7k0jHVOV7h0lHXOT7e0oHZOR7c0QI]Oo6a0SI_Om6>VIBj6<XIDh69[IGe67]IIc64`IL`61cIN_6ObI2^6KeI5h900000000O10000000O1000O1000000000000O10O1000000000O1000000000O10O1000j^6"}, "label": "aeroplane"}]}
{"id": 124532, "image": "COCO_train2014_000000124532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the delta airplane readying to take off\"."}], "task": "refering", "answer_template": "The \"the delta airplane readying to take off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 110, 111, 112, 125, 132, 133, 134, 135, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 213, 214, 215, 216, 217, 218, 219], "bbox": [0.014984375, 0.21284037558685445, 0.975953125, 0.6655399061032864], "iscrowd": 0, "area": 39809.8435, "rle": {"size": [426, 640], "counts": "Y\\41X=2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2O0O2N2N101N101N101N2N101N101N2O0O2O0O2O0O4M6I3N0O2UEoM[:Q2aESN^:n1aESN_:m1`ESN`:n1`ERN`:P2]ERNb:o1\\ESNc:\\2N000000YO^EgNa:Y1bEdN^:\\1eEaN[:^1hE`NX:`1iE_NV:b1kEZNX:f1iEVNZ:j1gERN\\:n1`000000000000000000000000O100000000000000000O10000000000000O1000000000000000000000000000000O100000O10000000000000000000000000O1000000000000000000000000000O10O100000000000000000000000000000000O101O001O001O001O001O00O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000O1O1O1O1O1O10000000000000000001O2N2N2N1O2N2N2N2N1O2N2N000000000000O10000000000000000000N2N2O1N2N101N2N2O1N2O1N2N2O1N2N2O_MUFe1i9YN[Fg1c9WNaFh1`9UNcFk1]9QNgFo1Y9nMjFR2U9kMoFU2Q9gMSGY2m8dMVG\\2l900000000000000O1000000001O1O001O1O001O1O00001O00001O00O100O10000O1000000C[EUNe:k1[EUNe:j1\\EVNd:i1]EWNc:_1[E^N33a:_1^E[N26`:_1_EZN17`:`1`EVN1:_:`1iE_NW:a1iE_NW:a1iE_NW:a1jE^NV:b1jE[NY:f1fEVN^:j1a00000001O0000000O101O00000000001O00000000001O0000000000LkDTNT;l15000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000O100000000O100000000O1000000O100000000O100000O001O1O1O001O1O1O001O1O1O001O1O1O0O2O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O00WOZFPNe9o1^FPNa9n1bFRN]9m1eFRN[9l1hFTNW9j1lFVNS9i1PGUNo8j1TGVNk8i1WGWNh8g1[GYNd8f1^GYNb8e1bGZN]8d1fG\\NY8c1iG]NV8a1mG^NS8a1oG_NP8_1TH`Nl7^1VHbNj7[1YHdNh7Y1[HgNe7W1]HiNc7T1aHkN_7S1cHlN^7Q1eHhNb7V1`HjN`7S1cHmN]7P1fHPOZ7n0hHROX7k0jHVOV7h0lHXOT7e0oHZOR7c0QI]Oo6a0SI_Om6>VIBj6<XIDh69[IGe67]IIc64`IL`61cIN_6ObI2^6KeI5h900000000O10000000O1000O1000000000000O10O1000000000O1000000000O10O1000j^6"}, "label": "the delta airplane readying to take off"}]}
{"id": 42614, "image": "COCO_train2014_000000042614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult elephant\"."}], "task": "refering", "answer_template": "The \"an adult elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 291, 292, 294, 295, 296, 310, 311, 312, 314, 315, 316, 317, 318], "bbox": [0.45109374999999996, 0.4538302752293578, 0.8991250000000001, 0.8659862385321101], "iscrowd": 0, "area": 31460.453850000005, "rle": {"size": [436, 640], "counts": "jZk32`=3M3M2M4M3M2N3L?B?A?A>B3L4M3M3M2M4M3M3M3L4L4L4L4LSO_FmM^9P2kFnMQ9n1ZGoMb8n1gGPNU8l1UHRNg7j1cHTNY7i1QITNn6g1ZIWNe6i1^IUNa6k1bIRN]6o1fIoMY6Q2kIlMT6T2oIjMP6V2f210000O1000000O10O1000O10000O10001O0O10001O0O1`GcMi5]2UJeMk5\\2RJfMm5[2PJhMP6X2nIjMR6W2jIlMV6T2hInMW6S2fIPNZ6P2dIRN\\6o1_IUNa6k1iHjNW7W1TH^Ol7b0PHBP8?lGDS8=iGGW8k20000O10O1000O10000O1000D<L4M3L4L4L4M3L4M3M3M4L3M3M3M3M3M3M3M3M3M3O11O010O001O1O1O1O1O1O1O1O1O1O1O101N1O1O1O2N3M2N3M2N3M2N3M2N3N1N3M1O2N2N2N2N1O2N2N2N2N1O]NZGZNe8e1]G[Na8e1`G\\N^8d1dG\\NZ8d1hG[NW8e1kG[NR8f1oG[No7e1SHZNl7f1VHXNj7h1XHVNh7j1ZHTNe7m1\\HRNd7n1^HPNb7o1aHoM_7Q2cHmM]7S2dHlM\\7T2fHjMY7W2iHgMW7Y2kHeMU7[2lHdMT7\\2nHbMR7^2PI`Mo6a2SI]Mm6c2TI\\Ml6d2VIZMj6f2XIXMh6h2ZIVMe6k2\\ITMd6l2^IRMb6n2`IPM`6P3bInL^6R3cImL\\6T3fIjLZ6V3hIhLX6X3jIgLU6Y3mIeLS6[3nIdLQ6]3S20000000O100000000O2O0000001O01O01O00100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100_GnK^7S4_HoKa7R4[HQLf7o3WHSLi7n3SHULm7l3PHVLQ8j3kGYLU8g3iG\\LV8e3fG^LZ8c3cG_L^8[400001O00001N100O2N100O1aNfGdM[8j1VHUNk7c1]H]Nc7[1fHcN]7S1kHmNX7h0PIWOT7>TIBP72YILk6OWI1l6MSI2Q7NmH1V70fHO^71_HOd71ZHMk73QHMR83kGH]89RGXOmN7[:a0bF\\OROLf:i0RF^O]:b0^EBd:>UEFm::nDIT;7gDL[;5^DOg;NTD3S<c06J7I7I7I7I7Ib[k0"}, "label": "an adult elephant"}]}
{"id": 42614, "image": "COCO_train2014_000000042614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant is walking near desert area\"."}], "task": "refering", "answer_template": "The \"the elephant is walking near desert area\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 291, 292, 294, 295, 296, 310, 311, 312, 314, 315, 316, 317, 318], "bbox": [0.45109374999999996, 0.4538302752293578, 0.8991250000000001, 0.8659862385321101], "iscrowd": 0, "area": 31460.453850000005, "rle": {"size": [436, 640], "counts": "jZk32`=3M3M2M4M3M2N3L?B?A?A>B3L4M3M3M2M4M3M3M3L4L4L4L4LSO_FmM^9P2kFnMQ9n1ZGoMb8n1gGPNU8l1UHRNg7j1cHTNY7i1QITNn6g1ZIWNe6i1^IUNa6k1bIRN]6o1fIoMY6Q2kIlMT6T2oIjMP6V2f210000O1000000O10O1000O10000O10001O0O10001O0O1`GcMi5]2UJeMk5\\2RJfMm5[2PJhMP6X2nIjMR6W2jIlMV6T2hInMW6S2fIPNZ6P2dIRN\\6o1_IUNa6k1iHjNW7W1TH^Ol7b0PHBP8?lGDS8=iGGW8k20000O10O1000O10000O1000D<L4M3L4L4L4M3L4M3M3M4L3M3M3M3M3M3M3M3M3M3O11O010O001O1O1O1O1O1O1O1O1O1O1O101N1O1O1O2N3M2N3M2N3M2N3M2N3N1N3M1O2N2N2N2N1O2N2N2N2N1O]NZGZNe8e1]G[Na8e1`G\\N^8d1dG\\NZ8d1hG[NW8e1kG[NR8f1oG[No7e1SHZNl7f1VHXNj7h1XHVNh7j1ZHTNe7m1\\HRNd7n1^HPNb7o1aHoM_7Q2cHmM]7S2dHlM\\7T2fHjMY7W2iHgMW7Y2kHeMU7[2lHdMT7\\2nHbMR7^2PI`Mo6a2SI]Mm6c2TI\\Ml6d2VIZMj6f2XIXMh6h2ZIVMe6k2\\ITMd6l2^IRMb6n2`IPM`6P3bInL^6R3cImL\\6T3fIjLZ6V3hIhLX6X3jIgLU6Y3mIeLS6[3nIdLQ6]3S20000000O100000000O2O0000001O01O01O00100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100_GnK^7S4_HoKa7R4[HQLf7o3WHSLi7n3SHULm7l3PHVLQ8j3kGYLU8g3iG\\LV8e3fG^LZ8c3cG_L^8[400001O00001N100O2N100O1aNfGdM[8j1VHUNk7c1]H]Nc7[1fHcN]7S1kHmNX7h0PIWOT7>TIBP72YILk6OWI1l6MSI2Q7NmH1V70fHO^71_HOd71ZHMk73QHMR83kGH]89RGXOmN7[:a0bF\\OROLf:i0RF^O]:b0^EBd:>UEFm::nDIT;7gDL[;5^DOg;NTD3S<c06J7I7I7I7I7Ib[k0"}, "label": "the elephant is walking near desert area"}]}
{"id": 403064, "image": "COCO_train2014_000000403064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue toyota car driving\"."}], "task": "refering", "answer_template": "The \"a blue toyota car driving\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 289, 290], "bbox": [0.351609375, 0.4560625, 0.627421875, 0.7442291666666667], "iscrowd": 0, "area": 18804.31685, "rle": {"size": [480, 640], "counts": "dhY3]1U<`1L4K3N2N2M3N2N2M3N2N2M2O2N2M3N2N2M3N20000000000000O1000000E;D<00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N2N3M2N2N2N2M3N2N2N3M2N2N2N2N2N2N3M2N2>BN2M3N2N3L3N2N2N2M3N2M4D;E;D<E;EQY_3"}, "label": "a blue toyota car driving"}]}
{"id": 403064, "image": "COCO_train2014_000000403064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back end of a dark blue car driving down the road with license plate in view\"."}], "task": "refering", "answer_template": "The \"the back end of a dark blue car driving down the road with license plate in view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 289, 290], "bbox": [0.351609375, 0.4560625, 0.627421875, 0.7442291666666667], "iscrowd": 0, "area": 18804.31685, "rle": {"size": [480, 640], "counts": "dhY3]1U<`1L4K3N2N2M3N2N2M3N2N2M2O2N2M3N2N2M3N20000000000000O1000000E;D<00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N2N3M2N2N2N2M3N2N2N3M2N2N2N2N2N2N3M2N2>BN2M3N2N3L3N2N2N2M3N2M4D;E;D<E;EQY_3"}, "label": "the back end of a dark blue car driving down the road with license plate in view"}]}
{"id": 386684, "image": "COCO_train2014_000000386684.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of blue skies\"."}], "task": "refering", "answer_template": "The \"a pair of blue skies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [344, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 413, 423, 424, 425, 427, 428, 429, 433, 434, 435, 436, 447, 448, 454, 455, 456, 457, 458], "bbox": [0.411640625, 0.7325858951175407, 1.0, 0.9887703435804701], "iscrowd": 0, "area": 12726.785200000004, "rle": {"size": [553, 640], "counts": "XR_45Ta04L2N1O001O001O001O001O001O001O001O001O00001O0000001O000O101O00000O2O0000001N1000O1000O100000O10O100000O10O10000002M3N3M2N3M3M2M4M2N3MmUg00RjXO2N2O2M2N3M2O1N3M2N3M2O1N3N1O1N3N1O2N000O010000O01000O10O10O1000O010000O10O10O100O01000O10O10O1000O010000O01000O10O10O10O010000O01000O01000O10O10O10O10O1000O0100O01000O10O10O10O10O1000O01000[AROn<n0SCROm<n0RCSOn<m0RCSOn<l0SCTOl<m0SCTOm<k0TCUOl<k0SCVOl<k0TCUOl<j0TCWOl<i0TCWOl<i0SCXOl<h0UCXOk<h0TCYOl<g0TCYOk<g0UCZOk<f0UCZOk<f0TC[Ok<e0VC[Oj<e0UC\\Ok<d0UC\\Ok<c0UC^Oj<c0VC]Oj<b0VC_Oj<a0VC_Oi<b0WC^Oi<a0WC@i<`0WC@i<`0VCAi<?XCAh<?WCBi<>WCBh<>YCBg<>XCCh<=XCCh<<XCEg<<YCDg<<YCDg<;YCFf<;ZCEf<;ZCEf<:ZCGf<9ZCGe<9[CHe<8[CHe<8[CHd<8\\CId<7\\CId<7\\CId<6\\CKc<6]CJc<6\\CKd<4]CLb<5^CKb<4^CMb<3^CMb<3^CMa<3_CNa<2_CNa<2_CN`<2`CO`<1`CO`<1`CO_<1bCO^<1aC0_<ObC1]<0cC0]<0cC0]<OcC2\\<OdC1\\<OdC1\\<NeC2Z<OeC2[<NeC2[<MfC3Y<NgC2Y<MgC4Y<LgC4X<MhC3X<LiC4W<LhC5W<LiC3X<LiC4W<LiC4V<LjC5V<KjC4W<LiC4V<LkC4U<LjC4W<LiC4V<LkC4U<LkC3U<NjC3V<LkC4U<LkC3U<MlC3T<MkC4U<LkC3U<MlC3T<MlC3T<MkC3U<MlC3T<MlC3T<MlC2T<NlC3T<MlC3T<LmC3S<NmC2S<NlC3T<LmC3S<NmC2S<NmC2S<MmC3S<NmC2S<MnC3R<MnC2R<OmC2S<MnC3R<MnC2R<OnC1R<NnC3Q<NoC1R<OnC1R<NoC1Q<0nC1R<NoC2Q<NoC1Q<0oC0Q<OoC2Q<NoC1Q<0nC1R<NoC2Q<NoC1Q<0nC1R<NoC2Q<NnC2R<NoC2Q<NoC2P<OoC1R<NoC2P<OoC1R<OnC1Q<OPD1P<OoC1R<OnC1Q<OoC2Q<NoC1Q<OPD1P<OoC2P<OPD0Q<OoC2P<OPD1P<OPD0P<0PD1P<OPD1o;OQD1P<OPD1P<OoC2P<NQD1P<OPD1o;0PD0Q<OPD1o;0PD1P<OPD0P<0QD0R<MmC4X<FiC9]<BbC?c<\\Oj2JeB"}, "label": "a pair of blue skies"}]}
{"id": 386684, "image": "COCO_train2014_000000386684.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of blue skies\"."}], "task": "refering", "answer_template": "The \"a pair of blue skies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [344, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 413, 423, 424, 425, 427, 428, 429, 433, 434, 435, 436, 447, 448, 454, 455, 456, 457, 458], "bbox": [0.411640625, 0.7325858951175407, 1.0, 0.9887703435804701], "iscrowd": 0, "area": 12726.785200000004, "rle": {"size": [553, 640], "counts": "XR_45Ta04L2N1O001O001O001O001O001O001O001O001O00001O0000001O000O101O00000O2O0000001N1000O1000O100000O10O100000O10O10000002M3N3M2N3M3M2M4M2N3MmUg00RjXO2N2O2M2N3M2O1N3M2N3M2O1N3N1O1N3N1O2N000O010000O01000O10O10O1000O010000O10O10O100O01000O10O10O1000O010000O01000O10O10O10O010000O01000O01000O10O10O10O10O1000O0100O01000O10O10O10O10O1000O01000[AROn<n0SCROm<n0RCSOn<m0RCSOn<l0SCTOl<m0SCTOm<k0TCUOl<k0SCVOl<k0TCUOl<j0TCWOl<i0TCWOl<i0SCXOl<h0UCXOk<h0TCYOl<g0TCYOk<g0UCZOk<f0UCZOk<f0TC[Ok<e0VC[Oj<e0UC\\Ok<d0UC\\Ok<c0UC^Oj<c0VC]Oj<b0VC_Oj<a0VC_Oi<b0WC^Oi<a0WC@i<`0WC@i<`0VCAi<?XCAh<?WCBi<>WCBh<>YCBg<>XCCh<=XCCh<<XCEg<<YCDg<<YCDg<;YCFf<;ZCEf<;ZCEf<:ZCGf<9ZCGe<9[CHe<8[CHe<8[CHd<8\\CId<7\\CId<7\\CId<6\\CKc<6]CJc<6\\CKd<4]CLb<5^CKb<4^CMb<3^CMb<3^CMa<3_CNa<2_CNa<2_CN`<2`CO`<1`CO`<1`CO_<1bCO^<1aC0_<ObC1]<0cC0]<0cC0]<OcC2\\<OdC1\\<OdC1\\<NeC2Z<OeC2[<NeC2[<MfC3Y<NgC2Y<MgC4Y<LgC4X<MhC3X<LiC4W<LhC5W<LiC3X<LiC4W<LiC4V<LjC5V<KjC4W<LiC4V<LkC4U<LjC4W<LiC4V<LkC4U<LkC3U<NjC3V<LkC4U<LkC3U<MlC3T<MkC4U<LkC3U<MlC3T<MlC3T<MkC3U<MlC3T<MlC3T<MlC2T<NlC3T<MlC3T<LmC3S<NmC2S<NlC3T<LmC3S<NmC2S<NmC2S<MmC3S<NmC2S<MnC3R<MnC2R<OmC2S<MnC3R<MnC2R<OnC1R<NnC3Q<NoC1R<OnC1R<NoC1Q<0nC1R<NoC2Q<NoC1Q<0oC0Q<OoC2Q<NoC1Q<0nC1R<NoC2Q<NoC1Q<0nC1R<NoC2Q<NnC2R<NoC2Q<NoC2P<OoC1R<NoC2P<OoC1R<OnC1Q<OPD1P<OoC1R<OnC1Q<OoC2Q<NoC1Q<OPD1P<OoC2P<OPD0Q<OoC2P<OPD1P<OPD0P<0PD1P<OPD1o;OQD1P<OPD1P<OoC2P<NQD1P<OPD1o;0PD0Q<OPD1o;0PD1P<OPD0P<0QD0R<MmC4X<FiC9]<BbC?c<\\Oj2JeB"}, "label": "a pair of blue skies"}]}
{"id": 411261, "image": "COCO_train2014_000000411261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of sausage that is not under the english muffin\"."}], "task": "refering", "answer_template": "The \"the piece of sausage that is not under the english muffin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 164, 165, 166, 167, 168, 183, 184, 185, 186, 187, 188, 202, 203, 204, 205, 206, 221, 222, 223, 224], "bbox": [0.13588, 0.6494133333333333, 0.48706, 0.9623466666666667], "iscrowd": 0, "area": 10383.501900000001, "rle": {"size": [375, 500], "counts": "oTi01d;4M3M3L4M3M3M3L4M3N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N1N3N2N2N001O1O001000O01000O10O1000O01000O10O10O10O10O1000O10O1000O01000O010000O00100O010O1O10O0100O10O01O10O010O010O00010O0010O01O01O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O0O101O000O2O00001N10001O0O10001O0O10001O0O10001O0O10001O0O1000001N1000001N100O101N100O101O0O100O2O0O1N8E=C=Dm]m2"}, "label": "the piece of sausage that is not under the english muffin"}]}
{"id": 263810, "image": "COCO_train2014_000000263810.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman kneeling behind the stroller\"."}], "task": "refering", "answer_template": "The \"the woman kneeling behind the stroller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 102, 103, 104, 105, 116, 117, 118, 130, 131, 132, 133, 144, 145, 146, 147, 158, 159, 160, 161, 173, 174, 175], "bbox": [0.2674412532637076, 0.27421875, 0.5230026109660575, 0.566296875], "iscrowd": 0, "area": 10658.635650000002, "rle": {"size": [640, 383], "counts": "QlQ22nc01O1O001O1O1O1N2O1R]OFTb0;i]OIUb08i]OJVb07h]OLVb05g]ONXb0P1M2L4L3E<VOQNY_OW2]`0k0J7M1N2N2O2M2O1N2N2O2M2N2M3M3N101O00100000O10000000SAlKn=T4oAQLo=o3QBSLm=m3RBULm=k3SBWLk=i3TBYLk=g3UBZLj=f3VB[Li=e3WB\\Lh=d3WB^Lh=b3XB_Lg=a3YB_Lg=a3YB_Lg=a3XB`Lh=`3XB`Lh=`3XB`Lh=`3WBbLh=_3WBaLi=_3VBbLj=f1gAUO?UOj=b1kAZO9UOm=^1lA^O6TOn=\\1oAB0ROQ>[1PBGIPOX>W1PBl0P>ROSBn0m=POTBQ1k=nNVBT1h=kNZBU1h=gNYBX1k=dNVB[1n=aNTB]1o=`NRB_1Q>_NoAa1Q>^NQB`1Q>^NPBa1Q>^NPBb1P>]NRBa1P>\\NSBc1m=[NUBe1l=XNWBf1j=YNWBg1j=WNXBg1k=lMl@IZ1Z2k=jMo@IX1\\2j=hMSAHV1]2i=gMhBV2d?L3MY1gNWRb3"}, "label": "the woman kneeling behind the stroller"}]}
{"id": 263810, "image": "COCO_train2014_000000263810.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman blue is kneeling beside a baby in a stroller while smiling\"."}], "task": "refering", "answer_template": "The \"a woman blue is kneeling beside a baby in a stroller while smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 102, 103, 104, 105, 116, 117, 118, 130, 131, 132, 133, 144, 145, 146, 147, 158, 159, 160, 161, 173, 174, 175], "bbox": [0.2674412532637076, 0.27421875, 0.5230026109660575, 0.566296875], "iscrowd": 0, "area": 10658.635650000002, "rle": {"size": [640, 383], "counts": "QlQ22nc01O1O001O1O1O1N2O1R]OFTb0;i]OIUb08i]OJVb07h]OLVb05g]ONXb0P1M2L4L3E<VOQNY_OW2]`0k0J7M1N2N2O2M2O1N2N2O2M2N2M3M3N101O00100000O10000000SAlKn=T4oAQLo=o3QBSLm=m3RBULm=k3SBWLk=i3TBYLk=g3UBZLj=f3VB[Li=e3WB\\Lh=d3WB^Lh=b3XB_Lg=a3YB_Lg=a3YB_Lg=a3XB`Lh=`3XB`Lh=`3XB`Lh=`3WBbLh=_3WBaLi=_3VBbLj=f1gAUO?UOj=b1kAZO9UOm=^1lA^O6TOn=\\1oAB0ROQ>[1PBGIPOX>W1PBl0P>ROSBn0m=POTBQ1k=nNVBT1h=kNZBU1h=gNYBX1k=dNVB[1n=aNTB]1o=`NRB_1Q>_NoAa1Q>^NQB`1Q>^NPBa1Q>^NPBb1P>]NRBa1P>\\NSBc1m=[NUBe1l=XNWBf1j=YNWBg1j=WNXBg1k=lMl@IZ1Z2k=jMo@IX1\\2j=hMSAHV1]2i=gMhBV2d?L3MY1gNWRb3"}, "label": "a woman blue is kneeling beside a baby in a stroller while smiling"}]}
{"id": 263810, "image": "COCO_train2014_000000263810.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby smiling in a stroller\"."}], "task": "refering", "answer_template": "The \"baby smiling in a stroller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 141, 142, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 175, 176, 185, 186, 187, 188, 189, 199, 200, 201, 202], "bbox": [0.07733681462140993, 0.40975, 0.5936292428198434, 0.644140625], "iscrowd": 0, "area": 8973.380100000002, "rle": {"size": [640, 383], "counts": "jPc0>Zc09N3M2[]OVOQb0l0l]OWORb0k0k]OYOQb0j0l]OYOSb0[1N2N2O1N100O100O100000000]^ORNTa0n1j^OUNUa0k1i^OXNVa0h1h^O[NVa0Y2O00O101O00001O00001O0O101N2O1N2K5K5O1Bo]OkNRb0R1P^OmNSb0o0n]OROVb0d0P^O\\OSb01j]OK44eb0L\\]O3eb0H`]O6Uc0O2N3N3O0O01O1O010]]O2Ra0Nm^O8n`0IQ_O:l`0GR_O;m`0GS^OKd0`0Xa0Fo]OMi0?Wa0Ki^O6Va0Ji^O8Va0Hj^O9Ta0Hm^O8Ra0Hn^O8\\`0@V_O8`0:V`0_OY_O8a0<R`0[OR_OJ:?c0=P`0VOU_OO7?d0=P`0SOU_O26?f0<R`0WOW_O=g0>P`0UOY_O=f0`0P`0PO]_Oa0a0a0_`0@`_Oa0_`0A^_Oa0a`0@^_Oa0a`0A\\_O`0d`0AZ_Oa0e`0@Z_O`0f`0\\Ok^O\\O<Y1i`0ZOm^O^O7Y1l`0XOo^O@2Y1o`0VOQ_OCLX1Sa0SOT_OFFY1Va0nNW_O[1j`0bNY_O]1k`0^NW_Oa1da000001O00000001O01O00000000000000000000000001O0O10000O10000O10000O10000O101O0O100000001N1000000O101O00000O101O0O101N4M3M1N010O100O100O100O100O00100O1O010O10O01O1O0O10001O0Kk\\O@Uc0`0500001N1O2N1010O1O010N2O1O2M2O1N2O2M2O1NUUQ3"}, "label": "baby smiling in a stroller"}]}
{"id": 263810, "image": "COCO_train2014_000000263810.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adorable , goofy baby sitting in the stroller holding an apple\"."}], "task": "refering", "answer_template": "The \"the adorable , goofy baby sitting in the stroller holding an apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 141, 142, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 175, 176, 185, 186, 187, 188, 189, 199, 200, 201, 202], "bbox": [0.07733681462140993, 0.40975, 0.5936292428198434, 0.644140625], "iscrowd": 0, "area": 8973.380100000002, "rle": {"size": [640, 383], "counts": "jPc0>Zc09N3M2[]OVOQb0l0l]OWORb0k0k]OYOQb0j0l]OYOSb0[1N2N2O1N100O100O100000000]^ORNTa0n1j^OUNUa0k1i^OXNVa0h1h^O[NVa0Y2O00O101O00001O00001O0O101N2O1N2K5K5O1Bo]OkNRb0R1P^OmNSb0o0n]OROVb0d0P^O\\OSb01j]OK44eb0L\\]O3eb0H`]O6Uc0O2N3N3O0O01O1O010]]O2Ra0Nm^O8n`0IQ_O:l`0GR_O;m`0GS^OKd0`0Xa0Fo]OMi0?Wa0Ki^O6Va0Ji^O8Va0Hj^O9Ta0Hm^O8Ra0Hn^O8\\`0@V_O8`0:V`0_OY_O8a0<R`0[OR_OJ:?c0=P`0VOU_OO7?d0=P`0SOU_O26?f0<R`0WOW_O=g0>P`0UOY_O=f0`0P`0PO]_Oa0a0a0_`0@`_Oa0_`0A^_Oa0a`0@^_Oa0a`0A\\_O`0d`0AZ_Oa0e`0@Z_O`0f`0\\Ok^O\\O<Y1i`0ZOm^O^O7Y1l`0XOo^O@2Y1o`0VOQ_OCLX1Sa0SOT_OFFY1Va0nNW_O[1j`0bNY_O]1k`0^NW_Oa1da000001O00000001O01O00000000000000000000000001O0O10000O10000O10000O10000O101O0O100000001N1000000O101O00000O101O0O101N4M3M1N010O100O100O100O100O00100O1O010O10O01O1O0O10001O0Kk\\O@Uc0`0500001N1O2N1010O1O010N2O1O2M2O1N2O2M2O1NUUQ3"}, "label": "the adorable , goofy baby sitting in the stroller holding an apple"}]}
{"id": 50820, "image": "COCO_train2014_000000050820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a checkered , bright shirt playing with a kite\"."}], "task": "refering", "answer_template": "The \"a little boy in a checkered , bright shirt playing with a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 269, 270, 271, 292, 293, 294, 314, 315, 316, 317, 338, 339, 340], "bbox": [0.659328125, 0.6949414519906323, 0.83084375, 0.9998360655737705], "iscrowd": 0, "area": 6575.703949999999, "rle": {"size": [427, 640], "counts": "R[`52W=6K4M0O10000O10000O0O2O001O00100O1O1O10O01O1O1O1O001N101N2N101O001O00O1006Jb0_O4K;F1YOmMTFT2c9WNUFP2e9VNWFm1g9VNWFl1g9VNVFm1g9m0N100O1O100O10000O1000000O100000000001O00002N1VM\\Fk1e9SN\\Fm1e9RN[Fn1f9PN[FP2f9oMYFR2h9kMZFU2h9aMWFJ2e2]:00XO_EfNa:Y1aEfN_:Y1cEfN\\:W1iEhNW:g0[FXOe9a0cF^O\\9<lFCS9:VGAh8`0ZG_Oe8b0]G\\Oa8e0bGYO]8h0gGTOW8m0W1VOXEj0h:ZOSEh0m:ZOPEf0Q;\\OkDe0V;]OgDb0\\;^OaDb0b;a03M3N101N2N2K5L4L8H;E3M4LWg\\1"}, "label": "a little boy in a checkered , bright shirt playing with a kite"}]}
{"id": 50820, "image": "COCO_train2014_000000050820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the white plaid jacket\"."}], "task": "refering", "answer_template": "The \"the boy in the white plaid jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 269, 270, 271, 292, 293, 294, 314, 315, 316, 317, 338, 339, 340], "bbox": [0.659328125, 0.6949414519906323, 0.83084375, 0.9998360655737705], "iscrowd": 0, "area": 6575.703949999999, "rle": {"size": [427, 640], "counts": "R[`52W=6K4M0O10000O10000O0O2O001O00100O1O1O10O01O1O1O1O001N101N2N101O001O00O1006Jb0_O4K;F1YOmMTFT2c9WNUFP2e9VNWFm1g9VNWFl1g9VNVFm1g9m0N100O1O100O10000O1000000O100000000001O00002N1VM\\Fk1e9SN\\Fm1e9RN[Fn1f9PN[FP2f9oMYFR2h9kMZFU2h9aMWFJ2e2]:00XO_EfNa:Y1aEfN_:Y1cEfN\\:W1iEhNW:g0[FXOe9a0cF^O\\9<lFCS9:VGAh8`0ZG_Oe8b0]G\\Oa8e0bGYO]8h0gGTOW8m0W1VOXEj0h:ZOSEh0m:ZOPEf0Q;\\OkDe0V;]OgDb0\\;^OaDb0b;a03M3N101N2N2K5L4L8H;E3M4LWg\\1"}, "label": "the boy in the white plaid jacket"}]}
{"id": 444037, "image": "COCO_train2014_000000444037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse on right\"."}], "task": "refering", "answer_template": "The \"horse on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 209, 228, 229, 230, 231, 232, 251, 252, 253, 254, 273, 274, 275, 276, 295, 296, 297, 298, 318, 319, 320], "bbox": [0.3763562091503268, 0.4050653594771242, 0.5750490196078432, 0.670751633986928], "iscrowd": 0, "area": 10261.306150000006, "rle": {"size": [612, 612], "counts": "UhY43Tb0n0E:N2N2M3N2M4N1O100000001O000001N1O100O2N1O1O2N1O1O2N100O1O23L5K6KS1lN4L4L4M2M4M3L3N3L4XAWLn=l3gAVL_>l3XAVLl>j2XAmMn>R2dAWMa>j2m000O01000O10001N10000O1000000O0j0VO101N2N1OROj@PNU?k1TARNj>k1]ASNa>i1hAnM\\>o1lAfM[>U2[14M3M3M3M3M4K6K6I8I7H5L4K5K2OTOWAZMk>f2[ARMi>l2\\AmLh>Q3g0N1O2N2N1Oh@VMY>h2eA\\MZ>j2]AZMb>l2UAYMi>b3O1OO1O2N2O1N3M3J5cN]AlMP?k1Y1@a0K4L5K4L5K4K6J6J5LUSk4"}, "label": "horse on right"}]}
{"id": 444037, "image": "COCO_train2014_000000444037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horses pulling wagon\"."}], "task": "refering", "answer_template": "The \"horses pulling wagon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 209, 228, 229, 230, 231, 232, 251, 252, 253, 254, 273, 274, 275, 276, 295, 296, 297, 298, 318, 319, 320], "bbox": [0.3763562091503268, 0.4050653594771242, 0.5750490196078432, 0.670751633986928], "iscrowd": 0, "area": 10261.306150000006, "rle": {"size": [612, 612], "counts": "UhY43Tb0n0E:N2N2M3N2M4N1O100000001O000001N1O100O2N1O1O2N1O1O2N100O1O23L5K6KS1lN4L4L4M2M4M3L3N3L4XAWLn=l3gAVL_>l3XAVLl>j2XAmMn>R2dAWMa>j2m000O01000O10001N10000O1000000O0j0VO101N2N1OROj@PNU?k1TARNj>k1]ASNa>i1hAnM\\>o1lAfM[>U2[14M3M3M3M3M4K6K6I8I7H5L4K5K2OTOWAZMk>f2[ARMi>l2\\AmLh>Q3g0N1O2N2N1Oh@VMY>h2eA\\MZ>j2]AZMb>l2UAYMi>b3O1OO1O2N2O1N3M3J5cN]AlMP?k1Y1@a0K4L5K4L5K4K6J6J5LUSk4"}, "label": "horses pulling wagon"}]}
{"id": 91784, "image": "COCO_train2014_000000091784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe whose head does not go above the tree\"."}], "task": "refering", "answer_template": "The \"the giraffe whose head does not go above the tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 125, 146, 147, 148, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308], "bbox": [0.264125, 0.3236, 0.458125, 0.9393176470588236], "iscrowd": 0, "area": 17206.562249999995, "rle": {"size": [425, 640], "counts": "Y]V27o<a0@?@`0A>B>A;F8G8I8H7H3N0O2O11N101O1O0O2O001O1O0O2O001O0O2ZOf0G8K61N3^FfMQ8\\2gGTNo7n1jG^Nn7e1kG`NU8b1cGdN\\8V3O0O10000O2N1O1N2O2M2O1N2O1N3N0O2O1N2O001N2O1N2J5A`0EoNgIhKR6[4aJVK[5j4RKmJm4Q5YKkJf4S5aKjJ^4T5hKiJW4V5m12O1@?O20e0[O9G9G:F9G:F1O1O1O1O20N5K1ON2O0O2N2O10jJQK;;mMd4e1XK8f0cMR4S2^K3g0fMl3S2dK0f0iMg3U2iKKg0hMd3Z2mKGS65TJDo59XJ@k5<]J^OU6M`J@f58h3YOUDKP<6a01N2O001O0O01O10O01O1O1O010O1O3M9G_k_4"}, "label": "the giraffe whose head does not go above the tree"}]}
{"id": 91784, "image": "COCO_train2014_000000091784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter giraffe\"."}], "task": "refering", "answer_template": "The \"the shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 125, 146, 147, 148, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308], "bbox": [0.264125, 0.3236, 0.458125, 0.9393176470588236], "iscrowd": 0, "area": 17206.562249999995, "rle": {"size": [425, 640], "counts": "Y]V27o<a0@?@`0A>B>A;F8G8I8H7H3N0O2O11N101O1O0O2O001O1O0O2O001O0O2ZOf0G8K61N3^FfMQ8\\2gGTNo7n1jG^Nn7e1kG`NU8b1cGdN\\8V3O0O10000O2N1O1N2O2M2O1N2O1N3N0O2O1N2O001N2O1N2J5A`0EoNgIhKR6[4aJVK[5j4RKmJm4Q5YKkJf4S5aKjJ^4T5hKiJW4V5m12O1@?O20e0[O9G9G:F9G:F1O1O1O1O20N5K1ON2O0O2N2O10jJQK;;mMd4e1XK8f0cMR4S2^K3g0fMl3S2dK0f0iMg3U2iKKg0hMd3Z2mKGS65TJDo59XJ@k5<]J^OU6M`J@f58h3YOUDKP<6a01N2O001O0O01O10O01O1O1O010O1O3M9G_k_4"}, "label": "the shorter giraffe"}]}
{"id": 91784, "image": "COCO_train2014_000000091784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"picture of a longer giraffe than the one standing behind him\"."}], "task": "refering", "answer_template": "The \"picture of a longer giraffe than the one standing behind him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 82, 83, 105, 106, 128, 129, 130, 152, 153, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 290, 291, 292, 313, 314, 315, 336, 337, 338], "bbox": [0.5414218749999999, 0.06418823529411764, 0.7478125, 0.9966117647058822], "iscrowd": 0, "area": 23274.110150000004, "rle": {"size": [425, 640], "counts": "bR`41X=1O1O2M2O1O001N10001O00001O01O0001O00001O010O4J7ROYOnDn0l:^OgDi0R;l0K6J6VIiMW2]2aM^Nf1h1SNQOU1T1dN]On0i0jN]OT1g0dN^O[1f0\\NAb1c0VNCU24bM3Y24_M2]24ZM3a24WM2d25RM3j23mL4n23iL4S32dL5k2_5KVHSMc6d1\\I^Nd6a1[IbNd6]1[IeNe6Y1[IjNc6W1\\IjNd6X1YIjNf6X1WIjNg6X1eG[Nm0>Z7]1bGYNT1<U7a1`GWNZ1:R7Q2nHQNl6S2TInMh6U2^HjM_O3n7V2\\HmN_7V1`HlN[7X1dHkNW7X1hHjNT7Y1kHjNR7W1lHlNS7T1lHoNR7Q1mHROP7o0oHSOP7m0oHVOo6j0PIXOS7d0lH_OV7Z32N3N2M3M3M3M10L3N2M3N3M2M3N2M2O2M2O2N2M2O2M2O2N1N3N2J5C>^Oa0O2N1O2N1O1O1O1;E=C=C1O00O2O0O100O100O100O100O1mMnHnMS7j1ZJoLk5f2e2D=C=Bo0ROP1POi`R2"}, "label": "picture of a longer giraffe than the one standing behind him"}]}
{"id": 509579, "image": "COCO_train2014_000000509579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a seagull on a car mirror\"."}], "task": "refering", "answer_template": "The \"a seagull on a car mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 71, 72, 73, 87, 88, 89, 103, 104, 105, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 139, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 181, 182, 183, 184, 185, 186, 199, 200, 201, 215, 217, 231, 232, 233, 234, 246, 247, 248], "bbox": [0.34487858719646797, 0.14153125, 0.7139735099337747, 0.6769999999999999], "iscrowd": 0, "area": 27236.2958, "rle": {"size": [640, 453], "counts": "dhQ34hc0=D=C<C=D3L4N3M2N2N3M2O1N3M2N2N3N0O1O1O1O1O101O00000000001O2N2N2N2O1N2N2L4_BiLR:]3mElLe9Z3YFfLb9a3]F^L_9h3`FWL[9P4dFPLW9V4gFjKV7IhGc4Q1cKV76[G\\4^1]KV7`0QGX4g1YKV7g0hFV4Q2RKV7P1_FS4Z2lJV7Y1VFP4c2fJW7P7gHPIY7R7fHnHZ7S7eHlH[7\\7]HdHc7h7RHWHo7U8eGjG[8]7UFbIT1QOg8]7VFlIg0gNS9^7WFVJ:[N_9`7XF`JMoMl9a7XFeJf9\\5[FcJe9]5\\FbJd9^5]FbJb9^5_FaJa9_5`FaJ_9_5iFYJW9g5XGkIg8U6hG\\IX8d6WHmHi7S7bHcH]7]7dHbH\\7^7dHcH[7]7fHbHZ7^7fHcHY7]7hHbHX7^7iHbHV7^7jHcHU7]7lHbHT7^7mHbHR7^7nHbHR7^7oHbHP7^7PIcHo6]7RIbHn6^7SIbHl6^7VIaHi6_7XI`Hh6`7ZI_He6a7]I^Hb6b7_I]Hb6b7`I]H_6c7bI\\H_6c7P3O001O001O001O1O4L4L8HQ1oN2N3M2N2N3M27I8I6I5L2M3]CaI^;a6ZDcIg;`6PDdIS<`4YDXL[O]O]<Y4\\DlLg;Q3]DkLe;S3`DhLc;U3aDgLa;W3cDeL`;X3eDcL];[3gDaLZ;]3jD`LW;^3nD^LS;`3QE\\LQ;a3UEZLm:d3XEWLj:g3ZEVLh:f3k2N2M3N2N2M3N2N2M3N3M2M3N2M3O1O1N2O1N3N1N3N3L3N2N2M2N3L4M2N3L4M2M4M2NTPa2"}, "label": "a seagull on a car mirror"}]}
{"id": 509579, "image": "COCO_train2014_000000509579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bird sitting on the side mirror of a vehicle\"."}], "task": "refering", "answer_template": "The \"a white bird sitting on the side mirror of a vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 71, 72, 73, 87, 88, 89, 103, 104, 105, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 139, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 181, 182, 183, 184, 185, 186, 199, 200, 201, 215, 217, 231, 232, 233, 234, 246, 247, 248], "bbox": [0.34487858719646797, 0.14153125, 0.7139735099337747, 0.6769999999999999], "iscrowd": 0, "area": 27236.2958, "rle": {"size": [640, 453], "counts": "dhQ34hc0=D=C<C=D3L4N3M2N2N3M2O1N3M2N2N3N0O1O1O1O1O101O00000000001O2N2N2N2O1N2N2L4_BiLR:]3mElLe9Z3YFfLb9a3]F^L_9h3`FWL[9P4dFPLW9V4gFjKV7IhGc4Q1cKV76[G\\4^1]KV7`0QGX4g1YKV7g0hFV4Q2RKV7P1_FS4Z2lJV7Y1VFP4c2fJW7P7gHPIY7R7fHnHZ7S7eHlH[7\\7]HdHc7h7RHWHo7U8eGjG[8]7UFbIT1QOg8]7VFlIg0gNS9^7WFVJ:[N_9`7XF`JMoMl9a7XFeJf9\\5[FcJe9]5\\FbJd9^5]FbJb9^5_FaJa9_5`FaJ_9_5iFYJW9g5XGkIg8U6hG\\IX8d6WHmHi7S7bHcH]7]7dHbH\\7^7dHcH[7]7fHbHZ7^7fHcHY7]7hHbHX7^7iHbHV7^7jHcHU7]7lHbHT7^7mHbHR7^7nHbHR7^7oHbHP7^7PIcHo6]7RIbHn6^7SIbHl6^7VIaHi6_7XI`Hh6`7ZI_He6a7]I^Hb6b7_I]Hb6b7`I]H_6c7bI\\H_6c7P3O001O001O001O1O4L4L8HQ1oN2N3M2N2N3M27I8I6I5L2M3]CaI^;a6ZDcIg;`6PDdIS<`4YDXL[O]O]<Y4\\DlLg;Q3]DkLe;S3`DhLc;U3aDgLa;W3cDeL`;X3eDcL];[3gDaLZ;]3jD`LW;^3nD^LS;`3QE\\LQ;a3UEZLm:d3XEWLj:g3ZEVLh:f3k2N2M3N2N2M3N2N2M3N3M2M3N2M3O1O1N2O1N3N1N3N3L3N2N2M2N3L4M2N3L4M2M4M2NTPa2"}, "label": "a white bird sitting on the side mirror of a vehicle"}]}
{"id": 272022, "image": "COCO_train2014_000000272022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with a blonde mane and a large white patch on its face and white on its front legs kicking up dust\"."}], "task": "refering", "answer_template": "The \"a horse with a blonde mane and a large white patch on its face and white on its front legs kicking up dust\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 238, 239, 257, 258, 259, 260, 261, 280, 281, 282, 284], "bbox": [0.180125, 0.23056206088992975, 0.41068750000000004, 0.8201873536299766], "iscrowd": 0, "area": 20451.359350000006, "rle": {"size": [427, 640], "counts": "kU`16d<a0_Ob0E;M2N3M3M2N3N2M3lGkMo4W2kJoMR5U2cHnMW15T6X3oHWMn6e4N1N2O1O1N2O1N2O1N2O100OhMXIiMn6R2W2O2N1O2N101O001O1O101N1O1O1O101N1000TGPN]6P2aIUN\\6`4N2M2N2N1O01O00000000001O00000000000`MZIUMl7c2b1O100O1O1O100O001O0001O000M3L5N1N3M3N2M3N2N3N2N2N2N2N2N2N1O2YIjKU4Y4eKoKV4T4fKPLX4R4gKoKV4T4iKnKT4T4kKmKR4V4mKkKQ4V4oKkKn3W4QLjKn3W4QLkKm3W4QLjKm3X4RLjKl3W4SLjKl3W4SLkKl3U4SLlKm3T4RLmKo3S4PLmKQ4R4nKoKR4Q4mKoKU4P4hKSLY4l3dKVL]4j3`KYLa4W62O0O2O0O2N1M4L3nLeJRN^5g1iJUNU5j1SKQNo4n1UKnMm4P2WKkMl4S2WKiMk4V2YKfMi4X2[KcMh4[2\\K`Mg4[2`K`Mb4Y2gKdM[4T2oKgMT4e1bLVN`3e1hLVN[3f1jLWNX3e1Z4Ji0YO6I7I7I7JaRm4"}, "label": "a horse with a blonde mane and a large white patch on its face and white on its front legs kicking up dust"}]}
{"id": 272022, "image": "COCO_train2014_000000272022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse on the left pulling some sort of carriage\"."}], "task": "refering", "answer_template": "The \"a brown horse on the left pulling some sort of carriage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 238, 239, 257, 258, 259, 260, 261, 280, 281, 282, 284], "bbox": [0.180125, 0.23056206088992975, 0.41068750000000004, 0.8201873536299766], "iscrowd": 0, "area": 20451.359350000006, "rle": {"size": [427, 640], "counts": "kU`16d<a0_Ob0E;M2N3M3M2N3N2M3lGkMo4W2kJoMR5U2cHnMW15T6X3oHWMn6e4N1N2O1O1N2O1N2O1N2O100OhMXIiMn6R2W2O2N1O2N101O001O1O101N1O1O1O101N1000TGPN]6P2aIUN\\6`4N2M2N2N1O01O00000000001O00000000000`MZIUMl7c2b1O100O1O1O100O001O0001O000M3L5N1N3M3N2M3N2N3N2N2N2N2N2N2N1O2YIjKU4Y4eKoKV4T4fKPLX4R4gKoKV4T4iKnKT4T4kKmKR4V4mKkKQ4V4oKkKn3W4QLjKn3W4QLkKm3W4QLjKm3X4RLjKl3W4SLjKl3W4SLkKl3U4SLlKm3T4RLmKo3S4PLmKQ4R4nKoKR4Q4mKoKU4P4hKSLY4l3dKVL]4j3`KYLa4W62O0O2O0O2N1M4L3nLeJRN^5g1iJUNU5j1SKQNo4n1UKnMm4P2WKkMl4S2WKiMk4V2YKfMi4X2[KcMh4[2\\K`Mg4[2`K`Mb4Y2gKdM[4T2oKgMT4e1bLVN`3e1hLVN[3f1jLWNX3e1Z4Ji0YO6I7I7I7JaRm4"}, "label": "a brown horse on the left pulling some sort of carriage"}]}
{"id": 272022, "image": "COCO_train2014_000000272022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse on the right\"."}], "task": "refering", "answer_template": "The \"the horse on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 265], "bbox": [0.40290625, 0.2605620608899298, 0.67484375, 0.7463700234192037], "iscrowd": 0, "area": 20611.967200000006, "rle": {"size": [427, 640], "counts": "[h[39Q==B=C=D=YEYNP9T2YF`N]9o2H9F:G8G3N001O001O00mMjGSOU8m0WHhNh7X1`H`N`7`1iHXNV7h1SIoMm6Q2bIaM]6^2Y2O2N1O2N1O1N2O11O001O1O1O1O1O1O1O1O1O1O100bFcMl7^2aGXN[8i1RGmNj8n2L4L3M2N000O100O10O010gNjG]MU8`2XHWMh7e2eHRMZ7k2SIlLm6Q3g1M4L3M40O5L4K6K4L4K5L4K5mGnKe6W4UIjKj6[4PIfKP7^4kHdKU7`4eHbKZ7b4bH_K]7f4]H\\Kc7g4XH[Kg7T510O1O100O10O01O100O1O10O0100O1O100N1O2OO0100O010O0011O1N2O1O1N2O1O1N2O1O1N2O1gMTHnNQ8T1VHXNU8k1PHbM[8`2lGkL`8W3i03M2N0000O0100O100O010O11O1O1N2O1O1O1O1N2O1O2N1O0L5B=C>C=J5O2N1N3N00001O0WOTD3m;LXD0i;N\\DNd;1aDK`;3eDI[;6iDFX;9n0N1N2O2N1O1Olcf2"}, "label": "the horse on the right"}]}
{"id": 272022, "image": "COCO_train2014_000000272022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown horse on the right\"."}], "task": "refering", "answer_template": "The \"the brown horse on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 265], "bbox": [0.40290625, 0.2605620608899298, 0.67484375, 0.7463700234192037], "iscrowd": 0, "area": 20611.967200000006, "rle": {"size": [427, 640], "counts": "[h[39Q==B=C=D=YEYNP9T2YF`N]9o2H9F:G8G3N001O001O00mMjGSOU8m0WHhNh7X1`H`N`7`1iHXNV7h1SIoMm6Q2bIaM]6^2Y2O2N1O2N1O1N2O11O001O1O1O1O1O1O1O1O1O1O100bFcMl7^2aGXN[8i1RGmNj8n2L4L3M2N000O100O10O010gNjG]MU8`2XHWMh7e2eHRMZ7k2SIlLm6Q3g1M4L3M40O5L4K6K4L4K5L4K5mGnKe6W4UIjKj6[4PIfKP7^4kHdKU7`4eHbKZ7b4bH_K]7f4]H\\Kc7g4XH[Kg7T510O1O100O10O01O100O1O10O0100O1O100N1O2OO0100O010O0011O1N2O1O1N2O1O1N2O1O1N2O1gMTHnNQ8T1VHXNU8k1PHbM[8`2lGkL`8W3i03M2N0000O0100O100O010O11O1O1N2O1O1O1O1N2O1O2N1O0L5B=C>C=J5O2N1N3N00001O0WOTD3m;LXD0i;N\\DNd;1aDK`;3eDI[;6iDFX;9n0N1N2O2N1O1Olcf2"}, "label": "the brown horse on the right"}]}
{"id": 206486, "image": "COCO_train2014_000000206486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a blue surfboard\"."}], "task": "refering", "answer_template": "The \"a man holding a blue surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 39, 40, 41, 62, 63, 64, 65, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 158, 159, 174, 175, 197, 198, 220, 221, 243, 266, 289, 290, 314, 315, 316, 338, 339, 361, 362, 384, 385, 407, 408, 431, 432, 454, 455, 478], "bbox": [0.5806250000000001, 0.01798076923076923, 0.9509062500000001, 0.9168589743589743], "iscrowd": 0, "area": 28397.428199999995, "rle": {"size": [624, 640], "counts": "enR7c0fb0\\1eN>A>@<D;D;D<F;J5K5J7J5J5LcN[BWMb=f2fBXMW=e2RCXMl<d2]CZM`<b2hC\\MU<b2SD[Mk;a2^D\\M`;`2hD^MU;_2TE^Mj:^2^EaM`:Z2hEdMV:Z1oFeNP9[1RGdNl8\\1XGbNg8^1[GaNc8`1_G_N`8a1bG^N]8b1fG\\NX8d1kGZNU8f1mGXNR8i1PHUNP8k1SHSNl7m1j4010O10000O10O0100O1000O2O0000000O2O0O1N2O1O2M2O1O1O11O1O2N2hDZNc4h1\\KZNa4h1^K[N^4g1aK[N\\4g1cK\\NY4f1gK[NV4f1jK\\NS4f1lK]NP4e1oK]Nb2HSHl1[5_NY2NXHd1_5aNP24]H\\1c5dNg17cHV1f5fNc18bHT1k5gN_1:aHP1P6jNY1=aHj0W6kNT1P3lNTMo0m2QOWMi0k2WOZMc0g2]O]M=e2C^M9c2G_M6b2JaM1a2ObMM_23iMCY2=nM[OS2e0TNSOm1m0WNmNk1S1YNhNh1X1]NbNd1_1_N]Na1c1dNXN\\1h1hNTNX1l1lNoMU1Q2nNlMR1T2QOiMo0W2TOfMl0Z2VOcMk0]2XO`Mg0a2\\O\\Md0d2_OYMa0g2BUM?k2DRM<n2FPM:o2JnL6R3MjL4V3nLhIn0o2T2Y3lLnIn0g2V2[3jLUJn0]2X2^3hL[Jo0S2Z2a3fLaJP1k1Z2d3cLhJQ1b1\\2f3aLmJS1Z1\\2i3_LSKT1P1^2m3\\LXKV1h0^2P4ZL^KV1a0_2Q4VLgKZ15`2S4RLPL^1K_2U4nKYLa1A`2V4kK`Lc1ZOa2V4gKiLd1QOe2V4bKRMe1iNh2W4\\KXMi1bNi2X4XK_Mk1YNm2\\8PMeGo2_8mLbGR3a8lL^GU3e8gL\\GX3g8eLZGZ3j8bLXG\\3k8aLXG1PNd1m:VNlHh1Y7TNgHk1^7PNdHn1a7mMaHQ2d7jM^HT2g7gM[HW2a<N1O1O010O00001N100O2N1O1O100O1O2N101N2N1O2O1N2O[OXNb_Og1^`0[Nb_Oc1^`0`Na_O_1^`0cNb_OZ1_`0iN`_OT1b`0mN^_OQ1b`0QO__OS1Z`0oNf_OR1X`0nNj_OQ1U`0oNm_OP1R`0POo_OP1P`0POR@o0l?ROU@n0j?TOV@k0j?VOV@h0k?YOT@g0l?YOU@f0k?ZOU@f0k?ZOV@e0k?ZOV@e0j?\\OU@d0k?^OT@a0l?AR@?n?CQ@<o?Hn_O6S`0Mk_O2V`01g_ONY`06d_OH]`0<__ODa`0`0[_O_Of`0d0W_O\\Oi`0f10O2O0O10000O100O2O001N3N1N3N2N2M3N2M3N2N1N3N2M2O2N1N3N1N3N1N3N1O3L<E;D]mc0"}, "label": "a man holding a blue surfboard"}]}
{"id": 206486, "image": "COCO_train2014_000000206486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy carrying a blue surfboard\"."}], "task": "refering", "answer_template": "The \"a boy carrying a blue surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 39, 40, 41, 62, 63, 64, 65, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 158, 159, 174, 175, 197, 198, 220, 221, 243, 266, 289, 290, 314, 315, 316, 338, 339, 361, 362, 384, 385, 407, 408, 431, 432, 454, 455, 478], "bbox": [0.5806250000000001, 0.01798076923076923, 0.9509062500000001, 0.9168589743589743], "iscrowd": 0, "area": 28397.428199999995, "rle": {"size": [624, 640], "counts": "enR7c0fb0\\1eN>A>@<D;D;D<F;J5K5J7J5J5LcN[BWMb=f2fBXMW=e2RCXMl<d2]CZM`<b2hC\\MU<b2SD[Mk;a2^D\\M`;`2hD^MU;_2TE^Mj:^2^EaM`:Z2hEdMV:Z1oFeNP9[1RGdNl8\\1XGbNg8^1[GaNc8`1_G_N`8a1bG^N]8b1fG\\NX8d1kGZNU8f1mGXNR8i1PHUNP8k1SHSNl7m1j4010O10000O10O0100O1000O2O0000000O2O0O1N2O1O2M2O1O1O11O1O2N2hDZNc4h1\\KZNa4h1^K[N^4g1aK[N\\4g1cK\\NY4f1gK[NV4f1jK\\NS4f1lK]NP4e1oK]Nb2HSHl1[5_NY2NXHd1_5aNP24]H\\1c5dNg17cHV1f5fNc18bHT1k5gN_1:aHP1P6jNY1=aHj0W6kNT1P3lNTMo0m2QOWMi0k2WOZMc0g2]O]M=e2C^M9c2G_M6b2JaM1a2ObMM_23iMCY2=nM[OS2e0TNSOm1m0WNmNk1S1YNhNh1X1]NbNd1_1_N]Na1c1dNXN\\1h1hNTNX1l1lNoMU1Q2nNlMR1T2QOiMo0W2TOfMl0Z2VOcMk0]2XO`Mg0a2\\O\\Md0d2_OYMa0g2BUM?k2DRM<n2FPM:o2JnL6R3MjL4V3nLhIn0o2T2Y3lLnIn0g2V2[3jLUJn0]2X2^3hL[Jo0S2Z2a3fLaJP1k1Z2d3cLhJQ1b1\\2f3aLmJS1Z1\\2i3_LSKT1P1^2m3\\LXKV1h0^2P4ZL^KV1a0_2Q4VLgKZ15`2S4RLPL^1K_2U4nKYLa1A`2V4kK`Lc1ZOa2V4gKiLd1QOe2V4bKRMe1iNh2W4\\KXMi1bNi2X4XK_Mk1YNm2\\8PMeGo2_8mLbGR3a8lL^GU3e8gL\\GX3g8eLZGZ3j8bLXG\\3k8aLXG1PNd1m:VNlHh1Y7TNgHk1^7PNdHn1a7mMaHQ2d7jM^HT2g7gM[HW2a<N1O1O010O00001N100O2N1O1O100O1O2N101N2N1O2O1N2O[OXNb_Og1^`0[Nb_Oc1^`0`Na_O_1^`0cNb_OZ1_`0iN`_OT1b`0mN^_OQ1b`0QO__OS1Z`0oNf_OR1X`0nNj_OQ1U`0oNm_OP1R`0POo_OP1P`0POR@o0l?ROU@n0j?TOV@k0j?VOV@h0k?YOT@g0l?YOU@f0k?ZOU@f0k?ZOV@e0k?ZOV@e0j?\\OU@d0k?^OT@a0l?AR@?n?CQ@<o?Hn_O6S`0Mk_O2V`01g_ONY`06d_OH]`0<__ODa`0`0[_O_Of`0d0W_O\\Oi`0f10O2O0O10000O100O2O001N3N1N3N2N2M3N2M3N2N1N3N2M2O2N1N3N1N3N1N3N1O3L<E;D]mc0"}, "label": "a boy carrying a blue surfboard"}]}
{"id": 206486, "image": "COCO_train2014_000000206486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing a pink shirt\"."}], "task": "refering", "answer_template": "The \"the woman wearing a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 74, 95, 96, 97, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 306, 323, 324, 325, 326, 327, 348, 349, 350, 371, 372, 373, 394, 395, 396, 417, 418, 419, 440, 441, 442, 464, 465, 487, 488], "bbox": [0.053234375, 0.12246794871794872, 0.351375, 0.9769070512820511], "iscrowd": 0, "area": 50856.53345000002, "rle": {"size": [624, 640], "counts": "[od0i0_b0c1^NS1lNe0\\O>C:F:E;F7I7M2N3M3M2N3M3M2N3M2N3M2N3M2M3N2N2N2O1O1O100O1O01O0001O01O00001L3O12O1N2N3N1N2N2O1UOhH\\EY7X:VIdEk6P:cInE^6f9PJVFQ6h9m1O1N2N2O1N3M2O1N2]J_FU1b9bNlFW1V9aNRG]1P9ZNXGf1j8PN_Gn1c8jMdGU2^8bMkG]2V8[MQHd2Q8SMWHm2j7jL^HU3d7cLcH\\3`7ZLhHf3b;N1O2O1N3N1N2N2O1N3M2O3L8J5J4MO01O010O0010O010O01O010O0010O0100O0010O01O010O01O000O2O0O1O2O0YNmLPCU3f<VMXCl2\\<aM`Cc2T<iMhC[2W9gL\\Gi1[Oc1n3jL]O;fLe1K[1d3]MWO3eLb1>Q1f3cMdNe3`MlNl3aMYNn3fMdNQ4aMlMn8U2SGaMV9VNnEP3l0`N`9YNiEX3h0VNm:j1VEkMR;V2oDaMX;`2iDWM_;h2cDnLe;S3\\DdLk;\\3W2O2N1N3N1N3N1O2M3N4L6J6J6J6J6J1O100O101N100O100O10001O000O1000001OQ_ObNi?_1n_OfNU`0^1c_OcN``0a1X_O`Nk`0b1o^O^NUa0a1h^O`N[a0i13L4M3M3L4M3L4L5K5K4M3L4J7IXal7"}, "label": "the woman wearing a pink shirt"}]}
{"id": 206486, "image": "COCO_train2014_000000206486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady not holding a surfboard\"."}], "task": "refering", "answer_template": "The \"a lady not holding a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 74, 95, 96, 97, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 306, 323, 324, 325, 326, 327, 348, 349, 350, 371, 372, 373, 394, 395, 396, 417, 418, 419, 440, 441, 442, 464, 465, 487, 488], "bbox": [0.053234375, 0.12246794871794872, 0.351375, 0.9769070512820511], "iscrowd": 0, "area": 50856.53345000002, "rle": {"size": [624, 640], "counts": "[od0i0_b0c1^NS1lNe0\\O>C:F:E;F7I7M2N3M3M2N3M3M2N3M2N3M2N3M2M3N2N2N2O1O1O100O1O01O0001O01O00001L3O12O1N2N3N1N2N2O1UOhH\\EY7X:VIdEk6P:cInE^6f9PJVFQ6h9m1O1N2N2O1N3M2O1N2]J_FU1b9bNlFW1V9aNRG]1P9ZNXGf1j8PN_Gn1c8jMdGU2^8bMkG]2V8[MQHd2Q8SMWHm2j7jL^HU3d7cLcH\\3`7ZLhHf3b;N1O2O1N3N1N2N2O1N3M2O3L8J5J4MO01O010O0010O010O01O010O0010O0100O0010O01O010O01O000O2O0O1O2O0YNmLPCU3f<VMXCl2\\<aM`Cc2T<iMhC[2W9gL\\Gi1[Oc1n3jL]O;fLe1K[1d3]MWO3eLb1>Q1f3cMdNe3`MlNl3aMYNn3fMdNQ4aMlMn8U2SGaMV9VNnEP3l0`N`9YNiEX3h0VNm:j1VEkMR;V2oDaMX;`2iDWM_;h2cDnLe;S3\\DdLk;\\3W2O2N1N3N1N3N1O2M3N4L6J6J6J6J6J1O100O101N100O100O10001O000O1000001OQ_ObNi?_1n_OfNU`0^1c_OcN``0a1X_O`Nk`0b1o^O^NUa0a1h^O`N[a0i13L4M3M3L4M3L4L5K5K4M3L4J7IXal7"}, "label": "a lady not holding a surfboard"}]}
{"id": 485016, "image": "COCO_train2014_000000485016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair\"."}], "task": "refering", "answer_template": "The \"a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 154, 155, 156, 172, 173, 174, 175, 178, 179, 190, 191, 192, 193, 196, 197, 210, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 251], "bbox": [0.5516796875, 0.32437499999999997, 0.9834375, 0.6074375], "iscrowd": 0, "area": 12740.920950000002, "rle": {"size": [640, 512], "counts": "`^`54lc0:F3M2N2O1N1O2N2O1N3M3M4L3M3M4L6J3M3M3M3M4L2N1O100O001O1O1O001O1O1O1O002N1O2N1O1O2O0O1O1O1O0000001O00001O010O1O001O001O001O0O2h@mLa=T3^BmLa=S3^BoLa=R3^BoLa=R3]BoLc=R3\\BnLd=S3ZBnLe=T3ZBlLf=U3XBlLh=8_Al1h0mMh=NiBV1]OmNk?T1T@lNm?T1Q@mNo?T1P@lNP`0U1n_OkNS`0V1l_OjNT`0W1j_OiNW`0X1g_OiNZ`0W1e_OhN\\`0Y1b_OhN^`0Y1`_OgNa`0Z1^_OfNa`0\\1]_OdNd`0\\1[_OeNe`0\\1Z_OcNg`0^1W_OcNi`0^1U_ObNl`0_1S_OaNm`0`1Q_O`NPa0a1n^O`NQa0b1n^O]NSa0d1k^O]NUa0d1i^O\\NXa0e1g^O[NYa0U2100O10O012N5K5K5K6J8HO:G?@:G5J101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O2O001N100O2O0O2O001N101N101N101M2N3N1N3N1N3M2O2M2O2M4L4M3L5Ljm9OXRF2M2O2N1N1Q^OHV`08j_OHU`09j_OHhN=``0Kg@HkN<]`0Mg@HmN;[`0Ng@GoN:Y`00g@FQO;W`0OU@2SOEb09V`00U@9FGT`00W@7GHR`01W@7HHP`02X@5HIP`02Y@3IKm?2Z@3JJl?3[@2JKj?4[@0LLi?4\\@OLMg?4]@OLMg?7[@JO0e?8Z@H2Od?;Y@E31c?>V@@82b?b0S@[O<3`?d0S@XO=3b?f0o_OWO?3b?g0o_OTOa03a?k0l_OROc03b?k0k_OXO=Ki?m0j_O=W`0Cj_O<X`0Bh_O=\\`0@e_O?^`0]Od_Oa0_`0\\Ob_Od0a`0WOb_Og0b`0mNg_OP1ga0Kje4"}, "label": "a chair"}]}
{"id": 59034, "image": "COCO_train2014_000000059034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unoccupied wooden bench by a body of water\"."}], "task": "refering", "answer_template": "The \"an unoccupied wooden bench by a body of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 283, 284, 285, 286, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.14765625, 0.6512291666666666, 0.45301562500000003, 0.9867916666666666], "iscrowd": 0, "area": 18847.810300000005, "rle": {"size": [480, 640], "counts": "Zo\\1`0o=a0_Oa0I7N2O1N2O1N2O1O1N2O1O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N2N2N2N3M2N2N3M1O00PO]NcDc1X;cNgD]1T;iNkDW1P;oNnDR1l:UOSEk0h:ZOXEf0c:@XEd0c:BWEc0d:CVEb0d:EVE`0e:P2K5K5J6K5L400001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O00001O00000I8H7I7I7H9H7I7I7I8G8I7I7I8H7I7H8I7IbgS5"}, "label": "an unoccupied wooden bench by a body of water"}]}
{"id": 59034, "image": "COCO_train2014_000000059034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the full bench behind the woman\"."}], "task": "refering", "answer_template": "The \"the full bench behind the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 283, 284, 285, 286, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.14765625, 0.6512291666666666, 0.45301562500000003, 0.9867916666666666], "iscrowd": 0, "area": 18847.810300000005, "rle": {"size": [480, 640], "counts": "Zo\\1`0o=a0_Oa0I7N2O1N2O1N2O1O1N2O1O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N2N2N2N3M2N2N3M1O00PO]NcDc1X;cNgD]1T;iNkDW1P;oNnDR1l:UOSEk0h:ZOXEf0c:@XEd0c:BWEc0d:CVEb0d:EVE`0e:P2K5K5J6K5L400001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O00001O00000I8H7I7I7H9H7I7I7I8G8I7I7I8H7I7H8I7IbgS5"}, "label": "the full bench behind the woman"}]}
{"id": 59034, "image": "COCO_train2014_000000059034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench the woman is sitting on\"."}], "task": "refering", "answer_template": "The \"the bench the woman is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 286, 287, 309, 310, 332, 333, 355, 356, 357, 359, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.354015625, 0.6452291666666666, 0.7220625, 1.0], "iscrowd": 0, "area": 9193.76275, "rle": {"size": [480, 640], "counts": "ifZ32l>3N2O1000000O01N2O0O2O1N10VS61hlI1N3O0010O010O01O01O01O010O010O00010O0010O01O010O0TDJn86kF2U9NcF9]9G\\Fa0c9@UFh0k9WOoEo0Q:ROgEV1Y:iN`E_1_:bNYEe1g:[NREm1n:RNiDX2V;iMZDf2f;;0O01O10O01O2O3L4M3L3M4M3L4L4M3L2N01O00001O00001O001O00002N3M3M3M3M3M3M3M3M3M>B`0@6J7I7I7I6J7I3M3M2N3M2N2N3M1OO1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O00Qob2"}, "label": "the bench the woman is sitting on"}]}
{"id": 59034, "image": "COCO_train2014_000000059034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a side of the bench that a woman is sitting on\"."}], "task": "refering", "answer_template": "The \"a side of the bench that a woman is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 286, 287, 309, 310, 332, 333, 355, 356, 357, 359, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.354015625, 0.6452291666666666, 0.7220625, 1.0], "iscrowd": 0, "area": 9193.76275, "rle": {"size": [480, 640], "counts": "ifZ32l>3N2O1000000O01N2O0O2O1N10VS61hlI1N3O0010O010O01O01O01O010O010O00010O0010O01O010O0TDJn86kF2U9NcF9]9G\\Fa0c9@UFh0k9WOoEo0Q:ROgEV1Y:iN`E_1_:bNYEe1g:[NREm1n:RNiDX2V;iMZDf2f;;0O01O10O01O2O3L4M3L3M4M3L4L4M3L2N01O00001O00001O001O00002N3M3M3M3M3M3M3M3M3M>B`0@6J7I7I7I6J7I3M3M2N3M2N2N3M1OO1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O00Qob2"}, "label": "a side of the bench that a woman is sitting on"}]}
{"id": 18075, "image": "COCO_train2014_000000018075.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"awoman with an i love the rain umbrella\"."}], "task": "refering", "answer_template": "The \"awoman with an i love the rain umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188], "bbox": [0.2259718969555035, 0.32810937500000004, 0.6410538641686183, 0.54365625], "iscrowd": 0, "area": 18884.349700000002, "rle": {"size": [640, 427], "counts": "]Xl12mc0?B>A?A8H1N2O2N1N2O1O1N2O1O2M2O1O1N3N2N1N3N1O2M2O2N1N3N101N2N101N1O2O0O2O0O2N101N2N101N1O2O0O2O0O2N101N2N101N1O2O0O2O0O2N101N2N100O100000O100000000000000000000O1000000000000000000O1000000000000000001N100000001O00000000001N100000001O000000000O2O00000000001O0000001N2O1O001O1O1N2N101N2O1N2O1N101N2N2N2N101N2N2N2N101N2N2N4L4M3L4L4L4L4L4M3L3M4L4L4M3L4L4L4L4@Y[o2"}, "label": "awoman with an i love the rain umbrella"}]}
{"id": 18075, "image": "COCO_train2014_000000018075.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella i love rain\"."}], "task": "refering", "answer_template": "The \"umbrella i love rain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188], "bbox": [0.2259718969555035, 0.32810937500000004, 0.6410538641686183, 0.54365625], "iscrowd": 0, "area": 18884.349700000002, "rle": {"size": [640, 427], "counts": "]Xl12mc0?B>A?A8H1N2O2N1N2O1O1N2O1O2M2O1O1N3N2N1N3N1O2M2O2N1N3N101N2N101N1O2O0O2O0O2N101N2N101N1O2O0O2O0O2N101N2N101N1O2O0O2O0O2N101N2N100O100000O100000000000000000000O1000000000000000000O1000000000000000001N100000001O00000000001N100000001O000000000O2O00000000001O0000001N2O1O001O1O1N2N101N2O1N2O1N101N2N2N2N101N2N2N2N101N2N2N4L4M3L4L4L4L4L4M3L3M4L4L4M3L4L4L4L4@Y[o2"}, "label": "umbrella i love rain"}]}
{"id": 239263, "image": "COCO_train2014_000000239263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow vehicle with the back window open\"."}], "task": "refering", "answer_template": "The \"a yellow vehicle with the back window open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 256], "bbox": [0.0753125, 0.42679906542056073, 0.2891875, 0.7443691588785046], "iscrowd": 0, "area": 10215.67475, "rle": {"size": [428, 640], "counts": "lRe02Z=000VFOX62gIOY62eIOZ62dI0\\60cI1]6O^I7`6JZI<f6ERIb0n6^OkHi0T7XOfHn0Z7RO_HU1a7lNWH[1h7fNRH`1n7`NQHb1m7_NSHa1m7`NRH`1n7`NRH`1n7`NRH`1n7`NRGF;j1c8`NPGH=h1c8aNnFH`0g1a8aNmFKa0d1b8aNlFLb0c1b8aNkFMc0b1b8bNhFOe0_1c8bNgF0f0^1c8bNfF1g0]1c8bNdF3i0[1c8bNcF4j0[1b8VO^Gj0b8VO^Gj0b8VO^Gj0b8VO^Gj0b8VO_Gi0a8WO_Gi0a8WO_Gj0`8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gj0`8VO`Gj0`8VO`Gj0`8VO`Gj0`8VO`Gj0_8WObGh0^8XObGh0^8XObGh0^8XObGi0]8WOcGi0]8WOcGi0]8WOcGi0\\8XOdGh0\\8XOdGh0\\8XOdGh0\\8XOdGi0Z8YOeGf0\\8ZOdGe0]8[OcGd0]8]OdGa0]8_OcGa0]8_OcG`0]8AcG>^8BbG=_8CaG<`8D`G;`8F`G9a8G_G9a8G_G8a8I_G6b8J^G6b8J^G5c8K]G5c8K^G4b8L^G4b8L^G4b8L^G3c8N\\G2d8N\\G2d8N\\G2d8N\\G1e8O[G1e8O[G1f8NZG2f8NZG1g8OYG1g8OYG1g8OZG0f80ZGOg81YGOg81YGOg82XGNh82XGMi83WGMi83WGMi83WGLk83UGMk83UGLl85SGKm85SGJn86RGJn86SGHn88RGHn88RGGo89QGGo8:PGEQ9;oFEQ9;oFDS9;mFES9;mFDT9<lFDT9<lFAW9`0hF\\O\\9d0dFWOa9i0_FSOe9m0\\FnNh9R1XFjNl9V1TFfNP:[1P101N1L4L4K5L4L4L4K5L4L4Knom5"}, "label": "a yellow vehicle with the back window open"}]}
{"id": 239263, "image": "COCO_train2014_000000239263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow vehical with hatch open\"."}], "task": "refering", "answer_template": "The \"yellow vehical with hatch open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 256], "bbox": [0.0753125, 0.42679906542056073, 0.2891875, 0.7443691588785046], "iscrowd": 0, "area": 10215.67475, "rle": {"size": [428, 640], "counts": "lRe02Z=000VFOX62gIOY62eIOZ62dI0\\60cI1]6O^I7`6JZI<f6ERIb0n6^OkHi0T7XOfHn0Z7RO_HU1a7lNWH[1h7fNRH`1n7`NQHb1m7_NSHa1m7`NRH`1n7`NRH`1n7`NRH`1n7`NRGF;j1c8`NPGH=h1c8aNnFH`0g1a8aNmFKa0d1b8aNlFLb0c1b8aNkFMc0b1b8bNhFOe0_1c8bNgF0f0^1c8bNfF1g0]1c8bNdF3i0[1c8bNcF4j0[1b8VO^Gj0b8VO^Gj0b8VO^Gj0b8VO^Gj0b8VO_Gi0a8WO_Gi0a8WO_Gj0`8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gi0a8WO_Gj0`8VO`Gj0`8VO`Gj0`8VO`Gj0`8VO`Gj0_8WObGh0^8XObGh0^8XObGh0^8XObGi0]8WOcGi0]8WOcGi0]8WOcGi0\\8XOdGh0\\8XOdGh0\\8XOdGh0\\8XOdGi0Z8YOeGf0\\8ZOdGe0]8[OcGd0]8]OdGa0]8_OcGa0]8_OcG`0]8AcG>^8BbG=_8CaG<`8D`G;`8F`G9a8G_G9a8G_G8a8I_G6b8J^G6b8J^G5c8K]G5c8K^G4b8L^G4b8L^G4b8L^G3c8N\\G2d8N\\G2d8N\\G2d8N\\G1e8O[G1e8O[G1f8NZG2f8NZG1g8OYG1g8OYG1g8OZG0f80ZGOg81YGOg81YGOg82XGNh82XGMi83WGMi83WGMi83WGLk83UGMk83UGLl85SGKm85SGJn86RGJn86SGHn88RGHn88RGGo89QGGo8:PGEQ9;oFEQ9;oFDS9;mFES9;mFDT9<lFDT9<lFAW9`0hF\\O\\9d0dFWOa9i0_FSOe9m0\\FnNh9R1XFjNl9V1TFfNP:[1P101N1L4L4K5L4L4L4K5L4L4Knom5"}, "label": "yellow vehical with hatch open"}]}
{"id": 239263, "image": "COCO_train2014_000000239263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow car with the pink hood\"."}], "task": "refering", "answer_template": "The \"the yellow car with the pink hood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 131, 132, 144, 145, 146, 147, 148, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 274, 275, 297, 298, 320, 321, 344], "bbox": [0.2645, 0.3595560747663551, 0.999375, 0.9865186915887849], "iscrowd": 0, "area": 40958.991850000006, "rle": {"size": [428, 640], "counts": "[mV22T=5L5J6J6jNZOcEh0[:[OeEe0W:_OiEa0S:CmE=P:FoE;m9JRF5l9NTF2h92XFNe95[FKa99_FF_9=aFC[9a0fF^OW9e0iF[OS9i0mFWOR9j0nFUOR9m0nFROP9P1PGPOo8Q1QGoNn8R1RGmNn8T1SGjNm8W1SGhNm8Y1SGfNm8[1SGeNl8\\1UGbNj8`1VG_Nj8c1UG\\Nk8e1UGZNk8g1VGXNi8i1WGVNh8l1XGSNh8n1YGPNf8R2ZGmMf8T2ZGkMf8V2ZGjMd8Y2\\GeMe8[2[GdMf8\\2ZGcMg8]2T10000000000001O0000000000000000001O0000000000000000001OO1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1OPQe0OQoZO2M2O1N2O1N2O1O1N2O1N3N1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2O1O2M2O1N2O1N2O1O1N2O1N3N1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1O1N3N1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2MgKRGZ4n83O0001O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001OO1O2N1O100O1O2N1O1O1O1O2O01O001O01O01O001O01O01O001O010O00001O010O001O0001n0QO>B1O1O101N1O1O1O2O0001O010O001O001O001O001O001O1O001O001O001O0013L5L3L4M4K2OO0O2N1O2N1O2O1N1O3M2N2N2O2M2N2NlG"}, "label": "the yellow car with the pink hood"}]}
{"id": 239263, "image": "COCO_train2014_000000239263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark red vehicle with a black top\"."}], "task": "refering", "answer_template": "The \"a dark red vehicle with a black top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 131, 132, 144, 145, 146, 147, 148, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 274, 275, 297, 298, 320, 321, 344], "bbox": [0.2645, 0.3595560747663551, 0.999375, 0.9865186915887849], "iscrowd": 0, "area": 40958.991850000006, "rle": {"size": [428, 640], "counts": "[mV22T=5L5J6J6jNZOcEh0[:[OeEe0W:_OiEa0S:CmE=P:FoE;m9JRF5l9NTF2h92XFNe95[FKa99_FF_9=aFC[9a0fF^OW9e0iF[OS9i0mFWOR9j0nFUOR9m0nFROP9P1PGPOo8Q1QGoNn8R1RGmNn8T1SGjNm8W1SGhNm8Y1SGfNm8[1SGeNl8\\1UGbNj8`1VG_Nj8c1UG\\Nk8e1UGZNk8g1VGXNi8i1WGVNh8l1XGSNh8n1YGPNf8R2ZGmMf8T2ZGkMf8V2ZGjMd8Y2\\GeMe8[2[GdMf8\\2ZGcMg8]2T10000000000001O0000000000000000001O0000000000000000001OO1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1OPQe0OQoZO2M2O1N2O1N2O1O1N2O1N3N1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2O1O2M2O1N2O1N2O1O1N2O1N3N1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1O1N3N1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2MgKRGZ4n83O0001O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001O01O000000001O000001OO1O2N1O100O1O2N1O1O1O1O2O01O001O01O01O001O01O01O001O010O00001O010O001O0001n0QO>B1O1O101N1O1O1O2O0001O010O001O001O001O001O001O1O001O001O001O0013L5L3L4M4K2OO0O2N1O2N1O2O1N1O3M2N2N2O2M2N2NlG"}, "label": "a dark red vehicle with a black top"}]}
{"id": 239263, "image": "COCO_train2014_000000239263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow convertable car\"."}], "task": "refering", "answer_template": "The \"yellow convertable car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 213, 214, 215, 216, 217, 218, 219, 230, 231, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 253, 254, 255, 256, 257, 258, 259, 260, 261, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.523714953271028, 0.9769375, 0.9890420560747664], "iscrowd": 0, "area": 71499.61770000002, "rle": {"size": [428, 640], "counts": "d7a5j71O1O2N1O1O1O100O1O1O1O1O1O1O100O10000O100O10000O10000O[LRIn0n6nNWIQ1i6nNYIQ1f6oN]Io0c6PO_IP1`6oNbIP1^6oNdIP1\\6nNgIQ1Y6nNjIP1V6oNlIP1T6oNoIo0Q6PORJn0n5QOUJm0k5ROXJl0h5QO]Jm0c5QObJl0^5QOgJm0Y5QOkJm0U5POoJo0Q5oNRKP1n4mNUKS1k4kNXKT1h4iN[KW1e4gN]KY1c4eN^K\\1b4aNaK_1_4_NcKa1]4\\NfKd1Z4\\NfKd1Z4\\NfKd1Z4\\NfKd1Z4\\NeKe1[4ZNfKf1Z4ZNfKf1Z4ZNfKf1Z4ZNfKf1Z4ZNfKe1[4[NdKe1]4[NcKd1^4\\NbKc1_4]NaKb1`4]NaKb1`4^N_Kb1b4^N^Ka1c4_N]K`1d4`N[K`1f4`NZK_1h4`NXK_1i4aNVK_1k4aNUK^1l4bNTK]1m4bNSK^1n4bNRK]1o4cNQK\\1P5dNoJ[1S5eNmJZ1T5fNlJY1U5gNjJY1W5gNiJX1Y5gNgJX1Z5hNeJX1\\5hNdJV1^5kNaJT1`5lN_JT1b5lN^JS1c5mN]JR1d5nN[JQ1g5oNYJP1h5POWJP1j5POVJo0k5QOTJo0n5POQJo0Q6QOoIn0R6ROmIn0T6SOjIm0W6SOiIl0X6TOgIk0[6UOdIk0]6UOcIj0^6VOaIj0`6VO_Ij0b6VO^Ih0d6XO[Ih0f6XOYIh0i6WOWIh0j6XOUIh0l6YORIf0P7ZOPIe0Q7[OnHe0S7[OlHd0V7\\OjHc0W7]OhHb0Z7^OeHb0\\7^OcHa0_7_OaH?a7A^H?d7@[H?g7AYH>h7BWH>j7BUH>`NDY8NWI<`NNS8F\\I<`N6n7^ObI;_N?i7VOgI:aNe0d7QOkI5dNm0`7nNkI2gNS1]7kNlIOhNY1[7hNlI0gN[1\\7eNmIOfN_1\\7bNmI0eNa1]7_NnIOfNc1[7^NnI0fNd1[7\\NoIOfNg1Z7ZNoI0gNg1Y7YNPJOgNj1Y7VNoI1hNj1X7UNPJ0hNm1W7SNPJ1iNl1W7SNPJ0iNo1V7QNQJ0iNo1V7QNPJ0jNP2V7PNPJ0jNQ2U7oMPJ0kNR2U7nMPJ0kNR2U7nMoI1kNS2U7lMPJ0kNU2U7kMoI1lNT2U7kMoI0lNW2T7iMoI1mNV2T7iMoI1mNV2T7jMnIOoNX2R7iMnI0oNX2S7hMnIOPOY2R7hMmI0QOX2R7hMmIOQOZ2R7gMmINRO[2Q7hMkIMUO[2P7hMkILUO]2P7gMjILWO]2o6gMjIKXO^2o6fMiIKXOa2n6eMhIJ[Oa2m6eMhII\\Ob2l6eMhIH]Oc2k6eMgII^Ob2k6eMgIH_Oc2j6eMfIHAc2i6fMeIFCd2h6fMeIFCd2h6fMdIFEd2g6fMdIEFe2f6fMcIFGd2f6fMcIEHe2e6gMaIDKe2e6fM`IEKe2e6fM`IDLf2d6fM_IDNf2c6fM_ICOg2b6gM]IC1f2b6gM]IB2g2a6gM]I@4i2_6gM\\I_O7j2]6gM\\I^O8k2\\6gM[I]O;l2[6gMYI[O>n2Y6gMYIYOa0o2V6hMXIYOc0o2U6hMXIWOe0Q3S6hMdJX2\\5iMcJW2]5iMdJV2\\5jMdJV2\\5jMdJV2]5iMcJW2]5iMcJW2]5jMbJV2^5jMcJU2]5kMcJU2]5kMcJU2]5kMcJU2]5kMcJU2]5lMcJS2]5mMcJS2]5mMcJS2]5mMdJR2\\5nMdJR2\\5oMcJQ2]5oMdJP2\\5PNdJP2\\5PNdJP2\\5PNeJo1[5QNeJo1[5RNdJn1\\5RNeJm1[5SNeJm1\\5SNcJm1]5SNdJl1\\5TNdJl1\\5UNcJk1]5UNdJj1\\5WNcJi1]5WNdJh1]5WNdJh1\\5YNdJf1\\5ZNeJe1[5[NfJd1Z5]NfJb1[5]NfJb1Z5^NgJa1Y5`NgJ_1Y5aNhJ^1X5bNiJ]1X5cNiJ[1W5eNlJX1T5hNnJV1R5kNPKR1Q5mNQKQ1o4oNTKn0l4SOVKj0j4VOZKf0f4ZO^Kb0c4^O_K?a4AcK;]4EfK8Z4IhK4X4LlK0U4OnKNR43PLJP46SLGm39VLDj3=XLmNXL<a7f0\\L\\Od3d0[L\\Of3e0XL\\Oh3d0VL^Ok3a0TL@l3a0SL_Om3a0SL_Om3a0SL_Om3b0RL^Oo3a0QL_Oo3a0QL_Oo3b0PL^OP4c0PL\\OP4e0hKjNdLa0e7e0dKlNhL>d7g0aKnNkL:d7h0`KoNlL9d7i0]KQOoL5e7j0ZKSOPM3f7k0WKTOTM0e7m0UKUOUMNf7n0RKWOXMJg7o0oJXOZMIg7P1mJYO\\MFg7R1kJYO^MEg7S1iJZO`MBh7S1hJ\\O`M@h7U1fJ\\OcM^Og7W1dJ]OeM[Oh7X1aJ_OhMVOh7\\1^JLb56[JKe56YJKh56UJKk56TJJl58QJIo58oIIR67lIJT68hIJX67fIJZ68cII^67`IJ`68]IIc68ZIJf68WIIj68SIIm69PIHP7:mHGT7:iHGW7;gHEY7=dHD\\7>aHC`7>]HCc7?ZHBf7a0VH@j7b0SH_On7b0PH^OP8d0mG]OT8d0iG]OX8k21O1O1O1O1O1O1O1O2N1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OO1001O000000000000001O000000000000000000001O0000000000000000001O000000000iMaFa0_9AbF<^9EdF8\\9IeF5\\9LeF1[90fFNZ93gFKY97gFGY9:hFDX9=jF@V9a0kF]OU9e0kFYOU9h0lFVOU9k0kFSOU9P1jFnNV9U1iFiNW9Z1hFdNX9_1gF_NY9d1eF[N[9g1eFWN[9l1dFRN]9P2bFnM^9U2aFiM_9Z2`FdM`9_2_F_Ma9c2^F\\Mb9f2^FXMc9h2^FVMb9l2^FRMb9P3^FnLb9T3]FkLc9V3]FiLd9_300000O1100O000001O000000001O00000000000000000000000000000000000O10000000001O0000000000001O0000000000001O00000000001O00000O100000000001O00000000000O10000000000000000000000O1001O000000001O0001O0001O0000001O00000000001O000000000000001O00000000O100O1O1O1O1O1O1O1O1O1O100O100O100O100O100001O00001O00001O00001O00001O001O1O1O1O1O3M2N2N6jMnEf0]:jNhEV1`:^NeEa1\\:[NgEe1P;0001O1O2N1O1O1O001O1O3M00O100001O1O1O1O1O2N1O2N1O2M3N1N2O1N2M3M7IUl5"}, "label": "yellow convertable car"}]}
{"id": 239263, "image": "COCO_train2014_000000239263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow and black convertible car with an orange striped surfboard in the back seat\"."}], "task": "refering", "answer_template": "The \"a yellow and black convertible car with an orange striped surfboard in the back seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 213, 214, 215, 216, 217, 218, 219, 230, 231, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 253, 254, 255, 256, 257, 258, 259, 260, 261, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.523714953271028, 0.9769375, 0.9890420560747664], "iscrowd": 0, "area": 71499.61770000002, "rle": {"size": [428, 640], "counts": "d7a5j71O1O2N1O1O1O100O1O1O1O1O1O1O100O10000O100O10000O10000O[LRIn0n6nNWIQ1i6nNYIQ1f6oN]Io0c6PO_IP1`6oNbIP1^6oNdIP1\\6nNgIQ1Y6nNjIP1V6oNlIP1T6oNoIo0Q6PORJn0n5QOUJm0k5ROXJl0h5QO]Jm0c5QObJl0^5QOgJm0Y5QOkJm0U5POoJo0Q5oNRKP1n4mNUKS1k4kNXKT1h4iN[KW1e4gN]KY1c4eN^K\\1b4aNaK_1_4_NcKa1]4\\NfKd1Z4\\NfKd1Z4\\NfKd1Z4\\NfKd1Z4\\NeKe1[4ZNfKf1Z4ZNfKf1Z4ZNfKf1Z4ZNfKf1Z4ZNfKe1[4[NdKe1]4[NcKd1^4\\NbKc1_4]NaKb1`4]NaKb1`4^N_Kb1b4^N^Ka1c4_N]K`1d4`N[K`1f4`NZK_1h4`NXK_1i4aNVK_1k4aNUK^1l4bNTK]1m4bNSK^1n4bNRK]1o4cNQK\\1P5dNoJ[1S5eNmJZ1T5fNlJY1U5gNjJY1W5gNiJX1Y5gNgJX1Z5hNeJX1\\5hNdJV1^5kNaJT1`5lN_JT1b5lN^JS1c5mN]JR1d5nN[JQ1g5oNYJP1h5POWJP1j5POVJo0k5QOTJo0n5POQJo0Q6QOoIn0R6ROmIn0T6SOjIm0W6SOiIl0X6TOgIk0[6UOdIk0]6UOcIj0^6VOaIj0`6VO_Ij0b6VO^Ih0d6XO[Ih0f6XOYIh0i6WOWIh0j6XOUIh0l6YORIf0P7ZOPIe0Q7[OnHe0S7[OlHd0V7\\OjHc0W7]OhHb0Z7^OeHb0\\7^OcHa0_7_OaH?a7A^H?d7@[H?g7AYH>h7BWH>j7BUH>`NDY8NWI<`NNS8F\\I<`N6n7^ObI;_N?i7VOgI:aNe0d7QOkI5dNm0`7nNkI2gNS1]7kNlIOhNY1[7hNlI0gN[1\\7eNmIOfN_1\\7bNmI0eNa1]7_NnIOfNc1[7^NnI0fNd1[7\\NoIOfNg1Z7ZNoI0gNg1Y7YNPJOgNj1Y7VNoI1hNj1X7UNPJ0hNm1W7SNPJ1iNl1W7SNPJ0iNo1V7QNQJ0iNo1V7QNPJ0jNP2V7PNPJ0jNQ2U7oMPJ0kNR2U7nMPJ0kNR2U7nMoI1kNS2U7lMPJ0kNU2U7kMoI1lNT2U7kMoI0lNW2T7iMoI1mNV2T7iMoI1mNV2T7jMnIOoNX2R7iMnI0oNX2S7hMnIOPOY2R7hMmI0QOX2R7hMmIOQOZ2R7gMmINRO[2Q7hMkIMUO[2P7hMkILUO]2P7gMjILWO]2o6gMjIKXO^2o6fMiIKXOa2n6eMhIJ[Oa2m6eMhII\\Ob2l6eMhIH]Oc2k6eMgII^Ob2k6eMgIH_Oc2j6eMfIHAc2i6fMeIFCd2h6fMeIFCd2h6fMdIFEd2g6fMdIEFe2f6fMcIFGd2f6fMcIEHe2e6gMaIDKe2e6fM`IEKe2e6fM`IDLf2d6fM_IDNf2c6fM_ICOg2b6gM]IC1f2b6gM]IB2g2a6gM]I@4i2_6gM\\I_O7j2]6gM\\I^O8k2\\6gM[I]O;l2[6gMYI[O>n2Y6gMYIYOa0o2V6hMXIYOc0o2U6hMXIWOe0Q3S6hMdJX2\\5iMcJW2]5iMdJV2\\5jMdJV2\\5jMdJV2]5iMcJW2]5iMcJW2]5jMbJV2^5jMcJU2]5kMcJU2]5kMcJU2]5kMcJU2]5kMcJU2]5lMcJS2]5mMcJS2]5mMcJS2]5mMdJR2\\5nMdJR2\\5oMcJQ2]5oMdJP2\\5PNdJP2\\5PNdJP2\\5PNeJo1[5QNeJo1[5RNdJn1\\5RNeJm1[5SNeJm1\\5SNcJm1]5SNdJl1\\5TNdJl1\\5UNcJk1]5UNdJj1\\5WNcJi1]5WNdJh1]5WNdJh1\\5YNdJf1\\5ZNeJe1[5[NfJd1Z5]NfJb1[5]NfJb1Z5^NgJa1Y5`NgJ_1Y5aNhJ^1X5bNiJ]1X5cNiJ[1W5eNlJX1T5hNnJV1R5kNPKR1Q5mNQKQ1o4oNTKn0l4SOVKj0j4VOZKf0f4ZO^Kb0c4^O_K?a4AcK;]4EfK8Z4IhK4X4LlK0U4OnKNR43PLJP46SLGm39VLDj3=XLmNXL<a7f0\\L\\Od3d0[L\\Of3e0XL\\Oh3d0VL^Ok3a0TL@l3a0SL_Om3a0SL_Om3a0SL_Om3b0RL^Oo3a0QL_Oo3a0QL_Oo3b0PL^OP4c0PL\\OP4e0hKjNdLa0e7e0dKlNhL>d7g0aKnNkL:d7h0`KoNlL9d7i0]KQOoL5e7j0ZKSOPM3f7k0WKTOTM0e7m0UKUOUMNf7n0RKWOXMJg7o0oJXOZMIg7P1mJYO\\MFg7R1kJYO^MEg7S1iJZO`MBh7S1hJ\\O`M@h7U1fJ\\OcM^Og7W1dJ]OeM[Oh7X1aJ_OhMVOh7\\1^JLb56[JKe56YJKh56UJKk56TJJl58QJIo58oIIR67lIJT68hIJX67fIJZ68cII^67`IJ`68]IIc68ZIJf68WIIj68SIIm69PIHP7:mHGT7:iHGW7;gHEY7=dHD\\7>aHC`7>]HCc7?ZHBf7a0VH@j7b0SH_On7b0PH^OP8d0mG]OT8d0iG]OX8k21O1O1O1O1O1O1O1O2N1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OO1001O000000000000001O000000000000000000001O0000000000000000001O000000000iMaFa0_9AbF<^9EdF8\\9IeF5\\9LeF1[90fFNZ93gFKY97gFGY9:hFDX9=jF@V9a0kF]OU9e0kFYOU9h0lFVOU9k0kFSOU9P1jFnNV9U1iFiNW9Z1hFdNX9_1gF_NY9d1eF[N[9g1eFWN[9l1dFRN]9P2bFnM^9U2aFiM_9Z2`FdM`9_2_F_Ma9c2^F\\Mb9f2^FXMc9h2^FVMb9l2^FRMb9P3^FnLb9T3]FkLc9V3]FiLd9_300000O1100O000001O000000001O00000000000000000000000000000000000O10000000001O0000000000001O0000000000001O00000000001O00000O100000000001O00000000000O10000000000000000000000O1001O000000001O0001O0001O0000001O00000000001O000000000000001O00000000O100O1O1O1O1O1O1O1O1O1O100O100O100O100O100001O00001O00001O00001O00001O001O1O1O1O1O3M2N2N6jMnEf0]:jNhEV1`:^NeEa1\\:[NgEe1P;0001O1O2N1O1O1O001O1O3M00O100001O1O1O1O1O2N1O2N1O2M3N1N2O1N2M3M7IUl5"}, "label": "a yellow and black convertible car with an orange striped surfboard in the back seat"}]}
{"id": 296613, "image": "COCO_train2014_000000296613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chocolate cake next to the ' july ' sign\"."}], "task": "refering", "answer_template": "The \"the chocolate cake next to the ' july ' sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 287], "bbox": [0.31710937499999997, 0.20604166666666668, 0.74496875, 0.7116250000000001], "iscrowd": 0, "area": 49478.66464999999, "rle": {"size": [480, 640], "counts": "jYo26i>=B=D<C=D<D<C4M3L5L3L4M4L3L4M4K4M3M4K4M3L5L3M3L5L3L4M4L3L4M4K4M3L5L3M4K4M3L5L3N2O2M2O1N3NI6H9O0100O010O010O010O010O10O01O010O010O010O10O010O010O010O0100O010O010O010O010O1O010O010O010O010O10O010O010O010O0100O010O010O010O0010O0100O010O010OO2O001O001O001O0O01000000000000O10O1000000000O100000O1000O1000000000O10O10000000000O10O1000000000O100O1O010O1O1O100O1O10O01O100O1O0010O000010O0001O01O0001O01O01O01O01O0001O01O00010O0000010O00010O0000101N2N2N2O0O2N2N2O1N2N101N2N2N2O0O2N2N2O1N2N1O2O1N2N2O0O2N2M3K5L4K4M4K5L4K5L3L5L4K5L7I7H9H8G9H8G9H8G9H7H9H8G9HPW\\2"}, "label": "the chocolate cake next to the ' july ' sign"}]}
{"id": 296613, "image": "COCO_train2014_000000296613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cutting chocolate cake kept in a plate with fork and also kept on the table\"."}], "task": "refering", "answer_template": "The \"a cutting chocolate cake kept in a plate with fork and also kept on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 287], "bbox": [0.31710937499999997, 0.20604166666666668, 0.74496875, 0.7116250000000001], "iscrowd": 0, "area": 49478.66464999999, "rle": {"size": [480, 640], "counts": "jYo26i>=B=D<C=D<D<C4M3L5L3L4M4L3L4M4K4M3M4K4M3L5L3M3L5L3L4M4L3L4M4K4M3L5L3M4K4M3L5L3N2O2M2O1N3NI6H9O0100O010O010O010O010O10O01O010O010O010O10O010O010O010O0100O010O010O010O010O1O010O010O010O010O10O010O010O010O0100O010O010O010O0010O0100O010O010OO2O001O001O001O0O01000000000000O10O1000000000O100000O1000O1000000000O10O10000000000O10O1000000000O100O1O010O1O1O100O1O10O01O100O1O0010O000010O0001O01O0001O01O01O01O01O0001O01O00010O0000010O00010O0000101N2N2N2O0O2N2N2O1N2N101N2N2N2O0O2N2N2O1N2N1O2O1N2N2O0O2N2M3K5L4K4M4K5L4K5L3L5L4K5L7I7H9H8G9H8G9H8G9H7H9H8G9HPW\\2"}, "label": "a cutting chocolate cake kept in a plate with fork and also kept on the table"}]}
{"id": 18089, "image": "COCO_train2014_000000018089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut with burnt cheese\"."}], "task": "refering", "answer_template": "The \"donut with burnt cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 281, 282, 283], "bbox": [0.4037029288702929, 0.44965625, 0.7201882845188284, 0.710546875], "iscrowd": 0, "area": 19371.4246, "rle": {"size": [640, 478], "counts": "`oh36`c0<G:J5L5K4K5L5K4K6K4L4K5L4L3L5M3L3M4L4M2M4L4L3N3M3N2N1N3N1O1N101O0O2O000O2O001N101O0O2O001N10001N101N101N101N100O2O0O2O0O2O0O2O0O101O001O00001O001O001O00001O010O0000L4001O0000000010O001O0010O01O001O001O001O0010O01O1O1O001O1O001O10O02N1O2N1O1O2N1O1O100O1O1N2N3M2N2O1N3M2N3M2K6J5K6K6K5J6K5J6K6I7J5I8F9Hd\\c2"}, "label": "donut with burnt cheese"}]}
{"id": 18089, "image": "COCO_train2014_000000018089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut that has some red crushed topping\"."}], "task": "refering", "answer_template": "The \"a donut that has some red crushed topping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 281, 282, 283], "bbox": [0.4037029288702929, 0.44965625, 0.7201882845188284, 0.710546875], "iscrowd": 0, "area": 19371.4246, "rle": {"size": [640, 478], "counts": "`oh36`c0<G:J5L5K4K5L5K4K6K4L4K5L4L3L5M3L3M4L4M2M4L4L3N3M3N2N1N3N1O1N101O0O2O000O2O001N101O0O2O001N10001N101N101N101N100O2O0O2O0O2O0O2O0O101O001O00001O001O001O00001O010O0000L4001O0000000010O001O0010O01O001O001O001O0010O01O1O1O001O1O001O10O02N1O2N1O1O2N1O1O100O1O1N2N3M2N2O1N3M2N3M2K6J5K6K6K5J6K5J6K6I7J5I8F9Hd\\c2"}, "label": "a donut that has some red crushed topping"}]}
{"id": 18089, "image": "COCO_train2014_000000018089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"top light - brown donut in box\"."}], "task": "refering", "answer_template": "The \"top light - brown donut in box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 225, 226, 227], "bbox": [0.19947698744769873, 0.376109375, 0.4970083682008368, 0.5940937500000001], "iscrowd": 0, "area": 13771.335750000002, "rle": {"size": [640, 478], "counts": "mek11hc0:F:F:F:E;J5M4L3L4M3L5L3M3L4M2N1N3N2M2O2N2O0O2O0O2O1N101N2O0O101N101N100O2O0O101N101N10001N10001O001O0O101O001O000O2O00001O000O101O0000000O101O0000000O2O00000000001O01O000001O0001O0O1N2N2M4M2M4M2N3L3N3M2N3M3M2N3L3N3M2N3M3M2N3M2N3M2M4M3M3M2N3M3M3N2O0O2N2N2N2N1O2N2O1N3M3M2N3M3M3M3N1N3MTge4"}, "label": "top light - brown donut in box"}]}
{"id": 18089, "image": "COCO_train2014_000000018089.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut on the top left\"."}], "task": "refering", "answer_template": "The \"the donut on the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 225, 226, 227], "bbox": [0.19947698744769873, 0.376109375, 0.4970083682008368, 0.5940937500000001], "iscrowd": 0, "area": 13771.335750000002, "rle": {"size": [640, 478], "counts": "mek11hc0:F:F:F:E;J5M4L3L4M3L5L3M3L4M2N1N3N2M2O2N2O0O2O0O2O1N101N2O0O101N101N100O2O0O101N101N10001N10001O001O0O101O001O000O2O00001O000O101O0000000O101O0000000O2O00000000001O01O000001O0001O0O1N2N2M4M2M4M2N3L3N3M2N3M3M2N3L3N3M2N3M3M2N3M2N3M2M4M3M3M2N3M3M3N2O0O2N2N2N2N1O2N2O1N3M3M2N3M3M3M3N1N3MTge4"}, "label": "the donut on the top left"}]}
{"id": 517805, "image": "COCO_train2014_000000517805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black lady\"."}], "task": "refering", "answer_template": "The \"black lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317, 334, 335, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 383, 384, 385, 386], "bbox": [0.53209375, 0.5524583333333334, 0.822703125, 1.0], "iscrowd": 0, "area": 11433.771899999996, "rle": {"size": [480, 640], "counts": "jWP52l>3M7I3O101N3N0O2N100O2OOIYOUBh0g=\\OYBc0f=_OYBb0f=_OYBa0h=_OXBa0h=_OWBa0i=_OXB`0i=@VB`0k=@UB?k=BTB?l=ASB?n=9001O0013MO010O1O1K500O100O1O100O1O15K00O1O00O100001O001O010O2N1O01O000010O0001O0000010O00001O01O01O000O11O10O00000O2O00000010O00000001O00000010O00O101Ne0ZO2D<00000000000000fL[NZId1`6fN\\IZ1b6jNYHYOZOm1[8nNXHVO]Ol1Y8WORHnNEk1W8^OmGiNLi1U8AmGhNMg1U8BmGhNNf1T8DkGhN1d1R8GjGhN3a1R8JfGhN8^1P8NdGgN;[1P8V1PHjNn7X1RHhNm7Z1RHfNm7[1SHeNl7\\1THdNj7_1UHaNj7`1VH`Ni7a1WH_Nh7b1XHhMWOZO_8P3ZHdMYO\\O\\8Q3\\H_M\\O_OW8S3]H\\M^OAS8U3_HXM@Co7W3aHVMABn7X3aHVMABn7X3aHVMBSOH5U8a3bHWMBROH5T8b3bHWMBROH5S8c3cHVMCQOG6S8c3cHWMBPOH6S8c3cHWMCPOG5S8d3cHWMCPOG5S8d3cHWMCPOG5S8d3cHoLUO]O>KG5S8d3cHkLYOB9JI4R8d3dHkLZOB7KI4Q8e3fHgL\\OD=0a7d3gHgL\\OD>0_7e3gHfL\\OF>O_7d3iHdL\\OH?N]7d3iHeL\\OIj8b3kGcL\\OJk8P2gF\\OS1iN\\OKj8o1jFZOQ1lNZOKm8l1mFZOm0mNZOMl8k1oFYOl0oNXOMn8i1QGYOj0QOVONP9f1QG[Oj0POh8c1aF[Oh0ROh8`1cF]Oe0SOi8]1dF_Oc0UOj8Y1fFA`0VOk8V1hFC=WOl8m0oFK6XOm8i0PGN3YOn8g0RGM1\\On8d0TGON]OY;`0hD@Y;?fDB[;=eDC];:dDF5Jd9?WFG4Kf9=VFF4Oi98SFG43m94mEG76P:0ZG3g;2O11O2N3M2N2NSod1"}, "label": "black lady"}]}
{"id": 517805, "image": "COCO_train2014_000000517805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl being hugged by a man in a black tshirt\"."}], "task": "refering", "answer_template": "The \"a young girl being hugged by a man in a black tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317, 334, 335, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 383, 384, 385, 386], "bbox": [0.53209375, 0.5524583333333334, 0.822703125, 1.0], "iscrowd": 0, "area": 11433.771899999996, "rle": {"size": [480, 640], "counts": "jWP52l>3M7I3O101N3N0O2N100O2OOIYOUBh0g=\\OYBc0f=_OYBb0f=_OYBa0h=_OXBa0h=_OWBa0i=_OXB`0i=@VB`0k=@UB?k=BTB?l=ASB?n=9001O0013MO010O1O1K500O100O1O100O1O15K00O1O00O100001O001O010O2N1O01O000010O0001O0000010O00001O01O01O000O11O10O00000O2O00000010O00000001O00000010O00O101Ne0ZO2D<00000000000000fL[NZId1`6fN\\IZ1b6jNYHYOZOm1[8nNXHVO]Ol1Y8WORHnNEk1W8^OmGiNLi1U8AmGhNMg1U8BmGhNNf1T8DkGhN1d1R8GjGhN3a1R8JfGhN8^1P8NdGgN;[1P8V1PHjNn7X1RHhNm7Z1RHfNm7[1SHeNl7\\1THdNj7_1UHaNj7`1VH`Ni7a1WH_Nh7b1XHhMWOZO_8P3ZHdMYO\\O\\8Q3\\H_M\\O_OW8S3]H\\M^OAS8U3_HXM@Co7W3aHVMABn7X3aHVMABn7X3aHVMBSOH5U8a3bHWMBROH5T8b3bHWMBROH5S8c3cHVMCQOG6S8c3cHWMBPOH6S8c3cHWMCPOG5S8d3cHWMCPOG5S8d3cHWMCPOG5S8d3cHoLUO]O>KG5S8d3cHkLYOB9JI4R8d3dHkLZOB7KI4Q8e3fHgL\\OD=0a7d3gHgL\\OD>0_7e3gHfL\\OF>O_7d3iHdL\\OH?N]7d3iHeL\\OIj8b3kGcL\\OJk8P2gF\\OS1iN\\OKj8o1jFZOQ1lNZOKm8l1mFZOm0mNZOMl8k1oFYOl0oNXOMn8i1QGYOj0QOVONP9f1QG[Oj0POh8c1aF[Oh0ROh8`1cF]Oe0SOi8]1dF_Oc0UOj8Y1fFA`0VOk8V1hFC=WOl8m0oFK6XOm8i0PGN3YOn8g0RGM1\\On8d0TGON]OY;`0hD@Y;?fDB[;=eDC];:dDF5Jd9?WFG4Kf9=VFF4Oi98SFG43m94mEG76P:0ZG3g;2O11O2N3M2N2NSod1"}, "label": "a young girl being hugged by a man in a black tshirt"}]}
{"id": 517805, "image": "COCO_train2014_000000517805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with sunglasses\"."}], "task": "refering", "answer_template": "The \"man with sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 379, 380, 381, 382, 383], "bbox": [0.48539062499999996, 0.41304166666666664, 0.82415625, 0.9887708333333334], "iscrowd": 0, "area": 28160.062950000018, "rle": {"size": [480, 640], "counts": "eWb43j>3L4M3L4M3L5L3M3L4M3L4M3L4K5L4K5K5K5K5K5K5K5K5aN_1L4K5K5L4M3NQNcF_N[9a1hF^NW9c1jF\\NT9f1mFYNQ9i1PGVNn8m1QGSNn8n1SGQNk8Q2VGnMh8T2YGkMe8X2[GgMc8[2^GdMa8]2`GbM^8a2aG_M]8c2dG\\MZ8^2RG^Le0S1X8`2UG\\Lc0S1V8c2XGYLc0S1S8f2\\GULa0U1Q8h2_GRL`0V1o7j2cGoK>V1n7l2\\HTMb7n2^HRM`7P3aHoL]7S3cHmL]7S3cHmL\\7U3dHjL[7X3dHhL\\7Y3cHgL\\7[3dHdL[7^3dHbL\\7_3cHaL\\7a3dH^L[7d3dH\\L\\7f3bHZL]7h3cHWL]7j3bHVL]7l3bHUL\\7m3dHRL\\7o3cHQL]7P4cHoK]7R4bHnK^7S4bHlK_7T4aHkK_7V4aHiK_7X4aHgK`7X4aHgK_7Y4aHgK_7Y4bHfK_7Y4bHfK^7Z4cHeK]7\\4cHcK^7\\4cHcK]7]4cHcK]7]4cHcK]7]4dHbK]7]4cHcK]7]4cHcK]7^4bHbK_7]4bHbK^7^4bHbK^7^4bHbK_7]4bHbK^7^4bHbK^7^4bHbK^7^4bHbK_7^4aHaK_7_4aHaK_7_4aHaK`7^4aHaK_7_4aHaK^7`4bH`K]7a4cH_K[7c4fH\\KX7g4gHYKX7h4hHXKV7j4kHUKT7l4lHTKR7n4nHRKP7P5PIPKg6Y5ZIfJb6^5^IbJ^6c5aI]J[6g5fIXJW6k5iIUJW6k5iIUJV6l5jITJU6m5lIRJT6n5lIRJS6o5mIQJR6Q6nInIP6T6PJlIo5U6QJkIn5V6RJjIl5X6UJgIj5Z6VJfIi5[6WJeIg5^6YJbIe5_6[JaId5`6\\J`Ic5a6]J_Ib5b6_J]I`5T6^IPJT1J]5W6gIgIg7Z6>XObGPK]8P5dGPK[8P5fGPKY8g2lGbNLgNW8c2THcNFjNU8_2]HcN_OoNR8\\2eHaNZOSOQ8Z2iH`NWOVOo7Z2nH\\NTOZOn7X2SIZNPO^On7V2VIXNmNBm7U2YIUNlNFk7S2^ISNhNKi7Q2cIPNeNOi7i1iGaMm1d0bN2h7h1iGbMQ2`0_N6g7g1iGbMV2OiNh0X7f1iGcMY3g0n4e1iGcM[3i0l4b1iGeM\\3i0k4`1jGfM]3j0i4_1jGgM^3j0h4^1\\KbNd4]1]KcNd4[1]KfNb4Y1_KgNa4W1aKiN_4V1bKjN_4T1bKlN_4R1bKnN^4Q1cKoN^4o0cKRO]4l0dKTO]4i0eKWO[4h0fKXO[4f0fKZO[4d0fK\\OZ4c0gK^OY4`0hK@Y4=iKCX48lKHW41mKOU4KoK5T4DPL=Q4]OSLc0o3WOTLj0Q90O100O2O0O100O01000O100O100O10O010000O100O01O0O2N1O2O0QO\\Bb0T>L01N1O1002O1N3MUcd1"}, "label": "man with sunglasses"}]}
{"id": 517805, "image": "COCO_train2014_000000517805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man hugging a girl near a giraffe\"."}], "task": "refering", "answer_template": "The \"a man hugging a girl near a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 379, 380, 381, 382, 383], "bbox": [0.48539062499999996, 0.41304166666666664, 0.82415625, 0.9887708333333334], "iscrowd": 0, "area": 28160.062950000018, "rle": {"size": [480, 640], "counts": "eWb43j>3L4M3L4M3L5L3M3L4M3L4M3L4K5L4K5K5K5K5K5K5K5K5aN_1L4K5K5L4M3NQNcF_N[9a1hF^NW9c1jF\\NT9f1mFYNQ9i1PGVNn8m1QGSNn8n1SGQNk8Q2VGnMh8T2YGkMe8X2[GgMc8[2^GdMa8]2`GbM^8a2aG_M]8c2dG\\MZ8^2RG^Le0S1X8`2UG\\Lc0S1V8c2XGYLc0S1S8f2\\GULa0U1Q8h2_GRL`0V1o7j2cGoK>V1n7l2\\HTMb7n2^HRM`7P3aHoL]7S3cHmL]7S3cHmL\\7U3dHjL[7X3dHhL\\7Y3cHgL\\7[3dHdL[7^3dHbL\\7_3cHaL\\7a3dH^L[7d3dH\\L\\7f3bHZL]7h3cHWL]7j3bHVL]7l3bHUL\\7m3dHRL\\7o3cHQL]7P4cHoK]7R4bHnK^7S4bHlK_7T4aHkK_7V4aHiK_7X4aHgK`7X4aHgK_7Y4aHgK_7Y4bHfK_7Y4bHfK^7Z4cHeK]7\\4cHcK^7\\4cHcK]7]4cHcK]7]4cHcK]7]4dHbK]7]4cHcK]7]4cHcK]7^4bHbK_7]4bHbK^7^4bHbK^7^4bHbK_7]4bHbK^7^4bHbK^7^4bHbK^7^4bHbK_7^4aHaK_7_4aHaK_7_4aHaK`7^4aHaK_7_4aHaK^7`4bH`K]7a4cH_K[7c4fH\\KX7g4gHYKX7h4hHXKV7j4kHUKT7l4lHTKR7n4nHRKP7P5PIPKg6Y5ZIfJb6^5^IbJ^6c5aI]J[6g5fIXJW6k5iIUJW6k5iIUJV6l5jITJU6m5lIRJT6n5lIRJS6o5mIQJR6Q6nInIP6T6PJlIo5U6QJkIn5V6RJjIl5X6UJgIj5Z6VJfIi5[6WJeIg5^6YJbIe5_6[JaId5`6\\J`Ic5a6]J_Ib5b6_J]I`5T6^IPJT1J]5W6gIgIg7Z6>XObGPK]8P5dGPK[8P5fGPKY8g2lGbNLgNW8c2THcNFjNU8_2]HcN_OoNR8\\2eHaNZOSOQ8Z2iH`NWOVOo7Z2nH\\NTOZOn7X2SIZNPO^On7V2VIXNmNBm7U2YIUNlNFk7S2^ISNhNKi7Q2cIPNeNOi7i1iGaMm1d0bN2h7h1iGbMQ2`0_N6g7g1iGbMV2OiNh0X7f1iGcMY3g0n4e1iGcM[3i0l4b1iGeM\\3i0k4`1jGfM]3j0i4_1jGgM^3j0h4^1\\KbNd4]1]KcNd4[1]KfNb4Y1_KgNa4W1aKiN_4V1bKjN_4T1bKlN_4R1bKnN^4Q1cKoN^4o0cKRO]4l0dKTO]4i0eKWO[4h0fKXO[4f0fKZO[4d0fK\\OZ4c0gK^OY4`0hK@Y4=iKCX48lKHW41mKOU4KoK5T4DPL=Q4]OSLc0o3WOTLj0Q90O100O2O0O100O01000O100O100O10O010000O100O01O0O2N1O2O0QO\\Bb0T>L01N1O1002O1N3MUcd1"}, "label": "a man hugging a girl near a giraffe"}]}
{"id": 231087, "image": "COCO_train2014_000000231087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana on the right\"."}], "task": "refering", "answer_template": "The \"the banana on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 38, 39, 40, 41, 42, 60, 61, 62, 63, 64, 83, 84, 85, 86, 105, 106, 107, 108, 127, 128, 129, 130, 149, 150, 151, 171, 172], "bbox": [0.7105882352941176, 0.0, 0.9223529411764706, 0.33882352941176475], "iscrowd": 0, "area": 16741.209600000006, "rle": {"size": [612, 612], "counts": "\\oS81Sc02N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N3M4L4L4M3L4L4L4L4L4L3M4L4L4Li2WM?A1O0000000000001O0000000000001O000000000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2N2N2M3N2M4M2M3M3N3L4M4K4M3L5L3L4L5L3L4M4K4M8G;B?WOh0YOg0XOjcl0"}, "label": "the banana on the right"}]}
{"id": 231087, "image": "COCO_train2014_000000231087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana is sitting on the outer right side of the grill\"."}], "task": "refering", "answer_template": "The \"a banana is sitting on the outer right side of the grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 38, 39, 40, 41, 42, 60, 61, 62, 63, 64, 83, 84, 85, 86, 105, 106, 107, 108, 127, 128, 129, 130, 149, 150, 151, 171, 172], "bbox": [0.7105882352941176, 0.0, 0.9223529411764706, 0.33882352941176475], "iscrowd": 0, "area": 16741.209600000006, "rle": {"size": [612, 612], "counts": "\\oS81Sc02N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N3M4L4L4M3L4L4L4L4L4L3M4L4L4Li2WM?A1O0000000000001O0000000000001O000000000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2N2N2M3N2M4M2M3M3N3L4M4K4M3L5L3L4L5L3L4M4K4M8G;B?WOh0YOg0XOjcl0"}, "label": "a banana is sitting on the outer right side of the grill"}]}
{"id": 231087, "image": "COCO_train2014_000000231087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second banana from the right on the grill\"."}], "task": "refering", "answer_template": "The \"the second banana from the right on the grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 168, 169, 170, 190, 191, 192], "bbox": [0.5035294117647059, 0.0, 0.7811764705882354, 0.3952941176470588], "iscrowd": 0, "area": 20913.292799999996, "rle": {"size": [612, 612], "counts": "leh52Rc02N1O1O1O2N1O1O2N1O1O2M2O1O2N1O1O1O2N1O1O2N1O1N3N1O1O2N1O1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O1O2N1O1N3N1O2N1O1O3M2N3M2N3M2N3M2M4M2N3M2N3M2N3M2N3M4L5K5J6K5K5K5K5K5K3M3M3M3M4L3L4M3M3M3M3M3M4L3M3M3M3M3L4M4L1O1O1O100100O001O1O1O001N2N1N3N2N2N1O2N2N2M2O2N2N2N1O2N2M3N1O2N2N2N2N1N3N2N2N2N2N2M2N3M3M3M3M3M3M2N3M3M4L4L4L3M4L4L4L4L3L5H8I7I6J7I<Ba0_O`0@`0_Olm_2"}, "label": "the second banana from the right on the grill"}]}
{"id": 231087, "image": "COCO_train2014_000000231087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana second from the right on the grill\"."}], "task": "refering", "answer_template": "The \"the banana second from the right on the grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 168, 169, 170, 190, 191, 192], "bbox": [0.5035294117647059, 0.0, 0.7811764705882354, 0.3952941176470588], "iscrowd": 0, "area": 20913.292799999996, "rle": {"size": [612, 612], "counts": "leh52Rc02N1O1O1O2N1O1O2N1O1O2M2O1O2N1O1O1O2N1O1O2N1O1N3N1O1O2N1O1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O1O2N1O1N3N1O2N1O1O3M2N3M2N3M2N3M2M4M2N3M2N3M2N3M2N3M4L5K5J6K5K5K5K5K5K3M3M3M3M4L3L4M3M3M3M3M3M4L3M3M3M3M3L4M4L1O1O1O100100O001O1O1O001N2N1N3N2N2N1O2N2N2M2O2N2N2N1O2N2M3N1O2N2N2N2N1N3N2N2N2N2N2M2N3M3M3M3M3M3M2N3M3M4L4L4L3M4L4L4L4L3L5H8I7I6J7I<Ba0_O`0@`0_Olm_2"}, "label": "the banana second from the right on the grill"}]}
{"id": 231087, "image": "COCO_train2014_000000231087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a first yellow color banana is ready to serve\"."}], "task": "refering", "answer_template": "The \"a first yellow color banana is ready to serve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 26, 27, 28, 29, 48, 49, 50, 51, 70, 71, 72, 73, 92, 93, 94, 95, 114, 115, 116, 117, 136, 137, 138, 139, 140, 159, 160, 161, 162, 163, 181, 182, 183, 184, 185, 186, 204, 205, 206, 207, 208, 227, 228, 229, 230], "bbox": [0.1909640522875817, 0.0, 0.4908986928104575, 0.4998856209150327], "iscrowd": 0, "area": 26613.084049999998, "rle": {"size": [612, 612], "counts": "`oU2^1da0g0XO6K4K6J6K4K6J5L5J6J5L5J5K6K5J5L5J6J5L5J5K6K5M2N3M2N3M3M2N3M2N3M3M2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N]J^Ee2a:WMlEa2S:ZMZF_2e9\\MhF]2W9^MVG[2i8aMbGY2]8bMkG]2S8^MTH`2l7[M\\Hc2c7YMcHe2]7VMkHh2T7SMTIj2l6QM[In2d6oLbIn2^6oLhIo2W6nLnIP3R6mLTJQ3k5mLYJQ3g5lL_JR3`5kLfJR3Z5lLjJS3U5jLQKS3o4jLWKT3h4iL]KU3c4iLbKU3X:N3N1N100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10O01O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2M3N2M3M3N2M3M3N2M6K6I7I7J\\Vj5"}, "label": "a first yellow color banana is ready to serve"}]}
{"id": 231087, "image": "COCO_train2014_000000231087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the farthest left banana of the bunch\"."}], "task": "refering", "answer_template": "The \"the farthest left banana of the bunch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 26, 27, 28, 29, 48, 49, 50, 51, 70, 71, 72, 73, 92, 93, 94, 95, 114, 115, 116, 117, 136, 137, 138, 139, 140, 159, 160, 161, 162, 163, 181, 182, 183, 184, 185, 186, 204, 205, 206, 207, 208, 227, 228, 229, 230], "bbox": [0.1909640522875817, 0.0, 0.4908986928104575, 0.4998856209150327], "iscrowd": 0, "area": 26613.084049999998, "rle": {"size": [612, 612], "counts": "`oU2^1da0g0XO6K4K6J6K4K6J5L5J6J5L5J5K6K5J5L5J6J5L5J5K6K5M2N3M2N3M3M2N3M2N3M3M2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N]J^Ee2a:WMlEa2S:ZMZF_2e9\\MhF]2W9^MVG[2i8aMbGY2]8bMkG]2S8^MTH`2l7[M\\Hc2c7YMcHe2]7VMkHh2T7SMTIj2l6QM[In2d6oLbIn2^6oLhIo2W6nLnIP3R6mLTJQ3k5mLYJQ3g5lL_JR3`5kLfJR3Z5lLjJS3U5jLQKS3o4jLWKT3h4iL]KU3c4iLbKU3X:N3N1N100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10O01O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2M3N2M3M3N2M3M3N2M6K6I7I7J\\Vj5"}, "label": "the farthest left banana of the bunch"}]}
{"id": 280241, "image": "COCO_train2014_000000280241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left hand most sheep\"."}], "task": "refering", "answer_template": "The \"left hand most sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 255, 259, 260, 261, 276, 278, 279, 281, 282, 284, 285, 301, 308], "bbox": [0.002484375, 0.33091334894613583, 0.49031249999999993, 0.884192037470726], "iscrowd": 0, "area": 37301.26799999999, "rle": {"size": [427, 640], "counts": "]T1Z1o;1N100O2O00001O0100O1O1N2N2O1N2N1O2O10000001N1O2M2O1N3N1N2O2M2O2N2O0O2N2N2N1O2N2N1O2N2N1O2N2N2N3M5L4K5K5K5L5K4K5L4L4L4L4K5L5K4L6J5J6K5K6J2M2O1O1O1O1N2L400000000OYNoHVMQ7h2[InLe6o2gIhLY6W3QJ`Lo5`3[JVLe5i3o10000O100O10O10O100O10000O100O10000O10O010000O10000O1000000O1000000001O00000000001O0O100000001O0000000000001O00000000001O01O\\HTLi5l3VJULj5k3TJXLk5h3TJYLl5g3SJ[Lm5e3QJ\\Lo5d3PJ^Lo5b3oI`LQ6`3nIbLQ6^3nIdLR6[3mIfLS6[3jIhLU6X3jIhLW6X3hIgLZ6Y3eIgL]6X3bIhL_6X3_IhLc6X3\\IhLe6X3ZIgLh6Y3WIgLk6X3SIiLn6W3QIhLQ7X3nHhLS7X3lHhLU7X3iHiLY7V3fHjL[7V3dHiL^7W3aHiL`7W3_HiLb7W3\\HjLe7V3ZHjLg7V3XHjLi7V3VHjLk7V3SHkLn7U3oGlLS8T3iGoLX8Q3dGRM]8n2`GTMa8j31O1O1O1N2O1O1O1O1O1O1N2O1N2O1N2O1O1N2O1001O1O1O1O010O1N2L4L4N2N2N2N1O3M2N2N3M2N2M4L3L4M4L3M3M4M2O1N3M2O1N3N1NiNUIQLj6n3\\IoKa6Q4fIkKW6U4nIiKP6V4SJjKj5V4ZJiKd5V4_JjK^5V4eJjKX5V4kJkKR5T4QKlKm4S4VKmKg4S4]KlKa4S4bKmK\\4R4gKkKZ4T4iKjKW4V4kKgKU4Y4nKeKR4Z4RLbKo3]4TLaKl3^4WL`Ki3_4ZL^Kf3b4]L\\Kc3c4c2O1O1O1O1O1O1O1O1M3M3M3L5L5K5K5J5L5K4L4L4L3N3hN[EEi:9YEDj::WECm:<R1N2N2N200O1O1O001O1O1O010O1O1O0O2N2O2M3M3K5GdhW4"}, "label": "left hand most sheep"}]}
{"id": 280241, "image": "COCO_train2014_000000280241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark sheep is standing to the right of the other sheep and behind the dog\"."}], "task": "refering", "answer_template": "The \"a dark sheep is standing to the right of the other sheep and behind the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 255, 259, 260, 261, 276, 278, 279, 281, 282, 284, 285, 301, 308], "bbox": [0.002484375, 0.33091334894613583, 0.49031249999999993, 0.884192037470726], "iscrowd": 0, "area": 37301.26799999999, "rle": {"size": [427, 640], "counts": "]T1Z1o;1N100O2O00001O0100O1O1N2N2O1N2N1O2O10000001N1O2M2O1N3N1N2O2M2O2N2O0O2N2N2N1O2N2N1O2N2N1O2N2N2N3M5L4K5K5K5L5K4K5L4L4L4L4K5L5K4L6J5J6K5K6J2M2O1O1O1O1N2L400000000OYNoHVMQ7h2[InLe6o2gIhLY6W3QJ`Lo5`3[JVLe5i3o10000O100O10O10O100O10000O100O10000O10O010000O10000O1000000O1000000001O00000000001O0O100000001O0000000000001O00000000001O01O\\HTLi5l3VJULj5k3TJXLk5h3TJYLl5g3SJ[Lm5e3QJ\\Lo5d3PJ^Lo5b3oI`LQ6`3nIbLQ6^3nIdLR6[3mIfLS6[3jIhLU6X3jIhLW6X3hIgLZ6Y3eIgL]6X3bIhL_6X3_IhLc6X3\\IhLe6X3ZIgLh6Y3WIgLk6X3SIiLn6W3QIhLQ7X3nHhLS7X3lHhLU7X3iHiLY7V3fHjL[7V3dHiL^7W3aHiL`7W3_HiLb7W3\\HjLe7V3ZHjLg7V3XHjLi7V3VHjLk7V3SHkLn7U3oGlLS8T3iGoLX8Q3dGRM]8n2`GTMa8j31O1O1O1N2O1O1O1O1O1O1N2O1N2O1N2O1O1N2O1001O1O1O1O010O1N2L4L4N2N2N2N1O3M2N2N3M2N2M4L3L4M4L3M3M4M2O1N3M2O1N3N1NiNUIQLj6n3\\IoKa6Q4fIkKW6U4nIiKP6V4SJjKj5V4ZJiKd5V4_JjK^5V4eJjKX5V4kJkKR5T4QKlKm4S4VKmKg4S4]KlKa4S4bKmK\\4R4gKkKZ4T4iKjKW4V4kKgKU4Y4nKeKR4Z4RLbKo3]4TLaKl3^4WL`Ki3_4ZL^Kf3b4]L\\Kc3c4c2O1O1O1O1O1O1O1O1M3M3M3L5L5K5K5J5L5K4L4L4L3N3hN[EEi:9YEDj::WECm:<R1N2N2N200O1O1O001O1O1O010O1O1O0O2N2O2M3M3K5GdhW4"}, "label": "a dark sheep is standing to the right of the other sheep and behind the dog"}]}
{"id": 280241, "image": "COCO_train2014_000000280241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two sheep running the ground\"."}], "task": "refering", "answer_template": "The \"two sheep running the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 149, 150, 171, 172, 173, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 261, 262, 263, 264, 285, 286, 287, 288, 309], "bbox": [0.376046875, 0.39365339578454334, 0.552796875, 0.8803981264637003], "iscrowd": 0, "area": 9630.0258, "rle": {"size": [427, 640], "counts": "fiT32X=3N2M3M2N3M1O1N3N1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O1O2M2O1O1O6J6J3M4J6J6J6Ie0\\O4L3M3MfN_FPO_9Q1gFkNU9V1QGfNl8\\1YG`Ne8b1_GZN^8i1fGSNR7]OTIc2NmMl6BRId25gMh6HoHd2<`Md6OkHd2d0ZMa6f3cIWL[6j3hISLX6m3c100001O0000001O2N2N2N2N1O2SOeKeH^4Y7cKeH`4Y7bKdHa4Y7aKeHb4Y7`KdHc4Z7fK\\H\\4c7h0NoNaHXL\\7\\3SIdLk6Y3[IfLc6Y3`IgL^6W3gIlLS6R3QJPMk5n2YJPMf5Q3]JmLb5R3aJmL^5R3fJkLZ5T3iJjLX5U3jJiLV5V3nJgLR5X3QKfLo4Y3TKeLm4S3[KlLe4m2bKSM^4k2eKTM\\4i2gKVMY4h2iKWMZ4[1kHEn2oNZ4Y1jHlN26l2E[4U1jHlN72g2K]4T1eHkNj30b3T1fHhNj34b3R1hLmN[3P1eLoN_3n0bLQOa3m0_LROc3l0_4O001O1O1M3N2N2N2NnQg3"}, "label": "two sheep running the ground"}]}
{"id": 280241, "image": "COCO_train2014_000000280241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep closest to the dog\"."}], "task": "refering", "answer_template": "The \"the sheep closest to the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 149, 150, 171, 172, 173, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 261, 262, 263, 264, 285, 286, 287, 288, 309], "bbox": [0.376046875, 0.39365339578454334, 0.552796875, 0.8803981264637003], "iscrowd": 0, "area": 9630.0258, "rle": {"size": [427, 640], "counts": "fiT32X=3N2M3M2N3M1O1N3N1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O1O2M2O1O1O6J6J3M4J6J6J6Ie0\\O4L3M3MfN_FPO_9Q1gFkNU9V1QGfNl8\\1YG`Ne8b1_GZN^8i1fGSNR7]OTIc2NmMl6BRId25gMh6HoHd2<`Md6OkHd2d0ZMa6f3cIWL[6j3hISLX6m3c100001O0000001O2N2N2N2N1O2SOeKeH^4Y7cKeH`4Y7bKdHa4Y7aKeHb4Y7`KdHc4Z7fK\\H\\4c7h0NoNaHXL\\7\\3SIdLk6Y3[IfLc6Y3`IgL^6W3gIlLS6R3QJPMk5n2YJPMf5Q3]JmLb5R3aJmL^5R3fJkLZ5T3iJjLX5U3jJiLV5V3nJgLR5X3QKfLo4Y3TKeLm4S3[KlLe4m2bKSM^4k2eKTM\\4i2gKVMY4h2iKWMZ4[1kHEn2oNZ4Y1jHlN26l2E[4U1jHlN72g2K]4T1eHkNj30b3T1fHhNj34b3R1hLmN[3P1eLoN_3n0bLQOa3m0_LROc3l0_4O001O1O1M3N2N2N2NnQg3"}, "label": "the sheep closest to the dog"}]}
{"id": 108210, "image": "COCO_train2014_000000108210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lufthansa airplane\"."}], "task": "refering", "answer_template": "The \"lufthansa airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 121, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 224], "bbox": [0.12684375, 0.2715294117647059, 0.831203125, 0.6355764705882353], "iscrowd": 0, "area": 24398.652899999997, "rle": {"size": [425, 640], "counts": "TjQ11X=2N2N2N2N1O000O100000000000000000000O100000000000000000000O10000000000^MEVH;j7GTH9l7IRH7n7KPH5P8MnG3R8OkG2T81jGOV83hGMX85fGKZ87dGI\\89bGG^8<_GDa8>]GBc8`0[G@e8b0YG^Og8d0WG\\Oh8g0VGYOj8i0SGXOm8j0QGVOo8l0oFUOP9m0nFSOR9o0lFQOT9Q1jFoNV9S1hFnNW9T1gFlNY9V1eFjNZ9Y1dFgN]9Z1aFgN_9Z1^FgNc9Z1[FfNf9\\1WFdNj9]1TFdNl9]1RFcNo9^1oEbNR:_1lEaNV:_1hEaNc0Ai8o1bFaN=GR9i1_F`N8NZ9W2dFkM]9T2cFlM^9T2bFkM_9T2aFlM`9S2`FlMc9R2]FnMd9R2[FnMf9Q2ZFoMg9P2ZFoMg9Q2XFoMi9P2WFPNj9o1VFQNk9n1UFRNm9m1RFSNo9l1RFRNP:m1PFSNQ:m1nESNS:l1mETNT:k1lEUNU:k1jEUNW:j1jEUNX:i1hEWNY:h1gEXNZ:h1eEXN\\:g1dEYN]:f1cEYN_:g1`EYNa:f1`EYNa:f1_EZNc:e1\\E[Ne:d1[E\\Nf:c1ZE]Ng:b1YE^Nh:b1XE]Ni:b1WE^Nj:a1VE^Nl:b1SE^No:`1QE`NP;_1PEaNQ;_1oD`NR;_1nDaNS;^1mDbNT;^1kDbNV;]1jDcNW;\\1iDdNY;Z1hDeNY;[1fDdN\\;[1dDeN];Z1cDfN^;b11O1G`DgNa;`12N1O1O1N2O1O1O0000O1000000000000000001O000O10001O00000000001O000000001O000000001O000000000010O0000000001O000000001O000000001O1O1O1O1O1O1O100O1O1O1O001O1O001O001O1O000000000000000000000001O000000000000000O1000000000000000000000000000001O000000000O1000000000000000000000000000000000001O00000O1000000000000000000000000000000000000000000000000000000000000000000000000000000O10000O10000O1CfDkNZ;U1gDjNY;V1hDhNY;X1iDfNW;Z1jDeNV;[1kDcNV;]1lDaNT;_1mD`NS;`1;00000000001O0000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000001O00000000000001O000000001O000000001O000000001O000000001O00000002O3M3M3L4M3MM2N3L4M2N3L4M3M2N3N2O0O2O1N101O1N2O0O2O1N101N2O0O2O1N1O2N2N2N1O2N2N1OUc\\1"}, "label": "lufthansa airplane"}]}
{"id": 108210, "image": "COCO_train2014_000000108210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane sitting on the tarmac\"."}], "task": "refering", "answer_template": "The \"an airplane sitting on the tarmac\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 121, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 224], "bbox": [0.12684375, 0.2715294117647059, 0.831203125, 0.6355764705882353], "iscrowd": 0, "area": 24398.652899999997, "rle": {"size": [425, 640], "counts": "TjQ11X=2N2N2N2N1O000O100000000000000000000O100000000000000000000O10000000000^MEVH;j7GTH9l7IRH7n7KPH5P8MnG3R8OkG2T81jGOV83hGMX85fGKZ87dGI\\89bGG^8<_GDa8>]GBc8`0[G@e8b0YG^Og8d0WG\\Oh8g0VGYOj8i0SGXOm8j0QGVOo8l0oFUOP9m0nFSOR9o0lFQOT9Q1jFoNV9S1hFnNW9T1gFlNY9V1eFjNZ9Y1dFgN]9Z1aFgN_9Z1^FgNc9Z1[FfNf9\\1WFdNj9]1TFdNl9]1RFcNo9^1oEbNR:_1lEaNV:_1hEaNc0Ai8o1bFaN=GR9i1_F`N8NZ9W2dFkM]9T2cFlM^9T2bFkM_9T2aFlM`9S2`FlMc9R2]FnMd9R2[FnMf9Q2ZFoMg9P2ZFoMg9Q2XFoMi9P2WFPNj9o1VFQNk9n1UFRNm9m1RFSNo9l1RFRNP:m1PFSNQ:m1nESNS:l1mETNT:k1lEUNU:k1jEUNW:j1jEUNX:i1hEWNY:h1gEXNZ:h1eEXN\\:g1dEYN]:f1cEYN_:g1`EYNa:f1`EYNa:f1_EZNc:e1\\E[Ne:d1[E\\Nf:c1ZE]Ng:b1YE^Nh:b1XE]Ni:b1WE^Nj:a1VE^Nl:b1SE^No:`1QE`NP;_1PEaNQ;_1oD`NR;_1nDaNS;^1mDbNT;^1kDbNV;]1jDcNW;\\1iDdNY;Z1hDeNY;[1fDdN\\;[1dDeN];Z1cDfN^;b11O1G`DgNa;`12N1O1O1N2O1O1O0000O1000000000000000001O000O10001O00000000001O000000001O000000001O000000000010O0000000001O000000001O000000001O1O1O1O1O1O1O100O1O1O1O001O1O001O001O1O000000000000000000000001O000000000000000O1000000000000000000000000000001O000000000O1000000000000000000000000000000000001O00000O1000000000000000000000000000000000000000000000000000000000000000000000000000000O10000O10000O1CfDkNZ;U1gDjNY;V1hDhNY;X1iDfNW;Z1jDeNV;[1kDcNV;]1lDaNT;_1mD`NS;`1;00000000001O0000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000001O00000000000001O000000001O000000001O000000001O000000001O00000002O3M3M3L4M3MM2N3L4M2N3L4M3M2N3N2O0O2O1N101O1N2O0O2O1N101N2O0O2O1N1O2N2N2N1O2N2N1OUc\\1"}, "label": "an airplane sitting on the tarmac"}]}
{"id": 165555, "image": "COCO_train2014_000000165555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rose loveseat facing a window\"."}], "task": "refering", "answer_template": "The \"a rose loveseat facing a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 207, 208, 209, 210, 211, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350], "bbox": [0.0, 0.5076875, 0.37496874999999996, 0.943], "iscrowd": 0, "area": 37020.938949999996, "rle": {"size": [480, 640], "counts": "T8Y4g:1O1O001N2O1O001O1O1N101O1O001N2O1O001O1O0O2O1O1O001N2O1O001O1O0O2O1O1O001N2O001O1O1N101O1O1O0O2O1O001O1N2O001O1O0O2O1O1O001O1O1O001O1O001N2O1O001O1O000O101O0000000O10000000001OO10O10O100O10000O10O10O100O10000O101O1N101N2O1O1N2O1O1N2O3L4M3M3L4M3M3L4M1N101O0O101O0O101O0O2O0O101O0O101O0O2O0O101O0O101O0O2O0O101O0O10000O100O01000O1000O0100O1000O010000O100O01000O1000O0100O1000O010000O10O010000O10O10O100O100O001O100O1O001O1O1O1O010O1O1O1O1O1O1O100O1O102N2YNkD_OW;1ZELh:1[EMg:0]EMe:0_EMb:1bEL`:1dEL^:1eEM]:1fEL\\:1hEK[:2iEJZ:3jEJh<MoSk5"}, "label": "a rose loveseat facing a window"}]}
{"id": 165555, "image": "COCO_train2014_000000165555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch farthest from the windows\"."}], "task": "refering", "answer_template": "The \"the couch farthest from the windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 207, 208, 209, 210, 211, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350], "bbox": [0.0, 0.5076875, 0.37496874999999996, 0.943], "iscrowd": 0, "area": 37020.938949999996, "rle": {"size": [480, 640], "counts": "T8Y4g:1O1O001N2O1O001O1O1N101O1O001N2O1O001O1O0O2O1O1O001N2O1O001O1O0O2O1O1O001N2O001O1O1N101O1O1O0O2O1O001O1N2O001O1O0O2O1O1O001O1O1O001O1O001N2O1O001O1O000O101O0000000O10000000001OO10O10O100O10000O10O10O100O10000O101O1N101N2O1O1N2O1O1N2O3L4M3M3L4M3M3L4M1N101O0O101O0O101O0O2O0O101O0O101O0O2O0O101O0O101O0O2O0O101O0O10000O100O01000O1000O0100O1000O010000O100O01000O1000O0100O1000O010000O10O010000O10O10O100O100O001O100O1O001O1O1O1O010O1O1O1O1O1O1O100O1O102N2YNkD_OW;1ZELh:1[EMg:0]EMe:0_EMb:1bEL`:1dEL^:1eEM]:1fEL\\:1hEK[:2iEJZ:3jEJh<MoSk5"}, "label": "the couch farthest from the windows"}]}
{"id": 165555, "image": "COCO_train2014_000000165555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rose colored couch\"."}], "task": "refering", "answer_template": "The \"a rose colored couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 245, 246, 247], "bbox": [0.55684375, 0.4200625, 0.83903125, 0.6101041666666667], "iscrowd": 0, "area": 11095.613, "rle": {"size": [480, 640], "counts": "laW52m>2M2O3M2N2N3N1N4L9H?A1O00000O10001O0000000O101O000O10000O2O00000O10001O0O10000O101XCWNW<i1bC^N^<b1[CfNc<l100000001O00000O101O000000001O000000010O000001O01O0001O01O000000001O000O100010O1O1OO1O2O000000001O0000000010O00000001O000000001O000000001O01N100O10000O100O10000O105J01N2O0010O100O101O0O100O101O0O100O100O1O2N1O1M3G9G9H8M3O2O0O1O100O1O100O2N100O1O100O2N101N1O2O0O2N101N1N3J[R`1"}, "label": "a rose colored couch"}]}
{"id": 165555, "image": "COCO_train2014_000000165555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red loveseat with 3 large pillows\"."}], "task": "refering", "answer_template": "The \"red loveseat with 3 large pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 245, 246, 247], "bbox": [0.55684375, 0.4200625, 0.83903125, 0.6101041666666667], "iscrowd": 0, "area": 11095.613, "rle": {"size": [480, 640], "counts": "laW52m>2M2O3M2N2N3N1N4L9H?A1O00000O10001O0000000O101O000O10000O2O00000O10001O0O10000O101XCWNW<i1bC^N^<b1[CfNc<l100000001O00000O101O000000001O000000010O000001O01O0001O01O000000001O000O100010O1O1OO1O2O000000001O0000000010O00000001O000000001O000000001O01N100O10000O100O10000O105J01N2O0010O100O101O0O100O101O0O100O100O1O2N1O1M3G9G9H8M3O2O0O1O100O1O100O2N100O1O100O2N101N1O2O0O2N101N1N3J[R`1"}, "label": "red loveseat with 3 large pillows"}]}
{"id": 165555, "image": "COCO_train2014_000000165555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch located directly across from a fire place\"."}], "task": "refering", "answer_template": "The \"a white couch located directly across from a fire place\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 227, 228, 229, 242, 243, 244, 245, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387], "bbox": [0.52034375, 0.5055625, 1.0, 0.9852083333333332], "iscrowd": 0, "area": 40347.219249999995, "rle": {"size": [480, 640], "counts": "i\\l43k>7J6J7H7J5K2M3N2N2M4M2O1O1O2N1O1O1O1O2N8H9G:F9G9G9G:F3M1N101O001O001O1O001O010O001O1O010O001O010O1O001O010O001O1O010O001O1O010O001O0010O01O000010O0001O00010O00001O01O01O000010O0001O00010O00001O01O01O000010O0001O00010O00001O01O01O000010O0001O01O01O000010O000000001O000001O01O000000001O0000000010O00000001O000000001O000001O000O1000000000000000001O00000000000O10000000001O000000000000O1O10O0100O1O100O1O100O1O1O10O0O10000O2O00000O2O00000O101O000O01000O1000O10O100000O01000000O10O10O100000O011O001O0O2N2N101N1O2N1O2O0O2N1O2O0O2O0O2N101N101N1O2O1N101O0O2O001N101O0O2O001N101O0O2O0O2M2O1O2M2O2A>B>K6K5J7J6J6M3N2M2N3M3I\\H"}, "label": "a white couch located directly across from a fire place"}]}
{"id": 165555, "image": "COCO_train2014_000000165555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white colored sofa in a room near two red colored sofas\"."}], "task": "refering", "answer_template": "The \"white colored sofa in a room near two red colored sofas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 227, 228, 229, 242, 243, 244, 245, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387], "bbox": [0.52034375, 0.5055625, 1.0, 0.9852083333333332], "iscrowd": 0, "area": 40347.219249999995, "rle": {"size": [480, 640], "counts": "i\\l43k>7J6J7H7J5K2M3N2N2M4M2O1O1O2N1O1O1O1O2N8H9G:F9G9G9G:F3M1N101O001O001O1O001O010O001O1O010O001O010O1O001O010O001O1O010O001O1O010O001O0010O01O000010O0001O00010O00001O01O01O000010O0001O00010O00001O01O01O000010O0001O00010O00001O01O01O000010O0001O01O01O000010O000000001O000001O01O000000001O0000000010O00000001O000000001O000001O000O1000000000000000001O00000000000O10000000001O000000000000O1O10O0100O1O100O1O100O1O1O10O0O10000O2O00000O2O00000O101O000O01000O1000O10O100000O01000000O10O10O100000O011O001O0O2N2N101N1O2N1O2O0O2N1O2O0O2O0O2N101N101N1O2O1N101O0O2O001N101O0O2O001N101O0O2O0O2M2O1O2M2O2A>B>K6K5J7J6J6M3N2M2N3M3I\\H"}, "label": "white colored sofa in a room near two red colored sofas"}]}
{"id": 181941, "image": "COCO_train2014_000000181941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white ski that ' s pointing upwards in the top left\"."}], "task": "refering", "answer_template": "The \"the white ski that ' s pointing upwards in the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 52, 53, 74, 75], "bbox": [0.15740625, 0.07318658280922431, 0.33240625, 0.4176939203354298], "iscrowd": 0, "area": 911.5092999999991, "rle": {"size": [477, 640], "counts": "dW_11j>3N1N3O13Jc]d0N_b[O100O1O100O1O100O002N2O1N2N3N1N1O01O01O010O00010O0010O00010O00010O01O010O0010O010O00010O010O100O100O100O101N1O2O1N2O1N4M2MkkV6"}, "label": "the white ski that ' s pointing upwards in the top left"}]}
{"id": 181941, "image": "COCO_train2014_000000181941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white ski that ' s pointing upwards in the top left\"."}], "task": "refering", "answer_template": "The \"the white ski that ' s pointing upwards in the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 52, 53, 74, 75], "bbox": [0.15740625, 0.07318658280922431, 0.33240625, 0.4176939203354298], "iscrowd": 0, "area": 911.5092999999991, "rle": {"size": [477, 640], "counts": "dW_11j>3N1N3O13Jc]d0N_b[O100O1O100O1O100O002N2O1N2N3N1N1O01O01O010O00010O0010O00010O00010O01O010O0010O010O00010O010O100O100O100O101N1O2O1N2O1N4M2MkkV6"}, "label": "the white ski that ' s pointing upwards in the top left"}]}
{"id": 321206, "image": "COCO_train2014_000000321206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe whose head is in the barn\"."}], "task": "refering", "answer_template": "The \"giraffe whose head is in the barn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 103, 104, 105, 127, 128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 198, 200, 221, 223], "bbox": [0.49190625, 0.23362997658079626, 0.735890625, 0.6687822014051522], "iscrowd": 0, "area": 8682.994750000002, "rle": {"size": [427, 640], "counts": "g^S43W=5K6J5K5K5K3N0O1O1O001001O1O1N2O1O1O1N2O1O100O100O10O10O10000O100O0100O010O1O010O010O1O010O10O0100O0010O0100O00100O010O00100O011N7J6I7I6K5J7J5J6K5J6XF_M[8f2^GhMV8^2bGgMZ8_2`GaMa8d2XG]Mh8c31O1O001O001N2B=D=B=f0[OWNbFHZ97nFAR9?WGYOi8f0_GSO`8m0hGlNX8S1QHdNP8[1XH^Ng7b1aHWN_7h1R210O1O001O001O1O001O001O001O010O001O1O001O001O001O001O02O1O1N2O0O2aEjNf8W1kFWOT9k0\\FEc9Q20L5K4L5K5K4L5KRFgMR9U2bFZN`9a1TFoNl9W23M2N3M3M2N3N1N3K5WOh0YOh0XOfYV2"}, "label": "giraffe whose head is in the barn"}]}
{"id": 321206, "image": "COCO_train2014_000000321206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe whose head is not visible\"."}], "task": "refering", "answer_template": "The \"the giraffe whose head is not visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 103, 104, 105, 127, 128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 198, 200, 221, 223], "bbox": [0.49190625, 0.23362997658079626, 0.735890625, 0.6687822014051522], "iscrowd": 0, "area": 8682.994750000002, "rle": {"size": [427, 640], "counts": "g^S43W=5K6J5K5K5K3N0O1O1O001001O1O1N2O1O1O1N2O1O100O100O10O10O10000O100O0100O010O1O010O010O1O010O10O0100O0010O0100O00100O010O00100O011N7J6I7I6K5J7J5J6K5J6XF_M[8f2^GhMV8^2bGgMZ8_2`GaMa8d2XG]Mh8c31O1O001O001N2B=D=B=f0[OWNbFHZ97nFAR9?WGYOi8f0_GSO`8m0hGlNX8S1QHdNP8[1XH^Ng7b1aHWN_7h1R210O1O001O001O1O001O001O001O010O001O1O001O001O001O001O02O1O1N2O0O2aEjNf8W1kFWOT9k0\\FEc9Q20L5K4L5K5K4L5KRFgMR9U2bFZN`9a1TFoNl9W23M2N3M3M2N3N1N3K5WOh0YOh0XOfYV2"}, "label": "the giraffe whose head is not visible"}]}
{"id": 321206, "image": "COCO_train2014_000000321206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue shirt feeding a giraffe in a zoo\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue shirt feeding a giraffe in a zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 163, 164, 165, 166, 186, 187, 188, 189, 190, 209, 210, 211, 212, 219, 220, 232, 233, 234, 235, 236, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329], "bbox": [0.09517187499999999, 0.4343325526932085, 0.59090625, 0.9999297423887589], "iscrowd": 0, "area": 27737.3642, "rle": {"size": [427, 640], "counts": "^ei0<_<a0M3N2M3N2M3N2M4M2M3N2N2N2N1O2N2lEUNe8m1XG[Na8g1\\GbN]8^1aGkNW8W1gGQOR8P1kGSOS8o0jGTOU8m0hGWOU8j0iGYOV8h0gG\\OW8e0gG]OW8e0fG_OX8b0eGAY8`0eGDY8=cGJX8j2K4K5L4K6K2N2M3N3L3N2M4M2M3N3L3N2N3L3N2M4M2M3O2M2O1N0001O001O00001O001O00001O001O001O1O1O001O1O1O1O1O1O001O1O1O1O001O001O1_KbIS2^6eMkIZ2V6^MRJa2n5^MTJa2m5\\MVJc2j5\\MXJc2i5[MZJc2g5ZM\\Je2e5WL[Ie0R1S3d5RL`Ii0n0T3P6kLQJT3Q6iLQJV3P6hLRJW3o5gLSJX3n5gLSJX3o5eLSJZ3o5cLSJ\\3P6`LRJ_3P6_LQJ`3Q6]LQJb3Q6[LQJe3k7O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O00001O001O001\\NPFKP:5ZFAf9?dFWO]9i0lFmNT9S1WGbNj8^1a100010O01O000O2O001O001O00001O001O00O1O2N1O1N2O1O1O2N1O1O1O1O101O01O000O10000O101N1O2O1N2N101N2N4M3L3M2N10O00010O0001O01O01O01O01O00001O001O00001O00001O0000001O00001O0000001O00001O00001O00000010O0001O00001O00000010O00010O0001O01O00010O0001O010O1O01000000O10O10O1000000O10000O1000000O1000O010000O10000O10O1O001O1O001O1O001O1O001O1O001N2O1N2N3M2O1N^o\\3"}, "label": "a woman wearing a blue shirt feeding a giraffe in a zoo"}]}
{"id": 321206, "image": "COCO_train2014_000000321206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing blue t - shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing blue t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 163, 164, 165, 166, 186, 187, 188, 189, 190, 209, 210, 211, 212, 219, 220, 232, 233, 234, 235, 236, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329], "bbox": [0.09517187499999999, 0.4343325526932085, 0.59090625, 0.9999297423887589], "iscrowd": 0, "area": 27737.3642, "rle": {"size": [427, 640], "counts": "^ei0<_<a0M3N2M3N2M3N2M4M2M3N2N2N2N1O2N2lEUNe8m1XG[Na8g1\\GbN]8^1aGkNW8W1gGQOR8P1kGSOS8o0jGTOU8m0hGWOU8j0iGYOV8h0gG\\OW8e0gG]OW8e0fG_OX8b0eGAY8`0eGDY8=cGJX8j2K4K5L4K6K2N2M3N3L3N2M4M2M3N3L3N2N3L3N2M4M2M3O2M2O1N0001O001O00001O001O00001O001O001O1O1O001O1O1O1O1O1O001O1O1O1O001O001O1_KbIS2^6eMkIZ2V6^MRJa2n5^MTJa2m5\\MVJc2j5\\MXJc2i5[MZJc2g5ZM\\Je2e5WL[Ie0R1S3d5RL`Ii0n0T3P6kLQJT3Q6iLQJV3P6hLRJW3o5gLSJX3n5gLSJX3o5eLSJZ3o5cLSJ\\3P6`LRJ_3P6_LQJ`3Q6]LQJb3Q6[LQJe3k7O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O00001O001O001\\NPFKP:5ZFAf9?dFWO]9i0lFmNT9S1WGbNj8^1a100010O01O000O2O001O001O00001O001O00O1O2N1O1N2O1O1O2N1O1O1O1O101O01O000O10000O101N1O2O1N2N101N2N4M3L3M2N10O00010O0001O01O01O01O01O00001O001O00001O00001O0000001O00001O0000001O00001O00001O00000010O0001O00001O00000010O00010O0001O01O00010O0001O010O1O01000000O10O10O1000000O10000O1000000O1000O010000O10000O10O1O001O1O001O1O001O1O001O1O001N2O1N2N3M2O1N^o\\3"}, "label": "a woman wearing blue t - shirt"}]}
{"id": 321206, "image": "COCO_train2014_000000321206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe is being fed\"."}], "task": "refering", "answer_template": "The \"a giraffe is being fed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 33, 34, 35, 36, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 62, 63, 64, 65, 66, 67, 68, 79, 80, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 297, 298], "bbox": [0.449796875, 0.0, 1.0, 0.8853864168618267], "iscrowd": 0, "area": 42380.45849999999, "rle": {"size": [427, 640], "counts": "`Uh35P=:G8G9H8H8G9H8G6K5H8H7I8I70O1000000O10000O1000000O1000000O2O000O1000000O10000O10000M3J6J6J6K5M3O1O1O1O1O100O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O11O1O1O001O1O001O1O1O001O1N2O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O000010O01O001O001O00001O001O001O001O000010O01O001O001O00001O001O001O00001O0010O01O001O00001O001O001O001O01O01O001O001O001O00001O0010O01O000eFjMa7W2\\H[NT7e1kHcNn6^1oHkNj6U1TISOf6n0WI[Ob6e0\\ID]6=`IKZ65dI4U6LiI=Q6ClIe0n5[OPJn0i5SOTJV1e5jNZJ]1`5dN]Jk4V2UKhM[5j1fJSNf5c1ZJ[Nf5g1ZJWNf5k1k22N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2N2N2N2N2N3J5D<C=SOm0bN_1VOi0E;2N2N2O1N2N2N2N2N2O1N3M2N2N2O1O1O1O1O3M2N2M3N3M2N2N2N3M2N5J7J6J6J6J6J6J6J6J5J7J6JR1nNN2M3N2N2M2N3VNnLVIV3e6RMUIQ3e6XMTIl2g6]MQIg2j6aMPIb2j6gMoH]2l6lMlHX2P7oMjHU2P7W2J5L5J6XN`I^Ld6]3jIUL\\6f3PJnKU6l3i1K4K6K5K4@a0hNX1gNjN"}, "label": "a giraffe is being fed"}]}
{"id": 321206, "image": "COCO_train2014_000000321206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe being fed by a young woman\"."}], "task": "refering", "answer_template": "The \"a giraffe being fed by a young woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 33, 34, 35, 36, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 62, 63, 64, 65, 66, 67, 68, 79, 80, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 297, 298], "bbox": [0.449796875, 0.0, 1.0, 0.8853864168618267], "iscrowd": 0, "area": 42380.45849999999, "rle": {"size": [427, 640], "counts": "`Uh35P=:G8G9H8H8G9H8G6K5H8H7I8I70O1000000O10000O1000000O1000000O2O000O1000000O10000O10000M3J6J6J6K5M3O1O1O1O1O100O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O11O1O1O001O1O001O1O1O001O1N2O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O000010O01O001O001O00001O001O001O001O000010O01O001O001O00001O001O001O00001O0010O01O001O00001O001O001O001O01O01O001O001O001O00001O0010O01O000eFjMa7W2\\H[NT7e1kHcNn6^1oHkNj6U1TISOf6n0WI[Ob6e0\\ID]6=`IKZ65dI4U6LiI=Q6ClIe0n5[OPJn0i5SOTJV1e5jNZJ]1`5dN]Jk4V2UKhM[5j1fJSNf5c1ZJ[Nf5g1ZJWNf5k1k22N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2N2N2N2N2N3J5D<C=SOm0bN_1VOi0E;2N2N2O1N2N2N2N2N2O1N3M2N2N2O1O1O1O1O3M2N2M3N3M2N2N2N3M2N5J7J6J6J6J6J6J6J6J5J7J6JR1nNN2M3N2N2M2N3VNnLVIV3e6RMUIQ3e6XMTIl2g6]MQIg2j6aMPIb2j6gMoH]2l6lMlHX2P7oMjHU2P7W2J5L5J6XN`I^Ld6]3jIUL\\6f3PJnKU6l3i1K4K6K5K4@a0hNX1gNjN"}, "label": "a giraffe being fed by a young woman"}]}
{"id": 542391, "image": "COCO_train2014_000000542391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch against wall with windows behind it\"."}], "task": "refering", "answer_template": "The \"couch against wall with windows behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 101, 102, 115, 116, 117, 118, 119, 120, 121, 133, 134, 135, 136, 137, 138, 139, 154, 155, 156], "bbox": [0.40772, 0.37976689976689976, 0.7388800000000001, 0.568018648018648], "iscrowd": 0, "area": 9908.132800000003, "rle": {"size": [429, 500], "counts": "fde28T=2O1hCHX;:_DNa;2WD6h;LoC<Q<a0N100O100O101N1O10O0NEgNhDY1X;jNdDW1\\;lNaDT1_;=000000010O0000000000000001O00000000000001O010O1O1O1O00000001M2O10000001O000001O0000000001O000001O0000000O10000010O4L00000000000001O0O100000000000001O00000000000000001N100O1N200O10001N10000000000000001O00000000000000001N100000000000001O000001O003M22K2O0O10O01O03N2N1N3N1O0_Oc0D>K100O2FcCA_<=:N2Maff1"}, "label": "couch against wall with windows behind it"}]}
{"id": 542391, "image": "COCO_train2014_000000542391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green couch against a wall with two windows\"."}], "task": "refering", "answer_template": "The \"a green couch against a wall with two windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 101, 102, 115, 116, 117, 118, 119, 120, 121, 133, 134, 135, 136, 137, 138, 139, 154, 155, 156], "bbox": [0.40772, 0.37976689976689976, 0.7388800000000001, 0.568018648018648], "iscrowd": 0, "area": 9908.132800000003, "rle": {"size": [429, 500], "counts": "fde28T=2O1hCHX;:_DNa;2WD6h;LoC<Q<a0N100O100O101N1O10O0NEgNhDY1X;jNdDW1\\;lNaDT1_;=000000010O0000000000000001O00000000000001O010O1O1O1O00000001M2O10000001O000001O0000000001O000001O0000000O10000010O4L00000000000001O0O100000000000001O00000000000000001N100O1N200O10001N10000000000000001O00000000000000001N100000000000001O000001O003M22K2O0O10O01O03N2N1N3N1O0_Oc0D>K100O2FcCA_<=:N2Maff1"}, "label": "a green couch against a wall with two windows"}]}
{"id": 566968, "image": "COCO_train2014_000000566968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sliver cup\"."}], "task": "refering", "answer_template": "The \"a sliver cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 236], "bbox": [0.18076562499999999, 0.30087499999999995, 0.382171875, 0.5947083333333333], "iscrowd": 0, "area": 10570.36345, "rle": {"size": [480, 640], "counts": "kbf1;\\>=J6J7H7J6J5J5L5K4K5L2O0O1O1O2N100O1QD_M_;c2`D^M_;b2`DaM^;_2bDbM];^2cDcM\\;]2cDfM[;Z2eDgMZ;Y2fDiMW;X2hDjMW;U2jDmMT;S2kDoMT;Q2lDPNS;P2mDRNR;m1nDTNQ;l1oDVNP;i1QEWNn:i1REYNm:f1TEZNk:f1UE[Nk:d1UE^Ni:a1YE_Ng:U1dEmNZ:o0kEQOU:l0mEVOQ:g0RFZOn9c0UF^Oi9?ZFBf9<\\FDc9;^FGa97aFI^97bFK]93dFN[91gFOY90hF0W9OjF2V9MkF3T9MlF4T9JnF6Q9JPG6P9IPG8P9FRG:m8FSG;m8DTG<k8DVG<j8BWG?h8AYG?g8@YGa0g8^OZGb0f8]O[Gc0e8\\O[Ge0e8ZO\\Gf0d8XO]Gi0c8VO]Gk0c8TO]Gm0c8RO]Go0c8PO^Go0c8QO\\GP1d8oN\\GR1d8mN\\GS1e8mNZGT1f8kNZGU1g8kNYGU1g8jNYGW1g8iNXGW1i8hNWGY1h8iNVGW1k8jNTGV1l8jNSGV1n8kNPGV1Q9jNmFW1T9iNjFW1X9hNgFY1Y9hNeFX1]9hN\\F^1e9bNSFd1n9\\NlEj1U:VNcEQ2^:oMZEW2g:iMSE]2n:i02N2N101N2N1O2N2N2N2N3M2O2L3M4M2M4L3M4M2M3M4L3K6J6K6I7I7I7H9D;Fmnh5"}, "label": "a sliver cup"}]}
{"id": 566968, "image": "COCO_train2014_000000566968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"steel bowl containing sauce\"."}], "task": "refering", "answer_template": "The \"steel bowl containing sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 236], "bbox": [0.18076562499999999, 0.30087499999999995, 0.382171875, 0.5947083333333333], "iscrowd": 0, "area": 10570.36345, "rle": {"size": [480, 640], "counts": "kbf1;\\>=J6J7H7J6J5J5L5K4K5L2O0O1O1O2N100O1QD_M_;c2`D^M_;b2`DaM^;_2bDbM];^2cDcM\\;]2cDfM[;Z2eDgMZ;Y2fDiMW;X2hDjMW;U2jDmMT;S2kDoMT;Q2lDPNS;P2mDRNR;m1nDTNQ;l1oDVNP;i1QEWNn:i1REYNm:f1TEZNk:f1UE[Nk:d1UE^Ni:a1YE_Ng:U1dEmNZ:o0kEQOU:l0mEVOQ:g0RFZOn9c0UF^Oi9?ZFBf9<\\FDc9;^FGa97aFI^97bFK]93dFN[91gFOY90hF0W9OjF2V9MkF3T9MlF4T9JnF6Q9JPG6P9IPG8P9FRG:m8FSG;m8DTG<k8DVG<j8BWG?h8AYG?g8@YGa0g8^OZGb0f8]O[Gc0e8\\O[Ge0e8ZO\\Gf0d8XO]Gi0c8VO]Gk0c8TO]Gm0c8RO]Go0c8PO^Go0c8QO\\GP1d8oN\\GR1d8mN\\GS1e8mNZGT1f8kNZGU1g8kNYGU1g8jNYGW1g8iNXGW1i8hNWGY1h8iNVGW1k8jNTGV1l8jNSGV1n8kNPGV1Q9jNmFW1T9iNjFW1X9hNgFY1Y9hNeFX1]9hN\\F^1e9bNSFd1n9\\NlEj1U:VNcEQ2^:oMZEW2g:iMSE]2n:i02N2N101N2N1O2N2N2N2N3M2O2L3M4M2M4L3M4M2M3M4L3K6J6K6I7I7I7H9D;Fmnh5"}, "label": "steel bowl containing sauce"}]}
{"id": 566968, "image": "COCO_train2014_000000566968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver tray with butter on it\"."}], "task": "refering", "answer_template": "The \"a silver tray with butter on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 100, 101, 102, 103, 125], "bbox": [0.264609375, 0.0, 0.6438124999999999, 0.303375], "iscrowd": 0, "area": 23360.004450000004, "rle": {"size": [480, 640], "counts": "fX_21g>9F;F9M3N2N3M2N2N2N3M2N2N2N3M2O1N2N3M2N2N2N3M2N2N2N2N1O2N1O2N1O2O00002N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O00001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001OO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1K5J6J6IVlZ3"}, "label": "a silver tray with butter on it"}]}
{"id": 566968, "image": "COCO_train2014_000000566968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a container of cheese sitting next to various other cheeses\"."}], "task": "refering", "answer_template": "The \"a container of cheese sitting next to various other cheeses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 100, 101, 102, 103, 125], "bbox": [0.264609375, 0.0, 0.6438124999999999, 0.303375], "iscrowd": 0, "area": 23360.004450000004, "rle": {"size": [480, 640], "counts": "fX_21g>9F;F9M3N2N3M2N2N2N3M2N2N2N3M2O1N2N3M2N2N2N3M2N2N2N2N1O2N1O2N1O2O00002N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O00001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001OO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1K5J6J6IVlZ3"}, "label": "a container of cheese sitting next to various other cheeses"}]}
{"id": 566968, "image": "COCO_train2014_000000566968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spoon in a bowl of jelly\"."}], "task": "refering", "answer_template": "The \"a spoon in a bowl of jelly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 144, 145, 166, 167, 168, 189, 190, 211, 212, 233, 234, 255, 256, 277, 278, 300], "bbox": [0.06121875, 0.3446666666666667, 0.33332812500000003, 0.7868541666666666], "iscrowd": 0, "area": 6192.341, "rle": {"size": [480, 640], "counts": "\\db05f>=D4O1O1O1O100O1O1O2N0001O00000000001O000000001O00000000001O000000O100O1000O010O010O10O010O01000O010O010O0100O010O010000O010O1000O0100O10O1000000O1000001N1000000O1000000O101O000O1000000O10001O0O1000000O1000000O2O00000O100000N1YOh0O2O1N2O1N2O1N101N2N2O1N2O1N2ON12O1N2N101N2O00100O010O1O10O01O10O01O100O00100O01O1N2N2N2N2O1N3M2N2N2N2O1M4L7I6I7J6J6J6JXoW6"}, "label": "a spoon in a bowl of jelly"}]}
{"id": 566968, "image": "COCO_train2014_000000566968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spoon lying in a silver bowl of brown liquid\"."}], "task": "refering", "answer_template": "The \"a spoon lying in a silver bowl of brown liquid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 144, 145, 166, 167, 168, 189, 190, 211, 212, 233, 234, 255, 256, 277, 278, 300], "bbox": [0.06121875, 0.3446666666666667, 0.33332812500000003, 0.7868541666666666], "iscrowd": 0, "area": 6192.341, "rle": {"size": [480, 640], "counts": "\\db05f>=D4O1O1O1O100O1O1O2N0001O00000000001O000000001O00000000001O000000O100O1000O010O010O10O010O01000O010O010O0100O010O010000O010O1000O0100O10O1000000O1000001N1000000O1000000O101O000O1000000O10001O0O1000000O1000000O2O00000O100000N1YOh0O2O1N2O1N2O1N101N2N2O1N2O1N2ON12O1N2N101N2O00100O010O1O10O01O10O01O100O00100O01O1N2N2N2N2O1N3M2N2N2N2O1M4L7I6I7J6J6J6JXoW6"}, "label": "a spoon lying in a silver bowl of brown liquid"}]}
{"id": 345781, "image": "COCO_train2014_000000345781.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt feeding leafs to a giraffe\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt feeding leafs to a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 220, 221, 228, 229, 230, 242, 243, 244, 250, 251, 252, 264, 265, 266, 267, 271, 272, 273, 286, 287, 288, 292, 293, 294, 308, 309, 310, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 378, 396, 397, 398, 399, 418, 419, 420, 421, 440, 441, 442, 443, 462, 463, 464, 465], "bbox": [0.002238562091503268, 0.4326307189542483, 0.49326797385620913, 0.9864379084967321], "iscrowd": 0, "area": 38537.40355000001, "rle": {"size": [612, 612], "counts": "aU1;[?^3bL^3cM]2001O0000000000001O000000000000010O0000000000001O0000000000001O0000000000010O0000000000001O0000002N1O2N1O1O2N1O2N1O101N1O2N1O1O2N1O2N1O2N1O1O2N1O2UJ^EP3c:gLdEZ3]:]LkEc3W:SLQFm3P:jKXFV4j9`K]Fa4d9VKdFk4\\9lJlFT5V9bJQG_5Z;000O10QMo2O1O001O10O01O1O00100O1O001O10O01O1O001O100O001O1O010O1O001O100O001O1O010O1O001O1O10O01O1O00100O1O001O10O01O1O00100O1O001O1O010O1O000001O000O10001O000000001O00000O10001O000000001O00000O101O000000001O0000000O101O000000001O0000000O2O00000000001O0000000O2O0000000000O100O1O10O01O100O1O100O1O100O1O010O100O1O100O1O100O00100O1O100O100O1O10O0100001O1O001O001O0O2O1O0010O010O10O010O10O10O010O01000O010O010O0100O001M3N2M3M3M3M3M3Mk0VOioh5"}, "label": "a woman in a white shirt feeding leafs to a giraffe"}]}
{"id": 345781, "image": "COCO_train2014_000000345781.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl feeding a giraffe\"."}], "task": "refering", "answer_template": "The \"a girl feeding a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 220, 221, 228, 229, 230, 242, 243, 244, 250, 251, 252, 264, 265, 266, 267, 271, 272, 273, 286, 287, 288, 292, 293, 294, 308, 309, 310, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 378, 396, 397, 398, 399, 418, 419, 420, 421, 440, 441, 442, 443, 462, 463, 464, 465], "bbox": [0.002238562091503268, 0.4326307189542483, 0.49326797385620913, 0.9864379084967321], "iscrowd": 0, "area": 38537.40355000001, "rle": {"size": [612, 612], "counts": "aU1;[?^3bL^3cM]2001O0000000000001O000000000000010O0000000000001O0000000000001O0000000000010O0000000000001O0000002N1O2N1O1O2N1O2N1O101N1O2N1O1O2N1O2N1O2N1O1O2N1O2UJ^EP3c:gLdEZ3]:]LkEc3W:SLQFm3P:jKXFV4j9`K]Fa4d9VKdFk4\\9lJlFT5V9bJQG_5Z;000O10QMo2O1O001O10O01O1O00100O1O001O10O01O1O001O100O001O1O010O1O001O100O001O1O010O1O001O1O10O01O1O00100O1O001O10O01O1O00100O1O001O1O010O1O000001O000O10001O000000001O00000O10001O000000001O00000O101O000000001O0000000O101O000000001O0000000O2O00000000001O0000000O2O0000000000O100O1O10O01O100O1O100O1O100O1O010O100O1O100O1O100O00100O1O100O100O1O10O0100001O1O001O001O0O2O1O0010O010O10O010O10O10O010O01000O010O010O0100O001M3N2M3M3M3M3M3Mk0VOioh5"}, "label": "a girl feeding a giraffe"}]}
{"id": 345781, "image": "COCO_train2014_000000345781.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid watching a giraffe at the zoo\"."}], "task": "refering", "answer_template": "The \"a kid watching a giraffe at the zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 328, 329, 350, 351, 372, 373, 393, 394, 395, 414, 415, 416, 417, 436, 437, 438, 439, 458, 459, 460, 461, 480, 481, 482, 483], "bbox": [0.8266830065359477, 0.5880392156862745, 1.0, 1.0], "iscrowd": 0, "area": 16010.297049999992, "rle": {"size": [612, 612], "counts": "[P_91Qc02M3N2N2N2N2M3WNFg@;V?Oc@1[?8_@G_?<`@D^??a@A]?b0b@^O]?d0a@]O^?f0_@[O`?g0^@ZOa?h0]@YOa?j0\\@XOc?j0[@WOZ?ZO[@a19VO\\?[OX@a1:UO]?]OV@`1;TO^?W1`@jN_?X1_@iN`?Y1]@iNb?Y1\\@hNc?]2N2O1O1N2O1N2O1N2O1N2O1N2O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100kN[LkBe3k<eLUC[3j<gLVCX3i<iLWCW3h<kLWCU3h<mLXCR3g<oLYCQ3f<QMYCo2f<RM[Cm2d<UM[Ck2d<VM\\Cj2j;dKfDd1_Oh2i;iKdD`1Cg2f;nKdD]1Ee2d;TLdDW1He2b;XLdDT1Id2a;_N_Da1`;aN_D_1`;bNaD]1_;dN`D\\1_;eNaD[1^;gNaDY1^;hNbDX1^;iNaDW1^;jNbDV1];Y300O1O1O100O100000000O100000000O1000000"}, "label": "a kid watching a giraffe at the zoo"}]}
{"id": 345781, "image": "COCO_train2014_000000345781.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toddler girl looks behind her at the giraffe\"."}], "task": "refering", "answer_template": "The \"a toddler girl looks behind her at the giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 328, 329, 350, 351, 372, 373, 393, 394, 395, 414, 415, 416, 417, 436, 437, 438, 439, 458, 459, 460, 461, 480, 481, 482, 483], "bbox": [0.8266830065359477, 0.5880392156862745, 1.0, 1.0], "iscrowd": 0, "area": 16010.297049999992, "rle": {"size": [612, 612], "counts": "[P_91Qc02M3N2N2N2N2M3WNFg@;V?Oc@1[?8_@G_?<`@D^??a@A]?b0b@^O]?d0a@]O^?f0_@[O`?g0^@ZOa?h0]@YOa?j0\\@XOc?j0[@WOZ?ZO[@a19VO\\?[OX@a1:UO]?]OV@`1;TO^?W1`@jN_?X1_@iN`?Y1]@iNb?Y1\\@hNc?]2N2O1O1N2O1N2O1N2O1N2O1N2O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100kN[LkBe3k<eLUC[3j<gLVCX3i<iLWCW3h<kLWCU3h<mLXCR3g<oLYCQ3f<QMYCo2f<RM[Cm2d<UM[Ck2d<VM\\Cj2j;dKfDd1_Oh2i;iKdD`1Cg2f;nKdD]1Ee2d;TLdDW1He2b;XLdDT1Id2a;_N_Da1`;aN_D_1`;bNaD]1_;dN`D\\1_;eNaD[1^;gNaDY1^;hNbDX1^;iNaDW1^;jNbDV1];Y300O1O1O100O100000000O100000000O1000000"}, "label": "a toddler girl looks behind her at the giraffe"}]}
{"id": 272058, "image": "COCO_train2014_000000272058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rectangle white table\"."}], "task": "refering", "answer_template": "The \"rectangle white table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 127, 128, 129, 140, 141, 142, 152, 153, 154, 155, 166, 167, 168, 179, 180, 181, 193, 194, 205, 206, 207, 218, 219, 220], "bbox": [0.73408, 0.47416, 1.0, 0.93708], "iscrowd": 0, "area": 13974.470500000001, "rle": {"size": [500, 375], "counts": "PcV41a?3N1N3M2N3M3M2O2M3M3M4L3VDPOm8S1RGnNj8U1UGmN[75PGQ1e1kNY7?eFj0P2iNX7i0[Fa0]2gNV7d2hH^MU7d2jH^MS7e2lH\\MR7e2mH]MP7f2oH[Mn6g2QI[Mm6f2SI[Mj6g2UI[Mh6g2XIYMg6h2XIXMg6k2XIUMf6m2YISMg6n2YIRMe6P3ZIQMd6Q3[IoLe6R3ZIoLd6S3[InLc6T3]IkLc6V3\\IkLb6W3]IjLa6X3^IhLb6Y3]IhLa6[3^IeLa6]3]IcLc6_3[IbLe6^3YIdLf6^3VIdLj6^3RIeLn6\\3mHhLR7Y3kHiLU7Y3gHjLX7X3dHkL\\7V3`HlL`7U3]HmLc7U3YHlLh7V3SHmLn7T3nGmLS8T3jGmLW8U3eGmL\\8T3`GmLa8U3[GlLf8Q50O2N100O2N100O1O100O1O100O1YMoFjNR9V1RGfNn8Z1VGaNk8_1ZG[Ng8d1^GWNc8i1aGRN`8n1dGmM]8S2hGhMX8W2mGdMT8\\2PH_MQ8a2THYMm7g2WHTMj7l2ZHoLg7P3^HkLc7U3bHfL^7Z3fHaL[7_3iH\\LX7c3]20O100O01000OF;[Oe0[OTI"}, "label": "rectangle white table"}]}
{"id": 272058, "image": "COCO_train2014_000000272058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty table\"."}], "task": "refering", "answer_template": "The \"empty table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 127, 128, 129, 140, 141, 142, 152, 153, 154, 155, 166, 167, 168, 179, 180, 181, 193, 194, 205, 206, 207, 218, 219, 220], "bbox": [0.73408, 0.47416, 1.0, 0.93708], "iscrowd": 0, "area": 13974.470500000001, "rle": {"size": [500, 375], "counts": "PcV41a?3N1N3M2N3M3M2O2M3M3M4L3VDPOm8S1RGnNj8U1UGmN[75PGQ1e1kNY7?eFj0P2iNX7i0[Fa0]2gNV7d2hH^MU7d2jH^MS7e2lH\\MR7e2mH]MP7f2oH[Mn6g2QI[Mm6f2SI[Mj6g2UI[Mh6g2XIYMg6h2XIXMg6k2XIUMf6m2YISMg6n2YIRMe6P3ZIQMd6Q3[IoLe6R3ZIoLd6S3[InLc6T3]IkLc6V3\\IkLb6W3]IjLa6X3^IhLb6Y3]IhLa6[3^IeLa6]3]IcLc6_3[IbLe6^3YIdLf6^3VIdLj6^3RIeLn6\\3mHhLR7Y3kHiLU7Y3gHjLX7X3dHkL\\7V3`HlL`7U3]HmLc7U3YHlLh7V3SHmLn7T3nGmLS8T3jGmLW8U3eGmL\\8T3`GmLa8U3[GlLf8Q50O2N100O2N100O1O100O1O100O1YMoFjNR9V1RGfNn8Z1VGaNk8_1ZG[Ng8d1^GWNc8i1aGRN`8n1dGmM]8S2hGhMX8W2mGdMT8\\2PH_MQ8a2THYMm7g2WHTMj7l2ZHoLg7P3^HkLc7U3bHfL^7Z3fHaL[7_3iH\\LX7c3]20O100O01000OF;[Oe0[OTI"}, "label": "empty table"}]}
{"id": 272058, "image": "COCO_train2014_000000272058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cake on a white table\"."}], "task": "refering", "answer_template": "The \"a white cake on a white table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.0, 0.5777, 0.88888, 0.9853600000000001], "iscrowd": 0, "area": 51315.32805, "rle": {"size": [500, 375], "counts": "`:]3U<3N2N2O1O1N2O1N2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O2N1N2O1O1N2O1N2O1O0O10000O10001N100O10000O10000O10000O100O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10001N1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000001O0000001O00001O0000001O00001O0000001O0000010O00001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000010O00001O0O1O1O2N1O1O2M2O1O1O2N1O1O2N1N2O1O2N1O1O2N1O1N2O2N1O1O1O2N1N2N3M2N2N2N3M2N2N3M2N2N2N3M2N3M2N3N1O2N1O2N1O2N101N1O2N1O2N1100O10O0100O11O001O001O1N101O001O001O001O0O2O0O2N2N101N1O2N1O2N1O2N1O2N2M2M4L3M4L3N3L3M4L3M4L4M3L6JdSd0"}, "label": "a white cake on a white table"}]}
{"id": 272058, "image": "COCO_train2014_000000272058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a birthday cake\"."}], "task": "refering", "answer_template": "The \"a birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.0, 0.5777, 0.88888, 0.9853600000000001], "iscrowd": 0, "area": 51315.32805, "rle": {"size": [500, 375], "counts": "`:]3U<3N2N2O1O1N2O1N2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O2N1N2O1O1N2O1N2O1O0O10000O10001N100O10000O10000O10000O100O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10001N1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000001O0000001O00001O0000001O00001O0000001O0000010O00001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000010O00001O0O1O1O2N1O1O2M2O1O1O2N1O1O2N1N2O1O2N1O1O2N1O1N2O2N1O1O1O2N1N2N3M2N2N2N3M2N2N3M2N2N2N3M2N3M2N3N1O2N1O2N1O2N101N1O2N1O2N1100O10O0100O11O001O001O1N101O001O001O001O0O2O0O2N2N101N1O2N1O2N1O2N1O2N2M2M4L3M4L3N3L3M4L3M4L4M3L6JdSd0"}, "label": "a birthday cake"}]}
{"id": 132793, "image": "COCO_train2014_000000132793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large lantern floor light closest to desk\"."}], "task": "refering", "answer_template": "The \"large lantern floor light closest to desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 90, 91, 108, 109, 126, 127, 144, 145, 162, 163, 180, 181, 198, 199], "bbox": [0.002, 0.3840840840840841, 0.10244, 0.9650150150150151], "iscrowd": 0, "area": 6663.827600000001, "rle": {"size": [333, 500], "counts": "Zb07g9`0B=XMoNoKW1g3VOoKP1UOjNW3:[MS1UOkN_36RM[1fNSOW4FiLm2V3WM^LR3a3RMRLX3m3lLgK]3X4\\100000000000O10000000000000000O100000000eKgKb2Y4eLdLW3\\3cLmLZ3S3aLUM[3l2_L^M]3b2[LiMb3d4M5K6J6J7I6I8G9F;F:E<D;E;E^ga4"}, "label": "large lantern floor light closest to desk"}]}
{"id": 132793, "image": "COCO_train2014_000000132793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall , textured white vase stands next to a painted black - and - white portrait of audrey hepburn\"."}], "task": "refering", "answer_template": "The \"a tall , textured white vase stands next to a painted black - and - white portrait of audrey hepburn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 90, 91, 108, 109, 126, 127, 144, 145, 162, 163, 180, 181, 198, 199], "bbox": [0.002, 0.3840840840840841, 0.10244, 0.9650150150150151], "iscrowd": 0, "area": 6663.827600000001, "rle": {"size": [333, 500], "counts": "Zb07g9`0B=XMoNoKW1g3VOoKP1UOjNW3:[MS1UOkN_36RM[1fNSOW4FiLm2V3WM^LR3a3RMRLX3m3lLgK]3X4\\100000000000O10000000000000000O100000000eKgKb2Y4eLdLW3\\3cLmLZ3S3aLUM[3l2_L^M]3b2[LiMb3d4M5K6J6J7I6I8G9F;F:E<D;E;E^ga4"}, "label": "a tall , textured white vase stands next to a painted black - and - white portrait of audrey hepburn"}]}
{"id": 280257, "image": "COCO_train2014_000000280257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the middle with its head turned very slightly to the right\"."}], "task": "refering", "answer_template": "The \"the zebra in the middle with its head turned very slightly to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 62, 63, 64, 79, 80, 81, 82, 97, 98, 99, 100, 115, 116, 117, 118, 134, 135, 152, 153, 170, 171], "bbox": [0.4185, 0.22636636636636634, 0.6075, 0.8051951951951952], "iscrowd": 0, "area": 10854.24375, "rle": {"size": [333, 500], "counts": "_ST27l9?K5I6I6I7J6J6I7J7H7J6J6M3M3L4M3L4M3Mk0TOa0@0O2O001N100O2OO0O2N2N1O2M2O2A?ZOe0\\Oe0J510001N1O1O2]IZM\\5g2\\JdM`5]2XJnMe5`3L5K:F7I5K6J6K4K1O2N1O1O2O0O10O2bNcKjL]4Q3lKjLU4R3SLjLm3R3[LiLf3R3cLjL^3Q3jLjLW3Q3VMfLk2V3aMbL_2Y3]2M4N100O1O101N8H4LO2OfM^Hj1a7WNcHe1]7\\NeHb1Y7_NjH^1U7dNnHX1R7hNQIU1n6lNUIP1k6ROWIg0R7TOWI?Q7^OcIIf62o1L4Kgdo1"}, "label": "the zebra in the middle with its head turned very slightly to the right"}]}
{"id": 280257, "image": "COCO_train2014_000000280257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing in the center\"."}], "task": "refering", "answer_template": "The \"a zebra standing in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 62, 63, 64, 79, 80, 81, 82, 97, 98, 99, 100, 115, 116, 117, 118, 134, 135, 152, 153, 170, 171], "bbox": [0.4185, 0.22636636636636634, 0.6075, 0.8051951951951952], "iscrowd": 0, "area": 10854.24375, "rle": {"size": [333, 500], "counts": "_ST27l9?K5I6I6I7J6J6I7J7H7J6J6M3M3L4M3L4M3Mk0TOa0@0O2O001N100O2OO0O2N2N1O2M2O2A?ZOe0\\Oe0J510001N1O1O2]IZM\\5g2\\JdM`5]2XJnMe5`3L5K:F7I5K6J6K4K1O2N1O1O2O0O10O2bNcKjL]4Q3lKjLU4R3SLjLm3R3[LiLf3R3cLjL^3Q3jLjLW3Q3VMfLk2V3aMbL_2Y3]2M4N100O1O101N8H4LO2OfM^Hj1a7WNcHe1]7\\NeHb1Y7_NjH^1U7dNnHX1R7hNQIU1n6lNUIP1k6ROWIg0R7TOWI?Q7^OcIIf62o1L4Kgdo1"}, "label": "a zebra standing in the center"}]}
{"id": 280257, "image": "COCO_train2014_000000280257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left\"."}], "task": "refering", "answer_template": "The \"the zebra on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 55, 56, 57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 96, 109, 110, 111, 112, 128, 129, 130, 146, 147, 164, 165], "bbox": [0.06848, 0.18864864864864864, 0.34416, 0.8257657657657658], "iscrowd": 0, "area": 14388.589999999995, "rle": {"size": [333, 500], "counts": "_V;4V:9F;E;F4K5L4K5M3M3M3M2M4M3M3M3M3N2UJ`Mf3b2XL`Mg3a2VLaMi3b2SL`Ml3c2PL_MP4a3nJaLQ5^4O10O001O0O1N3L3NVKbKk3]4eKTLZ4P501O0O10000O2O00O0M4K5L3L5K5L3]NjJlMX5W2kJaMV5c2lJVMX5l2iJmL[5T3hJdL\\5]3fJ\\L^5i3`0?A>B9G4K6K4M3M3M3M4L5K5KO1N2O2N0H9E:O1N3N1O2M3N1N3]OPKgKR5T4e0L5K5L3O2N2N1O2N2N1O2N2N3N2M3N2N1O2N3N2M3`M\\HR2g7hM]HX2l7O1O1QNPHa1P8^NSH_1n7`NVH\\1k7cN^HU1b7jNkHi0V7VOnHf0S7XOZI=f6B\\I<e6C\\I<\\8O1N110000O100O100O100O100O100O10O101O1O1N1O2OX`Z3"}, "label": "the zebra on the left"}]}
{"id": 280257, "image": "COCO_train2014_000000280257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left with its face visible\"."}], "task": "refering", "answer_template": "The \"the zebra on the left with its face visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 55, 56, 57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 96, 109, 110, 111, 112, 128, 129, 130, 146, 147, 164, 165], "bbox": [0.06848, 0.18864864864864864, 0.34416, 0.8257657657657658], "iscrowd": 0, "area": 14388.589999999995, "rle": {"size": [333, 500], "counts": "_V;4V:9F;E;F4K5L4K5M3M3M3M2M4M3M3M3M3N2UJ`Mf3b2XL`Mg3a2VLaMi3b2SL`Ml3c2PL_MP4a3nJaLQ5^4O10O001O0O1N3L3NVKbKk3]4eKTLZ4P501O0O10000O2O00O0M4K5L3L5K5L3]NjJlMX5W2kJaMV5c2lJVMX5l2iJmL[5T3hJdL\\5]3fJ\\L^5i3`0?A>B9G4K6K4M3M3M3M4L5K5KO1N2O2N0H9E:O1N3N1O2M3N1N3]OPKgKR5T4e0L5K5L3O2N2N1O2N2N1O2N2N3N2M3N2N1O2N3N2M3`M\\HR2g7hM]HX2l7O1O1QNPHa1P8^NSH_1n7`NVH\\1k7cN^HU1b7jNkHi0V7VOnHf0S7XOZI=f6B\\I<e6C\\I<\\8O1N110000O100O100O100O100O100O10O101O1O1N1O2OX`Z3"}, "label": "the zebra on the left with its face visible"}]}
{"id": 280257, "image": "COCO_train2014_000000280257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra to the right backwards\"."}], "task": "refering", "answer_template": "The \"zebra to the right backwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 84, 85, 86, 87, 102, 103, 104, 105, 120, 121, 122, 123, 138, 139, 140, 141, 156, 157, 158, 159, 175, 176, 177, 193, 195], "bbox": [0.68696, 0.3056156156156156, 0.8665400000000001, 0.8471771771771772], "iscrowd": 0, "area": 11177.164650000004, "rle": {"size": [333, 500], "counts": "cf_31U:9F:F:F:F:K5L4L4L4L4L4LW1iN6K6I6J7I6L3M3M3L]OcJ[LZ5a3mJ_LP5^3UKaLj4[3]KaLd4\\3aK`La4_3bK^L_4d3bKXL_4j3Q12N2N2M3N2QKjKc3X4TLoKl3S4jKULV4m3aKZL_4h3WKaLh4b40000000001N100000M3M3L4L4L4M4K4L5L3N3M2N3M3M2N3M23N1O2M2O2M2O2N6I:G9F01O1O001UNmKZMS4`2UL^Ml3\\2\\LaMe3X2dLeM\\3V2kLhMV3R2RMkMn2P2ZMmMl2h1[MVNk2`1X3F;E;E;F`de0"}, "label": "zebra to the right backwards"}]}
{"id": 280257, "image": "COCO_train2014_000000280257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra , which you cannot see the face of\"."}], "task": "refering", "answer_template": "The \"zebra , which you cannot see the face of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 84, 85, 86, 87, 102, 103, 104, 105, 120, 121, 122, 123, 138, 139, 140, 141, 156, 157, 158, 159, 175, 176, 177, 193, 195], "bbox": [0.68696, 0.3056156156156156, 0.8665400000000001, 0.8471771771771772], "iscrowd": 0, "area": 11177.164650000004, "rle": {"size": [333, 500], "counts": "cf_31U:9F:F:F:F:K5L4L4L4L4L4LW1iN6K6I6J7I6L3M3M3L]OcJ[LZ5a3mJ_LP5^3UKaLj4[3]KaLd4\\3aK`La4_3bK^L_4d3bKXL_4j3Q12N2N2M3N2QKjKc3X4TLoKl3S4jKULV4m3aKZL_4h3WKaLh4b40000000001N100000M3M3L4L4L4M4K4L5L3N3M2N3M3M2N3M23N1O2M2O2M2O2N6I:G9F01O1O001UNmKZMS4`2UL^Ml3\\2\\LaMe3X2dLeM\\3V2kLhMV3R2RMkMn2P2ZMmMl2h1[MVNk2`1X3F;E;E;F`de0"}, "label": "zebra , which you cannot see the face of"}]}
{"id": 59079, "image": "COCO_train2014_000000059079.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bear on the right is showing teeth\"."}], "task": "refering", "answer_template": "The \"the bear on the right is showing teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 34, 35, 36, 37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321], "bbox": [0.398875, 0.042699999999999995, 0.9971875000000001, 0.948325], "iscrowd": 0, "area": 89620.75585000002, "rle": {"size": [400, 640], "counts": "WjS31\\<3L4M3M4L3M3L4M3O2M2N2N2N2N3N1N2N2N2N3N1N2N2N2N2O2M2N2N2N2O2M2M3M3M3M4M2M3M3M3M4L3M3VOgLZH\\3d7j0L3M3M4L3O1N3N1O1N2O2N1N2O2N1N2O2N1N2O2N1O1N3N1O1O2N1O2N1O2N1O2N1O2N1O2N2M2O2N1O2N1O2N1O2N1O2N3M3M3M4L3M3M3L4M3N2N2N2N2N2N3cMoG6R8FQH9Q8DPH<R8@PH`0R8\\OQHc0Q8YOQHg0Q8UOQHk0Q8QORHn0P8jNVHV1l7aN[H_1h7XN^Hi1c7oMdHP2n80001O00001O00000010O0001O00001O0000001O01O01O00001O0000001O00010O00001O0000001O000010O0001O0000001O00000001O0O1000000000000000001N1000000000001O1O0O2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O001N2O1O1O1O1O1O1O1O1N2O1O1O1O001O1O1O001N2O1O001O1O1O001O1O1O10O01O1O1O00100O1O001O1O100O001O1O1O010O1O1O1O00100O1O1O1O1O10O01O1O1O1O100O1O001O1O1O100O1O1O001O100O1O1O1O1O010O1O1O1O1O100O001O1O1O100O1O1O1O1O2O0O2N1O2N2N101N1O2N1O2N101N2N1O2N1O2O0O2N2N1O2N101N1O2N1O2N2O0O2N1O2N1O2N2O0O2N1O2N1O2O1N1O2N1O2N101N1O2N2N1O2O0O2N1O2N2K4F;D;Eb0^Oe0[Od0]Od0[Oe0[O^`0"}, "label": "the bear on the right is showing teeth"}]}
{"id": 59079, "image": "COCO_train2014_000000059079.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bear with its mouth open , playing in the water\"."}], "task": "refering", "answer_template": "The \"a brown bear with its mouth open , playing in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 34, 35, 36, 37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321], "bbox": [0.398875, 0.042699999999999995, 0.9971875000000001, 0.948325], "iscrowd": 0, "area": 89620.75585000002, "rle": {"size": [400, 640], "counts": "WjS31\\<3L4M3M4L3M3L4M3O2M2N2N2N2N3N1N2N2N2N3N1N2N2N2N2O2M2N2N2N2O2M2M3M3M3M4M2M3M3M3M4L3M3VOgLZH\\3d7j0L3M3M4L3O1N3N1O1N2O2N1N2O2N1N2O2N1N2O2N1O1N3N1O1O2N1O2N1O2N1O2N1O2N1O2N2M2O2N1O2N1O2N1O2N1O2N3M3M3M4L3M3M3L4M3N2N2N2N2N2N3cMoG6R8FQH9Q8DPH<R8@PH`0R8\\OQHc0Q8YOQHg0Q8UOQHk0Q8QORHn0P8jNVHV1l7aN[H_1h7XN^Hi1c7oMdHP2n80001O00001O00000010O0001O00001O0000001O01O01O00001O0000001O00010O00001O0000001O000010O0001O0000001O00000001O0O1000000000000000001N1000000000001O1O0O2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O001N2O1O1O1O1O1O1O1O1N2O1O1O1O001O1O1O001N2O1O001O1O1O001O1O1O10O01O1O1O00100O1O001O1O100O001O1O1O010O1O1O1O00100O1O1O1O1O10O01O1O1O1O100O1O001O1O1O100O1O1O001O100O1O1O1O1O010O1O1O1O1O100O001O1O1O100O1O1O1O1O2O0O2N1O2N2N101N1O2N1O2N101N2N1O2N1O2O0O2N2N1O2N101N1O2N1O2N2O0O2N1O2N1O2N2O0O2N1O2N1O2O1N1O2N1O2N101N1O2N2N1O2O0O2N1O2N2K4F;D;Eb0^Oe0[Od0]Od0[Oe0[O^`0"}, "label": "a brown bear with its mouth open , playing in the water"}]}
{"id": 59079, "image": "COCO_train2014_000000059079.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby bear to the left\"."}], "task": "refering", "answer_template": "The \"the baby bear to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285], "bbox": [0.015453125000000002, 0.31010000000000004, 0.5870781249999999, 0.9101], "iscrowd": 0, "area": 46840.397300000004, "rle": {"size": [400, 640], "counts": "jS44\\<5J6K5J5L5J6K3L101N101N101N2O0010O01O001O1O001O001O001O001O001O5K4L4L4K4L3N1N3M3N1N2N3N1N2N2O1N3M2O1N2N3N1N2N2N2N3M1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N101N1O2N1O2N101N101N101N100O2O0O100O2N1O1O1N3N1O1VOWJlJk5R5[JhJf5W5_JcJc5\\5bJ_J_5a5i0O0000001O0000001O00001O0000001O0000001O0000001O0009H9F8H1O2O1N1O2N2O0O2N2N101N1O100O2N1O100O1O1O100O1O100O1O2N100O1O1O100O1O1O100O2N100O1O1O10O0001O01O0000010O00000001O0O1000001O00001O001O0O101O00001O001O00001N101O00001O000000O1O001O100O1O1O1O1O0010O01O001O000010O101N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O2M2O0O1O100O100O1O100O100O100O100O100000O01000O100O01000O10O017I=C4L0O1000000O101O000O100000000O1000000O1000O10O010000O10O10O100O010O100O10O0100O10O1000000O1000000O100O1O1O100O1O3@`0K5K5J6J7I6JVlV3"}, "label": "the baby bear to the left"}]}
{"id": 59079, "image": "COCO_train2014_000000059079.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear in water engaging with a larger bear with his teeth beared\"."}], "task": "refering", "answer_template": "The \"a bear in water engaging with a larger bear with his teeth beared\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285], "bbox": [0.015453125000000002, 0.31010000000000004, 0.5870781249999999, 0.9101], "iscrowd": 0, "area": 46840.397300000004, "rle": {"size": [400, 640], "counts": "jS44\\<5J6K5J5L5J6K3L101N101N101N2O0010O01O001O1O001O001O001O001O001O5K4L4L4K4L3N1N3M3N1N2N3N1N2N2O1N3M2O1N2N3N1N2N2N2N3M1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N101N1O2N1O2N101N101N101N100O2O0O100O2N1O1O1N3N1O1VOWJlJk5R5[JhJf5W5_JcJc5\\5bJ_J_5a5i0O0000001O0000001O00001O0000001O0000001O0000001O0009H9F8H1O2O1N1O2N2O0O2N2N101N1O100O2N1O100O1O1O100O1O100O1O2N100O1O1O100O1O1O100O2N100O1O1O10O0001O01O0000010O00000001O0O1000001O00001O001O0O101O00001O001O00001N101O00001O000000O1O001O100O1O1O1O1O0010O01O001O000010O101N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O2M2O0O1O100O100O1O100O100O100O100O100000O01000O100O01000O10O017I=C4L0O1000000O101O000O100000000O1000000O1000O10O010000O10O10O100O010O100O10O0100O10O1000000O1000000O100O1O1O100O1O3@`0K5K5J6J7I6JVlV3"}, "label": "a bear in water engaging with a larger bear with his teeth beared"}]}
{"id": 42696, "image": "COCO_train2014_000000042696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leftmost elephant\"."}], "task": "refering", "answer_template": "The \"the leftmost elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349, 350, 351], "bbox": [0.0438125, 0.23145833333333332, 0.387640625, 0.9258333333333333], "iscrowd": 0, "area": 48845.91165, "rle": {"size": [480, 640], "counts": "_Y=2j>5L4K4L5L4K4L5K5L3L5N2N101N2N101N2O0O2O1N101N2O0O2O1N101N101mClMW;V2dDRNX;n1dDYNY;h1bD`NZ;a1aDgN[;Y1aDnN];^2L4L3M4L3M4L4L3M4L4L4K6J5L5J6J5K6K4K6J6J7J6I7I7I7I7UJ_Hk3f7mK^HR4h7fK[HY4k7^KYHa4m7WKVHg4Q8QKSHm4S8jJQHU5n8O1O1O0O100000000O010000000O1000000000O101O0000000000000O100000000004L3M4L4L3M1O0000000O10000000000001O0N2L4L5K4N21O01O01O0010O0001O01O01O01O100O101N1O100O1O2O0O1O101N1N2O0O2O0O101N100O2O000O2O0O101J5D<10O1O010O10O0100O1O010O10O01O1M3L3N3lM\\IgKh6U4aIcKa6Z4hI]K\\6_4mIYKV6c4SJTKQ6h4XJPKj5m4_JjJe5R5W2L3N3L4L4L3L5J6K4K6K5K5J6K7H9H7I8G8H8G:F9H8G:F9H9Feag5"}, "label": "the leftmost elephant"}]}
{"id": 42696, "image": "COCO_train2014_000000042696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one elephant is coming in front from other elephants\"."}], "task": "refering", "answer_template": "The \"one elephant is coming in front from other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349, 350, 351], "bbox": [0.0438125, 0.23145833333333332, 0.387640625, 0.9258333333333333], "iscrowd": 0, "area": 48845.91165, "rle": {"size": [480, 640], "counts": "_Y=2j>5L4K4L5L4K4L5K5L3L5N2N101N2N101N2O0O2O1N101N2O0O2O1N101N101mClMW;V2dDRNX;n1dDYNY;h1bD`NZ;a1aDgN[;Y1aDnN];^2L4L3M4L3M4L4L3M4L4L4K6J5L5J6J5K6K4K6J6J7J6I7I7I7I7UJ_Hk3f7mK^HR4h7fK[HY4k7^KYHa4m7WKVHg4Q8QKSHm4S8jJQHU5n8O1O1O0O100000000O010000000O1000000000O101O0000000000000O100000000004L3M4L4L3M1O0000000O10000000000001O0N2L4L5K4N21O01O01O0010O0001O01O01O01O100O101N1O100O1O2O0O1O101N1N2O0O2O0O101N100O2O000O2O0O101J5D<10O1O010O10O0100O1O010O10O01O1M3L3N3lM\\IgKh6U4aIcKa6Z4hI]K\\6_4mIYKV6c4SJTKQ6h4XJPKj5m4_JjJe5R5W2L3N3L4L4L3L5J6K4K6K5K5J6K7H9H7I8G8H8G:F9H8G:F9H9Feag5"}, "label": "one elephant is coming in front from other elephants"}]}
{"id": 42696, "image": "COCO_train2014_000000042696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant whose head is not shown\"."}], "task": "refering", "answer_template": "The \"the elephant whose head is not shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 265, 266, 267, 283, 284, 285, 288, 289, 290, 307, 311, 312], "bbox": [0.338765625, 0.37752083333333336, 0.648875, 0.7752708333333334], "iscrowd": 0, "area": 26110.1528, "rle": {"size": [480, 640], "counts": "aQV39c>>B=C9G4L4L4L4L3M4M3L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4]NmKVHV4j7nKnGV4R8nKfGU4[8oK]GU4c8oKUGU4k8nKnFV4R9l000000O1I7I7I7I7I7I7I7I6K6K5000000001O00000000001O00000000001O000O1000001O00000000001O00000000001O01O00000001O00000000001O00000000001O00000000001O00000001O01O000001O0001O01O00000010O000eFWL_7i3VHbLk7]3jGnLV8R3_GYMa8h2SGcMm8X4001O0000001O0O10001O000000001O0000001O0000001N100000001O0000001O0000001O0000001N2O2N1O2N1O2N1O1O2H7G:E:G:E:I8J5L5K4L5J5L4L5K4K6K4L5K4KhWY3"}, "label": "the elephant whose head is not shown"}]}
{"id": 42696, "image": "COCO_train2014_000000042696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant hiding its face\"."}], "task": "refering", "answer_template": "The \"an elephant hiding its face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 265, 266, 267, 283, 284, 285, 288, 289, 290, 307, 311, 312], "bbox": [0.338765625, 0.37752083333333336, 0.648875, 0.7752708333333334], "iscrowd": 0, "area": 26110.1528, "rle": {"size": [480, 640], "counts": "aQV39c>>B=C9G4L4L4L4L3M4M3L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4]NmKVHV4j7nKnGV4R8nKfGU4[8oK]GU4c8oKUGU4k8nKnFV4R9l000000O1I7I7I7I7I7I7I7I6K6K5000000001O00000000001O00000000001O000O1000001O00000000001O00000000001O01O00000001O00000000001O00000000001O00000000001O00000001O01O000001O0001O01O00000010O000eFWL_7i3VHbLk7]3jGnLV8R3_GYMa8h2SGcMm8X4001O0000001O0O10001O000000001O0000001O0000001N100000001O0000001O0000001O0000001N2O2N1O2N1O2N1O1O2H7G:E:G:E:I8J5L5K4L5J5L4L5K4K6K4L5K4KhWY3"}, "label": "an elephant hiding its face"}]}
{"id": 42696, "image": "COCO_train2014_000000042696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a elephant facing the camera with its young child leaning up against it\"."}], "task": "refering", "answer_template": "The \"a elephant facing the camera with its young child leaning up against it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316], "bbox": [0.635390625, 0.2667291666666667, 0.9707812499999999, 0.8060624999999999], "iscrowd": 0, "area": 30647.98484999999, "rle": {"size": [480, 640], "counts": "mon57c>7I7H8K5L4M3L5K4M3L4M3L4L4M3M3O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2O1N3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3fGoJ`6T5oHjKa6Y4nHXLo6k3`HeL^7_5O2N1O2N100O2O010O010O010O0010O010O01O010O01O010O01O010OB?ZOe0N25L2M01O001O01O01O001O00001O010O001O00001O0010O0001O001O0010O00^IdG`6^801O00001QKdG;h04d7BjG_Ob1d0c6NYIaNNU1j6:UIgNVO3J`0k7g0RIlNnN?5Ck7S1oHQOfNk0a0eNk7_1kHV1f7kNWHb1^7_N`Hj1Z7UNeHk1\\7UNdHj1]7VNbHj1_7VN`Hj1a7VN^Hj1c7UN]Hk1e7RN]Hm1d7QN]Ho1d7oM]Hi1l7UNUHQ1f8mNZGQ1j8nNUGQ1n8mNSGP1R9mNnFQ1V9mNjFQ1Z9mNgFQ1\\9mNdFQ1`9mN`FQ1d9nN\\FS1e9jN[FX1U:VNkEl1d;1N3N1N2N2O1N`0@T1mNO01O100O01000000O1000000O2O00000O100O1O1O1O1O1kNkD^NV;b1oDXNS;g1SEQNP;n1VEkMl:T2YEfMi:[2n01O1O1O001O1O1O1O1O1N110000N2M3N2nNfCZO\\<a0jC[OY<a0lC\\OV<`0nC^OT<;TDAo;4]DIe;LgD1V=M4M[b8"}, "label": "a elephant facing the camera with its young child leaning up against it"}]}
{"id": 42696, "image": "COCO_train2014_000000042696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant with the baby elephant\"."}], "task": "refering", "answer_template": "The \"the elephant with the baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316], "bbox": [0.635390625, 0.2667291666666667, 0.9707812499999999, 0.8060624999999999], "iscrowd": 0, "area": 30647.98484999999, "rle": {"size": [480, 640], "counts": "mon57c>7I7H8K5L4M3L5K4M3L4M3L4L4M3M3O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2O1N3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3fGoJ`6T5oHjKa6Y4nHXLo6k3`HeL^7_5O2N1O2N100O2O010O010O010O0010O010O01O010O01O010O01O010OB?ZOe0N25L2M01O001O01O01O001O00001O010O001O00001O0010O0001O001O0010O00^IdG`6^801O00001QKdG;h04d7BjG_Ob1d0c6NYIaNNU1j6:UIgNVO3J`0k7g0RIlNnN?5Ck7S1oHQOfNk0a0eNk7_1kHV1f7kNWHb1^7_N`Hj1Z7UNeHk1\\7UNdHj1]7VNbHj1_7VN`Hj1a7VN^Hj1c7UN]Hk1e7RN]Hm1d7QN]Ho1d7oM]Hi1l7UNUHQ1f8mNZGQ1j8nNUGQ1n8mNSGP1R9mNnFQ1V9mNjFQ1Z9mNgFQ1\\9mNdFQ1`9mN`FQ1d9nN\\FS1e9jN[FX1U:VNkEl1d;1N3N1N2N2O1N`0@T1mNO01O100O01000000O1000000O2O00000O100O1O1O1O1O1kNkD^NV;b1oDXNS;g1SEQNP;n1VEkMl:T2YEfMi:[2n01O1O1O001O1O1O1O1O1N110000N2M3N2nNfCZO\\<a0jC[OY<a0lC\\OV<`0nC^OT<;TDAo;4]DIe;LgD1V=M4M[b8"}, "label": "the elephant with the baby elephant"}]}
{"id": 526029, "image": "COCO_train2014_000000526029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black suitcase\"."}], "task": "refering", "answer_template": "The \"the black suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.687921875, 0.0718125, 1.0, 0.9263958333333332], "iscrowd": 0, "area": 71623.10829999998, "rle": {"size": [480, 640], "counts": "Ve^63d>:G:F:E;F:E:G:F:E;F:F9F9H8H8G9H8G9H8H8G9H8H8G9H8H8oLQIhMX7o1YIaMo6X2`IXMh6`2hIoLa6j2nIfLZ6R3n2I7H8N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O100001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000000000001O0000000000000001O0000000001O000000000000h1XNo1QNo1QNo1QNn1RNo1QNSC"}, "label": "the black suitcase"}]}
{"id": 526029, "image": "COCO_train2014_000000526029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a solid black carry - on suitcase\"."}], "task": "refering", "answer_template": "The \"a solid black carry - on suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.687921875, 0.0718125, 1.0, 0.9263958333333332], "iscrowd": 0, "area": 71623.10829999998, "rle": {"size": [480, 640], "counts": "Ve^63d>:G:F:E;F:E:G:F:E;F:F9F9H8H8G9H8G9H8H8G9H8H8G9H8H8oLQIhMX7o1YIaMo6X2`IXMh6`2hIoLa6j2nIfLZ6R3n2I7H8N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O100001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000000000001O0000000000000001O0000000001O000000000000h1XNo1QNo1QNo1QNn1RNo1QNSC"}, "label": "a solid black carry - on suitcase"}]}
{"id": 526029, "image": "COCO_train2014_000000526029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bag with its zip open placed in the middle\"."}], "task": "refering", "answer_template": "The \"bag with its zip open placed in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.228703125, 0.1179375, 0.729828125, 0.8735416666666667], "iscrowd": 0, "area": 101923.09545000002, "rle": {"size": [480, 640], "counts": "ecT24l=S1H8H9H7J6K6J5J6K6J5K5K6I6K5K6J5K5J7J5K5K6I6K5K6J5K5J7J5K5K5K6I6K5K6J5cJhHi2^7SMfHh2`7SMeHg2`7VMdHe2`7WMcHg2a7TM`Hl2c7QM^Hn2e7nL\\HR3h7jLXHV3k7fLVHZ3m7bLTH^3P8^LPHb3S8ZLnGf3U8VLlGj3X8RLhGn3[8oKeGQ4^8lKbGT4b8hK^GX4e8eK[G[4h8bKXG^4e90000000002N2N1O2M3N1O1O1O001O00001O001O00001O001O001O00001O00000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000N2O1O1O1N2O1O1N2O1O1O1N3N2N1O2L4H8H7I8H8H8H8H7I8I7H8H7I8H8H8H8H7I8H8H`0A`0_Oa0_Oa0_Oa0_Oa0_Oa1_Ngm`2"}, "label": "bag with its zip open placed in the middle"}]}
{"id": 526029, "image": "COCO_train2014_000000526029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey suitcase in the middle\"."}], "task": "refering", "answer_template": "The \"grey suitcase in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.228703125, 0.1179375, 0.729828125, 0.8735416666666667], "iscrowd": 0, "area": 101923.09545000002, "rle": {"size": [480, 640], "counts": "ecT24l=S1H8H9H7J6K6J5J6K6J5K5K6I6K5K6J5K5J7J5K5K6I6K5K6J5K5J7J5K5K5K6I6K5K6J5cJhHi2^7SMfHh2`7SMeHg2`7VMdHe2`7WMcHg2a7TM`Hl2c7QM^Hn2e7nL\\HR3h7jLXHV3k7fLVHZ3m7bLTH^3P8^LPHb3S8ZLnGf3U8VLlGj3X8RLhGn3[8oKeGQ4^8lKbGT4b8hK^GX4e8eK[G[4h8bKXG^4e90000000002N2N1O2M3N1O1O1O001O00001O001O00001O001O001O00001O00000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000N2O1O1O1N2O1O1N2O1O1O1N3N2N1O2L4H8H7I8H8H8H8H7I8I7H8H7I8H8H8H8H7I8H8H`0A`0_Oa0_Oa0_Oa0_Oa0_Oa1_Ngm`2"}, "label": "grey suitcase in the middle"}]}
{"id": 353999, "image": "COCO_train2014_000000353999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a furry jacket\"."}], "task": "refering", "answer_template": "The \"a girl in a furry jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.0026562499999999998, 0.2815384615384615, 0.51875, 0.986498673740053], "iscrowd": 0, "area": 46391.0702, "rle": {"size": [377, 640], "counts": "QS15\\;8I7H8I7M300O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O10000000000O100000000000000000000000000O100000000O1M3N2N2N2N2N2N2O100O1O100O1O100O1O100O1O100O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N20000O1000000O10000O10000O10000O1000000O10000O100000000000000000000000000000000000000000000000000000000001O00001O00O1M3M3M3K5L4L4XNPLbKU4Z4VL[Kn3b4]LRKg3k4dLhJa3U5c1M3L4M3M3M3M3M3M3L4M3M3M3N2O1N2O1O1O1N2O1O1O1N2O1O100O100O100O1O100O100O100O100O100O1000000000000000000000000000000000000000000000000001O00000000001O000000001O000000001O000000006J7I7I7I4L3M3M4L2N1O1O1O001O1O1N2L4L4L4L3N3L4L4L4L4L4L4L3M4L4K4M4L3M4M3N1N3M2O2M3M2O2M3N3L3M4M3L4VOfGoM]8l1mGmMV8n1PHmMT8m1P1J5J7I7I7K5J6K5K5K3M3M3M3MTVa3"}, "label": "a girl in a furry jacket"}]}
{"id": 353999, "image": "COCO_train2014_000000353999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing animal print jacket with her painted face against a man ' s chest\"."}], "task": "refering", "answer_template": "The \"a woman wearing animal print jacket with her painted face against a man ' s chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.0026562499999999998, 0.2815384615384615, 0.51875, 0.986498673740053], "iscrowd": 0, "area": 46391.0702, "rle": {"size": [377, 640], "counts": "QS15\\;8I7H8I7M300O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O10000000000O100000000000000000000000000O100000000O1M3N2N2N2N2N2N2O100O1O100O1O100O1O100O1O100O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N20000O1000000O10000O10000O10000O1000000O10000O100000000000000000000000000000000000000000000000000000000001O00001O00O1M3M3M3K5L4L4XNPLbKU4Z4VL[Kn3b4]LRKg3k4dLhJa3U5c1M3L4M3M3M3M3M3M3L4M3M3M3N2O1N2O1O1O1N2O1O1O1N2O1O100O100O100O1O100O100O100O100O100O1000000000000000000000000000000000000000000000000001O00000000001O000000001O000000001O000000006J7I7I7I4L3M3M4L2N1O1O1O001O1O1N2L4L4L4L3N3L4L4L4L4L4L4L3M4L4K4M4L3M4M3N1N3M2O2M3M2O2M3N3L3M4M3L4VOfGoM]8l1mGmMV8n1PHmMT8m1P1J5J7I7I7K5J6K5K5K3M3M3M3MTVa3"}, "label": "a woman wearing animal print jacket with her painted face against a man ' s chest"}]}
{"id": 353999, "image": "COCO_train2014_000000353999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing smudged clown make - up and a dark gray suit\"."}], "task": "refering", "answer_template": "The \"a man wearing smudged clown make - up and a dark gray suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.381234375, 0.003819628647214854, 0.99809375, 0.9858355437665783], "iscrowd": 0, "area": 83910.45384999999, "rle": {"size": [377, 640], "counts": "Wki29_;b0]O8I6J6I7J6J6I7J4L4K6K4K5M3M1O2N1O2N2N2O5J7I7I6J2N2N1O1O2N1O2N1O1O2N1O2N1O1O1EVKeIk4W6XKiIi4S6ZKmIg4o5\\KQJe4l5]KTJd4h5_KXJb4d5aK\\J`4`5cK`J^4]5dKcJ]4Y5fKgJ[4U5hKkJY4R5iKnJX4n4kKRKV4j4mKVKT4f4oKZKR4c4PL]KQ4_4RLaKo3[4TLeKl3Y4VLgKk3U4XLkKi3Q4ZLoKg3o3ZLQLg3l3[LTLf3j3[LVLf3g3\\LYLe3e3\\L[Le3b3]L^Ld3`3]L`Ld3]3^LcLc3Z3_LfLa3Y3`LgLa3V3aLjL`3_2^JfMS2K`3Y2cJiMo1N`3R2hJmMi11`3m1lJoMf14`3h1nJRNc16`3d1cM\\N_2^1eMbN\\2Z1gMfN[2T1iMlNX2P1kMPOW2j0mMVOT2f0oMZOS2`0QN@P2<SNDo14WNLj1M\\N3e6000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O2N4lEZOm8j0hFZO[9j0ZF[Oh9Y12M3N2N2M300000jGgNl5Y1SJoNf5Q1YJWO`5i0_J^O[5b0dJAZ5?eJCZ5=eJE[5:dJG\\59dJH[5;aJG^5<^JFa5=[JEd5?WJCh5a0SJAl5d0nI]OR6g0iI[OV6j0dIXO[6l0`IVO_6o0[ISOd6Q1WIQOh6T1SIlNm6Q3O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O001O001O001O001O010O001O001O001O001O001O00001O001N101O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O0O2O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O0YNg4"}, "label": "a man wearing smudged clown make - up and a dark gray suit"}]}
{"id": 353999, "image": "COCO_train2014_000000353999.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man is dressed as a clown for halloween with his face painted and with his lady companion who is also face painted as a clown\"."}], "task": "refering", "answer_template": "The \"the man is dressed as a clown for halloween with his face painted and with his lady companion who is also face painted as a clown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.381234375, 0.003819628647214854, 0.99809375, 0.9858355437665783], "iscrowd": 0, "area": 83910.45384999999, "rle": {"size": [377, 640], "counts": "Wki29_;b0]O8I6J6I7J6J6I7J4L4K6K4K5M3M1O2N1O2N2N2O5J7I7I6J2N2N1O1O2N1O2N1O1O2N1O2N1O1O1EVKeIk4W6XKiIi4S6ZKmIg4o5\\KQJe4l5]KTJd4h5_KXJb4d5aK\\J`4`5cK`J^4]5dKcJ]4Y5fKgJ[4U5hKkJY4R5iKnJX4n4kKRKV4j4mKVKT4f4oKZKR4c4PL]KQ4_4RLaKo3[4TLeKl3Y4VLgKk3U4XLkKi3Q4ZLoKg3o3ZLQLg3l3[LTLf3j3[LVLf3g3\\LYLe3e3\\L[Le3b3]L^Ld3`3]L`Ld3]3^LcLc3Z3_LfLa3Y3`LgLa3V3aLjL`3_2^JfMS2K`3Y2cJiMo1N`3R2hJmMi11`3m1lJoMf14`3h1nJRNc16`3d1cM\\N_2^1eMbN\\2Z1gMfN[2T1iMlNX2P1kMPOW2j0mMVOT2f0oMZOS2`0QN@P2<SNDo14WNLj1M\\N3e6000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O2N4lEZOm8j0hFZO[9j0ZF[Oh9Y12M3N2N2M300000jGgNl5Y1SJoNf5Q1YJWO`5i0_J^O[5b0dJAZ5?eJCZ5=eJE[5:dJG\\59dJH[5;aJG^5<^JFa5=[JEd5?WJCh5a0SJAl5d0nI]OR6g0iI[OV6j0dIXO[6l0`IVO_6o0[ISOd6Q1WIQOh6T1SIlNm6Q3O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O001O001O001O001O010O001O001O001O001O001O00001O001N101O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O0O2O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O0YNg4"}, "label": "the man is dressed as a clown for halloween with his face painted and with his lady companion who is also face painted as a clown"}]}
{"id": 566992, "image": "COCO_train2014_000000566992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white , tallest refrigerator\"."}], "task": "refering", "answer_template": "The \"a white , tallest refrigerator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 123, 124, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264], "bbox": [0.3656875, 0.29327868852459016, 0.52240625, 0.7899297423887588], "iscrowd": 0, "area": 14363.466049999997, "rle": {"size": [427, 640], "counts": "`fQ3<m<d0^O9G000000000000000O100000000000000000000000000O1000000000>B`0@`0@`0@`0@`0@`0@?A`0@`0@;E00000000O100000000O100000000O100000000O10000000000O100000000O100000O10O100000000O100000000O100000000O100000o0QOS1mNR1mNT1mNS1mNoYo3"}, "label": "a white , tallest refrigerator"}]}
{"id": 566992, "image": "COCO_train2014_000000566992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bigger fridge in the photo\"."}], "task": "refering", "answer_template": "The \"the bigger fridge in the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 123, 124, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264], "bbox": [0.3656875, 0.29327868852459016, 0.52240625, 0.7899297423887588], "iscrowd": 0, "area": 14363.466049999997, "rle": {"size": [427, 640], "counts": "`fQ3<m<d0^O9G000000000000000O100000000000000000000000000O1000000000>B`0@`0@`0@`0@`0@`0@?A`0@`0@;E00000000O100000000O100000000O100000000O10000000000O100000000O100000O10O100000000O100000000O100000000O100000o0QOS1mNR1mNT1mNS1mNoYo3"}, "label": "the bigger fridge in the photo"}]}
{"id": 141015, "image": "COCO_train2014_000000141015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closer white chair\"."}], "task": "refering", "answer_template": "The \"the closer white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 265, 266, 298, 299, 300, 301, 314, 315, 316, 317, 332, 333, 334], "bbox": [0.6132888888888889, 0.7114833333333334, 0.8897111111111111, 0.9994833333333334], "iscrowd": 0, "area": 2811.2192999999947, "rle": {"size": [600, 450], "counts": "WUR54_b0;F4N3OO0001RAAb;?ZDEf;;VDIj;7RDMo;3mCOT<1hC3X<MeC6[<JfC1_<NbCJe<6\\CGf<9[CEg<:PC^OPN6Q?<nB@QN2R?>mBBoMOV?>jBFoMLV?>kBHnMJW?>jBJnMHX?=iBNnMDY?>jBOlMCZ?=kB5T=KnB3S=LnB4R=LnB4Q=LQC2P=MQC3n<MSC3m<MSC3m<LUC3gMJc>3fC4fM5Y>ERD6eM5Y>ERD7kMLT>LSD7VN_Oi=9QD8h<GYC:f<FZC:c?0O0010O0001O01O0001O01O01O01O0001O01O01O0001O0000010O00001O0000001O0000001O000002ONj]OCSb0=n]ODPb0<Q^OEna0:R^OOea01\\^O<Va0Dk^Oc0m`0^OT_Oh0e`0WOk_O=Q`0C_@Ma?3b101O00000010O0001O0000001N2M2N3N2M2113N2M2N3M31N3N2NO1N1O2ON1M3LPgl0"}, "label": "the closer white chair"}]}
{"id": 263896, "image": "COCO_train2014_000000263896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog responded to his owner ' s command\"."}], "task": "refering", "answer_template": "The \"a brown dog responded to his owner ' s command\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 297, 298, 321, 344, 367, 390], "bbox": [0.7455, 0.3833682008368201, 1.0, 0.9951464435146443], "iscrowd": 0, "area": 19570.3665, "rle": {"size": [478, 640], "counts": "Qln62k>7I4L4L3M4M3L5L4L3M2N1O2N1N3N1O2N1O001O1O1O1O1O1O100O001O1O001O001O001O1O001O001O001O001O1O001O001O1O001O001N2O00001O000O2O00001O000O2O0O2O1Nf0[O1N2O1N101O1N2O1N1N3N2M2O2M2O2M3O001O1O0O2O1O001O0O2O1O1O001N2O1O001O1N2O001O1O1N101O1O1O0O2O1O001O1N2O001O1O1O1O1N2O1O1O1O1O1N2O001O1O2N2M2O2N1O2O1N2N2N2O2M2N3N1N3M:F:G<C=C>B<E2MM6ZNdK"}, "label": "a brown dog responded to his owner ' s command"}]}
{"id": 263896, "image": "COCO_train2014_000000263896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog next to the table\"."}], "task": "refering", "answer_template": "The \"the dog next to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 297, 298, 321, 344, 367, 390], "bbox": [0.7455, 0.3833682008368201, 1.0, 0.9951464435146443], "iscrowd": 0, "area": 19570.3665, "rle": {"size": [478, 640], "counts": "Qln62k>7I4L4L3M4M3L5L4L3M2N1O2N1N3N1O2N1O001O1O1O1O1O1O100O001O1O001O001O001O1O001O001O001O001O1O001O001O1O001O001N2O00001O000O2O00001O000O2O0O2O1Nf0[O1N2O1N101O1N2O1N1N3N2M2O2M2O2M3O001O1O0O2O1O001O0O2O1O1O001N2O1O001O1N2O001O1O1N101O1O1O0O2O1O001O1N2O001O1O1O1O1N2O1O1O1O1O1N2O001O1O2N2M2O2N1O2O1N2N2N2O2M2N3N1N3M:F:G<C=C>B<E2MM6ZNdK"}, "label": "the dog next to the table"}]}
{"id": 378586, "image": "COCO_train2014_000000378586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wing of a plane in front of other planes\"."}], "task": "refering", "answer_template": "The \"a wing of a plane in front of other planes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 182, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.0, 0.5061826697892271, 0.9545468749999999, 0.9839110070257611], "iscrowd": 0, "area": 83534.14385, "rle": {"size": [427, 640], "counts": "o:U2R9T200000000001O00000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10O10000000000000O10000000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000mM^H]Na7g30nM_HZNa7f1`HYN`7g1`HYN`7g1aHXN_7h1bHWN^7j1aHVN^7k1cHTN]7l1dHSN\\7m1dHSN\\7m1eHRN[7n1fHQNZ7P2eHPN[7P2fHoMY7R2hHmMX7S2iHlMW7T2iHlMW7T2jHkMV7V2jHiMV7W2jHiMV7W2kHhMT7Y2mHfMS7Z2mHfMS7Z2nHeMR7\\2nHcMR7]2nHcMR7]2oHbMQ7^2PIaMo6`2QI]MR7c2oHYMT7h2lHTMW7l2iHRMY7n2hHnL[7R3fHjL]7V3cHgL_7Z3bHcL`7^3`H^Lc7b3]H[Lf7e3[HWLh7i3YHTLi7l3WHQLl7o3g02O00O1O100O00100O1O100O1O00100O1O100O1O010O1O100O1O10O01O100O1O1O10O01O100O1O100O00100O1O100O001O100O1O100O00100O1O100O1O010O1000000000O1000O1AiEjMW:U2kEjMU:U2lEkMT:S2oElMP:T2QFlMo9S2SFlMm9S2TFlMm9S2UFlMk9R2WFnMi9Q2YFnMg9Q2ZFoMe9Q2]FnMc9Q2_FnMa9P2aFoM`9P2bFoM^9P2cFPN]9o1eFPN[9o1fFQNZ9m1iFRNW9m1jFSNW9k1U1O1O1N2O1O1O1O1N2O1O1000000O1000000000000000001N1BUD\\Ok;c0VD]Oj;c0WD\\Oi;c0XD]Oh;b0ZD]Of;c0ZD]Of;b0\\D]Od;b0^D]Ob;c0^D]Ob;b0`D]O`;b0aD]O`;c0aD\\O_;c0bD]O^;b0dD]O\\;b0fD]OZ;c0fD]OZ;b0hD]OX;b0iD^OW;b0jD]OV;b0kD^OU;a0k000O2N100O1O1O100O1O100O1O100O1O100O1O1O100O1O10Tl;"}, "label": "a wing of a plane in front of other planes"}]}
{"id": 378586, "image": "COCO_train2014_000000378586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wing of a gray / blue airplane\"."}], "task": "refering", "answer_template": "The \"the wing of a gray / blue airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 182, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.0, 0.5061826697892271, 0.9545468749999999, 0.9839110070257611], "iscrowd": 0, "area": 83534.14385, "rle": {"size": [427, 640], "counts": "o:U2R9T200000000001O00000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10O10000000000000O10000000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000mM^H]Na7g30nM_HZNa7f1`HYN`7g1`HYN`7g1aHXN_7h1bHWN^7j1aHVN^7k1cHTN]7l1dHSN\\7m1dHSN\\7m1eHRN[7n1fHQNZ7P2eHPN[7P2fHoMY7R2hHmMX7S2iHlMW7T2iHlMW7T2jHkMV7V2jHiMV7W2jHiMV7W2kHhMT7Y2mHfMS7Z2mHfMS7Z2nHeMR7\\2nHcMR7]2nHcMR7]2oHbMQ7^2PIaMo6`2QI]MR7c2oHYMT7h2lHTMW7l2iHRMY7n2hHnL[7R3fHjL]7V3cHgL_7Z3bHcL`7^3`H^Lc7b3]H[Lf7e3[HWLh7i3YHTLi7l3WHQLl7o3g02O00O1O100O00100O1O100O1O00100O1O100O1O010O1O100O1O10O01O100O1O1O10O01O100O1O100O00100O1O100O001O100O1O100O00100O1O100O1O010O1000000000O1000O1AiEjMW:U2kEjMU:U2lEkMT:S2oElMP:T2QFlMo9S2SFlMm9S2TFlMm9S2UFlMk9R2WFnMi9Q2YFnMg9Q2ZFoMe9Q2]FnMc9Q2_FnMa9P2aFoM`9P2bFoM^9P2cFPN]9o1eFPN[9o1fFQNZ9m1iFRNW9m1jFSNW9k1U1O1O1N2O1O1O1O1N2O1O1000000O1000000000000000001N1BUD\\Ok;c0VD]Oj;c0WD\\Oi;c0XD]Oh;b0ZD]Of;c0ZD]Of;b0\\D]Od;b0^D]Ob;c0^D]Ob;b0`D]O`;b0aD]O`;c0aD\\O_;c0bD]O^;b0dD]O\\;b0fD]OZ;c0fD]OZ;b0hD]OX;b0iD^OW;b0jD]OV;b0kD^OU;a0k000O2N100O1O1O100O1O100O1O100O1O100O1O1O100O1O10Tl;"}, "label": "the wing of a gray / blue airplane"}]}
{"id": 378586, "image": "COCO_train2014_000000378586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tail end of plane with sas on back\"."}], "task": "refering", "answer_template": "The \"tail end of plane with sas on back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 99, 100, 101, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211], "bbox": [0.0, 0.2296018735362998, 0.475, 0.6195081967213115], "iscrowd": 0, "area": 21403.414349999992, "rle": {"size": [427, 640], "counts": "X7n0U;X100000000000000000001O000000000000000000000000000000001O000000000001O000000000000000001OO10000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O00000000000000O1000001N100000000O1000000O100000000O1000000O100000000O2O00000O100000000O1000000O010000O10O1000O1000O010000O0100000O1O001O1O1O001O1O1O1O001O100O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O010O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1OZOaFaM_9]2dFcM[9[2hFeMW9Y2lFgMS9X2oFgMQ9W2RGiMm8U2UGlMj8R2YGnMf8P2]GoMc8o1`GQN_8m1dGSN[8k1hGUNX8h1kGWNV8f1mGZNS8c1PH]NP8`1RHaNo7[1THdNm7Z1UHfNk7W1XHiNh7T1[HlNe7Q1^HnNc7o0`HjNZNFV9]1cHmNe7o0]HQOd7l0_HTOa7i0bHWO^7f0eHYO\\7d0gH\\OY7a0jH_OV7>mHAT7<oHDR79PIGP76RIJo63TIMl60WI0i6MZI3f6J]I5m90000O10O10O10000O10000O10000O2O000O2O0000Qm[4"}, "label": "tail end of plane with sas on back"}]}
{"id": 378586, "image": "COCO_train2014_000000378586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plane with sas on tail\"."}], "task": "refering", "answer_template": "The \"plane with sas on tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 99, 100, 101, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211], "bbox": [0.0, 0.2296018735362998, 0.475, 0.6195081967213115], "iscrowd": 0, "area": 21403.414349999992, "rle": {"size": [427, 640], "counts": "X7n0U;X100000000000000000001O000000000000000000000000000000001O000000000001O000000000000000001OO10000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O00000000000000O1000001N100000000O1000000O100000000O1000000O100000000O2O00000O100000000O1000000O010000O10O1000O1000O010000O0100000O1O001O1O1O001O1O1O1O001O100O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O010O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1OZOaFaM_9]2dFcM[9[2hFeMW9Y2lFgMS9X2oFgMQ9W2RGiMm8U2UGlMj8R2YGnMf8P2]GoMc8o1`GQN_8m1dGSN[8k1hGUNX8h1kGWNV8f1mGZNS8c1PH]NP8`1RHaNo7[1THdNm7Z1UHfNk7W1XHiNh7T1[HlNe7Q1^HnNc7o0`HjNZNFV9]1cHmNe7o0]HQOd7l0_HTOa7i0bHWO^7f0eHYO\\7d0gH\\OY7a0jH_OV7>mHAT7<oHDR79PIGP76RIJo63TIMl60WI0i6MZI3f6J]I5m90000O10O10O10000O10000O10000O2O000O2O0000Qm[4"}, "label": "plane with sas on tail"}]}
{"id": 370400, "image": "COCO_train2014_000000370400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an l - shaped cow print couch with plastic covering . there are two yellow pillows and one red\"."}], "task": "refering", "answer_template": "The \"an l - shaped cow print couch with plastic covering . there are two yellow pillows and one red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0, 0.2826666666666667, 0.45482812499999997, 0.9992916666666667], "iscrowd": 0, "area": 77147.13055000002, "rle": {"size": [480, 640], "counts": "g5Y9f510000O10000O10000O10000O10000O100O10000O10000O01000O10000O10000O10000O10000O100O10000O10000O10000O1000000O10000000000O100000000O100000000O10000000000O100000000O10000000000O100000000O100000000O10000000000O100000000O10000000`KVKjNj4f5ObKWKgNi4Y1XKfNh4[1WKeNi4[1XKdNh4\\1XKdNg4^1YKaNg4_1ZK`Nf4`1ZK`Nf4a1ZK^Nf4b1ZK^Nf4b1[K]Nd4e1\\KZNd4f1\\KZNd4f1]KYNc4g1]KYNc4h1]KWNb4j1_KUNa4k1_KUNa4l1_KSNa4m1_KSNa4m1`KRN_4P2aKoM_4Q2aKoM_4Q2bKnM^4S2aKmM_4S2bKlM^4T2cKkM\\4W2cKiM]4W2dKhM\\4X2dKhM\\4Y2dKfM\\4Z2dKfM[4[2fKdMZ4\\2gKcMY4]2gKcMY4]2hKbMX4]2iKcMW4]2jKbMU4^2mKaMS4_2mKaMS4^2oKaMQ4_2oKaMQ4^2QLaMn3`2SL_Mm3`2TL`Ml3`2UL_Mk3`2VL`Mj3`2WL_Mh3a2ZL^Mf3b2ZL^Mf3a2\\L^Md3b2\\L^Md3b2]L]Mc3b2_L]M`3d2`L\\M`3c2bL\\M^3d2bL\\M^3c2dL\\M\\3d2eL[MZ3e2gL[MY3e2hLZMX3e2iL[MW3e2jLZMV3e2lLZMT3f2lLZMS3f2oLYMQ3g2oLYMQ3f2QMYMo2g2RMXMn2g2SMYMl2h2UMWMk2h2VMXMj2h2WMWMi2h2YMWMg2i2YMWMf2i2\\MVMd2j2\\MVMd2i2^MVMb2j2^MVMb2i2`MVM`2j2aMUM^2l2bMTM^2k2dMTM\\2l2dMTM\\2k2fMTMZ2l2gMSMX2m2iMSMW2m2jMRMV2m2kMSMU2m2lMRMT2m2nMRMR2n2nMRMQ2n2QNQMo1o2QNQMo1n2SNQMm1o2TNPMl1o2UNQMk1o2VNPMk1n2VNRMj1n2WNQMj1m2XNRMi1m2WNSMi1l2YNSMh1l2XNTMi1j2YNUMg1k2ZNTMg1j2ZNVMg1i2ZNVMf1i2[NWMf1h2[NWMe1h2]NWMd1h2\\NXMV2V2kMiMi2b1XM^NW3S1jLlNV3S1lLlNU3S1kLmNU3R1mLmNT3Q1mLoNS3Q1nLnNR3Q1oLoNR3P1oLoNQ3P1QMoNP3P1PMPOP3o0RMPOo2n0RMROn2n0SMQOm2n0UMQOl2n0TMROl2m0VMROk2l0VMTOj2l0WMSOj2k0XMTOh2l0XMTOh2k0ZMTOg2k0YMUOg2j0[MUOf2i0\\MVOd2j0\\MVOe2h0]MWOc2i0]MWOc2h0_MWOb2g0`MXO`2h0`MXOa2f0aMYO_2g0aMYO`2e0bMZO_2d0cM[O^2d0bM\\O^2c0dM\\O]2c0cM]O^2a0dM^O]2a0dM^O]2`0dM_O^2?dM@\\2`0dM@]2>eMA\\2>eMA\\2=eMC\\2;fMD[2;eME\\29fMFZ2:gMEZ2SOmHf0j47Z2mNTIi0c49Z2iNZIj0]4=Q3^OQMa0R3\\OnLd0U3XOmLg0W3UOjLj0Y3ROhLn0[3nNgLQ1]3kNcLU1e8O1O001OVUS5"}, "label": "an l - shaped cow print couch with plastic covering . there are two yellow pillows and one red"}]}
{"id": 370400, "image": "COCO_train2014_000000370400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow patterned couch with two yellow pillows and one red pillow\"."}], "task": "refering", "answer_template": "The \"a cow patterned couch with two yellow pillows and one red pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0, 0.2826666666666667, 0.45482812499999997, 0.9992916666666667], "iscrowd": 0, "area": 77147.13055000002, "rle": {"size": [480, 640], "counts": "g5Y9f510000O10000O10000O10000O10000O100O10000O10000O01000O10000O10000O10000O10000O100O10000O10000O10000O1000000O10000000000O100000000O100000000O10000000000O100000000O10000000000O100000000O100000000O10000000000O100000000O10000000`KVKjNj4f5ObKWKgNi4Y1XKfNh4[1WKeNi4[1XKdNh4\\1XKdNg4^1YKaNg4_1ZK`Nf4`1ZK`Nf4a1ZK^Nf4b1ZK^Nf4b1[K]Nd4e1\\KZNd4f1\\KZNd4f1]KYNc4g1]KYNc4h1]KWNb4j1_KUNa4k1_KUNa4l1_KSNa4m1_KSNa4m1`KRN_4P2aKoM_4Q2aKoM_4Q2bKnM^4S2aKmM_4S2bKlM^4T2cKkM\\4W2cKiM]4W2dKhM\\4X2dKhM\\4Y2dKfM\\4Z2dKfM[4[2fKdMZ4\\2gKcMY4]2gKcMY4]2hKbMX4]2iKcMW4]2jKbMU4^2mKaMS4_2mKaMS4^2oKaMQ4_2oKaMQ4^2QLaMn3`2SL_Mm3`2TL`Ml3`2UL_Mk3`2VL`Mj3`2WL_Mh3a2ZL^Mf3b2ZL^Mf3a2\\L^Md3b2\\L^Md3b2]L]Mc3b2_L]M`3d2`L\\M`3c2bL\\M^3d2bL\\M^3c2dL\\M\\3d2eL[MZ3e2gL[MY3e2hLZMX3e2iL[MW3e2jLZMV3e2lLZMT3f2lLZMS3f2oLYMQ3g2oLYMQ3f2QMYMo2g2RMXMn2g2SMYMl2h2UMWMk2h2VMXMj2h2WMWMi2h2YMWMg2i2YMWMf2i2\\MVMd2j2\\MVMd2i2^MVMb2j2^MVMb2i2`MVM`2j2aMUM^2l2bMTM^2k2dMTM\\2l2dMTM\\2k2fMTMZ2l2gMSMX2m2iMSMW2m2jMRMV2m2kMSMU2m2lMRMT2m2nMRMR2n2nMRMQ2n2QNQMo1o2QNQMo1n2SNQMm1o2TNPMl1o2UNQMk1o2VNPMk1n2VNRMj1n2WNQMj1m2XNRMi1m2WNSMi1l2YNSMh1l2XNTMi1j2YNUMg1k2ZNTMg1j2ZNVMg1i2ZNVMf1i2[NWMf1h2[NWMe1h2]NWMd1h2\\NXMV2V2kMiMi2b1XM^NW3S1jLlNV3S1lLlNU3S1kLmNU3R1mLmNT3Q1mLoNS3Q1nLnNR3Q1oLoNR3P1oLoNQ3P1QMoNP3P1PMPOP3o0RMPOo2n0RMROn2n0SMQOm2n0UMQOl2n0TMROl2m0VMROk2l0VMTOj2l0WMSOj2k0XMTOh2l0XMTOh2k0ZMTOg2k0YMUOg2j0[MUOf2i0\\MVOd2j0\\MVOe2h0]MWOc2i0]MWOc2h0_MWOb2g0`MXO`2h0`MXOa2f0aMYO_2g0aMYO`2e0bMZO_2d0cM[O^2d0bM\\O^2c0dM\\O]2c0cM]O^2a0dM^O]2a0dM^O]2`0dM_O^2?dM@\\2`0dM@]2>eMA\\2>eMA\\2=eMC\\2;fMD[2;eME\\29fMFZ2:gMEZ2SOmHf0j47Z2mNTIi0c49Z2iNZIj0]4=Q3^OQMa0R3\\OnLd0U3XOmLg0W3UOjLj0Y3ROhLn0[3nNgLQ1]3kNcLU1e8O1O001OVUS5"}, "label": "a cow patterned couch with two yellow pillows and one red pillow"}]}
{"id": 386784, "image": "COCO_train2014_000000386784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a men wearing blue shirt with white t - shirt & eat the pizza\"."}], "task": "refering", "answer_template": "The \"a men wearing blue shirt with white t - shirt & eat the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 194, 210, 211, 212, 213, 214], "bbox": [0.14865625, 0.0033750000000000004, 0.547296875, 0.5732083333333333], "iscrowd": 0, "area": 40629.86860000001, "rle": {"size": [480, 640], "counts": "Xh\\1<c>4K6J6J2N2N2O1N3M2N2N2N3M2N3J6E;L4L4L4N2M3N1O0O100O2O000O101N1000001O01O0000010O000001O000000001N100O101N1QEoLP:Q3oETMl9m2SFUMl9l2RFVMl9k2SFWMk9j2TFXMj9i2aEmM]:\\3N2N2O1O1O1O1O1O1N2O1O1O1O1O1O1O10O01O1O1O1O100O1O1O100O1O1O1O100O1O100O00100O100O01000000O01WNiJiIV5n5YKmIS3oN_MQ7ElIk2KhLX6`0kIh2h7[MWHe2g7_MWHa2h7aMWH_2h7cMWH]2h7fMUH[2j7gMUHY2j7jMTHV2j7oMSHQ2l7RNRHn1m7VNoGk1P8ZNlGf1S8d2O1O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000001O00000000000000000000000000000000006JS1mN00O1O1O1O1O1O1N2M3O1O100O100O100O100O100O100O100O1SOm0M3M3M3M3M3N2M3000000000008H<D6J1O2N1O2N1O3M4\\KlEo1KOj:JYE5k:EXE:k:BVE>k:bNlD:<S1j:\\NPE>9U1[;hNfDX1^;bNeD]1`<O1O1O1O1O1O1O5K3M1VOj0I7N2M^iW4"}, "label": "a men wearing blue shirt with white t - shirt & eat the pizza"}]}
{"id": 386784, "image": "COCO_train2014_000000386784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy on the left with the pizza in his mouth\"."}], "task": "refering", "answer_template": "The \"the boy on the left with the pizza in his mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 194, 210, 211, 212, 213, 214], "bbox": [0.14865625, 0.0033750000000000004, 0.547296875, 0.5732083333333333], "iscrowd": 0, "area": 40629.86860000001, "rle": {"size": [480, 640], "counts": "Xh\\1<c>4K6J6J2N2N2O1N3M2N2N2N3M2N3J6E;L4L4L4N2M3N1O0O100O2O000O101N1000001O01O0000010O000001O000000001N100O101N1QEoLP:Q3oETMl9m2SFUMl9l2RFVMl9k2SFWMk9j2TFXMj9i2aEmM]:\\3N2N2O1O1O1O1O1O1N2O1O1O1O1O1O1O10O01O1O1O1O100O1O1O100O1O1O1O100O1O100O00100O100O01000000O01WNiJiIV5n5YKmIS3oN_MQ7ElIk2KhLX6`0kIh2h7[MWHe2g7_MWHa2h7aMWH_2h7cMWH]2h7fMUH[2j7gMUHY2j7jMTHV2j7oMSHQ2l7RNRHn1m7VNoGk1P8ZNlGf1S8d2O1O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000001O00000000000000000000000000000000006JS1mN00O1O1O1O1O1O1N2M3O1O100O100O100O100O100O100O100O1SOm0M3M3M3M3M3N2M3000000000008H<D6J1O2N1O2N1O3M4\\KlEo1KOj:JYE5k:EXE:k:BVE>k:bNlD:<S1j:\\NPE>9U1[;hNfDX1^;bNeD]1`<O1O1O1O1O1O1O5K3M1VOj0I7N2M^iW4"}, "label": "the boy on the left with the pizza in his mouth"}]}
{"id": 386784, "image": "COCO_train2014_000000386784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy in a black shirt with pizza slice in his hand\"."}], "task": "refering", "answer_template": "The \"guy in a black shirt with pizza slice in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 40, 41, 42, 43, 44, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.65225, 0.008979166666666667, 1.0, 0.5640416666666667], "iscrowd": 0, "area": 35204.84519999998, "rle": {"size": [480, 640], "counts": "[TT61n>2O0K6M2O2M2N3N1N3N2M3N3L4M3M3L2O1O1N2O1N1O2O1N2O1N2N2O1N2O1N2N101N2O1N2N2O1N2O1O1O1O010O1O1O1O1O1O1O1O2N1O2N1O2O0O2N2O0gDQM`:Q3\\ERMc:P3YESMf:P3UESMk:n2QEUMn:_3O2mNSLQGm3l7ZMoGg2l7bMoG_2k7kMPHV2l7RNPHn1l7[NnGf1P8`NkGa1T8f2O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O2O0O100O101N10000O101N10000O101N11O0000010O00000010O000001O01O00010O00001O01O0001O00010O00001O0001O01O000010O000001O0001O000010O01O1O001O10O01mJfH^2[7]MiHc2X7YMlHf2T7WMPIi2P7SMSIm2n6nLWIQ3k6jLYIU3i6fLZIZ3i6`L[I^3h6^LZI_3k6\\LYIa3\\7iKhHT4_7dKdH[4R9N2O010O01O01O01O010O0010O00010O00010O00010O00010O0TMlEl0T:TOQFh0o9VOWFe0i9[OXFd0h9[OZFe0f9ZO\\Fd0d9\\O]Fc0c9\\O`Fc0`9\\OcFa0]9^OgF?Y9AiF>W9@lF>T9BmF=S9BPG=P9B]G1c8N_G1b8N^G3b8K`G3b8L_G0e8N\\GOi8OXGMo8NSGOR9NaFAdN5c6"}, "label": "guy in a black shirt with pizza slice in his hand"}]}
{"id": 386784, "image": "COCO_train2014_000000386784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the black shirt eating pizza\"."}], "task": "refering", "answer_template": "The \"the man in the black shirt eating pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 40, 41, 42, 43, 44, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.65225, 0.008979166666666667, 1.0, 0.5640416666666667], "iscrowd": 0, "area": 35204.84519999998, "rle": {"size": [480, 640], "counts": "[TT61n>2O0K6M2O2M2N3N1N3N2M3N3L4M3M3L2O1O1N2O1N1O2O1N2O1N2N2O1N2O1N2N101N2O1N2N2O1N2O1O1O1O010O1O1O1O1O1O1O1O2N1O2N1O2O0O2N2O0gDQM`:Q3\\ERMc:P3YESMf:P3UESMk:n2QEUMn:_3O2mNSLQGm3l7ZMoGg2l7bMoG_2k7kMPHV2l7RNPHn1l7[NnGf1P8`NkGa1T8f2O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O2O0O100O101N10000O101N10000O101N11O0000010O00000010O000001O01O00010O00001O01O0001O00010O00001O0001O01O000010O000001O0001O000010O01O1O001O10O01mJfH^2[7]MiHc2X7YMlHf2T7WMPIi2P7SMSIm2n6nLWIQ3k6jLYIU3i6fLZIZ3i6`L[I^3h6^LZI_3k6\\LYIa3\\7iKhHT4_7dKdH[4R9N2O010O01O01O01O010O0010O00010O00010O00010O00010O0TMlEl0T:TOQFh0o9VOWFe0i9[OXFd0h9[OZFe0f9ZO\\Fd0d9\\O]Fc0c9\\O`Fc0`9\\OcFa0]9^OgF?Y9AiF>W9@lF>T9BmF=S9BPG=P9B]G1c8N_G1b8N^G3b8K`G3b8L_G0e8N\\GOi8OXGMo8NSGOR9NaFAdN5c6"}, "label": "the man in the black shirt eating pizza"}]}
{"id": 517869, "image": "COCO_train2014_000000517869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shirtless man doing pushups on a skateboard\"."}], "task": "refering", "answer_template": "The \"a shirtless man doing pushups on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 259, 260, 261, 262, 263, 277, 278, 279, 286], "bbox": [0.029921874999999997, 0.5762295081967214, 0.47121874999999996, 0.8677049180327868], "iscrowd": 0, "area": 12387.293700000002, "rle": {"size": [427, 640], "counts": "oe81Y=10000O10000O10000O10000O10000O10000QNMeF4Y9OfF1X92gFNV95iFLU97jFIS9;lFER9>lFCS9?lFAT9?lFAT9`0iFBV9`0hFAX9`0eFB[9?bFC^9=`FE`9?[FAf9c0TF^Om9f0mE[OS:j0gEVO[:i0eEWO\\:h0cEYO^:f0aE[O9ZOa9Z1UF\\O<ZO_9Y1UF]O<[O_9W1TF_O>ZO^9V1SFA`0YO\\9V1SFBa0YO\\9T1SFBc0ZOZ9S1RFDd0YO[9R1oEGg0WOZ9Q1mEJj0UOY9P1kEMl0TOY9n1iFRNV9n1jFSNV9l1kFTNU9k1kFVNT9i1nFVNS9i1nFWNQ9i1oFXNQ9g1PGYNo8g1QGZNo8b1UG^Nj8_1ZGaNf8\\1\\GeNc8X1aGhN_8T1dGlN\\8Q1hGoNX8m0kGTOT8j0oGVOQ8g0RHYOm7h0RHYOn7g0RHYOm7g0SHZOm7f0RHZOo7f0oG\\OP8e0mG^OS8a0lGAS8`0kGBU8>hGEW8>eGD[8?`GAa8b0ZG@g8c0TG]Om8g0nFYOT9j0gFWOY9l0bFUO`9n0[FROf9R1UFoNl9T1nEmNS:T22O0O2N101N2N2O1N2N2O1N2N2O1lNZEAh:?^EYOd:f0cESO^:m0gElN[:T1n00O1000O100000000O100000000O10000000000O1001O1O1O2N1O1O1O1O1N2O1O1O1O1O2N1O1O1O1O1O1O1O1O1O10001O000000000000000000001O00000000000000000000010O01O1N1O2N2M2O2M3N1N3N2M3N1N3N2N1N3N2M2O2M2O02O0O1O1O2N1O01M2O1O1N2O1O1N3N1O1N2000001O00000000001O0000000000001O00000000001O00000000001O000000000000000000fCDe;<ZDEf;;YDFg;;WDFj;:TDGl;:RDGn;:PDGP<:hCMX<c00O100O100O100O100O10000O100M4K4L4L4L4LgQ]4"}, "label": "a shirtless man doing pushups on a skateboard"}]}
{"id": 517869, "image": "COCO_train2014_000000517869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on skateboard and road\"."}], "task": "refering", "answer_template": "The \"man on skateboard and road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 259, 260, 261, 262, 263, 277, 278, 279, 286], "bbox": [0.029921874999999997, 0.5762295081967214, 0.47121874999999996, 0.8677049180327868], "iscrowd": 0, "area": 12387.293700000002, "rle": {"size": [427, 640], "counts": "oe81Y=10000O10000O10000O10000O10000O10000QNMeF4Y9OfF1X92gFNV95iFLU97jFIS9;lFER9>lFCS9?lFAT9?lFAT9`0iFBV9`0hFAX9`0eFB[9?bFC^9=`FE`9?[FAf9c0TF^Om9f0mE[OS:j0gEVO[:i0eEWO\\:h0cEYO^:f0aE[O9ZOa9Z1UF\\O<ZO_9Y1UF]O<[O_9W1TF_O>ZO^9V1SFA`0YO\\9V1SFBa0YO\\9T1SFBc0ZOZ9S1RFDd0YO[9R1oEGg0WOZ9Q1mEJj0UOY9P1kEMl0TOY9n1iFRNV9n1jFSNV9l1kFTNU9k1kFVNT9i1nFVNS9i1nFWNQ9i1oFXNQ9g1PGYNo8g1QGZNo8b1UG^Nj8_1ZGaNf8\\1\\GeNc8X1aGhN_8T1dGlN\\8Q1hGoNX8m0kGTOT8j0oGVOQ8g0RHYOm7h0RHYOn7g0RHYOm7g0SHZOm7f0RHZOo7f0oG\\OP8e0mG^OS8a0lGAS8`0kGBU8>hGEW8>eGD[8?`GAa8b0ZG@g8c0TG]Om8g0nFYOT9j0gFWOY9l0bFUO`9n0[FROf9R1UFoNl9T1nEmNS:T22O0O2N101N2N2O1N2N2O1N2N2O1lNZEAh:?^EYOd:f0cESO^:m0gElN[:T1n00O1000O100000000O100000000O10000000000O1001O1O1O2N1O1O1O1O1N2O1O1O1O1O2N1O1O1O1O1O1O1O1O1O10001O000000000000000000001O00000000000000000000010O01O1N1O2N2M2O2M3N1N3N2M3N1N3N2N1N3N2M2O2M2O02O0O1O1O2N1O01M2O1O1N2O1O1N3N1O1N2000001O00000000001O0000000000001O00000000001O00000000001O000000000000000000fCDe;<ZDEf;;YDFg;;WDFj;:TDGl;:RDGn;:PDGP<:hCMX<c00O100O100O100O100O10000O100M4K4L4L4L4LgQ]4"}, "label": "man on skateboard and road"}]}
{"id": 313071, "image": "COCO_train2014_000000313071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a cell phone to her ear\"."}], "task": "refering", "answer_template": "The \"a woman holding a cell phone to her ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 288, 289, 290, 291, 292, 305, 306, 307, 314, 315], "bbox": [0.25675, 0.21845833333333334, 0.7111562499999999, 0.8085625000000001], "iscrowd": 0, "area": 38378.024, "rle": {"size": [480, 640], "counts": "kf]21n>100O1O100O2O0O100O100O2N100O100O2O0O1O100O101N100O1O100O2O0O100O100O2N100O100O2O0O1O100O101N100O100O01O001O1O001O001O001O1O001O001O001O1O001O001O1O001OO1L4M4K4L4M3L4L4M4L3O2M2N2N3M2O2M2N3M2O2M2N3M2N3N1N3M2N3M2L5K4L5L3L4L5K4L5L3L5K4L5K4M4L3L5L3M3L5L3M3L5L3M3L4M4L31O00100O1O00100O1O00100O1O00100O1O00100000O101O000O101O000O101O000O101O000O101O000O101O000O101O002M4M4L3M3L5L3M4L3L5L3M4L3L5L3M1O01N1O2N1O2N2N1O2N2O0O2N1O2N2N1O2N1O2O1N101O1N101O0O2O1O1N3N2N1N3N2N2M3N2N2M3N2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N2N2M3N2N2O1nKXFj2h9oL^FR3a9gLfFY3[9`LkFa3U9WLRGi3T:0O1O00100O001O00100O001O010O1O0000O2N2N=C=C=C=C=C=C<D=C=B`[f2"}, "label": "a woman holding a cell phone to her ear"}]}
{"id": 313071, "image": "COCO_train2014_000000313071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman in a grey shirt holds a phone conversation at the dinner table\"."}], "task": "refering", "answer_template": "The \"a blonde woman in a grey shirt holds a phone conversation at the dinner table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 288, 289, 290, 291, 292, 305, 306, 307, 314, 315], "bbox": [0.25675, 0.21845833333333334, 0.7111562499999999, 0.8085625000000001], "iscrowd": 0, "area": 38378.024, "rle": {"size": [480, 640], "counts": "kf]21n>100O1O100O2O0O100O100O2N100O100O2O0O1O100O101N100O1O100O2O0O100O100O2N100O100O2O0O1O100O101N100O100O01O001O1O001O001O001O1O001O001O001O1O001O001O1O001OO1L4M4K4L4M3L4L4M4L3O2M2N2N3M2O2M2N3M2O2M2N3M2N3N1N3M2N3M2L5K4L5L3L4L5K4L5L3L5K4L5K4M4L3L5L3M3L5L3M3L5L3M3L4M4L31O00100O1O00100O1O00100O1O00100O1O00100000O101O000O101O000O101O000O101O000O101O000O101O000O101O002M4M4L3M3L5L3M4L3L5L3M4L3L5L3M1O01N1O2N1O2N2N1O2N2O0O2N1O2N2N1O2N1O2O1N101O1N101O0O2O1O1N3N2N1N3N2N2M3N2N2M3N2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N2N2M3N2N2O1nKXFj2h9oL^FR3a9gLfFY3[9`LkFa3U9WLRGi3T:0O1O00100O001O00100O001O010O1O0000O2N2N=C=C=C=C=C=C<D=C=B`[f2"}, "label": "a blonde woman in a grey shirt holds a phone conversation at the dinner table"}]}
{"id": 313071, "image": "COCO_train2014_000000313071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl eating next to her mom\"."}], "task": "refering", "answer_template": "The \"a little girl eating next to her mom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 318], "bbox": [0.77190625, 0.3641875, 0.9994375, 0.7934166666666668], "iscrowd": 0, "area": 21837.299650000004, "rle": {"size": [480, 640], "counts": "[mW73o<9\\DNc;9QDMo;:cCM\\<\\101N100O100O2O0O100O1O101N100O100O100O100O100O101N3cM_MWHb2a7oMUHS2c7]NRHf1f7jNPHX1m7TOhGn0U8VOgGk0X8ZOdGf0[8^ObGb0]8B_G?`8C_G=`8D`G;`8G^G:a8G_G9`8H`G8_8J_G7`8J`G6_8L`G4_8M`G4_8MaG3^8O`G2`8N`G2`8N`G2`8O^G2b8N^G1c80\\GNf8T32O1N1O2O1O1O1O1O100O1O1O1O1O100O1O100O1000000O10000O1000000O10000O10000O10000000000000000000000000001N10001O001O00001O001O001O00001O001O00001O001O001O3M4L4L4L6J8H8H4L2N1O1O2hLhDd2Y;WMcEQ2_:iMjG2W8JTL"}, "label": "a little girl eating next to her mom"}]}
{"id": 313071, "image": "COCO_train2014_000000313071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing red earring is eating something sitting besides to a lady\"."}], "task": "refering", "answer_template": "The \"a little girl wearing red earring is eating something sitting besides to a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 318], "bbox": [0.77190625, 0.3641875, 0.9994375, 0.7934166666666668], "iscrowd": 0, "area": 21837.299650000004, "rle": {"size": [480, 640], "counts": "[mW73o<9\\DNc;9QDMo;:cCM\\<\\101N100O100O2O0O100O1O101N100O100O100O100O100O101N3cM_MWHb2a7oMUHS2c7]NRHf1f7jNPHX1m7TOhGn0U8VOgGk0X8ZOdGf0[8^ObGb0]8B_G?`8C_G=`8D`G;`8G^G:a8G_G9`8H`G8_8J_G7`8J`G6_8L`G4_8M`G4_8MaG3^8O`G2`8N`G2`8N`G2`8O^G2b8N^G1c80\\GNf8T32O1N1O2O1O1O1O1O100O1O1O1O1O100O1O100O1000000O10000O1000000O10000O10000O10000000000000000000000000001N10001O001O00001O001O001O00001O001O00001O001O001O3M4L4L4L6J8H8H4L2N1O1O2hLhDd2Y;WMcEQ2_:iMjG2W8JTL"}, "label": "a little girl wearing red earring is eating something sitting besides to a lady"}]}
{"id": 313071, "image": "COCO_train2014_000000313071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the black sweatshirt\"."}], "task": "refering", "answer_template": "The \"the man in the black sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0, 0.22870833333333335, 0.56165625, 0.9865416666666665], "iscrowd": 0, "area": 87490.11815000001, "rle": {"size": [480, 640], "counts": "m=j0`<f1[Ne1[Ne1aN_11O001O1O1O1O1O1O010O1O1O1O1O001O1O1O1O1O1O001O1O1O1O00O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2K5L4L4L4K5L4L4L4K5L4L4L4K5L4L4L4O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000O100000000O10000000000O100000000O100000000O10001O0000TMXL`Jg3`5_L[Ja3e5dLVJ\\3j5iLQJW3n5jLRJV3n5iLSJV3m5kLSJU3m5kLSJU3m5jLTJV3k5kLUJU3k5jLVJV3i5kLWJT3j5lLVJT3j5kLWJU3h5lLXJT3h5kLYJU3f5lLZJS3g5mLYJS3g5lLZJT3e5mL[JS3e5lL\\JT3c5mL]JS3c5mL]JT3b5lL^JT3a5mL_JT3`5lL`JT3_5mLaJS3_5mLaJT3^5mLaJS3^5nLbJR3^5nLbJS3\\5nLdJR3\\5nLdJS3[5nLdJR3[5oLeJQ3[5oLeJR3Z5nLfJR3Y5oLgJQ3Y5oLgJR3W5PMhJP3X5PMhJQ3W5oLiJQ3V5PMjJP3V5PMjJQ3S5RMlJn2S5SMmJn2Q5SMoJm2o4UMQKk2n4VMRKk2l4VMTKj2k4XMTKh2k4YMUKh2i4YMWKg2h4ZMXKg2f4ZMZKf2f4ZMZKf2e4\\MZKe2d4\\M\\Kd2c4]M]Kc2b4^M^Kc2a4]M_Kc2`4_M_Kb2_4_MaKa2^4`MbK`2]4aMcK`2\\4`MdK`2[4aMeK_2Z4bMfK_2X4bMhK^2W4cMiK^2V4bMjK^2V4`MlK`2T4_MmKb2S4\\MnKd2R4ZMPLf2P4YMQLh2n3WMSLi2m3VMTLk2k3SMWLo2f3QM[LP3d3oL]LS3a3kLaLV3^3iLcLX3\\3gLeL[3Y3cLiL^3V3aLkLa3S3^LnLc3P3]LQMd3n2ZLTMh3j2WLWMj3h2ULYMm3\\2ZLfMg3m1dLTN]3_1nLbNT3V1PMlNQ3R1oLoNS3n0mLSOT3k0kLWOW3f0iL[OX3d0gL]OZ3a0fL@\\3=cLE^39bLH`35`LLa33]LOd3M^L4d3G^L:c3A_La0c3ZO_Lg0b3UO`Ll0a3PObLP1`3kNcLU1^3gNdLZ1^3`NfL`1[3\\NhLd1Z3WNhLj1Y3RNjLn1Y3lMiLU2Y3gMiLY2Z3bMhL^2[3]MfLd2\\3XMfLh2]3SMeLm2^3nLcLS3_3iLcLW3`3dLaL]3b3^L`Lc3a3YLaLg3b3TL_Lm3d3nK^LR4d3jK^LV4U7100O10000O10000O01O0O2O1O0O2O1N101O0O2O1N101O1N10001N100O2O000O2O0O101O0O101O0O1010O0001O01O0001O2O0O1O1O2N100O1O2N100O1O2N100O2N2N3M2O2M2N2N3N1N3M2N3M2O3J7H8I7H9F9AT\\S4"}, "label": "the man in the black sweatshirt"}]}
{"id": 313071, "image": "COCO_train2014_000000313071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy wearing black coat and necklace with short hair\"."}], "task": "refering", "answer_template": "The \"boy wearing black coat and necklace with short hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0, 0.22870833333333335, 0.56165625, 0.9865416666666665], "iscrowd": 0, "area": 87490.11815000001, "rle": {"size": [480, 640], "counts": "m=j0`<f1[Ne1[Ne1aN_11O001O1O1O1O1O1O010O1O1O1O1O001O1O1O1O1O1O001O1O1O1O00O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2K5L4L4L4K5L4L4L4K5L4L4L4K5L4L4L4O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000O100000000O10000000000O100000000O100000000O10001O0000TMXL`Jg3`5_L[Ja3e5dLVJ\\3j5iLQJW3n5jLRJV3n5iLSJV3m5kLSJU3m5kLSJU3m5jLTJV3k5kLUJU3k5jLVJV3i5kLWJT3j5lLVJT3j5kLWJU3h5lLXJT3h5kLYJU3f5lLZJS3g5mLYJS3g5lLZJT3e5mL[JS3e5lL\\JT3c5mL]JS3c5mL]JT3b5lL^JT3a5mL_JT3`5lL`JT3_5mLaJS3_5mLaJT3^5mLaJS3^5nLbJR3^5nLbJS3\\5nLdJR3\\5nLdJS3[5nLdJR3[5oLeJQ3[5oLeJR3Z5nLfJR3Y5oLgJQ3Y5oLgJR3W5PMhJP3X5PMhJQ3W5oLiJQ3V5PMjJP3V5PMjJQ3S5RMlJn2S5SMmJn2Q5SMoJm2o4UMQKk2n4VMRKk2l4VMTKj2k4XMTKh2k4YMUKh2i4YMWKg2h4ZMXKg2f4ZMZKf2f4ZMZKf2e4\\MZKe2d4\\M\\Kd2c4]M]Kc2b4^M^Kc2a4]M_Kc2`4_M_Kb2_4_MaKa2^4`MbK`2]4aMcK`2\\4`MdK`2[4aMeK_2Z4bMfK_2X4bMhK^2W4cMiK^2V4bMjK^2V4`MlK`2T4_MmKb2S4\\MnKd2R4ZMPLf2P4YMQLh2n3WMSLi2m3VMTLk2k3SMWLo2f3QM[LP3d3oL]LS3a3kLaLV3^3iLcLX3\\3gLeL[3Y3cLiL^3V3aLkLa3S3^LnLc3P3]LQMd3n2ZLTMh3j2WLWMj3h2ULYMm3\\2ZLfMg3m1dLTN]3_1nLbNT3V1PMlNQ3R1oLoNS3n0mLSOT3k0kLWOW3f0iL[OX3d0gL]OZ3a0fL@\\3=cLE^39bLH`35`LLa33]LOd3M^L4d3G^L:c3A_La0c3ZO_Lg0b3UO`Ll0a3PObLP1`3kNcLU1^3gNdLZ1^3`NfL`1[3\\NhLd1Z3WNhLj1Y3RNjLn1Y3lMiLU2Y3gMiLY2Z3bMhL^2[3]MfLd2\\3XMfLh2]3SMeLm2^3nLcLS3_3iLcLW3`3dLaL]3b3^L`Lc3a3YLaLg3b3TL_Lm3d3nK^LR4d3jK^LV4U7100O10000O10000O01O0O2O1O0O2O1N101O0O2O1N101O1N10001N100O2O000O2O0O101O0O101O0O1010O0001O01O0001O2O0O1O1O2N100O1O2N100O1O2N100O2N2N3M2O2M2N2N3N1N3M2N3M2O3J7H8I7H9F9AT\\S4"}, "label": "boy wearing black coat and necklace with short hair"}]}
{"id": 337648, "image": "COCO_train2014_000000337648.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy on his knees , throwing a ball\"."}], "task": "refering", "answer_template": "The \"the boy on his knees , throwing a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 23, 25, 26, 27, 28, 41, 42, 43, 44, 45, 46, 60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 138, 148, 149, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 183, 184, 185, 186, 188, 189, 190], "bbox": [0.13056, 0.06448571428571429, 0.68898, 0.9139142857142857], "iscrowd": 0, "area": 30405.290549999994, "rle": {"size": [350, 500], "counts": "i_f07g:0O1O100O2N100O1O100O100O1O101N1O100O1O100O1O101N1O100O10000O10000O010O1000O010000O0YFVO\\9j0]F^Ob9P1N2O1N2O1O1N2O000O10000O1000000000000000000001O1O1O1O1eNjFl0e9O1O1O1O1ON2O1N2O1N2O1N2N2`IgNZ3Z1bLlNZ3V1hLkNT3V1nLjNn2X1TMhNh2Z1YMhNb2Z1VM\\NZM<[5Z1ZM\\N[M<V5Z1]M^N[M:T5Y1dM\\NXM<Q5Y1iM[NSM=S5X1mMoNR2R1PNmNo1R1TNlNk1T1XNjNh1V1ZNiNd1W1]NiNc1W1^NiN[1\\1fNdNR1d1oN\\Nk0h1WOXNf0i1QOfMmLb0o3j1SOjMjL=P4i1WOdNf0^1YOeNd0[1]OgN`0Z1_OhN?X1CiN:W1JgN4Z1OeNN[16dNH\\1;cNB]1j0YNTOh1n0XNoNh1T1XNiNh1Y1YNdNh1^1WN`Ni1c1WNZNj1g1WNVNi1m1WNPNj1R2UNlMk1V2VNhMi1[2WNbMj1`2UN^Mk1e2UNXMl1i2UNTMk1o2UNnLl1T3SNjLm1X3TNeLl1^3TN_Lm1c3RN[Ln1g3QNXLP2j3nMVLQ2m3nMQLS2l3PNTLo1m3QNRLo1o3RNPLo1n3RNjKXOfNh2b3]NoLCh0Y2W2UNPMCi0Y2U2VNQMBh0Z2T2WNRM@i0[2S2VNTM_Oh0]2R2VNUM^Og0^2R2UNVM^Oh0^2o1WNXM9Jb1l2WNXMQ4f2PLWMT4g2S2N2M3N2N2N2M3N2O1O1N101O1O1O1N101O1O1O1N101O001O0O2O001O001O0O2O0000000001O01O01O001O001O0]MYNoKg1n3^NQLb16WNP1<gN]18XNo0`0fNY1:XNm0k0_Nm0d0XNm0n0\\Nk0e0YNn0n0[Ni0g0ZNm0P1YNf0j0ZNl0S1XNd0j0[Nk0U1YN`0l0[Nj0Y1WN=n0[Nh0\\1XN9o0]Ng0^1WN6Q1\\Ng0a1VN5P1\\Ni0b1TN5n0[Ne0k1[NMl0ZNf0n1[NKj0ZNi0n1[NKg0ZNl0o1ZNJf0YNn0P2ZNId0ZNMXODi2HHc0YN0AXOb23Fa0ZN3HnN\\2;C:cN=k2WOcN2lNf0c2VObNHWOR1Y2SO`NIZOT1W2QO`NI[OV1W2nNg0R1c4000001N1000O100000000000O10O100000000000O1XOYF8g9D]F<b9@cF`0]9\\OgFd0j901O00000O10001O00000O10000O2O000O1O1N2O2N1O1O1N2O1O1OkSe1"}, "label": "the boy on his knees , throwing a ball"}]}
{"id": 337648, "image": "COCO_train2014_000000337648.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing green t - shirt playing on a ground\"."}], "task": "refering", "answer_template": "The \"a boy wearing green t - shirt playing on a ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 23, 25, 26, 27, 28, 41, 42, 43, 44, 45, 46, 60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 138, 148, 149, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 183, 184, 185, 186, 188, 189, 190], "bbox": [0.13056, 0.06448571428571429, 0.68898, 0.9139142857142857], "iscrowd": 0, "area": 30405.290549999994, "rle": {"size": [350, 500], "counts": "i_f07g:0O1O100O2N100O1O100O100O1O101N1O100O1O100O1O101N1O100O10000O10000O010O1000O010000O0YFVO\\9j0]F^Ob9P1N2O1N2O1O1N2O000O10000O1000000000000000000001O1O1O1O1eNjFl0e9O1O1O1O1ON2O1N2O1N2O1N2N2`IgNZ3Z1bLlNZ3V1hLkNT3V1nLjNn2X1TMhNh2Z1YMhNb2Z1VM\\NZM<[5Z1ZM\\N[M<V5Z1]M^N[M:T5Y1dM\\NXM<Q5Y1iM[NSM=S5X1mMoNR2R1PNmNo1R1TNlNk1T1XNjNh1V1ZNiNd1W1]NiNc1W1^NiN[1\\1fNdNR1d1oN\\Nk0h1WOXNf0i1QOfMmLb0o3j1SOjMjL=P4i1WOdNf0^1YOeNd0[1]OgN`0Z1_OhN?X1CiN:W1JgN4Z1OeNN[16dNH\\1;cNB]1j0YNTOh1n0XNoNh1T1XNiNh1Y1YNdNh1^1WN`Ni1c1WNZNj1g1WNVNi1m1WNPNj1R2UNlMk1V2VNhMi1[2WNbMj1`2UN^Mk1e2UNXMl1i2UNTMk1o2UNnLl1T3SNjLm1X3TNeLl1^3TN_Lm1c3RN[Ln1g3QNXLP2j3nMVLQ2m3nMQLS2l3PNTLo1m3QNRLo1o3RNPLo1n3RNjKXOfNh2b3]NoLCh0Y2W2UNPMCi0Y2U2VNQMBh0Z2T2WNRM@i0[2S2VNTM_Oh0]2R2VNUM^Og0^2R2UNVM^Oh0^2o1WNXM9Jb1l2WNXMQ4f2PLWMT4g2S2N2M3N2N2N2M3N2O1O1N101O1O1O1N101O1O1O1N101O001O0O2O001O001O0O2O0000000001O01O01O001O001O0]MYNoKg1n3^NQLb16WNP1<gN]18XNo0`0fNY1:XNm0k0_Nm0d0XNm0n0\\Nk0e0YNn0n0[Ni0g0ZNm0P1YNf0j0ZNl0S1XNd0j0[Nk0U1YN`0l0[Nj0Y1WN=n0[Nh0\\1XN9o0]Ng0^1WN6Q1\\Ng0a1VN5P1\\Ni0b1TN5n0[Ne0k1[NMl0ZNf0n1[NKj0ZNi0n1[NKg0ZNl0o1ZNJf0YNn0P2ZNId0ZNMXODi2HHc0YN0AXOb23Fa0ZN3HnN\\2;C:cN=k2WOcN2lNf0c2VObNHWOR1Y2SO`NIZOT1W2QO`NI[OV1W2nNg0R1c4000001N1000O100000000000O10O100000000000O1XOYF8g9D]F<b9@cF`0]9\\OgFd0j901O00000O10001O00000O10000O2O000O1O1N2O2N1O1O1N2O1O1OkSe1"}, "label": "a boy wearing green t - shirt playing on a ground"}]}
{"id": 345842, "image": "COCO_train2014_000000345842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a brown , wood chair pulled up to the table\"."}], "task": "refering", "answer_template": "The \"the back of a brown , wood chair pulled up to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 104, 105, 106, 117, 118, 130, 131, 132, 144, 145, 146, 157, 158], "bbox": [0.044071038251366115, 0.41936, 0.23502732240437157, 0.69462], "iscrowd": 0, "area": 4478.770150000001, "rle": {"size": [500, 366], "counts": "bQ82]?:G9G9G9J6K5K5K4K6K5K5K5K5K5K5K4M4K5K5K5L4K00O101oL\\Do1e;QNfDd1Z;\\NQEY1o:gN\\En0e:POgEe0Y:[OnE>Q:CoE=Q:CoE=P:CQF=o9CPF>o9CQF=o9CQF=n9CSF=l9DSF=m9CSF=l9CUF=k9CUF=j9DUF5VNWOe;d0UF5ZNSO`;h0WF4R:LmE5R:LnE3S:MmE3R:MoE2R:NmE3R:NnE1S:OmE1R:OoE0R:0mE1R:OoE0R:OoE1P:OQF0P:OPF2o9NRF1o9NRF2m9NTF1m9NSF3l9MUF2l9MUF3P:GPF9X:_OiEa0]:XOdEg0e<0O010e]X4"}, "label": "the back of a brown , wood chair pulled up to the table"}]}
{"id": 345842, "image": "COCO_train2014_000000345842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one wooden chair near the table\"."}], "task": "refering", "answer_template": "The \"there is one wooden chair near the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 104, 105, 106, 117, 118, 130, 131, 132, 144, 145, 146, 157, 158], "bbox": [0.044071038251366115, 0.41936, 0.23502732240437157, 0.69462], "iscrowd": 0, "area": 4478.770150000001, "rle": {"size": [500, 366], "counts": "bQ82]?:G9G9G9J6K5K5K4K6K5K5K5K5K5K5K4M4K5K5K5L4K00O101oL\\Do1e;QNfDd1Z;\\NQEY1o:gN\\En0e:POgEe0Y:[OnE>Q:CoE=Q:CoE=P:CQF=o9CPF>o9CQF=o9CQF=n9CSF=l9DSF=m9CSF=l9CUF=k9CUF=j9DUF5VNWOe;d0UF5ZNSO`;h0WF4R:LmE5R:LnE3S:MmE3R:MoE2R:NmE3R:NnE1S:OmE1R:OoE0R:0mE1R:OoE0R:OoE1P:OQF0P:OPF2o9NRF1o9NRF2m9NTF1m9NSF3l9MUF2l9MUF3P:GPF9X:_OiEa0]:XOdEg0e<0O010e]X4"}, "label": "there is one wooden chair near the table"}]}
{"id": 26367, "image": "COCO_train2014_000000026367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the air in the middle of a skateboard trick\"."}], "task": "refering", "answer_template": "The \"a man in the air in the middle of a skateboard trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 101, 102, 103, 124, 125, 126, 127, 149, 150, 151, 173, 174], "bbox": [0.33021875, 0.005187793427230047, 0.577453125, 0.5142957746478873], "iscrowd": 0, "area": 13725.832149999998, "rle": {"size": [426, 640], "counts": "Wkg21P=9F:O2O0O1000000O2O00000O1ROYOYEg0f:[OYEe0e:]O[Ed0b:_O_E?_:DaE;]:HcE7[:LeE3Y:0gEOW:3jELT:7lEIS:8mEGC_Oo9k0^FECBm9j0`FCBFl9g0cFBAHk9g0dF@@Lj9e0fF^O_O0i9c0hF]O^O1i9c0iF[O]O5h9`0kF[O]O7f9>mF[O\\O9f9<oFZOZO=e99QGZOZO?c97SGZOYOa0c95TG[OXOb0b93VG[OWOe0a90XG[OVOg0a9NYG[OVOi0_9L[G[OUOk0_9J\\G[OUOm0]9H^G[OTOP1\\9E`G\\OSOP1\\9DaGm0ESNS8P1XHo0BRNV8o0XHP1@SNW8m0YHR1]OSNY8k0ZHT1ZORN\\8j0ZHU1XOSN]8h0[HW1UORN`8g0[Hf1d7ZN\\Hg1c7ZN\\Hf1d7ZN\\Hg1c7YN^Hf1b7\\N\\He1c7^NWHe1i7^NRHe1m7]NnGg1Q8\\NiGg1W8g1O001N2O001O001O1O00O1O100O1O100O3M4M3L4L3N2M2N3N1N3O001N10001O001O001O00001O001O001O001O0kNcF\\N]9b1eF^N\\9^1hF`NY9^1iFbNX9[1jFeNV9X1mFhNT9U1nFkNR9R1QGnNP9o0RGQOn8m0UGROk8k0XGUOi8h0YGXOg8f0[GZOf8b0]G^Oc8`0_G@b8<aGD_89eGF\\85hGKX82kGNU8OnG1S8JQH6o7GTH8n7DUH<k7@ZH`0f7\\O]He0b7WObHj0]7SOfHm0l91O1O100N1O2N2N1O001O001O001O00001OL4J6J6J6K5J6J^b`3"}, "label": "a man in the air in the middle of a skateboard trick"}]}
{"id": 26367, "image": "COCO_train2014_000000026367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"kid with glasses and white t - shirt performing a jump on his skateboard\"."}], "task": "refering", "answer_template": "The \"kid with glasses and white t - shirt performing a jump on his skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 101, 102, 103, 124, 125, 126, 127, 149, 150, 151, 173, 174], "bbox": [0.33021875, 0.005187793427230047, 0.577453125, 0.5142957746478873], "iscrowd": 0, "area": 13725.832149999998, "rle": {"size": [426, 640], "counts": "Wkg21P=9F:O2O0O1000000O2O00000O1ROYOYEg0f:[OYEe0e:]O[Ed0b:_O_E?_:DaE;]:HcE7[:LeE3Y:0gEOW:3jELT:7lEIS:8mEGC_Oo9k0^FECBm9j0`FCBFl9g0cFBAHk9g0dF@@Lj9e0fF^O_O0i9c0hF]O^O1i9c0iF[O]O5h9`0kF[O]O7f9>mF[O\\O9f9<oFZOZO=e99QGZOZO?c97SGZOYOa0c95TG[OXOb0b93VG[OWOe0a90XG[OVOg0a9NYG[OVOi0_9L[G[OUOk0_9J\\G[OUOm0]9H^G[OTOP1\\9E`G\\OSOP1\\9DaGm0ESNS8P1XHo0BRNV8o0XHP1@SNW8m0YHR1]OSNY8k0ZHT1ZORN\\8j0ZHU1XOSN]8h0[HW1UORN`8g0[Hf1d7ZN\\Hg1c7ZN\\Hf1d7ZN\\Hg1c7YN^Hf1b7\\N\\He1c7^NWHe1i7^NRHe1m7]NnGg1Q8\\NiGg1W8g1O001N2O001O001O1O00O1O100O1O100O3M4M3L4L3N2M2N3N1N3O001N10001O001O001O00001O001O001O001O0kNcF\\N]9b1eF^N\\9^1hF`NY9^1iFbNX9[1jFeNV9X1mFhNT9U1nFkNR9R1QGnNP9o0RGQOn8m0UGROk8k0XGUOi8h0YGXOg8f0[GZOf8b0]G^Oc8`0_G@b8<aGD_89eGF\\85hGKX82kGNU8OnG1S8JQH6o7GTH8n7DUH<k7@ZH`0f7\\O]He0b7WObHj0]7SOfHm0l91O1O100N1O2N2N1O001O001O001O00001OL4J6J6J6K5J6J^b`3"}, "label": "kid with glasses and white t - shirt performing a jump on his skateboard"}]}
{"id": 190216, "image": "COCO_train2014_000000190216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed aimal in a blue coat and red hat\"."}], "task": "refering", "answer_template": "The \"a stuffed aimal in a blue coat and red hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0016875000000000002, 0.3550625, 0.46517187499999996, 0.9887708333333334], "iscrowd": 0, "area": 67330.83714999998, "rle": {"size": [480, 640], "counts": "Tf09U>b0UG8Q24_K\\2b4dM[K_2e4`MXKc2i4]MTKf2k4ZMSKh2n4XMoJk2Q5TMmJn2T5RMiJQ3V5PMlI]OoNe3V7mLeIU4[6kKbIW4_6hKaIY4^6hKbIW4_6hKaIX4`6hK`IX4`6gKaIX4_6iK`IX4`6hK`IW4a6hK_IY4a6gK_IX4a6hK_IY4a6gK_IX4b6gK^IZ4b6fK^IY4b6gK^IY4c6gK]IY4c6fK]IZ4d6fK\\IZ4c6gK\\IY4e6fKYI]4g6cKWI^4e0bJd4o0eJb4?gJk4h0dJb4:nJR5?bJf44SKZ57_Jh40YKa5N^Jk4I_Kh5G\\Jl4FdKn5_O[Jc6e5]IXJf6h5ZIVJh6i5XIUJk6k5UISJm6m5RIQJQ7o5P1O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1001O000000000000000000000000001O00000000000000000000000000001O00001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O1O001O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O2hLoH]NS7b1TIVNm6i1[IoMg6YOdHn0l0Ab6_OdHn0Q1\\O]6DcHo0X1UOW6KbHo0^1nNQ62cHn0d1hNk58cHn0j1bNe5>bHP1o1[Na5d0aHo0V2UNZ5k0bHo0[2nMU5Q1bHo0a2hMo4W1bHo0a:mNbES1^:jNfET1\\:hNhEV1Z:eNkEZ1V:bNnE\\1T:`NoE_1R:]NSFb1n9ZNVFd1l9XNXFg1a;O2N2O1N2N10O01O010O010O0010O01O010O010O0010O01O010O0010O010OO2O001N101O001N101O001N101O0O2UOioo4"}, "label": "a stuffed aimal in a blue coat and red hat"}]}
{"id": 190216, "image": "COCO_train2014_000000190216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear with a red hat and blue jacket\"."}], "task": "refering", "answer_template": "The \"teddy bear with a red hat and blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0016875000000000002, 0.3550625, 0.46517187499999996, 0.9887708333333334], "iscrowd": 0, "area": 67330.83714999998, "rle": {"size": [480, 640], "counts": "Tf09U>b0UG8Q24_K\\2b4dM[K_2e4`MXKc2i4]MTKf2k4ZMSKh2n4XMoJk2Q5TMmJn2T5RMiJQ3V5PMlI]OoNe3V7mLeIU4[6kKbIW4_6hKaIY4^6hKbIW4_6hKaIX4`6hK`IX4`6gKaIX4_6iK`IX4`6hK`IW4a6hK_IY4a6gK_IX4a6hK_IY4a6gK_IX4b6gK^IZ4b6fK^IY4b6gK^IY4c6gK]IY4c6fK]IZ4d6fK\\IZ4c6gK\\IY4e6fKYI]4g6cKWI^4e0bJd4o0eJb4?gJk4h0dJb4:nJR5?bJf44SKZ57_Jh40YKa5N^Jk4I_Kh5G\\Jl4FdKn5_O[Jc6e5]IXJf6h5ZIVJh6i5XIUJk6k5UISJm6m5RIQJQ7o5P1O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1001O000000000000000000000000001O00000000000000000000000000001O00001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O1O001O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O2hLoH]NS7b1TIVNm6i1[IoMg6YOdHn0l0Ab6_OdHn0Q1\\O]6DcHo0X1UOW6KbHo0^1nNQ62cHn0d1hNk58cHn0j1bNe5>bHP1o1[Na5d0aHo0V2UNZ5k0bHo0[2nMU5Q1bHo0a2hMo4W1bHo0a:mNbES1^:jNfET1\\:hNhEV1Z:eNkEZ1V:bNnE\\1T:`NoE_1R:]NSFb1n9ZNVFd1l9XNXFg1a;O2N2O1N2N10O01O010O010O0010O01O010O010O0010O01O010O0010O010OO2O001N101O001N101O001N101O0O2UOioo4"}, "label": "teddy bear with a red hat and blue jacket"}]}
{"id": 190216, "image": "COCO_train2014_000000190216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear with purple colour cap\"."}], "task": "refering", "answer_template": "The \"a teddy bear with purple colour cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 205, 206, 228], "bbox": [0.562921875, 0.0449375, 0.97921875, 0.5460625], "iscrowd": 0, "area": 33350.71075, "rle": {"size": [480, 640], "counts": "X\\Y51o>1N1O1O1O100O1O1O1O100O1O2N1O100O1O1O1O100O1O1O100O2N1O1O010O1O100O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O100ROiNRDX1l;POmCQ1Q<WOiCi0U<_OdCb0Z<S1N2N2N1O2N2M3N2N2N2N2N2N2N2N2M3N2N2N2N200O1O010O100O1O100O100O2N100O1O100O101N1O100O100O1O101N100O1O100O1O101N100O1O100O10001O7hFWKk7R5fGTKZ8T5XGRKh8c5000000000000000000000000000000000000000000000000001O000000000000000000000000000000000O100000000000000000000000000000001O000000000000000O10001O1O1O1O1O001O1O1O1O1O1O1O001N2O1O1O1O1O1O001O1O1O1O1O1O1O0013MP1PO1O10O010O1ZMjG`NU8W1UHgNm7V1VHgNl7W1XHeNk7X1XHeNk7X1XHeNj7Y1ZHcNi7Z1ZHcNi7[1YHcNh7Z1]HbNf7[1]HbNe7\\1_H`Nd7\\1`HaNc7\\1eH\\N]7a1mHVNV7g1RIRNo6l1U3L4M2N3L4M2N3L4M2N3Lko5"}, "label": "a teddy bear with purple colour cap"}]}
{"id": 190216, "image": "COCO_train2014_000000190216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear wearing a pink hat\"."}], "task": "refering", "answer_template": "The \"a bear wearing a pink hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 205, 206, 228], "bbox": [0.562921875, 0.0449375, 0.97921875, 0.5460625], "iscrowd": 0, "area": 33350.71075, "rle": {"size": [480, 640], "counts": "X\\Y51o>1N1O1O1O100O1O1O1O100O1O2N1O100O1O1O1O100O1O1O100O2N1O1O010O1O100O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O100ROiNRDX1l;POmCQ1Q<WOiCi0U<_OdCb0Z<S1N2N2N1O2N2M3N2N2N2N2N2N2N2N2M3N2N2N2N200O1O010O100O1O100O100O2N100O1O100O101N1O100O100O1O101N100O1O100O1O101N100O1O100O10001O7hFWKk7R5fGTKZ8T5XGRKh8c5000000000000000000000000000000000000000000000000001O000000000000000000000000000000000O100000000000000000000000000000001O000000000000000O10001O1O1O1O1O001O1O1O1O1O1O1O001N2O1O1O1O1O1O001O1O1O1O1O1O1O0013MP1PO1O10O010O1ZMjG`NU8W1UHgNm7V1VHgNl7W1XHeNk7X1XHeNk7X1XHeNj7Y1ZHcNi7Z1ZHcNi7[1YHcNh7Z1]HbNf7[1]HbNe7\\1_H`Nd7\\1`HaNc7\\1eH\\N]7a1mHVNV7g1RIRNo6l1U3L4M2N3L4M2N3L4M2N3Lko5"}, "label": "a bear wearing a pink hat"}]}
{"id": 83725, "image": "COCO_train2014_000000083725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe to the right looking to its left\"."}], "task": "refering", "answer_template": "The \"giraffe to the right looking to its left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 150, 151, 152, 166, 167, 168, 182, 198, 213, 214, 215, 229, 230, 231, 232, 245, 246, 247, 248, 262, 263, 264, 278, 279, 280, 294, 295, 296, 310, 311, 312], "bbox": [0.3493654266958424, 0.359125, 0.5592122538293217, 0.87309375], "iscrowd": 0, "area": 16417.4949, "rle": {"size": [640, 457], "counts": "^\\T34jc05L4L4K5J7H7I7J6I7I7I8H7R@mMn93eGV2nM^NR:\\O]EFd1g2^NmNX:eNdE1Z1d2\\N\\O[:PNlE:d0P6T9fIVF_8e9aGYFV8R:kGkEm7`:RH^Ei7i:XHTEk7k:UHSEm7m:SHQEP8n:a010O3M2N2NnG[El6c:RIbEl6\\:TIiEj6T:UIQFi6m9VIXFQ7]9oHhFk6l8UHdFn0e0g6j8[H`Fn0W1T6[8oH]Fl0Q2[5d7jIZFj0Z2S5_7SJUFk0]2P5_7VJPFn0a2i4a7YJlEP1d2d4c7bK]H[4e7eK\\HX4f7hKZHW4h7iKXHV4h7jKYHU4g7lKYHS4h7lKYHS4g7mKZHR4h7mKXHR4i7mKXHR4i7nKWHQ4k7mKUHS4l7lKUHS4l7mKTHR4m7mKTHR4m7mKTHQ4n7oKRHP4o7oKRHP4o7PLQHn3Q8QLoGo3Q8QLPHn3Q8RLoGl3S8SLoGk3R8TLoGj3S8VLmGi3T8VLnGh3S8WLnGi3R8WLnGh3S8WLoGg3R8XLoGg3Q8ZLPHe3P8ZLQHe3P8ZLQHe3P8[LQHd3o7[LRHd3o7\\LQHc3P8\\LRHb3o7]LRHc3n7]LSHa3m7_LTH`3m7_LTHa3l7_LUH<dLh1Y;jMTHJbMQ2Z:UNVJ_1k5`N`Jj0k5UOb7O001N101L3M4L3MT[m3"}, "label": "giraffe to the right looking to its left"}]}
{"id": 83725, "image": "COCO_train2014_000000083725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest of the two giraffes standing in its natural habbitat looking off in the distance\"."}], "task": "refering", "answer_template": "The \"the tallest of the two giraffes standing in its natural habbitat looking off in the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 150, 151, 152, 166, 167, 168, 182, 198, 213, 214, 215, 229, 230, 231, 232, 245, 246, 247, 248, 262, 263, 264, 278, 279, 280, 294, 295, 296, 310, 311, 312], "bbox": [0.3493654266958424, 0.359125, 0.5592122538293217, 0.87309375], "iscrowd": 0, "area": 16417.4949, "rle": {"size": [640, 457], "counts": "^\\T34jc05L4L4K5J7H7I7J6I7I7I8H7R@mMn93eGV2nM^NR:\\O]EFd1g2^NmNX:eNdE1Z1d2\\N\\O[:PNlE:d0P6T9fIVF_8e9aGYFV8R:kGkEm7`:RH^Ei7i:XHTEk7k:UHSEm7m:SHQEP8n:a010O3M2N2NnG[El6c:RIbEl6\\:TIiEj6T:UIQFi6m9VIXFQ7]9oHhFk6l8UHdFn0e0g6j8[H`Fn0W1T6[8oH]Fl0Q2[5d7jIZFj0Z2S5_7SJUFk0]2P5_7VJPFn0a2i4a7YJlEP1d2d4c7bK]H[4e7eK\\HX4f7hKZHW4h7iKXHV4h7jKYHU4g7lKYHS4h7lKYHS4g7mKZHR4h7mKXHR4i7mKXHR4i7nKWHQ4k7mKUHS4l7lKUHS4l7mKTHR4m7mKTHR4m7mKTHQ4n7oKRHP4o7oKRHP4o7PLQHn3Q8QLoGo3Q8QLPHn3Q8RLoGl3S8SLoGk3R8TLoGj3S8VLmGi3T8VLnGh3S8WLnGi3R8WLnGh3S8WLoGg3R8XLoGg3Q8ZLPHe3P8ZLQHe3P8ZLQHe3P8[LQHd3o7[LRHd3o7\\LQHc3P8\\LRHb3o7]LRHc3n7]LSHa3m7_LTH`3m7_LTHa3l7_LUH<dLh1Y;jMTHJbMQ2Z:UNVJ_1k5`N`Jj0k5UOb7O001N101L3M4L3MT[m3"}, "label": "the tallest of the two giraffes standing in its natural habbitat looking off in the distance"}]}
{"id": 91917, "image": "COCO_train2014_000000091917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black hat riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"man in black hat riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [293, 314, 315, 316, 336, 337, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 402, 403, 404, 405, 425, 426, 427, 447, 448], "bbox": [0.2612581699346405, 0.6204901960784314, 0.4482026143790849, 0.9583333333333334], "iscrowd": 0, "area": 10835.615900000004, "rle": {"size": [612, 612], "counts": "[RP35kb09H8H7H9H7I4M4M2N2N2N2N2N1O2N1O2N1O2N1O2N1O1WOdMi@^2U?hMe@Y2Y?nMa@T2]?RN]@o1b?k0k@dL^>^3TAPMk>e3N2O1N2O1O2M20000000000000000010O000000000000kBcKc;[1nBd1[1VMf;U1RCc1U1[Mi;P1UCe1n0_Mm;j0XC7ILo0VOP<e0[C2L3f0ZOS<?^C3I3b0_OX<8aC4E5?C[<2cC5D5;G_<LfC7_O69Kb<FhC8]O76Nf<@kC9ZO813j<YOPD:UO;M6n<cNaDm0dN9J:i>]O]A9G>l>XO^AP2b>PN^AQ2a>PN^AP2c>oM^AQ2a>oM_AQ2a>PN_AP2a>oM_AQ2a>oM_AQ2a>oM`AP2a>PN^AQ2a>oM`AP2`>PN`AP2a>PN_Ao1a>QN_Ao1a>QN_Ao1b>PN_AP2`>QN_Ao1a>QN`An1a>QN_Ao1a>QN_Ao1a>RN_An1a>QN_Ak1e>UN\\Ae1j>[NUA`1P?`NQA[1T?dNl@S1^?lNb@k0g?VOY@?R`0@n_O6]`0Id_OMf`03Y_OCQa0=o^OYO\\a0f0b0O010O1O0001O00O2L3N2M3M4M2M3MjaY6"}, "label": "man in black hat riding a motorcycle"}]}
{"id": 91917, "image": "COCO_train2014_000000091917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in tan shirt riding a street bike with a boy sitting on the back\"."}], "task": "refering", "answer_template": "The \"man in tan shirt riding a street bike with a boy sitting on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [293, 314, 315, 316, 336, 337, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 402, 403, 404, 405, 425, 426, 427, 447, 448], "bbox": [0.2612581699346405, 0.6204901960784314, 0.4482026143790849, 0.9583333333333334], "iscrowd": 0, "area": 10835.615900000004, "rle": {"size": [612, 612], "counts": "[RP35kb09H8H7H9H7I4M4M2N2N2N2N2N1O2N1O2N1O2N1O2N1O1WOdMi@^2U?hMe@Y2Y?nMa@T2]?RN]@o1b?k0k@dL^>^3TAPMk>e3N2O1N2O1O2M20000000000000000010O000000000000kBcKc;[1nBd1[1VMf;U1RCc1U1[Mi;P1UCe1n0_Mm;j0XC7ILo0VOP<e0[C2L3f0ZOS<?^C3I3b0_OX<8aC4E5?C[<2cC5D5;G_<LfC7_O69Kb<FhC8]O76Nf<@kC9ZO813j<YOPD:UO;M6n<cNaDm0dN9J:i>]O]A9G>l>XO^AP2b>PN^AQ2a>PN^AP2c>oM^AQ2a>oM_AQ2a>PN_AP2a>oM_AQ2a>oM_AQ2a>oM`AP2a>PN^AQ2a>oM`AP2`>PN`AP2a>PN_Ao1a>QN_Ao1a>QN_Ao1b>PN_AP2`>QN_Ao1a>QN`An1a>QN_Ao1a>QN_Ao1a>RN_An1a>QN_Ak1e>UN\\Ae1j>[NUA`1P?`NQA[1T?dNl@S1^?lNb@k0g?VOY@?R`0@n_O6]`0Id_OMf`03Y_OCQa0=o^OYO\\a0f0b0O010O1O0001O00O2L3N2M3M4M2M3MjaY6"}, "label": "man in tan shirt riding a street bike with a boy sitting on the back"}]}
{"id": 91917, "image": "COCO_train2014_000000091917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle carrying two people\"."}], "task": "refering", "answer_template": "The \"a black motorcycle carrying two people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [362, 383, 384, 400, 401, 402, 403, 405, 406, 407, 408, 422, 423, 424, 425, 426, 427, 428, 429, 430, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475], "bbox": [0.15504901960784315, 0.7483169934640523, 0.6179738562091504, 0.9887581699346405], "iscrowd": 0, "area": 17840.74609999999, "rle": {"size": [612, 612], "counts": "U[i11Rc0100O2N1O1O1O101N1O1O1O101N1O1O1kMAdA`0\\>BbA?\\>CbA>^>D_A=a>D]A=b>F[A;e>GYA:f>GXA:h>HUA9j>ITA8l>JQA7o>Jo@8o>Ko@5Q?Mk@5U?Lh@6W?Me@5[?Lb@6^?L^@6b?LZ@6g?JU@9k?IQ@:o?Fn_O<R`0Fj_O<W`0X1001O001O0010O`Nh_O4Y`0Ji_O5W`0Jk_O5V`0Hm_O7S`0Ho_O7R`0GP@8P`0GR@8o?ET@:l?EV@;j?CX@<h?B[@=f?A\\@>d?A^@>b?A`@>a?_Ob@`0^?_Od@`0]?^Oe@a0[?]Oh@c0X?[Oj@d0V?[Ol@d0U?ZOm@e0S?YOo@f0S?XOl_O0?h0e?WOl_O5<d0Y`0@d_O`0[`0Db_O<^`0F`_O:_`0J^_O6a`0N\\_O2c`0O]_O1b`00^_O0a`02^_ONa`03__OM``04`_OL_`05a_OK_`05a_OK_`05a_OK_`05a_OK_`05a_OK_`05a_OK_`05`_OL``05__OKa`05^_OLb`04]_OMc`03\\_ONd`02\\_ONd`02[_OOe`01Z_O0f`00Y_O1g`0OY_O1g`00W_O1i`0OV_O2j`0NU_O3k`0MU_O3k`0R1100000000000000000000000001O000000000CW_O_Ni`0\\1\\_OdNd`0W1a_OiN_`0S1e_OmN[`0n0j_OROV`0j0n_OVOR`0f0R@ZOn?a0W@_Oi?=[@Ce?9_@Ga?4d@L\\?0h@0X?Ll@4Ra000O100000000000000000000000BKm]O5Rb0Mm]O3Qb01m]OOQb04n]OLQb06n]OJPb09o]OGoa0=o]OCoa0`0P^O@oa0b0P^O^Ona0m0N2O1N3M2N2O1N3N1N2O1N3N1N2O1N3N1N2O2_NPNjAR2T>SNhAn1V>WNeAk1X>\\NcAf1Z>_NaAc1]>bN_A_1_>fN\\A\\1b>jNYAX1d>mNWAU1g>e1N25L0OL4L50O1O1O100`LYAa2h>]MZAb2g>\\MZAd2g>ZM[Af2e>XM]Ag2d>XM\\Ah2e>VM]Ai2d>UM^Aj2c>TM_Ak2b>SM_Am2b>QM`An2a>QM`An2a>PM`AP3a>nLaAQ3`>mLbAR3_>lLcAS3W?01O1O001O1O001O001O001O1O001O001O1O001O001O001O1O001O010O001O1O001O001O001O1O001O001O1O001O001O001O1O0RNT_Od1m`0[NT_Od1l`0\\NV_Ob1k`0\\NX_Ob1h`0^NZ_O`1g`0_NZ_O`1g`0_N[_O_1e`0`N^_O^1c`0aN__O]1a`0cN`_O\\1a`0cNa_O[1Va0N2N1O2N2N2J5L5J6K5K4KXY[4"}, "label": "a black motorcycle carrying two people"}]}
{"id": 91917, "image": "COCO_train2014_000000091917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcycle\"."}], "task": "refering", "answer_template": "The \"motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [362, 383, 384, 400, 401, 402, 403, 405, 406, 407, 408, 422, 423, 424, 425, 426, 427, 428, 429, 430, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475], "bbox": [0.15504901960784315, 0.7483169934640523, 0.6179738562091504, 0.9887581699346405], "iscrowd": 0, "area": 17840.74609999999, "rle": {"size": [612, 612], "counts": "U[i11Rc0100O2N1O1O1O101N1O1O1O101N1O1O1kMAdA`0\\>BbA?\\>CbA>^>D_A=a>D]A=b>F[A;e>GYA:f>GXA:h>HUA9j>ITA8l>JQA7o>Jo@8o>Ko@5Q?Mk@5U?Lh@6W?Me@5[?Lb@6^?L^@6b?LZ@6g?JU@9k?IQ@:o?Fn_O<R`0Fj_O<W`0X1001O001O0010O`Nh_O4Y`0Ji_O5W`0Jk_O5V`0Hm_O7S`0Ho_O7R`0GP@8P`0GR@8o?ET@:l?EV@;j?CX@<h?B[@=f?A\\@>d?A^@>b?A`@>a?_Ob@`0^?_Od@`0]?^Oe@a0[?]Oh@c0X?[Oj@d0V?[Ol@d0U?ZOm@e0S?YOo@f0S?XOl_O0?h0e?WOl_O5<d0Y`0@d_O`0[`0Db_O<^`0F`_O:_`0J^_O6a`0N\\_O2c`0O]_O1b`00^_O0a`02^_ONa`03__OM``04`_OL_`05a_OK_`05a_OK_`05a_OK_`05a_OK_`05a_OK_`05a_OK_`05`_OL``05__OKa`05^_OLb`04]_OMc`03\\_ONd`02\\_ONd`02[_OOe`01Z_O0f`00Y_O1g`0OY_O1g`00W_O1i`0OV_O2j`0NU_O3k`0MU_O3k`0R1100000000000000000000000001O000000000CW_O_Ni`0\\1\\_OdNd`0W1a_OiN_`0S1e_OmN[`0n0j_OROV`0j0n_OVOR`0f0R@ZOn?a0W@_Oi?=[@Ce?9_@Ga?4d@L\\?0h@0X?Ll@4Ra000O100000000000000000000000BKm]O5Rb0Mm]O3Qb01m]OOQb04n]OLQb06n]OJPb09o]OGoa0=o]OCoa0`0P^O@oa0b0P^O^Ona0m0N2O1N3M2N2O1N3N1N2O1N3N1N2O1N3N1N2O2_NPNjAR2T>SNhAn1V>WNeAk1X>\\NcAf1Z>_NaAc1]>bN_A_1_>fN\\A\\1b>jNYAX1d>mNWAU1g>e1N25L0OL4L50O1O1O100`LYAa2h>]MZAb2g>\\MZAd2g>ZM[Af2e>XM]Ag2d>XM\\Ah2e>VM]Ai2d>UM^Aj2c>TM_Ak2b>SM_Am2b>QM`An2a>QM`An2a>PM`AP3a>nLaAQ3`>mLbAR3_>lLcAS3W?01O1O001O1O001O001O001O1O001O001O1O001O001O001O1O001O010O001O1O001O001O001O1O001O001O1O001O001O001O1O0RNT_Od1m`0[NT_Od1l`0\\NV_Ob1k`0\\NX_Ob1h`0^NZ_O`1g`0_NZ_O`1g`0_N[_O_1e`0`N^_O^1c`0aN__O]1a`0cN`_O\\1a`0cNa_O[1Va0N2N1O2N2N2J5L5J6K5K4KXY[4"}, "label": "motorcycle"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small evergreen tree in a solid gray container\"."}], "task": "refering", "answer_template": "The \"a small evergreen tree in a solid gray container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 230, 231, 232, 233, 247, 248, 249, 250, 264, 265, 266, 267, 282, 283, 300], "bbox": [0.529875, 0.516, 0.7569374999999999, 0.76196875], "iscrowd": 0, "area": 6869.7891, "rle": {"size": [640, 480], "counts": "hho42nc00O1O1O2nMLU@5k?0P@0P`04k_OMV`06e_OLZ`0a0X_O@h`0a0W_O_Oi`0a0V_O@j`0a0U_O@j`0a0U_O_Of`0f0Z_O[Oa`0j0^_OWO``0k0^_OVOc`0j0\\_OWOd`0h0\\_OYOd`0g0[_OYO]OAQa0V1`_O[O^OAQa0T1`_O\\O]OCRa0P1__O_O_OCQa0n0X_O3j`0KT_O8j`0IT_O9k`0GU_O;j`0EU_O<j`0ET_O>j`0BV_O?i`0BV_O`0m`0\\OQ_Og0Pa0U1O1O1N2O1N101O1N2OO010O01O001O001O1O001O001O001OmLc_Oc25PMP`0<k_Of2OSMU`05n_OQ3P`0nLQ@T3m?jLU@W3j?dL\\@=D_2``0^Mb_Oc2^`0ZMc_Og2]`0WMe_Oj2[`0SMg_On2Y`0oLi_OQ1Hk0j`0UOV_Om0h`0ROY_OP1e`0PO\\_OP1b`0PO__OR1``0mN`_OR1c`0kN__Ok0l`0SOT_Od0Wa0ZOj^O<aa0C^^OKJDRb0`0T^OMUb03k]OLVb0Hd]OO5:Wb0Gf]OL4<Wb0Hh]OI1?Wb0Hj]OGOa0Wb0HT^O7ma0IS^O7ma0IS^O6na0JR^O6na0JS^O4na0LR^O4ma0MT^O2la0NT^O1ma0OS^O1ob0O1O1O1O1O1O1N2OO100O100O100OO2M2NPgX2"}, "label": "a small evergreen tree in a solid gray container"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant in a big grey pot behind two people\"."}], "task": "refering", "answer_template": "The \"a plant in a big grey pot behind two people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 230, 231, 232, 233, 247, 248, 249, 250, 264, 265, 266, 267, 282, 283, 300], "bbox": [0.529875, 0.516, 0.7569374999999999, 0.76196875], "iscrowd": 0, "area": 6869.7891, "rle": {"size": [640, 480], "counts": "hho42nc00O1O1O2nMLU@5k?0P@0P`04k_OMV`06e_OLZ`0a0X_O@h`0a0W_O_Oi`0a0V_O@j`0a0U_O@j`0a0U_O_Of`0f0Z_O[Oa`0j0^_OWO``0k0^_OVOc`0j0\\_OWOd`0h0\\_OYOd`0g0[_OYO]OAQa0V1`_O[O^OAQa0T1`_O\\O]OCRa0P1__O_O_OCQa0n0X_O3j`0KT_O8j`0IT_O9k`0GU_O;j`0EU_O<j`0ET_O>j`0BV_O?i`0BV_O`0m`0\\OQ_Og0Pa0U1O1O1N2O1N101O1N2OO010O01O001O001O1O001O001O001OmLc_Oc25PMP`0<k_Of2OSMU`05n_OQ3P`0nLQ@T3m?jLU@W3j?dL\\@=D_2``0^Mb_Oc2^`0ZMc_Og2]`0WMe_Oj2[`0SMg_On2Y`0oLi_OQ1Hk0j`0UOV_Om0h`0ROY_OP1e`0PO\\_OP1b`0PO__OR1``0mN`_OR1c`0kN__Ok0l`0SOT_Od0Wa0ZOj^O<aa0C^^OKJDRb0`0T^OMUb03k]OLVb0Hd]OO5:Wb0Gf]OL4<Wb0Hh]OI1?Wb0Hj]OGOa0Wb0HT^O7ma0IS^O7ma0IS^O6na0JR^O6na0JS^O4na0LR^O4ma0MT^O2la0NT^O1ma0OS^O1ob0O1O1O1O1O1O1N2OO100O100O100OO2M2NPgX2"}, "label": "a plant in a big grey pot behind two people"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl\"."}], "task": "refering", "answer_template": "The \"the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 263, 264, 280, 281, 282, 297, 298, 299, 300, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 364, 365, 366, 367, 368, 381, 382, 383, 384, 385], "bbox": [0.44804166666666667, 0.640234375, 0.7241666666666667, 0.9952500000000001], "iscrowd": 0, "area": 21336.033450000003, "rle": {"size": [640, 480], "counts": "cnV4>Pc0l0\\Oe0@:G4L4L4L4L4N2bL^M]Ec2b:bMYE_2e:gMVEZ2h:lMSEU2l:oMoDS2o:PNoDQ2o:SNnDn1Q;TNmDm1Q;VNmDk1R;WNlDj1T;WNiDk1V;WN_DS2a;nMPD`2P<bM`Cl2_<VM\\Cn2d<SMZCn2e<TMZCm2e<TMYCm2f<TMZCl2f<TMYCm2g<TMXCl2g<UMXCl2h<UMWCk2h<WMVCj2j<V2000000001O00000000001O00000000001O000000001O1O2N2N2N2N2N3M2N2N2N2N2N2N2N2O1N2N1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1nJYAi4P?N101O1N101O0O2O0N3M2N3M2O4K5K5K6K4K6J5K5K6K4K6J5I7I8H7I8H7I7I8H7I8H7Iena2"}, "label": "the girl"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray and blue jacket worn by a young woman\"."}], "task": "refering", "answer_template": "The \"a gray and blue jacket worn by a young woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 263, 264, 280, 281, 282, 297, 298, 299, 300, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 364, 365, 366, 367, 368, 381, 382, 383, 384, 385], "bbox": [0.44804166666666667, 0.640234375, 0.7241666666666667, 0.9952500000000001], "iscrowd": 0, "area": 21336.033450000003, "rle": {"size": [640, 480], "counts": "cnV4>Pc0l0\\Oe0@:G4L4L4L4L4N2bL^M]Ec2b:bMYE_2e:gMVEZ2h:lMSEU2l:oMoDS2o:PNoDQ2o:SNnDn1Q;TNmDm1Q;VNmDk1R;WNlDj1T;WNiDk1V;WN_DS2a;nMPD`2P<bM`Cl2_<VM\\Cn2d<SMZCn2e<TMZCm2e<TMYCm2f<TMZCl2f<TMYCm2g<TMXCl2g<UMXCl2h<UMWCk2h<WMVCj2j<V2000000001O00000000001O00000000001O000000001O1O2N2N2N2N2N3M2N2N2N2N2N2N2N2O1N2N1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1nJYAi4P?N101O1N101O0O2O0N3M2N3M2O4K5K5K6K4K6J5K5K6K4K6J5I7I8H7I8H7I7I8H7I8H7Iena2"}, "label": "a gray and blue jacket worn by a young woman"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest plant\"."}], "task": "refering", "answer_template": "The \"tallest plant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 172, 173, 189, 190, 191, 205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 256, 257, 258, 259, 273, 274, 275, 276, 290, 291, 292, 293, 307, 308, 309, 325, 326, 342, 343, 359, 360], "bbox": [0.06974999999999999, 0.40721875, 0.30512500000000004, 0.9209687500000001], "iscrowd": 0, "area": 20703.613149999994, "rle": {"size": [640, 480], "counts": "nod02jc0h0WO?B9H8I7J=C7H4M4L4L4K5L4K4n@_L_=f3]BZLb=j3[BXLc=k3ZBXLb=l3[BXL`=l3]BWL_=m3_BUL]=o3`BTL[=Q4bBRLZ=Q4dBRLW=S4fBPLV=W4dBlKW=]4bBfKZ=c5SD^IY80`GW7NPIb8J^GZ7GRIk8F[G[7GPIn8GXG[7InHo8HWGZ7IoHP9ITGZ7KmHQ9KQGY7MmHR9KoFZ7NkHS9MmFX7OlHT9MlFX7OlHT9NkFV70mHV9MiFW70lHW9NhFV70mHX9NgFV7NnH[9MfFU7NoH\\9NdFT7MPI_9McFS7MQI`9MbFR7MRIa9LbFS7JSId9KaFR7JTIe9K`FR7ITIg9L^FP7IVIi9K]FT8c9mG[FS8e9nGZFR8f9PHXFP8h9QHWFo7i9T1000000001O2N1N3N2M3M3M3M3M3I7C=D;hNSEQIY;i6S1K5K5J6K5K5J6H7J7H9G;D;F:J7J4L4L3aMh_OQ1\\`0kNg_OR1\\`0lNe_OR1_`0jNc_OT1``0jNa_OT1c`0hN`_OU1c`0iN]_OV1g`0fN[_O5ka0IU^O6na0GT^O7oa0GR^O7Qb0Go]O8Tb0En]O:Tb0Dl]O;Wb0Bk]O<Xb0Bi]O<nb0O2N1O2O1N2N2Nf[`6"}, "label": "tallest plant"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall potted evergreen tree to the right of the man in the striped shirt\"."}], "task": "refering", "answer_template": "The \"a tall potted evergreen tree to the right of the man in the striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 172, 173, 189, 190, 191, 205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 256, 257, 258, 259, 273, 274, 275, 276, 290, 291, 292, 293, 307, 308, 309, 325, 326, 342, 343, 359, 360], "bbox": [0.06974999999999999, 0.40721875, 0.30512500000000004, 0.9209687500000001], "iscrowd": 0, "area": 20703.613149999994, "rle": {"size": [640, 480], "counts": "nod02jc0h0WO?B9H8I7J=C7H4M4L4L4K5L4K4n@_L_=f3]BZLb=j3[BXLc=k3ZBXLb=l3[BXL`=l3]BWL_=m3_BUL]=o3`BTL[=Q4bBRLZ=Q4dBRLW=S4fBPLV=W4dBlKW=]4bBfKZ=c5SD^IY80`GW7NPIb8J^GZ7GRIk8F[G[7GPIn8GXG[7InHo8HWGZ7IoHP9ITGZ7KmHQ9KQGY7MmHR9KoFZ7NkHS9MmFX7OlHT9MlFX7OlHT9NkFV70mHV9MiFW70lHW9NhFV70mHX9NgFV7NnH[9MfFU7NoH\\9NdFT7MPI_9McFS7MQI`9MbFR7MRIa9LbFS7JSId9KaFR7JTIe9K`FR7ITIg9L^FP7IVIi9K]FT8c9mG[FS8e9nGZFR8f9PHXFP8h9QHWFo7i9T1000000001O2N1N3N2M3M3M3M3M3I7C=D;hNSEQIY;i6S1K5K5J6K5K5J6H7J7H9G;D;F:J7J4L4L3aMh_OQ1\\`0kNg_OR1\\`0lNe_OR1_`0jNc_OT1``0jNa_OT1c`0hN`_OU1c`0iN]_OV1g`0fN[_O5ka0IU^O6na0GT^O7oa0GR^O7Qb0Go]O8Tb0En]O:Tb0Dl]O;Wb0Bk]O<Xb0Bi]O<nb0O2N1O2O1N2N2Nf[`6"}, "label": "a tall potted evergreen tree to the right of the man in the striped shirt"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"annoyed man in mime shirt\"."}], "task": "refering", "answer_template": "The \"annoyed man in mime shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 243, 244, 245, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 344, 345, 346, 347, 361, 362, 363, 364, 378, 379, 380, 381], "bbox": [0.22172916666666667, 0.561796875, 0.48239583333333336, 0.9887656249999999], "iscrowd": 0, "area": 22708.807199999996, "rle": {"size": [640, 480], "counts": "UhR21dc0?A>B?A>B?D<L3L3N3VAWMV<l2[CaMe<a2lBlMS=W2_BUNa=m1QB_Nn=e1kA_NU>c1fA`NY>d300O100O100O100O100O100O100O100O100O010O100O100O100O100O1jNSJYDm5e;YJVDh5h;^JmCi5Q<]JeCh5Z<n001I6J7O0100O100O10O0101N2003M3M4K4M3M3M3M3M2N000000000001O1O1O1O1O1O1O1O1O2N2N2N2N3M2^IZDf4h;QK`Dn4b;kJcDU5_;eJfDZ5\\;aJgD_5[;\\JiDX4nNbL\\<POjDZ4QObLf=[3bB`L_=Z3kBaLU=Z3TCbLm<X3]CcLc<X3gCcLZ<V3PDfLQ<T3`2K6M2N3N2M2O2N1N3N1N3N3M6I6K5K5J6K6I6K5Kg`j4"}, "label": "annoyed man in mime shirt"}]}
{"id": 419599, "image": "COCO_train2014_000000419599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 243, 244, 245, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 344, 345, 346, 347, 361, 362, 363, 364, 378, 379, 380, 381], "bbox": [0.22172916666666667, 0.561796875, 0.48239583333333336, 0.9887656249999999], "iscrowd": 0, "area": 22708.807199999996, "rle": {"size": [640, 480], "counts": "UhR21dc0?A>B?A>B?D<L3L3N3VAWMV<l2[CaMe<a2lBlMS=W2_BUNa=m1QB_Nn=e1kA_NU>c1fA`NY>d300O100O100O100O100O100O100O100O100O010O100O100O100O100O1jNSJYDm5e;YJVDh5h;^JmCi5Q<]JeCh5Z<n001I6J7O0100O100O10O0101N2003M3M4K4M3M3M3M3M2N000000000001O1O1O1O1O1O1O1O1O2N2N2N2N3M2^IZDf4h;QK`Dn4b;kJcDU5_;eJfDZ5\\;aJgD_5[;\\JiDX4nNbL\\<POjDZ4QObLf=[3bB`L_=Z3kBaLU=Z3TCbLm<X3]CcLc<X3gCcLZ<V3PDfLQ<T3`2K6M2N3N2M2O2N1N3N1N3N3M6I6K5K5J6K6I6K5Kg`j4"}, "label": "man"}]}
{"id": 50961, "image": "COCO_train2014_000000050961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the industrial kitchen stove\"."}], "task": "refering", "answer_template": "The \"the industrial kitchen stove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.54715625, 0.5449791666666666, 1.0, 1.0], "iscrowd": 0, "area": 48787.16649999999, "rle": {"size": [480, 640], "counts": "PjT51n>101O0O101O001N10001N101O2M2O1O2M2O1O2M2O1O3L3N00001N10000O2O0000001O00001O0000001O00010OS5mJ0000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O001O001O001O001O001O001O001O001O001O001OiNSK[Hm4`7iKPHV4l7RLPHn3o7VLoGi3Q8ZLlGf3T8]LjGb3V8d1O10000000000O10000000000000000000000001O00000000001O0000000000001O0000000000001O000000001O00001O00001O00001O001O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O00001O00001O000M4J5J6Jc0^O]F"}, "label": "the industrial kitchen stove"}]}
{"id": 50961, "image": "COCO_train2014_000000050961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a stove with four burners\"."}], "task": "refering", "answer_template": "The \"this is a stove with four burners\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.54715625, 0.5449791666666666, 1.0, 1.0], "iscrowd": 0, "area": 48787.16649999999, "rle": {"size": [480, 640], "counts": "PjT51n>101O0O101O001N10001N101O2M2O1O2M2O1O2M2O1O3L3N00001N10000O2O0000001O00001O0000001O00010OS5mJ0000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O001O001O001O001O001O001O001O001O001O001OiNSK[Hm4`7iKPHV4l7RLPHn3o7VLoGi3Q8ZLlGf3T8]LjGb3V8d1O10000000000O10000000000000000000000001O00000000001O0000000000001O0000000000001O000000001O00001O00001O00001O001O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O00001O00001O000M4J5J6Jc0^O]F"}, "label": "this is a stove with four burners"}]}
{"id": 50961, "image": "COCO_train2014_000000050961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"oven with yellow caution\"."}], "task": "refering", "answer_template": "The \"oven with yellow caution\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.26818749999999997, 0.5799166666666667, 0.6205624999999999, 0.99575], "iscrowd": 0, "area": 28812.603100000004, "rle": {"size": [480, 640], "counts": "em`27b>8O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O1N101N101N101N101N101N101N1012M3N2N2M3N1N3N1N10O010O0010O010O010O01O010O10O01O010O001O001O001O001O>B?A`0@?A`0@?A`0@?A5K00O1000000O1000000000000000000000000000O1000000000000O100O100O10000O100O100O100O100O100O100O100J6J6J6I7N20000001O0000001O0000001O00001O0000001O0000001O00010O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O00001O2N2N3M2N2N00001O00001O001O00001O00001O00001YNhFdMX9j1SH]Mn7P2j2]Oc0^Ofca3"}, "label": "oven with yellow caution"}]}
{"id": 50961, "image": "COCO_train2014_000000050961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall oven with a pan and copper pot on it\"."}], "task": "refering", "answer_template": "The \"a tall oven with a pan and copper pot on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.26818749999999997, 0.5799166666666667, 0.6205624999999999, 0.99575], "iscrowd": 0, "area": 28812.603100000004, "rle": {"size": [480, 640], "counts": "em`27b>8O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O1N101N101N101N101N101N101N1012M3N2N2M3N1N3N1N10O010O0010O010O010O01O010O10O01O010O001O001O001O001O>B?A`0@?A`0@?A`0@?A5K00O1000000O1000000000000000000000000000O1000000000000O100O100O10000O100O100O100O100O100O100O100J6J6J6I7N20000001O0000001O0000001O00001O0000001O0000001O00010O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O00001O2N2N3M2N2N00001O00001O001O00001O00001O00001YNhFdMX9j1SH]Mn7P2j2]Oc0^Ofca3"}, "label": "a tall oven with a pan and copper pot on it"}]}
{"id": 50961, "image": "COCO_train2014_000000050961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"oven fixed on the top of the gas stove\"."}], "task": "refering", "answer_template": "The \"oven fixed on the top of the gas stove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173], "bbox": [0.28884375, 0.07995833333333334, 0.6891875000000001, 0.42906249999999996], "iscrowd": 0, "area": 31033.517299999992, "rle": {"size": [480, 640], "counts": "dif25k>9G8H6I2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O4L<D;E000000000000000O101O000000000000000000000O101O000000000000000000000O101O00000O100O100O100O10O01O100O100O100O00100O100O100O100O00100O100O100O100O00100O100O100O100O00100O100O100O10O01O100O100O100O10O01O100O100O11O000O10000000000000001N100000000000000O101O0000000000000O10001O00000000000O1000001O000000000O10000000001O000O100000000000001O0O10000000000000001N10000000X2hMZUm2"}, "label": "oven fixed on the top of the gas stove"}]}
{"id": 534292, "image": "COCO_train2014_000000534292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in hat\"."}], "task": "refering", "answer_template": "The \"woman in hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 127, 128, 129, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 214, 226, 227, 228, 229, 243, 244, 245, 246, 247, 260, 261, 263, 264, 278, 295], "bbox": [0.3093125, 0.28715625, 0.6726666666666666, 0.78490625], "iscrowd": 0, "area": 26490.88625, "rle": {"size": [640, 480], "counts": "dkl21nc01N3M3N2M2N3N2M3M2N3TOl0Q_OaNi>R2QAUNf>S2TATNc>S2XAUN_>R2ZAVN]>R2]AUNZ>S2`ATNW>S2dATNU>Q4M2M3M3M3N3PCTJe;o5mC]JR<i600O1O2O0OmMmC[LR<c3[DSLd;k3hDlKW;R4UEfKi:Y4cE]K\\:a4PFVKo9k4[FkJd9U5d20O1O001O1O10O10O10000O10000O10000O10000O2O00000O10000O1001O000000001O01O00001SO^J[Ca5b<WKiBi4T=T1L3N2M4L3M4O2M4M3M3LdNoCmJo;S5QDoJl;R5TDmJm;R5UDmJk;S5UDlJk;U5UDjJl;V5TDjJl;V5UDhJk;X5VDhJj;X5VDgJj;Z5VDeJk;Q4\\DeKJ9j;P4aDcKF<j;P4dDaKB>k;P4gD_K_O`0j;P4lD\\KZOc0k;P4oDYKWOg0j;o3eEPL\\:o3fEPLZ:o3gEQLY:n3hEQLY:n3m2N101O1O1O100001O1O1O1O1O1O010O10O001O001O010O001O001O001O001O001O1O2N1O2N2\\OZ@jLg?T3`@gLa?V3g@cL[?Z3g0M4M3M2N3L4M2N4L=B>C<D3L4M3M2N3L4M3M3K5J6K5J_kQ3"}, "label": "woman in hat"}]}
{"id": 534292, "image": "COCO_train2014_000000534292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a floral print dress bathing a child\"."}], "task": "refering", "answer_template": "The \"a woman in a floral print dress bathing a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 127, 128, 129, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 214, 226, 227, 228, 229, 243, 244, 245, 246, 247, 260, 261, 263, 264, 278, 295], "bbox": [0.3093125, 0.28715625, 0.6726666666666666, 0.78490625], "iscrowd": 0, "area": 26490.88625, "rle": {"size": [640, 480], "counts": "dkl21nc01N3M3N2M2N3N2M3M2N3TOl0Q_OaNi>R2QAUNf>S2TATNc>S2XAUN_>R2ZAVN]>R2]AUNZ>S2`ATNW>S2dATNU>Q4M2M3M3M3N3PCTJe;o5mC]JR<i600O1O2O0OmMmC[LR<c3[DSLd;k3hDlKW;R4UEfKi:Y4cE]K\\:a4PFVKo9k4[FkJd9U5d20O1O001O1O10O10O10000O10000O10000O10000O2O00000O10000O1001O000000001O01O00001SO^J[Ca5b<WKiBi4T=T1L3N2M4L3M4O2M4M3M3LdNoCmJo;S5QDoJl;R5TDmJm;R5UDmJk;S5UDlJk;U5UDjJl;V5TDjJl;V5UDhJk;X5VDhJj;X5VDgJj;Z5VDeJk;Q4\\DeKJ9j;P4aDcKF<j;P4dDaKB>k;P4gD_K_O`0j;P4lD\\KZOc0k;P4oDYKWOg0j;o3eEPL\\:o3fEPLZ:o3gEQLY:n3hEQLY:n3m2N101O1O1O100001O1O1O1O1O1O010O10O001O001O010O001O001O001O001O001O1O2N1O2N2\\OZ@jLg?T3`@gLa?V3g@cL[?Z3g0M4M3M2N3L4M2N4L=B>C<D3L4M3M2N3L4M3M3K5J6K5J_kQ3"}, "label": "a woman in a floral print dress bathing a child"}]}
{"id": 534292, "image": "COCO_train2014_000000534292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy on the far right\"."}], "task": "refering", "answer_template": "The \"the boy on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 283, 284, 285, 286, 300, 301, 302, 318, 319], "bbox": [0.6517499999999999, 0.5210625, 0.8447916666666667, 0.79228125], "iscrowd": 0, "area": 11519.857799999998, "rle": {"size": [640, 480], "counts": "WST63ic05L4L5K4K5L3L5K5K4L5K5K5ZOe0K5L4H9G8N2N3N1O100O2N100QO_Lk@90X3T?bLh@93V3S?VMl@j2S?WMm@i2S?XMl@i2R?XMn@h2R?YMm@g2R?ZMn@f2P?\\MPAe2m>^MRAb2l>`MTA`2j>bMVA^2h>dMYA[2e>hMYAY2e>jMYAW2g>a1O1O100O100O1O100O100001O000000000000000000001N10001O00001O001N101O001O0YLVAW2k>gMWAY2j>eMWA[2i>dMXA\\2i>bMXA]2i>aM[A]2f>aMcA^1W?`NSAP1U?nNTA;\\?Cn@Af?<Q2N2N2N3M?@dl]1"}, "label": "the boy on the far right"}]}
{"id": 534292, "image": "COCO_train2014_000000534292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little boy who is wearing suspenders is patiently waiting for his bath\"."}], "task": "refering", "answer_template": "The \"the little boy who is wearing suspenders is patiently waiting for his bath\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 283, 284, 285, 286, 300, 301, 302, 318, 319], "bbox": [0.6517499999999999, 0.5210625, 0.8447916666666667, 0.79228125], "iscrowd": 0, "area": 11519.857799999998, "rle": {"size": [640, 480], "counts": "WST63ic05L4L5K4K5L3L5K5K4L5K5K5ZOe0K5L4H9G8N2N3N1O100O2N100QO_Lk@90X3T?bLh@93V3S?VMl@j2S?WMm@i2S?XMl@i2R?XMn@h2R?YMm@g2R?ZMn@f2P?\\MPAe2m>^MRAb2l>`MTA`2j>bMVA^2h>dMYA[2e>hMYAY2e>jMYAW2g>a1O1O100O100O1O100O100001O000000000000000000001N10001O00001O001N101O001O0YLVAW2k>gMWAY2j>eMWA[2i>dMXA\\2i>bMXA]2i>aM[A]2f>aMcA^1W?`NSAP1U?nNTA;\\?Cn@Af?<Q2N2N2N3M?@dl]1"}, "label": "the little boy who is wearing suspenders is patiently waiting for his bath"}]}
{"id": 534292, "image": "COCO_train2014_000000534292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child wrapped in a towel to the left of a wash tub\"."}], "task": "refering", "answer_template": "The \"a child wrapped in a towel to the left of a wash tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 207, 208, 209, 223, 224, 225, 226, 240, 241, 242, 243, 257, 258, 259, 260, 274, 275, 276, 277, 290, 291, 292, 293, 294, 308, 309, 310, 311, 326, 327, 328, 342, 343, 360], "bbox": [0.10908333333333334, 0.47701562500000005, 0.35520833333333335, 0.9213906250000001], "iscrowd": 0, "area": 20781.4186, "rle": {"size": [640, 480], "counts": "SoP11kc06I7I7I7J6I7I7J6I7I7J6G9F:F:F:F:F:K5RBgKQ<Z4nCfKQ<\\4mCeKS<\\4lCeKS<\\4kCeKT<]4kCcKU<^4iCdKU<^4jCbKV<_4iCaKW<`4gCbKW<`4hC`KX<a4fC`KZ<`4fCaKX<a4fC`KZ<a4eC_KZ<c4eC^KZ<c4dC^K\\<c4cC]K\\<e4bC]K]<d4bC\\K]<f4aC[K_<f4`C[K_<f4`CZK_<h4_CYKa<h4]CZKb<g4XC^Kg<d4TC`Kd;YOQEW5VOfKc;_OoDm4XOiKc;GnDj5m:aJmD`5m:lJkDU5U;k1000M3N1O2M3N2N2M3N1O2M3N2N2N101N2O1O1O1O1O1O1O1O1O001O100O100O1O10N12O2M2O1N2N2O1N3fJoBa3R=\\LQCb3Q=[LQCd3Q=YLRCf3o<WLSCh3o<ULTCj3n<RLTCm3n<PLUCn3o<mKSCR4U=eKnBX4Z=`KhB_4`=XKcBe4]>O2M2hLk@c1X?VNQAf1P?TNYAg1j>SN]Aj1d>PNeAk1^>nMkAn1V>mMQBo1T`0L4K4G:@`0_Oa0@PfP6"}, "label": "a child wrapped in a towel to the left of a wash tub"}]}
{"id": 534292, "image": "COCO_train2014_000000534292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little boy on the bench on the far left wrapped in a towl\"."}], "task": "refering", "answer_template": "The \"the little boy on the bench on the far left wrapped in a towl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 207, 208, 209, 223, 224, 225, 226, 240, 241, 242, 243, 257, 258, 259, 260, 274, 275, 276, 277, 290, 291, 292, 293, 294, 308, 309, 310, 311, 326, 327, 328, 342, 343, 360], "bbox": [0.10908333333333334, 0.47701562500000005, 0.35520833333333335, 0.9213906250000001], "iscrowd": 0, "area": 20781.4186, "rle": {"size": [640, 480], "counts": "SoP11kc06I7I7I7J6I7I7J6I7I7J6G9F:F:F:F:F:K5RBgKQ<Z4nCfKQ<\\4mCeKS<\\4lCeKS<\\4kCeKT<]4kCcKU<^4iCdKU<^4jCbKV<_4iCaKW<`4gCbKW<`4hC`KX<a4fC`KZ<`4fCaKX<a4fC`KZ<a4eC_KZ<c4eC^KZ<c4dC^K\\<c4cC]K\\<e4bC]K]<d4bC\\K]<f4aC[K_<f4`C[K_<f4`CZK_<h4_CYKa<h4]CZKb<g4XC^Kg<d4TC`Kd;YOQEW5VOfKc;_OoDm4XOiKc;GnDj5m:aJmD`5m:lJkDU5U;k1000M3N1O2M3N2N2M3N1O2M3N2N2N101N2O1O1O1O1O1O1O1O1O001O100O100O1O10N12O2M2O1N2N2O1N3fJoBa3R=\\LQCb3Q=[LQCd3Q=YLRCf3o<WLSCh3o<ULTCj3n<RLTCm3n<PLUCn3o<mKSCR4U=eKnBX4Z=`KhB_4`=XKcBe4]>O2M2hLk@c1X?VNQAf1P?TNYAg1j>SN]Aj1d>PNeAk1^>nMkAn1V>mMQBo1T`0L4K4G:@`0_Oa0@PfP6"}, "label": "the little boy on the bench on the far left wrapped in a towl"}]}
{"id": 337691, "image": "COCO_train2014_000000337691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding a hotdog\"."}], "task": "refering", "answer_template": "The \"a hand holding a hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 92, 93, 94, 95, 96, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 156, 157, 158, 159], "bbox": [0.0, 0.34832, 0.44344, 0.71236], "iscrowd": 0, "area": 21795.803799999998, "rle": {"size": [500, 375], "counts": "f:>Y>m0SOm0VOi001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O1O2N1O2M3N1O2N1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1N101O1O1O001O1O1O001O1O1O1O001O1000O10O100000O0100000O01000000O010000O010000O010000O01000O10O1000O10O10O1000O10O1000O01000O10O1000O010001O0O10001N1000000O2O000O101O000O101J5K6J6J6I7J5K6J6J6I7J6J6J5J7J6J6J6J6I7J6J5K6J6IdkU3"}, "label": "a hand holding a hotdog"}]}
{"id": 337691, "image": "COCO_train2014_000000337691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white hand that is holding a burnt hot dog\"."}], "task": "refering", "answer_template": "The \"a white hand that is holding a burnt hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 92, 93, 94, 95, 96, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 156, 157, 158, 159], "bbox": [0.0, 0.34832, 0.44344, 0.71236], "iscrowd": 0, "area": 21795.803799999998, "rle": {"size": [500, 375], "counts": "f:>Y>m0SOm0VOi001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O1O2N1O2M3N1O2N1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1N101O1O1O001O1O1O001O1O1O1O001O1000O10O100000O0100000O01000000O010000O010000O010000O01000O10O1000O10O10O1000O10O1000O01000O10O1000O010001O0O10001N1000000O2O000O101O000O101J5K6J6J6I7J5K6J6J6I7J6J6J5J7J6J6J6J6I7J6J5K6J6IdkU3"}, "label": "a white hand that is holding a burnt hot dog"}]}
{"id": 337691, "image": "COCO_train2014_000000337691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black pants\"."}], "task": "refering", "answer_template": "The \"black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 39, 40, 41, 42, 43, 52, 53, 54, 55, 65, 66, 67, 68, 69, 78, 79, 80, 81], "bbox": [0.00328, 0.00452, 0.3753066666666666, 0.39218000000000003], "iscrowd": 0, "area": 21254.9761, "rle": {"size": [500, 375], "counts": "^d0Z1[:o3N2O1O1O100O1000000O0100000O1000000O1000000O1000000O1000000O10O1000O1000000O1000000O1000000O100000000O10O10000000O100000000O10000000000O100000000O10O10000O1O010O1O100O00100O1O100SNRFSNn9n1RFQNo9n1RFQNo9n1SFQNm9o1SFPNn9o1TFoMl9Q2VFnMj9Q2XFmMi9S2WFlMj9S2XFlMh9S2ZFkMh9S2ZFkMg9U2ZFjMf9U2\\FiMe9V2]FhMd9X2]FfMd9Y2^FfMb9Y2`FeMa9Z2bFcM`9\\2aFcM_9\\2cFbM^9]2dFbM\\9[2hFdMX9Y2lFeMV9W2nFhMR9U2RGiMo8T2UGkMl8Q2XGmMi8P2\\GnMd8o1`GoMa8n1cGPN_8l1eGRN\\8k1hGRNZ8k1jGSNW8j1mGSNV8i1l2M3M4L3M3M3M4L5K4L5K5K5K5K5KdWb3"}, "label": "black pants"}]}
{"id": 337691, "image": "COCO_train2014_000000337691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in black pants\"."}], "task": "refering", "answer_template": "The \"a person in black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 39, 40, 41, 42, 43, 52, 53, 54, 55, 65, 66, 67, 68, 69, 78, 79, 80, 81], "bbox": [0.00328, 0.00452, 0.3753066666666666, 0.39218000000000003], "iscrowd": 0, "area": 21254.9761, "rle": {"size": [500, 375], "counts": "^d0Z1[:o3N2O1O1O100O1000000O0100000O1000000O1000000O1000000O1000000O10O1000O1000000O1000000O1000000O100000000O10O10000000O100000000O10000000000O100000000O10O10000O1O010O1O100O00100O1O100SNRFSNn9n1RFQNo9n1RFQNo9n1SFQNm9o1SFPNn9o1TFoMl9Q2VFnMj9Q2XFmMi9S2WFlMj9S2XFlMh9S2ZFkMh9S2ZFkMg9U2ZFjMf9U2\\FiMe9V2]FhMd9X2]FfMd9Y2^FfMb9Y2`FeMa9Z2bFcM`9\\2aFcM_9\\2cFbM^9]2dFbM\\9[2hFdMX9Y2lFeMV9W2nFhMR9U2RGiMo8T2UGkMl8Q2XGmMi8P2\\GnMd8o1`GoMa8n1cGPN_8l1eGRN\\8k1hGRNZ8k1jGSNW8j1mGSNV8i1l2M3M4L3M3M3M4L5K4L5K5K5K5K5KdWb3"}, "label": "a person in black pants"}]}
{"id": 337691, "image": "COCO_train2014_000000337691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman ' s leg in fishnet stockings\"."}], "task": "refering", "answer_template": "The \"a woman ' s leg in fishnet stockings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 19, 20, 21, 22, 23, 24, 25, 35, 36, 37, 38], "bbox": [0.4074933333333333, 0.0, 0.9977599999999999, 0.14158], "iscrowd": 0, "area": 4999.238, "rle": {"size": [500, 375], "counts": "nfZ21c?0O2O0O2O1O0O2O001N2O0O2O001N2O001N101N2O001O1O001O001O1O001O001O1N101O001O1O001O001O1O001O001O1O001O001N2O001O1O001O001O1O000100O001O1O00100O001O001O10O01O1O001O1O010O1O001O00100O0N3N2M2N3M3M2N3M2O2M2N2N3M2NR\\58icJ:E;E5LO00001O001O0O101O00001O00001O00001N10001O00001I6J6I5Lje=c0jYBO1O0001O0000000O2O000000001O000000001O000000001O0000000O2O000000001O00000000001N1007I8H9Hc="}, "label": "a woman ' s leg in fishnet stockings"}]}
{"id": 329501, "image": "COCO_train2014_000000329501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the police officer that is to the outermost right of the two officers in the photo\"."}], "task": "refering", "answer_template": "The \"the police officer that is to the outermost right of the two officers in the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 227, 228, 229, 244, 245, 246, 247, 261, 262, 263, 278, 279, 280, 295, 296, 312, 313, 329, 330, 346, 347], "bbox": [0.35483333333333333, 0.535765625, 0.5584583333333334, 0.927125], "iscrowd": 0, "area": 10312.0159, "rle": {"size": [640, 480], "counts": "YhZ31cc0`0@?A?A?A`0A>K5K6J5K5K5J7J5K5E;D=C<D;E;J5N2O1N3N1N2O1N3N1TO]IYD`6i;bIUDZ6o;iIoCS6U<nIiCn5[<TJdCh5_<ZJ_Cb5f<`JXC[5l<gJRCV5Q=lJnBo4V=YKbBc4c=i03L4M4K4L4M4K5K5L4K5L4K5L4L4L4L4L4L4K5L4^Mh_OX1\\`0eNg_OW1]`0gNf_OT1^`0jNd_OR1``0YN^_O26_1b`0\\NR@_1S`0^NQ@[1]a0J7I6J7N1010O01O0010O02N2N2OO000O2O0O100O21N1O1O2O0O2N101N1O2N2OVST4"}, "label": "the police officer that is to the outermost right of the two officers in the photo"}]}
{"id": 329501, "image": "COCO_train2014_000000329501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man closest to the signs\"."}], "task": "refering", "answer_template": "The \"man closest to the signs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 227, 228, 229, 244, 245, 246, 247, 261, 262, 263, 278, 279, 280, 295, 296, 312, 313, 329, 330, 346, 347], "bbox": [0.35483333333333333, 0.535765625, 0.5584583333333334, 0.927125], "iscrowd": 0, "area": 10312.0159, "rle": {"size": [640, 480], "counts": "YhZ31cc0`0@?A?A?A`0A>K5K6J5K5K5J7J5K5E;D=C<D;E;J5N2O1N3N1N2O1N3N1TO]IYD`6i;bIUDZ6o;iIoCS6U<nIiCn5[<TJdCh5_<ZJ_Cb5f<`JXC[5l<gJRCV5Q=lJnBo4V=YKbBc4c=i03L4M4K4L4M4K5K5L4K5L4K5L4L4L4L4L4L4K5L4^Mh_OX1\\`0eNg_OW1]`0gNf_OT1^`0jNd_OR1``0YN^_O26_1b`0\\NR@_1S`0^NQ@[1]a0J7I6J7N1010O01O0010O02N2N2OO000O2O0O100O21N1O1O2O0O2N101N1O2N2OVST4"}, "label": "man closest to the signs"}]}
{"id": 1822, "image": "COCO_train2014_000000001822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest chair\"."}], "task": "refering", "answer_template": "The \"the closest chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 190, 191, 192, 193, 214, 215, 216, 217, 237, 238, 239, 240, 259, 260, 262, 263], "bbox": [0.27868750000000003, 0.4406666666666667, 0.46510937500000005, 0.7130833333333333], "iscrowd": 0, "area": 7372.441649999999, "rle": {"size": [480, 640], "counts": "hSd24k>4K5L4M3M3M3M3M3M3cDPOd8S1UGQOk8R1nFROR9Q1gFSOY9P1`FTO`9o0XFVOh9m0QFWOo9l0jEWOW:l0bEVOa:l0WESOo:P1jDoN[;T1^DjNh;Q26K5N200O1000000O1O100O2N100O1O100O100O1O100O100000001O0i0WO00O1000000XOSD]Nm;c1`DoMa;Q2g01O0100O10O0101N3M3N2M3N101N2O1N2O0OO2N=Bh0YO000O1000E:D=C=G801N2O001N2O011N2N2N2O1N2O1000O1000002N2N2N24M2M3M3M3M3L5J5J6J6JQPP5"}, "label": "the closest chair"}]}
{"id": 1822, "image": "COCO_train2014_000000001822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair , different from the other chairs\"."}], "task": "refering", "answer_template": "The \"a wooden chair , different from the other chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 190, 191, 192, 193, 214, 215, 216, 217, 237, 238, 239, 240, 259, 260, 262, 263], "bbox": [0.27868750000000003, 0.4406666666666667, 0.46510937500000005, 0.7130833333333333], "iscrowd": 0, "area": 7372.441649999999, "rle": {"size": [480, 640], "counts": "hSd24k>4K5L4M3M3M3M3M3M3cDPOd8S1UGQOk8R1nFROR9Q1gFSOY9P1`FTO`9o0XFVOh9m0QFWOo9l0jEWOW:l0bEVOa:l0WESOo:P1jDoN[;T1^DjNh;Q26K5N200O1000000O1O100O2N100O1O100O100O1O100O100000001O0i0WO00O1000000XOSD]Nm;c1`DoMa;Q2g01O0100O10O0101N3M3N2M3N101N2O1N2O0OO2N=Bh0YO000O1000E:D=C=G801N2O001N2O011N2N2N2O1N2O1000O1000002N2N2N24M2M3M3M3M3L5J5J6J6JQPP5"}, "label": "a wooden chair , different from the other chairs"}]}
{"id": 1822, "image": "COCO_train2014_000000001822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair to the far right of the couch with gold trim\"."}], "task": "refering", "answer_template": "The \"a chair to the far right of the couch with gold trim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 272, 273], "bbox": [0.7285781250000001, 0.376625, 0.9448125, 0.6883125], "iscrowd": 0, "area": 9847.328650000001, "rle": {"size": [480, 640], "counts": "cgj61l>5K4L5L3L4L1O2N1O2L3J7H7I7M41N101N101N101N12O0O100O100O101N100O100O100O1O2O0O100O100O101N100O100O10001O01O0000000010O00000001O000001O01O0000000010O00000001O0001O0001O0004L7J7H8H2N00010O00000001O01J5H8IXOeDYNT;m1YEgMe:Z2hEZMV:h2P10O1O100O2O0O100O1O6K5J7J6J10O010000O10O0100O10O102H7B>A?B>B>B>B?A>B]W`0"}, "label": "a chair to the far right of the couch with gold trim"}]}
{"id": 1822, "image": "COCO_train2014_000000001822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a golden edged chair that faces the television\"."}], "task": "refering", "answer_template": "The \"this is a golden edged chair that faces the television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 272, 273], "bbox": [0.7285781250000001, 0.376625, 0.9448125, 0.6883125], "iscrowd": 0, "area": 9847.328650000001, "rle": {"size": [480, 640], "counts": "cgj61l>5K4L5L3L4L1O2N1O2L3J7H7I7M41N101N101N101N12O0O100O100O101N100O100O100O1O2O0O100O100O101N100O100O10001O01O0000000010O00000001O000001O01O0000000010O00000001O0001O0001O0004L7J7H8H2N00010O00000001O01J5H8IXOeDYNT;m1YEgMe:Z2hEZMV:h2P10O1O100O2O0O100O1O6K5J7J6J10O010000O10O0100O10O102H7B>A?B>B>B>B?A>B]W`0"}, "label": "this is a golden edged chair that faces the television"}]}
{"id": 395040, "image": "COCO_train2014_000000395040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"snow shooter on cross country skis in a competition\"."}], "task": "refering", "answer_template": "The \"snow shooter on cross country skis in a competition\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 195, 196, 218, 219, 241, 242, 263, 264, 265, 266, 286, 287, 288, 289], "bbox": [0.346203125, 0.23976470588235296, 0.6670468749999999, 0.8420000000000001], "iscrowd": 0, "area": 24038.892400000004, "rle": {"size": [425, 640], "counts": "jXl28i<:F9H9N1N2O1N2O001N2]EeNV9\\1eFiNY9X1bFmN^9T1]FQOb9P1[FROd9P1[FPOe9P1[FQOc9Q1\\FoNd9R1[FnNd9T1[FmNd9S1\\FmNd9T1[FmNd9T1[FlNe9T1[FlNe9T1[FlNe9U1ZFkNf9U1ZFkNf9U1ZFlNe9T1[FlNe9U1ZFkNf9U1ZFkNf9U1ZFkNf9U1[FjNe9W1ZFiNf9W1[FhNe9X1[FhNe9X1[FhNe9Y1[FgNd9Y1\\FgNd9Y1]FfNc9Z1]FfNc9[1\\FeNd9S1eFlN[9k0nFUOR9m0mFROS9Q1jFoNV9T1gFmNX9V1fFiNZ9Y1dFgN[9]1cFbN]9a1`F_N`9d1]F\\Nc9f1\\FYNd9h1[FXNe9i1ZFXNe9i1[FVNd9l1[FTNe9m1ZFSNf9n1ZFQNf9P2YFPNg9Q2XFPNg9Q2XFoMh9R2WFnMi9S2VFmMi9U2VFkMj9V2UFjMk9W2TFjMk9W2TFiMl9X2SFhMm9j20000O2lIdLY2\\3eMfL[2Z3cMhL]2X3aMkL^2U3`MnL_2R3`MQM^2P3`MSM^2m2bMUM[2l2dMWMZ2i2eMZMY2f2gM[MX2e2gM^MW2c2hM_MV2a2iM`MW2`2iM`MW2`2hMaMW2`2hMbMW2^2iMbMW2_2gMbMY2^2gMbMY2^2fMcMZ2]2mJaLZ2S1i2\\2fJlL]2h0m2R3nLoLR3T3kLmLT3W3gLjLY3Z3bLgL^3]3^LcLb3a3YL`Lf3e3VL[Lj3i3QLXLo3a601O0000000000001O0000O1000000O1000000O100WLfJCY5>kJ^OU5a0PKZOQ5f0SKVOm4i0XKSOh4m0\\KoNd4P1aKkN`4U1eKfN[4Z1oK\\NQ4c1[LRNe3n1cLiM^3V2iLdMW3\\2oL^MQ3a2UMZMk2f2[MSMf2l2_MPMb2n2bMoL^2P3gMlLY2S3kMiLV2W3mMfLT2X3QNdLo1[3UNbLk1]3YN_Li1_3Q400O1O2N1O100O2N1O1O2O0O1O1O2N100O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N2N3M2N3M2M4M2N3M2N3M2N2N3L3M4K4M4L3L5L3M3M4K4M4LZXh2"}, "label": "snow shooter on cross country skis in a competition"}]}
{"id": 395040, "image": "COCO_train2014_000000395040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and black ski suit on a woman\"."}], "task": "refering", "answer_template": "The \"a red and black ski suit on a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 195, 196, 218, 219, 241, 242, 263, 264, 265, 266, 286, 287, 288, 289], "bbox": [0.346203125, 0.23976470588235296, 0.6670468749999999, 0.8420000000000001], "iscrowd": 0, "area": 24038.892400000004, "rle": {"size": [425, 640], "counts": "jXl28i<:F9H9N1N2O1N2O001N2]EeNV9\\1eFiNY9X1bFmN^9T1]FQOb9P1[FROd9P1[FPOe9P1[FQOc9Q1\\FoNd9R1[FnNd9T1[FmNd9S1\\FmNd9T1[FmNd9T1[FlNe9T1[FlNe9T1[FlNe9U1ZFkNf9U1ZFkNf9U1ZFlNe9T1[FlNe9U1ZFkNf9U1ZFkNf9U1ZFkNf9U1[FjNe9W1ZFiNf9W1[FhNe9X1[FhNe9X1[FhNe9Y1[FgNd9Y1\\FgNd9Y1]FfNc9Z1]FfNc9[1\\FeNd9S1eFlN[9k0nFUOR9m0mFROS9Q1jFoNV9T1gFmNX9V1fFiNZ9Y1dFgN[9]1cFbN]9a1`F_N`9d1]F\\Nc9f1\\FYNd9h1[FXNe9i1ZFXNe9i1[FVNd9l1[FTNe9m1ZFSNf9n1ZFQNf9P2YFPNg9Q2XFPNg9Q2XFoMh9R2WFnMi9S2VFmMi9U2VFkMj9V2UFjMk9W2TFjMk9W2TFiMl9X2SFhMm9j20000O2lIdLY2\\3eMfL[2Z3cMhL]2X3aMkL^2U3`MnL_2R3`MQM^2P3`MSM^2m2bMUM[2l2dMWMZ2i2eMZMY2f2gM[MX2e2gM^MW2c2hM_MV2a2iM`MW2`2iM`MW2`2hMaMW2`2hMbMW2^2iMbMW2_2gMbMY2^2gMbMY2^2fMcMZ2]2mJaLZ2S1i2\\2fJlL]2h0m2R3nLoLR3T3kLmLT3W3gLjLY3Z3bLgL^3]3^LcLb3a3YL`Lf3e3VL[Lj3i3QLXLo3a601O0000000000001O0000O1000000O1000000O100WLfJCY5>kJ^OU5a0PKZOQ5f0SKVOm4i0XKSOh4m0\\KoNd4P1aKkN`4U1eKfN[4Z1oK\\NQ4c1[LRNe3n1cLiM^3V2iLdMW3\\2oL^MQ3a2UMZMk2f2[MSMf2l2_MPMb2n2bMoL^2P3gMlLY2S3kMiLV2W3mMfLT2X3QNdLo1[3UNbLk1]3YN_Li1_3Q400O1O2N1O100O2N1O1O2O0O1O1O2N100O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N2N3M2N3M2M4M2N3M2N3M2N2N3L3M4K4M4L3L5L3M3M4K4M4LZXh2"}, "label": "a red and black ski suit on a woman"}]}
{"id": 198434, "image": "COCO_train2014_000000198434.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman leads a white horse by the reins while a car drives by\"."}], "task": "refering", "answer_template": "The \"a young woman leads a white horse by the reins while a car drives by\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 88, 89, 110, 111, 112, 113, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 225, 226, 227, 228, 249, 250, 272, 273, 274, 295, 296, 297, 318, 319, 320, 341, 342], "bbox": [0.674578125, 0.1367605633802817, 1.0, 0.9928169014084507], "iscrowd": 0, "area": 27487.575800000002, "rle": {"size": [426, 640], "counts": "emc51X=1O1O1O2N1N2O1O1O1O2O00000O1O101N100O100O1O2O0O10O0100O1O100O100O1O100O100O100O1O100O100O100O1O100O010O010O001N101N1O2O0O2O0O2N101N1O2O0O2N101N101N2N2N2N1O2M3N2N2M3kEkMS9W2gFSNR9o1hF[NS9g1gFcNT9^1fFlNV9f2N2M3N2N2M3N3M3M5J6K5K5I7H8H?Ag0UNZIgKU7`1aH3k8CXG;Q9[OQGe0X9POkFn0^9hNdFX1d9^N^Fa1l9TNWFk1Q:kMQFT2c:01N101N100O2N100O100001O0O1000001O00000000001O0O101O4L4L4L4L4L2N3M7I>B5J2O00001G8XLZLbMQ4X2]LSMo3i2]LcLP4W3]LULo3f3^LfKn3V4f2J5M4N101O001O0O101lNoGmLQ8n1Y2D;100O11O1O2N1O1O1N2000O1O2N1O2N101N1O2N1O2O0O001O000001O000000001O2N2N2N1O2NZJ"}, "label": "a young woman leads a white horse by the reins while a car drives by"}]}
{"id": 198434, "image": "COCO_train2014_000000198434.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with the pony tail holding the horses reighn in both hands\"."}], "task": "refering", "answer_template": "The \"the girl with the pony tail holding the horses reighn in both hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 88, 89, 110, 111, 112, 113, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 225, 226, 227, 228, 249, 250, 272, 273, 274, 295, 296, 297, 318, 319, 320, 341, 342], "bbox": [0.674578125, 0.1367605633802817, 1.0, 0.9928169014084507], "iscrowd": 0, "area": 27487.575800000002, "rle": {"size": [426, 640], "counts": "emc51X=1O1O1O2N1N2O1O1O1O2O00000O1O101N100O100O1O2O0O10O0100O1O100O100O1O100O100O100O1O100O100O100O1O100O010O010O001N101N1O2O0O2O0O2N101N1O2O0O2N101N101N2N2N2N1O2M3N2N2M3kEkMS9W2gFSNR9o1hF[NS9g1gFcNT9^1fFlNV9f2N2M3N2N2M3N3M3M5J6K5K5I7H8H?Ag0UNZIgKU7`1aH3k8CXG;Q9[OQGe0X9POkFn0^9hNdFX1d9^N^Fa1l9TNWFk1Q:kMQFT2c:01N101N100O2N100O100001O0O1000001O00000000001O0O101O4L4L4L4L4L2N3M7I>B5J2O00001G8XLZLbMQ4X2]LSMo3i2]LcLP4W3]LULo3f3^LfKn3V4f2J5M4N101O001O0O101lNoGmLQ8n1Y2D;100O11O1O2N1O1O1N2000O1O2N1O2N101N1O2N1O2O0O001O000001O000000001O2N2N2N1O2NZJ"}, "label": "the girl with the pony tail holding the horses reighn in both hands"}]}
{"id": 321318, "image": "COCO_train2014_000000321318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy wearing healmet with bat on the ground\"."}], "task": "refering", "answer_template": "The \"boy wearing healmet with bat on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 79, 80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 284, 285, 286, 289, 307, 308, 309, 312, 313], "bbox": [0.36507812500000003, 0.09049411764705882, 0.6220000000000001, 0.9321176470588235], "iscrowd": 0, "area": 30232.607950000005, "rle": {"size": [425, 640], "counts": "o_Q33S=3M3M4L3M4L3M3M4L3M4M2M3M3M2N3M2N3M2N3M3M2N3M2O2M2E<C=D;E<M2M4M2N3L4M2N3L3N3M2M4M3L3M4L3M4bN\\1N2M3N2N1N3oMTIeMn6\\2XI[Mj6f2\\IPMg6Q3`IeLa6]3eIZL]6f3jIPLY6Q4^11O1N2O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O0O2O1O1O1O1NjJRHW5l7iJWHV5o7O0O2O1O0O2O1O001N20O1000001N10000O10000O101O0001O0104L3M3M3L4M3M3M3M3M3L4MO1O1N101N2O1O0O2O1O1N2O2N2M3N1NZKeJa1Y5[NRK`1l4\\N_K^1_4_NmKZ1R4bNYLW1g3dNeLU1Z3hNQMQ1o2kN\\Mn0d2nNhMi0X2TOYO_Og0=_OA`0;GC99LF462HM1=LDLf02Z6O2K5L3M4L4L4K4M4LVRT3"}, "label": "boy wearing healmet with bat on the ground"}]}
{"id": 321318, "image": "COCO_train2014_000000321318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one child with helmet and cap both on his head holding a bat\"."}], "task": "refering", "answer_template": "The \"there is one child with helmet and cap both on his head holding a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 79, 80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 284, 285, 286, 289, 307, 308, 309, 312, 313], "bbox": [0.36507812500000003, 0.09049411764705882, 0.6220000000000001, 0.9321176470588235], "iscrowd": 0, "area": 30232.607950000005, "rle": {"size": [425, 640], "counts": "o_Q33S=3M3M4L3M4L3M3M4L3M4M2M3M3M2N3M2N3M2N3M3M2N3M2O2M2E<C=D;E<M2M4M2N3L4M2N3L3N3M2M4M3L3M4L3M4bN\\1N2M3N2N1N3oMTIeMn6\\2XI[Mj6f2\\IPMg6Q3`IeLa6]3eIZL]6f3jIPLY6Q4^11O1N2O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O0O2O1O1O1O1NjJRHW5l7iJWHV5o7O0O2O1O0O2O1O001N20O1000001N10000O10000O101O0001O0104L3M3M3L4M3M3M3M3M3L4MO1O1N101N2O1O0O2O1O1N2O2N2M3N1NZKeJa1Y5[NRK`1l4\\N_K^1_4_NmKZ1R4bNYLW1g3dNeLU1Z3hNQMQ1o2kN\\Mn0d2nNhMi0X2TOYO_Og0=_OA`0;GC99LF462HM1=LDLf02Z6O2K5L3M4L4L4K4M4LVRT3"}, "label": "there is one child with helmet and cap both on his head holding a bat"}]}
{"id": 321318, "image": "COCO_train2014_000000321318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing jeans and carrying a black bag\"."}], "task": "refering", "answer_template": "The \"a person wearing jeans and carrying a black bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 64, 65, 66, 87, 88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 180, 181, 203, 204, 226, 227], "bbox": [0.7924375, 0.09804705882352942, 0.9330468749999999, 0.6259058823529412], "iscrowd": 0, "area": 11407.080200000002, "rle": {"size": [425, 640], "counts": "Rab6>g<5L4K5L4K5K5YOhNZEZ1Z:TObEn0U:]OgEe0X:^OcEe0[:@`Eb0_:U1N5L;D<E:E3N1N2O2M2O4^IcKP4b4kK^KU4f4fK[KZ4j4`KXK_4l4\\KUKd4P5WKPKi4U5QKlJP5W5kJjJU5[5]JnJc5V5nIUKR6Q61O1O1O2N1O101N3M3M3M2N3M2N1O00001O0001O01iNXIkKh6Y4R14L4M3L]NTHlMi7o1_HRN^7h1jHYNR7b1UI`Nh6\\1^IeN`6X1cIgN^6W1eIgN\\6V1hIgN[6V1gIiNZ6T1jIiNX6?^HbN^1m0V6>fH[NV1U1V6?aJ^Oa5a0aJ\\Ob5a0aJ\\Oa5c0aJZOa5d0bJYO`5f0bJWOa5g0aJVO_5j0cJTO^5l0dJQO\\5n0a3M3L3N3M2N5J6K6J5K6IWga0"}, "label": "a person wearing jeans and carrying a black bag"}]}
{"id": 321318, "image": "COCO_train2014_000000321318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a black coat and hood , with a black purse and blue jeans on the other side of the fence\"."}], "task": "refering", "answer_template": "The \"a person with a black coat and hood , with a black purse and blue jeans on the other side of the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 64, 65, 66, 87, 88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 180, 181, 203, 204, 226, 227], "bbox": [0.7924375, 0.09804705882352942, 0.9330468749999999, 0.6259058823529412], "iscrowd": 0, "area": 11407.080200000002, "rle": {"size": [425, 640], "counts": "Rab6>g<5L4K5L4K5K5YOhNZEZ1Z:TObEn0U:]OgEe0X:^OcEe0[:@`Eb0_:U1N5L;D<E:E3N1N2O2M2O4^IcKP4b4kK^KU4f4fK[KZ4j4`KXK_4l4\\KUKd4P5WKPKi4U5QKlJP5W5kJjJU5[5]JnJc5V5nIUKR6Q61O1O1O2N1O101N3M3M3M2N3M2N1O00001O0001O01iNXIkKh6Y4R14L4M3L]NTHlMi7o1_HRN^7h1jHYNR7b1UI`Nh6\\1^IeN`6X1cIgN^6W1eIgN\\6V1hIgN[6V1gIiNZ6T1jIiNX6?^HbN^1m0V6>fH[NV1U1V6?aJ^Oa5a0aJ\\Ob5a0aJ\\Oa5c0aJZOa5d0bJYO`5f0bJWOa5g0aJVO_5j0cJTO^5l0dJQO\\5n0a3M3L3N3M2N5J6K6J5K6IWga0"}, "label": "a person with a black coat and hood , with a black purse and blue jeans on the other side of the fence"}]}
{"id": 534311, "image": "COCO_train2014_000000534311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow with head down on the ground\"."}], "task": "refering", "answer_template": "The \"cow with head down on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.4875, 0.5612777777777778, 0.789921875, 0.9204166666666668], "iscrowd": 0, "area": 17360.698800000002, "rle": {"size": [360, 640], "counts": "io]31S;2QOOhF3X90cF2^9O]F4c9OXF4g9OSF5m9MmE7R:`000O1O100O20O0000000001O00000000010O0000000O1O1N3M2O1N2N2O1N2N201N10000O100O100O101O0O100O100O10001N1O1N2O1O2M2O1N2O2N1O101N100O2O0O2O0O101N101M2O1O2N1O1O2N1N3N1O1O2N1O1N3N1O1O2N1O1O2N1O101N1O1000000000000001O000000001O0000001O0000001O0000001O000000000000000000000000O10000O10O0100O10O10O1001O0O2O001O0O2O001O0O2O001N1O2O0O2N2N2O0O2N2N2O1N2N2M3M3M3M3M3M3M3M3N3L8H8I7H<_Of0VO[k^1"}, "label": "cow with head down on the ground"}]}
{"id": 534311, "image": "COCO_train2014_000000534311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow eating grass\"."}], "task": "refering", "answer_template": "The \"cow eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.4875, 0.5612777777777778, 0.789921875, 0.9204166666666668], "iscrowd": 0, "area": 17360.698800000002, "rle": {"size": [360, 640], "counts": "io]31S;2QOOhF3X90cF2^9O]F4c9OXF4g9OSF5m9MmE7R:`000O1O100O20O0000000001O00000000010O0000000O1O1N3M2O1N2N2O1N2N201N10000O100O100O101O0O100O100O10001N1O1N2O1O2M2O1N2O2N1O101N100O2O0O2O0O101N101M2O1O2N1O1O2N1N3N1O1O2N1O1N3N1O1O2N1O1O2N1O101N1O1000000000000001O000000001O0000001O0000001O0000001O000000000000000000000000O10000O10O0100O10O10O1001O0O2O001O0O2O001O0O2O001N1O2O0O2N2N2O0O2N2N2O1N2N2M3M3M3M3M3M3M3M3N3L8H8I7H<_Of0VO[k^1"}, "label": "cow eating grass"}]}
{"id": 534311, "image": "COCO_train2014_000000534311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third cow from the front\"."}], "task": "refering", "answer_template": "The \"the third cow from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 239, 240], "bbox": [0.41479687500000006, 0.5346111111111111, 0.593984375, 0.8148055555555557], "iscrowd": 0, "area": 6973.4510999999975, "rle": {"size": [360, 640], "counts": "l\\m22i:>L5J5L4L5K>A?C=C7I2M3N3M2N2O100O2MnLnGR3S8100000O1nNnGlNT8T1PHfNR8[1QH_NQ8a1THWNo7i1UHQNm7o1f000010O000000001O00000000010O00000000001O7I4L0010O01O00O1O1O1O1O2N1O100L4L4L4O10001O00000O100000000O2O000000000O101O0000000O2O001O1O1O1N2O1O1N2O1N101N2N2O1N2O2M3L4L5K4L4L4LlUk2"}, "label": "the third cow from the front"}]}
{"id": 534311, "image": "COCO_train2014_000000534311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bull next to the left of a bull laying down\"."}], "task": "refering", "answer_template": "The \"a bull next to the left of a bull laying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 187, 189, 190, 191, 192, 193, 210, 211, 213, 214, 215, 216, 233, 234, 236, 237, 238, 239, 260, 261, 262], "bbox": [0.13525, 0.12583333333333332, 0.558703125, 0.9078611111111111], "iscrowd": 0, "area": 40782.600150000006, "rle": {"size": [360, 640], "counts": "Tin05P;6J6J7I7^N]OTHj0W73[H3[7?VHI_7X2E:Ba0[Og0]JeJa4k5PKYJo4Q6M3N3L5L3N1O1O1O1O001O1O[N]LgKb3m3oLoKP3j3\\MRLc2l3dMPL[2m3mMoKR2o3UNlKl1P4\\NlKd1Q4cNkK\\1R4o2M3N21O1O1O1N2O1O1O1O1O1N2O1N2O2M2N2N2O1N2N2N2O1O1O100O1O1O00O1000000002O7H7I8G9H5J4M3M2N3M3N2M3N2M3N2M8I5JO2N1N3N1O2PNXL_Lg3`3WMfKf2X4_MhK^2W4fMiKX2U4lMiKS2V4PNhKP2X4SNeKm1[4VNbKk1]4XN`Kh1`4[N]Ke1c4\\N\\Kd1d4]N[Kd1d4^NZKb1f4_NYKa1g4`NXK`1h4bNVK_1i4bNVK^1l4aNSK_1o4\\22O2M2N2N3N1N2N5K6K3L4L4L4YLWJg1l5jMaJV2Y7O1O1O1O010O00000000000ObLTMbNl2_1WMfLO0k2Z3[M^L05e2^3^MVL39_2b3aMmK8=W2f3fMeK9b0P2k3ZORLf0o3\\OnKd0R4@jK`0W4AgK?Y4DdK<]4F`K9b4I[K7e4KYK5h4MUK3l4T3010O001O1kKoJQ2Q5hMXKW2h4ZLUKe0;o2`4XLSMa3b5H000000000000O10000000001nNfLTJ[3b5QM[JP3[5]MaJc2W5iMeJX2n4WNoJj1f4eNUK\\1f4lNVKU1e4SOWKm0f4m2K4M2N1O4M2_OUIbLo6]3QIaLQ7_3PI]LT7b3lH\\LV7d3kHYLS7l35O1O1O2N1OO2N1N2O1007JN10000O2O000O01O4L0O2O1O1O016KO0000O1M3POU1B9H9L310O010O0010O0001O0mN\\Fb0d9]O]Fc0c9[O_Fe0a9ZO`Ff0`9ZO`Ff0`9YOaFg0_9YO`Fh0`9WOaFi0_9WOaFi0_9WOaFi0^9WOcFi0\\9XOdFh0[9XOfFh0Y9YOhFf0W9[OiFe0V9[OkFd0W9[OjFd0W9ZOjFf0X9XOhFh0Y9WOhFg0m900O2O0O101N100O2O0O5L4K5Lo[S3"}, "label": "a bull next to the left of a bull laying down"}]}
{"id": 534311, "image": "COCO_train2014_000000534311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and brown bull sitting next to a cow\"."}], "task": "refering", "answer_template": "The \"a white and brown bull sitting next to a cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 187, 189, 190, 191, 192, 193, 210, 211, 213, 214, 215, 216, 233, 234, 236, 237, 238, 239, 260, 261, 262], "bbox": [0.13525, 0.12583333333333332, 0.558703125, 0.9078611111111111], "iscrowd": 0, "area": 40782.600150000006, "rle": {"size": [360, 640], "counts": "Tin05P;6J6J7I7^N]OTHj0W73[H3[7?VHI_7X2E:Ba0[Og0]JeJa4k5PKYJo4Q6M3N3L5L3N1O1O1O1O001O1O[N]LgKb3m3oLoKP3j3\\MRLc2l3dMPL[2m3mMoKR2o3UNlKl1P4\\NlKd1Q4cNkK\\1R4o2M3N21O1O1O1N2O1O1O1O1O1N2O1N2O2M2N2N2O1N2N2N2O1O1O100O1O1O00O1000000002O7H7I8G9H5J4M3M2N3M3N2M3N2M3N2M8I5JO2N1N3N1O2PNXL_Lg3`3WMfKf2X4_MhK^2W4fMiKX2U4lMiKS2V4PNhKP2X4SNeKm1[4VNbKk1]4XN`Kh1`4[N]Ke1c4\\N\\Kd1d4]N[Kd1d4^NZKb1f4_NYKa1g4`NXK`1h4bNVK_1i4bNVK^1l4aNSK_1o4\\22O2M2N2N3N1N2N5K6K3L4L4L4YLWJg1l5jMaJV2Y7O1O1O1O010O00000000000ObLTMbNl2_1WMfLO0k2Z3[M^L05e2^3^MVL39_2b3aMmK8=W2f3fMeK9b0P2k3ZORLf0o3\\OnKd0R4@jK`0W4AgK?Y4DdK<]4F`K9b4I[K7e4KYK5h4MUK3l4T3010O001O1kKoJQ2Q5hMXKW2h4ZLUKe0;o2`4XLSMa3b5H000000000000O10000000001nNfLTJ[3b5QM[JP3[5]MaJc2W5iMeJX2n4WNoJj1f4eNUK\\1f4lNVKU1e4SOWKm0f4m2K4M2N1O4M2_OUIbLo6]3QIaLQ7_3PI]LT7b3lH\\LV7d3kHYLS7l35O1O1O2N1OO2N1N2O1007JN10000O2O000O01O4L0O2O1O1O016KO0000O1M3POU1B9H9L310O010O0010O0001O0mN\\Fb0d9]O]Fc0c9[O_Fe0a9ZO`Ff0`9ZO`Ff0`9YOaFg0_9YO`Fh0`9WOaFi0_9WOaFi0_9WOaFi0^9WOcFi0\\9XOdFh0[9XOfFh0Y9YOhFf0W9[OiFe0V9[OkFd0W9[OjFd0W9ZOjFf0X9XOhFh0Y9WOhFg0m900O2O0O101N100O2O0O5L4K5Lo[S3"}, "label": "a white and brown bull sitting next to a cow"}]}
{"id": 124711, "image": "COCO_train2014_000000124711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich has bacon on it\"."}], "task": "refering", "answer_template": "The \"the sandwich has bacon on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 116, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.023390625, 0.2597280334728034, 0.49457812500000004, 0.8615062761506275], "iscrowd": 0, "area": 62197.96265, "rle": {"size": [478, 640], "counts": "eW72j>3M3M4L3M3M3N2O1N2O0O5L3L4M4K3N0O2O0O101N1O2O0O1mNT1M2M4M2M3N3L3aNhLVGZ3j8hLQG[3P9gLkF[3U9hLeF[3[9hL`FZ3`9iLZFZ3f9iLUFY3k9jLoEY3Q:iLkEY3V:g001O00000001OdF\\K^8d4_G`K`8`4]GeKa8[4[GkKc8U4ZGoKe8Q4WGULg8k3VGZLi8e3TG_Lk8l4N2N7I8H7J`1_N2N1O2N2N2N101N2N2N2N0000O2O00000000001O000O100000001O0000000O101O0000000000001N10000000001O00000O1000001O000000000O101O00000000001O0O10000000001O00000O10001O00000000000O2O0000000000001O0O100000001O0000000O10001O00000000001N1000000000001O000O1000001O000000000O101O0000000000001N10000000001O00000O1000001O0000001O0O101O001O001O001O0O101O001O001N10001O001N101O00001O0O2O001O001N10001O001O0O2O00001O0O2O001O00001N101O001O0O101O001O001N101mJoHW2X:D:F5J6K5K5K5K5K5J6K5K5K4L3L4M4L3M4L3M3L5L3M3Mn\\f4"}, "label": "the sandwich has bacon on it"}]}
{"id": 124711, "image": "COCO_train2014_000000124711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with bacon sticking out on a white plate\"."}], "task": "refering", "answer_template": "The \"a sandwich with bacon sticking out on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 116, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.023390625, 0.2597280334728034, 0.49457812500000004, 0.8615062761506275], "iscrowd": 0, "area": 62197.96265, "rle": {"size": [478, 640], "counts": "eW72j>3M3M4L3M3M3N2O1N2O0O5L3L4M4K3N0O2O0O101N1O2O0O1mNT1M2M4M2M3N3L3aNhLVGZ3j8hLQG[3P9gLkF[3U9hLeF[3[9hL`FZ3`9iLZFZ3f9iLUFY3k9jLoEY3Q:iLkEY3V:g001O00000001OdF\\K^8d4_G`K`8`4]GeKa8[4[GkKc8U4ZGoKe8Q4WGULg8k3VGZLi8e3TG_Lk8l4N2N7I8H7J`1_N2N1O2N2N2N101N2N2N2N0000O2O00000000001O000O100000001O0000000O101O0000000000001N10000000001O00000O1000001O000000000O101O00000000001O0O10000000001O00000O10001O00000000000O2O0000000000001O0O100000001O0000000O10001O00000000001N1000000000001O000O1000001O000000000O101O0000000000001N10000000001O00000O1000001O0000001O0O101O001O001O001O0O101O001O001N10001O001N101O00001O0O2O001O001N10001O001O0O2O00001O0O2O001O00001N101O001O0O101O001O001N101mJoHW2X:D:F5J6K5K5K5K5K5J6K5K5K4L3L4M4L3M4L3M3L5L3M3Mn\\f4"}, "label": "a sandwich with bacon sticking out on a white plate"}]}
{"id": 345897, "image": "COCO_train2014_000000345897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant that is showing you his butt\"."}], "task": "refering", "answer_template": "The \"an elephant that is showing you his butt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342, 359, 360, 361, 362, 363, 364, 365, 382, 383, 384, 385, 386, 387, 388, 405, 406, 408, 409, 410, 411], "bbox": [0.55134375, 0.25, 0.9982031250000001, 0.98873046875], "iscrowd": 0, "area": 82690.71309999996, "rle": {"size": [512, 640], "counts": "kf`5<Y?`0B?D<E:G6J6K5L4VC`Mo;c2fCjMV<T3L3L5L4K4M4K5L6I=D;E1N3N2M2O2M2O2N2M2Oe0ZOh0YO=B4M2N3L4M2M4M2N3L3N3L4Ml1TN0O100O10000O1000000O1000000O10000000001O0000000000000000000000000000000000000000F:B>K5J6J6J6J6K5J6L42N2M3M3M4L3N2M3M3M3N2M3M3M3M3N1N3M2N3M2O2M2N3M2N3N2M2N3M2O2M2N3M2N3N1O2N1O2N1O2N00000000000000000000000000000O100000O1000000000000000000000001O00001O00001O00001O00001O00001N101O00001O00001O00001O00001O00001O001O001O001O001O001N101O001O001O001O001O00001O0WMQLaIo3]6`LVI`3i6nLjHS3T7[M_He2`7iMSHW2k7P3O2N1O1N2O2N1O1O1N3N1O1O1N3N2N2N2M3N1O2M3N2N2N2M3N2N1O2M3N2N2N2M3M3M2O2M3M3M3M3N2M3M2N3M3`KXEd3k:YL\\Ea3f:]L`E\\3c:cLaEW3b:gLdER3^:mLhEl2[:RMjEh2Y:WMlEc2V:[MoE[2X:cMmEP2]:nMgEf1b<C<D=B=Do7"}, "label": "an elephant that is showing you his butt"}]}
{"id": 345897, "image": "COCO_train2014_000000345897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on left side\"."}], "task": "refering", "answer_template": "The \"elephant on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 305, 306, 323, 324, 325, 326, 328, 329, 346, 347, 348, 349, 351, 352, 370, 371, 372, 374, 375, 376, 397, 398], "bbox": [0.01978125, 0.2247265625, 0.523140625, 0.98427734375], "iscrowd": 0, "area": 72597.11335, "rle": {"size": [512, 640], "counts": "kd63l?4L3M3M4VBFg;=VD;R;HkDP1]:TO^Ee1j9^NSFQ2]9SN`FX2U9lMgF_2n8dMnFg2h8\\MUGo2`8UM\\GV3\\1_Kd4^1mI^3AiL`6LlI]5R6gJiI]5T6fJiI\\5U6gJhI[5U6jJgIX5W6Z2N3M2O1O1N2OO1000O1000O10O10000O0100000O10O1000O1000O010J6B>]Ob0^Oc0]Oc0N2O010000O100000O01000000O3Nn0ROm0mHgGe4j9O10000O1000000O1000000O10000001O00000001O01O000000001O00000000001O00000000001O000gMaKQI_4n5eLoI[3Q6hLlIY3R6jLlIV3T6mLiIS3W6PMfIP3Y6TMQITNFh4Y7VMoHZN@`4`7ZMmHZ3R7iLjHX3U7lLhHT3X7nLeHS3[7PMbHP3]7TM`Hm2_7j2O10000O101N100O10001N100O100O2N2N1O2N2N2N2O1N1O2N2N2N2N1O1O1O1O14L>B>RJgHY2h7VM\\Hh2R8gLRHX3\\8XLgGg3h8hK[GV4T9YKRGd4[:J6J001O001O001O001O001O001O001O001OO1N2N2O1N2O1N2O1N2O1N2N2O100000O1O1O00100O1O1000O100PL]NcIc1i5TOTJl0X5KeJ5g4a0WK_OU4X1gKiNV4]1gKdNX4_1eKaNZ4c600O100O10000O100O100O2O0O100O100O10000O100O100O101N100O10000O100O100O100O101N100O1O100O100O1O100O1O2O1N2O1N2N101N2O1N2N2O1N1O2O1N2O1N2N2O1N101N2N2O1N2O1N2N101M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M[jh4"}, "label": "elephant on left side"}]}
{"id": 345897, "image": "COCO_train2014_000000345897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a brown elephant\"."}], "task": "refering", "answer_template": "The \"the front of a brown elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 305, 306, 323, 324, 325, 326, 328, 329, 346, 347, 348, 349, 351, 352, 370, 371, 372, 374, 375, 376, 397, 398], "bbox": [0.01978125, 0.2247265625, 0.523140625, 0.98427734375], "iscrowd": 0, "area": 72597.11335, "rle": {"size": [512, 640], "counts": "kd63l?4L3M3M4VBFg;=VD;R;HkDP1]:TO^Ee1j9^NSFQ2]9SN`FX2U9lMgF_2n8dMnFg2h8\\MUGo2`8UM\\GV3\\1_Kd4^1mI^3AiL`6LlI]5R6gJiI]5T6fJiI\\5U6gJhI[5U6jJgIX5W6Z2N3M2O1O1N2OO1000O1000O10O10000O0100000O10O1000O1000O010J6B>]Ob0^Oc0]Oc0N2O010000O100000O01000000O3Nn0ROm0mHgGe4j9O10000O1000000O1000000O10000001O00000001O01O000000001O00000000001O00000000001O000gMaKQI_4n5eLoI[3Q6hLlIY3R6jLlIV3T6mLiIS3W6PMfIP3Y6TMQITNFh4Y7VMoHZN@`4`7ZMmHZ3R7iLjHX3U7lLhHT3X7nLeHS3[7PMbHP3]7TM`Hm2_7j2O10000O101N100O10001N100O100O2N2N1O2N2N2N2O1N1O2N2N2N2N1O1O1O1O14L>B>RJgHY2h7VM\\Hh2R8gLRHX3\\8XLgGg3h8hK[GV4T9YKRGd4[:J6J001O001O001O001O001O001O001O001OO1N2N2O1N2O1N2O1N2O1N2N2O100000O1O1O00100O1O1000O100PL]NcIc1i5TOTJl0X5KeJ5g4a0WK_OU4X1gKiNV4]1gKdNX4_1eKaNZ4c600O100O10000O100O100O2O0O100O100O10000O100O100O101N100O10000O100O100O100O101N100O1O100O100O1O100O1O2O1N2O1N2N101N2O1N2N2O1N1O2O1N2O1N2N2O1N101N2N2O1N2O1N2N101M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M[jh4"}, "label": "the front of a brown elephant"}]}
{"id": 558890, "image": "COCO_train2014_000000558890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit with a pink tie\"."}], "task": "refering", "answer_template": "The \"a man in a suit with a pink tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.56621875, 0.10666666666666666, 0.9602812499999999, 0.9866666666666667], "iscrowd": 0, "area": 56741.65615000001, "rle": {"size": [426, 640], "counts": "Tof41Q=;hNX1F:J5K4iNW1]Oc0[Oe0gNX101N2O1N2O1RIPJ_6^600O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O100YMiHZOX76ZIHf66]IHd66_IIa65cIH^66eII[65hIIY65jIJV65lIIU67lIHT67nIGS68oIoMfNZ1[7f0QJmMgN[1Y7g0RJkMiN\\1U7g0TJkMkN[1R7i0fJVOZ5h0jJUOW5i0mJUOS5j0SKPOn4n0VKoNk4P1YKlNh4R1]KkNc4S1aKjN`4U1cKhN^4W1eKfN\\4Z1eKdN\\4\\1eKbN\\4^1fK`NZ4_1hK_NY4a1iK\\NX4d1iKZNX4f1f30000000000000000001O0000O100000000O10000000000O100000000O11O00001O00001O001O00001O00001O01O01O002N2N3M2N2N3M2N3jFlMP7V2lHmMS7V2iHmMU7V2hHkMW7W2fHlMX7W2eHjMZ7Y2cHiM[7Y2bHjM\\7]2]HeMa7_2ZHeMc7]2[HfMb7\\2XHiMg7Y2QHnMn7h3O001O1O001O1O001O1O001O00001O00001O0000001O0000001cKcKK]4NmKOS4JXL2h3GbL6_3BlL:T3CRM:n2DXM8h2F\\M8e2F_M7a2GdM6\\2IhM4X2JmM3T2JPN4P2KTN2l1LYN1g1M]N1c1N`N0a1McN1]1OeNO[10hNNX12jNLW12mNKS14POJP16ROHo06TOFn0:SODo0:TOCm0=TOAm0>UO@m0>VO^Ol0b0UO[On0c0UOZOl0f0UOUOP1i0SOQOQ1n0g5O0O2N1O101N1O2N101N1O2N101N1O1O2O001N101O001O0O2O1O1O1O1O1N2O1O1O1Ojc:"}, "label": "a man in a suit with a pink tie"}]}
{"id": 558890, "image": "COCO_train2014_000000558890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a grey suite wearing a pink tie\"."}], "task": "refering", "answer_template": "The \"a man in a grey suite wearing a pink tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.56621875, 0.10666666666666666, 0.9602812499999999, 0.9866666666666667], "iscrowd": 0, "area": 56741.65615000001, "rle": {"size": [426, 640], "counts": "Tof41Q=;hNX1F:J5K4iNW1]Oc0[Oe0gNX101N2O1N2O1RIPJ_6^600O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O100YMiHZOX76ZIHf66]IHd66_IIa65cIH^66eII[65hIIY65jIJV65lIIU67lIHT67nIGS68oIoMfNZ1[7f0QJmMgN[1Y7g0RJkMiN\\1U7g0TJkMkN[1R7i0fJVOZ5h0jJUOW5i0mJUOS5j0SKPOn4n0VKoNk4P1YKlNh4R1]KkNc4S1aKjN`4U1cKhN^4W1eKfN\\4Z1eKdN\\4\\1eKbN\\4^1fK`NZ4_1hK_NY4a1iK\\NX4d1iKZNX4f1f30000000000000000001O0000O100000000O10000000000O100000000O11O00001O00001O001O00001O00001O01O01O002N2N3M2N2N3M2N3jFlMP7V2lHmMS7V2iHmMU7V2hHkMW7W2fHlMX7W2eHjMZ7Y2cHiM[7Y2bHjM\\7]2]HeMa7_2ZHeMc7]2[HfMb7\\2XHiMg7Y2QHnMn7h3O001O1O001O1O001O1O001O00001O00001O0000001O0000001cKcKK]4NmKOS4JXL2h3GbL6_3BlL:T3CRM:n2DXM8h2F\\M8e2F_M7a2GdM6\\2IhM4X2JmM3T2JPN4P2KTN2l1LYN1g1M]N1c1N`N0a1McN1]1OeNO[10hNNX12jNLW12mNKS14POJP16ROHo06TOFn0:SODo0:TOCm0=TOAm0>UO@m0>VO^Ol0b0UO[On0c0UOZOl0f0UOUOP1i0SOQOQ1n0g5O0O2N1O101N1O2N101N1O2N101N1O1O2O001N101O001O0O2O1O1O1O1O1N2O1O1O1Ojc:"}, "label": "a man in a grey suite wearing a pink tie"}]}
{"id": 558890, "image": "COCO_train2014_000000558890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red - haired person with a blue collared shirt\"."}], "task": "refering", "answer_template": "The \"a red - haired person with a blue collared shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.85409375, 0.4741549295774648, 0.9991875, 0.9887558685446011], "iscrowd": 0, "area": 17050.339650000005, "rle": {"size": [426, 640], "counts": "]jS7`1[;b0J5M4K5L3M4K5L3M4K5L4K5K5J7J5J6K6J5L3M2O2M2jHVKj5k4RJXKn5i4nIZKQ6i4kIYKU6h4gI[KX6g4dI\\K\\6e4aI]K_6e4\\I^Kc6d4ZI^Kf6c4WI_Kh6c4VI^Kj6d4RI^Kn6_5O100O100O10000O100O100O100O100O100O100O10000O100O100O10000O10000O1000000O010000000001O0000000000001O00000000O100O1P7"}, "label": "a red - haired person with a blue collared shirt"}]}
{"id": 558890, "image": "COCO_train2014_000000558890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person head bottom right\"."}], "task": "refering", "answer_template": "The \"person head bottom right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.85409375, 0.4741549295774648, 0.9991875, 0.9887558685446011], "iscrowd": 0, "area": 17050.339650000005, "rle": {"size": [426, 640], "counts": "]jS7`1[;b0J5M4K5L3M4K5L3M4K5L4K5K5J7J5J6K6J5L3M2O2M2jHVKj5k4RJXKn5i4nIZKQ6i4kIYKU6h4gI[KX6g4dI\\K\\6e4aI]K_6e4\\I^Kc6d4ZI^Kf6c4WI_Kh6c4VI^Kj6d4RI^Kn6_5O100O100O10000O100O100O100O100O100O100O10000O100O100O10000O10000O1000000O010000000001O0000000000001O00000000O100O1P7"}, "label": "person head bottom right"}]}
{"id": 419627, "image": "COCO_train2014_000000419627.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man who is skateboarding on top of what appears to be a mattress\"."}], "task": "refering", "answer_template": "The \"the man who is skateboarding on top of what appears to be a mattress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 56, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 222, 223], "bbox": [0.47006249999999994, 0.001976470588235294, 0.86403125, 0.6581647058823529], "iscrowd": 0, "area": 20665.88255, "rle": {"size": [425, 640], "counts": "WPm32U=5L3L3N2M3O1O7I0000000000O2O001N1O2O1N1O10O01O00100O010000O01000O100O01000O10O10O100O01000O1000O0100O01000O01000O10O10O10O01000O01]H]O`2c0^M_Ob2a0]M@b2`0^MBa2>^MCa2>jIDR2OT4<cIOT2IX48]I8U2D]45[I<R2Cc41YI`0o1Cg4LXIf0l1Ak4JVIh0m1@m4HSIl0l1_OP5ERIo0l1_OQ5BQIR1i1@V5^OPIT1d1C[5ZOPIT1`1H_5TOPIV1[1Kd5POQIV1W1Mg5nNQIV1V1Oh5kNQIX1T1Oj5jNQIY1R1Ol5iNRIX1P12l5gNSIY1o01n5fNRI[1m01P6eNSIZ1k04P6cNTI[1i04S6j0kI]NZOVOj6_2iI]N_OSOf6b2hI]NCPOd6d2hI^NElNc6g2eI_NJiN_6j2dI_NOeN\\6m2cI`N3aNZ6P3`IaN8^NV6S3_IaN=ZNT6U3]IcN`0WNR6]5UJ]Jj5a5o0N3L4M2M4M2N3L4M2M4M2N3L3N3O1O0O2O001O1N1101N101O1N101O1N101O0SJcKm2_4mLhKQ3Y4hLnKV3S4^L[L`3f3QLjLm3W3[K`Md4a2SKhMm4Y2jJoMV5OmJP2R5oMQKP2o4mMUKR2j4mMYKR2f4nM\\KR2b4mMaKR2^4nMdKQ2[4oMgKP2X4oMkKP2U4oMmKP2R4PNPLo1o3PNTLo1k3QNWLn1i3QNYLS1iNTMn4i1[LQ1iNUMk4j1^Lo0iNVMi4j1`Ln0iNWMf4k1cLm0hNWMd4l1iLh0eN[Ma4l1UMFQNbN;K_4l1_MZOZN`NI9]4m1eMROoNQO[3l1mMiNnNZOT3l1UN]NPOFk2l1^NoMnN5c2k1Q1TNnNk1U1SNkNm1W1RNiNl1Y1SNgNl1\\1SNcNm1_1QNaNn1b1PN_Nn1e1PNZNP2`6N2O1O1O1O100O100O10000O2O00000O10002N1N2O1O1O1N2O2N1O1N2O1N2O7H100O100O1O100O100O100O100O10000O1000O0100O1000O010000O100O0100000000O10000001O1N2O0O2N2NgRT1"}, "label": "the man who is skateboarding on top of what appears to be a mattress"}]}
{"id": 419627, "image": "COCO_train2014_000000419627.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person skateboarding\"."}], "task": "refering", "answer_template": "The \"a person skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 56, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 222, 223], "bbox": [0.47006249999999994, 0.001976470588235294, 0.86403125, 0.6581647058823529], "iscrowd": 0, "area": 20665.88255, "rle": {"size": [425, 640], "counts": "WPm32U=5L3L3N2M3O1O7I0000000000O2O001N1O2O1N1O10O01O00100O010000O01000O100O01000O10O10O100O01000O1000O0100O01000O01000O10O10O10O01000O01]H]O`2c0^M_Ob2a0]M@b2`0^MBa2>^MCa2>jIDR2OT4<cIOT2IX48]I8U2D]45[I<R2Cc41YI`0o1Cg4LXIf0l1Ak4JVIh0m1@m4HSIl0l1_OP5ERIo0l1_OQ5BQIR1i1@V5^OPIT1d1C[5ZOPIT1`1H_5TOPIV1[1Kd5POQIV1W1Mg5nNQIV1V1Oh5kNQIX1T1Oj5jNQIY1R1Ol5iNRIX1P12l5gNSIY1o01n5fNRI[1m01P6eNSIZ1k04P6cNTI[1i04S6j0kI]NZOVOj6_2iI]N_OSOf6b2hI]NCPOd6d2hI^NElNc6g2eI_NJiN_6j2dI_NOeN\\6m2cI`N3aNZ6P3`IaN8^NV6S3_IaN=ZNT6U3]IcN`0WNR6]5UJ]Jj5a5o0N3L4M2M4M2N3L4M2M4M2N3L3N3O1O0O2O001O1N1101N101O1N101O1N101O0SJcKm2_4mLhKQ3Y4hLnKV3S4^L[L`3f3QLjLm3W3[K`Md4a2SKhMm4Y2jJoMV5OmJP2R5oMQKP2o4mMUKR2j4mMYKR2f4nM\\KR2b4mMaKR2^4nMdKQ2[4oMgKP2X4oMkKP2U4oMmKP2R4PNPLo1o3PNTLo1k3QNWLn1i3QNYLS1iNTMn4i1[LQ1iNUMk4j1^Lo0iNVMi4j1`Ln0iNWMf4k1cLm0hNWMd4l1iLh0eN[Ma4l1UMFQNbN;K_4l1_MZOZN`NI9]4m1eMROoNQO[3l1mMiNnNZOT3l1UN]NPOFk2l1^NoMnN5c2k1Q1TNnNk1U1SNkNm1W1RNiNl1Y1SNgNl1\\1SNcNm1_1QNaNn1b1PN_Nn1e1PNZNP2`6N2O1O1O1O100O100O10000O2O00000O10002N1N2O1O1O1N2O2N1O1N2O1N2O7H100O100O1O100O100O100O100O10000O1000O0100O1000O010000O100O0100000000O10000001O1N2O0O2N2NgRT1"}, "label": "a person skateboarding"}]}
{"id": 509740, "image": "COCO_train2014_000000509740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced food close to us next to a pie\"."}], "task": "refering", "answer_template": "The \"sliced food close to us next to a pie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296], "bbox": [0.72809375, 0.7431388888888888, 0.9998593750000001, 0.9903333333333333], "iscrowd": 0, "area": 9551.4132, "rle": {"size": [360, 640], "counts": "ZTT56Q;2M2O2N2N2O1N2O1N2O1N2O0O2O1N101O1N2O0O2O1N101N2N101N2O1O1N2O1N2O1O1N2O0O101N100O2N100O2O0O101N100O100O1O2O0O100O10001N10000O10001NOVGfMj8Z22O10O01000O10000O10000O1000O010000O1000O01000O010O01000O01000O010O10000O100O100O100O100O100O2O0O101N100O2O0O101N101O0O2O1N101O0O2O000O2O0O10000O2O000O101O0O10000O2O0O10001N101N101N100O2O0O2O0O2O0O101N102L4M4L7I7I^F"}, "label": "sliced food close to us next to a pie"}]}
{"id": 509740, "image": "COCO_train2014_000000509740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana bread\"."}], "task": "refering", "answer_template": "The \"the banana bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296], "bbox": [0.72809375, 0.7431388888888888, 0.9998593750000001, 0.9903333333333333], "iscrowd": 0, "area": 9551.4132, "rle": {"size": [360, 640], "counts": "ZTT56Q;2M2O2N2N2O1N2O1N2O1N2O0O2O1N101O1N2O0O2O1N101N2N101N2O1O1N2O1N2O1O1N2O0O101N100O2N100O2O0O101N100O100O1O2O0O100O10001N10000O10001NOVGfMj8Z22O10O01000O10000O10000O1000O010000O1000O01000O010O01000O01000O010O10000O100O100O100O100O100O2O0O101N100O2O0O101N101O0O2O1N101O0O2O000O2O0O10000O2O000O101O0O10000O2O0O10001N101N101N100O2O0O2O0O2O0O101N102L4M4L7I7I^F"}, "label": "the banana bread"}]}
{"id": 296747, "image": "COCO_train2014_000000296747.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in tan shirt and gray shorts\"."}], "task": "refering", "answer_template": "The \"man in tan shirt and gray shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 101, 102, 103, 124, 125, 126, 147, 148, 149, 193, 194, 195, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 332, 333, 334, 353, 354, 356, 357], "bbox": [0.32390625, 0.12101952277657267, 0.5317812500000001, 0.9890238611713665], "iscrowd": 0, "area": 20681.100150000002, "rle": {"size": [461, 640], "counts": "fXm21Z>:F3L5L2O1N2N2O1N2O1N2N2O100O100O100O100O100gLE`H;Z75\\HK`7?ZHAc7b0]H^Oh6]1XIcNh6]1XIcNh6]1YIbNg6^1YIcNf6]1ZIcNg6\\1ZIcNf6]1[IbNf6]1ZIcNa5b2`J]M]5f2cJZMZ5i2gJVMU5n2kJRMk4X3VKgLg4\\3ZKcLb4a3_K^L[4h3gKVLS4P4nKoKo3T4RLkKm3V4TLiKj3Y4WLfKg3S4cLlK[3S4iLlKT3f2^JdMc2El2g2gJ]Mc2Jc2h2oNVMn0i2XOUMe0k2@SM=m2HQM5P33jLIW3?dL_OP3T6\\Od02`MaMPIa2o6`MoHc2n6`MoHd2n6]MPIh2l6ZMQIj2l6WMRIm2k6UMTIn2S5iL`J:<Q3P5YMoIIo0Q3P5YMmIGR1T3n4XMlIFU1U3l4XMkIDX1X3j4WMiID[1Y3i4VNWKm1f4TNXKo1f4QNYKR2f4mMYKZ2a4gM^KZ2a4fM^K[2b4eM]K]2b4dM\\K]2d4cM\\K^2d4aM[K`2e4`MZKb2e4_MZKa2g4^MXKd2h4[MWKj2e4WMZKj2f4UMZKm2e4RM\\Kn2d4RM[Ko2e4PM[KR3c4nL]KS3c4lL]KX3`4hL_K`3Z4_LgKj3P4ULPLX4c3iK\\L[4a3dK_L`4^3_KbLc4]3]KcLd4\\3[KdLg4Z3YKfLi4Y3VKgLl4Y3SKfLo4Z3oJgLR5Y3mJfLS5\\3kJdLU5^3jJaLV5a3iJ^LW5c3iJ]LV5e3hJ[LX5g3gJXLY5j3fJULZ5m3eJRL[5Q4bJPL^5Q4aJnK_5U4_JjKi1TO`0U5eMfKj1[O<R5gMcKl1AmMBZ1\\5lNaKk1GjM@[1Z5hNaJYOm0i2NfM^O\\1X5hNbJ\\Oj0h26aMZO`1V5gNdJ_Oe0h2=]MXOa1T5gNfJB`0g2e0ZMTOc1S5hNcKY2S1mN\\3gNcKX2V1oNY3fNbKX2Z1POU3fNbKX2]1QOS3cNaK09=W1o0`6nNcIP1P:]OPCCU=0YCJh=MgcV4"}, "label": "man in tan shirt and gray shorts"}]}
{"id": 296747, "image": "COCO_train2014_000000296747.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in grey shorts holding some type of cane\"."}], "task": "refering", "answer_template": "The \"a man in grey shorts holding some type of cane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 101, 102, 103, 124, 125, 126, 147, 148, 149, 193, 194, 195, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 332, 333, 334, 353, 354, 356, 357], "bbox": [0.32390625, 0.12101952277657267, 0.5317812500000001, 0.9890238611713665], "iscrowd": 0, "area": 20681.100150000002, "rle": {"size": [461, 640], "counts": "fXm21Z>:F3L5L2O1N2N2O1N2O1N2N2O100O100O100O100O100gLE`H;Z75\\HK`7?ZHAc7b0]H^Oh6]1XIcNh6]1XIcNh6]1YIbNg6^1YIcNf6]1ZIcNg6\\1ZIcNf6]1[IbNf6]1ZIcNa5b2`J]M]5f2cJZMZ5i2gJVMU5n2kJRMk4X3VKgLg4\\3ZKcLb4a3_K^L[4h3gKVLS4P4nKoKo3T4RLkKm3V4TLiKj3Y4WLfKg3S4cLlK[3S4iLlKT3f2^JdMc2El2g2gJ]Mc2Jc2h2oNVMn0i2XOUMe0k2@SM=m2HQM5P33jLIW3?dL_OP3T6\\Od02`MaMPIa2o6`MoHc2n6`MoHd2n6]MPIh2l6ZMQIj2l6WMRIm2k6UMTIn2S5iL`J:<Q3P5YMoIIo0Q3P5YMmIGR1T3n4XMlIFU1U3l4XMkIDX1X3j4WMiID[1Y3i4VNWKm1f4TNXKo1f4QNYKR2f4mMYKZ2a4gM^KZ2a4fM^K[2b4eM]K]2b4dM\\K]2d4cM\\K^2d4aM[K`2e4`MZKb2e4_MZKa2g4^MXKd2h4[MWKj2e4WMZKj2f4UMZKm2e4RM\\Kn2d4RM[Ko2e4PM[KR3c4nL]KS3c4lL]KX3`4hL_K`3Z4_LgKj3P4ULPLX4c3iK\\L[4a3dK_L`4^3_KbLc4]3]KcLd4\\3[KdLg4Z3YKfLi4Y3VKgLl4Y3SKfLo4Z3oJgLR5Y3mJfLS5\\3kJdLU5^3jJaLV5a3iJ^LW5c3iJ]LV5e3hJ[LX5g3gJXLY5j3fJULZ5m3eJRL[5Q4bJPL^5Q4aJnK_5U4_JjKi1TO`0U5eMfKj1[O<R5gMcKl1AmMBZ1\\5lNaKk1GjM@[1Z5hNaJYOm0i2NfM^O\\1X5hNbJ\\Oj0h26aMZO`1V5gNdJ_Oe0h2=]MXOa1T5gNfJB`0g2e0ZMTOc1S5hNcKY2S1mN\\3gNcKX2V1oNY3fNbKX2Z1POU3fNbKX2]1QOS3cNaK09=W1o0`6nNcIP1P:]OPCCU=0YCJh=MgcV4"}, "label": "a man in grey shorts holding some type of cane"}]}
{"id": 337704, "image": "COCO_train2014_000000337704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smallest male wearing white shirt and white and black cap\"."}], "task": "refering", "answer_template": "The \"smallest male wearing white shirt and white and black cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 304, 305, 306, 307, 308], "bbox": [0.24681250000000002, 0.21347417840375588, 0.451734375, 0.9101173708920187], "iscrowd": 0, "area": 19857.26645, "rle": {"size": [426, 640], "counts": "VSR21X=1N3RL8WJIg5:WJGg5<VJGg5<WJEg5>VJDg5`0WJAg5b0WJ@f5c0WJ_Og5d0WJ]Og5e0XJ\\Of5g0WJ[Og5h0WJZOe5j0VJZOh5i0SJ[Oh5k0TJXOh5m0SJWOh5o0TJTOh5Q1SJTOg5S1SJQOi5V1QJmNj5[1oIiNm5_1lIdN[5X2_JkM_5Z2\\JiMb5[2YJgMf5]2UJeMj5`2PJbMn5n4O\\OQJoIn5Q6TJoIi5R6ZJlIe5S6^JlIa5T6bJjI\\5V6gJiIX5V6kJiIT5V6PKiIm4V6WKiIh4V6\\KhIc4W6`KhI^4X6fKfIZ4X6iKgIW4W6mKhIR4V6QLiIo3U6d1N]OQJoIm5R6YJjId5V6dJdIZ5]6lJ^IS5b6j000O10O1000O10O101O000001O3N2M3M3M4L3M`1aN9F1O1001O1O0VKSL7n3@_L;b3[OkLa0U3XOWMd0i2\\O[M?e2B_M9b2GaM5`2KbM2^2OdMN]22eMK[26gMGZ29hMDY2<iMAW2]NaKd0Z2l0V2^NaKg0[2h0U2^NcKj0]2b0P2cNdKl0b2:S2`N]KW1e22[2YNQKf1j2oNQMBa5[OfJT2o2mMfMb0R5b1]N[Nd1e1^NXNc1h1]NWNc1j1]NVNc1j1\\NVNd1k1\\NTNe1Q2UNoMk1V2QNiMP2[2lMdMT2b2fM^M[2n0lI<e3eN`2c0XJl0R3aNf27fJY1b2_Ni2LRKc1U2aN[4^1eKaN]4]1dKaN_4^1`KbNb4\\1_KcNc4\\1\\KdNe4[1\\KdNf4[1ZKdNh4Z1XKfNj4Y1VKfNl4X1TKhNn4W1RKhNR5T1nJlNV5Q1e3L4M3K5J6J6Jdha4"}, "label": "smallest male wearing white shirt and white and black cap"}]}
{"id": 337704, "image": "COCO_train2014_000000337704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in a white shirt and white trucker hat who has just tossed a flying disc\"."}], "task": "refering", "answer_template": "The \"a young man in a white shirt and white trucker hat who has just tossed a flying disc\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 304, 305, 306, 307, 308], "bbox": [0.24681250000000002, 0.21347417840375588, 0.451734375, 0.9101173708920187], "iscrowd": 0, "area": 19857.26645, "rle": {"size": [426, 640], "counts": "VSR21X=1N3RL8WJIg5:WJGg5<VJGg5<WJEg5>VJDg5`0WJAg5b0WJ@f5c0WJ_Og5d0WJ]Og5e0XJ\\Of5g0WJ[Og5h0WJZOe5j0VJZOh5i0SJ[Oh5k0TJXOh5m0SJWOh5o0TJTOh5Q1SJTOg5S1SJQOi5V1QJmNj5[1oIiNm5_1lIdN[5X2_JkM_5Z2\\JiMb5[2YJgMf5]2UJeMj5`2PJbMn5n4O\\OQJoIn5Q6TJoIi5R6ZJlIe5S6^JlIa5T6bJjI\\5V6gJiIX5V6kJiIT5V6PKiIm4V6WKiIh4V6\\KhIc4W6`KhI^4X6fKfIZ4X6iKgIW4W6mKhIR4V6QLiIo3U6d1N]OQJoIm5R6YJjId5V6dJdIZ5]6lJ^IS5b6j000O10O1000O10O101O000001O3N2M3M3M4L3M`1aN9F1O1001O1O0VKSL7n3@_L;b3[OkLa0U3XOWMd0i2\\O[M?e2B_M9b2GaM5`2KbM2^2OdMN]22eMK[26gMGZ29hMDY2<iMAW2]NaKd0Z2l0V2^NaKg0[2h0U2^NcKj0]2b0P2cNdKl0b2:S2`N]KW1e22[2YNQKf1j2oNQMBa5[OfJT2o2mMfMb0R5b1]N[Nd1e1^NXNc1h1]NWNc1j1]NVNc1j1\\NVNd1k1\\NTNe1Q2UNoMk1V2QNiMP2[2lMdMT2b2fM^M[2n0lI<e3eN`2c0XJl0R3aNf27fJY1b2_Ni2LRKc1U2aN[4^1eKaN]4]1dKaN_4^1`KbNb4\\1_KcNc4\\1\\KdNe4[1\\KdNf4[1ZKdNh4Z1XKfNj4Y1VKfNl4X1TKhNn4W1RKhNR5T1nJlNV5Q1e3L4M3K5J6J6Jdha4"}, "label": "a young man in a white shirt and white trucker hat who has just tossed a flying disc"}]}
{"id": 337704, "image": "COCO_train2014_000000337704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male wearing an orange shirt and black shorts running a game of frisbee\"."}], "task": "refering", "answer_template": "The \"a male wearing an orange shirt and black shorts running a game of frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 61, 62, 63, 84, 85, 86, 87, 108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 226, 227, 228, 229, 245, 246, 250, 251, 252, 268, 269, 274, 275, 291, 292], "bbox": [0.65215625, 0.11460093896713615, 0.9797343749999999, 0.8314553990610328], "iscrowd": 0, "area": 19496.137049999994, "rle": {"size": [426, 640], "counts": "ZW^51X=1O2O0O1O1O2O0hGJ[37cLMZ35dLNY33eL1Y3OfL4]NDf19K7WNFm15J8QNIT20I;kMJ]2KG>jMG^2LFa0jMB`2OCc0jM_Oc2NBe0jM\\OKIa07i1f0jMZOK74JV2g0jMWOKf0H^Ob2f0jMVOJU1]OoNo2h0hMTOJc1RObN[3i0hMQOJS2eNUNh3h0hMPOIX3_2iMfMoNJZ3_2iMfMmNJ[3`2jMdMjNK_3`2hMdMiNJa3b2gMcMgNJd3b2gMcMeNIf3d2gMaMcNIi3e2eMaMaNIl3e2eMaM_NIm3f2eM_M^NJP4f2cM_M]NIR4g2dM^MYNJU4h2dM\\MWNJW4j2dMZMTNK[4k2bMXMSNLW1WO:5jNb3b1TMSNLX1_ON9QOZ3f1PMRNMZ1FB>ROV3Q2eLRNO[1MUOP5^1RKQN0]14iNR5i1hJQN0^1<]NT5T2^JPN0a1n6?PIPN0c1R7=lHoM1e1V7<gHoM0g1[7;cHnM0h1`7:^HmM0l1c78ZHmM1l1g78VHkM2n1k77QHkM1P2P86mGiM2R2S8j22M2N1O2O1N1O2N2O0O2N2N101N2N1O2O1N1O2001N3N1O2M2O1O0O01O1O001O10O01O1O2N2N2N2N2N2N2N2N2N2O1O100O1O100O1O101N1O100O1O100O1O1O1O1N2O1N5L6I7J4K4M3L5L3L4M4K4M3L5L:Ec0^OO001O001O1O1O00100O1O1O00100O1O1O010O1O1O1O010O1O1O100O0100000O10O1000O1000O10O10000O1000000O1000000O100O1N2O1N2N2O1N2N2O04L5K5K5K5K5K5K5K5K5JnR5"}, "label": "a male wearing an orange shirt and black shorts running a game of frisbee"}]}
{"id": 337704, "image": "COCO_train2014_000000337704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt is running towards the other players\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt is running towards the other players\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 61, 62, 63, 84, 85, 86, 87, 108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 226, 227, 228, 229, 245, 246, 250, 251, 252, 268, 269, 274, 275, 291, 292], "bbox": [0.65215625, 0.11460093896713615, 0.9797343749999999, 0.8314553990610328], "iscrowd": 0, "area": 19496.137049999994, "rle": {"size": [426, 640], "counts": "ZW^51X=1O2O0O1O1O2O0hGJ[37cLMZ35dLNY33eL1Y3OfL4]NDf19K7WNFm15J8QNIT20I;kMJ]2KG>jMG^2LFa0jMB`2OCc0jM_Oc2NBe0jM\\OKIa07i1f0jMZOK74JV2g0jMWOKf0H^Ob2f0jMVOJU1]OoNo2h0hMTOJc1RObN[3i0hMQOJS2eNUNh3h0hMPOIX3_2iMfMoNJZ3_2iMfMmNJ[3`2jMdMjNK_3`2hMdMiNJa3b2gMcMgNJd3b2gMcMeNIf3d2gMaMcNIi3e2eMaMaNIl3e2eMaM_NIm3f2eM_M^NJP4f2cM_M]NIR4g2dM^MYNJU4h2dM\\MWNJW4j2dMZMTNK[4k2bMXMSNLW1WO:5jNb3b1TMSNLX1_ON9QOZ3f1PMRNMZ1FB>ROV3Q2eLRNO[1MUOP5^1RKQN0]14iNR5i1hJQN0^1<]NT5T2^JPN0a1n6?PIPN0c1R7=lHoM1e1V7<gHoM0g1[7;cHnM0h1`7:^HmM0l1c78ZHmM1l1g78VHkM2n1k77QHkM1P2P86mGiM2R2S8j22M2N1O2O1N1O2N2O0O2N2N101N2N1O2O1N1O2001N3N1O2M2O1O0O01O1O001O10O01O1O2N2N2N2N2N2N2N2N2N2O1O100O1O100O1O101N1O100O1O100O1O1O1O1N2O1N5L6I7J4K4M3L5L3L4M4K4M3L5L:Ec0^OO001O001O1O1O00100O1O1O00100O1O1O010O1O1O1O010O1O1O100O0100000O10O1000O1000O10O10000O1000000O1000000O100O1N2O1N2N2O1N2N2O04L5K5K5K5K5K5K5K5K5JnR5"}, "label": "a man in an orange shirt is running towards the other players"}]}
{"id": 337704, "image": "COCO_train2014_000000337704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player in orange between two players in white\"."}], "task": "refering", "answer_template": "The \"player in orange between two players in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 97, 98, 99, 119, 120, 121, 140, 141, 142, 143, 144, 147, 163, 164, 166, 167, 170, 171, 188, 189, 190, 192, 194, 211, 212, 215, 216, 233, 234, 235, 239, 256, 257, 262, 279, 285, 286, 302, 309], "bbox": [0.085671875, 0.13657276995305165, 0.47943749999999996, 0.8984507042253521], "iscrowd": 0, "area": 14482.135200000012, "rle": {"size": [426, 640], "counts": "lRg01X=001D0]C1c<0[C10M]<2bC11M^<3`C00Na<2^C10Mb<4\\C00Nd<2[C01Ne<>1O`C@X<`0eCA\\<?aCD`<b01O1O001N2:FN101N10UHYOU3f0jL\\OV3b0jL@U3`0kLAU3=kLET3;lLFT39lLHS38mLIR38mLIR37nLKP35QMKn25RMLm24SMMl23TMNk22UMOj21VM0i21UM1j2OTM4k2LTM6k2JTM8k2HSM;l2FQM=n2CPM`0o2@mLe0R3[OgLm0X3SObLT1]3mN[L[1d3eNVLb1i3^NULe1j3[NTLh1k3YNRLj1m3VNQLm1n3SNQL\\1SM^Nj67RLY1^MYN`6>QLX1g4iNWKV1k4jNTKU1n4kNQKU1P5kNoJU1S5kNkJU1V5kNiJU1X5kNgJU1Z5kNeJT1^5kNaJU1`5lN^JS1d5mN[JS1g5lNXJS1i5nNVJR1k5nNTJQ1n5ZNjG=X2V1Q6]NjG:T2U1W6aNfG9S2R1[6eNdG7Q2o0`6kN_G5P2o0b6mN_G2P2o0c6PO^GOn1Q1e6QO]GMn1T1c6PO`GJl1X1c6nNbGIk1[1a6mNeGFi1_1_6nNoIS1n5POPJS1m5oNRJR1k5POTJS1h5POWJQ1f5ROXJP1g5QOXJQ1f5POXJR1g5oNWJT1f5nNYJS1f5nNXJU1f5lNYJU1f5lNXJV1h5jNWJX1g5iNWJY1i5gNVJ[1h5eNWJ]1i5cNVJ^1i5cNUJ`1j5`NTJb1k5_NRJb1P6^NlId1U6]NhId1Z6\\NcIe1^6\\N^Ij0kNgNh7`0dHQOMe10fN`7d0cHW13RN[7h0bHT17PNY7l0`HS19nMX7P1_HP1=lMV7T1]Hn0`0kMT7X1\\Hl0b0iMS7\\1ZHj0j8VOVGh0P8QNWHX1Ie0P8ZNPHS10a0P8bNhGQ18:P8b0PH\\OP8g0oGWOP8l0PHROo7Q1PHkNS8W1lGcN[8]1eG\\Nb8d1]GVNi8l1UGfMZ9[2dF_Mc9b2[FYMl9g2RFXMP:o23M2M3N3M2O2N1O2N101N1O2N1O2N101N101cNQE=R;^OPEb0T;YOnDe0X;UOjDj0k;01O0O2UOoC;Q<^OXD`0i;_OYD>h;A`D8a;GQ10Yc73c\\H4L5LD[C7m<1N2O2M<D2N2N2N2N2N2O1O1O1N2O2N1O1O1N8IO1O1N2O0O2N2O0OkLgNYJY1b5lN]JU1]5QObJo0]5TOaJm0_5SOaJm0`5RO_Jn0b5RO]Jo0d5PO\\Jo0f5POZJP1f5PO]Jn0c5RO^Jl0c5SO_Jk0a5UOaJj0\\5XOgJe0X5\\OjJb0T5@lJa0Q5AoJ>P5BRK=n4CSK;m4]OjGI[3h0j4@PHCW3k0j4B`K=_4DbK:^4GcK8\\4HgK5Z4KgK4X4MiK1X4LjGCo3`0V4ClG4OIo3?V4CnG4P48R4CQH4m39^4DbK<`4C`K<h8O1O0O2O2N2N3M2NUUZ4"}, "label": "player in orange between two players in white"}]}
{"id": 337704, "image": "COCO_train2014_000000337704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3 people on left\"."}], "task": "refering", "answer_template": "The \"3 people on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 97, 98, 99, 119, 120, 121, 140, 141, 142, 143, 144, 147, 163, 164, 166, 167, 170, 171, 188, 189, 190, 192, 194, 211, 212, 215, 216, 233, 234, 235, 239, 256, 257, 262, 279, 285, 286, 302, 309], "bbox": [0.085671875, 0.13657276995305165, 0.47943749999999996, 0.8984507042253521], "iscrowd": 0, "area": 14482.135200000012, "rle": {"size": [426, 640], "counts": "lRg01X=001D0]C1c<0[C10M]<2bC11M^<3`C00Na<2^C10Mb<4\\C00Nd<2[C01Ne<>1O`C@X<`0eCA\\<?aCD`<b01O1O001N2:FN101N10UHYOU3f0jL\\OV3b0jL@U3`0kLAU3=kLET3;lLFT39lLHS38mLIR38mLIR37nLKP35QMKn25RMLm24SMMl23TMNk22UMOj21VM0i21UM1j2OTM4k2LTM6k2JTM8k2HSM;l2FQM=n2CPM`0o2@mLe0R3[OgLm0X3SObLT1]3mN[L[1d3eNVLb1i3^NULe1j3[NTLh1k3YNRLj1m3VNQLm1n3SNQL\\1SM^Nj67RLY1^MYN`6>QLX1g4iNWKV1k4jNTKU1n4kNQKU1P5kNoJU1S5kNkJU1V5kNiJU1X5kNgJU1Z5kNeJT1^5kNaJU1`5lN^JS1d5mN[JS1g5lNXJS1i5nNVJR1k5nNTJQ1n5ZNjG=X2V1Q6]NjG:T2U1W6aNfG9S2R1[6eNdG7Q2o0`6kN_G5P2o0b6mN_G2P2o0c6PO^GOn1Q1e6QO]GMn1T1c6PO`GJl1X1c6nNbGIk1[1a6mNeGFi1_1_6nNoIS1n5POPJS1m5oNRJR1k5POTJS1h5POWJQ1f5ROXJP1g5QOXJQ1f5POXJR1g5oNWJT1f5nNYJS1f5nNXJU1f5lNYJU1f5lNXJV1h5jNWJX1g5iNWJY1i5gNVJ[1h5eNWJ]1i5cNVJ^1i5cNUJ`1j5`NTJb1k5_NRJb1P6^NlId1U6]NhId1Z6\\NcIe1^6\\N^Ij0kNgNh7`0dHQOMe10fN`7d0cHW13RN[7h0bHT17PNY7l0`HS19nMX7P1_HP1=lMV7T1]Hn0`0kMT7X1\\Hl0b0iMS7\\1ZHj0j8VOVGh0P8QNWHX1Ie0P8ZNPHS10a0P8bNhGQ18:P8b0PH\\OP8g0oGWOP8l0PHROo7Q1PHkNS8W1lGcN[8]1eG\\Nb8d1]GVNi8l1UGfMZ9[2dF_Mc9b2[FYMl9g2RFXMP:o23M2M3N3M2O2N1O2N101N1O2N1O2N101N101cNQE=R;^OPEb0T;YOnDe0X;UOjDj0k;01O0O2UOoC;Q<^OXD`0i;_OYD>h;A`D8a;GQ10Yc73c\\H4L5LD[C7m<1N2O2M<D2N2N2N2N2N2O1O1O1N2O2N1O1O1N8IO1O1N2O0O2N2O0OkLgNYJY1b5lN]JU1]5QObJo0]5TOaJm0_5SOaJm0`5RO_Jn0b5RO]Jo0d5PO\\Jo0f5POZJP1f5PO]Jn0c5RO^Jl0c5SO_Jk0a5UOaJj0\\5XOgJe0X5\\OjJb0T5@lJa0Q5AoJ>P5BRK=n4CSK;m4]OjGI[3h0j4@PHCW3k0j4B`K=_4DbK:^4GcK8\\4HgK5Z4KgK4X4MiK1X4LjGCo3`0V4ClG4OIo3?V4CnG4P48R4CQH4m39^4DbK<`4C`K<h8O1O0O2O2N2N3M2NUUZ4"}, "label": "3 people on left"}]}
{"id": 411441, "image": "COCO_train2014_000000411441.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white ship in the harbor\"."}], "task": "refering", "answer_template": "The \"a blue and white ship in the harbor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174], "bbox": [0.47757812499999996, 0.2659166666666667, 0.9585156250000001, 0.45875], "iscrowd": 0, "area": 6978.20045, "rle": {"size": [480, 640], "counts": "Qd_41o>001O0O2O001O001O1O0O2O001O001_A1o=`000O10000000000O1000000N2N2N2M3N2N3M2M3NoNiB=m=0O1000000O10000O1000000O10000O10000O1000000O10000J6O1O1O1O1O1O1O1O100O1O1O1O0ON3N2N2N1O2N2N20O2O1O1O1N2O1O1O0O2O1N2O1N2O1O1N10jNRC<l<^O\\Cb0c<WOeCi0Y<QOoCo0Q=0000O101O00000000000O1000000000000O10001O000000000O1000000000001O0O1000000000001O000000000000001O00000000000000001O0000000000000000000000002N2N3M3M3M2N00O100O1O100O100O100O1O100O10O01O100O100001O1O1O1O1O1O1O1O00000000000O100000000000000000000000000000000000000O10000000000000000000000000000000000000000O01O100O1O100000000O100000000000000000000O10O100000000000000000O10000000000000000000N101O1N2O1O1O1N2O1Og_<"}, "label": "a blue and white ship in the harbor"}]}
{"id": 411441, "image": "COCO_train2014_000000411441.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ship is anchored at the port\"."}], "task": "refering", "answer_template": "The \"the ship is anchored at the port\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174], "bbox": [0.47757812499999996, 0.2659166666666667, 0.9585156250000001, 0.45875], "iscrowd": 0, "area": 6978.20045, "rle": {"size": [480, 640], "counts": "Qd_41o>001O0O2O001O001O1O0O2O001O001_A1o=`000O10000000000O1000000N2N2N2M3N2N3M2M3NoNiB=m=0O1000000O10000O1000000O10000O10000O1000000O10000J6O1O1O1O1O1O1O1O100O1O1O1O0ON3N2N2N1O2N2N20O2O1O1O1N2O1O1O0O2O1N2O1N2O1O1N10jNRC<l<^O\\Cb0c<WOeCi0Y<QOoCo0Q=0000O101O00000000000O1000000000000O10001O000000000O1000000000001O0O1000000000001O000000000000001O00000000000000001O0000000000000000000000002N2N3M3M3M2N00O100O1O100O100O100O1O100O10O01O100O100001O1O1O1O1O1O1O1O00000000000O100000000000000000000000000000000000000O10000000000000000000000000000000000000000O01O100O1O100000000O100000000000000000000O10O100000000000000000O10000000000000000000N101O1N2O1O1O1N2O1Og_<"}, "label": "the ship is anchored at the port"}]}
{"id": 157491, "image": "COCO_train2014_000000157491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bench\"."}], "task": "refering", "answer_template": "The \"the white bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 42, 60, 61, 65, 66, 83, 84, 85, 86, 87, 88, 107, 110, 111], "bbox": [0.640453125, 0.07864583333333333, 0.87303125, 0.29662499999999997], "iscrowd": 0, "area": 4692.553100000001, "rle": {"size": [480, 640], "counts": "gXP63i>4M4K4M3L5L3L4L5I6JH`B]OY=h0hBXOV=j0jBVOV=i0lBWOS=h0nBXOR=g0oBYOQ=g0oBZOP=e0RCZOo<d0RC]Ob0JW;h0XDb0h;^OXDc0g;\\OZDd0f;[O\\Dd0d;[O]Df0b;ZO^Df0b;YO_Dh0`;WObDh0_;VObDk0];UOcDk0];TOdDl0\\;SOfDm0Y;QOiDG_Oe0h;@mDKM4U;NQENX<1iCOW<OlC0U<NlC2T<MmC3S<LnC4R<LnC4R<LnC4R<MnC2R<NnC2R<NnC2R<OmC1S<OmC1T<NlC2T<OkC1U<OkC1U<OkC1U<0jC0V<0jC0V<0jC0V<1iCOW<1iCOX<0hC0X<1gCOY<1gCOY<0hC0X<OiC1W<NjC2V<MkC3U<LlC4T<LlC4U<JlC6T<ImC7S<HnC8R<GoC9Q<FPD:P<FPD:[=000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000]1cN000000000000000000O100eNjBk0W=UOPCd0P=\\OXC;g<G`C2X<6oCBi;g0P18H9G8O10000000000000000000000000YOg0WO^nU1"}, "label": "the white bench"}]}
{"id": 157491, "image": "COCO_train2014_000000157491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bench with black and white bag on it\"."}], "task": "refering", "answer_template": "The \"white bench with black and white bag on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 42, 60, 61, 65, 66, 83, 84, 85, 86, 87, 88, 107, 110, 111], "bbox": [0.640453125, 0.07864583333333333, 0.87303125, 0.29662499999999997], "iscrowd": 0, "area": 4692.553100000001, "rle": {"size": [480, 640], "counts": "gXP63i>4M4K4M3L5L3L4L5I6JH`B]OY=h0hBXOV=j0jBVOV=i0lBWOS=h0nBXOR=g0oBYOQ=g0oBZOP=e0RCZOo<d0RC]Ob0JW;h0XDb0h;^OXDc0g;\\OZDd0f;[O\\Dd0d;[O]Df0b;ZO^Df0b;YO_Dh0`;WObDh0_;VObDk0];UOcDk0];TOdDl0\\;SOfDm0Y;QOiDG_Oe0h;@mDKM4U;NQENX<1iCOW<OlC0U<NlC2T<MmC3S<LnC4R<LnC4R<LnC4R<MnC2R<NnC2R<NnC2R<OmC1S<OmC1T<NlC2T<OkC1U<OkC1U<OkC1U<0jC0V<0jC0V<0jC0V<1iCOW<1iCOX<0hC0X<1gCOY<1gCOY<0hC0X<OiC1W<NjC2V<MkC3U<LlC4T<LlC4U<JlC6T<ImC7S<HnC8R<GoC9Q<FPD:P<FPD:[=000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000]1cN000000000000000000O100eNjBk0W=UOPCd0P=\\OXC;g<G`C2X<6oCBi;g0P18H9G8O10000000000000000000000000YOg0WO^nU1"}, "label": "white bench with black and white bag on it"}]}
{"id": 329528, "image": "COCO_train2014_000000329528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dog is sitting on sofa between two dogs\"."}], "task": "refering", "answer_template": "The \"a black dog is sitting on sofa between two dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 122, 123, 124, 125, 145, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 238, 239, 240, 261, 262, 263, 283, 284, 306, 307], "bbox": [0.31060937499999997, 0.27441666666666664, 0.52878125, 0.7895625000000001], "iscrowd": 0, "area": 15425.572300000002, "rle": {"size": [480, 640], "counts": "Z]m21o>0O2N101N101N1O2O0O2O0010O010O010OgBJh;7jC6_4_OZ2:VI8[4E]24VI9V4Jc2MVI:S4Ng2GUI<o34j2AVI<j39P3ZOUI>g3=S3VOTI>d3a0Y3PORI`0a3c0^3nNoHa0]3e0f3iNlHe0W3f0n3fNiHg0R3g0W4aNfHj0m2i0^4^NdHk0g2k0f4ZNbHP1_2i0R5WN^HU1?dNh0S2]6TN\\HY14hNm0m1g6QNXH_1GmNS1e1Q7PNTHX2e0J[7mMPHZ2>Le7jMmG\\26MP8hMiG\\20O[8dMeG[5]8eJcGW5b8iJ\\GR5j8nJVGl4Q9SKnFh4X9YKfFb4a9]K^Fa4e9:3L5L3M4K4M4L3L4M4L3M4N]FTLT8k3kGZLS8c3lGcLQ8\\3nGiLo7U3PHoLP8o2oGTMP8k2oGWMQ8i2mGYMR8h2lGZMT8f2jG]MU8c2iG_MV8b2hG`MX8a2fG`MZ8`2dGcM[8]2cGeM\\8\\2bGfM^8Z2`GgMa8Z2]GgMc8Y2\\GgMd8Z2ZGgMg8Y2XGhMh8X2VGiMl8V2SGjMn8W2PGjMP9V2nFSMk9m2TFRMn9n2QFRMQ:m2mETMT:l2kETMV:i32O1O2N1N2O1O1O2M2O1O2N2M3N2N2N2N2N2N2N2O1N2N2N2N2N3M1O2N2O1N1O2N2N2N1O2N2O1N1O2M3L4L5K5K5L3L5K5K4M4K5Kk[]4"}, "label": "a black dog is sitting on sofa between two dogs"}]}
{"id": 329528, "image": "COCO_train2014_000000329528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an awake black dog sitting on a towel on a couch\"."}], "task": "refering", "answer_template": "The \"an awake black dog sitting on a towel on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 122, 123, 124, 125, 145, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 238, 239, 240, 261, 262, 263, 283, 284, 306, 307], "bbox": [0.31060937499999997, 0.27441666666666664, 0.52878125, 0.7895625000000001], "iscrowd": 0, "area": 15425.572300000002, "rle": {"size": [480, 640], "counts": "Z]m21o>0O2N101N101N1O2O0O2O0010O010O010OgBJh;7jC6_4_OZ2:VI8[4E]24VI9V4Jc2MVI:S4Ng2GUI<o34j2AVI<j39P3ZOUI>g3=S3VOTI>d3a0Y3PORI`0a3c0^3nNoHa0]3e0f3iNlHe0W3f0n3fNiHg0R3g0W4aNfHj0m2i0^4^NdHk0g2k0f4ZNbHP1_2i0R5WN^HU1?dNh0S2]6TN\\HY14hNm0m1g6QNXH_1GmNS1e1Q7PNTHX2e0J[7mMPHZ2>Le7jMmG\\26MP8hMiG\\20O[8dMeG[5]8eJcGW5b8iJ\\GR5j8nJVGl4Q9SKnFh4X9YKfFb4a9]K^Fa4e9:3L5L3M4K4M4L3L4M4L3M4N]FTLT8k3kGZLS8c3lGcLQ8\\3nGiLo7U3PHoLP8o2oGTMP8k2oGWMQ8i2mGYMR8h2lGZMT8f2jG]MU8c2iG_MV8b2hG`MX8a2fG`MZ8`2dGcM[8]2cGeM\\8\\2bGfM^8Z2`GgMa8Z2]GgMc8Y2\\GgMd8Z2ZGgMg8Y2XGhMh8X2VGiMl8V2SGjMn8W2PGjMP9V2nFSMk9m2TFRMn9n2QFRMQ:m2mETMT:l2kETMV:i32O1O2N1N2O1O1O2M2O1O2N2M3N2N2N2N2N2N2N2O1N2N2N2N2N3M1O2N2O1N1O2N2N2N1O2N2O1N1O2M3L4L5K5K5L3L5K5K4M4K5Kk[]4"}, "label": "an awake black dog sitting on a towel on a couch"}]}
{"id": 329528, "image": "COCO_train2014_000000329528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old golden lab that is tired\"."}], "task": "refering", "answer_template": "The \"an old golden lab that is tired\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 221, 222, 224, 225, 244, 245, 267, 268], "bbox": [0.5000312499999999, 0.29575, 0.87053125, 0.6758541666666666], "iscrowd": 0, "area": 19877.62505, "rle": {"size": [480, 640], "counts": "\\Uf41n>2N2O2M2N2O1N3M2N2O1N2O2M2O1O1TCUO`;P1[DQOe;R1WDoNi;T1SDmNl;X1nCkNQ<X1hClNW<o10000O10000O100O10000O100O0010O010O0010O10O0100O1O010001O000001O0O105K000000000000000O10001O0000000000O0O2O1NjEbMn7]2RHkMg7T2XHPNf7n1[HUNc7j1]HXNb7h1\\H[Nc7f1[H]Nb7j1VHYNi7i1SH[Nk7g1QH\\Nn7f1mG_NQ8c1kG_NU8c1fG`NZ8b1bG`N^8b1^G`Nb8b1YGaNf8b1TGbNl8`1mFeNS9]1fFhNZ9_3000000000N1N3N2L4I7I7M300000OK6J6K5N2cN[E]Ng:b1bEUN_:k1TF`Mo9_2Y1O1000001O0O100000001OlC_Mi;b2UD`Mj;`2UDaMk;`2RDcMm;]2RDeMn;f20000000O101O000000001O0000001O00000O2EoCeMQ<[2:0001O00RDcMY;^2eDeMY;[2fDhMX;X2fDkMZ;U2cDnM\\;U2]DTN^;g2001O01O001O1O1OO101O000O10000O10001N1O1N2O1N200O101O001N101O001N101O0O2O00000O1WOPDbNP<U1R1J6K5K5L4K5M3L4M3M3000000000O100001O001O1O001O1O001O1N1O2N2N1N3N1OggV1"}, "label": "an old golden lab that is tired"}]}
{"id": 329528, "image": "COCO_train2014_000000329528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the completely all brown dog\"."}], "task": "refering", "answer_template": "The \"the completely all brown dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 221, 222, 224, 225, 244, 245, 267, 268], "bbox": [0.5000312499999999, 0.29575, 0.87053125, 0.6758541666666666], "iscrowd": 0, "area": 19877.62505, "rle": {"size": [480, 640], "counts": "\\Uf41n>2N2O2M2N2O1N3M2N2O1N2O2M2O1O1TCUO`;P1[DQOe;R1WDoNi;T1SDmNl;X1nCkNQ<X1hClNW<o10000O10000O100O10000O100O0010O010O0010O10O0100O1O010001O000001O0O105K000000000000000O10001O0000000000O0O2O1NjEbMn7]2RHkMg7T2XHPNf7n1[HUNc7j1]HXNb7h1\\H[Nc7f1[H]Nb7j1VHYNi7i1SH[Nk7g1QH\\Nn7f1mG_NQ8c1kG_NU8c1fG`NZ8b1bG`N^8b1^G`Nb8b1YGaNf8b1TGbNl8`1mFeNS9]1fFhNZ9_3000000000N1N3N2L4I7I7M300000OK6J6K5N2cN[E]Ng:b1bEUN_:k1TF`Mo9_2Y1O1000001O0O100000001OlC_Mi;b2UD`Mj;`2UDaMk;`2RDcMm;]2RDeMn;f20000000O101O000000001O0000001O00000O2EoCeMQ<[2:0001O00RDcMY;^2eDeMY;[2fDhMX;X2fDkMZ;U2cDnM\\;U2]DTN^;g2001O01O001O1O1OO101O000O10000O10001N1O1N2O1N200O101O001N101O001N101O0O2O00000O1WOPDbNP<U1R1J6K5K5L4K5M3L4M3M3000000000O100001O001O1O001O1O001O1N1O2N2N1N3N1OggV1"}, "label": "the completely all brown dog"}]}
{"id": 419645, "image": "COCO_train2014_000000419645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is red colored truck in between the other trucks\"."}], "task": "refering", "answer_template": "The \"there is red colored truck in between the other trucks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.146796875, 0.1932482598607889, 0.9352500000000001, 0.8674245939675175], "iscrowd": 0, "area": 90934.01249999998, "rle": {"size": [431, 640], "counts": "hhW16o<;F;D<D;F;D;E<L4N1O2N1O0001O0001O0001O000001O01OeEYMn9[3I8H7I4L100O1O1O1O100O1O1O1O100O1O1O001O010O001O001O001O001O00001O00N2O1O001O1O1N2O1O1O1O1O1N2O1O1O1C=1O001O00001O001O00001O001N10001O001O00001O001O010O00001O001O000010O01O00001O001O000010O01O00001O001O00010O001O00001O00010O0O2A>C=C>C<O2O0O1O2O0O101N1O2O0O101N1O101N101N100O2N101N10001O00001O010O00001O0000000000O1000000000000000001O000000000001O00000000000000001O4L4L5K4L4L4K5L4L4L4L3M1O1O1O1O001O1N2O1O00001O0000001O00001N10001O0000001O00001O000O101O00001O00001O0000001N10001O00001O0000001O0O101OO1N2M3N3M2N2N2N2N2O10001O0001O0000000001O000000000000001O000000000000001O00000000001O001O1O1O001O1O001O1O001O1O001O1O1O001^HdJo6]5oHoJf6Q5YIZK]6g4bIdKS6]4kIiKP6W4oIjKQ6W4nIiKR6X4lIjKR6W4mIjKS6W4lIiKT6i500000001O00000O1000001O00000000001N1000000000001O00000O101O000000001O00000O101O0000001O0000000O2O0000001O0000000O2O000O100O2O000O101O1N101N101O1N101N2O001N2O001N2O0O2O1O0O2O001N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O0O2O001N2O0O2O001N101N101O0O2O1O0O2O0O4M6J7H:G>A`0A>B7H2O1O0O2O1N101O1N101O1N101NR\\a0"}, "label": "there is red colored truck in between the other trucks"}]}
{"id": 419645, "image": "COCO_train2014_000000419645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shiny red vintage pickup truck\"."}], "task": "refering", "answer_template": "The \"a shiny red vintage pickup truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.146796875, 0.1932482598607889, 0.9352500000000001, 0.8674245939675175], "iscrowd": 0, "area": 90934.01249999998, "rle": {"size": [431, 640], "counts": "hhW16o<;F;D<D;F;D;E<L4N1O2N1O0001O0001O0001O000001O01OeEYMn9[3I8H7I4L100O1O1O1O100O1O1O1O100O1O1O001O010O001O001O001O001O00001O00N2O1O001O1O1N2O1O1O1O1O1N2O1O1O1C=1O001O00001O001O00001O001N10001O001O00001O001O010O00001O001O000010O01O00001O001O000010O01O00001O001O00010O001O00001O00010O0O2A>C=C>C<O2O0O1O2O0O101N1O2O0O101N1O101N101N100O2N101N10001O00001O010O00001O0000000000O1000000000000000001O000000000001O00000000000000001O4L4L5K4L4L4K5L4L4L4L3M1O1O1O1O001O1N2O1O00001O0000001O00001N10001O0000001O00001O000O101O00001O00001O0000001N10001O00001O0000001O0O101OO1N2M3N3M2N2N2N2N2O10001O0001O0000000001O000000000000001O000000000000001O00000000001O001O1O1O001O1O001O1O001O1O001O1O1O001^HdJo6]5oHoJf6Q5YIZK]6g4bIdKS6]4kIiKP6W4oIjKQ6W4nIiKR6X4lIjKR6W4mIjKS6W4lIiKT6i500000001O00000O1000001O00000000001N1000000000001O00000O101O000000001O00000O101O0000001O0000000O2O0000001O0000000O2O000O100O2O000O101O1N101N101O1N101N2O001N2O001N2O0O2O1O0O2O001N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O0O2O001N2O0O2O001N101N101O0O2O1O0O2O0O4M6J7H:G>A`0A>B7H2O1O0O2O1N101O1N101O1N101NR\\a0"}, "label": "a shiny red vintage pickup truck"}]}
{"id": 362301, "image": "COCO_train2014_000000362301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboarder wearing a lime green coat\"."}], "task": "refering", "answer_template": "The \"a snowboarder wearing a lime green coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243], "bbox": [0.372484375, 0.25077083333333333, 0.60234375, 0.6333333333333333], "iscrowd": 0, "area": 15599.697800000002, "rle": {"size": [480, 640], "counts": "Vf_31o>2M3M3N2M3M3N2M3M4M2M4L3N3L3N3L1O101YESOX7n0fHWOW7i0gHYOY7h0cH\\O\\7d0bH^O^7b0`H@_7b0^HAa7?\\HDd7<ZHFf7:XHIg78VHJi77UHLk73RH0n71oG1R8NkG6T8JhG:Y8FcG>\\8B`Ga0b8^O[Ge0e8[OWGh0k8WOQGm0o8SOnFo0T9POhFT1X9lNeFV1]9iN_F[1a9eN\\F]1f9bNVF^1n9bNoEe0l:\\OPEf0R;b12N2N2N2N1O2N2N2N2N2M2O2N2N2N2N1OoE^Mh7`2ZHaMe7^2[HcMe7\\2\\HeMc7Z2^HfMb7Y2^HhMa7X2`HiM_7U2cHkM]7T2jFaMV1=o7Q2iFlMQ13V8P2gFWNk0K]8m1gF\\Nj0G^8m1fFbNi0Aa8l1eFhNf0^Od8i1dFoNe0XOg8h1bFVOd0SOi8f1bF[Ob0POl8d1`FBa0kNm8c1aFG?fNP9g3lFZLT9i3iFXLV9j3gFWLY9l3dFUL[9h4O1100O010O10O01O010O10O01N2N101N1O2O1N1O2O1N1O2O1N1O2O0O2N2eMkEIU:6[FZOg9d0]FYOd9f0_FVOb9i0bFTO_9j0eFRO\\9n0fFPO[9n0iFnNY9P1kFmNU9S1nFjNS9T1QGhNQ9V1SGgNm8Y1UGdNm8[1VGbNk8]1XG_Ni8`1[G]Nf8b1]GZNd8f1^GXNc8f1aGWN`8h1\\202M3N3M2M3L4I8I6I7J7H7Jdlf3"}, "label": "a snowboarder wearing a lime green coat"}]}
{"id": 362301, "image": "COCO_train2014_000000362301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in bright green coat\"."}], "task": "refering", "answer_template": "The \"person in bright green coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243], "bbox": [0.372484375, 0.25077083333333333, 0.60234375, 0.6333333333333333], "iscrowd": 0, "area": 15599.697800000002, "rle": {"size": [480, 640], "counts": "Vf_31o>2M3M3N2M3M3N2M3M4M2M4L3N3L3N3L1O101YESOX7n0fHWOW7i0gHYOY7h0cH\\O\\7d0bH^O^7b0`H@_7b0^HAa7?\\HDd7<ZHFf7:XHIg78VHJi77UHLk73RH0n71oG1R8NkG6T8JhG:Y8FcG>\\8B`Ga0b8^O[Ge0e8[OWGh0k8WOQGm0o8SOnFo0T9POhFT1X9lNeFV1]9iN_F[1a9eN\\F]1f9bNVF^1n9bNoEe0l:\\OPEf0R;b12N2N2N2N1O2N2N2N2N2M2O2N2N2N2N1OoE^Mh7`2ZHaMe7^2[HcMe7\\2\\HeMc7Z2^HfMb7Y2^HhMa7X2`HiM_7U2cHkM]7T2jFaMV1=o7Q2iFlMQ13V8P2gFWNk0K]8m1gF\\Nj0G^8m1fFbNi0Aa8l1eFhNf0^Od8i1dFoNe0XOg8h1bFVOd0SOi8f1bF[Ob0POl8d1`FBa0kNm8c1aFG?fNP9g3lFZLT9i3iFXLV9j3gFWLY9l3dFUL[9h4O1100O010O10O01O010O10O01N2N101N1O2O1N1O2O1N1O2O1N1O2O0O2N2eMkEIU:6[FZOg9d0]FYOd9f0_FVOb9i0bFTO_9j0eFRO\\9n0fFPO[9n0iFnNY9P1kFmNU9S1nFjNS9T1QGhNQ9V1SGgNm8Y1UGdNm8[1VGbNk8]1XG_Ni8`1[G]Nf8b1]GZNd8f1^GXNc8f1aGWN`8h1\\202M3N3M2M3L4I8I6I7J7H7Jdlf3"}, "label": "person in bright green coat"}]}
{"id": 304958, "image": "COCO_train2014_000000304958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white wooden chair next to the table\"."}], "task": "refering", "answer_template": "The \"a small white wooden chair next to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 206], "bbox": [0.8747031249999999, 0.19633333333333336, 1.0, 0.6933611111111111], "iscrowd": 0, "area": 5684.894199999998, "rle": {"size": [360, 640], "counts": "PQU62T;3N2M3M3M4M2M3M3M3M3N2M3M3M3M3N0O000001O000000mFeN[8Z1hGhNf7NnGY1>jNb7OmGW1d0kN[72mGS1j0lNW73lGQ1o0mNS74kGo0T1oNm6]1VIcNh6\\1ZIeNc6[1_IfN_6Y1cIgN[6[1fIeNX6^1fIbNX6c1eI]NZ6f1eIZNY6i1eIWNY6n1dIRN[6Q2cIoM[6U2dIkMZ6Y2cIgM[6P1oHJd0VO\\6Q1RII`0WO[6S1VIG=UO\\6U1YIF9UO^6V1ZIE6UO`6W1\\ID1VOc6V1]IDOUOd6W1_IDKUOg6W1_IEGTOk6W1_IEDUOl6V1cID@UOn6W1cIE\\OTOQ7X1dIDYOUOS7W1eIDVOUOU7W1gIOZ61hILX65iIJW65kIIU68lIJQ66PJLl55VJMf52\\J0`51aJ2[5NfJ4V5MkJ6Q5IRK5m4LTK3l4MUK2j4NXK1h4OYK0f41[KNe42]KLb44`KK`45`KK_46bKeNgN7g5S1cKeNVOIW6b0cJCk2"}, "label": "a small white wooden chair next to the table"}]}
{"id": 304958, "image": "COCO_train2014_000000304958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white dinning room chair\"."}], "task": "refering", "answer_template": "The \"a white dinning room chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 206], "bbox": [0.8747031249999999, 0.19633333333333336, 1.0, 0.6933611111111111], "iscrowd": 0, "area": 5684.894199999998, "rle": {"size": [360, 640], "counts": "PQU62T;3N2M3M3M4M2M3M3M3M3N2M3M3M3M3N0O000001O000000mFeN[8Z1hGhNf7NnGY1>jNb7OmGW1d0kN[72mGS1j0lNW73lGQ1o0mNS74kGo0T1oNm6]1VIcNh6\\1ZIeNc6[1_IfN_6Y1cIgN[6[1fIeNX6^1fIbNX6c1eI]NZ6f1eIZNY6i1eIWNY6n1dIRN[6Q2cIoM[6U2dIkMZ6Y2cIgM[6P1oHJd0VO\\6Q1RII`0WO[6S1VIG=UO\\6U1YIF9UO^6V1ZIE6UO`6W1\\ID1VOc6V1]IDOUOd6W1_IDKUOg6W1_IEGTOk6W1_IEDUOl6V1cID@UOn6W1cIE\\OTOQ7X1dIDYOUOS7W1eIDVOUOU7W1gIOZ61hILX65iIJW65kIIU68lIJQ66PJLl55VJMf52\\J0`51aJ2[5NfJ4V5MkJ6Q5IRK5m4LTK3l4MUK2j4NXK1h4OYK0f41[KNe42]KLb44`KK`45`KK_46bKeNgN7g5S1cKeNVOIW6b0cJCk2"}, "label": "a white dinning room chair"}]}
{"id": 124751, "image": "COCO_train2014_000000124751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man without a shirt lays on a bed next to another man who is performing tricks on a bed\"."}], "task": "refering", "answer_template": "The \"a man without a shirt lays on a bed next to another man who is performing tricks on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 32, 33, 54, 55, 56, 77, 78, 93, 94, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 163, 164, 165], "bbox": [0.0006562499999999999, 0.04010416666666667, 0.45570312500000004, 0.431125], "iscrowd": 0, "area": 11337.657949999997, "rle": {"size": [480, 640], "counts": "W51o>9F;F00000O010000O101O0O2O01O01O01O010O010O00010O0ZO_OnBa0k<GSC9k<KSC6k<LUC3i<1UCOi<5VCKh<7WCIh<:VCGg<=XCBg<a0WC@g<b0YC]Of<f0XCZOh<h0VCYOi<h0WCWOi<k0UCVOj<l0TCTOk<m0VCROj<o0UCROk<n0UCQOl<n0TCSOm<_1100O1O10O001O0010O0001N101N101O0O101N101N1000UN\\CZ1d<bNaC\\1_<`NgC_1\\<ZNiCe1j<O1O1O1O2M2O2N1O2N00000O1000000000000O1000000000000O01000000000O1000000000000O1000000000000O1000000000000O1000O10000000000O10O01O100O1O100O1O11O2M3N2N1O1OO1N2N2O1N2O0O2O1N2N2O1N2O1N2O1N1O2O1O1N2O1O1O1O1O1N101O1O1O1O1O0O0100O100O102M2O1N2O2M2O1N2O2M2O1N3N1N2O1N3N1N2N2J7H7I7I7JSj90RVF3M3M3M3M3M3M3M3M3M2ON10000O10000O10000O10000O2O0O10000O1000L3L5L4K4L5N21O010O01O100O100O0011O1O1O1N3N1O3M6J6I7J6J6J6JoQS5"}, "label": "a man without a shirt lays on a bed next to another man who is performing tricks on a bed"}]}
{"id": 124751, "image": "COCO_train2014_000000124751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is laying on a bed wearing a cap\"."}], "task": "refering", "answer_template": "The \"a man is laying on a bed wearing a cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 32, 33, 54, 55, 56, 77, 78, 93, 94, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 163, 164, 165], "bbox": [0.0006562499999999999, 0.04010416666666667, 0.45570312500000004, 0.431125], "iscrowd": 0, "area": 11337.657949999997, "rle": {"size": [480, 640], "counts": "W51o>9F;F00000O010000O101O0O2O01O01O01O010O010O00010O0ZO_OnBa0k<GSC9k<KSC6k<LUC3i<1UCOi<5VCKh<7WCIh<:VCGg<=XCBg<a0WC@g<b0YC]Of<f0XCZOh<h0VCYOi<h0WCWOi<k0UCVOj<l0TCTOk<m0VCROj<o0UCROk<n0UCQOl<n0TCSOm<_1100O1O10O001O0010O0001N101N101O0O101N101N1000UN\\CZ1d<bNaC\\1_<`NgC_1\\<ZNiCe1j<O1O1O1O2M2O2N1O2N00000O1000000000000O1000000000000O01000000000O1000000000000O1000000000000O1000000000000O1000O10000000000O10O01O100O1O100O1O11O2M3N2N1O1OO1N2N2O1N2O0O2O1N2N2O1N2O1N2O1N1O2O1O1N2O1O1O1O1O1N101O1O1O1O1O0O0100O100O102M2O1N2O2M2O1N2O2M2O1N3N1N2O1N3N1N2N2J7H7I7I7JSj90RVF3M3M3M3M3M3M3M3M3M2ON10000O10000O10000O10000O2O0O10000O1000L3L5L4K4L5N21O010O01O100O100O0011O1O1O1N3N1O3M6J6I7J6J6J6JoQS5"}, "label": "a man is laying on a bed wearing a cap"}]}
{"id": 124751, "image": "COCO_train2014_000000124751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with a red , blue and green bedspread\"."}], "task": "refering", "answer_template": "The \"a bed with a red , blue and green bedspread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 192, 197, 198, 199, 200, 201, 202, 207, 208, 209, 210, 211, 212, 213, 214, 215, 219, 220, 221, 222, 223, 224, 225, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.40675, 1.0, 0.9910208333333334], "iscrowd": 0, "area": 137163.3116, "rle": {"size": [480, 640], "counts": "\\=`1V8Z5O10000O1000000O01000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O0100000O100000000O100000000O10000000000O100000000O10000000000O100000000O100000000O10000000000O100000000O10000000000O100000000O100000000O1000000001O1O1O2N1O1O1O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O00O100O100O100OoIgG]5Y8cJgG]5Y8cJhG\\5W8eJjGZ5V8fJjGZ5U8gJlGX5T8hJlGX5S8iJnGV5R8jJoGU5P8lJPHT5P8lJQHS5n7nJRHR5n7nJSHQ5m7oJTHP5k7QKUHo4k7QKVHn4i7SKWHm4i7SKXHl4g7UKZHj4f7UK[Hk4d7VK]Hi4c7WK]Hi4c7WK^Hh4b7XK_Hg4o7kJQHU5\\8^JeGa5m801O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O0O2O001O1O001O001O1O001O001O00001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000oNaKbG^4\\8fKbGZ4\\8iKcGW4[8mKcGS4[8PLdGP4Z8SLeGm3Y8WLeGi3Y8ZLgGe3W8_LgGa3W8bLhG^3V8eLiG[3U8iLiGW3U8lLjGT3T8oLkGQ3R8TMmGj2R8YMmGg2Q8]MmGc2Q8`MnG`2P8cMoG]2o7gMoGY2o7jMPHV2n7mMRHR2l7RNRHn1l7UNSHk1l7XNRHh1m7f2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O001O001O1O001O001O001O001O001O001O1O001O001O001O0O2O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O0O2O1O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O005K[1eN\\1dN_C"}, "label": "a bed with a red , blue and green bedspread"}]}
{"id": 124751, "image": "COCO_train2014_000000124751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed the man in the black shirt is jumping on\"."}], "task": "refering", "answer_template": "The \"the bed the man in the black shirt is jumping on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 192, 197, 198, 199, 200, 201, 202, 207, 208, 209, 210, 211, 212, 213, 214, 215, 219, 220, 221, 222, 223, 224, 225, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.40675, 1.0, 0.9910208333333334], "iscrowd": 0, "area": 137163.3116, "rle": {"size": [480, 640], "counts": "\\=`1V8Z5O10000O1000000O01000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O0100000O100000000O100000000O10000000000O100000000O10000000000O100000000O100000000O10000000000O100000000O10000000000O100000000O100000000O1000000001O1O1O2N1O1O1O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O00O100O100O100OoIgG]5Y8cJgG]5Y8cJhG\\5W8eJjGZ5V8fJjGZ5U8gJlGX5T8hJlGX5S8iJnGV5R8jJoGU5P8lJPHT5P8lJQHS5n7nJRHR5n7nJSHQ5m7oJTHP5k7QKUHo4k7QKVHn4i7SKWHm4i7SKXHl4g7UKZHj4f7UK[Hk4d7VK]Hi4c7WK]Hi4c7WK^Hh4b7XK_Hg4o7kJQHU5\\8^JeGa5m801O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O0O2O001O1O001O001O1O001O001O00001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000oNaKbG^4\\8fKbGZ4\\8iKcGW4[8mKcGS4[8PLdGP4Z8SLeGm3Y8WLeGi3Y8ZLgGe3W8_LgGa3W8bLhG^3V8eLiG[3U8iLiGW3U8lLjGT3T8oLkGQ3R8TMmGj2R8YMmGg2Q8]MmGc2Q8`MnG`2P8cMoG]2o7gMoGY2o7jMPHV2n7mMRHR2l7RNRHn1l7UNSHk1l7XNRHh1m7f2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O001O001O1O001O001O001O001O001O001O1O001O001O001O0O2O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O0O2O1O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O005K[1eN\\1dN_C"}, "label": "the bed the man in the black shirt is jumping on"}]}
{"id": 124751, "image": "COCO_train2014_000000124751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a balding man wearing a black tshirt and white shorts jumping on to a bed\"."}], "task": "refering", "answer_template": "The \"a balding man wearing a black tshirt and white shorts jumping on to a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 52, 53, 56, 57, 58, 72, 73, 74, 75, 76, 77, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 103, 104, 105, 106, 118, 119, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 263, 264, 265], "bbox": [0.0960625, 0.0022500000000000003, 0.6539375000000001, 0.6696666666666666], "iscrowd": 0, "area": 44164.806200000006, "rle": {"size": [480, 640], "counts": "kUm02n>1O2ZAL\\>4dAL\\>5bAM]>3cAM]>3bAN^>2bAOZ>4eAMT>:lAFT>:kAGU>9kAHT>8kAIU>a00001O000GVO\\Bj0d=XOZBi0e=YOYBg0g=\\OVBd0j=90001O00000000000001O0000000000bBoNm<Q1PCQOQ=o0lBSOU=m0hBUOY=k0dBWO]=V10O1O1O100O2N1O100O1O001O1O001O1N101O001N2O001O1N101O0O2O10O10000O01000O100O10O10O10000O10O010000O1000O0100O10000O01000O10000O010O10000O10OfD]OY8c0dGCX8>gGGU89jGJS87kGLT84kGNT82kG0S80mG1S8OkG4T8LkG6S8KlG7S8IlG9S8GkG<S8DmG>R8BmG`0Q8AmGb0R8^OnGd0P8\\OPHe0n7\\ORHe0m7ZOTHg0k7YOUHi0h7XOXHi0g7WOYHj0e7WO[Hi0e7WO[Hj0d7UO]Hk0b7VO^Hj0b7VO^Hk0a7UO_Hk0`7VO`Hj0`7VO`Hk0_7TObHl0]7UOcHk0]7VObHj0]7YOaHg0_7\\O^Hd0c7]O[Hc0f7^OXHb0i7@TH`0m7ARH>o7CoG=R8DlG<U8FhG:Y8GeG9\\8HbG8_8I_G7b8K[G5f8LXG4i8MUG3k8OSG1n81oFOR92lFNU93iFNW94gFKZ96dFJ7QNW8V2`GI:TNS8T2aGI;XNVOEh8\\2dGG>ZNo7P2aGFa0ZNn7R2^GDd0ZNo7T2ZGBg0ZNP8V2VGAj0YNP8\\4PHdKQ8[4oGeKR8Z4oGfKP8Z4PHfKQ8[4mGfKS8[4lGdKU8]4iGcKW8_4gGbKY8^4fGbK[8`4cG_K]8c4aG^K_8c4_G]Ka8f4]GZKc8g4[GYKf8h4XGXKh8a50O1O010O010O1O010O1O010O1O100O101N1O100O1O2O0O100O2N100O1O2O0O100O1O2O0O1O01O00001O001O000[MmJ]KT5`4SK[Km4c4[KWKe4f4cKUK^4i4hKRKX4m4nKmJT4Q5SLhJn3W5XLcJi3[5^L_Jd3_5bL[J_3d5hLUJY3j5mLPJU3n5QMlIP3S6VMgIl2W6o2O1O1N2O1O1O1O1O1O1O1O1OfNXJ^If5d6[J[Id5e6^JZIa5f6aJYI^5g6dJXI[5h6gJWIX5i6jJVIb4]7`KbH`4^7bK`H_4^7cKaH]4^7eKaH[4^7gKaHY4^7iKaHX4]7jKbHV4]7lKbHT4^7mKaHS4^7PL`HQ4^7QLaHo3^7SLaHm3^7ULaHk3^7WLaHj3n0[Ka4l0_Jj3n0]Ka4k0`Jh3l0aKb4i0`Jg3k0cKd4h0_Jg3i0eKf4f0_Jf3h0hKg4d0`Jd3f0lKh4b0`Jc3e0nKk4`0^Jd3d0oKl4?_Jb3b0SLm4=_Ja3b0ULm4i5QKYJn4i5PKYJn4h5QKZJm4g5QK[Jo4e5PK]Jn4e5oJ^Jo4c5PK_Jn4b5PK`Jo4a5PKaJn4a5oJbJo4h701N2N2N1O2N2O0O2N2N2N1O2N2O1N1O2N2N1O1O2O0O1O2N2N2N2N2O1N2N2N2N2N2O2M2N3M2N2N3M2O1N3M2N2N4J6I8H7I7I7H8G:F9F:G[^W3"}, "label": "a balding man wearing a black tshirt and white shorts jumping on to a bed"}]}
{"id": 124751, "image": "COCO_train2014_000000124751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 52, 53, 56, 57, 58, 72, 73, 74, 75, 76, 77, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 103, 104, 105, 106, 118, 119, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 263, 264, 265], "bbox": [0.0960625, 0.0022500000000000003, 0.6539375000000001, 0.6696666666666666], "iscrowd": 0, "area": 44164.806200000006, "rle": {"size": [480, 640], "counts": "kUm02n>1O2ZAL\\>4dAL\\>5bAM]>3cAM]>3bAN^>2bAOZ>4eAMT>:lAFT>:kAGU>9kAHT>8kAIU>a00001O000GVO\\Bj0d=XOZBi0e=YOYBg0g=\\OVBd0j=90001O00000000000001O0000000000bBoNm<Q1PCQOQ=o0lBSOU=m0hBUOY=k0dBWO]=V10O1O1O100O2N1O100O1O001O1O001O1N101O001N2O001O1N101O0O2O10O10000O01000O100O10O10O10000O10O010000O1000O0100O10000O01000O10000O010O10000O10OfD]OY8c0dGCX8>gGGU89jGJS87kGLT84kGNT82kG0S80mG1S8OkG4T8LkG6S8KlG7S8IlG9S8GkG<S8DmG>R8BmG`0Q8AmGb0R8^OnGd0P8\\OPHe0n7\\ORHe0m7ZOTHg0k7YOUHi0h7XOXHi0g7WOYHj0e7WO[Hi0e7WO[Hj0d7UO]Hk0b7VO^Hj0b7VO^Hk0a7UO_Hk0`7VO`Hj0`7VO`Hk0_7TObHl0]7UOcHk0]7VObHj0]7YOaHg0_7\\O^Hd0c7]O[Hc0f7^OXHb0i7@TH`0m7ARH>o7CoG=R8DlG<U8FhG:Y8GeG9\\8HbG8_8I_G7b8K[G5f8LXG4i8MUG3k8OSG1n81oFOR92lFNU93iFNW94gFKZ96dFJ7QNW8V2`GI:TNS8T2aGI;XNVOEh8\\2dGG>ZNo7P2aGFa0ZNn7R2^GDd0ZNo7T2ZGBg0ZNP8V2VGAj0YNP8\\4PHdKQ8[4oGeKR8Z4oGfKP8Z4PHfKQ8[4mGfKS8[4lGdKU8]4iGcKW8_4gGbKY8^4fGbK[8`4cG_K]8c4aG^K_8c4_G]Ka8f4]GZKc8g4[GYKf8h4XGXKh8a50O1O010O010O1O010O1O010O1O100O101N1O100O1O2O0O100O2N100O1O2O0O100O1O2O0O1O01O00001O001O000[MmJ]KT5`4SK[Km4c4[KWKe4f4cKUK^4i4hKRKX4m4nKmJT4Q5SLhJn3W5XLcJi3[5^L_Jd3_5bL[J_3d5hLUJY3j5mLPJU3n5QMlIP3S6VMgIl2W6o2O1O1N2O1O1O1O1O1O1O1O1OfNXJ^If5d6[J[Id5e6^JZIa5f6aJYI^5g6dJXI[5h6gJWIX5i6jJVIb4]7`KbH`4^7bK`H_4^7cKaH]4^7eKaH[4^7gKaHY4^7iKaHX4]7jKbHV4]7lKbHT4^7mKaHS4^7PL`HQ4^7QLaHo3^7SLaHm3^7ULaHk3^7WLaHj3n0[Ka4l0_Jj3n0]Ka4k0`Jh3l0aKb4i0`Jg3k0cKd4h0_Jg3i0eKf4f0_Jf3h0hKg4d0`Jd3f0lKh4b0`Jc3e0nKk4`0^Jd3d0oKl4?_Jb3b0SLm4=_Ja3b0ULm4i5QKYJn4i5PKYJn4h5QKZJm4g5QK[Jo4e5PK]Jn4e5oJ^Jo4c5PK_Jn4b5PK`Jo4a5PKaJn4a5oJbJo4h701N2N2N1O2N2O0O2N2N2N1O2N2O1N1O2N2N1O1O2O0O1O2N2N2N2N2O1N2N2N2N2N2O2M2N3M2N2N3M2O1N3M2N2N4J6I8H7I7I7H8G:F9F:G[^W3"}, "label": "a man in a black shirt"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small child plays with toy on sofa\"."}], "task": "refering", "answer_template": "The \"small child plays with toy on sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 113, 114, 115, 131, 132, 133, 150, 151, 169, 170, 187], "bbox": [0.30716000000000004, 0.4443733333333333, 0.46740000000000004, 0.7887733333333332], "iscrowd": 0, "area": 4005.012200000001, "rle": {"size": [375, 500], "counts": "Qch14b;1O2N1O2N101O1O001O1O0O2O1O1O101N1O2ZO]OTFf0g9_OVFc0f9FSF<j9GSF<j9k0M3M3N2M2N3N2M2O2N1O2N2N1O2N2N2N3N1N3M3N1O2N6I5LPMnGZ2S8aMRH_2o7[MVHd2_810O012K6Jd0\\O1O2M3L3N3N1N3N3L3O2VOdE4Q;O1O1N2O2N1O1O1O1O101N1O1O1MfTQ3"}, "label": "small child plays with toy on sofa"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby boy standing in front of a sofa\"."}], "task": "refering", "answer_template": "The \"a baby boy standing in front of a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 113, 114, 115, 131, 132, 133, 150, 151, 169, 170, 187], "bbox": [0.30716000000000004, 0.4443733333333333, 0.46740000000000004, 0.7887733333333332], "iscrowd": 0, "area": 4005.012200000001, "rle": {"size": [375, 500], "counts": "Qch14b;1O2N1O2N101O1O001O1O0O2O1O1O101N1O2ZO]OTFf0g9_OVFc0f9FSF<j9GSF<j9k0M3M3N2M2N3N2M2O2N1O2N2N1O2N2N2N3N1N3M3N1O2N6I5LPMnGZ2S8aMRH_2o7[MVHd2_810O012K6Jd0\\O1O2M3L3N3N1N3N3L3O2VOdE4Q;O1O1N2O2N1O1O1O1O101N1O1O1MfTQ3"}, "label": "a baby boy standing in front of a sofa"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cushion type wooden chair where a man is sitting\"."}], "task": "refering", "answer_template": "The \"cushion type wooden chair where a man is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 161, 175, 176, 177, 178, 179, 193, 197], "bbox": [0.7642599999999999, 0.5996533333333334, 1.0, 0.81248], "iscrowd": 0, "area": 1425.00795, "rle": {"size": [375, 500], "counts": "^a\\43X;<M3L4M3M3L5LK6M4L3L5L5L5O010O1O010N1M\\d:0d[E310O00001O01O01O00010O001O00010O00001O01O01O000010O01O00010O00001O01O01O001O01O01O000010OlDIh:c000010NN0O2N1O1O4L4MFbEAZ:`0iE_OT:c0nE]On9d0TF]Oh9d0[F[Ob9e0bF[O[9e0hF[OV9c0mFAo8=TGJc84aGK^82gGNW:IRG"}, "label": "cushion type wooden chair where a man is sitting"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on couch using laptop\"."}], "task": "refering", "answer_template": "The \"man on couch using laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 72, 73, 74, 90, 91, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 165, 166, 182, 183, 184, 185, 201], "bbox": [0.0016799999999999999, 0.25, 0.29222000000000004, 0.8828800000000001], "iscrowd": 0, "area": 15068.592800000002, "rle": {"size": [375, 500], "counts": "]`0T2c92M3M3M4L3M3M3M4L2N2N100O1O1O1O2N1O1K5K5K5K5M4L3N4K6J5M3N2M4M2N2M2O0O100O10000O10000000000000fL`I]1a6aNiIV1X6iNPJo0P6QORJm0o5RORJm0o5QOSJn0n5QOSJn0n5QOSJn0m5QOUJn0l5POVJo0k5nNYJP1g5oN]Jn0d5oN`Jo0`5POcJn0^5oNfJo0Z5oNkJn0V5iNSKV1n4gNTKY1l4eNVK[1k4bNWK^1i4`NYK`1j4[NXKe1k4VNXKb0dM>W7kNWKg0cM=f9B[F>e9A\\F?d9A\\F?d9@]F`0c9@]F`0c9_O^Fa0b9_O_F`0a9_O`Fa0`9_O`Fa0U1oNk6`0PHa0n0YOo65THb0f0DR7JXHb0>MY7AYHb066`7WO\\Hb0Ka0h7mN]H]2b7bM_H_2a7`M_H`2a7`M_Ha2`7^MaHc2^7]MbHd2]7[MdHf2[7ZMfHe2Z7ZMgHg2X7YMhHh2W7WMjHj2U7UMlHl2T7SMlHm2T7RMmHo2S7PMnHP3R7nLoHS3P7mLPIT3P7jLQIU3Q7iLPIW3h7O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1OO1O2N2N2N2N2N2NBaGhM]8X2eGhMZ8V2jGiMW8T2kGlMV8Q2lGoMU8m1nGSNS8j1PHUNQ8g1RHXNP8e1RHZNP8b1SH]No7`1TH^Nn7V1]HiNe7?RI@P7O`I0U9N1O1OQ[Q4"}, "label": "man on couch using laptop"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a white couch on his computer\"."}], "task": "refering", "answer_template": "The \"a man sitting on a white couch on his computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 72, 73, 74, 90, 91, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 165, 166, 182, 183, 184, 185, 201], "bbox": [0.0016799999999999999, 0.25, 0.29222000000000004, 0.8828800000000001], "iscrowd": 0, "area": 15068.592800000002, "rle": {"size": [375, 500], "counts": "]`0T2c92M3M3M4L3M3M3M4L2N2N100O1O1O1O2N1O1K5K5K5K5M4L3N4K6J5M3N2M4M2N2M2O0O100O10000O10000000000000fL`I]1a6aNiIV1X6iNPJo0P6QORJm0o5RORJm0o5QOSJn0n5QOSJn0n5QOSJn0m5QOUJn0l5POVJo0k5nNYJP1g5oN]Jn0d5oN`Jo0`5POcJn0^5oNfJo0Z5oNkJn0V5iNSKV1n4gNTKY1l4eNVK[1k4bNWK^1i4`NYK`1j4[NXKe1k4VNXKb0dM>W7kNWKg0cM=f9B[F>e9A\\F?d9A\\F?d9@]F`0c9@]F`0c9_O^Fa0b9_O_F`0a9_O`Fa0`9_O`Fa0U1oNk6`0PHa0n0YOo65THb0f0DR7JXHb0>MY7AYHb066`7WO\\Hb0Ka0h7mN]H]2b7bM_H_2a7`M_H`2a7`M_Ha2`7^MaHc2^7]MbHd2]7[MdHf2[7ZMfHe2Z7ZMgHg2X7YMhHh2W7WMjHj2U7UMlHl2T7SMlHm2T7RMmHo2S7PMnHP3R7nLoHS3P7mLPIT3P7jLQIU3Q7iLPIW3h7O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1OO1O2N2N2N2N2N2NBaGhM]8X2eGhMZ8V2jGiMW8T2kGlMV8Q2lGoMU8m1nGSNS8j1PHUNQ8g1RHXNP8e1RHZNP8b1SH]No7`1TH^Nn7V1]HiNe7?RI@P7O`I0U9N1O1OQ[Q4"}, "label": "a man sitting on a white couch on his computer"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing the yellow shirt\"."}], "task": "refering", "answer_template": "The \"man wearing the yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 88, 89, 105, 106, 107, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 174, 175, 176, 177, 192, 193, 194, 195, 212], "bbox": [0.69872, 0.28370666666666666, 1.0, 0.8724533333333334], "iscrowd": 0, "area": 13100.220850000002, "rle": {"size": [375, 500], "counts": "k^P44`;3M4L3M4M2M3M4L3M4L3M4L3N3L3M4L3M4L3M4L3M3M4L3N3M2N3N1O2M2O2N1O2M2O2O00010O0010O01O010O010O0010O010O00010O01O010O01M2N2N3M2N3M2N2O2M2N2N3M2N3L3M3N3L3M3N3K4L5L3L4L5L3L]NVObHg0^7_O^H`0a7F[H:d7OUH0i7;nGEP8f0gGZOX8P1`GnN`8\\1WGdNi8U2O1O1O001O1O1O100O001O1O1O0002O1N1O2O1N101NUOWMiHi2Q7bMkH^2o6lMlHT2P7VNjHj1V7[NeHf1[7^NaHa1`7Z1010O00100O010O00100O002O0O1O101N100O1O2O0O100O2O0O2O2M2XNi1[OoL"}, "label": "man wearing the yellow shirt"}]}
{"id": 419664, "image": "COCO_train2014_000000419664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a yellow t - shirt\"."}], "task": "refering", "answer_template": "The \"a man in a yellow t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 88, 89, 105, 106, 107, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 174, 175, 176, 177, 192, 193, 194, 195, 212], "bbox": [0.69872, 0.28370666666666666, 1.0, 0.8724533333333334], "iscrowd": 0, "area": 13100.220850000002, "rle": {"size": [375, 500], "counts": "k^P44`;3M4L3M4M2M3M4L3M4L3M4L3N3L3M4L3M4L3M4L3M3M4L3N3M2N3N1O2M2O2N1O2M2O2O00010O0010O01O010O010O0010O010O00010O01O010O01M2N2N3M2N3M2N2O2M2N2N3M2N3L3M3N3L3M3N3K4L5L3L4L5L3L]NVObHg0^7_O^H`0a7F[H:d7OUH0i7;nGEP8f0gGZOX8P1`GnN`8\\1WGdNi8U2O1O1O001O1O1O100O001O1O1O0002O1N1O2O1N101NUOWMiHi2Q7bMkH^2o6lMlHT2P7VNjHj1V7[NeHf1[7^NaHa1`7Z1010O00100O010O00100O002O0O1O101N100O1O2O0O100O2O0O2O2M2XNi1[OoL"}, "label": "a man in a yellow t - shirt"}]}
{"id": 321363, "image": "COCO_train2014_000000321363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a bald man with glasses eating a huge pink frosted donut\"."}], "task": "refering", "answer_template": "The \"this is a bald man with glasses eating a huge pink frosted donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.0601875, 0.09425, 0.9330416666666667, 0.974609375], "iscrowd": 0, "area": 165692.21125000002, "rle": {"size": [640, 480], "counts": "\\gb01mc02N2O1]JV3nClLn;X3QDiLk;[3SDgLj;\\3TDfLh;^3WDcLe;a3YDaLc;c3\\D^L`;f3^D\\L^;h3aDYL[;k3cDWLZ;l3dDVLX;n3gDSLU;Q4iDQLS;S4lDnKP;V4nDlKn:X4PEjKl:Z4SEgKi:]4UEeKh:^4WEcKe:a4YEaKc:c4\\E^K`:f4^E\\K^:h4`EZK\\:i4dEXKX:l4fEVKW:m4hETKT:P5jERKR:R5lEPKP:T5oEmJm9W5QFkJk9Y5TFhJh9\\5VFfJg9]5XFdJd9`5ZFbJb9b5\\F`J`9d5_F]J]9g5aF[J[9i5dFXJX9l5fFVJW9m5gFUJU9o5jFRJR9R6lFPJP9T6oFmIm8W6QGkIn8d8000000O10000O10000O10000O1000000O10000O10000O10000O10000O1000000O10000O2O000O10000O10000O100O1O100O1O100O1O1O100O1O100O1O1O100O1O10000eMVD`Lj;_3YD^Lh;`3\\D^Ld;a3_D]La;b3bD\\L^;c3fDZLZ;d3jDZLV;e3mDXLT;g3oDWLQ;h3REVLn:h3VEVLj:i3YEULg:j3\\ESLe:l3^ERLb:l3bERL^:m3eEQL[:n3hEPLX:o3kEoKU:o3oEoKQ:P4RFnKn9Q4UFmKk9R4XFlKh9S4[FkKe9S4_FkKa9T4bFjK^9U4eFiK[9V4hFhKX9V4lFhKT9W4oFgKQ9X4RGfKn8U4YGiKg8R4`GlK`8n3hGPLX8l3nGRLR8k3S4M3M3M3M3M3M3M3O1O1O1O1O1O1O1O1O1O1O100O100O1O100O100O1O100O1O100O100O100O100000000000000000000O1000000000001O0000000O100000000000000000000O100000000000000000000O1001O001O001O001O001O001O001O001O001O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O2N1O1O2N2eC`KR9b4jF`KV9c4dFaK[9b4`FaK_9b4\\F`Kd9b4WFbKh9a4SFaKm9b4nEaKQ:b4iEaKW:a4eEbKZ:a4aEbK^:a4\\EbKd:a4WEbKh:`4TEbKl:a4nDcKQ;`4jDcKU;`4fDbKZ;a4`DcK_;_4]DcKc;`4XDcKg;[601O1O001O001O001O000000001O01O0001O00001O00001O00001O00001O00001O000XMiHgHX7X7lHdHU7[7nHbHS7]7PI`HQ7_7RI^Ho6b7SI[Hm6e7WIWHj6h7YIUHh6j7[ISHf6l7]IQHd6o7^InGb6R8bIjG_6U8dIhG]6W8fIfG[6Y8hIdGY6\\8iIaGW6_8f2O1O1O10O01O1O1O10O01O1O1O1O10O01O1O1O100O1O2N2N2O1N2N2N1O2O1N2N2N2N101N2N2N2N2N2Od0[O?A7I6J7J6I6J7I7I6K6I6J7I7I6Kh1WNb0^Oi_c0"}, "label": "this is a bald man with glasses eating a huge pink frosted donut"}]}
{"id": 321363, "image": "COCO_train2014_000000321363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man eating a donut\"."}], "task": "refering", "answer_template": "The \"the man eating a donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.0601875, 0.09425, 0.9330416666666667, 0.974609375], "iscrowd": 0, "area": 165692.21125000002, "rle": {"size": [640, 480], "counts": "\\gb01mc02N2O1]JV3nClLn;X3QDiLk;[3SDgLj;\\3TDfLh;^3WDcLe;a3YDaLc;c3\\D^L`;f3^D\\L^;h3aDYL[;k3cDWLZ;l3dDVLX;n3gDSLU;Q4iDQLS;S4lDnKP;V4nDlKn:X4PEjKl:Z4SEgKi:]4UEeKh:^4WEcKe:a4YEaKc:c4\\E^K`:f4^E\\K^:h4`EZK\\:i4dEXKX:l4fEVKW:m4hETKT:P5jERKR:R5lEPKP:T5oEmJm9W5QFkJk9Y5TFhJh9\\5VFfJg9]5XFdJd9`5ZFbJb9b5\\F`J`9d5_F]J]9g5aF[J[9i5dFXJX9l5fFVJW9m5gFUJU9o5jFRJR9R6lFPJP9T6oFmIm8W6QGkIn8d8000000O10000O10000O10000O1000000O10000O10000O10000O10000O1000000O10000O2O000O10000O10000O100O1O100O1O100O1O1O100O1O100O1O1O100O1O10000eMVD`Lj;_3YD^Lh;`3\\D^Ld;a3_D]La;b3bD\\L^;c3fDZLZ;d3jDZLV;e3mDXLT;g3oDWLQ;h3REVLn:h3VEVLj:i3YEULg:j3\\ESLe:l3^ERLb:l3bERL^:m3eEQL[:n3hEPLX:o3kEoKU:o3oEoKQ:P4RFnKn9Q4UFmKk9R4XFlKh9S4[FkKe9S4_FkKa9T4bFjK^9U4eFiK[9V4hFhKX9V4lFhKT9W4oFgKQ9X4RGfKn8U4YGiKg8R4`GlK`8n3hGPLX8l3nGRLR8k3S4M3M3M3M3M3M3M3O1O1O1O1O1O1O1O1O1O1O100O100O1O100O100O1O100O1O100O100O100O100000000000000000000O1000000000001O0000000O100000000000000000000O100000000000000000000O1001O001O001O001O001O001O001O001O001O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O2N1O1O2N2eC`KR9b4jF`KV9c4dFaK[9b4`FaK_9b4\\F`Kd9b4WFbKh9a4SFaKm9b4nEaKQ:b4iEaKW:a4eEbKZ:a4aEbK^:a4\\EbKd:a4WEbKh:`4TEbKl:a4nDcKQ;`4jDcKU;`4fDbKZ;a4`DcK_;_4]DcKc;`4XDcKg;[601O1O001O001O001O000000001O01O0001O00001O00001O00001O00001O00001O000XMiHgHX7X7lHdHU7[7nHbHS7]7PI`HQ7_7RI^Ho6b7SI[Hm6e7WIWHj6h7YIUHh6j7[ISHf6l7]IQHd6o7^InGb6R8bIjG_6U8dIhG]6W8fIfG[6Y8hIdGY6\\8iIaGW6_8f2O1O1O10O01O1O1O10O01O1O1O1O10O01O1O1O100O1O2N2N2O1N2N2N1O2O1N2N2N2N101N2N2N2N2N2Od0[O?A7I6J7J6I6J7I7I6K6I6J7I7I6Kh1WNb0^Oi_c0"}, "label": "the man eating a donut"}]}
{"id": 100182, "image": "COCO_train2014_000000100182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat is parking between a two boats\"."}], "task": "refering", "answer_template": "The \"a boat is parking between a two boats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 145, 146, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173], "bbox": [0.308859375, 0.04269320843091335, 0.6776875, 0.5123653395784543], "iscrowd": 0, "area": 36792.81480000001, "rle": {"size": [427, 640], "counts": "_eb26R=6J7H7J6J6J7H7J6J6J7H7J6J6J7H7J6J7I6I7J6J5K3L4M4O001O0O101O001O00001O001O00001N101O00001O001O00001O001O0O101O001O00001O001O00001N101O00001O001O00001O001N10001O001O00001O001O000O2O00001O001O00001O001O0O101O001O00001O001O00001N101OO10000O10000O100O1000O010000O100O10000O100O10O10O10000O100O10000O100O01000O10000O100O10000O10O010000O10000O100O10000O010O10000O2O1O0O2O0O2O1O0O2O1N101O0O2O1O0O2O1N101O0O2O1O0O2O0O2O1M2O2N2N1O2M2O2N2N1O2M3N2N6J6J6I8I6J6I7I8H7H8I7I8H7H8I7I7H9Hghe2"}, "label": "a boat is parking between a two boats"}]}
{"id": 100182, "image": "COCO_train2014_000000100182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white boat docked between two other white boats\"."}], "task": "refering", "answer_template": "The \"white boat docked between two other white boats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 145, 146, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173], "bbox": [0.308859375, 0.04269320843091335, 0.6776875, 0.5123653395784543], "iscrowd": 0, "area": 36792.81480000001, "rle": {"size": [427, 640], "counts": "_eb26R=6J7H7J6J6J7H7J6J6J7H7J6J6J7H7J6J7I6I7J6J5K3L4M4O001O0O101O001O00001O001O00001N101O00001O001O00001O001O0O101O001O00001O001O00001N101O00001O001O00001O001N10001O001O00001O001O000O2O00001O001O00001O001O0O101O001O00001O001O00001N101OO10000O10000O100O1000O010000O100O10000O100O10O10O10000O100O10000O100O01000O10000O100O10000O10O010000O10000O100O10000O010O10000O2O1O0O2O0O2O1O0O2O1N101O0O2O1O0O2O1N101O0O2O1O0O2O0O2O1M2O2N2N1O2M2O2N2N1O2M3N2N6J6J6I8I6J6I7I8H7H8I7I8H7H8I7I7H9Hghe2"}, "label": "white boat docked between two other white boats"}]}
{"id": 100182, "image": "COCO_train2014_000000100182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat furthest to the right\"."}], "task": "refering", "answer_template": "The \"the boat furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 180, 181, 182], "bbox": [0.6731874999999999, 0.04044496487119438, 1.0, 0.4988758782201405], "iscrowd": 0, "area": 33343.899249999995, "rle": {"size": [427, 640], "counts": "Qkc55Q=8I6K6L3M4K4M4L3L4M4L3L5J5J7I6J7J5J7I6J7J5J6J4L4M3L5K4L4M3L5O0O10001O0000001O00000O2O0000001O0000001O0O10001O0000001O00000O2O0000001O000000001N1000001O0000001O000O10001O0000001O00000O2O000000001O0000001N1000001O000000001O0O10001O0000001O00000O101O0000001O0000001N100000001O0000001O000O10001O0000000000000O2O0000000000O1N2O1O010O100O100O100O100O010O1O100O100O100O10O0100O101N101N1O101N100O2O0O2O0O101N100O2O0O2O0^Ob0WNj1VNo0"}, "label": "the boat furthest to the right"}]}
{"id": 100182, "image": "COCO_train2014_000000100182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat in the shore with a blue cloth on it\"."}], "task": "refering", "answer_template": "The \"a boat in the shore with a blue cloth on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 180, 181, 182], "bbox": [0.6731874999999999, 0.04044496487119438, 1.0, 0.4988758782201405], "iscrowd": 0, "area": 33343.899249999995, "rle": {"size": [427, 640], "counts": "Qkc55Q=8I6K6L3M4K4M4L3L4M4L3L5J5J7I6J7J5J7I6J7J5J6J4L4M3L5K4L4M3L5O0O10001O0000001O00000O2O0000001O0000001O0O10001O0000001O00000O2O0000001O000000001N1000001O0000001O000O10001O0000001O00000O2O000000001O0000001N1000001O000000001O0O10001O0000001O00000O101O0000001O0000001N100000001O0000001O000O10001O0000000000000O2O0000000000O1N2O1O010O100O100O100O100O010O1O100O100O100O10O0100O101N101N1O101N100O2O0O2O0O101N100O2O0O2O0^Ob0WNj1VNo0"}, "label": "a boat in the shore with a blue cloth on it"}]}
{"id": 477015, "image": "COCO_train2014_000000477015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red truck\"."}], "task": "refering", "answer_template": "The \"a red truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295], "bbox": [0.736, 0.3199375, 1.0, 0.7138958333333333], "iscrowd": 0, "area": 25523.36254999999, "rle": {"size": [480, 640], "counts": "inl6g0R>d0B8G2N1O2N100O2M3N2M300O2O000O1000000O100bDlMP:T2kEPNV:Q2lD]NKFZ;X30O1O1O100OQEbL^:_3^EoLW:R3fEVMS:l2jE\\MP:e2mEcMm9j3N1O1O001O1O1O3M3M3M3M3M2N1O2N1O2N0000000000000001O0O10000O10000O10001O0O10001N10000O101O0O10001N100O10001N100O1O100O1C[F^Kf9b4[F\\Kf9e4[FXKf9h4[FVKf9j4\\FTKc9m481O0000001O0000000000000O101O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000O1000000000000O10000EZF\\Kg9Z4dFdK]9h3XGVLi8o2S2D;DWK"}, "label": "a red truck"}]}
{"id": 477015, "image": "COCO_train2014_000000477015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red colour tempo parking in the road\"."}], "task": "refering", "answer_template": "The \"a red colour tempo parking in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295], "bbox": [0.736, 0.3199375, 1.0, 0.7138958333333333], "iscrowd": 0, "area": 25523.36254999999, "rle": {"size": [480, 640], "counts": "inl6g0R>d0B8G2N1O2N100O2M3N2M300O2O000O1000000O100bDlMP:T2kEPNV:Q2lD]NKFZ;X30O1O1O100OQEbL^:_3^EoLW:R3fEVMS:l2jE\\MP:e2mEcMm9j3N1O1O001O1O1O3M3M3M3M3M2N1O2N1O2N0000000000000001O0O10000O10000O10001O0O10001N10000O101O0O10001N100O10001N100O1O100O1C[F^Kf9b4[F\\Kf9e4[FXKf9h4[FVKf9j4\\FTKc9m481O0000001O0000000000000O101O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000O1000000000000O10000EZF\\Kg9Z4dFdK]9h3XGVLi8o2S2D;DWK"}, "label": "a red colour tempo parking in the road"}]}
{"id": 550746, "image": "COCO_train2014_000000550746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a hat and suspenders\"."}], "task": "refering", "answer_template": "The \"a boy with a hat and suspenders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 138, 139, 140, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 341, 342, 343, 344, 345, 358, 359, 360, 361, 362, 375, 376, 377, 378, 379], "bbox": [0.006004184100418411, 0.22570312499999998, 0.40228033472803343, 0.9835625], "iscrowd": 0, "area": 55370.7665, "rle": {"size": [640, 478], "counts": "TX24gc06K5J5K6K5J6J6K5J6J6K5J6J5L5J6J6K5J6J6K5J6J6K5J5K6K5L4N2O^1bN2N2M3N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2MW1jNZ1fNY1gN5J10000SN_D[Ka;e4_D[Ka;e4`DYKa;f4aDYK_;g4bDXK^;h4cDWK];i4cDVK^;i4dDVK\\;j4eDUK[;k4fDTKZ;l4fDSK[;l4gDSKY;m4hDRKX;n4iDQKW;o4jDoJW;P5jDnJX;R5iDjJZ;V5gDgJ[;X5gDeJ[;X5iDdJZ;X5lDdJV;X5SEaJo:\\5XE]Jk:_5]E[Je:a5o1M3L4M3L5K4M3L41O0000000000000000001O00000000000000001O00000000000000001O0000000001O000001O000000000XDQK_8o4ZGXKf8h4SG`Kl8`4nFgKQ9Z4lFiKS9W4kFkKU9U4iFnKV9R4hFQLW9o3gFTLX9l3fFWLY9i3eFYL[9m3]FVLb9P4VFSLi9T4nEoKQ:W4gEkKY:X4bEkK]:U4aEnK^:S4_EPL`:P4^ERLo8gNUGW5KTLX8^OkG_4LTL`77bHe3MVLi6o0WIl2OVLQ6h1nIR21VL\\5_2aJ\\12VL\\5`2_J[14WL\\5`2]J[16VL\\5a2[JZ19UL]5b2WJZ1<UL]5b2TJ[1>TLZ2PNNj:gMXGW1SOQ1e9gMYG48S2^8iMfJU2Z5jMhJU2W5kMlJS2T5kMPKR2P5nMRKQ2m4oMVKo1j4PNYKn1f4RN\\Km1d4QN`Km1_4SNdKk1\\4TNfKk1Y4UNjKi1V4UNnKi1Q4WNQLh1n3XNULe1l3ZNWLa1k3_NWL]1l3aNXL[1i3eNZLV1i3iNZLR1h3nNZLn0`;K5K4L5K5K4LkYb5"}, "label": "a boy with a hat and suspenders"}]}
{"id": 550746, "image": "COCO_train2014_000000550746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a hat on and carrying something\"."}], "task": "refering", "answer_template": "The \"a man with a hat on and carrying something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 138, 139, 140, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 341, 342, 343, 344, 345, 358, 359, 360, 361, 362, 375, 376, 377, 378, 379], "bbox": [0.006004184100418411, 0.22570312499999998, 0.40228033472803343, 0.9835625], "iscrowd": 0, "area": 55370.7665, "rle": {"size": [640, 478], "counts": "TX24gc06K5J5K6K5J6J6K5J6J6K5J6J5L5J6J6K5J6J6K5J6J6K5J5K6K5L4N2O^1bN2N2M3N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2MW1jNZ1fNY1gN5J10000SN_D[Ka;e4_D[Ka;e4`DYKa;f4aDYK_;g4bDXK^;h4cDWK];i4cDVK^;i4dDVK\\;j4eDUK[;k4fDTKZ;l4fDSK[;l4gDSKY;m4hDRKX;n4iDQKW;o4jDoJW;P5jDnJX;R5iDjJZ;V5gDgJ[;X5gDeJ[;X5iDdJZ;X5lDdJV;X5SEaJo:\\5XE]Jk:_5]E[Je:a5o1M3L4M3L5K4M3L41O0000000000000000001O00000000000000001O00000000000000001O0000000001O000001O000000000XDQK_8o4ZGXKf8h4SG`Kl8`4nFgKQ9Z4lFiKS9W4kFkKU9U4iFnKV9R4hFQLW9o3gFTLX9l3fFWLY9i3eFYL[9m3]FVLb9P4VFSLi9T4nEoKQ:W4gEkKY:X4bEkK]:U4aEnK^:S4_EPL`:P4^ERLo8gNUGW5KTLX8^OkG_4LTL`77bHe3MVLi6o0WIl2OVLQ6h1nIR21VL\\5_2aJ\\12VL\\5`2_J[14WL\\5`2]J[16VL\\5a2[JZ19UL]5b2WJZ1<UL]5b2TJ[1>TLZ2PNNj:gMXGW1SOQ1e9gMYG48S2^8iMfJU2Z5jMhJU2W5kMlJS2T5kMPKR2P5nMRKQ2m4oMVKo1j4PNYKn1f4RN\\Km1d4QN`Km1_4SNdKk1\\4TNfKk1Y4UNjKi1V4UNnKi1Q4WNQLh1n3XNULe1l3ZNWLa1k3_NWL]1l3aNXL[1i3eNZLV1i3iNZLR1h3nNZLn0`;K5K4L5K5K4LkYb5"}, "label": "a man with a hat on and carrying something"}]}
{"id": 493407, "image": "COCO_train2014_000000493407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cow with the number 610 displayed around its neck\"."}], "task": "refering", "answer_template": "The \"a white cow with the number 610 displayed around its neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 61, 62, 63, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 122, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 147, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 378, 379, 380, 381, 383, 384, 385, 386], "bbox": [0.13610309278350516, 0.12993749999999998, 0.905360824742268, 0.988671875], "iscrowd": 0, "area": 148320.63865000007, "rle": {"size": [640, 485], "counts": "iZY12mc06K4K6K4K3N3M2_EUOi1m0nM^On1d0iMFT2=aMO[23mKlN^K[1b8K\\KDcKl0m8BmJ;gK=Y9[O]J]3_5eL^J_3_5cL^Jb3^5`L_Jd3^5^L_Jg3]5\\L`Jg3]5[L`Jj3\\5XLaJl3\\5VLaJo3[5TLaJP4\\5RLaJS4\\5nKaJV4]5kKaJY4\\5iK`J\\4]5eK`J_4]5cK`JZ5e4gJXKV6m3kIPLY6n3hI`JSN0X8_5gI\\JbNEj7n5eIXJP8g5QHUJS8j5nGQJW8m5kGnIZ8Q6gGkI^8S6i2O1O1O1O1O2M2O1O1O100O2N1O1O1O100O2N1O1O100O2N1N2L4L4M4K4L4L4L4L5K4M3M3N2M3[Li_O\\3^`0O100O1O100O1O1O100O1O100O00100O1O1O100O10000O10000000000O10000000000O100000000O1000dNQMgAo2Y>TMdAl2\\>WMaAi2^>[M_Ad2b>^M\\Ab2d>aMYA_2g>dMVA\\2i>hMTAX2l>kMQAU2o>mMo@R2R?oMm@Q2S?PNl@P2U?PNj@Q2V?oMi@Q2X?oMg@Q2Y?QNe@P2[?PNd@P2]?PNb@P2_?PN`@P2`?QN_@P2a?PN]@Q2d?PNZ@P2f?QNY@P2g?PNX@P2i?S110O0000001O00010O0000001O00001O00001O0000001O00001O2M3N2N2N2N3L3M3L4M3L4M2M3N2M3N101N2N2N101N2N2OO01OZKaBm2^=cLSC]3m<[L[Cd3e<VLbCj3]<oKkCQ4Q>10O100000O10O1000000O0100000O10O10O10O0L4M4KPLd@e3X?^Lj@a3n>fLTAZ3b>oL_AP3X>XMjAh2l=`MVB_2h=bM[B\\2e=dM\\B\\2d=bM^B]2b=bM`B]2`=bMbB^2^=aMcB^2]=aMeB^2[=aMgB_2Y=_MiB`2W=`MjB_2W=_MkBa2U=]MnBa2W=YMkBf2Y=VMhBj2\\=PMfBo2_=kLcBU3`=fLbBY3c=aL_B^3R?00O1O101N1O100O1O100O1O100O101N1O100O1OnDoLW5R3eJ\\MP5d2lJgMm4Y2SImLZNQ1_8R2VIRMSNT1c8j1YIXMlMU1g8c1[IeM_MR1Q9Y1^InMVMV1R9l0gIXNjLR1]9f0hIm0W6SOhIo0W6QOgIQ1Y6oNfIS1Y6nNeIT1Z6lNeIU1[6kNdIW1[6iNcIZ1\\6fNcI\\1]6cNcI]1]6dNbI]1]6cNdI\\1\\6dNdI]1Z6dNfI\\1X6fNhI[1V6gNjIY1T6iNeGhLZ1_4o6mN_GkLa1Y4n6oNTGSMo1m3l6KSI6j6MUI3j6NVI3h6OXI0f63YIMf64ZIMd65[IKc68]IGb6:^IF`6=_IC`6?_IA_6a0bI_O\\6c0cI]O\\6d0dI\\OZ6f0fIZOY6g0gIYOW6i0iIXOU6i0kIWOT6k0kIUOS6m0mISOR6n0nIROo5Q1QJPOk5T1TJlNi5W1XJhNe5\\1ZJdNc5_1]JaN`5c1_J]N^5f1bJZN[5i1eJWNX5m1gJRNW5Q2iJoMS5V2kJkMQ5Y2oJfMm4_2SKaMi4d2VK\\Me4i2ZKWMc4n2\\KRM_4S3aKlL\\4X3cKhLY2^KTNP8BaLZ2hKiMj7M^LZ2PL^Me77[L\\2XLRMa7a0VL]2bLgL[7l0RLPOeKj2W1nLU7W1PLlNkKl2j9WNZJjNRLm2d9YNYJfN[Lm2]9\\NXKb1i4]NZK`1f4`N\\K_OmJ=n94XKWO\\K0e9g0RKQOi6o0ZIiNk6W1P6N10001N101O1N2O1O1O0O2O1O1N2O1O1N2O1O1N2O2N3M3L4M2N_bl0"}, "label": "a white cow with the number 610 displayed around its neck"}]}
{"id": 493407, "image": "COCO_train2014_000000493407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cow with the numbers 610 on its collar being led by a woman wearing a white coat\"."}], "task": "refering", "answer_template": "The \"a white cow with the numbers 610 on its collar being led by a woman wearing a white coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 61, 62, 63, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 122, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 147, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 378, 379, 380, 381, 383, 384, 385, 386], "bbox": [0.13610309278350516, 0.12993749999999998, 0.905360824742268, 0.988671875], "iscrowd": 0, "area": 148320.63865000007, "rle": {"size": [640, 485], "counts": "iZY12mc06K4K6K4K3N3M2_EUOi1m0nM^On1d0iMFT2=aMO[23mKlN^K[1b8K\\KDcKl0m8BmJ;gK=Y9[O]J]3_5eL^J_3_5cL^Jb3^5`L_Jd3^5^L_Jg3]5\\L`Jg3]5[L`Jj3\\5XLaJl3\\5VLaJo3[5TLaJP4\\5RLaJS4\\5nKaJV4]5kKaJY4\\5iK`J\\4]5eK`J_4]5cK`JZ5e4gJXKV6m3kIPLY6n3hI`JSN0X8_5gI\\JbNEj7n5eIXJP8g5QHUJS8j5nGQJW8m5kGnIZ8Q6gGkI^8S6i2O1O1O1O1O2M2O1O1O100O2N1O1O1O100O2N1O1O100O2N1N2L4L4M4K4L4L4L4L5K4M3M3N2M3[Li_O\\3^`0O100O1O100O1O1O100O1O100O00100O1O1O100O10000O10000000000O10000000000O100000000O1000dNQMgAo2Y>TMdAl2\\>WMaAi2^>[M_Ad2b>^M\\Ab2d>aMYA_2g>dMVA\\2i>hMTAX2l>kMQAU2o>mMo@R2R?oMm@Q2S?PNl@P2U?PNj@Q2V?oMi@Q2X?oMg@Q2Y?QNe@P2[?PNd@P2]?PNb@P2_?PN`@P2`?QN_@P2a?PN]@Q2d?PNZ@P2f?QNY@P2g?PNX@P2i?S110O0000001O00010O0000001O00001O00001O0000001O00001O2M3N2N2N2N3L3M3L4M3L4M2M3N2M3N101N2N2N101N2N2OO01OZKaBm2^=cLSC]3m<[L[Cd3e<VLbCj3]<oKkCQ4Q>10O100000O10O1000000O0100000O10O10O10O0L4M4KPLd@e3X?^Lj@a3n>fLTAZ3b>oL_AP3X>XMjAh2l=`MVB_2h=bM[B\\2e=dM\\B\\2d=bM^B]2b=bM`B]2`=bMbB^2^=aMcB^2]=aMeB^2[=aMgB_2Y=_MiB`2W=`MjB_2W=_MkBa2U=]MnBa2W=YMkBf2Y=VMhBj2\\=PMfBo2_=kLcBU3`=fLbBY3c=aL_B^3R?00O1O101N1O100O1O100O1O100O101N1O100O1OnDoLW5R3eJ\\MP5d2lJgMm4Y2SImLZNQ1_8R2VIRMSNT1c8j1YIXMlMU1g8c1[IeM_MR1Q9Y1^InMVMV1R9l0gIXNjLR1]9f0hIm0W6SOhIo0W6QOgIQ1Y6oNfIS1Y6nNeIT1Z6lNeIU1[6kNdIW1[6iNcIZ1\\6fNcI\\1]6cNcI]1]6dNbI]1]6cNdI\\1\\6dNdI]1Z6dNfI\\1X6fNhI[1V6gNjIY1T6iNeGhLZ1_4o6mN_GkLa1Y4n6oNTGSMo1m3l6KSI6j6MUI3j6NVI3h6OXI0f63YIMf64ZIMd65[IKc68]IGb6:^IF`6=_IC`6?_IA_6a0bI_O\\6c0cI]O\\6d0dI\\OZ6f0fIZOY6g0gIYOW6i0iIXOU6i0kIWOT6k0kIUOS6m0mISOR6n0nIROo5Q1QJPOk5T1TJlNi5W1XJhNe5\\1ZJdNc5_1]JaN`5c1_J]N^5f1bJZN[5i1eJWNX5m1gJRNW5Q2iJoMS5V2kJkMQ5Y2oJfMm4_2SKaMi4d2VK\\Me4i2ZKWMc4n2\\KRM_4S3aKlL\\4X3cKhLY2^KTNP8BaLZ2hKiMj7M^LZ2PL^Me77[L\\2XLRMa7a0VL]2bLgL[7l0RLPOeKj2W1nLU7W1PLlNkKl2j9WNZJjNRLm2d9YNYJfN[Lm2]9\\NXKb1i4]NZK`1f4`N\\K_OmJ=n94XKWO\\K0e9g0RKQOi6o0ZIiNk6W1P6N10001N101O1N2O1O1O0O2O1O1N2O1O1N2O1O1N2O2N3M3L4M2N_bl0"}, "label": "a white cow with the numbers 610 on its collar being led by a woman wearing a white coat"}]}
{"id": 493407, "image": "COCO_train2014_000000493407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman guiding the cow\"."}], "task": "refering", "answer_template": "The \"the woman guiding the cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 49, 50, 67, 84, 101, 116, 117, 118, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 235, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 303, 304, 305, 320, 321, 322, 337, 338, 339, 354, 355, 356, 371, 372, 373, 388, 389, 390], "bbox": [0.6954639175257732, 0.0641875, 1.0, 0.989859375], "iscrowd": 0, "area": 44051.32535, "rle": {"size": [640, 485], "counts": "b^c6:cc09G9_^OMf>=k@5j>2h@7V?[2O0O1O2N100O2O00001O001O00001O00001O001O001O001O0100O010O100QM[@a1f?YNc@d1^?VNk@e1V?UNSAh1n>RN[Ai1f>lMhAQ2W`0L2O2M2N2N3M2N2N3LdMgNTBX1d=TOXBl0h=XOUBf0l=^OlAf0S>@eAb0\\>C[Aa0e>DRA?n>Gi@<X?I_@;a?IX@9k?m1O010O010O01O02O2YHSLVOP4QNgNn1\\1YK[1g4P7N100O2N1O1O2O2M3M3M3N2M3M3M3M2O1N2N1O2N100O100O10000O10000O10000O10000O100O10000O100O1O1jLoA=Q>AYB6h=GbB1_=LlBKU=2VCFj<8_C@b<>dC]O]<b0fC[O[<e0hCXOX<g0kCUOW<j0mCPOV<o0oCjNT<U1QDeNQ<Z1TD_No;`1Y300O1O1O100O1O1O100O1O11O2POhNT_O[1m;f3oC[LS7g8iHYGY2i=cMb@"}, "label": "the woman guiding the cow"}]}
{"id": 493407, "image": "COCO_train2014_000000493407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt holding a leash attached to a cow\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt holding a leash attached to a cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 49, 50, 67, 84, 101, 116, 117, 118, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 235, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 303, 304, 305, 320, 321, 322, 337, 338, 339, 354, 355, 356, 371, 372, 373, 388, 389, 390], "bbox": [0.6954639175257732, 0.0641875, 1.0, 0.989859375], "iscrowd": 0, "area": 44051.32535, "rle": {"size": [640, 485], "counts": "b^c6:cc09G9_^OMf>=k@5j>2h@7V?[2O0O1O2N100O2O00001O001O00001O00001O001O001O001O0100O010O100QM[@a1f?YNc@d1^?VNk@e1V?UNSAh1n>RN[Ai1f>lMhAQ2W`0L2O2M2N2N3M2N2N3LdMgNTBX1d=TOXBl0h=XOUBf0l=^OlAf0S>@eAb0\\>C[Aa0e>DRA?n>Gi@<X?I_@;a?IX@9k?m1O010O010O01O02O2YHSLVOP4QNgNn1\\1YK[1g4P7N100O2N1O1O2O2M3M3M3N2M3M3M3M2O1N2N1O2N100O100O10000O10000O10000O10000O100O10000O100O1O1jLoA=Q>AYB6h=GbB1_=LlBKU=2VCFj<8_C@b<>dC]O]<b0fC[O[<e0hCXOX<g0kCUOW<j0mCPOV<o0oCjNT<U1QDeNQ<Z1TD_No;`1Y300O1O1O100O1O1O100O1O11O2POhNT_O[1m;f3oC[LS7g8iHYGY2i=cMb@"}, "label": "a woman in a white shirt holding a leash attached to a cow"}]}
{"id": 18276, "image": "COCO_train2014_000000018276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a portion of a multi - decker sandwich that is to the left of some french fries\"."}], "task": "refering", "answer_template": "The \"a portion of a multi - decker sandwich that is to the left of some french fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 184, 185], "bbox": [0.0, 0.21508196721311476, 0.33209374999999997, 0.568688524590164], "iscrowd": 0, "area": 16223.428099999997, "rle": {"size": [427, 640], "counts": "g48h1M]9b0kEIT:e0TEFj:^1O2O1N2N1O2O1N1O2O1N2N101N2N2O0O2N2O1N1O2O1N1O2N2O1N1O2O0O2N01O0O101O1O0O2O1O1O1N2O1O1O1N101O1O0O2O1O001N2O001O001O1O001O1O00100O001O001O1O001O1O001O1O001O001O1O010000O01000O10000O010O10000O10O10O10000O100O01000O100000O2O000000000O1000000000000O10000000000O10000000000O1000000000O01000O10000O1000O0100O10000O1000O010000O10000O10O10O10000O1001O2M3N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N1N3M2O2M2N3M2O1NQZb5"}, "label": "a portion of a multi - decker sandwich that is to the left of some french fries"}]}
{"id": 18276, "image": "COCO_train2014_000000018276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bread in the left most side of the image\"."}], "task": "refering", "answer_template": "The \"bread in the left most side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 184, 185], "bbox": [0.0, 0.21508196721311476, 0.33209374999999997, 0.568688524590164], "iscrowd": 0, "area": 16223.428099999997, "rle": {"size": [427, 640], "counts": "g48h1M]9b0kEIT:e0TEFj:^1O2O1N2N1O2O1N1O2O1N2N101N2N2O0O2N2O1N1O2O1N1O2N2O1N1O2O0O2N01O0O101O1O0O2O1O1O1N2O1O1O1N101O1O0O2O1O001N2O001O001O1O001O1O00100O001O001O1O001O1O001O1O001O001O1O010000O01000O10000O010O10000O10O10O10000O100O01000O100000O2O000000000O1000000000000O10000000000O10000000000O1000000000O01000O10000O1000O0100O10000O1000O010000O10000O10O10O10000O1001O2M3N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N1N3M2O2M2N3M2O1NQZb5"}, "label": "bread in the left most side of the image"}]}
{"id": 18276, "image": "COCO_train2014_000000018276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of sandwich to the left of the fries\"."}], "task": "refering", "answer_template": "The \"the piece of sandwich to the left of the fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284], "bbox": [0.023984375, 0.4, 0.47678125000000005, 0.8359484777517565], "iscrowd": 0, "area": 39809.3176, "rle": {"size": [427, 640], "counts": "ko62W=3N1N2N2K6J5K5K5L5J5K5K6K4K5K5K6J5L4K5K6J5L4O1O2M2O1O2N1O1O1O2N1O1O1O2M2O1O2N1O1O1O2N1O1O1N2O2N1O1O1O1O1O1O1O1N2O1O1O1O2N1O1O100O10000000000O101O00000O10000000000O1000001O0O10000000000O100000000O2O000000000O10000000000O101O00000O10000000000O1000001O0O10000000000O100000000O2O000000000O100000000O101O0000000O100000000O1000001O000O10000000000O100000001N10000000000O1000000000001O01O000000001N1O100O1O101N1O100O1O2O0O1O100O1O2O0O1O1O101N1O100O1O101N1O100O1O2O2M2N3N1N3M2O2M2N2O2M2N3N1N3M2O2M2N3M2O2M2N2O2M2N3N1N3M2O2M2N3N1N2N3N1N3M2O2M2N3N1N3M2O2M2N2N3N1N3M2O2M2N3N1N3M3N6I8H7J7HV][4"}, "label": "the piece of sandwich to the left of the fries"}]}
{"id": 18276, "image": "COCO_train2014_000000018276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of layered sandwich in front of other pieces\"."}], "task": "refering", "answer_template": "The \"a piece of layered sandwich in front of other pieces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284], "bbox": [0.023984375, 0.4, 0.47678125000000005, 0.8359484777517565], "iscrowd": 0, "area": 39809.3176, "rle": {"size": [427, 640], "counts": "ko62W=3N1N2N2K6J5K5K5L5J5K5K6K4K5K5K6J5L4K5K6J5L4O1O2M2O1O2N1O1O1O2N1O1O1O2M2O1O2N1O1O1O2N1O1O1N2O2N1O1O1O1O1O1O1O1N2O1O1O1O2N1O1O100O10000000000O101O00000O10000000000O1000001O0O10000000000O100000000O2O000000000O10000000000O101O00000O10000000000O1000001O0O10000000000O100000000O2O000000000O100000000O101O0000000O100000000O1000001O000O10000000000O100000001N10000000000O1000000000001O01O000000001N1O100O1O101N1O100O1O2O0O1O100O1O2O0O1O1O101N1O100O1O101N1O100O1O2O2M2N3N1N3M2O2M2N2O2M2N3N1N3M2O2M2N3M2O2M2N2O2M2N3N1N3M2O2M2N3N1N2N3N1N3M2O2M2N3N1N3M2O2M2N2N3N1N3M2O2M2N3N1N3M3N6I8H7J7HV][4"}, "label": "a piece of layered sandwich in front of other pieces"}]}
{"id": 18276, "image": "COCO_train2014_000000018276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a triangle sandwich portion to the bottom right\"."}], "task": "refering", "answer_template": "The \"a triangle sandwich portion to the bottom right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 85, 86, 87, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.5965625, 0.18355971896955503, 0.9992812499999999, 0.7105854800936767], "iscrowd": 0, "area": 37492.55849999999, "rle": {"size": [427, 640], "counts": "_ao42X=3M2N3M2N3M3L3N3M2N3M3M2N3M2N3ROiNcEY1[:PO\\ES1a:WOTEm0h:m0N3M2M4M3M2N3L2O1O1N2O1O1N2O1O1N2O1VOcLeG^3Z8cLeG^3Z8dLdG]3Z8fLeGZ3Z8gLeGZ3Z8hLeGX3Y8kLeGV3Z8kLfGU3Y8mLeGT3Z8nLeGR3Y8QMeGP3Z8QMfGn2Z8TMdGm2Z8VMeGj2Z8WMeGj2Z8XMeGh2Z8ZMdGg2Z8\\MeGd2Z8W1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O0O2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N101O1N2O1O1N21O1O001O1O1O001O1O1O001O1O1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1O001N2O1O001O1O1N101O1O1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1O001N2O1O001O1O1N101O1O1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1N101O1N2O001N2O1O0O2O1O0O2O1O1N101O1N2O001N2O1O0O2O1O1N101O1N2O001N2O1O0O2O1O1N101O1N2O001N2O1O0O2O1O1N101O1mNS1iNUK"}, "label": "a triangle sandwich portion to the bottom right"}]}
{"id": 173925, "image": "COCO_train2014_000000173925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis player behind the one in foremost front\"."}], "task": "refering", "answer_template": "The \"the tennis player behind the one in foremost front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 28, 29, 37, 38, 45, 46, 47, 53, 54, 55, 61, 62, 69, 70, 77, 78, 85, 86, 93, 94, 98, 99, 102, 106, 107, 110, 113, 114, 117, 118, 122, 125, 126, 127], "bbox": [0.16792207792207792, 0.11668, 0.901038961038961, 0.8777999999999999], "iscrowd": 0, "area": 14751.870699999996, "rle": {"size": [500, 231], "counts": "k]c04^?6J6J3M3M3O1O1O001O1O0010O0001O001O00010O00001O00001N1O1O2O0O1N2L4K5K6K4K5L4K5K5L4K6J5L4001UO]CcNd<T1iCgNX<Q1TDjNm;S1YDhNj;T1]DfNi;S1b1I7I>Bk`8F`_G1N2O2N1N3N1O1N3]OEjA<V>GfA:Y>JbA9]>?O2O0O1O2O0O1010O01O00001O00001O00001O001O000100OmD4^6LZIj0Y6VO_I_1T6aNdIm1V6TNaIU2]6lMZI]2]2VL0^1ZMd2e2QLO\\1ZMe2d2QL2[1ZMc2c2SLl0b0aLZ3a2ULP1a0_L\\3[2VLV1?_L]3V2UL\\1?^L^3Q2ULa1>^L_3l1TLg1>]L_3h1TLm1=[L_3e1VLP2=ZL^3b1VLU2?XLY1dNDl2fNX2?WLQ1kNJc2gN\\2a0TLg0SO2Z2gN_2c0SL=[O9P2hNc2d0QL4C`0f1hNg2h0nKIKi0\\1gNl2i0lK@3P1R1hNo2k0kKVO:X1h0iNT3k0iKbNl0m10iN\\3j0mMc0`NeNc3k0iMg0[N`Nn3k0cMk0WN]NV4j0`MS1oMUNb4j0[M]1eMlMP5j0WMe1]McM]5j0RMS3o2SMjLn2W3XMbLh2^3^MZLd2g3bMoKa2R4eMdK^2\\4mMVKV2k4UNfJn1Z5V41O0mK`JRNa5i1iJQNW5k1SKoMn4l1]KmMe4m1eKlM^4W1cLmLSNL\\5l2PMnLPNNS5h2[MRMgM4P5g1^NnMgL9m4e1bNkMfL>k4b1eNiMeLc0h4`1:^NI^1;_NKY19eNMR18lNNk07QO1f0f7GUR;"}, "label": "the tennis player behind the one in foremost front"}]}
{"id": 173925, "image": "COCO_train2014_000000173925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white hat and black shoes holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man wearing a white hat and black shoes holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 18, 19, 20, 26, 27, 28, 29, 34, 35, 36, 37, 42, 43, 44, 45, 49, 50, 51, 52, 53, 54, 57, 58, 59, 60, 61, 62, 65, 66, 67, 68, 69, 70, 73, 74, 75, 76, 77, 78, 82, 83, 84, 85, 90, 91, 92, 93, 99, 100, 101, 107, 108, 109, 115, 116, 117, 123, 124, 125, 131, 132, 133], "bbox": [0.1605194805194805, 0.09214, 0.8074458874458874, 0.9303400000000001], "iscrowd": 0, "area": 33352.34715, "rle": {"size": [500, 231], "counts": "iYb02_?6J5K6G9F9G:I7K4K4L1O1O2N1O2L3LZOSCoNj<S1cCbNX<a1j02oNR1dEbMd7]2YHjMd7U2YHSNc7l1nFaN]OJb9e1`F<]9P3L3M3M2XNkIPJV6o5^J\\Id5d6`JWIb5g6bJTI`5k6eJPI\\5P7gJkH[5T7jJgHW5X7c100O1O101N1O1O100O1O2O2N3M3M2O2M2N3M<jLaF_On90TF0Q:HRF9R:@oEa0U:XOnEh0W:QOkEo0Y:jNjEV1Z:cNhE^1\\:[NgEf1\\:SNgEo1[:jMhEW2[:_MjEc2c;2N1O2M2O001N2O001O0O2O1O0`DdLh:]3TEfLk:[3SEhLl:Y3QEiLo:X3nDkLQ;U3mDmLS;T3jDnLW;Q3gDRMX;e310O01O010O10O010O01O010O010O00100O010O010O\\KTL`Ml3Z2mLTMS3e2aMmL_2m2lM_KnMIW4n2I_LhL?a3]2`0gLZLi0W3l1V1PMmKQ1m2f1S3VNoLa1\\3\\NeL[1f3bN[LU1Q4gNPLT1W4hNjKZ1W4cNjK_1V4^NkKc1V4YNlKi1T4TNmKm1T4PNlKQ2fMPNg8Q2VGPNk8Q2RGQNS9k1jFVNc9_1YFdNj9Y1TFiNn9V1lEROS:m0gE[OX:d0bED]:h0iDC\\;X26I7I7J6I9G9H8G7I7H8F:F:F9G:FQXe0"}, "label": "a man wearing a white hat and black shoes holding a tennis racket"}]}
{"id": 173925, "image": "COCO_train2014_000000173925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player wearing a white hat\"."}], "task": "refering", "answer_template": "The \"a tennis player wearing a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 18, 19, 20, 26, 27, 28, 29, 34, 35, 36, 37, 42, 43, 44, 45, 49, 50, 51, 52, 53, 54, 57, 58, 59, 60, 61, 62, 65, 66, 67, 68, 69, 70, 73, 74, 75, 76, 77, 78, 82, 83, 84, 85, 90, 91, 92, 93, 99, 100, 101, 107, 108, 109, 115, 116, 117, 123, 124, 125, 131, 132, 133], "bbox": [0.1605194805194805, 0.09214, 0.8074458874458874, 0.9303400000000001], "iscrowd": 0, "area": 33352.34715, "rle": {"size": [500, 231], "counts": "iYb02_?6J5K6G9F9G:I7K4K4L1O1O2N1O2L3LZOSCoNj<S1cCbNX<a1j02oNR1dEbMd7]2YHjMd7U2YHSNc7l1nFaN]OJb9e1`F<]9P3L3M3M2XNkIPJV6o5^J\\Id5d6`JWIb5g6bJTI`5k6eJPI\\5P7gJkH[5T7jJgHW5X7c100O1O101N1O1O100O1O2O2N3M3M2O2M2N3M<jLaF_On90TF0Q:HRF9R:@oEa0U:XOnEh0W:QOkEo0Y:jNjEV1Z:cNhE^1\\:[NgEf1\\:SNgEo1[:jMhEW2[:_MjEc2c;2N1O2M2O001N2O001O0O2O1O0`DdLh:]3TEfLk:[3SEhLl:Y3QEiLo:X3nDkLQ;U3mDmLS;T3jDnLW;Q3gDRMX;e310O01O010O10O010O01O010O010O00100O010O010O\\KTL`Ml3Z2mLTMS3e2aMmL_2m2lM_KnMIW4n2I_LhL?a3]2`0gLZLi0W3l1V1PMmKQ1m2f1S3VNoLa1\\3\\NeL[1f3bN[LU1Q4gNPLT1W4hNjKZ1W4cNjK_1V4^NkKc1V4YNlKi1T4TNmKm1T4PNlKQ2fMPNg8Q2VGPNk8Q2RGQNS9k1jFVNc9_1YFdNj9Y1TFiNn9V1lEROS:m0gE[OX:d0bED]:h0iDC\\;X26I7I7J6I9G9H8G7I7H8F:F:F9G:FQXe0"}, "label": "a tennis player wearing a white hat"}]}
{"id": 223078, "image": "COCO_train2014_000000223078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with snow slidder and bending to pose for a picture\"."}], "task": "refering", "answer_template": "The \"a man with snow slidder and bending to pose for a picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 166, 167, 168, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 263, 264, 276, 277, 279, 280, 281, 292, 293, 295, 296, 297, 308, 309, 312, 313, 324, 325, 328, 329, 340, 341, 344], "bbox": [0.2521428571428571, 0.40449999999999997, 0.5976839826839827, 0.9393281250000001], "iscrowd": 0, "area": 30920.506350000003, "rle": {"size": [640, 462], "counts": "aPY21lc08H9H8G9G8H8H9H5J5L4K5L5K4cK^MUGg2f8_MUGe2g8aMRGe2i8`MRGb2k8dMPG_2m8gMlF]2Q9hMjF[2R9kMiFX2T9nMeFV2X9oMcFT2Z9RN_FQ2^9UN\\Fo1a9WNVFo1g9VNRFm1l9YNlEl1R:ZNeEj1Z:[N_Ei1_:]NXEg1g:^NQEf1P;^NhDf1X;`3010O1O1J6G8I8lNT1M3M3L3N3L4M3L4M2N3L4M3L4M2M3N2N2M3N2M3N2M3N2001O1O1O2N1O1K5H8G9M300O101O0O100O10000O101N10000O10000O100O2O000O100O10000O2O00000001O00XCiIg;W6oCVJn;k5gCcJU<g6RF^HY7e7_HbH^7b7YHeHe7^7THiHi7[7oGkHP8W7hGQIU8S7ZG[Ie8P90O101N100O]1dN001N2O0O2O1N101O1N101N2O0O2M3M2N3UJmEf1T:YNRF`1Q:^NUF[1n9cNXFV1j9iN\\FP1g9nN_Fl0c9ROcFg0_9XOgFa0\\9\\OkF=X9@oF9S9ETG4o8JWG0k8N[GIi86]G@i8>]GXOi8f0\\GQOi8n0]GhNi8V1]G`NR9U1ZRd3"}, "label": "a man with snow slidder and bending to pose for a picture"}]}
{"id": 223078, "image": "COCO_train2014_000000223078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with skiis poles\"."}], "task": "refering", "answer_template": "The \"man with skiis poles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 166, 167, 168, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 263, 264, 276, 277, 279, 280, 281, 292, 293, 295, 296, 297, 308, 309, 312, 313, 324, 325, 328, 329, 340, 341, 344], "bbox": [0.2521428571428571, 0.40449999999999997, 0.5976839826839827, 0.9393281250000001], "iscrowd": 0, "area": 30920.506350000003, "rle": {"size": [640, 462], "counts": "aPY21lc08H9H8G9G8H8H9H5J5L4K5L5K4cK^MUGg2f8_MUGe2g8aMRGe2i8`MRGb2k8dMPG_2m8gMlF]2Q9hMjF[2R9kMiFX2T9nMeFV2X9oMcFT2Z9RN_FQ2^9UN\\Fo1a9WNVFo1g9VNRFm1l9YNlEl1R:ZNeEj1Z:[N_Ei1_:]NXEg1g:^NQEf1P;^NhDf1X;`3010O1O1J6G8I8lNT1M3M3L3N3L4M3L4M2N3L4M3L4M2M3N2N2M3N2M3N2M3N2001O1O1O2N1O1K5H8G9M300O101O0O100O10000O101N10000O10000O100O2O000O100O10000O2O00000001O00XCiIg;W6oCVJn;k5gCcJU<g6RF^HY7e7_HbH^7b7YHeHe7^7THiHi7[7oGkHP8W7hGQIU8S7ZG[Ie8P90O101N100O]1dN001N2O0O2O1N101O1N101N2O0O2M3M2N3UJmEf1T:YNRF`1Q:^NUF[1n9cNXFV1j9iN\\FP1g9nN_Fl0c9ROcFg0_9XOgFa0\\9\\OkF=X9@oF9S9ETG4o8JWG0k8N[GIi86]G@i8>]GXOi8f0\\GQOi8n0]GhNi8V1]G`NR9U1ZRd3"}, "label": "man with skiis poles"}]}
{"id": 223078, "image": "COCO_train2014_000000223078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child in yellow dress posing while skeing with a man\"."}], "task": "refering", "answer_template": "The \"a child in yellow dress posing while skeing with a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 218, 219, 220, 221, 233, 234, 235, 236, 237, 249, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 282, 283, 284, 285, 298, 299, 300, 301, 314, 315, 316, 317, 330, 331, 332, 333, 346, 347, 348, 349, 362, 364], "bbox": [0.5834415584415584, 0.550671875, 0.907922077922078, 0.989859375], "iscrowd": 0, "area": 22155.416100000006, "rle": {"size": [640, 462], "counts": "eTY51oc0:Ee0\\O0O100O100O2O0O100O100O100O100O2TCgNo6Y1mHkNR7V1jHoNU7Q1gHSOX7n0eHUO[7k0aHYO^7h0^H\\Ob7d0ZH@e7a0^EVO2=`:=SELNKn:;gDa0LXOV;>cDo0IgN^;`0_D^1DVNg;b0[D]3_;iLXD`3f;^2O2O0O10000OO2N101N2N101N101N2N101N101N2N101N1O2L4D;E<G8K7H8I6J7H7J7I7H7J7K401000^ATLn<l3TBWMh=e4PDQJe9T6iEYJX:k5WEaJj:d5cDiJ];m61O1O0O2O001O1O001O0O2O001O001O0M4M2N3M2M4M2N3M2M4M2N3M2M4oKSDKo;BoD1T;NXEDj:=`E[N^;d1_313L4M3M2M4M2M1N101N2O1N2O0O2O1N2O0O2O1N2O0O2O1N2O1N101N2O1N1M4L4L4L3M4Lj^j0"}, "label": "a child in yellow dress posing while skeing with a man"}]}
{"id": 223078, "image": "COCO_train2014_000000223078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child skiier in a yellow jacket with black pants\"."}], "task": "refering", "answer_template": "The \"a young child skiier in a yellow jacket with black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 218, 219, 220, 221, 233, 234, 235, 236, 237, 249, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 282, 283, 284, 285, 298, 299, 300, 301, 314, 315, 316, 317, 330, 331, 332, 333, 346, 347, 348, 349, 362, 364], "bbox": [0.5834415584415584, 0.550671875, 0.907922077922078, 0.989859375], "iscrowd": 0, "area": 22155.416100000006, "rle": {"size": [640, 462], "counts": "eTY51oc0:Ee0\\O0O100O100O2O0O100O100O100O100O2TCgNo6Y1mHkNR7V1jHoNU7Q1gHSOX7n0eHUO[7k0aHYO^7h0^H\\Ob7d0ZH@e7a0^EVO2=`:=SELNKn:;gDa0LXOV;>cDo0IgN^;`0_D^1DVNg;b0[D]3_;iLXD`3f;^2O2O0O10000OO2N101N2N101N101N2N101N101N2N101N1O2L4D;E<G8K7H8I6J7H7J7I7H7J7K401000^ATLn<l3TBWMh=e4PDQJe9T6iEYJX:k5WEaJj:d5cDiJ];m61O1O0O2O001O1O001O0O2O001O001O0M4M2N3M2M4M2N3M2M4M2N3M2M4oKSDKo;BoD1T;NXEDj:=`E[N^;d1_313L4M3M2M4M2M1N101N2O1N2O0O2O1N2O0O2O1N2O0O2O1N2O1N101N2O1N1M4L4L4L3M4Lj^j0"}, "label": "a young child skiier in a yellow jacket with black pants"}]}
{"id": 468836, "image": "COCO_train2014_000000468836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with yellow shirt\"."}], "task": "refering", "answer_template": "The \"a man with yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 219, 220, 221, 242, 243, 244, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338], "bbox": [0.475078125, 0.5831294117647059, 0.7152499999999999, 1.0], "iscrowd": 0, "area": 17605.199, "rle": {"size": [425, 640], "counts": "Zan35R=7H9G9G9G6K4K5K5K5L4K5K5K5L2M3M3M3M3N200O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O1RO]LVHc3d7fLXH[3c7mLZHS3`7VM\\Hk2c7YMZHg2e7]MWHd2g7aMUH`2j7dMSH\\2l7a1O1O1N2O1O1000000000000000000000000O1000000000000000000000000001O2N2N2N2N2N2N2PLlGg2V8TMoGj2R8QMTHm2n7mLXHQ3j7jL\\HS3f7gL`HW3b7dLcHZ3f80001O0000001O00001O0000001O0000001O002N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3gNSDP1o;lN]Dj0[<F=CQT[2"}, "label": "a man with yellow shirt"}]}
{"id": 468836, "image": "COCO_train2014_000000468836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in yellow t shirt wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man in yellow t shirt wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 219, 220, 221, 242, 243, 244, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338], "bbox": [0.475078125, 0.5831294117647059, 0.7152499999999999, 1.0], "iscrowd": 0, "area": 17605.199, "rle": {"size": [425, 640], "counts": "Zan35R=7H9G9G9G6K4K5K5K5L4K5K5K5L2M3M3M3M3N200O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O1RO]LVHc3d7fLXH[3c7mLZHS3`7VM\\Hk2c7YMZHg2e7]MWHd2g7aMUH`2j7dMSH\\2l7a1O1O1N2O1O1000000000000000000000000O1000000000000000000000000001O2N2N2N2N2N2N2PLlGg2V8TMoGj2R8QMTHm2n7mLXHQ3j7jL\\HS3f7gL`HW3b7dLcHZ3f80001O0000001O00001O0000001O0000001O002N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3gNSDP1o;lN]Dj0[<F=CQT[2"}, "label": "a man in yellow t shirt wearing glasses"}]}
{"id": 468836, "image": "COCO_train2014_000000468836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy is flying his kite\"."}], "task": "refering", "answer_template": "The \"the boy is flying his kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 166, 167, 168, 169, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.15140625000000002, 0.30416470588235295, 0.43171875000000004, 0.9858823529411764], "iscrowd": 0, "area": 36418.55199999999, "rle": {"size": [425, 640], "counts": "jaX16l<<C=D;E<DV1iN8N2O1N2O2M2N2Oj1UN2O1N2N2O100O100O1O100O100O100O1O1O1O100O1O1O1O1O1O100O1O1nMoIkLR6T3nIlLS6R3oImLR6R3nInLS6Q3nIoLR6o2PJPMQ6o2oIQMR6n2oIQMR6m2oISMR6l2SKoKn4o3WKmKj4R4XKlKi4S4YKkKh4S4\\KjKe4U4]KiKd4V4^KhKc4V4`KhKa4W4bKfK_4Y4cKfK]4Z4Y2O1O100O1O100O1O10000000000001O0000000000000000000000000000000000001O0000000000000000000000002N1O1O2N1O2N1O2N1O1O2oHnKg4S4VKoKj4S4SKoKl4R4QKPLo4R4nJPLQ5Q4lJQLT5P4hJTLW5n3cJVL]5k3^JYLb5i3XJ[Lh5c50000000000000001O001O00001O001O001O001O001O001O002N1O2VLdHc1^7\\NfH_1[7`NiH\\1Y7cNjHY1X7eNlHW1U7hNoHWO9Kj6m0SKnNP5o0TKmNn4Q1VKkNl4S1YKhNh4W1\\KeNf4Y1_KbNc4\\1f3O2M2iN]De0T<N3N2N1N3N2N5Jaif4"}, "label": "the boy is flying his kite"}]}
{"id": 468836, "image": "COCO_train2014_000000468836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a brown shirt is looking at the sky\"."}], "task": "refering", "answer_template": "The \"a boy in a brown shirt is looking at the sky\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 166, 167, 168, 169, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.15140625000000002, 0.30416470588235295, 0.43171875000000004, 0.9858823529411764], "iscrowd": 0, "area": 36418.55199999999, "rle": {"size": [425, 640], "counts": "jaX16l<<C=D;E<DV1iN8N2O1N2O2M2N2Oj1UN2O1N2N2O100O100O1O100O100O100O1O1O1O100O1O1O1O1O1O100O1O1nMoIkLR6T3nIlLS6R3oImLR6R3nInLS6Q3nIoLR6o2PJPMQ6o2oIQMR6n2oIQMR6m2oISMR6l2SKoKn4o3WKmKj4R4XKlKi4S4YKkKh4S4\\KjKe4U4]KiKd4V4^KhKc4V4`KhKa4W4bKfK_4Y4cKfK]4Z4Y2O1O100O1O100O1O10000000000001O0000000000000000000000000000000000001O0000000000000000000000002N1O1O2N1O2N1O2N1O1O2oHnKg4S4VKoKj4S4SKoKl4R4QKPLo4R4nJPLQ5Q4lJQLT5P4hJTLW5n3cJVL]5k3^JYLb5i3XJ[Lh5c50000000000000001O001O00001O001O001O001O001O001O002N1O2VLdHc1^7\\NfH_1[7`NiH\\1Y7cNjHY1X7eNlHW1U7hNoHWO9Kj6m0SKnNP5o0TKmNn4Q1VKkNl4S1YKhNh4W1\\KeNf4Y1_KbNc4\\1f3O2M2iN]De0T<N3N2N1N3N2N5Jaif4"}, "label": "a boy in a brown shirt is looking at the sky"}]}
{"id": 83815, "image": "COCO_train2014_000000083815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a passenger train with blue and white paint reflecting lights\"."}], "task": "refering", "answer_template": "The \"a passenger train with blue and white paint reflecting lights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 276, 277], "bbox": [0.0015, 0.38875878220140514, 0.45128124999999997, 0.8763934426229509], "iscrowd": 0, "area": 34270.6492, "rle": {"size": [427, 640], "counts": "je0W3l6W3010O100O100O101N100O100O1O100O100O101N100O100O100O100O100O2N100O100O100O100O100O2O0O100O1O100O100O100O101N100O100O100O100O1O101N100O100O100O100O100O2O0O100O1O100O100O100O101N100O100O100O100O1O101N100O100O100O100O100O2O0O1O100O100O100O100O2O0O100O100O100O1O100O101N100O100O100O100O100O2O0O1O100O100O100O100O2O0O100O100O100O1O100O101N100O100O100O100O100O2O0O1O100O100O100O100O2O0O100O100O1O100O100O2O0O100O100O100O100O100O2N100O100O100O100O100O2O0O100O100O1O100O100O2O0O100O100O100O100O100O2N100O100O100O100O100O2O0O100O1O100O100O100O2O0O100O100O100O100O1O2O0O100O100O101Nf0[OPUb4"}, "label": "a passenger train with blue and white paint reflecting lights"}]}
{"id": 83815, "image": "COCO_train2014_000000083815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lit up train to the left of another train\"."}], "task": "refering", "answer_template": "The \"a lit up train to the left of another train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 276, 277], "bbox": [0.0015, 0.38875878220140514, 0.45128124999999997, 0.8763934426229509], "iscrowd": 0, "area": 34270.6492, "rle": {"size": [427, 640], "counts": "je0W3l6W3010O100O100O101N100O100O1O100O100O101N100O100O100O100O100O2N100O100O100O100O100O2O0O100O1O100O100O100O101N100O100O100O100O1O101N100O100O100O100O100O2O0O100O1O100O100O100O101N100O100O100O100O1O101N100O100O100O100O100O2O0O1O100O100O100O100O2O0O100O100O100O1O100O101N100O100O100O100O100O2O0O1O100O100O100O100O2O0O100O100O100O1O100O101N100O100O100O100O100O2O0O1O100O100O100O100O2O0O100O100O1O100O100O2O0O100O100O100O100O100O2N100O100O100O100O100O2O0O100O100O1O100O100O2O0O100O100O100O100O100O2N100O100O100O100O100O2O0O100O1O100O100O100O2O0O100O100O100O100O1O2O0O100O100O101Nf0[OPUb4"}, "label": "a lit up train to the left of another train"}]}
{"id": 92009, "image": "COCO_train2014_000000092009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a polar bear seeing over the railing with small black eyes\"."}], "task": "refering", "answer_template": "The \"a polar bear seeing over the railing with small black eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 284, 285, 286, 287, 288, 300, 301, 302, 303, 307, 308, 309, 310, 311, 323, 324, 325, 326, 330, 331, 332, 333, 334], "bbox": [0.0656875, 0.39274004683840746, 0.8884374999999999, 1.0], "iscrowd": 0, "area": 67138.58685, "rle": {"size": [427, 640], "counts": "fma03R=6J6J6J6J6K5J6J6J6J6J6J6]OgMoE^2o9cMRF]2l9eMTF[2i9gMXFY2f9iMZFW2d9jM]FV2`9mM`FS2^9oMbFQ2\\9PNeFP2X9SNhFm1V9TNkFl1S9VNmFj1P9XNQGh1m8ZNSGf1k8\\NUGd1h8^NXGc1f8_NZGa1c8aN^G_1`8cN`G]1^8eNbG[1Z8iNfGW1R8QOnGo0k7XOUHh0d7^O]Hb0[7FeH:T7MlH3l65TIKe6<[ID]6d0cI\\OV6k0jIUOo5R1QJnNg5Z1YJfN`5a1`J_NX5i1hJWNQ5o1PKQNh4W2XKiMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2YKhMg4W2ZKiMf4W2ZKiMf4W2ZKiMf4W2ZKUM[OfN[5U4ZKUM[OfN[5U4ZKUM\\OeNZ5V4ZKUM\\OeNZ5V4ZKUM]OdNY5W4ZKUM]OdNY5W4ZKUM]OdNY5W4ZKUM^OcNX5W4[KVM]OcNX5W4[KVM^ObNW5X4[KVM^ObNW5X4[KVM_OaNV5Y4[KVM_OaNV5Y4[KVM@`NU5Z4[KRL^O62^OU5Z4[KeKKc0E^OU5Z4YLXLcN]OT5[4ZLWLbN^OT5[4ZLWLcN[OU5]4YLXLbNZOV5^4YLWLbNXOY5_4ULYLbNVO_5]4oK\\LcNVOb5[4lK^Le4^3[KbLi4Z3XKeLk4X3UKhLo4T3QKkLT5Q3mJnLV5n2kJRMY5j2gJVM]5f2dJYM_5d2aJ[Md5a2\\J_Mh5]2YJbMj5[2VJeMh5]2ZJaMQ4T4c201O0O2O001O001O001O001O000O4M4L4L4L3M4L4L4L4L3M4K`NmLnHo2R7UMnHg2R7^MlH_2T7fMkHV2U7nMkHn1V7VNhHg1X7]NdHc1\\7bN^H_1b7eNYH\\1g7iNTHW1l7mNoGS1S8QOgGP1Y8TObGm0^8XO]Gh0c8P20000001O000O1000001O0000000000001N1000000000001O0000000O10001O0000000000001O0O10000000001O000000000O101O00kIPLo2P4nLVLo2j3nL]Ln2c3oL`LQ3`3lLdLS3\\3kLgLU3X3hLkLX3T3mJnKS1R1o3P3nJRLl0Q1V4m2nJVLe0Q1\\4i2oJZL>Q1b4e2PK^L7P1i4b2oJbLLV1U5W2oJgL_O[1b5n1oJnNQ5R1oJnNQ5R1oJnNQ5R1oJnNQ5Q1oJPOQ5P1oJPOR5o0nJQOR5o0nJQOR5o0nJQOR5o0nJQOR5o0mJROS5n0mJROS5n0mJROT5m0lJSOT5m0lJSOT5l0lJUOT5k0lJUOT5k0lJUOT5k0lJUOT5k0lJUOU5j0kJVOU5j0jJWOV5i0jJWOV5i0jJWOV5i0jJWOV5i0jJWOV5h0kJXOV5g0iJZOW5f0iJZOW5f0iJZOW5f0iJZOW5f0iJZOW5f0iJZOW5f0hJ[OX5e0hJ[OX5e0hJ[OX5e0hJ[OX5d0iJ\\OW5d0hJ]OW5d0iJ\\OW5d0iJ\\OW5d0iJ\\OW5d0iJ\\OW5d0iJ\\OW5d0hJ]OX5c0hJ]OW5d0iJ\\OW5d0iJ[OX5e0hJYOZ5f0gJWO\\5i0cJUO`5k0`JSOb5m0^JPOd5Q1\\JmNf5S1ZJjNi5V1VJiNl5W1TJfNo5Z1QJcNR6]1nIaNT6_1lI_NU6a1lI^NU6b1jI_N>fMj4k3hJ_N;iMm4h3hJ_N8lMP5e3hJ_N5PNR5a3iJ_N2SNT5_3jJ^N0UNV5]3iJ_NNWNY5Z3iJ_NKZN\\5W3iJ_NH]N_5T3iJ^NFaNa5P3jJ_NCcNc5n2jJ_OU5b0jJ_OV5a0jJ_OV5a0jJ_OV5a0jJ_OV5a0jJ_OV5a0iJ@W5`0iJ@V5a0jJ_OV5a0jJ_OV5`0kJ@U5`0kJ@U5`0jJAV5?jJAV5?jJAU5`0kJ@U5`0kJ@U5`0kJ@U5`0jJAV5?jJAV5?jJAV5>kJBT5?lJAU5>kJBU5>jJDU5<kJDV5;jJEV5;jJEW5:iJFW5:hJeNH_N`5l2hJXN8iNQ5n2gJkMi0TO`4Q3gJ^MY1^OQ4R3gJSMi1G`3V3gJlLS2KW3X3eJkLZ2JQ3[3eJhL`2Jk2^3eJeLg2Ie2a3dJbLn2J^2d3dJ_Lk8`3UG]Ln8c3QG[LR9e3;O0O2E:[Oe0[Of0ZOe0ZOWi`0GQU_Of201000O100O100O010O100O100O10O0100O10000O10O0100O100O100O010000000O10O1000000000000O1000000001O00001O001O00001O00001O001O00001N101O01O01O00001O0010O0001O000010O01O00001O0010O0001O001O01O01O001O000010O01O00001O0010O0001O001O01O01O001O002N100O1O1O1O1O1O1O2N001N2O1O1O1O1N101O1O1N2O1O1N1O2M3M3N2M3N2M3N1N3M3N2M3N2M3N1NTim0"}, "label": "a polar bear seeing over the railing with small black eyes"}]}
{"id": 92009, "image": "COCO_train2014_000000092009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"polar bear\"."}], "task": "refering", "answer_template": "The \"polar bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 284, 285, 286, 287, 288, 300, 301, 302, 303, 307, 308, 309, 310, 311, 323, 324, 325, 326, 330, 331, 332, 333, 334], "bbox": [0.0656875, 0.39274004683840746, 0.8884374999999999, 1.0], "iscrowd": 0, "area": 67138.58685, "rle": {"size": [427, 640], "counts": "fma03R=6J6J6J6J6K5J6J6J6J6J6J6]OgMoE^2o9cMRF]2l9eMTF[2i9gMXFY2f9iMZFW2d9jM]FV2`9mM`FS2^9oMbFQ2\\9PNeFP2X9SNhFm1V9TNkFl1S9VNmFj1P9XNQGh1m8ZNSGf1k8\\NUGd1h8^NXGc1f8_NZGa1c8aN^G_1`8cN`G]1^8eNbG[1Z8iNfGW1R8QOnGo0k7XOUHh0d7^O]Hb0[7FeH:T7MlH3l65TIKe6<[ID]6d0cI\\OV6k0jIUOo5R1QJnNg5Z1YJfN`5a1`J_NX5i1hJWNQ5o1PKQNh4W2XKiMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2ZKgMf4Y2YKhMg4W2ZKiMf4W2ZKiMf4W2ZKiMf4W2ZKUM[OfN[5U4ZKUM[OfN[5U4ZKUM\\OeNZ5V4ZKUM\\OeNZ5V4ZKUM]OdNY5W4ZKUM]OdNY5W4ZKUM]OdNY5W4ZKUM^OcNX5W4[KVM]OcNX5W4[KVM^ObNW5X4[KVM^ObNW5X4[KVM_OaNV5Y4[KVM_OaNV5Y4[KVM@`NU5Z4[KRL^O62^OU5Z4[KeKKc0E^OU5Z4YLXLcN]OT5[4ZLWLbN^OT5[4ZLWLcN[OU5]4YLXLbNZOV5^4YLWLbNXOY5_4ULYLbNVO_5]4oK\\LcNVOb5[4lK^Le4^3[KbLi4Z3XKeLk4X3UKhLo4T3QKkLT5Q3mJnLV5n2kJRMY5j2gJVM]5f2dJYM_5d2aJ[Md5a2\\J_Mh5]2YJbMj5[2VJeMh5]2ZJaMQ4T4c201O0O2O001O001O001O001O000O4M4L4L4L3M4L4L4L4L3M4K`NmLnHo2R7UMnHg2R7^MlH_2T7fMkHV2U7nMkHn1V7VNhHg1X7]NdHc1\\7bN^H_1b7eNYH\\1g7iNTHW1l7mNoGS1S8QOgGP1Y8TObGm0^8XO]Gh0c8P20000001O000O1000001O0000000000001N1000000000001O0000000O10001O0000000000001O0O10000000001O000000000O101O00kIPLo2P4nLVLo2j3nL]Ln2c3oL`LQ3`3lLdLS3\\3kLgLU3X3hLkLX3T3mJnKS1R1o3P3nJRLl0Q1V4m2nJVLe0Q1\\4i2oJZL>Q1b4e2PK^L7P1i4b2oJbLLV1U5W2oJgL_O[1b5n1oJnNQ5R1oJnNQ5R1oJnNQ5R1oJnNQ5Q1oJPOQ5P1oJPOR5o0nJQOR5o0nJQOR5o0nJQOR5o0nJQOR5o0mJROS5n0mJROS5n0mJROT5m0lJSOT5m0lJSOT5l0lJUOT5k0lJUOT5k0lJUOT5k0lJUOT5k0lJUOU5j0kJVOU5j0jJWOV5i0jJWOV5i0jJWOV5i0jJWOV5i0jJWOV5h0kJXOV5g0iJZOW5f0iJZOW5f0iJZOW5f0iJZOW5f0iJZOW5f0iJZOW5f0hJ[OX5e0hJ[OX5e0hJ[OX5e0hJ[OX5d0iJ\\OW5d0hJ]OW5d0iJ\\OW5d0iJ\\OW5d0iJ\\OW5d0iJ\\OW5d0iJ\\OW5d0hJ]OX5c0hJ]OW5d0iJ\\OW5d0iJ[OX5e0hJYOZ5f0gJWO\\5i0cJUO`5k0`JSOb5m0^JPOd5Q1\\JmNf5S1ZJjNi5V1VJiNl5W1TJfNo5Z1QJcNR6]1nIaNT6_1lI_NU6a1lI^NU6b1jI_N>fMj4k3hJ_N;iMm4h3hJ_N8lMP5e3hJ_N5PNR5a3iJ_N2SNT5_3jJ^N0UNV5]3iJ_NNWNY5Z3iJ_NKZN\\5W3iJ_NH]N_5T3iJ^NFaNa5P3jJ_NCcNc5n2jJ_OU5b0jJ_OV5a0jJ_OV5a0jJ_OV5a0jJ_OV5a0jJ_OV5a0iJ@W5`0iJ@V5a0jJ_OV5a0jJ_OV5`0kJ@U5`0kJ@U5`0jJAV5?jJAV5?jJAU5`0kJ@U5`0kJ@U5`0kJ@U5`0jJAV5?jJAV5?jJAV5>kJBT5?lJAU5>kJBU5>jJDU5<kJDV5;jJEV5;jJEW5:iJFW5:hJeNH_N`5l2hJXN8iNQ5n2gJkMi0TO`4Q3gJ^MY1^OQ4R3gJSMi1G`3V3gJlLS2KW3X3eJkLZ2JQ3[3eJhL`2Jk2^3eJeLg2Ie2a3dJbLn2J^2d3dJ_Lk8`3UG]Ln8c3QG[LR9e3;O0O2E:[Oe0[Of0ZOe0ZOWi`0GQU_Of201000O100O100O010O100O100O10O0100O10000O10O0100O100O100O010000000O10O1000000000000O1000000001O00001O001O00001O00001O001O00001N101O01O01O00001O0010O0001O000010O01O00001O0010O0001O001O01O01O001O000010O01O00001O0010O0001O001O01O01O001O002N100O1O1O1O1O1O1O2N001N2O1O1O1O1N101O1O1N2O1O1N1O2M3M3N2M3N2M3N1N3M3N2M3N2M3N1NTim0"}, "label": "polar bear"}]}
{"id": 92009, "image": "COCO_train2014_000000092009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bear legs\"."}], "task": "refering", "answer_template": "The \"brown bear legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 298, 314, 315, 316, 317, 318, 337, 338, 339, 340, 341, 342], "bbox": [0.67334375, 0.4837704918032787, 1.0, 0.9992505854800937], "iscrowd": 0, "area": 25253.821350000002, "rle": {"size": [427, 640], "counts": "ZSd53V=6J7I3N1N2N3M2O1N2N2N2O2M2N2kMmNPHU1m7ROnGo0P8WOjGk0S8[OhGg0V8_OdGd0X8AeG`0Y8DcG>[8EbG=[8HaG:]8J`G7_8K^G8a8J[G8d8KXG7h8JVG7j8KRG7n8KnF7Q9R201O0O10000O1000000O10XORGSMn8l2TGSMl8m2TGSMl8l2UGTMl8k2UGTMk8k2VGUMk8i2WGVMi8j2WGVMj8h2XGWMh8i2XGWMi8g2XGYMh8f2ZGYMg8f2YGZMg8e2[GZMf8e2ZG[Mf8d2[G\\Mf8b2\\G]Md8c2\\G\\MOCT8o2oG]MLEU8n2oG\\MKHV8k2PH]MIIW8i2RH]MELY8g2RH\\MENY8e2TH\\MA1[8c2gH^MY7a2fHaM[7]2eHdM[7\\2dHeM\\7Z2cHhM]7W2bHkM^7U2aHlM_7S2`HoM`7Q2^HQNb7n1]HUNb7k1\\HWNd7h1\\HYNd7g1ZH[Nf7d1YH^Ng7b1WH`Ni7a30000000000001O00000000000000000000000000000000000\\O[GbLd8^3^GaLb8]3`GcL`8[3bGeL]8Z3fGeLZ8Y3hGgL_6AkJh3hNfL]6CjJe3kNhL[6DiJb3nNjLY6EgJ`3SOjLV6HeJ]3VOkLU6IcJ[3[OkLR6KbJX3^OmLP6L`JV3CmLl5O_JS3GnLj51[JR3LmLi52XJQ32lLf54VJo26mLd55SJn2<lLa58PJk2a0mL_59mIk2e0lL^5:jIj2k0kL[5<hIi2n0kLY5>fIg2T1jLV5a0bIf2Y1iLU5b0`Ie2]1iLR5c0^Id2b1iLP5d0[Id2h1fLm4X4WKfKi4Y4ZKeKf4Y4^KeKa4[4X2N2O1N2O1O1O1O001N2O1O1O1O1O1O001O100O1O100O10O0100O100O100O10O0100O1000000000O100000000000000O10000000O10000i0VO10O10000O100000O010000O10000O10aI"}, "label": "brown bear legs"}]}
{"id": 92009, "image": "COCO_train2014_000000092009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bear\"."}], "task": "refering", "answer_template": "The \"a brown bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 298, 314, 315, 316, 317, 318, 337, 338, 339, 340, 341, 342], "bbox": [0.67334375, 0.4837704918032787, 1.0, 0.9992505854800937], "iscrowd": 0, "area": 25253.821350000002, "rle": {"size": [427, 640], "counts": "ZSd53V=6J7I3N1N2N3M2O1N2N2N2O2M2N2kMmNPHU1m7ROnGo0P8WOjGk0S8[OhGg0V8_OdGd0X8AeG`0Y8DcG>[8EbG=[8HaG:]8J`G7_8K^G8a8J[G8d8KXG7h8JVG7j8KRG7n8KnF7Q9R201O0O10000O1000000O10XORGSMn8l2TGSMl8m2TGSMl8l2UGTMl8k2UGTMk8k2VGUMk8i2WGVMi8j2WGVMj8h2XGWMh8i2XGWMi8g2XGYMh8f2ZGYMg8f2YGZMg8e2[GZMf8e2ZG[Mf8d2[G\\Mf8b2\\G]Md8c2\\G\\MOCT8o2oG]MLEU8n2oG\\MKHV8k2PH]MIIW8i2RH]MELY8g2RH\\MENY8e2TH\\MA1[8c2gH^MY7a2fHaM[7]2eHdM[7\\2dHeM\\7Z2cHhM]7W2bHkM^7U2aHlM_7S2`HoM`7Q2^HQNb7n1]HUNb7k1\\HWNd7h1\\HYNd7g1ZH[Nf7d1YH^Ng7b1WH`Ni7a30000000000001O00000000000000000000000000000000000\\O[GbLd8^3^GaLb8]3`GcL`8[3bGeL]8Z3fGeLZ8Y3hGgL_6AkJh3hNfL]6CjJe3kNhL[6DiJb3nNjLY6EgJ`3SOjLV6HeJ]3VOkLU6IcJ[3[OkLR6KbJX3^OmLP6L`JV3CmLl5O_JS3GnLj51[JR3LmLi52XJQ32lLf54VJo26mLd55SJn2<lLa58PJk2a0mL_59mIk2e0lL^5:jIj2k0kL[5<hIi2n0kLY5>fIg2T1jLV5a0bIf2Y1iLU5b0`Ie2]1iLR5c0^Id2b1iLP5d0[Id2h1fLm4X4WKfKi4Y4ZKeKf4Y4^KeKa4[4X2N2O1N2O1O1O1O001N2O1O1O1O1O1O001O100O1O100O10O0100O100O100O10O0100O1000000000O100000000000000O10000000O10000i0VO10O10000O100000O010000O10000O10aI"}, "label": "a brown bear"}]}
{"id": 247660, "image": "COCO_train2014_000000247660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white shorts with an event smock and dark jacket\"."}], "task": "refering", "answer_template": "The \"a man in white shorts with an event smock and dark jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 291, 292, 293, 314, 315, 316, 337, 338, 339, 360, 361, 362, 383, 384, 385], "bbox": [0.6011875, 0.30362976406533576, 0.7982812499999999, 0.8368965517241379], "iscrowd": 0, "area": 18634.65745, "rle": {"size": [551, 640], "counts": "l\\_63Sa02M4L3N3M2O2M3L3N3M2N100N2O1N1N3N2O0001O001O00M3h@oN[>Q1ZAD\\>h1L4K5L;E9F3O0O2N3M5L3K5L4K6K4K5XFZKf6l4SIWKP6JXHT5b1UKR6JZHR6c0WJP7I\\HU6`0SJR7J\\HW6>PJT72UHT6c0jIW76RHT6c0gIZ7:mGT6a0gI`7Z7ZHjHd7[7WHfHi7]7PHgHQ8W80000000O100000000O10000_MPHkKQ8S4UH\\J9Mb7f5ZHVJa0IV7P6_HPJb0Ko6U6WJhIi5X6ZIgIYOO^7[6kGiIn04R8`6jGbIU8a6gGbIW8b6cGaIEDX8n6oGoIR8BbGj55fJZ8[OfG[7\\8`HiG^7R9N2O0O1O1O00001O00001lN_F`Ja9Z5gFdJZ9V5oFfJP9X5XGcJh8Y5bG^Ja8_5e1TOkDbKY;S4VEeKn:W4S1YM[Ch0j<WOaC9e<G\\CSOJHU=T1RCmN0IT=Y1eC\\NQO6^=_1\\C^NW=c1dB^N_=d1ZB_Ng=c1TB^NCLP>i1WB[NJMP>j22N3M3L3N3L2N3M3M3M3M3M4L4G:J5J6[Of0DTUU2"}, "label": "a man in white shorts with an event smock and dark jacket"}]}
{"id": 247660, "image": "COCO_train2014_000000247660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a grey umbrella following a woman\"."}], "task": "refering", "answer_template": "The \"a man holding a grey umbrella following a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 291, 292, 293, 314, 315, 316, 337, 338, 339, 360, 361, 362, 383, 384, 385], "bbox": [0.6011875, 0.30362976406533576, 0.7982812499999999, 0.8368965517241379], "iscrowd": 0, "area": 18634.65745, "rle": {"size": [551, 640], "counts": "l\\_63Sa02M4L3N3M2O2M3L3N3M2N100N2O1N1N3N2O0001O001O00M3h@oN[>Q1ZAD\\>h1L4K5L;E9F3O0O2N3M5L3K5L4K6K4K5XFZKf6l4SIWKP6JXHT5b1UKR6JZHR6c0WJP7I\\HU6`0SJR7J\\HW6>PJT72UHT6c0jIW76RHT6c0gIZ7:mGT6a0gI`7Z7ZHjHd7[7WHfHi7]7PHgHQ8W80000000O100000000O10000_MPHkKQ8S4UH\\J9Mb7f5ZHVJa0IV7P6_HPJb0Ko6U6WJhIi5X6ZIgIYOO^7[6kGiIn04R8`6jGbIU8a6gGbIW8b6cGaIEDX8n6oGoIR8BbGj55fJZ8[OfG[7\\8`HiG^7R9N2O0O1O1O00001O00001lN_F`Ja9Z5gFdJZ9V5oFfJP9X5XGcJh8Y5bG^Ja8_5e1TOkDbKY;S4VEeKn:W4S1YM[Ch0j<WOaC9e<G\\CSOJHU=T1RCmN0IT=Y1eC\\NQO6^=_1\\C^NW=c1dB^N_=d1ZB_Ng=c1TB^NCLP>i1WB[NJMP>j22N3M3L3N3L2N3M3M3M3M3M4L4G:J5J6[Of0DTUU2"}, "label": "a man holding a grey umbrella following a woman"}]}
{"id": 247660, "image": "COCO_train2014_000000247660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a golfer with pink pants walking\"."}], "task": "refering", "answer_template": "The \"a golfer with pink pants walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 172, 173, 174, 195, 196, 197, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 309, 310, 311, 312, 333, 334, 335, 356, 357, 358, 379, 380, 381, 402, 403, 404, 425, 426, 427], "bbox": [0.45171875000000006, 0.3446823956442831, 0.6373125, 0.9239019963702358], "iscrowd": 0, "area": 22849.7776, "rle": {"size": [551, 640], "counts": "bjk45e`0b0_Oa0B>E;E7I8H7I8H7K5N3M3VFaLg5c3UJ^Lj5f3RJ\\Ll5h3QJYLm5k3oIWLo5m3mIULQ6n3lITLR6P4ZG[LV1F^7c5YH_Je7j5RHXJl7Q6kGPJT8Y6cGiI[8c7N1O1O2N1O1O1O1O1O1O2N1O100VO_GmHb8Q7bGlH_8R7eGkH\\8S7hGjHY8T7kGjHV8S7nGjHS8T7QHiHP8W7RHfHo7Z7k03K6K5J6J6K5jMV2G9G8H9O16J6J6J6J6J6TGPJT6W6UIcJb6c5fHoJZ7_7000O2O2N1O1O1O2N1O1O2Nh0XO1O1O1N2O1G9F:E;F:F:E;F:F:F:G9G9G9H8G:F9G:F9C>C=L3L5L3L6K4K5L4K5L5K4K5L4K6K4K5L4K5Lial3"}, "label": "a golfer with pink pants walking"}]}
{"id": 247660, "image": "COCO_train2014_000000247660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a golf ball and golf club\"."}], "task": "refering", "answer_template": "The \"a woman holding a golf ball and golf club\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 172, 173, 174, 195, 196, 197, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 309, 310, 311, 312, 333, 334, 335, 356, 357, 358, 379, 380, 381, 402, 403, 404, 425, 426, 427], "bbox": [0.45171875000000006, 0.3446823956442831, 0.6373125, 0.9239019963702358], "iscrowd": 0, "area": 22849.7776, "rle": {"size": [551, 640], "counts": "bjk45e`0b0_Oa0B>E;E7I8H7I8H7K5N3M3VFaLg5c3UJ^Lj5f3RJ\\Ll5h3QJYLm5k3oIWLo5m3mIULQ6n3lITLR6P4ZG[LV1F^7c5YH_Je7j5RHXJl7Q6kGPJT8Y6cGiI[8c7N1O1O2N1O1O1O1O1O1O2N1O100VO_GmHb8Q7bGlH_8R7eGkH\\8S7hGjHY8T7kGjHV8S7nGjHS8T7QHiHP8W7RHfHo7Z7k03K6K5J6J6K5jMV2G9G8H9O16J6J6J6J6J6TGPJT6W6UIcJb6c5fHoJZ7_7000O2O2N1O1O1O2N1O1O2Nh0XO1O1O1N2O1G9F:E;F:F:E;F:F:F:G9G9G9H8G:F9G:F9C>C=L3L5L3L6K4K5L4K5L5K4K5L4K6K4K5L4K5Lial3"}, "label": "a woman holding a golf ball and golf club"}]}
{"id": 67438, "image": "COCO_train2014_000000067438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is not licking a girl ' s hand , but looking at the girl\"."}], "task": "refering", "answer_template": "The \"a giraffe that is not licking a girl ' s hand , but looking at the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 292, 293, 294, 295, 316, 317, 318, 339, 340, 341, 342, 363, 364, 365], "bbox": [0.570171875, 0.5680184331797236, 0.91228125, 0.9895391705069125], "iscrowd": 0, "area": 13296.840499999998, "rle": {"size": [434, 640], "counts": "doj41`=002M2O2N2N2M3N2N2N2M20O0100O1O10O0100O100O1O010O100O1O010O100O101N1O100O2O0O1O101N100O100O2N100O101N1O1O101N1O1O1O2N100O1O2N1O1O010O1O100000O10000000001N100000000O1000001O000O10000002N101N1O2N2N1O2N1O2N2O1N4L4L4L4L4L3M4M3N2N2M3N2N2M2O2N2M3N2M3NN2O0O2N2O0O2N2O0O2N101N2N101N2N101N2N1O2O0O2N2O0O2N2O0O2N2O0O27H9HO1N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2M3N1O2N2N1Ohgg0"}, "label": "a giraffe that is not licking a girl ' s hand , but looking at the girl"}]}
{"id": 67438, "image": "COCO_train2014_000000067438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe looking over a fence next to an adult giraffe\"."}], "task": "refering", "answer_template": "The \"a baby giraffe looking over a fence next to an adult giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 292, 293, 294, 295, 316, 317, 318, 339, 340, 341, 342, 363, 364, 365], "bbox": [0.570171875, 0.5680184331797236, 0.91228125, 0.9895391705069125], "iscrowd": 0, "area": 13296.840499999998, "rle": {"size": [434, 640], "counts": "doj41`=002M2O2N2N2M3N2N2N2M20O0100O1O10O0100O100O1O010O100O1O010O100O101N1O100O2O0O1O101N100O100O2N100O101N1O1O101N1O1O1O2N100O1O2N1O1O010O1O100000O10000000001N100000000O1000001O000O10000002N101N1O2N2N1O2N1O2N2O1N4L4L4L4L4L3M4M3N2N2M3N2N2M2O2N2M3N2M3NN2O0O2N2O0O2N2O0O2N101N2N101N2N101N2N1O2O0O2N2O0O2N2O0O2N2O0O27H9HO1N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2M3N1O2N2N1Ohgg0"}, "label": "a baby giraffe looking over a fence next to an adult giraffe"}]}
{"id": 67438, "image": "COCO_train2014_000000067438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a blonde girls head that is being held by a woman\"."}], "task": "refering", "answer_template": "The \"the back of a blonde girls head that is being held by a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 301, 302, 322, 323, 324, 345, 346], "bbox": [0.0, 0.3256451612903226, 0.22134375, 0.9787096774193548], "iscrowd": 0, "area": 24775.745249999993, "rle": {"size": [434, 640], "counts": "P:`0R<P1POQ1XOh0N1XHkLQ5X3`JVM`5j2`JWM_5j2_JWMa5i2_JXM`5i2^JXMg4CnIT3[1YM_4ORJi2]1YM]46SJa2_1YM[4=RJZ2b1ZMY4b0SJT2c1ZMW4i0RJm1f1[MU4\\4jKdKS4_4lKbKQ4b4nK^KP4d4oK]Ko3d4RL\\Kl3e4TL\\Kk3e4UL[Ki3g4VLZKh3h4XLXKf3j4YLWKf3j4ZLVKd3l4\\LTKc3l4]LUKb3l4^LTKa3m4^LTKb3l4^LTKa3m4^LTKb3l4^LTKa3m4^LTKb3l4^LTKb3l4]LUKc3j4^LVKa3k4^LVKb3j4^LVKa3k4^LVKb3j4^LVKa3k4^LVKb3j4^LVKa3j4_LWK`3j4`LVK`3j4_LWK`3i4aLWK^3j4aLWK_3i4aLWK^3j4aLWK^3i4cLWK\\3j4cLWK]3i4cLSKa3m4^LPKf3P5ZLmJi3S5VLjJn3V5RL`JX4a5fK^J\\4b5dK\\J^4d5aK[Ja4f5]KYJd4h5[KWJg4j5VKVJm4i5RKVJP5j5oJUJS5l5kJSJX5l5gJSJ[5m5dJRJ^5o5`JPJb5P6\\JPJg5P6WJPJj5R6SJlIP6U6nIiIV6X6hIeI[6\\6=2N2O0O0010O00100O10O01O1O1O1O1O1O2M3N2N2N2N1O2N2N2M3N6J3M4L4K5L3M4L4K5L3M4K5L2N2N2M3N2M4L4L4L3L5L4L4L3L4M3M3L5L3M3M3L5L3M3M3M3Nb0]O8Hdhb6"}, "label": "the back of a blonde girls head that is being held by a woman"}]}
{"id": 67438, "image": "COCO_train2014_000000067438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl with long hair\"."}], "task": "refering", "answer_template": "The \"little girl with long hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 301, 302, 322, 323, 324, 345, 346], "bbox": [0.0, 0.3256451612903226, 0.22134375, 0.9787096774193548], "iscrowd": 0, "area": 24775.745249999993, "rle": {"size": [434, 640], "counts": "P:`0R<P1POQ1XOh0N1XHkLQ5X3`JVM`5j2`JWM_5j2_JWMa5i2_JXM`5i2^JXMg4CnIT3[1YM_4ORJi2]1YM]46SJa2_1YM[4=RJZ2b1ZMY4b0SJT2c1ZMW4i0RJm1f1[MU4\\4jKdKS4_4lKbKQ4b4nK^KP4d4oK]Ko3d4RL\\Kl3e4TL\\Kk3e4UL[Ki3g4VLZKh3h4XLXKf3j4YLWKf3j4ZLVKd3l4\\LTKc3l4]LUKb3l4^LTKa3m4^LTKb3l4^LTKa3m4^LTKb3l4^LTKa3m4^LTKb3l4^LTKb3l4]LUKc3j4^LVKa3k4^LVKb3j4^LVKa3k4^LVKb3j4^LVKa3k4^LVKb3j4^LVKa3j4_LWK`3j4`LVK`3j4_LWK`3i4aLWK^3j4aLWK_3i4aLWK^3j4aLWK^3i4cLWK\\3j4cLWK]3i4cLSKa3m4^LPKf3P5ZLmJi3S5VLjJn3V5RL`JX4a5fK^J\\4b5dK\\J^4d5aK[Ja4f5]KYJd4h5[KWJg4j5VKVJm4i5RKVJP5j5oJUJS5l5kJSJX5l5gJSJ[5m5dJRJ^5o5`JPJb5P6\\JPJg5P6WJPJj5R6SJlIP6U6nIiIV6X6hIeI[6\\6=2N2O0O0010O00100O10O01O1O1O1O1O1O2M3N2N2N2N1O2N2N2M3N6J3M4L4K5L3M4L4K5L3M4K5L2N2N2M3N2M4L4L4L3L5L4L4L3L4M3M3L5L3M3M3L5L3M3M3M3Nb0]O8Hdhb6"}, "label": "little girl with long hair"}]}
{"id": 10094, "image": "COCO_train2014_000000010094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blurry person in the background in red and grey\"."}], "task": "refering", "answer_template": "The \"the blurry person in the background in red and grey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 168, 169, 170, 185, 186, 187, 188, 203, 204, 205, 206, 221, 222, 223, 224, 239, 240, 241, 242, 257, 258, 259, 275, 276, 277], "bbox": [0.28548, 0.47414, 0.49398000000000003, 0.90398], "iscrowd": 0, "area": 14060.934849999998, "rle": {"size": [500, 500], "counts": "UTV22_?4M4L3L4M3hCAW9c0dFBW9b0QEYOc0:X:`0QE]O`08[:>PEC<5_:<QEF93b:;PEI71d::QEM3Nh:8PE21Jl:Y3000000O10001O0O1000000O101O000O1000000O2O00000O102N1O2M3N1O2M2M4L4L3M4L3L101O1O1O001O1N2O1O001002N2M3L4M2M4M3L4M3M3L3N3L4M3L4M3M2M4M3L4N2N2fLcD[2];cMfD[2\\;cMfD\\2[;bMgD\\2[;aMhD^2Y;`MiD^2Y;_MjD`2V;_MlD_2V;_MmDP1]O1g;lNoDo0A0b;oNoDn0EO];POQEm0INX;SOQEk0NNQ;VOSEh03Mn;0XDLi;0^DKd;1cDK_;0hDKZ;1W2McWk3"}, "label": "the blurry person in the background in red and grey"}]}
{"id": 10094, "image": "COCO_train2014_000000010094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry man with a white t - shirt and blue jeans on wearing a red baseball cap\"."}], "task": "refering", "answer_template": "The \"a blurry man with a white t - shirt and blue jeans on wearing a red baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 165, 166, 167, 182, 183, 184, 200, 201, 202, 203, 218, 219, 220, 221, 236, 237, 238, 239, 254, 255, 256, 257, 272, 273, 274, 275, 290, 291, 292], "bbox": [0.10809999999999999, 0.45045999999999997, 0.30856, 0.94144], "iscrowd": 0, "area": 17834.80885, "rle": {"size": [500, 500], "counts": "fej0d0l>P1QO8G:F9H8J6J6oCfM^:`2SElMi:b3M4M3L3N3L2N2O1N2O1N2O1N2O0O2O1N2O1N1O10000O10000O1mN^JTHb5g7jJRHV5h7PKXHP5c7UK\\Hl4c7VK\\Hj4d7VK\\Hj4c7WK[Hk4d7WKZHj4e7WKYHk4g7UKXHl4g7UKWHm4h7UKSHo4m7\\10000000000000O1M2N3M3J6J6I7M3O1O1O1O1O1O010000001O001N101O00001N2_KgFe2Z9YMjFd2W9[MlFb2U9\\MnFb2R9\\MRGb2n8]MTGa2m8^MVG`2j8^MYGa2g8^M[Ga2e8^MaG]2_8^MnGZ2S8QMjHe2X7XMnHd2R7ZMTIb2m6\\M\\IHhMj1m8\\NWJb1j5]NXJb1i5[NZJc1h5[NZJP1Z6oNokX5"}, "label": "a blurry man with a white t - shirt and blue jeans on wearing a red baseball cap"}]}
{"id": 10094, "image": "COCO_train2014_000000010094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a member of the audience standing up wearing a light tan shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"a member of the audience standing up wearing a light tan shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 165, 166, 167, 182, 183, 184, 200, 201, 202, 203, 218, 219, 220, 221, 236, 237, 238, 239, 254, 255, 256, 257, 272, 273, 274, 275, 290, 291, 292], "bbox": [0.10809999999999999, 0.45045999999999997, 0.30856, 0.94144], "iscrowd": 0, "area": 17834.80885, "rle": {"size": [500, 500], "counts": "fej0d0l>P1QO8G:F9H8J6J6oCfM^:`2SElMi:b3M4M3L3N3L2N2O1N2O1N2O1N2O0O2O1N2O1N1O10000O10000O1mN^JTHb5g7jJRHV5h7PKXHP5c7UK\\Hl4c7VK\\Hj4d7VK\\Hj4c7WK[Hk4d7WKZHj4e7WKYHk4g7UKXHl4g7UKWHm4h7UKSHo4m7\\10000000000000O1M2N3M3J6J6I7M3O1O1O1O1O1O010000001O001N101O00001N2_KgFe2Z9YMjFd2W9[MlFb2U9\\MnFb2R9\\MRGb2n8]MTGa2m8^MVG`2j8^MYGa2g8^M[Ga2e8^MaG]2_8^MnGZ2S8QMjHe2X7XMnHd2R7ZMTIb2m6\\M\\IHhMj1m8\\NWJb1j5]NXJb1i5[NZJc1h5[NZJP1Z6oNokX5"}, "label": "a member of the audience standing up wearing a light tan shirt and blue jeans"}]}
{"id": 477040, "image": "COCO_train2014_000000477040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the right\"."}], "task": "refering", "answer_template": "The \"the chair on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 300, 301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356], "bbox": [0.5056147540983607, 0.7426875, 1.0, 0.9001250000000001], "iscrowd": 0, "area": 17663.800700000007, "rle": {"size": [640, 488], "counts": "okj45jc09H7H7J2M100O100O100O100O1O10O0100O4M5J6K0O010O100O1O101N100O100O100O101N100O100O10000O2O000O10000O10001N10000O1000001N10000O10000O101O0O10000O1000001O000O10000000001O00000000000O101O000000000000001O00000O1000001O000000000000001N100001O00000001O000000000000001O0000000001O0001O000000000000001O000000000000010O00000001O0001O01O0000001O0001O01O0000001O0001O01O0000001O0001O01O0000001O0001O01O0000001O0001O0001O00000010O000001O00001O01O01O001O001O00010O001O0H9^Oa0A?10O001O001O001O2O1C`@"}, "label": "the chair on the right"}]}
{"id": 477040, "image": "COCO_train2014_000000477040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair on the right\"."}], "task": "refering", "answer_template": "The \"chair on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 300, 301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356], "bbox": [0.5056147540983607, 0.7426875, 1.0, 0.9001250000000001], "iscrowd": 0, "area": 17663.800700000007, "rle": {"size": [640, 488], "counts": "okj45jc09H7H7J2M100O100O100O100O1O10O0100O4M5J6K0O010O100O1O101N100O100O100O101N100O100O10000O2O000O10000O10001N10000O1000001N10000O10000O101O0O10000O1000001O000O10000000001O00000000000O101O000000000000001O00000O1000001O000000000000001N100001O00000001O000000000000001O0000000001O0001O000000000000001O000000000000010O00000001O0001O01O0000001O0001O01O0000001O0001O01O0000001O0001O01O0000001O0001O01O0000001O0001O0001O00000010O000001O00001O01O01O001O001O00010O001O0H9^Oa0A?10O001O001O001O2O1C`@"}, "label": "chair on the right"}]}
{"id": 477040, "image": "COCO_train2014_000000477040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown chair to the left of another chair\"."}], "task": "refering", "answer_template": "The \"brown chair to the left of another chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330], "bbox": [0.0, 0.7049375, 0.4509016393442623, 0.8397656250000001], "iscrowd": 0, "area": 14263.690500000008, "rle": {"size": [640, 488], "counts": "n>T1Qc0L4KA?O1000000O10001O0O1000000O1000000O2O0000000O1000000O2O00000O1000000O101O000O1000000O10001O0O100000000O1000001N1000000O1000000O2O00000O1000000O101O00000000000000001O00000O1000000000001O000000000000000O2O00000000000000001O000000000O1000001O0000000000000000001O0O100000000001O01O0001O0000001O0001O01O00001O00000010O000001O0000001O0001O01O00001O00000010O000001O0000001O01O0001O00001O0000010O0000001O0000001O01O0001O00001O000J6G:F9G9G9G:F`QW5"}, "label": "brown chair to the left of another chair"}]}
{"id": 477040, "image": "COCO_train2014_000000477040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leftmost chair\"."}], "task": "refering", "answer_template": "The \"the leftmost chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330], "bbox": [0.0, 0.7049375, 0.4509016393442623, 0.8397656250000001], "iscrowd": 0, "area": 14263.690500000008, "rle": {"size": [640, 488], "counts": "n>T1Qc0L4KA?O1000000O10001O0O1000000O1000000O2O0000000O1000000O2O00000O1000000O101O000O1000000O10001O0O100000000O1000001N1000000O1000000O2O00000O1000000O101O00000000000000001O00000O1000000000001O000000000000000O2O00000000000000001O000000000O1000001O0000000000000000001O0O100000000001O01O0001O0000001O0001O01O00001O00000010O000001O0000001O0001O01O00001O00000010O000001O0000001O01O0001O00001O0000010O0000001O0000001O01O0001O00001O000J6G:F9G9G9G:F`QW5"}, "label": "the leftmost chair"}]}
{"id": 313206, "image": "COCO_train2014_000000313206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right with no stripes on its legs\"."}], "task": "refering", "answer_template": "The \"the zebra on the right with no stripes on its legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 340, 341, 342, 360, 361, 362, 363, 364, 365], "bbox": [0.6226406250000001, 0.12949561403508772, 0.9644375000000001, 1.0], "iscrowd": 0, "area": 58504.638599999984, "rle": {"size": [456, 640], "counts": "aZa51U>8H8I7H8I7H8K5K6J5K5K4L5nETN]7P2hGhNW8]1mF\\OQ9P3N2N1O2N2N1O2N2N1O2N2N3M2M4J6I7J5K6J6J6J6J=C=C=C=C=B?Ba0_Oc0]Ob0^Ob0^O4L2N2N2N2N2001O1O1O1O1O1O1O1O1O1O1O1O1O1O1E;VOk0I6fL[LjKk3U4RMiJU3U5hMiI^2V6l2N2O1N1O1O1O1O100OO100O101N100O100O2O0O1O100O2O0O100O1`JUIc3k6kK\\J7hNg1n6`M]L_2h7M2N3N1N3N0O100O100O100O1O010O100O100O100O100O10O1O1O2M2N2O1N2POoMhER2h9eNoE]1n9`1L4M3L4L4M3L4M3L4M3L_M[GGb8nNWGK?V1V8oNkGWO7j1k7mN^IR1^6mNhIR1T6mNQJS1o5hNWJW1c6kMaIU2Y900000000000O10000000001kMoMQHR2f7[NUHe1c7hNXHY1a7ROXGkN?T2T8LhG5S83iGMR8;kGFP8b0lG_Oo7h0nGXOn7P1mGROm7h3K5K4L5M3M3M2N3M3M3M2N3M3M2N3M3M3M2N3L4M4L4L3M4L4L4K5L4L4L4L4L4L4L3L5L4L4H8B>C=C=C=CjS:"}, "label": "the zebra on the right with no stripes on its legs"}]}
{"id": 313206, "image": "COCO_train2014_000000313206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smaller zebra with little striping on its legs standing next to a larger zebra\"."}], "task": "refering", "answer_template": "The \"a smaller zebra with little striping on its legs standing next to a larger zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 340, 341, 342, 360, 361, 362, 363, 364, 365], "bbox": [0.6226406250000001, 0.12949561403508772, 0.9644375000000001, 1.0], "iscrowd": 0, "area": 58504.638599999984, "rle": {"size": [456, 640], "counts": "aZa51U>8H8I7H8I7H8K5K6J5K5K4L5nETN]7P2hGhNW8]1mF\\OQ9P3N2N1O2N2N1O2N2N1O2N2N3M2M4J6I7J5K6J6J6J6J=C=C=C=C=B?Ba0_Oc0]Ob0^Ob0^O4L2N2N2N2N2001O1O1O1O1O1O1O1O1O1O1O1O1O1O1E;VOk0I6fL[LjKk3U4RMiJU3U5hMiI^2V6l2N2O1N1O1O1O1O100OO100O101N100O100O2O0O1O100O2O0O100O1`JUIc3k6kK\\J7hNg1n6`M]L_2h7M2N3N1N3N0O100O100O100O1O010O100O100O100O100O10O1O1O2M2N2O1N2POoMhER2h9eNoE]1n9`1L4M3L4L4M3L4M3L4M3L_M[GGb8nNWGK?V1V8oNkGWO7j1k7mN^IR1^6mNhIR1T6mNQJS1o5hNWJW1c6kMaIU2Y900000000000O10000000001kMoMQHR2f7[NUHe1c7hNXHY1a7ROXGkN?T2T8LhG5S83iGMR8;kGFP8b0lG_Oo7h0nGXOn7P1mGROm7h3K5K4L5M3M3M2N3M3M3M2N3M3M2N3M3M3M2N3L4M4L4L3M4L4L4K5L4L4L4L4L4L4L3L5L4L4H8B>C=C=C=CjS:"}, "label": "a smaller zebra with little striping on its legs standing next to a larger zebra"}]}
{"id": 313206, "image": "COCO_train2014_000000313206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the back that is cut off\"."}], "task": "refering", "answer_template": "The \"the zebra in the back that is cut off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 29, 30, 31, 32, 33, 34, 35, 52, 53, 54, 55, 56, 57, 76, 77], "bbox": [0.280203125, 0.0018640350877192982, 0.56840625, 0.24232456140350878], "iscrowd": 0, "area": 11625.566099999996, "rle": {"size": [456, 640], "counts": "mV`25W<LaE8^:0WE4h:4mD0R;8dDK[;>YDFf;b0oCBP<\\1O1O1O1O1O1O1O1O1O1O001O:F=C4L00O100O10000O100O100O100O100O10000O100O100]NmD0T;0REJn:6YEBh:?]E[Oc:e0dETO\\:l0jEmNW:S1PFfNP:Z1VF_Nk9b1]100000000001O000000001O000000001O0000001O3M3M2N3M3M1O00O100O100O100O100O100O100O1O100O100O100O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000M3M3L4M3M3M4L3M3M3M3M3L4M3M3M5K4Lnlj3"}, "label": "the zebra in the back that is cut off"}]}
{"id": 313206, "image": "COCO_train2014_000000313206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in the background behind two other zebras , looking away from them\"."}], "task": "refering", "answer_template": "The \"a zebra in the background behind two other zebras , looking away from them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 29, 30, 31, 32, 33, 34, 35, 52, 53, 54, 55, 56, 57, 76, 77], "bbox": [0.280203125, 0.0018640350877192982, 0.56840625, 0.24232456140350878], "iscrowd": 0, "area": 11625.566099999996, "rle": {"size": [456, 640], "counts": "mV`25W<LaE8^:0WE4h:4mD0R;8dDK[;>YDFf;b0oCBP<\\1O1O1O1O1O1O1O1O1O1O001O:F=C4L00O100O10000O100O100O100O100O10000O100O100]NmD0T;0REJn:6YEBh:?]E[Oc:e0dETO\\:l0jEmNW:S1PFfNP:Z1VF_Nk9b1]100000000001O000000001O000000001O0000001O3M3M2N3M3M1O00O100O100O100O100O100O100O1O100O100O100O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000M3M3L4M3M3M4L3M3M3M3M3L4M3M3M5K4Lnlj3"}, "label": "a zebra in the background behind two other zebras , looking away from them"}]}
{"id": 313206, "image": "COCO_train2014_000000313206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra to the left\"."}], "task": "refering", "answer_template": "The \"zebra to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 24, 25, 26, 27, 28, 34, 48, 49, 50, 51, 52, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 264, 265, 266, 281, 282, 283, 284, 287, 288, 289, 304, 305, 306, 307, 310, 311, 327, 328, 329, 330, 333, 334, 350, 351, 352, 353, 356, 357], "bbox": [0.070453125, 0.006732456140350877, 0.64365625, 0.9887719298245614], "iscrowd": 0, "area": 92961.79745000003, "rle": {"size": [456, 640], "counts": "nQd01W>2N3M2O2M2N2UGDf3=ULLg35SL2j3ORL6l3LnK9Q4HkK<U4EeK`0Z4B`Kd0^4]O^Kg0a4[OYKj0g4VOUKn0j4SOQKS1m4mNoJX1P5hNZIDEh1Q7cNXI4VO^1a7_NVIc0hNR1Q8[NTIY3k6gLSI\\3m6dLPI_3o6aLnHc3R7\\LlHg3T7YLiHj3W7ULfHn3Z7RLdHP4]7PL`HS4`7lK]HW4d7hKZHZ4g7fKWH\\4h7dKVH^4k7aKSHa4n7_KPHc4P8\\KnGf4S8YKkGi4V8e0N1N2O1N101N1O100O00L5K410O01O010O2O0O2N1100O010O010O1O010O10]HhJU6X5fISKU6l4gI[KW6e4eIaKX6_4fIfKX6X6N1N4M3M3M3M001O1O0O2O1O1O1O2N2N2N1N3N2M2O2M3N1N3N1N2O2M2O1N33NK5L4L4L4L4L4K:G;E;E:Fa0^Oc0^O:F000008Hf0ZO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1E:F;XNh1[Od0hNY1K4M4O001O1O010O5K7I5M1O2O0OO2M2aJ^IT3e6eLfIS3]6hLmIQ3V6hLSJS3Y6^LmI^3c8L4L3L5L3L5L4L3L4M000O10000O10000O10000O10000O10QOo0[Od0]Od0G8I7I7I7I7J5J7I7I7I7M22O1O1O1N2O100O100O1O100O1O100O10O01O100O100O1O100O1O100O100O1O100O100O10000000O010000000000O10000000001N3N2N2N3M2N2N1N3oMSHcMo7k1aHUN`7h1bHXN`7c1cH]N_7_1cHaN_7[1cHdN`7W1cHiN_7P1fHPO]7h0hHXOZ7?mHAU76RIJP7MWI2m6YOeIg0_6nNhIR1]6cNiI]1\\6XNkIg1[900O100001H7L5M2N2N3L3N3M2RNaMQHa2k7hMmG\\2m7oMkGS2T8SNeGP2b8hNjFl0^9PNbGl1`:N2M3N3L3M01O001O001O00001O001O001gMRNVHo1e7]NcGCQOQ2W9gN]GBVOg1Y9SOUGf1f8fNnF_1n8P2K5K4L5K5L4K5J5K6J6J6J5K6J6J6J6J5L5L4L4L4L3M4L4L4M3L3M4L4L4L3M4L4L4L4L3N3I:F:F9G:F:F:F:F9G:F:F:F:Fe\\U3"}, "label": "zebra to the left"}]}
{"id": 313206, "image": "COCO_train2014_000000313206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller of a pair of zebras\"."}], "task": "refering", "answer_template": "The \"the taller of a pair of zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 24, 25, 26, 27, 28, 34, 48, 49, 50, 51, 52, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 264, 265, 266, 281, 282, 283, 284, 287, 288, 289, 304, 305, 306, 307, 310, 311, 327, 328, 329, 330, 333, 334, 350, 351, 352, 353, 356, 357], "bbox": [0.070453125, 0.006732456140350877, 0.64365625, 0.9887719298245614], "iscrowd": 0, "area": 92961.79745000003, "rle": {"size": [456, 640], "counts": "nQd01W>2N3M2O2M2N2UGDf3=ULLg35SL2j3ORL6l3LnK9Q4HkK<U4EeK`0Z4B`Kd0^4]O^Kg0a4[OYKj0g4VOUKn0j4SOQKS1m4mNoJX1P5hNZIDEh1Q7cNXI4VO^1a7_NVIc0hNR1Q8[NTIY3k6gLSI\\3m6dLPI_3o6aLnHc3R7\\LlHg3T7YLiHj3W7ULfHn3Z7RLdHP4]7PL`HS4`7lK]HW4d7hKZHZ4g7fKWH\\4h7dKVH^4k7aKSHa4n7_KPHc4P8\\KnGf4S8YKkGi4V8e0N1N2O1N101N1O100O00L5K410O01O010O2O0O2N1100O010O010O1O010O10]HhJU6X5fISKU6l4gI[KW6e4eIaKX6_4fIfKX6X6N1N4M3M3M3M001O1O0O2O1O1O1O2N2N2N1N3N2M2O2M3N1N3N1N2O2M2O1N33NK5L4L4L4L4L4K:G;E;E:Fa0^Oc0^O:F000008Hf0ZO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1E:F;XNh1[Od0hNY1K4M4O001O1O010O5K7I5M1O2O0OO2M2aJ^IT3e6eLfIS3]6hLmIQ3V6hLSJS3Y6^LmI^3c8L4L3L5L3L5L4L3L4M000O10000O10000O10000O10000O10QOo0[Od0]Od0G8I7I7I7I7J5J7I7I7I7M22O1O1O1N2O100O100O1O100O1O100O10O01O100O100O1O100O1O100O100O1O100O100O10000000O010000000000O10000000001N3N2N2N3M2N2N1N3oMSHcMo7k1aHUN`7h1bHXN`7c1cH]N_7_1cHaN_7[1cHdN`7W1cHiN_7P1fHPO]7h0hHXOZ7?mHAU76RIJP7MWI2m6YOeIg0_6nNhIR1]6cNiI]1\\6XNkIg1[900O100001H7L5M2N2N3L3N3M2RNaMQHa2k7hMmG\\2m7oMkGS2T8SNeGP2b8hNjFl0^9PNbGl1`:N2M3N3L3M01O001O001O00001O001O001gMRNVHo1e7]NcGCQOQ2W9gN]GBVOg1Y9SOUGf1f8fNnF_1n8P2K5K4L5K5L4K5J5K6J6J6J5K6J6J6J6J5L5L4L4L4L3M4L4L4M3L3M4L4L4L3M4L4L4L4L3N3I:F:F9G:F:F:F:F9G:F:F:F:Fe\\U3"}, "label": "the taller of a pair of zebras"}]}
{"id": 223095, "image": "COCO_train2014_000000223095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle cruise ship in the back\"."}], "task": "refering", "answer_template": "The \"the middle cruise ship in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 147], "bbox": [0.322, 0.23809411764705882, 0.6185937499999999, 0.41731764705882357], "iscrowd": 0, "area": 5553.3508, "rle": {"size": [425, 640], "counts": "Rde21X=1O2N1eC0[;2`D3^;N^D6a;K[D8e;JWD8i;ISD:m;a0001O000000YOlNYET1g:oNWEP1i:ROUEn0k:TOSEl0m:VOQEj0P;^OfDc0Z;f000001O1O2N2N2N2N1O2N2N2N2N1O2N2N0000O10000000001O00000000000GkN^DU1b;lN]DT1c;lN]DU1b;mN\\DS1d;9O2O001O0M3N3O0007CZDiNe;X1\\DhNc;X1_DfNa;Z1`DeN`;[17000000000O100000O1O1O1O100O00FiNdDV1];jNcDU1^;kN`DV1W;jNiD1LY1[;hNgD]1Z;dNeDT1NlN^;2bDR11jN];_1eD_N\\;a1600O100O100O100O10000O101N3N3L3N2M3N0O1O101N1O100O1O2O0O1O2N2O2M3M2O2M2O01O0000000MKRC5l<5O2O00000OHXC1j<LXC3o<001O00000000001O01O0000000000000000O1M3L40000;E0000L4O100004LM3O12N2N2NN2M32N2N2N1OkVW3"}, "label": "the middle cruise ship in the back"}]}
{"id": 223095, "image": "COCO_train2014_000000223095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest of several cruise - liners on the water\"."}], "task": "refering", "answer_template": "The \"the tallest of several cruise - liners on the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 147], "bbox": [0.322, 0.23809411764705882, 0.6185937499999999, 0.41731764705882357], "iscrowd": 0, "area": 5553.3508, "rle": {"size": [425, 640], "counts": "Rde21X=1O2N1eC0[;2`D3^;N^D6a;K[D8e;JWD8i;ISD:m;a0001O000000YOlNYET1g:oNWEP1i:ROUEn0k:TOSEl0m:VOQEj0P;^OfDc0Z;f000001O1O2N2N2N2N1O2N2N2N2N1O2N2N0000O10000000001O00000000000GkN^DU1b;lN]DT1c;lN]DU1b;mN\\DS1d;9O2O001O0M3N3O0007CZDiNe;X1\\DhNc;X1_DfNa;Z1`DeN`;[17000000000O100000O1O1O1O100O00FiNdDV1];jNcDU1^;kN`DV1W;jNiD1LY1[;hNgD]1Z;dNeDT1NlN^;2bDR11jN];_1eD_N\\;a1600O100O100O100O10000O101N3N3L3N2M3N0O1O101N1O100O1O2O0O1O2N2O2M3M2O2M2O01O0000000MKRC5l<5O2O00000OHXC1j<LXC3o<001O00000000001O01O0000000000000000O1M3L40000;E0000L4O100004LM3O12N2N2NN2M32N2N2N1OkVW3"}, "label": "the tallest of several cruise - liners on the water"}]}
{"id": 223095, "image": "COCO_train2014_000000223095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cruise ship with tourists in the foreground and two cruise ships in the background\"."}], "task": "refering", "answer_template": "The \"a cruise ship with tourists in the foreground and two cruise ships in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.43723437499999995, 0.2876470588235294, 0.9296875, 0.4696705882352942], "iscrowd": 0, "area": 15183.731299999992, "rle": {"size": [425, 640], "counts": "W\\d31W=2O1N1O2N1O2O1N1O2N1O2O0O2N1O2O0O2N1O2N101N1O2N1O2O0O1O1O1O100O1O2N1O100O1O1O100O00O1O1O2N100000001O0O1000001O0003M3N2N2N2N2O1N1O001O0000000001O000001O00000001O000001O00000001O00000001O00000001O000001O00000001O00000001O000001O00000001O00000001N1O1O100O1O1O1O2M2O1N2N2O1N2N2O1N3O02N3M2N3M2N3M2N3L3M3N2M3N1NO2M2N2NN23M3M3M3M3M3M3O12N3M2N3M2N3M2N3M2N3L3N2N1O000000000000000000000000000000O2O1O001O1O001O1O001O1O1O0000O1O1O1O1O100O1O1O1O1O1O1N2O100000000001O00000SNPEa1P;^NQEb1o:\\NSEd1m:[NUEd1k:[NVEe1j:YNXEg1h:XNYEh1g:WN[Eh1S;0001N1000001O000000001O00000000001O000000001O00000000001O0000001O001O1O0O2O001O1O001O001O001O1O001O001O1O001N101O1N3N1N2O1O1N3N1N2O1O2M2O1N2O2N1N2O1N2O2Nn_b0"}, "label": "a cruise ship with tourists in the foreground and two cruise ships in the background"}]}
{"id": 223095, "image": "COCO_train2014_000000223095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white cruise ship closest to the beach front\"."}], "task": "refering", "answer_template": "The \"the white cruise ship closest to the beach front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.43723437499999995, 0.2876470588235294, 0.9296875, 0.4696705882352942], "iscrowd": 0, "area": 15183.731299999992, "rle": {"size": [425, 640], "counts": "W\\d31W=2O1N1O2N1O2O1N1O2N1O2O0O2N1O2O0O2N1O2N101N1O2N1O2O0O1O1O1O100O1O2N1O100O1O1O100O00O1O1O2N100000001O0O1000001O0003M3N2N2N2N2O1N1O001O0000000001O000001O00000001O000001O00000001O00000001O00000001O000001O00000001O00000001O000001O00000001O00000001N1O1O100O1O1O1O2M2O1N2N2O1N2N2O1N3O02N3M2N3M2N3M2N3L3M3N2M3N1NO2M2N2NN23M3M3M3M3M3M3O12N3M2N3M2N3M2N3M2N3L3N2N1O000000000000000000000000000000O2O1O001O1O001O1O001O1O1O0000O1O1O1O1O100O1O1O1O1O1O1N2O100000000001O00000SNPEa1P;^NQEb1o:\\NSEd1m:[NUEd1k:[NVEe1j:YNXEg1h:XNYEh1g:WN[Eh1S;0001N1000001O000000001O00000000001O000000001O00000000001O0000001O001O1O0O2O001O1O001O001O001O1O001O001O1O001N101O1N3N1N2O1O1N3N1N2O1O2M2O1N2O2N1N2O1N2O2Nn_b0"}, "label": "the white cruise ship closest to the beach front"}]}
{"id": 272255, "image": "COCO_train2014_000000272255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car of the cross walk\"."}], "task": "refering", "answer_template": "The \"a white car of the cross walk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241], "bbox": [0.4736902485659656, 0.2580625, 1.0, 0.5376406250000001], "iscrowd": 0, "area": 26597.453949999992, "rle": {"size": [640, 523], "counts": "`Xk42nc02N1O2N2N2M3N2N1O2N2N2N2M2O2N2N2N2N2M2O2N2N2N1O0O2O000000000O101O0000000000001O00000000001O000bN]N_@c1a?^N^@c1a?^N^@e1_?\\N`@g1\\?[Nc@g1[?ZNd@i1Y?XNf@k1W?VNh@m1T?UNk@m1S?TNl@l1T?UNk@k1T?WNk@i1U?XNj@h1V?YNi@h1V?YNi@g1V?[Nh@f1X?[Ne@g1[?ZNb@h1^?ZN^@i1`?YN]@i1c?XNZ@j1f?WNW@k1i?VNU@k1j?WNT@k1k?VNT@j1l?WNR@j1m?XNR@h1n?U101O0O1000001N10000O10O1000O1000O10O1000O10O10000O0100000O10O1000O10O10O100000O01000O10O0100O100O10O0100O100O100O010O10000001O0O2O001O001O0O2O001O0000O10O0100O100O010O10O010O010O01N2L3N3L3M4L4N101N1O2N2N1O2N1O2O010000O10O010000O010O1000O0100O100O01000O10O010M3M2O2M2N3M2O2N2N1O2N1O2N2N1O2O10O10O100000000O1000O10O1000000O1002N1]_OfM]?]2`@cMa?_2[@cMe?^2X@dMh?^2T@cMm?^2Q@cMo?_2m_OcMS`0^2j_OcMW`0_2f_ObMZ`0_2c_ObM]`0a2__OaMa`0o20O10000O100N2N2N101N2N2N2N200001O0O100001O001O010RNeL"}, "label": "a white car of the cross walk"}]}
{"id": 272255, "image": "COCO_train2014_000000272255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car sitting at an intersection\"."}], "task": "refering", "answer_template": "The \"a white car sitting at an intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241], "bbox": [0.4736902485659656, 0.2580625, 1.0, 0.5376406250000001], "iscrowd": 0, "area": 26597.453949999992, "rle": {"size": [640, 523], "counts": "`Xk42nc02N1O2N2N2M3N2N1O2N2N2N2M2O2N2N2N2N2M2O2N2N2N1O0O2O000000000O101O0000000000001O00000000001O000bN]N_@c1a?^N^@c1a?^N^@e1_?\\N`@g1\\?[Nc@g1[?ZNd@i1Y?XNf@k1W?VNh@m1T?UNk@m1S?TNl@l1T?UNk@k1T?WNk@i1U?XNj@h1V?YNi@h1V?YNi@g1V?[Nh@f1X?[Ne@g1[?ZNb@h1^?ZN^@i1`?YN]@i1c?XNZ@j1f?WNW@k1i?VNU@k1j?WNT@k1k?VNT@j1l?WNR@j1m?XNR@h1n?U101O0O1000001N10000O10O1000O1000O10O1000O10O10000O0100000O10O1000O10O10O100000O01000O10O0100O100O10O0100O100O100O010O10000001O0O2O001O001O0O2O001O0000O10O0100O100O010O10O010O010O01N2L3N3L3M4L4N101N1O2N2N1O2N1O2O010000O10O010000O010O1000O0100O100O01000O10O010M3M2O2M2N3M2O2N2N1O2N1O2N2N1O2O10O10O100000000O1000O10O1000000O1002N1]_OfM]?]2`@cMa?_2[@cMe?^2X@dMh?^2T@cMm?^2Q@cMo?_2m_OcMS`0^2j_OcMW`0_2f_ObMZ`0_2c_ObM]`0a2__OaMa`0o20O10000O100N2N2N101N2N2N2N200001O0O100001O001O010RNeL"}, "label": "a white car sitting at an intersection"}]}
{"id": 272255, "image": "COCO_train2014_000000272255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in profile with a black dress and a gray umbrella\"."}], "task": "refering", "answer_template": "The \"a woman in profile with a black dress and a gray umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 287, 288, 289, 306, 307, 308, 309, 325, 326, 327, 328, 344, 345, 346, 347, 363, 364, 365, 366, 382, 383, 384, 385, 401, 402, 403, 404, 420, 421, 422, 423], "bbox": [0.12736137667304015, 0.6463281249999999, 0.29131931166347996, 1.0], "iscrowd": 0, "area": 14595.931950000004, "rle": {"size": [640, 523], "counts": "fZZ15ic07J4L0Q]OB\\b0?b]OEh3OY:=lAHb33`:6lAK[37i:MjA1R3;U;CfA6l2?\\:lNdC>RO;e2c0b:iNeCX1_17i:fNgCS1o0g0W;ZNiCP18]1m;gMiCm4W<UKhCj4W<XKhCh4W<YKiCg4V<[KjCd4V<]KiCc4W<]KhCd4X<]KgCc4Z<\\KeCe4[<]KbCd4^<]K`Cd4a<\\K]Ce4c<\\K[Ce4f<[KXCf4h<[KVCf4j<X11O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O001O001O001O002N2N2N2YL\\Bn0f=kNiBk0Y=oNmBo0U=jNRCT1o<gNWC0T>HRB1a>\\OfA=n>POXAi0Ta0I7I7H8IfZW7"}, "label": "a woman in profile with a black dress and a gray umbrella"}]}
{"id": 272255, "image": "COCO_train2014_000000272255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a woman in black holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 287, 288, 289, 306, 307, 308, 309, 325, 326, 327, 328, 344, 345, 346, 347, 363, 364, 365, 366, 382, 383, 384, 385, 401, 402, 403, 404, 420, 421, 422, 423], "bbox": [0.12736137667304015, 0.6463281249999999, 0.29131931166347996, 1.0], "iscrowd": 0, "area": 14595.931950000004, "rle": {"size": [640, 523], "counts": "fZZ15ic07J4L0Q]OB\\b0?b]OEh3OY:=lAHb33`:6lAK[37i:MjA1R3;U;CfA6l2?\\:lNdC>RO;e2c0b:iNeCX1_17i:fNgCS1o0g0W;ZNiCP18]1m;gMiCm4W<UKhCj4W<XKhCh4W<YKiCg4V<[KjCd4V<]KiCc4W<]KhCd4X<]KgCc4Z<\\KeCe4[<]KbCd4^<]K`Cd4a<\\K]Ce4c<\\K[Ce4f<[KXCf4h<[KVCf4j<X11O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O001O001O001O002N2N2N2YL\\Bn0f=kNiBk0Y=oNmBo0U=jNRCT1o<gNWC0T>HRB1a>\\OfA=n>POXAi0Ta0I7I7H8IfZW7"}, "label": "a woman in black holding an umbrella"}]}
{"id": 272255, "image": "COCO_train2014_000000272255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the red striped shirt in the crosswalk\"."}], "task": "refering", "answer_template": "The \"the man in the red striped shirt in the crosswalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 37, 53, 54, 56, 72, 73, 75, 91, 92, 109, 110, 111, 112, 113, 131, 149, 168, 169, 170, 187, 188, 205, 206, 207, 223, 224, 225, 226, 243, 244, 245, 262, 263, 264, 281, 282, 283, 291, 310, 311, 312, 329, 330, 331, 348, 349, 350, 367, 368, 386, 387, 405, 406, 424, 425], "bbox": [0.31011472275334606, 0.05709375, 1.0, 0.97725], "iscrowd": 0, "area": 17713.248850000007, "rle": {"size": [640, 523], "counts": "dYV33mc02N2N1O11O1O0Ok\\O<Pb0Ef]Od0Zb0[Oa]Ok0_b0;10R_OiNP?X1n@mNo>U1l@QO2JR=W1hBTOMI\\=V1bBWOFJi=Q1]B[O_OJU>l0ZB_OWOJa>f0VBj0k=VOTBj0n=UOPBl0Q>TOmAl0V>TOgAm0Z>TOcAm0]>TObAl0_>UO^Ak0d>UOZAl0e>VOYAk0g>VOWAk0i>UOVAZOK7P?`0TAVO:Nc>l0RASOj0EU>Y1o@oNU1Bm=_1XCYNm<c1j2L4M2Lg_OmNo=S1lAAe=>WBFj=:QBJP>7iABd>a0TARO^?e23N2N2N101O1O1O1N2O1F:K5L4M2M4ZOf0N2N2O1O1N101O1O1O1O10O2O001O1OIhV[3XOmZdLMe:7VENPLJW>=gE9n91PF3j7VOlEk0Y21h7WOlEj0[20g7YOlEg0\\22f7YOlEf0^22d7[OlEb0a24]M\\OS80lGa0e24TMGW8DnGb0g24nLG`8E\\GISOh0S44iLEi8HQGQ1]3DdLBS9CTFI`0c1g3_O^LBZ9DRFH>`1k3CWLB_9DRFG=\\1n3HPLBe9BRFG<[1m3LmK_Ok9CPFG;Y1m30iK^OR:AnEH;V1m35dKfN08]:MjEJ7V1n3=d6_N[EN3T1o3?X7^NbDX1W4:P7\\NcD31Y1a0iNm2`1a7YNUE5I11Y1=oNS3Z1`7iNnDF1W15[O[3P1`7iNoDS16XOZ3m0X7WN[Ec0Mc0>KW1QO0h1m8\\N`E>O<=6T1nN3h1g8YOdE153\\1kN5c2k7_N_Fa1`1^M6c2j7_NaF^1_1`M5e2j7W1QHRL7h2g7V1RHQL8j2e7U1SHQL8k2d7T1THQL8n2a7Q1WHQL8j3f64RIRL8i3g6HfGaL\\1M7g3j6WOZHRMe016`3Q7lMmGWO=YO?55]3T7iLWGf0f0BR2BlN]3U7gLZGe0f0@S2HfN]3W7dL]Ge0R4ZOYMm2i7PMoFh0P4\\OWMU2Q9\\NkI^OSMg1b9UNRE:^4InLg1l9`NVII^L0WOg1X;_NSII[L5XOb1\\;cNnHGZL8[OS1i;kNeHIUL<\\Oe0Ua0oN\\_Oa0]O`0Va0PO[_Oi1f`0WNX_Oj1h`0VNW_Ol1i`0TNU_Om1k`0SNT_Oo1l`0QNS_Oo1S?nMXB3eNQ2P?mM\\B2bNV2j>lMeBMaNc2^>`MQCN`Ne2[>^MVCM]Nh2Z>[MZCM\\Ni2X>[M\\CM[Nj2V>YM`CMZNf3Y=^L^DLXNX4g<mKQEKXN^4a<fKXEMVN]4a<gKYEMUN\\4b<fKZENTN\\4b<fKZEOSN[4b<fKZE1TNV4e<hKWE3TNd3U=YLgD4TNd01V1h=QNTD5TNa03e0W>eNcC6TN;50n>_OiB6UN861m>BhB5VN574k>BgB6XN276i>CgB6[NN68i>DeB7^NJ5OS?1YB7aNF40T?2WB9[OAb>6RB:\\O^Oc>9PB9[?Gd@:[?Gd@:b=ETB19;_=HWBN9;\\=KZBK:9W=1^BH;5U=7\\BH`0OS=;VBMh0FR=>WBJh0GQ=`0XBGi0Gn<d0YBEi0Eo<g0XBCk0BP=k0UBBm0_OP=P1SB@o0^Oo<S1QB_OR`0c0m_O\\OQ`0^OP_OW1n0[OR`0^OP_OW1n0ZOR`0AP_OU1l0[OT`0@Q_OT1k0\\OS`0BR_OP1l0]Ok?@[_O40m0j0_Od?H`_ON2i0j0Bb?5d_O8j0G]?2k_O5h0KZ?5o_OJg02Z?6R@Cd06[?8\\@VO8b0R>"}, "label": "the man in the red striped shirt in the crosswalk"}]}
{"id": 124804, "image": "COCO_train2014_000000124804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair being sat in by an older man\"."}], "task": "refering", "answer_template": "The \"a chair being sat in by an older man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 282, 283, 284, 285, 308, 309, 329, 330, 331, 332, 352, 353, 354, 355, 378], "bbox": [0.27978125, 0.6045, 0.49214062500000005, 0.9842708333333334], "iscrowd": 0, "area": 4690.393600000002, "rle": {"size": [480, 640], "counts": "oWd23m>5K3M0O2O0000000O1000000O100mBDb;<^DLZ;4fD3R;NnD9l:FTE:l:FTE:l:FTE:l:FUE9k:GUE9k:GUE9k:GUE9k:HTE8l:HTE8l:HTE8m:GSE9m:GSE9m:GTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8m:GSE9m:GSE9m:GTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8l:ISE7n:HRE8n:HRE8n:HSE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7n:HRE8n:HRE8n:HSE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7n:IQE7o:IRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6o:IQE7o:IQE7o:IRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6o:IQE7o:IQE7o:JQE5o:KQE5o:KQE5o:KQE5o:KQE5o:KQE5o:3iDMW;c0YD]Oh;R1lCjNT<f1cCSN[:KaGb2Z8_MeGQ3k7QMTH\\3_7eLeHU3[7kLkHP3T7PMSIi2m6XMXIb2h6^M^I\\2b6eMcIU2]6kMjIn1V6SNoIh1P6XNVJb1k5]N\\J\\1d5eNaJU1_5kNgJo0Y5ROmJg0S5YOSKb0l4^OZK<f4E`K4`4LfKNZ43kKGV48Y51O00001O00001O01O0001O0000M3L4LnQh4"}, "label": "a chair being sat in by an older man"}]}
{"id": 124804, "image": "COCO_train2014_000000124804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the man is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the man is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 282, 283, 284, 285, 308, 309, 329, 330, 331, 332, 352, 353, 354, 355, 378], "bbox": [0.27978125, 0.6045, 0.49214062500000005, 0.9842708333333334], "iscrowd": 0, "area": 4690.393600000002, "rle": {"size": [480, 640], "counts": "oWd23m>5K3M0O2O0000000O1000000O100mBDb;<^DLZ;4fD3R;NnD9l:FTE:l:FTE:l:FTE:l:FUE9k:GUE9k:GUE9k:GUE9k:HTE8l:HTE8l:HTE8m:GSE9m:GSE9m:GTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8m:GSE9m:GSE9m:GTE8l:HTE8l:HTE8l:HTE8l:HTE8l:HTE8l:ISE7n:HRE8n:HRE8n:HSE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7n:HRE8n:HRE8n:HSE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7m:ISE7n:IQE7o:IRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6o:IQE7o:IQE7o:IRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6n:JRE6o:IQE7o:IQE7o:JQE5o:KQE5o:KQE5o:KQE5o:KQE5o:KQE5o:3iDMW;c0YD]Oh;R1lCjNT<f1cCSN[:KaGb2Z8_MeGQ3k7QMTH\\3_7eLeHU3[7kLkHP3T7PMSIi2m6XMXIb2h6^M^I\\2b6eMcIU2]6kMjIn1V6SNoIh1P6XNVJb1k5]N\\J\\1d5eNaJU1_5kNgJo0Y5ROmJg0S5YOSKb0l4^OZK<f4E`K4`4LfKNZ43kKGV48Y51O00001O00001O01O0001O0000M3L4LnQh4"}, "label": "the chair the man is sitting in"}]}
{"id": 124804, "image": "COCO_train2014_000000124804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older gentleman with a white dress shirt and a light colored tie on\"."}], "task": "refering", "answer_template": "The \"an older gentleman with a white dress shirt and a light colored tie on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 145, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 334, 355, 356, 357, 378, 379], "bbox": [0.190453125, 0.3101041666666667, 0.559546875, 0.9775208333333334], "iscrowd": 0, "area": 29510.8638, "rle": {"size": [480, 640], "counts": "S^i1=Y>;N3L4M2M4L3N3L4M2M4]N`N`Ec1[:cNbE`1Y:eNcE_1W:gNfE[1U:kNgEY1S:mNiEV1R:POkES1Q:POlET1Q:nNmET1Q:oNkEU1S:lNkEV1S:lNjEX1T:jNhEZ1U:hNiEZ1U:k1N2M2O2M2O2M3M2O2M2O2M3M2O2N1O2N2N1O2N1O2N1O00001O00000000000000000010OfGlIh7T6XHPJd7P6\\HSJa7m5_HWJ]7i5cH[JY7e5gH_JU7a5kHcJQ7]5oHeJo6\\5PIdJP7\\5PIeJo6[5QIfJn6Z5RIgJm6Y5SIgJm6Y5SIhJl6X5TIiJn6T5RIlJQ7Q5oHPKS7m4mHTKU7j4jHVKY7g4gHZK[7c4fH]K\\7`4dH`K_7]4aHdKa7Y4_HhKc7JQHY3<mLf7@VHb34oLh7WOZHi3NQMj7ROZHl3LSMl7mNZHo3JTMg8k2YGVMg8i2YGXMg8g2YGXMi8g2WGYMj8f2VGZMk8e2UG[Ml8d2TG\\Mm8c2SG]Mn8b2RG^Mo8a2QG_MP9`2PG`MQ9_2oFaMR9^2oF`MS9_2mFaMT9^2lFbMU9]2kFcMV9\\2jFdMW9[2iFeMX9Z2hFfMZ9X2fFhM[9W2eFiM\\9V2dFjM]9U2cFjM`9T2`FlMa9S2_FmMb9R2^FnMc9Q2]FoMe9o1[FQNf9n1ZFRNg9m1YFSNh9l1XFTNj9j1VFVNk9i1UFVNm9i1SFWNn9h1SFWNn9h1RFXNP:f1PFZNQ:e1oE[NR:d1nE\\NS:c1mE]NU:a1kE_NV:`1jE_NX:`1hE`NY:_1gEaN[:]1eEcN\\:\\1dEdN]:[1cEeN^:Z1bEfN_:Y1aEgNa:W1_EiNb:V1^EjNc:U1]EjNe:U1[EkNg:S1YEmNh:R1XEnNi:Q1WEoNj:P1WEoNk:o0UEQOl:n0TEROm:m0SESOn:l0RETOo:k0QETOQ;k0WEmNj:R1eE_N[:a1b101O001O00001O0007J6I8H7J6I8H5L0O10O01O100O1mElLX8T3dGPM]8P3]GUMc8l2XGYMg8h2TG\\Mm8c2oFaMQ9`2iFeMX9[2cFjM\\9W2_FmMa9S2[FQNf9o1UFUNk9a3O01O00e0\\Ol0SO9G2N101N1O2O0O1O2N101N1O1N3M2N3L3N2N3L3N3M2M3N3M2M4M2N2M4M2N3L3N2N3L3N3M2M3N3M2M4M2N3L3NhhS4"}, "label": "an older gentleman with a white dress shirt and a light colored tie on"}]}
{"id": 124804, "image": "COCO_train2014_000000124804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white shirt sitting in a chair\"."}], "task": "refering", "answer_template": "The \"a man wearing a white shirt sitting in a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 145, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 334, 355, 356, 357, 378, 379], "bbox": [0.190453125, 0.3101041666666667, 0.559546875, 0.9775208333333334], "iscrowd": 0, "area": 29510.8638, "rle": {"size": [480, 640], "counts": "S^i1=Y>;N3L4M2M4L3N3L4M2M4]N`N`Ec1[:cNbE`1Y:eNcE_1W:gNfE[1U:kNgEY1S:mNiEV1R:POkES1Q:POlET1Q:nNmET1Q:oNkEU1S:lNkEV1S:lNjEX1T:jNhEZ1U:hNiEZ1U:k1N2M2O2M2O2M3M2O2M2O2M3M2O2N1O2N2N1O2N1O2N1O00001O00000000000000000010OfGlIh7T6XHPJd7P6\\HSJa7m5_HWJ]7i5cH[JY7e5gH_JU7a5kHcJQ7]5oHeJo6\\5PIdJP7\\5PIeJo6[5QIfJn6Z5RIgJm6Y5SIgJm6Y5SIhJl6X5TIiJn6T5RIlJQ7Q5oHPKS7m4mHTKU7j4jHVKY7g4gHZK[7c4fH]K\\7`4dH`K_7]4aHdKa7Y4_HhKc7JQHY3<mLf7@VHb34oLh7WOZHi3NQMj7ROZHl3LSMl7mNZHo3JTMg8k2YGVMg8i2YGXMg8g2YGXMi8g2WGYMj8f2VGZMk8e2UG[Ml8d2TG\\Mm8c2SG]Mn8b2RG^Mo8a2QG_MP9`2PG`MQ9_2oFaMR9^2oF`MS9_2mFaMT9^2lFbMU9]2kFcMV9\\2jFdMW9[2iFeMX9Z2hFfMZ9X2fFhM[9W2eFiM\\9V2dFjM]9U2cFjM`9T2`FlMa9S2_FmMb9R2^FnMc9Q2]FoMe9o1[FQNf9n1ZFRNg9m1YFSNh9l1XFTNj9j1VFVNk9i1UFVNm9i1SFWNn9h1SFWNn9h1RFXNP:f1PFZNQ:e1oE[NR:d1nE\\NS:c1mE]NU:a1kE_NV:`1jE_NX:`1hE`NY:_1gEaN[:]1eEcN\\:\\1dEdN]:[1cEeN^:Z1bEfN_:Y1aEgNa:W1_EiNb:V1^EjNc:U1]EjNe:U1[EkNg:S1YEmNh:R1XEnNi:Q1WEoNj:P1WEoNk:o0UEQOl:n0TEROm:m0SESOn:l0RETOo:k0QETOQ;k0WEmNj:R1eE_N[:a1b101O001O00001O0007J6I8H7J6I8H5L0O10O01O100O1mElLX8T3dGPM]8P3]GUMc8l2XGYMg8h2TG\\Mm8c2oFaMQ9`2iFeMX9[2cFjM\\9W2_FmMa9S2[FQNf9o1UFUNk9a3O01O00e0\\Ol0SO9G2N101N1O2O0O1O2N101N1O1N3M2N3L3N2N3L3N3M2M3N3M2M4M2N2M4M2N3L3N2N3L3N3M2M3N3M2M4M2N3L3NhhS4"}, "label": "a man wearing a white shirt sitting in a chair"}]}
{"id": 124804, "image": "COCO_train2014_000000124804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green suit jacket\"."}], "task": "refering", "answer_template": "The \"a woman in a green suit jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 240, 241], "bbox": [0.383453125, 0.24775, 0.581078125, 0.6058541666666667], "iscrowd": 0, "area": 11281.7614, "rle": {"size": [480, 640], "counts": "VQc3h0V>8I7Ii0VO5L2N1N3N2N1N3N2N1N3O0O2O1N1O2O1N3M2O1N3N1N3M2O2M2N2OVNoDEk:?YE_Oa:g0`EXOZ:m0iEROQ:S1PFlNl9W1WFgNg9Z1[FeNd9\\1]FdN`9]1aFcN]9_1dF`N[9`1gF_NW9b1kF^NR9d1oF[NP9e1QG[Nm8g1TGXNj8j1WGVNg8k1YGUNf8m1YGSNg8m1YGSNf8o1ZGPNf8Q2YGoMf8R2ZGnMf8S2YGmMg8S2YGmMf8U2YGkMg8U2YGkMf8W2YGiMg8X2XGiMf8X2[GgMe8Z2ZGfMf8[2YGeMg8\\2YGcMh8]2XGbMh8_2XG`Mh8a2XG^Mi8b2WG]Mi8f2UGYMk8i2TGVMm8l2QGTMn8m2SGQMm8P3SGoLm8R3SGmLn8S3RGjLP9W3PGfLT9Y3lFdLY9Z3gFcL^9[3cF`Lc9^3]F_Li9^3o0N9G7H2O1O2N1O1N1000000F:TOl0K5K5L4N2NmBEe;8SD1n;MoC7R<GnC:T<ClC>U<@kC`0W<]OjCd0X<YOgCh0]<SOdCn0^<oNbCQ1V=1O001O0O101O0101N2O0O3M3N3L4L4L3M4L4L4L4La\\m3"}, "label": "a woman in a green suit jacket"}]}
{"id": 124804, "image": "COCO_train2014_000000124804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a pea green coat sits next to a man during a radio show\"."}], "task": "refering", "answer_template": "The \"a lady in a pea green coat sits next to a man during a radio show\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 240, 241], "bbox": [0.383453125, 0.24775, 0.581078125, 0.6058541666666667], "iscrowd": 0, "area": 11281.7614, "rle": {"size": [480, 640], "counts": "VQc3h0V>8I7Ii0VO5L2N1N3N2N1N3N2N1N3O0O2O1N1O2O1N3M2O1N3N1N3M2O2M2N2OVNoDEk:?YE_Oa:g0`EXOZ:m0iEROQ:S1PFlNl9W1WFgNg9Z1[FeNd9\\1]FdN`9]1aFcN]9_1dF`N[9`1gF_NW9b1kF^NR9d1oF[NP9e1QG[Nm8g1TGXNj8j1WGVNg8k1YGUNf8m1YGSNg8m1YGSNf8o1ZGPNf8Q2YGoMf8R2ZGnMf8S2YGmMg8S2YGmMf8U2YGkMg8U2YGkMf8W2YGiMg8X2XGiMf8X2[GgMe8Z2ZGfMf8[2YGeMg8\\2YGcMh8]2XGbMh8_2XG`Mh8a2XG^Mi8b2WG]Mi8f2UGYMk8i2TGVMm8l2QGTMn8m2SGQMm8P3SGoLm8R3SGmLn8S3RGjLP9W3PGfLT9Y3lFdLY9Z3gFcL^9[3cF`Lc9^3]F_Li9^3o0N9G7H2O1O2N1O1N1000000F:TOl0K5K5L4N2NmBEe;8SD1n;MoC7R<GnC:T<ClC>U<@kC`0W<]OjCd0X<YOgCh0]<SOdCn0^<oNbCQ1V=1O001O0O101O0101N2O0O3M3N3L4L4L3M4L4L4L4La\\m3"}, "label": "a lady in a pea green coat sits next to a man during a radio show"}]}
{"id": 206731, "image": "COCO_train2014_000000206731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear in front of all of the other bears\"."}], "task": "refering", "answer_template": "The \"a bear in front of all of the other bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 38, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 266, 267, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 293, 294, 295, 307, 308, 309, 310, 311, 312, 316, 317, 318, 330, 331, 332, 333, 334, 340, 341], "bbox": [0.123890625, 0.0002107728337236534, 0.8552500000000001, 0.98903981264637], "iscrowd": 0, "area": 84609.22844999998, "rle": {"size": [427, 640], "counts": "VVQ13S=7I7J7L3M3M3M2N2N2M3N3M2N2ON1O2M201O001O001O001O0010O010O010O010O010O010O100O010O10O0100O010O10O10O10000O1000000O01000O101O000O2O000O2O00001N101O0O101O001N101O0O2O00001N101OO1000000O101O00000O10001O0O100000000O1000000O100000000O0100O1O0O2O100O10000O1000000O10000O10000O10O1000O1000000O0100O1O10iITNn0l1jN]NV1c1`KVNY1`0W3Y1PKkNd1N\\3V1oJQO`1Ka3S1nJWO]1Gd3S1mJ_OV1@m3P1lJ1f0QO^4m0kJ><fNi4l0iJd07cNo4i0iJk01^Nh1[O<[1`M]14PNf1F@DhNY1Lc10lMh1L\\OEjNQ1Oi1MgMj13WODnNi02P2HcMl19TODPOa05`3c1cLQOZ1ZOV2`1iLQOR1]OW2^1oLoN^65iG^OR27V67jJDW58P4L4L400000000001O00000000001O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1N2O2N1O2N1O1O2I6@`0M4_OkMgEW2X:nMaEV2]:=O1O2N1O1O2N0000000001M2O1N2N3M2O1N2N3M2O2M3M3M3N2M2mNfD3];JeD4^;IdD6^;GdD7_;FcD8`;EbD9Y<O1N2N2N2O1N2N2N2O100000000000000000000000O10000\\Od0M3N2N5K5L4L5K4M3L5K4M3L4L5K4M3M3M4L3M3L2O000000L4L4K5L3M4L4N2O1N2N2O1N1010M2J7J5J9H8G:G9I6J4L4M2eNlK]IX4j3lKcM7TNP4W4mKdM=gMj3b4nKeMQ5W2UKiMh4o1cKPN\\4g1oKXNP4c1VL]Nj3`1ZL_Nf3_1]L`Nc3]1aLbN_3[1eLeNZ3X1jLgNV3V1mLjNS3S1QMlNo2R1TMmNl2P1XMoNh2n0\\MQOe2k0_MUOb2f0aMZO`2b0dM]O]2`0fM_O]2;fMEk2GXM9m2_OWM`0n2WOVMj0n2oNUMP1P3iNRMW1S3cNoL]1T3]NnLd1V3UNlLl1X3nMjLR2W7100O1O100O100O1O1O1O00100O001O00iMUEo1j:nMZER2f:jM]EV2m:000O010O10001N2O1N2O1N2O1O1N2O2M2O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1N3N2N1O2N2N1N2O0O100O100O100O1O100O100O100O100001O1O001N2O1O0O2O1N1O2N2N2N101N2N2N2N1O2O1N2N2N2N2N2O1N2N2N2M3N2M3N2M3N2M5J7I7I7H8I7F:Eb]V1"}, "label": "a bear in front of all of the other bears"}]}
{"id": 206731, "image": "COCO_train2014_000000206731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bear dressed as the statue of liberty with a shirt that says new york on it\"."}], "task": "refering", "answer_template": "The \"a white bear dressed as the statue of liberty with a shirt that says new york on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 38, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 266, 267, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 293, 294, 295, 307, 308, 309, 310, 311, 312, 316, 317, 318, 330, 331, 332, 333, 334, 340, 341], "bbox": [0.123890625, 0.0002107728337236534, 0.8552500000000001, 0.98903981264637], "iscrowd": 0, "area": 84609.22844999998, "rle": {"size": [427, 640], "counts": "VVQ13S=7I7J7L3M3M3M2N2N2M3N3M2N2ON1O2M201O001O001O001O0010O010O010O010O010O010O100O010O10O0100O010O10O10O10000O1000000O01000O101O000O2O000O2O00001N101O0O101O001N101O0O2O00001N101OO1000000O101O00000O10001O0O100000000O1000000O100000000O0100O1O0O2O100O10000O1000000O10000O10000O10O1000O1000000O0100O1O10iITNn0l1jN]NV1c1`KVNY1`0W3Y1PKkNd1N\\3V1oJQO`1Ka3S1nJWO]1Gd3S1mJ_OV1@m3P1lJ1f0QO^4m0kJ><fNi4l0iJd07cNo4i0iJk01^Nh1[O<[1`M]14PNf1F@DhNY1Lc10lMh1L\\OEjNQ1Oi1MgMj13WODnNi02P2HcMl19TODPOa05`3c1cLQOZ1ZOV2`1iLQOR1]OW2^1oLoN^65iG^OR27V67jJDW58P4L4L400000000001O00000000001O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1N2O2N1O2N1O1O2I6@`0M4_OkMgEW2X:nMaEV2]:=O1O2N1O1O2N0000000001M2O1N2N3M2O1N2N3M2O2M3M3M3N2M2mNfD3];JeD4^;IdD6^;GdD7_;FcD8`;EbD9Y<O1N2N2N2O1N2N2N2O100000000000000000000000O10000\\Od0M3N2N5K5L4L5K4M3L5K4M3L4L5K4M3M3M4L3M3L2O000000L4L4K5L3M4L4N2O1N2N2O1N1010M2J7J5J9H8G:G9I6J4L4M2eNlK]IX4j3lKcM7TNP4W4mKdM=gMj3b4nKeMQ5W2UKiMh4o1cKPN\\4g1oKXNP4c1VL]Nj3`1ZL_Nf3_1]L`Nc3]1aLbN_3[1eLeNZ3X1jLgNV3V1mLjNS3S1QMlNo2R1TMmNl2P1XMoNh2n0\\MQOe2k0_MUOb2f0aMZO`2b0dM]O]2`0fM_O]2;fMEk2GXM9m2_OWM`0n2WOVMj0n2oNUMP1P3iNRMW1S3cNoL]1T3]NnLd1V3UNlLl1X3nMjLR2W7100O1O100O100O1O1O1O00100O001O00iMUEo1j:nMZER2f:jM]EV2m:000O010O10001N2O1N2O1N2O1O1N2O2M2O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1N3N2N1O2N2N1N2O0O100O100O100O1O100O100O100O100001O1O001N2O1O0O2O1N1O2N2N2N101N2N2N2N1O2O1N2N2N2N2N2O1N2N2N2M3N2M3N2M3N2M5J7I7I7H8I7F:Eb]V1"}, "label": "a white bear dressed as the statue of liberty with a shirt that says new york on it"}]}
{"id": 206731, "image": "COCO_train2014_000000206731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear with cap near to eyes\"."}], "task": "refering", "answer_template": "The \"teddy bear with cap near to eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 18, 38, 39, 40, 41, 61, 62, 63, 64, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 205, 221, 224, 225, 226, 227, 228, 244, 245, 246, 248, 250, 251, 267, 268, 269, 290, 291, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.54525, 0.0, 0.9512812499999999, 1.0], "iscrowd": 0, "area": 44979.022500000014, "rle": {"size": [427, 640], "counts": "Qna46P=8H5K5K5N2N2O1N2O1N2O100O100O1O100O100O100O100O10000O100000000O100000000O1M3J6J6I7J6J6I7J6J6hNjLTH]3j7kLjG[3T8SMZGS3d8k0jLfK[MZ4_2VLXMi3e2\\LYMd3c2bL\\M]3`2hL_MX3]2nLaMR3[2SMeMl2W2ZMgMf2U2`MjM_2S2eMlM[2P2kMoMT2m1PNSNP2i1UNVNk1f1YN[Nf1a1_N^Na1_1cNaN\\1\\1gNdNY1Y1kNgNT1V1oNjNQ1R1TOnNk0o0XOQOh0l0\\OSOd0j0@VO?fNWMD]2f1<aN^MC[2l16\\NgMBV2R23XNV2h1iMSN]2l1cMoMb2R2\\4O0010O0100O10O01O10jHoMQ3P2YJZNm0Fj4Q2XJ\\Nk0Dm4P2WJ_Ni0BP5o1VJbNg0@S5n1UJeNe0^OU5n1UJgNc0\\OX5m1TJjNa0ZO[5l1SJlN?\\O\\5h1TJoN=[O]5h1TJPO<[O^5e1UJSO:ZO`5c1UJVO8ZOa5`1VJYO6YOb5_1WJ[O4YOc5^1VJ\\O4XOe5^1TJ]O4XOf5]1SJ^O4WOg5^1RJ]O5XOg5]1QJ^O5WOh5^1PJ^O5WOi5^1nI^O6VOk5^1lI_O5WOl5[1nIA3VOn5Y1nID1VOn5W1PJE0VOo5U1PJHNVOP6Q1RJf1k5o0VJ]Li5^3[JeLb5W3bJkL]5P3gJSMW5h2mJ[MP5a2TKaMk4Z2YKeMh4W2\\KhMf4S2^KlMd4o1`KPNa4l1cKRN`4i1dKVN^4e1fKZN[4b1iK]NY4^1kKaNV4]1lKbNV4_1hK_N[4c1bK\\N`4e1^KZNd4g1ZKXNi4i1TKUNo4l1oJQNU5P2iJnMZ5T2cJjM`5X2]JeMg5[2XJcMk5\\2UJbMn5^2QJ_MS6a2lI]MW6b2iI[M[6e2dIYM_6f2aIXMb6h2]IUMg6k2XISMl6k2TISMo6m2PIPMT7o2lHoLW7Q3hHlL\\7T3cHjL`7U3`HiLc7W3\\HfLh7Y3XHeLk7[3THcLo7\\3Q100O100O2NkFfLZ8Z3eGfLY8\\3gGeLc3b0Bk2kLcLZ3k0He2nLaLR3Q1Na2oL^Ll2X13\\2QM]Ld2^19W2SM[L]2e1>R2UMZLX2h1a0Q2VMkLa1X1W1o1XMXMP1k0f1o1ZM^Mf0e0m1P2]McM<?U2Q2^MhM2:^2P2`MnMH4c2R2fMRN]ONh2U2kM`0Q2CPN=_13bNM[16eNJ[15fNKZ15fNK[13fNMZ12gNM[12eNN[11fNO[12cNN]15`NL`18[NHi18SNIo15PNKT21lMOW2OhM2Z2LeM4_2I`M8b2F]M:g2CXM=j2BUM?l2@SM`0n2@QMa0P3^OoLb0S3]OlLc0V3\\OiLe0X3ZOgLf0[3YOdLh0]3WObLi0_3VOaLk0`3_42N2O100O1O001O10000000O1N101N2N2N2N2O4K8H6J2O1N3M2N2N010O000010O0001O000O2O000M4L3L4L5E:A`0@?@if<"}, "label": "teddy bear with cap near to eyes"}]}
{"id": 206731, "image": "COCO_train2014_000000206731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"torch that the bear behind is holidng\"."}], "task": "refering", "answer_template": "The \"torch that the bear behind is holidng\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 29, 30, 31, 52, 53, 54, 75, 76, 77, 97, 98, 99, 100, 119, 120, 121, 122, 123, 143, 144, 145, 146, 167], "bbox": [0.200859375, 0.0, 0.39334375, 0.4830444964871194], "iscrowd": 0, "area": 13399.738800000001, "rle": {"size": [427, 640], "counts": "nme19Q=5J5L3M3M2N3M2N3N0O2N101N101O00001O00001O0001O01O00001OO1000M2G:N2O0O2nDcNU:Z2N3N1dMVMUJl2j5VMSJl2m5VMoIl2R6TMkIn2V6RMgIP3Z6QMdIo2]6QMaIP3`6PM_IP3b6PM\\IQ3e6PMYIP3h6PMVIQ3l6nLSIR3n6nLPIS3Q7nLmHR3T7RMeHP3\\7TMYHR3h7RMnGR3T8TM_GR3b8j02O2M2O1N2O1N2N3M2M3N2M3M4M2M^F`L_9f3ROgLhGZ3V8kLeGW3W8Q1L3VOQKUIQ5h6VKQIn4j6[KoHh4o6k00O100000000O1000000O1000001N2O1O1O1O1N01N2M30O100000000TL`Ik0b6ROhIb0]6[OlI9Z6DPJ]Oh6`0bInNi6o0aIeNd6X1i2M3oNZD:i;DiDIZ;4P1N2M3N2N^iQ5"}, "label": "torch that the bear behind is holidng"}]}
{"id": 206731, "image": "COCO_train2014_000000206731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"torch held by the second bear\"."}], "task": "refering", "answer_template": "The \"torch held by the second bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 29, 30, 31, 52, 53, 54, 75, 76, 77, 97, 98, 99, 100, 119, 120, 121, 122, 123, 143, 144, 145, 146, 167], "bbox": [0.200859375, 0.0, 0.39334375, 0.4830444964871194], "iscrowd": 0, "area": 13399.738800000001, "rle": {"size": [427, 640], "counts": "nme19Q=5J5L3M3M2N3M2N3N0O2N101N101O00001O00001O0001O01O00001OO1000M2G:N2O0O2nDcNU:Z2N3N1dMVMUJl2j5VMSJl2m5VMoIl2R6TMkIn2V6RMgIP3Z6QMdIo2]6QMaIP3`6PM_IP3b6PM\\IQ3e6PMYIP3h6PMVIQ3l6nLSIR3n6nLPIS3Q7nLmHR3T7RMeHP3\\7TMYHR3h7RMnGR3T8TM_GR3b8j02O2M2O1N2O1N2N3M2M3N2M3M4M2M^F`L_9f3ROgLhGZ3V8kLeGW3W8Q1L3VOQKUIQ5h6VKQIn4j6[KoHh4o6k00O100000000O1000000O1000001N2O1O1O1O1N01N2M30O100000000TL`Ik0b6ROhIb0]6[OlI9Z6DPJ]Oh6`0bInNi6o0aIeNd6X1i2M3oNZD:i;DiDIZ;4P1N2M3N2N^iQ5"}, "label": "torch held by the second bear"}]}
{"id": 206731, "image": "COCO_train2014_000000206731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear upside down\"."}], "task": "refering", "answer_template": "The \"teddy bear upside down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229], "bbox": [0.8096249999999999, 0.0022482435597189696, 1.0, 0.660679156908665], "iscrowd": 0, "area": 21281.021799999995, "rle": {"size": [427, 640], "counts": "SUh61X=000O10001L5I8H7I8H7kLNfH=T7n0ZG@`8h201O001O001O00001O001O001O00010O010O01000O010O10O10O02O001N2O001N2O0O0100O010O1O010O10O01VOYKoHf4Q7\\KmHe4R7]KlHd4S7^KlHa4S7cKiH_4V7cKfH`4Y7bKdH_4\\7cKaH_4^7cK_H_4`7g0O1O1O2N1O1O1O1O1O1O1O1O1N2O1O1O1O1O3M3M4L3M3M3gJUIl3V8O1O1O1O1O1O001O1O1O1O1O1N2O1O1OO1O1O1O1O1O1N2Oh0XOe0[Od0YOh0YOU2kMg0XOh0YOVJ"}, "label": "teddy bear upside down"}]}
{"id": 468876, "image": "COCO_train2014_000000468876.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown piece of wooden furniture next to the desk\"."}], "task": "refering", "answer_template": "The \"brown piece of wooden furniture next to the desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 42, 43, 44, 65, 66, 67, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 180, 181, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 252], "bbox": [0.83953125, 0.010125, 1.0, 0.6272500000000001], "iscrowd": 0, "area": 18007.413749999996, "rle": {"size": [480, 640], "counts": "gok7;`>5B?A>B>B>B?A>B>B?A>B>B>B?B=B>B>B?A>B>B>L5O0000001O0000001O0000001O0000001O000000WOZJYHe5g7j01O0000[LYJQNg5o1`KkL_4U3lK`LT4`3mK_LS4a3mK_LS4a3mK`LR4a3mK_LS4a3mK_LS4a3mK_LS4a3mK`LR4`3nK`LR4a3mK_LS4a3mK_LS4a3nK_Ll6f0TIZOl6f0TIZOl6g0SIYOm6g0SIZOl6f0TIZOl6f0TIZOl6f0TIZOl6f0TI[Ok6f0TIZOl6f0TIZOl6f0TIZOm6e0SI\\Ol6d0TI\\Ol6e0SI[Om6e0SI[Om6e0SI[Om6e0SI\\Om6c0SI]On6c0QI]Oo6c0QI]OP7b0PI_OP7`0PI@P7`0oHAR7>nHBS7>kHDU7;jHFV7:iHGX78gHIZ76eHL[74cHM^72`H0a7O^H2c7M\\H5d7J[H7f7IXH8i7GVH:k7ESH>m7ARH`0o7_OPHb0Q8^OnGb0S8]OnGc0R8\\OnGd0S8[OnGd0Z4"}, "label": "brown piece of wooden furniture next to the desk"}]}
{"id": 468876, "image": "COCO_train2014_000000468876.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair to the right of the girl in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the chair to the right of the girl in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 42, 43, 44, 65, 66, 67, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 180, 181, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 252], "bbox": [0.83953125, 0.010125, 1.0, 0.6272500000000001], "iscrowd": 0, "area": 18007.413749999996, "rle": {"size": [480, 640], "counts": "gok7;`>5B?A>B>B>B?A>B>B?A>B>B>B?B=B>B>B?A>B>B>L5O0000001O0000001O0000001O0000001O000000WOZJYHe5g7j01O0000[LYJQNg5o1`KkL_4U3lK`LT4`3mK_LS4a3mK_LS4a3mK`LR4a3mK_LS4a3mK_LS4a3mK_LS4a3mK`LR4`3nK`LR4a3mK_LS4a3mK_LS4a3nK_Ll6f0TIZOl6f0TIZOl6g0SIYOm6g0SIZOl6f0TIZOl6f0TIZOl6f0TIZOl6f0TI[Ok6f0TIZOl6f0TIZOl6f0TIZOm6e0SI\\Ol6d0TI\\Ol6e0SI[Om6e0SI[Om6e0SI[Om6e0SI\\Om6c0SI]On6c0QI]Oo6c0QI]OP7b0PI_OP7`0PI@P7`0oHAR7>nHBS7>kHDU7;jHFV7:iHGX78gHIZ76eHL[74cHM^72`H0a7O^H2c7M\\H5d7J[H7f7IXH8i7GVH:k7ESH>m7ARH`0o7_OPHb0Q8^OnGb0S8]OnGc0R8\\OnGd0S8[OnGd0Z4"}, "label": "the chair to the right of the girl in the right hand picture"}]}
{"id": 214924, "image": "COCO_train2014_000000214924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a marble tabletop under a bottle of beer\"."}], "task": "refering", "answer_template": "The \"a marble tabletop under a bottle of beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0005, 0.5825208333333334, 0.41342187499999994, 0.994625], "iscrowd": 0, "area": 34617.4951, "rle": {"size": [480, 640], "counts": "h;8h>]2bMa0_O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100L4K5J6K5K5N2000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O10000000000000000000000000000000000000dKVGb2j8]MWGc2i8\\MXGd2h8ZMZGf2f8YM[Gg2e8XM\\Gh2e8UM]Gk2c8TM^Gl2b8SM_Gm2a8QMaGo2_8PMbGP3^8oLcGQ3]8mLeGS3[8lLfGT3Z8kLgGU3Y8iLiGW3W8hLjGX3V8gLlGX3T8fLnGZ3R8eLoG[3R8cLoG]3Q8aLQH_3o7`LRH`3n7_LSHa3m7]LUHc3k7\\LVHd3j7[LWHe3i7YLYHg3g7XLZHh3f7WL[Hi3g7SL[Hm3h7nKZHR4h7hK\\HX4V9gM`E2`:FjE8W:@QF?o9YO[Fe0e9TOdFj0\\9TOhFj0X9UOjFj0V9UOmFi0S9VOPGh0P9XORGf0n8YOTGf0l8YOWGe0i8ZOZGd0f8[O]Gc0c8\\O_Gc0a8\\ObGb0^8]OeGa0[8^OhG`0X8_OjG`0V8_OmG?S8@PH>Q8APH>P8ARH>n7BSH=m7BUH=k7CWH;i7DYH;g7EZH:f7E\\H:d7E^H:b7F_H9a7FaH9_7GcH7]7HeH7[7IfH6Z7IhH6X7JiH5W7JjH5W7JkH5V7IkH7U7HmH7S7HnH8R7GPI8P7GQI9l:O001O001O001O001O001O001O001O001O00UX`5"}, "label": "a marble tabletop under a bottle of beer"}]}
{"id": 214924, "image": "COCO_train2014_000000214924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left side of a shiny white marbled table with a beer sitting on it\"."}], "task": "refering", "answer_template": "The \"left side of a shiny white marbled table with a beer sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0005, 0.5825208333333334, 0.41342187499999994, 0.994625], "iscrowd": 0, "area": 34617.4951, "rle": {"size": [480, 640], "counts": "h;8h>]2bMa0_O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100L4K5J6K5K5N2000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O10000000000000000000000000000000000000dKVGb2j8]MWGc2i8\\MXGd2h8ZMZGf2f8YM[Gg2e8XM\\Gh2e8UM]Gk2c8TM^Gl2b8SM_Gm2a8QMaGo2_8PMbGP3^8oLcGQ3]8mLeGS3[8lLfGT3Z8kLgGU3Y8iLiGW3W8hLjGX3V8gLlGX3T8fLnGZ3R8eLoG[3R8cLoG]3Q8aLQH_3o7`LRH`3n7_LSHa3m7]LUHc3k7\\LVHd3j7[LWHe3i7YLYHg3g7XLZHh3f7WL[Hi3g7SL[Hm3h7nKZHR4h7hK\\HX4V9gM`E2`:FjE8W:@QF?o9YO[Fe0e9TOdFj0\\9TOhFj0X9UOjFj0V9UOmFi0S9VOPGh0P9XORGf0n8YOTGf0l8YOWGe0i8ZOZGd0f8[O]Gc0c8\\O_Gc0a8\\ObGb0^8]OeGa0[8^OhG`0X8_OjG`0V8_OmG?S8@PH>Q8APH>P8ARH>n7BSH=m7BUH=k7CWH;i7DYH;g7EZH:f7E\\H:d7E^H:b7F_H9a7FaH9_7GcH7]7HeH7[7IfH6Z7IhH6X7JiH5W7JjH5W7JkH5V7IkH7U7HmH7S7HnH8R7GPI8P7GQI9l:O001O001O001O001O001O001O001O001O00UX`5"}, "label": "left side of a shiny white marbled table with a beer sitting on it"}]}
{"id": 214924, "image": "COCO_train2014_000000214924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a potted tree\"."}], "task": "refering", "answer_template": "The \"a potted tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 120, 121, 122, 143, 144, 145, 166, 167, 189, 190, 212, 213], "bbox": [0.229734375, 0.0045000000000000005, 0.481421875, 0.5878333333333334], "iscrowd": 0, "area": 15750.373850000002, "rle": {"size": [480, 640], "counts": "ZmT2:f>d0\\Od0\\Od0\\Ob0^O0000000TDZOc0aNY7U2THB:nNn6`1hHB:VOf6X1PIB9YOe6U1RIB9\\Ob6R1UIC7_Oa6n0XIC7B^6k0[IC6F\\6g0^IC6IY6d0`ID6KX6a0bID5OV6=eID52S6:hIE35R66kIE34S67jIE25T66jIE24U67iIE14W67hIE14W67hIFO4Y66hIFO4Y66hIFN5Z65hIFM6[64hIFM6[64gIGM6\\63gIHL5]63gIHK6^62gIHK6^62gIHJ7`60fIIJ7`60fIX1IZMa6^1fIU1L]M^6^1fIQ10aMZ6^1fIm05dMU6_1fIj08gMR6_1fIf0<kMn5_1fIb0`0oMj5_1fI?d0QNf5`1eI<d8D\\G9g8GYG5k8KUG1o8XOhGe0[8aN_H[1e7lMTIQ2V:Ec0]Ob0_Oa0^O_]6EPcIf0ZOe0[Of0ZOf0ZO2N00000000000000000000000000000000000000000000000000000000000000000000000000000001N10000000cf4SNhZKc0NO2N2O1N1O2N2N2N1O2N2N2N1O2N2N101N2N2N1O2N2N2N1O2N2N2N1OPdk4"}, "label": "a potted tree"}]}
{"id": 214924, "image": "COCO_train2014_000000214924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in a yellow shirt is picking up a slice of pizza\"."}], "task": "refering", "answer_template": "The \"a young man in a yellow shirt is picking up a slice of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261], "bbox": [0.286328125, 0.018520833333333334, 0.8130468750000001, 0.6845416666666666], "iscrowd": 0, "area": 54396.414750000025, "rle": {"size": [480, 640], "counts": "nPf22m>5J6J7J5J6J4L4M2M4M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M2N2N2O0O2O0O2O0O2O0O2O0OQOSDmNm;R1WDlNg;U1\\DhNd;W1`DgN^;Z1eDcNZ;^1iD`NV;_1lDaNR;`1oD_NQ;`1QE`Nm:a1TE^Nl:a1VE_Nh:b1YE^Nf:a1\\E^Nc:c1^E]Na:c1`E\\N_:d1cE\\N\\:d1eE\\NY:d1iE[NW:e1jE[NT:f1mEYNS:f1oEZNo9g1RFYNm9g1TFXNk9i1VFWNi9i1YFUNf9l1[FSNe9m1\\FQNd9P2^FmMc9S2^FkMb9V2d10O1O100O1O1O100O1O010O0001G8O101O0000000000000001O0001O0000QDQNn:o1QEWNj:i1UE\\Ng:d1XE`Ne:`1[EbNc:^1\\EdNc:[1]EgNb:Y1]EiNb:W1]EkNb:U1]EmNc:R1\\EPOc:P1\\EROb:o0]ESOb:m0]EVOa:j0^EXO`:i0_EYO`:g0_E[O`:e0^E^O`:c0_E_O`:a0_EA`:?^ED`:=_EE`:;_EH_:8`EJ^:7`EL_:4`EN_:1aE1^:O`E4_:W2O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O100O1O1O1O2O00000000000000001O00000O1000000oMoJSJR5j5`KfI`4X6QLYIo3d6dLlH\\3R7Z2N2M3N2N2N2M4M2O100O1O100O100O1O101N100O1O100O100O100O2N100O100O100000000O101O000O1000000O1000000O2O000000000000000000001O00000001O001O000010O01O00001O00001O000010O000001O001O1O001O1O100O1O1bJgIS2Z6iMjIV2W6gMlIX2U6cMPJ\\2Q6^MTJb2m5XMYJg2h5SM^Jl2c5PMaJo2`5nLbJS3^5iLfJV3[5gLhJX3Y5dLkJ[3V5bLlJ^3T5_LPK`3Q5[LTKd3m4RL\\Kn3e4iKdKV4]4`KmK_4T4XKULh4R7N2O1O1O0O2O1O1O1N2O1O0O2O1O2N2N2N2N2N2N2N2N2N3M2N2N2N2N3M2N2N2N3M3N2M3M3M3M3M3M2N3M3M3M6J7G8I8G9H7H9H8GWog1"}, "label": "a young man in a yellow shirt is picking up a slice of pizza"}]}
{"id": 214924, "image": "COCO_train2014_000000214924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a yellow shirt getting a piece of pizza\"."}], "task": "refering", "answer_template": "The \"a man in a yellow shirt getting a piece of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261], "bbox": [0.286328125, 0.018520833333333334, 0.8130468750000001, 0.6845416666666666], "iscrowd": 0, "area": 54396.414750000025, "rle": {"size": [480, 640], "counts": "nPf22m>5J6J7J5J6J4L4M2M4M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M2N2N2O0O2O0O2O0O2O0O2O0OQOSDmNm;R1WDlNg;U1\\DhNd;W1`DgN^;Z1eDcNZ;^1iD`NV;_1lDaNR;`1oD_NQ;`1QE`Nm:a1TE^Nl:a1VE_Nh:b1YE^Nf:a1\\E^Nc:c1^E]Na:c1`E\\N_:d1cE\\N\\:d1eE\\NY:d1iE[NW:e1jE[NT:f1mEYNS:f1oEZNo9g1RFYNm9g1TFXNk9i1VFWNi9i1YFUNf9l1[FSNe9m1\\FQNd9P2^FmMc9S2^FkMb9V2d10O1O100O1O1O100O1O010O0001G8O101O0000000000000001O0001O0000QDQNn:o1QEWNj:i1UE\\Ng:d1XE`Ne:`1[EbNc:^1\\EdNc:[1]EgNb:Y1]EiNb:W1]EkNb:U1]EmNc:R1\\EPOc:P1\\EROb:o0]ESOb:m0]EVOa:j0^EXO`:i0_EYO`:g0_E[O`:e0^E^O`:c0_E_O`:a0_EA`:?^ED`:=_EE`:;_EH_:8`EJ^:7`EL_:4`EN_:1aE1^:O`E4_:W2O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O100O1O1O1O2O00000000000000001O00000O1000000oMoJSJR5j5`KfI`4X6QLYIo3d6dLlH\\3R7Z2N2M3N2N2N2M4M2O100O1O100O100O1O101N100O1O100O100O100O2N100O100O100000000O101O000O1000000O1000000O2O000000000000000000001O00000001O001O000010O01O00001O00001O000010O000001O001O1O001O1O100O1O1bJgIS2Z6iMjIV2W6gMlIX2U6cMPJ\\2Q6^MTJb2m5XMYJg2h5SM^Jl2c5PMaJo2`5nLbJS3^5iLfJV3[5gLhJX3Y5dLkJ[3V5bLlJ^3T5_LPK`3Q5[LTKd3m4RL\\Kn3e4iKdKV4]4`KmK_4T4XKULh4R7N2O1O1O0O2O1O1O1N2O1O0O2O1O2N2N2N2N2N2N2N2N2N3M2N2N2N2N3M2N2N2N3M3N2M3M3M3M3M3M2N3M3M3M6J7G8I8G9H7H9H8GWog1"}, "label": "a man in a yellow shirt getting a piece of pizza"}]}
{"id": 264076, "image": "COCO_train2014_000000264076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a purple shirt and shorts riding a skateboard down some stairs\"."}], "task": "refering", "answer_template": "The \"a guy in a purple shirt and shorts riding a skateboard down some stairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 189, 190, 191, 212, 213, 214], "bbox": [0.191546875, 0.24695, 0.396765625, 0.696875], "iscrowd": 0, "area": 7523.0027, "rle": {"size": [400, 640], "counts": "oU`14[<1O2N2N1O10N200N200001O1O0O1O100O100000O10O1O0O2O002O0O100O100O1jG@\\4`0`KD_4=]KGc49\\KHc49bJc0\\5^O_Jg0a5YO\\Jk0b5VO\\Jl0d5SO[IA@_1U7POkH=Be0c7ROfH\\2Y7lM`HT2`7QNVH^15jMe7\\3_HcLa7Z3fHYLC3g7b3PI\\Lo6c3c0bLhGY3Z8mLbGQ3`8`0O1O001O01@dGkL]8R3iGjLX8U3lGgLU8W3c0O1O1011fHmLS5`3_JaLb5_3[JcLf5^3mHaLm06W6b3SISMo6k2nHYMS7e2iH_M[7jNeHi4b7O00000B_HnKa7m3cHhKK8b7m3fHjKH:b7j3jHVLW7f3lHQLA7b7d3QIUL^O0O3b7d3UIYL]O3_7Z3]IbLUO5]7Y3eIgL[6Y3fIgLZ6V3hIiLY6a1]HbN0o0`1mNS6\\1oH@k8;]G_Oe8NSHLo73UHHm78_HZOc7e0T20100O100O101O0O1O1O10000000001O1O000O100000001O00001O0014K01O001KmCJS<5oCI20j;8TDLn;92N3M3Mbcf4"}, "label": "a guy in a purple shirt and shorts riding a skateboard down some stairs"}]}
{"id": 264076, "image": "COCO_train2014_000000264076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skateboarding\"."}], "task": "refering", "answer_template": "The \"a man skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 189, 190, 191, 212, 213, 214], "bbox": [0.191546875, 0.24695, 0.396765625, 0.696875], "iscrowd": 0, "area": 7523.0027, "rle": {"size": [400, 640], "counts": "oU`14[<1O2N2N1O10N200N200001O1O0O1O100O100000O10O1O0O2O002O0O100O100O1jG@\\4`0`KD_4=]KGc49\\KHc49bJc0\\5^O_Jg0a5YO\\Jk0b5VO\\Jl0d5SO[IA@_1U7POkH=Be0c7ROfH\\2Y7lM`HT2`7QNVH^15jMe7\\3_HcLa7Z3fHYLC3g7b3PI\\Lo6c3c0bLhGY3Z8mLbGQ3`8`0O1O001O01@dGkL]8R3iGjLX8U3lGgLU8W3c0O1O1011fHmLS5`3_JaLb5_3[JcLf5^3mHaLm06W6b3SISMo6k2nHYMS7e2iH_M[7jNeHi4b7O00000B_HnKa7m3cHhKK8b7m3fHjKH:b7j3jHVLW7f3lHQLA7b7d3QIUL^O0O3b7d3UIYL]O3_7Z3]IbLUO5]7Y3eIgL[6Y3fIgLZ6V3hIiLY6a1]HbN0o0`1mNS6\\1oH@k8;]G_Oe8NSHLo73UHHm78_HZOc7e0T20100O100O101O0O1O1O10000000001O1O000O100000001O00001O0014K01O001KmCJS<5oCI20j;8TDLn;92N3M3Mbcf4"}, "label": "a man skateboarding"}]}
{"id": 83866, "image": "COCO_train2014_000000083866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman ' s hair and shirt\"."}], "task": "refering", "answer_template": "The \"a woman ' s hair and shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.303359375, 1.0, 1.0], "iscrowd": 0, "area": 134322.34600000005, "rle": {"size": [640, 428], "counts": "f`0;fc0f0YOf0[Of0ZOc0\\O10O10O010O01O1O001O3M00000000000000000000000000001O00000000000000001O000000000000001O000000000000010O00000000000000001O000000O1O1O1O1O1O1O1O100O1O1O1O1O100O100O10000[JnL[JS3b5YMUJg2h5eMoI[2n5QNhIP2U6\\NbId1[6hN\\IX1b6SOTIn0i6^OnHb0o6JhH6U76aHK\\7a0[H_Ob7m0UHSOi7S1RHnNk7U1UHkNh7X1XHhNe7[1ZHfNc7\\1^HdN_7_1`HbN^7`1bH`N[7b1fH^NW7e1mGWOP8l0oGUOn7n0QHSOl7o0THROj7P1UHQOh7R1WHoNh7Q1XHPOf7R1YHoNe7S1ZHnNe7S1ZHnNd7V1ZHjNd7^1UHcNi7e1PH\\No7k1jGVNT8R2eGoMY8m6O1N2N2N2O1N2N2N2O1N2N2O1N200O1O1O100O1O2N100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O10000O10000O10000O100O10000O10000O100O10000000000O1000000000000000000O10000000000000000O1000001O000000000000000O1000000000000O100O1O100O1001O00000000001O00000000001O00000000001O00000000001O00000000001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O00100O1O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1N2O1[JaGG`85cGK^81eGO\\8MgG3Z8HkG7V8EmG;T8AoG?R8[OSHe0n7UOWHk0j7PO[Ho0g7jN^HV1d7dN`H\\1a7^NdHb1^7WNhHg1[7SNiHm1Y7lMlHT2V7fMnHZ2T7`MPI`2R7ZMSIe2o6UMUIk2m6oLWIQ3k6iLYIW3h6dL\\I\\3f6^L_Ia3c6ZL`If3b6TLbIl3`6nKdIR4^6hKgIW4[6cKiI]4Y6]KkIb4X6XKlIh4U6SKoIm4S6mJRJR5^:00000001O0000001O00000O101O000000001O0O1OXA"}, "label": "a woman ' s hair and shirt"}]}
{"id": 83866, "image": "COCO_train2014_000000083866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a oerson ' s head\"."}], "task": "refering", "answer_template": "The \"the back of a oerson ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.303359375, 1.0, 1.0], "iscrowd": 0, "area": 134322.34600000005, "rle": {"size": [640, 428], "counts": "f`0;fc0f0YOf0[Of0ZOc0\\O10O10O010O01O1O001O3M00000000000000000000000000001O00000000000000001O000000000000001O000000000000010O00000000000000001O000000O1O1O1O1O1O1O1O100O1O1O1O1O100O100O10000[JnL[JS3b5YMUJg2h5eMoI[2n5QNhIP2U6\\NbId1[6hN\\IX1b6SOTIn0i6^OnHb0o6JhH6U76aHK\\7a0[H_Ob7m0UHSOi7S1RHnNk7U1UHkNh7X1XHhNe7[1ZHfNc7\\1^HdN_7_1`HbN^7`1bH`N[7b1fH^NW7e1mGWOP8l0oGUOn7n0QHSOl7o0THROj7P1UHQOh7R1WHoNh7Q1XHPOf7R1YHoNe7S1ZHnNe7S1ZHnNd7V1ZHjNd7^1UHcNi7e1PH\\No7k1jGVNT8R2eGoMY8m6O1N2N2N2O1N2N2N2O1N2N2O1N200O1O1O100O1O2N100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O10000O10000O10000O100O10000O10000O100O10000000000O1000000000000000000O10000000000000000O1000001O000000000000000O1000000000000O100O1O100O1001O00000000001O00000000001O00000000001O00000000001O00000000001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O00100O1O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1N2O1[JaGG`85cGK^81eGO\\8MgG3Z8HkG7V8EmG;T8AoG?R8[OSHe0n7UOWHk0j7PO[Ho0g7jN^HV1d7dN`H\\1a7^NdHb1^7WNhHg1[7SNiHm1Y7lMlHT2V7fMnHZ2T7`MPI`2R7ZMSIe2o6UMUIk2m6oLWIQ3k6iLYIW3h6dL\\I\\3f6^L_Ia3c6ZL`If3b6TLbIl3`6nKdIR4^6hKgIW4[6cKiI]4Y6]KkIb4X6XKlIh4U6SKoIm4S6mJRJR5^:00000001O0000001O00000O101O000000001O0O1OXA"}, "label": "the back of a oerson ' s head"}]}
{"id": 83866, "image": "COCO_train2014_000000083866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of a mans head in a white hat\"."}], "task": "refering", "answer_template": "The \"the top of a mans head in a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 180, 181, 195, 196, 210], "bbox": [0.0, 0.3051875, 0.17175233644859814, 0.62725], "iscrowd": 0, "area": 10192.358950000003, "rle": {"size": [640, 428], "counts": "T6`4`?m1TNN1O2N1L5L4M2010O10O0100O01000O010O10O0N3M3N1N3M3N2M3M3N2M3N2M2N3N2M3M3N2N2M3N2M2O2N2M3N2N2M3N2M3N1O2M3N2N2L4M3L4M2M4M3L6K6J6I7J6I7H8H8I7H8H8HWcm6"}, "label": "the top of a mans head in a white hat"}]}
{"id": 83866, "image": "COCO_train2014_000000083866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a white cap on someone ' s head\"."}], "task": "refering", "answer_template": "The \"the back of a white cap on someone ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 180, 181, 195, 196, 210], "bbox": [0.0, 0.3051875, 0.17175233644859814, 0.62725], "iscrowd": 0, "area": 10192.358950000003, "rle": {"size": [640, 428], "counts": "T6`4`?m1TNN1O2N1L5L4M2010O10O0100O01000O010O10O0N3M3N1N3M3N2M3M3N2M3N2M2N3N2M3M3N2N2M3N2M2O2N2M3N2N2M3N2M3N1O2M3N2N2L4M3L4M2M4M3L6K6J6I7J6I7H8H8I7H8H8HWcm6"}, "label": "the back of a white cap on someone ' s head"}]}
{"id": 395169, "image": "COCO_train2014_000000395169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a notebook and stack of papers on the table\"."}], "task": "refering", "answer_template": "The \"a notebook and stack of papers on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 53, 68, 69, 70, 71, 87, 88, 89], "bbox": [0.46654, 0.0039039039039039042, 1.0, 0.39453453453453463], "iscrowd": 0, "area": 20235.050199999998, "rle": {"size": [333, 500], "counts": "_i[21[:1N3N1O2M3N1O2M3N1O2N2M2O2N2M2O2N2M2O2O1O001O1O001O1VGgN^8Y1\\GmNd8_1O00001O000010O000001O00001O00001O0000001O00001O00001O00000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000O1000000O100000000O1000000O100000000O1000000O10000001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O2N2N1O2N2N1O2N1O001O00001O0000001O01O01O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O0000E;A?@`0@`0A?@`0@`0A<"}, "label": "a notebook and stack of papers on the table"}]}
{"id": 395169, "image": "COCO_train2014_000000395169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stack of paper and notebooks adjacent to a apple\"."}], "task": "refering", "answer_template": "The \"a stack of paper and notebooks adjacent to a apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 53, 68, 69, 70, 71, 87, 88, 89], "bbox": [0.46654, 0.0039039039039039042, 1.0, 0.39453453453453463], "iscrowd": 0, "area": 20235.050199999998, "rle": {"size": [333, 500], "counts": "_i[21[:1N3N1O2M3N1O2M3N1O2N2M2O2N2M2O2N2M2O2O1O001O1O001O1VGgN^8Y1\\GmNd8_1O00001O000010O000001O00001O00001O0000001O00001O00001O00000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000O1000000O100000000O1000000O100000000O1000000O10000001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O2N2N1O2N2N1O2N1O001O00001O0000001O01O01O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O0000E;A?@`0@`0A?@`0@`0A<"}, "label": "a stack of paper and notebooks adjacent to a apple"}]}
{"id": 231337, "image": "COCO_train2014_000000231337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of blue and white m & ms\"."}], "task": "refering", "answer_template": "The \"a bowl of blue and white m & ms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237], "bbox": [0.4212121212121212, 0.54625, 0.8515850815850815, 0.690484375], "iscrowd": 0, "area": 11483.452199999998, "rle": {"size": [640, 429], "counts": "b_a39fc03M3d\\OCPc0g001O001O1O001O010O1O001O001O10OO2N1O2N2O0O2N101O001O001O1N101O001O001O1O0O2O001O001N101N2O001N3N1N3N1O2M2O1N3N1O2M2O2N1O2N00O1000000O2O00000O1000000O101O000O11O001O1O001O1O0O2O1OO010O1O001O1O001O1O010O10O0100O100O100O1000001O000000000000001O000000000000001O00000O10001O00000000001O0O1000001O000000001O000O10001O000001O01O01O00N2O2M2N2O1ZNU^OU1ma0gN`^Oo0Zb0M3N2M4L3N2M3N2M3N2M3NecW1"}, "label": "a bowl of blue and white m & ms"}]}
{"id": 231337, "image": "COCO_train2014_000000231337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dish of multiple shades of blue m & m ' s\"."}], "task": "refering", "answer_template": "The \"dish of multiple shades of blue m & m ' s\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237], "bbox": [0.4212121212121212, 0.54625, 0.8515850815850815, 0.690484375], "iscrowd": 0, "area": 11483.452199999998, "rle": {"size": [640, 429], "counts": "b_a39fc03M3d\\OCPc0g001O001O1O001O010O1O001O001O10OO2N1O2N2O0O2N101O001O001O1N101O001O001O1O0O2O001O001N101N2O001N3N1N3N1O2M2O1N3N1O2M2O2N1O2N00O1000000O2O00000O1000000O101O000O11O001O1O001O1O0O2O1OO010O1O001O1O001O1O010O10O0100O100O100O1000001O000000000000001O000000000000001O00000O10001O00000000001O0O1000001O000000001O000O10001O000001O01O01O00N2O2M2N2O1ZNU^OU1ma0gN`^Oo0Zb0M3N2M4L3N2M3N2M3N2M3NecW1"}, "label": "dish of multiple shades of blue m & m ' s"}]}
{"id": 231337, "image": "COCO_train2014_000000231337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green dish displaying a white type candy\"."}], "task": "refering", "answer_template": "The \"a green dish displaying a white type candy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 236, 237, 238, 239, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 294, 295, 296, 297], "bbox": [0.48785547785547784, 0.662859375, 0.9966666666666667, 0.8456718750000001], "iscrowd": 0, "area": 17586.145450000007, "rle": {"size": [640, 429], "counts": "]RS43hc08M3M6J4L4L3M4L3M001O000010OJ601N1O101N1O101N100O2N10001O00001O00001K400O2O000O2O00000O1000000O10004L9G8G3N001O1O010O010O10O010O01O10OO2N1O1O1O1O1O1O1O101N1O1O1O1O1O10000000001O000000000000000000000O2O000000000000000000001O000000000000000000O10000O10000O101O0O1000000O101O0O10000O101O0O1000000O2O000O1000001N10000O10001O0O10000O101M2O1O1N2O1N3N1\\Oh^O^NZa0b1o^OTNRa0k1b001O0O100O2O0O101N101N100O2O0O2O0O2O000O2O0O2O0O2O0O101N101N1N3N1O2M4M3M3L4M3M3L4M3Mci0"}, "label": "a green dish displaying a white type candy"}]}
{"id": 280490, "image": "COCO_train2014_000000280490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"omelette on plate\"."}], "task": "refering", "answer_template": "The \"omelette on plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289], "bbox": [0.18803125, 0.6831529411764705, 0.72525, 0.8269647058823529], "iscrowd": 0, "area": 14409.484100000005, "rle": {"size": [425, 640], "counts": "^ab12V=2O0O100O100O2O0O100O101N10000O100O2O0O100O100O2O0O100O101N100O100O101N10000O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O2O0O100O100O100O10000000000000000000000000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0O1000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000O1000000000G9000001O0000000000010O0000000000001O0000000000001O00000000000010O000000000001N10000O1000001N10000O1000001N10000O10000O2O00000O10000O2O00000O10000O2O00000O10000O2O00000O10000O2O0O1O1O1O100O2N1O1O100O1O1O2N100O1O1O1O101N1O1O1O100O1OogX2"}, "label": "omelette on plate"}]}
{"id": 280490, "image": "COCO_train2014_000000280490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"food on a white plate with red fruit in the background\"."}], "task": "refering", "answer_template": "The \"food on a white plate with red fruit in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289], "bbox": [0.18803125, 0.6831529411764705, 0.72525, 0.8269647058823529], "iscrowd": 0, "area": 14409.484100000005, "rle": {"size": [425, 640], "counts": "^ab12V=2O0O100O100O2O0O100O101N10000O100O2O0O100O100O2O0O100O101N100O100O101N10000O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O2O0O100O100O100O10000000000000000000000000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0O1000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000O1000000000G9000001O0000000000010O0000000000001O0000000000001O00000000000010O000000000001N10000O1000001N10000O1000001N10000O10000O2O00000O10000O2O00000O10000O2O00000O10000O2O00000O10000O2O0O1O1O1O100O2N1O1O100O1O1O2N100O1O1O1O101N1O1O1O100O1OogX2"}, "label": "food on a white plate with red fruit in the background"}]}
{"id": 460715, "image": "COCO_train2014_000000460715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green chair behind the man\"."}], "task": "refering", "answer_template": "The \"green chair behind the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 233, 234, 235, 256, 257, 279, 280, 281, 302, 303, 304, 307, 325, 326, 327, 330, 331, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.144296875, 0.5255, 0.431484375, 1.0], "iscrowd": 0, "area": 15523.272399999998, "rle": {"size": [480, 640], "counts": "S`[11k>9XN4nC3];b0\\DCW;l0bDYOQ;V1hDoNT;W1eDnNW;_2Mj0WOk0TOk0UO:G2M300O10000O100O100O101O0O100O10000O100O10eJXH^3h7`L[H_3d7aL]H_3c7_L`H`3_7`LbH`3^7_LdH`3^7]LdHb3d7TLaHi3]9N3M2N1O2N1O1O1O2N1O1O1O2N001O1O1O1O001O1PNSEKn:F`E:a:VOnEj0V:cNYF]1\\7dNnJ\\1k4nNSKQ1d4\\OXKd0_4H^K8a4FbK:\\4FfK:Y4DjK<R4ERL:n3DTL<m3@VL`0j3]OYLc0g3YO]Lg0c3VO`Lj0`3ROeLm0c3bNfL^1[3]NiLc1Z8XOnBBR=:YC_Og<=dC\\O]<?T1L4L4L4MgQ?NVn@8G9G9J6L4N2M3N2N2M3N3M2M3N2L4M3M3M3M3M3M3M3M3L4M3M3M3M3M31O5K6PM]DW2W<O001O001O001O001O001O00001O001O3M3_N_C=e=M2N2N2N2N3M2N2N2N2N2N2NRUZ5"}, "label": "green chair behind the man"}]}
{"id": 460715, "image": "COCO_train2014_000000460715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black single seater sofa behind standing man\"."}], "task": "refering", "answer_template": "The \"a black single seater sofa behind standing man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 233, 234, 235, 256, 257, 279, 280, 281, 302, 303, 304, 307, 325, 326, 327, 330, 331, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.144296875, 0.5255, 0.431484375, 1.0], "iscrowd": 0, "area": 15523.272399999998, "rle": {"size": [480, 640], "counts": "S`[11k>9XN4nC3];b0\\DCW;l0bDYOQ;V1hDoNT;W1eDnNW;_2Mj0WOk0TOk0UO:G2M300O10000O100O100O101O0O100O10000O100O10eJXH^3h7`L[H_3d7aL]H_3c7_L`H`3_7`LbH`3^7_LdH`3^7]LdHb3d7TLaHi3]9N3M2N1O2N1O1O1O2N1O1O1O2N001O1O1O1O001O1PNSEKn:F`E:a:VOnEj0V:cNYF]1\\7dNnJ\\1k4nNSKQ1d4\\OXKd0_4H^K8a4FbK:\\4FfK:Y4DjK<R4ERL:n3DTL<m3@VL`0j3]OYLc0g3YO]Lg0c3VO`Lj0`3ROeLm0c3bNfL^1[3]NiLc1Z8XOnBBR=:YC_Og<=dC\\O]<?T1L4L4L4MgQ?NVn@8G9G9J6L4N2M3N2N2M3N3M2M3N2L4M3M3M3M3M3M3M3M3L4M3M3M3M3M31O5K6PM]DW2W<O001O001O001O001O001O00001O001O3M3_N_C=e=M2N2N2N2N3M2N2N2N2N2N2NRUZ5"}, "label": "a black single seater sofa behind standing man"}]}
{"id": 452524, "image": "COCO_train2014_000000452524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa facing the laptop screen\"."}], "task": "refering", "answer_template": "The \"a sofa facing the laptop screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232], "bbox": [0.0871, 0.6408533333333333, 0.90968, 0.9849333333333334], "iscrowd": 0, "area": 31734.165549999998, "rle": {"size": [375, 500], "counts": "T_`01e;1M3L4L4L4L4L4L4M3O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1000000000000000000000000001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O1O001O1O001O1O001O1O001O001O1O1O1O001OO1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O10000O10000O10000O100O10000O10000O10000O10000O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O1O100O100O100O1O100O100O100O01000000000O100000000O1000000O100O100O100O100O100O1O1N2O1O1N2O1N2O1N2O1N2O1O1O10000O100O100O100O10000O100O100O100O10000O1000000000000000000000000000000000000000000000000001O00000000000000001O0000000000000000000000O100000000000000000000000000O1000O2O0000000000001O0000000000001O0000001O001O00001O00001O001O00001O001O00001N1N3M2M4M2N3M2M4oN_1]O7J4K5\\O[W`0"}, "label": "a sofa facing the laptop screen"}]}
{"id": 452524, "image": "COCO_train2014_000000452524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown sofa , brown glass table , with laptop setting in table\"."}], "task": "refering", "answer_template": "The \"a brown sofa , brown glass table , with laptop setting in table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232], "bbox": [0.0871, 0.6408533333333333, 0.90968, 0.9849333333333334], "iscrowd": 0, "area": 31734.165549999998, "rle": {"size": [375, 500], "counts": "T_`01e;1M3L4L4L4L4L4L4M3O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1000000000000000000000000001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O1O001O1O001O1O001O1O001O001O1O1O1O001OO1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O10000O10000O10000O100O10000O10000O10000O10000O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O1O100O100O100O1O100O100O100O01000000000O100000000O1000000O100O100O100O100O100O1O1N2O1O1N2O1N2O1N2O1N2O1O1O10000O100O100O100O10000O100O100O100O10000O1000000000000000000000000000000000000000000000000001O00000000000000001O0000000000000000000000O100000000000000000000000000O1000O2O0000000000001O0000000000001O0000001O001O00001O00001O001O00001O001O00001N1N3M2M4M2N3M2M4oN_1]O7J4K5\\O[W`0"}, "label": "a brown sofa , brown glass table , with laptop setting in table"}]}
{"id": 452524, "image": "COCO_train2014_000000452524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather sofa sitting behind a glass table\"."}], "task": "refering", "answer_template": "The \"a brown leather sofa sitting behind a glass table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 180, 181], "bbox": [0.00646, 0.4970133333333333, 0.32258, 0.8238933333333333], "iscrowd": 0, "area": 12408.103150000003, "rle": {"size": [375, 500], "counts": "oX1j3m71O0000000001O00000000000000000000000000000000000000000000000000000SOm000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O10000O2N1O1O100O1O100O10000000000O1000000000000O10000000000O100000001O2N3L3N3M3M2N3M0O1000000000O01000000000000O10000000000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O1ZOmnk3"}, "label": "a brown leather sofa sitting behind a glass table"}]}
{"id": 452524, "image": "COCO_train2014_000000452524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch that the back of the laptop is facing\"."}], "task": "refering", "answer_template": "The \"a couch that the back of the laptop is facing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 180, 181], "bbox": [0.00646, 0.4970133333333333, 0.32258, 0.8238933333333333], "iscrowd": 0, "area": 12408.103150000003, "rle": {"size": [375, 500], "counts": "oX1j3m71O0000000001O00000000000000000000000000000000000000000000000000000SOm000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O10000O2N1O1O100O1O100O10000000000O1000000000000O10000000000O100000001O2N3L3N3M3M2N3M0O1000000000O01000000000000O10000000000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O1ZOmnk3"}, "label": "a couch that the back of the laptop is facing"}]}
{"id": 239532, "image": "COCO_train2014_000000239532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray pier\"."}], "task": "refering", "answer_template": "The \"a gray pier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 207, 208, 209, 210, 211, 212, 213, 215, 216, 217, 218, 219, 220, 228, 229, 231, 232, 235, 236, 237, 238, 239, 240, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 259, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297], "bbox": [0.013468749999999998, 0.5820608899297424, 1.0, 0.8107728337236534], "iscrowd": 0, "area": 27992.20735000001, "rle": {"size": [427, 640], "counts": "RP43X=3M2N3M3M3M3M3M3M2N001O001O00010O001O00001O001O00001O001O00001O001O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O001O0000000000000000O1N2M3N2N2N2N2N3L3N2N2N2N2N2N2M3N2N2N2N2N2M3O11O00000000000000001O00000000000000001O00000000000000001O00000000000000001O0001O000000000001O00000000000000001O00000000000000001O000000004L5K5K6J5K6J5K5L50O1J6J6J5K0000001O0001O000001O000000001O000000010O00000000001OO1000000000001O0000000O10000000000000001O000O100000000000000000001N100000000000000O1N2N2N2M3N3M1O1O1O001N2O1O11N2O1O1O1O1N2O1O1O1O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1O1NKmNVDR1k;POSDo0n;ROQDm0P<60O2N2I7J6I7J6I6N3000000000O1000000000000O1000000000000O100000001O000100O00100O1O1O010O1O10O2O5K6J5K6I6J4L2N2O0O2N2N1O2N2N101N2N1O2N2N1O2O1N1O2N2N1O2N2O1N1O2N2N1OSNTE]1j:aNXE_1i:^NYEb1g:]NZEc1f:[N\\Ee1d:YN^Eg1b:WN`Ei1`:VNbEi1^:UNdEk1n:000000000000000000000001O00000000000000000000000000001O000000000000000000000000001N100000000000000000000000001O000000000000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O000000000000000000000000000000000000000O100000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000001O0000001O0000001O0000001O0000001O0000001O0O10001O0000001O0000001O0000001O0000001O00N2I7I7I7J6L40000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O100000000000000VOj0RO^H"}, "label": "a gray pier"}]}
{"id": 239532, "image": "COCO_train2014_000000239532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dock they are lounging on\"."}], "task": "refering", "answer_template": "The \"the dock they are lounging on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 207, 208, 209, 210, 211, 212, 213, 215, 216, 217, 218, 219, 220, 228, 229, 231, 232, 235, 236, 237, 238, 239, 240, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 259, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297], "bbox": [0.013468749999999998, 0.5820608899297424, 1.0, 0.8107728337236534], "iscrowd": 0, "area": 27992.20735000001, "rle": {"size": [427, 640], "counts": "RP43X=3M2N3M3M3M3M3M3M2N001O001O00010O001O00001O001O00001O001O00001O001O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O001O0000000000000000O1N2M3N2N2N2N2N3L3N2N2N2N2N2N2M3N2N2N2N2N2M3O11O00000000000000001O00000000000000001O00000000000000001O00000000000000001O0001O000000000001O00000000000000001O00000000000000001O000000004L5K5K6J5K6J5K5L50O1J6J6J5K0000001O0001O000001O000000001O000000010O00000000001OO1000000000001O0000000O10000000000000001O000O100000000000000000001N100000000000000O1N2N2N2M3N3M1O1O1O001N2O1O11N2O1O1O1O1N2O1O1O1O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1O1NKmNVDR1k;POSDo0n;ROQDm0P<60O2N2I7J6I7J6I6N3000000000O1000000000000O1000000000000O100000001O000100O00100O1O1O010O1O10O2O5K6J5K6I6J4L2N2O0O2N2N1O2N2N101N2N1O2N2N1O2O1N1O2N2N1O2N2O1N1O2N2N1OSNTE]1j:aNXE_1i:^NYEb1g:]NZEc1f:[N\\Ee1d:YN^Eg1b:WN`Ei1`:VNbEi1^:UNdEk1n:000000000000000000000001O00000000000000000000000000001O000000000000000000000000001N100000000000000000000000001O000000000000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O000000000000000000000000000000000000000O100000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000001O0000001O0000001O0000001O0000001O0000001O0O10001O0000001O0000001O0000001O0000001O00N2I7I7I7J6L40000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O100000000000000VOj0RO^H"}, "label": "the dock they are lounging on"}]}
{"id": 534443, "image": "COCO_train2014_000000534443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat at the end of a dock\"."}], "task": "refering", "answer_template": "The \"a boat at the end of a dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 216, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.40931249999999997, 0.5574473067915691, 0.8740937499999999, 0.7304918032786886], "iscrowd": 0, "area": 9987.616199999999, "rle": {"size": [427, 640], "counts": "Tn]31[=0O2O0O2N2O0O2O0O2N2N1O1O10O01O1O1O1O1O1O1O1O1O1O01N10000O100001O00001O00010O00001O00001O000000000001O00000000000000000000O10000000000000000O100000000000000O10000000000000000O100000O1000000000O1000000000O100000O100000000000O1000O100000000000000O01000000000000000O1000O100000000000O100000O01000000O1000000O1000O10O10000O1000000O1N1O2M3K5K5L4M3M3M3M3000O1000000000001O00000O1000000000001O0000000O10001O00001O0000001N1000001O0000001O00000O2O00001O0000001O2N3M2M2O1O2N1O1O000000000O100000000000000000000O1000000000000000001N100000001O000000001O0O10001O000000001O000O10001O000000000N2L4M3M3M3M3Ld`Q1"}, "label": "a boat at the end of a dock"}]}
{"id": 534443, "image": "COCO_train2014_000000534443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couple of boats tied to the end of a pier\"."}], "task": "refering", "answer_template": "The \"a couple of boats tied to the end of a pier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 216, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.40931249999999997, 0.5574473067915691, 0.8740937499999999, 0.7304918032786886], "iscrowd": 0, "area": 9987.616199999999, "rle": {"size": [427, 640], "counts": "Tn]31[=0O2O0O2N2O0O2O0O2N2N1O1O10O01O1O1O1O1O1O1O1O1O1O01N10000O100001O00001O00010O00001O00001O000000000001O00000000000000000000O10000000000000000O100000000000000O10000000000000000O100000O1000000000O1000000000O100000O100000000000O1000O100000000000000O01000000000000000O1000O100000000000O100000O01000000O1000000O1000O10O10000O1000000O1N1O2M3K5K5L4M3M3M3M3000O1000000000001O00000O1000000000001O0000000O10001O00001O0000001N1000001O0000001O00000O2O00001O0000001O2N3M2M2O1O2N1O1O000000000O100000000000000000000O1000000000000000001N100000001O000000001O0O10001O000000001O000O10001O000000000N2L4M3M3M3M3Ld`Q1"}, "label": "a couple of boats tied to the end of a pier"}]}
{"id": 1966, "image": "COCO_train2014_000000001966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing nike shoes in a bright neon green top is playing tennis\"."}], "task": "refering", "answer_template": "The \"a man wearing nike shoes in a bright neon green top is playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 78, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 194, 195, 196, 212, 213, 214, 215, 217, 218, 235, 236, 237, 238, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 326, 327, 328, 329, 330, 331, 350, 351, 352, 353], "bbox": [0.197640625, 0.14752083333333335, 0.5388437500000001, 0.9177916666666667], "iscrowd": 0, "area": 27770.737749999986, "rle": {"size": [480, 640], "counts": "kmk12m>2N1N3N1O2N1N3N2N1O2N1N3N1O2L4L3M4L3M4K4M4L4M2N3M2M4M3M2N3L3N3M2_MRNmGR2g7YNWHj1]7aNaHa1T7jNjHY1j6ROSIR1a6YO]Ii0Y6AeIb0X6@fIc0W6_OfIZOROmNV7k1fISO]OQOj6n1gIoNEROa6P2iInNJPO[6T2hIlN1POU6U2hIkN7nNo5Y2gIjN>kNi5]2fIiNe0gNd5b2cIhNn0dN]5f2bIgNU1`NX5k2`IfN[1^NS5[5QKbJn4]5VKaJj4^5YKaJe4_5^K_Jb4`5aK_J^4`5eK^J[4`5hK`JW4_5lK_JT4`5oK_JP4`5SL^Jm3`5j2O1000O0100O10000O100O100O01000O1o0QO6J6J6TIiId4\\6WKdIi4b6PK`Io4f6jJ\\IU5j6dJXIZ5o6_JSI`5S7YJoHf5V8O1O1O1O1O1O1O001N2OhIaJQ3^5nLfJP3Y5PMiJo2W5PMlJn2S5QMPKb0J_NV5n0SK1>jN^4U1VKAR1SOh3[1YK]OY1QO]3b1\\KYOa1nNS3h1PLcNX1]Oh2P2]LTNn0Le2o1jLfMd0;a2o1WMXM:j0_2m1;SNDn1;SNEk1=UNBk1a0SN_Ol1c0SN\\Ol1g0SNYOl1j0RNUOm1n0RNROm1P1RNoNm1U1QNkNn1W1PNjNn1Y1QNfNo1]1oMcNP2_1oMaNX2Y1gMgNa2R1^MmNd2S1[MmNe2S1[MmNf2R1ZMnNf2R1ZMnNf2R1ZMmNh2S1WMmNi2S1WMmNj2R1UMoNk2Q1UMnNm2Q1RMPOn2Q1PMPOP3P1oLQOR3n0mLSOS3m0lLSOV3m0gLUO[3i0dLXO^3f0aL[Oa3c0^L]Of3k54L4L4L4L4G9G9L4M3L4M4K4MkMdFjNZ9R1TGfNj8V1eGaNY8[1mGeNQ8W1SHkNk71RGROX1n0c70\\GhNV1Y1\\7MfG`NS1d1U7LgIoN]M`0j8b0iInN]M`0i8b0kInN]M>h8e0kIlN_M>e8g0kIlNaM<c8h0mIlNaM;a8j0nIjNcM;^8k0PJjNbM:^8m0PJiNcM9\\8m0SJjNbM8Z8m0VJjNbM7X8m0YJlN`M5W8n0ZJnN_M3W8n0\\JnN_M1V8o0^JPO]MOU8P1`JQOZMOV8o0bJROWMOW8m0eJSOTMOY8l0eJUORMNY8l0gJVOoLNZ8j0jJXOkLL]8k0iJYOkLI^8e0QKB`LGa8=YKMULEc88^K3oKCf85_K9kK_Oh84aK>fK\\Ok82cKb0d90O1O001O00N2N3M2N2N2N2O100O2O000O1O001O1N2Ok`Z4"}, "label": "a man wearing nike shoes in a bright neon green top is playing tennis"}]}
{"id": 1966, "image": "COCO_train2014_000000001966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male tennis player serving in a tennis match\"."}], "task": "refering", "answer_template": "The \"male tennis player serving in a tennis match\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 78, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 194, 195, 196, 212, 213, 214, 215, 217, 218, 235, 236, 237, 238, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 326, 327, 328, 329, 330, 331, 350, 351, 352, 353], "bbox": [0.197640625, 0.14752083333333335, 0.5388437500000001, 0.9177916666666667], "iscrowd": 0, "area": 27770.737749999986, "rle": {"size": [480, 640], "counts": "kmk12m>2N1N3N1O2N1N3N2N1O2N1N3N1O2L4L3M4L3M4K4M4L4M2N3M2M4M3M2N3L3N3M2_MRNmGR2g7YNWHj1]7aNaHa1T7jNjHY1j6ROSIR1a6YO]Ii0Y6AeIb0X6@fIc0W6_OfIZOROmNV7k1fISO]OQOj6n1gIoNEROa6P2iInNJPO[6T2hIlN1POU6U2hIkN7nNo5Y2gIjN>kNi5]2fIiNe0gNd5b2cIhNn0dN]5f2bIgNU1`NX5k2`IfN[1^NS5[5QKbJn4]5VKaJj4^5YKaJe4_5^K_Jb4`5aK_J^4`5eK^J[4`5hK`JW4_5lK_JT4`5oK_JP4`5SL^Jm3`5j2O1000O0100O10000O100O100O01000O1o0QO6J6J6TIiId4\\6WKdIi4b6PK`Io4f6jJ\\IU5j6dJXIZ5o6_JSI`5S7YJoHf5V8O1O1O1O1O1O1O001N2OhIaJQ3^5nLfJP3Y5PMiJo2W5PMlJn2S5QMPKb0J_NV5n0SK1>jN^4U1VKAR1SOh3[1YK]OY1QO]3b1\\KYOa1nNS3h1PLcNX1]Oh2P2]LTNn0Le2o1jLfMd0;a2o1WMXM:j0_2m1;SNDn1;SNEk1=UNBk1a0SN_Ol1c0SN\\Ol1g0SNYOl1j0RNUOm1n0RNROm1P1RNoNm1U1QNkNn1W1PNjNn1Y1QNfNo1]1oMcNP2_1oMaNX2Y1gMgNa2R1^MmNd2S1[MmNe2S1[MmNf2R1ZMnNf2R1ZMnNf2R1ZMmNh2S1WMmNi2S1WMmNj2R1UMoNk2Q1UMnNm2Q1RMPOn2Q1PMPOP3P1oLQOR3n0mLSOS3m0lLSOV3m0gLUO[3i0dLXO^3f0aL[Oa3c0^L]Of3k54L4L4L4L4G9G9L4M3L4M4K4MkMdFjNZ9R1TGfNj8V1eGaNY8[1mGeNQ8W1SHkNk71RGROX1n0c70\\GhNV1Y1\\7MfG`NS1d1U7LgIoN]M`0j8b0iInN]M`0i8b0kInN]M>h8e0kIlN_M>e8g0kIlNaM<c8h0mIlNaM;a8j0nIjNcM;^8k0PJjNbM:^8m0PJiNcM9\\8m0SJjNbM8Z8m0VJjNbM7X8m0YJlN`M5W8n0ZJnN_M3W8n0\\JnN_M1V8o0^JPO]MOU8P1`JQOZMOV8o0bJROWMOW8m0eJSOTMOY8l0eJUORMNY8l0gJVOoLNZ8j0jJXOkLL]8k0iJYOkLI^8e0QKB`LGa8=YKMULEc88^K3oKCf85_K9kK_Oh84aK>fK\\Ok82cKb0d90O1O001O00N2N3M2N2N2N2O100O2O000O1O001O1N2Ok`Z4"}, "label": "male tennis player serving in a tennis match"}]}
{"id": 165803, "image": "COCO_train2014_000000165803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black man cutting cake\"."}], "task": "refering", "answer_template": "The \"a black man cutting cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 142, 143, 144, 145, 146, 147, 148, 149, 153, 154, 155, 157, 158, 159, 160, 161, 162, 163, 164, 168, 169, 170, 172, 173, 174, 175, 176, 177, 178, 183, 184, 187, 188, 189, 190, 191, 192, 193, 198, 199, 202, 203, 204, 205, 206, 207, 212, 213, 214, 217, 218, 219, 220, 221, 222, 228, 229, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 266, 267, 281, 282], "bbox": [0.18861826697892273, 0.008609374999999999, 0.9902341920374709, 0.817609375], "iscrowd": 0, "area": 98139.30605000001, "rle": {"size": [640, 427], "counts": "bPc13lc02M3M4M2M3N2M4L3N2M3M4M2M3N1N3N2M3N2M2O2`MnNnAU1l=YOjAi0P>EeA>U>0aA1o=Y3E;I7N2N1O1O1O2N1N2N2M4M2M3N1N2N2N2N2N2K5N1O1O2N101O001NUNaCYL^<f3lCSLS<k3WDnKg;Q4dDfK[;Y4oD`KP;^4ZE[Kd:f4gEoJX:S5Y20N3N1O2O1O001O001O010O00aNgAZMX>f2PBRMP>o2VBkLj=U3\\BdLd=\\3cB^L]=b3iBWLW=j3PCoKP=P4a1O001N10001N10000O1000O010000O001O1O0O2O1O1O001O001O0000001O1O1O1O1O1O1O1O1O1O2N1O1O10WCSM`8R3oFmM^8W2SG<f6GjHk0P7ZO^HY1\\7kNSHh1g7\\NdGZ2W8j4M4L2O0O1O001O1O1O1O011N1O2N1O2O0O2N101N1O2O0O2N103L4L4L3N3L4L2O1N2N2N2N1O2N2N2N2N2N1O2N2N1O1O1O1O001O1O1O001O100O001O1O10O01O1O1O10O0100OXOi0K4L5K4L5L3L5K5K4L5L33N;lHUD\\5[=M3L4M3M3M1O1O001O1O001O1O001O001O001N101O001O001O001O001O1O001O001^K\\Af3d>WL_Ai3b>SLbAl3^>QLfAn3[>mKjAR4V>kKnAT4S>hKQBW4o=fKTBZ4j>O1O1O001O1O1000O011O3M2M0100O010O010O010O10O0M4K6J6G9G8QMSLbEV4e9XM]ER3V:d3H9H7I6K4N2N2O1N1O2N2O1N1O2N2M2N3M2N3L4M2N3M3M2N3M2N4K7J5K6J5K7I7I8G7J6J6J7I7I7I7I4L3M4J6K4L5J5L5K5J5L5POo0I8I7H9H;D;FR[2"}, "label": "a black man cutting cake"}]}
{"id": 165803, "image": "COCO_train2014_000000165803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a men in purple shirt celebrating with his friends\"."}], "task": "refering", "answer_template": "The \"a men in purple shirt celebrating with his friends\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 142, 143, 144, 145, 146, 147, 148, 149, 153, 154, 155, 157, 158, 159, 160, 161, 162, 163, 164, 168, 169, 170, 172, 173, 174, 175, 176, 177, 178, 183, 184, 187, 188, 189, 190, 191, 192, 193, 198, 199, 202, 203, 204, 205, 206, 207, 212, 213, 214, 217, 218, 219, 220, 221, 222, 228, 229, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 266, 267, 281, 282], "bbox": [0.18861826697892273, 0.008609374999999999, 0.9902341920374709, 0.817609375], "iscrowd": 0, "area": 98139.30605000001, "rle": {"size": [640, 427], "counts": "bPc13lc02M3M4M2M3N2M4L3N2M3M4M2M3N1N3N2M3N2M2O2`MnNnAU1l=YOjAi0P>EeA>U>0aA1o=Y3E;I7N2N1O1O1O2N1N2N2M4M2M3N1N2N2N2N2N2K5N1O1O2N101O001NUNaCYL^<f3lCSLS<k3WDnKg;Q4dDfK[;Y4oD`KP;^4ZE[Kd:f4gEoJX:S5Y20N3N1O2O1O001O001O010O00aNgAZMX>f2PBRMP>o2VBkLj=U3\\BdLd=\\3cB^L]=b3iBWLW=j3PCoKP=P4a1O001N10001N10000O1000O010000O001O1O0O2O1O1O001O001O0000001O1O1O1O1O1O1O1O1O1O2N1O1O10WCSM`8R3oFmM^8W2SG<f6GjHk0P7ZO^HY1\\7kNSHh1g7\\NdGZ2W8j4M4L2O0O1O001O1O1O1O011N1O2N1O2O0O2N101N1O2O0O2N103L4L4L3N3L4L2O1N2N2N2N1O2N2N2N2N2N1O2N2N1O1O1O1O001O1O1O001O100O001O1O10O01O1O1O10O0100OXOi0K4L5K4L5L3L5K5K4L5L33N;lHUD\\5[=M3L4M3M3M1O1O001O1O001O1O001O001O001N101O001O001O001O001O1O001O001^K\\Af3d>WL_Ai3b>SLbAl3^>QLfAn3[>mKjAR4V>kKnAT4S>hKQBW4o=fKTBZ4j>O1O1O001O1O1000O011O3M2M0100O010O010O010O10O0M4K6J6G9G8QMSLbEV4e9XM]ER3V:d3H9H7I6K4N2N2O1N1O2N2O1N1O2N2M2N3M2N3L4M2N3M3M2N3M2N4K7J5K6J5K7I7I8G7J6J6J7I7I7I7I4L3M4J6K4L5J5L5K5J5L5POo0I8I7H9H;D;FR[2"}, "label": "a men in purple shirt celebrating with his friends"}]}
{"id": 75697, "image": "COCO_train2014_000000075697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a patch of mutli - colored blue jeans\"."}], "task": "refering", "answer_template": "The \"a patch of mutli - colored blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 251, 252, 253, 266, 267, 268, 282, 283, 284, 297, 298, 299, 313, 314, 328, 329, 343, 344], "bbox": [0.7619859813084112, 0.651828125, 1.0, 0.9985156249999999], "iscrowd": 0, "area": 9471.452449999997, "rle": {"size": [640, 428], "counts": "RV\\62lc07I7J6I6J7J6I7I5K4M2M4L4M3L3M4L4M2L5N2000O0100000O01000O10O1000O0100000O01000O10O1000O01\\AZMl;f2oC]MS<c2hC`MZ<_2bCdM_<]2[CgMg<Y2TCjMn<U2nBnMT=R2gBQNZ=P2`BTNb=l1YBWNi=h1SB[Nn=f1mA]NU>c1eAaN]>_1^AdNd>[1XAhNi>Y1QAkNQ?U1j@nNX?l22N1O2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N2N3M3M4L3M4L3M4L4L3M4L3M4L3M4L3M4L4L3M4L6Jl0TO_\\O"}, "label": "a patch of mutli - colored blue jeans"}]}
{"id": 75697, "image": "COCO_train2014_000000075697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wood chair\"."}], "task": "refering", "answer_template": "The \"wood chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 192, 193, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 268, 269, 284, 299], "bbox": [0.8052570093457944, 0.5013124999999999, 1.0, 0.87990625], "iscrowd": 0, "area": 9927.834950000002, "rle": {"size": [640, 428], "counts": "\\og6n0lb0i0VO9H8H8K5N2M3M3M3N2M2N3N2M3M3M3N2M3M3N2M2N3M3N2M3MiKc@o3R`0F5KO1000O010000O10O10O01000O0O2N101N1O2O0O2N101N1O2O0O2N101N1O2O0O2N101N101O1N3N2N2M3N2N2M3N2N1N3N2N2N2OjNV1YNh1XNh1_OnC"}, "label": "wood chair"}]}
{"id": 75697, "image": "COCO_train2014_000000075697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of a rocking chair\"."}], "task": "refering", "answer_template": "The \"the arm of a rocking chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 192, 193, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 268, 269, 284, 299], "bbox": [0.8052570093457944, 0.5013124999999999, 1.0, 0.87990625], "iscrowd": 0, "area": 9927.834950000002, "rle": {"size": [640, 428], "counts": "\\og6n0lb0i0VO9H8H8K5N2M3M3M3N2M2N3N2M3M3M3N2M3M3N2M2N3M3N2M3MiKc@o3R`0F5KO1000O010000O10O10O01000O0O2N101N1O2O0O2N101N1O2O0O2N101N1O2O0O2N101N101O1N3N2N2M3N2N2M3N2N1N3N2N2N2OjNV1YNh1XNh1_OnC"}, "label": "the arm of a rocking chair"}]}
{"id": 305076, "image": "COCO_train2014_000000305076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hummingbird in flight with a green neck and head\"."}], "task": "refering", "answer_template": "The \"a hummingbird in flight with a green neck and head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 96, 97, 98, 99, 112, 113, 114, 129, 130], "bbox": [0.5883744855967078, 0.31041994750656166, 0.8872427983539094, 0.523517060367454], "iscrowd": 0, "area": 4003.505249999998, "rle": {"size": [381, 486], "counts": "\\cZ31k;1O1N2O1O1O1N2O1O1O1N1000001N2O001N2O1N101N3N1O2M2O1N3N1N2O2N1N3NG92N1O1O2N1O2N1O2N1O2N2N1O2N2N101NN2N2N2O1100O1O2N2N2N2NO1O1O10O1000000100O1O1O1O1O1O1O001O1O1M3M3M3L4M3L30100O2N1O2N101N1O2N1O2O1N1O2N1O2O0O2N110O0000001O0000000001O01O00000000000010O00000001O001O0010O01O001O001O001O0010O0`Zd0"}, "label": "a hummingbird in flight with a green neck and head"}]}
{"id": 305076, "image": "COCO_train2014_000000305076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird that has green on it\"."}], "task": "refering", "answer_template": "The \"the bird that has green on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 96, 97, 98, 99, 112, 113, 114, 129, 130], "bbox": [0.5883744855967078, 0.31041994750656166, 0.8872427983539094, 0.523517060367454], "iscrowd": 0, "area": 4003.505249999998, "rle": {"size": [381, 486], "counts": "\\cZ31k;1O1N2O1O1O1N2O1O1O1N1000001N2O001N2O1N101N3N1O2M2O1N3N1N2O2N1N3NG92N1O1O2N1O2N1O2N1O2N2N1O2N2N101NN2N2N2O1100O1O2N2N2N2NO1O1O10O1000000100O1O1O1O1O1O1O001O1O1M3M3M3L4M3L30100O2N1O2N101N1O2N1O2O1N1O2N1O2O0O2N110O0000001O0000000001O01O00000000000010O00000001O001O0010O01O001O001O001O0010O0`Zd0"}, "label": "the bird that has green on it"}]}
{"id": 305076, "image": "COCO_train2014_000000305076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the humming bird on the left\"."}], "task": "refering", "answer_template": "The \"the humming bird on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 159, 160], "bbox": [0.10644032921810699, 0.42913385826771655, 0.52, 0.682230971128609], "iscrowd": 0, "area": 8769.410250000003, "rle": {"size": [381, 486], "counts": "[ac01l;1O1O1O2N1O1O1O1O2N2N2N2N2N2N2N2N1O2N1O1O1O2N1O1O1N3N1O1O1O1O0000000000000000000000000000000000000000001O0000000000001O00000O1000001O00000000001O0000O10000O1000000O10000OCfNYFZ1f9gNZFY1e9gN\\FX1d9iN\\FW1c9iN^FW1`9kN`FT1]9oNdFQ1Z9QOfFo0Y9QOhFo0W9TOgFm0W9VOgFj0Y9XOeFh0Z9ZOeFf0Z9]OdFc0\\9^OdFa0\\9AbF?]9CbF=^9E`F;_9GaF8_9S100000000000010O1O10O0100O03N2N9G1N2O1O1O1O1N1000000O1000O1O001O01O00000000001O0000000000000O10000000000O100O1O100QOiE7X:HjE6W:JjE3Y:KiE3X:MiE2W:MlE0U:0lEOT:0nENS:2nEMR:3oEKR:4oELQ:4oELQ:4oEKS:3oELQ:4oEKR:4oELQ:4oEKR:5nEKR:4PFKP:5PFKQ:3PFMP:3PFMP:3QFLo93RFMo92QFNo91RFOn91RFOo90QF0o9ORF1n9OSF0n9NSF2n9MRF3n9MRF3l:0000N2O1Ocjf2"}, "label": "the humming bird on the left"}]}
{"id": 305076, "image": "COCO_train2014_000000305076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird with white underbelly\"."}], "task": "refering", "answer_template": "The \"bird with white underbelly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 159, 160], "bbox": [0.10644032921810699, 0.42913385826771655, 0.52, 0.682230971128609], "iscrowd": 0, "area": 8769.410250000003, "rle": {"size": [381, 486], "counts": "[ac01l;1O1O1O2N1O1O1O1O2N2N2N2N2N2N2N2N1O2N1O1O1O2N1O1O1N3N1O1O1O1O0000000000000000000000000000000000000000001O0000000000001O00000O1000001O00000000001O0000O10000O1000000O10000OCfNYFZ1f9gNZFY1e9gN\\FX1d9iN\\FW1c9iN^FW1`9kN`FT1]9oNdFQ1Z9QOfFo0Y9QOhFo0W9TOgFm0W9VOgFj0Y9XOeFh0Z9ZOeFf0Z9]OdFc0\\9^OdFa0\\9AbF?]9CbF=^9E`F;_9GaF8_9S100000000000010O1O10O0100O03N2N9G1N2O1O1O1O1N1000000O1000O1O001O01O00000000001O0000000000000O10000000000O100O1O100QOiE7X:HjE6W:JjE3Y:KiE3X:MiE2W:MlE0U:0lEOT:0nENS:2nEMR:3oEKR:4oELQ:4oELQ:4oEKS:3oELQ:4oEKR:4oELQ:4oEKR:5nEKR:4PFKP:5PFKQ:3PFMP:3PFMP:3QFLo93RFMo92QFNo91RFOn91RFOo90QF0o9ORF1n9OSF0n9NSF2n9MRF3n9MRF3l:0000N2O1Ocjf2"}, "label": "bird with white underbelly"}]}
{"id": 272310, "image": "COCO_train2014_000000272310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tail of small silver , white and red airplane\"."}], "task": "refering", "answer_template": "The \"the tail of small silver , white and red airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 83, 84, 85, 86, 87, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181], "bbox": [0.61925, 0.16104018912529552, 0.988421875, 0.5304018912529551], "iscrowd": 0, "area": 15461.130600000002, "rle": {"size": [423, 640], "counts": "PST55R=000PE2g8NYG2g8NYG1g80YG0g80YG0g80XG1h8OXG1h80WG0i80WG0i80WG0i80WG0i80WG0i80WG0i80WG0i80WG0i80VG1j8OVG1j8OVG1j8OVG1j8OVG1j8OVG1j8OVG1j8OVG0k80UG0k80TG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OSG2m8NSG2m8NSG2m8ORG1n8ORG1n8ORG1n8ORG1n8ORG1n8ORG0o80PG1P9OPG1P9OPG1P9OPG1]8IYF6Z11[8MYF2\\11Y80ZFO]11W83[FL^11V86ZFI`11V87YFHa11V89WFFb12V8:WFDc12V8;VFCd12V8=TFAf12U8?TF_Og12U8a0RF^Oh11V8m0hGSOX8n0gGROX8P1gGPOY8Q1fGnN[8T1bGmN]8U1bGkN^8V1aGjN_8W1`GiN_8Y1`GgN`8[1^GeNUOGl8e1nGdNTOJl8d1oGbNTOKn8c1mGbNSONP9`1kGcNSOOR9`1iGaNUO0R9_1hGaNWOOR9`1fGaNXO0R9_1eGaNYO0R9a1cG_N[O1R9`1bG_N\\O1S9`1`G_N]O2S9_1_G_N^O2S9`1^G^N_O3S9`1[G]NC3S9`1YG]ND3T9`1WG]NE4S9`1WG\\NF4T9`1UG\\NG5T9`1SG[NI5U9`1QG[NK5S9a1QGZNL5T9a1oFZNM6T9a1mFYNO6U9a1jFZN16T9a1hF[N44U9b1cF\\N83U9T2kFlMV9S2jFnMU9R2kFnMV9R2iFnMX9R2gFnMZ9Q2fFoMZ9R2eFoM[9Q2eFnM\\9R2cFnM^9Q2bFoM^9R2aFnM`9R2_FnMb9R2]FnMd9S2ZFmMf9T2YFlMh9T2WFlMj9h210O01O1O1O001O1O1O001O100O001O1O1O001O1O1O00100O1O001O1O10O101O000O1000000O10O1O10O01O1O10O01O1O01O0O100000O100O1O1O010O1WOcEgN^:Y1gEbNY:^1lE\\NU:d1mEYNT:g1oEUNR:k1e00O10000000000000000O100000001O00000O10000000000000000O10000000000000000O100000000000000O2O00000000000oDlMo:T2QElMo:V2O01000ch2"}, "label": "the tail of small silver , white and red airplane"}]}
{"id": 272310, "image": "COCO_train2014_000000272310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver plane with a red lightning bolt\"."}], "task": "refering", "answer_template": "The \"a silver plane with a red lightning bolt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 83, 84, 85, 86, 87, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181], "bbox": [0.61925, 0.16104018912529552, 0.988421875, 0.5304018912529551], "iscrowd": 0, "area": 15461.130600000002, "rle": {"size": [423, 640], "counts": "PST55R=000PE2g8NYG2g8NYG1g80YG0g80YG0g80XG1h8OXG1h80WG0i80WG0i80WG0i80WG0i80WG0i80WG0i80WG0i80WG0i80VG1j8OVG1j8OVG1j8OVG1j8OVG1j8OVG1j8OVG1j8OVG0k80UG0k80TG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OTG1l8OSG2m8NSG2m8NSG2m8ORG1n8ORG1n8ORG1n8ORG1n8ORG1n8ORG0o80PG1P9OPG1P9OPG1P9OPG1]8IYF6Z11[8MYF2\\11Y80ZFO]11W83[FL^11V86ZFI`11V87YFHa11V89WFFb12V8:WFDc12V8;VFCd12V8=TFAf12U8?TF_Og12U8a0RF^Oh11V8m0hGSOX8n0gGROX8P1gGPOY8Q1fGnN[8T1bGmN]8U1bGkN^8V1aGjN_8W1`GiN_8Y1`GgN`8[1^GeNUOGl8e1nGdNTOJl8d1oGbNTOKn8c1mGbNSONP9`1kGcNSOOR9`1iGaNUO0R9_1hGaNWOOR9`1fGaNXO0R9_1eGaNYO0R9a1cG_N[O1R9`1bG_N\\O1S9`1`G_N]O2S9_1_G_N^O2S9`1^G^N_O3S9`1[G]NC3S9`1YG]ND3T9`1WG]NE4S9`1WG\\NF4T9`1UG\\NG5T9`1SG[NI5U9`1QG[NK5S9a1QGZNL5T9a1oFZNM6T9a1mFYNO6U9a1jFZN16T9a1hF[N44U9b1cF\\N83U9T2kFlMV9S2jFnMU9R2kFnMV9R2iFnMX9R2gFnMZ9Q2fFoMZ9R2eFoM[9Q2eFnM\\9R2cFnM^9Q2bFoM^9R2aFnM`9R2_FnMb9R2]FnMd9S2ZFmMf9T2YFlMh9T2WFlMj9h210O01O1O1O001O1O1O001O100O001O1O1O001O1O1O00100O1O001O1O10O101O000O1000000O10O1O10O01O1O10O01O1O01O0O100000O100O1O1O010O1WOcEgN^:Y1gEbNY:^1lE\\NU:d1mEYNT:g1oEUNR:k1e00O10000000000000000O100000001O00000O10000000000000000O10000000000000000O100000000000000O2O00000000000oDlMo:T2QElMo:V2O01000ch2"}, "label": "a silver plane with a red lightning bolt"}]}
{"id": 575417, "image": "COCO_train2014_000000575417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray - haired man in a white shirt stands in an office looking down at the table that has pizzas on it\"."}], "task": "refering", "answer_template": "The \"a gray - haired man in a white shirt stands in an office looking down at the table that has pizzas on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 40, 41, 42, 63, 64, 65, 66, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 272, 273], "bbox": [0.7314375, 0.00014084507042253522, 0.9273906249999999, 0.7776760563380282], "iscrowd": 0, "area": 27592.991749999997, "rle": {"size": [426, 640], "counts": "QgR62W=3M3L4M3M4K4M3M3gJYOiMk0m1^OQNe0e1CZNa0\\1GaN=U1KjN8l01RO2d06\\OL;<DG1a0NBHg07[O_On0?UOWOS1i0oNnNY1Q1jNdN^1[1eN[Nd1d1^NRNj1n1XNiMP2U2oNeLZ1Z3[4O2O001N101N2O0010O010O0100O010O010O0010O0100O010O010O010O10O100O101N101O3L5L3L4M0O01000O100O100O100O10000O10O0100O10000O100O100O10000O010O100O100I7G9F:G9I7I7J6I6K6J6L4L4L4M3L4L4L6I;C<D<E;Dd0\\Of0ZOg0ZOe0ZOg0YOenb0"}, "label": "a gray - haired man in a white shirt stands in an office looking down at the table that has pizzas on it"}]}
{"id": 575417, "image": "COCO_train2014_000000575417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing in front of a pizza box looking at pizza\"."}], "task": "refering", "answer_template": "The \"a man standing in front of a pizza box looking at pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 40, 41, 42, 63, 64, 65, 66, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 272, 273], "bbox": [0.7314375, 0.00014084507042253522, 0.9273906249999999, 0.7776760563380282], "iscrowd": 0, "area": 27592.991749999997, "rle": {"size": [426, 640], "counts": "QgR62W=3M3L4M3M4K4M3M3gJYOiMk0m1^OQNe0e1CZNa0\\1GaN=U1KjN8l01RO2d06\\OL;<DG1a0NBHg07[O_On0?UOWOS1i0oNnNY1Q1jNdN^1[1eN[Nd1d1^NRNj1n1XNiMP2U2oNeLZ1Z3[4O2O001N101N2O0010O010O0100O010O010O0010O0100O010O010O010O10O100O101N101O3L5L3L4M0O01000O100O100O100O10000O10O0100O10000O100O100O10000O010O100O100I7G9F:G9I7I7J6I6K6J6L4L4L4M3L4L4L6I;C<D<E;Dd0\\Of0ZOg0ZOe0ZOg0YOenb0"}, "label": "a man standing in front of a pizza box looking at pizza"}]}
{"id": 575417, "image": "COCO_train2014_000000575417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the black shirt\"."}], "task": "refering", "answer_template": "The \"the woman in the black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 231, 232, 233, 255, 256], "bbox": [0.0, 0.24924882629107983, 0.194125, 0.780962441314554], "iscrowd": 0, "area": 14050.782700000002, "rle": {"size": [426, 640], "counts": "n3<Y10d9?iE1R1UO[7n0\\G1W1SO\\7o0WG2[1QO]7[2aHhM]7Y2aHiM^7X2aHiM^7X2`HjM`7V2^HmMa7o1aHSN_7j1cHWN\\7f1eH]N[7_1fHeNY7\\1cHfN^7`1ZHaNg7e1mG`NT8Z31N101O1O2N1O2M3O0O2N1O2N2N1O2N1016J8G7J3M4L3N3L101N10WK`Gf4`8YK`Gi4`81OXOaGeL^8W3fGkLW8S3lGnLR8P3PHSMm7k2VHVMh7g2\\HZMc7c2_H`M^7^2eHcMY7[2jHgMS7X2oHkMm6U2SInMj6R2WIQNf6n1[IUNa6k1_IYN]6g1dI\\NX6d1iI\\NV6d1kI[NU6d1lI\\NT6d1mIZNT6f1mIYNS6g1nIXNR6h1nIWNS6i1nIVNR6j1oIUNQ6k1oITNR6l1oISNQ6m1PJRNP6n1QJPNP6o1QJQNo5o1RJPNo5o1RJoMo5Q2QJoMo5Q2RJnMo5Q2RJmMP6R2QJmMP6R2PJnMQ6Q2PJmMR6R2oImMR6R2oImMR6R2nIoMR6o1PJPNQ6o1PJPNQ6o1oIQNR6n1oIQNR6n1oIQNQ6o1PJPNQ6o1oIRNQ6m1PJRNQ6m1PJRNQ6m1PJRNS6k1QJQNQ6l1WJmMk5Q2]JhMd5V2i2O1_NTEb0n:[OUEc0l:\\OVEb0k:\\OXEb0j:[OXEd0i:[OYEc0h:[O[Ec0g:ZO[Ee0f:ZO\\Ec0g:ZO\\Ed0f:YO\\Ef0g:WO[Eg0f;M3N1N3N2M2N3N2M3N1N3NQ^f6"}, "label": "the woman in the black shirt"}]}
{"id": 575417, "image": "COCO_train2014_000000575417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black is looking at something\"."}], "task": "refering", "answer_template": "The \"a woman in black is looking at something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 231, 232, 233, 255, 256], "bbox": [0.0, 0.24924882629107983, 0.194125, 0.780962441314554], "iscrowd": 0, "area": 14050.782700000002, "rle": {"size": [426, 640], "counts": "n3<Y10d9?iE1R1UO[7n0\\G1W1SO\\7o0WG2[1QO]7[2aHhM]7Y2aHiM^7X2aHiM^7X2`HjM`7V2^HmMa7o1aHSN_7j1cHWN\\7f1eH]N[7_1fHeNY7\\1cHfN^7`1ZHaNg7e1mG`NT8Z31N101O1O2N1O2M3O0O2N1O2N2N1O2N1016J8G7J3M4L3N3L101N10WK`Gf4`8YK`Gi4`81OXOaGeL^8W3fGkLW8S3lGnLR8P3PHSMm7k2VHVMh7g2\\HZMc7c2_H`M^7^2eHcMY7[2jHgMS7X2oHkMm6U2SInMj6R2WIQNf6n1[IUNa6k1_IYN]6g1dI\\NX6d1iI\\NV6d1kI[NU6d1lI\\NT6d1mIZNT6f1mIYNS6g1nIXNR6h1nIWNS6i1nIVNR6j1oIUNQ6k1oITNR6l1oISNQ6m1PJRNP6n1QJPNP6o1QJQNo5o1RJPNo5o1RJoMo5Q2QJoMo5Q2RJnMo5Q2RJmMP6R2QJmMP6R2PJnMQ6Q2PJmMR6R2oImMR6R2oImMR6R2nIoMR6o1PJPNQ6o1PJPNQ6o1oIQNR6n1oIQNR6n1oIQNQ6o1PJPNQ6o1oIRNQ6m1PJRNQ6m1PJRNQ6m1PJRNS6k1QJQNQ6l1WJmMk5Q2]JhMd5V2i2O1_NTEb0n:[OUEc0l:\\OVEb0k:\\OXEb0j:[OXEd0i:[OYEc0h:[O[Ec0g:ZO[Ee0f:ZO\\Ec0g:ZO\\Ed0f:YO\\Ef0g:WO[Eg0f;M3N1N3N2M2N3N2M3N1N3NQ^f6"}, "label": "a woman in black is looking at something"}]}
{"id": 485306, "image": "COCO_train2014_000000485306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with mobile & a necklace\"."}], "task": "refering", "answer_template": "The \"the girl with mobile & a necklace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5662812500000001, 0.007330210772833723, 1.0, 0.9915456674473068], "iscrowd": 0, "area": 97472.06145000002, "rle": {"size": [427, 640], "counts": "][g43[<m0UOk0N2N2N2N2N2N3M2N2M3N2N3M2N2N2N2N2N3L3N2N2N2N2N2N2N2M3XJ\\Lg1f3RNcLj1_3oMkLl1W3mMRMo1P3jMZMQ2h2iM`MS2b2fMgMV2\\2cMmMX2U2dMQNX2Q2cMWNY2j1cM\\NZ2e1cMaNY2`1fMcNW2^1gMfNV2[1hMjNT2W1kMlNR2U1lMPOP2Q1oMROn1o0PNVOl1k0SNXOj1i0TN[Oi1f0VND`1=^NH^1:_NJ_16`NN\\13bN2Z1OeN5W1LgN8V1IiN;T1EjN`0R1AmNc0o0^OoNf0m0\\OQOj0j0WOTOo0g0ROWOT1d0mNYO[1?hN_O^19fNEa12cNKd1M`N3e1D_N<X700000000001O00000O1000000000001O000000O100000O10O2O000000000O1000000000001O000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000O1000000000000000000O1000000000000000000000000000000000000001O000000000000000000000000000000000O10000000000000000O10000000000000000O1000000000001N101O00001O1N2O1O1O2N1O1O1O1N3N1O1O1O2N1O1O3M2N2N2N1O1O1O2N1O1O1O1O1O1O1O001O1O1O2N1O2N2N2N1O2N3M8H8H8H8H6J3M3M2N3dHZMj3l2gKUN\\3U2WLkNl2fM"}, "label": "the girl with mobile & a necklace"}]}
{"id": 485306, "image": "COCO_train2014_000000485306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with cell phone and black choaker necklace\"."}], "task": "refering", "answer_template": "The \"woman with cell phone and black choaker necklace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5662812500000001, 0.007330210772833723, 1.0, 0.9915456674473068], "iscrowd": 0, "area": 97472.06145000002, "rle": {"size": [427, 640], "counts": "][g43[<m0UOk0N2N2N2N2N2N3M2N2M3N2N3M2N2N2N2N2N3L3N2N2N2N2N2N2N2M3XJ\\Lg1f3RNcLj1_3oMkLl1W3mMRMo1P3jMZMQ2h2iM`MS2b2fMgMV2\\2cMmMX2U2dMQNX2Q2cMWNY2j1cM\\NZ2e1cMaNY2`1fMcNW2^1gMfNV2[1hMjNT2W1kMlNR2U1lMPOP2Q1oMROn1o0PNVOl1k0SNXOj1i0TN[Oi1f0VND`1=^NH^1:_NJ_16`NN\\13bN2Z1OeN5W1LgN8V1IiN;T1EjN`0R1AmNc0o0^OoNf0m0\\OQOj0j0WOTOo0g0ROWOT1d0mNYO[1?hN_O^19fNEa12cNKd1M`N3e1D_N<X700000000001O00000O1000000000001O000000O100000O10O2O000000000O1000000000001O000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000O1000000000000000000O1000000000000000000000000000000000000001O000000000000000000000000000000000O10000000000000000O10000000000000000O1000000000001N101O00001O1N2O1O1O2N1O1O1O1N3N1O1O1O2N1O1O3M2N2N2N1O1O1O2N1O1O1O1O1O1O1O001O1O1O2N1O2N2N2N1O2N3M8H8H8H8H6J3M3M2N3dHZMj3l2gKUN\\3U2WLkNl2fM"}, "label": "woman with cell phone and black choaker necklace"}]}
{"id": 485306, "image": "COCO_train2014_000000485306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in green looking at her friend ' s cellphone\"."}], "task": "refering", "answer_template": "The \"a woman in green looking at her friend ' s cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.13826697892271664, 0.589765625, 1.0], "iscrowd": 0, "area": 92138.17205, "rle": {"size": [427, 640], "counts": "f;k0_<k0TO2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2N2O1N20000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O1M3L4L4L4L4L4L4L4L4L4L4L4M3N2M3N2N2N2M3N2N2M3N2N2N2ZNVIhLm6U3]IbLe6Z3fI]L\\6_3oIWLT6f3VJQLl5k3l1L4M3L4L4M3L4M3L4L4M3L4L4O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O010O100O100O100O1O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O00001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O002N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N3M2N2oJgKl0[4kNnKS1V4bNSL\\1T4]LYKb0k0P3T4RL`K<e0`3S4hKgK6?P4[7N2N2N1O2N2N2N2N2N1O001O001O001O001O001O001O00001O001O001O001O001O001@?ROo0XOg0G:F9G:Fl[]3"}, "label": "a woman in green looking at her friend ' s cellphone"}]}
{"id": 485306, "image": "COCO_train2014_000000485306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.13826697892271664, 0.589765625, 1.0], "iscrowd": 0, "area": 92138.17205, "rle": {"size": [427, 640], "counts": "f;k0_<k0TO2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2N2O1N20000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O1M3L4L4L4L4L4L4L4L4L4L4L4M3N2M3N2N2N2M3N2N2M3N2N2N2ZNVIhLm6U3]IbLe6Z3fI]L\\6_3oIWLT6f3VJQLl5k3l1L4M3L4L4M3L4M3L4L4M3L4L4O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O010O100O100O100O1O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O00001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O002N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N3M2N2oJgKl0[4kNnKS1V4bNSL\\1T4]LYKb0k0P3T4RL`K<e0`3S4hKgK6?P4[7N2N2N1O2N2N2N2N2N1O001O001O001O001O001O001O00001O001O001O001O001O001@?ROo0XOg0G:F9G:Fl[]3"}, "label": "a woman in a green shirt"}]}
{"id": 550844, "image": "COCO_train2014_000000550844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange with a sticker on the peel\"."}], "task": "refering", "answer_template": "The \"an orange with a sticker on the peel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 256, 257, 258], "bbox": [0.09526562499999999, 0.3910810810810811, 0.33696875, 0.7103326403326403], "iscrowd": 0, "area": 16884.9505, "rle": {"size": [481, 640], "counts": "\\ml04h>7H9H7I7H9H7H8I8J5L4M4L3M3M4K4M3M3M1O2M3N2N1O2N2M3N1O2N2N1N3N2O1O001N2O1O001O1N2O001O1O1N101O1O0O2O1O00000O2O0000000O1000001O0O100000000O10001O000O101ON2N2N1O2O2M2N2N3M2N2O1N3M2N2N3N1N2N3M2O101O000O2O00001O0O10001O0O101O000O2O00001O0O101O000O2O00000O2O1O1O1O1N3N1O1O1N2O1O2N1N2O1O1O1N3N1O1O1O1N2O2N1O1N2L4UOk0UOgnV6"}, "label": "an orange with a sticker on the peel"}]}
{"id": 550844, "image": "COCO_train2014_000000550844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange fruit sports a dark blue sticker with yellow writing\"."}], "task": "refering", "answer_template": "The \"an orange fruit sports a dark blue sticker with yellow writing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 256, 257, 258], "bbox": [0.09526562499999999, 0.3910810810810811, 0.33696875, 0.7103326403326403], "iscrowd": 0, "area": 16884.9505, "rle": {"size": [481, 640], "counts": "\\ml04h>7H9H7I7H9H7H8I8J5L4M4L3M3M4K4M3M3M1O2M3N2N1O2N2M3N1O2N2N1N3N2O1O001N2O1O001O1N2O001O1O1N101O1O0O2O1O00000O2O0000000O1000001O0O100000000O10001O000O101ON2N2N1O2O2M2N2N3M2N2O1N3M2N2N3N1N2N3M2O101O000O2O00001O0O10001O0O101O000O2O00001O0O101O000O2O00000O2O1O1O1O1N3N1O1O1N2O1O2N1N2O1O1O1N3N1O1O1O1N2O2N1O1N2L4UOk0UOgnV6"}, "label": "an orange fruit sports a dark blue sticker with yellow writing"}]}
{"id": 550844, "image": "COCO_train2014_000000550844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange behind the tin\"."}], "task": "refering", "answer_template": "The \"an orange behind the tin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 167, 168, 169], "bbox": [0.232375, 0.1892723492723493, 0.42871875, 0.4758004158004158], "iscrowd": 0, "area": 8336.341049999999, "rle": {"size": [481, 640], "counts": "ZTV2?T>a0L4L4L4K5L3L4L4M3L5L3L3N1O2O1N1O2N1O2N1O1O2O0000000010O1O1O010O1O001O10O01O10O10000O2O0O10000O101N10000O100O101O0O10000O10O00100O001O001O1O001O010O1O0O2O1O0O2O0O2O1O0O[NbMjF]2U9eMjF[2V9eMkFY2V9hMjFW2W9iMhFV2Y9kMgFT2Y9lMgFS2Z9nMfFQ2Z9PNfFn1\\9RNcFn1]9SNcFk1^9UNcFj1]9WNbFh1_9YNaF=XOFX:M_F5GGk95a300100O1O00100O1O10O01O100O010O10N1O2M2O2N2N2M3NVj[5"}, "label": "an orange behind the tin"}]}
{"id": 550844, "image": "COCO_train2014_000000550844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange that is touching 2 other oranges and also touching the black bowl\"."}], "task": "refering", "answer_template": "The \"the orange that is touching 2 other oranges and also touching the black bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354], "bbox": [0.201125, 0.5931185031185031, 0.4728125, 0.9334927234927235], "iscrowd": 0, "area": 21741.699099999998, "rle": {"size": [481, 640], "counts": "Unl16i>8G9H8H7I5J5L5K4K6K5K4L4M2N3M2M3N2N3M2N2N2N3L3N2N2N2N3M2N2N2N3N1O0O2O1N2O001N2O0O2O1O0O2O1N101O1N101N2O001N2O0O2O1O1N101N10000O10000O100O10000O100O100000000O1000000000000O1000000000000O100000O100000O10O1000000000O0100000000O10O100001O0O10001O000O2O0O1O1O2O0O1O2N1O1O101N1O1O2N1O2O2M3M3M2N3M3M3N2M2N3M3M3M3N1N3M3M3M3M2O2M3M3M3M2N3N2I7I8G:G9F:G9F:GYnm4"}, "label": "the orange that is touching 2 other oranges and also touching the black bowl"}]}
{"id": 550844, "image": "COCO_train2014_000000550844.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange touching the black bowl\"."}], "task": "refering", "answer_template": "The \"an orange touching the black bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354], "bbox": [0.201125, 0.5931185031185031, 0.4728125, 0.9334927234927235], "iscrowd": 0, "area": 21741.699099999998, "rle": {"size": [481, 640], "counts": "Unl16i>8G9H8H7I5J5L5K4K6K5K4L4M2N3M2M3N2N3M2N2N2N3L3N2N2N2N3M2N2N2N3N1O0O2O1N2O001N2O0O2O1O0O2O1N101O1N101N2O001N2O0O2O1O1N101N10000O10000O100O10000O100O100000000O1000000000000O1000000000000O100000O100000O10O1000000000O0100000000O10O100001O0O10001O000O2O0O1O1O2O0O1O2N1O1O101N1O1O2N1O2O2M3M3M2N3M3M3N2M2N3M3M3M3N1N3M3M3M3M2O2M3M3M3M2N3N2I7I8G:G9F:G9F:GYnm4"}, "label": "an orange touching the black bowl"}]}
{"id": 10176, "image": "COCO_train2014_000000010176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt watching two other people play a video game\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt watching two other people play a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 112, 113, 134, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 250, 251], "bbox": [0.8458125000000001, 0.18783333333333332, 0.9728125000000001, 0.6296250000000001], "iscrowd": 0, "area": 8640.924449999995, "rle": {"size": [480, 640], "counts": "lVn72m>2N1O2N2M201O0O2O1O1N2lN@fCa0Y<AZCODa0P=LPCFM`0Q=;mBFP=>nBCn<Z1L4L4N2N2N5K8H4\\FRMo6Q3oHSMl6P3RIUMgM3b8k2dIbM[6a2cI_M\\6e2`I\\M`6g2]IZMa6j2\\IVMd6m2YISMg6Q3TIPMm6R3PInLP7U3mHlLR7W3jHjLV7Y3gHgLY7\\3cHeL]7^3`HbL`7a3\\H`Ld7d3WH^Lh7j3mGYLS8o3aGUL`8U5001bJbGY4]8fKlGR4T8lKoGS4R8kKoGV4R8fKQHX4R8dKQHY4S8bKQH[4YOWKg86THa4TO]Kn9`4QFcKP:Z4oEhKR:a43M3O0O2O001N2N1N3N2M3^O[EcLj:m2kEiLZ:o2P1H8G8M5K4D=F9K6H7A]B^Of=8f0FQj7"}, "label": "a man in a black shirt watching two other people play a video game"}]}
{"id": 10176, "image": "COCO_train2014_000000010176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 112, 113, 134, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 250, 251], "bbox": [0.8458125000000001, 0.18783333333333332, 0.9728125000000001, 0.6296250000000001], "iscrowd": 0, "area": 8640.924449999995, "rle": {"size": [480, 640], "counts": "lVn72m>2N1O2N2M201O0O2O1O1N2lN@fCa0Y<AZCODa0P=LPCFM`0Q=;mBFP=>nBCn<Z1L4L4N2N2N5K8H4\\FRMo6Q3oHSMl6P3RIUMgM3b8k2dIbM[6a2cI_M\\6e2`I\\M`6g2]IZMa6j2\\IVMd6m2YISMg6Q3TIPMm6R3PInLP7U3mHlLR7W3jHjLV7Y3gHgLY7\\3cHeL]7^3`HbL`7a3\\H`Ld7d3WH^Lh7j3mGYLS8o3aGUL`8U5001bJbGY4]8fKlGR4T8lKoGS4R8kKoGV4R8fKQHX4R8dKQHY4S8bKQH[4YOWKg86THa4TO]Kn9`4QFcKP:Z4oEhKR:a43M3O0O2O001N2N1N3N2M3^O[EcLj:m2kEiLZ:o2P1H8G8M5K4D=F9K6H7A]B^Of=8f0FQj7"}, "label": "a man in a black shirt"}]}
{"id": 10176, "image": "COCO_train2014_000000010176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a grey and white stripped sweater and jeans playing the wii\"."}], "task": "refering", "answer_template": "The \"a man with a grey and white stripped sweater and jeans playing the wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 85, 86, 87, 108, 109, 110, 131, 132, 133, 134, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 341, 342, 360, 361, 362, 364, 365, 383, 384, 385], "bbox": [0.67921875, 0.16141666666666668, 0.9151718750000001, 0.9861458333333334], "iscrowd": 0, "area": 35054.99769999999, "rle": {"size": [480, 640], "counts": "\\[\\6;`>5L4L4L4K5M4M2N2N2M3G9G9UFaNk5i1SJiNT5`1jJRO]4W1aKmNT4\\1jKhNj3b1ULaNf2b2ZM`M]2g2bM\\MW2i2iMYMQ2k2oMWMl1l2SNWMoMfMm2W5TOTMbMVNR3i4\\OQM_M_Nk2e4ElL]MgNf2`4LjL\\MnN^2\\46fLYMVOY2W4>^NWOg1h0`NoNc1Q1ZOQNg0o1[OoMe0Q2^OkMc0T2AhM`0X2CeM=[2FaM;_2H]M9c2I[M7e2R6000000OM4I7I7I7I7C=XOh0N2M3N1O3L4M3M2M4M3M3L4M2M4M3M3L4M2N3L4M4L5J5L5K4WIiHc5\\7RJmHo5S7gIUIZ6i71O0O2O001N10003L3N3L4M3M2M4M3M3L4M2M4M5K5J7J6I6K6J3LO2O1N1O2O0O2N2O0O2N101N2N101N2N3M2O1N2N3N1N2N2O2M2N3N3L3M4SK^IV1e6fNnIg0V6VO\\J7h5ElJGX56ZKXOi4d0nKWNd4f1^4L4M3L4M3L7J6I7J6J6I7J6IkQi0"}, "label": "a man with a grey and white stripped sweater and jeans playing the wii"}]}
{"id": 10176, "image": "COCO_train2014_000000010176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a striped shirt playing the wii\"."}], "task": "refering", "answer_template": "The \"a man in a striped shirt playing the wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 85, 86, 87, 108, 109, 110, 131, 132, 133, 134, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 341, 342, 360, 361, 362, 364, 365, 383, 384, 385], "bbox": [0.67921875, 0.16141666666666668, 0.9151718750000001, 0.9861458333333334], "iscrowd": 0, "area": 35054.99769999999, "rle": {"size": [480, 640], "counts": "\\[\\6;`>5L4L4L4K5M4M2N2N2M3G9G9UFaNk5i1SJiNT5`1jJRO]4W1aKmNT4\\1jKhNj3b1ULaNf2b2ZM`M]2g2bM\\MW2i2iMYMQ2k2oMWMl1l2SNWMoMfMm2W5TOTMbMVNR3i4\\OQM_M_Nk2e4ElL]MgNf2`4LjL\\MnN^2\\46fLYMVOY2W4>^NWOg1h0`NoNc1Q1ZOQNg0o1[OoMe0Q2^OkMc0T2AhM`0X2CeM=[2FaM;_2H]M9c2I[M7e2R6000000OM4I7I7I7I7C=XOh0N2M3N1O3L4M3M2M4M3M3L4M2M4M3M3L4M2N3L4M4L5J5L5K4WIiHc5\\7RJmHo5S7gIUIZ6i71O0O2O001N10003L3N3L4M3M2M4M3M3L4M2M4M5K5J7J6I6K6J3LO2O1N1O2O0O2N2O0O2N101N2N101N2N3M2O1N2N3N1N2N2O2M2N3N3L3M4SK^IV1e6fNnIg0V6VO\\J7h5ElJGX56ZKXOi4d0nKWNd4f1^4L4M3L4M3L7J6I7J6J6I7J6IkQi0"}, "label": "a man in a striped shirt playing the wii"}]}
{"id": 10176, "image": "COCO_train2014_000000010176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black top with blue pants and glasses holding a wii remote\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black top with blue pants and glasses holding a wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 98, 99, 101, 102, 103, 121, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286], "bbox": [0.27934375, 0.20710416666666667, 0.50959375, 0.7601666666666667], "iscrowd": 0, "area": 18701.789650000006, "rle": {"size": [480, 640], "counts": "cPd24l>6J7I4L3M2O1N3M2N2N3M2O1N31N5L10N3M2N1ON2N2N2O1N;E4L0001O0000001O0000000XCBn:?QEHh:8WEKg:5XELh:4WEMi:3WEMi:4UEMk:3UEMk:3UEMk:3UEMk:4TEMk:3UEMk:3UEMk:3UEMk:3VELj:5UEKk:5UEKk:5UEKk:5UEKk:6TEJl:6TEJl:6UEIk:7UEIk:7UEIk:8TEHl:8TEHl:8TEHl:8nDNR;3gD3Y;M`D:`;FZDk0[;UO_Da1Q;_NiDW2g:W3jMO10000000O01000000O1000000O1000O10O0100O1dN[J]Ie5\\6eJ`I[5`6iJ\\IX5d6jJZIV5e6nJWIS5i6PKTIo4m6TKoHm4P7b100O010O10O1O010O00100O010O002O0O1O100O100O101N101O1N3oI`H^4a7`K`H_4b7_K`H`4b7]K`Ha4b7]K_Hc4b7ZKaHd4b7YK`Hf4a7XK`Hg4b7WK`Ha2DfNm7gNaH^2HiNi7fN`H\\2NkNd7fNaHY22oN^7gNaHU27QOZ7hNaHQ2<TOU7iNiHf18[OU7mNULl0Q4RORLh0S4VOPLd0U4ZOnK?^9J5J7Jklb4"}, "label": "a woman wearing a black top with blue pants and glasses holding a wii remote"}]}
{"id": 10176, "image": "COCO_train2014_000000010176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with the joystick , playing the video gaes\"."}], "task": "refering", "answer_template": "The \"a girl with the joystick , playing the video gaes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 98, 99, 101, 102, 103, 121, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286], "bbox": [0.27934375, 0.20710416666666667, 0.50959375, 0.7601666666666667], "iscrowd": 0, "area": 18701.789650000006, "rle": {"size": [480, 640], "counts": "cPd24l>6J7I4L3M2O1N3M2N2N3M2O1N31N5L10N3M2N1ON2N2N2O1N;E4L0001O0000001O0000000XCBn:?QEHh:8WEKg:5XELh:4WEMi:3WEMi:4UEMk:3UEMk:3UEMk:3UEMk:4TEMk:3UEMk:3UEMk:3UEMk:3VELj:5UEKk:5UEKk:5UEKk:5UEKk:6TEJl:6TEJl:6UEIk:7UEIk:7UEIk:8TEHl:8TEHl:8TEHl:8nDNR;3gD3Y;M`D:`;FZDk0[;UO_Da1Q;_NiDW2g:W3jMO10000000O01000000O1000000O1000O10O0100O1dN[J]Ie5\\6eJ`I[5`6iJ\\IX5d6jJZIV5e6nJWIS5i6PKTIo4m6TKoHm4P7b100O010O10O1O010O00100O010O002O0O1O100O100O101N101O1N3oI`H^4a7`K`H_4b7_K`H`4b7]K`Ha4b7]K_Hc4b7ZKaHd4b7YK`Hf4a7XK`Hg4b7WK`Ha2DfNm7gNaH^2HiNi7fN`H\\2NkNd7fNaHY22oN^7gNaHU27QOZ7hNaHQ2<TOU7iNiHf18[OU7mNULl0Q4RORLh0S4VOPLd0U4ZOnK?^9J5J7Jklb4"}, "label": "a girl with the joystick , playing the video gaes"}]}
{"id": 10179, "image": "COCO_train2014_000000010179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player holding a bat looking off into the didtance\"."}], "task": "refering", "answer_template": "The \"baseball player holding a bat looking off into the didtance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 104, 105, 106, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386, 404, 405, 406, 407, 408, 409], "bbox": [0.48040625, 0.14699203187250998, 0.8857031249999998, 0.9888645418326694], "iscrowd": 0, "area": 67431.156, "rle": {"size": [502, 640], "counts": "dif41a?6K6I6K5J4L5L3L5L3L5K4M4K4M3L5L3L4L5L3L4M3L5K4M3L5gLeLRJ^3j5fLSJg32VLY26cMk3LWLX22kMm3EYLZ2KoMT4^OYL^2CSNZ4fNhLU3oNSN`4ZNjLc3fNRNe4TNiLh3dNQNj4PNdLo3bNPNP5iMaLV4`NoMU5eM^L[4]NoMa7P2`HoMa7Q2_HoMb7P2^HoMc7P2^HoMc7Q2`HlMa7R2eHhM\\7W2jHcMW7\\2oH_MR7_2UIZMl6e2[ITMf6k2`IPM`6o2eIlL[6T3jIgLW6W3nIfLR6Y3SJbLm5^3WJ^Lj5`3[J]Le5b3\\J]Ld5c3^J[Lc5c3_J]La5b3`J]L`5d3aJZL`5f3`JZL_5f3bJYL_5g3bJWL_5i3aJWL^5j3bJUL_5k3aJTL`5l3aJSL^5m3cJRL\\5P4R33N2M3N2M4M2M38I7H5K5Ll0SO<D01O00001O00001O001O1O001O1O001O001O1O001O001O1O001O0aDUMS:P3fEVMV:P3bEVMZ:Q3]ETM`:R3XEQMf:R4N1O1O1O2N1O1O0000001O00001O00001O0000001O00001O00000000001O0000000O1000001O0000000000001O0000001O0000001O000000001O0000001O001O00001O001N101O001O1O1O1QJYJm1h5QN^Jj1c5TNdJf1^5WNhJd1Y5ZNkJc1V5[NmJc1T5[NPKb1Q5]NQKa1P5]NTK`1m4_NUK_1l4`NWK]1k4`NXK]1j4bNXK[1j4dNZKW1h4hN[KT1g4jN]KS1e4kN_KP1d4nNYKQMXN^3P7@dJXM]Nd2b72nI^McNh1U8h0UIdMhN_1V8l0nHkMmNT1Y8o0gHQNSOj0Z8S1_HXNYO=^8Z1VH^N]O0d8b1jGbNk9`1nEeNS:\\1gEhNY:[31O001O001O000O1O1O2N100O1O1O1O1O101N1O1N3N1N2O2M2O1N3N1fLaD6OQ1c;fNeEV1]:hNhET1Z:iNkER1W:lNmEQ1U:lNPFo0R:oNPFP1R:mNQFP1\\<N3M3M5K5K3M4L4L3MP^S1"}, "label": "baseball player holding a bat looking off into the didtance"}]}
{"id": 10179, "image": "COCO_train2014_000000010179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a boston uniform holds a bat\"."}], "task": "refering", "answer_template": "The \"a man in a boston uniform holds a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 104, 105, 106, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386, 404, 405, 406, 407, 408, 409], "bbox": [0.48040625, 0.14699203187250998, 0.8857031249999998, 0.9888645418326694], "iscrowd": 0, "area": 67431.156, "rle": {"size": [502, 640], "counts": "dif41a?6K6I6K5J4L5L3L5L3L5K4M4K4M3L5L3L4L5L3L4M3L5K4M3L5gLeLRJ^3j5fLSJg32VLY26cMk3LWLX22kMm3EYLZ2KoMT4^OYL^2CSNZ4fNhLU3oNSN`4ZNjLc3fNRNe4TNiLh3dNQNj4PNdLo3bNPNP5iMaLV4`NoMU5eM^L[4]NoMa7P2`HoMa7Q2_HoMb7P2^HoMc7P2^HoMc7Q2`HlMa7R2eHhM\\7W2jHcMW7\\2oH_MR7_2UIZMl6e2[ITMf6k2`IPM`6o2eIlL[6T3jIgLW6W3nIfLR6Y3SJbLm5^3WJ^Lj5`3[J]Le5b3\\J]Ld5c3^J[Lc5c3_J]La5b3`J]L`5d3aJZL`5f3`JZL_5f3bJYL_5g3bJWL_5i3aJWL^5j3bJUL_5k3aJTL`5l3aJSL^5m3cJRL\\5P4R33N2M3N2M4M2M38I7H5K5Ll0SO<D01O00001O00001O001O1O001O1O001O001O1O001O001O1O001O0aDUMS:P3fEVMV:P3bEVMZ:Q3]ETM`:R3XEQMf:R4N1O1O1O2N1O1O0000001O00001O00001O0000001O00001O00000000001O0000000O1000001O0000000000001O0000001O0000001O000000001O0000001O001O00001O001N101O001O1O1O1QJYJm1h5QN^Jj1c5TNdJf1^5WNhJd1Y5ZNkJc1V5[NmJc1T5[NPKb1Q5]NQKa1P5]NTK`1m4_NUK_1l4`NWK]1k4`NXK]1j4bNXK[1j4dNZKW1h4hN[KT1g4jN]KS1e4kN_KP1d4nNYKQMXN^3P7@dJXM]Nd2b72nI^McNh1U8h0UIdMhN_1V8l0nHkMmNT1Y8o0gHQNSOj0Z8S1_HXNYO=^8Z1VH^N]O0d8b1jGbNk9`1nEeNS:\\1gEhNY:[31O001O001O000O1O1O2N100O1O1O1O1O101N1O1N3N1N2O2M2O1N3N1fLaD6OQ1c;fNeEV1]:hNhET1Z:iNkER1W:lNmEQ1U:lNPFo0R:oNPFP1R:mNQFP1\\<N3M3M5K5K3M4L4L3MP^S1"}, "label": "a man in a boston uniform holds a bat"}]}
{"id": 10179, "image": "COCO_train2014_000000010179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player with a boston jersey and a dark cap\"."}], "task": "refering", "answer_template": "The \"a baseball player with a boston jersey and a dark cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 122, 123, 124, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379, 380, 397, 398, 399, 400, 401, 402, 403], "bbox": [0.182890625, 0.1847609561752988, 0.535421875, 0.985199203187251], "iscrowd": 0, "area": 52552.47525000001, "rle": {"size": [502, 640], "counts": "Wdi11b?5L3M4K4M3M4K4M4L3M3N3M2N2N3M2M4M2N2N3M2N3L3N2N3M2N3M2M2O2N1O2N1O2N1N3N1O2N1O2N1N3N1O2N1O2N1N3N1O2N1O2N1O2M2O2WGWLb5j3WJ`Le5b3TJhLg5Y3SJQMi5P3PJYMl5h2mIcMo5^2jIkMR6V2hITNT6m1eI]NV6d1cIfNZ6[1_IoN\\6R1aIUO[6l0bIYOZ6h0cI_OY6b0dICX6>eIIW68fINU63hI1W60fI1Z6f40O0010O100O100O100O2O0O10000O100O100O100gM^G]Mc8a2bG\\M^8c2eGZM\\8e2hGXMX8f2mGVMT8i2oGUMQ8i2THSMm7l2VHRMj7l2[HPMf7o2]HoLc7P3aHlL`7R3dHlL\\7S3hHiLY7U3kHiLU7V3f20000001O000000001O000000001O0000001O000000001O0000001O1O001O1O001O1O001oN^LlEc3R:`LlE`3T:bLjE_3T:eLiE\\3V:fLhEZ3X:hLfEY3X:jLfEW3Y:kLeEU3Z:nLdES3[:mLeET3Y:mLgES3Y:lLhEU3j0cLe79aGU3O]M_8^ObGl4]8TKeGl4Z8TKfGl4Y8UKgGl4X8SKiGm4V8TKjGm4U8RKlGn4S8SKmGn4R8RKnGn4Q8RKPHo4o7QKQHo4o7QKQHP5m7PKTHP5l7PKTHQ5j7PKVHP5j7oJWHR5g7oJYHQ5g7oJYHR5e7nJ\\HR5d7nJ\\HS5b7mJ_HS5a7mJ_HT5_7mJaHS5_7lJbHU5\\7lJdHT5\\7lJdHU5Z7kJgHU5Y7kJgHV5W7kJiHU5W7jJjHW5T7jJlHV5T7jJlHW5S7hJnHX5Q7iJoHX5o2RJ1e0QMY5j2WJ^MG_1j0YNY5e2[J_MJ^1b0^NY5a2`J_ML]1;dNZ5\\2cJ^M0^13hNZ5X2hJ^M3\\1LnNZ5S2lJ_M5\\1DSO[5n1QK^M9[1\\OYO[5j1TK^M<[1UO]O[5f1YK]M`0Z1lND\\5`1bMhNRMH\\5\\1jMgNkLM\\5X1QNfNbL3]5S1YNeN[L8]5n0aNeNRL=]5j0iNdNjKc0^5d0QOdNaKh0`5?VOdNZKn0b58\\OfNRKR1c53=MEM>4DG?9CAa0?A[Ob0f0@TOd0l0^OoNg0o0T7L5K4L4L5K4L4Ldca4"}, "label": "a baseball player with a boston jersey and a dark cap"}]}
{"id": 10179, "image": "COCO_train2014_000000010179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the man on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 122, 123, 124, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379, 380, 397, 398, 399, 400, 401, 402, 403], "bbox": [0.182890625, 0.1847609561752988, 0.535421875, 0.985199203187251], "iscrowd": 0, "area": 52552.47525000001, "rle": {"size": [502, 640], "counts": "Wdi11b?5L3M4K4M3M4K4M4L3M3N3M2N2N3M2M4M2N2N3M2N3L3N2N3M2N3M2M2O2N1O2N1O2N1N3N1O2N1O2N1N3N1O2N1O2N1N3N1O2N1O2N1O2M2O2WGWLb5j3WJ`Le5b3TJhLg5Y3SJQMi5P3PJYMl5h2mIcMo5^2jIkMR6V2hITNT6m1eI]NV6d1cIfNZ6[1_IoN\\6R1aIUO[6l0bIYOZ6h0cI_OY6b0dICX6>eIIW68fINU63hI1W60fI1Z6f40O0010O100O100O100O2O0O10000O100O100O100gM^G]Mc8a2bG\\M^8c2eGZM\\8e2hGXMX8f2mGVMT8i2oGUMQ8i2THSMm7l2VHRMj7l2[HPMf7o2]HoLc7P3aHlL`7R3dHlL\\7S3hHiLY7U3kHiLU7V3f20000001O000000001O000000001O0000001O000000001O0000001O1O001O1O001O1O001oN^LlEc3R:`LlE`3T:bLjE_3T:eLiE\\3V:fLhEZ3X:hLfEY3X:jLfEW3Y:kLeEU3Z:nLdES3[:mLeET3Y:mLgES3Y:lLhEU3j0cLe79aGU3O]M_8^ObGl4]8TKeGl4Z8TKfGl4Y8UKgGl4X8SKiGm4V8TKjGm4U8RKlGn4S8SKmGn4R8RKnGn4Q8RKPHo4o7QKQHo4o7QKQHP5m7PKTHP5l7PKTHQ5j7PKVHP5j7oJWHR5g7oJYHQ5g7oJYHR5e7nJ\\HR5d7nJ\\HS5b7mJ_HS5a7mJ_HT5_7mJaHS5_7lJbHU5\\7lJdHT5\\7lJdHU5Z7kJgHU5Y7kJgHV5W7kJiHU5W7jJjHW5T7jJlHV5T7jJlHW5S7hJnHX5Q7iJoHX5o2RJ1e0QMY5j2WJ^MG_1j0YNY5e2[J_MJ^1b0^NY5a2`J_ML]1;dNZ5\\2cJ^M0^13hNZ5X2hJ^M3\\1LnNZ5S2lJ_M5\\1DSO[5n1QK^M9[1\\OYO[5j1TK^M<[1UO]O[5f1YK]M`0Z1lND\\5`1bMhNRMH\\5\\1jMgNkLM\\5X1QNfNbL3]5S1YNeN[L8]5n0aNeNRL=]5j0iNdNjKc0^5d0QOdNaKh0`5?VOdNZKn0b58\\OfNRKR1c53=MEM>4DG?9CAa0?A[Ob0f0@TOd0l0^OoNg0o0T7L5K4L4L5K4L4Ldca4"}, "label": "the man on the left in the right hand picture"}]}
{"id": 239559, "image": "COCO_train2014_000000239559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck which is missing driver door\"."}], "task": "refering", "answer_template": "The \"truck which is missing driver door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 193, 194, 209, 210, 211, 212, 213], "bbox": [0.024078125, 0.2617291666666667, 0.47592187500000005, 0.5654374999999999], "iscrowd": 0, "area": 25796.5294, "rle": {"size": [480, 640], "counts": "od81Pn00PA0P?0o@1N1SO0iB2V=1fB1X=4cBM]=7^BK`=9\\BIc=;YBGd=l0O1N3N1N2O2M2N2O2M2O1N2O1N2O1N2N2O1N2O1N2O2M2O0O1O100O10O0100O1O100O100O100O100O1O100O100O100O00100O100O10O0100O1O1O1O1O00100O00001O0010O0001O00001O010O001O1O001O100O1O1O001O10000O10002M3N2N2M3N2N2M3N001N10001N10001N10001O0O101O00001O000010O01O00001O00000000O2O0O10000O100O100O10000N2N1O2N2N2N2N2N2N2N2M2N3N2M3N2000000000O010000000000O1000001O001N10001O001O0O2O001O001O0O2O001O001O0O2O2N2N2M3N2N2N2M3N1O2N2M3N1O000O1000000O10O10001O2N2N2N2N2N1O1N101O1O001O1O001O1O0010O0000001O0000001O01O0001O0000001N100O1O2N1O1O2N1O100O2UOaCRO`<l0k0O1N3L3L4M3L5K4M3L4Lkkl4"}, "label": "truck which is missing driver door"}]}
{"id": 239559, "image": "COCO_train2014_000000239559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the jeep on the right that no doors\"."}], "task": "refering", "answer_template": "The \"the jeep on the right that no doors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 193, 194, 209, 210, 211, 212, 213], "bbox": [0.024078125, 0.2617291666666667, 0.47592187500000005, 0.5654374999999999], "iscrowd": 0, "area": 25796.5294, "rle": {"size": [480, 640], "counts": "od81Pn00PA0P?0o@1N1SO0iB2V=1fB1X=4cBM]=7^BK`=9\\BIc=;YBGd=l0O1N3N1N2O2M2N2O2M2O1N2O1N2O1N2N2O1N2O1N2O2M2O0O1O100O10O0100O1O100O100O100O100O1O100O100O100O00100O100O10O0100O1O1O1O1O00100O00001O0010O0001O00001O010O001O1O001O100O1O1O001O10000O10002M3N2N2M3N2N2M3N001N10001N10001N10001O0O101O00001O000010O01O00001O00000000O2O0O10000O100O100O10000N2N1O2N2N2N2N2N2N2N2M2N3N2M3N2000000000O010000000000O1000001O001N10001O001O0O2O001O001O0O2O001O001O0O2O2N2N2M3N2N2N2M3N1O2N2M3N1O000O1000000O10O10001O2N2N2N2N2N1O1N101O1O001O1O001O1O0010O0000001O0000001O01O0001O0000001N100O1O2N1O1O2N1O100O2UOaCRO`<l0k0O1N3L3L4M3L5K4M3L4Lkkl4"}, "label": "the jeep on the right that no doors"}]}
{"id": 239559, "image": "COCO_train2014_000000239559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck with doors and a visible licence plate\"."}], "task": "refering", "answer_template": "The \"the truck with doors and a visible licence plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 112, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 265, 269, 270, 271, 272], "bbox": [0.48953125000000003, 0.2539375, 0.9661406249999999, 0.7145416666666667], "iscrowd": 0, "area": 43454.414150000004, "rle": {"size": [480, 640], "counts": "lmb4l0S>3N1N2O1N2O1N2O2M5K5L4K5L5J5L4K5L5J3M3N1N2O1N2O1N2O1N2O1N2N2O1N3N1N010O100O10O01O100O10O0100O100O00100O10O0100O100O100O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O100O2O00000O10000O10000O10001O0O10000O10000O1000000O101O0O10000O1000000O10000O101O0O1000000O10000000000010O00000000001O000000000010O2N4L4L4L010O00001O00010O00001O0010O0001O000010O0001O001O01O01O00001O00010O001O000010O0N2O1O1N2O2N1N2O1O1N2O2N1N2O1SNm100O2N100O1O100O2O0O1O100O1O2O0O100O1O101N100O1O12N12N3M2N3M3N1N3M3M2N3M2NO2N1O1O1N2O1O1O1O1O2M2O1O1OO1O001O1O1N2O001O1O1O1O1O001bNmDnNT;m0b1K5L3L5K5L4K5K5LiT:"}, "label": "the truck with doors and a visible licence plate"}]}
{"id": 239559, "image": "COCO_train2014_000000239559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vehicle on the right with doors\"."}], "task": "refering", "answer_template": "The \"the vehicle on the right with doors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 112, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 265, 269, 270, 271, 272], "bbox": [0.48953125000000003, 0.2539375, 0.9661406249999999, 0.7145416666666667], "iscrowd": 0, "area": 43454.414150000004, "rle": {"size": [480, 640], "counts": "lmb4l0S>3N1N2O1N2O1N2O2M5K5L4K5L5J5L4K5L5J3M3N1N2O1N2O1N2O1N2O1N2N2O1N3N1N010O100O10O01O100O10O0100O100O00100O10O0100O100O100O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O100O2O00000O10000O10000O10001O0O10000O10000O1000000O101O0O10000O1000000O10000O101O0O1000000O10000000000010O00000000001O000000000010O2N4L4L4L010O00001O00010O00001O0010O0001O000010O0001O001O01O01O00001O00010O001O000010O0N2O1O1N2O2N1N2O1O1N2O2N1N2O1SNm100O2N100O1O100O2O0O1O100O1O2O0O100O1O101N100O1O12N12N3M2N3M3N1N3M3M2N3M2NO2N1O1O1N2O1O1O1O1O2M2O1O1OO1O001O1O1N2O001O1O1O1O1O001bNmDnNT;m0b1K5L3L5K5L4K5K5LiT:"}, "label": "the vehicle on the right with doors"}]}
{"id": 436168, "image": "COCO_train2014_000000436168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing glasses sitting at table\"."}], "task": "refering", "answer_template": "The \"man wearing glasses sitting at table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 92, 93, 94, 108, 109, 110, 111, 123, 124, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 145, 146, 147, 148, 159, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 196, 197], "bbox": [0.23254166666666667, 0.17553125, 0.7227708333333334, 0.5143906250000001], "iscrowd": 0, "area": 21546.9071, "rle": {"size": [640, 480], "counts": "jVV25jc05J6J3N1O2O0O2O001O0O2O001O0O2O001O0O2O001N101O000O2O00000O2O00000O2O000O2O00000O2O00000O2O000O101O0O101O0O10001N101O0O101O0O2O001N10001N101O0O2O000O2O001N5L5K5J6K5K2M0100O1O010001O000O100000000O1000000hNWMYAi2c>fMQA[2k>lMRAT2j>RNTAn1g>ZNVAf1i>]NUAc1j>`NSAa1m>bNQA]1n>fNQAY1n>m1O100O1O1O101N1O100O10000O10000O10000O2O2N2M3O1N2N1O0001O000001O01O000000001O0001O0001O0`LgAY1Y>fNhAZ1X>eNjAZ1V>eNlAZ1U>cNmA]1T>aNnA_1T>]NnAb1T>[NmAe1U>WNnAm1o=PNSBT2j=hMXB^2e=^M]Bg2_=UMdBo2Y=nLhBT3X=hLkBW3_=^LbBa3P?O1O0O2N2M2O2M3N1N3M3L3M4L4L3N3L4L3N3N2M3N2N2M3N2M2O2N2M3N2M3O1O100O010O100O100O100O010O1O100O100M3M2N3M3M3M3M3L4M3M2N3M3M3L4M3MXlb2"}, "label": "man wearing glasses sitting at table"}]}
{"id": 436168, "image": "COCO_train2014_000000436168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing dark dress near a woman sitting at dining table\"."}], "task": "refering", "answer_template": "The \"a man wearing dark dress near a woman sitting at dining table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 92, 93, 94, 108, 109, 110, 111, 123, 124, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 145, 146, 147, 148, 159, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 196, 197], "bbox": [0.23254166666666667, 0.17553125, 0.7227708333333334, 0.5143906250000001], "iscrowd": 0, "area": 21546.9071, "rle": {"size": [640, 480], "counts": "jVV25jc05J6J3N1O2O0O2O001O0O2O001O0O2O001O0O2O001N101O000O2O00000O2O00000O2O000O2O00000O2O00000O2O000O101O0O101O0O10001N101O0O101O0O2O001N10001N101O0O2O000O2O001N5L5K5J6K5K2M0100O1O010001O000O100000000O1000000hNWMYAi2c>fMQA[2k>lMRAT2j>RNTAn1g>ZNVAf1i>]NUAc1j>`NSAa1m>bNQA]1n>fNQAY1n>m1O100O1O1O101N1O100O10000O10000O10000O2O2N2M3O1N2N1O0001O000001O01O000000001O0001O0001O0`LgAY1Y>fNhAZ1X>eNjAZ1V>eNlAZ1U>cNmA]1T>aNnA_1T>]NnAb1T>[NmAe1U>WNnAm1o=PNSBT2j=hMXB^2e=^M]Bg2_=UMdBo2Y=nLhBT3X=hLkBW3_=^LbBa3P?O1O0O2N2M2O2M3N1N3M3L3M4L4L3N3L4L3N3N2M3N2N2M3N2M2O2N2M3N2M3O1O100O010O100O100O100O010O1O100O100M3M2N3M3M3M3M3L4M3M2N3M3M3L4M3MXlb2"}, "label": "a man wearing dark dress near a woman sitting at dining table"}]}
{"id": 436168, "image": "COCO_train2014_000000436168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with with shirt getting a slice of pizza\"."}], "task": "refering", "answer_template": "The \"woman with with shirt getting a slice of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 249, 250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 353, 354, 355], "bbox": [0.5599166666666666, 0.21632812499999998, 1.0, 0.9015625], "iscrowd": 0, "area": 48622.05180000001, "rle": {"size": [640, 480], "counts": "i^X55gc07J6M1N2N1O2O1N1O2N1O2O0010O100O100O010O11O001O01OO100000000O10000000O0100000000O0100000O1O00100oNTOl^Om0o`0XOP_Oi0k`0]OS_Oc0h`0DW_O<c`0J\\_O7^`00`_O1Z`06d_OJX`09j_OGU`09k_OHU`06m_OJR`06o_OJn?7U@Gh?<Y@Dc?>`@A\\?b0e@^OW?d0k@[OR?h0o@XOn>i0SAXOi>j0YAVOc>m0]ATO_>n0cAQOZ>R1fAoNV>S1lAmNP>V1PBkNl=W1UBiNi=X1YBhNc=Z1^BgN_=9\\AZOW1=c;jNdD`1hNUOQ1a0_;SOaDX1ROQOP1d0X;[O`DP1\\OnNm0h0Q;E_Dd0FmNl0j0m:MXD=0kNl0l0j:5QD4:jNm0l0g::nC2>fNn0o0e::lC3a0cN5MFT1f;;lC2j0ZNN8EQ1e;=kC2j1dNdNm0f;>jC3k1fNcNi0f;[2eEPMcNf0f;[2fERMcNc0e;\\2fEUMcN`0e;\\2gEWM_N`0j;Y2eE[M]N?l;W2fEoNY:Q1eEROY:P1eEQOZ:P1dETOZ:l0eEVOY:k0eEXOZ:i0dEZOY:g0eE\\OZ:d0eE^OY:c0eEAY:`0dECZ:>eEDZ:<dEHY::eEHZ:8dELY:5fEOW:1gE3V:OhE5U:KiE9T:HiE=U:ChEc0U:^OgEg0W:YOeEn0W:SOfER1X:_4Ld0\\Od0[Od0]Od0\\Od0[O7J00O10O100O010O100O10O01O100O10O0100O0100O010O01000O010O100O101O0O100O2O00O001O0O2O0101O1N2O1N2N2N4L6J2N1O1O1O1O1O1fF]HU6e7jI_HR6b7nI`Ho5a7PJcHl5^7TJbHk5_7TJcHj5^7VJcHh5_7VJbHi5_7WJbHf5`7YJaHd5b7\\J_H`5d7_J^H]5e7bJ\\H\\5f7cJ[HZ5h7fJXHW5k7hJVHU5m7jJUHT5l7kJUHS5m7lJTHS5m7mJSHk2VOfLh8>SHh2^OdL`8`0WHg2BcLX8c0ZHe2GaLQ8f0]He2J_Lj7HYGEX1n3d0hKm6f0YG\\O\\1Q4k0aKd6Q1]IY3[9gLgFT3\\9mLdFn2`9RMaFi2g4"}, "label": "woman with with shirt getting a slice of pizza"}]}
{"id": 436168, "image": "COCO_train2014_000000436168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white whirt , cutting a slice of pizza from a box , with a pair of scissors\"."}], "task": "refering", "answer_template": "The \"a woman in a white whirt , cutting a slice of pizza from a box , with a pair of scissors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 249, 250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 353, 354, 355], "bbox": [0.5599166666666666, 0.21632812499999998, 1.0, 0.9015625], "iscrowd": 0, "area": 48622.05180000001, "rle": {"size": [640, 480], "counts": "i^X55gc07J6M1N2N1O2O1N1O2N1O2O0010O100O100O010O11O001O01OO100000000O10000000O0100000000O0100000O1O00100oNTOl^Om0o`0XOP_Oi0k`0]OS_Oc0h`0DW_O<c`0J\\_O7^`00`_O1Z`06d_OJX`09j_OGU`09k_OHU`06m_OJR`06o_OJn?7U@Gh?<Y@Dc?>`@A\\?b0e@^OW?d0k@[OR?h0o@XOn>i0SAXOi>j0YAVOc>m0]ATO_>n0cAQOZ>R1fAoNV>S1lAmNP>V1PBkNl=W1UBiNi=X1YBhNc=Z1^BgN_=9\\AZOW1=c;jNdD`1hNUOQ1a0_;SOaDX1ROQOP1d0X;[O`DP1\\OnNm0h0Q;E_Dd0FmNl0j0m:MXD=0kNl0l0j:5QD4:jNm0l0g::nC2>fNn0o0e::lC3a0cN5MFT1f;;lC2j0ZNN8EQ1e;=kC2j1dNdNm0f;>jC3k1fNcNi0f;[2eEPMcNf0f;[2fERMcNc0e;\\2fEUMcN`0e;\\2gEWM_N`0j;Y2eE[M]N?l;W2fEoNY:Q1eEROY:P1eEQOZ:P1dETOZ:l0eEVOY:k0eEXOZ:i0dEZOY:g0eE\\OZ:d0eE^OY:c0eEAY:`0dECZ:>eEDZ:<dEHY::eEHZ:8dELY:5fEOW:1gE3V:OhE5U:KiE9T:HiE=U:ChEc0U:^OgEg0W:YOeEn0W:SOfER1X:_4Ld0\\Od0[Od0]Od0\\Od0[O7J00O10O100O010O100O10O01O100O10O0100O0100O010O01000O010O100O101O0O100O2O00O001O0O2O0101O1N2O1N2N2N4L6J2N1O1O1O1O1O1fF]HU6e7jI_HR6b7nI`Ho5a7PJcHl5^7TJbHk5_7TJcHj5^7VJcHh5_7VJbHi5_7WJbHf5`7YJaHd5b7\\J_H`5d7_J^H]5e7bJ\\H\\5f7cJ[HZ5h7fJXHW5k7hJVHU5m7jJUHT5l7kJUHS5m7lJTHS5m7mJSHk2VOfLh8>SHh2^OdL`8`0WHg2BcLX8c0ZHe2GaLQ8f0]He2J_Lj7HYGEX1n3d0hKm6f0YG\\O\\1Q4k0aKd6Q1]IY3[9gLgFT3\\9mLdFn2`9RMaFi2g4"}, "label": "a woman in a white whirt , cutting a slice of pizza from a box , with a pair of scissors"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lower right of couch and black arm of chair\"."}], "task": "refering", "answer_template": "The \"lower right of couch and black arm of chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 160, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.6985, 0.6716799999999999, 1.0, 0.9828266666666666], "iscrowd": 0, "area": 12117.09285, "rle": {"size": [375, 500], "counts": "WUP45_;3M3N200O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1000000000000000000000000O10000000000I7H8I7K5M3M3N2M3M3VOQNjGQ2R8ZNgGf1W8]NgGd1W8_NhGa1X8`NgG`1Y8o000O1000000000000000000000000000000O1O100O1O100O10000O100O10000001O001O00001O001O00001O001O00001O001O00000000O100001O00001O00001O00001O00001O001O001O001O001O001O001O001O001O001O001O001RMZJ"}, "label": "lower right of couch and black arm of chair"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray couch\"."}], "task": "refering", "answer_template": "The \"a gray couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 160, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.6985, 0.6716799999999999, 1.0, 0.9828266666666666], "iscrowd": 0, "area": 12117.09285, "rle": {"size": [375, 500], "counts": "WUP45_;3M3N200O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1000000000000000000000000O10000000000I7H8I7K5M3M3N2M3M3VOQNjGQ2R8ZNgGf1W8]NgGd1W8_NhGa1X8`NgG`1Y8o000O1000000000000000000000000000000O1O100O1O100O10000O100O10000001O001O00001O001O00001O001O00001O001O00000000O100001O00001O00001O00001O00001O001O001O001O001O001O001O001O001O001O001O001RMZJ"}, "label": "a gray couch"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a grey shirt\"."}], "task": "refering", "answer_template": "The \"a girl in a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 58, 59, 60, 73, 74, 75, 76, 77, 78, 91, 92, 93, 108, 109, 126, 127], "bbox": [0.038380000000000004, 0.2168, 0.35158, 0.58], "iscrowd": 0, "area": 4381.962949999999, "rle": {"size": [375, 500], "counts": "ZU72`;5L4K5TOk0N3O1O100O1O10000O100O101OiNZF=d9BeF8X9IoF3m8NZGMb86cGGY8:kGDT8=QH_Om7a0e10O001O0010O01000000O0100000000000000001O0001O00001O00001O001O0010O01O00001M2O1N2NLJkD5R;0mDNS;:100O100O1O100O10000000001O0000O2O1O1O001O1O1O100O100O101N100O10001O0O2O1O001O1O001O010O010O010OO1O00O101N101N1O101N100O2O0O2O001O10O01O001000O01000O1000O11O000000001O000O1lNiEh0X:UOUF?l9_OcF3fkf3"}, "label": "a girl in a grey shirt"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the couch wearing the gray top and smiling\"."}], "task": "refering", "answer_template": "The \"the woman on the couch wearing the gray top and smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 58, 59, 60, 73, 74, 75, 76, 77, 78, 91, 92, 93, 108, 109, 126, 127], "bbox": [0.038380000000000004, 0.2168, 0.35158, 0.58], "iscrowd": 0, "area": 4381.962949999999, "rle": {"size": [375, 500], "counts": "ZU72`;5L4K5TOk0N3O1O100O1O10000O100O101OiNZF=d9BeF8X9IoF3m8NZGMb86cGGY8:kGDT8=QH_Om7a0e10O001O0010O01000000O0100000000000000001O0001O00001O00001O001O0010O01O00001M2O1N2NLJkD5R;0mDNS;:100O100O1O100O10000000001O0000O2O1O1O001O1O1O100O100O101N100O10001O0O2O1O001O1O001O010O010O010OO1O00O101N101N1O101N100O2O0O2O001O10O01O001000O01000O1000O11O000000001O000O1lNiEh0X:UOUF?l9_OcF3fkf3"}, "label": "the woman on the couch wearing the gray top and smiling"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in maroon shirt looking at camera\"."}], "task": "refering", "answer_template": "The \"man in maroon shirt looking at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 60, 61, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 127, 128, 129, 146, 147], "bbox": [0.0751, 0.2018933333333333, 0.44186000000000003, 0.6433333333333333], "iscrowd": 0, "area": 11751.108600000003, "rle": {"size": [375, 500], "counts": "jR>9];:E6K5K5K4K6K5K5K2O1O001O000O2O001O00001N101O001O00001O001O00001O001O001O00001O001O001O0O1N3M4L2N10D;1O00001O010O1001N10001O0O101O0O2O001N2O2N1O1O101O0O101N100O2O0O10001O000O2O00001N1000001O00001O0000O100O10000O1O1O1O1O100O1O100O100O1O100O100O1000000O1000000O1000000O10000kNdNoG\\1k7mNRHS1m7\\10O0IhGQMX8n2lGoLT8P3:0O01O1O010O10O0100O0O2O1N2M4M2N2N3N1N2O2M2O2M2O1O2M2O3M3L4M6I4M3VOeE3_:IdE3a:HbE6R;M3L4M3LdQV3"}, "label": "man in maroon shirt looking at camera"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young couple sitting on a couch with their friends playing video games\"."}], "task": "refering", "answer_template": "The \"a young couple sitting on a couch with their friends playing video games\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 60, 61, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 127, 128, 129, 146, 147], "bbox": [0.0751, 0.2018933333333333, 0.44186000000000003, 0.6433333333333333], "iscrowd": 0, "area": 11751.108600000003, "rle": {"size": [375, 500], "counts": "jR>9];:E6K5K5K4K6K5K5K2O1O001O000O2O001O00001N101O001O00001O001O00001O001O001O00001O001O001O0O1N3M4L2N10D;1O00001O010O1001N10001O0O101O0O2O001N2O2N1O1O101O0O101N100O2O0O10001O000O2O00001N1000001O00001O0000O100O10000O1O1O1O1O100O1O100O100O1O100O100O1000000O1000000O1000000O10000kNdNoG\\1k7mNRHS1m7\\10O0IhGQMX8n2lGoLT8P3:0O01O1O010O10O0100O0O2O1N2M4M2N2N3N1N2O2M2O2M2O1O2M2O3M3L4M6I4M3VOeE3_:IdE3a:HbE6R;M3L4M3LdQV3"}, "label": "a young couple sitting on a couch with their friends playing video games"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown leather couch that four people are seated on\"."}], "task": "refering", "answer_template": "The \"brown leather couch that four people are seated on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 55, 56, 57, 58, 59, 61, 62, 64, 65, 66, 67, 72, 73, 74, 75, 79, 80, 82, 83, 84, 85, 88, 91, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 206, 207, 208, 209, 210, 211, 226, 227, 228], "bbox": [0.038700000000000005, 0.21936000000000003, 0.95, 0.9806400000000001], "iscrowd": 0, "area": 30807.406049999994, "rle": {"size": [375, 500], "counts": "jR71c;5J6J6J6N2M3N2M2O1N3N1O1O1N3N0000O0O2O1N2N101N2N2000000000000000000O1000O1000000000000000O1000001O00000000000O10000010O00001O00001OL4M3M4M2O1O1O1O100O1O1O1O1O2O0O1O1O100O100O1O10M2L5L3O1O2O0O2J5K5100O010O10000001O00001O000O101O0000000000O1O1N2O1N2O1N2O1N2N2N2N2O1N2N2N2NZoc0=Uj[O24e0DAHR13SOZO\\1f0dNROc1n0^NQOb1o0^NQOb1o0_NQOa1n0`NQO`1P1`NoN`1Q1`NoNa1P1_NPOa1P1^NROa1n0_NROa1o0^NQOc1n0\\NSOd1m0\\NSOd1m0\\NSOe1l0[NTOe1l0ZNUOf1l0YNTOh1k0XNUOh1k0XNUOh1l0WNTOi1l0WNTOj1k0VNUOj1o0QNROP2Q1lMoNT2T1iMlNW2X1dMiN]2Y1`MgN`2[1]MfNd2V5001O00001O001O001O00001O001O00001O001O000000001O00000000001O1O1O1O1O1O1O1O1O1dKjL5V3KnL1S3NQMNP30TMMm22SMNn21SMNn21RMOo21QMNP31PM0o20RMOn22QMNP31QMNo22QMNP31QMNo23PMMP33QMLP33QMMn23RMMn23SMLn23SMLm25RMKo24RMKn25SMJm26TMIm26TMIl27UMIj28WMFj29WMFi2:XMEh2;YMDh2;YMDg2<YMDh2<VMEj2;VMFi2:WMFj29VMGj29VMGj29VMGk29TMGl29TMGl29TMGl29TMGl29TMGl29TMGl29TMGl29TMHl27TMIl27TMIl27TMIl27TMIl27TMIl27TMIl27SMJm26SMJm26SMJm27QMJo26QMKn25QMLo24QMLP33PMMP33oLNR31nLOT3OkL2U3NkL2T3OlLN^O_Le3c3lLMA`Lc3c3lLK[36dLI]38cLF^3;bLB`3?_L@b3a0^L]Oc3d0\\L[Oe3f0[LXOe3j0[LTOf3m0YLROh3o0XLoNi3R1VLmNk3T1ULjNl3W1TLhNl3Z1RLeNo3\\1QLcNn3_1QL`NP4a1PL]NQ4e1nKZNR4g1mKXNT4i1lKVNT4k1kKTNV4n1iKPNX4Q2hKoMV4S2iKnMV4S2jKlMU4W2jKiMT4Y2kKhMR4[2nKeMP4]2PLbMn3b2PL_Mm3d2SL\\Mk3f2TLZMj3i2VLWMe3o2ZLQMb3c5L5K5L3L5L4K5L301N2O0O2O1N102M2O1N201O0O10001O0O10001O0O101O000O10000O10000O100O101O0O10000O100O10000O100O1000OO1O001O01O01O001O01O0N2L4nNS1H7O21N3N1N2O1N3N1N2O2M2O1N3C=M4L3M4L3M3K7H;F9G8H7HZP9"}, "label": "brown leather couch that four people are seated on"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather couch on which 4 people sit\"."}], "task": "refering", "answer_template": "The \"a brown leather couch on which 4 people sit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 55, 56, 57, 58, 59, 61, 62, 64, 65, 66, 67, 72, 73, 74, 75, 79, 80, 82, 83, 84, 85, 88, 91, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 206, 207, 208, 209, 210, 211, 226, 227, 228], "bbox": [0.038700000000000005, 0.21936000000000003, 0.95, 0.9806400000000001], "iscrowd": 0, "area": 30807.406049999994, "rle": {"size": [375, 500], "counts": "jR71c;5J6J6J6N2M3N2M2O1N3N1O1O1N3N0000O0O2O1N2N101N2N2000000000000000000O1000O1000000000000000O1000001O00000000000O10000010O00001O00001OL4M3M4M2O1O1O1O100O1O1O1O1O2O0O1O1O100O100O1O10M2L5L3O1O2O0O2J5K5100O010O10000001O00001O000O101O0000000000O1O1N2O1N2O1N2O1N2N2N2N2O1N2N2N2NZoc0=Uj[O24e0DAHR13SOZO\\1f0dNROc1n0^NQOb1o0^NQOb1o0_NQOa1n0`NQO`1P1`NoN`1Q1`NoNa1P1_NPOa1P1^NROa1n0_NROa1o0^NQOc1n0\\NSOd1m0\\NSOd1m0\\NSOe1l0[NTOe1l0ZNUOf1l0YNTOh1k0XNUOh1k0XNUOh1l0WNTOi1l0WNTOj1k0VNUOj1o0QNROP2Q1lMoNT2T1iMlNW2X1dMiN]2Y1`MgN`2[1]MfNd2V5001O00001O001O001O00001O001O00001O001O000000001O00000000001O1O1O1O1O1O1O1O1O1dKjL5V3KnL1S3NQMNP30TMMm22SMNn21SMNn21RMOo21QMNP31PM0o20RMOn22QMNP31QMNo22QMNP31QMNo23PMMP33QMLP33QMMn23RMMn23SMLn23SMLm25RMKo24RMKn25SMJm26TMIm26TMIl27UMIj28WMFj29WMFi2:XMEh2;YMDh2;YMDg2<YMDh2<VMEj2;VMFi2:WMFj29VMGj29VMGj29VMGk29TMGl29TMGl29TMGl29TMGl29TMGl29TMGl29TMGl29TMHl27TMIl27TMIl27TMIl27TMIl27TMIl27TMIl27SMJm26SMJm26SMJm27QMJo26QMKn25QMLo24QMLP33PMMP33oLNR31nLOT3OkL2U3NkL2T3OlLN^O_Le3c3lLMA`Lc3c3lLK[36dLI]38cLF^3;bLB`3?_L@b3a0^L]Oc3d0\\L[Oe3f0[LXOe3j0[LTOf3m0YLROh3o0XLoNi3R1VLmNk3T1ULjNl3W1TLhNl3Z1RLeNo3\\1QLcNn3_1QL`NP4a1PL]NQ4e1nKZNR4g1mKXNT4i1lKVNT4k1kKTNV4n1iKPNX4Q2hKoMV4S2iKnMV4S2jKlMU4W2jKiMT4Y2kKhMR4[2nKeMP4]2PLbMn3b2PL_Mm3d2SL\\Mk3f2TLZMj3i2VLWMe3o2ZLQMb3c5L5K5L3L5L4K5L301N2O0O2O1N102M2O1N201O0O10001O0O10001O0O101O000O10000O10000O100O101O0O10000O100O10000O100O1000OO1O001O01O01O001O01O0N2L4nNS1H7O21N3N1N2O1N3N1N2O2M2O1N3C=M4L3M4L3M3K7H;F9G8H7HZP9"}, "label": "a brown leather couch on which 4 people sit"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green beanie sitting on a couch\"."}], "task": "refering", "answer_template": "The \"a man in a green beanie sitting on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 62, 63, 64, 79, 80, 81, 82, 83, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 185, 186, 187, 203, 204, 205], "bbox": [0.22956000000000001, 0.20304, 0.61914, 0.9316533333333333], "iscrowd": 0, "area": 21527.289700000005, "rle": {"size": [375, 500], "counts": "ZZZ18U;<J6N2M3M5L6I7I6J3N2M3M3M2O1N2O001O1O1O001O1O1O001N2O1O001O1O1O1O1O1O1O1O2N1O1O1N2O1YHTMa6m2[I_M\\6f2[IcM`6k2oHZMo6g301O001O01O01O00001O1O1O1O1O1O1O1L2N2O1001O000O101O0000001O0O11O0O1O2M2O2N1O2O0O1O100OoNjIdLV6P3WJoLh5i2cJTM]5j2hJTMX5h2mJWMR5g2SKWMl4f2YKYMg4b2`K\\M_4`2gK]M[4c2gKUM^4m2cKlLb4U3n12O1N2O1N3O000O100000000O10O1000O100YO_LXIa3e6cLZI]3c6hLZIY3d6jL[IV3b6oL\\IQ3a6TM]Il2a6WM^Ih2b6[M[If2d6\\M[Id2d6^M[Ib2d6aMZI_2f6bMYI^2g6dMVI]2i6Y10O10000000000O01000000000O1000O100001N2O001O1O1N4M4L4L3M3L5K4hMWH<l7_OXH>l7^OWH?k7^OZH\\Oi8b0\\GUOj8j0R1O1O101N100O100O101N101O0O2O0O2O0O2O0O3N3L4M3L3N3L4MY^U2"}, "label": "a man in a green beanie sitting on a couch"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a eyeglasses with a bunnet\"."}], "task": "refering", "answer_template": "The \"a man wearing a eyeglasses with a bunnet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 62, 63, 64, 79, 80, 81, 82, 83, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 185, 186, 187, 203, 204, 205], "bbox": [0.22956000000000001, 0.20304, 0.61914, 0.9316533333333333], "iscrowd": 0, "area": 21527.289700000005, "rle": {"size": [375, 500], "counts": "ZZZ18U;<J6N2M3M5L6I7I6J3N2M3M3M2O1N2O001O1O1O001O1O1O001N2O1O001O1O1O1O1O1O1O1O2N1O1O1N2O1YHTMa6m2[I_M\\6f2[IcM`6k2oHZMo6g301O001O01O01O00001O1O1O1O1O1O1O1L2N2O1001O000O101O0000001O0O11O0O1O2M2O2N1O2O0O1O100OoNjIdLV6P3WJoLh5i2cJTM]5j2hJTMX5h2mJWMR5g2SKWMl4f2YKYMg4b2`K\\M_4`2gK]M[4c2gKUM^4m2cKlLb4U3n12O1N2O1N3O000O100000000O10O1000O100YO_LXIa3e6cLZI]3c6hLZIY3d6jL[IV3b6oL\\IQ3a6TM]Il2a6WM^Ih2b6[M[If2d6\\M[Id2d6^M[Ib2d6aMZI_2f6bMYI^2g6dMVI]2i6Y10O10000000000O01000000000O1000O100001N2O001O1O1N4M4L4L3M3L5K4hMWH<l7_OXH>l7^OWH?k7^OZH\\Oi8b0\\GUOj8j0R1O1O101N100O100O101N101O0O2O0O2O0O2O0O3N3L4M3L3N3L4MY^U2"}, "label": "a man wearing a eyeglasses with a bunnet"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair next to the couch that has designs on people on it\"."}], "task": "refering", "answer_template": "The \"the chair next to the couch that has designs on people on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 54, 55, 56, 57, 72, 74, 75, 90, 91, 108], "bbox": [0.0, 0.18808, 0.22078, 0.48696000000000006], "iscrowd": 0, "area": 4697.220049999999, "rle": {"size": [375, 500], "counts": "j4l0Z:a0O001O0010O01O001O0O2O0O2N101N1O2O0O101N1N2N2NnNfFMW92nFLP93UGKh85]GIa88bGF\\8=eGBX8`0jG^OV8c0kG[OU8e0mGYOR8h0PHWOo7j0ZHlNf7T1^11O001N2O1000000000001O0000000000010O000000000000001O00000001O0001O000000000000001O0001O00000001O000000000000001O01O00000M3H8I7N2O1O1N2O1M3M3N2M3M4LQh^4"}, "label": "the chair next to the couch that has designs on people on it"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an afghan on the couch that is onnocupied\"."}], "task": "refering", "answer_template": "The \"an afghan on the couch that is onnocupied\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 54, 55, 56, 57, 72, 74, 75, 90, 91, 108], "bbox": [0.0, 0.18808, 0.22078, 0.48696000000000006], "iscrowd": 0, "area": 4697.220049999999, "rle": {"size": [375, 500], "counts": "j4l0Z:a0O001O0010O01O001O0O2O0O2N101N1O2O0O101N1N2N2NnNfFMW92nFLP93UGKh85]GIa88bGF\\8=eGBX8`0jG^OV8c0kG[OU8e0mGYOR8h0PHWOo7j0ZHlNf7T1^11O001N2O1000000000001O0000000000010O000000000000001O00000001O0001O000000000000001O0001O00000001O000000000000001O01O00000M3H8I7N2O1O1N2O1M3M3N2M3M4LQh^4"}, "label": "an afghan on the couch that is onnocupied"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair to the right of a red leather couch\"."}], "task": "refering", "answer_template": "The \"a black chair to the right of a red leather couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.69296, 0.6624, 0.9966400000000001, 0.9890399999999999], "iscrowd": 0, "area": 12497.501999999993, "rle": {"size": [375, 500], "counts": "j]o34`;3L4L4O1O1N2O1O1N2O1O1N2O1000000O10000O10000O10000O10001O0O10000O10000O10000O1000000N2L4M3L4L4L4M3L4L4L4L4M3UOk0N2O1N200O10000O100O10000O100O10000O100O10000O10000O100O11O001O00001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O000010O01O00001O001O00001O00001O001O00001O001O00001O0000k0UOU1kNm<"}, "label": "a black chair to the right of a red leather couch"}]}
{"id": 305105, "image": "COCO_train2014_000000305105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of the arm and seat of a gray stuffed armchair\"."}], "task": "refering", "answer_template": "The \"part of the arm and seat of a gray stuffed armchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.69296, 0.6624, 0.9966400000000001, 0.9890399999999999], "iscrowd": 0, "area": 12497.501999999993, "rle": {"size": [375, 500], "counts": "j]o34`;3L4L4O1O1N2O1O1N2O1O1N2O1000000O10000O10000O10000O10001O0O10000O10000O10000O1000000N2L4M3L4L4L4M3L4L4L4L4M3UOk0N2O1N200O10000O100O10000O100O10000O100O10000O10000O100O11O001O00001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O000010O01O00001O001O00001O00001O001O00001O001O00001O0000k0UOU1kNm<"}, "label": "part of the arm and seat of a gray stuffed armchair"}]}
{"id": 305106, "image": "COCO_train2014_000000305106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing next to the drum set\"."}], "task": "refering", "answer_template": "The \"a woman standing next to the drum set\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 75, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332], "bbox": [0.23453125, 0.1822843822843823, 0.47057812499999996, 1.0], "iscrowd": 0, "area": 38813.78689999999, "rle": {"size": [429, 640], "counts": "[Qo14T=`0kCAb:n0iD\\OS;`1M4L4L3YJjM5Z2BoM9V2_ORN?o1YOZNf0g1QObNm0`1kNhNT1X1dNQOZ1Q1^NWOa1j0VN_Oi1b0oMFo1<hMMW24aM4]2NeLP1Z3QOaLT1]3nN^LW1a3jNZL[1d3gNVL_1i3bNRLc1l3_NoKf1P4[NkKj1T4VNhKo1V4SNeKR2Y4b3N2N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2A?0000O10000O10000O100O10000O10000O10000O100O10000O10000O11O001O001O1O001O001O1O001O001O1O001O001O1O001O1XGTMm6l2RIVMm6k2QIWMn6i2QIYMn6k2mHWMR7m2iHUMV7o2eHRM[7S3_HoL`7U3[HmLd7W3WHkLh7Z3RHhLm7\\3nGfLQ8^3jGdLU8Z401O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1ULdJC]5hNgJeN;]2o4lNWKnM8Q3b4oNgLl0Z3SOlLg0T3XORMc0o2[OXM?i2_O]M<d2hNZJAX3b1_2iN^JCY3^1Z2lN`JD\\3[1U2mNdJE^3X1n1QOhJD`3V1i1ROlJF`3S1e1SOPKGb3P1_1VORKHe3m0Z1WOVKIg3j0T1ZO<a0E[Oc0`0Z6K5K4L^c]4"}, "label": "a woman standing next to the drum set"}]}
{"id": 305106, "image": "COCO_train2014_000000305106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl singing on stage by holding an umberalla on her right shoulder\"."}], "task": "refering", "answer_template": "The \"a girl singing on stage by holding an umberalla on her right shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 75, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332], "bbox": [0.23453125, 0.1822843822843823, 0.47057812499999996, 1.0], "iscrowd": 0, "area": 38813.78689999999, "rle": {"size": [429, 640], "counts": "[Qo14T=`0kCAb:n0iD\\OS;`1M4L4L3YJjM5Z2BoM9V2_ORN?o1YOZNf0g1QObNm0`1kNhNT1X1dNQOZ1Q1^NWOa1j0VN_Oi1b0oMFo1<hMMW24aM4]2NeLP1Z3QOaLT1]3nN^LW1a3jNZL[1d3gNVL_1i3bNRLc1l3_NoKf1P4[NkKj1T4VNhKo1V4SNeKR2Y4b3N2N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2A?0000O10000O10000O100O10000O10000O10000O100O10000O10000O11O001O001O1O001O001O1O001O001O1O001O001O1O001O1XGTMm6l2RIVMm6k2QIWMn6i2QIYMn6k2mHWMR7m2iHUMV7o2eHRM[7S3_HoL`7U3[HmLd7W3WHkLh7Z3RHhLm7\\3nGfLQ8^3jGdLU8Z401O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1ULdJC]5hNgJeN;]2o4lNWKnM8Q3b4oNgLl0Z3SOlLg0T3XORMc0o2[OXM?i2_O]M<d2hNZJAX3b1_2iN^JCY3^1Z2lN`JD\\3[1U2mNdJE^3X1n1QOhJD`3V1i1ROlJF`3S1e1SOPKGb3P1_1VORKHe3m0Z1WOVKIg3j0T1ZO<a0E[Oc0`0Z6K5K4L^c]4"}, "label": "a girl singing on stage by holding an umberalla on her right shoulder"}]}
{"id": 305106, "image": "COCO_train2014_000000305106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman smiling with umbrella to the right\"."}], "task": "refering", "answer_template": "The \"woman smiling with umbrella to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 83, 84, 85, 86, 106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.569390625, 0.17752913752913752, 0.801359375, 0.9887645687645686], "iscrowd": 0, "area": 38545.58825, "rle": {"size": [429, 640], "counts": "Zih46U=3M4L3N3L3M4L3ZOg0hNW1J6I6K5J6J6oNQ1G9mIfKP3]4aLXLU3l3[LdLa3`3ZKeMc4d5M3N2M3M3N200O1000000O1000000O1N2L4L4L4L4L4L4L4N2M3N2M3N2N2M3N2M3N2001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O000YFTMm8l2eFeMX9Z3N3M2N3M2RMo2M2O0O2O000O2O0O2K4_NfHmL[7R3QIbLQ7\\3\\IWLe6h3b1O0O1O2F:D>B>D<H9H7I7H=D`0@`0_Ojod1"}, "label": "woman smiling with umbrella to the right"}]}
{"id": 305106, "image": "COCO_train2014_000000305106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl on the right singing with umbrella\"."}], "task": "refering", "answer_template": "The \"a girl on the right singing with umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 83, 84, 85, 86, 106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.569390625, 0.17752913752913752, 0.801359375, 0.9887645687645686], "iscrowd": 0, "area": 38545.58825, "rle": {"size": [429, 640], "counts": "Zih46U=3M4L3N3L3M4L3ZOg0hNW1J6I6K5J6J6oNQ1G9mIfKP3]4aLXLU3l3[LdLa3`3ZKeMc4d5M3N2M3M3N200O1000000O1000000O1N2L4L4L4L4L4L4L4N2M3N2M3N2N2M3N2M3N2001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O000YFTMm8l2eFeMX9Z3N3M2N3M2RMo2M2O0O2O000O2O0O2K4_NfHmL[7R3QIbLQ7\\3\\IWLe6h3b1O0O1O2F:D>B>D<H9H7I7H=D`0@`0_Ojod1"}, "label": "a girl on the right singing with umbrella"}]}
{"id": 305106, "image": "COCO_train2014_000000305106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far right umbrella\"."}], "task": "refering", "answer_template": "The \"the far right umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 58, 59, 60, 61, 62, 63, 64, 65, 82, 83, 86, 87, 88, 89, 106, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 156, 157, 158, 159, 160, 180, 181], "bbox": [0.551578125, 0.09999999999999999, 0.969390625, 0.49013986013986016], "iscrowd": 0, "area": 17019.8996, "rle": {"size": [429, 640], "counts": "knc42[=0O2O1O1N101N2O1O0O2O1N2O001N2O1N101O1N2O0O2O1O1N101N2O1O0O4M3L5L3M4K4M4K4M1OO0100O10000O100O10000O100ON3L4K4L5L3L5K5L3L5K4M4K50O100O101O0O100O2O000O101N10000O101O0000001O00000O2O0000001O000000001O0000001O0000001O0001O00010O000010O0000010O00011N5K6K4K6J5L5J3M2O1N1O2O1N2N1O2O1N1O2O1N2N101N2N101N2N2oElLf9T3UFSMh9\\3M2N3N2M2N3N0O1N1O1O2N1O1O2NQLQGb3n8]LTGc3k8\\LXGd3f8[L\\Ge3c8ZL`Ge3_8ZLcGg3[8XLhGg3W8XLlGh3R8XLoGh3P8WLSHh3l7WLVHj3h7UL[Hj3e8O010O0010O01O10O010O01O100O00100O1O010O1O10O0100O1O010O1O10O01O100O1O2O0O2N2M2N3M3N1N3M2O2M3M2O2M3M2N3N1N3M3N1N3M2O2M3M2N3N2M2O2N1O2N2N1O2M3N1O2N1O2N2N1O2N2N1O2N1O2N2M2O2N2N1O2N1OXV8"}, "label": "the far right umbrella"}]}
{"id": 305106, "image": "COCO_train2014_000000305106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella that the woman on the right is holding\"."}], "task": "refering", "answer_template": "The \"the umbrella that the woman on the right is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 58, 59, 60, 61, 62, 63, 64, 65, 82, 83, 86, 87, 88, 89, 106, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 156, 157, 158, 159, 160, 180, 181], "bbox": [0.551578125, 0.09999999999999999, 0.969390625, 0.49013986013986016], "iscrowd": 0, "area": 17019.8996, "rle": {"size": [429, 640], "counts": "knc42[=0O2O1O1N101N2O1O0O2O1N2O001N2O1N101O1N2O0O2O1O1N101N2O1O0O4M3L5L3M4K4M4K4M1OO0100O10000O100O10000O100ON3L4K4L5L3L5K5L3L5K4M4K50O100O101O0O100O2O000O101N10000O101O0000001O00000O2O0000001O000000001O0000001O0000001O0001O00010O000010O0000010O00011N5K6K4K6J5L5J3M2O1N1O2O1N2N1O2O1N1O2O1N2N101N2N101N2N2oElLf9T3UFSMh9\\3M2N3N2M2N3N0O1N1O1O2N1O1O2NQLQGb3n8]LTGc3k8\\LXGd3f8[L\\Ge3c8ZL`Ge3_8ZLcGg3[8XLhGg3W8XLlGh3R8XLoGh3P8WLSHh3l7WLVHj3h7UL[Hj3e8O010O0010O01O10O010O01O100O00100O1O010O1O10O0100O1O010O1O10O01O100O1O2O0O2N2M2N3M3N1N3M2O2M3M2O2M3M2N3N1N3M3N1N3M2O2M3M2N3N2M2O2N1O2N2N1O2M3N1O2N1O2N2N1O2N2N1O2N1O2N2M2O2N2N1O2N1OXV8"}, "label": "the umbrella that the woman on the right is holding"}]}
{"id": 395221, "image": "COCO_train2014_000000395221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange with a leaf in white bowl\"."}], "task": "refering", "answer_template": "The \"an orange with a leaf in white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 213], "bbox": [0.15390625, 0.33, 0.36403125, 0.6063466042154566], "iscrowd": 0, "area": 12999.307749999996, "rle": {"size": [427, 640], "counts": "o^Y1e0c<5K6K5J5K6J5L3L5K4N3M2M4M2N3M3M2N1N3N1O2N1O2N1N3N101N101O001N101O0O2O001N101O0O2O001O00001O00001O00001O0000001O00000000000000000000000000001O0000000000000001O00000000001O000000000O2O0O2O000O2N100O2N1O2O0O1O2O0O2N1O101N101N101N1O2O0O2N101N2N2N2N2N2N3N1N2N5J5L4L4L4Lm0RO^`Y5"}, "label": "an orange with a leaf in white bowl"}]}
{"id": 395221, "image": "COCO_train2014_000000395221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fruit in the white bowl\"."}], "task": "refering", "answer_template": "The \"the fruit in the white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 213], "bbox": [0.15390625, 0.33, 0.36403125, 0.6063466042154566], "iscrowd": 0, "area": 12999.307749999996, "rle": {"size": [427, 640], "counts": "o^Y1e0c<5K6K5J5K6J5L3L5K4N3M2M4M2N3M3M2N1N3N1O2N1O2N1N3N101N101O001N101O0O2O001N101O0O2O001O00001O00001O00001O0000001O00000000000000000000000000001O0000000000000001O00000000001O000000000O2O0O2O000O2N100O2N1O2O0O1O2O0O2N1O101N101N101N1O2O0O2N101N2N2N2N2N2N3N1N2N5J5L4L4L4Lm0RO^`Y5"}, "label": "the fruit in the white bowl"}]}
{"id": 165847, "image": "COCO_train2014_000000165847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black , cable - knit sweater\"."}], "task": "refering", "answer_template": "The \"a woman in a black , cable - knit sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 211, 212, 213, 228, 229, 230, 244, 245, 246, 247, 261, 262, 263, 264, 279, 280, 281, 296, 297, 298, 313, 314, 315, 330, 331, 332, 347, 348, 349, 363, 364, 365, 366], "bbox": [0.35722916666666665, 0.499328125, 0.5818541666666666, 0.9455625], "iscrowd": 0, "area": 14832.858800000002, "rle": {"size": [640, 480], "counts": "RY[32kc07J20O01000O01O100O100O100O1000000000OfACS9=mFDR9<mFFR9:nFGQ99nFHR98nFIQ97nFJR96nFJR96nFKQ95oFKQ95oFKQ95oFKQ95oFKP96PGJP96mFMS93gF3Y9MaF9_9GaEZ1^:fN[Ea1d:`NTEh1l:XNmDo1S;QNeDW2Z;jM_D]2a;cMWDe2h;\\MQDl2h9dL[F`0IP3i9hLXF9LR3k9SMhD[Ob0b0h0S3n9bMUE^Oj0T3o9aMPE@m0S3R:aMhDBS1P3T:aMbDDV1o2W:iNeE[1Z:fNmDYMOU4S;cNkD`MIQ4[;aNfDX2Z;W3O100000N2O0gMjDaK_;Z4oDnJ`;P5QERJ^;m5[1O1N2O2M3MPD\\Jh9b5XFhJa9U5_FRK]9k4bF_KW9o4YFXKc9Z5eElJ[:W5[EoJf:T5PESKQ;m61O1O2N1O1O2N1O2N1O3M2M4THQD`7[<L3N2L3M2\\MiC]MZ<`2SDPMR<o2[DYLR<d3XDhKU<U4P2F9H9G8H9F`0A8hN^^OJia0L^^OYO44ea0:T_O^OUhm3"}, "label": "a woman in a black , cable - knit sweater"}]}
{"id": 165847, "image": "COCO_train2014_000000165847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the middle that is holding a dog leash\"."}], "task": "refering", "answer_template": "The \"the woman in the middle that is holding a dog leash\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 211, 212, 213, 228, 229, 230, 244, 245, 246, 247, 261, 262, 263, 264, 279, 280, 281, 296, 297, 298, 313, 314, 315, 330, 331, 332, 347, 348, 349, 363, 364, 365, 366], "bbox": [0.35722916666666665, 0.499328125, 0.5818541666666666, 0.9455625], "iscrowd": 0, "area": 14832.858800000002, "rle": {"size": [640, 480], "counts": "RY[32kc07J20O01000O01O100O100O100O1000000000OfACS9=mFDR9<mFFR9:nFGQ99nFHR98nFIQ97nFJR96nFJR96nFKQ95oFKQ95oFKQ95oFKQ95oFKP96PGJP96mFMS93gF3Y9MaF9_9GaEZ1^:fN[Ea1d:`NTEh1l:XNmDo1S;QNeDW2Z;jM_D]2a;cMWDe2h;\\MQDl2h9dL[F`0IP3i9hLXF9LR3k9SMhD[Ob0b0h0S3n9bMUE^Oj0T3o9aMPE@m0S3R:aMhDBS1P3T:aMbDDV1o2W:iNeE[1Z:fNmDYMOU4S;cNkD`MIQ4[;aNfDX2Z;W3O100000N2O0gMjDaK_;Z4oDnJ`;P5QERJ^;m5[1O1N2O2M3MPD\\Jh9b5XFhJa9U5_FRK]9k4bF_KW9o4YFXKc9Z5eElJ[:W5[EoJf:T5PESKQ;m61O1O2N1O1O2N1O2N1O3M2M4THQD`7[<L3N2L3M2\\MiC]MZ<`2SDPMR<o2[DYLR<d3XDhKU<U4P2F9H9G8H9F`0A8hN^^OJia0L^^OYO44ea0:T_O^OUhm3"}, "label": "the woman in the middle that is holding a dog leash"}]}
{"id": 165847, "image": "COCO_train2014_000000165847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl wearing flip flop shoes\"."}], "task": "refering", "answer_template": "The \"the girl wearing flip flop shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 215, 216, 217, 232, 233, 234, 249, 250, 251, 265, 266, 267, 268, 283, 284, 285, 300, 301, 302, 317, 318, 319, 334, 335, 336, 351, 352, 353, 368, 369, 370, 385, 386, 387], "bbox": [0.6248541666666667, 0.475484375, 0.8085625000000001, 0.9966406250000001], "iscrowd": 0, "area": 16685.366749999997, "rle": {"size": [640, 480], "counts": "SSl51oc0001O0O2O0dJNjF3T9OjF2U91iF0T94jFLU95jFLU96jFJT99kFGT9:lFGR9;lFFS9<lFDR9>nFBQ9`0_BXOn39b9a0ZBAn3Ng9b0VBLl3Bl9f1mE[NR:m1eEUNZ:R2^EPN_:Z2YEhMc:a2UEaMh:i2PEXM^8]O`Gd3JQMb8JVG_3POULLc0i95RG[3POXLK:n98TG\\3mN]LH0S:;UG\\3kN`Mn9WOSG^3jN^MS:TOPGm5o8TJmFo5S9QJjFR6V9nIgFU6Y9lIcFW6\\9jIaFY6_9hI^FZ6b9fI[F]6e9cIXF`6h9aIUFa61UHY9Z1fFb6MWH\\9Y1eF_60YH[9X1dF^63[HX9X1cF]66[HW9X1cF[67^HV9W1bFZ6:_HT9X1`FX6=aHR9X1`FV6`0bHQ9X1^FU6a0dHR9V1\\Fj5o0PIe8V8hGjGY8h7SHYHn7Y7_HgHb7R7cHPI\\7o6eHQI\\7m6dHTI]7T7ZHlHg7]7nGcHS8c7eG_H\\8e7VGdHk8a7gFgHZ9]7YFkHg9g81O1N2eFoEe8V:UGkEk8_:jFcEU9e:O1O1N2fJbEP1^:SNeFg1\\9WNlFb1U9\\NSG]1n8aN[GW1f8gNdGP1]8nNmGh0U8VOUH=o7B[H1j7KcHFc72nH@\\73UI_OW73YI_OT71]IAo61dZi1"}, "label": "the girl wearing flip flop shoes"}]}
{"id": 165847, "image": "COCO_train2014_000000165847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with the glasses\"."}], "task": "refering", "answer_template": "The \"the girl with the glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 215, 216, 217, 232, 233, 234, 249, 250, 251, 265, 266, 267, 268, 283, 284, 285, 300, 301, 302, 317, 318, 319, 334, 335, 336, 351, 352, 353, 368, 369, 370, 385, 386, 387], "bbox": [0.6248541666666667, 0.475484375, 0.8085625000000001, 0.9966406250000001], "iscrowd": 0, "area": 16685.366749999997, "rle": {"size": [640, 480], "counts": "SSl51oc0001O0O2O0dJNjF3T9OjF2U91iF0T94jFLU95jFLU96jFJT99kFGT9:lFGR9;lFFS9<lFDR9>nFBQ9`0_BXOn39b9a0ZBAn3Ng9b0VBLl3Bl9f1mE[NR:m1eEUNZ:R2^EPN_:Z2YEhMc:a2UEaMh:i2PEXM^8]O`Gd3JQMb8JVG_3POULLc0i95RG[3POXLK:n98TG\\3mN]LH0S:;UG\\3kN`Mn9WOSG^3jN^MS:TOPGm5o8TJmFo5S9QJjFR6V9nIgFU6Y9lIcFW6\\9jIaFY6_9hI^FZ6b9fI[F]6e9cIXF`6h9aIUFa61UHY9Z1fFb6MWH\\9Y1eF_60YH[9X1dF^63[HX9X1cF]66[HW9X1cF[67^HV9W1bFZ6:_HT9X1`FX6=aHR9X1`FV6`0bHQ9X1^FU6a0dHR9V1\\Fj5o0PIe8V8hGjGY8h7SHYHn7Y7_HgHb7R7cHPI\\7o6eHQI\\7m6dHTI]7T7ZHlHg7]7nGcHS8c7eG_H\\8e7VGdHk8a7gFgHZ9]7YFkHg9g81O1N2eFoEe8V:UGkEk8_:jFcEU9e:O1O1N2fJbEP1^:SNeFg1\\9WNlFb1U9\\NSG]1n8aN[GW1f8gNdGP1]8nNmGh0U8VOUH=o7B[H1j7KcHFc72nH@\\73UI_OW73YI_OT71]IAo61dZi1"}, "label": "the girl with the glasses"}]}
{"id": 526301, "image": "COCO_train2014_000000526301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tan / gold car that is located directly behind and slightly to the right of the parking meter . it is next to both the red car and the van\"."}], "task": "refering", "answer_template": "The \"the tan / gold car that is located directly behind and slightly to the right of the parking meter . it is next to both the red car and the van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 88, 89, 107, 125, 143], "bbox": [0.82394, 0.22202666666666668, 1.0, 0.6163733333333333], "iscrowd": 0, "area": 3776.0694000000008, "rle": {"size": [375, 500], "counts": "gnf43d;2O0O1O00100O1O1O1O010O1O1O1O10O01O1O100O1O1O00100O1O1O1O010O1O1O1O1O1O101N1O2N100O2N101N1O2O0O101N1O2O0O1O2O0O2O0O2N2O2M2O2N1O2N1N3N1O2N1N3N1O2N1N3N1O2N1O2M2O2N7I8G9HO1O1O1eNbM"}, "label": "the tan / gold car that is located directly behind and slightly to the right of the parking meter . it is next to both the red car and the van"}]}
{"id": 526301, "image": "COCO_train2014_000000526301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rear of tan car\"."}], "task": "refering", "answer_template": "The \"rear of tan car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 88, 89, 107, 125, 143], "bbox": [0.82394, 0.22202666666666668, 1.0, 0.6163733333333333], "iscrowd": 0, "area": 3776.0694000000008, "rle": {"size": [375, 500], "counts": "gnf43d;2O0O1O00100O1O1O1O010O1O1O1O10O01O1O100O1O1O00100O1O1O1O010O1O1O1O1O1O101N1O2N100O2N101N1O2O0O101N1O2O0O1O2O0O2O0O2N2O2M2O2N1O2N1N3N1O2N1N3N1O2N1N3N1O2N1O2M2O2N7I8G9HO1O1O1eNbM"}, "label": "rear of tan car"}]}
{"id": 526301, "image": "COCO_train2014_000000526301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red four door car parked by a meter\"."}], "task": "refering", "answer_template": "The \"a red four door car parked by a meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 126, 127, 128, 129, 144, 145, 146, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 218, 219], "bbox": [0.0016799999999999999, 0.2827733333333334, 0.33876, 0.9884000000000001], "iscrowd": 0, "area": 24839.975499999997, "rle": {"size": [375, 500], "counts": "l?^2X9_3bL:E0010O010O0010O00010O01O01O010O2O1N101N2N101N2O0O2O1N1O2O1N101O1N101O1O001O1O001O1O001O1O001O1O001O1N101O1O001O1O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O^L`LmN`3m0nLlNQ3m0^MlNa2P1jMiNV2V1SNcNl1[1^N^Na1a1hNWNW1i1ROPNm0o1[OkMd0T2AhM?V2GfM9Y2LcM4\\21`MO_26]MJb2;ZMEd2a0XM_Og2f0UMZOj2k0RMUOm2o0PMQOo2T1mLlNQ3Z1kLfNT3o4O1O1O0O2O1O1O1O1N2O1O1O1N2N2O1N2N2N2N2N2N1O2N2N2N2O1N2N2N2N2N2N2N2N2O1N2O1N2N101N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O0O2N2O1N2O1N2N2O1NeSi3"}, "label": "a red four door car parked by a meter"}]}
{"id": 526301, "image": "COCO_train2014_000000526301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a parked red 4 door car\"."}], "task": "refering", "answer_template": "The \"a parked red 4 door car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 126, 127, 128, 129, 144, 145, 146, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 218, 219], "bbox": [0.0016799999999999999, 0.2827733333333334, 0.33876, 0.9884000000000001], "iscrowd": 0, "area": 24839.975499999997, "rle": {"size": [375, 500], "counts": "l?^2X9_3bL:E0010O010O0010O00010O01O01O010O2O1N101N2N101N2O0O2O1N1O2O1N101O1N101O1O001O1O001O1O001O1O001O1O001O1N101O1O001O1O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O^L`LmN`3m0nLlNQ3m0^MlNa2P1jMiNV2V1SNcNl1[1^N^Na1a1hNWNW1i1ROPNm0o1[OkMd0T2AhM?V2GfM9Y2LcM4\\21`MO_26]MJb2;ZMEd2a0XM_Og2f0UMZOj2k0RMUOm2o0PMQOo2T1mLlNQ3Z1kLfNT3o4O1O1O0O2O1O1O1O1N2O1O1O1N2N2O1N2N2N2N2N2N1O2N2N2N2O1N2N2N2N2N2N2N2N2O1N2O1N2N101N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O0O2N2O1N2O1N2N2O1NeSi3"}, "label": "a parked red 4 door car"}]}
{"id": 526301, "image": "COCO_train2014_000000526301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue sign and a mini van\"."}], "task": "refering", "answer_template": "The \"a blue sign and a mini van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 48, 49, 50, 51, 52, 53, 69, 70, 71, 89], "bbox": [0.6935, 0.08008, 1.0, 0.32165333333333335], "iscrowd": 0, "area": 4581.3944, "rle": {"size": [375, 500], "counts": "RTo35[;8H7I7O10001O0000001O0000001O0000001O000000001O0O10001O0000001O1O001O001O001O001O1O001O001O0S1nNO00100O001O001O001O001O00100O001O001O001O00001O010O00001O001O00001O0010O01O00001O001O00001O010O00001O001O00001O0010O01O00001O001O00001O0010O0001O001O00001O001O010O00001O001O00001O0010O0001O001O00001O001O010O000eM"}, "label": "a blue sign and a mini van"}]}
{"id": 526301, "image": "COCO_train2014_000000526301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gray minivan next to blue sign\"."}], "task": "refering", "answer_template": "The \"gray minivan next to blue sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 48, 49, 50, 51, 52, 53, 69, 70, 71, 89], "bbox": [0.6935, 0.08008, 1.0, 0.32165333333333335], "iscrowd": 0, "area": 4581.3944, "rle": {"size": [375, 500], "counts": "RTo35[;8H7I7O10001O0000001O0000001O0000001O000000001O0O10001O0000001O1O001O001O001O001O1O001O001O0S1nNO00100O001O001O001O001O00100O001O001O001O00001O010O00001O001O00001O0010O01O00001O001O00001O010O00001O001O00001O0010O01O00001O001O00001O0010O0001O001O00001O001O010O00001O001O00001O0010O0001O001O00001O001O010O000eM"}, "label": "gray minivan next to blue sign"}]}
{"id": 108510, "image": "COCO_train2014_000000108510.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long train with a yellow engine in front\"."}], "task": "refering", "answer_template": "The \"a long train with a yellow engine in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 234, 235, 236], "bbox": [0.176234375, 0.3572960372960373, 0.66428125, 0.7101165501165501], "iscrowd": 0, "area": 14284.5704, "rle": {"size": [429, 640], "counts": "Td_12Y=2M4M2M3N2M4M2N2N2N3M2N2N2K5C>E:O1O1N3N1N2O1O2O00000O2O00000O2OO100000O1000O10O100000O0100000000O0100000O1000O10O100000O10O100000O001O100O001O1O010O1O0010O01O1O010O1O001O10O01O00100O001O10O01O1O00100O001O010O1O001O10O01O1O010O1O001O010O1O00100O001O1O010O1O010O010O10O0100O01000O010O10O010O0100O010O10O0100O010O10O010O0100O010O10O10O10O010O0100O100O010O100O100O010O100O100O01000O100O10O0100O100O10O010000O100O010O100O100O010O10000O10O0100O100O10O10O100O10O010000O100O010O10000O10O0100O100O10O10O100O100O01000O100O010O10O10O10O0100O010O10O10O10O0100O01000O010O10O0100O01000O010O10O0100O10O10O10O0100O010O10001N100O100O10000Obmi2"}, "label": "a long train with a yellow engine in front"}]}
{"id": 108510, "image": "COCO_train2014_000000108510.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train with a yellow engine is on the tracks\"."}], "task": "refering", "answer_template": "The \"a train with a yellow engine is on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 234, 235, 236], "bbox": [0.176234375, 0.3572960372960373, 0.66428125, 0.7101165501165501], "iscrowd": 0, "area": 14284.5704, "rle": {"size": [429, 640], "counts": "Td_12Y=2M4M2M3N2M4M2N2N2N3M2N2N2K5C>E:O1O1N3N1N2O1O2O00000O2O00000O2OO100000O1000O10O100000O0100000000O0100000O1000O10O100000O10O100000O001O100O001O1O010O1O0010O01O1O010O1O001O10O01O00100O001O10O01O1O00100O001O010O1O001O10O01O1O010O1O001O010O1O00100O001O1O010O1O010O010O10O0100O01000O010O10O010O0100O010O10O0100O010O10O010O0100O010O10O10O10O010O0100O100O010O100O100O010O100O100O01000O100O10O0100O100O10O010000O100O010O100O100O010O10000O10O0100O100O10O10O100O10O010000O100O010O10000O10O0100O100O10O10O100O100O01000O100O010O10O10O10O0100O010O10O10O10O0100O01000O010O10O0100O01000O010O10O0100O10O10O10O0100O010O10001N100O100O10000Obmi2"}, "label": "a train with a yellow engine is on the tracks"}]}
{"id": 567268, "image": "COCO_train2014_000000567268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman smaling\"."}], "task": "refering", "answer_template": "The \"a woman smaling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 249, 250, 251, 268], "bbox": [0.28518828451882844, 0.15246875, 0.792531380753138, 0.66815625], "iscrowd": 0, "area": 40809.39104999999, "rle": {"size": [640, 478], "counts": "WYe21nc06J6J6J5K6J4L10O01O1O1O1O1O10O01O1O1O1O1O001O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100N2N2M4M2M3N3M2O1O2N1O1O2O0O2N1O1O2N1O1O2N1O1O2K4E;E<nMQ2O2M2O1O2M2O1N3N1O1N3N1O110N100O2N1O2O000O2O0O2O001N101N101O1O001O001N101O001O001O001O001O0O2O001O001O1O001O001O001O1O001O001O1O001O0010O01O1O0010O0100O0010O010O01000O2O0O101O0000001N1oJ[Df1e;SNgDh1Z;PNlDQ2T;hMREY2n:`MXE`2j:XM]Eh2c:UM_El2a:RM`Eo2a:mLbES3]:lLdET3]:jLdEW3\\:gLeEZ3Z:eLhE[3X:cLiE^3W:_LkEa3U:^LlEc3T:[LmEf3S:WLoEj3Q:SLQFn3n9QLSFP4m9mKVFR4k9kKWFV4h9hKZFY4f9eK[F\\4e9aK]F`4b9^K`Fb4a9TKhFm4U<100O1O001O1O1O100O001O1O1O1O010O1O1O1O001O100O1O001O1O1O100O001O1ON2H9Aa0jK]Ad2n>bLjAT3h?G9G:F`0@\\bm1"}, "label": "a woman smaling"}]}
{"id": 567268, "image": "COCO_train2014_000000567268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady wearing glasses\"."}], "task": "refering", "answer_template": "The \"lady wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 249, 250, 251, 268], "bbox": [0.28518828451882844, 0.15246875, 0.792531380753138, 0.66815625], "iscrowd": 0, "area": 40809.39104999999, "rle": {"size": [640, 478], "counts": "WYe21nc06J6J6J5K6J4L10O01O1O1O1O1O10O01O1O1O1O1O001O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100N2N2M4M2M3N3M2O1O2N1O1O2O0O2N1O1O2N1O1O2N1O1O2K4E;E<nMQ2O2M2O1O2M2O1N3N1O1N3N1O110N100O2N1O2O000O2O0O2O001N101N101O1O001O001N101O001O001O001O001O0O2O001O001O1O001O001O001O1O001O001O1O001O0010O01O1O0010O0100O0010O010O01000O2O0O101O0000001N1oJ[Df1e;SNgDh1Z;PNlDQ2T;hMREY2n:`MXE`2j:XM]Eh2c:UM_El2a:RM`Eo2a:mLbES3]:lLdET3]:jLdEW3\\:gLeEZ3Z:eLhE[3X:cLiE^3W:_LkEa3U:^LlEc3T:[LmEf3S:WLoEj3Q:SLQFn3n9QLSFP4m9mKVFR4k9kKWFV4h9hKZFY4f9eK[F\\4e9aK]F`4b9^K`Fb4a9TKhFm4U<100O1O001O1O1O100O001O1O1O1O010O1O1O1O001O100O1O001O1O1O100O001O1ON2H9Aa0jK]Ad2n>bLjAT3h?G9G:F`0@\\bm1"}, "label": "lady wearing glasses"}]}
{"id": 567268, "image": "COCO_train2014_000000567268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman\"."}], "task": "refering", "answer_template": "The \"a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 249, 250, 251, 268], "bbox": [0.28518828451882844, 0.15246875, 0.792531380753138, 0.66815625], "iscrowd": 0, "area": 40809.39104999999, "rle": {"size": [640, 478], "counts": "WYe21nc06J6J6J5K6J4L10O01O1O1O1O1O10O01O1O1O1O1O001O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100N2N2M4M2M3N3M2O1O2N1O1O2O0O2N1O1O2N1O1O2N1O1O2K4E;E<nMQ2O2M2O1O2M2O1N3N1O1N3N1O110N100O2N1O2O000O2O0O2O001N101N101O1O001O001N101O001O001O001O001O0O2O001O001O1O001O001O001O1O001O001O1O001O0010O01O1O0010O0100O0010O010O01000O2O0O101O0000001N1oJ[Df1e;SNgDh1Z;PNlDQ2T;hMREY2n:`MXE`2j:XM]Eh2c:UM_El2a:RM`Eo2a:mLbES3]:lLdET3]:jLdEW3\\:gLeEZ3Z:eLhE[3X:cLiE^3W:_LkEa3U:^LlEc3T:[LmEf3S:WLoEj3Q:SLQFn3n9QLSFP4m9mKVFR4k9kKWFV4h9hKZFY4f9eK[F\\4e9aK]F`4b9^K`Fb4a9TKhFm4U<100O1O001O1O1O100O001O1O1O1O010O1O1O1O001O100O1O001O1O1O100O001O1ON2H9Aa0jK]Ad2n>bLjAT3h?G9G:F`0@\\bm1"}, "label": "a woman"}]}
{"id": 264165, "image": "COCO_train2014_000000264165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the doctor is covering a small child with a sheet\"."}], "task": "refering", "answer_template": "The \"the doctor is covering a small child with a sheet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 44, 54, 55, 56, 57, 58, 59, 60, 61, 62, 70, 71, 72, 73, 74, 75, 76, 77, 78, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 100, 101, 102, 103, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 131, 132, 139, 140, 141, 142, 143, 146, 147, 156, 157, 158, 161, 162, 163, 172, 173, 174, 177, 178, 187, 188, 202, 203, 204, 218, 219, 234], "bbox": [0.07729545454545454, 0.090953125, 0.9609772727272727, 0.617328125], "iscrowd": 0, "area": 43981.572, "rle": {"size": [640, 440], "counts": "Yae02nc04K5K5L2M2N2O1N2O1N2N100O1O100O1O100O101N1O100O1O100O01_OX]OHib09X]OEhb0<Y]OAib0?X]O_Ohb0b0X]O\\Oib0e0X]OXOjb0h0600100O1N2N2O1N1O2O1N2N2O1N2N101N2O1N2N2O4K5K5L3L0001D]^O[Nca0\\1i^OaNWa0W1m0M200100O1O100O1000000O100000O10O100000000O100000000O10O100000O100000000O10000N3N1N2N2N3O00000O101O00000O2O00000O101O00000O100000000O10000001O0000000000000010O0000001O000oNQNR@o1l?`Ng_Oa1V`0iNb_OW1\\`0QO__On0``0Y101O00001N1000001O01O00010O001O01O01O0010O0001O010O001O100O1O1O100O1O100O1O1O100O1O11O001O1O001O001O001O1O001O001O1O001O010O001O1O001O010O1O001O001O0010O0001O001O001O001O00001O001O001O000nE[Li3f3RL`Ll3`3PLeLo3\\3lKjLS4U3kKnLT4S3iKPMV4P3hKRMX4n2fKTMZ4m2cKUM]4k2aKXM^4i2`KXMa4k2ZKVMf4m2UKVMj4n2QKSMo4Q3lJQMS5R3iJPMV5U3dJlL]5W3^JkLa5Y3ZJiLe5\\3UJdLm5_3nIaLS6d3gI\\L[6g3`IXLb6k3ZIiKnLSOi9V5VIgKVMnNf9]5PIeK]7]4`HcKa7_4]H`Ke7a4XH_Ki7a4VH_Kk7a4TH_Km7b4QH^KQ8a4nG`KR8`4mG`KT8`4kGaKV8^4iGbKX8^4hGbKX8^4gGbK[8^4cGcK]8]4bGdK^8\\4aGdK`8\\4^GfKc8Y4\\GgKe8Y4YGiKg8W4XGiKi8X4TGjKm8U4RGlKn8T4PGmKQ9S4nFnKR9R4lFoKV9P4iFQLW9o3fFTLZ9l3cFVL^9j3_FYLa9g3\\F[Lf9d3WF_Li9a3TFbLl9_3PFcLQ:]3lEfLU:Y3gEkLY:U3cEnL^:R3^ERMc:m2YEVMh:j2TEZMl:f2PE]MQ;c2kD_MX;`2dDbM^;^52M3L5K4L5K4L5K4M4B=M4L3M3N3N1O2M2O2N1N3N1O2O001O000010O01O001O001O001O2N1O2M3N2N2M3N3M2M3N3M2M3N1O2N2M3M3M3L4L5L3L4L4M3L4L3M4K5G9G9G8G:G9G9G8H9G9G9F:Gk^:"}, "label": "the doctor is covering a small child with a sheet"}]}
{"id": 264165, "image": "COCO_train2014_000000264165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and black bow tie leaning over a person in a bed\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and black bow tie leaning over a person in a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 44, 54, 55, 56, 57, 58, 59, 60, 61, 62, 70, 71, 72, 73, 74, 75, 76, 77, 78, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 100, 101, 102, 103, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 131, 132, 139, 140, 141, 142, 143, 146, 147, 156, 157, 158, 161, 162, 163, 172, 173, 174, 177, 178, 187, 188, 202, 203, 204, 218, 219, 234], "bbox": [0.07729545454545454, 0.090953125, 0.9609772727272727, 0.617328125], "iscrowd": 0, "area": 43981.572, "rle": {"size": [640, 440], "counts": "Yae02nc04K5K5L2M2N2O1N2O1N2N100O1O100O1O100O101N1O100O1O100O01_OX]OHib09X]OEhb0<Y]OAib0?X]O_Ohb0b0X]O\\Oib0e0X]OXOjb0h0600100O1N2N2O1N1O2O1N2N2O1N2N101N2O1N2N2O4K5K5L3L0001D]^O[Nca0\\1i^OaNWa0W1m0M200100O1O100O1000000O100000O10O100000000O100000000O10O100000O100000000O10000N3N1N2N2N3O00000O101O00000O2O00000O101O00000O100000000O10000001O0000000000000010O0000001O000oNQNR@o1l?`Ng_Oa1V`0iNb_OW1\\`0QO__On0``0Y101O00001N1000001O01O00010O001O01O01O0010O0001O010O001O100O1O1O100O1O100O1O1O100O1O11O001O1O001O001O001O1O001O001O1O001O010O001O1O001O010O1O001O001O0010O0001O001O001O001O00001O001O001O000nE[Li3f3RL`Ll3`3PLeLo3\\3lKjLS4U3kKnLT4S3iKPMV4P3hKRMX4n2fKTMZ4m2cKUM]4k2aKXM^4i2`KXMa4k2ZKVMf4m2UKVMj4n2QKSMo4Q3lJQMS5R3iJPMV5U3dJlL]5W3^JkLa5Y3ZJiLe5\\3UJdLm5_3nIaLS6d3gI\\L[6g3`IXLb6k3ZIiKnLSOi9V5VIgKVMnNf9]5PIeK]7]4`HcKa7_4]H`Ke7a4XH_Ki7a4VH_Kk7a4TH_Km7b4QH^KQ8a4nG`KR8`4mG`KT8`4kGaKV8^4iGbKX8^4hGbKX8^4gGbK[8^4cGcK]8]4bGdK^8\\4aGdK`8\\4^GfKc8Y4\\GgKe8Y4YGiKg8W4XGiKi8X4TGjKm8U4RGlKn8T4PGmKQ9S4nFnKR9R4lFoKV9P4iFQLW9o3fFTLZ9l3cFVL^9j3_FYLa9g3\\F[Lf9d3WF_Li9a3TFbLl9_3PFcLQ:]3lEfLU:Y3gEkLY:U3cEnL^:R3^ERMc:m2YEVMh:j2TEZMl:f2PE]MQ;c2kD_MX;`2dDbM^;^52M3L5K4L5K4L5K4M4B=M4L3M3N3N1O2M2O2N1N3N1O2O001O000010O01O001O001O001O2N1O2M3N2N2M3N3M2M3N3M2M3N1O2N2M3M3M3L4L5L3L4L4M3L4L3M4K5G9G9G8G:G9G9G8H9G9G9F:Gk^:"}, "label": "a man in a white shirt and black bow tie leaning over a person in a bed"}]}
{"id": 264165, "image": "COCO_train2014_000000264165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man curled up on a bed\"."}], "task": "refering", "answer_template": "The \"a man curled up on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 134, 135, 136, 137, 138, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 200, 201, 202, 203], "bbox": [0.3618636363636364, 0.334515625, 0.7842272727272728, 0.5443749999999999], "iscrowd": 0, "area": 16537.667699999995, "rle": {"size": [640, 440], "counts": "`eS31lc04M3L5L3L4M3M4K4M3L5L3L4M3M2M3N3L3N2M3N3M2M3N2M4M2O1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O1O2N1O100O2O0O100O2O0O1O101N100O101N100O101O0000001O01O0001O0000001O0000001O0001O01O0000010O01O0010O01O1O010O001O010O001O010O001O010O001O100O1O1O2O0O1O1O100O2N3M5K3M2M2O2N000O10000O1000O010000O10000O1000000O100O1O1O001N2O1O001O1O0O2O1O001O1N101O1O001O1N101O001O1N101M3J6K6J6I6K6I6K6I7JXTk1"}, "label": "a man curled up on a bed"}]}
{"id": 534502, "image": "COCO_train2014_000000534502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 110, 113], "bbox": [0.18392857142857144, 0.171421875, 0.5815476190476191, 0.32078125], "iscrowd": 0, "area": 7806.232299999998, "rle": {"size": [640, 420], "counts": "gY`11mc04M3M1O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O001O1O1O1O1O1O1O1O1O100O1O100O1O1J510O100000000000000O2O001O001O4L5J5L4L1O0000O1O10O0100O1O100O100O00100O100O100000O1000000000000O1000O1000000000OdNf^O4Za0Il^O4Ta0JQ_O3o`0LS_O3m`0KV_O4j`0JX_O6h`0H[_O7e`0G^_O8a`0Ga_O9_`0Ed_O9]`0Ef_O:Z`0Dh_O<X`0Bk_O=U`0An_O>S`0_Oo_Oa0S`0[OP@d0Q`0YOR@f0da001O1O001O1O001O1O010O1O1O010O1O1O010O1O00100O1O00100O1O0010N100O1O2O0001O001M7IQj]3"}, "label": "a woman in a blue shirt"}]}
{"id": 534502, "image": "COCO_train2014_000000534502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman taking a baby out the backseat\"."}], "task": "refering", "answer_template": "The \"a woman taking a baby out the backseat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 110, 113], "bbox": [0.18392857142857144, 0.171421875, 0.5815476190476191, 0.32078125], "iscrowd": 0, "area": 7806.232299999998, "rle": {"size": [640, 420], "counts": "gY`11mc04M3M1O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O001O1O1O1O1O1O1O1O1O100O1O100O1O1J510O100000000000000O2O001O001O4L5J5L4L1O0000O1O10O0100O1O100O100O00100O100O100000O1000000000000O1000O1000000000OdNf^O4Za0Il^O4Ta0JQ_O3o`0LS_O3m`0KV_O4j`0JX_O6h`0H[_O7e`0G^_O8a`0Ga_O9_`0Ed_O9]`0Ef_O:Z`0Dh_O<X`0Bk_O=U`0An_O>S`0_Oo_Oa0S`0[OP@d0Q`0YOR@f0da001O1O001O1O001O1O010O1O1O010O1O1O010O1O00100O1O00100O1O0010N100O1O2O0001O001M7IQj]3"}, "label": "a woman taking a baby out the backseat"}]}
{"id": 378864, "image": "COCO_train2014_000000378864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl with the least amount of content\"."}], "task": "refering", "answer_template": "The \"a bowl with the least amount of content\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 255, 256, 257], "bbox": [0.021156249999999998, 0.44558333333333333, 0.259109375, 0.6694375], "iscrowd": 0, "area": 12102.9853, "rle": {"size": [480, 640], "counts": "Uj64i>5L4K5L4L4L4L3N2M2O2N2N2O1N1O2N2O1N2N1O2O1N2N2O002M3N1N3N1O2M2O2M3N1O2M2O0O2O0O2O001N2O0O2O001O0O2O1O001N2O1O000O1000000O1000000O1000000O1000000O1000000O1000000000000000000000000000000000000000000000000000O100O1O100O1O100O1O100O1O101N1O101N1O2N100O2N1O2L3N2M4M2M4L3N2M4M2M4M2M3N3L3N3L3N2N3M2N3M5K7I6K6I^nm6"}, "label": "a bowl with the least amount of content"}]}
{"id": 378864, "image": "COCO_train2014_000000378864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl at the edge of the row with the least inside it\"."}], "task": "refering", "answer_template": "The \"a bowl at the edge of the row with the least inside it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 255, 256, 257], "bbox": [0.021156249999999998, 0.44558333333333333, 0.259109375, 0.6694375], "iscrowd": 0, "area": 12102.9853, "rle": {"size": [480, 640], "counts": "Uj64i>5L4K5L4L4L4L3N2M2O2N2N2O1N1O2N2O1N2N1O2O1N2N2O002M3N1N3N1O2M2O2M3N1O2M2O0O2O0O2O001N2O0O2O001O0O2O1O001N2O1O000O1000000O1000000O1000000O1000000O1000000O1000000000000000000000000000000000000000000000000000O100O1O100O1O100O1O100O1O101N1O101N1O2N100O2N1O2L3N2M4M2M4L3N2M4M2M4M2M3N3L3N3L3N2N3M2N3M5K7I6K6I^nm6"}, "label": "a bowl at the edge of the row with the least inside it"}]}
{"id": 59382, "image": "COCO_train2014_000000059382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby\"."}], "task": "refering", "answer_template": "The \"the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 72, 73, 74, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 140, 141, 142, 143, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292], "bbox": [0.038093749999999996, 0.22430555555555556, 0.9218437500000001, 1.0], "iscrowd": 0, "area": 109932.75995000007, "rle": {"size": [360, 640], "counts": "\\T91V;1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1fLSOXLn0e3WOXLj0d3\\OYLe0c3@\\L`0a3E\\L<`3J]L7_3O^L2^34_LM^37`LJ\\3<aLE[3a0bL@[3e0cL[OY3j0eLWOW3o0fLROV3T1gLmNV3X1gLiNU3\\1iLeNS3a1jL`NS3e1jL\\NR3j1kLWNQ3o1mLQNo2T2oLmMn2X2oLiMm2]2PMdMl2b2QM_Ml2e2RM\\Mm2g2PMZMn2j2oLWMo2m2oLSMP3o2nLRMP3R3mLPMP3T3mLmLR3V3kLkLS3Y3jLhLT3[3jLfLU3]3hLdLV3`3hL`LV3d3gL]LX3e3fL\\LX3k5N2O1N2N2O1N2N2O1N2N2O1N2O100000000000000000000003M2N1O0000000000000000000000001O0mIYMRNLW5l2eLeM[3b2]L_Mc3a2\\L`Md3a2ZL`Mf3`2ZL`Mf3`2YLaMg3`2XLaMg3_2YLbMf3^2YLdMf3]2YLdMf3\\2YLfMf3[2YLfMf3Z2YLhMf3X2YLkMe3V2YLlMf3T2YLnMf3R2YLQNe3P2YLUNc3l1[LWNc3j1[LYNc3j1WKcLl0e1l3P2QLSNm3n1QLTNn3m1PLVNn3j1QLZNl3f1TL]N@mMn3g3`L`N\\OlMT4d3`LhN`3X1_LiNa3W1_L\\N\\OWNT4]3_L]N@SNQ4a3]L\\NFPNm3d3XL`N0hMh3h3PLgNa4Y1\\KjNd4l31O1O100O100001O00000000001O2N1O2NN2N2N2O1N2O1O1O1O100O1O1O10000O1000000O100001O1O002N4L4L7I0001O0000000O10000O100O100O100O10000O10000O10000O10000000000000000O100000000000000O100O100jN_KdKb4\\4`KaKa4_4`K]Kc4b4R100000000O10000O100O100L4D<N2O1O1O100O10000O100O100O100001O0000001O00000000000000001O001O00000000000000GTIVLl6j3WIRLj6m39O100O1N2O100O100O1N2O1O1N2O1N2O1O1O1O1O100O1O1O100O1O1O1O100O1O2O0O100O1O100O10000O100O10000O10000O100O10000O100O10000O10000000000O10000000000O10000000000O1000000000000O1000000000000O10000001O000B_MUHa2k7bMQH_2o7eMlG\\2T8gMiGY2W8<00O100000000O100O10000O100O10000O100O1000000O101O001N101O001N101O001N2O001O001N101O001N101O0O2O001O0O2O001N101O1N101O001N101O0O2O001O0O2O001N101O0O2O1O001N101O0O2O001N101O001N101O0O2O001O1N101O0O2O001N102N1N2O1N2O2N1N2O1O1N2O2N1N2O1O1N2O2N1N2O1O1N2N3M2N2N2N2M3N3M2N2N2N2M3N3M2N2N2N2M3N3M2N2N2N2M3L5L3L4L4L4M6I8H8H8I7F9E<E;Di[a0"}, "label": "the baby"}]}
{"id": 59382, "image": "COCO_train2014_000000059382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby with wide eyes\"."}], "task": "refering", "answer_template": "The \"baby with wide eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 72, 73, 74, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 140, 141, 142, 143, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292], "bbox": [0.038093749999999996, 0.22430555555555556, 0.9218437500000001, 1.0], "iscrowd": 0, "area": 109932.75995000007, "rle": {"size": [360, 640], "counts": "\\T91V;1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1fLSOXLn0e3WOXLj0d3\\OYLe0c3@\\L`0a3E\\L<`3J]L7_3O^L2^34_LM^37`LJ\\3<aLE[3a0bL@[3e0cL[OY3j0eLWOW3o0fLROV3T1gLmNV3X1gLiNU3\\1iLeNS3a1jL`NS3e1jL\\NR3j1kLWNQ3o1mLQNo2T2oLmMn2X2oLiMm2]2PMdMl2b2QM_Ml2e2RM\\Mm2g2PMZMn2j2oLWMo2m2oLSMP3o2nLRMP3R3mLPMP3T3mLmLR3V3kLkLS3Y3jLhLT3[3jLfLU3]3hLdLV3`3hL`LV3d3gL]LX3e3fL\\LX3k5N2O1N2N2O1N2N2O1N2N2O1N2O100000000000000000000003M2N1O0000000000000000000000001O0mIYMRNLW5l2eLeM[3b2]L_Mc3a2\\L`Md3a2ZL`Mf3`2ZL`Mf3`2YLaMg3`2XLaMg3_2YLbMf3^2YLdMf3]2YLdMf3\\2YLfMf3[2YLfMf3Z2YLhMf3X2YLkMe3V2YLlMf3T2YLnMf3R2YLQNe3P2YLUNc3l1[LWNc3j1[LYNc3j1WKcLl0e1l3P2QLSNm3n1QLTNn3m1PLVNn3j1QLZNl3f1TL]N@mMn3g3`L`N\\OlMT4d3`LhN`3X1_LiNa3W1_L\\N\\OWNT4]3_L]N@SNQ4a3]L\\NFPNm3d3XL`N0hMh3h3PLgNa4Y1\\KjNd4l31O1O100O100001O00000000001O2N1O2NN2N2N2O1N2O1O1O1O100O1O1O10000O1000000O100001O1O002N4L4L7I0001O0000000O10000O100O100O100O10000O10000O10000O10000000000000000O100000000000000O100O100jN_KdKb4\\4`KaKa4_4`K]Kc4b4R100000000O10000O100O100L4D<N2O1O1O100O10000O100O100O100001O0000001O00000000000000001O001O00000000000000GTIVLl6j3WIRLj6m39O100O1N2O100O100O1N2O1O1N2O1N2O1O1O1O1O100O1O1O100O1O1O1O100O1O2O0O100O1O100O10000O100O10000O10000O100O10000O100O10000O10000000000O10000000000O10000000000O1000000000000O1000000000000O10000001O000B_MUHa2k7bMQH_2o7eMlG\\2T8gMiGY2W8<00O100000000O100O10000O100O10000O100O1000000O101O001N101O001N101O001N2O001O001N101O001N101O0O2O001O0O2O001N101O1N101O001N101O0O2O001O0O2O001N101O0O2O1O001N101O0O2O001N101O001N101O0O2O001O1N101O0O2O001N102N1N2O1N2O2N1N2O1O1N2O2N1N2O1O1N2O2N1N2O1O1N2N3M2N2N2N2M3N3M2N2N2N2M3N3M2N2N2N2M3N3M2N2N2N2M3L5L3L4L4L4M6I8H8H8I7F9E<E;Di[a0"}, "label": "baby with wide eyes"}]}
{"id": 59382, "image": "COCO_train2014_000000059382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand of a small child holding a brush\"."}], "task": "refering", "answer_template": "The \"the hand of a small child holding a brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 55, 56, 57, 58, 59, 69, 70, 78, 79, 80, 92], "bbox": [0.0, 0.0022500000000000003, 0.5940937500000001, 0.33483333333333337], "iscrowd": 0, "area": 18655.43565, "rle": {"size": [360, 640], "counts": "^1Z2a7]10O1O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O1O1N2N2N2N2O1N2N2N2N2O1N2N2N2O1O1O1N2O1O1O1O100001O000000001O000000001O01O2N2O1N3M2N2N2F:]Ohe6TOe[I001O00001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O00001O001O00001O001O00001O00001O001O00001O001OO1O1O1O1O1O1O1O1O1O1O1O1O10000001O0000001O00001O00001O0000001O00001O00001O0000001O00001O0000000000001O0000000000001O000001O001O010OkFaNa8_1^GeN`8Z1`GjN\\8V1dGmNZ8S1eGQOW8o0hGVOU8i0jG[OS8e0kG@S8?kGDT8=iGEW8;gGGZ88eGI[87fGH[87eGI[88eGG[89fGGZ88gGGY89hGFY89gGGY8:gGEZ8:gGDZ8<gGCY8=gGCZ8<gGBZ8?fG@[8?fG@Z8`0gG^O[8a0eG_O[8b0eG]O[8c0fG[O\\8d0eG[O[8e0fGZO[8f0dGZO\\8g0dGXO]8g0dGXO\\8i0cGWO]8j0aGXO_8h0`GXO`8i0^GXOc8g0\\GZOd8g0ZGZOg8f0XGZOh8g0VGZOj8c10O001O00001O001O001O0O101N101N101N100O2O0O2O0CgFjN[9f0dF]O2MZ9g0dFZO3NZ9h0dFXO4OY9h0dFWO40Y9j0bFTO71Y9a0dF\\OM073X9b0jFYO03X9c0UG\\Ol8d0TG\\Ol8e0SGZOo8e0j001O010O0010O01O01O01O010O001O010O00001O001O001N101O1O1O0O2O1O0O2O1O1N101O1N101OT[k2"}, "label": "the hand of a small child holding a brush"}]}
{"id": 59382, "image": "COCO_train2014_000000059382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child combing the babies hair\"."}], "task": "refering", "answer_template": "The \"a child combing the babies hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 55, 56, 57, 58, 59, 69, 70, 78, 79, 80, 92], "bbox": [0.0, 0.0022500000000000003, 0.5940937500000001, 0.33483333333333337], "iscrowd": 0, "area": 18655.43565, "rle": {"size": [360, 640], "counts": "^1Z2a7]10O1O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O1O1N2N2N2N2O1N2N2N2N2O1N2N2N2O1O1O1N2O1O1O1O100001O000000001O000000001O01O2N2O1N3M2N2N2F:]Ohe6TOe[I001O00001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O00001O001O00001O001O00001O00001O001O00001O001OO1O1O1O1O1O1O1O1O1O1O1O1O10000001O0000001O00001O00001O0000001O00001O00001O0000001O00001O0000000000001O0000000000001O000001O001O010OkFaNa8_1^GeN`8Z1`GjN\\8V1dGmNZ8S1eGQOW8o0hGVOU8i0jG[OS8e0kG@S8?kGDT8=iGEW8;gGGZ88eGI[87fGH[87eGI[88eGG[89fGGZ88gGGY89hGFY89gGGY8:gGEZ8:gGDZ8<gGCY8=gGCZ8<gGBZ8?fG@[8?fG@Z8`0gG^O[8a0eG_O[8b0eG]O[8c0fG[O\\8d0eG[O[8e0fGZO[8f0dGZO\\8g0dGXO]8g0dGXO\\8i0cGWO]8j0aGXO_8h0`GXO`8i0^GXOc8g0\\GZOd8g0ZGZOg8f0XGZOh8g0VGZOj8c10O001O00001O001O001O0O101N101N101N100O2O0O2O0CgFjN[9f0dF]O2MZ9g0dFZO3NZ9h0dFXO4OY9h0dFWO40Y9j0bFTO71Y9a0dF\\OM073X9b0jFYO03X9c0UG\\Ol8d0TG\\Ol8e0SGZOo8e0j001O010O0010O01O01O01O010O001O010O00001O001O001N101O1O1O0O2O1O0O2O1O1N101O1N101OT[k2"}, "label": "a child combing the babies hair"}]}
{"id": 395259, "image": "COCO_train2014_000000395259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bag on a stool\"."}], "task": "refering", "answer_template": "The \"a black bag on a stool\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 242, 261, 262, 263, 264], "bbox": [0.324546875, 0.3721041666666667, 0.52815625, 0.6845208333333335], "iscrowd": 0, "area": 14023.182199999997, "rle": {"size": [480, 640], "counts": "ShQ3l0o=7I7I8I6J7I6K5H9H7I7I7K5J6K4M3M4L2N3M3N101N2N00010O2N100O2O001O001N10001O000O10001O000000000000000000001O00O101O000O101O00000O2O0000001O0O10001O00001O0O2O1O001O1O001N2O001O2N2M3N2N2_LUEi2d;M3M3M3M3M2M4M3M2N2N1O1O2N1O2M2O1N3M2O1N2N2O2M2N2O1N2N2O1N2O2M2N3N1N3M3N1M:DnX]4"}, "label": "a black bag on a stool"}]}
{"id": 395259, "image": "COCO_train2014_000000395259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black travel bag\"."}], "task": "refering", "answer_template": "The \"black travel bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 242, 261, 262, 263, 264], "bbox": [0.324546875, 0.3721041666666667, 0.52815625, 0.6845208333333335], "iscrowd": 0, "area": 14023.182199999997, "rle": {"size": [480, 640], "counts": "ShQ3l0o=7I7I8I6J7I6K5H9H7I7I7K5J6K4M3M4L2N3M3N101N2N00010O2N100O2O001O001N10001O000O10001O000000000000000000001O00O101O000O101O00000O2O0000001O0O10001O00001O0O2O1O001O1O001N2O001O2N2M3N2N2_LUEi2d;M3M3M3M3M2M4M3M2N2N1O1O2N1O2M2O1N3M2O1N2N2O2M2N2O1N2N2O1N2O2M2N3N1N3M3N1M:DnX]4"}, "label": "black travel bag"}]}
{"id": 395259, "image": "COCO_train2014_000000395259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back luggage on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the back luggage on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 281, 282, 304, 305], "bbox": [0.0026406249999999997, 0.47572916666666665, 0.291703125, 0.8284791666666665], "iscrowd": 0, "area": 14349.100349999999, "rle": {"size": [480, 640], "counts": "RW1f0Z>00000001O000000000000000000000000000000000000000000000000ROn0[Oe00000O1000000000000000000000000001O0000001O00001O0000001O000000001O0000000001O0001OO100O10000O100O100N2O1N2N200O10000O1000000O10000O100000000O10000000000000000000000O11O00001O00001O001O1O1O1O001O1O1O001O000000001O0001O000010O000001O1O1O101N2Na0_Od0\\Oc0]O>B01O0001O000001O0001O01O1O1O1O010O1O1O3M4L5L3L4L5K4L4Li0WOS2cMboc6"}, "label": "the back luggage on the left in the right hand picture"}]}
{"id": 395259, "image": "COCO_train2014_000000395259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown suitcase with circles on the left\"."}], "task": "refering", "answer_template": "The \"a large brown suitcase with circles on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 281, 282, 304, 305], "bbox": [0.0026406249999999997, 0.47572916666666665, 0.291703125, 0.8284791666666665], "iscrowd": 0, "area": 14349.100349999999, "rle": {"size": [480, 640], "counts": "RW1f0Z>00000001O000000000000000000000000000000000000000000000000ROn0[Oe00000O1000000000000000000000000001O0000001O00001O0000001O000000001O0000000001O0001OO100O10000O100O100N2O1N2N200O10000O1000000O10000O100000000O10000000000000000000000O11O00001O00001O001O1O1O1O001O1O1O001O000000001O0001O000010O000001O1O1O101N2Na0_Od0\\Oc0]O>B01O0001O000001O0001O01O1O1O1O010O1O1O3M4L5L3L4L5K4L4Li0WOS2cMboc6"}, "label": "a large brown suitcase with circles on the left"}]}
{"id": 395259, "image": "COCO_train2014_000000395259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and pink opened piece of luggage\"."}], "task": "refering", "answer_template": "The \"a brown and pink opened piece of luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275], "bbox": [0.750484375, 0.370375, 1.0, 0.692625], "iscrowd": 0, "area": 13134.824750000003, "rle": {"size": [480, 640], "counts": "^YQ75k>1N2N101N2N2L3E<eNYOZDh0Z;EdD<Q;NPE1f::YEG[:e0eE[OP:o0PFSOk9Q1TFQOk9n0VFROi9o0VFSOi9l0XFUOf9l0YFVOf9j0ZFWOe9h0[FYOd9h0[FZOd9e0]F\\Ob9d0]F^Ob9a0_F@`9`0_FKW93jFNV91kFOU9OmF1S9NPG0P9ORG0n80SGOm80VGNj81XGNh81YGOg81YGOg80ZGOg80ZG0f8O[G0f8O[G0f80ZG0f8O\\GOe80\\G0d8O]G1c8O]G1c8N^G2b8M_G3b8K_G5a8K_G5a8J`G6`8IaG7_8HbG8^8GcG9]8GcG9]8FdG:\\8EeG;[8EeG;\\8CeG=[8BfG>Z8BfG>Z8AgG?Y8AgG`0X8_OiGa0W8^OjGb0V8^OjGb0V8]OkGc0V8\\OjGd0V8[OkGe0U8[OkGe0U8ZOlGf0T8ZOlGf0T8ZOlGf0T8ZOlGf0T8YOmGg0S8YOmGg0S8YOmGg0T8XOlGh0T8WOmGi0S8WOmGi0S8WOmGi0S8XOlGh0T8XOlGh0T8XOlGh0T8XOlGh0T8XOlGh0U8XOjGh0V8YOiGg0W8ZOhGf0X8[OgGe0Y8\\OfGd0Z8]OeGc0[8_OcGa0]8@bG`0^8AaG?_8B`G>a8B^G>b8C]G=c8D\\G<d8E[G;e8FZG:f8GYG9g8HXG8i8HVG8j8JTG6l8KSG5m8LRG4n8MQG3o8NPG2Q9NnF2R9OmF1S9OmF1S90lF0T91kFOU91kFOU92jFNW92hFNX92hFNX93gFMY94fFLZ94fFLZ95eFK[96dFJ]95cFK]96bFJ^97aFI_96bFJ^96bFJ^96bFJ^96bFJ_94bFL^94bFL^94bFL_93aFM`91aFO`90`F0a9O_F1b9N^F2c9L^F4c9K]F5f9HZF8h9\\ObFd0a9fNRGZ1Z;000000000000000000000000000000000XG"}, "label": "a brown and pink opened piece of luggage"}]}
{"id": 542718, "image": "COCO_train2014_000000542718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black suit & hat on wearing a white necktie\"."}], "task": "refering", "answer_template": "The \"a man with a black suit & hat on wearing a white necktie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 48, 49, 50, 51, 63, 64, 65, 66, 67, 78, 79, 80, 81, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 129, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.010070257611241218, 0.110125, 0.7494145199063232, 0.9845625], "iscrowd": 0, "area": 125772.17859999998, "rle": {"size": [640, 427], "counts": "Sk21kc0:E;F:F:E;F:E;F:E;F:E;F:E;F:E;F:E;F:E;F:E;H8I7H8I7I7H8I7I7H8I7I7H8I7I7I7J6I7I7I7J1N2N2N2O2M2N2O1N2N3M2O1N2N3M100O2N1O100O1O2O2M3M3M3N1N3M3M3N0O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1QNdBYM]=f2gBWMY=i2jBSMW=l2lBQMU=o2kBPMV=o2lBPMT=P3mBnLT=Q3mBnLT=Q3nBmLS=S3nBkLS=l1]BPNa02T=l1cBlM:6T=m1jBeM3=T=m1PC`MM`0U=o1VCYMGf0T=P2\\CTM@k0U=o1ZDnMh;Q2YDnMh;Q2ZDhMl;W2VDbMP<]2QD]MU<a2f2O1O1O1000000O10000000000000000000000O1000000000000000000000000O1000000000000000000O10000000000000000O1001O000000000000001O000000002N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2c@kMk<W2SCkMk<W2RClMl<V2RClMl<V2QClMn<V2PClMn<V2oBmMo<T2PCmMo<S2QCnMn<R2QCoMo<R2PCoMo<Q2PCPNP=Q2oBPNP=P2oBQNQ=P2nBQNQ=o1nBRNR=o1mBRNR=n1mBSNS=n1lBSNS=m1mBSNS=n1kBTNT=l1lBTNT=m1jBTNV=m1iBTNV=l1iBUNW=l1hBUNW=k1hBVNX=k1gBVNX=Y4000001O00001O00001O00001O00001O00001O0000001O00001O00001O001O2N1O2N1O2N1O1O2N1O2N1O2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N4L6J6J6J6J7QMaFoJe9j4`FRKf9f4aFUKe9d4`FXKf9`4aF[Ke9^4bF\\Ke9[4gF[K_9]4nFXKX9a4SGUKS9c4YGSKm8f4_GoJh8h4_3I7I7H8I7I8H7H8I7I7I7H9H7I7H8G9F:G:E:Gn\\R2"}, "label": "a man with a black suit & hat on wearing a white necktie"}]}
{"id": 542718, "image": "COCO_train2014_000000542718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black suit and a hat\"."}], "task": "refering", "answer_template": "The \"a man wearing a black suit and a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 48, 49, 50, 51, 63, 64, 65, 66, 67, 78, 79, 80, 81, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 129, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.010070257611241218, 0.110125, 0.7494145199063232, 0.9845625], "iscrowd": 0, "area": 125772.17859999998, "rle": {"size": [640, 427], "counts": "Sk21kc0:E;F:F:E;F:E;F:E;F:E;F:E;F:E;F:E;F:E;F:E;H8I7H8I7I7H8I7I7H8I7I7H8I7I7I7J6I7I7I7J1N2N2N2O2M2N2O1N2N3M2O1N2N3M100O2N1O100O1O2O2M3M3M3N1N3M3M3N0O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1QNdBYM]=f2gBWMY=i2jBSMW=l2lBQMU=o2kBPMV=o2lBPMT=P3mBnLT=Q3mBnLT=Q3nBmLS=S3nBkLS=l1]BPNa02T=l1cBlM:6T=m1jBeM3=T=m1PC`MM`0U=o1VCYMGf0T=P2\\CTM@k0U=o1ZDnMh;Q2YDnMh;Q2ZDhMl;W2VDbMP<]2QD]MU<a2f2O1O1O1000000O10000000000000000000000O1000000000000000000000000O1000000000000000000O10000000000000000O1001O000000000000001O000000002N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2c@kMk<W2SCkMk<W2RClMl<V2RClMl<V2QClMn<V2PClMn<V2oBmMo<T2PCmMo<S2QCnMn<R2QCoMo<R2PCoMo<Q2PCPNP=Q2oBPNP=P2oBQNQ=P2nBQNQ=o1nBRNR=o1mBRNR=n1mBSNS=n1lBSNS=m1mBSNS=n1kBTNT=l1lBTNT=m1jBTNV=m1iBTNV=l1iBUNW=l1hBUNW=k1hBVNX=k1gBVNX=Y4000001O00001O00001O00001O00001O00001O0000001O00001O00001O001O2N1O2N1O2N1O1O2N1O2N1O2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N4L6J6J6J6J7QMaFoJe9j4`FRKf9f4aFUKe9d4`FXKf9`4aF[Ke9^4bF\\Ke9[4gF[K_9]4nFXKX9a4SGUKS9c4YGSKm8f4_GoJh8h4_3I7I7H8I7I8H7H8I7I7I7H9H7I7H8G9F:G:E:Gn\\R2"}, "label": "a man wearing a black suit and a hat"}]}
{"id": 542718, "image": "COCO_train2014_000000542718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black haired woman wearing blue jeans and a black shirt\"."}], "task": "refering", "answer_template": "The \"a black haired woman wearing blue jeans and a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 103, 104, 117, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 164, 177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 328, 329, 344], "bbox": [0.7974941451990631, 0.215234375, 1.0, 0.9719218749999999], "iscrowd": 0, "area": 31252.539699999998, "rle": {"size": [640, 427], "counts": "U[e67hc07J6I7I7I7I7I7I7\\@XNc<o1TC[Nd<i1UCaNe<a1UCjNc<Y1WCROa<R1XCYO`<j0ZCA^<b0\\CH]<;^CN[<5_C5Z<N`C;Z<H`Cb0Z<A_Ci0[1ZLS9P3\\Eo0P1gLY9k6VFiI`9f8E8H9G9G8H9D8H7H9H7J7I7J5J7I5K4L3M4L3M4L3M3M4N1O2O0O101O0O2O0O2O0O101O0O2O0O101N101O0O2O0O101O01O001O1O001O1O001O1O001O001O1O\\K"}, "label": "a black haired woman wearing blue jeans and a black shirt"}]}
{"id": 542718, "image": "COCO_train2014_000000542718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman with dark hair , wearing a black shirt and jeans , standing with her arms folded\"."}], "task": "refering", "answer_template": "The \"a young woman with dark hair , wearing a black shirt and jeans , standing with her arms folded\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 103, 104, 117, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 164, 177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 328, 329, 344], "bbox": [0.7974941451990631, 0.215234375, 1.0, 0.9719218749999999], "iscrowd": 0, "area": 31252.539699999998, "rle": {"size": [640, 427], "counts": "U[e67hc07J6I7I7I7I7I7I7\\@XNc<o1TC[Nd<i1UCaNe<a1UCjNc<Y1WCROa<R1XCYO`<j0ZCA^<b0\\CH]<;^CN[<5_C5Z<N`C;Z<H`Cb0Z<A_Ci0[1ZLS9P3\\Eo0P1gLY9k6VFiI`9f8E8H9G9G8H9D8H7H9H7J7I7J5J7I5K4L3M4L3M4L3M3M4N1O2O0O101O0O2O0O2O0O101O0O2O0O101N101O0O2O0O101O01O001O1O001O1O001O1O001O001O1O\\K"}, "label": "a young woman with dark hair , wearing a black shirt and jeans , standing with her arms folded"}]}
{"id": 542718, "image": "COCO_train2014_000000542718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with long hair\"."}], "task": "refering", "answer_template": "The \"a person with long hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 135, 150, 165], "bbox": [0.0, 0.1550625, 0.24925058548009368, 0.53034375], "iscrowd": 0, "area": 11749.022400000002, "rle": {"size": [640, 427], "counts": "j8`1i?h2N2N2N3M2N2N2N2NK1MO2N1O2O07J6J6I7J5J7L4001O0O2M2O2N1O2M2O2N1O1O2M20001O0O2O00001O000010O01O0100O1O100O100O100O100O100O1O100O100O100O1^NS@QOm?k0W@TOj?h0[@VOf?f0^@ZOb?e0_@ZOb?f0^@ZOb?g0^@WOc?i0]@WOc?j0\\@UOe?k0\\@TOd?m0[@ROf?n0Z@QOg?o0Y@QOg?P1Y@nNh?R1X@nNh?S1W@lNj?T1W@kNj?U1U@jNn?T1R@lNP`0S1o_OlNT`0R1m_OmNT`05b_OF94X`03V@SOBj0Z`00g@N\\?0d@0]?Nd@1_?Ma@3a?Ja@4b?I_@7d?F\\@9h?CY@=k?^OV@a0n?\\OR@d0ea0O100O7J6Ii]X6"}, "label": "a person with long hair"}]}
{"id": 542718, "image": "COCO_train2014_000000542718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of a dark curly haired woman standing on a sidewalk\"."}], "task": "refering", "answer_template": "The \"back of a dark curly haired woman standing on a sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 135, 150, 165], "bbox": [0.0, 0.1550625, 0.24925058548009368, 0.53034375], "iscrowd": 0, "area": 11749.022400000002, "rle": {"size": [640, 427], "counts": "j8`1i?h2N2N2N3M2N2N2N2NK1MO2N1O2O07J6J6I7J5J7L4001O0O2M2O2N1O2M2O2N1O1O2M20001O0O2O00001O000010O01O0100O1O100O100O100O100O100O1O100O100O100O1^NS@QOm?k0W@TOj?h0[@VOf?f0^@ZOb?e0_@ZOb?f0^@ZOb?g0^@WOc?i0]@WOc?j0\\@UOe?k0\\@TOd?m0[@ROf?n0Z@QOg?o0Y@QOg?P1Y@nNh?R1X@nNh?S1W@lNj?T1W@kNj?U1U@jNn?T1R@lNP`0S1o_OlNT`0R1m_OmNT`05b_OF94X`03V@SOBj0Z`00g@N\\?0d@0]?Nd@1_?Ma@3a?Ja@4b?I_@7d?F\\@9h?CY@=k?^OV@a0n?\\OR@d0ea0O100O7J6Ii]X6"}, "label": "back of a dark curly haired woman standing on a sidewalk"}]}
{"id": 550911, "image": "COCO_train2014_000000550911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female tennis player\"."}], "task": "refering", "answer_template": "The \"female tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 154, 155, 156, 157, 158, 159, 160, 161, 162, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 193, 200, 201, 202, 203, 204, 205, 206, 207, 208, 214, 215, 216, 217, 218, 219, 221, 222, 229, 230, 231, 232, 236, 237, 245, 246, 251, 252, 253, 260, 261, 267, 268, 269, 275, 276, 283, 284, 289, 290, 291, 298, 299, 303, 304, 305, 306, 313, 314, 328, 329], "bbox": [0.21243559718969554, 0.263421875, 1.0, 0.9473437499999999], "iscrowd": 0, "area": 58638.236650000035, "rle": {"size": [640, 427], "counts": "l]i14hc06M2O0O2O0O2O0O1O100O1O1O100O1O100O10000O100jD\\OP3d0lL@S3a0mL@R3`0mLAS3?mLBQ3?oLAP3`0PMAm2a0RM_On2b0RM^Ol2d0TM\\Ok2e0TM\\Ok2e0UM[Oj2f0VMZOhNF[MQ1m3YOfNJZMn0o3YOXJHV23Am0Q4WOVJ3R2IEn0S4VOTJ:P2BHn0S4WORJ?Q2]OHm0U4WOPJe0P2XOIk0X4XOnIi0Z1iNbM<m2j0X4XO^H1k0n0k1eNeM<m2h0X4XO\\H9l0l0i1aNhM<l2g0Z45UI=h1[NlM=l2g0Z4<nH;k1VNPN<l2h0Z4b0hH:m1PNTN=l2h0Z4f0dH:n1lMVN>l2g0[4\\3^J_KZNb0i2d0^4`3[J[K]Nb0i2c0^4f3ZIUKL1Ab0j2c0^4j3RI[KOICa0l2b0_4o3iH`K4a08W2j6n8O1O1O1O1O10000000000O1000000000000O1000000000000000ZNZIRFf6f9cIYF]6a9kI]FU6^9RJ`Fn5[9aJ[F_5`9kJ[FU5c9PKZFP5d9TKZFl4d9YKYFg4e9^KUFe4j9i2N2M3L4K5G9G9G9B?O0O2N2O1N101N2N2O0O2N2O1N3M2O2M2O2M3M2O2M3M3N2M7J1N2O2M2O1N2O1N2O1N2O1N10O1O02O2M2O2M100O2O0O100O10O0O2003L4M1N2O1N101N010O013L4M1N2O1O1N2O1L3N3M110O000aKk@U4T?dKUAZ4Y?O101N1O2O0O2N101N1O2O0O1O2O0O2N1O1O1O1O2N1O1O1O10000O100N3N1O10000001O000000001O01O07J7I6I8I7H6K4K5L7I2M2O2M2N3L3N2M3N2O1O1O00000O1O001O10000O1000000O100O1O101O0000001O0O10001O0O101O1O0O2O0100O10000`MgBVNZ=a1nB`NR=X1UChNj1PNY8o2VFQO^1TN[8c2^FYOT1YN\\8W2hF@h0]N`8j1PGI?]Nc8b1UG25^Ng8W1\\G<J^Nl8m0bGe0@`Nm8b0mGo0SO`NQ98UHX1eNeNV9EdHf1QNhN[9WOoD6f3k1lMlN_9SOoD5g3m1gMmNd9POoD6f3n1dMnNg9mNQE5e3P2aMoNi9mNQE3f3R2]MPOl9jNSIU2PMQOV>P1gAROY>o0dARO]>o0aARO_>o0_AQOc>o0[AQOf>P1XAPOj>P1TAPOm>Q1PAPOQ?R1l@nNV?R1h@nNY?T1d@lN^?T1`@lNa?X1X@jNj?e23M2M4M3M2N4L3M3M3M3M4L3M3M3M3M4K4M3M3M4L3M4L3M2N2N3N1Nd^O"}, "label": "female tennis player"}]}
{"id": 550911, "image": "COCO_train2014_000000550911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis player\"."}], "task": "refering", "answer_template": "The \"the tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 154, 155, 156, 157, 158, 159, 160, 161, 162, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 193, 200, 201, 202, 203, 204, 205, 206, 207, 208, 214, 215, 216, 217, 218, 219, 221, 222, 229, 230, 231, 232, 236, 237, 245, 246, 251, 252, 253, 260, 261, 267, 268, 269, 275, 276, 283, 284, 289, 290, 291, 298, 299, 303, 304, 305, 306, 313, 314, 328, 329], "bbox": [0.21243559718969554, 0.263421875, 1.0, 0.9473437499999999], "iscrowd": 0, "area": 58638.236650000035, "rle": {"size": [640, 427], "counts": "l]i14hc06M2O0O2O0O2O0O1O100O1O1O100O1O100O10000O100jD\\OP3d0lL@S3a0mL@R3`0mLAS3?mLBQ3?oLAP3`0PMAm2a0RM_On2b0RM^Ol2d0TM\\Ok2e0TM\\Ok2e0UM[Oj2f0VMZOhNF[MQ1m3YOfNJZMn0o3YOXJHV23Am0Q4WOVJ3R2IEn0S4VOTJ:P2BHn0S4WORJ?Q2]OHm0U4WOPJe0P2XOIk0X4XOnIi0Z1iNbM<m2j0X4XO^H1k0n0k1eNeM<m2h0X4XO\\H9l0l0i1aNhM<l2g0Z45UI=h1[NlM=l2g0Z4<nH;k1VNPN<l2h0Z4b0hH:m1PNTN=l2h0Z4f0dH:n1lMVN>l2g0[4\\3^J_KZNb0i2d0^4`3[J[K]Nb0i2c0^4f3ZIUKL1Ab0j2c0^4j3RI[KOICa0l2b0_4o3iH`K4a08W2j6n8O1O1O1O1O10000000000O1000000000000O1000000000000000ZNZIRFf6f9cIYF]6a9kI]FU6^9RJ`Fn5[9aJ[F_5`9kJ[FU5c9PKZFP5d9TKZFl4d9YKYFg4e9^KUFe4j9i2N2M3L4K5G9G9G9B?O0O2N2O1N101N2N2O0O2N2O1N3M2O2M2O2M3M2O2M3M3N2M7J1N2O2M2O1N2O1N2O1N2O1N10O1O02O2M2O2M100O2O0O100O10O0O2003L4M1N2O1N101N010O013L4M1N2O1O1N2O1L3N3M110O000aKk@U4T?dKUAZ4Y?O101N1O2O0O2N101N1O2O0O1O2O0O2N1O1O1O1O2N1O1O1O10000O100N3N1O10000001O000000001O01O07J7I6I8I7H6K4K5L7I2M2O2M2N3L3N2M3N2O1O1O00000O1O001O10000O1000000O100O1O101O0000001O0O10001O0O101O1O0O2O0100O10000`MgBVNZ=a1nB`NR=X1UChNj1PNY8o2VFQO^1TN[8c2^FYOT1YN\\8W2hF@h0]N`8j1PGI?]Nc8b1UG25^Ng8W1\\G<J^Nl8m0bGe0@`Nm8b0mGo0SO`NQ98UHX1eNeNV9EdHf1QNhN[9WOoD6f3k1lMlN_9SOoD5g3m1gMmNd9POoD6f3n1dMnNg9mNQE5e3P2aMoNi9mNQE3f3R2]MPOl9jNSIU2PMQOV>P1gAROY>o0dARO]>o0aARO_>o0_AQOc>o0[AQOf>P1XAPOj>P1TAPOm>Q1PAPOQ?R1l@nNV?R1h@nNY?T1d@lN^?T1`@lNa?X1X@jNj?e23M2M4M3M2N4L3M3M3M3M4L3M3M3M3M4K4M3M3M4L3M4L3M2N2N3N1Nd^O"}, "label": "the tennis player"}]}
{"id": 550911, "image": "COCO_train2014_000000550911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in an aqua shirt sitting in the stands\"."}], "task": "refering", "answer_template": "The \"woman in an aqua shirt sitting in the stands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 18, 19, 20, 21, 33, 34, 35, 36, 49, 50, 51, 52, 64, 65, 66, 78, 79, 80, 81, 93, 94, 96], "bbox": [0.23749414519906323, 0.00271875, 0.4746135831381733, 0.29259375000000004], "iscrowd": 0, "area": 11341.960550000002, "rle": {"size": [640, 427], "counts": "Zio1391jb0<V]OGfb0:Z]OIcb07]]OM]3C[;`0XA0V3Db;<XA4l2Fl;6XA7a2KV<OYA9W2O`<HYA=k13l<@YA`0a17V=YOYAc0V1<`=ROZAf0k0?k=kNZAi0c0a0S>fNZAP1<:Z>fNZAZ120^>lN_AU13OU>UOhAl020n=\\OPBd020n=\\OPBd020n=\\OPBd020n=\\OPBd020n=\\OPBe01KS>@lAe01GW>DhAe01C[>HdAe01_O_>L`Ae01ZOd>1[Ae01VOh>5WAe01ROl>9SAe00oNQ?<o@f0OSOm>7TAf0OXOh>2YAf0O^Ob>L_Af0OC]>GdAf0OHX>BiAf0OHX>BiAf0OHX>BhAg00FY>CgAh0OEZ>CgAh0OEZ>CgAh0O[Od>M]Ah0NROo>6SAh0NhNY?`0i@h0N^Nc?j0_@h0NSNn?U1T@f0X`0ZOh_ONPa02P_OWOga0i0h0000000000000000^]OSOQb0m0o]OAba0`0^^OOSa01m^O=e`0C[_Oa0?UN^?Z1S@b0>^NU?P1^@a0>gNk>h0g@a0>QOa>>QAa0>ZOX>5ZAa0>[OW>4[Aa0>[OW>4[Aa0>[OW>3\\Ab0=[OW>3\\Ab0=[OW>3\\Ab0=[OW>3\\Ab0=[OW>3\\Ab0>ZOV>4\\Ab0>ZOV>4]Aa0=[OV>4]Aa0=[OV>4]Aa0=[OV>]OWA36V1<ZOW>[OYA54V1<ZOW>ZOZA63V1<ZOW>XO\\A81V1<ZOW>VO^A:OV1<ZOW>TO`A<MV1<ZOW>ROcA=JW1=YOV>ROdA>IW1=YOV>POfA`0HV1<ZOV>nNhAb0FV1<ZOV>lNjAd0DV1<ZOk>@i@V1<ZOk>@i@V1<ZOk>@i@V1<ZOl>_Oh@W1<ZOl>_Oh@X1;YOm>_Oh@X1;YOm>_Oh@X1<YOR?XOb@_1<YO\\?nNX@i1<SOn?l0R@nNW`0o0i_OkNba00^^OJob0J6JRl[4"}, "label": "woman in an aqua shirt sitting in the stands"}]}
{"id": 354318, "image": "COCO_train2014_000000354318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person ' s hand cutting a piece of cake\"."}], "task": "refering", "answer_template": "The \"a person ' s hand cutting a piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 153, 154, 155], "bbox": [0.59221875, 0.006744730679156909, 1.0, 0.42248243559718973], "iscrowd": 0, "area": 28678.2545, "rle": {"size": [427, 640], "counts": "YUn42Y=2M4L3N1N1O100O1O010O1O100O1O10O01O100O1O100O00100O1O100O00100O1O1cDUOZ:k0fEVOX:k0gEWOW:i0jEXOU:h0jEYOU:h0kEYOS:h0lEZOS:f0mE[OQ:f0oEZOP:g0oEZOQ:e0PF[Oo9f0PF[Oo9f0QFZOo9e0QF\\On9e0RF\\Ol9d0TF]Ol9c0TF]Ok9d0TF]Ok9c0VF]Oj9c0UF^Oj9b0WF_Og9b0XF_Oh9`0YF@f9a0YF@f9`0[F@e9`0ZFAe9`0[F@d9`0\\FBc9>]FBb9>^FCa9>_FBa9=_FDm8XOgFU1<Cj8[OjFR1<Cg8_OkFn0>Ec8@nFl0?D`8CQGh0?F]8ESGf0`0EZ8\\1fGeNW8^1iGbNT8`1lGaNQ8b1oG_Nm7d1RH]Nk7e1VH[Ng7h1XHYNe7i1\\HWNa7l1^HUN_7m1bHSN[7P2dHQNY7Q2hHoMU7T2jHmMR7W2nHiMo6Y2QIhMm6Z2SIfMm6Y2SIgMn6Y2RIgMn6X2RIiMn6W2RIiMn6V2RIkMn6U2QIlMo6T2PImMP7S2PImMP7S2oHmMR7T2lHmMT7S2kHnMU7S2jHmMV7S2iHnMW7R2hHoMX7R2fHoMZ7Q2eHPN[7Q2dHoM\\7Q2cHoM^7R2`HoM`7Q2_HPNa7Q2^HoMb7Q2]HPNc7Q2[HPNe7P2ZHQNf7P2YHPNg7P2XHPNi7Q2UHPNk7P2THQNl7g3000000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O1O2N100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1@hElMY:W1eFgN\\9;cGDY8"}, "label": "a person ' s hand cutting a piece of cake"}]}
{"id": 354318, "image": "COCO_train2014_000000354318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hands of the person cutting the cake\"."}], "task": "refering", "answer_template": "The \"the hands of the person cutting the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 153, 154, 155], "bbox": [0.59221875, 0.006744730679156909, 1.0, 0.42248243559718973], "iscrowd": 0, "area": 28678.2545, "rle": {"size": [427, 640], "counts": "YUn42Y=2M4L3N1N1O100O1O010O1O100O1O10O01O100O1O100O00100O1O100O00100O1O1cDUOZ:k0fEVOX:k0gEWOW:i0jEXOU:h0jEYOU:h0kEYOS:h0lEZOS:f0mE[OQ:f0oEZOP:g0oEZOQ:e0PF[Oo9f0PF[Oo9f0QFZOo9e0QF\\On9e0RF\\Ol9d0TF]Ol9c0TF]Ok9d0TF]Ok9c0VF]Oj9c0UF^Oj9b0WF_Og9b0XF_Oh9`0YF@f9a0YF@f9`0[F@e9`0ZFAe9`0[F@d9`0\\FBc9>]FBb9>^FCa9>_FBa9=_FDm8XOgFU1<Cj8[OjFR1<Cg8_OkFn0>Ec8@nFl0?D`8CQGh0?F]8ESGf0`0EZ8\\1fGeNW8^1iGbNT8`1lGaNQ8b1oG_Nm7d1RH]Nk7e1VH[Ng7h1XHYNe7i1\\HWNa7l1^HUN_7m1bHSN[7P2dHQNY7Q2hHoMU7T2jHmMR7W2nHiMo6Y2QIhMm6Z2SIfMm6Y2SIgMn6Y2RIgMn6X2RIiMn6W2RIiMn6V2RIkMn6U2QIlMo6T2PImMP7S2PImMP7S2oHmMR7T2lHmMT7S2kHnMU7S2jHmMV7S2iHnMW7R2hHoMX7R2fHoMZ7Q2eHPN[7Q2dHoM\\7Q2cHoM^7R2`HoM`7Q2_HPNa7Q2^HoMb7Q2]HPNc7Q2[HPNe7P2ZHQNf7P2YHPNg7P2XHPNi7Q2UHPNk7P2THQNl7g3000000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O1O2N100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1@hElMY:W1eFgN\\9;cGDY8"}, "label": "the hands of the person cutting the cake"}]}
{"id": 534543, "image": "COCO_train2014_000000534543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stack of empty paper plates on a table iwth a blue cover\"."}], "task": "refering", "answer_template": "The \"a stack of empty paper plates on a table iwth a blue cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 60, 61, 62, 63, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 160, 161, 162, 163], "bbox": [0.0025263157894736842, 0.085390625, 0.2977368421052632, 0.37528125], "iscrowd": 0, "area": 21772.0338, "rle": {"size": [640, 570], "counts": "Zk06ma0m1RNn1YNg1000O1000000000000O1000000000000O1000000000000O10000000O1000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O3N2N2N2N2N1O2M3N2N2N2N1O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2M3N1O2N2N2N2N2M3N1O2N2N2N2M3N1O2N2N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N1O2M3N2N2N2N1O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2M3N1O2N2NYii7"}, "label": "a stack of empty paper plates on a table iwth a blue cover"}]}
{"id": 534543, "image": "COCO_train2014_000000534543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty plate sitting net to a plate with chocolate cake\"."}], "task": "refering", "answer_template": "The \"an empty plate sitting net to a plate with chocolate cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 60, 61, 62, 63, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 160, 161, 162, 163], "bbox": [0.0025263157894736842, 0.085390625, 0.2977368421052632, 0.37528125], "iscrowd": 0, "area": 21772.0338, "rle": {"size": [640, 570], "counts": "Zk06ma0m1RNn1YNg1000O1000000000000O1000000000000O1000000000000O10000000O1000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O3N2N2N2N2N1O2M3N2N2N2N1O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2M3N1O2N2N2N2N2M3N1O2N2N2N2M3N1O2N2N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N1O2M3N2N2N2N1O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2M3N1O2N2NYii7"}, "label": "an empty plate sitting net to a plate with chocolate cake"}]}
{"id": 296984, "image": "COCO_train2014_000000296984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep with yellow tags in its ears that is holding its ears up higher than the other\"."}], "task": "refering", "answer_template": "The \"a sheep with yellow tags in its ears that is holding its ears up higher than the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 121, 122, 123, 124, 145], "bbox": [0.283875, 0.0021458333333333334, 0.9983906250000001, 0.36989583333333337], "iscrowd": 0, "area": 50679.71370000001, "rle": {"size": [480, 640], "counts": "S^e2;d>6I8H7I7I5K2N3M2O2M3M2N3M3M2gNZNVEi1g:Z1M2M4WN]LQHe3o7`LjGc3U8cLbGa3]8dL\\G^3d8bLZGa3e8`LWGh1ZO@^9iNUGi1^O\\O_9kNRGi1@[O^9mNQGh1AZO_9oNoFg1CYO^9QOnFf1EWO^9TOlFe1GVO]9UOlFe1GUO^9WOjFd1ITO]9YOhFd1LQO]9\\OfFc1NPO\\9^OeFb1OoN]9@cFa11nN\\9BaFa14kN\\9E]Fb18hN[9F[Fd1:eN]9[2dFdM\\9\\2eFbM\\9^2eFaM[9_2eF`M\\9`2eF_M[9a2fF^MZ9b2gF\\MZ9d2fF\\MZ9d2gF[MY9e2hFYMZ9f2gFYM^9b2bF^Mb9^2_F`Mg9[2ZFdMf9\\2[FcMe9]2[FcMf9\\2[FbMf9^2[FaMf9^2[FaMe9_2[F`Mf9`2[F_Mf9`2[F_Me9a2\\F]Me9c2[F]Mf9b2[F]Me9c2\\F[Mf9d2[F[Me9e2[F[Me9e2\\FYMf9f2[FYMe9g2\\FXMd9h2]FVMe9i2[FWMe9i2\\FVMe9i2\\FUMe9k2X101O00001OO10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100000000000000000000000000000000000000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000000000O10000000000000000O1000000000000000000O10000000000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O1000000000001O0O1000000000000O1000000000000O1000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1001O1O2N1O1O1O1O2N1O1O1O1O00O100O10000O100O100O100O10`1_N_="}, "label": "a sheep with yellow tags in its ears that is holding its ears up higher than the other"}]}
{"id": 296984, "image": "COCO_train2014_000000296984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep behind the other\"."}], "task": "refering", "answer_template": "The \"the sheep behind the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 121, 122, 123, 124, 145], "bbox": [0.283875, 0.0021458333333333334, 0.9983906250000001, 0.36989583333333337], "iscrowd": 0, "area": 50679.71370000001, "rle": {"size": [480, 640], "counts": "S^e2;d>6I8H7I7I5K2N3M2O2M3M2N3M3M2gNZNVEi1g:Z1M2M4WN]LQHe3o7`LjGc3U8cLbGa3]8dL\\G^3d8bLZGa3e8`LWGh1ZO@^9iNUGi1^O\\O_9kNRGi1@[O^9mNQGh1AZO_9oNoFg1CYO^9QOnFf1EWO^9TOlFe1GVO]9UOlFe1GUO^9WOjFd1ITO]9YOhFd1LQO]9\\OfFc1NPO\\9^OeFb1OoN]9@cFa11nN\\9BaFa14kN\\9E]Fb18hN[9F[Fd1:eN]9[2dFdM\\9\\2eFbM\\9^2eFaM[9_2eF`M\\9`2eF_M[9a2fF^MZ9b2gF\\MZ9d2fF\\MZ9d2gF[MY9e2hFYMZ9f2gFYM^9b2bF^Mb9^2_F`Mg9[2ZFdMf9\\2[FcMe9]2[FcMf9\\2[FbMf9^2[FaMf9^2[FaMe9_2[F`Mf9`2[F_Mf9`2[F_Me9a2\\F]Me9c2[F]Mf9b2[F]Me9c2\\F[Mf9d2[F[Me9e2[F[Me9e2\\FYMf9f2[FYMe9g2\\FXMd9h2]FVMe9i2[FWMe9i2\\FVMe9i2\\FUMe9k2X101O00001OO10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100000000000000000000000000000000000000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000000000O10000000000000000O1000000000000000000O10000000000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O1000000000001O0O1000000000000O1000000000000O1000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1001O1O2N1O1O1O1O2N1O1O1O1O00O100O10000O100O100O100O10`1_N_="}, "label": "the sheep behind the other"}]}
{"id": 296984, "image": "COCO_train2014_000000296984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the animal with both its left and right eye visible\"."}], "task": "refering", "answer_template": "The \"the animal with both its left and right eye visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 91, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 320, 321, 337, 338, 339, 340, 341, 360, 361, 362, 363, 385, 386], "bbox": [0.319359375, 0.22795833333333335, 1.0, 0.9978541666666667], "iscrowd": 0, "area": 92442.08049999995, "rle": {"size": [480, 640], "counts": "iVP31o>1O1O1O1O1O1O1O001O1N101O001O001O001O1O001O001O001O001O1O001O001O00000000000000000000000000000nBAc;?[DNZ;2dD7U;IiD;U;EjD?S;AkDd0R;\\OlDi0R1fNk7a0RGm0f0oNU84SGQ1=WO]8HTGV12_Og8[OVGZ1IEn8QOWG_1EER9lNWGc1BGT9fNYGl3d8TLZGP4d8PLZGQ4g8nKXGT4h8lKTGY4k8gKQG^4n8k0O001O1O1O001O1O001O1O1O00000000000000000000000000000000000000000000000000000000000000O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N3N1N2N2O2O0000000O2O0000001O0000001O0000001N100O101O0O100O10000O10000O100O101O0O100O100001O2N2N2N3M2N2N3M2N2N3M2N1O2N1O101N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O101N1O001O001O001O001O001O001O010gJaIo1_6PNbIP2_6mMdIR2\\6mMeIS2\\6jMgIU2Y6jMhIV2Y6gMjIX2V6gMkIY2V6dMmI[2T6cMmI]2S6aMPJ^2Q6`MPJ_2Q6_MRJ`2o5^MRJb2n5\\MUJb2m5\\MTJd2n5XMUJf2P6UMQJk2S6oLoIQ3U6jLlIU3X6fLjIZ3_6[LcId3l6mKUIS4Y7]KiHc4j8O100O10000O100O10000O100O10O10O10000O10O01000K4J7I6I8I7I6J_J]GS5]9K4ULTFY2R:]MSFc2Q:TMTFl2Q:kLSFU3Q:bLTF^3g:0001O0000000O10000000000000001O00000O100000O1O1O1N2O1M3N2N1O2N2N2N2M3N2N2N2N2N2M2O2M3M3M3N2M3M3M3M3N3L3M3M3M3N2L4L4L4L4L4L4L4L4L4M31O0000000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O000000000O2i0VOU1kNU1kNU1kNU1kNT1mNT1kNlD"}, "label": "the animal with both its left and right eye visible"}]}
{"id": 296984, "image": "COCO_train2014_000000296984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb posing for the camera with eyes wide open\"."}], "task": "refering", "answer_template": "The \"a lamb posing for the camera with eyes wide open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 91, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 320, 321, 337, 338, 339, 340, 341, 360, 361, 362, 363, 385, 386], "bbox": [0.319359375, 0.22795833333333335, 1.0, 0.9978541666666667], "iscrowd": 0, "area": 92442.08049999995, "rle": {"size": [480, 640], "counts": "iVP31o>1O1O1O1O1O1O1O001O1N101O001O001O001O1O001O001O001O001O1O001O001O00000000000000000000000000000nBAc;?[DNZ;2dD7U;IiD;U;EjD?S;AkDd0R;\\OlDi0R1fNk7a0RGm0f0oNU84SGQ1=WO]8HTGV12_Og8[OVGZ1IEn8QOWG_1EER9lNWGc1BGT9fNYGl3d8TLZGP4d8PLZGQ4g8nKXGT4h8lKTGY4k8gKQG^4n8k0O001O1O1O001O1O001O1O1O00000000000000000000000000000000000000000000000000000000000000O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N3N1N2N2O2O0000000O2O0000001O0000001O0000001N100O101O0O100O10000O10000O100O101O0O100O100001O2N2N2N3M2N2N3M2N2N3M2N1O2N1O101N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O101N1O001O001O001O001O001O001O010gJaIo1_6PNbIP2_6mMdIR2\\6mMeIS2\\6jMgIU2Y6jMhIV2Y6gMjIX2V6gMkIY2V6dMmI[2T6cMmI]2S6aMPJ^2Q6`MPJ_2Q6_MRJ`2o5^MRJb2n5\\MUJb2m5\\MTJd2n5XMUJf2P6UMQJk2S6oLoIQ3U6jLlIU3X6fLjIZ3_6[LcId3l6mKUIS4Y7]KiHc4j8O100O10000O100O10000O100O10O10O10000O10O01000K4J7I6I8I7I6J_J]GS5]9K4ULTFY2R:]MSFc2Q:TMTFl2Q:kLSFU3Q:bLTF^3g:0001O0000000O10000000000000001O00000O100000O1O1O1N2O1M3N2N1O2N2N2N2M3N2N2N2N2N2M2O2M3M3M3N2M3M3M3M3N3L3M3M3M3N2L4L4L4L4L4L4L4L4L4M31O0000000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O000000000O2i0VOU1kNU1kNU1kNU1kNT1mNT1kNlD"}, "label": "a lamb posing for the camera with eyes wide open"}]}
{"id": 133145, "image": "COCO_train2014_000000133145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fire truck next to the sidewalk with an american flag grill\"."}], "task": "refering", "answer_template": "The \"a fire truck next to the sidewalk with an american flag grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.13314062499999998, 0.017979166666666668, 0.883140625, 0.8696666666666667], "iscrowd": 0, "area": 154679.35534999997, "rle": {"size": [480, 640], "counts": "[UX17g>h0YOe0ZO2O1O1O10000O10000O1000000O2O000O10000O10000O1000000O10001N10000O10000O1000000O10000O10001N1000000O10000O1N2N2O1N2UOk0VKbMSL^OeMQ3l5cNaKjN\\Ne2f5TOdKaN^N\\2k5WOdKR2[4oMcKS2]4mMaKU2_4kM`KV2`4kM]KW2c4iM[KY2e4gMYK[2f4P400000000kNSGWLm8b3_GSLh8j3_GjKh8V4Q1000000000000O10000I7I7N201O00000000000O100000007I7I3M0000001O0000001O000O101O0000001O0000000000000000000O101O000000000000000000000000001O0O10000000000000000000001O000000000O10000000000000001O0000000000000O10000000001O000000000000000000000O10001O0000000000000000000000001N100000000000000000000000001O00000O1000000000000000001O0000000000000O1000000000001O0000000000000000FaMSD^2P<dMlC\\2V<62N2N2N00000000000000O10000000OEaMWD_2i;bMVD^2j;bMVD^2j;cMTD^2l;cMSD]2m;cMSD]2m;dMRD\\2n;eMQD[2o;eMQD[2o;fMPDZ2P<gMoCY2Q<hMmCY2S<gMmCY2S<:00000O1000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000O100000O100000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000000O10003M4L4L4L1000O100O101N10000O10O2N1O1O2N1O1N3N1On0RO000L3M4M3L3N3hJaKKc45_KaK2Y11Ea4b3]K]Ki2`0m1S4[K[KW30b1e4XKXKi8h4YGUKi8e4l0dNhEmMZ:S2YFWMk9h2W1O0O2O1O10001O01O01O00001O000010O01O00<Dh0XOg0YOh0XOVPS1"}, "label": "a fire truck next to the sidewalk with an american flag grill"}]}
{"id": 133145, "image": "COCO_train2014_000000133145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a firetruck with an american flag grill\"."}], "task": "refering", "answer_template": "The \"the front of a firetruck with an american flag grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.13314062499999998, 0.017979166666666668, 0.883140625, 0.8696666666666667], "iscrowd": 0, "area": 154679.35534999997, "rle": {"size": [480, 640], "counts": "[UX17g>h0YOe0ZO2O1O1O10000O10000O1000000O2O000O10000O10000O1000000O10001N10000O10000O1000000O10000O10001N1000000O10000O1N2N2O1N2UOk0VKbMSL^OeMQ3l5cNaKjN\\Ne2f5TOdKaN^N\\2k5WOdKR2[4oMcKS2]4mMaKU2_4kM`KV2`4kM]KW2c4iM[KY2e4gMYK[2f4P400000000kNSGWLm8b3_GSLh8j3_GjKh8V4Q1000000000000O10000I7I7N201O00000000000O100000007I7I3M0000001O0000001O000O101O0000001O0000000000000000000O101O000000000000000000000000001O0O10000000000000000000001O000000000O10000000000000001O0000000000000O10000000001O000000000000000000000O10001O0000000000000000000000001N100000000000000000000000001O00000O1000000000000000001O0000000000000O1000000000001O0000000000000000FaMSD^2P<dMlC\\2V<62N2N2N00000000000000O10000000OEaMWD_2i;bMVD^2j;bMVD^2j;cMTD^2l;cMSD]2m;cMSD]2m;dMRD\\2n;eMQD[2o;eMQD[2o;fMPDZ2P<gMoCY2Q<hMmCY2S<gMmCY2S<:00000O1000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000O100000O100000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000000O10003M4L4L4L1000O100O101N10000O10O2N1O1O2N1O1N3N1On0RO000L3M4M3L3N3hJaKKc45_KaK2Y11Ea4b3]K]Ki2`0m1S4[K[KW30b1e4XKXKi8h4YGUKi8e4l0dNhEmMZ:S2YFWMk9h2W1O0O2O1O10001O01O01O00001O000010O01O00<Dh0XOg0YOh0XOVPS1"}, "label": "the front of a firetruck with an american flag grill"}]}
{"id": 559132, "image": "COCO_train2014_000000559132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bride in wedding dress holding white flowers while walking with groom\"."}], "task": "refering", "answer_template": "The \"bride in wedding dress holding white flowers while walking with groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 63, 64, 65, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 226, 227, 228, 229, 230, 231, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 287, 288, 289, 290, 291, 302, 303, 304, 305, 306, 317, 318, 319, 320], "bbox": [0.09816470588235293, 0.1373125, 0.4989647058823529, 0.9280000000000002], "iscrowd": 0, "area": 56841.61185000001, "rle": {"size": [640, 425], "counts": "\\ej04hc07J6I7J6I7J6I7J6I7J6I7J6I7J6I7QB\\MY:j2PEjMm:\\2\\DXN`;o1jCdNS<b1VCROf<V4M3L4M4K4M3L4M3L4M3lLPHTJT8f5RHWJQ8d5THXJP8b5WHZJl7`5ZH\\Jj7^5]H^Jf7]5_H_Je7[5aHbJb7\\5aHcJ_7[5cHgJ[7X5fHjJY7T5iHlJV7S5kHoJS7P5nHRKP7m4RITKl6k4UIVKj6i4WIYKg6f4[I[Kc6d4^I^K`6a4aI`K^6_4dIbKZ6\\4hIeKW6Z4kIdKV6[4kIeKU6Z4lIeKU6Z4mIeKS6Z4nIfKS6X4nIgKS6X4UJaKk5\\4`J[Ka5`4lJXKT5c4YKTKh4f4fKRKZ4i4a5J6K5K5J6K5J6K5N2O101N1O100O1O100O100O1O100O1O100O101O01O00001O001O00001N10001O0O101O001N10001O0O101O001N10001O0O2O4L3M4K5L4L4L4K5L3M4K5G9F:G9G8H6I7J6J5J7J6J6J6I6dHeIn0a6oNbIk0d6QO`Ii0f6SO_Id0h6YO\\I?l6^OWI9R7CRI4W7HmH0Z7MjHJ_73eHDc78aH_Oh7=\\HZOl7c0XHUOR8f0QHQOY8j0lGmN^8m0fGjNe8P1_GgNl8T1WGdNR9W1P5J6J6K4K6J6J6JbmT4"}, "label": "bride in wedding dress holding white flowers while walking with groom"}]}
{"id": 559132, "image": "COCO_train2014_000000559132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a pic of married couple\"."}], "task": "refering", "answer_template": "The \"woman in a pic of married couple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 63, 64, 65, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 226, 227, 228, 229, 230, 231, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 287, 288, 289, 290, 291, 302, 303, 304, 305, 306, 317, 318, 319, 320], "bbox": [0.09816470588235293, 0.1373125, 0.4989647058823529, 0.9280000000000002], "iscrowd": 0, "area": 56841.61185000001, "rle": {"size": [640, 425], "counts": "\\ej04hc07J6I7J6I7J6I7J6I7J6I7J6I7J6I7QB\\MY:j2PEjMm:\\2\\DXN`;o1jCdNS<b1VCROf<V4M3L4M4K4M3L4M3L4M3lLPHTJT8f5RHWJQ8d5THXJP8b5WHZJl7`5ZH\\Jj7^5]H^Jf7]5_H_Je7[5aHbJb7\\5aHcJ_7[5cHgJ[7X5fHjJY7T5iHlJV7S5kHoJS7P5nHRKP7m4RITKl6k4UIVKj6i4WIYKg6f4[I[Kc6d4^I^K`6a4aI`K^6_4dIbKZ6\\4hIeKW6Z4kIdKV6[4kIeKU6Z4lIeKU6Z4mIeKS6Z4nIfKS6X4nIgKS6X4UJaKk5\\4`J[Ka5`4lJXKT5c4YKTKh4f4fKRKZ4i4a5J6K5K5J6K5J6K5N2O101N1O100O1O100O100O1O100O1O100O101O01O00001O001O00001N10001O0O101O001N10001O0O101O001N10001O0O2O4L3M4K5L4L4L4K5L3M4K5G9F:G9G8H6I7J6J5J7J6J6J6I6dHeIn0a6oNbIk0d6QO`Ii0f6SO_Id0h6YO\\I?l6^OWI9R7CRI4W7HmH0Z7MjHJ_73eHDc78aH_Oh7=\\HZOl7c0XHUOR8f0QHQOY8j0lGmN^8m0fGjNe8P1_GgNl8T1WGdNR9W1P5J6J6K4K6J6J6JbmT4"}, "label": "woman in a pic of married couple"}]}
{"id": 559132, "image": "COCO_train2014_000000559132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man in a suit holding hands with a bride\"."}], "task": "refering", "answer_template": "The \"a smiling man in a suit holding hands with a bride\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 54, 55, 56, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 171, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 264, 265, 266, 267, 279, 280, 281, 282, 294, 295, 296, 309, 310, 311, 325, 326, 340], "bbox": [0.43520000000000003, 0.10521875, 0.9370117647058824, 0.972609375], "iscrowd": 0, "area": 67067.63994999995, "rle": {"size": [640, 425], "counts": "mmc3P1]b0T1WO9I7J6I7K6L4L5I5K5K6J5K5K6J5L3MJl@fKR?\\4PAcKk>`4WA_Ke>e4]AYK_>j4dATKW>S565E;^O]JbBe5\\=bJ]B`5Y=i0O_Oa0O1O1001000O100J6J6L3M4_NkATMX>j2XBeLk=Z3^1O00l_OoL[?[4_O9kARKh<S5iB^Ko<P6I8Ig0XO5L4L3L5L4L3L5L4L3L5L4L3M3L3N2N5K5J5L5J6cNkEeHY:U7nEfHW:S7TFcHR:R7VE`Hn09P:k6]FTIi9f6XFZIm9`6UF^IP:\\6RFcIR:_1\\DR3d1^KT:W1_DX3b1]KS:X1[DY3e=dL\\B\\3g=aLZB]3i=`LXB_3k=_LUB`3n=]LTBb3n=\\LRBc3Q>ZLPBe3S>YLmAf3V>WLkAi3W>TLjAk3Y>SLgAi3^>ULcAj3`>TL`Ak3c>RL^An3d>mK_AR4X?0000O107I9G8H6J1O2N101N1O1O101N1O00100O1O01O00000O2O0N2O2N1N3N1O2O1N1o@lLV=U3gBnLf<3nAP3Y1PMe<7mAk2[1QMe<e3YC^Ld<e3VCaLg<b3RCeLk<`3mBgLo<]3kBgLS=f3^B_L_=R5N3M2N2L5L3M3M4L3L5L3M3L5F9I7N3M2N3M2N2N3I6J7J5K5J7J5K5K6J5K6J5K5D=\\Oc0D<D=@?I7DiBhI_=U67N2M310O0O1O2N1O1N3N1O1O1O2M2O2N3M4SK`AR4e>hKaAT4Z?J7I7I7H8I7A?H8H7H7J6J7A?YOhd`0"}, "label": "a smiling man in a suit holding hands with a bride"}]}
{"id": 559132, "image": "COCO_train2014_000000559132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man\"."}], "task": "refering", "answer_template": "The \"the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 54, 55, 56, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 171, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 264, 265, 266, 267, 279, 280, 281, 282, 294, 295, 296, 309, 310, 311, 325, 326, 340], "bbox": [0.43520000000000003, 0.10521875, 0.9370117647058824, 0.972609375], "iscrowd": 0, "area": 67067.63994999995, "rle": {"size": [640, 425], "counts": "mmc3P1]b0T1WO9I7J6I7K6L4L5I5K5K6J5K5K6J5L3MJl@fKR?\\4PAcKk>`4WA_Ke>e4]AYK_>j4dATKW>S565E;^O]JbBe5\\=bJ]B`5Y=i0O_Oa0O1O1001000O100J6J6L3M4_NkATMX>j2XBeLk=Z3^1O00l_OoL[?[4_O9kARKh<S5iB^Ko<P6I8Ig0XO5L4L3L5L4L3L5L4L3L5L4L3M3L3N2N5K5J5L5J6cNkEeHY:U7nEfHW:S7TFcHR:R7VE`Hn09P:k6]FTIi9f6XFZIm9`6UF^IP:\\6RFcIR:_1\\DR3d1^KT:W1_DX3b1]KS:X1[DY3e=dL\\B\\3g=aLZB]3i=`LXB_3k=_LUB`3n=]LTBb3n=\\LRBc3Q>ZLPBe3S>YLmAf3V>WLkAi3W>TLjAk3Y>SLgAi3^>ULcAj3`>TL`Ak3c>RL^An3d>mK_AR4X?0000O107I9G8H6J1O2N101N1O1O101N1O00100O1O01O00000O2O0N2O2N1N3N1O2O1N1o@lLV=U3gBnLf<3nAP3Y1PMe<7mAk2[1QMe<e3YC^Ld<e3VCaLg<b3RCeLk<`3mBgLo<]3kBgLS=f3^B_L_=R5N3M2N2L5L3M3M4L3L5L3M3L5F9I7N3M2N3M2N2N3I6J7J5K5J7J5K5K6J5K6J5K5D=\\Oc0D<D=@?I7DiBhI_=U67N2M310O0O1O2N1O1N3N1O1O1O2M2O2N3M4SK`AR4e>hKaAT4Z?J7I7I7H8I7A?H8H7H7J6J7A?YOhd`0"}, "label": "the man"}]}
{"id": 92197, "image": "COCO_train2014_000000092197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man doing skateboard jump\"."}], "task": "refering", "answer_template": "The \"man doing skateboard jump\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 92, 93, 94, 95, 96, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 152, 153, 154, 155, 156, 157, 171, 172], "bbox": [0.12352112676056337, 0.232078125, 0.5900469483568076, 0.49065624999999996], "iscrowd": 0, "area": 17672.828999999998, "rle": {"size": [640, 426], "counts": "jZQ1`0Vc0=L4L4K5L3M4L4L4L3L5L3M3M4L3POcMX@a2e?eMS@_2j?gMP@\\2m?kMl_OY2P`0mMj_OV2T`0m0M4L3M3N3L3M3M2O0O2N101O0O2O001O00000O1000[Ok@_LU?a3o@ZLR?g3QAULo>l3TAPLk>Q4YAkKg>V4a01O00000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O01O01O01O001O010O000010O01O000010O01O00010O001O01POi_OVNV`0h1l_OXNT`0f1n_OZNR`0d1Q@[No?c1S@^Nm?_1U@aNk?]1X@bNh?]1Y@dNf?Z1\\@fNe?W1^@hNb?V1`@jN`?T1b@mN]?o0h@POX?k0m@UOT?e0QA\\On>c0TA\\Ol>d0TA\\Ol>d0TA\\Om>d0SA\\Ol>d0TA\\Ol>d0TA\\Ol>d0UA\\Oj>d0VA\\Ok>c0UA]Ok>d0UA[Ok>e0UA\\Ok>c0UA]Ok>c0VA\\Ok>c0UA^Oj>b0VA^Oj>b0WA]Oj>b0VA^Oj>b0VA_Oj>`0VA@j>`0VA@j>`0VAAj>>VABj>>VABk>=UACk>=UADj><VADk>;UAEk>;UAFj>:VAFk>9UAGk>9WAEj>:WAFh>:YAEg>;[ACf><[ACe>=]ABc>=^ABb>>`A@`>`0aA_O`>`0aA@^>`0aAA`>>`AB`>>_AD`><`ADa>;^AFb>:^AFb>:]AGc>9]AGc>8^AHb>8]AIc>7]AIc>7\\AIe>7[AEi>:WABn>>RA]OT?b0k@ZOZ?f0f@UO_?k0a@QOc?o0g1O101O1O1O1N10000O0100O010O10O01000O001M3M3L3N3M3M3L3NQV]3"}, "label": "man doing skateboard jump"}]}
{"id": 92197, "image": "COCO_train2014_000000092197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in white shirt skateboarding on the ramp\"."}], "task": "refering", "answer_template": "The \"boy in white shirt skateboarding on the ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 92, 93, 94, 95, 96, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 152, 153, 154, 155, 156, 157, 171, 172], "bbox": [0.12352112676056337, 0.232078125, 0.5900469483568076, 0.49065624999999996], "iscrowd": 0, "area": 17672.828999999998, "rle": {"size": [640, 426], "counts": "jZQ1`0Vc0=L4L4K5L3M4L4L4L3L5L3M3M4L3POcMX@a2e?eMS@_2j?gMP@\\2m?kMl_OY2P`0mMj_OV2T`0m0M4L3M3N3L3M3M2O0O2N101O0O2O001O00000O1000[Ok@_LU?a3o@ZLR?g3QAULo>l3TAPLk>Q4YAkKg>V4a01O00000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O01O01O01O001O010O000010O01O000010O01O00010O001O01POi_OVNV`0h1l_OXNT`0f1n_OZNR`0d1Q@[No?c1S@^Nm?_1U@aNk?]1X@bNh?]1Y@dNf?Z1\\@fNe?W1^@hNb?V1`@jN`?T1b@mN]?o0h@POX?k0m@UOT?e0QA\\On>c0TA\\Ol>d0TA\\Ol>d0TA\\Om>d0SA\\Ol>d0TA\\Ol>d0TA\\Ol>d0UA\\Oj>d0VA\\Ok>c0UA]Ok>d0UA[Ok>e0UA\\Ok>c0UA]Ok>c0VA\\Ok>c0UA^Oj>b0VA^Oj>b0WA]Oj>b0VA^Oj>b0VA_Oj>`0VA@j>`0VA@j>`0VAAj>>VABj>>VABk>=UACk>=UADj><VADk>;UAEk>;UAFj>:VAFk>9UAGk>9WAEj>:WAFh>:YAEg>;[ACf><[ACe>=]ABc>=^ABb>>`A@`>`0aA_O`>`0aA@^>`0aAA`>>`AB`>>_AD`><`ADa>;^AFb>:^AFb>:]AGc>9]AGc>8^AHb>8]AIc>7]AIc>7\\AIe>7[AEi>:WABn>>RA]OT?b0k@ZOZ?f0f@UO_?k0a@QOc?o0g1O101O1O1O1N10000O0100O010O10O01000O001M3M3L3N3M3M3L3NQV]3"}, "label": "boy in white shirt skateboarding on the ramp"}]}
{"id": 239654, "image": "COCO_train2014_000000239654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat on the right with toilet roll in front of him\"."}], "task": "refering", "answer_template": "The \"the cat on the right with toilet roll in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 263, 264, 265, 286], "bbox": [0.443953125, 0.47168618266978923, 0.80728125, 0.8604215456674472], "iscrowd": 0, "area": 12676.504049999996, "rle": {"size": [427, 640], "counts": "ghf3<b<=O1O1O001O1O100TOSOYEm0f:UOYEl0e:VOZEi0f:XOZEg0f:[OYEe0f:]OYEb0h:_OWE`0i:BVE>i:CXE<g:FXE:g:HXE8g:JXE6g:KYE4g:NXE2g:0XE0g:1YEOg:2YEMf:4ZELe:5[EKd:7[EId:8\\EHc:9]EGb:;]EEb:<^EDa:=_ECa:=`EB`:?_EAa:?_EAa:d11O000000001O0000001O0000001O0000000010O001O010O1O00100O001O00100O001O010O1O0010O01O1O00100O001O010O1O0010O01O1000O0100O10O0100O10O0100O010O100O01000O10O0100O010O100O010O100O02O00000O101O0O10001O0O10000O2M2O1O1N3N1O1N2O1O2M2O1O101N1O1O100O2N1O100O2N1O2O4K4L5K5L0O001O0010ODgCGY<9mCAT<><10O010O0010O00010O010O01O010O00010O010O0010O01O100O1O100O100O1O100O10000O10000O101O2O1O2NO0O2O1N2O1N2O0O2ON1M4K4MjRc1"}, "label": "the cat on the right with toilet roll in front of him"}]}
{"id": 239654, "image": "COCO_train2014_000000239654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat touching toilet paper roll\"."}], "task": "refering", "answer_template": "The \"cat touching toilet paper roll\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 263, 264, 265, 286], "bbox": [0.443953125, 0.47168618266978923, 0.80728125, 0.8604215456674472], "iscrowd": 0, "area": 12676.504049999996, "rle": {"size": [427, 640], "counts": "ghf3<b<=O1O1O001O1O100TOSOYEm0f:UOYEl0e:VOZEi0f:XOZEg0f:[OYEe0f:]OYEb0h:_OWE`0i:BVE>i:CXE<g:FXE:g:HXE8g:JXE6g:KYE4g:NXE2g:0XE0g:1YEOg:2YEMf:4ZELe:5[EKd:7[EId:8\\EHc:9]EGb:;]EEb:<^EDa:=_ECa:=`EB`:?_EAa:?_EAa:d11O000000001O0000001O0000001O0000000010O001O010O1O00100O001O00100O001O010O1O0010O01O1O00100O001O010O1O0010O01O1000O0100O10O0100O10O0100O010O100O01000O10O0100O010O100O010O100O02O00000O101O0O10001O0O10000O2M2O1O1N3N1O1N2O1O2M2O1O101N1O1O100O2N1O100O2N1O2O4K4L5K5L0O001O0010ODgCGY<9mCAT<><10O010O0010O00010O010O01O010O00010O010O0010O01O100O1O100O100O1O100O10000O10000O101O2O1O2NO0O2O1N2O1N2O0O2ON1M4K4MjRc1"}, "label": "cat touching toilet paper roll"}]}
{"id": 239654, "image": "COCO_train2014_000000239654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat on the left\"."}], "task": "refering", "answer_template": "The \"the cat on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 80, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 214], "bbox": [0.291796875, 0.1722950819672131, 0.493578125, 0.6155737704918033], "iscrowd": 0, "area": 12470.466600000002, "rle": {"size": [427, 640], "counts": "UT^2i0c<`0PEiNj8Y1QGlNn8U1lFPOS9R1hFSOW9n0dFVO\\9k0_FYO`9h0[F\\Of9e0UF_Oh9k0mEXOT:o1101N1O10O0O1O2N1N2O2NG90000O100000000O02O1M3L4M4K4M3M4M2O1N2O2N1N2O2M2N2O2M2N2O2M2O2M2O1O2M2O1O2N1O2N1O1O2N1O1O2N1O2N100O2N1OK6J5M3M3M4L3M3M4F9G9K5K6J6J7J5M4M3M2N3L4M2O2N2M2N3L4L3M4L4K4L5J5M4N2O1N3N1N2O000O2O0O2O001N100O2O000O100O100K5FePW4"}, "label": "the cat on the left"}]}
{"id": 239654, "image": "COCO_train2014_000000239654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kitten standing by a toilet looking upwards\"."}], "task": "refering", "answer_template": "The \"a kitten standing by a toilet looking upwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 80, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 214], "bbox": [0.291796875, 0.1722950819672131, 0.493578125, 0.6155737704918033], "iscrowd": 0, "area": 12470.466600000002, "rle": {"size": [427, 640], "counts": "UT^2i0c<`0PEiNj8Y1QGlNn8U1lFPOS9R1hFSOW9n0dFVO\\9k0_FYO`9h0[F\\Of9e0UF_Oh9k0mEXOT:o1101N1O10O0O1O2N1N2O2NG90000O100000000O02O1M3L4M4K4M3M4M2O1N2O2N1N2O2M2N2O2M2N2O2M2O2M2O1O2M2O1O2N1O2N1O1O2N1O1O2N1O2N100O2N1OK6J5M3M3M4L3M3M4F9G9K5K6J6J7J5M4M3M2N3L4M2O2N2M2N3L4L3M4L4K4L5J5M4N2O1N3N1N2O000O2O0O2O001N100O2O000O100O100K5FePW4"}, "label": "a kitten standing by a toilet looking upwards"}]}
{"id": 419879, "image": "COCO_train2014_000000419879.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green painted stop light\"."}], "task": "refering", "answer_template": "The \"a green painted stop light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 81, 82, 83, 84, 96, 97, 98, 99, 111, 112, 113, 114, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 201, 202, 203], "bbox": [0.3790117647058824, 0.19775, 0.6734117647058825, 0.60225], "iscrowd": 0, "area": 23379.588349999995, "rle": {"size": [640, 425], "counts": "kmT33hc05K6[NE^_O`0[`0H__O>Y`0Ib_O<o>VOTAe0H;m>WOTAe0I9m>YORAe0L8k>YOVAc0K7j>\\OYA>M7c>B]A:N5d>B]A:N6c>B\\A:05c>CZA;14c>CZA:24c>CYA<23c>]1\\AdNc>]1]AdNa>^1]AcNb>_3M2O1O2N1O2N1O2N1O2Oe1[N00O101N1O100O2O0O101N101N101N100O2O001N101O0O101O0O2O001N10001N101OO010O1O010O100O00100O010O1O010O1O010O10O01O10O0100O00100O00100O010O1O010O10O01O10O01O10O0100O00100l2SM=D<D=B=D<D=C<C>C<D<C[af2"}, "label": "a green painted stop light"}]}
{"id": 419879, "image": "COCO_train2014_000000419879.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green stoplight with broadway blvd on top\"."}], "task": "refering", "answer_template": "The \"a green stoplight with broadway blvd on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 81, 82, 83, 84, 96, 97, 98, 99, 111, 112, 113, 114, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 201, 202, 203], "bbox": [0.3790117647058824, 0.19775, 0.6734117647058825, 0.60225], "iscrowd": 0, "area": 23379.588349999995, "rle": {"size": [640, 425], "counts": "kmT33hc05K6[NE^_O`0[`0H__O>Y`0Ib_O<o>VOTAe0H;m>WOTAe0I9m>YORAe0L8k>YOVAc0K7j>\\OYA>M7c>B]A:N5d>B]A:N6c>B\\A:05c>CZA;14c>CZA:24c>CYA<23c>]1\\AdNc>]1]AdNa>^1]AcNb>_3M2O1O2N1O2N1O2N1O2Oe1[N00O101N1O100O2O0O101N101N101N100O2O001N101O0O101O0O2O001N10001N101OO010O1O010O100O00100O010O1O010O1O010O10O01O10O0100O00100O00100O010O1O010O10O01O10O01O10O0100O00100l2SM=D<D=B=D<D=C<C>C<D<C[af2"}, "label": "a green stoplight with broadway blvd on top"}]}
{"id": 370728, "image": "COCO_train2014_000000370728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sliced orange facing straight up with a knife on top of it\"."}], "task": "refering", "answer_template": "The \"a sliced orange facing straight up with a knife on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 169, 170, 171, 172, 173, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 198, 199, 200, 201, 214, 215, 216, 217, 218, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 243, 244, 245, 246, 260, 261, 262, 263, 266, 267, 268, 283, 284, 285, 286, 288, 289, 290, 308], "bbox": [0.300515625, 0.24000000000000002, 0.789640625, 0.7740416666666666], "iscrowd": 0, "area": 39686.18925, "rle": {"size": [480, 640], "counts": "PZj21d><E;E;G9N2N2N2O1N2N2N2M2N3M3N2M3M3M3M3N2M3M3M3N2M3M3M3N2M3M3M3N1N3M2O2N1O101N1O2N1O2N1O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N101N1O2N1O1O2N101N1O2N1O2N1O2O0O1O000000O10001O000000000O101N1O1O2N1O1O2N1N2O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2M2O1O2N1O2N1O1N3N1O1O2N1O1O2N1N2O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2N1O101N100O2O0O2O0O101N1O2O0O2O0O2O0O2OiGmMb3R2]LQNc3n1\\LSNd3l1[LWNe3g1ZL\\Ne3d1ZL]Ng3a1XLbNg3^1WLeNi3Y1WLhNi3X1ULkNk3S1TLoNl3Q1SLQOm3m0RLVOm3j0QLXOP4f0PL\\Oo3c0PL@P4?nKCS4;mKGR49lKJT44kKNU42jK0V4NiK4W4LgK7Y4GgK;X4EfK=[4AdKb0[4^OdKd0\\4ZOcKh0]4WObKl0^4SO`Ko0`4PO^KT1b4kN\\KX1c4gN[K\\1f4cNWKa1h4^NWKe1i4ZNTKi1l4VNSKm1m4RNPKR2o4mMoJV2R5hMmJ[2R5eMkJ^2V5`MiJc2V5]MgJg2Y5WMeJl2[5TMcJo2]5oLaJU3^5iLbJY3_5cLaJa3^5]LaJg3_5ULbJm3^5QLaJS4_5iKbJY4^5dKbJ`4k74L3M4L4L3L3N3M3M3M3M3L4M3M3M3M3L4M3N12O1O1O1O2N1O1O1O1O2N1O1N2O2N1O1O1O1O2N1O1O1O1O2M2O1O1N2O2N1O1N2O2N1N2O1O1N3N1O1O1N2O2N1N2O1O1N3N1O1O1N3N1O1N2O1O2M2O1O1O1N3N1O4K5L3M4K4M4L3K6I6I:F9G9G:G8G:F9G\\So1"}, "label": "a sliced orange facing straight up with a knife on top of it"}]}
{"id": 370728, "image": "COCO_train2014_000000370728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of orange that the knife is resting on\"."}], "task": "refering", "answer_template": "The \"the piece of orange that the knife is resting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 169, 170, 171, 172, 173, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 198, 199, 200, 201, 214, 215, 216, 217, 218, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 243, 244, 245, 246, 260, 261, 262, 263, 266, 267, 268, 283, 284, 285, 286, 288, 289, 290, 308], "bbox": [0.300515625, 0.24000000000000002, 0.789640625, 0.7740416666666666], "iscrowd": 0, "area": 39686.18925, "rle": {"size": [480, 640], "counts": "PZj21d><E;E;G9N2N2N2O1N2N2N2M2N3M3N2M3M3M3M3N2M3M3M3N2M3M3M3N2M3M3M3N1N3M2O2N1O101N1O2N1O2N1O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N101N1O2N1O1O2N101N1O2N1O2N1O2O0O1O000000O10001O000000000O101N1O1O2N1O1O2N1N2O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2M2O1O2N1O2N1O1N3N1O1O2N1O1O2N1N2O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2N1O101N100O2O0O2O0O101N1O2O0O2O0O2O0O2OiGmMb3R2]LQNc3n1\\LSNd3l1[LWNe3g1ZL\\Ne3d1ZL]Ng3a1XLbNg3^1WLeNi3Y1WLhNi3X1ULkNk3S1TLoNl3Q1SLQOm3m0RLVOm3j0QLXOP4f0PL\\Oo3c0PL@P4?nKCS4;mKGR49lKJT44kKNU42jK0V4NiK4W4LgK7Y4GgK;X4EfK=[4AdKb0[4^OdKd0\\4ZOcKh0]4WObKl0^4SO`Ko0`4PO^KT1b4kN\\KX1c4gN[K\\1f4cNWKa1h4^NWKe1i4ZNTKi1l4VNSKm1m4RNPKR2o4mMoJV2R5hMmJ[2R5eMkJ^2V5`MiJc2V5]MgJg2Y5WMeJl2[5TMcJo2]5oLaJU3^5iLbJY3_5cLaJa3^5]LaJg3_5ULbJm3^5QLaJS4_5iKbJY4^5dKbJ`4k74L3M4L4L3L3N3M3M3M3M3L4M3M3M3M3L4M3N12O1O1O1O2N1O1O1O1O2N1O1N2O2N1O1O1O1O2N1O1O1O1O2M2O1O1N2O2N1O1N2O2N1N2O1O1N3N1O1O1N2O2N1N2O1O1N3N1O1O1N3N1O1N2O1O2M2O1O1O1N3N1O4K5L3M4K4M4L3K6I6I:F9G9G:G8G:F9G\\So1"}, "label": "the piece of orange that the knife is resting on"}]}
{"id": 370728, "image": "COCO_train2014_000000370728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of orange under another slice\"."}], "task": "refering", "answer_template": "The \"slice of orange under another slice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 167, 168, 169, 170, 190, 191, 192, 214], "bbox": [0.288203125, 0.19102083333333333, 0.632015625, 0.5820208333333333], "iscrowd": 0, "area": 17098.074450000004, "rle": {"size": [480, 640], "counts": "]_f21h>9G:G9F:F9H9F:G9H7I8H8I7H7I8L4K1O010O010O01O010O010O00010O010O100000000O2O0000000O101O00000O10001O00000000001O001O001O1O1O010O1O1O001O1O001O1O1O001O1O1O001O1O1O00100O1O001O10O01O1O010O1O001O10O01O00100O001O1O010O1O1O100O1O1O1O10O01O1O100O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O101N100O1O100O100O100O2N100O2O0O101N1O100O2O0O101N100O2N100O2O0O101N1O101N101N2O0O2N2O0O2O1N101N101N2N101N2O0O2O0O2N2O\\`^3"}, "label": "slice of orange under another slice"}]}
{"id": 370728, "image": "COCO_train2014_000000370728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange slice underneath everything\"."}], "task": "refering", "answer_template": "The \"the orange slice underneath everything\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 167, 168, 169, 170, 190, 191, 192, 214], "bbox": [0.288203125, 0.19102083333333333, 0.632015625, 0.5820208333333333], "iscrowd": 0, "area": 17098.074450000004, "rle": {"size": [480, 640], "counts": "]_f21h>9G:G9F:F9H9F:G9H7I8H8I7H7I8L4K1O010O010O01O010O010O00010O010O100000000O2O0000000O101O00000O10001O00000000001O001O001O1O1O010O1O1O001O1O001O1O1O001O1O1O001O1O1O00100O1O001O10O01O1O010O1O001O10O01O00100O001O1O010O1O1O100O1O1O1O10O01O1O100O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O101N100O1O100O100O100O2N100O2O0O101N1O100O2O0O101N100O2N100O2O0O101N1O101N101N2O0O2N2O0O2O1N101N101N2N101N2O0O2O0O2N2O\\`^3"}, "label": "the orange slice underneath everything"}]}
{"id": 370728, "image": "COCO_train2014_000000370728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom orange in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the bottom orange in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360, 361, 362, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.5039687500000001, 0.7095208333333333, 0.775375, 1.0], "iscrowd": 0, "area": 18653.606799999998, "rle": {"size": [480, 640], "counts": "_kg4a0T>;M3L4L4M3L4M3L4M3L4L4M3L4M3L4M3L4L4M3L4M3L4M3M3N2N2M30O01O1O1O1O1O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000O1000O10000000000O100000000O10000000000O100000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N2N1O2N2N1O2M3N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N7I6J6J7I6J7I6J7I6J6JXQS2"}, "label": "the bottom orange in the right hand picture"}]}
{"id": 264233, "image": "COCO_train2014_000000264233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing an orange sweater is smiling with her hands up\"."}], "task": "refering", "answer_template": "The \"a woman wearing an orange sweater is smiling with her hands up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 134, 135, 149, 150, 151, 152, 153, 156, 157, 158, 173, 174, 175, 176, 178, 179, 180, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 360, 361, 362, 363, 383, 384, 385, 386, 406, 407, 408, 409, 410, 429, 430, 431, 432, 433, 452, 453, 454, 455, 456], "bbox": [0.4875625, 0.19101549053356284, 0.9016875000000001, 0.959552495697074], "iscrowd": 0, "area": 40353.93594999998, "rle": {"size": [581, 640], "counts": "lTa52Pb05L4K5L5J5K5N2M3N2M3N2M3N2M4M2M3N2O1N2O1N21O5K5K5K5K4LO100O1O00100O1O1O100O001O1O100O1O1O010O1O1O100O1O001O100O1O1O10O01O1O1O100hBYNo9h1aEiN]:W1TEZOj:g0fDKX;5XD<f;EjCl0T<UO\\C\\1b<V2O2M2N2N2N2O1N3M2NO2M2N2O1N3M2O1N2O@_J]D_5`;hJ_DU5_;PK`Do4\\;WKcDg4[;_KcD_4Z;gKfDW4W;oKgDP4U;VLjDi3U;[LiDc3X;_LhD_3W;dLhD[3W;j100O00100O1O10O01aGTIe5l6WJhIZ5W6bJ^Jm4b5nJkJj4T5SKQKm4n4oJWKP5i4kJ]KT5c4gJcKY5[4cJkK\\5U4^JRLa5n3YJYLf5n3nIXLR6P4_IWL`6g70100O1OO2O0O100O2O7H8I8G9H8G8RLbGgMd8V2_GcMb8_2cGZM]8i2gGoLZ8S3kGfLU8]3nG\\LS8g3QHRLo7P4VHhKk7[4THcKl7`4SHiJWOWOf8S6SHaJAUO\\8]6SHXJJTOS8g6SHoIj8T6UGgIm8]6RG^IQ9d6oFVIT9n6kFmHX9U7k03L5L4K5L4K5L4K5L4L4K5L3L5L4K5L4K5L4L4K5L1N10]GlGV7R8jH]Hh6b7WInH\\6o6eIaIm5]6SJRJ_5l5bJcJQ5Z5oJVKc4i4]KfK]OVM[3Q7YM[Me2c2[MaMb2]2^MhM`2U2aMPN[2o1eMUNY2h1hM\\NU2c1jMbNT2[1mMiNQ2T1PNPOm1o0SNVOj1g0VN^Og1a0YNCe1<ZNHd16\\NNa11^N5_1IaN;\\1EcN?[1_OeNe0X1ZOgNl0V1ROjNR1T1mNkNW1R1hNnN\\1o0cNPOc1k0^NTOf1h0ZNXOg1g0YNYOh1e0YNZOi1e0XNZOi1e0WN[Oj1d0VN\\Ok1c0UN\\Om1c0TN\\Om1c0SN]On1b0RN^On1a0SN^Oo1a0RN^Oo1a0QN_OP2`0PN@Q2?PN_OR2`0nM@S2?mMAT2>lMBU2=lMAV2=kMCV2<jMDV2<jMDW2;jMCX2<hMDY2;gMEZ2:gMEZ2:fME\\29eMG\\28dMH]27dMH]27bMIb24\\MNg2OXM2k2KTM6o2GPM:S3BlL?Y3]OfLd0]3YObLh0a3UO^Ll0e3QOZLP1j3lNTLV1\\:13L3N3M2N3M2N2M4M2N3M2N3M2M3N3MPcS1"}, "label": "a woman wearing an orange sweater is smiling with her hands up"}]}
{"id": 264233, "image": "COCO_train2014_000000264233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in orange sweater with arms in the air\"."}], "task": "refering", "answer_template": "The \"woman in orange sweater with arms in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 134, 135, 149, 150, 151, 152, 153, 156, 157, 158, 173, 174, 175, 176, 178, 179, 180, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 360, 361, 362, 363, 383, 384, 385, 386, 406, 407, 408, 409, 410, 429, 430, 431, 432, 433, 452, 453, 454, 455, 456], "bbox": [0.4875625, 0.19101549053356284, 0.9016875000000001, 0.959552495697074], "iscrowd": 0, "area": 40353.93594999998, "rle": {"size": [581, 640], "counts": "lTa52Pb05L4K5L5J5K5N2M3N2M3N2M3N2M4M2M3N2O1N2O1N21O5K5K5K5K4LO100O1O00100O1O1O100O001O1O100O1O1O010O1O1O100O1O001O100O1O1O10O01O1O1O100hBYNo9h1aEiN]:W1TEZOj:g0fDKX;5XD<f;EjCl0T<UO\\C\\1b<V2O2M2N2N2N2O1N3M2NO2M2N2O1N3M2O1N2O@_J]D_5`;hJ_DU5_;PK`Do4\\;WKcDg4[;_KcD_4Z;gKfDW4W;oKgDP4U;VLjDi3U;[LiDc3X;_LhD_3W;dLhD[3W;j100O00100O1O10O01aGTIe5l6WJhIZ5W6bJ^Jm4b5nJkJj4T5SKQKm4n4oJWKP5i4kJ]KT5c4gJcKY5[4cJkK\\5U4^JRLa5n3YJYLf5n3nIXLR6P4_IWL`6g70100O1OO2O0O100O2O7H8I8G9H8G8RLbGgMd8V2_GcMb8_2cGZM]8i2gGoLZ8S3kGfLU8]3nG\\LS8g3QHRLo7P4VHhKk7[4THcKl7`4SHiJWOWOf8S6SHaJAUO\\8]6SHXJJTOS8g6SHoIj8T6UGgIm8]6RG^IQ9d6oFVIT9n6kFmHX9U7k03L5L4K5L4K5L4K5L4L4K5L3L5L4K5L4K5L4L4K5L1N10]GlGV7R8jH]Hh6b7WInH\\6o6eIaIm5]6SJRJ_5l5bJcJQ5Z5oJVKc4i4]KfK]OVM[3Q7YM[Me2c2[MaMb2]2^MhM`2U2aMPN[2o1eMUNY2h1hM\\NU2c1jMbNT2[1mMiNQ2T1PNPOm1o0SNVOj1g0VN^Og1a0YNCe1<ZNHd16\\NNa11^N5_1IaN;\\1EcN?[1_OeNe0X1ZOgNl0V1ROjNR1T1mNkNW1R1hNnN\\1o0cNPOc1k0^NTOf1h0ZNXOg1g0YNYOh1e0YNZOi1e0XNZOi1e0WN[Oj1d0VN\\Ok1c0UN\\Om1c0TN\\Om1c0SN]On1b0RN^On1a0SN^Oo1a0RN^Oo1a0QN_OP2`0PN@Q2?PN_OR2`0nM@S2?mMAT2>lMBU2=lMAV2=kMCV2<jMDV2<jMDW2;jMCX2<hMDY2;gMEZ2:gMEZ2:fME\\29eMG\\28dMH]27dMH]27bMIb24\\MNg2OXM2k2KTM6o2GPM:S3BlL?Y3]OfLd0]3YObLh0a3UO^Ll0e3QOZLP1j3lNTLV1\\:13L3N3M2N3M2N2M4M2N3M2N3M2M3N3MPcS1"}, "label": "woman in orange sweater with arms in the air"}]}
{"id": 264233, "image": "COCO_train2014_000000264233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 107, 118, 119, 120, 129, 130, 142, 143, 144, 145, 147, 148, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334, 335, 336, 337, 350, 351, 352, 353, 354, 356, 357, 358, 359, 360, 373, 374, 375, 376, 380, 381, 382, 383, 396, 397, 398, 399, 404, 405, 420, 421], "bbox": [0.142796875, 0.2, 0.687484375, 0.8719104991394148], "iscrowd": 0, "area": 43505.235299999986, "rle": {"size": [581, 640], "counts": "_hc11oa06I7L4M3N2N2O1O1N2O1O1N2O1O1O1O1O1O1O1O001O1O1O1O1O010001O00001N1000001O000O101O000000O0100000O10O1000O100000O010000_FPOl1P1RNYOf1h0XN_Oc1`0\\NF`1:_NK]15aNO\\12bN1]1OaN4^1LaN5_1JbN7\\1JcN7]1IbN9]1GcN:[1GdN:\\1EdN=[1CdN>\\1BdN?Z1AfNa0Y1_OfNb0Z1^OeNd0Z1\\OfNd0Y1\\OgNf0X1ZOgNg0Y1YOfNh0Z1WOgNi0X1XOgNi0Y1WOfNj0Y1WOgNi0Y1VOgNk0Y1UOfNl0Y1UOfNl0Z1TOfNl0Y1TOgNm0Y1SOfNn0Z1ROeNo0Z1ROfNn0Z1ROeNn0[1SOeNm0[1SOfNl0Z1TOfNl0Y1VOgNi0Y1WOgNi0X1XOiNg0W1YOiNf0X1ZOhNe0X1\\OiNb0X1_OeNa0\\1@bNa0_1_O_Nb0b1^O\\Nb0e1_OYNb0h1_OUNb0k1_OSNa0o1_OoMb0Q2_OmMb0T2^OjMb0X2_OeMa0\\2@bM`0`2A\\M`0e2AYM`0h2AUM?m2APM`0Q3BlL?U3AiL?X3CdL?]3B`L>b3B\\L>e3DWL=k3CSL=o3DnK<S4EjK<X4FcK<^4F]K;e4FXK:i4IRK8P5JkJ7W5JeJ7\\5L`J5a5NYJ3i53nINT67`INa69RIKQ7;bHIa7?PHFQ8[1iFhNZ9o43M3M2N3M3M2N3M3M2N100O1O1O10001N10B>L3M4K5O0101O1O1N101O1N101O1N2O001O1N2O001N2O001N2O1O001N2O1O001O3M4L4L3M4L4M2O2N2N1O2O1N2SJ[DV5e;fJ]D[5e;_J^Db5d;YJ]Dj5c;RJ_Do5k;2N1N2O2N1N3N1O1O2M2O2N1O1N3NYKQFY1n9WNdFi1Z9UNjFj1T9UNQGi1n8UNVGk1h8SN\\Gl1c8RNaGm1]8QNiGn1U8PNoGo1P8oMTHo1k7PNYHo1f7oM^Ho1a7PNbHo1]7QNfHn1X7RNkHl1T7TNnHl1P7TNSIj1m6UNUIk1i6VNYIh1f6XN\\Ih1b6XNaIf1^6ZNeId1Z6\\NhId1V6]NlIa1T6^NnIb1P6^NSJ`1m5_NVJ`1j5^NXJa1j5]NXJb1h5\\NZJc1h5ZNZJe1g5YN[Jg1e5XN[Jh1h5TNZJl1g5QN[Jn1g5oM[JQ2f5lM\\JS2g5jMZJV2k:O1O100O100O10O0100O1000XIRNXMm1h2UNWMk1g2YNWMg1g2\\NXMc1g2aNWM_1g2bNZM^1d2bN^M]1a2cNaM]1]2cNdM^1Z2bNhM]1W2dNjM\\1T2eNmM[1Q2eNQNZ1n1gNSNY1k1hNVNX1i1gNYNX1f1iNZNX1d1iN]NW1a1iNaNV1^1kNbNV1\\1kNeNU1Y1kNiNU1V1kNjNU1U1lNlNT1R1lNPOT1o0lNQOU1m0lNROU1m0kNSOW1k0jNTOX1k0gNUO[1i0fNVO\\1h0dNXO]1h0bNYO^1f0cNYO]1g0cNYO^1f0bNZO_1f0aNYO`1f0`NZOa1e0_N[Ob1e0^NZOb1f0^NZOc1e0^N[Ob1d0^N\\Oc1d0\\N]Od1b0]N]Oc1c0]N]Od1c0[N^Oe1>^NBc1:`NG`16bNJ^16bNK^15aNK^16bNJ^16bNK\\17cNI]17cNJ[17eNJY18fNHY19gNHW19hNKS18jNLS15lNOo03PO1l01QO4j0NUO5d00ZO5>1@382F233L1L61NG86n8M2M4M_[a3"}, "label": "a man wearing a green shirt"}]}
{"id": 264233, "image": "COCO_train2014_000000264233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 107, 118, 119, 120, 129, 130, 142, 143, 144, 145, 147, 148, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334, 335, 336, 337, 350, 351, 352, 353, 354, 356, 357, 358, 359, 360, 373, 374, 375, 376, 380, 381, 382, 383, 396, 397, 398, 399, 404, 405, 420, 421], "bbox": [0.142796875, 0.2, 0.687484375, 0.8719104991394148], "iscrowd": 0, "area": 43505.235299999986, "rle": {"size": [581, 640], "counts": "_hc11oa06I7L4M3N2N2O1O1N2O1O1N2O1O1O1O1O1O1O1O001O1O1O1O1O010001O00001N1000001O000O101O000000O0100000O10O1000O100000O010000_FPOl1P1RNYOf1h0XN_Oc1`0\\NF`1:_NK]15aNO\\12bN1]1OaN4^1LaN5_1JbN7\\1JcN7]1IbN9]1GcN:[1GdN:\\1EdN=[1CdN>\\1BdN?Z1AfNa0Y1_OfNb0Z1^OeNd0Z1\\OfNd0Y1\\OgNf0X1ZOgNg0Y1YOfNh0Z1WOgNi0X1XOgNi0Y1WOfNj0Y1WOgNi0Y1VOgNk0Y1UOfNl0Y1UOfNl0Z1TOfNl0Y1TOgNm0Y1SOfNn0Z1ROeNo0Z1ROfNn0Z1ROeNn0[1SOeNm0[1SOfNl0Z1TOfNl0Y1VOgNi0Y1WOgNi0X1XOiNg0W1YOiNf0X1ZOhNe0X1\\OiNb0X1_OeNa0\\1@bNa0_1_O_Nb0b1^O\\Nb0e1_OYNb0h1_OUNb0k1_OSNa0o1_OoMb0Q2_OmMb0T2^OjMb0X2_OeMa0\\2@bM`0`2A\\M`0e2AYM`0h2AUM?m2APM`0Q3BlL?U3AiL?X3CdL?]3B`L>b3B\\L>e3DWL=k3CSL=o3DnK<S4EjK<X4FcK<^4F]K;e4FXK:i4IRK8P5JkJ7W5JeJ7\\5L`J5a5NYJ3i53nINT67`INa69RIKQ7;bHIa7?PHFQ8[1iFhNZ9o43M3M2N3M3M2N3M3M2N100O1O1O10001N10B>L3M4K5O0101O1O1N101O1N101O1N2O001O1N2O001N2O001N2O1O001N2O1O001O3M4L4L3M4L4M2O2N2N1O2O1N2SJ[DV5e;fJ]D[5e;_J^Db5d;YJ]Dj5c;RJ_Do5k;2N1N2O2N1N3N1O1O2M2O2N1O1N3NYKQFY1n9WNdFi1Z9UNjFj1T9UNQGi1n8UNVGk1h8SN\\Gl1c8RNaGm1]8QNiGn1U8PNoGo1P8oMTHo1k7PNYHo1f7oM^Ho1a7PNbHo1]7QNfHn1X7RNkHl1T7TNnHl1P7TNSIj1m6UNUIk1i6VNYIh1f6XN\\Ih1b6XNaIf1^6ZNeId1Z6\\NhId1V6]NlIa1T6^NnIb1P6^NSJ`1m5_NVJ`1j5^NXJa1j5]NXJb1h5\\NZJc1h5ZNZJe1g5YN[Jg1e5XN[Jh1h5TNZJl1g5QN[Jn1g5oM[JQ2f5lM\\JS2g5jMZJV2k:O1O100O100O10O0100O1000XIRNXMm1h2UNWMk1g2YNWMg1g2\\NXMc1g2aNWM_1g2bNZM^1d2bN^M]1a2cNaM]1]2cNdM^1Z2bNhM]1W2dNjM\\1T2eNmM[1Q2eNQNZ1n1gNSNY1k1hNVNX1i1gNYNX1f1iNZNX1d1iN]NW1a1iNaNV1^1kNbNV1\\1kNeNU1Y1kNiNU1V1kNjNU1U1lNlNT1R1lNPOT1o0lNQOU1m0lNROU1m0kNSOW1k0jNTOX1k0gNUO[1i0fNVO\\1h0dNXO]1h0bNYO^1f0cNYO]1g0cNYO^1f0bNZO_1f0aNYO`1f0`NZOa1e0_N[Ob1e0^NZOb1f0^NZOc1e0^N[Ob1d0^N\\Oc1d0\\N]Od1b0]N]Oc1c0]N]Od1c0[N^Oe1>^NBc1:`NG`16bNJ^16bNK^15aNK^16bNJ^16bNK\\17cNI]17cNJ[17eNJY18fNHY19gNHW19hNKS18jNLS15lNOo03PO1l01QO4j0NUO5d00ZO5>1@382F233L1L61NG86n8M2M4M_[a3"}, "label": "man"}]}
{"id": 231466, "image": "COCO_train2014_000000231466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue boat with a dog on it\"."}], "task": "refering", "answer_template": "The \"a blue boat with a dog on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280], "bbox": [0.0015, 0.5641822429906542, 0.5256562499999999, 0.8457242990654206], "iscrowd": 0, "area": 21375.187000000005, "rle": {"size": [428, 640], "counts": "af0R1Z<S1mN000000O10O1000000000O100000000000000O0100000000000O100000000000O10O1000000000000O1000000000O1000O100000000000000O100000O100000O10000000000000O01000000000000O1000O1000000000O10000000O1000O1000000TNeDg1[;WNgDi1Y;VNhDj1];0O0O2I7N2O1O1N2OWNPEW1P;hNSEW1l:jNTEV1m:hNTEX1l:hNTEX1l:fNUE[1l:aNWE_1i:^NXEd1h:YNYEh1T;2N2N2N2N2O1HlMVEU2j:7001O1O1O1O1O2M2oMPEd1\\;01O00O1O1O1L3O20000O100000O1000O1000000000O010000000000O01000000000O100000O1000O10000000000O10000000O10O10000000000O1000000000O010000000000O10000000000O01000000000O10000000000O10O100000O10000000000O100000O1000O10000000000O10000000O10O1000000000O010O1O1O10HkMYET2h:mMVES2j:oMUEQ2k:PNTEo1m:RNREm1n:90O1O011N2N4MO01O100O00100O00100O00100O00100O1O010O1O00100O001O1N101O1N101O1N2O001N2O002N1N2O1O1N2O2N1N2O1O1N2O2N1N2O1N2O1N3M2O1N2O1N]jn3"}, "label": "a blue boat with a dog on it"}]}
{"id": 231466, "image": "COCO_train2014_000000231466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white boat with a dog on it\"."}], "task": "refering", "answer_template": "The \"a blue and white boat with a dog on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280], "bbox": [0.0015, 0.5641822429906542, 0.5256562499999999, 0.8457242990654206], "iscrowd": 0, "area": 21375.187000000005, "rle": {"size": [428, 640], "counts": "af0R1Z<S1mN000000O10O1000000000O100000000000000O0100000000000O100000000000O10O1000000000000O1000000000O1000O100000000000000O100000O100000O10000000000000O01000000000000O1000O1000000000O10000000O1000O1000000TNeDg1[;WNgDi1Y;VNhDj1];0O0O2I7N2O1O1N2OWNPEW1P;hNSEW1l:jNTEV1m:hNTEX1l:hNTEX1l:fNUE[1l:aNWE_1i:^NXEd1h:YNYEh1T;2N2N2N2N2O1HlMVEU2j:7001O1O1O1O1O2M2oMPEd1\\;01O00O1O1O1L3O20000O100000O1000O1000000000O010000000000O01000000000O100000O1000O10000000000O10000000O10O10000000000O1000000000O010000000000O10000000000O01000000000O10000000000O10O100000O10000000000O100000O1000O10000000000O10000000O10O1000000000O010O1O1O10HkMYET2h:mMVES2j:oMUEQ2k:PNTEo1m:RNREm1n:90O1O011N2N4MO01O100O00100O00100O00100O00100O1O010O1O00100O001O1N101O1N101O1N2O001N2O002N1N2O1O1N2O2N1N2O1O1N2O2N1N2O1N2O1N3M2O1N2O1N]jn3"}, "label": "a blue and white boat with a dog on it"}]}
{"id": 43049, "image": "COCO_train2014_000000043049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow whose head is partially visible\"."}], "task": "refering", "answer_template": "The \"the cow whose head is partially visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 325, 326, 327, 328, 330, 331, 332, 345, 346, 350], "bbox": [0.0032187500000000003, 0.5957083333333333, 0.482265625, 0.9225833333333334], "iscrowd": 0, "area": 32159.447050000002, "rle": {"size": [480, 640], "counts": "TZ14h>9H8fB^O_;OcCj0a0@m;F^CR16B]<c11O1N2O1O1O1ON2N2N2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2O1N2N2N2O1N2N10M2N2N3M2N2N3M2O1N3M21O1O2O0O2N100O2N101N1O101N1010O1O100O010O11O001O001O0O2O001O001O001O0O2O001O001O001O0O2O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O0000N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O1O1001O0000000000001O0O1000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O00000TEbLV:^3gEdLZ:\\3cEfL^:Z3^EiLc:U3\\EmLe:P3[ERMf:l2YEUMi:h2WEZMj:c2VE_Mk:^2TEeMm:X2SEjMn:S2REoMo:n1QESNQ;j1oDXNR;f1lD]NU;`1kDbNV;[1jDgNW;V1hDmNY;P1gDQO[;l0eDVO\\;g0cD\\O^;a0bDA_;=`DEa;8^DKc;2]D0d;M[D5g;HYD:h;CXD9o;T1PEbMn9V2nEWNQ:a1lEjNT:m0jE_OU:9gE3Y:EbEh0^:PO\\E_1c:YNWEU2i:R1O00H9_Ob0^Oa0B?D<C=D;D=D<I78Hd0[O]1dNm0SOK5K4K6B>B>C=B=C>C=BVkj4"}, "label": "the cow whose head is partially visible"}]}
{"id": 43049, "image": "COCO_train2014_000000043049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left side of a black cow eating grass in front of another cow\"."}], "task": "refering", "answer_template": "The \"the left side of a black cow eating grass in front of another cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 325, 326, 327, 328, 330, 331, 332, 345, 346, 350], "bbox": [0.0032187500000000003, 0.5957083333333333, 0.482265625, 0.9225833333333334], "iscrowd": 0, "area": 32159.447050000002, "rle": {"size": [480, 640], "counts": "TZ14h>9H8fB^O_;OcCj0a0@m;F^CR16B]<c11O1N2O1O1O1ON2N2N2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2O1N2N2N2O1N2N10M2N2N3M2N2N3M2O1N3M21O1O2O0O2N100O2N101N1O101N1010O1O100O010O11O001O001O0O2O001O001O001O0O2O001O001O001O0O2O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O0000N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O1O1001O0000000000001O0O1000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O00000TEbLV:^3gEdLZ:\\3cEfL^:Z3^EiLc:U3\\EmLe:P3[ERMf:l2YEUMi:h2WEZMj:c2VE_Mk:^2TEeMm:X2SEjMn:S2REoMo:n1QESNQ;j1oDXNR;f1lD]NU;`1kDbNV;[1jDgNW;V1hDmNY;P1gDQO[;l0eDVO\\;g0cD\\O^;a0bDA_;=`DEa;8^DKc;2]D0d;M[D5g;HYD:h;CXD9o;T1PEbMn9V2nEWNQ:a1lEjNT:m0jE_OU:9gE3Y:EbEh0^:PO\\E_1c:YNWEU2i:R1O00H9_Ob0^Oa0B?D<C=D;D=D<I78Hd0[O]1dNm0SOK5K4K6B>B>C=B=C>C=BVkj4"}, "label": "the left side of a black cow eating grass in front of another cow"}]}
{"id": 534573, "image": "COCO_train2014_000000534573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"frontier airlines plane with an owl on the back wing\"."}], "task": "refering", "answer_template": "The \"frontier airlines plane with an owl on the back wing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 70, 71, 72, 73, 95, 96, 97, 103, 106, 107, 108, 109, 110, 111, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 178, 193], "bbox": [0.027359375000000002, 0.18183060109289617, 0.860890625, 0.6286338797814207], "iscrowd": 0, "area": 29831.211199999998, "rle": {"size": [366, 640], "counts": "c`61];00001O000O2O0000001O0O1000000000000O1000O100000O10000000000O01000000000O10000000O10O10000000000O1I70O1000000002M3N3M3M3M2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3M3hGaM[7b2aHdMZ7^2dHcM[7`2bH`M^7c2^H^Mb7d2\\H]Mc7f2YH[Mg7h2VHXMj7j2THVMl7W3O00001O00001O0000001O00001O00001O00000010eL]Hi2c7VM`Hh2`7WMbHi2]7RMjHl2V7TMlHj2T7VMnHi2P7XMQIg2o6YMSIe2m6ZMVId2j6\\MWId2h6\\MZIb2f6^M\\I`2d6`M]I`2b6`M`I^2_6cMcI[2]6eMdI[2[6eMgIY2Y6gMiIW2W6iMkIU2U6kMlIU2S6jMPJT2o5mMSJQ2m5oMTJP2l5PNVJo1i5QNYJm1g5SNZJl1f5TN\\Jj1d5VN^Jh1a5YN`Jg1_5YNcJe1]5[NeJc1[5]NfJb1Z5]NgJd1X5\\NhJd1X5\\NhJd1W5]NiJc1W5]NiJc1W5]NjJc1U5]NkJc1U5]NkJc1U5]NkJc1U5]NkJd1T5\\NmJc1S5]NmJc1T5[NmJe1S5[NmJe1S5ZNnJg1R5VNPKj1[7O0001O0001O0001O000001O01O000010O0001O001O01O000000001O01O00000001O0000000010O0000000001O0000000010O0000000001O000000001O0O100000000000000000000000000000O10000000000lFYNa8g1XG`Nh8`1UGcNk8]1SGeNm8[1RGfNn8Z1PGgNQ9Y1nFhNR9j1000000000000O10O100000000_ORGbNn8]1VG^Nl8a1XGZNj8f1?00000000000000000O1000O100000L4O1001O00000000000O10001O0000000000000000ZO^N`Gb1`8kNSGU1m8e02O0000007I<D0000000000000O1000001O00N2N2N2N22N1O1O1N2O000000000000000000O1000O100000000000000000000000O01000000000LYNaFg1_94000000lFSNg8m1VGWNi8i1TG[Nk8e1RG_Nm8`1RGcNl8Q2O1O00000000O10000O10O1000O1@SG_Nm8a1UG]Nk8c1VG\\Nj8c1YGZNh8f1YGYNg8g1`GRN`8n1`00O10O100000000O10000000O010000000dFUNU9S2O000000000O1000001OO1000[OnMmGR2T8SNgGm1Z8WNaGi1_8b02O1O2M2N3M2M9H6J1N20O1000000O1000O10O1O1O100O1O1O1000O101SNdFf1\\9YNeFg1\\9VNeFk1_900000000_N_FU1`9kNaFT1`9lNaFS1_9lNbFT1^9lNbFT1^9kNdFT1\\9kNeFU1[9kNeFU1Z9kNhFT1X9lNiFS1W9mNjF[1m8dNTG_1i8aNXG^1h8bNYG]1g8cNZG\\1e8eN[G[1e8eN\\GZ1d8eN^GZ1b8fN_GY1a8gN_GY1a8gN`GX1_8iNbGV1^8iNdGV1\\8jNeGT1\\8lNeGQ1]8oNdGm0_8SObGk0_8UObGi0^8WOeGe0]8[OdGd0\\8\\OeGb0\\8^OeGa0[8_OfG`0Z8_OiG>W8CkG;U8ElG9U8GmG7S8IoGLZ84`1O10O100000O1000O100000O100000O10O10000000O10O100000000O011O00001O0O101OQfo0"}, "label": "frontier airlines plane with an owl on the back wing"}]}
{"id": 534573, "image": "COCO_train2014_000000534573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plane\"."}], "task": "refering", "answer_template": "The \"plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 70, 71, 72, 73, 95, 96, 97, 103, 106, 107, 108, 109, 110, 111, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 178, 193], "bbox": [0.027359375000000002, 0.18183060109289617, 0.860890625, 0.6286338797814207], "iscrowd": 0, "area": 29831.211199999998, "rle": {"size": [366, 640], "counts": "c`61];00001O000O2O0000001O0O1000000000000O1000O100000O10000000000O01000000000O10000000O10O10000000000O1I70O1000000002M3N3M3M3M2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3M3hGaM[7b2aHdMZ7^2dHcM[7`2bH`M^7c2^H^Mb7d2\\H]Mc7f2YH[Mg7h2VHXMj7j2THVMl7W3O00001O00001O0000001O00001O00001O00000010eL]Hi2c7VM`Hh2`7WMbHi2]7RMjHl2V7TMlHj2T7VMnHi2P7XMQIg2o6YMSIe2m6ZMVId2j6\\MWId2h6\\MZIb2f6^M\\I`2d6`M]I`2b6`M`I^2_6cMcI[2]6eMdI[2[6eMgIY2Y6gMiIW2W6iMkIU2U6kMlIU2S6jMPJT2o5mMSJQ2m5oMTJP2l5PNVJo1i5QNYJm1g5SNZJl1f5TN\\Jj1d5VN^Jh1a5YN`Jg1_5YNcJe1]5[NeJc1[5]NfJb1Z5]NgJd1X5\\NhJd1X5\\NhJd1W5]NiJc1W5]NiJc1W5]NjJc1U5]NkJc1U5]NkJc1U5]NkJc1U5]NkJd1T5\\NmJc1S5]NmJc1T5[NmJe1S5[NmJe1S5ZNnJg1R5VNPKj1[7O0001O0001O0001O000001O01O000010O0001O001O01O000000001O01O00000001O0000000010O0000000001O0000000010O0000000001O000000001O0O100000000000000000000000000000O10000000000lFYNa8g1XG`Nh8`1UGcNk8]1SGeNm8[1RGfNn8Z1PGgNQ9Y1nFhNR9j1000000000000O10O100000000_ORGbNn8]1VG^Nl8a1XGZNj8f1?00000000000000000O1000O100000L4O1001O00000000000O10001O0000000000000000ZO^N`Gb1`8kNSGU1m8e02O0000007I<D0000000000000O1000001O00N2N2N2N22N1O1O1N2O000000000000000000O1000O100000000000000000000000O01000000000LYNaFg1_94000000lFSNg8m1VGWNi8i1TG[Nk8e1RG_Nm8`1RGcNl8Q2O1O00000000O10000O10O1000O1@SG_Nm8a1UG]Nk8c1VG\\Nj8c1YGZNh8f1YGYNg8g1`GRN`8n1`00O10O100000000O10000000O010000000dFUNU9S2O000000000O1000001OO1000[OnMmGR2T8SNgGm1Z8WNaGi1_8b02O1O2M2N3M2M9H6J1N20O1000000O1000O10O1O1O100O1O1O1000O101SNdFf1\\9YNeFg1\\9VNeFk1_900000000_N_FU1`9kNaFT1`9lNaFS1_9lNbFT1^9lNbFT1^9kNdFT1\\9kNeFU1[9kNeFU1Z9kNhFT1X9lNiFS1W9mNjF[1m8dNTG_1i8aNXG^1h8bNYG]1g8cNZG\\1e8eN[G[1e8eN\\GZ1d8eN^GZ1b8fN_GY1a8gN_GY1a8gN`GX1_8iNbGV1^8iNdGV1\\8jNeGT1\\8lNeGQ1]8oNdGm0_8SObGk0_8UObGi0^8WOeGe0]8[OdGd0\\8\\OeGb0\\8^OeGa0[8_OfG`0Z8_OiG>W8CkG;U8ElG9U8GmG7S8IoGLZ84`1O10O100000O1000O100000O100000O10O10000000O10O100000000O011O00001O0O101OQfo0"}, "label": "plane"}]}
{"id": 198704, "image": "COCO_train2014_000000198704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver van that is about to cross the horse path\"."}], "task": "refering", "answer_template": "The \"the silver van that is about to cross the horse path\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320], "bbox": [0.54634375, 0.6002083333333333, 0.9504374999999999, 0.8222083333333334], "iscrowd": 0, "area": 20721.623349999998, "rle": {"size": [480, 640], "counts": "[]T5;a>6J5K6K4M2N3M2N2O2N10000O2O000O2O000O2O0O10001N10001N2O001N101O0O2O001N10000O1000O010000O10O10O100O00100O011N100O2N2O0O2O1N101N1O100O101N100O1O100O100O100O00100O100O100O1O100OO2O0O2O0O101N101O0100O1O100O100O100O10000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O10000000001O1O100O1O1O1O00001O010O00001O0000000001N100O10000O100O2N100O1O1O1O11O2O1N1O1O001O001O0010O00000001OO10001O00000000000O102N1O2N1O2N1O2L3N2N3L3N3M2M4M2N3M2N2O2N1O2N1N7I?A`0@gd>"}, "label": "the silver van that is about to cross the horse path"}]}
{"id": 198704, "image": "COCO_train2014_000000198704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver minivan near man with horse\"."}], "task": "refering", "answer_template": "The \"silver minivan near man with horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320], "bbox": [0.54634375, 0.6002083333333333, 0.9504374999999999, 0.8222083333333334], "iscrowd": 0, "area": 20721.623349999998, "rle": {"size": [480, 640], "counts": "[]T5;a>6J5K6K4M2N3M2N2O2N10000O2O000O2O000O2O0O10001N10001N2O001N101O0O2O001N10000O1000O010000O10O10O100O00100O011N100O2N2O0O2O1N101N1O100O101N100O1O100O100O100O00100O100O100O1O100OO2O0O2O0O101N101O0100O1O100O100O100O10000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O10000000001O1O100O1O1O1O00001O010O00001O0000000001N100O10000O100O2N100O1O1O1O11O2O1N1O1O001O001O0010O00000001OO10001O00000000000O102N1O2N1O2N1O2L3N2N3L3N3M2M4M2N3M2N2O2N1O2N1N7I?A`0@gd>"}, "label": "silver minivan near man with horse"}]}
{"id": 198704, "image": "COCO_train2014_000000198704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rear of a black subaru legacy with a yelklow license plate\"."}], "task": "refering", "answer_template": "The \"the rear of a black subaru legacy with a yelklow license plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0016875000000000002, 0.7550625, 0.22246875000000005, 0.9887708333333334], "iscrowd": 0, "area": 14431.1394, "rle": {"size": [480, 640], "counts": "cj0Z1f=k1UN0000000000001O00000O10000000000000000000000000000000000000000000000O1000000000000000000000000O1000000O1000000O1000000O1000000O1000000O10001O0O11O00000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000001O0000001O00a0_O2N2N2N2N2N2O1N5K8H7I7I6FbPY7"}, "label": "the rear of a black subaru legacy with a yelklow license plate"}]}
{"id": 198704, "image": "COCO_train2014_000000198704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car with a yellow license plate\"."}], "task": "refering", "answer_template": "The \"a black car with a yellow license plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0016875000000000002, 0.7550625, 0.22246875000000005, 0.9887708333333334], "iscrowd": 0, "area": 14431.1394, "rle": {"size": [480, 640], "counts": "cj0Z1f=k1UN0000000000001O00000O10000000000000000000000000000000000000000000000O1000000000000000000000000O1000000O1000000O1000000O1000000O1000000O10001O0O11O00000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000001O0000001O00a0_O2N2N2N2N2N2O1N5K8H7I7I6FbPY7"}, "label": "a black car with a yellow license plate"}]}
{"id": 346161, "image": "COCO_train2014_000000346161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a plaid shirt about to cut a piece of pizza\"."}], "task": "refering", "answer_template": "The \"a man with a plaid shirt about to cut a piece of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 151, 153, 154, 155, 163, 164, 165, 167, 168, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 202, 203, 204, 205, 215, 216, 217, 218], "bbox": [0.5108055555555555, 0.45023437499999996, 1.0, 0.7422968749999999], "iscrowd": 0, "area": 15426.011999999997, "rle": {"size": [640, 360], "counts": "h[c39ec07J3M4L4L4L4L4L4K6K4L4L5K4L4L3M2M4M2N3M2N3M2N2M4M2N2N3M2N2N3M2M3N2N2N2N2N2N2N2M3N2N2N2N2000O010O100O010O100O010O100N1O2N2N2M210O1O001O10O01O001O10O01O1O0010O01O1O0UOi_OnMW`0P2j_OPNW`0o1j_OPNV`0P2k_OoMV`0P2k_OPNT`0P2m_OoMT`0o1n_OPNS`0o1n_OPNR`0P2n_OQNQ`0o1P@PNQ`0o1P@PNP`0o1R@oMP`0P2Q@oMo?Q2R@mMP`0R2Q@mMo?S2S@kMm?U2U@hMm?V2V@hMj?X2X@eMj?Z2X@dMh?[2P1O0O_NeM`AZ2_>hM`AW2`>jM`AV2_>lMaAR2_>PN`AP2^>SNbAk1^>VNbAj1]>XNcAf1]>\\NbAb1_>`N`A]1b>dN_AY1a>jN^AS1d>oN\\Am0f>UOYAh0i>ZOWAd0i>]OWAa0j>AVA=j>EVA9j>IUA5m>KTA3m>NRA0o>2QALP?5o@JQ?7PAFR?;n@CS?>l@AT?a0l@\\OV?d0j@ZOX?g0h@UOZ?m0e@QO]?P1c@lN_?U1a@iNa?X1_@dNd?]1\\@_Ng?b1[1001O1O1O001O1O1O001O1O1O001O1O10M3N1O2N2O1O1O001O1O1O10O01O2N3M3M3M2H9G[F"}, "label": "a man with a plaid shirt about to cut a piece of pizza"}]}
{"id": 346161, "image": "COCO_train2014_000000346161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing black and white checks shirt\"."}], "task": "refering", "answer_template": "The \"a person wearing black and white checks shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 151, 153, 154, 155, 163, 164, 165, 167, 168, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 202, 203, 204, 205, 215, 216, 217, 218], "bbox": [0.5108055555555555, 0.45023437499999996, 1.0, 0.7422968749999999], "iscrowd": 0, "area": 15426.011999999997, "rle": {"size": [640, 360], "counts": "h[c39ec07J3M4L4L4L4L4L4K6K4L4L5K4L4L3M2M4M2N3M2N3M2N2M4M2N2N3M2N2N3M2M3N2N2N2N2N2N2N2M3N2N2N2N2000O010O100O010O100O010O100N1O2N2N2M210O1O001O10O01O001O10O01O1O0010O01O1O0UOi_OnMW`0P2j_OPNW`0o1j_OPNV`0P2k_OoMV`0P2k_OPNT`0P2m_OoMT`0o1n_OPNS`0o1n_OPNR`0P2n_OQNQ`0o1P@PNQ`0o1P@PNP`0o1R@oMP`0P2Q@oMo?Q2R@mMP`0R2Q@mMo?S2S@kMm?U2U@hMm?V2V@hMj?X2X@eMj?Z2X@dMh?[2P1O0O_NeM`AZ2_>hM`AW2`>jM`AV2_>lMaAR2_>PN`AP2^>SNbAk1^>VNbAj1]>XNcAf1]>\\NbAb1_>`N`A]1b>dN_AY1a>jN^AS1d>oN\\Am0f>UOYAh0i>ZOWAd0i>]OWAa0j>AVA=j>EVA9j>IUA5m>KTA3m>NRA0o>2QALP?5o@JQ?7PAFR?;n@CS?>l@AT?a0l@\\OV?d0j@ZOX?g0h@UOZ?m0e@QO]?P1c@lN_?U1a@iNa?X1_@dNd?]1\\@_Ng?b1[1001O1O1O001O1O1O001O1O1O001O1O10M3N1O2N2O1O1O001O1O1O10O01O2N3M3M3M2H9G[F"}, "label": "a person wearing black and white checks shirt"}]}
{"id": 346161, "image": "COCO_train2014_000000346161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing dress with flowers on it\"."}], "task": "refering", "answer_template": "The \"woman wearing dress with flowers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 32, 33, 34, 35, 44, 45, 46, 47, 48, 49, 57, 58, 59, 60, 61, 62, 69, 70, 71, 72, 73, 74, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 137, 138, 139, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 157, 158, 159, 160, 161, 162, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 214, 215, 221, 222, 223, 224, 227, 228, 229, 234, 235, 241, 242, 254, 255, 267], "bbox": [0.0, 0.042703124999999995, 1.0, 0.8809062500000001], "iscrowd": 0, "area": 72765.18595, "rle": {"size": [640, 360], "counts": "j<j2Ta03N1N2N2N3M2N2N2N3M2N2O1N3M2N2N2N2N3M2N2N2N3N1N2N2N2N3M2N2M3M3N3L3M3N2M3M4M2M3N2N2N3M2N2N2O1N2N2N1O1O2N100O1O2N1O1O2N100O1O2N100O2N1O100O2N100O2N100O1O2O0O1O2O0O1O1O2O0O1O2O0O1O101N1O101N1N2L4L5K4M3L5K4L4M3L5M2000001O000O2O000000O100O100O100O100O1O2N1O1O1O2N1O1O1O2N100O2N1O1O1O2N11O2N2O1N2N3N1N2N2N2O2M2N2J6H8H9F9J6M3N2N3L3N3L4M3L4M3L4M3M3L4M2MXLPHeKn7U4^HhK^7U4jHiKT7W4QIgKl6X4YIfKd6[4aIbK]6^4gIaKU6`4PJ^Km5b4WJ]Ke5d4`JZK]5e4hJZKT5g4PKXKm4h4XKVKe4j4_KUK]4l4hKRKU4m4PLQKm3P5XLnJe3R5_LPKZ3Q5jLSKn2m4WMUKa2m4aMXKW2h4lM\\Km1d4VNaKa1a4aNaKZ1_4iNbKR1_4QOcKj0^4WOdKd0]4_OdK=\\4FdK7]4KcK1^42bKK^48bKD`4>`K_O`4d0_KYOb4i0_KTOb4n0^KnNe4S1[KjNh4V1XKgNj4Z1VKbNn4^1nJcNU5^1gJbN[5`1aJaNa5`1ZJbNh5`1SJaNn5a1nI`NT6b1fI`N\\6a1`I`Na6c1ZI]Nh6e1TI\\Nm6f1nH\\N`3[LSN\\5XNZNd3^LRNZ5VNYNg3aLQNY5RNXNl3bLQNX5oMWNo3eLPNX5kMTNT4hLoM[5dMnM\\4kLnM^5\\MiMe4lLnMl9Q2XFmMg9R2]FlMb9S2bFkM]9T2hFiMW9V2nFgMQ9X2TGeMk8Z2YGeMe8Z2`GcM_8\\2fGaMY8^2lG_MS8`2RH]Ml7c2VH]Mi7b2YH^Mf7a2\\H_Md7^2`H`Ma7^2aHbM^7]2dHcM\\7[2fHdM[7Y2hHgMW7X2kHgMV7W2lHiMT7T2oHkMR7S2QIlMn6S2TImMl6Q2VInMk6o1XIQNg6n1[IQNf6m1\\ISNd6k1^ITNb6j1aIVN_6h1dIVN]6h1eIXNZ6g1hIXNY6e1jI[NV6c1lI\\NU6b1nI]NR6`1QJ_NQ6^1QJbNo5[1TJdNm5Y1WJfNi5W1ZJhNg5T1]JlNd5P1_JoNe5k0^JUOf5c0_J[Of5>]JBf58]JGh5N_J2e5DbJ:b5]OdJc0`5ROgJm0^5iNhJV1S<O0O100O2OO0100O10O01000O0100O10O01000O0100O10O01000O0100O10O01000O0100O10O0100O00100O1O010O1O010O1O10O01O10O01O100O010O1O010O1O10O01O10O01O100O00100O0010^D"}, "label": "woman wearing dress with flowers on it"}]}
{"id": 346161, "image": "COCO_train2014_000000346161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman in a gray dress\"."}], "task": "refering", "answer_template": "The \"a blonde woman in a gray dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 32, 33, 34, 35, 44, 45, 46, 47, 48, 49, 57, 58, 59, 60, 61, 62, 69, 70, 71, 72, 73, 74, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 137, 138, 139, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 157, 158, 159, 160, 161, 162, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 214, 215, 221, 222, 223, 224, 227, 228, 229, 234, 235, 241, 242, 254, 255, 267], "bbox": [0.0, 0.042703124999999995, 1.0, 0.8809062500000001], "iscrowd": 0, "area": 72765.18595, "rle": {"size": [640, 360], "counts": "j<j2Ta03N1N2N2N3M2N2N2N3M2N2O1N3M2N2N2N2N3M2N2N2N3N1N2N2N2N3M2N2M3M3N3L3M3N2M3M4M2M3N2N2N3M2N2N2O1N2N2N1O1O2N100O1O2N1O1O2N100O1O2N100O2N1O100O2N100O2N100O1O2O0O1O2O0O1O1O2O0O1O2O0O1O101N1O101N1N2L4L5K4M3L5K4L4M3L5M2000001O000O2O000000O100O100O100O100O1O2N1O1O1O2N1O1O1O2N100O2N1O1O1O2N11O2N2O1N2N3N1N2N2N2O2M2N2J6H8H9F9J6M3N2N3L3N3L4M3L4M3L4M3M3L4M2MXLPHeKn7U4^HhK^7U4jHiKT7W4QIgKl6X4YIfKd6[4aIbK]6^4gIaKU6`4PJ^Km5b4WJ]Ke5d4`JZK]5e4hJZKT5g4PKXKm4h4XKVKe4j4_KUK]4l4hKRKU4m4PLQKm3P5XLnJe3R5_LPKZ3Q5jLSKn2m4WMUKa2m4aMXKW2h4lM\\Km1d4VNaKa1a4aNaKZ1_4iNbKR1_4QOcKj0^4WOdKd0]4_OdK=\\4FdK7]4KcK1^42bKK^48bKD`4>`K_O`4d0_KYOb4i0_KTOb4n0^KnNe4S1[KjNh4V1XKgNj4Z1VKbNn4^1nJcNU5^1gJbN[5`1aJaNa5`1ZJbNh5`1SJaNn5a1nI`NT6b1fI`N\\6a1`I`Na6c1ZI]Nh6e1TI\\Nm6f1nH\\N`3[LSN\\5XNZNd3^LRNZ5VNYNg3aLQNY5RNXNl3bLQNX5oMWNo3eLPNX5kMTNT4hLoM[5dMnM\\4kLnM^5\\MiMe4lLnMl9Q2XFmMg9R2]FlMb9S2bFkM]9T2hFiMW9V2nFgMQ9X2TGeMk8Z2YGeMe8Z2`GcM_8\\2fGaMY8^2lG_MS8`2RH]Ml7c2VH]Mi7b2YH^Mf7a2\\H_Md7^2`H`Ma7^2aHbM^7]2dHcM\\7[2fHdM[7Y2hHgMW7X2kHgMV7W2lHiMT7T2oHkMR7S2QIlMn6S2TImMl6Q2VInMk6o1XIQNg6n1[IQNf6m1\\ISNd6k1^ITNb6j1aIVN_6h1dIVN]6h1eIXNZ6g1hIXNY6e1jI[NV6c1lI\\NU6b1nI]NR6`1QJ_NQ6^1QJbNo5[1TJdNm5Y1WJfNi5W1ZJhNg5T1]JlNd5P1_JoNe5k0^JUOf5c0_J[Of5>]JBf58]JGh5N_J2e5DbJ:b5]OdJc0`5ROgJm0^5iNhJV1S<O0O100O2OO0100O10O01000O0100O10O01000O0100O10O01000O0100O10O01000O0100O10O0100O00100O1O010O1O010O1O10O01O10O01O100O010O1O010O1O10O01O10O01O100O00100O0010^D"}, "label": "a blonde woman in a gray dress"}]}
{"id": 346161, "image": "COCO_train2014_000000346161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a whole cheese pizza\"."}], "task": "refering", "answer_template": "The \"a whole cheese pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 255, 256, 257, 258, 259, 268, 269, 270, 271, 272, 282, 283, 284, 285, 297, 298], "bbox": [0.6272222222222222, 0.804125, 0.99875, 0.9883906249999999], "iscrowd": 0, "area": 10300.221599999999, "rle": {"size": [640, 360], "counts": "[i]44ic04L4K6K4L5J5L4L5M2O2M2N2O2M2N3N1N2O2M2N3N1N2N3N1N3N1N2N3N0O2O00001O000010O0001O000010O0001O000010O0001O000010O0001O000010O0001O000010O0001O000010O00010O00100O0010O01O010O1O010O00100O0010O01O010O1O010O0010O01O10O01O010O00100O1O010O1O100O1O100O00100O1O100O1O10O01O1@`0]NP@"}, "label": "a whole cheese pizza"}]}
{"id": 346161, "image": "COCO_train2014_000000346161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a whole cheese pizza in a pizza box\"."}], "task": "refering", "answer_template": "The \"a whole cheese pizza in a pizza box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 255, 256, 257, 258, 259, 268, 269, 270, 271, 272, 282, 283, 284, 285, 297, 298], "bbox": [0.6272222222222222, 0.804125, 0.99875, 0.9883906249999999], "iscrowd": 0, "area": 10300.221599999999, "rle": {"size": [640, 360], "counts": "[i]44ic04L4K6K4L5J5L4L5M2O2M2N2O2M2N3N1N2O2M2N3N1N2N3N1N3N1N2N3N0O2O00001O000010O0001O000010O0001O000010O0001O000010O0001O000010O0001O000010O0001O000010O00010O00100O0010O01O010O1O010O00100O0010O01O010O1O010O0010O01O10O01O010O00100O1O010O1O100O1O100O00100O1O100O1O10O01O1@`0]NP@"}, "label": "a whole cheese pizza in a pizza box"}]}
{"id": 174137, "image": "COCO_train2014_000000174137.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a solid orange tabby cat\"."}], "task": "refering", "answer_template": "The \"a solid orange tabby cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216], "bbox": [0.232125, 0.23264583333333333, 0.5319843750000001, 0.5858333333333333], "iscrowd": 0, "area": 17894.647050000003, "rle": {"size": [480, 640], "counts": "USV2:d>6J3M3N2N2N2N2N1N1O1O1O2O0O1N2002N2N2N2E;L5M2N2N01N101_Oa0M3M3N2M2N4K4K5J7I6ROn0M3N3L3O1O2N1O2N101N1O2N1O100O1O1O100O1O1O00100O1O1O1O100O1O10000O10000O101O000O10000O1000000O1000000O1000000O1000000O101O00000O1000001O0O100000001N100000000O2OO100O10O10O10000O10O02M2M3E<E:M3O2N1O1O2M2N2N3I6J7M2O1N3N1N2N2O1N3M2O1N2N2N2O1O10000000O101N1O1O1O1O2N100O100O100O100001O1O1N4K5L?A4K5L4L4K4K6J6JX_\\4"}, "label": "a solid orange tabby cat"}]}
{"id": 174137, "image": "COCO_train2014_000000174137.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange cat sitting on bed next to suit case\"."}], "task": "refering", "answer_template": "The \"orange cat sitting on bed next to suit case\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216], "bbox": [0.232125, 0.23264583333333333, 0.5319843750000001, 0.5858333333333333], "iscrowd": 0, "area": 17894.647050000003, "rle": {"size": [480, 640], "counts": "USV2:d>6J3M3N2N2N2N2N1N1O1O1O2O0O1N2002N2N2N2E;L5M2N2N01N101_Oa0M3M3N2M2N4K4K5J7I6ROn0M3N3L3O1O2N1O2N101N1O2N1O100O1O1O100O1O1O00100O1O1O1O100O1O10000O10000O101O000O10000O1000000O1000000O1000000O1000000O101O00000O1000001O0O100000001N100000000O2OO100O10O10O10000O10O02M2M3E<E:M3O2N1O1O2M2N2N3I6J7M2O1N3N1N2N2O1N3M2O1N2N2N2O1O10000000O101N1O1O1O1O2N100O100O100O100001O1O1N4K5L?A4K5L4L4K4K6J6JX_\\4"}, "label": "orange cat sitting on bed next to suit case"}]}
{"id": 493626, "image": "COCO_train2014_000000493626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with one foot on the group and another inside a plane\"."}], "task": "refering", "answer_template": "The \"a man with one foot on the group and another inside a plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 103, 104, 123, 124, 126, 127, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 286, 287, 288, 290, 291, 309, 310, 311, 332, 333, 334, 355, 356, 357, 378, 379], "bbox": [0.38289062500000004, 0.23692007797270956, 0.7031562499999999, 0.9244249512670566], "iscrowd": 0, "area": 28175.450249999998, "rle": {"size": [513, 640], "counts": "Pmj3=_?8H7I8I7H8J6K4L2N2ZEQNX7Q2fHQNX7R2dHRNX7Q2eHRNY7P2dHTNY7n1dHUNZ7m1cHWNZ7l1cHWN[7j1bHYN^7g1_H]N`7c1]H`Nc7`1ZHdNe7]1WHfNi7Z1UHiNj7W1SHmNl7S1QHPOo7P1nGTOQ8m0kGVOU8j0hGZOY8B[FZOZ1V1^8VO_FDP1Z1c8hNdFNf0^1h8YNjF9;`1P9lMlFe00a1`:_N]Ec1e:\\NXEe1j:[NSEg1n:YNoDi1R;XNiDj1Y;VNdDl1];TN`Dn1a;RN\\Do1f;QNWDQ2j;PNRDR2o;n02O1O2M2O1N2O1O1N2O1O1N2OUG`NW3_1YLROg3l0jKFU49[JX1e5gNnIg1R6WNbIX2]6gMVIg2j6Y3O0000O1N2O1O1O1O1N2O001O001N10YOg0M3N1N3M300O10O100000001N2O2N1F:C>A>B>C>A>B>O2N1O2O0O2N2O0O2N2O1N1O2O1N2M2Nh0YOb0]O2N3M3N1N3M3M2O2M3N101N2O0O2O001N2O0O3N<D<C<E<C=D<D8G100O10000O100O10000O100O1000VAj0i<UOTCo0l<QOoBT1Q=n00000000000O1000000000O100000000000000000002N9G8H9G6J6J6J6J7I5K2N2N2N2N2N3M2N2N2N2N2N2N3M_jn2"}, "label": "a man with one foot on the group and another inside a plane"}]}
{"id": 493626, "image": "COCO_train2014_000000493626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"businessman posing in front of an airplane door\"."}], "task": "refering", "answer_template": "The \"businessman posing in front of an airplane door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 103, 104, 123, 124, 126, 127, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 286, 287, 288, 290, 291, 309, 310, 311, 332, 333, 334, 355, 356, 357, 378, 379], "bbox": [0.38289062500000004, 0.23692007797270956, 0.7031562499999999, 0.9244249512670566], "iscrowd": 0, "area": 28175.450249999998, "rle": {"size": [513, 640], "counts": "Pmj3=_?8H7I8I7H8J6K4L2N2ZEQNX7Q2fHQNX7R2dHRNX7Q2eHRNY7P2dHTNY7n1dHUNZ7m1cHWNZ7l1cHWN[7j1bHYN^7g1_H]N`7c1]H`Nc7`1ZHdNe7]1WHfNi7Z1UHiNj7W1SHmNl7S1QHPOo7P1nGTOQ8m0kGVOU8j0hGZOY8B[FZOZ1V1^8VO_FDP1Z1c8hNdFNf0^1h8YNjF9;`1P9lMlFe00a1`:_N]Ec1e:\\NXEe1j:[NSEg1n:YNoDi1R;XNiDj1Y;VNdDl1];TN`Dn1a;RN\\Do1f;QNWDQ2j;PNRDR2o;n02O1O2M2O1N2O1O1N2O1O1N2OUG`NW3_1YLROg3l0jKFU49[JX1e5gNnIg1R6WNbIX2]6gMVIg2j6Y3O0000O1N2O1O1O1O1N2O001O001N10YOg0M3N1N3M300O10O100000001N2O2N1F:C>A>B>C>A>B>O2N1O2O0O2N2O0O2N2O1N1O2O1N2M2Nh0YOb0]O2N3M3N1N3M3M2O2M3N101N2O0O2O001N2O0O3N<D<C<E<C=D<D8G100O10000O100O10000O100O1000VAj0i<UOTCo0l<QOoBT1Q=n00000000000O1000000000O100000000000000000002N9G8H9G6J6J6J6J7I5K2N2N2N2N2N3M2N2N2N2N2N2N3M_jn2"}, "label": "businessman posing in front of an airplane door"}]}
{"id": 493626, "image": "COCO_train2014_000000493626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the plain with the white hat\"."}], "task": "refering", "answer_template": "The \"the man in the plain with the white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 245, 246, 247, 248, 269, 270, 271, 293, 294], "bbox": [0.5630625, 0.2878362573099415, 0.8177031250000001, 0.7061988304093567], "iscrowd": 0, "area": 13379.097250000006, "rle": {"size": [513, 640], "counts": "cPe54l?2N3M3L4M2N3M3L4M1O2N2M3N2N101O1O1N101O1O1N2O1O2N2M3N2N3M2XCRNX;Q2dDQN\\;Q2aDPN_;R2]DQNb;Q2ZDQNf;R2UDPNj;S2RDoMn;S2oCoMP<0dC_18cNi<^1SCdNm<_1nBcNR=W2O0000001O0000001O01O01O0000001O0000001O00101N1O1O1O2N1O1O1O2N1O1O101N1O1O1O2N1O1O1O2N1O100O2N1O1O1O2N11O00010O000010O01O00010O01O10O3N8G9H1N101O0O2O0O2O0O2O0O2O001N101N101N101N1N2N3M2O2M2N3M2N3N1N3M2N3M2N3N1N3M2N3L3M4M2M3M4L3M4M2M4L3M4M2M4L3MkXj1"}, "label": "the man in the plain with the white hat"}]}
{"id": 493626, "image": "COCO_train2014_000000493626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man inside of the plane\"."}], "task": "refering", "answer_template": "The \"a man inside of the plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 245, 246, 247, 248, 269, 270, 271, 293, 294], "bbox": [0.5630625, 0.2878362573099415, 0.8177031250000001, 0.7061988304093567], "iscrowd": 0, "area": 13379.097250000006, "rle": {"size": [513, 640], "counts": "cPe54l?2N3M3L4M2N3M3L4M1O2N2M3N2N101O1O1N101O1O1N2O1O2N2M3N2N3M2XCRNX;Q2dDQN\\;Q2aDPN_;R2]DQNb;Q2ZDQNf;R2UDPNj;S2RDoMn;S2oCoMP<0dC_18cNi<^1SCdNm<_1nBcNR=W2O0000001O0000001O01O01O0000001O0000001O00101N1O1O1O2N1O1O1O2N1O1O101N1O1O1O2N1O1O1O2N1O100O2N1O1O1O2N11O00010O000010O01O00010O01O10O3N8G9H1N101O0O2O0O2O0O2O0O2O001N101N101N101N1N2N3M2O2M2N3M2N3N1N3M2N3M2N3N1N3M2N3L3M4M2M3M4L3M4M2M4L3M4M2M4L3MkXj1"}, "label": "a man inside of the plane"}]}
{"id": 346178, "image": "COCO_train2014_000000346178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a baseball player wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 150, 151, 174, 175, 197, 198, 221, 222], "bbox": [0.536453125, 0.05292740046838408, 0.69875, 0.6317564402810304], "iscrowd": 0, "area": 10920.275599999999, "rle": {"size": [427, 640], "counts": "ea_46S=5kDFQ9?lFDn8b0oFBk8c0RG@i8e0SG_Og8g0VG\\Oe8h0YGZOc8S1RGlNm8b1fF_NX9g1dFYNZ9n1aFRN]9T2_FlM_9[2[FfMc9R3N2N2N2N2N2N2N2O1O5K4L5K5J5L5K5K4L5K5J5L5K3M1O1O1O1O1O1O1NmL`J_N_5\\1SKXNl4g1`KoM_4o1fKPNX4o1lKoMS4P2QLnMn3Q2WLlMh3R2]LlMc3S2`LoM[3R2gLPNU3o1nLRNn2n1UMTNg2k1\\MWN_2i1dMYNX2f1kMZNR2f1QNXNo1g1TNVNl1j1XNSNh1l1[NRNe1m1^NQNa1n1cNPN]1o1eNPNZ1o1iNPNW1o1jNQNU1n1nNPNS1o1oNPNP1o1SOPNm0o1TOQNk0o1WOPNi0S2UOlMk0W2ROiMn0[2oNdMR1^2lNaMU1a2iN^MW1e2gNZMZ1S3ZNkLg1U3ZNjLe1S1ZJj0S4PNd1T1\\Jj0l9UOYFg0g9XO_Fc0`9]OfF>Z9AmF8T9FoF9P9GRG8n8GSG8n8GSG8m8HTG8l8GUG8l8FWG8l8EUG:n8CSG=n:O2N100N2J6J7I6Jk[`2"}, "label": "a baseball player wearing a black shirt"}]}
{"id": 346178, "image": "COCO_train2014_000000346178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing black t shirt walking on a road\"."}], "task": "refering", "answer_template": "The \"a man wearing black t shirt walking on a road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 150, 151, 174, 175, 197, 198, 221, 222], "bbox": [0.536453125, 0.05292740046838408, 0.69875, 0.6317564402810304], "iscrowd": 0, "area": 10920.275599999999, "rle": {"size": [427, 640], "counts": "ea_46S=5kDFQ9?lFDn8b0oFBk8c0RG@i8e0SG_Og8g0VG\\Oe8h0YGZOc8S1RGlNm8b1fF_NX9g1dFYNZ9n1aFRN]9T2_FlM_9[2[FfMc9R3N2N2N2N2N2N2N2O1O5K4L5K5J5L5K5K4L5K5J5L5K3M1O1O1O1O1O1O1NmL`J_N_5\\1SKXNl4g1`KoM_4o1fKPNX4o1lKoMS4P2QLnMn3Q2WLlMh3R2]LlMc3S2`LoM[3R2gLPNU3o1nLRNn2n1UMTNg2k1\\MWN_2i1dMYNX2f1kMZNR2f1QNXNo1g1TNVNl1j1XNSNh1l1[NRNe1m1^NQNa1n1cNPN]1o1eNPNZ1o1iNPNW1o1jNQNU1n1nNPNS1o1oNPNP1o1SOPNm0o1TOQNk0o1WOPNi0S2UOlMk0W2ROiMn0[2oNdMR1^2lNaMU1a2iN^MW1e2gNZMZ1S3ZNkLg1U3ZNjLe1S1ZJj0S4PNd1T1\\Jj0l9UOYFg0g9XO_Fc0`9]OfF>Z9AmF8T9FoF9P9GRG8n8GSG8n8GSG8m8HTG8l8GUG8l8FWG8l8EUG:n8CSG=n:O2N100N2J6J7I6Jk[`2"}, "label": "a man wearing black t shirt walking on a road"}]}
{"id": 346178, "image": "COCO_train2014_000000346178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man running towards a base , covered in dirt\"."}], "task": "refering", "answer_template": "The \"a man running towards a base , covered in dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 197, 198, 216, 217, 219, 220, 221, 238, 239, 244, 261, 262], "bbox": [0.34784375, 0.12135831381733021, 0.634203125, 0.8134894613583139], "iscrowd": 0, "area": 17734.449250000005, "rle": {"size": [427, 640], "counts": "bYm22W=3M3M3N1N3O1O1O1O1O010O1O1O1O001M3M3M3M2N3M3L4M3M2N3M3M3L4M3M2N3M3M3L3NO1O10O0N2O2N1N2cM]MTJe2m5\\MnIf2S6\\MiIe2Y6\\McIf2]6\\M^If2d6[MXIf2i6\\MSIf2n6[MmHg2i4RMSL8POg2n4VMbJHb0=Ih2S5XM]JK`06Lh2X5[MWJ0=N1h2\\5_MPJ4<F3j2a5`MlI88@6j2g5mNlIZN6m2o5DjIa0V6X33M2O2M3M2O2N1O2O1O0O2O1N101O0O2O1O1N3N1N3N1O2M2O2N2N110O0010O03N4L3M4L3N2M000001O00000O2O000000O100jLjGU1W8jNkGU1T8kNnGS1S8lNoGR1R8mNPHR1P8lNSHR1n7mNTHQ1m7nNUHQ1k7nNVHQ1k7nNWHP1i7POYHo0j7lNYHR1j7kNXHS1l7iNVHV1l7gNVHW1n7eNTHY1o7dNSH[1P8aNRH]1Q8_NRH_1R8]NPHb1R8[NPHc1T8YNZI:h6CXI=l6_OTIa0P7ZOQIg0i900001O00001O00001O00001O00001O00001O001O00001O00001O1O2N1O1N3N1O1O1O2N11O100O1O1O1N2I7J6I7J6I7JVZQ3"}, "label": "a man running towards a base , covered in dirt"}]}
{"id": 346178, "image": "COCO_train2014_000000346178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in a dirty dodger ' s jersey\"."}], "task": "refering", "answer_template": "The \"a baseball player in a dirty dodger ' s jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 197, 198, 216, 217, 219, 220, 221, 238, 239, 244, 261, 262], "bbox": [0.34784375, 0.12135831381733021, 0.634203125, 0.8134894613583139], "iscrowd": 0, "area": 17734.449250000005, "rle": {"size": [427, 640], "counts": "bYm22W=3M3M3N1N3O1O1O1O1O010O1O1O1O001M3M3M3M2N3M3L4M3M2N3M3M3L4M3M2N3M3M3L3NO1O10O0N2O2N1N2cM]MTJe2m5\\MnIf2S6\\MiIe2Y6\\McIf2]6\\M^If2d6[MXIf2i6\\MSIf2n6[MmHg2i4RMSL8POg2n4VMbJHb0=Ih2S5XM]JK`06Lh2X5[MWJ0=N1h2\\5_MPJ4<F3j2a5`MlI88@6j2g5mNlIZN6m2o5DjIa0V6X33M2O2M3M2O2N1O2O1O0O2O1N101O0O2O1O1N3N1N3N1O2M2O2N2N110O0010O03N4L3M4L3N2M000001O00000O2O000000O100jLjGU1W8jNkGU1T8kNnGS1S8lNoGR1R8mNPHR1P8lNSHR1n7mNTHQ1m7nNUHQ1k7nNVHQ1k7nNWHP1i7POYHo0j7lNYHR1j7kNXHS1l7iNVHV1l7gNVHW1n7eNTHY1o7dNSH[1P8aNRH]1Q8_NRH_1R8]NPHb1R8[NPHc1T8YNZI:h6CXI=l6_OTIa0P7ZOQIg0i900001O00001O00001O00001O00001O00001O001O00001O00001O1O2N1O1N3N1O1O1O2N11O100O1O1O1N2I7J6I7J6I7JVZQ3"}, "label": "a baseball player in a dirty dodger ' s jersey"}]}
{"id": 518215, "image": "COCO_train2014_000000518215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle giraffee\"."}], "task": "refering", "answer_template": "The \"middle giraffee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 77, 100, 123, 124, 125, 146, 147, 148, 149, 170, 171, 172, 193, 194, 196, 216, 217, 219, 220, 239, 240, 242, 243, 262, 263, 265, 266], "bbox": [0.27978125, 0.1215, 0.57303125, 0.6900416666666666], "iscrowd": 0, "area": 11048.871150000003, "rle": {"size": [480, 640], "counts": "joc2;d>1O1O1O1O100O1O1O1O1O1O1O1O001O1O1N2N2N2M3N2O100O1O100O100001N3N201N2O0O100O10O0100O0010O01001N3N1N3N1N3M5K8H8H6J01O000000000O101O0000006J8H00001O00001O00kBgNb<Y1ZCSO^<m0^CYOa<c10001O01O0001O0kDbMg9^2oEPNl9Q2jE^NQ:c1dEkNX:U1^EYO3bNh9e4010O1O001O1O1O001O1O001O1O1O001O1gMlEDT::VF^Ok9?`FXO`9g0iFQOX9m0QGkNo8U1YGcNh8[1_201O0O2O0O2O001N2O001N101M2H9G8G:M2N3M2N3M2N3M2N3M2N3N1Ngd67Q[I9G9G:E:G9G7I7I6J7K4L4L1N2O1O1N21O001O001O0O2O4L4SNfCl0]<hNUDo0o;cNcDP1R=B>A`eo3"}, "label": "middle giraffee"}]}
{"id": 518215, "image": "COCO_train2014_000000518215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe extending its neck behind another giraffe ' s neck\"."}], "task": "refering", "answer_template": "The \"a giraffe extending its neck behind another giraffe ' s neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 98, 99, 100, 101, 119, 120, 121, 122, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236], "bbox": [0.1449375, 0.1646041666666667, 0.431453125, 0.64775], "iscrowd": 0, "area": 11754.360099999996, "rle": {"size": [480, 640], "counts": "Sk[1Q1e=f0ZO;D<E<F9L4L4MO01O1O100OjNV1A`01O11N10YDmNb9Y301N101N2O0O2O1N101N@`0D=K4M4K4M4L3L5L3M3L400O100010O1O001O1O001O10O01O1O001O1O1O010O1O00100O00100UEYN\\8i1WGcNh8^1SGgNm8Y1oFkNP9W1kFmNT9T1hFPOX9P1cFUO\\9m0_FXO_9i0]F[Oc9Q3O1O10VOcFQL\\9l3lFPLT9k3UGQLj8l3^GQL`8l3iGoKW8l3`1M4K4M4L3nMaD`0d;_O_D;d;E`D4d;K`DMd;3_DGd;8aDAc;>aD\\Oa;d0bDVOb;i0bDQO`;o0dDkN_;T1V101O001O001O00001O001O001O001O001O00001O001O001O001O001O1O3M4L3M4L3M4L3MVi20jVM3L5L3M3L4M1N101M2N2N3N12O0O2000O10O010000O0101O0O2O000O2N101N1O2N1O101N1O3M4M4KZ`Z5"}, "label": "a giraffe extending its neck behind another giraffe ' s neck"}]}
{"id": 518215, "image": "COCO_train2014_000000518215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large giraffe to the left of two others\"."}], "task": "refering", "answer_template": "The \"a large giraffe to the left of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 98, 99, 100, 101, 119, 120, 121, 122, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236], "bbox": [0.1449375, 0.1646041666666667, 0.431453125, 0.64775], "iscrowd": 0, "area": 11754.360099999996, "rle": {"size": [480, 640], "counts": "Sk[1Q1e=f0ZO;D<E<F9L4L4MO01O1O100OjNV1A`01O11N10YDmNb9Y301N101N2O0O2O1N101N@`0D=K4M4K4M4L3L5L3M3L400O100010O1O001O1O001O10O01O1O001O1O1O010O1O00100O00100UEYN\\8i1WGcNh8^1SGgNm8Y1oFkNP9W1kFmNT9T1hFPOX9P1cFUO\\9m0_FXO_9i0]F[Oc9Q3O1O10VOcFQL\\9l3lFPLT9k3UGQLj8l3^GQL`8l3iGoKW8l3`1M4K4M4L3nMaD`0d;_O_D;d;E`D4d;K`DMd;3_DGd;8aDAc;>aD\\Oa;d0bDVOb;i0bDQO`;o0dDkN_;T1V101O001O001O00001O001O001O001O001O00001O001O001O001O001O1O3M4L3M4L3M4L3MVi20jVM3L5L3M3L4M1N101M2N2N3N12O0O2000O10O010000O0101O0O2O000O2N101N1O2N1O101N1O3M4M4KZ`Z5"}, "label": "a large giraffe to the left of two others"}]}
{"id": 305224, "image": "COCO_train2014_000000305224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in jeans sitting on a green chair next to a lady with a laptop on a pink bean bag chair\"."}], "task": "refering", "answer_template": "The \"a person in jeans sitting on a green chair next to a lady with a laptop on a pink bean bag chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 60, 61, 62, 83, 84, 85, 106, 107, 108], "bbox": [0.361828125, 0.0015625, 0.753515625, 0.3060208333333333], "iscrowd": 0, "area": 13750.720549999998, "rle": {"size": [480, 640], "counts": "Qh\\31o>1O7I:F6J1O1O1O1O2N1O1O1O1O00000000001O00000000001O00000000001O00000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O0001O000000000001O00000000000000001O00000000000000001O00000000O1000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O10000000]BS1S;]101O001O00010O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00SO[E_Me:Z2dEeM[:S2oEkMQ:n1WFQNi9h1_FXN`9a1gF_NY9Z1oFeNQ9S1XGmNg8l0aGSO_8f0iGYOW8`0PHAo78YHGg71bHNY;08G0010O01O010O0010O01O100O100O1O10]mY2"}, "label": "a person in jeans sitting on a green chair next to a lady with a laptop on a pink bean bag chair"}]}
{"id": 305224, "image": "COCO_train2014_000000305224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person sitting on a light yellow color couch\"."}], "task": "refering", "answer_template": "The \"a person sitting on a light yellow color couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 60, 61, 62, 83, 84, 85, 106, 107, 108], "bbox": [0.361828125, 0.0015625, 0.753515625, 0.3060208333333333], "iscrowd": 0, "area": 13750.720549999998, "rle": {"size": [480, 640], "counts": "Qh\\31o>1O7I:F6J1O1O1O1O2N1O1O1O1O00000000001O00000000001O00000000001O00000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O0001O000000000001O00000000000000001O00000000000000001O00000000O1000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O10000000]BS1S;]101O001O00010O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O00SO[E_Me:Z2dEeM[:S2oEkMQ:n1WFQNi9h1_FXN`9a1gF_NY9Z1oFeNQ9S1XGmNg8l0aGSO_8f0iGYOW8`0PHAo78YHGg71bHNY;08G0010O01O010O0010O01O100O100O1O10]mY2"}, "label": "a person sitting on a light yellow color couch"}]}
{"id": 305224, "image": "COCO_train2014_000000305224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light colored cushion a man is sitting in partially obscured by a laptop\"."}], "task": "refering", "answer_template": "The \"light colored cushion a man is sitting in partially obscured by a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 56, 57, 58, 59, 60, 75, 76, 77, 80, 81, 82, 83, 99, 104], "bbox": [0.268515625, 0.0, 0.6240781249999999, 0.2543125], "iscrowd": 0, "area": 12421.718300000004, "rle": {"size": [480, 640], "counts": "Pd`28h>:F;E:F00001O000lClN\\:T1YEYOe:g0oDGo:9nDLP;4oDOo:1PE2n:OPE3o:MoD7o:IPE:n:FQE=m:CQEa0m:_OREd0l:\\OSEf0l:ZOSEi0k:WOSEm0k:SOTEn0l:ROREP1n:POPER1P;nNnDT1R;lNlDV1V;Y11O2N1O2N1O1O000000O10000O10000O1000000O10000ON3M3L5L5J5L4L4L5K4L4L4M3L5K4L4L4M4K4K5K5L5J4L5K4L5O001O1O001O010O001O1O0010O01O1O0010O001O1O1O100O1O1O001O0000001O00001O00001O00001O00001O0000001O00001O00001O0000001O000001O01O0000001O0O10001O002N4L3M4L4L4L6J9G9G001O001O001O001O001O001O000000000000000000000O1000000000O100000000000000000000000000000000000000000000O10O01O101N2O2M2O2M2O2M2O2M2O2M2O:Ed0]OZl`3"}, "label": "light colored cushion a man is sitting in partially obscured by a laptop"}]}
{"id": 305224, "image": "COCO_train2014_000000305224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow seat with someone in blue jeans sitting on it\"."}], "task": "refering", "answer_template": "The \"a yellow seat with someone in blue jeans sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 56, 57, 58, 59, 60, 75, 76, 77, 80, 81, 82, 83, 99, 104], "bbox": [0.268515625, 0.0, 0.6240781249999999, 0.2543125], "iscrowd": 0, "area": 12421.718300000004, "rle": {"size": [480, 640], "counts": "Pd`28h>:F;E:F00001O000lClN\\:T1YEYOe:g0oDGo:9nDLP;4oDOo:1PE2n:OPE3o:MoD7o:IPE:n:FQE=m:CQEa0m:_OREd0l:\\OSEf0l:ZOSEi0k:WOSEm0k:SOTEn0l:ROREP1n:POPER1P;nNnDT1R;lNlDV1V;Y11O2N1O2N1O1O000000O10000O10000O1000000O10000ON3M3L5L5J5L4L4L5K4L4L4M3L5K4L4L4M4K4K5K5L5J4L5K4L5O001O1O001O010O001O1O0010O01O1O0010O001O1O1O100O1O1O001O0000001O00001O00001O00001O00001O0000001O00001O00001O0000001O000001O01O0000001O0O10001O002N4L3M4L4L4L6J9G9G001O001O001O001O001O001O000000000000000000000O1000000000O100000000000000000000000000000000000000000000O10O01O101N2O2M2O2M2O2M2O2M2O2M2O:Ed0]OZl`3"}, "label": "a yellow seat with someone in blue jeans sitting on it"}]}
{"id": 305224, "image": "COCO_train2014_000000305224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an extra large , adult - sized , hot pink bean bag\"."}], "task": "refering", "answer_template": "The \"an extra large , adult - sized , hot pink bean bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 218, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 334, 353, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380], "bbox": [0.17778125, 0.21745833333333334, 0.8166718749999999, 0.9828749999999999], "iscrowd": 0, "area": 35429.809349999996, "rle": {"size": [480, 640], "counts": "Whe12l>3M4L3M3M3M3L5L3M3M3M4L3M3M3M3M4L2N2N3N101O01O01O001O01O01O001O01O01O001O01O01O00100O1O1O1O010O1O1O1O100O001O100O1O1O1O010O1O1O1O10O01O1O100O1O1O00100O1O1O1O10O01O1O1O10O01O1O100O001O1O100O001O1O1O001O100O1O001O1O1O1O010O1O1O1O001O1O100O001O1O1O1O010O1O1O1O001O1O100O001O1O1O001O100O1O001O1O1O1O001N2N2O1N101N2O1N2O0O2N2O1N2O0O2O1N2O1N101N2N2O0O2O0O100O2O0O100O1O101N100O100O2O0O100O1O101N100O100O1lKUJSOl5i0WJVOj5g0ZJWOg5f0]JYOc5d0`J[Oa5b0cJ]O]5`0fJ_O\\5=hJBb2YLVOR4\\NCZ2eLTOf3eNES2oLROY3oNFk1[MPOn2WO2Y1WM^Oe2\\Oc0a0mL2_2_OT1KbLe0X2Ce1SOXLZ1Q2FQ48nKJR46lKMT41mKOS41mK0R4OoK2Q4MoK3Q4MoK4Q4JQL5o3KQL6o3HRL9m3GSL9n3ETL:l3FTL:m3ESL:n3ESL;m3ESL;n3CTL<l3DTL<m3BTL>l3BTL=n3BSL=m3BTL>m3ASL?m3@TL`0l3@UL?l3_OUL`0l3@TL`0m3_OSLa0m3^OTLb0m3]OTLb0l3]OULc0l3\\OTLc0m3\\OTLd0l3\\OULc0l3\\OTLd0l3[OULe0l3ZOTLf0l3YOULf0m3YOTLf0l3YOULg0l3XOTLh0^LeN]6c0UMh0YLkNa6<WMi0RLROg64XMh0mKZOj6MZMi0fKAo6F[Mi0aKGS7_O]MR2c2mM^MS2a2mM_MT2`2kMaMV2^2jMbMW2^2gMcMZ2\\2fMeMX2\\2gMeMX2]2gMcMY2]2gMcMX2^2gMcMX2^2hMcMV2_2hMbMW2_2iMaMV2`2iMaMV2`2jMaMU2`2jM`MU2a2jM`MU2a2kM_MT2c2jM^MU2c2kM^MS2c2lM^MT2b2lM^MS2d2lM\\MS2e2lM\\MS2e2mM\\MQ2e2nM\\MQ2f2nMZMQ2k700O2N1O1O1O2N100O1O1O2N1O1O100O2N1O1O101O000001O01O0001O0001O01O0000010O0001O100O001O100O001O100O001O100O001O10O01O1O10O01O1O10O01O1O10O01O100O001O100O001O10O01O100O001O100O1O1O010O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O10O01O10ldf1"}, "label": "an extra large , adult - sized , hot pink bean bag"}]}
{"id": 305224, "image": "COCO_train2014_000000305224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink bean bag chair\"."}], "task": "refering", "answer_template": "The \"a pink bean bag chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 218, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 334, 353, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380], "bbox": [0.17778125, 0.21745833333333334, 0.8166718749999999, 0.9828749999999999], "iscrowd": 0, "area": 35429.809349999996, "rle": {"size": [480, 640], "counts": "Whe12l>3M4L3M3M3M3L5L3M3M3M4L3M3M3M3M4L2N2N3N101O01O01O001O01O01O001O01O01O001O01O01O00100O1O1O1O010O1O1O1O100O001O100O1O1O1O010O1O1O1O10O01O1O100O1O1O00100O1O1O1O10O01O1O1O10O01O1O100O001O1O100O001O1O1O001O100O1O001O1O1O1O010O1O1O1O001O1O100O001O1O1O1O010O1O1O1O001O1O100O001O1O1O001O100O1O001O1O1O1O001N2N2O1N101N2O1N2O0O2N2O1N2O0O2O1N2O1N101N2N2O0O2O0O100O2O0O100O1O101N100O100O2O0O100O1O101N100O100O1lKUJSOl5i0WJVOj5g0ZJWOg5f0]JYOc5d0`J[Oa5b0cJ]O]5`0fJ_O\\5=hJBb2YLVOR4\\NCZ2eLTOf3eNES2oLROY3oNFk1[MPOn2WO2Y1WM^Oe2\\Oc0a0mL2_2_OT1KbLe0X2Ce1SOXLZ1Q2FQ48nKJR46lKMT41mKOS41mK0R4OoK2Q4MoK3Q4MoK4Q4JQL5o3KQL6o3HRL9m3GSL9n3ETL:l3FTL:m3ESL:n3ESL;m3ESL;n3CTL<l3DTL<m3BTL>l3BTL=n3BSL=m3BTL>m3ASL?m3@TL`0l3@UL?l3_OUL`0l3@TL`0m3_OSLa0m3^OTLb0m3]OTLb0l3]OULc0l3\\OTLc0m3\\OTLd0l3\\OULc0l3\\OTLd0l3[OULe0l3ZOTLf0l3YOULf0m3YOTLf0l3YOULg0l3XOTLh0^LeN]6c0UMh0YLkNa6<WMi0RLROg64XMh0mKZOj6MZMi0fKAo6F[Mi0aKGS7_O]MR2c2mM^MS2a2mM_MT2`2kMaMV2^2jMbMW2^2gMcMZ2\\2fMeMX2\\2gMeMX2]2gMcMY2]2gMcMX2^2gMcMX2^2hMcMV2_2hMbMW2_2iMaMV2`2iMaMV2`2jMaMU2`2jM`MU2a2jM`MU2a2kM_MT2c2jM^MU2c2kM^MS2c2lM^MT2b2lM^MS2d2lM\\MS2e2lM\\MS2e2mM\\MQ2e2nM\\MQ2f2nMZMQ2k700O2N1O1O1O2N100O1O1O2N1O1O100O2N1O1O101O000001O01O0001O0001O01O0000010O0001O100O001O100O001O100O001O100O001O10O01O1O10O01O1O10O01O1O10O01O100O001O100O001O10O01O100O001O100O1O1O010O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O10O01O10ldf1"}, "label": "a pink bean bag chair"}]}
{"id": 51278, "image": "COCO_train2014_000000051278.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large white truck with a red tank in its bed\"."}], "task": "refering", "answer_template": "The \"a large white truck with a red tank in its bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 209, 211, 212, 213], "bbox": [0.09132812500000001, 0.3833255269320843, 0.321609375, 0.6357142857142857], "iscrowd": 0, "area": 9919.944450000003, "rle": {"size": [427, 640], "counts": "S]h05S=;G6J5\\O_OUDe0e;e0K2M4M2N3N2N1N2O1O101N1O1O10001OO1O2N1O1O1O100O100O101O00000O10000000O1O1M3E<K4L4O1000001O01O1XMgE\\2e:O1O0010O2O1N1O2N2N100O00001O00001N1000001O000000001O0000O1O1N2M3M3L4M3O100000001O0000001O0000001O00000000001O000000001O000004L5L0O000101N1O2N3M3M3M2N2N2N2N2N2M3N1N3L3K8K5K4L2N1O2O1N2N107H5Kagd5"}, "label": "a large white truck with a red tank in its bed"}]}
{"id": 51278, "image": "COCO_train2014_000000051278.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck with a white cab and a red tanker\"."}], "task": "refering", "answer_template": "The \"a truck with a white cab and a red tanker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 209, 211, 212, 213], "bbox": [0.09132812500000001, 0.3833255269320843, 0.321609375, 0.6357142857142857], "iscrowd": 0, "area": 9919.944450000003, "rle": {"size": [427, 640], "counts": "S]h05S=;G6J5\\O_OUDe0e;e0K2M4M2N3N2N1N2O1O101N1O1O10001OO1O2N1O1O1O100O100O101O00000O10000000O1O1M3E<K4L4O1000001O01O1XMgE\\2e:O1O0010O2O1N1O2N2N100O00001O00001N1000001O000000001O0000O1O1N2M3M3L4M3O100000001O0000001O0000001O00000000001O000000001O000004L5L0O000101N1O2N3M3M3M2N2N2N2N2N2M3N1N3L3K8K5K4L2N1O2O1N2N107H5Kagd5"}, "label": "a truck with a white cab and a red tanker"}]}
{"id": 542799, "image": "COCO_train2014_000000542799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman hitting a tennis ball with a racket\"."}], "task": "refering", "answer_template": "The \"a woman hitting a tennis ball with a racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 19, 20, 21, 22, 23, 35, 36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 73, 74, 75, 84, 85, 86, 87, 88, 89, 90, 91, 92, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 338, 339, 340, 341, 342, 343, 346, 347, 348, 349, 350, 351, 354, 355, 356, 357, 358, 363, 364, 365, 366, 367], "bbox": [0.0991685393258427, 0.03240625, 0.9786516853932584, 1.0], "iscrowd": 0, "area": 158825.81720000002, "rle": {"size": [640, 445], "counts": "mjk02jc08I8G8I7H9H7I8G8I5gEkM_2[2\\MhMa2]2ZMfMb2`2ZMbMc2d2WM_Me2g2VM\\Mg2i2TMZMh2l2TMVMi2o2RMTMk2Q3PMRMl2T3PMoLl2U3PMnLn2V3mLmLP3X3lLjLQ3[3jLhLP3`3kLcLP3d3lL^Ln2j3mLYLm2o3nLTLm2S4nLPLl2X4PMjKk2]4PMgKi2a4RMbKi2e4SM]Kg2i4VMZKd2l4YMWKb2n4\\MTK^2R5_MQK_2Q5^MRKa2o4\\MTKc2m4[MUKc2m4ZMVKe2k4XMYKf2h4XMZKg2g4WM[Kh2f4WM[Kh2f4WM[Kh2f4VM\\Kj2d4UM]Kj2d4UM]Kj2d4TM^Kl2b4SM_Kl2b4SM_Km2a4QMbKm2_4RMbKm2_4RMbKVOiL_1e7YOdKWOoLX1^7@dKVOXMR1T7GeKUO`Ml0l6MfKVOgMd0d65fKUOoM?[6;gKUOWN7S6b0hKUO[N5m5e0iKTO`N2h5i0jKSOcNOd5l0kKSOfNN_5n0lKROkNKZ5R1lKROoNHU5T1nKROSOEP5X1nKROVOBm4[1nKQO[O@g4]1PLSO]O[Od4a1PLSO@YO`4c1QLTOCTO]4f1RLUOFPOY4j1SLUOHmNU4l1ULVOR8i0oGVOR8i0oGWOQ8g0QHXOP8g0QHYOo7f0RHYOo7e0SH[Om7d0TH[Om7d0TH[Om7c0UH]Ok7b0WH\\Oj7c0WH]Oi7a0YH^Oh7a0YH_Og7`0ZH_Og7?[HAd7?]H@d7?]H@d7>^HBb7=_HBb7=_HCa7;bHC_7<bHD^7;cHD^7:dHE]7:dHF\\79eHG[77fHJZ75fHLZ73fHNZ70hH1W7MjH4V7IlH8T7FnH:R7CPI>o6ARI`0n6]OUId0j6ZOWIg0i6VOYIk0g6SO[Im0k<1O1O2O0O1O100O1O2O0O1O1O100O2N100O1O100O2N1O101N2N2O1N2N101N2N2O1N2N2N2O0O2N2O1N2N2O1N1O2N2O1N2N2O1N1O2O1N2N2N2O1N2N2O1N2N2O1N3M2O1N2N2N2O1O1O1O1O1O1O1O1O100O001O1O1O1O1O1M3N2N2N2N2N2N2M3N1O2N2N2N2N2M3N2O1O010O100O100O1O10O0100O100O100O010O100O01O001O001O00001N101O1O1O2N1O2N1O2N1O1O2M201O001N101O00001O001O001O001N101O00001O001O001O0O2O001O01O010O010N1O2N1O2N1O2N1O2N2`A\\K`=c5O2N1O1O2N1O1O2PGeIW4\\6gKgIW4Z6hKgIW4[6fKhIX4Y6fKiIY4X6fKjIX4X6eKjIZ4X6cKkI[4]6]KdIb4e6TK^Ij4j6mJXIR5m6iJUIU5n6fJUIY5n6cJUI[5n6`JUI_5n6\\JVIb5m6ZJUIe5n6VJVIh5m6SJVIl5m6PJVIn5m6mIWIQ6l6jI[IQ6i6jI\\IR6g6iI_IS6d6hIcIS6Y:K6K4L4L3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M4L3M3M3M3M3M2N1O2N1jM\\_On0e`0PO`_Ol0a`0QOe_Ok0]`0ROh_Oj0Y`0TOl_Oh0U`0VOP@f0R`0VOT@f0m?XOX@d0i?ZO\\@b0e?\\O`@`0b?]Oc@?^?^Oh@>Y?@l@<V?Ao@;^a0M4L4LRP5"}, "label": "a woman hitting a tennis ball with a racket"}]}
{"id": 542799, "image": "COCO_train2014_000000542799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman hitting a ball with her bat while playing teenis\"."}], "task": "refering", "answer_template": "The \"a woman hitting a ball with her bat while playing teenis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 19, 20, 21, 22, 23, 35, 36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 73, 74, 75, 84, 85, 86, 87, 88, 89, 90, 91, 92, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 338, 339, 340, 341, 342, 343, 346, 347, 348, 349, 350, 351, 354, 355, 356, 357, 358, 363, 364, 365, 366, 367], "bbox": [0.0991685393258427, 0.03240625, 0.9786516853932584, 1.0], "iscrowd": 0, "area": 158825.81720000002, "rle": {"size": [640, 445], "counts": "mjk02jc08I8G8I7H9H7I8G8I5gEkM_2[2\\MhMa2]2ZMfMb2`2ZMbMc2d2WM_Me2g2VM\\Mg2i2TMZMh2l2TMVMi2o2RMTMk2Q3PMRMl2T3PMoLl2U3PMnLn2V3mLmLP3X3lLjLQ3[3jLhLP3`3kLcLP3d3lL^Ln2j3mLYLm2o3nLTLm2S4nLPLl2X4PMjKk2]4PMgKi2a4RMbKi2e4SM]Kg2i4VMZKd2l4YMWKb2n4\\MTK^2R5_MQK_2Q5^MRKa2o4\\MTKc2m4[MUKc2m4ZMVKe2k4XMYKf2h4XMZKg2g4WM[Kh2f4WM[Kh2f4WM[Kh2f4VM\\Kj2d4UM]Kj2d4UM]Kj2d4TM^Kl2b4SM_Kl2b4SM_Km2a4QMbKm2_4RMbKm2_4RMbKVOiL_1e7YOdKWOoLX1^7@dKVOXMR1T7GeKUO`Ml0l6MfKVOgMd0d65fKUOoM?[6;gKUOWN7S6b0hKUO[N5m5e0iKTO`N2h5i0jKSOcNOd5l0kKSOfNN_5n0lKROkNKZ5R1lKROoNHU5T1nKROSOEP5X1nKROVOBm4[1nKQO[O@g4]1PLSO]O[Od4a1PLSO@YO`4c1QLTOCTO]4f1RLUOFPOY4j1SLUOHmNU4l1ULVOR8i0oGVOR8i0oGWOQ8g0QHXOP8g0QHYOo7f0RHYOo7e0SH[Om7d0TH[Om7d0TH[Om7c0UH]Ok7b0WH\\Oj7c0WH]Oi7a0YH^Oh7a0YH_Og7`0ZH_Og7?[HAd7?]H@d7?]H@d7>^HBb7=_HBb7=_HCa7;bHC_7<bHD^7;cHD^7:dHE]7:dHF\\79eHG[77fHJZ75fHLZ73fHNZ70hH1W7MjH4V7IlH8T7FnH:R7CPI>o6ARI`0n6]OUId0j6ZOWIg0i6VOYIk0g6SO[Im0k<1O1O2O0O1O100O1O2O0O1O1O100O2N100O1O100O2N1O101N2N2O1N2N101N2N2O1N2N2N2O0O2N2O1N2N2O1N1O2N2O1N2N2O1N1O2O1N2N2N2O1N2N2O1N2N2O1N3M2O1N2N2N2O1O1O1O1O1O1O1O1O100O001O1O1O1O1O1M3N2N2N2N2N2N2M3N1O2N2N2N2N2M3N2O1O010O100O100O1O10O0100O100O100O010O100O01O001O001O00001N101O1O1O2N1O2N1O2N1O1O2M201O001N101O00001O001O001O001N101O00001O001O001O0O2O001O01O010O010N1O2N1O2N1O2N1O2N2`A\\K`=c5O2N1O1O2N1O1O2PGeIW4\\6gKgIW4Z6hKgIW4[6fKhIX4Y6fKiIY4X6fKjIX4X6eKjIZ4X6cKkI[4]6]KdIb4e6TK^Ij4j6mJXIR5m6iJUIU5n6fJUIY5n6cJUI[5n6`JUI_5n6\\JVIb5m6ZJUIe5n6VJVIh5m6SJVIl5m6PJVIn5m6mIWIQ6l6jI[IQ6i6jI\\IR6g6iI_IS6d6hIcIS6Y:K6K4L4L3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M4L3M3M3M3M3M2N1O2N1jM\\_On0e`0PO`_Ol0a`0QOe_Ok0]`0ROh_Oj0Y`0TOl_Oh0U`0VOP@f0R`0VOT@f0m?XOX@d0i?ZO\\@b0e?\\O`@`0b?]Oc@?^?^Oh@>Y?@l@<V?Ao@;^a0M4L4LRP5"}, "label": "a woman hitting a ball with her bat while playing teenis"}]}
{"id": 542799, "image": "COCO_train2014_000000542799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blurred spectator wearing white in the stands area that is visible to the right of her ponytail\"."}], "task": "refering", "answer_template": "The \"the blurred spectator wearing white in the stands area that is visible to the right of her ponytail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 27, 28, 29, 30, 31, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 91, 92, 93, 94], "bbox": [0.6625168539325843, 0.0, 0.9449213483146067, 0.25828125], "iscrowd": 0, "area": 14975.959, "rle": {"size": [640, 445], "counts": "X_h56fc08H8H8H8H8H8H9I5K3N2M3M3N1N3N2M3N2M2O2N2N2N1O2N2N2N2N1O2N2N1O1O2N100O2N1O2N1O1O2BTLe@m3W?[Ld@f3X?d0K4L5K4L400000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O11O00001O001O00001O001O00001O0000YMSA<l>CaA1_>NfAOY>0kAMU>2oAKR>3RBKn=3WBIk=4YBJh=3\\BJf=3^BJc=4aBJ`=3eBI]=4gBIZ=5jBHW=5nBEW=7nBCV=;nB@U==PC]OT=`0QCZOT=c0V3M3K6JZc?"}, "label": "the blurred spectator wearing white in the stands area that is visible to the right of her ponytail"}]}
{"id": 542799, "image": "COCO_train2014_000000542799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white blur in the upper right corner\"."}], "task": "refering", "answer_template": "The \"the white blur in the upper right corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 27, 28, 29, 30, 31, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 91, 92, 93, 94], "bbox": [0.6625168539325843, 0.0, 0.9449213483146067, 0.25828125], "iscrowd": 0, "area": 14975.959, "rle": {"size": [640, 445], "counts": "X_h56fc08H8H8H8H8H8H9I5K3N2M3M3N1N3N2M3N2M2O2N2N2N1O2N2N2N2N1O2N2N1O1O2N100O2N1O2N1O1O2BTLe@m3W?[Ld@f3X?d0K4L5K4L400000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O11O00001O001O00001O001O00001O0000YMSA<l>CaA1_>NfAOY>0kAMU>2oAKR>3RBKn=3WBIk=4YBJh=3\\BJf=3^BJc=4aBJ`=3eBI]=4gBIZ=5jBHW=5nBEW=7nBCV=;nB@U==PC]OT=`0QCZOT=c0V3M3K6JZc?"}, "label": "the white blur in the upper right corner"}]}
{"id": 182353, "image": "COCO_train2014_000000182353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with her arm around a man\"."}], "task": "refering", "answer_template": "The \"a woman with her arm around a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 181, 182, 183, 184, 196, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 276, 277, 278, 279, 280, 295, 296, 297, 298, 314, 315, 316, 317, 333, 334, 335, 336, 353, 354], "bbox": [0.3251320754716981, 0.43128301886792453, 0.7413773584905661, 1.0], "iscrowd": 0, "area": 28189.892499999994, "rle": {"size": [530, 530], "counts": "Pki29X`08H8H9G8H2N1O1O1O1000000001O00001O01O0001O00001O0000001O00001O0001O01O00001O0000001O00001O0001O01O00001O00000000O1O100O100O2OO01O100O010O1O010O10O01O10O0100O00100O010O1O010O10O01O100O010O1O0O2O1O0O2O1O0O2O1O0O2O1N101O1N101O1N2O001N2O0O2N2N1O2N2N1O2N;E<^FhLT5d3WJjLe5a3fImLU6_3WInLe6]3fHQMV7Z3UHSMi7n5L4M3L4M3L4M3L4M3L4M2N2N01O1O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001gNVHhIj7W6]HcId7[6bH`I_7^6gH]IY7b6mHYIT7e6RIVIn6j6WIQIj6m6_1N1O2N2N2N1N3N2M2N6J8I7H8H8H8H8I8E:F:G9F:F:G9F:F:F<E<C=C=D<Cd`V2"}, "label": "a woman with her arm around a man"}]}
{"id": 182353, "image": "COCO_train2014_000000182353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"people hugging\"."}], "task": "refering", "answer_template": "The \"people hugging\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 181, 182, 183, 184, 196, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 276, 277, 278, 279, 280, 295, 296, 297, 298, 314, 315, 316, 317, 333, 334, 335, 336, 353, 354], "bbox": [0.3251320754716981, 0.43128301886792453, 0.7413773584905661, 1.0], "iscrowd": 0, "area": 28189.892499999994, "rle": {"size": [530, 530], "counts": "Pki29X`08H8H9G8H2N1O1O1O1000000001O00001O01O0001O00001O0000001O00001O0001O01O00001O0000001O00001O0001O01O00001O00000000O1O100O100O2OO01O100O010O1O010O10O01O10O0100O00100O010O1O010O10O01O100O010O1O0O2O1O0O2O1O0O2O1O0O2O1N101O1N101O1N2O001N2O0O2N2N1O2N2N1O2N;E<^FhLT5d3WJjLe5a3fImLU6_3WInLe6]3fHQMV7Z3UHSMi7n5L4M3L4M3L4M3L4M3L4M2N2N01O1O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001gNVHhIj7W6]HcId7[6bH`I_7^6gH]IY7b6mHYIT7e6RIVIn6j6WIQIj6m6_1N1O2N2N2N1N3N2M2N6J8I7H8H8H8H8I8E:F:G9F:F:G9F:F:F<E<C=C=D<Cd`V2"}, "label": "people hugging"}]}
{"id": 182353, "image": "COCO_train2014_000000182353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a white jumper\"."}], "task": "refering", "answer_template": "The \"a person in a white jumper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 183, 184, 185, 195, 196, 197, 198, 199, 216, 217, 233, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 257, 271, 272, 273, 274, 275, 276, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352], "bbox": [0.2747169811320755, 0.3538490566037736, 0.7802264150943395, 1.0], "iscrowd": 0, "area": 46443.53555000001, "rle": {"size": [530, 530], "counts": "em[28X`09F:iJ^OnIj0j51[I8\\6a0jHIm6=mHKj6<oHNh68RI1f64TI4f6OTI;e6HVI`0d6CVIg0c6\\OXIm0a6WOYIQ1a6ROYIX1`6kN^I[1[6hNfIi1e5ZN[J\\2l4hMQKW2o4lMnJT2R5lMnJT2R5kMoJU2Q5jMQKU2o4jMRKV2n4jMRKV2n4iMSKW2n4gMTKX2l4hMTKX2l4gMUKY2k4fMVKZ2j4fMWKY2i4fMXKZ2h4eMYK[2g4dMZK\\2f4dM[K[2e4eM[K[2e4eM[K[2e4eM\\KZ2d4gM[KX2f4hMZKX2f4hMZKX2f4iMZKV2f4jMZKV2f4jMZKV2f4kMYKU2g4kMZKT2f4lMZKT2f4mMYKS2g4mMYKS2g4mMZKR2f4oMYKQ2g4oMYKQ2g4oMYKQ2g4oMZKP2f4QNYKo1g4QNYKo1g4QNYKo1g4RNYKm1g4SNYKm1g4SNYKm1g4TNXKl1h4TNYKk1g4UNYKk1g4VNXKj1h4VNXKj1h4VNYKi1g4XNXKe1k4[NUK\\1T5dNlJR1^5oNbJg0g5YOYJ>P6BPJ;S6FlI:T6FmI9S6GmI9S6GmI:R6GmI9S6GnI8R6HnI8R6HnI8R6HnI8R6HoI8LPLk4g3ZK9XOcL^5T3ZK9fNUMP6b2ZKb1f4]N\\Kb1e4]N[Kc1e4]NZKd1f4[N[Ke1e4[NZKf1f4YN[Kg1e4YN[Kg1e4YNZKh1f4WN[Ki1e4WNZKj1f4UN[Kj1f4UN[Kk1e4UNZKl1f4SN[Km1e4RN[Ko1e4PN\\KP2d4oM]KQ2c4nM]KS2c4lM^KT2b4kM^KV2b4iM_KW2a4hM`KX2`4gM`KZ2`4eMaK[2_4dMaK]2_4bMbK^2^4aMcK_2]4`McKa2]4^MdKb2\\4^McKc2]4\\MdKd2\\4[MeKe2[4ZMeKg2[4XMfKh2Z4WMfKj2Z4UMgKk2Y4TMhKl2X4SMhKn2X4QMiKo2W4PMiKQ3W4nLjKR3V4mLkKS3U4lLkKU3U4jLlKV3T4iLlKX3T4gLmKY3S4fLnKZ3R4fLmK[3S4dLnK\\3R4cLnK^3R4aLoK_3Q4`LPL\\3T4cLlKV3\\4iLeKo2c4PM]Ki2k4VMVKb2R5[MQK]2W5`MkJ^2X5^MlJa2U5\\MmJc2U5YMoJf2S5VMPKi2Q5SMRKl2P5SMQKl2P5SMPKm2Q5RMUKg2m4XM_KZ2d4fMgKi1_4VNlKZ1Z4eNlKP1Z4oNlKf0Z4ZOjK<\\4CiK3]4LhKI_46fK@`4`0eKUOa4j0dKlNb4S1_5O100O1O11O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O10O01O00001O0O1O2O0O1O2O0O1O2O0O1O2O000100O00100O1O010O1O10O01O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O0K6J8H8H8H8H8HWjk1"}, "label": "a person in a white jumper"}]}
{"id": 182353, "image": "COCO_train2014_000000182353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white puffy and striped sweater\"."}], "task": "refering", "answer_template": "The \"a white puffy and striped sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 183, 184, 185, 195, 196, 197, 198, 199, 216, 217, 233, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 257, 271, 272, 273, 274, 275, 276, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352], "bbox": [0.2747169811320755, 0.3538490566037736, 0.7802264150943395, 1.0], "iscrowd": 0, "area": 46443.53555000001, "rle": {"size": [530, 530], "counts": "em[28X`09F:iJ^OnIj0j51[I8\\6a0jHIm6=mHKj6<oHNh68RI1f64TI4f6OTI;e6HVI`0d6CVIg0c6\\OXIm0a6WOYIQ1a6ROYIX1`6kN^I[1[6hNfIi1e5ZN[J\\2l4hMQKW2o4lMnJT2R5lMnJT2R5kMoJU2Q5jMQKU2o4jMRKV2n4jMRKV2n4iMSKW2n4gMTKX2l4hMTKX2l4gMUKY2k4fMVKZ2j4fMWKY2i4fMXKZ2h4eMYK[2g4dMZK\\2f4dM[K[2e4eM[K[2e4eM[K[2e4eM\\KZ2d4gM[KX2f4hMZKX2f4hMZKX2f4iMZKV2f4jMZKV2f4jMZKV2f4kMYKU2g4kMZKT2f4lMZKT2f4mMYKS2g4mMYKS2g4mMZKR2f4oMYKQ2g4oMYKQ2g4oMYKQ2g4oMZKP2f4QNYKo1g4QNYKo1g4QNYKo1g4RNYKm1g4SNYKm1g4SNYKm1g4TNXKl1h4TNYKk1g4UNYKk1g4VNXKj1h4VNXKj1h4VNYKi1g4XNXKe1k4[NUK\\1T5dNlJR1^5oNbJg0g5YOYJ>P6BPJ;S6FlI:T6FmI9S6GmI9S6GmI:R6GmI9S6GnI8R6HnI8R6HnI8R6HnI8R6HoI8LPLk4g3ZK9XOcL^5T3ZK9fNUMP6b2ZKb1f4]N\\Kb1e4]N[Kc1e4]NZKd1f4[N[Ke1e4[NZKf1f4YN[Kg1e4YN[Kg1e4YNZKh1f4WN[Ki1e4WNZKj1f4UN[Kj1f4UN[Kk1e4UNZKl1f4SN[Km1e4RN[Ko1e4PN\\KP2d4oM]KQ2c4nM]KS2c4lM^KT2b4kM^KV2b4iM_KW2a4hM`KX2`4gM`KZ2`4eMaK[2_4dMaK]2_4bMbK^2^4aMcK_2]4`McKa2]4^MdKb2\\4^McKc2]4\\MdKd2\\4[MeKe2[4ZMeKg2[4XMfKh2Z4WMfKj2Z4UMgKk2Y4TMhKl2X4SMhKn2X4QMiKo2W4PMiKQ3W4nLjKR3V4mLkKS3U4lLkKU3U4jLlKV3T4iLlKX3T4gLmKY3S4fLnKZ3R4fLmK[3S4dLnK\\3R4cLnK^3R4aLoK_3Q4`LPL\\3T4cLlKV3\\4iLeKo2c4PM]Ki2k4VMVKb2R5[MQK]2W5`MkJ^2X5^MlJa2U5\\MmJc2U5YMoJf2S5VMPKi2Q5SMRKl2P5SMQKl2P5SMPKm2Q5RMUKg2m4XM_KZ2d4fMgKi1_4VNlKZ1Z4eNlKP1Z4oNlKf0Z4ZOjK<\\4CiK3]4LhKI_46fK@`4`0eKUOa4j0dKlNb4S1_5O100O1O11O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O10O01O00001O0O1O2O0O1O2O0O1O2O0O1O2O000100O00100O1O010O1O10O01O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O0K6J8H8H8H8H8HWjk1"}, "label": "a white puffy and striped sweater"}]}
{"id": 436306, "image": "COCO_train2014_000000436306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle banana in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the middle banana in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 195, 196, 213, 214, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264], "bbox": [0.265671875, 0.31243750000000003, 0.556765625, 0.5192187500000001], "iscrowd": 0, "area": 7973.177449999998, "rle": {"size": [640, 640], "counts": "d`Z31nc01O2N2N2N2O1N3M4L3M4M3N1O2M3M100O1O010O1O100O1O10O01O100O1O100O00100O010O00010O0010O0010O00010O010O00010O00010O010O00010O0001O00001O0001O0001O000000001O00000010O00000001O000000000000O1000000O01000O1000O010000O10O1000O1000O10O10000O0100000O1000O01O1O1O001O00001O00001O00001O00000010O0001O0O101O000O101O00001N10000N2N2N2N2N1N3L4M3N2M3N20O2O1O1O1\\NX_OOh`0O`_OK^`05j_OET`0;T@_Ol?`0[@YOg?f0k1O0001O0002N2O2M`Ya5"}, "label": "the middle banana in the right hand picture"}]}
{"id": 436306, "image": "COCO_train2014_000000436306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana in the middle , tied to elbow\"."}], "task": "refering", "answer_template": "The \"banana in the middle , tied to elbow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 195, 196, 213, 214, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264], "bbox": [0.265671875, 0.31243750000000003, 0.556765625, 0.5192187500000001], "iscrowd": 0, "area": 7973.177449999998, "rle": {"size": [640, 640], "counts": "d`Z31nc01O2N2N2N2O1N3M4L3M4M3N1O2M3M100O1O010O1O100O1O10O01O100O1O100O00100O010O00010O0010O0010O00010O010O00010O00010O010O00010O0001O00001O0001O0001O000000001O00000010O00000001O000000000000O1000000O01000O1000O010000O10O1000O1000O10O10000O0100000O1000O01O1O1O001O00001O00001O00001O00000010O0001O0O101O000O101O00001N10000N2N2N2N2N1N3L4M3N2M3N20O2O1O1O1\\NX_OOh`0O`_OK^`05j_OET`0;T@_Ol?`0[@YOg?f0k1O0001O0002N2O2M`Ya5"}, "label": "banana in the middle , tied to elbow"}]}
{"id": 477266, "image": "COCO_train2014_000000477266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a red plaid shirt and baseball cap sticking his tongue out\"."}], "task": "refering", "answer_template": "The \"a guy in a red plaid shirt and baseball cap sticking his tongue out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 38, 39, 40, 41, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202], "bbox": [0.00754, 0.1574550898203593, 0.31062, 0.9791317365269461], "iscrowd": 0, "area": 32056.11714999999, "rle": {"size": [334, 500], "counts": "Z^1[1Q9n1SNm1SNc0\\O2O1O1N2O1O1N20000000O10000000000000001O00000000000000000000O1I7H8H7K6N2N2N2N2N2N2O1N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O001O1N2O1O1O1O1O1O1O1O1O10000O1000000000000000O10O10000000000000000001O00001O001O00001O00001N101O00001O001O000\\OlMnHU2l6RNRIn1i6YNUIh1f6^NXIb1c6eN[I[1`6mN]IT1]6SOaIm0[6f1J5K5K6K4K6J5K5gMlJTOY5j0jJoN[5o0iJjN]5S1gJgN]5W1gJbN^5]1dJ]Nb5`1bJYNc5e1aJUNd5i1_JPNf5n1^JkMg5S2`1I8I7H8H8H7J7H8H8HU]`3"}, "label": "a guy in a red plaid shirt and baseball cap sticking his tongue out"}]}
{"id": 477266, "image": "COCO_train2014_000000477266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with his mouth open wide\"."}], "task": "refering", "answer_template": "The \"a man with his mouth open wide\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 38, 39, 40, 41, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202], "bbox": [0.00754, 0.1574550898203593, 0.31062, 0.9791317365269461], "iscrowd": 0, "area": 32056.11714999999, "rle": {"size": [334, 500], "counts": "Z^1[1Q9n1SNm1SNc0\\O2O1O1N2O1O1N20000000O10000000000000001O00000000000000000000O1I7H8H7K6N2N2N2N2N2N2O1N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O001O1N2O1O1O1O1O1O1O1O1O10000O1000000000000000O10O10000000000000000001O00001O001O00001O00001N101O00001O001O000\\OlMnHU2l6RNRIn1i6YNUIh1f6^NXIb1c6eN[I[1`6mN]IT1]6SOaIm0[6f1J5K5K6K4K6J5K5gMlJTOY5j0jJoN[5o0iJjN]5S1gJgN]5W1gJbN^5]1dJ]Nb5`1bJYNc5e1aJUNd5i1_JPNf5n1^JkMg5S2`1I8I7H8H8H7J7H8H8HU]`3"}, "label": "a man with his mouth open wide"}]}
{"id": 477266, "image": "COCO_train2014_000000477266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women was seeing cell phone\"."}], "task": "refering", "answer_template": "The \"a women was seeing cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 50, 51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.56124, 0.1438323353293413, 0.99132, 0.9885928143712575], "iscrowd": 0, "area": 38894.86595000001, "rle": {"size": [334, 500], "counts": "Sok25S:6I7J6J6J6J6J6J6J5K6M3M3N2M3M3M3M3M3M3O1N2O1N2O0O2O1N2O1N2N2O1N2O1O100O1O10O01O100O1O100O1O100O1O100O2N100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O2N100O100O1O100O1O100O1O100O100O1O100N2N2_OXKZKj4]4`K_Kc4X4gKeK[4S4T1J6I7K5M3N3L3M3M3N2M3M3N2M3K5K5K5K5L400O1O1O100O1O1O100O1O100O10000000000000000000000O100000000000000000000000000000000000O10000000000000000]L]N[Nc1c1`N\\Na1`1bN`N_1\\1eNcN\\1Z1gNeNZ1W1iNiNX1S1lNlNU1Q1mNoNT1m0POROQ1k0ROTOo0i0SOWOn0g0TOXOm0e0UO[Ol0c0VO\\Ok0a0WO_Oj0?XO@i0=YOCh0;YOEh08[OGf07[OIf04]OKd03]OMd0N@2a0JC5>FF:;BI=8^OLb05ZONf03UO2j0ORO4n0MmN8R1IjN:V1KaN9_10UN4j15iMOW2:]MJb2?QMEo2m3I8H7IeR1"}, "label": "a women was seeing cell phone"}]}
{"id": 477266, "image": "COCO_train2014_000000477266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a brown shirt looking at her phone\"."}], "task": "refering", "answer_template": "The \"a woman in a brown shirt looking at her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 50, 51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.56124, 0.1438323353293413, 0.99132, 0.9885928143712575], "iscrowd": 0, "area": 38894.86595000001, "rle": {"size": [334, 500], "counts": "Sok25S:6I7J6J6J6J6J6J6J5K6M3M3N2M3M3M3M3M3M3O1N2O1N2O0O2O1N2O1N2N2O1N2O1O100O1O10O01O100O1O100O1O100O1O100O2N100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O2N100O100O1O100O1O100O1O100O100O1O100N2N2_OXKZKj4]4`K_Kc4X4gKeK[4S4T1J6I7K5M3N3L3M3M3N2M3M3N2M3K5K5K5K5L400O1O1O100O1O1O100O1O100O10000000000000000000000O100000000000000000000000000000000000O10000000000000000]L]N[Nc1c1`N\\Na1`1bN`N_1\\1eNcN\\1Z1gNeNZ1W1iNiNX1S1lNlNU1Q1mNoNT1m0POROQ1k0ROTOo0i0SOWOn0g0TOXOm0e0UO[Ol0c0VO\\Ok0a0WO_Oj0?XO@i0=YOCh0;YOEh08[OGf07[OIf04]OKd03]OMd0N@2a0JC5>FF:;BI=8^OLb05ZONf03UO2j0ORO4n0MmN8R1IjN:V1KaN9_10UN4j15iMOW2:]MJb2?QMEo2m3I8H7IeR1"}, "label": "a woman in a brown shirt looking at her phone"}]}
{"id": 477266, "image": "COCO_train2014_000000477266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the middle with the polka dot shirt\"."}], "task": "refering", "answer_template": "The \"the girl in the middle with the polka dot shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 24, 25, 26, 27, 42, 43, 44, 45, 60, 61, 62, 63, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208], "bbox": [0.22416, 0.06062874251497006, 0.6426000000000001, 1.0], "iscrowd": 0, "area": 43661.82150000001, "rle": {"size": [334, 500], "counts": "ghT13W:6I7J7I6J6J6J7I6J6K5K6J5aIeMe4_2WKcMf4b2VK`Mg4f2TK\\Mi4i2RKZMl4j2PKXMm4m2oJUMn4Q3mJQMQ5S3kJoLR5V3jJlLS5Y3iJiLV5[3eJgLZ5\\3bJfL]5V4O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O100PN[JWOe5e0bJXO^5d0jJXOV5e0PKXOP5d0XKWOi4e0^KXOb4d0fKXOZ4e0mKWOS4e0ULWOk3e0\\LXOd3f0bLUO_3k0eLQO[3n0jLnNV3Q1oLkNQ3T1TMhNl2W1V3O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100000001O000000000000000000000001O00000000000000000000000000000000000000000000000000001O1O2N2N1O2N2N1O2N2O0O2N2N1O2N2N3M5K5K5K5K5K4L5K5K4L4L3M4L4L4L4L4L4J8Gj0UOa0@4K4M4K5L4L4K5L4K5L4K5L4K5L4K5L4L4K5L3L5L4K5L4K5M4L4K5L3M4L4L4K5L3M4L4L4KkUj1"}, "label": "the girl in the middle with the polka dot shirt"}]}
{"id": 477266, "image": "COCO_train2014_000000477266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a dotted black dress holds a green beer bottle\"."}], "task": "refering", "answer_template": "The \"a woman wearing a dotted black dress holds a green beer bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 24, 25, 26, 27, 42, 43, 44, 45, 60, 61, 62, 63, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208], "bbox": [0.22416, 0.06062874251497006, 0.6426000000000001, 1.0], "iscrowd": 0, "area": 43661.82150000001, "rle": {"size": [334, 500], "counts": "ghT13W:6I7J7I6J6J6J7I6J6K5K6J5aIeMe4_2WKcMf4b2VK`Mg4f2TK\\Mi4i2RKZMl4j2PKXMm4m2oJUMn4Q3mJQMQ5S3kJoLR5V3jJlLS5Y3iJiLV5[3eJgLZ5\\3bJfL]5V4O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O100PN[JWOe5e0bJXO^5d0jJXOV5e0PKXOP5d0XKWOi4e0^KXOb4d0fKXOZ4e0mKWOS4e0ULWOk3e0\\LXOd3f0bLUO_3k0eLQO[3n0jLnNV3Q1oLkNQ3T1TMhNl2W1V3O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100000001O000000000000000000000001O00000000000000000000000000000000000000000000000000001O1O2N2N1O2N2N1O2N2O0O2N2N1O2N2N3M5K5K5K5K5K4L5K5K4L4L3M4L4L4L4L4L4J8Gj0UOa0@4K4M4K5L4L4K5L4K5L4K5L4K5L4K5L4L4K5L3L5L4K5L4K5M4L4K5L3M4L4L4K5L3M4L4L4KkUj1"}, "label": "a woman wearing a dotted black dress holds a green beer bottle"}]}
{"id": 59483, "image": "COCO_train2014_000000059483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange glimmering in light making it look darker and lighter in spots\"."}], "task": "refering", "answer_template": "The \"an orange glimmering in light making it look darker and lighter in spots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 240, 241, 242, 243, 264, 265], "bbox": [0.297015625, 0.33778554778554776, 0.7100625, 0.7766899766899766], "iscrowd": 0, "area": 23279.112549999998, "rle": {"size": [429, 640], "counts": "Tk_22Y=2M4M2M4M2M4M2M3N3L3N2M3N2O1O100000010O0DiNbDW1\\;kNdDU1[;lNeDT1Z;mNfDS1Y;oNfDQ1X;QOhDo0W;ROiDn0V;SOjDn0T;SOlDm0S;TOmDl0R;UOnDk0P;WOQEh0n:YOREg0m:ZOSEf0l:[OTEe0k:\\OUEd0j:^OUEc0h:_OXEa0g:@YE`0f:AZE?e:B[E>d:C\\E=c:D]E<b:E^E<`:E`E;`:EaE:^:GbE9]:HcE8]:HcE6^:KbE1a:0_EMc:3^EJe:6[EFh:;XEBj:?VE^Om:b0SE[Oo:f0QEVOR;k0nDROT;o0mDmNV;S1a000100O100OO101O0O2O001O02N101N2O0TFgNc7[1`GdNSO`0\\9m0bG_O^8b0QGnN[O?c9d0PGPO]O<b9e0nFQO@:a9g0lFROC6a9h0lFSOC5_9j0nFPOD5^9l0mFPOE3^9m0nFPOC4^9m0nFPOD2]9o0PGnND2\\9Q1oFnNE0\\9S1oFmNE0Z9T1RGkND1Z9U1QGkNE0Y9V1RGjNEOY9W1RGjNFNW9Z1SGhNFNV9Z1TGiNEMW9[1TGgNFNU9\\1UGfNJIP9b1VGgN5ZOe8P2VGhN_9Y1`FjN\\9W1eFlNW9U1hFnNU9R1lFVOk8k0UG[Oc8g0\\GZOc8f0^GZOa8g0^GZOa8g0_GYO_8h0aGYO^8h0bGXO]8h0dGXO[8i0dGXOZ8j0fGUOZ8k0fGVOY8k0gGUOX8l0gGUOX8k0iGUOV8l0iGUOV8k0kGUOS8m0lGTOS8m0mGSOR8m0nGTOQ8m0oGSOP8m0PHSOP8n0PHROo7n0QHSOm7n0THROk7o0THROk7n0VHROi7o0WHQOh7o0XHROg7o0YHQOf7o0ZHROd7P1\\HoNd7Q1\\HPOc7Q1]HoNb7Q1^HPOa7Q1_HQO^7o0bHSO\\7n0dHTOY7l0gHVOW7k0iHVOU7j0kHYOR7h0\\JiMd5W2bJcM^5^2hJZMZ5e2lJUMT5k2SKnLm4R3g2OjFQMP8o2`GmMT8S2lGYNh7g1YHeNZ7[1gHmNP7S1PImNP7S1QImNn6S1SImNl6S1TInNk6R1VImNj6R1XInNg6S1XIiNl6X1TIcNP7]1QI]NT7d1lHWNY7i1fHSN^7n1bHlMc7U2]HfMg7[2XHaMl7`2TH^Mm7c2SH[Mn7f2QHYMP8h2PHVMQ8k2oGRMS8o2lGPMU8P3lGnLU8S3kGkLV8V3iGiLX8X3hGfLY8[3gGbL[8_3eG_L\\8b3h00O1O1O100O1O100O100OO2N1O1O2N1O2N1O101N1O2N1O2N1O1O2N1O2O0O1O2O0O2O0O2O0O101N2N2O0O2O1N101N2O0O2N2O1N101N2O0O2I7M2O2M3M3N2M3M3M3N3L3M3N3L3O1N3N1N2O2M2O1N3M2N2NXg]2"}, "label": "an orange glimmering in light making it look darker and lighter in spots"}]}
{"id": 59483, "image": "COCO_train2014_000000059483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange with dried stem on tip\"."}], "task": "refering", "answer_template": "The \"orange with dried stem on tip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 338, 339, 340, 341], "bbox": [0.552015625, 0.5082284382284382, 0.9794375000000001, 0.9882284382284382], "iscrowd": 0, "area": 39395.30545, "rle": {"size": [429, 640], "counts": "bVd44W==C=C=C;E3M3M2N3M3M2N3M3M3M2N3M3M2N3M3N1O2N2M2O2N2M2O2N2N1N3N1O1O2N1N2O1O1O2N1O1N2O2N1O1O1N3N1O1O1O2N1O1O101N1O1O101N1O1O101N1O1O101N1O1O100O2N100O101N100O101N1O100O2O0O100O2O0O101N2O0O2N101N101N101O1O001O001O001O001O1O001N101O001O001O00001O0000001O01O000010O01O00010O000010O0001O0O2O0O101O0O101N10001N100O101N10000O101N10000O2O0O10001N100O10001N100O101N100O101N100O1O2O0O100O1O2O0O100O2N100O101N1O100O101N100O1O2O0O100O2N100O101N2N2O1N2N2N1O2N2N2N2N2O1N1O2N2N2N2N2N2N2O0O2N2N2N2N2N2N1O2O1N2N2N2N2N2N1O2N2M3N2M3N2M2O2M3N2M3M3N2M3N1N3N5J7J7H7I7J6I]T5"}, "label": "orange with dried stem on tip"}]}
{"id": 59483, "image": "COCO_train2014_000000059483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lowest orange in the bowl\"."}], "task": "refering", "answer_template": "The \"the lowest orange in the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 338, 339, 340, 341], "bbox": [0.552015625, 0.5082284382284382, 0.9794375000000001, 0.9882284382284382], "iscrowd": 0, "area": 39395.30545, "rle": {"size": [429, 640], "counts": "bVd44W==C=C=C;E3M3M2N3M3M2N3M3M3M2N3M3M2N3M3N1O2N2M2O2N2M2O2N2N1N3N1O1O2N1N2O1O1O2N1O1N2O2N1O1O1N3N1O1O1O2N1O1O101N1O1O101N1O1O101N1O1O101N1O1O100O2N100O101N100O101N1O100O2O0O100O2O0O101N2O0O2N101N101N101O1O001O001O001O001O1O001N101O001O001O00001O0000001O01O000010O01O00010O000010O0001O0O2O0O101O0O101N10001N100O101N10000O101N10000O2O0O10001N100O10001N100O101N100O101N100O1O2O0O100O1O2O0O100O2N100O101N1O100O101N100O1O2O0O100O2N100O101N2N2O1N2N2N1O2N2N2N2N2O1N1O2N2N2N2N2N2N2O0O2N2N2N2N2N2N1O2O1N2N2N2N2N2N1O2N2M3N2M3N2M2O2M3N2M3M3N2M3N1N3N5J7J7H7I7J6I]T5"}, "label": "the lowest orange in the bowl"}]}
{"id": 157793, "image": "COCO_train2014_000000157793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smiling boy with blue sweater behind the baby\"."}], "task": "refering", "answer_template": "The \"smiling boy with blue sweater behind the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 31, 32, 33, 34, 35, 36, 37, 38, 39, 46, 47, 48, 49, 50, 51, 52, 53, 54, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 90, 91, 92, 93, 94, 95, 98, 99, 100, 105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 168, 176, 180, 181, 182, 183, 190, 191, 192, 195, 196, 197, 198, 206, 207, 210, 211, 221, 222, 225], "bbox": [0.007816901408450705, 0.0026093749999999997, 0.8294131455399061, 0.682296875], "iscrowd": 0, "area": 55183.59170000002, "rle": {"size": [640, 426], "counts": "XU23Vc0o0POQ1POo0QOo0ROo0D;D<D=D;D<E<C<E;D=C<E;D7J1N2O0O2N101N2O0O2O1N1O2O0O2002M2O1O2M2O2M2O1O2M2O1O20O100O1O2O0O100O2N100O2N100O101N1O100O2O0O1O101N1O100O2O0O1O2OnLSFQLl9c3cF[L]9X3SGeLl8P3bGoL\\8e2SHYMm7[2bHbM]7R2RIlMm6h1cIVNg4_MbIo3U2`NY4cM_Il3\\2_NU4fM^Ik3`2\\NR4jM\\Ik3g2WNl3PNZIj3o2QNg3VNYIh3W3lM`3]NWIh3^3fM[3dNUIf3e3bMU3iNTIf3m3[Mo2PORIe3U4VMi2VOQId3[4RMc2\\OoHc3d4kL]2CmHc3k4eLX2IlHa3R5aLR20iH`3[5[Lk16iH_3a5VLf1<gH^3i5QL`1b0eH^3n=dLPB\\3P>eLnA\\3R>eLmAZ3T>hLiAY3W>hLgAY3Y>hLfAX3Z>jLcAV3^>kLaAU3_>lL_AU3a>U10O1000000O1000000O1000000O1000000O100000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O1O1N2O1N2O1N20000O100000000O1000000O100000000O100000000O100001O001O1O001O001O001O1OlKYAo2g>fLeAZ3Z>\\LQBc3V?100O0100O06KO001O1O001O001O001O1O001O001O1O010O001O001O1O001O001O1O2N3M2N3M2N3N1O2M2O1N3N1N3N1O2M2O2M2O2N1N3N0O2O0O10000O100O100O10000O2O0O10000O1ZOl]O_OTb05[^OGfa0Mj^OOVa0EZ_O6\\b0Mi`4NV_K4L5K5L3L5K4L5K5L3L5K5K4M4K4L5K5L1N3M2N3M22N101N2O0O2O0O2N2O0O2O0O2O1N1O2O1N101N1O2O1N101N101NRY]1"}, "label": "smiling boy with blue sweater behind the baby"}]}
{"id": 157793, "image": "COCO_train2014_000000157793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy smiling and holding the small baby\"."}], "task": "refering", "answer_template": "The \"a boy smiling and holding the small baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 31, 32, 33, 34, 35, 36, 37, 38, 39, 46, 47, 48, 49, 50, 51, 52, 53, 54, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 90, 91, 92, 93, 94, 95, 98, 99, 100, 105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 168, 176, 180, 181, 182, 183, 190, 191, 192, 195, 196, 197, 198, 206, 207, 210, 211, 221, 222, 225], "bbox": [0.007816901408450705, 0.0026093749999999997, 0.8294131455399061, 0.682296875], "iscrowd": 0, "area": 55183.59170000002, "rle": {"size": [640, 426], "counts": "XU23Vc0o0POQ1POo0QOo0ROo0D;D<D=D;D<E<C<E;D=C<E;D7J1N2O0O2N101N2O0O2O1N1O2O0O2002M2O1O2M2O2M2O1O2M2O1O20O100O1O2O0O100O2N100O2N100O101N1O100O2O0O1O101N1O100O2O0O1O2OnLSFQLl9c3cF[L]9X3SGeLl8P3bGoL\\8e2SHYMm7[2bHbM]7R2RIlMm6h1cIVNg4_MbIo3U2`NY4cM_Il3\\2_NU4fM^Ik3`2\\NR4jM\\Ik3g2WNl3PNZIj3o2QNg3VNYIh3W3lM`3]NWIh3^3fM[3dNUIf3e3bMU3iNTIf3m3[Mo2PORIe3U4VMi2VOQId3[4RMc2\\OoHc3d4kL]2CmHc3k4eLX2IlHa3R5aLR20iH`3[5[Lk16iH_3a5VLf1<gH^3i5QL`1b0eH^3n=dLPB\\3P>eLnA\\3R>eLmAZ3T>hLiAY3W>hLgAY3Y>hLfAX3Z>jLcAV3^>kLaAU3_>lL_AU3a>U10O1000000O1000000O1000000O1000000O100000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O1O1N2O1N2O1N20000O100000000O1000000O100000000O100000000O100001O001O1O001O001O001O1OlKYAo2g>fLeAZ3Z>\\LQBc3V?100O0100O06KO001O1O001O001O001O1O001O001O1O010O001O001O1O001O001O1O2N3M2N3M2N3N1O2M2O1N3N1N3N1O2M2O2M2O2N1N3N0O2O0O10000O100O100O10000O2O0O10000O1ZOl]O_OTb05[^OGfa0Mj^OOVa0EZ_O6\\b0Mi`4NV_K4L5K5L3L5K4L5K5L3L5K5K4M4K4L5K5L1N3M2N3M22N101N2O0O2O0O2N2O0O2O0O2O1N1O2O1N101N1O2O1N101N101NRY]1"}, "label": "a boy smiling and holding the small baby"}]}
{"id": 157793, "image": "COCO_train2014_000000157793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girl with the bear tied to her back\"."}], "task": "refering", "answer_template": "The \"the little girl with the bear tied to her back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 285, 287, 288, 289, 290, 291, 292, 293, 294, 295, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 330, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.0, 0.26009375, 0.724225352112676, 0.988796875], "iscrowd": 0, "area": 98925.91875, "rle": {"size": [640, 426], "counts": "\\=n0f4In81lE\\1S:^NfDQ3W;jLbCd4]<`1O1N2O1O1N2O1O1N2O1O1N2O1OoMTCPMk<n2]CmLa<S3fChLY<W3nCdLQ<Z3WDaLg;_3_D]L`;a3hD[LV;d3QEWLm:h3[ESLd:l3cEoK\\:o3lElKR:T4UFgKj9W4]FcKd9\\4o2O2O1N101N2O0O2O1N2O3L3N3M3L3N3L4M2M4M3M2O2O1NWBSMa:m2XE]Mf:b2SEhMk:V2oDSNe:W2TESNj:k1oD`No:_1jDkNS;T1gDUOX;j0aD@\\;?^DK`;4ZD4e;JUD`0h;@QDj0m;TOmCV1P<jNiC_1V8kKZKd2ZLk1T8hKaKV7Z4PIdKP7W4WIhKh6S4_IkKa6Q4eInKZ6m3nIQLQ6k3UJSLk5h3\\JWLc5d3dJZL\\5b3jJ]LU5^3SK_Lm4\\3ZKaLg4[3_KcLa4X3gKdLZ4Y3lKdLT4[3PLaLQ4_3SL]Lm3b3WLZLj3f3YLVLh3i3]LSLc3m3aLnK`3Q4eLkK[3T4jLgKW3Y4mLbKT3]4QM_Ko2a4VMYKk2f4ZMVKf2j4^MQKc2n4bMmJ_2S5eMiJ[2V5jMeJW2[5mMaJS2^5RN]Jo1c5UNXJl1g5W600O1O100O101N100O100O100O100O1O100O100O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O1O001O1O001O1O001O001O1O00100O001O1O001O1O001O0YFQJ]5P6bJPJ^5Q6aJPJ^5P6bJPJ^5R6`JoI_5S6_JmIa5U6]JlIb5V6\\JjId5W6[JjId5X6ZJhIf5Z6XJgIg5[6WJeIi5]6UJdIj5^6QJeIo5]6lIfIT6\\6fIiIY6Y6bIjI^6X6]IlIb6U6ZInIf6T6UIPJj6R6QIQJo6Q6lHSJS7o5jHRJV7P6gHRJX7P6eHQJ[7Q6aHRJ^7P6_HQJa7Q6\\HQJc7P6[HQJe7Q6WHRJh7P6UHQJk7Q6RHQJm7Q6PHPJP8S6kGoIU8S6hGoIW8T6eGmI[8U6bGmI]8V6^GlIb8V6[GlId8W6XGjIh8Y6TGiIk8Y6QGiIo8e8O001O001O001O00001O001O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N1O0lM]CkLd<S3bChL_<W3eCeL\\<Y3iCcLW<]3nC^LS<`3RD\\Lo;c3VDXLj;g3[DULf;j3^DRLc;l3cDoK];Q4gDkKZ;S4lDhKU;W4oDeKQ;Z4TE_KP;`4Y2N101A?VOi0E<C=D;D=C=^Oa0YOfgX2"}, "label": "the little girl with the bear tied to her back"}]}
{"id": 157793, "image": "COCO_train2014_000000157793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby with a teddy bear tied to their back\"."}], "task": "refering", "answer_template": "The \"baby with a teddy bear tied to their back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 285, 287, 288, 289, 290, 291, 292, 293, 294, 295, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 330, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.0, 0.26009375, 0.724225352112676, 0.988796875], "iscrowd": 0, "area": 98925.91875, "rle": {"size": [640, 426], "counts": "\\=n0f4In81lE\\1S:^NfDQ3W;jLbCd4]<`1O1N2O1O1N2O1O1N2O1O1N2O1OoMTCPMk<n2]CmLa<S3fChLY<W3nCdLQ<Z3WDaLg;_3_D]L`;a3hD[LV;d3QEWLm:h3[ESLd:l3cEoK\\:o3lElKR:T4UFgKj9W4]FcKd9\\4o2O2O1N101N2O0O2O1N2O3L3N3M3L3N3L4M2M4M3M2O2O1NWBSMa:m2XE]Mf:b2SEhMk:V2oDSNe:W2TESNj:k1oD`No:_1jDkNS;T1gDUOX;j0aD@\\;?^DK`;4ZD4e;JUD`0h;@QDj0m;TOmCV1P<jNiC_1V8kKZKd2ZLk1T8hKaKV7Z4PIdKP7W4WIhKh6S4_IkKa6Q4eInKZ6m3nIQLQ6k3UJSLk5h3\\JWLc5d3dJZL\\5b3jJ]LU5^3SK_Lm4\\3ZKaLg4[3_KcLa4X3gKdLZ4Y3lKdLT4[3PLaLQ4_3SL]Lm3b3WLZLj3f3YLVLh3i3]LSLc3m3aLnK`3Q4eLkK[3T4jLgKW3Y4mLbKT3]4QM_Ko2a4VMYKk2f4ZMVKf2j4^MQKc2n4bMmJ_2S5eMiJ[2V5jMeJW2[5mMaJS2^5RN]Jo1c5UNXJl1g5W600O1O100O101N100O100O100O100O1O100O100O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O1O001O1O001O1O001O001O1O00100O001O1O001O1O001O0YFQJ]5P6bJPJ^5Q6aJPJ^5P6bJPJ^5R6`JoI_5S6_JmIa5U6]JlIb5V6\\JjId5W6[JjId5X6ZJhIf5Z6XJgIg5[6WJeIi5]6UJdIj5^6QJeIo5]6lIfIT6\\6fIiIY6Y6bIjI^6X6]IlIb6U6ZInIf6T6UIPJj6R6QIQJo6Q6lHSJS7o5jHRJV7P6gHRJX7P6eHQJ[7Q6aHRJ^7P6_HQJa7Q6\\HQJc7P6[HQJe7Q6WHRJh7P6UHQJk7Q6RHQJm7Q6PHPJP8S6kGoIU8S6hGoIW8T6eGmI[8U6bGmI]8V6^GlIb8V6[GlId8W6XGjIh8Y6TGiIk8Y6QGiIo8e8O001O001O001O00001O001O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N1O0lM]CkLd<S3bChL_<W3eCeL\\<Y3iCcLW<]3nC^LS<`3RD\\Lo;c3VDXLj;g3[DULf;j3^DRLc;l3cDoK];Q4gDkKZ;S4lDhKU;W4oDeKQ;Z4TE_KP;`4Y2N101A?VOi0E<C=D;D=C=^Oa0YOfgX2"}, "label": "baby with a teddy bear tied to their back"}]}
{"id": 338025, "image": "COCO_train2014_000000338025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde fluffy dog looking out the window of a passenger car with a small blue and pink bow on top of its head\"."}], "task": "refering", "answer_template": "The \"a blonde fluffy dog looking out the window of a passenger car with a small blue and pink bow on top of its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309], "bbox": [0.348234375, 0.39975, 0.7186718750000001, 0.7823333333333333], "iscrowd": 0, "area": 30006.749300000003, "rle": {"size": [480, 640], "counts": "dlX34h>3M4L4K5L4L4M3N2M3N2N1N3N2N2M3N2M3N2N2M3N1O2M3N2N2M3N2M3N2K5G9G8H9H8VObLiEg3U:c0O1N2O1O1N2O0O2O1N2O1N2O1O1N2O1N101N2O1O1N2O1O100000O1000000000000000000O10000000000000000O100000000000000O10000000000000001N100000001O00000O2O000000001O00000O2O000000001O000O101O000000001O000O101O0000001O00000O100000001O00000O1000001O0000000O10001O000000000O101O1O1O001O1O1N2O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1N2O1O1O001O1O1O1N4M4L4L4L4L4L4L4K5L4L4L4L4L4L4K4M2N2N3M2N2N3M2M3N2N3M2N2N2N3M2M3N3M2N2N2N3M2N2MVQd2"}, "label": "a blonde fluffy dog looking out the window of a passenger car with a small blue and pink bow on top of its head"}]}
{"id": 338025, "image": "COCO_train2014_000000338025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white dog inside the car\"."}], "task": "refering", "answer_template": "The \"white dog inside the car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309], "bbox": [0.348234375, 0.39975, 0.7186718750000001, 0.7823333333333333], "iscrowd": 0, "area": 30006.749300000003, "rle": {"size": [480, 640], "counts": "dlX34h>3M4L4K5L4L4M3N2M3N2N1N3N2N2M3N2M3N2N2M3N1O2M3N2N2M3N2M3N2K5G9G8H9H8VObLiEg3U:c0O1N2O1O1N2O0O2O1N2O1N2O1O1N2O1N101N2O1O1N2O1O100000O1000000000000000000O10000000000000000O100000000000000O10000000000000001N100000001O00000O2O000000001O00000O2O000000001O000O101O000000001O000O101O0000001O00000O100000001O00000O1000001O0000000O10001O000000000O101O1O1O001O1O1N2O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1N2O1O1O001O1O1O1N4M4L4L4L4L4L4L4K5L4L4L4L4L4L4K4M2N2N3M2N2N3M2M3N2N3M2N2N2N3M2M3N3M2N2N2N3M2N2MVQd2"}, "label": "white dog inside the car"}]}
{"id": 338025, "image": "COCO_train2014_000000338025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beautiful black puppy\"."}], "task": "refering", "answer_template": "The \"beautiful black puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 237], "bbox": [0.15678125, 0.10685416666666667, 0.654125, 0.6277708333333333], "iscrowd": 0, "area": 40420.212449999985, "rle": {"size": [480, 640], "counts": "_o^11k>7I8H7H8I8H7I8K4M3L5YC`Nf;^2N3M2O1N3N1N2N2O2M2O1N3M2O1N2O2M2N2O2M2O1N2O2M2N2O2M2O1N3M2O1N2O2M2N2O2M2O1N2O2M2N2O2M2O1N2N3N1N2O2O0O1O101N1O101N1O1O101N1O101N1O1O101N1O101N1O101N1O1O101N1O101N1O100O2N1O101N1O100O2N1O101N1O100OO1O2M2O1O1O2N1N2O2N1O1N3N1O1O1N3N1O1O2N1N2O1O2N1N2O2N1O1O2M2O1O1O2M2O1O2N1N2O2N1O1O1N3N1O1O2M2O1O1O2M2O1N3M2M3N3M2M3N2N3M2M3N3M2M3N3M2M3N2N3L3N201O00001O0000001O000010O000001O00001O00001O0001O01O00001O00001O0000001O01O01O0000001O00001O000010O000001O00001O00001O0001O01O00001O0000001O00001O01O01O000N2O2N1O1O2N1O1O2M2O1O1O2N1O1O2N1N2O1O2N1O1O2N1O1N3N1O1O1O2N1O1O2M2O1O2N1O1O1O2N1N2O2N1O1O1O2N1O1O2M2O1O2N1O1O1O2N1N2O2NQnW3"}, "label": "beautiful black puppy"}]}
{"id": 510061, "image": "COCO_train2014_000000510061.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in thick yellow pants snowboarding on a mountain\"."}], "task": "refering", "answer_template": "The \"a man in thick yellow pants snowboarding on a mountain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 76, 77, 78, 99, 100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 171, 192], "bbox": [0.32815625000000004, 0.10859484777517564, 0.48109375, 0.5582669789227166], "iscrowd": 0, "area": 9033.43555, "rle": {"size": [427, 640], "counts": "feg23T=5L5jDHQ9>lF0d84YG:[8HbG=Y8FcG?Y8DhFTO;]1i8@jFZO5[1m8^OlF]O0Y1Q9\\OlFBKV1V9[OlFb1Q9_1L8I0N2N2M3N2M3N1N300O100O1O10O0100ON3M4L3hNkGXMX8f2X1O0O2O0O2O1N102NXOiL^GT3_8SMaGj2[8]MdG`2]8VNoFh1Q9X1100O10000O10O0101N101RNgFHZ9OQGNP9NWG1f8_NnFo0c0d0Z8XN[Gl0c0k0d8nNbGP1`8jNgGS1\\8fNkGW1X8bNnG\\1Q:000O10O11N10001N100010O2O00100O0M4M3L3N2M2O2M3N2M3N2M2O2N3L8IiWZ4"}, "label": "a man in thick yellow pants snowboarding on a mountain"}]}
{"id": 510061, "image": "COCO_train2014_000000510061.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboarder grinding down a rail\"."}], "task": "refering", "answer_template": "The \"a snowboarder grinding down a rail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 76, 77, 78, 99, 100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 171, 192], "bbox": [0.32815625000000004, 0.10859484777517564, 0.48109375, 0.5582669789227166], "iscrowd": 0, "area": 9033.43555, "rle": {"size": [427, 640], "counts": "feg23T=5L5jDHQ9>lF0d84YG:[8HbG=Y8FcG?Y8DhFTO;]1i8@jFZO5[1m8^OlF]O0Y1Q9\\OlFBKV1V9[OlFb1Q9_1L8I0N2N2M3N2M3N1N300O100O1O10O0100ON3M4L3hNkGXMX8f2X1O0O2O0O2O1N102NXOiL^GT3_8SMaGj2[8]MdG`2]8VNoFh1Q9X1100O10000O10O0101N101RNgFHZ9OQGNP9NWG1f8_NnFo0c0d0Z8XN[Gl0c0k0d8nNbGP1`8jNgGS1\\8fNkGW1X8bNnG\\1Q:000O10O11N10001N100010O2O00100O0M4M3L3N2M2O2M3N2M3N2M2O2N3L8IiWZ4"}, "label": "a snowboarder grinding down a rail"}]}
{"id": 18542, "image": "COCO_train2014_000000018542.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toddler brushing teeth in sink\"."}], "task": "refering", "answer_template": "The \"toddler brushing teeth in sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 421, 422, 423, 424, 425, 426, 427, 428, 429, 445, 446, 447, 448, 449, 450], "bbox": [0.2940625, 0.37304, 0.722, 0.887648], "iscrowd": 0, "area": 46808.69464999999, "rle": {"size": [625, 640], "counts": "VWc31_c03M4L3M4L3M3M1O2N2N2N1O2O1O100O010O100O00100O10O0100O10O0100O1O010O100O010O100O10O01O1N2N1N3N2M3N1N3N2M3N2M2O2M3N2oNP101O001O00000O101O0000001N1000001O00000O101O0000001N1000001O00000O101O000O1O2N1O1O1O2N1O1O1O001O1O1O1O1O1O1O1O01000O10000O100O01000O10000O10000O0hLRLoFn3b8bL\\G_3b8dL]G\\3a8hL]GX3b8jL\\GW3b8mL\\GS3b8PM]GP3a8SM]Gm2c8VM[Gj2c8YM[Gh2c8[M\\Ge2b8_M\\Ga2c8aM[G`2c8dMZG]2d8fMYG\\2e8gMYGZ2f8iMVGX2i8kMUGV2i8nMSGT2k8oMSGR2l8PNQGR2m8RNPGo1n8TNPGm1P9TNmFn1R9UNkFl1T9VNiFk1W9XNfFi1Y9YNdFi1[9YNcFh1\\9[N`Fg1_9[N_Ff1`9]N\\Fe1c9]N[Fd1d9^NZFb1g9`NUFb1j9`NTFa1k9S4O1O1O100000O10000000000000O10O100000000000000O10O100000000000O10000O100O010O10000O100O100O100O10O10O100O3N1N2O2N1N3N1N2O2M2O1O2M2O2M2O1UJQEQ3R;lLPES3Q;kLQET3P;jLRET3Q;iLQEV3P;hLREV3Q;gLQEX3P;fLREX3P;fLREY3P;dLREZ3P;dLRE[3P;bLRE]3o:`LTE^3n:`LTE_3m:_LUE_3m:_LUE`3l:^LVE`3l:^LVEa3j:^LXEa3i:]LYEa3i:]LYEb3h:\\L[Ea3g:]L\\Ea3d:^L_E_3c:_L`E_3a:_LcEQ3k:mLXEY2`;fMcD_1W<_NlCT1`<jNcCm0e<QO^Cg0i<WOZCa0m<]OVC;R=BQC6W=GlB0]=MfBKb=2Q3L4MV[\\3"}, "label": "toddler brushing teeth in sink"}]}
{"id": 18542, "image": "COCO_train2014_000000018542.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby sitting in a bathroom sink with a toothbrush\"."}], "task": "refering", "answer_template": "The \"a baby sitting in a bathroom sink with a toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 421, 422, 423, 424, 425, 426, 427, 428, 429, 445, 446, 447, 448, 449, 450], "bbox": [0.2940625, 0.37304, 0.722, 0.887648], "iscrowd": 0, "area": 46808.69464999999, "rle": {"size": [625, 640], "counts": "VWc31_c03M4L3M4L3M3M1O2N2N2N1O2O1O100O010O100O00100O10O0100O10O0100O1O010O100O010O100O10O01O1N2N1N3N2M3N1N3N2M3N2M2O2M3N2oNP101O001O00000O101O0000001N1000001O00000O101O0000001N1000001O00000O101O000O1O2N1O1O1O2N1O1O1O001O1O1O1O1O1O1O1O01000O10000O100O01000O10000O10000O0hLRLoFn3b8bL\\G_3b8dL]G\\3a8hL]GX3b8jL\\GW3b8mL\\GS3b8PM]GP3a8SM]Gm2c8VM[Gj2c8YM[Gh2c8[M\\Ge2b8_M\\Ga2c8aM[G`2c8dMZG]2d8fMYG\\2e8gMYGZ2f8iMVGX2i8kMUGV2i8nMSGT2k8oMSGR2l8PNQGR2m8RNPGo1n8TNPGm1P9TNmFn1R9UNkFl1T9VNiFk1W9XNfFi1Y9YNdFi1[9YNcFh1\\9[N`Fg1_9[N_Ff1`9]N\\Fe1c9]N[Fd1d9^NZFb1g9`NUFb1j9`NTFa1k9S4O1O1O100000O10000000000000O10O100000000000000O10O100000000000O10000O100O010O10000O100O100O100O10O10O100O3N1N2O2N1N3N1N2O2M2O1O2M2O2M2O1UJQEQ3R;lLPES3Q;kLQET3P;jLRET3Q;iLQEV3P;hLREV3Q;gLQEX3P;fLREX3P;fLREY3P;dLREZ3P;dLRE[3P;bLRE]3o:`LTE^3n:`LTE_3m:_LUE_3m:_LUE`3l:^LVE`3l:^LVEa3j:^LXEa3i:]LYEa3i:]LYEb3h:\\L[Ea3g:]L\\Ea3d:^L_E_3c:_L`E_3a:_LcEQ3k:mLXEY2`;fMcD_1W<_NlCT1`<jNcCm0e<QO^Cg0i<WOZCa0m<]OVC;R=BQC6W=GlB0]=MfBKb=2Q3L4MV[\\3"}, "label": "a baby sitting in a bathroom sink with a toothbrush"}]}
{"id": 18542, "image": "COCO_train2014_000000018542.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with blue color t - shirt and black color trousers\"."}], "task": "refering", "answer_template": "The \"man with blue color t - shirt and black color trousers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 76, 77, 78, 79, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216], "bbox": [0.224859375, 0.061408000000000004, 0.48276562500000003, 0.426256], "iscrowd": 0, "area": 21620.33195, "rle": {"size": [625, 640], "counts": "mRh29Sc0c0]O=C9G8H9G8H9G8H8H9G7ZM^LVEe3g:^LVEe3g:_LXEa3e:bL[E^3b:eL^E[3_:hLaEX3\\:kLdEU3Y:nLgER3V:RMiEn2T:UMlEk2S:VMmEj2S:VMmEj2T:UMlEk2T:UMlEk2T:UMlEk2T:VMkEj2U:VMkEi2V:WMjEi2V:WMjEi2V:WMjEi2V:WMjEi2V:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEf2Y:ZMgEf2Y:ZMgEf2Z:YMfEg2Z:YMfEg2Z:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMRF]2n9cM[FT2e9lMaFn1_9SN_Fn1a9RN_Fn1a9SN^Fm1b9TN]Fk1d9UN\\Fk1d9VN[Fj1e9WNZFi1f9YOWEh0i:XOUEj0k:VOSEl0m:TOcD\\1];dNcD\\1^;cNbD]1^;cNbD]1^;cNbD]1^;bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];aNdD_1\\;aNdD_1\\;aNdD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];cNbD]1^;cNbD]1^;cNbD]1^;cNbD]1^;dNaD\\1_;dNaD\\1_;dNaD\\1`;cN`D]1`;dN_D\\1a;dN_D\\1a;dN_D[1b;eN^D[1b;eN^D[1b;eN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;bN_D^1a;aN`D^1a;aN`D_1`;_NiDW1[;gN]E9l:FlEZO]:e0SFbNW:]1o3O1O1O1O1O1O100000000O1000000000000001O0001O00000001O000005K:F6J5K5H8CcnY6"}, "label": "man with blue color t - shirt and black color trousers"}]}
{"id": 18542, "image": "COCO_train2014_000000018542.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 76, 77, 78, 79, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216], "bbox": [0.224859375, 0.061408000000000004, 0.48276562500000003, 0.426256], "iscrowd": 0, "area": 21620.33195, "rle": {"size": [625, 640], "counts": "mRh29Sc0c0]O=C9G8H9G8H9G8H8H9G7ZM^LVEe3g:^LVEe3g:_LXEa3e:bL[E^3b:eL^E[3_:hLaEX3\\:kLdEU3Y:nLgER3V:RMiEn2T:UMlEk2S:VMmEj2S:VMmEj2T:UMlEk2T:UMlEk2T:UMlEk2T:VMkEj2U:VMkEi2V:WMjEi2V:WMjEi2V:WMjEi2V:WMjEi2V:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:XMiEh2W:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEg2X:YMhEf2Y:ZMgEf2Y:ZMgEf2Z:YMfEg2Z:YMfEg2Z:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMgEh2Y:XMRF]2n9cM[FT2e9lMaFn1_9SN_Fn1a9RN_Fn1a9SN^Fm1b9TN]Fk1d9UN\\Fk1d9VN[Fj1e9WNZFi1f9YOWEh0i:XOUEj0k:VOSEl0m:TOcD\\1];dNcD\\1^;cNbD]1^;cNbD]1^;cNbD]1^;bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];aNdD_1\\;aNdD_1\\;aNdD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];bNcD^1];cNbD]1^;cNbD]1^;cNbD]1^;cNbD]1^;dNaD\\1_;dNaD\\1_;dNaD\\1`;cN`D]1`;dN_D\\1a;dN_D\\1a;dN_D[1b;eN^D[1b;eN^D[1b;eN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;dN]D\\1c;bN_D^1a;aN`D^1a;aN`D_1`;_NiDW1[;gN]E9l:FlEZO]:e0SFbNW:]1o3O1O1O1O1O1O100000000O1000000000000001O0001O00000001O000005K:F6J5K5H8CcnY6"}, "label": "a man in a blue shirt"}]}
{"id": 18542, "image": "COCO_train2014_000000018542.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in green in the mirror\"."}], "task": "refering", "answer_template": "The \"a woman in green in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 80, 81, 82, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 174, 175, 194, 195, 217], "bbox": [0.45918749999999997, 0.065264, 0.631125, 0.42913599999999996], "iscrowd": 0, "area": 10556.964300000001, "rle": {"size": [625, 640], "counts": "cdc5o1^a04K5K5L4K4M4K5K5L4K5L3eLjLXFX3e9SMRFn2Q:WMjEg2X:^McE_2`:gMZEd1[;oNRDn0Q<QOPDl0S<TOmCi0V<VOkCg0X<YOhCd0[<[OfCb0]<^OcC`0`<_O`C?b<@_C>c<B]C<e<C\\C;f<EZC9h<FYC8i<HWC6k<JUC3n<LSC2o<NQC1P=OPC1P=OPC0Q=0oB0Q=OPC0Q=0oBOR=1nBOS=0mBOT=1lBOT=1lBNU=2kBNU=2kBMV=3jBLW=3jBMV=3jBLW=4iBLW=4iBMV=3jBMV=3jBMV=3jBMV=3jBNU=2kBNU=1lBOU=0kB0U=0kB1T=OlB1T=OlB1T=OlB1T=0kB1T=0kB0U=0kB0U=1jBOV=2iBNW=3hBNW=3hBMX=3hBMX=4gBLY=5fBLY=5fBK[=5dBL[=4eBMZ=4eBMZ=4eBL[=5dBL[=4eBMZ=4fBLY=5fBKZ=6eBKZ=7dBJ[=9bBH]=;`BE`==^BDa=?\\BBc=>]BCb==^BDb=<]BDc=<]BEb=;^BFa=;^BFa=lNdBBKb1a=iNPDX1o;dNUD]1j;`NYDa1f;[N^De1b;XNaDi1^;SNfDn1^>10O01O1O1O1O001?Af0ZOf0YOjj_4"}, "label": "a woman in green in the mirror"}]}
{"id": 18542, "image": "COCO_train2014_000000018542.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult in a green shirt\"."}], "task": "refering", "answer_template": "The \"an adult in a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 80, 81, 82, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 174, 175, 194, 195, 217], "bbox": [0.45918749999999997, 0.065264, 0.631125, 0.42913599999999996], "iscrowd": 0, "area": 10556.964300000001, "rle": {"size": [625, 640], "counts": "cdc5o1^a04K5K5L4K4M4K5K5L4K5L3eLjLXFX3e9SMRFn2Q:WMjEg2X:^McE_2`:gMZEd1[;oNRDn0Q<QOPDl0S<TOmCi0V<VOkCg0X<YOhCd0[<[OfCb0]<^OcC`0`<_O`C?b<@_C>c<B]C<e<C\\C;f<EZC9h<FYC8i<HWC6k<JUC3n<LSC2o<NQC1P=OPC1P=OPC0Q=0oB0Q=OPC0Q=0oBOR=1nBOS=0mBOT=1lBOT=1lBNU=2kBNU=2kBMV=3jBLW=3jBMV=3jBLW=4iBLW=4iBMV=3jBMV=3jBMV=3jBMV=3jBNU=2kBNU=1lBOU=0kB0U=0kB1T=OlB1T=OlB1T=OlB1T=0kB1T=0kB0U=0kB0U=1jBOV=2iBNW=3hBNW=3hBMX=3hBMX=4gBLY=5fBLY=5fBK[=5dBL[=4eBMZ=4eBMZ=4eBL[=5dBL[=4eBMZ=4fBLY=5fBKZ=6eBKZ=7dBJ[=9bBH]=;`BE`==^BDa=?\\BBc=>]BCb==^BDb=<]BDc=<]BEb=;^BFa=;^BFa=lNdBBKb1a=iNPDX1o;dNUD]1j;`NYDa1f;[N^De1b;XNaDi1^;SNfDn1^>10O01O1O1O1O001?Af0ZOf0YOjj_4"}, "label": "an adult in a green shirt"}]}
{"id": 239745, "image": "COCO_train2014_000000239745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of skis that no one is wearing\"."}], "task": "refering", "answer_template": "The \"a pair of skis that no one is wearing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 97, 98, 99, 106, 107, 108, 109], "bbox": [0.6092086330935252, 0.48064, 1.0, 0.6140399999999999], "iscrowd": 0, "area": 1475.3868999999995, "rle": {"size": [500, 278], "counts": "bgc21c?000dA0S=1lB0T=0jB2U=OkB1U=OkB1U=OkB2S=OmB1S=OnB0Q=2nBNR=2nBNR=2nBNQ=4nBMQ=3oBMP=5oBKR=4oBKQ=5oBKR=5mBLR=4nBLR=5mBKT=4lBLT=4mBKT=5kBKU=5kBLT=5kBKV=4jBLV=4jBLW=4hBLX=4iBLV=5iBKX=4hBLX=4hBLX=5gBKZ=4fBLZ=5fBKZ=4fBLZ=4fBLZ=5eBK\\=4dBL\\=5cBL]=3cBM]=4cBK]=5cBK^=4bBL^=5aBK`=4`BM_=4`BL`=4`BLa=3`BL`=5_BKb=4^BMa=4^BLb=4^BLc=3]BMc=4]BKd=4\\BLd=5[BLd=4\\BLe=3[BMe=4ZBLf=4ZBLg=4YBLf=4ZBLg=3YBMg=4XBLh=O]B1d>OS^1NkaN4O2ZAMi=3VBNj=3UBMl=2TBNl=3SBNm=1SBOm=1TBNl=3SBMn=2RBNn=3QBNo=1QBOo=2QBMP>2PBNP>2PBOo=2PBNQ>1oAOQ>2oAMR>2nANR>3mANS>1mAOS>1mAOS>2mAMT>2lANT>3kANU>1kAOU>2jANV>2kAMV>2jAOU>2jANW>1iAOW>2hANY>1hANX>3gANX>2hANd5"}, "label": "a pair of skis that no one is wearing"}]}
{"id": 239745, "image": "COCO_train2014_000000239745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the corner of some skis who ' s rider is out of the picture\"."}], "task": "refering", "answer_template": "The \"the corner of some skis who ' s rider is out of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 97, 98, 99, 106, 107, 108, 109], "bbox": [0.6092086330935252, 0.48064, 1.0, 0.6140399999999999], "iscrowd": 0, "area": 1475.3868999999995, "rle": {"size": [500, 278], "counts": "bgc21c?000dA0S=1lB0T=0jB2U=OkB1U=OkB1U=OkB2S=OmB1S=OnB0Q=2nBNR=2nBNR=2nBNQ=4nBMQ=3oBMP=5oBKR=4oBKQ=5oBKR=5mBLR=4nBLR=5mBKT=4lBLT=4mBKT=5kBKU=5kBLT=5kBKV=4jBLV=4jBLW=4hBLX=4iBLV=5iBKX=4hBLX=4hBLX=5gBKZ=4fBLZ=5fBKZ=4fBLZ=4fBLZ=5eBK\\=4dBL\\=5cBL]=3cBM]=4cBK]=5cBK^=4bBL^=5aBK`=4`BM_=4`BL`=4`BLa=3`BL`=5_BKb=4^BMa=4^BLb=4^BLc=3]BMc=4]BKd=4\\BLd=5[BLd=4\\BLe=3[BMe=4ZBLf=4ZBLg=4YBLf=4ZBLg=3YBMg=4XBLh=O]B1d>OS^1NkaN4O2ZAMi=3VBNj=3UBMl=2TBNl=3SBNm=1SBOm=1TBNl=3SBMn=2RBNn=3QBNo=1QBOo=2QBMP>2PBNP>2PBOo=2PBNQ>1oAOQ>2oAMR>2nANR>3mANS>1mAOS>1mAOS>2mAMT>2lANT>3kANU>1kAOU>2jANV>2kAMV>2jAOU>2jANW>1iAOW>2hANY>1hANX>3gANX>2hANd5"}, "label": "the corner of some skis who ' s rider is out of the picture"}]}
{"id": 264324, "image": "COCO_train2014_000000264324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult sitting outside of the tub\"."}], "task": "refering", "answer_template": "The \"the adult sitting outside of the tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 229, 252], "bbox": [0.858484375, 0.255625, 1.0, 0.6645625000000001], "iscrowd": 0, "area": 8323.048250000003, "rle": {"size": [480, 640], "counts": "^bQ81m>3CMjA5R>OkA4Q>0lA2k=FVB<LOj=JXBf0f=:00O2N101N100O100O1O100O100O100O00100O100O2O0O1O100001O000000000001XC`No;`1kChNS<Y1kCkNS<U1lCmNR<T1mCmNR<S1mCnNS<S1lCnNS<S1lClNU<U1iCjNY<W1fCgN\\<Z1cCeN_<o1N4L4L4L4L4L4L4L4M3L4L3M4L4L3M3M3M3N2M3M3M3M3M3M3M3N1N3N2N1O2N2N1O2N2N1O3M2N2N4QL"}, "label": "the adult sitting outside of the tub"}]}
{"id": 264324, "image": "COCO_train2014_000000264324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person near a tub a kid in tub\"."}], "task": "refering", "answer_template": "The \"a person near a tub a kid in tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 229, 252], "bbox": [0.858484375, 0.255625, 1.0, 0.6645625000000001], "iscrowd": 0, "area": 8323.048250000003, "rle": {"size": [480, 640], "counts": "^bQ81m>3CMjA5R>OkA4Q>0lA2k=FVB<LOj=JXBf0f=:00O2N101N100O100O1O100O100O100O00100O100O2O0O1O100001O000000000001XC`No;`1kChNS<Y1kCkNS<U1lCmNR<T1mCmNR<S1mCnNS<S1lCnNS<S1lClNU<U1iCjNY<W1fCgN\\<Z1cCeN_<o1N4L4L4L4L4L4L4L4M3L4L3M4L4L3M3M3M3N2M3M3M3M3M3M3M3N1N3N2N1O2N2N1O2N2N1O3M2N2N4QL"}, "label": "a person near a tub a kid in tub"}]}
{"id": 264324, "image": "COCO_train2014_000000264324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby sitting in a turquoise tub\"."}], "task": "refering", "answer_template": "The \"baby sitting in a turquoise tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 290, 291, 306, 307, 308, 309, 310, 313, 329, 330, 331, 332, 333, 334, 356, 357], "bbox": [0.2841875, 0.2734375, 0.66928125, 0.9393541666666666], "iscrowd": 0, "area": 37031.105599999995, "rle": {"size": [480, 640], "counts": "]de2j0S>4L5K4M3L4L5J5K5K6J5K5K5K6L3N2M3N3L3N2M3M4M2M3N2M4M2M3N2M3N3L3aNhKRH[4l7jKoGX4^6cKlI;_OV4b6jKdI5ES4d6UL[IJNT4e6^LRI_O7U4d6iLiHTO`0V4e6`MYIc2c6_M[Id2c6^M[Id2b6_M\\Ic2b6_M\\Ic2a6`M^Ia2`6aM_I_2_6cM`I^2_6dM`I[2a6fM]I[2b6R3000O0100O10O10O100O10O10O1000O0100O10O10O10001N101O001O001N101O001O1O000O1000O100O1000O10000100O001O1O1O1O1O001O1O100O010O1O100O2O1N101N2O0O2hNYISJi6i5ZIWJf6f5]IYJe6b5_I^Jb6^5aIbJ_6\\5cIdJ^6Y5dIgJ]6V5eIjJ[6T5gIlJZ6R5gInJZ6o4hIQKX6m4jISK[6IjHP4m0WLe6ZOfHX4g0^Lc6WOnHT4`0fLa6TOQIT4`0iL_6POTIT4`0lLV7Q3mHQMR77gGU1Z1dNn66lGP1Y1kNk63QHn0U1oNi62VHk0R1TOh60ZHh0o0XOf61]Hd0m0\\Of6O`Ha0l0@c60bH>k0Cb60dH;k0E`60dH;m0F^60dH:o0F\\60dH;P1F[60cH:S1HX6OdH9U1HV6OdHTO_Ob0g1:W61aHSOF=d1=V63`HROK:_1`0V66^HPO08]1`0V69\\HnN37\\1a0V6;YHlN96Y1a0V6?UHjN?4V1b0W6a0SHhNb03V1b0V6c0QHhNf0OU1e0U6d0PHgNi0LT1g0U6g0lGfNm0IT1i0U6g0iGgN\\3`0l4i0hGfN]3`0m4i0eGgN_3?m4k0cGfN`3>o4k0aGhN`3;P5m0`GhN`3:R5n0^GhN_39T5o0]GhN_38V5o0[GjN^36X5Q1YGiN`34Y5R1WGjN`33Z5S1VGkN_31]5S1TGlN_30^5U1RGlN_3Na5U1PGmN`3La5W1PGmN^3Kd5V1oFoN]3Jf5V1mFQO\\3Hh5W1lFQO^3Eh5X1kFTO^3Ah5[1jFTO15o;f0PD9h;EYD;h;DXD<i;CXD;j;DVD<k;BWD=i;CWD=j;BWD=j;AXD>i;AWD>k;AVD>k;@WD?j;@VD`0k;_OVD`0j;_OXD`0i;^OXDc0i;ZOYDe0i;XOXDh0k;TOXDk0i;RO\\Dj0g;RO^Dj0d;SOaDj0n<L3M4M2M3M3N2M2N3NVBMh<0UC2i=4M3L4M30O100000O10O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O00nPS3"}, "label": "baby sitting in a turquoise tub"}]}
{"id": 264324, "image": "COCO_train2014_000000264324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby in a tub with toys taking a bath\"."}], "task": "refering", "answer_template": "The \"a baby in a tub with toys taking a bath\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 290, 291, 306, 307, 308, 309, 310, 313, 329, 330, 331, 332, 333, 334, 356, 357], "bbox": [0.2841875, 0.2734375, 0.66928125, 0.9393541666666666], "iscrowd": 0, "area": 37031.105599999995, "rle": {"size": [480, 640], "counts": "]de2j0S>4L5K4M3L4L5J5K5K6J5K5K5K6L3N2M3N3L3N2M3M4M2M3N2M4M2M3N2M3N3L3aNhKRH[4l7jKoGX4^6cKlI;_OV4b6jKdI5ES4d6UL[IJNT4e6^LRI_O7U4d6iLiHTO`0V4e6`MYIc2c6_M[Id2c6^M[Id2b6_M\\Ic2b6_M\\Ic2a6`M^Ia2`6aM_I_2_6cM`I^2_6dM`I[2a6fM]I[2b6R3000O0100O10O10O100O10O10O1000O0100O10O10O10001N101O001O001N101O001O1O000O1000O100O1000O10000100O001O1O1O1O1O001O1O100O010O1O100O2O1N101N2O0O2hNYISJi6i5ZIWJf6f5]IYJe6b5_I^Jb6^5aIbJ_6\\5cIdJ^6Y5dIgJ]6V5eIjJ[6T5gIlJZ6R5gInJZ6o4hIQKX6m4jISK[6IjHP4m0WLe6ZOfHX4g0^Lc6WOnHT4`0fLa6TOQIT4`0iL_6POTIT4`0lLV7Q3mHQMR77gGU1Z1dNn66lGP1Y1kNk63QHn0U1oNi62VHk0R1TOh60ZHh0o0XOf61]Hd0m0\\Of6O`Ha0l0@c60bH>k0Cb60dH;k0E`60dH;m0F^60dH:o0F\\60dH;P1F[60cH:S1HX6OdH9U1HV6OdHTO_Ob0g1:W61aHSOF=d1=V63`HROK:_1`0V66^HPO08]1`0V69\\HnN37\\1a0V6;YHlN96Y1a0V6?UHjN?4V1b0W6a0SHhNb03V1b0V6c0QHhNf0OU1e0U6d0PHgNi0LT1g0U6g0lGfNm0IT1i0U6g0iGgN\\3`0l4i0hGfN]3`0m4i0eGgN_3?m4k0cGfN`3>o4k0aGhN`3;P5m0`GhN`3:R5n0^GhN_39T5o0]GhN_38V5o0[GjN^36X5Q1YGiN`34Y5R1WGjN`33Z5S1VGkN_31]5S1TGlN_30^5U1RGlN_3Na5U1PGmN`3La5W1PGmN^3Kd5V1oFoN]3Jf5V1mFQO\\3Hh5W1lFQO^3Eh5X1kFTO^3Ah5[1jFTO15o;f0PD9h;EYD;h;DXD<i;CXD;j;DVD<k;BWD=i;CWD=j;BWD=j;AXD>i;AWD>k;AVD>k;@WD?j;@VD`0k;_OVD`0j;_OXD`0i;^OXDc0i;ZOYDe0i;XOXDh0k;TOXDk0i;RO\\Dj0g;RO^Dj0d;SOaDj0n<L3M4M2M3M3N2M2N3NVBMh<0UC2i=4M3L4M30O100000O10O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O00nPS3"}, "label": "a baby in a tub with toys taking a bath"}]}
{"id": 534663, "image": "COCO_train2014_000000534663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white panam airplane\"."}], "task": "refering", "answer_template": "The \"a white panam airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 94, 95, 117, 118, 119, 120, 141, 142, 143, 144, 145, 146, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 222], "bbox": [0.085140625, 0.2584295612009238, 0.773859375, 0.6404387990762125], "iscrowd": 0, "area": 26658.55765000001, "rle": {"size": [433, 640], "counts": "g[g03^=2N3M2N3M2N3M2N3M2N3M2N3M2YEnNk8U1RGlNm8V1QGkNn8X1oFjNo8X1oFiNP9Z1mFgNR9[1lFgNR9\\1kFfNS9\\1kFgNR9\\1kFfNT9[1jFfNU9]1hFcNY9^1eFbN\\9`1aFaN_9`1_F`Nb9b1[F^Nf9c1XF^Ng9e1VF[Nk9f1SFZNn9h1oEXNR:i1lEXNT:j1iEVNX:k1fEUNZ:a2100O1O1O10O01O1O100O1O10O01O1O100O1O00100O1O1O10O01O1O100O1O010O1O1O100O1O00100O1O1O10O01O1O100O1O010O1O1O100O001O100O1O1O10O01O1O100O1O00100O1O100O001O100O1O1O010O1O1O100O1O00100O1O01O0000001O0001O01O0000001O00000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O000000001O4M5J2N1O000000000000000000001O000000000000000000001O000000000000000000001O00000000000000CjD_NV;a1jD_NV;a1kD^NU;b1kD_NT;a1lD_NT;a1mD^NS;b1nD]NR;c1PE[NP;e1REYNn:g1TEWNl:i1UEVNk:j1WETNi:l1>0000000000000000000000000O10000000000O10000000000O10000000001N1000000000000O10000000000O10000000000O1000000000000001O0O1O1O100O1O1O1O10001O00000000000000000000001O00000000000000000000001O00000000000000000001O01O0000000000000000eDWNn:i1oDZNQ;f1lD]NT;c1jD`NV;_1gDdNY;j10001O00001O0000N2M3M4L3M3M3O100O100O100O101N2O1N101N2O1N2O001N2O1N101N2O1N2O0O2O1N2O0O2O1N2O1N101N2O1N101N2O1N2O0O2O1N2OhRm1"}, "label": "a white panam airplane"}]}
{"id": 534663, "image": "COCO_train2014_000000534663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plane coming in for a landing\"."}], "task": "refering", "answer_template": "The \"a white plane coming in for a landing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 94, 95, 117, 118, 119, 120, 141, 142, 143, 144, 145, 146, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 222], "bbox": [0.085140625, 0.2584295612009238, 0.773859375, 0.6404387990762125], "iscrowd": 0, "area": 26658.55765000001, "rle": {"size": [433, 640], "counts": "g[g03^=2N3M2N3M2N3M2N3M2N3M2N3M2YEnNk8U1RGlNm8V1QGkNn8X1oFjNo8X1oFiNP9Z1mFgNR9[1lFgNR9\\1kFfNS9\\1kFgNR9\\1kFfNT9[1jFfNU9]1hFcNY9^1eFbN\\9`1aFaN_9`1_F`Nb9b1[F^Nf9c1XF^Ng9e1VF[Nk9f1SFZNn9h1oEXNR:i1lEXNT:j1iEVNX:k1fEUNZ:a2100O1O1O10O01O1O100O1O10O01O1O100O1O00100O1O1O10O01O1O100O1O010O1O1O100O1O00100O1O1O10O01O1O100O1O010O1O1O100O001O100O1O1O10O01O1O100O1O00100O1O100O001O100O1O1O010O1O1O100O1O00100O1O01O0000001O0001O01O0000001O00000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O000000001O4M5J2N1O000000000000000000001O000000000000000000001O000000000000000000001O00000000000000CjD_NV;a1jD_NV;a1kD^NU;b1kD_NT;a1lD_NT;a1mD^NS;b1nD]NR;c1PE[NP;e1REYNn:g1TEWNl:i1UEVNk:j1WETNi:l1>0000000000000000000000000O10000000000O10000000000O10000000001N1000000000000O10000000000O10000000000O1000000000000001O0O1O1O100O1O1O1O10001O00000000000000000000001O00000000000000000000001O00000000000000000001O01O0000000000000000eDWNn:i1oDZNQ;f1lD]NT;c1jD`NV;_1gDdNY;j10001O00001O0000N2M3M4L3M3M3O100O100O100O101N2O1N101N2O1N2O001N2O1N101N2O1N2O0O2O1N2O0O2O1N2O1N101N2O1N101N2O1N2O0O2O1N2OhRm1"}, "label": "a white plane coming in for a landing"}]}
{"id": 157834, "image": "COCO_train2014_000000157834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white stuffed teddy bear beside the vase with a red ribbon\"."}], "task": "refering", "answer_template": "The \"a white stuffed teddy bear beside the vase with a red ribbon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 174, 175, 176, 177, 186, 187, 188, 189, 190, 199, 200, 201, 202, 212, 213, 214, 215, 225, 226, 227, 228], "bbox": [0.2899733333333333, 0.6905800000000001, 0.64872, 0.98878], "iscrowd": 0, "area": 11645.249549999997, "rle": {"size": [500, 375], "counts": "dTh11b?1O1O100O1O100O1O100O1O100dMNUE3f:6VEJh:9VEHi:9VEHi::VEFi:;VEFi:<VEEh:<REJm:8lDNT;2fD4Y;NbD6];K_D9a;HZD<e;EWD?i;ASDc0l;^OjCl0V<Y1O100O100O100O100O100O100O10000O10000O10000O10000O100001O001O001O001O000000_OhLfDX3X;nLdDR3Y;UMcDk2Z;i0N2N2O1N2N2N2O1O100000000000000000000000000000000000J6I8J5L4L4L4L5K4L4K5K5K5M4M2N2N2YOPChNR=W1QCeNQ=Z1SCaNo<_1c0O101L3F:H8H9A>00O2O0O3N_SP2"}, "label": "a white stuffed teddy bear beside the vase with a red ribbon"}]}
{"id": 157834, "image": "COCO_train2014_000000157834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear is placed next to a flower vase\"."}], "task": "refering", "answer_template": "The \"a teddy bear is placed next to a flower vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 174, 175, 176, 177, 186, 187, 188, 189, 190, 199, 200, 201, 202, 212, 213, 214, 215, 225, 226, 227, 228], "bbox": [0.2899733333333333, 0.6905800000000001, 0.64872, 0.98878], "iscrowd": 0, "area": 11645.249549999997, "rle": {"size": [500, 375], "counts": "dTh11b?1O1O100O1O100O1O100O1O100dMNUE3f:6VEJh:9VEHi:9VEHi::VEFi:;VEFi:<VEEh:<REJm:8lDNT;2fD4Y;NbD6];K_D9a;HZD<e;EWD?i;ASDc0l;^OjCl0V<Y1O100O100O100O100O100O100O10000O10000O10000O10000O100001O001O001O001O000000_OhLfDX3X;nLdDR3Y;UMcDk2Z;i0N2N2O1N2N2N2O1O100000000000000000000000000000000000J6I8J5L4L4L4L5K4L4K5K5K5M4M2N2N2YOPChNR=W1QCeNQ=Z1SCaNo<_1c0O101L3F:H8H9A>00O2O0O3N_SP2"}, "label": "a teddy bear is placed next to a flower vase"}]}
{"id": 157834, "image": "COCO_train2014_000000157834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vase with red ribbon\"."}], "task": "refering", "answer_template": "The \"vase with red ribbon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 151, 152, 164, 165, 177, 178, 190, 191, 203, 204, 215, 216, 217, 228, 229, 230], "bbox": [0.576, 0.58392, 0.7668, 0.98814], "iscrowd": 0, "area": 7659.834499999999, "rle": {"size": [500, 375], "counts": "dmY33`?6I6J7J5J2O1N2O1N2O1O1O001O2N1fNVOiCk0k;GmC;h;6QDKc;f0VD]O];c2E;J6J6J7POo0O100O100O2O0000000O2O0000000O2O0000O010O10O01_MdFjN]9h0SGUOo8=_G@c82mGKU8GYH6i7\\OgHa0[7QOSIl0o6fNaIW1_6\\NPJa1U:M4K4M4L1N2O1L4L3M4L4K5L3M4Ko`Z1"}, "label": "vase with red ribbon"}]}
{"id": 157834, "image": "COCO_train2014_000000157834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red ribbon tied around a glass vase with flowers\"."}], "task": "refering", "answer_template": "The \"red ribbon tied around a glass vase with flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 151, 152, 164, 165, 177, 178, 190, 191, 203, 204, 215, 216, 217, 228, 229, 230], "bbox": [0.576, 0.58392, 0.7668, 0.98814], "iscrowd": 0, "area": 7659.834499999999, "rle": {"size": [500, 375], "counts": "dmY33`?6I6J7J5J2O1N2O1N2O1O1O001O2N1fNVOiCk0k;GmC;h;6QDKc;f0VD]O];c2E;J6J6J7POo0O100O100O2O0000000O2O0000000O2O0000O010O10O01_MdFjN]9h0SGUOo8=_G@c82mGKU8GYH6i7\\OgHa0[7QOSIl0o6fNaIW1_6\\NPJa1U:M4K4M4L1N2O1L4L3M4L4K5L3M4Ko`Z1"}, "label": "red ribbon tied around a glass vase with flowers"}]}
{"id": 157834, "image": "COCO_train2014_000000157834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear to the left of purple ribbon\"."}], "task": "refering", "answer_template": "The \"bear to the left of purple ribbon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 157, 158, 159, 160, 169, 170, 171, 172, 173, 183, 184, 185, 196, 197, 198, 208, 209, 210, 211, 221, 222, 223, 224], "bbox": [0.053919999999999996, 0.6606799999999999, 0.3296, 0.98876], "iscrowd": 0, "area": 11713.3998, "rle": {"size": [500, 375], "counts": "[W:4^??UL^OhGd0W8I\\G8c8LXG6h8KUG7j8KTG6k8LSG5l8MRG4n8NoF3P9OnF2Q90mF1R91mFOS92aF9^9I^F:a9G\\F<d9CZF`0f9@WFc0i9]OTFf0l9ZOQFi0o9WOnEl0R:TOkEo0U:QOhER1X:nNeEU1[:kNcEW1]:hNaE[1_:eN^E^1b:bN[E`1f:`NWEc1i:]NTEf1l:ZNSEf1n:ZNQEg1o:YNPEg1Q;XNoDi1Q;WNnDi1S;WNlDj1T;VNkDj1V;VNiDk1W;UNhDk1Y;UNfDl1Z;TNeDl1\\;W1O1O1O1O1O1O0100000O01O1O1O1O1O1O1N2OH8B>O1O1O1N2O001O1O1O1O100O11N100000001O00000O101O000YObEXL^:e2^EjM`0VOR:m2fEgMT;U2TEeMm:X2[EbMf:[2bE_M_:^2^1QOSCROP=n0TCkNo<T1UCfNn<Z1VC_Nm<a1c0O100O1000O0_Wj3"}, "label": "bear to the left of purple ribbon"}]}
{"id": 157834, "image": "COCO_train2014_000000157834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white teddy bear tied to a vase with purple lace\"."}], "task": "refering", "answer_template": "The \"white teddy bear tied to a vase with purple lace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 157, 158, 159, 160, 169, 170, 171, 172, 173, 183, 184, 185, 196, 197, 198, 208, 209, 210, 211, 221, 222, 223, 224], "bbox": [0.053919999999999996, 0.6606799999999999, 0.3296, 0.98876], "iscrowd": 0, "area": 11713.3998, "rle": {"size": [500, 375], "counts": "[W:4^??UL^OhGd0W8I\\G8c8LXG6h8KUG7j8KTG6k8LSG5l8MRG4n8NoF3P9OnF2Q90mF1R91mFOS92aF9^9I^F:a9G\\F<d9CZF`0f9@WFc0i9]OTFf0l9ZOQFi0o9WOnEl0R:TOkEo0U:QOhER1X:nNeEU1[:kNcEW1]:hNaE[1_:eN^E^1b:bN[E`1f:`NWEc1i:]NTEf1l:ZNSEf1n:ZNQEg1o:YNPEg1Q;XNoDi1Q;WNnDi1S;WNlDj1T;VNkDj1V;VNiDk1W;UNhDk1Y;UNfDl1Z;TNeDl1\\;W1O1O1O1O1O1O0100000O01O1O1O1O1O1O1N2OH8B>O1O1O1N2O001O1O1O1O100O11N100000001O00000O101O000YObEXL^:e2^EjM`0VOR:m2fEgMT;U2TEeMm:X2[EbMf:[2bE_M_:^2^1QOSCROP=n0TCkNo<T1UCfNn<Z1VC_Nm<a1c0O100O1000O0_Wj3"}, "label": "white teddy bear tied to a vase with purple lace"}]}
{"id": 354445, "image": "COCO_train2014_000000354445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of two leather suitcases\"."}], "task": "refering", "answer_template": "The \"the top of two leather suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258], "bbox": [0.009171875, 0.0022520107238605896, 0.7661718750000001, 0.9146112600536193], "iscrowd": 0, "area": 70265.61775, "rle": {"size": [373, 640], "counts": "g\\22c;1N2N2N2O0O2N2N2O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N101N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N3N1N3M2N3M2N2N2N2N2N1O2N2O1N2N2N1O2N2N2N2N2N1O2O1N2N2N2N2N1O2N2N2N2N2N101N2N2N2N2N1O0000000001O0000000000000001O000001O000000000000000000001O000000000001O00000001O00000000000000000000001O000001O00000000000001O000000000000000000000010O00000000000000000001O000000000000000001O0001O000000000000000000001O0000001O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1ORPg1"}, "label": "the top of two leather suitcases"}]}
{"id": 354445, "image": "COCO_train2014_000000354445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather suitcase with silver clasps\"."}], "task": "refering", "answer_template": "The \"a leather suitcase with silver clasps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258], "bbox": [0.009171875, 0.0022520107238605896, 0.7661718750000001, 0.9146112600536193], "iscrowd": 0, "area": 70265.61775, "rle": {"size": [373, 640], "counts": "g\\22c;1N2N2N2O0O2N2N2O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N101N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N3N1N3M2N3M2N2N2N2N2N1O2N2O1N2N2N1O2N2N2N2N2N1O2O1N2N2N2N2N1O2N2N2N2N2N101N2N2N2N2N1O0000000001O0000000000000001O000001O000000000000000000001O000000000001O00000001O00000000000000000000001O000001O00000000000001O000000000000000000000010O00000000000000000001O000000000000000001O0001O000000000000000000001O0000001O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1ORPg1"}, "label": "a leather suitcase with silver clasps"}]}
{"id": 354445, "image": "COCO_train2014_000000354445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lower brown suitcasre\"."}], "task": "refering", "answer_template": "The \"the lower brown suitcasre\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 136, 137, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.628125, 0.3440750670241287, 1.0, 1.0], "iscrowd": 0, "area": 29487.365199999986, "rle": {"size": [373, 640], "counts": "^ib41c;1N2O1O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O001O00000000000000000000001O000000000O100000000000001O0000000001O1O1O1O1OQL"}, "label": "the lower brown suitcasre"}]}
{"id": 354445, "image": "COCO_train2014_000000354445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown smooth brief case\"."}], "task": "refering", "answer_template": "The \"brown smooth brief case\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 136, 137, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.628125, 0.3440750670241287, 1.0, 1.0], "iscrowd": 0, "area": 29487.365199999986, "rle": {"size": [373, 640], "counts": "^ib41c;1N2O1O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O001O00000000000000000000001O000000000O100000000000001O0000000001O1O1O1O1OQL"}, "label": "brown smooth brief case"}]}
{"id": 452750, "image": "COCO_train2014_000000452750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting with plate , spoon and knife with other hand\"."}], "task": "refering", "answer_template": "The \"a woman sitting with plate , spoon and knife with other hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 137, 138, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.0, 0.3316816816816817, 0.73962, 0.984924924924925], "iscrowd": 0, "area": 40991.9437, "rle": {"size": [333, 500], "counts": "o8V1i6^2cM]2I700000001O000001O000000000000001O00000000000000001O00000000001O001O1O001O001O001O001O00O1O1O1N2M3M3M3N2O100O100O100O1O100O100O100O100001O2N2N1O2N1\\JSLf4P4WKQLh4Q4VKQLh4R4UKoKk4S4RKoKl4Q4TKPLk4P4UKRLi4n3WKTLg4l3YKULf4k3ZKVLe4j3[KWLd4j3[KXLc4h3]KYLb4g3^KZLa4f3_K\\L_4d3`K^L_4b3aK_L^4a3bK`L]4a3bK_L^4a3bK`L]4`3cK`L]4b3aK_L^4f3]KZLc4k3XKVLg4h401O00001O000000001O000000001O00000000001O000000000000000000000000000000001O001O001O001O00001O001O002N101N2N2N1O2N2N1O2N2N2N2N2N2N2N2N2N3M4L1O0000000000000000O1O1O1O1O1O1N2O1O1O001O000O20O100O1O2O0O100O1O2mMnHi0R7UOoHk0R7SOPIm0P7ROQIm0P7QOQIP1P7mNRIR1o6lNRIT1o6jNSIV1m6hNTIX1o6dNSI\\1P7_NQIa1S7[NnHd1V7WNkHj1W7RNkHm1j710O00100O010O0010O010O01O010O010O00100O010O00100O100O00100O10O01O100O10O0100O100O01000O100O010O100O010O10000O010O100O10O01O100O10O01O10O010O01O10O01O10O0100O010O010O10O0100O010O01000O010O10O0100O100O10000O100000000O101OO100O1O1O1O010O1O1000000O0100000O10000O10O1000O1000O10O01O101N3N1N3N1N3M2O2M2L5K4M4K4M4KjQZ1"}, "label": "a woman sitting with plate , spoon and knife with other hand"}]}
{"id": 452750, "image": "COCO_train2014_000000452750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black shirt with buttons holding a plate standing in the kitchen\"."}], "task": "refering", "answer_template": "The \"a woman in a black shirt with buttons holding a plate standing in the kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 137, 138, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.0, 0.3316816816816817, 0.73962, 0.984924924924925], "iscrowd": 0, "area": 40991.9437, "rle": {"size": [333, 500], "counts": "o8V1i6^2cM]2I700000001O000001O000000000000001O00000000000000001O00000000001O001O1O001O001O001O001O00O1O1O1N2M3M3M3N2O100O100O100O1O100O100O100O100001O2N2N1O2N1\\JSLf4P4WKQLh4Q4VKQLh4R4UKoKk4S4RKoKl4Q4TKPLk4P4UKRLi4n3WKTLg4l3YKULf4k3ZKVLe4j3[KWLd4j3[KXLc4h3]KYLb4g3^KZLa4f3_K\\L_4d3`K^L_4b3aK_L^4a3bK`L]4a3bK_L^4a3bK`L]4`3cK`L]4b3aK_L^4f3]KZLc4k3XKVLg4h401O00001O000000001O000000001O00000000001O000000000000000000000000000000001O001O001O001O00001O001O002N101N2N2N1O2N2N1O2N2N2N2N2N2N2N2N2N3M4L1O0000000000000000O1O1O1O1O1O1N2O1O1O001O000O20O100O1O2O0O100O1O2mMnHi0R7UOoHk0R7SOPIm0P7ROQIm0P7QOQIP1P7mNRIR1o6lNRIT1o6jNSIV1m6hNTIX1o6dNSI\\1P7_NQIa1S7[NnHd1V7WNkHj1W7RNkHm1j710O00100O010O0010O010O01O010O010O00100O010O00100O100O00100O10O01O100O10O0100O100O01000O100O010O100O010O10000O010O100O10O01O100O10O01O10O010O01O10O01O10O0100O010O010O10O0100O010O01000O010O10O0100O100O10000O100000000O101OO100O1O1O1O010O1O1000000O0100000O10000O10O1000O1000O10O01O101N3N1N3N1N3M2O2M2L5K4M4K4M4KjQZ1"}, "label": "a woman in a black shirt with buttons holding a plate standing in the kitchen"}]}
{"id": 452750, "image": "COCO_train2014_000000452750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand graced with a bracelet reached over toward the table where the ice cream sits\"."}], "task": "refering", "answer_template": "The \"a hand graced with a bracelet reached over toward the table where the ice cream sits\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 18, 19, 20, 21, 22, 23, 36, 37], "bbox": [0.0015, 0.0, 0.38314, 0.25393393393393393], "iscrowd": 0, "area": 7198.8102, "rle": {"size": [333, 500], "counts": "^;S1Y9b0_O0O01O001O010O010O010O0010O010O0100O0010O010O01O010O10O0100O00100O10O01O10O0100O10O01O10O0100O010O00100O010O1O010O10O010O00001O001O000100O100O100O100O01000O10000O10000O10000O10000O10000000000000000000000000000O10000000000000000001O0000000000001O00000000000000000000O1000000000000O100000000000000O1000O100000000000O1O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100M3M;EoTT3"}, "label": "a hand graced with a bracelet reached over toward the table where the ice cream sits"}]}
{"id": 452750, "image": "COCO_train2014_000000452750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand and wrist with a watch on the wrist\"."}], "task": "refering", "answer_template": "The \"the hand and wrist with a watch on the wrist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 18, 19, 20, 21, 22, 23, 36, 37], "bbox": [0.0015, 0.0, 0.38314, 0.25393393393393393], "iscrowd": 0, "area": 7198.8102, "rle": {"size": [333, 500], "counts": "^;S1Y9b0_O0O01O001O010O010O010O0010O010O0100O0010O010O01O010O10O0100O00100O10O01O10O0100O10O01O10O0100O010O00100O010O1O010O10O010O00001O001O000100O100O100O100O01000O10000O10000O10000O10000O10000000000000000000000000000O10000000000000000001O0000000000001O00000000000000000000O1000000000000O100000000000000O1000O100000000000O1O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100M3M;EoTT3"}, "label": "the hand and wrist with a watch on the wrist"}]}
{"id": 370831, "image": "COCO_train2014_000000370831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch with a coffee table in front of it\"."}], "task": "refering", "answer_template": "The \"a brown couch with a coffee table in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 366, 367, 390], "bbox": [0.761296875, 0.39570833333333333, 1.0, 0.9806458333333333], "iscrowd": 0, "area": 30245.728849999992, "rle": {"size": [480, 640], "counts": "mbT75Z>b0iM\\O_EV1a:ROeDa1Z;S100000000000001O00000bNWMdFi2[9^M_Fb2`9iMhEe2W:S1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001N2O1O1N2O001N2O1O1O1N101O1N2O1O1O0O2O1O1N2O10O3N2N2N1OO1N101N2O1N2O1O0O2O1O1O1O10O01O1O1O1O1O001O1O1a0_Oi0WOi0VOj0WOhF"}, "label": "a brown couch with a coffee table in front of it"}]}
{"id": 370831, "image": "COCO_train2014_000000370831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch that has a forest green pillow on it\"."}], "task": "refering", "answer_template": "The \"a brown couch that has a forest green pillow on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 366, 367, 390], "bbox": [0.761296875, 0.39570833333333333, 1.0, 0.9806458333333333], "iscrowd": 0, "area": 30245.728849999992, "rle": {"size": [480, 640], "counts": "mbT75Z>b0iM\\O_EV1a:ROeDa1Z;S100000000000001O00000bNWMdFi2[9^M_Fb2`9iMhEe2W:S1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001N2O1O1N2O001N2O1O1O1N101O1N2O1O1O0O2O1O1N2O10O3N2N2N1OO1N101N2O1N2O1O0O2O1O1O1O10O01O1O1O1O1O001O1O1a0_Oi0WOi0VOj0WOhF"}, "label": "a brown couch that has a forest green pillow on it"}]}
{"id": 370831, "image": "COCO_train2014_000000370831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the floor plant on the right side of the tv\"."}], "task": "refering", "answer_template": "The \"the floor plant on the right side of the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 38, 39, 40, 41, 42, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 200, 201, 223, 224], "bbox": [0.658984375, 0.03333333333333333, 0.841015625, 0.5659166666666666], "iscrowd": 0, "area": 17548.2995, "rle": {"size": [480, 640], "counts": "`kU63k>7J5K6J5K6J5K6I6K6J7I7I7I7I6J7H6K4M2^FPMo6T3`H_M[7c2]HeMa7]2WHkMf7Y2QHnMn7T2jGRNV8P2bGVN]8n1ZGRNl8T4M3L4M3Ma0^Ob0_O3M3L1000000O10000O10000O10000O10000O10000O10_NUImJk6R5\\I_Jn6`5XImIU7S6l0O101N10001N100O2O000O2O0O101O0O101O0O101N1L4cNmFjLW9S3nFfLV9X3nFcLU9[3PG^LU9^3PG^LR9`3QG^LQ9_3SG_Ln8^3UG`Lm8]3VGaLl8[3YGbLi8W3^GgLd8S3h1J6I6M4M3L4L4N2O1N2O1O1O1O1N2O001O1O1O1N2O9Ga0_Oa0_ORg_1"}, "label": "the floor plant on the right side of the tv"}]}
{"id": 370831, "image": "COCO_train2014_000000370831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large , leafy houseplant next to a television set\"."}], "task": "refering", "answer_template": "The \"a large , leafy houseplant next to a television set\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 38, 39, 40, 41, 42, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 200, 201, 223, 224], "bbox": [0.658984375, 0.03333333333333333, 0.841015625, 0.5659166666666666], "iscrowd": 0, "area": 17548.2995, "rle": {"size": [480, 640], "counts": "`kU63k>7J5K6J5K6J5K6I6K6J7I7I7I7I6J7H6K4M2^FPMo6T3`H_M[7c2]HeMa7]2WHkMf7Y2QHnMn7T2jGRNV8P2bGVN]8n1ZGRNl8T4M3L4M3Ma0^Ob0_O3M3L1000000O10000O10000O10000O10000O10000O10_NUImJk6R5\\I_Jn6`5XImIU7S6l0O101N10001N100O2O000O2O0O101O0O101O0O101N1L4cNmFjLW9S3nFfLV9X3nFcLU9[3PG^LU9^3PG^LR9`3QG^LQ9_3SG_Ln8^3UG`Lm8]3VGaLl8[3YGbLi8W3^GgLd8S3h1J6I6M4M3L4L4N2O1N2O1O1O1O1N2O001O1O1O1N2O9Ga0_Oa0_ORg_1"}, "label": "a large , leafy houseplant next to a television set"}]}
{"id": 370831, "image": "COCO_train2014_000000370831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black leather sofa near window with white and red crocheted blanket\"."}], "task": "refering", "answer_template": "The \"black leather sofa near window with white and red crocheted blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 322, 323], "bbox": [0.0, 0.34831249999999997, 0.244375, 0.8539375], "iscrowd": 0, "area": 25314.40285000001, "rle": {"size": [480, 640], "counts": "e55j>:G:E;UDB_8j0bF\\1l7Z4aN100O100O1O100O101N1O100O100O1O100O1001O010O00001N101O000O2O001N10001O0O101O001N10001N1O2N1O1O2N2N2N3M3M2N3M2N3M2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1N2O1O1O2N2N3M2N2N3K<oNjBDmQS7"}, "label": "black leather sofa near window with white and red crocheted blanket"}]}
{"id": 370831, "image": "COCO_train2014_000000370831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch with a red and white blanket across the back of it\"."}], "task": "refering", "answer_template": "The \"a couch with a red and white blanket across the back of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 322, 323], "bbox": [0.0, 0.34831249999999997, 0.244375, 0.8539375], "iscrowd": 0, "area": 25314.40285000001, "rle": {"size": [480, 640], "counts": "e55j>:G:E;UDB_8j0bF\\1l7Z4aN100O100O1O100O101N1O100O100O1O100O1001O010O00001N101O000O2O001N10001O0O101O001N10001N1O2N1O1O2N2N2N3M3M2N3M2N3M2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1N2O1O1O2N2N3M2N2N3K<oNjBDmQS7"}, "label": "a couch with a red and white blanket across the back of it"}]}
{"id": 84114, "image": "COCO_train2014_000000084114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with brown hair and black marks on each cheek and a glove on his left hand\"."}], "task": "refering", "answer_template": "The \"a boy with brown hair and black marks on each cheek and a glove on his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 89, 90, 91, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 136, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 218, 219, 220, 221, 234, 235, 236, 237, 250, 251, 252, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 346, 348, 349, 361, 362, 363, 364, 365], "bbox": [0.5189277899343545, 0.15621875000000002, 0.9196498905908096, 1.0], "iscrowd": 0, "area": 51684.39454999997, "rle": {"size": [640, 457], "counts": "S\\d42jc05K4L5J6K5K4L5J5L5M2O2M2O2M2N3N2M2O2M2N3O1N3N2jMeMbB_2Z=dMdB^2Y=dMeB_2X=dMfB^2W=dMhB^2U=dMiB`2S=cMkB_2n<gMhA^Og0n2Y=lMjA[Ol0k2f5oLa0Q1kHWOm0k2d5PMb0P1iHXOo0k2c5oLc0_1hIe1j4hLkJ8`6]1jIe1h4iLkJ7a6]1jIf1f4jLlJ5b6\\1lIg1c4kLkJ6c6Y1nIi1`4lLlJ4e6W1oIk1\\3[LgLc0WO2X3YOOm1^Mm1Z3^LdL`0[O0U3\\O3j1^MP2W3aLaLZ1_2^N:h1_MR2S3dL^L_1Z2VNe0g1^MS2R3gL[Lb1U2PNo0e1_MU2n2_OXNjL[1c1]MX2l2DRNcLe1a1]M[2h2KkM[LP2`1\\M]2e21fMSLY2`1ZM_2e27]MlKd2^1ZMb2a2=WMeKm2^1ZMc2^2^18PLXMf2\\2[1<oKXMi2X2Y1`0oKWMj2U2Y1d0nKVMl2Q2X1i0mKTMo2n1V1n0oKPMn2l1V1T1PLlLm2[1`JROe6f2d1UNjGPOe6k2d1QNmGmNa6R3e1mMiNS2Z1iMgNW2\\1dMfN\\2^1_McNb2_1ZMbNf2b1UM_Nk2d1kLcNV3S900004L00000000000]NZMlAf2f=mMUBT2`=WN^Bj1k9hLmHb1VMg1j9mLhG2^N^1Ne1h9PMiG5WN]5l9dJhG8RNW5R:gJhG[6T8kIhGW6T8nIhGU6T8QJRGg6l8X2O1N2001mF`Fc7b9WG`Fb07U8m9[G\\Fd8a:0REWGd:m0YEf6MbHl:a0\\EV7h:eH[EZ7i:_H[Ea7M\\HHOm9MaFj7BmHR:TO]FR9X:100O0100O1O4M4K5RJfDl2_;kLfDT3_;cJfD[1OR4d<VK`Ck4P=`JTC`5f=000000001O000000010O00001O1O001SIZJ^Og5>fJXOZ5e0kJYOV5c0PKZOP5c0UK[Ok4c0XK\\Oh4c0ZK\\Of4c0^KSNHZKj4b6bKcMT6[2PJlKcNdNi7^5iIhKgNWNn7o5`I^JZN0Y9a5aHWJ\\N5U9c5cIZJa6d5aIZJ`6d5bIZJa6d5bIXJ`6f5dIVJ_6g5fIUJ\\6h5P4N2M2O2N1O1N3N1O1O2N1N4M3E:M4M3L3N3L5L6I6TOm0kNU1D?G6K2O0O2O1M4K=D9F]kf0"}, "label": "a boy with brown hair and black marks on each cheek and a glove on his left hand"}]}
{"id": 84114, "image": "COCO_train2014_000000084114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in baseball attire with eye black on\"."}], "task": "refering", "answer_template": "The \"a young boy in baseball attire with eye black on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 89, 90, 91, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 136, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 218, 219, 220, 221, 234, 235, 236, 237, 250, 251, 252, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 346, 348, 349, 361, 362, 363, 364, 365], "bbox": [0.5189277899343545, 0.15621875000000002, 0.9196498905908096, 1.0], "iscrowd": 0, "area": 51684.39454999997, "rle": {"size": [640, 457], "counts": "S\\d42jc05K4L5J6K5K4L5J5L5M2O2M2O2M2N3N2M2O2M2N3O1N3N2jMeMbB_2Z=dMdB^2Y=dMeB_2X=dMfB^2W=dMhB^2U=dMiB`2S=cMkB_2n<gMhA^Og0n2Y=lMjA[Ol0k2f5oLa0Q1kHWOm0k2d5PMb0P1iHXOo0k2c5oLc0_1hIe1j4hLkJ8`6]1jIe1h4iLkJ7a6]1jIf1f4jLlJ5b6\\1lIg1c4kLkJ6c6Y1nIi1`4lLlJ4e6W1oIk1\\3[LgLc0WO2X3YOOm1^Mm1Z3^LdL`0[O0U3\\O3j1^MP2W3aLaLZ1_2^N:h1_MR2S3dL^L_1Z2VNe0g1^MS2R3gL[Lb1U2PNo0e1_MU2n2_OXNjL[1c1]MX2l2DRNcLe1a1]M[2h2KkM[LP2`1\\M]2e21fMSLY2`1ZM_2e27]MlKd2^1ZMb2a2=WMeKm2^1ZMc2^2^18PLXMf2\\2[1<oKXMi2X2Y1`0oKWMj2U2Y1d0nKVMl2Q2X1i0mKTMo2n1V1n0oKPMn2l1V1T1PLlLm2[1`JROe6f2d1UNjGPOe6k2d1QNmGmNa6R3e1mMiNS2Z1iMgNW2\\1dMfN\\2^1_McNb2_1ZMbNf2b1UM_Nk2d1kLcNV3S900004L00000000000]NZMlAf2f=mMUBT2`=WN^Bj1k9hLmHb1VMg1j9mLhG2^N^1Ne1h9PMiG5WN]5l9dJhG8RNW5R:gJhG[6T8kIhGW6T8nIhGU6T8QJRGg6l8X2O1N2001mF`Fc7b9WG`Fb07U8m9[G\\Fd8a:0REWGd:m0YEf6MbHl:a0\\EV7h:eH[EZ7i:_H[Ea7M\\HHOm9MaFj7BmHR:TO]FR9X:100O0100O1O4M4K5RJfDl2_;kLfDT3_;cJfD[1OR4d<VK`Ck4P=`JTC`5f=000000001O000000010O00001O1O001SIZJ^Og5>fJXOZ5e0kJYOV5c0PKZOP5c0UK[Ok4c0XK\\Oh4c0ZK\\Of4c0^KSNHZKj4b6bKcMT6[2PJlKcNdNi7^5iIhKgNWNn7o5`I^JZN0Y9a5aHWJ\\N5U9c5cIZJa6d5aIZJ`6d5bIZJa6d5bIXJ`6f5dIVJ_6g5fIUJ\\6h5P4N2M2O2N1O1N3N1O1O2N1N4M3E:M4M3L3N3L5L6I6TOm0kNU1D?G6K2O0O2O1M4K=D9F]kf0"}, "label": "a young boy in baseball attire with eye black on"}]}
{"id": 84114, "image": "COCO_train2014_000000084114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black baseball glove\"."}], "task": "refering", "answer_template": "The \"a black baseball glove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 230, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 309], "bbox": [0.12568927789934353, 0.62615625, 0.460109409190372, 0.8366875], "iscrowd": 0, "area": 9619.068150000005, "rle": {"size": [640, 457], "counts": "ofT13lc03M2N2N3N1O1N3N1O1N3N1N3N1N2N2O0O1O2O0O2N100O2O0O2N101N11O2O1N2O0O2O1N2N10O0100O010O00100O010O010O1O010O00000O100000O1000000000000O10O01gNAQ_O?o`0Bo^O?Pa0Dm^O=Sa0Dl^O<Sa0Gj^O:Va0Gi^O9Va0Ii^O7Wa0Kh^O4Wa0Nh^O2Xa00f^O0Ya02f^OM[a04d^ONYa05b^O0\\a0\\1M4M2M2O1N2O1N3E:F:N2N2N100O1O1O1O1O1O100O1O1O00100001O0000001O0O1O2N1O101N1O1O2N1O1TOk_OmMV`0R2m_OjMT`0U2Q@fMR`0X2Q@dMS`0X2Q@dMS`0W2R@eMR`0V2n0N1O1O2N1O2O0O3M3M4L3M4L3M4K4K6K5J5L5J5L5Jkmi4"}, "label": "a black baseball glove"}]}
{"id": 84114, "image": "COCO_train2014_000000084114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with a black baseball glove\"."}], "task": "refering", "answer_template": "The \"boy with a black baseball glove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 230, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 309], "bbox": [0.12568927789934353, 0.62615625, 0.460109409190372, 0.8366875], "iscrowd": 0, "area": 9619.068150000005, "rle": {"size": [640, 457], "counts": "ofT13lc03M2N2N3N1O1N3N1O1N3N1N3N1N2N2O0O1O2O0O2N100O2O0O2N101N11O2O1N2O0O2O1N2N10O0100O010O00100O010O010O1O010O00000O100000O1000000000000O10O01gNAQ_O?o`0Bo^O?Pa0Dm^O=Sa0Dl^O<Sa0Gj^O:Va0Gi^O9Va0Ii^O7Wa0Kh^O4Wa0Nh^O2Xa00f^O0Ya02f^OM[a04d^ONYa05b^O0\\a0\\1M4M2M2O1N2O1N3E:F:N2N2N100O1O1O1O1O1O100O1O1O00100001O0000001O0O1O2N1O101N1O1O2N1O1TOk_OmMV`0R2m_OjMT`0U2Q@fMR`0X2Q@dMS`0X2Q@dMS`0W2R@eMR`0V2n0N1O1O2N1O2O0O3M3M4L3M4L3M4K4K6K5J5L5J5L5Jkmi4"}, "label": "boy with a black baseball glove"}]}
{"id": 84114, "image": "COCO_train2014_000000084114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy holding his glove with both hands\"."}], "task": "refering", "answer_template": "The \"the boy holding his glove with both hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 67, 68, 69, 70, 83, 84, 85, 86, 99, 100, 101, 102, 115, 116, 117, 118, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360], "bbox": [0.0, 0.134015625, 0.6655142231947483, 0.994375], "iscrowd": 0, "area": 102052.15335000001, "rle": {"size": [640, 457], "counts": "];R1kb04L5L3LQ1oN4L4L4L4L3M4L4L4M3L4L4L4L3M4L4L4L4L4L4M2M4L4L4L4M3N2N1O2N2N2N1O1O2N1N2O1O1O1O2N1O100O100O101N100O100O101ZGPJY3Q6bKTK^4l4]JYLb5h3bITM^6l2XI^Mg6c2nGXJ6_3l7d7O100O100O100O100O100O100O100O100O100O100O100O1[MmC]MS<^2\\DWMe;d2jDRMV;i2XEmLi:m2^ERMb:i2dEUM]:f2iEYMW:b2oE\\MR:a2RF]Mo9a2TF^Ml9`2WF^Mj9`2YF^Mh9`2[F_Me9_2^F_Mc9^2aF`M`9^2cFaM]9]2fFaM[9]2iF`MX9_2kFYM[9f2P400O1O1O100O1O100O1O1O101N1O100O1O1O10000000000000000000000O100000000000000000000001O001O00001O001O001O001O001O001O001O00001O1O1O001O1O1O1O001O1O1O001lBcLc9^3YFiLc9X3ZFoLa9R3[FUMa9k2\\F\\Ma9d2[FcMa9^2]FhM^9Y2`FmM[9Z2^FfMb9`2WFaMi9f2PFZMP:l2iEVMV:P3aESM_:T3TERMl:T3iDQMW;S3`DSM_;R3WDSMi;b50001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O2N1O1XK\\HRMe7l2`IPLb6m3cJoJ^5o4fKnI[4P6m4M2M3N2N3M2N2M3N3M2N2N3M2M3N2N3M2N2M3O2M2N2N3N1N3M4L4L4M3L3M4L4M3L4L3M4M3L4L3M4L4M3L4L3M4M3L4L4L3L5I7J6I7I6K6I7I7J6IgWo2"}, "label": "the boy holding his glove with both hands"}]}
{"id": 84114, "image": "COCO_train2014_000000084114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young baseball player wearing a turtleneck underlayer and holding his baseball glove in his hands\"."}], "task": "refering", "answer_template": "The \"a young baseball player wearing a turtleneck underlayer and holding his baseball glove in his hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 67, 68, 69, 70, 83, 84, 85, 86, 99, 100, 101, 102, 115, 116, 117, 118, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360], "bbox": [0.0, 0.134015625, 0.6655142231947483, 0.994375], "iscrowd": 0, "area": 102052.15335000001, "rle": {"size": [640, 457], "counts": "];R1kb04L5L3LQ1oN4L4L4L4L3M4L4L4M3L4L4L4L3M4L4L4L4L4L4M2M4L4L4L4M3N2N1O2N2N2N1O1O2N1N2O1O1O1O2N1O100O100O101N100O100O101ZGPJY3Q6bKTK^4l4]JYLb5h3bITM^6l2XI^Mg6c2nGXJ6_3l7d7O100O100O100O100O100O100O100O100O100O100O100O1[MmC]MS<^2\\DWMe;d2jDRMV;i2XEmLi:m2^ERMb:i2dEUM]:f2iEYMW:b2oE\\MR:a2RF]Mo9a2TF^Ml9`2WF^Mj9`2YF^Mh9`2[F_Me9_2^F_Mc9^2aF`M`9^2cFaM]9]2fFaM[9]2iF`MX9_2kFYM[9f2P400O1O1O100O1O100O1O1O101N1O100O1O1O10000000000000000000000O100000000000000000000001O001O00001O001O001O001O001O001O001O00001O1O1O001O1O1O1O001O1O1O001lBcLc9^3YFiLc9X3ZFoLa9R3[FUMa9k2\\F\\Ma9d2[FcMa9^2]FhM^9Y2`FmM[9Z2^FfMb9`2WFaMi9f2PFZMP:l2iEVMV:P3aESM_:T3TERMl:T3iDQMW;S3`DSM_;R3WDSMi;b50001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O2N1O1XK\\HRMe7l2`IPLb6m3cJoJ^5o4fKnI[4P6m4M2M3N2N3M2N2M3N3M2N2N3M2M3N2N3M2N2M3O2M2N2N3N1N3M4L4L4M3L3M4L4M3L4L3M4M3L4L3M4L4M3L4L3M4M3L4L4L3L5I7J6I7I6K6I7I7J6IgWo2"}, "label": "a young baseball player wearing a turtleneck underlayer and holding his baseball glove in his hands"}]}
{"id": 75924, "image": "COCO_train2014_000000075924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone ' s hand brushing the teeth of the child\"."}], "task": "refering", "answer_template": "The \"someone ' s hand brushing the teeth of the child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375], "bbox": [0.023979166666666666, 0.22696875, 0.6711666666666666, 0.9707812499999999], "iscrowd": 0, "area": 66266.37735000001, "rle": {"size": [640, 480], "counts": "aj7l8U6o4O100O100O1O100O100O1O100O1O100O100O1O10O0100O1O100O00100O100O1O010O10O01O10O01O10O0100O010000O10000O100O1000O010000O1000000O10000O10000eI[JZNf5e1\\JZNd5e1]J[Nc5d1_JZNb5e1_J[Na5d1aJ[N_5d1bJ[N_5d1cJ[N]5d1dJ\\N[5c1hJ[NY5d1hJ\\NX5c1jJ[NW5d1jJ\\NV5c1kJ\\NV5c1lJ\\NT5gMUJ8h0P2T5fMWJ8f0R2S5dMZJ7f0S2Q5dM]J6c0V2P5bM`J6a0W2P5`MdJ6>Y2n4_MgJ6<Z2n4^MjJ59]2m4]MlJ49]2l4_MlJ29_2k4^MoJ17_2k4_MPK15`2k4_MRKN4b2k4_MSKN3b2j4`MUKL1c2j4aMWKKOd2j4aMXKI0d2i4cMYKGNe2j4dMZKELg2j4dM\\KCKg2j4fM]K@Ji2j4gM^K^OIi2j4iM`K[OFj2l4kM`KYODk2m4kMbKWOCk2m4nMbKUOAk2o4PNcKRO^Om2P5QNdKPO]Ol2Q5SNeKnN[On2R5SNeKmNYOo2S5TNfKkNXOo2S5VNhKhNUOP3U5WNiKgNROQ3V5XNjKdNROR3U5ZNkKbNPOR3W5\\NkK`NoNR3W5^NmK]NlNT3X5^NoK[NjNU3Y5`NoKYNiNU3Y5bNPLVNhNW3Y5cNQLPNjN[3W5dNSLjMkN`3S5fNTLeMlNd3Q5gNUL_MnNi3n4hNVLYMQOm3j4iNgNW1Y1iNgNV1Z1jNgNT1Z1lNfNS1[1lNgNR1Z1nNgNo0\\1POdNo0]1QOdNl0^1SOcNk0_1UObNh0`1XO`Nf0b1ZO_Nd0b1\\O_Na0c1^O^N`0d1@]N>d1B]N;e1E[N9g1F[N7g1IZN5g1JZN4h1KZNGS29nMAX2=iM_O[2a0fMYO_2f0cMUOa2j0`MTOb2l0_MROb2m0`MQOa2o0_MoNc2P1_MnNb2Q1`MlNb2T1^MjNc2V1_MgNc2Y1^MeNc2Z1^MdNd2\\1]MaNe2^1]M`Nd2_1]MaNc2_1^M`Nb2_1`M`N_2a1aM_N_2`1cM_N]2a1dM^N\\2a1eM_N[2a1fM^NZ2b1gM]NY2c1gM]NX2c1jM\\NV2d1kM[NU2e1kM[NU2d1mMZNT2f1mMYNS2g1mMYNS2g1nMXNR2g1PNXNo1i1RNVNn1j1RNVNn1i1TNVNl1j1UNTNl1l1TNTNl1k1VNSNk1m1VNRNi1o1WNQNi1n1YNPNh1P2YNoMg1Q2YNnMh1R2S8O10O10O10000O10000O1000001N10000O10000O10000O1000000O10000O10000O10000O1000000O101O0O10000O10000O1000001N10000O2O000O101O0O101O000O101O0O2O001N10001O0O2O00001O010O001O00001O001O00001O0O2O000O2O0O2O0O2O0O101N101N1O101N101N100O2OebR3"}, "label": "someone ' s hand brushing the teeth of the child"}]}
{"id": 75924, "image": "COCO_train2014_000000075924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"arm and hand of adult helping child\"."}], "task": "refering", "answer_template": "The \"arm and hand of adult helping child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375], "bbox": [0.023979166666666666, 0.22696875, 0.6711666666666666, 0.9707812499999999], "iscrowd": 0, "area": 66266.37735000001, "rle": {"size": [640, 480], "counts": "aj7l8U6o4O100O100O1O100O100O1O100O1O100O100O1O10O0100O1O100O00100O100O1O010O10O01O10O01O10O0100O010000O10000O100O1000O010000O1000000O10000O10000eI[JZNf5e1\\JZNd5e1]J[Nc5d1_JZNb5e1_J[Na5d1aJ[N_5d1bJ[N_5d1cJ[N]5d1dJ\\N[5c1hJ[NY5d1hJ\\NX5c1jJ[NW5d1jJ\\NV5c1kJ\\NV5c1lJ\\NT5gMUJ8h0P2T5fMWJ8f0R2S5dMZJ7f0S2Q5dM]J6c0V2P5bM`J6a0W2P5`MdJ6>Y2n4_MgJ6<Z2n4^MjJ59]2m4]MlJ49]2l4_MlJ29_2k4^MoJ17_2k4_MPK15`2k4_MRKN4b2k4_MSKN3b2j4`MUKL1c2j4aMWKKOd2j4aMXKI0d2i4cMYKGNe2j4dMZKELg2j4dM\\KCKg2j4fM]K@Ji2j4gM^K^OIi2j4iM`K[OFj2l4kM`KYODk2m4kMbKWOCk2m4nMbKUOAk2o4PNcKRO^Om2P5QNdKPO]Ol2Q5SNeKnN[On2R5SNeKmNYOo2S5TNfKkNXOo2S5VNhKhNUOP3U5WNiKgNROQ3V5XNjKdNROR3U5ZNkKbNPOR3W5\\NkK`NoNR3W5^NmK]NlNT3X5^NoK[NjNU3Y5`NoKYNiNU3Y5bNPLVNhNW3Y5cNQLPNjN[3W5dNSLjMkN`3S5fNTLeMlNd3Q5gNUL_MnNi3n4hNVLYMQOm3j4iNgNW1Y1iNgNV1Z1jNgNT1Z1lNfNS1[1lNgNR1Z1nNgNo0\\1POdNo0]1QOdNl0^1SOcNk0_1UObNh0`1XO`Nf0b1ZO_Nd0b1\\O_Na0c1^O^N`0d1@]N>d1B]N;e1E[N9g1F[N7g1IZN5g1JZN4h1KZNGS29nMAX2=iM_O[2a0fMYO_2f0cMUOa2j0`MTOb2l0_MROb2m0`MQOa2o0_MoNc2P1_MnNb2Q1`MlNb2T1^MjNc2V1_MgNc2Y1^MeNc2Z1^MdNd2\\1]MaNe2^1]M`Nd2_1]MaNc2_1^M`Nb2_1`M`N_2a1aM_N_2`1cM_N]2a1dM^N\\2a1eM_N[2a1fM^NZ2b1gM]NY2c1gM]NX2c1jM\\NV2d1kM[NU2e1kM[NU2d1mMZNT2f1mMYNS2g1mMYNS2g1nMXNR2g1PNXNo1i1RNVNn1j1RNVNn1i1TNVNl1j1UNTNl1l1TNTNl1k1VNSNk1m1VNRNi1o1WNQNi1n1YNPNh1P2YNoMg1Q2YNnMh1R2S8O10O10O10000O10000O1000001N10000O10000O10000O1000000O10000O10000O10000O1000000O101O0O10000O10000O1000001N10000O2O000O101O0O101O000O101O0O2O001N10001O0O2O00001O010O001O00001O001O00001O0O2O000O2O0O2O0O2O0O101N101N1O101N101N100O2OebR3"}, "label": "arm and hand of adult helping child"}]}
{"id": 75924, "image": "COCO_train2014_000000075924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child getting their teeth brushed\"."}], "task": "refering", "answer_template": "The \"a young child getting their teeth brushed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.2528125, 0.27532812500000003, 0.9870208333333333, 0.97921875], "iscrowd": 0, "area": 115231.57525, "rle": {"size": [640, 480], "counts": "bW\\21mc02N2N2M3N2N2N2M3N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2kHXNPLj1j3kNeKW1U4^OZKd0`41PK0j4;oJGk4c0PK^Oj4k0RKVOh4T1SKmNg4\\1UKeNe4e1UK]Ne4l1WKUNc4U2XKlMb4]2ZKdM`4f2ZK\\Md4g2ZKZMd4j2YKWMf4k2XKVMf4m2XKTMg4o2UKSMi4P3UKQMj4Q3TKPMj4T3SKmLl4U3QKmLm4W3PKjLo4X3oJiLo4Z3oJgLP5\\3mJeLQ5^3lJdLS5^3kJcLS5a3jJ`LU5b3iJ_LU5e3hJ\\LW5f3fJ\\LX5g3fJZLY5i3dJXLZ5k3dJVL[5l3bJVL\\5m3bJTL]5m3bJTL\\5o3bJRL]5P4kIgLU6Z3jIfLV6[3hIfLW6[3iIeLW6\\3gIeLY6\\3fIdLZ6]3dIdL[6^3cIcL]6]3cIcL]6^3aIcL_6^3`IbL_6`3_IaLa6`3^I`Lb6`3]IaLc6`3\\I`Ld6a3ZI`Le6b3ZI^Lf6c3XI^Lh6b3XI^Lh6c3VI^Li6e3TI\\Ll6g3QIYLo6\\800O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O10000O100O10000O100O100O10000O100O100O10000O100O11O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2RKlFlNU9P1PGnNU9k0oFSOW9d0mF[OX9>lF@Y98lFFY93kFK[9LjF2[9GhF8]9AgF=_9ZOfFd0_9UOeFi0`9oNeFo0`9jNcFU1c9bNbF\\1c9]NaFa1e9VN`Fh1f9PN]Fo1i9hM[FW2k9aMXF^2m9ZMXFd2n9TMUFk2Q:lLSFS3S:eLPFZ3V:]LoEa3Y=O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1L5WOh0XOh0XOcW3"}, "label": "a young child getting their teeth brushed"}]}
{"id": 75924, "image": "COCO_train2014_000000075924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child\"."}], "task": "refering", "answer_template": "The \"child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.2528125, 0.27532812500000003, 0.9870208333333333, 0.97921875], "iscrowd": 0, "area": 115231.57525, "rle": {"size": [640, 480], "counts": "bW\\21mc02N2N2M3N2N2N2M3N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2kHXNPLj1j3kNeKW1U4^OZKd0`41PK0j4;oJGk4c0PK^Oj4k0RKVOh4T1SKmNg4\\1UKeNe4e1UK]Ne4l1WKUNc4U2XKlMb4]2ZKdM`4f2ZK\\Md4g2ZKZMd4j2YKWMf4k2XKVMf4m2XKTMg4o2UKSMi4P3UKQMj4Q3TKPMj4T3SKmLl4U3QKmLm4W3PKjLo4X3oJiLo4Z3oJgLP5\\3mJeLQ5^3lJdLS5^3kJcLS5a3jJ`LU5b3iJ_LU5e3hJ\\LW5f3fJ\\LX5g3fJZLY5i3dJXLZ5k3dJVL[5l3bJVL\\5m3bJTL]5m3bJTL\\5o3bJRL]5P4kIgLU6Z3jIfLV6[3hIfLW6[3iIeLW6\\3gIeLY6\\3fIdLZ6]3dIdL[6^3cIcL]6]3cIcL]6^3aIcL_6^3`IbL_6`3_IaLa6`3^I`Lb6`3]IaLc6`3\\I`Ld6a3ZI`Le6b3ZI^Lf6c3XI^Lh6b3XI^Lh6c3VI^Li6e3TI\\Ll6g3QIYLo6\\800O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O10000O100O10000O100O100O10000O100O100O10000O100O11O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2RKlFlNU9P1PGnNU9k0oFSOW9d0mF[OX9>lF@Y98lFFY93kFK[9LjF2[9GhF8]9AgF=_9ZOfFd0_9UOeFi0`9oNeFo0`9jNcFU1c9bNbF\\1c9]NaFa1e9VN`Fh1f9PN]Fo1i9hM[FW2k9aMXF^2m9ZMXFd2n9TMUFk2Q:lLSFS3S:eLPFZ3V:]LoEa3Y=O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1L5WOh0XOh0XOcW3"}, "label": "child"}]}
{"id": 215191, "image": "COCO_train2014_000000215191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toilet in between two toilets\"."}], "task": "refering", "answer_template": "The \"toilet in between two toilets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 149, 172, 194, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312], "bbox": [0.42978125, 0.34158333333333335, 0.581453125, 0.7977500000000001], "iscrowd": 0, "area": 10214.969750000002, "rle": {"size": [480, 640], "counts": "YUQ42n>3M4K4M3M3L5L3M3L5_OPOjBR1S=QOkBP1S=TOiBn0U=VOhBl0U=a0N2N3N1O1N3N1O1fDPNa9S2XFRNh9o1QFXNn9j1jE]NU:d1dEcNZ:_1_EhN`:Z1YElNf:h2J6C=C=I7J6I7J6I7J6I6K5M30O10001O5K5K4L5J6K5K4L5K5K5K4K6K4L2N2N2N2N2M3N2N2N2N2N4K4M3M3L4M4L3L9H8H8G9Hi0WO2M4M2N3L3N2M4M2N2M4M2M3N3M2M3N3L3NbXm3"}, "label": "toilet in between two toilets"}]}
{"id": 215191, "image": "COCO_train2014_000000215191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle toilet\"."}], "task": "refering", "answer_template": "The \"the middle toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 149, 172, 194, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312], "bbox": [0.42978125, 0.34158333333333335, 0.581453125, 0.7977500000000001], "iscrowd": 0, "area": 10214.969750000002, "rle": {"size": [480, 640], "counts": "YUQ42n>3M4K4M3M3L5L3M3L5_OPOjBR1S=QOkBP1S=TOiBn0U=VOhBl0U=a0N2N3N1O1N3N1O1fDPNa9S2XFRNh9o1QFXNn9j1jE]NU:d1dEcNZ:_1_EhN`:Z1YElNf:h2J6C=C=I7J6I7J6I7J6I6K5M30O10001O5K5K4L5J6K5K4L5K5K5K4K6K4L2N2N2N2N2M3N2N2N2N2N4K4M3M3L4M4L3L9H8H8G9Hi0WO2M4M2N3L3N2M4M2N2M4M2M3N3M2M3N3L3NbXm3"}, "label": "the middle toilet"}]}
{"id": 215191, "image": "COCO_train2014_000000215191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet has it ' s seat up\"."}], "task": "refering", "answer_template": "The \"the toilet has it ' s seat up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 140, 141, 142, 143, 163, 164, 165, 166, 185, 186, 187, 188, 189, 209, 210, 211, 212, 230, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 281, 300, 301, 302, 303, 323, 324, 325], "bbox": [0.03378125, 0.34572916666666664, 0.26351562500000003, 0.8434583333333333], "iscrowd": 0, "area": 21659.6101, "rle": {"size": [480, 640], "counts": "Zc::b>5M4L4L4M2O2M3N1O2N2N1N3N2N1O2N1O1N2O001O1O0O2O1O1O0YDfNh9[1QFkNn9W1kEoNU:R1dETO\\:m0]EYOP9mNkGk1nN^OR9QOkGb1lNCT9UOjGR3R8WMiGi2S8`MhG_2V8gMgGY2V8nMfGR2W8TNfGl1V8[NgGe1V8bNfG^1X8gNeGY1Y8lNdGT1[8i2N2O1N2N2O1N2O1N2O1N2N2O1O100O10000O2O000O101OO010O10O01O10O0100O100O100O100O1O100O100O100O101N2N2O2M2O1N2O1N2O2M2O1N2N1gMaGQN_8l1fGQNZ8m1kGQNU8l1PHQNQ8l1THRNl7k1YHRNh7m1[HPNe7o1_HmMc7R2`HkMa7S2cHiM_7V2dHgM]7X2fHdM\\7Z2hHcMZ7[2iHaMY7]2kH`MV7]2QI]MQ7_2WI\\Mj6_2^I\\Md6d2`IWMa6i2k201N10000O10000O2O000O1000000L4I8H7I7I7F:SOm0M6K4LVcl6"}, "label": "the toilet has it ' s seat up"}]}
{"id": 215191, "image": "COCO_train2014_000000215191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toilet on the left with both lids up\"."}], "task": "refering", "answer_template": "The \"toilet on the left with both lids up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 140, 141, 142, 143, 163, 164, 165, 166, 185, 186, 187, 188, 189, 209, 210, 211, 212, 230, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 281, 300, 301, 302, 303, 323, 324, 325], "bbox": [0.03378125, 0.34572916666666664, 0.26351562500000003, 0.8434583333333333], "iscrowd": 0, "area": 21659.6101, "rle": {"size": [480, 640], "counts": "Zc::b>5M4L4L4M2O2M3N1O2N2N1N3N2N1O2N1O1N2O001O1O0O2O1O1O0YDfNh9[1QFkNn9W1kEoNU:R1dETO\\:m0]EYOP9mNkGk1nN^OR9QOkGb1lNCT9UOjGR3R8WMiGi2S8`MhG_2V8gMgGY2V8nMfGR2W8TNfGl1V8[NgGe1V8bNfG^1X8gNeGY1Y8lNdGT1[8i2N2O1N2N2O1N2O1N2O1N2N2O1O100O10000O2O000O101OO010O10O01O10O0100O100O100O100O1O100O100O100O101N2N2O2M2O1N2O1N2O2M2O1N2N1gMaGQN_8l1fGQNZ8m1kGQNU8l1PHQNQ8l1THRNl7k1YHRNh7m1[HPNe7o1_HmMc7R2`HkMa7S2cHiM_7V2dHgM]7X2fHdM\\7Z2hHcMZ7[2iHaMY7]2kH`MV7]2QI]MQ7_2WI\\Mj6_2^I\\Md6d2`IWMa6i2k201N10000O10000O2O000O1000000L4I8H7I7I7F:SOm0M6K4LVcl6"}, "label": "toilet on the left with both lids up"}]}
{"id": 215191, "image": "COCO_train2014_000000215191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toilet with the seat down\"."}], "task": "refering", "answer_template": "The \"toilet with the seat down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.71846875, 0.2613958333333333, 0.9845937499999999, 0.7380833333333333], "iscrowd": 0, "area": 21985.860499999995, "rle": {"size": [480, 640], "counts": "jhg61n>3M3N2M3M3N2N2N2N3M5K5K6J5K5K6J5K5K5K2M3N2N2N2M3N2N1O0O1000000O100O100O101N100O100O2O2M2N2N3M2N3M2N2N3M2N2O2jFZLm6h3lH_LR7d3fHcLY7^3_HjL`7W3YHPMe7S3SHTMl7m2oGXMP8j2lGXMT8i2iGZMU8i2gGYMY8h2dG[M[8f2bG\\M^8d4N101O001N110O0010O01O0010O01O010O001O010O0010O10O0100O10O01O100O1O010O1O101N1O2O1N1O2N2O1N3M3N3L3M3M4L3M4L8Hc0]O;E4L1N10000000O100000000000000000O100000000000001QNTEHl:7\\EBd:>bE[O_:e0eEWO[:h0iETOY:k0kEQOU:n0oEnNR:Q1QFlNP:R1UFkNl9S1XFiNi9U1[FiNe9U1`FgNa9W1dFfN\\9X1hFeNZ9X1X2O1N2N3M3M4M3L3L5Cm^4"}, "label": "toilet with the seat down"}]}
{"id": 215191, "image": "COCO_train2014_000000215191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet has the seat down\"."}], "task": "refering", "answer_template": "The \"the toilet has the seat down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.71846875, 0.2613958333333333, 0.9845937499999999, 0.7380833333333333], "iscrowd": 0, "area": 21985.860499999995, "rle": {"size": [480, 640], "counts": "jhg61n>3M3N2M3M3N2N2N2N3M5K5K6J5K5K6J5K5K5K2M3N2N2N2M3N2N1O0O1000000O100O100O101N100O100O2O2M2N2N3M2N3M2N2N3M2N2O2jFZLm6h3lH_LR7d3fHcLY7^3_HjL`7W3YHPMe7S3SHTMl7m2oGXMP8j2lGXMT8i2iGZMU8i2gGYMY8h2dG[M[8f2bG\\M^8d4N101O001N110O0010O01O0010O01O010O001O010O0010O10O0100O10O01O100O1O010O1O101N1O2O1N1O2N2O1N3M3N3L3M3M4L3M4L8Hc0]O;E4L1N10000000O100000000000000000O100000000000001QNTEHl:7\\EBd:>bE[O_:e0eEWO[:h0iETOY:k0kEQOU:n0oEnNR:Q1QFlNP:R1UFkNl9S1XFiNi9U1[FiNe9U1`FgNa9W1dFfN\\9X1hFeNZ9X1X2O1N2N3M3M4M3L3L5Cm^4"}, "label": "the toilet has the seat down"}]}
{"id": 239772, "image": "COCO_train2014_000000239772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black coat and holding a wine glass\"."}], "task": "refering", "answer_template": "The \"a man wearing a black coat and holding a wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 30, 31, 32, 33, 34, 35, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.13704687499999998, 0.0, 0.7921875, 0.9859016393442623], "iscrowd": 0, "area": 105753.04345000004, "rle": {"size": [427, 640], "counts": "cPU12V=4M3M3L4M3M4K4M3POSO^EP1_:ZOVEj0g:@oDc0o:m0N2N2N2N3N1N2O1O1O1O2N1O1O1O1O1O1N3N1M3L4L4L4L400000001O001O1O1O1O100O1O001O1O1O1O1O1O1O1O1O1Of0ZO1O1O001O1O1O010O1O1O001O1O1O1O1O1O1O1TMhJcMY5[2hJeMY5Z2gJfMZ5X2gJhMZ5]MgJ^40UNZ5WMkJd4LTNc5k1^JUNb5j1_JVNa5j1_JVNa5i1aJVN_5i1bJXN]5h1cJXN]5g1dJYN\\5f1fJYNZ5g1fJYNZ5f1gJZNY5e1hJ[NX5e1iJZNW5e1jJ[NV5d1kJ\\NU5c1lJ]NT5c1lJ]NT5b1nJ]NR5b1oJ^NQ5b1oJ^NQ5a1PK_NP5`1RK_Nn4a1RK_Nn4`1SK`Nm4_1TKaNl4_1UK`Nk4_1VKbNi4]1XKcNh4]1XKcNh4\\1ZKcNf4\\1[KdNe4\\1[KdNe4[1\\KeNd4Z1]KfNc4Z1^KeNb4hM]JT3R1TOa4hMaJo2o0YO`4hMdJk2m0]O_4hMgJg2l0@]4iMjJc2j0D\\4hMnJ`2g0H[4hMQK\\2e0LZ4hMTKY2c0NY4iMWKU2a02X4iMZKQ2?6W4iM]Km1=:V4iM`Ki1<=T4jMcKe1:b0R4iMfKb19e0Q4iMhK_18h0P4iMjK\\17k0o3iMlKY17m0m3jMnKV16P1l3jMPLS15S1k3jMQLQ15T1k3kMRLo04T1k3mMSLl03V1k3nMTLi02X1k3oMULf01[1j3oMVL1RN7P2g1i3QNaL7Gg1i3RNaL5Gh1i3SN`L4Hh1i3TN`L2Ih1h3VN_L1Jh1h3WN_LOJj1g3WN_LNKj1g3XN_LLKk1g3YN^LLLi1g3[N]LKOg1e3^N]LI1f1c3aN\\LH4e1`3cNSO[1n0eNTOX1m0hNUOU1l0kNVOR1k0nNVOP1k0POWOm0j0SOXOj0i0VOYOg0h0YOYOe0h0[OZOb0g0^O[O`0f0_O\\O?d0A\\O?d0A\\O?d0A\\O`0c0@]O`0c0@]O`0c0@]O`0c0@]O`0c0@]O`0c0@^O?b0A^O`0a0@_O`0a0_O@a0`0_O@a0`0_O@a0`0_O@a0`0_O@a0`0_O@b0?^OAb0?^OBa0>_OBa0>_OBa0>_OBa0>_OBa0>_OBb0=^OCb0=^OCb0=^OCb0=^OCb0=^ODa0<_ODb0;^OEb0;^OEb0;^OEb0;^OEb0;^OEb0;^OEb0;^OEc0:]OFc0:]OGb09^OGb09^OGb09^OGb09^OGb09^OGc08]OHc08]OHc08]OHc08]OIb07^OIb07^OIc06]OJc06]OJc06]OJc06]OJc06]OJc06]OJc06]OJc06]OKb05^OKb05^OKb05^OKa06_OJa06_OJa06_OJa06_OJa06_OJ4c0L^6O100000000000000000000000000000000000000000QD0c:0\\E7^:I`E?Z:AeEd0W:\\OgEg0X:YOgEj0W:VOgEm0X:TOfEo0X:SOdEP1_6UOXLNUMP1]6WO]LKRMQ1\\6YOaLIoLo0\\6]OdLFlLP1[6_OhLDiLo0Z6BlLX1P3lNoLU1l2POSMQ1h2TOWMl0e2YOZMh0d2YO\\Mh0b2YO^Mh0`2YO`Mh0^2YObMg0\\2[OdMf0Z2[OfMf0X2[OhMf0V2[OjMf0T2[OlMe0S2\\OmMe0P2]OPNd0n1]ORNd0l1]OTNd0j1]OVNc0i1^OWNc0g1^OYNc0d1_O\\Nb0b1_O^Nb0`1_O`Nb0^1_ObNa0]1@cNa0[1_OfNb0W1_OjNb0T1^OmNb0R1^OoNc0P1\\OQOd0o0\\OQOd0o0[OROf0m0YOTOg0l0XOUOi0j0VOWOj0j0TOWOm0h0SOXOm0h0ROYOn0g0PO[OQ1d0mN^OS1b0kN@V1`0fNCZ1=dNE]1:`NI`17^NKb15\\NMe12YN0g10YN0h10WN0i10WN0j1OUN2k1NUN2l1MTN3l1MTN3l1NSN2n1MQN4o1LQN4P2KPN5P2KPN5Q2JnM7R2JmM6S2JmM6T2IlM7T2IkM8V2GjM9V2GjM:V2EiM<W2EhM;X2EhM;Y2DgM<Y2DfM=[2BeM>[2BeM>\\2BcM>]2BbM?^2AbM?_2@aM`0`2_O`Ma0`2_O_Mb0b2^O]Mb0d2]O\\Mc0e2\\OZMe0f2[OZMe0g2ZOYMf0h2YOXMg0i2XOVMi0j2XOUMh0l2WOTMi0P3SOPMm0T3oNkLR1V3mNkLR1V3mNkLR1W3mNhLS1Y3lNhLS1Y3lNgLT1Z3kNgLT1Z3kNfLU1\\3iNeLV1\\3jNcLV1^3iNcLV1^3iNcLV1_3hNaLX1`3gNaLX1`3gN`LY1a3fN`LY1a3gN_LX1c3fN^LY1c3fN]LZ1d3eN]LZ1d3eN]LZ1d3eN]LZ1g3cNYL\\1k3`NVL_1m3^NTLa1P4[NQLd1S4XNmKh1V4UNkKj1Y4SNgKl1]4PNdKo1_4nMbKQ2b4kM_KT2e4hM\\KW2g4fMZKY2j4cMVK]2n4`MRK_2Q5^MPKa2T5[MmJd2W5XMjJg2Z5UMgJj2\\5SMeJl2_5QM`Jo2d5mL]JR3f5kL[JT3i5iLWJV3m5gLSJX3P6fLoIZ3R8O1O1O1O001O1O1O1O001O1O_2aMURg1"}, "label": "a man wearing a black coat and holding a wine glass"}]}
{"id": 239772, "image": "COCO_train2014_000000239772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a glass of wine while someone else poors it\"."}], "task": "refering", "answer_template": "The \"a man holding a glass of wine while someone else poors it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 30, 31, 32, 33, 34, 35, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.13704687499999998, 0.0, 0.7921875, 0.9859016393442623], "iscrowd": 0, "area": 105753.04345000004, "rle": {"size": [427, 640], "counts": "cPU12V=4M3M3L4M3M4K4M3POSO^EP1_:ZOVEj0g:@oDc0o:m0N2N2N2N3N1N2O1O1O1O2N1O1O1O1O1O1N3N1M3L4L4L4L400000001O001O1O1O1O100O1O001O1O1O1O1O1O1O1O1O1Of0ZO1O1O001O1O1O010O1O1O001O1O1O1O1O1O1O1TMhJcMY5[2hJeMY5Z2gJfMZ5X2gJhMZ5]MgJ^40UNZ5WMkJd4LTNc5k1^JUNb5j1_JVNa5j1_JVNa5i1aJVN_5i1bJXN]5h1cJXN]5g1dJYN\\5f1fJYNZ5g1fJYNZ5f1gJZNY5e1hJ[NX5e1iJZNW5e1jJ[NV5d1kJ\\NU5c1lJ]NT5c1lJ]NT5b1nJ]NR5b1oJ^NQ5b1oJ^NQ5a1PK_NP5`1RK_Nn4a1RK_Nn4`1SK`Nm4_1TKaNl4_1UK`Nk4_1VKbNi4]1XKcNh4]1XKcNh4\\1ZKcNf4\\1[KdNe4\\1[KdNe4[1\\KeNd4Z1]KfNc4Z1^KeNb4hM]JT3R1TOa4hMaJo2o0YO`4hMdJk2m0]O_4hMgJg2l0@]4iMjJc2j0D\\4hMnJ`2g0H[4hMQK\\2e0LZ4hMTKY2c0NY4iMWKU2a02X4iMZKQ2?6W4iM]Km1=:V4iM`Ki1<=T4jMcKe1:b0R4iMfKb19e0Q4iMhK_18h0P4iMjK\\17k0o3iMlKY17m0m3jMnKV16P1l3jMPLS15S1k3jMQLQ15T1k3kMRLo04T1k3mMSLl03V1k3nMTLi02X1k3oMULf01[1j3oMVL1RN7P2g1i3QNaL7Gg1i3RNaL5Gh1i3SN`L4Hh1i3TN`L2Ih1h3VN_L1Jh1h3WN_LOJj1g3WN_LNKj1g3XN_LLKk1g3YN^LLLi1g3[N]LKOg1e3^N]LI1f1c3aN\\LH4e1`3cNSO[1n0eNTOX1m0hNUOU1l0kNVOR1k0nNVOP1k0POWOm0j0SOXOj0i0VOYOg0h0YOYOe0h0[OZOb0g0^O[O`0f0_O\\O?d0A\\O?d0A\\O?d0A\\O`0c0@]O`0c0@]O`0c0@]O`0c0@]O`0c0@]O`0c0@^O?b0A^O`0a0@_O`0a0_O@a0`0_O@a0`0_O@a0`0_O@a0`0_O@a0`0_O@b0?^OAb0?^OBa0>_OBa0>_OBa0>_OBa0>_OBa0>_OBb0=^OCb0=^OCb0=^OCb0=^OCb0=^ODa0<_ODb0;^OEb0;^OEb0;^OEb0;^OEb0;^OEb0;^OEb0;^OEc0:]OFc0:]OGb09^OGb09^OGb09^OGb09^OGb09^OGc08]OHc08]OHc08]OHc08]OIb07^OIb07^OIc06]OJc06]OJc06]OJc06]OJc06]OJc06]OJc06]OJc06]OKb05^OKb05^OKb05^OKa06_OJa06_OJa06_OJa06_OJa06_OJ4c0L^6O100000000000000000000000000000000000000000QD0c:0\\E7^:I`E?Z:AeEd0W:\\OgEg0X:YOgEj0W:VOgEm0X:TOfEo0X:SOdEP1_6UOXLNUMP1]6WO]LKRMQ1\\6YOaLIoLo0\\6]OdLFlLP1[6_OhLDiLo0Z6BlLX1P3lNoLU1l2POSMQ1h2TOWMl0e2YOZMh0d2YO\\Mh0b2YO^Mh0`2YO`Mh0^2YObMg0\\2[OdMf0Z2[OfMf0X2[OhMf0V2[OjMf0T2[OlMe0S2\\OmMe0P2]OPNd0n1]ORNd0l1]OTNd0j1]OVNc0i1^OWNc0g1^OYNc0d1_O\\Nb0b1_O^Nb0`1_O`Nb0^1_ObNa0]1@cNa0[1_OfNb0W1_OjNb0T1^OmNb0R1^OoNc0P1\\OQOd0o0\\OQOd0o0[OROf0m0YOTOg0l0XOUOi0j0VOWOj0j0TOWOm0h0SOXOm0h0ROYOn0g0PO[OQ1d0mN^OS1b0kN@V1`0fNCZ1=dNE]1:`NI`17^NKb15\\NMe12YN0g10YN0h10WN0i10WN0j1OUN2k1NUN2l1MTN3l1MTN3l1NSN2n1MQN4o1LQN4P2KPN5P2KPN5Q2JnM7R2JmM6S2JmM6T2IlM7T2IkM8V2GjM9V2GjM:V2EiM<W2EhM;X2EhM;Y2DgM<Y2DfM=[2BeM>[2BeM>\\2BcM>]2BbM?^2AbM?_2@aM`0`2_O`Ma0`2_O_Mb0b2^O]Mb0d2]O\\Mc0e2\\OZMe0f2[OZMe0g2ZOYMf0h2YOXMg0i2XOVMi0j2XOUMh0l2WOTMi0P3SOPMm0T3oNkLR1V3mNkLR1V3mNkLR1W3mNhLS1Y3lNhLS1Y3lNgLT1Z3kNgLT1Z3kNfLU1\\3iNeLV1\\3jNcLV1^3iNcLV1^3iNcLV1_3hNaLX1`3gNaLX1`3gN`LY1a3fN`LY1a3gN_LX1c3fN^LY1c3fN]LZ1d3eN]LZ1d3eN]LZ1d3eN]LZ1g3cNYL\\1k3`NVL_1m3^NTLa1P4[NQLd1S4XNmKh1V4UNkKj1Y4SNgKl1]4PNdKo1_4nMbKQ2b4kM_KT2e4hM\\KW2g4fMZKY2j4cMVK]2n4`MRK_2Q5^MPKa2T5[MmJd2W5XMjJg2Z5UMgJj2\\5SMeJl2_5QM`Jo2d5mL]JR3f5kL[JT3i5iLWJV3m5gLSJX3P6fLoIZ3R8O1O1O1O001O1O1O1O001O1O_2aMURg1"}, "label": "a man holding a glass of wine while someone else poors it"}]}
{"id": 239772, "image": "COCO_train2014_000000239772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady holding a wine bottle shareing with others\"."}], "task": "refering", "answer_template": "The \"lady holding a wine bottle shareing with others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 92, 93, 94, 115, 116, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327], "bbox": [0.001859375, 0.0018032786885245902, 0.23084375000000001, 1.0], "iscrowd": 0, "area": 30294.15809999999, "rle": {"size": [427, 640], "counts": "\\=S5X8NWJi0bMVOW2R1iMmNU2V1kMjNU2V1kMiNV2W1jMiNV2W1jMhNW2X1iMhNW2X1iMgNX2Y1hMgNY2X1gMgNZ2Y1fMfN[2Z1eMfN[2Z1eMeN\\2[1dMeN\\2[1dMeN\\2[1dMdN]2\\1cMdN]2\\1cMdN]2\\1cMdN]2\\1cMcN^2]1bMcN_2\\1aMdN_2\\1aMdN_2\\1aMdN_2\\1aMcN`2]1`McN`2]1`MbNa2^1_MbNa2^1_MaNc2^1]MbNc2^1]MbNc2^1]MaNe2^1[MbNe2^1[MaNg2^1YMbNg2^1YMaNh2_1XMaNi2^1WMbNi2^1WMaNj2_1VMaNk2^1UMaNl2_1TMaNl2_1TM`Nm2`1SM`Nn2_1RM`No2`1QM`No2`1QM_NP3a1PM]NR3c1nL\\NS3d1nLYNT3g1mLWNT3i1oLRNS3n1oLoMR3Q2PMgMV3Y2lL]M\\3c2eL\\M[3d2fL[MZ3e2fL[MZ3e2gLZMY3f2hLYMX3g2iLXMW3h2jLWMV3i2kLVMU3j2lLUMT3k2mL^LBTNa3^5mL[LHTN[3a5nLVLNRNX3h5jLQLT4o3lKlKY4T4hKjKY4V4gKjKY4V4gKiKZ4W4gKhKY4X4gKgKZ4Y4gKfKY4Z4hKdKY4\\4iKbKW4^4jK`KW4`4kK]KV4c4kKZKV4g4lKUKV4k4kKSKV4U1iJ`1S1WMV4l0YKh1c0XMV4k0aKh1=XMU4i0YLPO^Ne1X1[NW46]M`1`NRNY45]Mj1^4mMhKS2g71O1O1N2L4O1O1O1O1O1001O2M1N3L5L3N3N00O1O001000000000000O010000O10000O010O101N4M2N2M2O1O0O2O0O2N1O2N2N3M>B2N;D4K5Knh\\6"}, "label": "lady holding a wine bottle shareing with others"}]}
{"id": 239772, "image": "COCO_train2014_000000239772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blond woman wears glasses on her head\"."}], "task": "refering", "answer_template": "The \"a blond woman wears glasses on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 92, 93, 94, 115, 116, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327], "bbox": [0.001859375, 0.0018032786885245902, 0.23084375000000001, 1.0], "iscrowd": 0, "area": 30294.15809999999, "rle": {"size": [427, 640], "counts": "\\=S5X8NWJi0bMVOW2R1iMmNU2V1kMjNU2V1kMiNV2W1jMiNV2W1jMhNW2X1iMhNW2X1iMgNX2Y1hMgNY2X1gMgNZ2Y1fMfN[2Z1eMfN[2Z1eMeN\\2[1dMeN\\2[1dMeN\\2[1dMdN]2\\1cMdN]2\\1cMdN]2\\1cMdN]2\\1cMcN^2]1bMcN_2\\1aMdN_2\\1aMdN_2\\1aMdN_2\\1aMcN`2]1`McN`2]1`MbNa2^1_MbNa2^1_MaNc2^1]MbNc2^1]MbNc2^1]MaNe2^1[MbNe2^1[MaNg2^1YMbNg2^1YMaNh2_1XMaNi2^1WMbNi2^1WMaNj2_1VMaNk2^1UMaNl2_1TMaNl2_1TM`Nm2`1SM`Nn2_1RM`No2`1QM`No2`1QM_NP3a1PM]NR3c1nL\\NS3d1nLYNT3g1mLWNT3i1oLRNS3n1oLoMR3Q2PMgMV3Y2lL]M\\3c2eL\\M[3d2fL[MZ3e2fL[MZ3e2gLZMY3f2hLYMX3g2iLXMW3h2jLWMV3i2kLVMU3j2lLUMT3k2mL^LBTNa3^5mL[LHTN[3a5nLVLNRNX3h5jLQLT4o3lKlKY4T4hKjKY4V4gKjKY4V4gKiKZ4W4gKhKY4X4gKgKZ4Y4gKfKY4Z4hKdKY4\\4iKbKW4^4jK`KW4`4kK]KV4c4kKZKV4g4lKUKV4k4kKSKV4U1iJ`1S1WMV4l0YKh1c0XMV4k0aKh1=XMU4i0YLPO^Ne1X1[NW46]M`1`NRNY45]Mj1^4mMhKS2g71O1O1N2L4O1O1O1O1O1001O2M1N3L5L3N3N00O1O001000000000000O010000O10000O010O101N4M2N2M2O1O0O2O0O2N1O2N2N3M>B2N;D4K5Knh\\6"}, "label": "a blond woman wears glasses on her head"}]}
{"id": 215201, "image": "COCO_train2014_000000215201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry grey car\"."}], "task": "refering", "answer_template": "The \"a blurry grey car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203], "bbox": [0.463796875, 0.262018779342723, 1.0, 0.5716901408450704], "iscrowd": 0, "area": 31782.035249999994, "rle": {"size": [426, 640], "counts": "iek31X=2O1N3N1N2O1N2O2M2O1N2O1N3N1N2O1N3M2O1N2O1N3N1N3N2N1O2N1O2N1O2N2N1O2N;E8H7I1O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O00000O1000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000O1000000000001O00000000000000001O0000001O1O001O1O001O1O1O001O1O001O1O001O1O1O001OO1000000000000000000O1000000000000000001O0O101O00001O00001O001N10001O00001O00001O000O2O00001O00001O00001N10001O00001N101N1O101N100O2N100O2N100O2N100O2N100O2O0O1O2O0O1O2O0O101N101O000O2O000O2O00001N10001O0O101O0O101O000O2O00001N10001N101O000O2O00001N10001N10001O0O101O000O2O000O2O00001N101O000O2O000O2O00007H;F:F:FmI"}, "label": "a blurry grey car"}]}
{"id": 215201, "image": "COCO_train2014_000000215201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver car that is blurred\"."}], "task": "refering", "answer_template": "The \"the silver car that is blurred\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203], "bbox": [0.463796875, 0.262018779342723, 1.0, 0.5716901408450704], "iscrowd": 0, "area": 31782.035249999994, "rle": {"size": [426, 640], "counts": "iek31X=2O1N3N1N2O1N2O2M2O1N2O1N3N1N2O1N3M2O1N2O1N3N1N3N2N1O2N1O2N1O2N2N1O2N;E8H7I1O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O00000O1000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000O1000000000001O00000000000000001O0000001O1O001O1O001O1O1O001O1O001O1O001O1O1O001OO1000000000000000000O1000000000000000001O0O101O00001O00001O001N10001O00001O00001O000O2O00001O00001O00001N10001O00001N101N1O101N100O2N100O2N100O2N100O2N100O2O0O1O2O0O1O2O0O101N101O000O2O000O2O00001N10001O0O101O0O101O000O2O00001N10001N101O000O2O00001N10001N10001O0O101O000O2O000O2O00001N101O000O2O000O2O00007H;F:F:FmI"}, "label": "the silver car that is blurred"}]}
{"id": 559267, "image": "COCO_train2014_000000559267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"kid under the umbrella wearing white pant and rose shoe\"."}], "task": "refering", "answer_template": "The \"kid under the umbrella wearing white pant and rose shoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 144, 145, 146, 147, 148, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 229, 230, 231], "bbox": [0.4280416666666667, 0.32823437499999997, 0.8190416666666667, 0.6017812499999999], "iscrowd": 0, "area": 17793.7553, "rle": {"size": [640, 480], "counts": "R^P41hc0;I6J4_O`0O2N3M3L2O1N3N1N2N2O1N2O1Ne0\\O8G4L4M3L4M3L5K4M3L2O1N2O1O0O2O1O1O1N2O1O1N2O1O1O10000O10000O10000OUOc@UM]?k2i@oLW?R3i@mLV?T3k@kLU?V3k@jLf>GVA_34lLa>f3_A[L\\>h3dARLFIa>Y4iAmKJG\\>^4iAkK_>U4aAkK_>V4`AjKa>U4_AkKa>V4^AiKc>W4`AfK`>Z4c01O010O00001O01O0001O000010O000001O00010O00000001O000000000O2O00000000001O00000O1O2N100O1O1O101N1O1O1O1L4L5B=00000001O00000001O01O0001OdMj_On0V`0QOl_On0T`0QOo_OROKOHY1_`0Dc@;]?Ed@;[?Dh@YOjNg0^`0Oj@WOlNh0Z`00XANi>0[AMe>2^AMa>1cAM]>2fANX>1kANT>1PBMP>1SBOk=0XBOg=AkB>]`0N2N100O1O1O100O1O0001O000001O0000000000001OO10000O2O001O001O001O01O00003LYQf1"}, "label": "kid under the umbrella wearing white pant and rose shoe"}]}
{"id": 559267, "image": "COCO_train2014_000000559267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl sitting under umbrella wearing white pant and pink shoes\"."}], "task": "refering", "answer_template": "The \"a girl sitting under umbrella wearing white pant and pink shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 144, 145, 146, 147, 148, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 229, 230, 231], "bbox": [0.4280416666666667, 0.32823437499999997, 0.8190416666666667, 0.6017812499999999], "iscrowd": 0, "area": 17793.7553, "rle": {"size": [640, 480], "counts": "R^P41hc0;I6J4_O`0O2N3M3L2O1N3N1N2N2O1N2O1Ne0\\O8G4L4M3L4M3L5K4M3L2O1N2O1O0O2O1O1O1N2O1O1N2O1O1O10000O10000O10000OUOc@UM]?k2i@oLW?R3i@mLV?T3k@kLU?V3k@jLf>GVA_34lLa>f3_A[L\\>h3dARLFIa>Y4iAmKJG\\>^4iAkK_>U4aAkK_>V4`AjKa>U4_AkKa>V4^AiKc>W4`AfK`>Z4c01O010O00001O01O0001O000010O000001O00010O00000001O000000000O2O00000000001O00000O1O2N100O1O1O101N1O1O1O1L4L5B=00000001O00000001O01O0001OdMj_On0V`0QOl_On0T`0QOo_OROKOHY1_`0Dc@;]?Ed@;[?Dh@YOjNg0^`0Oj@WOlNh0Z`00XANi>0[AMe>2^AMa>1cAM]>2fANX>1kANT>1PBMP>1SBOk=0XBOg=AkB>]`0N2N100O1O1O100O1O0001O000001O0000000000001OO10000O2O001O001O001O01O00003LYQf1"}, "label": "a girl sitting under umbrella wearing white pant and pink shoes"}]}
{"id": 559267, "image": "COCO_train2014_000000559267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl on the left side of the umbrella wearing light green socks\"."}], "task": "refering", "answer_template": "The \"little girl on the left side of the umbrella wearing light green socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 207, 208, 210, 211], "bbox": [0.07747916666666667, 0.29221875, 0.5507291666666667, 0.539390625], "iscrowd": 0, "area": 20389.591299999996, "rle": {"size": [640, 480], "counts": "S[g02kc07H8J6K5J4M3M3M2N3L4N2N2N2N2M3N1O2N2N1O2M3N1O2N1O2O1N1O2O1N1O2O0O2O1N100000000O10000M3E;]Oc000O1O1O100O1O1O1O2N1e@QLc>P4[AQLe>P4XARLh>U4o@nKP?S4n@nKS?Q4k@QLU?]401O000000000001O000001O0000000001O0L4L4L4L4A[@_Li?a3Y@ZLj?f3801N1000000O100000001O0000000000000000001O00000000000000001O000O1000000000001N100O10000O100O101N100O10000O100O1011h_OiLc?Z3U@mLi?U3Q@PMn?b3O1O1O2N1O1O100O2N1O1O1O2NbMg@Am?UOaA?g>VOhA`0_>UOlAe0i`0L5K]]O3\\a0Ib^O>^a0^O^^Oj0aa0SO\\^OT1Qb0J^]O\\O\\b0c0m]O[Ona0f0W^OYOda0i0`^OVOb`0M`_Oa0@Bc0OV`04b_O?g0\\Oa?:d_O>n0VOX?a0g_O;T1SOP?g0g_O:\\1mNg>V2\\14O21N3N1N2OK42O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1O101N1000000O101O0O1000000O2^OX]OJhb04c0N2N`iV4"}, "label": "little girl on the left side of the umbrella wearing light green socks"}]}
{"id": 559267, "image": "COCO_train2014_000000559267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child crouched under an umbrella in the grass , wearing a green sleeved top and green socks\"."}], "task": "refering", "answer_template": "The \"a young child crouched under an umbrella in the grass , wearing a green sleeved top and green socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 207, 208, 210, 211], "bbox": [0.07747916666666667, 0.29221875, 0.5507291666666667, 0.539390625], "iscrowd": 0, "area": 20389.591299999996, "rle": {"size": [640, 480], "counts": "S[g02kc07H8J6K5J4M3M3M2N3L4N2N2N2N2M3N1O2N2N1O2M3N1O2N1O2O1N1O2O1N1O2O0O2O1N100000000O10000M3E;]Oc000O1O1O100O1O1O1O2N1e@QLc>P4[AQLe>P4XARLh>U4o@nKP?S4n@nKS?Q4k@QLU?]401O000000000001O000001O0000000001O0L4L4L4L4A[@_Li?a3Y@ZLj?f3801N1000000O100000001O0000000000000000001O00000000000000001O000O1000000000001N100O10000O100O101N100O10000O100O1011h_OiLc?Z3U@mLi?U3Q@PMn?b3O1O1O2N1O1O100O2N1O1O1O2NbMg@Am?UOaA?g>VOhA`0_>UOlAe0i`0L5K]]O3\\a0Ib^O>^a0^O^^Oj0aa0SO\\^OT1Qb0J^]O\\O\\b0c0m]O[Ona0f0W^OYOda0i0`^OVOb`0M`_Oa0@Bc0OV`04b_O?g0\\Oa?:d_O>n0VOX?a0g_O;T1SOP?g0g_O:\\1mNg>V2\\14O21N3N1N2OK42O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1O101N1000000O101O0O1000000O2^OX]OJhb04c0N2N`iV4"}, "label": "a young child crouched under an umbrella in the grass , wearing a green sleeved top and green socks"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a round cake with a hole in the center\"."}], "task": "refering", "answer_template": "The \"a round cake with a hole in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 191, 192], "bbox": [0.12115625000000001, 0.2535680751173709, 0.48014062500000004, 0.55018779342723], "iscrowd": 0, "area": 22028.685199999996, "rle": {"size": [426, 640], "counts": "_cP14R=6UCJZ<;dCIU<<hCIT<j0L4L5K3L3N2N1O2N2N2N1O2N2N2N1O2N2O0O2N2N2O0O2N2N2O0O2N2N2O0O2N2O0O2O1O1N10001N10000O101O000O101O0O1000O01000000O1000000O10O1000O1000000O1000000O100000O01000000O1000000000000000000001O0000000O100000000000000000010O0001O001O00001O00001O100O001O1O1O1O1O010O00000000000001O0000000O10001N10000O10000O101O0O10000O10000O2O000O10001O0O10000O2O000O10001N10000O101O0O10001N101O0O2O001N2O0O2O0O2N2O0O2N101N1O2O1N2O2M2N2O1N3M2O1N2N2O2M2N2O1N3N1N3M4M3L4L4M3L4G;Ff[Z4"}, "label": "a round cake with a hole in the center"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plate of cake in table\"."}], "task": "refering", "answer_template": "The \"plate of cake in table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 191, 192], "bbox": [0.12115625000000001, 0.2535680751173709, 0.48014062500000004, 0.55018779342723], "iscrowd": 0, "area": 22028.685199999996, "rle": {"size": [426, 640], "counts": "_cP14R=6UCJZ<;dCIU<<hCIT<j0L4L5K3L3N2N1O2N2N2N1O2N2N2N1O2N2O0O2N2N2O0O2N2N2O0O2N2N2O0O2N2O0O2O1O1N10001N10000O101O000O101O0O1000O01000000O1000000O10O1000O1000000O1000000O100000O01000000O1000000000000000000001O0000000O100000000000000000010O0001O001O00001O00001O100O001O1O1O1O1O010O00000000000001O0000000O10001N10000O10000O101O0O10000O10000O2O000O10001O0O10000O2O000O10001N10000O101O0O10001N101O0O2O001N2O0O2O0O2N2O0O2N101N1O2O1N2O2M2N2O1N3M2O1N2N2O2M2N2O1N3N1N3M4M3L4L4M3L4G;Ff[Z4"}, "label": "plate of cake in table"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bowl of chips between the bread and pie but further away from bottles\"."}], "task": "refering", "answer_template": "The \"white bowl of chips between the bread and pie but further away from bottles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.0450625, 0.5195305164319248, 0.3726875, 0.8632863849765258], "iscrowd": 0, "area": 23621.412249999998, "rle": {"size": [426, 640], "counts": "_[<3U=4K5L4L5K4L4L4K5L5K4L4L4L4K6K4L4L4L4L5L3M3N1N2O0O2O0O101N100O2O0O2O0O101N100O2O0O2O0O101N101N100O2O0O101N101N100O2O0O2O0O101N100O2O0O10O0100O10000O100O010O100O100O100O01000O1000000000O010000000000O10O10000000O100000O1000O1000000000O100000000000000O100000000000000O10000000000000O010001O0000001O0O101O0000001O0O101N100O101N100O2N100O101N100O2O0O101N1O100O2O1N101N2O0O2O1N2N101N2O0N3M3M2O2M3M2N3M3M2O2M3M4L6J6J6K4K6J6J5KhaV5"}, "label": "white bowl of chips between the bread and pie but further away from bottles"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate of crackers to the left of a pie\"."}], "task": "refering", "answer_template": "The \"a plate of crackers to the left of a pie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.0450625, 0.5195305164319248, 0.3726875, 0.8632863849765258], "iscrowd": 0, "area": 23621.412249999998, "rle": {"size": [426, 640], "counts": "_[<3U=4K5L4L5K4L4L4K5L5K4L4L4L4K6K4L4L4L4L5L3M3N1N2O0O2O0O101N100O2O0O2O0O101N100O2O0O2O0O101N101N100O2O0O101N101N100O2O0O2O0O101N100O2O0O10O0100O10000O100O010O100O100O100O01000O1000000000O010000000000O10O10000000O100000O1000O1000000000O100000000000000O100000000000000O10000000000000O010001O0000001O0O101O0000001O0O101N100O101N100O2N100O101N100O2O0O101N1O100O2O1N101N2O0O2O1N2N101N2O0N3M3M2O2M3M2N3M3M2O2M3M4L6J6J6K4K6J6J5KhaV5"}, "label": "a plate of crackers to the left of a pie"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza in table ready to serve\"."}], "task": "refering", "answer_template": "The \"pizza in table ready to serve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 297, 298, 310, 311, 312, 313, 314, 315, 336, 337, 338], "bbox": [0.387890625, 0.4306572769953052, 0.9933125, 0.9655399061032865], "iscrowd": 0, "area": 51251.46734999999, "rle": {"size": [426, 640], "counts": "]^W32P=:F:G8G:F:G8L5M3M2O2M3M2N3N2M2N3N2M2N3M3N1N3M3M2O2M2N3M3N1N3M2O0O2N2N2O1N1O2N2O1N2N101N2N2N2O0O2O1N2O1O0O2O1N2O1O0O2O1N2O1N101O1N2O1N101O1N2O1N2O001N2O1N2O0O2N2O1N2N101N2N2O0O1O2O0O1O101N10010O0000001O00001O0000001O00001O0000001O01O01O0000001O01O01O0001O01O000010O000001O01O01O0000010O00000010O0001O0001O01O00001O0000001O^JZI`4f6`KZI`4f6_K[Ib4d6]K]Ic4c6]K]Ic4b6]K_Ic4a6]K_Id4`6[KaIe4_6ZKbIf4^6ZKbIg4]6XKdIh4\\6XKdIh4\\6WKeIi4[6VKfIk4Y6UKgIk4Y6TKhIl4X6TKhIl0G[2`6hLjIn0IV2]6kLkIo0MQ2X6PMkIo02l1R6UMmIP15g1m5YMnIP1:b1h5]MoIR1=\\1d5aMPJS1`0Y1_5dMQJS1e0T1Z5iMQJT1i0n0V5nMQJT1n0i0Q5SNQJT1S1e0k4WNRJU1W1?g4\\NRJU1\\1:b4`NSJW1^15_4dNSJW1c10Z4iNSJW1h1JV4oNRJX1l1DR4TORJX1Q2_Om3YORJY1U2XOj3_OQJY1Y2TOf3CQJY1^2oNa3HQJZ1b2hN^3NPJZ1g2cNY33PJZ1l2^NS39QJZ1P3WNP3?PJZ1T3SNl2b0QJ\\1d9dN\\F\\1d9dN\\F\\1d9dN\\F]1c9cN]F]1c9cN]F^1b9bN^F^1b9bN^F^1b9bN^F_1a9aN_F_1a9aN_F`1`9`N`F`1`9_NaFa1_9_NaFb1^9^NbFb1_9]NaFc1_9]NaFd1^9\\NbFd1^9\\NbFe1]9[NcFe1]9[NcFe1]9[NcFf1]9YNcFg1]9YNcFg1]9YNcFh1\\9WNeFi1[9WNeFi1[9WNeFj1Z9VNfFj1Z9VNfFk1Z9TNfFl1Z9TNfFl1Z9TNfFm1Y9SNgFm1Y9SNgFm1Y9SNgFn1Y9QNgFo1Y9QNfFQ2Y9oMgFQ2Z9nMfFR2Z9nMfFS2Z9lMfFT2Z9lMfFT2Z9lMfFU2Z9jMfFV2Z9jMfFV2Z9jMfFW2Z9hMfFX2Z9hMfFY2Z9fMfFZ2Z9fMfFZ2Z9fMfF[2Z9dMfF\\2Z9dMfF\\2Z9dMfF]2Z9bMfF^2Z9bMeF`2[9_MeFa2[9_MeFa2[9_MeFb2[9]MeFc2[9]MeFc2[9]MeFd2Z9\\MfFd2[9[MeFe2[9[MeFf2Z9ZMfFf2Z9ZMfFg2Y9YMgFg2Y9YMgFg2Y9YMgFh2X9XMhFh2Y9WMgFi2Y9WMgFj2X9VMhFj2X9VMhFk2W9UMiFk2W9UMiFk2W9UMiFl2V9TMjFl2W9SMiFm2W9SMiFn2V9RMjFn2V9RMiFo2W9QMiFP3V9PMjFP3V9PMjFQ3U9oLkFQ3V9nLjFR3V9nLjFS3U9mLkFS3U9mLkFS3U9mLkFT3T9lLlFT3T9lLlFU3S9kLmFU3d9001O00001O001O00001O00001O00001O001O00001O00001O001O000010O0001O01O01O0010O00010O0001O0101N1O100O1O2O0O1O100O2O0O1O100O2N100O101M2O1O1O1O2N1O1N2O1O2N1O2N2M3N2N2N2N2N2N2M3N2XOh0UOo\\1"}, "label": "pizza in table ready to serve"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pie with a black topping\"."}], "task": "refering", "answer_template": "The \"a pie with a black topping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 297, 298, 310, 311, 312, 313, 314, 315, 336, 337, 338], "bbox": [0.387890625, 0.4306572769953052, 0.9933125, 0.9655399061032865], "iscrowd": 0, "area": 51251.46734999999, "rle": {"size": [426, 640], "counts": "]^W32P=:F:G8G:F:G8L5M3M2O2M3M2N3N2M2N3N2M2N3M3N1N3M3M2O2M2N3M3N1N3M2O0O2N2N2O1N1O2N2O1N2N101N2N2N2O0O2O1N2O1O0O2O1N2O1O0O2O1N2O1N101O1N2O1N101O1N2O1N2O001N2O1N2O0O2N2O1N2N101N2N2O0O1O2O0O1O101N10010O0000001O00001O0000001O00001O0000001O01O01O0000001O01O01O0001O01O000010O000001O01O01O0000010O00000010O0001O0001O01O00001O0000001O^JZI`4f6`KZI`4f6_K[Ib4d6]K]Ic4c6]K]Ic4b6]K_Ic4a6]K_Id4`6[KaIe4_6ZKbIf4^6ZKbIg4]6XKdIh4\\6XKdIh4\\6WKeIi4[6VKfIk4Y6UKgIk4Y6TKhIl4X6TKhIl0G[2`6hLjIn0IV2]6kLkIo0MQ2X6PMkIo02l1R6UMmIP15g1m5YMnIP1:b1h5]MoIR1=\\1d5aMPJS1`0Y1_5dMQJS1e0T1Z5iMQJT1i0n0V5nMQJT1n0i0Q5SNQJT1S1e0k4WNRJU1W1?g4\\NRJU1\\1:b4`NSJW1^15_4dNSJW1c10Z4iNSJW1h1JV4oNRJX1l1DR4TORJX1Q2_Om3YORJY1U2XOj3_OQJY1Y2TOf3CQJY1^2oNa3HQJZ1b2hN^3NPJZ1g2cNY33PJZ1l2^NS39QJZ1P3WNP3?PJZ1T3SNl2b0QJ\\1d9dN\\F\\1d9dN\\F\\1d9dN\\F]1c9cN]F]1c9cN]F^1b9bN^F^1b9bN^F^1b9bN^F_1a9aN_F_1a9aN_F`1`9`N`F`1`9_NaFa1_9_NaFb1^9^NbFb1_9]NaFc1_9]NaFd1^9\\NbFd1^9\\NbFe1]9[NcFe1]9[NcFe1]9[NcFf1]9YNcFg1]9YNcFg1]9YNcFh1\\9WNeFi1[9WNeFi1[9WNeFj1Z9VNfFj1Z9VNfFk1Z9TNfFl1Z9TNfFl1Z9TNfFm1Y9SNgFm1Y9SNgFm1Y9SNgFn1Y9QNgFo1Y9QNfFQ2Y9oMgFQ2Z9nMfFR2Z9nMfFS2Z9lMfFT2Z9lMfFT2Z9lMfFU2Z9jMfFV2Z9jMfFV2Z9jMfFW2Z9hMfFX2Z9hMfFY2Z9fMfFZ2Z9fMfFZ2Z9fMfF[2Z9dMfF\\2Z9dMfF\\2Z9dMfF]2Z9bMfF^2Z9bMeF`2[9_MeFa2[9_MeFa2[9_MeFb2[9]MeFc2[9]MeFc2[9]MeFd2Z9\\MfFd2[9[MeFe2[9[MeFf2Z9ZMfFf2Z9ZMfFg2Y9YMgFg2Y9YMgFg2Y9YMgFh2X9XMhFh2Y9WMgFi2Y9WMgFj2X9VMhFj2X9VMhFk2W9UMiFk2W9UMiFk2W9UMiFl2V9TMjFl2W9SMiFm2W9SMiFn2V9RMjFn2V9RMiFo2W9QMiFP3V9PMjFP3V9PMjFQ3U9oLkFQ3V9nLjFR3V9nLjFS3U9mLkFS3U9mLkFS3U9mLkFT3T9lLlFT3T9lLlFU3S9kLmFU3d9001O00001O001O00001O00001O00001O001O00001O00001O001O000010O0001O01O01O0010O00010O0001O0101N1O100O1O2O0O1O100O2O0O1O100O2N100O101M2O1O1O1O2N1O1N2O1O2N1O2N2M3N2N2N2N2N2N2M3N2XOh0UOo\\1"}, "label": "a pie with a black topping"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of wine rests on the table behind an elderberry pie\"."}], "task": "refering", "answer_template": "The \"a glass of wine rests on the table behind an elderberry pie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 39, 40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 131, 132, 133, 154, 155, 156], "bbox": [0.698140625, 0.011807511737089202, 0.84065625, 0.42415492957746476], "iscrowd": 0, "area": 11618.717249999992, "rle": {"size": [426, 640], "counts": "kRj55P=6I7J7H7J6I7J6J6J6J7I6J6K5J6J6J6J7L<C<E3M3M3M3L2O2N2N2N2N2M3N2N2N2N1100000000000O100000O1000000000000000O10O1000000000000000000O0100000K5K5K5K5K5K5K5K5J6K4L5K5K5K5K5K<D;E<C=D<D<D<D\\\\Z1"}, "label": "a glass of wine rests on the table behind an elderberry pie"}]}
{"id": 59556, "image": "COCO_train2014_000000059556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottle of wine that is to the right , closer to the paper towels\"."}], "task": "refering", "answer_template": "The \"the bottle of wine that is to the right , closer to the paper towels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 39, 40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 131, 132, 133, 154, 155, 156], "bbox": [0.698140625, 0.011807511737089202, 0.84065625, 0.42415492957746476], "iscrowd": 0, "area": 11618.717249999992, "rle": {"size": [426, 640], "counts": "kRj55P=6I7J7H7J6I7J6J6J6J7I6J6K5J6J6J6J7L<C<E3M3M3M3L2O2N2N2N2N2M3N2N2N2N1100000000000O100000O1000000000000000O10O1000000000000000000O0100000K5K5K5K5K5K5K5K5J6K4L5K5K5K5K5K<D;E<C=D<D<D<D\\\\Z1"}, "label": "the bottle of wine that is to the right , closer to the paper towels"}]}
{"id": 444583, "image": "COCO_train2014_000000444583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a camo hat and camo pants\"."}], "task": "refering", "answer_template": "The \"a man in a camo hat and camo pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 65, 66, 67, 68, 78, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 320, 321, 322, 323, 335, 336], "bbox": [0.20337264150943396, 0.13482812500000002, 0.6933018867924527, 0.9990937499999999], "iscrowd": 0, "area": 74056.7434, "rle": {"size": [640, 424], "counts": "`ne13lc0>iNBQ^Ol0aa0VO^^OV1[`0fNj_O5Ic1P`0[NT@2KQ2c?oM\\@MA4>T2b?lM]@70P2a?iM^@8MU2b?cMa@9JX2d?_M[@MK=3]2d?ZM[@a00Z2a?UM^@`01`2]?QMa@>3e2Y?mLc@>4e2Z?eMd@\\2]?cMc@]2]?cMb@^2_?mLa@7Om2a?hL_@1100IO_3b?eL^@80G4\\3^?dL^@;1B3`3`?`L\\@`05Q3`?RM]@o2b?SM]@m2b?i0;E1O1O001O1O0O2O1NRGYLf1e3[J^MlMQOEc0W8FTG_2m2_NRN=c8Q1UIjNVN7g8_N[Fn1`2E\\N0U9<YHL`NIT9=UH3dNCT9;RH:cNAY96nGa0VNLi9DkGd3R8]LgGk3W8VLcGR4Z8PL_GW4_8g3N1O2O0O9H>A3N1N3N2M2O000O100O100^NZDiJg;V5[DaINP1g;R5_DlIN0NP1f;]O]D\\4K^Kf0H^OP1d;WOlDX4@bKi0MYO1Lj0f;WOZEV4g0aL];c2YEhLdM;W=h2`EWMe:d2]E[Mg:b2YE\\Mk:b2UE]Mo:_2RE_MP;a2PE^MP;b2PE]MQ;d2nD\\MR;d2nD\\MR;e2mDZMT;f2lDZMT;g2kDXMV;i2iDWMW;i2hDWMY;j2fDVMZ;i2gDVMZ;j2fDVMZ;j2n20000O10000000O0100000000000000001O000UOTMY@V3e?oLV@R3i?RMT@n2l?VMo_Ok2P`0d0O000O101O0O101O0O101O1N101O1N2O001N2O1O1N1O2N2O1N1^B\\Kg;f4YD^Ka;d4_D_K\\;c4cDbKX;_4jChK_OJe<c4eChKEDe<h4aChKHAe<l4\\CgKO]Od<o4XChK2YOe<Y6YChIe<j6O2O1N1O2N1O2O1N1O2N101N2O0O2O0O2O0O100`LQHnJQ8Q5PHmJQ8R5RHlJn7S5THkJm7T5UHkJk7T5WH\\JJbMQ8P8UH]J[8\\5cGYHHV2g8Y5dGfHFn1h8W5bGPIHe1i8[6[G_Ig8a6]GZId8d6bGVI`8h6fGSI[8l6kGmHX8R7lGiHU8W7oGcHS8]7QH^HP8a7UHYHn7d7YHVHh7i7]HRHd7m7aHmGb7Q8cHjGa7R8Y201M2N3L3N3M3I7I7SOl0O2M3A?^Ob0\\Od0[Oe0[OR1nN^1bN_o`2"}, "label": "a man in a camo hat and camo pants"}]}
{"id": 444583, "image": "COCO_train2014_000000444583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in army hat and suit jacket using loud speaker\"."}], "task": "refering", "answer_template": "The \"man in army hat and suit jacket using loud speaker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 65, 66, 67, 68, 78, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 320, 321, 322, 323, 335, 336], "bbox": [0.20337264150943396, 0.13482812500000002, 0.6933018867924527, 0.9990937499999999], "iscrowd": 0, "area": 74056.7434, "rle": {"size": [640, 424], "counts": "`ne13lc0>iNBQ^Ol0aa0VO^^OV1[`0fNj_O5Ic1P`0[NT@2KQ2c?oM\\@MA4>T2b?lM]@70P2a?iM^@8MU2b?cMa@9JX2d?_M[@MK=3]2d?ZM[@a00Z2a?UM^@`01`2]?QMa@>3e2Y?mLc@>4e2Z?eMd@\\2]?cMc@]2]?cMb@^2_?mLa@7Om2a?hL_@1100IO_3b?eL^@80G4\\3^?dL^@;1B3`3`?`L\\@`05Q3`?RM]@o2b?SM]@m2b?i0;E1O1O001O1O0O2O1NRGYLf1e3[J^MlMQOEc0W8FTG_2m2_NRN=c8Q1UIjNVN7g8_N[Fn1`2E\\N0U9<YHL`NIT9=UH3dNCT9;RH:cNAY96nGa0VNLi9DkGd3R8]LgGk3W8VLcGR4Z8PL_GW4_8g3N1O2O0O9H>A3N1N3N2M2O000O100O100^NZDiJg;V5[DaINP1g;R5_DlIN0NP1f;]O]D\\4K^Kf0H^OP1d;WOlDX4@bKi0MYO1Lj0f;WOZEV4g0aL];c2YEhLdM;W=h2`EWMe:d2]E[Mg:b2YE\\Mk:b2UE]Mo:_2RE_MP;a2PE^MP;b2PE]MQ;d2nD\\MR;d2nD\\MR;e2mDZMT;f2lDZMT;g2kDXMV;i2iDWMW;i2hDWMY;j2fDVMZ;i2gDVMZ;j2fDVMZ;j2n20000O10000000O0100000000000000001O000UOTMY@V3e?oLV@R3i?RMT@n2l?VMo_Ok2P`0d0O000O101O0O101O0O101O1N101O1N2O001N2O1O1N1O2N2O1N1^B\\Kg;f4YD^Ka;d4_D_K\\;c4cDbKX;_4jChK_OJe<c4eChKEDe<h4aChKHAe<l4\\CgKO]Od<o4XChK2YOe<Y6YChIe<j6O2O1N1O2N1O2O1N1O2N101N2O0O2O0O2O0O100`LQHnJQ8Q5PHmJQ8R5RHlJn7S5THkJm7T5UHkJk7T5WH\\JJbMQ8P8UH]J[8\\5cGYHHV2g8Y5dGfHFn1h8W5bGPIHe1i8[6[G_Ig8a6]GZId8d6bGVI`8h6fGSI[8l6kGmHX8R7lGiHU8W7oGcHS8]7QH^HP8a7UHYHn7d7YHVHh7i7]HRHd7m7aHmGb7Q8cHjGa7R8Y201M2N3L3N3M3I7I7SOl0O2M3A?^Ob0\\Od0[Oe0[OR1nN^1bN_o`2"}, "label": "man in army hat and suit jacket using loud speaker"}]}
{"id": 444583, "image": "COCO_train2014_000000444583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a walking man in a dirty white tank top , black gym shorts and white sandals\"."}], "task": "refering", "answer_template": "The \"a walking man in a dirty white tank top , black gym shorts and white sandals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 101, 102, 115, 116, 117, 130, 131, 132, 145, 146, 147, 160, 161, 162, 175, 176, 177, 178, 190, 193, 194, 205, 208], "bbox": [0.6848349056603774, 0.22385937500000003, 0.9511320754716981, 0.591546875], "iscrowd": 0, "area": 9320.724699999997, "rle": {"size": [640, 424], "counts": "RSf56^4Kl:9fD0[;1ZD8g;JmC?T<BdCd0]<]OaCc0a<]O\\Cd0f<\\OXCe0h<\\OVCd0l<[ORCf0P=ZOnBg0S=YOkBg0W=YORB3WNf0h?WOkAa1V>`NdAe1]>ZN^Aj1d>VNVAm1m>SNm@o1W?QNe@P2^?PN]@e1\\OWN[`04U@h1DnMZ`09n_Ol1]`0WN\\_Om1e`0f0M2VN\\MUBj2g=[MTBh2m=YMPBi2P>YMmAg2U>YMiAh2W>YMfAi2[>XMbAj2]>WMaAk2_>UM^Am2b>TM[Ao2e>V1O10O10001O0001O2N1O2N1O1O0003M2N3YKWAQ4l>gKZAX4X?O1O2N0101N100O2O0O10mLWAn0j>kNiAi0Y>mN[Bf0g=QOeBj0^=mNjBR1W=eNRCY1m?N3N1N2N10O01O1O010O00100Ob]OUOfa0k0V^O]Oga0c0T^OAma0>o]OERb0;k]OFXb0P12M3N2M3N3L3N2M3M3M1O2N2N1O2N2N2M2O2N2N1O2N2N2N1Ooh<"}, "label": "a walking man in a dirty white tank top , black gym shorts and white sandals"}]}
{"id": 444583, "image": "COCO_train2014_000000444583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black shorts and a white shirt walking down the sidewalk\"."}], "task": "refering", "answer_template": "The \"a man in black shorts and a white shirt walking down the sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 101, 102, 115, 116, 117, 130, 131, 132, 145, 146, 147, 160, 161, 162, 175, 176, 177, 178, 190, 193, 194, 205, 208], "bbox": [0.6848349056603774, 0.22385937500000003, 0.9511320754716981, 0.591546875], "iscrowd": 0, "area": 9320.724699999997, "rle": {"size": [640, 424], "counts": "RSf56^4Kl:9fD0[;1ZD8g;JmC?T<BdCd0]<]OaCc0a<]O\\Cd0f<\\OXCe0h<\\OVCd0l<[ORCf0P=ZOnBg0S=YOkBg0W=YORB3WNf0h?WOkAa1V>`NdAe1]>ZN^Aj1d>VNVAm1m>SNm@o1W?QNe@P2^?PN]@e1\\OWN[`04U@h1DnMZ`09n_Ol1]`0WN\\_Om1e`0f0M2VN\\MUBj2g=[MTBh2m=YMPBi2P>YMmAg2U>YMiAh2W>YMfAi2[>XMbAj2]>WMaAk2_>UM^Am2b>TM[Ao2e>V1O10O10001O0001O2N1O2N1O1O0003M2N3YKWAQ4l>gKZAX4X?O1O2N0101N100O2O0O10mLWAn0j>kNiAi0Y>mN[Bf0g=QOeBj0^=mNjBR1W=eNRCY1m?N3N1N2N10O01O1O010O00100Ob]OUOfa0k0V^O]Oga0c0T^OAma0>o]OERb0;k]OFXb0P12M3N2M3N3L3N2M3M3M1O2N2N1O2N2N2M2O2N2N1O2N2N2N1Ooh<"}, "label": "a man in black shorts and a white shirt walking down the sidewalk"}]}
{"id": 559271, "image": "COCO_train2014_000000559271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant standing behind a tree\"."}], "task": "refering", "answer_template": "The \"a baby elephant standing behind a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 322, 323], "bbox": [0.006015625, 0.2786448598130841, 0.20437500000000003, 0.966285046728972], "iscrowd": 0, "area": 24857.598050000004, "rle": {"size": [428, 640], "counts": "cj1>m4f0o1NmKe1P4oNlIe2P6^2M4L3M3M4L3M4L3M3M4M2M4L3M3M4L3N0O10000O10000O10000O100O10000O101OO0M4M3L3N3L4L3N3L3M4M3L3M4M3N110O10O01000O010O0100O010O10O10O10O0100O010O10O10O010O10O0100O01000O0O2M3N1O2M2O2N2M2aIdIn5T8hNe0[O3L4M3M3L4M3M3L4M3M3L4M3M3M3L4M3L4M3L4M4L3L5L3L5I6J7J5J7I6K6I6Jl`d6"}, "label": "a baby elephant standing behind a tree"}]}
{"id": 559271, "image": "COCO_train2014_000000559271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby of an elephant\"."}], "task": "refering", "answer_template": "The \"a baby of an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 322, 323], "bbox": [0.006015625, 0.2786448598130841, 0.20437500000000003, 0.966285046728972], "iscrowd": 0, "area": 24857.598050000004, "rle": {"size": [428, 640], "counts": "cj1>m4f0o1NmKe1P4oNlIe2P6^2M4L3M3M4L3M4L3M3M4M2M4L3M3M4L3N0O10000O10000O10000O100O10000O101OO0M4M3L3N3L4L3N3L3M4M3L3M4M3N110O10O01000O010O0100O010O10O10O10O0100O010O10O10O010O10O0100O01000O0O2M3N1O2M2O2N2M2aIdIn5T8hNe0[O3L4M3M3L4M3M3L4M3M3L4M3M3M3L4M3L4M3L4M4L3L5L3L5I6J7J5J7I6K6I6Jl`d6"}, "label": "a baby of an elephant"}]}
{"id": 559271, "image": "COCO_train2014_000000559271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant in front of other elephants with trunk visible\"."}], "task": "refering", "answer_template": "The \"an elephant in front of other elephants with trunk visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 220, 234, 235, 236, 237, 238, 239, 240, 257, 259, 260, 261, 262, 263, 284], "bbox": [0.1544375, 0.19955607476635515, 0.60578125, 0.8385747663551402], "iscrowd": 0, "area": 47345.013049999994, "rle": {"size": [428, 640], "counts": "d_Y1b0j<6J5K6J6J6QDBd:d1O1O1O001O1O1O001O1O1O001O1O1O001O1O1XISMo2n2iLXMX3h2`L`M_3b2]KaMVN4^6\\2WKcNi4^1RKfNn4Z1PKgNQ5Z1kJhNU5Z1hJgNY5[1bJhN^5[1]JfNc5]1VJgNk5\\1mIhNS6[1gIhNZ6Z1_IjNa6P41O001O010O001O1O010O0100O1000O10O100000O1000O1O1O10O01O100O1O00100O1O10O100000000O0101O001O001Of0ZO0O010000O10000O1N1O2M3N2M30000000O101O00000000000000001O00000O100001OiJQIQ4P7`K^I`4h7O000000010O0000001O0001N1O100O100O1O2^Oa0XOh0WOi0I71O0001O0001O00105J6J6K5J6J6K5J7J1N00O2O0O2N100O2O0O1O2O0O2O0O1O2O0O101N1O2O0O1oLcIRO_6m0iIkNW6V1PJaNQ6_1WJYNj5f1]JRNd5o1cJhM_5W2iJaMW5`2PKWMQ5i2WKoLj4P3]KhLd4Y3cK^L^4b3jKVLW4j3e200010O00010O0010O00010O010O010O10000O0100000O10O10O1000O10O10000O010O1O1O1O001N4M4L5K4L4K5K5L4K6K4K5K3N2M3N3L3N2O101N1O1O100O2N1O100O2O0O100O101O0O10002M3N2M3N2N2M3N1N3M3M3L4M3MP_Y3"}, "label": "an elephant in front of other elephants with trunk visible"}]}
{"id": 559271, "image": "COCO_train2014_000000559271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in between the other two\"."}], "task": "refering", "answer_template": "The \"the elephant in between the other two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 220, 234, 235, 236, 237, 238, 239, 240, 257, 259, 260, 261, 262, 263, 284], "bbox": [0.1544375, 0.19955607476635515, 0.60578125, 0.8385747663551402], "iscrowd": 0, "area": 47345.013049999994, "rle": {"size": [428, 640], "counts": "d_Y1b0j<6J5K6J6J6QDBd:d1O1O1O001O1O1O001O1O1O001O1O1O001O1O1XISMo2n2iLXMX3h2`L`M_3b2]KaMVN4^6\\2WKcNi4^1RKfNn4Z1PKgNQ5Z1kJhNU5Z1hJgNY5[1bJhN^5[1]JfNc5]1VJgNk5\\1mIhNS6[1gIhNZ6Z1_IjNa6P41O001O010O001O1O010O0100O1000O10O100000O1000O1O1O10O01O100O1O00100O1O10O100000000O0101O001O001Of0ZO0O010000O10000O1N1O2M3N2M30000000O101O00000000000000001O00000O100001OiJQIQ4P7`K^I`4h7O000000010O0000001O0001N1O100O100O1O2^Oa0XOh0WOi0I71O0001O0001O00105J6J6K5J6J6K5J7J1N00O2O0O2N100O2O0O1O2O0O2O0O1O2O0O101N1O2O0O1oLcIRO_6m0iIkNW6V1PJaNQ6_1WJYNj5f1]JRNd5o1cJhM_5W2iJaMW5`2PKWMQ5i2WKoLj4P3]KhLd4Y3cK^L^4b3jKVLW4j3e200010O00010O0010O00010O010O010O10000O0100000O10O10O1000O10O10000O010O1O1O1O001N4M4L5K4L4K5K5L4K6K4K5K3N2M3N3L3N2O101N1O1O100O2N1O100O2O0O100O101O0O10002M3N2M3N2N2M3N1N3M3M3L4M3MP_Y3"}, "label": "the elephant in between the other two"}]}
{"id": 559271, "image": "COCO_train2014_000000559271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant which is with half head\"."}], "task": "refering", "answer_template": "The \"elephant which is with half head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314], "bbox": [0.46948437500000006, 0.3109579439252336, 0.83028125, 0.9226869158878503], "iscrowd": 0, "area": 35143.771499999995, "rle": {"size": [428, 640], "counts": "adm31R=<D<E;N1O2O1N2N2O1N2O1N3M2O1N2hF[Nd6f1UI`Nl6`1nHeNR7\\1gHjNZ7V1_HoNb7S1WHQOj7P1oGUOS8n0dGVO]8P1WGUOj8P1gFYOZ9\\2O100000000000000000000O10000O100O010_LnKgMS4X2oKgMQ4Y2PLfMo3Z2SLdMn3\\2SLcMm3\\2ULbMl3^2ULaMj3`2cLRM^3m2cLSM]3m2cLTM[3l2kLPMT3P3RMjLn2U3YMfLe2[3]MdLb2\\3`McL_2]3dMaL[2_3gM`LX2`3jM^LV2b3lM]LS2c3oM\\LP2c3SN\\Ll1d3WNZLi1e3YNZLh1d3ZNZLh1d3ZN[Lg1c3[N\\Le1c3]N\\Lb1d3`N`L[1_3fNjLP1V3ROmLj0R3WOmLi0S3YOkLh0T3YOkLg0U3E_L<`3LVL5k3OoK1S43gKNZ49]KGe4`0RKAo4\\42O1VO_JgIc5W6`JgIa5X6j0O1N2O1O1N2O1N2O1O1N2O1NnIlIP5S6PKUJi4i5XKXJg4h5XKZJh4e5XK\\Jh4b5YK^Jg4b5YK_Jg4_5YKcJg4\\5YKeJf4Z5[KgJe4W5\\KiJe4U5\\KlJd4S5[KoJe4P5[KQKe4n4[KRKf4n4YKSKg4l4XKUKi4j4RK\\Kn4d4lJaKU5^4fJhKZ5W4bJmK_5S4^JPLb5o3[JTLf5k3XJXLh5e5100O1000000O1000000O2O000O1000000O10000O2N1O1O1O1N2O1O1N2O2M2O1O1N2O1O1N2O1N2O2N1]MUIcNm6[1ZI]Ni6b1^IUNe6j1kIdMX6[2UJUMP6i2]J`Ln5^3Q2O2N1O2M2O1O2L3L4M4L3M4M2N3M2N3N101O1O001O0010O10000O101N10000O100O2O0O2O1O1N101N2O0O2O000O2O001O000O01N2M3N2YNYEf0i:WOYEh0k:TOVEk0o:oNSEP1R;jNPET1e;Od[]1"}, "label": "elephant which is with half head"}]}
{"id": 559271, "image": "COCO_train2014_000000559271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an african elephant standing in the right , eating\"."}], "task": "refering", "answer_template": "The \"an african elephant standing in the right , eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314], "bbox": [0.46948437500000006, 0.3109579439252336, 0.83028125, 0.9226869158878503], "iscrowd": 0, "area": 35143.771499999995, "rle": {"size": [428, 640], "counts": "adm31R=<D<E;N1O2O1N2N2O1N2O1N3M2O1N2hF[Nd6f1UI`Nl6`1nHeNR7\\1gHjNZ7V1_HoNb7S1WHQOj7P1oGUOS8n0dGVO]8P1WGUOj8P1gFYOZ9\\2O100000000000000000000O10000O100O010_LnKgMS4X2oKgMQ4Y2PLfMo3Z2SLdMn3\\2SLcMm3\\2ULbMl3^2ULaMj3`2cLRM^3m2cLSM]3m2cLTM[3l2kLPMT3P3RMjLn2U3YMfLe2[3]MdLb2\\3`McL_2]3dMaL[2_3gM`LX2`3jM^LV2b3lM]LS2c3oM\\LP2c3SN\\Ll1d3WNZLi1e3YNZLh1d3ZNZLh1d3ZN[Lg1c3[N\\Le1c3]N\\Lb1d3`N`L[1_3fNjLP1V3ROmLj0R3WOmLi0S3YOkLh0T3YOkLg0U3E_L<`3LVL5k3OoK1S43gKNZ49]KGe4`0RKAo4\\42O1VO_JgIc5W6`JgIa5X6j0O1N2O1O1N2O1N2O1O1N2O1NnIlIP5S6PKUJi4i5XKXJg4h5XKZJh4e5XK\\Jh4b5YK^Jg4b5YK_Jg4_5YKcJg4\\5YKeJf4Z5[KgJe4W5\\KiJe4U5\\KlJd4S5[KoJe4P5[KQKe4n4[KRKf4n4YKSKg4l4XKUKi4j4RK\\Kn4d4lJaKU5^4fJhKZ5W4bJmK_5S4^JPLb5o3[JTLf5k3XJXLh5e5100O1000000O1000000O2O000O1000000O10000O2N1O1O1O1N2O1O1N2O2M2O1O1N2O1O1N2O1N2O2N1]MUIcNm6[1ZI]Ni6b1^IUNe6j1kIdMX6[2UJUMP6i2]J`Ln5^3Q2O2N1O2M2O1O2L3L4M4L3M4M2N3M2N3N101O1O001O0010O10000O101N10000O100O2O0O2O1O1N101N2O0O2O000O2O001O000O01N2M3N2YNYEf0i:WOYEh0k:TOVEk0o:oNSEP1R;jNPET1e;Od[]1"}, "label": "an african elephant standing in the right , eating"}]}
{"id": 247979, "image": "COCO_train2014_000000247979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tall white coffee mug\"."}], "task": "refering", "answer_template": "The \"tall white coffee mug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 271, 272, 273, 286, 287, 288, 301, 302, 303, 304, 316, 317, 318, 319], "bbox": [0.08756440281030445, 0.664484375, 0.4479625292740046, 0.952125], "iscrowd": 0, "area": 14968.347950000005, "rle": {"size": [640, 427], "counts": "Yag02nc0f0[Od0[Oe0[Of0ZOe0\\Oe0ZOe0[Of0ZO:GO0000001O0000001O0000001O0000001O0000001O00000000000000001O000000000000000000001O00000000000000000000001O0000000000000O100O1O100O2lLeAa0[>^OQB6P>I]BJd=6gB_OY=`0T300O100O100O1O2O0O10000O10000O101O1N1O2N2N1N3N2N2N2N3L4L4J5K6K5K5N100O1O000000001O0000001O000001O01O00000100O1O1O100O1O1O00100O1O1O12N1O2N2N1O2N2N1OfPc4"}, "label": "tall white coffee mug"}]}
{"id": 247979, "image": "COCO_train2014_000000247979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white decorative coffee cup\"."}], "task": "refering", "answer_template": "The \"a white decorative coffee cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 271, 272, 273, 286, 287, 288, 301, 302, 303, 304, 316, 317, 318, 319], "bbox": [0.08756440281030445, 0.664484375, 0.4479625292740046, 0.952125], "iscrowd": 0, "area": 14968.347950000005, "rle": {"size": [640, 427], "counts": "Yag02nc0f0[Od0[Oe0[Of0ZOe0\\Oe0ZOe0[Of0ZO:GO0000001O0000001O0000001O0000001O0000001O00000000000000001O000000000000000000001O00000000000000000000001O0000000000000O100O1O100O2lLeAa0[>^OQB6P>I]BJd=6gB_OY=`0T300O100O100O1O2O0O10000O10000O101O1N1O2N2N1N3N2N2N2N3L4L4J5K6K5K5N100O1O000000001O0000001O000001O01O00000100O1O1O100O1O1O00100O1O1O12N1O2N2N1O2N2N1OfPc4"}, "label": "a white decorative coffee cup"}]}
{"id": 247979, "image": "COCO_train2014_000000247979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mug with a picture of a dog on it\"."}], "task": "refering", "answer_template": "The \"a mug with a picture of a dog on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 319, 320, 321, 322, 323, 324, 334, 335, 336, 337], "bbox": [0.2719672131147541, 0.76778125, 0.6383372365339578, 0.983171875], "iscrowd": 0, "area": 18072.07585, "rle": {"size": [640, 427], "counts": "eRY2h0Rb0W1iNV1VOj001O00001O0O2O00001O001O00001O00000000000000000000000O10000000000000000O100000000000000O110O0000000001O000000001O000000000000001O0O100000000000000O10000000001N10000000000O1000000000000000000O1N2K5K5000001O0eLj_Ol2W`0oLP@n2a`00000000000001O0000000000000000000000O100O2O0O2O0O2O0O2O0O2O0O2O1N2O0O2O1N2M3M2N4M3L4L4M4I6J6I7C=EWgo2"}, "label": "a mug with a picture of a dog on it"}]}
{"id": 247979, "image": "COCO_train2014_000000247979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black mug with a picture of a dog on it\"."}], "task": "refering", "answer_template": "The \"a black mug with a picture of a dog on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 319, 320, 321, 322, 323, 324, 334, 335, 336, 337], "bbox": [0.2719672131147541, 0.76778125, 0.6383372365339578, 0.983171875], "iscrowd": 0, "area": 18072.07585, "rle": {"size": [640, 427], "counts": "eRY2h0Rb0W1iNV1VOj001O00001O0O2O00001O001O00001O00000000000000000000000O10000000000000000O100000000000000O110O0000000001O000000001O000000000000001O0O100000000000000O10000000001N10000000000O1000000000000000000O1N2K5K5000001O0eLj_Ol2W`0oLP@n2a`00000000000001O0000000000000000000000O100O2O0O2O0O2O0O2O0O2O0O2O1N2O0O2O1N2M3M2N4M3L4L4M4I6J6I7C=EWgo2"}, "label": "a black mug with a picture of a dog on it"}]}
{"id": 297131, "image": "COCO_train2014_000000297131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white umbrella in the rain\"."}], "task": "refering", "answer_template": "The \"a blue and white umbrella in the rain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 146, 148, 149, 150, 171, 172, 173], "bbox": [0.298546875, 0.363875, 0.547484375, 0.6584375], "iscrowd": 0, "area": 7709.526600000001, "rle": {"size": [320, 640], "counts": "Xjk18g95L3L3N2M2O1N2N2O1N3N1N2O1N2O1O1N3N001N2O001N2O001N2O1O0O2O1O0O2O1O0O2OO1O010O10O0100O0I8J6L3N3N2N1O201O0O100000000O1000000000001O01O003M2N3M1O1O1O2N1O100O1O1O100O100O100O1O100O100O1O101N101J5N2MTNVHh1g7YNYHh1o72O0O2N100O2N3N3L1O01O0001O01O00001O00100O1O1O1O1O1O101N1O1O1O1O1O100O1O1N2O1O1O1N2O2N2N3L3N2N2N3M2M3N2N3M2M3N2L5K4M3Lo]j2"}, "label": "a blue and white umbrella in the rain"}]}
{"id": 297131, "image": "COCO_train2014_000000297131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open blue and white umbrella\"."}], "task": "refering", "answer_template": "The \"an open blue and white umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 146, 148, 149, 150, 171, 172, 173], "bbox": [0.298546875, 0.363875, 0.547484375, 0.6584375], "iscrowd": 0, "area": 7709.526600000001, "rle": {"size": [320, 640], "counts": "Xjk18g95L3L3N2M2O1N2N2O1N3N1N2O1N2O1O1N3N001N2O001N2O001N2O1O0O2O1O0O2O1O0O2OO1O010O10O0100O0I8J6L3N3N2N1O201O0O100000000O1000000000001O01O003M2N3M1O1O1O2N1O100O1O1O100O100O100O1O100O100O1O101N101J5N2MTNVHh1g7YNYHh1o72O0O2N100O2N3N3L1O01O0001O01O00001O00100O1O1O1O1O1O101N1O1O1O1O1O100O1O1N2O1O1O1N2O2N2N3L3N2N2N3M2M3N2N3M2M3N2L5K4M3Lo]j2"}, "label": "an open blue and white umbrella"}]}
{"id": 297131, "image": "COCO_train2014_000000297131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person under a white umbrela\"."}], "task": "refering", "answer_template": "The \"a person under a white umbrela\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239], "bbox": [0.34334375, 0.48890625, 0.47162500000000007, 1.0], "iscrowd": 0, "area": 9078.94075, "rle": {"size": [320, 640], "counts": "^nT26f98H9G6J4L4WIlNb4X1ZKnNa4T1\\KQOb4P1ZKVOd4j0XK[Of4f0VK^Oi4c0SKBj4?SKEj4>RKGk4;QKIn48bJ8]5H]Ja0a5_O\\Ji0_5WO_JP1\\5POaJX1Z5l1LO0000O1O1O1O1O100O1O100O1O100O1O100O11O000000000000001O0000001O2N2N2N3M3TOm0J5L4L3M2O0O2N2BZI`Mf6]2dI[M^6b2b0M2N3YOPIZNS7d1e0O2M3N2M6K6J5J5L7I<CP]Y3"}, "label": "a person under a white umbrela"}]}
{"id": 297131, "image": "COCO_train2014_000000297131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person standing outside holding a bright blue umbrella\"."}], "task": "refering", "answer_template": "The \"person standing outside holding a bright blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239], "bbox": [0.34334375, 0.48890625, 0.47162500000000007, 1.0], "iscrowd": 0, "area": 9078.94075, "rle": {"size": [320, 640], "counts": "^nT26f98H9G6J4L4WIlNb4X1ZKnNa4T1\\KQOb4P1ZKVOd4j0XK[Of4f0VK^Oi4c0SKBj4?SKEj4>RKGk4;QKIn48bJ8]5H]Ja0a5_O\\Ji0_5WO_JP1\\5POaJX1Z5l1LO0000O1O1O1O1O100O1O100O1O100O1O100O11O000000000000001O0000001O2N2N2N3M3TOm0J5L4L3M2O0O2N2BZI`Mf6]2dI[M^6b2b0M2N3YOPIZNS7d1e0O2M3N2M6K6J5J5L7I<CP]Y3"}, "label": "person standing outside holding a bright blue umbrella"}]}
{"id": 133295, "image": "COCO_train2014_000000133295.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog on the right facing the camera\"."}], "task": "refering", "answer_template": "The \"the dog on the right facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 210, 211], "bbox": [0.6067400000000001, 0.31234666666666666, 0.95898, 0.8876533333333333], "iscrowd": 0, "area": 24009.9527, "rle": {"size": [375, 500], "counts": "lW_31e;3M3L4M2nLFdJ<m46mJLQ5:jJH`M^Oa7P1dJIjMWO`7V1_JH`5<oI3n52jI3T63cI2\\62[I5b60oH<P7V2N3N1O1N3N2M2O1O1N2O0O2O1N2O0O2O1N2N010O00N2O21O2O0O3N1N3M2O2M3M3N2M4M5K5J4M2N2N1O000O2O000000000001O01O000001O000000001O0000001O000N2O1O1O1O1N2O1O1N2N2O1N2N3N1N2O100O2N100O1O101O0O101O0O2O1O000O101O001O1O001O1N2O1O1O1O1O1bM`ICa6:fI@\\6=jI]OY6`0nIVOX6g0oInNX6o0Z2M3M3N2O2M4M4L1O1N2O1O0O1000O10O10O1000[OPFBP:?QF@o9`0RF_Om9b0SF]On9c0SF\\Om9d0SF[Oo9d0RFVOS:j0>N10001O0O100000001N101N4K3M3M3M3Mm`7"}, "label": "the dog on the right facing the camera"}]}
{"id": 133295, "image": "COCO_train2014_000000133295.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog furthest to the right\"."}], "task": "refering", "answer_template": "The \"the dog furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 210, 211], "bbox": [0.6067400000000001, 0.31234666666666666, 0.95898, 0.8876533333333333], "iscrowd": 0, "area": 24009.9527, "rle": {"size": [375, 500], "counts": "lW_31e;3M3L4M2nLFdJ<m46mJLQ5:jJH`M^Oa7P1dJIjMWO`7V1_JH`5<oI3n52jI3T63cI2\\62[I5b60oH<P7V2N3N1O1N3N2M2O1O1N2O0O2O1N2O0O2O1N2N010O00N2O21O2O0O3N1N3M2O2M3M3N2M4M5K5J4M2N2N1O000O2O000000000001O01O000001O000000001O0000001O000N2O1O1O1O1N2O1O1N2N2O1N2N3N1N2O100O2N100O1O101O0O101O0O2O1O000O101O001O1O001O1N2O1O1O1O1O1bM`ICa6:fI@\\6=jI]OY6`0nIVOX6g0oInNX6o0Z2M3M3N2O2M4M4L1O1N2O1O0O1000O10O10O1000[OPFBP:?QF@o9`0RF_Om9b0SF]On9c0SF\\Om9d0SF[Oo9d0RFVOS:j0>N10001O0O100000001N101N4K3M3M3M3Mm`7"}, "label": "the dog furthest to the right"}]}
{"id": 133295, "image": "COCO_train2014_000000133295.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown dog in lying in front of the black dog\"."}], "task": "refering", "answer_template": "The \"the brown dog in lying in front of the black dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 90, 91, 92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189], "bbox": [0.00152, 0.35202666666666665, 0.63508, 0.7814399999999999], "iscrowd": 0, "area": 34459.2225, "rle": {"size": [375, 500], "counts": "ha0i0m:5K2N2N2N2N2N2M3K5H8I6J7M3M4M3M3L4M2N1N101O1N2O1O1O1O10000O100O10000O10000O10000O1000O0100O10000O10001N101O001O001O001O00100O001O001O0010O010O01O010O010O1O010O010O0010O010O01O010O010O00100O010O0001O001O00001O001O00000000000000000000001O00000000000000000000000010O00000000000000000000000001O0000000000000000000000001O0001O00000000000001O0000SHcLa7\\3_HhL]7X3bHjL]7W3bHiL^7W3bHjL^7U3aHlL_7T3aHmL^7S3aHnL`7Q3`HPM_7P3aHPM_7P3`HRM`7m2`HSM`7m2`HTM_7l2`HUM`7k2`HVM`7i2`HXM_7i2_HXMa7h2_HXMb7^300001O0000001O00001O00001O0000001O00001O00001O0001O0001O000000001O00000000001O000000001O000000001O000000001O00000000001O000000001O000000001O000O1000001O00000O101O0000001O001N2O001O1O001O1N101O1O001O1O001N2O001O1O001N1O2N2N2N2N3M3M2N3L3N3M3M4L5K6J6J5K6H7B?A_]R2"}, "label": "the brown dog in lying in front of the black dog"}]}
{"id": 133295, "image": "COCO_train2014_000000133295.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color dog is lying between two other dogs in the bed\"."}], "task": "refering", "answer_template": "The \"a black color dog is lying between two other dogs in the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 117, 118, 119], "bbox": [0.29794, 0.2950666666666667, 0.67802, 0.5455466666666667], "iscrowd": 0, "area": 9139.196100000001, "rle": {"size": [375, 500], "counts": "Vff12c;2N2N2M3N2N10100O1O10O0100O011N>TEROU:Q1fEROY:W1O0O10000000001O0001O000000010O00000001O01O01O0000001O00000O2O00001O0000001O01O000010O01O0001O0001O0001O01O001O010O001O010O1O1O1O100O2N1O0001O01O000001O01O1O2N01O0O10001N10001N10001N1O100O2N1O101N1O101N1O101N100O2O000O2O1N101N2O001N101O1N101O1O1O1N2O1O001O1N2O1O1O1O1N2O100O1OO101O000O101M2H8F:K6K4M4M2N3M2O2N1O2M2O2N2M3M2N3M3M3M2N3N2M3N3L3N2Mcjj1"}, "label": "a black color dog is lying between two other dogs in the bed"}]}
{"id": 133295, "image": "COCO_train2014_000000133295.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black dog laying down behind brown dogs\"."}], "task": "refering", "answer_template": "The \"black dog laying down behind brown dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 117, 118, 119], "bbox": [0.29794, 0.2950666666666667, 0.67802, 0.5455466666666667], "iscrowd": 0, "area": 9139.196100000001, "rle": {"size": [375, 500], "counts": "Vff12c;2N2N2M3N2N10100O1O10O0100O011N>TEROU:Q1fEROY:W1O0O10000000001O0001O000000010O00000001O01O01O0000001O00000O2O00001O0000001O01O000010O01O0001O0001O0001O01O001O010O001O010O1O1O1O100O2N1O0001O01O000001O01O1O2N01O0O10001N10001N10001N1O100O2N1O101N1O101N1O101N100O2O000O2O1N101N2O001N101O1N101O1O1O1N2O1O001O1N2O1O1O1O1N2O100O1OO101O000O101M2H8F:K6K4M4M2N3M2O2N1O2M2O2N2M3M2N3M3M3M2N3N2M3N3L3N2Mcjj1"}, "label": "black dog laying down behind brown dogs"}]}
{"id": 264371, "image": "COCO_train2014_000000264371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and black toothbrush\"."}], "task": "refering", "answer_template": "The \"white and black toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 210, 211, 212], "bbox": [0.0, 0.6689825581395349, 0.8240599999999999, 0.9723546511627907], "iscrowd": 0, "area": 28042.808250000002, "rle": {"size": [344, 500], "counts": "e79Z:6O0O2O000O2O001N101O0O2O00001N101O0O2O000O2O001N101O0O2O000O2O0001O0O101O0000001N10001O0000001N10000010O001O001O001O1O1O1O2N1O1N2O001O00001O00001N10001O000O2O0000000O2O0000000O101O000O10010O0100O100O100O1O10O0O2O001O001O001O001O001O001O001O001O0O101O001O001O0000001O0000001O0000000O2O0000001O0000001O0000001O000000001N1000001O0000001O0000001O000000001N1000001O0000001O0000001O000001O01O0000001O0000001O0001O01O000000001O0000001O01O0001O0000001O0000000010O000001O0000001O00000010O00000001O0000001O0000010O0000001O000000001O0001O01O0000001O000000000001O00000000000000000000001O00000000000O10000000001O00000000000000000000001O000000000O1000000000001O0000000000001O00000000001O0000000O10001O00000000001O0000000000001O00000000001O000O100000001O00000000001O00000000000O2O000O10000O101O0O100O10001N10000O10000O2O000O8I:F9F;F:E;FVXm0"}, "label": "white and black toothbrush"}]}
{"id": 264371, "image": "COCO_train2014_000000264371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black oral - b toothbrush\"."}], "task": "refering", "answer_template": "The \"a black oral - b toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 210, 211, 212], "bbox": [0.0, 0.6689825581395349, 0.8240599999999999, 0.9723546511627907], "iscrowd": 0, "area": 28042.808250000002, "rle": {"size": [344, 500], "counts": "e79Z:6O0O2O000O2O001N101O0O2O00001N101O0O2O000O2O001N101O0O2O000O2O0001O0O101O0000001N10001O0000001N10000010O001O001O001O1O1O1O2N1O1N2O001O00001O00001N10001O000O2O0000000O2O0000000O101O000O10010O0100O100O100O1O10O0O2O001O001O001O001O001O001O001O001O0O101O001O001O0000001O0000001O0000000O2O0000001O0000001O0000001O000000001N1000001O0000001O0000001O000000001N1000001O0000001O0000001O000001O01O0000001O0000001O0001O01O000000001O0000001O01O0001O0000001O0000000010O000001O0000001O00000010O00000001O0000001O0000010O0000001O000000001O0001O01O0000001O000000000001O00000000000000000000001O00000000000O10000000001O00000000000000000000001O000000000O1000000000001O0000000000001O00000000001O0000000O10001O00000000001O0000000000001O00000000001O000O100000001O00000000001O00000000000O2O000O10000O101O0O100O10001N10000O10000O2O000O8I:F9F;F:E;FVXm0"}, "label": "a black oral - b toothbrush"}]}
{"id": 264371, "image": "COCO_train2014_000000264371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue toothbrush\"."}], "task": "refering", "answer_template": "The \"the blue toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 129, 130, 131, 147], "bbox": [0.16016, 0.15409883720930231, 0.99998, 0.6757848837209303], "iscrowd": 0, "area": 26026.203799999996, "rle": {"size": [344, 500], "counts": "oRk03b:4M3N1N3M3N1N3N1O1O1O1O1O2M2O1O1O1O1O1O1N2I7O1O001O1O1O001O100O010O10O0100O10O0100O010O010000O01000O1000O010000O1N101O1O1O1N2O1O1O1O1O1O100O1O10O01O1O100O1O1O1O001J6O100O1000O1O001O1O10O0100000O100000O100O1O1O100O1O1O100O1O2N2O2M2N3M2O1N1O101N101O00001O0000000000000000000O10000000000000O10O1000O10O10O100000O0100000O01000000O0100000O0100000O10O1000O10O1000O1000O010O100O1000O0100O100O1000O0100O100O100O01000O100O100O01000O100O100O010O10000O100O010O10000O100O01000O100O1000O010000O100O10O10O10000O010O10000O10O010000O1000O0100O10000O010O10000O10O10O10000O01000O10000O01000O10000O0100000O1000O010000O1000O010000O10O10O10000O10O10O10000O10O10O10000O10O1000O10000O01000O1000O10O10000O10O10O10000O10O1000O1000O010000O1000O01000000O10O10O10000O10O10O100000O010000O1000O01000000O10O10O1000000O01000O1000O10O10000O10O1000O10000O010000O010O1O010YMR1"}, "label": "the blue toothbrush"}]}
{"id": 264371, "image": "COCO_train2014_000000264371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue and green toothbrush in the center\"."}], "task": "refering", "answer_template": "The \"the blue and green toothbrush in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 129, 130, 131, 147], "bbox": [0.16016, 0.15409883720930231, 0.99998, 0.6757848837209303], "iscrowd": 0, "area": 26026.203799999996, "rle": {"size": [344, 500], "counts": "oRk03b:4M3N1N3M3N1N3N1O1O1O1O1O2M2O1O1O1O1O1O1N2I7O1O001O1O1O001O100O010O10O0100O10O0100O010O010000O01000O1000O010000O1N101O1O1O1N2O1O1O1O1O1O100O1O10O01O1O100O1O1O1O001J6O100O1000O1O001O1O10O0100000O100000O100O1O1O100O1O1O100O1O2N2O2M2N3M2O1N1O101N101O00001O0000000000000000000O10000000000000O10O1000O10O10O100000O0100000O01000000O0100000O0100000O10O1000O10O1000O1000O010O100O1000O0100O100O1000O0100O100O100O01000O100O100O01000O100O100O010O10000O100O010O10000O100O01000O100O1000O010000O100O10O10O10000O010O10000O10O010000O1000O0100O10000O010O10000O10O10O10000O01000O10000O01000O10000O0100000O1000O010000O1000O010000O10O10O10000O10O10O10000O10O10O10000O10O1000O10000O01000O1000O10O10000O10O10O10000O10O1000O1000O010000O1000O01000000O10O10O10000O10O10O100000O010000O1000O01000000O10O10O1000000O01000O1000O10O10000O10O1000O10000O010000O010O1O010YMR1"}, "label": "the blue and green toothbrush in the center"}]}
{"id": 264371, "image": "COCO_train2014_000000264371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toothbrush on the right\"."}], "task": "refering", "answer_template": "The \"toothbrush on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 61, 62, 63, 64, 65, 66, 67, 78, 79, 80, 81, 96], "bbox": [0.36632, 0.004680232558139535, 0.9989000000000001, 0.45212209302325584], "iscrowd": 0, "area": 11322.5759, "rle": {"size": [344, 500], "counts": "ecm11f:4L4L5L3L1O0001N1000001O00000100O1O00100O1D;10O1O10O0100O100O1O10O0100O100O100O100O01000O10000O01000O10O10O100O10O10O1000O010000O01000O10O10O01000O0100O100O2O000O1O2N1O1O1O2N1O1O001O1O001O001000O0100O01000O010O01000O01000O010O01000O01000O010O10O100O01000O10000O01000O100O10O01O100O1O1O1O100O1O1O1O100O1O1O1O100O1O00100O001O10O10000O10O10O10000O100000O010000O10000O10O10O10000O10000O0100000O01000O01000O0100O01000O01000O01000O01000O10O10O10O01000O01000O01000O01000O0100O010O100O010O10O10O10O0100O010O10O0100O01000O010O100O010O10O0100O010O10O10O10O0100O010O10O0100O10O10O10O0100O010O10O0100O01000O010O100O010O10O0100O01000O010;E=B>Ca9"}, "label": "toothbrush on the right"}]}
{"id": 264371, "image": "COCO_train2014_000000264371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toothbrush with blue / green handlegrip and blue and green bristles\"."}], "task": "refering", "answer_template": "The \"a toothbrush with blue / green handlegrip and blue and green bristles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 61, 62, 63, 64, 65, 66, 67, 78, 79, 80, 81, 96], "bbox": [0.36632, 0.004680232558139535, 0.9989000000000001, 0.45212209302325584], "iscrowd": 0, "area": 11322.5759, "rle": {"size": [344, 500], "counts": "ecm11f:4L4L5L3L1O0001N1000001O00000100O1O00100O1D;10O1O10O0100O100O1O10O0100O100O100O100O01000O10000O01000O10O10O100O10O10O1000O010000O01000O10O10O01000O0100O100O2O000O1O2N1O1O1O2N1O1O001O1O001O001000O0100O01000O010O01000O01000O010O01000O01000O010O10O100O01000O10000O01000O100O10O01O100O1O1O1O100O1O1O1O100O1O1O1O100O1O00100O001O10O10000O10O10O10000O100000O010000O10000O10O10O10000O10000O0100000O01000O01000O0100O01000O01000O01000O01000O10O10O10O01000O01000O01000O01000O0100O010O100O010O10O10O10O0100O010O10O0100O01000O010O100O010O10O0100O010O10O10O10O0100O010O10O0100O10O10O10O0100O010O10O0100O01000O010O100O010O10O0100O01000O010;E=B>Ca9"}, "label": "a toothbrush with blue / green handlegrip and blue and green bristles"}]}
{"id": 223415, "image": "COCO_train2014_000000223415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a white shirt\"."}], "task": "refering", "answer_template": "The \"a boy with a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 50, 51, 52, 73, 74, 75, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 141, 142, 143, 144, 165, 166, 167, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304, 305, 306, 326, 327, 328, 329, 351], "bbox": [0.092703125, 0.07864583333333333, 0.3438125, 0.8921250000000001], "iscrowd": 0, "area": 28839.139600000006, "rle": {"size": [480, 640], "counts": "`ik01o>0O2N2O0O2N101N2N101N2N1O2O1N1O2O1N1O2O0O2N2N101N2N101N2N101N2N1O2O0O2N2O0O2N2O0O2O1N101N101N2O0O2O1N101N2O0O2O1UJiMjNY2m0VNlNn1h0aNPOe1d0cN[Ob18fNG`1LhN4\\1AkN>[1UOmNk0X1gNQOX1T1XNXOh1m0hM_OV2g0ZMDe2b0kLIU3;\\L0c3g500dNWF`Mj9]2]F_Mc9^2cF^M_9^2hF^MX9_2oF\\MR9a2UG[Mk8d2ZGWMg8i2\\GTMd8l2`GoLa8Q3g101O000000000000001N10000O1000000O2O00000000001O00000000001O0000000010O00001O2O1N2N3M2Na0^O;F:F6I6K5Jc0^Oh0XO]1bNb1mKeG_N`<N1N3N2M3N1O2M3N1O2M3N2N1N3N2M3N1O2M3N2N1N3N1N2O2NSnT6"}, "label": "a boy with a white shirt"}]}
{"id": 223415, "image": "COCO_train2014_000000223415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white shirt and a helmet on standing on his skateboard\"."}], "task": "refering", "answer_template": "The \"a man with a white shirt and a helmet on standing on his skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 50, 51, 52, 73, 74, 75, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 141, 142, 143, 144, 165, 166, 167, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304, 305, 306, 326, 327, 328, 329, 351], "bbox": [0.092703125, 0.07864583333333333, 0.3438125, 0.8921250000000001], "iscrowd": 0, "area": 28839.139600000006, "rle": {"size": [480, 640], "counts": "`ik01o>0O2N2O0O2N101N2N101N2N1O2O1N1O2O1N1O2O0O2N2N101N2N101N2N101N2N1O2O0O2N2O0O2N2O0O2O1N101N101N2O0O2O1N101N2O0O2O1UJiMjNY2m0VNlNn1h0aNPOe1d0cN[Ob18fNG`1LhN4\\1AkN>[1UOmNk0X1gNQOX1T1XNXOh1m0hM_OV2g0ZMDe2b0kLIU3;\\L0c3g500dNWF`Mj9]2]F_Mc9^2cF^M_9^2hF^MX9_2oF\\MR9a2UG[Mk8d2ZGWMg8i2\\GTMd8l2`GoLa8Q3g101O000000000000001N10000O1000000O2O00000000001O00000000001O0000000010O00001O2O1N2N3M2Na0^O;F:F6I6K5Jc0^Oh0XO]1bNb1mKeG_N`<N1N3N2M3N1O2M3N1O2M3N2N1N3N2M3N1O2M3N2N1N3N1N2O2NSnT6"}, "label": "a man with a white shirt and a helmet on standing on his skateboard"}]}
{"id": 166073, "image": "COCO_train2014_000000166073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy holding a tennis racquet\"."}], "task": "refering", "answer_template": "The \"a little boy holding a tennis racquet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338], "bbox": [0.43726562500000005, 0.040437956204379566, 0.7706250000000001, 0.9865206812652068], "iscrowd": 0, "area": 55229.47549999999, "rle": {"size": [411, 640], "counts": "mb`3l0l;4L3M3M4L3M4L3N2N3M2N3M2N2N3M2N3M2N2N3M2N3L3N2N3hJUM<m2@WM=m2_OVM?l2]OXMa0j2\\OYMlNG[NS3e2YM\\N=hN\\2i2ZMlMR1UOf1k2[MaMb1^OU1n2\\MdMd1XOQ1Q3^MfMg1ROn0S3_MjMi1iNm0Z3]MlMf5P2]JoMd5n1_JQNb5l1aJSNa5i1bJWN^5f1eJYN\\5QOVIT2a1jN[5oN\\In1\\1ROY5nNbIh1X1YOU6?nI@X68kIG[61hIO]6IfI6`6BcI=c6[O`Id0f6SO^Il0Y90000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000001O01O001O00001O001O001O00001O001O0XGSOb5n0\\JSOd5m0[JUOd5l0ZJUOf5k0YJWOf5i0YJXOg5i0WJYOh5g0WJZOi5g0UJZOk5f0TJ\\Ok5e0SJ\\Om5d0SJ]Ol5c0SJ^Om5c0QJ_On5a0QJ@o5a0oI@Q6`0nIBQ6?mIBS6>lIDS6=kIDU6<jIFU6:jIGV6:hIGX69gIIX68eHj0[7\\2O001O00001O001O001O00001O001O001O00001O001O001O00001O1O1O1O1O1M3K5K5K5K5K5K5L4M3L4M3L4M3L4M3M3L4M3L4M3L4L5L3L4M3gMbHPOb7m0jHgNY7V1mHcNW7Z1oH_NU7^1RI[NQ7b1UIWNo6f1WITNl6i1U2N2M3M3M3M6J7I8H7I8H7IUij1"}, "label": "a little boy holding a tennis racquet"}]}
{"id": 166073, "image": "COCO_train2014_000000166073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy holding the tennis racket\"."}], "task": "refering", "answer_template": "The \"the boy holding the tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338], "bbox": [0.43726562500000005, 0.040437956204379566, 0.7706250000000001, 0.9865206812652068], "iscrowd": 0, "area": 55229.47549999999, "rle": {"size": [411, 640], "counts": "mb`3l0l;4L3M3M4L3M4L3N2N3M2N3M2N2N3M2N3M2N2N3M2N3L3N2N3hJUM<m2@WM=m2_OVM?l2]OXMa0j2\\OYMlNG[NS3e2YM\\N=hN\\2i2ZMlMR1UOf1k2[MaMb1^OU1n2\\MdMd1XOQ1Q3^MfMg1ROn0S3_MjMi1iNm0Z3]MlMf5P2]JoMd5n1_JQNb5l1aJSNa5i1bJWN^5f1eJYN\\5QOVIT2a1jN[5oN\\In1\\1ROY5nNbIh1X1YOU6?nI@X68kIG[61hIO]6IfI6`6BcI=c6[O`Id0f6SO^Il0Y90000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000001O01O001O00001O001O001O00001O001O0XGSOb5n0\\JSOd5m0[JUOd5l0ZJUOf5k0YJWOf5i0YJXOg5i0WJYOh5g0WJZOi5g0UJZOk5f0TJ\\Ok5e0SJ\\Om5d0SJ]Ol5c0SJ^Om5c0QJ_On5a0QJ@o5a0oI@Q6`0nIBQ6?mIBS6>lIDS6=kIDU6<jIFU6:jIGV6:hIGX69gIIX68eHj0[7\\2O001O00001O001O001O00001O001O001O00001O001O001O00001O1O1O1O1O1M3K5K5K5K5K5K5L4M3L4M3L4M3L4M3M3L4M3L4M3L4L5L3L4M3gMbHPOb7m0jHgNY7V1mHcNW7Z1oH_NU7^1RI[NQ7b1UIWNo6f1WITNl6i1U2N2M3M3M3M6J7I8H7I8H7IUij1"}, "label": "the boy holding the tennis racket"}]}
{"id": 166073, "image": "COCO_train2014_000000166073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in background wearing a striped shirt\"."}], "task": "refering", "answer_template": "The \"person in background wearing a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 34, 37, 38, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 146, 147, 148, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 283, 284, 285, 306, 307, 308, 329, 330], "bbox": [0.307390625, 0.002238442822384428, 0.6595, 0.9550608272506084], "iscrowd": 0, "area": 24695.97245, "rle": {"size": [411, 640], "counts": "Z^_22g<2N2M3N2O1N3N1N2O1N2O1N2O1N3M2O1N2O1N2O1N2O1N3UMoNiIR1R6\\OcIf0X6H^I9]65XIMc6a0RIAi6m0mHTOm6[1hHgNS7g1cH[NW7f3L5K5^L^JFg51bJJc5MgJL^5LkJOZ5InJ2W5ESK5R5BWK9m4_O\\K<i4\\O`K=f4ZOcKa0b4VOgKe0]4XOhKb0]4]OdK>a4B_K9f4G[K2j4OVKLo44QKGT59mJAX5?hJ\\O]5d0\\31^NJ_F6`9c1O1O1N2O1O1O1O1O1O1O1N2N2M3N2M3N2M3N2M3N2M3N2M3N2M3[MYMTKi2U4oMgKT2X4nMeKT2Y4nMdKU2[4mMbKU2\\4nM`KU2_4mM^KU2a4lM\\KW2b4lM[KV2d4lMXKW2g4jMWKX2g4kMUKX2j4jMSKX2k4jMRKY2m4iMPKY2o4j2N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1mLTHU1n7jNeIC\\6<jI^OW6a0nIYOT6f0QJUOP6j0VJPOk5o0Q3O1O1O1O1O1O1O1O100O2O000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O10000O100001O001O00001O001O00001O001O00001O001O00001O001O0O101O001O00001O001O001O001O001O001O001O001O1O02O0O2O1N2O1N:GW_g2"}, "label": "person in background wearing a striped shirt"}]}
{"id": 166073, "image": "COCO_train2014_000000166073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person behind little tennis player\"."}], "task": "refering", "answer_template": "The \"person behind little tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 34, 37, 38, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 146, 147, 148, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 283, 284, 285, 306, 307, 308, 329, 330], "bbox": [0.307390625, 0.002238442822384428, 0.6595, 0.9550608272506084], "iscrowd": 0, "area": 24695.97245, "rle": {"size": [411, 640], "counts": "Z^_22g<2N2M3N2O1N3N1N2O1N2O1N2O1N3M2O1N2O1N2O1N2O1N3UMoNiIR1R6\\OcIf0X6H^I9]65XIMc6a0RIAi6m0mHTOm6[1hHgNS7g1cH[NW7f3L5K5^L^JFg51bJJc5MgJL^5LkJOZ5InJ2W5ESK5R5BWK9m4_O\\K<i4\\O`K=f4ZOcKa0b4VOgKe0]4XOhKb0]4]OdK>a4B_K9f4G[K2j4OVKLo44QKGT59mJAX5?hJ\\O]5d0\\31^NJ_F6`9c1O1O1N2O1O1O1O1O1O1O1N2N2M3N2M3N2M3N2M3N2M3N2M3N2M3[MYMTKi2U4oMgKT2X4nMeKT2Y4nMdKU2[4mMbKU2\\4nM`KU2_4mM^KU2a4lM\\KW2b4lM[KV2d4lMXKW2g4jMWKX2g4kMUKX2j4jMSKX2k4jMRKY2m4iMPKY2o4j2N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1mLTHU1n7jNeIC\\6<jI^OW6a0nIYOT6f0QJUOP6j0VJPOk5o0Q3O1O1O1O1O1O1O1O100O2O000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O10000O100001O001O00001O001O00001O001O00001O001O00001O001O0O101O001O00001O001O001O001O001O001O001O001O1O02O0O2O1N2O1N:GW_g2"}, "label": "person behind little tennis player"}]}
{"id": 526521, "image": "COCO_train2014_000000526521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with trunk down\"."}], "task": "refering", "answer_template": "The \"elephant with trunk down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.4528125, 0.3575117370892019, 1.0, 0.9901643192488263], "iscrowd": 0, "area": 60378.11825000001, "rle": {"size": [426, 640], "counts": "jmh3`0f<6M3N1N3N2M3N2N1N3N1N2O0O2O1O0O2O1N2O001N2O0O2O1N101O1N2O0O2O1N1O2O1N1O2N1O1O101N1O1O1O2N100O1O1O2N1O100O1O2ROYMVGh2j8]MPGd2P9aMjF`2V9eMeF\\2Y9jMaFW2_9j000O101O000O101O000O103M2M4M2N3M3L3N3M2N3hHWKh5l4RJXKn5k4jIZKV6h4dI\\K[6h4^I\\Kb6f4XI^Kh6e4QI_Ko6^5O1000000O10000O100O100O10000O100O100O10000O100O100O100O100O1O1O10O01O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O100O1000000O1000000O11O001O1O001O1O001O1O001O1O001O3M3M2N3M3M2N3M2N3M2N3M2N3M2N2N2N1O1O2N1O1O2N1O2N1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O00000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000001O6mNT1B>A?lN\\G"}, "label": "elephant with trunk down"}]}
{"id": 526521, "image": "COCO_train2014_000000526521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with small tusks\"."}], "task": "refering", "answer_template": "The \"an elephant with small tusks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.4528125, 0.3575117370892019, 1.0, 0.9901643192488263], "iscrowd": 0, "area": 60378.11825000001, "rle": {"size": [426, 640], "counts": "jmh3`0f<6M3N1N3N2M3N2N1N3N1N2O0O2O1O0O2O1N2O001N2O0O2O1N101O1N2O0O2O1N1O2O1N1O2N1O1O101N1O1O1O2N100O1O1O2N1O100O1O2ROYMVGh2j8]MPGd2P9aMjF`2V9eMeF\\2Y9jMaFW2_9j000O101O000O101O000O103M2M4M2N3M3L3N3M2N3hHWKh5l4RJXKn5k4jIZKV6h4dI\\K[6h4^I\\Kb6f4XI^Kh6e4QI_Ko6^5O1000000O10000O100O100O10000O100O100O10000O100O100O100O100O1O1O10O01O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O100O1000000O1000000O11O001O1O001O1O001O1O001O1O001O3M3M2N3M3M2N3M2N3M2N3M2N3M2N2N2N1O1O2N1O1O2N1O2N1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O00000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000001O6mNT1B>A?lN\\G"}, "label": "an elephant with small tusks"}]}
{"id": 526521, "image": "COCO_train2014_000000526521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with trunk up\"."}], "task": "refering", "answer_template": "The \"elephant with trunk up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0044843750000000005, 0.10572769953051643, 0.73875, 1.0], "iscrowd": 0, "area": 88531.3374, "rle": {"size": [426, 640], "counts": "Xa1m3^9O00001O0000010O0000001OO1M3M3N2M3M3M3N2M3M3N2M3M3M3N2M3M3N3L3O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O2N1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000000000O100000O10O10000000000O10000000000O10000000000O1000O100000O10000000000000000000000000000]NUGgNk8U1aGcN_8X1kGcNT8Z1PHfNP8V1THjNl7R1XHnNh7P1ZHPOf7o0[HQOe7n0\\HROd7m0]HSOc7l0^HTOb7k0_HUOa7j0`HVO`7i0aHWO_7h0bHWO_7h0bHXO_7f0bHZO^7e0cH[O]7d0dH\\O\\7c0eH]O[7b0eH_O[7`0fH@Z7`0fH@Z7?gHAY7>hHBX7>hHBX7>hHBX7>gHCY7=fHDm3mNiM^1YNGl3POhMY1ZNIk3SOiMT1YNLk3UOjMo0YNNl3WOhMk0ZN0n3WOeMj0ZN2P4WOdMg0\\N2P4YOmKGMn062o3]O\\K88:=1o3_OkJj0d0Fb01n3d1ZKZNh03n3h1TKVNn02m3o1oJoMU11l3R2mJmMW11k3g3ULYLk3g3ULYLj3h3WLVLj3j3VLVLi3k3WLULi3k3XLTLh3l3XLTLg3m3YLSLg3m3ZLRLe3o3[LQLe3o3[LPLe3Q4\\LnKd3R4]LmKb3T4_LkKa3U4_LkK`3V4aLiK_3W4bLhK]3Z4bLfK^3Z4cLdK]3]4dLbK\\3^4eLaKZ3`4fL`KZ3`4gL_KX3b4iL]KW3c4jL\\KU3e4kLZKV3f4kLYKT3h4mLWKS3i4mLWKR3j4oLUKQ3k4PMTKo2m4QMSKo2m4RMRKm2P5SMnJn2R5SMmJl2U5SMkJm2V5SMiJl2X5UMgJk2Z5TMfJk2[5VMdJV1[OaNR69aJW1^O^NR6;`JW1_O]NR6<^JW1@\\NS6>\\JV1B[NS6>[JW1CYNS6a0YJV1DYNT6a0VJW1GVNU6b0TJX1HUNT6d0SJW1JSNU6f0oIX1LRNU6g0nIW1NPNV6h0lIX1OoMU6j0kIW10nMW6k0gIX13kMW6n0eIW15jMW6o0cIW17hMX6P1aIX17gMY6R1^IX1:eMY6S1\\IX1<cMY6V1ZIW1>bMY6V1XIY1`0_MY6a4gI^K[6a4fI^KZ6b4gI\\K[6c4fI[K\\6c4fI\\KZ6d4fI[K\\6d4eIZK\\6f4eIYK\\6f4eIXK\\6h4T1O0O2O0O2N2N1O2O1N2N2N2N2N2O1N2O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O:Fc0\\O>jNgDAd;3\\DNg;NZD2e;OZD2f;M[D2e;N[D3d;M]D3b;M^D4b;L^D4a;L_D5`;KaD4_;LaD5_;JbD6];KbD6];JdD6[;JeD6\\;IeD7Z;IfD8Y;HhD8W;IhD8X;GiD8W;HiD9V;GkD9T;GlD:T;EmD;R;DoD<Q;DPE<o:CRE>n:ARE`0m:_OTEb0k:^OUEc0j:\\OXEc0i:[OXEe0h:[OXEe0h:ZOYEf0g:ZOYEf0h:XOYEh0g:XOYEh0g:WOZEi0f:WOZEi0g:UOZEk0f:TO\\Ek0d:UO\\Ek0e:SO\\Em0f:QOZEP1_;000000000001O0000000000000000001O0000000000000000001O001N2M2O2N2M3N2M3N2N2M2O2M3N2NT[U2"}, "label": "elephant with trunk up"}]}
{"id": 526521, "image": "COCO_train2014_000000526521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant in front of another\"."}], "task": "refering", "answer_template": "The \"an elephant in front of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0044843750000000005, 0.10572769953051643, 0.73875, 1.0], "iscrowd": 0, "area": 88531.3374, "rle": {"size": [426, 640], "counts": "Xa1m3^9O00001O0000010O0000001OO1M3M3N2M3M3M3N2M3M3N2M3M3M3N2M3M3N3L3O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O2N1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000000000O100000O10O10000000000O10000000000O10000000000O1000O100000O10000000000000000000000000000]NUGgNk8U1aGcN_8X1kGcNT8Z1PHfNP8V1THjNl7R1XHnNh7P1ZHPOf7o0[HQOe7n0\\HROd7m0]HSOc7l0^HTOb7k0_HUOa7j0`HVO`7i0aHWO_7h0bHWO_7h0bHXO_7f0bHZO^7e0cH[O]7d0dH\\O\\7c0eH]O[7b0eH_O[7`0fH@Z7`0fH@Z7?gHAY7>hHBX7>hHBX7>hHBX7>gHCY7=fHDm3mNiM^1YNGl3POhMY1ZNIk3SOiMT1YNLk3UOjMo0YNNl3WOhMk0ZN0n3WOeMj0ZN2P4WOdMg0\\N2P4YOmKGMn062o3]O\\K88:=1o3_OkJj0d0Fb01n3d1ZKZNh03n3h1TKVNn02m3o1oJoMU11l3R2mJmMW11k3g3ULYLk3g3ULYLj3h3WLVLj3j3VLVLi3k3WLULi3k3XLTLh3l3XLTLg3m3YLSLg3m3ZLRLe3o3[LQLe3o3[LPLe3Q4\\LnKd3R4]LmKb3T4_LkKa3U4_LkK`3V4aLiK_3W4bLhK]3Z4bLfK^3Z4cLdK]3]4dLbK\\3^4eLaKZ3`4fL`KZ3`4gL_KX3b4iL]KW3c4jL\\KU3e4kLZKV3f4kLYKT3h4mLWKS3i4mLWKR3j4oLUKQ3k4PMTKo2m4QMSKo2m4RMRKm2P5SMnJn2R5SMmJl2U5SMkJm2V5SMiJl2X5UMgJk2Z5TMfJk2[5VMdJV1[OaNR69aJW1^O^NR6;`JW1_O]NR6<^JW1@\\NS6>\\JV1B[NS6>[JW1CYNS6a0YJV1DYNT6a0VJW1GVNU6b0TJX1HUNT6d0SJW1JSNU6f0oIX1LRNU6g0nIW1NPNV6h0lIX1OoMU6j0kIW10nMW6k0gIX13kMW6n0eIW15jMW6o0cIW17hMX6P1aIX17gMY6R1^IX1:eMY6S1\\IX1<cMY6V1ZIW1>bMY6V1XIY1`0_MY6a4gI^K[6a4fI^KZ6b4gI\\K[6c4fI[K\\6c4fI\\KZ6d4fI[K\\6d4eIZK\\6f4eIYK\\6f4eIXK\\6h4T1O0O2O0O2N2N1O2O1N2N2N2N2N2O1N2O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O:Fc0\\O>jNgDAd;3\\DNg;NZD2e;OZD2f;M[D2e;N[D3d;M]D3b;M^D4b;L^D4a;L_D5`;KaD4_;LaD5_;JbD6];KbD6];JdD6[;JeD6\\;IeD7Z;IfD8Y;HhD8W;IhD8X;GiD8W;HiD9V;GkD9T;GlD:T;EmD;R;DoD<Q;DPE<o:CRE>n:ARE`0m:_OTEb0k:^OUEc0j:\\OXEc0i:[OXEe0h:[OXEe0h:ZOYEf0g:ZOYEf0h:XOYEh0g:XOYEh0g:WOZEi0f:WOZEi0g:UOZEk0f:TO\\Ek0d:UO\\Ek0e:SO\\Em0f:QOZEP1_;000000000001O0000000000000000001O0000000000000000001O001N2M2O2N2M3N2M3N2N2M2O2M3N2NT[U2"}, "label": "an elephant in front of another"}]}
{"id": 526523, "image": "COCO_train2014_000000526523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue spaghetti strap tank next to young girl in pink top\"."}], "task": "refering", "answer_template": "The \"woman in blue spaghetti strap tank next to young girl in pink top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 217, 218, 219, 220, 241, 242, 243, 264, 265], "bbox": [0.46120312500000005, 0.0006323185011709602, 0.693375, 0.763887587822014], "iscrowd": 0, "area": 31257.153150000006, "rle": {"size": [427, 640], "counts": "ZXk3=j<8I7I6I6M3M3M3M2N2O1N2O1N2O1O1jDWNg:Z2M3TMbMeJa2Q5jMjJ[2m4mMPKT2k4QNTKP2g4UNXKl1c4ZNlJUOYNb2g6]NlJV2o4nMnJV2m4oMnJV2l4PNjJY2R5kMgJ\\2T5jMiJX2S5lMkJV2Q5oMlJS2P5RNmJP2`3eKULMg0a2Bn1^3kKVLHg0a2Cm1j2^KQM1O>IGh0`2Cm1l2aKlLO2h0a0l1Bn1P3bKgLN4i0`0k1@P2T3fKjLb0`0i1Ao1V3fKhLc0`0i1An1X3kLVMU6j2kIlL\\N1i7T3lIhL_N1e7X3lIeLbN1c7Z3R21O1000000001O00001O001O001O001O0000O1000000O1000000N2M3N2N2N2N2N3M2N2N2O1N2N2O1N2N2M3M3N2M3K5J6M3N2O1N2O1N2N2N2N2N2G9M3N2M3N2N2M3N2O1O1O1[I_I_6b6_I`I9Oj5g6VJ[Ih5f6WJ[Ii5d6WJ]I2KY5i6dJ\\I2MY5h6dJ\\IJL315Mn4m6mJ_IOH24m4g6PK_IOFO9R5c6oJ_IO4R5c7QKjGk4W8SKjGm4Z80000O10O10O2K4M4L3M4L3M4L4kNjFoM[9o1kFhMZ9V2lFXMb9e2d0N5I8H6J6J8ZOl0TOlda2"}, "label": "woman in blue spaghetti strap tank next to young girl in pink top"}]}
{"id": 526523, "image": "COCO_train2014_000000526523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with full black tops\"."}], "task": "refering", "answer_template": "The \"a woman with full black tops\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 217, 218, 219, 220, 241, 242, 243, 264, 265], "bbox": [0.46120312500000005, 0.0006323185011709602, 0.693375, 0.763887587822014], "iscrowd": 0, "area": 31257.153150000006, "rle": {"size": [427, 640], "counts": "ZXk3=j<8I7I6I6M3M3M3M2N2O1N2O1N2O1O1jDWNg:Z2M3TMbMeJa2Q5jMjJ[2m4mMPKT2k4QNTKP2g4UNXKl1c4ZNlJUOYNb2g6]NlJV2o4nMnJV2m4oMnJV2l4PNjJY2R5kMgJ\\2T5jMiJX2S5lMkJV2Q5oMlJS2P5RNmJP2`3eKULMg0a2Bn1^3kKVLHg0a2Cm1j2^KQM1O>IGh0`2Cm1l2aKlLO2h0a0l1Bn1P3bKgLN4i0`0k1@P2T3fKjLb0`0i1Ao1V3fKhLc0`0i1An1X3kLVMU6j2kIlL\\N1i7T3lIhL_N1e7X3lIeLbN1c7Z3R21O1000000001O00001O001O001O001O0000O1000000O1000000N2M3N2N2N2N2N3M2N2N2O1N2N2O1N2N2M3M3N2M3K5J6M3N2O1N2O1N2N2N2N2N2G9M3N2M3N2N2M3N2O1O1O1[I_I_6b6_I`I9Oj5g6VJ[Ih5f6WJ[Ii5d6WJ]I2KY5i6dJ\\I2MY5h6dJ\\IJL315Mn4m6mJ_IOH24m4g6PK_IOFO9R5c6oJ_IO4R5c7QKjGk4W8SKjGm4Z80000O10O10O2K4M4L3M4L3M4L4kNjFoM[9o1kFhMZ9V2lFXMb9e2d0N5I8H6J6J8ZOl0TOlda2"}, "label": "a woman with full black tops"}]}
{"id": 526523, "image": "COCO_train2014_000000526523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in white eating something\"."}], "task": "refering", "answer_template": "The \"a little girl in white eating something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 78, 79, 80, 100, 101, 102, 103, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 263, 284], "bbox": [0.28871875, 0.10302107728337237, 0.51909375, 0.8092974238875879], "iscrowd": 0, "area": 21846.938949999996, "rle": {"size": [427, 640], "counts": "k]]22X=3M2N3M3M3M2N3M3M3M2N3M3M2N3M3M3M2O2M3N2M2O0O2N2O1N101M3N2N1N3N2N2M2O2M3N2N0O2O1N2O1N101O1N2oNRMdGo2Z8UMbGl2]8YM_Gh2_8\\M]Gf2a8^M\\Gc2b8aM[G`2c8dMYG]2f8hMVGY2h8V1N2O1N1O2N2N2N2N2TNlK_JU4\\5VL]Jl3`5[LZJg3c5aLUJb3i5dLoI`3n5o1N3M3N2O10O10O1000O0100O10O10O1000O010000O01000O1nNnHkKQ7U4S1O100O0010J5K5K6I6F:F;D;E;I8M2M4M3M3L4M201O100O1O100O100O1O100O1O101N1O1N2N2O1N2WMcHB_7;fHA\\7;iHCZ78kHDY77kHFY74mHHW72oHKT7JVI3n6B]I:g6_O_I>e6[ObIb0e9MkXP4"}, "label": "a little girl in white eating something"}]}
{"id": 526523, "image": "COCO_train2014_000000526523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child in magenta socks to the left of the woman\"."}], "task": "refering", "answer_template": "The \"the child in magenta socks to the left of the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 78, 79, 80, 100, 101, 102, 103, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 263, 284], "bbox": [0.28871875, 0.10302107728337237, 0.51909375, 0.8092974238875879], "iscrowd": 0, "area": 21846.938949999996, "rle": {"size": [427, 640], "counts": "k]]22X=3M2N3M3M3M2N3M3M3M2N3M3M2N3M3M3M2O2M3N2M2O0O2N2O1N101M3N2N1N3N2N2M2O2M3N2N0O2O1N2O1N101O1N2oNRMdGo2Z8UMbGl2]8YM_Gh2_8\\M]Gf2a8^M\\Gc2b8aM[G`2c8dMYG]2f8hMVGY2h8V1N2O1N1O2N2N2N2N2TNlK_JU4\\5VL]Jl3`5[LZJg3c5aLUJb3i5dLoI`3n5o1N3M3N2O10O10O1000O0100O10O10O1000O010000O01000O1nNnHkKQ7U4S1O100O0010J5K5K6I6F:F;D;E;I8M2M4M3M3L4M201O100O1O100O100O1O100O1O101N1O1N2N2O1N2WMcHB_7;fHA\\7;iHCZ78kHDY77kHFY74mHHW72oHKT7JVI3n6B]I:g6_O_I>e6[ObIb0e9MkXP4"}, "label": "the child in magenta socks to the left of the woman"}]}
{"id": 526523, "image": "COCO_train2014_000000526523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde haired girl eating food from her right hand\"."}], "task": "refering", "answer_template": "The \"a blonde haired girl eating food from her right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 28, 29, 30, 51, 52, 53, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 257], "bbox": [0.166421875, 0.04269320843091335, 0.36732812500000006, 0.7595550351288057], "iscrowd": 0, "area": 22843.996750000002, "rle": {"size": [427, 640], "counts": "Rj\\1c0g<3M2N3M3M2N3M3M3M2N3M3M3M2N3M7G<C<_ElMV9[3H8G9H7H7J6J6I7J6I7I3M4L3N2L5K4N2100O1OTNbInL^6R3gIiLX6X3nIbLQ6^3UJ]Lk5c3[JWLd5i3bJRL]5o3iJkKW5T4UKbKi4_4cKUK\\4m4P20O2N2O1N1O2O11N2O001N2O1O001N2O010O1O10O0100O0001O0O101O0O2O1O1N2O2M2O1O1N2O1O1N2O1N2O2N1N3N1YOh0XOh0K4K5L4M3N2M3N3L3N2M3NP1oN5L3L4K4J6J6J6K5K5K5K5K5K5K6I6J6J6J6H8E<DmhX5"}, "label": "a blonde haired girl eating food from her right hand"}]}
{"id": 526523, "image": "COCO_train2014_000000526523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a flowered top sits and eats a snack\"."}], "task": "refering", "answer_template": "The \"a girl in a flowered top sits and eats a snack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 28, 29, 30, 51, 52, 53, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 257], "bbox": [0.166421875, 0.04269320843091335, 0.36732812500000006, 0.7595550351288057], "iscrowd": 0, "area": 22843.996750000002, "rle": {"size": [427, 640], "counts": "Rj\\1c0g<3M2N3M3M2N3M3M3M2N3M3M3M2N3M7G<C<_ElMV9[3H8G9H7H7J6J6I7J6I7I3M4L3N2L5K4N2100O1OTNbInL^6R3gIiLX6X3nIbLQ6^3UJ]Lk5c3[JWLd5i3bJRL]5o3iJkKW5T4UKbKi4_4cKUK\\4m4P20O2N2O1N1O2O11N2O001N2O1O001N2O010O1O10O0100O0001O0O101O0O2O1O1N2O2M2O1O1N2O1O1N2O1N2O2N1N3N1YOh0XOh0K4K5L4M3N2M3N3L3N2M3NP1oN5L3L4K4J6J6J6K5K5K5K5K5K5K6I6J6J6J6H8E<DmhX5"}, "label": "a girl in a flowered top sits and eats a snack"}]}
{"id": 256190, "image": "COCO_train2014_000000256190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with a blue t - shirt and her friend with a black tank top on her right\"."}], "task": "refering", "answer_template": "The \"a lady with a blue t - shirt and her friend with a black tank top on her right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.360390625, 0.2181875, 0.75584375, 0.5298750000000001], "iscrowd": 0, "area": 19100.704999999994, "rle": {"size": [480, 640], "counts": "Xn\\3110l>3ONUA1i>OWA3g>MYA3g>4OJZA2e>O[A1e>6O0YOIeB1^O0l=3bBNB0k=5\\B6d=MUB8o<\\OiCl0XOYOn<NfCR1[<POaCR1^<QO^CP1c<ROZCf0GVOn<:TCj0m<e0O010O010O01O0100O010O11N8I0e0[OO0010O010O01O010O1O100OFgMnCY2P<kMnCT2Q<PNnCn1Q<UNPDi1n;ZNVD`1i;bNVD^1h;eNWD\\1g;fNXDZ1g;gNYDY1e;jNYDW1g;iNYDW1i;fNXDZ1j;bNWD_1k;^NTDd1n;XNSDi1o;TNQDm1P<QNhCY2o;mMRDT2f;SN[Dm1c;SN^Dn1d;oM\\DR2g;iMZDX2i;dMXD\\2k;_MTDd2o;53M3M3M2N3NkNT1TOl000O100O100O10000O100O10000O1000000000000001O000000001O1O1O2N1O1O1O2N1PLbE\\3_:[LkEc3l:M2N2N2N1O1O2N3M3M2N2N2N2N0000001O001O001O001O001O2N;E00000000000000O100O100000000000000000000000000000OK6M300000O1000001N10001O000O2O00001O1N1000MjM\\CV2e<12O1O001N2O1O001O1N2O001O1N2O001O1N101O0O2O0O2O002M2O1N3N1O1N3N0O10000O10000O3N001N100O100O2O000O100O2O0O101O1N2M3N2N1N3N3M5KWmX2"}, "label": "a lady with a blue t - shirt and her friend with a black tank top on her right"}]}
{"id": 256190, "image": "COCO_train2014_000000256190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table top and the woman in blue with glasses sitting in the middle\"."}], "task": "refering", "answer_template": "The \"the table top and the woman in blue with glasses sitting in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.360390625, 0.2181875, 0.75584375, 0.5298750000000001], "iscrowd": 0, "area": 19100.704999999994, "rle": {"size": [480, 640], "counts": "Xn\\3110l>3ONUA1i>OWA3g>MYA3g>4OJZA2e>O[A1e>6O0YOIeB1^O0l=3bBNB0k=5\\B6d=MUB8o<\\OiCl0XOYOn<NfCR1[<POaCR1^<QO^CP1c<ROZCf0GVOn<:TCj0m<e0O010O010O01O0100O010O11N8I0e0[OO0010O010O01O010O1O100OFgMnCY2P<kMnCT2Q<PNnCn1Q<UNPDi1n;ZNVD`1i;bNVD^1h;eNWD\\1g;fNXDZ1g;gNYDY1e;jNYDW1g;iNYDW1i;fNXDZ1j;bNWD_1k;^NTDd1n;XNSDi1o;TNQDm1P<QNhCY2o;mMRDT2f;SN[Dm1c;SN^Dn1d;oM\\DR2g;iMZDX2i;dMXD\\2k;_MTDd2o;53M3M3M2N3NkNT1TOl000O100O100O10000O100O10000O1000000000000001O000000001O1O1O2N1O1O1O2N1PLbE\\3_:[LkEc3l:M2N2N2N1O1O2N3M3M2N2N2N2N0000001O001O001O001O001O2N;E00000000000000O100O100000000000000000000000000000OK6M300000O1000001N10001O000O2O00001O1N1000MjM\\CV2e<12O1O001N2O1O001O1N2O001O1N2O001O1N101O0O2O0O2O002M2O1N3N1O1N3N0O10000O10000O3N001N100O100O2O000O100O2O0O101O1N2M3N2N1N3N3M5KWmX2"}, "label": "the table top and the woman in blue with glasses sitting in the middle"}]}
{"id": 116927, "image": "COCO_train2014_000000116927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in mid air on the slopes wearing a black coat\"."}], "task": "refering", "answer_template": "The \"a skier in mid air on the slopes wearing a black coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 96, 97, 98, 111, 112, 113, 114, 127, 128, 129, 142, 143, 144, 158], "bbox": [0.392887323943662, 0.23109375000000001, 0.632887323943662, 0.45318749999999997], "iscrowd": 0, "area": 6403.342100000002, "rle": {"size": [640, 426], "counts": "lUY35hc04L5L3L4N3M2O1O001O000O100000001O00000000011N2N2N2N1N3M2N3M2O2N101N2N4M3L00100O01000Q^OUNea0n1Y^OVNba0W2L6F9K5L5K4L2M3N2NjNd_OkNZ`0U1k_OhNR`0X1T@eNj?[1[@bNc?^1b@_N[?c1i@ZNV?f1n@WNQ?i1SASNm>m1WAWNa>j1bASN]>m1gAPNX>P2lAlMT>U2oAhMP>X2TBeMk=`2TB]Ml=h2c1mMR_OZ1j`0_N\\_Of1``0RNg_OR2Ra01O1N101O5K8G01O1O1O001O1O1O1N101O1iNd^OH]a07g^OEYa0:k^OCVa0<n^O@Sa0>Q_O_OPa0`0T_O\\Om`0b0W_OYOl`0f0S_OVOTa0g0T1N2Dn\\OCZc05>HaYQ3"}, "label": "a skier in mid air on the slopes wearing a black coat"}]}
{"id": 116927, "image": "COCO_train2014_000000116927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in the air making an x with his skis\"."}], "task": "refering", "answer_template": "The \"a skier in the air making an x with his skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 96, 97, 98, 111, 112, 113, 114, 127, 128, 129, 142, 143, 144, 158], "bbox": [0.392887323943662, 0.23109375000000001, 0.632887323943662, 0.45318749999999997], "iscrowd": 0, "area": 6403.342100000002, "rle": {"size": [640, 426], "counts": "lUY35hc04L5L3L4N3M2O1O001O000O100000001O00000000011N2N2N2N1N3M2N3M2O2N101N2N4M3L00100O01000Q^OUNea0n1Y^OVNba0W2L6F9K5L5K4L2M3N2NjNd_OkNZ`0U1k_OhNR`0X1T@eNj?[1[@bNc?^1b@_N[?c1i@ZNV?f1n@WNQ?i1SASNm>m1WAWNa>j1bASN]>m1gAPNX>P2lAlMT>U2oAhMP>X2TBeMk=`2TB]Ml=h2c1mMR_OZ1j`0_N\\_Of1``0RNg_OR2Ra01O1N101O5K8G01O1O1O001O1O1O1N101O1iNd^OH]a07g^OEYa0:k^OCVa0<n^O@Sa0>Q_O_OPa0`0T_O\\Om`0b0W_OYOl`0f0S_OVOTa0g0T1N2Dn\\OCZc05>HaYQ3"}, "label": "a skier in the air making an x with his skis"}]}
{"id": 559301, "image": "COCO_train2014_000000559301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large cow stands next to a calf\"."}], "task": "refering", "answer_template": "The \"a large cow stands next to a calf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 246, 247, 269, 270], "bbox": [0.46187500000000004, 0.29785547785547783, 0.853265625, 0.7631934731934731], "iscrowd": 0, "area": 27832.067550000007, "rle": {"size": [429, 640], "counts": "oXl32V=3M4L3N2M3N2M3N2N2N2N2O1TN[ORGe0o8M^Fc0R92YFd0Q9W2E4K2O1N2O1O1O1O1O1O100OE`GjK`8o3mGlKR8m3j0N3M2N2O1N2N3M3M3M3N2M4O01O1O100O1O2N2N2M;F4L3L5L3M4K4M3M2N2M3N2N5L6J1N01O100O1hNRHoLo7j2ZHSMf7h2bHUM^7Z2UIeMj6\\2XIaMh6_2[I^Me6c2]IZMc6g2_IVMa6j2`IVM_6k2bISM_6l2cIRM]6o2dIoL\\6Q3fImLZ6T3fIlLY6U3hIiLX6W3jIgLV6Z3n1001O1O0010O01O00000001O01O00000000010O00000000010O000000000001O00000000001N10000000001O0000000000001O0O100000001O0000000O1000001O000000000O11O101M5L3M4L3L5dGWLl6m3nHULR7n3iHTLW7P4dHQL[7S4`HoK`7V4YHlKg7n4N10000000001O00000000000@lJgHU5V7QKfHo4Z7RKeHn4[7SKdHm4]7SKbHm4^7TKbHk4_7UK`Hl4`7`0001O0TO_HlKb7k3lHoKT7j3ZIPLf6g3hISLX6g3TJSLm5j3VJULk5i3WJVLh5j3m100O101O000O1000001O0O10000F:E;N3O2M3M2O2M3M2O2M2O2M3M2O2N2M2O2N2M2O2N1N3N1H7I8G8I8O001O001O1O3M2KnWW1"}, "label": "a large cow stands next to a calf"}]}
{"id": 559301, "image": "COCO_train2014_000000559301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown cow standing in the grass\"."}], "task": "refering", "answer_template": "The \"a large brown cow standing in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 246, 247, 269, 270], "bbox": [0.46187500000000004, 0.29785547785547783, 0.853265625, 0.7631934731934731], "iscrowd": 0, "area": 27832.067550000007, "rle": {"size": [429, 640], "counts": "oXl32V=3M4L3N2M3N2M3N2N2N2N2O1TN[ORGe0o8M^Fc0R92YFd0Q9W2E4K2O1N2O1O1O1O1O1O100OE`GjK`8o3mGlKR8m3j0N3M2N2O1N2N3M3M3M3N2M4O01O1O100O1O2N2N2M;F4L3L5L3M4K4M3M2N2M3N2N5L6J1N01O100O1hNRHoLo7j2ZHSMf7h2bHUM^7Z2UIeMj6\\2XIaMh6_2[I^Me6c2]IZMc6g2_IVMa6j2`IVM_6k2bISM_6l2cIRM]6o2dIoL\\6Q3fImLZ6T3fIlLY6U3hIiLX6W3jIgLV6Z3n1001O1O0010O01O00000001O01O00000000010O00000000010O000000000001O00000000001N10000000001O0000000000001O0O100000001O0000000O1000001O000000000O11O101M5L3M4L3L5dGWLl6m3nHULR7n3iHTLW7P4dHQL[7S4`HoK`7V4YHlKg7n4N10000000001O00000000000@lJgHU5V7QKfHo4Z7RKeHn4[7SKdHm4]7SKbHm4^7TKbHk4_7UK`Hl4`7`0001O0TO_HlKb7k3lHoKT7j3ZIPLf6g3hISLX6g3TJSLm5j3VJULk5i3WJVLh5j3m100O101O000O1000001O0O10000F:E;N3O2M3M2O2M3M2O2M2O2M3M2O2N2M2O2N2M2O2N1N3N1H7I8G8I8O001O001O1O3M2KnWW1"}, "label": "a large brown cow standing in the grass"}]}
{"id": 452806, "image": "COCO_train2014_000000452806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white table holding plates\"."}], "task": "refering", "answer_template": "The \"a small white table holding plates\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [318, 319, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 386, 387], "bbox": [0.7065, 0.776625, 1.0, 0.984421875], "iscrowd": 0, "area": 8034.700500000002, "rle": {"size": [640, 480], "counts": "l[d61lc04M4L3M4L301O1O001O001O001O001f_OZOW=f0aBA`=`0WBGj=9oAMR>4eA3[>N^A8c>ITA>m>Bk@e0V?\\Ob@j0_?WOX@P1i?ROm_OU1T`0lNd_OZ1]`0T11O1O1O1O1O1O1O1O00001O0000001O00001O00001O00001O0L4K6J5K5L5J5K5L5J5K5K6L31O01O01O001O01O01O0010O0001O010O00001O010O000010O01O00010O001O00010O001O01O01O001O01O01O0010O0001O010O00001O010O000010O01O00010O001O00010O001O01O01OG:^Oa0_OX_O"}, "label": "a small white table holding plates"}]}
{"id": 452806, "image": "COCO_train2014_000000452806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white table\"."}], "task": "refering", "answer_template": "The \"the white table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [318, 319, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 386, 387], "bbox": [0.7065, 0.776625, 1.0, 0.984421875], "iscrowd": 0, "area": 8034.700500000002, "rle": {"size": [640, 480], "counts": "l[d61lc04M4L3M4L301O1O001O001O001O001f_OZOW=f0aBA`=`0WBGj=9oAMR>4eA3[>N^A8c>ITA>m>Bk@e0V?\\Ob@j0_?WOX@P1i?ROm_OU1T`0lNd_OZ1]`0T11O1O1O1O1O1O1O1O00001O0000001O00001O00001O00001O0L4K6J5K5L5J5K5L5J5K5K6L31O01O01O001O01O01O0010O0001O010O00001O010O000010O01O00010O001O00010O001O01O01O001O01O01O0010O0001O010O00001O010O000010O01O00010O001O00010O001O01O01OG:^Oa0_OX_O"}, "label": "the white table"}]}
{"id": 59593, "image": "COCO_train2014_000000059593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in ballcap on floor watching tv\"."}], "task": "refering", "answer_template": "The \"man in ballcap on floor watching tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.12168749999999999, 0.46230046948356807, 0.5752499999999999, 1.0], "iscrowd": 0, "area": 36635.84889999999, "rle": {"size": [426, 640], "counts": "bkP13R=5K5L4N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O100O100O100O100O100O100O10000001O001O001O001O001O00DgLhFX3V9RMbFn2\\9a0N2N2O1N2N2N2N2N2N2N2N2N2N2O1N2O1O10000O100O10000O100O100O10000O100O10000O100000000000000000000O100000000000000000000SOXKWIi4f6[KYIe4f6\\KYIe4g6\\KXId4h6]KVId4i6]KWIc4g6`KWIa4e6dKWI_4d6gKXI\\4d6W1K5M300O100O100O100O100O100O100O1O100O100O1000000001O000000000000001O0000000000001O002N1O1lJhIT3Y6gLkIY3V6eLlIZ3U6eLkIZ3W6dLjIZ3Y6eLhIW3\\6gLeIW3^6gLdIU3`6jL`IS3d6kL^IQ3f6mL[IQ3h6nLYIn2_7\\LbHa3a7YLeHd3^7VLiHg3^8M4L3M4M2M3M4M2M4L3M3N3O0010O100O3N2M3N2M3N100O100O100O101N101N2O1N2O1N101N2OO0O2N1O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1nMaEW1a:fNcEW1^:gNfEV1\\:gNgEW1[:fNiEW1X:gNkEW1W;M3M2N3M3M2N5K5K5K5Kjg`3"}, "label": "man in ballcap on floor watching tv"}]}
{"id": 59593, "image": "COCO_train2014_000000059593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black hat and a beige shirt sits next to a girl playing wii\"."}], "task": "refering", "answer_template": "The \"a man with a black hat and a beige shirt sits next to a girl playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.12168749999999999, 0.46230046948356807, 0.5752499999999999, 1.0], "iscrowd": 0, "area": 36635.84889999999, "rle": {"size": [426, 640], "counts": "bkP13R=5K5L4N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O100O100O100O100O100O100O10000001O001O001O001O001O00DgLhFX3V9RMbFn2\\9a0N2N2O1N2N2N2N2N2N2N2N2N2N2O1N2O1O10000O100O10000O100O100O10000O100O10000O100000000000000000000O100000000000000000000SOXKWIi4f6[KYIe4f6\\KYIe4g6\\KXId4h6]KVId4i6]KWIc4g6`KWIa4e6dKWI_4d6gKXI\\4d6W1K5M300O100O100O100O100O100O100O1O100O100O1000000001O000000000000001O0000000000001O002N1O1lJhIT3Y6gLkIY3V6eLlIZ3U6eLkIZ3W6dLjIZ3Y6eLhIW3\\6gLeIW3^6gLdIU3`6jL`IS3d6kL^IQ3f6mL[IQ3h6nLYIn2_7\\LbHa3a7YLeHd3^7VLiHg3^8M4L3M4M2M3M4M2M4L3M3N3O0010O100O3N2M3N2M3N100O100O100O101N101N2O1N2O1N101N2OO0O2N1O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1nMaEW1a:fNcEW1^:gNfEV1\\:gNgEW1[:fNiEW1X:gNkEW1W;M3M2N3M3M2N5K5K5K5Kjg`3"}, "label": "a man with a black hat and a beige shirt sits next to a girl playing wii"}]}
{"id": 59593, "image": "COCO_train2014_000000059593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mother sitting on a chair watching her daughter play video games\"."}], "task": "refering", "answer_template": "The \"mother sitting on a chair watching her daughter play video games\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 136, 137, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 269, 270, 271, 272, 273, 294, 295, 296, 317, 318, 319], "bbox": [0.69703125, 0.24269953051643192, 1.0, 0.9101173708920187], "iscrowd": 0, "area": 21872.267849999997, "rle": {"size": [426, 640], "counts": "iji51Y=1O2kBOl<9O0O2O001N101O2N2M1000000O10O1000O1O100O1O100O100O1O100O1O1O1N2O1O0O2O1O1N2O1O1N2O1O1O1N2N2O0O1O1O100ODaDPO^;R1eDjN[;W1gDfNX;[1:2N101N2N2N2oFYN]6h1cIZNZ6h1fIYNW6i1hIYNU6i1jIXNT6j1kIXNR6j1nIWNo5k1PJWNm5T2iInMT6e2ZI]Mc6W3kHjLS7m3VHTLh7i4NNTHhJk7Y5UHgJi7[5XHdJg7]5YHcJf7^5ZHbJe7_53O0100O10000O010O100O1000O0100O100O010O10000O10O010TM]H1b70dHJR7?UI[Ok6e0[IUOd6k0dImN\\6T1jIfNU6Z1RJ`Nn5`1XJYNh5h1^JQNb5o1fJiMZ5X2lJaMU5^2RK[Mn4f2XKSMh4m2_KlLa4U3j2O01O00100O0010O01O00100O1O1O101N1O1mNbLZH^3b7gL]HZ3a7hL]HY3b7iL]HW3b7kL]HU3b7mL\\HU3b7mL]HS3b7PMhGd3W8h0O1O2N1O1O1O1O2N1O1O10000O10000O100O10000O100O10001N11O1O001O1O001O001O1O001O1O001O10O01O0TOYM"}, "label": "mother sitting on a chair watching her daughter play video games"}]}
{"id": 59593, "image": "COCO_train2014_000000059593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in black shirt looking bored\"."}], "task": "refering", "answer_template": "The \"woman in black shirt looking bored\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 136, 137, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 269, 270, 271, 272, 273, 294, 295, 296, 317, 318, 319], "bbox": [0.69703125, 0.24269953051643192, 1.0, 0.9101173708920187], "iscrowd": 0, "area": 21872.267849999997, "rle": {"size": [426, 640], "counts": "iji51Y=1O2kBOl<9O0O2O001N101O2N2M1000000O10O1000O1O100O1O100O100O1O100O1O1O1N2O1O0O2O1O1N2O1O1N2O1O1O1N2N2O0O1O1O100ODaDPO^;R1eDjN[;W1gDfNX;[1:2N101N2N2N2oFYN]6h1cIZNZ6h1fIYNW6i1hIYNU6i1jIXNT6j1kIXNR6j1nIWNo5k1PJWNm5T2iInMT6e2ZI]Mc6W3kHjLS7m3VHTLh7i4NNTHhJk7Y5UHgJi7[5XHdJg7]5YHcJf7^5ZHbJe7_53O0100O10000O010O100O1000O0100O100O010O10000O10O010TM]H1b70dHJR7?UI[Ok6e0[IUOd6k0dImN\\6T1jIfNU6Z1RJ`Nn5`1XJYNh5h1^JQNb5o1fJiMZ5X2lJaMU5^2RK[Mn4f2XKSMh4m2_KlLa4U3j2O01O00100O0010O01O00100O1O1O101N1O1mNbLZH^3b7gL]HZ3a7hL]HY3b7iL]HW3b7kL]HU3b7mL\\HU3b7mL]HS3b7PMhGd3W8h0O1O2N1O1O1O1O2N1O1O10000O10000O100O10000O100O10001N11O1O001O1O001O001O1O001O1O001O10O01O0TOYM"}, "label": "woman in black shirt looking bored"}]}
{"id": 59593, "image": "COCO_train2014_000000059593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a brown outfit holding a game controller\"."}], "task": "refering", "answer_template": "The \"a girl in a brown outfit holding a game controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 243, 244, 266, 267, 289, 290, 312, 313, 335, 336], "bbox": [0.50984375, 0.4022065727699531, 0.65490625, 1.0], "iscrowd": 0, "area": 12587.6284, "rle": {"size": [426, 640], "counts": "VkW41S=8M201WCMW<d0O100O1O100N2N2M2O2N2O1O1O1O0001N101N2N2O1N2NnNoNnEP1o9[OjEc0U:FeE8Y:NeE1Z:1eENZ:5cEM\\:6aEJ_:8^EI<AQ9h0aFHJ5c96_Fn0_9TO_Fn0_9`1N1O2VJTLR2o3lMTLQ2n3mMULP2m3oMULo1l3eLQLQO6X4i3cLYLPOO]4i3XKmK?d0GFb4j3PKRN8UNh4j3hJ\\N;kMm4f6nJZIR5j6iJWIW5^7000000000000O100001O001O00001O001O001O001_IaJU5_5hJeJW5\\5eJgJ[5Y5cJjJ\\5V5aJmJ_5T5^JoJa5R5\\JPKd5P5YJTKf5m4VJVKj5j4TJYKk5h4QJ[Kb3POQN1`Nd5J_K_3WOQNGeNd5GaK^3_OoM_OhNc5IbKY3HPNUOjNQ7`1]IUOg0;nNmNP7o0QJA<=P6MlI\\O`0a0e5M`LMb3NhLLY3NRMKQ3Omkk2"}, "label": "a girl in a brown outfit holding a game controller"}]}
{"id": 59593, "image": "COCO_train2014_000000059593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl dressed in brown playing a video game\"."}], "task": "refering", "answer_template": "The \"a little girl dressed in brown playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 243, 244, 266, 267, 289, 290, 312, 313, 335, 336], "bbox": [0.50984375, 0.4022065727699531, 0.65490625, 1.0], "iscrowd": 0, "area": 12587.6284, "rle": {"size": [426, 640], "counts": "VkW41S=8M201WCMW<d0O100O1O100N2N2M2O2N2O1O1O1O0001N101N2N2O1N2NnNoNnEP1o9[OjEc0U:FeE8Y:NeE1Z:1eENZ:5cEM\\:6aEJ_:8^EI<AQ9h0aFHJ5c96_Fn0_9TO_Fn0_9`1N1O2VJTLR2o3lMTLQ2n3mMULP2m3oMULo1l3eLQLQO6X4i3cLYLPOO]4i3XKmK?d0GFb4j3PKRN8UNh4j3hJ\\N;kMm4f6nJZIR5j6iJWIW5^7000000000000O100001O001O00001O001O001O001_IaJU5_5hJeJW5\\5eJgJ[5Y5cJjJ\\5V5aJmJ_5T5^JoJa5R5\\JPKd5P5YJTKf5m4VJVKj5j4TJYKk5h4QJ[Kb3POQN1`Nd5J_K_3WOQNGeNd5GaK^3_OoM_OhNc5IbKY3HPNUOjNQ7`1]IUOg0;nNmNP7o0QJA<=P6MlI\\O`0a0e5M`LMb3NhLLY3NRMKQ3Omkk2"}, "label": "a little girl dressed in brown playing a video game"}]}
{"id": 321737, "image": "COCO_train2014_000000321737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra bear a person with a green hat\"."}], "task": "refering", "answer_template": "The \"a zebra bear a person with a green hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 83, 84, 85, 86, 103, 104, 105, 106, 124, 125, 126, 144, 145, 146, 164, 165, 166], "bbox": [0.15409090909090908, 0.26294277929155313, 0.34505454545454545, 0.6776566757493188], "iscrowd": 0, "area": 7866.516050000001, "rle": {"size": [367, 550], "counts": "\\cn07k92SG9g8LoF=m8ObF9\\9n0O1N0100000001N101O0O2N1N3M3M2011N2O1N2O1N2O1N3N2M4WHVNf5n1gI`NY6f1nHmNS7f2101O001H8H7H9G8I8G8M4N101N1O20O01O010O000001O00001O4L=C=C;E0000O1J6^Oc0]Ob0N2000000000000001O001O001O1O001O1O1O1O1O1O1O1]G\\NY7i1]H_N`7j2N01iMfH6[7_OSI=n6YO_IXOB<P71kI]Od7?h1N2N2N2N3M2IUmP4"}, "label": "a zebra bear a person with a green hat"}]}
{"id": 321737, "image": "COCO_train2014_000000321737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra directly behind the one in front\"."}], "task": "refering", "answer_template": "The \"zebra directly behind the one in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 83, 84, 85, 86, 103, 104, 105, 106, 124, 125, 126, 144, 145, 146, 164, 165, 166], "bbox": [0.15409090909090908, 0.26294277929155313, 0.34505454545454545, 0.6776566757493188], "iscrowd": 0, "area": 7866.516050000001, "rle": {"size": [367, 550], "counts": "\\cn07k92SG9g8LoF=m8ObF9\\9n0O1N0100000001N101O0O2N1N3M3M2011N2O1N2O1N2O1N3N2M4WHVNf5n1gI`NY6f1nHmNS7f2101O001H8H7H9G8I8G8M4N101N1O20O01O010O000001O00001O4L=C=C;E0000O1J6^Oc0]Ob0N2000000000000001O001O001O1O001O1O1O1O1O1O1O1]G\\NY7i1]H_N`7j2N01iMfH6[7_OSI=n6YO_IXOB<P71kI]Od7?h1N2N2N2N3M2IUmP4"}, "label": "zebra directly behind the one in front"}]}
{"id": 321737, "image": "COCO_train2014_000000321737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra , almost head - on , shown with other zebras but appears closer to the camera\"."}], "task": "refering", "answer_template": "The \"a zebra , almost head - on , shown with other zebras but appears closer to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 52, 70, 71, 72, 86, 87, 88, 89, 90, 91, 92, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 206, 207, 208, 209, 210, 211, 228, 229, 230, 231], "bbox": [0.32989090909090907, 0.18201634877384196, 0.6462727272727272, 0.9258310626702998], "iscrowd": 0, "area": 22950.74965, "rle": {"size": [367, 550], "counts": "ePQ24W;m0RO8I7jGYOY5o0PJ5_50WJ?^5FYJ`0f5DQJc0m5AkIe0R6^2N1O2M3N2N1N3NO2G9fNWJfLP6S3[JcLl5U3_J`Li5X3Z1I8M2O2O1O001001N101O1N101O0O2O1O0010O01O10O01O00100O001O10O01O00010O0iHjLU6W3fISMT6n2fI]MU6b2gIgMi0nNS4\\3nJQNb0mN^4m4UK]Kh4g5O2N2M3N2MO2O1O0O10000XNaK_L`4Q3WLhLi3g2nLSMR3b2_MWMb2d2kMTMU2k2XNiLg1X3X300000000O100000O1000O10000000O1O010O100O1O100O010O1O100O1O102gHfLk4b0RKl3j4XLQKm3j4YLoJm3m4_1L4oK\\Io2g6hL^IY3e6]L`If3T72N2M3N2NO001O1OoN[LXJe3f3QL`Mg0cNW3j3YL[Mk0eNk2m3aLUMP5h2ZKlLk4Q3aKbLc4[3j1M3D<I7N2N23M3M2N3M2N1O1O1O1O1O1O0000000N101O1O1O1N2O1O1O1N101O001O0O1002jLZKAf49^KHd41dKL]4OjKNn7N_fU2"}, "label": "a zebra , almost head - on , shown with other zebras but appears closer to the camera"}]}
{"id": 321737, "image": "COCO_train2014_000000321737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a statue of a zebra staring straight ahead\"."}], "task": "refering", "answer_template": "The \"a statue of a zebra staring straight ahead\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 52, 70, 71, 72, 86, 87, 88, 89, 90, 91, 92, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 206, 207, 208, 209, 210, 211, 228, 229, 230, 231], "bbox": [0.32989090909090907, 0.18201634877384196, 0.6462727272727272, 0.9258310626702998], "iscrowd": 0, "area": 22950.74965, "rle": {"size": [367, 550], "counts": "ePQ24W;m0RO8I7jGYOY5o0PJ5_50WJ?^5FYJ`0f5DQJc0m5AkIe0R6^2N1O2M3N2N1N3NO2G9fNWJfLP6S3[JcLl5U3_J`Li5X3Z1I8M2O2O1O001001N101O1N101O0O2O1O0010O01O10O01O00100O001O10O01O00010O0iHjLU6W3fISMT6n2fI]MU6b2gIgMi0nNS4\\3nJQNb0mN^4m4UK]Kh4g5O2N2M3N2MO2O1O0O10000XNaK_L`4Q3WLhLi3g2nLSMR3b2_MWMb2d2kMTMU2k2XNiLg1X3X300000000O100000O1000O10000000O1O010O100O1O100O010O1O100O1O102gHfLk4b0RKl3j4XLQKm3j4YLoJm3m4_1L4oK\\Io2g6hL^IY3e6]L`If3T72N2M3N2NO001O1OoN[LXJe3f3QL`Mg0cNW3j3YL[Mk0eNk2m3aLUMP5h2ZKlLk4Q3aKbLc4[3j1M3D<I7N2N23M3M2N3M2N1O1O1O1O1O1O0000000N101O1O1O1N2O1O1O1N101O001O0O1002jLZKAf49^KHd41dKL]4OjKNn7N_fU2"}, "label": "a statue of a zebra staring straight ahead"}]}
{"id": 362699, "image": "COCO_train2014_000000362699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl jumping up to swing her racket , as seen on the left side of duplicate image\"."}], "task": "refering", "answer_template": "The \"girl jumping up to swing her racket , as seen on the left side of duplicate image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 97, 98, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 167, 168, 190, 191], "bbox": [0.16328125, 0.046156249999999996, 0.396390625, 0.77590625], "iscrowd": 0, "area": 9513.621550000003, "rle": {"size": [320, 640], "counts": "[PQ16c98I7L4L4J5L3100O10O0100O1M3M4L3L4O100000001O4K100000O10O10O100000O01000000XNLcI5b3BbMi0eNDj3GVMR1jNWOP4KlLZ1mNlNW4JfLd1oNbNZ4K`Ll1TOYN\\4KULX2^OlM^4LiKd2GaM`4L]KP31TMa4M]KQ31RMb4M\\KT3OPMe4L[KV3OmLg4MYKX3OkLg4NYKZ3NhLi4NXK\\3MgLk4MWK_3LdLm4MVKR4i4PLUKP4j4d0O1O1O1O0O2O10O1O1N101N2N1O2N2N2N2N2N2]NiJ_NV5[OiJQ17En4XOmJn0;Kf4VOQKi0`01]4TOUKf0d07U4QOYKc0h0=m3oN\\K?n0b0d3mN`K<R1l0Y3eNgK:X1U1j2^NQL8]1]1\\2XNZL6a1g1iNmMS21hM8a1n1YNmM]35]NQ2kMoMh30`Nd2`1\\MbNc2]1\\MgNa2Y1_MjN^2V1bMmN[2S1eMQOX2o0gMUOU2k0kMYOQ2g0oM]Om1c0SNAi1?VNF1bLS1g3lN>P1BPOb0l0^OTOf0h0ZOXOj0e0UO[OX18hNHX18hNHX18hNHX19gNFZ1KkJ9k3LZ1KkJ9k3LZ1JmJ9i3MZ1JnJ9g3M[1IPK9e3N[1IQK8d3O[1HRK9c3O[1HSK9a3N]1HTK5GLh37]1HUK3HMf38k1I[N7g500000000O100000000000000O11O1O1O1O1O1O1O1000O0010O01O1L3L]lh3"}, "label": "girl jumping up to swing her racket , as seen on the left side of duplicate image"}]}
{"id": 362699, "image": "COCO_train2014_000000362699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman who is jumping\"."}], "task": "refering", "answer_template": "The \"the woman who is jumping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 97, 98, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 167, 168, 190, 191], "bbox": [0.16328125, 0.046156249999999996, 0.396390625, 0.77590625], "iscrowd": 0, "area": 9513.621550000003, "rle": {"size": [320, 640], "counts": "[PQ16c98I7L4L4J5L3100O10O0100O1M3M4L3L4O100000001O4K100000O10O10O100000O01000000XNLcI5b3BbMi0eNDj3GVMR1jNWOP4KlLZ1mNlNW4JfLd1oNbNZ4K`Ll1TOYN\\4KULX2^OlM^4LiKd2GaM`4L]KP31TMa4M]KQ31RMb4M\\KT3OPMe4L[KV3OmLg4MYKX3OkLg4NYKZ3NhLi4NXK\\3MgLk4MWK_3LdLm4MVKR4i4PLUKP4j4d0O1O1O1O0O2O10O1O1N101N2N1O2N2N2N2N2N2]NiJ_NV5[OiJQ17En4XOmJn0;Kf4VOQKi0`01]4TOUKf0d07U4QOYKc0h0=m3oN\\K?n0b0d3mN`K<R1l0Y3eNgK:X1U1j2^NQL8]1]1\\2XNZL6a1g1iNmMS21hM8a1n1YNmM]35]NQ2kMoMh30`Nd2`1\\MbNc2]1\\MgNa2Y1_MjN^2V1bMmN[2S1eMQOX2o0gMUOU2k0kMYOQ2g0oM]Om1c0SNAi1?VNF1bLS1g3lN>P1BPOb0l0^OTOf0h0ZOXOj0e0UO[OX18hNHX18hNHX18hNHX19gNFZ1KkJ9k3LZ1KkJ9k3LZ1JmJ9i3MZ1JnJ9g3M[1IPK9e3N[1IQK8d3O[1HRK9c3O[1HSK9a3N]1HTK5GLh37]1HUK3HMf38k1I[N7g500000000O100000000000000O11O1O1O1O1O1O1O1000O0010O01O1L3L]lh3"}, "label": "the woman who is jumping"}]}
{"id": 362699, "image": "COCO_train2014_000000362699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the left picture with a yellow racket\"."}], "task": "refering", "answer_template": "The \"the girl in the left picture with a yellow racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 47, 48, 70, 71, 93, 94, 95, 116, 117, 139, 140, 141, 162, 163, 185, 186, 208, 209, 231, 232], "bbox": [0.03828125, 0.09234375, 0.16215625, 0.97071875], "iscrowd": 0, "area": 11742.393999999998, "rle": {"size": [320, 640], "counts": "[S84i95K5K4ZKP1VNVO\\OBQOQ2]1`N]OKaNT2n1TN\\O4XNP2X2oM\\Oa3`0bL_O`3=cLB_39eLE^38eLF]36fLI\\34fLJ]32fLM]3OeL0]3LfL2^3JdL5`3FbL9b3AaL=d3^O^La0a60O01OnNjN\\IU1e6QOUIn0l6SOSIm0m6TOQIm0o6TOPIl0P7UOoHk0Q7WOmHi0S7XOkHi0U7XOjHh0V7YOiHg0W7ZOhHf0X7\\OfHd0Z7]OeHc0[7^OdHb0_6WOcI8Na0[6@aI04`0Y6D`IM6`0X6c0gI]OW6e0iI[OV6h0hIXOV6k0iIUOU6n0iISOU6]2\\KXLj06@d3EYLe0<@]3IZLa0?BY3L[L<a0FU3M]L:a0ET3ObL7;HT30fL48HT3YOQL4i0f03KU3XOTL3i0f00KX3VOjMj0PONZ3SOjMm0nNLh52ZJJi55XJHj58XJDk5;VJBm5=UJ_On5`0SJ]OP6c0QJYOQ6g0e1O1O010O001O0010ON5K4K5L4Lk[W5"}, "label": "the girl in the left picture with a yellow racket"}]}
{"id": 362699, "image": "COCO_train2014_000000362699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl on the left in the picture on the left\"."}], "task": "refering", "answer_template": "The \"the girl on the left in the picture on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 47, 48, 70, 71, 93, 94, 95, 116, 117, 139, 140, 141, 162, 163, 185, 186, 208, 209, 231, 232], "bbox": [0.03828125, 0.09234375, 0.16215625, 0.97071875], "iscrowd": 0, "area": 11742.393999999998, "rle": {"size": [320, 640], "counts": "[S84i95K5K4ZKP1VNVO\\OBQOQ2]1`N]OKaNT2n1TN\\O4XNP2X2oM\\Oa3`0bL_O`3=cLB_39eLE^38eLF]36fLI\\34fLJ]32fLM]3OeL0]3LfL2^3JdL5`3FbL9b3AaL=d3^O^La0a60O01OnNjN\\IU1e6QOUIn0l6SOSIm0m6TOQIm0o6TOPIl0P7UOoHk0Q7WOmHi0S7XOkHi0U7XOjHh0V7YOiHg0W7ZOhHf0X7\\OfHd0Z7]OeHc0[7^OdHb0_6WOcI8Na0[6@aI04`0Y6D`IM6`0X6c0gI]OW6e0iI[OV6h0hIXOV6k0iIUOU6n0iISOU6]2\\KXLj06@d3EYLe0<@]3IZLa0?BY3L[L<a0FU3M]L:a0ET3ObL7;HT30fL48HT3YOQL4i0f03KU3XOTL3i0f00KX3VOjMj0PONZ3SOjMm0nNLh52ZJJi55XJHj58XJDk5;VJBm5=UJ_On5`0SJ]OP6c0QJYOQ6g0e1O1O010O001O0010ON5K4K5L4Lk[W5"}, "label": "the girl on the left in the picture on the left"}]}
{"id": 215243, "image": "COCO_train2014_000000215243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller horse\"."}], "task": "refering", "answer_template": "The \"smaller horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 175, 176, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 265, 266, 267, 269, 270, 271, 288, 289, 290, 292, 293, 294, 311, 312, 313, 315, 316], "bbox": [0.55321875, 0.39408333333333334, 0.8887031249999999, 0.8155833333333333], "iscrowd": 0, "area": 18957.101249999985, "rle": {"size": [480, 640], "counts": "UZV51m>2UN3`D1T;=gDEW;a0bDB];b0^D@a;e0QDEn;g1N2O1O1aN^M_Fc2^9bM^F`2_9eM^F\\2`9iM\\FX2a9mM\\FT2a9QNXFT2e9RNSFS2]9_N\\F4U;i14M3M2M4M3L3N2N101O1N101ORD`M^;^2`DoMX1]O]8d2XF[NP1TOh8`2UFhNg0lNT9P4cFSL\\9U4ZFnKf9f400O100O10O1O0N2C>B=J7J5K5L5J5N3N2N2M4M2N2M3N2N2N2O1N2O1N3N1N2N2O1000001O001N2O001O001O1N110O10O00010O001O00001O001O00001O001O0O1gEkMg7V2THnMk7S2QHQNo7R2jGRNU8f2oF_MQ9X4O01000ON3M2N3N1N3M2N3M2O2M2N3mNS1O001O001OnEVMP8i2PHZMm7f2SH\\Ml7d2SH^Ml7d2QH^Mm7e2PH]Mo7e2nG]MP8f2mG\\MR8f2kG\\MT8f2iG[MV8h2gGYMY8i2cGYM\\8j2aGWM_8k2ZGZMe8i2SG]Mm8e2lF`MS9c2fFbMY9S4ORMcF7a9HdFO`91dFG`99eF^O`9a0dFXO_9h0eFPO_9o0gFhN^9W1fFaN_9]1gFZN]9e1iFSN[9l1jFkM[9P2m1H8L4O1O1O001O1O1O1O1O100O001O1O11O1O001O1N2N101N2O1N101N2N2O0O2O1N2O0O2N2O1N2OoM[Ch1c<UNaCl1^<PNeCS2c<1N2O2M2OQNcC[1\\<dNjCW1V<hNPDT1P<jNVDR1i;lN^DP1b;nNeDl0[;SOkDi0U;UOQEg0h<L4L4K6K5KcPQ1"}, "label": "smaller horse"}]}
{"id": 215243, "image": "COCO_train2014_000000215243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby mule\"."}], "task": "refering", "answer_template": "The \"a baby mule\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 175, 176, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 265, 266, 267, 269, 270, 271, 288, 289, 290, 292, 293, 294, 311, 312, 313, 315, 316], "bbox": [0.55321875, 0.39408333333333334, 0.8887031249999999, 0.8155833333333333], "iscrowd": 0, "area": 18957.101249999985, "rle": {"size": [480, 640], "counts": "UZV51m>2UN3`D1T;=gDEW;a0bDB];b0^D@a;e0QDEn;g1N2O1O1aN^M_Fc2^9bM^F`2_9eM^F\\2`9iM\\FX2a9mM\\FT2a9QNXFT2e9RNSFS2]9_N\\F4U;i14M3M2M4M3L3N2N101O1N101ORD`M^;^2`DoMX1]O]8d2XF[NP1TOh8`2UFhNg0lNT9P4cFSL\\9U4ZFnKf9f400O100O10O1O0N2C>B=J7J5K5L5J5N3N2N2M4M2N2M3N2N2N2O1N2O1N3N1N2N2O1000001O001N2O001O001O1N110O10O00010O001O00001O001O00001O001O0O1gEkMg7V2THnMk7S2QHQNo7R2jGRNU8f2oF_MQ9X4O01000ON3M2N3N1N3M2N3M2O2M2N3mNS1O001O001OnEVMP8i2PHZMm7f2SH\\Ml7d2SH^Ml7d2QH^Mm7e2PH]Mo7e2nG]MP8f2mG\\MR8f2kG\\MT8f2iG[MV8h2gGYMY8i2cGYM\\8j2aGWM_8k2ZGZMe8i2SG]Mm8e2lF`MS9c2fFbMY9S4ORMcF7a9HdFO`91dFG`99eF^O`9a0dFXO_9h0eFPO_9o0gFhN^9W1fFaN_9]1gFZN]9e1iFSN[9l1jFkM[9P2m1H8L4O1O1O001O1O1O1O1O100O001O1O11O1O001O1N2N101N2O1N101N2N2O0O2O1N2O0O2N2O1N2OoM[Ch1c<UNaCl1^<PNeCS2c<1N2O2M2OQNcC[1\\<dNjCW1V<hNPDT1P<jNVDR1i;lN^DP1b;nNeDl0[;SOkDi0U;UOQEg0h<L4L4K6K5KcPQ1"}, "label": "a baby mule"}]}
{"id": 215243, "image": "COCO_train2014_000000215243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult horse standing in a meadow while a foal is walking away from it\"."}], "task": "refering", "answer_template": "The \"an adult horse standing in a meadow while a foal is walking away from it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 282, 283, 284, 302, 303, 304, 306, 307], "bbox": [0.10943750000000001, 0.2695208333333333, 0.482828125, 0.8038541666666668], "iscrowd": 0, "area": 35559.2875, "rle": {"size": [480, 640], "counts": "oPQ1g0`=k0K6I6K6I6J7J5J7J5J7M2M4M9F:F:G3L5L4K5K4M4K5L3L5K5L4eG\\JQ7j5bH^J^7f5UHcJj7Z600O1O100O10fMVH]L5Dc7P4\\HTL<EW7X4`HmKd0Cl6a4cHeKk0Ca6i4gH]Ki8c4[GVKg8l4e05LaNYFhMc9Z2_FfM^9Z2eFfMX9[2jFeMS9\\2oFdMn8\\2UGRMBAY9^3WGlLECT9a3ZGfLHFn8e3\\G_LMGh8k3]GXL\\9h3gFSLA`0m8^3hGVMj7k2nGiMe7W2UHVNd7k1UHaNe7_1RHQOg7[4I2NO1000000O1RObHZJ^7b5kHYJU7b5UIYJk6b5]IZJd6a5bI^J_6\\5hIbJX6Y5nIfJR6V5TJgJn5W5TJhJl5V5XJhJh5V5S2M2N2O1N3M2M3H9F9O100010O0001O00000010O0001O000010O000001O002NO2N1N2O1O1O1O2N1O1O1O7I7H7J7aFUK\\8Q5YGaKZ8i5I8H7I7J6J2N0000000000]OmHYIT7_6UI_Ik6Z6_IcIa6\\4iIULM]OZ6]4TJgKIJS6^4iK_KW4`4U3O100O1O2N1O1001O001O001O010O001O001Oi1WN9G9G8H9H2M00000000001O0001M2K5YOg0`N`1L4L4L4L4K5L5K4L3M1O1N101N6K5J6K6[Od0O101O0O1000O1O00100Oa0_O2N1O2O0O2N1O000001O00O1O2O0O100O1OO2PMbDb10fN_;GhD[1OiN[;KlDT1OlNV;OQEn0OnNQ;4VEg0NPOn:7aE;^;EoDLT;3l1O101N100O2O0Oo_k4"}, "label": "an adult horse standing in a meadow while a foal is walking away from it"}]}
{"id": 215243, "image": "COCO_train2014_000000215243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a adult brown horse\"."}], "task": "refering", "answer_template": "The \"a adult brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 282, 283, 284, 302, 303, 304, 306, 307], "bbox": [0.10943750000000001, 0.2695208333333333, 0.482828125, 0.8038541666666668], "iscrowd": 0, "area": 35559.2875, "rle": {"size": [480, 640], "counts": "oPQ1g0`=k0K6I6K6I6J7J5J7J5J7M2M4M9F:F:G3L5L4K5K4M4K5L3L5K5L4eG\\JQ7j5bH^J^7f5UHcJj7Z600O1O100O10fMVH]L5Dc7P4\\HTL<EW7X4`HmKd0Cl6a4cHeKk0Ca6i4gH]Ki8c4[GVKg8l4e05LaNYFhMc9Z2_FfM^9Z2eFfMX9[2jFeMS9\\2oFdMn8\\2UGRMBAY9^3WGlLECT9a3ZGfLHFn8e3\\G_LMGh8k3]GXL\\9h3gFSLA`0m8^3hGVMj7k2nGiMe7W2UHVNd7k1UHaNe7_1RHQOg7[4I2NO1000000O1RObHZJ^7b5kHYJU7b5UIYJk6b5]IZJd6a5bI^J_6\\5hIbJX6Y5nIfJR6V5TJgJn5W5TJhJl5V5XJhJh5V5S2M2N2O1N3M2M3H9F9O100010O0001O00000010O0001O000010O000001O002NO2N1N2O1O1O1O2N1O1O1O7I7H7J7aFUK\\8Q5YGaKZ8i5I8H7I7J6J2N0000000000]OmHYIT7_6UI_Ik6Z6_IcIa6\\4iIULM]OZ6]4TJgKIJS6^4iK_KW4`4U3O100O1O2N1O1001O001O001O010O001O001Oi1WN9G9G8H9H2M00000000001O0001M2K5YOg0`N`1L4L4L4L4K5L5K4L3M1O1N101N6K5J6K6[Od0O101O0O1000O1O00100Oa0_O2N1O2O0O2N1O000001O00O1O2O0O100O1OO2PMbDb10fN_;GhD[1OiN[;KlDT1OlNV;OQEn0OnNQ;4VEg0NPOn:7aE;^;EoDLT;3l1O101N100O2O0Oo_k4"}, "label": "a adult brown horse"}]}
{"id": 75982, "image": "COCO_train2014_000000075982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player serving a ball to his colleague wearing white t shirt and black color trouser\"."}], "task": "refering", "answer_template": "The \"a tennis player serving a ball to his colleague wearing white t shirt and black color trouser\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 97, 98, 99, 119, 120, 121, 140, 141, 142, 143, 161, 162, 163, 164, 165, 183, 184, 185, 186, 187, 206, 207, 208, 228, 229, 230, 250, 251, 252, 271, 272, 273, 293, 294, 295], "bbox": [0.346344262295082, 0.23385749385749388, 0.5202786885245901, 0.9282309582309582], "iscrowd": 0, "area": 15471.230350000002, "rle": {"size": [407, 610], "counts": "XRd21f<1kGOb33ZLOe34VL0i31TL1k32QL0o31nK2Q40kK2T41gK3X4NeK4Z4ObK3\\40aK3\\41_K2`41\\K2a41\\K1b41\\K1b40]K3`4N^K5`4L_K7^4IbK9\\4HcK;Z4FdK=Z4DeK>Y4BgKa0V4@iKb0V4^OiKe0T4[OkKh0T4XOkKj0T4VOkKm0S4SOlKo0R4QOmKS1Q4mNnKU1Q4SOfKo0Y4KlJ8Q5c0TJ_Ok5h3N1O2M2O1O2N1O2M2O1O2N1O2NO0N3N2N2N2N2N1O2N2N2N2N2N2NO010O10000O10000O1002N3M2N3L4M2N3M2N3M2N3M3M2N3M2N3M3L4M3M3M4L3M3M3M3M3M3M3M7I<C=D;[Oc0WOi0XOkZd3"}, "label": "a tennis player serving a ball to his colleague wearing white t shirt and black color trouser"}]}
{"id": 75982, "image": "COCO_train2014_000000075982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a white headband , black shorts , and white shirt playing tennis\"."}], "task": "refering", "answer_template": "The \"a person with a white headband , black shorts , and white shirt playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 97, 98, 99, 119, 120, 121, 140, 141, 142, 143, 161, 162, 163, 164, 165, 183, 184, 185, 186, 187, 206, 207, 208, 228, 229, 230, 250, 251, 252, 271, 272, 273, 293, 294, 295], "bbox": [0.346344262295082, 0.23385749385749388, 0.5202786885245901, 0.9282309582309582], "iscrowd": 0, "area": 15471.230350000002, "rle": {"size": [407, 610], "counts": "XRd21f<1kGOb33ZLOe34VL0i31TL1k32QL0o31nK2Q40kK2T41gK3X4NeK4Z4ObK3\\40aK3\\41_K2`41\\K2a41\\K1b41\\K1b40]K3`4N^K5`4L_K7^4IbK9\\4HcK;Z4FdK=Z4DeK>Y4BgKa0V4@iKb0V4^OiKe0T4[OkKh0T4XOkKj0T4VOkKm0S4SOlKo0R4QOmKS1Q4mNnKU1Q4SOfKo0Y4KlJ8Q5c0TJ_Ok5h3N1O2M2O1O2N1O2M2O1O2N1O2NO0N3N2N2N2N2N1O2N2N2N2N2N2NO010O10000O10000O1002N3M2N3L4M2N3M2N3M2N3M3M2N3M2N3M3L4M3M3M4L3M3M3M3M3M3M3M7I<C=D;[Oc0WOi0XOkZd3"}, "label": "a person with a white headband , black shorts , and white shirt playing tennis"}]}
{"id": 501972, "image": "COCO_train2014_000000501972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fire truck is red\"."}], "task": "refering", "answer_template": "The \"the fire truck is red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.122359375, 0.5885647058823529, 0.79521875, 0.9297647058823529], "iscrowd": 0, "area": 46116.89, "rle": {"size": [425, 640], "counts": "]dP11X=4L5K4L5K4L4L00001_EVOa8j0[GZOe8f0WG^Oh8c0SGCl8=PGGP99kFLU94gF0Y90eF3Z9McF6]9J`F9`9G^F;b9E[F?d9AYFb0g9^OVFe0j9[OTFh0k9XORFk0n9UOoEn0Q:ROlEQ1T:oNjET1U:P1O10001O001O1O1O1O1O1O1O1O1O00001O000O2O0000001O00001O0000001O000000000O100000N2O1O1O1O1N2I7H8I7L4M3N2N2N110O100000000O100000000O11RFkMf8Y2TGiMl8\\2nFeMR9`2eFdM[9a2YFfMf9o2000000000000000O1000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O10O100000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000O100000000000O100000000000002N3M3M2N3M3M1O00001N1000001O0000001O0000001O0000001O0O100000O10000O10000O1000000O10000O01000O1000000001O00000000001O00000O10001O0000000000001O0000000000001N10000000O1O1O1O100O1O1O1O1O1O1O1O00100O1O1O100000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O1000000001O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1N3N3M2N2M4M2N2N5J:G9G6J1N3N1O2N2N1NXPf1"}, "label": "the fire truck is red"}]}
{"id": 501972, "image": "COCO_train2014_000000501972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red fire truck with man sitting on side\"."}], "task": "refering", "answer_template": "The \"red fire truck with man sitting on side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.122359375, 0.5885647058823529, 0.79521875, 0.9297647058823529], "iscrowd": 0, "area": 46116.89, "rle": {"size": [425, 640], "counts": "]dP11X=4L5K4L5K4L4L00001_EVOa8j0[GZOe8f0WG^Oh8c0SGCl8=PGGP99kFLU94gF0Y90eF3Z9McF6]9J`F9`9G^F;b9E[F?d9AYFb0g9^OVFe0j9[OTFh0k9XORFk0n9UOoEn0Q:ROlEQ1T:oNjET1U:P1O10001O001O1O1O1O1O1O1O1O1O00001O000O2O0000001O00001O0000001O000000000O100000N2O1O1O1O1N2I7H8I7L4M3N2N2N110O100000000O100000000O11RFkMf8Y2TGiMl8\\2nFeMR9`2eFdM[9a2YFfMf9o2000000000000000O1000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O10O100000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000O100000000000O100000000000002N3M3M2N3M3M1O00001N1000001O0000001O0000001O0000001O0O100000O10000O10000O1000000O10000O01000O1000000001O00000000001O00000O10001O0000000000001O0000000000001N10000000O1O1O1O100O1O1O1O1O1O1O1O00100O1O1O100000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O1000000001O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1N3N3M2N2M4M2N2N5J:G9G6J1N3N1O2N2N1NXPf1"}, "label": "red fire truck with man sitting on side"}]}
{"id": 26836, "image": "COCO_train2014_000000026836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a duck is flying higher than two other ducks\"."}], "task": "refering", "answer_template": "The \"a duck is flying higher than two other ducks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 60, 61, 62, 63, 64, 85, 86, 87, 88, 89, 90, 108, 109], "bbox": [0.57584375, 0.08348946135831381, 0.929078125, 0.3060421545667448], "iscrowd": 0, "area": 4263.047000000001, "rle": {"size": [427, 640], "counts": "oli42Y=001O100O010O100O010O10O0010O0010O01O001O010O000010O01O001O010O001O010O001O00010O001O01O01O010O0010O00010O01O01O01O010O0010O01O10O01O100O010O1O010O100O010O1RD^OR;a0kDDS;=kDH]OH^;a0TE0j:OVEOm:1QENQ;3nDNR;1mD0T;0kD1T;0fD5[;J`D>^;B`Da0^;@`D`0b;d01O00100O10O0100N01N2O10000000000002N3M0O2O00001N1005LO000O2ZOYD^Oo;a0TD\\Om;b0VD[Om;c0VDZOk;`0d0M201O0000000000O10000000000000000000001O0000000000000000001O0000001O0001O010O0010O01O10O00010O0001O01O01O000010O01O00010O001O00010O001O01O01O001O01O01O0010O01O00001O000000000O10001O00000O01000LlB1T=OlB0T=0mBOi_c0"}, "label": "a duck is flying higher than two other ducks"}]}
{"id": 26836, "image": "COCO_train2014_000000026836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"duck flying highest of the three\"."}], "task": "refering", "answer_template": "The \"duck flying highest of the three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 60, 61, 62, 63, 64, 85, 86, 87, 88, 89, 90, 108, 109], "bbox": [0.57584375, 0.08348946135831381, 0.929078125, 0.3060421545667448], "iscrowd": 0, "area": 4263.047000000001, "rle": {"size": [427, 640], "counts": "oli42Y=001O100O010O100O010O10O0010O0010O01O001O010O000010O01O001O010O001O010O001O00010O001O01O01O010O0010O00010O01O01O01O010O0010O01O10O01O100O010O1O010O100O010O1RD^OR;a0kDDS;=kDH]OH^;a0TE0j:OVEOm:1QENQ;3nDNR;1mD0T;0kD1T;0fD5[;J`D>^;B`Da0^;@`D`0b;d01O00100O10O0100N01N2O10000000000002N3M0O2O00001N1005LO000O2ZOYD^Oo;a0TD\\Om;b0VD[Om;c0VDZOk;`0d0M201O0000000000O10000000000000000000001O0000000000000000001O0000001O0001O010O0010O01O10O00010O0001O01O01O000010O01O00010O001O00010O001O01O01O001O01O01O0010O01O00001O000000000O10001O00000O01000LlB1T=OlB0T=0mBOi_c0"}, "label": "duck flying highest of the three"}]}
{"id": 26836, "image": "COCO_train2014_000000026836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"duck closest to water with white chest\"."}], "task": "refering", "answer_template": "The \"duck closest to water with white chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 168, 169, 170, 171, 172, 173, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219], "bbox": [0.30353125, 0.4411943793911007, 0.747359375, 0.6686651053864168], "iscrowd": 0, "area": 6312.549650000001, "rle": {"size": [427, 640], "counts": "X`a23T=5O1O1O1O1I6101O001O1O1O1N101N10O001O001O001O001O000001O000000000000000000000000000000001O0000000000000000000001O0000001O0001O01O000O10001O0000001O0000001O0000001O0000001O0000010O0001O010O0O2OkCEZ;:gDJIGl:>\\E:f:EYE<g:CZE<h:BXE>i:CVE<k:DUE;m:EQE=n:DoD?Q;AkDc0T;]OiDg0V;ZOfDj0Z;VObDn0];a001N00000000001O000000000000000000000O1M3M4L3N200O1O1O100O1O11N2O1O1O1N7K6K1ON1OJ7H7O2N1O2N1O1ROUD;l;F\\DFKNk;<j01O000000000001O0000000001O0001O000000000000001O0000000001O01O000000001O001O00001O00010O00010O000010O01O01O01O01O01O01O01O00010O0010O00010O00010O00010O0001O01O00001O00000000000000O010000000000000O100000000000000O010O100O1O100O100O100O100O100O[=0jlR2"}, "label": "duck closest to water with white chest"}]}
{"id": 26836, "image": "COCO_train2014_000000026836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the duck that is closest to the water\"."}], "task": "refering", "answer_template": "The \"the duck that is closest to the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 168, 169, 170, 171, 172, 173, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219], "bbox": [0.30353125, 0.4411943793911007, 0.747359375, 0.6686651053864168], "iscrowd": 0, "area": 6312.549650000001, "rle": {"size": [427, 640], "counts": "X`a23T=5O1O1O1O1I6101O001O1O1O1N101N10O001O001O001O001O000001O000000000000000000000000000000001O0000000000000000000001O0000001O0001O01O000O10001O0000001O0000001O0000001O0000001O0000010O0001O010O0O2OkCEZ;:gDJIGl:>\\E:f:EYE<g:CZE<h:BXE>i:CVE<k:DUE;m:EQE=n:DoD?Q;AkDc0T;]OiDg0V;ZOfDj0Z;VObDn0];a001N00000000001O000000000000000000000O1M3M4L3N200O1O1O100O1O11N2O1O1O1N7K6K1ON1OJ7H7O2N1O2N1O1ROUD;l;F\\DFKNk;<j01O000000000001O0000000001O0001O000000000000001O0000000001O01O000000001O001O00001O00010O00010O000010O01O01O01O01O01O01O01O00010O0010O00010O00010O00010O0001O01O00001O00000000000000O010000000000000O100000000000000O010O100O1O100O100O100O100O100O[=0jlR2"}, "label": "the duck that is closest to the water"}]}
{"id": 256215, "image": "COCO_train2014_000000256215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue and white boat with two people on the front\"."}], "task": "refering", "answer_template": "The \"the blue and white boat with two people on the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.623703125, 0.3281041666666667, 1.0, 0.7488958333333334], "iscrowd": 0, "area": 35741.927200000006, "rle": {"size": [480, 640], "counts": "XVk52m>4L3N3M2N2M4M2N2M4M2N2N3L3N3M2N2M4M2N2M4M2N3M2M3N3M2N2N3L3N000000O1000000000O10001O1O1O001O1O1N2001O00O100O1L3eFUM`6m2^IUMa6l2^IUMa6l2]IVMb6k2]IVMb6k2\\IWMc6j2[IXMe6h2ZIYMe6h2YIZMf6g2YIZMf6g2XI[Mg6e2XI]Mg6d2XI]Mg6d2WI^Mh6c2WI^Mh6c2VI_Mi6b2VI_Mi6b2UI`Mj6a2TIaMk6`2TI`Mm6`2QIaMo6`2PI`MP7`2oHaMQ7`2mHaMS7`2lH`MT7a2jH`MV7a2iH_MW7b2gH_MY7b2fH^MZ7c2dH^M\\7c2bH]M_7d2`H\\M`7e2^H\\Mc7c2]H]Mc7d2[H]Me7d2YH]Mg7d2XH\\Mh7e2VH\\Mj7e2UH[Mk7f2SH[Mm7f2RHZMn7g2PHZMP8g2nGYMS8h2lGXMU8g2jGZMV8g2iGYMW8h2gGYMY8h2fGXMZ8i2dGXM\\8i2bGXM^8i2aGWM_8j2_GWMa8j2^GVMb8k2\\GVMe8j2YGVMh8j2XGVMh8a4000000000000001O00000O1000000000001O000000000O10000000000000000000000O101O0000000000000000000O10000000000000001O0000000O10000000000000000000000O10001O00000000000000O10O100000000000O100000000000O10O10000000000000000O10O100000000000O1000000000O1000O10000000000nJ"}, "label": "the blue and white boat with two people on the front"}]}
{"id": 256215, "image": "COCO_train2014_000000256215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front end of a yacht that is black and white sitting in the water\"."}], "task": "refering", "answer_template": "The \"the front end of a yacht that is black and white sitting in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.623703125, 0.3281041666666667, 1.0, 0.7488958333333334], "iscrowd": 0, "area": 35741.927200000006, "rle": {"size": [480, 640], "counts": "XVk52m>4L3N3M2N2M4M2N2M4M2N2N3L3N3M2N2M4M2N2M4M2N3M2M3N3M2N2N3L3N000000O1000000000O10001O1O1O001O1O1N2001O00O100O1L3eFUM`6m2^IUMa6l2^IUMa6l2]IVMb6k2]IVMb6k2\\IWMc6j2[IXMe6h2ZIYMe6h2YIZMf6g2YIZMf6g2XI[Mg6e2XI]Mg6d2XI]Mg6d2WI^Mh6c2WI^Mh6c2VI_Mi6b2VI_Mi6b2UI`Mj6a2TIaMk6`2TI`Mm6`2QIaMo6`2PI`MP7`2oHaMQ7`2mHaMS7`2lH`MT7a2jH`MV7a2iH_MW7b2gH_MY7b2fH^MZ7c2dH^M\\7c2bH]M_7d2`H\\M`7e2^H\\Mc7c2]H]Mc7d2[H]Me7d2YH]Mg7d2XH\\Mh7e2VH\\Mj7e2UH[Mk7f2SH[Mm7f2RHZMn7g2PHZMP8g2nGYMS8h2lGXMU8g2jGZMV8g2iGYMW8h2gGYMY8h2fGXMZ8i2dGXM\\8i2bGXM^8i2aGWM_8j2_GWMa8j2^GVMb8k2\\GVMe8j2YGVMh8j2XGVMh8a4000000000000001O00000O1000000000001O000000000O10000000000000000000000O101O0000000000000000000O10000000000000001O0000000O10000000000000000000000O10001O00000000000000O10O100000000000O100000000000O10O10000000000000000O10O100000000000O1000000000O1000O10000000000nJ"}, "label": "the front end of a yacht that is black and white sitting in the water"}]}
{"id": 256215, "image": "COCO_train2014_000000256215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large boat painted to look like a fish with teeth and named godzilla\"."}], "task": "refering", "answer_template": "The \"a large boat painted to look like a fish with teeth and named godzilla\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 108, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241], "bbox": [0.00540625, 0.18370833333333333, 0.9986875000000002, 0.6186666666666667], "iscrowd": 0, "area": 30904.388349999994, "rle": {"size": [480, 640], "counts": "m`32n>00000000001O0O2O001O001O001O0O2O1O001O001O001N101O1O001O001O0O2O001O001O1O001N101O001O001O1O0O2O001O001O001O1N101O001O001O001N2O001O001O001O0O2O001O1O001O001N101O00000000000O100000000000O10O100000000000000O100000000000000O100000000000000O010000000000000O100000000000000000000000000000000000O1000000000000000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000001N10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000ORObN\\D^1d;bN\\D^1d;bN\\D^1c;cN]D]1c;cN]D]1c;cN]D]1c;cN]D]1b;dN^D\\1b;dN^D\\1b;dN^D\\1a;eN_D[1a;eN_D[1a;eN_D[1`;eNaD[1_;eNaDZ1`;fN`DZ1_;gNaDY1_;gNaDY1_;gNaDY1^;hNbDX1i:cNlD5;X1h:dNmD4;X1h:dNlD5<W1g:eNmD4<W1g:eNmD4<W1g:eNlD5=V1f:fNmD4=V1f:fNmD4=V1e:gNnD3=V1e:gNmD4>T1e:iNmD3>T1e:iNmD3>T1d:jNmD3?S1d:jNmD3?S1d:kNlD2`0S1c:lNlD2a0R1c:lNlD2b0Q1a:nNmD1b0Q1a:nNlD2c0P1`:oNmD1c0P1`:TOhDLh0P1`:H`E8_:IaE7_:IaE7^:JbE6^:JbE5^:LbE4^:LbE4]:McE3]:McE3]:McE3\\:NdE2\\:NdE2[:OeE1[:OeE1Z:0fE0Z:0fE0Y:1gEOW:3iEMV:4jELU:4lEKS:7mEIR:8nEHQ:9oEGo9;QFEn9<RFDn9<RFDm9=SFCm9=SFCl9>TFBl9>TFBk9?UFAY9gNWFh1`0AW9iNXFg1a0@T9kN\\Fe1`0@Q9nN^Fd1`0]OQ9PO`Fd1>\\OQ9SO^Fc1`0ZOR9UO\\Fb1a0YOR9XO[Fa1a0WOT9ZOXFa1c0UOT9]OWF_1d0TOU9_OTF_1f0ROV9i1iFWNV9k1iFUNW9l1hFTNW9n1hFRNX9o1gFQNY9P2fFPNY9S2eFmM[9T2dFkM]9V2bFjM]9X2bFhM^9Y2_FiM`9Y2^FhMb9Y2[FiMe9X2XFjMg9h300O10000O100OaL]Fc1c9]N`F_1a9aNbF\\1]9eNeFY1[9gNhFV1W9kNlFR1f8SMVGj18P1b8VMVGi1;o0^8ZMUGb1f0Q1T8^MVGV1T1Y1f7aMUGU1Z1W1`7fMVGP1^1X1\\7hMVGn0c1W1V7mMWGi0h1W1Q7PNXGf0l1W1k6UNXGc0P2V1i6VNXGc0R2T1g6XNWGc0V2Q1d6]NSGc0]2m0a6_NoFg0b2h0_6bNlFg0i2d0\\6EhI8Y6FkI7U6HmI7T6FoI9Q6SOhF2Y3k0P6oNjF6W3j0P6kNnF:R3k0P6hNQG=P3j0^6VObIj0^6VOcIi0]6XObIh0_6WOaIh0`6CVI<j6GSI9m6IRI6o6JPI6P7KPI4P7MPI2P7OoH1R7NoH1Q70oHOQ72oHMQ74oHKR75nHJR76oHIQ78oHGQ7:oHDS7<mHCS7>mHAS7?mHAS7?nH@S7?nH@T7>mHAT7>mHAU7=lHBV7<kHCV7<kHCU7=lHBS7EjE`0U3JQ7FnE<Q3NQ7FoE;Q3NP7FQF:Q3No6HQF9Q3NY;O1O1O1O1O1O1O001O1O1O1O^Q50_nJ3A?0010O0000000000000000000010O0000000000000000000010O0000000000000000000010O00000000000000000000010O00000000000000000001O01O00000000000000000001O01O000000000000000001O0001O000000000000000O1000001O0000000000000O1000001O0000000000000O100000001O00000000000O10000000001O0000000O100000000000001O00001NV:"}, "label": "a large boat painted to look like a fish with teeth and named godzilla"}]}
{"id": 256215, "image": "COCO_train2014_000000256215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large graffetti yacht\"."}], "task": "refering", "answer_template": "The \"large graffetti yacht\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 108, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241], "bbox": [0.00540625, 0.18370833333333333, 0.9986875000000002, 0.6186666666666667], "iscrowd": 0, "area": 30904.388349999994, "rle": {"size": [480, 640], "counts": "m`32n>00000000001O0O2O001O001O001O0O2O1O001O001O001N101O1O001O001O0O2O001O001O1O001N101O001O001O1O0O2O001O001O001O1N101O001O001O001N2O001O001O001O0O2O001O1O001O001N101O00000000000O100000000000O10O100000000000000O100000000000000O100000000000000O010000000000000O100000000000000000000000000000000000O1000000000000000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000001N10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000ORObN\\D^1d;bN\\D^1d;bN\\D^1c;cN]D]1c;cN]D]1c;cN]D]1c;cN]D]1b;dN^D\\1b;dN^D\\1b;dN^D\\1a;eN_D[1a;eN_D[1a;eN_D[1`;eNaD[1_;eNaDZ1`;fN`DZ1_;gNaDY1_;gNaDY1_;gNaDY1^;hNbDX1i:cNlD5;X1h:dNmD4;X1h:dNlD5<W1g:eNmD4<W1g:eNmD4<W1g:eNlD5=V1f:fNmD4=V1f:fNmD4=V1e:gNnD3=V1e:gNmD4>T1e:iNmD3>T1e:iNmD3>T1d:jNmD3?S1d:jNmD3?S1d:kNlD2`0S1c:lNlD2a0R1c:lNlD2b0Q1a:nNmD1b0Q1a:nNlD2c0P1`:oNmD1c0P1`:TOhDLh0P1`:H`E8_:IaE7_:IaE7^:JbE6^:JbE5^:LbE4^:LbE4]:McE3]:McE3]:McE3\\:NdE2\\:NdE2[:OeE1[:OeE1Z:0fE0Z:0fE0Y:1gEOW:3iEMV:4jELU:4lEKS:7mEIR:8nEHQ:9oEGo9;QFEn9<RFDn9<RFDm9=SFCm9=SFCl9>TFBl9>TFBk9?UFAY9gNWFh1`0AW9iNXFg1a0@T9kN\\Fe1`0@Q9nN^Fd1`0]OQ9PO`Fd1>\\OQ9SO^Fc1`0ZOR9UO\\Fb1a0YOR9XO[Fa1a0WOT9ZOXFa1c0UOT9]OWF_1d0TOU9_OTF_1f0ROV9i1iFWNV9k1iFUNW9l1hFTNW9n1hFRNX9o1gFQNY9P2fFPNY9S2eFmM[9T2dFkM]9V2bFjM]9X2bFhM^9Y2_FiM`9Y2^FhMb9Y2[FiMe9X2XFjMg9h300O10000O100OaL]Fc1c9]N`F_1a9aNbF\\1]9eNeFY1[9gNhFV1W9kNlFR1f8SMVGj18P1b8VMVGi1;o0^8ZMUGb1f0Q1T8^MVGV1T1Y1f7aMUGU1Z1W1`7fMVGP1^1X1\\7hMVGn0c1W1V7mMWGi0h1W1Q7PNXGf0l1W1k6UNXGc0P2V1i6VNXGc0R2T1g6XNWGc0V2Q1d6]NSGc0]2m0a6_NoFg0b2h0_6bNlFg0i2d0\\6EhI8Y6FkI7U6HmI7T6FoI9Q6SOhF2Y3k0P6oNjF6W3j0P6kNnF:R3k0P6hNQG=P3j0^6VObIj0^6VOcIi0]6XObIh0_6WOaIh0`6CVI<j6GSI9m6IRI6o6JPI6P7KPI4P7MPI2P7OoH1R7NoH1Q70oHOQ72oHMQ74oHKR75nHJR76oHIQ78oHGQ7:oHDS7<mHCS7>mHAS7?mHAS7?nH@S7?nH@T7>mHAT7>mHAU7=lHBV7<kHCV7<kHCU7=lHBS7EjE`0U3JQ7FnE<Q3NQ7FoE;Q3NP7FQF:Q3No6HQF9Q3NY;O1O1O1O1O1O1O001O1O1O1O^Q50_nJ3A?0010O0000000000000000000010O0000000000000000000010O0000000000000000000010O00000000000000000000010O00000000000000000001O01O00000000000000000001O01O000000000000000001O0001O000000000000000O1000001O0000000000000O1000001O0000000000000O100000001O00000000000O10000000001O0000000O100000000000001O00001NV:"}, "label": "large graffetti yacht"}]}
{"id": 67807, "image": "COCO_train2014_000000067807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white horse and white / brown horse nuzzling each other\"."}], "task": "refering", "answer_template": "The \"white horse and white / brown horse nuzzling each other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 122, 123, 124, 125, 126, 145, 146, 147, 148, 168, 169, 171, 192, 193, 194], "bbox": [0.33384375, 0.2939906103286385, 0.530109375, 0.585093896713615], "iscrowd": 0, "area": 5383.86, "rle": {"size": [426, 640], "counts": "SVi2V1P<7J7I6L5L3L3N3L4M2M4M1OeNfEGX:<bFhN^9Y1X1c0fDTNR:n1iEVNV:l1eEXNY:k1bEYN]:[2N10000O100000XOhEaNU:\\1PFfNl9V1[FhNd9U1bFhN^9T1iFiNX9S1mFjNT9R1SGkNm8R1YGkNg8U1[GiNe8W1h10O00000001O01O00000001O0001O000001O0001O000O1O1^EeNU9[1iFiNT9X1kFkNR9W1kFmNR9g1ZF\\Nd9e2O1O1O@ZF^Mg9]2_FaMd9X2bFfMb9Q2Q1I8H7oN\\D1l;JZDNn;KYDOl;LYD1mW3OcTM4M2N2N2M101O001N2O001O0O2O001O0O2O001O0O2O0K6L3O0018G3N0O01O01O2M3N2M2N1J6O1O1OPSm3"}, "label": "white horse and white / brown horse nuzzling each other"}]}
{"id": 67807, "image": "COCO_train2014_000000067807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse behind a brown and white spotted horse\"."}], "task": "refering", "answer_template": "The \"a white horse behind a brown and white spotted horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 122, 123, 124, 125, 126, 145, 146, 147, 148, 168, 169, 171, 192, 193, 194], "bbox": [0.33384375, 0.2939906103286385, 0.530109375, 0.585093896713615], "iscrowd": 0, "area": 5383.86, "rle": {"size": [426, 640], "counts": "SVi2V1P<7J7I6L5L3L3N3L4M2M4M1OeNfEGX:<bFhN^9Y1X1c0fDTNR:n1iEVNV:l1eEXNY:k1bEYN]:[2N10000O100000XOhEaNU:\\1PFfNl9V1[FhNd9U1bFhN^9T1iFiNX9S1mFjNT9R1SGkNm8R1YGkNg8U1[GiNe8W1h10O00000001O01O00000001O0001O000001O0001O000O1O1^EeNU9[1iFiNT9X1kFkNR9W1kFmNR9g1ZF\\Nd9e2O1O1O@ZF^Mg9]2_FaMd9X2bFfMb9Q2Q1I8H7oN\\D1l;JZDNn;KYDOl;LYD1mW3OcTM4M2N2N2M101O001N2O001O0O2O001O0O2O001O0O2O0K6L3O0018G3N0O01O01O2M3N2M2N1J6O1O1OPSm3"}, "label": "a white horse behind a brown and white spotted horse"}]}
{"id": 157920, "image": "COCO_train2014_000000157920.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a catchers mitt and a yellow ball\"."}], "task": "refering", "answer_template": "The \"a man holding a catchers mitt and a yellow ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294], "bbox": [0.692296875, 0.24137214137214136, 0.826875, 0.7635135135135135], "iscrowd": 0, "area": 12978.197199999995, "rle": {"size": [481, 640], "counts": "e[`64i>8H8H7I6J6K5J6TDcNl9c1oE`NP:c1mE^NQ:g1jE[Ni6IWKo1mMYN`67_Kc1nMVNc6:[Kd1nMTNf6<XKc1oMRNh6=WKd1nMPNj6>nI[O^O[2FnMm6>kI^O_OY2FlMo6?jI^O_OZ2EiMR7`0iI_O]O\\2DgMU7`0hI_OXOc2G_MY7`0fIAnNl20TM[7`0fIQ4Z6PLdIP4\\6RLbIo3^6RL`Ie1TOXO\\7SO_Ie1YOUOW7XOYIh1CnNT7[OPIn10cNQ7_OgHT2<[Nm6W3WIfLi6Y3[IcLf6[3_IbLa6_3kITLU6m3oImKR6U4PJgKP6Z4SJbKm5_4TJ_Kl5c4SJ[KP6d4QJZKQ6f4nIYKS6f4oIWKT6g4mIXKU6f4lIYKU6f4mIWKV6g4T2O0O2N001O001O03N3M3M3M3M4M2M3M3M3]GcJ\\7_5YHhJh7[5QHhJP8[5iGhJX8P61O2N1O2N1`MfGXN[8b1kG\\NV8_1PH_NR8\\1SHbNn7[1VHcNl7R1_HlNb7GYG@[1g0^7C_GAV14fNYOh8o0bG_O[1KaNCf8o0dG^Oi1ZOQN6e8P1fGZOY2CX6k0eG]OV2D^6f0bGAl;7ZDDn;4ioc1"}, "label": "a man holding a catchers mitt and a yellow ball"}]}
{"id": 157920, "image": "COCO_train2014_000000157920.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding a baseball and glove walking on a field with two other baseball players\"."}], "task": "refering", "answer_template": "The \"man holding a baseball and glove walking on a field with two other baseball players\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294], "bbox": [0.692296875, 0.24137214137214136, 0.826875, 0.7635135135135135], "iscrowd": 0, "area": 12978.197199999995, "rle": {"size": [481, 640], "counts": "e[`64i>8H8H7I6J6K5J6TDcNl9c1oE`NP:c1mE^NQ:g1jE[Ni6IWKo1mMYN`67_Kc1nMVNc6:[Kd1nMTNf6<XKc1oMRNh6=WKd1nMPNj6>nI[O^O[2FnMm6>kI^O_OY2FlMo6?jI^O_OZ2EiMR7`0iI_O]O\\2DgMU7`0hI_OXOc2G_MY7`0fIAnNl20TM[7`0fIQ4Z6PLdIP4\\6RLbIo3^6RL`Ie1TOXO\\7SO_Ie1YOUOW7XOYIh1CnNT7[OPIn10cNQ7_OgHT2<[Nm6W3WIfLi6Y3[IcLf6[3_IbLa6_3kITLU6m3oImKR6U4PJgKP6Z4SJbKm5_4TJ_Kl5c4SJ[KP6d4QJZKQ6f4nIYKS6f4oIWKT6g4mIXKU6f4lIYKU6f4mIWKV6g4T2O0O2N001O001O03N3M3M3M3M4M2M3M3M3]GcJ\\7_5YHhJh7[5QHhJP8[5iGhJX8P61O2N1O2N1`MfGXN[8b1kG\\NV8_1PH_NR8\\1SHbNn7[1VHcNl7R1_HlNb7GYG@[1g0^7C_GAV14fNYOh8o0bG_O[1KaNCf8o0dG^Oi1ZOQN6e8P1fGZOY2CX6k0eG]OV2D^6f0bGAl;7ZDDn;4ioc1"}, "label": "man holding a baseball and glove walking on a field with two other baseball players"}]}
{"id": 313569, "image": "COCO_train2014_000000313569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a blue with red flower shirt\"."}], "task": "refering", "answer_template": "The \"a person in a blue with red flower shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 350, 351, 352, 353, 354, 355, 356, 357, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.22903125000000002, 0.04933333333333333, 0.7359218750000001, 1.0], "iscrowd": 0, "area": 83457.10975, "rle": {"size": [480, 640], "counts": "cSU2f0R>f0B?@?A?B>A`0@?A?B?@?A?B>B?@?B>B>A>C00000O2O000O10000O10000000000001O0000000000000000001OO100000000O100000000O1fLoGER88WHAi7=_H]Oa7`0iHYOW7e0oHVOR7h0RIVOn6g0WIWOi6g0[IWOe6f0_IYOa6e0cIXO^6f0fIWO[6f0jIWOW6g0lIVOV6h0nITOT6j0PJSOQ6l0RJZNPM<P9Y1\\JdNf5Z1]JdNd5[1^JbNd5]1^JaNc5^1^JbNb5\\1aJcN_5\\1cJcN]5\\1T4O1N2O1O1N2O10000O10001N10000O10000O10000O10000O100000000O1000000000000O10000000000O100000001O00000000000001O00000000001O00000000001O00000000001O0lF^Oj3b0ULAi3?ULEi3<ULGi3:TLJj36TLNj33TL0j31SL3k3NRL6m3IRL:l3GQL<n3EPL=o3DnK?Q4AmKc0Q4^OmKd0R4]OkKf0T4[OjKg0U4YOiKk0U4VOhKm0W4TOgKn0X4SOeKP1Z4POdKS1[4nNbKU1]4lN`KV1`4kN]KX1b4hN[K[1e4fNWK^1h4cNTK`1l4aNoJd1P5_NjJe1U5^NeJe1[5^N^Jg1a5\\NYJg1g5[NTJi1k5ZNnIj1R6YNhIk1W6XNbIl1^6WN\\Im1c6VNUIo1k6TNlHS2S7oMfHV2Z7h201O01O01O00001O00001O00001O00001O00001VNi1I8M2O1N3N10001O001O1O001N2O001O1O1O001O00000O2eKhEl3X:SLjEl3W:TLiEk3W:ULjEj3W:ULkEi3U:WLlEh3U:WLlEh3T:XLjEj3W:ULfEm3a:nK\\ET4e:3N2M3N2M3N2O10C<C>M3M3M2N3M3M3L8I3L4L3\\O4oMSDi1T<XNkCb1[<_NeCY1b<iN]Co0k<QOTCh0S=YOmBb0X=^OhB?Z=BeB>\\=BdB>[=CeB<\\=DcB=^=BbB=`=C^B>b=C]B=c=D\\B<c=F\\B:d=F\\B:c=F^B9c=G\\B:c=G]B9e=E[B;f=DZB=g=AYB?g=BXB>g=BZB?e=A[B?d=B\\B>d=B\\B>d=B\\B?c=A]B?c=A]BN13d=M[BO24d=MYBO44c=LYB044c=LYB043e=LWB152d=MWB062d=8\\BHe=7[BHg=7ZBGf=:ZBFe=;[BDe=<\\BCd=>\\BBd=>\\BAe=?\\B@d=`0>O2O001O0O1N2O2M2N2O3L]o^2"}, "label": "a person in a blue with red flower shirt"}]}
{"id": 313569, "image": "COCO_train2014_000000313569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady with wii remote\"."}], "task": "refering", "answer_template": "The \"lady with wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 350, 351, 352, 353, 354, 355, 356, 357, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.22903125000000002, 0.04933333333333333, 0.7359218750000001, 1.0], "iscrowd": 0, "area": 83457.10975, "rle": {"size": [480, 640], "counts": "cSU2f0R>f0B?@?A?B>A`0@?A?B?@?A?B>B?@?B>B>A>C00000O2O000O10000O10000000000001O0000000000000000001OO100000000O100000000O1fLoGER88WHAi7=_H]Oa7`0iHYOW7e0oHVOR7h0RIVOn6g0WIWOi6g0[IWOe6f0_IYOa6e0cIXO^6f0fIWO[6f0jIWOW6g0lIVOV6h0nITOT6j0PJSOQ6l0RJZNPM<P9Y1\\JdNf5Z1]JdNd5[1^JbNd5]1^JaNc5^1^JbNb5\\1aJcN_5\\1cJcN]5\\1T4O1N2O1O1N2O10000O10001N10000O10000O10000O10000O100000000O1000000000000O10000000000O100000001O00000000000001O00000000001O00000000001O00000000001O0lF^Oj3b0ULAi3?ULEi3<ULGi3:TLJj36TLNj33TL0j31SL3k3NRL6m3IRL:l3GQL<n3EPL=o3DnK?Q4AmKc0Q4^OmKd0R4]OkKf0T4[OjKg0U4YOiKk0U4VOhKm0W4TOgKn0X4SOeKP1Z4POdKS1[4nNbKU1]4lN`KV1`4kN]KX1b4hN[K[1e4fNWK^1h4cNTK`1l4aNoJd1P5_NjJe1U5^NeJe1[5^N^Jg1a5\\NYJg1g5[NTJi1k5ZNnIj1R6YNhIk1W6XNbIl1^6WN\\Im1c6VNUIo1k6TNlHS2S7oMfHV2Z7h201O01O01O00001O00001O00001O00001O00001VNi1I8M2O1N3N10001O001O1O001N2O001O1O1O001O00000O2eKhEl3X:SLjEl3W:TLiEk3W:ULjEj3W:ULkEi3U:WLlEh3U:WLlEh3T:XLjEj3W:ULfEm3a:nK\\ET4e:3N2M3N2M3N2O10C<C>M3M3M2N3M3M3L8I3L4L3\\O4oMSDi1T<XNkCb1[<_NeCY1b<iN]Co0k<QOTCh0S=YOmBb0X=^OhB?Z=BeB>\\=BdB>[=CeB<\\=DcB=^=BbB=`=C^B>b=C]B=c=D\\B<c=F\\B:d=F\\B:c=F^B9c=G\\B:c=G]B9e=E[B;f=DZB=g=AYB?g=BXB>g=BZB?e=A[B?d=B\\B>d=B\\B>d=B\\B?c=A]B?c=A]BN13d=M[BO24d=MYBO44c=LYB044c=LYB043e=LWB152d=MWB062d=8\\BHe=7[BHg=7ZBGf=:ZBFe=;[BDe=<\\BCd=>\\BBd=>\\BAe=?\\B@d=`0>O2O001O0O1N2O2M2N2O3L]o^2"}, "label": "lady with wii remote"}]}
{"id": 313569, "image": "COCO_train2014_000000313569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a flower lei and a white sailor cup is with a friend\"."}], "task": "refering", "answer_template": "The \"a man with a flower lei and a white sailor cup is with a friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 25, 26, 27, 28, 29, 49, 50, 51, 52, 71, 72, 73, 74, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 278, 279, 280, 281, 302], "bbox": [0.013453125, 0.017041666666666667, 0.27242187500000004, 0.777125], "iscrowd": 0, "area": 41374.21085, "rle": {"size": [480, 640], "counts": "S^4^1Q<c1^Ob0M3M3M3L4M3M3M3M2M4M2N3M2N3L3N3M2N3M2M4M2N3M2N3L3N3M3M2N3L3N3M2N3M2M4M2N3M2N3L3N3M1O2N2M2O2N2N1O2N2N1O2N2O0O2N2N2N1O2N2N1O2O1N1O2N2N1O2N2mMiGZMX8e2UHnLm7Q3^HdLc7[3hHYLY7f3TImKn6R4Q2N101O1N101O1N2O0O2O1O0O2O1O0O010O0100O00100O010O010O10O0101N100O101N1O100O2O0O100O2O01O00010OZHlK`4T4`KmK`4R4`KoK_4R4`KPL^4P4bKQL^4n3bKSL]4m3dKSL[4n3dKTL[4k3eKVLZ4j3fKWLZ4i3eKXLZ4h3fKZLX4f3hK[LX4d3iK\\LV4e3iK\\LW4c3iK_LU4a3kK`LT4`3lKaLkN6a2Y3cNbLkNQ1f1]2_OdLiNl1l0`1;dLhNh21d0W1eLgNc3VOIQ2fLgNS4hNVO_2kLgNU4dNQOX2RLdMj0_1Y4_NkNY2]MWOP8d0UH[Ok7e0WHXOk7g0VHXOj7h0WHWOj7g0XHXOh7h0ZHVOg7i0ZHUOg7k0ZHTOg7j0[HUOe7k0]HSOd7l0]HSOc7;PICR7IbI4`6HfI2_6JeI0`6JgIMdcj6"}, "label": "a man with a flower lei and a white sailor cup is with a friend"}]}
{"id": 313569, "image": "COCO_train2014_000000313569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sailor in the white hat\"."}], "task": "refering", "answer_template": "The \"the sailor in the white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 25, 26, 27, 28, 29, 49, 50, 51, 52, 71, 72, 73, 74, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 278, 279, 280, 281, 302], "bbox": [0.013453125, 0.017041666666666667, 0.27242187500000004, 0.777125], "iscrowd": 0, "area": 41374.21085, "rle": {"size": [480, 640], "counts": "S^4^1Q<c1^Ob0M3M3M3L4M3M3M3M2M4M2N3M2N3L3N3M2N3M2M4M2N3M2N3L3N3M3M2N3L3N3M2N3M2M4M2N3M2N3L3N3M1O2N2M2O2N2N1O2N2N1O2N2O0O2N2N2N1O2N2N1O2O1N1O2N2N1O2N2mMiGZMX8e2UHnLm7Q3^HdLc7[3hHYLY7f3TImKn6R4Q2N101O1N101O1N2O0O2O1O0O2O1O0O010O0100O00100O010O010O10O0101N100O101N1O100O2O0O100O2O01O00010OZHlK`4T4`KmK`4R4`KoK_4R4`KPL^4P4bKQL^4n3bKSL]4m3dKSL[4n3dKTL[4k3eKVLZ4j3fKWLZ4i3eKXLZ4h3fKZLX4f3hK[LX4d3iK\\LV4e3iK\\LW4c3iK_LU4a3kK`LT4`3lKaLkN6a2Y3cNbLkNQ1f1]2_OdLiNl1l0`1;dLhNh21d0W1eLgNc3VOIQ2fLgNS4hNVO_2kLgNU4dNQOX2RLdMj0_1Y4_NkNY2]MWOP8d0UH[Ok7e0WHXOk7g0VHXOj7h0WHWOj7g0XHXOh7h0ZHVOg7i0ZHUOg7k0ZHTOg7j0[HUOe7k0]HSOd7l0]HSOc7;PICR7IbI4`6HfI2_6JeI0`6JgIMdcj6"}, "label": "the sailor in the white hat"}]}
{"id": 493793, "image": "COCO_train2014_000000493793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bananas\"."}], "task": "refering", "answer_template": "The \"bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 114, 115, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 189, 206, 207, 208], "bbox": [0.1929575163398693, 0.13264705882352942, 0.6500490196078432, 0.4506862745098039], "iscrowd": 0, "area": 19362.4548, "rle": {"size": [612, 612], "counts": "^eV21Qc02N3M1O2N2N101N101O001N2O001O2M2O100O101N100O2O0O1O2O0O100O101N100O1O101N100O1O101N1O1O1O1O2N1O1O1O1N3M2O1N3N1O2O0O2O000O2O0O2O001N100O2O0O2O000O2O0O101O001N10001O001N10001O001N3N1O1N2O100O100O100O10000OSMP@h2W`0O010O1O010O0101O1N3N1N2N2M2O2M2O1gAeLe<]3TChLm<Y3mBjLT=Y3eBkL\\=`3SBeLm=X42O1N101N101O1N101N2O0O2O0O2N2O2M2N2O1N2000a0_OO100O0001O1O001O001O1N101N101N2N101N101N2O0O2N101N1O2O0O1O2O0O2N1O1O2N1O1O1O001O1O1O1J6K4O200O100O10O0100O100O1O010O100O100O010O100O100O100O2O0Ob^O^N\\a0b1c^O_N]a0`1c^OaN]a0^1c^OcN]a0b11O2N2N001O001O00001O01O00000000001O00000000001O000000001O0000C_^OXO`a0g0b^OXO^a0g0c^OZO\\a0e0f^OZOZa0e0g^O[OYa0d0h^O]OWa0a0k^O_OUa0?l^OCSa0;o^OEQa09Q_OGo`07R_OKm`02V_OMk`0H`_O8ea01O0O2O10O101N2M3M2O2M3N1N3NTfo3"}, "label": "bananas"}]}
{"id": 493793, "image": "COCO_train2014_000000493793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bananas\"."}], "task": "refering", "answer_template": "The \"bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 114, 115, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 189, 206, 207, 208], "bbox": [0.1929575163398693, 0.13264705882352942, 0.6500490196078432, 0.4506862745098039], "iscrowd": 0, "area": 19362.4548, "rle": {"size": [612, 612], "counts": "^eV21Qc02N3M1O2N2N101N101O001N2O001O2M2O100O101N100O2O0O1O2O0O100O101N100O1O101N100O1O101N1O1O1O1O2N1O1O1O1N3M2O1N3N1O2O0O2O000O2O0O2O001N100O2O0O2O000O2O0O101O001N10001O001N10001O001N3N1O1N2O100O100O100O10000OSMP@h2W`0O010O1O010O0101O1N3N1N2N2M2O2M2O1gAeLe<]3TChLm<Y3mBjLT=Y3eBkL\\=`3SBeLm=X42O1N101N101O1N101N2O0O2O0O2N2O2M2N2O1N2000a0_OO100O0001O1O001O001O1N101N101N2N101N101N2O0O2N101N1O2O0O1O2O0O2N1O1O2N1O1O1O001O1O1O1J6K4O200O100O10O0100O100O1O010O100O100O010O100O100O100O2O0Ob^O^N\\a0b1c^O_N]a0`1c^OaN]a0^1c^OcN]a0b11O2N2N001O001O00001O01O00000000001O00000000001O000000001O0000C_^OXO`a0g0b^OXO^a0g0c^OZO\\a0e0f^OZOZa0e0g^O[OYa0d0h^O]OWa0a0k^O_OUa0?l^OCSa0;o^OEQa09Q_OGo`07R_OKm`02V_OMk`0H`_O8ea01O0O2O10O101N2M3M2O2M3N1N3NTfo3"}, "label": "bananas"}]}
{"id": 493793, "image": "COCO_train2014_000000493793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange in front of the bananas\"."}], "task": "refering", "answer_template": "The \"the orange in front of the bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 134, 135, 136, 156, 157, 158, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 269], "bbox": [0.08684640522875817, 0.2366013071895425, 0.3935457516339869, 0.5486601307189543], "iscrowd": 0, "area": 21156.299400000004, "rle": {"size": [612, 612], "counts": "lmo08Yb0i0C<E7I7I7K5M3M4L3M3L4M3M3M3L4M4L3M3M3M3M3N1N3M3M3N2M2O2N2N2N2N2N1O2O1N2N2N1O2N2O1N2N101N2N2O0O2O1N2N101N2O1N1O2O0O100O101N10000O2O000O101O0O10UKVD\\2j;RMhDn2X;kLoDV3o:iLSEW3m:dLXE\\3X=O100000000000000000000001O00000000000000000000000O101O00000000O1000000O1000000O1000000O1000000O10000O1000000O100O010O100O1O100O1O100O100O1O100O1O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3N2N2N2N2N2N2M3M3H9Gj0VOR`m6"}, "label": "the orange in front of the bananas"}]}
{"id": 493793, "image": "COCO_train2014_000000493793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange next to the bananas\"."}], "task": "refering", "answer_template": "The \"the orange next to the bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 134, 135, 136, 156, 157, 158, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 269], "bbox": [0.08684640522875817, 0.2366013071895425, 0.3935457516339869, 0.5486601307189543], "iscrowd": 0, "area": 21156.299400000004, "rle": {"size": [612, 612], "counts": "lmo08Yb0i0C<E7I7I7K5M3M4L3M3L4M3M3M3L4M4L3M3M3M3M3N1N3M3M3N2M2O2N2N2N2N2N1O2O1N2N2N1O2N2O1N2N101N2N2O0O2O1N2N101N2O1N1O2O0O100O101N10000O2O000O101O0O10UKVD\\2j;RMhDn2X;kLoDV3o:iLSEW3m:dLXE\\3X=O100000000000000000000001O00000000000000000000000O101O00000000O1000000O1000000O1000000O1000000O10000O1000000O100O010O100O1O100O1O100O100O1O100O1O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3N2N2N2N2N2N2M3M3H9Gj0VOR`m6"}, "label": "the orange next to the bananas"}]}
{"id": 485602, "image": "COCO_train2014_000000485602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow train travelling under a bridge and along down the railroad tracks\"."}], "task": "refering", "answer_template": "The \"a yellow train travelling under a bridge and along down the railroad tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 81, 103, 104, 125, 126, 127, 148, 149, 150, 171, 172, 173, 194, 195, 196, 217, 218, 219], "bbox": [0.44037499999999996, 0.1859718969555035, 0.5571249999999999, 0.6514519906323184], "iscrowd": 0, "area": 10985.124450000003, "rle": {"size": [427, 640], "counts": "cee36o<6bELRO>Z6FcJ=bNNk6EbJm0SN@Z7CbJ^1cMQOj7AbJa2X5_MgJj2Q5VMoJl2o4TMPKn2o4RMPKP3o4PMQKQ3n4oLQKR3o4nLQKS3n4mLQKU3n4lLQKU3n4kLbJDgNb3g6jL`JHgN_3h6iL^JMgN[3k6hL\\J1aN]3S7bL[Jn3e5RLZJo3f5QLYJP4g5PLYJP4h5oKWJR4i5nKVJS4j5mKbIFI_4e6kK`IHEc4k6gK\\IHCg4Q7bKYIj4h6VKUIl4k6UKnHQ5R7d00000001O0000001O1O001O001O1O1O1O2N1O2N3M3M2N3M4L6J5K5K4L5K4K6K4PLnF^3a9N1O2N1O2N1O2N101N5^OW2gMgie3"}, "label": "a yellow train travelling under a bridge and along down the railroad tracks"}]}
{"id": 485602, "image": "COCO_train2014_000000485602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long train running on the first track\"."}], "task": "refering", "answer_template": "The \"a long train running on the first track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 81, 103, 104, 125, 126, 127, 148, 149, 150, 171, 172, 173, 194, 195, 196, 217, 218, 219], "bbox": [0.44037499999999996, 0.1859718969555035, 0.5571249999999999, 0.6514519906323184], "iscrowd": 0, "area": 10985.124450000003, "rle": {"size": [427, 640], "counts": "cee36o<6bELRO>Z6FcJ=bNNk6EbJm0SN@Z7CbJ^1cMQOj7AbJa2X5_MgJj2Q5VMoJl2o4TMPKn2o4RMPKP3o4PMQKQ3n4oLQKR3o4nLQKS3n4mLQKU3n4lLQKU3n4kLbJDgNb3g6jL`JHgN_3h6iL^JMgN[3k6hL\\J1aN]3S7bL[Jn3e5RLZJo3f5QLYJP4g5PLYJP4h5oKWJR4i5nKVJS4j5mKbIFI_4e6kK`IHEc4k6gK\\IHCg4Q7bKYIj4h6VKUIl4k6UKnHQ5R7d00000001O0000001O1O001O001O1O1O1O2N1O2N3M3M2N3M4L6J5K5K4L5K4K6K4PLnF^3a9N1O2N1O2N1O2N101N5^OW2gMgie3"}, "label": "a long train running on the first track"}]}
{"id": 452837, "image": "COCO_train2014_000000452837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of soda with ice cubes\"."}], "task": "refering", "answer_template": "The \"a glass of soda with ice cubes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 22, 23, 24, 25, 26, 44, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 88, 89, 90, 91, 92, 93, 94, 110, 111, 112, 113, 114, 115, 116, 132, 133, 134, 135, 136, 137, 138, 139, 154, 155, 156, 157, 158, 159, 160, 161, 176, 177, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 291, 292, 293], "bbox": [0.0022549019607843133, 0.01573529411764706, 0.4247222222222222, 0.6359477124183006], "iscrowd": 0, "area": 57467.22295, "rle": {"size": [612, 612], "counts": "ci0=b>V4iNW11OO1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N101N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O100000000000000000000001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O0O2O1O1O1O001O1O1N2O0O2N2O1N1O2O1N2N2O0O2N2N2N2O0O2N2N2N1O2O1N2N2N1O2N2O1N2N1O2N2N2O1N1O2N2N2N101N2N2N2L3L5K5K5K5K4L5K5K5L3L5K5K5L4K4M4L4K5L4L3L5L4L4K5L3M4K5L4M2N3M3M3L4M2N3M3M3M3L3N3M3M3M3M2M4M3M3M2N3M3L4MQSb6"}, "label": "a glass of soda with ice cubes"}]}
{"id": 452837, "image": "COCO_train2014_000000452837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a full glass of soda next to a plate of pasta\"."}], "task": "refering", "answer_template": "The \"a full glass of soda next to a plate of pasta\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 22, 23, 24, 25, 26, 44, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 88, 89, 90, 91, 92, 93, 94, 110, 111, 112, 113, 114, 115, 116, 132, 133, 134, 135, 136, 137, 138, 139, 154, 155, 156, 157, 158, 159, 160, 161, 176, 177, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 291, 292, 293], "bbox": [0.0022549019607843133, 0.01573529411764706, 0.4247222222222222, 0.6359477124183006], "iscrowd": 0, "area": 57467.22295, "rle": {"size": [612, 612], "counts": "ci0=b>V4iNW11OO1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N101N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O100000000000000000000001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O0O2O1O1O1O001O1O1N2O0O2N2O1N1O2O1N2N2O0O2N2N2N2O0O2N2N2N1O2O1N2N2N1O2N2O1N2N1O2N2N2O1N1O2N2N2N101N2N2N2L3L5K5K5K5K4L5K5K5L3L5K5K5L4K4M4L4K5L4L3L5L4L4K5L3M4K5L4M2N3M3M3L4M2N3M3M3M3L3N3M3M3M3M2M4M3M3M2N3M3L4MQSb6"}, "label": "a full glass of soda next to a plate of pasta"}]}
{"id": 452837, "image": "COCO_train2014_000000452837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottle with label on it\"."}], "task": "refering", "answer_template": "The \"bottle with label on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 185, 186, 187, 207, 208], "bbox": [0.1820261437908497, 0.0028104575163398694, 0.6606699346405229, 0.42303921568627456], "iscrowd": 0, "area": 40888.16679999999, "rle": {"size": [612, 612], "counts": "RnR21Sc02N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N101N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2O0O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N101N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N101N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N3M2N3M0000000000000000000000000000000000O1O2N1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O3M3M4M2M3M3M3M3M3M3M3M3M4L3N2M3M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N3M2M3N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N3M2M3N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N3L3N2M3M3N2M3M3N[Tl3"}, "label": "bottle with label on it"}]}
{"id": 452837, "image": "COCO_train2014_000000452837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a bottle of beer\"."}], "task": "refering", "answer_template": "The \"the side of a bottle of beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 185, 186, 187, 207, 208], "bbox": [0.1820261437908497, 0.0028104575163398694, 0.6606699346405229, 0.42303921568627456], "iscrowd": 0, "area": 40888.16679999999, "rle": {"size": [612, 612], "counts": "RnR21Sc02N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N101N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2O0O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N101N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N101N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N3M2N3M0000000000000000000000000000000000O1O2N1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O3M3M4M2M3M3M3M3M3M3M3M3M4L3N2M3M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N3M2M3N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N3M2M3N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N3L3N2M3M3N2M3M3N[Tl3"}, "label": "the side of a bottle of beer"}]}
{"id": 411877, "image": "COCO_train2014_000000411877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra facing away from camera\"."}], "task": "refering", "answer_template": "The \"zebra facing away from camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 137, 139], "bbox": [0.0872599531615925, 0.164515625, 0.3927166276346604, 0.41578125000000005], "iscrowd": 0, "area": 12680.560050000002, "rle": {"size": [640, 427], "counts": "_Si01ok1a0_gNT1lNT1kNe0\\O3M3M3M3M2M4M3M3M3M3M3L4M3M1OfNl@iMS?U2RAhMn>V2WAhMg>X2]AeMc>Y2cAcM]>[2hAcMW>[2nAbMR>\\2SBaMm=]2XBaMf=^2_B_Ma=_2eB^MZ=`2kB]MU=b2oB[MQ=c2TC[Mk<c2Y2O2N201O00000000001O000O1000001O01O9G8H9G9H8G8H5K1O1O100O1OnM_@F`?:e@BZ?>k@[OX?d0m@lN^?T1h@[Nc?d1c@mMg?S2S10001O0o1QN<D4L00010OZOi@cLW?Z3PAbLP?Z3XAcLg>Y3aAcL`>Y3hAbLY>Y3Y1M4K5K5L4K4N3N2M3N2M2Ok^OPN]`0m1X_OaNh`0\\20000O100000000O11O001O001O00001O001@?_Ob0G8L5K4L5K7I9G:F:F`fQ5"}, "label": "zebra facing away from camera"}]}
{"id": 411877, "image": "COCO_train2014_000000411877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra is eating facing the wall\"."}], "task": "refering", "answer_template": "The \"the zebra is eating facing the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 137, 139], "bbox": [0.0872599531615925, 0.164515625, 0.3927166276346604, 0.41578125000000005], "iscrowd": 0, "area": 12680.560050000002, "rle": {"size": [640, 427], "counts": "_Si01ok1a0_gNT1lNT1kNe0\\O3M3M3M3M2M4M3M3M3M3M3L4M3M1OfNl@iMS?U2RAhMn>V2WAhMg>X2]AeMc>Y2cAcM]>[2hAcMW>[2nAbMR>\\2SBaMm=]2XBaMf=^2_B_Ma=_2eB^MZ=`2kB]MU=b2oB[MQ=c2TC[Mk<c2Y2O2N201O00000000001O000O1000001O01O9G8H9G9H8G8H5K1O1O100O1OnM_@F`?:e@BZ?>k@[OX?d0m@lN^?T1h@[Nc?d1c@mMg?S2S10001O0o1QN<D4L00010OZOi@cLW?Z3PAbLP?Z3XAcLg>Y3aAcL`>Y3hAbLY>Y3Y1M4K5K5L4K4N3N2M3N2M2Ok^OPN]`0m1X_OaNh`0\\20000O100000000O11O001O001O00001O001@?_Ob0G8L5K4L5K7I9G:F:F`fQ5"}, "label": "the zebra is eating facing the wall"}]}
{"id": 411877, "image": "COCO_train2014_000000411877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with another zebra standing immediately behind it\"."}], "task": "refering", "answer_template": "The \"a zebra with another zebra standing immediately behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 110, 111, 112, 114, 115, 116, 117, 125, 126, 127, 129, 130, 131, 140, 141, 142, 145, 146], "bbox": [0.36569086651053867, 0.17089062500000002, 0.8276346604215457, 0.41312499999999996], "iscrowd": 0, "area": 17296.200649999995, "rle": {"size": [640, 427], "counts": "TgQ3e0_b0U1F3M2N2]NTNn@n1j>]NQAe1j>cNQA_1k>iNn@4^OJa?;i@JMGV?i0d@@=Cn>Q1`@[Oj0_Of>Z2_AaMb>^2bA[Mb>d2^1O0O101O000O2O001O1N101O1O0O2O1O00n^OSNS`0k1n_O_Ni?`1V@fNe?Z1Z@lNb?S1^@TO\\?m0a@[OY?f0d@^OZ?b0g@^OX?c0h@^OV?e0g@]OV?h0f@YOY?j0d@kM;`0o>U2UAjMj>U2XAjMh>U2[AjMd>T2_AkMa>T2bAkM]>T2fAjMZ>T2iAlMV>S2mAkMR>U2PBjMP>U2SBiMn=U2UBiMk=V2YBgMg=W2]BgMc=X2`BfM`=R2jBlMV=T2mBeMW=\\2V2001O1O01O01O0000010O000001O01O0001O000000001O0000001O0000000000000010O00001O00001O0000000000000000O100000000O100000000:F:m^O_MS`0[3L8H1O1O0O1000O1N2O1ZOf0M3N2N101N3N1N2O2N1N2O2N1O1O2OU_OaMU`0^2i_OgMU`0Y2g_OlMX`0Q3N6V@[Lk>l3i@ZLV?X41N10001O00100O1O1O1O2N1O1M3M3]O[@fLi?W3k0Dl0UO>B2N2E;L3M4M3L4M21O0O101N1O1O001OF:02N1O2N1L4M4N1N3NWR^1"}, "label": "a zebra with another zebra standing immediately behind it"}]}
{"id": 411877, "image": "COCO_train2014_000000411877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zibera in the zoe\"."}], "task": "refering", "answer_template": "The \"zibera in the zoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 110, 111, 112, 114, 115, 116, 117, 125, 126, 127, 129, 130, 131, 140, 141, 142, 145, 146], "bbox": [0.36569086651053867, 0.17089062500000002, 0.8276346604215457, 0.41312499999999996], "iscrowd": 0, "area": 17296.200649999995, "rle": {"size": [640, 427], "counts": "TgQ3e0_b0U1F3M2N2]NTNn@n1j>]NQAe1j>cNQA_1k>iNn@4^OJa?;i@JMGV?i0d@@=Cn>Q1`@[Oj0_Of>Z2_AaMb>^2bA[Mb>d2^1O0O101O000O2O001O1N101O1O0O2O1O00n^OSNS`0k1n_O_Ni?`1V@fNe?Z1Z@lNb?S1^@TO\\?m0a@[OY?f0d@^OZ?b0g@^OX?c0h@^OV?e0g@]OV?h0f@YOY?j0d@kM;`0o>U2UAjMj>U2XAjMh>U2[AjMd>T2_AkMa>T2bAkM]>T2fAjMZ>T2iAlMV>S2mAkMR>U2PBjMP>U2SBiMn=U2UBiMk=V2YBgMg=W2]BgMc=X2`BfM`=R2jBlMV=T2mBeMW=\\2V2001O1O01O01O0000010O000001O01O0001O000000001O0000001O0000000000000010O00001O00001O0000000000000000O100000000O100000000:F:m^O_MS`0[3L8H1O1O0O1000O1N2O1ZOf0M3N2N101N3N1N2O2N1N2O2N1O1O2OU_OaMU`0^2i_OgMU`0Y2g_OlMX`0Q3N6V@[Lk>l3i@ZLV?X41N10001O00100O1O1O1O2N1O1M3M3]O[@fLi?W3k0Dl0UO>B2N2E;L3M4M3L4M21O0O101N1O1O001OF:02N1O2N1L4M4N1N3NWR^1"}, "label": "zibera in the zoe"}]}
{"id": 10471, "image": "COCO_train2014_000000010471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with a tusk standing between two other elephants\"."}], "task": "refering", "answer_template": "The \"an elephant with a tusk standing between two other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 218, 219, 235, 236, 237, 238, 241, 242, 258, 259, 260, 261, 264, 265], "bbox": [0.10414062500000001, 0.20850117096018736, 0.57540625, 0.7799297423887587], "iscrowd": 0, "area": 44242.52205, "rle": {"size": [427, 640], "counts": "gTl0P1o;>M2M3N3L3M4M2M4M2M3N3L3N3L3N3L3N2M3M2N2M3M3N2M2O2M3MB[F\\Mc9f2]FZM`9h2aFXM]9i2dFWMY9k2hFUMV9l2kFTMR9o2nFQMo8Q3RGoLl8R3UGnLh8T3YGlLe8U3\\GkLb8W3^GhLa8Y3`GgL_8Y3bGgL]8Y3dGgL[8Y3fGgLX8[3iGdLV8\\3kGdLT8\\3mGdLR8\\3oGdLo7^3QHbLn7^3SHbLl7^3UHbLi7_3XHaLg7_3ZH`Lf7a3ZH_Le7a3\\H^Lc7b3_H^L`7a3bH^L]7a3fH_LX7a3jH^LU7b3mH^LQ7b3QI]Ln6b3XI[Lf6e3`IVL`6i3b1N3L3N2N2K5LRJjLo1Q3UNQMk1n2TNSMl1m2SNTMm1m2RNSMn1m2QNTMo1m2oMUMP2k2oMVMQ2k2mMVMT2i2kMXMU2i2iMXMW2h2hMYMY2g2PMPNo2P2RLoNo3Q1ZKEf4;WKHj48RKKn45PKNP51mJ2S5OiJ4W5LgJ6Z5JbJ9^5H^J<a5D]J>c5CZJ?e5CWJ`0i5ATJa0l5_ORJc0n5^OoIe0o5\\300000001NCSJ^Im5]6dJWI\\5d6h0K4K6K5K5K4L5N2M3M2N3M3N2M3M2N3N2M3M2N3M3N2O07J5K5K5K5J7J5K5L4K6K4^IVJZ5n5^JUJb5P6VJSJk5f6001O001O1O001O001O001O1O001O001O00001O000lMUKhKl4U4XKjKg4T4^KiKc4T4bKiK^4T4hKiKX4U4mKhKT4U4PLiKP4U4ULhKk3V4ZLgKg3V4h2M3N3M20001O010O001O0010O01O001O010O001O0001O0O1000001N1000000O2O00000O101O000O1000001N1000000O2O00001N10001O0O2O00001O001O00001O001OO10[IaK^4_4`KdK_4[4aKhK]4X4aKkK^4U4aKmK^4R4aKQL^4o3`KTL_4k3mIdKY1c0h4j3oIeKU1d0k4g3PJgKR1d0m4d3QJkKn0b0R5c3PJlKk0c0U5`3PJPLh0`0Y5`3nISLe0?]5^3nInMS6Q2mIPNS6P2mIoMT6Q2kIPNU6o1lIPNU6P2jIQNV6n1kIQNV6o1iIRNX6m1hISNY6j1hIWNY6g1hIYNX6e1iI\\NX6b1iI^NX6_1iIbNX6\\1iIdNX6Y1iIhNX6V1hIkNY6R1hIoNX6P1hIQOY6l0hIUOY6i0gIXOZ6e0gI\\OZ6]OkJd0V5XOlJi0U5SOmJn0S5nNPKS1Q5iNUKT1l4hN[KT1f4hN`KU1a4gNfKU1[4fNlKW1U4eNRLW1n3fNXLW1i3eN^LW1c3hN`LU1a3iNdLS1]3lNiLm0Y3ROoLf0R3XOVM`0l2_O[M9f2FcM2_2LhMLZ23`5N2O1O1N2O2NPYa3"}, "label": "an elephant with a tusk standing between two other elephants"}]}
{"id": 10471, "image": "COCO_train2014_000000010471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray elephant with large white tusks\"."}], "task": "refering", "answer_template": "The \"a gray elephant with large white tusks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 218, 219, 235, 236, 237, 238, 241, 242, 258, 259, 260, 261, 264, 265], "bbox": [0.10414062500000001, 0.20850117096018736, 0.57540625, 0.7799297423887587], "iscrowd": 0, "area": 44242.52205, "rle": {"size": [427, 640], "counts": "gTl0P1o;>M2M3N3L3M4M2M4M2M3N3L3N3L3N3L3N2M3M2N2M3M3N2M2O2M3MB[F\\Mc9f2]FZM`9h2aFXM]9i2dFWMY9k2hFUMV9l2kFTMR9o2nFQMo8Q3RGoLl8R3UGnLh8T3YGlLe8U3\\GkLb8W3^GhLa8Y3`GgL_8Y3bGgL]8Y3dGgL[8Y3fGgLX8[3iGdLV8\\3kGdLT8\\3mGdLR8\\3oGdLo7^3QHbLn7^3SHbLl7^3UHbLi7_3XHaLg7_3ZH`Lf7a3ZH_Le7a3\\H^Lc7b3_H^L`7a3bH^L]7a3fH_LX7a3jH^LU7b3mH^LQ7b3QI]Ln6b3XI[Lf6e3`IVL`6i3b1N3L3N2N2K5LRJjLo1Q3UNQMk1n2TNSMl1m2SNTMm1m2RNSMn1m2QNTMo1m2oMUMP2k2oMVMQ2k2mMVMT2i2kMXMU2i2iMXMW2h2hMYMY2g2PMPNo2P2RLoNo3Q1ZKEf4;WKHj48RKKn45PKNP51mJ2S5OiJ4W5LgJ6Z5JbJ9^5H^J<a5D]J>c5CZJ?e5CWJ`0i5ATJa0l5_ORJc0n5^OoIe0o5\\300000001NCSJ^Im5]6dJWI\\5d6h0K4K6K5K5K4L5N2M3M2N3M3N2M3M2N3N2M3M2N3M3N2O07J5K5K5K5J7J5K5L4K6K4^IVJZ5n5^JUJb5P6VJSJk5f6001O001O1O001O001O001O1O001O001O00001O000lMUKhKl4U4XKjKg4T4^KiKc4T4bKiK^4T4hKiKX4U4mKhKT4U4PLiKP4U4ULhKk3V4ZLgKg3V4h2M3N3M20001O010O001O0010O01O001O010O001O0001O0O1000001N1000000O2O00000O101O000O1000001N1000000O2O00001N10001O0O2O00001O001O00001O001OO10[IaK^4_4`KdK_4[4aKhK]4X4aKkK^4U4aKmK^4R4aKQL^4o3`KTL_4k3mIdKY1c0h4j3oIeKU1d0k4g3PJgKR1d0m4d3QJkKn0b0R5c3PJlKk0c0U5`3PJPLh0`0Y5`3nISLe0?]5^3nInMS6Q2mIPNS6P2mIoMT6Q2kIPNU6o1lIPNU6P2jIQNV6n1kIQNV6o1iIRNX6m1hISNY6j1hIWNY6g1hIYNX6e1iI\\NX6b1iI^NX6_1iIbNX6\\1iIdNX6Y1iIhNX6V1hIkNY6R1hIoNX6P1hIQOY6l0hIUOY6i0gIXOZ6e0gI\\OZ6]OkJd0V5XOlJi0U5SOmJn0S5nNPKS1Q5iNUKT1l4hN[KT1f4hN`KU1a4gNfKU1[4fNlKW1U4eNRLW1n3fNXLW1i3eN^LW1c3hN`LU1a3iNdLS1]3lNiLm0Y3ROoLf0R3XOVM`0l2_O[M9f2FcM2_2LhMLZ23`5N2O1O1N2O2NPYa3"}, "label": "a gray elephant with large white tusks"}]}
{"id": 10471, "image": "COCO_train2014_000000010471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant facign away\"."}], "task": "refering", "answer_template": "The \"an elephant facign away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 162, 163, 185, 186, 208, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257], "bbox": [0.025437500000000002, 0.2061592505854801, 0.203453125, 0.7734192037470726], "iscrowd": 0, "area": 14711.9924, "rle": {"size": [427, 640], "counts": "di65U=d0\\O`0@9G9G7I3N3L4L4L3M4L4L3N3L4L4L3M4L4L7I7I7I7VIaKZ4g4_KZK`4m4ZKSKe4T5TKlJl4[5mJfJR5a5hJ_JW5i5aJXJ^5R6XJoIg5Z6QJeIo5d6hI]IW6l6O1O1O1O1O1O1O1O1O1O1O1O10000000bMTJ]Ml5^2ZJaMf5Y2aJfM_5U2gJjMY5Q2mJnMS5l1TKSNl4h1ZKWNf4d1`K[N`4_1gK`NY4[1mKdNS4W1SLhNn3Q1ZLmNf3Q1]LnNc3o0aLPO`3m0cLSO\\3k0gLTOY3j0jLUOW3h0lLWOT3g0oLXOQ3f0RMZOn2b0VM]Oj2a0YM^Og2`0\\M_Od2?^MAc2<_ME`29bMG^27dMI]23fMMZ21hM0W2NkM2V2KlM5T2HoM8Q2FQN:P2BSN>m1@UN`0k1]OXNc0i1YOZNh0e1VO]Nj0c1SOaNl0_1SObNm0_1QObNo0^1POcNP1]1nNeNR1\\1lNeNT1[1kNfNU1[1iNfNW1[1gNfNY1Z1fNhNY1Y1dNiN\\1W1cNjN]1W1aNkN^1V1`NkN`1U1_NlNa1U1]NmNb1S1\\NoNd1\\601O00001N6K5K5J6K6Jj[d6"}, "label": "an elephant facign away"}]}
{"id": 10471, "image": "COCO_train2014_000000010471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant with its tail facing the camera\"."}], "task": "refering", "answer_template": "The \"the elephant with its tail facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 162, 163, 185, 186, 208, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257], "bbox": [0.025437500000000002, 0.2061592505854801, 0.203453125, 0.7734192037470726], "iscrowd": 0, "area": 14711.9924, "rle": {"size": [427, 640], "counts": "di65U=d0\\O`0@9G9G7I3N3L4L4L3M4L4L3N3L4L4L3M4L4L7I7I7I7VIaKZ4g4_KZK`4m4ZKSKe4T5TKlJl4[5mJfJR5a5hJ_JW5i5aJXJ^5R6XJoIg5Z6QJeIo5d6hI]IW6l6O1O1O1O1O1O1O1O1O1O1O1O10000000bMTJ]Ml5^2ZJaMf5Y2aJfM_5U2gJjMY5Q2mJnMS5l1TKSNl4h1ZKWNf4d1`K[N`4_1gK`NY4[1mKdNS4W1SLhNn3Q1ZLmNf3Q1]LnNc3o0aLPO`3m0cLSO\\3k0gLTOY3j0jLUOW3h0lLWOT3g0oLXOQ3f0RMZOn2b0VM]Oj2a0YM^Og2`0\\M_Od2?^MAc2<_ME`29bMG^27dMI]23fMMZ21hM0W2NkM2V2KlM5T2HoM8Q2FQN:P2BSN>m1@UN`0k1]OXNc0i1YOZNh0e1VO]Nj0c1SOaNl0_1SObNm0_1QObNo0^1POcNP1]1nNeNR1\\1lNeNT1[1kNfNU1[1iNfNW1[1gNfNY1Z1fNhNY1Y1dNiN\\1W1cNjN]1W1aNkN^1V1`NkN`1U1_NlNa1U1]NmNb1S1\\NoNd1\\601O00001N6K5K5J6K6Jj[d6"}, "label": "the elephant with its tail facing the camera"}]}
{"id": 10471, "image": "COCO_train2014_000000010471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant\"."}], "task": "refering", "answer_template": "The \"baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 201, 203, 204, 216, 217, 219, 220, 221, 222, 223, 224, 226, 227, 242, 243, 244, 245, 246, 247, 250, 251, 265, 266, 267, 268, 269, 270], "bbox": [0.42623437500000005, 0.2903981264637002, 0.9237187500000001, 0.8037704918032788], "iscrowd": 0, "area": 33450.505750000004, "rle": {"size": [427, 640], "counts": "mRb34V=3M10O01O1O11O000000010O01O1OO1000000O10000O2OO010O01000O010O010O010O10O010O010O010O0100O010O0100O0100O001O001O001N2O001O00000O1000O10kN1oDOo:a0bD@];W1N5K6J6K5J6J5K6K8G8H9G9H7I8J`0_Oe0\\O3L2O1N2O0O100O100O1O100O100O100OnMmHnMS7R2kIPMT6R3QJhLo5X3WJbLh5_3^JZLb5g3cJTL]5l3gJPLX5Q4Q24K5K<lG\\Kk6b5O10O0O2N2N2N3M2J6F:G9L44L3M3M3M3M4bHWKS6k4hIXKW6k4dIXK[6k4`IXK_6k4[IYKe6j4UIZKi6d500001O0000001O0000001O0O10001O000000O100OBUI^Jj6a5\\I[Jd6d5aIXJ^6g5d0N1O1O2H7I7J7J5K6I6N201N101N2O2M2O1N2N2KZIWLU4\\3QLlLm3T3lKUMR4k2gK^MW4b2bKfM^4Y2[KnMe4R2mJ\\NS5d1XJQOh5P1cID]6e31O0000001O00000000O100O10000O100O1000WOhI_JW6S5ZJkJf5f4jJYKW5a4PK]KP5_4UK`Kl4Z4ZKdKg4W4`KfKb4T4dKiK^4S4_2K6J5K6K4K6J5K6K4K5K6O0010O01O010O01O01O010O010O0001OM3M4O000000001O0000000100O1O1O1O2O1N2N2N2O2N1O100O010O01O10O01O010O06K4K5L3M3L4MO0010O010O0100O010O2lNTFdNn9Q1^FlNc9o0dFnN^9m0hFPOZ9k0lFQOW9l0kFSOW9k0kFSOW9k0jFSOY98iEEo02Z9MQF1e02W:MiE2X:LjE4W:JjE5W:JjE6W:HjE8d;010O10O10OO2O1O2N2NXTd0"}, "label": "baby elephant"}]}
{"id": 10471, "image": "COCO_train2014_000000010471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller of the three elephants\"."}], "task": "refering", "answer_template": "The \"the smaller of the three elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 201, 203, 204, 216, 217, 219, 220, 221, 222, 223, 224, 226, 227, 242, 243, 244, 245, 246, 247, 250, 251, 265, 266, 267, 268, 269, 270], "bbox": [0.42623437500000005, 0.2903981264637002, 0.9237187500000001, 0.8037704918032788], "iscrowd": 0, "area": 33450.505750000004, "rle": {"size": [427, 640], "counts": "mRb34V=3M10O01O1O11O000000010O01O1OO1000000O10000O2OO010O01000O010O010O010O10O010O010O010O0100O010O0100O0100O001O001O001N2O001O00000O1000O10kN1oDOo:a0bD@];W1N5K6J6K5J6J5K6K8G8H9G9H7I8J`0_Oe0\\O3L2O1N2O0O100O100O1O100O100O100OnMmHnMS7R2kIPMT6R3QJhLo5X3WJbLh5_3^JZLb5g3cJTL]5l3gJPLX5Q4Q24K5K<lG\\Kk6b5O10O0O2N2N2N3M2J6F:G9L44L3M3M3M3M4bHWKS6k4hIXKW6k4dIXK[6k4`IXK_6k4[IYKe6j4UIZKi6d500001O0000001O0000001O0O10001O000000O100OBUI^Jj6a5\\I[Jd6d5aIXJ^6g5d0N1O1O2H7I7J7J5K6I6N201N101N2O2M2O1N2N2KZIWLU4\\3QLlLm3T3lKUMR4k2gK^MW4b2bKfM^4Y2[KnMe4R2mJ\\NS5d1XJQOh5P1cID]6e31O0000001O00000000O100O10000O100O1000WOhI_JW6S5ZJkJf5f4jJYKW5a4PK]KP5_4UK`Kl4Z4ZKdKg4W4`KfKb4T4dKiK^4S4_2K6J5K6K4K6J5K6K4K5K6O0010O01O010O01O01O010O010O0001OM3M4O000000001O0000000100O1O1O1O2O1N2N2N2O2N1O100O010O01O10O01O010O06K4K5L3M3L4MO0010O010O0100O010O2lNTFdNn9Q1^FlNc9o0dFnN^9m0hFPOZ9k0lFQOW9l0kFSOW9k0kFSOW9k0jFSOY98iEEo02Z9MQF1e02W:MiE2X:LjE4W:JjE5W:JjE6W:HjE8d;010O10O10OO2O1O2N2NXTd0"}, "label": "the smaller of the three elephants"}]}
{"id": 157926, "image": "COCO_train2014_000000157926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ferry carrying trucks\"."}], "task": "refering", "answer_template": "The \"a ferry carrying trucks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 63, 64, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.006, 0.141203125, 0.99775, 0.986140625], "iscrowd": 0, "area": 230301.21175, "rle": {"size": [640, 480], "counts": "_]2U2i>R3nLR3QOo0O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N3M2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100000000000000000000O1000000000000000000000000000001O0000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O1001O001O00001O001O001O001O001O01O01O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001Oe9[F]5cJ00Zd0"}, "label": "a ferry carrying trucks"}]}
{"id": 157926, "image": "COCO_train2014_000000157926.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"to boats carrying vegetables travelling in the water\"."}], "task": "refering", "answer_template": "The \"to boats carrying vegetables travelling in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 63, 64, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.006, 0.141203125, 0.99775, 0.986140625], "iscrowd": 0, "area": 230301.21175, "rle": {"size": [640, 480], "counts": "_]2U2i>R3nLR3QOo0O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N3M2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100000000000000000000O1000000000000000000000000000001O0000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O1001O001O00001O001O001O001O001O01O01O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001Oe9[F]5cJ00Zd0"}, "label": "to boats carrying vegetables travelling in the water"}]}
{"id": 231657, "image": "COCO_train2014_000000231657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mutton food in the plate\"."}], "task": "refering", "answer_template": "The \"a mutton food in the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.00196, 0.0013066666666666667, 1.0, 0.9882933333333334], "iscrowd": 0, "area": 181051.50299999997, "rle": {"size": [375, 500], "counts": "i;P;g00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000001O0000000000000000"}, "label": "a mutton food in the plate"}]}
{"id": 231657, "image": "COCO_train2014_000000231657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow plate and silverware\"."}], "task": "refering", "answer_template": "The \"a yellow plate and silverware\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.00196, 0.0013066666666666667, 1.0, 0.9882933333333334], "iscrowd": 0, "area": 181051.50299999997, "rle": {"size": [375, 500], "counts": "i;P;g00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000001O0000000000000000"}, "label": "a yellow plate and silverware"}]}
{"id": 231657, "image": "COCO_train2014_000000231657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver fork on the plate\"."}], "task": "refering", "answer_template": "The \"a silver fork on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 65, 83, 84, 101, 102, 103, 120, 121, 122, 139, 140, 141, 158, 159, 160, 177, 178, 196, 197], "bbox": [0.54748, 0.18341333333333334, 0.96012, 0.82464], "iscrowd": 0, "area": 7488.841700000002, "rle": {"size": [375, 500], "counts": "TiT31g;001O001N1N2N3M2N3M2M4M2N2N3M2M4M2O2O0O101N101N101O0O101N101N10001N101N101N10001N101O01O001N101O001O001N101O001O001N101O001O001N100001O0000001O0000001O00001O0000001OPOkE=V:]OoEc0Q:XOUFg0^:1O0000001O0000010O0000001O00000010O000001O0000001O01O01O0000001O0001O01O0000001O0000010O0000001O00000010O000001O0000001O01O01O0000001O0000010O0000001O0000010O1O001O1O001O1O010O001O1O001OO1O101N1O1O1O1O100O1O1O2N1O1O1O1N2N2N2N2N3M4M3L4L5KSQ7"}, "label": "a silver fork on the plate"}]}
{"id": 231657, "image": "COCO_train2014_000000231657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork on a plate\"."}], "task": "refering", "answer_template": "The \"a fork on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 65, 83, 84, 101, 102, 103, 120, 121, 122, 139, 140, 141, 158, 159, 160, 177, 178, 196, 197], "bbox": [0.54748, 0.18341333333333334, 0.96012, 0.82464], "iscrowd": 0, "area": 7488.841700000002, "rle": {"size": [375, 500], "counts": "TiT31g;001O001N1N2N3M2N3M2M4M2N2N3M2M4M2O2O0O101N101N101O0O101N101N10001N101N101N10001N101O01O001N101O001O001N101O001O001N101O001O001N100001O0000001O0000001O00001O0000001OPOkE=V:]OoEc0Q:XOUFg0^:1O0000001O0000010O0000001O00000010O000001O0000001O01O01O0000001O0001O01O0000001O0000010O0000001O00000010O000001O0000001O01O01O0000001O0000010O0000001O0000010O1O001O1O001O1O010O001O1O001OO1O101N1O1O1O1O100O1O1O2N1O1O1O1N2N2N2N2N3M4M3L4L5KSQ7"}, "label": "a fork on a plate"}]}
{"id": 100586, "image": "COCO_train2014_000000100586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right monitor of a 3 screen computer set up\"."}], "task": "refering", "answer_template": "The \"the right monitor of a 3 screen computer set up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136], "bbox": [0.66015625, 0.10741666666666667, 0.949984375, 0.44044444444444447], "iscrowd": 0, "area": 20121.395549999994, "rle": {"size": [360, 640], "counts": "Pkd4>[:?A?A?B>A?A?H800000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000001O0001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000000001O00000001O0000000000000000000000000000000000001O00K5F:F:F:F:G9F:F:F:F:G9FQW;"}, "label": "the right monitor of a 3 screen computer set up"}]}
{"id": 100586, "image": "COCO_train2014_000000100586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rightmost computer monitor\"."}], "task": "refering", "answer_template": "The \"the rightmost computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136], "bbox": [0.66015625, 0.10741666666666667, 0.949984375, 0.44044444444444447], "iscrowd": 0, "area": 20121.395549999994, "rle": {"size": [360, 640], "counts": "Pkd4>[:?A?A?B>A?A?H800000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000001O0001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000000001O00000001O0000000000000000000000000000000000001O00K5F:F:F:F:G9F:F:F:F:G9FQW;"}, "label": "the rightmost computer monitor"}]}
{"id": 100586, "image": "COCO_train2014_000000100586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the largest of the three monitors\"."}], "task": "refering", "answer_template": "The \"the largest of the three monitors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152], "bbox": [0.327875, 0.08319444444444445, 0.660140625, 0.5236111111111111], "iscrowd": 0, "area": 31414.94545, "rle": {"size": [360, 640], "counts": "^kY2_3i7_1aN0000000000O10000000000O1000000O10000O100O100O10000O100001O00001O0000001O0000001O0000001O00000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O1000000000000000000000000G9_Oa0^Ob0_Oa0_Oa0_Oa0_Oa0^OTd\\2"}, "label": "the largest of the three monitors"}]}
{"id": 100586, "image": "COCO_train2014_000000100586.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pc monitor between another two\"."}], "task": "refering", "answer_template": "The \"a pc monitor between another two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152], "bbox": [0.327875, 0.08319444444444445, 0.660140625, 0.5236111111111111], "iscrowd": 0, "area": 31414.94545, "rle": {"size": [360, 640], "counts": "^kY2_3i7_1aN0000000000O10000000000O1000000O10000O100O100O10000O100001O00001O0000001O0000001O0000001O00000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O1000000000000000000000000G9_Oa0^Ob0_Oa0_Oa0_Oa0_Oa0^OTd\\2"}, "label": "a pc monitor between another two"}]}
{"id": 207083, "image": "COCO_train2014_000000207083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 110, 111, 112, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.7281875, 0.21275000000000002, 1.0, 0.7451875], "iscrowd": 0, "area": 29880.43795000001, "rle": {"size": [480, 640], "counts": "^fj62k>4M3L4L4L4L3N3N2O1O1N2O1O1N2O001O1N2O1O1O1N2O0G:E;E;F:G9WOXMoDQ3d:m0M3M3N2M3M2N3M3O1N2N2N2O0O2N2N2N2N2M2O2N2N2N2L4K4L5K5K5K5M3N101N2O1N2O1N101N2O1O1N2O001O1O1N2O1O001O1N101O001O0O2O1O010O0100O0100O010O010O010O10O010O10O010O10O010O010O2N4L4L3M4L9Gc0]O1O001O001O00001O001O001O001O001N100O2O001N101N101O0O2O1N101O1N2O001N2O1N101O1N101N2O1O2M2O2M2O1O2M2O1O2M2OhI"}, "label": "a man wearing a black shirt"}]}
{"id": 207083, "image": "COCO_train2014_000000207083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black t shirt and jeans seated o floor with back to camera talking on phone\"."}], "task": "refering", "answer_template": "The \"man in black t shirt and jeans seated o floor with back to camera talking on phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 110, 111, 112, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.7281875, 0.21275000000000002, 1.0, 0.7451875], "iscrowd": 0, "area": 29880.43795000001, "rle": {"size": [480, 640], "counts": "^fj62k>4M3L4L4L4L3N3N2O1O1N2O1O1N2O001O1N2O1O1O1N2O0G:E;E;F:G9WOXMoDQ3d:m0M3M3N2M3M2N3M3O1N2N2N2O0O2N2N2N2N2M2O2N2N2N2L4K4L5K5K5K5M3N101N2O1N2O1N101N2O1O1N2O001O1O1N2O1O001O1N101O001O0O2O1O010O0100O0100O010O010O010O10O010O10O010O10O010O010O2N4L4L3M4L9Gc0]O1O001O001O00001O001O001O001O001N100O2O001N101N101O0O2O1N101O1N2O001N2O1N101O1N101N2O1O2M2O2M2O1O2M2O1O2M2OhI"}, "label": "man in black t shirt and jeans seated o floor with back to camera talking on phone"}]}
{"id": 207083, "image": "COCO_train2014_000000207083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white shirt , jeans , and a black backpack\"."}], "task": "refering", "answer_template": "The \"a man wearing a white shirt , jeans , and a black backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 42, 43, 44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 136, 137], "bbox": [0.81925, 0.0011250000000000001, 0.981421875, 0.3592291666666666], "iscrowd": 0, "area": 9504.6899, "rle": {"size": [480, 640], "counts": "aSf77e>8I6I7J6L2OO100O1O100O010O1O100O100O1O1001O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N1oCSNm:n1PEVNn:k1oDXNP;i1mD[NQ;f1lD]NS;d1iD`NV;a1gDcNW;^1fDeNY;\\1dDhNZ;Y1cDjN\\;_2N5K4L4L4L4L5K4L4L4L4L3M1O001O00001O00001O0000001O0002O1N1O2N101N2N101N2N1O2O1N1O3N5J6SNUEYOP;>aEUOe:b0kEROZ:e0VFnNn9k0aFiNc9o0^2H7I7H8Iba5"}, "label": "a man wearing a white shirt , jeans , and a black backpack"}]}
{"id": 207083, "image": "COCO_train2014_000000207083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a white t shirt and blue jeans with a back pack on his back\"."}], "task": "refering", "answer_template": "The \"man wearing a white t shirt and blue jeans with a back pack on his back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 42, 43, 44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 136, 137], "bbox": [0.81925, 0.0011250000000000001, 0.981421875, 0.3592291666666666], "iscrowd": 0, "area": 9504.6899, "rle": {"size": [480, 640], "counts": "aSf77e>8I6I7J6L2OO100O1O100O010O1O100O100O1O1001O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N1oCSNm:n1PEVNn:k1oDXNP;i1mD[NQ;f1lD]NS;d1iD`NV;a1gDcNW;^1fDeNY;\\1dDhNZ;Y1cDjN\\;_2N5K4L4L4L4L5K4L4L4L4L3M1O001O00001O00001O0000001O0002O1N1O2N101N2N101N2N1O2O1N1O3N5J6SNUEYOP;>aEUOe:b0kEROZ:e0VFnNn9k0aFiNc9o0^2H7I7H8Iba5"}, "label": "man wearing a white t shirt and blue jeans with a back pack on his back"}]}
{"id": 207083, "image": "COCO_train2014_000000207083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a nerdy female with a tube she is filling with posters and dirty feet\"."}], "task": "refering", "answer_template": "The \"a nerdy female with a tube she is filling with posters and dirty feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 98, 99, 100, 103, 104, 120, 121, 122, 123, 124, 126, 127, 143, 144, 145, 146, 147, 149, 166, 167, 168, 169, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 266, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 359, 360, 361, 362], "bbox": [0.2225625, 0.17335416666666664, 0.7887500000000001, 0.9176041666666667], "iscrowd": 0, "area": 42005.99999999999, "rle": {"size": [480, 640], "counts": "SiR23^>i0A=C;F:G7J6I5K101N1O2N1O2M2O2N1O7J7H8H8I7G6F9G9M4M2N3M2N3M2N3_O`0M4M2M4M2N1O1O2N1lGcId7^6XHeIh7\\6UHgIk7Y6RHiIn7e6O00001O001O01O01O0010O0003N2N3M2N3M2O2M2N3N1N:F0001N100O2O0M3L5K4L4L5aMYH`Mk7]2XHTM0[Nl7^4YHUMR8h2QHXMo7d2TH]Mk7b2WH]Mj7`2YH`Mg7]2[HcMe7\\2^HcMb7Z2dHbM]7\\2dHeM[7Y2gHhMY7V2iHiMW7V2jHkMW7R2jHnMY7n1hHSN[7h1fHXN]7d1eH\\N^7_1cHaN`7QOWG`1Y1A]8:cGH`83`GNb8O^G3a8L_G5a8I`G9`8E`G<`8C`G?_8@aGa0_8^OaGd0_8ZOaGh0^8WObGj0^8UObGm0]8ROcGo0^8oNbGS1]8lNcGU1]8jNbGY1^8eNbG\\1^8cNbG_1^8_NbGc1]8\\NcGe1^8YNbGi1]8UNdGl1]8RNeGn1Z8QNhGn1Y8PNjGo1U8PNnGo1R8oMQHo1P8oMTHo1l7oMWHo1j7oMYHP2h7mM\\HP2e7nM^HQ2b7mM_HS2c7jM_HV2a7hMaHX2_7fMZGNN\\2j8cMXG3NZ2Q:gMPFX2n9kMRFU2l9mMSFS2l9oMTFQ2j9QNVFn1i9SNWFn1g9TNYFl1P1dM^6b0bHAoNo1P2PN]6b0lH_1d0QN_6b0lH]1d0RN_6b0mH]1b0SN_6b0nH[1b0TN_6c0nHZ1a0UN_6c0oHX1a0VN_6c0PIX1?WN_6c0RIU1>YN_6d0RIT1=YN`6e0SIQ1<\\N_6e0TIP1;\\Nc6b0SIQ19_Nc6a0SIQ18_Nf6`0RIP16bNi6=RIQ13cNl6;QIR12eNm68RIS1OfNo68QIR1OZNiNLX8g0QIR1L\\NnNIV8h0QIS1I]NQOGV8h0PIT1H]NUODU8k0mHU1G]NXOAU8m0mHT1D_N]O]OS8P1lHU1B_Nm7>bHR1@`N\\6\\OYJT1JQ1_O`NU6\\OWI1Y1S1LQ1]O_NX6]OUI2V1S11o0YOaN[6\\OTI2U1S13o0WO`N_6\\ORI3Q1S19o0ROXNj6DiH3i0Z1b0h0POXNc6@WI3K37LJf1\\1`0lNYNf6ASI2N25k1X18iNZNi6_OTI1M42l1\\17bN]No6[OUI1L5Ok1`1S1`5mLTI1L[2a1g0_5lLTI3LY2a1h0_5mLTI2KX2d1i0\\5mLUI3KV2d1j0\\5mLUI^2d1e0W5mLUIZ2g1i0T5oLSIU2m1l0P5PMRIo1S2P1l4QMPIl1X2S1h4SO[Km0e4oN`KQ1`4_NSIaNa2o2]4[NYI`N^2U3Y4fM\\IVOl2S3i3bMmL^2c2PMgI?i3`2a2RMeI<l3b2_2SMcI9Q4c2]2UM_I7W4d2[2UMZI7]4d2Z2cMgM^2Y2_MiMa2X2[MlMe2U2WMmMi2T2SMoMn2Q2nLSNQ3n61O0010O01O0010O100O100O100O2N100O1O100O1O100O2N100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O00100O1O010O1O100O00100O1O10O01O100O001O100O1O001O10O01O00100O001O010001N100O10000O10O001N2O001N101O0O1001O000000000000000000000000001O01O00O100O100O100O1O100O2O0O100O1O10000O1000000O10O1N2O0O2O2M3M2N3L4J6C=N2N3M;E6K5Jaln1"}, "label": "a nerdy female with a tube she is filling with posters and dirty feet"}]}
{"id": 207083, "image": "COCO_train2014_000000207083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl sitting on the ground\"."}], "task": "refering", "answer_template": "The \"a girl sitting on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 98, 99, 100, 103, 104, 120, 121, 122, 123, 124, 126, 127, 143, 144, 145, 146, 147, 149, 166, 167, 168, 169, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 266, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 359, 360, 361, 362], "bbox": [0.2225625, 0.17335416666666664, 0.7887500000000001, 0.9176041666666667], "iscrowd": 0, "area": 42005.99999999999, "rle": {"size": [480, 640], "counts": "SiR23^>i0A=C;F:G7J6I5K101N1O2N1O2M2O2N1O7J7H8H8I7G6F9G9M4M2N3M2N3M2N3_O`0M4M2M4M2N1O1O2N1lGcId7^6XHeIh7\\6UHgIk7Y6RHiIn7e6O00001O001O01O01O0010O0003N2N3M2N3M2O2M2N3N1N:F0001N100O2O0M3L5K4L4L5aMYH`Mk7]2XHTM0[Nl7^4YHUMR8h2QHXMo7d2TH]Mk7b2WH]Mj7`2YH`Mg7]2[HcMe7\\2^HcMb7Z2dHbM]7\\2dHeM[7Y2gHhMY7V2iHiMW7V2jHkMW7R2jHnMY7n1hHSN[7h1fHXN]7d1eH\\N^7_1cHaN`7QOWG`1Y1A]8:cGH`83`GNb8O^G3a8L_G5a8I`G9`8E`G<`8C`G?_8@aGa0_8^OaGd0_8ZOaGh0^8WObGj0^8UObGm0]8ROcGo0^8oNbGS1]8lNcGU1]8jNbGY1^8eNbG\\1^8cNbG_1^8_NbGc1]8\\NcGe1^8YNbGi1]8UNdGl1]8RNeGn1Z8QNhGn1Y8PNjGo1U8PNnGo1R8oMQHo1P8oMTHo1l7oMWHo1j7oMYHP2h7mM\\HP2e7nM^HQ2b7mM_HS2c7jM_HV2a7hMaHX2_7fMZGNN\\2j8cMXG3NZ2Q:gMPFX2n9kMRFU2l9mMSFS2l9oMTFQ2j9QNVFn1i9SNWFn1g9TNYFl1P1dM^6b0bHAoNo1P2PN]6b0lH_1d0QN_6b0lH]1d0RN_6b0mH]1b0SN_6b0nH[1b0TN_6c0nHZ1a0UN_6c0oHX1a0VN_6c0PIX1?WN_6c0RIU1>YN_6d0RIT1=YN`6e0SIQ1<\\N_6e0TIP1;\\Nc6b0SIQ19_Nc6a0SIQ18_Nf6`0RIP16bNi6=RIQ13cNl6;QIR12eNm68RIS1OfNo68QIR1OZNiNLX8g0QIR1L\\NnNIV8h0QIS1I]NQOGV8h0PIT1H]NUODU8k0mHU1G]NXOAU8m0mHT1D_N]O]OS8P1lHU1B_Nm7>bHR1@`N\\6\\OYJT1JQ1_O`NU6\\OWI1Y1S1LQ1]O_NX6]OUI2V1S11o0YOaN[6\\OTI2U1S13o0WO`N_6\\ORI3Q1S19o0ROXNj6DiH3i0Z1b0h0POXNc6@WI3K37LJf1\\1`0lNYNf6ASI2N25k1X18iNZNi6_OTI1M42l1\\17bN]No6[OUI1L5Ok1`1S1`5mLTI1L[2a1g0_5lLTI3LY2a1h0_5mLTI2KX2d1i0\\5mLUI3KV2d1j0\\5mLUI^2d1e0W5mLUIZ2g1i0T5oLSIU2m1l0P5PMRIo1S2P1l4QMPIl1X2S1h4SO[Km0e4oN`KQ1`4_NSIaNa2o2]4[NYI`N^2U3Y4fM\\IVOl2S3i3bMmL^2c2PMgI?i3`2a2RMeI<l3b2_2SMcI9Q4c2]2UM_I7W4d2[2UMZI7]4d2Z2cMgM^2Y2_MiMa2X2[MlMe2U2WMmMi2T2SMoMn2Q2nLSNQ3n61O0010O01O0010O100O100O100O2N100O1O100O1O100O2N100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O00100O1O010O1O100O00100O1O10O01O100O001O100O1O001O10O01O00100O001O010001N100O10000O10O001N2O001N101O0O1001O000000000000000000000000001O01O00O100O100O100O1O100O2O0O100O1O10000O1000000O10O1N2O0O2O2M3M2N3L4J6C=N2N3M;E6K5Jaln1"}, "label": "a girl sitting on the ground"}]}
{"id": 329963, "image": "COCO_train2014_000000329963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girrafe furthest to the left\"."}], "task": "refering", "answer_template": "The \"girrafe furthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 51, 52, 53, 73, 74, 95, 96, 97, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 186, 187, 188, 210, 230, 231, 233, 253, 256, 276, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.0134375, 0.07023529411764706, 0.31487499999999996, 0.9893176470588236], "iscrowd": 0, "area": 16139.056049999996, "rle": {"size": [425, 640], "counts": "RQ4a1g;6K4L4K5L4K5L5K4K5L4L4K5L4KgMYFV1c9gNhFT1U9iNVGS1e8kNfGn0l4YNZMe0UNj0a4bNZM`0`Nf0U4kN\\M;jNb0i3UO]M5POc0c3YO]M0VOd0\\3^O^MK[Od0W3B;;EG=6BKa02_OOb0O]O3e0JZO8h0EXO<i0BVO?m0^OSOc0h62N3N1N2O1N3M2O1N2O2M2N2O1N3N1N2N2hG]Nh4e1WKaNa4a1_KeNZ4\\1fKkNQ4X1nKnNj3T1WLRO`3P1`LXOV3j0jL]OaMcN]4R2QNCUM^Nj4Q2QNHiL[NU5o1RN3m1OSN0l13TNKk17UNHk19UNFj1<VNCi1`0VN_Oi1c0WN\\Oh1f0XNYOg1i0ZNUOf1m0YNROf1P1ZNoNe1S1[NlNd1V1\\NhNd1[1[NdNe1_1YN`Nf1c1ZN[Ne1i1YNVNf1n1XNQNg1R2XNmMg1W2WNhMh1[2WNdMh1`2VN_Mi1e2VNYMi1j2^NmLb1V3eN`LZ1d3mNbKb1a4a3]1cN3M4L4L3M4L3N3KfMSJQMW6m2R201N101N101N101O0O101N1`NmFlNT9S1UGcNl8]1^GYNc8g1eGoM\\8Q2nGeMS8Z2]10010O0010O01O0O2N1O2N100O2N1O2O0O2N1O2O0O2N1O1O2O0O1O1O100O1O1O10I63N2O1N2O2N1N2O1O1NO2O0O2O11N5L4K3N10O2O0O2O001N101N1017H01O001O1OTOPD?P<@TD>k;@YD>g;A\\D=d;A`D=`;BdD;\\;DgD:Y;DkD:T;FPE7S<N2N3MnSf5"}, "label": "girrafe furthest to the left"}]}
{"id": 329963, "image": "COCO_train2014_000000329963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first giraffe on the left in front of the fence\"."}], "task": "refering", "answer_template": "The \"first giraffe on the left in front of the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 51, 52, 53, 73, 74, 95, 96, 97, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 186, 187, 188, 210, 230, 231, 233, 253, 256, 276, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.0134375, 0.07023529411764706, 0.31487499999999996, 0.9893176470588236], "iscrowd": 0, "area": 16139.056049999996, "rle": {"size": [425, 640], "counts": "RQ4a1g;6K4L4K5L4K5L5K4K5L4L4K5L4KgMYFV1c9gNhFT1U9iNVGS1e8kNfGn0l4YNZMe0UNj0a4bNZM`0`Nf0U4kN\\M;jNb0i3UO]M5POc0c3YO]M0VOd0\\3^O^MK[Od0W3B;;EG=6BKa02_OOb0O]O3e0JZO8h0EXO<i0BVO?m0^OSOc0h62N3N1N2O1N3M2O1N2O2M2N2O1N3N1N2N2hG]Nh4e1WKaNa4a1_KeNZ4\\1fKkNQ4X1nKnNj3T1WLRO`3P1`LXOV3j0jL]OaMcN]4R2QNCUM^Nj4Q2QNHiL[NU5o1RN3m1OSN0l13TNKk17UNHk19UNFj1<VNCi1`0VN_Oi1c0WN\\Oh1f0XNYOg1i0ZNUOf1m0YNROf1P1ZNoNe1S1[NlNd1V1\\NhNd1[1[NdNe1_1YN`Nf1c1ZN[Ne1i1YNVNf1n1XNQNg1R2XNmMg1W2WNhMh1[2WNdMh1`2VN_Mi1e2VNYMi1j2^NmLb1V3eN`LZ1d3mNbKb1a4a3]1cN3M4L4L3M4L3N3KfMSJQMW6m2R201N101N101N101O0O101N1`NmFlNT9S1UGcNl8]1^GYNc8g1eGoM\\8Q2nGeMS8Z2]10010O0010O01O0O2N1O2N100O2N1O2O0O2N1O2O0O2N1O1O2O0O1O1O100O1O1O10I63N2O1N2O2N1N2O1O1NO2O0O2O11N5L4K3N10O2O0O2O001N101N1017H01O001O1OTOPD?P<@TD>k;@YD>g;A\\D=d;A`D=`;BdD;\\;DgD:Y;DkD:T;FPE7S<N2N3MnSf5"}, "label": "first giraffe on the left in front of the fence"}]}
{"id": 329963, "image": "COCO_train2014_000000329963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing in the distance and to the right of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe standing in the distance and to the right of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 78, 79, 80, 100, 101, 123, 124, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 330], "bbox": [0.204890625, 0.14769411764705884, 0.501375, 1.0], "iscrowd": 0, "area": 21939.847699999995, "rle": {"size": [425, 640], "counts": "cdf12T=6J7I6mC@T;g0YDHd;P1M4M2NO2M2N3M2O24K4_EZNZ9i1bF]NX9h1dF]NW9f1eF_Ni0Ic6m1`H=W7FcH>[7F]Ha0a7g2L3M1O1O1O1O1O100O100O100O100O1O1000RM`ISO_6l0iInNW6P1VJeNj5X1]JdNc5Z1cJbN]5\\1iJ`NW5]1QK^No4`1WK\\Ni4a1^K[Nb4b1eKZNY4f1mKVNS4g1TLUNk3i1\\LSNc3l1cLPN\\3o1jLmMU3T2lLkMT3U2mLjMR3X2nLgMQ3Z2oLfMP3[2QMdMn2]2SMbMm2^2TMaMk2a2TM_Mk2b2VM]Mi2d2WM\\Mi2d2WM\\Mh2e2XM[Mg2f2m32N100O1O1O1O1O2N1O1O1O1`GQMK4Q5l2_J`N[OROQ6Z6L2M1O1O1O1O1O1O1O10O01O100O1O1O100O1O100N2N2N2N2NIjJ_HT5k5QKVK_5]4[KnJm4k4g1N20201N2O1O1BSHbKn7`4;2N2N2N2M3N^NoGQNm7S2\\HcMa7b2gHTMX7n2QIhLo6Y3_1N2O2N2N1N3N2NCeLQGX3R9kLjFT3Y9oLcFn2a976I7H7J6F;JNiM^ER2f:QNXEi1n:YNPEa1U;55M3M3M3M3M300O11O01O2N1O2N101N1O00001O010O002N2N3M2N2O0O1O1O1O2N100O1O1O1O1O1N2N2N2N3LaYT4"}, "label": "a giraffe standing in the distance and to the right of another giraffe"}]}
{"id": 329963, "image": "COCO_train2014_000000329963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shorter giraffe in the background\"."}], "task": "refering", "answer_template": "The \"shorter giraffe in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 78, 79, 80, 100, 101, 123, 124, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 330], "bbox": [0.204890625, 0.14769411764705884, 0.501375, 1.0], "iscrowd": 0, "area": 21939.847699999995, "rle": {"size": [425, 640], "counts": "cdf12T=6J7I6mC@T;g0YDHd;P1M4M2NO2M2N3M2O24K4_EZNZ9i1bF]NX9h1dF]NW9f1eF_Ni0Ic6m1`H=W7FcH>[7F]Ha0a7g2L3M1O1O1O1O1O100O100O100O100O1O1000RM`ISO_6l0iInNW6P1VJeNj5X1]JdNc5Z1cJbN]5\\1iJ`NW5]1QK^No4`1WK\\Ni4a1^K[Nb4b1eKZNY4f1mKVNS4g1TLUNk3i1\\LSNc3l1cLPN\\3o1jLmMU3T2lLkMT3U2mLjMR3X2nLgMQ3Z2oLfMP3[2QMdMn2]2SMbMm2^2TMaMk2a2TM_Mk2b2VM]Mi2d2WM\\Mi2d2WM\\Mh2e2XM[Mg2f2m32N100O1O1O1O1O2N1O1O1O1`GQMK4Q5l2_J`N[OROQ6Z6L2M1O1O1O1O1O1O1O10O01O100O1O1O100O1O100N2N2N2N2NIjJ_HT5k5QKVK_5]4[KnJm4k4g1N20201N2O1O1BSHbKn7`4;2N2N2N2M3N^NoGQNm7S2\\HcMa7b2gHTMX7n2QIhLo6Y3_1N2O2N2N1N3N2NCeLQGX3R9kLjFT3Y9oLcFn2a976I7H7J6F;JNiM^ER2f:QNXEi1n:YNPEa1U;55M3M3M3M3M300O11O01O2N1O2N101N1O00001O010O002N2N3M2N2O0O1O1O1O2N100O1O1O1O1O1N2N2N2N3LaYT4"}, "label": "shorter giraffe in the background"}]}
{"id": 403705, "image": "COCO_train2014_000000403705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"short giraffe\"."}], "task": "refering", "answer_template": "The \"short giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 128, 129, 130, 131, 132, 133, 153, 154, 155, 156, 177, 178, 179, 180, 200, 201, 202, 223, 224, 225, 246, 247], "bbox": [0.5853125, 0.2594166666666667, 0.8385781250000001, 0.6448125], "iscrowd": 0, "area": 11337.314900000007, "rle": {"size": [480, 640], "counts": "Wm_5o0P>2O0O2O0O2O0O2O0O2O0O20N4L5L0O1O1O1C=00000O2O0000000000000000001N10000000000000001O0O10000000000000001O000O1000000O10001O0O1000000O10000WB_OS=[1K5L5K4L4K4M4L3M5J7J6J6J6I7J5WF\\Ll7j3kGZLT8l3dGVL[8Q4\\GRLd8T4SGoKm8m1nFALeNV9k1TG4m8KZGNg82_GGb88eG@]8`0iGPOb8n0eGdNc8`1_GSNh8R2YGaMn8d2e16K4K5K6J7J:E9G5K6J4K100O00100O10BUFeKk9X4`0M4L3L4M4L3L5L3M3L5L3SOaDVNc;e1fDSN];i1lDnMX;o1o0K4L4L57H<E;D=C7J1N3M2O20O000K6J5J7I6K5J7I6K5J7J5J6J7J5J7I6K5Jha`1"}, "label": "short giraffe"}]}
{"id": 403705, "image": "COCO_train2014_000000403705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shorter giraffe\"."}], "task": "refering", "answer_template": "The \"shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 128, 129, 130, 131, 132, 133, 153, 154, 155, 156, 177, 178, 179, 180, 200, 201, 202, 223, 224, 225, 246, 247], "bbox": [0.5853125, 0.2594166666666667, 0.8385781250000001, 0.6448125], "iscrowd": 0, "area": 11337.314900000007, "rle": {"size": [480, 640], "counts": "Wm_5o0P>2O0O2O0O2O0O2O0O2O0O20N4L5L0O1O1O1C=00000O2O0000000000000000001N10000000000000001O0O10000000000000001O000O1000000O10001O0O1000000O10000WB_OS=[1K5L5K4L4K4M4L3M5J7J6J6J6I7J5WF\\Ll7j3kGZLT8l3dGVL[8Q4\\GRLd8T4SGoKm8m1nFALeNV9k1TG4m8KZGNg82_GGb88eG@]8`0iGPOb8n0eGdNc8`1_GSNh8R2YGaMn8d2e16K4K5K6J7J:E9G5K6J4K100O00100O10BUFeKk9X4`0M4L3L4M4L3L5L3M3L5L3SOaDVNc;e1fDSN];i1lDnMX;o1o0K4L4L57H<E;D=C7J1N3M2O20O000K6J5J7I6K5J7I6K5J7J5J6J7J5J7I6K5Jha`1"}, "label": "shorter giraffe"}]}
{"id": 403705, "image": "COCO_train2014_000000403705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe out of the two giraffes is eating the leaves of a small tree\"."}], "task": "refering", "answer_template": "The \"a tall giraffe out of the two giraffes is eating the leaves of a small tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 121, 122, 123, 124, 128, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239], "bbox": [0.24493749999999997, 0.20833333333333334, 0.5844531249999999, 0.6497708333333333], "iscrowd": 0, "area": 14592.580099999997, "rle": {"size": [480, 640], "counts": "chY2T2j<9G1OiMdCl1\\<XNlC]1S<gNUDn0i;VO`D>`;FhDOW;5QE_O5HV9m0TH@h7d0kGIQ8;aG3[8?jFNQ9g0QFGl9j2L3M]OaFhK]9]3`GaLa8l2SHSMm7m2UHPMk7Q3R2000O2O002M4M3M3L3N3M3L4M3M2M4M3MPObEaMZ:Y2PFgMm9R2]FmM`9m1kFRNR9g1XGYNe8b1cG^NY8`1mG`NP8]1VHbNh7Z1_HfN]7X1iHhNT7U1RIjNl6V1XIeNg6]1\\I_Nd6b1_IZNa6g1bIUN]6l1gIPNY6Q2Z31N2O1O0O2O1O1QFgMV7Z2eHoMU7R2hHVNS7k1hH^NS7c1hHfNR7\\1iHlNS7U1hHROU7o0fHXOV7i0fH`NfNCb8m1dHaNTO[OX8T2_HcNDPOn7]2YHeNm8[1nFgNU9X1gFjN\\9V1_FlNc9U1XFmNk9S1QFnNR:Y1bEiNa:h23M3M3M3M3M3M3M3L^FhLd7U3_GmM_8]4M2N3M2N3M2NYNaGPM\\8c2VHYMi7[2hH_MX7U2ZIfMf6n1kImMU6i1ZJRNf5k1aJPN`5l1iJoMW5n1QKmMo4P2YKkMg4R2U4mNZCIi<7_C^Od<a0dCUO_<k0k00000O100000000O1000O10O1000000O100000000O1000000O100000000O1000O10O1000000O100000000O1000000O100000000O1000O10O1000000O100000000O1000000O100000000O10O1000O10000001O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2M100000O1M3L4M2M3N2XOTB;P>CXB2gPm3"}, "label": "a tall giraffe out of the two giraffes is eating the leaves of a small tree"}]}
{"id": 403705, "image": "COCO_train2014_000000403705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left side of the image with it ' s head highest of the other giraffe\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left side of the image with it ' s head highest of the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 121, 122, 123, 124, 128, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239], "bbox": [0.24493749999999997, 0.20833333333333334, 0.5844531249999999, 0.6497708333333333], "iscrowd": 0, "area": 14592.580099999997, "rle": {"size": [480, 640], "counts": "chY2T2j<9G1OiMdCl1\\<XNlC]1S<gNUDn0i;VO`D>`;FhDOW;5QE_O5HV9m0TH@h7d0kGIQ8;aG3[8?jFNQ9g0QFGl9j2L3M]OaFhK]9]3`GaLa8l2SHSMm7m2UHPMk7Q3R2000O2O002M4M3M3L3N3M3L4M3M2M4M3MPObEaMZ:Y2PFgMm9R2]FmM`9m1kFRNR9g1XGYNe8b1cG^NY8`1mG`NP8]1VHbNh7Z1_HfN]7X1iHhNT7U1RIjNl6V1XIeNg6]1\\I_Nd6b1_IZNa6g1bIUN]6l1gIPNY6Q2Z31N2O1O0O2O1O1QFgMV7Z2eHoMU7R2hHVNS7k1hH^NS7c1hHfNR7\\1iHlNS7U1hHROU7o0fHXOV7i0fH`NfNCb8m1dHaNTO[OX8T2_HcNDPOn7]2YHeNm8[1nFgNU9X1gFjN\\9V1_FlNc9U1XFmNk9S1QFnNR:Y1bEiNa:h23M3M3M3M3M3M3M3L^FhLd7U3_GmM_8]4M2N3M2N3M2NYNaGPM\\8c2VHYMi7[2hH_MX7U2ZIfMf6n1kImMU6i1ZJRNf5k1aJPN`5l1iJoMW5n1QKmMo4P2YKkMg4R2U4mNZCIi<7_C^Od<a0dCUO_<k0k00000O100000000O1000O10O1000000O100000000O1000000O100000000O1000O10O1000000O100000000O1000000O100000000O1000O10O1000000O100000000O1000000O100000000O10O1000O10000001O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2M100000O1M3L4M2M3N2XOTB;P>CXB2gPm3"}, "label": "the giraffe on the left side of the image with it ' s head highest of the other giraffe"}]}
{"id": 215289, "image": "COCO_train2014_000000215289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby goat rubbing his head on the neck of another goat\"."}], "task": "refering", "answer_template": "The \"a baby goat rubbing his head on the neck of another goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 103, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 303, 304, 306, 307, 308, 323, 324, 326, 327, 329, 330, 331, 346, 349, 350, 351, 353, 354, 355, 373, 374, 377], "bbox": [0.04978125, 0.2700208333333334, 0.6038593750000001, 0.9809375], "iscrowd": 0, "area": 56276.30189999999, "rle": {"size": [480, 640], "counts": "Q\\?=b>`0_Ob0_O9WLcN]Ia1_6POPIS1m6ZOgHh0V7E]H>`7OSH4i7:kGHR8e0aG^O\\8l0[GVOb8l0\\GWOa8l0\\GWOa8l0]GVO`8l0^GXO]8h3M3M3MLkGbIR8Z6VHeIg7V6bHiIZ7S6oHlIn6P6ZIoId6k5eITJY6j5_1O2M2O1N2N2O1N2O1N2O1N3N1N2O1N2O1N2O1N3N2M2O2M3N200O10O0100O100O00100O100O10O0100O100O10O100000001O1N2O2N1O1O1O1N2O1O2N1cGWKa6j4ZI[Ke6f4VI`Kg6b4TIcKk6^4PIgKo6Z4kHmKS7U4gHPLX7Q4cHUL[7l3ZH_Le7`5N1O2N1O1O2N1N3N1O1O1O001O1O1O1O1O1O1O1O1O1O00dNoHjJP7Q5VIoJi6l4^ISKa6i4eIVKZ6g4jIYKU6e4oIZKP6c4TJ]Kk5`4ZJ_Ke5_4^JaKa5\\4dJcK[5[4hJeKW5X4nJfKR5W4SKhKl4V4WKjKh4S4]KlKb4R4aKnK^4R4dKmK[4R4gKnKX4R4jKmKU4R4mKnKR4R4PLmKo3S4SLkKm3T4ULkKk3U4WLiKi3W4XLhKh3W4[LgKe3Y4\\LfKd3Z4^LdKb3[4`LdK`3\\4bLbK^3^4dL`K\\3_4fL`KZ3`4hL^KX3a4jL^KV3b4lL\\KT3d4mL[KS3d4PMZKP3f4QMYKo2g4QMYKo2f4SMYKm2g4SMYKm2g4SMYKm2f4UMYKk2g4UMYKk2g4VMXKj2g4WMYKi2g4WMYKi2f4i30000O10000O10000O10000O100O11O2N1O1N3N1O1oGXKh5j4QJ\\Kn5e4lIaKS6g6O001O1O1O1O0O2O00M3N1N3N2N2M2O2M3K5J6K4L5J6K5K4L5J6>B`0_Oa0@8H3M4K5L3L5K5J5L4L1N3N2M3NhMeInKY6m3TJmKi5o3eJjKQ4YNVLh56iKa3cNVL_5g0gKP3nNULW5W1eKc2WOSLo4a1iK\\2ZOoKi4m1kKU2\\OlKe4U2nKP2^OhKc4\\2lKm1CcKa4c2kKk1E_K_4j2jKg1I\\K]4o2iKf1JXK]4V3fKc1OTKZ4\\3fKa1Q5cNmJ^1P5eNoJ[1P5hNmJZ1S5fNlJZ1V5fNhJ[1Y5dNfJ]1[5cNbJ^1`5aN_J`1b5`N\\J`1f5_NYJb1g5_NVJc1k5h3M4I6J7I7I6J5M4N2N1O2N2O0O2N2N1O2N2N1O2O1N1O2N2N1O2N2O0O2N1O2N2N2N2O1N2N2N2N2O1N2N2N2N3N2M3M3M2N3N2M3M3N2O1O10O01O1O1O100O1O1O010O1O1O100O1O1O010O1O2N1O101N1O1O2O0O1N3M2N2N3M2N2N3M2N2N3M2N2N3M2Nokf3"}, "label": "a baby goat rubbing his head on the neck of another goat"}]}
{"id": 215289, "image": "COCO_train2014_000000215289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shaggy deer with very small horns standing on grass in a pen\"."}], "task": "refering", "answer_template": "The \"a shaggy deer with very small horns standing on grass in a pen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 103, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 303, 304, 306, 307, 308, 323, 324, 326, 327, 329, 330, 331, 346, 349, 350, 351, 353, 354, 355, 373, 374, 377], "bbox": [0.04978125, 0.2700208333333334, 0.6038593750000001, 0.9809375], "iscrowd": 0, "area": 56276.30189999999, "rle": {"size": [480, 640], "counts": "Q\\?=b>`0_Ob0_O9WLcN]Ia1_6POPIS1m6ZOgHh0V7E]H>`7OSH4i7:kGHR8e0aG^O\\8l0[GVOb8l0\\GWOa8l0\\GWOa8l0]GVO`8l0^GXO]8h3M3M3MLkGbIR8Z6VHeIg7V6bHiIZ7S6oHlIn6P6ZIoId6k5eITJY6j5_1O2M2O1N2N2O1N2O1N2O1N3N1N2O1N2O1N2O1N3N2M2O2M3N200O10O0100O100O00100O100O10O0100O100O10O100000001O1N2O2N1O1O1O1N2O1O2N1cGWKa6j4ZI[Ke6f4VI`Kg6b4TIcKk6^4PIgKo6Z4kHmKS7U4gHPLX7Q4cHUL[7l3ZH_Le7`5N1O2N1O1O2N1N3N1O1O1O001O1O1O1O1O1O1O1O1O1O00dNoHjJP7Q5VIoJi6l4^ISKa6i4eIVKZ6g4jIYKU6e4oIZKP6c4TJ]Kk5`4ZJ_Ke5_4^JaKa5\\4dJcK[5[4hJeKW5X4nJfKR5W4SKhKl4V4WKjKh4S4]KlKb4R4aKnK^4R4dKmK[4R4gKnKX4R4jKmKU4R4mKnKR4R4PLmKo3S4SLkKm3T4ULkKk3U4WLiKi3W4XLhKh3W4[LgKe3Y4\\LfKd3Z4^LdKb3[4`LdK`3\\4bLbK^3^4dL`K\\3_4fL`KZ3`4hL^KX3a4jL^KV3b4lL\\KT3d4mL[KS3d4PMZKP3f4QMYKo2g4QMYKo2f4SMYKm2g4SMYKm2g4SMYKm2f4UMYKk2g4UMYKk2g4VMXKj2g4WMYKi2g4WMYKi2f4i30000O10000O10000O10000O100O11O2N1O1N3N1O1oGXKh5j4QJ\\Kn5e4lIaKS6g6O001O1O1O1O0O2O00M3N1N3N2N2M2O2M3K5J6K4L5J6K5K4L5J6>B`0_Oa0@8H3M4K5L3L5K5J5L4L1N3N2M3NhMeInKY6m3TJmKi5o3eJjKQ4YNVLh56iKa3cNVL_5g0gKP3nNULW5W1eKc2WOSLo4a1iK\\2ZOoKi4m1kKU2\\OlKe4U2nKP2^OhKc4\\2lKm1CcKa4c2kKk1E_K_4j2jKg1I\\K]4o2iKf1JXK]4V3fKc1OTKZ4\\3fKa1Q5cNmJ^1P5eNoJ[1P5hNmJZ1S5fNlJZ1V5fNhJ[1Y5dNfJ]1[5cNbJ^1`5aN_J`1b5`N\\J`1f5_NYJb1g5_NVJc1k5h3M4I6J7I7I6J5M4N2N1O2N2O0O2N2N1O2N2N1O2O1N1O2N2N1O2N2O0O2N1O2N2N2N2O1N2N2N2N2O1N2N2N2N3N2M3M3M2N3N2M3M3N2O1O10O01O1O1O100O1O1O010O1O1O100O1O1O010O1O2N1O101N1O1O2O0O1N3M2N2N3M2N2N3M2N2N3M2N2N3M2Nokf3"}, "label": "a shaggy deer with very small horns standing on grass in a pen"}]}
{"id": 215289, "image": "COCO_train2014_000000215289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep to the right of another sheep\"."}], "task": "refering", "answer_template": "The \"a sheep to the right of another sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 334, 335, 336], "bbox": [0.440328125, 0.18243749999999997, 0.7258125, 0.8706041666666666], "iscrowd": 0, "area": 19201.92245000002, "rle": {"size": [480, 640], "counts": "SZT44k>2N2M3N1O2M3M3N1N3N101N2N1O1O0010O100O100O1O100O100O10000000000O10iEDj5;QJLn54PJ0m51RJ2l5MSJ7k5ISJ;k5DTJ`0j5@TJd0i5\\OUJi0i5WOTJn0j5ROSJU1i5lNRJ\\1TMRNW8b0cJP2Z5PNdJU2Z5kMcJZ2^NgMh4NhL`2XNhMo4IfLd2SNhMX5CcLj2nMgM_5@bLl2hMiMf5[O`LQ3cMgMm5XO`Li4_3XK`Lh4a3XK^Lh4n2cJPKe0Q2i4o2dJnJd0R2h4P3gJkJb0T2g4P3jJjJ`0U2f4Q3kJiJ?V2e4Q3oJgJ=W2d4R3PKfJ=V2m0fM67\\OW5VOdJ<W2k0kM2?WOj4AeJ<V2j0oMNf0QOa4KdJ=U2i0WNEj0POW45dJ=S2j0_NWOQ1SOi3`0fJ;P2k0n6jNSG;n1l0n6iNVG:k1n0n6hNYG9h1P1n6gN\\G7f1S1n6eN^G7c1U1n6dN`G8`1V1n6bNdG7]1\\1k6\\NiG8[1`1h6WNoG9X1d1e6SNSH:W1g1c6nMWH;U1k1`6kMZH;U1n1nNXMd6?YI<T1Q2fNZMP76VI?S1U2_N[MW71WI`0R1k2f5eLVIb0S1j2f5ZNWJi1h5WNVJk1i5VNUJl1j5UNTJm1l5TNRJm1JZL_4k1dKm1oNWMZ5o0dKb3Z4`LdKb3Y4aLcKc3Z4_LdKc3Z4`LcKc3Z4_LdKd3X4`LeKb3Y4`LeKb3Y4aL[Kj3d4WLYKk3f4WLXKk3g4BfKRJd0k0e3j4SLVJd0h0X3m4\\LTJg0g0m2U5_LRJh0e0i2X5dLRJa0e0k2W5hNhJW1X5kNfJV1Y5lNfJT1X5oNgJQ1X5POhJP1W5ROhJn0W5\\O`Jd0^5A]Ja0V5k4O1O2M2O1O2N1O2N1O1@a0I6M3M4L3M3M3M3H9H7O2ROTDdNo;Y1VDcNm;Z1XDaNk;\\1ZD\\Nk;b1k0M3N2N2N1N3B>I5E<O01O00O2N11O1O10O01O10O01O100O001O100O010000O100O10001N100N2O1O2M2O5JPna2"}, "label": "a sheep to the right of another sheep"}]}
{"id": 379136, "image": "COCO_train2014_000000379136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby in teal is holding her hands up while looking at an elephant\"."}], "task": "refering", "answer_template": "The \"a baby in teal is holding her hands up while looking at an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 148, 162, 163, 164, 165, 166, 167, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 255, 256, 257, 258, 279, 280, 302, 303], "bbox": [0.071484375, 0.23593457943925233, 0.46364062500000003, 0.9250467289719626], "iscrowd": 0, "area": 34242.55330000001, "rle": {"size": [428, 640], "counts": "f]c08R=8H8H8H8H8H8H8H8H8H7I7I8H7I8H7I8H6J4J6H7H9H8G8I8G9H7O2M3M3N1N3N2M2N3N2M2O2M3N1N3M3N2M2O:E>B=D0O101N101N1O2O0O101N101NElK`GS4a8oK^GP4a8SL]Gm3b8WL[Gi3d8ZL[Ge3e8a0O1O100O1O1O100O1O1O100O1O1O100O1O1N2N2O1N2N3M2O1N2N2N2O1N2N2N2O1N2N2N2N2N2N3M3N1N3M3M3M3M3M3M3M3M2O2M3M3M3M3M3M3M3M2O2O1O1O1O1O1O1O1O001O1O1cMcGH_83fGL[8OjG0W8KnG4S8FSH8o7CVH<j7@\\H>h7ZO]Hd0i7SO\\Hl0i7jN]HU1h7bN]H\\1j7ZN[He1g9O2N1N2O1O1N2O1O2M0100O10O0100O010O10O0100O10O10O10O0100O010O100O010O1O0O2O1O001N2O1O001O10O10000O001O1O1O0O2O1O1O1N101O1O0O101O001N10001O0O2O00001N101O001O000010O01O01O02N100O2N100O2N100O2N100O2N100O2N4M6I7JmT_4"}, "label": "a baby in teal is holding her hands up while looking at an elephant"}]}
{"id": 379136, "image": "COCO_train2014_000000379136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby wearing a turquoise top waving at an elephant\"."}], "task": "refering", "answer_template": "The \"a baby wearing a turquoise top waving at an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 148, 162, 163, 164, 165, 166, 167, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 255, 256, 257, 258, 279, 280, 302, 303], "bbox": [0.071484375, 0.23593457943925233, 0.46364062500000003, 0.9250467289719626], "iscrowd": 0, "area": 34242.55330000001, "rle": {"size": [428, 640], "counts": "f]c08R=8H8H8H8H8H8H8H8H8H7I7I8H7I8H7I8H6J4J6H7H9H8G8I8G9H7O2M3M3N1N3N2M2N3N2M2O2M3N1N3M3N2M2O:E>B=D0O101N101N1O2O0O101N101NElK`GS4a8oK^GP4a8SL]Gm3b8WL[Gi3d8ZL[Ge3e8a0O1O100O1O1O100O1O1O100O1O1O100O1O1N2N2O1N2N3M2O1N2N2N2O1N2N2N2O1N2N2N2N2N2N3M3N1N3M3M3M3M3M3M3M3M2O2M3M3M3M3M3M3M3M2O2O1O1O1O1O1O1O1O001O1O1cMcGH_83fGL[8OjG0W8KnG4S8FSH8o7CVH<j7@\\H>h7ZO]Hd0i7SO\\Hl0i7jN]HU1h7bN]H\\1j7ZN[He1g9O2N1N2O1O1N2O1O2M0100O10O0100O010O10O0100O10O10O10O0100O010O100O010O1O0O2O1O001N2O1O001O10O10000O001O1O1O0O2O1O1O1N101O1O0O101O001N10001O0O2O00001N101O001O000010O01O01O02N100O2N100O2N100O2N100O2N100O2N4M6I7JmT_4"}, "label": "a baby wearing a turquoise top waving at an elephant"}]}
{"id": 379136, "image": "COCO_train2014_000000379136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person holding the toddler\"."}], "task": "refering", "answer_template": "The \"the person holding the toddler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 46, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 326], "bbox": [0.0, 0.0025, 0.20842187499999998, 0.9867990654205606], "iscrowd": 0, "area": 22708.544200000004, "rle": {"size": [428, 640], "counts": "1Y3S:c6]IY3gL00000000000000000001O000000000000000000000QD8Z:HmDW1m:i0K6J6J6J3M3M2N3M2N3M3aGdLh6^3UIfLh6]3SIhLj6Z3RIkLk6X3oHmLo6U3^H_M_7Z4M3M3M3M3lH`JV6`6K5K5K5K5K6J5K5K5K6J5K6J5K6J5K6J5K5K2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N8iNPEBX;OoD1Y;@nD`0o;01O0000001O0000001O0000001O0000001O0000001O00001O0000001O00001O1O2M2O1O2N1O1O2N1O2N1O1O2N1OP`c6"}, "label": "the person holding the toddler"}]}
{"id": 379136, "image": "COCO_train2014_000000379136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of the person holding the baby\"."}], "task": "refering", "answer_template": "The \"the arm of the person holding the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 46, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 326], "bbox": [0.0, 0.0025, 0.20842187499999998, 0.9867990654205606], "iscrowd": 0, "area": 22708.544200000004, "rle": {"size": [428, 640], "counts": "1Y3S:c6]IY3gL00000000000000000001O000000000000000000000QD8Z:HmDW1m:i0K6J6J6J3M3M2N3M2N3M3aGdLh6^3UIfLh6]3SIhLj6Z3RIkLk6X3oHmLo6U3^H_M_7Z4M3M3M3M3lH`JV6`6K5K5K5K5K6J5K5K5K6J5K6J5K6J5K6J5K5K2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N8iNPEBX;OoD1Y;@nD`0o;01O0000001O0000001O0000001O0000001O0000001O00001O0000001O00001O1O2M2O1O2N1O1O2N1O2N1O1O2N1OP`c6"}, "label": "the arm of the person holding the baby"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy on a computer that is more tan and has shorter hair than the other\"."}], "task": "refering", "answer_template": "The \"a guy on a computer that is more tan and has shorter hair than the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 103, 104, 105, 126, 127, 128, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 224, 240, 241, 242, 246, 247, 250, 269, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298], "bbox": [0.4131875, 0.14760416666666665, 0.9953593749999999, 0.7701666666666666], "iscrowd": 0, "area": 25654.193300000003, "rle": {"size": [480, 640], "counts": "hnk31i>=C<G:H8YBfNX=d1M3M3M3M3L4M3M3M2N3L4M3M3M3M3N2M3N2M2N3N2M3M3N2M3N2M00100O100O1O100O1O100O00100O1gMeLlH\\3R7iLjHX3T7nLfHT3X7QMcHQ3[7TM`Hn2^7WM]Hk2a7[2O010O1O100O1O100O1O100O00100O1000000000000000O10000000000000001O000O101O000:G0OO2N1M4L4M2M4M2M4L4[L]Gi0e8oNdGm0`8mNeGP1^8nNdGP1_8lNdGQ1_8lNeGP1_8mNcGQ1`8kNcGR1`8lNbGQ1b8kNbGR1`8kNcGR1a8kNbGR1a8jNbGS1c8hNaGT1c8]NlG`1X8_NgG_1]8`NcG_1^8_NcGa1_8\\NcGc1`8ZNaGe1a8XN`Gg1d8UN^Gj1d8TN\\Gl1m:O001O1N101O001O2N2M2O2N2N2N1O2M3N2N3M2N2M4M2N3M2N2N3L3N2N2N2N2NlT>1PkA300000O100000000O2O000000000O10001OjN4XCLf<5[CJd<8\\CHb<:^CF`<<`CD^<>bCB\\<`0P12N1O21N1O2O000100O0hB_OR<`0nCAR<=oCDP<<PDDQ<:PDFQ<9PDGo;8RDHo;7QDIP<5QDKP<4PDMS<NnC2`=0010O0001O00001O00001N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101N10000O2O0O101N100O2O000O2O0O101N100O2O000O2O0O101N100O2O000O100O101O0001O1O0010O01O001O1O010O001O001O10O01O001O001O10O01O001O00100O001O001O010O000010O0001O00010O000010O0001O0001O01O00010O00001O01O01O0001O01O0000aR1"}, "label": "a guy on a computer that is more tan and has shorter hair than the other"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the right , closest to the piano , using one of two apple laptops\"."}], "task": "refering", "answer_template": "The \"the man on the right , closest to the piano , using one of two apple laptops\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 103, 104, 105, 126, 127, 128, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 224, 240, 241, 242, 246, 247, 250, 269, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298], "bbox": [0.4131875, 0.14760416666666665, 0.9953593749999999, 0.7701666666666666], "iscrowd": 0, "area": 25654.193300000003, "rle": {"size": [480, 640], "counts": "hnk31i>=C<G:H8YBfNX=d1M3M3M3M3L4M3M3M2N3L4M3M3M3M3N2M3N2M2N3N2M3M3N2M3N2M00100O100O1O100O1O100O00100O1gMeLlH\\3R7iLjHX3T7nLfHT3X7QMcHQ3[7TM`Hn2^7WM]Hk2a7[2O010O1O100O1O100O1O100O00100O1000000000000000O10000000000000001O000O101O000:G0OO2N1M4L4M2M4M2M4L4[L]Gi0e8oNdGm0`8mNeGP1^8nNdGP1_8lNdGQ1_8lNeGP1_8mNcGQ1`8kNcGR1`8lNbGQ1b8kNbGR1`8kNcGR1a8kNbGR1a8jNbGS1c8hNaGT1c8]NlG`1X8_NgG_1]8`NcG_1^8_NcGa1_8\\NcGc1`8ZNaGe1a8XN`Gg1d8UN^Gj1d8TN\\Gl1m:O001O1N101O001O2N2M2O2N2N2N1O2M3N2N3M2N2M4M2N3M2N2N3L3N2N2N2N2NlT>1PkA300000O100000000O2O000000000O10001OjN4XCLf<5[CJd<8\\CHb<:^CF`<<`CD^<>bCB\\<`0P12N1O21N1O2O000100O0hB_OR<`0nCAR<=oCDP<<PDDQ<:PDFQ<9PDGo;8RDHo;7QDIP<5QDKP<4PDMS<NnC2`=0010O0001O00001O00001N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101N10000O2O0O101N100O2O000O2O0O101N100O2O000O2O0O101N100O2O000O100O101O0001O1O0010O01O001O1O010O001O001O10O01O001O001O10O01O001O00100O001O001O010O000010O0001O00010O000010O0001O0001O01O00010O00001O01O01O0001O01O0000aR1"}, "label": "the man on the right , closest to the piano , using one of two apple laptops"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black hair sitting closest to the remotes\"."}], "task": "refering", "answer_template": "The \"a man with black hair sitting closest to the remotes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 268, 269, 270, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.109375, 0.10647916666666667, 1.0, 0.9865208333333334], "iscrowd": 0, "area": 81733.18435, "rle": {"size": [480, 640], "counts": "oPQ1:e>:E014L3M3L5L1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O2N1O3L6K5K5K3M100O2O0O101N100O101N1UM[MZIf2d6_MYIb2c6dMYI]2e6hMXIX2e6nMWIS2g6RNVIo1g6WNTIj1i6dNlH\\1R7hNkHZ1R7jNlHV1S7lNkHU1T7nNjHR1ROPMc7P2YIR1ROoLe7Q2WIP1SOQMd7R2VIn0UORMc7S2VIl0UOSMc7W2RIg0ZOSMd7Z2mHd0^OTMc7^2jH>BVMb7a2gH;EUMc7e2dH6HWMc7h2_H2MXMb7k2]Ha0b7`3O1N1O2N2O0O2O1O001N2O1O001O1N101O1O0O2O1O001O1N10100O00100O1O010O1O010O1O101N1O1O100O1O2O0O1O100O1O2O0000dJhIo1Y6lMmIR2S6jMSJT2m5hMYJW2g5dM]J\\2c5`M`Ja2`5[MdJe2\\5WMgJj2Z5RMiJn2W5nLlJT3S5hLQKX3P5cLSK^3m4^LWKb3i4ZLZKg3f4UL]Kl3d4oK`KQ4`4SKYJCZ1\\5\\4mJ`LS5`3jJbLW5_3eJdL[5\\3bJfL_5^60OO2N1O2N10eFjJl8U5SGPKk8n4UGVKh8i4XG\\Ke8b4ZGcKc8]4\\GgKb8W4^GnK^8Q4aGTL]8j3cGZLZ8e3fG`LW8_3gGfLV8Y3jGkLT8S3lGQMQ8n2PHSMo7m2QHSMP8k2RHUMm7j2THVMm7h2THXMl7g2UHZMk7e2UH[Mk7d2WH[Mj7d2VH]Mj7b2VH^Mk7a2UH`Ml7^2THbMm7]2THbMn7\\2RHeMn7Z2RHfMP8X2PHiMQ8U2oGkMR8T2oGkMS8S2mGnMS8Q2mGoMU8o1kGQNV8n1jGSNW8k1jGTNX8k1gGVNY8i1gGWN[8g1eGYN\\8f1dG[N]8c1dG\\N]8c1cG^N^8`1bG`N_8_1aGaNa8]1_GdNb8Z1_GeNb8Z1^GfNd8X1\\GiNd8V1\\GjNf8T1ZGmNf8R1[GmNg8Q1YGoNi8o0WGROi8m0WGSOk8l0TGTOm8k0TGUOm8i0SGWOn8h0RGYOo8e0QG[OP9c0QG]OQ9a0PG_OQ9?oFAR9=oFDR9:nFFS98nFHT96mFJT93mFMU91kFOW9OiF2Y9JhF6Z9HgF8[9DfF<\\9BdF>_9_OaFb0`9[OaFe0a9YO`Fg0b9UO_Fk0c9SO]Fm0f9QOYFP1g9oNYFQ1g9oNZFP1f9POZFQ1e9oN[FQ1e9POZFQ1e9oN[FQ1e9oN\\FP1d9PO\\FQ1c9PO\\FP1d9PO\\FQ1b9PO^FP1b9PO_Fo0a9RO^Fo0a9QO_FP1`9PO`FP1`9PO`FQ1_9POaFP1^9PObFQ1]9oNcFR1\\9nNdFR1\\9nNdFS1Z9oNfFQ1Y9oNgFR1X9nNhFS1W9mNiFS1W9nNhFS1W9mNjFS1U9mNkFT1T9lNlFU1S9lNlFT1T9lNlFU1R9lNoFT1P9lNPGU1o8lNPGU1o8kNQGV1n8iNSGW1m8hNUGX1j8gNWGZ1h8eNYG\\1f8cN[G^1c8bN^G^1b8iM]F2S1T2`8jM_FOR1W2_8kM_FLS1Y2^8kM`FJS1[2]8kMaFIR1\\2]8kMcFFR1^2[8mMcFCS1`2Z8mMdFAS1b2X8nMfF^OS1d2W8oMgFZOS1g2V8oMhFXOS1i2U8oMWHR2h7oMVHR2j7nMVHR2j7nMUHS2k7mMTHT2k7nMTHR2l7mMTHT2l7fMZHZ2f7_M`Hb2`7WMgHi2g9100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O101N100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O2O0O1O100O100O1MfK`EZ4`:fK_E[4a:eK_E[4a:200000000000000O100000000000000000000000000000M3K5K5M30000000000000O1000001O00001O0000001O000010O00010O000010O00010O00010O01O00010O00010O00010O000010O00010O00010O0001O010O00001O00001O01O01O00001O00001O00001O00001O001O00001O00001O00001O000000001O000000001O00000000001O000000001O00000000001O000001O01O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000001O01O00000M3ZOg0XOg0ZOf0YOUE"}, "label": "a man with black hair sitting closest to the remotes"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the left\"."}], "task": "refering", "answer_template": "The \"the man on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 268, 269, 270, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.109375, 0.10647916666666667, 1.0, 0.9865208333333334], "iscrowd": 0, "area": 81733.18435, "rle": {"size": [480, 640], "counts": "oPQ1:e>:E014L3M3L5L1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O2N1O3L6K5K5K3M100O2O0O101N100O101N1UM[MZIf2d6_MYIb2c6dMYI]2e6hMXIX2e6nMWIS2g6RNVIo1g6WNTIj1i6dNlH\\1R7hNkHZ1R7jNlHV1S7lNkHU1T7nNjHR1ROPMc7P2YIR1ROoLe7Q2WIP1SOQMd7R2VIn0UORMc7S2VIl0UOSMc7W2RIg0ZOSMd7Z2mHd0^OTMc7^2jH>BVMb7a2gH;EUMc7e2dH6HWMc7h2_H2MXMb7k2]Ha0b7`3O1N1O2N2O0O2O1O001N2O1O001O1N101O1O0O2O1O001O1N10100O00100O1O010O1O010O1O101N1O1O100O1O2O0O1O100O1O2O0000dJhIo1Y6lMmIR2S6jMSJT2m5hMYJW2g5dM]J\\2c5`M`Ja2`5[MdJe2\\5WMgJj2Z5RMiJn2W5nLlJT3S5hLQKX3P5cLSK^3m4^LWKb3i4ZLZKg3f4UL]Kl3d4oK`KQ4`4SKYJCZ1\\5\\4mJ`LS5`3jJbLW5_3eJdL[5\\3bJfL_5^60OO2N1O2N10eFjJl8U5SGPKk8n4UGVKh8i4XG\\Ke8b4ZGcKc8]4\\GgKb8W4^GnK^8Q4aGTL]8j3cGZLZ8e3fG`LW8_3gGfLV8Y3jGkLT8S3lGQMQ8n2PHSMo7m2QHSMP8k2RHUMm7j2THVMm7h2THXMl7g2UHZMk7e2UH[Mk7d2WH[Mj7d2VH]Mj7b2VH^Mk7a2UH`Ml7^2THbMm7]2THbMn7\\2RHeMn7Z2RHfMP8X2PHiMQ8U2oGkMR8T2oGkMS8S2mGnMS8Q2mGoMU8o1kGQNV8n1jGSNW8k1jGTNX8k1gGVNY8i1gGWN[8g1eGYN\\8f1dG[N]8c1dG\\N]8c1cG^N^8`1bG`N_8_1aGaNa8]1_GdNb8Z1_GeNb8Z1^GfNd8X1\\GiNd8V1\\GjNf8T1ZGmNf8R1[GmNg8Q1YGoNi8o0WGROi8m0WGSOk8l0TGTOm8k0TGUOm8i0SGWOn8h0RGYOo8e0QG[OP9c0QG]OQ9a0PG_OQ9?oFAR9=oFDR9:nFFS98nFHT96mFJT93mFMU91kFOW9OiF2Y9JhF6Z9HgF8[9DfF<\\9BdF>_9_OaFb0`9[OaFe0a9YO`Fg0b9UO_Fk0c9SO]Fm0f9QOYFP1g9oNYFQ1g9oNZFP1f9POZFQ1e9oN[FQ1e9POZFQ1e9oN[FQ1e9oN\\FP1d9PO\\FQ1c9PO\\FP1d9PO\\FQ1b9PO^FP1b9PO_Fo0a9RO^Fo0a9QO_FP1`9PO`FP1`9PO`FQ1_9POaFP1^9PObFQ1]9oNcFR1\\9nNdFR1\\9nNdFS1Z9oNfFQ1Y9oNgFR1X9nNhFS1W9mNiFS1W9nNhFS1W9mNjFS1U9mNkFT1T9lNlFU1S9lNlFT1T9lNlFU1R9lNoFT1P9lNPGU1o8lNPGU1o8kNQGV1n8iNSGW1m8hNUGX1j8gNWGZ1h8eNYG\\1f8cN[G^1c8bN^G^1b8iM]F2S1T2`8jM_FOR1W2_8kM_FLS1Y2^8kM`FJS1[2]8kMaFIR1\\2]8kMcFFR1^2[8mMcFCS1`2Z8mMdFAS1b2X8nMfF^OS1d2W8oMgFZOS1g2V8oMhFXOS1i2U8oMWHR2h7oMVHR2j7nMVHR2j7nMUHS2k7mMTHT2k7nMTHR2l7mMTHT2l7fMZHZ2f7_M`Hb2`7WMgHi2g9100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O101N100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O2O0O1O100O100O1MfK`EZ4`:fK_E[4a:eK_E[4a:200000000000000O100000000000000000000000000000M3K5K5M30000000000000O1000001O00001O0000001O000010O00010O000010O00010O00010O01O00010O00010O00010O000010O00010O00010O0001O010O00001O00001O01O01O00001O00001O00001O00001O001O00001O00001O00001O000000001O000000001O00000000001O000000001O00000000001O000001O01O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000001O01O00000M3ZOg0XOg0ZOf0YOUE"}, "label": "the man on the left"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple computer farthest from the camera\"."}], "task": "refering", "answer_template": "The \"the apple computer farthest from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 228, 246, 247, 248, 249, 250, 269, 270, 271, 272], "bbox": [0.7039375, 0.41806249999999995, 0.9429062500000001, 0.6889791666666666], "iscrowd": 0, "area": 9403.609749999998, "rle": {"size": [480, 640], "counts": "ofc62m>1N3M2N2N2O1N2N3M2N2N2O1N2N3M2O10000000001O000O1000001O00000000001O00000000001O0N2N3N1000001O01O01O00O101N10000O100O2OL4M2M4M2M4L4M2M4M3L3N3L3M4M3L3N3L3M4M3L3N3O001O10O010O10O10O100O10000O10O10O100O10000O10O010000O100O10O10O100O1O1O1O1O001O1O100O1M3M3M2N3M3M3M3N2M3M3M2N3M3M3M3M3M3N1N3M3M3M3M3M3M3M3M3GPUa0"}, "label": "the apple computer farthest from the camera"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver laptop closest to the piano\"."}], "task": "refering", "answer_template": "The \"a silver laptop closest to the piano\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 228, 246, 247, 248, 249, 250, 269, 270, 271, 272], "bbox": [0.7039375, 0.41806249999999995, 0.9429062500000001, 0.6889791666666666], "iscrowd": 0, "area": 9403.609749999998, "rle": {"size": [480, 640], "counts": "ofc62m>1N3M2N2N2O1N2N3M2N2N2O1N2N3M2O10000000001O000O1000001O00000000001O00000000001O0N2N3N1000001O01O01O00O101N10000O100O2OL4M2M4M2M4L4M2M4M3L3N3L3M4M3L3N3L3M4M3L3N3O001O10O010O10O10O100O10000O10O10O100O10000O10O010000O100O10O10O100O1O1O1O1O001O1O100O1M3M3M2N3M3M3M3N2M3M3M2N3M3M3M3M3M3N1N3M3M3M3M3M3M3M3M3GPUa0"}, "label": "a silver laptop closest to the piano"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple laptop to the left of another apple laptop\"."}], "task": "refering", "answer_template": "The \"an apple laptop to the left of another apple laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311], "bbox": [0.33978125000000003, 0.4557083333333333, 0.7800156250000001, 0.7936041666666667], "iscrowd": 0, "area": 15984.139099999999, "rle": {"size": [480, 640], "counts": "nQV31o>8G100O10001O0O1000001O000000001O0000001O000O10001O000000001N100000001O0001O01O000000001O000000001O0000001O000000001O000O101O000O100O2O000O101O0O1000001O00000010O000000010OCoNkBQ1b=0DoNjBQ1U=oNkBQ1U=POjBP1V=POkBo0U=ROjBn0V=ROkBn0T=ROlBn0U=ROjBn0W=QOjBn0W=ROhBo0X=QOhBn0Y=ROfBn0Z=ROgBm0Z=SOeBm0[=TOcBm0^=RObBn0^=SOaBm0_=TO`Bl0a=TO^Bm0b=SO\\Bn0e=61O1O2M100O2O000O10O001N1N3N2N1N3N1N3N1O2M3N1N3N1O2N1O2M2O2N2N1N3N1N3N1N3N1N3N1N2O2M2O1O2N1O1O1O2M2N3M2O2O1N110O1O010000O10O100000O010000000O010O100O10O0100O1000O0100O1O1O010O1O1O1O010O1O1O10O01O100O10O01O100O10O100000000O10O100000O100000000O01N2N2N2N1N3N2N2N1O2M3N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2O1N2O0O2N2O1N1O2N2N2N2N101N2N2N2N1O2N2N2N1O2N2N2M5KVlQ2"}, "label": "an apple laptop to the left of another apple laptop"}]}
{"id": 133384, "image": "COCO_train2014_000000133384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple laptop nearest the remote controls\"."}], "task": "refering", "answer_template": "The \"an apple laptop nearest the remote controls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311], "bbox": [0.33978125000000003, 0.4557083333333333, 0.7800156250000001, 0.7936041666666667], "iscrowd": 0, "area": 15984.139099999999, "rle": {"size": [480, 640], "counts": "nQV31o>8G100O10001O0O1000001O000000001O0000001O000O10001O000000001N100000001O0001O01O000000001O000000001O0000001O000000001O000O101O000O100O2O000O101O0O1000001O00000010O000000010OCoNkBQ1b=0DoNjBQ1U=oNkBQ1U=POjBP1V=POkBo0U=ROjBn0V=ROkBn0T=ROlBn0U=ROjBn0W=QOjBn0W=ROhBo0X=QOhBn0Y=ROfBn0Z=ROgBm0Z=SOeBm0[=TOcBm0^=RObBn0^=SOaBm0_=TO`Bl0a=TO^Bm0b=SO\\Bn0e=61O1O2M100O2O000O10O001N1N3N2N1N3N1N3N1O2M3N1N3N1O2N1O2M2O2N2N1N3N1N3N1N3N1N3N1N2O2M2O1O2N1O1O1O2M2N3M2O2O1N110O1O010000O10O100000O010000000O010O100O10O0100O1000O0100O1O1O010O1O1O1O010O1O1O10O01O100O10O01O100O10O100000000O10O100000O100000000O01N2N2N2N1N3N2N2N1O2M3N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2O1N2O0O2N2O1N1O2N2N2N2N101N2N2N2N1O2N2N2N1O2N2N2M5KVlQ2"}, "label": "an apple laptop nearest the remote controls"}]}
{"id": 452873, "image": "COCO_train2014_000000452873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s hands being clasped together\"."}], "task": "refering", "answer_template": "The \"a man ' s hands being clasped together\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 29, 30, 31, 32, 33, 34, 53, 54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 123, 124, 125, 147, 148], "bbox": [0.26882812500000003, 0.0026041666666666665, 0.522078125, 0.4155833333333333], "iscrowd": 0, "area": 14521.2544, "rle": {"size": [480, 640], "counts": "md`22k>4M4L4K4M4L4K4M4L4K4N2O0O01O100O00100O1O10O01O100O2N100O1O1000mCPOV:P1gEWOV:h0hE_OS:a0jEFR::lELP:3nE4n9LoE<l9DRFb0j9^OSFi0j9VOTFP1h9QOTFV1h9mNRFV1n9mNmEV1R:mNhEW1W:lNcEX1\\:kN_EX1`:d110O1O1O10O100000O10O100000000O010000000O01O1O1O1O001O1O1O1O001O1O1O_LhE[2X:V10_LhE[2W:eMjE[2U:eMkE\\2T:dMmE\\2R:eMnE:<0f9FoE9<1d9FPF:<0c9FRF9<1a9FTF9;1`9GTF8=0`9GTF8=1^9GVF8<1^9FWF8<2]9EWF:;2]9EXF8<3\\9DYF8<3\\9DXF:<2[9DZF9<3Z9C[F9<4Y9CZF:=3X9C\\F9=3X9C\\F:<3X9B]F:<4V9B^F:=4U9B^F:=4U9A_F:=4T9B_F;=3T9A`F;=4S9@aF;=5Q9AaF;>4Q9@bF;=5R9_ObF<<5Q9_OdF;<6P9^OdF<=6o8^OdF<=5o8_OeF;=5o8_OdF==3P9_OdF==3o8@eF=<2P9AcF=>1P9AcF>=0P9BdF==0P9BdF><OQ9BcF?=NP9DcF=>NP9DcF>=MQ9DbF?=KS9FaF?<GW9I^F?=DX9J]Fc0<^OZ9N\\Fc0;\\O\\9N\\Ff09XO^90[Fg09TO_93ZFj08oNa95YFk0X;ROkDn0T;QOmDo0T;nNoDQ1a<0O0100O10O0100O010O1O1N1O2N2N1O2O1N1O2N9GXZ_4"}, "label": "a man ' s hands being clasped together"}]}
{"id": 452873, "image": "COCO_train2014_000000452873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child ' s hand holded by an adult hand\"."}], "task": "refering", "answer_template": "The \"a child ' s hand holded by an adult hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 29, 30, 31, 32, 33, 34, 53, 54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 123, 124, 125, 147, 148], "bbox": [0.26882812500000003, 0.0026041666666666665, 0.522078125, 0.4155833333333333], "iscrowd": 0, "area": 14521.2544, "rle": {"size": [480, 640], "counts": "md`22k>4M4L4K4M4L4K4M4L4K4N2O0O01O100O00100O1O10O01O100O2N100O1O1000mCPOV:P1gEWOV:h0hE_OS:a0jEFR::lELP:3nE4n9LoE<l9DRFb0j9^OSFi0j9VOTFP1h9QOTFV1h9mNRFV1n9mNmEV1R:mNhEW1W:lNcEX1\\:kN_EX1`:d110O1O1O10O100000O10O100000000O010000000O01O1O1O1O001O1O1O1O001O1O1O_LhE[2X:V10_LhE[2W:eMjE[2U:eMkE\\2T:dMmE\\2R:eMnE:<0f9FoE9<1d9FPF:<0c9FRF9<1a9FTF9;1`9GTF8=0`9GTF8=1^9GVF8<1^9FWF8<2]9EWF:;2]9EXF8<3\\9DYF8<3\\9DXF:<2[9DZF9<3Z9C[F9<4Y9CZF:=3X9C\\F9=3X9C\\F:<3X9B]F:<4V9B^F:=4U9B^F:=4U9A_F:=4T9B_F;=3T9A`F;=4S9@aF;=5Q9AaF;>4Q9@bF;=5R9_ObF<<5Q9_OdF;<6P9^OdF<=6o8^OdF<=5o8_OeF;=5o8_OdF==3P9_OdF==3o8@eF=<2P9AcF=>1P9AcF>=0P9BdF==0P9BdF><OQ9BcF?=NP9DcF=>NP9DcF>=MQ9DbF?=KS9FaF?<GW9I^F?=DX9J]Fc0<^OZ9N\\Fc0;\\O\\9N\\Ff09XO^90[Fg09TO_93ZFj08oNa95YFk0X;ROkDn0T;QOmDo0T;nNoDQ1a<0O0100O10O0100O010O1O1N1O2N2N1O2O1N1O2N9GXZ_4"}, "label": "a child ' s hand holded by an adult hand"}]}
{"id": 452873, "image": "COCO_train2014_000000452873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the torso and hands of a person\"."}], "task": "refering", "answer_template": "The \"the torso and hands of a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 28, 29, 30, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 51, 52, 53, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.21573437499999998, 0.0, 0.9977499999999999, 0.99775], "iscrowd": 0, "area": 161324.5888, "rle": {"size": [480, 640], "counts": "ifP22l>4L4L4M3L5K4L4L4L5L3GgNeB^1X=cNeBa1Y=5N3M2N3M2N1O001O1O1O001O1O1O001TDkMk:V2eDZNZ;g2O1O2N_McD`1\\;^NgDb1X;[NmDd1R;ZNQEf1n:XNUEh1j:VNZEi1e:UN^Ek1a:RNcEo1[:oMiEP2V:nMmER2R:mMQFR2n9mMSFT2l9kMTFW2k9gMWFZ2h9eMXF]2g9bMYF`2f9_MZFc2Q;0O2N2N2N2O1N1O2O1N2N2O1N2O0O2NbNaD^O^;?hD_OW;=QE@n:=YE@f:=`E@a:<fEAY:<mEBR:;SFDm98XFGg97\\FIc94bFK^92fFL]9OfF1]9IgF6\\9DhF;[9_OhFa0m;O1N101O1OO1aKA\\J?c5B]J=c5E\\J;c5F\\J;c5F]J:b5G^J9a5H_J8`5I_J8`5I`J6`5K`J5_5LaJ4^5MaJ4^5LcJ4\\5MdJ3[5MfJ2Z5NgJ2X5NhJ3W5MjJ3U5NkJ2T5NmJ2R5NnJ2R5NoJ3o4MRK3m4NSK3k4MUK5i4KXK6f4J[K6d4J]K7a4J_K7_4IbK8\\4HeK8Z4HgK9W4GjK:T4FmK=o3DUJKbNf0S7_O[JKcNh0o6]O^JLcNf0o6^O^JMcNd0o6@]JLeNc0n6A]JMeNa0n6CnH]O0`04?n6EeHE885=n6FfHD7:5:o6IdHD7:68o6KcHC8<46Q7KcHD7<54Q7MbHD8<42R7ObHB8>40R71aHB9>3NS72aHB9?3KT75_HB:>3JT77^HB:?3FV79]HB;?2CX7=[HA:`03^O[7a0XHA;`02[O]7e0UHA<?2WO`7i0RHB;?3ROc7m0oGB<?1UOa7k0RHA;`02VO^7i0UHA<`00YO\\7g0WHA<?2[OX7e0ZHB<>1^OV7b0]HB=>OAT7`0_HB=>0:Y6F[IA=?Ne0o5\\OeIA=?Nl0h5TOmIB=>Nn0e5SOoIA>?MP1c5PORJB>>MR1_5oNWJ@>?KU1]5mNYJ@>?KX1Z5iN]J@??JZ1W5hN_J@??J\\1U5eNbJA?>I_1S5cNeJ_O?`0Gb1R5`NgJ_O`0?Eg1P5\\NjJ^Oa0`0Bk1P5WNmJ_Oa0?_OP2P5SNoJ_Ob0>]OT2o4PNPK_Od0>ZOX2o4lMPKAg0<WO\\2n4gMRKCi0a0nNZ2T5cMRKEk0b4P4jKRKGm0b4n3hKQKIQ1b4k3fKQKKS1c4h3bKSKMU1d4d3oL\\LS3a3nL^LT3`3lL`LU3_3lL`LU3_3kLaLV3^3kLaLV3^3kLaLW3]3iLcLX3\\3iLdLW3[3iLeLX3Z3iLeLX3Z3iLeLX3Z3iLeLY3Y3hLfLY3Y3hLfLY3Y3hLgLX3X3iLgLX3X3hLhLZ3V3gLiLZ3V3hLgLZ3X3hLdL\\3Z3gLaL`3Z3bLbLd3Z3_L`Li3[3R4L5K4L5K2N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1Oh0XO001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000lH`0hN@eMW67\\6mMT2lM8"}, "label": "the torso and hands of a person"}]}
{"id": 452873, "image": "COCO_train2014_000000452873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stomach of a man wearing a shirt and tie\"."}], "task": "refering", "answer_template": "The \"the stomach of a man wearing a shirt and tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 28, 29, 30, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 51, 52, 53, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.21573437499999998, 0.0, 0.9977499999999999, 0.99775], "iscrowd": 0, "area": 161324.5888, "rle": {"size": [480, 640], "counts": "ifP22l>4L4L4M3L5K4L4L4L5L3GgNeB^1X=cNeBa1Y=5N3M2N3M2N1O001O1O1O001O1O1O001TDkMk:V2eDZNZ;g2O1O2N_McD`1\\;^NgDb1X;[NmDd1R;ZNQEf1n:XNUEh1j:VNZEi1e:UN^Ek1a:RNcEo1[:oMiEP2V:nMmER2R:mMQFR2n9mMSFT2l9kMTFW2k9gMWFZ2h9eMXF]2g9bMYF`2f9_MZFc2Q;0O2N2N2N2O1N1O2O1N2N2O1N2O0O2NbNaD^O^;?hD_OW;=QE@n:=YE@f:=`E@a:<fEAY:<mEBR:;SFDm98XFGg97\\FIc94bFK^92fFL]9OfF1]9IgF6\\9DhF;[9_OhFa0m;O1N101O1OO1aKA\\J?c5B]J=c5E\\J;c5F\\J;c5F]J:b5G^J9a5H_J8`5I_J8`5I`J6`5K`J5_5LaJ4^5MaJ4^5LcJ4\\5MdJ3[5MfJ2Z5NgJ2X5NhJ3W5MjJ3U5NkJ2T5NmJ2R5NnJ2R5NoJ3o4MRK3m4NSK3k4MUK5i4KXK6f4J[K6d4J]K7a4J_K7_4IbK8\\4HeK8Z4HgK9W4GjK:T4FmK=o3DUJKbNf0S7_O[JKcNh0o6]O^JLcNf0o6^O^JMcNd0o6@]JLeNc0n6A]JMeNa0n6CnH]O0`04?n6EeHE885=n6FfHD7:5:o6IdHD7:68o6KcHC8<46Q7KcHD7<54Q7MbHD8<42R7ObHB8>40R71aHB9>3NS72aHB9?3KT75_HB:>3JT77^HB:?3FV79]HB;?2CX7=[HA:`03^O[7a0XHA;`02[O]7e0UHA<?2WO`7i0RHB;?3ROc7m0oGB<?1UOa7k0RHA;`02VO^7i0UHA<`00YO\\7g0WHA<?2[OX7e0ZHB<>1^OV7b0]HB=>OAT7`0_HB=>0:Y6F[IA=?Ne0o5\\OeIA=?Nl0h5TOmIB=>Nn0e5SOoIA>?MP1c5PORJB>>MR1_5oNWJ@>?KU1]5mNYJ@>?KX1Z5iN]J@??JZ1W5hN_J@??J\\1U5eNbJA?>I_1S5cNeJ_O?`0Gb1R5`NgJ_O`0?Eg1P5\\NjJ^Oa0`0Bk1P5WNmJ_Oa0?_OP2P5SNoJ_Ob0>]OT2o4PNPK_Od0>ZOX2o4lMPKAg0<WO\\2n4gMRKCi0a0nNZ2T5cMRKEk0b4P4jKRKGm0b4n3hKQKIQ1b4k3fKQKKS1c4h3bKSKMU1d4d3oL\\LS3a3nL^LT3`3lL`LU3_3lL`LU3_3kLaLV3^3kLaLV3^3kLaLW3]3iLcLX3\\3iLdLW3[3iLeLX3Z3iLeLX3Z3iLeLX3Z3iLeLY3Y3hLfLY3Y3hLfLY3Y3hLgLX3X3iLgLX3X3hLhLZ3V3gLiLZ3V3hLgLZ3X3hLdL\\3Z3gLaL`3Z3bLbLd3Z3_L`Li3[3R4L5K4L5K2N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1Oh0XO001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000lH`0hN@eMW67\\6mMT2lM8"}, "label": "the stomach of a man wearing a shirt and tie"}]}
{"id": 542988, "image": "COCO_train2014_000000542988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right who has more dark stripes\"."}], "task": "refering", "answer_template": "The \"the zebra on the right who has more dark stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 202, 214, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 259, 260, 261, 262, 265, 266, 267, 283, 284, 285, 288, 289, 290, 306, 307, 308, 311, 312, 313], "bbox": [0.278875, 0.29894613583138174, 0.8292343750000001, 0.9236065573770492], "iscrowd": 0, "area": 28992.54049999999, "rle": {"size": [427, 640], "counts": "l`Z21X=4L4L5K4L4K5I8H7K4N2N2N000100O1O100O1O100O1O1N24Lg0YO<D;E2N1O2N1O1O2QFlLc9U3[FmLd9U3YFlLg9U3VFnLi9[3O000O0100O0dNPG_NP9l0gGQOZ8m0iGROW8l0lGROX8h0mGVOU;N2O2M2O1O2MlU6j0ZiI<C<D5K6KUOeDKV;1RE1h:L_E4]:HlE6Q:FWF9e9DcF;Y9AQG=k8_O^G`0_8]OhGa0U8_OQH?b7\\O]F5V2>[7^O\\F0b2a0P7@YI?e6C]I:b6G_I9`6GbI7]6JeI4Z6MhI1W60jIOU62mILh5LTG:U3He5d0\\J\\Ob5f0_JXO`5j0^33N1N3L4M2M4L4M2M4M3L3O2N2M2O2N1O2N1N3N1O1M4K4O2O0100O001O001O1O002N1O1O1O100O1O1O1N1000000000O100O1VOfMaF\\2_9fM^F[2a9gM]FZ2c9hMZFY2f9iMUFZ2k9hMoE\\2Q:?0O1000000000000O100000000O1000000O1000000OYHUMP5k2lJZMS5f2jJ^MU5b2fJdMY5\\2cJhM]5X2\\JPNb5Q2VJYNh5g1RJ`NfNaNQ7o2TJkNl5U1nIQOR6P1iITOW6l0dIYO\\6f0`InN@\\NQ7f2\\IgNNbNe6h2ZIfN6`N`6j2WIfN=_N\\6l2RIfNf0]NX6j4mITKR6j4RJUKn5i4WJTKh5k4]1N2N2M3N1013M2_HXK[6m4UI_Kj6c5N1O1O1N3N1O2mHiIn6Z60001H7MZO_IhJ_6e4ZJWKc5i4dJSKZ5l4kJRKS5n4PKQKn4n4VKQKj4]4SJ]KW14e4X4TLgKk3T4\\LkKd3g3jI\\Lg2L^3h3nLVLS3j3V300010O010O00010O010O01O0N2O2N1N3N1O2N1O2N1O2O001O1O001O001N101O1N101O0O2O001N2O001O1N101O1O000O2O0000002M2O2N2N001O0O10001O101O0000O1O101O001O1O10O0100O00100O10O0100O010O10nMXE_1i:`NYE_1g:`NZE`1f:_N[E`1f:_N\\E`1d:_N]Ea1c:^N^Ea1c:]N`Eb1`:]NaEc1_:\\NbEc1_:[NdEd1\\:ZNfEf1Z:YNhEg1n:2OO0O2N2N2N2N2N2M3N2N1O1O001O1O1O1O1N2O1N2O1N2M3L9H7HlW]1"}, "label": "the zebra on the right who has more dark stripes"}]}
{"id": 542988, "image": "COCO_train2014_000000542988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra looking forward\"."}], "task": "refering", "answer_template": "The \"the zebra looking forward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 202, 214, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 259, 260, 261, 262, 265, 266, 267, 283, 284, 285, 288, 289, 290, 306, 307, 308, 311, 312, 313], "bbox": [0.278875, 0.29894613583138174, 0.8292343750000001, 0.9236065573770492], "iscrowd": 0, "area": 28992.54049999999, "rle": {"size": [427, 640], "counts": "l`Z21X=4L4L5K4L4K5I8H7K4N2N2N000100O1O100O1O100O1O1N24Lg0YO<D;E2N1O2N1O1O2QFlLc9U3[FmLd9U3YFlLg9U3VFnLi9[3O000O0100O0dNPG_NP9l0gGQOZ8m0iGROW8l0lGROX8h0mGVOU;N2O2M2O1O2MlU6j0ZiI<C<D5K6KUOeDKV;1RE1h:L_E4]:HlE6Q:FWF9e9DcF;Y9AQG=k8_O^G`0_8]OhGa0U8_OQH?b7\\O]F5V2>[7^O\\F0b2a0P7@YI?e6C]I:b6G_I9`6GbI7]6JeI4Z6MhI1W60jIOU62mILh5LTG:U3He5d0\\J\\Ob5f0_JXO`5j0^33N1N3L4M2M4L4M2M4M3L3O2N2M2O2N1O2N1N3N1O1M4K4O2O0100O001O001O1O002N1O1O1O100O1O1O1N1000000000O100O1VOfMaF\\2_9fM^F[2a9gM]FZ2c9hMZFY2f9iMUFZ2k9hMoE\\2Q:?0O1000000000000O100000000O1000000O1000000OYHUMP5k2lJZMS5f2jJ^MU5b2fJdMY5\\2cJhM]5X2\\JPNb5Q2VJYNh5g1RJ`NfNaNQ7o2TJkNl5U1nIQOR6P1iITOW6l0dIYO\\6f0`InN@\\NQ7f2\\IgNNbNe6h2ZIfN6`N`6j2WIfN=_N\\6l2RIfNf0]NX6j4mITKR6j4RJUKn5i4WJTKh5k4]1N2N2M3N1013M2_HXK[6m4UI_Kj6c5N1O1O1N3N1O2mHiIn6Z60001H7MZO_IhJ_6e4ZJWKc5i4dJSKZ5l4kJRKS5n4PKQKn4n4VKQKj4]4SJ]KW14e4X4TLgKk3T4\\LkKd3g3jI\\Lg2L^3h3nLVLS3j3V300010O010O00010O010O01O0N2O2N1N3N1O2N1O2N1O2O001O1O001O001N101O1N101O0O2O001N2O001O1N101O1O000O2O0000002M2O2N2N001O0O10001O101O0000O1O101O001O1O10O0100O00100O10O0100O010O10nMXE_1i:`NYE_1g:`NZE`1f:_N[E`1f:_N\\E`1d:_N]Ea1c:^N^Ea1c:]N`Eb1`:]NaEc1_:\\NbEc1_:[NdEd1\\:ZNfEf1Z:YNhEg1n:2OO0O2N2N2N2N2N2M3N2N1O1O001O1O1O1O1N2O1N2O1N2M3L9H7HlW]1"}, "label": "the zebra looking forward"}]}
{"id": 542988, "image": "COCO_train2014_000000542988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra\"."}], "task": "refering", "answer_template": "The \"a zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 118, 119, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 236, 237, 238, 239, 254, 255, 259, 260, 261, 262, 278, 283, 284, 285, 301, 306, 307, 308], "bbox": [0.0255625, 0.24679156908665104, 0.52234375, 0.9241217798594848], "iscrowd": 0, "area": 39742.53145, "rle": {"size": [427, 640], "counts": "`m66T=4K5L2N110O0010O001O0000N2M3M3N0N2N3M2N110O100O10O001O1O001O1O1O12N5J6UFoNT7U1hHYOj6l0QIC`6b0[IH[6=`I<g5ITJ:j5IRJ:m5InI:P6JkI8T6LfI8X6KcI9[6K^I:a6IYI<KlLa6k2_I=KoLb6_6O2O0O2N1000000O10kNgIXKY6k3hJQLW5n3PKmKP5P4VKmKj4Q4ZKnKe4o3`KoK_4P4eKmK\\4P4iKnKW4o3oKnKP4Q4SLnKm3P4WLnKi3P4ZLoKf3o3]LPLc3n3aLPL_3n3P3N1O2N2N2N2N2N1O2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1O1O2O0O1O100O1100O001O1O100O001O1O1O10O01O1O100O001O100O1O010O000010O000010O0000010O0000010O0001O000001O01O0000001O000000001O000000001O0001cFnL_8R3^GUM_8j2]G]M`8c2]GdM_8\\2^GiM`8W2\\GPNa8P2[GUNd8k1WG[Nh8T300000000O0100N2N2N2M3N1012N1Og0YOo0QO2M3N1O0O2O1O1SIjIc6W6ZIlId6_6O00O0100O1O\\N^IaLa6\\3WJPLg5o3bJjK^5T4jJgKU5V4SKfKl4Y4XKeKg4Z4]KdKb4]4_KbK`4j4[J^KBFS6R5nIbK0YOP6_6PJ_Io5c6RJ[Im5h692eIWIQ6R7O1O1N2O1O1O1O100O1O1O10O01O1O1O100O1OUNfJgKY5V4mJhKS5T4TKiKk4U4ZKiKe4T4aKjK_4Q4iKkKW4R4oKlKP4P4XLmKg3P4_LnKa3n3eLPLZ3m3mLPLR3m3TMQLl2k3ZMSLe2j3aMTL^2l3fMQLZ2m3iMRLV2n3lMQLS2n3PNPLZ1LkKT4m2oKX11gKn3T3QLT14dKk3Z3PLR19^Kg3c3oKP1a4SO^Kn0^4VOaKk0\\4YObKi0Y4W4M3M3L5G7J6M3N2N1O1O1O1O1O1O100O1O1O1O1O200001O0O2000O10O01001O1N101O1OO100O1OO1O1O1O1OdMoE`1P:`NSF]1m9cNVF[1h9fN\\FU1e9lN]FP1e9oN^Fm0d9SO^FUOb0>R9<fG_O\\8`0gG\\O[8`0lG[OY8=`2Hj_o3"}, "label": "a zebra"}]}
{"id": 542988, "image": "COCO_train2014_000000542988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra looking towards the camera\"."}], "task": "refering", "answer_template": "The \"a zebra looking towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 118, 119, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 236, 237, 238, 239, 254, 255, 259, 260, 261, 262, 278, 283, 284, 285, 301, 306, 307, 308], "bbox": [0.0255625, 0.24679156908665104, 0.52234375, 0.9241217798594848], "iscrowd": 0, "area": 39742.53145, "rle": {"size": [427, 640], "counts": "`m66T=4K5L2N110O0010O001O0000N2M3M3N0N2N3M2N110O100O10O001O1O001O1O1O12N5J6UFoNT7U1hHYOj6l0QIC`6b0[IH[6=`I<g5ITJ:j5IRJ:m5InI:P6JkI8T6LfI8X6KcI9[6K^I:a6IYI<KlLa6k2_I=KoLb6_6O2O0O2N1000000O10kNgIXKY6k3hJQLW5n3PKmKP5P4VKmKj4Q4ZKnKe4o3`KoK_4P4eKmK\\4P4iKnKW4o3oKnKP4Q4SLnKm3P4WLnKi3P4ZLoKf3o3]LPLc3n3aLPL_3n3P3N1O2N2N2N2N2N1O2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1O1O2O0O1O100O1100O001O1O100O001O1O1O10O01O1O100O001O100O1O010O000010O000010O0000010O0000010O0001O000001O01O0000001O000000001O000000001O0001cFnL_8R3^GUM_8j2]G]M`8c2]GdM_8\\2^GiM`8W2\\GPNa8P2[GUNd8k1WG[Nh8T300000000O0100N2N2N2M3N1012N1Og0YOo0QO2M3N1O0O2O1O1SIjIc6W6ZIlId6_6O00O0100O1O\\N^IaLa6\\3WJPLg5o3bJjK^5T4jJgKU5V4SKfKl4Y4XKeKg4Z4]KdKb4]4_KbK`4j4[J^KBFS6R5nIbK0YOP6_6PJ_Io5c6RJ[Im5h692eIWIQ6R7O1O1N2O1O1O1O100O1O1O10O01O1O1O100O1OUNfJgKY5V4mJhKS5T4TKiKk4U4ZKiKe4T4aKjK_4Q4iKkKW4R4oKlKP4P4XLmKg3P4_LnKa3n3eLPLZ3m3mLPLR3m3TMQLl2k3ZMSLe2j3aMTL^2l3fMQLZ2m3iMRLV2n3lMQLS2n3PNPLZ1LkKT4m2oKX11gKn3T3QLT14dKk3Z3PLR19^Kg3c3oKP1a4SO^Kn0^4VOaKk0\\4YObKi0Y4W4M3M3L5G7J6M3N2N1O1O1O1O1O1O100O1O1O1O1O200001O0O2000O10O01001O1N101O1OO100O1OO1O1O1O1OdMoE`1P:`NSF]1m9cNVF[1h9fN\\FU1e9lN]FP1e9oN^Fm0d9SO^FUOb0>R9<fG_O\\8`0gG\\O[8`0lG[OY8=`2Hj_o3"}, "label": "a zebra looking towards the camera"}]}
{"id": 567566, "image": "COCO_train2014_000000567566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of bench unoccupied by woman\"."}], "task": "refering", "answer_template": "The \"half of bench unoccupied by woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 234, 235, 236, 237, 238, 239, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 312, 313, 327, 328], "bbox": [0.5793208430913349, 0.608984375, 1.0, 0.9303437499999999], "iscrowd": 0, "area": 19774.4594, "rle": {"size": [640, 427], "counts": "]jj41oc00O2N1O2N1O1O2O0O2N1O2N1O2O0O2N1O2N1O2N14M0ON3M2N3M2N3N1N3M2N3M2N2N3M2N3M2N3N1010O01O010O001O010O0010O01O010O0010O01b@PNj<P2RCTNn<l1mBYNT=g1gB]NY=c1cBbN\\=_1^BfNc=Y1YBkNg=V1TBnNl=R1oATOQ>l0jAXOV>h0fA\\OZ>e0`A@`>`0\\AEd>;WAIi>7RANn>3m@2S?Mi@7W?Y20010O0001O01O0001O01O01O01O01O0001O01O00010O000010N1M3N3M2M3N2N3M2M3N2N3M2M3N2N3L3N2N3M2M3O101O01O00010O000010O00010O0001O01O01O00010O000O1K6J5L4K6J5L4K6J5L4K6J5K5L5J5K5L5J5K5L4K6I6J6J7I6J6J7I6K5J7I6J6J7I6J6J^C"}, "label": "half of bench unoccupied by woman"}]}
{"id": 567566, "image": "COCO_train2014_000000567566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty part of a park bench a woman is sitting on\"."}], "task": "refering", "answer_template": "The \"the empty part of a park bench a woman is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 234, 235, 236, 237, 238, 239, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 312, 313, 327, 328], "bbox": [0.5793208430913349, 0.608984375, 1.0, 0.9303437499999999], "iscrowd": 0, "area": 19774.4594, "rle": {"size": [640, 427], "counts": "]jj41oc00O2N1O2N1O1O2O0O2N1O2N1O2O0O2N1O2N1O2N14M0ON3M2N3M2N3N1N3M2N3M2N2N3M2N3M2N3N1010O01O010O001O010O0010O01O010O0010O01b@PNj<P2RCTNn<l1mBYNT=g1gB]NY=c1cBbN\\=_1^BfNc=Y1YBkNg=V1TBnNl=R1oATOQ>l0jAXOV>h0fA\\OZ>e0`A@`>`0\\AEd>;WAIi>7RANn>3m@2S?Mi@7W?Y20010O0001O01O0001O01O01O01O01O0001O01O00010O000010N1M3N3M2M3N2N3M2M3N2N3M2M3N2N3L3N2N3M2M3O101O01O00010O000010O00010O0001O01O01O00010O000O1K6J5L4K6J5L4K6J5L4K6J5K5L5J5K5L5J5K5L4K6I6J6J7I6J6J7I6K5J7I6J6J7I6J6J^C"}, "label": "the empty part of a park bench a woman is sitting on"}]}
{"id": 567566, "image": "COCO_train2014_000000567566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"public bench with noone sitting on it\"."}], "task": "refering", "answer_template": "The \"public bench with noone sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 181, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230], "bbox": [0.05063231850117096, 0.5078125, 0.4311709601873537, 0.69625], "iscrowd": 0, "area": 12549.862550000002, "rle": {"size": [640, 427], "counts": "ZT>9^c09G:F:F9O2O001O1O001O001O001O1O001O001O001O1O001O001O1O001O010O001O1O001O001O1O001O001O001O1O001O001N1N3N2N1O2N1O2M3N1O2N1O2M2O1O2N1O1O1N2O1O1O2N1N2O1O1O1O100010O000000001O01O0001O000000010O000000001O01O0001O000000010N1O1N2O1O2N1O1O1O1N3N1O1O1O2N1O1N2O1N3M2M3N2N2M4M2N2N2M4M2N2M3N2N3L3N2N2M3N3M2M3N2N3L3N2N2N2M4M2N2M3N2N3L\\ag4"}, "label": "public bench with noone sitting on it"}]}
{"id": 567566, "image": "COCO_train2014_000000567566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty black bench in the park\"."}], "task": "refering", "answer_template": "The \"the empty black bench in the park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 181, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230], "bbox": [0.05063231850117096, 0.5078125, 0.4311709601873537, 0.69625], "iscrowd": 0, "area": 12549.862550000002, "rle": {"size": [640, 427], "counts": "ZT>9^c09G:F:F9O2O001O1O001O001O001O1O001O001O001O1O001O001O1O001O010O001O1O001O001O1O001O001O001O1O001O001N1N3N2N1O2N1O2M3N1O2N1O2M2O1O2N1O1O1N2O1O1O2N1N2O1O1O1O100010O000000001O01O0001O000000010O000000001O01O0001O000000010N1O1N2O1O2N1O1O1O1N3N1O1O1O2N1O1N2O1N3M2M3N2N2M4M2N2N2M4M2N2M3N2N3L3N2N2M3N3M2M3N2N3L3N2N2N2M4M2N2M3N2N3L\\ag4"}, "label": "the empty black bench in the park"}]}
{"id": 174357, "image": "COCO_train2014_000000174357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red and yellow surf board\"."}], "task": "refering", "answer_template": "The \"the red and yellow surf board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 31, 32, 33, 44, 45, 46, 57, 58, 59, 69, 70, 71, 72, 82, 83, 84, 85], "bbox": [0.36256000000000005, 0.06966, 0.58128, 0.38652], "iscrowd": 0, "area": 9248.327099999999, "rle": {"size": [500, 375], "counts": "mbR24Y?7I7H8I7H8I7I7H8I7H8J6N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2N2N2O1O1N2O1O1N2O1O1O1N200O10000O10000O10000O100O10000O10000O10000O1000O3E;E:F;F:E;Eh0XO;E:F;E;E:F;EVc\\2"}, "label": "the red and yellow surf board"}]}
{"id": 174357, "image": "COCO_train2014_000000174357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and yellow surfboard in the back of a car\"."}], "task": "refering", "answer_template": "The \"a red and yellow surfboard in the back of a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 31, 32, 33, 44, 45, 46, 57, 58, 59, 69, 70, 71, 72, 82, 83, 84, 85], "bbox": [0.36256000000000005, 0.06966, 0.58128, 0.38652], "iscrowd": 0, "area": 9248.327099999999, "rle": {"size": [500, 375], "counts": "mbR24Y?7I7H8I7H8I7I7H8I7H8J6N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2N2N2O1O1N2O1O1N2O1O1O1N200O10000O10000O10000O100O10000O10000O10000O1000O3E;E:F;F:E;Eh0XO;E:F;E;E:F;EVc\\2"}, "label": "a red and yellow surfboard in the back of a car"}]}
{"id": 92439, "image": "COCO_train2014_000000092439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather couch\"."}], "task": "refering", "answer_template": "The \"a black leather couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274], "bbox": [0.589375, 0.3943125, 0.983625, 0.6944375], "iscrowd": 0, "area": 19007.73705, "rle": {"size": [480, 640], "counts": "hn`52f>9G8H9F9E;J7O0O2O0O2O0O101N101O001O000O2O001O001O000O2O00001O00000O2O00001O0000001N1000001O00001N1000001O0000001OeNdC2\\<JhC6X<FlC;S<AQD?o;]OUDc0k;YOYDg0j;QO[DP1j<0000001O00001O0oBeN]<[1[CmNe<S1ZCoNe<R1ZCnNf<R1ZCnNf<R1ZCnNf<R1ZCoNf<Q1XCPOh<P1XCPOh<T1TCmNk<d10001O00000010O0001O00010O0000010O00001O01O01O0001O01O00010O00001O01O0001N10001O00001O0001O01O0010O00010O0001O010O000010O00010O00010O010O0010O01O01O010O01O010OO3I7I7I7J6N000O1000000O1000000O1000000O1000O010O1000O0100O10O010000O010O10O10O100O010O10O10O10O0100O01000O010O10O01000O0100O011N10000O2O0O101O0O1M3K6K4K5L4K6K4K5L5J5K5L4K6K4K5L5J5L4K5LW_4"}, "label": "a black leather couch"}]}
{"id": 92439, "image": "COCO_train2014_000000092439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large black leather sofa on left side of room\"."}], "task": "refering", "answer_template": "The \"large black leather sofa on left side of room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.45510416666666664, 0.497265625, 1.0], "iscrowd": 0, "area": 58964.12790000001, "rle": {"size": [480, 640], "counts": "f9X1Z;_2G9O2N1O1O2M2O1O1O2N1N2O2N1O1O1N3N1O1O2M2O1O2N1O1N2O2N1O1O2M21O00100O00100O0010O01O1O010O1O010O00M3N3M2N2N3L3N2N3M2N2M3N3M2N2N3L3N2N3M200O101O000O2O00000O2O000O2O000O2M2O1O1O2N1O1N3N1O1O1O2N1N2O2N1O1O1O000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000001O1O001O1O001O001O1O001O2N2N2N3M3M;E;E8H001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001N101N101N101N101O0O2O0O2O0O2O0O2O00O0100O10O0100O10O010000O010O100O010O1000O0100O10O0100O10O10O100O010O100O01000O100O010O100O010O1000O0100O10O0100O10O10O100O2O1N101N2O001N101N2O0O2O1N101O1N101N2O0O2O1N101O0O2Oi0VOT1mNR1mN[af4"}, "label": "large black leather sofa on left side of room"}]}
{"id": 92439, "image": "COCO_train2014_000000092439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa with a white covering on the seat\"."}], "task": "refering", "answer_template": "The \"a sofa with a white covering on the seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.45510416666666664, 0.497265625, 1.0], "iscrowd": 0, "area": 58964.12790000001, "rle": {"size": [480, 640], "counts": "f9X1Z;_2G9O2N1O1O2M2O1O1O2N1N2O2N1O1O1N3N1O1O2M2O1O2N1O1N2O2N1O1O2M21O00100O00100O0010O01O1O010O1O010O00M3N3M2N2N3L3N2N3M2N2M3N3M2N2N3L3N2N3M200O101O000O2O00000O2O000O2O000O2M2O1O1O2N1O1N3N1O1O1O2N1N2O2N1O1O1O000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000001O1O001O1O001O001O1O001O2N2N2N3M3M;E;E8H001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001N101N101N101N101O0O2O0O2O0O2O0O2O00O0100O10O0100O10O010000O010O100O010O1000O0100O10O0100O10O10O100O010O100O01000O100O010O100O010O1000O0100O10O0100O10O10O100O2O1N101N2O001N101N2O0O2O1N101O1N101N2O0O2O1N101O0O2Oi0VOT1mNR1mN[af4"}, "label": "a sofa with a white covering on the seat"}]}
{"id": 493851, "image": "COCO_train2014_000000493851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lounge chair under the umbrella , and closest towards the camera\"."}], "task": "refering", "answer_template": "The \"the lounge chair under the umbrella , and closest towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 280, 281, 282, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 331, 332, 333, 351], "bbox": [0.1695, 0.6755208333333333, 0.512375, 0.9062083333333334], "iscrowd": 0, "area": 9459.998800000003, "rle": {"size": [480, 640], "counts": "]^c14j>4N2N2N2N2O1N1O2N2N2N2N2N2N2N1O2N2N2O1N2N2N2N1O1O1O1hN`NnDa1P;bNnD_1P;dNnD\\1R;fNlDZ1U;gNiDY1W;iNhDV1X;nNdDR1];oNaDQ1_;QO_Dn0b;TO\\Dl0d;VOZDj0g;WOWDh0j;ZOSDe0o;]OnCc0S<_OjCa0X<@dCa0]<P11N3N11O2N101N2N1O2O1N2O0O2O1O001O1O0100N2O0O2N2N2N1O2M3N2N1O2N2N2N1N3M3M3L3N3M2N1N2O1O1O0O2O1O1O1N2O1O1O1N2O1O001N2O1O1N2O1O1O1N2O1O01000000000O1000000000O1000O1000000000000000O011O1O1O1O1O1O1O1O1O1N2O1O1O1O1O0000O100O1O001N2O1O1N2O1O1O11O000O10001O1O1O1O2N1O1O1O2M2O1O1O2N1O100O2N1O100O2N1O100O000001O0O101O0N2N2N3M2O1N2N3M2N2N3M2N2N2O2M2N2N2N3M2N4LXla4"}, "label": "the lounge chair under the umbrella , and closest towards the camera"}]}
{"id": 493851, "image": "COCO_train2014_000000493851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lounge chair on the right\"."}], "task": "refering", "answer_template": "The \"lounge chair on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 280, 281, 282, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 331, 332, 333, 351], "bbox": [0.1695, 0.6755208333333333, 0.512375, 0.9062083333333334], "iscrowd": 0, "area": 9459.998800000003, "rle": {"size": [480, 640], "counts": "]^c14j>4N2N2N2N2O1N1O2N2N2N2N2N2N2N1O2N2N2O1N2N2N2N1O1O1O1hN`NnDa1P;bNnD_1P;dNnD\\1R;fNlDZ1U;gNiDY1W;iNhDV1X;nNdDR1];oNaDQ1_;QO_Dn0b;TO\\Dl0d;VOZDj0g;WOWDh0j;ZOSDe0o;]OnCc0S<_OjCa0X<@dCa0]<P11N3N11O2N101N2N1O2O1N2O0O2O1O001O1O0100N2O0O2N2N2N1O2M3N2N1O2N2N2N1N3M3M3L3N3M2N1N2O1O1O0O2O1O1O1N2O1O1O1N2O1O001N2O1O1N2O1O1O1N2O1O01000000000O1000000000O1000O1000000000000000O011O1O1O1O1O1O1O1O1O1N2O1O1O1O1O0000O100O1O001N2O1O1N2O1O1O11O000O10001O1O1O1O2N1O1O1O2M2O1O1O2N1O100O2N1O100O2N1O100O000001O0O101O0N2N2N3M2O1N2N3M2N2N3M2N2N2O2M2N2N2N3M2N4LXla4"}, "label": "lounge chair on the right"}]}
{"id": 493851, "image": "COCO_train2014_000000493851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beach chair nearest the tree / shade\"."}], "task": "refering", "answer_template": "The \"beach chair nearest the tree / shade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [256, 257, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 326], "bbox": [0.0965, 0.6532291666666666, 0.40354687499999997, 0.8411041666666667], "iscrowd": 0, "area": 3457.123700000002, "rle": {"size": [480, 640], "counts": "n\\m03k>3M4N2N2O0O2N2O1N1O2N2O_OhA:V>GkA8T>HnA8Q>HPB8n=HTB8j=IWB6i=IYB7e=J\\B6b=J`B6^=KcB4]=LdB4T<DaD8]O4P<FaD7_O4n;GaD5C3k;J_D5GOj;N]D4JLj;2XD5NGj;6UD61Bj;:RD62Al;;oC71_OQ<<kC81[OV<?fC90YO\\<?aC:1VO`<b0\\CK1I0Jd<X1^CnNNFf<]1\\CnNh<R1ZCmNe<S1\\CnNb<S1_CmN`<R1bCnN\\<R1eCnNZ<S1gCmNX<R1jCnNU<Q1lCPOS<n0PDQOQ<m0QDSOn;l0SDUOm;h0VDWOj;h0XDVOi;i0XDVOj;h0XDVOi;i0YDUOh;j0YDVOh;h0ZDVOg;i0YDWOi;g0VDYOl;f0SD[On;d0RD\\OP<c0nC]OT<b0nC\\OS<c0oC[OS<c0V1O2N1O1O2N1O2N1O100O0000O10000000000000O10000O102N3Lj`>0W_A2N2M2O2N000000O1000O10000000000000000O100000000000000O100000000000001N1000000000000000100O10O001O001N101O001O0O2O0001O001O005LO00000O101O1O1M3N2N2Mgfb5"}, "label": "beach chair nearest the tree / shade"}]}
{"id": 493851, "image": "COCO_train2014_000000493851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue lounge chair on left\"."}], "task": "refering", "answer_template": "The \"blue lounge chair on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [256, 257, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 326], "bbox": [0.0965, 0.6532291666666666, 0.40354687499999997, 0.8411041666666667], "iscrowd": 0, "area": 3457.123700000002, "rle": {"size": [480, 640], "counts": "n\\m03k>3M4N2N2O0O2N2O1N1O2N2O_OhA:V>GkA8T>HnA8Q>HPB8n=HTB8j=IWB6i=IYB7e=J\\B6b=J`B6^=KcB4]=LdB4T<DaD8]O4P<FaD7_O4n;GaD5C3k;J_D5GOj;N]D4JLj;2XD5NGj;6UD61Bj;:RD62Al;;oC71_OQ<<kC81[OV<?fC90YO\\<?aC:1VO`<b0\\CK1I0Jd<X1^CnNNFf<]1\\CnNh<R1ZCmNe<S1\\CnNb<S1_CmN`<R1bCnN\\<R1eCnNZ<S1gCmNX<R1jCnNU<Q1lCPOS<n0PDQOQ<m0QDSOn;l0SDUOm;h0VDWOj;h0XDVOi;i0XDVOj;h0XDVOi;i0YDUOh;j0YDVOh;h0ZDVOg;i0YDWOi;g0VDYOl;f0SD[On;d0RD\\OP<c0nC]OT<b0nC\\OS<c0oC[OS<c0V1O2N1O1O2N1O2N1O100O0000O10000000000000O10000O102N3Lj`>0W_A2N2M2O2N000000O1000O10000000000000000O100000000000000O100000000000001N1000000000000000100O10O001O001N101O001O0O2O0001O001O005LO00000O101O1O1M3N2N2Mgfb5"}, "label": "blue lounge chair on left"}]}
{"id": 157981, "image": "COCO_train2014_000000157981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat\"."}], "task": "refering", "answer_template": "The \"the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 32, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 373, 374, 375, 376, 377], "bbox": [0.07903125, 0.08172916666666666, 0.4258125, 0.9763541666666667], "iscrowd": 0, "area": 68173.41324999997, "rle": {"size": [480, 640], "counts": "bXh0:c>;F6I7H7H8H9YOh0A>H8J6I7H9H7I7H:G8H3L4M3L4M3M3L4M3M3L4M3M3L4L4L4L4L301N2N2O1N2O0O2N2O1O1O1N101O1O1O1N2O001O1O1N2O001O1O1O0O2O1O010O0010O01O00100O1O1O1O1N2O1N2M3M3L4M3M3[Oe0M3L4L4M3L4L4L4L4L4L4K5L4K5ZOf0J6L4L4M3L4N2jMjC`1X<]NnC^1T<_NRD]1k<N000O10000000000O1001O001O001N101O1O001O00001O00000O10000000000000000000O010000000000000O1000O1000000000000000000000000000000000000000000000000000000000000001N100O1O100O100O2O000000aJPOnLP1m2WOQMi0i2^OVMb0d2FZM:[NdNm0Y1g03iMYOW1l0m0MdM@X1j0hMZNj1]1mNJX1f0lMYNm1W1fN5Z1b0mMXNP2S1bN;Y1b0PNUNS2o0^Na0X1b0RNSNV2j0]Ng0R1U2>VM\\Nn0j0T2h0oLYNV1e0R2o0iLYN]1=R2V1dLWNc1:P2\\1_LTNP20j1g1R7M3L4M6Jf0YO;E;@]c[5"}, "label": "the cat"}]}
{"id": 157981, "image": "COCO_train2014_000000157981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a siamese cat sitting on a bathroom counter\"."}], "task": "refering", "answer_template": "The \"a siamese cat sitting on a bathroom counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 32, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 373, 374, 375, 376, 377], "bbox": [0.07903125, 0.08172916666666666, 0.4258125, 0.9763541666666667], "iscrowd": 0, "area": 68173.41324999997, "rle": {"size": [480, 640], "counts": "bXh0:c>;F6I7H7H8H9YOh0A>H8J6I7H9H7I7H:G8H3L4M3L4M3M3L4M3M3L4M3M3L4L4L4L4L301N2N2O1N2O0O2N2O1O1O1N101O1O1O1N2O001O1O1N2O001O1O1O0O2O1O010O0010O01O00100O1O1O1O1N2O1N2M3M3L4M3M3[Oe0M3L4L4M3L4L4L4L4L4L4K5L4K5ZOf0J6L4L4M3L4N2jMjC`1X<]NnC^1T<_NRD]1k<N000O10000000000O1001O001O001N101O1O001O00001O00000O10000000000000000000O010000000000000O1000O1000000000000000000000000000000000000000000000000000000000000001N100O1O100O100O2O000000aJPOnLP1m2WOQMi0i2^OVMb0d2FZM:[NdNm0Y1g03iMYOW1l0m0MdM@X1j0hMZNj1]1mNJX1f0lMYNm1W1fN5Z1b0mMXNP2S1bN;Y1b0PNUNS2o0^Na0X1b0RNSNV2j0]Ng0R1U2>VM\\Nn0j0T2h0oLYNV1e0R2o0iLYN]1=R2V1dLWNc1:P2\\1_LTNP20j1g1R7M3L4M6Jf0YO;E;@]c[5"}, "label": "a siamese cat sitting on a bathroom counter"}]}
{"id": 272670, "image": "COCO_train2014_000000272670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear glass on table sitting next to orange juice\"."}], "task": "refering", "answer_template": "The \"clear glass on table sitting next to orange juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 329, 330, 331, 332], "bbox": [0.24093749999999997, 0.27358333333333335, 0.493515625, 0.8616666666666666], "iscrowd": 0, "area": 37145.38185, "rle": {"size": [480, 640], "counts": "X\\X2W2b<;F:E;E;E;F;D;K5L4K5L4L4K6K4L4K5L4L4K6K4K5L4L2M2O2N1N3N1O1N3N1O2M2O2N1N3N1O1N3N1N3N1O2M2O1O1N101O0O2O001N101O0O2O0O2O001N101O1N101O0O20O01O001O001O001O010O001O0000000001N1000000O1000001N1000000O1000001N1000000O1000001N1000000O1000000O2O1O1N2O2N1O1N2O1O2N1N2O1N3L3N2M3M5L4K5L4K6J5L4K5L5J5K5L4K5L5J5L4K5K5L5H7E;E;Eb0]Of0[Oe0[Oe0[Oe0ZOXcg4"}, "label": "clear glass on table sitting next to orange juice"}]}
{"id": 272670, "image": "COCO_train2014_000000272670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass jar to the left of a bottle of orange juice\"."}], "task": "refering", "answer_template": "The \"a glass jar to the left of a bottle of orange juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 329, 330, 331, 332], "bbox": [0.24093749999999997, 0.27358333333333335, 0.493515625, 0.8616666666666666], "iscrowd": 0, "area": 37145.38185, "rle": {"size": [480, 640], "counts": "X\\X2W2b<;F:E;E;E;F;D;K5L4K5L4L4K6K4L4K5L4L4K6K4K5L4L2M2O2N1N3N1O1N3N1O2M2O2N1N3N1O1N3N1N3N1O2M2O1O1N101O0O2O001N101O0O2O0O2O001N101O1N101O0O20O01O001O001O001O010O001O0000000001N1000000O1000001N1000000O1000001N1000000O1000001N1000000O1000000O2O1O1N2O2N1O1N2O1O2N1N2O1N3L3N2M3M5L4K5L4K6J5L4K5L5J5K5L4K5L5J5L4K5K5L5H7E;E;Eb0]Of0[Oe0[Oe0[Oe0ZOXcg4"}, "label": "a glass jar to the left of a bottle of orange juice"}]}
{"id": 272670, "image": "COCO_train2014_000000272670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white cup behind glass\"."}], "task": "refering", "answer_template": "The \"white cup behind glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 140, 141, 142, 143, 164, 165, 166, 188], "bbox": [0.12782812500000001, 0.20729166666666668, 0.423328125, 0.4739583333333333], "iscrowd": 0, "area": 11115.3144, "rle": {"size": [480, 640], "counts": "WcV1l0k=a0B>D6M3N1O1O1N2O1O1O1O1N2O1O1O1O0O100000000O101O00000000000000000000000000000000O0100000O10000000\\OlM\\DS2e;PNXDP2g;e0O100O00100O1O010O1O100O010O100O10O0O2J6J5J7J5K6I6K6J5M4O1O1O001N2O1O001O10O01O100O00100O1O010O1O10O01O100O1000000O100000000O100000000O100000000O10000000000O1000000000001O0O1000000000000O1000001O00001N11O01O00001O01O01O0000010O00001O010O001O1O010O001O10O01O1O010O103M2M4M7Hgj\\5"}, "label": "white cup behind glass"}]}
{"id": 272670, "image": "COCO_train2014_000000272670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup just behind a glass\"."}], "task": "refering", "answer_template": "The \"a cup just behind a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 140, 141, 142, 143, 164, 165, 166, 188], "bbox": [0.12782812500000001, 0.20729166666666668, 0.423328125, 0.4739583333333333], "iscrowd": 0, "area": 11115.3144, "rle": {"size": [480, 640], "counts": "WcV1l0k=a0B>D6M3N1O1O1N2O1O1O1O1N2O1O1O1O0O100000000O101O00000000000000000000000000000000O0100000O10000000\\OlM\\DS2e;PNXDP2g;e0O100O00100O1O010O1O100O010O100O10O0O2J6J5J7J5K6I6K6J5M4O1O1O001N2O1O001O10O01O100O00100O1O010O1O10O01O100O1000000O100000000O100000000O100000000O10000000000O1000000000001O0O1000000000000O1000001O00001N11O01O00001O01O01O0000010O00001O010O001O1O010O001O10O01O1O010O103M2M4M7Hgj\\5"}, "label": "a cup just behind a glass"}]}
{"id": 477471, "image": "COCO_train2014_000000477471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white minature horse pulling a small carriage\"."}], "task": "refering", "answer_template": "The \"a black and white minature horse pulling a small carriage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 313, 314, 315, 334, 336, 338], "bbox": [0.45, 0.37597916666666664, 0.84678125, 0.8577083333333333], "iscrowd": 0, "area": 32746.147200000003, "rle": {"size": [480, 640], "counts": "gZW42l>5K5K6`M9]EL^:8_EM[:7aENZ:6bE0X:4dE2V:2gE6P:m2_O2J6J6K5O1O1O1O1OjNQG]Lo8a3VG\\Li8c3\\GZLc8e3bGXL]8f3iGWLV8h3oGULQ8i3c1O1O20O3M3N2N1O0O2O001O001O00001O001O0000001O00a0_O3M2SFVK`9m4ZFXKd9S5O2N1O2O0O2N2N000000SNfFgL0m0[9\\2eFgL2j0Y9_2eFiL3d0Y9c2eFiL4a0W9f2fFiL5>U9i2gFiL6;S9l2hFiL78Q9o2iFhL87P9P3hFiL;4m8S3iFiL;2l8V3iFgL=0k8Y3iFgL>Mi8\\3jFfLX:Z3hEgLW:Y3jEfLV:Z3jEfLV:Z3kEfLT:Z3lEfLT:[3lEeLS:[3mEeLS:[3nEeLQ:[3oEeLQ:[3PFXL\\:h3>00001O0000000O10001O00000O1000001O0O10000000001N0100ObFTLj7l3THWLk7i3SHYLm7f3QH^Ln7b3PHaLo7^3PHdLP8\\3nGgLQ8Y3lGkLS8T3lGoLS8Q3kGQMU8n2jGUMU8k2iGXMV8h2iGZMV8e2iG]MW8c2gG`MX8_2hG]M\\8d2bGWMe8i2YGXMh8h2WGWMj8l2RGUMo8m2mFSMT9Q3gFoL[9S3`FoL`9W42O1N1O2N1O2M2O2N1O2M2O21N3M2O2M2O2M2O2M2O2M2O2MiNcFnL\\9Q3hFmLV9S3mFlLQ9T3SGkL5oNR8U4lGkLHZOZ8l3PH]Mo7b2UH\\Mi7d2ZH\\Md7c2_H\\M`7c2dH[MZ7e2iH[MU7d2oHZMP7f2TIUMl6k2YIQMg6n2_ImLa6Q3eIjL[6V3jIfLV6X3QJcLo5[3WJaLh5_3]J\\Ld5b3cJYL]5f3hJVLW5i3Q3O1N101N2N10OA?0002N2N2N3M2N2N3M2N2O1N3M21O100N2O2N1O1O1O1N2O2N2N2N2N2N1N3N2N2N2N1O1N101O001O001N101O001O001O00O8H9H7H6J5K6K4K6cNYC6l<B[C:k<]O[C`0j<XO]Cc0c=L5K5K4L]e]1"}, "label": "a black and white minature horse pulling a small carriage"}]}
{"id": 477471, "image": "COCO_train2014_000000477471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white horse pulling a cart with a man in it\"."}], "task": "refering", "answer_template": "The \"a black and white horse pulling a cart with a man in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 313, 314, 315, 334, 336, 338], "bbox": [0.45, 0.37597916666666664, 0.84678125, 0.8577083333333333], "iscrowd": 0, "area": 32746.147200000003, "rle": {"size": [480, 640], "counts": "gZW42l>5K5K6`M9]EL^:8_EM[:7aENZ:6bE0X:4dE2V:2gE6P:m2_O2J6J6K5O1O1O1O1OjNQG]Lo8a3VG\\Li8c3\\GZLc8e3bGXL]8f3iGWLV8h3oGULQ8i3c1O1O20O3M3N2N1O0O2O001O001O00001O001O0000001O00a0_O3M2SFVK`9m4ZFXKd9S5O2N1O2O0O2N2N000000SNfFgL0m0[9\\2eFgL2j0Y9_2eFiL3d0Y9c2eFiL4a0W9f2fFiL5>U9i2gFiL6;S9l2hFiL78Q9o2iFhL87P9P3hFiL;4m8S3iFiL;2l8V3iFgL=0k8Y3iFgL>Mi8\\3jFfLX:Z3hEgLW:Y3jEfLV:Z3jEfLV:Z3kEfLT:Z3lEfLT:[3lEeLS:[3mEeLS:[3nEeLQ:[3oEeLQ:[3PFXL\\:h3>00001O0000000O10001O00000O1000001O0O10000000001N0100ObFTLj7l3THWLk7i3SHYLm7f3QH^Ln7b3PHaLo7^3PHdLP8\\3nGgLQ8Y3lGkLS8T3lGoLS8Q3kGQMU8n2jGUMU8k2iGXMV8h2iGZMV8e2iG]MW8c2gG`MX8_2hG]M\\8d2bGWMe8i2YGXMh8h2WGWMj8l2RGUMo8m2mFSMT9Q3gFoL[9S3`FoL`9W42O1N1O2N1O2M2O2N1O2M2O21N3M2O2M2O2M2O2M2O2M2O2MiNcFnL\\9Q3hFmLV9S3mFlLQ9T3SGkL5oNR8U4lGkLHZOZ8l3PH]Mo7b2UH\\Mi7d2ZH\\Md7c2_H\\M`7c2dH[MZ7e2iH[MU7d2oHZMP7f2TIUMl6k2YIQMg6n2_ImLa6Q3eIjL[6V3jIfLV6X3QJcLo5[3WJaLh5_3]J\\Ld5b3cJYL]5f3hJVLW5i3Q3O1N101N2N10OA?0002N2N2N3M2N2N3M2N2O1N3M21O100N2O2N1O1O1O1N2O2N2N2N2N2N1N3N2N2N2N1O1N101O001O001N101O001O001O00O8H9H7H6J5K6K4K6cNYC6l<B[C:k<]O[C`0j<XO]Cc0c=L5K5K4L]e]1"}, "label": "a black and white horse pulling a cart with a man in it"}]}
{"id": 297251, "image": "COCO_train2014_000000297251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy\"."}], "task": "refering", "answer_template": "The \"boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 211, 212, 213, 234, 235, 236, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328, 350, 351, 373], "bbox": [0.12471874999999999, 0.505625, 0.303375, 0.9820208333333333], "iscrowd": 0, "area": 13171.808649999997, "rle": {"size": [480, 640], "counts": "nkU12l>4M2M4L4M3L3N3L4L3N3L2O1N101N2N101N2O0O2N2O0O2O1O1O001O10O01O1VNYNZFg1d9aNVF`1h9gNRFZ1k9nNoER1P:UOjEl0T:[OfEf0W:^OgEb0X:BeE?Y:DeE=Z:EeE;Z:GeE8[:JdE6\\:KcE5\\:NbE2^:OaE3\\:OcE3[:NdE5X:MhE4U:NjE4T:MkE5R:NlE4R:MmE6o9LPF6n9KQF7l9KSF7k9b2L3N2N2M3N3M2M3N2N3L3eGQJa7Q6ZHSJe7o5VHUJh7n5THUJk7n5oGVJP8^6N2O1O0OO2N2N1O2N2N4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3SNjFlMY9n1P2I7J6J6TOm0J5K5K5L4K5K5K5L5J\\i`6"}, "label": "boy"}]}
{"id": 297251, "image": "COCO_train2014_000000297251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby in a red t - shirt sitting in his mothers back papoose\"."}], "task": "refering", "answer_template": "The \"a baby in a red t - shirt sitting in his mothers back papoose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 211, 212, 213, 234, 235, 236, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328, 350, 351, 373], "bbox": [0.12471874999999999, 0.505625, 0.303375, 0.9820208333333333], "iscrowd": 0, "area": 13171.808649999997, "rle": {"size": [480, 640], "counts": "nkU12l>4M2M4L4M3L3N3L4L3N3L2O1N101N2N101N2O0O2N2O0O2O1O1O001O10O01O1VNYNZFg1d9aNVF`1h9gNRFZ1k9nNoER1P:UOjEl0T:[OfEf0W:^OgEb0X:BeE?Y:DeE=Z:EeE;Z:GeE8[:JdE6\\:KcE5\\:NbE2^:OaE3\\:OcE3[:NdE5X:MhE4U:NjE4T:MkE5R:NlE4R:MmE6o9LPF6n9KQF7l9KSF7k9b2L3N2N2M3N3M2M3N2N3L3eGQJa7Q6ZHSJe7o5VHUJh7n5THUJk7n5oGVJP8^6N2O1O0OO2N2N1O2N2N4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M4L3SNjFlMY9n1P2I7J6J6TOm0J5K5K5L4K5K5K5L5J\\i`6"}, "label": "a baby in a red t - shirt sitting in his mothers back papoose"}]}
{"id": 297251, "image": "COCO_train2014_000000297251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with sunglasses and a black shirt carrying kid on her back at zoo\"."}], "task": "refering", "answer_template": "The \"woman with sunglasses and a black shirt carrying kid on her back at zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 357, 358, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.24362499999999998, 0.3794375, 0.5992031250000001, 1.0], "iscrowd": 0, "area": 38605.598300000005, "rle": {"size": [480, 640], "counts": "e_Y23k>6K5J6J6K3L3M2N3M3L3N3L4M2N2^DfN[9]1YFVO^9k0WFGa9;TF0k91iE;V:X2N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2O1N2N2eNdJQI]5f6QKSIQ5g6YKTIh4g6aKSIa4h6hKRIZ4h6R2K5J6O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O11O001O00001O001O000000O10000O10000O11O001O1O001O001O1O001O1O001O1O001O001O1O2N1O2N1O2N2N1O2N1O2N2N2N4L5K4L5K4L4[J_Hc3f7RLbHl3b7jKfHT4_7aKjH\\4Z7ZKnHd4V7SKQIk4d8N1SLVFb2k9VM^Fh2o:N2N2N2N2N2N2N2N2N2N2N2N2hMoC]1S<ZNUDe1m;QN[Do1\\<O1O001O1O001O1O001O1O0000001O000000001O00000000001O00000000001O001O001O001O001O001O001O001O001]OmBWOS=i0c0N1O101N1O101J5HcQh3"}, "label": "woman with sunglasses and a black shirt carrying kid on her back at zoo"}]}
{"id": 297251, "image": "COCO_train2014_000000297251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown haired woman wearing a black shirt and an infant carrier with a baby in it standing in front of giraffes\"."}], "task": "refering", "answer_template": "The \"a brown haired woman wearing a black shirt and an infant carrier with a baby in it standing in front of giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 357, 358, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.24362499999999998, 0.3794375, 0.5992031250000001, 1.0], "iscrowd": 0, "area": 38605.598300000005, "rle": {"size": [480, 640], "counts": "e_Y23k>6K5J6J6K3L3M2N3M3L3N3L4M2N2^DfN[9]1YFVO^9k0WFGa9;TF0k91iE;V:X2N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2O1N2N2eNdJQI]5f6QKSIQ5g6YKTIh4g6aKSIa4h6hKRIZ4h6R2K5J6O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O11O001O00001O001O000000O10000O10000O11O001O1O001O001O1O001O1O001O1O001O001O1O2N1O2N1O2N2N1O2N1O2N2N2N4L5K4L5K4L4[J_Hc3f7RLbHl3b7jKfHT4_7aKjH\\4Z7ZKnHd4V7SKQIk4d8N1SLVFb2k9VM^Fh2o:N2N2N2N2N2N2N2N2N2N2N2N2hMoC]1S<ZNUDe1m;QN[Do1\\<O1O001O1O001O1O001O1O0000001O000000001O00000000001O00000000001O001O001O001O001O001O001O001O001]OmBWOS=i0c0N1O101N1O101J5HcQh3"}, "label": "a brown haired woman wearing a black shirt and an infant carrier with a baby in it standing in front of giraffes"}]}
{"id": 2342, "image": "COCO_train2014_000000002342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe stands next to another giraffe and puts his head over a railing\"."}], "task": "refering", "answer_template": "The \"a giraffe stands next to another giraffe and puts his head over a railing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 147, 166, 167, 168, 190, 191, 213, 214, 215, 236, 237, 259, 260, 282, 283, 305, 306, 328, 329, 330, 351, 352, 353, 354, 373, 374, 375, 376, 377, 378], "bbox": [0.167609375, 0.26129166666666664, 0.467265625, 0.9881458333333334], "iscrowd": 0, "area": 21408.82119999999, "rle": {"size": [480, 640], "counts": "cZb16i>1O1O2M2O1O1O1O1O1N2O2N10000O10000O101O0O2O001N101O0O2O001N010000O2O00000O2O000O1O2N100O1O2N1O1O02N5Jd0\\O6J7J6I3lJoMUMT2g2oMWMU2e2mMYMV2c2mMWM[2d2hMTMa2i2aMRMf2i2]MWMf2c2]M]Mg2\\2\\MdMi2U2YMkMl2m1WMRNQ3d1RM\\NT8\\LnGd3X8SLkGm3Z8kKiGU4a9O12N101N2N2N1O2N2N2N000000001OO100O100O100O1K5K5001O1O1O1O2N3M3M3ZJ^Kh0d4VOiK=Z4@RL4Q4I\\LJg33eLA]3=kL[OX3SNaKO_1g1Q3WNmKGX1l1l2ZNRLD[1k1d2^NULD_1g1]2aNYLE`1e1W2dN]LDa1d1R2fNaLCb1c1m1iNdLAb1d1j1kNfL^Ob1f1h1kNkLZO`1i1e1kNR1T1nNkNT1T1lNjNX1T1hNjN[1U1eNiN_1U1aNjNa1U1_NlNa1S1_NmNc1Q1]NoNd1P1\\NPOf1n0ZNQOh1n0XNROi1m0WNSOk1k0UNUOl1j0TNUOo1i0QNWOP2h0PNXOQ2g0PNWOS2h0lMXOU2g0lMXOV2f0jMZOW2e0jMYOX2f0hMZOZ2d0fM[O\\2d0eMYO_2e0aMZO`2f0aMXOa2g0_MXOb2h0_MVOc2i0^MTOe2k0\\MSOe2m0\\MQOf2n0[MPOf2P1f5O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O002N3M2N3M2NW\\o4"}, "label": "a giraffe stands next to another giraffe and puts his head over a railing"}]}
{"id": 2342, "image": "COCO_train2014_000000002342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe to the left of the other giraffe\"."}], "task": "refering", "answer_template": "The \"the giraffe to the left of the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 147, 166, 167, 168, 190, 191, 213, 214, 215, 236, 237, 259, 260, 282, 283, 305, 306, 328, 329, 330, 351, 352, 353, 354, 373, 374, 375, 376, 377, 378], "bbox": [0.167609375, 0.26129166666666664, 0.467265625, 0.9881458333333334], "iscrowd": 0, "area": 21408.82119999999, "rle": {"size": [480, 640], "counts": "cZb16i>1O1O2M2O1O1O1O1O1N2O2N10000O10000O101O0O2O001N101O0O2O001N010000O2O00000O2O000O1O2N100O1O2N1O1O02N5Jd0\\O6J7J6I3lJoMUMT2g2oMWMU2e2mMYMV2c2mMWM[2d2hMTMa2i2aMRMf2i2]MWMf2c2]M]Mg2\\2\\MdMi2U2YMkMl2m1WMRNQ3d1RM\\NT8\\LnGd3X8SLkGm3Z8kKiGU4a9O12N101N2N2N1O2N2N2N000000001OO100O100O100O1K5K5001O1O1O1O2N3M3M3ZJ^Kh0d4VOiK=Z4@RL4Q4I\\LJg33eLA]3=kL[OX3SNaKO_1g1Q3WNmKGX1l1l2ZNRLD[1k1d2^NULD_1g1]2aNYLE`1e1W2dN]LDa1d1R2fNaLCb1c1m1iNdLAb1d1j1kNfL^Ob1f1h1kNkLZO`1i1e1kNR1T1nNkNT1T1lNjNX1T1hNjN[1U1eNiN_1U1aNjNa1U1_NlNa1S1_NmNc1Q1]NoNd1P1\\NPOf1n0ZNQOh1n0XNROi1m0WNSOk1k0UNUOl1j0TNUOo1i0QNWOP2h0PNXOQ2g0PNWOS2h0lMXOU2g0lMXOV2f0jMZOW2e0jMYOX2f0hMZOZ2d0fM[O\\2d0eMYO_2e0aMZO`2f0aMXOa2g0_MXOb2h0_MVOc2i0^MTOe2k0\\MSOe2m0\\MQOf2n0[MPOf2P1f5O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O002N3M2N3M2NW\\o4"}, "label": "the giraffe to the left of the other giraffe"}]}
{"id": 2342, "image": "COCO_train2014_000000002342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is the giraffe on the right who is looking towards the camera\"."}], "task": "refering", "answer_template": "The \"this is the giraffe on the right who is looking towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 270, 271, 293, 294, 295, 316, 317, 318, 339, 340, 341, 342, 363, 364, 365, 366, 386, 387, 388, 389, 390], "bbox": [0.657296875, 0.39775, 1.0, 0.9842708333333332], "iscrowd": 0, "area": 18458.995099999996, "rle": {"size": [480, 640], "counts": "XbU61o>2N2N1O2N2N2N1O2N2O0O1bBAY<?eCC[<=cCE]<<`CG_<9^CJb<7[CKe<5YCMg<3WCOi<2SC1m<OQC3n<NPC4m<0PC3l<0RC2k<S1M3M3ZOf0000001N10000000000O101O000000000O10001O02N2N2N2N2N23M2N3L4K5J5L5J6K5K4K6K5J6J5K6J6I7J6J7I7I7IhJZG^4_8bKbGd4W8\\KjGj4o7VKRHP5g7PKZHV5a7hJ_H6NR4_7eKdH81V4V7aKiH93Z4P7ZKnH;6^4g6UKTI=8a4_6PKZI?9e4Y6iJ^Ib0=d4T6hJ`Id0>b4T6gJ_If0a0`4Q6hJ_Ig0d0^4n5iJ_Ih0f0]4m5hJ]Ik0j0Z4j5iJ]Im0l0W4W6hKlIV4T6iKPJT4P6kKTJR4l5mKWJQ4h5oK\\Jo3d5PL^Jn3c5PL`Jn3a5PLbJn3^5QLeJm3\\5QLgJm3Z5RLiJk3X5RLlJl3U5RLnJl3S5RLPKl3P5SLSKk3n4RLVKl3k4RLXKl3i4RLZKl3o7N2O1N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2O1N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2O1N2N1O1O1O001O1O1O1O1O001O1O100O001O1O1O1O1O001O1O1O1O001O1O100O1O001O1O1O1O001O1O1O1O1O00100O1O1O001O1O1O1O1O001O1O1O1O00100O1O1O1O001O1O1O1O001O1O9"}, "label": "this is the giraffe on the right who is looking towards the camera"}]}
{"id": 2342, "image": "COCO_train2014_000000002342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the right\"."}], "task": "refering", "answer_template": "The \"the giraffe on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 270, 271, 293, 294, 295, 316, 317, 318, 339, 340, 341, 342, 363, 364, 365, 366, 386, 387, 388, 389, 390], "bbox": [0.657296875, 0.39775, 1.0, 0.9842708333333332], "iscrowd": 0, "area": 18458.995099999996, "rle": {"size": [480, 640], "counts": "XbU61o>2N2N1O2N2N2N1O2N2O0O1bBAY<?eCC[<=cCE]<<`CG_<9^CJb<7[CKe<5YCMg<3WCOi<2SC1m<OQC3n<NPC4m<0PC3l<0RC2k<S1M3M3ZOf0000001N10000000000O101O000000000O10001O02N2N2N2N2N23M2N3L4K5J5L5J6K5K4K6K5J6J5K6J6I7J6J7I7I7IhJZG^4_8bKbGd4W8\\KjGj4o7VKRHP5g7PKZHV5a7hJ_H6NR4_7eKdH81V4V7aKiH93Z4P7ZKnH;6^4g6UKTI=8a4_6PKZI?9e4Y6iJ^Ib0=d4T6hJ`Id0>b4T6gJ_If0a0`4Q6hJ_Ig0d0^4n5iJ_Ih0f0]4m5hJ]Ik0j0Z4j5iJ]Im0l0W4W6hKlIV4T6iKPJT4P6kKTJR4l5mKWJQ4h5oK\\Jo3d5PL^Jn3c5PL`Jn3a5PLbJn3^5QLeJm3\\5QLgJm3Z5RLiJk3X5RLlJl3U5RLnJl3S5RLPKl3P5SLSKk3n4RLVKl3k4RLXKl3i4RLZKl3o7N2O1N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2O1N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2O1N2N1O1O1O001O1O1O1O1O001O1O100O001O1O1O1O1O001O1O1O1O001O1O100O1O001O1O1O1O001O1O1O1O1O00100O1O1O001O1O1O1O1O001O1O1O1O00100O1O1O1O001O1O1O1O001O1O9"}, "label": "the giraffe on the right"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue pants and white shirt holding surfboard\"."}], "task": "refering", "answer_template": "The \"woman in blue pants and white shirt holding surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 81, 82, 99, 100, 101, 117, 118, 119, 135, 136, 153, 154, 171, 172, 189, 190, 207, 208], "bbox": [0.49888, 0.2472, 0.6269600000000001, 0.9078666666666666], "iscrowd": 0, "area": 8913.532550000002, "rle": {"size": [375, 500], "counts": "_\\k2R1b:<D=QOZNVGR2`8XNTGU2a8m0bMfLfKg3k3_LRLe3i3^LULe3g3aLTLb3g3eLTL\\3j3iLSLV3l3oLPLR3n3_MbKa2]4bMaK^2^4fM_KZ2a4iM\\KW2c4lM[KT2d4nM[KR2=hK`3X2RLP2<jKa3X2RLo1:kKc3Y2PLm1;mKc3X2QLk1;nKc3Y2QLk18nKf3Y2QLj15PLj3g6RL[In3W700000SOTLoIl3n5[LnIe3o5bLmI^3P6fLoIZ3n5jLQJV3l5mLSJT3j5PMTJQ3g5TMYJl2`5\\M^Jf2[5`MdJa2U5gMiJ[2P5kMPKU2i4SNVKn1c4YN[Kh1_4^NaKc1g0lK_2i2hL\\1e0QLS12POg2XOW1a0ULe0k0mN`68jHk0g8nNaGQ1d9O100O001O1O1O1O100M4M4K5L4K7ImYT2"}, "label": "woman in blue pants and white shirt holding surfboard"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a white shirt and blue pants\"."}], "task": "refering", "answer_template": "The \"woman in a white shirt and blue pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 81, 82, 99, 100, 101, 117, 118, 119, 135, 136, 153, 154, 171, 172, 189, 190, 207, 208], "bbox": [0.49888, 0.2472, 0.6269600000000001, 0.9078666666666666], "iscrowd": 0, "area": 8913.532550000002, "rle": {"size": [375, 500], "counts": "_\\k2R1b:<D=QOZNVGR2`8XNTGU2a8m0bMfLfKg3k3_LRLe3i3^LULe3g3aLTLb3g3eLTL\\3j3iLSLV3l3oLPLR3n3_MbKa2]4bMaK^2^4fM_KZ2a4iM\\KW2c4lM[KT2d4nM[KR2=hK`3X2RLP2<jKa3X2RLo1:kKc3Y2PLm1;mKc3X2QLk1;nKc3Y2QLk18nKf3Y2QLj15PLj3g6RL[In3W700000SOTLoIl3n5[LnIe3o5bLmI^3P6fLoIZ3n5jLQJV3l5mLSJT3j5PMTJQ3g5TMYJl2`5\\M^Jf2[5`MdJa2U5gMiJ[2P5kMPKU2i4SNVKn1c4YN[Kh1_4^NaKc1g0lK_2i2hL\\1e0QLS12POg2XOW1a0ULe0k0mN`68jHk0g8nNaGQ1d9O100O001O1O1O1O100M4M4K5L4K7ImYT2"}, "label": "woman in a white shirt and blue pants"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt , dark shorts with a red stripe and a straw hat at a party\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt , dark shorts with a red stripe and a straw hat at a party\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 96, 97, 98, 99, 112, 113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 148, 149, 150, 151, 166, 167, 168, 169, 170, 185, 186, 187, 188, 203, 204, 205, 206, 221, 222], "bbox": [0.22986, 0.34696000000000005, 0.5453, 0.9868], "iscrowd": 0, "area": 18986.983799999998, "rle": {"size": [375, 500], "counts": "][Z16V;=J5L5K4M4K4L5K4L5K5K4L5L3L5K4L5M2N3M3N1N3M2N3N1N3M2iIiLl3Y3fKSMY4Q3XK[Mg4j2iJbMV5b2[JjMe5U4N1O100O1O101N1O1O01M2O1N3M2O1N3M2N2O1N3M2N2ObJkJ^4R5aKRK^4l4bKWK]4h4aK\\K_4a4aKbK^4^4_KfK`4Z4]KjKb4V4\\KmKd4R4YKRLf4n3XKULg4k3VKYLj4f3TK]Ll4c3PKaLP5^3mJfLS5Y3kJjLU5l4000\\LkJf0U5ZOmJd0S5\\OoJa0R5_OoJ>S5BoJ;R5FoJ6S5JnJ3T5MnJOU50mJLU54lJIV57lJEV5;lJAV5`0jJ]OX5f0gJVO[5m0dJoN^5T1aJhNa5[1]JbNe5a1ZJZNi5i1VJSNl5P2RJmMP6V2oIfMS6]2lI_MV6d2hIYM[6i2dISMj5QOcJo3BlLm5YO]Jn3DfLQ6@WJm3I]LR6IRJm3b6TL]Io3`6RL_IQ4^6PLaIS4\\6nKcIU4Z6lKeIl15Jo5i0WJZOa5f0dJ\\OU5c0QKZOm4e0YKUOh4k0^KoNb4P1dKjN]4U1iKeNW4m0]LnNc3l0fLPOZ3m0lLPOT3l0TMQOl2k0ZMROf2k0aMQO`2m0dMQO[2P1hMlNY2S1jMjNV2W1mMfNR2Z1QNcNP2]1SN^Nn1b1UNYNn1g1P401O010O100O100O100O1O010M3L4M3M310O1O101N1[OeEL\\:GRF8g:O101N1O2O0O2N101N10[Zc2"}, "label": "a man in a blue shirt , dark shorts with a red stripe and a straw hat at a party"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a light blue shirt , straw hat with a child ' s float around his waist\"."}], "task": "refering", "answer_template": "The \"a man in a light blue shirt , straw hat with a child ' s float around his waist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 96, 97, 98, 99, 112, 113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 148, 149, 150, 151, 166, 167, 168, 169, 170, 185, 186, 187, 188, 203, 204, 205, 206, 221, 222], "bbox": [0.22986, 0.34696000000000005, 0.5453, 0.9868], "iscrowd": 0, "area": 18986.983799999998, "rle": {"size": [375, 500], "counts": "][Z16V;=J5L5K4M4K4L5K4L5K5K4L5L3L5K4L5M2N3M3N1N3M2N3N1N3M2iIiLl3Y3fKSMY4Q3XK[Mg4j2iJbMV5b2[JjMe5U4N1O100O1O101N1O1O01M2O1N3M2O1N3M2N2O1N3M2N2ObJkJ^4R5aKRK^4l4bKWK]4h4aK\\K_4a4aKbK^4^4_KfK`4Z4]KjKb4V4\\KmKd4R4YKRLf4n3XKULg4k3VKYLj4f3TK]Ll4c3PKaLP5^3mJfLS5Y3kJjLU5l4000\\LkJf0U5ZOmJd0S5\\OoJa0R5_OoJ>S5BoJ;R5FoJ6S5JnJ3T5MnJOU50mJLU54lJIV57lJEV5;lJAV5`0jJ]OX5f0gJVO[5m0dJoN^5T1aJhNa5[1]JbNe5a1ZJZNi5i1VJSNl5P2RJmMP6V2oIfMS6]2lI_MV6d2hIYM[6i2dISMj5QOcJo3BlLm5YO]Jn3DfLQ6@WJm3I]LR6IRJm3b6TL]Io3`6RL_IQ4^6PLaIS4\\6nKcIU4Z6lKeIl15Jo5i0WJZOa5f0dJ\\OU5c0QKZOm4e0YKUOh4k0^KoNb4P1dKjN]4U1iKeNW4m0]LnNc3l0fLPOZ3m0lLPOT3l0TMQOl2k0ZMROf2k0aMQO`2m0dMQO[2P1hMlNY2S1jMjNV2W1mMfNR2Z1QNcNP2]1SN^Nn1b1UNYNn1g1P401O010O100O100O100O1O010M3L4M3M310O1O101N1[OeEL\\:GRF8g:O101N1O2O0O2N101N10[Zc2"}, "label": "a man in a light blue shirt , straw hat with a child ' s float around his waist"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfboard the woman in a white shirt and blue capris is holding\"."}], "task": "refering", "answer_template": "The \"the surfboard the woman in a white shirt and blue capris is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 45, 46, 63, 64, 154, 170, 171, 172, 188, 190], "bbox": [0.46702, 0.09085333333333333, 0.58874, 0.8487466666666666], "iscrowd": 0, "area": 2082.0268499999997, "rle": {"size": [375, 500], "counts": "one2:Z;5L3M2N3N1N3M2OLXE[Of:d07O010O1O010000003BTELQ;Mib06S]O<F9J7I2N6M4K4N10002N200O1O11O2N2M3N2N3M3M_LiNeLT1X3POiLn0S3WOQMe0k2^OYM?d2D`M7`2IdM3\\2LkMNT21TNIl15\\NEX1d0k40D;O3L3N3N05KeV[2"}, "label": "the surfboard the woman in a white shirt and blue capris is holding"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white surfboard that the women in light purple is holding\"."}], "task": "refering", "answer_template": "The \"the white surfboard that the women in light purple is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 33, 34, 35, 51, 52, 53, 69, 87, 88, 105, 106], "bbox": [0.8429800000000001, 0.0, 1.0, 0.42117333333333334], "iscrowd": 0, "area": 6001.019549999999, "rle": {"size": [375, 500], "counts": "WYj44^;W1jN8G4L5K5K4L5L4J5L5J5M4M3N1N3M3N1N3NeNkGXOR8e0VHXOi7e0_HXO^7h0gHVOW7k0mHQOQ7P1UIlNi6T1]IhN`6Y1gIaNX6_1oI\\Nn5f1XJTNg5m1_JnM^5T2Q22M2O1N2001000000O1000001O0O1000000000000000000000000000010O000000011N2O001O1O1O1N1O2N2O1N2N2N1F^O"}, "label": "the white surfboard that the women in light purple is holding"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman grabing her hat\"."}], "task": "refering", "answer_template": "The \"a woman grabing her hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 58, 59, 60, 76, 77, 78, 94, 95, 96, 112, 130], "bbox": [0.2196, 0.20946666666666666, 0.37668, 0.5968533333333333], "iscrowd": 0, "area": 5288.393600000001, "rle": {"size": [375, 500], "counts": "S^X1=W;b0^Oa0^O9H7J5K4L4L4L4L2N01O1`NRG7o8ISG4n8KTG3m8NTGOm82TGKm86TGHl89TGEm8<TGAl8a0UG\\Ol8e0TGYOl7[OiH]1\\OUOk7AfH\\1@POm7E`H\\1CmNo7IZH\\1HfNCI\\87VH[1LcNBK^87QH]1O`NBL_89mG\\13\\NBO_8:iG]17XNA1a8;dG]1=WNP8=aG]1>VNS8\\2lGdMU8\\2jGdMW8\\2b0OWGeMW8[2iGeMW8\\2hGdMW8]2a0000O10100O1O2O00001O0000001O000000001O01O1N3N1N4I7J6H8H:G4K5K6K4J6H8O1O2N1ORUb3"}, "label": "a woman grabing her hat"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a hat and a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a hat and a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 58, 59, 60, 76, 77, 78, 94, 95, 96, 112, 130], "bbox": [0.2196, 0.20946666666666666, 0.37668, 0.5968533333333333], "iscrowd": 0, "area": 5288.393600000001, "rle": {"size": [375, 500], "counts": "S^X1=W;b0^Oa0^O9H7J5K4L4L4L4L2N01O1`NRG7o8ISG4n8KTG3m8NTGOm82TGKm86TGHl89TGEm8<TGAl8a0UG\\Ol8e0TGYOl7[OiH]1\\OUOk7AfH\\1@POm7E`H\\1CmNo7IZH\\1HfNCI\\87VH[1LcNBK^87QH]1O`NBL_89mG\\13\\NBO_8:iG]17XNA1a8;dG]1=WNP8=aG]1>VNS8\\2lGdMU8\\2jGdMW8\\2b0OWGeMW8[2iGeMW8\\2hGdMW8]2a0000O10100O1O2O00001O0000001O000000001O01O1N3N1N4I7J6H8H:G4K5K6K4J6H8O1O2N1ORUb3"}, "label": "a woman in a hat and a white shirt"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with a maroon t - shirt showing her back\"."}], "task": "refering", "answer_template": "The \"a lady with a maroon t - shirt showing her back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 69, 70, 71, 88, 89, 105, 106, 107, 122, 123, 124, 125, 140, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197, 213, 214, 215, 231, 232, 233], "bbox": [0.8143400000000001, 0.19808, 1.0, 0.9881599999999998], "iscrowd": 0, "area": 21418.041050000003, "rle": {"size": [375, 500], "counts": "bWe4>V;5K4K5L4L4K5L4M3N2N2M3lH_No3c1fKoNQ4X1_KZOX4o0VKCa4a0SK0c44RK8j4KkJa0Q5BdJj0X5o2K4K5L1O1O10000000000O1000000jNYI\\Mg6]2cI_M^6^2hI_MX6`2mI\\MS6c2QJZMo5d2VJYMj5f2[JVMe5h2`JTMa5k2f1O1N2O1N2O1O1N2O1O100O10000O10M2N3L4L3N3L4M2j0WO000O1001O001O001O001O001O001O001O001O001O001O00010O001OZM"}, "label": "a lady with a maroon t - shirt showing her back"}]}
{"id": 338214, "image": "COCO_train2014_000000338214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old lady with grey hair , in dark maroon clothing\"."}], "task": "refering", "answer_template": "The \"an old lady with grey hair , in dark maroon clothing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 69, 70, 71, 88, 89, 105, 106, 107, 122, 123, 124, 125, 140, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197, 213, 214, 215, 231, 232, 233], "bbox": [0.8143400000000001, 0.19808, 1.0, 0.9881599999999998], "iscrowd": 0, "area": 21418.041050000003, "rle": {"size": [375, 500], "counts": "bWe4>V;5K4K5L4L4K5L4M3N2N2M3lH_No3c1fKoNQ4X1_KZOX4o0VKCa4a0SK0c44RK8j4KkJa0Q5BdJj0X5o2K4K5L1O1O10000000000O1000000jNYI\\Mg6]2cI_M^6^2hI_MX6`2mI\\MS6c2QJZMo5d2VJYMj5f2[JVMe5h2`JTMa5k2f1O1N2O1N2O1O1N2O1O100O10000O10M2N3L4L3N3L4M2j0WO000O1001O001O001O001O001O001O001O001O001O001O00010O001OZM"}, "label": "an old lady with grey hair , in dark maroon clothing"}]}
{"id": 280873, "image": "COCO_train2014_000000280873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"keyboard closet to laptop\"."}], "task": "refering", "answer_template": "The \"keyboard closet to laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203], "bbox": [0.184421875, 0.348375, 0.8491406250000001, 0.5285416666666667], "iscrowd": 0, "area": 22131.939899999994, "rle": {"size": [480, 640], "counts": "c`g14k>9G9G9G9G8H10O01O1O1O001O1O001O1O001O10O01O10O10000O10000O10O10O10000O100000000000O1000000000000000000O1000O10000000000000O10000000O10000O10000O0100000O10000O10000O0100000O10000O100O100O01000O100N2N2O1N2N2O0100000000000000000000O10000000O100O1O1O1O010O1O1O100O1O10000O01000O1000000O10000O10O1000O10000O10000O1000000O01000O10000O100000000000000001N1000000000001O00000000001O000O1000001O001O001O1O001O001O001N2O001O001O001O1O001O001O1O001N101O001O1O001O001O001O1O001O001N101O5K6JO10000O100O100O10000O010O100O10000O1000000000000000O10000000001O00000000000000000O100000000000000000000000000000000000000000000000000000000001O000000000001O000000000000000000000001O0000000000000000O100000000000000000000000000O1000000000000000001O0000000O100000000000000000000000000O1000000000000000000000000000000000000000000O2O2N1O2N1O2N000004L4L5K4L4L5K4L5K4L4L5K4L4L5ERX]1"}, "label": "keyboard closet to laptop"}]}
{"id": 280873, "image": "COCO_train2014_000000280873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer keyboard in between another keyboard and a laptop\"."}], "task": "refering", "answer_template": "The \"a computer keyboard in between another keyboard and a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203], "bbox": [0.184421875, 0.348375, 0.8491406250000001, 0.5285416666666667], "iscrowd": 0, "area": 22131.939899999994, "rle": {"size": [480, 640], "counts": "c`g14k>9G9G9G9G8H10O01O1O1O001O1O001O1O001O10O01O10O10000O10000O10O10O10000O100000000000O1000000000000000000O1000O10000000000000O10000000O10000O10000O0100000O10000O10000O0100000O10000O100O100O01000O100N2N2O1N2N2O0100000000000000000000O10000000O100O1O1O1O010O1O1O100O1O10000O01000O1000000O10000O10O1000O10000O10000O1000000O01000O10000O100000000000000001N1000000000001O00000000001O000O1000001O001O001O1O001O001O001N2O001O001O001O1O001O001O1O001N101O001O1O001O001O001O1O001O001N101O5K6JO10000O100O100O10000O010O100O10000O1000000000000000O10000000001O00000000000000000O100000000000000000000000000000000000000000000000000000000001O000000000001O000000000000000000000001O0000000000000000O100000000000000000000000000O1000000000000000001O0000000O100000000000000000000000000O1000000000000000000000000000000000000000000O2O2N1O2N1O2N000004L4L5K4L4L5K4L5K4L4L5K4L4L5ERX]1"}, "label": "a computer keyboard in between another keyboard and a laptop"}]}
{"id": 280873, "image": "COCO_train2014_000000280873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby reaching for mouse and keyboard\"."}], "task": "refering", "answer_template": "The \"baby reaching for mouse and keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 237, 238, 239, 240, 253, 254, 255, 256, 257, 260, 261, 262, 263, 276, 277, 278, 279, 280, 283, 284, 285, 299, 300, 301, 302, 303, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0033750000000000004, 0.27347916666666666, 0.46854687500000003, 0.9790833333333334], "iscrowd": 0, "area": 62724.4034, "rle": {"size": [480, 640], "counts": "X[1^1h8j4eK[41O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N>B?A?A4L00O100OcMeGWN[8g1hGXNW8f1lGZNT8c1oG]NQ8a1QH_Nn7_1UHaNk7\\1XHdNg7Y1]HgNc7_OmGDe0m0]7_OXHVO`0[1X7^OaHjN<h1R7^OkH]N9T2l6^O^Jb0a5^O`Jb0`5]OaJc0^5]OcJc0]5]OcJc0\\5]OeJc0[5\\OfJd0Y5\\OhJd0X5[OjJd0U5\\OlJd0T5\\OlJd0S5\\OnJd0R5[OoJe0P5[OQKe0o4[OQKe0o4ZORKf0n4YOSKg0m4XOTKh0l4WOUKi0j4XOVKh0j4WOWKi0i4VOXKj0h4UOYKk0g4UOYKk0f4UO[Kk0e4TO\\Kl0d4SO^Kl0b4TO^Kl0b4SO`Kl0_4TObKl0^4SOdKl0\\4TOeKk0[4TOfKl0Z4SOhKl0W4TOjKl0V4TOkKk0U4TOlKl0T4SOmKm0S4ROnKn0Q4ROPLn0P4ROPLn0P4QOQLo0o3PORLP1n3oNSLQ1l3oNULQ1k3oNULQ1k3nNVLR1j3mNWLS1h3mNYLS1g3lNZLT1f3lNZLT1f3kN[LU1d3lN]LS1c3mN]LS1c3mN]LS1b3nN^LR1b3mN`LR1`3nN`LR1`3nN`LR1_3oNbLP1^3PObLP1^3PObLP1^3oNdLP1[3QOeLo0[3QOeLo0[3QOfLn0Z3ROfLn0Y3SOgLm0Y3ROlKj1T4VNhKn1X4RNdKR2[4oMbKT2^4lMdIoNb0Y3j5hMdI]O1n2[6dMeIAIo2a6aMgIACQ3f6^MgIC_OQ3j6\\MgIFZOP3o6ZMgIHVOP3R7YMiIIPOP3W7VMjILkNP3Z7UMkIMgNP3^7SMkIo3U6QLjIP4U6PLlIP4T6]KbH2Y1b4U6ZKcH4X1c4T6XKfH4W1d4S6VKhH6T1e4S6TKkH6R1g4S6PKnH9o0g4S6nJPI:m0i4R6lJSI;j0j4S6iJUI<i0k4Q6hJXI<f0n4f6RKVIR5j6mJSIW5m6iJQIY5n6gJSIY5m6gJSIY5m6fJTIZ5l6eJUI[5k6eJUI[5k6dJVI\\5j6cJWI]5i6bJWI_5i6_JYIa5g6^JZIb5f6]J[Ic5e6[J]Ie5b6[J_Ie5n70000000000000000000000000000000000000001O00001PLfFX2Z9fMjFX2V9gMnFV2R9hMRGV2o8hMTGV2l8hMXGV2h8iM[GU2e8iM`GT2a8jMbGT2^8jMfGT2[8jMhGT2X8jMlGT2U8jMnGT2S8iMPHV2P8iMRHV2o7gMTHX2m7fMVHX2j7fMYHY2h7eMZHZ2g7dM\\HW2g7gM\\Ho1n7oMTHg1V8VNmG_1]8`NfGU1d8hN_Gn0k8POWGe0h;F;ESPo4"}, "label": "baby reaching for mouse and keyboard"}]}
{"id": 280873, "image": "COCO_train2014_000000280873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby and her hands touching mouse and keyboard\"."}], "task": "refering", "answer_template": "The \"baby and her hands touching mouse and keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 237, 238, 239, 240, 253, 254, 255, 256, 257, 260, 261, 262, 263, 276, 277, 278, 279, 280, 283, 284, 285, 299, 300, 301, 302, 303, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0033750000000000004, 0.27347916666666666, 0.46854687500000003, 0.9790833333333334], "iscrowd": 0, "area": 62724.4034, "rle": {"size": [480, 640], "counts": "X[1^1h8j4eK[41O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N>B?A?A4L00O100OcMeGWN[8g1hGXNW8f1lGZNT8c1oG]NQ8a1QH_Nn7_1UHaNk7\\1XHdNg7Y1]HgNc7_OmGDe0m0]7_OXHVO`0[1X7^OaHjN<h1R7^OkH]N9T2l6^O^Jb0a5^O`Jb0`5]OaJc0^5]OcJc0]5]OcJc0\\5]OeJc0[5\\OfJd0Y5\\OhJd0X5[OjJd0U5\\OlJd0T5\\OlJd0S5\\OnJd0R5[OoJe0P5[OQKe0o4[OQKe0o4ZORKf0n4YOSKg0m4XOTKh0l4WOUKi0j4XOVKh0j4WOWKi0i4VOXKj0h4UOYKk0g4UOYKk0f4UO[Kk0e4TO\\Kl0d4SO^Kl0b4TO^Kl0b4SO`Kl0_4TObKl0^4SOdKl0\\4TOeKk0[4TOfKl0Z4SOhKl0W4TOjKl0V4TOkKk0U4TOlKl0T4SOmKm0S4ROnKn0Q4ROPLn0P4ROPLn0P4QOQLo0o3PORLP1n3oNSLQ1l3oNULQ1k3oNULQ1k3nNVLR1j3mNWLS1h3mNYLS1g3lNZLT1f3lNZLT1f3kN[LU1d3lN]LS1c3mN]LS1c3mN]LS1b3nN^LR1b3mN`LR1`3nN`LR1`3nN`LR1_3oNbLP1^3PObLP1^3PObLP1^3oNdLP1[3QOeLo0[3QOeLo0[3QOfLn0Z3ROfLn0Y3SOgLm0Y3ROlKj1T4VNhKn1X4RNdKR2[4oMbKT2^4lMdIoNb0Y3j5hMdI]O1n2[6dMeIAIo2a6aMgIACQ3f6^MgIC_OQ3j6\\MgIFZOP3o6ZMgIHVOP3R7YMiIIPOP3W7VMjILkNP3Z7UMkIMgNP3^7SMkIo3U6QLjIP4U6PLlIP4T6]KbH2Y1b4U6ZKcH4X1c4T6XKfH4W1d4S6VKhH6T1e4S6TKkH6R1g4S6PKnH9o0g4S6nJPI:m0i4R6lJSI;j0j4S6iJUI<i0k4Q6hJXI<f0n4f6RKVIR5j6mJSIW5m6iJQIY5n6gJSIY5m6gJSIY5m6fJTIZ5l6eJUI[5k6eJUI[5k6dJVI\\5j6cJWI]5i6bJWI_5i6_JYIa5g6^JZIb5f6]J[Ic5e6[J]Ie5b6[J_Ie5n70000000000000000000000000000000000000001O00001PLfFX2Z9fMjFX2V9gMnFV2R9hMRGV2o8hMTGV2l8hMXGV2h8iM[GU2e8iM`GT2a8jMbGT2^8jMfGT2[8jMhGT2X8jMlGT2U8jMnGT2S8iMPHV2P8iMRHV2o7gMTHX2m7fMVHX2j7fMYHY2h7eMZHZ2g7dM\\HW2g7gM\\Ho1n7oMTHg1V8VNmG_1]8`NfGU1d8hN_Gn0k8POWGe0h;F;ESPo4"}, "label": "baby and her hands touching mouse and keyboard"}]}
{"id": 280873, "image": "COCO_train2014_000000280873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black computer keyboard with rounded edges\"."}], "task": "refering", "answer_template": "The \"black computer keyboard with rounded edges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.207296875, 0.48539583333333336, 0.839328125, 0.7640416666666667], "iscrowd": 0, "area": 38122.62744999999, "rle": {"size": [480, 640], "counts": "Yfn16d>6I7J6I7J6J6I7J6O100O1O1O1O1O1O1O1O001O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O100O1O100O100O1O100O100O1O100O100O10000O1000eNZD\\Of;`0_D_O`;>eDA[;;jDDV;8oDGP;7SEGo:9RECQ;=PE^OT;a0nDZOU;g0lDTOX;l0iDoN[;P1fDlN];U1dDgN_;Y1T11O1N2O1O1O0O2O1O1N2O1O_Oa0O1O1N2O1N2O1O1O10000O1000000000000000000000001O00000000000000000000000000000O1000000000O100_CXOj:h0TE_Og:a0VDWOa0>T;;VD[Oe0a0P;4UD@k0b0k:<TEJh:6WEKi:4VENj:1VE0j:OUE3k:MTE4l:KTE6l:ISE9m:FSE;m:EQE=o:CPE>P;CmD?S;BkD?U;BhD`0X;BeD?[;BcD?];B`D`0`;A^D`0b;_1000000000000000000000000000000000000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O100O100O100O100O100O100O100O100O100O100O1000000000000000000000000001O000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000O100O2O0O100O100O100O100O100O100O100O100O100O100O100O3N5J5L4K5L4K5L4K5L5J5L4K5L4K5G9B?Ago_1"}, "label": "black computer keyboard with rounded edges"}]}
{"id": 280873, "image": "COCO_train2014_000000280873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"area of keyboard where baby can reach\"."}], "task": "refering", "answer_template": "The \"area of keyboard where baby can reach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.207296875, 0.48539583333333336, 0.839328125, 0.7640416666666667], "iscrowd": 0, "area": 38122.62744999999, "rle": {"size": [480, 640], "counts": "Yfn16d>6I7J6I7J6J6I7J6O100O1O1O1O1O1O1O1O001O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O100O1O100O100O1O100O100O1O100O100O10000O1000eNZD\\Of;`0_D_O`;>eDA[;;jDDV;8oDGP;7SEGo:9RECQ;=PE^OT;a0nDZOU;g0lDTOX;l0iDoN[;P1fDlN];U1dDgN_;Y1T11O1N2O1O1O0O2O1O1N2O1O_Oa0O1O1N2O1N2O1O1O10000O1000000000000000000000001O00000000000000000000000000000O1000000000O100_CXOj:h0TE_Og:a0VDWOa0>T;;VD[Oe0a0P;4UD@k0b0k:<TEJh:6WEKi:4VENj:1VE0j:OUE3k:MTE4l:KTE6l:ISE9m:FSE;m:EQE=o:CPE>P;CmD?S;BkD?U;BhD`0X;BeD?[;BcD?];B`D`0`;A^D`0b;_1000000000000000000000000000000000000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O100O100O100O100O100O100O100O100O100O100O1000000000000000000000000001O000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000O100O2O0O100O100O100O100O100O100O100O100O100O100O100O3N5J5L4K5L4K5L4K5L5J5L4K5L4K5G9B?Ago_1"}, "label": "area of keyboard where baby can reach"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in green pants sleeping on the bench\"."}], "task": "refering", "answer_template": "The \"the man in green pants sleeping on the bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237], "bbox": [0.157, 0.1456267409470752, 0.37374999999999997, 0.8045682451253483], "iscrowd": 0, "area": 18494.095899999997, "rle": {"size": [359, 640], "counts": "aeS15Q;4L3M2O2N1N3N00001O0O10001N1O1O101N001O10O01O1O0J7XOh0I6K6I6TMgMfL_2W3gM_L`2^3eM[L`2d3dMUL`2i3cMRLa2m3`MoKc2Q4^MkKf2T4[MhKi2W4XMeKl2Z4UMaKQ3]4oL_KV3`4kL[K[3c4fLYK_3e4aLXKc3g4^LVKd3j4\\LSKh3l4XLRKj3g4ZKUKm03j3`4bK[Kd05j3_4fL_K\\3`4dL`K]3`4bL`K^3`4cL^Ka26dL[4l4lKoJV4m4nKRKR4k4RLTKn3h4WLVKk3f4ZLXKf3e4`LXK`3d4fLZK[3a4kL]KV3]4QMaKo2Y4S2DWITLo6i38O2M3N1N3H80eJeLi2[3VMgLTNOQ4Y3iMlLTNOP4T3jMPMVNOm3o2mMUMTNOl3k2oMZMSNNl3f2PN_MTNKl3d2PNdMRNIm3c2oMgMSNGn3`2oMlMRNDo3]2oMRNQNBo3[2oMWNQN_Oo3Y2oM[NPN]OQ4V2nM`NQNZOP4V2nMcNPNXOR4S2nMgNPNVOR4Q2nMlNnMTOS4P2mMFS28mMJS25lMMS23lMOT20kM2U2MiM6V2MfM5Z2O`M4^21[M2e22VMOj24RMMm2:kLHU3b0WLHh3l3000000000001O000000000001AVL_Ij3[6\\LeId3o5iLPJX3l5lLSJV3h5mLXJT3d5PM[JQ3a5SM_Jm2\\5WMdJj2X5ZMgJh2T5[MlJf2P5^MoJc2l4aMTKf2b4_M\\Ka2a4dM]K\\2`4W2L5L6J7cLoHY2W7]MTI[2R8I7H9\\N[G8R:F:F_]\\4"}, "label": "the man in green pants sleeping on the bench"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man is sleeping on road table\"."}], "task": "refering", "answer_template": "The \"the man is sleeping on road table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237], "bbox": [0.157, 0.1456267409470752, 0.37374999999999997, 0.8045682451253483], "iscrowd": 0, "area": 18494.095899999997, "rle": {"size": [359, 640], "counts": "aeS15Q;4L3M2O2N1N3N00001O0O10001N1O1O101N001O10O01O1O0J7XOh0I6K6I6TMgMfL_2W3gM_L`2^3eM[L`2d3dMUL`2i3cMRLa2m3`MoKc2Q4^MkKf2T4[MhKi2W4XMeKl2Z4UMaKQ3]4oL_KV3`4kL[K[3c4fLYK_3e4aLXKc3g4^LVKd3j4\\LSKh3l4XLRKj3g4ZKUKm03j3`4bK[Kd05j3_4fL_K\\3`4dL`K]3`4bL`K^3`4cL^Ka26dL[4l4lKoJV4m4nKRKR4k4RLTKn3h4WLVKk3f4ZLXKf3e4`LXK`3d4fLZK[3a4kL]KV3]4QMaKo2Y4S2DWITLo6i38O2M3N1N3H80eJeLi2[3VMgLTNOQ4Y3iMlLTNOP4T3jMPMVNOm3o2mMUMTNOl3k2oMZMSNNl3f2PN_MTNKl3d2PNdMRNIm3c2oMgMSNGn3`2oMlMRNDo3]2oMRNQNBo3[2oMWNQN_Oo3Y2oM[NPN]OQ4V2nM`NQNZOP4V2nMcNPNXOR4S2nMgNPNVOR4Q2nMlNnMTOS4P2mMFS28mMJS25lMMS23lMOT20kM2U2MiM6V2MfM5Z2O`M4^21[M2e22VMOj24RMMm2:kLHU3b0WLHh3l3000000000001O000000000001AVL_Ij3[6\\LeId3o5iLPJX3l5lLSJV3h5mLXJT3d5PM[JQ3a5SM_Jm2\\5WMdJj2X5ZMgJh2T5[MlJf2P5^MoJc2l4aMTKf2b4_M\\Ka2a4dM]K\\2`4W2L5L6J7cLoHY2W7]MTI[2R8I7H9\\N[G8R:F:F_]\\4"}, "label": "the man is sleeping on road table"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a empty wooden chair near by man who sleeping table\"."}], "task": "refering", "answer_template": "The \"a empty wooden chair near by man who sleeping table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 193, 194, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286], "bbox": [0.222859375, 0.5596935933147632, 0.4533125, 0.9637604456824512], "iscrowd": 0, "area": 9657.0937, "rle": {"size": [359, 640], "counts": "a^b15m:6F9B>D=N100O2O001N10001N100O2O000O2O000O2O001N10001O00001O00001O00001O0000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O000000000000001O00001O2N2N2N4L5K6J4L1O2N1N2O1O2NN2O0O2O1O1N2O0O2PNROfIo0W6[O`Ig0\\6@_Ib0^6E\\I<a6KYI8d6NVI5i62nH1Q74hHOW75cHM]7o1O1N2O1O001O1O1O1O1N2O1O1N2OaN^1G=C=C=C^`j3"}, "label": "a empty wooden chair near by man who sleeping table"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"very edge of a black wooden bench\"."}], "task": "refering", "answer_template": "The \"very edge of a black wooden bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 193, 194, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286], "bbox": [0.222859375, 0.5596935933147632, 0.4533125, 0.9637604456824512], "iscrowd": 0, "area": 9657.0937, "rle": {"size": [359, 640], "counts": "a^b15m:6F9B>D=N100O2O001N10001N100O2O000O2O000O2O001N10001O00001O00001O00001O0000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O000000000000001O00001O2N2N2N4L5K6J4L1O2N1N2O1O2NN2O0O2O1O1N2O0O2PNROfIo0W6[O`Ig0\\6@_Ib0^6E\\I<a6KYI8d6NVI5i62nH1Q74hHOW75cHM]7o1O1N2O1O001O1O1O1O1N2O1O1N2OaN^1G=C=C=C^`j3"}, "label": "very edge of a black wooden bench"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a striped shirt looking at a man sleeping\"."}], "task": "refering", "answer_template": "The \"a lady in a striped shirt looking at a man sleeping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 39, 40, 41, 62, 63, 64, 65, 86, 87, 88, 110, 111, 133, 134], "bbox": [0.71634375, 0.05838440111420613, 0.86415625, 0.4420612813370474], "iscrowd": 0, "area": 5526.130199999996, "rle": {"size": [359, 640], "counts": "ZcP51Q;8H8[EFW:e002M3N2N000O1000001O0O10001O2M3N2N2N2N2M3N1O200O100O10_NdFU1[9jNeFX1[9fNeF\\1Z9cNfF_1Y9_NhFb1Y9\\NgFf1X9YNhFi1Z92O1N2N2N00000001O0O1000000000001O00001O1O2N1O1OeGXNX7g1]HeNc7`21O0000000O2O001O1O001O001N2O001N1M4L4L3M4L3M4L4L3M4L3M4L4L3M4L3AZ]n0"}, "label": "a lady in a striped shirt looking at a man sleeping"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman looking at a man on a bench\"."}], "task": "refering", "answer_template": "The \"a woman looking at a man on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 39, 40, 41, 62, 63, 64, 65, 86, 87, 88, 110, 111, 133, 134], "bbox": [0.71634375, 0.05838440111420613, 0.86415625, 0.4420612813370474], "iscrowd": 0, "area": 5526.130199999996, "rle": {"size": [359, 640], "counts": "ZcP51Q;8H8[EFW:e002M3N2N000O1000001O0O10001O2M3N2N2N2N2M3N1O200O100O10_NdFU1[9jNeFX1[9fNeF\\1Z9cNfF_1Y9_NhFb1Y9\\NgFf1X9YNhFi1Z92O1N2N2N00000001O0O1000000000001O00001O1O2N1O1OeGXNX7g1]HeNc7`21O0000000O2O001O1O001O001N2O001N1M4L4L3M4L3M4L4L3M4L3M4L4L3M4L3AZ]n0"}, "label": "a woman looking at a man on a bench"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit lying on a bench\"."}], "task": "refering", "answer_template": "The \"a man in a suit lying on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 269, 270], "bbox": [0.5872499999999999, 0.2884679665738162, 0.80746875, 0.9106963788300836], "iscrowd": 0, "area": 14249.712849999993, "rle": {"size": [359, 640], "counts": "jnS47o:<D2O0O1O100O1O10O010O1O010O1O010O1O10O01000O1000O10O104L=C=C=C4L0000O10000000O1O102M3M3M2O2M3M3M3N2M3N2N2N2WH[Mk6f3K4L5J6K4L5K3L5L3M4K4M4M2M4M2M2O0O2O0O100OO2O0O2WLnJ[1S5cNVKV1j4hN^KQ1c4mN`KR1a4kNcKS1]4kNgKS1Y4lNjKQ1X4lNkKS1U4kNoKS1Q4lNRLQ1P4lNTLQ1l3oNVLP1j3nNZLo0f3QO]Ll0d3RO_Ll0a3SOcLj0^3UOeLh0[3WOiLg0U3ZOmLd0Q3]OSM`0k2AYM<f2E\\M8c2HbM4^2LfM1X20lMLS25PNHP28TNEj1<ZN@f1`0_N\\O_1e0fNVOZ1j0kNQOT1P1fLeN;7m2U1_LPOe0Gl2]1WLSOo0[Oi2g1QLTOX1QOf2o1kKVO\\6n0]IXO`6l0YIZOd6j0TI]Oj6b2M3M3N2M2N100O100OUNnIcNQ6U1\\JgNd5X1`JeN_5[1fJaNY5_1kJ^NU5a1PK[NP5d1TKYNl4f1YKVNh4h1^KSNc4l1cKnM]4Q2iKjMX4T2oKfMQ4Y2`2NbP[1"}, "label": "a man in a suit lying on a bench"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a statue of a man laying on a bench\"."}], "task": "refering", "answer_template": "The \"a statue of a man laying on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 269, 270], "bbox": [0.5872499999999999, 0.2884679665738162, 0.80746875, 0.9106963788300836], "iscrowd": 0, "area": 14249.712849999993, "rle": {"size": [359, 640], "counts": "jnS47o:<D2O0O1O100O1O10O010O1O010O1O010O1O10O01000O1000O10O104L=C=C=C4L0000O10000000O1O102M3M3M2O2M3M3M3N2M3N2N2N2WH[Mk6f3K4L5J6K4L5K3L5L3M4K4M4M2M4M2M2O0O2O0O100OO2O0O2WLnJ[1S5cNVKV1j4hN^KQ1c4mN`KR1a4kNcKS1]4kNgKS1Y4lNjKQ1X4lNkKS1U4kNoKS1Q4lNRLQ1P4lNTLQ1l3oNVLP1j3nNZLo0f3QO]Ll0d3RO_Ll0a3SOcLj0^3UOeLh0[3WOiLg0U3ZOmLd0Q3]OSM`0k2AYM<f2E\\M8c2HbM4^2LfM1X20lMLS25PNHP28TNEj1<ZN@f1`0_N\\O_1e0fNVOZ1j0kNQOT1P1fLeN;7m2U1_LPOe0Gl2]1WLSOo0[Oi2g1QLTOX1QOf2o1kKVO\\6n0]IXO`6l0YIZOd6j0TI]Oj6b2M3M3N2M2N100O100OUNnIcNQ6U1\\JgNd5X1`JeN_5[1fJaNY5_1kJ^NU5a1PK[NP5d1TKYNl4f1YKVNh4h1^KSNc4l1cKnM]4Q2iKjMX4T2oKfMQ4Y2`2NbP[1"}, "label": "a statue of a man laying on a bench"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bench with a person lying on it\"."}], "task": "refering", "answer_template": "The \"a black bench with a person lying on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 109, 110, 111, 133, 134, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 269, 270, 271, 272, 293, 294], "bbox": [0.7078125, 0.24774373259052923, 0.8977499999999999, 0.9407520891364903], "iscrowd": 0, "area": 13312.887849999994, "rle": {"size": [359, 640], "counts": "TTo41V;1O0O2TL9VLHh3c0mK_OR4d0kK\\OT4h0iKYOV4j0fKXOY4j0eKWOY4m0cKTO]4o0`KRO_4P1^KRO`4R1]KoNb4T1[KlN`1A\\Og1POjNd1_O[Oi1oNiNf1^OZOl1lNhNj1\\OXOP2kNdNn1\\OVOS2iNbNQ2[OWOS2eNdNT2YOXOT2aNdNW2XOYOU2\\NdN[2XOZOT2YNdN^2XOZOU2UNcNb2XOYOW2QNbNg2WOYOX2mM`Nl2YOWOW2jMaNP3XOWOX2fM`NT3XOWOY2bM_NX3XOWOZ2]M_N\\3XOXOY2ZM_N_3XOXOX2WMaNb3WOXOV2VMcNc3WOXOT2UMeNd3WOWOS2TMgNf3UOXOR2RMiNf3VOYOo1PMlNg3UO[Om1nLnNg3VO\\Oj1lLQOh3UO^Oh1jLROi3WO^Oe1iLTOj3WO^Oc1hLVOj3WO@a1fLXOj3XOA_1cLZOl3WOD\\1`L]Ol3XOFX1^L@l3XOIU1[LCl3YOLP1WLHn3WONn0TLKn3XOOk0SLMn3XONl0TLLn3YOLl0VLKn3YOJn0WLJo3YOHn0YLIo3YOGo0ZLGP4[OCP1]LEQ4ZOAR1]LER4ZO_OS1^LCS4ZO]OU1`LAS4[O[OU1bL@S4^OWOS1eL@T4@ROR1jL^OT4DmNo0oL]OU4GbNR1YMWOU4KSN^1cMgNZ4OeMh2[2\\MVMo2j2g20O10001O2M2O2M2O2M2O2N2M2O2M2O2\\OlJPKV5n4kJRKW5j4lJUKU5i4lJVKW5f4lJYKV5d4kJ[KW5b4lJ]KV5`4kJ_KW5^4lJaKV5\\4kJdKV5Y4mJeKV5X4kJhKW5T4lJjKV5T4kJlKW5P4lJnKV5P4kJPLW5l3lJRLV5l3kJTLW5h3X1L3L4M4K4M4K4L4I8H7H8I8H7H9H7I7H9H7I7Hmbf0"}, "label": "a black bench with a person lying on it"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a park bench with a man in green pants on it\"."}], "task": "refering", "answer_template": "The \"a park bench with a man in green pants on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 32, 54, 55, 77, 78, 79, 100, 121, 122, 144], "bbox": [0.26803125, 0.06052924791086351, 0.44954687499999996, 0.5211977715877437], "iscrowd": 0, "area": 3374.1375000000007, "rle": {"size": [359, 640], "counts": "P_l1?g:0O1O1O100O1O1O1O100O10000O10000O100O10000O100O26I7I7I7Ie\\30ZcL2N3M3M3M3M4M2N2N2N2N3M2M3N2N2N3N111N3L4K5Le0ZO8I4L^MfG[2V8eMmG]2n7cMTH_2h7aM[H`2a7`MaHc2Z7]MiHd2S7\\MPIe2l6[MVIa2l6dMRIU2R7nMmHm1V7TNlHf1W7[NkH_1X7bNiHX1[7iNgHQ1\\7POeHk0^7VOdHc0`7XOhHb0[7ZOkHa0X7[OoH?T7\\OSI>o8K3M2O200O010O0100O0100O010OcEKi96RFNo91mE3T:MfE8[:92N1O1N3N1O1O1O2N1O1O2N1O2N4LZZk3"}, "label": "a park bench with a man in green pants on it"}]}
{"id": 338218, "image": "COCO_train2014_000000338218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench the man in the green pants is on\"."}], "task": "refering", "answer_template": "The \"the bench the man in the green pants is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 32, 54, 55, 77, 78, 79, 100, 121, 122, 144], "bbox": [0.26803125, 0.06052924791086351, 0.44954687499999996, 0.5211977715877437], "iscrowd": 0, "area": 3374.1375000000007, "rle": {"size": [359, 640], "counts": "P_l1?g:0O1O1O100O1O1O1O100O10000O10000O100O10000O100O26I7I7I7Ie\\30ZcL2N3M3M3M3M4M2N2N2N2N3M2M3N2N2N3N111N3L4K5Le0ZO8I4L^MfG[2V8eMmG]2n7cMTH_2h7aM[H`2a7`MaHc2Z7]MiHd2S7\\MPIe2l6[MVIa2l6dMRIU2R7nMmHm1V7TNlHf1W7[NkH_1X7bNiHX1[7iNgHQ1\\7POeHk0^7VOdHc0`7XOhHb0[7ZOkHa0X7[OoH?T7\\OSI>o8K3M2O200O010O0100O0100O010OcEKi96RFNo91mE3T:MfE8[:92N1O1N3N1O1O1O2N1O1O2N1O2N4LZZk3"}, "label": "the bench the man in the green pants is on"}]}
{"id": 182571, "image": "COCO_train2014_000000182571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with one hand in his jeans pocket who is playing wii\"."}], "task": "refering", "answer_template": "The \"the man with one hand in his jeans pocket who is playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362], "bbox": [0.53715625, 0.23910179640718565, 0.8561406249999999, 0.8409780439121757], "iscrowd": 0, "area": 47526.81905, "rle": {"size": [501, 640], "counts": "T`X5:W?<D:F;G`0_O7J5J7I7I6J3M4M3L4M3L4L3N3N2N2N2N2O1N1O2N2N2O1N2N2O0O2N4L3iHbKS3`4cLhK\\3[4[LlKd3X4RLoKn3U4fKSLY4R4]KTLc4m3VKYLj4i3nJ]LQ5e3hJaLX5`3bJeL^5\\3^JgLa5\\3ZJgLf5Z3VJiLi5Y3SJjLm5X3nIkLR6V3jImLU6U3gInLY6S3aIRM_6R6O10000O100O10000O2O000O100O10000O100O10000O100O100O10000O100O10000O100O10000000000000000000000001O00000000000000001O001O1O001O001O001O001O2N3M2N3M3M2N3M3M2N2N00O1O1O100O1O1O100O1O1O100O10O2N100O101N100O100O2O0O100O101N1O1O1O2M2O1O1O1O2N1N2O1O2N1O2N1L5J5L4K6J5L5M2N3M2N4L3M3M3M4L3M3M3N3L7UKkFk2\\9eLSGT3m:I7J7H7J6I7ZOm0QOYh\\1"}, "label": "the man with one hand in his jeans pocket who is playing wii"}]}
{"id": 182571, "image": "COCO_train2014_000000182571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white shirt\"."}], "task": "refering", "answer_template": "The \"the man in the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362], "bbox": [0.53715625, 0.23910179640718565, 0.8561406249999999, 0.8409780439121757], "iscrowd": 0, "area": 47526.81905, "rle": {"size": [501, 640], "counts": "T`X5:W?<D:F;G`0_O7J5J7I7I6J3M4M3L4M3L4L3N3N2N2N2N2O1N1O2N2N2O1N2N2O0O2N4L3iHbKS3`4cLhK\\3[4[LlKd3X4RLoKn3U4fKSLY4R4]KTLc4m3VKYLj4i3nJ]LQ5e3hJaLX5`3bJeL^5\\3^JgLa5\\3ZJgLf5Z3VJiLi5Y3SJjLm5X3nIkLR6V3jImLU6U3gInLY6S3aIRM_6R6O10000O100O10000O2O000O100O10000O100O10000O100O100O10000O100O10000O100O10000000000000000000000001O00000000000000001O001O1O001O001O001O001O2N3M2N3M3M2N3M3M2N2N00O1O1O100O1O1O100O1O1O100O10O2N100O101N100O100O2O0O100O101N1O1O1O2M2O1O1O1O2N1N2O1O2N1O2N1L5J5L4K6J5L5M2N3M2N4L3M3M3M4L3M3M3N3L7UKkFk2\\9eLSGT3m:I7J7H7J6I7ZOm0QOYh\\1"}, "label": "the man in the white shirt"}]}
{"id": 182571, "image": "COCO_train2014_000000182571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a red night gown and a santa hat\"."}], "task": "refering", "answer_template": "The \"a woman wearing a red night gown and a santa hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 192, 193, 194, 215, 216, 217, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332, 352, 353, 354, 355], "bbox": [0.325609375, 0.41790419161676645, 0.47221874999999996, 0.8429740518962077], "iscrowd": 0, "area": 14262.227850000003, "rle": {"size": [501, 640], "counts": "eUV31b?2M3M3M3N2M3M3M3oM\\OoDf0m:AmDb0o:EgDb0U;EbDa0\\;DZDc0e;ARDe0l;_OlCg0S<^1O1O1O1N2N2N2[MRMaHP3]7UM_Hl2_7YM\\Hi2b7[MZHg2d7^MWHd2i7_MRHc2n7_MoGb2P8aMlGa2T8aMhGa2X8bMcG`2]8bM_G`2`8cM]G^2c8dMYG^2g8dMUG^2k8eMPG]2o8U2000000O1000000O11O0000000000001O00001O1O001O1O1O1O1O1O001O1O1O1O1O001SKbFa3_9\\LcFd3^9YLeFf3\\9WLgFh3Z9ULiFj3W9TLlFk3U9RLnFm3S9PLPGo3Q9mKSGR4n8kKTGU4n8gKWGV4k8eK[GX4T:M3M4L3M3M3ZNSE^NQ;Z1\\E]Ng:]1eEZN^:_1RTU5"}, "label": "a woman wearing a red night gown and a santa hat"}]}
{"id": 182571, "image": "COCO_train2014_000000182571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in santa hat\"."}], "task": "refering", "answer_template": "The \"woman in santa hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 192, 193, 194, 215, 216, 217, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332, 352, 353, 354, 355], "bbox": [0.325609375, 0.41790419161676645, 0.47221874999999996, 0.8429740518962077], "iscrowd": 0, "area": 14262.227850000003, "rle": {"size": [501, 640], "counts": "eUV31b?2M3M3M3N2M3M3M3oM\\OoDf0m:AmDb0o:EgDb0U;EbDa0\\;DZDc0e;ARDe0l;_OlCg0S<^1O1O1O1N2N2N2[MRMaHP3]7UM_Hl2_7YM\\Hi2b7[MZHg2d7^MWHd2i7_MRHc2n7_MoGb2P8aMlGa2T8aMhGa2X8bMcG`2]8bM_G`2`8cM]G^2c8dMYG^2g8dMUG^2k8eMPG]2o8U2000000O1000000O11O0000000000001O00001O1O001O1O1O1O1O1O001O1O1O1O1O001SKbFa3_9\\LcFd3^9YLeFf3\\9WLgFh3Z9ULiFj3W9TLlFk3U9RLnFm3S9PLPGo3Q9mKSGR4n8kKTGU4n8gKWGV4k8eK[GX4T:M3M4L3M3M3ZNSE^NQ;Z1\\E]Ng:]1eEZN^:_1RTU5"}, "label": "woman in santa hat"}]}
{"id": 125234, "image": "COCO_train2014_000000125234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women playing with the child\"."}], "task": "refering", "answer_template": "The \"a women playing with the child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 306, 307, 328, 329, 330], "bbox": [0.29085937500000003, 0.37768149882903984, 0.644703125, 0.9866744730679158], "iscrowd": 0, "area": 18735.397999999997, "rle": {"size": [427, 640], "counts": "`n]21Z=2M3N2M3N2N2M2O0O2O0O101N101O0O101N100O101N100O10M30O100O100`NYOYFj0d9^OTFe0h9EoE>m9KkE8R:b1L3N3E;I7O1O1O1O1O1O1O1O1O1O1O1N2M3N2N0000O101O2N1N3I6J7iNV1O2M3M2N3M3N2O00100O1O010O1O100O100O1O100O1O101O00000001O000000001O000001O01O000000000000001O000001OO100O1O100O1O100O1O100O1O1B>J6A?O1O1N2O1N2O2N1N2O1000000000000000000001O0000001O000000ROfLlGY3R8jLmGV3R8lLmGT3R8oLmGP3Q8SMmGn2R8TMbGX3[8m0A?00000000000000O100000000000000000M3M3N1N3M3L4L4L4L4L4K5L4M3M2N3M3N2M4L4L4L3N3M3M3M3M3M3M3N2M2N3M3M3M3M3N2M3M2N3L4L4L4L8G:G:Fk^n2"}, "label": "a women playing with the child"}]}
{"id": 125234, "image": "COCO_train2014_000000125234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a child by the hand sitting in a armchair\"."}], "task": "refering", "answer_template": "The \"a woman holding a child by the hand sitting in a armchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 306, 307, 328, 329, 330], "bbox": [0.29085937500000003, 0.37768149882903984, 0.644703125, 0.9866744730679158], "iscrowd": 0, "area": 18735.397999999997, "rle": {"size": [427, 640], "counts": "`n]21Z=2M3N2M3N2N2M2O0O2O0O101N101O0O101N100O101N100O10M30O100O100`NYOYFj0d9^OTFe0h9EoE>m9KkE8R:b1L3N3E;I7O1O1O1O1O1O1O1O1O1O1O1N2M3N2N0000O101O2N1N3I6J7iNV1O2M3M2N3M3N2O00100O1O010O1O100O100O1O100O1O101O00000001O000000001O000001O01O000000000000001O000001OO100O1O100O1O100O1O100O1O1B>J6A?O1O1N2O1N2O2N1N2O1000000000000000000001O0000001O000000ROfLlGY3R8jLmGV3R8lLmGT3R8oLmGP3Q8SMmGn2R8TMbGX3[8m0A?00000000000000O100000000000000000M3M3N1N3M3L4L4L4L4L4K5L4M3M2N3M3N2M4L4L4L3N3M3M3M3M3M3M3N2M2N3M3M3M3M3N2M3M2N3L4L4L4L8G:G:Fk^n2"}, "label": "a woman holding a child by the hand sitting in a armchair"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and silver knife in a plate setting\"."}], "task": "refering", "answer_template": "The \"a black and silver knife in a plate setting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [350, 351, 372, 373, 395, 417, 440, 463, 485, 486, 508], "bbox": [0.081484375, 0.645921875, 0.2930625, 1.0], "iscrowd": 0, "area": 2388.792900000002, "rle": {"size": [640, 640], "counts": "oSQ11nc01N2O1N2O1N2O1N2O1N2N2OO0100O100O100O100O1bN[Oa_Oe0_`0[Oa_Of0_`0ZO`_Of0``0ZO`_Of0``0[O__Oe0a`0[O__Of0``0[O__Oe0b`0ZO^_Of0b`0[O]_Oe0c`0[O]_Oe0c`0\\O\\_Od0e`0[O[_Oe0e`0\\OZ_Od0f`0]OY_Oc0g`0]OY_Oc0h`0]OW_Oc0i`0]OW_Oc0i`0^OW_Oa0i`0_OX_O`0i`0@W_O?i`0AW_O?i`0BW_O>h`0CX_O<i`0CX_O<h`0EX_O:h`0FY_O9g`0HY_O7g`0JY_O5h`0JY_O5g`0LX_O4h`0MX_O2h`0NY_O1h`0OX_O0h`00Y_OOg`02Y_OMg`04Y_OKh`04Y_OKg`06X_OJh`06Z_OIe`08\\_OFe`09]_OEc`0<_1001O010O001O010O001O010O1O0010O01O0010O01O001O010O001O010O001O0100O101N101N2O0O2O1N101N101N2O0O2000O2O0O2O1O0O2O0OO2O001N2O001N101O1N101N101OSgj8"}, "label": "a black and silver knife in a plate setting"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate full of food with green leave salad and a piece of meat on a flowered plate\"."}], "task": "refering", "answer_template": "The \"a plate full of food with green leave salad and a piece of meat on a flowered plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [356, 357, 358, 359, 360, 361, 378, 379, 380, 381, 382, 383, 384, 385, 386, 401, 402, 403, 404, 405, 406, 407, 408, 409, 424, 425, 426, 427, 428, 429, 430, 431, 432, 447, 448, 449, 450, 451, 452, 453, 454, 455, 470, 471, 472, 473, 474, 475, 476, 477, 478, 493, 494, 495, 496, 497, 498, 499, 500, 501, 516, 517, 518, 519, 520, 521, 522], "bbox": [0.427265625, 0.661, 0.7974218750000001, 1.0], "iscrowd": 0, "area": 40069.099949999996, "rle": {"size": [640, 640], "counts": "ag[55jc04K5L4L4K5L4L4K6K4P@hNP=]1]BSOb=Q1lA_OR>e0cADZ>a0cA_O\\>e0aA\\O]>h0`AXO^>m0_ATO_>P1^AQO_>T1]AmNb>W1[AjNc>Z1ZAgNc>^1ZAbNe>b1XA_Nf>e1WA[Ng>j1VAWNh>b3M2O2N2M2O2M2O2N2M2O2N2M2O2M2O2OO100O010O10O10O10O0100O10O0100O010O10O01000O0100O10O0100O010O10O0100O10O01000O0100O010O10O0100O10O01001N10000O1000000O10000O1000000010O0000000000001O00000000000010O0000000001O0000000000001O0001O00000001O0000000000001O000001O00O2O0000000000001N1000000000001O0O10000000001O000O10001O001O00001N101O0O2N100O2O0O2O0O1O2O0O2O0O2N100O2O0O2O0O2N100O2O0O2O0O2N100O2O0O2O0O1O2O0O2O0O2N100O2eMZ2D=C<D=CRj`2"}, "label": "a plate full of food with green leave salad and a piece of meat on a flowered plate"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a meal sitting on a plate with a floral pattern\"."}], "task": "refering", "answer_template": "The \"a meal sitting on a plate with a floral pattern\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [356, 357, 358, 359, 360, 361, 378, 379, 380, 381, 382, 383, 384, 385, 386, 401, 402, 403, 404, 405, 406, 407, 408, 409, 424, 425, 426, 427, 428, 429, 430, 431, 432, 447, 448, 449, 450, 451, 452, 453, 454, 455, 470, 471, 472, 473, 474, 475, 476, 477, 478, 493, 494, 495, 496, 497, 498, 499, 500, 501, 516, 517, 518, 519, 520, 521, 522], "bbox": [0.427265625, 0.661, 0.7974218750000001, 1.0], "iscrowd": 0, "area": 40069.099949999996, "rle": {"size": [640, 640], "counts": "ag[55jc04K5L4L4K5L4L4K6K4P@hNP=]1]BSOb=Q1lA_OR>e0cADZ>a0cA_O\\>e0aA\\O]>h0`AXO^>m0_ATO_>P1^AQO_>T1]AmNb>W1[AjNc>Z1ZAgNc>^1ZAbNe>b1XA_Nf>e1WA[Ng>j1VAWNh>b3M2O2N2M2O2M2O2N2M2O2N2M2O2M2O2OO100O010O10O10O10O0100O10O0100O010O10O01000O0100O10O0100O010O10O0100O10O01000O0100O010O10O0100O10O01001N10000O1000000O10000O1000000010O0000000000001O00000000000010O0000000001O0000000000001O0001O00000001O0000000000001O000001O00O2O0000000000001N1000000000001O0O10000000001O000O10001O001O00001N101O0O2N100O2O0O2O0O1O2O0O2O0O2N100O2O0O2O0O2N100O2O0O2O0O2N100O2O0O2O0O1O2O0O2O0O2N100O2eMZ2D=C<D=CRj`2"}, "label": "a meal sitting on a plate with a floral pattern"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"food on the brown plate at the top\"."}], "task": "refering", "answer_template": "The \"food on the brown plate at the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 106], "bbox": [0.545671875, 0.0, 0.8123437499999999, 0.18765625], "iscrowd": 0, "area": 12750.339549999999, "rle": {"size": [640, 640], "counts": "RTj61oc05K4L5K5K5K5T]OSO[b0Q1b]OPO\\b0U1a]OkN\\b0[1`]OfN^b0a1M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N001O1O1O1O001O1O1O001O1O1O001O1O1O1O001ON2M3N2N2M3N2M3N2M3N2M30000O100O10000O10000O10000O10000O10000O10000O10000O10000O10O010000O10000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O2O1O2N1N3N1O2M2O2N1O2M2O1O2M2O2N1M4M2N3L3N3L3N3L3N3M2M3N3L3NfnZ2"}, "label": "food on the brown plate at the top"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"meal on plate with brown rim\"."}], "task": "refering", "answer_template": "The \"meal on plate with brown rim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 106], "bbox": [0.545671875, 0.0, 0.8123437499999999, 0.18765625], "iscrowd": 0, "area": 12750.339549999999, "rle": {"size": [640, 640], "counts": "RTj61oc05K4L5K5K5K5T]OSO[b0Q1b]OPO\\b0U1a]OkN\\b0[1`]OfN^b0a1M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N001O1O1O1O001O1O1O001O1O1O001O1O1O1O001ON2M3N2N2M3N2M3N2M3N2M30000O100O10000O10000O10000O10000O10000O10000O10000O10000O10O010000O10000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O2O1O2N1N3N1O2M2O2N1O2M2O1O2M2O2N1M4M2N3L3N3L3N3L3N3M2M3N3L3NfnZ2"}, "label": "meal on plate with brown rim"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork that is between the two plates\"."}], "task": "refering", "answer_template": "The \"the fork that is between the two plates\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [350, 351, 372, 373, 374, 395, 396, 397, 417, 418, 419, 440, 441, 462, 463, 484, 485, 486, 506, 507, 508], "bbox": [0.02525, 0.6696093750000001, 0.29485937500000003, 1.0], "iscrowd": 0, "area": 6936.935800000002, "rle": {"size": [640, 640], "counts": "nW;2mc01O1N2O1O1N2O1O1N2O1O1N2O1O1O0O2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1O100O1O001O1O1OO1O1O1O1N2O1N2O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O0000O1O100O1O100O100O1O100O1O102M2O2O0O101N1O2O0OiNi]Oh0Wb0XOj]Oh0Tb0YOl]Og0Tb0YOm]Og0Qb0YOP^Og0Pb0YOP^Og0oa0YOQ^Oh0oa0WOQ^Oj0na0VOS^OLMa0Pb0CR^OMO?na0DT^OMO>ma0DT^OO0<ka0FT^OO0<ka0EV^ON0=ja0DV^O0O>ia0BY^O0N>ia0BX^O1NON0ka00Y^O1N0ONja00Z^O2MO1Nga02Z^O1N01Mga01[^O1N01Nea01\\^O2M02Mea00\\^O3NO1Oda00]^O1N10Oea0N]^O3M02Oca0N_^O2M100da0M^^O2N101ca0L_^O3N001ca0K`^O3M101ca0L_^O2O0O3ca0J`^O3M013aa0Jb^O2L203ba0Ia^O2N1O4ba0Ib^O2L114aa0Hc^O2L106`a0Hc^O1M115`a0Hc^O1M107_a0Gd^O2L018_a0Ee^O2K128]a0Fe^O1MO2:]a0Ee^O2KO3;\\a0Dg^O1K02;\\a0Df^O1L02;\\a0Df^O2LN2;^a0Dc^O4MM3:^a0Ea^O4NM3:`a0C_^O6OL2;ba0A]^O9NL38fa0AX^O;OL38oa0Lo]OK38na0Mo]OK38na0Nn]OJ57ma0Oo]OJ37Pb0Mm]OL46Rb0Kj]OO46Ub0If]O165Wb0Gc]O465Zb0D`]O774`b0M_]O3ab0M`]O2bb0L^]O4db0J]]O5fb0HZ]O8Wc0O0000O0011O2M3N3MV^i8"}, "label": "the fork that is between the two plates"}]}
{"id": 330040, "image": "COCO_train2014_000000330040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork on the right\"."}], "task": "refering", "answer_template": "The \"the fork on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [350, 351, 372, 373, 374, 395, 396, 397, 417, 418, 419, 440, 441, 462, 463, 484, 485, 486, 506, 507, 508], "bbox": [0.02525, 0.6696093750000001, 0.29485937500000003, 1.0], "iscrowd": 0, "area": 6936.935800000002, "rle": {"size": [640, 640], "counts": "nW;2mc01O1N2O1O1N2O1O1N2O1O1N2O1O1O0O2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1O100O1O001O1O1OO1O1O1O1N2O1N2O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O0000O1O100O1O100O100O1O100O1O102M2O2O0O101N1O2O0OiNi]Oh0Wb0XOj]Oh0Tb0YOl]Og0Tb0YOm]Og0Qb0YOP^Og0Pb0YOP^Og0oa0YOQ^Oh0oa0WOQ^Oj0na0VOS^OLMa0Pb0CR^OMO?na0DT^OMO>ma0DT^OO0<ka0FT^OO0<ka0EV^ON0=ja0DV^O0O>ia0BY^O0N>ia0BX^O1NON0ka00Y^O1N0ONja00Z^O2MO1Nga02Z^O1N01Mga01[^O1N01Nea01\\^O2M02Mea00\\^O3NO1Oda00]^O1N10Oea0N]^O3M02Oca0N_^O2M100da0M^^O2N101ca0L_^O3N001ca0K`^O3M101ca0L_^O2O0O3ca0J`^O3M013aa0Jb^O2L203ba0Ia^O2N1O4ba0Ib^O2L114aa0Hc^O2L106`a0Hc^O1M115`a0Hc^O1M107_a0Gd^O2L018_a0Ee^O2K128]a0Fe^O1MO2:]a0Ee^O2KO3;\\a0Dg^O1K02;\\a0Df^O1L02;\\a0Df^O2LN2;^a0Dc^O4MM3:^a0Ea^O4NM3:`a0C_^O6OL2;ba0A]^O9NL38fa0AX^O;OL38oa0Lo]OK38na0Mo]OK38na0Nn]OJ57ma0Oo]OJ37Pb0Mm]OL46Rb0Kj]OO46Ub0If]O165Wb0Gc]O465Zb0D`]O774`b0M_]O3ab0M`]O2bb0L^]O4db0J]]O5fb0HZ]O8Wc0O0000O0011O2M3N3MV^i8"}, "label": "the fork on the right"}]}
{"id": 567616, "image": "COCO_train2014_000000567616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange half on the left\"."}], "task": "refering", "answer_template": "The \"the orange half on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219], "bbox": [0.369046875, 0.18824761904761905, 0.6555625, 0.5034666666666666], "iscrowd": 0, "area": 23706.40345, "rle": {"size": [525, 640], "counts": "eVi35i?`0@a0_O`0L5L3M4L4L3M4L3M4L4L3M4L3M4L3M4L4L3M3N1O2N1O1O1O2N1O1O2N1O1O2N1O1O101N10001N100O2O000O101N100O2O000O2O0O100O10000O100O100O10000O100O100O10000O100O10000000000000000O10000000000000000O10000000000000000O100000000000O100O1O010O1O1O100O1O1O2O0O2N101N1O2N101N1O2N101N1O2O0O2N1O2O0O2N1O2O0O2N101N1O2N101N2N101N1O2O0O2N101N1O2O0O2N101N3L5K4M4K4L5K4L5L3L5K4L5K4L5L4K:Fg0YO`c`3"}, "label": "the orange half on the left"}]}
{"id": 567616, "image": "COCO_train2014_000000567616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange half on the edge of the bowl\"."}], "task": "refering", "answer_template": "The \"an orange half on the edge of the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219], "bbox": [0.369046875, 0.18824761904761905, 0.6555625, 0.5034666666666666], "iscrowd": 0, "area": 23706.40345, "rle": {"size": [525, 640], "counts": "eVi35i?`0@a0_O`0L5L3M4L4L3M4L3M4L4L3M4L3M4L3M4L4L3M3N1O2N1O1O1O2N1O1O2N1O1O2N1O1O101N10001N100O2O000O101N100O2O000O2O0O100O10000O100O100O10000O100O100O10000O100O10000000000000000O10000000000000000O10000000000000000O100000000000O100O1O010O1O1O100O1O1O2O0O2N101N1O2N101N1O2N101N1O2O0O2N1O2O0O2N1O2O0O2N101N1O2N101N2N101N1O2O0O2N101N1O2O0O2N101N3L5K4M4K4L5K4L5L3L5K4L5K4L5L4K:Fg0YO`c`3"}, "label": "an orange half on the edge of the bowl"}]}
{"id": 567616, "image": "COCO_train2014_000000567616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the orange with the least amount of peel showing\"."}], "task": "refering", "answer_template": "The \"the half of the orange with the least amount of peel showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338], "bbox": [0.53859375, 0.4311809523809524, 0.8149218749999999, 0.7835619047619048], "iscrowd": 0, "area": 25002.869499999997, "rle": {"size": [525, 640], "counts": "^Va56R`09G:F9G9G9G:G8I7J6J6J2O2M3M3M3M3N2M3M3M3M3N2M2N3N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O1O2N1O2N1O2N100O2O0O2O0O2O001N100O100O100O10000O100O100O10O10O1000000O100000000000000000O10000000000000000O100O10000O100O10000O100O01000O10001N100O2O000O2O0O101O0O1O2N100O2N1O1O2O0O1O2N1O101N1O1O2N2O2M2N2N2N2O2M2N2N2N2N2N3M2M3N3M3L5L3M4L3L5L4L3M4K4K6J5K6J6J7I8H8H9G8H8HoUl1"}, "label": "the half of the orange with the least amount of peel showing"}]}
{"id": 567616, "image": "COCO_train2014_000000567616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brighter orange half\"."}], "task": "refering", "answer_template": "The \"the brighter orange half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338], "bbox": [0.53859375, 0.4311809523809524, 0.8149218749999999, 0.7835619047619048], "iscrowd": 0, "area": 25002.869499999997, "rle": {"size": [525, 640], "counts": "^Va56R`09G:F9G9G9G:G8I7J6J6J2O2M3M3M3M3N2M3M3M3M3N2M2N3N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O1O2N1O2N1O2N100O2O0O2O0O2O001N100O100O100O10000O100O100O10O10O1000000O100000000000000000O10000000000000000O100O10000O100O10000O100O01000O10001N100O2O000O2O0O101O0O1O2N100O2N1O1O2O0O1O2N1O101N1O1O2N2O2M2N2N2N2O2M2N2N2N2N2N3M2M3N3M3L5L3M4L3L5L4L3M4K4K6J5K6J6J7I8H8H9G8H8HoUl1"}, "label": "the brighter orange half"}]}
{"id": 354631, "image": "COCO_train2014_000000354631.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant to the far right that is only partially in the picture\"."}], "task": "refering", "answer_template": "The \"the elephant to the far right that is only partially in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.5589999999999999, 0.0011241217798594848, 1.0, 0.9966276346604216], "iscrowd": 0, "area": 63658.214349999995, "rle": {"size": [427, 640], "counts": "RYe46U=1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O001O1O1O001O1O1O1O010O1O1O001O1O1O1O001O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1Oi0WO7I4L4L4L2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1fN]JjJd5U5eJbJ\\5]5eJbJ\\5^5dJaJ]5^5dJaJ]5^5dJaJ]5^5dJaJ]5^5dJaJ]5^5dJaJ]5_5cJ`J^5_5cJ`J^5_5cJ`J^5_5bJaJ_5^5aJbJ`5]5`JcJa5]5^JcJc5\\5]JdJd5[5\\JeJe5Z5[JfJf5Y5YJhJh5W5XJiJi5W5UJjJl5U5SJlJn5S5bIlJL1c6R5bInJH1g6P5bIPKD1k6m4dI]K]6b4dI]K]6b4eI\\K\\6c4eI\\K\\6c4fI[K[6c4kIXKV6g4oITKR6k4QJRKo5n4QJRKP6m4PJSKQ6k4PJUKQ6j4oIVKR6i4nIWKS6h4mIXKT6g4lIYKU6e4lI[KU6d4kI\\KV6c4jI]KW6a4jI_KW6`4iI`KX6^4iIbKX6]4hIcKY6[4hIeKY6Z4gIfKZ6X4gIhKZ6W4fIiK[6U4fIkK\\6S4dImK]6Q4dIoK^6o3bIQL`6l3aITLa6j3_IVLc6g3^IYLd6d3]I\\Ld6c3\\I]Lf6`3[I`Lg6^3YIbLi6[3XIeLj6Y3VIgLl6V3UIjLl6U3TIkLn6P3UIPMm6k2VIUMl6f2WIZMk6`2YI`Mi6]2XIcMj6Y2XIgMk6T2WIlMk6Q2VIoMl6m1VISNl6j1UIVNm6f1UIZNm6^1YIbNi6U1^IkNe6m0`ISOd6d0aI\\Od6;`IEd63`IMo9000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O"}, "label": "the elephant to the far right that is only partially in the picture"}]}
{"id": 272716, "image": "COCO_train2014_000000272716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the refrigerator , to the right of the girl and against the wall\"."}], "task": "refering", "answer_template": "The \"the refrigerator , to the right of the girl and against the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 20, 21, 22, 23, 24, 25, 26, 37, 38, 39, 40, 41, 42, 43, 58, 59, 60, 61, 75, 76, 77, 78, 92, 93, 94, 95, 109, 110, 111, 112, 127, 128, 129, 145, 146], "bbox": [0.20675, 0.00225, 0.6142291666666666, 0.37753125], "iscrowd": 0, "area": 20462.267199999995, "rle": {"size": [640, 480], "counts": "\\lm11oc05J6K4L5K5K5K5K5K4L5K5K2N00O10000000O10000000000000000000000000000000O10O100000000000000000000000000000000O1000O1000000000000000000000000000O100000000000O1000000000000000000000O1000000000000000O10000000000001O1N2O001O1O1O1O1O1O1O1O00?Ae0[Od0\\Oe0[O<D4K6K4L5K4L4L5K4L4L4L0000000000001O000O10000000001O000000000000001O1O001N2O1O1O001O1O1O001O1O1O1O001O1O1O1N1013M<D<D<C=D<D<D<D;D=D<D<D<D<C=D<D<D<C=Da\\c3"}, "label": "the refrigerator , to the right of the girl and against the wall"}]}
{"id": 272716, "image": "COCO_train2014_000000272716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a refrigerator handle next to a smaller freezer\"."}], "task": "refering", "answer_template": "The \"a refrigerator handle next to a smaller freezer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 20, 21, 22, 23, 24, 25, 26, 37, 38, 39, 40, 41, 42, 43, 58, 59, 60, 61, 75, 76, 77, 78, 92, 93, 94, 95, 109, 110, 111, 112, 127, 128, 129, 145, 146], "bbox": [0.20675, 0.00225, 0.6142291666666666, 0.37753125], "iscrowd": 0, "area": 20462.267199999995, "rle": {"size": [640, 480], "counts": "\\lm11oc05J6K4L5K5K5K5K5K4L5K5K2N00O10000000O10000000000000000000000000000000O10O100000000000000000000000000000000O1000O1000000000000000000000000000O100000000000O1000000000000000000000O1000000000000000O10000000000001O1N2O001O1O1O1O1O1O1O1O00?Ae0[Od0\\Oe0[O<D4K6K4L5K4L4L5K4L4L4L0000000000001O000O10000000001O000000000000001O1O001N2O1O1O001O1O1O001O1O1O1O001O1O1O1N1013M<D<D<C=D<D<D<D;D=D<D<D<D<C=D<D<D<C=Da\\c3"}, "label": "a refrigerator handle next to a smaller freezer"}]}
{"id": 272716, "image": "COCO_train2014_000000272716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white fridge with magnets on it\"."}], "task": "refering", "answer_template": "The \"a small white fridge with magnets on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 58, 68, 69, 70, 71, 72, 73, 74, 75, 85, 86, 87, 88, 89, 90, 91, 92, 102, 103, 104, 105, 106, 107, 108, 109, 110, 119, 120, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 205, 206, 207, 208, 209, 210, 211, 212, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 280, 281, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 344, 345, 346, 347, 348, 361, 362, 363, 364, 378, 379, 380, 381], "bbox": [0.007604166666666666, 0.09421874999999999, 0.5678541666666667, 0.9997343749999998], "iscrowd": 0, "area": 103495.61660000002, "rle": {"size": [640, 480], "counts": "ic2[4d?4M2N3M3L4M2N3M3L4M2N3M3L4M2N3L4M3M2N3L4M3M2N3L4M3M2N3L4M2N3L4M3M2M4M3M3L3N3M3L4M2N3L4M3M2N3L4M3M2M4M3M3L3N3M3L4M2N3L4M2N3L4M3M2M4M3M3L3N3M3L4M2N3M3M3M2M4M3M3M2N3M3M3M2N3M3M3M2N3M3L3N3M3M3M2N3M3M3M2N3M3M3M3hNSBVLP>h3PBXLS>e3mA[LV>b3jA^LY>`3gA_L\\>^3dAbL^>\\3bAdLa>Y3_AgLd>V3]AiLf>T3ZAlLi>Q3WAoLk>o2VAPMm>m2SASMQ?i2o@VMX?d2h@\\M]?_2d@`Ma?[2_@eMf?V2Z@jMl?P2T@PNQ`0k1P@TNU`0g1k_OYNZ`0b1f_O^N``0\\1`_OdNe`0W1\\_OhNi`0S1W_OmNk`0Q1U_OoNk`0Q1U_OnNl`0R1U_OmNk`0S1U_OmNk`0S1U_OmNk`0S1U_OmNk`0S1V_OlNj`0T1V_OlNj`0T1V_OlNj`0T1V_OlNj`0T1W_OkNi`0U1W_OkNi`0U1W_OkNi`0U1W_OkNi`0U1X_OiNi`0W1W_OiNi`0W1W_OiNi`0W1X_OhNh`0X1X_OhNh`0X1X_OhNh`0X1X_OhNh`0X1Y_OgNg`0Y1Y_OgNg`0Y1Y_OgNg`0Z1X_OfNh`0Z1X_OfNh`0Z1X_OeNi`0[1X_OcNi`0]1W_OaNk`0_1U_O`Nl`0`1T_O^Nn`0b1R_O\\NPa0d1P_O[NQa0e1P_OXNRa0h1n^OWNSa0i1m^OUNUa0k1k^OTNVa0l1?000000000000000000000000000O10000O100O100O10000O100O100O10000O101N100O1O1M3N3M2N3L3N3M2N3L4M2NU2kMc0\\O;F6J5K6I6K6J5K6I6K6YDXHl:n7oDWHk:n7REWHf:P8UEUHe:h8J7I6I8I6J6J7H7J6J7I6I8I6J4L3nJ[Ed0h:WO\\Eh0g:RO^El0e:nN`EP1c:jNbES1a:hNdEV1_:cNgE[1\\:_NiE5cM^Og<7kE8cM\\Of<6kE;fMZOa<6nE=fMYO_<4PF`0gMWO\\<3RFc0gMUO[<2SFf0hMSOW<1WFi0hMQOT<0YFl0W?L3N3M3M3M3M3M3L4M2NonP4"}, "label": "a small white fridge with magnets on it"}]}
{"id": 272716, "image": "COCO_train2014_000000272716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half - sized frigde door\"."}], "task": "refering", "answer_template": "The \"a half - sized frigde door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 58, 68, 69, 70, 71, 72, 73, 74, 75, 85, 86, 87, 88, 89, 90, 91, 92, 102, 103, 104, 105, 106, 107, 108, 109, 110, 119, 120, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 205, 206, 207, 208, 209, 210, 211, 212, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 280, 281, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 344, 345, 346, 347, 348, 361, 362, 363, 364, 378, 379, 380, 381], "bbox": [0.007604166666666666, 0.09421874999999999, 0.5678541666666667, 0.9997343749999998], "iscrowd": 0, "area": 103495.61660000002, "rle": {"size": [640, 480], "counts": "ic2[4d?4M2N3M3L4M2N3M3L4M2N3M3L4M2N3L4M3M2N3L4M3M2N3L4M3M2N3L4M2N3L4M3M2M4M3M3L3N3M3L4M2N3L4M3M2N3L4M3M2M4M3M3L3N3M3L4M2N3L4M2N3L4M3M2M4M3M3L3N3M3L4M2N3M3M3M2M4M3M3M2N3M3M3M2N3M3M3M2N3M3L3N3M3M3M2N3M3M3M2N3M3M3M3hNSBVLP>h3PBXLS>e3mA[LV>b3jA^LY>`3gA_L\\>^3dAbL^>\\3bAdLa>Y3_AgLd>V3]AiLf>T3ZAlLi>Q3WAoLk>o2VAPMm>m2SASMQ?i2o@VMX?d2h@\\M]?_2d@`Ma?[2_@eMf?V2Z@jMl?P2T@PNQ`0k1P@TNU`0g1k_OYNZ`0b1f_O^N``0\\1`_OdNe`0W1\\_OhNi`0S1W_OmNk`0Q1U_OoNk`0Q1U_OnNl`0R1U_OmNk`0S1U_OmNk`0S1U_OmNk`0S1U_OmNk`0S1V_OlNj`0T1V_OlNj`0T1V_OlNj`0T1V_OlNj`0T1W_OkNi`0U1W_OkNi`0U1W_OkNi`0U1W_OkNi`0U1X_OiNi`0W1W_OiNi`0W1W_OiNi`0W1X_OhNh`0X1X_OhNh`0X1X_OhNh`0X1X_OhNh`0X1Y_OgNg`0Y1Y_OgNg`0Y1Y_OgNg`0Z1X_OfNh`0Z1X_OfNh`0Z1X_OeNi`0[1X_OcNi`0]1W_OaNk`0_1U_O`Nl`0`1T_O^Nn`0b1R_O\\NPa0d1P_O[NQa0e1P_OXNRa0h1n^OWNSa0i1m^OUNUa0k1k^OTNVa0l1?000000000000000000000000000O10000O100O100O10000O100O100O10000O101N100O1O1M3N3M2N3L3N3M2N3L4M2NU2kMc0\\O;F6J5K6I6K6J5K6I6K6YDXHl:n7oDWHk:n7REWHf:P8UEUHe:h8J7I6I8I6J6J7H7J6J7I6I8I6J4L3nJ[Ed0h:WO\\Eh0g:RO^El0e:nN`EP1c:jNbES1a:hNdEV1_:cNgE[1\\:_NiE5cM^Og<7kE8cM\\Of<6kE;fMZOa<6nE=fMYO_<4PF`0gMWO\\<3RFc0gMUO[<2SFf0hMSOW<1WFi0hMQOT<0YFl0W?L3N3M3M3M3M3M3L4M2NonP4"}, "label": "a half - sized frigde door"}]}
{"id": 575826, "image": "COCO_train2014_000000575826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt and blue shorts playing tennis\"."}], "task": "refering", "answer_template": "The \"man in white shirt and blue shorts playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 104, 105, 106, 108, 109, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338, 358, 360, 361, 362, 383, 384, 385], "bbox": [0.53428125, 0.17280318091451294, 0.794125, 0.9022266401590457], "iscrowd": 0, "area": 26735.32345, "rle": {"size": [503, 640], "counts": "iTX5<[?g0mBA_9R1jETOT:T1`ERO^:V1VEQOg:W1mDoNQ;Y1dDlNZ;j2N3M3M3M2O2M3QOWKkFl4j1RKd44`Il4j1SKc44_Im4k1QKe43^Io4j1QKe41_IR5i1oJbNHo4a0dJj4i1PK`NIk4c0kJg4g1oJ`NJk4c0lJg4f1oJ_NKk4b0oJe4e1PK]NMl4c0nJc4f1PK]NMk4a6eLdI]NNl4a6eLcI[NOo4_6eL_JZ3c5dL_JZ3c5dL^JZ3d5eL]JZ3e5dL\\J[3f5cLZJY3SN]Ke7Y1XJZ3SN]Kg7W1VJ\\3SN]Ki7U1TJ^3SN\\Kl7T1RJ_3RN]Km7S1QJ`3RN]Ko7Q1oIb3U6\\LkId3T6]LlIc3Q6`LPJ_3n5cLRJ]3l5dL_IPOTO\\4Z7gL_IWOPOR4_7iL^IS4`6nKXI[4f6gKYIZ4e6hKYIZ4g6eKYI\\4g6dKXI^4g6bKXI`4g6`KWId4g6[KYIg4e6ZKZIi4d6WK[Ik4d6TK]In4a6RK^IP5=bIR5^1aJR5:bIS5\\1cJU56bIW5X1cJX51dI\\5R1dJ\\5KfIb5l0bJa5FhIh5e0cJa6]5^IbJd6_5YIbJi6]5VIbJl6`5PIaJQ7a5lH_JU7c5gH]J[7e5bH[J_7g5]HYJe7P1kGT3<lKj7P1jGjN<a3LdLP8P1jG_Nh0n3YOcLV8o0^I_2d6`M]I]2e6bM^I[2c6dM_I[2a6eM`I[2_6dMcI]2[6bMgI^2X6bMiI^2V6aMlI_2S6`MoI`2P6`MQJ`2o5^MSJb2l5]MVJc2i5\\MYJd2f5\\M[Jd2d5[M^Jf2]NYLo5o0eKi2ZNYLQ6n0fKi2VNZLU6l0gKj2QN[LX6k0iKj2mM\\LV6k0QLi2fM]LX6j0ULi2`M^LZ6i0XLk2VMbLa6b0\\Lc3c3\\LaLb3_3]LdLa3[3_LhL_3X3`LkL_3S3bLoL\\3Q3cLRM[3m2fLVMU1ZLH`6ROYMT1ZLG\\6VO\\MP1\\LGX6XO_Mn0]LGS6\\ObMk0^LFP6^OfMh0^LGk5AjMe0^LHh5CmMb0]LId5FSN=[LLb4nNXKj0n38[LN_4ROUKh0U45YLO]4UOSKh0Z40YL1Z4YOPKf0a4LXL3W4\\OnJf0f4HVL5V4^OlJe0l4DUL7S4AiJe0R5@UL8P4DgJe0W5[OTL;n3FdJe0^5VORL=m3Q15_NoK?l3R18[NnKb0j3R1<XNlKe0f3T1Z2kNbMY1_2fN^M\\1e2bNZM^1h2aNVMa1k2^NTMb1n2]NSMa1n2_NSM`1n2_NSM_1o2`NRM^1o2bNRM]1o2bNRM\\1P3cNQM[1Q3dNRMX1o2hNUMS1m2lNVMP1l2nNWMo0l2oNSMQ1o2nNPMR1j80002N1O2M2O2N1O2N2M2O2N1O2M2O2N2M3O1N3M2N3M^fP2"}, "label": "man in white shirt and blue shorts playing tennis"}]}
{"id": 575826, "image": "COCO_train2014_000000575826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player in a white shirt about to hit the ball\"."}], "task": "refering", "answer_template": "The \"a tennis player in a white shirt about to hit the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 104, 105, 106, 108, 109, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338, 358, 360, 361, 362, 383, 384, 385], "bbox": [0.53428125, 0.17280318091451294, 0.794125, 0.9022266401590457], "iscrowd": 0, "area": 26735.32345, "rle": {"size": [503, 640], "counts": "iTX5<[?g0mBA_9R1jETOT:T1`ERO^:V1VEQOg:W1mDoNQ;Y1dDlNZ;j2N3M3M3M2O2M3QOWKkFl4j1RKd44`Il4j1SKc44_Im4k1QKe43^Io4j1QKe41_IR5i1oJbNHo4a0dJj4i1PK`NIk4c0kJg4g1oJ`NJk4c0lJg4f1oJ_NKk4b0oJe4e1PK]NMl4c0nJc4f1PK]NMk4a6eLdI]NNl4a6eLcI[NOo4_6eL_JZ3c5dL_JZ3c5dL^JZ3d5eL]JZ3e5dL\\J[3f5cLZJY3SN]Ke7Y1XJZ3SN]Kg7W1VJ\\3SN]Ki7U1TJ^3SN\\Kl7T1RJ_3RN]Km7S1QJ`3RN]Ko7Q1oIb3U6\\LkId3T6]LlIc3Q6`LPJ_3n5cLRJ]3l5dL_IPOTO\\4Z7gL_IWOPOR4_7iL^IS4`6nKXI[4f6gKYIZ4e6hKYIZ4g6eKYI\\4g6dKXI^4g6bKXI`4g6`KWId4g6[KYIg4e6ZKZIi4d6WK[Ik4d6TK]In4a6RK^IP5=bIR5^1aJR5:bIS5\\1cJU56bIW5X1cJX51dI\\5R1dJ\\5KfIb5l0bJa5FhIh5e0cJa6]5^IbJd6_5YIbJi6]5VIbJl6`5PIaJQ7a5lH_JU7c5gH]J[7e5bH[J_7g5]HYJe7P1kGT3<lKj7P1jGjN<a3LdLP8P1jG_Nh0n3YOcLV8o0^I_2d6`M]I]2e6bM^I[2c6dM_I[2a6eM`I[2_6dMcI]2[6bMgI^2X6bMiI^2V6aMlI_2S6`MoI`2P6`MQJ`2o5^MSJb2l5]MVJc2i5\\MYJd2f5\\M[Jd2d5[M^Jf2]NYLo5o0eKi2ZNYLQ6n0fKi2VNZLU6l0gKj2QN[LX6k0iKj2mM\\LV6k0QLi2fM]LX6j0ULi2`M^LZ6i0XLk2VMbLa6b0\\Lc3c3\\LaLb3_3]LdLa3[3_LhL_3X3`LkL_3S3bLoL\\3Q3cLRM[3m2fLVMU1ZLH`6ROYMT1ZLG\\6VO\\MP1\\LGX6XO_Mn0]LGS6\\ObMk0^LFP6^OfMh0^LGk5AjMe0^LHh5CmMb0]LId5FSN=[LLb4nNXKj0n38[LN_4ROUKh0U45YLO]4UOSKh0Z40YL1Z4YOPKf0a4LXL3W4\\OnJf0f4HVL5V4^OlJe0l4DUL7S4AiJe0R5@UL8P4DgJe0W5[OTL;n3FdJe0^5VORL=m3Q15_NoK?l3R18[NnKb0j3R1<XNlKe0f3T1Z2kNbMY1_2fN^M\\1e2bNZM^1h2aNVMa1k2^NTMb1n2]NSMa1n2_NSM`1n2_NSM_1o2`NRM^1o2bNRM]1o2bNRM\\1P3cNQM[1Q3dNRMX1o2hNUMS1m2lNVMP1l2nNWMo0l2oNSMQ1o2nNPMR1j80002N1O2M2O2N1O2N2M2O2N1O2M2O2N2M3O1N3M2N3M^fP2"}, "label": "a tennis player in a white shirt about to hit the ball"}]}
{"id": 371029, "image": "COCO_train2014_000000371029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep on the end furthest from the pole\"."}], "task": "refering", "answer_template": "The \"the sheep on the end furthest from the pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 116, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279], "bbox": [0.043546875, 0.2607291666666667, 0.248375, 0.7575000000000001], "iscrowd": 0, "area": 20852.3578, "rle": {"size": [480, 640], "counts": "iZ=m0P>4L4L5cMgNkF^1S9eNhF_1W9bNfFb1X9`NdFe1Y9\\NcFj1Z9XNbFn1Z9SNcFR2Z9PNcFU2Y9lMeFV2Z9kMdFX2Z9iMeFY2Z9fMeF\\2Z9eMdF^2Z9cMdF`2Z9aMdFb2Z9^MeFf2Y9ZMfFj2V9VMiFo2S9QMlFS3Q9mLoFV3n8jLQG[3?WLm6?cH_31]L\\73oG[44nKl7EkGj5T8a0O2O0O1O100O1O100O100000dNhHPKX7[4bI_K_6a4fIXK\\6i4aHbK1_O_7o4WHkK_8[5O2O00000000001O0000O100O1O1O1DdHWJY6V5VJnJi5S5SJQKm5o4lHXLT7g3dHcL[7]3]HkLc7Y50001O0001O1O010O1O001O1mNjHZJV7a5RI[JP7_5WI_Jj6[5aI^J`6\\5f1J8J6K2N2cKnEj3m0TLk72YGg3d0cLP8G\\Gd3;PMY8\\O\\Gd31ZMc8RO\\GV5d8jJ\\GV5d8jJ\\GV5d8iJ^GV5b8jJ^GV5a8kJ_GU5a8jJ`GV5`8jJ`GV5`8iJbGV5^8jJbGV5^8iJcGW5]8hJeGV5\\8iJeGV5]8iJcG^4HmKe8DeGX4Z9gKgFS4_9lKbFo3c9PL_F^3R:aLQF\\3P:dLRFX3P:gLRFW3o9hLTFT3o9jLSFT3n:N2M4M2M3N2M4L3M4L3L5L4L3L5H8H;E?[OmXQ7"}, "label": "the sheep on the end furthest from the pole"}]}
{"id": 371029, "image": "COCO_train2014_000000371029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep to the left of two other sheeps\"."}], "task": "refering", "answer_template": "The \"a sheep to the left of two other sheeps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 116, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279], "bbox": [0.043546875, 0.2607291666666667, 0.248375, 0.7575000000000001], "iscrowd": 0, "area": 20852.3578, "rle": {"size": [480, 640], "counts": "iZ=m0P>4L4L5cMgNkF^1S9eNhF_1W9bNfFb1X9`NdFe1Y9\\NcFj1Z9XNbFn1Z9SNcFR2Z9PNcFU2Y9lMeFV2Z9kMdFX2Z9iMeFY2Z9fMeF\\2Z9eMdF^2Z9cMdF`2Z9aMdFb2Z9^MeFf2Y9ZMfFj2V9VMiFo2S9QMlFS3Q9mLoFV3n8jLQG[3?WLm6?cH_31]L\\73oG[44nKl7EkGj5T8a0O2O0O1O100O1O100O100000dNhHPKX7[4bI_K_6a4fIXK\\6i4aHbK1_O_7o4WHkK_8[5O2O00000000001O0000O100O1O1O1DdHWJY6V5VJnJi5S5SJQKm5o4lHXLT7g3dHcL[7]3]HkLc7Y50001O0001O1O010O1O001O1mNjHZJV7a5RI[JP7_5WI_Jj6[5aI^J`6\\5f1J8J6K2N2cKnEj3m0TLk72YGg3d0cLP8G\\Gd3;PMY8\\O\\Gd31ZMc8RO\\GV5d8jJ\\GV5d8jJ\\GV5d8iJ^GV5b8jJ^GV5a8kJ_GU5a8jJ`GV5`8jJ`GV5`8iJbGV5^8jJbGV5^8iJcGW5]8hJeGV5\\8iJeGV5]8iJcG^4HmKe8DeGX4Z9gKgFS4_9lKbFo3c9PL_F^3R:aLQF\\3P:dLRFX3P:gLRFW3o9hLTFT3o9jLSFT3n:N2M4M2M3N2M4L3M4L3L5L4L3L5H8H;E?[OmXQ7"}, "label": "a sheep to the left of two other sheeps"}]}
{"id": 371029, "image": "COCO_train2014_000000371029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep between two others\"."}], "task": "refering", "answer_template": "The \"a sheep between two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286], "bbox": [0.284875, 0.31760416666666663, 0.55203125, 0.7489166666666667], "iscrowd": 0, "area": 20764.995699999996, "rle": {"size": [480, 640], "counts": "[ne21o>0O2O001O001N2O001O001O0O2O1O001O0jB`0k:BQEc0m:]OPEh0n:YOmDm0Q;TOkDQ1S;oNjDV1S;lNiDY1U;gNfD`1X;aNbDf1\\;[N]Dm1a;k0M3N2M3N2M4M2M3N2M3N2M3mFhKZ7Z4bHhK]7_4ZHdKf7b4RH`Km7f4lG\\KT8j4cGYK\\8m4]GUKc8Q5UGQKj8c500O100O100O100000000001O000oMaGaM_8`2UHkLl7T3hHXLX7h3P2010O0000010O0001O010O1O1O010O1O00100O1O001003M6I7H8H8I7H8H8I6I6J1000O10001N10000O101]NWGmLi8Q3_GhLb8U3fGfLZ8X3nGbLR8\\3UH^Ll7`3\\HZLe7c3bHXL^7f3o1N2N2O2O0O100O10001N100O10001N100ORFgLW8Y3VGYMk8f2cFmM^9j3000O10000O2O000O2O2N1O1F:E<D;F:E<J5O1N2N3M2N2N2N3M2N2L5H7I7H8I8H7I7I7I8GhZV4"}, "label": "a sheep between two others"}]}
{"id": 371029, "image": "COCO_train2014_000000371029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep with black eyes , eyes and mouth\"."}], "task": "refering", "answer_template": "The \"a sheep with black eyes , eyes and mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286], "bbox": [0.284875, 0.31760416666666663, 0.55203125, 0.7489166666666667], "iscrowd": 0, "area": 20764.995699999996, "rle": {"size": [480, 640], "counts": "[ne21o>0O2O001O001N2O001O001O0O2O1O001O0jB`0k:BQEc0m:]OPEh0n:YOmDm0Q;TOkDQ1S;oNjDV1S;lNiDY1U;gNfD`1X;aNbDf1\\;[N]Dm1a;k0M3N2M3N2M4M2M3N2M3N2M3mFhKZ7Z4bHhK]7_4ZHdKf7b4RH`Km7f4lG\\KT8j4cGYK\\8m4]GUKc8Q5UGQKj8c500O100O100O100000000001O000oMaGaM_8`2UHkLl7T3hHXLX7h3P2010O0000010O0001O010O1O1O010O1O00100O1O001003M6I7H8H8I7H8H8I6I6J1000O10001N10000O101]NWGmLi8Q3_GhLb8U3fGfLZ8X3nGbLR8\\3UH^Ll7`3\\HZLe7c3bHXL^7f3o1N2N2O2O0O100O10001N100O10001N100ORFgLW8Y3VGYMk8f2cFmM^9j3000O10000O2O000O2O2N1O1F:E<D;F:E<J5O1N2N3M2N2N2N3M2N2L5H7I7H8I8H7I7I7I8GhZV4"}, "label": "a sheep with black eyes , eyes and mouth"}]}
{"id": 371029, "image": "COCO_train2014_000000371029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lamb on far right\"."}], "task": "refering", "answer_template": "The \"lamb on far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 177, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 294, 295], "bbox": [0.7145937499999999, 0.23947916666666666, 0.9109375, 0.72875], "iscrowd": 0, "area": 17530.6412, "rle": {"size": [480, 640], "counts": "XYg61n>2O001O1O001O1O001O1^CNT:2kE4P:MmE;m9FPFa0k9_OSFi0g9XOWFn0d9SOYFU1a9lN\\F[1_9eN_Fc1[9^NcFg1Y9ZNdFk1Y9UNeFo1Y9RNeFR2X9oMeFV2X9jMfF[2W9fMfF_2W9bMgFb2V9bMdFc2U9oM[FU2\\9n1M7Ib0]O`0A1O1O0O2O1O0O2O1OhNZHVKf7c4gHXKW7b4VIYKi6a4cIYK]6`4PJ[Ko5c4YJXKe5i4Y200000O100000f1YN01O10O01O100O00SHjIR7Q7101N2O1N2N2O1N3N1N2O1N2O1N2O1bNnGkKT8R4XHdKi7KcGi3P1QL_76cGd3Z1lKT7?dGa3o9`LQF^3P:bLPF\\3R:dLnEZ3T:gLlEV3W:iLiEU3Y:lLfES3[:mLeEQ3]:PMbEo2_:QMaEn2a:QM_En2b:SM]El2d:TM\\Ek2e:TM\\El3d9SL^FS4\\9kKeF\\4T9cKmFc4m8\\KTGi4g8VKZGj4f8UK[Gk4f8SK[Gm4e8RK\\Gm4e8RK]Gm4c8RK`Gl4a8RKbGl4^8SKfGj4Z8UKiGn3]O`Lj8AmGd3ChLa8CoG\\3IoL`9f2jFXMX9c2nFZMU9a2oF]MS9_2QG^MS9\\2RGbMP9Z2o1J5L5J5gNUC7P=CaCLd<0Y1Kc`j0"}, "label": "lamb on far right"}]}
{"id": 412002, "image": "COCO_train2014_000000412002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on left\"."}], "task": "refering", "answer_template": "The \"woman on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 169, 170, 186, 187, 188, 204, 205, 206, 222, 223, 224, 240, 241, 242, 258, 259, 260, 276, 277, 278, 294, 295, 296, 312, 313, 314, 330, 331, 332, 348, 349, 350, 366, 367, 368], "bbox": [0.3307421875, 0.36890625, 0.49798828125, 0.910734375], "iscrowd": 0, "area": 20385.62805, "rle": {"size": [640, 512], "counts": "[TZ31kc06I6J7I6J7hMVORAo0d>^OSAi0b>DUAb0a>LUA;_>3XA3^>:ZALT<h2cC_MZ<U5O2M2N3N1N3M2O1N3N1N2N2O2M2O1N2N3E:A?B?L3N2N1O1O2O0O2N1O1O2N100O2N1O1O2O0O2N1O1O2O000100O1O2O0GdFlE]9R:fFlE\\9R:eFnE[9Q:gFmEZ9Q:iFnEW9P:?M2O1YGoEc63jHP:`0RF^66nHk9a0SFY6:SId9a0WFR6?YI\\9c0YFi5g0_IR9e0\\Fa5];[JhD[5`;[JlDZ5T=XMgA\\Ob>a0TBbNW>Y1aBjMi=g0j@O\\a0GR_O1Ra0E\\_O3h`0Cc_O8Vb0L4L4KeWP5"}, "label": "woman on left"}]}
{"id": 412002, "image": "COCO_train2014_000000412002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with brown hair and a black and white outfit standing between two staircases with bearded man\"."}], "task": "refering", "answer_template": "The \"a woman with brown hair and a black and white outfit standing between two staircases with bearded man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 169, 170, 186, 187, 188, 204, 205, 206, 222, 223, 224, 240, 241, 242, 258, 259, 260, 276, 277, 278, 294, 295, 296, 312, 313, 314, 330, 331, 332, 348, 349, 350, 366, 367, 368], "bbox": [0.3307421875, 0.36890625, 0.49798828125, 0.910734375], "iscrowd": 0, "area": 20385.62805, "rle": {"size": [640, 512], "counts": "[TZ31kc06I6J7I6J7hMVORAo0d>^OSAi0b>DUAb0a>LUA;_>3XA3^>:ZALT<h2cC_MZ<U5O2M2N3N1N3M2O1N3N1N2N2O2M2O1N2N3E:A?B?L3N2N1O1O2O0O2N1O1O2N100O2N1O1O2O0O2N1O1O2O000100O1O2O0GdFlE]9R:fFlE\\9R:eFnE[9Q:gFmEZ9Q:iFnEW9P:?M2O1YGoEc63jHP:`0RF^66nHk9a0SFY6:SId9a0WFR6?YI\\9c0YFi5g0_IR9e0\\Fa5];[JhD[5`;[JlDZ5T=XMgA\\Ob>a0TBbNW>Y1aBjMi=g0j@O\\a0GR_O1Ra0E\\_O3h`0Cc_O8Vb0L4L4KeWP5"}, "label": "a woman with brown hair and a black and white outfit standing between two staircases with bearded man"}]}
{"id": 502114, "image": "COCO_train2014_000000502114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl with a pink umbrell is walking\"."}], "task": "refering", "answer_template": "The \"a little girl with a pink umbrell is walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 296, 316, 318, 319], "bbox": [0.741171875, 0.5336533957845433, 0.8860625000000001, 0.8951990632318502], "iscrowd": 0, "area": 5146.145149999996, "rle": {"size": [427, 640], "counts": "dPV65U=10001O0000001O00O010000cMLbG5^8O^G1b82[GNe82[GNe81[G0d81ZG1f8OXG2P8DkF6R1=R8_OkF3k0g0W7ZORHOJ0h0l0\\7]OhGJ2N?V1f7YObGF7K20BY1b8YO`GE9HOk1W8lN\\GF;CKR2^8^OcG`NJW2b8]OaGc0LjMb8g1^G>5gM^8W4<N3M2N2\\OoFfLZ9Y3<O101N101N101O0O2O1O001O1O001O1O0011O01SNnE>R:AQF>o9_OTF`0l9_OWF?j9_OXFa0g9^O[Fa0e9]O_Fa0a9[OdFd0]9mNkEJn0e0lNK[:CWG>`NOY:_O^G>ZN3W<HkC8X<BjC>b<000000000O2dC]Om;e0nC^OR<d0jC^OV<l0001O2N3L3L5K3N0O2N2O0O2O1M3N1N3NbRn0"}, "label": "a little girl with a pink umbrell is walking"}]}
{"id": 502114, "image": "COCO_train2014_000000502114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child under pink umbrella\"."}], "task": "refering", "answer_template": "The \"child under pink umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 296, 316, 318, 319], "bbox": [0.741171875, 0.5336533957845433, 0.8860625000000001, 0.8951990632318502], "iscrowd": 0, "area": 5146.145149999996, "rle": {"size": [427, 640], "counts": "dPV65U=10001O0000001O00O010000cMLbG5^8O^G1b82[GNe82[GNe81[G0d81ZG1f8OXG2P8DkF6R1=R8_OkF3k0g0W7ZORHOJ0h0l0\\7]OhGJ2N?V1f7YObGF7K20BY1b8YO`GE9HOk1W8lN\\GF;CKR2^8^OcG`NJW2b8]OaGc0LjMb8g1^G>5gM^8W4<N3M2N2\\OoFfLZ9Y3<O101N101N101O0O2O1O001O1O001O1O0011O01SNnE>R:AQF>o9_OTF`0l9_OWF?j9_OXFa0g9^O[Fa0e9]O_Fa0a9[OdFd0]9mNkEJn0e0lNK[:CWG>`NOY:_O^G>ZN3W<HkC8X<BjC>b<000000000O2dC]Om;e0nC^OR<d0jC^OV<l0001O2N3L3L5K3N0O2N2O0O2O1M3N1N3NbRn0"}, "label": "child under pink umbrella"}]}
{"id": 272745, "image": "COCO_train2014_000000272745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white coloured cushion type sofa is placed on a carpet\"."}], "task": "refering", "answer_template": "The \"a white coloured cushion type sofa is placed on a carpet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 136, 137, 138, 153, 154, 155, 156, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309], "bbox": [0.0, 0.33723437500000003, 0.23458158995815898, 0.8290312500000001], "iscrowd": 0, "area": 28152.4595, "rle": {"size": [640, 478], "counts": "k7j0g6<b4V1`GV1_8V500O10O0100O10O0100O010O100O010O100O010M3M2N3M3M3M2N3M3O10O10001N1000000O101O0O1000000O2O00000O10001N1000000O101O000O2O001N2O001O1N101O1N101O1O0O2O1O001N4M5K5J5L5K4L5J5L5K4L5J5L5K4K8I8H7I7H8I7I8G8I7IU1kNi0VO8I8H7I8G9H7I;D>CeXT7"}, "label": "a white coloured cushion type sofa is placed on a carpet"}]}
{"id": 272745, "image": "COCO_train2014_000000272745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white armchair in a living room area\"."}], "task": "refering", "answer_template": "The \"a white armchair in a living room area\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 136, 137, 138, 153, 154, 155, 156, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309], "bbox": [0.0, 0.33723437500000003, 0.23458158995815898, 0.8290312500000001], "iscrowd": 0, "area": 28152.4595, "rle": {"size": [640, 478], "counts": "k7j0g6<b4V1`GV1_8V500O10O0100O10O0100O010O100O010O100O010M3M2N3M3M3M2N3M3O10O10001N1000000O101O0O1000000O2O00000O10001N1000000O101O000O2O001N2O001O1N101O1N101O1O0O2O1O001N4M5K5J5L5K4L5J5L5K4L5J5L5K4K8I8H7I7H8I7I8G8I7IU1kNi0VO8I8H7I8G9H7I;D>CeXT7"}, "label": "a white armchair in a living room area"}]}
{"id": 272745, "image": "COCO_train2014_000000272745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair closest to the dog\"."}], "task": "refering", "answer_template": "The \"the chair closest to the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 293, 306, 307, 308, 309], "bbox": [0.0, 0.323765625, 0.2457112970711297, 0.827921875], "iscrowd": 0, "area": 28951.628999999994, "rle": {"size": [640, 478], "counts": "o7Y4g?Z4eK0100000O10O100000O0100000O010000000YOeGhE[8W:RH^En7`:j0O1N1O2N2N101O10O0100000O01000O10O10O1000O0100000O010000O011O0O2O001O0O2O1O0O2O001N101O1N101O001N101O1N104L4K6K4L5K4K5L5K4K6K4L4K6K4L5J5L4L5K4K6K4L4K6K4L5J5L\\1dN8H8G8I8H7H9H7I8G8I7I7H9H7I7I7HhTQ7"}, "label": "the chair closest to the dog"}]}
{"id": 272745, "image": "COCO_train2014_000000272745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a blue chair beside a dog\"."}], "task": "refering", "answer_template": "The \"the back of a blue chair beside a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 293, 306, 307, 308, 309], "bbox": [0.0, 0.323765625, 0.2457112970711297, 0.827921875], "iscrowd": 0, "area": 28951.628999999994, "rle": {"size": [640, 478], "counts": "o7Y4g?Z4eK0100000O10O100000O0100000O010000000YOeGhE[8W:RH^En7`:j0O1N1O2N2N101O10O0100000O01000O10O10O1000O0100000O010000O011O0O2O001O0O2O1O0O2O001N101O1N101O001N101O1N104L4K6K4L5K4K5L5K4K6K4L4K6K4L5J5L4L5K4K6K4L4K6K4L5J5L\\1dN8H8G8I8H7H9H7I8G8I7I7H9H7I7I7HhTQ7"}, "label": "the back of a blue chair beside a dog"}]}
{"id": 256364, "image": "COCO_train2014_000000256364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train in the background\"."}], "task": "refering", "answer_template": "The \"the train in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.7028125000000001, 0.3550625, 1.0, 0.5617916666666667], "iscrowd": 0, "area": 14883.6939, "rle": {"size": [480, 640], "counts": "iSc6;e>>A`0A?A:E10001O000000000O1000001O000000000O10001O000000000O1000001O000000000O1000001O000000000O10001O000000000O1000001O000000000O1000001O00gN_Z4VOUgK?A`0@?WOTNoC\\2P<;O00001O0O1000000000000000O1000000O1000000O1000O10O1000000O1000000O10000O1000000O100000O101O001O001O001O001O001O001O001O001O001N101O001O001O00O1000000000000000000000O1000000000O1000000000000000000000000000O100fJ"}, "label": "the train in the background"}]}
{"id": 256364, "image": "COCO_train2014_000000256364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue train with pink doors on right side\"."}], "task": "refering", "answer_template": "The \"blue train with pink doors on right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.7028125000000001, 0.3550625, 1.0, 0.5617916666666667], "iscrowd": 0, "area": 14883.6939, "rle": {"size": [480, 640], "counts": "iSc6;e>>A`0A?A:E10001O000000000O1000001O000000000O10001O000000000O1000001O000000000O1000001O000000000O10001O000000000O1000001O000000000O1000001O00gN_Z4VOUgK?A`0@?WOTNoC\\2P<;O00001O0O1000000000000000O1000000O1000000O1000O10O1000000O1000000O10000O1000000O100000O101O001O001O001O001O001O001O001O001O001N101O001O001O00O1000000000000000000000O1000000000O1000000000000000000000000000O100fJ"}, "label": "blue train with pink doors on right side"}]}
{"id": 256364, "image": "COCO_train2014_000000256364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a train with a yellow front on the tracks\"."}], "task": "refering", "answer_template": "The \"there is a train with a yellow front on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268], "bbox": [0.15709375, 0.3119375, 0.71115625, 0.6813125], "iscrowd": 0, "area": 41773.57295, "rle": {"size": [480, 640], "counts": "aa_1;d>2O1O0O2O001N101N110N10001O001N101O001N101O001N10001O001N101O001N101O001O0O101O001N101O001N101O001O0O101O001N101O001N101O001O0O2O00001N101O001N101O001O0O2O00001N101O001O0O2O001O0O2O00001N101O001O0O2O001O0O2O00001N101O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O000O2O001O0O2O001O0O2O001O0O2O00001N101O001N101O001N101O000O2O001N101O001N101O001N10001O0O2O001O0O2O001O0O2O00001N101O001N101O0O2O001O0O101O001N101O001N101O001N10001O0O2O001O0O2O001O001O0O2O00001O001N1000000000000O1000000000000O1000000000000O100000000000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O001O001O001O001O001O001O001O1O001O001O001HlEbKT:]4PF`KQ:^49O2N1O2N2N1O2Nd0\\Ol0TOl0TOm0SOZ^f2"}, "label": "there is a train with a yellow front on the tracks"}]}
{"id": 256364, "image": "COCO_train2014_000000256364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train with a yellow front and blue and red sides is at a station\"."}], "task": "refering", "answer_template": "The \"a train with a yellow front and blue and red sides is at a station\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268], "bbox": [0.15709375, 0.3119375, 0.71115625, 0.6813125], "iscrowd": 0, "area": 41773.57295, "rle": {"size": [480, 640], "counts": "aa_1;d>2O1O0O2O001N101N110N10001O001N101O001N101O001N10001O001N101O001N101O001O0O101O001N101O001N101O001O0O101O001N101O001N101O001O0O2O00001N101O001N101O001O0O2O00001N101O001O0O2O001O0O2O00001N101O001O0O2O001O0O2O00001N101O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O000O2O001O0O2O001O0O2O001O0O2O00001N101O001N101O001N101O000O2O001N101O001N101O001N10001O0O2O001O0O2O001O0O2O00001N101O001N101O0O2O001O0O101O001N101O001N101O001N10001O0O2O001O0O2O001O001O0O2O00001O001N1000000000000O1000000000000O1000000000000O100000000000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O001O001O001O001O001O001O001O1O001O001O001HlEbKT:]4PF`KQ:^49O2N1O2N2N1O2Nd0\\Ol0TOl0TOm0SOZ^f2"}, "label": "a train with a yellow front and blue and red sides is at a station"}]}
{"id": 256364, "image": "COCO_train2014_000000256364.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train with the yellow front\"."}], "task": "refering", "answer_template": "The \"the train with the yellow front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268], "bbox": [0.15709375, 0.3119375, 0.71115625, 0.6813125], "iscrowd": 0, "area": 41773.57295, "rle": {"size": [480, 640], "counts": "aa_1;d>2O1O0O2O001N101N110N10001O001N101O001N101O001N10001O001N101O001N101O001O0O101O001N101O001N101O001O0O101O001N101O001N101O001O0O2O00001N101O001N101O001O0O2O00001N101O001O0O2O001O0O2O00001N101O001O0O2O001O0O2O00001N101O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O000O2O001O0O2O001O0O2O001O0O2O00001N101O001N101O001N101O000O2O001N101O001N101O001N10001O0O2O001O0O2O001O0O2O00001N101O001N101O0O2O001O0O101O001N101O001N101O001N10001O0O2O001O0O2O001O001O0O2O00001O001N1000000000000O1000000000000O1000000000000O100000000000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O001O001O001O001O001O001O001O1O001O001O001HlEbKT:]4PF`KQ:^49O2N1O2N2N1O2Nd0\\Ol0TOl0TOm0SOZ^f2"}, "label": "the train with the yellow front"}]}
{"id": 330094, "image": "COCO_train2014_000000330094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a pink sweater holding a blender\"."}], "task": "refering", "answer_template": "The \"a woman in a pink sweater holding a blender\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 39, 40, 41, 42, 61, 62, 63, 64, 65, 66, 84, 85, 86, 87, 88, 89, 108, 109, 110, 111, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 198, 201, 202, 203, 204, 205, 206, 221, 224, 225, 226, 227, 228, 229, 243, 244, 247, 248, 249, 250, 251, 252, 266, 267, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 310, 311, 315, 316, 317, 318, 319, 320, 321, 333, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 410, 411, 412], "bbox": [0.496484375, 0.02140529531568228, 1.0, 1.0], "iscrowd": 0, "area": 70330.40074999999, "rle": {"size": [491, 640], "counts": "]lh44Y>MbB6[=NbB5Y=NeB5W=OgB3V=OhB5V=KiB7U=JiB9U=HiB;V=DjB>T=BkB`0U=@hBLO4X=MbC1]<LhC3X<IlC7S<IoC6P<JQD6n;KRD5n;JSD6m;ITD6m;ITD7l;IUD5k;JWD6_NMR=LaD5^N1g>NXA3P?1O001O001OO1O11O000000001OO10000O10000000000O10eLIbG7\\8KdG5Z8MfG2Z8OfG1X81hGOV83jGMT84mGKR87nGGR8;nGCR8?fI@X4k0]K@\\4g0ZKCe4>PKMn44hJ7V5KeJ:Y5HhJ7W5JiJ6U5LkJ4R5OoJ0o41RKOk44VKKh47XKIe4:\\KEb4=^KC_4`0bK_O\\4c0dK]OZ4d0gK\\OX4e0iKZOU4h0kKXOS4j0nKUOQ4i0SLVOk3h0YLXOe3g0_LXO`3i0aLVO^3k0cLTO\\3l0eLTOZ3m0gLROW3P1jLoNU3Q1lLoNS3j0UMVOi2;iMDS20]N0`1BQO>k0VOEj0^700000000000000000^OQOhBo0X=QOhBo0X=ROgBn0Y=ROgBn0Y=ROgBn0Y=SOgBl0Y=UOfBk0X=WOgBj0V=ZOiBf0U=]OjBc0V=]OjBc0V=^OhBc0X=]OhBc0X=]OhBc0X=^OfBc0W2ROkMLR9NoGa0mNc0W2VOeMKV9OnG=POc0V2^Ol6EkG;ROb0W2_Oj6ElG;ROb0W2@i6DmG;SOa0W2@h6FlG;TO?W2Cf6DnG;TO?R2Ij6^OoG;UO>j12Q7UOoGd1i0DQ7jNTHc1h0IQ7dNVHd1e0OQ7]NZHe1a05R7VN[Hg13lMCZ2m7RN]Hi13oM]OZ2P8nM_Hk13QNZO]2n7gMbHn16i0V7YMcHP24k0X7UMbHR22n0Z7PMcHT21o0[7mLbHV22P1Y7jLcHY24n0W7jLcHY26W1m6`LhH_2:T1k6@UIa0i6@VIc0g6^OYIc0e6^O[Ic0c6^O]Ic0b6]O^Ic0b6]O^Ic0a6^O_Ic0`6]O`Ic0`6]O`Ic0_6^OaIc0^6]ObIc0^6]ObIc0]6^OcIc0\\6]OdIc0\\6]OdIc0[6^OeIb0l4gKSLg3QOb0g4lKXLb3QOb0JdKb4=aL^3SOa0IeK^4a0fLY3SOa0IfKX4m0dLl2[Oa0HhKT4Z1^L]2Fa0HhKo3_1cLX2Fa0GjKj3c1hLR2Ga0GkKc3i1oLj1Gc0ElK_3m1TMd1Hn0o2cMYM_1Ho0i2hM^MY1Jo0b2nMcMS1KP1\\2RNhMo0LP1V2WNmMi0MQ1P2\\NQNd00P1j1`NVN`00Q1e1dNWN>4o0_1iN[N96o0Z1mN_N48o0T1QOcN19Q1m0SOiNL:T1c0XOROD;X15^O_O[O<Z11]OBYO=^1K`35k500000000000000O1000000000000001O00000000001O00001O001O1O001O1O001O1O1O2N2N2N1O2N1O2N1O1O2N1O2N1kDkNR8W1iGmNV8T1iGnNU8S1jGnNU8T1[GeNTN7`:V1YGfNUN5a:W1WGfNWN3b:Y1UGfNXN1b:[1TGeNYN0c:]1lFjN`NId:_1_FTOlN^Od:`1[FVOQOYOd:f1TFCl9d0kE^OU:j0`EXOa:i20O1O10000000000O100001O001O1O001O1O1O001O00001O00001O00001O00001O001`LPLcKP4[4VLYInN;1S1l4V5dL_ITOi0X4g5d3O2O0O2N101N101N1O101N2O1N2N3N1N2O2M2N2N2N3L3N2N4K5L4L4K5L4L5J6I7I7I6J6J5K5KTI"}, "label": "a woman in a pink sweater holding a blender"}]}
{"id": 330094, "image": "COCO_train2014_000000330094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a pink shirt holds a blender\"."}], "task": "refering", "answer_template": "The \"a woman in a pink shirt holds a blender\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 39, 40, 41, 42, 61, 62, 63, 64, 65, 66, 84, 85, 86, 87, 88, 89, 108, 109, 110, 111, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 198, 201, 202, 203, 204, 205, 206, 221, 224, 225, 226, 227, 228, 229, 243, 244, 247, 248, 249, 250, 251, 252, 266, 267, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 310, 311, 315, 316, 317, 318, 319, 320, 321, 333, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 410, 411, 412], "bbox": [0.496484375, 0.02140529531568228, 1.0, 1.0], "iscrowd": 0, "area": 70330.40074999999, "rle": {"size": [491, 640], "counts": "]lh44Y>MbB6[=NbB5Y=NeB5W=OgB3V=OhB5V=KiB7U=JiB9U=HiB;V=DjB>T=BkB`0U=@hBLO4X=MbC1]<LhC3X<IlC7S<IoC6P<JQD6n;KRD5n;JSD6m;ITD6m;ITD7l;IUD5k;JWD6_NMR=LaD5^N1g>NXA3P?1O001O001OO1O11O000000001OO10000O10000000000O10eLIbG7\\8KdG5Z8MfG2Z8OfG1X81hGOV83jGMT84mGKR87nGGR8;nGCR8?fI@X4k0]K@\\4g0ZKCe4>PKMn44hJ7V5KeJ:Y5HhJ7W5JiJ6U5LkJ4R5OoJ0o41RKOk44VKKh47XKIe4:\\KEb4=^KC_4`0bK_O\\4c0dK]OZ4d0gK\\OX4e0iKZOU4h0kKXOS4j0nKUOQ4i0SLVOk3h0YLXOe3g0_LXO`3i0aLVO^3k0cLTO\\3l0eLTOZ3m0gLROW3P1jLoNU3Q1lLoNS3j0UMVOi2;iMDS20]N0`1BQO>k0VOEj0^700000000000000000^OQOhBo0X=QOhBo0X=ROgBn0Y=ROgBn0Y=ROgBn0Y=SOgBl0Y=UOfBk0X=WOgBj0V=ZOiBf0U=]OjBc0V=]OjBc0V=^OhBc0X=]OhBc0X=]OhBc0X=^OfBc0W2ROkMLR9NoGa0mNc0W2VOeMKV9OnG=POc0V2^Ol6EkG;ROb0W2_Oj6ElG;ROb0W2@i6DmG;SOa0W2@h6FlG;TO?W2Cf6DnG;TO?R2Ij6^OoG;UO>j12Q7UOoGd1i0DQ7jNTHc1h0IQ7dNVHd1e0OQ7]NZHe1a05R7VN[Hg13lMCZ2m7RN]Hi13oM]OZ2P8nM_Hk13QNZO]2n7gMbHn16i0V7YMcHP24k0X7UMbHR22n0Z7PMcHT21o0[7mLbHV22P1Y7jLcHY24n0W7jLcHY26W1m6`LhH_2:T1k6@UIa0i6@VIc0g6^OYIc0e6^O[Ic0c6^O]Ic0b6]O^Ic0b6]O^Ic0a6^O_Ic0`6]O`Ic0`6]O`Ic0_6^OaIc0^6]ObIc0^6]ObIc0]6^OcIc0\\6]OdIc0\\6]OdIc0[6^OeIb0l4gKSLg3QOb0g4lKXLb3QOb0JdKb4=aL^3SOa0IeK^4a0fLY3SOa0IfKX4m0dLl2[Oa0HhKT4Z1^L]2Fa0HhKo3_1cLX2Fa0GjKj3c1hLR2Ga0GkKc3i1oLj1Gc0ElK_3m1TMd1Hn0o2cMYM_1Ho0i2hM^MY1Jo0b2nMcMS1KP1\\2RNhMo0LP1V2WNmMi0MQ1P2\\NQNd00P1j1`NVN`00Q1e1dNWN>4o0_1iN[N96o0Z1mN_N48o0T1QOcN19Q1m0SOiNL:T1c0XOROD;X15^O_O[O<Z11]OBYO=^1K`35k500000000000000O1000000000000001O00000000001O00001O001O1O001O1O001O1O1O2N2N2N1O2N1O2N1O1O2N1O2N1kDkNR8W1iGmNV8T1iGnNU8S1jGnNU8T1[GeNTN7`:V1YGfNUN5a:W1WGfNWN3b:Y1UGfNXN1b:[1TGeNYN0c:]1lFjN`NId:_1_FTOlN^Od:`1[FVOQOYOd:f1TFCl9d0kE^OU:j0`EXOa:i20O1O10000000000O100001O001O1O001O1O1O001O00001O00001O00001O00001O001`LPLcKP4[4VLYInN;1S1l4V5dL_ITOi0X4g5d3O2O0O2N101N101N1O101N2O1N2N3N1N2O2M2N2N2N3L3N2N4K5L4L4K5L4L5J6I7I7I6J6J5K5KTI"}, "label": "a woman in a pink shirt holds a blender"}]}
{"id": 330094, "image": "COCO_train2014_000000330094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a blue cut off shirt smiling at a woman holding a blender\"."}], "task": "refering", "answer_template": "The \"man in a blue cut off shirt smiling at a woman holding a blender\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401], "bbox": [0.0, 0.0, 0.47992187499999994, 0.9865376782077393], "iscrowd": 0, "area": 109015.09229999999, "rle": {"size": [491, 640], "counts": "k9j0R=b1_Na1aN_1_Ob0]Ob0_Oa0_Oa0H9L3L4M4K4M3L4M4K4M3M4K4M3L4M4K4M3L4M4K4M3L5L3L4M3L5L3L4M4K4M3L4M4K4M3L4M2M100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O1gMXCn1i<RNZCj1g<VN\\Cf1e<ZN]Cc1d<]N_C_1b<aNaC[1`<eNcCX1]<hNkCo0V<QOQDg0P<YOWD?j;AaDMf;3i1000001O0000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000]DJQ86mGMR83lG0S80jG5T8KjG8U8HhG=V8ChG`0W8@gGc0X8]OeGh0Y8XOeGk0Z8UOcGP1[8POcGS1\\8mNbGV1]8jN`GZ1_8fN_G^1_8bN^Gb1a8^N]Ge1b8[N\\Gh1c8XNZGl1e8TNYGP2e8PNZGR2e8nMZGT2e8lMZGV2e8jMZGX2e8hMZG[2d8eM[G]2d8cM[G_2d8aM[Ga2d8_M\\Gb2c8^M\\Ge2b8[M]Gg2b8YM^Gh2a8XMhF0\\Oj2k9VMgF5[Oi2j9YMaF4Ah2j9ZM]F3Eg2k9[MYFf3c9U1L4L3M4L4L4L4L3M4cI`Ig3d6QLdIk3`6mKgIP4]6hKjIU4Y6eKnIW4V6aKQJ\\4S6\\KTJa4P6XKWJd4m5TKZJi4i5PK^Jm4f5kJaJR5P8L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M4L3M4L3M4L4L3M4L3M4L3M4L3M4L4L3M4L3M4L3M4L]Vo4"}, "label": "man in a blue cut off shirt smiling at a woman holding a blender"}]}
{"id": 330094, "image": "COCO_train2014_000000330094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a blue and gray sleeveless shirt\"."}], "task": "refering", "answer_template": "The \"man in a blue and gray sleeveless shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401], "bbox": [0.0, 0.0, 0.47992187499999994, 0.9865376782077393], "iscrowd": 0, "area": 109015.09229999999, "rle": {"size": [491, 640], "counts": "k9j0R=b1_Na1aN_1_Ob0]Ob0_Oa0_Oa0H9L3L4M4K4M3L4M4K4M3M4K4M3L4M4K4M3L4M4K4M3L5L3L4M3L5L3L4M4K4M3L4M4K4M3L4M2M100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O1gMXCn1i<RNZCj1g<VN\\Cf1e<ZN]Cc1d<]N_C_1b<aNaC[1`<eNcCX1]<hNkCo0V<QOQDg0P<YOWD?j;AaDMf;3i1000001O0000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000]DJQ86mGMR83lG0S80jG5T8KjG8U8HhG=V8ChG`0W8@gGc0X8]OeGh0Y8XOeGk0Z8UOcGP1[8POcGS1\\8mNbGV1]8jN`GZ1_8fN_G^1_8bN^Gb1a8^N]Ge1b8[N\\Gh1c8XNZGl1e8TNYGP2e8PNZGR2e8nMZGT2e8lMZGV2e8jMZGX2e8hMZG[2d8eM[G]2d8cM[G_2d8aM[Ga2d8_M\\Gb2c8^M\\Ge2b8[M]Gg2b8YM^Gh2a8XMhF0\\Oj2k9VMgF5[Oi2j9YMaF4Ah2j9ZM]F3Eg2k9[MYFf3c9U1L4L3M4L4L4L4L3M4cI`Ig3d6QLdIk3`6mKgIP4]6hKjIU4Y6eKnIW4V6aKQJ\\4S6\\KTJa4P6XKWJd4m5TKZJi4i5PK^Jm4f5kJaJR5P8L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M4L3M4L3M4L4L3M4L3M4L3M4L3M4L4L3M4L3M4L3M4L]Vo4"}, "label": "man in a blue and gray sleeveless shirt"}]}
{"id": 379246, "image": "COCO_train2014_000000379246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple truck with gated sides driving on a road following other trucks\"."}], "task": "refering", "answer_template": "The \"a purple truck with gated sides driving on a road following other trucks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123], "bbox": [0.60612, 0.37610878661087865, 0.9215399999999999, 0.7589958158995816], "iscrowd": 0, "area": 11662.7172, "rle": {"size": [239, 500], "counts": "kkV22Y76lN0^J9`5T1F4M1O0O2O0O1JUNfJn1X5501N10O0O101N100011N100N2O1O1N2O1O1O1O1000000010O0O1000000O2O00000O100010O2N2N2N1O10O00000000001O01O00000001O000001O0001O0000000001N10000000O100O1M3O100O10000000O10002N3M1O00000000001O00000000001O00000000001O000000000000000000000000O1O1M3L4H8L41O00IeJZN[5e1fJ[NZ5d1fJ^NY5a1hJ_NX5`1iJ`NW5`1iJ`NW5`1iJ`NW5`1iJaNV5_1jJaNV5_1jJaNV5^1kJbNU5^1kJbNU5^1kJcNT5]1lJ`NW5`1iJ]NZ5c1fJ[N\\5e1700O1000000VOTQ9"}, "label": "a purple truck with gated sides driving on a road following other trucks"}]}
{"id": 379246, "image": "COCO_train2014_000000379246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue truck which is third in a line of trucks\"."}], "task": "refering", "answer_template": "The \"a blue truck which is third in a line of trucks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123], "bbox": [0.60612, 0.37610878661087865, 0.9215399999999999, 0.7589958158995816], "iscrowd": 0, "area": 11662.7172, "rle": {"size": [239, 500], "counts": "kkV22Y76lN0^J9`5T1F4M1O0O2O0O1JUNfJn1X5501N10O0O101N100011N100N2O1O1N2O1O1O1O1000000010O0O1000000O2O00000O100010O2N2N2N1O10O00000000001O01O00000001O000001O0001O0000000001N10000000O100O1M3O100O10000000O10002N3M1O00000000001O00000000001O00000000001O000000000000000000000000O1O1M3L4H8L41O00IeJZN[5e1fJ[NZ5d1fJ^NY5a1hJ_NX5`1iJ`NW5`1iJ`NW5`1iJ`NW5`1iJaNV5_1jJaNV5_1jJaNV5^1kJbNU5^1kJbNU5^1kJcNT5]1lJ`NW5`1iJ]NZ5c1fJ[N\\5e1700O1000000VOTQ9"}, "label": "a blue truck which is third in a line of trucks"}]}
{"id": 493936, "image": "COCO_train2014_000000493936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front most green bus\"."}], "task": "refering", "answer_template": "The \"the front most green bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413], "bbox": [0.198671875, 0.004491869918699187, 1.0, 0.9932520325203252], "iscrowd": 0, "area": 224092.96215000004, "rle": {"size": [492, 640], "counts": "gjn174Jk>;RAEl>?N2M3N2N2M3N2N1O2M3dNYOPDi0h;EPD=h;1PD1h;=oCFj;R2I7I7H7J7H8^Ob0D<D<D<D<H8L4L3M4L4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L4L3M4M3L4L4L5K4L5L4K4L5K5K4L5K5L3L4J6I6L5N101O010O000010O0O1O2N10001O0000O10000O10000O10000O10000O1000000O10000O10000O100000001O000000000000000000000001O000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000O1000000O100000000O1000000O1000000000000000000001O0000000000000000000000000000001O0000000000000001O0000000000000001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000001O00000000000000000000000001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O01O0001O00001O0O10001N100mBUOe;l0YDYOb;h0]D\\O`;d0_DA];?bDFY;`0aDEk:\\O^DT1a0a0n9bN]ER1?_1n8gM^FP1>]2n7aNlGc2m6fMmH_4e1`JD]O"}, "label": "the front most green bus"}]}
{"id": 493936, "image": "COCO_train2014_000000493936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bus with a black tire\"."}], "task": "refering", "answer_template": "The \"a green bus with a black tire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413], "bbox": [0.198671875, 0.004491869918699187, 1.0, 0.9932520325203252], "iscrowd": 0, "area": 224092.96215000004, "rle": {"size": [492, 640], "counts": "gjn174Jk>;RAEl>?N2M3N2N2M3N2N1O2M3dNYOPDi0h;EPD=h;1PD1h;=oCFj;R2I7I7H7J7H8^Ob0D<D<D<D<H8L4L3M4L4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L4L3M4M3L4L4L5K4L5L4K4L5K5K4L5K5L3L4J6I6L5N101O010O000010O0O1O2N10001O0000O10000O10000O10000O10000O1000000O10000O10000O100000001O000000000000000000000001O000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000O1000000O100000000O1000000O1000000000000000000001O0000000000000000000000000000001O0000000000000001O0000000000000001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000001O00000000000000000000000001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O01O0001O00001O0O10001N100mBUOe;l0YDYOb;h0]D\\O`;d0_DA];?bDFY;`0aDEk:\\O^DT1a0a0n9bN]ER1?_1n8gM^FP1>]2n7aNlGc2m6fMmH_4e1`JD]O"}, "label": "a green bus with a black tire"}]}
{"id": 493936, "image": "COCO_train2014_000000493936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"double decker bus next to the closest bus\"."}], "task": "refering", "answer_template": "The \"double decker bus next to the closest bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 28, 29, 30, 51, 52, 53, 74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 165, 166, 188, 189, 210, 211, 212, 233, 234, 235, 256, 257, 279, 280, 303], "bbox": [0.16312500000000002, 0.024268292682926827, 0.35226562499999997, 0.7353252032520324], "iscrowd": 0, "area": 12679.288949999996, "rle": {"size": [492, 640], "counts": "[Wb16R?Q2oM;E5K5K5J5L5K5K5K4L5K5K5K4L5K5K5K5J5K6J6J6J5K6J6J6J5K6J3LM4L4K5L4L5K5K6I6K5K6J5K501O00010O00001O001O3M4L3N2M4L3M4L3M4L3M4L3N2M4L3M4L3M4L3M4L3M3N3L3M4L3M4M2N3N1O2N1O1O2M2O2N1O2N1N3N1O1O2N1N3N1O2N1O2N1N2O2N1O2N1N3N1O2N1O1O2M2O2N1O2N1N3N1O1OZ\\W6"}, "label": "double decker bus next to the closest bus"}]}
{"id": 493936, "image": "COCO_train2014_000000493936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the 2nd bus from the camera\"."}], "task": "refering", "answer_template": "The \"the 2nd bus from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 28, 29, 30, 51, 52, 53, 74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 165, 166, 188, 189, 210, 211, 212, 233, 234, 235, 256, 257, 279, 280, 303], "bbox": [0.16312500000000002, 0.024268292682926827, 0.35226562499999997, 0.7353252032520324], "iscrowd": 0, "area": 12679.288949999996, "rle": {"size": [492, 640], "counts": "[Wb16R?Q2oM;E5K5K5J5L5K5K5K4L5K5K5K4L5K5K5K5J5K6J6J6J5K6J6J6J5K6J3LM4L4K5L4L5K5K6I6K5K6J5K501O00010O00001O001O3M4L3N2M4L3M4L3M4L3M4L3N2M4L3M4L3M4L3M4L3M3N3L3M4L3M4M2N3N1O2N1O1O2M2O2N1O2N1N3N1O1O2N1N3N1O2N1O2N1N2O2N1O2N1N3N1O2N1O1O2M2O2N1O2N1N3N1O1OZ\\W6"}, "label": "the 2nd bus from the camera"}]}
{"id": 125298, "image": "COCO_train2014_000000125298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left suitcase\"."}], "task": "refering", "answer_template": "The \"the left suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 202], "bbox": [0.0, 0.20861333333333335, 0.37416000000000005, 0.9119733333333333], "iscrowd": 0, "area": 32748.075800000006, "rle": {"size": [375, 500], "counts": "i6b3o5V2L4M3L4M3M4K4M3L4M3M3L4M3M3L4N2O2N1O1O1O1N2O1O1O1000001N10000O1000000O10000O2O00000O10000O10000O101O000O10000O10000000001O01O000000000000000001O0000000000000001O0001O000000000000000000001O0000000001O0000000dMaM`K_2X4gNkJX1l4ROSKn0e4ZO[Kf0]4CbK=V4LiK4o34QLLg3=XLC`3f0_LZO\\3l0cLTO[3n0eLROZ3P1eLPOZ3R1fLnNW3T1iLlNV3V1iLjNV3X1iLhNU3Z1kLfNT3o4O1N2O1O1N2O1O1N2O1O1N3N10000O100O100O100O10001N100O101N10000O2O0O100O2O000O101O000O101O000O101O000O101O000O101O000O102N6J6I7J5K6A]`b3"}, "label": "the left suitcase"}]}
{"id": 125298, "image": "COCO_train2014_000000125298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"medium sized pink suitcase\"."}], "task": "refering", "answer_template": "The \"medium sized pink suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 202], "bbox": [0.0, 0.20861333333333335, 0.37416000000000005, 0.9119733333333333], "iscrowd": 0, "area": 32748.075800000006, "rle": {"size": [375, 500], "counts": "i6b3o5V2L4M3L4M3M4K4M3L4M3M3L4M3M3L4N2O2N1O1O1O1N2O1O1O1000001N10000O1000000O10000O2O00000O10000O10000O101O000O10000O10000000001O01O000000000000000001O0000000000000001O0001O000000000000000000001O0000000001O0000000dMaM`K_2X4gNkJX1l4ROSKn0e4ZO[Kf0]4CbK=V4LiK4o34QLLg3=XLC`3f0_LZO\\3l0cLTO[3n0eLROZ3P1eLPOZ3R1fLnNW3T1iLlNV3V1iLjNV3X1iLhNU3Z1kLfNT3o4O1N2O1O1N2O1O1N2O1O1N3N10000O100O100O100O10001N100O101N10000O2O0O100O2O000O101O000O101O000O101O000O101O000O101O000O102N6J6I7J5K6A]`b3"}, "label": "medium sized pink suitcase"}]}
{"id": 125298, "image": "COCO_train2014_000000125298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a suitcase that is the largest of three suitcases\"."}], "task": "refering", "answer_template": "The \"a suitcase that is the largest of three suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 32, 45, 46, 47, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.52922, 0.09146666666666665, 0.9876400000000001, 0.8779733333333334], "iscrowd": 0, "area": 46169.989299999994, "rle": {"size": [375, 500], "counts": "nTQ3`0f:b0^Ob0^Oa0_Ob0M3O1N2O0O2O1N2O1N2O0O:G1N101N101O001O010O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O0010O01O001O010O1O3M3M3N3L3M3M3M3M3N2M>lL_JVO[6OfI1T7TOmHl0k7ZNWHg1j7VNWHj1k7QNXHo1j7kMZHU2h7dM]H\\2e7^M_Hc2Z8000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O0000001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N3M4L5K4L5K4L5K5K4L5K4L5K4L5K4L5K4L5K4L5K4L5K5I6K=C=C=C=C>B=C=C>B=Cbl1"}, "label": "a suitcase that is the largest of three suitcases"}]}
{"id": 125298, "image": "COCO_train2014_000000125298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pink suitcase on the right\"."}], "task": "refering", "answer_template": "The \"pink suitcase on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 32, 45, 46, 47, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.52922, 0.09146666666666665, 0.9876400000000001, 0.8779733333333334], "iscrowd": 0, "area": 46169.989299999994, "rle": {"size": [375, 500], "counts": "nTQ3`0f:b0^Ob0^Oa0_Ob0M3O1N2O0O2O1N2O1N2O0O:G1N101N101O001O010O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O0010O01O001O010O1O3M3M3N3L3M3M3M3M3N2M>lL_JVO[6OfI1T7TOmHl0k7ZNWHg1j7VNWHj1k7QNXHo1j7kMZHU2h7dM]H\\2e7^M_Hc2Z8000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O0000001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N3M4L5K4L5K4L5K5K4L5K4L5K4L5K4L5K4L5K4L5K4L5K5I6K=C=C=C=C>B=C=C>B=Cbl1"}, "label": "pink suitcase on the right"}]}
{"id": 452979, "image": "COCO_train2014_000000452979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shadow of a black car\"."}], "task": "refering", "answer_template": "The \"a shadow of a black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305], "bbox": [0.5161875, 0.56875, 0.962625, 0.7739687500000001], "iscrowd": 0, "area": 16286.056250000001, "rle": {"size": [640, 480], "counts": "U^k43jc03M3L4M3M3L5L3M3M201000000O0100000O10O100000O100000O10O100000000O1000O10O1000O1O10O100000000O10O1000O100000000O010000000O100000O10O1000001N2O001O1O1N101N2O1N2O1N2O0O2O1N2O1N2O0O2O1N100O100O100O10O0100O100O100O10O01O100O100O101N2O1N3N2M3N1N3N2M3N1N3N2M2O2M3N2M2O2M2O0O100O010O100O10O010000O10O0100O10O10O10O010O010O10O0100O100O100O11O0000000O101O00000000001O000000001N10001O0000001O0000001O0O10001O0000001O000000000O10001OmNS1jNV1jNjm:"}, "label": "a shadow of a black car"}]}
{"id": 452979, "image": "COCO_train2014_000000452979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark suv\"."}], "task": "refering", "answer_template": "The \"the dark suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305], "bbox": [0.5161875, 0.56875, 0.962625, 0.7739687500000001], "iscrowd": 0, "area": 16286.056250000001, "rle": {"size": [640, 480], "counts": "U^k43jc03M3L4M3M3L5L3M3M201000000O0100000O10O100000O100000O10O100000000O1000O10O1000O1O10O100000000O10O1000O100000000O010000000O100000O10O1000001N2O001O1O1N101N2O1N2O1N2O0O2O1N2O1N2O0O2O1N100O100O100O10O0100O100O100O10O01O100O100O101N2O1N3N2M3N1N3N2M3N1N3N2M2O2M3N2M2O2M2O0O100O010O100O10O010000O10O0100O10O10O10O010O010O10O0100O100O100O11O0000000O101O00000000001O000000001N10001O0000001O0000001O0O10001O0000001O000000000O10001OmNS1jNV1jNjm:"}, "label": "the dark suv"}]}
{"id": 141682, "image": "COCO_train2014_000000141682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large tv showing a wii game\"."}], "task": "refering", "answer_template": "The \"a large tv showing a wii game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 39, 40, 41, 42, 51, 52, 53, 54, 55, 56, 57, 58, 59, 68, 69, 70, 71, 72, 73, 74, 75, 76, 85, 86, 87, 88, 89, 90, 91, 92, 93, 102, 103, 104, 105, 106, 107, 108, 109, 110, 119, 120, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 172], "bbox": [0.012041666666666668, 0.11456249999999998, 0.5267083333333333, 0.43735937499999994], "iscrowd": 0, "area": 45232.20129999999, "rle": {"size": [640, 480], "counts": "Yj3a1_b0U2kMU2lM5JO100O1O10000000000000000000000001O00000000000000000000001O0000000001O0000000002N5K5K3M00000000000000000000000000000000000000O10000000001O0000000000000000000000000000000000000000000000000000000000000000001O0O1000000000000000000000000000000000000000000000000000000000000000001O0000000000000O10000000000000000000000000000000000@ZBfJf=[5_B_Ja=b5cBYJ]=g5>1O001OK5_Ob0L3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000000000000000000000000000000001O00000000000000000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O00100O1OE;YLVm]4"}, "label": "a large tv showing a wii game"}]}
{"id": 141682, "image": "COCO_train2014_000000141682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"big tv screen\"."}], "task": "refering", "answer_template": "The \"big tv screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 39, 40, 41, 42, 51, 52, 53, 54, 55, 56, 57, 58, 59, 68, 69, 70, 71, 72, 73, 74, 75, 76, 85, 86, 87, 88, 89, 90, 91, 92, 93, 102, 103, 104, 105, 106, 107, 108, 109, 110, 119, 120, 121, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 172], "bbox": [0.012041666666666668, 0.11456249999999998, 0.5267083333333333, 0.43735937499999994], "iscrowd": 0, "area": 45232.20129999999, "rle": {"size": [640, 480], "counts": "Yj3a1_b0U2kMU2lM5JO100O1O10000000000000000000000001O00000000000000000000001O0000000001O0000000002N5K5K3M00000000000000000000000000000000000000O10000000001O0000000000000000000000000000000000000000000000000000000000000000001O0O1000000000000000000000000000000000000000000000000000000000000000001O0000000000000O10000000000000000000000000000000000@ZBfJf=[5_B_Ja=b5cBYJ]=g5>1O001OK5_Ob0L3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000000000000000000000000000000001O00000000000000000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O00100O1OE;YLVm]4"}, "label": "big tv screen"}]}
{"id": 223603, "image": "COCO_train2014_000000223603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a old woman sit and turn back\"."}], "task": "refering", "answer_template": "The \"a old woman sit and turn back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 70, 71, 72, 86, 87, 88, 89, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 154, 155, 156, 172, 173], "bbox": [0.06122916666666667, 0.16609374999999998, 0.4139791666666666, 0.45099999999999996], "iscrowd": 0, "area": 13584.991, "rle": {"size": [640, 480], "counts": "Rob06ic04L5K5J6K4L5J6K3M1O2SOeNQ_O]1l`0jNo^OX1m`0nNo^OT1m`0SOn^On0o`0YOk^Oj0Qa0T1M4L3M4M2O2O001O1N101O001N101O0O2O001O1N101O0O2O0O2H7I8I7H7I8J501N101O0O2O0O2O1O0O2O001N100O01O10O01O1O1O101N2N2N2N2UOWA\\Lk>c3VA\\Lk>b3WA\\Lk>b3WA\\Lk>c3VA[Lk>d3WAZLk>e3VAZLk>d3VA[Ll>d3UAZLm>d3UAZLm>e3TAYLn>e3TAYLn>f3SAYLn>e3TAYLm>`0TAP2j`0N2O1N2N101N2O1N101N2O1N2O0O2O1N2N101XOk]OBWb0=P^O\\OQb0c0V^OUOla0j0e001O1O1O001O1O1O1O0O2O1O1O001O1O1O000000000000000000000000000000O10O10000001O0000001O0000000010N2N1O2N2N2N2N2O1N1O2Ng]_5"}, "label": "a old woman sit and turn back"}]}
{"id": 223603, "image": "COCO_train2014_000000223603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older woman sitting down and looking to the left\"."}], "task": "refering", "answer_template": "The \"an older woman sitting down and looking to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 70, 71, 72, 86, 87, 88, 89, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 154, 155, 156, 172, 173], "bbox": [0.06122916666666667, 0.16609374999999998, 0.4139791666666666, 0.45099999999999996], "iscrowd": 0, "area": 13584.991, "rle": {"size": [640, 480], "counts": "Rob06ic04L5K5J6K4L5J6K3M1O2SOeNQ_O]1l`0jNo^OX1m`0nNo^OT1m`0SOn^On0o`0YOk^Oj0Qa0T1M4L3M4M2O2O001O1N101O001N101O0O2O001O1N101O0O2O0O2H7I8I7H7I8J501N101O0O2O0O2O1O0O2O001N100O01O10O01O1O1O101N2N2N2N2UOWA\\Lk>c3VA\\Lk>b3WA\\Lk>b3WA\\Lk>c3VA[Lk>d3WAZLk>e3VAZLk>d3VA[Ll>d3UAZLm>d3UAZLm>e3TAYLn>e3TAYLn>f3SAYLn>e3TAYLm>`0TAP2j`0N2O1N2N101N2O1N101N2O1N2O0O2O1N2N101XOk]OBWb0=P^O\\OQb0c0V^OUOla0j0e001O1O1O001O1O1O1O0O2O1O1O001O1O1O000000000000000000000000000000O10O10000001O0000001O0000000010N2N1O2N2N2N2N2O1N1O2Ng]_5"}, "label": "an older woman sitting down and looking to the left"}]}
{"id": 223603, "image": "COCO_train2014_000000223603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in a living room\"."}], "task": "refering", "answer_template": "The \"a man sitting in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 153, 154, 170, 171, 172, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 273, 274, 275, 291], "bbox": [0.0017708333333333332, 0.379421875, 0.21539583333333334, 0.750265625], "iscrowd": 0, "area": 17915.068950000004, "rle": {"size": [640, 480], "counts": "^n0m1ja0`1WLc20O010O0100O010O01000O010O10O010O0100O010O010O10O10O010O10O0100O010O010O100O100O10000O1000000000000000000000000000000000N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N101N2N2N2N2N2M3N2M3N2M3L5I8VNo_OZOX`0:^@UOj?>l@QO\\?b0X2DPZ[7"}, "label": "a man sitting in a living room"}]}
{"id": 223603, "image": "COCO_train2014_000000223603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"side of person wearing gray shirt with black longsleeve shirt under , partially seen to left of umbrella\"."}], "task": "refering", "answer_template": "The \"side of person wearing gray shirt with black longsleeve shirt under , partially seen to left of umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 153, 154, 170, 171, 172, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 273, 274, 275, 291], "bbox": [0.0017708333333333332, 0.379421875, 0.21539583333333334, 0.750265625], "iscrowd": 0, "area": 17915.068950000004, "rle": {"size": [640, 480], "counts": "^n0m1ja0`1WLc20O010O0100O010O01000O010O10O010O0100O010O010O10O10O010O10O0100O010O010O100O100O10000O1000000000000000000000000000000000N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N101N2N2N2N2N2M3N2M3N2M3L5I8VNo_OZOX`0:^@UOj?>l@QO\\?b0X2DPZ[7"}, "label": "side of person wearing gray shirt with black longsleeve shirt under , partially seen to left of umbrella"}]}
{"id": 223603, "image": "COCO_train2014_000000223603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl underneath umbrella\"."}], "task": "refering", "answer_template": "The \"girl underneath umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 165, 166, 167, 168, 182, 183, 184, 198, 199, 200, 201, 215, 216, 217, 218, 232, 233, 234, 235, 249, 250, 251, 252, 253, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 318, 319, 320], "bbox": [0.6363958333333334, 0.270171875, 0.9510833333333334, 0.8172499999999999], "iscrowd": 0, "area": 33223.6905, "rle": {"size": [640, 480], "counts": "Rfo55hc03L5L3M4gMJb@;[?M\\@6c?0U@4h?1Q@4l?OP@7k?KR@:j?HR@?j?CR@c0j?^OR@h0k?ZOQ@l0X;hMRI^1bKQ1W;dMUIQ4g6RLWIQ4f6QLWIS4e6QLWIS4e6PLWIT4f6oKWIU4f6mKWIV4f6nKVIV4f6mKUF[OW2m4b7iKQF@Z2j4c7iKoEAZ2i4f7gKmEDX2h4k7eKiEGX2h4m7_LiGj3V8WLdGo3[8SL`GQ4^8QL^GR4b8nK[GV4d8jKZGX4e8iKZGX4f8iKXGY4g8gKXGZ4g8gKXGZ4h8gKVG[4h8fKWG[4i8eKWG[4i8fKUG\\4i8eKVG\\4j8eKTG\\4k8fKSG\\4l8dKSG]4m8dKQG]4n8dKRG]4m8dKQG]4n8dKQG]4o8dKoF^4P9bKoFa4n8aKPGb4n8^KQGe4l8]KRGf4l8\\KPGf4P9\\KkFg4T9\\KgFh4X9P30000000001O00000O101O00000000001O00000000001O000000001O01O00000001O0000M3L4M3L5L3L4[KfFcN[9X1kFgNU9U1PGiNQ9R1UGmNk8nNYFnNQ1R2g8mNfFaNi0a2a8kNmF[Nh0i2[8kNPGVNj0m2X8lNPGPNn0S3R8mNQGkMR1V3o7mNRGfMV1[3i7nNSGaMY1`3e7nNeIP1^6nNdIP1]6oNdIo0^6oNdIP1^6mNeIP1]6nNeIQ1\\6mNgIQ1[6kNhIS1\\6hNhIV1\\6eNgIX1]6cNgI[1R7kMQIR2`<O2M3M3M3M2N3M3M3N2M4M3M3L5L3M3L5L3M4K4M3M4K4M3L5L3MZm="}, "label": "girl underneath umbrella"}]}
{"id": 223603, "image": "COCO_train2014_000000223603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl underneath a minnie mouse umbrella\"."}], "task": "refering", "answer_template": "The \"a little girl underneath a minnie mouse umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 165, 166, 167, 168, 182, 183, 184, 198, 199, 200, 201, 215, 216, 217, 218, 232, 233, 234, 235, 249, 250, 251, 252, 253, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 318, 319, 320], "bbox": [0.6363958333333334, 0.270171875, 0.9510833333333334, 0.8172499999999999], "iscrowd": 0, "area": 33223.6905, "rle": {"size": [640, 480], "counts": "Rfo55hc03L5L3M4gMJb@;[?M\\@6c?0U@4h?1Q@4l?OP@7k?KR@:j?HR@?j?CR@c0j?^OR@h0k?ZOQ@l0X;hMRI^1bKQ1W;dMUIQ4g6RLWIQ4f6QLWIS4e6QLWIS4e6PLWIT4f6oKWIU4f6mKWIV4f6nKVIV4f6mKUF[OW2m4b7iKQF@Z2j4c7iKoEAZ2i4f7gKmEDX2h4k7eKiEGX2h4m7_LiGj3V8WLdGo3[8SL`GQ4^8QL^GR4b8nK[GV4d8jKZGX4e8iKZGX4f8iKXGY4g8gKXGZ4g8gKXGZ4h8gKVG[4h8fKWG[4i8eKWG[4i8fKUG\\4i8eKVG\\4j8eKTG\\4k8fKSG\\4l8dKSG]4m8dKQG]4n8dKRG]4m8dKQG]4n8dKQG]4o8dKoF^4P9bKoFa4n8aKPGb4n8^KQGe4l8]KRGf4l8\\KPGf4P9\\KkFg4T9\\KgFh4X9P30000000001O00000O101O00000000001O00000000001O000000001O01O00000001O0000M3L4M3L5L3L4[KfFcN[9X1kFgNU9U1PGiNQ9R1UGmNk8nNYFnNQ1R2g8mNfFaNi0a2a8kNmF[Nh0i2[8kNPGVNj0m2X8lNPGPNn0S3R8mNQGkMR1V3o7mNRGfMV1[3i7nNSGaMY1`3e7nNeIP1^6nNdIP1]6oNdIo0^6oNdIP1^6mNeIP1]6nNeIQ1\\6mNgIQ1[6kNhIS1\\6hNhIV1\\6eNgIX1]6cNgI[1R7kMQIR2`<O2M3M3M3M2N3M3M3N2M4M3M3L5L3M3L5L3M4K4M3M4K4M3L5L3MZm="}, "label": "a little girl underneath a minnie mouse umbrella"}]}
{"id": 117117, "image": "COCO_train2014_000000117117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on near side of screen\"."}], "task": "refering", "answer_template": "The \"zebra on near side of screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 104, 105, 106, 107, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.0242, 0.2748358208955224, 0.9997, 0.9881492537313432], "iscrowd": 0, "area": 79196.95475, "rle": {"size": [335, 500], "counts": "QV42Z:5J5L5J6J6K4K6K5J6K4K6L4L3M4K5L4L2N3M2M3N2PNUMZLn2c3UMZLm2d3VMYLl2e3WMXLl2e3WMWLl2f3XMWLj2g3YMULj2i3YMRLl2l3VMoKn2P4TMkKP3T4RMfKR3Z4PMaKT3^4nL]KV3b4lLYKW3g4Y1O1N2O001O1O0000001O00000010O001N2K5K401O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O1O001O1O1O010O1O1O001O1O1O010O1O1O010O1O2N101N101N11O10O0100O010O100O010O1O10O0100O010O100O010O1O10O0100O010O100O010O1O10O0100O010O100O00100O1O010O1O00100O1O010O1O1O010O1O010O1O10O01O1O10O03M4M2M3M4L3N2M4L3N2M4L3N2M4L3M10O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O00001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O002N1O2N1O1O2N1O1O2N1O1O2N00000000O100000000000O10O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O100000000000000001O000000000000cK"}, "label": "zebra on near side of screen"}]}
{"id": 117117, "image": "COCO_train2014_000000117117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing in front of another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra standing in front of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 104, 105, 106, 107, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.0242, 0.2748358208955224, 0.9997, 0.9881492537313432], "iscrowd": 0, "area": 79196.95475, "rle": {"size": [335, 500], "counts": "QV42Z:5J5L5J6J6K4K6K5J6K4K6L4L3M4K5L4L2N3M2M3N2PNUMZLn2c3UMZLm2d3VMYLl2e3WMXLl2e3WMWLl2f3XMWLj2g3YMULj2i3YMRLl2l3VMoKn2P4TMkKP3T4RMfKR3Z4PMaKT3^4nL]KV3b4lLYKW3g4Y1O1N2O001O1O0000001O00000010O001N2K5K401O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O10O01O1O1O001O1O1O010O1O1O001O1O1O010O1O1O010O1O2N101N101N11O10O0100O010O100O010O1O10O0100O010O100O010O1O10O0100O010O100O010O1O10O0100O010O100O00100O1O010O1O00100O1O010O1O1O010O1O010O1O10O01O1O10O03M4M2M3M4L3N2M4L3N2M4L3N2M4L3M10O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O00001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O002N1O2N1O1O2N1O1O2N1O1O2N00000000O100000000000O10O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O100000000000000001O000000000000cK"}, "label": "a zebra standing in front of another zebra"}]}
{"id": 117117, "image": "COCO_train2014_000000117117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 20, 21, 22, 23, 24, 25, 26, 38, 39, 40, 41, 42, 43, 44, 45, 46, 51, 52, 53, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120], "bbox": [0.11992, 0.042716417910447765, 0.99826, 0.5147462686567165], "iscrowd": 0, "area": 32750.989149999994, "rle": {"size": [335, 500], "counts": "Vgc01\\:2M3N1O2N2M3N2N1O2M3N2N2M3L3M4L4L4L4L3M4L4O100O10O10O100O1000O010000O10O010000O10000O100O10000O2O000O100O10000O2O000O100O10000O2O00000O10000O101O0O10000O101O000100O100O1O2O0O1O100O1O2O0O10O00010O00001O01O01O0010O0001O01O01O000010O01O00010O000010O0001O010O00001O01O01O00010O001O01O01O00010O001O00010O001O01O01O0010O0001O010O00001O010O000010O01O00010O001O010O00001O010O000010O01O00010O001O01O01O001O01O100O100O1O10O0100O1O100O100O100O100O100O01000O10O10O01000O01O1O010O001O1O00100O001O1O001O10O01O001O1O01O00000000001O00000000001O00000000001O000000000O10O100000O100000000O100000000O10O100000O100000000O10000000000O10O100000O100000000O100000000O1000O1000O100000000O100000000O1000O1000O100000000O100000000O1000O1000O10O100000O01000000O10O100000O10O1000O1000O10O100000O01000000O0100000O10O100000O0100000O0100000O0100000O0100000O0100000O01000000O0100000O0100000O101O1O1N2O1O1O1N2O1lNX9"}, "label": "a zebra behind another zebra"}]}
{"id": 117117, "image": "COCO_train2014_000000117117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that can be partially seen behind another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra that can be partially seen behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 20, 21, 22, 23, 24, 25, 26, 38, 39, 40, 41, 42, 43, 44, 45, 46, 51, 52, 53, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120], "bbox": [0.11992, 0.042716417910447765, 0.99826, 0.5147462686567165], "iscrowd": 0, "area": 32750.989149999994, "rle": {"size": [335, 500], "counts": "Vgc01\\:2M3N1O2N2M3N2N1O2M3N2N2M3L3M4L4L4L4L3M4L4O100O10O10O100O1000O010000O10O010000O10000O100O10000O2O000O100O10000O2O000O100O10000O2O00000O10000O101O0O10000O101O000100O100O1O2O0O1O100O1O2O0O10O00010O00001O01O01O0010O0001O01O01O000010O01O00010O000010O0001O010O00001O01O01O00010O001O01O01O00010O001O00010O001O01O01O0010O0001O010O00001O010O000010O01O00010O001O010O00001O010O000010O01O00010O001O01O01O001O01O100O100O1O10O0100O1O100O100O100O100O100O01000O10O10O01000O01O1O010O001O1O00100O001O1O001O10O01O001O1O01O00000000001O00000000001O00000000001O000000000O10O100000O100000000O100000000O10O100000O100000000O10000000000O10O100000O100000000O100000000O1000O1000O100000000O100000000O1000O1000O100000000O100000000O1000O1000O10O100000O01000000O10O100000O10O1000O1000O10O100000O01000000O0100000O10O100000O0100000O0100000O0100000O0100000O0100000O01000000O0100000O0100000O101O1O1N2O1O1O1N2O1lNX9"}, "label": "a zebra that can be partially seen behind another zebra"}]}
{"id": 420221, "image": "COCO_train2014_000000420221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two slices of pizza in a warming tray\"."}], "task": "refering", "answer_template": "The \"two slices of pizza in a warming tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 337, 338, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.629297385620915, 1.0, 0.9998366013071895], "iscrowd": 0, "area": 50331.57749999999, "rle": {"size": [612, 612], "counts": "`<[1ia0001O1O001O1O001O001O1O001O001O1O001O1O010O001O1N101O0O2O1N101O1N101N10001N10000O100O2O000O100O2O000O1000001O00000000001O000O1000001O000000001O0000000O101O00000000001O00000000001N100000001O00000000001O000O100000001O00000000001O000O100000001O0001O00000001O000001O000001O00000001O0001O0000000000010O00000001O0001O01O0000001O01O01O00000010O000001O0000010O0000001O00010O0000001O0001O01O00000010O000001O01O01O9G1O100O1O001O100O001O100O1O001O100O1O001O100O001O00100O001O00100O001O010O1O001O010O1O001O10O01O001O10O01O00100O001O00100O001O1O010O1O001O100N101O1O001O1O1O001N2O001O1O001O1O1O0O2O1OO100O100O1O100O100O100O1006J1O00000001O1O001O1O1N2O1O1O001O1O1b_OhNn>Y1PAhNP?Y1n@iNQ?X1m@iNS?W1l@jNT?W1j@kNU?V1i@kNW?V1g@kNY?V1e@kN[?V1d@kN[?V1c@kN]?U1b@lN^?U1`@lN`?U1^@mNa?T1]@mNd?S1Z@nNf?S1X@nNh?S1V@oNi?Q1V@POj?Q1T@POl?Q1R@QOn?o0P@ROP`0o0n_OROR`0o0l_OROT`0o0j_OSOU`0m0j_OTOV`0m10010O00001O0001O01O00000010O0001O0001O01O00001O01O01O0000010O00001O0001O01O00000001O0000000010O001O000010O01O001O010O001O001O010O001O001O000O2O001O001O001O001O001N101O00000000000000000O10000000O1000000000O100000000000000000000O10001O00001O0004L5L1N10O1O2O0O1O2O0O2N100O0001O000O2O00001O00001O000O2O00001O00001O00001N10001O00001O00001N1000001O000000001O0O1000001O0000001O0O1000001O000000001O0O10000000000000000O100000000000001O000O100000000000000O100000000O1000001O000O10000000000O1000O1000O10000000000O10000000000O1000O100n_O"}, "label": "two slices of pizza in a warming tray"}]}
{"id": 420221, "image": "COCO_train2014_000000420221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate with 2 slices of pizza\"."}], "task": "refering", "answer_template": "The \"a plate with 2 slices of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 337, 338, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.629297385620915, 1.0, 0.9998366013071895], "iscrowd": 0, "area": 50331.57749999999, "rle": {"size": [612, 612], "counts": "`<[1ia0001O1O001O1O001O001O1O001O001O1O001O1O010O001O1N101O0O2O1N101O1N101N10001N10000O100O2O000O100O2O000O1000001O00000000001O000O1000001O000000001O0000000O101O00000000001O00000000001N100000001O00000000001O000O100000001O00000000001O000O100000001O0001O00000001O000001O000001O00000001O0001O0000000000010O00000001O0001O01O0000001O01O01O00000010O000001O0000010O0000001O00010O0000001O0001O01O00000010O000001O01O01O9G1O100O1O001O100O001O100O1O001O100O1O001O100O001O00100O001O00100O001O010O1O001O010O1O001O10O01O001O10O01O00100O001O00100O001O1O010O1O001O100N101O1O001O1O1O001N2O001O1O001O1O1O0O2O1OO100O100O1O100O100O100O1006J1O00000001O1O001O1O1N2O1O1O001O1O1b_OhNn>Y1PAhNP?Y1n@iNQ?X1m@iNS?W1l@jNT?W1j@kNU?V1i@kNW?V1g@kNY?V1e@kN[?V1d@kN[?V1c@kN]?U1b@lN^?U1`@lN`?U1^@mNa?T1]@mNd?S1Z@nNf?S1X@nNh?S1V@oNi?Q1V@POj?Q1T@POl?Q1R@QOn?o0P@ROP`0o0n_OROR`0o0l_OROT`0o0j_OSOU`0m0j_OTOV`0m10010O00001O0001O01O00000010O0001O0001O01O00001O01O01O0000010O00001O0001O01O00000001O0000000010O001O000010O01O001O010O001O001O010O001O001O000O2O001O001O001O001O001N101O00000000000000000O10000000O1000000000O100000000000000000000O10001O00001O0004L5L1N10O1O2O0O1O2O0O2N100O0001O000O2O00001O00001O000O2O00001O00001O00001N10001O00001O00001N1000001O000000001O0O1000001O0000001O0O1000001O000000001O0O10000000000000000O100000000000001O000O100000000000000O100000000O1000001O000O10000000000O1000O1000O10000000000O10000000000O1000O100n_O"}, "label": "a plate with 2 slices of pizza"}]}
{"id": 10621, "image": "COCO_train2014_000000010621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man hold and walking with horse\"."}], "task": "refering", "answer_template": "The \"the man hold and walking with horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 117, 118, 134, 135, 136, 137, 152, 153, 154, 155, 171, 172, 189, 190, 207, 208], "bbox": [0.48352, 0.45140672782874625, 0.66122, 0.9886850152905199], "iscrowd": 0, "area": 7073.208250000001, "rle": {"size": [327, 500], "counts": "m_]25m97J7H8I6J4N2N2N2M3N1iHbNe5X1dJkNV5S1mJROm4m0UKXOf4e0]K@^4>cKHX49gKLT45jKNT44aJmNc0P1k44SJBj0<Q5V1iJkNU5\\1eJeNY5a1aJaN^5d1]J^N`5h1[JZNb5W3N2N101N2N1B]KXKc4g4?00O1O100O1O100O1O1000XOcKfK^4o3PLmKQ4Q4ULjKl3T4U1N1O2N3M3M2O2`LdIP3^6oLcIW16NZ6iNbIU1;LV6lNaIU1a0HR6PO^IV1g0CX69TJ\\On5b0_JQOd5n0k1O101N101N1O101N100O2N101N101N2O1N101O1O01O1O0000O1O2N:Fhge1"}, "label": "the man hold and walking with horse"}]}
{"id": 10621, "image": "COCO_train2014_000000010621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leading a horse\"."}], "task": "refering", "answer_template": "The \"a man leading a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 117, 118, 134, 135, 136, 137, 152, 153, 154, 155, 171, 172, 189, 190, 207, 208], "bbox": [0.48352, 0.45140672782874625, 0.66122, 0.9886850152905199], "iscrowd": 0, "area": 7073.208250000001, "rle": {"size": [327, 500], "counts": "m_]25m97J7H8I6J4N2N2N2M3N1iHbNe5X1dJkNV5S1mJROm4m0UKXOf4e0]K@^4>cKHX49gKLT45jKNT44aJmNc0P1k44SJBj0<Q5V1iJkNU5\\1eJeNY5a1aJaN^5d1]J^N`5h1[JZNb5W3N2N101N2N1B]KXKc4g4?00O1O100O1O100O1O1000XOcKfK^4o3PLmKQ4Q4ULjKl3T4U1N1O2N3M3M2O2`LdIP3^6oLcIW16NZ6iNbIU1;LV6lNaIU1a0HR6PO^IV1g0CX69TJ\\On5b0_JQOd5n0k1O101N101N1O101N100O2N101N101N2O1N101O1O01O1O0000O1O2N:Fhge1"}, "label": "a man leading a horse"}]}
{"id": 403841, "image": "COCO_train2014_000000403841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women plaing the tennis in the tv\"."}], "task": "refering", "answer_template": "The \"a women plaing the tennis in the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 150, 151, 152, 173, 174, 175, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 378, 379, 380, 381, 382, 383], "bbox": [0.45714062499999997, 0.33797916666666666, 0.680859375, 1.0], "iscrowd": 0, "area": 29714.910100000005, "rle": {"size": [480, 640], "counts": "ohY44j>3L3M4L4L4M2L4K5L4K5L4K5L4K5L4K5K5L4K3M3M3N1N3M100O1O010^Oa0]Od0K52M2O10TF_L\\8b3bGQMm7o2PHeM^7\\2_HkM]7T2aHQN^7n1_HWNa7g1UHeNj7Y400O1O100O1O100O100O1ROaIZI`6\\6W1F:E<I6M3M3N2M3O1O1O1O1N2O1O10000000000000001O0O100000000001O0000001O00001O0000001O00001O0000001O00004L4L4L4L3M4L4L4L4L3M4L4L4L4L3M4L4L2N2N2N2N2G9A?_Oa0\\Oe0jMcFPNS:l1e1L3L4M3L4L4D<E<C<E;D<E;ES[o2"}, "label": "a women plaing the tennis in the tv"}]}
{"id": 403841, "image": "COCO_train2014_000000403841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women playing t v tennis with her friend\"."}], "task": "refering", "answer_template": "The \"a women playing t v tennis with her friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 150, 151, 152, 173, 174, 175, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 378, 379, 380, 381, 382, 383], "bbox": [0.45714062499999997, 0.33797916666666666, 0.680859375, 1.0], "iscrowd": 0, "area": 29714.910100000005, "rle": {"size": [480, 640], "counts": "ohY44j>3L3M4L4L4M2L4K5L4K5L4K5L4K5L4K5K5L4K3M3M3N1N3M100O1O010^Oa0]Od0K52M2O10TF_L\\8b3bGQMm7o2PHeM^7\\2_HkM]7T2aHQN^7n1_HWNa7g1UHeNj7Y400O1O100O1O100O100O1ROaIZI`6\\6W1F:E<I6M3M3N2M3O1O1O1O1N2O1O10000000000000001O0O100000000001O0000001O00001O0000001O00001O0000001O00004L4L4L4L3M4L4L4L4L3M4L4L4L4L3M4L4L2N2N2N2N2G9A?_Oa0\\Oe0jMcFPNS:l1e1L3L4M3L4L4D<E<C<E;D<E;ES[o2"}, "label": "a women playing t v tennis with her friend"}]}
{"id": 403841, "image": "COCO_train2014_000000403841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man\"."}], "task": "refering", "answer_template": "The \"a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 178, 179, 180, 181, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 363, 364, 365, 366, 386, 387, 388, 389, 390], "bbox": [0.755109375, 0.2947916666666667, 0.9676874999999999, 0.9871458333333333], "iscrowd": 0, "area": 27222.394800000002, "rle": {"size": [480, 640], "counts": "agR71n>2N3N2M2N3M3dJOdK2Y46aKK\\4`0[KAb4j0VKVOh4o0TKROl4R1oJoNP5U1lJlNT5X1gJhNZ5[1bJfN]5]1aJcN_5^1`JbN`5_1_JaN`5b1^J^Nb5c1]J]Nc5d1\\J\\Nc5f1\\JZNd5h1ZJWNg5j1WJWNh5k1WJUNi5l1VJTNi5n1VJRNj5o1UJQNk5P2TJPNk5R2TJnMl5R2TJmMm5T2RJlMn5U2nHXMZOd0g7V2kH[M\\O?i7W2fHaM^O8l7Y2aHdMA3n7Z2]HiMANo2FR1d2ZLmMBJe1R1_2X3iKeK`0\\2g3Q2fKdK<b2m3n1cKaK9h2T4i1aK_K4o2[4e1]K]K1U3b4\\60000000000001O001O00001O00001O001O1O=C>B=C1OO100O100O18H1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1N2O1O1O1N2O1O1O3M3L4M4L3M3L5L3M3M4K4M3M4L4K8I6J6J7H7J6J7bMgFgN]9m0oFnNU9k0SGQOQ9g0VGVOm8c0[GYOi8?_G\\Oe8=bG@b88fGD^84Xc9"}, "label": "a man"}]}
{"id": 403841, "image": "COCO_train2014_000000403841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on right\"."}], "task": "refering", "answer_template": "The \"person on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 178, 179, 180, 181, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 363, 364, 365, 366, 386, 387, 388, 389, 390], "bbox": [0.755109375, 0.2947916666666667, 0.9676874999999999, 0.9871458333333333], "iscrowd": 0, "area": 27222.394800000002, "rle": {"size": [480, 640], "counts": "agR71n>2N3N2M2N3M3dJOdK2Y46aKK\\4`0[KAb4j0VKVOh4o0TKROl4R1oJoNP5U1lJlNT5X1gJhNZ5[1bJfN]5]1aJcN_5^1`JbN`5_1_JaN`5b1^J^Nb5c1]J]Nc5d1\\J\\Nc5f1\\JZNd5h1ZJWNg5j1WJWNh5k1WJUNi5l1VJTNi5n1VJRNj5o1UJQNk5P2TJPNk5R2TJnMl5R2TJmMm5T2RJlMn5U2nHXMZOd0g7V2kH[M\\O?i7W2fHaM^O8l7Y2aHdMA3n7Z2]HiMANo2FR1d2ZLmMBJe1R1_2X3iKeK`0\\2g3Q2fKdK<b2m3n1cKaK9h2T4i1aK_K4o2[4e1]K]K1U3b4\\60000000000001O001O00001O00001O001O1O=C>B=C1OO100O100O18H1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1N2O1O1O1N2O1O1O3M3L4M4L3M3L5L3M3M4K4M3M4L4K8I6J6J7H7J6J7bMgFgN]9m0oFnNU9k0SGQOQ9g0VGVOm8c0[GYOi8?_G\\Oe8=bG@b88fGD^84Xc9"}, "label": "person on right"}]}
{"id": 280968, "image": "COCO_train2014_000000280968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bed layed on by man\"."}], "task": "refering", "answer_template": "The \"bed layed on by man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 183, 184, 185, 186], "bbox": [0.16984, 0.6457333333333334, 0.63228, 0.82064], "iscrowd": 0, "area": 5981.859999999997, "rle": {"size": [375, 500], "counts": "f]o01d;2O0O2O1N2N2O1N101N2N2O1N2O0O2O1N2N2O10O10000000O10O1000000000O010000000000O01000000000O01000000000O10O10000000O1000O100000O1000O100000O100000O1000O10000000O010000000000O01000000000O01000000000O10O10000000O1000O100000O1000O100000O100000O1000O10000000O10O10000000O10O1000000000O010000000000O0100O1O1N2O001N2000000O010000000000O100000O1000O10000000000O10O10000000O1000000000O010000000000O100000O1000O10000000000O1000O100000O10000000000O01000000000O10000L4K4L5K5KkTS2"}, "label": "bed layed on by man"}]}
{"id": 280968, "image": "COCO_train2014_000000280968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pink fabric under the man\"."}], "task": "refering", "answer_template": "The \"the pink fabric under the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 183, 184, 185, 186], "bbox": [0.16984, 0.6457333333333334, 0.63228, 0.82064], "iscrowd": 0, "area": 5981.859999999997, "rle": {"size": [375, 500], "counts": "f]o01d;2O0O2O1N2N2O1N101N2N2O1N2O0O2O1N2N2O10O10000000O10O1000000000O010000000000O01000000000O01000000000O10O10000000O1000O100000O1000O100000O100000O1000O10000000O010000000000O01000000000O01000000000O10O10000000O1000O100000O1000O100000O100000O1000O10000000O10O10000000O10O1000000000O010000000000O0100O1O1N2O001N2000000O010000000000O100000O1000O10000000000O10O10000000O1000000000O010000000000O100000O1000O10000000000O1000O100000O10000000000O01000000000O10000L4K4L5K5KkTS2"}, "label": "the pink fabric under the man"}]}
{"id": 280968, "image": "COCO_train2014_000000280968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed\"."}], "task": "refering", "answer_template": "The \"the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.0016799999999999999, 0.6091466666666667, 1.0, 0.9844266666666666], "iscrowd": 0, "area": 54265.99965000001, "rle": {"size": [375, 500], "counts": "he07`;9G9G9G9F6K000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1O1N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2M3L4L4L4N2000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O000000000000000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000001O000000000000000000000O10000000000000000000000000001O00000000000000000000000000000000000000000000000000001O0000000000000000000000001O2N2N2N3M2N2N2\\Od0YOg0YOh0YOf0YOeH"}, "label": "the bed"}]}
{"id": 280968, "image": "COCO_train2014_000000280968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheet in front of the reclining man , in the foreground of the image on the bed\"."}], "task": "refering", "answer_template": "The \"a white sheet in front of the reclining man , in the foreground of the image on the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.0016799999999999999, 0.6091466666666667, 1.0, 0.9844266666666666], "iscrowd": 0, "area": 54265.99965000001, "rle": {"size": [375, 500], "counts": "he07`;9G9G9G9F6K000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O1O1N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2M3L4L4L4N2000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O000000000000000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000001O000000000000000000000O10000000000000000000000000001O00000000000000000000000000000000000000000000000000001O0000000000000000000000001O2N2N2N3M2N2N2\\Od0YOg0YOh0YOf0YOeH"}, "label": "a white sheet in front of the reclining man , in the foreground of the image on the bed"}]}
{"id": 559497, "image": "COCO_train2014_000000559497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sweet potatoes\"."}], "task": "refering", "answer_template": "The \"sweet potatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 356, 357], "bbox": [0.485765625, 0.7047071129707113, 0.820921875, 0.9139121338912134], "iscrowd": 0, "area": 8999.105749999999, "rle": {"size": [478, 640], "counts": "Zba4`0\\>3N3L3M4M2M3M4L3N3L2N101N2N2N2O0O2N2O1N200O010O1O010O1O010O100O00100O0001O0O101O001N10001O0O101N101N100O2O0O101N101N10001O00001O00001O001O00001O01O101N101O1O001O0O10000000000000000O100000000000000000O10001N10000O10000O101O0O1000000O1000001O0000000O1000000000000O10000000O1000000000000000000000000000000000000000O1000O01000001N10000O100O100O10000O100O101N100O2O0O2O0O2O0O1O100O100O100O1O100O10O0100O0100O010001N2O1O2M2O1O1N2O2NkZe1"}, "label": "sweet potatoes"}]}
{"id": 559497, "image": "COCO_train2014_000000559497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small pile of leafy greens on the counter and directly below the switchplate\"."}], "task": "refering", "answer_template": "The \"small pile of leafy greens on the counter and directly below the switchplate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 299, 300, 301, 302], "bbox": [0.030203124999999997, 0.3476359832635983, 0.318890625, 0.8038075313807532], "iscrowd": 0, "area": 21399.780199999997, "rle": {"size": [478, 640], "counts": "SW91k>:F:F2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2M3N2N1O2N2N2M3N2N2N1O3M2M3N3M2N2N3L3N3M2N2M4M2N2N3L3N2N3bN]1O2N100O2O0O1O2O0O1O2O0O100O1O100O1O100O100O1O100O1O100O100O1O100O100O00N3M2N2N3M2N3N101N101O001N2O001O0O2O1O001N101O1O0O2O001O1N1O2O0O2N2N101N1O2O1N1O2N101N2N1O2O0O2N2N101N2N3M2O1N2N2O2M2N2M3N2N3M2N2N2N2N2M4M2N2N2N4Lb0_O`0@1O2N1O1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N1O2N1O2K4H[Z[6"}, "label": "small pile of leafy greens on the counter and directly below the switchplate"}]}
{"id": 559497, "image": "COCO_train2014_000000559497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leafy green that has an orange rubberband around it and is lying down on the table and not propped up with the other leafy greens\"."}], "task": "refering", "answer_template": "The \"a leafy green that has an orange rubberband around it and is lying down on the table and not propped up with the other leafy greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 299, 300, 301, 302], "bbox": [0.030203124999999997, 0.3476359832635983, 0.318890625, 0.8038075313807532], "iscrowd": 0, "area": 21399.780199999997, "rle": {"size": [478, 640], "counts": "SW91k>:F:F2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2M3N2N1O2N2N2M3N2N2N1O3M2M3N3M2N2N3L3N3M2N2M4M2N2N3L3N2N3bN]1O2N100O2O0O1O2O0O1O2O0O100O1O100O1O100O100O1O100O1O100O100O1O100O100O00N3M2N2N3M2N3N101N101O001N2O001O0O2O1O001N101O1O0O2O001O1N1O2O0O2N2N101N1O2O1N1O2N101N2N1O2O0O2N2N101N2N3M2O1N2N2O2M2N2M3N2N3M2N2N2N2N2M4M2N2N2N4Lb0_O`0@1O2N1O1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N1O2N1O2K4H[Z[6"}, "label": "a leafy green that has an orange rubberband around it and is lying down on the table and not propped up with the other leafy greens"}]}
{"id": 305550, "image": "COCO_train2014_000000305550.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy or girl in a black shirt and jeans playing with a wii remote\"."}], "task": "refering", "answer_template": "The \"a young boy or girl in a black shirt and jeans playing with a wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 109, 110, 111, 112, 126, 127, 128, 129, 144, 145, 146, 161, 162, 163, 178, 179, 180, 181, 182, 183, 197, 198, 199, 213, 214, 231, 246, 247, 248, 263, 264, 265, 280, 281, 282, 297, 298, 314, 315, 316, 331, 332, 333, 348, 349], "bbox": [0.4524375, 0.23109375000000001, 0.7880208333333333, 0.914234375], "iscrowd": 0, "area": 20282.72725, "rle": {"size": [640, 480], "counts": "RjW41lc02M6K5J6K5L4L4M3M3hGmN^MV1[2POcMS1W2ROgMQ1R2UOlMn0n1WOoMm0j1YOTNj0e1]OYNe0a1@]Nc0\\1CcN?W1FgN=R1ImN9m0LQO7h0OWO5XOY1g0nNhN^1U1iN\\Na1c1eNmMf1Q2aN`Mi1_2]NRMm1l2[NmLf1R3`NhLa1W3fNbL[1\\3lNeJgK6]5U5oNbJdK9^5S5POcJcK8_5T5nNdJcK8_5T5oNcJcK8_5S5POdJbK8^5T5QOcJaK9_5R5QOeJaK7_5T5POeJbK6_5S5POgJaK6_5S5QOfJaK6_5S5POgJaK6_5R5QOhJaK5^5S5QOhJbK3_5S5QOiJ`K4_5S5QOiJaK3^5S5ROjJ`K3_5R5QOkJaK2^5S5ROjJaK2]5S5SOkJ`K1_5S5QOlJaK0^5S5ROmJaKO[5V5UOjJ`K0V5\\5YOdJbKOQ5c5[O^JdKOm4h5^OYJfKMh4Q6AQJhKMc4W6DlIiKM^4]6HfIkKLZ4c6IaImKLZ4f6mL[Hn0R1lMLX4l6iLXHR1P1nMJW4S7cLVHW1m0PNI[2SOjNT8@UHZ1k0RNHV2]OhNP8@SH`1g0SNHP2GiNk7^ORHe1e0TNFk11iNf7YOVHn1=VNDf1=hN`7RO[HY24XNC`1g0hNb8`1SGYNB[1R1hNZ8c1RG[N_OW1_1gNQ8e1QG]N]OU1i1eNj7h1PG_NYOR1U2eNb7j1oF`NWOn0^2gN]7j1nFbNSOk0h2iNW7i1nFdNPOg0Q3lNR7g1mF:X2mMQ7d1fF:_2RNQ7]1`F=d2UNS7W1YF?j2ZNS7P1SFa0P3^NP7n0PF?V3cNj6m0PF;]3gNc6m0PF7c3kN^6m0oE3i3oNY6m0nE0n3QOV6n0lELT4UOQ6n0kEH[4WOl5P1iEDb4ZOf5Q1hE@U`0?k_O\\O[`0c0e_OXOa`0g0__OUOf`0j0Z_OQOl`0n0T1O1O1O1O1O001O1O1N2O1O1O1O1O1OO10000000000O1000000000000O0100000000O010O10000O1000O010000O1000O0100000000O10000000000O1000000N2M3M3MR_o1"}, "label": "a young boy or girl in a black shirt and jeans playing with a wii remote"}]}
{"id": 305550, "image": "COCO_train2014_000000305550.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a younger boy playing a nintendo wii game\"."}], "task": "refering", "answer_template": "The \"a younger boy playing a nintendo wii game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 109, 110, 111, 112, 126, 127, 128, 129, 144, 145, 146, 161, 162, 163, 178, 179, 180, 181, 182, 183, 197, 198, 199, 213, 214, 231, 246, 247, 248, 263, 264, 265, 280, 281, 282, 297, 298, 314, 315, 316, 331, 332, 333, 348, 349], "bbox": [0.4524375, 0.23109375000000001, 0.7880208333333333, 0.914234375], "iscrowd": 0, "area": 20282.72725, "rle": {"size": [640, 480], "counts": "RjW41lc02M6K5J6K5L4L4M3M3hGmN^MV1[2POcMS1W2ROgMQ1R2UOlMn0n1WOoMm0j1YOTNj0e1]OYNe0a1@]Nc0\\1CcN?W1FgN=R1ImN9m0LQO7h0OWO5XOY1g0nNhN^1U1iN\\Na1c1eNmMf1Q2aN`Mi1_2]NRMm1l2[NmLf1R3`NhLa1W3fNbL[1\\3lNeJgK6]5U5oNbJdK9^5S5POcJcK8_5T5nNdJcK8_5T5oNcJcK8_5S5POdJbK8^5T5QOcJaK9_5R5QOeJaK7_5T5POeJbK6_5S5POgJaK6_5S5QOfJaK6_5S5POgJaK6_5R5QOhJaK5^5S5QOhJbK3_5S5QOiJ`K4_5S5QOiJaK3^5S5ROjJ`K3_5R5QOkJaK2^5S5ROjJaK2]5S5SOkJ`K1_5S5QOlJaK0^5S5ROmJaKO[5V5UOjJ`K0V5\\5YOdJbKOQ5c5[O^JdKOm4h5^OYJfKMh4Q6AQJhKMc4W6DlIiKM^4]6HfIkKLZ4c6IaImKLZ4f6mL[Hn0R1lMLX4l6iLXHR1P1nMJW4S7cLVHW1m0PNI[2SOjNT8@UHZ1k0RNHV2]OhNP8@SH`1g0SNHP2GiNk7^ORHe1e0TNFk11iNf7YOVHn1=VNDf1=hN`7RO[HY24XNC`1g0hNb8`1SGYNB[1R1hNZ8c1RG[N_OW1_1gNQ8e1QG]N]OU1i1eNj7h1PG_NYOR1U2eNb7j1oF`NWOn0^2gN]7j1nFbNSOk0h2iNW7i1nFdNPOg0Q3lNR7g1mF:X2mMQ7d1fF:_2RNQ7]1`F=d2UNS7W1YF?j2ZNS7P1SFa0P3^NP7n0PF?V3cNj6m0PF;]3gNc6m0PF7c3kN^6m0oE3i3oNY6m0nE0n3QOV6n0lELT4UOQ6n0kEH[4WOl5P1iEDb4ZOf5Q1hE@U`0?k_O\\O[`0c0e_OXOa`0g0__OUOf`0j0Z_OQOl`0n0T1O1O1O1O1O001O1O1N2O1O1O1O1O1OO10000000000O1000000000000O0100000000O010O10000O1000O010000O1000O0100000000O10000000000O1000000N2M3M3MR_o1"}, "label": "a younger boy playing a nintendo wii game"}]}
{"id": 305550, "image": "COCO_train2014_000000305550.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a gray top playing wii\"."}], "task": "refering", "answer_template": "The \"a woman in a gray top playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 73, 74, 75, 76, 89, 90, 91, 92, 93, 106, 107, 108, 109, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 208, 209, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 344, 345, 346, 347, 348, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382], "bbox": [0.2326666666666667, 0.1401875, 0.6800833333333334, 0.985828125], "iscrowd": 0, "area": 60773.296999999984, "rle": {"size": [640, 480], "counts": "o[V2?Wc0`0_Oa0_Oa0D;J7I7I7I7I6I8nKBUDe0c;CUDe0d;BUDe0c;R4kHhGW2_8WMmGh2Z8eLSHY3U8ULWHj3P8dK\\HZ4k7UKaHf3PNcKf95fHU4gMeKj9DkHe4^MdKo9UOoHT5VMeKU=X4oBeKS=Y4PCeKQ=X4SCfKn<W4VCfKl<X4VCgKk<V4YChKh<U4\\CjKd<S4`CkK`<T4cCkK]<R4gClKZ<R4iCmKW<P4lCoKU<o3nCPLR<m3RDQLo;m3TDRLl;k3V2N2M3M3N2M3N2M3N2N200O10000000O10000000000000O10000000000000000000000O100000O100000000000000000000000000000000SBkLo:U3PEoLm:Q3RESMk:m2TEWMi:i2WEZMf:f2YE[Mf:f2YE[Mg:e2YE\\Mf:e2XE\\Mh:d2WE]Mi:c2VE_Mi:b2VE^Mj:b2UE_Mk:b2SE`Ml:`2TEaMk:_2TEcMk:^2SEcMm:]2REeMm:\\2QEfMn2oNg2\\3YJgMP3mNf2]3XJhMQ3mNR13TKX3h0jMR3kNf0f0XKf2n0nMP3hN<X1^KS2T1SNn2bN4k1bKa1Z1XNk2^NL\\2gKo0`1]Ni2XNCP3lK=>bL<Q2b3RN[OW7_LQI9n1i3jMSOg7bLhH;P2l3bMiNP9SMgIe3TM[L6c2c<m0iCiNd<S1\\CdNQ=4YBjNf0i0^=9lAnNg0?j=>_ATOg05T>e0VAVO?nNEm0P?m0m@XO<oNHc0Z?T1b@ZO:POL8b?\\1Y@[O8SOMMl?c1P@]O5TOY`0]1b_O1_`0Na_O1b`0L__O3c`0K^_O5c`0J]_O5e`0I\\_O6f`0HZ_O9h`0EX_O:j`0DW_O;j`0EV_O;i`0EX_O:i`0FV_O:j`0FW_O:i`0EX_O:h`0GX_O9h`0FY_O9g`0HY_O7h`0HZ_O7e`0J]_O3d`0L^_O2b`0N`_O1``0Oa_OO_`01c_OM^`03c_OL\\`04f_OJ[`05n_OCQ`0=W@YOk?g0h1O0000000O101O00000000000O2O00000000001N10000000001O0O10000000001O0O100000001O0O100000001N2O1O1O1O1N2M4Kmmo2"}, "label": "a woman in a gray top playing wii"}]}
{"id": 305550, "image": "COCO_train2014_000000305550.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a gray top and black pants , playing a video game\"."}], "task": "refering", "answer_template": "The \"a woman in a gray top and black pants , playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 73, 74, 75, 76, 89, 90, 91, 92, 93, 106, 107, 108, 109, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 208, 209, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 344, 345, 346, 347, 348, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382], "bbox": [0.2326666666666667, 0.1401875, 0.6800833333333334, 0.985828125], "iscrowd": 0, "area": 60773.296999999984, "rle": {"size": [640, 480], "counts": "o[V2?Wc0`0_Oa0_Oa0D;J7I7I7I7I6I8nKBUDe0c;CUDe0d;BUDe0c;R4kHhGW2_8WMmGh2Z8eLSHY3U8ULWHj3P8dK\\HZ4k7UKaHf3PNcKf95fHU4gMeKj9DkHe4^MdKo9UOoHT5VMeKU=X4oBeKS=Y4PCeKQ=X4SCfKn<W4VCfKl<X4VCgKk<V4YChKh<U4\\CjKd<S4`CkK`<T4cCkK]<R4gClKZ<R4iCmKW<P4lCoKU<o3nCPLR<m3RDQLo;m3TDRLl;k3V2N2M3M3N2M3N2M3N2N200O10000000O10000000000000O10000000000000000000000O100000O100000000000000000000000000000000SBkLo:U3PEoLm:Q3RESMk:m2TEWMi:i2WEZMf:f2YE[Mf:f2YE[Mg:e2YE\\Mf:e2XE\\Mh:d2WE]Mi:c2VE_Mi:b2VE^Mj:b2UE_Mk:b2SE`Ml:`2TEaMk:_2TEcMk:^2SEcMm:]2REeMm:\\2QEfMn2oNg2\\3YJgMP3mNf2]3XJhMQ3mNR13TKX3h0jMR3kNf0f0XKf2n0nMP3hN<X1^KS2T1SNn2bN4k1bKa1Z1XNk2^NL\\2gKo0`1]Ni2XNCP3lK=>bL<Q2b3RN[OW7_LQI9n1i3jMSOg7bLhH;P2l3bMiNP9SMgIe3TM[L6c2c<m0iCiNd<S1\\CdNQ=4YBjNf0i0^=9lAnNg0?j=>_ATOg05T>e0VAVO?nNEm0P?m0m@XO<oNHc0Z?T1b@ZO:POL8b?\\1Y@[O8SOMMl?c1P@]O5TOY`0]1b_O1_`0Na_O1b`0L__O3c`0K^_O5c`0J]_O5e`0I\\_O6f`0HZ_O9h`0EX_O:j`0DW_O;j`0EV_O;i`0EX_O:i`0FV_O:j`0FW_O:i`0EX_O:h`0GX_O9h`0FY_O9g`0HY_O7h`0HZ_O7e`0J]_O3d`0L^_O2b`0N`_O1``0Oa_OO_`01c_OM^`03c_OL\\`04f_OJ[`05n_OCQ`0=W@YOk?g0h1O0000000O101O00000000000O2O00000000001N10000000001O0O10000000001O0O100000001O0O100000001N2O1O1O1O1N2M4Kmmo2"}, "label": "a woman in a gray top and black pants , playing a video game"}]}
{"id": 2448, "image": "COCO_train2014_000000002448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kitchen chair in the foreground on the right\"."}], "task": "refering", "answer_template": "The \"a kitchen chair in the foreground on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.390265625, 0.75525, 0.63115625, 0.9952083333333334], "iscrowd": 0, "area": 11459.286400000003, "rle": {"size": [480, 640], "counts": "bae34k>5L3L100000010O0001O000010O0001O01eBHl;9RD2d;OYD4f;MXD6g;JWD8h;OoC4P<8cCJ\\<_1N100O001O1O010SDdMT;]2eDiM[;X2_DmMa;S2]DoMc;R2[DoMf;P2XDRNh;c20000001O000000001O000000001O000000001O000000001OO100000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000O1000000O1000000O1000000O10000O10ZOZDSNe;l1iDgMV;Z2h0O1O010O010O1O001N2SO`CWOb<;TD\\On;9cD_O_;>b1N3N1O200O010O1O001O1O1O1N2N2KjX^3"}, "label": "a kitchen chair in the foreground on the right"}]}
{"id": 2448, "image": "COCO_train2014_000000002448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the right\"."}], "task": "refering", "answer_template": "The \"the chair on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.390265625, 0.75525, 0.63115625, 0.9952083333333334], "iscrowd": 0, "area": 11459.286400000003, "rle": {"size": [480, 640], "counts": "bae34k>5L3L100000010O0001O000010O0001O01eBHl;9RD2d;OYD4f;MXD6g;JWD8h;OoC4P<8cCJ\\<_1N100O001O1O010SDdMT;]2eDiM[;X2_DmMa;S2]DoMc;R2[DoMf;P2XDRNh;c20000001O000000001O000000001O000000001O000000001OO100000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000O1000000O1000000O1000000O10000O10ZOZDSNe;l1iDgMV;Z2h0O1O010O010O1O001N2SO`CWOb<;TD\\On;9cD_O_;>b1N3N1O200O010O1O001O1O1O1N2N2KjX^3"}, "label": "the chair on the right"}]}
{"id": 2448, "image": "COCO_train2014_000000002448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair under bar to left of other chair\"."}], "task": "refering", "answer_template": "The \"chair under bar to left of other chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [300, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375], "bbox": [0.067109375, 0.7608333333333334, 0.31879687500000004, 0.9890208333333333], "iscrowd": 0, "area": 12105.043500000005, "rle": {"size": [480, 640], "counts": "]`d0>c>0O0010O01O010O0010O01O010O0010O010O02N6K5J6J6K4K6QCfNl;`1dCjN]<m1O00100O0010O01O010nC`Mb;`2\\DbMd;_2YDcMh;]2UDeMk;[2SDgMm;f2000N2K5N20000001O000000000000000000000000001O00000000000000007I00O100000000N2O10000000O1000O10000000000O0100000000000O1000O100000O100002N1O1O2M2O1OO1GTD^Ml;a2VD^Mj;b2WD\\Mj;c2YD[Mf;f2[DYMe;f2;0000O010O100O010O1000O0100O100O010O14L3L5LnMSDU1i;gN\\D\\1`;aNdD`1Z;\\NkDc1T;ZNQEe1o:XNUEf1l:ZNTEf1l:YNVEi0VOFc;BXEf0[OC];FZEe0@_OV;L[Eb0F]Oo:0^E`0I[Oh:6`E<OYOa::bE;c;E^D9c;F_D7c;I^DMj;3UR\\6"}, "label": "chair under bar to left of other chair"}]}
{"id": 2448, "image": "COCO_train2014_000000002448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black colur chair\"."}], "task": "refering", "answer_template": "The \"a black colur chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [300, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375], "bbox": [0.067109375, 0.7608333333333334, 0.31879687500000004, 0.9890208333333333], "iscrowd": 0, "area": 12105.043500000005, "rle": {"size": [480, 640], "counts": "]`d0>c>0O0010O01O010O0010O01O010O0010O010O02N6K5J6J6K4K6QCfNl;`1dCjN]<m1O00100O0010O01O010nC`Mb;`2\\DbMd;_2YDcMh;]2UDeMk;[2SDgMm;f2000N2K5N20000001O000000000000000000000000001O00000000000000007I00O100000000N2O10000000O1000O10000000000O0100000000000O1000O100000O100002N1O1O2M2O1OO1GTD^Ml;a2VD^Mj;b2WD\\Mj;c2YD[Mf;f2[DYMe;f2;0000O010O100O010O1000O0100O100O010O14L3L5LnMSDU1i;gN\\D\\1`;aNdD`1Z;\\NkDc1T;ZNQEe1o:XNUEf1l:ZNTEf1l:YNVEi0VOFc;BXEf0[OC];FZEe0@_OV;L[Eb0F]Oo:0^E`0I[Oh:6`E<OYOa::bE;c;E^D9c;F_D7c;I^DMj;3UR\\6"}, "label": "a black colur chair"}]}
{"id": 297360, "image": "COCO_train2014_000000297360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich\"."}], "task": "refering", "answer_template": "The \"a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 307], "bbox": [0.06896875, 0.2532552693208431, 0.580234375, 0.8802107728337236], "iscrowd": 0, "area": 58033.97114999999, "rle": {"size": [427, 640], "counts": "Xdb01Z=0O2N2O1N2[NJSF7l9MoE5o90lE1T:3hELY:6dEK\\:6cEIb9GXFb03He9GWFb02Gh9HTFc03Eh9JTFc01Cl9JSFd00Bm9JSFe0NAo9KSFf0L_OQ:KSFg0J^OT:LPFi0I\\OV:LQFi0HZOX:MPFj0FZOZ:LPFl0DWO\\:NPFl0BWO^:MPFm0@VOa:NmEn0ATOa:0mEm0@SOd:2jEY1V:iNhEV1X:mNfES1X:QOfEn0Y:VOdEk0Y:ZOeEf0X:_OfEb0W:[1L4M2O2N2M3N2M3N1N3N2M3N2N2M2O2M3N2M3lGdKV7]4iHgKR7]4kHfKR7\\4lHfKR7\\4lHeKR7_4kHbKT7`4jHaKT7c4hH_KW7c4gH^KW7e4gH\\KX7g4eHZKY7\\501O1N2O1O1O100O1O102M2O2M2O2M2O2M2O2M3N0O2O1N101N2N101N2O0O010O100O100O100O100O100O1O100O100O10O010O0100O101N100O1O2O0O101N100O101O01O0001O0000010O00001O0001O01O00000010O000001O00010O0000001O00000010O0001O0000001O0000001O01O0001O00001O0000001O00O2O0O1O101N101N101N101N101N101N101N101N101N1O2O1N101N101N101N101N101N101N101N101N1O2O0O2O1N101N101N101N101N101N101N101N1O2O1N2O1N2O1N20O0100O100O100O100O10O0VI]Jm5c5hIeJ[6Z5ZIoJi6g53M3\\NQIiLQ7U3RIgLQ7V3SIfLP7X3SIdLP7Y3TIcLn6\\3VI_Lm6^3WI_Lk6_3XI]Lk6`3YI\\Lj6b3YIZLj6c3ZIYLi6e3[IVLh6g3`1N1N3N2N2M3J6E;F:F:L301O1O10O01O1O001O1O010O1O001O1O00100O0O2M3M2N3M3M2N3M3M2N3M6J7J5J7Ian_3"}, "label": "a sandwich"}]}
{"id": 297360, "image": "COCO_train2014_000000297360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a burger with tomato and bacon\"."}], "task": "refering", "answer_template": "The \"a burger with tomato and bacon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 307], "bbox": [0.06896875, 0.2532552693208431, 0.580234375, 0.8802107728337236], "iscrowd": 0, "area": 58033.97114999999, "rle": {"size": [427, 640], "counts": "Xdb01Z=0O2N2O1N2[NJSF7l9MoE5o90lE1T:3hELY:6dEK\\:6cEIb9GXFb03He9GWFb02Gh9HTFc03Eh9JTFc01Cl9JSFd00Bm9JSFe0NAo9KSFf0L_OQ:KSFg0J^OT:LPFi0I\\OV:LQFi0HZOX:MPFj0FZOZ:LPFl0DWO\\:NPFl0BWO^:MPFm0@VOa:NmEn0ATOa:0mEm0@SOd:2jEY1V:iNhEV1X:mNfES1X:QOfEn0Y:VOdEk0Y:ZOeEf0X:_OfEb0W:[1L4M2O2N2M3N2M3N1N3N2M3N2N2M2O2M3N2M3lGdKV7]4iHgKR7]4kHfKR7\\4lHfKR7\\4lHeKR7_4kHbKT7`4jHaKT7c4hH_KW7c4gH^KW7e4gH\\KX7g4eHZKY7\\501O1N2O1O1O100O1O102M2O2M2O2M2O2M2O2M3N0O2O1N101N2N101N2O0O010O100O100O100O100O100O1O100O100O10O010O0100O101N100O1O2O0O101N100O101O01O0001O0000010O00001O0001O01O00000010O000001O00010O0000001O00000010O0001O0000001O0000001O01O0001O00001O0000001O00O2O0O1O101N101N101N101N101N101N101N101N101N1O2O1N101N101N101N101N101N101N101N101N1O2O0O2O1N101N101N101N101N101N101N101N1O2O1N2O1N2O1N20O0100O100O100O100O10O0VI]Jm5c5hIeJ[6Z5ZIoJi6g53M3\\NQIiLQ7U3RIgLQ7V3SIfLP7X3SIdLP7Y3TIcLn6\\3VI_Lm6^3WI_Lk6_3XI]Lk6`3YI\\Lj6b3YIZLj6c3ZIYLi6e3[IVLh6g3`1N1N3N2N2M3J6E;F:F:L301O1O10O01O1O001O1O010O1O001O1O00100O0O2M3M2N3M3M2N3M3M2N3M6J7J5J7Ian_3"}, "label": "a burger with tomato and bacon"}]}
{"id": 190868, "image": "COCO_train2014_000000190868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange , brown , and yellow train on the right side of the train tracks\"."}], "task": "refering", "answer_template": "The \"an orange , brown , and yellow train on the right side of the train tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.047640625, 0.14382075471698114, 0.867953125, 0.860683962264151], "iscrowd": 0, "area": 99326.0955, "rle": {"size": [424, 640], "counts": "`Q=2V=6I7J6J6I6K000O2O000O2O00001N10000O2O000O2O000O2O00000O2O000O2O000O101O0O101O0O101O000O101O0O101O0O101O0O10001O0O101O0O101O0O10001N10001O0O101O0O10001N10001N1000001N100O2O0O101O0O100O2O000O2O0O101O0O100O2O000O2O0O101O0O100O2O000O2O0O101O0O100O2O000O2O0O10001N100O2O000O2O0O2O0O101O0O101N10001N101N10001N100O2O001N100O2O000O2O0O101O0O2O0O101O0O101N101O0O101N10001N101N10001N100O2O000O2O001N100O2O000O2O001N100O2O000O2O0O101O0O2O000O2O0O101O0O2O0O101O0O101O0O2O0O101O00001O00001O001O00001O00001O001O00001O0O101N100O2N101N100O2O0O1O2O0O2O0O101N100O2N101N100O2O0O1O2O0O101N101N1O101N100O2O0O2O0O1O2O0O101N101N1O101N100O2O0O101N1O2O0O101N100O2N101N10001N10001N10001O0O2O000O2O000O2O001N10000O10000O1000000O10000O101O0O10000O10000O10000000000000000000000000000O10000000001O000O10000O100O10000O100O10000O10000O100O10000O100O10001N10000O100O10000O100O10000O100001O0O2O00001O001O001O0O101O001O00001O001O0O2O00001O001O00001N101O001O001O001O001N101O001O001O001O0O2O001O001O001O001N101O001N1N3N1O2N1N3N1O2N1N3N1N3L3M4M2M4L3M4L3N3L3M4L4L9H7H8H8H8_Oa0[Oe0ZOg0ZOe0^Nb1A?BbQS1"}, "label": "an orange , brown , and yellow train on the right side of the train tracks"}]}
{"id": 190868, "image": "COCO_train2014_000000190868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train on the tracks\"."}], "task": "refering", "answer_template": "The \"a train on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.047640625, 0.14382075471698114, 0.867953125, 0.860683962264151], "iscrowd": 0, "area": 99326.0955, "rle": {"size": [424, 640], "counts": "`Q=2V=6I7J6J6I6K000O2O000O2O00001N10000O2O000O2O000O2O00000O2O000O2O000O101O0O101O0O101O000O101O0O101O0O101O0O10001O0O101O0O101O0O10001N10001O0O101O0O10001N10001N1000001N100O2O0O101O0O100O2O000O2O0O101O0O100O2O000O2O0O101O0O100O2O000O2O0O101O0O100O2O000O2O0O10001N100O2O000O2O0O2O0O101O0O101N10001N101N10001N100O2O001N100O2O000O2O0O101O0O2O0O101O0O101N101O0O101N10001N101N10001N100O2O000O2O001N100O2O000O2O001N100O2O000O2O0O101O0O2O000O2O0O101O0O2O0O101O0O101O0O2O0O101O00001O00001O001O00001O00001O001O00001O0O101N100O2N101N100O2O0O1O2O0O2O0O101N100O2N101N100O2O0O1O2O0O101N101N1O101N100O2O0O2O0O1O2O0O101N101N1O101N100O2O0O101N1O2O0O101N100O2N101N10001N10001N10001O0O2O000O2O000O2O001N10000O10000O1000000O10000O101O0O10000O10000O10000000000000000000000000000O10000000001O000O10000O100O10000O100O10000O10000O100O10000O100O10001N10000O100O10000O100O10000O100001O0O2O00001O001O001O0O101O001O00001O001O0O2O00001O001O00001N101O001O001O001O001N101O001O001O001O0O2O001O001O001O001N101O001N1N3N1O2N1N3N1O2N1N3N1N3L3M4M2M4L3M4L3N3L3M4L4L9H7H8H8H8_Oa0[Oe0ZOg0ZOe0^Nb1A?BbQS1"}, "label": "a train on the tracks"}]}
{"id": 477590, "image": "COCO_train2014_000000477590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a red passanger train\"."}], "task": "refering", "answer_template": "The \"the front of a red passanger train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.37415625, 0.305625, 0.664046875, 0.7505624999999999], "iscrowd": 0, "area": 35404.274549999995, "rle": {"size": [480, 640], "counts": "g[`33Q>o0QOn0ROm0SOm0SOl0F;N1N3N1O2M2O2N0O2O001N10001N2O1O1N3N1O1N2O2N1N2O1O1N3N1O1N2O2N1N2O1O2O000000001O0000001O0000000000000000O10000000000000000O100000O10000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O100000000000001O1N2O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1N200000000O10000O1O1O1N2O0O2O1N2O3L`0A?@`0YOg0WOj0VOi0XOh0WOY`T3"}, "label": "the front of a red passanger train"}]}
{"id": 477590, "image": "COCO_train2014_000000477590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red train\"."}], "task": "refering", "answer_template": "The \"red train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.37415625, 0.305625, 0.664046875, 0.7505624999999999], "iscrowd": 0, "area": 35404.274549999995, "rle": {"size": [480, 640], "counts": "g[`33Q>o0QOn0ROm0SOm0SOl0F;N1N3N1O2M2O2N0O2O001N10001N2O1O1N3N1O1N2O2N1N2O1O1N3N1O1N2O2N1N2O1O2O000000001O0000001O0000000000000000O10000000000000000O100000O10000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O100000000000001O1N2O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1N200000000O10000O1O1O1N2O0O2O1N2O3L`0A?@`0YOg0WOj0VOi0XOh0WOY`T3"}, "label": "red train"}]}
{"id": 477590, "image": "COCO_train2014_000000477590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a maintenance vehicle on a railway\"."}], "task": "refering", "answer_template": "The \"a maintenance vehicle on a railway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234], "bbox": [0.000375, 0.36745833333333333, 0.2145, 0.6292916666666667], "iscrowd": 0, "area": 10698.212650000001, "rle": {"size": [480, 640], "counts": "_7<c>Y1gN6K0O100O100O2O0O100O1O100O2O0O100O100O101N100O1O100O101N1001O001O1O1O010O1O1O1O001O100O1O0nNgj2K`VM`1`N00000000000O10000000000000000000000O100000000000000000000O1001O0000001O0000001O0000001O00000000A?TOl000000000O100000000O100000000O100000000O100000000O1>Bc0]Oc0]Od0\\Oc0]Oc0]OR`[7"}, "label": "a maintenance vehicle on a railway"}]}
{"id": 477590, "image": "COCO_train2014_000000477590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow machinery\"."}], "task": "refering", "answer_template": "The \"yellow machinery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234], "bbox": [0.000375, 0.36745833333333333, 0.2145, 0.6292916666666667], "iscrowd": 0, "area": 10698.212650000001, "rle": {"size": [480, 640], "counts": "_7<c>Y1gN6K0O100O100O2O0O100O1O100O2O0O100O100O101N100O1O100O101N1001O001O1O1O010O1O1O1O001O100O1O0nNgj2K`VM`1`N00000000000O10000000000000000000000O100000000000000000000O1001O0000001O0000001O0000001O00000000A?TOl000000000O100000000O100000000O100000000O100000000O1>Bc0]Oc0]Od0\\Oc0]Oc0]OR`[7"}, "label": "yellow machinery"}]}
{"id": 149916, "image": "COCO_train2014_000000149916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player catching a baseball\"."}], "task": "refering", "answer_template": "The \"a player catching a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 54, 55, 56, 72, 73, 74, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 192, 193, 194, 195, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 285, 286, 303, 304, 305, 308, 309, 310, 311, 325, 326, 327, 328, 332, 333, 334, 356, 357], "bbox": [0.140671875, 0.11485928705440901, 0.557078125, 0.84234521575985], "iscrowd": 0, "area": 36795.857450000025, "rle": {"size": [533, 640], "counts": "]n^15V`0;E:01N2N2N2N101N2N2N1O2\\IlNdMV1[2kNdMV1[2kNcMV1\\2kNcMW1\\2jNbMW1^2jN`MW1`2jN_MV1a2jN^MW1b2jN\\MW1c2jN\\MW1d2iN[MY1c2iN\\MW1d2iN[MX1d2iN[MX1d2iN[MX1e2iNYMX1f2iNYMX1g2hNYMX1f2iNYMY1f2hNXMY1g2hNXMY1h2gNWMZ1h2gNXMY1h2gNYMX1f2jNYMV1g2jNYMV1f2kN[MU1d2kN\\MU1c2lN]MT1c2lN^MS1a2nN_MR1a2nN_MR1a2nN`MQ1_2POaMP1_2POaMQ1]2POdMo0\\2QOeMn0Z2SOhMk0X2UOkMh0U2XOnMe0Q2\\ORNa0n1_OTN`0j1AYN=e1D^N:`1GbN7]1JiN1T11SOIk08[OCc0>B^O;d0F\\O8e0I[O5f0LZO1h01VONk03UOJm07SOSJHi3W1T2QORJIg3X1X2oNoIK]3a1f2cNlIMQ3l1T3WNjINf2V2a3lMgI0\\2_2n3`MeI3T2e2W4YMcI4T2d2Z4XM`I5U2d2[4XM^I6U2c2^4WM[I7W2b2^4XMYI7X2b2_4XMXI6X2d2`4VMWI5Y2f2`4UMWI5Y2f2a4UMUI5Y2g2b4UMUI3X2i2c4TMUI2X2\\2XMPM\\7b0SI2X2T2]5jM[H2X2m1d5QNUH0W2n1g5RNQH0W2m1j5SNoG0V2l1m5UNlGNX2k1n5WNkGMV2l1P6WNjGMU2k1T6XNfGNT2i1X6YNeGMR2i1[6[NcGLQ2g1^6]NaGMo1f1b6\\N`GMm1f1e6^N^GLk1e1i6_N]GLi1b1m6cNZGJh1b1P7dNXGKf1_1V7eNUGLd1]1Y7hNSGKb1[1^7jNQGJa1Z1`7mNoFJ^1W1g7nNkFL\\1T1l7POiFMY1Q1P8SOfFMX1n0U8UOdFNU1k0Y8WObFOS1h0_8YO]F1Q1e0d8ZO\\F0P1e0e8\\OZFOP1f0f8[OZF0o0d0h8\\OZFOm0f0j8[OXFOm0g0k8ZOXFOm0f0l8\\OVFOl0f0n8[OWFNk0f0P9[OUFOj0g0Q9[OTFNj0h0R9ZOUFMi0h0S9\\OSFLi0i0T9[OSFLi0h0V9[OQFMh0i0W9[OQFKh0j0W9[OQFKg0j0Y9\\OoEJg0k0Z9[OoEJg0j0\\9[OnEJe0l0]9ZOnEJe0k0^9[OmEJd0l0_9[OlEId0m0a9YOjEKe0k0b9ZOiEKd0l0c9YOhELe0j0d9ZOgELd0k0i7eN\\He0UOLe0k0j7dN]Hd0TOMe0j0k7fN[Hc0UOMd0k0l7eN[Hc0TONe0i0m7gNYHb0UONd0j0n7gNYHa0SOOe0j0o7gNXH`0TOOe0i0P8iNVH?UOOd0j0Q8iNUGAa0^29WOQ8lNPG18l1f0XOR8mNiF6>d1f0ZOS8HWGn0f0YOT8IVGQ1b0WOX8HVGV1=QO^8ITG]17kNe8ISGa12gNk8HRGh1M_NR9JoFm1IZNX9JnFh3R9ZLlFg3T9[LjFd3W9]LgFd3X9m10O01O010O`0Aa0_O=B9H9F9H1N2O001N2O0O2O0O2O1O0010O10O010000O010O10O]GZIX7f6dH_IZ7c6bHbI\\7`6^HfI_7\\6\\HiIc7X6YHmId7U6WHQJg7`7O1N2O0O2O1N101cNYGiJi8Q5[GnJh8l4\\GTKf8g4\\GYKg8a4]G_Ke8\\4^GcKd8X4_GiKc8R4`GmKc8m3aGSLa8h3aGYLa8a3cG^L`8]3cGcL_8W3eGhL^8S3dGoL]8k2gGUM[8e2iG\\MX8_2jGcMW8X2lGhMV8R2nGoMS8l1PHUNQ8e1RH\\NS8\\1PHcNY8Q1jGPO_8d0dG[Oe89^GHk8LXG3o8CTG>P9[OSGd0R9ROTGn0Y<010O0010O01O3M2O2M2N2N3N1N3M0001O00001O3M3M4L5K5K5K4L4J6J6I7J6IVlb4"}, "label": "a player catching a baseball"}]}
{"id": 149916, "image": "COCO_train2014_000000149916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player about to catch a ball\"."}], "task": "refering", "answer_template": "The \"the baseball player about to catch a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 54, 55, 56, 72, 73, 74, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 192, 193, 194, 195, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 285, 286, 303, 304, 305, 308, 309, 310, 311, 325, 326, 327, 328, 332, 333, 334, 356, 357], "bbox": [0.140671875, 0.11485928705440901, 0.557078125, 0.84234521575985], "iscrowd": 0, "area": 36795.857450000025, "rle": {"size": [533, 640], "counts": "]n^15V`0;E:01N2N2N2N101N2N2N1O2\\IlNdMV1[2kNdMV1[2kNcMV1\\2kNcMW1\\2jNbMW1^2jN`MW1`2jN_MV1a2jN^MW1b2jN\\MW1c2jN\\MW1d2iN[MY1c2iN\\MW1d2iN[MX1d2iN[MX1d2iN[MX1e2iNYMX1f2iNYMX1g2hNYMX1f2iNYMY1f2hNXMY1g2hNXMY1h2gNWMZ1h2gNXMY1h2gNYMX1f2jNYMV1g2jNYMV1f2kN[MU1d2kN\\MU1c2lN]MT1c2lN^MS1a2nN_MR1a2nN_MR1a2nN`MQ1_2POaMP1_2POaMQ1]2POdMo0\\2QOeMn0Z2SOhMk0X2UOkMh0U2XOnMe0Q2\\ORNa0n1_OTN`0j1AYN=e1D^N:`1GbN7]1JiN1T11SOIk08[OCc0>B^O;d0F\\O8e0I[O5f0LZO1h01VONk03UOJm07SOSJHi3W1T2QORJIg3X1X2oNoIK]3a1f2cNlIMQ3l1T3WNjINf2V2a3lMgI0\\2_2n3`MeI3T2e2W4YMcI4T2d2Z4XM`I5U2d2[4XM^I6U2c2^4WM[I7W2b2^4XMYI7X2b2_4XMXI6X2d2`4VMWI5Y2f2`4UMWI5Y2f2a4UMUI5Y2g2b4UMUI3X2i2c4TMUI2X2\\2XMPM\\7b0SI2X2T2]5jM[H2X2m1d5QNUH0W2n1g5RNQH0W2m1j5SNoG0V2l1m5UNlGNX2k1n5WNkGMV2l1P6WNjGMU2k1T6XNfGNT2i1X6YNeGMR2i1[6[NcGLQ2g1^6]NaGMo1f1b6\\N`GMm1f1e6^N^GLk1e1i6_N]GLi1b1m6cNZGJh1b1P7dNXGKf1_1V7eNUGLd1]1Y7hNSGKb1[1^7jNQGJa1Z1`7mNoFJ^1W1g7nNkFL\\1T1l7POiFMY1Q1P8SOfFMX1n0U8UOdFNU1k0Y8WObFOS1h0_8YO]F1Q1e0d8ZO\\F0P1e0e8\\OZFOP1f0f8[OZF0o0d0h8\\OZFOm0f0j8[OXFOm0g0k8ZOXFOm0f0l8\\OVFOl0f0n8[OWFNk0f0P9[OUFOj0g0Q9[OTFNj0h0R9ZOUFMi0h0S9\\OSFLi0i0T9[OSFLi0h0V9[OQFMh0i0W9[OQFKh0j0W9[OQFKg0j0Y9\\OoEJg0k0Z9[OoEJg0j0\\9[OnEJe0l0]9ZOnEJe0k0^9[OmEJd0l0_9[OlEId0m0a9YOjEKe0k0b9ZOiEKd0l0c9YOhELe0j0d9ZOgELd0k0i7eN\\He0UOLe0k0j7dN]Hd0TOMe0j0k7fN[Hc0UOMd0k0l7eN[Hc0TONe0i0m7gNYHb0UONd0j0n7gNYHa0SOOe0j0o7gNXH`0TOOe0i0P8iNVH?UOOd0j0Q8iNUGAa0^29WOQ8lNPG18l1f0XOR8mNiF6>d1f0ZOS8HWGn0f0YOT8IVGQ1b0WOX8HVGV1=QO^8ITG]17kNe8ISGa12gNk8HRGh1M_NR9JoFm1IZNX9JnFh3R9ZLlFg3T9[LjFd3W9]LgFd3X9m10O01O010O`0Aa0_O=B9H9F9H1N2O001N2O0O2O0O2O1O0010O10O010000O010O10O]GZIX7f6dH_IZ7c6bHbI\\7`6^HfI_7\\6\\HiIc7X6YHmId7U6WHQJg7`7O1N2O0O2O1N101cNYGiJi8Q5[GnJh8l4\\GTKf8g4\\GYKg8a4]G_Ke8\\4^GcKd8X4_GiKc8R4`GmKc8m3aGSLa8h3aGYLa8a3cG^L`8]3cGcL_8W3eGhL^8S3dGoL]8k2gGUM[8e2iG\\MX8_2jGcMW8X2lGhMV8R2nGoMS8l1PHUNQ8e1RH\\NS8\\1PHcNY8Q1jGPO_8d0dG[Oe89^GHk8LXG3o8CTG>P9[OSGd0R9ROTGn0Y<010O0010O01O3M2O2M2N2N3N1N3M0001O00001O3M3M4L5K5K5K4L4J6J6I7J6IVlb4"}, "label": "the baseball player about to catch a ball"}]}
{"id": 149916, "image": "COCO_train2014_000000149916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player with the number 27 running to base\"."}], "task": "refering", "answer_template": "The \"a baseball player with the number 27 running to base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 108, 109, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 249, 250, 251, 252, 266, 267, 268, 269, 270, 274, 275, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 358, 359, 360, 382, 383, 404, 405, 427, 428], "bbox": [0.5675625, 0.1874108818011257, 1.0, 0.9907317073170733], "iscrowd": 0, "area": 39751.41544999997, "rle": {"size": [533, 640], "counts": "ZUm5a0R`02M4M3L3L5L4\\OlNaAY1m7oNmMOlIU1U8nNnM`1o1bNQN_1l1cNSN^1k1eNTN\\1i1fNVN\\1f1gNYN[1d1hN[NX1c1jN\\NV1b1mN^NS1`1oN_NR1_1QO`Nn0^1UOaNl0]OVN\\LP1V4k0\\NZOXMN\\4g0\\M>TNmN_4f0eLX1gNTNc4e0cL\\1eNRNg4b0aLc1cNmMl4`0^Li1aNiMP5?\\Lm1_NgMT5<ZLT2]NbMX5`1SKS1A_M[5`1PKW1_O\\Ma5^1kJ`3T5bLjJ_3U5bLkJm2CPJa5U3kJj2HmI^5Z3iJi2JjI^5X:dJeE]5[:hJ`EX5`:>01O0O101N2N2N2M3N2N2N2N2N1O2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N12OO1O1O1O1O1O1N2M3[OTH^Ho7a7c0OTO_HfH`7W7fHgHX7W7nHfHS7V7SIgHo6T7WIiHk6R7ZIlHg6P7_ImHb6P7c1O1O100O1O2O0O2N2N101N2N10RO]IkGb6S8aImG^6P8eIPH[6n7hIRHW6m7kISHU6k7mITHS6j7oIUHR6j7PJUHP6j7RJTHo5k7SJSHn5l7SJTHn5j7TJSHn5k7VJQHl5m7YJnGi5P8\\JkGf5S8_1N2N3M2N3M3XJXGS2OTMk8f0YGT2OSMk8e0ZGU2NSMk8c0\\GW2MRMi8d0]GY2LPMj8c0^G\\2InLl8a0`Gm0XOXOm0gN]8Q1bGP1XOSO^:I]FU1XOnN\\:I`FY1XOiNY:KbF[1YOfNY:J`F`1ZObNZ:H^Fg1\\OZN[:J\\Fk1]OTN^:LWFP2_OlMa:NRFW2i<O1000O10O1000O010000O010000O1000000O2O00O1N4L3M4M2O2O0O2N1O2N1O2M2O2M2O1N2O1N2O1N2O001N2O0O10O0100O010O010O010O0100O0100O010O0010O010O01O010O010O1O010O0010O010O01O010O010O1O010O010O0010O010O01000O100O00100O0001O01O0001O000010O0001O0002O2M2N3N_Ob@K\\?6e@KX?5j@HW?9k@DV?;l@CT?>m@@S?`0PA\\OR?d0?0010O010O01O01O01fF"}, "label": "a baseball player with the number 27 running to base"}]}
{"id": 149916, "image": "COCO_train2014_000000149916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player in jersey number 27\"."}], "task": "refering", "answer_template": "The \"a player in jersey number 27\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 108, 109, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 249, 250, 251, 252, 266, 267, 268, 269, 270, 274, 275, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 358, 359, 360, 382, 383, 404, 405, 427, 428], "bbox": [0.5675625, 0.1874108818011257, 1.0, 0.9907317073170733], "iscrowd": 0, "area": 39751.41544999997, "rle": {"size": [533, 640], "counts": "ZUm5a0R`02M4M3L3L5L4\\OlNaAY1m7oNmMOlIU1U8nNnM`1o1bNQN_1l1cNSN^1k1eNTN\\1i1fNVN\\1f1gNYN[1d1hN[NX1c1jN\\NV1b1mN^NS1`1oN_NR1_1QO`Nn0^1UOaNl0]OVN\\LP1V4k0\\NZOXMN\\4g0\\M>TNmN_4f0eLX1gNTNc4e0cL\\1eNRNg4b0aLc1cNmMl4`0^Li1aNiMP5?\\Lm1_NgMT5<ZLT2]NbMX5`1SKS1A_M[5`1PKW1_O\\Ma5^1kJ`3T5bLjJ_3U5bLkJm2CPJa5U3kJj2HmI^5Z3iJi2JjI^5X:dJeE]5[:hJ`EX5`:>01O0O101N2N2N2M3N2N2N2N2N1O2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N12OO1O1O1O1O1O1N2M3[OTH^Ho7a7c0OTO_HfH`7W7fHgHX7W7nHfHS7V7SIgHo6T7WIiHk6R7ZIlHg6P7_ImHb6P7c1O1O100O1O2O0O2N2N101N2N10RO]IkGb6S8aImG^6P8eIPH[6n7hIRHW6m7kISHU6k7mITHS6j7oIUHR6j7PJUHP6j7RJTHo5k7SJSHn5l7SJTHn5j7TJSHn5k7VJQHl5m7YJnGi5P8\\JkGf5S8_1N2N3M2N3M3XJXGS2OTMk8f0YGT2OSMk8e0ZGU2NSMk8c0\\GW2MRMi8d0]GY2LPMj8c0^G\\2InLl8a0`Gm0XOXOm0gN]8Q1bGP1XOSO^:I]FU1XOnN\\:I`FY1XOiNY:KbF[1YOfNY:J`F`1ZObNZ:H^Fg1\\OZN[:J\\Fk1]OTN^:LWFP2_OlMa:NRFW2i<O1000O10O1000O010000O010000O1000000O2O00O1N4L3M4M2O2O0O2N1O2N1O2M2O2M2O1N2O1N2O1N2O001N2O0O10O0100O010O010O010O0100O0100O010O0010O010O01O010O010O1O010O0010O010O01O010O010O1O010O010O0010O010O01000O100O00100O0001O01O0001O000010O0001O0002O2M2N3N_Ob@K\\?6e@KX?5j@HW?9k@DV?;l@CT?>m@@S?`0PA\\OR?d0?0010O010O01O01O01fF"}, "label": "a player in jersey number 27"}]}
{"id": 305564, "image": "COCO_train2014_000000305564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sleeping golden retriever\"."}], "task": "refering", "answer_template": "The \"the sleeping golden retriever\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.086015625, 0.4750416666666667, 0.559859375, 0.6923541666666667], "iscrowd": 0, "area": 20327.818949999993, "rle": {"size": [480, 640], "counts": "ZRj01m>3M201N2N1O2O0O2N2O0L4O2M2O1N2O1O1N3N1N2O1N2O2N1N2O2M2O1N3N1O1O2O001O0000O0100000O10000O10000001O0O10000000000O2O00MkB\\NS=f1mBYNS=h1mBXNR=h1oBXNo<j1QCUNo<l1QCTNn<m11101O1N101N10O10O1000O010000O01000O200O1O00100O1O1O010O0000010O000001O0O10001O00000000O1000000O2O00001N1000001O0O2O00001N10001O0O2O00001N1O1ON1O3M24M00001O00001O001O0O2O001O001O00001O0O2O00001OCRDhMn;Y2SDeMm;[2VDbMj;_2WD]Mk;c291N10O10000000000O01000000O100000000O1000000O10001O0O100000O10O100000O1000001O0O1000000O100000O10O1000000O100000O10O1000001O001O00001N101O001O001O00001O001O001O001N10001O001O001O001O00001O001O0O2O1O1O1O1O1O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2N2N2M3N2M3N2M3N2M3N2M3N2MVmS4"}, "label": "the sleeping golden retriever"}]}
{"id": 305564, "image": "COCO_train2014_000000305564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog closest to teddy bear\"."}], "task": "refering", "answer_template": "The \"dog closest to teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.086015625, 0.4750416666666667, 0.559859375, 0.6923541666666667], "iscrowd": 0, "area": 20327.818949999993, "rle": {"size": [480, 640], "counts": "ZRj01m>3M201N2N1O2O0O2N2O0L4O2M2O1N2O1O1N3N1N2O1N2O2N1N2O2M2O1N3N1O1O2O001O0000O0100000O10000O10000001O0O10000000000O2O00MkB\\NS=f1mBYNS=h1mBXNR=h1oBXNo<j1QCUNo<l1QCTNn<m11101O1N101N10O10O1000O010000O01000O200O1O00100O1O1O010O0000010O000001O0O10001O00000000O1000000O2O00001N1000001O0O2O00001N10001O0O2O00001N1O1ON1O3M24M00001O00001O001O0O2O001O001O00001O0O2O00001OCRDhMn;Y2SDeMm;[2VDbMj;_2WD]Mk;c291N10O10000000000O01000000O100000000O1000000O10001O0O100000O10O100000O1000001O0O1000000O100000O10O1000000O100000O10O1000001O001O00001N101O001O001O00001O001O001O001N10001O001O001O001O00001O001O0O2O1O1O1O1O1O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2N2N2M3N2M3N2M3N2M3N2M3N2MVmS4"}, "label": "dog closest to teddy bear"}]}
{"id": 305564, "image": "COCO_train2014_000000305564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog with a black nose\"."}], "task": "refering", "answer_template": "The \"a dog with a black nose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 273, 274], "bbox": [0.49696875, 0.4262708333333334, 0.9515156250000001, 0.6707083333333334], "iscrowd": 0, "area": 21249.852499999986, "rle": {"size": [480, 640], "counts": "dYe43l>3L3N3M2M4M3M3L4M3M3L3N2N2M3N2O1O1O1N2O1O1O1O1O2M2O1O1O1O001N4M3M2N00000O1000O1000000O0100000O1000O10000000O10O100000000O0100000O100000O01000001N1000001N1000000O2O0000001O00001O001O00001O001O00001O00000000001O0000000O100000001O000000O1000O1000O100000000O10O100000O100000000O01000000000000000O10O100000000000000000000O10O1001O00001O00001O0000000000000001O00000000000000001O00000000000000001O00000001O0000010O0000001O000010O0010O01O010O0010O01O010O010O0010O010O0100O00100O010O1O0010O01O1O001O10O0100O01000O02O001N2OO1O003M6K4K5K3WOoB]OR=a0PC^OR=?QC@o<?RC@P=?QC_OP=a0QC\\OR=c0nB\\OS=d0SCTOP=k0QCSOP=m0c0O0O101O1N10100O1N1O2N2M2O2N2M5L4L^W>"}, "label": "a dog with a black nose"}]}
{"id": 305564, "image": "COCO_train2014_000000305564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog laying against pillow\"."}], "task": "refering", "answer_template": "The \"dog laying against pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 273, 274], "bbox": [0.49696875, 0.4262708333333334, 0.9515156250000001, 0.6707083333333334], "iscrowd": 0, "area": 21249.852499999986, "rle": {"size": [480, 640], "counts": "dYe43l>3L3N3M2M4M3M3L4M3M3L3N2N2M3N2O1O1O1N2O1O1O1O1O2M2O1O1O1O001N4M3M2N00000O1000O1000000O0100000O1000O10000000O10O100000000O0100000O100000O01000001N1000001N1000000O2O0000001O00001O001O00001O001O00001O00000000001O0000000O100000001O000000O1000O1000O100000000O10O100000O100000000O01000000000000000O10O100000000000000000000O10O1001O00001O00001O0000000000000001O00000000000000001O00000000000000001O00000001O0000010O0000001O000010O0010O01O010O0010O01O010O010O0010O010O0100O00100O010O1O0010O01O1O001O10O0100O01000O02O001N2OO1O003M6K4K5K3WOoB]OR=a0PC^OR=?QC@o<?RC@P=?QC_OP=a0QC\\OR=c0nB\\OS=d0SCTOP=k0QCSOP=m0c0O0O101O1N10100O1N1O2N2M2O2N2M5L4L^W>"}, "label": "dog laying against pillow"}]}
{"id": 354716, "image": "COCO_train2014_000000354716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a pink tinted leather couch that is next to the coffee table\"."}], "task": "refering", "answer_template": "The \"this is a pink tinted leather couch that is next to the coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.41612499999999997, 0.8021458333333332, 0.993546875, 0.9871041666666666], "iscrowd": 0, "area": 22886.046350000004, "rle": {"size": [480, 640], "counts": "URm31o>8H9G9F:G9G9G7I6I8I4L3M3MO010000000O010000001O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OmN@gC?W<DiC;U<HkC7S<LmC3Q<0oCOo;4QDKo;6QDJn;7RDHm;:SDEm;<SDCm;>SDAm;`0TD^Ol;c0TD\\Ol;e0TDZOl;g0TDXOl;h0TDXOl;i0SDWOm;j0QDWOo;j0PDVOo;l0PDTOP<m0oCSOQ<n0nCROR<o0mCQOS<P1kCQOU<o0kCQOU<P1jCPOU<R1jCnNV<S1iCmNV<U1iCkNV<W1iCiNV<Y1hChNX<Y1gCgNY<Y1gCgNY<Z1fCfNZ<[1eCeN\\<[1cCeN]<\\1bCdN^<]1`CdN`<]1_CcNb<]1]CcNc<]1]CcNc<^1\\CbNd<_1[CaNf<_1YCaNg<`1WCaNi<`1VC`Nj<a1UC_Nl<a1SC_Nm<a1SC_Nm<b1RC^Nn<j11O0000010O00000000O10000000000000000000000000000O100000000000000000000000O10O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O10O100000000000000000000000O10000000000000000000000000000O100000000000000000000000000O1000000000O1000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000000000000O1000000000O100000000000000000000000000O10000000000000000000000000000O1000000000000000O10000O1000OoNR1WO_`1"}, "label": "this is a pink tinted leather couch that is next to the coffee table"}]}
{"id": 354716, "image": "COCO_train2014_000000354716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink couch next to a black chair behind a wood table\"."}], "task": "refering", "answer_template": "The \"a pink couch next to a black chair behind a wood table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.41612499999999997, 0.8021458333333332, 0.993546875, 0.9871041666666666], "iscrowd": 0, "area": 22886.046350000004, "rle": {"size": [480, 640], "counts": "URm31o>8H9G9F:G9G9G7I6I8I4L3M3MO010000000O010000001O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OmN@gC?W<DiC;U<HkC7S<LmC3Q<0oCOo;4QDKo;6QDJn;7RDHm;:SDEm;<SDCm;>SDAm;`0TD^Ol;c0TD\\Ol;e0TDZOl;g0TDXOl;h0TDXOl;i0SDWOm;j0QDWOo;j0PDVOo;l0PDTOP<m0oCSOQ<n0nCROR<o0mCQOS<P1kCQOU<o0kCQOU<P1jCPOU<R1jCnNV<S1iCmNV<U1iCkNV<W1iCiNV<Y1hChNX<Y1gCgNY<Y1gCgNY<Z1fCfNZ<[1eCeN\\<[1cCeN]<\\1bCdN^<]1`CdN`<]1_CcNb<]1]CcNc<]1]CcNc<^1\\CbNd<_1[CaNf<_1YCaNg<`1WCaNi<`1VC`Nj<a1UC_Nl<a1SC_Nm<a1SC_Nm<b1RC^Nn<j11O0000010O00000000O10000000000000000000000000000O100000000000000000000000O10O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O10O100000000000000000000000O10000000000000000000000000000O100000000000000000000000000O1000000000O1000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000000000000O1000000000O100000000000000000000000000O10000000000000000000000000000O1000000000000000O10000O1000OoNR1WO_`1"}, "label": "a pink couch next to a black chair behind a wood table"}]}
{"id": 354716, "image": "COCO_train2014_000000354716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch nearest the windows\"."}], "task": "refering", "answer_template": "The \"the couch nearest the windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 300, 301, 302, 303], "bbox": [0.027421875000000002, 0.4516041666666667, 0.214515625, 0.8150625], "iscrowd": 0, "area": 17274.52105, "rle": {"size": [480, 640], "counts": "lf8;`>=D;D=C=D;D=D<C<E<C=C=D;D9J200O10000O10000O1000O10O10000O10000001O1O1O1O001O1N2O001O1O1O00000000001O000000000O100000000000000000001O0000000O10000000000000000000001O0000000O1000000000000000000000001O000O100000000000000XOl0ROR1E;E;F:E<D=C<D<Eb_[7"}, "label": "the couch nearest the windows"}]}
{"id": 354716, "image": "COCO_train2014_000000354716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown leather couch closest to the glass door\"."}], "task": "refering", "answer_template": "The \"brown leather couch closest to the glass door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 300, 301, 302, 303], "bbox": [0.027421875000000002, 0.4516041666666667, 0.214515625, 0.8150625], "iscrowd": 0, "area": 17274.52105, "rle": {"size": [480, 640], "counts": "lf8;`>=D;D=C=D;D=D<C<E<C=C=D;D9J200O10000O10000O1000O10O10000O10000001O1O1O1O001O1N2O001O1O1O00000000001O000000000O100000000000000000001O0000000O10000000000000000000001O0000000O1000000000000000000000001O000O100000000000000XOl0ROR1E;E;F:E<D=C<D<Eb_[7"}, "label": "brown leather couch closest to the glass door"}]}
{"id": 354716, "image": "COCO_train2014_000000354716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unoccupied black chair\"."}], "task": "refering", "answer_template": "The \"an unoccupied black chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 351, 352, 353, 354, 374, 375, 376, 377], "bbox": [0.211375, 0.5477500000000001, 0.4774062499999999, 0.9989375], "iscrowd": 0, "area": 22456.0878, "rle": {"size": [480, 640], "counts": "Tdo11n>2gM:UEHh:h0hDZOU;U1^DmN_;Y1[DjNb;\\1YDfNd;`1VDbNh;d1SD^Nk;g1PD[No;[201O1N2O1O1O1Ng0ZO1O1O0O2O1O2N1N2O0_FjKX8W4eGkKZ8W4cGkK]8U4cGkK]8V4aGkK_8V4`GkK_8U4`GlKLIi7\\4ZHkKLLh7Z4ZHkKNMf7X4\\HkKMNg7X4[HkKMMh7X4[HkKLOh7V4\\HkKLOh7W4[HjKL1h7U4\\HjKL1h7U4]HjKI2j7U4\\HiKI4j7S4]HiKI4j7S4]HiKH4l7T4[HhKI4l7T4\\HhKF4o7T4[HhKF4o7U4ZHgKF4Q8U4YHgKF4Q8U4YHgKE4S8V4WHfKF3T8W4WHfKC4U8W4XHeKC3V8Y4VHdKC4W8X4VHdKC3X8Y4UHdKB4Y8X4UHeKA2[8Z4THcK@4\\8Y4THcK_O4^8Y4SHcK_O4^8Z4RHbK_O4`8Z4QHcK^O3a8Z4RHbK\\O4c8[4PHaK]O3d8\\4oGaK\\O4e8[4oGaK\\O3f8]4mGoKS8Q4mGnKT8R4mGmKS8T4lGkKU8U4kGkKU8[4eGdK\\8h4XGXKi8^5100O1O100O2O0O100O100010O1O1O1O2N100O1O1O1O1O100O1O1OjN`GmK_8Q4fGmKY8P4iGRLW8k3kGVLT8g3oGaK_O>a8P4QHbK_O>_8P4RHbK@?]8n3SHcKA?]8l3QHfKB>^8k3PHgKC=^8k3oGgKE=\\8l3oGgKF<[8l3PHhKF:\\8m3nGiKG9[8n3nGiKG9[8n3mGjKI7[8n3lGkKJ6Z8o3lGkKK4Z8Q4kGkKL3Z8P4kGlKM3X8Q4kGlKM3X8Q4jGmKO0X8S4iGmK0OX8S4hGnK1NW8T4hGnK2MV8T4iGoK2LV8T4hGPL7FR8Z4gGPLT9P4kFPLW9o3iFQLW9S3^FWM;FW9R3mFjLL4W9?`G?7RO[:n0fEROY:o0fEROY:n0hEROW:n0kEQOT:o0PFmNP:T1TFhNk9Y1m12N1O1O1O1O1N2O1000_OoBTOP=j0d0O1O0O2O1O1N2O1N2O1O1N2N2O1N3M3M3N5Jafl4"}, "label": "an unoccupied black chair"}]}
{"id": 354716, "image": "COCO_train2014_000000354716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black colored rolling chair\"."}], "task": "refering", "answer_template": "The \"a black colored rolling chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 351, 352, 353, 354, 374, 375, 376, 377], "bbox": [0.211375, 0.5477500000000001, 0.4774062499999999, 0.9989375], "iscrowd": 0, "area": 22456.0878, "rle": {"size": [480, 640], "counts": "Tdo11n>2gM:UEHh:h0hDZOU;U1^DmN_;Y1[DjNb;\\1YDfNd;`1VDbNh;d1SD^Nk;g1PD[No;[201O1N2O1O1O1Ng0ZO1O1O0O2O1O2N1N2O0_FjKX8W4eGkKZ8W4cGkK]8U4cGkK]8V4aGkK_8V4`GkK_8U4`GlKLIi7\\4ZHkKLLh7Z4ZHkKNMf7X4\\HkKMNg7X4[HkKMMh7X4[HkKLOh7V4\\HkKLOh7W4[HjKL1h7U4\\HjKL1h7U4]HjKI2j7U4\\HiKI4j7S4]HiKI4j7S4]HiKH4l7T4[HhKI4l7T4\\HhKF4o7T4[HhKF4o7U4ZHgKF4Q8U4YHgKF4Q8U4YHgKE4S8V4WHfKF3T8W4WHfKC4U8W4XHeKC3V8Y4VHdKC4W8X4VHdKC3X8Y4UHdKB4Y8X4UHeKA2[8Z4THcK@4\\8Y4THcK_O4^8Y4SHcK_O4^8Z4RHbK_O4`8Z4QHcK^O3a8Z4RHbK\\O4c8[4PHaK]O3d8\\4oGaK\\O4e8[4oGaK\\O3f8]4mGoKS8Q4mGnKT8R4mGmKS8T4lGkKU8U4kGkKU8[4eGdK\\8h4XGXKi8^5100O1O100O2O0O100O100010O1O1O1O2N100O1O1O1O1O100O1O1OjN`GmK_8Q4fGmKY8P4iGRLW8k3kGVLT8g3oGaK_O>a8P4QHbK_O>_8P4RHbK@?]8n3SHcKA?]8l3QHfKB>^8k3PHgKC=^8k3oGgKE=\\8l3oGgKF<[8l3PHhKF:\\8m3nGiKG9[8n3nGiKG9[8n3mGjKI7[8n3lGkKJ6Z8o3lGkKK4Z8Q4kGkKL3Z8P4kGlKM3X8Q4kGlKM3X8Q4jGmKO0X8S4iGmK0OX8S4hGnK1NW8T4hGnK2MV8T4iGoK2LV8T4hGPL7FR8Z4gGPLT9P4kFPLW9o3iFQLW9S3^FWM;FW9R3mFjLL4W9?`G?7RO[:n0fEROY:o0fEROY:n0hEROW:n0kEQOT:o0PFmNP:T1TFhNk9Y1m12N1O1O1O1O1N2O1000_OoBTOP=j0d0O1O0O2O1O1N2O1N2O1O1N2N2O1N3M3M3N5Jafl4"}, "label": "a black colored rolling chair"}]}
{"id": 35230, "image": "COCO_train2014_000000035230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza on the right\"."}], "task": "refering", "answer_template": "The \"the pizza on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249], "bbox": [0.634203125, 0.330327868852459, 0.953546875, 0.6943793911007026], "iscrowd": 0, "area": 22716.57155, "rle": {"size": [427, 640], "counts": "i`Y53U=5K4L5K5K4L5K4L5K5K4L5M2O2M3M2N3M2N3M3M2N3M2N3M2N2O0O2N1O2N2N1O2N2N1O2N101N2O0O2O1N101N101N2O0O2O1N100O101N10000O101N100O100O2O0O100O101N1000001O000000001O000000001O000000001O000000001O00O1000000000O10000001O000000O100O001O1O100O1O1O100O1O1O100O1O1O10O01O1O100O11O001O001O1O001O001N2O0O2O0O2O0O2O1N2O1N101N2O1N101O1N2O0O2O1N2O1N101N2O1N101N2N2N2N1O2M3N2N2N1O2N2N2M3N1O2N2N2N2M2O2N2N2N1O2N2M3L4L3L5K6K:E;F;DgY<"}, "label": "the pizza on the right"}]}
{"id": 35230, "image": "COCO_train2014_000000035230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza in paper box and neare to the camera\"."}], "task": "refering", "answer_template": "The \"a pizza in paper box and neare to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249], "bbox": [0.634203125, 0.330327868852459, 0.953546875, 0.6943793911007026], "iscrowd": 0, "area": 22716.57155, "rle": {"size": [427, 640], "counts": "i`Y53U=5K4L5K5K4L5K4L5K5K4L5M2O2M3M2N3M2N3M3M2N3M2N3M2N2O0O2N1O2N2N1O2N2N1O2N101N2O0O2O1N101N101N2O0O2O1N100O101N10000O101N100O100O2O0O100O101N1000001O000000001O000000001O000000001O000000001O00O1000000000O10000001O000000O100O001O1O100O1O1O100O1O1O100O1O1O10O01O1O100O11O001O001O1O001O001N2O0O2O0O2O0O2O1N2O1N101N2O1N101O1N2O0O2O1N2O1N101N2O1N101N2N2N2N1O2M3N2N2N1O2N2N2M3N1O2N2N2N2M2O2N2N2N1O2N2M3L4L3L5K6K:E;F;DgY<"}, "label": "a pizza in paper box and neare to the camera"}]}
{"id": 35230, "image": "COCO_train2014_000000035230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in a paper box above the another one\"."}], "task": "refering", "answer_template": "The \"the pizza in a paper box above the another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 126, 127, 128], "bbox": [0.40271875, 0.09459016393442624, 0.68671875, 0.3603512880562061], "iscrowd": 0, "area": 14098.700150000002, "rle": {"size": [427, 640], "counts": "Rf[32X=2M3L5K4L4L4M3M3N2M3M3M4L3N2M3N2N2N2N2N2N2O2M2N2N2N2O1N2N2N1O100O1O2N100O10001O0000001N100000001O00000O101O000000001N1000001O000O1O101N100O100O2O0O1O101N100O100O2O0O1O100O2O0O10O0100001N2O1O1O1O1N2O1O1O000O10001O000O100000000O2O0000000O1000001O0O100000000O101O00000O2O0N_EZMa:e2`E[Ma:d2_E\\Ma:c25O0O2O000O2O0O2O0O101N101N100O2O0O2O1N2O1N2O1O1N2O1N2O1N2N2M4L3M3M3M4L3M3M4L3L4L4L5KZgc2"}, "label": "the pizza in a paper box above the another one"}]}
{"id": 35230, "image": "COCO_train2014_000000035230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in the box towards the back\"."}], "task": "refering", "answer_template": "The \"the pizza in the box towards the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 126, 127, 128], "bbox": [0.40271875, 0.09459016393442624, 0.68671875, 0.3603512880562061], "iscrowd": 0, "area": 14098.700150000002, "rle": {"size": [427, 640], "counts": "Rf[32X=2M3L5K4L4L4M3M3N2M3M3M4L3N2M3N2N2N2N2N2N2O2M2N2N2N2O1N2N2N1O100O1O2N100O10001O0000001N100000001O00000O101O000000001N1000001O000O1O101N100O100O2O0O1O101N100O100O2O0O1O100O2O0O10O0100001N2O1O1O1O1N2O1O1O000O10001O000O100000000O2O0000000O1000001O0O100000000O101O00000O2O0N_EZMa:e2`E[Ma:d2_E\\Ma:c25O0O2O000O2O0O2O0O101N101N100O2O0O2O1N2O1N2O1O1N2O1N2O1N2N2M4L3M3M3M4L3M3M4L3L4L4L5KZgc2"}, "label": "the pizza in the box towards the back"}]}
{"id": 428445, "image": "COCO_train2014_000000428445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large cow in the front to the right of the other large one\"."}], "task": "refering", "answer_template": "The \"large cow in the front to the right of the other large one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 96, 97, 98, 99, 112, 113, 127, 128, 129], "bbox": [0.457981220657277, 0.5908235294117647, 0.6709154929577464, 0.9330980392156862], "iscrowd": 0, "area": 4176.079499999999, "rle": {"size": [255, 426], "counts": "Yg`18_7;K4VI[OT6i0fIHIFf5n1M2O1O101N1OZOf0100jIkNb5j1O1O1OO1PO`JUO4B^5X1aJoNn5P1VJjNm5T1oJgN\\4^1]KgN`4^1YKfNf4`1PKbNQ5V200000jNdJYO:@R5X1hJQO9Fo4Z1lJiN9Kk4]1S11O100O001O1O1O00O2O00001N101O000O2O03M6J4M2M3M5KmMVJU2g53NhM[JX2e5aNdJ<\\5A[KLc44aKI]49dKD]4=dKA\\4?eK@Z4b0fK\\O[4d0gKZOX4g0iKXOW4i0iKUOX4k0jKSOV4n0]100001OWORIb0m6^OTIa0l6_OTIa0l6_OUI`0l6_OTI`0m6@\\I7d6I_I3b6LbILc64e0O00O1O02O0O101N10cfR1"}, "label": "large cow in the front to the right of the other large one"}]}
{"id": 526754, "image": "COCO_train2014_000000526754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white zebra looking upward with another zebra in the zoo\"."}], "task": "refering", "answer_template": "The \"black and white zebra looking upward with another zebra in the zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 146, 147, 148, 149, 165, 166, 167, 168, 184, 185, 186, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 306, 307, 308, 309, 310, 311, 312, 313, 314, 316, 317, 318, 325, 326, 327, 328, 329, 335, 336, 344, 345, 346, 347, 354, 355, 363, 364, 365, 366, 373, 374, 382, 383, 385, 386, 392, 393, 401, 402, 404, 405, 411, 412, 420, 421, 424, 431], "bbox": [0.10761638733705772, 0.27709375, 0.9147113594040969, 0.9836406249999999], "iscrowd": 0, "area": 80116.65205000002, "rle": {"size": [640, 537], "counts": "XeT1`0Xc0i1XN<D;E<G9F3]@nMW=@YB[1`06T<bNcCP1:a0m;aNQDf03m0f;_N^D<MY1^;^NlD2Ed1Y;]NYEG_OP2Q;\\NgEWO^Oa2d:ZNVFgN\\OS3X:YNbFWN]Od3n9TNPHP2l7PNSHT2k7jMVHZ2f7fMZH^2c7dMYH`2d7eMWH\\2i7hMRHZ2l7lMmGW2R8mMiGU2U8QNdGR2[8RN`GP2_8TN[Go1c8WNWGk1h8ZNRGg1m8`NkF;kN^MY:]2eF2ZO^Mo9f2`FKH\\Mg9o2ZFC6[M^9Y3TF[Oe0YMV9a3oETOR1XMm8R7]GkHb8n6iGoHW8h6UHUIj7c6bHZI^7`6kH]IU7a6PI\\Io6c6VIZIj6d6Z3N1O2N2N1O2N1O2N2O0O2N2N1O2N2N1O2N1O2N2N1O2O10O01O10O01O010O1O1O10]CeJa:Z5PDXLn;P6M3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N2N2N1O2M3N1O2N2N2NWMfEmKY:P4QFkKm9R4^FhK`9V4iFeKU9X4UGbKj8\\4_G_K_8^4kG]KS8a4VHZKh7c4bHWK]7f4jHXKT7f4PIZKn6c4WI]Kg6a4]I]Kc6`4aI_K_6^4fI`KZ6^4jIaKU6\\4PJbKP6\\4SJcKm5Z4XJdKh5X4^JfKb5W4cJgK]5V4gJiKY5T4lJjKT5R4RKlKn4Q4WKnKh4o3\\KPLd4m3aKQL`4j3fKTLZ4i3kKULU4i3nKVLR4k3oKSLQ4m3QLQLo3P4RLnKn3R4SLnKm3R4TLlKl3T4VLjKj3W4WLgKi3Z4WLeKi3[4n51O0010O01O001O1O0010O01O001O001O10O01O0000000010O00000000000010O0000000000001O000O10000000001O0000000000000O2O000000000000001O0000O10O100000O100000000O10000000O10O100000000O100000000O10O100000O100000000O10000000O0100000000O010000O10O10O10000O1000O10O10000O1000O010000O1000O010000O10000SD_LZ7a3TGUNi8k1mEBP:Z5L4M3M3L3N001N2O1O0O2O1O1N101N2O1O0O2O1O0OkMeFhIZ9R6oFlIP9o5XGoIf8l5cGQJ]8i5mGTJa5]NaKZ7WOVJW5jNYKk6HWJn4POZKc61ZJc4VOZK[6=ZJY4\\OZKV6d0[JP4B[KP6k0ZJi3H[Km5Q1XJc3MZKi5Y1VJ\\34WKf5d1QJV39TKf5k1nIQ3=PKe5U2jIm2`0lJe5]2hIg2d0hJe5g2bIb2j0dJc5P3`I\\2m0aJc5Y3\\IW2R1]Jb5a3YIR2U1YJe5Y:\\JaEg5a:YJ[Ei5g:YJSEi5o:XJlDj5W;]12N2N2N2N2N2O1N2N3M2N2O1N2N2N2O1N2N3M2O1N2N2N2O1N2N2N3M2N2O1N2N2N2N2N2N2N3M2N2N2G8I8G9H8K5M2O2M3M3M2N3H8C=B>SOl0K6K5J6K5J5L5K5L4K5K4L50000O10O1000O1000O0100000O01000000O0100000N2N1O2N2N2N2N2N4L4L\\bl0"}, "label": "black and white zebra looking upward with another zebra in the zoo"}]}
{"id": 526754, "image": "COCO_train2014_000000526754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with its mouth open , looking towards the camera\"."}], "task": "refering", "answer_template": "The \"a zebra with its mouth open , looking towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 146, 147, 148, 149, 165, 166, 167, 168, 184, 185, 186, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 306, 307, 308, 309, 310, 311, 312, 313, 314, 316, 317, 318, 325, 326, 327, 328, 329, 335, 336, 344, 345, 346, 347, 354, 355, 363, 364, 365, 366, 373, 374, 382, 383, 385, 386, 392, 393, 401, 402, 404, 405, 411, 412, 420, 421, 424, 431], "bbox": [0.10761638733705772, 0.27709375, 0.9147113594040969, 0.9836406249999999], "iscrowd": 0, "area": 80116.65205000002, "rle": {"size": [640, 537], "counts": "XeT1`0Xc0i1XN<D;E<G9F3]@nMW=@YB[1`06T<bNcCP1:a0m;aNQDf03m0f;_N^D<MY1^;^NlD2Ed1Y;]NYEG_OP2Q;\\NgEWO^Oa2d:ZNVFgN\\OS3X:YNbFWN]Od3n9TNPHP2l7PNSHT2k7jMVHZ2f7fMZH^2c7dMYH`2d7eMWH\\2i7hMRHZ2l7lMmGW2R8mMiGU2U8QNdGR2[8RN`GP2_8TN[Go1c8WNWGk1h8ZNRGg1m8`NkF;kN^MY:]2eF2ZO^Mo9f2`FKH\\Mg9o2ZFC6[M^9Y3TF[Oe0YMV9a3oETOR1XMm8R7]GkHb8n6iGoHW8h6UHUIj7c6bHZI^7`6kH]IU7a6PI\\Io6c6VIZIj6d6Z3N1O2N2N1O2N1O2N2O0O2N2N1O2N2N1O2N1O2N2N1O2O10O01O10O01O010O1O1O10]CeJa:Z5PDXLn;P6M3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N2N2N1O2M3N1O2N2N2NWMfEmKY:P4QFkKm9R4^FhK`9V4iFeKU9X4UGbKj8\\4_G_K_8^4kG]KS8a4VHZKh7c4bHWK]7f4jHXKT7f4PIZKn6c4WI]Kg6a4]I]Kc6`4aI_K_6^4fI`KZ6^4jIaKU6\\4PJbKP6\\4SJcKm5Z4XJdKh5X4^JfKb5W4cJgK]5V4gJiKY5T4lJjKT5R4RKlKn4Q4WKnKh4o3\\KPLd4m3aKQL`4j3fKTLZ4i3kKULU4i3nKVLR4k3oKSLQ4m3QLQLo3P4RLnKn3R4SLnKm3R4TLlKl3T4VLjKj3W4WLgKi3Z4WLeKi3[4n51O0010O01O001O1O0010O01O001O001O10O01O0000000010O00000000000010O0000000000001O000O10000000001O0000000000000O2O000000000000001O0000O10O100000O100000000O10000000O10O100000000O100000000O10O100000O100000000O10000000O0100000000O010000O10O10O10000O1000O10O10000O1000O010000O1000O010000O10000SD_LZ7a3TGUNi8k1mEBP:Z5L4M3M3L3N001N2O1O0O2O1O1N101N2O1O0O2O1O0OkMeFhIZ9R6oFlIP9o5XGoIf8l5cGQJ]8i5mGTJa5]NaKZ7WOVJW5jNYKk6HWJn4POZKc61ZJc4VOZK[6=ZJY4\\OZKV6d0[JP4B[KP6k0ZJi3H[Km5Q1XJc3MZKi5Y1VJ\\34WKf5d1QJV39TKf5k1nIQ3=PKe5U2jIm2`0lJe5]2hIg2d0hJe5g2bIb2j0dJc5P3`I\\2m0aJc5Y3\\IW2R1]Jb5a3YIR2U1YJe5Y:\\JaEg5a:YJ[Ei5g:YJSEi5o:XJlDj5W;]12N2N2N2N2N2O1N2N3M2N2O1N2N2N2O1N2N3M2O1N2N2N2O1N2N2N3M2N2O1N2N2N2N2N2N2N3M2N2N2G8I8G9H8K5M2O2M3M3M2N3H8C=B>SOl0K6K5J6K5J5L5K5L4K5K4L50000O10O1000O1000O0100000O01000000O0100000N2N1O2N2N2N2N2N4L4L\\bl0"}, "label": "a zebra with its mouth open , looking towards the camera"}]}
{"id": 526754, "image": "COCO_train2014_000000526754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the buttend of an unseen zebra\"."}], "task": "refering", "answer_template": "The \"the buttend of an unseen zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 224, 225, 226, 227, 243, 244, 245, 246, 262, 263, 264, 265, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 318, 319, 320, 321, 322, 337, 338, 339, 340, 356, 357, 358, 375, 376, 377, 395, 396, 414, 415, 433, 434, 435], "bbox": [0.7713221601489758, 0.43370312499999997, 1.0, 0.9820156250000001], "iscrowd": 0, "area": 26070.689900000012, "rle": {"size": [640, 537], "counts": "^US8g0Vc0\\2cM<D5L3L5L4K4L5L3L5K5L3L5aBUKd;n4UD[Ke;j4SD^Kh;g4PDbKk;a4nChKl;]4lClKo;W4jCQLQ<S6J6K5J6J5L7H9H7H4L4M3O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1OiMPG`Io8[6kGQIT8j6RHVIm7d6[H[Id7`6cH_I]7[6jHdIU7[6nHdIQ7[6RIdIm6\\6TIdIk6[6XIdIh6Z6[IeId6[6^IdIb6Z6aIeI^6Z6eIeI[6Z6gIeIX6Z6jIfIV6X6mIgIR6Y6PJfIP6X6SJgIl5X6WJgIi5X6YJgIf5X6\\JhId5V6_JiI`5W6bJhI^5V6eJiIZ5V6iJiIW5V6kJiIT5V6\\4N101N1O2O0O2N101O0O2O001O001O1O1O001O1N2O1O1O1O1O001O1O1O1N2O1O1O1O001O1O1O1O1O1N101O1O1O1O1O[G"}, "label": "the buttend of an unseen zebra"}]}
{"id": 526754, "image": "COCO_train2014_000000526754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back end of a zebra\"."}], "task": "refering", "answer_template": "The \"back end of a zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 224, 225, 226, 227, 243, 244, 245, 246, 262, 263, 264, 265, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 318, 319, 320, 321, 322, 337, 338, 339, 340, 356, 357, 358, 375, 376, 377, 395, 396, 414, 415, 433, 434, 435], "bbox": [0.7713221601489758, 0.43370312499999997, 1.0, 0.9820156250000001], "iscrowd": 0, "area": 26070.689900000012, "rle": {"size": [640, 537], "counts": "^US8g0Vc0\\2cM<D5L3L5L4K4L5L3L5K5L3L5aBUKd;n4UD[Ke;j4SD^Kh;g4PDbKk;a4nChKl;]4lClKo;W4jCQLQ<S6J6K5J6J5L7H9H7H4L4M3O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1OiMPG`Io8[6kGQIT8j6RHVIm7d6[H[Id7`6cH_I]7[6jHdIU7[6nHdIQ7[6RIdIm6\\6TIdIk6[6XIdIh6Z6[IeId6[6^IdIb6Z6aIeI^6Z6eIeI[6Z6gIeIX6Z6jIfIV6X6mIgIR6Y6PJfIP6X6SJgIl5X6WJgIi5X6YJgIf5X6\\JhId5V6_JiI`5W6bJhI^5V6eJiIZ5V6iJiIW5V6kJiIT5V6\\4N101N1O2O0O2N101O0O2O001O001O1O1O001O1N2O1O1O1O1O001O1O1O1N2O1O1O1O001O1O1O1O1O1N101O1O1O1O1O[G"}, "label": "back end of a zebra"}]}
{"id": 59816, "image": "COCO_train2014_000000059816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green couch across from a white couch\"."}], "task": "refering", "answer_template": "The \"a green couch across from a white couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.7035312499999999, 0.46463529411764704, 1.0, 0.9886117647058823], "iscrowd": 0, "area": 34295.74555, "rle": {"size": [425, 640], "counts": "bRk54k<:E;E<jNU10001O00001N1000001O000O2O001O1O1O1N2O1O1O1O1kN\\McGe2\\8bM\\G`2c8T1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O001N2O1O2N1O1O1O1N2O1O2N1O1O1O1N2O2N1O2N2N1O2M3N1O2N1O2N2N1N3N1O2N2N1O2N1N3N2N1O2N1O2N2M1000000000000O1000000000000O1000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O100001O0000000000000000000000000O101O00000000000000000000iI"}, "label": "a green couch across from a white couch"}]}
{"id": 59816, "image": "COCO_train2014_000000059816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green cushion couch with a pillow\"."}], "task": "refering", "answer_template": "The \"a green cushion couch with a pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.7035312499999999, 0.46463529411764704, 1.0, 0.9886117647058823], "iscrowd": 0, "area": 34295.74555, "rle": {"size": [425, 640], "counts": "bRk54k<:E;E<jNU10001O00001N1000001O000O2O001O1O1O1N2O1O1O1O1kN\\McGe2\\8bM\\G`2c8T1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O001N2O1O2N1O1O1O1N2O1O2N1O1O1O1N2O2N1O2N2N1O2M3N1O2N1O2N2N1N3N1O2N2N1O2N1N3N2N1O2N1O2N2M1000000000000O1000000000000O1000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O100001O0000000000000000000000000O101O00000000000000000000iI"}, "label": "a green cushion couch with a pillow"}]}
{"id": 59816, "image": "COCO_train2014_000000059816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch with a black and white pillow\"."}], "task": "refering", "answer_template": "The \"a white couch with a black and white pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.0014375, 0.3976235294117647, 0.38068749999999996, 0.9881411764705883], "iscrowd": 0, "area": 48057.47570000001, "rle": {"size": [425, 640], "counts": "bb0c0`2U4n2QMlKo2T4WNfJi1Z5S30000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000O10000000000000001O00000000000000000000001O001O1O001O1O001O2N1O2N2N1O2N1O2N2N2N2N2N3M2N2N3M3M1O00000000000000000000000000000000000000000000000000000000000O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O1O100O1O1O100^OSFiMl9W2VFhMi9X2YFgMf9Y2\\FfMc9Y2`FdMa9\\2aF_Mc9`2_F[Me9d2a0N2O001N2L4M6I?A<D:G9Fc0]OhRT5"}, "label": "a white couch with a black and white pillow"}]}
{"id": 59816, "image": "COCO_train2014_000000059816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch sitting next to a yellow couch\"."}], "task": "refering", "answer_template": "The \"a white couch sitting next to a yellow couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.0014375, 0.3976235294117647, 0.38068749999999996, 0.9881411764705883], "iscrowd": 0, "area": 48057.47570000001, "rle": {"size": [425, 640], "counts": "bb0c0`2U4n2QMlKo2T4WNfJi1Z5S30000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000O10000000000000001O00000000000000000000001O001O1O001O1O001O2N1O2N2N1O2N1O2N2N2N2N2N3M2N2N3M3M1O00000000000000000000000000000000000000000000000000000000000O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O1O100O1O1O100^OSFiMl9W2VFhMi9X2YFgMf9Y2\\FfMc9Y2`FdMa9\\2aF_Mc9`2_F[Me9d2a0N2O001N2L4M6I?A<D:G9Fc0]OhRT5"}, "label": "a white couch sitting next to a yellow couch"}]}
{"id": 59817, "image": "COCO_train2014_000000059817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the meter that you cannot see\"."}], "task": "refering", "answer_template": "The \"the meter that you cannot see\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 51], "bbox": [0.50522, 0.0013866666666666667, 0.88468, 0.20594666666666664], "iscrowd": 0, "area": 5286.077149999999, "rle": {"size": [375, 500], "counts": "ael22c;1N3N2N2N101N2O001N2O0O20000000O100000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O0011O001O001O001O001O001N2O2N3M2N3M2N3M2N7I7HjUe0"}, "label": "the meter that you cannot see"}]}
{"id": 59817, "image": "COCO_train2014_000000059817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green meter at the back\"."}], "task": "refering", "answer_template": "The \"a green meter at the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 51], "bbox": [0.50522, 0.0013866666666666667, 0.88468, 0.20594666666666664], "iscrowd": 0, "area": 5286.077149999999, "rle": {"size": [375, 500], "counts": "ael22c;1N3N2N2N101N2O001N2O0O20000000O100000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O0011O001O001O001O001O001N2O2N3M2N3M2N3M2N7I7HjUe0"}, "label": "a green meter at the back"}]}
{"id": 59817, "image": "COCO_train2014_000000059817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green parking meater with 03 : 55 on it\"."}], "task": "refering", "answer_template": "The \"a green parking meater with 03 : 55 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.0088, 0.04874666666666667, 0.9648199999999999, 1.0], "iscrowd": 0, "area": 111413.37844999997, "rle": {"size": [375, 500], "counts": "^i11d;>C=B9H1N2O1N2O1N2O1N2N2O1N3N1N2O1N2O1N2O1N2O1N2O1N2O2M2O1O1N2O1O1O1N2O1O1N2O1O2M2O1O1O1N2O1O1N2O1O1O1N3N1O1N2O1O1O1N2O1O1N2O1O1N3N1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1000000O100000O01000000O1000000O1000000O1000000O100000000O1000000O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2M3L4L4QOYJkKj5Q4jJ[KZ5b4o0L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4N2O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O2O0O100O100O100O100O100O100O100O100O10000000000O100000000000O1O100O1O100O010O1O100O100O1O100O1O100O010O1O100O1O100O100O100O100O010O100O100O100O100O100O10O0100O100O100O100O100O100O10O0100O100O100O100O100O100O010O100O100O2O0O100O100O2O0O100O100O2O0O100O100O3N2M101N101N101N100O2O0O2N1O2O0O2N1O101N1O2O0O2N1O2O0O2N1O101N1O2O0O2N1O2O0O2N100O2O0O2O0O2O001N100O2O0O2O0O2O001N101N100O2O0O2O0O2O1O1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O1M4L3M3N3L3M3M3N3L3M3M4M2M3M3M4M2M3M3M4L3N2M4L3M3N2M4L3M3N3L3M3M6K6G8H9F9G:F9H9F9G:F9Hn]6"}, "label": "a green parking meater with 03 : 55 on it"}]}
{"id": 59817, "image": "COCO_train2014_000000059817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pay meter\"."}], "task": "refering", "answer_template": "The \"a pay meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.0088, 0.04874666666666667, 0.9648199999999999, 1.0], "iscrowd": 0, "area": 111413.37844999997, "rle": {"size": [375, 500], "counts": "^i11d;>C=B9H1N2O1N2O1N2O1N2N2O1N3N1N2O1N2O1N2O1N2O1N2O1N2O2M2O1O1N2O1O1O1N2O1O1N2O1O2M2O1O1O1N2O1O1N2O1O1O1N3N1O1N2O1O1O1N2O1O1N2O1O1N3N1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1000000O100000O01000000O1000000O1000000O1000000O100000000O1000000O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2M3L4L4QOYJkKj5Q4jJ[KZ5b4o0L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4N2O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O2O0O100O100O100O100O100O100O100O100O10000000000O100000000000O1O100O1O100O010O1O100O100O1O100O1O100O010O1O100O1O100O100O100O100O010O100O100O100O100O100O10O0100O100O100O100O100O100O10O0100O100O100O100O100O100O010O100O100O2O0O100O100O2O0O100O100O2O0O100O100O3N2M101N101N101N100O2O0O2N1O2O0O2N1O101N1O2O0O2N1O2O0O2N1O101N1O2O0O2N1O2O0O2N100O2O0O2O0O2O001N100O2O0O2O0O2O001N101N100O2O0O2O0O2O1O1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O1M4L3M3N3L3M3M3N3L3M3M4M2M3M3M4M2M3M3M4L3N2M4L3M3N2M4L3M3N3L3M3M6K6G8H9F9G:F9H9F9G:F9Hn]6"}, "label": "a pay meter"}]}
{"id": 321960, "image": "COCO_train2014_000000321960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop on the left\"."}], "task": "refering", "answer_template": "The \"the laptop on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 352, 369, 370, 371, 372, 373, 374, 375, 392, 393, 394, 395, 396, 397, 398, 414, 415, 416, 417, 418, 419, 420, 421, 437, 438, 439, 440, 441, 442, 443, 460, 461, 462, 463, 464, 465, 466, 483, 484, 485, 486, 487, 488, 489], "bbox": [0.040109375, 0.42646875, 0.32887500000000003, 0.955875], "iscrowd": 0, "area": 57444.63255, "rle": {"size": [640, 640], "counts": "Zh`0j2R>T3mLS3^Nb1001O00000O101O0000000000000000000000000000000000000000000000000O100000001O0000000000000000000000000000000000000000000O10000000000000000000001O0000000000000000000000000000000O100000000000000000000000000000001O0000000000000000000O10000000000000000000000000000000000000000000001O0O100O100O100O100O100O100O1O1K5J6K5J6K5J6K5J6K5J6K5J6K5J6K5J6K5J7J5J6A?XOh0WOi0WOi0XOh0WOi0XOh0WOQ`\\8"}, "label": "the laptop on the left"}]}
{"id": 321960, "image": "COCO_train2014_000000321960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop on the left\"."}], "task": "refering", "answer_template": "The \"the laptop on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 352, 369, 370, 371, 372, 373, 374, 375, 392, 393, 394, 395, 396, 397, 398, 414, 415, 416, 417, 418, 419, 420, 421, 437, 438, 439, 440, 441, 442, 443, 460, 461, 462, 463, 464, 465, 466, 483, 484, 485, 486, 487, 488, 489], "bbox": [0.040109375, 0.42646875, 0.32887500000000003, 0.955875], "iscrowd": 0, "area": 57444.63255, "rle": {"size": [640, 640], "counts": "Zh`0j2R>T3mLS3^Nb1001O00000O101O0000000000000000000000000000000000000000000000000O100000001O0000000000000000000000000000000000000000000O10000000000000000000001O0000000000000000000000000000000O100000000000000000000000000000001O0000000000000000000O10000000000000000000000000000000000000000000001O0O100O100O100O100O100O100O1O1K5J6K5J6K5J6K5J6K5J6K5J6K5J6K5J6K5J7J5J6A?XOh0WOi0WOi0XOh0WOi0XOh0WOQ`\\8"}, "label": "the laptop on the left"}]}
{"id": 281003, "image": "COCO_train2014_000000281003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old type oven in a kitchen\"."}], "task": "refering", "answer_template": "The \"an old type oven in a kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.295703125, 0.7587353629976581, 0.723546875, 1.0], "iscrowd": 0, "area": 26059.6895, "rle": {"size": [427, 640], "counts": "ZT_23X=;E;E;E;E;E;E<D:F2N2M2O0000000000000000000000000000000000000000000000000000000000000000O10000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000O10000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000000004L?A>B?A?A>B?Ak\\Y2"}, "label": "an old type oven in a kitchen"}]}
{"id": 281003, "image": "COCO_train2014_000000281003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lower part of the oven\"."}], "task": "refering", "answer_template": "The \"lower part of the oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.295703125, 0.7587353629976581, 0.723546875, 1.0], "iscrowd": 0, "area": 26059.6895, "rle": {"size": [427, 640], "counts": "ZT_23X=;E;E;E;E;E;E<D:F2N2M2O0000000000000000000000000000000000000000000000000000000000000000O10000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000O10000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000000004L?A>B?A?A>B?Ak\\Y2"}, "label": "lower part of the oven"}]}
{"id": 436649, "image": "COCO_train2014_000000436649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in his kitchen\"."}], "task": "refering", "answer_template": "The \"a man sitting in his kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 105, 106, 127, 128, 129, 130, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245, 267, 268], "bbox": [0.45403125, 0.28502777777777777, 0.795828125, 0.9285], "iscrowd": 0, "area": 16857.862349999996, "rle": {"size": [360, 640], "counts": "S`V36m:6L4O1N2O2M2O1N101O001O00000O10O2O2N1O1O1O1O2N1O0000000010O000000000000000001O01OOO2O1N2O1N101N2N2O1N101N2N2O0O2O1N2N100O010N1O101O003Mj0WOi0WO9F4M3lH_L[6e3aI^L]6e3^I_L_6e3]I^La6e3[I^Lb6V4N2M3N2M2WOcKfJ^4Y5hK`JZ4_5i0N2N2N3M2N2N3M2O1O101N1O100O2O0O1O100O2O0O100O1O100O2O1N10YJcKc4\\4ZKfKg4Z4VKhKj4Y4TKiKk4[4oJgKQ5[4kJgKU5[4gJhKY5Y4cJiK]5Z4]JjKe3]OSNa5l1jJjMW5U2nJfMR5Z2nJfMS5X2mJiMS5W2lJiMU5V2kJkMV5T2iJlMX5S2iJlMY5S2fJnM[5P2eJoM^5P2aJoMb5o1_JoMd5P2[JnMi5P2WJoMl5P2SJoMP6P2oIoMS6P2nInMU6P2kIoMX6o1hIPN[6n1eIQN^6m1cIQN`6l1bIRNa6l1`ISNb6k1_ISNd6k1]ITNf6i1[IVNi6c1[I[Nj6^1[I_Nj6Z1ZIdNk6h0fIVO^6h0bITOf6g0[IQOP7l0h1N2M3N001N2O1O1N10000O2O0O10000O10000000000O10000000000000000O01O100O1O1O1O10O01O1K4101N10000O100O10001N100O100O2N1O2N1O2N2N10bk]1"}, "label": "a man sitting in his kitchen"}]}
{"id": 436649, "image": "COCO_train2014_000000436649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaning back in a kitchen\"."}], "task": "refering", "answer_template": "The \"a man leaning back in a kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 105, 106, 127, 128, 129, 130, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245, 267, 268], "bbox": [0.45403125, 0.28502777777777777, 0.795828125, 0.9285], "iscrowd": 0, "area": 16857.862349999996, "rle": {"size": [360, 640], "counts": "S`V36m:6L4O1N2O2M2O1N101O001O00000O10O2O2N1O1O1O1O2N1O0000000010O000000000000000001O01OOO2O1N2O1N101N2N2O1N101N2N2O0O2O1N2N100O010N1O101O003Mj0WOi0WO9F4M3lH_L[6e3aI^L]6e3^I_L_6e3]I^La6e3[I^Lb6V4N2M3N2M2WOcKfJ^4Y5hK`JZ4_5i0N2N2N3M2N2N3M2O1O101N1O100O2O0O1O100O2O0O100O1O100O2O1N10YJcKc4\\4ZKfKg4Z4VKhKj4Y4TKiKk4[4oJgKQ5[4kJgKU5[4gJhKY5Y4cJiK]5Z4]JjKe3]OSNa5l1jJjMW5U2nJfMR5Z2nJfMS5X2mJiMS5W2lJiMU5V2kJkMV5T2iJlMX5S2iJlMY5S2fJnM[5P2eJoM^5P2aJoMb5o1_JoMd5P2[JnMi5P2WJoMl5P2SJoMP6P2oIoMS6P2nInMU6P2kIoMX6o1hIPN[6n1eIQN^6m1cIQN`6l1bIRNa6l1`ISNb6k1_ISNd6k1]ITNf6i1[IVNi6c1[I[Nj6^1[I_Nj6Z1ZIdNk6h0fIVO^6h0bITOf6g0[IQOP7l0h1N2M3N001N2O1O1N10000O2O0O10000O10000000000O10000000000000000O01O100O1O1O1O10O01O1K4101N10000O100O10001N100O100O2N1O2N1O2N2N10bk]1"}, "label": "a man leaning back in a kitchen"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza closest to the camera\"."}], "task": "refering", "answer_template": "The \"the pizza closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.09, 0.6258750000000001, 0.8576406249999999, 0.9882291666666667], "iscrowd": 0, "area": 72149.28899999999, "rle": {"size": [480, 640], "counts": "]dk0:Z><E;D<E;E;D<E;D<O1O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O101N10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000000000O10000000000000000000000O1000000000000000001O000O10000000000000000000000O1000000001O00000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000001O0000000000000000000000001O000000001O00000000001O00000000001O0001O0001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O001O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2K4K5J6K6I?A?B>A?BfWZ1"}, "label": "the pizza closest to the camera"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a thin crust pizza with pepperoni\"."}], "task": "refering", "answer_template": "The \"a thin crust pizza with pepperoni\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.09, 0.6258750000000001, 0.8576406249999999, 0.9882291666666667], "iscrowd": 0, "area": 72149.28899999999, "rle": {"size": [480, 640], "counts": "]dk0:Z><E;D<E;E;D<E;D<O1O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O101N10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000000000O10000000000000000000000O1000000000000000001O000O10000000000000000000000O1000000001O00000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000001O0000000000000000000000001O000000001O00000000001O00000000001O0001O0001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O001O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2K4K5J6K6I?A?B>A?BfWZ1"}, "label": "a thin crust pizza with pepperoni"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bottle of wine\"."}], "task": "refering", "answer_template": "The \"a green bottle of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 49, 50, 72, 73, 95, 96, 97, 118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 190, 211, 212], "bbox": [0.15337499999999998, 0.08370833333333333, 0.266296875, 0.5623541666666667], "iscrowd": 0, "area": 10391.484949999998, "rle": {"size": [480, 640], "counts": "eS^16Z>e0[Oe0dLeNZHo1`7WNSHV2h7nMkG`2P8eMbGi2X8\\M\\GQ3_8TMSGZ3h8kLjFc3Q9[1J6K5K5K5J6K5K000O1000001N11O1O1O1O1^JTHn3n7aKbH_4^7RKRIl4c8N2N2N1O2N2O1N1O2N2N2N2M2O2N2N2M2O2N3M3L4M3M3M3M3M3L4L4M3L4L4M3L4L4M3L4L8I:E<D;F;ChRl6"}, "label": "a green bottle of wine"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine bottle\"."}], "task": "refering", "answer_template": "The \"wine bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 49, 50, 72, 73, 95, 96, 97, 118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 190, 211, 212], "bbox": [0.15337499999999998, 0.08370833333333333, 0.266296875, 0.5623541666666667], "iscrowd": 0, "area": 10391.484949999998, "rle": {"size": [480, 640], "counts": "eS^16Z>e0[Oe0dLeNZHo1`7WNSHV2h7nMkG`2P8eMbGi2X8\\M\\GQ3_8TMSGZ3h8kLjFc3Q9[1J6K5K5K5J6K5K000O1000001N11O1O1O1O1^JTHn3n7aKbH_4^7RKRIl4c8N2N2N1O2N2O1N1O2N2N2N2M2O2N2N2M2O2N3M3L4M3M3M3M3M3L4L4M3L4L4M3L4L4M3L4L8I:E<D;F;ChRl6"}, "label": "wine bottle"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"very functional knife , black with 3 red stripes on the handle , which will be useful to cut the pizza\"."}], "task": "refering", "answer_template": "The \"very functional knife , black with 3 red stripes on the handle , which will be useful to cut the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 299, 300], "bbox": [0.0110625, 0.6167916666666666, 0.370234375, 0.7843125000000001], "iscrowd": 0, "area": 4167.036100000001, "rle": {"size": [480, 640], "counts": "[d31n>1O2N2M201N2O001O0O2O1O0O2O001N2O001N2O000O10O1000O01000O01000O01000O01000O010000O010O10O10O10O10O10O10O10O10O10O10O1000O01000O01000O0100O01000O01000O10O10O10O10O10O10O10O10O10O10O100O01000O01000O01000O01000O01000O01000O10O10O10000O100O10000O1000O00100O010O10O0100O0010O0100O010O1O010O010O10O0100O0010O01000O010O1000O01000O0100O01000O010O1000O01000O0100O01000O1000O0100O1000O0100O1000O010000O100O10000O10000O1M3M3M3M300001O000000000001O01O01O001O000010O000dRm5"}, "label": "very functional knife , black with 3 red stripes on the handle , which will be useful to cut the pizza"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tall glass bottle of dark yellow liquid and water\"."}], "task": "refering", "answer_template": "The \"tall glass bottle of dark yellow liquid and water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 43, 44, 65, 66, 67, 88, 89, 90, 91, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 228, 249], "bbox": [0.8022500000000001, 0.0022500000000000003, 0.964046875, 0.5977499999999999], "iscrowd": 0, "area": 18927.5535, "rle": {"size": [480, 640], "counts": "\\g`72j>4K6K5J6K5J5L5J5L5J5ZN^NjEg1k9fNnE^1i9nNoEX1g9TORFP1e9\\OSFj0b9CWFa0`9KXF;^9h2F;F9G:J5O2N101N1O2N101N1O2N100O1O00100O1O1O10O01O1O100O1O0XOi0WOi00000000O10000000000O10000000000O100000O1000O1K5J6J6J6J6J6J6J6J6J6J9GX1hN<D8H8H9F9H8H8H8H9G8H8H8H9G8H8H8H9G8H8Hbf:"}, "label": "tall glass bottle of dark yellow liquid and water"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear bottle on table next to the parmesan cheese\"."}], "task": "refering", "answer_template": "The \"clear bottle on table next to the parmesan cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 43, 44, 65, 66, 67, 88, 89, 90, 91, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 228, 249], "bbox": [0.8022500000000001, 0.0022500000000000003, 0.964046875, 0.5977499999999999], "iscrowd": 0, "area": 18927.5535, "rle": {"size": [480, 640], "counts": "\\g`72j>4K6K5J6K5J5L5J5L5J5ZN^NjEg1k9fNnE^1i9nNoEX1g9TORFP1e9\\OSFj0b9CWFa0`9KXF;^9h2F;F9G:J5O2N101N1O2N101N1O2N100O1O00100O1O1O10O01O1O100O1O0XOi0WOi00000000O10000000000O10000000000O100000O1000O1K5J6J6J6J6J6J6J6J6J6J9GX1hN<D8H8H9F9H8H8H8H9G8H8H8H9G8H8H8H9G8H8Hbf:"}, "label": "clear bottle on table next to the parmesan cheese"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey table with a pizza and wine glasses on it\"."}], "task": "refering", "answer_template": "The \"a grey table with a pizza and wine glasses on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.3483958333333333, 0.998828125, 0.9836874999999999], "iscrowd": 0, "area": 171748.1124, "rle": {"size": [480, 640], "counts": "W=a1Y;V2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000i9"}, "label": "a grey table with a pizza and wine glasses on it"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table full of food and drinks\"."}], "task": "refering", "answer_template": "The \"a table full of food and drinks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.3483958333333333, 0.998828125, 0.9836874999999999], "iscrowd": 0, "area": 171748.1124, "rle": {"size": [480, 640], "counts": "W=a1Y;V2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000i9"}, "label": "a table full of food and drinks"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork on the yellow napkin to the right of the pizza\"."}], "task": "refering", "answer_template": "The \"a fork on the yellow napkin to the right of the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 271, 272, 295, 296, 319, 320, 342, 343, 344, 366, 367], "bbox": [0.800953125, 0.5750624999999999, 0.98928125, 0.9394791666666665], "iscrowd": 0, "area": 3073.370050000003, "rle": {"size": [480, 640], "counts": "eg`75k>6K5N201O0YO@hB`0U=AmB=R=EoB6S=LkB3V=OgB1[=0bB0k<HWC9LOn<IUC:IMT=JRC;GKX=JQCi0R=XOmBe0U=\\OkBb0h<XOXC7O>k<\\OVC60:m<@SC7O6P=DQC6N5R=FPC6J4Y=FmB7G2^=HkB6D2d=HgBd0Z=]OfB`0\\=AdB<_=DaB:`=H_B5c=LZB5h=KTB6n=KoA6R>;3N1N3N1N1010O01BfA4b>101O1O001O0O2000O0100O010O1000O0100O010O1O1O01?A0O01000O100000O010000O1O000O101O0O10001N10000O20O001O001O001N101N101N1O2N1O2O0O2N1O2N2O0O2N1N2O2M2O2M2N3N2MUk2"}, "label": "a fork on the yellow napkin to the right of the pizza"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife to the right of the pizza\"."}], "task": "refering", "answer_template": "The \"the knife to the right of the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 294, 318, 341, 342, 365, 366, 389], "bbox": [0.7483125, 0.6516875, 0.9505625, 0.9730416666666667], "iscrowd": 0, "area": 2655.4112500000006, "rle": {"size": [480, 640], "counts": "njP7c0\\>1O1O1O2N1001O1O2O0O1O1O101N1O1O1O2O0O10001O0000001O0000001O0001O0001O0000001O0000001O0001O01O0000001O0001O0001O00O10001O1O1O1O1O1O1O1O2N3M2N2N2N2N0O1000001O00000000001O00000000001O0O100000001O0000000000001O0000000O1010O00000O2N1O100O2N1O1O2O0O1O1O2N1O101Noa>"}, "label": "the knife to the right of the pizza"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"knife with blade hidden by pizza\"."}], "task": "refering", "answer_template": "The \"knife with blade hidden by pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 294, 318, 341, 342, 365, 366, 389], "bbox": [0.7483125, 0.6516875, 0.9505625, 0.9730416666666667], "iscrowd": 0, "area": 2655.4112500000006, "rle": {"size": [480, 640], "counts": "njP7c0\\>1O1O1O2N1001O1O2O0O1O1O101N1O1O1O2O0O10001O0000001O0000001O0001O0001O0000001O0000001O0001O01O0000001O0001O0001O00O10001O1O1O1O1O1O1O1O2N3M2N2N2N2N0O1000001O00000000001O00000000001O0O100000001O0000000000001O0000000O1010O00000O2N1O100O2N1O1O2O0O1O1O2N1O101Noa>"}, "label": "knife with blade hidden by pizza"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with the necklace\"."}], "task": "refering", "answer_template": "The \"the girl with the necklace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 129, 145], "bbox": [0.296625, 0.0, 0.6438124999999999, 0.40449999999999997], "iscrowd": 0, "area": 32310.15660000001, "rle": {"size": [480, 640], "counts": "eSi2>`>d0]Od0[Od0]Od0[Od0]Oc0\\Ob0_OF9G:F900101N100O2O0O101N100O2O0O2O1N2N101N2O1N2O1N2O01000000000000O1000000000000O1000000000000001O001O001O001O1O001O001O001O1O001O001O000000001O00000000000000000000000000000000000000000000000000000O1000000000001O00000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1001O00000000001O00000000001O0001O1O1O100O1O1O1O1O100O1O1O1O1O010O1O1O1O1O100O1O1O1O1O100O1O1O1O1O010O1O1O1Of0ZOn0SOn0QOo0QOZgZ3"}, "label": "the girl with the necklace"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black sweater , a blue and white top and a necklace with blonde hair\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black sweater , a blue and white top and a necklace with blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 129, 145], "bbox": [0.296625, 0.0, 0.6438124999999999, 0.40449999999999997], "iscrowd": 0, "area": 32310.15660000001, "rle": {"size": [480, 640], "counts": "eSi2>`>d0]Od0[Od0]Od0[Od0]Oc0\\Ob0_OF9G:F900101N100O2O0O101N100O2O0O2O1N2N101N2O1N2O1N2O01000000000000O1000000000000O1000000000000001O001O001O001O1O001O001O001O1O001O001O000000001O00000000000000000000000000000000000000000000000000000O1000000000001O00000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1001O00000000001O00000000001O0001O1O1O100O1O1O1O1O100O1O1O1O1O010O1O1O1O1O100O1O1O1O1O100O1O1O1O1O010O1O1O1Of0ZOn0SOn0QOo0QOZgZ3"}, "label": "a woman wearing a black sweater , a blue and white top and a necklace with blonde hair"}]}
{"id": 469427, "image": "COCO_train2014_000000469427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty metal table to the left\"."}], "task": "refering", "answer_template": "The \"empty metal table to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 75, 98, 121, 144, 145], "bbox": [0.134125, 0.1129375, 0.3070625, 0.41177083333333336], "iscrowd": 0, "area": 3561.9584500000014, "rle": {"size": [480, 640], "counts": "^\\X11n>2N2O1N100O1O100O100O1O100O100OM3L5LmU<3TjC;E8HO100000000000000O100000000000000O1000000000000O100000000000000O10O10001O<D<dDmNS8_1VGlNi8b1^FjNb9b1fEiN[:n20000O010000\\L^Ej2b:VMeEc2Z:^MlE\\2T:cMSFW2m9iMZFP2f9PN`Fk1_9UNhFd1W9]NoF]1Q9cNUGW1k8iN\\GP1d8oNcGl0\\8TOjGf0V8ZOQH?n7BXH8j7F\\H4W8YOPHa0dj_6"}, "label": "empty metal table to the left"}]}
{"id": 166328, "image": "COCO_train2014_000000166328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog with a green frisbee in its mouth\"."}], "task": "refering", "answer_template": "The \"a brown dog with a green frisbee in its mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 93, 94, 95, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 264, 265, 266, 267, 284, 285, 286, 287, 289, 290, 308, 309], "bbox": [0.0535625, 0.14103211009174313, 0.7361875, 0.8386009174311927], "iscrowd": 0, "area": 64469.18185000004, "rle": {"size": [436, 640], "counts": "]c>1b=4K5L4K5L4K5L4K5L4K4M4K5L4K5L4L4M2N3L5L4L5K4L4L4L3M000000000O1000001O0000000000000O1000001O01O00000fLQFo2o9PM\\Ff2c:F6J1O100O1O1O1O1O100O100O100O100O100O10O001O001O1O001O1O001O001]EoMa9R2[FQNe9o1XFTNh9m1UFUNk9l1QFWNo9i1nEZNR:g1kE[NU:e1iE]NW:d1eE_N[:b1bE`N^:[20001O004L1O00001O0O2N1O101N1O1O100O00100O001O10O01O10OhF\\La8e3[G_Ld8b3YGaLg8^3WGeLh8\\3UGgLj8Y3TGjLl8V3PGnLo8R3oFQMP9j300O1O100O1O010O1O010O1O00100O00100O001O010O1O010O1O00100O001O10O01O1O102N1N3N2N1O2N2N1O2N1O2N2N1O2M3N1O2N1O2N2N1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O001N2O1O1O1O100O1O1O100O2N1O1O2O0O1O1O2O0O1O2N10O01O1O010O1O0010O0bNnI[KQ6c4QJ^Kn5`4UJ_Kl5^4VJcKi5[4ZJeKe5Y4]JgKc5W4`JiK`5T4bJlK^5S4dJlK\\5R4fJ`KYO_OQ6o4iJaKVO@Q6m4kJcKTO@R6i4nJeKROAP6h4PKgKPOAP6f4SKhKnNAo5e4UKiKmNBn5c4XKjKkNBm5c4ZKjKiNCm5d4YKhKlNCk5f4YKfKlNDk5f4YKfKlNDj5h4ZKcKmNDi5i4ZKbKnNEh5j4ZK`KROBd5o4YK_KX5b4iJ\\KX5e4gJ[KY5e4hJZKX5g4gJXKZ5i4iJSKV5n4nJnJR5S5QKhJP5X5f11N2O001O0O2O1O001N101N2N1mH\\JV6e5fI^JY6d5bIaJ\\6`5cIaJ\\6a5bIaJ]6`5`IbJ_6_5`IbJ_6`5_IaJ`6a5^I`Jb6`5\\IbJc6`5\\I`Jc6b5\\I^Jc6c5]I\\Jc6f5\\IZJc6g5]IYJa6^6O1N2O1N101N2N2O1N2O0O2O1N1O200O10O10O1000O010000O01000O01000O10O1O010O10O0100O010O1O010O10OWNTJkKm5S4YJiKg5U4_JmKZ5R4iJSLQ5k3RKYLi4g3XKZLe4f3]K\\L`4c3bK^L\\4a3fK`LW4`3kKaLS4^3oKdLn3[3TLfLi3Z3YLgLe3X3]LgLc3Y3^LgLa3X3aLhL]3X3eLhLZ3V3iLjLV3T3mLlLR3R3QMnLn2P3UMoLj2P3ZMmLg2Q3]MmLc2Q3aMlL`2Q3eMlL[2S3iMkLW2S3mMjLT2T3PNjLP2d2dNYM]1f2fNXMY1i2iNTMX1l2jNRMV1n2lNoLU1P3nNnLR1R3CXL=i3g4`MUG7k8IWG4j8LYG1g8O[GOe80^GNb82aGJa85aGH`88bGDa8;aGAa8?aG]Ob8a0`G\\Ob8d0_GYOd8f0]GWOe8i0\\GSOh8l0ZGPOi8o0XGnNj8R1WGkNl8S1VGjNl8V1UGgNn8Y1f101O001O1O00001O00001POdCg0[<WOgCi0Y<UOiCk0W<SOkCm0^<01O0000001O0001O01O000000001O0O1O1N3N1N3N1O2M3N2M3N1N3NfkW2"}, "label": "a brown dog with a green frisbee in its mouth"}]}
{"id": 166328, "image": "COCO_train2014_000000166328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog with the green frisbee\"."}], "task": "refering", "answer_template": "The \"the dog with the green frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 93, 94, 95, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 264, 265, 266, 267, 284, 285, 286, 287, 289, 290, 308, 309], "bbox": [0.0535625, 0.14103211009174313, 0.7361875, 0.8386009174311927], "iscrowd": 0, "area": 64469.18185000004, "rle": {"size": [436, 640], "counts": "]c>1b=4K5L4K5L4K5L4K5L4K4M4K5L4K5L4L4M2N3L5L4L5K4L4L4L3M000000000O1000001O0000000000000O1000001O01O00000fLQFo2o9PM\\Ff2c:F6J1O100O1O1O1O1O100O100O100O100O100O10O001O001O1O001O1O001O001]EoMa9R2[FQNe9o1XFTNh9m1UFUNk9l1QFWNo9i1nEZNR:g1kE[NU:e1iE]NW:d1eE_N[:b1bE`N^:[20001O004L1O00001O0O2N1O101N1O1O100O00100O001O10O01O10OhF\\La8e3[G_Ld8b3YGaLg8^3WGeLh8\\3UGgLj8Y3TGjLl8V3PGnLo8R3oFQMP9j300O1O100O1O010O1O010O1O00100O00100O001O010O1O010O1O00100O001O10O01O1O102N1N3N2N1O2N2N1O2N1O2N2N1O2M3N1O2N1O2N2N1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O001N2O1O1O1O100O1O1O100O2N1O1O2O0O1O1O2O0O1O2N10O01O1O010O1O0010O0bNnI[KQ6c4QJ^Kn5`4UJ_Kl5^4VJcKi5[4ZJeKe5Y4]JgKc5W4`JiK`5T4bJlK^5S4dJlK\\5R4fJ`KYO_OQ6o4iJaKVO@Q6m4kJcKTO@R6i4nJeKROAP6h4PKgKPOAP6f4SKhKnNAo5e4UKiKmNBn5c4XKjKkNBm5c4ZKjKiNCm5d4YKhKlNCk5f4YKfKlNDk5f4YKfKlNDj5h4ZKcKmNDi5i4ZKbKnNEh5j4ZK`KROBd5o4YK_KX5b4iJ\\KX5e4gJ[KY5e4hJZKX5g4gJXKZ5i4iJSKV5n4nJnJR5S5QKhJP5X5f11N2O001O0O2O1O001N101N2N1mH\\JV6e5fI^JY6d5bIaJ\\6`5cIaJ\\6a5bIaJ]6`5`IbJ_6_5`IbJ_6`5_IaJ`6a5^I`Jb6`5\\IbJc6`5\\I`Jc6b5\\I^Jc6c5]I\\Jc6f5\\IZJc6g5]IYJa6^6O1N2O1N101N2N2O1N2O0O2O1N1O200O10O10O1000O010000O01000O01000O10O1O010O10O0100O010O1O010O10OWNTJkKm5S4YJiKg5U4_JmKZ5R4iJSLQ5k3RKYLi4g3XKZLe4f3]K\\L`4c3bK^L\\4a3fK`LW4`3kKaLS4^3oKdLn3[3TLfLi3Z3YLgLe3X3]LgLc3Y3^LgLa3X3aLhL]3X3eLhLZ3V3iLjLV3T3mLlLR3R3QMnLn2P3UMoLj2P3ZMmLg2Q3]MmLc2Q3aMlL`2Q3eMlL[2S3iMkLW2S3mMjLT2T3PNjLP2d2dNYM]1f2fNXMY1i2iNTMX1l2jNRMV1n2lNoLU1P3nNnLR1R3CXL=i3g4`MUG7k8IWG4j8LYG1g8O[GOe80^GNb82aGJa85aGH`88bGDa8;aGAa8?aG]Ob8a0`G\\Ob8d0_GYOd8f0]GWOe8i0\\GSOh8l0ZGPOi8o0XGnNj8R1WGkNl8S1VGjNl8V1UGgNn8Y1f101O001O1O00001O00001POdCg0[<WOgCi0Y<UOiCk0W<SOkCm0^<01O0000001O0001O01O000000001O0O1O1N3N1N3N1O2M3N2M3N1N3NfkW2"}, "label": "the dog with the green frisbee"}]}
{"id": 559544, "image": "COCO_train2014_000000559544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bathtub\"."}], "task": "refering", "answer_template": "The \"a white bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222], "bbox": [0.35503125, 0.436953125, 0.7596875000000001, 0.6720833333333333], "iscrowd": 0, "area": 17409.2042, "rle": {"size": [384, 640], "counts": "YZe21m;4L4K5L4L4L4L3O2N2M3N2M3N1N2O001N100O2O0O101O0O101N10000O2O0O100O101O0O100O101N10000O100000000O100000000O100[FPNZ9P2cFSN]9m1aFVN^9j1`FXN_9i1_FYNa9S200000001O0000000000001O0000000000001O0000000000001O0001O00000001O00000000000O2O0000000001O01O00000001O00010O1O001O000001O0000000000001O0000000001O01O01O001O00LbMfF^2Z95O0000000000000000000000000000000000D<1O0000001O00012M0000000001O0000000O2O0000000000001O00000000001O0O10000000001O0000000000001O000O1000001O0000000O10001O001N101O001O1O0O2O001O1O0O2O001O1O0O2O001O001N2O001O001O1N101M2N3M3N1N3M6J5L4K5K5LWai1"}, "label": "a white bathtub"}]}
{"id": 559544, "image": "COCO_train2014_000000559544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bathtub\"."}], "task": "refering", "answer_template": "The \"a bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222], "bbox": [0.35503125, 0.436953125, 0.7596875000000001, 0.6720833333333333], "iscrowd": 0, "area": 17409.2042, "rle": {"size": [384, 640], "counts": "YZe21m;4L4K5L4L4L4L3O2N2M3N2M3N1N2O001N100O2O0O101O0O101N10000O2O0O100O101O0O100O101N10000O100000000O100000000O100[FPNZ9P2cFSN]9m1aFVN^9j1`FXN_9i1_FYNa9S200000001O0000000000001O0000000000001O0000000000001O0001O00000001O00000000000O2O0000000001O01O00000001O00010O1O001O000001O0000000000001O0000000001O01O01O001O00LbMfF^2Z95O0000000000000000000000000000000000D<1O0000001O00012M0000000001O0000000O2O0000000000001O00000000001O0O10000000001O0000000000001O000O1000001O0000000O10001O001N101O001O1O0O2O001O1O0O2O001O1O0O2O001O001N2O001O001O1N101M2N3M3N1N3M6J5L4K5K5LWai1"}, "label": "a bathtub"}]}
{"id": 387513, "image": "COCO_train2014_000000387513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three carrot sticks in bowl\"."}], "task": "refering", "answer_template": "The \"three carrot sticks in bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 100, 101, 102, 123, 124, 125, 146, 147, 148, 170, 171, 173, 174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 264, 265, 266, 267, 268, 283, 284, 288, 289, 290, 306, 307, 312, 313, 314, 329, 330, 336, 337, 338], "bbox": [0.301953125, 0.1613785046728972, 0.713578125, 0.9992056074766354], "iscrowd": 0, "area": 30245.29404999999, "rle": {"size": [428, 640], "counts": "^Ra21W=4M3L4L4M3L4M4K4M3L4L4M3L4M3L4M4K4L4M3L4M3L4M3L4M4K4L4M3L4M3L4M3L4L5L3L4M3L4M3L2OO00001O001O001O00001O00bLPLeMo3_2]MQLc2S4lM[KS2i4PNQKn1T5UNgJi1]5ZN]Je1g5_NRJ`1R6cNiI[1Z6h24L4L4L4L4L4L4L3M4L4L4L4L4LL4ZOf0G94L4L4L4L4L4N22N2N2N2N2N2N2N2N2N2N1O2N1O1O2N1O2N1O1O[L]Ic0a6XOaJL^5O[K_Oc4<dKC[48lKGS44SLMk3O[L0d3KcL4\\3HiL9U3CQM<n2@XM`0g2[O^Me0a2YOcMf0\\2XOhMh0V2VOmMj0S2RORNn0l1POXNo0g1oN\\NQ1c1mNaNS1^1iNfNW1Y1eNmN[1Q1bNTO]1k0`NYO`1g0\\N^Od1`0YNEf1:VNKk13RN2m1h5O010O10O01O10O0100O00100O010OcMXNUIh1j6\\NTId1j6`NTI_1l6dNQI\\1n6hNPIW1o6mNoHR1P7ROmHn0S7UOkHj0T7ZOjHe0U7_OhHa0W7CgH<Y7GeH9Y7KdH5[7OcH0]73aHL^78_HH`7;_HD`7`0^H_Ob7d0[H\\Od7h0ZHWOe7m0YHROf7R1WHnNi7U1UHjNj7Z1SHfNl7^1RHbNm7a1QH^Nn7f1oGZNP8j1nGUNQ8o1mGPNS8S2jGmMU8W2iGhMV8\\2hGcMW8a2fG_MZ8j3O1O100O2N1O100O1O1O100O1O100001O2O1N2N2N2N2N2N2N2kM_GZOc8c0_G\\Ob8b0_G^Oc8?^GAd8<]GDe88^GHc85^GKd82]GNe8O\\G1f8L\\G3f8I\\G7f8F[G:f8D[G;h8A[G>g8_OZG`0i8\\OYGc0j8YOXGg0j8VOWGi0l8SOWGl0k8QOVGn0m8nNUGQ1n8kNTGU1m8iNUGU1n8gNTGX1o8eNRG[1P9aNRG^1Q9_NQG_1_:00O101N1O100O2N100O101N1O1O1O1N3M2N2O1N2N2O2M2N2N2O1N2N3M2O1N2N2N2OhR\\2"}, "label": "three carrot sticks in bowl"}]}
{"id": 387513, "image": "COCO_train2014_000000387513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"thin slices of carrots\"."}], "task": "refering", "answer_template": "The \"thin slices of carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 100, 101, 102, 123, 124, 125, 146, 147, 148, 170, 171, 173, 174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 264, 265, 266, 267, 268, 283, 284, 288, 289, 290, 306, 307, 312, 313, 314, 329, 330, 336, 337, 338], "bbox": [0.301953125, 0.1613785046728972, 0.713578125, 0.9992056074766354], "iscrowd": 0, "area": 30245.29404999999, "rle": {"size": [428, 640], "counts": "^Ra21W=4M3L4L4M3L4M4K4M3L4L4M3L4M3L4M4K4L4M3L4M3L4M3L4M4K4L4M3L4M3L4M3L4L5L3L4M3L4M3L2OO00001O001O001O00001O00bLPLeMo3_2]MQLc2S4lM[KS2i4PNQKn1T5UNgJi1]5ZN]Je1g5_NRJ`1R6cNiI[1Z6h24L4L4L4L4L4L4L3M4L4L4L4L4LL4ZOf0G94L4L4L4L4L4N22N2N2N2N2N2N2N2N2N2N1O2N1O1O2N1O2N1O1O[L]Ic0a6XOaJL^5O[K_Oc4<dKC[48lKGS44SLMk3O[L0d3KcL4\\3HiL9U3CQM<n2@XM`0g2[O^Me0a2YOcMf0\\2XOhMh0V2VOmMj0S2RORNn0l1POXNo0g1oN\\NQ1c1mNaNS1^1iNfNW1Y1eNmN[1Q1bNTO]1k0`NYO`1g0\\N^Od1`0YNEf1:VNKk13RN2m1h5O010O10O01O10O0100O00100O010OcMXNUIh1j6\\NTId1j6`NTI_1l6dNQI\\1n6hNPIW1o6mNoHR1P7ROmHn0S7UOkHj0T7ZOjHe0U7_OhHa0W7CgH<Y7GeH9Y7KdH5[7OcH0]73aHL^78_HH`7;_HD`7`0^H_Ob7d0[H\\Od7h0ZHWOe7m0YHROf7R1WHnNi7U1UHjNj7Z1SHfNl7^1RHbNm7a1QH^Nn7f1oGZNP8j1nGUNQ8o1mGPNS8S2jGmMU8W2iGhMV8\\2hGcMW8a2fG_MZ8j3O1O100O2N1O100O1O1O100O1O100001O2O1N2N2N2N2N2N2N2kM_GZOc8c0_G\\Ob8b0_G^Oc8?^GAd8<]GDe88^GHc85^GKd82]GNe8O\\G1f8L\\G3f8I\\G7f8F[G:f8D[G;h8A[G>g8_OZG`0i8\\OYGc0j8YOXGg0j8VOWGi0l8SOWGl0k8QOVGn0m8nNUGQ1n8kNTGU1m8iNUGU1n8gNTGX1o8eNRG[1P9aNRG^1Q9_NQG_1_:00O101N1O100O2N100O101N1O1O1O1N3M2N2O1N2N2O2M2N2N2O1N2N3M2O1N2N2N2OhR\\2"}, "label": "thin slices of carrots"}]}
{"id": 387513, "image": "COCO_train2014_000000387513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large stalk of broccoli in the white bowl\"."}], "task": "refering", "answer_template": "The \"a large stalk of broccoli in the white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 340, 341], "bbox": [0.645140625, 0.5082242990654205, 0.939390625, 0.9694158878504672], "iscrowd": 0, "area": 24810.2163, "rle": {"size": [428, 640], "counts": "bn\\51Z=3M2N2N2N3M2N2N3M2N2O2M2N2N2N3M2N2N3bMhNcHZ1W7lNgHV1T7POiHR1R7SOlHP1P7SOoHn0P7SOnHo0P7SOnHP1P7QOnHQ1Q7QOlHQ1S7POkHS1S7nNkHT1T7mNjHU1U7lNiHV1V7lNgHW1W7jNgHX1X7iNfHX1Z7iNeHW1[7jNcHW1]7jNaHW1_7kN^HW1a7jN\\HX1d7iNYHY1g7hNVHZ1j7gNRH\\1n7eNoG]1Q8eNkG]1U8dNgG`1X8aNeGa1[8`NbGb1]8g10000O100O2O0O100O10000O2O0O100O100O101O0O100O100O101O0O1000000O101O000O10000000010O01O01000O010O0100O04M2M4M3M2M4M2NO001N1O2O1N101N2N101N101N2O0O2O1N101N010O010O01O010O0010O01O010O0010O010O01O01O01O01O01O010O00010O01O01O01O01O1jLSHm0m7nNZHo0g7mN^HQ1c7jNdHT1\\7hNjHU1W7fNPIW1Q7eNTIW1o6dNXIX1j6dN\\IW1g6eN_IW1d6fN_IW1c6gNaIT1b6iNcIS1_6kNeIP1^6mNgIo0[6oNhIn0Z6oNkIl0X6ROlIj0V6SOoIi0S6UOPJg0T6UOQJg0]9K5L4L3L5LVo?"}, "label": "a large stalk of broccoli in the white bowl"}]}
{"id": 387513, "image": "COCO_train2014_000000387513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoli in stirfry\"."}], "task": "refering", "answer_template": "The \"a piece of broccoli in stirfry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 340, 341], "bbox": [0.645140625, 0.5082242990654205, 0.939390625, 0.9694158878504672], "iscrowd": 0, "area": 24810.2163, "rle": {"size": [428, 640], "counts": "bn\\51Z=3M2N2N2N3M2N2N3M2N2O2M2N2N2N3M2N2N3bMhNcHZ1W7lNgHV1T7POiHR1R7SOlHP1P7SOoHn0P7SOnHo0P7SOnHP1P7QOnHQ1Q7QOlHQ1S7POkHS1S7nNkHT1T7mNjHU1U7lNiHV1V7lNgHW1W7jNgHX1X7iNfHX1Z7iNeHW1[7jNcHW1]7jNaHW1_7kN^HW1a7jN\\HX1d7iNYHY1g7hNVHZ1j7gNRH\\1n7eNoG]1Q8eNkG]1U8dNgG`1X8aNeGa1[8`NbGb1]8g10000O100O2O0O100O10000O2O0O100O100O101O0O100O100O101O0O1000000O101O000O10000000010O01O01000O010O0100O04M2M4M3M2M4M2NO001N1O2O1N101N2N101N101N2O0O2O1N101N010O010O01O010O0010O01O010O0010O010O01O01O01O01O01O010O00010O01O01O01O01O1jLSHm0m7nNZHo0g7mN^HQ1c7jNdHT1\\7hNjHU1W7fNPIW1Q7eNTIW1o6dNXIX1j6dN\\IW1g6eN_IW1d6fN_IW1c6gNaIT1b6iNcIS1_6kNeIP1^6mNgIo0[6oNhIn0Z6oNkIl0X6ROlIj0V6SOoIi0S6UOPJg0T6UOQJg0]9K5L4L3L5LVo?"}, "label": "a piece of broccoli in stirfry"}]}
{"id": 289211, "image": "COCO_train2014_000000289211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple book on table\"."}], "task": "refering", "answer_template": "The \"purple book on table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [307, 308, 309, 310, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 374, 375, 376, 377], "bbox": [0.0011458333333333333, 0.78959375, 0.3115416666666667, 0.999], "iscrowd": 0, "area": 10571.768699999999, "rle": {"size": [640, 480], "counts": "oU1>cc0?A:FO1O1O010O1O100O1O2N101N1O101N1O2N101N1O2O0O1O2N100O2N1O10hMHh@6W?Mh@2W?2g@NW?5i@IV?;g@EX?>g@AX?b0f@_OX?e0f@ZOY?i0e@WOZ?m0b@UO\\?n0a@SO^?Q1^@POa?S1\\@nNb?V1[@lNc?X1Y@iNf?Z1W@gNh?]1T@eNj?^1S@cNl?a1P@`No?c1n_O^NQ`0e1l_O]NR`0k1f_OVNY`0i2O1O2N1O1O2N1O1O1O2N1O1O2N1O11O010O000010O01O01O01O0010O00010O0001O0O2M2O1O2M2O2N1N2O2N1N3N1O1N3N1O1N3N1O2M2O1O2M2O2N1N2O2N1N2O2N1N3N1O1N3N1O2M2O1O2M2O2N1N2O2M2O1O2M2O2N1N2O2N1N3N1O1N3N1O2[Okg]6"}, "label": "purple book on table"}]}
{"id": 289211, "image": "COCO_train2014_000000289211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple book on a table\"."}], "task": "refering", "answer_template": "The \"a purple book on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [307, 308, 309, 310, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 374, 375, 376, 377], "bbox": [0.0011458333333333333, 0.78959375, 0.3115416666666667, 0.999], "iscrowd": 0, "area": 10571.768699999999, "rle": {"size": [640, 480], "counts": "oU1>cc0?A:FO1O1O010O1O100O1O2N101N1O101N1O2N101N1O2O0O1O2N100O2N1O10hMHh@6W?Mh@2W?2g@NW?5i@IV?;g@EX?>g@AX?b0f@_OX?e0f@ZOY?i0e@WOZ?m0b@UO\\?n0a@SO^?Q1^@POa?S1\\@nNb?V1[@lNc?X1Y@iNf?Z1W@gNh?]1T@eNj?^1S@cNl?a1P@`No?c1n_O^NQ`0e1l_O]NR`0k1f_OVNY`0i2O1O2N1O1O2N1O1O1O2N1O1O2N1O11O010O000010O01O01O01O0010O00010O0001O0O2M2O1O2M2O2N1N2O2N1N3N1O1N3N1O1N3N1O2M2O1O2M2O2N1N2O2N1N2O2N1N3N1O1N3N1O2M2O1O2M2O2N1N2O2M2O1O2M2O2N1N2O2N1N3N1O1N3N1O2[Okg]6"}, "label": "a purple book on a table"}]}
{"id": 371134, "image": "COCO_train2014_000000371134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the airplane on the runway\"."}], "task": "refering", "answer_template": "The \"the airplane on the runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 108, 109, 110, 128, 129, 151, 152, 170, 171, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 289], "bbox": [0.0165, 0.27566744730679155, 0.9960781249999998, 0.8194847775175644], "iscrowd": 0, "area": 31877.088549999993, "rle": {"size": [427, 640], "counts": "fk44V=2M200000000001O00000001O000000000000000000001O0001O0000000000000001O000000000000000000010O000000000000000000001O000000000001O00000001O000000000000000000001O0001O0000000000000001O000000000000000001O01O000000000000000000001O000000000001O00000001O000000000000000000001O0001O0000000000000001O000000000000000001O01O000000000000000000001O000000000001O00000001F9F:O1oCTOa;]1M2N2N3N1N2N2N3M2N2O2N1O1O2N2N1\\EjMR:X2lEiMS:X2lEiMS:Y2lEgMT:Z2jEgMV:Y2iEiMV:f2O100000000000000O10\\OPFmMP:m1[FmMf9Q2aFhMa9X2i000O10O100\\EiMT:V2jEmMV:S2gEPNY:P2dESN\\:m1aE_NV:]2N1N3N1O2N1O00001O00O001O1O1O1O1O1O11N2O1O1O[ORFnMl9R2UFoMi9R2XFmMg9S2ZFmMe9S2\\FmMc9S2^FjMd9W2]F^Ml9b2>1O1O2M2O10000O10000O2O000O100VMnL[KR3c4PM\\KQ3a4RM_Kn2_4TMaKm2^4SMbKm2_4RMaKn2_4RMaKn2_4RM`Ko2`4QM`Ko2`4QM`Ko2`4QM`Ko2`4RM_Kn2a4RM_Ko2`4QM`Ko2`4QM`Ko2`4QM`Ko2`4QM`Ko2`4QM`KP3_4PMaKQ3^4oLbKR3]4nLcKS3\\4mLcKT3]4lLcKS3^4mLcKR3]4nLdKQ3]4nLdKP3]4PMdKo2\\4QMeKn2[4RMeKn2[4RMfKl2[4TMfKk2Z4UMgKj2Y4WMgKg2Z4YMgKf2Y4ZMgKf2Y4ZMeKh2[4XMcKi2^4WMeKf2[4ZMhKc2X4]MiKb2W4^MiKa2X4_MgKb2Y4^MgKb2Y4^MgKb2Z4]MeKc2\\4]MdKc2\\4]MdKc2\\4]MdKc2\\4]McKd2]4\\McKc2^4]MbKc2^4]MbKc2^4]MaKd2_4\\MaKc2`4^M^Kc2b4]M\\Ke2d4[MYKh2g4XMXKh2i4XMXKg2h4YMYKf2g4ZMZKe2f4[M[Kd2e4\\M\\Kb2f4]MZKc2f4]MZKc2f4]MYKd2g4\\MYKc2h4]MXKc2h4]MXKc2h4]MWKd2i4\\MWKd2i4\\MVKd2k4\\MTKe2l4[MTKe2l4[MSKf2m4ZMRKf2o4[MoJg2P5YMoJi2P5WMoJk2P5UMoJm2P5SMPKn2RNkL[67bKo2QNQMY6OeKR3oMVMW6HiKT3nMWMW6EjKV3lMWMZ6CiKX3jMWM]6AhKR4Y4nKgKR4Y4nKfKS4Z4mKeKT4[4lKdKU4\\4kKcKV4]4jKbKV4_4jK`KW4`4iK_KX4a4hK_KW4b4iK]KW4d4iK[KW4f4jKXKV4i4jKRKT3ZNgMe6UOkJZ3aNaMd6TOfJ`3hN\\M`6TOcJe3oNVM]6UO_Ji3WOQMY6VO[Jn3^OjLW6YOUJR4FbLV6\\OoIW4L\\LU6^OiI[44ULS6AcI_4<mKR6D]Id4c0eKQ6Z4\\I^Ke03R6^4ZI_Ke0OT6`4YIaKe0JU6d4WIbKf0EV6g4VIdKf7\\4[HdKc7\\4_HdK`7\\4aHcK^7]4dHcK[7]4fHcKY7\\4hHeKW7i2ZH\\N`0kNV7[2hHhN4mNT7m1VIUOGnNR7_1fICXOnNR7Q1TJ2iNmNS7c0aJa0\\NlNR7;kJi0SNlNR7:lJj0RNlNR7:lJj0RNlNQ7;mJi0RNlNQ7;mJj0PNkNT7;lJj0PNkNS7<mJi0PNkNS7<lJj0QNjNS7<lJj0QNjN[74dJR1QNjN_70`JV1QNjN_70`JV1QNjN_7OaJW1PNjN_7OaJV1QNkN^7O`JW1RNjN^7O`JW1RNjN^7O`JW1QNkN_7N`JW1QNkN_7N`JW1QNjN`7O_JW1QNjN_70`JV1QNjN_70_JW1RNiN_70_JV1SNjN^70`JU1RNkN^7OaJV1QNkN^7OaJV1QNkN^7ObJU1PNlN^7ObJU1PNlN^7ObJU1oMmN_7NbJU1oMmN_7NcJS1oMoN^7NcJR1PNPO]7NcJQ1QNPO]7ObJo0SNRO[7OcJm0SNTOY7OeJk0TNVOW7OeJi0VNXOU7OfJg0VNZOT7OfJe0XN\\OR7OfJd0YN]OQ7OgJa0YNAP7NiJ>XNDo6NjJ;YNGm6NkJ9YNIl6NmJ5YNMj6NnJ2ZNOi6NPK0XN2h6NQKMYN5f6NRKKYN7e6NTKGYN;b6OdL1\\3OdL1\\3OcL2]3NcL2]3NcL2]3NcL2]3NcL2]3MdL3\\3MdL3\\3MdL2]3NcL2]3NcL2]3NcL2]3NcL2]3NbL3]3NcL2]3NcL2]3NcL2]3NcL2]3MdL3\\3MdL3\\3MdL3\\3MdL2]3NcL2]3NcL2]3NcL2]3NbL3^3MbL3^3MbL3^3MbL3]3MdL3\\3MdL3\\3MdL3\\3MdL3\\3MdL3\\3MdL2]3NcL2]3NcL2]3NbL3^3MbL3^3MbL3^3LcL4]3LcL4]3LcL4]3LcL4\\3MdL3\\3MdL3\\3MdL3\\3MdL3\\3MdL2]3NbL3^3MbL3^3LcL4]3LcL4]3LcL4]3LcL4]3LcL4]3LcL4]3LcL4\\3MdL3\\3MdL3\\3McL4]3LcL3^3MbL3^3LcL4]3LcL4]3LcL4]3LcL4]3KdL5\\3KdL5\\3JeL6[3JeL6[3IfL7Z3IeL8]3EdL;^3CbL<W800000000000000000O10000000O100O1O10000O100000000000000000000000000O100000O1000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10O10000000000000000000000000O10000000000000003MZo0"}, "label": "the airplane on the runway"}]}
{"id": 371134, "image": "COCO_train2014_000000371134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is a picture of a white plane on a runway that appears to be taking off\"."}], "task": "refering", "answer_template": "The \"it is a picture of a white plane on a runway that appears to be taking off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 108, 109, 110, 128, 129, 151, 152, 170, 171, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 289], "bbox": [0.0165, 0.27566744730679155, 0.9960781249999998, 0.8194847775175644], "iscrowd": 0, "area": 31877.088549999993, "rle": {"size": [427, 640], "counts": "fk44V=2M200000000001O00000001O000000000000000000001O0001O0000000000000001O000000000000000000010O000000000000000000001O000000000001O00000001O000000000000000000001O0001O0000000000000001O000000000000000001O01O000000000000000000001O000000000001O00000001O000000000000000000001O0001O0000000000000001O000000000000000001O01O000000000000000000001O000000000001O00000001F9F:O1oCTOa;]1M2N2N3N1N2N2N3M2N2O2N1O1O2N2N1\\EjMR:X2lEiMS:X2lEiMS:Y2lEgMT:Z2jEgMV:Y2iEiMV:f2O100000000000000O10\\OPFmMP:m1[FmMf9Q2aFhMa9X2i000O10O100\\EiMT:V2jEmMV:S2gEPNY:P2dESN\\:m1aE_NV:]2N1N3N1O2N1O00001O00O001O1O1O1O1O1O11N2O1O1O[ORFnMl9R2UFoMi9R2XFmMg9S2ZFmMe9S2\\FmMc9S2^FjMd9W2]F^Ml9b2>1O1O2M2O10000O10000O2O000O100VMnL[KR3c4PM\\KQ3a4RM_Kn2_4TMaKm2^4SMbKm2_4RMaKn2_4RMaKn2_4RM`Ko2`4QM`Ko2`4QM`Ko2`4QM`Ko2`4RM_Kn2a4RM_Ko2`4QM`Ko2`4QM`Ko2`4QM`Ko2`4QM`Ko2`4QM`KP3_4PMaKQ3^4oLbKR3]4nLcKS3\\4mLcKT3]4lLcKS3^4mLcKR3]4nLdKQ3]4nLdKP3]4PMdKo2\\4QMeKn2[4RMeKn2[4RMfKl2[4TMfKk2Z4UMgKj2Y4WMgKg2Z4YMgKf2Y4ZMgKf2Y4ZMeKh2[4XMcKi2^4WMeKf2[4ZMhKc2X4]MiKb2W4^MiKa2X4_MgKb2Y4^MgKb2Y4^MgKb2Z4]MeKc2\\4]MdKc2\\4]MdKc2\\4]MdKc2\\4]McKd2]4\\McKc2^4]MbKc2^4]MbKc2^4]MaKd2_4\\MaKc2`4^M^Kc2b4]M\\Ke2d4[MYKh2g4XMXKh2i4XMXKg2h4YMYKf2g4ZMZKe2f4[M[Kd2e4\\M\\Kb2f4]MZKc2f4]MZKc2f4]MYKd2g4\\MYKc2h4]MXKc2h4]MXKc2h4]MWKd2i4\\MWKd2i4\\MVKd2k4\\MTKe2l4[MTKe2l4[MSKf2m4ZMRKf2o4[MoJg2P5YMoJi2P5WMoJk2P5UMoJm2P5SMPKn2RNkL[67bKo2QNQMY6OeKR3oMVMW6HiKT3nMWMW6EjKV3lMWMZ6CiKX3jMWM]6AhKR4Y4nKgKR4Y4nKfKS4Z4mKeKT4[4lKdKU4\\4kKcKV4]4jKbKV4_4jK`KW4`4iK_KX4a4hK_KW4b4iK]KW4d4iK[KW4f4jKXKV4i4jKRKT3ZNgMe6UOkJZ3aNaMd6TOfJ`3hN\\M`6TOcJe3oNVM]6UO_Ji3WOQMY6VO[Jn3^OjLW6YOUJR4FbLV6\\OoIW4L\\LU6^OiI[44ULS6AcI_4<mKR6D]Id4c0eKQ6Z4\\I^Ke03R6^4ZI_Ke0OT6`4YIaKe0JU6d4WIbKf0EV6g4VIdKf7\\4[HdKc7\\4_HdK`7\\4aHcK^7]4dHcK[7]4fHcKY7\\4hHeKW7i2ZH\\N`0kNV7[2hHhN4mNT7m1VIUOGnNR7_1fICXOnNR7Q1TJ2iNmNS7c0aJa0\\NlNR7;kJi0SNlNR7:lJj0RNlNR7:lJj0RNlNQ7;mJi0RNlNQ7;mJj0PNkNT7;lJj0PNkNS7<mJi0PNkNS7<lJj0QNjNS7<lJj0QNjN[74dJR1QNjN_70`JV1QNjN_70`JV1QNjN_7OaJW1PNjN_7OaJV1QNkN^7O`JW1RNjN^7O`JW1RNjN^7O`JW1QNkN_7N`JW1QNkN_7N`JW1QNjN`7O_JW1QNjN_70`JV1QNjN_70_JW1RNiN_70_JV1SNjN^70`JU1RNkN^7OaJV1QNkN^7OaJV1QNkN^7ObJU1PNlN^7ObJU1PNlN^7ObJU1oMmN_7NbJU1oMmN_7NcJS1oMoN^7NcJR1PNPO]7NcJQ1QNPO]7ObJo0SNRO[7OcJm0SNTOY7OeJk0TNVOW7OeJi0VNXOU7OfJg0VNZOT7OfJe0XN\\OR7OfJd0YN]OQ7OgJa0YNAP7NiJ>XNDo6NjJ;YNGm6NkJ9YNIl6NmJ5YNMj6NnJ2ZNOi6NPK0XN2h6NQKMYN5f6NRKKYN7e6NTKGYN;b6OdL1\\3OdL1\\3OcL2]3NcL2]3NcL2]3NcL2]3NcL2]3MdL3\\3MdL3\\3MdL2]3NcL2]3NcL2]3NcL2]3NcL2]3NbL3]3NcL2]3NcL2]3NcL2]3NcL2]3MdL3\\3MdL3\\3MdL3\\3MdL2]3NcL2]3NcL2]3NcL2]3NbL3^3MbL3^3MbL3^3MbL3]3MdL3\\3MdL3\\3MdL3\\3MdL3\\3MdL3\\3MdL2]3NcL2]3NcL2]3NbL3^3MbL3^3MbL3^3LcL4]3LcL4]3LcL4]3LcL4\\3MdL3\\3MdL3\\3MdL3\\3MdL3\\3MdL2]3NbL3^3MbL3^3LcL4]3LcL4]3LcL4]3LcL4]3LcL4]3LcL4]3LcL4\\3MdL3\\3MdL3\\3McL4]3LcL3^3MbL3^3LcL4]3LcL4]3LcL4]3LcL4]3KdL5\\3KdL5\\3JeL6[3JeL6[3IfL7Z3IeL8]3EdL;^3CbL<W800000000000000000O10000000O100O1O10000O100000000000000000000000000O100000O1000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10O10000000000000000000000000O10000000000000003MZo0"}, "label": "it is a picture of a white plane on a runway that appears to be taking off"}]}
{"id": 371134, "image": "COCO_train2014_000000371134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plane landing\"."}], "task": "refering", "answer_template": "The \"a plane landing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 72, 73, 74, 75, 76, 77, 78, 79, 98, 99, 100], "bbox": [0.13071875, 0.11836065573770492, 0.46706249999999994, 0.2986651053864169], "iscrowd": 0, "area": 3549.037299999999, "rle": {"size": [427, 640], "counts": "`SS13X=3M3O1O01O0N2O1N1O01O01O01O000001O01O0000000000001O0000000001O000001O0000000KMRC3n<MRC3n<MRC3n<MRC3n<MRC4m<LTC3l<MTC3l<MTC3l<MTC3l<MTC3l<MTC3l<MTC3l<NSC3l<MTC3l<MTC3l<MTC3m<LSC5l<KTC5l<KTC5l<KUC4k<LUC8g<HYC7h<IXC7h<IXC6i<JWC5j<LUC4k<LUC4k<LUC4k<LUC4k<LUC4k<LUC4k<LUC4k<LUC5j<KWC4i<LWC4i<LWC4i<LWC4i<LWC4i<MVC7f<IZC7f<IZC5Z;2fEIPO5Z;a0WEZO_O4[;R1fDjNO4[;[1dDeN_;b13M4L3M3M3M4O0O1N2N2N0O00100O14L0O00O0O31N1O00GkC[OV<d09O10000000O100000000000000N210OO11OO1O1000O100000OeC[OR<e0mC]OR<b0oC^OQ<c0oC\\OQ<d0oCYOT<o0O4MGmCZOS<d0oC\\OQ<d0oC\\OQ<c0<00O10000001O1O0bCZOW<m01O001L301O0O2O1O2N5K3M0O0GUC4k<LVC2k<NVC0k<171N01O1O100O0100000O10O1000O1000O10O100000O010000000O01000000O0100000O10O1000O1000K5D;2O0O;EaS^4"}, "label": "a plane landing"}]}
{"id": 371134, "image": "COCO_train2014_000000371134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plane in air\"."}], "task": "refering", "answer_template": "The \"plane in air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 72, 73, 74, 75, 76, 77, 78, 79, 98, 99, 100], "bbox": [0.13071875, 0.11836065573770492, 0.46706249999999994, 0.2986651053864169], "iscrowd": 0, "area": 3549.037299999999, "rle": {"size": [427, 640], "counts": "`SS13X=3M3O1O01O0N2O1N1O01O01O01O000001O01O0000000000001O0000000001O000001O0000000KMRC3n<MRC3n<MRC3n<MRC3n<MRC4m<LTC3l<MTC3l<MTC3l<MTC3l<MTC3l<MTC3l<MTC3l<NSC3l<MTC3l<MTC3l<MTC3m<LSC5l<KTC5l<KTC5l<KUC4k<LUC8g<HYC7h<IXC7h<IXC6i<JWC5j<LUC4k<LUC4k<LUC4k<LUC4k<LUC4k<LUC4k<LUC4k<LUC5j<KWC4i<LWC4i<LWC4i<LWC4i<LWC4i<MVC7f<IZC7f<IZC5Z;2fEIPO5Z;a0WEZO_O4[;R1fDjNO4[;[1dDeN_;b13M4L3M3M3M4O0O1N2N2N0O00100O14L0O00O0O31N1O00GkC[OV<d09O10000000O100000000000000N210OO11OO1O1000O100000OeC[OR<e0mC]OR<b0oC^OQ<c0oC\\OQ<d0oCYOT<o0O4MGmCZOS<d0oC\\OQ<d0oC\\OQ<c0<00O10000001O1O0bCZOW<m01O001L301O0O2O1O2N5K3M0O0GUC4k<LVC2k<NVC0k<171N01O1O100O0100000O10O1000O1000O10O100000O010000000O01000000O0100000O10O1000O1000K5D;2O0O;EaS^4"}, "label": "plane in air"}]}
{"id": 35265, "image": "COCO_train2014_000000035265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player holding black baseball bat\"."}], "task": "refering", "answer_template": "The \"baseball player holding black baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 60, 83, 84, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 387, 388, 389], "bbox": [0.540265625, 0.00020833333333333335, 0.94128125, 0.9979791666666668], "iscrowd": 0, "area": 49047.406449999995, "rle": {"size": [480, 640], "counts": "QWR53j>4L4K5M2O2M3N2N2M2O2N2M3N2N1ZN6VDK[;c0eD^OX;d0hD]OU;e0kD[O_:hNjEo1GYO\\:lNjEm1JWOY:POjEk1MVOV:ROjEj10TOS:VOjEh13ROP:ZOjEf16POm9k1SFVNj9l1VFTNg9o1ZFPNc9S2]FmM`9V2aFiM]9Y2eFfMT9`2mF_Ml8h2VGVMn6POQJQ4SOmLf6]OkIm3@fL^6IdIi3N^LX64^Ie3HTK`0R1U6a5\\I_I;Q1S6f5cI[I5P1S6k5hIXI1n0P6Q6PJZJj5l5WJRJg5Q6YJoIc5U6^JjI_5Y6aJfI]5\\6eJcIW5a6jJ`IQ5c6oJcIi4_6XKfIa4[6aKiIY4W6jKlIo3U6TLnIf3R6^LPJX3T6kLoIk2U6XMnI^2V6fMkIR2X6QNiIg1[6\\NfI\\1^6hNbIP1b6SO`Ic0e6@\\I8h6KYIMk66^IZOf6i0g34L3M3M3M4L1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N001O1O1O1O001O000O1000000000000000000000001O00000000000aHmKQ4S4jKRLV4o3eKTL\\4l3dKUL[4k3eKVLZ4j3fKWLY4i3hKWLW4i3iKXLV4h3jKYLU4g3kKZLT4g3kK[LS4e3mK]LQ4c3PL]Lo3c3QL_LZ2\\OTLU4b1aLX2\\OTLS4d1cLV2\\OTLQ4f1eLT2\\OTLo3i1eLQ2^OULm3j1gLo1^OULk3l1iLm1^OULi3m1lLl1]OULg3m1PMl1[OULe3n1RMl17XKg2k2UMk1_4TNfKh1Z4VNmKe1S4ZNRLb1n3]NXL^1h3`N^L\\1b3cNdLX1\\3gNiLU1W3hNQMS1o2kNXMP1h2mN_Mo0c2mNhMj0Y2QOSNg0o1VO[Na0f1^OcN9_1D\\OBf0<^O@c0>B^Oa000N3OOO5LM37HL67FK96EM95DN:3EN:3DO;1EO;1EO;1E0;OE1;OE2:OD2<ND2<ND3;ME3;ME3<LD5;MB4>MA3?O_O2`0O_O1a00^O1a01]OOd01ZO0f01YO0f0U70001OlB]Ok;d0kCFT<:eCM\\<2bC0^<0aC1`<N_C3a<M]C5c<K\\C6e<JYC7g<IXC8h<HVC8m<GRC7Q=InB6T=JkB4Y=KeB4^=M`B1d=N[B1g=e0O10O01O10O01O10O0100O00100O010gBdNl<]1PCfNQ=Y1lBiNU=a1001O001N10001O001N10001O003L4M2N3M2N2M2O2N2N2N1O2K5FU\\a0"}, "label": "baseball player holding black baseball bat"}]}
{"id": 35265, "image": "COCO_train2014_000000035265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white baseball uniform holding his bat behind his shoulder\"."}], "task": "refering", "answer_template": "The \"a man in a white baseball uniform holding his bat behind his shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 60, 83, 84, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 387, 388, 389], "bbox": [0.540265625, 0.00020833333333333335, 0.94128125, 0.9979791666666668], "iscrowd": 0, "area": 49047.406449999995, "rle": {"size": [480, 640], "counts": "QWR53j>4L4K5M2O2M3N2N2M2O2N2M3N2N1ZN6VDK[;c0eD^OX;d0hD]OU;e0kD[O_:hNjEo1GYO\\:lNjEm1JWOY:POjEk1MVOV:ROjEj10TOS:VOjEh13ROP:ZOjEf16POm9k1SFVNj9l1VFTNg9o1ZFPNc9S2]FmM`9V2aFiM]9Y2eFfMT9`2mF_Ml8h2VGVMn6POQJQ4SOmLf6]OkIm3@fL^6IdIi3N^LX64^Ie3HTK`0R1U6a5\\I_I;Q1S6f5cI[I5P1S6k5hIXI1n0P6Q6PJZJj5l5WJRJg5Q6YJoIc5U6^JjI_5Y6aJfI]5\\6eJcIW5a6jJ`IQ5c6oJcIi4_6XKfIa4[6aKiIY4W6jKlIo3U6TLnIf3R6^LPJX3T6kLoIk2U6XMnI^2V6fMkIR2X6QNiIg1[6\\NfI\\1^6hNbIP1b6SO`Ic0e6@\\I8h6KYIMk66^IZOf6i0g34L3M3M3M4L1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N001O1O1O1O001O000O1000000000000000000000001O00000000000aHmKQ4S4jKRLV4o3eKTL\\4l3dKUL[4k3eKVLZ4j3fKWLY4i3hKWLW4i3iKXLV4h3jKYLU4g3kKZLT4g3kK[LS4e3mK]LQ4c3PL]Lo3c3QL_LZ2\\OTLU4b1aLX2\\OTLS4d1cLV2\\OTLQ4f1eLT2\\OTLo3i1eLQ2^OULm3j1gLo1^OULk3l1iLm1^OULi3m1lLl1]OULg3m1PMl1[OULe3n1RMl17XKg2k2UMk1_4TNfKh1Z4VNmKe1S4ZNRLb1n3]NXL^1h3`N^L\\1b3cNdLX1\\3gNiLU1W3hNQMS1o2kNXMP1h2mN_Mo0c2mNhMj0Y2QOSNg0o1VO[Na0f1^OcN9_1D\\OBf0<^O@c0>B^Oa000N3OOO5LM37HL67FK96EM95DN:3EN:3DO;1EO;1EO;1E0;OE1;OE2:OD2<ND2<ND3;ME3;ME3<LD5;MB4>MA3?O_O2`0O_O1a00^O1a01]OOd01ZO0f01YO0f0U70001OlB]Ok;d0kCFT<:eCM\\<2bC0^<0aC1`<N_C3a<M]C5c<K\\C6e<JYC7g<IXC8h<HVC8m<GRC7Q=InB6T=JkB4Y=KeB4^=M`B1d=N[B1g=e0O10O01O10O01O10O0100O00100O010gBdNl<]1PCfNQ=Y1lBiNU=a1001O001N10001O001N10001O003L4M2N3M2N2M2O2N2N2N1O2K5FU\\a0"}, "label": "a man in a white baseball uniform holding his bat behind his shoulder"}]}
{"id": 35265, "image": "COCO_train2014_000000035265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 87, 88, 89, 109, 110, 111, 112, 113, 133, 134, 135, 136, 137, 156, 157, 158, 160, 178, 179, 180, 181, 201, 202, 203, 204], "bbox": [0.7550625, 0.12135416666666667, 0.9994375, 0.5303333333333333], "iscrowd": 0, "area": 11530.675950000004, "rle": {"size": [480, 640], "counts": "TaR71m>2N2N2N2N2`DGW8;gGHV8:iGGU8;iGHU89jGIT88jGKU85jGMT84jGOU81jG0U81iG2V8NiG4U8MiG6V8JiG7V8KgG7X8KfG6Z8KcG7\\8LaG5_8L^G6a8L]G5c8LZG6e8LYG5f8NVG4j8MTG4j8OgF?T9GkF9Q9MmF3n83QGM^8e0`G\\O^8h0`GXO_8k0_GUO_8n0`GRO^8R1_GnNa8T1^GlN`8X1^GgNb8[1\\GfNd8[1[GeNe8[1[GdNf8]1YGcNf8_1XGaNi8_1UGcNk8^1SGbNn8_1oFcNP9^1nFcNS9^1fFhNZ9^310O0101N11O3M3M2O2M3M1O01O2N1O2N101N1QKkFQ4W9kKTGl3m8RLTGn3n8oKSGR4m8lKUGS4l8kKUGU4m8hKTGV2A]O\\9[NTGW2EZOY9\\NSGX2JYOS9]NUGX2NTOj9j0\\FPOf9o0`FjNc9S1W2N1N3M3N1N3N2M2A`0M30O01O1O010O1O10O01O10O01O1O10O01O100O1O100O00100O1O1O100O00100O11O0000001O000010O000001O01O01O000001N100O1O1O2N1O1O1O2N1O100N2O2M2OZ9"}, "label": "a man in an orange shirt"}]}
{"id": 35265, "image": "COCO_train2014_000000035265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurred man in orange watching the game\"."}], "task": "refering", "answer_template": "The \"a blurred man in orange watching the game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 87, 88, 89, 109, 110, 111, 112, 113, 133, 134, 135, 136, 137, 156, 157, 158, 160, 178, 179, 180, 181, 201, 202, 203, 204], "bbox": [0.7550625, 0.12135416666666667, 0.9994375, 0.5303333333333333], "iscrowd": 0, "area": 11530.675950000004, "rle": {"size": [480, 640], "counts": "TaR71m>2N2N2N2N2`DGW8;gGHV8:iGGU8;iGHU89jGIT88jGKU85jGMT84jGOU81jG0U81iG2V8NiG4U8MiG6V8JiG7V8KgG7X8KfG6Z8KcG7\\8LaG5_8L^G6a8L]G5c8LZG6e8LYG5f8NVG4j8MTG4j8OgF?T9GkF9Q9MmF3n83QGM^8e0`G\\O^8h0`GXO_8k0_GUO_8n0`GRO^8R1_GnNa8T1^GlN`8X1^GgNb8[1\\GfNd8[1[GeNe8[1[GdNf8]1YGcNf8_1XGaNi8_1UGcNk8^1SGbNn8_1oFcNP9^1nFcNS9^1fFhNZ9^310O0101N11O3M3M2O2M3M1O01O2N1O2N101N1QKkFQ4W9kKTGl3m8RLTGn3n8oKSGR4m8lKUGS4l8kKUGU4m8hKTGV2A]O\\9[NTGW2EZOY9\\NSGX2JYOS9]NUGX2NTOj9j0\\FPOf9o0`FjNc9S1W2N1N3M3N1N3N2M2A`0M30O01O1O010O1O10O01O10O01O1O10O01O100O1O100O00100O1O1O100O00100O11O0000001O000010O000001O01O01O000001N100O1O1O2N1O1O1O2N1O100N2O2M2OZ9"}, "label": "a blurred man in orange watching the game"}]}
{"id": 35265, "image": "COCO_train2014_000000035265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"batter on left\"."}], "task": "refering", "answer_template": "The \"batter on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376], "bbox": [0.1045, 0.2831458333333333, 0.542703125, 1.0], "iscrowd": 0, "area": 63382.17715000002, "rle": {"size": [480, 640], "counts": "Teo04j>5K5J6K5K5J5L5K5J6K5J6K5K5K5_FjMR6X2`IWN]6l1TIbNj6a1hHmNU7V1\\HXOa7l0PHCm7`0dGNP1TMl5[6nIhIR6[6gIiIX6Z6cIiI]6Z6\\IjId6X6WIkIi6X6QIkIn6X6lHlIT7P7000000O1000000O10001O0O100000000O1000000O1000000O100000000O2O00000O1000000O100000000O1000000O1000001N100000000O1000000O1000000O100000000O2O00000O1000000O100000000^OgG\\JZ8Y5SHeJm7U5ZHjJf7T5^HjJb7T5bHiJ_7U5dHjJ\\7T5Y1N3M2N2N2N2N2N200O1O1O1O100O1O1O1O100O1O1O2N100000000000000000000000000000000O1000000000000000000000@kKUFU4j9SLoEm3o9c0O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O0O2O1N2O1N2O1N2O1M3J6J6K5J6J6K5J6J7J7H7I8J6K4L5K4M4N2O0O2N101N2N101N2N1`KbF]3_9aLfFZ3[9eLjFW3W9gLnFT3T9iLRGR3n8nLTGQ3l8nLUGQ3l8nLVGP3j8oLXGQ3h8nLZGP3g8oLZGP3f8oL\\GQ3d8nL^GP3c8oL^GQ3a8oLaGo2`8oLbGP3_8oLcGP3\\8PMeGo2\\8PMfGn2[8PMgGo2Z8PMhGm2Y8SMhGl2Y8RMjGk2X8TMiGk2W8UMkGh2W8WMjGh2W8VMkGh2V8XMlGf2U8YMlGe2V8YMmGe2S8[MnGd2Y:M2O2M2O2M2O2M2O1N3N1N3N1N3N1N3N1N3N1N3N1N3N1N2J7D;FPRY4"}, "label": "batter on left"}]}
{"id": 35265, "image": "COCO_train2014_000000035265.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player holding up a baseball bat , he is wearing a jersey with the number 21 on it\"."}], "task": "refering", "answer_template": "The \"the baseball player holding up a baseball bat , he is wearing a jersey with the number 21 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376], "bbox": [0.1045, 0.2831458333333333, 0.542703125, 1.0], "iscrowd": 0, "area": 63382.17715000002, "rle": {"size": [480, 640], "counts": "Teo04j>5K5J6K5K5J5L5K5J6K5J6K5K5K5_FjMR6X2`IWN]6l1TIbNj6a1hHmNU7V1\\HXOa7l0PHCm7`0dGNP1TMl5[6nIhIR6[6gIiIX6Z6cIiI]6Z6\\IjId6X6WIkIi6X6QIkIn6X6lHlIT7P7000000O1000000O10001O0O100000000O1000000O1000000O100000000O2O00000O1000000O100000000O1000000O1000001N100000000O1000000O1000000O100000000O2O00000O1000000O100000000^OgG\\JZ8Y5SHeJm7U5ZHjJf7T5^HjJb7T5bHiJ_7U5dHjJ\\7T5Y1N3M2N2N2N2N2N200O1O1O1O100O1O1O1O100O1O1O2N100000000000000000000000000000000O1000000000000000000000@kKUFU4j9SLoEm3o9c0O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O0O2O1N2O1N2O1N2O1M3J6J6K5J6J6K5J6J7J7H7I8J6K4L5K4M4N2O0O2N101N2N101N2N1`KbF]3_9aLfFZ3[9eLjFW3W9gLnFT3T9iLRGR3n8nLTGQ3l8nLUGQ3l8nLVGP3j8oLXGQ3h8nLZGP3g8oLZGP3f8oL\\GQ3d8nL^GP3c8oL^GQ3a8oLaGo2`8oLbGP3_8oLcGP3\\8PMeGo2\\8PMfGn2[8PMgGo2Z8PMhGm2Y8SMhGl2Y8RMjGk2X8TMiGk2W8UMkGh2W8WMjGh2W8VMkGh2V8XMlGf2U8YMlGe2V8YMmGe2S8[MnGd2Y:M2O2M2O2M2O2M2O1N3N1N3N1N3N1N3N1N3N1N3N1N3N1N2J7D;FPRY4"}, "label": "the baseball player holding up a baseball bat , he is wearing a jersey with the number 21 on it"}]}
{"id": 231878, "image": "COCO_train2014_000000231878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella that the woman in gray is holding that is facing the ground\"."}], "task": "refering", "answer_template": "The \"the umbrella that the woman in gray is holding that is facing the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 271, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 363, 364], "bbox": [0.702203125, 0.59904296875, 0.9475156250000001, 0.8605078125], "iscrowd": 0, "area": 9784.188950000002, "rle": {"size": [512, 640], "counts": "lkP71P`01O1N2O4J8H7J3L1O2O0O1O2N101N1O2O0O2N101N1O2O0O2N1O2O0O2N101N1O2O0O2N101N1`NTNkDn1R;WNjDj1U;WNkDi1S;SNhCNT1Q2S;QNoDNgNR2Z<oMQET2n:lMSES2n:mMQET2n:lMRET2n:mMQET2o:kMPEV2Q;kMmDU2T;lMjDT2W;nMeDS2\\;nMbDR2_;oM_DQ2b;QNZDP2g;QNWDo1j;oMWDR2k<0O10000JdMgB]2Z=41O100O1O1O00000000O100000O01000000O1000000O1000000O01001O2M2O2N1O2MFoBhMP=X2VCdMh<[2?0O0100O1O010O100OO2N1103L100O1O2O0O100O101N100O1O101N100O100O2O0O1O100O2O0O100O2N100O100O2O0O1M3M3M4L3M3M3M3K6I6J5KYe`0"}, "label": "the umbrella that the woman in gray is holding that is facing the ground"}]}
{"id": 231878, "image": "COCO_train2014_000000231878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella with a red handle being held with a girl in high - top sneakers\"."}], "task": "refering", "answer_template": "The \"umbrella with a red handle being held with a girl in high - top sneakers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 271, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 363, 364], "bbox": [0.702203125, 0.59904296875, 0.9475156250000001, 0.8605078125], "iscrowd": 0, "area": 9784.188950000002, "rle": {"size": [512, 640], "counts": "lkP71P`01O1N2O4J8H7J3L1O2O0O1O2N101N1O2O0O2N101N1O2O0O2N1O2O0O2N101N1O2O0O2N101N1`NTNkDn1R;WNjDj1U;WNkDi1S;SNhCNT1Q2S;QNoDNgNR2Z<oMQET2n:lMSES2n:mMQET2n:lMRET2n:mMQET2o:kMPEV2Q;kMmDU2T;lMjDT2W;nMeDS2\\;nMbDR2_;oM_DQ2b;QNZDP2g;QNWDo1j;oMWDR2k<0O10000JdMgB]2Z=41O100O1O1O00000000O100000O01000000O1000000O1000000O01001O2M2O2N1O2MFoBhMP=X2VCdMh<[2?0O0100O1O010O100OO2N1103L100O1O2O0O100O101N100O1O101N100O100O2O0O1O100O2O0O100O2N100O100O2O0O1M3M3M4L3M3M3M3K6I6J5KYe`0"}, "label": "umbrella with a red handle being held with a girl in high - top sneakers"}]}
{"id": 231878, "image": "COCO_train2014_000000231878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a transparent umbrella being held up\"."}], "task": "refering", "answer_template": "The \"a transparent umbrella being held up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 200], "bbox": [0.58096875, 0.31513671875, 0.8102031249999999, 0.53712890625], "iscrowd": 0, "area": 4127.402150000002, "rle": {"size": [512, 640], "counts": "iVj51n?1O1O2N1N2O1O1O1O1O1N2O2N1O1O1O1N2O1O1O1O2O0O1O100O100OG:M2010O100O1O100O010O100O100O10O1000O102N3M2N3L4M2N2N0O100000O0100000000@mNRBT1n=?5K:F00000:F?A?AfNZCB^<<d1O1N2N2000000000000000000000001O01O010O0010O01O01O0O101O00001O0000001O0O101O00001O00001O1N2O001O1O1O0O2O1O0O2O1O1N101O1N2O001N2O001O1N2O001Ndil1"}, "label": "a transparent umbrella being held up"}]}
{"id": 231878, "image": "COCO_train2014_000000231878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear umbrella held over two women\"."}], "task": "refering", "answer_template": "The \"clear umbrella held over two women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 200], "bbox": [0.58096875, 0.31513671875, 0.8102031249999999, 0.53712890625], "iscrowd": 0, "area": 4127.402150000002, "rle": {"size": [512, 640], "counts": "iVj51n?1O1O2N1N2O1O1O1O1O1N2O2N1O1O1O1N2O1O1O1O2O0O1O100O100OG:M2010O100O1O100O010O100O100O10O1000O102N3M2N3L4M2N2N0O100000O0100000000@mNRBT1n=?5K:F00000:F?A?AfNZCB^<<d1O1N2N2000000000000000000000001O01O010O0010O01O01O0O101O00001O0000001O0O101O00001O00001O1N2O001O1O1O0O2O1O0O2O1O1N101O1N2O001N2O001O1N2O001Ndil1"}, "label": "clear umbrella held over two women"}]}
{"id": 559566, "image": "COCO_train2014_000000559566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that the child is riding on\"."}], "task": "refering", "answer_template": "The \"the horse that the child is riding on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 125, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 150, 151, 152, 153, 154, 155, 156, 157, 158, 164, 165, 166, 168, 169, 170, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 269, 270, 271, 272, 273, 285, 286, 287, 292, 293, 294, 295, 308, 309, 310, 315, 316, 317, 318, 331, 332, 333, 339, 340, 341, 354, 355, 361, 362, 363, 364, 384, 385, 386], "bbox": [0.15903125, 0.19987124463519312, 0.91678125, 0.9686051502145923], "iscrowd": 0, "area": 68942.01199999999, "rle": {"size": [466, 640], "counts": "`c^13]>5K5L4K5K5L4K3M2O2M2N2O2M2N2O2M2N2O2M2N3N1N2N3N1N2N3N1N2N3N0O2N2N101N1O2O1N1O2O0O2N2QO\\MoEd2T:hM\\EZ2e:h02N2M3N2N2N2N2N1O2M3N2N2N2N2N2N2M3N2N2N2N1O001N2O001O001O1O0O2O1O0011O0O2O100O010O1O10O01O100O00100O10O01O100O00100O2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O2M2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1\\JWK[1l4`NXK_1i4^NZK`1h4]N[Ka1g4[N\\Ke1e4XN^Kf1h4RN\\Kl1j4mMYKR2l4gMVKX2P5`MQKa2U5XMlJi2Z5oLfJR3`5fLaJ\\3d5]L\\Jd3j5TLXJl3n5mKRJU4S6dKmI]4<dJ_4g0VKf4;PKX43]Ko4:PK_4HXKX58SKh5n4XJTKe5n4ZJTKc5m4]JUKa5k4_JWK^5k4aJWK\\5j4cJYKZ5i4eJYKX5h4hJZKV5f4jJ\\KCkNe4k5gK\\KAUO\\4`5RL^K@RO]4a5SL_K]OQO`4a5RL_K]OPO`4b5SL_K[OPOa4b5TL_KYOPOb4c5TL^KYOPOa4c5UL_KXOoNc4b5UL`KWOnNc4d5UL_KVOnNb2_OWOU6QO_KUOmN^20oNf5]O^KTOmNZ2a0iNT5I_KSOlNV2S1`Nc46`KROlNR2b1YNT4b0ULP1d6PO\\Ij0l6TOTIg0S7WOmHe0Y7XOhHQ1Q7mNoH\\1j6bNWIe1c6XN^IQ2[6gMkIb2VO]Ma7k2oG\\MR8m2]GZMd8o2kFXMV9Q42N1O1O00O100000000000000000O01J6K5J6JUMSG>h8HWHXOc7o0]IPN^6V2n21O1O1O1O1O2O0O1O1O1O1O1O1O1O1O2O0O10000O10000000001N10000000000O1000001O000O10000000000O101OO100O1O010O100O1O10O01O100O1O10O0100O1O100O00100O1O01O0010O0001O01O01O00010O001O01O01O01O01O00010O001O01O01O00010O001O01O01O000100O12N4L3N3L4L4L4L4L4L4L4L4L4M2dJmJS1V5bNRK^1P5YNWKg1l4RNWKo1l4mMUKS2n4iMSKW2o4fMQK[2R5aMoJ_2T5]MmJc2V5ZMjJg2W5VMiJk2Z5QMgJo2\\5mLeJS3]5jLcJW3`5eLaJ[3b5aL_J_3d5]L\\Jd3f5ZLZJf3i5VLXJj3k5RLVJn3m5hKYJY4i5YKcJg4`5jJlJV5W710000O1000000N2N2N2N2N2N2N2N2[LRJdNP6^1QJ[NS6f1PJoMW6S2jIcM]6^2eIXMb6j2`IkLg6V3[I`Ll6b3UIULQ7l3RIiKU7Y4nH[KY7f4nHkJY7V5Q12N1O2`J^Jl1c5QN_Jo1c5nM_JQ2b5mM_JS2c5iM_JW2b5gM`JX2b5eM_J[2b5bMaJ]2a5`M`J`2e5ZM]Je2m5oLUJQ3T6eLnIZ3\\6[LeIe3d6mKaIS4i6]K\\Id4m6lJYIU5R80O1O100O1O1fJiJY1W5bNSKX1n4cN[KX1f4cNdKW1]4dNlKX1T4cNVLW1k3dN_LV1c3dNfLX1[3bNoLX1S3lNlLo0U3YOhLb0Y3EhL0[37gL]O\\3k0eLiN^3^1dLVN^3R2eLaM^3g2cLmL`3Z3bL[L`3g3dLPL^3Q4gLgK\\3Y4iL_KZ3a4\\30O101O000O1000000O1000000O100000N2L3N3M3L4M2M4M3L4M2N3L4M3L3N3L4C=_O`0@a0_Oa0_Oa0_OYng0"}, "label": "the horse that the child is riding on"}]}
{"id": 559566, "image": "COCO_train2014_000000559566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse tethered to a fence\"."}], "task": "refering", "answer_template": "The \"a white horse tethered to a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 125, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 150, 151, 152, 153, 154, 155, 156, 157, 158, 164, 165, 166, 168, 169, 170, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 269, 270, 271, 272, 273, 285, 286, 287, 292, 293, 294, 295, 308, 309, 310, 315, 316, 317, 318, 331, 332, 333, 339, 340, 341, 354, 355, 361, 362, 363, 364, 384, 385, 386], "bbox": [0.15903125, 0.19987124463519312, 0.91678125, 0.9686051502145923], "iscrowd": 0, "area": 68942.01199999999, "rle": {"size": [466, 640], "counts": "`c^13]>5K5L4K5K5L4K3M2O2M2N2O2M2N2O2M2N2O2M2N3N1N2N3N1N2N3N1N2N3N0O2N2N101N1O2O1N1O2O0O2N2QO\\MoEd2T:hM\\EZ2e:h02N2M3N2N2N2N2N1O2M3N2N2N2N2N2N2M3N2N2N2N1O001N2O001O001O1O0O2O1O0011O0O2O100O010O1O10O01O100O00100O10O01O100O00100O2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O2M2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1\\JWK[1l4`NXK_1i4^NZK`1h4]N[Ka1g4[N\\Ke1e4XN^Kf1h4RN\\Kl1j4mMYKR2l4gMVKX2P5`MQKa2U5XMlJi2Z5oLfJR3`5fLaJ\\3d5]L\\Jd3j5TLXJl3n5mKRJU4S6dKmI]4<dJ_4g0VKf4;PKX43]Ko4:PK_4HXKX58SKh5n4XJTKe5n4ZJTKc5m4]JUKa5k4_JWK^5k4aJWK\\5j4cJYKZ5i4eJYKX5h4hJZKV5f4jJ\\KCkNe4k5gK\\KAUO\\4`5RL^K@RO]4a5SL_K]OQO`4a5RL_K]OPO`4b5SL_K[OPOa4b5TL_KYOPOb4c5TL^KYOPOa4c5UL_KXOoNc4b5UL`KWOnNc4d5UL_KVOnNb2_OWOU6QO_KUOmN^20oNf5]O^KTOmNZ2a0iNT5I_KSOlNV2S1`Nc46`KROlNR2b1YNT4b0ULP1d6PO\\Ij0l6TOTIg0S7WOmHe0Y7XOhHQ1Q7mNoH\\1j6bNWIe1c6XN^IQ2[6gMkIb2VO]Ma7k2oG\\MR8m2]GZMd8o2kFXMV9Q42N1O1O00O100000000000000000O01J6K5J6JUMSG>h8HWHXOc7o0]IPN^6V2n21O1O1O1O1O2O0O1O1O1O1O1O1O1O1O2O0O10000O10000000001N10000000000O1000001O000O10000000000O101OO100O1O010O100O1O10O01O100O1O10O0100O1O100O00100O1O01O0010O0001O01O01O00010O001O01O01O01O01O00010O001O01O01O00010O001O01O01O000100O12N4L3N3L4L4L4L4L4L4L4L4L4M2dJmJS1V5bNRK^1P5YNWKg1l4RNWKo1l4mMUKS2n4iMSKW2o4fMQK[2R5aMoJ_2T5]MmJc2V5ZMjJg2W5VMiJk2Z5QMgJo2\\5mLeJS3]5jLcJW3`5eLaJ[3b5aL_J_3d5]L\\Jd3f5ZLZJf3i5VLXJj3k5RLVJn3m5hKYJY4i5YKcJg4`5jJlJV5W710000O1000000N2N2N2N2N2N2N2N2[LRJdNP6^1QJ[NS6f1PJoMW6S2jIcM]6^2eIXMb6j2`IkLg6V3[I`Ll6b3UIULQ7l3RIiKU7Y4nH[KY7f4nHkJY7V5Q12N1O2`J^Jl1c5QN_Jo1c5nM_JQ2b5mM_JS2c5iM_JW2b5gM`JX2b5eM_J[2b5bMaJ]2a5`M`J`2e5ZM]Je2m5oLUJQ3T6eLnIZ3\\6[LeIe3d6mKaIS4i6]K\\Id4m6lJYIU5R80O1O100O1O1fJiJY1W5bNSKX1n4cN[KX1f4cNdKW1]4dNlKX1T4cNVLW1k3dN_LV1c3dNfLX1[3bNoLX1S3lNlLo0U3YOhLb0Y3EhL0[37gL]O\\3k0eLiN^3^1dLVN^3R2eLaM^3g2cLmL`3Z3bL[L`3g3dLPL^3Q4gLgK\\3Y4iL_KZ3a4\\30O101O000O1000000O1000000O100000N2L3N3M3L4M2M4M3L4M2N3L4M3L3N3L4C=_O`0@a0_Oa0_Oa0_OYng0"}, "label": "a white horse tethered to a fence"}]}
{"id": 559566, "image": "COCO_train2014_000000559566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girl with the furry boots is sitting on a horse\"."}], "task": "refering", "answer_template": "The \"the little girl with the furry boots is sitting on a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 81, 82, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 194, 195, 216, 217, 218], "bbox": [0.421203125, 0.08092274678111588, 0.629734375, 0.5605364806866953], "iscrowd": 0, "area": 11856.02895, "rle": {"size": [466, 640], "counts": "`Sk32`>2N1O2N2M2O1O00001O0O100000001O000O1000001O0QMM[F2WO1^:L\\Fe0cN@Q;K\\FX1e9gNZFZ1f9fNZF[1e9dNbEV2^:jMbEV2^:jMaEW2_:iM`EX2`:gMaEY2_:gM`E[2_:eM`E\\2`:dM`E\\2_:dMaE]2]:eMcE[2\\:fMcE[2\\:fMcE\\2[:dMfE\\2Y:eMfE\\2Y:eMgE[2X:eMhE\\2W:eMhEZ2X:gMiEV2Y:kMfET2[:lMfEP2]:PNcEm1`:TN_Eh1e:XN\\Ed1g:\\NYEa1j:`NVE[1R9VNTH`0jNW1S9^NPH<lNS1U9fNlG7POP1T9nNiG2SOn0T9VOaG1ZOf0V9^OZGOA`0U9EWGLD=U9IWGKD9W9MSGKF5[9OnFMG2^91jFNHN`94gFOIMa94cF2LJ`94`F60G`93^F72G`91^F92F_94\\F65F`98VF`0MXOl9<SFc1m9a100O10000O10O010000001O001O001O0O2O00001O001O001O0O3N2N2N2N2N2N2N2N3M5J7J5K5K4L2M3M3L4M3M3M4K4M3M3M3L4M4L6J6I7J5K6J6I7J6J_f[3"}, "label": "the little girl with the furry boots is sitting on a horse"}]}
{"id": 559566, "image": "COCO_train2014_000000559566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid on a white horse\"."}], "task": "refering", "answer_template": "The \"a kid on a white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 81, 82, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 194, 195, 216, 217, 218], "bbox": [0.421203125, 0.08092274678111588, 0.629734375, 0.5605364806866953], "iscrowd": 0, "area": 11856.02895, "rle": {"size": [466, 640], "counts": "`Sk32`>2N1O2N2M2O1O00001O0O100000001O000O1000001O0QMM[F2WO1^:L\\Fe0cN@Q;K\\FX1e9gNZFZ1f9fNZF[1e9dNbEV2^:jMbEV2^:jMaEW2_:iM`EX2`:gMaEY2_:gM`E[2_:eM`E\\2`:dM`E\\2_:dMaE]2]:eMcE[2\\:fMcE[2\\:fMcE\\2[:dMfE\\2Y:eMfE\\2Y:eMgE[2X:eMhE\\2W:eMhEZ2X:gMiEV2Y:kMfET2[:lMfEP2]:PNcEm1`:TN_Eh1e:XN\\Ed1g:\\NYEa1j:`NVE[1R9VNTH`0jNW1S9^NPH<lNS1U9fNlG7POP1T9nNiG2SOn0T9VOaG1ZOf0V9^OZGOA`0U9EWGLD=U9IWGKD9W9MSGKF5[9OnFMG2^91jFNHN`94gFOIMa94cF2LJ`94`F60G`93^F72G`91^F92F_94\\F65F`98VF`0MXOl9<SFc1m9a100O10000O10O010000001O001O001O0O2O00001O001O001O0O3N2N2N2N2N2N2N2N3M5J7J5K5K4L2M3M3L4M3M3M4K4M3M3M3L4M4L6J6I7J5K6J6I7J6J_f[3"}, "label": "a kid on a white horse"}]}
{"id": 510418, "image": "COCO_train2014_000000510418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the last two fighter planes in the back of the group\"."}], "task": "refering", "answer_template": "The \"the last two fighter planes in the back of the group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 166, 167, 181, 182, 183, 184, 198, 199, 200, 214, 215, 216, 230, 231, 232, 245, 246, 247, 248, 262, 263, 264], "bbox": [0.3636547085201794, 0.380046875, 0.5615695067264573, 0.72625], "iscrowd": 0, "area": 8091.697199999999, "rle": {"size": [640, 446], "counts": "QbU32m20V>6]A0b>3`AJ[>7dAMX>4hA0S>1lA3P>MPB5o=KQB5n=LQB4o=MPB4o=MPB4o=MPB4o=MPB4P>LoA4Q>MnA4Q>MnA4Q>MnA4Q>MnA3R=WOhAf0U14R=ZOeAb0Y15Q=^OaA>2]O=h0P>A\\A;6C7b0V>A\\A;6B8b0V>B[A;6B8b0IUOb==WB:5C8b0IZO]=8\\B:5C8b0H[O^=7\\B:5C8b0H\\O]=6]B:6B7b0I\\O]=7\\B:6B7b0I\\O]=7\\B:6B7b0I\\O]=7]B95C7b0I\\O^=6\\B:5C7b0I\\O[=9]B88A7b0I\\OW=>]B7;^OLm04QOT=k1kB40RNQ=n1eB:9iMn<o4RCQKj<T5UCmJm<R64L5J5L4L5K4M3L5L3L4lL`A`0d>\\O_Aa0f>ZO^Ac0e>YO_Ac0b>bNXA]O9n1`>^N^AC6i0ABk>2_AB8g0F]Od>9^AC<c0C@d>9^AD<a0B@f>;]AC<a0B_Og><[AD<a0B^Oi><ZAD<a0B^Oh>=ZAD=`0A_Oi><ZAD<a0B]Oi>>ZAC<a0A^Oj>=YAE;`0C]Oj>=YAE;`0B^Oj>=ZAD;`0B^Oj>=YAE;?C_Oj><YAE;?C_Oi>=ZAD:`0C_Oj><YAE;?C_Oj><YAD<`0A_Ok>=eB4aN^Ok>=eB4aN^Oj>>fB3cN\\Oh>`0eB4Y>KhA4X>LhA4Y>KhA4Y>KhA4X>LhA4Y>KhA4X>LhA3Z>LgA3Z>LgA2Z>JjA6Sa0L4M3LkRj3"}, "label": "the last two fighter planes in the back of the group"}]}
{"id": 510418, "image": "COCO_train2014_000000510418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the last two planes going up in the air\"."}], "task": "refering", "answer_template": "The \"the last two planes going up in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 166, 167, 181, 182, 183, 184, 198, 199, 200, 214, 215, 216, 230, 231, 232, 245, 246, 247, 248, 262, 263, 264], "bbox": [0.3636547085201794, 0.380046875, 0.5615695067264573, 0.72625], "iscrowd": 0, "area": 8091.697199999999, "rle": {"size": [640, 446], "counts": "QbU32m20V>6]A0b>3`AJ[>7dAMX>4hA0S>1lA3P>MPB5o=KQB5n=LQB4o=MPB4o=MPB4o=MPB4o=MPB4P>LoA4Q>MnA4Q>MnA4Q>MnA4Q>MnA3R=WOhAf0U14R=ZOeAb0Y15Q=^OaA>2]O=h0P>A\\A;6C7b0V>A\\A;6B8b0V>B[A;6B8b0IUOb==WB:5C8b0IZO]=8\\B:5C8b0H[O^=7\\B:5C8b0H\\O]=6]B:6B7b0I\\O]=7\\B:6B7b0I\\O]=7\\B:6B7b0I\\O]=7]B95C7b0I\\O^=6\\B:5C7b0I\\O[=9]B88A7b0I\\OW=>]B7;^OLm04QOT=k1kB40RNQ=n1eB:9iMn<o4RCQKj<T5UCmJm<R64L5J5L4L5K4M3L5L3L4lL`A`0d>\\O_Aa0f>ZO^Ac0e>YO_Ac0b>bNXA]O9n1`>^N^AC6i0ABk>2_AB8g0F]Od>9^AC<c0C@d>9^AD<a0B@f>;]AC<a0B_Og><[AD<a0B^Oi><ZAD<a0B^Oh>=ZAD=`0A_Oi><ZAD<a0B]Oi>>ZAC<a0A^Oj>=YAE;`0C]Oj>=YAE;`0B^Oj>=ZAD;`0B^Oj>=YAE;?C_Oj><YAE;?C_Oi>=ZAD:`0C_Oj><YAE;?C_Oj><YAD<`0A_Ok>=eB4aN^Ok>=eB4aN^Oj>>fB3cN\\Oh>`0eB4Y>KhA4X>LhA4Y>KhA4Y>KhA4X>LhA4Y>KhA4X>LhA3Z>LgA3Z>LgA2Z>JjA6Sa0L4M3LkRj3"}, "label": "the last two planes going up in the air"}]}
{"id": 10728, "image": "COCO_train2014_000000010728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle containing oil and garlic cloves\"."}], "task": "refering", "answer_template": "The \"a bottle containing oil and garlic cloves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 182, 183], "bbox": [0.822640625, 0.0011250000000000001, 1.0, 0.43806249999999997], "iscrowd": 0, "area": 17605.79955, "rle": {"size": [480, 640], "counts": "XUg77e>6J7I6J6J7J5J7I6K6K4L3M2N3N2M2N3M3M3N1N3M3M2N3N2M2N3M3M2O2M3M2N3M3N2M2N3M3M2O2M3M2N3M3N1N3M3M3M2O2M3M2N3M3N1N3M3M2N3N2M3M2N3M3N1O2O1O001O1O001O1O00000000000000000000000000000000O010000000000000O1O1O1O1O100O1O1O1O100O1O1O1O10O"}, "label": "a bottle containing oil and garlic cloves"}]}
{"id": 10728, "image": "COCO_train2014_000000010728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"jar of pickles sitting on a table\"."}], "task": "refering", "answer_template": "The \"jar of pickles sitting on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 182, 183], "bbox": [0.822640625, 0.0011250000000000001, 1.0, 0.43806249999999997], "iscrowd": 0, "area": 17605.79955, "rle": {"size": [480, 640], "counts": "XUg77e>6J7I6J6J7J5J7I6K6K4L3M2N3N2M2N3M3M3N1N3M3M2N3N2M2N3M3M2O2M3M2N3M3N2M2N3M3M2O2M3M2N3M3N1N3M3M3M2O2M3M2N3M3N1N3M3M2N3N2M3M2N3M3N1O2O1O001O1O001O1O00000000000000000000000000000000O010000000000000O1O1O1O1O100O1O1O1O100O1O1O1O10O"}, "label": "jar of pickles sitting on a table"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a patch of dirt and red and violet flowers inside of a toliet\"."}], "task": "refering", "answer_template": "The \"a patch of dirt and red and violet flowers inside of a toliet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 130, 131, 132, 133, 147, 148, 149, 150, 166], "bbox": [0.19296000000000002, 0.47672000000000003, 0.40604, 0.7362400000000001], "iscrowd": 0, "area": 5869.744650000001, "rle": {"size": [375, 500], "counts": "bhS11d;1N3G0hD28Kc:>ZEEe:f010O1O100O2N101N101@nNXFS1g9ROTFP1i9ROVFP1h9POXFQ1f9QOYFQ1e9oN[FR1d9oN[FS1b9nN^FS1a9nN]FU1a9lN^FU1\\9QOcFQ1Z9h0N1O2N1O2N2O1N3NO100O1O1O2N1O1N3M2O1N2N2O1O1N2O1O0010000O100O1LZFWNf9i1ZFWNf9h1\\FWNd9i1\\FVNd9k1\\FUNd9k1\\FUNd9j1]FVNc9j1]FVNc9j1]FVNc9j1]FVNd9i1\\FWNd9i1\\FXNd9g1]FXNc9h1]FXNd9g1\\FYNe9f17O1O1O1O0000000000O1AbNeF^1[9hN_FX1a9?000000001O1O1O1O1N3L3XN\\FY1T:M3L4L4L4M3M3M4M9F_a\\3"}, "label": "a patch of dirt and red and violet flowers inside of a toliet"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plants in toilet\"."}], "task": "refering", "answer_template": "The \"plants in toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 130, 131, 132, 133, 147, 148, 149, 150, 166], "bbox": [0.19296000000000002, 0.47672000000000003, 0.40604, 0.7362400000000001], "iscrowd": 0, "area": 5869.744650000001, "rle": {"size": [375, 500], "counts": "bhS11d;1N3G0hD28Kc:>ZEEe:f010O1O100O2N101N101@nNXFS1g9ROTFP1i9ROVFP1h9POXFQ1f9QOYFQ1e9oN[FR1d9oN[FS1b9nN^FS1a9nN]FU1a9lN^FU1\\9QOcFQ1Z9h0N1O2N1O2N2O1N3NO100O1O1O2N1O1N3M2O1N2N2O1O1N2O1O0010000O100O1LZFWNf9i1ZFWNf9h1\\FWNd9i1\\FVNd9k1\\FUNd9k1\\FUNd9j1]FVNc9j1]FVNc9j1]FVNc9j1]FVNd9i1\\FWNd9i1\\FXNd9g1]FXNc9h1]FXNd9g1\\FYNe9f17O1O1O1O0000000000O1AbNeF^1[9hN_FX1a9?000000001O1O1O1O1N3L3XN\\FY1T:M3L4L4L4M3M3M4M9F_a\\3"}, "label": "plants in toilet"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow flowers in a terra cotta pot\"."}], "task": "refering", "answer_template": "The \"yellow flowers in a terra cotta pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 231], "bbox": [0.70628, 0.5538133333333334, 0.9685999999999999, 0.9372266666666668], "iscrowd": 0, "area": 14089.49245, "rle": {"size": [375, 500], "counts": "g`Q45`;>B=C>B9G7I8H8H7I7I3M2N2N3N1O1O2O0O1O2O0O2N100O2N1O1O2N1N2O2M2O1N3N1O101O000010O0001O01O01O000010O0001O00010O00010O010O01O01O01O010O0010O000000O1O00100O001O010O1O00100O001O1O010O1O001O1O001O001O1O001O1O2N4L4M2M4L3M4L4L3M2N3M2N3L3N3M2N3M3M2N3L3N3M2N3M2N3M7I<C=Deb5"}, "label": "yellow flowers in a terra cotta pot"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown flower pot with 3 yellow flowers in it\"."}], "task": "refering", "answer_template": "The \"a brown flower pot with 3 yellow flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 231], "bbox": [0.70628, 0.5538133333333334, 0.9685999999999999, 0.9372266666666668], "iscrowd": 0, "area": 14089.49245, "rle": {"size": [375, 500], "counts": "g`Q45`;>B=C>B9G7I8H8H7I7I3M2N2N3N1O1O2O0O1O2O0O2N100O2N1O1O2N1N2O2M2O1N3N1O101O000010O0001O01O01O000010O0001O00010O00010O010O01O01O01O010O0010O000000O1O00100O001O010O1O00100O001O1O010O1O001O1O001O001O1O001O1O2N4L4M2M4L3M4L4L3M2N3M2N3L3N3M2N3M3M2N3L3N3M2N3M2N3M7I<C=Deb5"}, "label": "a brown flower pot with 3 yellow flowers in it"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flowers planted in a toilet that is used outside as a planter\"."}], "task": "refering", "answer_template": "The \"flowers planted in a toilet that is used outside as a planter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 202, 203, 204, 205, 222], "bbox": [0.14221999999999999, 0.478, 0.46388, 0.9339733333333333], "iscrowd": 0, "area": 18729.69265, "rle": {"size": [375, 500], "counts": "dWj02c;9G9H7H9G3M4L3N3L3M4L4L3N3L3M4M2N3M3M2O2M2N3N0O2N1O2O0O2N2O0O2N1O2O0O2N2O0O2N1O2O0O2O1N101N101O0O2O1^HWLS7j3eH_LZ7o3O101O0O101N100O2O001N100O2O0O101O0O2O0O101N10001N100O2O0O2O01O01O00010O0010O00010O00010O0001O010O00010O00010O00O2N100O1O2N1O1O2N1O101N1O1O2N1O1N2N3iNeHoM\\7P2fHmM\\7R2fHkM]7S2eHjM]7U2eHiM\\7V2fHgM\\7X2fHeM]7Y2eHdM]7[2eHbM]7]2eH`M]7_2eH^M^7`2dH^M]7a2eH\\M]7c2m0N1O1N2O2N1O1O1O2N1O1O1O1M4L3M3L4M9G<D<C=D^mQ3"}, "label": "flowers planted in a toilet that is used outside as a planter"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl and bottom part of toilet\"."}], "task": "refering", "answer_template": "The \"bowl and bottom part of toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 202, 203, 204, 205, 222], "bbox": [0.14221999999999999, 0.478, 0.46388, 0.9339733333333333], "iscrowd": 0, "area": 18729.69265, "rle": {"size": [375, 500], "counts": "dWj02c;9G9H7H9G3M4L3N3L3M4L4L3N3L3M4M2N3M3M2O2M2N3N0O2N1O2O0O2N2O0O2N1O2O0O2N2O0O2N1O2O0O2O1N101N101O0O2O1^HWLS7j3eH_LZ7o3O101O0O101N100O2O001N100O2O0O101O0O2O0O101N10001N100O2O0O2O01O01O00010O0010O00010O00010O0001O010O00010O00010O00O2N100O1O2N1O1O2N1O101N1O1O2N1O1N2N3iNeHoM\\7P2fHmM\\7R2fHkM]7S2eHjM]7U2eHiM\\7V2fHgM\\7X2fHeM]7Y2eHdM]7[2eHbM]7]2eH`M]7_2eH^M^7`2dH^M]7a2eH\\M]7c2m0N1O1N2O2N1O1O1O2N1O1O1O1M4L3M3L4M9G<D<C=D^mQ3"}, "label": "bowl and bottom part of toilet"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet seat and the toilet tank\"."}], "task": "refering", "answer_template": "The \"the toilet seat and the toilet tank\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 27, 28, 40, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117], "bbox": [0.2309, 0.09213333333333333, 0.6117999999999999, 0.5438133333333334], "iscrowd": 0, "area": 22425.150100000003, "rle": {"size": [375, 500], "counts": "[aZ11f;1N101N101O00010O001\\E5Z9K^Fb0^9]OYFQ1b9POTF^1g9>L3_Ob0K5J6J6K1N3M2N2O2M2N2O2M2N3N1N2N3N1N2N3M2O1N2N2O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O000100O1O100O1O100O1O100O1O100O1O10L3M3N3O0000]KZIU4f6jK\\IV4c6jK^IT4c6lK_IP4c6PL^Im3d6SL]Ij3e6VL]If3f6YL[Id3g6ZL]Ib3e6\\L^Ia3d6\\L`Ia3b6]LbI`3[7N2N2N2N2N200001O0000000001O00000001O00000000000000000^OjLcHW3W7nLjHQ3V7oLjHQ3V7oLjHQ3V7PMiHP3W7PMiHP3X7oLhHQ3X7oLhHQ3X7oLhHQ3X7oLhHP3Y7PMgHo2Z7QMfHm2]7RMcHm2^7SMaHm2`7SM_Hl2c7TM]Hk2d7UM[Hk2f7VMXHi2j7WMUHi2m7VMSHi2n7WMQHh2Q8XMnGh2S8XMlGh2U8XMkGf2W8ZMhGf2Z8YMeGg2\\8YMcGf2_8ZMaGe2`8[M_Ga2f8O9G9G:F9G9H8G9G9G:FooV2"}, "label": "the toilet seat and the toilet tank"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet is being used as a planter for flowers\"."}], "task": "refering", "answer_template": "The \"the toilet is being used as a planter for flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 27, 28, 40, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117], "bbox": [0.2309, 0.09213333333333333, 0.6117999999999999, 0.5438133333333334], "iscrowd": 0, "area": 22425.150100000003, "rle": {"size": [375, 500], "counts": "[aZ11f;1N101N101O00010O001\\E5Z9K^Fb0^9]OYFQ1b9POTF^1g9>L3_Ob0K5J6J6K1N3M2N2O2M2N2O2M2N3N1N2N3N1N2N3M2O1N2N2O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O000100O1O100O1O100O1O100O1O100O1O10L3M3N3O0000]KZIU4f6jK\\IV4c6jK^IT4c6lK_IP4c6PL^Im3d6SL]Ij3e6VL]If3f6YL[Id3g6ZL]Ib3e6\\L^Ia3d6\\L`Ia3b6]LbI`3[7N2N2N2N2N200001O0000000001O00000001O00000000000000000^OjLcHW3W7nLjHQ3V7oLjHQ3V7oLjHQ3V7PMiHP3W7PMiHP3X7oLhHQ3X7oLhHQ3X7oLhHQ3X7oLhHP3Y7PMgHo2Z7QMfHm2]7RMcHm2^7SMaHm2`7SM_Hl2c7TM]Hk2d7UM[Hk2f7VMXHi2j7WMUHi2m7VMSHi2n7WMQHh2Q8XMnGh2S8XMlGh2U8XMkGf2W8ZMhGf2Z8YMeGg2\\8YMcGf2_8ZMaGe2`8[M_Ga2f8O9G9G:F9G9H8G9G9G:FooV2"}, "label": "the toilet is being used as a planter for flowers"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pot holding a green plant with no flowers behind and above a pot with yellow flowers\"."}], "task": "refering", "answer_template": "The \"a pot holding a green plant with no flowers behind and above a pot with yellow flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 121, 122, 123, 124], "bbox": [0.72978, 0.17010666666666666, 0.9691, 0.5094133333333334], "iscrowd": 0, "area": 10790.297250000001, "rle": {"size": [375, 500], "counts": "ZhU4118Z;?D;D4L4M3L4M3L4M;D<E4K101N2N101N101O0O2O0O2O0O2J5N3M3M201O0010O01O0010O01O10O01O02N2N2N0000001O00O1000001O0000000N2CgLZH[3f7;O1000000000001O000000000001O0O1O1O1O1O1O1O1O1N2O1N3M3N2RMeG\\2^8bMhGW2[8fMiGW2Y8gMhGW2[8fMgGX2\\8fMeGY2m8N2N2O1N2N3M3N2M3M3N2N2M3N2N2N2M3N2N2N2M3N2N3L3N2N2@h\\5"}, "label": "a pot holding a green plant with no flowers behind and above a pot with yellow flowers"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan pot of flowers has yet to bloom\"."}], "task": "refering", "answer_template": "The \"a tan pot of flowers has yet to bloom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 121, 122, 123, 124], "bbox": [0.72978, 0.17010666666666666, 0.9691, 0.5094133333333334], "iscrowd": 0, "area": 10790.297250000001, "rle": {"size": [375, 500], "counts": "ZhU4118Z;?D;D4L4M3L4M3L4M;D<E4K101N2N101N101O0O2O0O2O0O2J5N3M3M201O0010O01O0010O01O10O01O02N2N2N0000001O00O1000001O0000000N2CgLZH[3f7;O1000000000001O000000000001O0O1O1O1O1O1O1O1O1N2O1N3M3N2RMeG\\2^8bMhGW2[8fMiGW2Y8gMhGW2[8fMgGX2\\8fMeGY2m8N2N2O1N2N3M3N2M3M3N2N2M3N2N2N2M3N2N2N2M3N2N3L3N2N2@h\\5"}, "label": "a tan pot of flowers has yet to bloom"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tank on the back of a white toilet , which has red flowers growing in it\"."}], "task": "refering", "answer_template": "The \"the tank on the back of a white toilet , which has red flowers growing in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 22, 23, 24, 25, 26, 27, 28, 44, 45, 46, 62, 63, 64, 81, 82, 98, 99, 100, 116, 117], "bbox": [0.26282, 0.055946666666666665, 0.61048, 0.4746666666666667], "iscrowd": 0, "area": 8308.552049999998, "rle": {"size": [375, 500], "counts": "oP`17^;000100O1O010O1000000O1000000001O00001O00001O001O00010O001O001O1O1O1O0101O00001O0000001O0001O01O001O3M4L01O0001O000001O000000001O0000001N100O101N2O001O001O1O010O001O1O001O10O101O1N2O1OO1O002N1O2M2O2N1O2N2N1N3N1N3M4L5L4_HcN_4c1YK`Nf4e1TK]Nk4h1nJ[NR5j1fJYNY5l1aJVN^5o1[JTNe5Q2SJSNl5Q2nIQNS6R2fIQNZ6W2[IlMf6b31O001O1O001N2O1O1O1O1O2N1O1O000O10001O0000000000000O2M2O1N3N1O1N2O2N1N2O2N1N2O1O2M2O1O2M2O1O1J7H7I7I8F9YOg0]OP\\W2"}, "label": "the tank on the back of a white toilet , which has red flowers growing in it"}]}
{"id": 133609, "image": "COCO_train2014_000000133609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the water tank of a toilet , outdoors and filled with potted plants\"."}], "task": "refering", "answer_template": "The \"the water tank of a toilet , outdoors and filled with potted plants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 22, 23, 24, 25, 26, 27, 28, 44, 45, 46, 62, 63, 64, 81, 82, 98, 99, 100, 116, 117], "bbox": [0.26282, 0.055946666666666665, 0.61048, 0.4746666666666667], "iscrowd": 0, "area": 8308.552049999998, "rle": {"size": [375, 500], "counts": "oP`17^;000100O1O010O1000000O1000000001O00001O00001O001O00010O001O001O1O1O1O0101O00001O0000001O0001O01O001O3M4L01O0001O000001O000000001O0000001N100O101N2O001O001O1O010O001O1O001O10O101O1N2O1OO1O002N1O2M2O2N1O2N2N1N3N1N3M4L5L4_HcN_4c1YK`Nf4e1TK]Nk4h1nJ[NR5j1fJYNY5l1aJVN^5o1[JTNe5Q2SJSNl5Q2nIQNS6R2fIQNZ6W2[IlMf6b31O001O1O001N2O1O1O1O1O2N1O1O000O10001O0000000000000O2M2O1N3N1O1N2O2N1N2O2N1N2O1O2M2O1O2M2O1O1J7H7I7I8F9YOg0]OP\\W2"}, "label": "the water tank of a toilet , outdoors and filled with potted plants"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3 tiered white and blue rounded square cake\"."}], "task": "refering", "answer_template": "The \"3 tiered white and blue rounded square cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 307, 308], "bbox": [0.24025, 0.1955058823529412, 0.5387187499999999, 0.8831529411764707], "iscrowd": 0, "area": 43011.3833, "rle": {"size": [425, 640], "counts": "SWP24o<7I6K6N1O2O0O2_NOhE1l9b0lE@g9l0YFTOQ9b1nF`NR8_2nGaMQ8`2nGbMQ8]2PHcMo7^2PHdMo7\\2QHeMm7\\2RHfMm7Z2SHgMk7Z2UHgMj7Y2UHiMi7X2WHiMh7W2WHkMg7]2RHdMm7m3N2O1N2N101N2O1G9XOh0^Ob0N2M3M2O2M3N2M3M2O2M3M3N2M2O2M3M3N2O001O1O100O1O001O1O1O1O001O100O1O001O1O1O1O0000010O0000000000000001O0000000000000000001O00000000000000000O2N100O100O100O100O100O100O100O100O100O2N100O100O100O100O100O100O1O1O1O100O1O1O1O1O101N1O1O1O1O1O101N<Dn0RO1O2N100O2N1O2N1O1O2O0O2N1O2N1O101N1O2N1SMcHM^71fHL\\71gHMZ7SOcHoN7m1W7QOiHkN3R2U7QOPIdN0X2R7QOmIm0X6nNlIn0X6mNlIP1Y6jNlIS1W6hNnIl0_6mNnI9e6Be3K6JfXj3"}, "label": "3 tiered white and blue rounded square cake"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3 tier cake\"."}], "task": "refering", "answer_template": "The \"3 tier cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 307, 308], "bbox": [0.24025, 0.1955058823529412, 0.5387187499999999, 0.8831529411764707], "iscrowd": 0, "area": 43011.3833, "rle": {"size": [425, 640], "counts": "SWP24o<7I6K6N1O2O0O2_NOhE1l9b0lE@g9l0YFTOQ9b1nF`NR8_2nGaMQ8`2nGbMQ8]2PHcMo7^2PHdMo7\\2QHeMm7\\2RHfMm7Z2SHgMk7Z2UHgMj7Y2UHiMi7X2WHiMh7W2WHkMg7]2RHdMm7m3N2O1N2N101N2O1G9XOh0^Ob0N2M3M2O2M3N2M3M2O2M3M3N2M2O2M3M3N2O001O1O100O1O001O1O1O1O001O100O1O001O1O1O1O0000010O0000000000000001O0000000000000000001O00000000000000000O2N100O100O100O100O100O100O100O100O100O2N100O100O100O100O100O100O1O1O1O100O1O1O1O1O101N1O1O1O1O1O101N<Dn0RO1O2N100O2N1O2N1O1O2O0O2N1O2N1O101N1O2N1SMcHM^71fHL\\71gHMZ7SOcHoN7m1W7QOiHkN3R2U7QOPIdN0X2R7QOmIm0X6nNlIn0X6mNlIP1Y6jNlIS1W6hNnIl0_6mNnI9e6Be3K6JfXj3"}, "label": "3 tier cake"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left most person\"."}], "task": "refering", "answer_template": "The \"left most person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 79, 80, 81, 103, 104, 126, 127, 149, 150, 151, 172, 173, 174, 175, 196, 197, 219], "bbox": [0.439140625, 0.09757647058823529, 0.614703125, 0.6140941176470588], "iscrowd": 0, "area": 10144.548200000003, "rle": {"size": [425, 640], "counts": "RTe34R=3M3M3M3M4M2N2N2N2N2N3M2N2N2O1O2N1N2O1O1O1O2N1O1O1iEfN]8\\1QGYOl8g0bFN[9S2M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N2N2N3M2N2N2O2N1ORMaH2]7LgH4X7IlH8R7ERI;m6BWI>h6_O\\Ib0b6[ObIe0\\6YOhIg0W6VOnIi0Q6VOQJk0l5TOWJl0h5RO[Jn0c5RO`Jm0_5QOdJP1Y5oNkJn0V5oNoJm0R5QORKm0R5mNTKo0o4lNVKP1n4kNXKQ1k4jNZKR1j4iN[KT1g4hN]KV1f4bN`K\\1c4\\NeKb1]4WNkKf1m7N2O1N10000O10O1O010N2O001O001O10I7F:G9M4O0O2N101N1O2N101N2N1OZZV3"}, "label": "left most person"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with long black hair standing in front of a large cake\"."}], "task": "refering", "answer_template": "The \"a woman with long black hair standing in front of a large cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 79, 80, 81, 103, 104, 126, 127, 149, 150, 151, 172, 173, 174, 175, 196, 197, 219], "bbox": [0.439140625, 0.09757647058823529, 0.614703125, 0.6140941176470588], "iscrowd": 0, "area": 10144.548200000003, "rle": {"size": [425, 640], "counts": "RTe34R=3M3M3M3M4M2N2N2N2N2N3M2N2N2O1O2N1N2O1O1O1O2N1O1O1iEfN]8\\1QGYOl8g0bFN[9S2M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N2N2N3M2N2N2O2N1ORMaH2]7LgH4X7IlH8R7ERI;m6BWI>h6_O\\Ib0b6[ObIe0\\6YOhIg0W6VOnIi0Q6VOQJk0l5TOWJl0h5RO[Jn0c5RO`Jm0_5QOdJP1Y5oNkJn0V5oNoJm0R5QORKm0R5mNTKo0o4lNVKP1n4kNXKQ1k4jNZKR1j4iN[KT1g4hN]KV1f4bN`K\\1c4\\NeKb1]4WNkKf1m7N2O1N10000O10O1O010N2O001O001O10I7F:G9M4O0O2N101N1O2N101N2N1OZZV3"}, "label": "a woman with long black hair standing in front of a large cake"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing with black suit\"."}], "task": "refering", "answer_template": "The \"a man standing with black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 41, 42, 43, 63, 64, 65, 66, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 246, 247, 248], "bbox": [0.674515625, 0.008988235294117646, 0.9072968749999999, 0.7370823529411764], "iscrowd": 0, "area": 27350.801399999997, "rle": {"size": [425, 640], "counts": "X`c5?i<a0^O7J6I7J5J7J5J2O2M2O2RMfMiJ\\2V5nM`JT2^5VNWJl1h5^NnId1Q6eNeI\\1Z6oNZIT1d6VORIk0m6_OhHd0W7E_H<`7h2N1O2O0O2N1O2N101N1O2N101N1O2N101N1O2N101N1O2N101N1O2N1O2O0O1O2N101N1O2N1N3N1O2M2O2N1N3N1oNbHVL`7h3fHQL\\7n3iHmKY7Q4lHjKU7U4PIeKS7Y4o0O2N101N1O2N1O2N1O2N1O2N101N1OO1O100O1O1O1O2N2N2N3M2N2O1N3N1O1N2O1O2N1N2fJkL?V3[ORMb0P3\\OSM`0o2_OTM>m2BUM;l2DWM8l2FWM7j2IXM4i2KZM1h2N[MOg20\\MLe23^MJc25`MGb29`MDa2;bMB`2<cM@_2`0cMZOa2e0bMSOd2k0`MmNg2P1]MhNi2V1SNkMR2R2d4N3M2N3M3L3N3M2N3L3N3M2N3M3K4L5J5L5J5LZlh0"}, "label": "a man standing with black suit"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a formally dressed gentleman stands beside an ornate blue - and - white cake and matching cookies\"."}], "task": "refering", "answer_template": "The \"a formally dressed gentleman stands beside an ornate blue - and - white cake and matching cookies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 41, 42, 43, 63, 64, 65, 66, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 246, 247, 248], "bbox": [0.674515625, 0.008988235294117646, 0.9072968749999999, 0.7370823529411764], "iscrowd": 0, "area": 27350.801399999997, "rle": {"size": [425, 640], "counts": "X`c5?i<a0^O7J6I7J5J7J5J2O2M2O2RMfMiJ\\2V5nM`JT2^5VNWJl1h5^NnId1Q6eNeI\\1Z6oNZIT1d6VORIk0m6_OhHd0W7E_H<`7h2N1O2O0O2N1O2N101N1O2N101N1O2N101N1O2N101N1O2N101N1O2N1O2O0O1O2N101N1O2N1N3N1O2M2O2N1N3N1oNbHVL`7h3fHQL\\7n3iHmKY7Q4lHjKU7U4PIeKS7Y4o0O2N101N1O2N1O2N1O2N1O2N101N1OO1O100O1O1O1O2N2N2N3M2N2O1N3N1O1N2O1O2N1N2fJkL?V3[ORMb0P3\\OSM`0o2_OTM>m2BUM;l2DWM8l2FWM7j2IXM4i2KZM1h2N[MOg20\\MLe23^MJc25`MGb29`MDa2;bMB`2<cM@_2`0cMZOa2e0bMSOd2k0`MmNg2P1]MhNi2V1SNkMR2R2d4N3M2N3M3L3N3M2N3L3N3M2N3M3K4L5J5L5J5LZlh0"}, "label": "a formally dressed gentleman stands beside an ornate blue - and - white cake and matching cookies"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tray of blue - and - white frosted , heart - shaped cookies\"."}], "task": "refering", "answer_template": "The \"a tray of blue - and - white frosted , heart - shaped cookies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294], "bbox": [0.565578125, 0.6359529411764705, 0.8326875000000001, 0.8269647058823529], "iscrowd": 0, "area": 8932.711899999997, "rle": {"size": [425, 640], "counts": "i`f41X=2N2N1O2M3N1O2N2N1O2M3N2N1O2N2M2O2N2N1O2N2M3N1O2N2N1O2M3N1O2N2N2M2O2N2N1O2N0O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001N2O1O^R\\1"}, "label": "a tray of blue - and - white frosted , heart - shaped cookies"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"heart - shaped cookies on edge of table\"."}], "task": "refering", "answer_template": "The \"heart - shaped cookies on edge of table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294], "bbox": [0.565578125, 0.6359529411764705, 0.8326875000000001, 0.8269647058823529], "iscrowd": 0, "area": 8932.711899999997, "rle": {"size": [425, 640], "counts": "i`f41X=2N2N1O2M3N1O2N2N1O2M3N2N1O2N2M2O2N2N1O2N2M3N1O2N2N1O2M3N1O2N2N2M2O2N2N1O2N0O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001N2O1O^R\\1"}, "label": "heart - shaped cookies on edge of table"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman dressed in black wearing glasses\"."}], "task": "refering", "answer_template": "The \"a woman dressed in black wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 82, 83, 84, 105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 197, 198, 199, 221, 222], "bbox": [0.5643281250000001, 0.1828470588235294, 0.6827500000000001, 0.6333411764705883], "iscrowd": 0, "area": 10114.551249999999, "rle": {"size": [425, 640], "counts": "aPf47^<d0H8I7H8I8O0O01000O1]EVNg9h1\\F\\N^9a1fFeNT9X1oFmNl8R1UGSOe8m0[GWOZ7ZOeH_11WOX7D_HU18YOW7KXHm0`0YOV71UHf0e0ZOS74UHc0g0ZOR77UH>i0\\OQ79TH;k0]On6`0PH6o0[Oo6\\2mHdMS7]2kHeMS7\\2lHeMT7\\2kHdMT7]2kHeMT7\\2jHeMU7\\2kHdMU7]2iHeMU7k2\\HUMd7Y4N10001O00001O00001O0O101O001O00010O01O010O0010O01O10O01O010O1oK]H[2c7cM`H[2a7cMaH[2a7cMaHW2e7fM^HT2h7jMZH`1^8\\NeGP1R9jNPGn0[9cNQGU1f:H9G<EZSd2"}, "label": "a woman dressed in black wearing glasses"}]}
{"id": 485868, "image": "COCO_train2014_000000485868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"taking women weared spectacles\"."}], "task": "refering", "answer_template": "The \"taking women weared spectacles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 82, 83, 84, 105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 197, 198, 199, 221, 222], "bbox": [0.5643281250000001, 0.1828470588235294, 0.6827500000000001, 0.6333411764705883], "iscrowd": 0, "area": 10114.551249999999, "rle": {"size": [425, 640], "counts": "aPf47^<d0H8I7H8I8O0O01000O1]EVNg9h1\\F\\N^9a1fFeNT9X1oFmNl8R1UGSOe8m0[GWOZ7ZOeH_11WOX7D_HU18YOW7KXHm0`0YOV71UHf0e0ZOS74UHc0g0ZOR77UH>i0\\OQ79TH;k0]On6`0PH6o0[Oo6\\2mHdMS7]2kHeMS7\\2lHeMT7\\2kHdMT7]2kHeMT7\\2jHeMU7\\2kHdMU7]2iHeMU7k2\\HUMd7Y4N10001O00001O00001O0O101O001O00010O01O010O0010O01O10O01O010O1oK]H[2c7cM`H[2a7cMaH[2a7cMaHW2e7fM^HT2h7jMZH`1^8\\NeGP1R9jNPGn0[9cNQGU1f:H9G<EZSd2"}, "label": "taking women weared spectacles"}]}
{"id": 174574, "image": "COCO_train2014_000000174574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player in front\"."}], "task": "refering", "answer_template": "The \"tennis player in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 141, 142, 143, 144, 165, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331, 349, 350, 351, 352, 353, 354], "bbox": [0.15621875000000002, 0.30786343612334804, 0.43040625, 1.0], "iscrowd": 0, "area": 38521.94419999998, "rle": {"size": [454, 640], "counts": "Pg\\1;X=f0B>SJlN_NY1S1UOhNQ1T1QOiNT1R1POiNU1S1nNjNW1Q1mNjNW1T1kNhNY1U1jNgMQOVMY2P5iNYMh2e2ZMmLU3P3nLhLX3W3kL`L\\3_3gLXL`3g3cLQLc3n3_LoKb3Q4cLiKX3]4lL^KT3c4nLZKS3g4nLVKR3k4PMRKP3o4QMoJP3R5PMlJP3V5PMhJP3g0fJX3Z2oKQ3h0gJX3Y2oKo2h0kJY3V2mKP3i0lJZ3T2kKQ3i0nJ\\3R2hKQ3k0oJ\\3R2fKo2n0QK\\3c7bL`H^3_7aLcH_3\\7_LgHa3V7`LlH`3P7aLSI_3j6bLXI^3e6bL^I^3`6bLbI^3^6`LeI_3Z6_LiIb3V6\\LlId3S6[LoIe3Q6YLQJg3n5XLTJh3l5VLVJj3i5ULYJk3g5SL[Jm3d5RL^Jn3b5PL`JP4_5PLbJP4^5nKdJR4[5mKgJS4Y5kKiJV4U5iKmJW4S5gKoJY4P5fKRKZ4n4dKTK\\4k4cKWK]4i4bKXK^4X7000000000000000010O00UHbKn5^4PJeKo5[4PJfKP6Z4oIgKQ6Y4nIiKQ6X4lIjKT6V4kIkKU6X4gIjKX6Y4dIhK\\6\\4^IfKb6^4YIdKf6`4UIaKk6c4hHfKX7^4[HlKb7]5M3M4M2M3N3L30001O010O0010O00010O01O010O00010O0010O01O010O12O1N101NO2N1O2N1O2M2O2N2N3M2N3M2N3M3M2N3M2M4M2N3M2N3M2N3M2N3M2N3M2M4M2N1O2N1O1O2N1O2N1O2M2O2N1O1O2N1O2N2M4M2L5L3L8IS1lNT1mNg]Q5"}, "label": "tennis player in front"}]}
{"id": 174574, "image": "COCO_train2014_000000174574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older woman wearing a purple shirt and glasses\"."}], "task": "refering", "answer_template": "The \"an older woman wearing a purple shirt and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 141, 142, 143, 144, 165, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331, 349, 350, 351, 352, 353, 354], "bbox": [0.15621875000000002, 0.30786343612334804, 0.43040625, 1.0], "iscrowd": 0, "area": 38521.94419999998, "rle": {"size": [454, 640], "counts": "Pg\\1;X=f0B>SJlN_NY1S1UOhNQ1T1QOiNT1R1POiNU1S1nNjNW1Q1mNjNW1T1kNhNY1U1jNgMQOVMY2P5iNYMh2e2ZMmLU3P3nLhLX3W3kL`L\\3_3gLXL`3g3cLQLc3n3_LoKb3Q4cLiKX3]4lL^KT3c4nLZKS3g4nLVKR3k4PMRKP3o4QMoJP3R5PMlJP3V5PMhJP3g0fJX3Z2oKQ3h0gJX3Y2oKo2h0kJY3V2mKP3i0lJZ3T2kKQ3i0nJ\\3R2hKQ3k0oJ\\3R2fKo2n0QK\\3c7bL`H^3_7aLcH_3\\7_LgHa3V7`LlH`3P7aLSI_3j6bLXI^3e6bL^I^3`6bLbI^3^6`LeI_3Z6_LiIb3V6\\LlId3S6[LoIe3Q6YLQJg3n5XLTJh3l5VLVJj3i5ULYJk3g5SL[Jm3d5RL^Jn3b5PL`JP4_5PLbJP4^5nKdJR4[5mKgJS4Y5kKiJV4U5iKmJW4S5gKoJY4P5fKRKZ4n4dKTK\\4k4cKWK]4i4bKXK^4X7000000000000000010O00UHbKn5^4PJeKo5[4PJfKP6Z4oIgKQ6Y4nIiKQ6X4lIjKT6V4kIkKU6X4gIjKX6Y4dIhK\\6\\4^IfKb6^4YIdKf6`4UIaKk6c4hHfKX7^4[HlKb7]5M3M4M2M3N3L30001O010O0010O00010O01O010O00010O0010O01O010O12O1N101NO2N1O2N1O2M2O2N2N3M2N3M2N3M3M2N3M2M4M2N3M2N3M2N3M2N3M2N3M2M4M2N1O2N1O1O2N1O2N1O2M2O2N1O1O2N1O2N2M4M2L5L3L8IS1lNT1mNg]Q5"}, "label": "an older woman wearing a purple shirt and glasses"}]}
{"id": 174574, "image": "COCO_train2014_000000174574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing beige shirt and black pants hitting a tennis ball\"."}], "task": "refering", "answer_template": "The \"a woman wearing beige shirt and black pants hitting a tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 86, 87, 88, 110, 111, 133, 134, 135, 156, 157, 179, 180, 202, 203, 225, 226], "bbox": [0.767125, 0.1545374449339207, 0.8883125, 0.6184140969162996], "iscrowd": 0, "area": 5259.5309000000025, "rle": {"size": [454, 640], "counts": "Sii6<o<l00001N10001N^OXCGh<3^CMa<NeC3Z<GlC9T<BQD>e0Dc:=VEIj:7PEOP;1jD5V;LcD:];F]D`0c;@XDk0b;VOWDV1c;h0O2N1O2M3M2O2fFUMS7m2eHYMT7KVGR3\\1XM\\7IWGU3U1SMc7^3UHbLk7g3lGZLR8o3fGQLZ8W4^GjKa8m4O2O001O0100YN`GgMb8U2cGiM]8T2hG`MDSOf8Y3kG_Ma8]2eG^M_8^2fG]M]8`2hG[M\\8W1cFAU1TO[8Y1dF]OW1UOY8[1eFZOW1VOW8^1gFVOQ;h0SEQOQ;o0QElNS;L\\Dm0c0QO`;7iC;T=BPC=o<AUC=l<_OXCa0X=N2N101N01O010O00010O0010O01O00O2N1O1O2O6IbZo0"}, "label": "a woman wearing beige shirt and black pants hitting a tennis ball"}]}
{"id": 174574, "image": "COCO_train2014_000000174574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a cream shirt and black yoga pants playing tennis and serving a ball\"."}], "task": "refering", "answer_template": "The \"a woman in a cream shirt and black yoga pants playing tennis and serving a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 86, 87, 88, 110, 111, 133, 134, 135, 156, 157, 179, 180, 202, 203, 225, 226], "bbox": [0.767125, 0.1545374449339207, 0.8883125, 0.6184140969162996], "iscrowd": 0, "area": 5259.5309000000025, "rle": {"size": [454, 640], "counts": "Sii6<o<l00001N10001N^OXCGh<3^CMa<NeC3Z<GlC9T<BQD>e0Dc:=VEIj:7PEOP;1jD5V;LcD:];F]D`0c;@XDk0b;VOWDV1c;h0O2N1O2M3M2O2fFUMS7m2eHYMT7KVGR3\\1XM\\7IWGU3U1SMc7^3UHbLk7g3lGZLR8o3fGQLZ8W4^GjKa8m4O2O001O0100YN`GgMb8U2cGiM]8T2hG`MDSOf8Y3kG_Ma8]2eG^M_8^2fG]M]8`2hG[M\\8W1cFAU1TO[8Y1dF]OW1UOY8[1eFZOW1VOW8^1gFVOQ;h0SEQOQ;o0QElNS;L\\Dm0c0QO`;7iC;T=BPC=o<AUC=l<_OXCa0X=N2N101N01O010O00010O0010O01O00O2N1O1O2O6IbZo0"}, "label": "a woman in a cream shirt and black yoga pants playing tennis and serving a ball"}]}
{"id": 51706, "image": "COCO_train2014_000000051706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a third baseman is waiting for a play with the shortstop in the background\"."}], "task": "refering", "answer_template": "The \"a third baseman is waiting for a play with the shortstop in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 336, 337, 338, 358, 359, 360, 381, 382, 383, 384, 404, 405], "bbox": [0.567859375, 0.336421875, 0.773296875, 0.76671875], "iscrowd": 0, "area": 16444.426500000005, "rle": {"size": [640, 640], "counts": "]VS71kc06J6m@FW:a0dECX:b0cEAZ:d0bE^O[:g0`E\\O]:i0fAZOT30T;^1gDeNW;]1fDfNX;\\1eDgNY;[1dDhN[;Y1cDiN[;Y1bDjN];W1`DlN_;U1^DnNa;S1\\DPOc;Q1ZDSOd;n0YDUOf;l0XDVOg;k0VDXOi;i0TDZOk;g0RD\\Om;e0PD^Oo;c0nC@R<c0hC@W<e0cC]O\\<g0]C]Ob<g0XCgNYOgN^=g2QCeNCcN[=l2lBcNK`NW=S3fB`N5]NS=Z5oBeJP=[5RCdJm<\\5UCcJj<\\5YCcJf<]5\\CbJb<_5aC_J^<a5cC_J\\<a5fC^JY<c5hC\\JW<f5iCYJU<k5jCTJU<o5jCPJU<S6jClIU<W6jCiIT<Y6lCfIS<]6kCcIU<_6jC`IU<b6d0O0100O100O100O10O0100O100O100O100O0100000O10000O10lLQCWOo<e0\\CbNDWNQ=S3fCVNCfNf<Q3lCoMGPO]<l2QDRNEQOZ<R2oB[NR1a0GQOW<R2VCXNm0b0HTOU<P2[CWNh0b0LVOQ<o1`CUNd0c0OXO[;NkCP2<TN`0b01[OW;3iCk1e0QN:c05]OQ;8iCe1l0PN6a09Ak:<hC`1\\2`Ne;3mA[1`2_Nd;:iAV1f2\\Nc;\\2^DaMd;^2^D_Me;_2o2N2N2N1O2N2N2N2N1FU^O_Nma0[1]^O_Nea0`1>O100O001O1O1O1O100O2O1N2O1N2N2O2M101ZOP]O9Qc0FQ]O9ob0GR]O7ob0HS]ONM0Pc01g0O0O2N10mPk2"}, "label": "a third baseman is waiting for a play with the shortstop in the background"}]}
{"id": 51706, "image": "COCO_train2014_000000051706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player with a long sleeved black undershirt waits for the play\"."}], "task": "refering", "answer_template": "The \"a baseball player with a long sleeved black undershirt waits for the play\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 336, 337, 338, 358, 359, 360, 381, 382, 383, 384, 404, 405], "bbox": [0.567859375, 0.336421875, 0.773296875, 0.76671875], "iscrowd": 0, "area": 16444.426500000005, "rle": {"size": [640, 640], "counts": "]VS71kc06J6m@FW:a0dECX:b0cEAZ:d0bE^O[:g0`E\\O]:i0fAZOT30T;^1gDeNW;]1fDfNX;\\1eDgNY;[1dDhN[;Y1cDiN[;Y1bDjN];W1`DlN_;U1^DnNa;S1\\DPOc;Q1ZDSOd;n0YDUOf;l0XDVOg;k0VDXOi;i0TDZOk;g0RD\\Om;e0PD^Oo;c0nC@R<c0hC@W<e0cC]O\\<g0]C]Ob<g0XCgNYOgN^=g2QCeNCcN[=l2lBcNK`NW=S3fB`N5]NS=Z5oBeJP=[5RCdJm<\\5UCcJj<\\5YCcJf<]5\\CbJb<_5aC_J^<a5cC_J\\<a5fC^JY<c5hC\\JW<f5iCYJU<k5jCTJU<o5jCPJU<S6jClIU<W6jCiIT<Y6lCfIS<]6kCcIU<_6jC`IU<b6d0O0100O100O100O10O0100O100O100O100O0100000O10000O10lLQCWOo<e0\\CbNDWNQ=S3fCVNCfNf<Q3lCoMGPO]<l2QDRNEQOZ<R2oB[NR1a0GQOW<R2VCXNm0b0HTOU<P2[CWNh0b0LVOQ<o1`CUNd0c0OXO[;NkCP2<TN`0b01[OW;3iCk1e0QN:c05]OQ;8iCe1l0PN6a09Ak:<hC`1\\2`Ne;3mA[1`2_Nd;:iAV1f2\\Nc;\\2^DaMd;^2^D_Me;_2o2N2N2N1O2N2N2N2N1FU^O_Nma0[1]^O_Nea0`1>O100O001O1O1O1O100O2O1N2O1N2N2O2M101ZOP]O9Qc0FQ]O9ob0GR]O7ob0HS]ONM0Pc01g0O0O2N10mPk2"}, "label": "a baseball player with a long sleeved black undershirt waits for the play"}]}
{"id": 51706, "image": "COCO_train2014_000000051706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player farthest from the camera\"."}], "task": "refering", "answer_template": "The \"the baseball player farthest from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 76, 77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 167, 168, 170, 171, 191], "bbox": [0.27146875000000004, 0.062671875, 0.46693750000000006, 0.352546875], "iscrowd": 0, "area": 9517.411499999998, "rle": {"size": [640, 640], "counts": "mn\\31nc01O100O1N3N1N2O1N2O1N2O1N2O1N2O2M2O1O1N2N2VMUO]Bm0b=^ORBd0m=G`@WOa0T1n>0Q@SOo0n0P?=m@FQ?=k@ET?=i@EV?=g@EX?<f@FY?;e@[O_OSOk?c1d@[OCQOh?e1c@ZOGQOe?g1a@YOLoNc?h1_@YO1nN_?l1\\@VO7oN[?Z3g@dLY?[3i@dLW?\\3k@aLV?^3f000O100O100O100O100O2N101N101N101N101N1\\OY_OPNg`0R2^_OfMd`0[2?101N2O0O2O1N2O0O2O1N101N2O0_Ob0N2M2O2N_Nl_OYOS`0e0Q@ZOn?c0V@\\Oj?b0Y@^Of?`0]@@b?=b@B^?<e@S1k=jNYBU1g=iN\\BV1c=iN`BV1`=gNcBY1]=eNeB[1[=cNhB\\1Y=`NjB`1W=]NlBc1T=ZNnBf1S=VNQCi1P=nMXCR2i<gM^CX2c<bMbC^2_<aMbC^2_<`MbC`2^<`MbC?fM[1i>TNcC>gM]1g>TNbC=kM]1c>VNcC;lMN5e0]>POcC;QO<l?B^@6c?G`@8`?Fc@9^?Dd@<aa000O1O001O1O1O001O1O1O001L_md6"}, "label": "the baseball player farthest from the camera"}]}
{"id": 51706, "image": "COCO_train2014_000000051706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player on top of picture\"."}], "task": "refering", "answer_template": "The \"player on top of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 76, 77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 167, 168, 170, 171, 191], "bbox": [0.27146875000000004, 0.062671875, 0.46693750000000006, 0.352546875], "iscrowd": 0, "area": 9517.411499999998, "rle": {"size": [640, 640], "counts": "mn\\31nc01O100O1N3N1N2O1N2O1N2O1N2O1N2O2M2O1O1N2N2VMUO]Bm0b=^ORBd0m=G`@WOa0T1n>0Q@SOo0n0P?=m@FQ?=k@ET?=i@EV?=g@EX?<f@FY?;e@[O_OSOk?c1d@[OCQOh?e1c@ZOGQOe?g1a@YOLoNc?h1_@YO1nN_?l1\\@VO7oN[?Z3g@dLY?[3i@dLW?\\3k@aLV?^3f000O100O100O100O100O2N101N101N101N101N1\\OY_OPNg`0R2^_OfMd`0[2?101N2O0O2O1N2O0O2O1N101N2O0_Ob0N2M2O2N_Nl_OYOS`0e0Q@ZOn?c0V@\\Oj?b0Y@^Of?`0]@@b?=b@B^?<e@S1k=jNYBU1g=iN\\BV1c=iN`BV1`=gNcBY1]=eNeB[1[=cNhB\\1Y=`NjB`1W=]NlBc1T=ZNnBf1S=VNQCi1P=nMXCR2i<gM^CX2c<bMbC^2_<aMbC^2_<`MbC`2^<`MbC?fM[1i>TNcC>gM]1g>TNbC=kM]1c>VNcC;lMN5e0]>POcC;QO<l?B^@6c?G`@8`?Fc@9^?Dd@<aa000O1O001O1O1O001O1O1O001L_md6"}, "label": "player on top of picture"}]}
{"id": 51706, "image": "COCO_train2014_000000051706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mr datz with the number 28\"."}], "task": "refering", "answer_template": "The \"mr datz with the number 28\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [339, 340, 341, 362, 363, 364, 383, 384, 385, 386, 387, 405, 406, 407, 408, 409, 428, 429, 430, 431, 432, 433, 450, 451, 452, 453, 454, 455, 456, 474, 475, 476, 477, 478, 479, 497, 498, 499, 500, 501, 502, 521, 522, 523, 524, 525], "bbox": [0.5962500000000001, 0.612296875, 0.87165625, 0.9973281249999999], "iscrowd": 0, "area": 27308.8516, "rle": {"size": [640, 640], "counts": "jX_74ic05K5K5K5K5K5M3M3N2M3M3M3N2M3M3L4M3M3M2N2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N1N3N2N2N2N2M3N2N2N2M3N2N2N2N2Mj0WO100000000O100000000O1000000O100000000O10001O000O100000000O100000000O100gN_JSDa5l;bJRD^5l;fJRDZ5m;jJoCW5o;mJoCS5P<PKmCQ5Q<SK_C[5_<T1O1N2O1N2O1N2O1N200O2O0000000000000000000000000O100000000000000000000000SKhCR2Y<lMnCn1R<RNSDi1m;VNYDe1h;YN[De1e;ZN]Df1b;YN`Df1a;XNaDg1_;XNcDg1^;WNdDh1\\;WNfDh1[;WNfDh1[;VNhDh1Y;VNiDi1W;VNkDi1V;UNlDj1U;TNmDk1T;SNnDl1R;RNQEm1P;QNSEm1n:QNTEn1o:nMSEQ2Q;iMREV2n:fMWEY2j:aM\\E^2d:^MaEa2`:YMgEe2Y:WMlEh2T:WMnEc2W:\\MkE^2[:_MhE[2]:dMeEV2`:iMbEQ2[>J5J6J6J6J6J6J6D=D;D[gb1"}, "label": "mr datz with the number 28"}]}
{"id": 51706, "image": "COCO_train2014_000000051706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player named datz\"."}], "task": "refering", "answer_template": "The \"a baseball player named datz\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [339, 340, 341, 362, 363, 364, 383, 384, 385, 386, 387, 405, 406, 407, 408, 409, 428, 429, 430, 431, 432, 433, 450, 451, 452, 453, 454, 455, 456, 474, 475, 476, 477, 478, 479, 497, 498, 499, 500, 501, 502, 521, 522, 523, 524, 525], "bbox": [0.5962500000000001, 0.612296875, 0.87165625, 0.9973281249999999], "iscrowd": 0, "area": 27308.8516, "rle": {"size": [640, 640], "counts": "jX_74ic05K5K5K5K5K5M3M3N2M3M3M3N2M3M3L4M3M3M2N2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N1N3N2N2N2N2M3N2N2N2M3N2N2N2N2Mj0WO100000000O100000000O1000000O100000000O10001O000O100000000O100000000O100gN_JSDa5l;bJRD^5l;fJRDZ5m;jJoCW5o;mJoCS5P<PKmCQ5Q<SK_C[5_<T1O1N2O1N2O1N2O1N200O2O0000000000000000000000000O100000000000000000000000SKhCR2Y<lMnCn1R<RNSDi1m;VNYDe1h;YN[De1e;ZN]Df1b;YN`Df1a;XNaDg1_;XNcDg1^;WNdDh1\\;WNfDh1[;WNfDh1[;VNhDh1Y;VNiDi1W;VNkDi1V;UNlDj1U;TNmDk1T;SNnDl1R;RNQEm1P;QNSEm1n:QNTEn1o:nMSEQ2Q;iMREV2n:fMWEY2j:aM\\E^2d:^MaEa2`:YMgEe2Y:WMlEh2T:WMnEc2W:\\MkE^2[:_MhE[2]:dMeEV2`:iMbEQ2[>J5J6J6J6J6J6J6D=D;D[gb1"}, "label": "a baseball player named datz"}]}
{"id": 543233, "image": "COCO_train2014_000000543233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the row of bleacher seat the starbucks cup is sitting on\"."}], "task": "refering", "answer_template": "The \"the row of bleacher seat the starbucks cup is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 140, 141, 162, 163, 164, 165, 185, 186, 187, 188, 189, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.08165625, 0.393231850117096, 0.526453125, 0.9843793911007025], "iscrowd": 0, "area": 33336.64245, "rle": {"size": [427, 640], "counts": "ole04l<>B?A>B>B>J6M4L3M3M3M4L3M3M3M3M4L3M3M3M4K4M3M3M3M4L3M3M3M4L3M3M3M3M4L3N2N2N3M2N2N2N2N3M2N2O1N3M2N2N2N2N3M2N2N2N3N1N2N1OO1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N6J;E;E:F000000O100000000000000O100000000O1M3N2N2M3N2N2N2001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O][n3"}, "label": "the row of bleacher seat the starbucks cup is sitting on"}]}
{"id": 543233, "image": "COCO_train2014_000000543233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench with coffee on it\"."}], "task": "refering", "answer_template": "The \"a wooden bench with coffee on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 140, 141, 162, 163, 164, 165, 185, 186, 187, 188, 189, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.08165625, 0.393231850117096, 0.526453125, 0.9843793911007025], "iscrowd": 0, "area": 33336.64245, "rle": {"size": [427, 640], "counts": "ole04l<>B?A>B>B>J6M4L3M3M3M4L3M3M3M3M4L3M3M3M4K4M3M3M3M4L3M3M3M4L3M3M3M3M4L3N2N2N3M2N2N2N2N3M2N2O1N3M2N2N2N2N3M2N2N2N3N1N2N1OO1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O2N6J;E;E:F000000O100000000000000O100000000O1M3N2N2M3N2N2N2001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O][n3"}, "label": "a wooden bench with coffee on it"}]}
{"id": 543233, "image": "COCO_train2014_000000543233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the upper level seating bench in a gym\"."}], "task": "refering", "answer_template": "The \"the upper level seating bench in a gym\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.19940625, 0.3213583138173302, 1.0, 0.9887587822014052], "iscrowd": 0, "area": 65635.60979999999, "rle": {"size": [427, 640], "counts": "]`e14W=4L5K4K5L5K4L4L4L2N2N3M2M2O1O1O001O1O001O1N2O001O1O001O1O0O2O1O00100O1O010O1O010O1O010O1O10O01O1O010O1O010O1O010O1O10O01O10O01O10O01O100O00100O00100O00100O1O010O1O010O1O010O1O010O1O10O01O10O01O1O010O1O10O01O10O01O10O01O10O01O100O00100O00100O00100O1O010O1O010O1O010O1O010O1O10O01O10O01O10O01O100O001O10O01O10O01O10O01O100O00100O00100O00100O1O010O1O010O1O010O1O010O1O10O01O10O01O10O01O100O00100O001O10O01O10O01O100O00100O00100O00100O1O010O1O010O1O010O1O010O1O10O01O10O01O10O01O100O00100O00100O001O10O01O100O00100O00100O00100O1O010O1O010O1O010O1O010O1O10O01O10O01O10O01O100O00100O00100O0010O01O10O01O00100O0010O01O10O01O010O1O010O1O010O00100O0010O01O10O01O010O1O010O00100O0010O01O10O01O010O1O010O001O10O01O010O0001O00001O00001O00001O00001O00001O00001O000010O0001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00010O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O000PNkH"}, "label": "the upper level seating bench in a gym"}]}
{"id": 117250, "image": "COCO_train2014_000000117250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier with a white hat on\"."}], "task": "refering", "answer_template": "The \"a skier with a white hat on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 60, 61, 62, 83, 84, 85, 106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 175, 176, 177, 198, 199, 200, 221, 222, 223, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 293, 313, 314], "bbox": [0.599015625, 0.07843096234309624, 0.7704531250000001, 0.7918410041841004], "iscrowd": 0, "area": 17756.033499999994, "rle": {"size": [478, 640], "counts": "cSc51i>6J6J5K6J6G9SOm0SOl0ZOg0G9G9G8H9F9H8H7I8H8H8G9D;E<D<D<D<D<DRO[JYI]5h6lJWIn4h6ZKWIa4g6gKXIS4f6VLYIe3f6bLXI\\3e6kLZIR3d6UMZIi2b6^M\\I`2c6dM\\IZ2d6jMZIT2g6nMWIQ2i6RNVIm1Q7nMnHR2_7cM^H^2P8UMoGk2^8iLaGW3l8]LRGd3R9ZLlFf3T9[LjFe3W9\\LhFd3Y9\\LfFd3Z9^LcFc3]9^LbFb3^9_LaFa3_9o000001N11O1O001O100O001O1O1O013aJbJSNa1ZOR4]2fJlMj1ZOe3UOjJc20bNR2YOW4m1PJ]NP2EU4g1TJWNl12V4`1WJRNh1=W4Z1`MfNd2U1^MjNe2S1\\MmNf2P1ZMPOj2l0WMSOl2j0UMUOo2g0QMZOQ3c0PM\\OS3a0nL^OS3a0mL_OS3a0nL_OR3`0oL_OR3`0nL@R3`0oL_OR3`0nL@R3a0nL_OR3`0oL_OQ3a0oL_OR3`0oL_OQ3a0PM_Oo2a0QM]OR3b0e5000001O0000001O0O101N6J5L4Ki^T2"}, "label": "a skier with a white hat on"}]}
{"id": 117250, "image": "COCO_train2014_000000117250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"both snowskate of a man with a lady beside\"."}], "task": "refering", "answer_template": "The \"both snowskate of a man with a lady beside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 286, 287, 311, 313, 334, 335, 336, 337, 338, 358, 359, 361, 362, 382, 383, 385, 386, 387], "bbox": [0.43842187499999996, 0.6883054393305439, 0.8419374999999999, 0.9818200836820083], "iscrowd": 0, "area": 3462.972100000001, "rle": {"size": [478, 640], "counts": "n_S41m>2M2O1O2M2O101O000010O00010O00010O0000\\OFdB:Z=GfB9Y=HgB8W=JjB6T=KlB5S=LmB4Q=NPC2n<ORCOo<1RCLR=3nBJU=6j00001O3M3MnU<OQjC2N1O2N2N2N2O100O1000O10000000O100000O1000O100000000000O010000000000O10O1000000000O100000O1000lNCeC>Z<BgC>X<ChC=W<CjC=U<DkC<T<ElC:T<FlC;S<FmC:R<GnC9Q<GPD9o;HQD8n;HSD8l;ITD7k;JTD7k;IVD6j;KVD5h;MXD3g;MZD3f;MZD3f;LZD5e;L[D4e;L[D4e;K\\D4d;M\\D3d;M\\D3c;M^D3b;M]D4c;K^D5a;L_D4a;L_D4a;K`D5_;LaD3`;M_D4`;LaD4_;LaD4_;KbD4^;MbD2_;NaD1`;NaD1_;0`D0a;0_DOa;1`DNa;2_DMb;2_DMa;4_DKb;5]DKd;4]DKc;6]DId;7\\DHe;7\\DHd;9\\DFe;9\\DFd;;[DEf;;ZDDg;;^11000O010O01000O01000O0100O0100O01000O01000O010O01000O01000O010O01000O01000O010O01000O01000O010O01000O01000O010O10O10O01000O010O10O10O01000O010O10O01O001O1O001O1O001O001O1O001O1O001Onc_1"}, "label": "both snowskate of a man with a lady beside"}]}
{"id": 117250, "image": "COCO_train2014_000000117250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and blue snow skis on snow\"."}], "task": "refering", "answer_template": "The \"white and blue snow skis on snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 286, 287, 311, 313, 334, 335, 336, 337, 338, 358, 359, 361, 362, 382, 383, 385, 386, 387], "bbox": [0.43842187499999996, 0.6883054393305439, 0.8419374999999999, 0.9818200836820083], "iscrowd": 0, "area": 3462.972100000001, "rle": {"size": [478, 640], "counts": "n_S41m>2M2O1O2M2O101O000010O00010O00010O0000\\OFdB:Z=GfB9Y=HgB8W=JjB6T=KlB5S=LmB4Q=NPC2n<ORCOo<1RCLR=3nBJU=6j00001O3M3MnU<OQjC2N1O2N2N2N2O100O1000O10000000O100000O1000O100000000000O010000000000O10O1000000000O100000O1000lNCeC>Z<BgC>X<ChC=W<CjC=U<DkC<T<ElC:T<FlC;S<FmC:R<GnC9Q<GPD9o;HQD8n;HSD8l;ITD7k;JTD7k;IVD6j;KVD5h;MXD3g;MZD3f;MZD3f;LZD5e;L[D4e;L[D4e;K\\D4d;M\\D3d;M\\D3c;M^D3b;M]D4c;K^D5a;L_D4a;L_D4a;K`D5_;LaD3`;M_D4`;LaD4_;LaD4_;KbD4^;MbD2_;NaD1`;NaD1_;0`D0a;0_DOa;1`DNa;2_DMb;2_DMa;4_DKb;5]DKd;4]DKc;6]DId;7\\DHe;7\\DHd;9\\DFe;9\\DFd;;[DEf;;ZDDg;;^11000O010O01000O01000O0100O0100O01000O01000O010O01000O01000O010O01000O01000O010O01000O01000O010O01000O01000O010O10O10O01000O010O10O10O01000O010O10O01O001O1O001O1O001O001O1O001O1O001Onc_1"}, "label": "white and blue snow skis on snow"}]}
{"id": 117250, "image": "COCO_train2014_000000117250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with arm around woman\"."}], "task": "refering", "answer_template": "The \"man with arm around woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312], "bbox": [0.477859375, 0.027071129707112968, 0.6399375, 0.8167782426778243], "iscrowd": 0, "area": 26825.776549999995, "rle": {"size": [478, 640], "counts": "ko^45c>9H8XGH^2a0eL:U3NnKm0k3[OYK`1a4hNbJS2X5UNlId2n5eMfIe2T6bMaIh2X6aM\\Ii2b6k2N2N2N2N2M3N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2M2O1O001OaNoJRIQ5m6[KiHd4V7hK_HX4`7mK]HR4b7RL[Hn3d7WLYHh3f7]LVHn1=bM\\7e0THh1g0\\MT7P1SHd1n0VMn6[1PH`1X1nLg6Q5^IiJa6X5a12N2O1N2NfKiFQ3U9bL[G]3c8VLmGi3Q8iKaHU4U9O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1OYOcLnE\\3l9kLTFT3f9SMYFm2a9ZM^Ff2\\9aMcF_2V9iMiFW2Q9PNnFP2f8]NYGc1V8oNiGR1d7AZH`0T73iHOg6b0VIAj6Y401]JTHS4k7hKZHX4g7bK^HV4j7fK[HP4o7jKVHm3S8dK]HR4\\9F;F9G:F9G9F;F9G:F9G:EgY[3"}, "label": "man with arm around woman"}]}
{"id": 117250, "image": "COCO_train2014_000000117250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in brown pants and shirt skiing\"."}], "task": "refering", "answer_template": "The \"a man in brown pants and shirt skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312], "bbox": [0.477859375, 0.027071129707112968, 0.6399375, 0.8167782426778243], "iscrowd": 0, "area": 26825.776549999995, "rle": {"size": [478, 640], "counts": "ko^45c>9H8XGH^2a0eL:U3NnKm0k3[OYK`1a4hNbJS2X5UNlId2n5eMfIe2T6bMaIh2X6aM\\Ii2b6k2N2N2N2N2M3N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2M2O1O001OaNoJRIQ5m6[KiHd4V7hK_HX4`7mK]HR4b7RL[Hn3d7WLYHh3f7]LVHn1=bM\\7e0THh1g0\\MT7P1SHd1n0VMn6[1PH`1X1nLg6Q5^IiJa6X5a12N2O1N2NfKiFQ3U9bL[G]3c8VLmGi3Q8iKaHU4U9O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1OYOcLnE\\3l9kLTFT3f9SMYFm2a9ZM^Ff2\\9aMcF_2V9iMiFW2Q9PNnFP2f8]NYGc1V8oNiGR1d7AZH`0T73iHOg6b0VIAj6Y401]JTHS4k7hKZHX4g7bK^HV4j7fK[HP4o7jKVHm3S8dK]HR4\\9F;F9G:F9G9F;F9G:F9G:EgY[3"}, "label": "a man in brown pants and shirt skiing"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard of the computer that is charging\"."}], "task": "refering", "answer_template": "The \"the keyboard of the computer that is charging\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307], "bbox": [0.5381045751633987, 0.5044934640522876, 0.9704411764705881, 0.6370261437908497], "iscrowd": 0, "area": 19411.610050000003, "rle": {"size": [612, 612], "counts": "imT62Rc0c0]Oc0]Od0\\Oa0_O000000000000001O00000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003N3L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M4L3M3M3M4L3MT\\:"}, "label": "the keyboard of the computer that is charging"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"keyboard on right laptop\"."}], "task": "refering", "answer_template": "The \"keyboard on right laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307], "bbox": [0.5381045751633987, 0.5044934640522876, 0.9704411764705881, 0.6370261437908497], "iscrowd": 0, "area": 19411.610050000003, "rle": {"size": [612, 612], "counts": "imT62Rc0c0]Oc0]Od0\\Oa0_O000000000000001O00000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003N3L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M4L3M3M3M4L3MT\\:"}, "label": "keyboard on right laptop"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver mac computer with a blank screen\"."}], "task": "refering", "answer_template": "The \"a silver mac computer with a blank screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.0, 0.17977124183006535, 0.4674183006535948, 0.7505555555555555], "iscrowd": 0, "area": 96513.46074999998, "rle": {"size": [612, 612], "counts": "a39jb0d0]Od0\\Oc0lCT1j2_OcKa1\\4SOfJX1Z5[OoIi0Q6KWI9i6:aHI^7k0kGYOV8j5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000001O0O101O000O2O00Q2oMl1RNh1WN8G9[MicR6"}, "label": "a silver mac computer with a blank screen"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blank screen of a mac book air\"."}], "task": "refering", "answer_template": "The \"blank screen of a mac book air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.0, 0.17977124183006535, 0.4674183006535948, 0.7505555555555555], "iscrowd": 0, "area": 96513.46074999998, "rle": {"size": [612, 612], "counts": "a39jb0d0]Od0\\Oc0lCT1j2_OcKa1\\4SOfJX1Z5[OoIi0Q6KWI9i6:aHI^7k0kGYOV8j5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000001O0O101O000O2O00Q2oMl1RNh1WN8G9[MicR6"}, "label": "blank screen of a mac book air"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard on the left\"."}], "task": "refering", "answer_template": "The \"the keyboard on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.013480392156862746, 0.49003267973856207, 0.44269607843137254, 0.6338562091503267], "iscrowd": 0, "area": 18907.94560000001, "rle": {"size": [612, 612], "counts": "SU51nb05L4K5L4K5K5L4K4M4K5L4K5L4K5L4K5L4M300000000000O100000O100000000000000000000000O1000000000000000000000000000000O100000000000O1000000000000000O1000000000000000000000000000000O100000000000000000O100000000000O1000000000000000000000000000000O100000000000000000000000O100000O10000000000000000000000000000O100000000000000000000000000000O01000000000000000000000000000000O1000000000000000000000000000000O10O1000000000000000000000000000O1000000000000000000000000000000O10000000O10000000000000000000O1000000000003M7I6J7I6J7I6J7I6J7H7JT^[6"}, "label": "the keyboard on the left"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"keyboard of the laptop in the left side of the image\"."}], "task": "refering", "answer_template": "The \"keyboard of the laptop in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.013480392156862746, 0.49003267973856207, 0.44269607843137254, 0.6338562091503267], "iscrowd": 0, "area": 18907.94560000001, "rle": {"size": [612, 612], "counts": "SU51nb05L4K5L4K5K5L4K4M4K5L4K5L4K5L4K5L4M300000000000O100000O100000000000000000000000O1000000000000000000000000000000O100000000000O1000000000000000O1000000000000000000000000000000O100000000000000000O100000000000O1000000000000000000000000000000O100000000000000000000000O100000O10000000000000000000000000000O100000000000000000000000000000O01000000000000000000000000000000O1000000000000000000000000000000O10O1000000000000000000000000000O1000000000000000000000000000000O10000000O10000000000000000000O1000000000003M7I6J7I6J7I6J7I6J7H7JT^[6"}, "label": "keyboard of the laptop in the left side of the image"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop computer with a background of a galaxy\"."}], "task": "refering", "answer_template": "The \"a laptop computer with a background of a galaxy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373], "bbox": [0.5033660130718954, 0.1926143790849673, 1.0, 0.7724019607843137], "iscrowd": 0, "area": 100235.90579999998, "rle": {"size": [612, 612], "counts": "m[h5S1ca0V2kMV2iMo0RO=B>B>C=B>C=B>C=B>C=L40000000000000000O100000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000O10000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000dJUH_Ok7JnH4R7VOgIg0Y6bN`J\\1`5nMYKo1g4ZMRLd2n3fLkLW3U3RLdMl3\\2^K]N_4V8N2N3M2O2M2N3M2N2N3M2N3M2N3M2N2NcD"}, "label": "a laptop computer with a background of a galaxy"}]}
{"id": 559618, "image": "COCO_train2014_000000559618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop with a picture of the night sky as a wallpaper that has a pop up message on it\"."}], "task": "refering", "answer_template": "The \"the laptop with a picture of the night sky as a wallpaper that has a pop up message on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373], "bbox": [0.5033660130718954, 0.1926143790849673, 1.0, 0.7724019607843137], "iscrowd": 0, "area": 100235.90579999998, "rle": {"size": [612, 612], "counts": "m[h5S1ca0V2kMV2iMo0RO=B>B>C=B>C=B>C=B>C=L40000000000000000O100000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000O10000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000dJUH_Ok7JnH4R7VOgIg0Y6bN`J\\1`5nMYKo1g4ZMRLd2n3fLkLW3U3RLdMl3\\2^K]N_4V8N2N3M2O2M2N3M2N2N3M2N3M2N3M2N2NcD"}, "label": "the laptop with a picture of the night sky as a wallpaper that has a pop up message on it"}]}
{"id": 412167, "image": "COCO_train2014_000000412167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tail of a very large elephant walking away from the watering hole\"."}], "task": "refering", "answer_template": "The \"the tail of a very large elephant walking away from the watering hole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 184, 186, 187, 207, 209, 210], "bbox": [0.0, 0.003134851138353765, 0.17328125, 0.4846234676007006], "iscrowd": 0, "area": 21482.5919, "rle": {"size": [571, 640], "counts": "2_8\\93M00O100O100O100O100O10000jMjFfKW9S4SGjKm8Q4\\GkKd8P4dGlK]8o3lGmKT8Q4QHlKo7R4VHjKk7U4XHiKh7U4]HgKd7W4aHfK`7X4dHdK]7\\4hH_KX7a4Q30000O10000001O001O001O0000001O000000001O001O1O1O1O1O1O2^F_Jf6b5WIaJh6a5TIbJk6`5QIbJo6`5mHcJR7_5lHaJT7a5iHaJV7a5gH`JY7b5dH_J\\7c5aH^J_7b5_H`Ja7_5^HcJb7]5[HfJe7Y5ZHiJf7W5XHkJh7U5`GcK`8\\4cGbK]8^4dGaK]8^4eG`K]8^4dGaK]8]4fGaK[8^4eGbK\\8]4dGcK]8\\4cGdK_8Y4aGhK`8W4`GiKa8V4_GjKb8U4^GkKc8S4^GmKd8Q4\\GoKe8P4[GoKg8P4YGPLh8n3YGRLi8l3WGTLj8j3WGVLk8h3UGWLm8g3TGYLn8e3RG[Lo8c3RG]LP9a3PG_LQ9_3PG`LS9^3mFbLT9U3TGkLn8i2\\GVMf8`2cG`M_8_2`GaMa8`2]G_Mf8a2XG_MS9W2mFhMb9k1]FTN]<SOQD=j>M3M3L3M4L3LVgV9"}, "label": "the tail of a very large elephant walking away from the watering hole"}]}
{"id": 412167, "image": "COCO_train2014_000000412167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big gray elephant wtih a long tail\"."}], "task": "refering", "answer_template": "The \"a big gray elephant wtih a long tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 184, 186, 187, 207, 209, 210], "bbox": [0.0, 0.003134851138353765, 0.17328125, 0.4846234676007006], "iscrowd": 0, "area": 21482.5919, "rle": {"size": [571, 640], "counts": "2_8\\93M00O100O100O100O100O10000jMjFfKW9S4SGjKm8Q4\\GkKd8P4dGlK]8o3lGmKT8Q4QHlKo7R4VHjKk7U4XHiKh7U4]HgKd7W4aHfK`7X4dHdK]7\\4hH_KX7a4Q30000O10000001O001O001O0000001O000000001O001O1O1O1O1O1O2^F_Jf6b5WIaJh6a5TIbJk6`5QIbJo6`5mHcJR7_5lHaJT7a5iHaJV7a5gH`JY7b5dH_J\\7c5aH^J_7b5_H`Ja7_5^HcJb7]5[HfJe7Y5ZHiJf7W5XHkJh7U5`GcK`8\\4cGbK]8^4dGaK]8^4eG`K]8^4dGaK]8]4fGaK[8^4eGbK\\8]4dGcK]8\\4cGdK_8Y4aGhK`8W4`GiKa8V4_GjKb8U4^GkKc8S4^GmKd8Q4\\GoKe8P4[GoKg8P4YGPLh8n3YGRLi8l3WGTLj8j3WGVLk8h3UGWLm8g3TGYLn8e3RG[Lo8c3RG]LP9a3PG_LQ9_3PG`LS9^3mFbLT9U3TGkLn8i2\\GVMf8`2cG`M_8_2`GaMa8`2]G_Mf8a2XG_MS9W2mFhMb9k1]FTN]<SOQD=j>M3M3L3M4L3LVgV9"}, "label": "a big gray elephant wtih a long tail"}]}
{"id": 412167, "image": "COCO_train2014_000000412167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest elephant in the image\"."}], "task": "refering", "answer_template": "The \"the smallest elephant in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 282, 283, 284], "bbox": [0.10109375000000001, 0.2746584938704028, 0.404, 0.6141856392294222], "iscrowd": 0, "area": 22624.22345, "rle": {"size": [571, 640], "counts": "[^T14ba08I6K5N2N3M2N2M3N00000O100000001N1000000O1000000O100O100O010O100O010O100O101N101N101N2O0O2O0O2O1N102N3M2M3N00O100O1O1O00100O1O1O100000O2O0000000000000O5L6J6J6J6J6J6J6I6K5K6J5K5K5K6I6K5K5K6J1O1O1O1N2O00001O0O1000000O10000O10000O10000O2O000O10000O10000O100O1O100O1O2O0O1O1ObN_12N2N1O2N2N2N101N2N1O2N2N2N1O2N2O1N1O2N2N1O2N2N2O001O1O0O01N2O1O1O1O001N2O1O1O1O001N2O1O1O1O001N2O7I>B>B>B>\\Oc0QOP1POP1POUVd6"}, "label": "the smallest elephant in the image"}]}
{"id": 412167, "image": "COCO_train2014_000000412167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small elephant drinks water between two other elephants\"."}], "task": "refering", "answer_template": "The \"a small elephant drinks water between two other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 282, 283, 284], "bbox": [0.10109375000000001, 0.2746584938704028, 0.404, 0.6141856392294222], "iscrowd": 0, "area": 22624.22345, "rle": {"size": [571, 640], "counts": "[^T14ba08I6K5N2N3M2N2M3N00000O100000001N1000000O1000000O100O100O010O100O010O100O101N101N101N2O0O2O0O2O1N102N3M2M3N00O100O1O1O00100O1O1O100000O2O0000000000000O5L6J6J6J6J6J6J6I6K5K6J5K5K5K6I6K5K5K6J1O1O1O1N2O00001O0O1000000O10000O10000O10000O2O000O10000O10000O100O1O100O1O2O0O1O1ObN_12N2N1O2N2N2N101N2N1O2N2N2N1O2N2O1N1O2N2N1O2N2N2O001O1O0O01N2O1O1O1O001N2O1O1O1O001N2O1O1O1O001N2O7I>B>B>B>\\Oc0QOP1POP1POUVd6"}, "label": "a small elephant drinks water between two other elephants"}]}
{"id": 412167, "image": "COCO_train2014_000000412167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant to the right of the group\"."}], "task": "refering", "answer_template": "The \"the elephant to the right of the group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 218, 219, 221, 222, 223, 224, 241, 244, 245, 246, 247, 263, 264, 265, 267, 268, 269, 270, 290, 291], "bbox": [0.44115624999999997, 0.22618213660245184, 0.9118437499999998, 0.6115936952714537], "iscrowd": 0, "area": 34721.86710000001, "rle": {"size": [571, 640], "counts": "f_m41[a0?B?H7J6I7I8KVOX@Fi?8m0N1O2N2O16J6J8H;D=D:F8H7I7ZCUMP:R3\\E]Mc:j2iDdMV;b2WDmMg;T4O1O1O1O1O1O2N1O1O1O1O1O1O2N1O100O1O1O1O2N1O10000000TNaE\\L_:b3eE]LZ:b3iE\\LV:c3nE[LR:d3RFYLn9e3VFYLj9e3ZFYLf9f3]FWLd9g3`FSLc9m3`FnKc9P4bFiKb9V4R2N2O1N2N1101O001O001O00O1O2O0O1O1O2O0O10001O0000001O0000000O2O00001O001O00001O001O001O001O001O00^DbLg8^3UGfLk8Z3PGkLP9U3kFPMU9o2hFUMX9k2cFZM]9f2^F_Mb9d2WFaMh9b2PFcMQ:^2hEgMX:\\2aEhM_:c2QEbMo:e40001O000000000000001O0001O00000001OVOhDQKX;k4nDSKR;i4SEVKm:g4XEWKh:e4]EZKc:b4cEZK_:b4fE[K\\:b4iE[KX:a4nE[KT:^4h1F:G9J62N1O1O2N1O1O2N1O1O2N1O1O2O1N3M4L3M3M4L3M3M3N3L7I7I7I7I2N000000000000\\L_Dn0a;ROfDg0[;XOlDa0T;_OSE:m:FZE3f:MaEL_:4hEEX:;nE_OR:a0VFWOj9i0`FmN`9S1kFbNU9^1VGWNj8j1^GmMb8S2_GlMa8T2`GkM`8U2aGjM_8V2bGiM^8W2cGhM]8X2dGfM]8Z2dGcM^8\\2b3O0000000000001O001O001O0O2O001O0O2N1O2M201O001O001O001N101O000O2N1O2N1O2O0O2N101N1O2O0O2N11O010O001O001O010O001O001O001O001O000000001O00000000000001N1000000O1O2O1N2N2N2O1N2N2N2N3K6J6J6JYco0"}, "label": "the elephant to the right of the group"}]}
{"id": 412167, "image": "COCO_train2014_000000412167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant on the right\"."}], "task": "refering", "answer_template": "The \"the elephant on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 218, 219, 221, 222, 223, 224, 241, 244, 245, 246, 247, 263, 264, 265, 267, 268, 269, 270, 290, 291], "bbox": [0.44115624999999997, 0.22618213660245184, 0.9118437499999998, 0.6115936952714537], "iscrowd": 0, "area": 34721.86710000001, "rle": {"size": [571, 640], "counts": "f_m41[a0?B?H7J6I7I8KVOX@Fi?8m0N1O2N2O16J6J8H;D=D:F8H7I7ZCUMP:R3\\E]Mc:j2iDdMV;b2WDmMg;T4O1O1O1O1O1O2N1O1O1O1O1O1O2N1O100O1O1O1O2N1O10000000TNaE\\L_:b3eE]LZ:b3iE\\LV:c3nE[LR:d3RFYLn9e3VFYLj9e3ZFYLf9f3]FWLd9g3`FSLc9m3`FnKc9P4bFiKb9V4R2N2O1N2N1101O001O001O00O1O2O0O1O1O2O0O10001O0000001O0000000O2O00001O001O00001O001O001O001O001O00^DbLg8^3UGfLk8Z3PGkLP9U3kFPMU9o2hFUMX9k2cFZM]9f2^F_Mb9d2WFaMh9b2PFcMQ:^2hEgMX:\\2aEhM_:c2QEbMo:e40001O000000000000001O0001O00000001OVOhDQKX;k4nDSKR;i4SEVKm:g4XEWKh:e4]EZKc:b4cEZK_:b4fE[K\\:b4iE[KX:a4nE[KT:^4h1F:G9J62N1O1O2N1O1O2N1O1O2N1O1O2O1N3M4L3M3M4L3M3M3N3L7I7I7I7I2N000000000000\\L_Dn0a;ROfDg0[;XOlDa0T;_OSE:m:FZE3f:MaEL_:4hEEX:;nE_OR:a0VFWOj9i0`FmN`9S1kFbNU9^1VGWNj8j1^GmMb8S2_GlMa8T2`GkM`8U2aGjM_8V2bGiM^8W2cGhM]8X2dGfM]8Z2dGcM^8\\2b3O0000000000001O001O001O0O2O001O0O2N1O2M201O001O001O001N101O000O2N1O2N1O2O0O2N101N1O2O0O2N11O010O001O001O010O001O001O001O001O000000001O00000000000001N1000000O1O2O1N2N2N2O1N2N2N2N3K6J6J6JYco0"}, "label": "the elephant on the right"}]}
{"id": 518664, "image": "COCO_train2014_000000518664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating from a tall food container with two giraffes in the background\"."}], "task": "refering", "answer_template": "The \"a giraffe eating from a tall food container with two giraffes in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 58, 59, 60, 80, 81, 82, 103, 104, 126, 127, 148, 149, 150, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 259, 260, 261, 263, 264, 282, 283, 284, 286, 287, 305, 306, 307, 309, 310, 329, 330, 332, 333, 352, 353, 354, 355, 356, 378, 379], "bbox": [0.294015625, 0.091875, 0.6234375, 0.9580625], "iscrowd": 0, "area": 28271.943849999978, "rle": {"size": [480, 640], "counts": "k_h28l=j0]Ob02N3N1N3MbDiNX9V1bFROe8d1TGgNR8o1eG]NY8b1_GiN^8X1YGSOe8l0RG@l8?lFLQ95eF8X9i2M3N2N2ML5H8G8I8H9F:G:F99H:E;F:E<E:E;F9F2N101N10mMeGaMZ8P2_HhM`7i1XInMg6e1PJSNo5l1[JkMd5U2fJcMY5\\2QK[Mn4d2]KTMb4k2i301O1O1O0O2O1O1O001N2O1O001O1N2O10O1000000O1000000O2O00000O1000000O1000000O1RHXMc3j2VL]Mh3d2RLbMm3_2mKhMQ4Y2iKmMV4T2dKSNZ4n1`KYN^4h1[K_Nd4b1VKeNh4\\1RKjNm4W1mJPOP5R1jJTOU5m0eJZOY5g0`JA^5a0[JEd5<SJNk53fI<Y6EXIk0f6h3O2N1OM3H7H9H8H8G9H8H8G9H8H8G9H7^2bMT1lN2N2N2M4M2N2N2NmM]JmJa5o4mJiJQ5S5^KdJ`4X5nK`Jj3b5dLWJS3k5[MmI]2U6QNcIg1_6gNYIQ1i6^32N2O32N2O1N2N2NkNYIWJe6n5\\InIb6V6_IgI_6]6bI`IT5VOoK]7nNZIP5CiKX7XOQIm4OcKT7AjHk49\\KQ7JcHi4c0UKn63\\Hh4o8YKmFf4V9[KgFe4[9\\KbFc4a9<3L3N2M4M2N2M2O1N3N1N3N1ZOg0POo0QOo0F;O2N100O1O1O2N1O1O1O2N100O1O2N1O1O1O1O2N1O100O2N1O1O1O2N1O1O1O101N1O1O1O2N1O\\m`3"}, "label": "a giraffe eating from a tall food container with two giraffes in the background"}]}
{"id": 518664, "image": "COCO_train2014_000000518664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe stands away from another one near a rope fence\"."}], "task": "refering", "answer_template": "The \"a tall giraffe stands away from another one near a rope fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 58, 59, 60, 80, 81, 82, 103, 104, 126, 127, 148, 149, 150, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 259, 260, 261, 263, 264, 282, 283, 284, 286, 287, 305, 306, 307, 309, 310, 329, 330, 332, 333, 352, 353, 354, 355, 356, 378, 379], "bbox": [0.294015625, 0.091875, 0.6234375, 0.9580625], "iscrowd": 0, "area": 28271.943849999978, "rle": {"size": [480, 640], "counts": "k_h28l=j0]Ob02N3N1N3MbDiNX9V1bFROe8d1TGgNR8o1eG]NY8b1_GiN^8X1YGSOe8l0RG@l8?lFLQ95eF8X9i2M3N2N2ML5H8G8I8H9F:G:F99H:E;F:E<E:E;F9F2N101N10mMeGaMZ8P2_HhM`7i1XInMg6e1PJSNo5l1[JkMd5U2fJcMY5\\2QK[Mn4d2]KTMb4k2i301O1O1O0O2O1O1O001N2O1O001O1N2O10O1000000O1000000O2O00000O1000000O1000000O1RHXMc3j2VL]Mh3d2RLbMm3_2mKhMQ4Y2iKmMV4T2dKSNZ4n1`KYN^4h1[K_Nd4b1VKeNh4\\1RKjNm4W1mJPOP5R1jJTOU5m0eJZOY5g0`JA^5a0[JEd5<SJNk53fI<Y6EXIk0f6h3O2N1OM3H7H9H8H8G9H8H8G9H8H8G9H7^2bMT1lN2N2N2M4M2N2N2NmM]JmJa5o4mJiJQ5S5^KdJ`4X5nK`Jj3b5dLWJS3k5[MmI]2U6QNcIg1_6gNYIQ1i6^32N2O32N2O1N2N2NkNYIWJe6n5\\InIb6V6_IgI_6]6bI`IT5VOoK]7nNZIP5CiKX7XOQIm4OcKT7AjHk49\\KQ7JcHi4c0UKn63\\Hh4o8YKmFf4V9[KgFe4[9\\KbFc4a9<3L3N2M4M2N2M2O1N3N1N3N1ZOg0POo0QOo0F;O2N100O1O1O2N1O1O1O2N100O1O2N1O1O1O1O2N1O100O2N1O1O1O2N1O1O1O101N1O1O1O2N1O\\m`3"}, "label": "a tall giraffe stands away from another one near a rope fence"}]}
{"id": 420363, "image": "COCO_train2014_000000420363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flat box full of plants labeled wegman ' s nursery\"."}], "task": "refering", "answer_template": "The \"a flat box full of plants labeled wegman ' s nursery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 134, 135, 136, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173], "bbox": [0.46396, 0.28540540540540543, 0.94138, 0.7865165165165166], "iscrowd": 0, "area": 27976.069350000005, "rle": {"size": [333, 500], "counts": "Yc[23V:5K6K5J5L5L7H:F9H2M3iH]Na5e1\\JeN\\5]1_JhN^5[1_JhN^5[1`JgN^5[1`JgN]5[1bJgN[5\\1cJgNZ5[1dJgNY5\\1fJeNW5^1gJeNU5]1jJfNR5]1lJfNQ5\\1mJgNS5X1kJkNT5U1jJnNV5Q1hJROX5l0gJWOX5i0fJZOZ5e0dJ^O\\5a0bJB]5h2000O001O1O001O1O0010O0001O001O001O00001O001O001N100000K401N2O001000O1000000O10000O1000000O10000O11O000b0^O5K01N1O1O101N001O1O1@?0100O010O100O010O1O01000000O1000000O10000O101O000O1000000O10000O10001O0O10000O1000000O10000O1000000O101O000O10000O1000000O10000O10000000000000004L7I8eL`IY2l6UMZIk2T7000001O00000000001O001O1O1O001O1O1O1O0O2O1O1O0O2O1O1O0O2O1O1N2O001O1N2O1O0O2O1O1O1N2O001O1O1O1O1O1O001O1O1O100O001O1O1O001O1O1M2J7I7Ia0_OdW9"}, "label": "a flat box full of plants labeled wegman ' s nursery"}]}
{"id": 420363, "image": "COCO_train2014_000000420363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a set of flowers\"."}], "task": "refering", "answer_template": "The \"a set of flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 134, 135, 136, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173], "bbox": [0.46396, 0.28540540540540543, 0.94138, 0.7865165165165166], "iscrowd": 0, "area": 27976.069350000005, "rle": {"size": [333, 500], "counts": "Yc[23V:5K6K5J5L5L7H:F9H2M3iH]Na5e1\\JeN\\5]1_JhN^5[1_JhN^5[1`JgN^5[1`JgN]5[1bJgN[5\\1cJgNZ5[1dJgNY5\\1fJeNW5^1gJeNU5]1jJfNR5]1lJfNQ5\\1mJgNS5X1kJkNT5U1jJnNV5Q1hJROX5l0gJWOX5i0fJZOZ5e0dJ^O\\5a0bJB]5h2000O001O1O001O1O0010O0001O001O001O00001O001O001N100000K401N2O001000O1000000O10000O1000000O10000O11O000b0^O5K01N1O1O101N001O1O1@?0100O010O100O010O1O01000000O1000000O10000O101O000O1000000O10000O10001O0O10000O1000000O10000O1000000O101O000O10000O1000000O10000O10000000000000004L7I8eL`IY2l6UMZIk2T7000001O00000000001O001O1O1O001O1O1O1O0O2O1O1O0O2O1O1O0O2O1O1N2O001O1N2O1O0O2O1O1O1N2O001O1O1O1O1O1O001O1O1O100O001O1O1O001O1O1M2J7I7Ia0_OdW9"}, "label": "a set of flowers"}]}
{"id": 420363, "image": "COCO_train2014_000000420363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pink flowers in a green pot\"."}], "task": "refering", "answer_template": "The \"pink flowers in a green pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 39, 40, 41, 42, 43, 44, 56, 57, 58, 59, 61, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 111, 112, 113, 114, 115, 116, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 202, 203, 204, 205], "bbox": [0.1347, 0.07639639639639641, 0.55076, 1.0], "iscrowd": 0, "area": 30642.073699999997, "rle": {"size": [333, 500], "counts": "Rme01_9m001O000O1000001HPOPGQ1P9POoFP1P9QOPGo0P9ROoFn0Q9SOnFn0P9SOoFn0Q97O100O101N100O100XHjNY6V1hIZOf5h0[J@Z5a0gJ@V5a0lJ_OR5a0oJ@n4a0TK@h4b0XK_Oe4b0]K^O`4c0aK_O[4b0gK^OW4b0jK_OS4b0oK^On3d0RL]Om3b0UL]Oj3c0WL]Oh3c0ZL\\Of3c0[L]Od3d0\\L\\Oc3d0_L[Oa3f0]L[Ob3j0WLYOh3k0RLXOm3m0lKVOT4m0fKVOY4o0aKSO]4R1aKmN]4Y1_KgN_4_1^K`Na4d1\\K\\Nb4j1ZKVN[4`N_K_34PN[4eN]Ka34jM^4c2`K\\M_4Z4PLXJY3i5eL[JX3f5gL^JU3c5jL`JR3b5lLcJP3^5oLfJn0ZOd0Q6]NhJj0YOi0P6[NlJf0WOn0n5[NnJa0WOS1m5[NPK<UOY1k5[NSK7TO]1k5ZNVK3QOb1k5ZNUK0SOf1i5YNkKe1_6O100O15Kc0]O2N000000000001n0QO3M3M3M3M3M3M3M3M3M3M2N2NN2N2N2N2N2O1001O00001O00001OO1O100O1O1TMhIb1Y6^NiITONo1Y6lNlIQOOP2V6nNnImN0S2S6PO[Jn0e5QO^Jl0c5TO^Jj0c5UO`Jj0^5WOdJh0[5WOgJi0X5WOjJh0U5WOnJh0Q5XOPKh0o4WOTKh0k4XOWKg0h4YO[Ke0d4[O^Kd0a4[OcKb0]4^OeKa0Z4_OiKGWNNo5;lKDXN0k5;QLBVN2h5<TL_OWN4d5=XL[OWNIn5k0mIUOe14S4JkJm0\\OYOf1LS4OkJl0[O\\Og1GS41kJl0[O_Oe7EQIk0ZODa7BSIl0[OE`7_OTIm0[OG^7]OVIl0\\OH]7]OVIl0\\OH^7\\OUIm0\\OHg01R5[OiJn0]OGi00o4\\OjJm0_OGi00l4]OkJm0_OGk0Oj4]OkJn0_OGm0Of4\\OmJn1>fNc4_OlJl1b0eNa4BiJj1e0fNS1TOY1?lLh1e0iNR1ROg0NoMc0^Of1f0mNR1lNm0NmMo21\\NQ1iNQ1MlMo2O_NQ1fNU1LjMQ3O^Nn0gNY1LhMT3MZNo0gN\\1LhMW3IWNo0hNa1JfMY3IVNl0hNf1JdMZ3ITNk0iNh1KcMY3ISNi0kNI7d0BPO[3IQNh0lNC>l0[OoN\\3IoMf0nN]Oe0V1ROnN]3HnMe0=f0ZNkN]3IlMc0?i0YNjN^3IjMa0a0>kM^O<Il7j0hG\\O=Jj7j0iG]O>Gh7m0kG[O>Gg75kG>1E>He75WH2H1<Ge76\\HKE8<Ec77\\HLG7;Fa77\\HLJ7]8NiGJL7[8NiGJN9X8MiGJ18V8NhGJ47T8NhGJ78o7OaH1_7ObH0^7OdH1[7OfH0Z70gH0X7OiH1W7OjH0V70kH0U7NmH1T7NlH3T7LmH3g8O10O01O100O010O1O10O0]XY2"}, "label": "pink flowers in a green pot"}]}
{"id": 420363, "image": "COCO_train2014_000000420363.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flowers planted in pot\"."}], "task": "refering", "answer_template": "The \"flowers planted in pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 39, 40, 41, 42, 43, 44, 56, 57, 58, 59, 61, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 111, 112, 113, 114, 115, 116, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 202, 203, 204, 205], "bbox": [0.1347, 0.07639639639639641, 0.55076, 1.0], "iscrowd": 0, "area": 30642.073699999997, "rle": {"size": [333, 500], "counts": "Rme01_9m001O000O1000001HPOPGQ1P9POoFP1P9QOPGo0P9ROoFn0Q9SOnFn0P9SOoFn0Q97O100O101N100O100XHjNY6V1hIZOf5h0[J@Z5a0gJ@V5a0lJ_OR5a0oJ@n4a0TK@h4b0XK_Oe4b0]K^O`4c0aK_O[4b0gK^OW4b0jK_OS4b0oK^On3d0RL]Om3b0UL]Oj3c0WL]Oh3c0ZL\\Of3c0[L]Od3d0\\L\\Oc3d0_L[Oa3f0]L[Ob3j0WLYOh3k0RLXOm3m0lKVOT4m0fKVOY4o0aKSO]4R1aKmN]4Y1_KgN_4_1^K`Na4d1\\K\\Nb4j1ZKVN[4`N_K_34PN[4eN]Ka34jM^4c2`K\\M_4Z4PLXJY3i5eL[JX3f5gL^JU3c5jL`JR3b5lLcJP3^5oLfJn0ZOd0Q6]NhJj0YOi0P6[NlJf0WOn0n5[NnJa0WOS1m5[NPK<UOY1k5[NSK7TO]1k5ZNVK3QOb1k5ZNUK0SOf1i5YNkKe1_6O100O15Kc0]O2N000000000001n0QO3M3M3M3M3M3M3M3M3M3M2N2NN2N2N2N2N2O1001O00001O00001OO1O100O1O1TMhIb1Y6^NiITONo1Y6lNlIQOOP2V6nNnImN0S2S6PO[Jn0e5QO^Jl0c5TO^Jj0c5UO`Jj0^5WOdJh0[5WOgJi0X5WOjJh0U5WOnJh0Q5XOPKh0o4WOTKh0k4XOWKg0h4YO[Ke0d4[O^Kd0a4[OcKb0]4^OeKa0Z4_OiKGWNNo5;lKDXN0k5;QLBVN2h5<TL_OWN4d5=XL[OWNIn5k0mIUOe14S4JkJm0\\OYOf1LS4OkJl0[O\\Og1GS41kJl0[O_Oe7EQIk0ZODa7BSIl0[OE`7_OTIm0[OG^7]OVIl0\\OH]7]OVIl0\\OH^7\\OUIm0\\OHg01R5[OiJn0]OGi00o4\\OjJm0_OGi00l4]OkJm0_OGk0Oj4]OkJn0_OGm0Of4\\OmJn1>fNc4_OlJl1b0eNa4BiJj1e0fNS1TOY1?lLh1e0iNR1ROg0NoMc0^Of1f0mNR1lNm0NmMo21\\NQ1iNQ1MlMo2O_NQ1fNU1LjMQ3O^Nn0gNY1LhMT3MZNo0gN\\1LhMW3IWNo0hNa1JfMY3IVNl0hNf1JdMZ3ITNk0iNh1KcMY3ISNi0kNI7d0BPO[3IQNh0lNC>l0[OoN\\3IoMf0nN]Oe0V1ROnN]3HnMe0=f0ZNkN]3IlMc0?i0YNjN^3IjMa0a0>kM^O<Il7j0hG\\O=Jj7j0iG]O>Gh7m0kG[O>Gg75kG>1E>He75WH2H1<Ge76\\HKE8<Ec77\\HLG7;Fa77\\HLJ7]8NiGJL7[8NiGJN9X8MiGJ18V8NhGJ47T8NhGJ78o7OaH1_7ObH0^7OdH1[7OfH0Z70gH0X7OiH1W7OjH0V70kH0U7NmH1T7NlH3T7LmH3g8O10O01O100O010O1O10O0]XY2"}, "label": "flowers planted in pot"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a remote on the far right\"."}], "task": "refering", "answer_template": "The \"a remote on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 366, 367], "bbox": [0.753890625, 0.16609561752988047, 1.0, 0.8889043824701196], "iscrowd": 0, "area": 32650.770950000002, "rle": {"size": [502, 640], "counts": "o[\\71e?2N2N2N2N2N2N2N2N2N2N4L4L5K4L5K5K4L5K4L5K4L5K5K4L5K4L4L3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M4N10001N10001O00001OO1M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3L4M3M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3MQH"}, "label": "a remote on the far right"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the remote control to the left of the gray cox one\"."}], "task": "refering", "answer_template": "The \"the remote control to the left of the gray cox one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304], "bbox": [0.003015625, 0.14149402390438248, 0.26303125, 0.7751593625498008], "iscrowd": 0, "area": 51169.957350000004, "rle": {"size": [502, 640], "counts": "_W1R4b8R3TOl0TOl0TOl0L400000000000O1000000000000000000000000000000O1000000000000000000000000000000O0100000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O0100000000000000000000000000000O1000000000000000000000000000000O1000000000000000`2`Mi2WMPUW7"}, "label": "the remote control to the left of the gray cox one"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the remote control on the far left with a big circular white play button\"."}], "task": "refering", "answer_template": "The \"the remote control on the far left with a big circular white play button\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304], "bbox": [0.003015625, 0.14149402390438248, 0.26303125, 0.7751593625498008], "iscrowd": 0, "area": 51169.957350000004, "rle": {"size": [502, 640], "counts": "_W1R4b8R3TOl0TOl0TOl0L400000000000O1000000000000000000000000000000O1000000000000000000000000000000O0100000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O0100000000000000000000000000000O1000000000000000000000000000000O1000000000000000`2`Mi2WMPUW7"}, "label": "the remote control on the far left with a big circular white play button"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver cox cable remote\"."}], "task": "refering", "answer_template": "The \"a silver cox cable remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 335, 352, 353, 354, 355, 356, 357, 358], "bbox": [0.255578125, 0.1393227091633466, 0.609875, 0.8808964143426294], "iscrowd": 0, "area": 67117.59315, "rle": {"size": [502, 640], "counts": "me`27Z?9H8H8H7I8G9H8H8H8G9H7I8H8G9H8H8H8G8I8H8H8fKeJmMc5o1PKYMY5c2ZKfLn4V3eKSLb4k3PL^KX4^4[LjJn3R5_3L4L4L3M4L4L4M3L00001O00000001O00000000000000001O000000000000001O000001O00000001O00000000000000001O000000000000001O00000001O00000001O000000000000001O000000000000001O00000001O00000001O000000000000001O00000000000000001O000001O00000001O000000000000001O00000000000000001O000001O00000001O00000000000000001O000000001O7I8H8H7I8H7I8I6I8H8H7I8H7I8H7I8H8H7I8H7I7I6J6J6I7G9G9F:G9G9F:G9G9F:G9G9G9F:G9G9F:G9GYVj3"}, "label": "a silver cox cable remote"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver remote control with a black buttons and curved sides , sitting between two black remotes\"."}], "task": "refering", "answer_template": "The \"a silver remote control with a black buttons and curved sides , sitting between two black remotes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 335, 352, 353, 354, 355, 356, 357, 358], "bbox": [0.255578125, 0.1393227091633466, 0.609875, 0.8808964143426294], "iscrowd": 0, "area": 67117.59315, "rle": {"size": [502, 640], "counts": "me`27Z?9H8H8H7I8G9H8H8H8G9H7I8H8G9H8H8H8G8I8H8H8fKeJmMc5o1PKYMY5c2ZKfLn4V3eKSLb4k3PL^KX4^4[LjJn3R5_3L4L4L3M4L4L4M3L00001O00000001O00000000000000001O000000000000001O000001O00000001O00000000000000001O000000000000001O00000001O00000001O000000000000001O000000000000001O00000001O00000001O000000000000001O00000000000000001O000001O00000001O000000000000001O00000000000000001O000001O00000001O00000000000000001O000000001O7I8H8H7I8H7I8I6I8H8H7I8H7I8H7I8H8H7I8H7I7I6J6J6I7G9G9F:G9G9F:G9G9F:G9G9G9F:G9G9F:G9GYVj3"}, "label": "a silver remote control with a black buttons and curved sides , sitting between two black remotes"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"samsung remote control\"."}], "task": "refering", "answer_template": "The \"samsung remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 359, 360, 361, 362, 363, 364, 365, 383, 384, 385], "bbox": [0.55271875, 0.12163346613545817, 0.9042656250000001, 0.9014342629482072], "iscrowd": 0, "area": 58087.32509999999, "rle": {"size": [502, 640], "counts": "Yc]5h0n>R1nNR1nNS1mN7I5K4L4L5K4L5K4L5K4L5K4L5K4L4L5K4L5L3L5K4L5K4L5K4L4L5K4L5K4L5K4L5K4L5K4L4L5K4L5K4L5K4L5K4L4L5K4L5K4L4L1O1O2N1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1OO1000000000000000000O10000000000000000O10001O000000003M3M3L4M3M3M4L3M3M3M3M3L4M3M4L3M3M3M3M3M3L4M3M4L3M3M3M3M3M3M3L4M4L3M3M3M3M3M3M3L4M4L3M3M3M3M3M3M3M3L5L3M3M3M3M3M3M3M4K4M3M3M3M3M3M3M3M4L3L4M3M3M3M3M3M4L4L4K5L4L5K4L4L4L4L4L4L4K5L5K4L4L4L4L4L4L4K6K4L4L4L4L4L4L4L5K4K5L4LY_m0"}, "label": "samsung remote control"}]}
{"id": 420366, "image": "COCO_train2014_000000420366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"samsung remote\"."}], "task": "refering", "answer_template": "The \"samsung remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 359, 360, 361, 362, 363, 364, 365, 383, 384, 385], "bbox": [0.55271875, 0.12163346613545817, 0.9042656250000001, 0.9014342629482072], "iscrowd": 0, "area": 58087.32509999999, "rle": {"size": [502, 640], "counts": "Yc]5h0n>R1nNR1nNS1mN7I5K4L4L5K4L5K4L5K4L5K4L5K4L4L5K4L5L3L5K4L5K4L5K4L4L5K4L5K4L5K4L5K4L5K4L4L5K4L5K4L5K4L5K4L4L5K4L5K4L4L1O1O2N1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1OO1000000000000000000O10000000000000000O10001O000000003M3M3L4M3M3M4L3M3M3M3M3L4M3M4L3M3M3M3M3M3L4M3M4L3M3M3M3M3M3M3L4M4L3M3M3M3M3M3M3L4M4L3M3M3M3M3M3M3M3L5L3M3M3M3M3M3M3M4K4M3M3M3M3M3M3M3M4L3L4M3M3M3M3M3M4L4L4K5L4L5K4L4L4L4L4L4L4K5L5K4L4L4L4L4L4L4K6K4L4L4L4L4L4L4L5K4K5L4LY_m0"}, "label": "samsung remote"}]}
{"id": 248337, "image": "COCO_train2014_000000248337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an end section of a table with a red square and a purple border\"."}], "task": "refering", "answer_template": "The \"an end section of a table with a red square and a purple border\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.28526562499999997, 0.5506572769953052, 0.864453125, 0.8467605633802817], "iscrowd": 0, "area": 38054.25780000001, "rle": {"size": [426, 640], "counts": "f\\\\2>_<?@`0J5O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2M3N2N1O2N2N2N2N1O2N2N2N2N2N00000000000000O1000000000000000000000000000000000O1000000000O100000000000000000000000000000000000000000000O100000000000000000O1000000000000000000000000000O100000000000000000000000000000000000000000000O01000000000000000000000000000000000000000000000O1000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000O1000000000O1000000000000000000000000000000000O1000000000000000000000000000000000001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O0O2O001O00001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O0O2O0UOWEUOj:j0k0O2O0O2O0O2O0O2O0O2O3L8I7HclS1"}, "label": "an end section of a table with a red square and a purple border"}]}
{"id": 248337, "image": "COCO_train2014_000000248337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple and red desk with two bears sitting at it\"."}], "task": "refering", "answer_template": "The \"a purple and red desk with two bears sitting at it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.28526562499999997, 0.5506572769953052, 0.864453125, 0.8467605633802817], "iscrowd": 0, "area": 38054.25780000001, "rle": {"size": [426, 640], "counts": "f\\\\2>_<?@`0J5O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2M3N2N1O2N2N2N2N1O2N2N2N2N2N00000000000000O1000000000000000000000000000000000O1000000000O100000000000000000000000000000000000000000000O100000000000000000O1000000000000000000000000000O100000000000000000000000000000000000000000000O01000000000000000000000000000000000000000000000O1000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000O1000000000O1000000000000000000000000000000000O1000000000000000000000000000000000001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O0O2O001O00001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O0O2O0UOWEUOj:j0k0O2O0O2O0O2O0O2O0O2O3L8I7HclS1"}, "label": "a purple and red desk with two bears sitting at it"}]}
{"id": 207381, "image": "COCO_train2014_000000207381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue and yellow train\"."}], "task": "refering", "answer_template": "The \"the blue and yellow train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 311, 312], "bbox": [0.14028125, 0.06068235294117647, 0.775984375, 0.9236000000000001], "iscrowd": 0, "area": 103268.09690000003, "rle": {"size": [425, 640], "counts": "_aU1`0X<f0ZOh0UEUN[9\\2aFhM]9Y2`FjM^9X2_FlM^9V2_FlM`9W2\\FlMc9U2ZFmMe9l2O2O0O1O2O0O1O2N100O2N100O3M2O2M3M3M2O2M2N2O2M2N3N1N2N2N3N1N2N2O2M2N100O001O10O03M2O1N2N2O1N2N2N100O1O100O1O100O2N2N101N2N2O0O2N2O1N1O2N2O1N2N2O1N2N101N1O2N101N1O2O0O2N101N1O3M2O1N3M2O1N3M2O1N2N1O2O1N1O2O1N2N101N2O0O2O1N2N101N2O0O2N2O1N101N2O1N1O2O1N102M3M2O2M3N2M2N101N101N101N1O2O1N101N1O2O0O2O0O2O0O2N101N101N1O2O0O2O001N101O0O2O001N101O0O2O001N101O001N101O0O2O001N101O0O2O000O10N2N1O2N2N1O2N2N2O0O2N2O1N2O1N2N2O1N2000000000000000000001O00000000000000000000001O00001O0000001O00001O00001O00001O001O001O001O00001O001O001O00001O001O010O00001O0010O0K5K5K5E;F:H9N100O1O100O101N1O100O100O2N1000000000001N100YGlKY8T4cGPL]8P4`GTL_8l3]GYLb8W401O001O001O1O001O001O001N101O1O001O001O001O1O001O001O001O1N101O001O001O001O1O001O001D;ZOg0H8N1O2O0O2N1O2N1O2O1N1O2N5K=C=C>C<C>C6K0O2O00000O;F1cNQF\\OP:c0TFZOk9f0YFVOg9j0YFWOg9DnE>;Ng9]OYFa0O3h9UOdFd0D8j:CUE>o:^OQEc0R;YOmDh0W;ROkDn0Y;mNgDU1g;O3N2M3N3K4J6K5K5K5KnRk1"}, "label": "the blue and yellow train"}]}
{"id": 207381, "image": "COCO_train2014_000000207381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue train with a yellow facing\"."}], "task": "refering", "answer_template": "The \"a blue train with a yellow facing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 311, 312], "bbox": [0.14028125, 0.06068235294117647, 0.775984375, 0.9236000000000001], "iscrowd": 0, "area": 103268.09690000003, "rle": {"size": [425, 640], "counts": "_aU1`0X<f0ZOh0UEUN[9\\2aFhM]9Y2`FjM^9X2_FlM^9V2_FlM`9W2\\FlMc9U2ZFmMe9l2O2O0O1O2O0O1O2N100O2N100O3M2O2M3M3M2O2M2N2O2M2N3N1N2N2N3N1N2N2O2M2N100O001O10O03M2O1N2N2O1N2N2N100O1O100O1O100O2N2N101N2N2O0O2N2O1N1O2N2O1N2N2O1N2N101N1O2N101N1O2O0O2N101N1O3M2O1N3M2O1N3M2O1N2N1O2O1N1O2O1N2N101N2O0O2O1N2N101N2O0O2N2O1N101N2O1N1O2O1N102M3M2O2M3N2M2N101N101N101N1O2O1N101N1O2O0O2O0O2O0O2N101N101N1O2O0O2O001N101O0O2O001N101O0O2O001N101O001N101O0O2O001N101O0O2O000O10N2N1O2N2N1O2N2N2O0O2N2O1N2O1N2N2O1N2000000000000000000001O00000000000000000000001O00001O0000001O00001O00001O00001O001O001O001O00001O001O001O00001O001O010O00001O0010O0K5K5K5E;F:H9N100O1O100O101N1O100O100O2N1000000000001N100YGlKY8T4cGPL]8P4`GTL_8l3]GYLb8W401O001O001O1O001O001O001N101O1O001O001O001O1O001O001O001O1N101O001O001O001O1O001O001D;ZOg0H8N1O2O0O2N1O2N1O2O1N1O2N5K=C=C>C<C>C6K0O2O00000O;F1cNQF\\OP:c0TFZOk9f0YFVOg9j0YFWOg9DnE>;Ng9]OYFa0O3h9UOdFd0D8j:CUE>o:^OQEc0R;YOmDh0W;ROkDn0Y;mNgDU1g;O3N2M3N3K4J6K5K5K5KnRk1"}, "label": "a blue train with a yellow facing"}]}
{"id": 207381, "image": "COCO_train2014_000000207381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train to the far right you cant see all of\"."}], "task": "refering", "answer_template": "The \"the train to the far right you cant see all of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 272, 273, 274, 275], "bbox": [0.723484375, 0.09865882352941176, 0.9730624999999999, 0.7608470588235294], "iscrowd": 0, "area": 26642.216549999986, "rle": {"size": [425, 640], "counts": "kWP61X=8Ga0@`0@`0@6J2M2O2N2N2N1O2M3N2N1O2N5J2O1VF[Mn8f2eFiMX9X3L4L4K5L3M6J8H7I7H8lH^JS6i5`I_J^6X6WJ`Ie4c6RKcIn4e6dJaI[5W70000000D<[IQJZ5d6bJaIg4V7mJVIU3^9`MPU=@VmBj0UOk0VOj0[Ke4XOh01OO10000O100O10000O10000O10O10O10000O10000O10000O10000O10000O10000O10O10O10000O10000O100O10000O10000O10000O1000O010000O10000O100`0@001O0000001O0000001TOk0ZNf1ZNf1[Nf1YNaQ7"}, "label": "the train to the far right you cant see all of"}]}
{"id": 207381, "image": "COCO_train2014_000000207381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white train in background\"."}], "task": "refering", "answer_template": "The \"white train in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 272, 273, 274, 275], "bbox": [0.723484375, 0.09865882352941176, 0.9730624999999999, 0.7608470588235294], "iscrowd": 0, "area": 26642.216549999986, "rle": {"size": [425, 640], "counts": "kWP61X=8Ga0@`0@`0@6J2M2O2N2N2N1O2M3N2N1O2N5J2O1VF[Mn8f2eFiMX9X3L4L4K5L3M6J8H7I7H8lH^JS6i5`I_J^6X6WJ`Ie4c6RKcIn4e6dJaI[5W70000000D<[IQJZ5d6bJaIg4V7mJVIU3^9`MPU=@VmBj0UOk0VOj0[Ke4XOh01OO10000O100O10000O10000O10O10O10000O10000O10000O10000O10000O10000O10O10O10000O10000O100O10000O10000O10000O1000O010000O10000O100`0@001O0000001O0000001TOk0ZNf1ZNf1[Nf1YNaQ7"}, "label": "white train in background"}]}
{"id": 92694, "image": "COCO_train2014_000000092694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with short hair and a grey shirt holding up a wii remote looking to the side\"."}], "task": "refering", "answer_template": "The \"a man with short hair and a grey shirt holding up a wii remote looking to the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.011796875, 0.2876458333333333, 0.57303125, 0.9887708333333334], "iscrowd": 0, "area": 68788.20679999999, "rle": {"size": [480, 640], "counts": "^T46h>7I7I7H8I6J7I7I7H8I7I6J7I7H8I2N2N2N2M3N2N2N2N2M3N2N2N2O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O00100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1gLkJfLV5W3RKcLo4Y3ZKbLf4Z3dK_L]4^3kK\\LV4`3SLZLn3b3[LYLe3f3aLTL`3k3fLoK[3P4lLiKU3V4QMeKo2[4XM]Ki2b4^MWKc2h4dMQK]2n4jMlJV2S5RNeJo1Z5l3O100O001O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O2N1O1O1O11O1O000000000000000000O1000000000000O1000001O00000O100000000000000O1000000000000O1000000000000O101O00000000000O10000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1On0RO4L4L4L4L4L2N1O1O001O1O1TJRId3o6QKYI2Ml4j6nJ`JR5a5mJ`JR5a5lJaJS5`5kJaJU5`5jJaJU5`5iJbJV5^5iJcJX5]5fJeJY5\\5fJeJY5\\5eJeJZ5]5dJeJ[5\\5dJeJZ5\\5eJeJZ5]5dJeJ[5\\5cJfJ[5`701N2N101N2O0O2O0O2N2O0O2O1N1O2O1N101N2N2O0O2O1N2N2O0O2O1N2N2O0O2O1N2N101N2O1N2N101N2O1N2O0O2N2O1N2O0O2N2O1O00100O1O1O001O1O100O001O1O100O010O1O10O0100O100O00100O100O010O1O100O010O1O001O1O001O001O1O001O1O001O1N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2M3L4M7I6J7I6I8I6JSbo3"}, "label": "a man with short hair and a grey shirt holding up a wii remote looking to the side"}]}
{"id": 92694, "image": "COCO_train2014_000000092694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a grey shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.011796875, 0.2876458333333333, 0.57303125, 0.9887708333333334], "iscrowd": 0, "area": 68788.20679999999, "rle": {"size": [480, 640], "counts": "^T46h>7I7I7H8I6J7I7I7H8I7I6J7I7H8I2N2N2N2M3N2N2N2N2M3N2N2N2O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O00100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1gLkJfLV5W3RKcLo4Y3ZKbLf4Z3dK_L]4^3kK\\LV4`3SLZLn3b3[LYLe3f3aLTL`3k3fLoK[3P4lLiKU3V4QMeKo2[4XM]Ki2b4^MWKc2h4dMQK]2n4jMlJV2S5RNeJo1Z5l3O100O001O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O2N1O1O1O11O1O000000000000000000O1000000000000O1000001O00000O100000000000000O1000000000000O1000000000000O101O00000000000O10000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1On0RO4L4L4L4L4L2N1O1O001O1O1TJRId3o6QKYI2Ml4j6nJ`JR5a5mJ`JR5a5lJaJS5`5kJaJU5`5jJaJU5`5iJbJV5^5iJcJX5]5fJeJY5\\5fJeJY5\\5eJeJZ5]5dJeJ[5\\5dJeJZ5\\5eJeJZ5]5dJeJ[5\\5cJfJ[5`701N2N101N2O0O2O0O2N2O0O2O1N1O2O1N101N2N2O0O2O1N2N2O0O2O1N2N2O0O2O1N2N101N2O1N2N101N2O1N2O0O2N2O1N2O0O2N2O1O00100O1O1O001O1O100O001O1O100O010O1O10O0100O100O00100O100O010O1O100O010O1O001O1O001O001O1O001O1O001O1N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2M3L4M7I6J7I6I8I6JSbo3"}, "label": "a man wearing a grey shirt"}]}
{"id": 92694, "image": "COCO_train2014_000000092694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in blue jeans and a red shirt\"."}], "task": "refering", "answer_template": "The \"the man in blue jeans and a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 196, 197, 198, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343, 359, 360, 361, 362, 363, 364, 365, 366, 384, 385, 386, 387, 388, 389], "bbox": [0.50534375, 0.4237916666666666, 0.9625625, 0.9719166666666667], "iscrowd": 0, "area": 41440.49110000001, "rle": {"size": [480, 640], "counts": "\\gg41i>8G8I8H7H9L3N3M2N3M2M3N3M2N3M2M4M2N3M2N3L3XN^MoFc2o8gMhFZ2W9nMaFS2^9VNlEX2R:\\1O001N2O1O0O2O1O1O001O1O1O00100000O1000000001O0000001O00001O000000101N3M3M2O2M2NQ1oN010O1O1O100O1O1O100O001O010O1O0010O012M2O2M2O1N3N1NZ1gNO0O2O001O001O001N10001O001O0O101O001O00001N101O00001O001N10001O001O000O2O001O00001O0O2O001O1O2N1O1N2O2N1O1O1O2N1N2O1OoJ]Gg3a8YLaGf3^8YLdGf3\\8ZLeGe3Z8[LhGd3X8\\LjGb3V8^LkGa3U8^LmGa3S8_LoG_3Q8`LQH_3n7bLSH\\3n7cLTH\\3l7dLVHZ3j7eLXHZ3h7fLYHY3g7fL[HY3d7hL^HV3b7jL_HU3a7jLaHU3_7kLbHT3^7kLeHS3Z7nLgHQ3Y7nLiHQ3W7oLjHP3V7oLmHo2S7QMnHn2R7QMPIn2o6SMSIk2m6UMTIj2l6UMVIj2j6VMWIi2i6VMZIh2f6XM[Ig2d6YM^If2b6ZM_Ie2a6ZMbId2^6\\McIc2]6\\MeIc2[6]MfIb2Y6_MiI_2W6`MkI_2U6aMlI^2T6aMoI]2Q6cMPJ\\2o5dMSJ[2m5eMTJZ2l5eMWJY2i5gMXJX2h5gM\\JV2d5jM^JT2a5mMbJP2^5oMeJo1[5QNhJl1X5SNjJl1V5TNlJj1T5UNnJj1Q5WNoJi1Q5VNPKj1P5VNPKj1P5UNQKk1o4UNQKk1o4UNQKk1n4UNSKk1m4UNSKk1m4TNTKl1l4TNTKl1l4SNUKm1k4SNUKm1j4SNWKm1i4SNWKm1i4RNXKn1h4RNXKn1h4RNXKn1g4RNZKn1f4RNZKn1f4QN[Ko1e4QN[Ko1e4PN\\KP2d4PN\\KP2c4PN^KP2b4PN^KP2b4oM_KQ2a4oM_KQ2a4oM_KQ2a4nM`KR2_4oMaKQ2_4nMbKR2^4nMbKR2^4mMcKS2]4mMcKS2]4lMdKT2[4mMeKS2[4lMfKT2Z4lMfKT2Z4lMfKT2Z4kMgKU2X4lMhKT2X4lMhKT2X4mMgKS2Y4mMgKS2Y4mMgKS2Y4nMfKR2Y4oMgKP2Z4PNgKo1Y4RNfKn1Z4RNfKn1Z4SNeKm1[4SNeKm1[4SNeKm1[4TNdKl1\\4TNdKl1\\4TNdKl1[4VNdKj1\\4VNdKj1\\4WNcKi1]4WNcKi1]4WNcKi1]4XNbKh1^4XNbKh1^4XNbKh1^4YNcKe1]4[NeKc1[4]NgKa1Y4`NhK^1X4bNjK\\1V4eNjKZ1W4eNkKY1U4gNmKW1T4iNmKU1S4kNoKS1R4lNPLo0S4ROnKi0V4VOkKg0W4ZOjKa0Z4^OhK=[4CgKOf40\\K@R5?i4O00?Aio:"}, "label": "the man in blue jeans and a red shirt"}]}
{"id": 92694, "image": "COCO_train2014_000000092694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt playing wii\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 196, 197, 198, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343, 359, 360, 361, 362, 363, 364, 365, 366, 384, 385, 386, 387, 388, 389], "bbox": [0.50534375, 0.4237916666666666, 0.9625625, 0.9719166666666667], "iscrowd": 0, "area": 41440.49110000001, "rle": {"size": [480, 640], "counts": "\\gg41i>8G8I8H7H9L3N3M2N3M2M3N3M2N3M2M4M2N3M2N3L3XN^MoFc2o8gMhFZ2W9nMaFS2^9VNlEX2R:\\1O001N2O1O0O2O1O1O001O1O1O00100000O1000000001O0000001O00001O000000101N3M3M2O2M2NQ1oN010O1O1O100O1O1O100O001O010O1O0010O012M2O2M2O1N3N1NZ1gNO0O2O001O001O001N10001O001O0O101O001O00001N101O00001O001N10001O001O000O2O001O00001O0O2O001O1O2N1O1N2O2N1O1O1O2N1N2O1OoJ]Gg3a8YLaGf3^8YLdGf3\\8ZLeGe3Z8[LhGd3X8\\LjGb3V8^LkGa3U8^LmGa3S8_LoG_3Q8`LQH_3n7bLSH\\3n7cLTH\\3l7dLVHZ3j7eLXHZ3h7fLYHY3g7fL[HY3d7hL^HV3b7jL_HU3a7jLaHU3_7kLbHT3^7kLeHS3Z7nLgHQ3Y7nLiHQ3W7oLjHP3V7oLmHo2S7QMnHn2R7QMPIn2o6SMSIk2m6UMTIj2l6UMVIj2j6VMWIi2i6VMZIh2f6XM[Ig2d6YM^If2b6ZM_Ie2a6ZMbId2^6\\McIc2]6\\MeIc2[6]MfIb2Y6_MiI_2W6`MkI_2U6aMlI^2T6aMoI]2Q6cMPJ\\2o5dMSJ[2m5eMTJZ2l5eMWJY2i5gMXJX2h5gM\\JV2d5jM^JT2a5mMbJP2^5oMeJo1[5QNhJl1X5SNjJl1V5TNlJj1T5UNnJj1Q5WNoJi1Q5VNPKj1P5VNPKj1P5UNQKk1o4UNQKk1o4UNQKk1n4UNSKk1m4UNSKk1m4TNTKl1l4TNTKl1l4SNUKm1k4SNUKm1j4SNWKm1i4SNWKm1i4RNXKn1h4RNXKn1h4RNXKn1g4RNZKn1f4RNZKn1f4QN[Ko1e4QN[Ko1e4PN\\KP2d4PN\\KP2c4PN^KP2b4PN^KP2b4oM_KQ2a4oM_KQ2a4oM_KQ2a4nM`KR2_4oMaKQ2_4nMbKR2^4nMbKR2^4mMcKS2]4mMcKS2]4lMdKT2[4mMeKS2[4lMfKT2Z4lMfKT2Z4lMfKT2Z4kMgKU2X4lMhKT2X4lMhKT2X4mMgKS2Y4mMgKS2Y4mMgKS2Y4nMfKR2Y4oMgKP2Z4PNgKo1Y4RNfKn1Z4RNfKn1Z4SNeKm1[4SNeKm1[4SNeKm1[4TNdKl1\\4TNdKl1\\4TNdKl1[4VNdKj1\\4VNdKj1\\4WNcKi1]4WNcKi1]4WNcKi1]4XNbKh1^4XNbKh1^4XNbKh1^4YNcKe1]4[NeKc1[4]NgKa1Y4`NhK^1X4bNjK\\1V4eNjKZ1W4eNkKY1U4gNmKW1T4iNmKU1S4kNoKS1R4lNPLo0S4ROnKi0V4VOkKg0W4ZOjKa0Z4^OhK=[4CgKOf40\\K@R5?i4O00?Aio:"}, "label": "a man in a red shirt playing wii"}]}
{"id": 92694, "image": "COCO_train2014_000000092694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two white game controls\"."}], "task": "refering", "answer_template": "The \"two white game controls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 262, 285, 286, 308], "bbox": [0.40325, 0.4563333333333333, 0.567421875, 0.7887083333333333], "iscrowd": 0, "area": 1013.5399000000002, "rle": {"size": [480, 640], "counts": "]Yi3=h=k0C<00100O010O010O1O100O100O2N3G^BPOc=n09N9G;E1O1O2N1J[A0f>O[]m02hbRO10O10O10O10O10O10O1VAMg>62N1O2N2NncQ4"}, "label": "two white game controls"}]}
{"id": 150044, "image": "COCO_train2014_000000150044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of the train that says 2a74 on the top\"."}], "task": "refering", "answer_template": "The \"the front of the train that says 2a74 on the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294], "bbox": [0.6623593750000001, 0.43427083333333333, 0.8342656250000001, 0.7308958333333333], "iscrowd": 0, "area": 13562.040399999993, "rle": {"size": [480, 640], "counts": "kPW66`>=D<D<D=C>C>I6J7J5J7I4L2O1N2N3M2O1N2N3M100O1O1O100O2O000O10000000001O0O1000000000001N1000000000000O1000000000000O1000000000000O11O00001O001O00001N100O2O0O101O0O101N101N10001N100O2O0O2O1N2O1O1N2N2N2N2N3M2N2O1N2N2N2N2N2Nd^a1"}, "label": "the front of the train that says 2a74 on the top"}]}
{"id": 150044, "image": "COCO_train2014_000000150044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train on the right\"."}], "task": "refering", "answer_template": "The \"the train on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294], "bbox": [0.6623593750000001, 0.43427083333333333, 0.8342656250000001, 0.7308958333333333], "iscrowd": 0, "area": 13562.040399999993, "rle": {"size": [480, 640], "counts": "kPW66`>=D<D<D=C>C>I6J7J5J7I4L2O1N2N3M2O1N2N3M100O1O1O100O2O000O10000000001O0O1000000000001N1000000000000O1000000000000O1000000000000O11O00001O001O00001N100O2O0O101O0O101N101N10001N100O2O0O2O1N2O1O1N2N2N2N2N3M2N2O1N2N2N2N2N2Nd^a1"}, "label": "the train on the right"}]}
{"id": 150044, "image": "COCO_train2014_000000150044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train in the middle\"."}], "task": "refering", "answer_template": "The \"train in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286, 287, 304, 308], "bbox": [0.10892187499999999, 0.3296875, 0.654390625, 0.7898125], "iscrowd": 0, "area": 43639.79819999999, "rle": {"size": [480, 640], "counts": "lSQ17h>9F4L3N2M3N2N2N1O2N2N1O1O001000O100O10000O100000000001O00000000000000001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1nNYNiDg1d:oNYER1`:WO]Ej0]:]O`Ed0\\:C^Ea0^:S2M3N1N3N2M3N1O2M3N2M5L3M3M3M3M2N100O001O1O1O0000O1000000O100001O1O0O2M2N3@aJ^G`5a8gJXGY5g8a0N1O201O0O10000000001O0O10000000001O0O1000000001O00000000000TJiGo4X8oJoGk4S8QKSHk4n7SKTHl4m7RKUHm4l7PKWHo4k7nJWHQ5j7mJXHR5l8O001O1O001O1O001O1O1O\\O[KmFe4R9_KjFb4U9aKiF_4W9dKfF\\4Y9gKeFY4Z9g000O1000001O000000000000001O00001O00001O001O000010O000O10000O101O3L8I7I2M100000O01N2N2N20O0100O10000O010O100O11O1O1O1O1O2NO1O010O1O1O100O1O1O100O1O1O1000000000000001O00000001N10001N101N1O1O2N101N1O1O2O001N10001N101O0O101O001N101O0O2O00001O1N101O1O1N101O1O2M3N1O2M3N1O001N101N101N101N101N101N101N2O0O2O0O2N101N1O2O0O2N2O0O2N3M4M2M3N001N101O0O2O001N101O0O2O001N100O2M2N3M4L3UOUB=\\>J7I6JmYW3"}, "label": "train in the middle"}]}
{"id": 150044, "image": "COCO_train2014_000000150044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train engine\"."}], "task": "refering", "answer_template": "The \"train engine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286, 287, 304, 308], "bbox": [0.10892187499999999, 0.3296875, 0.654390625, 0.7898125], "iscrowd": 0, "area": 43639.79819999999, "rle": {"size": [480, 640], "counts": "lSQ17h>9F4L3N2M3N2N2N1O2N2N1O1O001000O100O10000O100000000001O00000000000000001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1nNYNiDg1d:oNYER1`:WO]Ej0]:]O`Ed0\\:C^Ea0^:S2M3N1N3N2M3N1O2M3N2M5L3M3M3M3M2N100O001O1O1O0000O1000000O100001O1O0O2M2N3@aJ^G`5a8gJXGY5g8a0N1O201O0O10000000001O0O10000000001O0O1000000001O00000000000TJiGo4X8oJoGk4S8QKSHk4n7SKTHl4m7RKUHm4l7PKWHo4k7nJWHQ5j7mJXHR5l8O001O1O001O1O001O1O1O\\O[KmFe4R9_KjFb4U9aKiF_4W9dKfF\\4Y9gKeFY4Z9g000O1000001O000000000000001O00001O00001O001O000010O000O10000O101O3L8I7I2M100000O01N2N2N20O0100O10000O010O100O11O1O1O1O1O2NO1O010O1O1O100O1O1O100O1O1O1000000000000001O00000001N10001N101N1O1O2N101N1O1O2O001N10001N101O0O101O001N101O0O2O00001O1N101O1O1N101O1O2M3N1O2M3N1O001N101N101N101N101N101N101N2O0O2O0O2N101N1O2O0O2N2O0O2N3M4M2M3N001N101O0O2O001N101O0O2O001N100O2M2N3M4L3UOUB=\\>J7I6JmYW3"}, "label": "train engine"}]}
{"id": 150044, "image": "COCO_train2014_000000150044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a trained marked dr1 which was standing in the track near the green painted train\"."}], "task": "refering", "answer_template": "The \"a trained marked dr1 which was standing in the track near the green painted train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255], "bbox": [0.00190625, 0.43972916666666667, 0.213203125, 0.7112916666666665], "iscrowd": 0, "area": 12343.686349999998, "rle": {"size": [480, 640], "counts": "de0o0Q>k2VMO00NXLlDh3V;0O1002NO100001O0000001O00001O000000000000000000O100000E;E;0000000000000000=C4L00O2O00000O1000001N10000001O100O2N1O1O1O100O2N]OjD]MU;b2UEVMj:h2aEPM^:o2j0N3N1O1N2O1O1N3N1N2O1O1N2O1O2M2O1O1000001O000O100000001O0O100000001O0O100000001O0O10000000001O00000001O01O0000000000010O0000000_N\\Cf0_]\\7"}, "label": "a trained marked dr1 which was standing in the track near the green painted train"}]}
{"id": 191005, "image": "COCO_train2014_000000191005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe holding head highest\"."}], "task": "refering", "answer_template": "The \"giraffe holding head highest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 126, 127, 128, 150, 151, 152, 173, 174, 175, 198], "bbox": [0.435609375, 0.2671529411764706, 0.6540000000000001, 0.5798352941176471], "iscrowd": 0, "area": 5857.580849999999, "rle": {"size": [425, 640], "counts": "emc32V=2M2O2M3N1N2O1N2N200O1O1O1O100O1O1000O2O001O0010O10O0100O100O100000O01000000O1000O1000O1000000O10O100000O100000O0100000003L=D<D;D2N10O00001O000010O0001O00001O00010O00001O000010O0001O001O1O1O1O010O1O1O1O1O1O1O001O1O1O1O1O001000O2O1N2O0O2O1N2O001N2O0O2OO01000O010O010M2F;F:F9G:E:Gcak2"}, "label": "giraffe holding head highest"}]}
{"id": 191005, "image": "COCO_train2014_000000191005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe talking with anothergiraffe\"."}], "task": "refering", "answer_template": "The \"giraffe talking with anothergiraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 126, 127, 128, 150, 151, 152, 173, 174, 175, 198], "bbox": [0.435609375, 0.2671529411764706, 0.6540000000000001, 0.5798352941176471], "iscrowd": 0, "area": 5857.580849999999, "rle": {"size": [425, 640], "counts": "emc32V=2M2O2M3N1N2O1N2N200O1O1O1O100O1O1000O2O001O0010O10O0100O100O100000O01000000O1000O1000O1000000O10O100000O100000O0100000003L=D<D;D2N10O00001O000010O0001O00001O00010O00001O000010O0001O001O1O1O1O010O1O1O1O1O1O1O001O1O1O1O1O001000O2O1N2O0O2O1N2O001N2O0O2OO01000O010O010M2F;F:F9G:E:Gcak2"}, "label": "giraffe talking with anothergiraffe"}]}
{"id": 191005, "image": "COCO_train2014_000000191005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing alone in front of two other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe standing alone in front of two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 244, 245, 246, 267, 268, 269], "bbox": [0.6098125, 0.3170823529411764, 0.79525, 0.7953176470588235], "iscrowd": 0, "area": 10390.161049999997, "rle": {"size": [425, 640], "counts": "TVR51U=3M3M4L3M4L3L4K5L4K4M3L5K4M4K4M2MN2M4MN2O1N21O1O1N2O16J8H8H8L=C4L4LMTFoLh9n2:M3N2H8A?H8H9N33bEaNk8b1PGlNe8W1VGSOa8Q1ZGXO_8j0\\GA\\8b0_GGY8=bGMV85fG4S8g2I101N2O1OSNZHWNe7f1eHUNZ7g1oHUNP7g1ZITNd6i1eIUNX6k1nIQNo5Q2UJlMi5U2gHbMj03]6[3hI]LW6g3aIPLRO6[7o3[IRL[OLY7V4TIULDAX7^4lHYLKWOX7g5iHVJV7m511O1O1OWOPI[Ko6b4XI[Kf6c4bIZK\\6c4kIZKT6d4RJWKP6e4\\1L3TOgGfL]8W3mG^LW8_3k0M3M4M3L4L4QMlEd2^:M3L4M4K4A?01O000O101O000N3M2M3N3M3M3M3L4M3M3M2M1L5L304K5LQWf1"}, "label": "a giraffe standing alone in front of two other giraffes"}]}
{"id": 191005, "image": "COCO_train2014_000000191005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe bent over pile of brush\"."}], "task": "refering", "answer_template": "The \"giraffe bent over pile of brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 244, 245, 246, 267, 268, 269], "bbox": [0.6098125, 0.3170823529411764, 0.79525, 0.7953176470588235], "iscrowd": 0, "area": 10390.161049999997, "rle": {"size": [425, 640], "counts": "TVR51U=3M3M4L3M4L3L4K5L4K4M3L5K4M4K4M2MN2M4MN2O1N21O1O1N2O16J8H8H8L=C4L4LMTFoLh9n2:M3N2H8A?H8H9N33bEaNk8b1PGlNe8W1VGSOa8Q1ZGXO_8j0\\GA\\8b0_GGY8=bGMV85fG4S8g2I101N2O1OSNZHWNe7f1eHUNZ7g1oHUNP7g1ZITNd6i1eIUNX6k1nIQNo5Q2UJlMi5U2gHbMj03]6[3hI]LW6g3aIPLRO6[7o3[IRL[OLY7V4TIULDAX7^4lHYLKWOX7g5iHVJV7m511O1O1OWOPI[Ko6b4XI[Kf6c4bIZK\\6c4kIZKT6d4RJWKP6e4\\1L3TOgGfL]8W3mG^LW8_3k0M3M4M3L4L4QMlEd2^:M3L4M4K4A?01O000O101O000N3M2M3N3M3M3M3L4M3M3M2M1L5L304K5LQWf1"}, "label": "giraffe bent over pile of brush"}]}
{"id": 191005, "image": "COCO_train2014_000000191005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first giraffe on the left who is looking towards the right\"."}], "task": "refering", "answer_template": "The \"first giraffe on the left who is looking towards the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 164, 165, 166, 167], "bbox": [0.16815625, 0.28811764705882353, 0.459265625, 0.5437176470588235], "iscrowd": 0, "area": 6655.545050000001, "rle": {"size": [425, 640], "counts": "eP]1a0c<6G:MO1M3L4N2N3O06J5K6J6J6J4L10O01O010O01O10O010O01O1O0F;M21000O02O1N3N1N2O1N10000O101N100O1O1O2O0O1O1O101N001O010O1O1O1TERN]:P2^ESNa:Q2XERNh:W2010O00010O2N100O1O2O0O100O2O00000L4L4L4L4L4G9I7K5L4000000O1000000O1000000O100000O010000O10O1000O100000O01000000O1000000O01000O010000O010O0100O010O10O01O010O010O0010O0100O1O100O10000001O000000001M2L4K5N30O1O1O1O5K2N1O2N1O1O1O1O1O2N2O001N2O0O2N2M_^_4"}, "label": "first giraffe on the left who is looking towards the right"}]}
{"id": 502300, "image": "COCO_train2014_000000502300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"computer monitor above laptop screen\"."}], "task": "refering", "answer_template": "The \"computer monitor above laptop screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175], "bbox": [0.37190625, 0.1974375, 0.634265625, 0.4731458333333334], "iscrowd": 0, "area": 19981.119899999998, "rle": {"size": [480, 640], "counts": "Ue_3d0\\>X1hNX1gNf0[O1O2N1O1O0000000000O10000000000000000O1000000000000000O010000000000000000O10000000000000000O10000000000000000O0100000000000000000O10000000000000000O10000000000000000O01000000000000000O10000000000000000O10000000000000000O10O10000000000000O10000000000000000O10000000000000000O1000O100000000000O10000000000000000O10000000001O00m0SOh1WN\\`]3"}, "label": "computer monitor above laptop screen"}]}
{"id": 502300, "image": "COCO_train2014_000000502300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large computer monitor directly above a laptop monitor\"."}], "task": "refering", "answer_template": "The \"a large computer monitor directly above a laptop monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175], "bbox": [0.37190625, 0.1974375, 0.634265625, 0.4731458333333334], "iscrowd": 0, "area": 19981.119899999998, "rle": {"size": [480, 640], "counts": "Ue_3d0\\>X1hNX1gNf0[O1O2N1O1O0000000000O10000000000000000O1000000000000000O010000000000000000O10000000000000000O10000000000000000O0100000000000000000O10000000000000000O10000000000000000O01000000000000000O10000000000000000O10000000000000000O10O10000000000000O10000000000000000O10000000000000000O1000O100000000000O10000000000000000O10000000001O00m0SOh1WN\\`]3"}, "label": "a large computer monitor directly above a laptop monitor"}]}
{"id": 510493, "image": "COCO_train2014_000000510493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb in side profile in a green field next to another lamb and a large sheep\"."}], "task": "refering", "answer_template": "The \"a lamb in side profile in a green field next to another lamb and a large sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282], "bbox": [0.1375625, 0.44344262295081965, 0.40207812499999995, 0.8384543325526932], "iscrowd": 0, "area": 12167.069899999997, "rle": {"size": [427, 640], "counts": "SmT14V=9G7I1O1O1O100O1O1O1M3M2N3N2N2O1N110000000000O100000000O100000000O12N1O2N1O2N1O2M3M4L4M3L5VGQNo5R2mISNQ6n1kIWNR6k1iIZNU6i1fIYNZ6i1bIYN_6DhGf1d1iNd6_OkGi1f0ZN_O`0P8\\OmGk1L6W8nMoGX4R8eKQHZ4o7eKTHY4l7eKWHZ4i7eKYHZ4]8O1O2N1N2O1O1O2M20O001O1O10O01O1O001O1O0000000000YNUGmNm8h0bGSO^8e0lGYOT8d0QHZOo7c0UH\\Ok7b0YHiNZ8W1hG`N_8`1f1000000000001O00000000O1000000O1000001fE^Nh8d1UG^Nk8b1SGaNl8`1PGdNo8]1nFeNR9]1jFfNT9]1hFeNX9]1cFfN]9[1`FgN`9[1\\FgNd9[1WFhNi9Y1TFiNl9Y1PFiNP:Y21O00000O1O1O2N1O1N2^OlElMU:Q2TFhMn9U2e0M3M3M4N200000O01000000dNgDf0X;VOoDh0P;TOVEk0i:PO^EP1OROb9JdFT1^O\\On:`11O2N1O1O1O2N1O1O2N1N2O2N5K7H7POgCb0h<I7I_\\o4"}, "label": "a lamb in side profile in a green field next to another lamb and a large sheep"}]}
{"id": 510493, "image": "COCO_train2014_000000510493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young sheep standing up\"."}], "task": "refering", "answer_template": "The \"a young sheep standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282], "bbox": [0.1375625, 0.44344262295081965, 0.40207812499999995, 0.8384543325526932], "iscrowd": 0, "area": 12167.069899999997, "rle": {"size": [427, 640], "counts": "SmT14V=9G7I1O1O1O100O1O1O1M3M2N3N2N2O1N110000000000O100000000O100000000O12N1O2N1O2N1O2M3M4L4M3L5VGQNo5R2mISNQ6n1kIWNR6k1iIZNU6i1fIYNZ6i1bIYN_6DhGf1d1iNd6_OkGi1f0ZN_O`0P8\\OmGk1L6W8nMoGX4R8eKQHZ4o7eKTHY4l7eKWHZ4i7eKYHZ4]8O1O2N1N2O1O1O2M20O001O1O10O01O1O001O1O0000000000YNUGmNm8h0bGSO^8e0lGYOT8d0QHZOo7c0UH\\Ok7b0YHiNZ8W1hG`N_8`1f1000000000001O00000000O1000000O1000001fE^Nh8d1UG^Nk8b1SGaNl8`1PGdNo8]1nFeNR9]1jFfNT9]1hFeNX9]1cFfN]9[1`FgN`9[1\\FgNd9[1WFhNi9Y1TFiNl9Y1PFiNP:Y21O00000O1O1O2N1O1N2^OlElMU:Q2TFhMn9U2e0M3M3M4N200000O01000000dNgDf0X;VOoDh0P;TOVEk0i:PO^EP1OROb9JdFT1^O\\On:`11O2N1O1O1O2N1O1O2N1N2O2N5K7H7POgCb0h<I7I_\\o4"}, "label": "a young sheep standing up"}]}
{"id": 428576, "image": "COCO_train2014_000000428576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and grey marble table with wooden chairs and place settings\"."}], "task": "refering", "answer_template": "The \"a white and grey marble table with wooden chairs and place settings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 275, 276, 277, 278, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 324, 325], "bbox": [0.0059019607843137254, 0.581765625, 0.5196862745098039, 0.8241250000000001], "iscrowd": 0, "area": 21119.17575000001, "rle": {"size": [640, 510], "counts": "[[2T1d`0X2H710O01000O0100O010O10O10O10O01000O010O0100O01000O010O10O01000O0100O010O10O10O10O01000O010O0100O01000O010O10O01000O0100O010O10O10O10O01000O0100O010O01000O010O10O01000O0100O010O10O10O10O0100O01000O010O01000O010O10O01000O0100O010O2O000O100O100O10000O2O0O10000O100O100O10001N100O100O10000O101N100O101O0O101N10001N100O2O0O10001N100O2O0O101O0O101N100O2O000O101N10001N100O2O0O101O0O100O2O0O101O0O101N100O2O000O2O0O100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O101N1O100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O10Ugh4"}, "label": "a white and grey marble table with wooden chairs and place settings"}]}
{"id": 428576, "image": "COCO_train2014_000000428576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white color table have three cup & soccer\"."}], "task": "refering", "answer_template": "The \"a white color table have three cup & soccer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 275, 276, 277, 278, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 324, 325], "bbox": [0.0059019607843137254, 0.581765625, 0.5196862745098039, 0.8241250000000001], "iscrowd": 0, "area": 21119.17575000001, "rle": {"size": [640, 510], "counts": "[[2T1d`0X2H710O01000O0100O010O10O10O10O01000O010O0100O01000O010O10O01000O0100O010O10O10O10O01000O010O0100O01000O010O10O01000O0100O010O10O10O10O01000O0100O010O01000O010O10O01000O0100O010O10O10O10O0100O01000O010O01000O010O10O01000O0100O010O2O000O100O100O10000O2O0O10000O100O100O10001N100O100O10000O101N100O101O0O101N10001N100O2O0O10001N100O2O0O101O0O101N100O2O000O101N10001N100O2O0O101O0O100O2O0O101O0O101N100O2O000O2O0O100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O101N1O100O2O0O101N100O2O0O101N100O101N100O2O0O101N100O2O0O10Ugh4"}, "label": "a white color table have three cup & soccer"}]}
{"id": 428576, "image": "COCO_train2014_000000428576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"set of chairs\"."}], "task": "refering", "answer_template": "The \"set of chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 259, 260, 277, 278, 279, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 344, 345, 346, 347, 348, 349, 350, 351, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 378, 379, 380, 381, 382, 383, 384, 385, 387, 396, 397, 398, 399, 400, 401, 403, 405, 406], "bbox": [0.000980392156862745, 0.574078125, 0.5635686274509805, 1.0], "iscrowd": 0, "area": 33001.588050000006, "rle": {"size": [640, 510], "counts": "[W1e0db0g0[Oe000000000000000000000001O0000000000000000000000O1O1O1O1O1O1O1O1O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100L4D<E;D<E;J60000O10000O10000O1000000O10000O10000O1000000O10000O10000001O00001O000O2NN[Kk@e4U?[Kk@d4U?]Kl@a4U?`Kj@_4V?bKj@^4U?cKl@[4U?fKj@Y4V?hKj@W4W?iKj@V4U?lKj@S4V?nKj@Q4W?PLi@o3V?RLj@m3W?SLi@l3X?ULg@j3Z?VLg@i3Z?VLf@i3[?XLd@g3]?YLd@e3]?[Lc@e3^?[La@d3`?\\La@b3`?_L_@`3b?`L^@`3c?_L^@_3c?bL\\@]3e?cL[@\\3g?cLZ@\\3f?eLY@Z3h?fLX@Y3i?gLX@W3j?iLU@W3k?iLU@V3l?jLT@U3m?lLS@S3Z?_LXA=^OV3X?]LZAV4a>lK^A<_Oi04V1k>eMbA<_Oh0<T1^>iMhAa12c0Q>nMlA9_Oe0k0R1f=PNPB9_Od0S1P1Y=TNUB7@b1<Mn=ZNWB5_Oh1?Cj=bNWB2_Ol1f0YOd=iNWBZ2<fN\\=QOXBY2c0_NU=YOXBX2h0YNo<@YBW2o0RNh<GYBW2V1kM`<O[BU2\\1eMY<7ZBU2b1^MS<>[BT2i?mMW@S2i?nMW@Q2h?PNX@Q2g?oMY@Q2f?PNZ@P2f?QNZ@n1c=aMjCb0bNo1_=cMoC>bNo1\\=fMQD=aNn1Z=iMUD9aNo1U=lMYD6bNn1R=oM\\D3aNo1T=nMYD5bNm1U=nMXD6cNm1T=mMYD6bNn1U=lMXD8bNl1V=lMXD8aNm1W=lMVD8aNo1X=iMWD9^NP2[=gMVD:]NQ2^=dMUD;\\NS2^=bMUD<]NR2^=cM[D\\3b;dLVEi2e:WM[D<^N`2Q=TMaD;_Ng2j<oLeD;aNk2f<iLiD<`NR3a<bLnD<bNW3[<]LSE<bN]3U<WLXE=cNb3o;RL]E;eNh3i;mKaE<fNm3c;gKgE<eNU22`N_;oNiEG\\Oi2FXNV;WOhEG]Oi2OPNl:AfEFXOFCR3m0hMb:JfEESOMHl2V1_MY:3fEDmN4Oe2_1XMn9;gECmN6Nd2j1oLd9e0fEBnN5Oe2_=dMdCBnN5Od2a=dMaCDnN41c2a=eMaCClN62a2b=fM`CClN61b2c=fM_CBmN52b2c=gM^CBnN41b2d=hM]CCmN23c2c=hM^CBlN33b2d=iM]CBlN24b2e=jMZCBmN23b2g=jMYCBmN23b2g=jMZCAlN24b2g=kMYCBlN04b2h=mMWCAmNO5c2g=mMWCAmNO4c2i=mMVCAmNO4b2j=nMVC@lNO5c2j=mMUCAlNO5b2k=oMSCAlNM7b2k=PNRCAmNL5c2m=PNRC@lNL6d2l=PNRC@lNL5d2n=PNQC@lNL5c2o=RNoB_OmNK5e2P>PNnBAlNJ6d2Q>QNnB@kNJ6e2R>QNmB@lNI5e2S>SNkB_OmNH5g2S>RNkB_OmNH4g2U>RNhBXOkN64I4f2V>SN`BI^OU2R>RNYB0Em1T>SNQB5Jh1V>SNoAM@A;^2W>TNnA9Gd1[>SNnAW3S>iLmAV3T>kLjAV3V>jLjAU3X>fLlAY3U>`LQB`3P>ZLQCk2o<oLXCo2m>O100O110nMS_OS1i`0kN__OS1]`0lNi_OS1S`0lNT@R1g?mN`@R1[?nNl@o0P?POWAo0e>PObAn0Y>QOoAm0l=SOZBl0a=SOgBk0U=TOQCj0k<UO]Ci0^<WOhCh0S<WOUDg0e?O2N1N3N1O2N1O2N2N1K6J5K6JZhZ4"}, "label": "set of chairs"}]}
{"id": 428576, "image": "COCO_train2014_000000428576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and gray island\"."}], "task": "refering", "answer_template": "The \"a white and gray island\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 231, 232, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286], "bbox": [0.6156078431372548, 0.442125, 0.9558039215686274, 0.6897499999999999], "iscrowd": 0, "area": 19790.091349999995, "rle": {"size": [640, 510], "counts": "UbT6Y1cb0Q2RN1000001N1000000O10001O0O1000000O101O000O1000000O101O0O1000000O101O000O1000000O2O00000O1000001N1000000O1000001N1000000O10001N1000000O10001O0O1000000@mKRAT4m>oKQAR4n>QLo@o3P?ULm@k3R?b02N1O2N1O2O1N1O2N1O000001O000O100000001O0O100000_Ol@VLT?c3h0O02O00000009HO0000001O000000001O0000010O000000001O0001O0001O00000010O0000000PMg_OZ2X`0fMj_OX2V`0hMl_OV2T`0jMm_OU2T`0jMl_OW2S`0iMn_OV2R`0jMo_OY1ImNX`0Jo_OY1KkNV`0LP@X1MiNS`0OQ@W1NfNS`0DR@L0g1MdNU`0Go_ON1e1NaNT`0Kn_OO3b1M`NT`0Nm_O04a1a`0_N[_O07^1^`0aN\\_O19[1m`0eNU_OY1k`0gNX_OV1h`0jNZ_OT1f`0mN\\_OP1d`0PO^_Oo0a`0QOb_Ol0^`0TOd_Oj0\\`0VOg_Og0Y`0YOj_Oc0W`0^Oj_O>X`0Al_O:Pb0L5Koo="}, "label": "a white and gray island"}]}
{"id": 428576, "image": "COCO_train2014_000000428576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and silver island in a kitchen with a centerpiece and two modern chairs\"."}], "task": "refering", "answer_template": "The \"white and silver island in a kitchen with a centerpiece and two modern chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 231, 232, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286], "bbox": [0.6156078431372548, 0.442125, 0.9558039215686274, 0.6897499999999999], "iscrowd": 0, "area": 19790.091349999995, "rle": {"size": [640, 510], "counts": "UbT6Y1cb0Q2RN1000001N1000000O10001O0O1000000O101O000O1000000O101O0O1000000O101O000O1000000O2O00000O1000001N1000000O1000001N1000000O10001N1000000O10001O0O1000000@mKRAT4m>oKQAR4n>QLo@o3P?ULm@k3R?b02N1O2N1O2O1N1O2N1O000001O000O100000001O0O100000_Ol@VLT?c3h0O02O00000009HO0000001O000000001O0000010O000000001O0001O0001O00000010O0000000PMg_OZ2X`0fMj_OX2V`0hMl_OV2T`0jMm_OU2T`0jMl_OW2S`0iMn_OV2R`0jMo_OY1ImNX`0Jo_OY1KkNV`0LP@X1MiNS`0OQ@W1NfNS`0DR@L0g1MdNU`0Go_ON1e1NaNT`0Kn_OO3b1M`NT`0Nm_O04a1a`0_N[_O07^1^`0aN\\_O19[1m`0eNU_OY1k`0gNX_OV1h`0jNZ_OT1f`0mN\\_OP1d`0PO^_Oo0a`0QOb_Ol0^`0TOd_Oj0\\`0VOg_Og0Y`0YOj_Oc0W`0^Oj_O>X`0Al_O:Pb0L5Koo="}, "label": "white and silver island in a kitchen with a centerpiece and two modern chairs"}]}
{"id": 281122, "image": "COCO_train2014_000000281122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plastic and metal chair pushed under a desk\"."}], "task": "refering", "answer_template": "The \"a plastic and metal chair pushed under a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 102, 103, 104, 119, 120, 136, 137, 153, 154, 155, 170, 171, 172, 187, 188, 190, 204, 207, 224, 225], "bbox": [0.003, 0.24943749999999998, 0.2546875, 0.62021875], "iscrowd": 0, "area": 8830.604350000001, "rle": {"size": [640, 480], "counts": "^n05Vb0e1[Nd1aN`1O010O10O010O10O010O010O0100O001O001O0010O01O1O001O001O001O010O1O001O001O001O1O010O1O1O1O1O001O1]Md@a0]?YOn@a0S?YOXAa0i>XObAd0]>WOnAc0S>WOXBc0i=VOcBd0^=VOlBe0U=UOVCe0k<TOaCf0kP=\\Oh_C7J5J3M0001O000010O0001O000010O0001O000010O0001O00001O01O01O00001O01O01O000L4Lm[o6"}, "label": "a plastic and metal chair pushed under a desk"}]}
{"id": 281122, "image": "COCO_train2014_000000281122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the smaller chair\"."}], "task": "refering", "answer_template": "The \"the back of the smaller chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 102, 103, 104, 119, 120, 136, 137, 153, 154, 155, 170, 171, 172, 187, 188, 190, 204, 207, 224, 225], "bbox": [0.003, 0.24943749999999998, 0.2546875, 0.62021875], "iscrowd": 0, "area": 8830.604350000001, "rle": {"size": [640, 480], "counts": "^n05Vb0e1[Nd1aN`1O010O10O010O10O010O010O0100O001O001O0010O01O1O001O001O001O010O1O001O001O001O1O010O1O1O1O1O001O1]Md@a0]?YOn@a0S?YOXAa0i>XObAd0]>WOnAc0S>WOXBc0i=VOcBd0^=VOlBe0U=UOVCe0k<TOaCf0kP=\\Oh_C7J5J3M0001O000010O0001O000010O0001O000010O0001O00001O01O01O00001O01O01O000L4Lm[o6"}, "label": "the back of the smaller chair"}]}
{"id": 281122, "image": "COCO_train2014_000000281122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange handle scissor on a gray chair\"."}], "task": "refering", "answer_template": "The \"orange handle scissor on a gray chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 22, 23, 24, 25, 26, 27, 28, 29, 30, 39, 40, 41, 42, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174, 175, 176, 177, 178, 179, 180, 182, 189, 192, 193, 194, 195, 196, 199, 206, 210, 211, 213, 216, 217, 223, 227, 228, 230, 233, 234, 240, 241, 244, 245, 247, 250, 257, 258, 261, 262, 264, 267, 274, 275, 281, 284, 291, 292, 298, 301, 309, 315, 326, 332, 349, 365, 366, 382, 383], "bbox": [0.1135625, 0.008843750000000001, 0.7930416666666666, 0.9852968749999998], "iscrowd": 0, "area": 76720.3887, "rle": {"size": [640, 480], "counts": "acR1<cc02O1N3M2N2N2O1N2N3M2N2O1N2N<De0[Od0]Od0[Oe0[Oe0\\Oc0_APKk<[6H7IdJnBR4i<PLeCh3T<XLZDb3^;_LPEY3h:hLfEQ3S:PM[Fi2]9XMQG`2g8aMgGY2Q8hM]HP2[7QNTIh1d6YNhIa1Q6_NRJb1i5`NYJ]1g5dN[JZ1d5fN_JW1`5kNbJR1^5nNeJP1Z5QOhJl0W5VOkJg0U5YOnJd0Q5^OQK?o4ATK<l4EVK8i4I[K3e4N^KNb42bKJ]48gKCY4=kK^OU4c0oKXOR4i0QLSOo3m0ULnNk3T1XLhNh3X1]LbNd3_1_L]N`3d1dLWN]3j1hLnMZ3R2Y7101N101O0O2O001N101O0O101N101O0O2O001N101O0O101N101O0O2O001N101O000O2O001O0O2O001O0O2O000O2O0iJdLjI]3T6oLaIQ3^6ZMWIh2h6bMnH^2Q7mMeHS2[7XNZHi1e7`NQHa1n7dNnG\\1R8gNkGY1U8jNhGV1W8nNeGT1Z8PObGP1^8SO_Gm0`8WO]Gi0c8ZOYGg0g8\\OVGe0h8@TG`0l8CPG>P9EmF;R9IkF7U9MgF4X9OdF2[92bFN^95_FKa98\\FHc9=XFEg9?UFAk9c0QF]OU1ULh6d4nGWOQ1VMY6g3aHTOm0jMV6e7\\IdHc6b7oHgHQ7T:00O2O000O2O001N1000iLYIUIf6l6hIfHX6Z7mI`HT6`7RJZHm5g7XJTHh5l7]JoGc5Q8bJjG]5W8iJaGY5`8kJnFb5R9cJZFk5g9W2000O1000000000000O100000001O000000000000000000000O10000000001O0000000kN^FlGb9k7hFTHX9c7SG[Hm8]7\\GbHd8U7fGjHZ8n6PHoHQ8h6YHWIh7_6bH`Ia7U6gHkI_7f5jHZJ[7X5mHiJP9U3UGkL`;;eDEk?0000000000000000000\\FI_O7`0U2UMkMj2Q4YKQLf4f5_I`J^6n5QIWJn6U6aHQJ\\7X6YHmIe7\\6oGiIP8^6aGkI^8]6QGnIm8c8O1O1N2O1O1O2N1O1O1N2O1O2N1O1O1O1O1N2O2N1O1O1O1O1N3N1O1O1O1O1O1O2M2O1O1O1O1O1O2N1N2O1O1O1O2N1O1N2O2N110000OB^C^IP7I\\Nf6gJaIg6NcN^6hJeI_63hNV6lJfIW69nNo5lJiIP6>TOf5oJlIg5c0ZO_5RKmI_5j0_OV5TKQJW5o0Ej4ZKWJk4T1K^4aK_J^4X12V3cLaKV3^18k2hLgKj2b1`0b2jLmK`2f1f0Y2oLVLo1g1S1m1TMaL^1f1`1d1WMkLn0f1k1[1\\MTM>f1W2P1aM^MNf1c2g0cMiM_Oe1n2>hMRNoNe1Z33mM]N^Nd1g3JTNcNlMf1P4C^NfNZMj1Y4\\OgNiNgLn1c4TOPOnNTLP2m4oNXOQO`Kk1c5POWOSOiJj1\\6nNUO]2W1`MSOn1Y1nMQO`1[1[NoNS1^1iNlNe0`1WOjN7c1ChNIf13dNfN]2V1Y8K5L4L4K5L5J5L4L]^n1"}, "label": "orange handle scissor on a gray chair"}]}
{"id": 281122, "image": "COCO_train2014_000000281122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey chair with scissors on it and a pile of debris under it\"."}], "task": "refering", "answer_template": "The \"a grey chair with scissors on it and a pile of debris under it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 22, 23, 24, 25, 26, 27, 28, 29, 30, 39, 40, 41, 42, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174, 175, 176, 177, 178, 179, 180, 182, 189, 192, 193, 194, 195, 196, 199, 206, 210, 211, 213, 216, 217, 223, 227, 228, 230, 233, 234, 240, 241, 244, 245, 247, 250, 257, 258, 261, 262, 264, 267, 274, 275, 281, 284, 291, 292, 298, 301, 309, 315, 326, 332, 349, 365, 366, 382, 383], "bbox": [0.1135625, 0.008843750000000001, 0.7930416666666666, 0.9852968749999998], "iscrowd": 0, "area": 76720.3887, "rle": {"size": [640, 480], "counts": "acR1<cc02O1N3M2N2N2O1N2N3M2N2O1N2N<De0[Od0]Od0[Oe0[Oe0\\Oc0_APKk<[6H7IdJnBR4i<PLeCh3T<XLZDb3^;_LPEY3h:hLfEQ3S:PM[Fi2]9XMQG`2g8aMgGY2Q8hM]HP2[7QNTIh1d6YNhIa1Q6_NRJb1i5`NYJ]1g5dN[JZ1d5fN_JW1`5kNbJR1^5nNeJP1Z5QOhJl0W5VOkJg0U5YOnJd0Q5^OQK?o4ATK<l4EVK8i4I[K3e4N^KNb42bKJ]48gKCY4=kK^OU4c0oKXOR4i0QLSOo3m0ULnNk3T1XLhNh3X1]LbNd3_1_L]N`3d1dLWN]3j1hLnMZ3R2Y7101N101O0O2O001N101O0O101N101O0O2O001N101O0O101N101O0O2O001N101O000O2O001O0O2O001O0O2O000O2O0iJdLjI]3T6oLaIQ3^6ZMWIh2h6bMnH^2Q7mMeHS2[7XNZHi1e7`NQHa1n7dNnG\\1R8gNkGY1U8jNhGV1W8nNeGT1Z8PObGP1^8SO_Gm0`8WO]Gi0c8ZOYGg0g8\\OVGe0h8@TG`0l8CPG>P9EmF;R9IkF7U9MgF4X9OdF2[92bFN^95_FKa98\\FHc9=XFEg9?UFAk9c0QF]OU1ULh6d4nGWOQ1VMY6g3aHTOm0jMV6e7\\IdHc6b7oHgHQ7T:00O2O000O2O001N1000iLYIUIf6l6hIfHX6Z7mI`HT6`7RJZHm5g7XJTHh5l7]JoGc5Q8bJjG]5W8iJaGY5`8kJnFb5R9cJZFk5g9W2000O1000000000000O100000001O000000000000000000000O10000000001O0000000kN^FlGb9k7hFTHX9c7SG[Hm8]7\\GbHd8U7fGjHZ8n6PHoHQ8h6YHWIh7_6bH`Ia7U6gHkI_7f5jHZJ[7X5mHiJP9U3UGkL`;;eDEk?0000000000000000000\\FI_O7`0U2UMkMj2Q4YKQLf4f5_I`J^6n5QIWJn6U6aHQJ\\7X6YHmIe7\\6oGiIP8^6aGkI^8]6QGnIm8c8O1O1N2O1O1O2N1O1O1N2O1O2N1O1O1O1O1N2O2N1O1O1O1O1N3N1O1O1O1O1O1O2M2O1O1O1O1O1O2N1N2O1O1O1O2N1O1N2O2N110000OB^C^IP7I\\Nf6gJaIg6NcN^6hJeI_63hNV6lJfIW69nNo5lJiIP6>TOf5oJlIg5c0ZO_5RKmI_5j0_OV5TKQJW5o0Ej4ZKWJk4T1K^4aK_J^4X12V3cLaKV3^18k2hLgKj2b1`0b2jLmK`2f1f0Y2oLVLo1g1S1m1TMaL^1f1`1d1WMkLn0f1k1[1\\MTM>f1W2P1aM^MNf1c2g0cMiM_Oe1n2>hMRNoNe1Z33mM]N^Nd1g3JTNcNlMf1P4C^NfNZMj1Y4\\OgNiNgLn1c4TOPOnNTLP2m4oNXOQO`Kk1c5POWOSOiJj1\\6nNUO]2W1`MSOn1Y1nMQO`1[1[NoNS1^1iNlNe0`1WOjN7c1ChNIf13dNfN]2V1Y8K5L4L4K5L5J5L4L]^n1"}, "label": "a grey chair with scissors on it and a pile of debris under it"}]}
{"id": 412195, "image": "COCO_train2014_000000412195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl woman hanging on to the dog , she is wearing a green headband\"."}], "task": "refering", "answer_template": "The \"the girl woman hanging on to the dog , she is wearing a green headband\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 80, 81, 82, 83, 98, 99, 100, 101, 114, 116, 117, 118, 119, 131, 132, 133, 135, 136, 149, 150, 151, 153, 154, 155, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208], "bbox": [0.27333999999999997, 0.25039039039039035, 0.62708, 0.9866666666666667], "iscrowd": 0, "area": 22911.8458, "rle": {"size": [333, 500], "counts": "]k\\1?i95K5K5K5K5K5K5K5K5K5K5K5K5K5M3M3M3M3L4M3M3M3L4M3M3M3M3L4M3M3M3L4M31O1O1O1O2N1O1O1O1O1O010O1O001O1O001O1O1O001O000000O10000O100O10000O100O10000O1001O1O2N2N1nLdIQ2^6gMeI\\2]6]MeIf2\\6SMhIo2Z6jLhI;I\\2k6[MbIc2V7N3M2NQLdMOZ2NjM4S2IQN8m1FVN:i1C[N=d1@`N?a1]OcNc0\\1ZOhNf0W1WOmNh0S1UOQOi0P1VOROiMiNS2U23TOiMhNR2U24UOhMhNR2T26UOgMiNP2R29XOeMgNP2R2:[OcMeNQ2P2<]OaMdNQ2P2=_O_MdNQ2n1?B]MaNR2n1a0C[MaNQ2m1c0FYM^NR2m1d0HWM]NS2l1f0IUM]NR2k1h0`0UOBj0a0SO@m0a0QO@n0c0oN^OP1d0nN]OS1c0kN^OU1d0hN]OX1d0hN[OX1g0fNYO[1g0eNXO[1j0cNVO]1m1aMRN`2n1`MQN`2P2`MoM`2Q2aMnM_2S2`MmMa2R2`MmM`2T2`MkM`2V2`MiM`2W2aMhM_2Y2aMfM`2Y2aMfM_2Z1XMTM;a1]2X1\\MUM;`1Y2[1^MQM=b1V2]1]MoLa0a1R2`1_MkLc0c1n1b1aMhLc0d1m1d1aMfLd0e1k1e1cMbLe0h1i1f1cM_Lf0j1g1g1oNXNQ1h1oNXNR1g1WOPNj0o1WOPNj0P2VOoMk0P2VOoMk0P2WOnMj0Q2WOlMl0S2UOhMQ1W2oLUMQ1>V2\\2iLYMo06]2`2dL\\Mn0Nd2e2^L^Mb5`2_JaMa5]2`JdM`5Z2bJfM^5X2cJiM]5U2dJlM\\5S2dJnM\\5o1fJRNZ5j1iJVNX5f1kJiNg4l0cKQOa4l0aKQOc4n0]KPOf4o0ZKnNj4P1XKlNl4S1TKjNP5U1QKgNS5X1mJcNY5g0bIWOV1L^5k0aIUOQ1Je5o0Y2SOm0LVfl1"}, "label": "the girl woman hanging on to the dog , she is wearing a green headband"}]}
{"id": 412195, "image": "COCO_train2014_000000412195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman with a green headband is holding a black and white puppy on her shoulders\"."}], "task": "refering", "answer_template": "The \"a blonde woman with a green headband is holding a black and white puppy on her shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 80, 81, 82, 83, 98, 99, 100, 101, 114, 116, 117, 118, 119, 131, 132, 133, 135, 136, 149, 150, 151, 153, 154, 155, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208], "bbox": [0.27333999999999997, 0.25039039039039035, 0.62708, 0.9866666666666667], "iscrowd": 0, "area": 22911.8458, "rle": {"size": [333, 500], "counts": "]k\\1?i95K5K5K5K5K5K5K5K5K5K5K5K5K5M3M3M3M3L4M3M3M3L4M3M3M3M3L4M3M3M3L4M31O1O1O1O2N1O1O1O1O1O010O1O001O1O001O1O1O001O000000O10000O100O10000O100O10000O1001O1O2N2N1nLdIQ2^6gMeI\\2]6]MeIf2\\6SMhIo2Z6jLhI;I\\2k6[MbIc2V7N3M2NQLdMOZ2NjM4S2IQN8m1FVN:i1C[N=d1@`N?a1]OcNc0\\1ZOhNf0W1WOmNh0S1UOQOi0P1VOROiMiNS2U23TOiMhNR2U24UOhMhNR2T26UOgMiNP2R29XOeMgNP2R2:[OcMeNQ2P2<]OaMdNQ2P2=_O_MdNQ2n1?B]MaNR2n1a0C[MaNQ2m1c0FYM^NR2m1d0HWM]NS2l1f0IUM]NR2k1h0`0UOBj0a0SO@m0a0QO@n0c0oN^OP1d0nN]OS1c0kN^OU1d0hN]OX1d0hN[OX1g0fNYO[1g0eNXO[1j0cNVO]1m1aMRN`2n1`MQN`2P2`MoM`2Q2aMnM_2S2`MmMa2R2`MmM`2T2`MkM`2V2`MiM`2W2aMhM_2Y2aMfM`2Y2aMfM_2Z1XMTM;a1]2X1\\MUM;`1Y2[1^MQM=b1V2]1]MoLa0a1R2`1_MkLc0c1n1b1aMhLc0d1m1d1aMfLd0e1k1e1cMbLe0h1i1f1cM_Lf0j1g1g1oNXNQ1h1oNXNR1g1WOPNj0o1WOPNj0P2VOoMk0P2VOoMk0P2WOnMj0Q2WOlMl0S2UOhMQ1W2oLUMQ1>V2\\2iLYMo06]2`2dL\\Mn0Nd2e2^L^Mb5`2_JaMa5]2`JdM`5Z2bJfM^5X2cJiM]5U2dJlM\\5S2dJnM\\5o1fJRNZ5j1iJVNX5f1kJiNg4l0cKQOa4l0aKQOc4n0]KPOf4o0ZKnNj4P1XKlNl4S1TKjNP5U1QKgNS5X1mJcNY5g0bIWOV1L^5k0aIUOQ1Je5o0Y2SOm0LVfl1"}, "label": "a blonde woman with a green headband is holding a black and white puppy on her shoulders"}]}
{"id": 59947, "image": "COCO_train2014_000000059947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bicycle chained up between 2 other bikes\"."}], "task": "refering", "answer_template": "The \"a red bicycle chained up between 2 other bikes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 150, 151, 152, 153, 167, 168, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 203, 204, 205, 206, 207, 208, 209, 221, 223, 224, 225, 226], "bbox": [0.29124, 0.5784745762711865, 0.6265799999999999, 0.9853954802259888], "iscrowd": 0, "area": 10254.304299999998, "rle": {"size": [354, 500], "counts": "jgb13m:3N3N1O1N3kFD`7=YHHh79UHGo5JfJb0WOET6IeJc0TODY6JbJc0ROD]6I`Jf0oNAc6J\\Jg0nN@`6GPI3`1h0mN^Oe6HmH2_1f1f5XNjH3]1e1k5bNRJ_1P6aNnI^1U6cNgI^1[6cNaI]1c6cNYI^1j6eNlH^1W7U12M3N2M3N2M3gNkGAX8`0mGnN_8R1i01O0O101N1OSOeF5Z9FlF9S9EPG<n8AVG?i8_OZGa0e8]O^Gd0a8ZOaGf0^8ZOdGf0Z8ZOgGf0X8ZOiGg0T8ZOnGf0P8YOSHh0j7XOWHi0g7WO[Hj0b7UOaHk0]7UOfHi0Y7VOkHh0T7XOoHf0P7YOSIg0k6YOWIf0h6YO[If0d6ZO^Id0a6]O`Ic0_6]ObIb0^6]OdIb0\\6^OeIa0[6_OfIa0Y6_OhI`0X6@iI`0V6@kI?^NAR70`J`0[NDS7LcJh0PN^O\\7JeJh0lM_O`7IdJi0jM_Ob7HdJT1\\5lNcJU1]5kNcJU1]5jNdJV1\\5jNcJV1e7M01N1O4M3M3M3M3N2M8H?`NcM`Ig1SOROY7ZO^Ic1BmNQ7AYIg1IcNn6GVIk1MYNn6MQIo14nMl65jHT2k7lMQHZ2l7SNVH`1i7]NYHe1e7YN]Hi1a7UNaHm1]7oMgHT2V7lMjHV2T7iMmHY2Q7gMoHY2Q7gMoHY2Q7fMPI[2o6eMQI[2o6eMQI[2o6dMRI\\2n6dMRI\\2n6cMSI]2m6cMSI]2m6bMTI^2h70O101O00000000000000001OO10000000000000000O1000000000000O100O100O100O10001N2O5J3N1N2O0O2O1N2O2M2N2N3M2N3M2N4J5J7Hi[P2"}, "label": "a red bicycle chained up between 2 other bikes"}]}
{"id": 59947, "image": "COCO_train2014_000000059947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bike is parked on the sidewalk\"."}], "task": "refering", "answer_template": "The \"a red bike is parked on the sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 150, 151, 152, 153, 167, 168, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 203, 204, 205, 206, 207, 208, 209, 221, 223, 224, 225, 226], "bbox": [0.29124, 0.5784745762711865, 0.6265799999999999, 0.9853954802259888], "iscrowd": 0, "area": 10254.304299999998, "rle": {"size": [354, 500], "counts": "jgb13m:3N3N1O1N3kFD`7=YHHh79UHGo5JfJb0WOET6IeJc0TODY6JbJc0ROD]6I`Jf0oNAc6J\\Jg0nN@`6GPI3`1h0mN^Oe6HmH2_1f1f5XNjH3]1e1k5bNRJ_1P6aNnI^1U6cNgI^1[6cNaI]1c6cNYI^1j6eNlH^1W7U12M3N2M3N2M3gNkGAX8`0mGnN_8R1i01O0O101N1OSOeF5Z9FlF9S9EPG<n8AVG?i8_OZGa0e8]O^Gd0a8ZOaGf0^8ZOdGf0Z8ZOgGf0X8ZOiGg0T8ZOnGf0P8YOSHh0j7XOWHi0g7WO[Hj0b7UOaHk0]7UOfHi0Y7VOkHh0T7XOoHf0P7YOSIg0k6YOWIf0h6YO[If0d6ZO^Id0a6]O`Ic0_6]ObIb0^6]OdIb0\\6^OeIa0[6_OfIa0Y6_OhI`0X6@iI`0V6@kI?^NAR70`J`0[NDS7LcJh0PN^O\\7JeJh0lM_O`7IdJi0jM_Ob7HdJT1\\5lNcJU1]5kNcJU1]5jNdJV1\\5jNcJV1e7M01N1O4M3M3M3M3N2M8H?`NcM`Ig1SOROY7ZO^Ic1BmNQ7AYIg1IcNn6GVIk1MYNn6MQIo14nMl65jHT2k7lMQHZ2l7SNVH`1i7]NYHe1e7YN]Hi1a7UNaHm1]7oMgHT2V7lMjHV2T7iMmHY2Q7gMoHY2Q7gMoHY2Q7fMPI[2o6eMQI[2o6eMQI[2o6dMRI\\2n6dMRI\\2n6cMSI]2m6cMSI]2m6bMTI^2h70O101O00000000000000001OO10000000000000000O1000000000000O100O100O100O10001N2O5J3N1N2O0O2O1N2O2M2N2N3M2N3M2N4J5J7Hi[P2"}, "label": "a red bike is parked on the sidewalk"}]}
{"id": 59947, "image": "COCO_train2014_000000059947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bike parked ahead of a red bike and a silver bike\"."}], "task": "refering", "answer_template": "The \"white bike parked ahead of a red bike and a silver bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 218, 219, 220, 221], "bbox": [0.0060999999999999995, 0.6297457627118644, 0.31670000000000004, 0.9802824858757062], "iscrowd": 0, "area": 8210.736900000002, "rle": {"size": [354, 500], "counts": "mZ16j:<E;D6J1O2O0O1O1O2O0O10010O01O00001O001O00000000000000XNiNfIW1\\6iNbIW1_6jN^IW1d6iNYIX1h6iNVIV1m6kNoHV1R7kNkHU1X7kNbHX1`7iNZH[1h7eNRH^1P8cNnG\\1U8dNiG\\1X8eNeG[1^8eN`Gh0MZOe8O\\Gd0[9\\O`Fg0a9<2N101O0101NUOZF;f9B]F?a9_ObFa0^9\\OeFe0Y9[OhFe0X9ZOiFg0U9YOlFh0S9WOnFi0Q9WOPGj0o8TOSGl0l8TOUGm0i8TOWGl0i8SOXGn0f8SOZGm0f8RO[Go0c8QO^GP1a8PO_GP1a8oN`GR1`8mN`GS15lN_70]HU10mN\\8S1bGnN_8S1]GoNd8S1WGoNj8R1SGoNn8S1mFoNT9_11O1N2N1N3N2M11N1O2O0O2N1O2O0[OZNnGh1P8[NmGg1R8ZNlGi1T8WNjGl1T8TNjGo1U8RNhG;JT1]8aNgG;NU1Z8aNfG92W1W8`NeG88X1R8nNPHS1o7kNSHV1l7hNWHX1i7dNZH]1e7aN]H`1b7^NaHb1^7^NbHc1]7\\NdHe1[7[NfHe1Y7[NfHg1Y7XNhHh1X7XNgHi1Z7VNfHk1Y7TNgHm1Y7SNgHn1Z7PNfHP2U8101N100O2O0O2O000O2O0O010O10gM\\GS2e8mM[GS2d8nM]GP2d8oM]GQ2b8PN_Go1a8QN_Gn1a8SN`Gl1`8SNaGl1`8TNaGk1^8VNbGj1^8VNcGh1]8XNdGh1\\8XNeGg1Z8ZNfGe1[8[NfGb1\\8^NeG_1\\8aNfG\\1\\8dNeGY1]8gNcGW1_8iNbGT1`8iNdGT1^8jNeGS1W9M3M3M3Cene3"}, "label": "white bike parked ahead of a red bike and a silver bike"}]}
{"id": 59947, "image": "COCO_train2014_000000059947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bike on left\"."}], "task": "refering", "answer_template": "The \"white bike on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 218, 219, 220, 221], "bbox": [0.0060999999999999995, 0.6297457627118644, 0.31670000000000004, 0.9802824858757062], "iscrowd": 0, "area": 8210.736900000002, "rle": {"size": [354, 500], "counts": "mZ16j:<E;D6J1O2O0O1O1O2O0O10010O01O00001O001O00000000000000XNiNfIW1\\6iNbIW1_6jN^IW1d6iNYIX1h6iNVIV1m6kNoHV1R7kNkHU1X7kNbHX1`7iNZH[1h7eNRH^1P8cNnG\\1U8dNiG\\1X8eNeG[1^8eN`Gh0MZOe8O\\Gd0[9\\O`Fg0a9<2N101O0101NUOZF;f9B]F?a9_ObFa0^9\\OeFe0Y9[OhFe0X9ZOiFg0U9YOlFh0S9WOnFi0Q9WOPGj0o8TOSGl0l8TOUGm0i8TOWGl0i8SOXGn0f8SOZGm0f8RO[Go0c8QO^GP1a8PO_GP1a8oN`GR1`8mN`GS15lN_70]HU10mN\\8S1bGnN_8S1]GoNd8S1WGoNj8R1SGoNn8S1mFoNT9_11O1N2N1N3N2M11N1O2O0O2N1O2O0[OZNnGh1P8[NmGg1R8ZNlGi1T8WNjGl1T8TNjGo1U8RNhG;JT1]8aNgG;NU1Z8aNfG92W1W8`NeG88X1R8nNPHS1o7kNSHV1l7hNWHX1i7dNZH]1e7aN]H`1b7^NaHb1^7^NbHc1]7\\NdHe1[7[NfHe1Y7[NfHg1Y7XNhHh1X7XNgHi1Z7VNfHk1Y7TNgHm1Y7SNgHn1Z7PNfHP2U8101N100O2O0O2O000O2O0O010O10gM\\GS2e8mM[GS2d8nM]GP2d8oM]GQ2b8PN_Go1a8QN_Gn1a8SN`Gl1`8SNaGl1`8TNaGk1^8VNbGj1^8VNcGh1]8XNdGh1\\8XNeGg1Z8ZNfGe1[8[NfGb1\\8^NeG_1\\8aNfG\\1\\8dNeGY1]8gNcGW1_8iNbGT1`8iNdGT1^8jNeGS1W9M3M3M3Cene3"}, "label": "white bike on left"}]}
{"id": 59947, "image": "COCO_train2014_000000059947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike behind the red bike\"."}], "task": "refering", "answer_template": "The \"the bike behind the red bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 159, 160, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 232, 233], "bbox": [0.61976, 0.647316384180791, 1.0, 0.987090395480226], "iscrowd": 0, "area": 11454.5506, "rle": {"size": [354, 500], "counts": "P_[3b0^:4K4M4K5M3M3M3M3M2O0O2N1O1O2O0O2N100O2O000O2O0O2O000O2O000000000000000bNPNiIQ2W6PNhIP2W6RNhIn1[6PNdIP2S6PNlH0P1P2S6RNkH0k0MSOR2X7RNfH2n0P2_6WN_Ii1c6]NTId1n6\\NPIc1R7]NjHf1X7YNeHi1]7VNaHj1`7VN_Hk1b7TN^Hk1d7UN[Hk1f7TNUHN5]1f7eNTHK:_1d7oN`HP1`7nNcHo0`7oNaHP1a7oN`Ho0a7RO^Hm0c7TO]Hj0d7WO[Hg0h7fNiG;?m0Q8SOoGl0R8UOnGi0S8XOmGf0T8[OmGb0U8^OnG<T8EoG6R8JQH0R81QHGS8:PH[OW8f0R11O00100O1O1O1O001O1O1O1O1O00100O1O1O1O001O1O1O1O1O00100bFlNl8U1QGnNn8S1oFQOo8b1O001O001O0000001O01O00000000SOQNgHo1Y7QNgHo1Y7PNhHP2X7PNhHP2X7PNhHo1Y7RNfHm1[7SNcHn1^7RN^HP2d7PNYHn1l7RNPHS2o7PNiG4Ne1X8cNoGZ1P8dNVHY1i7dN^HY1a7dNcH]1[7bNgH^1X7bNiH^1V7bNkH_1S7aNnH_1Q7aNPI_1o6`NTI^1l6bNUI^1j6bNWI^1h6bNXI^1h6bNWI`1h6^NTHKT1h1g6]NUHKT1i1f6\\NVHKT1i1f6[NWHLR1k1f6YN]Ih1b6WN_Ii1m71O1N101O1O1O0O2O1O001O1O0000000000000001O0000001O00001O001O000O2O0O2O0O2O1N2O0O2O1NTG"}, "label": "the bike behind the red bike"}]}
{"id": 59947, "image": "COCO_train2014_000000059947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the last bike on the right\"."}], "task": "refering", "answer_template": "The \"the last bike on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 159, 160, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 232, 233], "bbox": [0.61976, 0.647316384180791, 1.0, 0.987090395480226], "iscrowd": 0, "area": 11454.5506, "rle": {"size": [354, 500], "counts": "P_[3b0^:4K4M4K5M3M3M3M3M2O0O2N1O1O2O0O2N100O2O000O2O0O2O000O2O000000000000000bNPNiIQ2W6PNhIP2W6RNhIn1[6PNdIP2S6PNlH0P1P2S6RNkH0k0MSOR2X7RNfH2n0P2_6WN_Ii1c6]NTId1n6\\NPIc1R7]NjHf1X7YNeHi1]7VNaHj1`7VN_Hk1b7TN^Hk1d7UN[Hk1f7TNUHN5]1f7eNTHK:_1d7oN`HP1`7nNcHo0`7oNaHP1a7oN`Ho0a7RO^Hm0c7TO]Hj0d7WO[Hg0h7fNiG;?m0Q8SOoGl0R8UOnGi0S8XOmGf0T8[OmGb0U8^OnG<T8EoG6R8JQH0R81QHGS8:PH[OW8f0R11O00100O1O1O1O001O1O1O1O1O00100O1O1O1O001O1O1O1O1O00100bFlNl8U1QGnNn8S1oFQOo8b1O001O001O0000001O01O00000000SOQNgHo1Y7QNgHo1Y7PNhHP2X7PNhHP2X7PNhHo1Y7RNfHm1[7SNcHn1^7RN^HP2d7PNYHn1l7RNPHS2o7PNiG4Ne1X8cNoGZ1P8dNVHY1i7dN^HY1a7dNcH]1[7bNgH^1X7bNiH^1V7bNkH_1S7aNnH_1Q7aNPI_1o6`NTI^1l6bNUI^1j6bNWI^1h6bNXI^1h6bNWI`1h6^NTHKT1h1g6]NUHKT1i1f6\\NVHKT1i1f6[NWHLR1k1f6YN]Ih1b6WN_Ii1m71O1N101O1O1O0O2O1O001O1O0000000000000001O0000001O00001O001O000O2O0O2O0O2O1N2O0O2O1NTG"}, "label": "the last bike on the right"}]}
{"id": 68139, "image": "COCO_train2014_000000068139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food truck called docs\"."}], "task": "refering", "answer_template": "The \"a food truck called docs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 234], "bbox": [0.4228267973856209, 0.127859477124183, 0.8522712418300652, 0.4636928104575163], "iscrowd": 0, "area": 39379.36215000001, "rle": {"size": [612, 612], "counts": "ooj45ib07H8I7H7J7H8L4[OaNc_O`1Q`0mNn_OR1h?YOW@h0]?Db@=\\?Dd@=Z?Ee@<Z?Df@=X?Dh@<W?Fh@;W?Ej@;T?Gk@:T?Fl@;R?Gm@>m>CSAd0f>\\OZAk0^>WOaAj0^>VObAj0]>XOcAh0[>YOeAg0[>YOeAg0Z>[OeAe0[>[OeAf0Y>\\OfAd0Y>]OgAc0Y>]OgAc0V>_MfAP24a0U>AkA?U>BjA>V>BjA>V>CiA>V>BjA>V>BjA>V>CiA=W>CjA=V>CiA=W>CiA=W>CiA=W>DhA=W>CiA=W>X2000\\MiA;W>Z2O0000001O00000010OWOSK[Cm4e<TKZCl4f<UKYCl4f<UKYCk4g<VKXCj4h<WKWCi4j<XKTCi4k<XKTCh4l<YKSCg4m<ZKRCf4n<[KQCf4n<[KQCe4o<\\KPCd4P=]KoBc4Q=_KmBb4R=_KmBa4S=`KlB`4T=aKlB^4T=cKkB^4T=cKkB]4U=dKjB\\4V=eKiB[4W=i001O0001O01O0000001O0000001O0000001O0000001O0000010O0000001O0000001O0000001O0000001O000001O000000000000001O1O1N1O2O1O010O01O1O1O001O00XMRC[Om<f0RC[Om<e0SC[Om<e0TC[Oj<g0UCYOk<g0VCYOj<f0VCYOk<g0UCXOl<h0TCXOl<h0TCWOm<i0SCVOn<j0RCVOn<j0RCUOo<k0QCUOo<k0QCTOP=l0PCSOR=l0nBTOR=l0nBTOR=l0nBSOS=m0mBSOS=m0mBROT=n0lBROT=n0lBROU=m0kBQOW=o0iBnN[=Q1eBmN]=S1cBlN_=S1aBlN`=T1`BkNb=T1^BlNb=T1^BlNc=R1^BnNc=P1^BPOb=o0_BQOb=n0^B`NT>_1i1N2L4N2N3M2O10000O110O000002N4L9H1N1O1SAiMk<X2fBVN[=j1cBWN_=h1_BYNc=f1\\BZNf=d1YB]Ni=l1kAUNW>k1fAVN[>j1cAWN]>X3O2O000000000O100000000O100000000O11O2N2O0O2N2N1O2N1O2N2N1O2N3M3M3M2O0O1O1N1N3M3K4\\Nl@UOX?h0n@POW?f0i1L3D=L4K4M4L=C[de1"}, "label": "a food truck called docs"}]}
{"id": 68139, "image": "COCO_train2014_000000068139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"doc ' s foodtruck\"."}], "task": "refering", "answer_template": "The \"doc ' s foodtruck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 234], "bbox": [0.4228267973856209, 0.127859477124183, 0.8522712418300652, 0.4636928104575163], "iscrowd": 0, "area": 39379.36215000001, "rle": {"size": [612, 612], "counts": "ooj45ib07H8I7H7J7H8L4[OaNc_O`1Q`0mNn_OR1h?YOW@h0]?Db@=\\?Dd@=Z?Ee@<Z?Df@=X?Dh@<W?Fh@;W?Ej@;T?Gk@:T?Fl@;R?Gm@>m>CSAd0f>\\OZAk0^>WOaAj0^>VObAj0]>XOcAh0[>YOeAg0[>YOeAg0Z>[OeAe0[>[OeAf0Y>\\OfAd0Y>]OgAc0Y>]OgAc0V>_MfAP24a0U>AkA?U>BjA>V>BjA>V>CiA>V>BjA>V>BjA>V>CiA=W>CjA=V>CiA=W>CiA=W>CiA=W>DhA=W>CiA=W>X2000\\MiA;W>Z2O0000001O00000010OWOSK[Cm4e<TKZCl4f<UKYCl4f<UKYCk4g<VKXCj4h<WKWCi4j<XKTCi4k<XKTCh4l<YKSCg4m<ZKRCf4n<[KQCf4n<[KQCe4o<\\KPCd4P=]KoBc4Q=_KmBb4R=_KmBa4S=`KlB`4T=aKlB^4T=cKkB^4T=cKkB]4U=dKjB\\4V=eKiB[4W=i001O0001O01O0000001O0000001O0000001O0000001O0000010O0000001O0000001O0000001O0000001O000001O000000000000001O1O1N1O2O1O010O01O1O1O001O00XMRC[Om<f0RC[Om<e0SC[Om<e0TC[Oj<g0UCYOk<g0VCYOj<f0VCYOk<g0UCXOl<h0TCXOl<h0TCWOm<i0SCVOn<j0RCVOn<j0RCUOo<k0QCUOo<k0QCTOP=l0PCSOR=l0nBTOR=l0nBTOR=l0nBSOS=m0mBSOS=m0mBROT=n0lBROT=n0lBROU=m0kBQOW=o0iBnN[=Q1eBmN]=S1cBlN_=S1aBlN`=T1`BkNb=T1^BlNb=T1^BlNc=R1^BnNc=P1^BPOb=o0_BQOb=n0^B`NT>_1i1N2L4N2N3M2O10000O110O000002N4L9H1N1O1SAiMk<X2fBVN[=j1cBWN_=h1_BYNc=f1\\BZNf=d1YB]Ni=l1kAUNW>k1fAVN[>j1cAWN]>X3O2O000000000O100000000O100000000O11O2N2O0O2N2N1O2N1O2N2N1O2N3M3M3M2O0O1O1N1N3M3K4\\Nl@UOX?h0n@POW?f0i1L3D=L4K4M4L=C[de1"}, "label": "doc ' s foodtruck"}]}
{"id": 428587, "image": "COCO_train2014_000000428587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the front of the line\"."}], "task": "refering", "answer_template": "The \"the elephant in the front of the line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245], "bbox": [0.41375, 0.44154166666666667, 0.6597500000000001, 0.6480625], "iscrowd": 0, "area": 9453.433100000002, "rle": {"size": [480, 640], "counts": "``l33j>4M3L4L3M3F:UOj0B?O1N2O1N101N2O1NHhClMW<Q2=N30O01O10O01O10O01O10O01O10O013M2M4M2N2M4M2N3L3N3M2M2O000O10000OQOUDkNk;U1[DeNe;[1`D`Nc;]1cD^Nc;[1aDaN_;_1aDaN_;_1bD`N^;`1cD_N];`1dD^N^;b1cD\\N^;d1n01000000000O100000O100000O100000000000000O100001O1O1O1O010cCYN_;h1XD`Nh;a1oChNP<T20000000000N3N1O11O001O0010O01]NnCJ0VOR<o0WDBNYOl;T1^DYOQ<f0XDPOk;o0R10100O010O10O0L6N2O0O10000O10000O100001O001O001O001O0010O100O100Ob0_O3_CTNh;n1SDRNR<o1hCRN\\<n1^CSNg<Q24L3L5L4K`0Ac0\\O<EjlU3"}, "label": "the elephant in the front of the line"}]}
{"id": 428587, "image": "COCO_train2014_000000428587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with a red cloth draped over its back , carring a couple wearing white shirts\"."}], "task": "refering", "answer_template": "The \"elephant with a red cloth draped over its back , carring a couple wearing white shirts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245], "bbox": [0.41375, 0.44154166666666667, 0.6597500000000001, 0.6480625], "iscrowd": 0, "area": 9453.433100000002, "rle": {"size": [480, 640], "counts": "``l33j>4M3L4L3M3F:UOj0B?O1N2O1N101N2O1NHhClMW<Q2=N30O01O10O01O10O01O10O01O10O013M2M4M2N2M4M2N3L3N3M2M2O000O10000OQOUDkNk;U1[DeNe;[1`D`Nc;]1cD^Nc;[1aDaN_;_1aDaN_;_1bD`N^;`1cD_N];`1dD^N^;b1cD\\N^;d1n01000000000O100000O100000O100000000000000O100001O1O1O1O010cCYN_;h1XD`Nh;a1oChNP<T20000000000N3N1O11O001O0010O01]NnCJ0VOR<o0WDBNYOl;T1^DYOQ<f0XDPOk;o0R10100O010O10O0L6N2O0O10000O10000O100001O001O001O001O0010O100O100Ob0_O3_CTNh;n1SDRNR<o1hCRN\\<n1^CSNg<Q24L3L5L4K`0Ac0\\O<EjlU3"}, "label": "elephant with a red cloth draped over its back , carring a couple wearing white shirts"}]}
{"id": 494128, "image": "COCO_train2014_000000494128.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red snowboard held by a man with another man upside down\"."}], "task": "refering", "answer_template": "The \"a red snowboard held by a man with another man upside down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 209, 210], "bbox": [0.07584375, 0.34568750000000004, 0.47359375000000004, 0.5614166666666667], "iscrowd": 0, "area": 11559.9006, "rle": {"size": [480, 640], "counts": "^Vg0?^>5L3N2M4M2M3N3L3N2M4M2M3N1N2O1O01000000O01000O100000O01000000O1000O10O1000000O0100000O1000O10O10000O10O1000O100000O01000000O10O1000O1000000O01000O100000O01000000O10O1000O10000O0100000O1000O10O10000O10O1000O100000O01000000O1000O01000000O10O1000O10000O0100000O1000O10O1000000O0100000O1000O10O100000000O010000000O10O1000O100000O10O10O100O10O10O1000O010000O10O10O1000O010000O10O10O10000O01000O10O10O10000O01000O10O010000O10O10O10000O01000O10O10O10000O01000O10O10O1000O2O001O0O101O0O2O00001N1N3M2N2N5K4K6K4L5K5KVim4"}, "label": "a red snowboard held by a man with another man upside down"}]}
{"id": 494128, "image": "COCO_train2014_000000494128.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red snowboard on a man ' s feet\"."}], "task": "refering", "answer_template": "The \"a red snowboard on a man ' s feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 209, 210], "bbox": [0.07584375, 0.34568750000000004, 0.47359375000000004, 0.5614166666666667], "iscrowd": 0, "area": 11559.9006, "rle": {"size": [480, 640], "counts": "^Vg0?^>5L3N2M4M2M3N3L3N2M4M2M3N1N2O1O01000000O01000O100000O01000000O1000O10O1000000O0100000O1000O10O10000O10O1000O100000O01000000O10O1000O1000000O01000O100000O01000000O10O1000O10000O0100000O1000O10O10000O10O1000O100000O01000000O1000O01000000O10O1000O10000O0100000O1000O10O1000000O0100000O1000O10O100000000O010000000O10O1000O100000O10O10O100O10O10O1000O010000O10O10O1000O010000O10O10O10000O01000O10O10O10000O01000O10O010000O10O10O10000O01000O10O10O10000O01000O10O10O1000O2O001O0O101O0O2O00001N1N3M2N2N5K4K6K4L5K5KVim4"}, "label": "a red snowboard on a man ' s feet"}]}
{"id": 485940, "image": "COCO_train2014_000000485940.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cream colored laundry bag hanging in front of the red bag\"."}], "task": "refering", "answer_template": "The \"cream colored laundry bag hanging in front of the red bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 90, 91, 113, 114, 136, 159, 160, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 272, 273, 274, 294, 295, 296, 297, 318, 319, 320, 341, 342, 343, 365, 366], "bbox": [0.818421875, 0.049937499999999996, 0.987484375, 0.9274166666666666], "iscrowd": 0, "area": 18510.717950000002, "rle": {"size": [480, 640], "counts": "fne77e>9G9F;F9G9H5K5K5K5K5K5L4K5K9G?A4L4M4K4L3M4L3M4L3M4K5L3L5L3M4K4M4K5L3L5]Ob0O2N101O001O1O0O2N100O2O0O2O0O1O2O0010O0001O0010O01O0L4L5J5K6J5K5K6J5K6K4K5K6J5K6J5K5K6K4K6J5LoLnK\\Kn3\\4aLbKZ3V4VMeIgN]1o3f4jMeIbN_1_3j0cLh0l1gL]N`1P3b0`Me0`2RN<k0eMh0b2hM3R1lMl0l6oN[Ig0m6YOVI<R7CRI2V7NmHG[79gH]Oa7c0bHROf7m0U30000O10000O8I7I7I7HQg3"}, "label": "cream colored laundry bag hanging in front of the red bag"}]}
{"id": 485940, "image": "COCO_train2014_000000485940.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear bag hanging on wall\"."}], "task": "refering", "answer_template": "The \"clear bag hanging on wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 90, 91, 113, 114, 136, 159, 160, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 272, 273, 274, 294, 295, 296, 297, 318, 319, 320, 341, 342, 343, 365, 366], "bbox": [0.818421875, 0.049937499999999996, 0.987484375, 0.9274166666666666], "iscrowd": 0, "area": 18510.717950000002, "rle": {"size": [480, 640], "counts": "fne77e>9G9F;F9G9H5K5K5K5K5K5L4K5K9G?A4L4M4K4L3M4L3M4L3M4K5L3L5L3M4K4M4K5L3L5]Ob0O2N101O001O1O0O2N100O2O0O2O0O1O2O0010O0001O0010O01O0L4L5J5K6J5K5K6J5K6K4K5K6J5K6J5K5K6K4K6J5LoLnK\\Kn3\\4aLbKZ3V4VMeIgN]1o3f4jMeIbN_1_3j0cLh0l1gL]N`1P3b0`Me0`2RN<k0eMh0b2hM3R1lMl0l6oN[Ig0m6YOVI<R7CRI2V7NmHG[79gH]Oa7c0bHROf7m0U30000O10000O8I7I7I7HQg3"}, "label": "clear bag hanging on wall"}]}
{"id": 182837, "image": "COCO_train2014_000000182837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfer that is not lying on his board\"."}], "task": "refering", "answer_template": "The \"the surfer that is not lying on his board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 165, 166, 188, 189, 190, 210, 211, 212, 213, 214, 234, 235, 236, 237], "bbox": [0.16609374999999998, 0.42533957845433257, 0.357265625, 0.7138173302107729], "iscrowd": 0, "area": 6150.990249999999, "rle": {"size": [427, 640], "counts": "bl\\17P=5K5L5K4lNJnD:l:V1L4L4L4O1N3N0O2VOeMbF[2\\9hMcFX2\\9jMbFW2\\9lMcFU2[9lMdFU2[9mMcFT2]9nM`FS2`9nM_FR2`9PN^FR2a9l00000O101O00O100000000O10000ZM_Fb1b9[NaFc1`9]N`F<6AZ93`F6?DR94`F5c0Em86aF3d0Fk87aF3d0Em87_F3e0FS9_OSF97b0c0F\\9^OXFk0<H^9ZOXFc0d03f9MYF3i9NUF2k9OSF1o90oE0Q:1mEOU:1jEOW:1gEO[:1dEO_:0^E0e:OZE1h:NWE1l:NRE2R;LmD4U;LiD3Y;MfD3\\;LcD3`;L^D5c;J]D5f;IZD7g;IXD6j;IVD7j;IUD7m;HSD8m;HSD7o;IPD6Q<JoC7P<IoC9Q<FoC;P<FoC;Q<DPD=n;CRD>n;ARDa0l;_OTDb0l;^OTDb0k;^OWDb0h;]OZDb0e;^O]Da0c;_O^D`0a;@`D`0_;@cD=_;BbD;`;F`D8b;G_D6c;J^D4d;K]D2e;N\\D0f;0ZDNg;2ZDLf;5h0000O1000O10O10O1000000O1000000000O01000000MlBOT=0mB0S=0fa[5"}, "label": "the surfer that is not lying on his board"}]}
{"id": 182837, "image": "COCO_train2014_000000182837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black and blue wet suit\"."}], "task": "refering", "answer_template": "The \"a woman in a black and blue wet suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 165, 166, 188, 189, 190, 210, 211, 212, 213, 214, 234, 235, 236, 237], "bbox": [0.16609374999999998, 0.42533957845433257, 0.357265625, 0.7138173302107729], "iscrowd": 0, "area": 6150.990249999999, "rle": {"size": [427, 640], "counts": "bl\\17P=5K5L5K4lNJnD:l:V1L4L4L4O1N3N0O2VOeMbF[2\\9hMcFX2\\9jMbFW2\\9lMcFU2[9lMdFU2[9mMcFT2]9nM`FS2`9nM_FR2`9PN^FR2a9l00000O101O00O100000000O10000ZM_Fb1b9[NaFc1`9]N`F<6AZ93`F6?DR94`F5c0Em86aF3d0Fk87aF3d0Em87_F3e0FS9_OSF97b0c0F\\9^OXFk0<H^9ZOXFc0d03f9MYF3i9NUF2k9OSF1o90oE0Q:1mEOU:1jEOW:1gEO[:1dEO_:0^E0e:OZE1h:NWE1l:NRE2R;LmD4U;LiD3Y;MfD3\\;LcD3`;L^D5c;J]D5f;IZD7g;IXD6j;IVD7j;IUD7m;HSD8m;HSD7o;IPD6Q<JoC7P<IoC9Q<FoC;P<FoC;Q<DPD=n;CRD>n;ARDa0l;_OTDb0l;^OTDb0k;^OWDb0h;]OZDb0e;^O]Da0c;_O^D`0a;@`D`0_;@cD=_;BbD;`;F`D8b;G_D6c;J^D4d;K]D2e;N\\D0f;0ZDNg;2ZDLf;5h0000O1000O10O10O1000000O1000000000O01000000MlBOT=0mB0S=0fa[5"}, "label": "a woman in a black and blue wet suit"}]}
{"id": 346678, "image": "COCO_train2014_000000346678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a climbing bear\"."}], "task": "refering", "answer_template": "The \"a climbing bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 356, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.448640625, 0.55475, 0.8651250000000001, 0.9881458333333332], "iscrowd": 0, "area": 35331.577950000006, "rle": {"size": [480, 640], "counts": "clV41l>6J6K5J6K5J6M3L3M4M3L4M2M3N2M3M3N2M3N2N2O1N2O2M2O1N100O100O010O1000O0100O10O010000O010O100O01001M4L4L4L3M4L4L2N1O2N2N1O100O1O1O1O100O1O1O1O1O1WNSL_Hm3_7VL_Hk3^7ZL_Hg3_7]L^Hd3`7`L]Ha3a7cL]H]3a7gL\\HZ3a7jL]HW3a7mL\\HT3e7mLXHT3h7nLVHR3j7PMSHQ3n7PMoGR3P8oLnGR3S8SMfGn2Z8YM_Gg2a8l11O00001O00001O00001O00001O00000000000000O10000000000000000000000000000O100000000000000000000O100O10000O100O10000O100O10000O100O1001O000000001O000000001O0000001O0001O0001O0000001Oh0XO2N2N2N2N1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2QOjEPMW:e2RF\\Mo9X2\\FhMe9d1mF]NT9U1XGlNj8S1UGmNl8R1SGoNn8Q1PGPOQ9P1nFPOR9P1mFQOT9o0kFQOV9n0iFSOX9m0fFTO[9k0eFUO[9l0cFUO^9k0`FVOa9i0_FWOb9i0\\FXOd9h0[FYOf9g0YFYOh9f0XFZOi9f0WFYOi9h0VFXOk9g0VFXOk9h0TFXOm9g0TFXOl9i0SFWOm9i0TFVOk9l0TFTOk9n0UFQOj9P1VFPOj9Q1VFnNi9S1WFmNh9T1YFkNf9V1[FiNd9W1_FgNa9Y1aFeN`9Z1bFdN^9\\1eFaN[9_1gF_NZ9_1iF_NW9a1lF\\NT9d1nFZNR9f1PGXNQ9g1RGVNn8i1U20001O0O2O001O1O0O2O1O0O2M3M2O2M2N3Nn0QOPnW1"}, "label": "a climbing bear"}]}
{"id": 346678, "image": "COCO_train2014_000000346678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear climbing rocks\"."}], "task": "refering", "answer_template": "The \"a bear climbing rocks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 356, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.448640625, 0.55475, 0.8651250000000001, 0.9881458333333332], "iscrowd": 0, "area": 35331.577950000006, "rle": {"size": [480, 640], "counts": "clV41l>6J6K5J6K5J6M3L3M4M3L4M2M3N2M3M3N2M3N2N2O1N2O2M2O1N100O100O010O1000O0100O10O010000O010O100O01001M4L4L4L3M4L4L2N1O2N2N1O100O1O1O1O100O1O1O1O1O1WNSL_Hm3_7VL_Hk3^7ZL_Hg3_7]L^Hd3`7`L]Ha3a7cL]H]3a7gL\\HZ3a7jL]HW3a7mL\\HT3e7mLXHT3h7nLVHR3j7PMSHQ3n7PMoGR3P8oLnGR3S8SMfGn2Z8YM_Gg2a8l11O00001O00001O00001O00001O00000000000000O10000000000000000000000000000O100000000000000000000O100O10000O100O10000O100O10000O100O1001O000000001O000000001O0000001O0001O0001O0000001Oh0XO2N2N2N2N1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2QOjEPMW:e2RF\\Mo9X2\\FhMe9d1mF]NT9U1XGlNj8S1UGmNl8R1SGoNn8Q1PGPOQ9P1nFPOR9P1mFQOT9o0kFQOV9n0iFSOX9m0fFTO[9k0eFUO[9l0cFUO^9k0`FVOa9i0_FWOb9i0\\FXOd9h0[FYOf9g0YFYOh9f0XFZOi9f0WFYOi9h0VFXOk9g0VFXOk9h0TFXOm9g0TFXOl9i0SFWOm9i0TFVOk9l0TFTOk9n0UFQOj9P1VFPOj9Q1VFnNi9S1WFmNh9T1YFkNf9V1[FiNd9W1_FgNa9Y1aFeN`9Z1bFdN^9\\1eFaN[9_1gF_NZ9_1iF_NW9a1lF\\NT9d1nFZNR9f1PGXNQ9g1RGVNn8i1U20001O0O2O001O1O0O2O1O0O2M3M2O2M2N3Nn0QOPnW1"}, "label": "a bear climbing rocks"}]}
{"id": 346678, "image": "COCO_train2014_000000346678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear left\"."}], "task": "refering", "answer_template": "The \"bear left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 216, 217, 218, 235, 236], "bbox": [0.047093750000000004, 0.1585, 0.541, 0.6187083333333334], "iscrowd": 0, "area": 44783.660200000006, "rle": {"size": [480, 640], "counts": "[V>4a>=C>D<E:J7M1N101O000O101O000O2O006J5Jc0^O10O10000000000000O1N2N2O1N010O01O010O10O2O1O1N2O1O1N10O00100O001O100O1O1O1O100001O1O2N1O2N1O1O2N1O2N1OO1O1O0O2O1O2N1N2M3M3M4M2M3O12N2N2M2O2M3N2N1N3bFcLU7_3gHcLX7`3cHcL]7_3_HcL`7`3[HcLe7_3XHbLh7`3THbLk7`3SHaLm7a3oGaLP8b3mG_LS8c3jG^LU8e3gG]LY8f3cG[L]8i3]GYLb8l3YGTLh8P4RGRLm8S4nFnKR9P5O1000O0100O10O010000O010001O0O10001OnFeJ_8T6J5K1O01O0000001O0000001O0000001O0000001O001O1O1O1O1O1O1O1O000VOdGPK\\8j4WHiJj7o4W1J6L41O0000000000000000O100000000O100000000O10000000000O1000000000O100000000000000000000000001O00001O1O1O1O001O1O1O1O1N101O100O1O1O00100gFhJh8Y5UGiJk8X5RGiJo8X5mFkJS9T5lFnJT9R5jFPKW9[5100000001N1000000O2O0000000O2O00000O101O00001N2O1O1O1N2O1O1O1O1N2O001O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O001N2O1O1O1O1O1RO[E^Mg:^2_E_Mb:]2cEaM^:\\2fEbM[:Z2kEcMV:Y2oEeMR:W2SFfMo9V2WFeMl9X2VFgMl9U2WFjMj9S2d1L6K4K6K5J6K4L5H8I7I6ImdY4"}, "label": "bear left"}]}
{"id": 346678, "image": "COCO_train2014_000000346678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear standing in the water with a fish in its mouth\"."}], "task": "refering", "answer_template": "The \"a bear standing in the water with a fish in its mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 216, 217, 218, 235, 236], "bbox": [0.047093750000000004, 0.1585, 0.541, 0.6187083333333334], "iscrowd": 0, "area": 44783.660200000006, "rle": {"size": [480, 640], "counts": "[V>4a>=C>D<E:J7M1N101O000O101O000O2O006J5Jc0^O10O10000000000000O1N2N2O1N010O01O010O10O2O1O1N2O1O1N10O00100O001O100O1O1O1O100001O1O2N1O2N1O1O2N1O2N1OO1O1O0O2O1O2N1N2M3M3M4M2M3O12N2N2M2O2M3N2N1N3bFcLU7_3gHcLX7`3cHcL]7_3_HcL`7`3[HcLe7_3XHbLh7`3THbLk7`3SHaLm7a3oGaLP8b3mG_LS8c3jG^LU8e3gG]LY8f3cG[L]8i3]GYLb8l3YGTLh8P4RGRLm8S4nFnKR9P5O1000O0100O10O010000O010001O0O10001OnFeJ_8T6J5K1O01O0000001O0000001O0000001O0000001O001O1O1O1O1O1O1O1O000VOdGPK\\8j4WHiJj7o4W1J6L41O0000000000000000O100000000O100000000O10000000000O1000000000O100000000000000000000000001O00001O1O1O1O001O1O1O1O1N101O100O1O1O00100gFhJh8Y5UGiJk8X5RGiJo8X5mFkJS9T5lFnJT9R5jFPKW9[5100000001N1000000O2O0000000O2O00000O101O00001N2O1O1O1N2O1O1O1O1N2O001O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O001N2O1O1O1O1O1RO[E^Mg:^2_E_Mb:]2cEaM^:\\2fEbM[:Z2kEcMV:Y2oEeMR:W2SFfMo9V2WFeMl9X2VFgMl9U2WFjMj9S2d1L6K4K6K5J6K4L5H8I7I6ImdY4"}, "label": "a bear standing in the water with a fish in its mouth"}]}
{"id": 199225, "image": "COCO_train2014_000000199225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid a flannel shirt riding a skateboard\"."}], "task": "refering", "answer_template": "The \"a kid a flannel shirt riding a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283], "bbox": [0.19534374999999998, 0.07770491803278688, 0.35012499999999996, 0.8682435597189696], "iscrowd": 0, "area": 21767.59915, "rle": {"size": [427, 640], "counts": "lYd16m<<C<D<D=D;D<D=D9F:F:M3M2O2M3N2N2N101N2N2O1N2N6K5J5K6K5kJQK?J_MZ5Y1bKU1_OXMY5e0mKR2POTM^5IXLQ3_NRMP8k2UHRMk7m2ZHnLh7o2[1O1N3O001N101O000O2O00000O010001O0000000000001O000000001O2N1O2jFPMm7Q3PHRMo7P3mGSMP8P3nGQMP8S3lGPMR8R3kGQMR8S3kGoLS8S3jGoLT8V4L4I7F9E<D<D<D;E<D<D;E<D<D<M2O3M3M2N3M3M9G9G9F9G:A?A_Z]5"}, "label": "a kid a flannel shirt riding a skateboard"}]}
{"id": 199225, "image": "COCO_train2014_000000199225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black and white checkered shirt with black pants and a grey hat and tennis shoes standing on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man wearing a black and white checkered shirt with black pants and a grey hat and tennis shoes standing on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283], "bbox": [0.19534374999999998, 0.07770491803278688, 0.35012499999999996, 0.8682435597189696], "iscrowd": 0, "area": 21767.59915, "rle": {"size": [427, 640], "counts": "lYd16m<<C<D<D=D;D<D=D9F:F:M3M2O2M3N2N2N101N2N2O1N2N6K5J5K6K5kJQK?J_MZ5Y1bKU1_OXMY5e0mKR2POTM^5IXLQ3_NRMP8k2UHRMk7m2ZHnLh7o2[1O1N3O001N101O000O2O00000O010001O0000000000001O000000001O2N1O2jFPMm7Q3PHRMo7P3mGSMP8P3nGQMP8S3lGPMR8R3kGQMR8S3kGoLS8S3jGoLT8V4L4I7F9E<D<D<D;E<D<D;E<D<D<M2O3M3M2N3M3M9G9G9F9G:A?A_Z]5"}, "label": "a man wearing a black and white checkered shirt with black pants and a grey hat and tennis shoes standing on a skateboard"}]}
{"id": 109114, "image": "COCO_train2014_000000109114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwich hidden behind a napkin\"."}], "task": "refering", "answer_template": "The \"half of a sandwich hidden behind a napkin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 16, 17, 18, 22, 23, 24, 25, 26, 27, 28, 29, 30, 38, 39, 40, 45, 46, 47, 48, 49, 50, 51, 52, 61, 62, 63, 67, 68, 69, 70, 71, 72, 73, 74, 84, 85, 92, 93, 94, 95, 96, 116, 117, 118], "bbox": [0.0, 0.0, 0.8716176470588235, 0.2590032679738562], "iscrowd": 0, "area": 28946.85734999999, "rle": {"size": [612, 612], "counts": "31Sc03M3M3M3M2N3L4M3M3M3M1O1O00000V^OPO`a0P1_^OSO_a0m0a^OVO\\a0k0b^OXO\\a0h0d^O[OYa0e0f^O^OXa0b0h^O_OWa0a0h^OAWa0?h^OBXa0?g^OAYa0?f^OCYa0=g^OCYa0U1O001O000O2O001O001O001O001O3M2N3M2N3M2N3M2N1O1O001O1O1O001N2O1O1O001O1O1O001O1O1O001O01O01O01O01O00010O0001O0O100O2O000O2O0O101O0O10001N100O2O000O101N01000O100O10000O10S@^MX?a2g@aMY?_2e@dMZ?\\2d@fM\\?Y2c@jM\\?V2b@lM^?T2`@oM^?Q2a@QN_?o1_@TN`?k1_@WNa?i1]@ZNb?d201O001O001O001O001O0O101O001O001O001O001O001O001N101O001O001O001O001O00001O001N101O001O001O001O001O001O001O0O2O001O00001O001O001O001O001N101O001O001O001O001O001O00000O100000000000000000000000000O1005K:E:G:E;E:F;F9F;E;E6J7J5J6J7I6K5J7I6JiZV43UeiK9G:F9G:F9G:F3M001O001O1O001O001O001O001O001O001O001N101O001O001O001O1O001O001O001O0100O2O0O3N2M3N2M3N2M3N2M3N2M4M2N2M3N2M3N2M3N<C>CTT_1"}, "label": "half of a sandwich hidden behind a napkin"}]}
{"id": 109114, "image": "COCO_train2014_000000109114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of sandwich under napkin on a white plate\"."}], "task": "refering", "answer_template": "The \"half of sandwich under napkin on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 16, 17, 18, 22, 23, 24, 25, 26, 27, 28, 29, 30, 38, 39, 40, 45, 46, 47, 48, 49, 50, 51, 52, 61, 62, 63, 67, 68, 69, 70, 71, 72, 73, 74, 84, 85, 92, 93, 94, 95, 96, 116, 117, 118], "bbox": [0.0, 0.0, 0.8716176470588235, 0.2590032679738562], "iscrowd": 0, "area": 28946.85734999999, "rle": {"size": [612, 612], "counts": "31Sc03M3M3M3M2N3L4M3M3M3M1O1O00000V^OPO`a0P1_^OSO_a0m0a^OVO\\a0k0b^OXO\\a0h0d^O[OYa0e0f^O^OXa0b0h^O_OWa0a0h^OAWa0?h^OBXa0?g^OAYa0?f^OCYa0=g^OCYa0U1O001O000O2O001O001O001O001O3M2N3M2N3M2N3M2N1O1O001O1O1O001N2O1O1O001O1O1O001O1O1O001O01O01O01O01O00010O0001O0O100O2O000O2O0O101O0O10001N100O2O000O101N01000O100O10000O10S@^MX?a2g@aMY?_2e@dMZ?\\2d@fM\\?Y2c@jM\\?V2b@lM^?T2`@oM^?Q2a@QN_?o1_@TN`?k1_@WNa?i1]@ZNb?d201O001O001O001O001O0O101O001O001O001O001O001O001N101O001O001O001O001O00001O001N101O001O001O001O001O001O001O0O2O001O00001O001O001O001O001N101O001O001O001O001O001O00000O100000000000000000000000000O1005K:E:G:E;E:F;F9F;E;E6J7J5J6J7I6K5J7I6JiZV43UeiK9G:F9G:F9G:F3M001O001O1O001O001O001O001O001O001O001N101O001O001O001O1O001O001O001O0100O2O0O3N2M3N2M3N2M3N2M3N2M4M2N2M3N2M3N2M3N<C>CTT_1"}, "label": "half of sandwich under napkin on a white plate"}]}
{"id": 109114, "image": "COCO_train2014_000000109114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwhich split in half\"."}], "task": "refering", "answer_template": "The \"the sandwhich split in half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 160, 161, 162, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 447, 448, 449, 450, 451, 452, 453, 454, 455, 469, 470, 471, 472, 473, 474, 475, 476], "bbox": [0.24161764705882355, 0.20155228758169932, 0.8970915032679738, 0.9845588235294117], "iscrowd": 0, "area": 130800.80985, "rle": {"size": [612, 612], "counts": "Pih25ib08I7I8G8I7K5J6K5J7J5J6K5J6K5mNXMbAP3X>ZM]Am2\\>^MXAj2b>Q1J7H7J6I8I6I7J6J5J6K6I6K6J5J6K6I6K6I6K5K6I6K6J5J6K6I6K6J5J6K6J5XIfE`4^:nJRFo4R:`J^F^5e9QJkFm5X;M3N2M2O2M3N2N2M3N2M2O2M3N2N2M3N0O100O10000O010O100O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O1O10000O10000O10000O12N1O1O2N1O2N1O2N1O2N1O2N1O1O2M2O2N1O2N1O2N1O1O2N1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O00001O0O2O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O1O1O1O1O1O1O2N1O1O1N2N2N2N2N2N3M2N2N2N2O1N2N2N2N2N1O2N1O2N1O2N1O2N1O2M2O2N1O2M2O2N1O2M2O1O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1N2O2N1O2N1N3N1O2N1N3N1O2N2N2M3N3M2N2M3N2N2N2N3L3N2N2N2N2M3N3M2NkCYJP;d5gDgJZ;W5\\DUKd;V60O10001N10000O10000O2O0O10000cLoDeNR;Y1PEgNP;W1QEjNo:U1REhNR;U1PEhNS;V1oDgNT;X1oDcNT;[1PEaNR;^1PE^NS;`1Z3N3N1N2O3L4L4M3L5K4M3L4M3L4L4L4L5K4L4L4K5L4L4K5L5K4L4K5LUXU1"}, "label": "the sandwhich split in half"}]}
{"id": 109114, "image": "COCO_train2014_000000109114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwhich you can see the most of\"."}], "task": "refering", "answer_template": "The \"the half of the sandwhich you can see the most of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 160, 161, 162, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 447, 448, 449, 450, 451, 452, 453, 454, 455, 469, 470, 471, 472, 473, 474, 475, 476], "bbox": [0.24161764705882355, 0.20155228758169932, 0.8970915032679738, 0.9845588235294117], "iscrowd": 0, "area": 130800.80985, "rle": {"size": [612, 612], "counts": "Pih25ib08I7I8G8I7K5J6K5J7J5J6K5J6K5mNXMbAP3X>ZM]Am2\\>^MXAj2b>Q1J7H7J6I8I6I7J6J5J6K6I6K6J5J6K6I6K6I6K5K6I6K6J5J6K6I6K6J5J6K6J5XIfE`4^:nJRFo4R:`J^F^5e9QJkFm5X;M3N2M2O2M3N2N2M3N2M2O2M3N2N2M3N0O100O10000O010O100O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O1O10000O10000O10000O12N1O1O2N1O2N1O2N1O2N1O2N1O1O2M2O2N1O2N1O2N1O1O2N1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O00001O0O2O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O1O1O1O1O1O1O2N1O1O1N2N2N2N2N2N3M2N2N2N2O1N2N2N2N2N1O2N1O2N1O2N1O2N1O2M2O2N1O2M2O2N1O2M2O1O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1N2O2N1O2N1N3N1O2N1N3N1O2N2N2M3N3M2N2M3N2N2N2N3L3N2N2N2N2M3N3M2NkCYJP;d5gDgJZ;W5\\DUKd;V60O10001N10000O10000O2O0O10000cLoDeNR;Y1PEgNP;W1QEjNo:U1REhNR;U1PEhNS;V1oDgNT;X1oDcNT;[1PEaNR;^1PE^NS;`1Z3N3N1N2O3L4L4M3L5K4M3L4M3L4L4L4L5K4L4L4K5L4L4K5L5K4L4K5LUXU1"}, "label": "the half of the sandwhich you can see the most of"}]}
{"id": 494138, "image": "COCO_train2014_000000494138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue\"."}], "task": "refering", "answer_template": "The \"a man in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 51, 52, 73, 74, 75, 96, 97, 98, 120, 121, 143, 144, 166, 167, 168, 191, 192], "bbox": [0.18246875, 0.1092235294117647, 0.36343749999999997, 0.5934588235294118], "iscrowd": 0, "area": 6653.830900000001, "rle": {"size": [425, 640], "counts": "Te`12V=3M3L4M4L4L4L5K4L4K5L4L4L2NO1WO^DKc;6cDA^;b0gDVOZ;l0`0104K7J`0_Of0ZO8I2M3N2M3M3N2YOjLVGY3g8kLVGW3h8lLTGW3i8mLSGV3k8mLRGU3l8f0M3N2N2M2O1N3N1O1N3N1O0O2O0O22N2\\OfGTL]8e3fG_LY8a1dGB7QOT8U1nGB3YOQ8l0UHDNAS8?XHHIJo77aHJCOk72kHLYO3l7JUI0nN7l7F\\I2gN8m7CbI2`N=m7_OgI2[N?n7^OlI1VN`0n7]OQJ1SN>l7_OUJ3QN:k7AWJ5PN7h7C[J6oM3g7E\\J9nMNg7G_J;lMIl:6U1O10O0100O1O10O01O1O001O000010O01O00001O001O06J5L3L3N2M3M2K22N2N2N2N2N1O2N2N2N2M3N3M2NoRY5"}, "label": "a man in blue"}]}
{"id": 494138, "image": "COCO_train2014_000000494138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt , who is behind the man who is catching the frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt , who is behind the man who is catching the frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 51, 52, 73, 74, 75, 96, 97, 98, 120, 121, 143, 144, 166, 167, 168, 191, 192], "bbox": [0.18246875, 0.1092235294117647, 0.36343749999999997, 0.5934588235294118], "iscrowd": 0, "area": 6653.830900000001, "rle": {"size": [425, 640], "counts": "Te`12V=3M3L4M4L4L4L5K4L4K5L4L4L2NO1WO^DKc;6cDA^;b0gDVOZ;l0`0104K7J`0_Of0ZO8I2M3N2M3M3N2YOjLVGY3g8kLVGW3h8lLTGW3i8mLSGV3k8mLRGU3l8f0M3N2N2M2O1N3N1O1N3N1O0O2O0O22N2\\OfGTL]8e3fG_LY8a1dGB7QOT8U1nGB3YOQ8l0UHDNAS8?XHHIJo77aHJCOk72kHLYO3l7JUI0nN7l7F\\I2gN8m7CbI2`N=m7_OgI2[N?n7^OlI1VN`0n7]OQJ1SN>l7_OUJ3QN:k7AWJ5PN7h7C[J6oM3g7E\\J9nMNg7G_J;lMIl:6U1O10O0100O1O10O01O1O001O000010O01O00001O001O06J5L3L3N2M3M2K22N2N2N2N2N1O2N2N2N2M3N3M2NoRY5"}, "label": "a man in a blue shirt , who is behind the man who is catching the frisbee"}]}
{"id": 494138, "image": "COCO_train2014_000000494138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ultimate frisbee player catching the disc on a grass field\"."}], "task": "refering", "answer_template": "The \"an ultimate frisbee player catching the disc on a grass field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 98, 99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 168, 169, 170, 171, 192, 193, 194, 216, 217, 239, 240], "bbox": [0.264125, 0.17752941176470588, 0.47006250000000005, 0.7011294117647059], "iscrowd": 0, "area": 11497.770249999998, "rle": {"size": [425, 640], "counts": "nXV22U=4L4L4L4K5L4L3M4L4L4nNTO`EQ1]:P1M3M3M3M3M3M3M3M3M3M3N2M3N2O1N2O1O1N2O1O1O0O1000O10O10000O100000O01000000O14L4K6K5K00O1O1O1O001O1O1O1O1O1O010O13M5K6J2N001O001O001O001O1O001O001O001O1O1O1O1O2N1O2N101N1O1O2N1O2N101N1O1OQERNd:Y21N3N1N3M2O1N3M2O2M2N3N3L4L4M3L4L4M3L4L4M4K4L4M4K5L4KS]\\4"}, "label": "an ultimate frisbee player catching the disc on a grass field"}]}
{"id": 494138, "image": "COCO_train2014_000000494138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing black shorts green shirt catching a white freesbie\"."}], "task": "refering", "answer_template": "The \"a man wearing black shorts green shirt catching a white freesbie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 98, 99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 168, 169, 170, 171, 192, 193, 194, 216, 217, 239, 240], "bbox": [0.264125, 0.17752941176470588, 0.47006250000000005, 0.7011294117647059], "iscrowd": 0, "area": 11497.770249999998, "rle": {"size": [425, 640], "counts": "nXV22U=4L4L4L4K5L4L3M4L4L4nNTO`EQ1]:P1M3M3M3M3M3M3M3M3M3M3N2M3N2O1N2O1O1N2O1O1O0O1000O10O10000O100000O01000000O14L4K6K5K00O1O1O1O001O1O1O1O1O1O010O13M5K6J2N001O001O001O001O1O001O001O001O1O1O1O1O2N1O2N101N1O1O2N1O2N101N1O1OQERNd:Y21N3N1N3M2O1N3M2O2M2N3N3L4L4M3L4L4M3L4L4M4K4L4M4K5L4KS]\\4"}, "label": "a man wearing black shorts green shirt catching a white freesbie"}]}
{"id": 469567, "image": "COCO_train2014_000000469567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"reddish - orange livingroom couch with four pillows and a throw blanket\"."}], "task": "refering", "answer_template": "The \"reddish - orange livingroom couch with four pillows and a throw blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 182, 183, 184, 185, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 273, 274, 275], "bbox": [0.05, 0.514296875, 0.3979058823529412, 0.7912343749999999], "iscrowd": 0, "area": 17483.746249999997, "rle": {"size": [640, 425], "counts": "Xb=1mc03N2M3N2M3NOOO12N3M2N2O2M2N3M3M3M3M3M3N2M3M3M3M3M4L3M3N2M3M4L3N2M3M3N3L3N3L4M3L4L5L4K9n_OVL`?W4L3O1000000O10000001O0000001O0000001O00001O001O010O00]@jK]?`4L4K4M3M2N1N1000000O2O0000000O1000000O100O100O100O100O100O1O100O100O100O1O1001O1O1O7I4L1O1O1O00001O00001O001O1O1O1M4M1N2I8I6J6J6I7J6J6J6I7J6J6J6H8E=D;D>CQeo4"}, "label": "reddish - orange livingroom couch with four pillows and a throw blanket"}]}
{"id": 469567, "image": "COCO_train2014_000000469567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange love seat with four pillows and a white blanket\"."}], "task": "refering", "answer_template": "The \"an orange love seat with four pillows and a white blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 182, 183, 184, 185, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 273, 274, 275], "bbox": [0.05, 0.514296875, 0.3979058823529412, 0.7912343749999999], "iscrowd": 0, "area": 17483.746249999997, "rle": {"size": [640, 425], "counts": "Xb=1mc03N2M3N2M3NOOO12N3M2N2O2M2N3M3M3M3M3M3N2M3M3M3M3M4L3M3N2M3M4L3N2M3M3N3L3N3L4M3L4L5L4K9n_OVL`?W4L3O1000000O10000001O0000001O0000001O00001O001O010O00]@jK]?`4L4K4M3M2N1N1000000O2O0000000O1000000O100O100O100O100O100O1O100O100O100O1O1001O1O1O7I4L1O1O1O00001O00001O001O1O1O1M4M1N2I8I6J6J6I7J6J6J6I7J6J6J6H8E=D;D>CQeo4"}, "label": "an orange love seat with four pillows and a white blanket"}]}
{"id": 469567, "image": "COCO_train2014_000000469567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a 3 seat couch with 6 pillows on it\"."}], "task": "refering", "answer_template": "The \"a 3 seat couch with 6 pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 192, 204, 205, 206, 207, 208, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 310, 311, 312, 313], "bbox": [0.6296235294117647, 0.53125, 0.9801176470588234, 0.8890156250000001], "iscrowd": 0, "area": 25759.95575, "rle": {"size": [640, 425], "counts": "djW5b1^b0?A>B9H7H9G9G9G6J5K5K5K5K6oAlKk;e4YCkKa<f54K52N2N1O1O2N100O1O0000000000010O00000000000001O00000000000O101OO10O100000O10O100000O0100000000O010000000O1000000O10O1001O0O10000000001O0000000000000000000000003M;E<D6J2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2M5L4K4M4K5KQ1PO3L4M4K4M3L3N1N3N2M3M3M3M3M3M3M3M3M4L3M4L3M4L3Mi_4"}, "label": "a 3 seat couch with 6 pillows on it"}]}
{"id": 469567, "image": "COCO_train2014_000000469567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark tan three cushion couch on the right side of a living room\"."}], "task": "refering", "answer_template": "The \"dark tan three cushion couch on the right side of a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 192, 204, 205, 206, 207, 208, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 310, 311, 312, 313], "bbox": [0.6296235294117647, 0.53125, 0.9801176470588234, 0.8890156250000001], "iscrowd": 0, "area": 25759.95575, "rle": {"size": [640, 425], "counts": "djW5b1^b0?A>B9H7H9G9G9G6J5K5K5K5K6oAlKk;e4YCkKa<f54K52N2N1O1O2N100O1O0000000000010O00000000000001O00000000000O101OO10O100000O10O100000O0100000000O010000000O1000000O10O1001O0O10000000001O0000000000000000000000003M;E<D6J2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2M5L4K4M4K5KQ1PO3L4M4K4M3L3N1N3N2M3M3M3M3M3M3M3M3M4L3M4L3M4L3Mi_4"}, "label": "dark tan three cushion couch on the right side of a living room"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle between two other motorcycles\"."}], "task": "refering", "answer_template": "The \"a motorcycle between two other motorcycles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 222, 223, 224, 225, 240, 241, 242, 258, 259, 260], "bbox": [0.31592000000000003, 0.40089156626506023, 0.6318400000000001, 0.9707228915662651], "iscrowd": 0, "area": 20702.7021, "rle": {"size": [415, 500], "counts": "dZP23h<6J7J6YNCbFc0Q9IjF<R9HkF:T9GkF;R9HkF:T9IhF:U9IiF8l8fNhFT198o8fNdFV1:6Q9fNaFW1:6U9n1O00001O001VHiLj5W3nIUMn5k2oI\\Mm5e2PJ`Mm5`2QJbMP6^2mIdMS6\\2kIfMU6Z2iIhMW6Y2eIjM[6V2dIlM[6S2eInM\\6P2dIQN\\6n1dISN\\6l1dIUN\\6k1cIVN]6i1cIXN^6f1bI[N^6d1bI]N^6b1bI_N^6\\1gIdNY6X1kIhNU6Y1iIiNW6V1iIjNW6V1[IXOk5iM]Jo2A_OP6hM[Jj2AAR6kMYJd2AEU6jMYJX5e5kJZJU5e5mJZJIKc4j5eK\\JD0d4c5jK\\J@5d4^5mK^J[O9f4X5QL^JVOa0e4Q5VL`Kg3`4YLaKf3_4[L_Kf3a4[L[Ki3d4WLYKl3g4ULUKn3k4SLQKP4P5oKmJT4U5kKhJW4Y5iKeJX4]5fKcJX4a5gK]JX4g5gKXJW4l5gKSJY4o5gKPJW4R6iKnIU4U6kKiIT4Z6lKeIR4^6mKaIR4b6nK]IQ4d6PLZIo3i6PLVIo3m6QLRIm3P7SLoHl3S7ULlHi3U7XLjHh3W7YLgHf3[7[LcHd3_7l01O2N2O1N1O010O10O01O010O1O010O10O01000O0100000O2000O10O01M3L4L4M2M4L5L3L4L4L4L4M4K4WOi0L4L4M301O0000000000001O01OO100O2N2N1O2N1O2N2TNgEo0^;F8HN3K43M3M3M3M3M3M5KW[Z2"}, "label": "a motorcycle between two other motorcycles"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcyle in the center\"."}], "task": "refering", "answer_template": "The \"motorcyle in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 222, 223, 224, 225, 240, 241, 242, 258, 259, 260], "bbox": [0.31592000000000003, 0.40089156626506023, 0.6318400000000001, 0.9707228915662651], "iscrowd": 0, "area": 20702.7021, "rle": {"size": [415, 500], "counts": "dZP23h<6J7J6YNCbFc0Q9IjF<R9HkF:T9GkF;R9HkF:T9IhF:U9IiF8l8fNhFT198o8fNdFV1:6Q9fNaFW1:6U9n1O00001O001VHiLj5W3nIUMn5k2oI\\Mm5e2PJ`Mm5`2QJbMP6^2mIdMS6\\2kIfMU6Z2iIhMW6Y2eIjM[6V2dIlM[6S2eInM\\6P2dIQN\\6n1dISN\\6l1dIUN\\6k1cIVN]6i1cIXN^6f1bI[N^6d1bI]N^6b1bI_N^6\\1gIdNY6X1kIhNU6Y1iIiNW6V1iIjNW6V1[IXOk5iM]Jo2A_OP6hM[Jj2AAR6kMYJd2AEU6jMYJX5e5kJZJU5e5mJZJIKc4j5eK\\JD0d4c5jK\\J@5d4^5mK^J[O9f4X5QL^JVOa0e4Q5VL`Kg3`4YLaKf3_4[L_Kf3a4[L[Ki3d4WLYKl3g4ULUKn3k4SLQKP4P5oKmJT4U5kKhJW4Y5iKeJX4]5fKcJX4a5gK]JX4g5gKXJW4l5gKSJY4o5gKPJW4R6iKnIU4U6kKiIT4Z6lKeIR4^6mKaIR4b6nK]IQ4d6PLZIo3i6PLVIo3m6QLRIm3P7SLoHl3S7ULlHi3U7XLjHh3W7YLgHf3[7[LcHd3_7l01O2N2O1N1O010O10O01O010O1O010O10O01000O0100000O2000O10O01M3L4L4M2M4L5L3L4L4L4L4M4K4WOi0L4L4M301O0000000000001O01OO100O2N2N1O2N1O2N2TNgEo0^;F8HN3K43M3M3M3M3M3M5KW[Z2"}, "label": "motorcyle in the center"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on the left , standing in front of his motorcycle\"."}], "task": "refering", "answer_template": "The \"a man on the left , standing in front of his motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 146, 163, 164, 181, 182, 199, 200, 217, 218], "bbox": [0.018920000000000003, 0.27342168674698797, 0.166, 0.8582168674698795], "iscrowd": 0, "area": 10892.530700000003, "rle": {"size": [415, 500], "counts": "Vj3c0[<5J6K5J6J5L3L5L4K5L4K5L3M4M3L5K4M4K4L4M4UIfLh3^3cJcL34Y5^3ZJSMHEn5[3PJfM^OUO`6Z3hI[NW6i1`I]N`6Q4O2N1VOj0O1O2N100O1O1O1O100000O10001O00001O00000bNWKeJk4o4aKPK_4l4eKTK[4h4jKWKW4d4nK[KR4a4RL_Kn3]4WLbKi3[4[LdKf3W4cLdK^3W4iLeKZ3U4lLiKU3P4TMmKm2m0XJ`1`3aMZ2d0cJh1P9WNTGf1l8YNWGe1j8YNZGd1f8[N^Gb1b8]N`Gb1a8\\NaGd1^8[NdGd1]8ZNeGe1]8XNeGg1]8VNeGi1]8TNeGk1hSY5"}, "label": "a man on the left , standing in front of his motorcycle"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with blue helmet\"."}], "task": "refering", "answer_template": "The \"man with blue helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 146, 163, 164, 181, 182, 199, 200, 217, 218], "bbox": [0.018920000000000003, 0.27342168674698797, 0.166, 0.8582168674698795], "iscrowd": 0, "area": 10892.530700000003, "rle": {"size": [415, 500], "counts": "Vj3c0[<5J6K5J6J5L3L5L4K5L4K5L3M4M3L5K4M4K4L4M4UIfLh3^3cJcL34Y5^3ZJSMHEn5[3PJfM^OUO`6Z3hI[NW6i1`I]N`6Q4O2N1VOj0O1O2N100O1O1O1O100000O10001O00001O00000bNWKeJk4o4aKPK_4l4eKTK[4h4jKWKW4d4nK[KR4a4RL_Kn3]4WLbKi3[4[LdKf3W4cLdK^3W4iLeKZ3U4lLiKU3P4TMmKm2m0XJ`1`3aMZ2d0cJh1P9WNTGf1l8YNWGe1j8YNZGd1f8[N^Gb1b8]N`Gb1a8\\NaGd1^8[NdGd1]8ZNeGe1]8XNeGg1]8VNeGi1]8TNeGk1hSY5"}, "label": "man with blue helmet"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike on the left\"."}], "task": "refering", "answer_template": "The \"the bike on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 128, 129, 130, 131, 146, 147, 148, 164, 165, 166, 182, 183, 200], "bbox": [0.12084, 0.3919036144578313, 0.30886, 0.7868192771084337], "iscrowd": 0, "area": 7664.332000000003, "rle": {"size": [415, 500], "counts": "ebh01j<5K5QOFRE`0e0@j85XF`0l0]Ok89oE?S1ZOn8X1PGiNl6MoI\\1ROhNo6MnI^1kNjNW7JlI^1dNPO^7ElI]1^NTOg7@jI_2Y6aMdI`2^6_M_Ib2d6^MUIf2n6ZMkHj2V7WMdHl2_7YMVHi2m7^MhGd2[8m03N1M2O2M3M3N2M3N2OM3N2M3O10010O0001O00001O001O01O01O00100O1O1O2N100O1O1O2N100O1O2N1O101WN[FN2eNf9Z1YF03dNf9Z1XF0S:LoE3T:JmE5V:IkE5W:JiE4[:SOhE5Lg0^:QOhE8Jd0b:ROfE:G`0h:SOcE=E;_;DeD4_;MfDI];8e01OCbC9]<HdC7\\<IdC8[<HeC8Z<HhC7X<IhC8V<IkC6T<JmC7Q<JPD4P<MPD1Q<OPDOQ<2PDJR<7:12M5LSQ\\4"}, "label": "the bike on the left"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle , located behind a man wearing a black jacket and blue and white pants\"."}], "task": "refering", "answer_template": "The \"a motorcycle , located behind a man wearing a black jacket and blue and white pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 128, 129, 130, 131, 146, 147, 148, 164, 165, 166, 182, 183, 200], "bbox": [0.12084, 0.3919036144578313, 0.30886, 0.7868192771084337], "iscrowd": 0, "area": 7664.332000000003, "rle": {"size": [415, 500], "counts": "ebh01j<5K5QOFRE`0e0@j85XF`0l0]Ok89oE?S1ZOn8X1PGiNl6MoI\\1ROhNo6MnI^1kNjNW7JlI^1dNPO^7ElI]1^NTOg7@jI_2Y6aMdI`2^6_M_Ib2d6^MUIf2n6ZMkHj2V7WMdHl2_7YMVHi2m7^MhGd2[8m03N1M2O2M3M3N2M3N2OM3N2M3O10010O0001O00001O001O01O01O00100O1O1O2N100O1O1O2N100O1O2N1O101WN[FN2eNf9Z1YF03dNf9Z1XF0S:LoE3T:JmE5V:IkE5W:JiE4[:SOhE5Lg0^:QOhE8Jd0b:ROfE:G`0h:SOcE=E;_;DeD4_;MfDI];8e01OCbC9]<HdC7\\<IdC8[<HeC8Z<HhC7X<IhC8V<IkC6T<JmC7Q<JPD4P<MPD1Q<OPDOQ<2PDJR<7:12M5LSQ\\4"}, "label": "a motorcycle , located behind a man wearing a black jacket and blue and white pants"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vacant blue motorcycle on the far right of the photo\"."}], "task": "refering", "answer_template": "The \"the vacant blue motorcycle on the far right of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 193, 194, 195, 210, 211, 212, 213, 228, 229, 230, 231, 246, 247, 248, 265], "bbox": [0.70648, 0.4649638554216868, 0.9685, 0.948289156626506], "iscrowd": 0, "area": 10765.482150000003, "rle": {"size": [415, 500], "counts": "m\\_46\\<d0E;F5J6J6K4L5L4M3L3M4M2M3N1O2QO]M^Ge2a8\\M^Gf2_8\\M`Ge2^8]MbGd2\\8]McGd2\\8]McGe2[8\\MeGd2Z8]MeGe2Y8\\MfGe2V7PMXI<Af2T7SMWI9\\OCLP3`7VMVI9ZOEOm2_7WMWI8WOG3j2_7XMUId0ZOV2a7VMRIf0[OW2d7RMoHQ4X6gKlI8JS4Y6oKdIN2T4Z6nKdIN1T4[6oKdIMEG6^4`6oKfIKDJ3\\4e6oKcIKEK1\\4h6mKbILDNN[4n6iK_IOE_4n6bK\\IOF_4o6cKXIOI^4W7]KnH5J^4[7]KhH6M]4^7\\KbH9OZ4b7d02M3L2O1N2N1O100O1O2M2M3N2M3M4L4J6N2@[GhLc8X3bGdL]8n2\\GRMLJ=6Y8i2dGVM71S8j2eGTM;1o7k2gGSM<2k7k2iGSM>1h7l2kGRM>3e7k2lHUMR7k2PIYMj6g2XIWMh6i2ZIVMe6h2^IWMc6f2aIWM`6f2o1M3M3L4M2bNaE:b:DaEVOLo0d:JfE4\\:EXE@?i0`:\\OUEK<g0`:\\OVEM<e0^:_OUEL`0a0\\:DSELc0=Z:GSEMe08Y:KQEOi02V:OQEOl0NU:2oD0o0KR:5oDOX<1iCNW<2jCMV<3:100M3M3M3M3M3O1OAQD0o;HYD8f;IZD8e;H[D8e;G\\D9c;H]D8c;G^D9Y;J[DM<9X;1iDNW;1jDOU;2kDNT;3lDMS;5lDKT;5lDKU;4kDLU;4kDLU;4kDLU;4kDKW;3kDKW;4iDKY;4gDLZ;OjD0T<N1O]e6"}, "label": "the vacant blue motorcycle on the far right of the photo"}]}
{"id": 526912, "image": "COCO_train2014_000000526912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lone motorcycle to to the right of others\"."}], "task": "refering", "answer_template": "The \"a lone motorcycle to to the right of others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 193, 194, 195, 210, 211, 212, 213, 228, 229, 230, 231, 246, 247, 248, 265], "bbox": [0.70648, 0.4649638554216868, 0.9685, 0.948289156626506], "iscrowd": 0, "area": 10765.482150000003, "rle": {"size": [415, 500], "counts": "m\\_46\\<d0E;F5J6J6K4L5L4M3L3M4M2M3N1O2QO]M^Ge2a8\\M^Gf2_8\\M`Ge2^8]MbGd2\\8]McGd2\\8]McGe2[8\\MeGd2Z8]MeGe2Y8\\MfGe2V7PMXI<Af2T7SMWI9\\OCLP3`7VMVI9ZOEOm2_7WMWI8WOG3j2_7XMUId0ZOV2a7VMRIf0[OW2d7RMoHQ4X6gKlI8JS4Y6oKdIN2T4Z6nKdIN1T4[6oKdIMEG6^4`6oKfIKDJ3\\4e6oKcIKEK1\\4h6mKbILDNN[4n6iK_IOE_4n6bK\\IOF_4o6cKXIOI^4W7]KnH5J^4[7]KhH6M]4^7\\KbH9OZ4b7d02M3L2O1N2N1O100O1O2M2M3N2M3M4L4J6N2@[GhLc8X3bGdL]8n2\\GRMLJ=6Y8i2dGVM71S8j2eGTM;1o7k2gGSM<2k7k2iGSM>1h7l2kGRM>3e7k2lHUMR7k2PIYMj6g2XIWMh6i2ZIVMe6h2^IWMc6f2aIWM`6f2o1M3M3L4M2bNaE:b:DaEVOLo0d:JfE4\\:EXE@?i0`:\\OUEK<g0`:\\OVEM<e0^:_OUEL`0a0\\:DSELc0=Z:GSEMe08Y:KQEOi02V:OQEOl0NU:2oD0o0KR:5oDOX<1iCNW<2jCMV<3:100M3M3M3M3M3O1OAQD0o;HYD8f;IZD8e;H[D8e;G\\D9c;H]D8c;G^D9Y;J[DM<9X;1iDNW;1jDOU;2kDNT;3lDMS;5lDKT;5lDKU;4kDLU;4kDLU;4kDLU;4kDKW;3kDKW;4iDKY;4gDLZ;OjD0T<N1O]e6"}, "label": "a lone motorcycle to to the right of others"}]}
{"id": 117319, "image": "COCO_train2014_000000117319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue blazer ready to ski\"."}], "task": "refering", "answer_template": "The \"a man in blue blazer ready to ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 149, 150, 173, 174, 197, 219, 242, 264, 265, 266, 287, 288, 310, 311, 332, 333, 334, 335], "bbox": [0.46453125, 0.2784375, 0.5895312500000001, 0.8772500000000001], "iscrowd": 0, "area": 7658.425499999998, "rle": {"size": [480, 640], "counts": "SS\\42l>2N2N2N2N2N2N2N2N2N2O1N2N1O2N2N2N2N2N2N2N2N2N2TIkNKW1H\\O0f0NAL`02GG;7LB6=0\\O2d03UONl08lNJT19\\N4d1LkMe0U2\\OYMT1h2mNfLd1Z3]NSLU2m3kMgKa2Y4`MfK`2Z4aMeK^2\\4cMcK]2]4dMbK\\2^4dMbK\\2^4eMaKZ2`4gM_KY2a4hM^KX2c4hM\\Kd1`N^LT6n1\\K`1dNbLP6o1[K\\1hNdLo5P2XKX1mNhLk5Q2WKS1ROlLg5R2WKn0UOPMe5Q2VKn0WOoLd5T2UKj0YORMc5T2TKh0ZOTMc5T2SKe0\\OWMb5R1jIYOX1]2]OXMc5i0PJBP1Z2@YMb5f0TJFj0f1hNmMi01b5b0XJIe0e1POfMc0:f58XJ2?c17SNT52\\J6i0T1HcNk68PJ?XOXOj68SJ9TO_Ok67VJ3POFl66[JJkN0k65cLJ`35`LJb35]LKe34ZLLh33WLLn30RL0R4MmK3W4JhK5b9O1O2N1O][k3"}, "label": "a man in blue blazer ready to ski"}]}
{"id": 117319, "image": "COCO_train2014_000000117319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing skiis , a blue jacket , and an orange hat\"."}], "task": "refering", "answer_template": "The \"a man wearing skiis , a blue jacket , and an orange hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 149, 150, 173, 174, 197, 219, 242, 264, 265, 266, 287, 288, 310, 311, 332, 333, 334, 335], "bbox": [0.46453125, 0.2784375, 0.5895312500000001, 0.8772500000000001], "iscrowd": 0, "area": 7658.425499999998, "rle": {"size": [480, 640], "counts": "SS\\42l>2N2N2N2N2N2N2N2N2N2O1N2N1O2N2N2N2N2N2N2N2N2N2TIkNKW1H\\O0f0NAL`02GG;7LB6=0\\O2d03UONl08lNJT19\\N4d1LkMe0U2\\OYMT1h2mNfLd1Z3]NSLU2m3kMgKa2Y4`MfK`2Z4aMeK^2\\4cMcK]2]4dMbK\\2^4dMbK\\2^4eMaKZ2`4gM_KY2a4hM^KX2c4hM\\Kd1`N^LT6n1\\K`1dNbLP6o1[K\\1hNdLo5P2XKX1mNhLk5Q2WKS1ROlLg5R2WKn0UOPMe5Q2VKn0WOoLd5T2UKj0YORMc5T2TKh0ZOTMc5T2SKe0\\OWMb5R1jIYOX1]2]OXMc5i0PJBP1Z2@YMb5f0TJFj0f1hNmMi01b5b0XJIe0e1POfMc0:f58XJ2?c17SNT52\\J6i0T1HcNk68PJ?XOXOj68SJ9TO_Ok67VJ3POFl66[JJkN0k65cLJ`35`LJb35]LKe34ZLLh33WLLn30RL0R4MmK3W4JhK5b9O1O2N1O][k3"}, "label": "a man wearing skiis , a blue jacket , and an orange hat"}]}
{"id": 117319, "image": "COCO_train2014_000000117319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy in the green coat in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the guy in the green coat in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 262, 263, 264, 285, 286, 287, 308, 309, 310, 330, 331, 332, 333, 353, 354, 355, 356, 377, 378, 379], "bbox": [0.362359375, 0.2719166666666667, 0.67584375, 0.9595416666666667], "iscrowd": 0, "area": 28241.2019, "rle": {"size": [480, 640], "counts": "lm\\32l>4L4L4L5K4L4M3N2N2mBC^;a0kC7Q<_1M4L4iHRMT2R3bMVM\\2m2ZM\\Md2h2RM`Mk2d2lLdMR3^2fLjMX3W2`LSN]3n1\\LZNb3h1VL`Nh3a1QLfNn3[1bKUO\\4n0RKCm4>cJ2\\5OQJd0n5X4O1O1O001N101O00001O000O100000000000000L4TOl0M3N2N20000000000000000001O0O10000000000UOZK\\Gf4b8\\K^Gd4_8_K`Gb4^8aKaG`4[8cKeG]4Y8eKgG\\4V8fKjGZ4S8iKmGX4P8kKoGV4m7mKSHT4j7nKUHT4h7nKXHR4f7PLZHQ4c7RL\\Ho3a7SL_Hn3]7ULcHl3Z7VLfHk3W7WLiHi3U7ZLiHh3T7ZLlHg3Q7[LoHf34jJQ6b1kIe3@_K`6n0PJb4j5aKUJb4f5`KZJd4`5^K`Je4Z5^KeJg4U5[KkJh4P5[KoJi4k4YKTKk4_1VIR1Q2\\Mo4_1bIa0V9^O\\GEP9:l2O001O1O1O001O1O1O1O01000O1000O010000O10O10O1000O010000O10O10O100O1O001O1O10O01O000000001O00000000001O000000000010O00000001O00000000001O0000000000000000000O100000001N1000000O101O1O1O1N3N1O1O2N1N2O2NnhP3"}, "label": "the guy in the green coat in the right hand picture"}]}
{"id": 117319, "image": "COCO_train2014_000000117319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in a light green coat and white pants\"."}], "task": "refering", "answer_template": "The \"person in a light green coat and white pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 262, 263, 264, 285, 286, 287, 308, 309, 310, 330, 331, 332, 333, 353, 354, 355, 356, 377, 378, 379], "bbox": [0.362359375, 0.2719166666666667, 0.67584375, 0.9595416666666667], "iscrowd": 0, "area": 28241.2019, "rle": {"size": [480, 640], "counts": "lm\\32l>4L4L4L5K4L4M3N2N2mBC^;a0kC7Q<_1M4L4iHRMT2R3bMVM\\2m2ZM\\Md2h2RM`Mk2d2lLdMR3^2fLjMX3W2`LSN]3n1\\LZNb3h1VL`Nh3a1QLfNn3[1bKUO\\4n0RKCm4>cJ2\\5OQJd0n5X4O1O1O001N101O00001O000O100000000000000L4TOl0M3N2N20000000000000000001O0O10000000000UOZK\\Gf4b8\\K^Gd4_8_K`Gb4^8aKaG`4[8cKeG]4Y8eKgG\\4V8fKjGZ4S8iKmGX4P8kKoGV4m7mKSHT4j7nKUHT4h7nKXHR4f7PLZHQ4c7RL\\Ho3a7SL_Hn3]7ULcHl3Z7VLfHk3W7WLiHi3U7ZLiHh3T7ZLlHg3Q7[LoHf34jJQ6b1kIe3@_K`6n0PJb4j5aKUJb4f5`KZJd4`5^K`Je4Z5^KeJg4U5[KkJh4P5[KoJi4k4YKTKk4_1VIR1Q2\\Mo4_1bIa0V9^O\\GEP9:l2O001O1O1O001O1O1O1O01000O1000O010000O10O10O1000O010000O10O10O100O1O001O1O10O01O000000001O00000000001O000000000010O00000001O00000000001O0000000000000000000O100000001N1000000O101O1O1O1N3N1O1O2N1N2O2NnhP3"}, "label": "person in a light green coat and white pants"}]}
{"id": 117319, "image": "COCO_train2014_000000117319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis that belong to the man wearing a green and white coat\"."}], "task": "refering", "answer_template": "The \"the skis that belong to the man wearing a green and white coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [358, 359, 360, 361, 374, 375, 376, 380, 381], "bbox": [0.23932812499999997, 0.8915416666666667, 0.711265625, 0.99775], "iscrowd": 0, "area": 2005.3924000000027, "rle": {"size": [480, 640], "counts": "jdX21o>00000000000O2O0000000000000O101O000000000O1000001O0000000O100000001O000O100000000000000O100000000000O0100000000000000O100000000000000O10000001O001O001N101O001O001O001O001O001O001O001O00X`\\17a_cN00000O10O100000O10000000O10O100000000O10O100000O10000000O0100000000O10O100000O10000000O10O100000000O10O100000O10000000O0100000000O1000O10O100O1O100O10O0100O1O100O100O010O1000000O1000O3N2N2N1N2O0O010O010O010O010O00102M3N2M\\Yf2"}, "label": "the skis that belong to the man wearing a green and white coat"}]}
{"id": 117319, "image": "COCO_train2014_000000117319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis being worn by a skier wearing a green and white jacket\"."}], "task": "refering", "answer_template": "The \"skis being worn by a skier wearing a green and white jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [358, 359, 360, 361, 374, 375, 376, 380, 381], "bbox": [0.23932812499999997, 0.8915416666666667, 0.711265625, 0.99775], "iscrowd": 0, "area": 2005.3924000000027, "rle": {"size": [480, 640], "counts": "jdX21o>00000000000O2O0000000000000O101O000000000O1000001O0000000O100000001O000O100000000000000O100000000000O0100000000000000O100000000000000O10000001O001O001N101O001O001O001O001O001O001O001O00X`\\17a_cN00000O10O100000O10000000O10O100000000O10O100000O10000000O0100000000O10O100000O10000000O10O100000000O10O100000O10000000O0100000000O1000O10O100O1O100O10O0100O1O100O100O010O1000000O1000O3N2N2N1N2O0O010O010O010O010O00102M3N2M\\Yf2"}, "label": "skis being worn by a skier wearing a green and white jacket"}]}
{"id": 289357, "image": "COCO_train2014_000000289357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man surfing , but not the tip of the board\"."}], "task": "refering", "answer_template": "The \"a young man surfing , but not the tip of the board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 214, 217, 218], "bbox": [0.262671875, 0.19720657276995307, 0.6790625, 0.6680516431924883], "iscrowd": 0, "area": 26685.4439, "rle": {"size": [426, 640], "counts": "VRV21W=6J5L4K5K6K4K5K5L5J5K6K5J6J5K3N1N3M2O2M3M2O2M2O2N101N2O0O2O0O2O0O2N101N2O0O2O0O2O0dF`Lm8b3QG_Lo8b3oF^LQ9c3nF^LR9c3lF^LS9c3lF^LT9c3jF]LW9c3hF^LW9k30000O01000O1000O010000O01000O1000O010000O10O10O1000O010000O10O10O10000O01001N100000001O0000001O0O1000001O0000001O00000O101O0000001O002N2N2N2J6J6J6J6J6J6K5N2M3N2NO10000000O1000000000O01000000000000001O0K5F:G9J6N3M2N3M2O2M2WMdGb0_8YOfGd0\\8XOhGf0Z8YOhGd0[8YOgGe0[8ZOgGc0\\8[OeGc0]8\\OeGc0\\8ZOfGg0Y8XOiGi0V8ROnGo0R8mNRHS1o7hNTHZ1k7bNYH^1g7^N\\Hd1d7VNaHj1_7RNdHP2\\7kMhHU2Y7fMjH\\2R9O010O001O001O0jNRE0o:ORE0n:1RENo:1QE0n:0SEOn:1RENn:2SEMn:2SEMm:4SEKn:4SEKm:5SEKm:6SEIn:6SEJl:6UEIl:7TEHl:8UEGl:8UEGk::TEFm:9VEDj:<YEAh:?YE_Og:a0\\E]Oc:c0`EZOa:f0`EXO`:h0cEUO^:j0eESO[:n0fEPO[:o0P101O01O01O010O001O010O001O01O01O010O001O010O001O010O1O10O01O1O100O00100O1O1O010O1O1O100O1O100O1O2M2M3M3M3NWRe2"}, "label": "a young man surfing , but not the tip of the board"}]}
{"id": 289357, "image": "COCO_train2014_000000289357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man surfing on some rough tides\"."}], "task": "refering", "answer_template": "The \"a young man surfing on some rough tides\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 152, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 214, 217, 218], "bbox": [0.262671875, 0.19720657276995307, 0.6790625, 0.6680516431924883], "iscrowd": 0, "area": 26685.4439, "rle": {"size": [426, 640], "counts": "VRV21W=6J5L4K5K6K4K5K5L5J5K6K5J6J5K3N1N3M2O2M3M2O2M2O2N101N2O0O2O0O2O0O2N101N2O0O2O0O2O0dF`Lm8b3QG_Lo8b3oF^LQ9c3nF^LR9c3lF^LS9c3lF^LT9c3jF]LW9c3hF^LW9k30000O01000O1000O010000O01000O1000O010000O10O10O1000O010000O10O10O10000O01001N100000001O0000001O0O1000001O0000001O00000O101O0000001O002N2N2N2J6J6J6J6J6J6K5N2M3N2NO10000000O1000000000O01000000000000001O0K5F:G9J6N3M2N3M2O2M2WMdGb0_8YOfGd0\\8XOhGf0Z8YOhGd0[8YOgGe0[8ZOgGc0\\8[OeGc0]8\\OeGc0\\8ZOfGg0Y8XOiGi0V8ROnGo0R8mNRHS1o7hNTHZ1k7bNYH^1g7^N\\Hd1d7VNaHj1_7RNdHP2\\7kMhHU2Y7fMjH\\2R9O010O001O001O0jNRE0o:ORE0n:1RENo:1QE0n:0SEOn:1RENn:2SEMn:2SEMm:4SEKn:4SEKm:5SEKm:6SEIn:6SEJl:6UEIl:7TEHl:8UEGl:8UEGk::TEFm:9VEDj:<YEAh:?YE_Og:a0\\E]Oc:c0`EZOa:f0`EXO`:h0cEUO^:j0eESO[:n0fEPO[:o0P101O01O01O010O001O010O001O01O01O010O001O010O001O010O1O10O01O1O100O00100O1O1O010O1O1O100O1O100O1O2M2M3M3M3NWRe2"}, "label": "a young man surfing on some rough tides"}]}
{"id": 84558, "image": "COCO_train2014_000000084558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black aircraft hanging from the ceiling in an airplane facility\"."}], "task": "refering", "answer_template": "The \"black aircraft hanging from the ceiling in an airplane facility\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 75, 76, 77], "bbox": [0.003359375, 0.004488517745302714, 0.47429687499999995, 0.2], "iscrowd": 0, "area": 21634.070999999993, "rle": {"size": [479, 640], "counts": "RP1k0T<o1100O100O10000O100O100O100O100000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000XObCnN_<R1fCiNZ<X1jCcNV<]1oC^NQ<b1f0001O000000001O000000001O000000001O000000001O000000001O0000001N100000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O3M3M3M1O1O001O1O001O001O1O0000O10000000000000000O100000000000O1000000000000000000000000000000000000000O100000000O100000002M3N2N2N2N2M3N2N2N0O10000000O0100N2N2N1N3ZOPC[OR=d0VCROm<n0b000O01000O10000O010O10000O10O104L5K4L4K6K4L2N1O1O0O2O1O1O001O1N10000000000O01000000000O01bN^C>c<A`C<a<CaC<_<CdC:]<EeC9]<EfC8\\<FfC8\\<EgC:Z<DhC:Y<EiC9Y<EiC9`=N2N2M3CRTm4"}, "label": "black aircraft hanging from the ceiling in an airplane facility"}]}
{"id": 84558, "image": "COCO_train2014_000000084558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black plane hanging from the roof\"."}], "task": "refering", "answer_template": "The \"a black plane hanging from the roof\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 75, 76, 77], "bbox": [0.003359375, 0.004488517745302714, 0.47429687499999995, 0.2], "iscrowd": 0, "area": 21634.070999999993, "rle": {"size": [479, 640], "counts": "RP1k0T<o1100O100O10000O100O100O100O100000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000XObCnN_<R1fCiNZ<X1jCcNV<]1oC^NQ<b1f0001O000000001O000000001O000000001O000000001O000000001O0000001N100000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O3M3M3M1O1O001O1O001O001O1O0000O10000000000000000O100000000000O1000000000000000000000000000000000000000O100000000O100000002M3N2N2N2N2M3N2N2N0O10000000O0100N2N2N1N3ZOPC[OR=d0VCROm<n0b000O01000O10000O010O10000O10O104L5K4L4K6K4L2N1O1O0O2O1O1O001O1N10000000000O01000000000O01bN^C>c<A`C<a<CaC<_<CdC:]<EeC9]<EfC8\\<FfC8\\<EgC:Z<DhC:Y<EiC9Y<EiC9`=N2N2M3CRTm4"}, "label": "a black plane hanging from the roof"}]}
{"id": 84558, "image": "COCO_train2014_000000084558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane parked inside is green and white\"."}], "task": "refering", "answer_template": "The \"an airplane parked inside is green and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273], "bbox": [0.31310937499999997, 0.2296868475991649, 0.957734375, 0.6950104384133611], "iscrowd": 0, "area": 52732.63130000001, "rle": {"size": [479, 640], "counts": "_Vn25j>4K6K4L5K4L5J3N1O1O001N2O1O1O1O1N2O1O1O001O1N2O1O1O1O1N2O001O1O1O10000O10000O0100000O100nDPNT9P2iFRNW9o1eFTN\\9l1`FWNa9i1[FZNf9e1XF]Nh9d1TF_Nm9a1oEbNR:^1kEdNU:]1gEfNZ:Z1bEiN_:W1^EkNc:U1YEoNf:R1VEQOk:o0RESOo:l0nDWOR;j0jDYOW;g0fD[O[;T21O001O1OlNaLhF_3X9aLiF^3W9cLiF\\3W9dLiF]3V9cLjF]3V9dLiF\\3W9dLiF\\3V9fLiFZ3W9fLiFZ3W9fLiFZ3W9gLhFY3X9gLhFY3X9hLgFX3Y9hLgFX3Y9hLgFX3Y9iLfFW3Z9iLfFX3Y9iLfFW3Z9iLfFW3Z9iLfFW3Z9jLeFV3[9jLdFW3\\9jLcFV3]9jLcFV3]9jLcFV3]9kLbFU3^9kLbFU3^9lLaFT3_9lLaFT3_9lLaFU3^9lLaFT3_9lLaFT3_9mL`FS3`9mL`FS3`9mL`FS3`9nL_FR3a9nL_FR3a9oL^FQ3b9oL^FQ3a9PM_FP3a9QM^Fo2b9QM]FP3c9QM\\Fo2n7^LUIc0mNP3m7^LUIb0nNP3m7_LTIb0nNo2o7_LRIb0oNo2o7`LQIa0lNS3S8]LPIa0eNY3[8WLoHd4Q7]KnHc4R7^KmHb4T7^KkHb4U7_KjHa4V7`KiH`4W7aKhH_4X7bKgH_4X7bKgH^4Z7aKfH_4Z7bKeH^4[7bKeH^4[7cKdH]4\\7cKdH]4\\7dKcH\\4^7cKbH]4^7dKaH\\4_7dKaH\\4_7dKaH\\4_7eK`H[4`7eK`H\\4`7dK^H]4b7cK^H]4b7dK]H\\4c7dK]H\\4c7eK\\H[4d7eK\\H[4e7eKZH[4f7eKZH[4f7fKYHZ4g7fKYHZ4g7gKXHY4h7gKXHZ4h7eKXH[4h7fKWHZ4i7fKWHZ4i7gKVHY4j7gKVHY4j7hKUHX4l7gKTHY4l7hKSHX4m7hKSHX4m7iKRHW4n7iKRHX4m7iKRHW4o7hKQHX4o7iKPHW4P8iKPHW4P8iKPHW4P8jKoGV4Q8jKoGV4R8jKmGV4S8jKlGW4T8jKkGW4T8iKlGW4T8jKkGV4U8jKkGV4V8jKiGV4W8jKiGV4W8kKhGU4X8kKhGU4X8kKhGU4X8lKgGT4Z8kKfGV4Y8kKfGU4Z8kKfGU4Z8lKeGT4[8lKeGT4[8mKdGS4]8lKcGT4]8mKbGS4^8mKbGS4^8nKaGR4_8nKaGS4^8nKaGR4`8mK`GS4`8mK`GS4`8nK_GR4a8nK_GR4a8oK^GQ4b8oK^GQ4c8oK\\GQ4d8oK\\GQ4d8PL[GP4e8PLZGR4e8T1001O00000000001O01O0000000001O00000001O01O00000000001O000001O0001O00000000001O01O00000001O00000000010O00000000001O00000001O01O000000000aKVGi2k8UMWGj2i8VMWGj2i8UMYGk2f8TM\\Gk2d8UM\\Gk2d8TM^Gk2c8TM^Gk2b8TM_Gl2a8TM`Gk2`8TMbGj2_8VMaGj2_8UMcGj2^8UMdGi2\\8VMgGg2Z8YMhGe2X8ZMkGd2U8\\MmGb2S8]MPHa2Q8^MQH_2P8`MSH^2m7bMUH\\2k7cMXH[2h7eMZHX2g7gM\\HW2e7hM\\HW2d7hM]HX2c7hM]HW2d7hM^HW2b7iM^HW2b7hM`HW2a7hM_HX2a7gM`HX2a7hM`HW2`7hMaHX2_7hMaHX2_7hMbHV2`7hMaHX2_7hMbHW2^7hMcHX2]7hMcHW2^7hMdHW2\\7iMdHW2]7gMdHY2\\7gMeHX2[7gMfHX2[7hMfHW2Z7hMgHX2Y7hMgHX2Z7fMhHX2Y7hMgHX2Y7gMhHY2X7gMiHX2W7gMjHX2W7hMjHW2W7gMjHY2V7gMjHY2V7fMlHY2T7gMlHX2U7gMlHY2T7gMmHX2T7fMmHZ2S7fMmHY2T7fMnHY2R7gMnHY2R7fMPIY2P7gMPIX2R7gMnHY2R7fMPIY2P7gMPIY2P7fMQIZ2o6fMRIX2o6gMRIY2o6fMRIY2n6fMSIZ2m6fMSIY2n6fMTIY2l6gMTIY2l6fMUIZ2i9N101O00001N10001O0O2O00001N10001O000O2O001O0O101O000O2O001O000O2O00001N101O000O2O00001O0O101O001N102N2N3L4M3M3M3M3L4M3M3M2M4M3MiZ<"}, "label": "an airplane parked inside is green and white"}]}
{"id": 84558, "image": "COCO_train2014_000000084558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white airplane sitting on display\"."}], "task": "refering", "answer_template": "The \"a green and white airplane sitting on display\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273], "bbox": [0.31310937499999997, 0.2296868475991649, 0.957734375, 0.6950104384133611], "iscrowd": 0, "area": 52732.63130000001, "rle": {"size": [479, 640], "counts": "_Vn25j>4K6K4L5K4L5J3N1O1O001N2O1O1O1O1N2O1O1O001O1N2O1O1O1O1N2O001O1O1O10000O10000O0100000O100nDPNT9P2iFRNW9o1eFTN\\9l1`FWNa9i1[FZNf9e1XF]Nh9d1TF_Nm9a1oEbNR:^1kEdNU:]1gEfNZ:Z1bEiN_:W1^EkNc:U1YEoNf:R1VEQOk:o0RESOo:l0nDWOR;j0jDYOW;g0fD[O[;T21O001O1OlNaLhF_3X9aLiF^3W9cLiF\\3W9dLiF]3V9cLjF]3V9dLiF\\3W9dLiF\\3V9fLiFZ3W9fLiFZ3W9fLiFZ3W9gLhFY3X9gLhFY3X9hLgFX3Y9hLgFX3Y9hLgFX3Y9iLfFW3Z9iLfFX3Y9iLfFW3Z9iLfFW3Z9iLfFW3Z9jLeFV3[9jLdFW3\\9jLcFV3]9jLcFV3]9jLcFV3]9kLbFU3^9kLbFU3^9lLaFT3_9lLaFT3_9lLaFU3^9lLaFT3_9lLaFT3_9mL`FS3`9mL`FS3`9mL`FS3`9nL_FR3a9nL_FR3a9oL^FQ3b9oL^FQ3a9PM_FP3a9QM^Fo2b9QM]FP3c9QM\\Fo2n7^LUIc0mNP3m7^LUIb0nNP3m7_LTIb0nNo2o7_LRIb0oNo2o7`LQIa0lNS3S8]LPIa0eNY3[8WLoHd4Q7]KnHc4R7^KmHb4T7^KkHb4U7_KjHa4V7`KiH`4W7aKhH_4X7bKgH_4X7bKgH^4Z7aKfH_4Z7bKeH^4[7bKeH^4[7cKdH]4\\7cKdH]4\\7dKcH\\4^7cKbH]4^7dKaH\\4_7dKaH\\4_7dKaH\\4_7eK`H[4`7eK`H\\4`7dK^H]4b7cK^H]4b7dK]H\\4c7dK]H\\4c7eK\\H[4d7eK\\H[4e7eKZH[4f7eKZH[4f7fKYHZ4g7fKYHZ4g7gKXHY4h7gKXHZ4h7eKXH[4h7fKWHZ4i7fKWHZ4i7gKVHY4j7gKVHY4j7hKUHX4l7gKTHY4l7hKSHX4m7hKSHX4m7iKRHW4n7iKRHX4m7iKRHW4o7hKQHX4o7iKPHW4P8iKPHW4P8iKPHW4P8jKoGV4Q8jKoGV4R8jKmGV4S8jKlGW4T8jKkGW4T8iKlGW4T8jKkGV4U8jKkGV4V8jKiGV4W8jKiGV4W8kKhGU4X8kKhGU4X8kKhGU4X8lKgGT4Z8kKfGV4Y8kKfGU4Z8kKfGU4Z8lKeGT4[8lKeGT4[8mKdGS4]8lKcGT4]8mKbGS4^8mKbGS4^8nKaGR4_8nKaGS4^8nKaGR4`8mK`GS4`8mK`GS4`8nK_GR4a8nK_GR4a8oK^GQ4b8oK^GQ4c8oK\\GQ4d8oK\\GQ4d8PL[GP4e8PLZGR4e8T1001O00000000001O01O0000000001O00000001O01O00000000001O000001O0001O00000000001O01O00000001O00000000010O00000000001O00000001O01O000000000aKVGi2k8UMWGj2i8VMWGj2i8UMYGk2f8TM\\Gk2d8UM\\Gk2d8TM^Gk2c8TM^Gk2b8TM_Gl2a8TM`Gk2`8TMbGj2_8VMaGj2_8UMcGj2^8UMdGi2\\8VMgGg2Z8YMhGe2X8ZMkGd2U8\\MmGb2S8]MPHa2Q8^MQH_2P8`MSH^2m7bMUH\\2k7cMXH[2h7eMZHX2g7gM\\HW2e7hM\\HW2d7hM]HX2c7hM]HW2d7hM^HW2b7iM^HW2b7hM`HW2a7hM_HX2a7gM`HX2a7hM`HW2`7hMaHX2_7hMaHX2_7hMbHV2`7hMaHX2_7hMbHW2^7hMcHX2]7hMcHW2^7hMdHW2\\7iMdHW2]7gMdHY2\\7gMeHX2[7gMfHX2[7hMfHW2Z7hMgHX2Y7hMgHX2Z7fMhHX2Y7hMgHX2Y7gMhHY2X7gMiHX2W7gMjHX2W7hMjHW2W7gMjHY2V7gMjHY2V7fMlHY2T7gMlHX2U7gMlHY2T7gMmHX2T7fMmHZ2S7fMmHY2T7fMnHY2R7gMnHY2R7fMPIY2P7gMPIX2R7gMnHY2R7fMPIY2P7gMPIY2P7fMQIZ2o6fMRIX2o6gMRIY2o6fMRIY2n6fMSIZ2m6fMSIY2n6fMTIY2l6gMTIY2l6fMUIZ2i9N101O00001N10001O0O2O00001N10001O000O2O001O0O101O000O2O001O000O2O00001N101O000O2O00001O0O101O001N102N2N3L4M3M3M3M3L4M3M3M2M4M3MiZ<"}, "label": "a green and white airplane sitting on display"}]}
{"id": 51793, "image": "COCO_train2014_000000051793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man carrying a surfboard\"."}], "task": "refering", "answer_template": "The \"man carrying a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 89, 90, 91, 106, 107, 108, 124, 125, 141, 142, 159, 176, 193, 210, 227, 243], "bbox": [0.26366666666666666, 0.19775, 0.4134791666666667, 0.6292187499999999], "iscrowd": 0, "area": 7406.025149999998, "rle": {"size": [640, 480], "counts": "`a_22nc02N3M3M3M2N100mNH\\^O9ca0?e]OBYb0T1N1O2N2N2O1N2N2O1N2O1N3N2M3M3N2M2O2M3N2M3N2M2O2M3ZDkL]6W3`InL\\6U3aIkL`6V3]IiLe6Y3XIgLj6Z3SIfLn6\\3oHdLS7^3hHcLY7_3^HgLd7Z3THjLo7W3bFbL^N=Q;S3mEdMZN^Ok;P3WEROj:dNTDf0>W1`;oMWDg4n;SKWDj4l;QKYDl4k;nJZDo4i;lJ]DP5j;gJ[DV5Z=L4M3M2N3M3L4[Oe0L4K5K5L5J6K5UNk1C=C=BlP`5"}, "label": "man carrying a surfboard"}]}
{"id": 51793, "image": "COCO_train2014_000000051793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shirtless man holding a surfboard\"."}], "task": "refering", "answer_template": "The \"a shirtless man holding a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 89, 90, 91, 106, 107, 108, 124, 125, 141, 142, 159, 176, 193, 210, 227, 243], "bbox": [0.26366666666666666, 0.19775, 0.4134791666666667, 0.6292187499999999], "iscrowd": 0, "area": 7406.025149999998, "rle": {"size": [640, 480], "counts": "`a_22nc02N3M3M3M2N100mNH\\^O9ca0?e]OBYb0T1N1O2N2N2O1N2N2O1N2O1N3N2M3M3N2M2O2M3N2M3N2M2O2M3ZDkL]6W3`InL\\6U3aIkL`6V3]IiLe6Y3XIgLj6Z3SIfLn6\\3oHdLS7^3hHcLY7_3^HgLd7Z3THjLo7W3bFbL^N=Q;S3mEdMZN^Ok;P3WEROj:dNTDf0>W1`;oMWDg4n;SKWDj4l;QKYDl4k;nJZDo4i;lJ]DP5j;gJ[DV5Z=L4M3M2N3M3L4[Oe0L4K5K5L5J6K5UNk1C=C=BlP`5"}, "label": "a shirtless man holding a surfboard"}]}
{"id": 51793, "image": "COCO_train2014_000000051793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teal and blue surfboard\"."}], "task": "refering", "answer_template": "The \"a teal and blue surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 34, 35, 51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325, 340, 341, 357, 358, 374], "bbox": [0.0, 0.042828125, 0.3592708333333333, 0.9785], "iscrowd": 0, "area": 64944.298449999995, "rle": {"size": [640, 480], "counts": "l0fb0[1O1O2N1O1O1O1O2N1O1O1M3N3L3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N3L3N2M3N2M4L3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N3L3N2M3N3L3N2M3N2M4L3N2M3N2M4M2M3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N2M4M2M3M3N2M4M2M3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N3L3N2M4M2M3M4M2M3N3L3N2M4M2M3N3L3N2M4M2M3N3L3N2M4M2M4M2M3N3L3N2M4M2M3M4M3L3N3L4M2M4M3L3L5K4L5K5J5LQ1oNT1lNS1mNdUP6"}, "label": "a teal and blue surfboard"}]}
{"id": 51793, "image": "COCO_train2014_000000051793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"surfboard\"."}], "task": "refering", "answer_template": "The \"surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 34, 35, 51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325, 340, 341, 357, 358, 374], "bbox": [0.0, 0.042828125, 0.3592708333333333, 0.9785], "iscrowd": 0, "area": 64944.298449999995, "rle": {"size": [640, 480], "counts": "l0fb0[1O1O2N1O1O1O1O2N1O1O1M3N3L3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N3L3N2M3N2M4L3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N3L3N2M3N3L3N2M3N2M4L3N2M3N2M4M2M3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N2M4M2M3M3N2M4M2M3N2M3N2M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3N3L3N2M4M2M3M4M2M3N3L3N2M4M2M3N3L3N2M4M2M3N3L3N2M4M2M4M2M3N3L3N2M4M2M3M4M3L3N3L4M2M4M3L3L5K4L5K5J5LQ1oNT1lNS1mNdUP6"}, "label": "surfboard"}]}
{"id": 289367, "image": "COCO_train2014_000000289367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player\"."}], "task": "refering", "answer_template": "The \"a tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 60, 61, 62, 78, 79, 80, 81, 95, 96, 97, 98, 99, 116, 117, 118, 134, 135, 136, 152, 153, 154, 171, 172, 173, 174, 189, 190, 191], "bbox": [0.28196, 0.21741741741741744, 0.6832999999999999, 0.9057057057057057], "iscrowd": 0, "area": 12311.395900000005, "rle": {"size": [333, 500], "counts": "oo]13Y:7J1N01O0100O10O0100O10000O01000O100O01000O10000O01000O10000O010O100O0000001O000nN_OYHa0g7AWH`0h7AWH?i7CUH=k7ESH;n7EQH<n7FPH:P8GPH8P8JnG6R8KiGDBb0d8NfG@Ga0c84`G]OL?e88ZGZO0>f8>[GAe8`0ZGAf8>ZGBf8S102M2N2O2M2O1N2N2O1N2N2O1N2O1N2N20000O01000O10000hIbM_4_2_KeM_4[2^KjM`4V2^KnM_4S2^KRN`4n1^KTNb4l1[KVNf4j1XKXNg4i1VKYNW5Z1hJhN`5P1^JROa5o0\\JSOe5m0YJUOf5l0WJWOh5j0VJWOk5i0RJZOm5i0oIYOQ6j0iIXOW6`2O10O0100O010O00100O010jJnKo3S4nKoKS4P4jKRLW4o3eKTL[4m3bKUL_4l3\\KVLe4m3UKVLl4l3iJ\\LW5\\42O1O0O2O1O1N2N2O1N2N2TL]Jk2e5nLaJQ3a5hLfJW3[5bLlJ\\3T6N1O2N2O1N1K6eMnHY1S7eNhIa0Y6^OUJ5j5JXJ5i5JWJ6j5HXJ8h5GXJ9h5FZJ9g5FYJ;g5C[J<f5CZJ>g5_O[J`0g5^OZJa0h5\\OYJe0h5YOYJf0g7010O01ORGZOU8g0iG[OW8d0iG\\OX8c0gG_OX8b0fG@Z8?fGAZ8`0dGB\\8=cGD]8W100O0100O001N1O2O0O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N2N2M4MV^c1"}, "label": "a tennis player"}]}
{"id": 289367, "image": "COCO_train2014_000000289367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player wearing white shirt and black skirt\"."}], "task": "refering", "answer_template": "The \"tennis player wearing white shirt and black skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 60, 61, 62, 78, 79, 80, 81, 95, 96, 97, 98, 99, 116, 117, 118, 134, 135, 136, 152, 153, 154, 171, 172, 173, 174, 189, 190, 191], "bbox": [0.28196, 0.21741741741741744, 0.6832999999999999, 0.9057057057057057], "iscrowd": 0, "area": 12311.395900000005, "rle": {"size": [333, 500], "counts": "oo]13Y:7J1N01O0100O10O0100O10000O01000O100O01000O10000O01000O10000O010O100O0000001O000nN_OYHa0g7AWH`0h7AWH?i7CUH=k7ESH;n7EQH<n7FPH:P8GPH8P8JnG6R8KiGDBb0d8NfG@Ga0c84`G]OL?e88ZGZO0>f8>[GAe8`0ZGAf8>ZGBf8S102M2N2O2M2O1N2N2O1N2N2O1N2O1N2N20000O01000O10000hIbM_4_2_KeM_4[2^KjM`4V2^KnM_4S2^KRN`4n1^KTNb4l1[KVNf4j1XKXNg4i1VKYNW5Z1hJhN`5P1^JROa5o0\\JSOe5m0YJUOf5l0WJWOh5j0VJWOk5i0RJZOm5i0oIYOQ6j0iIXOW6`2O10O0100O010O00100O010jJnKo3S4nKoKS4P4jKRLW4o3eKTL[4m3bKUL_4l3\\KVLe4m3UKVLl4l3iJ\\LW5\\42O1O0O2O1O1N2N2O1N2N2TL]Jk2e5nLaJQ3a5hLfJW3[5bLlJ\\3T6N1O2N2O1N1K6eMnHY1S7eNhIa0Y6^OUJ5j5JXJ5i5JWJ6j5HXJ8h5GXJ9h5FZJ9g5FYJ;g5C[J<f5CZJ>g5_O[J`0g5^OZJa0h5\\OYJe0h5YOYJf0g7010O01ORGZOU8g0iG[OW8d0iG\\OX8c0gG_OX8b0fG@Z8?fGAZ8`0dGB\\8=cGD]8W100O0100O001N1O2O0O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N2N2M4MV^c1"}, "label": "tennis player wearing white shirt and black skirt"}]}
{"id": 313950, "image": "COCO_train2014_000000313950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog which is lying on the bed with standing another one\"."}], "task": "refering", "answer_template": "The \"a dog which is lying on the bed with standing another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.0, 0.5791764705882353, 0.348796875, 0.8458588235294118], "iscrowd": 0, "area": 17649.952100000002, "rle": {"size": [425, 640], "counts": "Z9f1f:m0C=1O0O100000000000000000000000001O00000000000000000000000000001O000000000000000001O000001O000001O000000000000001O000000000000001O00000000000000001O0000000000000000001O000000000O1O1O1N2N2N2N2N2001O1O1O001O1O1O1N2O1O1O1O1O1O001O001O00001O001O00001O001O010O000000000000010O000000000001O000001O00000000000000001O2N0000O1O1O1O1O1N2O1N2O1N2OO010O00010101N1OdN\\FUOd9ROSFb1`0VO\\9XOYF^1=YOh9b0ZF]Ok9=XFAl9GiF9T;O001O01O00010O0001O1O10O01O1O001O1O00100O1O001O1O001O10O01O1O1O001O1O001O1N2O00Tg\\5"}, "label": "a dog which is lying on the bed with standing another one"}]}
{"id": 313950, "image": "COCO_train2014_000000313950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beautiful cute dog laying on a bed\"."}], "task": "refering", "answer_template": "The \"a beautiful cute dog laying on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.0, 0.5791764705882353, 0.348796875, 0.8458588235294118], "iscrowd": 0, "area": 17649.952100000002, "rle": {"size": [425, 640], "counts": "Z9f1f:m0C=1O0O100000000000000000000000001O00000000000000000000000000001O000000000000000001O000001O000001O000000000000001O000000000000001O00000000000000001O0000000000000000001O000000000O1O1O1N2N2N2N2N2001O1O1O001O1O1O1N2O1O1O1O1O1O001O001O00001O001O00001O001O010O000000000000010O000000000001O000001O00000000000000001O2N0000O1O1O1O1O1N2O1N2O1N2OO010O00010101N1OdN\\FUOd9ROSFb1`0VO\\9XOYF^1=YOh9b0ZF]Ok9=XFAl9GiF9T;O001O01O00010O0001O1O10O01O1O001O1O00100O1O001O1O001O10O01O1O1O001O1O001O1N2O00Tg\\5"}, "label": "a beautiful cute dog laying on a bed"}]}
{"id": 313950, "image": "COCO_train2014_000000313950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chihuahua sitting atop a bed\"."}], "task": "refering", "answer_template": "The \"a chihuahua sitting atop a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 130, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295], "bbox": [0.6895, 0.3064470588235294, 0.8889843749999999, 0.8326588235294118], "iscrowd": 0, "area": 14137.093750000007, "rle": {"size": [425, 640], "counts": "eUg52U=9G6L4L4L4L2N2N3M2N2O1N2O011N4L4M1lEmNm7T1nFaN0l0P9c0nFgNJl0V9>mF9R9GkF=S9DjFa0S9P2L3L3M3M3N3L3L3L5L4UHkJZ7X5cHhJ]7[5`HeJ_7c5000O10O10000000O1000M2O2N2N2M3M3N3C=N2OO00001O0H91O12M3N2N5J101O1N10;]JcHc4R8O01O2O1O1O1O0O18G:G4K3N2M2M4L4M2M4M3L3N2N2O1O2M1O101hMPF:1YOR:5UF`03mNn9:YFe0]:SOkEk0W:QOQFj0Y;M3N1N3N6JO1O1O010O0010O010O100O0010O01O01O01O010M2KmUm0"}, "label": "a chihuahua sitting atop a bed"}]}
{"id": 313950, "image": "COCO_train2014_000000313950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white dog sitting upright on a bed watching another dog\"."}], "task": "refering", "answer_template": "The \"a black and white dog sitting upright on a bed watching another dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 130, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295], "bbox": [0.6895, 0.3064470588235294, 0.8889843749999999, 0.8326588235294118], "iscrowd": 0, "area": 14137.093750000007, "rle": {"size": [425, 640], "counts": "eUg52U=9G6L4L4L4L2N2N3M2N2O1N2O011N4L4M1lEmNm7T1nFaN0l0P9c0nFgNJl0V9>mF9R9GkF=S9DjFa0S9P2L3L3M3M3N3L3L3L5L4UHkJZ7X5cHhJ]7[5`HeJ_7c5000O10O10000000O1000M2O2N2N2M3M3N3C=N2OO00001O0H91O12M3N2N5J101O1N10;]JcHc4R8O01O2O1O1O1O0O18G:G4K3N2M2M4L4M2M4M3L3N2N2O1O2M1O101hMPF:1YOR:5UF`03mNn9:YFe0]:SOkEk0W:QOQFj0Y;M3N1N3N6JO1O1O010O0010O010O100O0010O01O01O01O010M2KmUm0"}, "label": "a black and white dog sitting upright on a bed watching another dog"}]}
{"id": 313950, "image": "COCO_train2014_000000313950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small black and tan dog sitting up\"."}], "task": "refering", "answer_template": "The \"small black and tan dog sitting up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 130, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295], "bbox": [0.68971875, 0.30576470588235294, 0.891453125, 0.8283529411764705], "iscrowd": 0, "area": 14453.87275, "rle": {"size": [425, 640], "counts": "RVg54e<a0O2N1N3N1O2M2O1O2N1O2O0XDkN[;V1`DoN_;Q1]DUOa;[1[F[Nc7g1UHbNh7_1PHkNl7X1lGPOR8Q1fGYOW8i0_GB^8?XGLf86PG5m8W2N9H1N10000O1L4M3L5L<D0O10000O101N10000N2N2M3M3M3M3O1N2N2O1N2N2N2O0O1O02O3M4L4M2M4L4M3L3N3L3N2QK]HS4f7iK^HU4c7gKbHV4EkK\\7KRIi4o6SKVIk4f7N2YKfG[4f8N3N1N3M2N2N2N2N2N2N2N2N2O1N<D2N1O1O2M2iMjEX1Y:dNjE[1W:bNmE[1U:bNoE\\1S:`NQF]1o:N3N2M3N1ZOYDGk;5XDIj;5YDGk;6XDFk;8XDEh;;[DAe;`0b00O10O000O2N101N1O2N100O2O001O1O2N[hl0"}, "label": "small black and tan dog sitting up"}]}
{"id": 313950, "image": "COCO_train2014_000000313950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog sitting on a bed\"."}], "task": "refering", "answer_template": "The \"a dog sitting on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 130, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295], "bbox": [0.68971875, 0.30576470588235294, 0.891453125, 0.8283529411764705], "iscrowd": 0, "area": 14453.87275, "rle": {"size": [425, 640], "counts": "RVg54e<a0O2N1N3N1O2M2O1O2N1O2O0XDkN[;V1`DoN_;Q1]DUOa;[1[F[Nc7g1UHbNh7_1PHkNl7X1lGPOR8Q1fGYOW8i0_GB^8?XGLf86PG5m8W2N9H1N10000O1L4M3L5L<D0O10000O101N10000N2N2M3M3M3M3O1N2N2O1N2N2N2O0O1O02O3M4L4M2M4L4M3L3N3L3N2QK]HS4f7iK^HU4c7gKbHV4EkK\\7KRIi4o6SKVIk4f7N2YKfG[4f8N3N1N3M2N2N2N2N2N2N2N2N2O1N<D2N1O1O2M2iMjEX1Y:dNjE[1W:bNmE[1U:bNoE\\1S:`NQF]1o:N3N2M3N1ZOYDGk;5XDIj;5YDGk;6XDFk;8XDEh;;[DAe;`0b00O10O000O2N101N1O2N100O2O001O1O2N[hl0"}, "label": "a dog sitting on a bed"}]}
{"id": 330342, "image": "COCO_train2014_000000330342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady is cutting birthday / wedding cake\"."}], "task": "refering", "answer_template": "The \"the lady is cutting birthday / wedding cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 24, 25, 26, 27, 28, 36, 37, 38, 39, 40, 48, 49, 50, 51, 52, 60, 61, 62, 63, 72, 73, 74, 84, 85, 86, 87, 96, 97, 98, 99, 108, 109, 110, 111, 120, 121, 122, 123, 132, 133, 134, 135, 144, 145, 146, 147, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209], "bbox": [0.0033633633633633636, 0.08522, 0.48705705705705704, 0.9845600000000001], "iscrowd": 0, "area": 45835.89640000001, "rle": {"size": [500, 333], "counts": "[i0a5U6m3O1O2N2N3M2N3L3N3M2N2N2N2N2N2N2`NeEeM]:U2mEeMU:U2TFfMn9T2\\FeMg9V2bFdMa9U2hFfMZ9T2PGfMR9k1`GPNb8P2a2O100000000O1000000O100000000O10001O000O1000000O100000000O100000000O1000000O100000001N100000000O1000]FbNP5^1hJQOQ5n0iJ_OQ5a0oJ@P5`0oJCo4=QKCo4=PKDP5;QKDP5<PKDnNmNS3_1nMEmNQOR3Z1QNEkNVOQ3U1SNEkN[Oo2P1VNEiN@n2l0XNDVMiNb0k0m3i0YNDWMlN>k0P4e0[NCWMQO9k0S4b0[NCXMTO5k0V4>]NCWMXO1l0X4:_NBWM\\OIP1^44`NBWM^O_OW1h4IbNBVMj1V4eNbNDUMj1W4cNbNJPMf1[4aNdN0jLc1_4^NdN7fL^1d4[NeNm3Y1TLdNP4Y1RLdNR4Z1nKeNU4Y1lKdNX4Y1iKeN[4Y1fKdN_4Y1bKdNb4Z1^KdNf4Y1\\KdNh4Z1XKdNl4Z1UKbNP5\\1PKaNT5]1nJ_NV5UObHo0X2IZ5XO_Hm0X2G]5\\O\\Hj0X2Fb5_OWHi0X2De5CTHg0W2Ci5FQHe0W2Al5JnGb0W2AP6LjGa0Z2[OQ64fG?`2QOP6`0aG=f2dNR6o0YG:T<EnC8S<HoC5R<KoC3R<MPD0Q<0QDLR<3PDJQ<6QDGP<8RDFo;:SDBP<=RD@o;`0SD]On;c0SD[On;e0SDYOn;g0RDWOP<i0PDVOR<i0nCVOS<j0mCUOT<k0lCSOV<m0jCQOX<o0hCPOY<P1gCnN\\<Q1dCnN]<R1S100000010O0001O0002O1N3H7G9G`Rc2"}, "label": "the lady is cutting birthday / wedding cake"}]}
{"id": 330342, "image": "COCO_train2014_000000330342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bride on her wedding day cutting the wedding cake with her groom\"."}], "task": "refering", "answer_template": "The \"a bride on her wedding day cutting the wedding cake with her groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 24, 25, 26, 27, 28, 36, 37, 38, 39, 40, 48, 49, 50, 51, 52, 60, 61, 62, 63, 72, 73, 74, 84, 85, 86, 87, 96, 97, 98, 99, 108, 109, 110, 111, 120, 121, 122, 123, 132, 133, 134, 135, 144, 145, 146, 147, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209], "bbox": [0.0033633633633633636, 0.08522, 0.48705705705705704, 0.9845600000000001], "iscrowd": 0, "area": 45835.89640000001, "rle": {"size": [500, 333], "counts": "[i0a5U6m3O1O2N2N3M2N3L3N3M2N2N2N2N2N2N2`NeEeM]:U2mEeMU:U2TFfMn9T2\\FeMg9V2bFdMa9U2hFfMZ9T2PGfMR9k1`GPNb8P2a2O100000000O1000000O100000000O10001O000O1000000O100000000O100000000O1000000O100000001N100000000O1000]FbNP5^1hJQOQ5n0iJ_OQ5a0oJ@P5`0oJCo4=QKCo4=PKDP5;QKDP5<PKDnNmNS3_1nMEmNQOR3Z1QNEkNVOQ3U1SNEkN[Oo2P1VNEiN@n2l0XNDVMiNb0k0m3i0YNDWMlN>k0P4e0[NCWMQO9k0S4b0[NCXMTO5k0V4>]NCWMXO1l0X4:_NBWM\\OIP1^44`NBWM^O_OW1h4IbNBVMj1V4eNbNDUMj1W4cNbNJPMf1[4aNdN0jLc1_4^NdN7fL^1d4[NeNm3Y1TLdNP4Y1RLdNR4Z1nKeNU4Y1lKdNX4Y1iKeN[4Y1fKdN_4Y1bKdNb4Z1^KdNf4Y1\\KdNh4Z1XKdNl4Z1UKbNP5\\1PKaNT5]1nJ_NV5UObHo0X2IZ5XO_Hm0X2G]5\\O\\Hj0X2Fb5_OWHi0X2De5CTHg0W2Ci5FQHe0W2Al5JnGb0W2AP6LjGa0Z2[OQ64fG?`2QOP6`0aG=f2dNR6o0YG:T<EnC8S<HoC5R<KoC3R<MPD0Q<0QDLR<3PDJQ<6QDGP<8RDFo;:SDBP<=RD@o;`0SD]On;c0SD[On;e0SDYOn;g0RDWOP<i0PDVOR<i0nCVOS<j0mCUOT<k0lCSOV<m0jCQOX<o0hCPOY<P1gCnN\\<Q1dCnN]<R1S100000010O0001O0002O1N3H7G9G`Rc2"}, "label": "a bride on her wedding day cutting the wedding cake with her groom"}]}
{"id": 330342, "image": "COCO_train2014_000000330342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black suit helping his wife cut their cake\"."}], "task": "refering", "answer_template": "The \"a man wearing a black suit helping his wife cut their cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 29, 30, 31, 32, 41, 42, 43, 44, 52, 53, 54, 55, 56, 57, 58, 63, 64, 65, 66, 67, 68, 69, 70, 75, 76, 77, 78, 79, 80, 81, 82, 83, 87, 88, 89, 90, 91, 92, 93, 94, 95, 99, 100, 101, 102, 103, 104, 105, 106, 107, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 135, 136, 137, 138, 139, 140, 141, 142, 143, 147, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 197, 198, 199, 200, 201], "bbox": [0.2637837837837838, 0.08896, 0.9773273273273274, 0.9448200000000001], "iscrowd": 0, "area": 71255.11284999999, "rle": {"size": [500, 333], "counts": "`T[16[?6L4L5K7I6J6Jc0]Oa0^Ob0_O`1`N5K4L5K5K4L5K5J5L5K?A2N3N1N2N2N2N3M2N2N2O1N3M2N2N2N2N3N1N2O1N2N3N1N2O1N2O2M2O1N2O1N3N1N2N2O1N3_NUHRKo7^4dH[K`7b4gHVK]7h4hHQK\\7m4h1M2N2N3M2N2M3N3M2N2fLmDP2\\;^MPE_2W<K5K2O1O1N101O1O1O001O00001O000000000000000000000O100001N100O100O1N2N2N2J7I61O2N1O1O2N3M2N3N1N0000010O01O001O001O001O0O2N2I\\M[Cf2d<7N1O1O1O1N4M3M2N3M3M3M4L4L2N2N1QEPLX:R4bERL^:c4O000000000O101O00000O1000000O2O000L4K5L4N2O1N2O1O100O100O20O001O00100O001O010O001O010O001OE<I6K5M5K5N2M2XN]I]Je6_5]I`Jf6]5\\IbJf6Z5]IeJe6W5^IgJd6W5^IhJd6V5]IhJf6U5[IgIJa0R7a5VIkIKa0X7[5PIPJ2=V7[5kI[J[6^5Q2M4L3L9H7H8G:F9F;F9G9H9G8G:G8H8H9G8H8H9G_c3"}, "label": "a man wearing a black suit helping his wife cut their cake"}]}
{"id": 330342, "image": "COCO_train2014_000000330342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a groom wearing a dark suit\"."}], "task": "refering", "answer_template": "The \"a groom wearing a dark suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 29, 30, 31, 32, 41, 42, 43, 44, 52, 53, 54, 55, 56, 57, 58, 63, 64, 65, 66, 67, 68, 69, 70, 75, 76, 77, 78, 79, 80, 81, 82, 83, 87, 88, 89, 90, 91, 92, 93, 94, 95, 99, 100, 101, 102, 103, 104, 105, 106, 107, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 135, 136, 137, 138, 139, 140, 141, 142, 143, 147, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 197, 198, 199, 200, 201], "bbox": [0.2637837837837838, 0.08896, 0.9773273273273274, 0.9448200000000001], "iscrowd": 0, "area": 71255.11284999999, "rle": {"size": [500, 333], "counts": "`T[16[?6L4L5K7I6J6Jc0]Oa0^Ob0_O`1`N5K4L5K5K4L5K5J5L5K?A2N3N1N2N2N2N3M2N2N2O1N3M2N2N2N2N3N1N2O1N2N3N1N2O1N2O2M2O1N2O1N3N1N2N2O1N3_NUHRKo7^4dH[K`7b4gHVK]7h4hHQK\\7m4h1M2N2N3M2N2M3N3M2N2fLmDP2\\;^MPE_2W<K5K2O1O1N101O1O1O001O00001O000000000000000000000O100001N100O100O1N2N2N2J7I61O2N1O1O2N3M2N3N1N0000010O01O001O001O001O0O2N2I\\M[Cf2d<7N1O1O1O1N4M3M2N3M3M3M4L4L2N2N1QEPLX:R4bERL^:c4O000000000O101O00000O1000000O2O000L4K5L4N2O1N2O1O100O100O20O001O00100O001O010O001O010O001OE<I6K5M5K5N2M2XN]I]Je6_5]I`Jf6]5\\IbJf6Z5]IeJe6W5^IgJd6W5^IhJd6V5]IhJf6U5[IgIJa0R7a5VIkIKa0X7[5PIPJ2=V7[5kI[J[6^5Q2M4L3L9H7H8G:F9F;F9G9H9G8G:G8H8H9G8H8H9G_c3"}, "label": "a groom wearing a dark suit"}]}
{"id": 141927, "image": "COCO_train2014_000000141927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an air india plane parked near a terminal\"."}], "task": "refering", "answer_template": "The \"an air india plane parked near a terminal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 155, 156, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 249, 250, 263, 264, 265], "bbox": [0.117453125, 0.31245833333333334, 1.0, 0.6681666666666667], "iscrowd": 0, "area": 25844.711750000013, "rle": {"size": [480, 640], "counts": "fkS12n>3M2N2N2M4M2N2N2N3M2N2N2N3M2N2N2N000000000O0100000000000000O10000000000000O10O100000000000000O10000000000000O010000000000O1N2N2N2M3N2N2N2M3N2N1O20000000000000000O100000000000O100000000000O10000000000000000000O10O1000000000000000000000000O1000O1000000000001O2N3M2N3M3M2N3M2N1O1N101O1O1O001O1O1O001O1O1O001O1O1N101O1O001O1O1O001O1O1O00000000000O1000000000001O0000000000000000000O100000000000001O000000000000000O103M2N2N3M2N3M2N3M2N2N2N00O100000O1000O10000000000O10000000000O1000O100000O1000000000000O10000000000O01000000000O10000000000O10000000O10O10000000000O1000000M3M3M3M3M3M2N3M3N2000000000000O10000000000000O10O10000000000000000O1000000000O10000000O10000000000000000O100000O1000000000O1000000000000000000O10O10000000000000O1000000000000000O10O10000000000000000O100000000000O1000O100000000000N2O1N2O0O2O1N2O1N2N2O1N2O0O2O1N2O1N2N2O1N2O0O2O1N2O2M2N2O2M2O2M2O1N3N1N3MXO`DTN_;m1aDSN];o1cDRN[;o1eDQNY;Q2gDPNW;m2NUOkDoMS;S2mDnMQ;S2oDmMo:U2QElMW=KiB5W=JjB6V=JjBCO8V=5kBC07V=5jBD08U=4kBD17T=4lBE16S=5lBE26Q=5mBE26Q=5mBD47n<5nBD56m<5oBE47m<3oBF56l<4oBF65k<5oBF75i<5PCF75i<5oBF:5f<4QCG:4e<5QCG:4e<5QCG;4c<5RCG<3c<5QCH=3a<4SCI<3a<4SCH>3_<5SCH?3]<5TCH?3]<5TCH`03[<4VCI`02Z<5VCIa01Z<5UCJa02Y<4VCIc02W<5VCId02U<4XCJc02U<4XCJd01T<5XCJe01R<5YCJe01R<5XCKg00Q<4XCKi00o;4YCLi0On;5YCLi00m;4ZCLj0Ol;5ZCLk0Oj;5[CLk0Oj;4\\CMk0Ni;5\\CLm0Nh;5[CMn0Mg;6[CMn0Lh;7ZCMo0Kg;7[CNo0Jf;8[CNo0Jf;8[CMo=3QBMo=3QBMo=2RBNo=1QBOo=1QBOo=1PB0P>0PBOQ>1oAOQ>0PB0P>0PB0P>0PB0Q>OoA1Q>OoA1Q>NPB1Q>OoA1Q>OoA1Q>OoA1Q>OoA1Q>NPB2Q>MoA3Q>MoA2R>NnA2R>NnA2R>MoA3Q>MoA3Q>MoA3Q>MnA4_>O1000000000000O1000000000000O10000000O1000O10000000000O100000000000oi20^]M"}, "label": "an air india plane parked near a terminal"}]}
{"id": 141927, "image": "COCO_train2014_000000141927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the airplane parked at the terminal\"."}], "task": "refering", "answer_template": "The \"the airplane parked at the terminal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 155, 156, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 249, 250, 263, 264, 265], "bbox": [0.117453125, 0.31245833333333334, 1.0, 0.6681666666666667], "iscrowd": 0, "area": 25844.711750000013, "rle": {"size": [480, 640], "counts": "fkS12n>3M2N2N2M4M2N2N2N3M2N2N2N3M2N2N2N000000000O0100000000000000O10000000000000O10O100000000000000O10000000000000O010000000000O1N2N2N2M3N2N2N2M3N2N1O20000000000000000O100000000000O100000000000O10000000000000000000O10O1000000000000000000000000O1000O1000000000001O2N3M2N3M3M2N3M2N1O1N101O1O1O001O1O1O001O1O1O001O1O1N101O1O001O1O1O001O1O1O00000000000O1000000000001O0000000000000000000O100000000000001O000000000000000O103M2N2N3M2N3M2N3M2N2N2N00O100000O1000O10000000000O10000000000O1000O100000O1000000000000O10000000000O01000000000O10000000000O10000000O10O10000000000O1000000M3M3M3M3M3M2N3M3N2000000000000O10000000000000O10O10000000000000000O1000000000O10000000O10000000000000000O100000O1000000000O1000000000000000000O10O10000000000000O1000000000000000O10O10000000000000000O100000000000O1000O100000000000N2O1N2O0O2O1N2O1N2N2O1N2O0O2O1N2O1N2N2O1N2O0O2O1N2O2M2N2O2M2O2M2O1N3N1N3MXO`DTN_;m1aDSN];o1cDRN[;o1eDQNY;Q2gDPNW;m2NUOkDoMS;S2mDnMQ;S2oDmMo:U2QElMW=KiB5W=JjB6V=JjBCO8V=5kBC07V=5jBD08U=4kBD17T=4lBE16S=5lBE26Q=5mBE26Q=5mBD47n<5nBD56m<5oBE47m<3oBF56l<4oBF65k<5oBF75i<5PCF75i<5oBF:5f<4QCG:4e<5QCG:4e<5QCG;4c<5RCG<3c<5QCH=3a<4SCI<3a<4SCH>3_<5SCH?3]<5TCH?3]<5TCH`03[<4VCI`02Z<5VCIa01Z<5UCJa02Y<4VCIc02W<5VCId02U<4XCJc02U<4XCJd01T<5XCJe01R<5YCJe01R<5XCKg00Q<4XCKi00o;4YCLi0On;5YCLi00m;4ZCLj0Ol;5ZCLk0Oj;5[CLk0Oj;4\\CMk0Ni;5\\CLm0Nh;5[CMn0Mg;6[CMn0Lh;7ZCMo0Kg;7[CNo0Jf;8[CNo0Jf;8[CMo=3QBMo=3QBMo=2RBNo=1QBOo=1QBOo=1PB0P>0PBOQ>1oAOQ>0PB0P>0PB0P>0PB0Q>OoA1Q>OoA1Q>NPB1Q>OoA1Q>OoA1Q>OoA1Q>OoA1Q>NPB2Q>MoA3Q>MoA2R>NnA2R>NnA2R>MoA3Q>MoA3Q>MoA3Q>MnA4_>O1000000000000O1000000000000O10000000O1000O10000000000O100000000000oi20^]M"}, "label": "the airplane parked at the terminal"}]}
{"id": 174698, "image": "COCO_train2014_000000174698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fully visible chair sitting at the round table\"."}], "task": "refering", "answer_template": "The \"a fully visible chair sitting at the round table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 256, 257, 277, 278, 279, 280, 299, 300, 301, 302, 303, 323, 325, 326], "bbox": [0.03509375, 0.6693144208037826, 0.230140625, 1.0], "iscrowd": 0, "area": 6142.629400000001, "rle": {"size": [423, 640], "counts": "kk95Q=4L5J:G7I1O1O1O1N2O2NBRDEm;<YD^Of;c0`DWO_;j0=2O001O1O1O1O1O2O6K5J6J6K5I2O0000POTEEl:;]E\\Oc:d0fETOY:l0Q1000O1000000000000000000O1000000000O1000000000000000000000000000001O000000A?L5K8G9H8H9I6I7J6I7J3M2O01000O10O0100O10O0100O10RMjFj1U9TNWGd1h8ZNcG^1[8dNmGT1R8XNlF:\\1T1g7cNoF7f1k0Z7nNSG5k1e0Q7VOVG3o1b0j6[OZG1Q2>e6@cGJn1a0_6D^J7b5HcJ6Z5KUGL_36Z5NWGOa3OW52WG3`<LaC7\\<IcCM04^<ObC0M1`<1cCLN2`<2bCKO3`<1bCKN3b<2cCN^<1bCN_<3_CMc<5YCLh<82O6Ie[[6"}, "label": "a fully visible chair sitting at the round table"}]}
{"id": 174698, "image": "COCO_train2014_000000174698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair was infront of table\"."}], "task": "refering", "answer_template": "The \"a chair was infront of table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 256, 257, 277, 278, 279, 280, 299, 300, 301, 302, 303, 323, 325, 326], "bbox": [0.03509375, 0.6693144208037826, 0.230140625, 1.0], "iscrowd": 0, "area": 6142.629400000001, "rle": {"size": [423, 640], "counts": "kk95Q=4L5J:G7I1O1O1O1N2O2NBRDEm;<YD^Of;c0`DWO_;j0=2O001O1O1O1O1O2O6K5J6J6K5I2O0000POTEEl:;]E\\Oc:d0fETOY:l0Q1000O1000000000000000000O1000000000O1000000000000000000000000000001O000000A?L5K8G9H8H9I6I7J6I7J3M2O01000O10O0100O10O0100O10RMjFj1U9TNWGd1h8ZNcG^1[8dNmGT1R8XNlF:\\1T1g7cNoF7f1k0Z7nNSG5k1e0Q7VOVG3o1b0j6[OZG1Q2>e6@cGJn1a0_6D^J7b5HcJ6Z5KUGL_36Z5NWGOa3OW52WG3`<LaC7\\<IcCM04^<ObC0M1`<1cCLN2`<2bCKO3`<1bCKN3b<2cCN^<1bCN_<3_CMc<5YCLh<82O6Ie[[6"}, "label": "a chair was infront of table"}]}
{"id": 494190, "image": "COCO_train2014_000000494190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mirror and the side of the car\"."}], "task": "refering", "answer_template": "The \"the mirror and the side of the car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.009416666666666665, 0.998828125, 0.985875], "iscrowd": 0, "area": 228509.17569999996, "rle": {"size": [480, 640], "counts": "T3l4T:i6VI1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O010O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O10000000000001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001mKUOoIk0Q6UOoIk0Q6UOoIk0Q6UOoIk0Q6VOnIj0R6VOnIk0P6VOPJj0P6WOoIi0Q6WOnIj0R6VOnIj0R6WOmIi0S6WOmIj0R6VOnIj0R6WOmIi0S6WOmIi0S6WOmIi0S6WOmIi0R6YOmIh0R6XOnIh0R6XOnIh0R6YOmIg0S6YOmIg0S6YOmIg0S6ZOlIg0S6YOmIg0S6YOmIg0S6ZOlIf0S6[OmIe0S6[OmIe0S6[OmIf0R6[OmIe0S6[OmIe0S6[OmIe0S6\\OlId0T6\\OlId0T6\\OlIe0S6\\OlId0S6]OmIc0S6]OmIc0S6^OlIb0T6^OlIb0T6^OlIc0S6]OmIc0S6^OlIb0T6^OlIb0T6^OlIb0T6_OkIa0T6@lIa0S6_OmIa0S6@lI`0T6@lI`0T6@lI`0T6AkI?U6AkI`0T6@lI`0T6@lI`0T6AkI?U6AkI?T6BlI>T6CkI>T6BlI>T6BlI>T6CkI=U6CkI=U6CkI>T6CkI=U6CkI=U6CkI=T6DlI<T6EkI;U6EkI<T6DlI<T6EkI;U6EkI;U6EkI;U6FjI:V6FjI;U6EkI;T6GkI9U6GkI9U6GkI9U6GkI9U6HjI9U6GkI9U6GkI9U6HjI8V6HjI8V6HjI8U6JjI7U6IkI7U6IkI7U6JjI6V6JjI6V6JjI6V6JjI7U6JjI6V6JjI6V6JjI6U6LjI4V6LjI4V6LjI5U6LjI4V6LjI4V6LjI4V6MiI3W6MiI3W6MiI4V6LjI4V6MiI3V6NjI2V6NjI2V6OiI1W6OiI2V6NjI2V6OiI1W6OiI1W6OiI1W60hI0X60hI1V60jI0V60jI0V61iIOW61iIOW61iI0V61iIOW61iIOW61iIOW62hINX62hINW63iINV63iIMW63iIMW63iIMW63iIMW64hILX64hIMW63iIMW64hILX64hILW65iIKW66hIJX66hIKW65iIKW66hIJX66hIJX66hIJX66hIJX67gIJX66hIJX66hIJW68hIHX68hIHX68hIHX69fIIY67gIIY67gIIY68fIHZ68fIHZ68fIHZ68fIIX69gIGY69gIGY69gIGY6:fIFZ6:fIFZ6:fIGY6:fIFZ6:fIFZ6:fIFZ6;eIEZ6<fIDZ6<fIEY6;gIEY6<fIDZ6<fIDZ6<fIDZ6=eIC[6=eIDZ6<fIDZ6=eICZ6>fIBZ6>fIBZ6?eIBZ6>fIBZ6>fIBZ6>fIBZ6?eIA[6?eIA[6?eIBZ6?eIA[6?eIAZ6`0fI@Z6a0eI_O[6a0eI_O[6a0eI@Z6a0eI_O[6a0eI_O[6a0eI_O[6a0eI_O[6b0dI^O\\6b0dI_OZ6b0fI^OZ6c0eI]O[6c0eI]O[6c0eI]O[6d0dI\\O\\6d0dI]O[6c0eI]O[6d0dI\\O\\6d0dI\\O\\6d0dI\\O[6e0eI[O[6f0dI[O[6e0eI[O[6e0eI[O[6f0dIZO\\6f0dIZO\\6f0dIZO\\6g0cIZO\\6f0dIZO\\6f0dIZO[6h0dIXO\\6h0dIXO\\6h0dIXO\\6h0dIYO[6h0dIXO\\6h0dIXO\\6h0dIXO\\6i0cIWO]6i0cIWO]6i0cIXO[6j0dIVO\\6j0dIVO\\6j0dIVO\\6k0cIUO]6k0cIUO]6k0cIVO\\6j0dIVO\\6k0cIUO]6k0cIUO]6k0cIUO]6l0bITO]6m0cITO\\6l0dITO\\6m0cISO]6m0cISO]6m0cISO]6n0bISO]6m0cISO]6m0cISO]6m0cISO]6n0bIRO]6o0cIQO]6o0cIRO\\6o0cIQO]6o0cIQO]6o0cIQO]6P1bIPO^6P1bIPO^6P1bIQO]6P1bIPO^6P1bIPO]6Q1cIoN]6Q1cIoN]6R1bInN^6R1bIoN]6Q1cIoN]6R1bInN^6R1bInN^6R1bInN^6S1aImN_6S1aInN]6S1cImN]6T1bIlN^6T1bIlN^6T1bIlN^6T1bIlN^6U1aIlN^6T1bIlN^6T1bIlN^6U1aIkN_6U1aIkN_6U1aIkN^6W1aIjN^6V1bIjN^6V1bIjN^6W1aIiN_6W1aIiN_6W1aIiN_6W1aIjN^6W1aIiN_6W1aIiN_6W1aIiN^6Y1aIgN_6Y1aIgN_6Y1aIhN^6Y1aIgN_6Y1aIgN_6Y1`IhN`6Y1_IgNa6Y1_IgNa6Y1_IhN`6X1`IhN_6Z1`IfN`6Z1`IfN`6Z1`IgN_6Z1`IgN_6Y1aIhN^6X1bIiN]6X1bIiN]6W1cIjN\\6V1dIkN[6V1dIkNZ6V1fIkNY6U1gIlNX6T1hImNW6T1hImNW6S1iInNV6R1jIoNU6R1jIoNU6Q1kIPOT6P1lIQOS6P1lIPOS6Q1mIPOR6P1nIQOQ6P1nIQOQ6o0oIROP6n0PJSOo5m0QJTOn5m0QJTOn5l0RJUOm5k0SJVOl5k0SJVOl5j0TJWOj5j0VJWOi5j0VJWOh5j0XJWOf5j0ZJWOe5j0ZJWOd5j0\\JZO_5g0aJ^OZ5b0fJBU5`0jJDQ5=oJGm49SKKh47WKMd44\\K0_41aK4Z4MeK7V4JjK:Q4GoK=m3DRL`0i3AWLc0d3^O\\Lf0_3[OaLi0[3XOdLm0V3TOjLP1Q3QOoLS1m2nNRMV1i2kNWMY1d2hN\\M\\1`2eN_M_1\\2bNdMc1V2^NjMf1Q2\\NnMh1n1XNRNl1i1UNWNo1d1RN\\NR2`1oM_NU2\\1lMdNX2W1iMiN\\2R1eMmN_2n0bMROb2i0_MWOe2d0]M[Og2a0YM_Ok2<VMDn27TMHQ33oLMU3NlL2X3IiL7Q9O00001O00001O00001O0000001O001O1O2N<"}, "label": "the mirror and the side of the car"}]}
{"id": 494190, "image": "COCO_train2014_000000494190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a side - view mirror\"."}], "task": "refering", "answer_template": "The \"a side - view mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.009416666666666665, 0.998828125, 0.985875], "iscrowd": 0, "area": 228509.17569999996, "rle": {"size": [480, 640], "counts": "T3l4T:i6VI1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O010O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O10000000000001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001mKUOoIk0Q6UOoIk0Q6UOoIk0Q6UOoIk0Q6VOnIj0R6VOnIk0P6VOPJj0P6WOoIi0Q6WOnIj0R6VOnIj0R6WOmIi0S6WOmIj0R6VOnIj0R6WOmIi0S6WOmIi0S6WOmIi0S6WOmIi0R6YOmIh0R6XOnIh0R6XOnIh0R6YOmIg0S6YOmIg0S6YOmIg0S6ZOlIg0S6YOmIg0S6YOmIg0S6ZOlIf0S6[OmIe0S6[OmIe0S6[OmIf0R6[OmIe0S6[OmIe0S6[OmIe0S6\\OlId0T6\\OlId0T6\\OlIe0S6\\OlId0S6]OmIc0S6]OmIc0S6^OlIb0T6^OlIb0T6^OlIc0S6]OmIc0S6^OlIb0T6^OlIb0T6^OlIb0T6_OkIa0T6@lIa0S6_OmIa0S6@lI`0T6@lI`0T6@lI`0T6AkI?U6AkI`0T6@lI`0T6@lI`0T6AkI?U6AkI?T6BlI>T6CkI>T6BlI>T6BlI>T6CkI=U6CkI=U6CkI>T6CkI=U6CkI=U6CkI=T6DlI<T6EkI;U6EkI<T6DlI<T6EkI;U6EkI;U6EkI;U6FjI:V6FjI;U6EkI;T6GkI9U6GkI9U6GkI9U6GkI9U6HjI9U6GkI9U6GkI9U6HjI8V6HjI8V6HjI8U6JjI7U6IkI7U6IkI7U6JjI6V6JjI6V6JjI6V6JjI7U6JjI6V6JjI6V6JjI6U6LjI4V6LjI4V6LjI5U6LjI4V6LjI4V6LjI4V6MiI3W6MiI3W6MiI4V6LjI4V6MiI3V6NjI2V6NjI2V6OiI1W6OiI2V6NjI2V6OiI1W6OiI1W6OiI1W60hI0X60hI1V60jI0V60jI0V61iIOW61iIOW61iI0V61iIOW61iIOW61iIOW62hINX62hINW63iINV63iIMW63iIMW63iIMW63iIMW64hILX64hIMW63iIMW64hILX64hILW65iIKW66hIJX66hIKW65iIKW66hIJX66hIJX66hIJX66hIJX67gIJX66hIJX66hIJW68hIHX68hIHX68hIHX69fIIY67gIIY67gIIY68fIHZ68fIHZ68fIHZ68fIIX69gIGY69gIGY69gIGY6:fIFZ6:fIFZ6:fIGY6:fIFZ6:fIFZ6:fIFZ6;eIEZ6<fIDZ6<fIEY6;gIEY6<fIDZ6<fIDZ6<fIDZ6=eIC[6=eIDZ6<fIDZ6=eICZ6>fIBZ6>fIBZ6?eIBZ6>fIBZ6>fIBZ6>fIBZ6?eIA[6?eIA[6?eIBZ6?eIA[6?eIAZ6`0fI@Z6a0eI_O[6a0eI_O[6a0eI@Z6a0eI_O[6a0eI_O[6a0eI_O[6a0eI_O[6b0dI^O\\6b0dI_OZ6b0fI^OZ6c0eI]O[6c0eI]O[6c0eI]O[6d0dI\\O\\6d0dI]O[6c0eI]O[6d0dI\\O\\6d0dI\\O\\6d0dI\\O[6e0eI[O[6f0dI[O[6e0eI[O[6e0eI[O[6f0dIZO\\6f0dIZO\\6f0dIZO\\6g0cIZO\\6f0dIZO\\6f0dIZO[6h0dIXO\\6h0dIXO\\6h0dIXO\\6h0dIYO[6h0dIXO\\6h0dIXO\\6h0dIXO\\6i0cIWO]6i0cIWO]6i0cIXO[6j0dIVO\\6j0dIVO\\6j0dIVO\\6k0cIUO]6k0cIUO]6k0cIVO\\6j0dIVO\\6k0cIUO]6k0cIUO]6k0cIUO]6l0bITO]6m0cITO\\6l0dITO\\6m0cISO]6m0cISO]6m0cISO]6n0bISO]6m0cISO]6m0cISO]6m0cISO]6n0bIRO]6o0cIQO]6o0cIRO\\6o0cIQO]6o0cIQO]6o0cIQO]6P1bIPO^6P1bIPO^6P1bIQO]6P1bIPO^6P1bIPO]6Q1cIoN]6Q1cIoN]6R1bInN^6R1bIoN]6Q1cIoN]6R1bInN^6R1bInN^6R1bInN^6S1aImN_6S1aInN]6S1cImN]6T1bIlN^6T1bIlN^6T1bIlN^6T1bIlN^6U1aIlN^6T1bIlN^6T1bIlN^6U1aIkN_6U1aIkN_6U1aIkN^6W1aIjN^6V1bIjN^6V1bIjN^6W1aIiN_6W1aIiN_6W1aIiN_6W1aIjN^6W1aIiN_6W1aIiN_6W1aIiN^6Y1aIgN_6Y1aIgN_6Y1aIhN^6Y1aIgN_6Y1aIgN_6Y1`IhN`6Y1_IgNa6Y1_IgNa6Y1_IhN`6X1`IhN_6Z1`IfN`6Z1`IfN`6Z1`IgN_6Z1`IgN_6Y1aIhN^6X1bIiN]6X1bIiN]6W1cIjN\\6V1dIkN[6V1dIkNZ6V1fIkNY6U1gIlNX6T1hImNW6T1hImNW6S1iInNV6R1jIoNU6R1jIoNU6Q1kIPOT6P1lIQOS6P1lIPOS6Q1mIPOR6P1nIQOQ6P1nIQOQ6o0oIROP6n0PJSOo5m0QJTOn5m0QJTOn5l0RJUOm5k0SJVOl5k0SJVOl5j0TJWOj5j0VJWOi5j0VJWOh5j0XJWOf5j0ZJWOe5j0ZJWOd5j0\\JZO_5g0aJ^OZ5b0fJBU5`0jJDQ5=oJGm49SKKh47WKMd44\\K0_41aK4Z4MeK7V4JjK:Q4GoK=m3DRL`0i3AWLc0d3^O\\Lf0_3[OaLi0[3XOdLm0V3TOjLP1Q3QOoLS1m2nNRMV1i2kNWMY1d2hN\\M\\1`2eN_M_1\\2bNdMc1V2^NjMf1Q2\\NnMh1n1XNRNl1i1UNWNo1d1RN\\NR2`1oM_NU2\\1lMdNX2W1iMiN\\2R1eMmN_2n0bMROb2i0_MWOe2d0]M[Og2a0YM_Ok2<VMDn27TMHQ33oLMU3NlL2X3IiL7Q9O00001O00001O00001O0000001O001O1O2N<"}, "label": "a side - view mirror"}]}
{"id": 338549, "image": "COCO_train2014_000000338549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a double decker bus in front of another double decker bus\"."}], "task": "refering", "answer_template": "The \"a double decker bus in front of another double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 84, 85, 86, 87, 88, 103, 104, 105, 107, 108, 109, 110, 111, 126, 127, 130, 131, 132, 133, 134, 156, 157], "bbox": [0.49387499999999995, 0.012317596566523606, 0.8624218750000001, 0.4137982832618026], "iscrowd": 0, "area": 25832.142250000004, "rle": {"size": [466, 640], "counts": "Tn_48j=`0@`0@a0@?@`0@a0B=H8I8M2O101O00001O0000001O0O101O0000O1K5M3N2M3O1O001N2O1O1O1O1O1N2O1O1O001O1gMQEm0Q;QOSEl0n:SOUEj0l:UOVEi0k:VOXEg0i:XOZEd0h:[OZEa0i:^OZE=i:BZE:h:EZE7h:IYE5i:JYE4h:KYE3i:LYE1i:OWE0j:0WEMk:3VEJl:6XEEi:;ZE@h:`0[E[Og:e0\\EWOe:i0b10O10000000000000000000000O1000001O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O<D:F3M2N3L3N3M2N3M3M3XFSMe7P3THSMm7P3mGSMS8P3gGSMY8P3bGRM^8Q3\\GQMf8P3VGRMj8P3QGSMo8P3lFRMU9o2fFSM[9o2aFSM`9n2[FUMe9m2WFUMi9i32O00001O0O101O000O2O00001O001O00001O001O00001O001O00001O001O000010O01O000010O01O000010O01O001O01O01O001O01O01O001O01O01O001O01O100O100O1O100O100O10O02N101N101N1O101N101N1O2OO0O2O00001N101O00001N101O000O2O0N3N1N3N1N2O2M2K6mNR1nNR1nNlPX1"}, "label": "a double decker bus in front of another double decker bus"}]}
{"id": 338549, "image": "COCO_train2014_000000338549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red color double tucker bus on the road\"."}], "task": "refering", "answer_template": "The \"a red color double tucker bus on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 84, 85, 86, 87, 88, 103, 104, 105, 107, 108, 109, 110, 111, 126, 127, 130, 131, 132, 133, 134, 156, 157], "bbox": [0.49387499999999995, 0.012317596566523606, 0.8624218750000001, 0.4137982832618026], "iscrowd": 0, "area": 25832.142250000004, "rle": {"size": [466, 640], "counts": "Tn_48j=`0@`0@a0@?@`0@a0B=H8I8M2O101O00001O0000001O0O101O0000O1K5M3N2M3O1O001N2O1O1O1O1O1N2O1O1O001O1gMQEm0Q;QOSEl0n:SOUEj0l:UOVEi0k:VOXEg0i:XOZEd0h:[OZEa0i:^OZE=i:BZE:h:EZE7h:IYE5i:JYE4h:KYE3i:LYE1i:OWE0j:0WEMk:3VEJl:6XEEi:;ZE@h:`0[E[Og:e0\\EWOe:i0b10O10000000000000000000000O1000001O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O<D:F3M2N3L3N3M2N3M3M3XFSMe7P3THSMm7P3mGSMS8P3gGSMY8P3bGRM^8Q3\\GQMf8P3VGRMj8P3QGSMo8P3lFRMU9o2fFSM[9o2aFSM`9n2[FUMe9m2WFUMi9i32O00001O0O101O000O2O00001O001O00001O001O00001O001O00001O001O000010O01O000010O01O000010O01O001O01O01O001O01O01O001O01O01O001O01O100O100O1O100O100O10O02N101N101N1O101N101N1O2OO0O2O00001N101O00001N101O000O2O0N3N1N3N1N2O2M2K6mNR1nNR1nNlPX1"}, "label": "a red color double tucker bus on the road"}]}
{"id": 223871, "image": "COCO_train2014_000000223871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"kid sitting on couch\"."}], "task": "refering", "answer_template": "The \"kid sitting on couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 179, 180, 181, 182, 192, 193, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.006, 0.24456250000000002, 0.8449375, 0.986140625], "iscrowd": 0, "area": 105040.2665, "rle": {"size": [640, 480], "counts": "\\^2:[c0=C=B>F:M3M3M3M3M3M3M3M3M2N1N3N1N2O2M2O1N3N1O2M2O1N3N1N2O2M2O2M200O2O000O100O10000O10000O10000O100O10000O10000O10000O100O10000O1N2N2M3N2M3M2O2M3N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N200O1O100O1O1O100O1O1O10O01O1O10000000000001O000000000O100000001O0000000000000000VMZIZHf6e7]IYHc6g7^IYH`6g7cIWH]6h7eIWH[6i7fIVHZ6i7iIUHW6k7jITHV6k7mITHR6k7PJUHo5k7RJTHn5k7UJTHj5k7XJTHh5l7ZJSHe5l7]JSHc5m7^JSHa5l7bJSH]5l7eJSHZ5n7hJQHW5n7kJQHU5n7nJQHQ5o7PKQHo4n7SKQHm4Q8RKoGm4U8QKiGo4Z8oJfGP5^8mJbGR5`8nJ^GR5c8nJ]GQ5d8oJ[GQ5g8oJXGP5i8PKVGP5k8PKUGo4l8QKTGm4o8SKoFm4R9SKnFl4S9UKkFk4W9VKgFi4Z9XKeFg4\\9ZKbFf4_9\\K_Fc4c9]K[Fc4f9^KYFa4h9aKUF_4l9bKSF]4o9dKoE[4R:fKlEZ4U:gKjEX4W:jKfEU4]:kKbET4_:mK`ER4a:PL\\EP4e:oK\\EP4f:oKYEQ4h:nKYEQ4h:oKXEP4i:PLVEP4l:nKUEP4m:PLREP4o:oKREo3P;QLoDo3S;PLmDn3U;QLlDn3U;RLiDo3X;PLeDR4^;mK_DU4`;mK\\DU4d;lKZDV4d;mKXDU4h;lKVDV4i;lKTDV4j;mKRDU4n;lKPDV4o;m1O2O001N2O001N101O0O2O1O0O2O001N101N101O1N101N101N101N2O0O2O0O2O001N2O0PMWGiJk8T5dG_J\\8_5SHSJo7k5_HhIa7V6nH\\IT7b6RIYIo6d6YIUIh6i6`IQIa6m6fIlH[6R7mIhHT6V7[3N3M2O2M3N1N3N110O010O10O010O01O010O10O010O010O010O0100O010O010O010O2O1N2O1O1O001O1O1O001O1O00100O001O1O001O1O0000000001O00000O100000000000001O000000000000000O1000001O0cFfIT5Z6kJhIT5X6jJkIU5V6hJmIW5T6gJnIX5S6eJPJZ5Q6cJQJ]5o5bJSJ]5n5`JUJ_5l5^JWJa5j5^JWJa5j5^JWJ`5k5_JVJ`5k5_JVJ`5k5_JUJa5l5^JUJa5l5^JUJ^5o5aJRJ[5R6dJoIX5U6fJmIV5W6hJkIT5Y6jJiIS5Z6kJhIQ5\\6mJfIo4]6PKeIl4`6QKcIj4c6RK_Ik4f6QK\\Il4i6PKYIm4l6oJWIm4o6lJUIQ5P7jJSIS5R7gJRIU5T7eJQIW5U7cJnHZ5W7`JmH]5X7^JkH_5Z7[JkHa5l:M3M3M3M4L3M3K5I8I6I7J6I8I9Fc0^Oa0^Ob0SOn0VNfl]1"}, "label": "kid sitting on couch"}]}
{"id": 223871, "image": "COCO_train2014_000000223871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid eating pizza and winking\"."}], "task": "refering", "answer_template": "The \"a kid eating pizza and winking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 179, 180, 181, 182, 192, 193, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.006, 0.24456250000000002, 0.8449375, 0.986140625], "iscrowd": 0, "area": 105040.2665, "rle": {"size": [640, 480], "counts": "\\^2:[c0=C=B>F:M3M3M3M3M3M3M3M3M2N1N3N1N2O2M2O1N3N1O2M2O1N3N1N2O2M2O2M200O2O000O100O10000O10000O10000O100O10000O10000O10000O100O10000O1N2N2M3N2M3M2O2M3N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N200O1O100O1O1O100O1O1O10O01O1O10000000000001O000000000O100000001O0000000000000000VMZIZHf6e7]IYHc6g7^IYH`6g7cIWH]6h7eIWH[6i7fIVHZ6i7iIUHW6k7jITHV6k7mITHR6k7PJUHo5k7RJTHn5k7UJTHj5k7XJTHh5l7ZJSHe5l7]JSHc5m7^JSHa5l7bJSH]5l7eJSHZ5n7hJQHW5n7kJQHU5n7nJQHQ5o7PKQHo4n7SKQHm4Q8RKoGm4U8QKiGo4Z8oJfGP5^8mJbGR5`8nJ^GR5c8nJ]GQ5d8oJ[GQ5g8oJXGP5i8PKVGP5k8PKUGo4l8QKTGm4o8SKoFm4R9SKnFl4S9UKkFk4W9VKgFi4Z9XKeFg4\\9ZKbFf4_9\\K_Fc4c9]K[Fc4f9^KYFa4h9aKUF_4l9bKSF]4o9dKoE[4R:fKlEZ4U:gKjEX4W:jKfEU4]:kKbET4_:mK`ER4a:PL\\EP4e:oK\\EP4f:oKYEQ4h:nKYEQ4h:oKXEP4i:PLVEP4l:nKUEP4m:PLREP4o:oKREo3P;QLoDo3S;PLmDn3U;QLlDn3U;RLiDo3X;PLeDR4^;mK_DU4`;mK\\DU4d;lKZDV4d;mKXDU4h;lKVDV4i;lKTDV4j;mKRDU4n;lKPDV4o;m1O2O001N2O001N101O0O2O1O0O2O001N101N101O1N101N101N101N2O0O2O0O2O001N2O0PMWGiJk8T5dG_J\\8_5SHSJo7k5_HhIa7V6nH\\IT7b6RIYIo6d6YIUIh6i6`IQIa6m6fIlH[6R7mIhHT6V7[3N3M2O2M3N1N3N110O010O10O010O01O010O10O010O010O010O0100O010O010O010O2O1N2O1O1O001O1O1O001O1O00100O001O1O001O1O0000000001O00000O100000000000001O000000000000000O1000001O0cFfIT5Z6kJhIT5X6jJkIU5V6hJmIW5T6gJnIX5S6eJPJZ5Q6cJQJ]5o5bJSJ]5n5`JUJ_5l5^JWJa5j5^JWJa5j5^JWJ`5k5_JVJ`5k5_JVJ`5k5_JUJa5l5^JUJa5l5^JUJ^5o5aJRJ[5R6dJoIX5U6fJmIV5W6hJkIT5Y6jJiIS5Z6kJhIQ5\\6mJfIo4]6PKeIl4`6QKcIj4c6RK_Ik4f6QK\\Il4i6PKYIm4l6oJWIm4o6lJUIQ5P7jJSIS5R7gJRIU5T7eJQIW5U7cJnHZ5W7`JmH]5X7^JkH_5Z7[JkHa5l:M3M3M3M4L3M3K5I8I6I7J6I8I9Fc0^Oa0^Ob0SOn0VNfl]1"}, "label": "a kid eating pizza and winking"}]}
{"id": 223871, "image": "COCO_train2014_000000223871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting next to a child eating a slice of pizza\"."}], "task": "refering", "answer_template": "The \"a woman sitting next to a child eating a slice of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 17, 18, 34, 35, 36, 51, 52, 53, 68, 69, 85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291], "bbox": [0.001, 0.001875, 0.24570833333333333, 0.785140625], "iscrowd": 0, "area": 42358.284449999985, "rle": {"size": [640, 480], "counts": "1c0]c0h9XFV1jNV1jNV1jNU1kNc0]O000000O2O00000O100000000O1000000O1000000O2O1O1O1O1dD\\KY7f4fH[KY7f4fH[KY7g4dH[K[7g4cHZK[7i4cHWK]7m4^HUKa7Q5YHoJg7W5SHjJk7]5nGdJR8b5hG^JW8d5hG]JW8c5hG^JW8c5iG^JV8c5hG^JW8c5iG^JV8c5hG^JW8c5hG_JW8a5hG`JW8b5fGaJY8_5fGcJY8]5fGdJY8^5eGdJZ8]5cGeJ\\8]5bGeJ]8\\5aGfJ^8[5`GfJ_8\\5^GgJa8Z5\\GiJb8Y5[GjJd8W5XGlJg8V5VGmJi8T5TGoJk8R5RGPKm8R5oFRKP9P5kFTKS9Q5eFSK[9P5]FVKa9f7O0N3N2N1O2N1O2N2N1O2N2N1N3M2O2M3M2N3M3M3M3N3L3M4L4L4L3M4K6K4L5J5L5K4bMTCbMQ=U2WChMl<o1\\CoMh<h1_CVNd<a1fC\\N]<[1kCbNX<U1RDgNR<P1WDlNl;k0]DROg;e0`DYOR<OVDO^<XOaDNl?I7J`kQ7"}, "label": "a woman sitting next to a child eating a slice of pizza"}]}
{"id": 223871, "image": "COCO_train2014_000000223871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman was sitting on sofa\"."}], "task": "refering", "answer_template": "The \"a woman was sitting on sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 17, 18, 34, 35, 36, 51, 52, 53, 68, 69, 85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291], "bbox": [0.001, 0.001875, 0.24570833333333333, 0.785140625], "iscrowd": 0, "area": 42358.284449999985, "rle": {"size": [640, 480], "counts": "1c0]c0h9XFV1jNV1jNV1jNU1kNc0]O000000O2O00000O100000000O1000000O1000000O2O1O1O1O1dD\\KY7f4fH[KY7f4fH[KY7g4dH[K[7g4cHZK[7i4cHWK]7m4^HUKa7Q5YHoJg7W5SHjJk7]5nGdJR8b5hG^JW8d5hG]JW8c5hG^JW8c5iG^JV8c5hG^JW8c5iG^JV8c5hG^JW8c5hG_JW8a5hG`JW8b5fGaJY8_5fGcJY8]5fGdJY8^5eGdJZ8]5cGeJ\\8]5bGeJ]8\\5aGfJ^8[5`GfJ_8\\5^GgJa8Z5\\GiJb8Y5[GjJd8W5XGlJg8V5VGmJi8T5TGoJk8R5RGPKm8R5oFRKP9P5kFTKS9Q5eFSK[9P5]FVKa9f7O0N3N2N1O2N1O2N2N1O2N2N1N3M2O2M3M2N3M3M3M3N3L3M4L4L4L3M4K6K4L5J5L5K4bMTCbMQ=U2WChMl<o1\\CoMh<h1_CVNd<a1fC\\N]<[1kCbNX<U1RDgNR<P1WDlNl;k0]DROg;e0`DYOR<OVDO^<XOaDNl?I7J`kQ7"}, "label": "a woman was sitting on sofa"}]}
{"id": 313983, "image": "COCO_train2014_000000313983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"biege foal\"."}], "task": "refering", "answer_template": "The \"biege foal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 283, 284, 301, 302, 303, 306, 307, 322, 323, 324, 325, 326, 329, 330], "bbox": [0.000296875, 0.7452927400468384, 0.39528125, 1.0], "iscrowd": 0, "area": 6705.18145, "rle": {"size": [427, 640], "counts": "b<i0b<000000000001O0000000000000000000000000000000000000000001O00000000000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O10000000000O100000000O10000000000O10000006Je0[OcWo0BjhPO3M2N2N2N2N3N1N4L3M4L2N3M2N3M1O2N1M4L4K4M4L8G5L3M3L4M4L3L410O1000O0100O10N2O0O2O1O2M2O2SNXES1i:kN[ER1g:kN]EQ1e:jNbES1`:kNcER1_:nNcEn0^:SOcEj0_:VObEe0a:\\O`E>k;J6K5J6JloP5"}, "label": "biege foal"}]}
{"id": 313983, "image": "COCO_train2014_000000313983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mini pony\"."}], "task": "refering", "answer_template": "The \"the mini pony\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 283, 284, 301, 302, 303, 306, 307, 322, 323, 324, 325, 326, 329, 330], "bbox": [0.000296875, 0.7452927400468384, 0.39528125, 1.0], "iscrowd": 0, "area": 6705.18145, "rle": {"size": [427, 640], "counts": "b<i0b<000000000001O0000000000000000000000000000000000000000001O00000000000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O10000000000O100000000O10000000000O10000006Je0[OcWo0BjhPO3M2N2N2N2N3N1N4L3M4L2N3M2N3M1O2N1M4L4K4M4L8G5L3M3L4M4L3L410O1000O0100O10N2O0O2O1O2M2O2SNXES1i:kN[ER1g:kN]EQ1e:jNbES1`:kNcER1_:nNcEn0^:SOcEj0_:VObEe0a:\\O`E>k;J6K5J6JloP5"}, "label": "the mini pony"}]}
{"id": 313983, "image": "COCO_train2014_000000313983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown horse\"."}], "task": "refering", "answer_template": "The \"a dark brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 326, 327, 328], "bbox": [0.002859375, 0.35536299765807966, 0.599890625, 0.9819672131147541], "iscrowd": 0, "area": 65151.460399999996, "rle": {"size": [427, 640], "counts": "^Q1i5b70001O0O10001O000000001O000O101O0000001O00000O2O000000000000O1000O010000O10000O1000O01000000O10000O10O10O10000O10000O1000O010000O10000O100O010O10O0100O10O010000O010O100O010O10O0100O10O010000O010O100O010O10O0100O10O010000O010O10O0100O4M;D;F;D4M000O100O100O100O100O100O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10O1N2N1O2N2N1O2N2N2N1O2O1N1O2N2N2N1O2N2N1O2N2N20O01O1O010O1O10O01O1O010O1O1O010O1O00100M3M2N3L4M2N3M3L4M2N3M3L3N3M3M201N2O1O1N3N1O2N1N3N1O2N1N3N1O1O1N3N1O1O1N2O1O1O1O1O2O0O10000O100O10000O2O0O10000O100O10001N100O10000O100O101O0O100O10000O11O2N2O0O2N2N2N1O2O1N2N1O2N2O0O001O1O1O010O1O1O00100O1O001O100O001O1O10O01O1O1O100O1O1O1O1O101N1O1O1O100O1O1O1O2O0O1O1O1O10WKfG`4Y8^KkGa4T8^KoGa4P8^KSHb4k7\\KYHc4Y8O2N1O1N2M3N3L3N2M3N2M4M2M3N2M3M3XMUFl1n9PNWFk1l9SNXFi1j9UNYFg1j9WNZFe1h9YN\\Fb1h9[N\\Fa1f9\\N_F_1d9_N_F^1k:K5H8H8H8H7Ia^Z3"}, "label": "a dark brown horse"}]}
{"id": 313983, "image": "COCO_train2014_000000313983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hourse\"."}], "task": "refering", "answer_template": "The \"hourse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 326, 327, 328], "bbox": [0.002859375, 0.35536299765807966, 0.599890625, 0.9819672131147541], "iscrowd": 0, "area": 65151.460399999996, "rle": {"size": [427, 640], "counts": "^Q1i5b70001O0O10001O000000001O000O101O0000001O00000O2O000000000000O1000O010000O10000O1000O01000000O10000O10O10O10000O10000O1000O010000O10000O100O010O10O0100O10O010000O010O100O010O10O0100O10O010000O010O100O010O10O0100O10O010000O010O10O0100O4M;D;F;D4M000O100O100O100O100O100O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10O1N2N1O2N2N1O2N2N2N1O2O1N1O2N2N2N1O2N2N1O2N2N20O01O1O010O1O10O01O1O010O1O1O010O1O00100M3M2N3L4M2N3M3L4M2N3M3L3N3M3M201N2O1O1N3N1O2N1N3N1O2N1N3N1O1O1N3N1O1O1N2O1O1O1O1O2O0O10000O100O10000O2O0O10000O100O10001N100O10000O100O101O0O100O10000O11O2N2O0O2N2N2N1O2O1N2N1O2N2O0O001O1O1O010O1O1O00100O1O001O100O001O1O10O01O1O1O100O1O1O1O1O101N1O1O1O100O1O1O1O2O0O1O1O1O10WKfG`4Y8^KkGa4T8^KoGa4P8^KSHb4k7\\KYHc4Y8O2N1O1N2M3N3L3N2M3N2M4M2M3N2M3M3XMUFl1n9PNWFk1l9SNXFi1j9UNYFg1j9WNZFe1h9YN\\Fb1h9[N\\Fa1f9\\N_F_1d9_N_F^1k:K5H8H8H8H7Ia^Z3"}, "label": "hourse"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a serving table in front of a man with white t - shirt\"."}], "task": "refering", "answer_template": "The \"a serving table in front of a man with white t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 254, 255, 256, 278, 279, 300, 301], "bbox": [0.040453125, 0.48764583333333333, 0.4028125, 0.8224791666666665], "iscrowd": 0, "area": 9180.041300000003, "rle": {"size": [480, 640], "counts": "k]<1n>1O2N2N1O2N1O2O1N101O001O1O>A?BM3M3M3M3O0:G9G9G9G9G9F:lDSMc9V3TFQMe9T3VFRMd9n2[FWMa9i2^FWMc9i2ZFZMe9g2XF\\Mh9d2UF^Ml9=kEh06nNo9:UF?HYOT:9^FZ1b9gNhFn0W9SOTGa0m8@]G5c8LgGIY87RH]Oo7d0[HQOe7P1eHeNZ7\\1U31O1O001O1O0O2O1O001O1O0O5L:F;E:F:F;D2O0010O2N101N2N1VNcD0^;MkDKW;3QEEP;9YE@h:>`EZOb:c0gEUOZ:j0nEoNS:o0UFiNl9U1P200O10000O100O10000O10000O1000000O10000O101O0O1000000O10000O1000O01000000O10000O10000O100000O01000000O1000000O1000000O01O1N2O1N101N2O1O0O2O1N2O0011O000O2O00001O0O1HUB[Ol=b0YB[Og=d0;O2M2O1N3O000O10000O10001N1000000O10000O101O0O12N2O1N2O1N2N2O1N2O1N2OO001O00001O01O0000O101N100O100O2O001N101N101N101O0O2O0O2O0O2O00lXc5"}, "label": "a serving table in front of a man with white t - shirt"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty table with seating for two\"."}], "task": "refering", "answer_template": "The \"an empty table with seating for two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 381, 382, 383, 384, 385, 386, 387, 389, 390], "bbox": [0.40112500000000006, 0.6265833333333333, 0.9994375000000002, 0.9906458333333333], "iscrowd": 0, "area": 44931.708999999995, "rle": {"size": [480, 640], "counts": "_jh31n>1O2N2N1O2N1O2N1O2O1N101O001N101O001N2O00g0YO1N2O1O1O0O2OR2nM000O1000000O1000000O1000000O1000000O1000000OO2ZOf0YOg0N11000000O2O001O0O2O001O0O2O001O0O2O001O0O2O001O0O2O001O1N101O001N101O001N101O001N101N1O2N101N1O2N2N1YO[LSFf3l9lLaEU3_:g0O001N101O0010O01O010O0O2N1N3N1O2O001O1O0010O01O0100O10d0[O3NN101O001N101O0O2O001O0O2O1O001N101O001N101O000O10000O1000O10O1000000O10000O1000000O1000000O10000O100000O01000000001O001O00001O00001O001O00001O00001O001N10001O001O00001O00001O001O00001O00001O001O00001N10001O001O00001O0OB?1N10000O2O0O10001N100O10O1O001O001O001O1N101O02O000O100O10000O2O0O100O10000O101N10000O101N10000O2O0O100iLZES2g:kM[EU2f:gM]EY2c:eM_EZ2b:dMaE[2_:cMcE\\2^:aMeE_2\\:^MgEa2Y:]MiEb2X:\\MjEd2V:YMmEf2U:WMnEh2R:VMPFj2S;O101N100O101O0O101N10000O2O0O10001N1008H<E:E6J0010O00001O0000001O0000001O0000001O00001O0001O01M2N2K5I8G8I7H8H9H7O100O2N1M3L5L3M3L4M4K4M300N3G8GeD"}, "label": "an empty table with seating for two"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dining table with no one sitting at it and no pillows\"."}], "task": "refering", "answer_template": "The \"a dining table with no one sitting at it and no pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 381, 382, 383, 384, 385, 386, 387, 389, 390], "bbox": [0.40112500000000006, 0.6265833333333333, 0.9994375000000002, 0.9906458333333333], "iscrowd": 0, "area": 44931.708999999995, "rle": {"size": [480, 640], "counts": "_jh31n>1O2N2N1O2N1O2N1O2O1N101O001N101O001N2O00g0YO1N2O1O1O0O2OR2nM000O1000000O1000000O1000000O1000000O1000000OO2ZOf0YOg0N11000000O2O001O0O2O001O0O2O001O0O2O001O0O2O001O0O2O001O1N101O001N101O001N101O001N101N1O2N101N1O2N2N1YO[LSFf3l9lLaEU3_:g0O001N101O0010O01O010O0O2N1N3N1O2O001O1O0010O01O0100O10d0[O3NN101O001N101O0O2O001O0O2O1O001N101O001N101O000O10000O1000O10O1000000O10000O1000000O1000000O10000O100000O01000000001O001O00001O00001O001O00001O00001O001N10001O001O00001O00001O001O00001O00001O001O00001N10001O001O00001O0OB?1N10000O2O0O10001N100O10O1O001O001O001O1N101O02O000O100O10000O2O0O100O10000O101N10000O101N10000O2O0O100iLZES2g:kM[EU2f:gM]EY2c:eM_EZ2b:dMaE[2_:cMcE\\2^:aMeE_2\\:^MgEa2Y:]MiEb2X:\\MjEd2V:YMmEf2U:WMnEh2R:VMPFj2S;O101N100O101O0O101N10000O2O0O10001N1008H<E:E6J0010O00001O0000001O0000001O0000001O00001O0001O01M2N2K5I8G8I7H8H9H7O100O2N1M3L5L3M3L4M4K4M300N3G8GeD"}, "label": "a dining table with no one sitting at it and no pillows"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green couch with people sitting on it and pink pillows atop it as well\"."}], "task": "refering", "answer_template": "The \"a green couch with people sitting on it and pink pillows atop it as well\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 220, 221, 225, 226, 227, 228, 229, 243, 244, 248, 249, 250, 251, 252, 267, 271, 272, 273, 274, 275, 296, 297, 298, 321], "bbox": [0.45345312499999996, 0.40370833333333334, 0.9993749999999999, 0.7793541666666667], "iscrowd": 0, "area": 22919.35899999998, "rle": {"size": [480, 640], "counts": "RcX42n>2N2N2O1N2N2N2N3M=C^1cN2M1OO1N2O1N2O2M2O1N2O1N2O1N3N1N2O1N2O1O2M200O10000O101O0O10000O100O2O000O10000O101N10000O100O101N10000O100O2O0O10000O1000001N1000000000001O0OhCYOY:g0eE]OY:c0eEAY:?eEDZ:<dEF\\::bEH^:8`EK`:4^ENb:3ZE0f:0XE3g:NVE4j:LTE6l:KPE9o:GlD>U;AgDc0Y;^ObDg0];YO_Dk0a;VOUDS1k;Q101O012M9G2N1O001O1O100O1O000000O101]MjDY1W;aNUET1P;fN]Em0i:SO]Eb0h:^O]2O101O0O1000000O10001N1000000001O01O000000000000001O01O00001O00010O0000001O00010O00001O00000010O00000001O0001O0001O00000000010O0000O1000001O0000000O10001O000000000000001O0000hCNb92kEa0U:_OXET1h:b1001O0001O00000001O0001O0000000001M2O1O1O1O1O1O2N1O1O1N200000010O00000001O0001O0001O00000001O0001O0000000001O01aD`M\\:`2`EeM_:[2]EiMc:X2XElMh:T2UEPNj:P2TERNm:n1PETNP;l1nDVNR;j1lDYNS;h1jDZNV;h201O00010O0000001O00010O00001O0000010O00001O00001O01O0001O00001O01O01O0000001O01O01O00001O0000010O00001O000010O000001O00001O01O01O0000001;DW1iNW1jNV1iN^D"}, "label": "a green couch with people sitting on it and pink pillows atop it as well"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long green seat with two pink pillows on it\"."}], "task": "refering", "answer_template": "The \"a long green seat with two pink pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 220, 221, 225, 226, 227, 228, 229, 243, 244, 248, 249, 250, 251, 252, 267, 271, 272, 273, 274, 275, 296, 297, 298, 321], "bbox": [0.45345312499999996, 0.40370833333333334, 0.9993749999999999, 0.7793541666666667], "iscrowd": 0, "area": 22919.35899999998, "rle": {"size": [480, 640], "counts": "RcX42n>2N2N2O1N2N2N2N3M=C^1cN2M1OO1N2O1N2O2M2O1N2O1N2O1N3N1N2O1N2O1O2M200O10000O101O0O10000O100O2O000O10000O101N10000O100O101N10000O100O2O0O10000O1000001N1000000000001O0OhCYOY:g0eE]OY:c0eEAY:?eEDZ:<dEF\\::bEH^:8`EK`:4^ENb:3ZE0f:0XE3g:NVE4j:LTE6l:KPE9o:GlD>U;AgDc0Y;^ObDg0];YO_Dk0a;VOUDS1k;Q101O012M9G2N1O001O1O100O1O000000O101]MjDY1W;aNUET1P;fN]Em0i:SO]Eb0h:^O]2O101O0O1000000O10001N1000000001O01O000000000000001O01O00001O00010O0000001O00010O00001O00000010O00000001O0001O0001O00000000010O0000O1000001O0000000O10001O000000000000001O0000hCNb92kEa0U:_OXET1h:b1001O0001O00000001O0001O0000000001M2O1O1O1O1O1O2N1O1O1N200000010O00000001O0001O0001O00000001O0001O0000000001O01aD`M\\:`2`EeM_:[2]EiMc:X2XElMh:T2UEPNj:P2TERNm:n1PETNP;l1nDVNR;j1lDYNS;h1jDZNV;h201O00010O0000001O00010O00001O0000010O00001O00001O01O0001O00001O01O01O0000001O01O01O00001O0000010O00001O000010O000001O00001O01O01O0000001;DW1iNW1jNV1iN^D"}, "label": "a long green seat with two pink pillows on it"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unoccupied table next to an occupied table\"."}], "task": "refering", "answer_template": "The \"an unoccupied table next to an occupied table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 303, 304, 306, 329, 352, 375], "bbox": [0.17825, 0.5506666666666666, 0.5885625, 0.9878958333333333], "iscrowd": 0, "area": 17796.457650000004, "rle": {"size": [480, 640], "counts": "cge13i>5K5K4L5O1O1N2O1O1O001N2O1O1O5J?B>B>B=C0O2O001O1N101O001N101O1O001N10TNkDNT;3_EZO`:g0RFfNm9[1l11O001N101O001O0O2O001O1N101O001O0O2O001O0O2O1O001N101O001O0O2O001O1N101O001O0O2O02O2M3N1N09Ga0_Oa0_Oa0_Oa0_Oa0_O`0@001O001O001O001O001O001O1O001O[LWJSNi5l1f3000000000000O10000000000O10000000O1O1JiMcCV2]<lMaCU2_<kM`CV2_<6O100O00100O1O100O00100O100O1O12N21O2NN2M3N2N1O2M3N2N2M3ESC_No<_1RC`Nf=h0[BVOf=i0]BUOc=k0^BTOa=m0`BRO`=n0900O010000000O0100YOUOYCk0e<ZOXCe0h<_OUCa0i<CTC>k<FRC:m<JPC6n<OoB0Q=P1N2O1O001O1O100O010001O001N10001O0O2O00001O0O101O001N10001O001N10001O000O2O001O0O1013L3NN100O2N100O2N101N1O1O2O0O2N100O2N100O2N1O2M2M3N3L3M4M2M3O2O00001O001O00001Oeoj3"}, "label": "an unoccupied table next to an occupied table"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table far away , with two people dining\"."}], "task": "refering", "answer_template": "The \"a table far away , with two people dining\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 230, 231, 232, 233, 253, 254], "bbox": [0.000578125, 0.36487499999999995, 0.193765625, 0.6765833333333333], "iscrowd": 0, "area": 5137.0021000000015, "rle": {"size": [480, 640], "counts": "\\71f>:O0010O01O<EO00001OO10001O0O1WOZOXCg0e<[O[CR1W<POhC]1k;dNTDi1_;WN`DV2T;kMkD_2k:bMTEc2g:^MXEc2g:_MVEc2h:`MVE`2j:cMSE]2m:gMoDZ2P;k00[MSEW1m:PO`E=_:BUFKk95hFXOX9h0\\GeNb8[1^240000O100000O1000000000M30O0100C=0RNVCf1V=L1O001N101O001O001O0O1000O100JcNiB]1W=60O10O1000000N2L310OZNSC[1`1bNc93[En0Q1ZOZ9GfEo0Q1Dm8\\OUFo0n0Fl8ZOWFP1n0Ej8[OYFo0n0Ff9:[FEf9:gFYOY9g0`20001O000O101O0O1GoACQ>=:N10000O2O00000O2O0000001O0000001I6O10001O0O1000[cb7"}, "label": "a table far away , with two people dining"}]}
{"id": 10881, "image": "COCO_train2014_000000010881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table next to a window\"."}], "task": "refering", "answer_template": "The \"a table next to a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 230, 231, 232, 233, 253, 254], "bbox": [0.000578125, 0.36487499999999995, 0.193765625, 0.6765833333333333], "iscrowd": 0, "area": 5137.0021000000015, "rle": {"size": [480, 640], "counts": "\\71f>:O0010O01O<EO00001OO10001O0O1WOZOXCg0e<[O[CR1W<POhC]1k;dNTDi1_;WN`DV2T;kMkD_2k:bMTEc2g:^MXEc2g:_MVEc2h:`MVE`2j:cMSE]2m:gMoDZ2P;k00[MSEW1m:PO`E=_:BUFKk95hFXOX9h0\\GeNb8[1^240000O100000O1000000000M30O0100C=0RNVCf1V=L1O001N101O001O001O0O1000O100JcNiB]1W=60O10O1000000N2L310OZNSC[1`1bNc93[En0Q1ZOZ9GfEo0Q1Dm8\\OUFo0n0Fl8ZOWFP1n0Ej8[OYFo0n0Ff9:[FEf9:gFYOY9g0`20001O000O101O0O1GoACQ>=:N10000O2O00000O2O0000001O0000001I6O10001O0O1000[cb7"}, "label": "a table next to a window"}]}
{"id": 256643, "image": "COCO_train2014_000000256643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player at bat swings\"."}], "task": "refering", "answer_template": "The \"the baseball player at bat swings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 195, 196, 197, 219, 220, 221, 242, 243, 244, 263, 264, 265, 267, 268, 286, 287, 290, 291], "bbox": [0.4675625, 0.48281690140845074, 0.689828125, 0.8293192488262912], "iscrowd": 0, "area": 6506.818100000001, "rle": {"size": [426, 640], "counts": "Rgl32S=7L4N1N3N2M3N0O100O100O100O100O1OG:O001O0010O010O010]MGTH9m7HRH8m7JQH7n7JRH5o7KPH6o7LPH3P8NoG3Q8NmG2S8OmG1R81lGOU81kGOT82kGOT82lGMU84iGMV84iGLk7@bFe0c1Kc7KeF<g1Ie7i0[HWOd7j0[HVOf7k0XHVOg7k0YHUOg7k0XHUOi7k0UHWOk7l0PHUOQ8m0kGUOU8n0eGTO\\8n0QGBP9`0mFAS9b0hF_OY9c0cF^O^9d0_F\\Ob9U21O1O1O1O1O2N2N2N2N4L4L4L1O1O1O100O1000000000100O0VN`Ej0a:SOfEh0[:UOkEg0V:UOQFg0P:VOWFf0j9VOZFh0U;00O10O10O10O10O100O01000O10O10O100O0101000000O10=CO10O0100O1N1O2M3N1O2N2N2M2O2N2N2N2M3N001O1N1O2N2N2N2N2Ne^b2"}, "label": "the baseball player at bat swings"}]}
{"id": 256643, "image": "COCO_train2014_000000256643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left handed batter after a full swing\"."}], "task": "refering", "answer_template": "The \"left handed batter after a full swing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 195, 196, 197, 219, 220, 221, 242, 243, 244, 263, 264, 265, 267, 268, 286, 287, 290, 291], "bbox": [0.4675625, 0.48281690140845074, 0.689828125, 0.8293192488262912], "iscrowd": 0, "area": 6506.818100000001, "rle": {"size": [426, 640], "counts": "Rgl32S=7L4N1N3N2M3N0O100O100O100O100O1OG:O001O0010O010O010]MGTH9m7HRH8m7JQH7n7JRH5o7KPH6o7LPH3P8NoG3Q8NmG2S8OmG1R81lGOU81kGOT82kGOT82lGMU84iGMV84iGLk7@bFe0c1Kc7KeF<g1Ie7i0[HWOd7j0[HVOf7k0XHVOg7k0YHUOg7k0XHUOi7k0UHWOk7l0PHUOQ8m0kGUOU8n0eGTO\\8n0QGBP9`0mFAS9b0hF_OY9c0cF^O^9d0_F\\Ob9U21O1O1O1O1O2N2N2N2N4L4L4L1O1O1O100O1000000000100O0VN`Ej0a:SOfEh0[:UOkEg0V:UOQFg0P:VOWFf0j9VOZFh0U;00O10O10O10O10O100O01000O10O10O100O0101000000O10=CO10O0100O1N1O2M3N1O2N2N2M2O2N2N2N2M3N001O1N1O2N2N2N2N2Ne^b2"}, "label": "left handed batter after a full swing"}]}
{"id": 256643, "image": "COCO_train2014_000000256643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire officiating a baseball game\"."}], "task": "refering", "answer_template": "The \"an umpire officiating a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 210, 211, 232, 233, 234, 235, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281], "bbox": [0.031484374999999995, 0.5336384976525822, 0.253359375, 0.8557042253521127], "iscrowd": 0, "area": 6896.949300000001, "rle": {"size": [426, 640], "counts": "Qe81Y=3M3L4M3M2N0O101O0O1000001N10O10O10000O10000O100O1000O010000O10000O10000O01000O01000O01000O10O10O10O10O10O10O1000O01I6I7H8H8J601O0000000N2M300000001O00000O10000O100O1000O01O1O1O1N20O1000O3N1O1O1O1O100O2O0O100O1O1O1O4K5N2M3N2M3M3N2M4LaM_FY1^9cNhFCIR1_9UOnFCG[1_9fNRGHDb1`9[NRGMCf1R;01N2O0O2O2M2N2M3M4M2M?A4L2N2O2M2N2N3M2N]`V6"}, "label": "an umpire officiating a baseball game"}]}
{"id": 256643, "image": "COCO_train2014_000000256643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire kneeling and wearing dark colors\"."}], "task": "refering", "answer_template": "The \"the umpire kneeling and wearing dark colors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 210, 211, 232, 233, 234, 235, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281], "bbox": [0.031484374999999995, 0.5336384976525822, 0.253359375, 0.8557042253521127], "iscrowd": 0, "area": 6896.949300000001, "rle": {"size": [426, 640], "counts": "Qe81Y=3M3L4M3M2N0O101O0O1000001N10O10O10000O10000O100O1000O010000O10000O10000O01000O01000O01000O10O10O10O10O10O10O1000O01I6I7H8H8J601O0000000N2M300000001O00000O10000O100O1000O01O1O1O1N20O1000O3N1O1O1O1O100O2O0O100O1O1O1O4K5N2M3N2M3M3N2M4LaM_FY1^9cNhFCIR1_9UOnFCG[1_9fNRGHDb1`9[NRGMCf1R;01N2O0O2O2M2N2M3M4M2M?A4L2N2O2M2N2N3M2N]`V6"}, "label": "the umpire kneeling and wearing dark colors"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slab of cornbread in a glass baking dish\"."}], "task": "refering", "answer_template": "The \"a slab of cornbread in a glass baking dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 207, 208], "bbox": [0.0036875, 0.30212500000000003, 0.258625, 0.5484375], "iscrowd": 0, "area": 10475.223849999998, "rle": {"size": [480, 640], "counts": "aR1:f>R3nL:GO00000O2O0000001O000O10001O00000O2O0000001O000O10001O00000O2O0O100O2N100O100O2O0O1O101N100O101N1O100O2O0O100O1O2O0O100O2N100O101N100O1O101N100O101N1O100O2O0O100O1O2O0O100O2N100O101N100O1O101N100O101N1O100O2O0O100O1O2O0O100O2N100O101N100O1O101N100O101N1O100O2O0O100O2N100O100O2N100O101N100O1O2O0O100O101N1O100O2O0O100O[nn6"}, "label": "a slab of cornbread in a glass baking dish"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a casserole in a rectangular glass pan\"."}], "task": "refering", "answer_template": "The \"a casserole in a rectangular glass pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 207, 208], "bbox": [0.0036875, 0.30212500000000003, 0.258625, 0.5484375], "iscrowd": 0, "area": 10475.223849999998, "rle": {"size": [480, 640], "counts": "aR1:f>R3nL:GO00000O2O0000001O000O10001O00000O2O0000001O000O10001O00000O2O0O100O2N100O100O2O0O1O101N100O101N1O100O2O0O100O1O2O0O100O2N100O101N100O1O101N100O101N1O100O2O0O100O1O2O0O100O2N100O101N100O1O101N100O101N1O100O2O0O100O1O2O0O100O2N100O101N100O1O101N100O101N1O100O2O0O100O2N100O100O2N100O101N100O1O2O0O100O101N1O100O2O0O100O[nn6"}, "label": "a casserole in a rectangular glass pan"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half filled skillet with salads and a shortbreat\"."}], "task": "refering", "answer_template": "The \"a half filled skillet with salads and a shortbreat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250], "bbox": [0.5011875, 0.22260416666666666, 1.0, 0.6253124999999999], "iscrowd": 0, "area": 42173.13875, "rle": {"size": [480, 640], "counts": "\\df45f>8H9G8I7H9G8K5J6J7K10O010O000010O01O0010O100O100O100O2O0O1O100O100O100O100O100O100O1000000O10001O00000O10000000000O1000000000000O100000001O0O10000000000O1000000000000O10000000000O101O0000001O1N101O1O001O0O2O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O5K4L5K5K4L5K5K3M1O2N2N1O2N00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O0010O01O001O0000001O0001O00000001O0000000001O01O0000000000001O01O000000000001O000001O000001O00000000kNU101O0O1000000O1000001O0O1000000O1000000O2O000001O00000001O00001O0001O01O000010O1O010O0010O01O10O01O010O00100O1O1O2O0O1O100O1O101NO101O0O101O000O2O00001N1000001XNcM"}, "label": "a half filled skillet with salads and a shortbreat"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vegetable medley in the frying pan\"."}], "task": "refering", "answer_template": "The \"vegetable medley in the frying pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250], "bbox": [0.5011875, 0.22260416666666666, 1.0, 0.6253124999999999], "iscrowd": 0, "area": 42173.13875, "rle": {"size": [480, 640], "counts": "\\df45f>8H9G8I7H9G8K5J6J7K10O010O000010O01O0010O100O100O100O2O0O1O100O100O100O100O100O100O1000000O10001O00000O10000000000O1000000000000O100000001O0O10000000000O1000000000000O10000000000O101O0000001O1N101O1O001O0O2O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O5K4L5K5K4L5K5K3M1O2N2N1O2N00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O0010O01O001O0000001O0001O00000001O0000000001O01O0000000000001O01O000000000001O000001O000001O00000000kNU101O0O1000000O1000001O0O1000000O1000000O2O000001O00000001O00001O0001O01O000010O1O010O0010O01O10O01O010O00100O1O1O2O0O1O100O1O101NO101O0O101O000O2O00001N1000001XNcM"}, "label": "vegetable medley in the frying pan"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the meat on the white plate\"."}], "task": "refering", "answer_template": "The \"the meat on the white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105], "bbox": [0.33404687499999997, 0.029083333333333336, 0.837828125, 0.25725], "iscrowd": 0, "area": 24903.9607, "rle": {"size": [480, 640], "counts": "U\\T32l>3N3L3N3L3M4M2M3N1N3MSORBm0o=101N101N101N2O0O2O0O2O0O2O1N101N101N101N101G8I7L4O2N1O10001O0000001O000000001O0000010O0001O010O000010O0001O0010O0000001O0O10001O0000001N10001O001O0O2O001O001O0O2O001O001O000O2O001O001O0O2O001O001O0O10000000001N1000000000000O101O0000000O10000000000O10000000O1000O010000O1000O010000O10000O10000O10000O1000000O10000000000O1000000000000O1000001O000O1000000000000O10000000000O1000000000001N10000000000O2O00000000001N10000000000O2O00000000000O2O00000000001N10000000000O2O00000000001N10000000000O2O0000000O101N100O100O10001N100O100O101N100O100O100O2O0O100O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O001N101N101N101N101N101N101N101N101N9H:ERf`1"}, "label": "the meat on the white plate"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate that only has roast beef on it\"."}], "task": "refering", "answer_template": "The \"the plate that only has roast beef on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105], "bbox": [0.33404687499999997, 0.029083333333333336, 0.837828125, 0.25725], "iscrowd": 0, "area": 24903.9607, "rle": {"size": [480, 640], "counts": "U\\T32l>3N3L3N3L3M4M2M3N1N3MSORBm0o=101N101N101N2O0O2O0O2O0O2O1N101N101N101N101G8I7L4O2N1O10001O0000001O000000001O0000010O0001O010O000010O0001O0010O0000001O0O10001O0000001N10001O001O0O2O001O001O0O2O001O001O000O2O001O001O0O2O001O001O0O10000000001N1000000000000O101O0000000O10000000000O10000000O1000O010000O1000O010000O10000O10000O10000O1000000O10000000000O1000000000000O1000001O000O1000000000000O10000000000O1000000000001N10000000000O2O00000000001N10000000000O2O00000000000O2O00000000001N10000000000O2O00000000001N10000000000O2O0000000O101N100O100O10001N100O100O101N100O100O100O2O0O100O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O001N101N101N101N101N101N101N101N101N9H:ERf`1"}, "label": "the plate that only has roast beef on it"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass casserole dish full of food\"."}], "task": "refering", "answer_template": "The \"a glass casserole dish full of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 232, 253, 254], "bbox": [0.0, 0.2471875, 0.304234375, 0.6979166666666666], "iscrowd": 0, "area": 27069.71335, "rle": {"size": [480, 640], "counts": "g3g6Z8O0001O0001N10000O2O000O100O2O000O10001N10000O101O0O10001N10000O101O0O10000O2O0O10000O2O000O1O1N3N1O1N2O2N1N2O1N2O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1N2O100O2O0O1O100O2O0O100O1O2O0O100O101N1O100O101N1O100O101N10000O100O2O000O100O2O0O10000O2O0O100O10001N100O100O2O000O100O2O0O10000O101N100O10001N100O100O2O0O1O1O1O1O2N1O1O1O2N100O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2O0O1O1O2N2N1O2N1O2N2N1O2N1N3N2M;Fc0\\OS]`6"}, "label": "a glass casserole dish full of food"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rectangle glass baking dish with food in it\"."}], "task": "refering", "answer_template": "The \"a rectangle glass baking dish with food in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 232, 253, 254], "bbox": [0.0, 0.2471875, 0.304234375, 0.6979166666666666], "iscrowd": 0, "area": 27069.71335, "rle": {"size": [480, 640], "counts": "g3g6Z8O0001O0001N10000O2O000O100O2O000O10001N10000O101O0O10001N10000O101O0O10000O2O0O10000O2O000O1O1N3N1O1N2O2N1N2O1N2O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1N2O100O2O0O1O100O2O0O100O1O2O0O100O101N1O100O101N1O100O101N10000O100O2O000O100O2O0O10000O2O0O100O10001N100O100O2O000O100O2O0O10000O101N100O10001N100O100O2O0O1O1O1O1O2N1O1O1O2N100O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2O0O1O1O2N2N1O2N1O2N2N1O2N1N3N2M;Fc0\\OS]`6"}, "label": "a rectangle glass baking dish with food in it"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 slices of meatloaf on blue plate in the right hand picture\"."}], "task": "refering", "answer_template": "The \"2 slices of meatloaf on blue plate in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 332, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360], "bbox": [0.46734375000000006, 0.6512916666666667, 0.8029375, 0.9007708333333333], "iscrowd": 0, "area": 19651.830299999994, "rle": {"size": [480, 640], "counts": "\\a\\47`>e0@4M4K5K4M4M3M2O2M3M3N1N3M3N1N3M3M2M4M2N3M2M3N2N2N2O10001O000O10000000001O000000000O100000000000000000000O1000000000000000000O100000000000000000000O101O000000000O1000000000000O1000000000000O101O000000000O100000000000000O1000000000000O10O1000000O10000O10000O1000000O10001N10000O1000000O10000O10001N10000O1O1O1O1O1O1N2O1O1O2N1O1O1O0O1010O00100O10O10O1000O10O1000O10O1000O102N2N1N3N2N2N2M3N2N2N2M3N1O2N2M3N2M4L3M4M2M4L3N3L4L3N3I6Fcfj1"}, "label": "2 slices of meatloaf on blue plate in the right hand picture"}]}
{"id": 502407, "image": "COCO_train2014_000000502407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark meat to the right of the broccoli\"."}], "task": "refering", "answer_template": "The \"the dark meat to the right of the broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 332, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360], "bbox": [0.46734375000000006, 0.6512916666666667, 0.8029375, 0.9007708333333333], "iscrowd": 0, "area": 19651.830299999994, "rle": {"size": [480, 640], "counts": "\\a\\47`>e0@4M4K5K4M4M3M2O2M3M3N1N3M3N1N3M3M2M4M2N3M2M3N2N2N2O10001O000O10000000001O000000000O100000000000000000000O1000000000000000000O100000000000000000000O101O000000000O1000000000000O1000000000000O101O000000000O100000000000000O1000000000000O10O1000000O10000O10000O1000000O10001N10000O1000000O10000O10001N10000O1O1O1O1O1O1N2O1O1O2N1O1O1O0O1010O00100O10O10O1000O10O1000O10O1000O102N2N1N3N2N2N2M3N2N2N2M3N1O2N2M3N2M4L3M4M2M4L3N3L4L3N3I6Fcfj1"}, "label": "the dark meat to the right of the broccoli"}]}
{"id": 84615, "image": "COCO_train2014_000000084615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse in field touched by hand\"."}], "task": "refering", "answer_template": "The \"a horse in field touched by hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 361, 370, 371, 372, 373, 374, 376, 377, 378, 379, 380, 381, 382, 383, 384, 394, 395, 396, 399, 400, 401, 402, 403, 404, 405, 406, 407, 422, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.07590625, 0.0012801484230055657, 0.7662812499999999, 0.9690909090909091], "iscrowd": 0, "area": 128239.12440000003, "rle": {"size": [539, 640], "counts": "QWj03g`03M4L3M4M2M3M4L3M4L3M3M4L3M4M2M3M4L2N1O1O1O1O1O10001O00000TKnMoJR2o4oMRKQ2l4PNUKP2i4QNXKo1g4RNYKn1e4SN\\Km1b4TN_Kl1_4VNaKj1]4WNdKi1Z4XNgKh1W4ZNiKf1U4[NlKe1R4\\NoKd1o3]NRLd1k3^NULb1c0oM[M`0S2a1a0SNWM=Y2`1?WNSM;^2^1>\\NnL7e2]1<`NjL4k2\\1:dNfL2P3Z19hNbLOV3Y17mN]LK]3X15QOYLHc3W13UOULFh3U12YOQLCn3T10^OlK_OU4S1NBfK_O\\4o0MFiJ5[55Ki15XNJi16WNIi18WNGi19XNFh1;XNDi1<WNYLoM[1j3\\2XNYLSNU1e3c2XNXLYNn0`3i2XNYL]Nh0[3P3XNXLcNa0U3X3XNWLhN;Q3]3XNXLlN5l2d3XNWLRONi2i3UNXLXOHf2QOiK[4Z2lLYOEf2POjK_4W2lLZOCh2nNhKd4V2kL[OAi2mNhKf4U2lL\\O^Oi2mNgKj4T2kL]O\\OS3j3`MjL^OZOS3k3`MkL^OXOR3n3`MjL@UOP3R4`MiLASOP3S4`MjLAQOo2V4`MiLBoNo2X4_MiLDlNm2[4`MiLDhNn2`4^MhLEeNo2c4\\MhLGaNn2g4\\MhLG]No2l4ZMgLHZNP3o4XMfLR5Z3oJfLQ5Z3oJfLP5Z3QKfLn4Z3SKfLm4Z3SKfLl4Z3UKfLj4[3VKeLj4Z3WKfLh4[3XKeLh4Z3YKfLf4[3ZKeLe4[3\\KeLd4[3\\KeLc4[3^KeLa4[3`KeL`4[3`KeL_4[3bKgL\\4Y3cKiL[4W3fKkLW4V3iKlLU4o1UIRMg2Q1Q4j1\\IUMc2R1o3f1bIWM`2U1l3_1hI]M\\2V1i3Z1oI_MY2X1g3T1UJcMU2[1c3o0[JgMR2\\1`3l0`JgMQ2_1]3i0cJhMP2`1[3h0fJgMP2c1W3f0iJhMP2d1U3c0lJhMP2f1T3`0mJjMo1h1S3<nJlMP2j1Q38PKnMm1n1Q33SKnMk1Q2S3ORKPNj1S2U3KRKRNf1V2Y3FRKSNd1Y2Z3CRKUNb1Z2[3ATKTN`1]2\\3^OUKUN\\1`2_3ZOVKVNY1b2a3WOVKXNW1c2b3UOXKWNU1g2b3QOZKXNQ1j2e3mNZKZNo0k2g3jN[K[Nl0m2i3gN\\K[Nj0P3i3eN]K\\Ng0R3l3aN^K]Nd0T3n3^N_K^N`0W3Q4ZN_K`N>X3S4VNaKaN:\\3T4RNdKbN6^3V4nMfKdN1a3Y4iMgKgNNb3[4eMiKhNJf3]4`MkKjNFh3^4]MmKlNBj3a4XMoKnN^Ol3c4TMQLoNZOP4g4mLPLTOWOQ4j4hLQLWOSOS4n4bLQL[OnNV4R5\\LQL_OkNW4V5VLQLBgN[4Y5PLRLEcN]4]5jKRLI^N`4a5eKQLL\\Na4e5_KQLOXNe4h5YKRL2TNg4k5TKRL6PNh4o5PKSL8lMi4S6lJSL:hMm4V6fJSL>eMm4Z6bJSLa0`Mo4_6]JTLc0[MQ5c6YJULd0VMV5f6SJVLg0RMW5j6oIWLk6n6eLWFAj96cFF^9OQGKQ9K\\G1d8FjG6V8DSH8n7C[H8e7DcH9]7BlH9U7BSI:m6_NkEh0a3e0d6^NSFe0a3i0m6PO\\Ik0h6mN\\IS1g6fN]I[1Q;0000SOg@9X?Gj@7V?Il@5T?Ko@2Q?NQA1n>OUANk>2WALi>5XAJg>6\\AGd>9^AEb>;aAB_>>cAA\\>?fA_OZ>a0P100001O00000000000000001O00000000000000001O00000000000000001O0000000000000m@Gd=o1I:Fn0RO5K1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O5K;E;E;E;E;E;E:F;E:FO1N2O1N2O1N2O1N2O1O1O1O100O1O100O100O100O1O100O100O100O1O100O10000000000000000000000000000001O000000000000000000000000000O2O00000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001K4M4L3M4K5L3M4L3M4K4010O100O100O100OL5H7H9I6L5K5K6J5K6J5K6J5K6J5L5J5L5J5L5K4K6K=Ba0@a0_O?@:G8H8G9H8G:G8H`c^2"}, "label": "a horse in field touched by hand"}]}
{"id": 84615, "image": "COCO_train2014_000000084615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that is getting pet\"."}], "task": "refering", "answer_template": "The \"the horse that is getting pet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 361, 370, 371, 372, 373, 374, 376, 377, 378, 379, 380, 381, 382, 383, 384, 394, 395, 396, 399, 400, 401, 402, 403, 404, 405, 406, 407, 422, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.07590625, 0.0012801484230055657, 0.7662812499999999, 0.9690909090909091], "iscrowd": 0, "area": 128239.12440000003, "rle": {"size": [539, 640], "counts": "QWj03g`03M4L3M4M2M3M4L3M4L3M3M4L3M4M2M3M4L2N1O1O1O1O1O10001O00000TKnMoJR2o4oMRKQ2l4PNUKP2i4QNXKo1g4RNYKn1e4SN\\Km1b4TN_Kl1_4VNaKj1]4WNdKi1Z4XNgKh1W4ZNiKf1U4[NlKe1R4\\NoKd1o3]NRLd1k3^NULb1c0oM[M`0S2a1a0SNWM=Y2`1?WNSM;^2^1>\\NnL7e2]1<`NjL4k2\\1:dNfL2P3Z19hNbLOV3Y17mN]LK]3X15QOYLHc3W13UOULFh3U12YOQLCn3T10^OlK_OU4S1NBfK_O\\4o0MFiJ5[55Ki15XNJi16WNIi18WNGi19XNFh1;XNDi1<WNYLoM[1j3\\2XNYLSNU1e3c2XNXLYNn0`3i2XNYL]Nh0[3P3XNXLcNa0U3X3XNWLhN;Q3]3XNXLlN5l2d3XNWLRONi2i3UNXLXOHf2QOiK[4Z2lLYOEf2POjK_4W2lLZOCh2nNhKd4V2kL[OAi2mNhKf4U2lL\\O^Oi2mNgKj4T2kL]O\\OS3j3`MjL^OZOS3k3`MkL^OXOR3n3`MjL@UOP3R4`MiLASOP3S4`MjLAQOo2V4`MiLBoNo2X4_MiLDlNm2[4`MiLDhNn2`4^MhLEeNo2c4\\MhLGaNn2g4\\MhLG]No2l4ZMgLHZNP3o4XMfLR5Z3oJfLQ5Z3oJfLP5Z3QKfLn4Z3SKfLm4Z3SKfLl4Z3UKfLj4[3VKeLj4Z3WKfLh4[3XKeLh4Z3YKfLf4[3ZKeLe4[3\\KeLd4[3\\KeLc4[3^KeLa4[3`KeL`4[3`KeL_4[3bKgL\\4Y3cKiL[4W3fKkLW4V3iKlLU4o1UIRMg2Q1Q4j1\\IUMc2R1o3f1bIWM`2U1l3_1hI]M\\2V1i3Z1oI_MY2X1g3T1UJcMU2[1c3o0[JgMR2\\1`3l0`JgMQ2_1]3i0cJhMP2`1[3h0fJgMP2c1W3f0iJhMP2d1U3c0lJhMP2f1T3`0mJjMo1h1S3<nJlMP2j1Q38PKnMm1n1Q33SKnMk1Q2S3ORKPNj1S2U3KRKRNf1V2Y3FRKSNd1Y2Z3CRKUNb1Z2[3ATKTN`1]2\\3^OUKUN\\1`2_3ZOVKVNY1b2a3WOVKXNW1c2b3UOXKWNU1g2b3QOZKXNQ1j2e3mNZKZNo0k2g3jN[K[Nl0m2i3gN\\K[Nj0P3i3eN]K\\Ng0R3l3aN^K]Nd0T3n3^N_K^N`0W3Q4ZN_K`N>X3S4VNaKaN:\\3T4RNdKbN6^3V4nMfKdN1a3Y4iMgKgNNb3[4eMiKhNJf3]4`MkKjNFh3^4]MmKlNBj3a4XMoKnN^Ol3c4TMQLoNZOP4g4mLPLTOWOQ4j4hLQLWOSOS4n4bLQL[OnNV4R5\\LQL_OkNW4V5VLQLBgN[4Y5PLRLEcN]4]5jKRLI^N`4a5eKQLL\\Na4e5_KQLOXNe4h5YKRL2TNg4k5TKRL6PNh4o5PKSL8lMi4S6lJSL:hMm4V6fJSL>eMm4Z6bJSLa0`Mo4_6]JTLc0[MQ5c6YJULd0VMV5f6SJVLg0RMW5j6oIWLk6n6eLWFAj96cFF^9OQGKQ9K\\G1d8FjG6V8DSH8n7C[H8e7DcH9]7BlH9U7BSI:m6_NkEh0a3e0d6^NSFe0a3i0m6PO\\Ik0h6mN\\IS1g6fN]I[1Q;0000SOg@9X?Gj@7V?Il@5T?Ko@2Q?NQA1n>OUANk>2WALi>5XAJg>6\\AGd>9^AEb>;aAB_>>cAA\\>?fA_OZ>a0P100001O00000000000000001O00000000000000001O00000000000000001O0000000000000m@Gd=o1I:Fn0RO5K1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O5K;E;E;E;E;E;E:F;E:FO1N2O1N2O1N2O1N2O1O1O1O100O1O100O100O100O1O100O100O100O1O100O10000000000000000000000000000001O000000000000000000000000000O2O00000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001K4M4L3M4K5L3M4L3M4K4010O100O100O100OL5H7H9I6L5K5K6J5K6J5K6J5K6J5L5J5L5J5L5K4K6K=Ba0@a0_O?@:G8H8G9H8G:G8H`c^2"}, "label": "the horse that is getting pet"}]}
{"id": 84615, "image": "COCO_train2014_000000084615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"largest , closest horse\"."}], "task": "refering", "answer_template": "The \"largest , closest horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 361, 370, 371, 372, 373, 374, 376, 377, 378, 379, 380, 381, 382, 383, 384, 394, 395, 396, 399, 400, 401, 402, 403, 404, 405, 406, 407, 422, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.07590625, 0.0012801484230055657, 0.7662812499999999, 0.9690909090909091], "iscrowd": 0, "area": 128239.12440000003, "rle": {"size": [539, 640], "counts": "QWj03g`03M4L3M4M2M3M4L3M4L3M3M4L3M4M2M3M4L2N1O1O1O1O1O10001O00000TKnMoJR2o4oMRKQ2l4PNUKP2i4QNXKo1g4RNYKn1e4SN\\Km1b4TN_Kl1_4VNaKj1]4WNdKi1Z4XNgKh1W4ZNiKf1U4[NlKe1R4\\NoKd1o3]NRLd1k3^NULb1c0oM[M`0S2a1a0SNWM=Y2`1?WNSM;^2^1>\\NnL7e2]1<`NjL4k2\\1:dNfL2P3Z19hNbLOV3Y17mN]LK]3X15QOYLHc3W13UOULFh3U12YOQLCn3T10^OlK_OU4S1NBfK_O\\4o0MFiJ5[55Ki15XNJi16WNIi18WNGi19XNFh1;XNDi1<WNYLoM[1j3\\2XNYLSNU1e3c2XNXLYNn0`3i2XNYL]Nh0[3P3XNXLcNa0U3X3XNWLhN;Q3]3XNXLlN5l2d3XNWLRONi2i3UNXLXOHf2QOiK[4Z2lLYOEf2POjK_4W2lLZOCh2nNhKd4V2kL[OAi2mNhKf4U2lL\\O^Oi2mNgKj4T2kL]O\\OS3j3`MjL^OZOS3k3`MkL^OXOR3n3`MjL@UOP3R4`MiLASOP3S4`MjLAQOo2V4`MiLBoNo2X4_MiLDlNm2[4`MiLDhNn2`4^MhLEeNo2c4\\MhLGaNn2g4\\MhLG]No2l4ZMgLHZNP3o4XMfLR5Z3oJfLQ5Z3oJfLP5Z3QKfLn4Z3SKfLm4Z3SKfLl4Z3UKfLj4[3VKeLj4Z3WKfLh4[3XKeLh4Z3YKfLf4[3ZKeLe4[3\\KeLd4[3\\KeLc4[3^KeLa4[3`KeL`4[3`KeL_4[3bKgL\\4Y3cKiL[4W3fKkLW4V3iKlLU4o1UIRMg2Q1Q4j1\\IUMc2R1o3f1bIWM`2U1l3_1hI]M\\2V1i3Z1oI_MY2X1g3T1UJcMU2[1c3o0[JgMR2\\1`3l0`JgMQ2_1]3i0cJhMP2`1[3h0fJgMP2c1W3f0iJhMP2d1U3c0lJhMP2f1T3`0mJjMo1h1S3<nJlMP2j1Q38PKnMm1n1Q33SKnMk1Q2S3ORKPNj1S2U3KRKRNf1V2Y3FRKSNd1Y2Z3CRKUNb1Z2[3ATKTN`1]2\\3^OUKUN\\1`2_3ZOVKVNY1b2a3WOVKXNW1c2b3UOXKWNU1g2b3QOZKXNQ1j2e3mNZKZNo0k2g3jN[K[Nl0m2i3gN\\K[Nj0P3i3eN]K\\Ng0R3l3aN^K]Nd0T3n3^N_K^N`0W3Q4ZN_K`N>X3S4VNaKaN:\\3T4RNdKbN6^3V4nMfKdN1a3Y4iMgKgNNb3[4eMiKhNJf3]4`MkKjNFh3^4]MmKlNBj3a4XMoKnN^Ol3c4TMQLoNZOP4g4mLPLTOWOQ4j4hLQLWOSOS4n4bLQL[OnNV4R5\\LQL_OkNW4V5VLQLBgN[4Y5PLRLEcN]4]5jKRLI^N`4a5eKQLL\\Na4e5_KQLOXNe4h5YKRL2TNg4k5TKRL6PNh4o5PKSL8lMi4S6lJSL:hMm4V6fJSL>eMm4Z6bJSLa0`Mo4_6]JTLc0[MQ5c6YJULd0VMV5f6SJVLg0RMW5j6oIWLk6n6eLWFAj96cFF^9OQGKQ9K\\G1d8FjG6V8DSH8n7C[H8e7DcH9]7BlH9U7BSI:m6_NkEh0a3e0d6^NSFe0a3i0m6PO\\Ik0h6mN\\IS1g6fN]I[1Q;0000SOg@9X?Gj@7V?Il@5T?Ko@2Q?NQA1n>OUANk>2WALi>5XAJg>6\\AGd>9^AEb>;aAB_>>cAA\\>?fA_OZ>a0P100001O00000000000000001O00000000000000001O00000000000000001O0000000000000m@Gd=o1I:Fn0RO5K1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O5K;E;E;E;E;E;E:F;E:FO1N2O1N2O1N2O1N2O1O1O1O100O1O100O100O100O1O100O100O100O1O100O10000000000000000000000000000001O000000000000000000000000000O2O00000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001K4M4L3M4K5L3M4L3M4K4010O100O100O100OL5H7H9I6L5K5K6J5K6J5K6J5K6J5L5J5L5J5L5K4K6K=Ba0@a0_O?@:G8H8G9H8G:G8H`c^2"}, "label": "largest , closest horse"}]}
{"id": 60043, "image": "COCO_train2014_000000060043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with short blonde hair wearing a blue sweater reaching for condiments\"."}], "task": "refering", "answer_template": "The \"a woman with short blonde hair wearing a blue sweater reaching for condiments\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 70, 71, 72, 73, 74, 87, 88, 89, 90, 91, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 138, 139, 140, 141, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 331, 340, 341, 342, 343, 344, 345, 346, 347, 348, 357, 358, 359, 360, 361, 362, 365], "bbox": [0.007770833333333334, 0.165328125, 0.5347708333333333, 0.9537656250000002], "iscrowd": 0, "area": 72152.98945000001, "rle": {"size": [640, 480], "counts": "dm2k0`b0e0K5K6J5K6K4K5K6J5cAZM[;k2_DeMR;a2gDoMi:W2QEYN`:m1YEcNW:c1cEmNm9Y1mEnNm9X1oEkNk9[1SFgNg9`1VFbNe9c1XF`Nb9f1\\F\\N_9j1^FWN]9o1aFSNY9S2eFoMV9W2fFlMU9Y2iFiMW9X2fFjMY9W2eFkMZ9V2dFlM[9V2aFlM_9U2_FmMa9S2]FoMb9S2[FoMd9R2ZFPNe9V6O1O010O1O1O1O100O001O1O1O100O1O1O001O100O1dM]EZKd:a4bE]K_:_4gE_KX:^4nE_KS:\\4TFaKm9[4XFcKi9X4^FeKc9W4cFgK]9U4hFiKY9T4lFiKU9T4PGiKP9V4TGgKm8V4WGiKi8T4\\GiKe8U4_GhKb8U4fGeK[8X4PH_KQ8_4ZHXKe7f4fHQK[7m4m3M3M3N2N20O01000000000000O0100000000000O01000000000O1000O1000000O010000000O1000O10O10000000O01000000O1000O1000O1000001N2O001O1O1N101O1O1N2O0]EfK`5[4^JiK_5X4`JkK]5V4bJlK[5U4eJnKX5S4fJQLW5P4hJSLU5n3jJULR5l3nJVLP5k3nJYLo4h3PK[Lm4f3PK_Lm4b3PKcLm4]3QKgLl4[3QKjLl4W3PKoLm4R3PKSMm4m2QKWMm4j2oJ\\Mm4f2PK_Mm4b2PKcMm4]2QKhMl4Y2PKlMn4U2oJPNn4Q2UJVJoNn3h6m1SJ]JPOl3j6h1nIeJTOh3k6d1jIlJWOc3R4kKVNg7cMaLf3YLgNl;h0gDWOY;7ZEGg:GkE9U:WO\\Fh0c9TOcFk0]9POhFP1X9kNnFS1S9hNRGX1n8cNXG\\1g8dN[G[1e8eN[G[1e8eN\\GZ1d8fN]GX1c8iN]GW1c8hN_GW1a8iN`GV1`8jN`GV1_8kNbGT1^8lNcGR1^8nNbGR1^8nNcGQ1]8oNdGP1[8QOeGo0[8QOfGm0[8SOeGm0[8SOfGl0Y8UOhGj0X8fN\\CL]4^1W8dN^CN\\4]1V8cN`C0[4[1U8dNbC1Y4[1U8cNdC1X4[1T8bNfC3W4Z1S8aNhC5U4Z1R8`NkC6T4X1R8aNkC7T4W1Q8`NmC9R4W1Q8^NoC;Q4V1f8jN[GU1e8kN[GU1e8kN\\GU1c8kN]GU1c8kN^GU1a8kN`GT1`8lN`GU1`8jNaGU1_8kNbGU1]8jNdGV1\\8jNeGV1Z8jNgGU1Y8kNgGV1X8jNiGV1V8jNkGU1U8kNkGV1T8jNmGU1S8kNnGU1Q8kNoGU1Q8kNPHU1o7kNQHU1o7kNRHU1n7jNSHU1m7kNSHV1l7jNUHU1k7kNVHU1i7kNWHU1j7jNWHV1h7jNYHU1h7jN`Ho0_7POkHg0V7XOSI`0l6@]I7d6HeI0[6Ok601O001O00Zd[4"}, "label": "a woman with short blonde hair wearing a blue sweater reaching for condiments"}]}
{"id": 60043, "image": "COCO_train2014_000000060043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing wearing grey button up sweater\"."}], "task": "refering", "answer_template": "The \"woman standing wearing grey button up sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 70, 71, 72, 73, 74, 87, 88, 89, 90, 91, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 138, 139, 140, 141, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 331, 340, 341, 342, 343, 344, 345, 346, 347, 348, 357, 358, 359, 360, 361, 362, 365], "bbox": [0.007770833333333334, 0.165328125, 0.5347708333333333, 0.9537656250000002], "iscrowd": 0, "area": 72152.98945000001, "rle": {"size": [640, 480], "counts": "dm2k0`b0e0K5K6J5K6K4K5K6J5cAZM[;k2_DeMR;a2gDoMi:W2QEYN`:m1YEcNW:c1cEmNm9Y1mEnNm9X1oEkNk9[1SFgNg9`1VFbNe9c1XF`Nb9f1\\F\\N_9j1^FWN]9o1aFSNY9S2eFoMV9W2fFlMU9Y2iFiMW9X2fFjMY9W2eFkMZ9V2dFlM[9V2aFlM_9U2_FmMa9S2]FoMb9S2[FoMd9R2ZFPNe9V6O1O010O1O1O1O100O001O1O1O100O1O1O001O100O1dM]EZKd:a4bE]K_:_4gE_KX:^4nE_KS:\\4TFaKm9[4XFcKi9X4^FeKc9W4cFgK]9U4hFiKY9T4lFiKU9T4PGiKP9V4TGgKm8V4WGiKi8T4\\GiKe8U4_GhKb8U4fGeK[8X4PH_KQ8_4ZHXKe7f4fHQK[7m4m3M3M3N2N20O01000000000000O0100000000000O01000000000O1000O1000000O010000000O1000O10O10000000O01000000O1000O1000O1000001N2O001O1O1N101O1O1N2O0]EfK`5[4^JiK_5X4`JkK]5V4bJlK[5U4eJnKX5S4fJQLW5P4hJSLU5n3jJULR5l3nJVLP5k3nJYLo4h3PK[Lm4f3PK_Lm4b3PKcLm4]3QKgLl4[3QKjLl4W3PKoLm4R3PKSMm4m2QKWMm4j2oJ\\Mm4f2PK_Mm4b2PKcMm4]2QKhMl4Y2PKlMn4U2oJPNn4Q2UJVJoNn3h6m1SJ]JPOl3j6h1nIeJTOh3k6d1jIlJWOc3R4kKVNg7cMaLf3YLgNl;h0gDWOY;7ZEGg:GkE9U:WO\\Fh0c9TOcFk0]9POhFP1X9kNnFS1S9hNRGX1n8cNXG\\1g8dN[G[1e8eN[G[1e8eN\\GZ1d8fN]GX1c8iN]GW1c8hN_GW1a8iN`GV1`8jN`GV1_8kNbGT1^8lNcGR1^8nNbGR1^8nNcGQ1]8oNdGP1[8QOeGo0[8QOfGm0[8SOeGm0[8SOfGl0Y8UOhGj0X8fN\\CL]4^1W8dN^CN\\4]1V8cN`C0[4[1U8dNbC1Y4[1U8cNdC1X4[1T8bNfC3W4Z1S8aNhC5U4Z1R8`NkC6T4X1R8aNkC7T4W1Q8`NmC9R4W1Q8^NoC;Q4V1f8jN[GU1e8kN[GU1e8kN\\GU1c8kN]GU1c8kN^GU1a8kN`GT1`8lN`GU1`8jNaGU1_8kNbGU1]8jNdGV1\\8jNeGV1Z8jNgGU1Y8kNgGV1X8jNiGV1V8jNkGU1U8kNkGV1T8jNmGU1S8kNnGU1Q8kNoGU1Q8kNPHU1o7kNQHU1o7kNRHU1n7jNSHU1m7kNSHV1l7jNUHU1k7kNVHU1i7kNWHU1j7jNWHV1h7jNYHU1h7jN`Ho0_7POkHg0V7XOSI`0l6@]I7d6HeI0[6Ok601O001O00Zd[4"}, "label": "woman standing wearing grey button up sweater"}]}
{"id": 60043, "image": "COCO_train2014_000000060043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black drinking from a cup\"."}], "task": "refering", "answer_template": "The \"a woman in black drinking from a cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 102, 103, 104, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 187, 204, 221, 238, 255, 272, 289, 306, 323, 340], "bbox": [0.003, 0.24718749999999998, 0.13183333333333333, 0.8943749999999999], "iscrowd": 0, "area": 8098.860700000002, "rle": {"size": [640, 480], "counts": "Zm06cc0c1]Nc1]Nd1\\Nc1\\NRMdC4i:_OTFX1e8kNgGo0R8SOZHg0]7\\OoH=j6EbI4Z6KRJNi52cJGY57TKAi4>]KROn4m0WKdNT5[1RKWNV5g1RKPNR5n1UKhMo4W2XK`Mj4_2^KXMc4i2cKnL`4R3T6010O3N3L4O1010O000000000010O0000000000010O000001O1O101N1O1O1O1O1O100O1O1kM[_On0g`0WN]@^1^a0DW]T8"}, "label": "a woman in black drinking from a cup"}]}
{"id": 60043, "image": "COCO_train2014_000000060043.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with brunette hair and a black shirt\"."}], "task": "refering", "answer_template": "The \"a woman with brunette hair and a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 102, 103, 104, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 187, 204, 221, 238, 255, 272, 289, 306, 323, 340], "bbox": [0.003, 0.24718749999999998, 0.13183333333333333, 0.8943749999999999], "iscrowd": 0, "area": 8098.860700000002, "rle": {"size": [640, 480], "counts": "Zm06cc0c1]Nc1]Nd1\\Nc1\\NRMdC4i:_OTFX1e8kNgGo0R8SOZHg0]7\\OoH=j6EbI4Z6KRJNi52cJGY57TKAi4>]KROn4m0WKdNT5[1RKWNV5g1RKPNR5n1UKhMo4W2XK`Mj4_2^KXMc4i2cKnL`4R3T6010O3N3L4O1010O000000000010O0000000000010O000001O1O101N1O1O1O1O1O100O1O1kM[_On0g`0WN]@^1^a0DW]T8"}, "label": "a woman with brunette hair and a black shirt"}]}
{"id": 453260, "image": "COCO_train2014_000000453260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white toliet\"."}], "task": "refering", "answer_template": "The \"the white toliet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 251, 252, 253, 254, 268, 269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 321, 322, 330, 331, 332, 333, 334, 335, 336, 337, 338, 347, 348, 349, 350, 351, 352, 353, 354, 355, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.45543750000000005, 0.586515625, 0.9797708333333334, 0.986515625], "iscrowd": 0, "area": 34434.12575000001, "rle": {"size": [640, 480], "counts": "S]Y4o0Qc0c0\\O3N3L4M3L4M2M4L4M3L3N3L4M1N100O1O100O100O100O100O1O100O100O100O100O1O100O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000L4K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5M3O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O12N1O1O1O1YNmC_KT<]4PDbKQ<Z4SDeKo;V4UDiKl;S4XDlKi;P4ZDPLg;l3]DSLd;i3`DVLa;f3cDYL_;a3fD^L[;^3iDaLX;[3kDeLV;W3nDhLS;T3QEkLP;Q3TEmLn:o2VEjLR;Q3REhLU;T3oDeLX;W3kDcL\\;Y3hD`L_;[3f2L4L5K4L4L4L4L4L4L5K4K5L4LPk5"}, "label": "the white toliet"}]}
{"id": 453260, "image": "COCO_train2014_000000453260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white toilet including bowl , seat and raised lid\"."}], "task": "refering", "answer_template": "The \"the white toilet including bowl , seat and raised lid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 251, 252, 253, 254, 268, 269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 321, 322, 330, 331, 332, 333, 334, 335, 336, 337, 338, 347, 348, 349, 350, 351, 352, 353, 354, 355, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.45543750000000005, 0.586515625, 0.9797708333333334, 0.986515625], "iscrowd": 0, "area": 34434.12575000001, "rle": {"size": [640, 480], "counts": "S]Y4o0Qc0c0\\O3N3L4M3L4M2M4L4M3L3N3L4M1N100O1O100O100O100O100O1O100O100O100O100O1O100O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000L4K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5M3O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O12N1O1O1O1YNmC_KT<]4PDbKQ<Z4SDeKo;V4UDiKl;S4XDlKi;P4ZDPLg;l3]DSLd;i3`DVLa;f3cDYL_;a3fD^L[;^3iDaLX;[3kDeLV;W3nDhLS;T3QEkLP;Q3TEmLn:o2VEjLR;Q3REhLU;T3oDeLX;W3kDcL\\;Y3hD`L_;[3f2L4L5K4L4L4L4L4L4L5K4K5L4LPk5"}, "label": "the white toilet including bowl , seat and raised lid"}]}
{"id": 264846, "image": "COCO_train2014_000000264846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue shirt and striped tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue shirt and striped tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 216], "bbox": [0.284859375, 0.015081967213114755, 0.424296875, 0.6600468384074941], "iscrowd": 0, "area": 14804.633600000001, "rle": {"size": [427, 640], "counts": "\\T\\21Q==B>B=D=B=D:E;E<SMXMjJR3S5dMoIg2n5PNTIZ2i6W2L3M3M4L2N2N2N2N2N2N2N2O100O1O1O00M4L3M3M3M4L3K5J^Oa0N35K7I7H8I7I7I8G8I7K58H9G8I7H9G8H9G8I8GeISJ`5e5SJiJ^6e4VIgKj6^5O010O10O01O10O\\ObI[J_6]5iIbJW6T5TJlJl5[4mJdKT5c3eK]LZ4R3XLmLi3Q3XLoLh3o2[LPMf3m2]LRMd3l2^LSMb3k2aLTM`3j2bLUMd3c2_L\\Mm3W2ULhMW4j1lKUN_4_1cK`Ni4R1ZKmNR5f0oJZO]58fJFf5N\\J1`9O1O\\ci4"}, "label": "a man wearing a blue shirt and striped tie"}]}
{"id": 264846, "image": "COCO_train2014_000000264846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gentleman wearing white slacks , blue shirt , and striped tie stands at the rear of a grass tennis court\"."}], "task": "refering", "answer_template": "The \"a gentleman wearing white slacks , blue shirt , and striped tie stands at the rear of a grass tennis court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 216], "bbox": [0.284859375, 0.015081967213114755, 0.424296875, 0.6600468384074941], "iscrowd": 0, "area": 14804.633600000001, "rle": {"size": [427, 640], "counts": "\\T\\21Q==B>B=D=B=D:E;E<SMXMjJR3S5dMoIg2n5PNTIZ2i6W2L3M3M4L2N2N2N2N2N2N2N2O100O1O1O00M4L3M3M3M4L3K5J^Oa0N35K7I7H8I7I7I8G8I7K58H9G8I7H9G8H9G8I8GeISJ`5e5SJiJ^6e4VIgKj6^5O010O10O01O10O\\ObI[J_6]5iIbJW6T5TJlJl5[4mJdKT5c3eK]LZ4R3XLmLi3Q3XLoLh3o2[LPMf3m2]LRMd3l2^LSMb3k2aLTM`3j2bLUMd3c2_L\\Mm3W2ULhMW4j1lKUN_4_1cK`Ni4R1ZKmNR5f0oJZO]58fJFf5N\\J1`9O1O\\ci4"}, "label": "a gentleman wearing white slacks , blue shirt , and striped tie stands at the rear of a grass tennis court"}]}
{"id": 264846, "image": "COCO_train2014_000000264846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman playing tennis\"."}], "task": "refering", "answer_template": "The \"a woman playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 150, 151, 152, 173, 174, 175, 196, 197, 198, 199, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 314, 315, 317], "bbox": [0.5247499999999999, 0.38651053864168616, 0.820109375, 0.9056206088992974], "iscrowd": 0, "area": 13799.42685, "rle": {"size": [427, 640], "counts": "kZ\\4i0b<0O001O100O001O10O01O1O10O01O1_NROdFo0\\9XO]Fg0d9@TFa0k9FnE;R:LfE5Y:2aEM`::XEGg:_100O01000O10O0100O10O010000O010O1001O1000O1000000hGaMc5_2UJiMj5Y2lIPNT6P2dIXN\\6h1[I`Nf6`1RIhNn6X1nHlNR7T1jHPOU7Q1gHSOY7m0cHVO^7j0^HZOb7g0YH]Og7c0UHAj7`0RHDn7l210000CQH]Ko7a4RH`Km7^4VHaKk7\\4WHeKi7Y4XHhKh7U4[HkKe7S4\\HnKc7P4_HPLb7m3aHSL_7k3bHVL^7g3dHZL\\7c3gH]LY7a3hH`LY7\\3iHdLY7Y3iHgLX7U3jHlLV73PHT2l0jMU7MSHW2j0lMT7GWHZ2g0oMT7BZH\\2c0SNk7m1UHSNk7n1SHSNn7l1RHTNn7m1PHSNQ8m1nGTNZO\\O[8`2[HTNXO]O_8_2WHUNYO\\Oa8_2UHVNXO\\Od8_2SHTNYO]Of8^2PHVNXO]Oi8]2nGWNWO]Ol8]2lGVNWO]Oo8\\2iG^NX8c1fG]N[8c1eG\\N\\8e1bGZNa8e1^G[Nc8e1]GZNd8g1ZGYNh8f1WGZNj8g1UGXNl8h1SGXNo8g1PGYNQ9g1oFXNR9g1nFYNS9f1nFYNT9f1kFZNV9e1jF[NW9d1jF[NX9d1gF\\NZ9c1hF[NY9d1iFZNY9e1gFZNZ9e1hFYNY9f1hFYNY9g1hFWNZ9g1gFXNP;1REMn:4SEJm:6TEIl:7VEGj::VEEj:;XECi:<XECh:>R100001O000010O0001O00001O000010O01O00001O000010O0001O00001O00100O1O1O1O1O1O1O1O100N2O001O1O1N2O1O1O1O1N2O1O1O000O01O1Ol0TOVc_1"}, "label": "a woman playing tennis"}]}
{"id": 264846, "image": "COCO_train2014_000000264846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female tennis player in white outfit\"."}], "task": "refering", "answer_template": "The \"female tennis player in white outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 150, 151, 152, 173, 174, 175, 196, 197, 198, 199, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 314, 315, 317], "bbox": [0.5247499999999999, 0.38651053864168616, 0.820109375, 0.9056206088992974], "iscrowd": 0, "area": 13799.42685, "rle": {"size": [427, 640], "counts": "kZ\\4i0b<0O001O100O001O10O01O1O10O01O1_NROdFo0\\9XO]Fg0d9@TFa0k9FnE;R:LfE5Y:2aEM`::XEGg:_100O01000O10O0100O10O010000O010O1001O1000O1000000hGaMc5_2UJiMj5Y2lIPNT6P2dIXN\\6h1[I`Nf6`1RIhNn6X1nHlNR7T1jHPOU7Q1gHSOY7m0cHVO^7j0^HZOb7g0YH]Og7c0UHAj7`0RHDn7l210000CQH]Ko7a4RH`Km7^4VHaKk7\\4WHeKi7Y4XHhKh7U4[HkKe7S4\\HnKc7P4_HPLb7m3aHSL_7k3bHVL^7g3dHZL\\7c3gH]LY7a3hH`LY7\\3iHdLY7Y3iHgLX7U3jHlLV73PHT2l0jMU7MSHW2j0lMT7GWHZ2g0oMT7BZH\\2c0SNk7m1UHSNk7n1SHSNn7l1RHTNn7m1PHSNQ8m1nGTNZO\\O[8`2[HTNXO]O_8_2WHUNYO\\Oa8_2UHVNXO\\Od8_2SHTNYO]Of8^2PHVNXO]Oi8]2nGWNWO]Ol8]2lGVNWO]Oo8\\2iG^NX8c1fG]N[8c1eG\\N\\8e1bGZNa8e1^G[Nc8e1]GZNd8g1ZGYNh8f1WGZNj8g1UGXNl8h1SGXNo8g1PGYNQ9g1oFXNR9g1nFYNS9f1nFYNT9f1kFZNV9e1jF[NW9d1jF[NX9d1gF\\NZ9c1hF[NY9d1iFZNY9e1gFZNZ9e1hFYNY9f1hFYNY9g1hFWNZ9g1gFXNP;1REMn:4SEJm:6TEIl:7VEGj::VEEj:;XECi:<XECh:>R100001O000010O0001O00001O000010O01O00001O000010O0001O00001O00100O1O1O1O1O1O1O1O100N2O001O1O1N2O1O1O1O1N2O1O1O000O01O1Ol0TOVc_1"}, "label": "female tennis player in white outfit"}]}
{"id": 43664, "image": "COCO_train2014_000000043664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza on the plate near the knife\"."}], "task": "refering", "answer_template": "The \"the pizza on the plate near the knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 320, 341, 342, 343], "bbox": [0.08246875000000001, 0.2741686182669789, 0.9610468749999999, 0.9887587822014051], "iscrowd": 0, "area": 79885.38949999999, "rle": {"size": [427, 640], "counts": "[Zf06n<8I7N3M2O1O1N3N1O1N2O2N1N2O2N1N2O1N3N1O1N2O2N0O2O1O1N101O1N101O1N101O1N101O1N2O001N2O001N2O0O2O1O1N101O1N101O1N101O1N2O001N2O001N2O001N2O1O0O2O1O0O2O1O0O2O1N2O000O2O000O10001N10000O101O0O10000O2O000O101O0O10000O2O0O10000O2O000O10001N10000O101O0O10000O2O000O10001N10000O2O0O1000001N100000001N1000000O2O00000O101O00001N101O000O2O001O000O2O00001N101O00001N101O000O2O00001N101O00001N101O000O2O00001O0O2O00001N101O00001N10001O0O2O00001N101O00000O1000000O10001O000O1000000O1000001O0O1000000O1000000O2O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101N10000001O00001O0001O01O0000001O00010O0001O01O01O00010O00010O000010O0001O01O01O01O01O00010O001O02N2O1N2N2O2M2O1N2N2O1N2O1N2N3N1N2O1N2N2O1N2O1N3M2O1N2N2O1N2O1N2N3N1N2O1N2N2O1N2O1N3M2O1N2N100O1O2N100O1O1O100O1O1O100O1O2N100O1O1O100O1O1O100O2N1O100O1O1O100O1O1O101N1O100O1O1O100O1O1O100O2N1O100O1O1O100O1O1O101N1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O1MWDbNk;\\1300O101N100O2N100O101N101N2N2O0O2O1N2O0O2N2O1N101N2L3N3L4M3LhP:"}, "label": "the pizza on the plate near the knife"}]}
{"id": 43664, "image": "COCO_train2014_000000043664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pizza on a plate with a knife next to it\"."}], "task": "refering", "answer_template": "The \"a slice of pizza on a plate with a knife next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 320, 341, 342, 343], "bbox": [0.08246875000000001, 0.2741686182669789, 0.9610468749999999, 0.9887587822014051], "iscrowd": 0, "area": 79885.38949999999, "rle": {"size": [427, 640], "counts": "[Zf06n<8I7N3M2O1O1N3N1O1N2O2N1N2O2N1N2O1N3N1O1N2O2N0O2O1O1N101O1N101O1N101O1N101O1N2O001N2O001N2O0O2O1O1N101O1N101O1N101O1N2O001N2O001N2O001N2O1O0O2O1O0O2O1O0O2O1N2O000O2O000O10001N10000O101O0O10000O2O000O101O0O10000O2O0O10000O2O000O10001N10000O101O0O10000O2O000O10001N10000O2O0O1000001N100000001N1000000O2O00000O101O00001N101O000O2O001O000O2O00001N101O00001N101O000O2O00001N101O00001N101O000O2O00001O0O2O00001N101O00001N10001O0O2O00001N101O00000O1000000O10001O000O1000000O1000001O0O1000000O1000000O2O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101O0O100O10000O100O101N10000001O00001O0001O01O0000001O00010O0001O01O01O00010O00010O000010O0001O01O01O01O01O00010O001O02N2O1N2N2O2M2O1N2N2O1N2O1N2N3N1N2O1N2N2O1N2O1N3M2O1N2N2O1N2O1N2N3N1N2O1N2N2O1N2O1N3M2O1N2N100O1O2N100O1O1O100O1O1O100O1O2N100O1O1O100O1O1O100O2N1O100O1O1O100O1O1O101N1O100O1O1O100O1O1O100O2N1O100O1O1O100O1O1O101N1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O1MWDbNk;\\1300O101N100O2N100O101N101N2N2O0O2O1N2O0O2N2O1N101N2L3N3L4M3LhP:"}, "label": "a slice of pizza on a plate with a knife next to it"}]}
{"id": 43664, "image": "COCO_train2014_000000043664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 slices of pizza in a blurry shot\"."}], "task": "refering", "answer_template": "The \"2 slices of pizza in a blurry shot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 48, 49, 50, 51, 52, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 77, 78, 79, 80, 81, 82, 83, 84, 85], "bbox": [0.0609375, 0.0, 0.774609375, 0.26107728337236535], "iscrowd": 0, "area": 31412.093699999998, "rle": {"size": [427, 640], "counts": "]X`0=n<k0UO=C1O1O001O1O001O1O001O00001O001O001O00001O001O001O00001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O000000001O000000001O000000001O00000000001O000000001O000000001O01O000001O000000001O00000000N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O10000O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000001O1O1O001O1lEhNS8Y1jGiNV8W1hGkNX8V1dGmN\\8T1aGnN_8S1^GPOa8P1\\GSOd8n0YGTOg8m0VGUOj8l0RGWOn8i0PGZOo8g0nF[OR9f0jF]OV9c0hF_OX9b0eF@[9a0aFB_9>_FE`9<]FFc9;YFHg98WFJi97TFKl9j101O000000000000000000000000000000000000000000mLVFe2j9ZMYFd2g9ZM^Fc2b9\\MbFb2]9\\MgFb2Y9]MhFc2X9[MjFe2n901O0000000000000000001O0000000000000000001O0000000000000000001O00000O1000000000001O0000000000000O10001O0000000000000000001N1000000000000O10001O00000O10000000000O101O000000000O100000001O0O1000000000000O101O0000000O10000000001N1000000000000O10001O000O10000O10000O101O0O1000000O10000O101O0O10000O1000000O101O0O10000O10000O10000O100O001O1O1O010O1O1O00100O1O001O1O10O01O1O1O00100O1O001O1O10O01O1O1N4K4L4L5J5L[Pl1"}, "label": "2 slices of pizza in a blurry shot"}]}
{"id": 35473, "image": "COCO_train2014_000000035473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train that is partially brown\"."}], "task": "refering", "answer_template": "The \"train that is partially brown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 319, 320, 321, 344], "bbox": [0.741609375, 0.12195833333333334, 1.0, 0.8537083333333334], "iscrowd": 0, "area": 41610.57075, "rle": {"size": [480, 640], "counts": "Wjn6h2V<W1iN=D;D6K2O1N100O1O10N1N3M2O25J9H3L3N1N1O2O1N2O0O100O100O100O1O100O100O100O100O3N2M4L4M2M3N3L2O0O2O0O2N100O2O0O2N1O2N1O2N1O2N101N10001O001O0O2^OcH^I^7a6fH\\I[7b6fH\\I\\7d6dHZI_7e6_H\\Ib7d6]H[If7c6<1O100O9HO0010O010O01N1O2N1O2M2O2N1O2M2O2N1O2O1N1O2N101N2N1O2O1N1VOVG]Kl8`4YG\\Kh8c4[G[Kf8d4[GZKg8f4YGXKh8h4ZGVKg8j4YGTKh8m4XGQKj8n4XGoJj8o4XGPKh8P5e0O10O100O1000O0100O10000OO2O1N101N1O2O1N101N1O2O0O1O2O0O2O0O2N101N100O2N101N101N1O2O0OmM^LkHf1"}, "label": "train that is partially brown"}]}
{"id": 35473, "image": "COCO_train2014_000000035473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a trolley car ahead of a bus\"."}], "task": "refering", "answer_template": "The \"a trolley car ahead of a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 319, 320, 321, 344], "bbox": [0.741609375, 0.12195833333333334, 1.0, 0.8537083333333334], "iscrowd": 0, "area": 41610.57075, "rle": {"size": [480, 640], "counts": "Wjn6h2V<W1iN=D;D6K2O1N100O1O10N1N3M2O25J9H3L3N1N1O2O1N2O0O100O100O100O1O100O100O100O100O3N2M4L4M2M3N3L2O0O2O0O2N100O2O0O2N1O2N1O2N1O2N101N10001O001O0O2^OcH^I^7a6fH\\I[7b6fH\\I\\7d6dHZI_7e6_H\\Ib7d6]H[If7c6<1O100O9HO0010O010O01N1O2N1O2M2O2N1O2M2O2N1O2O1N1O2N101N2N1O2O1N1VOVG]Kl8`4YG\\Kh8c4[G[Kf8d4[GZKg8f4YGXKh8h4ZGVKg8j4YGTKh8m4XGQKj8n4XGoJj8o4XGPKh8P5e0O10O100O1000O0100O10000OO2O1N101N1O2O1N101N1O2O0O1O2O0O2O0O2N101N100O2N101N101N1O2O0OmM^LkHf1"}, "label": "a trolley car ahead of a bus"}]}
{"id": 281237, "image": "COCO_train2014_000000281237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza with spinach and tomatoes\"."}], "task": "refering", "answer_template": "The \"the pizza with spinach and tomatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.094375, 0.298875, 0.9977499999999999, 0.9797708333333333], "iscrowd": 0, "area": 157380.68759999998, "rle": {"size": [480, 640], "counts": "_\\l02h>d0\\O;E;I6J7I6K6I6K6I7I6K6I6J7J6I6J7J5J4L3N2M3M3N2M3M2O2M3M3N2M3N2O0O2O1N2O1N2N2O0O2O1N2O1N2O1N2N101N2O1N2O1N2N101N2O1N2O0O2O1N1O2O1N101N2O0O2N2O0O2O1O0O2O001N2O001N2O001N2O001O1N101O1N101O1N101O1N101O0O2O1O0O2O1O0O2O1O001N2O001N2O001N2O001N2O001N2O001N101O1N101O1O0O2O1O0O2O1O0O2O1O0O2O1O0O2O1O0O2O001N2O001N2O001N2O001N2O000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O11O000000000000001O000000000000001O0000000000001O000000000001O01O000000000000001O0000000000001O000000000000001O0000000000000000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000001O00001O00001O001O00001O00001O001O00001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O010O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1B=UOl0TOl0TOk0UOl0TOl0TOk0UOb7"}, "label": "the pizza with spinach and tomatoes"}]}
{"id": 281237, "image": "COCO_train2014_000000281237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with fresh mozzarella and halved cherry tomatoes\"."}], "task": "refering", "answer_template": "The \"a pizza with fresh mozzarella and halved cherry tomatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.094375, 0.298875, 0.9977499999999999, 0.9797708333333333], "iscrowd": 0, "area": 157380.68759999998, "rle": {"size": [480, 640], "counts": "_\\l02h>d0\\O;E;I6J7I6K6I6K6I7I6K6I6J7J6I6J7J5J4L3N2M3M3N2M3M2O2M3M3N2M3N2O0O2O1N2O1N2N2O0O2O1N2O1N2O1N2N101N2O1N2O1N2N101N2O1N2O0O2O1N1O2O1N101N2O0O2N2O0O2O1O0O2O001N2O001N2O001N2O001O1N101O1N101O1N101O1N101O0O2O1O0O2O1O0O2O1O001N2O001N2O001N2O001N2O001N2O001N101O1N101O1O0O2O1O0O2O1O0O2O1O0O2O1O0O2O1O0O2O001N2O001N2O001N2O001N2O000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O11O000000000000001O000000000000001O0000000000001O000000000001O01O000000000000001O0000000000001O000000000000001O0000000000000000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000001O00001O00001O001O00001O00001O001O00001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O010O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1B=UOl0TOl0TOk0UOl0TOl0TOk0UOb7"}, "label": "a pizza with fresh mozzarella and halved cherry tomatoes"}]}
{"id": 19096, "image": "COCO_train2014_000000019096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the longest wood skis in the scene being carried\"."}], "task": "refering", "answer_template": "The \"the longest wood skis in the scene being carried\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 83, 84, 85, 86, 100, 101], "bbox": [0.5555399999999999, 0.29839879154078547, 0.8613599999999999, 0.4696374622356495], "iscrowd": 0, "area": 1166.288099999999, "rle": {"size": [331, 500], "counts": "WPj26T:10O0100O01000O01000O0100O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O100O010O1Mkb20S]M40100O1O010O1000O01000000O0100000O0100000O10O1000O10O1000O10O1000O0100000O010000O01000O10O10O1000O010000O0100O10O10O10O01000O0100O01000O010O1000O01000O0100O01000O10O01000O010000O010004L0O101O1O1Ni`f0"}, "label": "the longest wood skis in the scene being carried"}]}
{"id": 19096, "image": "COCO_train2014_000000019096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long skis the man on the right is holding\"."}], "task": "refering", "answer_template": "The \"the long skis the man on the right is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 83, 84, 85, 86, 100, 101], "bbox": [0.5555399999999999, 0.29839879154078547, 0.8613599999999999, 0.4696374622356495], "iscrowd": 0, "area": 1166.288099999999, "rle": {"size": [331, 500], "counts": "WPj26T:10O0100O01000O01000O0100O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O100O010O1Mkb20S]M40100O1O010O1000O01000000O0100000O0100000O10O1000O10O1000O10O1000O0100000O010000O01000O10O10O1000O010000O0100O10O10O10O01000O0100O01000O010O1000O01000O0100O01000O10O01000O010000O010004L0O101O1O1Ni`f0"}, "label": "the long skis the man on the right is holding"}]}
{"id": 510617, "image": "COCO_train2014_000000510617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"side of passenger bus , parked behind another passenger bus\"."}], "task": "refering", "answer_template": "The \"side of passenger bus , parked behind another passenger bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 184, 185, 186, 207, 208, 230, 231], "bbox": [0.0015, 0.14156908665105386, 0.367328125, 0.7168618266978923], "iscrowd": 0, "area": 27379.109800000006, "rle": {"size": [427, 640], "counts": "W?U2W;W4hKY1gN00000001N10000000001O000O1000001O0000000O2O00000000001O0O100000001O00000O10001O000000000O2QOn0ROn0O1000001N10000O10001O0O10000O101O000O1000001N10000O10001O0O10000O101O000O1O1N2O2N1N2O1N2O1O2M2O1N2O1O1N3N1N2O1O1N2O2M2O1N2O1O1N3N1O1000001N10000000001O000O1000001O0000000O101O00000000001N10000O1O1O2M2O1O1O1N2O2N1O1N2O1O1F;YOf0O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N100O1O1O1O2N1O1O1O10001O00000000001O00000000001O00000000001O00000000001O000000001O00000O100O2N1O1O1O1O1O2N_VY5"}, "label": "side of passenger bus , parked behind another passenger bus"}]}
{"id": 510617, "image": "COCO_train2014_000000510617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus that is closest to the building\"."}], "task": "refering", "answer_template": "The \"the bus that is closest to the building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 184, 185, 186, 207, 208, 230, 231], "bbox": [0.0015, 0.14156908665105386, 0.367328125, 0.7168618266978923], "iscrowd": 0, "area": 27379.109800000006, "rle": {"size": [427, 640], "counts": "W?U2W;W4hKY1gN00000001N10000000001O000O1000001O0000000O2O00000000001O0O100000001O00000O10001O000000000O2QOn0ROn0O1000001N10000O10001O0O10000O101O000O1000001N10000O10001O0O10000O101O000O1O1N2O2N1N2O1N2O1O2M2O1N2O1O1N3N1N2O1O1N2O2M2O1N2O1O1N3N1O1000001N10000000001O000O1000001O0000000O101O00000000001N10000O1O1O2M2O1O1O1N2O2N1O1N2O1O1F;YOf0O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N100O1O1O1O2N1O1O1O10001O00000000001O00000000001O00000000001O00000000001O000000001O00000O100O2N1O1O1O1O1O2N_VY5"}, "label": "the bus that is closest to the building"}]}
{"id": 510617, "image": "COCO_train2014_000000510617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow , orange , and white vehicle with headlights\"."}], "task": "refering", "answer_template": "The \"a yellow , orange , and white vehicle with headlights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 307], "bbox": [0.07910937500000001, 0.21498829039812645, 0.9045156249999999, 0.8749414519906323], "iscrowd": 0, "area": 95507.62079999998, "rle": {"size": [427, 640], "counts": "Zoe01Z=1O000O1M4G8H8G9G:A>I70001O0O10001N1000000O2O000O101O000O101O0O100O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1N3N1N2O1O2M2O1N2O2N1O1O101N1O100O2N100O100O2N100O1O2O00000O2O0000001N1000001O0O10001O0O1O1N3M2N2M3N2N3N1N2N2O2M2N2O1N3M2O1N2N3N1M3L4L5K4L4L4O2N1O1O1O1O2N100O1O2N1O10001O0O10001N10001N3N2N1N2O1O1O1N2O1O2M2O1O1N2O1O1N2O1O2N1N2O1O1N2O1O1N2O1O2M2OO1000O10000O10000O1000O01000000O10000O10000O10O10O1000000O10000O1000O010000O1000000O01O1O001O1O1O2N1O2N1O2N1O2N1O2N1O1M3N2N2M4M2N2M3N2WOSJnIS6R6c001O01O0000000010O000001O0001O0001O00000010O000001O0000010O0000001O0001O01O0000001O01O0001O00000010O01O001O0001O0O100000000O2O0000000O10001O000O100000001N100000000O101O00000O1000001O0O100000000O2O0000000O101O0000000O10001O000O100000001N100000000O101O00000O1000001O0O100000000O2O0000000O101O00000O1000001O0O100000000O2O0000000O10001O000O100000001N100000000O101O00000O1000001O0O100000001N100000000O101O00000O1000001O000O100000001O01O1O1O2N1O100O1O2N1O1O1O101NO100O10000O2O0O10000O101N100H8E;O1O101O0000000O10001O00001O1N101O1O1O0O2O1O001O1O1N101O1O1O1N4M2M4G9lMT2B>C<CQYi0"}, "label": "a yellow , orange , and white vehicle with headlights"}]}
{"id": 510617, "image": "COCO_train2014_000000510617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter old style school bus with a hose in front of it on the ground\"."}], "task": "refering", "answer_template": "The \"the shorter old style school bus with a hose in front of it on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 307], "bbox": [0.07910937500000001, 0.21498829039812645, 0.9045156249999999, 0.8749414519906323], "iscrowd": 0, "area": 95507.62079999998, "rle": {"size": [427, 640], "counts": "Zoe01Z=1O000O1M4G8H8G9G:A>I70001O0O10001N1000000O2O000O101O000O101O0O100O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1N3N1N2O1O2M2O1N2O2N1O1O101N1O100O2N100O100O2N100O1O2O00000O2O0000001N1000001O0O10001O0O1O1N3M2N2M3N2N3N1N2N2O2M2N2O1N3M2O1N2N3N1M3L4L5K4L4L4O2N1O1O1O1O2N100O1O2N1O10001O0O10001N10001N3N2N1N2O1O1O1N2O1O2M2O1O1N2O1O1N2O1O2N1N2O1O1N2O1O1N2O1O2M2OO1000O10000O10000O1000O01000000O10000O10000O10O10O1000000O10000O1000O010000O1000000O01O1O001O1O1O2N1O2N1O2N1O2N1O2N1O1M3N2N2M4M2N2M3N2WOSJnIS6R6c001O01O0000000010O000001O0001O0001O00000010O000001O0000010O0000001O0001O01O0000001O01O0001O00000010O01O001O0001O0O100000000O2O0000000O10001O000O100000001N100000000O101O00000O1000001O0O100000000O2O0000000O101O0000000O10001O000O100000001N100000000O101O00000O1000001O0O100000000O2O0000000O101O00000O1000001O0O100000000O2O0000000O10001O000O100000001N100000000O101O00000O1000001O0O100000001N100000000O101O00000O1000001O000O100000001O01O1O1O2N1O100O1O2N1O1O1O101NO100O10000O2O0O10000O101N100H8E;O1O101O0000000O10001O00001O1N101O1O1O0O2O1O001O1O1N101O1O1O1N4M2M4G9lMT2B>C<CQYi0"}, "label": "the shorter old style school bus with a hose in front of it on the ground"}]}
{"id": 469658, "image": "COCO_train2014_000000469658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small truck being lifted by a heavier truck\"."}], "task": "refering", "answer_template": "The \"a small truck being lifted by a heavier truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 201, 202], "bbox": [0.550390625, 0.1461968085106383, 0.86321875, 0.633563829787234], "iscrowd": 0, "area": 24707.856350000013, "rle": {"size": [376, 640], "counts": "W[Q4l0k:3N1N3M2O:Ee0[O:F2J5K5K5J6O1O1O2N100O1O1O1O1O101O00000001O000001O0N2N2M3000001O000002N2N06K5I7H8I7H7J1001O00002N1O000000N2O1N2O1001O00001O1O1O0000O1O1O1N2O1001O0000002N1O1O000000000000O10000000000000000000000000000000000000000000000O100gIWKc5i4QJcKo5S50000000O10000000000000000O100000000000O1000O10000000000000000O6K9G:F:F9G:F:F9G:E;F9G:F:F9G00000O1000O100000000000O10000000000000000O10000000000000000O100000;Ea0_OZSP1"}, "label": "a small truck being lifted by a heavier truck"}]}
{"id": 469658, "image": "COCO_train2014_000000469658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck on the bed of another truck\"."}], "task": "refering", "answer_template": "The \"a truck on the bed of another truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 201, 202], "bbox": [0.550390625, 0.1461968085106383, 0.86321875, 0.633563829787234], "iscrowd": 0, "area": 24707.856350000013, "rle": {"size": [376, 640], "counts": "W[Q4l0k:3N1N3M2O:Ee0[O:F2J5K5K5J6O1O1O2N100O1O1O1O1O101O00000001O000001O0N2N2M3000001O000002N2N06K5I7H8I7H7J1001O00002N1O000000N2O1N2O1001O00001O1O1O0000O1O1O1N2O1001O0000002N1O1O000000000000O10000000000000000000000000000000000000000000000O100gIWKc5i4QJcKo5S50000000O10000000000000000O100000000000O1000O10000000000000000O6K9G:F:F9G:F:F9G:E;F9G:F:F9G00000O1000O100000000000O10000000000000000O10000000000000000O100000;Ea0_OZSP1"}, "label": "a truck on the bed of another truck"}]}
{"id": 469658, "image": "COCO_train2014_000000469658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white mercedes flatbed semi truck carrying another white mercedes semi truck cab\"."}], "task": "refering", "answer_template": "The \"a white mercedes flatbed semi truck carrying another white mercedes semi truck cab\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.14070312499999998, 0.11851063829787235, 0.896484375, 0.9160372340425532], "iscrowd": 0, "area": 86347.26985000001, "rle": {"size": [376, 640], "counts": "QXQ1f3\\4f300O1O1O1O1O1O2N1O1O1O1O100O1O1O1O1O2N1O1O1O1O1O100O1O1O1O1O2O000O10000000000O10001O001O1N2O1O1O1O1O1N2O1O1O00O1O001O1O1O1O1O010O10000000O10000000000O10000000000O10000000000O10000000000O10000000000O101O0000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000001O001O00001O001O00001O001O00001O00001O001O00001O0000O100O2O0O100O100O100O100O100O100O1O100O1O1N2O1N2N2O1N2N2N2O1N2N2O100O106J9G8G9H9G8G9H9G8H8G6K6J5K6I7J5K6I6K5K000O2O00000O1000000O10000O1000000O1000000O101O0O1000000O1000000O100000014K6J5K6J6K4K4L0001O000000001O0000001O000000001O0000001O0000000000O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1O1O10000000000O10000000000O010000000O1jNYF;g9VOhFj0S:00O1000O1000O10000000O0100000000O10O100000O100000O10O1000000O010000000O1000O1000O10000000O0100000000O\\EVO\\:j0dEWOZ:j0eEWO[:i0dEYO[:g0dEZO\\:e0dE]O[:c0dE^O\\:b0cE_O]:a0bEA\\:?dEB\\:>cED\\:l000000000O01000000O1000O10O1000000O01000N2M3N2M3N1N4N2O0O2O1O1N2O1O1N101O1N2O1O2M4M4Lfog0"}, "label": "a white mercedes flatbed semi truck carrying another white mercedes semi truck cab"}]}
{"id": 469658, "image": "COCO_train2014_000000469658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"heaver bros . ltd truck\"."}], "task": "refering", "answer_template": "The \"heaver bros . ltd truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.14070312499999998, 0.11851063829787235, 0.896484375, 0.9160372340425532], "iscrowd": 0, "area": 86347.26985000001, "rle": {"size": [376, 640], "counts": "QXQ1f3\\4f300O1O1O1O1O1O2N1O1O1O1O100O1O1O1O1O2N1O1O1O1O1O100O1O1O1O1O2O000O10000000000O10001O001O1N2O1O1O1O1O1N2O1O1O00O1O001O1O1O1O1O010O10000000O10000000000O10000000000O10000000000O10000000000O10000000000O101O0000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000001O001O00001O001O00001O001O00001O00001O001O00001O0000O100O2O0O100O100O100O100O100O100O1O100O1O1N2O1N2N2O1N2N2N2O1N2N2O100O106J9G8G9H9G8G9H9G8H8G6K6J5K6I7J5K6I6K5K000O2O00000O1000000O10000O1000000O1000000O101O0O1000000O1000000O100000014K6J5K6J6K4K4L0001O000000001O0000001O000000001O0000001O0000000000O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1O1O10000000000O10000000000O010000000O1jNYF;g9VOhFj0S:00O1000O1000O10000000O0100000000O10O100000O100000O10O1000000O010000000O1000O1000O10000000O0100000000O\\EVO\\:j0dEWOZ:j0eEWO[:i0dEYO[:g0dEZO\\:e0dE]O[:c0dE^O\\:b0cE_O]:a0bEA\\:?dEB\\:>cED\\:l000000000O01000000O1000O10O1000000O01000N2M3N2M3N1N4N2O0O2O1O1N2O1O1N101O1N2O1O2M4M4Lfog0"}, "label": "heaver bros . ltd truck"}]}
{"id": 576157, "image": "COCO_train2014_000000576157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large gray elephant with 2 young elephants next to him / her\"."}], "task": "refering", "answer_template": "The \"a large gray elephant with 2 young elephants next to him / her\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 233, 234, 235, 236, 237, 238, 239, 254, 256, 257, 258, 259, 261, 262, 276, 277, 279, 280, 281, 282, 284, 285, 303, 304, 305], "bbox": [0.023875, 0.303364705882353, 0.516328125, 0.8831529411764707], "iscrowd": 0, "area": 46290.09644999997, "rle": {"size": [425, 640], "counts": "R_63_3NR6<hIFR6c0hI_OT6j0eIYOU6P1fIQOU6Y1dIjNW6^1cIeNY6c1aI^NT58WIb1_1YNX5T3aJnL]5X3^JiLa5\\3YJeLg5_3TJcLk5a3QJ_Lo5e3lI\\LS6j3gIXLX6l3cIUL]6o3_IQL`6T4[ImKe6W4VIkKh6[500O2N100010iMhI\\MX6Q2^JlMc5n1eJoM\\5l1jJPNY5m1lJkMY5U2jJdM[5[2gJ^M_5a2dJVMc5h2aJnLg5Q3X2f0[O00N2N2N2N2N3N10000000000000001O000000O1000000O100000O1O001O1O1O1O00100O1O1O001O1O1O1O1;D<E<D<C<D;F6I7I7J6I7I7J6I7kIVI^5Q7[JPId5W7VJiHj5[7O1O100000000000000000000000000000000000O10000fMZ2H83M3M3M4K4M3M3M3M3M3M3M4K4M3M3M3SJaIj4b6RK_In4d6nJ]IR5f6jJ[IV5h6fJYIY5k6cJVI]5m6_JTIa5o6\\JQId5Z70000000000O10O10000000000000O10ZM^JeMb5[2eJ^M[5b2mJVMS5j2TKoLl4Q3\\KgLd4Y3cK`L]4`3kKXLU4h3RLQLm3Q4ZLgKf3Y4bL^K_3b4hLWKX3i4d20000000O101O000000000001O0001O000001O00000001O01O00000000001O01O00000001O000001O00010O0000001O01O01O00000010O02N3XIPKQ5R5hJSKX5o4bJUK_5m4YJXKg5j4RJ\\Km5g4lI]KT6e4eI`K[6g510O0000001O01_MRJeMn5P2bJkM^5S2jJhMU5U2SKfMm4W2\\KcMe4Z2bKaM^4\\2kK_MT4^2TL]Ml3a2\\LYMd3f2bLUM_3i2hLRMX3m2mLnLS3Q3d3O0O2O001N2O0O2O0O2O0O2O0O2O1N101N101M2O2N2N1O2N2M3N2N2N2N2N2O1N2O001N2O1N2O1O1N2O1N2O1O1N2O1N2O2N2M4M2M4M2N2M4M2N3L3N3Lf^P4"}, "label": "a large gray elephant with 2 young elephants next to him / her"}]}
{"id": 576157, "image": "COCO_train2014_000000576157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the largest elephant is eating grass\"."}], "task": "refering", "answer_template": "The \"the largest elephant is eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 233, 234, 235, 236, 237, 238, 239, 254, 256, 257, 258, 259, 261, 262, 276, 277, 279, 280, 281, 282, 284, 285, 303, 304, 305], "bbox": [0.023875, 0.303364705882353, 0.516328125, 0.8831529411764707], "iscrowd": 0, "area": 46290.09644999997, "rle": {"size": [425, 640], "counts": "R_63_3NR6<hIFR6c0hI_OT6j0eIYOU6P1fIQOU6Y1dIjNW6^1cIeNY6c1aI^NT58WIb1_1YNX5T3aJnL]5X3^JiLa5\\3YJeLg5_3TJcLk5a3QJ_Lo5e3lI\\LS6j3gIXLX6l3cIUL]6o3_IQL`6T4[ImKe6W4VIkKh6[500O2N100010iMhI\\MX6Q2^JlMc5n1eJoM\\5l1jJPNY5m1lJkMY5U2jJdM[5[2gJ^M_5a2dJVMc5h2aJnLg5Q3X2f0[O00N2N2N2N2N3N10000000000000001O000000O1000000O100000O1O001O1O1O1O00100O1O1O001O1O1O1O1;D<E<D<C<D;F6I7I7J6I7I7J6I7kIVI^5Q7[JPId5W7VJiHj5[7O1O100000000000000000000000000000000000O10000fMZ2H83M3M3M4K4M3M3M3M3M3M3M4K4M3M3M3SJaIj4b6RK_In4d6nJ]IR5f6jJ[IV5h6fJYIY5k6cJVI]5m6_JTIa5o6\\JQId5Z70000000000O10O10000000000000O10ZM^JeMb5[2eJ^M[5b2mJVMS5j2TKoLl4Q3\\KgLd4Y3cK`L]4`3kKXLU4h3RLQLm3Q4ZLgKf3Y4bL^K_3b4hLWKX3i4d20000000O101O000000000001O0001O000001O00000001O01O00000000001O01O00000001O000001O00010O0000001O01O01O00000010O02N3XIPKQ5R5hJSKX5o4bJUK_5m4YJXKg5j4RJ\\Km5g4lI]KT6e4eI`K[6g510O0000001O01_MRJeMn5P2bJkM^5S2jJhMU5U2SKfMm4W2\\KcMe4Z2bKaM^4\\2kK_MT4^2TL]Ml3a2\\LYMd3f2bLUM_3i2hLRMX3m2mLnLS3Q3d3O0O2O001N2O0O2O0O2O0O2O0O2O1N101N101M2O2N2N1O2N2M3N2N2N2N2N2O1N2O001N2O1N2O1O1N2O1N2O1O1N2O1N2O2N2M4M2M4M2N2M4M2N3L3N3Lf^P4"}, "label": "the largest elephant is eating grass"}]}
{"id": 576157, "image": "COCO_train2014_000000576157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant walking ot the right with other elephants next to it\"."}], "task": "refering", "answer_template": "The \"baby elephant walking ot the right with other elephants next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316], "bbox": [0.530234375, 0.47421176470588233, 0.7616250000000001, 0.9014352941176471], "iscrowd": 0, "area": 18734.424349999998, "rle": {"size": [425, 640], "counts": "h[]4:k<:F:F9G:G8H6L3M3M4L3L4M3M4hEgMY9Y3K4L8H8VGjKR8g4O100O100O2N100O1O100O1O100O2N100O1O100O1O100O100O1O100O100O100O01O10O010O01O10O01N2M3N2M4N101N1O2O0O2VOi0O1O2O0O2N100O2N100O2N101N2N2O1N2N21O5K6J5oFXLV8n3]GZLb8X4O000000O10O000O2O0000000O10000000eNeHhL[7f2YIXMf6h2^IUMb6j2P2O1N1N3M3O02O2N1iEcM^9^3G7I0O2O0RGQLa8o3]GUL`8m3\\GULe8k3YGWLg8U41O2N1O1hNUGQNk8_1jG[NW8a0[GSOg09o7=hGmN`0d0i7:[IDf6:]IDV7HmH5[7CgH<U:N2N2N2N2M4M[To1"}, "label": "baby elephant walking ot the right with other elephants next to it"}]}
{"id": 576157, "image": "COCO_train2014_000000576157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant wanders away from its sibling and mother\"."}], "task": "refering", "answer_template": "The \"a baby elephant wanders away from its sibling and mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316], "bbox": [0.530234375, 0.47421176470588233, 0.7616250000000001, 0.9014352941176471], "iscrowd": 0, "area": 18734.424349999998, "rle": {"size": [425, 640], "counts": "h[]4:k<:F:F9G:G8H6L3M3M4L3L4M3M4hEgMY9Y3K4L8H8VGjKR8g4O100O100O2N100O1O100O1O100O2N100O1O100O1O100O100O1O100O100O100O01O10O010O01O10O01N2M3N2M4N101N1O2O0O2VOi0O1O2O0O2N100O2N100O2N101N2N2O1N2N21O5K6J5oFXLV8n3]GZLb8X4O000000O10O000O2O0000000O10000000eNeHhL[7f2YIXMf6h2^IUMb6j2P2O1N1N3M3O02O2N1iEcM^9^3G7I0O2O0RGQLa8o3]GUL`8m3\\GULe8k3YGWLg8U41O2N1O1hNUGQNk8_1jG[NW8a0[GSOg09o7=hGmN`0d0i7:[IDf6:]IDV7HmH5[7CgH<U:N2N2N2N2M4M[To1"}, "label": "a baby elephant wanders away from its sibling and mother"}]}
{"id": 576157, "image": "COCO_train2014_000000576157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant closest to large elephant\"."}], "task": "refering", "answer_template": "The \"baby elephant closest to large elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 309, 310], "bbox": [0.40073437500000003, 0.49023529411764705, 0.581125, 0.9071764705882351], "iscrowd": 0, "area": 12219.85995, "rle": {"size": [425, 640], "counts": "\\`Z3h0_<6J6J7J5J7I6J7J6I7I7I7J6K5J6K4K5K5L3L100O2O0O1O2O0O101N_ORGiLm8W3UGhLj8Y3VGgLj8Y3WGfLh8[3XGdLh8]3YGbLf8_3[G_Le8b3[G^Le8b3\\G\\Ld8e3\\GZLd8g3]GWLc8j3=1O0O2N<E5K6K2M0100O1J61N2O1O1O8G3N1O000O2O000O2O0O1O100O1O10O01O100O01H8H8F:F:I7I7J6N2O100mMdF9\\9BPG7P9C\\G7c8DjG5V8IRH1n7MYHNg71]HLc73aHJ_74fHJY75jHJV74nHJQ74TIKj64ZIJe65_I@i6?Q3O2O0O100O2N101N1O2O0O2N101N1Oo]_3"}, "label": "baby elephant closest to large elephant"}]}
{"id": 576157, "image": "COCO_train2014_000000576157.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small elephant behind another small elephant\"."}], "task": "refering", "answer_template": "The \"a small elephant behind another small elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 309, 310], "bbox": [0.40073437500000003, 0.49023529411764705, 0.581125, 0.9071764705882351], "iscrowd": 0, "area": 12219.85995, "rle": {"size": [425, 640], "counts": "\\`Z3h0_<6J6J7J5J7I6J7J6I7I7I7J6K5J6K4K5K5L3L100O2O0O1O2O0O101N_ORGiLm8W3UGhLj8Y3VGgLj8Y3WGfLh8[3XGdLh8]3YGbLf8_3[G_Le8b3[G^Le8b3\\G\\Ld8e3\\GZLd8g3]GWLc8j3=1O0O2N<E5K6K2M0100O1J61N2O1O1O8G3N1O000O2O000O2O0O1O100O1O10O01O100O01H8H8F:F:I7I7J6N2O100mMdF9\\9BPG7P9C\\G7c8DjG5V8IRH1n7MYHNg71]HLc73aHJ_74fHJY75jHJV74nHJQ74TIKj64ZIJe65_I@i6?Q3O2O0O100O2N101N1O2O0O2N101N1Oo]_3"}, "label": "a small elephant behind another small elephant"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass of water next to the saucer with the cup on it\"."}], "task": "refering", "answer_template": "The \"the glass of water next to the saucer with the cup on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 48, 49, 50, 71, 72, 73, 94, 95, 96, 117, 118, 119, 141, 142, 143, 164, 165], "bbox": [0.078125, 0.07162500000000001, 0.224609375, 0.4231875], "iscrowd": 0, "area": 10456.25, "rle": {"size": [480, 640], "counts": "]_g02m>4L5L3L4L5K4L4L5L3M3M4L3L4M4L4L5K5J7J5K5K5K5K5J7J5K5K5K3L4M3M3M2N3M3M3M3M3M2N3M3M3M3M00000000000000000000000000000001O00001O00001O00001O003M7I7I6J7I7I6J7I7I6J7I7I6J7I7I6J7I7I6I8H8H7I8H8HnYX7"}, "label": "the glass of water next to the saucer with the cup on it"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"water in a tall , clear glass\"."}], "task": "refering", "answer_template": "The \"water in a tall , clear glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 48, 49, 50, 71, 72, 73, 94, 95, 96, 117, 118, 119, 141, 142, 143, 164, 165], "bbox": [0.078125, 0.07162500000000001, 0.224609375, 0.4231875], "iscrowd": 0, "area": 10456.25, "rle": {"size": [480, 640], "counts": "]_g02m>4L5L3L4L5K4L4L5L3M3M4L3L4M4L4L5K5J7J5K5K5K5K5J7J5K5K5K3L4M3M3M2N3M3M3M3M3M2N3M3M3M3M00000000000000000000000000000001O00001O00001O00001O003M7I7I6J7I7I6J7I7I6J7I7I6J7I7I6J7I7I6I8H8H7I8H8HnYX7"}, "label": "water in a tall , clear glass"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of noodle soup\"."}], "task": "refering", "answer_template": "The \"a bowl of noodle soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.2983125, 0.39775, 0.7668593749999999, 0.8112291666666667], "iscrowd": 0, "area": 47408.58045, "rle": {"size": [480, 640], "counts": "eji26a>;E;E<D;F:M3M4L3M3M3L4M4L3M3M1O2N2M3N2N1O2N2N2M3L4M2N3L4nNULkFn3S9WLgFl3X9XLbFj3^9ZL[Fj3d9YLWFj3h9h0O1O1O001O1O1O1O1O001N2O1O1O001O1O1O1N2O001O1O1O1O0O2O1O1O1O1O001N2O00O1O100O1O10O0100000000000O101O000000000O1000000000000O1000001O000000000001O0000000000000001O000001O00000000000001O00000001O00000000000001O00000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O0000000000000000000000000000000000000000O100000000000000000000O1000O10000000000O1O1O001O100O1O1O1O1O1O100O1O001O1O1O2O1N1O2N1O2N1O2O1N1O2N3M3M3M2O2M3M3M3M2N3M3N2M2N3M3M3M2N3N2M3M3M2Nn0RO5K2O2M2N2N3M2N3M2O1N3M2N2N3M2N2O2M2N3M3M3M3M3N1N3M3M3M3MVPV2"}, "label": "a bowl of noodle soup"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bowl in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the white bowl in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.2983125, 0.39775, 0.7668593749999999, 0.8112291666666667], "iscrowd": 0, "area": 47408.58045, "rle": {"size": [480, 640], "counts": "eji26a>;E;E<D;F:M3M4L3M3M3L4M4L3M3M1O2N2M3N2N1O2N2N2M3L4M2N3L4nNULkFn3S9WLgFl3X9XLbFj3^9ZL[Fj3d9YLWFj3h9h0O1O1O001O1O1O1O1O001N2O1O1O001O1O1O1N2O001O1O1O1O0O2O1O1O1O1O001N2O00O1O100O1O10O0100000000000O101O000000000O1000000000000O1000001O000000000001O0000000000000001O000001O00000000000001O00000001O00000000000001O00000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O0000000000000000000000000000000000000000O100000000000000000000O1000O10000000000O1O1O001O100O1O1O1O1O1O100O1O001O1O1O2O1N1O2N1O2N1O2O1N1O2N3M3M3M2O2M3M3M3M2N3M3N2M2N3M3M3M2N3N2M3M3M2Nn0RO5K2O2M2N2N3M2N3M2O1N3M2N2N3M2N2O2M2N3M3M3M3M3N1N3M3M3M3MVPV2"}, "label": "the white bowl in the right hand picture"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lone fork laying on a gray - and - black checkered place mat\"."}], "task": "refering", "answer_template": "The \"a lone fork laying on a gray - and - black checkered place mat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 210, 211, 233, 234, 256, 278, 301, 323, 324, 346, 347], "bbox": [0.043359375, 0.5030625, 0.20364062499999996, 0.9073749999999999], "iscrowd": 0, "area": 3270.2482999999997, "rle": {"size": [480, 640], "counts": "Xa=8h>1N1O2N1O101N101M2N101N2O1N2N2O1N2N2N2O1N1O2N101N1O1O2O0O1O2N1O1O1O1O001O1O00K5J6K5N2O1N2N2O1N2O1N2O1N2N2O1N2O1O1000000000O100O100O1N2M3N2O1O1O1O3M2M4M2M4MUO_B<^=BgB?V=^OPCa0m<]OXCc0f<ZO_Cg0^<WOgCh0W<UOiCQ1i<9F;F[OTCIn<0WC1j<H[C8g<BXCd0i<UOVCQ1W=6I_OeBJV=2SCNn<H\\C7f<^OdCa0^=OD<]Oc0O1i0WOfi^7"}, "label": "a lone fork laying on a gray - and - black checkered place mat"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long fork on a place setting wit ha checkered placemat\"."}], "task": "refering", "answer_template": "The \"a long fork on a place setting wit ha checkered placemat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 210, 211, 233, 234, 256, 278, 301, 323, 324, 346, 347], "bbox": [0.043359375, 0.5030625, 0.20364062499999996, 0.9073749999999999], "iscrowd": 0, "area": 3270.2482999999997, "rle": {"size": [480, 640], "counts": "Xa=8h>1N1O2N1O101N101M2N101N2O1N2N2O1N2N2N2O1N1O2N101N1O1O2O0O1O2N1O1O1O1O001O1O00K5J6K5N2O1N2N2O1N2O1N2O1N2N2O1N2O1O1000000000O100O100O1N2M3N2O1O1O1O3M2M4M2M4MUO_B<^=BgB?V=^OPCa0m<]OXCc0f<ZO_Cg0^<WOgCh0W<UOiCQ1i<9F;F[OTCIn<0WC1j<H[C8g<BXCd0i<UOVCQ1W=6I_OeBJV=2SCNn<H\\C7f<^OdCa0^=OD<]Oc0O1i0WOfi^7"}, "label": "a long fork on a place setting wit ha checkered placemat"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver knife next to the bowl\"."}], "task": "refering", "answer_template": "The \"a silver knife next to the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 247, 248, 271, 272, 295, 296, 319, 320, 343, 344, 366, 367, 390], "bbox": [0.7621249999999999, 0.5698958333333334, 0.999953125, 0.9953541666666667], "iscrowd": 0, "area": 4116.101699999997, "rle": {"size": [480, 640], "counts": "cPU78h>4K4M2N2N2N2N2O1N101N101O00010O0010O00010O01O01O01O010O000010O0001O001O0001O0000000000001O000000000O2O000O1O101N100O1O101N100O1000001O000000000000000001O00000001O000001O00000001O00000001O000001O0000000001O000010O01O00001O001O01O01O001O001O001O1O001O001O0010O01O001O001O1O001O1O0010O01O1O001O010O1O001O1ZB"}, "label": "a silver knife next to the bowl"}]}
{"id": 379564, "image": "COCO_train2014_000000379564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife next to the spoon\"."}], "task": "refering", "answer_template": "The \"the knife next to the spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 247, 248, 271, 272, 295, 296, 319, 320, 343, 344, 366, 367, 390], "bbox": [0.7621249999999999, 0.5698958333333334, 0.999953125, 0.9953541666666667], "iscrowd": 0, "area": 4116.101699999997, "rle": {"size": [480, 640], "counts": "cPU78h>4K4M2N2N2N2N2O1N101N101O00010O0010O00010O01O01O01O010O000010O0001O001O0001O0000000000001O000000000O2O000O1O101N100O1O101N100O1000001O000000000000000001O00000001O000001O00000001O00000001O000001O0000000001O000010O01O00001O001O01O01O001O001O001O1O001O001O0010O01O001O001O1O001O1O0010O01O1O001O010O1O001O1ZB"}, "label": "the knife next to the spoon"}]}
{"id": 420524, "image": "COCO_train2014_000000420524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza cutter slicing a small pizza\"."}], "task": "refering", "answer_template": "The \"pizza cutter slicing a small pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290], "bbox": [0.1439375, 0.3479391100702576, 0.8725937500000001, 0.8153629976580795], "iscrowd": 0, "area": 71368.23400000001, "rle": {"size": [427, 640], "counts": "jcV14j<=D=B>B=J7N2N101N2N101N2N101N2N1O2O1N1O2O1N1O2O0O2N2O0O2N2N101N2N101N2N101N2N1O2O1N1O2O1N1O2O1N1O2O1N1O2N2O0O2N2O0O2N101N2N101N2N1O2O1O010O1O001O10O01O1O010O1O00100O001O10O01O1O00100O001O010O0000010O00000001O0000010O0000001OO101O000000000O10001O000000000O10001O0000000O10001O0000000O1000001O0000000O1000001O00000O1000001O00000O100000001O00000O100000001O000O100000001O000O10000000001O000O100000001O000O10000000001O0O1000000000001N1000000000001O0O1000000000001N1000000000000O2O0000000000001N1000000000000O100000000000O0100000000000000O100000O100000O1000000000000O10O1000000000000000001O000O101O000000001O0000001O0O10001O0000001O000000001N1000001O0000001O000000001N1000001O0000001O0000000O2O0000001O0000001O00000O101O0000001O0000001O0O100O2O0O101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O1N101N101N101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O0O2O1N102N1N3N2M3N1N3N2M3N1N3M3K5K4L5K5K5K4L5L4K5K5K4L5K5K5K^nQ1"}, "label": "pizza cutter slicing a small pizza"}]}
{"id": 420524, "image": "COCO_train2014_000000420524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza is being cut by a pizza wheel\"."}], "task": "refering", "answer_template": "The \"a pizza is being cut by a pizza wheel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290], "bbox": [0.1439375, 0.3479391100702576, 0.8725937500000001, 0.8153629976580795], "iscrowd": 0, "area": 71368.23400000001, "rle": {"size": [427, 640], "counts": "jcV14j<=D=B>B=J7N2N101N2N101N2N101N2N1O2O1N1O2O1N1O2O0O2N2O0O2N2N101N2N101N2N101N2N1O2O1N1O2O1N1O2O1N1O2O1N1O2N2O0O2N2O0O2N101N2N101N2N1O2O1O010O1O001O10O01O1O010O1O00100O001O10O01O1O00100O001O010O0000010O00000001O0000010O0000001OO101O000000000O10001O000000000O10001O0000000O10001O0000000O1000001O0000000O1000001O00000O1000001O00000O100000001O00000O100000001O000O100000001O000O10000000001O000O100000001O000O10000000001O0O1000000000001N1000000000001O0O1000000000001N1000000000000O2O0000000000001N1000000000000O100000000000O0100000000000000O100000O100000O1000000000000O10O1000000000000000001O000O101O000000001O0000001O0O10001O0000001O000000001N1000001O0000001O000000001N1000001O0000001O0000000O2O0000001O0000001O00000O101O0000001O0000001O0O100O2O0O101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O1N101N101N101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O1N101N101O0O2O0O2O0O2O0O2O1N102N1N3N2M3N1N3N2M3N1N3M3K5K4L5K5K5K4L5L4K5K5K4L5K5K5K^nQ1"}, "label": "a pizza is being cut by a pizza wheel"}]}
{"id": 420524, "image": "COCO_train2014_000000420524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand that ' s holding the pizza cutter\"."}], "task": "refering", "answer_template": "The \"the hand that ' s holding the pizza cutter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 34, 35, 36, 37, 57, 58, 59], "bbox": [0.38082812499999996, 0.004637002341920374, 0.6626875, 0.20913348946135832], "iscrowd": 0, "area": 8278.801850000002, "rle": {"size": [427, 640], "counts": "hhU3`0h<4N2N2N2N2O1N2N2N001O001O001O00100O001O001O0100O010O10000O100O010O100O100O010O10000O100O0100000000000000000001O0000000001O000001O0000000000000000001O000000000000003M:F:F:F5LO01O00001O001O00001O00001O001O00001O00001O00001O010O00001O00001O001O0000O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O2M2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2OSRj2"}, "label": "the hand that ' s holding the pizza cutter"}]}
{"id": 420524, "image": "COCO_train2014_000000420524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fingers holding a pizza cutter\"."}], "task": "refering", "answer_template": "The \"fingers holding a pizza cutter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 34, 35, 36, 37, 57, 58, 59], "bbox": [0.38082812499999996, 0.004637002341920374, 0.6626875, 0.20913348946135832], "iscrowd": 0, "area": 8278.801850000002, "rle": {"size": [427, 640], "counts": "hhU3`0h<4N2N2N2N2O1N2N2N001O001O001O00100O001O001O0100O010O10000O100O010O100O100O010O10000O100O0100000000000000000001O0000000001O000001O0000000000000000001O000000000000003M:F:F:F5LO01O00001O001O00001O00001O001O00001O00001O00001O010O00001O00001O001O0000O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O2M2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2OSRj2"}, "label": "fingers holding a pizza cutter"}]}
{"id": 420524, "image": "COCO_train2014_000000420524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with pale legs\"."}], "task": "refering", "answer_template": "The \"a person with pale legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142], "bbox": [0.0, 0.0011241217798594848, 0.29453125, 0.4718501170960187], "iscrowd": 0, "area": 32668.32325, "rle": {"size": [427, 640], "counts": "R1Q1o;T2bM^2WNd0\\O00000000O100000000O10000000000O100000000O100000000O1000O100000O100000000O10000000000O100000000O10000000000O100000000O1000O100000O100000001N10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000001O000000000001O0000000O100O1O1O100O1O100O1O100O1O1O100O100O1O100O1O100O100O1O100001O0000001O000011N2N2N2O1N2N2N2O1N2N2O1N2N2N9H>A`0@7I3N3L3M4M2kNU1D=E:F;Elnk5"}, "label": "a person with pale legs"}]}
{"id": 420524, "image": "COCO_train2014_000000420524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in the left side of the image\"."}], "task": "refering", "answer_template": "The \"man in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142], "bbox": [0.0, 0.0011241217798594848, 0.29453125, 0.4718501170960187], "iscrowd": 0, "area": 32668.32325, "rle": {"size": [427, 640], "counts": "R1Q1o;T2bM^2WNd0\\O00000000O100000000O10000000000O100000000O100000000O1000O100000O100000000O10000000000O100000000O10000000000O100000000O1000O100000O100000001N10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000001O000000000001O0000000O100O1O1O100O1O100O1O100O1O1O100O100O1O100O1O100O100O1O100001O0000001O000011N2N2N2O1N2N2N2O1N2N2O1N2N2N9H>A`0@7I3N3L3M4M2kNU1D=E:F;Elnk5"}, "label": "man in the left side of the image"}]}
{"id": 551607, "image": "COCO_train2014_000000551607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suitcase under the white bag\"."}], "task": "refering", "answer_template": "The \"the suitcase under the white bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 259, 260], "bbox": [0.205625, 0.31197916666666664, 0.431453125, 0.6625416666666666], "iscrowd": 0, "area": 18787.87695, "rle": {"size": [480, 640], "counts": "]Sn1P1o<T1lNT1C;N2O1N2N2O1N1O2O1N2N2O0O2N2N2O1N1010000000O100000O1000O100000000O10000O10O0100O100O100O10O001O1O0O2O001N101O0O2O001N101O001N0100O010O10O0100O010O10O0100O10O0100O010O10O0100O010O10O0100O10O0100O010O10O0100O010O100O010O10O0100O010O10O0100O010O100O010O101N2O1N2O1N2O1N2`MPFNQ:LZGmNh8m0i2K5J6J6K5Jo^Z5"}, "label": "the suitcase under the white bag"}]}
{"id": 551607, "image": "COCO_train2014_000000551607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traveller bag holding white cover in its head\"."}], "task": "refering", "answer_template": "The \"traveller bag holding white cover in its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 259, 260], "bbox": [0.205625, 0.31197916666666664, 0.431453125, 0.6625416666666666], "iscrowd": 0, "area": 18787.87695, "rle": {"size": [480, 640], "counts": "]Sn1P1o<T1lNT1C;N2O1N2N2O1N1O2O1N2N2O0O2N2N2O1N1010000000O100000O1000O100000000O10000O10O0100O100O100O10O001O1O0O2O001N101O0O2O001N101O001N0100O010O10O0100O010O10O0100O10O0100O010O10O0100O010O10O0100O10O0100O010O10O0100O010O100O010O10O0100O010O10O0100O010O100O010O101N2O1N2O1N2O1N2`MPFNQ:LZGmNh8m0i2K5J6J6K5Jo^Z5"}, "label": "traveller bag holding white cover in its head"}]}
{"id": 551607, "image": "COCO_train2014_000000551607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue luggage with wheels\"."}], "task": "refering", "answer_template": "The \"blue luggage with wheels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 201, 202], "bbox": [0.653453125, 0.12870833333333334, 0.8476874999999999, 0.4983125], "iscrowd": 0, "area": 17298.768249999997, "rle": {"size": [480, 640], "counts": "TTT68[>>G8H8H9D;^Oc0B=G9H9F9H801N1O2N1O1O2N100O2N1O2N1O101N1O1O2N1O1O2O0O101O0O101O00001O0O101O001O00001N10001O00001O00010O00001O01O01O00010O000010O01O0000101N2N1O2O0ON2O1O2N1O1O1O1O2N1N200O2N1O100O2N1O101N1O101N1O1O2O0O1O2O0O2N1O101N1O2O2L4D;E<D;F;Db0^Oh0XOi0WO\\[]1"}, "label": "blue luggage with wheels"}]}
{"id": 551607, "image": "COCO_train2014_000000551607.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue luggage paralel to the manholes\"."}], "task": "refering", "answer_template": "The \"a blue luggage paralel to the manholes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 201, 202], "bbox": [0.653453125, 0.12870833333333334, 0.8476874999999999, 0.4983125], "iscrowd": 0, "area": 17298.768249999997, "rle": {"size": [480, 640], "counts": "TTT68[>>G8H8H9D;^Oc0B=G9H9F9H801N1O2N1O1O2N100O2N1O2N1O101N1O1O2N1O1O2O0O101O0O101O00001O0O101O001O00001N10001O00001O00010O00001O01O01O00010O000010O01O0000101N2N1O2O0ON2O1O2N1O1O1O1O2N1N200O2N1O100O2N1O101N1O101N1O1O2O0O1O2O0O2N1O101N1O2O2L4D;E<D;F;Db0^Oh0XOi0WO\\[]1"}, "label": "a blue luggage paralel to the manholes"}]}
{"id": 10936, "image": "COCO_train2014_000000010936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver car parked behind a silver suv\"."}], "task": "refering", "answer_template": "The \"a silver car parked behind a silver suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123], "bbox": [0.0538125, 0.17029166666666665, 0.38509375, 0.325], "iscrowd": 0, "area": 9774.724499999997, "rle": {"size": [480, 640], "counts": "cQ`01m>?A3M3M2N2N2N3M2N2O101N10001O000O101O00000O2O0000001O0O10001O000O101O00001O0O10001O000O101O00000O2O0000001O0O102N1O2N2M2O2N1O1O0O10O100000000O10O10000000O1000000O1N2N1O2N200000000001O000O100000000000000001O00001O000010O01O00001O001O00001O001O00001O001O01O01O001O00001O001O00001O00001O001O00010O001O1O1O001O1O1O001O1O1O001O100O00100O1O00100O1O010O001N101O001O001O0O2O001O001O000O2O001O001O001O0O2O001O001O001O001O001O0M4L3Niah5"}, "label": "a silver car parked behind a silver suv"}]}
{"id": 10936, "image": "COCO_train2014_000000010936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver car on side of road\"."}], "task": "refering", "answer_template": "The \"silver car on side of road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123], "bbox": [0.0538125, 0.17029166666666665, 0.38509375, 0.325], "iscrowd": 0, "area": 9774.724499999997, "rle": {"size": [480, 640], "counts": "cQ`01m>?A3M3M2N2N2N3M2N2O101N10001O000O101O00000O2O0000001O0O10001O000O101O00001O0O10001O000O101O00000O2O0000001O0O102N1O2N2M2O2N1O1O0O10O100000000O10O10000000O1000000O1N2N1O2N200000000001O000O100000000000000001O00001O000010O01O00001O001O00001O001O00001O001O01O01O001O00001O001O00001O00001O001O00010O001O1O1O001O1O1O001O1O1O001O100O00100O1O00100O1O010O001N101O001O001O0O2O001O001O000O2O001O001O001O0O2O001O001O001O001O001O0M4L3Niah5"}, "label": "silver car on side of road"}]}
{"id": 10936, "image": "COCO_train2014_000000010936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suv on run\"."}], "task": "refering", "answer_template": "The \"the suv on run\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 158], "bbox": [0.6258125, 0.17258333333333334, 0.940328125, 0.376875], "iscrowd": 0, "area": 12717.356250000003, "rle": {"size": [480, 640], "counts": "RSl5a0_>1O0O2O0J7I6I8M2O2M2O2N1O2N1O2N1N2O1O1O1O1O1O1N2O1O1O1O1O1O010000000000000000000000O10001O0O1000000O10000O1000000O10000001O001O010O001O001O001O001O00001O0000000000000001O00000000000000000001O00001O00001O000010O0001O001O00001O000010O1O100O00100O100O1O10O01O1O1O1O1O1O001O1O000000001O000000001O000000001O0001O00010O1O1O1O1O100O1O1O1O10O000000001O000000001O000O1O1O1N3N1O3L4M3L4M3M3L4N2M2N3N2M3N2M3N7H9GUea0"}, "label": "the suv on run"}]}
{"id": 10936, "image": "COCO_train2014_000000010936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"suv parked by side of field\"."}], "task": "refering", "answer_template": "The \"suv parked by side of field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 158], "bbox": [0.6258125, 0.17258333333333334, 0.940328125, 0.376875], "iscrowd": 0, "area": 12717.356250000003, "rle": {"size": [480, 640], "counts": "RSl5a0_>1O0O2O0J7I6I8M2O2M2O2N1O2N1O2N1N2O1O1O1O1O1O1N2O1O1O1O1O1O010000000000000000000000O10001O0O1000000O10000O1000000O10000001O001O010O001O001O001O001O00001O0000000000000001O00000000000000000001O00001O00001O000010O0001O001O00001O000010O1O100O00100O100O1O10O01O1O1O1O1O1O001O1O000000001O000000001O000000001O0001O00010O1O1O1O1O100O1O1O1O10O000000001O000000001O000O1O1O1N3N1O3L4M3L4M3M3L4N2M2N3N2M3N2M3N7H9GUea0"}, "label": "suv parked by side of field"}]}
{"id": 166585, "image": "COCO_train2014_000000166585.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"standing dog\"."}], "task": "refering", "answer_template": "The \"standing dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 243, 244, 245, 246], "bbox": [0.430078125, 0.322125, 0.762078125, 0.6259583333333334], "iscrowd": 0, "area": 15218.443949999999, "rle": {"size": [480, 640], "counts": "aRQ4:c>8I5J4N1O1O10000O1000O10O10O19G2M2O00001N1000001N1000000O10000O10000O10O10O100000O10000c0\\O3N2N2N2N3M2N3M3L5L3XDcMg:[301O00000000010O01O001O1N101O001N1H9H7I8G8N30O1000O0100O10=B4M:E100O0100O001O0010O01O0O2O001PO`D]Nb;c1kDoMU;R2nDiMT;W2PEdMP;]2j00O01000O010000O010O10000O010O10000O01000O100O010O1=C00O00100O1O00100O1O001O100O001O1O00100O1O001O1O010N2O0O2JZCoMh<R2XClMh<V211OO00100O100O010O10O01O010O010O1O000^NTCP1m<POTCn0l<ROVCk0l<UOUCi0k<XOWCe0i<[OZCb0g<^OZC`0f<A\\C<d<D_C9b<G`C6`<JcC3]<NfCNZ<2iCKW<5mCGS<9QDCo;=Z10O1O1O1N2O1N2O1O0001N1N2N3I8J5MRQW2"}, "label": "standing dog"}]}
{"id": 166585, "image": "COCO_train2014_000000166585.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog looking forward from a boat\"."}], "task": "refering", "answer_template": "The \"a dog looking forward from a boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 243, 244, 245, 246], "bbox": [0.430078125, 0.322125, 0.762078125, 0.6259583333333334], "iscrowd": 0, "area": 15218.443949999999, "rle": {"size": [480, 640], "counts": "aRQ4:c>8I5J4N1O1O10000O1000O10O10O19G2M2O00001N1000001N1000000O10000O10000O10O10O100000O10000c0\\O3N2N2N2N3M2N3M3L5L3XDcMg:[301O00000000010O01O001O1N101O001N1H9H7I8G8N30O1000O0100O10=B4M:E100O0100O001O0010O01O0O2O001PO`D]Nb;c1kDoMU;R2nDiMT;W2PEdMP;]2j00O01000O010000O010O10000O010O10000O01000O100O010O1=C00O00100O1O00100O1O001O100O001O1O00100O1O001O1O010N2O0O2JZCoMh<R2XClMh<V211OO00100O100O010O10O01O010O010O1O000^NTCP1m<POTCn0l<ROVCk0l<UOUCi0k<XOWCe0i<[OZCb0g<^OZC`0f<A\\C<d<D_C9b<G`C6`<JcC3]<NfCNZ<2iCKW<5mCGS<9QDCo;=Z10O1O1O1N2O1N2O1O0001N1N2N3I8J5MRQW2"}, "label": "a dog looking forward from a boat"}]}
{"id": 576188, "image": "COCO_train2014_000000576188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the soccer player the number 8 on his shorts\"."}], "task": "refering", "answer_template": "The \"the soccer player the number 8 on his shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 58, 59, 60, 81, 82, 83, 84, 85, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 173, 174, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 244, 245, 246, 247, 268, 269, 270, 271, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 338, 339, 340], "bbox": [0.5137343750000001, 0.12326291079812206, 0.880046875, 0.9465258215962441], "iscrowd": 0, "area": 31305.351949999997, "rle": {"size": [426, 640], "counts": "S_Y42X=1O1O1O1O1O1O2N1O1O1O1O1O2N1O2N1O2N2N1O2N001bLYOXJf0b5A]J?]5HbJ8X50eJ2U54iJMW55eJM[54bJN]57^JKa5;WJGi5:TJHk5:SJHl59QJIo58oIIQ68lIKR67lIJT67iIKW65hIMW64fIMZ65dIJ^67`IIa67]IId68[IHf68XIHj68UIHl68oHKR76jHLX75bHO_71]H1e7OYH2i7MTH4o7KoG6R8KkG5X8MbG5`8L\\G4g8MTG5n8Q23N2M3N2N1O2N2M3N2N2N2N6J7I6J5K20O1000000O10000hH\\Mi3d2SLhMd3Y2XLlMg3T2ULQNi3P2SLUNk3l1QLYNn3g1nK^NP4c1lKaNS4`1iKeNV4\\1dKjNZ4X1aK]Om3d0oK_OR4b0iKBV4?gKDX4>dKE\\4<aKF^4;_KGb4:ZKIe49WKJh47VKKi47SKKn49kJJT5<dJG[5`0]JBc5S4O001O010O1O010O001O010O00100O00101eKUHS3l7lLTHT3m7jLTHV3m7iLTHW3m7fLTHZ3m7eLSH[3n7cLSH^3n7_LSHa3n7^LSHa3n7]LSHc3o7[LQHf3o7XLRHh3P8VLPHj3S6iK`K;^Nm3Q6hKcK8]NP4P6iKdK5\\NR4P6jKdK2]NT4o5jKfK0[NW4n5jKgKM\\NY4m5jKVLV4j3kKULV4j3kKULU4k3kKULU4k3lKTLT4l3mKSLT4l3lKTLT4l3mKSLS4m3mKSLT4l3mKSLS4m3nKSLQ4m3oKSLR4l3oKSLQ4m3PLRLQ4m3oKSLR4l3oKSLQ4m3oKSLR4l3oKkJB8`4l4nKlJC7_4m4oKkJD6^4n4nKlJF4]4o4mKmJG3\\4P5mKmJI1[4Q5mKmJI1[4Q5lKnJKOY4S5lKnJMMX4T5lKnJMMX4l4SLWKGKZ4b4WLcK@J]4X4ZLnK[OH`4m3^LZLSOHc4b3bLfLlNGf4X3eLQMfNFi4m2jL\\M^NFl4b2nLhMWNEP5W2PMTNQNDS5n1RM^NRN]OP5n1RMdNZ4U1iKkN[4n0hKQO^4g0fKXO^4?gKA]46hKJ\\4MjK2Z4EkK;Y4[OmKe0\\801O000010O000001O01O0000000O2O0000005K6I7J6J6J001N101O1O001O0O2O001O1O00M2L5K5K5K5L4L4L:F:EYfo0"}, "label": "the soccer player the number 8 on his shorts"}]}
{"id": 576188, "image": "COCO_train2014_000000576188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soccer play in maroon , gold and white pursuing the ball controlled by opponent in front of him\"."}], "task": "refering", "answer_template": "The \"a soccer play in maroon , gold and white pursuing the ball controlled by opponent in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 58, 59, 60, 81, 82, 83, 84, 85, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 173, 174, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 244, 245, 246, 247, 268, 269, 270, 271, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 338, 339, 340], "bbox": [0.5137343750000001, 0.12326291079812206, 0.880046875, 0.9465258215962441], "iscrowd": 0, "area": 31305.351949999997, "rle": {"size": [426, 640], "counts": "S_Y42X=1O1O1O1O1O1O2N1O1O1O1O1O2N1O2N1O2N2N1O2N001bLYOXJf0b5A]J?]5HbJ8X50eJ2U54iJMW55eJM[54bJN]57^JKa5;WJGi5:TJHk5:SJHl59QJIo58oIIQ68lIKR67lIJT67iIKW65hIMW64fIMZ65dIJ^67`IIa67]IId68[IHf68XIHj68UIHl68oHKR76jHLX75bHO_71]H1e7OYH2i7MTH4o7KoG6R8KkG5X8MbG5`8L\\G4g8MTG5n8Q23N2M3N2N1O2N2M3N2N2N2N6J7I6J5K20O1000000O10000hH\\Mi3d2SLhMd3Y2XLlMg3T2ULQNi3P2SLUNk3l1QLYNn3g1nK^NP4c1lKaNS4`1iKeNV4\\1dKjNZ4X1aK]Om3d0oK_OR4b0iKBV4?gKDX4>dKE\\4<aKF^4;_KGb4:ZKIe49WKJh47VKKi47SKKn49kJJT5<dJG[5`0]JBc5S4O001O010O1O010O001O010O00100O00101eKUHS3l7lLTHT3m7jLTHV3m7iLTHW3m7fLTHZ3m7eLSH[3n7cLSH^3n7_LSHa3n7^LSHa3n7]LSHc3o7[LQHf3o7XLRHh3P8VLPHj3S6iK`K;^Nm3Q6hKcK8]NP4P6iKdK5\\NR4P6jKdK2]NT4o5jKfK0[NW4n5jKgKM\\NY4m5jKVLV4j3kKULV4j3kKULU4k3kKULU4k3lKTLT4l3mKSLT4l3lKTLT4l3mKSLS4m3mKSLT4l3mKSLS4m3nKSLQ4m3oKSLR4l3oKSLQ4m3PLRLQ4m3oKSLR4l3oKSLQ4m3oKSLR4l3oKkJB8`4l4nKlJC7_4m4oKkJD6^4n4nKlJF4]4o4mKmJG3\\4P5mKmJI1[4Q5mKmJI1[4Q5lKnJKOY4S5lKnJMMX4T5lKnJMMX4l4SLWKGKZ4b4WLcK@J]4X4ZLnK[OH`4m3^LZLSOHc4b3bLfLlNGf4X3eLQMfNFi4m2jL\\M^NFl4b2nLhMWNEP5W2PMTNQNDS5n1RM^NRN]OP5n1RMdNZ4U1iKkN[4n0hKQO^4g0fKXO^4?gKA]46hKJ\\4MjK2Z4EkK;Y4[OmKe0\\801O000010O000001O01O0000000O2O0000005K6I7J6J6J001N101O1O001O0O2O001O1O00M2L5K5K5K5L4L4L:F:EYfo0"}, "label": "a soccer play in maroon , gold and white pursuing the ball controlled by opponent in front of him"}]}
{"id": 576188, "image": "COCO_train2014_000000576188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player with a red shirt and green shoes\"."}], "task": "refering", "answer_template": "The \"a player with a red shirt and green shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 60, 61, 62, 63, 83, 84, 85, 86, 87, 109, 110, 133, 156], "bbox": [0.586703125, 0.0002112676056338028, 0.82346875, 0.47896713615023473], "iscrowd": 0, "area": 11130.193250000004, "rle": {"size": [426, 640], "counts": "eQl41W=2N2N2N2N2M3N2N2N2N2M4M2N2N2N001N1000001O02N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N]E0P8OhG:X8DaGf0^82bF6]9l101O00001O001O00001O00001O001O00001O00001O001O00001O00001O001ON2L4L4M3L4O100000O101O2N2N1O2N2N2N1O2N2N1O2N1O1O001O001O001O001O001O001O10O01O001O001O001O001OkMdF<\\9DiF8V9GQG6l8JYG4d8KbG3[8MkG0R80SHNk70\\HMa73dHLX74mHJP75WIHf6]O^G4P2>_6XOjG6j1c0Y6ROUH8d1g0T6kNaH;]1k0R7QOQIQ1n6lNTIU1a91N2O1O1O1O2N1O1O1O1005H8H8G9D<Dbj^1"}, "label": "a player with a red shirt and green shoes"}]}
{"id": 576188, "image": "COCO_train2014_000000576188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the back side of the camera with white jersey with number 3\"."}], "task": "refering", "answer_template": "The \"the man in the back side of the camera with white jersey with number 3\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 60, 61, 62, 63, 83, 84, 85, 86, 87, 109, 110, 133, 156], "bbox": [0.586703125, 0.0002112676056338028, 0.82346875, 0.47896713615023473], "iscrowd": 0, "area": 11130.193250000004, "rle": {"size": [426, 640], "counts": "eQl41W=2N2N2N2N2M3N2N2N2N2M4M2N2N2N001N1000001O02N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N]E0P8OhG:X8DaGf0^82bF6]9l101O00001O001O00001O00001O001O00001O00001O001O00001O00001O001ON2L4L4M3L4O100000O101O2N2N1O2N2N2N1O2N2N1O2N1O1O001O001O001O001O001O001O10O01O001O001O001O001OkMdF<\\9DiF8V9GQG6l8JYG4d8KbG3[8MkG0R80SHNk70\\HMa73dHLX74mHJP75WIHf6]O^G4P2>_6XOjG6j1c0Y6ROUH8d1g0T6kNaH;]1k0R7QOQIQ1n6lNTIU1a91N2O1O1O1O2N1O1O1O1005H8H8G9D<Dbj^1"}, "label": "the man in the back side of the camera with white jersey with number 3"}]}
{"id": 576188, "image": "COCO_train2014_000000576188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man whith black short\"."}], "task": "refering", "answer_template": "The \"a man whith black short\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 54, 55, 76, 77, 78, 100, 101, 102, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 265, 266, 284, 285, 286, 289, 290, 308, 309, 310, 312, 313, 314, 332, 333, 336, 337], "bbox": [0.28718750000000004, 0.12133802816901408, 0.703015625, 0.9865258215962441], "iscrowd": 0, "area": 31742.027649999996, "rle": {"size": [426, 640], "counts": "cg\\2<k<8K1O1O00000000000000000000000000000bL_OSJ`0i5HTJ8f52VJN_5`0^J@^5P1VJoNo0^OX1g1eMkNQ1AY1d1fMkNo0DZ1a1gMkNm0GZ1_1hMjNl0K[1[1iMjNj0N\\1Y1iMiNi01]1V1jMiNg04]1T1lMgNf08]1Q1lMhNe0:^1n0mMhNc0=_1k0nMhNa0a0^1h0QNgN?d0_1e0RNfN>h0_1b0RNgN=j0`1?SNgN:n0b1;TNgN8Q1b1:UNdN8U1b17VNdN6X1c14VNeN6Y1c15TNbN8[1c17QN]N<^1b19nMYN?`1b1;kMUNb0b1b1=gMRNf0c1b1?dMmMj0f1a1a0aMiMm0h1a1c0^MeMP1j1a1e0[MaMS1l1a1g0WM]MX1n1`1i0TMYM[1P2`1k0QMUM^1R2`1m0nLQMa1U2_1n0lLlLe1X2^1P1hLiLi1Y2^1R1eLeLl1[2^1T1bLaLo1]2^1h1aNZN^1e1bN]N]1c1cN^N\\1b1cN`N\\1`1cNbN\\1]1dNeN[1[1dNgN\\1X1cNiN_1U1`NmNa1P1_N^MUND_3k2`MjKDh1^OC_3i2XMTLI`1BB_3h2PM\\LN[1DA_3e2iLgL2T1F@b3b2_LQM8m0I@a3UOZLn3:^ML^O\\4Q3fKaM0^O]4m2bKfM2\\O_4l2]KiM4\\Oa4i2YKkM8[O`4i2WKmM:YO_4k2UKlM>YO\\4R5eKmJ\\4Q5fKoJY4Q5gKoJY4Q5hKoJX4o4jKPKV4o4lKQKS4n4oKQKR4m4PLSKo3k4SLUKm3j4ULVK0fNj2Q6YMXKLhNk2o5[MYKHjNl2l5^MYKElNn2h5_M]KAlNP3f5aM]K^OnNQ3d5cM^KZOoNT3`5fM_KUOROU3^5hM`KQOSOW3\\5jM`KnNUOY3Z5jMbKjNUO\\3X5lMgKbNROc3V5lMlK[NoNi3U5mM]LU2a3lM^LV2_3lMaLU2]3lMbLV2\\3kMcLV2\\3kMcLW2[3jMdLX2Y3jMfLX2X3iMgLY2W3hMhLY2W3hMhLZ2U3hMjLZ2T3gMkL[2S3fMlL\\2R3eMmL]2Q3dMnL]2P3eMoL]2o2eMoL]2o2dMPM^2n2cMQM_2l2cMSM^2l2cMSM_2k2bMTM`2j2aMUMa2i2`MVMb2g2`MXMb2f2_MXMc2g2^MXMd2f2]MYMe2d2]M[Me2c2[M]Mg2a2ZM^Mh2a2XM]Mk2b2UM]Mm2b2SM]Mo2c2PM[MS3d2mL[MU3d2kL[MW3e2hLYM[3f2eLRMd3m2\\LkLm3U3U3O1O100O100O1O100O01O001O00001O0O2O001O001O00001O00dMUFY1j9dNZF[1f9bN_F\\1a9aNcF^1]9_NgFa1X9\\NmFb1S9ZNRGe1m8YNXGf1g8WN]Gh1c8UNaGj1_8SNfGk1Z8RNjGn1j9O2N1O2O0O2N101N1O2N101N2N101N1O2N10O00001O000O101O00001N1000001O0O10001O000O2O00000O2O00001O01O001O0O2N1O2N1O2N101N1O2N1O2N2N101N1O2N1O2N1O2N2N1O2N2N2N2N1O2N2N2N2N1OZd^2"}, "label": "a man whith black short"}]}
{"id": 576188, "image": "COCO_train2014_000000576188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soccer player playing the ball away from another soccer player\"."}], "task": "refering", "answer_template": "The \"a soccer player playing the ball away from another soccer player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 54, 55, 76, 77, 78, 100, 101, 102, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 265, 266, 284, 285, 286, 289, 290, 308, 309, 310, 312, 313, 314, 332, 333, 336, 337], "bbox": [0.28718750000000004, 0.12133802816901408, 0.703015625, 0.9865258215962441], "iscrowd": 0, "area": 31742.027649999996, "rle": {"size": [426, 640], "counts": "cg\\2<k<8K1O1O00000000000000000000000000000bL_OSJ`0i5HTJ8f52VJN_5`0^J@^5P1VJoNo0^OX1g1eMkNQ1AY1d1fMkNo0DZ1a1gMkNm0GZ1_1hMjNl0K[1[1iMjNj0N\\1Y1iMiNi01]1V1jMiNg04]1T1lMgNf08]1Q1lMhNe0:^1n0mMhNc0=_1k0nMhNa0a0^1h0QNgN?d0_1e0RNfN>h0_1b0RNgN=j0`1?SNgN:n0b1;TNgN8Q1b1:UNdN8U1b17VNdN6X1c14VNeN6Y1c15TNbN8[1c17QN]N<^1b19nMYN?`1b1;kMUNb0b1b1=gMRNf0c1b1?dMmMj0f1a1a0aMiMm0h1a1c0^MeMP1j1a1e0[MaMS1l1a1g0WM]MX1n1`1i0TMYM[1P2`1k0QMUM^1R2`1m0nLQMa1U2_1n0lLlLe1X2^1P1hLiLi1Y2^1R1eLeLl1[2^1T1bLaLo1]2^1h1aNZN^1e1bN]N]1c1cN^N\\1b1cN`N\\1`1cNbN\\1]1dNeN[1[1dNgN\\1X1cNiN_1U1`NmNa1P1_N^MUND_3k2`MjKDh1^OC_3i2XMTLI`1BB_3h2PM\\LN[1DA_3e2iLgL2T1F@b3b2_LQM8m0I@a3UOZLn3:^ML^O\\4Q3fKaM0^O]4m2bKfM2\\O_4l2]KiM4\\Oa4i2YKkM8[O`4i2WKmM:YO_4k2UKlM>YO\\4R5eKmJ\\4Q5fKoJY4Q5gKoJY4Q5hKoJX4o4jKPKV4o4lKQKS4n4oKQKR4m4PLSKo3k4SLUKm3j4ULVK0fNj2Q6YMXKLhNk2o5[MYKHjNl2l5^MYKElNn2h5_M]KAlNP3f5aM]K^OnNQ3d5cM^KZOoNT3`5fM_KUOROU3^5hM`KQOSOW3\\5jM`KnNUOY3Z5jMbKjNUO\\3X5lMgKbNROc3V5lMlK[NoNi3U5mM]LU2a3lM^LV2_3lMaLU2]3lMbLV2\\3kMcLV2\\3kMcLW2[3jMdLX2Y3jMfLX2X3iMgLY2W3hMhLY2W3hMhLZ2U3hMjLZ2T3gMkL[2S3fMlL\\2R3eMmL]2Q3dMnL]2P3eMoL]2o2eMoL]2o2dMPM^2n2cMQM_2l2cMSM^2l2cMSM_2k2bMTM`2j2aMUMa2i2`MVMb2g2`MXMb2f2_MXMc2g2^MXMd2f2]MYMe2d2]M[Me2c2[M]Mg2a2ZM^Mh2a2XM]Mk2b2UM]Mm2b2SM]Mo2c2PM[MS3d2mL[MU3d2kL[MW3e2hLYM[3f2eLRMd3m2\\LkLm3U3U3O1O100O100O1O100O01O001O00001O0O2O001O001O00001O00dMUFY1j9dNZF[1f9bN_F\\1a9aNcF^1]9_NgFa1X9\\NmFb1S9ZNRGe1m8YNXGf1g8WN]Gh1c8UNaGj1_8SNfGk1Z8RNjGn1j9O2N1O2O0O2N101N1O2N101N2N101N1O2N10O00001O000O101O00001N1000001O0O10001O000O2O00000O2O00001O01O001O0O2N1O2N1O2N101N1O2N1O2N2N101N1O2N1O2N1O2N2N1O2N2N2N2N1O2N2N2N2N1OZd^2"}, "label": "a soccer player playing the ball away from another soccer player"}]}
{"id": 314051, "image": "COCO_train2014_000000314051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"holstein laying on the ground\"."}], "task": "refering", "answer_template": "The \"holstein laying on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 290, 291, 292, 293, 294, 295], "bbox": [0.479484375, 0.4011875, 0.907125, 0.7381249999999999], "iscrowd": 0, "area": 27932.23529999999, "rle": {"size": [480, 640], "counts": "`V`43k>3L4M4[BE_<>[CGc<=VCHh<;RCJk<;nBKo<T1N2N2M3N2N2EkMiCX2T<jMiCY2U<iMiCY2U<:M3N2N2N3M2N2O11O000001O0000000000010O00H[DUMe;k2]DSMc;m2_DQMa;o28O100000000O10O10O10000O10O10O10000O10O10O10000O10O100000000O2O0000001O0O1000001O000O10001O000O10001O00000O101O0000001]DiLZ;X3eDiL[;W3dDjL\\;V3cDlL\\;T3dDlL\\;U3bDlL];\\30000000001N1000000000001O000000000000001O000000000000001O0000000000001O00000000000000000O100O100O10000O010O100O100O100O100O10OFSLcEm3\\:ULcEk3\\:WLbEj3]:XLbEh3]:[LaEe3_:\\L`Ed3`:]L_Ec3a:>001O0O100000000000000000000O10001O000000000000000O100000N2O1O1N2O1N2O1O0O2O1N2O1O10O010O100O100O100O100O100O1dNoEiMQ:V2cFVM^9h2_1M3N2M3M3N2M3M3N3L3M3N2M3M3H7G9G9H8G9N20000O10000000O100000000000O10004L3M4L4L3Mo]k0"}, "label": "holstein laying on the ground"}]}
{"id": 314051, "image": "COCO_train2014_000000314051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow laying down in grass\"."}], "task": "refering", "answer_template": "The \"a cow laying down in grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 290, 291, 292, 293, 294, 295], "bbox": [0.479484375, 0.4011875, 0.907125, 0.7381249999999999], "iscrowd": 0, "area": 27932.23529999999, "rle": {"size": [480, 640], "counts": "`V`43k>3L4M4[BE_<>[CGc<=VCHh<;RCJk<;nBKo<T1N2N2M3N2N2EkMiCX2T<jMiCY2U<iMiCY2U<:M3N2N2N3M2N2O11O000001O0000000000010O00H[DUMe;k2]DSMc;m2_DQMa;o28O100000000O10O10O10000O10O10O10000O10O10O10000O10O100000000O2O0000001O0O1000001O000O10001O000O10001O00000O101O0000001]DiLZ;X3eDiL[;W3dDjL\\;V3cDlL\\;T3dDlL\\;U3bDlL];\\30000000001N1000000000001O000000000000001O000000000000001O0000000000001O00000000000000000O100O100O10000O010O100O100O100O100O10OFSLcEm3\\:ULcEk3\\:WLbEj3]:XLbEh3]:[LaEe3_:\\L`Ed3`:]L_Ec3a:>001O0O100000000000000000000O10001O000000000000000O100000N2O1O1N2O1N2O1O0O2O1N2O1O10O010O100O100O100O100O100O1dNoEiMQ:V2cFVM^9h2_1M3N2M3M3N2M3M3N3L3M3N2M3M3H7G9G9H8G9N20000O10000000O100000000000O10004L3M4L4L3Mo]k0"}, "label": "a cow laying down in grass"}]}
{"id": 314051, "image": "COCO_train2014_000000314051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow standing in a field next to a cow laying down\"."}], "task": "refering", "answer_template": "The \"a cow standing in a field next to a cow laying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 233, 234, 235, 237, 238], "bbox": [0.126125, 0.20464583333333333, 0.35481250000000003, 0.6538541666666667], "iscrowd": 0, "area": 19848.524900000004, "rle": {"size": [480, 640], "counts": "eRV12m>2N3L3N2O1N3YCHb:9XENf:2TE6j:JPE>o:AlDg0Q;YOlDl0S;SOiDS1U;mNgDk1f:TNWER2f:nMWEX2g:gMXE]2e:cMYEc2c:]M[Eh2`:YM`Ek2[:VMdEo2X:YM`Em2d1]Lm6d4mH]KR7i4hHXKW7m4aHXK]7m4WH[Kh7j4jG`KU8f5O2N1000O10O1000nMlGWMU8i2SHoLl7R3[HgLe7X3cHaL]7_3kHRL\\7n3j110O1000000O101O000O2O1N1O2O1N2N2N00N3M20011O1N3ZGhK`6d4PIbKn6Q5[HUKe7NdG]4a0jKm7FbGb4<kKV8^O`Gl5d8oI]GQ6j8O001O0000000000000SOTGdKm8^4TG^Kn8b4UGYKl8_4_G]Kc8`4o0N3M3M2N3M2M4M2N3001N1000IZESLg:m3[EQLd:Q45001N1O1N1011N101O1N2O0O21O1O0OM40000O0O2M2L5K4O2O0000001N14L7J6In0RO<D1002N2M4M2RL]FY2g9VMmFc2V9lL^Gm2a:WMZDf0A8P=@ZC:j<^O`C;e<]OR]Q6"}, "label": "a cow standing in a field next to a cow laying down"}]}
{"id": 314051, "image": "COCO_train2014_000000314051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large black and white cow with chain on neck staring at camera\"."}], "task": "refering", "answer_template": "The \"large black and white cow with chain on neck staring at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 233, 234, 235, 237, 238], "bbox": [0.126125, 0.20464583333333333, 0.35481250000000003, 0.6538541666666667], "iscrowd": 0, "area": 19848.524900000004, "rle": {"size": [480, 640], "counts": "eRV12m>2N3L3N2O1N3YCHb:9XENf:2TE6j:JPE>o:AlDg0Q;YOlDl0S;SOiDS1U;mNgDk1f:TNWER2f:nMWEX2g:gMXE]2e:cMYEc2c:]M[Eh2`:YM`Ek2[:VMdEo2X:YM`Em2d1]Lm6d4mH]KR7i4hHXKW7m4aHXK]7m4WH[Kh7j4jG`KU8f5O2N1000O10O1000nMlGWMU8i2SHoLl7R3[HgLe7X3cHaL]7_3kHRL\\7n3j110O1000000O101O000O2O1N1O2O1N2N2N00N3M20011O1N3ZGhK`6d4PIbKn6Q5[HUKe7NdG]4a0jKm7FbGb4<kKV8^O`Gl5d8oI]GQ6j8O001O0000000000000SOTGdKm8^4TG^Kn8b4UGYKl8_4_G]Kc8`4o0N3M3M2N3M2M4M2N3001N1000IZESLg:m3[EQLd:Q45001N1O1N1011N101O1N2O0O21O1O0OM40000O0O2M2L5K4O2O0000001N14L7J6In0RO<D1002N2M4M2RL]FY2g9VMmFc2V9lL^Gm2a:WMZDf0A8P=@ZC:j<^O`C;e<]OR]Q6"}, "label": "large black and white cow with chain on neck staring at camera"}]}
{"id": 10948, "image": "COCO_train2014_000000010948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with green tops singing a song\"."}], "task": "refering", "answer_template": "The \"a woman with green tops singing a song\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 38, 39, 40, 41, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221], "bbox": [0.07098, 0.1101063829787234, 0.43092, 0.9865159574468084], "iscrowd": 0, "area": 37652.815749999994, "rle": {"size": [376, 500], "counts": "hU=2b;e0\\O`0@3L4M3M3L4M3M3M3M3M3M3M3M3\\M\\NQKg1c4@^Jd0V5U3G9J6J6K5O1O1O1O100O1O1O1O1O100O1O2N1O1O1YNlHjNT7o0VImNk6m0^IoNc6k0fIQO[6l0kIPOV6o0nIlNT6S1PJiNQ6V1SJfNn5Y1VJbNl5\\1YJ`Nh5_1\\J]Ne5b1Y2O1O1O1N2O1O1O1O10000O1000000O10000O1000001N10000O10000000000000000000000000000000000000000000000000000000000000VMeNWK[1h4oNoJR1o4SOnJl0Q5YOkJh0S5^OhJb0W5J^J7`50ZJ0e55WJLg58VJHh5=UJCi5b0TJ_Oi5f0TJZOj5k0SJVOj5P1RJPOl5U1QJlNX1\\Ne2n2oKgN[1\\Nc2\\3eK[Nd1ZNe2_3eKZNc1XNe2a3gKYNb1WNd2f5WM\\JROJR3[6]MlI_O0l2\\6_MeIC3j2[6aMcIC4i2\\6cMaIA6i2\\6dM`IA6h2^6aMZJ\\2i5aMYJ]2j5aMXJ]2j5bMWJ^2i5aMXJ^2i5`MYJ_2i5^MZJa2f5]M^J`2\\73M3M3M4L3M3M2N1O2N2N2O0O2N2N2N2O0O1000O10O01O10O01M2O2M3N1N3N1N3K5J5J7I6ROo0O4L?AT`X3"}, "label": "a woman with green tops singing a song"}]}
{"id": 10948, "image": "COCO_train2014_000000010948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl was sinning a song holdintg with in a mobile\"."}], "task": "refering", "answer_template": "The \"a girl was sinning a song holdintg with in a mobile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 38, 39, 40, 41, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221], "bbox": [0.07098, 0.1101063829787234, 0.43092, 0.9865159574468084], "iscrowd": 0, "area": 37652.815749999994, "rle": {"size": [376, 500], "counts": "hU=2b;e0\\O`0@3L4M3M3L4M3M3M3M3M3M3M3M3\\M\\NQKg1c4@^Jd0V5U3G9J6J6K5O1O1O1O100O1O1O1O1O100O1O2N1O1O1YNlHjNT7o0VImNk6m0^IoNc6k0fIQO[6l0kIPOV6o0nIlNT6S1PJiNQ6V1SJfNn5Y1VJbNl5\\1YJ`Nh5_1\\J]Ne5b1Y2O1O1O1N2O1O1O1O10000O1000000O10000O1000001N10000O10000000000000000000000000000000000000000000000000000000000000VMeNWK[1h4oNoJR1o4SOnJl0Q5YOkJh0S5^OhJb0W5J^J7`50ZJ0e55WJLg58VJHh5=UJCi5b0TJ_Oi5f0TJZOj5k0SJVOj5P1RJPOl5U1QJlNX1\\Ne2n2oKgN[1\\Nc2\\3eK[Nd1ZNe2_3eKZNc1XNe2a3gKYNb1WNd2f5WM\\JROJR3[6]MlI_O0l2\\6_MeIC3j2[6aMcIC4i2\\6cMaIA6i2\\6dM`IA6h2^6aMZJ\\2i5aMYJ]2j5aMXJ]2j5bMWJ^2i5aMXJ^2i5`MYJ_2i5^MZJa2f5]M^J`2\\73M3M3M4L3M3M2N1O2N2N2O0O2N2N2N2O0O1000O10O01O10O01M2O2M3N1N3N1N3K5J5J7I6ROo0O4L?AT`X3"}, "label": "a girl was sinning a song holdintg with in a mobile"}]}
{"id": 10948, "image": "COCO_train2014_000000010948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown shirt holds a pink wii control\"."}], "task": "refering", "answer_template": "The \"a man in a brown shirt holds a pink wii control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 44, 45, 46, 50, 62, 63, 64, 65, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 205, 206, 207, 208, 209, 223, 224, 225, 226, 227], "bbox": [0.41064, 0.12515957446808512, 0.8179, 0.9858244680851064], "iscrowd": 0, "area": 32379.451600000008, "rle": {"size": [376, 500], "counts": "Vc[2U1V:=C=C=D<C=E;UNgLSK]3b2^LkMc0]OS3g2^L^Ml0Ej2k2_L[Mj0Ek2o2_LWMj0Dk2U3]LSMl0Bk2Z3jN`LZ1`3hNYL]1g3Z3O100O100O100O10000O100O100O10000O100O100O100O100\\OPH_Mo7j1kHUNS7k1PISNo6l1UISNi6m1ZIQNe6o1`13N1O2N1N3N1O1O2N1N3N1O2N1N2012N2N3M2N2N3M2N2N3M2N2O1N3M2O1N3N1N2N2OO001O00001O00001O2N2N1O2eGPNj6R2QIQNo6Q2lHRNT7R2eHQN[7Y2QHQNo7l2000000000000VK^M>b2]OkM;U2@UN;l1_O^N<b1^OhN>X1]ORO>n0]O[O?e0\\OD`0=ZOLb04XO6d0JWO?e0AXOf0d0ZOYOn0b0SO[OT1`0lN_OZ1<fND_17aNHe13[NMj1NWN1o1IQN6V2DjM<]2]OcMc0c2WO]Mi0i2QOXMm0o2mNQMS1j60000O2O000O10001O0O1000001O000000001O0O100000001N1O1O1O1O101O01O00000001O00O1O2O0O1O1O1O00000001O000000XOSOfFm0Z9\\O]Fd0d9BUF>k9DSF<m9FRF9n9IPF7Q:g0001O1O1M3N3L3N2M4L5K4L5K4M3M3M2N3M3M`ZQ1"}, "label": "a man in a brown shirt holds a pink wii control"}]}
{"id": 10948, "image": "COCO_train2014_000000010948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is playing wii\"."}], "task": "refering", "answer_template": "The \"a man is playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 44, 45, 46, 50, 62, 63, 64, 65, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 205, 206, 207, 208, 209, 223, 224, 225, 226, 227], "bbox": [0.41064, 0.12515957446808512, 0.8179, 0.9858244680851064], "iscrowd": 0, "area": 32379.451600000008, "rle": {"size": [376, 500], "counts": "Vc[2U1V:=C=C=D<C=E;UNgLSK]3b2^LkMc0]OS3g2^L^Ml0Ej2k2_L[Mj0Ek2o2_LWMj0Dk2U3]LSMl0Bk2Z3jN`LZ1`3hNYL]1g3Z3O100O100O100O10000O100O100O10000O100O100O100O100\\OPH_Mo7j1kHUNS7k1PISNo6l1UISNi6m1ZIQNe6o1`13N1O2N1N3N1O1O2N1N3N1O2N1N2012N2N3M2N2N3M2N2N3M2N2O1N3M2O1N3N1N2N2OO001O00001O00001O2N2N1O2eGPNj6R2QIQNo6Q2lHRNT7R2eHQN[7Y2QHQNo7l2000000000000VK^M>b2]OkM;U2@UN;l1_O^N<b1^OhN>X1]ORO>n0]O[O?e0\\OD`0=ZOLb04XO6d0JWO?e0AXOf0d0ZOYOn0b0SO[OT1`0lN_OZ1<fND_17aNHe13[NMj1NWN1o1IQN6V2DjM<]2]OcMc0c2WO]Mi0i2QOXMm0o2mNQMS1j60000O2O000O10001O0O1000001O000000001O0O100000001N1O1O1O1O101O01O00000001O00O1O2O0O1O1O1O00000001O000000XOSOfFm0Z9\\O]Fd0d9BUF>k9DSF<m9FRF9n9IPF7Q:g0001O1O1M3N3L3N2M4L5K4L5K4M3M3M2N3M3M`ZQ1"}, "label": "a man is playing wii"}]}
{"id": 404183, "image": "COCO_train2014_000000404183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bus that is parked next to two other buses\"."}], "task": "refering", "answer_template": "The \"a blue bus that is parked next to two other buses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284, 301, 305], "bbox": [0.08932812500000001, 0.31685416666666666, 0.36742187499999995, 0.8], "iscrowd": 0, "area": 33655.47185, "rle": {"size": [480, 640], "counts": "XRk09Q>g0M4L3J6POP1POQ1POo0@`0H8H8G7J7I6J7H8I6J7000O10000O100O1000O01O1J6K5K5K50000000O1000001O001O3M4L4L000O10000000000000000000000000000000000000000O100000O100O1O1O100O1O1O1O1000000O100000000O100000O10O100000000O100000000O10001O0016Ic0^O0O01000O100O10N1O2N1N3N2N2N1O2M3N2N2O1N101N2O1N2O0O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1O001O1O1O1O001O1O1O1O1O001O1O1K5L3L<D;F;YOg0UOk0UOj0[Of0^O]cm5"}, "label": "a blue bus that is parked next to two other buses"}]}
{"id": 404183, "image": "COCO_train2014_000000404183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue bus\"."}], "task": "refering", "answer_template": "The \"the blue bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284, 301, 305], "bbox": [0.08932812500000001, 0.31685416666666666, 0.36742187499999995, 0.8], "iscrowd": 0, "area": 33655.47185, "rle": {"size": [480, 640], "counts": "XRk09Q>g0M4L3J6POP1POQ1POo0@`0H8H8G7J7I6J7H8I6J7000O10000O100O1000O01O1J6K5K5K50000000O1000001O001O3M4L4L000O10000000000000000000000000000000000000000O100000O100O1O1O100O1O1O1O1000000O100000000O100000O10O100000000O100000000O10001O0016Ic0^O0O01000O100O10N1O2N1N3N2N2N1O2M3N2N2O1N101N2O1N2O0O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1O001O1O1O1O001O1O1O1O1O001O1O1K5L3L<D;F;YOg0UOk0UOj0[Of0^O]cm5"}, "label": "the blue bus"}]}
{"id": 404183, "image": "COCO_train2014_000000404183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a creme and green color double decker bus\"."}], "task": "refering", "answer_template": "The \"a creme and green color double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319], "bbox": [0.6235937500000001, 0.3084166666666667, 0.9202187500000001, 0.7848333333333334], "iscrowd": 0, "area": 38676.35130000001, "rle": {"size": [480, 640], "counts": "`[k5>U=^1bN]1gNZ1N2N2M3N1O2N2N2N1N3N2N2N1O2M3N2N1O2N2N2M2O2N2N2N1N3N2N2N1O2N2M3N1O1O1O1N101O1O1O1O1N2O1O1O1O2M2O01000000000000000000000001O0000000O1000000000000000000000000000000O10001O0000000000000000000000000O1000000000000000001O0000000000000O1000000000000000000000000000000O2O00000000000000000000000000000O100000000000001O000000000000000O1000000000000000000000009Gb0^Ob0^Ob0^Oa0^Oc0^Ob0^Ob0^Ob0^Oa0_Ob0^Ob0^Ojag0"}, "label": "a creme and green color double decker bus"}]}
{"id": 404183, "image": "COCO_train2014_000000404183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange double - decker bus displaying ipswich\"."}], "task": "refering", "answer_template": "The \"an orange double - decker bus displaying ipswich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319], "bbox": [0.6235937500000001, 0.3084166666666667, 0.9202187500000001, 0.7848333333333334], "iscrowd": 0, "area": 38676.35130000001, "rle": {"size": [480, 640], "counts": "`[k5>U=^1bN]1gNZ1N2N2M3N1O2N2N2N1N3N2N2N1O2M3N2N1O2N2N2M2O2N2N2N1N3N2N2N1O2N2M3N1O1O1O1N101O1O1O1O1N2O1O1O1O2M2O01000000000000000000000001O0000000O1000000000000000000000000000000O10001O0000000000000000000000000O1000000000000000001O0000000000000O1000000000000000000000000000000O2O00000000000000000000000000000O100000000000001O000000000000000O1000000000000000000000009Gb0^Ob0^Ob0^Oa0^Oc0^Ob0^Ob0^Ob0^Oa0_Ob0^Ob0^Ojag0"}, "label": "an orange double - decker bus displaying ipswich"}]}
{"id": 404183, "image": "COCO_train2014_000000404183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white double decker bus\"."}], "task": "refering", "answer_template": "The \"a red and white double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 310, 311, 312], "bbox": [0.401578125, 0.3151041666666667, 0.60859375, 0.8077500000000001], "iscrowd": 0, "area": 25564.501700000008, "rle": {"size": [480, 640], "counts": "ejh33g>7J5J7I7J5@a0[Oe0[Od0[Of0[Od0\\Oe0[Oe0[Od0@a0I6J7MkN_HPK`7P5W1101O000O2O001O0O101O000O4M5K5K4K1000000O1000000O1000001O01O0001O0000001O0000001O000001O01O0000000000000000000001O000O100000000000000000001O0O10000000000000000000000O2O00000O1N2O1NlGaIk7^6SHgIk7W6THmIk7Q6THSJk7l5SHXJl7f5SH^Jn7_5PHeJT8S63L3M4L4L3L5K5K4L5K4L5K5fM\\FnNh94YGEl8\\OUH=P8eNoHU1e:I7ATkd3"}, "label": "a red and white double decker bus"}]}
{"id": 404183, "image": "COCO_train2014_000000404183.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and beige bus between two other busses\"."}], "task": "refering", "answer_template": "The \"a brown and beige bus between two other busses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 310, 311, 312], "bbox": [0.401578125, 0.3151041666666667, 0.60859375, 0.8077500000000001], "iscrowd": 0, "area": 25564.501700000008, "rle": {"size": [480, 640], "counts": "ejh33g>7J5J7I7J5@a0[Oe0[Od0[Of0[Od0\\Oe0[Oe0[Od0@a0I6J7MkN_HPK`7P5W1101O000O2O001O0O101O000O4M5K5K4K1000000O1000000O1000001O01O0001O0000001O0000001O000001O01O0000000000000000000001O000O100000000000000000001O0O10000000000000000000000O2O00000O1N2O1NlGaIk7^6SHgIk7W6THmIk7Q6THSJk7l5SHXJl7f5SH^Jn7_5PHeJT8S63L3M4L4L3L5K5K4L5K4L5K5fM\\FnNh94YGEl8\\OUH=P8eNoHU1e:I7ATkd3"}, "label": "a brown and beige bus between two other busses"}]}
{"id": 363224, "image": "COCO_train2014_000000363224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large vehicle moving down the highway\"."}], "task": "refering", "answer_template": "The \"a large vehicle moving down the highway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 288, 290, 291], "bbox": [0.43392187499999996, 0.554215456674473, 0.682234375, 0.8168384074941453], "iscrowd": 0, "area": 11955.41865, "rle": {"size": [427, 640], "counts": "ZVd38n<5K5O2jC^Oa;b0ZDHa;8ZD1b;OZD7e;h02N2N3M^Oa0E<M3M3N1N3M2N3M2N3O001O0010O0000001N1000001O0O100O2N100O1O2N1O100100O2N2N2O1N5K2N101N1O2N100O000001O00001O00001O00001O00001O0000001O00001O000O2M2O1N3M2N2N3M20001O0000001O01O01O00001O00001O00001O0001O01O00001O00001O3M4L5K4L2N01O01O0001O2O1N2N1O2N10O0000001O002N1O2N1O2N1N3N1O2N1O1O0000O13M>B>B>BnYd2"}, "label": "a large vehicle moving down the highway"}]}
{"id": 363224, "image": "COCO_train2014_000000363224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a highway worker driving down the road\"."}], "task": "refering", "answer_template": "The \"a highway worker driving down the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 288, 290, 291], "bbox": [0.43392187499999996, 0.554215456674473, 0.682234375, 0.8168384074941453], "iscrowd": 0, "area": 11955.41865, "rle": {"size": [427, 640], "counts": "ZVd38n<5K5O2jC^Oa;b0ZDHa;8ZD1b;OZD7e;h02N2N3M^Oa0E<M3M3N1N3M2N3M2N3O001O0010O0000001N1000001O0O100O2N100O1O2N1O100100O2N2N2O1N5K2N101N1O2N100O000001O00001O00001O00001O00001O0000001O00001O000O2M2O1N3M2N2N3M20001O0000001O01O01O00001O00001O00001O0001O01O00001O00001O3M4L5K4L2N01O01O0001O2O1N2N1O2N10O0000001O002N1O2N1O2N1N3N1O2N1O1O0000O13M>B>B>BnYd2"}, "label": "a highway worker driving down the road"}]}
{"id": 183007, "image": "COCO_train2014_000000183007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"car on left hand back\"."}], "task": "refering", "answer_template": "The \"car on left hand back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.000875, 0.27215625, 0.4514583333333333, 0.52078125], "iscrowd": 0, "area": 17980.5577, "rle": {"size": [640, 480], "counts": "U:2\\?2_Da0a;BiCQ1V<SOTC`1l<cN^BQ2`=QNTBZ2l=fMQB]2o=cMnA`2Q>aMQAA?R3`>]MQAI4m2j>[MRA0Hi2V?WMRA\\3m>eLSA[3m>eLSA[3m>eLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3l>fLTAZ3m>eLSA\\3l>dLTA\\3l>dLTA\\3l>dLTA\\3l>dLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLUA[3k>eLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVAZ3j>fLVA[3i>eLWA[3i>eLWA[3i>eLWA[3i>eLWA[3i>eLWA[3i>eLXAZ3h>fLXAZ3h>fLXAZ3h>fLXAZ3h>fLXAZ3h>fLXAZ3h>fLXAZ3h>fLXAZ3h>fLXAW2HUNP?DXAj15bNc>DXA\\1c0POU>DXAo0P1]Oh=DXAa0^1KZ=DXA3l19l<DXAOP2=i<CWA0P2=i<CWA0P2=i<CWA0P2=i<CWA0P2=i<CWA0P2=i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1>i<CXAOo1?h<BYAOo1?h<BYAOo1?h<BYAOo1?h<BYAOo1?h<BYAOo1?h<BYAOP2>g<CYAOP2>g<CYAOP2>g<CYAOP2>g<CZANo1?g<CZANo1?g<CZANo1?g<CZANo1?g<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<CZAMP2`0f<C[ALo1a0f<C[ALo1a0f<C[ALo1a0f<C[ALo1a0f<C[ALo1a0f<C[ALo1a0f<C[ALo1a0f<C[ALo1b0e<B\\ALo1b0f<A[AMP2a0e<B[AMP2a0f<AZANP2a0f<AZANP2a0g<@YAOP2a0g<@YAOP2a0g<@YAOP2a0h<_OXA0P2a0h<_OXA0P2a0i<^OWA0Q2b0h<^OWA0Q2b0i<]OWA0P2c0i<]OWA0P2c0j<\\OVA1P2c0j<\\OVA1P2c0j<\\OVA1P2c0k<[OUA2P2c0k<[OUA2P2c0l<ZOTA3P2c0l<ZOTA3P2c0m<YOSA4P2c0m<YOSA4P2c0m<YOSA4P2c0n<XORA5P2c0n<XORA5P2c0o<WOQA6P2c0o<WOQA6P2c0P=VOPA7P2c0P=VOPA7P2c0Q=UOo@8P2c0Q=UOo@8P2c0Q=UOPA7P2c0Q=UOo@8P2c0Q=UOo@8P2c0R=TOn@9P2d0Q=SOo@9P2d0d=\\O\\Bd0d=\\O\\Bd0d=\\O\\Bd0d=\\O\\Bd0d=\\O\\Bd0d=\\O\\Bd0d=[O]Be0c=[O]Be0c=[O]Be0d=ZO\\Bf0f=XOZBh0g=WOYBi0h=VOXBj0i=UOWBk0j=TOVBl0k=SOUBm0_`0000000000000000000000000000000O1O1N2N2N2M3N2N;EZRT5"}, "label": "car on left hand back"}]}
{"id": 273120, "image": "COCO_train2014_000000273120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a woman holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 80, 81, 82, 83, 84, 104, 105, 106, 126, 127, 128, 129, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 357, 358, 360, 361, 383, 384, 405, 406, 407], "bbox": [0.48896875, 0.12390438247011953, 0.770109375, 0.9602589641434263], "iscrowd": 0, "area": 30802.766200000002, "rle": {"size": [502, 640], "counts": "lbi41e?2M2N3M2N2O2M2N2cF@a3c0ZLAd3a0XLCQ2LPJa0l3Go1NRJ;m3Kl10TJ6n3Mj12UJ2n30h14WJLo34h12XJJm38j10VJIm3;l1NTJGn3?m1MQJEo3b0n1MoIBQ4d0o1MmI^OR4i0cM]NU2^1SL]OR4l0cM[NW2`1PLYOT4P1cMYNX2a1mKWOU4S1dMVNZ2d1hKTOW4V1eMSN\\2b2N^OcMSN]2`2MAdMPN_2_2JEeMnM`2\\2IJeMkMb2[2FNfMhMc2[2D1fMgMe2W2D5eMeMg2V2YOa0nM[Mh2S2oNP1VNnLj2S2dN\\1_NdLl2P2YNj1hNWLo2n1nMX2QOkKP3n1cMe2ZO_KR3l1XMR3CTKU3i1mLa3KhJW3g1bLn35\\JX3g1WL[4>oI[3P9cLQG]3Q9_LQGa3P9]LPGc3S9ZLmFg3T9WLlFj3V9RLlFn3]:00100O1O0UMRLPJo3n5TLoIm3o5WLnIk3o5YLoIh3o5\\LmIe3Q6_LlIc3R6`LlIa3Q6cLlI^3S6eLjI]3S6gLkIZ3S6iLjIX3T6mLhIU3V6nLhIR3V6RMgIo2X6TMeIm2Y6WMeIi2Z6ZMcIg2[6]MbId2]6_MaIa2]6cM`I]2`6fM]I[2a6iM]IW2n6`MoHa26[KS6j1dIl26_K`6Z1YIW34dKl6l0mHa34hKo6g0jHb34lKR7b0hHb33QLV7<dHd33ULY76bHf36ULX75_Hg39XLV71_Hh3:[LV7L]Hj3=^LT7i4lHZKR7d4nH`KP7_4oHeKo6Y4RIjKl6U4SIoKj6P4VISLi6l3VIXLh6g0lH0=d0]5XOXJ5:d0_5TOXJ89e0`5oNYJ=7e0`5jN[Ja05f0`5gN\\Jd03f0b5bN]Jh01g0c5]N^Jm0Of0d5ZN]JQ1Of0e5UN^JV1Lf0g5QN^JY1Kh0f5lMaJ]1Ig0g5hMbJa1Gh0h5dMbJe1Eh0j5_McJi1Dh0j5\\McJm1Bh0l5WMdJQ2@j0k5RMgJU2]Oj0`3RL`Nl0dNX2]Oj0Y3ZLeN`0gN]2ZOj0T3aLjN4jNa2XOk0n2hLoNIlNe2VOk0i2oLTO]OoNi2UOk0b2XMWOQOTOm2ROl0\\2^M]OcNXOT3oNl0V2eMBUN]O[3kNk0P2mMT1X1lLl0j1TNW1Q1nLk0f1\\NY1i0QMk0a1eNZ1a0UMi0\\1nN\\19XMj0V1UO_1a4\\NgKa1Y4_NjK^1V4aNnK\\1R4dNRLX1o3fNULW1k3hNYLU1g3kN\\LR1d3mN`LP1`3POcLl0_3ROfLj0Z3VOiLg0W3XOmLd0T3\\OoL`0R3_ORM>o2@VM;k2EXM8h2G[M6f2J\\M3e2L]M3d2L^M1c2N_M0b2OaMO_21cML^23dMK^24dMJ\\25fMI[27gMGY28iMFX2:jMCX2;jMDV2;mMBT2>b6O101N100O100O2N100O100O2O0O1O100O2O0O100O10_jW2"}, "label": "a woman holding a tennis racket"}]}
{"id": 273120, "image": "COCO_train2014_000000273120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman is playing tennis in the tennis court\"."}], "task": "refering", "answer_template": "The \"a woman is playing tennis in the tennis court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 80, 81, 82, 83, 84, 104, 105, 106, 126, 127, 128, 129, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 357, 358, 360, 361, 383, 384, 405, 406, 407], "bbox": [0.48896875, 0.12390438247011953, 0.770109375, 0.9602589641434263], "iscrowd": 0, "area": 30802.766200000002, "rle": {"size": [502, 640], "counts": "lbi41e?2M2N3M2N2O2M2N2cF@a3c0ZLAd3a0XLCQ2LPJa0l3Go1NRJ;m3Kl10TJ6n3Mj12UJ2n30h14WJLo34h12XJJm38j10VJIm3;l1NTJGn3?m1MQJEo3b0n1MoIBQ4d0o1MmI^OR4i0cM]NU2^1SL]OR4l0cM[NW2`1PLYOT4P1cMYNX2a1mKWOU4S1dMVNZ2d1hKTOW4V1eMSN\\2b2N^OcMSN]2`2MAdMPN_2_2JEeMnM`2\\2IJeMkMb2[2FNfMhMc2[2D1fMgMe2W2D5eMeMg2V2YOa0nM[Mh2S2oNP1VNnLj2S2dN\\1_NdLl2P2YNj1hNWLo2n1nMX2QOkKP3n1cMe2ZO_KR3l1XMR3CTKU3i1mLa3KhJW3g1bLn35\\JX3g1WL[4>oI[3P9cLQG]3Q9_LQGa3P9]LPGc3S9ZLmFg3T9WLlFj3V9RLlFn3]:00100O1O0UMRLPJo3n5TLoIm3o5WLnIk3o5YLoIh3o5\\LmIe3Q6_LlIc3R6`LlIa3Q6cLlI^3S6eLjI]3S6gLkIZ3S6iLjIX3T6mLhIU3V6nLhIR3V6RMgIo2X6TMeIm2Y6WMeIi2Z6ZMcIg2[6]MbId2]6_MaIa2]6cM`I]2`6fM]I[2a6iM]IW2n6`MoHa26[KS6j1dIl26_K`6Z1YIW34dKl6l0mHa34hKo6g0jHb34lKR7b0hHb33QLV7<dHd33ULY76bHf36ULX75_Hg39XLV71_Hh3:[LV7L]Hj3=^LT7i4lHZKR7d4nH`KP7_4oHeKo6Y4RIjKl6U4SIoKj6P4VISLi6l3VIXLh6g0lH0=d0]5XOXJ5:d0_5TOXJ89e0`5oNYJ=7e0`5jN[Ja05f0`5gN\\Jd03f0b5bN]Jh01g0c5]N^Jm0Of0d5ZN]JQ1Of0e5UN^JV1Lf0g5QN^JY1Kh0f5lMaJ]1Ig0g5hMbJa1Gh0h5dMbJe1Eh0j5_McJi1Dh0j5\\McJm1Bh0l5WMdJQ2@j0k5RMgJU2]Oj0`3RL`Nl0dNX2]Oj0Y3ZLeN`0gN]2ZOj0T3aLjN4jNa2XOk0n2hLoNIlNe2VOk0i2oLTO]OoNi2UOk0b2XMWOQOTOm2ROl0\\2^M]OcNXOT3oNl0V2eMBUN]O[3kNk0P2mMT1X1lLl0j1TNW1Q1nLk0f1\\NY1i0QMk0a1eNZ1a0UMi0\\1nN\\19XMj0V1UO_1a4\\NgKa1Y4_NjK^1V4aNnK\\1R4dNRLX1o3fNULW1k3hNYLU1g3kN\\LR1d3mN`LP1`3POcLl0_3ROfLj0Z3VOiLg0W3XOmLd0T3\\OoL`0R3_ORM>o2@VM;k2EXM8h2G[M6f2J\\M3e2L]M3d2L^M1c2N_M0b2OaMO_21cML^23dMK^24dMJ\\25fMI[27gMGY28iMFX2:jMCX2;jMDV2;mMBT2>b6O101N100O100O2N100O100O2O0O1O100O2O0O100O10_jW2"}, "label": "a woman is playing tennis in the tennis court"}]}
{"id": 559840, "image": "COCO_train2014_000000559840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown haired girl wearing a black and white shirt sitting at a table with food in front of her\"."}], "task": "refering", "answer_template": "The \"a brown haired girl wearing a black and white shirt sitting at a table with food in front of her\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 274, 275, 276, 277, 282, 283, 291, 292, 293], "bbox": [0.12827083333333333, 0.127734375, 0.8858958333333333, 0.785453125], "iscrowd": 0, "area": 75060.46900000001, "rle": {"size": [640, 480], "counts": "_VW1f0Xc0>C1N2O1N3N1N2O1N2O1N2O1N3N1N2O1O10000O2O00000O1000000O2O00000O1000000O2O00000O1000001OO01000O10O0100O10O001O00001O000000O10000O1OgMPNeBo1Z=UNdBj1[=ZNcBe1\\=_NbB`1]=dNaB\\1]=hNaBX1]=lNaBT1]=POaBP1]=TOaBm0\\=WObBi0\\=[ObBe0\\=_ObBa0\\=CbB=\\=GbB9]=JaB6]=NaB2]=2aBN]=6aBJ]=:aBF]=>aBPOo=T1nAnNo=V1oAkNn=Y1oAhNo=\\1oAeNn=_1oAcNn=a1PB`Nm=d1QB]Nl=g1QBZNm=j1QBWNl=m1QBUNl=o1RBRNk=R2RBPNl=R401000000O10O10O100000O01000O010000O0100000O01000O10O1000O10000O10000QNVJhEj5m9dJoE]5f9RKVFn4_9`K^F`4V9PLeFQ4Q9\\LlFd3S9aLiF_3U9gLfFZ3Y9jLdFV3Z9PMbFP3]9TM`Fl2^9ZM]Fg2b9]M[Fc2d9bMXF^2f9o3O1N2O1N2O1O1N20000O100O10000O100O10000O100O100O10000O100O10000O100O01000O100O10000O100O10000000000001O0000000000001O0000000000001OPG\\Ed8P;0001O000010O0001O000010O0001O00001O02N2O0O2N2N101N2N2O0O2N2O0O2N2O1kHgF_3Z9^LhFb3Z9ZLhFg3Y9ULiFk3Y9RLhFn3Y9oKiFQ4Y9kKiFT4Z9PK[FeN=[6Z9mJaF_N8c6Y9kJfFZN3j6Y9iJ_GW5c8fJ^GY5e8dJ\\G\\5f8`J\\G_5f8_J[Ga5g8\\JZGc5i8ZJXGf5j8VJYGh5o8PJRGP6U9hIlFX6[9_IgF`6a9XI`Fh6\\;O100O101N100O10000O100O101N100O100O2O1O0O2O0O2O1N101N2O0O2O0O2N2O0O2N2O0O2O0O2N2O0O2N2O0O2O0O2N2O0O2N2O0O2O0O2N2O1N3M2O1N2O2M2N2O1N3M2O1N2O2M2N2O1N3M6K4K6K4K6J5L5J5K5L5I6I8H7I8H7I7I8G8I8HcPR1"}, "label": "a brown haired girl wearing a black and white shirt sitting at a table with food in front of her"}]}
{"id": 559840, "image": "COCO_train2014_000000559840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child sitting at a table ready to eat wearing a black tshirt\"."}], "task": "refering", "answer_template": "The \"a child sitting at a table ready to eat wearing a black tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 274, 275, 276, 277, 282, 283, 291, 292, 293], "bbox": [0.12827083333333333, 0.127734375, 0.8858958333333333, 0.785453125], "iscrowd": 0, "area": 75060.46900000001, "rle": {"size": [640, 480], "counts": "_VW1f0Xc0>C1N2O1N3N1N2O1N2O1N2O1N3N1N2O1O10000O2O00000O1000000O2O00000O1000000O2O00000O1000001OO01000O10O0100O10O001O00001O000000O10000O1OgMPNeBo1Z=UNdBj1[=ZNcBe1\\=_NbB`1]=dNaB\\1]=hNaBX1]=lNaBT1]=POaBP1]=TOaBm0\\=WObBi0\\=[ObBe0\\=_ObBa0\\=CbB=\\=GbB9]=JaB6]=NaB2]=2aBN]=6aBJ]=:aBF]=>aBPOo=T1nAnNo=V1oAkNn=Y1oAhNo=\\1oAeNn=_1oAcNn=a1PB`Nm=d1QB]Nl=g1QBZNm=j1QBWNl=m1QBUNl=o1RBRNk=R2RBPNl=R401000000O10O10O100000O01000O010000O0100000O01000O10O1000O10000O10000QNVJhEj5m9dJoE]5f9RKVFn4_9`K^F`4V9PLeFQ4Q9\\LlFd3S9aLiF_3U9gLfFZ3Y9jLdFV3Z9PMbFP3]9TM`Fl2^9ZM]Fg2b9]M[Fc2d9bMXF^2f9o3O1N2O1N2O1O1N20000O100O10000O100O10000O100O100O10000O100O10000O100O01000O100O10000O100O10000000000001O0000000000001O0000000000001OPG\\Ed8P;0001O000010O0001O000010O0001O00001O02N2O0O2N2N101N2N2O0O2N2O0O2N2O1kHgF_3Z9^LhFb3Z9ZLhFg3Y9ULiFk3Y9RLhFn3Y9oKiFQ4Y9kKiFT4Z9PK[FeN=[6Z9mJaF_N8c6Y9kJfFZN3j6Y9iJ_GW5c8fJ^GY5e8dJ\\G\\5f8`J\\G_5f8_J[Ga5g8\\JZGc5i8ZJXGf5j8VJYGh5o8PJRGP6U9hIlFX6[9_IgF`6a9XI`Fh6\\;O100O101N100O10000O100O101N100O100O2O1O0O2O0O2O1N101N2O0O2O0O2N2O0O2N2O0O2O0O2N2O0O2N2O0O2O0O2N2O0O2N2O0O2O0O2N2O1N3M2O1N2O2M2N2O1N3M2O1N2O2M2N2O1N3M6K4K6K4K6J5L5J5K5L5I6I8H7I8H7I7I8G8I8HcPR1"}, "label": "a child sitting at a table ready to eat wearing a black tshirt"}]}
{"id": 527073, "image": "COCO_train2014_000000527073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a stuffed bear ' s head can be seen in the middle of two others\"."}], "task": "refering", "answer_template": "The \"the back of a stuffed bear ' s head can be seen in the middle of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 95, 96, 97, 98, 99, 113, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 170, 171], "bbox": [0.2879, 0.4010510510510511, 0.59662, 0.8203003003003003], "iscrowd": 0, "area": 13380.254649999999, "rle": {"size": [333, 500], "counts": "mo^13W:7J5J7I7J5J7I7J5J4M3N2O1O1O001O0O2O001O1O001N101O001O001N101O001O1O0O2O001O001O0O2O001O001O1N101O001O001N101O001O001N2O001O001O0O2O001O001O1N101O001O001N101O001O001N2O001O001O0010O100O010O10O10O100O010O100O010O10O010000O010O100O010O100O010O0100O010O0010O01O010ON3J5L4K6J5L5J5L4K6J5L4K6K4K6K4L4L5J5L4L5K4L5Kl`Q2"}, "label": "the back of a stuffed bear ' s head can be seen in the middle of two others"}]}
{"id": 527073, "image": "COCO_train2014_000000527073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear just past wagon writing on side\"."}], "task": "refering", "answer_template": "The \"teddy bear just past wagon writing on side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 95, 96, 97, 98, 99, 113, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 170, 171], "bbox": [0.2879, 0.4010510510510511, 0.59662, 0.8203003003003003], "iscrowd": 0, "area": 13380.254649999999, "rle": {"size": [333, 500], "counts": "mo^13W:7J5J7I7J5J7I7J5J4M3N2O1O1O001O0O2O001O1O001N101O001O001N101O001O1O0O2O001O001O0O2O001O001O1N101O001O001N101O001O001N2O001O001O0O2O001O001O1N101O001O001N101O001O001N2O001O001O0010O100O010O10O10O100O010O100O010O10O010000O010O100O010O100O010O0100O010O0010O01O010ON3J5L4K6J5L5J5L4K6J5L4K6K4K6K4L4L5J5L4L5K4L5Kl`Q2"}, "label": "teddy bear just past wagon writing on side"}]}
{"id": 551651, "image": "COCO_train2014_000000551651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is getting ready to surf in water\"."}], "task": "refering", "answer_template": "The \"a man is getting ready to surf in water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 21, 22, 23, 24, 38, 39, 40, 41, 54, 55, 56, 68, 69, 70, 71, 72, 73, 83, 84, 85, 86, 87, 88, 89, 90, 98, 99, 100, 102, 103, 104, 105, 106, 114, 115, 116, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 136, 137, 138, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 275, 276, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 328, 329, 330, 331, 339, 340, 341, 342, 344, 345, 346, 347, 360, 361], "bbox": [0.16160270880361174, 0.018109375, 0.7498194130925507, 0.968890625], "iscrowd": 0, "area": 85193.56724999998, "rle": {"size": [640, 443], "counts": "eV]1:bc0;]_O]OW=o0_BUO[=V1[BoN_=]1VBhNd=c1SB`Nh=k1nAZNm=Q2iASNQ>X2eAmMU>_2aAdMZ>g2\\A^M_>U4M3M2N3M3M3M2N3M1O1O1N3N1O1O1O2N1O1RIiI2X6XO_Jf0b5ZN`Kd1a4TNgKj1Z4nMPLo1Q4jMWLT2j3fM^LW2d3dMbLY2_3cMfL\\2Y3aMmL\\2S3aMRM]2o2_MWM^2[NVJS2Y3H^2TN\\JR2S3O_2nMaJQ2m27`2eMfJS2h2;a2aMjJR2c2a0a2\\MnJR2_2e0b2XMRKQ2Y2l0c2RMWKP2T2Q1d2nLZKP2P2U1e2jL^KP2j1Z1g2cLcKQ2c1`1i2^LfKQ2^1f1j2XLkKP2W1m1m2RLoKo1Q1T2o2kKRLP2l0Y2Q3fKVLo1e0`2U3_KXLP2`0f2V3YK]Lo19m2Z3QKaLQ2NV3_3hJeLQ2G]3d3`JhLQ2@d3g3ZJlLP2YOl3j3RJoLQ2SOR4n3kIRMc2ZNg3c4eIUMa:k2\\EYMd:e2[E^Md:b2ZE`Mf:_2YEdMf:\\2XEgMh:h504L:F:F:G9F5K5K3NK4K5L4K6K4K4L4M2M3N3L3[KiDa0Z;nMhDPOOk2\\;TNlDmNJh2^;XNQEkNCf2_;]NVEiN]Oc2a;aNZEhNWO_2c;hN^EeNPO\\2e;oNaEaNlNY2g;VObE]NjNU2g;@dEVNiNQ2g;IfERNeNm1h;3hEkMdNi1g;>iEeMdNd1g;g0iFPOZ9R1jFeNZ9[1kF\\NX9f1lFQNW9P2mFgMX9Z2U41O2O1N101UOj0_Mb2L3N3M2M4M3M2M4M3M3L4M3M3L4O2a0^Oa0_Oa0_O;E;E9H2M4L4L3M4L4L3N3L4L3M4L4L4M2M4L4M2M4M3L4M2N3L6K8G:G4L0O100O10000O2O0O1000Y@dNi<[1RCkNm<V1lBQOR=R1gBTOX=Q1]BVOa=P1UBVOj=o0lAWOS>o0aAXO]>W3O1N2O1O1N2O1N2O1O1N2O1aGnJm1S5mM[Kk1f4nMgKk1Z4nMTLi1n3PN`Lh1a3QNlLg1V3QNYMg1h2SNdMf1]2VNnMa1S2\\NXN\\1i1`NPOi0R1SO3KN17MJO;NGN>0CLa02AJd04CBb0<6iNOT1k0PNYOo1`1UMeNh2\\9N2M2O2M3N1O2M3N2M2O2M3N2M3N1N3N2N2M3N2M3N2M3N2M3N2M2O2M3N1N3M3N1N3N2M3N1N3M3N1N3N2M2N3N2M2O2M3N1NPjT2"}, "label": "a man is getting ready to surf in water"}]}
{"id": 551651, "image": "COCO_train2014_000000551651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player carrying a snow skating board\"."}], "task": "refering", "answer_template": "The \"a player carrying a snow skating board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 21, 22, 23, 24, 38, 39, 40, 41, 54, 55, 56, 68, 69, 70, 71, 72, 73, 83, 84, 85, 86, 87, 88, 89, 90, 98, 99, 100, 102, 103, 104, 105, 106, 114, 115, 116, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 136, 137, 138, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 275, 276, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 328, 329, 330, 331, 339, 340, 341, 342, 344, 345, 346, 347, 360, 361], "bbox": [0.16160270880361174, 0.018109375, 0.7498194130925507, 0.968890625], "iscrowd": 0, "area": 85193.56724999998, "rle": {"size": [640, 443], "counts": "eV]1:bc0;]_O]OW=o0_BUO[=V1[BoN_=]1VBhNd=c1SB`Nh=k1nAZNm=Q2iASNQ>X2eAmMU>_2aAdMZ>g2\\A^M_>U4M3M2N3M3M3M2N3M1O1O1N3N1O1O1O2N1O1RIiI2X6XO_Jf0b5ZN`Kd1a4TNgKj1Z4nMPLo1Q4jMWLT2j3fM^LW2d3dMbLY2_3cMfL\\2Y3aMmL\\2S3aMRM]2o2_MWM^2[NVJS2Y3H^2TN\\JR2S3O_2nMaJQ2m27`2eMfJS2h2;a2aMjJR2c2a0a2\\MnJR2_2e0b2XMRKQ2Y2l0c2RMWKP2T2Q1d2nLZKP2P2U1e2jL^KP2j1Z1g2cLcKQ2c1`1i2^LfKQ2^1f1j2XLkKP2W1m1m2RLoKo1Q1T2o2kKRLP2l0Y2Q3fKVLo1e0`2U3_KXLP2`0f2V3YK]Lo19m2Z3QKaLQ2NV3_3hJeLQ2G]3d3`JhLQ2@d3g3ZJlLP2YOl3j3RJoLQ2SOR4n3kIRMc2ZNg3c4eIUMa:k2\\EYMd:e2[E^Md:b2ZE`Mf:_2YEdMf:\\2XEgMh:h504L:F:F:G9F5K5K3NK4K5L4K6K4K4L4M2M3N3L3[KiDa0Z;nMhDPOOk2\\;TNlDmNJh2^;XNQEkNCf2_;]NVEiN]Oc2a;aNZEhNWO_2c;hN^EeNPO\\2e;oNaEaNlNY2g;VObE]NjNU2g;@dEVNiNQ2g;IfERNeNm1h;3hEkMdNi1g;>iEeMdNd1g;g0iFPOZ9R1jFeNZ9[1kF\\NX9f1lFQNW9P2mFgMX9Z2U41O2O1N101UOj0_Mb2L3N3M2M4M3M2M4M3M3L4M3M3L4O2a0^Oa0_Oa0_O;E;E9H2M4L4L3M4L4L3N3L4L3M4L4L4M2M4L4M2M4M3L4M2N3L6K8G:G4L0O100O10000O2O0O1000Y@dNi<[1RCkNm<V1lBQOR=R1gBTOX=Q1]BVOa=P1UBVOj=o0lAWOS>o0aAXO]>W3O1N2O1O1N2O1N2O1O1N2O1aGnJm1S5mM[Kk1f4nMgKk1Z4nMTLi1n3PN`Lh1a3QNlLg1V3QNYMg1h2SNdMf1]2VNnMa1S2\\NXN\\1i1`NPOi0R1SO3KN17MJO;NGN>0CLa02AJd04CBb0<6iNOT1k0PNYOo1`1UMeNh2\\9N2M2O2M3N1O2M3N2M2O2M3N2M3N1N3N2N2M3N2M3N2M3N2M3N2M2O2M3N1N3M3N1N3N2M3N1N3M3N1N3N2M2N3N2M2O2M3N1NPjT2"}, "label": "a player carrying a snow skating board"}]}
{"id": 35558, "image": "COCO_train2014_000000035558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman standing\"."}], "task": "refering", "answer_template": "The \"the woman standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 40, 41, 42, 56, 57, 58, 59, 74, 75, 76, 90, 91, 92, 93, 94, 106, 107, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 156, 157, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 262, 263, 264], "bbox": [0.22479166666666667, 0.065453125, 0.7185, 0.70071875], "iscrowd": 0, "area": 52235.484800000006, "rle": {"size": [640, 480], "counts": "mgS29Xc0`0@`0D=J5L5K4K5L4L4K5L4L4L4N2N2N2N2N2N2O1N2N2N6J5K6J3M2N2N2N100O2N1O001O3N2M4[AaKc=_5N3N1N2O1N2O1N2O2M2O1N2O1N2O1N3N1N3XNXITFi6i9[ITFh6S9TIoE8l0e6R9]IiE1R1e6R9cIbELY1c6S9UJkFm5R9VJkFl5S9WJjFl5S9VJkFl5R9XJlFj5Q9YJlFi5R9ZJkFi5R9ZJlFg5R9\\JkFg5R9\\JkFf5R9^JlFd5Q9h2M4L3M3M3L4M3N3M2N2N2N2M3N3M2N2N2N2N2N3M2M3N2N2N2N2N3M000O0100O100O100O100O2O000O100O2O0O10001N100O102N2M3N2eFRH`6o7\\ITHc6o7XIUHg6m7UIUHk6m7QIUHn6n7oHTHP7n7lHTHS7n7jHTHV7n7gHTHX7n7dHTH[7o7bHRH^7o7`HSH_7o7]HSHb7o7\\HRHd7P8XHSHg7n7QHYHn7j7hG^HX8c7`GeH^8^7ZGhHf8Y9O2O0O101N101N101N101N2O0O2N101N1lJ\\Dl1e;UN_Dd1b;]NcDSNGm1g;1fDiMHS2d;5hD_MIZ2`;8kDSMLb2[;<iECX:>hE_OZ:c0eE[O\\:f0dEWO^:j0cERO_:o0cEmN]:V1dEeN]:]1eE_N[:d1eEXN\\:k1eEQN[:S2gEfM[:\\2gE_M[:b2gEXM^:g2eESM^:m2dEmL`:S3bEhLb:W3aE]Lh:c3l201O2O1N3N1O1O2M5L<DM2O2N2M2O2N1O11O0010O01O010M7I<D<C<E;D5K4L4L4L4L3L2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1NQPd2"}, "label": "the woman standing"}]}
{"id": 35558, "image": "COCO_train2014_000000035558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue striped shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 40, 41, 42, 56, 57, 58, 59, 74, 75, 76, 90, 91, 92, 93, 94, 106, 107, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 156, 157, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 262, 263, 264], "bbox": [0.22479166666666667, 0.065453125, 0.7185, 0.70071875], "iscrowd": 0, "area": 52235.484800000006, "rle": {"size": [640, 480], "counts": "mgS29Xc0`0@`0D=J5L5K4K5L4L4K5L4L4L4N2N2N2N2N2N2O1N2N2N6J5K6J3M2N2N2N100O2N1O001O3N2M4[AaKc=_5N3N1N2O1N2O1N2O2M2O1N2O1N2O1N3N1N3XNXITFi6i9[ITFh6S9TIoE8l0e6R9]IiE1R1e6R9cIbELY1c6S9UJkFm5R9VJkFl5S9WJjFl5S9VJkFl5R9XJlFj5Q9YJlFi5R9ZJkFi5R9ZJlFg5R9\\JkFg5R9\\JkFf5R9^JlFd5Q9h2M4L3M3M3L4M3N3M2N2N2N2M3N3M2N2N2N2N2N3M2M3N2N2N2N2N3M000O0100O100O100O100O2O000O100O2O0O10001N100O102N2M3N2eFRH`6o7\\ITHc6o7XIUHg6m7UIUHk6m7QIUHn6n7oHTHP7n7lHTHS7n7jHTHV7n7gHTHX7n7dHTH[7o7bHRH^7o7`HSH_7o7]HSHb7o7\\HRHd7P8XHSHg7n7QHYHn7j7hG^HX8c7`GeH^8^7ZGhHf8Y9O2O0O101N101N101N101N2O0O2N101N1lJ\\Dl1e;UN_Dd1b;]NcDSNGm1g;1fDiMHS2d;5hD_MIZ2`;8kDSMLb2[;<iECX:>hE_OZ:c0eE[O\\:f0dEWO^:j0cERO_:o0cEmN]:V1dEeN]:]1eE_N[:d1eEXN\\:k1eEQN[:S2gEfM[:\\2gE_M[:b2gEXM^:g2eESM^:m2dEmL`:S3bEhLb:W3aE]Lh:c3l201O2O1N3N1O1O2M5L<DM2O2N2M2O2N1O11O0010O01O010M7I<D<C<E;D5K4L4L4L4L3L2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1NQPd2"}, "label": "a woman wearing a blue striped shirt"}]}
{"id": 232167, "image": "COCO_train2014_000000232167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer screen\"."}], "task": "refering", "answer_template": "The \"a white computer screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176], "bbox": [0.39453125, 0.08985416666666667, 0.673828125, 0.4596458333333333], "iscrowd": 0, "area": 27151.5625, "rle": {"size": [480, 640], "counts": "Vif3>S>?A?A?A?A?A?A?A?A?A?K50010O0000000001O000000001O00000000001O000001O0001O00000000001O00000000001O00000000010O000000001O00000000001O00000000001O01O00000001O00000000001O00000000001O000001O0001O000000001O00000000001O00000000010O00000000001O00000000001O00000000001O01O00000001O000000001O00000000001O00000001O01O00000000001O00000000001O00000000010O0D<XOh0XOh0XOh0XOi0WO`mQ3"}, "label": "a white computer screen"}]}
{"id": 232167, "image": "COCO_train2014_000000232167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white monitor to the right in front of the white keyboard\"."}], "task": "refering", "answer_template": "The \"the white monitor to the right in front of the white keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176], "bbox": [0.39453125, 0.08985416666666667, 0.673828125, 0.4596458333333333], "iscrowd": 0, "area": 27151.5625, "rle": {"size": [480, 640], "counts": "Vif3>S>?A?A?A?A?A?A?A?A?A?K50010O0000000001O000000001O00000000001O000001O0001O00000000001O00000000001O00000000010O000000001O00000000001O00000000001O01O00000001O00000000001O00000000001O000001O0001O000000001O00000000001O00000000010O00000000001O00000000001O00000000001O01O00000001O000000001O00000000001O00000001O01O00000000001O00000000001O00000000010O0D<XOh0XOh0XOh0XOi0WO`mQ3"}, "label": "the white monitor to the right in front of the white keyboard"}]}
{"id": 232167, "image": "COCO_train2014_000000232167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer keyboard on top of a table\"."}], "task": "refering", "answer_template": "The \"a white computer keyboard on top of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 374, 375, 376, 377, 378, 379], "bbox": [0.267984375, 0.705625, 0.738203125, 0.9820208333333333], "iscrowd": 0, "area": 26803.875399999997, "rle": {"size": [480, 640], "counts": "mo`25k>5J5L5J6J5L5J6K5K4K6K5J5L5K2M3N2M2O2N2M3N1N3N2N2M3N1N10000O100O10000O101N10000O100O10000O100O10000O100O10000O100O1000000O10000000000O100000O1000O10000000O010000000000O01000000000O10O10000000000O100000000000O01000000000000O10000000000O0100000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O100000O01000000O10O1000O01000O01000O010000O010000O1000O010000O1000000O10O10O10000O10000O10O1000O10000O1000O010000O1O100O1O1O101N1O100O1O2N100O1O101N1O1O101N1O2O0O2N1O2O0O1O2O0O2N101N1O2O0O2O001N101O0O2O000O2O001N101O0O2O001N2O1O1N2O1N2O1O1N2O1O1N2O1O2M2O2N1N2O2N1N3N1O2M2O2N1N2Oh[^2"}, "label": "a white computer keyboard on top of a table"}]}
{"id": 232167, "image": "COCO_train2014_000000232167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white keyboard\"."}], "task": "refering", "answer_template": "The \"white keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 374, 375, 376, 377, 378, 379], "bbox": [0.267984375, 0.705625, 0.738203125, 0.9820208333333333], "iscrowd": 0, "area": 26803.875399999997, "rle": {"size": [480, 640], "counts": "mo`25k>5J5L5J6J5L5J6K5K4K6K5J5L5K2M3N2M2O2N2M3N1N3N2N2M3N1N10000O100O10000O101N10000O100O10000O100O10000O100O10000O100O1000000O10000000000O100000O1000O10000000O010000000000O01000000000O10O10000000000O100000000000O01000000000000O10000000000O0100000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O100000O01000000O10O1000O01000O01000O010000O010000O1000O010000O1000000O10O10O10000O10000O10O1000O10000O1000O010000O1O100O1O1O101N1O100O1O2N100O1O101N1O1O101N1O2O0O2N1O2O0O1O2O0O2N101N1O2O0O2O001N101O0O2O000O2O001N101O0O2O001N2O1O1N2O1N2O1O1N2O1O1N2O1O2M2O2N1N2O2N1N3N1O2M2O2N1N2Oh[^2"}, "label": "white keyboard"}]}
{"id": 404205, "image": "COCO_train2014_000000404205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a taller sheep in front of the two smaller sheep\"."}], "task": "refering", "answer_template": "The \"a taller sheep in front of the two smaller sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 92, 93, 94, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 254], "bbox": [0.0014375, 0.0012850467289719626, 0.08896875000000001, 0.7545093457943925], "iscrowd": 0, "area": 11753.1584, "rle": {"size": [428, 640], "counts": "P>d2U:j2iMc0]O1O2N1O2N2N101O001O1O001O0O2hKbI_1a6mMPJS2e6gLlIY3T801O0000001O0000001O00000000O1O2N4K4M4L3M3lKbKTOb4e0gNUL\\1e3jNXLY1c3lNYLY1`3mN]LV1]3QO_LR1[3TObLP1W3g4J7J5J6J7I6J6KP1oNmac7"}, "label": "a taller sheep in front of the two smaller sheep"}]}
{"id": 404205, "image": "COCO_train2014_000000404205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"adult sheep\"."}], "task": "refering", "answer_template": "The \"adult sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 92, 93, 94, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 254], "bbox": [0.0014375, 0.0012850467289719626, 0.08896875000000001, 0.7545093457943925], "iscrowd": 0, "area": 11753.1584, "rle": {"size": [428, 640], "counts": "P>d2U:j2iMc0]O1O2N1O2N2N101O001O1O001O0O2hKbI_1a6mMPJS2e6gLlIY3T801O0000001O0000001O00000000O1O2N4K4M4L3M3lKbKTOb4e0gNUL\\1e3jNXLY1c3lNYLY1`3mN]LV1]3QO_LR1[3TObLP1W3g4J7J5J6J7I6J6KP1oNmac7"}, "label": "adult sheep"}]}
{"id": 404205, "image": "COCO_train2014_000000404205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb with its tail up\"."}], "task": "refering", "answer_template": "The \"a lamb with its tail up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 127, 128, 129, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 269, 284, 285, 287, 288, 292], "bbox": [0.2775625, 0.3698831775700935, 0.7607968749999999, 0.8623598130841122], "iscrowd": 0, "area": 33891.71670000001, "rle": {"size": [428, 640], "counts": "UcZ2;P=3M2N1O2A_OoCd0o;CjC>U<>M3N2N1O2M3N2N1O1N3N1O1O1O100O100O100O101N100O1O100O1000100O00100O1O010O1O001O100O001O1O1O100M3M3N2N2O1N2O1N2O1O1O1^HSMc4n2[KSMe4?eHi1d2jMf47mHo1[2lMh4ORIV2T2lMj4HYI\\2k1nMb5S2\\JnMe5R2YJoMg5R2WJoMj5Q2TJPNl5P2SJQNn5o1PJRNP6o1mISNT6m1jITNV6m1gIUNZ6j1dIXN]6h1`IZN`6g1^IZNc6e1[I^Nd6c1YI^Ni6b1UI^Nk6c1^HROd7o0ZHQOg7o0XHPOi7R1THoNm7R1QHmNQ8S1nGmNR8U1lGjNV8V1hGjNZ8W1dGiN\\8Y1bGfN`8Z1_GeNc8\\1[GdNf8^1VGbNk8n22O1N1O2O1N1O2O1N2N1O2N2N101N2N1O2N2N13N3M2N3L3N2N3L3N3M1N2O1O1N101O1O1N2O1O1N2O002M3N2N1N3N2N2N1N3N2N2M4M3M3L5L3M2N1N10001N1000O010O10O01O100O010O1O010O100O1O1O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O0O2O1O1N2O1N2O1O1N2O1O1N2O1N2UNaGhN`8X1dGcN]8]1gG]N[8c1gGZNZ8f1hGWNY8i1jGRNX8o1iGnMX8R2jGjMY8U2jGgMX8X2jGeMW8]2jG^MY8b2hG[MZ8e2gGXMZ8i2hGRM[8n2R10O10000O10000O10000O101O0O10000O10000O10000O2O001N2M2N3NQMkFh1S9XNVGb1g8_N_G]1_8cNfGY1X8hNmGU1P8lNVHo0i7QO\\Hl0a7UOdHg0Z7ZOiHe0U7[OlHf0Q7ZOQIf0m6YOVIj0f6UO\\In0_6QOdIR1X6mNkIU1Q6jNQJY1k5eNXJ^1c5bN_Jb1\\5\\NgJg1U5XNnJj1n4UNVKl1f4RN_Ko1]4PNhKo1U4oMPLP2n3PNULm1j3TNZLh1f3XN]Le1c3[N`Lb1`3^NdL^1\\3bNgL[1Y3eNkLW1U3hNoLU1Q3kNRMR1m2oNWMk0k2UOXMf0j2ZOYMa0i2_O[M;g2E\\M6f2J]M1e2N`MKc25Z50O100N2O1O0O2O0010O101O0O4M5JSio1"}, "label": "a lamb with its tail up"}]}
{"id": 404205, "image": "COCO_train2014_000000404205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tredi bear animals in the ground\"."}], "task": "refering", "answer_template": "The \"a tredi bear animals in the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 127, 128, 129, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 269, 284, 285, 287, 288, 292], "bbox": [0.2775625, 0.3698831775700935, 0.7607968749999999, 0.8623598130841122], "iscrowd": 0, "area": 33891.71670000001, "rle": {"size": [428, 640], "counts": "UcZ2;P=3M2N1O2A_OoCd0o;CjC>U<>M3N2N1O2M3N2N1O1N3N1O1O1O100O100O100O101N100O1O100O1000100O00100O1O010O1O001O100O001O1O1O100M3M3N2N2O1N2O1N2O1O1O1^HSMc4n2[KSMe4?eHi1d2jMf47mHo1[2lMh4ORIV2T2lMj4HYI\\2k1nMb5S2\\JnMe5R2YJoMg5R2WJoMj5Q2TJPNl5P2SJQNn5o1PJRNP6o1mISNT6m1jITNV6m1gIUNZ6j1dIXN]6h1`IZN`6g1^IZNc6e1[I^Nd6c1YI^Ni6b1UI^Nk6c1^HROd7o0ZHQOg7o0XHPOi7R1THoNm7R1QHmNQ8S1nGmNR8U1lGjNV8V1hGjNZ8W1dGiN\\8Y1bGfN`8Z1_GeNc8\\1[GdNf8^1VGbNk8n22O1N1O2O1N1O2O1N2N1O2N2N101N2N1O2N2N13N3M2N3L3N2N3L3N3M1N2O1O1N101O1O1N2O1O1N2O002M3N2N1N3N2N2N1N3N2N2M4M3M3L5L3M2N1N10001N1000O010O10O01O100O010O1O010O100O1O1O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O0O2O1O1N2O1N2O1O1N2O1O1N2O1N2UNaGhN`8X1dGcN]8]1gG]N[8c1gGZNZ8f1hGWNY8i1jGRNX8o1iGnMX8R2jGjMY8U2jGgMX8X2jGeMW8]2jG^MY8b2hG[MZ8e2gGXMZ8i2hGRM[8n2R10O10000O10000O10000O101O0O10000O10000O10000O2O001N2M2N3NQMkFh1S9XNVGb1g8_N_G]1_8cNfGY1X8hNmGU1P8lNVHo0i7QO\\Hl0a7UOdHg0Z7ZOiHe0U7[OlHf0Q7ZOQIf0m6YOVIj0f6UO\\In0_6QOdIR1X6mNkIU1Q6jNQJY1k5eNXJ^1c5bN_Jb1\\5\\NgJg1U5XNnJj1n4UNVKl1f4RN_Ko1]4PNhKo1U4oMPLP2n3PNULm1j3TNZLh1f3XN]Le1c3[N`Lb1`3^NdL^1\\3bNgL[1Y3eNkLW1U3hNoLU1Q3kNRMR1m2oNWMk0k2UOXMf0j2ZOYMa0i2_O[M;g2E\\M6f2J]M1e2N`MKc25Z50O100N2O1O0O2O0010O101O0O4M5JSio1"}, "label": "a tredi bear animals in the ground"}]}
{"id": 404205, "image": "COCO_train2014_000000404205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lamb that is behind the other lamb . it is slightly to the upper lefthand side of the other\"."}], "task": "refering", "answer_template": "The \"the lamb that is behind the other lamb . it is slightly to the upper lefthand side of the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 167, 190, 191, 192, 213, 214, 215], "bbox": [0.144640625, 0.2400467289719626, 0.478296875, 0.6704439252336448], "iscrowd": 0, "area": 15181.60785, "rle": {"size": [428, 640], "counts": "aoV13Y=1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O>B2N1O2N1O2M2O1O2N2N3M4K2O1O001O0O2O001O00000O0100O100O1000000001O00001O0000001O00001O0000001N10001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M2N3Me1[NN2dMnG\\OT8`0[HSOg7i0dHnN^7P1fHlN\\7S1hHhNZ7X1jHbNX7^1kH]NW7b1iH]NY7c1^HdNd7\\1]HaNoN@R8o1PI_NdN4V8]1VIZOj6f0VIZOj6e0WI[Oh6f0YIYOg6g0YIXOh6g0YIYOg6g0YIYOg6g0YIYOf6g0\\IWOe6i0[IWOe6i0[IWOe6h0\\IXOc6i0]IWOc6i0]IVOhNlNY7m1QJVObNQO]7i1QJVO_NTO_7g1RJUO]NWO`7c1TJVOZNYOb7a1TJUOYN\\Oc7_1UJTOVN_Od7]1WJTOSN_Oh7]1UJEm5;TJCm5=SJAn5?TJ^On5b0SJ\\On5d0SJYOo5f0SJWOn5j0RJUOo5k0RJROP6m0RJPOP6P1QJnNo5S1RJkNo5T1SJiNo5W1QJcNU6]1PJYNU6f1i20000O10000O100O10000O100O10000OnD[N_:^2K3M0000000001O0001O0001O00001O001O001O001O001O001O001O001O001O00001O0O2O001O001O001O001K4K6K6L8K5K6J5J>C3M2N2N2N2Ne][4"}, "label": "the lamb that is behind the other lamb . it is slightly to the upper lefthand side of the other"}]}
{"id": 404205, "image": "COCO_train2014_000000404205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep standing the lawn backside\"."}], "task": "refering", "answer_template": "The \"a sheep standing the lawn backside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 167, 190, 191, 192, 213, 214, 215], "bbox": [0.144640625, 0.2400467289719626, 0.478296875, 0.6704439252336448], "iscrowd": 0, "area": 15181.60785, "rle": {"size": [428, 640], "counts": "aoV13Y=1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O>B2N1O2N1O2M2O1O2N2N3M4K2O1O001O0O2O001O00000O0100O100O1000000001O00001O0000001O00001O0000001N10001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M2N3Me1[NN2dMnG\\OT8`0[HSOg7i0dHnN^7P1fHlN\\7S1hHhNZ7X1jHbNX7^1kH]NW7b1iH]NY7c1^HdNd7\\1]HaNoN@R8o1PI_NdN4V8]1VIZOj6f0VIZOj6e0WI[Oh6f0YIYOg6g0YIXOh6g0YIYOg6g0YIYOg6g0YIYOf6g0\\IWOe6i0[IWOe6i0[IWOe6h0\\IXOc6i0]IWOc6i0]IVOhNlNY7m1QJVObNQO]7i1QJVO_NTO_7g1RJUO]NWO`7c1TJVOZNYOb7a1TJUOYN\\Oc7_1UJTOVN_Od7]1WJTOSN_Oh7]1UJEm5;TJCm5=SJAn5?TJ^On5b0SJ\\On5d0SJYOo5f0SJWOn5j0RJUOo5k0RJROP6m0RJPOP6P1QJnNo5S1RJkNo5T1SJiNo5W1QJcNU6]1PJYNU6f1i20000O10000O100O10000O100O10000OnD[N_:^2K3M0000000001O0001O0001O00001O001O001O001O001O001O001O001O001O00001O0O2O001O001O001O001K4K6K6L8K5K6J5J>C3M2N2N2N2Ne][4"}, "label": "a sheep standing the lawn backside"}]}
{"id": 396014, "image": "COCO_train2014_000000396014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow with a green marker on its ear\"."}], "task": "refering", "answer_template": "The \"a cow with a green marker on its ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 326, 327], "bbox": [0.094109375, 0.11152224824355972, 0.8194374999999999, 0.9695316159250585], "iscrowd": 0, "area": 83622.26204999998, "rle": {"size": [427, 640], "counts": "^\\i02X=2N2M2O1O1N3N1O1O1O1O1N101O1O1O1O2N1O1O1O1N2O2N1O1O2N1O1O2N1O2N1O1O1O2N1O1O1O1O1O100O2N100O1O010O1O100O1O1O1O001O001O002M2O1O1O2O0O2O1N4M3L3N2M3N2M2O2M2O2N1O1O1N101O1OO10eMnL\\JS3c5RMYJm2f5XMWJh2g5dMoI[2P6iMmIX2R6kMlIU2R6nMlIR2T6QNjIo1U6SNjIl1W6UNgIl1Y6TNgIk1Y6VNfIk1Y6WNfIi1Y6XNgIg1Y6ZNfIg1Z6YNfIf1Z6[NeIf1Z6[NfId1Z6^NeIb1Y6`NfI`1Z6aNfI^1Z6cNeI]1[6dNeIZ1\\6hNbIX1]6jNcIU1]6lNbIS1_6oNoHRN3n2n6QOlHVN3i2P7TOkHVN2f2S7GlH8T7KiH5V7Q3O1N3M4M3L2O1N2O1N1000O0100O100O10000O100jN^IdKb6o3oImKQ6R4SJkKm5U4WJgKi5Y4ZJcKg5]4\\J`Kd5`4c10000000000000001O00000000010O01O001O0010O01O001O010O1O010O1OO1O2O0000001O001EdGfK\\8Z4dGfK]8Y4dGfK\\8X4fGhK[8T4?01O00001O0100O010O100O100O1O100000O100000O10O100000000O01000000O100001O0000001O0000O1O1O100O10000O10000O102M2O1O0O10000O2O00000O10001O00O010O100001N1000001O1N3N0oGoJh7Q5WHSKf7m4ZHTK5LR7P5iHUK3MR7P5jHTK4L0Mf6S5VIUK4KJ2l6`5XIgJh6n500O2O0000000O1000001O0O100000000O1000001O000O10000000000000001O0000O10O10000000000000O1O10O1000O10000000O10000000001O0O01O1N2O100O010000000000000000000000000000000000000000000000000000O1N2O1O1O1O1O1O2O0O1O100O101N100O2O0O2N1O2M2O2M3N1O2M4M3L3N3L3N2N2M3N2M3N3M1O2N2O1O1N2O001N2O1O1N2O2N1N3M3M3M3M4L3L4M3M4L3M8H8G3O1N2N2O1N2O2M2N2O1N2N2O1N2N2M3N2N2B?M2N3M2N3M2N2O2M2O2M3M4M2M3N2M\\U`1"}, "label": "a cow with a green marker on its ear"}]}
{"id": 396014, "image": "COCO_train2014_000000396014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow with the green tag on it ' s left ear\"."}], "task": "refering", "answer_template": "The \"the cow with the green tag on it ' s left ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 326, 327], "bbox": [0.094109375, 0.11152224824355972, 0.8194374999999999, 0.9695316159250585], "iscrowd": 0, "area": 83622.26204999998, "rle": {"size": [427, 640], "counts": "^\\i02X=2N2M2O1O1N3N1O1O1O1O1N101O1O1O1O2N1O1O1O1N2O2N1O1O2N1O1O2N1O2N1O1O1O2N1O1O1O1O1O100O2N100O1O010O1O100O1O1O1O001O001O002M2O1O1O2O0O2O1N4M3L3N2M3N2M2O2M2O2N1O1O1N101O1OO10eMnL\\JS3c5RMYJm2f5XMWJh2g5dMoI[2P6iMmIX2R6kMlIU2R6nMlIR2T6QNjIo1U6SNjIl1W6UNgIl1Y6TNgIk1Y6VNfIk1Y6WNfIi1Y6XNgIg1Y6ZNfIg1Z6YNfIf1Z6[NeIf1Z6[NfId1Z6^NeIb1Y6`NfI`1Z6aNfI^1Z6cNeI]1[6dNeIZ1\\6hNbIX1]6jNcIU1]6lNbIS1_6oNoHRN3n2n6QOlHVN3i2P7TOkHVN2f2S7GlH8T7KiH5V7Q3O1N3M4M3L2O1N2O1N1000O0100O100O10000O100jN^IdKb6o3oImKQ6R4SJkKm5U4WJgKi5Y4ZJcKg5]4\\J`Kd5`4c10000000000000001O00000000010O01O001O0010O01O001O010O1O010O1OO1O2O0000001O001EdGfK\\8Z4dGfK]8Y4dGfK\\8X4fGhK[8T4?01O00001O0100O010O100O100O1O100000O100000O10O100000000O01000000O100001O0000001O0000O1O1O100O10000O10000O102M2O1O0O10000O2O00000O10001O00O010O100001N1000001O1N3N0oGoJh7Q5WHSKf7m4ZHTK5LR7P5iHUK3MR7P5jHTK4L0Mf6S5VIUK4KJ2l6`5XIgJh6n500O2O0000000O1000001O0O100000000O1000001O000O10000000000000001O0000O10O10000000000000O1O10O1000O10000000O10000000001O0O01O1N2O100O010000000000000000000000000000000000000000000000000000O1N2O1O1O1O1O1O2O0O1O100O101N100O2O0O2N1O2M2O2M3N1O2M4M3L3N3L3N2N2M3N2M3N3M1O2N2O1O1N2O001N2O1O1N2O2N1N3M3M3M3M4L3L4M3M4L3M8H8G3O1N2N2O1N2O2M2N2O1N2N2O1N2N2M3N2N2B?M2N3M2N3M2N2O2M2O2M3M4M2M3N2M\\U`1"}, "label": "the cow with the green tag on it ' s left ear"}]}
{"id": 396014, "image": "COCO_train2014_000000396014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color cow standing in the left side\"."}], "task": "refering", "answer_template": "The \"white color cow standing in the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188], "bbox": [0.0, 0.0044496487119437935, 0.1965625, 0.5893911007025762], "iscrowd": 0, "area": 21753.68825, "rle": {"size": [427, 640], "counts": "2Z1Q<d2\\Mj0VO00000000001O000000001O00000000001O000000010O00000000001O000000001O000000000QJWK\\3j4UKdLk4n5000O10000O1000000O10000O1000000O100cN]1H8gNhHVLa7i3Q1O1O1O1O1O1O1O1O2M2O1O1O1O1N2O1O1O1O1N2O1O1O1O11O7I8H7I7I8H7I8H7I7I8H7I7I8H7I8H7I7I8H0000O10000O100O100O100eNfJeJZ5W5UK]Jl4`5]KZJd4b5eKWJ\\4f5f1L4TLoJ[OT5:aKYOY5EV^f6"}, "label": "white color cow standing in the left side"}]}
{"id": 396014, "image": "COCO_train2014_000000396014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow that is standing to the left of a cow that is sitting down\"."}], "task": "refering", "answer_template": "The \"a cow that is standing to the left of a cow that is sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188], "bbox": [0.0, 0.0044496487119437935, 0.1965625, 0.5893911007025762], "iscrowd": 0, "area": 21753.68825, "rle": {"size": [427, 640], "counts": "2Z1Q<d2\\Mj0VO00000000001O000000001O00000000001O000000010O00000000001O000000001O000000000QJWK\\3j4UKdLk4n5000O10000O1000000O10000O1000000O100cN]1H8gNhHVLa7i3Q1O1O1O1O1O1O1O1O2M2O1O1O1O1N2O1O1O1O1N2O1O1O1O11O7I8H7I7I8H7I8H7I7I8H7I7I8H7I8H7I7I8H0000O10000O100O100O100eNfJeJZ5W5UK]Jl4`5]KZJd4b5eKWJ\\4f5f1L4TLoJ[OT5:aKYOY5EV^f6"}, "label": "a cow that is standing to the left of a cow that is sitting down"}]}
{"id": 486125, "image": "COCO_train2014_000000486125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gentleman wearing glasses tying a necktie\"."}], "task": "refering", "answer_template": "The \"a gentleman wearing glasses tying a necktie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 331], "bbox": [0.032046875, 0.09797647058823529, 0.48728125, 1.0], "iscrowd": 0, "area": 59350.14920000002, "rle": {"size": [425, 640], "counts": "YP97n<8I7J5K5K4L2N3M2M4M2N3N1O1N2O2N0O2O001O1N101O0O2O001O1N101O001N2O001O0O2O001O1O001O0O2O001O1O00001O0O101O001O0O101O00001N101O000000000001O01O01O001O000O2MnL_MaK_2a4cM]K[2e4fM[KX2g4iMWKX2h4kMUKU2X4aMaI>T2P2Y4kM]I6X2o1Y4PN\\I3Y2Q2W4nM^I3Z2S2R4nMbIO[2Y2m3jMfIO[2]2i3fMkIM\\2b2b3[N]Lk1]3VNbLo1Y3RNgLQ2U3QNiLR2T3TNfLo1W3VNeLl1X3YNdLj1X3ZNdLi1Y3\\NcLf1Z3^NcLd1gNRLQ4]2VMU2g2mMWMU2g2nMWMT2f2oMXMR2f2PNXMS2f2oMXMS2e2PNYMQ2f2PNYMR2d2QNYMS2d2mM\\MU2DaJZ2\\3QNU2_OiJ=1h0R3ZOW2[OnJ91o0l2\\Oo2^O[JV1f2\\OQ3XO]J[1d2[OP4e0PL[OP4d0RL[On3e0SLYOn3f0VLWOj3i0XLUOh3l0YLROg3n0[LoNf3Q1`LhNa3X1X4O1O10000001O1O00001O1OO1000000000000001O00001O00001O001O1O001O001O0000001O0000001O00001O00001O2N1O1O2N1O1O1O5K1[GlMS6T2kIPNS6Q2kIRNS6o1jIUNT6k1cIhMYNa0Q8h1cIdN[6]1cIfN[6[1cIiNZ6X1dIkNZ6W1bImN\\6U1`ImN`6U1\\ImNd6\\1^HUM6b1\\7j1\\HYNd7k1THYNl7b30000000001O001O0000000000001O00000000001O0000000000001O00000000001O00001O001O00001O001O001aN[JXKe5d4`J[Ka5X4lJgKT5X4nJgKS5W4QKfKP5X4TKeKl4Z4XKcKi4[4[KbKe4]4_K`Kb4^4bK_K_4_4jKYKW4d4PLWKQ4f4TLWKn3e4VLYKk3d4a2K4K6J5N3N1O2M2O2N1N3N1O1N3N3M2M3M3N2M4K4M3L4M3L4L4L4L4L6I6K5J7J5K6I8I8G;DYjW4"}, "label": "a gentleman wearing glasses tying a necktie"}]}
{"id": 486125, "image": "COCO_train2014_000000486125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man trying to wear a tie\"."}], "task": "refering", "answer_template": "The \"the man trying to wear a tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 331], "bbox": [0.032046875, 0.09797647058823529, 0.48728125, 1.0], "iscrowd": 0, "area": 59350.14920000002, "rle": {"size": [425, 640], "counts": "YP97n<8I7J5K5K4L2N3M2M4M2N3N1O1N2O2N0O2O001O1N101O0O2O001O1N101O001N2O001O0O2O001O1O001O0O2O001O1O00001O0O101O001O0O101O00001N101O000000000001O01O01O001O000O2MnL_MaK_2a4cM]K[2e4fM[KX2g4iMWKX2h4kMUKU2X4aMaI>T2P2Y4kM]I6X2o1Y4PN\\I3Y2Q2W4nM^I3Z2S2R4nMbIO[2Y2m3jMfIO[2]2i3fMkIM\\2b2b3[N]Lk1]3VNbLo1Y3RNgLQ2U3QNiLR2T3TNfLo1W3VNeLl1X3YNdLj1X3ZNdLi1Y3\\NcLf1Z3^NcLd1gNRLQ4]2VMU2g2mMWMU2g2nMWMT2f2oMXMR2f2PNXMS2f2oMXMS2e2PNYMQ2f2PNYMR2d2QNYMS2d2mM\\MU2DaJZ2\\3QNU2_OiJ=1h0R3ZOW2[OnJ91o0l2\\Oo2^O[JV1f2\\OQ3XO]J[1d2[OP4e0PL[OP4d0RL[On3e0SLYOn3f0VLWOj3i0XLUOh3l0YLROg3n0[LoNf3Q1`LhNa3X1X4O1O10000001O1O00001O1OO1000000000000001O00001O00001O001O1O001O001O0000001O0000001O00001O00001O2N1O1O2N1O1O1O5K1[GlMS6T2kIPNS6Q2kIRNS6o1jIUNT6k1cIhMYNa0Q8h1cIdN[6]1cIfN[6[1cIiNZ6X1dIkNZ6W1bImN\\6U1`ImN`6U1\\ImNd6\\1^HUM6b1\\7j1\\HYNd7k1THYNl7b30000000001O001O0000000000001O00000000001O0000000000001O00000000001O00001O001O00001O001O001aN[JXKe5d4`J[Ka5X4lJgKT5X4nJgKS5W4QKfKP5X4TKeKl4Z4XKcKi4[4[KbKe4]4_K`Kb4^4bK_K_4_4jKYKW4d4PLWKQ4f4TLWKn3e4VLYKk3d4a2K4K6J5N3N1O2M2O2N1N3N1O1N3N3M2M3M3N2M4K4M3L4M3L4L4L4L4L6I6K5J7J5K6I8I8G;DYjW4"}, "label": "the man trying to wear a tie"}]}
{"id": 486125, "image": "COCO_train2014_000000486125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair with arms between a table and dresser\"."}], "task": "refering", "answer_template": "The \"a black chair with arms between a table and dresser\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 265, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340], "bbox": [0.5605781249999999, 0.7042823529411765, 0.8286562499999999, 1.0], "iscrowd": 0, "area": 12786.495600000002, "rle": {"size": [425, 640], "counts": "\\Ze46m<:G8I7L3N1N3aElNd8W1YGjNg8W1WGoNc8T1YGoNf8R1XGQOhNCh9_1]GTOb8n0\\GSOd8o0ZGQOe8Q1ZGoNf8S1XGmNg8W1VGiNj8Z1SGfNl8^1RGbNl8b1QG^No8e1nF[NQ9i1lFWNT9l1iFTNV9R2eFnMZ9Z2_FfMa9Q30000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O11O0000000000000000001O00000000000000001O000000001O1O1O1O1O1O1O0nN[F`Nf9]1\\FVN2Dc9S2]FSN5K_9o1^FVN3K`9l1_FXN1Ma9d0YF36\\O0Mb9IaG=oNMMNo;2QDNo;2PDOP<1PDOP<1oC0Q<0oC1P<OPD1P<OoC2P<NQD2d;D\\D;O2d;FZD921c;IYD632d;9\\DGc;9]DHb;9^DGa;9`DGa;7`DI`;6bDI^;5dDK\\;4eDL[;2gDNY;0iD0W;0iD0X;OhD1X;OhD1X;OiD0W;0iD0W;0jDOV;0lDOT;1lDOU;0lDOT;1mDNS;2nDMR;3nDMR;4nDKR;6nDIR;8mDHT;8lDGT;9mDFS;;lDET;<lDCT;>lDAT;`0kD@U;a0kD^OU;b0lD]OU;b0kD^OU;c0kD\\OU;d0lD[OT;e0lD[OU;d0h010O1O00100O001O10O01N2M2O2N2M2O2N2MhX]1"}, "label": "a black chair with arms between a table and dresser"}]}
{"id": 486125, "image": "COCO_train2014_000000486125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black computer chair\"."}], "task": "refering", "answer_template": "The \"a black computer chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 265, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340], "bbox": [0.5605781249999999, 0.7042823529411765, 0.8286562499999999, 1.0], "iscrowd": 0, "area": 12786.495600000002, "rle": {"size": [425, 640], "counts": "\\Ze46m<:G8I7L3N1N3aElNd8W1YGjNg8W1WGoNc8T1YGoNf8R1XGQOhNCh9_1]GTOb8n0\\GSOd8o0ZGQOe8Q1ZGoNf8S1XGmNg8W1VGiNj8Z1SGfNl8^1RGbNl8b1QG^No8e1nF[NQ9i1lFWNT9l1iFTNV9R2eFnMZ9Z2_FfMa9Q30000000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O11O0000000000000000001O00000000000000001O000000001O1O1O1O1O1O1O0nN[F`Nf9]1\\FVN2Dc9S2]FSN5K_9o1^FVN3K`9l1_FXN1Ma9d0YF36\\O0Mb9IaG=oNMMNo;2QDNo;2PDOP<1PDOP<1oC0Q<0oC1P<OPD1P<OoC2P<NQD2d;D\\D;O2d;FZD921c;IYD632d;9\\DGc;9]DHb;9^DGa;9`DGa;7`DI`;6bDI^;5dDK\\;4eDL[;2gDNY;0iD0W;0iD0X;OhD1X;OhD1X;OiD0W;0iD0W;0jDOV;0lDOT;1lDOU;0lDOT;1mDNS;2nDMR;3nDMR;4nDKR;6nDIR;8mDHT;8lDGT;9mDFS;;lDET;<lDCT;>lDAT;`0kD@U;a0kD^OU;b0lD]OU;b0kD^OU;c0kD\\OU;d0lD[OT;e0lD[OU;d0h010O1O00100O001O10O01N2M2O2N2M2O2N2MhX]1"}, "label": "a black computer chair"}]}
{"id": 264944, "image": "COCO_train2014_000000264944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white old fashioned style river boat with decorative yellow edging\"."}], "task": "refering", "answer_template": "The \"a white old fashioned style river boat with decorative yellow edging\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 62, 79, 80, 81, 82, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 245, 246, 247, 258, 259, 260, 261, 263, 264, 265, 268, 269, 270, 282, 283, 284, 286, 287, 288, 291, 292], "bbox": [0.21364062499999997, 0.12154566744730678, 0.7616406250000001, 0.8741686182669789], "iscrowd": 0, "area": 58168.255600000004, "rle": {"size": [427, 640], "counts": "Q\\i1<=FS<:iCI[;4PE4_OMZ;6VE;i:GVE9j:GVE9j:HUE7k:JUE6k:KkD>U;l0O010O1000O06K4L5J6K5K5J7J5K6I6K4K2O001N2O001N2O001N2O001N2O0O2O001N2O001N2O001N2O001N2O0O2O1O0O2O0O2O1N101O1N101N2O001N2O0O2O1ZOgJXIY5^6iJPI35L3Z5e6RKVIMMR5j6SKXI[5h6fJVI\\5h6d0O2O1N101N2O001O10O0M3M3M4K4M3M4K4M3WNfH^M]7b2nHoLW7P3TIbLo6^3\\ITLg6l3]1O0O101O0000001O00001O00DVLVGk3j8;O100O1O100O1O1O0I8H8I7J6KVK]HV4^7[KWId4j7N1O1O0000O1000000O10O0100O1O1lNkKRIU4P7R16J5K6J3M3M3M3M3L4M3M4L1O00]Oc0<EnM[KdKm4_3SKcJ3o1X5k2kJQK0T2c5_2gJYKBS2W6X2`K`Lj4U3f2GnFZLR9g392M2ON1M4L42M3N0cFXLU9h3iFZLX9d3fF`L9O_8\\40000O010O0100O001O00O1O1O2N1O1O1N2O2N1O1O1WNYKmJi4R5`KdJa4[5hK\\JY4c5QLSJP4l5YLkIi3S6aLcI`3\\6h1O1O2N1O2N3M3N0O1O001O2N1O2N2N2O1O001O1O1O1O001O1O1O001O1O1O1O1OCiJhHU5X7SKaHl4^7YK_Hf4`7c0O04M4LO1N2O1O1O09H6J4ZKlGm3l8M001N1000001O000O2OO100O1O1O001001O2N1O1OO001O2M2O2N2M2O2N7I6I8I6J7I6I8I6J7H7J7I6J7H7J2N1O0O2O1O0O2O001O0O2O^MgJSMY5b2TK]Mj4_2]K_Mc4\\2cKdM\\4W2kKgMU4T2RLkMl3P2[LoMe3l1bLSN]3h1jLVNV3e1PM[Nn2a1YM]Ng2]1aMbN^2W1jMhNU2S1SNlNl1m0]NROb1g0gNWOY1c0nN]OP1=YOAg08BG=3JL5N41KH>6BDe0<ZO]On0>UO\\OS19\\O\\Ol0b0o5O101N1O101N1O100O2N100O2N100O3MX`o1"}, "label": "a white old fashioned style river boat with decorative yellow edging"}]}
{"id": 264944, "image": "COCO_train2014_000000264944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow and white ferry at the boat dock\"."}], "task": "refering", "answer_template": "The \"a yellow and white ferry at the boat dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 62, 79, 80, 81, 82, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 245, 246, 247, 258, 259, 260, 261, 263, 264, 265, 268, 269, 270, 282, 283, 284, 286, 287, 288, 291, 292], "bbox": [0.21364062499999997, 0.12154566744730678, 0.7616406250000001, 0.8741686182669789], "iscrowd": 0, "area": 58168.255600000004, "rle": {"size": [427, 640], "counts": "Q\\i1<=FS<:iCI[;4PE4_OMZ;6VE;i:GVE9j:GVE9j:HUE7k:JUE6k:KkD>U;l0O010O1000O06K4L5J6K5K5J7J5K6I6K4K2O001N2O001N2O001N2O001N2O0O2O001N2O001N2O001N2O001N2O0O2O1O0O2O0O2O1N101O1N101N2O001N2O0O2O1ZOgJXIY5^6iJPI35L3Z5e6RKVIMMR5j6SKXI[5h6fJVI\\5h6d0O2O1N101N2O001O10O0M3M3M4K4M3M4K4M3WNfH^M]7b2nHoLW7P3TIbLo6^3\\ITLg6l3]1O0O101O0000001O00001O00DVLVGk3j8;O100O1O100O1O1O0I8H8I7J6KVK]HV4^7[KWId4j7N1O1O0000O1000000O10O0100O1O1lNkKRIU4P7R16J5K6J3M3M3M3M3L4M3M4L1O00]Oc0<EnM[KdKm4_3SKcJ3o1X5k2kJQK0T2c5_2gJYKBS2W6X2`K`Lj4U3f2GnFZLR9g392M2ON1M4L42M3N0cFXLU9h3iFZLX9d3fF`L9O_8\\40000O010O0100O001O00O1O1O2N1O1O1N2O2N1O1O1WNYKmJi4R5`KdJa4[5hK\\JY4c5QLSJP4l5YLkIi3S6aLcI`3\\6h1O1O2N1O2N3M3N0O1O001O2N1O2N2N2O1O001O1O1O1O001O1O1O001O1O1O1O1OCiJhHU5X7SKaHl4^7YK_Hf4`7c0O04M4LO1N2O1O1O09H6J4ZKlGm3l8M001N1000001O000O2OO100O1O1O001001O2N1O1OO001O2M2O2N2M2O2N7I6I8I6J7I6I8I6J7H7J7I6J7H7J2N1O0O2O1O0O2O001O0O2O^MgJSMY5b2TK]Mj4_2]K_Mc4\\2cKdM\\4W2kKgMU4T2RLkMl3P2[LoMe3l1bLSN]3h1jLVNV3e1PM[Nn2a1YM]Ng2]1aMbN^2W1jMhNU2S1SNlNl1m0]NROb1g0gNWOY1c0nN]OP1=YOAg08BG=3JL5N41KH>6BDe0<ZO]On0>UO\\OS19\\O\\Ol0b0o5O101N1O101N1O100O2N100O2N100O3MX`o1"}, "label": "a yellow and white ferry at the boat dock"}]}
{"id": 305905, "image": "COCO_train2014_000000305905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two stacks of suitcases on a table\"."}], "task": "refering", "answer_template": "The \"two stacks of suitcases on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 282], "bbox": [0.09253125, 0.4402122641509434, 0.47682812499999994, 0.8168867924528302], "iscrowd": 0, "area": 28570.34835, "rle": {"size": [424, 640], "counts": "`fh04h<<D<J7N101N1O101N1O2N100iNeN`F]1^9ROSFo0m9^OeEd0Z:C_E=a:U1O2O001O0O101O001N10001O001N1000^NiE4V:_1O_NkE2T:a1000^NlE1R:c10^NnEOQ:1oEOQ:d1O_NPFMo9f10^NQFLn94RFLn95RFJm97SFJl96TFJl97TFIk97UFIj99VFFj9:VFGi99XFFh9;WFDi9=WFCi9=XFBh9?WFAi9?XF_Oh9c0WF]Oi9c0XF\\Oh9d0XF[Oi9f0VFZOj9f0WFYOh9h0XFWOi9j0WFUOi9k0WFUOi9l0VFSOj9n0WFQOi9o0WFQOi9P1WFoNi9Q1WFnNj9R1VFnNj9S1UFmNk9S1UFlNm9S1SFmNm9S1SFmNm9S1TFkNm9U1SFkNm9U1SFaNW:_1iElMW18S8i1fGeNY8Z1[GTOd8^1iFcNW9]1iFdNV9]1jFcNU9^1jFbNW9]1jFcNU9^1kFbNT9lNWGQ2DSOU9kNoF[2MjNS9`1lF`NT9a1lF_NT9`1mF`NR9a1mF_NS9b1mF^NR9b1nF_NQ9b1oF]NQ9iNnFi21^NQ9c1oF^NP9c1PG\\NP9e1oF\\NP9d1QG\\Nn8e1RGZNn8g1QGZNn8f1SGZNl8g1SGYNm8h1SGXNk8i1VGWNi8j1VGVNj8k1VGUNi8k1WGVNh8k1XGTNh8m1XGSNg8m1YGTNf8m1ZGRNf8Y3NgN\\GQNc8P2]GoMc8\\3O1O001O1O0O1000O10000O10000O0VM`Gn0`8m10O10000O1VM^Gn0b8l100O0100VM^Gn0b8k110000O1WM\\Gn0d8RO\\Gn0c8k110000WM\\Gm0e8SO[Gm0e8SO[Gm0e8j1100XMZGn0e8SO[Gl0f8TOZGl0f8TOZGl0f8j11YMYGm0g8SOYGm0g8SOYGl0h8TOXGl0g8k10YMYGl0h8TOXGl0h8TOXGl0h8TOXGl0h8TOXGk0i8TOWGm0i8SOWGm0i8SOWGl0i8UOWGk0i8UOWGk0i8TOXGk0i8UOVGl0j8TOVGl0j8TOVGk0k8UOUGk0k8UOUGk0k8TOVGk0k8UOTGl0l8TOTGl0m8SOSGl0n8TORGl0n8SOSGm0n8ROQGn0P9ROPGn0P9ROPGn0Q9QOoFn0R9QOoFo0Q9QOoFo0R9POmFP1T9POlFP1U9oNkFQ1U9nNlFQ1U9oNkFQ1V9nNjFR1V9nNiFS1W9mNiFR1Y9lNhFT1X9lNhFT1X9lNhFS1Z9lNfFT1Z9lNeFU1[9jNfFU1\\9jNdFV1\\9jNdFV1\\9jNdFU1^9jNbFV1^9iNbFX1^9hNbFW1`9hN`FX1`9hN`FX1`9hN`FW1b9gN^FZ1b9fN^FZ1c9eN]FZ1d9fN\\FZ1d9fN\\FZ1e9dN\\F[1e9eNZF\\1f9\\NbFd1_9kMQGT2P:21O0O10001N1iMoEW1Q;01O0O2O001N101OX^Z4"}, "label": "two stacks of suitcases on a table"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut touching two another and partly visible\"."}], "task": "refering", "answer_template": "The \"a donut touching two another and partly visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 48, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 95, 96], "bbox": [0.1711764705882353, 0.03715686274509804, 0.5765686274509804, 0.19031045751633988], "iscrowd": 0, "area": 16768.87225, "rle": {"size": [612, 612], "counts": "lhn18lb09G1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O00000000000000O10000000000O100000000000000O1000O10000000000000O10000000000000000O1000000000000000000O10000000000000000O1000000000000000000O100000000O1000000O10000O10000O1000000O10000O10000O1000000O100M3N2M3M3N2M3O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2M3N2N2M3Nk[k4"}, "label": "a donut touching two another and partly visible"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one honey glazed donut behind two more honey glazed donuts\"."}], "task": "refering", "answer_template": "The \"one honey glazed donut behind two more honey glazed donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 48, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 95, 96], "bbox": [0.1711764705882353, 0.03715686274509804, 0.5765686274509804, 0.19031045751633988], "iscrowd": 0, "area": 16768.87225, "rle": {"size": [612, 612], "counts": "lhn18lb09G1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O00000000000000O10000000000O100000000000000O1000O10000000000000O10000000000000000O1000000000000000000O10000000000000000O1000000000000000000O100000000O1000000O10000O10000O1000000O10000O10000O1000000O100M3N2M3M3N2M3O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2M3N2N2M3Nk[k4"}, "label": "one honey glazed donut behind two more honey glazed donuts"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut to the top right of the chocolate donut\"."}], "task": "refering", "answer_template": "The \"the donut to the top right of the chocolate donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 63, 77, 78, 79, 80, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193], "bbox": [0.5, 0.04279411764705882, 0.8896405228758171, 0.3963888888888889], "iscrowd": 0, "area": 40682.28365, "rle": {"size": [612, 612], "counts": "bof5b1`a04K5L3M4K4M4K4M4L3L5L3L5L3M4K4M4L4M2N3L3N3L3N3M2M4M2N3L3N3M2M4M3L3N3M2M4M2N3L2O2M2O2O000O2O00001O001O0O101O001O00001O0O2O00001O001O000O2O001O00001O001O000O2O001O00001O00001O001O00001O001O00001O00001O00000000000000000000001O00000O100000000000001O000000000001O00000000010O00000000001O00000001OO2O000O100O101O0O100O10001N100O10000O101N10000O100O2O000O100O101O0O100O1N2O2N1N2O1O1N2O2N1N2O1O1N3N2N1O2M3N2N1N3N2N1N3N2N2M2O2N2N2M2M4H8I6K6L4L4M3L5K6J6K4K6J5L5J6J5L5J5K6J5L5J6J5L5J5K]aX1"}, "label": "the donut to the top right of the chocolate donut"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut to the right side\"."}], "task": "refering", "answer_template": "The \"the donut to the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 63, 77, 78, 79, 80, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193], "bbox": [0.5, 0.04279411764705882, 0.8896405228758171, 0.3963888888888889], "iscrowd": 0, "area": 40682.28365, "rle": {"size": [612, 612], "counts": "bof5b1`a04K5L3M4K4M4K4M4L3L5L3L5L3M4K4M4L4M2N3L3N3L3N3M2M4M2N3L3N3M2M4M3L3N3M2M4M2N3L2O2M2O2O000O2O00001O001O0O101O001O00001O0O2O00001O001O000O2O001O00001O001O000O2O001O00001O00001O001O00001O001O00001O00001O00000000000000000000001O00000O100000000000001O000000000001O00000000010O00000000001O00000001OO2O000O100O101O0O100O10001N100O10000O101N10000O100O2O000O100O101O0O100O1N2O2N1N2O1O1N2O2N1N2O1O1N3N2N1O2M3N2N1N3N2N1N3N2N2M2O2N2N2M2M4H8I6K6L4L4M3L5K6J6K4K6J5L5J6J5L5J5K6J5L5J6J5L5J5K]aX1"}, "label": "the donut to the right side"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate donut\"."}], "task": "refering", "answer_template": "The \"a chocolate donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 206, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235], "bbox": [0.357140522875817, 0.33875816993464053, 0.7019281045751634, 0.4966503267973856], "iscrowd": 0, "area": 11088.416000000003, "rle": {"size": [612, 612], "counts": "WTS43ob04L4L4M3M2N1O100O2N1O1O1O101N1O1O1O1O2M2O1O1O1O2N1O1O1O101N1O1O100O2N100O1O1O2O000O100O2O000O100O2O000000001O000000001O0O1000001O000000001O000000001O000000001N100000010O000000001O0000000010O00000001O000001O010O001O0010O01O001O010O001O00010O001O00010O001O00010O001O00010O001O0010O0001O0010O0001O0010O01O00100O1O001O100O001O1O010O1O1O1O101N1O1O1O100O1O1O1O2O0O1O1O1O101N1O1O1O101N1O1O1O101N1O1O1O101N1O2N2N2O1N3M2N2N2O1N_R]3"}, "label": "a chocolate donut"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate glazed doughnut behind a cellphone\"."}], "task": "refering", "answer_template": "The \"a chocolate glazed doughnut behind a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 206, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235], "bbox": [0.357140522875817, 0.33875816993464053, 0.7019281045751634, 0.4966503267973856], "iscrowd": 0, "area": 11088.416000000003, "rle": {"size": [612, 612], "counts": "WTS43ob04L4L4M3M2N1O100O2N1O1O1O101N1O1O1O1O2M2O1O1O1O2N1O1O1O101N1O1O100O2N100O1O1O2O000O100O2O000O100O2O000000001O000000001O0O1000001O000000001O000000001O000000001N100000010O000000001O0000000010O00000001O000001O010O001O0010O01O001O010O001O00010O001O00010O001O00010O001O00010O001O0010O0001O0010O0001O0010O01O00100O1O001O100O001O1O010O1O1O1O101N1O1O1O100O1O1O1O2O0O1O1O1O101N1O1O1O101N1O1O1O101N1O1O1O101N1O2N2N2O1N3M2N2N2O1N_R]3"}, "label": "a chocolate glazed doughnut behind a cellphone"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut on the left side\"."}], "task": "refering", "answer_template": "The \"the donut on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 72, 88, 89, 90, 91, 92, 93, 94, 95, 110, 111, 112, 113, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 159, 160, 161, 162, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 204, 205, 221, 222, 223, 244], "bbox": [0.040539215686274506, 0.13851307189542483, 0.3963888888888889, 0.514640522875817], "iscrowd": 0, "area": 34505.8376, "rle": {"size": [612, 612], "counts": "cQ?V5m=3N2M2O2M2O2N1N3N2M101N101O0O2O0O2O1N101O0O2O0O2O0O2O001N2O0O2O0O2O001N101N101N2O001N101N101N101O0O2O1OO1N1N3M3N2M3N2M3N2M2N3N2M3N2N2M3N2M2O2N2M3N2N2M30O0100O100O10000O100O100O100O10000O2O0O100O10000O100O100O10000O10010O00000001O0000001O000000001O0001O0001O01O0010O01O000010O01O001O010O001O010O00001O010O001O0010O01O000010O01O0010O01O001O010O000O2N1O2N1O2N2M3N3M2M4M2N3M2M4M2N3L3N3M2M4M2N3L3N3M2M4M2N3L3N3L5L4K6K4K6K4K6K4K5L5J5L5J5LYnl6"}, "label": "the donut on the left side"}]}
{"id": 35571, "image": "COCO_train2014_000000035571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glazed donut just to the left of phone\"."}], "task": "refering", "answer_template": "The \"glazed donut just to the left of phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 72, 88, 89, 90, 91, 92, 93, 94, 95, 110, 111, 112, 113, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 159, 160, 161, 162, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 204, 205, 221, 222, 223, 244], "bbox": [0.040539215686274506, 0.13851307189542483, 0.3963888888888889, 0.514640522875817], "iscrowd": 0, "area": 34505.8376, "rle": {"size": [612, 612], "counts": "cQ?V5m=3N2M2O2M2O2N1N3N2M101N101O0O2O0O2O1N101O0O2O0O2O0O2O001N2O0O2O0O2O001N101N101N2O001N101N101N101O0O2O1OO1N1N3M3N2M3N2M3N2M2N3N2M3N2N2M3N2M2O2N2M3N2N2M30O0100O100O10000O100O100O100O10000O2O0O100O10000O100O100O10000O10010O00000001O0000001O000000001O0001O0001O01O0010O01O000010O01O001O010O001O010O00001O010O001O0010O01O000010O01O0010O01O001O010O000O2N1O2N1O2N2M3N3M2M4M2N3M2M4M2N3L3N3M2M4M2N3L3N3M2M4M2N3L3N3L5L4K6K4K6K4K6K4K5L5J5L5J5LYnl6"}, "label": "glazed donut just to the left of phone"}]}
{"id": 256760, "image": "COCO_train2014_000000256760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic light with red arrow pointing left\"."}], "task": "refering", "answer_template": "The \"traffic light with red arrow pointing left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167], "bbox": [0.07116, 0.5501856763925729, 0.38970000000000005, 0.7142440318302387], "iscrowd": 0, "area": 8052.30285, "rle": {"size": [377, 500], "counts": "g_=4b;4L4L4L4L4L4K5L4L3L5O1O1O1O1O1O1O1O1O1O1O1N2O001O000000000000000000O1000000000000000000000000O1000000000000000000000O1000O100000000000000000000000000O1000000000000000000000000O10000000000000O100000000000O1000000000000000000000000O100000000000000000000000000O100000O10000001O1O1O1O1O1O2M2O1O1O1O1O1O1L4L4L4L4L5J5L4L4LUR`3"}, "label": "traffic light with red arrow pointing left"}]}
{"id": 256760, "image": "COCO_train2014_000000256760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of a traffic light with a red arrow pointing left lit up\"."}], "task": "refering", "answer_template": "The \"a picture of a traffic light with a red arrow pointing left lit up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167], "bbox": [0.07116, 0.5501856763925729, 0.38970000000000005, 0.7142440318302387], "iscrowd": 0, "area": 8052.30285, "rle": {"size": [377, 500], "counts": "g_=4b;4L4L4L4L4L4K5L4L3L5O1O1O1O1O1O1O1O1O1O1O1N2O001O000000000000000000O1000000000000000000000000O1000000000000000000000O1000O100000000000000000000000000O1000000000000000000000000O10000000000000O100000000000O1000000000000000000000000O100000000000000000000000000O100000O10000001O1O1O1O1O1O2M2O1O1O1O1O1O1L4L4L4L4L5J5L4L4LUR`3"}, "label": "a picture of a traffic light with a red arrow pointing left lit up"}]}
{"id": 256760, "image": "COCO_train2014_000000256760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green signal light on road ]\"."}], "task": "refering", "answer_template": "The \"green signal light on road ]\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158], "bbox": [0.4975, 0.5311936339522546, 0.8090599999999999, 0.6940318302387267], "iscrowd": 0, "area": 8125.1744, "rle": {"size": [377, 500], "counts": "hlk22e;3N3L3M4M2M3N3L3N3L3M3N3L3N3L3M3O2N1O2N1O1O0000O010000000000000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000000000000000O100000000000000000000000000000000000000000O10000000O10000000000000000000000000000000000000000000000000000O100000000000O10002N1N2N3M2N2M4M2N3M2M3N3M2N2Me0\\OjgR1"}, "label": "green signal light on road ]"}]}
{"id": 256760, "image": "COCO_train2014_000000256760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic light with green arrows\"."}], "task": "refering", "answer_template": "The \"traffic light with green arrows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158], "bbox": [0.4975, 0.5311936339522546, 0.8090599999999999, 0.6940318302387267], "iscrowd": 0, "area": 8125.1744, "rle": {"size": [377, 500], "counts": "hlk22e;3N3L3M4M2M3N3L3N3L3M3N3L3N3L3M3O2N1O2N1O1O0000O010000000000000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000000000000000O100000000000000000000000000000000000000000O10000000O10000000000000000000000000000000000000000000000000000O100000000000O10002N1N2N3M2N2M4M2N3M2M3N3M2N2Me0\\OjgR1"}, "label": "traffic light with green arrows"}]}
{"id": 240378, "image": "COCO_train2014_000000240378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young man\"."}], "task": "refering", "answer_template": "The \"young man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 20, 21, 22, 23, 24, 25, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 120, 122, 123, 124, 125, 131, 132, 133, 134, 136, 137, 138, 139, 146, 147, 148, 150, 151, 152, 153, 160, 161, 162, 164, 165, 166, 167, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 201, 202, 203, 204, 205, 206, 207, 208, 209, 215, 216, 217, 218, 219, 220, 221, 222, 223, 229, 230, 231, 232, 233, 234, 235, 236, 237, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 271, 272, 273, 274, 275, 276, 277, 278, 279, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.38621483375959076, 0.0, 0.9996163682864451, 1.0], "iscrowd": 0, "area": 106657.20080000002, "rle": {"size": [640, 391], "counts": "Scn2:bc0e0\\Od0SFbN2g1BcN8_1BlN7V1DTO6m0D^O7b0AJ;7XO:c0FPOj0m0VOeN\\1W1dNTLU4h3kKYLW4d3jK\\LY4`3gKaL\\4[3dKgL]4V3cKkL`4Q3aKoLb4m2^KTMd4i2\\KYMQ5X2oJiM]5j1cJWN]5h1dJXN\\5g1dJ[N[5d1eJ]KnMa1]7R3dJ^KPN_1\\7R3dJ_KQN_1[7Q3eJ`KQN^1Z7Q3eJaKSN^1W7P3fJcKSN]1W7o2fJdKUN\\1U7o2gJdKUN]1T7cN\\JS1;nNVN\\1R7bNhJi00YOWN\\1Q7`NmJg0K^OXNZ1P7`NPKf0G]O]N]1l6^NTKe0D\\OaN`1g6^NWKc0A]OdNb1d6\\N\\Ka0\\O^OiNd1_6\\N_K?XOAmNc1\\6[NcK>lNLVO[1[6YNgK=_N8AQ1Y6YNkK:SNd0Ji0X6WNoK9mMi0Nf0V6WNRL7jMl0Of0U6UNUL8eMn03e0R6TNYL6bMQ15d0P6SN]L5^MT16d0o5QNaL5YMV19c0m5QNdL3VMY1:c0l5oMhL3QMZ1>c0i5oMkL1nL]1?c0h5nMnLOkL`1`0c0g5mMPMOiL`1a0d0f5lMSMMfLc1c0c0d5lMVMLbLe1e0c0c5kMcNb1kKc0b5jMcNc1lKc0a5iMcNd1mKc0`5iMcNc1oKb0_5jMbNd1PLb0^5iMbNe1QLb0]5hMbNf1RLb0\\5gMbNg1TLa0Z5hMbNf1ULa0Z5hMcNe1TLc0Y5hMcNd1ULd0X5gMeNc1TLf0W5fMgNb1TLg0U5fMhNb1TLg0U5gMhN`1TLi0T5gMhN`1RLk0V5eMiN^1PLo0W5cMiN^1nKQ1Y5aMjN\\1lKT1[5`MjN[1iKW1]5^MkNY1gK[1^5\\MlN8oI;e1S2`5ZMAb0mJV2c5XM_Ob0mJX2d5VM^Oc0lJX2g5UM]Ob0kJ[2h5SM\\Oc0jJ[2k5RM[Ob0iJ^2l5PMZOc0gJ_2P6nLYOb0fJb2Q6lLXOb0fJc2S6kLWOb0dJe2U6iLVOb0dJf2W6hLTOc0bJh2Z6eLTOb0aJj2\\6dLROb0aJl2]6bLROb0_Jn2_6`LQOb0_Jo2a6_LPOb0\\JR3d6\\LoNb0\\JT3e6ZLoNT1hIc2Z7YLmNY1cI`2`7WLlN^1^I\\2g7VLkN]5U1cJjN^5V1bJhN_5Y1aJfN`5Z1`JeNa5[1_JdNa5]1_JbNb5^1^JaNc5_1]J`Nc5a1]J^Nc5c1]J]N`5f1`JYN^5j1bJVN\\5l1dJTNY5o1gJPNX5R2hJmMV5V2jJiMT5Z2lJeMS5]2nJaMP5b2PK\\MP5f2PKYMn4j2SKVMi4m2WKSMe4Q3\\KnLa4U3_KlL]4W3cKiLY4[3hKdLU4_3kKaLQ4c3PL]Ll3f3TLZL^3T4cLkKP3b4PM^Kf2m4YMTK[2W5fMhJR2`5nM`Jn1d5SN\\Jh1h5XNXJc1m5^NRJ_1Q6aNPJ[1S6eNmIX1V6iNjIS1Y6mNgIP1\\6QOcIl0`6TObIg0a6ZO`Ia0c6A]I;e6F[I7g6JZI2h60RF2EJ[:5mE6J]O\\:?eE9S=HjB<T=FgB`0V=AgBc0W=_OdBf0Z=[OcBi0[=XObBm0[=UO`BP1^=SO]BQ1a=ROQB\\1l=gNfAf1X>]NYAP2f>g1N2N2N1O2N2N2N2N6J00001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O1O1O2N1O1cEc1UO_NdNX5J"}, "label": "young man"}]}
{"id": 240378, "image": "COCO_train2014_000000240378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man , wearing a white vest\"."}], "task": "refering", "answer_template": "The \"a smiling man , wearing a white vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 20, 21, 22, 23, 24, 25, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 120, 122, 123, 124, 125, 131, 132, 133, 134, 136, 137, 138, 139, 146, 147, 148, 150, 151, 152, 153, 160, 161, 162, 164, 165, 166, 167, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 201, 202, 203, 204, 205, 206, 207, 208, 209, 215, 216, 217, 218, 219, 220, 221, 222, 223, 229, 230, 231, 232, 233, 234, 235, 236, 237, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 271, 272, 273, 274, 275, 276, 277, 278, 279, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.38621483375959076, 0.0, 0.9996163682864451, 1.0], "iscrowd": 0, "area": 106657.20080000002, "rle": {"size": [640, 391], "counts": "Scn2:bc0e0\\Od0SFbN2g1BcN8_1BlN7V1DTO6m0D^O7b0AJ;7XO:c0FPOj0m0VOeN\\1W1dNTLU4h3kKYLW4d3jK\\LY4`3gKaL\\4[3dKgL]4V3cKkL`4Q3aKoLb4m2^KTMd4i2\\KYMQ5X2oJiM]5j1cJWN]5h1dJXN\\5g1dJ[N[5d1eJ]KnMa1]7R3dJ^KPN_1\\7R3dJ_KQN_1[7Q3eJ`KQN^1Z7Q3eJaKSN^1W7P3fJcKSN]1W7o2fJdKUN\\1U7o2gJdKUN]1T7cN\\JS1;nNVN\\1R7bNhJi00YOWN\\1Q7`NmJg0K^OXNZ1P7`NPKf0G]O]N]1l6^NTKe0D\\OaN`1g6^NWKc0A]OdNb1d6\\N\\Ka0\\O^OiNd1_6\\N_K?XOAmNc1\\6[NcK>lNLVO[1[6YNgK=_N8AQ1Y6YNkK:SNd0Ji0X6WNoK9mMi0Nf0V6WNRL7jMl0Of0U6UNUL8eMn03e0R6TNYL6bMQ15d0P6SN]L5^MT16d0o5QNaL5YMV19c0m5QNdL3VMY1:c0l5oMhL3QMZ1>c0i5oMkL1nL]1?c0h5nMnLOkL`1`0c0g5mMPMOiL`1a0d0f5lMSMMfLc1c0c0d5lMVMLbLe1e0c0c5kMcNb1kKc0b5jMcNc1lKc0a5iMcNd1mKc0`5iMcNc1oKb0_5jMbNd1PLb0^5iMbNe1QLb0]5hMbNf1RLb0\\5gMbNg1TLa0Z5hMbNf1ULa0Z5hMcNe1TLc0Y5hMcNd1ULd0X5gMeNc1TLf0W5fMgNb1TLg0U5fMhNb1TLg0U5gMhN`1TLi0T5gMhN`1RLk0V5eMiN^1PLo0W5cMiN^1nKQ1Y5aMjN\\1lKT1[5`MjN[1iKW1]5^MkNY1gK[1^5\\MlN8oI;e1S2`5ZMAb0mJV2c5XM_Ob0mJX2d5VM^Oc0lJX2g5UM]Ob0kJ[2h5SM\\Oc0jJ[2k5RM[Ob0iJ^2l5PMZOc0gJ_2P6nLYOb0fJb2Q6lLXOb0fJc2S6kLWOb0dJe2U6iLVOb0dJf2W6hLTOc0bJh2Z6eLTOb0aJj2\\6dLROb0aJl2]6bLROb0_Jn2_6`LQOb0_Jo2a6_LPOb0\\JR3d6\\LoNb0\\JT3e6ZLoNT1hIc2Z7YLmNY1cI`2`7WLlN^1^I\\2g7VLkN]5U1cJjN^5V1bJhN_5Y1aJfN`5Z1`JeNa5[1_JdNa5]1_JbNb5^1^JaNc5_1]J`Nc5a1]J^Nc5c1]J]N`5f1`JYN^5j1bJVN\\5l1dJTNY5o1gJPNX5R2hJmMV5V2jJiMT5Z2lJeMS5]2nJaMP5b2PK\\MP5f2PKYMn4j2SKVMi4m2WKSMe4Q3\\KnLa4U3_KlL]4W3cKiLY4[3hKdLU4_3kKaLQ4c3PL]Ll3f3TLZL^3T4cLkKP3b4PM^Kf2m4YMTK[2W5fMhJR2`5nM`Jn1d5SN\\Jh1h5XNXJc1m5^NRJ_1Q6aNPJ[1S6eNmIX1V6iNjIS1Y6mNgIP1\\6QOcIl0`6TObIg0a6ZO`Ia0c6A]I;e6F[I7g6JZI2h60RF2EJ[:5mE6J]O\\:?eE9S=HjB<T=FgB`0V=AgBc0W=_OdBf0Z=[OcBi0[=XObBm0[=UO`BP1^=SO]BQ1a=ROQB\\1l=gNfAf1X>]NYAP2f>g1N2N2N1O2N2N2N2N6J00001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O1O1O2N1O1cEc1UO_NdNX5J"}, "label": "a smiling man , wearing a white vest"}]}
{"id": 240378, "image": "COCO_train2014_000000240378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady fixing the guy ' s tie for him\"."}], "task": "refering", "answer_template": "The \"the lady fixing the guy ' s tie for him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 28, 29, 30, 42, 43, 44, 45, 56, 57, 58, 59, 70, 71, 72, 73, 84, 85, 86, 87, 98, 99, 100, 101, 104, 105, 107, 108, 112, 113, 114, 118, 119, 120, 121, 122, 126, 127, 128, 129, 131, 132, 133, 134, 135, 136, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 252, 253, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313], "bbox": [0.011023017902813298, 0.0561875, 0.7503580562659846, 0.9820156250000001], "iscrowd": 0, "area": 89977.27244999999, "rle": {"size": [640, 391], "counts": "]i2o0Q`0l9UFQ5kKS2O2N1O1O1O1O1O11O001O001O1O001O00100O001O001O1O001O001O001O1O001O010O1O001O001O1O001O001O1O001O00100O001O001O1O001O001O002N1O1O1O2O0O1O1O2gDUMY5l2UJeMk5\\2cIVN\\6k1WIaNi6a1TIaNk6`1RIbNn6_1PIbNP7_1mHdNR7^1kHcNU7^1hHdNX7]1fHeNY7\\1dHfN\\7\\1aHeN_7\\1^HgNb7Y1\\HhNd7Y1YHjNf7X1WHiNi7X1THjNl7W1RHkNm7V1PHlNP8V1mGkNS8V1kGlNT8U1gGoNY8R1bGRO^8P1[GVOd8k0WGYOi8m0lFYOS9X1XFlNh9f1bE^N^:c50O0001O001O00001O00001O00001O0010O0001O00001O00001O00O10000O100O2O000O10XLQG_Ln8n2jGnLV8^2bH^M^7Z2nHbMQ7[2WIaMi6[2^IbMb6Z2fIbMY6[2oIaMQ6[2WJaMh5\\2`J`M`5\\2hJ`MX5\\2oJaMP5\\2XK`Mh4\\2`KaM^4\\2jK`MV4\\2RL`Mm3]2[L_Md3_2cL]M\\3d2hLXMW3i2nLRMQ3o2SMmLl2T3YMgLf2Z3_MaL`2`3dM\\L[2e3jMVLU2k3oMQLP2P4UNkKj1V4o61O1O1O1O0O2N2O1N2O1N2N2O1N2O1N2N2O1L4J6J5F;G90O10O01000O010O10O10O010O10O100O2O000O101O0O100O101O0O10001N100_JPBQ5`>0003L5L5K0O10O00100O010O010O10O0100O0010O0100O010O10O010O0100O010O010O10O0M4L3L5L3M4L3M4K4M4L4L3M4K4M4L4L3M4K4M4K5L3L5K4M4K5L3L5L3L5L4K?B`Qm1"}, "label": "the lady fixing the guy ' s tie for him"}]}
{"id": 240378, "image": "COCO_train2014_000000240378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman buttoning a man ' s shirt\"."}], "task": "refering", "answer_template": "The \"a woman buttoning a man ' s shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 28, 29, 30, 42, 43, 44, 45, 56, 57, 58, 59, 70, 71, 72, 73, 84, 85, 86, 87, 98, 99, 100, 101, 104, 105, 107, 108, 112, 113, 114, 118, 119, 120, 121, 122, 126, 127, 128, 129, 131, 132, 133, 134, 135, 136, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 252, 253, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313], "bbox": [0.011023017902813298, 0.0561875, 0.7503580562659846, 0.9820156250000001], "iscrowd": 0, "area": 89977.27244999999, "rle": {"size": [640, 391], "counts": "]i2o0Q`0l9UFQ5kKS2O2N1O1O1O1O1O11O001O001O1O001O00100O001O001O1O001O001O001O1O001O010O1O001O001O1O001O001O1O001O00100O001O001O1O001O001O002N1O1O1O2O0O1O1O2gDUMY5l2UJeMk5\\2cIVN\\6k1WIaNi6a1TIaNk6`1RIbNn6_1PIbNP7_1mHdNR7^1kHcNU7^1hHdNX7]1fHeNY7\\1dHfN\\7\\1aHeN_7\\1^HgNb7Y1\\HhNd7Y1YHjNf7X1WHiNi7X1THjNl7W1RHkNm7V1PHlNP8V1mGkNS8V1kGlNT8U1gGoNY8R1bGRO^8P1[GVOd8k0WGYOi8m0lFYOS9X1XFlNh9f1bE^N^:c50O0001O001O00001O00001O00001O0010O0001O00001O00001O00O10000O100O2O000O10XLQG_Ln8n2jGnLV8^2bH^M^7Z2nHbMQ7[2WIaMi6[2^IbMb6Z2fIbMY6[2oIaMQ6[2WJaMh5\\2`J`M`5\\2hJ`MX5\\2oJaMP5\\2XK`Mh4\\2`KaM^4\\2jK`MV4\\2RL`Mm3]2[L_Md3_2cL]M\\3d2hLXMW3i2nLRMQ3o2SMmLl2T3YMgLf2Z3_MaL`2`3dM\\L[2e3jMVLU2k3oMQLP2P4UNkKj1V4o61O1O1O1O0O2N2O1N2O1N2N2O1N2O1N2N2O1L4J6J5F;G90O10O01000O010O10O10O010O10O100O2O000O101O0O100O101O0O10001N100_JPBQ5`>0003L5L5K0O10O00100O010O010O10O0100O0010O0100O010O10O010O0100O010O010O10O0M4L3L5L3M4L3M4K4M4L4L3M4K4M4L4L3M4K4M4K5L3L5K4M4K5L3L5L3L5L4K?B`Qm1"}, "label": "a woman buttoning a man ' s shirt"}]}
{"id": 232186, "image": "COCO_train2014_000000232186.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoli , lying on top of two pasta shells\"."}], "task": "refering", "answer_template": "The \"a piece of broccoli , lying on top of two pasta shells\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 163, 164, 165], "bbox": [0.07675, 0.11084705882352941, 0.46359374999999997, 0.5117882352941177], "iscrowd": 0, "area": 17456.10295, "rle": {"size": [425, 640], "counts": "n_d0l0_<0N2N1O1O1N2O1O1O1O2N0O2O00001O00000O2O001O001O0O2O00001O001O0O2O00001O001N101N001O10O01O01O01O010O00010O00010O001O01O01O01O01O010O00010O00010O0010O01O010O1O010O010O010O010O010O10O010O010O010O010O01O10O010O010O01O010O10O01O010O0010O01O010O1O010O0010O01M2M4M3M2F;K4L5L3N3M2N3N1N3O10O01O0010O01000O1000O0100O10000O10O10O10001N100O101O0O10001N100O101O0O101N10001N101O0O2O0O2O001N10001N101N1O2N1O2N1O2O001N101O001N101O0O2O001N10001O0O2O001N101O0O2O001N101O001N101O0O2O001N2O1O2N1N2O1O2M2O1O4K6K5K5JSX^4"}, "label": "a piece of broccoli , lying on top of two pasta shells"}]}
{"id": 232186, "image": "COCO_train2014_000000232186.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"piece of broccoli on top of the noodles\"."}], "task": "refering", "answer_template": "The \"piece of broccoli on top of the noodles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 163, 164, 165], "bbox": [0.07675, 0.11084705882352941, 0.46359374999999997, 0.5117882352941177], "iscrowd": 0, "area": 17456.10295, "rle": {"size": [425, 640], "counts": "n_d0l0_<0N2N1O1O1N2O1O1O1O2N0O2O00001O00000O2O001O001O0O2O00001O001O0O2O00001O001N101N001O10O01O01O01O010O00010O00010O001O01O01O01O01O010O00010O00010O0010O01O010O1O010O010O010O010O010O10O010O010O010O010O01O10O010O010O01O010O10O01O010O0010O01O010O1O010O0010O01M2M4M3M2F;K4L5L3N3M2N3N1N3O10O01O0010O01000O1000O0100O10000O10O10O10001N100O101O0O10001N100O101O0O101N10001N101O0O2O0O2O001N10001N101N1O2N1O2N1O2O001N101O001N101O0O2O001N10001O0O2O001N101O0O2O001N101O001N101O0O2O001N2O1O2N1N2O1O2M2O1O4K6K5K5JSX^4"}, "label": "piece of broccoli on top of the noodles"}]}
{"id": 543490, "image": "COCO_train2014_000000543490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pigeon in foreground looking right\"."}], "task": "refering", "answer_template": "The \"pigeon in foreground looking right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 104, 105, 106, 107, 127, 128, 129, 130, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 333, 334, 335, 336, 337, 338], "bbox": [0.240078125, 0.08512035010940919, 0.751, 0.9407221006564551], "iscrowd": 0, "area": 52955.124699999986, "rle": {"size": [457, 640], "counts": "YTU24S>2N2YBIZ=9fBGX=;gBEX==gBDW=?gBBX=`0gB@W=d0fB]OY=j0N2O1O1N2O1O1N2O0O2O1O1O1O1O1N101O1O1O1O0O2O1O1N2O1O1N200O1O1O001O1O001O1O0O2N2N1O2N2N101O1O1N101O1O0M4J6K4O2N2O0O1O2O0O2O0O1O2O0O2O0O101N100N2O2N100O100O10O01000O010N1100O2O1N2O1O1N2O1O1O0O2O1O1N2O001O1O1O001O1N101O1O1O00000O1N3N1N2O2M3O001O1O1O1O1O1O1O1O001O1O1O001O1O1O1O1O1O2N1O1O101N1O2N1O2N2N101N2N1O2N1O2N2M2O2N3L3N3M2N1N3^ITId5m6[JSIe5o6YJRIe5R7WJPIh5U7SJlHl5Y7nIiHQ6g7O2O1O1N2O1O00POSJfIk5W6[JhId5U6bJhI^5V6gJhIW5X6lJgIS5Y6PKeIo4[6SKdIl4[6WKdIg4]6\\K`Ic4a6_K^I`4b6bK]I\\4e6dK[IZ4g6gKXIW4j6iKVIU4m6kKRIS4P7mKPIQ4R7PLmHo3T7QLlHm3W7RLjHk3Y7ULfHi3]7VLcHh3`7XL_Hg3b7YL^He3e7ZL[Hd3h7[LXHc3j7^LUHa3m7^LSH`3P8_LPH_3R8aLnG]3U8bLkG\\3W8eLhGZ3Z8eLfGY3\\8gLdGW3_8hLaGV3b8iL^GU3d8kL\\GS3g8lLYGQ3j8oLWGn2m8PMSGn2S9nLmFP3Y9lLgFQ3]8^LXH?ZOQ3X8oL[HL]OS3X8TM^HFYOT3Z8WMaHAUOV3[8[McH[OROY3Z8_MaH[OTOU3[8aM`H[OTOS3]8dM]H[OUOP3^8gM\\He2d7]M\\Hb2d7`M[HVOWOi2_8RNZHUOWOh2^8VNZHROXOg2^8XNZHQOXOe2_8\\NXHnNZOe2^8^NXHmNYOd2`8aNVHkNZOc2_8dNWHhN[Oc2^8gNVHfN\\Ob2^8iNWHcN\\Oc2]8kNWHbN\\Ob2]8nNWH_N[Oc2]8POXH\\N\\Oc2\\8SOWHZN]Ob2\\8UOXHWN]Od2Z8WOXHUN^Oc2Z8YOYHSN]Oc2Z8\\OXHPN_Oc2X8_ObH`0^7BaH=_7DaH<^7FaH9_7HaH7^7LaH3^7ObH1\\72cHM]74cHK\\78cHG]7:cHF\\7;dHD\\7>cHB[7`0eH@Z7b0eH^OZ7c0fH]OY7d0gH\\OX7f0gHZOW7h0iHXOV7j0iHVOV7k0jHUOU7m0jHSOT7o0lHQOS7P1mHPOR7R1mHnNQ7T1oHlNm3dM_Nc3cMiNe1Oe0Y1fMhN_18f0R1kMfN]1<d0Q1nMcN]1`0`0o0SNaN\\1d0=m0VN_N\\1h09k0[N]N[1m04h0aN[N[1Q1Ng0fNYNZ1V1Jb0lNXNY1[1E`0POVN[1^1_O>VOTNZ1c1ZO<[ORNZ1f1UO:APNZ1j1PO7FoMZ1n1iN7LkM[1R2^N:6fM[1S7dNmH\\1S7dNmH[1T7eNlH[1U7dNlH[1T7eNlH[1U7cNnH[1R7eNoHZ1Q7fNQIX1P7gNQIX1o6hNRIW1o6hNSIV1m6jNTIV1l6hNWIV1k6hNVIW1P:N1O2N101N102M4L4M2M3L3N3L4MmkV2"}, "label": "pigeon in foreground looking right"}]}
{"id": 543490, "image": "COCO_train2014_000000543490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three pegion standing together\"."}], "task": "refering", "answer_template": "The \"three pegion standing together\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 104, 105, 106, 107, 127, 128, 129, 130, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 333, 334, 335, 336, 337, 338], "bbox": [0.240078125, 0.08512035010940919, 0.751, 0.9407221006564551], "iscrowd": 0, "area": 52955.124699999986, "rle": {"size": [457, 640], "counts": "YTU24S>2N2YBIZ=9fBGX=;gBEX==gBDW=?gBBX=`0gB@W=d0fB]OY=j0N2O1O1N2O1O1N2O0O2O1O1O1O1O1N101O1O1O1O0O2O1O1N2O1O1N200O1O1O001O1O001O1O0O2N2N1O2N2N101O1O1N101O1O0M4J6K4O2N2O0O1O2O0O2O0O1O2O0O2O0O101N100N2O2N100O100O10O01000O010N1100O2O1N2O1O1N2O1O1O0O2O1O1N2O001O1O1O001O1N101O1O1O00000O1N3N1N2O2M3O001O1O1O1O1O1O1O1O001O1O1O001O1O1O1O1O1O2N1O1O101N1O2N1O2N2N101N2N1O2N1O2N2M2O2N3L3N3M2N1N3^ITId5m6[JSIe5o6YJRIe5R7WJPIh5U7SJlHl5Y7nIiHQ6g7O2O1O1N2O1O00POSJfIk5W6[JhId5U6bJhI^5V6gJhIW5X6lJgIS5Y6PKeIo4[6SKdIl4[6WKdIg4]6\\K`Ic4a6_K^I`4b6bK]I\\4e6dK[IZ4g6gKXIW4j6iKVIU4m6kKRIS4P7mKPIQ4R7PLmHo3T7QLlHm3W7RLjHk3Y7ULfHi3]7VLcHh3`7XL_Hg3b7YL^He3e7ZL[Hd3h7[LXHc3j7^LUHa3m7^LSH`3P8_LPH_3R8aLnG]3U8bLkG\\3W8eLhGZ3Z8eLfGY3\\8gLdGW3_8hLaGV3b8iL^GU3d8kL\\GS3g8lLYGQ3j8oLWGn2m8PMSGn2S9nLmFP3Y9lLgFQ3]8^LXH?ZOQ3X8oL[HL]OS3X8TM^HFYOT3Z8WMaHAUOV3[8[McH[OROY3Z8_MaH[OTOU3[8aM`H[OTOS3]8dM]H[OUOP3^8gM\\He2d7]M\\Hb2d7`M[HVOWOi2_8RNZHUOWOh2^8VNZHROXOg2^8XNZHQOXOe2_8\\NXHnNZOe2^8^NXHmNYOd2`8aNVHkNZOc2_8dNWHhN[Oc2^8gNVHfN\\Ob2^8iNWHcN\\Oc2]8kNWHbN\\Ob2]8nNWH_N[Oc2]8POXH\\N\\Oc2\\8SOWHZN]Ob2\\8UOXHWN]Od2Z8WOXHUN^Oc2Z8YOYHSN]Oc2Z8\\OXHPN_Oc2X8_ObH`0^7BaH=_7DaH<^7FaH9_7HaH7^7LaH3^7ObH1\\72cHM]74cHK\\78cHG]7:cHF\\7;dHD\\7>cHB[7`0eH@Z7b0eH^OZ7c0fH]OY7d0gH\\OX7f0gHZOW7h0iHXOV7j0iHVOV7k0jHUOU7m0jHSOT7o0lHQOS7P1mHPOR7R1mHnNQ7T1oHlNm3dM_Nc3cMiNe1Oe0Y1fMhN_18f0R1kMfN]1<d0Q1nMcN]1`0`0o0SNaN\\1d0=m0VN_N\\1h09k0[N]N[1m04h0aN[N[1Q1Ng0fNYNZ1V1Jb0lNXNY1[1E`0POVN[1^1_O>VOTNZ1c1ZO<[ORNZ1f1UO:APNZ1j1PO7FoMZ1n1iN7LkM[1R2^N:6fM[1S7dNmH\\1S7dNmH[1T7eNlH[1U7dNlH[1T7eNlH[1U7cNnH[1R7eNoHZ1Q7fNQIX1P7gNQIX1o6hNRIW1o6hNSIV1m6jNTIV1l6hNWIV1k6hNVIW1P:N1O2N101N102M4L4M2M3L3N3L4MmkV2"}, "label": "three pegion standing together"}]}
{"id": 543490, "image": "COCO_train2014_000000543490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pigeon looking down , to the left of two other pigeons\"."}], "task": "refering", "answer_template": "The \"a pigeon looking down , to the left of two other pigeons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238, 260], "bbox": [0.24281250000000001, 0.29514223194748357, 0.46653125000000006, 0.7307439824945295], "iscrowd": 0, "area": 14896.232549999997, "rle": {"size": [457, 640], "counts": "Q]U29k=7J6J6M4K4M3M2K6G=I4L2N1O2O1N2N3M2N3M2M4J5M4L4L4M4K6K2M2O2N3L4H8K5L4L4L3M4I6M4M2M3M3N2M3N2M3N2N2N2N2N2N2N2O1O1OL3N3N2N2N2O1N2N2N2M3J7K4O1O2N1O2N1O3M2N2N2M3N2M2N3N2N1O2O1N2N1O2N2N1O2M4M2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M2O3L3N2N3L3N2M3N3L3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N2N2M2O2M3N1N3N1O2O]Zi4"}, "label": "a pigeon looking down , to the left of two other pigeons"}]}
{"id": 543490, "image": "COCO_train2014_000000543490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dove is sitting near another two doves\"."}], "task": "refering", "answer_template": "The \"a dove is sitting near another two doves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238, 260], "bbox": [0.24281250000000001, 0.29514223194748357, 0.46653125000000006, 0.7307439824945295], "iscrowd": 0, "area": 14896.232549999997, "rle": {"size": [457, 640], "counts": "Q]U29k=7J6J6M4K4M3M2K6G=I4L2N1O2O1N2N3M2N3M2M4J5M4L4L4M4K6K2M2O2N3L4H8K5L4L4L3M4I6M4M2M3M3N2M3N2M3N2N2N2N2N2N2N2O1O1OL3N3N2N2N2O1N2N2N2M3J7K4O1O2N1O2N1O3M2N2N2M3N2M2N3N2N1O2O1N2N1O2N2N1O2M4M2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M2O3L3N2N3L3N2M3N3L3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N2N2M2O2M3N1N3N1O2O]Zi4"}, "label": "a dove is sitting near another two doves"}]}
{"id": 543490, "image": "COCO_train2014_000000543490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pegeon at the middle of another two\"."}], "task": "refering", "answer_template": "The \"a pegeon at the middle of another two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 78, 79, 80, 81, 82, 103, 104, 126, 127, 149, 153, 154, 176, 177, 178, 200, 201, 202, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 297, 298, 311, 312, 314, 315, 334, 335], "bbox": [0.42474999999999996, 0.1174617067833698, 0.98953125, 0.8933260393873085], "iscrowd": 0, "area": 25219.056200000006, "rle": {"size": [457, 640], "counts": "j_i32U>5L3M1O1N2O001N2O001O0^ONmB4R=b0O0O2O001O001N101O1O1N2O1O1O1O101O00010O1O100O1O2N100O1O2O3L4L4REYNlN@R:[2hFZN^9i1WF]Nj9o22O2N1O1N2O2N1O1N3N1O1N2O2N1O1N3N1O1O1N3N1O1O1N4M2N3M3L4MlKmMULQ2l3PNULl1l3TNWLi1j3WNXLf1h3ZNZLd1g3\\N[L`1g3_N\\L]1e3dN\\LY1f3fN]LV1d3kN^LQ1d3oN^Lm0d3SO]Lj0d3WO^Le0d3[O^La0c3@_L<c3D^L9c3H_L4a3MbLO_32bLK^39bLC^3`0S54L3M3N02O1N2N2N4L4M3L4K\\=O\\B>L4K2O1O1O1O1O2M2O00O100O10O0100O100O10O2O00001O000O2O00001O000O101O00001N100O2O0O103L4M3L4M3L4L4M5J5L5lDcMR:b2gEbMU7G^Kl2UMbMZ7G\\K_3c4fLXK\\3g4hLSKZ3m4iLnJY3S5jLgJY3X5kLaJX3_5kL\\JX3c5lLVJW3k5lLoIU3R6nLhIQ3\\6SM^Il2f6VMTIi2P7[MjHc2Z7aM_H^2e7eMUH[2o7hMkGW2X8i14M3L5L3L4M4M2O1O101N2N2N2O1N3M2O1O1N2O1O1O1N2O2N1O1N2O1O1N2O1O1O1N3N1O1O1N2O1O1N2O1O1O0O200O100O010O1O100O001O1O1O1O001O1N2O1O001O1O1O1O001O2N1O1N3N1O1O2N1O1O2N1O1O2N1N2N1N3N2M2O2O1O0100O01000O01000O010O10O10O10O10O10O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001SOXC7i<HYC6g<J\\C3e<L]C2d<M^C1c<N]C2d<M]C2c<N]C2d<M]C2d<M]C2d<M\\C3d<M]C3c<L]C4d<K]C5c<J]C6d<I]C6c<K\\C6i<DXC;[=10O00100O010O0010O010O01O010O10O01O10O0100O1O1N101O1N2O1O0OQi2"}, "label": "a pegeon at the middle of another two"}]}
{"id": 84744, "image": "COCO_train2014_000000084744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"larger sheep in the back\"."}], "task": "refering", "answer_template": "The \"larger sheep in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 145, 146, 147, 148, 149, 158, 159, 161, 162, 171, 172, 175], "bbox": [0.21104, 0.55404, 0.5441066666666667, 0.75214], "iscrowd": 0, "area": 7146.790150000001, "rle": {"size": [500, 375], "counts": "TmV11b?3cNNSC5h<2RC2i<5RC6c<OUCc0\\<[10O001O01O01O001@?O1O20O010O<E3L2O1O0O2O00QOVCTOj<k0XCTOg<l0[CROf<l0]CROd<m0^CROb<m0`CQOa<n0cCoN\\<R1iChNX<X1nCbNR<_1Q11N101O001N3N00001O0O1000000000000000O1000000000000000000O2O0000001O0000000O2O000000001O00000O20O001O0O5L3M1O00`0@;E7JO01O10O0O2O0O1RO]CkNd<P1kCdNV<[1R101O0O2O0O2M2O2M2N2N3M2O2N2N2N2N3L3N5GPAEV?2bVc2"}, "label": "larger sheep in the back"}]}
{"id": 84744, "image": "COCO_train2014_000000084744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ewe letting her lamb nurse\"."}], "task": "refering", "answer_template": "The \"a ewe letting her lamb nurse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 145, 146, 147, 148, 149, 158, 159, 161, 162, 171, 172, 175], "bbox": [0.21104, 0.55404, 0.5441066666666667, 0.75214], "iscrowd": 0, "area": 7146.790150000001, "rle": {"size": [500, 375], "counts": "TmV11b?3cNNSC5h<2RC2i<5RC6c<OUCc0\\<[10O001O01O01O001@?O1O20O010O<E3L2O1O0O2O00QOVCTOj<k0XCTOg<l0[CROf<l0]CROd<m0^CROb<m0`CQOa<n0cCoN\\<R1iChNX<X1nCbNR<_1Q11N101O001N3N00001O0O1000000000000000O1000000000000000000O2O0000001O0000000O2O000000001O00000O20O001O0O5L3M1O00`0@;E7JO01O10O0O2O0O1RO]CkNd<P1kCdNV<[1R101O0O2O0O2M2O2M2N2N3M2O2N2N2N2N3L3N5GPAEV?2bVc2"}, "label": "a ewe letting her lamb nurse"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tall man\"."}], "task": "refering", "answer_template": "The \"the tall man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 131, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 287, 288, 289, 290, 291, 292, 310, 311, 312, 314, 315], "bbox": [0.50228125, 0.11639344262295083, 0.7261875, 0.9213817330210773], "iscrowd": 0, "area": 21595.107649999994, "rle": {"size": [427, 640], "counts": "^dV43W=1O2M2O2N2N1N3M2L5J6K4K6K4L5J6K4K6K4K6K4K6K5J6J5L3L5bJ`MMc20cMIb27`MBd2>_M\\Oe2c0^MVOf2k0\\MhNo2W1SM\\NX3d1kLoM`3Q2bLbMh3^2[LVMo3i2TLnLR4S3oKdLY4[3jK\\L\\4d3gKRL^3EbKY4R1iK^37[KP4g6[LTId3h6eLSIi2Z7bMaHn1j7[NQH^1Q8lNkGT1P8V2K4K6K5J5L5J6K4O2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O15K<E;G9F:F5K3N2M3M3M3M3N2M3M3M3N24K5I5KNjJjJX2V5aMSK^2l4ZM^Ke2`4UMiKi2W4PMQLP3m3jL\\LU3c3dLfLZ3Z3bLlL]3R3cLQM[3o2dLSM\\3k2dLXM[3g2cL]M[3c2dL`MWOQOS2^3eNIT18kNNo03PO4i0MVO9d0H[O?>BAe08\\OGj02XOLP1MQO2P1MQO2P1MRO1o0NRO1o0NSO0m0OVOOk00WONj01XOMi02XOLj03XOKh04[OJf05\\OIe06]OHd07]OHd07^OGb08AFa08AEa0:@Ea09BE`09Z6O1N2O2N1N2OYcX2"}, "label": "the tall man"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black belt\"."}], "task": "refering", "answer_template": "The \"a man wearing a black belt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 131, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 287, 288, 289, 290, 291, 292, 310, 311, 312, 314, 315], "bbox": [0.50228125, 0.11639344262295083, 0.7261875, 0.9213817330210773], "iscrowd": 0, "area": 21595.107649999994, "rle": {"size": [427, 640], "counts": "^dV43W=1O2M2O2N2N1N3M2L5J6K4K6K4L5J6K4K6K4K6K4K6K5J6J5L3L5bJ`MMc20cMIb27`MBd2>_M\\Oe2c0^MVOf2k0\\MhNo2W1SM\\NX3d1kLoM`3Q2bLbMh3^2[LVMo3i2TLnLR4S3oKdLY4[3jK\\L\\4d3gKRL^3EbKY4R1iK^37[KP4g6[LTId3h6eLSIi2Z7bMaHn1j7[NQH^1Q8lNkGT1P8V2K4K6K5J5L5J6K4O2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O15K<E;G9F:F5K3N2M3M3M3M3N2M3M3M3N24K5I5KNjJjJX2V5aMSK^2l4ZM^Ke2`4UMiKi2W4PMQLP3m3jL\\LU3c3dLfLZ3Z3bLlL]3R3cLQM[3o2dLSM\\3k2dLXM[3g2cL]M[3c2dL`MWOQOS2^3eNIT18kNNo03PO4i0MVO9d0H[O?>BAe08\\OGj02XOLP1MQO2P1MQO2P1MRO1o0NRO1o0NSO0m0OVOOk00WONj01XOMi02XOLj03XOKh04[OJf05\\OIe06]OHd07]OHd07^OGb08AFa08AEa0:@Ea09BE`09Z6O1N2O2N1N2OYcX2"}, "label": "a man wearing a black belt"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man next to a taller man\"."}], "task": "refering", "answer_template": "The \"a man next to a taller man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 304, 305, 306, 308, 309, 328], "bbox": [0.24965625, 0.16063231850117096, 0.46057812499999995, 0.9498829039812647], "iscrowd": 0, "area": 23132.140949999997, "rle": {"size": [427, 640], "counts": "RSS26R=8G8H9H2M4M2M4L3N3L3M4M2M4M3L3M4M2[KiMTN[2a1PN[NS2\\1WNaNl1Y1[NbNh1]1ZN`Ni1_1XN]Nk1b1XNYNl1f1UNZMk2d2WM[Li3e3XLYLj3e3YLXLh3h3YLVLi3h3YLVLh3j3ZLSLf3m3\\LQLe3P4[LmKf3U4ZLiKg3X4YLfKg3[4`L]Ka3d4fLSKZ3o4mLhJS3Z5Z25J6K4K6K5K5J6KZOfIdJU6_5QJ^Jj5d5\\JYJ^5j5hJSJS5o5SKnIh4T6^KiI\\4[6iKaIS4c6U15L4K500010O1O001O1O0010O01O1O001O5L3N2N2N2O1N2N2N2O1N3M2N2N6K5J6J5K6K1N1O13M3M4L3L4M2QK[JR2c5kMcJT2[5jMiJV2U5gMPKX2o4eMVKZ2i4cM]K[2b4cMbK]2\\4`MiK_2V4^MoKa2P4]MULa2j3\\M[Lc2d3ZMaLf2]3XMgLg2X3VMnLh2S3SMRMk2Q3PMSMn2P3lLUMR3n2hLXMU3l2eLXMX3T3ZLQMd3]6N3M3M5J7J5K6J5J7J5K6J5J7I6J7E:C>A>CQa_4"}, "label": "a man next to a taller man"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a hankercheif and dark jeans\"."}], "task": "refering", "answer_template": "The \"a man wearing a hankercheif and dark jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 304, 305, 306, 308, 309, 328], "bbox": [0.24965625, 0.16063231850117096, 0.46057812499999995, 0.9498829039812647], "iscrowd": 0, "area": 23132.140949999997, "rle": {"size": [427, 640], "counts": "RSS26R=8G8H9H2M4M2M4L3N3L3M4M2M4M3L3M4M2[KiMTN[2a1PN[NS2\\1WNaNl1Y1[NbNh1]1ZN`Ni1_1XN]Nk1b1XNYNl1f1UNZMk2d2WM[Li3e3XLYLj3e3YLXLh3h3YLVLi3h3YLVLh3j3ZLSLf3m3\\LQLe3P4[LmKf3U4ZLiKg3X4YLfKg3[4`L]Ka3d4fLSKZ3o4mLhJS3Z5Z25J6K4K6K5K5J6KZOfIdJU6_5QJ^Jj5d5\\JYJ^5j5hJSJS5o5SKnIh4T6^KiI\\4[6iKaIS4c6U15L4K500010O1O001O1O0010O01O1O001O5L3N2N2N2O1N2N2N2O1N3M2N2N6K5J6J5K6K1N1O13M3M4L3L4M2QK[JR2c5kMcJT2[5jMiJV2U5gMPKX2o4eMVKZ2i4cM]K[2b4cMbK]2\\4`MiK_2V4^MoKa2P4]MULa2j3\\M[Lc2d3ZMaLf2]3XMgLg2X3VMnLh2S3SMRMk2Q3PMSMn2P3lLUMR3n2hLXMU3l2eLXMX3T3ZLQMd3]6N3M3M5J7J5K6J5J7J5K6J5J7I6J7E:C>A>CQa_4"}, "label": "a man wearing a hankercheif and dark jeans"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue rimmed motorcycle\"."}], "task": "refering", "answer_template": "The \"a blue rimmed motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 187, 188, 189, 190, 210, 211, 212, 213, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 263, 264, 284, 285, 286, 287], "bbox": [0.130140625, 0.33845433255269325, 0.5041249999999999, 0.8608665105386417], "iscrowd": 0, "area": 12596.53025, "rle": {"size": [427, 640], "counts": "]lR12Y=6J5QODeDa0T;BkD?Q;DoD<P;EoD=Q;BoD?P;APE`0P;@oD`0Q;@oDa0Q;^OoDc0P;]OoDe0Q;ZOnDg0R;YOkDk0U;TOhDP1W;POgDS1X;mNeDV1[;jNcDY1\\;gNaD]1^;bNaDa1^;601O1O001D;O2O0O2N2O010O001O010O001O0010O000O1O2O0O100O2N100O101N1000O10001O0000000O101O000000000O2O000001O1O100O001O]MiMSJV2l5kMUJU2j5kMWJT2S5kMmH1Q2S2R5mMkH1S2R2R5mMkH1S2R2R5nMjH0S2S2S5nMhH1S2S2T5lMiH1R2T2U5lMhH0R2U2V5lMfH0S2U2W5kMfH0R2V2X5kMeHOR2W2Z5jMbH0T2W2Y5iMcH1R2W2[5iMbH0R2X2\\5iM`H0S2X2]5hM`H0R2Y2^5hM_HOR2W2b5kMZHJOMT2W2k5PNTHHY2Q2k5UN]Je1j5UNaG0h2c1l8J6I7J6I7I7J6Igh6M`WI5K4L4K5L4L5K4K5L4L3M3M4K4M3M3M4L3O22N1O2N2N1OUE^No9_1cEQO]:n12O1N1O2O0O2O1N1O3N5J6K5J6J6K5J7J5J6J6K5J6K5J6Khg53TXJ7J7H8H7I8I7H7I8H7J7H7I1O010O001O010O1O^M^E[2b:bMbE^2e:010000N2N2N2N2N101N2N2N2N2N2N1O2N2N2N3N2M4L4K4J7J5J7J6I6KnkS4"}, "label": "a blue rimmed motorcycle"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle on the left\"."}], "task": "refering", "answer_template": "The \"the motorcycle on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 187, 188, 189, 190, 210, 211, 212, 213, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 263, 264, 284, 285, 286, 287], "bbox": [0.130140625, 0.33845433255269325, 0.5041249999999999, 0.8608665105386417], "iscrowd": 0, "area": 12596.53025, "rle": {"size": [427, 640], "counts": "]lR12Y=6J5QODeDa0T;BkD?Q;DoD<P;EoD=Q;BoD?P;APE`0P;@oD`0Q;@oDa0Q;^OoDc0P;]OoDe0Q;ZOnDg0R;YOkDk0U;TOhDP1W;POgDS1X;mNeDV1[;jNcDY1\\;gNaD]1^;bNaDa1^;601O1O001D;O2O0O2N2O010O001O010O001O0010O000O1O2O0O100O2N100O101N1000O10001O0000000O101O000000000O2O000001O1O100O001O]MiMSJV2l5kMUJU2j5kMWJT2S5kMmH1Q2S2R5mMkH1S2R2R5mMkH1S2R2R5nMjH0S2S2S5nMhH1S2S2T5lMiH1R2T2U5lMhH0R2U2V5lMfH0S2U2W5kMfH0R2V2X5kMeHOR2W2Z5jMbH0T2W2Y5iMcH1R2W2[5iMbH0R2X2\\5iM`H0S2X2]5hM`H0R2Y2^5hM_HOR2W2b5kMZHJOMT2W2k5PNTHHY2Q2k5UN]Je1j5UNaG0h2c1l8J6I7J6I7I7J6Igh6M`WI5K4L4K5L4L5K4K5L4L3M3M4K4M3M3M4L3O22N1O2N2N1OUE^No9_1cEQO]:n12O1N1O2O0O2O1N1O3N5J6K5J6J6K5J7J5J6J6K5J6K5J6Khg53TXJ7J7H8H7I8I7H7I8H7J7H7I1O010O001O010O1O^M^E[2b:bMbE^2e:010000N2N2N2N2N101N2N2N2N2N2N1O2N2N2N3N2M4L4K4J7J5J7J6I6KnkS4"}, "label": "the motorcycle on the left"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle behind the taller man\"."}], "task": "refering", "answer_template": "The \"the motorcycle behind the taller man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 176, 177, 178, 194, 195, 196, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 269, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 317, 318, 319], "bbox": [0.45278124999999997, 0.465175644028103, 0.9205625, 0.8966276346604216], "iscrowd": 0, "area": 23873.400899999993, "rle": {"size": [427, 640], "counts": "oUi37R=6K2M2O2M3N2M3N1N3N2M3N2M3L3N3M3M3L4M2N3L5L3M4K4N3L3N2N3L3N3L3N3L3N3M3L4M3L3011O2N1O2N2M2O2K5I7I4L4L3M3M4L3N2M4L3M3M4L3M3M3M4Ai`5SO_`J3N2M3N2J6I7J6J6J6K6K4L4M3L4L5L3L4L4O2O0000001O00000O1102M2N3M3N2M3L4M3L4M3L6K6I7J5J7Jd0[OZT10fkN3M3N2M2N3M3M3M2O2M5K8I7I7I7I6J7J6I7I5K4L4M3L4L1OO101O0O10000O2M2J6K5POhEgN^:[1h03M2N3N1N2N3M2N2N3M2N2Nf0ZO4L4L4L4L4L5K4L4L4L5K4L4L3M00XKTHR4d8N3M2O2M2O1N3M2O1N100O100O1O10O0100O1O100O100O0010O001O001O01O01O001O00001O010O00000000000001O00000000000000001O000000000000001O000000000000001O1O101N1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1N2M3M2O2M3N2M3M2O2M3N2M3M3N4K:Ga0^OZnd0"}, "label": "the motorcycle behind the taller man"}]}
{"id": 420620, "image": "COCO_train2014_000000420620.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle to the right of another\"."}], "task": "refering", "answer_template": "The \"a motorcycle to the right of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 176, 177, 178, 194, 195, 196, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 269, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 317, 318, 319], "bbox": [0.45278124999999997, 0.465175644028103, 0.9205625, 0.8966276346604216], "iscrowd": 0, "area": 23873.400899999993, "rle": {"size": [427, 640], "counts": "oUi37R=6K2M2O2M3N2M3N1N3N2M3N2M3L3N3M3M3L4M2N3L5L3M4K4N3L3N2N3L3N3L3N3L3N3M3L4M3L3011O2N1O2N2M2O2K5I7I4L4L3M3M4L3N2M4L3M3M4L3M3M3M4Ai`5SO_`J3N2M3N2J6I7J6J6J6K6K4L4M3L4L5L3L4L4O2O0000001O00000O1102M2N3M3N2M3L4M3L4M3L6K6I7J5J7Jd0[OZT10fkN3M3N2M2N3M3M3M2O2M5K8I7I7I7I6J7J6I7I5K4L4M3L4L1OO101O0O10000O2M2J6K5POhEgN^:[1h03M2N3N1N2N3M2N2N3M2N2Nf0ZO4L4L4L4L4L5K4L4L4L5K4L4L3M00XKTHR4d8N3M2O2M2O1N3M2O1N100O100O1O10O0100O1O100O100O0010O001O001O01O01O001O00001O010O00000000000001O00000000000000001O000000000000001O000000000000001O1O101N1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1N2M3M2O2M3N2M3M2O2M3N2M3M3N4K:Ga0^OZnd0"}, "label": "a motorcycle to the right of another"}]}
{"id": 133905, "image": "COCO_train2014_000000133905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch - bed with backpack\"."}], "task": "refering", "answer_template": "The \"couch - bed with backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 235, 236, 237, 241, 242, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 328, 329, 330], "bbox": [0.22389062499999998, 0.4811943793911007, 0.916453125, 0.9979859484777517], "iscrowd": 0, "area": 26570.069349999998, "rle": {"size": [427, 640], "counts": "Tnk13o<;D<G9N2N2N2N2N`1`NO02O1O1O1O2N1O1O1N2O2N1O1O1O1O2N1N2O1O1O1O2N1O1O1N2O2N00`MkFk0T9oNlFX1S9fNkF_1T9`NjFd1U9ZNiFk1U9UNiFo1V9oMkFS2T9lMmFT2S9jMnFX2Q9gMPGZ2o8eMQG]2n8aMTG`2k8_MUGc2i8^MWGb2i8]MWGd2i8\\MWGd2i8[MXGe2h8[MXGe2h8ZMZGe2f8[MZGe2f8ZM[Gf2e8TMaGl2`8lLgGT3X90000000001O000000001O000000001O00000000001O000000001O000000001O000O10001O0gNmEZOS:d0SFXOm9f0YFVOg9h0`FSOa9j0dFSO\\9j0jFSOV9k0nFSOR9k0QGTOo8j0UGTOk8i0ZGROi8m0ZGoNh8P1k1O0O101N2O1N2N101N2O1N2O1N2O1N2O0O2O1O_OGQD8o;IPD6Q<KoC4Q<MnC3Q<0mC0S<5iCJW<;dCE[<f00000O100000000000000001O001O001O001O0O2O001O00O100O100H8N2O100O1O001O1^OkNmDV1m:ROPEo0o:SOoDn0Q;SOnDm0R;TOlDm0T;TOkDl0U;UOiDl0W;UOgDk0Z;VOeDh0];a02N2O0O2N2N2N2N2N2I7I7J7J5K5K5KXWm06ehRO4K5L4G9H8C=J4N2M3N21O1O1O1N2O1O1O1O1O1O1O001O1O001O001O001O00aNeNoFZ1o8iNPGV1Q9lNnFR1R9POmFm0U9VOiF`0`9B_F8g9JWF5i9MWF2h91VFNj94UFKl97RFIm99RFFn9=PFCo9a0nE^OR:e0lE[OR:j0kEVOT:n0fETOY:W1[ElNb:P2M20100O10O10O100O010O1O1O00100O1O1O00100O1O1000O10O10000000000O100000000O10000000N2M2O3M2M4M2N201N100O2O0O2O000101N1O100O2N100O0001O00001O00001O00001O00001O0000001O00001O0O2O001O1O0O2O0100O100O10000O10O010O010O10O01O010O100O1O001O1O1O1O1O002N2O2D;A?D<D=D^ke0"}, "label": "couch - bed with backpack"}]}
{"id": 133905, "image": "COCO_train2014_000000133905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with a black backpack on it\"."}], "task": "refering", "answer_template": "The \"a bed with a black backpack on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 235, 236, 237, 241, 242, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 328, 329, 330], "bbox": [0.22389062499999998, 0.4811943793911007, 0.916453125, 0.9979859484777517], "iscrowd": 0, "area": 26570.069349999998, "rle": {"size": [427, 640], "counts": "Tnk13o<;D<G9N2N2N2N2N`1`NO02O1O1O1O2N1O1O1N2O2N1O1O1O1O2N1N2O1O1O1O2N1O1O1N2O2N00`MkFk0T9oNlFX1S9fNkF_1T9`NjFd1U9ZNiFk1U9UNiFo1V9oMkFS2T9lMmFT2S9jMnFX2Q9gMPGZ2o8eMQG]2n8aMTG`2k8_MUGc2i8^MWGb2i8]MWGd2i8\\MWGd2i8[MXGe2h8[MXGe2h8ZMZGe2f8[MZGe2f8ZM[Gf2e8TMaGl2`8lLgGT3X90000000001O000000001O000000001O00000000001O000000001O000000001O000O10001O0gNmEZOS:d0SFXOm9f0YFVOg9h0`FSOa9j0dFSO\\9j0jFSOV9k0nFSOR9k0QGTOo8j0UGTOk8i0ZGROi8m0ZGoNh8P1k1O0O101N2O1N2N101N2O1N2O1N2O1N2O0O2O1O_OGQD8o;IPD6Q<KoC4Q<MnC3Q<0mC0S<5iCJW<;dCE[<f00000O100000000000000001O001O001O001O0O2O001O00O100O100H8N2O100O1O001O1^OkNmDV1m:ROPEo0o:SOoDn0Q;SOnDm0R;TOlDm0T;TOkDl0U;UOiDl0W;UOgDk0Z;VOeDh0];a02N2O0O2N2N2N2N2N2I7I7J7J5K5K5KXWm06ehRO4K5L4G9H8C=J4N2M3N21O1O1O1N2O1O1O1O1O1O1O001O1O001O001O001O00aNeNoFZ1o8iNPGV1Q9lNnFR1R9POmFm0U9VOiF`0`9B_F8g9JWF5i9MWF2h91VFNj94UFKl97RFIm99RFFn9=PFCo9a0nE^OR:e0lE[OR:j0kEVOT:n0fETOY:W1[ElNb:P2M20100O10O10O100O010O1O1O00100O1O1O00100O1O1000O10O10000000000O100000000O10000000N2M2O3M2M4M2N201N100O2O0O2O000101N1O100O2N100O0001O00001O00001O00001O00001O0000001O00001O0O2O001O1O0O2O0100O100O10000O10O010O010O10O01O010O100O1O001O1O1O1O1O002N2O2D;A?D<D=D^ke0"}, "label": "a bed with a black backpack on it"}]}
{"id": 314130, "image": "COCO_train2014_000000314130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tug boat moving down the river\"."}], "task": "refering", "answer_template": "The \"tug boat moving down the river\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246], "bbox": [0.146875, 0.3537352245862884, 0.714890625, 0.732434988179669], "iscrowd": 0, "area": 29584.99864999999, "rle": {"size": [423, 640], "counts": "gSW13S=6K2K3M3M2O2M3M3M3O1O1O001O100O100O100O0100000000000O010O10000O10001N1O100000000000000000001O00000000000000000001O01O0000000001O001O1O000001DjNeDV1X;nNgDR1V;SOgDn0V;VOiDj0U;YOiDh0T;g0O1000000000001O0001O0000000000M3K5K56K1N2jMPEQ2Q;nMoDR2S;000000000000000O2O0000000000000000000000001O00000000000000000000001O0000000000000000000O101O0000000000O10000000000O1000000000000000O10O1000000000O1000000000000NiMTEW2k:jMUEV2j:4O100O1O1O1O1O1O100O1O1O2N1O1O100O1O1O1O1O1O101N1O1O100O100O100O100O2O0O1^OiLTGW3j8oLQGR3m8SMPGm2n8VMPGk2n8WMQGj2m8XMRGi2l8YMSGh2k8ZMTGg2j8l0O1N2O1N2O1N2O1N101L4O1000000000000O1000000000O10000000000000002N1O2N001O0O100000002N2N1O1O1O1O001O0000000O100000000000000000002\\MVGg0l8WOTGi0o8VNVG0Lj1P9jM^G;Ck1i9VNWFj1i9UNXFk1g9VNYFj1g9UNZFk1f9TN\\Fk1d9UN\\Fk1d9TN]Fl1`:O00000010O0001O000010O000001O01O01O00000010O0001O00010O000O101N100O2O0000001O00001O000O2O0000001O00001O0000001O00001O0J6\\Oe0J=_OUnZ2"}, "label": "tug boat moving down the river"}]}
{"id": 314130, "image": "COCO_train2014_000000314130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tug boat on water\"."}], "task": "refering", "answer_template": "The \"tug boat on water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246], "bbox": [0.146875, 0.3537352245862884, 0.714890625, 0.732434988179669], "iscrowd": 0, "area": 29584.99864999999, "rle": {"size": [423, 640], "counts": "gSW13S=6K2K3M3M2O2M3M3M3O1O1O001O100O100O100O0100000000000O010O10000O10001N1O100000000000000000001O00000000000000000001O01O0000000001O001O1O000001DjNeDV1X;nNgDR1V;SOgDn0V;VOiDj0U;YOiDh0T;g0O1000000000001O0001O0000000000M3K5K56K1N2jMPEQ2Q;nMoDR2S;000000000000000O2O0000000000000000000000001O00000000000000000000001O0000000000000000000O101O0000000000O10000000000O1000000000000000O10O1000000000O1000000000000NiMTEW2k:jMUEV2j:4O100O1O1O1O1O1O100O1O1O2N1O1O100O1O1O1O1O1O101N1O1O100O100O100O100O2O0O1^OiLTGW3j8oLQGR3m8SMPGm2n8VMPGk2n8WMQGj2m8XMRGi2l8YMSGh2k8ZMTGg2j8l0O1N2O1N2O1N2O1N101L4O1000000000000O1000000000O10000000000000002N1O2N001O0O100000002N2N1O1O1O1O001O0000000O100000000000000000002\\MVGg0l8WOTGi0o8VNVG0Lj1P9jM^G;Ck1i9VNWFj1i9UNXFk1g9VNYFj1g9UNZFk1f9TN\\Fk1d9UN\\Fk1d9TN]Fl1`:O00000010O0001O000010O000001O01O01O00000010O0001O00010O000O101N100O2O0000001O00001O000O2O0000001O00001O0000001O00001O0J6\\Oe0J=_OUnZ2"}, "label": "tug boat on water"}]}
{"id": 322324, "image": "COCO_train2014_000000322324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table on which a mug of beer is sitting on\"."}], "task": "refering", "answer_template": "The \"a table on which a mug of beer is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.003515625, 0.5223430962343096, 0.9999375, 0.9811715481171548], "iscrowd": 0, "area": 78237.71114999999, "rle": {"size": [478, 640], "counts": "m[12f>8H8G9H8H8H8G9H8J6M1O100O100O1O100O100O100O1O100O100O1UM_MXIb2h6aMUI_2k6dMQI]2o6fMnHZ2Q7jMkHW2U7kMhHV2X7mMeHS2Y7XMfFh0m1Q2]7TN`Hl1`7WN\\Hj1d7YNYHg1f7]NVHd1j7^NSHc1m7`NPH`1P8cNlG^1T8eNiG[1W8hNeGY1Z8kNbGV1^8mN_GS1a8PO[GQ1e8ROXGn0h8UOTGl0l8WOQGi0o8ZOmFg0R9]OjFd0V9b20000000000O10000000000O100000000000O010000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000O100000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O100000mJlG\\3T8\\LTHd3l7TL]Hk3b7nKfHR4Z7fKnHZ4R7dKPI\\4P7dKQI[4o6dKRI\\4n6cKSI]4m6bKUI]4j6dKVI\\4j6cKWI]4i6bKXI^4h6aKZI^4f6bKZI^4f6aK[I_4e6`K]I_4c6`K^I`4c6_K]Ia4c6^K^Ib4b6]K`Ib4a6\\K`Id4`6\\K`Id4a6ZK`If4`6YKbIf4^6YKcIg4^6XKbIh4[8O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O000000O100000000O100000000O100000000OWLPFf2P:ZMXF^2h9bMXF^2h9bMXF^2g9bMZF^2f9bMZF^2f9bMZF^2f9bMZF^2f9bMZF^2e9cM[F]2e9cM[F]2e9bM\\F^2d9bM\\F^2d9bM\\F^2c9cM]F]2c9cM]F]2c9cM]F<@f0S:nN]F;Fb0m9RO^F;K>f9XO_F91:`9]O_F876Y9C`F6=2Q9JbF4b0Mi82eF0h0I`89iFMh0J]8;kFJi0KY8>nFFk0KX8>mFFl0LZ8;jFHm0MS:2nENR:2nENR:1oEOQ:ORF0n9E]F;c9ZOhFf0h;000001O000000001O00000000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O00000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O2N1[NSOWEo0b:]OXEd0b:GXE:h:KRE6n:OkD3V;1dD0\\;5^DLb;9XDHh;g11O000000001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1]Oc0UOPD"}, "label": "a table on which a mug of beer is sitting on"}]}
{"id": 322324, "image": "COCO_train2014_000000322324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with two beers on it\"."}], "task": "refering", "answer_template": "The \"a table with two beers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.003515625, 0.5223430962343096, 0.9999375, 0.9811715481171548], "iscrowd": 0, "area": 78237.71114999999, "rle": {"size": [478, 640], "counts": "m[12f>8H8G9H8H8H8G9H8J6M1O100O100O1O100O100O100O1O100O100O1UM_MXIb2h6aMUI_2k6dMQI]2o6fMnHZ2Q7jMkHW2U7kMhHV2X7mMeHS2Y7XMfFh0m1Q2]7TN`Hl1`7WN\\Hj1d7YNYHg1f7]NVHd1j7^NSHc1m7`NPH`1P8cNlG^1T8eNiG[1W8hNeGY1Z8kNbGV1^8mN_GS1a8PO[GQ1e8ROXGn0h8UOTGl0l8WOQGi0o8ZOmFg0R9]OjFd0V9b20000000000O10000000000O100000000000O010000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000O100000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O100000mJlG\\3T8\\LTHd3l7TL]Hk3b7nKfHR4Z7fKnHZ4R7dKPI\\4P7dKQI[4o6dKRI\\4n6cKSI]4m6bKUI]4j6dKVI\\4j6cKWI]4i6bKXI^4h6aKZI^4f6bKZI^4f6aK[I_4e6`K]I_4c6`K^I`4c6_K]Ia4c6^K^Ib4b6]K`Ib4a6\\K`Id4`6\\K`Id4a6ZK`If4`6YKbIf4^6YKcIg4^6XKbIh4[8O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O000000O100000000O100000000O100000000OWLPFf2P:ZMXF^2h9bMXF^2h9bMXF^2g9bMZF^2f9bMZF^2f9bMZF^2f9bMZF^2f9bMZF^2e9cM[F]2e9cM[F]2e9bM\\F^2d9bM\\F^2d9bM\\F^2c9cM]F]2c9cM]F]2c9cM]F<@f0S:nN]F;Fb0m9RO^F;K>f9XO_F91:`9]O_F876Y9C`F6=2Q9JbF4b0Mi82eF0h0I`89iFMh0J]8;kFJi0KY8>nFFk0KX8>mFFl0LZ8;jFHm0MS:2nENR:2nENR:1oEOQ:ORF0n9E]F;c9ZOhFf0h;000001O000000001O00000000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O00000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O2N1[NSOWEo0b:]OXEd0b:GXE:h:KRE6n:OkD3V;1dD0\\;5^DLb;9XDHh;g11O000000001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1]Oc0UOPD"}, "label": "a table with two beers on it"}]}
{"id": 125724, "image": "COCO_train2014_000000125724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two pairs of feet , both wearing high socks , behind a man sitting on a skateboard in motion\"."}], "task": "refering", "answer_template": "The \"two pairs of feet , both wearing high socks , behind a man sitting on a skateboard in motion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 25, 26, 27, 28, 29, 43, 44, 45, 46, 62, 63], "bbox": [0.3628, 0.0, 0.6405, 0.30405333333333334], "iscrowd": 0, "area": 5626.8025, "rle": {"size": [375, 500], "counts": "kdR21f;1O1O3M2N2N2N2N2N2N2N2N2N2N4L4L4L4L6J4L2N2N1O2N2N2N2N2N2NlM_FT2_9lMbFV2[9jMfFW2X9iMiFX2U9hMlFY2R9gMoFZ2o8fMRG[2l8eMTG]2i8dMZG[2d8eM`GX2_8hMeGV2Y8jMjGU2T8lMoGR2o7nMTHQ2j7oMZHn1e7RN]Hn1a7RNaHn1]7RNeHn1Y7RNiHn1U7RNmHm1R7SNQIl1m6TNUIk1j6UNYIi1f6WN]If1c6\\N]Ic1b6gNVIW1j6jNVIV1i6kNXIS1h6nNYIo0h6ROYIk0h6VOYIg0h6ZOXId0i6]OXI`0i6AXI<i6EWI9j6HWI5j6KWI3j6NWI0i61WIMj63WIKj66TIJm67RIIn68PIHQ78oHGR7:lHFU7;jHEV7;jHEW7;hHDY7=fHC[7=dHB]7?bHA_7?`HAa7?^H@d7`0ZH@h7`0WH_Ok7a0TH_Om7a0RH^OQ8a0nG^OT8b0kG^OV8b0iG]OX8d0gG[O[8e0dGZO^8f0aGXOb8g0^GXOd8h0[GVOh8j0XGTOi8m0WGQOk8o0TGoNo8o0n0]O[^4H[bK5J7I6J6L21O001O001O1O00000O100O1O2O0O2N1O2N1N4L3L5LelQ2"}, "label": "two pairs of feet , both wearing high socks , behind a man sitting on a skateboard in motion"}]}
{"id": 125724, "image": "COCO_train2014_000000125724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with white socks with red lining\"."}], "task": "refering", "answer_template": "The \"the man with white socks with red lining\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 25, 26, 27, 28, 29, 43, 44, 45, 46, 62, 63], "bbox": [0.3628, 0.0, 0.6405, 0.30405333333333334], "iscrowd": 0, "area": 5626.8025, "rle": {"size": [375, 500], "counts": "kdR21f;1O1O3M2N2N2N2N2N2N2N2N2N2N4L4L4L4L6J4L2N2N1O2N2N2N2N2N2NlM_FT2_9lMbFV2[9jMfFW2X9iMiFX2U9hMlFY2R9gMoFZ2o8fMRG[2l8eMTG]2i8dMZG[2d8eM`GX2_8hMeGV2Y8jMjGU2T8lMoGR2o7nMTHQ2j7oMZHn1e7RN]Hn1a7RNaHn1]7RNeHn1Y7RNiHn1U7RNmHm1R7SNQIl1m6TNUIk1j6UNYIi1f6WN]If1c6\\N]Ic1b6gNVIW1j6jNVIV1i6kNXIS1h6nNYIo0h6ROYIk0h6VOYIg0h6ZOXId0i6]OXI`0i6AXI<i6EWI9j6HWI5j6KWI3j6NWI0i61WIMj63WIKj66TIJm67RIIn68PIHQ78oHGR7:lHFU7;jHEV7;jHEW7;hHDY7=fHC[7=dHB]7?bHA_7?`HAa7?^H@d7`0ZH@h7`0WH_Ok7a0TH_Om7a0RH^OQ8a0nG^OT8b0kG^OV8b0iG]OX8d0gG[O[8e0dGZO^8f0aGXOb8g0^GXOd8h0[GVOh8j0XGTOi8m0WGQOk8o0TGoNo8o0n0]O[^4H[bK5J7I6J6L21O001O001O1O00000O100O1O2O0O2N1O2N1N4L3L5LelQ2"}, "label": "the man with white socks with red lining"}]}
{"id": 125724, "image": "COCO_train2014_000000125724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing red and white shorts on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man wearing red and white shorts on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 48, 49, 50, 58, 59, 65, 66, 67, 68, 75, 76, 77, 83, 84, 85, 86, 92, 93, 94, 95, 96, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191], "bbox": [0.15506, 0.09677333333333334, 0.8325799999999999, 0.8405866666666667], "iscrowd": 0, "area": 38240.36925000001, "rle": {"size": [375, 500], "counts": "ffl03c;3L4L4L3M4M3L4L3M4M3L3N3M3M3M1O2N1O1O1O101N100O1O100O2N100O100O2N100O100O2O0O10000O2ZGkMh7U2UHQNg7Q2UHUNh7m1SHZNh7j1RH\\Nk7g1PH^Nl7h2O1N2O1N2O2N1N2O1N2O1O2N2N1O2N2M2O2N2N1O2M3N1O2N2N1101N2M2O2M3N2M3M5L3L5L3L100O100O1O100O100O100O1O100[LaHP3S810O0010O010O0010O0010O010O00010O010O01O01O010O01O010O01O01O010O01O01O010O010O00010O010O0010O0010O0010O010O00010O010O00001N101O001O0O101O001O000O2O001O000O2O001O000O2O001O00001N101O00001N10000000O0100O100O10O0100O1O10O0100O100O00100O10OO1O1O1O2N1O1O1O1O2O0O1O1O1O2M2O1O1N2O2M10000O10000000000000000000000000000000O1000O1N2O1O1N2O1O11O000O2O0O1O2O0O101N100O2O0000001N10001O1O2M3N1O2N2N2M3N2N2N1N3N2N3M2N2M3N3M2N2N3L3N2N2N3M2M3N2N3N1O1O2O0O1O3N3L4L4L4M3L4L3N2M3M3M3N2M3M3N2M3M3Kken0"}, "label": "a man wearing red and white shorts on a skateboard"}]}
{"id": 125724, "image": "COCO_train2014_000000125724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man wearing a blue cap and white and red boxer shorts sitting on a skateboard moving down a street\"."}], "task": "refering", "answer_template": "The \"a young man wearing a blue cap and white and red boxer shorts sitting on a skateboard moving down a street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 48, 49, 50, 58, 59, 65, 66, 67, 68, 75, 76, 77, 83, 84, 85, 86, 92, 93, 94, 95, 96, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191], "bbox": [0.15506, 0.09677333333333334, 0.8325799999999999, 0.8405866666666667], "iscrowd": 0, "area": 38240.36925000001, "rle": {"size": [375, 500], "counts": "ffl03c;3L4L4L3M4M3L4L3M4M3L3N3M3M3M1O2N1O1O1O101N100O1O100O2N100O100O2N100O100O2O0O10000O2ZGkMh7U2UHQNg7Q2UHUNh7m1SHZNh7j1RH\\Nk7g1PH^Nl7h2O1N2O1N2O2N1N2O1N2O1O2N2N1O2N2M2O2N2N1O2M3N1O2N2N1101N2M2O2M3N2M3M5L3L5L3L100O100O1O100O100O100O1O100[LaHP3S810O0010O010O0010O0010O010O00010O010O01O01O010O01O010O01O01O010O01O01O010O010O00010O010O0010O0010O0010O010O00010O010O00001N101O001O0O101O001O000O2O001O000O2O001O000O2O001O00001N101O00001N10000000O0100O100O10O0100O1O10O0100O100O00100O10OO1O1O1O2N1O1O1O1O2O0O1O1O1O2M2O1O1N2O2M10000O10000000000000000000000000000000O1000O1N2O1O1N2O1O11O000O2O0O1O2O0O101N100O2O0000001N10001O1O2M3N1O2N2N2M3N2N2N1N3N2N3M2N2M3N3M2N2N3L3N2N2N3M2M3N2N3N1O1O2O0O1O3N3L4L4L4M3L4L3N2M3M3M3N2M3M3N2M3M3Kken0"}, "label": "a young man wearing a blue cap and white and red boxer shorts sitting on a skateboard moving down a street"}]}
{"id": 527139, "image": "COCO_train2014_000000527139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the booth to the left of the vase\"."}], "task": "refering", "answer_template": "The \"the back of the booth to the left of the vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368, 369, 370, 371, 391, 392, 393, 394, 414, 415, 416, 437, 438, 439, 460, 461, 462, 463, 483, 484, 485, 486, 487], "bbox": [0.007734375, 0.6420198675496688, 0.207046875, 1.0], "iscrowd": 0, "area": 20334.083150000002, "rle": {"size": [604, 640], "counts": "o]3Y3Z<Y3000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000hLSDAm;<aDYO_;c0PEROP;k0_EiNa:S1mEcNS:Y1RFfNn9W1VFhNj9T1[FkNd9S1`FlN`9P1eFoN[9m0jFROV9k0nFTOR9h0SGWOm8g0VGXOj8g0XGXOh8g0ZGXOf8g0\\GXOd8g0^GXOb8g0`GXO`8f0dGXO\\8g0fGXOZ8g0hGXOX8g0jGXOV8g0lGXOT8g0nGXOR8g0PHXOP8g0RHXOn7g0THXOl7g0VHXOj7g0XHXOh7g0ZHXOf7g0\\HXOd7g0^HXOc7f0_HYOa7f0aHYO_7f0cHYO]7e0fHZOZ7d0iH[OW7c0lH\\OW7?lH@X7;jHDZ76iHI[71hHN\\7LgH3]7GfH8P=O1O2N2N1O2NfQ[9"}, "label": "the back of the booth to the left of the vase"}]}
{"id": 527139, "image": "COCO_train2014_000000527139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pillos behind the antiqus\"."}], "task": "refering", "answer_template": "The \"pillos behind the antiqus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368, 369, 370, 371, 391, 392, 393, 394, 414, 415, 416, 437, 438, 439, 460, 461, 462, 463, 483, 484, 485, 486, 487], "bbox": [0.007734375, 0.6420198675496688, 0.207046875, 1.0], "iscrowd": 0, "area": 20334.083150000002, "rle": {"size": [604, 640], "counts": "o]3Y3Z<Y3000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000hLSDAm;<aDYO_;c0PEROP;k0_EiNa:S1mEcNS:Y1RFfNn9W1VFhNj9T1[FkNd9S1`FlN`9P1eFoN[9m0jFROV9k0nFTOR9h0SGWOm8g0VGXOj8g0XGXOh8g0ZGXOf8g0\\GXOd8g0^GXOb8g0`GXO`8f0dGXO\\8g0fGXOZ8g0hGXOX8g0jGXOV8g0lGXOT8g0nGXOR8g0PHXOP8g0RHXOn7g0THXOl7g0VHXOj7g0XHXOh7g0ZHXOf7g0\\HXOd7g0^HXOc7f0_HYOa7f0aHYO_7f0cHYO]7e0fHZOZ7d0iH[OW7c0lH\\OW7?lH@X7;jHDZ76iHI[71hHN\\7LgH3]7GfH8P=O1O2N2N1O2NfQ[9"}, "label": "pillos behind the antiqus"}]}
{"id": 527139, "image": "COCO_train2014_000000527139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase that is shorter and rounder\"."}], "task": "refering", "answer_template": "The \"a vase that is shorter and rounder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 371, 372, 373, 374, 375, 376, 377, 378, 379, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 463, 464, 465, 466, 467, 468, 469, 470, 471, 487, 488, 489, 490, 491, 492, 493], "bbox": [0.11907812499999999, 0.5585099337748344, 0.5262499999999999, 0.9914238410596026], "iscrowd": 0, "area": 52445.33110000002, "rle": {"size": [604, 640], "counts": "\\Z]15ab0a0^O=D8I7J6J5L5J6K5M3M3L3N3M2N3M2QMQMnEQ3P:QMnER3P:oLnES3P:oLmEU3Q:lLmEV3Q:lLmEW3Q:jLmEX3R:iLlEZ3Q:hLmEZ3Q:hLmE[3m9jLPFY3k9mLRFV3k9lLSFV3l9kLRFX3l9iLRFY3n9gLPF\\3n9eLPF]3o9dLoE^3P:cLmEa3R:_LlEc3S:^LkEe3S:\\LkEf3T:[LiEi3V:WLgEl3Y:TLdEP4[:PLbES4_:lK_EV4a:jK\\EX4f:gKWE\\4j:cKSE_4n:aKPE`4R;_KkDc4V;]KgDe4[;ZKbDh4_;XK^Dj4c;VK[Dk4e;VKXDl4h;UKUDm4k;R1000000000000O2O00000000000000000N2O1N2O1N2N2O1N2O2M2N2O1N2O1000000000000000000000001O0000000000000000000001O0000000010O00000001O000000001O000000001O0001O0001OO10000000O10O10000000000O01000000000O1000O100000O1000000O01000O10000O10O10O10000O100O10O10O10000O10000O01000O100O11O001N2O1N101N2O0O2O1O0O2O1N101N2O0O2O1N101O1YJgDh3Z;VLkDg3V;WLmDf3U;XLoDe3R;YLREc3o:[LUEc3l:[LVEc3k:\\LWEc3i:\\LYEb3h:\\L[Ec3e:\\L]Ec3d:[L^Ec3c:\\L_Eb3b:\\LaEb3`:]LaEb3`:\\LcEb3^:[LfEc3[:[LhEc3Y:[LjEb3Y:[LjEc3W:ZLmEd3f<O2N2N2N2N2N2N3M2N2N3M2N2N3J5L4L5J5L4L5J5L5J>C<D=B=D]Wb5"}, "label": "a vase that is shorter and rounder"}]}
{"id": 527139, "image": "COCO_train2014_000000527139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a short , wide glass vase with pink and purple flowers\"."}], "task": "refering", "answer_template": "The \"a short , wide glass vase with pink and purple flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 371, 372, 373, 374, 375, 376, 377, 378, 379, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 463, 464, 465, 466, 467, 468, 469, 470, 471, 487, 488, 489, 490, 491, 492, 493], "bbox": [0.11907812499999999, 0.5585099337748344, 0.5262499999999999, 0.9914238410596026], "iscrowd": 0, "area": 52445.33110000002, "rle": {"size": [604, 640], "counts": "\\Z]15ab0a0^O=D8I7J6J5L5J6K5M3M3L3N3M2N3M2QMQMnEQ3P:QMnER3P:oLnES3P:oLmEU3Q:lLmEV3Q:lLmEW3Q:jLmEX3R:iLlEZ3Q:hLmEZ3Q:hLmE[3m9jLPFY3k9mLRFV3k9lLSFV3l9kLRFX3l9iLRFY3n9gLPF\\3n9eLPF]3o9dLoE^3P:cLmEa3R:_LlEc3S:^LkEe3S:\\LkEf3T:[LiEi3V:WLgEl3Y:TLdEP4[:PLbES4_:lK_EV4a:jK\\EX4f:gKWE\\4j:cKSE_4n:aKPE`4R;_KkDc4V;]KgDe4[;ZKbDh4_;XK^Dj4c;VK[Dk4e;VKXDl4h;UKUDm4k;R1000000000000O2O00000000000000000N2O1N2O1N2N2O1N2O2M2N2O1N2O1000000000000000000000001O0000000000000000000001O0000000010O00000001O000000001O000000001O0001O0001OO10000000O10O10000000000O01000000000O1000O100000O1000000O01000O10000O10O10O10000O100O10O10O10000O10000O01000O100O11O001N2O1N101N2O0O2O1O0O2O1N101N2O0O2O1N101O1YJgDh3Z;VLkDg3V;WLmDf3U;XLoDe3R;YLREc3o:[LUEc3l:[LVEc3k:\\LWEc3i:\\LYEb3h:\\L[Ec3e:\\L]Ec3d:[L^Ec3c:\\L_Eb3b:\\LaEb3`:]LaEb3`:\\LcEb3^:[LfEc3[:[LhEc3Y:[LjEb3Y:[LjEc3W:ZLmEd3f<O2N2N2N2N2N2N3M2N2N3M2N2N3J5L4L5J5L4L5J5L5J>C<D=B=D]Wb5"}, "label": "a short , wide glass vase with pink and purple flowers"}]}
{"id": 297764, "image": "COCO_train2014_000000297764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana on right on the tray\"."}], "task": "refering", "answer_template": "The \"banana on right on the tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 197, 198, 199, 200, 220, 221, 222, 223, 244, 245, 246], "bbox": [0.5641406250000001, 0.33142857142857146, 0.87115625, 0.6779156908665106], "iscrowd": 0, "area": 16483.733200000002, "rle": {"size": [427, 640], "counts": "jgf4?h<>B7I7J6I7I7J6I8D9N2O1N2O1N1O2O1N2O1N2N2O1N2O1N1O2O0O2O1N1O2O0O2N101N2O0O2O0O2O001N1000000O10000O2O000O1000000O10001N1000000O10000O101O000O10000O1000000O10000O10000OjMVGMj83XGKh84ZGJg85[GJe86\\GId86_GH`88bGG^89cGF]89eGF[8:gGDY8;iGDW8;kGCU8>lGAT8?nG_OR8`0PH_OP8a0RH]On7c0THZOm7e0UHZOk7f0WHXOh7i0ZHUOf7j0\\HUOd7k0^HROc7n0_HPOa7P1`HoN`7P1cHlN^7U1dHeN`7[1X2O10000O1000000O101O0O10000O101O000O1000001O0O1000000O2O0000000O101O00000O10001O000O1000001O0O10000O2O0O1O101N1O100O2O0O1O2O0O1O101N1O101O0000001O0000001O00001O001O2N3M2L5I6IPQR1"}, "label": "banana on right on the tray"}]}
{"id": 297764, "image": "COCO_train2014_000000297764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana to the right of two others\"."}], "task": "refering", "answer_template": "The \"a banana to the right of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 197, 198, 199, 200, 220, 221, 222, 223, 244, 245, 246], "bbox": [0.5641406250000001, 0.33142857142857146, 0.87115625, 0.6779156908665106], "iscrowd": 0, "area": 16483.733200000002, "rle": {"size": [427, 640], "counts": "jgf4?h<>B7I7J6I7I7J6I8D9N2O1N2O1N1O2O1N2O1N2N2O1N2O1N1O2O0O2O1N1O2O0O2N101N2O0O2O0O2O001N1000000O10000O2O000O1000000O10001N1000000O10000O101O000O10000O1000000O10000O10000OjMVGMj83XGKh84ZGJg85[GJe86\\GId86_GH`88bGG^89cGF]89eGF[8:gGDY8;iGDW8;kGCU8>lGAT8?nG_OR8`0PH_OP8a0RH]On7c0THZOm7e0UHZOk7f0WHXOh7i0ZHUOf7j0\\HUOd7k0^HROc7n0_HPOa7P1`HoN`7P1cHlN^7U1dHeN`7[1X2O10000O1000000O101O0O10000O101O000O1000001O0O1000000O2O0000000O101O00000O10001O000O1000001O0O10000O2O0O1O101N1O100O2O0O1O2O0O1O101N1O101O0000001O0000001O00001O001O2N3M2L5I6IPQR1"}, "label": "a banana to the right of two others"}]}
{"id": 297764, "image": "COCO_train2014_000000297764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana in fruit bunch\"."}], "task": "refering", "answer_template": "The \"banana in fruit bunch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 170, 171, 172, 173, 193, 194, 195, 196, 217, 218, 219], "bbox": [0.41231249999999997, 0.28695550351288057, 0.775140625, 0.6532552693208431], "iscrowd": 0, "area": 12511.740650000003, "rle": {"size": [427, 640], "counts": "_Z^34S=4K6K5J6M3L3N3L4M3L4M2M4M3L4M2N3N2M3N2N1O2M3N2N2N1N3N2N2N1O2M3N2N2N1N3N2N1O001N2O10O01O100O00100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100O1O01O0N2O2N1O1N3N1O1O2M2O2N2N1N3N2N1O2M2O2M3K4L5L4K4L5K4M4K5K401O10O01O00100O001O10O01O1O010O001O10O01O1O010O001O10O01O1O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10001O0O2O001O0O101O001N101O001N1000000O1000000O1000001O000000000000000000000000000000000000000000001O2N1O1O1O101N1O1ORmk1"}, "label": "banana in fruit bunch"}]}
{"id": 297764, "image": "COCO_train2014_000000297764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle banana in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the middle banana in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 170, 171, 172, 173, 193, 194, 195, 196, 217, 218, 219], "bbox": [0.41231249999999997, 0.28695550351288057, 0.775140625, 0.6532552693208431], "iscrowd": 0, "area": 12511.740650000003, "rle": {"size": [427, 640], "counts": "_Z^34S=4K6K5J6M3L3N3L4M3L4M2M4M3L4M2N3N2M3N2N1O2M3N2N2N1N3N2N2N1O2M3N2N2N1N3N2N1O001N2O10O01O100O00100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100O1O01O0N2O2N1O1N3N1O1O2M2O2N2N1N3N2N1O2M2O2M3K4L5L4K4L5K4M4K5K401O10O01O00100O001O10O01O1O010O001O10O01O1O010O001O10O01O1O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10001O0O2O001O0O101O001N101O001N1000000O1000000O1000001O000000000000000000000000000000000000000000001O2N1O1O1O101N1O1ORmk1"}, "label": "the middle banana in the right hand picture"}]}
{"id": 256809, "image": "COCO_train2014_000000256809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 60, 61, 77, 78, 79, 94, 95, 96, 112, 113, 114, 129, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167, 168, 184], "bbox": [0.21252000000000001, 0.1685285285285285, 0.44898000000000005, 0.8516816816816817], "iscrowd": 0, "area": 8210.92165, "rle": {"size": [333, 500], "counts": "YgR16h9a0_Ob0]Ob0^Oc0]Ob0C=M01^OYI]Mg6a2c0N2N3L3N2O29F:F:FiNRIXOg6f0cIRO\\6m0mImNQ6R1XJhNf5V1dJcN\\5[1nJ^NP5a1YKXNf4g1P2N3N1O1O2O01O1O010O1O1O010O1O00100O00100O001O10O01O1O1YI`N`4`1nJTOP5m0^JF_5<nI9o5T2N2N3MEYJSLe5h3a0K6K5J6K5K5J66I<E;E:F;E;EiMZKlNa4T1]LoMc3o1oLbMP3]2TMaMk2^2XM`Mh2a1mJUO`2XOb2a1RKTO_2ZO]2a1YKPO_7o0eHnN[7Q1iHkNX7T1mHhNS7W1QIeNP7Z1Q1OXOdNcH[1`7eN^HZ1e7gNXHX1k7iNRHV1Q8kNlGT1W8mNfGR1]8oN`GP1c8QOZGn0i8SOTGl0f8oN^GX1f843N3L4L3M4L2N2O02O1O11O5K4L4L1ON2O1N2=C4L`ei2"}, "label": "the giraffe on the left"}]}
{"id": 256809, "image": "COCO_train2014_000000256809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller giraffe\"."}], "task": "refering", "answer_template": "The \"the taller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 60, 61, 77, 78, 79, 94, 95, 96, 112, 113, 114, 129, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167, 168, 184], "bbox": [0.21252000000000001, 0.1685285285285285, 0.44898000000000005, 0.8516816816816817], "iscrowd": 0, "area": 8210.92165, "rle": {"size": [333, 500], "counts": "YgR16h9a0_Ob0]Ob0^Oc0]Ob0C=M01^OYI]Mg6a2c0N2N3L3N2O29F:F:FiNRIXOg6f0cIRO\\6m0mImNQ6R1XJhNf5V1dJcN\\5[1nJ^NP5a1YKXNf4g1P2N3N1O1O2O01O1O010O1O1O010O1O00100O00100O001O10O01O1O1YI`N`4`1nJTOP5m0^JF_5<nI9o5T2N2N3MEYJSLe5h3a0K6K5J6K5K5J66I<E;E:F;E;EiMZKlNa4T1]LoMc3o1oLbMP3]2TMaMk2^2XM`Mh2a1mJUO`2XOb2a1RKTO_2ZO]2a1YKPO_7o0eHnN[7Q1iHkNX7T1mHhNS7W1QIeNP7Z1Q1OXOdNcH[1`7eN^HZ1e7gNXHX1k7iNRHV1Q8kNlGT1W8mNfGR1]8oN`GP1c8QOZGn0i8SOTGl0f8oN^GX1f843N3L4L3M4L2N2O02O1O11O5K4L4L1ON2O1N2=C4L`ei2"}, "label": "the taller giraffe"}]}
{"id": 256809, "image": "COCO_train2014_000000256809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shorter giraffe is standing in the middle of the field\"."}], "task": "refering", "answer_template": "The \"a shorter giraffe is standing in the middle of the field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 82, 83, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 151, 152, 153, 169, 170], "bbox": [0.42728, 0.2727327327327327, 0.65562, 0.7714414414414414], "iscrowd": 0, "area": 6307.175699999998, "rle": {"size": [333, 500], "counts": "chU2j0b9k0UO:G:E6J100O1O1O100O1O[Oe03N2M3M3N3L3M3M3N1N1O1O10O0QOkHhNT7U1ZIaNf6[1W1M3O1100O1O1O100O1O010O1O1O100O1O1fHXNn5i1jI_NV6a1bIgN]6Z1ZIoNe6R1SIWOk6j0mH^OS7n1O1O1O100O1O1OXOi0L3N2M3N3M2M3N2M4M2N2M3N3_O`000O100O100O1O100O100O100O100O100OEUOWGk0j8VOTGi0n8XOPGg0R9ZOlFf0U9[OjFc0X9^OfFa0\\9_OcFa0^952N2N1ON31N3M300ON3L32O1O2M01OO1O10ijg1"}, "label": "a shorter giraffe is standing in the middle of the field"}]}
{"id": 256809, "image": "COCO_train2014_000000256809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the right of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe to the right of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 82, 83, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 151, 152, 153, 169, 170], "bbox": [0.42728, 0.2727327327327327, 0.65562, 0.7714414414414414], "iscrowd": 0, "area": 6307.175699999998, "rle": {"size": [333, 500], "counts": "chU2j0b9k0UO:G:E6J100O1O1O100O1O[Oe03N2M3M3N3L3M3M3N1N1O1O10O0QOkHhNT7U1ZIaNf6[1W1M3O1100O1O1O100O1O010O1O1O100O1O1fHXNn5i1jI_NV6a1bIgN]6Z1ZIoNe6R1SIWOk6j0mH^OS7n1O1O1O100O1O1OXOi0L3N2M3N3M2M3N2M4M2N2M3N3_O`000O100O100O1O100O100O100O100O100OEUOWGk0j8VOTGi0n8XOPGg0R9ZOlFf0U9[OjFc0X9^OfFa0\\9_OcFa0^952N2N1ON31N3M300ON3L32O1O2M01OO1O10ijg1"}, "label": "a giraffe to the right of another giraffe"}]}
{"id": 68397, "image": "COCO_train2014_000000068397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fried egg in a cupcake paper\"."}], "task": "refering", "answer_template": "The \"a fried egg in a cupcake paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 267, 268], "bbox": [0.505625, 0.5099765807962529, 0.722421875, 0.7510070257611241], "iscrowd": 0, "area": 10887.9351, "rle": {"size": [427, 640], "counts": "o[W46P=7I7K5L4L3M3M4L3M3M3M3M3M3M3M2N3N100O2N100O2N100O2O0O1O100O1O101N1O101N1O101N101O001O1O1N2O0000000000O1000O100000000000000O10001O001O00001O001O001O1O0010O01O00001O000000O1O1O2N1O101O00001O00001N1000010O00010O00001O001O001O0O101O001O001N1O3N2M2O2M2O2M3M2N3M2N3M3M2N3L3M6J8G9HinY2"}, "label": "a fried egg in a cupcake paper"}]}
{"id": 68397, "image": "COCO_train2014_000000068397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an egg cup on a breakfast plate\"."}], "task": "refering", "answer_template": "The \"an egg cup on a breakfast plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 267, 268], "bbox": [0.505625, 0.5099765807962529, 0.722421875, 0.7510070257611241], "iscrowd": 0, "area": 10887.9351, "rle": {"size": [427, 640], "counts": "o[W46P=7I7K5L4L3M3M4L3M3M3M3M3M3M3M2N3N100O2N100O2N100O2O0O1O100O1O101N1O101N1O101N101O001O1O1N2O0000000000O1000O100000000000000O10001O001O00001O001O001O1O0010O01O00001O000000O1O1O2N1O101O00001O00001N1000010O00010O00001O001O001O0O101O001O001N1O3N2M2O2M2O2M3M2N3M2N3M3M2N3L3M6J8G9HinY2"}, "label": "an egg cup on a breakfast plate"}]}
{"id": 68397, "image": "COCO_train2014_000000068397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chocolate muffin\"."}], "task": "refering", "answer_template": "The \"the chocolate muffin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 215, 216, 217, 218, 239, 240, 241], "bbox": [0.349859375, 0.3893208430913349, 0.5274375, 0.6890632318501171], "iscrowd": 0, "area": 11565.01865, "rle": {"size": [427, 640], "counts": "jcm24R=8H8G9H8I7K4M4L4K5L4L4K5L4L4K5M3M3L4M3M1O1N101O0O101O0O2O001O0O2O001N101O0O2O001O0O2O001N101O0O2O001O0O10000O10000O1000000000000001O000000000000000O100000000000000000000O101N1O2O1N2O0O2N2O1N101M3L3M4L4L5K4L4L5J5J8H7I8H;E;E;E;EVgm3"}, "label": "the chocolate muffin"}]}
{"id": 68397, "image": "COCO_train2014_000000068397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown muffin on the white plate\"."}], "task": "refering", "answer_template": "The \"a brown muffin on the white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 215, 216, 217, 218, 239, 240, 241], "bbox": [0.349859375, 0.3893208430913349, 0.5274375, 0.6890632318501171], "iscrowd": 0, "area": 11565.01865, "rle": {"size": [427, 640], "counts": "jcm24R=8H8G9H8I7K4M4L4K5L4L4K5L4L4K5M3M3L4M3M1O1N101O0O101O0O2O001O0O2O001N101O0O2O001O0O2O001N101O0O2O001O0O10000O10000O1000000000000001O000000000000000O100000000000000000000O101N1O2O1N2O0O2N2O1N101M3L3M4L4L5K4L4L5J5J8H7I8H;E;E;E;EVgm3"}, "label": "a brown muffin on the white plate"}]}
{"id": 494382, "image": "COCO_train2014_000000494382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goose in center\"."}], "task": "refering", "answer_template": "The \"a goose in center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 143, 144, 145, 146, 160, 161, 177, 178, 192, 193, 194, 195, 196, 206, 208, 209, 210, 211, 212, 213, 222, 223, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 257, 260, 261, 262, 263, 278, 279, 295, 296], "bbox": [0.03491666666666667, 0.31484375, 0.64325, 0.7611875], "iscrowd": 0, "area": 25164.125350000002, "rle": {"size": [640, 480], "counts": "eP;1nc02N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2O0O1O2N100O2N1O1O2O0O1O2N100O2O0O2EgNm]OY1Rb0hNm]OZ1Qb0gNo]OY1Qb0fNP^OZ1oa0gNP^O[1na0fNR^OZ1ma0fNS^O\\1la0dNT^O\\1ka0eNU^O\\1ja0cNV^O^1ja0bNV^O_1ha0bNW^O_1ia0aNW^O_1ia0`NX^Oa1ga0_NX^Ob1ha0]NY^Od1ea0]NZ^Od1fa0\\NZ^Oe1ea0ZN\\^Oe1ea0[NZ^Oe1ga0[NY^Oe1ga0ZNY^Of1ha0ZNX^Of1ha0ZNX^Oe1ha0]NV^Oc1ka091N4L5L3L4M3L5K4M4K5K4M4K5K4M2M2N3N2MUf80lYG0O4M5K4L5K7I<D3M2N3M3M3L3N3M3M0000000001O00001OBVNf^Oj1Za0XNd^Oh1\\a0YNd^Of1\\a0[Nc^Oe1]a0\\Nc^Oc1]a0^Nb^Ob1]a0`Nb^O`1^a0aNb^O^1^a0cNa^O]1_a0dN`^O^1^a0cNa^O^1]a0e0N1O1O1O1O2M2O1O1O1O1O1N2O1O1O1O2N2M3N2N2N2N2N2N2M3N2N0000000O10000000O10O1000000000000O3N2N4L4L4L4k@cK_>a4YAcKg>k400000000000000001O0G9@`0F:J^NiLgBV3g<aMVC^2g<iMUCV2i<QNSCn1k<YNQCf1m<bNnB^1n<jNnBU1P=ROmBm0P=ZOlBe0R=BjB>S=X3N3L3N3M2N3O10O10O0100O1000O0100O100O010O1000O0100O100O100_JTDm2m;RMXDi2i;VM]Dc2e;]M_D^2b;bM_D[2c;eM^DY2c;fM_DW2c;iM^DU2c;kM_DQ2c;oM^Dn1d;RN]Dl1d;SN^Dj1d;VN]Dh1d;XN^Dd1e;[N\\Dc1f;\\N\\D`1f;_N\\D_1f;`N\\D\\1g;bN\\D[1e;eN\\DX1g;fN\\DV1g;hN[DV1f;iN]DS1f;lN[DR1f;mN]Do0f;PO[Dm0h;QO\\Di0g;WOeD:_;EREGQ;8]4O001O0O2O001N10001O001O0O2O001O001O0O2O000O2O0O101N100O2O0LSeZ3"}, "label": "a goose in center"}]}
{"id": 494382, "image": "COCO_train2014_000000494382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goos between two other geese\"."}], "task": "refering", "answer_template": "The \"a goos between two other geese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 143, 144, 145, 146, 160, 161, 177, 178, 192, 193, 194, 195, 196, 206, 208, 209, 210, 211, 212, 213, 222, 223, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 257, 260, 261, 262, 263, 278, 279, 295, 296], "bbox": [0.03491666666666667, 0.31484375, 0.64325, 0.7611875], "iscrowd": 0, "area": 25164.125350000002, "rle": {"size": [640, 480], "counts": "eP;1nc02N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2O0O1O2N100O2N1O1O2O0O1O2N100O2O0O2EgNm]OY1Rb0hNm]OZ1Qb0gNo]OY1Qb0fNP^OZ1oa0gNP^O[1na0fNR^OZ1ma0fNS^O\\1la0dNT^O\\1ka0eNU^O\\1ja0cNV^O^1ja0bNV^O_1ha0bNW^O_1ia0aNW^O_1ia0`NX^Oa1ga0_NX^Ob1ha0]NY^Od1ea0]NZ^Od1fa0\\NZ^Oe1ea0ZN\\^Oe1ea0[NZ^Oe1ga0[NY^Oe1ga0ZNY^Of1ha0ZNX^Of1ha0ZNX^Oe1ha0]NV^Oc1ka091N4L5L3L4M3L5K4M4K5K4M4K5K4M2M2N3N2MUf80lYG0O4M5K4L5K7I<D3M2N3M3M3L3N3M3M0000000001O00001OBVNf^Oj1Za0XNd^Oh1\\a0YNd^Of1\\a0[Nc^Oe1]a0\\Nc^Oc1]a0^Nb^Ob1]a0`Nb^O`1^a0aNb^O^1^a0cNa^O]1_a0dN`^O^1^a0cNa^O^1]a0e0N1O1O1O1O2M2O1O1O1O1O1N2O1O1O1O2N2M3N2N2N2N2N2N2M3N2N0000000O10000000O10O1000000000000O3N2N4L4L4L4k@cK_>a4YAcKg>k400000000000000001O0G9@`0F:J^NiLgBV3g<aMVC^2g<iMUCV2i<QNSCn1k<YNQCf1m<bNnB^1n<jNnBU1P=ROmBm0P=ZOlBe0R=BjB>S=X3N3L3N3M2N3O10O10O0100O1000O0100O100O010O1000O0100O100O100_JTDm2m;RMXDi2i;VM]Dc2e;]M_D^2b;bM_D[2c;eM^DY2c;fM_DW2c;iM^DU2c;kM_DQ2c;oM^Dn1d;RN]Dl1d;SN^Dj1d;VN]Dh1d;XN^Dd1e;[N\\Dc1f;\\N\\D`1f;_N\\D_1f;`N\\D\\1g;bN\\D[1e;eN\\DX1g;fN\\DV1g;hN[DV1f;iN]DS1f;lN[DR1f;mN]Do0f;PO[Dm0h;QO\\Di0g;WOeD:_;EREGQ;8]4O001O0O2O001N10001O001O0O2O001O001O0O2O000O2O0O101N100O2O0LSeZ3"}, "label": "a goos between two other geese"}]}
{"id": 494382, "image": "COCO_train2014_000000494382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown duck standing in water next facing the camera next to two larger ducks\"."}], "task": "refering", "answer_template": "The \"a brown duck standing in water next facing the camera next to two larger ducks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 206, 207, 208, 223, 224, 225, 241, 242, 257, 258, 259, 260, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 312, 325, 326, 327, 328, 342], "bbox": [0.09420833333333332, 0.48465625, 0.3797291666666667, 0.876515625], "iscrowd": 0, "area": 17783.383249999995, "rle": {"size": [640, 480], "counts": "ibl09dc0j0VO6I5L4K4M4L4K5L3N3N2N2N101N2N2N1O2N1O1O1O2N7J6I6WMcLdDd3T;aLhDf3P;`LkDf3o:^LmDf3n:_LnDc3o:bLmD_3R;eLjD\\3T;iLiDX2[NcMj<;fDS2dN^Md<d0dDP2kNYM_<k0bDm1SOTMY<T1`Di1[OoLS<]1^Df1BiLo;e1[Dc1JeLh;m1ZD_12`Lb;V2XD\\1d<hNYCY1f<jNXCV1f<nNWCT1g<nNWCS1i<oNTCS1j<POSCQ1m<POQCQ1o<Q3O00001O00001O000000000O1000SJSCg4m<YKVCd4j<\\KZC`4g<_K\\C^4g<_K]C]4e<aKcCW4`<fKhCR4Z<kKnCn3U<oKnCn3S<QLoCm3R<RLQDk3P<TLSDi3n;VLTDh3m;WLVDf3k;YLXDd3i;[LYDc3h;\\LXDd3i;[LWDe3j;YLXDf3i;YLWDg3j;XLVDh3k;WLUDi3l;VLTDj3m;ULSDk3n;TLSDk3n;TLRDm3n;RLRDo3n;PLRDR4m=1O1O2O0O1O20O1O010^OY@hLh?V3Z@hLh?U3[@iLf?U3]@iLe?U3^@iLc?T3`@jLb?S3a@kL`?T3b@jL`?S3i0N1N3M2O2M2O1N3M2O1N3N1N2N3N1N2N2N3K4L4K7J6J6J6J6J7I6J6JQii5"}, "label": "a brown duck standing in water next facing the camera next to two larger ducks"}]}
{"id": 494382, "image": "COCO_train2014_000000494382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the duck that is closest to the photographer\"."}], "task": "refering", "answer_template": "The \"the duck that is closest to the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 206, 207, 208, 223, 224, 225, 241, 242, 257, 258, 259, 260, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 312, 325, 326, 327, 328, 342], "bbox": [0.09420833333333332, 0.48465625, 0.3797291666666667, 0.876515625], "iscrowd": 0, "area": 17783.383249999995, "rle": {"size": [640, 480], "counts": "ibl09dc0j0VO6I5L4K4M4L4K5L3N3N2N2N101N2N2N1O2N1O1O1O2N7J6I6WMcLdDd3T;aLhDf3P;`LkDf3o:^LmDf3n:_LnDc3o:bLmD_3R;eLjD\\3T;iLiDX2[NcMj<;fDS2dN^Md<d0dDP2kNYM_<k0bDm1SOTMY<T1`Di1[OoLS<]1^Df1BiLo;e1[Dc1JeLh;m1ZD_12`Lb;V2XD\\1d<hNYCY1f<jNXCV1f<nNWCT1g<nNWCS1i<oNTCS1j<POSCQ1m<POQCQ1o<Q3O00001O00001O000000000O1000SJSCg4m<YKVCd4j<\\KZC`4g<_K\\C^4g<_K]C]4e<aKcCW4`<fKhCR4Z<kKnCn3U<oKnCn3S<QLoCm3R<RLQDk3P<TLSDi3n;VLTDh3m;WLVDf3k;YLXDd3i;[LYDc3h;\\LXDd3i;[LWDe3j;YLXDf3i;YLWDg3j;XLVDh3k;WLUDi3l;VLTDj3m;ULSDk3n;TLSDk3n;TLRDm3n;RLRDo3n;PLRDR4m=1O1O2O0O1O20O1O010^OY@hLh?V3Z@hLh?U3[@iLf?U3]@iLe?U3^@iLc?T3`@jLb?S3a@kL`?T3b@jL`?S3i0N1N3M2O2M2O1N3M2O1N3N1N2N3N1N2N2N3K4L4K7J6J6J6J6J7I6J6JQii5"}, "label": "the duck that is closest to the photographer"}]}
{"id": 494382, "image": "COCO_train2014_000000494382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the duck behind the 2 others\"."}], "task": "refering", "answer_template": "The \"the duck behind the 2 others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 131, 132, 133, 147, 148, 149, 165, 166, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 247, 248, 249, 264, 265, 266], "bbox": [0.5562708333333333, 0.28171875, 0.8774166666666666, 0.673125], "iscrowd": 0, "area": 18521.460800000004, "rle": {"size": [640, 480], "counts": "oXW55ec0:J2O1O1N2O0dL]300O10000O100O10XOh00O1000O010000O1000O010000O10000O1000000000O1006Z@[Le>o3o@TLP?^400000000000jNi@cMX?\\2PA\\MP?d2RAZMn>f2TAYMk>g2WAWMi>i2YAUMg>j2\\ATMd>l2^AkLi>U3Q10000000000O10000000000000mMQMUCo2h<]MoBb2o<dMnB\\2o<jMnBV2o<oMoBQ2n<UNoBk1n<[NoBe1P=^NnBb1P=cNmB]1R=gNkBY1T=jNjBU1V=oNgBQ1X=SOeBm0Y=WOeBi0[=YOcBf0]=]OaBc0^=_OaB`0`=B^B>a=D^B;b=H\\B8d=JZB5f=MYB3f=0XB0h=S301O010O0010O011O0O100O1O1O2WKTB\\3l=cLXBY3i=fLZBW3g=hL]BT3d=kL`BQ3b=nLaBo2_=PMeBl2\\=SMgBj2Z=UMiBh2Y=VMjBg2X=XMjBe2W=[MjBd2W=ZMlBc2V=\\MlBa2V=^MlB_2V=`MlB]2U=cMmBZ2U=dMnBZ2S=eMoBX2S=gMoBV2R=jMPCS2R=lMoBR2R=mMQCo1R=PNoBm1S=SNoBj1S=UNoBg1S=XNoBf1R=ZNPCb1S=]NnB`1T=`NnB]1T=aNnB\\1T=dNnBY1S=fNPCV1S=iNnBT1T=lNoBP1S=nNWCg0k<YO^C=c<BgC3\\<LmCIU<6l3O000O2O0O2O0O10eeT1"}, "label": "the duck behind the 2 others"}]}
{"id": 494382, "image": "COCO_train2014_000000494382.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far most duck\"."}], "task": "refering", "answer_template": "The \"the far most duck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 131, 132, 133, 147, 148, 149, 165, 166, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 247, 248, 249, 264, 265, 266], "bbox": [0.5562708333333333, 0.28171875, 0.8774166666666666, 0.673125], "iscrowd": 0, "area": 18521.460800000004, "rle": {"size": [640, 480], "counts": "oXW55ec0:J2O1O1N2O0dL]300O10000O100O10XOh00O1000O010000O1000O010000O10000O1000000000O1006Z@[Le>o3o@TLP?^400000000000jNi@cMX?\\2PA\\MP?d2RAZMn>f2TAYMk>g2WAWMi>i2YAUMg>j2\\ATMd>l2^AkLi>U3Q10000000000O10000000000000mMQMUCo2h<]MoBb2o<dMnB\\2o<jMnBV2o<oMoBQ2n<UNoBk1n<[NoBe1P=^NnBb1P=cNmB]1R=gNkBY1T=jNjBU1V=oNgBQ1X=SOeBm0Y=WOeBi0[=YOcBf0]=]OaBc0^=_OaB`0`=B^B>a=D^B;b=H\\B8d=JZB5f=MYB3f=0XB0h=S301O010O0010O011O0O100O1O1O2WKTB\\3l=cLXBY3i=fLZBW3g=hL]BT3d=kL`BQ3b=nLaBo2_=PMeBl2\\=SMgBj2Z=UMiBh2Y=VMjBg2X=XMjBe2W=[MjBd2W=ZMlBc2V=\\MlBa2V=^MlB_2V=`MlB]2U=cMmBZ2U=dMnBZ2S=eMoBX2S=gMoBV2R=jMPCS2R=lMoBR2R=mMQCo1R=PNoBm1S=SNoBj1S=UNoBg1S=XNoBf1R=ZNPCb1S=]NnB`1T=`NnB]1T=aNnB\\1T=dNnBY1S=fNPCV1S=iNnBT1T=lNoBP1S=nNWCg0k<YO^C=c<BgC3\\<LmCIU<6l3O000O2O0O2O0O10eeT1"}, "label": "the far most duck"}]}
{"id": 355119, "image": "COCO_train2014_000000355119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that isn ' t sticking it ' s tongue out\"."}], "task": "refering", "answer_template": "The \"the horse that isn ' t sticking it ' s tongue out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.37915625, 0.0, 1.0, 0.9894919786096257], "iscrowd": 0, "area": 102354.53600000001, "rle": {"size": [374, 640], "counts": "bkh2X1l3oNZ2]1[MnNX2`1bKdNH6Z4g1\\K`N1N]4g3[K^L`4`5K4K5L3L4M4K4E;F;\\Oc0G9M4M2N3N2O1O100000001OO1O100O1O1O3N1N2N3M5K2O2M2M2O00O10000O1O1O1O1O1N2N2N2L5K3M4M3L4O0O2N2N20O010N101O2N1O2Na0^O4K6L4M3L3N3L2VOi0N2M3N2N2N2O1N2N1O2N2O0O2N2N2N101N2O1N101N2O1N101N2O0O2O1O1O010O1O1O1O00100N2O1N2N2O0O2O1N2N2N2N101N2O1O1O1O1O001O1O1O1O1N1O2O1N2N2O1N2N2O1N1O200000000000000000O10001O0000001O0O101O1O001O1O0O2O1O001O1O1N101O1O001O1O1O1O1O1O001O100O1O1O1O001O1O1O1O1O2N100O1O1O2N1O5K5K4L3M4L3M3M2N2N2N1O2N2M2O2N2N2N1O2N2N1O001O1O001O0O2O1O001O1O001O001N2O1O1O1O1O1O2N1O2M3N2N1O2N1O2N1O2N2N1O1O1O1O1O1O1O2N1O2N1O1O2N1O2M2O1O1O1O2N1O1O5K7I6J2N2N2N3N0O2N1O2N1O2O0O2N1O1O2N101TOWEg0l:0001O0001O000000000000001O0000000000000000000000000000000000000000000000000000000000O2O000000000000000O10000000000000000O100000000000O10000000000O100000000N2M3M3N2M3N2M3O1N1O2O1N2O0OC"}, "label": "the horse that isn ' t sticking it ' s tongue out"}]}
{"id": 355119, "image": "COCO_train2014_000000355119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white horse standing to the right of another black and right horse\"."}], "task": "refering", "answer_template": "The \"a black and white horse standing to the right of another black and right horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.37915625, 0.0, 1.0, 0.9894919786096257], "iscrowd": 0, "area": 102354.53600000001, "rle": {"size": [374, 640], "counts": "bkh2X1l3oNZ2]1[MnNX2`1bKdNH6Z4g1\\K`N1N]4g3[K^L`4`5K4K5L3L4M4K4E;F;\\Oc0G9M4M2N3N2O1O100000001OO1O100O1O1O3N1N2N3M5K2O2M2M2O00O10000O1O1O1O1O1N2N2N2L5K3M4M3L4O0O2N2N20O010N101O2N1O2Na0^O4K6L4M3L3N3L2VOi0N2M3N2N2N2O1N2N1O2N2O0O2N2N2N101N2O1N101N2O1N101N2O0O2O1O1O010O1O1O1O00100N2O1N2N2O0O2O1N2N2N2N101N2O1O1O1O1O001O1O1O1O1N1O2O1N2N2O1N2N2O1N1O200000000000000000O10001O0000001O0O101O1O001O1O0O2O1O001O1O1N101O1O001O1O1O1O1O1O001O100O1O1O1O001O1O1O1O1O2N100O1O1O2N1O5K5K4L3M4L3M3M2N2N2N1O2N2M2O2N2N2N1O2N2N1O001O1O001O0O2O1O001O1O001O001N2O1O1O1O1O1O2N1O2M3N2N1O2N1O2N1O2N2N1O1O1O1O1O1O1O2N1O2N1O1O2N1O2M2O1O1O1O2N1O1O5K7I6J2N2N2N3N0O2N1O2N1O2O0O2N1O1O2N101TOWEg0l:0001O0001O000000000000001O0000000000000000000000000000000000000000000000000000000000O2O000000000000000O10000000000000000O100000000000O10000000000O100000000N2M3M3N2M3N2M3O1N1O2O1N2O0OC"}, "label": "a black and white horse standing to the right of another black and right horse"}]}
{"id": 355119, "image": "COCO_train2014_000000355119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two black and white horses standing side by side\"."}], "task": "refering", "answer_template": "The \"two black and white horses standing side by side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 4, 5, 6, 7, 8, 9, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 151, 152, 153, 162, 163, 164, 165, 167, 168, 169, 174, 175, 176, 177, 185, 186, 187, 191, 192, 196, 197, 198, 199, 200, 208, 209, 210, 214, 215, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287], "bbox": [0.0351875, 0.002299465240641711, 0.808078125, 0.9850802139037433], "iscrowd": 0, "area": 74629.64490000003, "rle": {"size": [374, 640], "counts": "l\\8b0U;>B1O1YHQOd4o0YKUOf4l0VKXOi4i0SKZOm4g0oJ]OQ5c0jJBU5>hJEX5<dJH[59aJK^56YIS1f6nNVIU1j6kNTIX1k6iNQI[1n6fNoH]1Q3gM6l0fL`1m2jM>f0aLe1j2kMe0?_Lj1d2mMn0_3kNfLV1[3bNkL_1U3[NQMd1Q3TNUMm1m2jMYMW2P3ZMVMf2R3kLTMV3U3[LPMf3a5O01O01N1O1N3N001O0000000O10000000000O2O00000000000O1000O001O1N2O001O1O1O0O2O1N2O0L5J6K4K6J6J5L5N1N3M2N3N1N3N101O0O10L4M2M4L3O22N2M2O2N2M2O2M2O3L4M3M3L4M2M2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1001O1O2N1O1O1N2O2N1O1O1O2N1O1O1O1O2N1O1N2O2N001O1O1O001O1N2O001O1O1O001O1O1O0O2O1O1TOlITLU6j3lIVLV6g3kIYLV6e3kI[LV6c3kI]LW6`3jI`LW6^3jIaLX6]3iIcLX6[3iIeLY6h0QJ<GlNY6b0eJ1SO]OY6<SKLeNHZ66XKN_NLZ60^K1YNOZ6KdK1TN4Z6DiK5nM7Z6_OoK6hM:[6ZOTL9bM=`9^ObFb0`9[OaFe0X:000000000000O1000000000000000000000000000000000000000000000000000000000fHk0WO`NNe05S2lNkMk02L]3nMSMR2@L\\50dJLe5O[JMl50TJKR64nIIU67kIFY69gIC^6<bIAb6>^I_Oe6a0[I\\Oi6c0WIZOm6e0_IkNf6T1n101O1O1O001O1O1O001O1O1O1O1O001O1O1O0000000000O10000000000O100000000O10000000000O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001N10000CUNPGl1o8<001O001O0010O01O001O001O001O001O001O001O02O2M2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2M3J6J6N2N2N2N2O1N2O10000001O00000000001O00000000001O000O1000001O00001O001O1O001O001O001O001O001N101O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O3M4L4L4K6K4L2N2N2M3N2N2N2N3M2M3N1O1O1O1N2O001O1O1N2O001O1O1N2O001O1O1O1N2O1O1O1O1N3N1O1O1N2O2N1O1O2M2O2N1O2N2M2O2N1O2M3N2N1O2N2MPd\\1"}, "label": "two black and white horses standing side by side"}]}
{"id": 355119, "image": "COCO_train2014_000000355119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the faces of the two black and white horses\"."}], "task": "refering", "answer_template": "The \"the faces of the two black and white horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 4, 5, 6, 7, 8, 9, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 151, 152, 153, 162, 163, 164, 165, 167, 168, 169, 174, 175, 176, 177, 185, 186, 187, 191, 192, 196, 197, 198, 199, 200, 208, 209, 210, 214, 215, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287], "bbox": [0.0351875, 0.002299465240641711, 0.808078125, 0.9850802139037433], "iscrowd": 0, "area": 74629.64490000003, "rle": {"size": [374, 640], "counts": "l\\8b0U;>B1O1YHQOd4o0YKUOf4l0VKXOi4i0SKZOm4g0oJ]OQ5c0jJBU5>hJEX5<dJH[59aJK^56YIS1f6nNVIU1j6kNTIX1k6iNQI[1n6fNoH]1Q3gM6l0fL`1m2jM>f0aLe1j2kMe0?_Lj1d2mMn0_3kNfLV1[3bNkL_1U3[NQMd1Q3TNUMm1m2jMYMW2P3ZMVMf2R3kLTMV3U3[LPMf3a5O01O01N1O1N3N001O0000000O10000000000O2O00000000000O1000O001O1N2O001O1O1O0O2O1N2O0L5J6K4K6J6J5L5N1N3M2N3N1N3N101O0O10L4M2M4L3O22N2M2O2N2M2O2M2O3L4M3M3L4M2M2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1001O1O2N1O1O1N2O2N1O1O1O2N1O1O1O1O2N1O1N2O2N001O1O1O001O1N2O001O1O1O001O1O1O0O2O1O1TOlITLU6j3lIVLV6g3kIYLV6e3kI[LV6c3kI]LW6`3jI`LW6^3jIaLX6]3iIcLX6[3iIeLY6h0QJ<GlNY6b0eJ1SO]OY6<SKLeNHZ66XKN_NLZ60^K1YNOZ6KdK1TN4Z6DiK5nM7Z6_OoK6hM:[6ZOTL9bM=`9^ObFb0`9[OaFe0X:000000000000O1000000000000000000000000000000000000000000000000000000000fHk0WO`NNe05S2lNkMk02L]3nMSMR2@L\\50dJLe5O[JMl50TJKR64nIIU67kIFY69gIC^6<bIAb6>^I_Oe6a0[I\\Oi6c0WIZOm6e0_IkNf6T1n101O1O1O001O1O1O001O1O1O1O1O001O1O1O0000000000O10000000000O100000000O10000000000O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001N10000CUNPGl1o8<001O001O0010O01O001O001O001O001O001O001O02O2M2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2M3J6J6N2N2N2N2O1N2O10000001O00000000001O00000000001O000O1000001O00001O001O1O001O001O001O001O001N101O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O3M4L4L4K6K4L2N2N2M3N2N2N2N3M2M3N1O1O1O1N2O001O1O1N2O001O1O1N2O001O1O1O1N2O1O1O1O1N3N1O1O1N2O2N1O1O2M2O2N1O2N2M2O2N1O2M3N2N1O2N2MPd\\1"}, "label": "the faces of the two black and white horses"}]}
{"id": 174896, "image": "COCO_train2014_000000174896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair that is behind the laptop\"."}], "task": "refering", "answer_template": "The \"a chair that is behind the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 35, 36, 42, 43, 44, 57, 58, 59, 66, 79, 80, 81, 89, 101, 102, 103, 104, 111, 112, 113, 114, 124, 125, 126, 127, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.40265625, 0.0040416666666666665, 1.0, 0.41331249999999997], "iscrowd": 0, "area": 20707.201600000015, "rle": {"size": [480, 640], "counts": "cRi34k>3M1O1O2N1N2O1O1O1O1O1N3N1O1O1O1N2O1O1O1O2N1N2O1O1O1O1N2O1O2N1O1O1N2O1O1O1O2M2O1O1O1O1O1N2O1O2N1O1O1N2O1O11O00010O001O001O000010O01ON2M3M3M3M3N2M4L3M3N2M3M3L4I7H8I8H7I7H8J600000001O000000O1L4K5L4K5K5K5B>B>B>B>C=B>B>B>D<000000000000001O00000000000000001O0O100000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O0O100000000000001O000000000000001O00000bFWOe4i0UK]Ok4c0oJDP5<jJJV56dJ1[5O_J8`5IYJ=g5CRJe0m5[OmIk0S6UOgIR1X6nNbIX1^6hN\\I_1c6aNWIe1i6[NQIl1n6TNlHR2T7oMeHX2Z7hM_H_2a7aMYHe2g7[MSHk2m7UMmGR3R8nLhGX3X8hLbG^3^8bLSGm3m8S10001O000000001O0000000000fLgGNX8LnG4R8FUH9k7A[H?e7[OaHe0_7UOgHk0Y7nNnHS1Q7gNVIX1j6bN\\I^1d6\\NbId1^6VNhIj1X6PNnIP2R6jMUJU2k5eM[J[2e5_MaJ`2P901O000O10001O0O1000001N10002N1N3N1O2N1N3N1O2N1N2O2M2O2M2O2M2O2M2O2M3N3M2M3N3L3N2M4M2M3N2M4MaJ"}, "label": "a chair that is behind the laptop"}]}
{"id": 174896, "image": "COCO_train2014_000000174896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty chair on the inside near the window\"."}], "task": "refering", "answer_template": "The \"the empty chair on the inside near the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 24, 25, 26, 27, 28, 29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 122, 123, 124], "bbox": [0.0016875000000000002, 0.0, 0.46517187499999996, 0.32135416666666666], "iscrowd": 0, "area": 32731.188599999998, "rle": {"size": [480, 640], "counts": "]a0f0Z>0O2O000O100O2O000O100O101N10000O101N1O100O100O2O0O100O1O2O0O100O100O2O0O100O1O2O0O100O1dN\\1O2N100O100O2N100O100O2N100O1000001O000000001O000hDcLm:]3SEcLm:]3REeLm:[3REfLn:[3QEeLo:[3PEfLP;Z3PEfLP;Z3oDgLQ;Y3oDhLP;Y3nDhLR;X3mDiLS;W3mDiLS;W3lDjLT;V3lDjLT;b3O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000O100000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001O0000001O0O1000001O0000001O0000001O0000001O0000O1O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2Oc2\\M2O2N1N2O1N2O1N2O1N2O0O2O1O1N2O1N2O_fP5"}, "label": "the empty chair on the inside near the window"}]}
{"id": 174896, "image": "COCO_train2014_000000174896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair which is not behind the laptop\"."}], "task": "refering", "answer_template": "The \"the chair which is not behind the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 24, 25, 26, 27, 28, 29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 122, 123, 124], "bbox": [0.0016875000000000002, 0.0, 0.46517187499999996, 0.32135416666666666], "iscrowd": 0, "area": 32731.188599999998, "rle": {"size": [480, 640], "counts": "]a0f0Z>0O2O000O100O2O000O100O101N10000O101N1O100O100O2O0O100O1O2O0O100O100O2O0O100O1O2O0O100O1dN\\1O2N100O100O2N100O100O2N100O1000001O000000001O000hDcLm:]3SEcLm:]3REeLm:[3REfLn:[3QEeLo:[3PEfLP;Z3PEfLP;Z3oDgLQ;Y3oDhLP;Y3nDhLR;X3mDiLS;W3mDiLS;W3lDjLT;V3lDjLT;b3O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000O100000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001O0000001O0O1000001O0000001O0000001O0000001O0000O1O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2Oc2\\M2O2N1N2O1N2O1N2O1N2O0O2O1O1N2O1N2O_fP5"}, "label": "the chair which is not behind the laptop"}]}
{"id": 174896, "image": "COCO_train2014_000000174896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large bowl contains bread and potatoes\"."}], "task": "refering", "answer_template": "The \"a large bowl contains bread and potatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.45168749999999996, 0.8202291666666667, 0.7870781250000001, 0.9887708333333334], "iscrowd": 0, "area": 13146.703400000002, "rle": {"size": [480, 640], "counts": "\\mW43j><C5L3M3L5L3M3M3L4M3M3L4M3M3M3L4M300O100O100O100O100O100O100O100O1O100O100000000001O00000000001O000000001O00000000001O00000000001O00000000001O000000001O00000000001O0001O000001O00000000001O000000001O000000000O2O00000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O00001O001O001O001O001O001O1O001O001O001O001O0O2H7J7I6I8I_Zo1"}, "label": "a large bowl contains bread and potatoes"}]}
{"id": 174896, "image": "COCO_train2014_000000174896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate of meat and potatos\"."}], "task": "refering", "answer_template": "The \"a plate of meat and potatos\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.45168749999999996, 0.8202291666666667, 0.7870781250000001, 0.9887708333333334], "iscrowd": 0, "area": 13146.703400000002, "rle": {"size": [480, 640], "counts": "\\mW43j><C5L3M3L5L3M3M3L4M3M3L4M3M3M3L4M300O100O100O100O100O100O100O100O1O100O100000000001O00000000001O000000001O00000000001O00000000001O00000000001O000000001O00000000001O0001O000001O00000000001O000000001O000000000O2O00000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O00001O001O001O001O001O001O1O001O001O001O001O0O2H7J7I6I8I_Zo1"}, "label": "a plate of meat and potatos"}]}
{"id": 404270, "image": "COCO_train2014_000000404270.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed blue bear\"."}], "task": "refering", "answer_template": "The \"a stuffed blue bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 28, 29, 30, 31, 32, 33, 41, 42, 43, 44, 45, 46, 55, 56, 57, 58, 59, 68, 69, 70, 71, 72, 113, 114, 119, 120, 126, 132, 133, 138, 139, 145, 146, 148, 149, 150, 151, 152, 162, 163, 164, 165, 175, 176, 177, 178, 188, 189, 190, 191, 201, 202, 203, 214, 215, 216, 217, 228, 229], "bbox": [0.17117333333333332, 0.0901, 0.7777866666666665, 0.9662200000000001], "iscrowd": 0, "area": 27316.818900000006, "rle": {"size": [500, 375], "counts": "o`o0g0k>2N2N3M2N2O1N3M2N2N3N1N2\\JeNoL\\1n2jNnLV1o2QOmLo0Q3VOlLj0R3\\OjLe0S3AjL>T3HhL8U3KkL5S3NlL2R31mLNR35mLKQ38oLGo2<PMCQ3>nLBR3?mLAR3`0nL_OS3b0lL^OS3d0mLZOT3g0kLXOV3i0iLVOW3l0hLSOY3n0fLQOZ3Q1fLmN[3T1dLkN\\3W1cLiN]3Y1aLfN`3[1_LdNa3_1]L`Nd3a1\\L]Nd3f1ZLYNg3h1XLWNi3k1ULSNm3n1RLPNP4R2nKlMT4V2kKgMW4Z2hKdMZ4^2dK`M^4a2aK]Ma4^2dK`M^4a2bK\\M`4e2`KYMa4h2`KTMb4m2R410O01O1O1O1O1O000000000001O000000000O2O00000000001O000000000000O10000O01000O100O10000kHmLh1T3YNlLf1T3ZNmLe1S3[NmLd1T3\\NnLb1R3^NPM`1P3`NRM]1o2bNTM\\1l2dNVMZ1j2fNXMW1i2iNYMU1f2lN\\MR1d2mN_MP1b2PO`Mn0`2RObMl0^2TOdMi0]2WOeMg0[2XOhMf0X2ZOjMc0W2]OkMa0U2_OmM`0R2@PN>P2BRN=l1C\\OVOd0j00cNO]1;YNEg1e0PNZOP2n0hMROX2n0iMQOW2P1hMPOW2R1iMlNX2T1hMlNX2U1hMjNX2W1gMiNY2W1hMhNX2Y1hMfNX2[1gMeNY2[1hMdNW2^1hMaNY2`1gM_NY2b1fM^NZ2b1iM[NW2f1kMWNU2j1mMSNS2m1PNPNo1R2TNiMm1X2UN]MS2c2PNRMX2o2gNiKa1X4m41O0O2O1O001O1O001O1O001O1O1O001O1O0O2O1OdMcEA\\:<iEDU::oEFo98UFHj99UFGj9:WFEh9=XFAh9`0YF_Of9c0ZF[Og9e0ZFZOe9h0[FWOd9j0]FTOc9n0]FQOb9P1_FnNb9S1^FlNa9U1`FjN_9X1bFeN^9\\1cFcN\\9_1dF_N\\9b1fF\\NZ9e1fFYNZ9h1hFVNW9l1iFSNV9n1kFPNU9R2lFlMT9T2mFjMS9X2nFfMQ9[2PGcMP9_2SG]Mn8b2\\GTMY7]OhHZ39nLn6JnHS3>iLb66UIk2k8WM\\Gb2b8aMeGM]N^1m9gNcHW1[7kNhHQ1W7ROjHl0T7VOnHg0Q7[OQIc0m6@SI>m6CTI<k6EWI8i6IXI6g6LYI2h6NYI0g61[IMd65\\IIg65ZIHi67YIFi6:ZI@i6?\\I[Of6c0_IWOe6f0V4M3N2N2M3N<DUYX1"}, "label": "a stuffed blue bear"}]}
{"id": 404270, "image": "COCO_train2014_000000404270.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sweater on the bear\"."}], "task": "refering", "answer_template": "The \"the sweater on the bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 72, 73, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 134, 135, 136, 137, 138, 147, 148, 149, 150], "bbox": [0.18354666666666666, 0.2935, 0.72728, 0.6415599999999999], "iscrowd": 0, "area": 26662.447899999996, "rle": {"size": [500, 375], "counts": "ZnQ12_?3M3M3M3M4M2M3M3M3M4L3N2M3M3M4L3M3N2M3M4L3M3M3N2M4L3M3M3N2M4L3M3O1O1O1O2N1N2O1O1O2N1O1O1O1O2M2O1O1O1O2N1O1O10001O0^EUL[9k3[F_Le9a3QFjLn9V3gEUMY:S40001O000000001O000000001O000000001O000000001O000000001O0000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000001O0O2O1N2O1N2O1N1O2O1N2O1N2O0O2N2O1N2O1N2O0O2N2O1N2O1N101N2O1N2N2O1N4M2M4M2L5H7I8H7J7H7I8H7J7H7I8I6J7H7J7Ilba1"}, "label": "the sweater on the bear"}]}
{"id": 404270, "image": "COCO_train2014_000000404270.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the grey teddy bear sitting on the windowsill\"."}], "task": "refering", "answer_template": "The \"the grey teddy bear sitting on the windowsill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 72, 73, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 134, 135, 136, 137, 138, 147, 148, 149, 150], "bbox": [0.18354666666666666, 0.2935, 0.72728, 0.6415599999999999], "iscrowd": 0, "area": 26662.447899999996, "rle": {"size": [500, 375], "counts": "ZnQ12_?3M3M3M3M4M2M3M3M3M4L3N2M3M3M4L3M3N2M3M4L3M3M3N2M4L3M3M3N2M4L3M3O1O1O1O2N1N2O1O1O2N1O1O1O1O2M2O1O1O1O2N1O1O10001O0^EUL[9k3[F_Le9a3QFjLn9V3gEUMY:S40001O000000001O000000001O000000001O000000001O000000001O0000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000001O0O2O1N2O1N2O1N1O2O1N2O1N2O0O2N2O1N2O1N2O0O2N2O1N2O1N101N2O1N2N2O1N4M2M4M2L5H7I8H7J7H7I8H7J7H7I8I6J7H7J7Ilba1"}, "label": "the grey teddy bear sitting on the windowsill"}]}
{"id": 518966, "image": "COCO_train2014_000000518966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in white hoodie on the right\"."}], "task": "refering", "answer_template": "The \"person in white hoodie on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 59, 74, 88, 89, 102, 103, 104, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 149, 164, 179, 194, 209, 224, 239], "bbox": [0.6691100702576112, 0.0980625, 1.0, 0.703375], "iscrowd": 0, "area": 13446.93795, "rle": {"size": [640, 427], "counts": "gnb52nc04L4L3M4M3L4L3M1O0010O0001O00001O01O01O00001O00001O00001O00001OO10000O10000O0100000O0100O010O10O01O100O010O100O00100O100O100O1O100O010N2M3M3M3M3M3M3M3M201N2O1O1N101O0O101N10001N100O2O01O01O000010O0001O00010O00010O4M6I7J6I7J5J>Cb0]Oe0\\Oe0[O6`NnIeDW6U;lIjDW6Q;mIiD[6Q;iIgDa6S;bIeDh6U;]IbDn6X;R1J6J5QFSGc8T9UGTGd8S9UGSGf8S9RGUGi8P9PGWGj8R:M2N3M2O2M2M4L5K6K5lM"}, "label": "person in white hoodie on the right"}]}
{"id": 518966, "image": "COCO_train2014_000000518966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mans back in a whitye hoodie\"."}], "task": "refering", "answer_template": "The \"a mans back in a whitye hoodie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 59, 74, 88, 89, 102, 103, 104, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 149, 164, 179, 194, 209, 224, 239], "bbox": [0.6691100702576112, 0.0980625, 1.0, 0.703375], "iscrowd": 0, "area": 13446.93795, "rle": {"size": [640, 427], "counts": "gnb52nc04L4L3M4M3L4L3M1O0010O0001O00001O01O01O00001O00001O00001O00001OO10000O10000O0100000O0100O010O10O01O100O010O100O00100O100O100O1O100O010N2M3M3M3M3M3M3M3M201N2O1O1N101O0O101N10001N100O2O01O01O000010O0001O00010O00010O4M6I7J6I7J5J>Cb0]Oe0\\Oe0[O6`NnIeDW6U;lIjDW6Q;mIiD[6Q;iIgDa6S;bIeDh6U;]IbDn6X;R1J6J5QFSGc8T9UGTGd8S9UGSGf8S9RGUGi8P9PGWGj8R:M2N3M2O2M2M4L5K6K5lM"}, "label": "a mans back in a whitye hoodie"}]}
{"id": 518966, "image": "COCO_train2014_000000518966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman with a red and white backpack helping to transport a surfboard\"."}], "task": "refering", "answer_template": "The \"a blonde woman with a red and white backpack helping to transport a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 64, 65, 66, 79, 80, 81, 93, 94, 95, 96, 97, 110, 111, 112, 125, 126, 127, 128, 140, 141, 142, 143, 155, 156, 157, 167, 169, 170, 171, 182, 183, 184, 185, 186, 197, 198, 199, 200, 212, 213, 214, 215, 228, 229, 230, 243, 244, 245, 258, 259, 260, 273, 274, 288, 289, 303, 304, 318, 319, 333, 334], "bbox": [0.17470725995316158, 0.14203125, 0.559672131147541, 0.9848437499999999], "iscrowd": 0, "area": 29838.738599999993, "rle": {"size": [640, 427], "counts": "WV_1>bc0`0A6I6J6K5J6K5J5K6K2M3M2O2M2N3N1N3M2O2M3VKmLaHU3\\7lLeHV3W7jLjHY3S7gLmH[3YOeLd31TM\\3SOYMT3\\OiM^3[LPMn0m0W4fNaN_3TLWMl0h0\\4cNdN`3oK^Mj0b0a4`NfNc3iKdMh0>e4\\NkNd3bKkMg06k4\\NlNe3^Kl0d5`KnNg3XKl0i5]KQOe8n0\\GROe8l0\\GTOd8k0]GVOc8h0_GWOa8h0`GXOa8f0`G[O_8d0bG\\O_8b0cG]O]8b0dG^O]8`0dG@\\8?fG_O[8`0fG@Z8?gG@Z8?hG@X8?iG@X8?iG@X8?iGAW8>kG@U80lFjKP1V4T8NoFjKn0W4b5iKaKS4SNkKl0X4\\5YLZKa3bNlKi0Y4X5iLRKo2QOmKf0[4S5WMlJ^2_OnKd0[4o4gMcJm1NoKa0]4j4VNPJh1h0cK?^4f4i2ZLVMc3j2`LUM]3l2fLTMX3l2kLRMS3P3nLPMo2R3SMlLl2U3VMiLh2Y3YMgLe2Z3]MdLa2^3aMaLV2g3lMWLn1o3SNQLk1o3XN[I[Mj0[4m5\\NSImM`0f3^6^NnH_N3S3P7`NgHmNLc2^7bN`HQOO]2b7dNYHUO3X2c7dNVHYO4W2c7bNSH]O7T2e7b1XHbNg7^1VHcNk7^1RHmMe8T2XGWM`9X6<J7I6J6J7H7J7I6EfDjG`;R8aDkGf;Q86L5K4L4M4K4L5K6L4L4L4L4jJdBb3a=ZLbBb3b=[L`Bb3d=YL`Bd3d=WL`Bf3d=VL_Ba3AiKT>a0_Bc3k=XLYBf3j=VLYBi3i=RL[Bl3h=oK\\BP4k>N2O1N2N2O1O3M2M3N3M2N3L3N3M3M2M4M3M3M2M5L8H6J3L2OO1000O1O2N2M3N2N1O1O1N2O1WO[]O6gb0Hd]OM^b02k]OCYb0:h0O1N2O1N3M2OYXe3"}, "label": "a blonde woman with a red and white backpack helping to transport a surfboard"}]}
{"id": 518966, "image": "COCO_train2014_000000518966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a woman in black with a ponytail\"."}], "task": "refering", "answer_template": "The \"the back of a woman in black with a ponytail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 64, 65, 66, 79, 80, 81, 93, 94, 95, 96, 97, 110, 111, 112, 125, 126, 127, 128, 140, 141, 142, 143, 155, 156, 157, 167, 169, 170, 171, 182, 183, 184, 185, 186, 197, 198, 199, 200, 212, 213, 214, 215, 228, 229, 230, 243, 244, 245, 258, 259, 260, 273, 274, 288, 289, 303, 304, 318, 319, 333, 334], "bbox": [0.17470725995316158, 0.14203125, 0.559672131147541, 0.9848437499999999], "iscrowd": 0, "area": 29838.738599999993, "rle": {"size": [640, 427], "counts": "WV_1>bc0`0A6I6J6K5J6K5J5K6K2M3M2O2M2N3N1N3M2O2M3VKmLaHU3\\7lLeHV3W7jLjHY3S7gLmH[3YOeLd31TM\\3SOYMT3\\OiM^3[LPMn0m0W4fNaN_3TLWMl0h0\\4cNdN`3oK^Mj0b0a4`NfNc3iKdMh0>e4\\NkNd3bKkMg06k4\\NlNe3^Kl0d5`KnNg3XKl0i5]KQOe8n0\\GROe8l0\\GTOd8k0]GVOc8h0_GWOa8h0`GXOa8f0`G[O_8d0bG\\O_8b0cG]O]8b0dG^O]8`0dG@\\8?fG_O[8`0fG@Z8?gG@Z8?hG@X8?iG@X8?iG@X8?iGAW8>kG@U80lFjKP1V4T8NoFjKn0W4b5iKaKS4SNkKl0X4\\5YLZKa3bNlKi0Y4X5iLRKo2QOmKf0[4S5WMlJ^2_OnKd0[4o4gMcJm1NoKa0]4j4VNPJh1h0cK?^4f4i2ZLVMc3j2`LUM]3l2fLTMX3l2kLRMS3P3nLPMo2R3SMlLl2U3VMiLh2Y3YMgLe2Z3]MdLa2^3aMaLV2g3lMWLn1o3SNQLk1o3XN[I[Mj0[4m5\\NSImM`0f3^6^NnH_N3S3P7`NgHmNLc2^7bN`HQOO]2b7dNYHUO3X2c7dNVHYO4W2c7bNSH]O7T2e7b1XHbNg7^1VHcNk7^1RHmMe8T2XGWM`9X6<J7I6J6J7H7J7I6EfDjG`;R8aDkGf;Q86L5K4L4M4K4L5K6L4L4L4L4jJdBb3a=ZLbBb3b=[L`Bb3d=YL`Bd3d=WL`Bf3d=VL_Ba3AiKT>a0_Bc3k=XLYBf3j=VLYBi3i=RL[Bl3h=oK\\BP4k>N2O1N2N2O1O3M2M3N3M2N3L3N3M3M2M4M3M3M2M5L8H6J3L2OO1000O1O2N2M3N2N1O1O1N2O1WO[]O6gb0Hd]OM^b02k]OCYb0:h0O1N2O1N3M2OYXe3"}, "label": "the back of a woman in black with a ponytail"}]}
{"id": 11065, "image": "COCO_train2014_000000011065.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman petting the horse on the head\"."}], "task": "refering", "answer_template": "The \"the woman petting the horse on the head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 64, 65, 66, 87, 88, 89, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 294, 295, 296, 317, 318, 319, 341, 342, 363, 364, 365], "bbox": [0.767625, 0.08296213808463251, 0.9752656249999999, 0.9865478841870824], "iscrowd": 0, "area": 29369.622499999998, "rle": {"size": [449, 640], "counts": "oag63e=9G9F9H9G9G9UKXNTMQ2i2WNiLR2T3VNdLm1Y3ZN`Li1]3_N[Ld1b3dNVL_1j3fNnK]1X2WMdL^1P1^1\\2TMcLa1l0]1a2SMcLa1h0_1e2PMbLd1d0_1i2nLbLe1a0`1m2kLaLh1=`1R3hLaLj18a1V3fLaLk16a1Y3dL`Ln12b1\\3aLbLn1MR2T3PLnLQ2Ga2m2_K\\MQ2AR3f2mJhMT2[Oa3_2\\JVNS2YOg3]2VJZNS2WOn3Y2oI`NU2TOS4W2hIeNU2ROY4T2cIjNT2PO`4Q2\\InNU2oNf4m1VITOU2bNU5V2fHWOo9c0RF]OS:>mEBW:6mEIY:NkE2Z:FiE:[:^OiEa0]:VOgEj0^:mNfET1]:dNgE\\1_;00001O0000001O0000001O0000001O00000O2O00002N2N2N2N3M2N2N2N3M2N2N2N3^Na1jMV2ZOf0N3L3TIXI[6Z7M3M2N3M3L3N3M3M2N3L4M2N3M3M3M2M4M3M3M5K4K6K5K4L5K4L5J6K4L5K4L5J6K4L5K4L5K5J5L5K4L5K5J5L5K4L5K4K6K5K[g6"}, "label": "the woman petting the horse on the head"}]}
{"id": 11065, "image": "COCO_train2014_000000011065.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the dark green jacket standing up\"."}], "task": "refering", "answer_template": "The \"the woman in the dark green jacket standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 64, 65, 66, 87, 88, 89, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 294, 295, 296, 317, 318, 319, 341, 342, 363, 364, 365], "bbox": [0.767625, 0.08296213808463251, 0.9752656249999999, 0.9865478841870824], "iscrowd": 0, "area": 29369.622499999998, "rle": {"size": [449, 640], "counts": "oag63e=9G9F9H9G9G9UKXNTMQ2i2WNiLR2T3VNdLm1Y3ZN`Li1]3_N[Ld1b3dNVL_1j3fNnK]1X2WMdL^1P1^1\\2TMcLa1l0]1a2SMcLa1h0_1e2PMbLd1d0_1i2nLbLe1a0`1m2kLaLh1=`1R3hLaLj18a1V3fLaLk16a1Y3dL`Ln12b1\\3aLbLn1MR2T3PLnLQ2Ga2m2_K\\MQ2AR3f2mJhMT2[Oa3_2\\JVNS2YOg3]2VJZNS2WOn3Y2oI`NU2TOS4W2hIeNU2ROY4T2cIjNT2PO`4Q2\\InNU2oNf4m1VITOU2bNU5V2fHWOo9c0RF]OS:>mEBW:6mEIY:NkE2Z:FiE:[:^OiEa0]:VOgEj0^:mNfET1]:dNgE\\1_;00001O0000001O0000001O0000001O00000O2O00002N2N2N2N3M2N2N2N3M2N2N2N3^Na1jMV2ZOf0N3L3TIXI[6Z7M3M2N3M3L3N3M3M2N3L4M2N3M3M3M2M4M3M3M5K4K6K5K4L5K4L5J6K4L5K4L5J6K4L5K4L5K5J5L5K4L5K5J5L5K4L5K4K6K5K[g6"}, "label": "the woman in the dark green jacket standing up"}]}
{"id": 11065, "image": "COCO_train2014_000000011065.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a brown hat leaning down to touch the foot of a horse\"."}], "task": "refering", "answer_template": "The \"a man wearing a brown hat leaning down to touch the foot of a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 281, 282, 283, 284, 300, 301, 304, 305, 307, 323, 327, 328, 330], "bbox": [0.0424375, 0.3836748329621381, 0.38940625, 0.90728285077951], "iscrowd": 0, "area": 28309.0418, "rle": {"size": [449, 640], "counts": "UV<1o=4SCMe;7oC3o;2eC6Z<i000O1O1O2N1O1M3N2M3N2N2N2M3N2N2N2N2M3NFaDVN];l1eDRNY;P2:3M2M2O1O1O1O1O1N2O1O11O001O1O0IhDfMS;MoD[22eMn:3nDX2Z;jMeDT2[;90O01L3L5K4M4K4L5K4N3L3N2M3N2N2N2N2N2N2O2M2O1N21O2N2N2N2N2N2O1N2N2N2N1O2N1N2O22M5K6K5J4M1N2O1N2O1N2O1N10fNfFmMZ9Q2jFlMW9S2lFkMT9T2oFiMR9V2QGhMo8W2SGgMn8W2UGhMk8W2WGgMj8X2XGgMh8X2ZGfMg8Y2[GfMe8X2^GfMc8Y2`GeM?_O`6k2XI]M:L[6f2WK]MdMKW6g2ULbMVM3d6Z2WL\\Ni3e1UL\\Nk3f1SLZNm3g1QL[Nn3f1QLZNo3g1oKZNQ4g1eKbN[4_1cKcN\\4_1aKbN_4_1^KcNb4^1\\KcNd4^1ZKcNf4^1WKdNi4^1TKcNl4^1RKcNn4^1mJgNR5Z1WJ\\Oi5k0nIWOR6n0gIUOX6Q1_IROa6W40001O00001O000000000000001O00ZObIiI_6Q6iIlIW6S6jImIW6Q6jIoIV6Q6jIoIW6o5jIRJU6m5lISJU6k5lIUJU6j5kIVJU6i5lIWJU6g5mIXJS6d5QJ\\JP6^5VJaJj5Z5[JeJg5U5_JiJb5R5cJmJ_5m4gJRKY5j4kJUKW5e4oJYKR5b4Z2J6UN`FnNe9R1aFcNe9\\1bFYNb9g1dFnMb9Q2eFcM`9]2S1O001O0O23L5L3M4K5L4K5L;E8HO100O001008Hb0^Oc0]O9G0O2O00001O1O1N3N1O2N1O2N1N01TKbJS1a5bNcKb0c4POgKn0h8M4L3L4M4K5L6I7JSR[5"}, "label": "a man wearing a brown hat leaning down to touch the foot of a horse"}]}
{"id": 11065, "image": "COCO_train2014_000000011065.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man bending over touching a horse foot\"."}], "task": "refering", "answer_template": "The \"a man bending over touching a horse foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 281, 282, 283, 284, 300, 301, 304, 305, 307, 323, 327, 328, 330], "bbox": [0.0424375, 0.3836748329621381, 0.38940625, 0.90728285077951], "iscrowd": 0, "area": 28309.0418, "rle": {"size": [449, 640], "counts": "UV<1o=4SCMe;7oC3o;2eC6Z<i000O1O1O2N1O1M3N2M3N2N2N2M3N2N2N2N2M3NFaDVN];l1eDRNY;P2:3M2M2O1O1O1O1O1N2O1O11O001O1O0IhDfMS;MoD[22eMn:3nDX2Z;jMeDT2[;90O01L3L5K4M4K4L5K4N3L3N2M3N2N2N2N2N2N2O2M2O1N21O2N2N2N2N2N2O1N2N2N2N1O2N1N2O22M5K6K5J4M1N2O1N2O1N2O1N10fNfFmMZ9Q2jFlMW9S2lFkMT9T2oFiMR9V2QGhMo8W2SGgMn8W2UGhMk8W2WGgMj8X2XGgMh8X2ZGfMg8Y2[GfMe8X2^GfMc8Y2`GeM?_O`6k2XI]M:L[6f2WK]MdMKW6g2ULbMVM3d6Z2WL\\Ni3e1UL\\Nk3f1SLZNm3g1QL[Nn3f1QLZNo3g1oKZNQ4g1eKbN[4_1cKcN\\4_1aKbN_4_1^KcNb4^1\\KcNd4^1ZKcNf4^1WKdNi4^1TKcNl4^1RKcNn4^1mJgNR5Z1WJ\\Oi5k0nIWOR6n0gIUOX6Q1_IROa6W40001O00001O000000000000001O00ZObIiI_6Q6iIlIW6S6jImIW6Q6jIoIV6Q6jIoIW6o5jIRJU6m5lISJU6k5lIUJU6j5kIVJU6i5lIWJU6g5mIXJS6d5QJ\\JP6^5VJaJj5Z5[JeJg5U5_JiJb5R5cJmJ_5m4gJRKY5j4kJUKW5e4oJYKR5b4Z2J6UN`FnNe9R1aFcNe9\\1bFYNb9g1dFnMb9Q2eFcM`9]2S1O001O0O23L5L3M4K5L4K5L;E8HO100O001008Hb0^Oc0]O9G0O2O00001O1O1N3N1O2N1O2N1N01TKbJS1a5bNcKb0c4POgKn0h8M4L3L4M4K5L6I7JSR[5"}, "label": "a man bending over touching a horse foot"}]}
{"id": 224060, "image": "COCO_train2014_000000224060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wii remote near the adidas shoes\"."}], "task": "refering", "answer_template": "The \"the wii remote near the adidas shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 234, 235, 236, 237, 238, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 287, 307, 308, 309, 310, 311, 331, 332, 333, 334, 355], "bbox": [0.207296875, 0.549875, 0.5309062499999999, 0.8959375], "iscrowd": 0, "area": 14526.584499999999, "rle": {"size": [480, 640], "counts": "Sdn14k>5L4K4L3M2N2O1N1O2N2O1N1O2N2N2O0O2N2N2N101N2N2N101N2N2N1O2O1N2N1O2N2O10O0100O10O010000O010O100O010O1000O0100O010O100O01000O10O0100O10O010000O010O100O010O1000O0100O10O0100O10O10O100O010O100O01000O10O0100O010O1000O0100O10O0100O10O10O100O010O100O01000O10O0100O10O010000O010O100O010O100O01000O10O0100O010O01000O010O010O10O010O10O10O010O0100O01000O01N1O2N1O2N1O2N2N1O2N1O2N1O2O0O2N2N1O2N1O2N1O2N1O2N2N1O2N1O3N9F:F:F;EfW\\4"}, "label": "the wii remote near the adidas shoes"}]}
{"id": 224060, "image": "COCO_train2014_000000224060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the controller closest to the shoe\"."}], "task": "refering", "answer_template": "The \"the controller closest to the shoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 234, 235, 236, 237, 238, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 287, 307, 308, 309, 310, 311, 331, 332, 333, 334, 355], "bbox": [0.207296875, 0.549875, 0.5309062499999999, 0.8959375], "iscrowd": 0, "area": 14526.584499999999, "rle": {"size": [480, 640], "counts": "Sdn14k>5L4K4L3M2N2O1N1O2N2O1N1O2N2N2O0O2N2N2N101N2N2N101N2N2N1O2O1N2N1O2N2O10O0100O10O010000O010O100O010O1000O0100O010O100O01000O10O0100O10O010000O010O100O010O1000O0100O10O0100O10O10O100O010O100O01000O10O0100O010O1000O0100O10O0100O10O10O100O010O100O01000O10O0100O10O010000O010O100O010O100O01000O10O0100O010O01000O010O010O10O010O10O10O010O0100O01000O01N1O2N1O2N1O2N2N1O2N1O2N1O2O0O2N2N1O2N1O2N1O2N1O2N2N1O2N1O3N9F:F:F;EfW\\4"}, "label": "the controller closest to the shoe"}]}
{"id": 224060, "image": "COCO_train2014_000000224060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the remote bwtween the hard drive and the other remote\"."}], "task": "refering", "answer_template": "The \"the remote bwtween the hard drive and the other remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 312, 313], "bbox": [0.310109375, 0.5296458333333333, 0.664046875, 0.8127916666666667], "iscrowd": 0, "area": 12477.227150000002, "rle": {"size": [480, 640], "counts": "_Rm25j>1O2N2N1O2N2O0O2O1N101N101N2O0O2O1N101O1N101N101N2O0O2O1N10100O00100O010O00100O00100O00100O0010O01O10O01O10O01O10O01O10O01O010O1O010O1O010O1O010O00100O00100O00100O0010O01O10O01O10O01O10O01O10O010O01O10O01O010O0010O01O010O0010O01O01O01O010O0010O01O010O0010O01O01O01O010O0010O01O010O0010O01O010O00010O0010O01O010O010O0010O01O01O01O010O0010O01O010O0010O01O100O00100O1O010O1O100O00100O1O010O1OO2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O2N1O1O2N1N9H;E;EV^T3"}, "label": "the remote bwtween the hard drive and the other remote"}]}
{"id": 224060, "image": "COCO_train2014_000000224060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wii controller on the right\"."}], "task": "refering", "answer_template": "The \"wii controller on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 312, 313], "bbox": [0.310109375, 0.5296458333333333, 0.664046875, 0.8127916666666667], "iscrowd": 0, "area": 12477.227150000002, "rle": {"size": [480, 640], "counts": "_Rm25j>1O2N2N1O2N2O0O2O1N101N101N2O0O2O1N101O1N101N101N2O0O2O1N10100O00100O010O00100O00100O00100O0010O01O10O01O10O01O10O01O10O01O010O1O010O1O010O1O010O00100O00100O00100O0010O01O10O01O10O01O10O01O10O010O01O10O01O010O0010O01O010O0010O01O01O01O010O0010O01O010O0010O01O01O01O010O0010O01O010O0010O01O010O00010O0010O01O010O010O0010O01O01O01O010O0010O01O010O0010O01O100O00100O1O010O1O100O00100O1O010O1OO2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O2N1O1O2N1N9H;E;EV^T3"}, "label": "wii controller on the right"}]}
{"id": 224060, "image": "COCO_train2014_000000224060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red on black adidas sneaker worn on the right foot\"."}], "task": "refering", "answer_template": "The \"a red on black adidas sneaker worn on the right foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 256, 277, 278, 279, 280, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.6808958333333333, 0.185390625, 0.9887708333333334], "iscrowd": 0, "area": 11450.631300000005, "rle": {"size": [480, 640], "counts": "T=1n>g1YN1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2M3N2N2M3N2N2M3N2N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1000000O1000000O1000000O10000000YNcEkN\\:S1WF]Ni9`1^F\\Nb9a1cF]N]9a1gF]NX9a1mF]NS9`1RG^Nn8_1WG_Ni8^1\\G`Nc8^1cG_N]8_1gG_NY8^1lG`NT8]1QHaNP8]1SHaNm7]1WHaNj7]1XHbNi7\\1YHcNg7[1[HeNf7Y1\\HfNe7X1]HgNd7W1^HhNb7V1`HjNa7T1aHkN`7S1bHlN^7R1Z3N2N2M2O2N2M^lc7"}, "label": "a red on black adidas sneaker worn on the right foot"}]}
{"id": 224060, "image": "COCO_train2014_000000224060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and black tennis shoe is next to a white ottoman\"."}], "task": "refering", "answer_template": "The \"a red and black tennis shoe is next to a white ottoman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 256, 277, 278, 279, 280, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.6808958333333333, 0.185390625, 0.9887708333333334], "iscrowd": 0, "area": 11450.631300000005, "rle": {"size": [480, 640], "counts": "T=1n>g1YN1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2M3N2N2M3N2N2M3N2N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1000000O1000000O1000000O10000000YNcEkN\\:S1WF]Ni9`1^F\\Nb9a1cF]N]9a1gF]NX9a1mF]NS9`1RG^Nn8_1WG_Ni8^1\\G`Nc8^1cG_N]8_1gG_NY8^1lG`NT8]1QHaNP8]1SHaNm7]1WHaNj7]1XHbNi7\\1YHcNg7[1[HeNf7Y1\\HfNe7X1]HgNd7W1^HhNb7V1`HjNa7T1aHkN`7S1bHlN^7R1Z3N2N2M2O2N2M^lc7"}, "label": "a red and black tennis shoe is next to a white ottoman"}]}
{"id": 248640, "image": "COCO_train2014_000000248640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat on the left looking out the window\"."}], "task": "refering", "answer_template": "The \"the cat on the left looking out the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0038906250000000004, 0.49414583333333334, 0.313640625, 0.9876666666666666], "iscrowd": 0, "area": 38771.78655, "rle": {"size": [480, 640], "counts": "`W17h>R2nMR2nMR1nN1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1000000O10000O1000O10O1000000O10000O1000000O1000000O10000O1000000O10000O10O1000O1000000O10000O1000000O1000000O10000000000000000O01000000000000000O1000000000000000000O10000000000000000O010000000O100000000O10000000000O100000000O100000000O10O10000000O100000000O100000000O10`N[HeKe7Y4iH[KW7b4XIRKh6l4fIhJZ6U5i1O2M3N2M3N1N3N2M3N2WLnEa2T:\\MZFX2h9dMfFo1]9nMPGf1R9VNUGd1n8XNYGb1j8ZN\\G`1g8]N`G\\1V;J7I7I6J7I6K6Ie^]6"}, "label": "the cat on the left looking out the window"}]}
{"id": 248640, "image": "COCO_train2014_000000248640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silhouette of a cat sitting to the left of a flower vase\"."}], "task": "refering", "answer_template": "The \"the silhouette of a cat sitting to the left of a flower vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0038906250000000004, 0.49414583333333334, 0.313640625, 0.9876666666666666], "iscrowd": 0, "area": 38771.78655, "rle": {"size": [480, 640], "counts": "`W17h>R2nMR2nMR1nN1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1000000O10000O1000O10O1000000O10000O1000000O1000000O10000O1000000O10000O10O1000O1000000O10000O1000000O1000000O10000000000000000O01000000000000000O1000000000000000000O10000000000000000O010000000O100000000O10000000000O100000000O100000000O10O10000000O100000000O100000000O10`N[HeKe7Y4iH[KW7b4XIRKh6l4fIhJZ6U5i1O2M3N2M3N1N3N2M3N2WLnEa2T:\\MZFX2h9dMfFo1]9nMPGf1R9VNUGd1n8XNYGb1j8ZN\\G`1g8]N`G\\1V;J7I7I6J7I6K6Ie^]6"}, "label": "the silhouette of a cat sitting to the left of a flower vase"}]}
{"id": 248640, "image": "COCO_train2014_000000248640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light cat with a bell around its neck , looking at a black cat\"."}], "task": "refering", "answer_template": "The \"a light cat with a bell around its neck , looking at a black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 388, 389, 390], "bbox": [0.635484375, 0.13333333333333333, 0.99515625, 0.98925], "iscrowd": 0, "area": 61780.87874999999, "rle": {"size": [480, 640], "counts": "PPo59e>7I4M4K4M3L5K3N2N1O2fNhNiDY1V;oNbDS1[;UO^Dl0`;\\1N1N3N1N3M2O2M3N2M3fNaLoFb3m8fLmF\\3o8mLiFV3R8[LWH`0EV3T8^LSH=FX3V8_LQH9F[3X8`LnG6I[3X8cLkG5JY3Z8fLhG4KW3\\8hLfG5JT3^8jLfG5HS3a8iLfGZ4Z8hKdGX4\\8jKbGV4^8lK`GU4_8lK`GT4`8V101O0000001O000O2O00000000000001O001O001O001O001O001O001O001O001O001O00001O00001O0000001O000100O2N100O2SHgIo6Z6mHlIP7V6lHnIS7S6iHQJU7Q6gHSJW7o5eHUJZ7l5bHXJS1XOP5_7jJgHU5Y7eJnHZ5S7_JTIDA^5Z7lJhIR5X6lJlIR5T6mJoIR5P6mJRJR5n5mJUJQ5k5mJYJQ5h5mJ[JR5d5mJ_JQ5a5nJaJQ5_5mJeJR5Z5mJiJQ5d7O2N101N1O2O0O2N1O2O0O2N1O2O0O1O2N1O2O0O1O1O1O1O101N1O1O1O100O1O1O1O100O1O2N100O1O100O2O0O2N2O1N1O2O1N101N1O01O00001O0O101O001O00001O001O1O001O001O001O1O001O001O0O2O1O001O001O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1N2O1O2N1O1O2N1O1O2N1O1O1O2N1OUX1"}, "label": "a light cat with a bell around its neck , looking at a black cat"}]}
{"id": 248640, "image": "COCO_train2014_000000248640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color cat standing tghe another cat\"."}], "task": "refering", "answer_template": "The \"white color cat standing tghe another cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 388, 389, 390], "bbox": [0.635484375, 0.13333333333333333, 0.99515625, 0.98925], "iscrowd": 0, "area": 61780.87874999999, "rle": {"size": [480, 640], "counts": "PPo59e>7I4M4K4M3L5K3N2N1O2fNhNiDY1V;oNbDS1[;UO^Dl0`;\\1N1N3N1N3M2O2M3N2M3fNaLoFb3m8fLmF\\3o8mLiFV3R8[LWH`0EV3T8^LSH=FX3V8_LQH9F[3X8`LnG6I[3X8cLkG5JY3Z8fLhG4KW3\\8hLfG5JT3^8jLfG5HS3a8iLfGZ4Z8hKdGX4\\8jKbGV4^8lK`GU4_8lK`GT4`8V101O0000001O000O2O00000000000001O001O001O001O001O001O001O001O001O001O00001O00001O0000001O000100O2N100O2SHgIo6Z6mHlIP7V6lHnIS7S6iHQJU7Q6gHSJW7o5eHUJZ7l5bHXJS1XOP5_7jJgHU5Y7eJnHZ5S7_JTIDA^5Z7lJhIR5X6lJlIR5T6mJoIR5P6mJRJR5n5mJUJQ5k5mJYJQ5h5mJ[JR5d5mJ_JQ5a5nJaJQ5_5mJeJR5Z5mJiJQ5d7O2N101N1O2O0O2N1O2O0O2N1O2O0O1O2N1O2O0O1O1O1O1O101N1O1O1O100O1O1O1O100O1O2N100O1O100O2O0O2N2O1N1O2O1N101N1O01O00001O0O101O001O00001O001O1O001O001O001O1O001O001O0O2O1O001O001O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1N2O1O2N1O1O2N1O1O2N1O1O1O2N1OUX1"}, "label": "white color cat standing tghe another cat"}]}
{"id": 84803, "image": "COCO_train2014_000000084803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a red parasol\"."}], "task": "refering", "answer_template": "The \"a woman holding a red parasol\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 236, 237, 238, 259, 260, 261, 262, 263, 264], "bbox": [0.259125, 0.41523364485981307, 0.501703125, 0.7906775700934578], "iscrowd": 0, "area": 10758.758249999997, "rle": {"size": [428, 640], "counts": "jcU26S=j0WO8[NoNaF_1R9hNgFZ1W9kNdFX1Z9iNcF[1Z9hNcFZ1[9hNcFY1\\9iNbFX1]9jN^FZ1a9Y1O001O1O2N2O2N7I1O1O1O0ZGTLl7m3mGYLS8`40001O00000000O10000O2O1O1N101O1O1O001N2O1O001O2M2O1O3M3L4M3M4K6K7I1O0O101O001O001NSFlLd9T3\\FlLd9T3\\FlLe9S3[FmLe9S3ZFnLf9Q3[FoLf9P3ZFPMf9P3ZFPMf9P3ZFPMf9P3ZFPMg9n2YFSMh9l2XFTMi9k1ZFfNM_Oj9i1\\FfNJAk9g1^FeNGEV:[1UFnNEGV:Z1WFnNCHV:Y1XFoNAIV:<hE6a0EAIV:;lE3?G@KT::WFJ71^OKS::\\FC77[OLS:8]GL`NLT:6]GN^NMV:2_GO\\NOX:K`G6XNO\\<0dC0\\<OeC1[<OeC1\\<NcC4\\<LcC5]<JdC7[<IdC8]<GcC9]<GbC:^<FbC9_<FaC;`<D`C<`<D`C<_<EaC;_<DbC<^<DbC;_<EaC;_<EaC;_<EaC;h<0O1000000O2O00000O100000000O100000CVC7k<FYC9n<O2N1Ofj0L\\UO6J40OO1O000001O0000001O0000001O001O2M3NVPU4"}, "label": "a woman holding a red parasol"}]}
{"id": 84803, "image": "COCO_train2014_000000084803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman boting\"."}], "task": "refering", "answer_template": "The \"a woman boting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 236, 237, 238, 259, 260, 261, 262, 263, 264], "bbox": [0.259125, 0.41523364485981307, 0.501703125, 0.7906775700934578], "iscrowd": 0, "area": 10758.758249999997, "rle": {"size": [428, 640], "counts": "jcU26S=j0WO8[NoNaF_1R9hNgFZ1W9kNdFX1Z9iNcF[1Z9hNcFZ1[9hNcFY1\\9iNbFX1]9jN^FZ1a9Y1O001O1O2N2O2N7I1O1O1O0ZGTLl7m3mGYLS8`40001O00000000O10000O2O1O1N101O1O1O001N2O1O001O2M2O1O3M3L4M3M4K6K7I1O0O101O001O001NSFlLd9T3\\FlLd9T3\\FlLe9S3[FmLe9S3ZFnLf9Q3[FoLf9P3ZFPMf9P3ZFPMf9P3ZFPMf9P3ZFPMg9n2YFSMh9l2XFTMi9k1ZFfNM_Oj9i1\\FfNJAk9g1^FeNGEV:[1UFnNEGV:Z1WFnNCHV:Y1XFoNAIV:<hE6a0EAIV:;lE3?G@KT::WFJ71^OKS::\\FC77[OLS:8]GL`NLT:6]GN^NMV:2_GO\\NOX:K`G6XNO\\<0dC0\\<OeC1[<OeC1\\<NcC4\\<LcC5]<JdC7[<IdC8]<GcC9]<GbC:^<FbC9_<FaC;`<D`C<`<D`C<_<EaC;_<DbC<^<DbC;_<EaC;_<EaC;_<EaC;h<0O1000000O2O00000O100000000O100000CVC7k<FYC9n<O2N1Ofj0L\\UO6J40OO1O000001O0000001O0000001O001O2M3NVPU4"}, "label": "a woman boting"}]}
{"id": 527173, "image": "COCO_train2014_000000527173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man leaning on bike on boat\"."}], "task": "refering", "answer_template": "The \"man leaning on bike on boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 44, 45, 61, 62, 63, 80, 81, 82, 99, 100, 117, 135], "bbox": [0.42744, 0.13395027624309394, 0.5796, 0.5607458563535912], "iscrowd": 0, "area": 3225.1443000000013, "rle": {"size": [362, 500], "counts": "eg[24U;6K0O1O1O1O2O0O1000000O100000001O0O10000000000O1000000O010XO^OhFb08^OP8<\\G6d0Al7?[G0i0Dh7=^G07\\O7:Q8:aG07\\O7=P1IT5?^IO6]O8?m0IT5<aIO3A:>j0GW5<aIOOE?<g0DY5=cI:=Hd0A\\5=cI:=K`0_O_5<eI:<N=\\Oa5=fI9<1:XOe5>fI8;56VOj5<eI::47VOj5<eI::38WOi5<eI::29XOh5<fI991:[Og5;fI990;\\Of5;fI99O<]Oe5:gI:8N<XOZOJ[6d0jI<4M>VOo5?`I>4K`7D^H`02KL^OQ75TI>1NICR7LXI3CI<d0GGQ7BRJ1YOe0DK_8^OoGf0A0X9OhF5U9IlF:R9EnF>Q9@nFd0P9[OPGh0n8WORGJMe0m8_OUGM5?e8CVGM>8]8IVGOe01V8NTG2m0Jo72UG4V:KjE5d:11O2N2N1N3NZ^Z2"}, "label": "man leaning on bike on boat"}]}
{"id": 527173, "image": "COCO_train2014_000000527173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man not wearing a hat standing in a boat\"."}], "task": "refering", "answer_template": "The \"a man not wearing a hat standing in a boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 44, 45, 61, 62, 63, 80, 81, 82, 99, 100, 117, 135], "bbox": [0.42744, 0.13395027624309394, 0.5796, 0.5607458563535912], "iscrowd": 0, "area": 3225.1443000000013, "rle": {"size": [362, 500], "counts": "eg[24U;6K0O1O1O1O2O0O1000000O100000001O0O10000000000O1000000O010XO^OhFb08^OP8<\\G6d0Al7?[G0i0Dh7=^G07\\O7:Q8:aG07\\O7=P1IT5?^IO6]O8?m0IT5<aIO3A:>j0GW5<aIOOE?<g0DY5=cI:=Hd0A\\5=cI:=K`0_O_5<eI:<N=\\Oa5=fI9<1:XOe5>fI8;56VOj5<eI::47VOj5<eI::38WOi5<eI::29XOh5<fI991:[Og5;fI990;\\Of5;fI99O<]Oe5:gI:8N<XOZOJ[6d0jI<4M>VOo5?`I>4K`7D^H`02KL^OQ75TI>1NICR7LXI3CI<d0GGQ7BRJ1YOe0DK_8^OoGf0A0X9OhF5U9IlF:R9EnF>Q9@nFd0P9[OPGh0n8WORGJMe0m8_OUGM5?e8CVGM>8]8IVGOe01V8NTG2m0Jo72UG4V:KjE5d:11O2N2N1N3NZ^Z2"}, "label": "a man not wearing a hat standing in a boat"}]}
{"id": 527173, "image": "COCO_train2014_000000527173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man rowing boat\"."}], "task": "refering", "answer_template": "The \"man rowing boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 113, 114, 115, 116, 131, 132, 133, 134, 150, 151, 152], "bbox": [0.31492000000000003, 0.33430939226519335, 0.5097200000000001, 0.6392265193370165], "iscrowd": 0, "area": 6073.554199999999, "rle": {"size": [362, 500], "counts": "`fg14^:m0H7I8H5K3M3M3M3M4L3N2M3O1O1O1O2N1O1O1O1O1@`001N10000O1000001N1000000000001O0O10000000000000bMWH[1i7cN[H[1e7cN_H[1a7cNdHZ1^7bNfH\\1\\7_NjH^1X7^NnH^1W7[NQI_1b8K5K1O001O00001O00001O001O000000000000O10O100[OmE3S:LnE4R:KPF4P:LPF4o9LSF3m9LTF4k9LWF3h9MYF3g9MZF2g9LZF4i9HYF7i9FXF:k9BWF=[:01O1O001O3M2N3M2N]\\f2"}, "label": "man rowing boat"}]}
{"id": 527173, "image": "COCO_train2014_000000527173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in a boat while wearing a hat\"."}], "task": "refering", "answer_template": "The \"a man sitting in a boat while wearing a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 113, 114, 115, 116, 131, 132, 133, 134, 150, 151, 152], "bbox": [0.31492000000000003, 0.33430939226519335, 0.5097200000000001, 0.6392265193370165], "iscrowd": 0, "area": 6073.554199999999, "rle": {"size": [362, 500], "counts": "`fg14^:m0H7I8H5K3M3M3M3M4L3N2M3O1O1O1O2N1O1O1O1O1@`001N10000O1000001N1000000000001O0O10000000000000bMWH[1i7cN[H[1e7cN_H[1a7cNdHZ1^7bNfH\\1\\7_NjH^1X7^NnH^1W7[NQI_1b8K5K1O001O00001O00001O001O000000000000O10O100[OmE3S:LnE4R:KPF4P:LPF4o9LSF3m9LTF4k9LWF3h9MYF3g9MZF2g9LZF4i9HYF7i9FXF:k9BWF=[:01O1O001O3M2N3M2N]\\f2"}, "label": "a man sitting in a boat while wearing a hat"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed brown teddy bear\"."}], "task": "refering", "answer_template": "The \"a stuffed brown teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 173, 174, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 297, 298, 299, 300, 301, 302, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334], "bbox": [0.4846389496717724, 0.4494375, 0.9283807439824945, 0.912359375], "iscrowd": 0, "area": 38292.323099999994, "rle": {"size": [640, 457], "counts": "k_Z41hc07I8H8H7F;D;D=D<G801N2O001O0O2O1O0^@WMl=k2RBXMl=i2RBZMl=f2TB]Mh=e2VB^Mh=c2VBaMg=_2XBdMf=]2XBgMd=Z2\\BhMb=Y2\\BkMa=V2]BmM`=T2_BPN^=Q2`BRN^=o1aBSN]=m1bBWNZ=k1dBXNZ=g1fB]NW=b1iBaNT=_1lBeNQ=[1oBgNo<X1QClNk<T1UCoNi<Q1VCROh<m0YCVOd<i0\\CZOa<g0_C\\O^<c0bC@\\<?eCDW<=hCFV<9jCKS<4nCNP<2oC1o;NRD5l0hLV9R3nE9c0lL_9k2nE<8QMj9b2nE`0OUMS:Z2nEe0EXM]:R2oEh0ZO^Mg:j1nEV2R:iMoEW2R:gMnEZ2R:fMnEZ2R:eMnE\\2R:cMoE]2Q:cMnE^2R:aMnE`2R:_MoEa2Q:_MnEb2R:]MoEc2Q:\\MoEe2R:ZMnEf2R:YMnEh2R:WMoEi2Q:VMoEk2Q:UMnEl2R:SMoEm2Q:RMoEo2Q:QMoEo2Q:PMoEQ3Q:nLPFR3Q:mLnES3S:lLnET3R:kLnEV3R:jLmEW3S:hLnEX3R:gLnEZ3R:fLnEZ3R:fLmE[3S:eLmE[3S:dLmE]3T:bLlE^3T:bLkE_3U:`LkEa3U:_LkEa3U:_LjEb3V:^LjEb3V:]LjEd3V:\\LjEd3V:\\LiEe3W:ZLjEf3W:YLhEh3X:XLgEi3Y:WLgEi3Y:VLgEk3Y:ULgEk3Y:ULfEl3Z:TLfEl3Z:SLfEm3[:SLeEl3]:SLdEk3^:SLcEl3^:TLbEk3`:TL`Ej3c:UL]Ej3d:UL]Ej3e:UL[Ej3f:VLZEi3h:ULYEj3i:ULWEj3j:VLVEh3m:WLSEh3o:VLREi3o:WLQEh3Q;WLoDh3S;WLmDh3T;WLmDh3U;WLkDg3X;XLiDg3W;XLjDi3V;VLjDk3U;ULkDl3U;SLkDn3U;PLlDQ4S;oKmDR4S;mKmDT4S;jKnDW4Q;iKoDX4Q;gKoD[4P;dKPE]4o:bKRE_4n:`KREa4m:_KSEb4m:]KSEd4m:ZKTEg4k:YKVEg4j:XKVEi4j:VKVEk4i:UKWEl4i:SKWEn4i:PKXEQ5g:oJYER5g:mJYET5f:lJZEU5f:jJZEW5f:hJZEY5e:gJ[EZ5e:dJ\\E]5d:bJ\\E^5d:bJ\\E^5e:aJ[E_5e:aJ\\E^5a:eJbEX5\\:jJfES5Y:nJkEo4R:TKPFj4n9XKTFf4j9\\KYFa4d9bK^F\\4`9fKcFW4[9kKgFS4V9oKnFn3P9TLRGj3l8XLWGe3g8]L[G`3g8_L\\G^3d8bL^G\\3b8dLaGY3`8eLcGY3]8gLfGV3Z8jLhGT3Y8kLkGQ3U8oLVHf2j7ZMbHZ2_7eMmHo1S7PNZIc1g6]NdIX1]6gNoIm0Q6SO[Ja0f5^OeJ7[5IlJ0T50oJMR51QKMo43SKKm45VKHk47WKGi49YKEg4;\\KAf4>\\K@d4?`K^O`4b0bK\\O_4c0cK[O]4e0fKXOZ4h0hKVOY4h0jKVOV4>YL_Oh34fLJf]d0"}, "label": "a stuffed brown teddy bear"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young kid holding a brown bear\"."}], "task": "refering", "answer_template": "The \"a young kid holding a brown bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 90, 91, 92, 93, 94, 106, 107, 108, 109, 110, 122, 123, 124, 125, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 190, 191, 203, 204, 205, 206, 207, 217, 218, 219, 220, 221, 222, 223, 233, 234, 237, 238, 239, 253, 254, 255, 270, 271, 286, 287, 302, 303, 329, 330, 331, 332, 345, 346, 347, 348], "bbox": [0.5809409190371991, 0.1845, 1.0, 0.9431406250000001], "iscrowd": 0, "area": 35147.842749999996, "rle": {"size": [640, 457], "counts": "[VV51mc05J5L5K5J6K4L2M31O1OoIPOSIo0l6VOPIj0o6ZOoHe0Q7_OlH?T7EjH:V7GjH8U7JkH5U7KlH4S7MnH2Q70oHOQ71PINo64QIKo65QIKn67QIIo67QIIn68RIHm6:RIFn6:RIFm6<RID_2[OZLQ1W1D]2ESLi0_1B]2OkK?h1B[2:cK5R2AY2d0\\KK\\2AX2m0SKBe2AX2T1lJZOm2BW2U1jJYOP3BV2U1jJYOP3BW2T1iJYOQ3CV2T1hJZOR3BV2T1hJYOS3C;9gJk0k1YOS3C9<hJh0l1XOT3D6`0iJc0l1YOV3D2d0kJ`0l1XOW3D0g0lJ=m1WOW3FNi0mJ:m1XOX3ELm0mJ6o1WOY3FHQ1oJ2P2VOZ3GFS1PKOo1XO[3FEU1PKMP2WO\\3GCX1oJKQ2VO]3GBZ1oJIR2UO^3H@\\1oJGR2UO`3H^O^1oJES2UO_3I^O_1PKBS2VO_3I]Oa1PK@T2VO_3I\\Oc1PK^OT2VOa3IZOe1PK\\OU2VOa3IYOg1PKZOV2VOa3IXOi1oJYOW2UOc3IVOk1nJXOY2TOc3IVOl1lJXO[2SOc3IUOn1kJXO\\2QOc3JUOm3WMYLe3JSOn3XMXLe3JROo3YMWLe3JROo3YMVLf3KPOP4YMVLg3JPOP4YMVLg3JPOQ4XMULh3JoNR4XMTLi3KoNR4WMSLj3KoNS4VMRLk3KnNU4VMPLl3KnNV4TMPLn3JnNW4SMoKo3JnNW4SMoKo3JmNY4SMmKP4JmNZ4QMmKR4ImN[4PMlKR4JmN\\4PMjKS4JmN\\4oLkKT4ImN^4mLiKV4IlN`4mLgKW4IlNa4lLfKX4IlNb4jLfK[4GjNe4jLdK]4EjNh4hLcK^4EjNh4hLcK_4DhNk4gLcK`4BiNk4gLcKa4AhNm4fLbKb4AiNl4eLcKc4@iNm4bLdKf4^OhNn4bLdKf4]OjNo4_LdKh4YOlNS5[LeKi4UOPOV5VLfKj4POSO[5RLeKk4mNWO^5mKfKl4hNZOc5hKfKa:Z4^EgKa:Z4^EgKb:Y4]EgKd:Y4ZEiKe:X4ZEiKf:U4[EmKc:S4\\EoKd:o3]ESLa:l3`EVL_:i3aEYL]:g3cE[L\\:d3dE]L\\:b3dE`LZ:_3fEdLY:[3gEfLX:Z3hEhLW:W3iEjLV:V3jElLU:S3kEnLT:Q3mEQMR:n2nESMR:l2mEVMR:i2oEYMP:f2PF[Mo9e2QF]Mn9a2SF`Mm9_2SFcMl9\\2TFeMm9X2TFlMi9kNkEg0<b0f9fNoEg0;g0d9_NRFi0;k0a9[NTFj0;o0_9UNWFk0:T1]9nMZFm0:Y1Z9hM]Fn09^1X9aMaFP17c1W9ZMbFS17l1P9nLkFU15Y2f8`LVGV14f2\\8QLaGY12R3T8bKlG[10^3j7TKYH]1M_3j7RK\\H^1J`3k7oJ]Ha1H`3k7nJ_Ha1Fa3l7lJ`Hb1Db3l7kJaHc1Ca3m7jJcHd1@b3n7hJdHe1^Oc3n7gJfHe1\\Od3o7eJfHg1[Od3o7dJhHg1WOg3Q8aJjHi1QOh3V8]JjHm1lNg3`:[L[Eg3f:[LUEg3l:[LoDg3Q;\\LjDf3W;\\LdDf3];\\L^De3c;\\21O1O1O001N2O1O1O1eKVD>k;AZD9h;F]D3e;LaDNa;1dDH_;7fDC\\;;jD^OY;a0lDYOV;f0oDSOT;k0REoNP;P1UEiNn:V1VEdNm:Z1YEVNX;d1j3B>F:E<@c0]O\\F"}, "label": "a young kid holding a brown bear"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy wearing a baseball cap\"."}], "task": "refering", "answer_template": "The \"a little boy wearing a baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 90, 91, 92, 93, 94, 106, 107, 108, 109, 110, 122, 123, 124, 125, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 190, 191, 203, 204, 205, 206, 207, 217, 218, 219, 220, 221, 222, 223, 233, 234, 237, 238, 239, 253, 254, 255, 270, 271, 286, 287, 302, 303, 329, 330, 331, 332, 345, 346, 347, 348], "bbox": [0.5809409190371991, 0.1845, 1.0, 0.9431406250000001], "iscrowd": 0, "area": 35147.842749999996, "rle": {"size": [640, 457], "counts": "[VV51mc05J5L5K5J6K4L2M31O1OoIPOSIo0l6VOPIj0o6ZOoHe0Q7_OlH?T7EjH:V7GjH8U7JkH5U7KlH4S7MnH2Q70oHOQ71PINo64QIKo65QIKn67QIIo67QIIn68RIHm6:RIFn6:RIFm6<RID_2[OZLQ1W1D]2ESLi0_1B]2OkK?h1B[2:cK5R2AY2d0\\KK\\2AX2m0SKBe2AX2T1lJZOm2BW2U1jJYOP3BV2U1jJYOP3BW2T1iJYOQ3CV2T1hJZOR3BV2T1hJYOS3C;9gJk0k1YOS3C9<hJh0l1XOT3D6`0iJc0l1YOV3D2d0kJ`0l1XOW3D0g0lJ=m1WOW3FNi0mJ:m1XOX3ELm0mJ6o1WOY3FHQ1oJ2P2VOZ3GFS1PKOo1XO[3FEU1PKMP2WO\\3GCX1oJKQ2VO]3GBZ1oJIR2UO^3H@\\1oJGR2UO`3H^O^1oJES2UO_3I^O_1PKBS2VO_3I]Oa1PK@T2VO_3I\\Oc1PK^OT2VOa3IZOe1PK\\OU2VOa3IYOg1PKZOV2VOa3IXOi1oJYOW2UOc3IVOk1nJXOY2TOc3IVOl1lJXO[2SOc3IUOn1kJXO\\2QOc3JUOm3WMYLe3JSOn3XMXLe3JROo3YMWLe3JROo3YMVLf3KPOP4YMVLg3JPOP4YMVLg3JPOQ4XMULh3JoNR4XMTLi3KoNR4WMSLj3KoNS4VMRLk3KnNU4VMPLl3KnNV4TMPLn3JnNW4SMoKo3JnNW4SMoKo3JmNY4SMmKP4JmNZ4QMmKR4ImN[4PMlKR4JmN\\4PMjKS4JmN\\4oLkKT4ImN^4mLiKV4IlN`4mLgKW4IlNa4lLfKX4IlNb4jLfK[4GjNe4jLdK]4EjNh4hLcK^4EjNh4hLcK_4DhNk4gLcK`4BiNk4gLcKa4AhNm4fLbKb4AiNl4eLcKc4@iNm4bLdKf4^OhNn4bLdKf4]OjNo4_LdKh4YOlNS5[LeKi4UOPOV5VLfKj4POSO[5RLeKk4mNWO^5mKfKl4hNZOc5hKfKa:Z4^EgKa:Z4^EgKb:Y4]EgKd:Y4ZEiKe:X4ZEiKf:U4[EmKc:S4\\EoKd:o3]ESLa:l3`EVL_:i3aEYL]:g3cE[L\\:d3dE]L\\:b3dE`LZ:_3fEdLY:[3gEfLX:Z3hEhLW:W3iEjLV:V3jElLU:S3kEnLT:Q3mEQMR:n2nESMR:l2mEVMR:i2oEYMP:f2PF[Mo9e2QF]Mn9a2SF`Mm9_2SFcMl9\\2TFeMm9X2TFlMi9kNkEg0<b0f9fNoEg0;g0d9_NRFi0;k0a9[NTFj0;o0_9UNWFk0:T1]9nMZFm0:Y1Z9hM]Fn09^1X9aMaFP17c1W9ZMbFS17l1P9nLkFU15Y2f8`LVGV14f2\\8QLaGY12R3T8bKlG[10^3j7TKYH]1M_3j7RK\\H^1J`3k7oJ]Ha1H`3k7nJ_Ha1Fa3l7lJ`Hb1Db3l7kJaHc1Ca3m7jJcHd1@b3n7hJdHe1^Oc3n7gJfHe1\\Od3o7eJfHg1[Od3o7dJhHg1WOg3Q8aJjHi1QOh3V8]JjHm1lNg3`:[L[Eg3f:[LUEg3l:[LoDg3Q;\\LjDf3W;\\LdDf3];\\L^De3c;\\21O1O1O001N2O1O1O1eKVD>k;AZD9h;F]D3e;LaDNa;1dDH_;7fDC\\;;jD^OY;a0lDYOV;f0oDSOT;k0REoNP;P1UEiNn:V1VEdNm:Z1YEVNX;d1j3B>F:E<@c0]O\\F"}, "label": "a little boy wearing a baseball cap"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with white bear and no hat\"."}], "task": "refering", "answer_template": "The \"boy with white bear and no hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 50, 51, 52, 53, 54, 66, 67, 68, 69, 70, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 160, 161, 164, 165, 212, 213, 226, 227, 228, 229, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 288, 289, 290, 291, 292, 293, 294, 304, 308, 309, 310, 323, 324, 325, 326, 336, 337, 338, 339, 340, 341, 342, 352, 353, 354, 355, 356, 357, 358], "bbox": [0.0031291028446389494, 0.12357812500000001, 0.4364551422319475, 0.9912968750000001], "iscrowd": 0, "area": 56965.894499999995, "rle": {"size": [640, 457], "counts": "ml01gc08H8I8gCYOo4o0]JGY5a0`JGW5b0hJ@n4h0RKYOU4_1jKcNU4^1jKdNS4]1lKeNR4\\1nKfNP4Z1oKhNlK]O^5m1eNgNkK^O_5k1fNiNiK]O_5k1gNjNiK[O_5m1gNiNiK[O_5l1gNlNhKXO`5m1hNlNgKXO_5m1iNmNgKWO_5l1jNlNhKXO]5m1kNkNhKYO\\5l1kNlNiKYOZ5l1mNjNjKZOX5m1mNjNlKYOV5m1nNjNlKZOT5l1QOiNlK[OR5m1QOiNmK[OQ5l1ROiNmK\\Oo4l1SOhNoK\\Om4m1TOgNoK]Ok4m1UOfNRL]Oh4m1VOfNRL]Og4n1WOeNRL^Oe4m1ZOdNSL_Oa4n1\\OcNSL_Oa4n1[OdNTL_O_4n1]ObNVL_O\\4P2^OaNVL_O[4Q2_O`NWL^OY4S2@^NXL_OW4T2@^NYL^OV4T2B]NZL^OS4V2C\\NZL^OR4W2D[NZL^OQ4X2EYN[L_Oo3Y2FXN[L_On3Z2FXN\\L^Om3[2GVN]L_Oj3]2ITN]L_Oi3]2KTN\\L_Oh3^2LRN]L@f3^2NRN\\L@d3_20RN]L^Oa3`24QN\\L_O]3a28QN[L^OY3d2<nM\\L^OU3e2`0mM\\L^OP3h2e0jM[L^Ok2k2k0gM[L^Of2n2o0dM\\L^Oa2Q3S1bM\\L]O]2S3Y1_M[L^OY2U3\\1]M\\L^OT2X3`1[M\\L]OQ2Z3K_LeLi0`1^Ol1]30iLVL<o1^Oh1^34SMjK2Z2]O]1h3`0bNTNfM[1h3b0bNSNfMZ1h3c0cNSNeMY1h3e0bNSNfMW1j3e0`NTNfMV1k3e0`NUNeMU1l3f0_NUNeMT1m3f0^NWNeMR1n3g0]NWNeMQ1o3h0\\NWNeMP1P4h0[NYNeMn0Q4i0ZNYNeMm0R4j0YNYNeMl0T4i0XN[NdMj0V4k0UN\\NeMh0W4k0UN]NdMg0X4l0TN\\NeMg0X4m0RN]NfMe0Y4m0RN]NfMe0Y4n0QN]NfMd0Z4o0PN\\NgMd0Z4o0oM^NgMb0\\4o0mM^NhMc0[4o0mM]NiMc0[4P1lM]NiMc0[4o0lM^NjMb0[4P1kM^NjMb0[4P1kM]NkMc0Z4o0lM^NkMa0[4P1jM^NlMb0Z4P1jM]NmMc0Z4o0iM^NmMc0[4m0iM_NmMc0\\4m0gM`NmMc0^4k0eMaNoMc0]4j0eMcNnMc0_4h0cMcNPNd0_4h0bMaNQNg0_4f0bM`NPNj0`4c0dM^NnMo0_4b0eM[NoMQ1]4d0fMWNoMU1[4c0jMTNlMY1[4b0kMQNlM]1Y4b0nMlMkMb1W4b0PNiMkMe1U4`0SNgMjMi1S4`0VNbMiMn1Q4?YN_MhMR2P4>[N\\MfMV2o3>]NXMfMZ2m3=`NUMfM]2k3=[1DdN;]1EcN;]1EdN9]1GcN9]1GcN9^1FcN8^1HcN7]1JcN4^1LbN4^1LcN2_1MbN2^1NcN1]1OcN0^11bNN^12cNM]13dNL]13dNL\\14eNLZ14hNJX16iNIW18iNGW19kNEU1;lNDT1<lNDT1<jNGU19iNKU15jNNT13iN2T1NjN6T1JjN;S1EkN?S1AkNd0R1\\OmNg0Q1YOmNl0P1UOmNo0Q1QOmNT1P1lNnNX1P1hNoN[1o0eNoN`1n0`NQOc1m0^NPOg1m0YNQOk1m0UNROn1l0RNROR2l0nMSOU2k0kMSOY2k0gMTO]2XLlJc2h2TOWMZN[5EmJl2a2TOYMYN]5AlJR3^2RO]MWN_5^OjJX3\\2PO]MXNa5ZOiJ^3Z2mN`MnLAS1R6]OgJd3W2kNbMPM@Q1U6YOeJk3U2iNcMQM@n0X6WOdJo3S2hNmMoMf6Y3^KeNPNoMc6\\3^KcNRNmMc6_3\\KaNUNmMa6a3\\K^NVNnMa6c3ZK\\NYNnM_6d3[KZNZNnM_6f3\\KUNYNRN]6g3`KPNWNUN\\6j3bKjMc6S2dIfM_6W2TJVMo5f2h5L5L4K5K6K4K6I6J7H7J7I]gP5"}, "label": "boy with white bear and no hat"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little boy on the left holding a white teddy bear\"."}], "task": "refering", "answer_template": "The \"the little boy on the left holding a white teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 50, 51, 52, 53, 54, 66, 67, 68, 69, 70, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 160, 161, 164, 165, 212, 213, 226, 227, 228, 229, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 288, 289, 290, 291, 292, 293, 294, 304, 308, 309, 310, 323, 324, 325, 326, 336, 337, 338, 339, 340, 341, 342, 352, 353, 354, 355, 356, 357, 358], "bbox": [0.0031291028446389494, 0.12357812500000001, 0.4364551422319475, 0.9912968750000001], "iscrowd": 0, "area": 56965.894499999995, "rle": {"size": [640, 457], "counts": "ml01gc08H8I8gCYOo4o0]JGY5a0`JGW5b0hJ@n4h0RKYOU4_1jKcNU4^1jKdNS4]1lKeNR4\\1nKfNP4Z1oKhNlK]O^5m1eNgNkK^O_5k1fNiNiK]O_5k1gNjNiK[O_5m1gNiNiK[O_5l1gNlNhKXO`5m1hNlNgKXO_5m1iNmNgKWO_5l1jNlNhKXO]5m1kNkNhKYO\\5l1kNlNiKYOZ5l1mNjNjKZOX5m1mNjNlKYOV5m1nNjNlKZOT5l1QOiNlK[OR5m1QOiNmK[OQ5l1ROiNmK\\Oo4l1SOhNoK\\Om4m1TOgNoK]Ok4m1UOfNRL]Oh4m1VOfNRL]Og4n1WOeNRL^Oe4m1ZOdNSL_Oa4n1\\OcNSL_Oa4n1[OdNTL_O_4n1]ObNVL_O\\4P2^OaNVL_O[4Q2_O`NWL^OY4S2@^NXL_OW4T2@^NYL^OV4T2B]NZL^OS4V2C\\NZL^OR4W2D[NZL^OQ4X2EYN[L_Oo3Y2FXN[L_On3Z2FXN\\L^Om3[2GVN]L_Oj3]2ITN]L_Oi3]2KTN\\L_Oh3^2LRN]L@f3^2NRN\\L@d3_20RN]L^Oa3`24QN\\L_O]3a28QN[L^OY3d2<nM\\L^OU3e2`0mM\\L^OP3h2e0jM[L^Ok2k2k0gM[L^Of2n2o0dM\\L^Oa2Q3S1bM\\L]O]2S3Y1_M[L^OY2U3\\1]M\\L^OT2X3`1[M\\L]OQ2Z3K_LeLi0`1^Ol1]30iLVL<o1^Oh1^34SMjK2Z2]O]1h3`0bNTNfM[1h3b0bNSNfMZ1h3c0cNSNeMY1h3e0bNSNfMW1j3e0`NTNfMV1k3e0`NUNeMU1l3f0_NUNeMT1m3f0^NWNeMR1n3g0]NWNeMQ1o3h0\\NWNeMP1P4h0[NYNeMn0Q4i0ZNYNeMm0R4j0YNYNeMl0T4i0XN[NdMj0V4k0UN\\NeMh0W4k0UN]NdMg0X4l0TN\\NeMg0X4m0RN]NfMe0Y4m0RN]NfMe0Y4n0QN]NfMd0Z4o0PN\\NgMd0Z4o0oM^NgMb0\\4o0mM^NhMc0[4o0mM]NiMc0[4P1lM]NiMc0[4o0lM^NjMb0[4P1kM^NjMb0[4P1kM]NkMc0Z4o0lM^NkMa0[4P1jM^NlMb0Z4P1jM]NmMc0Z4o0iM^NmMc0[4m0iM_NmMc0\\4m0gM`NmMc0^4k0eMaNoMc0]4j0eMcNnMc0_4h0cMcNPNd0_4h0bMaNQNg0_4f0bM`NPNj0`4c0dM^NnMo0_4b0eM[NoMQ1]4d0fMWNoMU1[4c0jMTNlMY1[4b0kMQNlM]1Y4b0nMlMkMb1W4b0PNiMkMe1U4`0SNgMjMi1S4`0VNbMiMn1Q4?YN_MhMR2P4>[N\\MfMV2o3>]NXMfMZ2m3=`NUMfM]2k3=[1DdN;]1EcN;]1EdN9]1GcN9]1GcN9^1FcN8^1HcN7]1JcN4^1LbN4^1LcN2_1MbN2^1NcN1]1OcN0^11bNN^12cNM]13dNL]13dNL\\14eNLZ14hNJX16iNIW18iNGW19kNEU1;lNDT1<lNDT1<jNGU19iNKU15jNNT13iN2T1NjN6T1JjN;S1EkN?S1AkNd0R1\\OmNg0Q1YOmNl0P1UOmNo0Q1QOmNT1P1lNnNX1P1hNoN[1o0eNoN`1n0`NQOc1m0^NPOg1m0YNQOk1m0UNROn1l0RNROR2l0nMSOU2k0kMSOY2k0gMTO]2XLlJc2h2TOWMZN[5EmJl2a2TOYMYN]5AlJR3^2RO]MWN_5^OjJX3\\2PO]MXNa5ZOiJ^3Z2mN`MnLAS1R6]OgJd3W2kNbMPM@Q1U6YOeJk3U2iNcMQM@n0X6WOdJo3S2hNmMoMf6Y3^KeNPNoMc6\\3^KcNRNmMc6_3\\KaNUNmMa6a3\\K^NVNnMa6c3ZK\\NYNnM_6d3[KZNZNnM_6f3\\KUNYNRN]6g3`KPNWNUN\\6j3bKjMc6S2dIfM_6W2TJVMo5f2h5L5L4K5K6K4K6I6J7H7J7I]gP5"}, "label": "the little boy on the left holding a white teddy bear"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white stuffed bear\"."}], "task": "refering", "answer_template": "The \"a white stuffed bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 160, 161, 162, 163, 164, 165, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 224, 225, 226, 240, 241, 242, 256, 257, 258], "bbox": [0.002560175054704595, 0.41503125, 0.36323851203501095, 0.7395625], "iscrowd": 0, "area": 20464.048700000003, "rle": {"size": [640, 457], "counts": "UP1d2V>V31O001O001N101O001O01N10000000O100O1O1O1O1000001O00001O2N1O100O000001OO1O100O1JZARKf>n4ZAQKg>o4ZAoJg>P5ZAPKf>P5ZAoJg>Q54O10NPKTAP5k>201O1O001O_NYAkM]>_2eA^M[>c2gAWM\\>j2fATMY>m2gARMX>P3hAoLX>R3iAlLW>U3iAjLW>V3kAhLU>Y3kAfLU>Z3mAeLQ>^3nAaLR>`3oA^LR>b3WBULi=k3XBSLi=m3XBRLh=n3U10000O1O1O001O1O1O1O1O1N2N2N2N2N2M3N2N3M3L4N2M4M3M4M2OO00010O00001O0000000000000000000000000000000000000001O0000000O1010O0010O001O1O1O1O001O1O1O004L2N1K5L4K5L5J3M4M3L4M2O4N1N3N1N3M4L5K8I7Hhae5"}, "label": "a white stuffed bear"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white teddy bear\"."}], "task": "refering", "answer_template": "The \"a white teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 160, 161, 162, 163, 164, 165, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 224, 225, 226, 240, 241, 242, 256, 257, 258], "bbox": [0.002560175054704595, 0.41503125, 0.36323851203501095, 0.7395625], "iscrowd": 0, "area": 20464.048700000003, "rle": {"size": [640, 457], "counts": "UP1d2V>V31O001O001N101O001O01N10000000O100O1O1O1O1000001O00001O2N1O100O000001OO1O100O1JZARKf>n4ZAQKg>o4ZAoJg>P5ZAPKf>P5ZAoJg>Q54O10NPKTAP5k>201O1O001O_NYAkM]>_2eA^M[>c2gAWM\\>j2fATMY>m2gARMX>P3hAoLX>R3iAlLW>U3iAjLW>V3kAhLU>Y3kAfLU>Z3mAeLQ>^3nAaLR>`3oA^LR>b3WBULi=k3XBSLi=m3XBRLh=n3U10000O1O1O001O1O1O1O1O1N2N2N2N2N2M3N2N3M3L4N2M4M3M4M2OO00010O00001O0000000000000000000000000000000000000001O0000000O1010O0010O001O1O1O1O001O1O1O004L2N1K5L4K5L5J3M4M3L4M2O4N1N3N1N3M4L5K8I7Hhae5"}, "label": "a white teddy bear"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male soldier holding two young boys in each arm\"."}], "task": "refering", "answer_template": "The \"a male soldier holding two young boys in each arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 55, 56, 57, 58, 70, 71, 72, 73, 74, 86, 87, 88, 89, 90, 103, 104, 105, 106, 119, 120, 121, 122, 134, 135, 136, 137, 138, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 181, 182, 183, 184, 197, 198, 199, 213, 214, 215, 229, 230, 231, 245, 246, 247, 262, 263, 264, 278, 279, 280, 294, 295, 296, 297, 310, 311, 312, 313, 327, 328, 329, 343, 344, 345, 358, 359, 360, 361], "bbox": [0.330109409190372, 0.11684375, 0.6754266958424509, 1.0], "iscrowd": 0, "area": 45368.0064, "rle": {"size": [640, 457], "counts": "Xen23lc03L4M3L4M3L3N3_^OCY?a0R@7h?LU@;c?IZ@=`?F^@?[?Eb@a0X?Be@c0U?Ah@d0S?_Ok@d0Q?_Ol@f0n>^Oo@g0l>\\OQAi0i>[OUAh0g>[OVAj0d>^2K5J5L5J6K5J6K4K6L4K5L3L5L4K5L3M4K5L4K5L3M4dLdGkJa8m4kGkJY8m4RHkJR8m4ZHlJj7l4`HmJe7MnGh1e0WNg7DgGQ2k0VNh7\\O`GZ2P1UNk7TOXGc2V1TNk7nNQGk2[1RNo7fNiFW3]1oM_9o1hFkMY9T2lFhMT9V2SGcMo8\\2VG_Mk8_2\\GZMf8e2_GUMc8i2[4N2O1N101N2O1N2O10jFfMf0Z2WOWN;i1CgNOY1NYOCg0;IWO6h0:jNFS1k0_NUO_1o0_NQO^1T1`NlN^1W1aNiN\\1\\1bNcN]1`1bN`N[1e1cN[N[1h1dNXNZ1l1dNTNY1P2fNoMY1U2eNkMY1X2fNhMY1[2eNeM[1\\2dNdM[1_2cNaM]1`2bN_M^1d2`N\\M`1e2_N[M`1g2_NYMa1i2\\NXMc1j2\\NUMe1m2YNSMf1o2YNQMg1Q3WNoLi1S3UNmLj1V3TNjLl1Y3QNfLo1]3oMcLQ2_3mMaLR2b3lM^LT2e3iM[LV2h3hMWLY2k3eMULZ2o3cMQL]2Q4aMoK^2T4`MlK`2V4^MjKb2Y4[MfKe2]4YMcKg2_4WMaKh2b4VM^Kj2e4SM[Kl2h4YKkHJ\\2n4k4oJSI0R2P5n4gJ[I5g1T5Q5\\JcI=\\1hNdMW6`7[JjIc0Y1R5o4RJRJi0o0U5Q5hI[JP1d0W5Z7fJfHZ5^7aJdH]5`7`J`H`5c7]J]Hb5g7[JYHe5j7XJWHf5m7WJSHi5Q8SJoGl5U8QJkGo5X8nIhGR6[8kIfGS6T;00O10001PMXBEl=8hBTOZ=k0mBmNU=Q1RCgNQ=W1VCbNm<[1YC_Ni<_1]CZNf<d1_CWNd<f1aCUNa<g1gCRN\\<k1jCPNY<k1nCoMU<n1Q3M4K4M3\\Oe0POo\\l2"}, "label": "a male soldier holding two young boys in each arm"}]}
{"id": 346950, "image": "COCO_train2014_000000346950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with short brown hair dressed in fatigues between two children\"."}], "task": "refering", "answer_template": "The \"a man with short brown hair dressed in fatigues between two children\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 55, 56, 57, 58, 70, 71, 72, 73, 74, 86, 87, 88, 89, 90, 103, 104, 105, 106, 119, 120, 121, 122, 134, 135, 136, 137, 138, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 181, 182, 183, 184, 197, 198, 199, 213, 214, 215, 229, 230, 231, 245, 246, 247, 262, 263, 264, 278, 279, 280, 294, 295, 296, 297, 310, 311, 312, 313, 327, 328, 329, 343, 344, 345, 358, 359, 360, 361], "bbox": [0.330109409190372, 0.11684375, 0.6754266958424509, 1.0], "iscrowd": 0, "area": 45368.0064, "rle": {"size": [640, 457], "counts": "Xen23lc03L4M3L4M3L3N3_^OCY?a0R@7h?LU@;c?IZ@=`?F^@?[?Eb@a0X?Be@c0U?Ah@d0S?_Ok@d0Q?_Ol@f0n>^Oo@g0l>\\OQAi0i>[OUAh0g>[OVAj0d>^2K5J5L5J6K5J6K4K6L4K5L3L5L4K5L3M4K5L4K5L3M4dLdGkJa8m4kGkJY8m4RHkJR8m4ZHlJj7l4`HmJe7MnGh1e0WNg7DgGQ2k0VNh7\\O`GZ2P1UNk7TOXGc2V1TNk7nNQGk2[1RNo7fNiFW3]1oM_9o1hFkMY9T2lFhMT9V2SGcMo8\\2VG_Mk8_2\\GZMf8e2_GUMc8i2[4N2O1N101N2O1N2O10jFfMf0Z2WOWN;i1CgNOY1NYOCg0;IWO6h0:jNFS1k0_NUO_1o0_NQO^1T1`NlN^1W1aNiN\\1\\1bNcN]1`1bN`N[1e1cN[N[1h1dNXNZ1l1dNTNY1P2fNoMY1U2eNkMY1X2fNhMY1[2eNeM[1\\2dNdM[1_2cNaM]1`2bN_M^1d2`N\\M`1e2_N[M`1g2_NYMa1i2\\NXMc1j2\\NUMe1m2YNSMf1o2YNQMg1Q3WNoLi1S3UNmLj1V3TNjLl1Y3QNfLo1]3oMcLQ2_3mMaLR2b3lM^LT2e3iM[LV2h3hMWLY2k3eMULZ2o3cMQL]2Q4aMoK^2T4`MlK`2V4^MjKb2Y4[MfKe2]4YMcKg2_4WMaKh2b4VM^Kj2e4SM[Kl2h4YKkHJ\\2n4k4oJSI0R2P5n4gJ[I5g1T5Q5\\JcI=\\1hNdMW6`7[JjIc0Y1R5o4RJRJi0o0U5Q5hI[JP1d0W5Z7fJfHZ5^7aJdH]5`7`J`H`5c7]J]Hb5g7[JYHe5j7XJWHf5m7WJSHi5Q8SJoGl5U8QJkGo5X8nIhGR6[8kIfGS6T;00O10001PMXBEl=8hBTOZ=k0mBmNU=Q1RCgNQ=W1VCbNm<[1YC_Ni<_1]CZNf<d1_CWNd<f1aCUNa<g1gCRN\\<k1jCPNY<k1nCoMU<n1Q3M4K4M3\\Oe0POo\\l2"}, "label": "a man with short brown hair dressed in fatigues between two children"}]}
{"id": 469832, "image": "COCO_train2014_000000469832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown sofa in the hall\"."}], "task": "refering", "answer_template": "The \"brown sofa in the hall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0016093750000000001, 0.5505416666666666, 0.380640625, 0.9827916666666667], "iscrowd": 0, "area": 36936.718949999995, "rle": {"size": [480, 640], "counts": "Xg0`6`80001O00001O00001N10001O00001O0000001O00001O00001O00001O00001O0O100000000000000O100000000000000O100000000000000O100000000000000O100000001O00000O10001O001O001O0O2O001O001O001O0O2O001O001O001N2O001O001O001O0O2O001O3M0000000O10001O000000000O100000000000000O100000000000000O10001O000000000O100000000000000O100000000000SFSKg9m4WFUKi9j4VFXKj9h4UFZKj9n410O00001O001O00001O0000L4M3L4L4O100O1O1O1O1O1O101SOhEoLY:n2QFjLP:S3ZFeLg9X3R1N2M3M3M3M3M3N2M3M3M3M3N2M3M3M3000001O00001O000O101N100O2O0O1O2O0O101N100O2N100O101N100O2N100O2O8G8I8G8H9HUYi5"}, "label": "brown sofa in the hall"}]}
{"id": 469832, "image": "COCO_train2014_000000469832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown couch in front of the girl who is playing video games\"."}], "task": "refering", "answer_template": "The \"a light brown couch in front of the girl who is playing video games\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0016093750000000001, 0.5505416666666666, 0.380640625, 0.9827916666666667], "iscrowd": 0, "area": 36936.718949999995, "rle": {"size": [480, 640], "counts": "Xg0`6`80001O00001O00001N10001O00001O0000001O00001O00001O00001O00001O0O100000000000000O100000000000000O100000000000000O100000000000000O100000001O00000O10001O001O001O0O2O001O001O001O0O2O001O001O001N2O001O001O001O0O2O001O3M0000000O10001O000000000O100000000000000O100000000000000O10001O000000000O100000000000000O100000000000SFSKg9m4WFUKi9j4VFXKj9h4UFZKj9n410O00001O001O00001O0000L4M3L4L4O100O1O1O1O1O1O101SOhEoLY:n2QFjLP:S3ZFeLg9X3R1N2M3M3M3M3M3N2M3M3M3M3N2M3M3M3000001O00001O000O101N100O2O0O1O2O0O101N100O2N100O101N100O2N100O2O8G8I8G8H9HUYi5"}, "label": "a light brown couch in front of the girl who is playing video games"}]}
{"id": 469832, "image": "COCO_train2014_000000469832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white sofa\"."}], "task": "refering", "answer_template": "The \"the white sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 273, 274, 275, 292, 295, 296, 297, 298, 314, 315, 317, 318, 319, 320, 321, 337, 338, 340, 341, 342, 343, 344, 361, 363, 364, 365, 366, 367, 389, 390], "bbox": [0.689734375, 0.5844583333333334, 1.0, 0.9580625], "iscrowd": 0, "area": 16668.770450000004, "rle": {"size": [480, 640], "counts": "`S_62^>i0\\Oc0_O?B8J6M3L4O14L4L5L3L3M4L4L4L3M4L5K7I6J6J6J5Kghi0KSWVO`0F:G9G9G8H4J7I6O101O001O001N101O001O001N101O001N100O101N100O100O10001N1O1M3M3M3M3M3J7I6N2mNeL`F\\3`9oLSFS3m9P100O2O0000001O0000000O2O0000001O00000O100000001O0000000O10000000001O00000O100000001O00000O2O0000001O0000000O2O0000001O0000001NJ7YOg0YOg0YOg0YOg0YOg0YOg0YOnG"}, "label": "the white sofa"}]}
{"id": 469832, "image": "COCO_train2014_000000469832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white couch with yellow cushions against wall\"."}], "task": "refering", "answer_template": "The \"white couch with yellow cushions against wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 273, 274, 275, 292, 295, 296, 297, 298, 314, 315, 317, 318, 319, 320, 321, 337, 338, 340, 341, 342, 343, 344, 361, 363, 364, 365, 366, 367, 389, 390], "bbox": [0.689734375, 0.5844583333333334, 1.0, 0.9580625], "iscrowd": 0, "area": 16668.770450000004, "rle": {"size": [480, 640], "counts": "`S_62^>i0\\Oc0_O?B8J6M3L4O14L4L5L3L3M4L4L4L3M4L5K7I6J6J6J5Kghi0KSWVO`0F:G9G9G8H4J7I6O101O001O001N101O001O001N101O001N100O101N100O100O10001N1O1M3M3M3M3M3J7I6N2mNeL`F\\3`9oLSFS3m9P100O2O0000001O0000000O2O0000001O00000O100000001O0000000O10000000001O00000O100000001O00000O2O0000001O0000000O2O0000001O0000001NJ7YOg0YOg0YOg0YOg0YOg0YOg0YOnG"}, "label": "white couch with yellow cushions against wall"}]}
{"id": 191305, "image": "COCO_train2014_000000191305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elaphant with the trunk\"."}], "task": "refering", "answer_template": "The \"the elaphant with the trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.20634375, 0.0033878504672897195, 0.8976875, 0.9898598130841121], "iscrowd": 0, "area": 155804.68310000002, "rle": {"size": [428, 640], "counts": "nYg16o<:[F0`5;SJg0n4CfJ]1\\4nNVKT2j3WNjKi2X3aM[L`3f2kLnLV4U2RL^M_4X2hK[Ma4a2W3L4L4L4M3L4L4L4L4N2M3M3M3N3L3M3N2M3M3N200O100O1O100O100O1O100O100O1O2O0O100O1O100O100O1O100O100O10000000000000001O00J6I7I7J6I7J6J6I7J6J6I7J6J7I6I7L4N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O100000000001O0000002N3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000cNMaE3a:4VELk:=kDCW;d0`D\\Ob;l0TDTOm;U12N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O0N3L4K4M4K5L3L5K5L3L5L4K5K4M4K5L3L6K4K6J6K4K6K4K6K5J5K6K4K6K5K4L5K4L5K5K4L5L3L5K5K4L5K4L5K5K4L5K4L5L4K4L5K4M4K5K4M4K5K4M4K4L5L4K4L5K4M<C>B>C=BToj0"}, "label": "the elaphant with the trunk"}]}
{"id": 191305, "image": "COCO_train2014_000000191305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on the left\"."}], "task": "refering", "answer_template": "The \"elephant on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0, 0.13707943925233645, 0.226921875, 0.9887616822429907], "iscrowd": 0, "area": 46933.191300000006, "rle": {"size": [428, 640], "counts": "i7\\5c2]50000000000000000000O100000000000000000000000000000000O100K5G9H8H8G9H8H8H8N200O1O10O01O1001O6J5K6J6J6J6J5K6J6J6J6J6J00000000000000000000001O00000000000000000000000000000O100O100O100O100O1O100O100O100O100O1O100O10001N10000000000000000O10000000000000000O1000000000UOWNSFi1m9`NjE`1R2XN^5b0WHV1T2cN`1;\\Oo1mNQNY1R1QOm0@[NR1j1eNL2eNk0a2ZNkNd0nNe0W6TOTJ>V6[OTJ8V6AUJ2T6FWJMS6Lcn^6"}, "label": "elephant on the left"}]}
{"id": 191305, "image": "COCO_train2014_000000191305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is an elephant on the left of the other elephant\"."}], "task": "refering", "answer_template": "The \"there is an elephant on the left of the other elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0, 0.13707943925233645, 0.226921875, 0.9887616822429907], "iscrowd": 0, "area": 46933.191300000006, "rle": {"size": [428, 640], "counts": "i7\\5c2]50000000000000000000O100000000000000000000000000000000O100K5G9H8H8G9H8H8H8N200O1O10O01O1001O6J5K6J6J6J6J5K6J6J6J6J6J00000000000000000000001O00000000000000000000000000000O100O100O100O100O1O100O100O100O100O1O100O10001N10000000000000000O10000000000000000O1000000000UOWNSFi1m9`NjE`1R2XN^5b0WHV1T2cN`1;\\Oo1mNQNY1R1QOm0@[NR1j1eNL2eNk0a2ZNkNd0nNe0W6TOTJ>V6[OTJ8V6AUJ2T6FWJMS6Lcn^6"}, "label": "there is an elephant on the left of the other elephant"}]}
{"id": 330572, "image": "COCO_train2014_000000330572.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 280, 281, 303, 304, 326], "bbox": [0.136296875, 0.3929166666666667, 0.376890625, 1.0], "iscrowd": 0, "area": 13437.49695, "rle": {"size": [408, 640], "counts": "_mR13[<;E;L4M4K4M3L:G3L4M3L4M3L4M3O1O1QFeM^9]2\\FnM^9S2\\FXN]9e2M2N3L3N2oH^LQN1]6d3ZKjLf4X3QKoLo4S3iJSMV5P3bJVM^5m2XJZMh5h2PJ^Mo5e2iIaMW6`4O10000O100O10000O01K4M4L3O_O[JRJe5n5]JQJb5P6_JnI`5P6dJoI\\5m5i0M201000O2O001N2O0XNnIjLT6V3nIgLS6X3PJeLP6\\3RJ`LP6_3SJ^Ln5b3TJZLn5e3UJTLP6l3RJmKS6R4PJgKU6Y4mI_KZ6_4iIZK\\6f4k0O2N101gLlGk1U8SNlGn1U8oMmGP2U8mMlGROMk2W8PNoGSOLb11@W8GnGVOLY1:Hm7FnGYOLS1g0F`7@iGI?Q19Da7ZOoGN:T16Cm8>SG@n8a0SG]On8b0SG\\Oo8d0QG[OP9e0PGYOR9g0nFXOS9h0nFVOS9j0mFTOV9k0jFTOW9l0iFROY9n0gFQO[9n0fFoN]9P1cFoN`9o0U100O5L0000O1000000000000000O1O2N1O1O1N2O1O1O1O1O1O1N2O1O1O001O1O1NRV5KTjJ1O1O1O2N1O1O1001O1N1M_fn4"}, "label": "a man wearing a white shirt"}]}
{"id": 330572, "image": "COCO_train2014_000000330572.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt with two women behind him on a moped\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt with two women behind him on a moped\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 280, 281, 303, 304, 326], "bbox": [0.136296875, 0.3929166666666667, 0.376890625, 1.0], "iscrowd": 0, "area": 13437.49695, "rle": {"size": [408, 640], "counts": "_mR13[<;E;L4M4K4M3L:G3L4M3L4M3L4M3O1O1QFeM^9]2\\FnM^9S2\\FXN]9e2M2N3L3N2oH^LQN1]6d3ZKjLf4X3QKoLo4S3iJSMV5P3bJVM^5m2XJZMh5h2PJ^Mo5e2iIaMW6`4O10000O100O10000O01K4M4L3O_O[JRJe5n5]JQJb5P6_JnI`5P6dJoI\\5m5i0M201000O2O001N2O0XNnIjLT6V3nIgLS6X3PJeLP6\\3RJ`LP6_3SJ^Ln5b3TJZLn5e3UJTLP6l3RJmKS6R4PJgKU6Y4mI_KZ6_4iIZK\\6f4k0O2N101gLlGk1U8SNlGn1U8oMmGP2U8mMlGROMk2W8PNoGSOLb11@W8GnGVOLY1:Hm7FnGYOLS1g0F`7@iGI?Q19Da7ZOoGN:T16Cm8>SG@n8a0SG]On8b0SG\\Oo8d0QG[OP9e0PGYOR9g0nFXOS9h0nFVOS9j0mFTOV9k0jFTOW9l0iFROY9n0gFQO[9n0fFoN]9P1cFoN`9o0U100O5L0000O1000000000000000O1O2N1O1O1N2O1O1O1O1O1O1N2O1O1O001O1O1NRV5KTjJ1O1O1O2N1O1O1001O1N1M_fn4"}, "label": "a man in a white shirt with two women behind him on a moped"}]}
{"id": 84821, "image": "COCO_train2014_000000084821.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suv\"."}], "task": "refering", "answer_template": "The \"a black suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 181, 182, 183], "bbox": [0.726703125, 0.049413145539906105, 0.99609375, 0.4988262910798122], "iscrowd": 0, "area": 18006.33395, "rle": {"size": [426, 640], "counts": "SiR61Y=001O000O1000000O1000000O100000001N1000000O100000000O10001O0O2O001O001N2O2N2N3L3N2N2N2N2M3N3M2N2M2O2N1O2N2M2O2M3M2N3M3M2N2N2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N2N2N2N3M2N2O1N3M2N2aGdLm6^3kHjLS7Y3dHnL[7T3]HSMb7o2VHXMi7U4O1O1O1O1O1O10000O100000000O2O0000000O10000000000O100000000O101O00000O100000000O10000000000O10001O000O100000000O100000000O10000000001N10O100O10OH9ZMf2ZMel0"}, "label": "a black suv"}]}
{"id": 84821, "image": "COCO_train2014_000000084821.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car behind a fire hydrant\"."}], "task": "refering", "answer_template": "The \"a car behind a fire hydrant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 181, 182, 183], "bbox": [0.726703125, 0.049413145539906105, 0.99609375, 0.4988262910798122], "iscrowd": 0, "area": 18006.33395, "rle": {"size": [426, 640], "counts": "SiR61Y=001O000O1000000O1000000O100000001N1000000O100000000O10001O0O2O001O001N2O2N2N3L3N2N2N2N2M3N3M2N2M2O2N1O2N2M2O2M3M2N3M3M2N2N2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N2N2N2N3M2N2O1N3M2N2aGdLm6^3kHjLS7Y3dHnL[7T3]HSMb7o2VHXMi7U4O1O1O1O1O1O10000O100000000O2O0000000O10000000000O100000000O101O00000O100000000O10000000000O10001O000O100000000O100000000O10000000001N10O100O10OH9ZMf2ZMel0"}, "label": "a car behind a fire hydrant"}]}
{"id": 207715, "image": "COCO_train2014_000000207715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark brown horse with black hair in the middle\"."}], "task": "refering", "answer_template": "The \"the dark brown horse with black hair in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 241, 242, 243], "bbox": [0.427265625, 0.339625, 0.6423593750000001, 0.6278333333333334], "iscrowd": 0, "area": 9816.33115, "rle": {"size": [480, 640], "counts": "fTP42h>n0UO`0C?@c0^O4K4L=D5J5L0O2N2O0O2OTOYE_Mf:X2V1M2O2N2N2M2O2N2N2N200000000000000000001O0000010O0010O0jDSNY9m1cFWN^9i1]F\\Nb9d1ZFaNe9_1XFdNh9]1UFfNj9Z1TFhNm9X1oElNP:W1gEPOX:Z1SEQOm:_22M3M3M3YOg0N1N3M3K5L4M2O2N2N20000O1000001O00000012M3M3M4L9G:G6I6J8H8H2N1O2N1O00O1^Nb1C=J6M3M3N3N100N2D=O0O100O2O0O2O0O2O0;E2O1N100O1O1O2N1O200O100O1O0O100O0O2NkNkBa0T=]OWC;h<D^C7b<DjC5W<GSD2n;Ld1M`U[3"}, "label": "the dark brown horse with black hair in the middle"}]}
{"id": 207715, "image": "COCO_train2014_000000207715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the middle of the three inside the fence\"."}], "task": "refering", "answer_template": "The \"the horse in the middle of the three inside the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 241, 242, 243], "bbox": [0.427265625, 0.339625, 0.6423593750000001, 0.6278333333333334], "iscrowd": 0, "area": 9816.33115, "rle": {"size": [480, 640], "counts": "fTP42h>n0UO`0C?@c0^O4K4L=D5J5L0O2N2O0O2OTOYE_Mf:X2V1M2O2N2N2M2O2N2N2N200000000000000000001O0000010O0010O0jDSNY9m1cFWN^9i1]F\\Nb9d1ZFaNe9_1XFdNh9]1UFfNj9Z1TFhNm9X1oElNP:W1gEPOX:Z1SEQOm:_22M3M3M3YOg0N1N3M3K5L4M2O2N2N20000O1000001O00000012M3M3M4L9G:G6I6J8H8H2N1O2N1O00O1^Nb1C=J6M3M3N3N100N2D=O0O100O2O0O2O0O2O0;E2O1N100O1O1O2N1O200O100O1O0O100O0O2NkNkBa0T=]OWC;h<D^C7b<DjC5W<GSD2n;Ld1M`U[3"}, "label": "the horse in the middle of the three inside the fence"}]}
{"id": 27495, "image": "COCO_train2014_000000027495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the braying mouth of a horse\"."}], "task": "refering", "answer_template": "The \"the braying mouth of a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.286765625, 0.0016159250585480091, 1.0, 0.9886651053864169], "iscrowd": 0, "area": 136520.15495000005, "rle": {"size": [427, 640], "counts": "gi\\2>i<5J7J5K6B=F;M2N3M2N3M2N2N3M3M3M3M3M3N2M3N2N2N4K5L5K4K6K4L5K4K4M2N2M3N2N2N0O2O1O1N101O1O0O2O1N1O2N2N2N1O2O0O001O1O1O1O1O001O1O1O1iK[Jh0e5VO^Jh0b5VOcJg0^5WOeJg0\\5VOhJi0X5TOmJj0S5TOPKk0P5ROTKm0l4QOXKm0h4PO\\Ko0c4mNdKQ1\\4^NVLa1j3YN]Lf1c3UNcLj1]3SNgLl1Y3RNjLm1V3PNnLo1R3nMRMQ2n2mMVMQ2i2mMZMS2f2jM^MU2b2iMaMV2_2gMeMX2\\2dMhM[2Y2bMjM]2W2_MlMa2U2[MoMd2Q2ZMRNe2o1YMSNf2m1XMUNh2l1VMUNj2k1TMWNl2j1RMWNn2i1PMZNo2Q6000000000000000000000O10000000000000000000000000000O100000000000000000000000000O1000000000000O10000O10000O1000000O10000O1000000O10000O10000O100O100O10000O100O100O100O100O100O1000000000000000000000000000000000000000000000000000G]FjLc9S3`FnL_9n2eFRM[9j2iFVMW9f2mFZMS9W2\\GiMd8U2^GkMb8R2aGnM_8P2cGPN]8m1fGSNZ8k1hGUNX8i1jGWNV8g1lGYNT8e1nG[NR8c1PH]NP8a1RH_Nn7_1THaNm7\\1UHdNk7Z1WHfNi7X1YHhNg7V1[HjNf7S1\\HmNd7Q1^HoNb7o0`HQOa7l0aHTOa7h0aHXOa7b0cH^O^7=fHC\\77hHIY72kHNY:000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O4L3M4L3M4L3L5L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M3M4L3M4L3M4L3M2N1O001QOcKnH^4o6dKQI]4k6fKUIZ4h6iKXIX4d6kK\\IV4g5bKSJ:5U4b5gKXJ55T4^5nK[J05S4]5PL]JN5S4[5SL^JK6R4[5VL]JJ6Q4Z5XL_JH6Q4Y5]MfJd2Y5]MfJc2Z5_MdJb2[5_MdJb2[5_MdJa2P4WKkK0h0Y2\\Oa2o3WKlK2i0V2ZOb2o3WKnK3h0T2ZOc2m3YKPL2i0Q2YOd2m3ZKQL3h0n1ZOf2k3ZKSL4h0k1YOh2i3[KVL4g0i1YOi2h3[KXL3i0h1VOj2h3\\KYL2j0h1TOk2f3]K\\L0l0g1QOm2e3]K^LOm0i1mNl2g3\\K^L0P1l6_2UIaLOQ1l6\\2VIcLNS1k6Y2WIdLNT1k6V2XIfLMV1j6S2ZIfLLY1i6P2[IgLL[1h6m1\\IhLL]1h6j1[IiLM_1g6g1\\IiLNa1f6e1\\IjLMd1f6b1\\IiLOg1d6_1]IjLOh1d6]1]IjL0k1c6Z1\\IkL1m1b6W1PJjNP6U1PJmNo5S1PJoNo5P1QJQOo5n0QJTOn5l0QJVOo5h0QJZOn5e0RJ]Om5c0RJ_Om5`0SJAm5>SJDl5<SJEn59QJJn55RJLn54SJLl53VJLj53XJMg53ZJLP6IRJ1Q7POPIh0W8YNjG4[1"}, "label": "the braying mouth of a horse"}]}
{"id": 27495, "image": "COCO_train2014_000000027495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown horse with its mouth open wide\"."}], "task": "refering", "answer_template": "The \"a large brown horse with its mouth open wide\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.286765625, 0.0016159250585480091, 1.0, 0.9886651053864169], "iscrowd": 0, "area": 136520.15495000005, "rle": {"size": [427, 640], "counts": "gi\\2>i<5J7J5K6B=F;M2N3M2N3M2N2N3M3M3M3M3M3N2M3N2N2N4K5L5K4K6K4L5K4K4M2N2M3N2N2N0O2O1O1N101O1O0O2O1N1O2N2N2N1O2O0O001O1O1O1O1O001O1O1O1iK[Jh0e5VO^Jh0b5VOcJg0^5WOeJg0\\5VOhJi0X5TOmJj0S5TOPKk0P5ROTKm0l4QOXKm0h4PO\\Ko0c4mNdKQ1\\4^NVLa1j3YN]Lf1c3UNcLj1]3SNgLl1Y3RNjLm1V3PNnLo1R3nMRMQ2n2mMVMQ2i2mMZMS2f2jM^MU2b2iMaMV2_2gMeMX2\\2dMhM[2Y2bMjM]2W2_MlMa2U2[MoMd2Q2ZMRNe2o1YMSNf2m1XMUNh2l1VMUNj2k1TMWNl2j1RMWNn2i1PMZNo2Q6000000000000000000000O10000000000000000000000000000O100000000000000000000000000O1000000000000O10000O10000O1000000O10000O1000000O10000O10000O100O100O10000O100O100O100O100O100O1000000000000000000000000000000000000000000000000000G]FjLc9S3`FnL_9n2eFRM[9j2iFVMW9f2mFZMS9W2\\GiMd8U2^GkMb8R2aGnM_8P2cGPN]8m1fGSNZ8k1hGUNX8i1jGWNV8g1lGYNT8e1nG[NR8c1PH]NP8a1RH_Nn7_1THaNm7\\1UHdNk7Z1WHfNi7X1YHhNg7V1[HjNf7S1\\HmNd7Q1^HoNb7o0`HQOa7l0aHTOa7h0aHXOa7b0cH^O^7=fHC\\77hHIY72kHNY:000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O4L3M4L3M4L3L5L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M3M4L3M4L3M4L3M2N1O001QOcKnH^4o6dKQI]4k6fKUIZ4h6iKXIX4d6kK\\IV4g5bKSJ:5U4b5gKXJ55T4^5nK[J05S4]5PL]JN5S4[5SL^JK6R4[5VL]JJ6Q4Z5XL_JH6Q4Y5]MfJd2Y5]MfJc2Z5_MdJb2[5_MdJb2[5_MdJa2P4WKkK0h0Y2\\Oa2o3WKlK2i0V2ZOb2o3WKnK3h0T2ZOc2m3YKPL2i0Q2YOd2m3ZKQL3h0n1ZOf2k3ZKSL4h0k1YOh2i3[KVL4g0i1YOi2h3[KXL3i0h1VOj2h3\\KYL2j0h1TOk2f3]K\\L0l0g1QOm2e3]K^LOm0i1mNl2g3\\K^L0P1l6_2UIaLOQ1l6\\2VIcLNS1k6Y2WIdLNT1k6V2XIfLMV1j6S2ZIfLLY1i6P2[IgLL[1h6m1\\IhLL]1h6j1[IiLM_1g6g1\\IiLNa1f6e1\\IjLMd1f6b1\\IiLOg1d6_1]IjLOh1d6]1]IjL0k1c6Z1\\IkL1m1b6W1PJjNP6U1PJmNo5S1PJoNo5P1QJQOo5n0QJTOn5l0QJVOo5h0QJZOn5e0RJ]Om5c0RJ_Om5`0SJAm5>SJDl5<SJEn59QJJn55RJLn54SJLl53VJLj53XJMg53ZJLP6IRJ1Q7POPIh0W8YNjG4[1"}, "label": "a large brown horse with its mouth open wide"}]}
{"id": 27495, "image": "COCO_train2014_000000027495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man who is on the yelling horse\"."}], "task": "refering", "answer_template": "The \"the man who is on the yelling horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 112, 113, 114, 135, 136, 137, 159, 160, 183], "bbox": [0.8310000000000001, 0.00477751756440281, 1.0, 0.4891569086651054], "iscrowd": 0, "area": 14998.378149999995, "rle": {"size": [427, 640], "counts": "Pkm68S=;E<D5K3M3M4L3M3M3M3M3M3M3M3M3M4L3M2N3M3M3M3M2N3M3M3L4M2N3M3M3M3M2N2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1ObM^2ZMd2"}, "label": "the man who is on the yelling horse"}]}
{"id": 27495, "image": "COCO_train2014_000000027495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in green sweater\"."}], "task": "refering", "answer_template": "The \"a man in green sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 112, 113, 114, 135, 136, 137, 159, 160, 183], "bbox": [0.8310000000000001, 0.00477751756440281, 1.0, 0.4891569086651054], "iscrowd": 0, "area": 14998.378149999995, "rle": {"size": [427, 640], "counts": "Pkm68S=;E<D5K3M3M4L3M3M3M3M3M3M3M3M3M4L3M2N3M3M3M3M2N3M3M3L4M2N3M3M3M3M2N2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1ObM^2ZMd2"}, "label": "a man in green sweater"}]}
{"id": 27495, "image": "COCO_train2014_000000027495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with a police saddle on\"."}], "task": "refering", "answer_template": "The \"a horse with a police saddle on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0015, 0.33932084309133487, 0.5472499999999999, 1.0], "iscrowd": 0, "area": 75232.32700000002, "rle": {"size": [427, 640], "counts": "^b0X8R5100O10000O100O100O100O100O100O100O10000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000000000000000O100000000000000000000000000000000001O0000001O00001O0000001O00001OO10000O10000O100002N3M2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N0000000000O10iLVJmNj5n0\\JQOd5j0bJUO^5f0hJYOX5c0mJ\\OS5?SK@m4;YKDg47_KHa42fKMZ4KnK5Q4EWL:i3^O`La0`3XOhLg0X3ROPMm0P3kNXMU1h2dN`M[1`2^NhMa1X2^NjMa1V2^NlMa1T2]NoMb1Q2]NQNb1n1^NTNa1l1]NXNa1h1^NZNa1g1]N[Nb1h1YN[Nf1h1VNZNi1j1RNXNm1k1nMWNR2l1jMVNU2m1gMTNY2_6000001O00000000001O00001O00VOiMaFV2_9kMaFT2_9nM`FQ2`9PN_FP2a9RN^Fm1b9TN^Fk1b9WN\\Fi1d9XN\\Fg1d9ZN\\Fe1d9\\N\\Fc1d9^N[Fb1e9_N[F`1e9`N\\F_1d9bN[F^1e9cN[F\\1e9eN[FZ1e9gN[FX1e9iNZFW1f9jNZFU1f9lNZFS1e9oNZFQ1f9POZFo0f9ROZFm0f9SO[Fl0e9UOZFk0f9VOZFi0f9XOZFg0f9ZOYFf0g9[OYFd0g9]OYFb0g9_OYF`0g9AXF?h9AYF>g9CYF<g9EXF;h9FXF9h9HWF8i9IWF6i9KVF5j9LVF3j9MWF2i9OVF1j90VFOi93VFMj94VFKj96UFJk97UFHk99TFGl99UFFk9;TFEl9<TFCl9=TFCl9>TFAl9?UF@k9`0UF@k9a0UF^Ok9b0UF^Ok9c0UF\\Ok9d0UFZOm9g0SFVOo9j0QFSOR:n0nEoNT:Q1lEmNU:T1lEhNW:Y1iEdNY:\\1gEaN\\:`1dE]N^:c1bE[N`:f1`EVNf:g1a1nN[bh3"}, "label": "a horse with a police saddle on"}]}
{"id": 27495, "image": "COCO_train2014_000000027495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown horse standing behind the horse whose mouth is wide opened\"."}], "task": "refering", "answer_template": "The \"a light brown horse standing behind the horse whose mouth is wide opened\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0015, 0.33932084309133487, 0.5472499999999999, 1.0], "iscrowd": 0, "area": 75232.32700000002, "rle": {"size": [427, 640], "counts": "^b0X8R5100O10000O100O100O100O100O100O100O10000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000000000000000O100000000000000000000000000000000001O0000001O00001O0000001O00001OO10000O10000O100002N3M2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N0000000000O10iLVJmNj5n0\\JQOd5j0bJUO^5f0hJYOX5c0mJ\\OS5?SK@m4;YKDg47_KHa42fKMZ4KnK5Q4EWL:i3^O`La0`3XOhLg0X3ROPMm0P3kNXMU1h2dN`M[1`2^NhMa1X2^NjMa1V2^NlMa1T2]NoMb1Q2]NQNb1n1^NTNa1l1]NXNa1h1^NZNa1g1]N[Nb1h1YN[Nf1h1VNZNi1j1RNXNm1k1nMWNR2l1jMVNU2m1gMTNY2_6000001O00000000001O00001O00VOiMaFV2_9kMaFT2_9nM`FQ2`9PN_FP2a9RN^Fm1b9TN^Fk1b9WN\\Fi1d9XN\\Fg1d9ZN\\Fe1d9\\N\\Fc1d9^N[Fb1e9_N[F`1e9`N\\F_1d9bN[F^1e9cN[F\\1e9eN[FZ1e9gN[FX1e9iNZFW1f9jNZFU1f9lNZFS1e9oNZFQ1f9POZFo0f9ROZFm0f9SO[Fl0e9UOZFk0f9VOZFi0f9XOZFg0f9ZOYFf0g9[OYFd0g9]OYFb0g9_OYF`0g9AXF?h9AYF>g9CYF<g9EXF;h9FXF9h9HWF8i9IWF6i9KVF5j9LVF3j9MWF2i9OVF1j90VFOi93VFMj94VFKj96UFJk97UFHk99TFGl99UFFk9;TFEl9<TFCl9=TFCl9>TFAl9?UF@k9`0UF@k9a0UF^Ok9b0UF^Ok9c0UF\\Ok9d0UFZOm9g0SFVOo9j0QFSOR:n0nEoNT:Q1lEmNU:T1lEhNW:Y1iEdNY:\\1gEaN\\:`1dE]N^:c1bE[N`:f1`EVNf:g1a1nN[bh3"}, "label": "a light brown horse standing behind the horse whose mouth is wide opened"}]}
{"id": 27510, "image": "COCO_train2014_000000027510.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a younf woman wearing a green shirt eating a piece of pizza with her legs up\"."}], "task": "refering", "answer_template": "The \"a younf woman wearing a green shirt eating a piece of pizza with her legs up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.00484375, 0.23870833333333333, 0.6209687500000001, 0.98925], "iscrowd": 0, "area": 85952.24194999997, "rle": {"size": [480, 640], "counts": "Pa1k:T410000O10000O10000O100O10000O1000O0100O10000O10000O100O10000O10000O100000000000000000000000000000000000000000000000000000000000000000000000000000O101O0000000000001O001O001O001O001O001O001O001O001O001O001O001cJlKLT43nKLS42oKMQ41RLNo30SLOm30ULOl3OVL0j3NZL0g3N[L1e3N]L1d3M^L2b3N_L1b3M`L2`3MbL2_3McL1]3NeL1\\3MfL2Z3NgL1Z3MhL2X3MjL2W3MjL2V3MmL1T3MnL2R3NoL1R3MPM2P3MRM2o2MRM2n2MTM2m2LVM1k2OVM0k2NWM1i2NZM0g2OZM0f2O]MOd2O^M0b2LcM3^2HhM6X2FmM9T2ARN>n1^OXN`0i1[O\\Nd0g1UO_Ni0e1oN`NP1d1iN\\NZ1g1aNWNc1m1VNRNn1R2lMmMW2Y2aMeMc2b2SM]MQ3j2fLWM[3P3\\LQMe3V3QLmLo3Z3hKgLY4o6100O1O100O1O100O1O100SOm0M3M_MmF[Oo8g0UGWOh8j0\\GTO`8n0eGoNW8S1mGkNo7W1UHgNg7[1\\HdNc7\\1_HcNb7[1`HdN`7Z1dHdN\\7[1fHdN[7Z1gHeNY7Y1jHfNV7Y1mHeNS7Y1PIfNQ7X1QIgNo6X1SIgNm6W1UIiNk6V1UIkNl6R1UIoNk6o0VIROj6l0VIVOk6g0VIZOj6c0XI^Oh6`0YIAk69UIIU7KlH6U7GlH:T7CmH?T7^OmHb0T7\\OmHe0T7XOmHi0V7ROjHP1Y7kNhHV1[7dNgH]1[7_NfHb1]7RNjHP2Y7cMPI^2h91_LfMZJ\\2d5kMaIFSNa2Z8mM`IDTN`2Z8oM`ICSN`2[8PN`IBRN`2\\8QN_IBSN^2]8RN^IBRN^2`8QN\\ICQN^2c8PNZIb2f6_MWIc2i6_MTIb2l6_MRIb2n6_MPIb2P7_MmHc2S7]MlHd2T7]MjHd2V7]MhHd2X7\\MfHf2Z7[MdHf2\\7[MbHf2]7[MbHf2^7[M`Hf2`7[M^Hf2b7ZM]Hg2c7ZMfGeN:S4P8dMmG]2S8dMkG]2U8eMhG\\2X8eMfG\\2Z8fMcG[2]8gM`GZ2`8S200000000O10000000000000000000]M_GgNa8V1aGkN_8S1bGnN^8o0dGRO\\8l0eGUO[8h0gGYOX8e0jG\\OV8b0kG_OU8>mGCS8;nGFR8BeH?\\7_OdHb0]7\\OcHe0^7YObHh0_7VOaHl0_7ROaHo0`7oN`HR1a7lN_HU1b7iN^HX1c7fN]H[1c7dN]H]1d7aN\\H`1e7^N[Hc1f7[NZHf1g7XNYHi1h7UNXHl1i7RNWHo1j7mMXHT2i7gMZHZ2g7aM\\H`2e7[M^Hf2c7UM`Hl2`7PMcHQ3^7kLdHV3]7eLeH]3\\7^LfHd3Y92N2N2M3N2N2N2N2M3N20000O2O00000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O2O01O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2L3L4K6K4L4K6K4K5L4L5J5L4L5JVaa3"}, "label": "a younf woman wearing a green shirt eating a piece of pizza with her legs up"}]}
{"id": 27510, "image": "COCO_train2014_000000027510.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing green t - shirt and a black pants holding a piece of pizza\"."}], "task": "refering", "answer_template": "The \"a women wearing green t - shirt and a black pants holding a piece of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.00484375, 0.23870833333333333, 0.6209687500000001, 0.98925], "iscrowd": 0, "area": 85952.24194999997, "rle": {"size": [480, 640], "counts": "Pa1k:T410000O10000O10000O100O10000O1000O0100O10000O10000O100O10000O10000O100000000000000000000000000000000000000000000000000000000000000000000000000000O101O0000000000001O001O001O001O001O001O001O001O001O001O001O001cJlKLT43nKLS42oKMQ41RLNo30SLOm30ULOl3OVL0j3NZL0g3N[L1e3N]L1d3M^L2b3N_L1b3M`L2`3MbL2_3McL1]3NeL1\\3MfL2Z3NgL1Z3MhL2X3MjL2W3MjL2V3MmL1T3MnL2R3NoL1R3MPM2P3MRM2o2MRM2n2MTM2m2LVM1k2OVM0k2NWM1i2NZM0g2OZM0f2O]MOd2O^M0b2LcM3^2HhM6X2FmM9T2ARN>n1^OXN`0i1[O\\Nd0g1UO_Ni0e1oN`NP1d1iN\\NZ1g1aNWNc1m1VNRNn1R2lMmMW2Y2aMeMc2b2SM]MQ3j2fLWM[3P3\\LQMe3V3QLmLo3Z3hKgLY4o6100O1O100O1O100O1O100SOm0M3M_MmF[Oo8g0UGWOh8j0\\GTO`8n0eGoNW8S1mGkNo7W1UHgNg7[1\\HdNc7\\1_HcNb7[1`HdN`7Z1dHdN\\7[1fHdN[7Z1gHeNY7Y1jHfNV7Y1mHeNS7Y1PIfNQ7X1QIgNo6X1SIgNm6W1UIiNk6V1UIkNl6R1UIoNk6o0VIROj6l0VIVOk6g0VIZOj6c0XI^Oh6`0YIAk69UIIU7KlH6U7GlH:T7CmH?T7^OmHb0T7\\OmHe0T7XOmHi0V7ROjHP1Y7kNhHV1[7dNgH]1[7_NfHb1]7RNjHP2Y7cMPI^2h91_LfMZJ\\2d5kMaIFSNa2Z8mM`IDTN`2Z8oM`ICSN`2[8PN`IBRN`2\\8QN_IBSN^2]8RN^IBRN^2`8QN\\ICQN^2c8PNZIb2f6_MWIc2i6_MTIb2l6_MRIb2n6_MPIb2P7_MmHc2S7]MlHd2T7]MjHd2V7]MhHd2X7\\MfHf2Z7[MdHf2\\7[MbHf2]7[MbHf2^7[M`Hf2`7[M^Hf2b7ZM]Hg2c7ZMfGeN:S4P8dMmG]2S8dMkG]2U8eMhG\\2X8eMfG\\2Z8fMcG[2]8gM`GZ2`8S200000000O10000000000000000000]M_GgNa8V1aGkN_8S1bGnN^8o0dGRO\\8l0eGUO[8h0gGYOX8e0jG\\OV8b0kG_OU8>mGCS8;nGFR8BeH?\\7_OdHb0]7\\OcHe0^7YObHh0_7VOaHl0_7ROaHo0`7oN`HR1a7lN_HU1b7iN^HX1c7fN]H[1c7dN]H]1d7aN\\H`1e7^N[Hc1f7[NZHf1g7XNYHi1h7UNXHl1i7RNWHo1j7mMXHT2i7gMZHZ2g7aM\\H`2e7[M^Hf2c7UM`Hl2`7PMcHQ3^7kLdHV3]7eLeH]3\\7^LfHd3Y92N2N2M3N2N2N2N2M3N20000O2O00000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O2O01O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2L3L4K6K4L4K6K4K5L4L5J5L4L5JVaa3"}, "label": "a women wearing green t - shirt and a black pants holding a piece of pizza"}]}
{"id": 576376, "image": "COCO_train2014_000000576376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy operating laptop in front of a television , wearing a white sweat shirt\"."}], "task": "refering", "answer_template": "The \"a guy operating laptop in front of a television , wearing a white sweat shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 216, 217, 218, 219, 228, 229, 230, 231, 232, 233, 234, 235, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.18442622950819673, 0.40960937499999994, 0.96903981264637, 0.98903125], "iscrowd": 0, "area": 77251.85685, "rle": {"size": [640, 427], "counts": "_ka18ac0=D;D<D=D;J6I7I8H7I7J7H:F9G9H9F8H7J6J7J5K6I8I1O0O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O1000000O10000O1000000O1]NeIaE[6P:UJoEk5e9bJZF^5Y9PKeFQ5P9[KoFe4n8_KQGa4k8eKSG[4j8iKTGX4i8lKVGT4f8QLYGo3d8UL[Gk3b8ZL[Gg3b8]L]Gc3_8cL_G]3^8gL`GZ3`8hL^GX3a8jL^GV3a8mL\\GT3d8nLYGS3f8oLXGR3h8PMUGQ3j8QMSGQ3l8RMQGo2o8Q4O1O100O100O1O100O1O100O100O00100O100O1O10000O100000000O100000000O100000000O100000000O100000000O100001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O0kHRGX3o8gLRGX3o8fLRGZ3n8fLSGY3n8eLTGZ3l8fLUGY3m8eLSG[3o8bLSG]3n8bLSG]3o8`LRG`3P9^LQGa3Q9\\LQGc3P9\\LPGd3R9YLPGf3R9XLoFg3S9VLnFj3S9ULnFj3T9TLmFk3U9RLlFn3V9PLjFP4X9mKiFS4X9lKiFS4Y9fKlFZ4V9_KoFa4S9XKRGh4o8RKVGn4P:hIUFW6j;00000001O000000001O00001O2N2N2N2N1O2N2N2N6J2N1O000000000000001O00000000000000001O000000000000000000000000000000000000001O1O001O1O001O1O001O1O0XOi0L4L3N3L4L3N3L4N131O00001O0O101O00001O000N3M0O2N2N1O2N2N101N2N1O2N2N1O2O1N1O2N2N1O2N2O3L6J6J5K6J6J3N3I7G9H8G8H9H\\a7"}, "label": "a guy operating laptop in front of a television , wearing a white sweat shirt"}]}
{"id": 576376, "image": "COCO_train2014_000000576376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person\"."}], "task": "refering", "answer_template": "The \"a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 216, 217, 218, 219, 228, 229, 230, 231, 232, 233, 234, 235, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.18442622950819673, 0.40960937499999994, 0.96903981264637, 0.98903125], "iscrowd": 0, "area": 77251.85685, "rle": {"size": [640, 427], "counts": "_ka18ac0=D;D<D=D;J6I7I8H7I7J7H:F9G9H9F8H7J6J7J5K6I8I1O0O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O1000000O10000O1000000O1]NeIaE[6P:UJoEk5e9bJZF^5Y9PKeFQ5P9[KoFe4n8_KQGa4k8eKSG[4j8iKTGX4i8lKVGT4f8QLYGo3d8UL[Gk3b8ZL[Gg3b8]L]Gc3_8cL_G]3^8gL`GZ3`8hL^GX3a8jL^GV3a8mL\\GT3d8nLYGS3f8oLXGR3h8PMUGQ3j8QMSGQ3l8RMQGo2o8Q4O1O100O100O1O100O1O100O100O00100O100O1O10000O100000000O100000000O100000000O100000000O100000000O100001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O0kHRGX3o8gLRGX3o8fLRGZ3n8fLSGY3n8eLTGZ3l8fLUGY3m8eLSG[3o8bLSG]3n8bLSG]3o8`LRG`3P9^LQGa3Q9\\LQGc3P9\\LPGd3R9YLPGf3R9XLoFg3S9VLnFj3S9ULnFj3T9TLmFk3U9RLlFn3V9PLjFP4X9mKiFS4X9lKiFS4Y9fKlFZ4V9_KoFa4S9XKRGh4o8RKVGn4P:hIUFW6j;00000001O000000001O00001O2N2N2N2N1O2N2N2N6J2N1O000000000000001O00000000000000001O000000000000000000000000000000000000001O1O001O1O001O1O001O1O0XOi0L4L3N3L4L3N3L4N131O00001O0O101O00001O000N3M0O2N2N1O2N2N101N2N1O2N2N1O2O1N1O2N2N1O2N2O3L6J6J5K6J6J3N3I7G9H8G8H9H\\a7"}, "label": "a person"}]}
{"id": 576376, "image": "COCO_train2014_000000576376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop in which the person is using\"."}], "task": "refering", "answer_template": "The \"the laptop in which the person is using\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 184, 185, 189, 190, 198, 199, 200, 204, 205, 213, 214, 215, 216, 219, 220, 228, 229, 230, 235, 236], "bbox": [0.2597423887587822, 0.528125, 0.7400234192037471, 0.689015625], "iscrowd": 0, "area": 7484.960000000001, "rle": {"size": [640, 427], "counts": "bfU2S1mb0P2PN0000000000000000001O00O1O1O100O1O1O1O100O1O1000000O1000000O10000O100O100O1O1RNY_Oc0g`0[O[_Od0f`0YO^_Of0b`0WOb_Og0_`0WOc_Oh0^`0UOe_Oj0\\`0TOf_Ok0[`0ROh_On0X`0POk_O3gNg0Qc010N101O0O101O0O1O1O2N100O2N5Kocg1JT\\XN500O1O1O1O1O1O0O1N3N1O2M2O2M2O1SOSOe^OP1i`0TOa^O5c0j0k`0@P_Od0Pa0]Ok^Oh0Ta0YOh^Ok0Wa0P1O001O00000000001O0000001O00001O001O0o^O`Mb`0a2W_OeMi`0i20000013Ld0\\Oe0[Oa0_O`0_O`0A`nT2"}, "label": "the laptop in which the person is using"}]}
{"id": 576376, "image": "COCO_train2014_000000576376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop screen that girl is working on while watching tv\"."}], "task": "refering", "answer_template": "The \"laptop screen that girl is working on while watching tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 184, 185, 189, 190, 198, 199, 200, 204, 205, 213, 214, 215, 216, 219, 220, 228, 229, 230, 235, 236], "bbox": [0.2597423887587822, 0.528125, 0.7400234192037471, 0.689015625], "iscrowd": 0, "area": 7484.960000000001, "rle": {"size": [640, 427], "counts": "bfU2S1mb0P2PN0000000000000000001O00O1O1O100O1O1O1O100O1O1000000O1000000O10000O100O100O1O1RNY_Oc0g`0[O[_Od0f`0YO^_Of0b`0WOb_Og0_`0WOc_Oh0^`0UOe_Oj0\\`0TOf_Ok0[`0ROh_On0X`0POk_O3gNg0Qc010N101O0O101O0O1O1O2N100O2N5Kocg1JT\\XN500O1O1O1O1O1O0O1N3N1O2M2O2M2O1SOSOe^OP1i`0TOa^O5c0j0k`0@P_Od0Pa0]Ok^Oh0Ta0YOh^Ok0Wa0P1O001O00000000001O0000001O00001O001O0o^O`Mb`0a2W_OeMi`0i20000013Ld0\\Oe0[Oa0_O`0_O`0A`nT2"}, "label": "laptop screen that girl is working on while watching tv"}]}
{"id": 576376, "image": "COCO_train2014_000000576376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a television set where something is playing\"."}], "task": "refering", "answer_template": "The \"a television set where something is playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146], "bbox": [0.21892271662763466, 0.14621875, 0.7679391100702577, 0.40689062500000006], "iscrowd": 0, "area": 37703.0405, "rle": {"size": [640, 427], "counts": "_[j13Tc0j0UOk0UOk0VOj0UOk0^Ob0O1O1O1O1O2N1O1O1O1O1O1O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O1O100O1O1O1O100Oj0VOc1]Nc1]NTem1"}, "label": "a television set where something is playing"}]}
{"id": 576376, "image": "COCO_train2014_000000576376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tv screen with blurred image\"."}], "task": "refering", "answer_template": "The \"tv screen with blurred image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 63, 64, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146], "bbox": [0.21892271662763466, 0.14621875, 0.7679391100702577, 0.40689062500000006], "iscrowd": 0, "area": 37703.0405, "rle": {"size": [640, 427], "counts": "_[j13Tc0j0UOk0UOk0VOj0UOk0^Ob0O1O1O1O1O2N1O1O1O1O1O1O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O1O100O1O1O1O100Oj0VOc1]Nc1]NTem1"}, "label": "tv screen with blurred image"}]}
{"id": 371577, "image": "COCO_train2014_000000371577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboarder turning on a ramps edge\"."}], "task": "refering", "answer_template": "The \"skateboarder turning on a ramps edge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 70, 71, 93, 94, 116, 117, 118, 139, 140, 141, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 191, 193, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239], "bbox": [0.052359375, 0.15314583333333334, 0.43581250000000005, 0.6418958333333333], "iscrowd": 0, "area": 20008.10055, "rle": {"size": [480, 640], "counts": "^S`0;^>;E<D;E<D<C<E<D;G:H8I6I7J6I8I6J6K3N1O001O1O001O1\\OkJ[GV5d8nJXGR5h8SKSGn4k8WKQGj4n8d001O1O1OjKXGT2h8dM`G]2`8ZMhGg2W8QMRHo2m7iL[HW3e7fL_HZ3a7bLbH_3]7^LfHb3Z7[LjHe3U7XLnHi3R7SLRIm3P900100O010O1O10O01O10O01O100O00100O1O01kL\\Em1d:jMfEU2Z:bMoE]2R:`MQF`2o9]MTFc2l9ZMWFe2j9XMYFg2g9SM`Fl2k:O2N001O1O1O001O1O001O1O1O001O001N2O001O001O001O1O001O001O001O001O1O0000000000001O000000001N100000001O0000001O0000001O03M2N2O1N100O1O100O100O0001N101O001N10000O101O000O101N100O1O2N1O1O1O2O0N2N2N3N1N2N2N201O0000001N10001O1O1O1N2O1O100O10000O10000O10000O10O01000O010O10O0100O0010O01O1O001O001O001O0O2M2N3M2N3M1O2N1010O010O00010O010O010O011N3N1N2O1N:VO_oX5"}, "label": "skateboarder turning on a ramps edge"}]}
{"id": 371577, "image": "COCO_train2014_000000371577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder wearing a white top touching the lip of the bowl with his board\"."}], "task": "refering", "answer_template": "The \"a skateboarder wearing a white top touching the lip of the bowl with his board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 70, 71, 93, 94, 116, 117, 118, 139, 140, 141, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 191, 193, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239], "bbox": [0.052359375, 0.15314583333333334, 0.43581250000000005, 0.6418958333333333], "iscrowd": 0, "area": 20008.10055, "rle": {"size": [480, 640], "counts": "^S`0;^>;E<D;E<D<C<E<D;G:H8I6I7J6I8I6J6K3N1O001O1O001O1\\OkJ[GV5d8nJXGR5h8SKSGn4k8WKQGj4n8d001O1O1OjKXGT2h8dM`G]2`8ZMhGg2W8QMRHo2m7iL[HW3e7fL_HZ3a7bLbH_3]7^LfHb3Z7[LjHe3U7XLnHi3R7SLRIm3P900100O010O1O10O01O10O01O100O00100O1O01kL\\Em1d:jMfEU2Z:bMoE]2R:`MQF`2o9]MTFc2l9ZMWFe2j9XMYFg2g9SM`Fl2k:O2N001O1O1O001O1O001O1O1O001O001N2O001O001O001O1O001O001O001O001O1O0000000000001O000000001N100000001O0000001O0000001O03M2N2O1N100O1O100O100O0001N101O001N10000O101O000O101N100O1O2N1O1O1O2O0N2N2N3N1N2N2N201O0000001N10001O1O1O1N2O1O100O10000O10000O10000O10O01000O010O10O0100O0010O01O1O001O001O001O0O2M2N3M2N3M1O2N1010O010O00010O010O010O011N3N1N2O1N:VO_oX5"}, "label": "a skateboarder wearing a white top touching the lip of the bowl with his board"}]}
{"id": 52090, "image": "COCO_train2014_000000052090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboard that the person is stepping on\"."}], "task": "refering", "answer_template": "The \"skateboard that the person is stepping on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 239, 240, 241, 264, 265, 266, 288, 289, 290], "bbox": [0.40815625000000005, 0.7052222222222222, 0.63903125, 0.9886666666666667], "iscrowd": 0, "area": 3519.1502499999965, "rle": {"size": [360, 640], "counts": "XPl21W;2N2N2M4M2N2N2M3N2N3M2O1N2O1N0SFRO`9o0^FUO`9j0`FXO^9h0bFXO_9h0_FYOa9g0_FXOc9h0\\FXOd9h0[FXOg9g0YFYOg9h0XFWOj9h0VFXOj9i0TFXOm9g0SFXOo9h0PFXOQ:g0nEYOT:f0lEZOU:l00O2O1O10000O10O010000O10000O010O10000O01000O100O10O10O100O10O10O10000O010O10000O10O10O100O01000O0100O0100O01000O01000O010O10O10O01000O010O10O102N6I2OO1O001O001O1N101O1O001O1O001O1O001O00100O00100O001O10O01O1N101N101N2O0O2N2O0O2O1N10lk`2"}, "label": "skateboard that the person is stepping on"}]}
{"id": 52090, "image": "COCO_train2014_000000052090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard being tipped on its back wheels\"."}], "task": "refering", "answer_template": "The \"a skateboard being tipped on its back wheels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 239, 240, 241, 264, 265, 266, 288, 289, 290], "bbox": [0.40815625000000005, 0.7052222222222222, 0.63903125, 0.9886666666666667], "iscrowd": 0, "area": 3519.1502499999965, "rle": {"size": [360, 640], "counts": "XPl21W;2N2N2M4M2N2N2M3N2N3M2O1N2O1N0SFRO`9o0^FUO`9j0`FXO^9h0bFXO_9h0_FYOa9g0_FXOc9h0\\FXOd9h0[FXOg9g0YFYOg9h0XFWOj9h0VFXOj9i0TFXOm9g0SFXOo9h0PFXOQ:g0nEYOT:f0lEZOU:l00O2O1O10000O10O010000O10000O010O10000O01000O100O10O10O100O10O10O10000O010O10000O10O10O100O01000O0100O0100O01000O01000O010O10O10O01000O010O10O102N6I2OO1O001O001O1N101O1O001O1O001O1O001O00100O00100O001O10O01O1N101N101N2O0O2N2O0O2O1N10lk`2"}, "label": "a skateboard being tipped on its back wheels"}]}
{"id": 543617, "image": "COCO_train2014_000000543617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with no shirt in jeans in doorway\"."}], "task": "refering", "answer_template": "The \"man with no shirt in jeans in doorway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 127, 128, 150, 151, 173, 174, 196, 197], "bbox": [0.493984375, 0.0016861826697892272, 0.625015625, 0.6089227166276348], "iscrowd": 0, "area": 12479.915249999998, "rle": {"size": [427, 640], "counts": "SlS43R=6I7I7I8H7I7J6I8H7I7I7N2O2N1O1O1O2N100O1O2_FjNj6W1TIoNg6R1WIUOc6l0\\IYO_6i0^I^O\\6c0bIDX6=fIIV67iIOQ62mI5m5MPJ9k5HSJ?g5BXJb0d5_OZJc0e5_OXJd0f5]OXJe0g5\\OXJf0g5_3O2N2N1O2N1O2J501OL4L4M3L4L4ZMf200O100O10000O101O0O10000O10000O1O1O1O1O1O1O1O1O1O1O1O1YOSGoLn8[2iGaMZ8i1]HTNf7U1QIhNQ7b0fIZOjXT3"}, "label": "man with no shirt in jeans in doorway"}]}
{"id": 543617, "image": "COCO_train2014_000000543617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man without shirt\"."}], "task": "refering", "answer_template": "The \"a man without shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 127, 128, 150, 151, 173, 174, 196, 197], "bbox": [0.493984375, 0.0016861826697892272, 0.625015625, 0.6089227166276348], "iscrowd": 0, "area": 12479.915249999998, "rle": {"size": [427, 640], "counts": "SlS43R=6I7I7I8H7I7J6I8H7I7I7N2O2N1O1O1O2N100O1O2_FjNj6W1TIoNg6R1WIUOc6l0\\IYO_6i0^I^O\\6c0bIDX6=fIIV67iIOQ62mI5m5MPJ9k5HSJ?g5BXJb0d5_OZJc0e5_OXJd0f5]OXJe0g5\\OXJf0g5_3O2N2N1O2N1O2J501OL4L4M3L4L4ZMf200O100O10000O101O0O10000O10000O1O1O1O1O1O1O1O1O1O1O1O1YOSGoLn8[2iGaMZ8i1]HTNf7U1QIhNQ7b0fIZOjXT3"}, "label": "a man without shirt"}]}
{"id": 543617, "image": "COCO_train2014_000000543617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person far left side standing striped shirt\"."}], "task": "refering", "answer_template": "The \"person far left side standing striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 300, 322, 323], "bbox": [0.0012812499999999998, 0.0034660421545667447, 0.191765625, 0.999016393442623], "iscrowd": 0, "area": 31915.06785, "rle": {"size": [427, 640], "counts": "^=a2\\1c5NP5mN001O00000000000000000000000000001O0000000000000000000O10bLNcI2Y65dIKY6:eIFW6`0nG@90e7f0jFd0^1eNd7`2YH`Mc7g2ZHYMb7o2ZHQMb7V3[HjLa7]3\\HcL`7e3\\H[L`7l3]HTL_7S4^HmK^7Z4_HfK]7a4`H_K\\7g4bHYKZ7n4cHRKY7U5dHkJX7\\5eHdJW7c5fH]JV7j5hHUJV7Q6O100O1O100O1O00100O1O100O1O1O100O100O100O10000O100O100O101O2M2O1N2O1O1N2O0O2O0O2N2O0O2N101N2O0O2eNkH]LV7b3RIWLo6g3ZIPLh6n3`IjKc6S4fIeK\\6X4_1L3M3M3M4L3K5H9G8H8H8XOi0L3L9Ha0^Oc0^Ok_g6"}, "label": "person far left side standing striped shirt"}]}
{"id": 543617, "image": "COCO_train2014_000000543617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bacj of somone dressed in shorts , a t - shirt , and vest standing behind a man working on his laptop\"."}], "task": "refering", "answer_template": "The \"the bacj of somone dressed in shorts , a t - shirt , and vest standing behind a man working on his laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 300, 322, 323], "bbox": [0.0012812499999999998, 0.0034660421545667447, 0.191765625, 0.999016393442623], "iscrowd": 0, "area": 31915.06785, "rle": {"size": [427, 640], "counts": "^=a2\\1c5NP5mN001O00000000000000000000000000001O0000000000000000000O10bLNcI2Y65dIKY6:eIFW6`0nG@90e7f0jFd0^1eNd7`2YH`Mc7g2ZHYMb7o2ZHQMb7V3[HjLa7]3\\HcL`7e3\\H[L`7l3]HTL_7S4^HmK^7Z4_HfK]7a4`H_K\\7g4bHYKZ7n4cHRKY7U5dHkJX7\\5eHdJW7c5fH]JV7j5hHUJV7Q6O100O1O100O1O00100O1O100O1O1O100O100O100O10000O100O100O101O2M2O1N2O1O1N2O0O2O0O2N2O0O2N101N2O0O2eNkH]LV7b3RIWLo6g3ZIPLh6n3`IjKc6S4fIeK\\6X4_1L3M3M3M4L3K5H9G8H8H8XOi0L3L9Ha0^Oc0^Ok_g6"}, "label": "the bacj of somone dressed in shorts , a t - shirt , and vest standing behind a man working on his laptop"}]}
{"id": 543617, "image": "COCO_train2014_000000543617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a hat with a fork in the hatband\"."}], "task": "refering", "answer_template": "The \"a man wearing a hat with a fork in the hatband\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.05546875, 0.04201405152224825, 0.61471875, 0.9858313817330211], "iscrowd": 0, "area": 79294.24644999998, "rle": {"size": [427, 640], "counts": "b[?]1e;:F:G9F;F9F:F:G9F:G:E:F:K5M3N3M1O1O1O1O001O1O1O1J[JeHf5Z7[JeHf5[7ZJdHg5[7500O010O1O100O100O1O100O010O1O100O100O1O10O0100O100O1O100O100O1O100O100O1O100O100O1O100O010O100O1O100O100O1O100O100O1O100O100fLUI3l6IYI5h6H[I8e6E^I:c6E_I:a6F_I9b6G_I8a6G`I8a6H_I7b6I_I6a6J_I5a6L`IVN]Om0T7n0^IQNBP1Q7o0^IlMET1n6P1]IhMIV1l6S1ZIeMLW1k6T1ZIcMLX1k6V1XIaMNW1l6X1WI_MNX1l6Z1UI]M0X1l6[1TI]M0V1n6]1SI[M0W1n6_1QIYM2W1n6`1QIWM2W1o6c1nHVM3V1P7d1nHTM3W1P7f1lHRM5V1Q7h1cIWN^6j1aITNa6l1_InMg6R2YIhMm6Y2RI`MU7^2mH\\MY7\\2oH^MW7h1c2O1O1N2O001O1N2O1O1O1N2O100O10000O10000O100O10000O10000O100O10000O01000O10000O100000O010O1O010O1O010O1O00100O1O010O1O010O1O100O1O1O100O1O101N1O100O1O100O1\\OXNdEi1\\:XNbEi1]:ZNaEf1_:\\N^Ee1a:]N]Ed1c:^N[Eb1d:`NZEa1f:aNWE`1h:bNWE^1i:b001N100O100O101O1N2O1N2O0O2O1O1N2O1N2O1N2O1N101O1N3N5J6K5J6K4L2M2O1N3N1N2mJ^K[1d4aN_K^1b4^NaKa1a4[NcKd1^4jLaK\\O4i3^4bLmK[OHR4\\4_LWLTO@\\4[4^L^LkN[Of4Y4]LYMa3i2]LXMc3i2[LYMd3i2ZLWMe3k2YLWMf3j2XLWMP2`MjN\\5TOUMh1hMTOT5ROVM`1oM]Om4QOUMY1WNFf4nNUMZ1VNHf4mNTMZ1WNHg4lNTMZ1VNJh4iNSM\\1VNJi4iNRM[1VNLh4hNTMZ1UNMh4hNTMZ1UNNh4gNTMY1UN0g4fNUMZ1TNOh4fNVMZ1RN0i4eNUM[1RNOj4eNVM[1PN0j4dNXM[1nM0k4dNXM\\1mM0l4cNXM\\1mMOl4dNXM]1lMOW80iG0X8OhG0Z8NgG2Z8MfG2\\8MdG3]8LcG4^8KbG4`8K`G5a8J_G5c8J]G6d8I\\G7e8H[G7g8HYG8h8GXG8j8GVG9k8FUG:l8ETG:n8ERG;o8DQG;Q9DnF=S9BmF=U9BkF>W9_OiFb0X9]OhFb0Z9]OfFc0^9YOaFg0e9TO[Fl0j9oNVFQ1Q;0100O01000O010O10O10O10O010000O010O01O10O01O0010O01O1O0100O2O2N2M2O2M2O2N2M2O2M3N1O2M3N1N3N2N1N`fV3"}, "label": "a man wearing a hat with a fork in the hatband"}]}
{"id": 543617, "image": "COCO_train2014_000000543617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with glasses and painted fingernails\"."}], "task": "refering", "answer_template": "The \"the man with glasses and painted fingernails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.05546875, 0.04201405152224825, 0.61471875, 0.9858313817330211], "iscrowd": 0, "area": 79294.24644999998, "rle": {"size": [427, 640], "counts": "b[?]1e;:F:G9F;F9F:F:G9F:G:E:F:K5M3N3M1O1O1O1O001O1O1O1J[JeHf5Z7[JeHf5[7ZJdHg5[7500O010O1O100O100O1O100O010O1O100O100O1O10O0100O100O1O100O100O1O100O100O1O100O100O1O100O010O100O1O100O100O1O100O100O1O100O100fLUI3l6IYI5h6H[I8e6E^I:c6E_I:a6F_I9b6G_I8a6G`I8a6H_I7b6I_I6a6J_I5a6L`IVN]Om0T7n0^IQNBP1Q7o0^IlMET1n6P1]IhMIV1l6S1ZIeMLW1k6T1ZIcMLX1k6V1XIaMNW1l6X1WI_MNX1l6Z1UI]M0X1l6[1TI]M0V1n6]1SI[M0W1n6_1QIYM2W1n6`1QIWM2W1o6c1nHVM3V1P7d1nHTM3W1P7f1lHRM5V1Q7h1cIWN^6j1aITNa6l1_InMg6R2YIhMm6Y2RI`MU7^2mH\\MY7\\2oH^MW7h1c2O1O1N2O001O1N2O1O1O1N2O100O10000O10000O100O10000O10000O100O10000O01000O10000O100000O010O1O010O1O010O1O00100O1O010O1O010O1O100O1O1O100O1O101N1O100O1O100O1\\OXNdEi1\\:XNbEi1]:ZNaEf1_:\\N^Ee1a:]N]Ed1c:^N[Eb1d:`NZEa1f:aNWE`1h:bNWE^1i:b001N100O100O101O1N2O1N2O0O2O1O1N2O1N2O1N2O1N101O1N3N5J6K5J6K4L2M2O1N3N1N2mJ^K[1d4aN_K^1b4^NaKa1a4[NcKd1^4jLaK\\O4i3^4bLmK[OHR4\\4_LWLTO@\\4[4^L^LkN[Of4Y4]LYMa3i2]LXMc3i2[LYMd3i2ZLWMe3k2YLWMf3j2XLWMP2`MjN\\5TOUMh1hMTOT5ROVM`1oM]Om4QOUMY1WNFf4nNUMZ1VNHf4mNTMZ1WNHg4lNTMZ1VNJh4iNSM\\1VNJi4iNRM[1VNLh4hNTMZ1UNMh4hNTMZ1UNNh4gNTMY1UN0g4fNUMZ1TNOh4fNVMZ1RN0i4eNUM[1RNOj4eNVM[1PN0j4dNXM[1nM0k4dNXM\\1mM0l4cNXM\\1mMOl4dNXM]1lMOW80iG0X8OhG0Z8NgG2Z8MfG2\\8MdG3]8LcG4^8KbG4`8K`G5a8J_G5c8J]G6d8I\\G7e8H[G7g8HYG8h8GXG8j8GVG9k8FUG:l8ETG:n8ERG;o8DQG;Q9DnF=S9BmF=U9BkF>W9_OiFb0X9]OhFb0Z9]OfFc0^9YOaFg0e9TO[Fl0j9oNVFQ1Q;0100O01000O010O10O10O10O010000O010O01O10O01O0010O01O1O0100O2O2N2M2O2M2O2N2M2O2M3N1O2M3N1N3N2N1N`fV3"}, "label": "the man with glasses and painted fingernails"}]}
{"id": 478092, "image": "COCO_train2014_000000478092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bus driving down a busy road\"."}], "task": "refering", "answer_template": "The \"a white bus driving down a busy road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 165, 166, 167, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 216, 217, 218], "bbox": [0.0033799999999999998, 0.6779733333333333, 0.33708, 0.9858399999999999], "iscrowd": 0, "area": 9013.43515, "rle": {"size": [375, 500], "counts": "dQ1;[;<D9H2M3N2M3N2M3N2M10O0100O100O100O100O100O100O100O100O010O100O100O100O100O100O100O100O10O0100O100O10O001O0O2O001O001O001N1O2N101N1O2N1O2N101N110O010O0100O010O010O10O010O0100O010O10O010O0100O101N101N100O2O0O2O0O2O0O101N100O10O0010O010O010O010O010O010O010O01O010O010O010O010O010O010O011N2O1N2N101N2O1N2O0O2O1N2O0O2O1N2O1N1O2O1L4E;F9Ffnh3"}, "label": "a white bus driving down a busy road"}]}
{"id": 478092, "image": "COCO_train2014_000000478092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a city bus entering into a intersection\"."}], "task": "refering", "answer_template": "The \"a city bus entering into a intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 165, 166, 167, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 216, 217, 218], "bbox": [0.0033799999999999998, 0.6779733333333333, 0.33708, 0.9858399999999999], "iscrowd": 0, "area": 9013.43515, "rle": {"size": [375, 500], "counts": "dQ1;[;<D9H2M3N2M3N2M3N2M10O0100O100O100O100O100O100O100O100O010O100O100O100O100O100O100O100O10O0100O100O10O001O0O2O001O001O001N1O2N101N1O2N1O2N101N110O010O0100O010O010O10O010O0100O010O10O010O0100O101N101N100O2O0O2O0O2O0O101N100O10O0010O010O010O010O010O010O010O01O010O010O010O010O010O010O011N2O1N2N101N2O1N2O0O2O1N2O0O2O1N2O1N1O2O1L4E;F9Ffnh3"}, "label": "a city bus entering into a intersection"}]}
{"id": 322445, "image": "COCO_train2014_000000322445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the hat on\"."}], "task": "refering", "answer_template": "The \"the man with the hat on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 20, 21, 22, 36, 37, 38, 39, 40, 53, 54, 55, 56, 70, 71, 72, 73, 74, 87, 88, 89, 90, 91, 92, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.011291666666666667, 0.032390625, 0.4537083333333333, 0.61321875], "iscrowd": 0, "area": 47308.29795000001, "rle": {"size": [640, 480], "counts": "S_35Wc0d0K6J5M3M3N3H7M3L4L5L3O1O1N3O000001OR_OjMR`0W2k_OkMV`0T2i_OmMV`0T2h_OmMY`0S2e_OoM[`0R2b_OPN^`0P2a_OQN_`0o1__ORNb`0n1\\_OTNd`0l1Z_OVNf`0j1Y_OWNg`0j1V_OWNj`0j1U_OWNk`0h1T_OZNl`0e1T_O\\Nl`0b1T_O_Nm`0`1S_OaNl`0^1T_OdNl`0[1T_OfNk`0Z1T_OgNm`0W1T_OjNl`0U1S_OmNl`0R1U_OoNk`0n0V_OSOj`0h0Y_O[Og`0?W_OIi`01T_O8^;]NmIP3R6RMjI0jJl2[;UMhI5mJc2Z;ZMeI9PKH0P2d;PNYI=SK]O3X2^;PNYIa0XK`1^;PNSIU3l6lLiH_3U7dL[F:C]3P:ZLWFk0UOS3c:ULTFU1fNn2V;oKPFP6P:RJmEP6R:PJQEC6_6i:oIPELKW6U;nInDl6R;TImDn6R;P1000001O0001O01O00001N10000`N_G`Gc8^8`G`G`8^8eG^G\\8a8gG]GX8c8kGZGT8h8mGWGQ8j8QHUGm7m8THQGm7o8THPGk7Q9VHmFk7S9VHlFi7U9XHjFg7W9ZHgFg7Y9ZHfFe7\\9ZHdFf7\\9[HcFd7^9\\HbFc7`9]H^Fd7b9\\H^Fc7d9]H[Fc7e9]H[Fc7e9^HZFb7g9Z110N2M3L4M3M3M3M2O2O100O10000O100O011O0O101N100O2O00001O00010O000010O0001O1O2O1N1O000001O0O101O1O001O001_GPGP6P9mITGR6m8jIVGV6j8gIZGX6h8YIfGf6\\8XIdGh6U9^HmFa7S9^HnFb7R9^HoFa7R9]HoFb7S9\\HoFc7R9\\HnFd7S9ZHoFe7R9YHPGf7Q9YHoFg7Q9XHQGg7o8XHRGh7n8XHSGg7m8XHTGh7l8XHUGg7l8WHVGh7j8WHXGh7h8XHXGh7h8WHZGh7f8WH\\Gh7d8XH]Gg7d8WH]Gh7d8WH^Gh7b8WH`Gh7]:N1O2bNmClJU<Q5\\D^Jg;`5hDQJY;m5_1M3M3M3J7J5K6J5jNQAUMU?h2QAQMS?l2RAmLT?n2SAjLS?R3SAeLT?W3n0L4L4L3M4L4L4L4L<D8H7G8H8ZOR]ONUc0LT]OITc01aaS5"}, "label": "the man with the hat on"}]}
{"id": 322445, "image": "COCO_train2014_000000322445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a cowboy hat , holding a floral print suitcase\"."}], "task": "refering", "answer_template": "The \"a man wearing a cowboy hat , holding a floral print suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 20, 21, 22, 36, 37, 38, 39, 40, 53, 54, 55, 56, 70, 71, 72, 73, 74, 87, 88, 89, 90, 91, 92, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.011291666666666667, 0.032390625, 0.4537083333333333, 0.61321875], "iscrowd": 0, "area": 47308.29795000001, "rle": {"size": [640, 480], "counts": "S_35Wc0d0K6J5M3M3N3H7M3L4L5L3O1O1N3O000001OR_OjMR`0W2k_OkMV`0T2i_OmMV`0T2h_OmMY`0S2e_OoM[`0R2b_OPN^`0P2a_OQN_`0o1__ORNb`0n1\\_OTNd`0l1Z_OVNf`0j1Y_OWNg`0j1V_OWNj`0j1U_OWNk`0h1T_OZNl`0e1T_O\\Nl`0b1T_O_Nm`0`1S_OaNl`0^1T_OdNl`0[1T_OfNk`0Z1T_OgNm`0W1T_OjNl`0U1S_OmNl`0R1U_OoNk`0n0V_OSOj`0h0Y_O[Og`0?W_OIi`01T_O8^;]NmIP3R6RMjI0jJl2[;UMhI5mJc2Z;ZMeI9PKH0P2d;PNYI=SK]O3X2^;PNYIa0XK`1^;PNSIU3l6lLiH_3U7dL[F:C]3P:ZLWFk0UOS3c:ULTFU1fNn2V;oKPFP6P:RJmEP6R:PJQEC6_6i:oIPELKW6U;nInDl6R;TImDn6R;P1000001O0001O01O00001N10000`N_G`Gc8^8`G`G`8^8eG^G\\8a8gG]GX8c8kGZGT8h8mGWGQ8j8QHUGm7m8THQGm7o8THPGk7Q9VHmFk7S9VHlFi7U9XHjFg7W9ZHgFg7Y9ZHfFe7\\9ZHdFf7\\9[HcFd7^9\\HbFc7`9]H^Fd7b9\\H^Fc7d9]H[Fc7e9]H[Fc7e9^HZFb7g9Z110N2M3L4M3M3M3M2O2O100O10000O100O011O0O101N100O2O00001O00010O000010O0001O1O2O1N1O000001O0O101O1O001O001_GPGP6P9mITGR6m8jIVGV6j8gIZGX6h8YIfGf6\\8XIdGh6U9^HmFa7S9^HnFb7R9^HoFa7R9]HoFb7S9\\HoFc7R9\\HnFd7S9ZHoFe7R9YHPGf7Q9YHoFg7Q9XHQGg7o8XHRGh7n8XHSGg7m8XHTGh7l8XHUGg7l8WHVGh7j8WHXGh7h8XHXGh7h8WHZGh7f8WH\\Gh7d8XH]Gg7d8WH]Gh7d8WH^Gh7b8WH`Gh7]:N1O2bNmClJU<Q5\\D^Jg;`5hDQJY;m5_1M3M3M3J7J5K6J5jNQAUMU?h2QAQMS?l2RAmLT?n2SAjLS?R3SAeLT?W3n0L4L4L3M4L4L4L4L<D8H7G8H8ZOR]ONUc0LT]OITc01aaS5"}, "label": "a man wearing a cowboy hat , holding a floral print suitcase"}]}
{"id": 314254, "image": "COCO_train2014_000000314254.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor that is on the right wall\"."}], "task": "refering", "answer_template": "The \"the monitor that is on the right wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 274, 275], "bbox": [0.8838125, 0.20519906323185014, 1.0, 0.7948009367681499], "iscrowd": 0, "area": 14853.769200000002, "rle": {"size": [427, 640], "counts": "ZT\\7>k<<F:E<E:F;D;F;D;F;D;F:E<E:F;D;F;D;F:E6K4L2M100O100O101N10000O100O101N10000O100O101N100O10000O2O0O100O10000O2O0O100O100O101O0O100O100O101O0O100`2`Mb3^LVG"}, "label": "the monitor that is on the right wall"}]}
{"id": 314254, "image": "COCO_train2014_000000314254.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a screen with an image of a plate of hot dogs\"."}], "task": "refering", "answer_template": "The \"a screen with an image of a plate of hot dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 76, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282], "bbox": [0.18784375, 0.17508196721311475, 0.68559375, 0.8131615925058547], "iscrowd": 0, "area": 71031.62849999999, "rle": {"size": [427, 640], "counts": "U[b15h<`0@`0@`0@`0_Oa0@`0@`0@`0@`0@`0@`0_O`0A<D<D<E;1O1OC=2N2N2N3L3N21O000010O01O00001O010O00001O00100O2N2N3M2N2O1N2N2N2N0001O000O1000001O0N2L4M3M3M4L3M3M3N20001O00000000001O0003M3M3M4M2M3M00000001O00000000001O00000000001O000000000YOh00O001O001O010O7I<D000001O0001O000001O00000000001O01O00000001O0000000YO\\JgId5Z6_JbIb5^6bJ^I]5c6fJYIZ5h6jJSIV5n6mJnHS5T7a0100O1O1O1O2N100O1O2N1O1O101N0000O10001O000000000O2O00000000001N1000000O1N3M2M3N2N20001O0O100000001O00000000001O000O1000001O00000000001O000O1000001O00000000001O00000O10001O00000000001O01O1O001O10O01O1O1O00100O001O1O00100O1O001O1O01O00000001O01O00000000001O000F:F:00O2O0O10000O101N10000O100O2O000O100O101O00000000001O00000000001O=Cl0TOl0TOk0UOl0TOl0TOl0TOc`c2"}, "label": "a screen with an image of a plate of hot dogs"}]}
{"id": 314254, "image": "COCO_train2014_000000314254.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a screen showing a plate of six hot dogs\"."}], "task": "refering", "answer_template": "The \"a screen showing a plate of six hot dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 76, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282], "bbox": [0.18784375, 0.17508196721311475, 0.68559375, 0.8131615925058547], "iscrowd": 0, "area": 71031.62849999999, "rle": {"size": [427, 640], "counts": "U[b15h<`0@`0@`0@`0_Oa0@`0@`0@`0@`0@`0@`0_O`0A<D<D<E;1O1OC=2N2N2N3L3N21O000010O01O00001O010O00001O00100O2N2N3M2N2O1N2N2N2N0001O000O1000001O0N2L4M3M3M4L3M3M3N20001O00000000001O0003M3M3M4M2M3M00000001O00000000001O00000000001O000000000YOh00O001O001O010O7I<D000001O0001O000001O00000000001O01O00000001O0000000YO\\JgId5Z6_JbIb5^6bJ^I]5c6fJYIZ5h6jJSIV5n6mJnHS5T7a0100O1O1O1O2N100O1O2N1O1O101N0000O10001O000000000O2O00000000001N1000000O1N3M2M3N2N20001O0O100000001O00000000001O000O1000001O00000000001O000O1000001O00000000001O00000O10001O00000000001O01O1O001O10O01O1O1O00100O001O1O00100O1O001O1O01O00000001O01O00000000001O000F:F:00O2O0O10000O101N10000O100O2O000O100O101O00000000001O00000000001O=Cl0TOl0TOk0UOl0TOl0TOl0TOc`c2"}, "label": "a screen showing a plate of six hot dogs"}]}
{"id": 355223, "image": "COCO_train2014_000000355223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby bear\"."}], "task": "refering", "answer_template": "The \"baby bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 398, 399, 400, 401, 402], "bbox": [0.32229687500000004, 0.4318688524590164, 0.860703125, 0.7853606557377049], "iscrowd": 0, "area": 49225.3269, "rle": {"size": [610, 640], "counts": "RTk3;db07I6J6K6I6J8I9F;E9G3N3L3M3M3N2M3N1O2O1N2O1O0O2O1O1N2O0O2O1O1N101N2O1O1N101N2O100O00101N100O1O100O1O100O2N100O100O1O1O1O2O0O1O1O1O1O100O1O2N1O1O100O1O2N1O100O1O1O101N100O100O100O101N100O100O100O10000O100O2O000O100O10000O10000O10000O100O10000O10O1O001O1O001O10O01O1O0010O0100O010O10O01000O10000O2O000O100O100O100O100O100O100O100O1O2O0O01M2N3M2N3M2N3N20O10000O10000O0100O00100O00100O00100O00100O00100O00100O001O10O01O10O01O100O1O100O100O1O101N1O10000000000001O0000000000000000000000010O0000000000000000O101O0000000O101O00000O10001O0O1O100O101N100O100O2N100O100O2O0O100O101N1O100O101N100O1O2N100O1O101N1O100O1O2O0O1O101N100O100000O1000000O1000000O100001O001gMd@j0\\?ROj@l0W?POo@m0Q?QOSAm0n>POWAm0i>QO[Am0h>nN\\AP1h>jN]AS1g>gNcAQ1`>jNoAi0W`0N2N2O1N2M2L5K7Jkdd1"}, "label": "baby bear"}]}
{"id": 355223, "image": "COCO_train2014_000000355223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small bear\"."}], "task": "refering", "answer_template": "The \"small bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 398, 399, 400, 401, 402], "bbox": [0.32229687500000004, 0.4318688524590164, 0.860703125, 0.7853606557377049], "iscrowd": 0, "area": 49225.3269, "rle": {"size": [610, 640], "counts": "RTk3;db07I6J6K6I6J8I9F;E9G3N3L3M3M3N2M3N1O2O1N2O1O0O2O1O1N2O0O2O1O1N101N2O1O1N101N2O100O00101N100O1O100O1O100O2N100O100O1O1O1O2O0O1O1O1O1O100O1O2N1O1O100O1O2N1O100O1O1O101N100O100O100O101N100O100O100O10000O100O2O000O100O10000O10000O10000O100O10000O10O1O001O1O001O10O01O1O0010O0100O010O10O01000O10000O2O000O100O100O100O100O100O100O100O1O2O0O01M2N3M2N3M2N3N20O10000O10000O0100O00100O00100O00100O00100O00100O00100O001O10O01O10O01O100O1O100O100O1O101N1O10000000000001O0000000000000000000000010O0000000000000000O101O0000000O101O00000O10001O0O1O100O101N100O100O2N100O100O2O0O100O101N1O100O101N100O1O2N100O1O101N1O100O1O2O0O1O101N100O100000O1000000O1000000O100001O001gMd@j0\\?ROj@l0W?POo@m0Q?QOSAm0n>POWAm0i>QO[Am0h>nN\\AP1h>jN]AS1g>gNcAQ1`>jNoAi0W`0N2N2O1N2M2L5K7Jkdd1"}, "label": "small bear"}]}
{"id": 355223, "image": "COCO_train2014_000000355223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"adult bear\"."}], "task": "refering", "answer_template": "The \"adult bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 391, 392, 393, 394, 395, 415, 416, 417], "bbox": [0.002140625, 0.0022459016393442627, 0.5161875, 0.844950819672131], "iscrowd": 0, "area": 115143.54520000001, "rle": {"size": [610, 640], "counts": "Sc0c1_a0l9TF`3`L1O0010O01O001O001O001O001O001O001O001O1O001O001O001O001O1O001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O00001O01O01O0000001O00001O0000000000000000000000000000000000000000O1O100O1O1O1O1O100O1O1O1O1O1O1M3N2M3N2M3M3N2M3N3L3M3N2M3K5K5J6K5J6A?M3M3L4M3M3M3eLbIoIa6R6aIkI_6V6bIhI^6X6eIeI[6\\6gIaIY6_6iI^IX6c6jIZIV6f6mIWIS6j6oISIQ6n6PJPIP6P7SJmHm5T7TJiHm5W7UJgHk5Z7VJdHj5]7XJ`Hh5`7ZJ^Hf5c7[J[He5e7]JXHd5h7^JVHb5k7_JSHa5m7aJQH_5P8aJoG_5Q8cJmG]5T8dJiG]5W8eJgG[5Z8fJkFS6U9m11O00001O001O0000000001O00000000000000000000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000001O1O100O1O1O00100O1O1O1O1O100O001O1O100O1O1O1O10O01O1O1O1O1O1O1O1O001O2N2N2N2N2N2N2N2N2N2N2N2N2O2M2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2O1N2N2N2N2M3N3L4M3L4M3L4M3L4M4K4M3L4M3L4M3L4M3L4M3oKbAW3b>bLfAZ3U?K5L4K4M4K5K5L4K5L4K5L4K5L4K5L4K5L3L5L4K7J8G_\\h5"}, "label": "adult bear"}]}
{"id": 355223, "image": "COCO_train2014_000000355223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bear is with her cub\"."}], "task": "refering", "answer_template": "The \"the bear is with her cub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 391, 392, 393, 394, 395, 415, 416, 417], "bbox": [0.002140625, 0.0022459016393442627, 0.5161875, 0.844950819672131], "iscrowd": 0, "area": 115143.54520000001, "rle": {"size": [610, 640], "counts": "Sc0c1_a0l9TF`3`L1O0010O01O001O001O001O001O001O001O001O1O001O001O001O001O1O001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O00001O01O01O0000001O00001O0000000000000000000000000000000000000000O1O100O1O1O1O1O100O1O1O1O1O1O1M3N2M3N2M3M3N2M3N3L3M3N2M3K5K5J6K5J6A?M3M3L4M3M3M3eLbIoIa6R6aIkI_6V6bIhI^6X6eIeI[6\\6gIaIY6_6iI^IX6c6jIZIV6f6mIWIS6j6oISIQ6n6PJPIP6P7SJmHm5T7TJiHm5W7UJgHk5Z7VJdHj5]7XJ`Hh5`7ZJ^Hf5c7[J[He5e7]JXHd5h7^JVHb5k7_JSHa5m7aJQH_5P8aJoG_5Q8cJmG]5T8dJiG]5W8eJgG[5Z8fJkFS6U9m11O00001O001O0000000001O00000000000000000000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000001O1O100O1O1O00100O1O1O1O1O100O001O1O100O1O1O1O10O01O1O1O1O1O1O1O1O001O2N2N2N2N2N2N2N2N2N2N2N2N2O2M2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2O1N2N2N2N2M3N3L4M3L4M3L4M3L4M4K4M3L4M3L4M3L4M3L4M3oKbAW3b>bLfAZ3U?K5L4K4M4K5K5L4K5L4K5L4K5L4K5L4K5L3L5L4K7J8G_\\h5"}, "label": "the bear is with her cub"}]}
{"id": 502679, "image": "COCO_train2014_000000502679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rear bicycle ride\"."}], "task": "refering", "answer_template": "The \"a rear bicycle ride\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 186, 187, 188, 209, 210, 211, 232, 233, 234, 254, 255, 256, 278, 279, 301, 302], "bbox": [0.076375, 0.3991022964509394, 0.21384375000000003, 0.8072651356993736], "iscrowd": 0, "area": 8155.98935, "rle": {"size": [479, 640], "counts": "\\Wg05g>7I7I7I:F9H6K4K6K5K3L3N1N3L4L4L4L4M3L4L4L4K5J6K5J6L4N2N2O1N2N2O001O1O000000000000000O100000001O00O001O1O001O1O001D<J5K6SOXE]Mk:a2h0L4L3M5J8B=0001O0000001N10001O00000O2M2N7I7I8H7I7I8HjR[7"}, "label": "a rear bicycle ride"}]}
{"id": 502679, "image": "COCO_train2014_000000502679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front tire of the bike that ' s hidden behind the red wheels in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the front tire of the bike that ' s hidden behind the red wheels in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 186, 187, 188, 209, 210, 211, 232, 233, 234, 254, 255, 256, 278, 279, 301, 302], "bbox": [0.076375, 0.3991022964509394, 0.21384375000000003, 0.8072651356993736], "iscrowd": 0, "area": 8155.98935, "rle": {"size": [479, 640], "counts": "\\Wg05g>7I7I7I:F9H6K4K6K5K3L3N1N3L4L4L4L4M3L4L4L4K5J6K5J6L4N2N2O1N2N2O001O1O000000000000000O100000001O00O001O1O001O1O001D<J5K6SOXE]Mk:a2h0L4L3M5J8B=0001O0000001N10001O00000O2M2N7I7I8H7I7I8HjR[7"}, "label": "the front tire of the bike that ' s hidden behind the red wheels in the right hand picture"}]}
{"id": 502679, "image": "COCO_train2014_000000502679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green bicycle has no basket\"."}], "task": "refering", "answer_template": "The \"the green bicycle has no basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 146, 147, 169, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 238, 239, 261, 262, 263, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379], "bbox": [0.267890625, 0.2910438413361169, 0.530765625, 0.9890396659707724], "iscrowd": 0, "area": 25914.676400000004, "rle": {"size": [479, 640], "counts": "g]`21g>9F:G9F:G9H8H8I7H8H7K6L4L2O1N2N2N2O1N2N2N2O1N1O2N2N2O1N2N200O1O100O1O1eITLc1l3\\NVLb1k3^NVL`1k3aNUL^1k3bNVL\\1k3eNULY1l3gNTLX1m3iNSLV1m3kNSLS1n3mNSLQ1n3PORLo0n3QOSLm0n3TORLj0o3VOQLj0o3WOQLg0P4ZOPLd0Q4]OoKb0Q4_OoK?R4AoK=R4DnK;R4FmK9T4HlK6U4KkK4U4LlK2U4OkKOV42jKMU45kKIV48iKGX49iKFW4;iKCX4>hK_OZ4b0fKYO^4h0bKTOa4l0`KQLDd1o4\\2\\KPLG_1P5b2YKmKI\\1R5h2UKkKKX1S5n2RKhKMU1U5S3oJgKNQ1V5Y3lJdK0n0X5_3gJcK2i0[5e3cJaK4g0Z5i3bJ^K6f0Z5l3aJ]K7c0Z5Q4_JZK9b0Z5U4]JXK<>Y5[4[JUK`0<V5`4YJTKe06T5f4XJRKh03R5l4VJPKj00R5Q5TJmJj02S5R5SJkJh03W5S5QJhJh04Y5T5PJgJf05[5U5nIeJf06^5V5lIcJe06a5Y5iI_Je09c5Z5gI\\Je09f5\\5]KUJd4m5V22N2N2N2M3N1O2N2dJaIT2a6XMXJa2h5[M\\Je2e5VM`Ji2`5TMdJk2]5PMhJo2X5mLlJS3U5iLoJV3Q5fLTKY3m4dLUK\\3l4bLVK]3j4bLXK]3i4aLXK_3h4aLYK^3h4`LZK_3f4`L[K`3f4^L\\Ka3d4_L]K`3d4^L]Kb3c4]L_Kb3b4\\L`Kc3`4]L`Kc3a4[LaKd3_4[LcKd3^4ZLcKf3]4ZLdKe3]4YLeKf3[4ZLeKf3\\4YLeKf3[4ZLfKe3[4ZLgKc3Z4]LgKa3[4^LfK`3[4_LgK_3[4`LfK^3[4bLgK[3[4dLfKZ3]4dLdKZ3`4cLaK[3b4cL`KZ3d4cL]K[3f4cL[K[3e4fL\\KX3e4gL]KW3d4iL^KS3d4mL]KQ3d4oL]Ko2c4RM^Kl2c4TM^Kj2b4VMaKf2a4ZMaKa2a4`MbK[2`4dMdKW2]4jMeKQ2d3hMUI7Z3l1c3oMQI5_3f1b3XNeH9m3Z1_3ZOcLa0`3_OeL8a3EgLLd31cL@g3=`5J7I6I7JoS\\4"}, "label": "the green bicycle has no basket"}]}
{"id": 502679, "image": "COCO_train2014_000000502679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bike closest to camera\"."}], "task": "refering", "answer_template": "The \"a green bike closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 146, 147, 169, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 238, 239, 261, 262, 263, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379], "bbox": [0.267890625, 0.2910438413361169, 0.530765625, 0.9890396659707724], "iscrowd": 0, "area": 25914.676400000004, "rle": {"size": [479, 640], "counts": "g]`21g>9F:G9F:G9H8H8I7H8H7K6L4L2O1N2N2N2O1N2N2N2O1N1O2N2N2O1N2N200O1O100O1O1eITLc1l3\\NVLb1k3^NVL`1k3aNUL^1k3bNVL\\1k3eNULY1l3gNTLX1m3iNSLV1m3kNSLS1n3mNSLQ1n3PORLo0n3QOSLm0n3TORLj0o3VOQLj0o3WOQLg0P4ZOPLd0Q4]OoKb0Q4_OoK?R4AoK=R4DnK;R4FmK9T4HlK6U4KkK4U4LlK2U4OkKOV42jKMU45kKIV48iKGX49iKFW4;iKCX4>hK_OZ4b0fKYO^4h0bKTOa4l0`KQLDd1o4\\2\\KPLG_1P5b2YKmKI\\1R5h2UKkKKX1S5n2RKhKMU1U5S3oJgKNQ1V5Y3lJdK0n0X5_3gJcK2i0[5e3cJaK4g0Z5i3bJ^K6f0Z5l3aJ]K7c0Z5Q4_JZK9b0Z5U4]JXK<>Y5[4[JUK`0<V5`4YJTKe06T5f4XJRKh03R5l4VJPKj00R5Q5TJmJj02S5R5SJkJh03W5S5QJhJh04Y5T5PJgJf05[5U5nIeJf06^5V5lIcJe06a5Y5iI_Je09c5Z5gI\\Je09f5\\5]KUJd4m5V22N2N2N2M3N1O2N2dJaIT2a6XMXJa2h5[M\\Je2e5VM`Ji2`5TMdJk2]5PMhJo2X5mLlJS3U5iLoJV3Q5fLTKY3m4dLUK\\3l4bLVK]3j4bLXK]3i4aLXK_3h4aLYK^3h4`LZK_3f4`L[K`3f4^L\\Ka3d4_L]K`3d4^L]Kb3c4]L_Kb3b4\\L`Kc3`4]L`Kc3a4[LaKd3_4[LcKd3^4ZLcKf3]4ZLdKe3]4YLeKf3[4ZLeKf3\\4YLeKf3[4ZLfKe3[4ZLgKc3Z4]LgKa3[4^LfK`3[4_LgK_3[4`LfK^3[4bLgK[3[4dLfKZ3]4dLdKZ3`4cLaK[3b4cL`KZ3d4cL]K[3f4cL[K[3e4fL\\KX3e4gL]KW3d4iL^KS3d4mL]KQ3d4oL]Ko2c4RM^Kl2c4TM^Kj2b4VMaKf2a4ZMaKa2a4`MbK[2`4dMdKW2]4jMeKQ2d3hMUI7Z3l1c3oMQI5_3f1b3XNeH9m3Z1_3ZOcLa0`3_OeL8a3EgLLd31cL@g3=`5J7I6I7JoS\\4"}, "label": "a green bike closest to camera"}]}
{"id": 502679, "image": "COCO_train2014_000000502679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cycle in the middle with other cycles\"."}], "task": "refering", "answer_template": "The \"cycle in the middle with other cycles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 118, 119, 122, 123, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 302, 303, 325, 326], "bbox": [0.12559375, 0.2584968684759916, 0.41409375, 0.8412734864300626], "iscrowd": 0, "area": 22243.52835, "rle": {"size": [479, 640], "counts": "]PV12l>2O2N2M3N1O2N2O0O2O1N101N2O1N13NnHHZO9c0H^O8`0JA6;LF57LK33NN3NO32JO80dNS1]1nN]NV1e1iNVN[1k1eNPN_1Q2bNiMb1Y2]NcMf1^2ZN]Mj1e2UNWMn1j2RNQMR2P3nMkLV2W3iMcL[2`3dM[L`2h3]MSLg2P4XMgKP3[4oL]KW3f4gLSK`3P5^LhJi3Z5VL[JS4g5lKUJX4l5hKPJZ4R6fKjI]4V6cKgI_4[6aKbI`4`6`K\\Ib4e6^KXIe4i6[KSIg4o6YKmHi4T7XKhHh4\\7XK`Hi4c7WKYHi4j7WKPHm4S8SKfGP5^8PK]GQ5f8PKUGQ5o8<3M3M3M3M3M3N2N2N2B>M3K500000000001O001N100O2N101N101N100O2O0O2O4K2O1N2O1O1N2O0O10O00100O10O01O10O010O00101N1O010O1O100O100O010O10000O100O01000O100O1000oJdF^4\\9bKfF]4Z9bKgF^4X9bKjFm0KR2[9PMlFb05]2o8PMmF9>g2d8PMPGNg0Q3Y8PMQGDR1Z3m7RMaHn2_7QMcHn2]7QMeHn2Z7RMgHn2Y7QMhHo2W7QMkHn2U7QMlHn2U7QMlHo2S7QMnHn2S7QMoHl2R7TMoHk2R7TMPIj2Q7UMPIi2Q7WMQIg2P7XMQIf2P7ZMRIb2Q7]MPIa2R7^MPI^2R7bMoH\\2S7cMoHY2S7gMnHW2o9L3N3L3N3L3N3M2N3M3M2N3M3M3M3O1N2O1L4M3L3MaU_5"}, "label": "cycle in the middle with other cycles"}]}
{"id": 502679, "image": "COCO_train2014_000000502679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle with red rims\"."}], "task": "refering", "answer_template": "The \"the bicycle with red rims\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 118, 119, 122, 123, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 302, 303, 325, 326], "bbox": [0.12559375, 0.2584968684759916, 0.41409375, 0.8412734864300626], "iscrowd": 0, "area": 22243.52835, "rle": {"size": [479, 640], "counts": "]PV12l>2O2N2M3N1O2N2O0O2O1N101N2O1N13NnHHZO9c0H^O8`0JA6;LF57LK33NN3NO32JO80dNS1]1nN]NV1e1iNVN[1k1eNPN_1Q2bNiMb1Y2]NcMf1^2ZN]Mj1e2UNWMn1j2RNQMR2P3nMkLV2W3iMcL[2`3dM[L`2h3]MSLg2P4XMgKP3[4oL]KW3f4gLSK`3P5^LhJi3Z5VL[JS4g5lKUJX4l5hKPJZ4R6fKjI]4V6cKgI_4[6aKbI`4`6`K\\Ib4e6^KXIe4i6[KSIg4o6YKmHi4T7XKhHh4\\7XK`Hi4c7WKYHi4j7WKPHm4S8SKfGP5^8PK]GQ5f8PKUGQ5o8<3M3M3M3M3M3N2N2N2B>M3K500000000001O001N100O2N101N101N100O2O0O2O4K2O1N2O1O1N2O0O10O00100O10O01O10O010O00101N1O010O1O100O100O010O10000O100O01000O100O1000oJdF^4\\9bKfF]4Z9bKgF^4X9bKjFm0KR2[9PMlFb05]2o8PMmF9>g2d8PMPGNg0Q3Y8PMQGDR1Z3m7RMaHn2_7QMcHn2]7QMeHn2Z7RMgHn2Y7QMhHo2W7QMkHn2U7QMlHn2U7QMlHo2S7QMnHn2S7QMoHl2R7TMoHk2R7TMPIj2Q7UMPIi2Q7WMQIg2P7XMQIf2P7ZMRIb2Q7]MPIa2R7^MPI^2R7bMoH\\2S7cMoHY2S7gMnHW2o9L3N3L3N3L3N3M2N3M3M2N3M3M3M3O1N2O1L4M3L3MaU_5"}, "label": "the bicycle with red rims"}]}
{"id": 486300, "image": "COCO_train2014_000000486300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller pizza\"."}], "task": "refering", "answer_template": "The \"the smaller pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275], "bbox": [0.606421875, 0.1362708333333333, 1.0, 0.6722916666666667], "iscrowd": 0, "area": 36868.6067, "rle": {"size": [480, 640], "counts": "goe51m>3N2M4M2M3N3L3N2N3L3N2M4M2M3N3L3N2M3O1N101O1N2O0O2O1N101N2O1O0O2O1N101N2O1N101N2O2N6I6K6I6K6I7J2M2O2N2M2O2M3N2M2O2M3N1O1O0000000O1000000000000000000O1000000000000000000O100000001O000000000O2O000000001O00000000001O0O100000001O00001O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N7I7I6J7H7J7I6J3M00001O001O00001O0O2O00001O001O00001O001O000O2O03N2M2N3N2M3N1N3M3N2M3N1N3M3N2M2O2M3M3N1N3M3N2M3M2N2N3M2N2M4M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M3L3N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N3MTH"}, "label": "the smaller pizza"}]}
{"id": 486300, "image": "COCO_train2014_000000486300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of pizza partly eaten on a plate with a blue color spill on it\"."}], "task": "refering", "answer_template": "The \"a piece of pizza partly eaten on a plate with a blue color spill on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275], "bbox": [0.606421875, 0.1362708333333333, 1.0, 0.6722916666666667], "iscrowd": 0, "area": 36868.6067, "rle": {"size": [480, 640], "counts": "goe51m>3N2M4M2M3N3L3N2N3L3N2M4M2M3N3L3N2M3O1N101O1N2O0O2O1N101N2O1O0O2O1N101N2O1N101N2O2N6I6K6I6K6I7J2M2O2N2M2O2M3N2M2O2M3N1O1O0000000O1000000000000000000O1000000000000000000O100000001O000000000O2O000000001O00000000001O0O100000001O00001O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N7I7I6J7H7J7I6J3M00001O001O00001O0O2O00001O001O00001O001O000O2O03N2M2N3N2M3N1N3M3N2M3N1N3M3N2M2O2M3M3N1N3M3N2M3M2N2N3M2N2M4M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M3L3N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N3MTH"}, "label": "a piece of pizza partly eaten on a plate with a blue color spill on it"}]}
{"id": 486300, "image": "COCO_train2014_000000486300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"food with smiling face on the left\"."}], "task": "refering", "answer_template": "The \"food with smiling face on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 301, 302, 303, 304, 305, 325, 326, 327, 349], "bbox": [0.0, 0.1929375, 0.464125, 0.9105833333333334], "iscrowd": 0, "area": 61936.345550000005, "rle": {"size": [480, 640], "counts": "]42m>6J6K5J5K6K5J6J5L5J6J5K6K5J6J5L3L2N3N1N3M2N3N1N3M4M4K5K5L5J5K5L4K5K5K6K4K5K5L8G9G9H8G9G9G9H8G8H2O0O2N2O1O0O2O1O1N101O1O1N101O1N2O001N2O1O001N2O1O0O2O1O1O0O2O1O1N101O1O1N3N1O1N3N1O1O2M2O2N1O1O2N1O1O2O0O2N1O1O2O0O2N1O1O2O0O1O2N1O2N100O2N1O1O2O0O1OO100N2O2N1O1N2O1O2N1N2O1O1O2M2O1O2N1N2O2N1O1N3N1O2M2O1O2N1N3N1O1O2M2O1O2N1N3N1O1O2M2O1O2M2O2M2O1N3N1N2O2M2O2M2O1N3N1O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2N1N3N1N2O2N10001N10001O000O101O00000O2O0000001N100O2N1O1N2O2M2N2O1N3N1N2O1N3N1N2O2M2O1N2N3N1N2O1N3N1N2O1N3N1N2O1M4K4M3L4M3L5L3L4M3L5L3M3O1N3N1N2O1O1N3N1O1N2O3L3N3M2M3N3L3N3M2M3N3L3N3M2M3N3L3N3I6J6J7I6JgcP5"}, "label": "food with smiling face on the left"}]}
{"id": 289696, "image": "COCO_train2014_000000289696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult grey long - haired cat sitting in an orange chair\"."}], "task": "refering", "answer_template": "The \"an adult grey long - haired cat sitting in an orange chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 142, 143, 144, 145, 166, 167, 168, 169, 189, 190, 191, 192, 213, 214, 215, 237, 238, 260, 261], "bbox": [0.209, 0.34752083333333333, 0.386390625, 0.7027916666666667], "iscrowd": 0, "area": 8307.8264, "rle": {"size": [480, 640], "counts": "PPo13k>3O1O1O001O1O1O1N1O2N1N3N2N1N3N2N1N3N2N1N3O1N2O001N2O1N2\\CeNa;[1\\DiNc;X1ZDkNd;W1ZDkNe;V1XDmNg;V2O2O1O1O0O2O1O1O001O01O000000001O000010O01O1O001O1O001O10O01O001N2O001N2O000O2O010O001O00aCoMo;i2H5L3oDeLU:`3bEfL]:P4O1N2000O101N100O1O2M2N2O1N3RMVE`1l:\\NXEb1j:ZNYEe1i:PNaEo1a:fMjEX2`;O001O1O0N4L4M2M4UOk0L4K8D]ng5"}, "label": "an adult grey long - haired cat sitting in an orange chair"}]}
{"id": 289696, "image": "COCO_train2014_000000289696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest cat sitting on the left chair\"."}], "task": "refering", "answer_template": "The \"the biggest cat sitting on the left chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 142, 143, 144, 145, 166, 167, 168, 169, 189, 190, 191, 192, 213, 214, 215, 237, 238, 260, 261], "bbox": [0.209, 0.34752083333333333, 0.386390625, 0.7027916666666667], "iscrowd": 0, "area": 8307.8264, "rle": {"size": [480, 640], "counts": "PPo13k>3O1O1O001O1O1O1N1O2N1N3N2N1N3N2N1N3N2N1N3O1N2O001N2O1N2\\CeNa;[1\\DiNc;X1ZDkNd;W1ZDkNe;V1XDmNg;V2O2O1O1O0O2O1O1O001O01O000000001O000010O01O1O001O1O001O10O01O001N2O001N2O000O2O010O001O00aCoMo;i2H5L3oDeLU:`3bEfL]:P4O1N2000O101N100O1O2M2N2O1N3RMVE`1l:\\NXEb1j:ZNYEe1i:PNaEo1a:fMjEX2`;O001O1O0N4L4M2M4UOk0L4K8D]ng5"}, "label": "the biggest cat sitting on the left chair"}]}
{"id": 289696, "image": "COCO_train2014_000000289696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden armchair with ginger cushions and a small gray kitten atop it\"."}], "task": "refering", "answer_template": "The \"a wooden armchair with ginger cushions and a small gray kitten atop it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 292, 293, 294, 295, 296, 311, 312, 318, 341, 364], "bbox": [0.534265625, 0.1993125, 0.964046875, 0.8959375], "iscrowd": 0, "area": 53709.354349999994, "rle": {"size": [480, 640], "counts": "oaP55i><E;E;E;D<\\DUN^9U2UFnMk9]2gEfMY:e2YE^Md:P3nDSMo:`3Mi0VO5L4L4L4L4L4L4L4L400O10VOi0TOm0G91N10000O2O000O100O2O000O101O0O10001N100O101O0O10000O2N1N2N2N3N1N2N2N3M2O1N2N2TNlJjIW5U6mJeIU5Z6PKaIQ5_6RK\\IP5d6SKVIQ5i6PKQIU5n6mJlHX5T7iJgH[5Y7fJaH`5^7aJ]Hc5b7n00001N1000000O2O00000O2O000O101O0000001O000000010O0000001O01O0001O00000010O00000001O0001O01O0000001O01O0001O000000010O0000001O0001O01O0000001O0001O0001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0001O00010O000010O0mJXHn2h7lL^HT3b7fLdH[3\\7]LkHc3U7WLQIi3o6PLXIQ4g6iK`IV4`6dKfI\\4[6]KkId4T6VKRJj4R800001J5K5K6J5K5K5K6J5K5K6J5K5K6K4L4M3M4K4M3L4M4K4M3L4M4K4M3L4M4K4M3L4M4K4M3M3L5L3L4M3M4L3N2M3N2M4M2M3N2M4L3J6lNT1lNU1K4O10001N100O101N1O1N2O2M2N4L3M4L4L4Mhd:"}, "label": "a wooden armchair with ginger cushions and a small gray kitten atop it"}]}
{"id": 289696, "image": "COCO_train2014_000000289696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a tiny kitten on it\"."}], "task": "refering", "answer_template": "The \"a chair with a tiny kitten on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 292, 293, 294, 295, 296, 311, 312, 318, 341, 364], "bbox": [0.534265625, 0.1993125, 0.964046875, 0.8959375], "iscrowd": 0, "area": 53709.354349999994, "rle": {"size": [480, 640], "counts": "oaP55i><E;E;E;D<\\DUN^9U2UFnMk9]2gEfMY:e2YE^Md:P3nDSMo:`3Mi0VO5L4L4L4L4L4L4L4L400O10VOi0TOm0G91N10000O2O000O100O2O000O101O0O10001N100O101O0O10000O2N1N2N2N3N1N2N2N3M2O1N2N2TNlJjIW5U6mJeIU5Z6PKaIQ5_6RK\\IP5d6SKVIQ5i6PKQIU5n6mJlHX5T7iJgH[5Y7fJaH`5^7aJ]Hc5b7n00001N1000000O2O00000O2O000O101O0000001O000000010O0000001O01O0001O00000010O00000001O0001O01O0000001O01O0001O000000010O0000001O0001O01O0000001O0001O0001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0001O00010O000010O0mJXHn2h7lL^HT3b7fLdH[3\\7]LkHc3U7WLQIi3o6PLXIQ4g6iK`IV4`6dKfI\\4[6]KkId4T6VKRJj4R800001J5K5K6J5K5K5K6J5K5K6J5K5K6K4L4M3M4K4M3L4M4K4M3L4M4K4M3L4M4K4M3L4M4K4M3M3L5L3L4M3M4L3N2M3N2M4M2M3N2M4L3J6lNT1lNU1K4O10001N100O101N1O1N2O2M2N4L3M4L4L4Mhd:"}, "label": "a chair with a tiny kitten on it"}]}
{"id": 289696, "image": "COCO_train2014_000000289696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair on wood floor with grey tabby cat\"."}], "task": "refering", "answer_template": "The \"a brown chair on wood floor with grey tabby cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 145, 146, 147, 162, 163, 164, 165, 166, 169, 170, 185, 186, 187, 188, 189, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 285, 301, 302, 308, 324, 325], "bbox": [0.059109375, 0.18810416666666668, 0.43912500000000004, 0.8679166666666667], "iscrowd": 0, "area": 43777.53959999998, "rle": {"size": [480, 640], "counts": "hma04k>5K3M2N2N1O2QDDW9>fFOo81oF<e8FXGf0]8\\O`Go0W8ROfGY1P8iNmG`1l7aNQHh1h7ZNTHP2d7QNSH^2e7eMVHd2d7]MYHh2d7dMmGc2o7jMbG\\2Y8Z2L4L4K5L3M4L4K5L3M4L3L5L3M4L3L5L3UJdGl4`8hJhGX5P9O2O3M4L3LfMmJiJS5W5QKeJo4\\5UK^Jk4c5[KWJe4i5aKPJ`4P6gKiIY4W6mKbIS4_6QL]Io3c6SL[Im3d6VLXIl3h6VLSIl3n6VLgHS4Y7j100000000000O100000O1000000000O100000000000O1000O10000000000000000O01000000000000^KdH`1\\7_NfH`1Y7`NiH_1W7`NkH^1V7aNlH^1T7aNnH^1R7aNPI^1P7`NSI_1m6`NUI_1k6`NWI_1i6`NYI_1g6`NZI`1e6aN\\I]1e6bN]I]1c6bN_I]1a6bNaI]1_6cNaI]1_6bNcI]1]6bNfI\\1Z6cNiI[1W6eNjIZ1V6eNlIY1U6gNlIX1T6gNmIY1S6gNnIX1R6gNoIY1P6hNQJW1o5hNRJX1n5hNRJX1n5gNSJZNdNT2Y7BTJTNnNS2o6ISJPNSOV2j6IUK7k4IVK6j4JWK5i4JXK6h4JYK5g4J[K5e4K[K5e4K\\K3e4L\\K4d4M\\K2c4O]K1c40]KOc41]KOc41]KOc42]KMc43]KMc44]KJd46\\KJd46]KIc48\\KHd48\\KHd49[KGe49[KFf4;YKEg4;YKEg4<XKDh4<XKCh4?WKAi4?WKAi4`0VK@j4a0UK^Ol4b0TK^Ol4c0RK^On4b0RK^Oo4b0PK]OQ5c0oJ]OQ5c0oJ]OQ5d0nJ\\OR5d0oJZOS5f0nJXOR5h0QKaM]MX1b7W1eKaN[4`1fK]N[4c1gK[NZ4e1gKYNY4g1iKWNW4j1kKRNV4n1lKPNT4Q2nKlMR4T2QL[M^4e2n31O1O1O010O1O1O1O012M4L3M3M6K5J6VE\\Ll9]4O1O1O100OgLkEh1T:RNUFl1j9nM^FQ2V1SM\\6d0gHW2i0XMa6:nH]2`0YMc68QIc25TMl68RIi2JoLU76TIo2AjL]75VIR3YOiLe71TIX3ROhLS8FmHe3lNdLo9]3oEcLR:[3oEeLR:Y3oEfLS:X3nEhLS:W3lEjLU:W3iEhLY:Z3dEfL]:\\3`EdL`:_3]E`Le:c3VE^Lk:W4E>WFTKa8f5000000000:FO001O1O1O1O3M5J6gNcFdLc9d2UGWMW9T2WGeMX9e1VGUNo<J5K^mW5"}, "label": "a brown chair on wood floor with grey tabby cat"}]}
{"id": 289696, "image": "COCO_train2014_000000289696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair with the bigger cat on it\"."}], "task": "refering", "answer_template": "The \"the chair with the bigger cat on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 145, 146, 147, 162, 163, 164, 165, 166, 169, 170, 185, 186, 187, 188, 189, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 285, 301, 302, 308, 324, 325], "bbox": [0.059109375, 0.18810416666666668, 0.43912500000000004, 0.8679166666666667], "iscrowd": 0, "area": 43777.53959999998, "rle": {"size": [480, 640], "counts": "hma04k>5K3M2N2N1O2QDDW9>fFOo81oF<e8FXGf0]8\\O`Go0W8ROfGY1P8iNmG`1l7aNQHh1h7ZNTHP2d7QNSH^2e7eMVHd2d7]MYHh2d7dMmGc2o7jMbG\\2Y8Z2L4L4K5L3M4L4K5L3M4L3L5L3M4L3L5L3UJdGl4`8hJhGX5P9O2O3M4L3LfMmJiJS5W5QKeJo4\\5UK^Jk4c5[KWJe4i5aKPJ`4P6gKiIY4W6mKbIS4_6QL]Io3c6SL[Im3d6VLXIl3h6VLSIl3n6VLgHS4Y7j100000000000O100000O1000000000O100000000000O1000O10000000000000000O01000000000000^KdH`1\\7_NfH`1Y7`NiH_1W7`NkH^1V7aNlH^1T7aNnH^1R7aNPI^1P7`NSI_1m6`NUI_1k6`NWI_1i6`NYI_1g6`NZI`1e6aN\\I]1e6bN]I]1c6bN_I]1a6bNaI]1_6cNaI]1_6bNcI]1]6bNfI\\1Z6cNiI[1W6eNjIZ1V6eNlIY1U6gNlIX1T6gNmIY1S6gNnIX1R6gNoIY1P6hNQJW1o5hNRJX1n5hNRJX1n5gNSJZNdNT2Y7BTJTNnNS2o6ISJPNSOV2j6IUK7k4IVK6j4JWK5i4JXK6h4JYK5g4J[K5e4K[K5e4K\\K3e4L\\K4d4M\\K2c4O]K1c40]KOc41]KOc41]KOc42]KMc43]KMc44]KJd46\\KJd46]KIc48\\KHd48\\KHd49[KGe49[KFf4;YKEg4;YKEg4<XKDh4<XKCh4?WKAi4?WKAi4`0VK@j4a0UK^Ol4b0TK^Ol4c0RK^On4b0RK^Oo4b0PK]OQ5c0oJ]OQ5c0oJ]OQ5d0nJ\\OR5d0oJZOS5f0nJXOR5h0QKaM]MX1b7W1eKaN[4`1fK]N[4c1gK[NZ4e1gKYNY4g1iKWNW4j1kKRNV4n1lKPNT4Q2nKlMR4T2QL[M^4e2n31O1O1O010O1O1O1O012M4L3M3M6K5J6VE\\Ll9]4O1O1O100OgLkEh1T:RNUFl1j9nM^FQ2V1SM\\6d0gHW2i0XMa6:nH]2`0YMc68QIc25TMl68RIi2JoLU76TIo2AjL]75VIR3YOiLe71TIX3ROhLS8FmHe3lNdLo9]3oEcLR:[3oEeLR:Y3oEfLS:X3nEhLS:W3lEjLU:W3iEhLY:Z3dEfL]:\\3`EdL`:_3]E`Le:c3VE^Lk:W4E>WFTKa8f5000000000:FO001O1O1O1O3M5J6gNcFdLc9d2UGWMW9T2WGeMX9e1VGUNo<J5K^mW5"}, "label": "the chair with the bigger cat on it"}]}
{"id": 396193, "image": "COCO_train2014_000000396193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of grapefruit on the right\"."}], "task": "refering", "answer_template": "The \"the half of grapefruit on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 333, 334], "bbox": [0.22134375, 0.18587822014051522, 0.7346875, 0.9505854800936767], "iscrowd": 0, "area": 77836.06589999999, "rle": {"size": [427, 640], "counts": "d]k16Q=8I7I6I8I7H8I6J7H8I7H7J7I5J5L4M3M3N2M3N2M3N2M4M2M3N2M3N2M3N2M3N2M3M3N2M3N2M4M2N2N2N2O1N2N2N2O1N2N3M2O1N2N2N2O1N3M2N2O1N2N2N2O2M2N2N2O1N2N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N101N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O0O2O1N2O1O1N2O1O0O2O1O1N2O001N101O0O2O0O2O001N101O0O2O001N101O0O2O001N10001N10010O00001O00001O00001O01O01O00001O00001O000010O000001O0000001O000000010O00000000001O000001O010O001O1O010O001O0010O01O001N101N101O1N101N101O0O2O0O2O001N101N2O001N101N101O0O2O1N101O1N101N101N2O001N2O1N101N2N2O0O2N2O1N1O2O1N2N101N2N2O1N1O2O1N2N101N2N2O0O2N2O1N1O2O1N2N2O1N2N2O2M2N2O1N2N2O1N2N2M3M3M3N2M4L3M3N2M3M3M3N2M3M4L4M4K4L4L4M4K4L4M4K4L4L5L3K5K6I6K5K6I6K5J?BP1oNbdV2"}, "label": "the half of grapefruit on the right"}]}
{"id": 396193, "image": "COCO_train2014_000000396193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"full half fruit\"."}], "task": "refering", "answer_template": "The \"full half fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 333, 334], "bbox": [0.22134375, 0.18587822014051522, 0.7346875, 0.9505854800936767], "iscrowd": 0, "area": 77836.06589999999, "rle": {"size": [427, 640], "counts": "d]k16Q=8I7I6I8I7H8I6J7H8I7H7J7I5J5L4M3M3N2M3N2M3N2M4M2M3N2M3N2M3N2M3N2M3M3N2M3N2M4M2N2N2N2O1N2N2N2O1N2N3M2O1N2N2N2O1N3M2N2O1N2N2N2O2M2N2N2O1N2N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N101N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O0O2O1N2O1O1N2O1O0O2O1O1N2O001N101O0O2O0O2O001N101O0O2O001N101O0O2O001N10001N10010O00001O00001O00001O01O01O00001O00001O000010O000001O0000001O000000010O00000000001O000001O010O001O1O010O001O0010O01O001N101N101O1N101N101O0O2O0O2O001N101N2O001N101N101O0O2O1N101O1N101N101N2O001N2O1N101N2N2O0O2N2O1N1O2O1N2N101N2N2O1N1O2O1N2N101N2N2O0O2N2O1N1O2O1N2N2O1N2N2O2M2N2O1N2N2O1N2N2M3M3M3N2M4L3M3N2M3M3M3N2M3M4L4M4K4L4L4M4K4L4M4K4L4L5L3K5K6I6K5K6I6K5J?BP1oNbdV2"}, "label": "full half fruit"}]}
{"id": 396193, "image": "COCO_train2014_000000396193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a piece of fruit facing away from the shot\"."}], "task": "refering", "answer_template": "The \"half of a piece of fruit facing away from the shot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.536463700234192, 0.28885937500000003, 0.9882435597189696], "iscrowd": 0, "area": 30809.227250000007, "rle": {"size": [427, 640], "counts": "i;\\1X9g2dN\\10000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10000000000001O000000000000001O01O000000000001O000000000000001O000000000000001O000000000UJeHf5`701O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M2N2N2N3M4L4J7H7J7H8H7J7H7I8I6YO]\\m5"}, "label": "half of a piece of fruit facing away from the shot"}]}
{"id": 396193, "image": "COCO_train2014_000000396193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blood orange with the purple insides\"."}], "task": "refering", "answer_template": "The \"blood orange with the purple insides\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.536463700234192, 0.28885937500000003, 0.9882435597189696], "iscrowd": 0, "area": 30809.227250000007, "rle": {"size": [427, 640], "counts": "i;\\1X9g2dN\\10000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10000000000001O000000000000001O01O000000000001O000000000000001O000000000000001O000000000UJeHf5`701O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M2N2N2N3M4L4J7H7J7H8H7J7H7I8I6YO]\\m5"}, "label": "blood orange with the purple insides"}]}
{"id": 396193, "image": "COCO_train2014_000000396193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"triangle slice of fruit on plate\"."}], "task": "refering", "answer_template": "The \"triangle slice of fruit on plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 140, 141, 142, 143, 163, 164, 165, 166], "bbox": [0.10570312500000001, 0.0025995316159250588, 0.42804687500000005, 0.5012880562060891], "iscrowd": 0, "area": 21162.7859, "rle": {"size": [427, 640], "counts": "aal08i<8G;F;E:G9L5K4L4L5K4L4L4L5K4L4L5K4L4M4M2M3N3M2M3N3M2M3N3M2M3N3M2M3N3M2M3N2N2M3N2N2M2O2N2M3N2N2M3N2N2M3N2N2M3N2N2M2O21O001O001O001O1O001O001O001N101O001O001O1O001O0K6G8I8G8I8I6O2M3N1N3M2O2M2O2M2O2M2N3N1N3N2M2O2M2O2M2N3N1N3N1O2N1O2O1N1O2N1O2N1O2N1O2N1O2N2N1O2O0O2N1O2N1O2N1O2N1O2N2N1O2O001N101O001N101O0O2O1O0O2O001O0O2O001N101O1N101O001N101O0O2O001N101O1O0O2O001N101O0O2O001N101N2N101N101N101N101N101N2N101N10enh4"}, "label": "triangle slice of fruit on plate"}]}
{"id": 396193, "image": "COCO_train2014_000000396193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of fruit with greenish yellow skin\"."}], "task": "refering", "answer_template": "The \"a piece of fruit with greenish yellow skin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 140, 141, 142, 143, 163, 164, 165, 166], "bbox": [0.10570312500000001, 0.0025995316159250588, 0.42804687500000005, 0.5012880562060891], "iscrowd": 0, "area": 21162.7859, "rle": {"size": [427, 640], "counts": "aal08i<8G;F;E:G9L5K4L4L5K4L4L4L5K4L4L5K4L4M4M2M3N3M2M3N3M2M3N3M2M3N3M2M3N3M2M3N2N2M3N2N2M2O2N2M3N2N2M3N2N2M3N2N2M3N2N2M2O21O001O001O001O1O001O001O001N101O001O001O1O001O0K6G8I8G8I8I6O2M3N1N3M2O2M2O2M2O2M2N3N1N3N2M2O2M2O2M2N3N1N3N1O2N1O2O1N1O2N1O2N1O2N1O2N1O2N2N1O2O0O2N1O2N1O2N1O2N1O2N2N1O2O001N101O001N101O0O2O1O0O2O001O0O2O001N101O1N101O001N101O0O2O001N101O1O0O2O001N101O0O2O001N101N2N101N101N101N101N101N2N101N10enh4"}, "label": "a piece of fruit with greenish yellow skin"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cream colored vase between green vase and blue vase\"."}], "task": "refering", "answer_template": "The \"cream colored vase between green vase and blue vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 194, 195, 196, 197, 218, 219, 220, 242, 243, 265, 266, 288], "bbox": [0.37753125, 0.024729166666666667, 0.67078125, 0.7236041666666666], "iscrowd": 0, "area": 38320.36644999998, "rle": {"size": [480, 640], "counts": "Yda32f>8F:F:F;E:F:F:D<A?XMoLRIa3m6bLjHd3V7_LbHf3^7]LYHi3g7ZLQHk3o7XLhGn3X8UL`GQ4_8RLXGT4h8oKPGV4P9m00000O2O000000001O00000000001N10000000001O00000gGSJ_7m5^HYJ_7g5_H_J]7b5_HdJ^7\\5`HjJ[7W5bHmJ]7S5aHPK^7P5`HRK`7W60001O0000001O000O101O0000001O0000001O00000O101O0000001O0000001O000O101O001O3M3M2N3M2N3M3M2N3M3M2N3[KWGm2l8oLUGQ3m8kLTGV3o8eLSG[3P9aLPG`3R9\\LPGd3S9XLmFi3V9RLlFo3U9nKkFS4X9hKjFX4Y9cKiF]4Y9`KgFa4Y5\\KZNN_Lf4W5_KfNb4Z1bKbN^4^1eK^N\\4b1hKZNX4e1lKWNU4i1oKSNQ4m1RLPNn3P2VLkMk3U2WLiMi3W2YLfMh3Z2ZLdMf3\\2\\LaMe3_2]L_Mc3`2`L]Mb3b2`L\\M`3d2bLZM^3f2dLWM]3i2eLUM[3k2gLRMZ3n2hLoLY3Q3iLlLX3S3kLjLV3V3lLgLU3Y3mLdLT3\\3oL`LR3`3PM]LQ3c3QMZLP3f3RMWLo2i3SMjKX3U4kLYKe3g4]LhJR4X5k201O000000000O100000000000000O1000000dK_FY3a9fLdFV3]9hLhFT3X9kLmFQ3S9mLSGn2n8QMWGk2i8TM[Gi2e8VM`Gf2`8XMfGd2Z8[MkGa2U8^MPH^2P8aMUH[2k7cM[HX2f7gM_HU2a7jMdHR2\\7mMiHo1W7oMoHm1R7QNSIk1m6TNXIh1h6WN]Ie1c6YNbI\\1g6bN^Ik0Q7TOTI9[7FP4N`lR3"}, "label": "cream colored vase between green vase and blue vase"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tall vase behind two others\"."}], "task": "refering", "answer_template": "The \"tall vase behind two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 194, 195, 196, 197, 218, 219, 220, 242, 243, 265, 266, 288], "bbox": [0.37753125, 0.024729166666666667, 0.67078125, 0.7236041666666666], "iscrowd": 0, "area": 38320.36644999998, "rle": {"size": [480, 640], "counts": "Yda32f>8F:F:F;E:F:F:D<A?XMoLRIa3m6bLjHd3V7_LbHf3^7]LYHi3g7ZLQHk3o7XLhGn3X8UL`GQ4_8RLXGT4h8oKPGV4P9m00000O2O000000001O00000000001N10000000001O00000gGSJ_7m5^HYJ_7g5_H_J]7b5_HdJ^7\\5`HjJ[7W5bHmJ]7S5aHPK^7P5`HRK`7W60001O0000001O000O101O0000001O0000001O00000O101O0000001O0000001O000O101O001O3M3M2N3M2N3M3M2N3M3M2N3[KWGm2l8oLUGQ3m8kLTGV3o8eLSG[3P9aLPG`3R9\\LPGd3S9XLmFi3V9RLlFo3U9nKkFS4X9hKjFX4Y9cKiF]4Y9`KgFa4Y5\\KZNN_Lf4W5_KfNb4Z1bKbN^4^1eK^N\\4b1hKZNX4e1lKWNU4i1oKSNQ4m1RLPNn3P2VLkMk3U2WLiMi3W2YLfMh3Z2ZLdMf3\\2\\LaMe3_2]L_Mc3`2`L]Mb3b2`L\\M`3d2bLZM^3f2dLWM]3i2eLUM[3k2gLRMZ3n2hLoLY3Q3iLlLX3S3kLjLV3V3lLgLU3Y3mLdLT3\\3oL`LR3`3PM]LQ3c3QMZLP3f3RMWLo2i3SMjKX3U4kLYKe3g4]LhJR4X5k201O000000000O100000000000000O1000000dK_FY3a9fLdFV3]9hLhFT3X9kLmFQ3S9mLSGn2n8QMWGk2i8TM[Gi2e8VM`Gf2`8XMfGd2Z8[MkGa2U8^MPH^2P8aMUH[2k7cM[HX2f7gM_HU2a7jMdHR2\\7mMiHo1W7oMoHm1R7QNSIk1m6TNXIh1h6WN]Ie1c6YNbI\\1g6bN^Ik0Q7TOTI9[7FP4N`lR3"}, "label": "tall vase behind two others"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase to the right of the blue vase\"."}], "task": "refering", "answer_template": "The \"the vase to the right of the blue vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.71965625, 0.0, 0.9994375, 0.8741666666666668], "iscrowd": 0, "area": 57560.352849999996, "rle": {"size": [480, 640], "counts": "QSh65k>5K5K5K6J5K5K5K6J5K5K5K5K6J5K5K5K6J5K5K5K6J5K5K5K6J5K5K5K5K6J5K5K5QGaJS8d5fGaJW8b5eGbJX8b5bGdJZ8T6M3M3M3M3M4L3XMkH_MX7Y2mHgMV7Q2nHPNU7i1oHWNT7a1SI]NQ7Z1UIeNn6U1VIjNm6R1UImNn6o0TIPOo6m0RIQOR7k0PITOT7g0mHYOV7c0lH\\OW7a0iH_OZ7=hHB[7:gHE[78fHH]75dHJ^73cHM_70cHO`7MbH2`7L`H4c7H_H7c7F^H:d7D]H;f7A\\H>f7_O[Ha0h7[OZHd0h7ZOXHf0j7WOXHh0k7TOWHk0k7ROVHn0m7oNTHP1n7mNSHS1o7jNRHV1Q8gNoGY1S8dNnG\\1U8`NlG`1V8]NkGc1W8[NiGe1Z8WNgGi1[8TNfGl1g:000000000000000000000000000000000000000000001O001O001O001O0010O01O1O001N101O0O2O001N101O1N101O0O2O001N101O001N2O000O01O10O01O1O010O1O001O10O01O1O010O1O1O1O010O1O1000000000000000000000000O"}, "label": "the vase to the right of the blue vase"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large colorful vase on the right\"."}], "task": "refering", "answer_template": "The \"a large colorful vase on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.71965625, 0.0, 0.9994375, 0.8741666666666668], "iscrowd": 0, "area": 57560.352849999996, "rle": {"size": [480, 640], "counts": "QSh65k>5K5K5K6J5K5K5K6J5K5K5K5K6J5K5K5K6J5K5K5K6J5K5K5K6J5K5K5K5K6J5K5K5QGaJS8d5fGaJW8b5eGbJX8b5bGdJZ8T6M3M3M3M3M4L3XMkH_MX7Y2mHgMV7Q2nHPNU7i1oHWNT7a1SI]NQ7Z1UIeNn6U1VIjNm6R1UImNn6o0TIPOo6m0RIQOR7k0PITOT7g0mHYOV7c0lH\\OW7a0iH_OZ7=hHB[7:gHE[78fHH]75dHJ^73cHM_70cHO`7MbH2`7L`H4c7H_H7c7F^H:d7D]H;f7A\\H>f7_O[Ha0h7[OZHd0h7ZOXHf0j7WOXHh0k7TOWHk0k7ROVHn0m7oNTHP1n7mNSHS1o7jNRHV1Q8gNoGY1S8dNnG\\1U8`NlG`1V8]NkGc1W8[NiGe1Z8WNgGi1[8TNfGl1g:000000000000000000000000000000000000000000001O001O001O001O0010O01O1O001N101O0O2O001N101O1N101O0O2O001N101O001N2O000O01O10O01O1O010O1O001O10O01O1O010O1O1O1O010O1O1000000000000000000000000O"}, "label": "a large colorful vase on the right"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small blue vase\"."}], "task": "refering", "answer_template": "The \"a small blue vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 358, 359, 360, 361, 362], "bbox": [0.5246562499999999, 0.39462499999999995, 0.7987656249999999, 0.9282499999999999], "iscrowd": 0, "area": 29235.018200000002, "rle": {"size": [480, 640], "counts": "Zkm4e0Q><N2M3M3M2N3M3N2M3M2N3M3M2O1N2N2O1N2N2aKgMULZ2j3jMQLX2n3mMlKV2Q4oMkKR2T4SNfKo1Y4UNbKm1]4UNaKl1^4VN`Kk1_4VN_Kl1`4VN^Kk1a4VN^Kk1PMXMT7o0jKj1kL]M[7j0iKV2U4mMiKS2W4oMgKQ2Y4PNfKQ2Y4QNeKo1[4SNcKm1]4TNbKm1]4UNaKk1_4WN]Kk1c4VNZKl1e4WNXKk1g4WNUKk1k4XNPKj1P5ZNkJg1U5]NbJh1^5\\NWJk1i5YNlIn1T6VN]IU2c6U3O0O10000000000000000O10000000000000001O0O1000000000000000000O100001O000001O0000000000000000000000000000000000010O0000000000000000000000000000000000[JeJe1[5WNkJh1T5UNRKh1n4TNXKi1i4SN^Ki1d4RNbKl1^4PNfKo1[4mMiKR2X4jMmKT2T4iMoKW2Q4hMPLW2Q4hMQLV2P4hMRLX2n3gMTLW2m3hMTLW2m3hMTLW2m3hMULW2k3gMWLX2j3gMWLX2j3gMXLX2h3gMYLX2h3gMZLn1P4QNQLn1P4PNRLo1o3PNSLn1n3QNSLn1n3QNSLn1o3PNSLn1n3PNTLo1m3PNULn1l3QNULn1l3QNULn1m3PNULn1l3QNULn1l3PNWLn1k3PNVLo1k3PNVLo1k3PNWLn1^8O1O3M2N3I6K5J7J5J6K6I6K5JoSl1"}, "label": "a small blue vase"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue vase to the right of the green vase\"."}], "task": "refering", "answer_template": "The \"the blue vase to the right of the green vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 358, 359, 360, 361, 362], "bbox": [0.5246562499999999, 0.39462499999999995, 0.7987656249999999, 0.9282499999999999], "iscrowd": 0, "area": 29235.018200000002, "rle": {"size": [480, 640], "counts": "Zkm4e0Q><N2M3M3M2N3M3N2M3M2N3M3M2O1N2N2O1N2N2aKgMULZ2j3jMQLX2n3mMlKV2Q4oMkKR2T4SNfKo1Y4UNbKm1]4UNaKl1^4VN`Kk1_4VN_Kl1`4VN^Kk1a4VN^Kk1PMXMT7o0jKj1kL]M[7j0iKV2U4mMiKS2W4oMgKQ2Y4PNfKQ2Y4QNeKo1[4SNcKm1]4TNbKm1]4UNaKk1_4WN]Kk1c4VNZKl1e4WNXKk1g4WNUKk1k4XNPKj1P5ZNkJg1U5]NbJh1^5\\NWJk1i5YNlIn1T6VN]IU2c6U3O0O10000000000000000O10000000000000001O0O1000000000000000000O100001O000001O0000000000000000000000000000000000010O0000000000000000000000000000000000[JeJe1[5WNkJh1T5UNRKh1n4TNXKi1i4SN^Ki1d4RNbKl1^4PNfKo1[4mMiKR2X4jMmKT2T4iMoKW2Q4hMPLW2Q4hMQLV2P4hMRLX2n3gMTLW2m3hMTLW2m3hMTLW2m3hMULW2k3gMWLX2j3gMWLX2j3gMXLX2h3gMYLX2h3gMZLn1P4QNQLn1P4PNRLo1o3PNSLn1n3QNSLn1n3QNSLn1o3PNSLn1n3PNTLo1m3PNULn1l3QNULn1l3QNULn1m3PNULn1l3QNULn1l3PNWLn1k3PNVLo1k3PNVLo1k3PNWLn1^8O1O3M2N3I6K5J7J5J6K6I6K5JoSl1"}, "label": "the blue vase to the right of the green vase"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green ceramic vase\"."}], "task": "refering", "answer_template": "The \"a green ceramic vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355], "bbox": [0.17359375, 0.4217916666666667, 0.5460625, 0.9273958333333333], "iscrowd": 0, "area": 43528.46975, "rle": {"size": [480, 640], "counts": "jYd1a2^<3N1N2N2O1N2N3N1N2N2O1N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2O1N2N2O2M2N2O1N2N2O2PF^K]9d4aF]K^9e4`F]K_9d4`F]K^9e4`F\\K`9f4]F\\Ka9f4]F\\Ka9f4^FZKb9S5N2kNbJeH_5[7bJdH_5Z7bJeH`5Y7bJeH_5Z7cJcH`5[7bJcH`5\\7aJbHa5\\7aJaHa5^7aJ`Ha5^7aJ`Ha5^7aJ`H`5_7bJ^Ha5a7`J]Hb5b7_J\\Hb5d7_JYHd5e7]JZHe5e7\\JYHe5g7k0O1O0000000O10000000001O00000O1000000000000000001N10000000000000000000000O2O0000000000000000000000001O00000000000000000000000O100001N1000000000001O00000O1000001O000000000O101O00000000001O000O100000001O000000000O101O00000aI^H`5b7\\JfH`5Z7\\JnH`5R7]JUI^5m6]JYIa5g6\\J]Ic5c6ZJ`Ie5n70001O0O10001O0O101O0O10001O0K5C=N3L3N2N3L3N2N2N3L3N2N2N3L3N2N3L3N2N2N3L3N2N2M4M2N3M5J6K6J5K5K5J6K6J5K5K5K5K6I6K5K5KTSX4"}, "label": "a green ceramic vase"}]}
{"id": 256930, "image": "COCO_train2014_000000256930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green ceramic vase on a white table\"."}], "task": "refering", "answer_template": "The \"a green ceramic vase on a white table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355], "bbox": [0.17359375, 0.4217916666666667, 0.5460625, 0.9273958333333333], "iscrowd": 0, "area": 43528.46975, "rle": {"size": [480, 640], "counts": "jYd1a2^<3N1N2N2O1N2N3N1N2N2O1N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2O1N2N2O2M2N2O1N2N2O2PF^K]9d4aF]K^9e4`F]K_9d4`F]K^9e4`F\\K`9f4]F\\Ka9f4]F\\Ka9f4^FZKb9S5N2kNbJeH_5[7bJdH_5Z7bJeH`5Y7bJeH_5Z7cJcH`5[7bJcH`5\\7aJbHa5\\7aJaHa5^7aJ`Ha5^7aJ`Ha5^7aJ`H`5_7bJ^Ha5a7`J]Hb5b7_J\\Hb5d7_JYHd5e7]JZHe5e7\\JYHe5g7k0O1O0000000O10000000001O00000O1000000000000000001N10000000000000000000000O2O0000000000000000000000001O00000000000000000000000O100001N1000000000001O00000O1000001O000000000O101O00000000001O000O100000001O000000000O101O00000aI^H`5b7\\JfH`5Z7\\JnH`5R7]JUI^5m6]JYIa5g6\\J]Ic5c6ZJ`Ie5n70001O0O10001O0O101O0O10001O0K5C=N3L3N2N3L3N2N2N3L3N2N2N3L3N2N3L3N2N2N3L3N2N2M4M2N3M5J6K6J5K5K5J6K6J5K5K5K5K6I6K5K5KTSX4"}, "label": "a green ceramic vase on a white table"}]}
{"id": 19374, "image": "COCO_train2014_000000019374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table in front of kids\"."}], "task": "refering", "answer_template": "The \"a table in front of kids\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 213, 214, 220, 230, 231, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 253, 254, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.003, 0.5483177570093458, 0.997859375, 0.9865186915887851], "iscrowd": 0, "area": 66682.64770000003, "rle": {"size": [428, 640], "counts": "fT1V3V:0O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100Q1oN6J2N3N2M2N3M3M2N3M3M3M2N3M3M1O000000001O0000000000000000001O000000000000000000001O0000000000000000001O0000000000000000001O0000O1L4L4M3L4M3L4L4M3L4L4000000O100000000O10000000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O1000000O1O1O1N2O1O1N2O1O1UMTMTKn2l4TMQKm2P5UMlJl2T5WMgJk2Y5WMdJj2]5XM_Ji2a5ZMnIT3R6nLYIe3g6^LbHV4_7k000001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001nKRHd2n7cLlH\\3b801O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001OO10000000000000000O100000000000000O100000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O1000000O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O10000001O0000001O000000001O00000010O000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O000000V2jM>"}, "label": "a table in front of kids"}]}
{"id": 19374, "image": "COCO_train2014_000000019374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown table top\"."}], "task": "refering", "answer_template": "The \"a brown table top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 213, 214, 220, 230, 231, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 253, 254, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.003, 0.5483177570093458, 0.997859375, 0.9865186915887851], "iscrowd": 0, "area": 66682.64770000003, "rle": {"size": [428, 640], "counts": "fT1V3V:0O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100Q1oN6J2N3N2M2N3M3M2N3M3M3M2N3M3M1O000000001O0000000000000000001O000000000000000000001O0000000000000000001O0000000000000000001O0000O1L4L4M3L4M3L4L4M3L4L4000000O100000000O10000000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O1000000O1O1O1N2O1O1N2O1O1UMTMTKn2l4TMQKm2P5UMlJl2T5WMgJk2Y5WMdJj2]5XM_Ji2a5ZMnIT3R6nLYIe3g6^LbHV4_7k000001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001nKRHd2n7cLlH\\3b801O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001OO10000000000000000O100000000000000O100000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O1000000O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O10000001O0000001O000000001O00000010O000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O000000V2jM>"}, "label": "a brown table top"}]}
{"id": 330671, "image": "COCO_train2014_000000330671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle parked in front of a blue truck\"."}], "task": "refering", "answer_template": "The \"a motorcycle parked in front of a blue truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 44, 45, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 181, 182, 183], "bbox": [0.773640625, 0.04269320843091335, 1.0, 0.5348243559718969], "iscrowd": 0, "area": 14941.319249999997, "rle": {"size": [427, 640], "counts": "Qb^62V=5J5L5J6K5J5L5J6K5K5K4L5K4L3O2O001O001O00001O001O001O001O001N101O00001O001O001O001O001O001O00001O0O2O001O001O0VOUMTGl2j8VMUGj2k8XMSGi2k8YMTGg2l8ZMSGg2k8[MTGd2l8^MSGa2n8`MQG_2o8cMPG]2o8eMPGZ2Q9gMnFX2R9kMkFU2V9lMgFV2X9lMeFU2[9mMbFT2_9k0010O001O1O010O001O010O[G[Le7e3\\H\\Lc7c3]H_La7b3_H_L`7`3aHaL^7^3cHcL[7]3fHdLY7\\3fHfLX7Z3iHgLV7X3kHiLS7W3nHjLQ7V3oHkLo6U3QIlLo6S3RIlLn6T3SIkLn6U3RIkLn6T3RIlLm6U3TIjLl6W3TIhLk6Z3TIfLl6\\3SIdLk6^3TIbLk6a3SI_Lm6b3RI^Lm6e3QI[Ln6g3RIXLn6g0XHk1i0_Mm6b0_HQ2b0]Mn6=gHV2:]Mo68mH\\23[Mo6:oH[21[Mo6;QI[2NZMQ7<QIZ2M[MP7<TIY2K[MP7=VIY2IYMQ7>WIY2GYMQ7?YIY2DXMR7`0[IX2BYMR7?]IY2_OXMT7?^IY2^OWMU7?^IZ2\\OWMV7?_I[2YOVMY7>_I\\2WOVM[7=_I^2TOVM]7;`I_2SOUM^7;`Ia2POTM`7;`Ib2oNSMb7:`Ic2mNSMd79`Ie2kNRMe78aIf2iNRMg77aIh2fNQMi77bIR3_6mLbIR3_6mLbIR3_6mLbIS3^6lLcIT3\\6lLeIT3[6kLfIU3Z6jLgIV3Y6iLhIW3W6hLkIX3U6bLQJ^3n70_M"}, "label": "a motorcycle parked in front of a blue truck"}]}
{"id": 330671, "image": "COCO_train2014_000000330671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle that is parked in front of the blue truck\"."}], "task": "refering", "answer_template": "The \"the motorcycle that is parked in front of the blue truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 44, 45, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 181, 182, 183], "bbox": [0.773640625, 0.04269320843091335, 1.0, 0.5348243559718969], "iscrowd": 0, "area": 14941.319249999997, "rle": {"size": [427, 640], "counts": "Qb^62V=5J5L5J6K5J5L5J6K5K5K4L5K4L3O2O001O001O00001O001O001O001O001N101O00001O001O001O001O001O001O00001O0O2O001O001O0VOUMTGl2j8VMUGj2k8XMSGi2k8YMTGg2l8ZMSGg2k8[MTGd2l8^MSGa2n8`MQG_2o8cMPG]2o8eMPGZ2Q9gMnFX2R9kMkFU2V9lMgFV2X9lMeFU2[9mMbFT2_9k0010O001O1O010O001O010O[G[Le7e3\\H\\Lc7c3]H_La7b3_H_L`7`3aHaL^7^3cHcL[7]3fHdLY7\\3fHfLX7Z3iHgLV7X3kHiLS7W3nHjLQ7V3oHkLo6U3QIlLo6S3RIlLn6T3SIkLn6U3RIkLn6T3RIlLm6U3TIjLl6W3TIhLk6Z3TIfLl6\\3SIdLk6^3TIbLk6a3SI_Lm6b3RI^Lm6e3QI[Ln6g3RIXLn6g0XHk1i0_Mm6b0_HQ2b0]Mn6=gHV2:]Mo68mH\\23[Mo6:oH[21[Mo6;QI[2NZMQ7<QIZ2M[MP7<TIY2K[MP7=VIY2IYMQ7>WIY2GYMQ7?YIY2DXMR7`0[IX2BYMR7?]IY2_OXMT7?^IY2^OWMU7?^IZ2\\OWMV7?_I[2YOVMY7>_I\\2WOVM[7=_I^2TOVM]7;`I_2SOUM^7;`Ia2POTM`7;`Ib2oNSMb7:`Ic2mNSMd79`Ie2kNRMe78aIf2iNRMg77aIh2fNQMi77bIR3_6mLbIR3_6mLbIR3_6mLbIS3^6lLcIT3\\6lLeIT3[6kLfIU3Z6jLgIV3Y6iLhIW3W6hLkIX3U6bLQJ^3n70_M"}, "label": "the motorcycle that is parked in front of the blue truck"}]}
{"id": 330671, "image": "COCO_train2014_000000330671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red motorbike on a jack with tools on the ground\"."}], "task": "refering", "answer_template": "The \"red motorbike on a jack with tools on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 74, 75, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 337, 338], "bbox": [0.20090625, 0.17079625292740047, 0.815625, 0.9415690866510539], "iscrowd": 0, "area": 71301.7643, "rle": {"size": [427, 640], "counts": "dPf1:P=3M2N2M4M2C]OoCe0\\;P1N1O2N1O1O1O1N3N1O1O2N3^NPNdGS2l7]NQHf1_7iN^H[1^7hNaHZ1]7gNaH\\1]7eNaH^1\\7dNbH_1\\7bNbHa1\\7`NbHc1[7_NdHc1Z7_NcHd1[7]NcHf1[7ZNdHh1Z7ZNdHh1[7YNcHh1]7YNbHh1\\7ZNbHh1]7YNaHh1_7YN_Hi1_7XN`Hi1`7R2O1N101O010001N2O0O10O010O10O010O01O010O01O001O001O00010O001O001O001O00010O001O001O002N2N1N3M3N1N3N2M2N3N1N3N2M2N101N101O1O00100O001O10O01O1O01N1M3N2M3N3L3N2M3N3L3O1O1O2N1O1O1O1O2N1O1N201O000000001O0000001O0000001O000000001O00000^OZJ_Ig5]6`J_I`5^6gJ^IY5^6oJ^IQ5_6o0O10O1000000O100000000O1000000O100000000O1000O10O100000000O100O1O1O00100O1O1O1O100O1O1O1O010O1O1O1O1O1N2N2N2NJ61O2N2N2N21O010O001O1O001O1O001O010O1ZI_J`5a5RJTKg5^6K6J6I2O00000O2O000O10001O0O10001N1000000O2O00000O10000O1000000O10000O1000000O1000000O10000O10000000000000O100000001O0O100O10000O100]JTKe2m4XMWKf2j4WMYKi2g4TM]Kj2d4SM_Kl2b4VLPKXO`0b4a4TLQKWOa0d4_4SLRKVOb0f4]4SLQKUOe0g4[4RLRKTOe0k4Y4PLSKROg0m4W4oKTKQOh0o4U4nKUKPOh0R5T4mKUKnNj0T5S4kK^LU4c3jK^LU4c3iK^LW4c3fK`LZ4`3bKcL^4^3]KgLb4^3UKfLk4n5O001N101O001O0O2O1O001N101O001O0O2O001O0O2O1O001O1N101O1O0O2O1N101N2N1O2O1N1O2O1N1O2O1N1O2O2M3M2O2M3M3N1N3M3N2M2N3N2M3M3N3L4L4M2M4L4G9G9G8H9H8G9Gbl`1"}, "label": "red motorbike on a jack with tools on the ground"}]}
{"id": 330671, "image": "COCO_train2014_000000330671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red motorcycle with the number 67 on it\"."}], "task": "refering", "answer_template": "The \"the red motorcycle with the number 67 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 74, 75, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 337, 338], "bbox": [0.20090625, 0.17079625292740047, 0.815625, 0.9415690866510539], "iscrowd": 0, "area": 71301.7643, "rle": {"size": [427, 640], "counts": "dPf1:P=3M2N2M4M2C]OoCe0\\;P1N1O2N1O1O1O1N3N1O1O2N3^NPNdGS2l7]NQHf1_7iN^H[1^7hNaHZ1]7gNaH\\1]7eNaH^1\\7dNbH_1\\7bNbHa1\\7`NbHc1[7_NdHc1Z7_NcHd1[7]NcHf1[7ZNdHh1Z7ZNdHh1[7YNcHh1]7YNbHh1\\7ZNbHh1]7YNaHh1_7YN_Hi1_7XN`Hi1`7R2O1N101O010001N2O0O10O010O10O010O01O010O01O001O001O00010O001O001O001O00010O001O001O002N2N1N3M3N1N3N2M2N3N1N3N2M2N101N101O1O00100O001O10O01O1O01N1M3N2M3N3L3N2M3N3L3O1O1O2N1O1O1O1O2N1O1N201O000000001O0000001O0000001O000000001O00000^OZJ_Ig5]6`J_I`5^6gJ^IY5^6oJ^IQ5_6o0O10O1000000O100000000O1000000O100000000O1000O10O100000000O100O1O1O00100O1O1O1O100O1O1O1O010O1O1O1O1O1N2N2N2NJ61O2N2N2N21O010O001O1O001O1O001O010O1ZI_J`5a5RJTKg5^6K6J6I2O00000O2O000O10001O0O10001N1000000O2O00000O10000O1000000O10000O1000000O1000000O10000O10000000000000O100000001O0O100O10000O100]JTKe2m4XMWKf2j4WMYKi2g4TM]Kj2d4SM_Kl2b4VLPKXO`0b4a4TLQKWOa0d4_4SLRKVOb0f4]4SLQKUOe0g4[4RLRKTOe0k4Y4PLSKROg0m4W4oKTKQOh0o4U4nKUKPOh0R5T4mKUKnNj0T5S4kK^LU4c3jK^LU4c3iK^LW4c3fK`LZ4`3bKcL^4^3]KgLb4^3UKfLk4n5O001N101O001O0O2O1O001N101O001O0O2O001O0O2O1O001O1N101O1O0O2O1N101N2N1O2O1N1O2O1N1O2O1N1O2O2M3M2O2M3M3N1N3M3N2M2N3N2M3M3N3L4L4M2M4L4G9G9G8H9H8G9Gbl`1"}, "label": "the red motorcycle with the number 67 on it"}]}
{"id": 330671, "image": "COCO_train2014_000000330671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leg with a teal shoe on it ' s foot\"."}], "task": "refering", "answer_template": "The \"a leg with a teal shoe on it ' s foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 162, 184, 185, 186, 208, 209, 231, 232, 233], "bbox": [0.0015, 0.30402810304449646, 0.14360937499999998, 0.7367681498829041], "iscrowd": 0, "area": 4160.027550000002, "rle": {"size": [427, 640], "counts": "gb0Z1i:Z12M3M3M3M3M3M3M3M3M3M3M3M3M2O1N2N2N2N2N1O00000000000001O01O01O00001O00001O000[EmNQ9T1jFoNW9Q1fFROZ9n0bFUO_9l0]FVOd9j0ZFXOf9h0XFYOi9h0SFZOo9e0oE\\OR:d0kE_OU:b0hE_OY:a0dEA]:?aEB`:?\\ECe:=YEDh:<VEDl:=PEEQ;Z11O1N2O1O1N2O1O2N1N2O1N2N2O2M2N3M3M2N3M3M2O2M2N3M3M2N3M2N3N2M2N3M2N2N]WT7"}, "label": "a leg with a teal shoe on it ' s foot"}]}
{"id": 330671, "image": "COCO_train2014_000000330671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leg of a parson with blue boot\"."}], "task": "refering", "answer_template": "The \"the leg of a parson with blue boot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 162, 184, 185, 186, 208, 209, 231, 232, 233], "bbox": [0.0015, 0.30402810304449646, 0.14360937499999998, 0.7367681498829041], "iscrowd": 0, "area": 4160.027550000002, "rle": {"size": [427, 640], "counts": "gb0Z1i:Z12M3M3M3M3M3M3M3M3M3M3M3M3M2O1N2N2N2N2N1O00000000000001O01O01O00001O00001O000[EmNQ9T1jFoNW9Q1fFROZ9n0bFUO_9l0]FVOd9j0ZFXOf9h0XFYOi9h0SFZOo9e0oE\\OR:d0kE_OU:b0hE_OY:a0dEA]:?aEB`:?\\ECe:=YEDh:<VEDl:=PEEQ;Z11O1N2O1O1N2O1O2N1N2O1N2N2O2M2N3M3M2N3M3M2O2M2N3M3M2N3M2N3N2M2N3M2N2N]WT7"}, "label": "the leg of a parson with blue boot"}]}
{"id": 453553, "image": "COCO_train2014_000000453553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a purple shirt placing a piece of chocolate cake on a plate\"."}], "task": "refering", "answer_template": "The \"a woman in a purple shirt placing a piece of chocolate cake on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 39, 40, 41, 42, 56, 57, 58, 59, 60, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 206, 207, 208, 209, 210, 211, 212, 213, 214, 223, 224, 225, 226, 227, 228, 229, 230, 231, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 263], "bbox": [0.10164583333333334, 0.06715625, 0.6248125000000001, 0.677015625], "iscrowd": 0, "area": 57907.491299999994, "rle": {"size": [640, 480], "counts": "omn02mc02N2N2N2N2N3M2N2M3N2N2N3M2N2N2N2N2Nd1\\N5F:iMhLUCe3a<nLjB_3k<SMaBY3V=o1F;E:F:F:G5J6J5K6N2000O0100O1000O010000O010O1000O010000O010O1000O010000O01000O100O01000O10O10O100O01000O10O10O100O01000O10O10O100O10O10O1000OWOj0H8H8H7O2O100O00100O10O0100O10O0100O100O010O100O010O100O010O1O10O01O1M3M2N3M3M3M2O200O1000O1000000O0100000O1000O010000O10000O01000O10000O10O3NUNVGdHf8[7\\GgHa8X7aGkH[8U7fGmHW8R7kGQIP8o6RHPIn7P7SHoHm7R7SHnHl7T7SHkHn7U7RHjHo7W7PHiHP8X7oGgHR8Y7nGfHS8[7lGhGJF[8b8kGeG1ET8g8jGbG7CP8k8iG`G<Al7P9iG[G`0Ah7T9WHjF6Oc7X9^IdFc6\\9Q21OdIbEU3^:gLgEY3X:dLlE[3T:bLPF^3o9_LUF`3k9]LYFc3f9ZL^Fe3b9XLbF\\O2^1]9SOdIh0]6TOkIh0U6UORJf0o5WOXJe0h5XO^Jd0c5YOdJc0]5YOjJb0W5[OPKa0P5\\OWK?n4YOYKc0P5ROWKj0P5lNWKo0R5fNRKY1k;O100O00100O100O10O0100O1O100O010O100O100O0O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2N2M2O2N2NhT`3"}, "label": "a woman in a purple shirt placing a piece of chocolate cake on a plate"}]}
{"id": 453553, "image": "COCO_train2014_000000453553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a purple coat adjusts a cake\"."}], "task": "refering", "answer_template": "The \"a woman in a purple coat adjusts a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 39, 40, 41, 42, 56, 57, 58, 59, 60, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 206, 207, 208, 209, 210, 211, 212, 213, 214, 223, 224, 225, 226, 227, 228, 229, 230, 231, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 263], "bbox": [0.10164583333333334, 0.06715625, 0.6248125000000001, 0.677015625], "iscrowd": 0, "area": 57907.491299999994, "rle": {"size": [640, 480], "counts": "omn02mc02N2N2N2N2N3M2N2M3N2N2N3M2N2N2N2N2Nd1\\N5F:iMhLUCe3a<nLjB_3k<SMaBY3V=o1F;E:F:F:G5J6J5K6N2000O0100O1000O010000O010O1000O010000O010O1000O010000O01000O100O01000O10O10O100O01000O10O10O100O01000O10O10O100O10O10O1000OWOj0H8H8H7O2O100O00100O10O0100O10O0100O100O010O100O010O100O010O1O10O01O1M3M2N3M3M3M2O200O1000O1000000O0100000O1000O010000O10000O01000O10000O10O3NUNVGdHf8[7\\GgHa8X7aGkH[8U7fGmHW8R7kGQIP8o6RHPIn7P7SHoHm7R7SHnHl7T7SHkHn7U7RHjHo7W7PHiHP8X7oGgHR8Y7nGfHS8[7lGhGJF[8b8kGeG1ET8g8jGbG7CP8k8iG`G<Al7P9iG[G`0Ah7T9WHjF6Oc7X9^IdFc6\\9Q21OdIbEU3^:gLgEY3X:dLlE[3T:bLPF^3o9_LUF`3k9]LYFc3f9ZL^Fe3b9XLbF\\O2^1]9SOdIh0]6TOkIh0U6UORJf0o5WOXJe0h5XO^Jd0c5YOdJc0]5YOjJb0W5[OPKa0P5\\OWK?n4YOYKc0P5ROWKj0P5lNWKo0R5fNRKY1k;O100O00100O100O10O0100O1O100O010O100O100O0O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2N2M2O2N2NhT`3"}, "label": "a woman in a purple coat adjusts a cake"}]}
{"id": 453553, "image": "COCO_train2014_000000453553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a multi colored scarf watches another woman\"."}], "task": "refering", "answer_template": "The \"a woman with a multi colored scarf watches another woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 80, 81, 82, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 182, 183, 184, 185, 200, 201, 202, 218], "bbox": [0.6328541666666666, 0.137125, 0.9535833333333332, 0.5328125], "iscrowd": 0, "area": 25686.78595000001, "rle": {"size": [640, 480], "counts": "cVn55hc06I7J6J6I7J7H7J6J5J5L4K5M3M4M2N2M3N3M2M4M2N3L3N3M2M4M2N3L3N3L3M4K4M2M3N2M3N2M3N2M3N2M3N2O001O1O1O1O2N2N2N2N2N3M2N2N2N2N2N2N3jBSJT<n6N1O2O001O001O0O2O001O1O001O001O001N101O001O001O1O000100O1O1O100O1O1O1O100O1O1O3M2O2M2N3M2dITC`5n<TJ]Cl5[=000000001O01O0O1O100O1O101N100O1O100O100O2N103L5L3L5K5L3L5K5L3K6C=B=D=B>F:H7J7H8I6I8I7QOla="}, "label": "a woman with a multi colored scarf watches another woman"}]}
{"id": 453553, "image": "COCO_train2014_000000453553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a scarf near a table\"."}], "task": "refering", "answer_template": "The \"a woman with a scarf near a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 80, 81, 82, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 182, 183, 184, 185, 200, 201, 202, 218], "bbox": [0.6328541666666666, 0.137125, 0.9535833333333332, 0.5328125], "iscrowd": 0, "area": 25686.78595000001, "rle": {"size": [640, 480], "counts": "cVn55hc06I7J6J6I7J7H7J6J5J5L4K5M3M4M2N2M3N3M2M4M2N3L3N3M2M4M2N3L3N3L3M4K4M2M3N2M3N2M3N2M3N2M3N2O001O1O1O1O2N2N2N2N2N3M2N2N2N2N2N2N3jBSJT<n6N1O2O001O001O0O2O001O1O001O001O001N101O001O001O1O000100O1O1O100O1O1O1O100O1O1O3M2O2M2N3M2dITC`5n<TJ]Cl5[=000000001O01O0O1O100O1O101N100O1O100O100O2N103L5L3L5K5L3L5K5L3K6C=B=D=B>F:H7J7H8I6I8I7QOla="}, "label": "a woman with a scarf near a table"}]}
{"id": 453553, "image": "COCO_train2014_000000453553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large chocolate cake\"."}], "task": "refering", "answer_template": "The \"large chocolate cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 302, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353], "bbox": [0.4876666666666667, 0.679609375, 0.7988125, 0.891015625], "iscrowd": 0, "area": 16005.095899999995, "rle": {"size": [640, 480], "counts": "agb4;\\c0c0^Ob0E3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2O1N2N2N2N101N2O1N101N2N2O0O2O1N2N2O0O2O1N2O0O2O0O10000O2O000O100O10000O10000O100O101O00000O10000000000O10000000001N1000000O100O100O100O100O100001O001O0O101O001O00001O001O00001O0O2O1O1O1O1O001O1O1O1O1O1O1O1N2O1O1O1O1O1O001O1O1O1N2N2N2N3M2N2N2M3N2N2N2N2N2N3L3SNk^Ol0Wa0POkVl1"}, "label": "large chocolate cake"}]}
{"id": 453553, "image": "COCO_train2014_000000453553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate cake\"."}], "task": "refering", "answer_template": "The \"a chocolate cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 302, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353], "bbox": [0.4876666666666667, 0.679609375, 0.7988125, 0.891015625], "iscrowd": 0, "area": 16005.095899999995, "rle": {"size": [640, 480], "counts": "agb4;\\c0c0^Ob0E3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2O1N2N2N2N101N2O1N101N2N2O0O2O1N2N2O0O2O1N2O0O2O0O10000O2O000O100O10000O10000O100O101O00000O10000000000O10000000001N1000000O100O100O100O100O100001O001O0O101O001O00001O001O00001O0O2O1O1O1O1O001O1O1O1O1O1O1O1N2O1O1O1O1O1O001O1O1O1N2N2N2N3M2N2N2M3N2N2N2N2N2N3L3SNk^Ol0Wa0POkVl1"}, "label": "a chocolate cake"}]}
{"id": 551873, "image": "COCO_train2014_000000551873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small elephant behind a larger elephant\"."}], "task": "refering", "answer_template": "The \"a small elephant behind a larger elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 264, 281, 283, 284], "bbox": [0.174265625, 0.42079625292740047, 0.5173125000000001, 0.8149414519906324], "iscrowd": 0, "area": 21724.9345, "rle": {"size": [427, 640], "counts": "Pn^14T=6K4L5K4L5J6K4L5K4K6K5L3L5L4L4L4K5L4L5K5J8I7I3M3L4M3M3dFbLe8V4L3M1O2N1O100O100O100O100O100O100O01O010N1N3L41O2N1O1O00001O001O000000N2O1N2O100WOoGWLR8d3VHXLk7f3XHYLh7f3\\HVLe7h3P1N2N2N2N2N2O1N20000O1000000O100001O1O1O1O100O:F7J4K6bGfK`7^4kGRLU8b4001O00001O001O00001O001O00001O00001O001O000EdGgK]8X4eGfK[8Z4hGcKY8\\4;0001O00001XOUGnLk8o2k0K4L4L5L3@a0O01O01O01O01O010O01O010O100O2N010O100O01O1O1O1L3L5I7O10O10O10000000O10O10003M3M3M4K7J5K6J5K1O1O1O1N2O1O001O1O1O1O10O1O2N1O1O2NN2O1O1O2N1O1N2O1O1O1O1O1N2O2N1O1O1O`aP4"}, "label": "a small elephant behind a larger elephant"}]}
{"id": 551873, "image": "COCO_train2014_000000551873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant with some hay\"."}], "task": "refering", "answer_template": "The \"a baby elephant with some hay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 264, 281, 283, 284], "bbox": [0.174265625, 0.42079625292740047, 0.5173125000000001, 0.8149414519906324], "iscrowd": 0, "area": 21724.9345, "rle": {"size": [427, 640], "counts": "Pn^14T=6K4L5K4L5J6K4L5K4K6K5L3L5L4L4L4K5L4L5K5J8I7I3M3L4M3M3dFbLe8V4L3M1O2N1O100O100O100O100O100O100O01O010N1N3L41O2N1O1O00001O001O000000N2O1N2O100WOoGWLR8d3VHXLk7f3XHYLh7f3\\HVLe7h3P1N2N2N2N2N2O1N20000O1000000O100001O1O1O1O100O:F7J4K6bGfK`7^4kGRLU8b4001O00001O001O00001O001O00001O00001O001O000EdGgK]8X4eGfK[8Z4hGcKY8\\4;0001O00001XOUGnLk8o2k0K4L4L5L3@a0O01O01O01O01O010O01O010O100O2N010O100O01O1O1O1L3L5I7O10O10O10000000O10O10003M3M3M4K7J5K6J5K1O1O1O1N2O1O001O1O1O1O10O1O2N1O1O2NN2O1O1O2N1O1N2O1O1O1O1O1N2O2N1O1O1O`aP4"}, "label": "a baby elephant with some hay"}]}
{"id": 166849, "image": "COCO_train2014_000000166849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a close - up of a horse ' s head\"."}], "task": "refering", "answer_template": "The \"a close - up of a horse ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 33, 34, 35, 36, 55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 312, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 352, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0, 0.034416666666666665, 0.8354843750000001, 0.98925], "iscrowd": 0, "area": 141190.45925000004, "rle": {"size": [480, 640], "counts": "d7T7j5R200001O00001O000oFUJm8l5RGTJn8l5RGTJn8l5RGUJm8l5RGTJn8l5RGTJn8l5RGUJm8l5SGSJm8m5SGSJm8m5SGTJl8P601O00001O00001O00001O00001O00001O0000001O000000001O0001O01O0000001O0000001O001O00001O00001O00001O001O00001O0000O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O2N100O100O100OoMTLPIl3n6WLQIh3o6[LoHe3P7]LoHb3P7aLoH^3Q7eLmH[3R7gLmHX3R7lLlHT3S7nLlHQ3T7RMjHn2T7UMkHj2U7XMjHh2U7[MiHe2U7^M\\Ho2d7W2O0O2O001N2O001N101O0O2O1O0O2O0O2O001N2O001O0O2O001O1O001N101O001O1O001N101O001O001O1N101O001O001O0O2O001O1O001N101O001O001O1N101O001O001O0O2O001O1O001O0O2O02O0aN[IbJg6Y5bIbJ_6Z5iI`JX6]5PJ^JQ6]5XJ]Ji5b5]JYJd5g5`JSJb5l5cJoI]5R6i1000O10O010000O01000O100O01000O10O10O100O10O10O1000O10O100000O10O1000O10O100000O0100000O01000000O0100000OM4L4L4L4M2O2N2N2M2O2N2N2N101N2O1O0O2O1O1O0O2O1O1N1107I7I8G1000O01000O010O1000O0100O010O10O10O100O010O10O10O10O103M2N3L4M1OO010000O10000O1000000O10000O1000O0100002N8H5K00O100O10003L5L3L4M3M3L5L3L4M3L5L3M3L3N2M2O2M3N1O2M3N1N3N1O2M3N1N3N2M2O2N1N3N2M2O2N2M2O1N2O0O2O1O0O2O1M2M4M3L3N3L3N3M3L3O2N2M2O2M3N1N3N2N1N3N2M2O2M3N1N3M3M2O2M3M3M3M4M2M3M4L3N2M3N3L3N2M4M2M3N2N3L3N2M3N3L3N2M4M2K9H8G9G9H8Gnj`1"}, "label": "a close - up of a horse ' s head"}]}
{"id": 166849, "image": "COCO_train2014_000000166849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with a white mane\"."}], "task": "refering", "answer_template": "The \"a horse with a white mane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 33, 34, 35, 36, 55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 312, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 352, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0, 0.034416666666666665, 0.8354843750000001, 0.98925], "iscrowd": 0, "area": 141190.45925000004, "rle": {"size": [480, 640], "counts": "d7T7j5R200001O00001O000oFUJm8l5RGTJn8l5RGTJn8l5RGUJm8l5RGTJn8l5RGTJn8l5RGUJm8l5SGSJm8m5SGSJm8m5SGTJl8P601O00001O00001O00001O00001O00001O0000001O000000001O0001O01O0000001O0000001O001O00001O00001O00001O001O00001O0000O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O2N100O100O100OoMTLPIl3n6WLQIh3o6[LoHe3P7]LoHb3P7aLoH^3Q7eLmH[3R7gLmHX3R7lLlHT3S7nLlHQ3T7RMjHn2T7UMkHj2U7XMjHh2U7[MiHe2U7^M\\Ho2d7W2O0O2O001N2O001N101O0O2O1O0O2O0O2O001N2O001O0O2O001O1O001N101O001O1O001N101O001O001O1N101O001O001O0O2O001O1O001N101O001O001O1N101O001O001O0O2O001O1O001O0O2O02O0aN[IbJg6Y5bIbJ_6Z5iI`JX6]5PJ^JQ6]5XJ]Ji5b5]JYJd5g5`JSJb5l5cJoI]5R6i1000O10O010000O01000O100O01000O10O10O100O10O10O1000O10O100000O10O1000O10O100000O0100000O01000000O0100000OM4L4L4L4M2O2N2N2M2O2N2N2N101N2O1O0O2O1O1O0O2O1O1N1107I7I8G1000O01000O010O1000O0100O010O10O10O100O010O10O10O10O103M2N3L4M1OO010000O10000O1000000O10000O1000O0100002N8H5K00O100O10003L5L3L4M3M3L5L3L4M3L5L3M3L3N2M2O2M3N1O2M3N1N3N1O2M3N1N3N2M2O2N1N3N2M2O2N2M2O1N2O0O2O1O0O2O1M2M4M3L3N3L3N3M3L3O2N2M2O2M3N1N3N2N1N3N2M2O2M3N1N3M3M2O2M3M3M3M4M2M3M4L3N2M3N3L3N2M4M2M3N2N3L3N2M3N3L3N2M4M2K9H8G9G9H8Gnj`1"}, "label": "a horse with a white mane"}]}
{"id": 76740, "image": "COCO_train2014_000000076740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman playing tennis on the right side but only in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the woman playing tennis on the right side but only in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 88, 89, 111, 112, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 225, 226, 227, 248, 249, 271, 272, 294, 295, 316, 317, 318, 319, 339, 340, 341, 342], "bbox": [0.717921875, 0.17776497695852536, 0.9383593750000001, 0.9249539170506911], "iscrowd": 0, "area": 17356.855399999997, "rle": {"size": [434, 640], "counts": "_fR65]=00000000000O1O1O1K6O01O1O1O1OfH1R1OkN5T1KiN8W1HiN8W1HhN:V1GiN:W1FhN;W1FhN<W1DhN=W1CiN>V1CiN?V1AiN`0V1AiN`0W1_OiNb0V1_OiNb0W1]OiNc0W1^OhNc0X1]OgNd0X1\\OhNc0Z1]OeNc0[1^OdNb0]1]OcNc0]1^ObNb0_1^O`Nb0`1^O`Nb0`1_O_Na0a1_OXLNhNc0Q5_OoK8kN8Y5AeK>oN1^5A]Ke0eN7o5VOVKj0dN2X6VOnJn0dNO`6SOgJT1bNLi6RO_JY1aNHQ7POZJW3T5]MiJc2Y5]MdJc2_5]M^Jc2e5^MVJc2m5]MQJe2P6ZMmIi2T6VMiIm2X6SMhIl2X6TMiIj2Y6UMhIj2Z6UMfIj2[6UMeIl2\\6RMeIP3Y6PMgIR3X6lLhIW3V6hLhI]3W6eLcI`3[6lLVIY3^5[KaK_6]4aIcKb6Z4^IeKe6X4]IgKf6V4ZIjKi6R4XInKf4nNTMR5VNPLe4SOTMi4`NlK[4@TM`4dNnKV4HUMU4jNPLP40TMl3QOQLj38TMYOeNc3a0iLf3`0jMa2dNkLc3i0eM[2lNhL_3R1aMT2TOfL\\3[1\\Mn1[OdLX3b1[Mj1W4YNgKg1X4\\NfKd1Y4_NeKa1Z4bNcK_1\\4cNcK]1[4gNbKZ1]4iN`KY1^4jN_KW1`4lN]KV1a4mN\\KT1c4oNZKR1e4QOXKQ1e4SOYKm0f4UOXKl0g4VOWKl0g4VOWKk0h4WOVKj0i4WOVKk0g4XOWKi0h4YOVKh0i4YOWKh0d4]OYKe0c4@VKg0e4^OTKh0h4S4L5L3O1N3M2]LoFd2R9VMVGf2m8TMZGi2i9M4K5I7I7I6L4O0O2N2O0O2N2N101N2N3N2M3M3N2G9F:G8G:F\\\\`0"}, "label": "the woman playing tennis on the right side but only in the right hand picture"}]}
{"id": 76740, "image": "COCO_train2014_000000076740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle behind the woman on the right\"."}], "task": "refering", "answer_template": "The \"a bicycle behind the woman on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 245, 246, 247, 249, 250, 251, 268, 269, 270, 272, 273, 274], "bbox": [0.665125, 0.55536866359447, 0.9354687500000001, 0.7432718894009216], "iscrowd": 0, "area": 5035.584099999999, "rle": {"size": [434, 640], "counts": "ijd55Y=5K5L4L4L4L4L4M3N2N2M3N2N1N3N2O1O1O1O001O1O1O1O1O001O1O1OO10000000000000000000000O1000001O0O100000O010M3L3M4M3O1O001O10O3M2N2O1N2N2N2N3L5J5K5L5G9F:G8Heeb0KhZ]O0O1J6L4M2O2M2N2N2N2O2M2N2N2N2N3NO0O2SOgC`0k<CYb2IV^M1N2O0000000aNN^E3a:0\\E0c:3ZENf:3WEOj:1SE1n:OoD2R;0jD2V;0gD1[;0`D1c;0TD6n;e04M5J4M3L4M4K7J9F\\Ra0"}, "label": "a bicycle behind the woman on the right"}]}
{"id": 76740, "image": "COCO_train2014_000000076740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bicycle in the right image\"."}], "task": "refering", "answer_template": "The \"bicycle in the right image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 245, 246, 247, 249, 250, 251, 268, 269, 270, 272, 273, 274], "bbox": [0.665125, 0.55536866359447, 0.9354687500000001, 0.7432718894009216], "iscrowd": 0, "area": 5035.584099999999, "rle": {"size": [434, 640], "counts": "ijd55Y=5K5L4L4L4L4L4M3N2N2M3N2N1N3N2O1O1O1O001O1O1O1O1O001O1O1OO10000000000000000000000O1000001O0O100000O010M3L3M4M3O1O001O10O3M2N2O1N2N2N2N3L5J5K5L5G9F:G8Heeb0KhZ]O0O1J6L4M2O2M2N2N2N2O2M2N2N2N2N3NO0O2SOgC`0k<CYb2IV^M1N2O0000000aNN^E3a:0\\E0c:3ZENf:3WEOj:1SE1n:OoD2R;0jD2V;0gD1[;0`D1c;0TD6n;e04M5J4M3L4M4K7J9F\\Ra0"}, "label": "bicycle in the right image"}]}
{"id": 76740, "image": "COCO_train2014_000000076740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the image of the woman playing tennis on the left\"."}], "task": "refering", "answer_template": "The \"the image of the woman playing tennis on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 76, 77, 78, 99, 100, 101, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 167, 168, 169, 170, 171, 191, 192, 193, 213, 214, 215, 236, 237, 238, 260, 282, 283, 284, 304, 305, 306, 307, 328, 330], "bbox": [0.2185, 0.18223502304147465, 0.447734375, 0.9309447004608294], "iscrowd": 0, "area": 17668.953350000014, "rle": {"size": [434, 640], "counts": "W`k11a=0O101N10000O2O0JKoB5R=MkB3S=8N10000O2O000000000dHK[15eNL[13eNOZ10eN2Z1NfN4Y1KgN6X1KgN7X1HhN9X1GgN9Y1HfN9Z1GeN:Z1FgN9Z1GeN:Z1GfN9Z1FfN:Z1GfN9Z1GeN:Z1FgN:Y1FfN:Z1GdN;\\1EcN<\\1DcN=^1CaN>^1C`N?`1A^N`0c1_O[Nd0d1]OYNf0f1[OVNh0i1YOUNj0h1YOkK0UOj0n4XOdK8YOc0Q5XO]K?^O;d4[NaJl0Y1d0[O:i4YN`Jj0X1j0XO7P5WN^Jh0W1P1TO6V5UN[Jf0W1U1SO\\OQ6lNbId0W1Z1oNZOX6KeJR1lNVO_6IbJm2^5SM`Jn2`5SM_Jn2`5SM^Jn2b5SM\\Jo2c5RM[Jo2g5nLZJS3g5jLYJW3i5fLWJ\\3i5cLWJ]3k5`LTJc3f4ZKcKP1Fh3h4YKbKl0En3i4WKaKn0Al3P5VK_K[1POc3`5SK`Kb6]4_IcKQ5POWL[5hNeKo4SOZLU5hNgKn4WOZLo4iNjKl4YO[Ll4iNkKk4\\O\\Lg4iNmKi4_O_Lb4hNnKi4C_L\\4\\O]KU49`LU4^OaKR4=bLj3AhKl3`0fLM^Ng2T1jLg3f0iL]OdNS3m0hLf3j0bM]2iNgLe3o0`MZ2kNfLe3R1_MX2mNdLc3X1^MT2o3QNmKo1R4WNiKi1V4]NeKc1Y4dNbK]1\\4hN`KX1_4nN\\KR1c4TOXKl0f4ZOTKh0k4ZORKh0m4YOQKi0m4YOQKi0n4XOPKi0P5YOmJi0Q5YOmJi0R5XOlJj0S5WOkJk0S5XOjJj0T5XOiJk0U5WOiJl0T5VOjJm0S5VOiJn0g0jKW3Z3PLP1d0kKX3W3QLR1d0iKY3P8eLRHX3P8gLQHW3R8gLPHV3R8iLPHT3S8kLmGS3U8lLmGQ3V8mLkGQ3X8mLjGP3Z8mLgGQ3]8lLeGQ3_8mLaGP3c9K4M4N1N3@^ElMd:T2>O001O001O0O2O002N3M3M3M3M3L4H8@`0[OS_e4"}, "label": "the image of the woman playing tennis on the left"}]}
{"id": 76740, "image": "COCO_train2014_000000076740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the left\"."}], "task": "refering", "answer_template": "The \"the woman on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 76, 77, 78, 99, 100, 101, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 167, 168, 169, 170, 171, 191, 192, 193, 213, 214, 215, 236, 237, 238, 260, 282, 283, 284, 304, 305, 306, 307, 328, 330], "bbox": [0.2185, 0.18223502304147465, 0.447734375, 0.9309447004608294], "iscrowd": 0, "area": 17668.953350000014, "rle": {"size": [434, 640], "counts": "W`k11a=0O101N10000O2O0JKoB5R=MkB3S=8N10000O2O000000000dHK[15eNL[13eNOZ10eN2Z1NfN4Y1KgN6X1KgN7X1HhN9X1GgN9Y1HfN9Z1GeN:Z1FgN9Z1GeN:Z1GfN9Z1FfN:Z1GfN9Z1GeN:Z1FgN:Y1FfN:Z1GdN;\\1EcN<\\1DcN=^1CaN>^1C`N?`1A^N`0c1_O[Nd0d1]OYNf0f1[OVNh0i1YOUNj0h1YOkK0UOj0n4XOdK8YOc0Q5XO]K?^O;d4[NaJl0Y1d0[O:i4YN`Jj0X1j0XO7P5WN^Jh0W1P1TO6V5UN[Jf0W1U1SO\\OQ6lNbId0W1Z1oNZOX6KeJR1lNVO_6IbJm2^5SM`Jn2`5SM_Jn2`5SM^Jn2b5SM\\Jo2c5RM[Jo2g5nLZJS3g5jLYJW3i5fLWJ\\3i5cLWJ]3k5`LTJc3f4ZKcKP1Fh3h4YKbKl0En3i4WKaKn0Al3P5VK_K[1POc3`5SK`Kb6]4_IcKQ5POWL[5hNeKo4SOZLU5hNgKn4WOZLo4iNjKl4YO[Ll4iNkKk4\\O\\Lg4iNmKi4_O_Lb4hNnKi4C_L\\4\\O]KU49`LU4^OaKR4=bLj3AhKl3`0fLM^Ng2T1jLg3f0iL]OdNS3m0hLf3j0bM]2iNgLe3o0`MZ2kNfLe3R1_MX2mNdLc3X1^MT2o3QNmKo1R4WNiKi1V4]NeKc1Y4dNbK]1\\4hN`KX1_4nN\\KR1c4TOXKl0f4ZOTKh0k4ZORKh0m4YOQKi0m4YOQKi0n4XOPKi0P5YOmJi0Q5YOmJi0R5XOlJj0S5WOkJk0S5XOjJj0T5XOiJk0U5WOiJl0T5VOjJm0S5VOiJn0g0jKW3Z3PLP1d0kKX3W3QLR1d0iKY3P8eLRHX3P8gLQHW3R8gLPHV3R8iLPHT3S8kLmGS3U8lLmGQ3V8mLkGQ3X8mLjGP3Z8mLgGQ3]8lLeGQ3_8mLaGP3c9K4M4N1N3@^ElMd:T2>O001O001O0O2O002N3M3M3M3M3L4H8@`0[OS_e4"}, "label": "the woman on the left"}]}
{"id": 183237, "image": "COCO_train2014_000000183237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaning over while sitting on a boat with a drink in hand\"."}], "task": "refering", "answer_template": "The \"a man leaning over while sitting on a boat with a drink in hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 86, 87, 88, 89, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.41481999999999997, 0.3101204819277108, 0.99674, 0.9887650602409638], "iscrowd": 0, "area": 40965.0237, "rle": {"size": [332, 500], "counts": "X\\S22m9?G9fGXOj6k0QIYOn6j0kH[OT7h0fH\\OX7h0bH\\O]7g0]H]Oa7g0XH^Og7d10000O1000000O1000000O1000000O100O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O10000O10000O10000O100O10000O10000O100O1000000O10000O10000O10001N1000000O10000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O100O100O100O100O100O100O100O1O100O100O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O10000O10000O1000000O10000O100I7H8L4O1N2O1O1O1O1N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1O1001O000000001O0000001O0001O0001O0000001O0000001O000000001N1O100O2N1O101N1O2N1O1bLXL7j3\\ObLc0_3\\OcLb0_3[OdLc0]3[OeLd0\\3YOhLg0X3UOkLm0S3POQMR1l2kNWMW1h2eN\\M]1a2`NcMa1\\2[NgMf1X2WNlMh1T2UNoMk1R2QNRNn1n1nMVNR2k1jMYNU2g1hM]NW2c1eMaN[2`1aMdN^2f1UM]Nk2d4N4L4L4L3MU="}, "label": "a man leaning over while sitting on a boat with a drink in hand"}]}
{"id": 183237, "image": "COCO_train2014_000000183237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person is sitting with glass\"."}], "task": "refering", "answer_template": "The \"a person is sitting with glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 86, 87, 88, 89, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.41481999999999997, 0.3101204819277108, 0.99674, 0.9887650602409638], "iscrowd": 0, "area": 40965.0237, "rle": {"size": [332, 500], "counts": "X\\S22m9?G9fGXOj6k0QIYOn6j0kH[OT7h0fH\\OX7h0bH\\O]7g0]H]Oa7g0XH^Og7d10000O1000000O1000000O1000000O100O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O10000O10000O10000O100O10000O10000O100O1000000O10000O10000O10001N1000000O10000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O100O100O100O100O100O100O100O1O100O100O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O10000O10000O1000000O10000O100I7H8L4O1N2O1O1O1O1N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1O1001O000000001O0000001O0001O0001O0000001O0000001O000000001N1O100O2N1O101N1O2N1O1bLXL7j3\\ObLc0_3\\OcLb0_3[OdLc0]3[OeLd0\\3YOhLg0X3UOkLm0S3POQMR1l2kNWMW1h2eN\\M]1a2`NcMa1\\2[NgMf1X2WNlMh1T2UNoMk1R2QNRNn1n1nMVNR2k1jMYNU2g1hM]NW2c1eMaN[2`1aMdN^2f1UM]Nk2d4N4L4L4L3MU="}, "label": "a person is sitting with glass"}]}
{"id": 183237, "image": "COCO_train2014_000000183237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on left\"."}], "task": "refering", "answer_template": "The \"woman on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 20, 21, 22, 37, 38, 39, 40, 55, 56, 57, 58, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 109, 110, 111, 127, 128, 129, 145, 146, 147, 163, 164, 165, 181, 182, 199, 200, 201], "bbox": [0.0, 0.04768072289156627, 0.30732, 0.9817771084337349], "iscrowd": 0, "area": 15124.591299999995, "rle": {"size": [332, 500], "counts": "^47T:3M2N1@EnF<S9EjF=U9DiF=W9DgF=Y9DeF=\\9CbF>^99O0000O2N000000000010N1O1O100O1O2O000010O0O10M3000000ULUOVNk0o0oNSL?i2b0l0YOYL7j2?l0\\OXL8j2<m0_OWL7j2:n0BVL9g25R1ElKa0P3IS1S1kNnNR1V1mNjNP1Z1nNgNo0]1nNeNcNSOg0\\2a0eN`N]Of0S2d0cN`NDf0m1e0cN]NKh0f1j0_NZN1f0e1o0]NTN5c0e0Y2[OjL8=f0m2W1dLQO`3j33M4L4LhL]Jn1`5SNSJ]O<V2]5]NXJA;P2Z5`N[JD:l1W5aN`JF7j1U5aNeJH5f1S5cNiJJ3c1Q5bNoJMOa1n4cNTKNM_1l4dNYKNK^1i4cN`K0F]1h4bNeK1C\\1g4bNjK1^O^1e4`NQL2ZO^1c4_NWL2VO]1n3mMWLb0i04RO\\1m3QNTL?P14nN[1n3TNQL=W13jNU1CYN\\44lK:_14fNP1H]NX45gK;d15eNh0NaNS49bK9j16bNd0c4]OnL0_Nb0c4\\OoL5]N>`4_OUM4[N<`4]OWM9XN:b4ZOnKDAN4l0J6e4[OkK:C7M3e4_OfK;F5N0f4CbK0@A7n00Ng4E_KOE_O4S1NIj4IZKNL]OOW1ODl4MVKL2\\OK^2n4\\OXKm0i4ROYKm0f4TO[Kk0e4UO\\Ki0b4b2N2O010000001O0100O10E:01N1O2I7I8G9G9ZMTIa1U7YNoHLLX1^7aNjH4MX1Z7aNkH50f0E_OS8HfH8\\O0n7F^I:c6C_I=X800000000001O00O1000000O100O10001O01O0000001O01OO10000000N2M2O2N2Me\\`3"}, "label": "woman on left"}]}
{"id": 183237, "image": "COCO_train2014_000000183237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman standing on the boat\"."}], "task": "refering", "answer_template": "The \"the woman standing on the boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 20, 21, 22, 37, 38, 39, 40, 55, 56, 57, 58, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 109, 110, 111, 127, 128, 129, 145, 146, 147, 163, 164, 165, 181, 182, 199, 200, 201], "bbox": [0.0, 0.04768072289156627, 0.30732, 0.9817771084337349], "iscrowd": 0, "area": 15124.591299999995, "rle": {"size": [332, 500], "counts": "^47T:3M2N1@EnF<S9EjF=U9DiF=W9DgF=Y9DeF=\\9CbF>^99O0000O2N000000000010N1O1O100O1O2O000010O0O10M3000000ULUOVNk0o0oNSL?i2b0l0YOYL7j2?l0\\OXL8j2<m0_OWL7j2:n0BVL9g25R1ElKa0P3IS1S1kNnNR1V1mNjNP1Z1nNgNo0]1nNeNcNSOg0\\2a0eN`N]Of0S2d0cN`NDf0m1e0cN]NKh0f1j0_NZN1f0e1o0]NTN5c0e0Y2[OjL8=f0m2W1dLQO`3j33M4L4LhL]Jn1`5SNSJ]O<V2]5]NXJA;P2Z5`N[JD:l1W5aN`JF7j1U5aNeJH5f1S5cNiJJ3c1Q5bNoJMOa1n4cNTKNM_1l4dNYKNK^1i4cN`K0F]1h4bNeK1C\\1g4bNjK1^O^1e4`NQL2ZO^1c4_NWL2VO]1n3mMWLb0i04RO\\1m3QNTL?P14nN[1n3TNQL=W13jNU1CYN\\44lK:_14fNP1H]NX45gK;d15eNh0NaNS49bK9j16bNd0c4]OnL0_Nb0c4\\OoL5]N>`4_OUM4[N<`4]OWM9XN:b4ZOnKDAN4l0J6e4[OkK:C7M3e4_OfK;F5N0f4CbK0@A7n00Ng4E_KOE_O4S1NIj4IZKNL]OOW1ODl4MVKL2\\OK^2n4\\OXKm0i4ROYKm0f4TO[Kk0e4UO\\Ki0b4b2N2O010000001O0100O10E:01N1O2I7I8G9G9ZMTIa1U7YNoHLLX1^7aNjH4MX1Z7aNkH50f0E_OS8HfH8\\O0n7F^I:c6C_I=X800000000001O00O1000000O100O10001O01O0000001O01OO10000000N2M2O2N2Me\\`3"}, "label": "the woman standing on the boat"}]}
{"id": 494534, "image": "COCO_train2014_000000494534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bicycle behind a bicyclist in blue\"."}], "task": "refering", "answer_template": "The \"a blue bicycle behind a bicyclist in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 72, 73, 74, 76, 77, 78, 90, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 162, 163, 164, 180, 181, 182, 198], "bbox": [0.008039999999999999, 0.25106666666666666, 0.38626000000000005, 0.87984], "iscrowd": 0, "area": 21577.528199999993, "rle": {"size": [375, 500], "counts": "nb11i4\\1g0eN]LT4e3lKRL[4o3gKPLW4R4jKmKT4U4mKjKQ4X4PLgKn3Z4SLfKm3Z4TLeKk3\\4VLcKj3\\4XLcKh3]4YLbKf3_4ZL`Kg3`4ZL_Kf3`4\\L_Kc3b4^L]Kb3c4_L\\Ka3d4_L\\K`3d4bL[K^3e4cLZK]3f4dLYK[3g4gLXKY3h4nLQKR3o4YMfJf2[5eMZJ[2e5l10O1000O01000O010000O01000PKVJR4j5nKXJQ4h5mK\\JQ4d5nK^JP4c5oK`Jo3_5PLdJo3\\5PLfJo3Z5oKjJn3W5QLkJn3U5QLnJm3R5QLPKn3Q5QLoJP4Q5oKPKQ4P5mKRKR4o4mKRKS4n4lKSKT4m4jKUKU4l4jKUKV4k4iKUKX4k4fKWKY4i4gKXKY4h4fKYKZ4R6O1000O010000O1000O01N2K5L4L4N2N1O2N2N2M3N2N1O2N2N2N2N2N2N1O2N2N1O2N20O01000O1000O10O10000O01000O1000O010000O10O10O100000O010000O10O10O00001N100O2O0O100O2OO01M3F:L40O0100O1O10O01001O1O2M2O2N1O2N1O2M2O2N2N1O2M2O2N1O2N1N3N2N1O2N2N4K4M4L4L3M4K4K6K4K6J6J5L5J5L5JUY`3"}, "label": "a blue bicycle behind a bicyclist in blue"}]}
{"id": 494534, "image": "COCO_train2014_000000494534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue bike\"."}], "task": "refering", "answer_template": "The \"blue bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 72, 73, 74, 76, 77, 78, 90, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 162, 163, 164, 180, 181, 182, 198], "bbox": [0.008039999999999999, 0.25106666666666666, 0.38626000000000005, 0.87984], "iscrowd": 0, "area": 21577.528199999993, "rle": {"size": [375, 500], "counts": "nb11i4\\1g0eN]LT4e3lKRL[4o3gKPLW4R4jKmKT4U4mKjKQ4X4PLgKn3Z4SLfKm3Z4TLeKk3\\4VLcKj3\\4XLcKh3]4YLbKf3_4ZL`Kg3`4ZL_Kf3`4\\L_Kc3b4^L]Kb3c4_L\\Ka3d4_L\\K`3d4bL[K^3e4cLZK]3f4dLYK[3g4gLXKY3h4nLQKR3o4YMfJf2[5eMZJ[2e5l10O1000O01000O010000O01000PKVJR4j5nKXJQ4h5mK\\JQ4d5nK^JP4c5oK`Jo3_5PLdJo3\\5PLfJo3Z5oKjJn3W5QLkJn3U5QLnJm3R5QLPKn3Q5QLoJP4Q5oKPKQ4P5mKRKR4o4mKRKS4n4lKSKT4m4jKUKU4l4jKUKV4k4iKUKX4k4fKWKY4i4gKXKY4h4fKYKZ4R6O1000O010000O1000O01N2K5L4L4N2N1O2N2N2M3N2N1O2N2N2N2N2N2N1O2N2N1O2N20O01000O1000O10O10000O01000O1000O010000O10O10O100000O010000O10O10O00001N100O2O0O100O2OO01M3F:L40O0100O1O10O01001O1O2M2O2N1O2N1O2M2O2N2N1O2M2O2N1O2N1N3N2N1O2N2N4K4M4L4L3M4K4K6K4K6J6J5L5J5L5JUY`3"}, "label": "blue bike"}]}
{"id": 494534, "image": "COCO_train2014_000000494534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"kids cycle parked on the lawn\"."}], "task": "refering", "answer_template": "The \"kids cycle parked on the lawn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 34, 35, 48, 49, 50, 51, 52, 53, 65, 67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 100, 101, 102, 104, 105, 106, 107, 119, 120, 123, 124], "bbox": [0.59392, 0.09208000000000001, 1.0, 0.4862666666666666], "iscrowd": 0, "area": 13636.6519, "rle": {"size": [375, 500], "counts": "mm\\31c;3N3M2N2M4M2N2N3J5I7H8J7J5J6J7J5J6J7O001NO2O1O1N2O1kMkFg1U9XNoFe1R9ZNQGb1Q9\\NSGa1n8]NVGb1i8]NZGa1f8]N^Gb1]9O1O1O1N01N2O1aN^NgHb1X7^NiHc1V7]NiHe1U7\\NkHc1V7\\NkHc1U7^NkHa1U7`NjH`1W7`NiH_1W7aNjH^1V7cNjH\\1W7dNhH\\1X7eNhHZ1Y7eNhHZ1X7gNgHY1Y7hNgHW1Z7hNgHW1Y7jNgHU1Y7lNfH:hN4c8BeH9QOLZ8KfH8YODP86gH5S8LlG4T8OjG0U83jGLV87hGGY8<dGD[8?dG@\\8c0bG\\O]8g0bGXO^8k0_GUOa8n0]GQOb8R1]GmNc8V1ZGjNf8X1VGiNj8[1PGhNP9[1jFhNU9[1fFhN^9f17I7I7I7I8H7H8KO001O1N2O0O2O1O1N1PFXOQ9i0iF^OU9c0dFDZ9>_FJ_96\\F1b9S1N3N1N3N2N1N3N1O2M3N1O2M2O2N2M2O2MmNT1]Nii3FbWLb0^Ob0SNZOWHW1e7lNlGe1S8P1N2O1N2O1O0O2O1N2O1N2O00100O10YOVHaMj7_2XH_Mg7b2\\H[Md7e2]HYMd7g2\\HYMc7i2]HVMc7j2]HVMb7k2_HTMa7l2`HSM`7m2`HRM`7o2aHPM_7P3aHoL_7R3bHjLa7V3_HgLd7a2_HhMl7Y2THgMl7Y2THgMl7Y2UHfMj7[2VHeMj7[2VHeMi7\\2XHcMh7]2XHcMh7]2XHdMf7]2ZHcMf7]2[HbMd7_2\\HaMd7_2_H^Ma7b2bH[M]7f2j0000O100O10000O102M3N2N1N3N2N2M3N2M3N2N0OM4L3M4L4Lf1ZNm0SOfL"}, "label": "kids cycle parked on the lawn"}]}
{"id": 494534, "image": "COCO_train2014_000000494534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle tire on the far right\"."}], "task": "refering", "answer_template": "The \"a bicycle tire on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197, 213, 214, 215], "bbox": [0.83294, 0.40941333333333335, 1.0, 0.9105866666666668], "iscrowd": 0, "area": 12391.013649999997, "rle": {"size": [375, 500], "counts": "gch42X;a0_O`0@a0_Oa0_Oa0_O`0K6K5K101N2N2O1N2N101N2N2N2O1N1O2O1N2N2O1N1O2O1QOdK`J^4_5gK[J[4d5eK\\J[4d5fKZJ\\4f5cKYJ_4f5aKYJ`4g5`KYJa4f5_KYJb4h5^KVJc4j5]KVJd4i5\\KVJe4j5[KUJf4l5YKTJg4l5ZKRJh4m5XKlIo4T6:10O00001O01O01O00010O000010O0001O01O000O100O2O0O100O2N100O101N100O101N1On1SNU2jMiF"}, "label": "a bicycle tire on the far right"}]}
{"id": 494534, "image": "COCO_train2014_000000494534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wheel of the bike to the right of the blue biker in the picture\"."}], "task": "refering", "answer_template": "The \"the wheel of the bike to the right of the blue biker in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197, 213, 214, 215], "bbox": [0.83294, 0.40941333333333335, 1.0, 0.9105866666666668], "iscrowd": 0, "area": 12391.013649999997, "rle": {"size": [375, 500], "counts": "gch42X;a0_O`0@a0_Oa0_Oa0_O`0K6K5K101N2N2O1N2N101N2N2N2O1N1O2O1N2N2O1N1O2O1QOdK`J^4_5gK[J[4d5eK\\J[4d5fKZJ\\4f5cKYJ_4f5aKYJ`4g5`KYJa4f5_KYJb4h5^KVJc4j5]KVJd4i5\\KVJe4j5[KUJf4l5YKTJg4l5ZKRJh4m5XKlIo4T6:10O00001O01O01O00010O000010O0001O01O000O100O2O0O100O2N100O101N100O101N1On1SNU2jMiF"}, "label": "the wheel of the bike to the right of the blue biker in the picture"}]}
{"id": 478148, "image": "COCO_train2014_000000478148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep nuzzling a baby sheep\"."}], "task": "refering", "answer_template": "The \"a sheep nuzzling a baby sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 240, 241, 242, 243, 244, 255, 256, 257, 258, 264, 265, 266, 267, 277, 278, 280, 281, 288, 289, 290, 300, 301, 303, 304, 311, 312, 313, 323, 324, 326, 327], "bbox": [0.003125, 0.00673076923076923, 0.665484375, 0.8718162393162393], "iscrowd": 0, "area": 112555.98495, "rle": {"size": [468, 640], "counts": "[m0Y1[=b2^Mb2^MZ1fN0000001O0000001O0000001O0000001O0000001O0000001O0000000000001O0000000YMTHfNl7W1YHgNg7W1]HgNc7W1aHgN`7U1eHiN[7U1iHiNW7T1mHkNS7S1oHmNQ7Q1RInNo6n0TIROl6k0XITOh6i0[IWOe6e0_I[Ob6a0bI^O^6?eIA[6<iICW69mIGT64PJLP6OUJ1k5KYJ5g5G]J9d5AaJ?_5]OeJc0\\5XOhJh0Y5ROlJn0V5lNnJT1S5fNRKZ1o4^NYKa1m80000000000000000000O1POPN\\EP2V:dNcE]1R:TOgEm0S:^OhEc0Q:FkE;o9X2K5N2M3N2N2N2M3O1O100O100002N1O1O2N1O2N1O3M2N3M2N3M4L4L4L4YLhEj2f;[O>B000000001O0000000000000000001O000000000000iKnMlKR2S4SNiKm1V4YNeKg1[4]NbKb1]4dN^K\\1a4iN[KW1e4mNWKS1h4SOSKm0m4XOnJh0Q5BfJ>Y5M]J3c58RJHm5c0iI]OW6b1kH]NT7g1iHYNW7j1fHVNY7m1eHSN[7P2bHPN^7S2_HmM`7V2^HjMb7Y2[HgMd7[2[HeMe7]2YHcMf7_2ZH`Mf7b2XH^Mh7c2WH]Mi7d2VH\\Mj7f2THZMl7g41O000000001O0000001O000000001O0000001O00O10000O100O10000O100O10000O100O10000O100O10000O10000O1000000O1000000O1000000O10000O1000000O100000000001O001O001O001O001O001O001O001O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O100001N1000001O0000001O000000001O0000001O000O101O000000001O0O10001O0000001O0O1000001O0000001O0000010O00000000O2O0000000O3N2N2N2N2M3N2N2N2N1N2O001O1O0O1000O101N1O101N100O2N100O2O0O101N1O2O0O2N2N2O1N2N2N2oHYIQ6i6kI[IT6k6cIYI[6P7ZIUIc6_7O2N2N101N2N1O2N1N3N1O1N101O001N2O001O0O2`LhGb0Y8\\OjGb0V8^OmG>U8AmG=S8BPH<Q8CQH;P8EPH9Q8InG6S8JnGOX82gGEa8=_GVN]O8\\9b1WGlMK6U9P2oFcM72Q9Z2S2O2N101N1O1N3M2M5L4K5L4L5B=@`0E^cQ3"}, "label": "a sheep nuzzling a baby sheep"}]}
{"id": 478148, "image": "COCO_train2014_000000478148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older sheep next to a baby\"."}], "task": "refering", "answer_template": "The \"an older sheep next to a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 240, 241, 242, 243, 244, 255, 256, 257, 258, 264, 265, 266, 267, 277, 278, 280, 281, 288, 289, 290, 300, 301, 303, 304, 311, 312, 313, 323, 324, 326, 327], "bbox": [0.003125, 0.00673076923076923, 0.665484375, 0.8718162393162393], "iscrowd": 0, "area": 112555.98495, "rle": {"size": [468, 640], "counts": "[m0Y1[=b2^Mb2^MZ1fN0000001O0000001O0000001O0000001O0000001O0000001O0000000000001O0000000YMTHfNl7W1YHgNg7W1]HgNc7W1aHgN`7U1eHiN[7U1iHiNW7T1mHkNS7S1oHmNQ7Q1RInNo6n0TIROl6k0XITOh6i0[IWOe6e0_I[Ob6a0bI^O^6?eIA[6<iICW69mIGT64PJLP6OUJ1k5KYJ5g5G]J9d5AaJ?_5]OeJc0\\5XOhJh0Y5ROlJn0V5lNnJT1S5fNRKZ1o4^NYKa1m80000000000000000000O1POPN\\EP2V:dNcE]1R:TOgEm0S:^OhEc0Q:FkE;o9X2K5N2M3N2N2N2M3O1O100O100002N1O1O2N1O2N1O3M2N3M2N3M4L4L4L4YLhEj2f;[O>B000000001O0000000000000000001O000000000000iKnMlKR2S4SNiKm1V4YNeKg1[4]NbKb1]4dN^K\\1a4iN[KW1e4mNWKS1h4SOSKm0m4XOnJh0Q5BfJ>Y5M]J3c58RJHm5c0iI]OW6b1kH]NT7g1iHYNW7j1fHVNY7m1eHSN[7P2bHPN^7S2_HmM`7V2^HjMb7Y2[HgMd7[2[HeMe7]2YHcMf7_2ZH`Mf7b2XH^Mh7c2WH]Mi7d2VH\\Mj7f2THZMl7g41O000000001O0000001O000000001O0000001O00O10000O100O10000O100O10000O100O10000O100O10000O10000O1000000O1000000O1000000O10000O1000000O100000000001O001O001O001O001O001O001O001O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O100001N1000001O0000001O000000001O0000001O000O101O000000001O0O10001O0000001O0O1000001O0000001O0000010O00000000O2O0000000O3N2N2N2N2M3N2N2N2N1N2O001O1O0O1000O101N1O101N100O2N100O2O0O101N1O2O0O2N2N2O1N2N2N2oHYIQ6i6kI[IT6k6cIYI[6P7ZIUIc6_7O2N2N101N2N1O2N1N3N1O1N101O001N2O001O0O2`LhGb0Y8\\OjGb0V8^OmG>U8AmG=S8BPH<Q8CQH;P8EPH9Q8InG6S8JnGOX82gGEa8=_GVN]O8\\9b1WGlMK6U9P2oFcM72Q9Z2S2O2N101N1O1N3M2M5L4K5L4L5B=@`0E^cQ3"}, "label": "an older sheep next to a baby"}]}
{"id": 478148, "image": "COCO_train2014_000000478148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a newborn lamb laying down in the hay\"."}], "task": "refering", "answer_template": "The \"a newborn lamb laying down in the hay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 224, 225, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.6117343749999999, 0.5541880341880342, 0.9348281249999999, 0.8728632478632479], "iscrowd": 0, "area": 21601.699699999997, "rle": {"size": [468, 640], "counts": "Xac52a>2RL1TI0k6<hHGV7:iHGV7;hHGV7;gHGX7;fHGW7<gHEZ7:fHF[7:cHH\\79cHG^79`HH`79_HGb79eFVOd0c0g86cF]Ob0=k87`FB`08Q96[FI`01U97XFN>MZ97PF5b0D_9;cE<i0ZOc9l1\\FTNe9m1YFTNg9m1WFSNj9n1TFRNm9o1QFQNP:o1kEVNU:k1bE\\N_:l21O100O1O0001O01O0001O0001O01OO101N100O100O2O0O1001O01O0001O01O01O0000010O00000O2O000O2O0O10001N10O10O10000O10000O100O10000O10O1N1N3M21O1O2N101N1O2N1O2N1O2N101N1O1O2O02O1N2O1M3N1O2N2N5K8HO100000000O1000000O010000000O_OXEXMh:h2YEWMg:i2[EUMe:j2^ETMa:m2aEQM_:o2c000O10O1000O10O11N10001O0010O00010O00010O00010O01O00001O001N101O001O0O2O001O2N2M3N2N2N2N2M3N1O2N2N2M3N2N2N2N2M2O2N1L5I6I8H7J7H8Ihkb0"}, "label": "a newborn lamb laying down in the hay"}]}
{"id": 478148, "image": "COCO_train2014_000000478148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby sheep lying next to it ' s mom\"."}], "task": "refering", "answer_template": "The \"a baby sheep lying next to it ' s mom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 224, 225, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.6117343749999999, 0.5541880341880342, 0.9348281249999999, 0.8728632478632479], "iscrowd": 0, "area": 21601.699699999997, "rle": {"size": [468, 640], "counts": "Xac52a>2RL1TI0k6<hHGV7:iHGV7;hHGV7;gHGX7;fHGW7<gHEZ7:fHF[7:cHH\\79cHG^79`HH`79_HGb79eFVOd0c0g86cF]Ob0=k87`FB`08Q96[FI`01U97XFN>MZ97PF5b0D_9;cE<i0ZOc9l1\\FTNe9m1YFTNg9m1WFSNj9n1TFRNm9o1QFQNP:o1kEVNU:k1bE\\N_:l21O100O1O0001O01O0001O0001O01OO101N100O100O2O0O1001O01O0001O01O01O0000010O00000O2O000O2O0O10001N10O10O10000O10000O100O10000O10O1N1N3M21O1O2N101N1O2N1O2N1O2N101N1O1O2O02O1N2O1M3N1O2N2N5K8HO100000000O1000000O010000000O_OXEXMh:h2YEWMg:i2[EUMe:j2^ETMa:m2aEQM_:o2c000O10O1000O10O11N10001O0010O00010O00010O00010O01O00001O001N101O001O0O2O001O2N2M3N2N2N2N2M3N1O2N2N2M3N2N2N2N2M2O2N1L5I6I8H7J7H8Ihkb0"}, "label": "a baby sheep lying next to it ' s mom"}]}
{"id": 52168, "image": "COCO_train2014_000000052168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large cow to the left of another\"."}], "task": "refering", "answer_template": "The \"a large cow to the left of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 149, 150, 151, 152, 153, 164, 165, 170, 171, 182, 183, 188], "bbox": [0.10321999999999999, 0.34408, 0.52742, 0.8408533333333333], "iscrowd": 0, "area": 24319.735149999997, "rle": {"size": [375, 500], "counts": "gVc09Y;8H9I6J6J7I6I7L4M3L4L4M2M2N2N2O2M2O1N2O1O1iHbMS5a2hJgMR5^2fJdMZ5c2^J^Mc5h2UJYMl5n2jITMW6V3]IkLc6R43L4L4M2M4LQImKb6P4ZIZLc6b3[IfLc6V3[IPMe6R4O000000001O00000O10000000O100XNkInMU6j1^JjMc5P2hJkMX5U2nJeMR5[2RK`Mo4`2TK]Ml4d2UKZMk4f2XKVMi4j2YKTMg4l2R21O0000000O2O000000001O0O101O00001O000O2O00001O00001O0000001O000O1000000000000000001O0000000000000001O0001O0001O0001O01O000000010O00001O01000O100000O10000O1O1O1O100O001O01O0001O0000001O0000001O0000000000001N1000000PIcLf5]3WJiLf5W3WJPMe5P3XJSMh5n2QJXMo5k2bIaM^6l3O10000000O11O1O2N1O2M2O00001O000O10L4K5L4K5K6K4K6K4K6bNaH]Nd7_1kHSNX7X1`HSNS12_6g1P2N4K5L4L6J6J6AnXf2"}, "label": "a large cow to the left of another"}]}
{"id": 52168, "image": "COCO_train2014_000000052168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow on the left\"."}], "task": "refering", "answer_template": "The \"the cow on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 149, 150, 151, 152, 153, 164, 165, 170, 171, 182, 183, 188], "bbox": [0.10321999999999999, 0.34408, 0.52742, 0.8408533333333333], "iscrowd": 0, "area": 24319.735149999997, "rle": {"size": [375, 500], "counts": "gVc09Y;8H9I6J6J7I6I7L4M3L4L4M2M2N2N2O2M2O1N2O1O1iHbMS5a2hJgMR5^2fJdMZ5c2^J^Mc5h2UJYMl5n2jITMW6V3]IkLc6R43L4L4M2M4LQImKb6P4ZIZLc6b3[IfLc6V3[IPMe6R4O000000001O00000O10000000O100XNkInMU6j1^JjMc5P2hJkMX5U2nJeMR5[2RK`Mo4`2TK]Ml4d2UKZMk4f2XKVMi4j2YKTMg4l2R21O0000000O2O000000001O0O101O00001O000O2O00001O00001O0000001O000O1000000000000000001O0000000000000001O0001O0001O0001O01O000000010O00001O01000O100000O10000O1O1O1O100O001O01O0001O0000001O0000001O0000000000001N1000000PIcLf5]3WJiLf5W3WJPMe5P3XJSMh5n2QJXMo5k2bIaM^6l3O10000000O11O1O2N1O2M2O00001O000O10L4K5L4K5K6K4K6K4K6bNaH]Nd7_1kHSNX7X1`HSNS12_6g1P2N4K5L4L6J6J6AnXf2"}, "label": "the cow on the left"}]}
{"id": 52168, "image": "COCO_train2014_000000052168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow to the far right\"."}], "task": "refering", "answer_template": "The \"cow to the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 156, 157, 160, 161, 175, 177, 178], "bbox": [0.6431, 0.40792, 0.96454, 0.7497333333333333], "iscrowd": 0, "area": 12234.019300000004, "rle": {"size": [375, 500], "counts": "QSf36`;5J3N2M3N3L4M;E5J5J5J5N3M3M3N1N2O1O001O0O2O001O001O0O2O00iNW11O000000000000Q2nElL_7m2fHXMW7d2kH]MV7j2`HYM`7_3000000000XNbHTNEi0i7R1fHQNJe0`7Y1[I_Nh6a1\\IUNj6k1ZIQNf6o1a10000001O001O001O00001O000000001O001O00010O00001O0010O00000000000000000000001O000O100000000O1000001N1000000O100001O000010O01O0QHlMZ6U2eImMZ6T2dImM[6T2eIlM[6V2VHhMY14`6X2^IiMb6W2]IjMc6R2`IoM`6Q2`IoM`6Q2^IQNb6o1\\ISNc6n1ZIUNf6l1WIVNi6k1TIVNm6n1lHUNT7n1gHSNZ7Q2_HRNa7Q2WHUNh7m200000000000O1O1N3N1001O00L4L4L4K6K4K6J5M4M3M2N3L4M4ZOdFjNc9;XG^Oaf6"}, "label": "cow to the far right"}]}
{"id": 52168, "image": "COCO_train2014_000000052168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow on right\"."}], "task": "refering", "answer_template": "The \"cow on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 156, 157, 160, 161, 175, 177, 178], "bbox": [0.6431, 0.40792, 0.96454, 0.7497333333333333], "iscrowd": 0, "area": 12234.019300000004, "rle": {"size": [375, 500], "counts": "QSf36`;5J3N2M3N3L4M;E5J5J5J5N3M3M3N1N2O1O001O0O2O001O001O0O2O00iNW11O000000000000Q2nElL_7m2fHXMW7d2kH]MV7j2`HYM`7_3000000000XNbHTNEi0i7R1fHQNJe0`7Y1[I_Nh6a1\\IUNj6k1ZIQNf6o1a10000001O001O001O00001O000000001O001O00010O00001O0010O00000000000000000000001O000O100000000O1000001N1000000O100001O000010O01O0QHlMZ6U2eImMZ6T2dImM[6T2eIlM[6V2VHhMY14`6X2^IiMb6W2]IjMc6R2`IoM`6Q2`IoM`6Q2^IQNb6o1\\ISNc6n1ZIUNf6l1WIVNi6k1TIVNm6n1lHUNT7n1gHSNZ7Q2_HRNa7Q2WHUNh7m200000000000O1O1N3N1001O00L4L4L4K6K4K6J5M4M3M2N3L4M4ZOdFjNc9;XG^Oaf6"}, "label": "cow on right"}]}
{"id": 379853, "image": "COCO_train2014_000000379853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in the blue in front\"."}], "task": "refering", "answer_template": "The \"man in the blue in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 247, 248, 249, 252, 253, 262, 263, 267, 268, 276, 277, 283, 290, 291, 292, 297, 298, 299, 312, 313], "bbox": [0.3661592505854801, 0.403234375, 0.9436065573770491, 0.8825156249999999], "iscrowd": 0, "area": 25037.216649999995, "rle": {"size": [640, 427], "counts": "odR33lc03M2N2O1N2N2N1O100O1O10O01O100O1O100O1O10O01O1O100O1O1O1O0O2N2N2N2N2N2O1N1O2N2M3N2M3M3M3M2N3M3N2M3M3M3M3N1N3M3M3M3M3N2M3MmNc_OiNZ`0X1Q@^Nm?d1^@RN_?o1X10dLQNhDQ2R;ZNhDf1W;aNcD`1\\;eN`DZ1`;lNZDT1f;ROUDn0k;WOoCi0R<\\OiCc0Y<@bCa0_<_O_Ca0b<@]C`0d<@ZC`0h<@WC0Z=0kAVN3W1g>c0UA_NJo0V>lNSBf1LhNBg0^>UOkA\\147Q>gNaAS1=6R>i0mAWOR>j0nAUOQ>m0oAROQ>o0oAQOP>P1PBoNP>R1PBnNo=S1QBoNm=Q1SBPOk=Q1UBPOj=P1VBROh=n0XBSOg=m0YBTOf=l0ZBVOc=k0]BVOb=j0^BXO`=h0`BYO_=g0aB[O]=e0cB]OZ=d0fB^OX=b0hB@V=`0jBBT=>lBDR=<nBEP=<PCFn<:RCGm<9SCGm<9SCGn<8RCHo<7QCHP=8PCGR=8nBFT=:lBEV=:jBDX=<hBCZ=<fBC]=;cBCa=;_BCe=;[BCi=;WBBn=<RBBS>;mACY>9gAD_>9aAEd>8\\AGh>6XAIi>7WAHj>8VAGj>:VAEk>;UADl><TADk>=UABl>>TAAm>?SA@m>a0SA^On>b0RA]Oo>c0QA\\Oo>e0QAZOP?f0PAYOQ?g0o@XOQ?i0o@VOR?j0n@UOS?k0m@TOS?m0m@ROT?n0l@QOU?o0k@POV?P1j@oNV?R1j@mNW?S1i@lNX?T1h@kNZ?T1f@kN\\?T1d@kN^?T1b@kN`?T1`@jNc?U1]@jNd?V1\\@hNg?W1Y@iNh?V1X@iNj?V1V@iNl?V1T@iNn?V1R@jNo?U1Q@jNQ`0U1o_OjNS`0U1m_OjNT`0V1l_OjNU`0U1k_OjNW`0U1i_OlNW`0S1i_OnNW`0Q1i_OPOW`0o0i_OSOV`0l0j_OUOV`0j0j_OWOJ_Oa?Z1e@3Z?Lf@6Y?Jf@8Y?Gg@:Y?Eg@=X?Bh@`0W?_Oi@c0V?\\Oj@f0U?YOk@i0S?WOm@k0R?TOn@n0Q?QOo@Q1P?nNPAT1o>kNQAW1n>hNRAZ1m>eNSA]1l>bNTA`1j>`NVAb1i>]NWAe1h>ZNXAh1g>WNYAk1f>TNZAn1e>QN[AP2e>oM[AS2d>lM\\AV2c>iM]AY2a>gM_A[2`>dM`A3EP1k>lN`A03l0\\>SOaAJa0n0o0[Nk:i0hCBj0\\1`0^Nk::TDBl0f13aNl:NZFb1fNbNQ;H]Fg1_NaNV;E]Fl1YNaNd>a1XA`Ni>b1RA`NP?b1j@`NX?a1a@cN`?_1Z@cNi?^1P@fNR`0[1h_OhNZ`0^22N2N2N2M4M5K4L5K5K4L5K4L5J5L5K5K2N3M2N3M2O2M2N3M4L4L4I7GP_>"}, "label": "man in the blue in front"}]}
{"id": 379853, "image": "COCO_train2014_000000379853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with buzzed hair playing tennis crouched down near the net holding a racket with yellow strings with both hands\"."}], "task": "refering", "answer_template": "The \"a man with buzzed hair playing tennis crouched down near the net holding a racket with yellow strings with both hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 247, 248, 249, 252, 253, 262, 263, 267, 268, 276, 277, 283, 290, 291, 292, 297, 298, 299, 312, 313], "bbox": [0.3661592505854801, 0.403234375, 0.9436065573770491, 0.8825156249999999], "iscrowd": 0, "area": 25037.216649999995, "rle": {"size": [640, 427], "counts": "odR33lc03M2N2O1N2N2N1O100O1O10O01O100O1O100O1O10O01O1O100O1O1O1O0O2N2N2N2N2N2O1N1O2N2M3N2M3M3M3M2N3M3N2M3M3M3M3N1N3M3M3M3M3N2M3MmNc_OiNZ`0X1Q@^Nm?d1^@RN_?o1X10dLQNhDQ2R;ZNhDf1W;aNcD`1\\;eN`DZ1`;lNZDT1f;ROUDn0k;WOoCi0R<\\OiCc0Y<@bCa0_<_O_Ca0b<@]C`0d<@ZC`0h<@WC0Z=0kAVN3W1g>c0UA_NJo0V>lNSBf1LhNBg0^>UOkA\\147Q>gNaAS1=6R>i0mAWOR>j0nAUOQ>m0oAROQ>o0oAQOP>P1PBoNP>R1PBnNo=S1QBoNm=Q1SBPOk=Q1UBPOj=P1VBROh=n0XBSOg=m0YBTOf=l0ZBVOc=k0]BVOb=j0^BXO`=h0`BYO_=g0aB[O]=e0cB]OZ=d0fB^OX=b0hB@V=`0jBBT=>lBDR=<nBEP=<PCFn<:RCGm<9SCGm<9SCGn<8RCHo<7QCHP=8PCGR=8nBFT=:lBEV=:jBDX=<hBCZ=<fBC]=;cBCa=;_BCe=;[BCi=;WBBn=<RBBS>;mACY>9gAD_>9aAEd>8\\AGh>6XAIi>7WAHj>8VAGj>:VAEk>;UADl><TADk>=UABl>>TAAm>?SA@m>a0SA^On>b0RA]Oo>c0QA\\Oo>e0QAZOP?f0PAYOQ?g0o@XOQ?i0o@VOR?j0n@UOS?k0m@TOS?m0m@ROT?n0l@QOU?o0k@POV?P1j@oNV?R1j@mNW?S1i@lNX?T1h@kNZ?T1f@kN\\?T1d@kN^?T1b@kN`?T1`@jNc?U1]@jNd?V1\\@hNg?W1Y@iNh?V1X@iNj?V1V@iNl?V1T@iNn?V1R@jNo?U1Q@jNQ`0U1o_OjNS`0U1m_OjNT`0V1l_OjNU`0U1k_OjNW`0U1i_OlNW`0S1i_OnNW`0Q1i_OPOW`0o0i_OSOV`0l0j_OUOV`0j0j_OWOJ_Oa?Z1e@3Z?Lf@6Y?Jf@8Y?Gg@:Y?Eg@=X?Bh@`0W?_Oi@c0V?\\Oj@f0U?YOk@i0S?WOm@k0R?TOn@n0Q?QOo@Q1P?nNPAT1o>kNQAW1n>hNRAZ1m>eNSA]1l>bNTA`1j>`NVAb1i>]NWAe1h>ZNXAh1g>WNYAk1f>TNZAn1e>QN[AP2e>oM[AS2d>lM\\AV2c>iM]AY2a>gM_A[2`>dM`A3EP1k>lN`A03l0\\>SOaAJa0n0o0[Nk:i0hCBj0\\1`0^Nk::TDBl0f13aNl:NZFb1fNbNQ;H]Fg1_NaNV;E]Fl1YNaNd>a1XA`Ni>b1RA`NP?b1j@`NX?a1a@cN`?_1Z@cNi?^1P@fNR`0[1h_OhNZ`0^22N2N2N2M4M5K4L5K5K4L5K4L5J5L5K5K2N3M2N3M2O2M2N3M4L4L4I7GP_>"}, "label": "a man with buzzed hair playing tennis crouched down near the net holding a racket with yellow strings with both hands"}]}
{"id": 183247, "image": "COCO_train2014_000000183247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bearded man playing with a wii controller\"."}], "task": "refering", "answer_template": "The \"a bearded man playing with a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 378, 379, 380, 382, 383, 384], "bbox": [0.268578125, 0.025895833333333333, 0.7297343749999999, 0.989875], "iscrowd": 0, "area": 62891.41185000001, "rle": {"size": [480, 640], "counts": "bl`25j>8I4L0O2O1N2O1O1N2O001N101O0O10001O00001N10001O00000O2O0000O1O1O001O2N1O1O1O1O1O1O1O1O100000000O100000000000O01000O10000O01000O10000O10O10O10000O1000O0100O10000O01000O10000O10O10O10000O1000O010000O10000O010O1O100O10O01O0O2O0O2N100O2N1O2O0O1O2K4I8H7I8\\KlMTL[2MhMC15^23GfMQNk1`2:DjMPNg1]2:JmMmMe1Z28ORNlMa1T2:6SNiM`1Q2;9SNkM^1l1>;SNlMQOlNk1m2P1=SNnMmNlNm1h2T1?QNPNiNoNn1b2W1a0QNaNe0n0Y1c0PNdNc0i0\\1e0PNdNb0f0_1g0nMfN`0c0a1i0nMgN>`0c1k0nMhN<=e1m0mMjN;8i1o0kMkN:6j1Q1kMoN50o1S1kMSOOKU2T1kMWOJD\\2V1iM\\OEGX2o0SN@^O2m1`0eNCXO?`10XO^47dKGn4FTK:i9000000O10000000000O100000000O10000000000O10000000O0O2N2N2M3N2N1O2M3N2N2N1N3N2N2N2M3N1O2M3N2N20O3O1N101N1O2O1N101N101N101N2O0O2O0O2N2O0O2O0O2O0O2O0O03nDjNT8[1fGfNZ8_1`GcN_8a1\\GaNc8d1WG]Nj8g1PG[No8j1kFXNT9m1fFTNZ9Q2`FQN_9T2WFRNh9S2oEQNQ:T2eERN[:R2[ETNd:T301O1O0010O01O1O001O1O010O1O001O1O00100O001O1O00100O1O2N1O2N1lJXKNj4MdKI\\44QLAP4;XM`Nj2[1cM[N^2a1oMUNS2g1YNbMU2Z2XNmLX2o2UNXL\\2d3b4L5J5J6J7I6K6I6J6J7I6L5L3^Ob0WORn`2"}, "label": "a bearded man playing with a wii controller"}]}
{"id": 183247, "image": "COCO_train2014_000000183247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing wii\"."}], "task": "refering", "answer_template": "The \"a man playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 378, 379, 380, 382, 383, 384], "bbox": [0.268578125, 0.025895833333333333, 0.7297343749999999, 0.989875], "iscrowd": 0, "area": 62891.41185000001, "rle": {"size": [480, 640], "counts": "bl`25j>8I4L0O2O1N2O1O1N2O001N101O0O10001O00001N10001O00000O2O0000O1O1O001O2N1O1O1O1O1O1O1O1O100000000O100000000000O01000O10000O01000O10000O10O10O10000O1000O0100O10000O01000O10000O10O10O10000O1000O010000O10000O010O1O100O10O01O0O2O0O2N100O2N1O2O0O1O2K4I8H7I8\\KlMTL[2MhMC15^23GfMQNk1`2:DjMPNg1]2:JmMmMe1Z28ORNlMa1T2:6SNiM`1Q2;9SNkM^1l1>;SNlMQOlNk1m2P1=SNnMmNlNm1h2T1?QNPNiNoNn1b2W1a0QNaNe0n0Y1c0PNdNc0i0\\1e0PNdNb0f0_1g0nMfN`0c0a1i0nMgN>`0c1k0nMhN<=e1m0mMjN;8i1o0kMkN:6j1Q1kMoN50o1S1kMSOOKU2T1kMWOJD\\2V1iM\\OEGX2o0SN@^O2m1`0eNCXO?`10XO^47dKGn4FTK:i9000000O10000000000O100000000O10000000000O10000000O0O2N2N2M3N2N1O2M3N2N2N1N3N2N2N2M3N1O2M3N2N20O3O1N101N1O2O1N101N101N101N2O0O2O0O2N2O0O2O0O2O0O2O0O03nDjNT8[1fGfNZ8_1`GcN_8a1\\GaNc8d1WG]Nj8g1PG[No8j1kFXNT9m1fFTNZ9Q2`FQN_9T2WFRNh9S2oEQNQ:T2eERN[:R2[ETNd:T301O1O0010O01O1O001O1O010O1O001O1O00100O001O1O00100O1O2N1O2N1lJXKNj4MdKI\\44QLAP4;XM`Nj2[1cM[N^2a1oMUNS2g1YNbMU2Z2XNmLX2o2UNXL\\2d3b4L5J5J6J7I6K6I6J6J7I6L5L3^Ob0WORn`2"}, "label": "a man playing wii"}]}
{"id": 142287, "image": "COCO_train2014_000000142287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large giraffe in the foreground\"."}], "task": "refering", "answer_template": "The \"a large giraffe in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 119, 120, 121, 122, 141, 142, 144, 145, 146, 168, 169, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 332, 335, 336, 337, 382], "bbox": [0.165546875, 0.2263541666666667, 0.699328125, 0.9876041666666666], "iscrowd": 0, "area": 30101.40245, "rle": {"size": [480, 640], "counts": "Zka16d>601N1O2O0O1O2OO01O01O01O010N101O001O1O1O1O001O1N2O1N101N2O1N2L3N300000O100000O1O0O2O1N2N2O0O2O1N2O1N13N3M3M3L4M2M2O1N101N2O1N2O0011O001O001N10001O001N101O00001N101O001O0O101O001O000O2O001O001N10001O001N101O00001N101O00001O0O2O001O000O2O00001O0O100000001O0O1000001O0O1000001O0O100000001N100000001O0O1000001O0O1000001O0fDiMi9X2nERNQ:n1gE[NX:e1_EeN`:[1XEoNf:g2O1N2O1O3M3M3L4M3M3M2N3L4M3M3M3M3L4MK5N2N2N2N1O2N2N1O1O2N1O1O1O2N100O101O0000001O00001O0O10001N2O001N2O1O1N10100O1O010O100O1O010O1O10O0100O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O0O2O1N101O1N2O0O2O1O1N101N2O1N1102N1N3N1O2N1N3N1O2M2O2M2N2OO0000001OYFXLV8h3hG[LW8e3fG`LY8_3dGeL[8[3cGhL\\8X3cGjL]8U3dGkL[8U3eGlLZ8T3fGmLY8S3hGlLY8S3gGjL\\8V3dGhL^8X3cGeL`8Z3`GdLb8\\3^GaLe8_3\\G^Lg8a3YG]Li8c3WGZLm8e3Y1N1O1O1O100O1N2N2N2M3N2M3N2N2M3N2N2M3N2M3N2M3M4M4K4M3L4L4M3L5K4M37Ik0UOYOg0YOg0ZOeei2"}, "label": "a large giraffe in the foreground"}]}
{"id": 142287, "image": "COCO_train2014_000000142287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large giraffe standing in front of a small giraffe\"."}], "task": "refering", "answer_template": "The \"a large giraffe standing in front of a small giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 119, 120, 121, 122, 141, 142, 144, 145, 146, 168, 169, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 332, 335, 336, 337, 382], "bbox": [0.165546875, 0.2263541666666667, 0.699328125, 0.9876041666666666], "iscrowd": 0, "area": 30101.40245, "rle": {"size": [480, 640], "counts": "Zka16d>601N1O2O0O1O2OO01O01O01O010N101O001O1O1O1O001O1N2O1N101N2O1N2L3N300000O100000O1O0O2O1N2N2O0O2O1N2O1N13N3M3M3L4M2M2O1N101N2O1N2O0011O001O001N10001O001N101O00001N101O001O0O101O001O000O2O001O001N10001O001N101O00001N101O00001O0O2O001O000O2O00001O0O100000001O0O1000001O0O1000001O0O100000001N100000001O0O1000001O0O1000001O0fDiMi9X2nERNQ:n1gE[NX:e1_EeN`:[1XEoNf:g2O1N2O1O3M3M3L4M3M3M2N3L4M3M3M3M3L4MK5N2N2N2N1O2N2N1O1O2N1O1O1O2N100O101O0000001O00001O0O10001N2O001N2O1O1N10100O1O010O100O1O010O1O10O0100O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O0O2O1N101O1N2O0O2O1O1N101N2O1N1102N1N3N1O2N1N3N1O2M2O2M2N2OO0000001OYFXLV8h3hG[LW8e3fG`LY8_3dGeL[8[3cGhL\\8X3cGjL]8U3dGkL[8U3eGlLZ8T3fGmLY8S3hGlLY8S3gGjL\\8V3dGhL^8X3cGeL`8Z3`GdLb8\\3^GaLe8_3\\G^Lg8a3YG]Li8c3WGZLm8e3Y1N1O1O1O100O1N2N2N2M3N2M3N2N2M3N2N2M3N2M3N2M3M4M4K4M3L4L4M3L5K4M37Ik0UOYOg0YOg0ZOeei2"}, "label": "a large giraffe standing in front of a small giraffe"}]}
{"id": 273362, "image": "COCO_train2014_000000273362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in shorts and a yellow shirt catching a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in shorts and a yellow shirt catching a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 81, 82, 83, 91, 92, 93, 94, 95, 102, 103, 104, 105, 106, 112, 113, 114, 115, 116, 117, 123, 124, 125, 126, 127, 128, 135, 136, 137, 138, 146, 147, 148, 149, 157, 158, 159, 160, 168, 169, 170, 171, 179, 180, 181, 182, 190, 191, 192, 202, 203, 213, 214, 224], "bbox": [0.25943037974683547, 0.2720625, 0.7100000000000001, 0.89453125], "iscrowd": 0, "area": 32516.077599999993, "rle": {"size": [640, 316], "counts": "eac1<\\c0e0[O;C=S_O_Nl>n1o@_Na>h1[AdNV>c1fAjNj=]1RBoN_=X1^BSOV=R1eBQOX=S1dBoNZ=U1bBlN]=X1_BjN_=Z1]BgNb=m3N1O2O1N3M4L4L3M4M2M4M3L3N3L3M4M3M2M4M2M4nDmGQ:V8]EeHV:o8D<C5L0O100O1O1O1O1O100O2N1O1O1O1O1O1O1F:\\Od0N2OCQHkDn7T;VHjDi7T;\\HjDc7T;bHjD]7U;hHiDW7T;nHjDQ7U;m0O11O010O1O1O100O10O0O2I7I6J7J5K6K4J8D;K6M2M4M2M4M2N3L3M4]Ob0E<K4M4K4L5L5L5L4L3SJUC_4o<\\KUCb4n<YKUCe4o<UKVCi4m<RKVCm4P>M3N1O2K4H9G9jNU1O2N2M2O2N1N3N2N1O2N1O2N2N1O2O0O2N3M3M3M3J6K5K5K5J6CmWi1"}, "label": "a man in shorts and a yellow shirt catching a frisbee"}]}
{"id": 273362, "image": "COCO_train2014_000000273362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow shirt\"."}], "task": "refering", "answer_template": "The \"man in yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 81, 82, 83, 91, 92, 93, 94, 95, 102, 103, 104, 105, 106, 112, 113, 114, 115, 116, 117, 123, 124, 125, 126, 127, 128, 135, 136, 137, 138, 146, 147, 148, 149, 157, 158, 159, 160, 168, 169, 170, 171, 179, 180, 181, 182, 190, 191, 192, 202, 203, 213, 214, 224], "bbox": [0.25943037974683547, 0.2720625, 0.7100000000000001, 0.89453125], "iscrowd": 0, "area": 32516.077599999993, "rle": {"size": [640, 316], "counts": "eac1<\\c0e0[O;C=S_O_Nl>n1o@_Na>h1[AdNV>c1fAjNj=]1RBoN_=X1^BSOV=R1eBQOX=S1dBoNZ=U1bBlN]=X1_BjN_=Z1]BgNb=m3N1O2O1N3M4L4L3M4M2M4M3L3N3L3M4M3M2M4M2M4nDmGQ:V8]EeHV:o8D<C5L0O100O1O1O1O1O100O2N1O1O1O1O1O1O1F:\\Od0N2OCQHkDn7T;VHjDi7T;\\HjDc7T;bHjD]7U;hHiDW7T;nHjDQ7U;m0O11O010O1O1O100O10O0O2I7I6J7J5K6K4J8D;K6M2M4M2M4M2N3L3M4]Ob0E<K4M4K4L5L5L5L4L3SJUC_4o<\\KUCb4n<YKUCe4o<UKVCi4m<RKVCm4P>M3N1O2K4H9G9jNU1O2N2M2O2N1N3N2N1O2N1O2N2N1O2O0O2N3M3M3M3J6K5K5K5J6CmWi1"}, "label": "man in yellow shirt"}]}
{"id": 52179, "image": "COCO_train2014_000000052179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holds a fork as she eats a pizza\"."}], "task": "refering", "answer_template": "The \"a woman holds a fork as she eats a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 64, 65, 66, 78, 79, 80, 81, 94, 95, 96, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 136, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 181], "bbox": [0.06077283372365339, 0.141890625, 0.4523419203747072, 0.5270312500000001], "iscrowd": 0, "area": 24980.965099999998, "rle": {"size": [640, 427], "counts": "ka`07\\c0>C<C>F:M3M2N3N2M3M3M2N3N2M3M3M2O2M3M3M2O1N2M3L4L3M4L4L4N2000O010000O100O01000O100O10000O100O01000O100O10000O100O10O1000O100000000O010000O01000O10O10O1000O010000O01000O01VNgLmBY3c<YMZCg2a<aM]C_2b<dM[C]2d<fMZCY2f<kMVCV2i<mMUCS2j<PNSCP2m<SNQCm1m<XNoBi1P=ZNnBe1R=^NlBb1S=f2O001O1O001O001O1O0100O10000O100O010O10000O1O100O001O100O100O100O101N100O10001N100O100O2O1N2aL]B?d=^O_Ba0b=[ObBc0`=ZOdBd0^=XOeBf0]=VOgBi0Z=TOhBk0[=QOhBn0Y=nNjBQ1X=kNlBT1V=gNnBW1T=eNRCX1P=bNWCZ1P=]NUC_1Y=RNkBl1a=fMcBX2d?M3N2N2M3J6Aj0UO[Qb4"}, "label": "a woman holds a fork as she eats a pizza"}]}
{"id": 52179, "image": "COCO_train2014_000000052179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling woman about to eat pizza and salad\"."}], "task": "refering", "answer_template": "The \"a smiling woman about to eat pizza and salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 64, 65, 66, 78, 79, 80, 81, 94, 95, 96, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 136, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 181], "bbox": [0.06077283372365339, 0.141890625, 0.4523419203747072, 0.5270312500000001], "iscrowd": 0, "area": 24980.965099999998, "rle": {"size": [640, 427], "counts": "ka`07\\c0>C<C>F:M3M2N3N2M3M3M2N3N2M3M3M2O2M3M3M2O1N2M3L4L3M4L4L4N2000O010000O100O01000O100O10000O100O01000O100O10000O100O10O1000O100000000O010000O01000O10O10O1000O010000O01000O01VNgLmBY3c<YMZCg2a<aM]C_2b<dM[C]2d<fMZCY2f<kMVCV2i<mMUCS2j<PNSCP2m<SNQCm1m<XNoBi1P=ZNnBe1R=^NlBb1S=f2O001O1O001O001O1O0100O10000O100O010O10000O1O100O001O100O100O100O101N100O10001N100O100O2O1N2aL]B?d=^O_Ba0b=[ObBc0`=ZOdBd0^=XOeBf0]=VOgBi0Z=TOhBk0[=QOhBn0Y=nNjBQ1X=kNlBT1V=gNnBW1T=eNRCX1P=bNWCZ1P=]NUC_1Y=RNkBl1a=fMcBX2d?M3N2N2M3J6Aj0UO[Qb4"}, "label": "a smiling woman about to eat pizza and salad"}]}
{"id": 52179, "image": "COCO_train2014_000000052179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a calzone\"."}], "task": "refering", "answer_template": "The \"a calzone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 294, 295, 296, 297, 310, 311], "bbox": [0.3658782201405152, 0.6284375, 0.8747540983606557, 0.88790625], "iscrowd": 0, "area": 18470.107, "rle": {"size": [640, 427], "counts": "nlQ31nc02O1N2N2N2O1N2N2N2O2N1N101O1O1N101O1O1N101O1O1N101O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O0O2O1O1O1O001O1O2N1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O2N1O1O2N1O1O101N1O1O2N1O1O2N100O2N1O2N1O2N2O0O2N2N2N1O2N2N2O0O2N2N2N1O2N2N100O1O1O010000000O01O1O1O1O1N1O2O1N2N2O0O2N2O1N2N101N2N2O0O2N2O1N1O2O1N2N101N2N2O0O2N2O1N1O2O0O2N2O0O2N2O0O2N101N2N101N2N1O2O0O2N2O0O2N3L5I8G8I7I8G8I7I8G8IkdP1"}, "label": "a calzone"}]}
{"id": 52179, "image": "COCO_train2014_000000052179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tasty calzone\"."}], "task": "refering", "answer_template": "The \"a tasty calzone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 294, 295, 296, 297, 310, 311], "bbox": [0.3658782201405152, 0.6284375, 0.8747540983606557, 0.88790625], "iscrowd": 0, "area": 18470.107, "rle": {"size": [640, 427], "counts": "nlQ31nc02O1N2N2N2O1N2N2N2O2N1N101O1O1N101O1O1N101O1O1N101O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O0O2O1O1O1O001O1O2N1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O2N1O1O2N1O1O101N1O1O2N1O1O2N100O2N1O2N1O2N2O0O2N2N2N1O2N2N2O0O2N2N2N1O2N2N100O1O1O010000000O01O1O1O1O1N1O2O1N2N2O0O2N2O1N2N101N2N2O0O2N2O1N1O2O1N2N101N2N2O0O2N2O1N1O2O0O2N2O0O2N2O0O2N101N2N101N2N1O2O0O2N2O0O2N3L5I8G8I7I8G8I7I8G8IkdP1"}, "label": "a tasty calzone"}]}
{"id": 478164, "image": "COCO_train2014_000000478164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bent over giraffe\"."}], "task": "refering", "answer_template": "The \"the bent over giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 171, 173, 186, 187, 188, 190, 191, 192, 193, 194, 195, 196, 209, 210, 213, 214, 215, 216, 217, 218, 232, 233, 236, 237, 239, 240, 241, 259, 260, 262, 263, 264, 265, 282, 283, 285, 286, 306, 308, 309, 310, 331, 332], "bbox": [0.098484375, 0.3910117647058824, 0.5715468749999999, 0.950564705882353], "iscrowd": 0, "area": 25039.61295000001, "rle": {"size": [425, 640], "counts": "i\\j02U=4M3L3M4M3_Oa0H8N2M3N2N2M3N3M3L301N2N00M4M2M3M2NO100O2O0O101N100O2O0O101N12N1O101N1O2N1O1O2N100O2N1O2O00001O00001O001O00001O000010O01O00001O000010O10O01000O10000O01000O1000O01000000O10O10O10000O010001O00000O2O0^DhNS;X1hDQOT;P1gDYOT;c1J6K5K6J5K5K5J8I6J6J6J6I8I6J6J9G:fHcJT6f5[I`Jf6Q61N2O1N2O0O2O1N2`MQIeNP7Z1`IUNb6j1nIfMR6[2]JTMe5k2kJdLW5[3\\2009F:G9G9H8G9VIRKh4V5bJWK_5Q5lI[KT6m51O001O001O0_LdI5]6JkINV61RJGn59ZJ@f5?bJYO_5e0jJSOW5l0`L]Ma3b2d3O1O1O1O1O2N1N2O1O1O1O1O1O1O1N3N0011N3M3N1N3N1N3M3N1N3N2M2N3N1N3N2M2iHbL_4a3\\KaLe4`3VKcLj4`3PKdLP5^3kJdLU5^3fJeLZ5l4RIWKo6a5001N100O2O0O2O000O2kLeH:[7KkHIW7;nH[OR7j0TIjNm6[1_24K4M3L4L4M3L4M3L4L4M32N2N2N1PNSG^Oi8:cGFZ81RHOk7GaH:[7ClH=Q7CSI<i6D\\I;a6EcI7\\6JiIOY61kIJV67mICT6`0oIXOT6j0nIoNT6T1nIeNU6\\1oI\\NS63[GZ1\\;eNPER1m:nN^Ej0`:UOjEd0T:[OnEf0o9ZOSFf0l9YOVFh0g9XO[Fi0c9VO_Fk0^9UOdFl0X9UOjFk0S9VOoFj0m8XOUGf0j8[OXGc0g8^O[Ga0c8@_G>`8CbG;]8FeG8Z8HiG6V8KlG3S8NoG1o70RHOm72UHLj75WHJh77YHHf78]HGa79aHF^7:dHE[7;hHCW7=kHBT7?mH@R7`0QI_Om6a0UI^Oj6b0XI]Og6c0o20001O1O010O1O1O001O1O1O010O1O001O1O1O001M3K\\`a3"}, "label": "the bent over giraffe"}]}
{"id": 478164, "image": "COCO_train2014_000000478164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lighter colored giraffe whose leaning its head into branches\"."}], "task": "refering", "answer_template": "The \"a lighter colored giraffe whose leaning its head into branches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 171, 173, 186, 187, 188, 190, 191, 192, 193, 194, 195, 196, 209, 210, 213, 214, 215, 216, 217, 218, 232, 233, 236, 237, 239, 240, 241, 259, 260, 262, 263, 264, 265, 282, 283, 285, 286, 306, 308, 309, 310, 331, 332], "bbox": [0.098484375, 0.3910117647058824, 0.5715468749999999, 0.950564705882353], "iscrowd": 0, "area": 25039.61295000001, "rle": {"size": [425, 640], "counts": "i\\j02U=4M3L3M4M3_Oa0H8N2M3N2N2M3N3M3L301N2N00M4M2M3M2NO100O2O0O101N100O2O0O101N12N1O101N1O2N1O1O2N100O2N1O2O00001O00001O001O00001O000010O01O00001O000010O10O01000O10000O01000O1000O01000000O10O10O10000O010001O00000O2O0^DhNS;X1hDQOT;P1gDYOT;c1J6K5K6J5K5K5J8I6J6J6J6I8I6J6J9G:fHcJT6f5[I`Jf6Q61N2O1N2O0O2O1N2`MQIeNP7Z1`IUNb6j1nIfMR6[2]JTMe5k2kJdLW5[3\\2009F:G9G9H8G9VIRKh4V5bJWK_5Q5lI[KT6m51O001O001O0_LdI5]6JkINV61RJGn59ZJ@f5?bJYO_5e0jJSOW5l0`L]Ma3b2d3O1O1O1O1O2N1N2O1O1O1O1O1O1O1N3N0011N3M3N1N3N1N3M3N1N3N2M2N3N1N3N2M2iHbL_4a3\\KaLe4`3VKcLj4`3PKdLP5^3kJdLU5^3fJeLZ5l4RIWKo6a5001N100O2O0O2O000O2kLeH:[7KkHIW7;nH[OR7j0TIjNm6[1_24K4M3L4L4M3L4M3L4L4M32N2N2N1PNSG^Oi8:cGFZ81RHOk7GaH:[7ClH=Q7CSI<i6D\\I;a6EcI7\\6JiIOY61kIJV67mICT6`0oIXOT6j0nIoNT6T1nIeNU6\\1oI\\NS63[GZ1\\;eNPER1m:nN^Ej0`:UOjEd0T:[OnEf0o9ZOSFf0l9YOVFh0g9XO[Fi0c9VO_Fk0^9UOdFl0X9UOjFk0S9VOoFj0m8XOUGf0j8[OXGc0g8^O[Ga0c8@_G>`8CbG;]8FeG8Z8HiG6V8KlG3S8NoG1o70RHOm72UHLj75WHJh77YHHf78]HGa79aHF^7:dHE[7;hHCW7=kHBT7?mH@R7`0QI_Om6a0UI^Oj6b0XI]Og6c0o20001O1O010O1O1O001O1O1O010O1O001O1O1O001M3K\\`a3"}, "label": "a lighter colored giraffe whose leaning its head into branches"}]}
{"id": 478164, "image": "COCO_train2014_000000478164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking to the right\"."}], "task": "refering", "answer_template": "The \"a giraffe looking to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 105, 106, 107, 108, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 219, 220, 242, 243, 265, 288], "bbox": [0.44917187500000005, 0.1932470588235294, 0.7237500000000001, 0.8359529411764706], "iscrowd": 0, "area": 10200.417149999994, "rle": {"size": [425, 640], "counts": "fTh31V=4L4M3L4L4L4Lh0XO2O1N2O2O000000001O00000000001O0O1000001O0000000000O1O100O1O1O1O100O001O10oF`NX6`1gIcNV6^1jIeNR6\\1mIgNQ6Y1nIjNo5X1PJkNm5U1RJnNk5S1TJPOi5Q1WJQOg5P1WJROg5o0YJoNi5Q1VJoNj5R1UJnNk5S1UJkNm5V1]HTNm0f0f6W1]HTNk0d0j6X1[HVNi0a0[OHZ6b1bIVNg0?PO8g6T1aIWNe0h1i51bIXNd0h1j5ObIZN`0j1o5KaI]NK\\2d6WO`Ie1`6\\N^Id1b6]N^I:BkNP7l0\\I81`Nd6X1ZI8=VNY6c1XI7j0lMn5n1VI5X8KgG5Y8MeG2]8NbG2_8O_GlNf0WOm7n1[GjNk9V1TFiNm9X22N2M3N2N001O1O1O001]NhE5Y:KSFGo9:]FXOc9i0iFjNX9W1\\10O2O001O0O2OO100O010O1O10O0100O1O010O100O00100O10O01O100O100O100O1O100O100O1O100O10O01O100O100O110100O100O00100O1O100O1O10O01O100OXOSD5l;E[D;e;]OcDc0R<O1O1N2O1N2N2N2O1N2N2N2OoZY2"}, "label": "a giraffe looking to the right"}]}
{"id": 478164, "image": "COCO_train2014_000000478164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe walking away and looking off into the distance\"."}], "task": "refering", "answer_template": "The \"a giraffe walking away and looking off into the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 105, 106, 107, 108, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 219, 220, 242, 243, 265, 288], "bbox": [0.44917187500000005, 0.1932470588235294, 0.7237500000000001, 0.8359529411764706], "iscrowd": 0, "area": 10200.417149999994, "rle": {"size": [425, 640], "counts": "fTh31V=4L4M3L4L4L4Lh0XO2O1N2O2O000000001O00000000001O0O1000001O0000000000O1O100O1O1O1O100O001O10oF`NX6`1gIcNV6^1jIeNR6\\1mIgNQ6Y1nIjNo5X1PJkNm5U1RJnNk5S1TJPOi5Q1WJQOg5P1WJROg5o0YJoNi5Q1VJoNj5R1UJnNk5S1UJkNm5V1]HTNm0f0f6W1]HTNk0d0j6X1[HVNi0a0[OHZ6b1bIVNg0?PO8g6T1aIWNe0h1i51bIXNd0h1j5ObIZN`0j1o5KaI]NK\\2d6WO`Ie1`6\\N^Id1b6]N^I:BkNP7l0\\I81`Nd6X1ZI8=VNY6c1XI7j0lMn5n1VI5X8KgG5Y8MeG2]8NbG2_8O_GlNf0WOm7n1[GjNk9V1TFiNm9X22N2M3N2N001O1O1O001]NhE5Y:KSFGo9:]FXOc9i0iFjNX9W1\\10O2O001O0O2OO100O010O1O10O0100O1O010O100O00100O10O01O100O100O100O1O100O100O1O100O10O01O100O100O110100O100O00100O1O100O1O10O01O100OXOSD5l;E[D;e;]OcDc0R<O1O1N2O1N2N2N2O1N2N2N2OoZY2"}, "label": "a giraffe walking away and looking off into the distance"}]}
{"id": 265173, "image": "COCO_train2014_000000265173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing the glasses\"."}], "task": "refering", "answer_template": "The \"the woman wearing the glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.32534375, 0.48021077283372365, 0.581734375, 0.9858313817330212], "iscrowd": 0, "area": 23046.56184999999, "rle": {"size": [427, 640], "counts": "_Sg23Q=8H9G9G8J7L4K5K4L4L5L3L4L4L4L4L4M3L4L4L5K5K4L5K4L3N2O1O100O1O1O1O100O1O1O1TN]LoId3o5dLjI]3T6iLhIW3V6PMdIQ3Z6UM_In2a6VMXIm2g6l100O100O100O100O100O100O100O101N100000000000000000000000000000000000000001O00001O001O00001O00001O2N3M2N2N2N3M2N2N3]JoHf4m7N1O1O2N1O1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O2N1O1O1O1O5K6J5K6J5K6J5K6J5K6J5K6J4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3KX[_3"}, "label": "the woman wearing the glasses"}]}
{"id": 265173, "image": "COCO_train2014_000000265173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing black who gaze towards a display\"."}], "task": "refering", "answer_template": "The \"a woman wearing black who gaze towards a display\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.32534375, 0.48021077283372365, 0.581734375, 0.9858313817330212], "iscrowd": 0, "area": 23046.56184999999, "rle": {"size": [427, 640], "counts": "_Sg23Q=8H9G9G8J7L4K5K4L4L5L3L4L4L4L4L4M3L4L4L5K5K4L5K4L3N2O1O100O1O1O1O100O1O1O1TN]LoId3o5dLjI]3T6iLhIW3V6PMdIQ3Z6UM_In2a6VMXIm2g6l100O100O100O100O100O100O100O101N100000000000000000000000000000000000000001O00001O001O00001O00001O2N3M2N2N2N3M2N2N3]JoHf4m7N1O1O2N1O1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O2N1O1O1O1O5K6J5K6J5K6J5K6J5K6J5K6J4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3KX[_3"}, "label": "a woman wearing black who gaze towards a display"}]}
{"id": 265173, "image": "COCO_train2014_000000265173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car parked at the side of the road right next to parking sign\"."}], "task": "refering", "answer_template": "The \"black car parked at the side of the road right next to parking sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.834796875, 0.6714051522248243, 1.0, 1.0], "iscrowd": 0, "area": 12984.425900000006, "rle": {"size": [427, 640], "counts": "[Ro6b0e<4B>N2N2N2L4O1O1eNVOUFj0k9AiE`0W:HaE8]:\\1D<N2O1O1O1O1O100O1O1O1O1N2O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O100O10000O1000000O10000O1000000O10000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000"}, "label": "black car parked at the side of the road right next to parking sign"}]}
{"id": 142299, "image": "COCO_train2014_000000142299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black horse with white stripe on forehead pulling wagon\"."}], "task": "refering", "answer_template": "The \"black horse with white stripe on forehead pulling wagon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 211, 213, 214, 215, 216, 234, 237, 238, 239, 257, 260, 261, 262, 283, 284, 306, 307, 330], "bbox": [0.18812500000000001, 0.20817330210772833, 0.542, 0.9452927400468384], "iscrowd": 0, "area": 32239.873349999998, "rle": {"size": [427, 640], "counts": "`Vb11X=9H8G8dGD[3b0VLIi3;dJmNmNQ1]68XJWOjNl0l62\\IoNVOc0Oi0^7OlHm1R7W2N101N1O10YNnHWMQ7f2YISMg6i2dIQMZ6l2_J_La5]3RKULn4g3\\2K6J5E<E;M23N1N3N1O1N3N1N2O1N200O1O1O100O1O1O010O1O010O1O00100O001O10O01O1O01O000010O000001O01O=C5K8H4L3M3M2N3M3M2N3M2M4M3M2N3L9mI\\Jo3l5fKXJY4P6\\KUJb4S6QKSJm4[6aJlI]5R7O2N1N3N1O001N2O1O1O1NaNQKdJo4[5VKbJi4^5ZK_Jf4a5]K]Ja4d5aKYJ`4g5bKWJ^4i5eKTJ[4l5gKQJY4o5jKoIV4Q6lKmIT4S6nKjIS4U6PLiIo3X6TLeIl3Z6WLcIh3_6ZL_Ie3a6^L]Ia3c6bLZI]3h6b1000O2O0000010O01O00010O00001O00001N100O2O1N2D;G:E;D<K5K4M4[Oe0L4M3M2N3N2M3N1N3N1UOYFnM^9\\OlFd21bM[9KaF`2W:bMeE_2\\:bMaE`2_:600000000001O1O2O01O001O001O1N101N2N2O0O2N2O0O1O100O1M3N2M3N2M3N2M3N2M3M3O100O100000O01c0\\O3M4L4L2N2N2O0N3N2M3N2M3nMZE[1g:ZNkE]1U;M4K4M4K5L4K5K:F[oi3"}, "label": "black horse with white stripe on forehead pulling wagon"}]}
{"id": 142299, "image": "COCO_train2014_000000142299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"biggest black horse\"."}], "task": "refering", "answer_template": "The \"biggest black horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 211, 213, 214, 215, 216, 234, 237, 238, 239, 257, 260, 261, 262, 283, 284, 306, 307, 330], "bbox": [0.18812500000000001, 0.20817330210772833, 0.542, 0.9452927400468384], "iscrowd": 0, "area": 32239.873349999998, "rle": {"size": [427, 640], "counts": "`Vb11X=9H8G8dGD[3b0VLIi3;dJmNmNQ1]68XJWOjNl0l62\\IoNVOc0Oi0^7OlHm1R7W2N101N1O10YNnHWMQ7f2YISMg6i2dIQMZ6l2_J_La5]3RKULn4g3\\2K6J5E<E;M23N1N3N1O1N3N1N2O1N200O1O1O100O1O1O010O1O010O1O00100O001O10O01O1O01O000010O000001O01O=C5K8H4L3M3M2N3M3M2N3M2M4M3M2N3L9mI\\Jo3l5fKXJY4P6\\KUJb4S6QKSJm4[6aJlI]5R7O2N1N3N1O001N2O1O1O1NaNQKdJo4[5VKbJi4^5ZK_Jf4a5]K]Ja4d5aKYJ`4g5bKWJ^4i5eKTJ[4l5gKQJY4o5jKoIV4Q6lKmIT4S6nKjIS4U6PLiIo3X6TLeIl3Z6WLcIh3_6ZL_Ie3a6^L]Ia3c6bLZI]3h6b1000O2O0000010O01O00010O00001O00001N100O2O1N2D;G:E;D<K5K4M4[Oe0L4M3M2N3N2M3N1N3N1UOYFnM^9\\OlFd21bM[9KaF`2W:bMeE_2\\:bMaE`2_:600000000001O1O2O01O001O001O1N101N2N2O0O2N2O0O1O100O1M3N2M3N2M3N2M3N2M3M3O100O100000O01c0\\O3M4L4L2N2N2O0N3N2M3N2M3nMZE[1g:ZNkE]1U;M4K4M4K5L4K5K:F[oi3"}, "label": "biggest black horse"}]}
{"id": 52192, "image": "COCO_train2014_000000052192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with trench coat and red boats walking in the rain\"."}], "task": "refering", "answer_template": "The \"woman with trench coat and red boats walking in the rain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 81, 82, 83, 95, 96, 97, 98, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 139, 141, 142, 143, 144, 156, 157, 158, 159, 170, 171, 172, 173, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 246, 247, 261, 262, 276, 277, 291, 292, 293], "bbox": [0.2977830188679245, 0.20954687500000002, 0.6566509433962264, 0.87028125], "iscrowd": 0, "area": 30996.069500000005, "rle": {"size": [640, 424], "counts": "jo^2`0]c02O1O1N101O1N2O1N2O1O1N101O1N2O1O1O1O0O2O1O2N1P@k0U;WOiDk0U;VOkDl0S;UOlDl0S;TOmDn0R;ROlDP1T;oNlDS1S;mNlDT1S;lNmDV1R;jNlDX1T;gNlD[1R;fNmD[1S;dNmD^1R;bNmD`1R;_NmDc1R;^NmDd1R;[NnDf1dNTM\\;V1oEh1ZN`Mb;h0SFj1QNlMi;9UFY3g9fLYF_3c9aL\\Fd3a9ZL^Fl3^9TLaFQ4[9nKeFX4V9hKiF]4T9cKjFb4Q9_KmFg4m8\\KQGi4aNoIi9[1dGk4]NoIj9X1hGZ5BXIV7a1UI\\5jNkIo7k0VId6h6_IVIb6h6`IWIb6Q6VJlIl5R6VJmIk5Q6XJmIi5R6YJaIT6\\6nIVI`6h6cIjHj6U7Q3N2N3N100O10001O000O101O00000O101O0000001O000000nNZIiDg6S;^IlDb6P;cInD^6n:gIQEZ6j:kIUEV6f:nIYEU6a:PJ^EQ6]:TJaEn5Z:WJeEj5l9eJRF]5^9SKaFn4n8cKQG^4=jH[7P3VHX4;mHW7R3]HR48QIS7T3cHm36TIo6V3jHg33XIn6U3nHd30\\IP7Q3nHe3MaIR7k2PIk4o6VKoHl4P7UKoHk4Q7UKoHk4Q7VKmHk4S7R4O00001cK\\C\\1d<bNcCU1a<lNbCh0g<WO^C=i<D[COm<1WCCR==RCVOV=j0nBjNZ=W1jB]N]=d1`2O01O0101N3N2M3N2N2M3N2M3N1N3N2L4L4L4K5L4L4L4L4L4L4L4Ljoj2"}, "label": "woman with trench coat and red boats walking in the rain"}]}
{"id": 52192, "image": "COCO_train2014_000000052192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with red rain boots\"."}], "task": "refering", "answer_template": "The \"woman with red rain boots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 81, 82, 83, 95, 96, 97, 98, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 139, 141, 142, 143, 144, 156, 157, 158, 159, 170, 171, 172, 173, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 246, 247, 261, 262, 276, 277, 291, 292, 293], "bbox": [0.2977830188679245, 0.20954687500000002, 0.6566509433962264, 0.87028125], "iscrowd": 0, "area": 30996.069500000005, "rle": {"size": [640, 424], "counts": "jo^2`0]c02O1O1N101O1N2O1N2O1O1N101O1N2O1O1O1O0O2O1O2N1P@k0U;WOiDk0U;VOkDl0S;UOlDl0S;TOmDn0R;ROlDP1T;oNlDS1S;mNlDT1S;lNmDV1R;jNlDX1T;gNlD[1R;fNmD[1S;dNmD^1R;bNmD`1R;_NmDc1R;^NmDd1R;[NnDf1dNTM\\;V1oEh1ZN`Mb;h0SFj1QNlMi;9UFY3g9fLYF_3c9aL\\Fd3a9ZL^Fl3^9TLaFQ4[9nKeFX4V9hKiF]4T9cKjFb4Q9_KmFg4m8\\KQGi4aNoIi9[1dGk4]NoIj9X1hGZ5BXIV7a1UI\\5jNkIo7k0VId6h6_IVIb6h6`IWIb6Q6VJlIl5R6VJmIk5Q6XJmIi5R6YJaIT6\\6nIVI`6h6cIjHj6U7Q3N2N3N100O10001O000O101O00000O101O0000001O000000nNZIiDg6S;^IlDb6P;cInD^6n:gIQEZ6j:kIUEV6f:nIYEU6a:PJ^EQ6]:TJaEn5Z:WJeEj5l9eJRF]5^9SKaFn4n8cKQG^4=jH[7P3VHX4;mHW7R3]HR48QIS7T3cHm36TIo6V3jHg33XIn6U3nHd30\\IP7Q3nHe3MaIR7k2PIk4o6VKoHl4P7UKoHk4Q7UKoHk4Q7VKmHk4S7R4O00001cK\\C\\1d<bNcCU1a<lNbCh0g<WO^C=i<D[COm<1WCCR==RCVOV=j0nBjNZ=W1jB]N]=d1`2O01O0101N3N2M3N2N2M3N2M3N1N3N2L4L4L4K5L4L4L4L4L4L4L4Ljoj2"}, "label": "woman with red rain boots"}]}
{"id": 52192, "image": "COCO_train2014_000000052192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella the lady with red shoes holding\"."}], "task": "refering", "answer_template": "The \"umbrella the lady with red shoes holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 78, 79, 80, 81, 83, 84, 93, 94, 95, 96, 98, 99, 100, 109, 110, 114, 115], "bbox": [0.2340566037735849, 0.17978125, 0.69875, 0.33482812500000003], "iscrowd": 0, "area": 9467.057349999995, "rle": {"size": [640, 424], "counts": "jQn11lc04L3N2M4L3M4L3M4M2M4L3M4L3M3N3L3M4L3O2O001O01O01O001O0O2O0O3N1N2O1N2O1N3N1NO2N2N1O21N4M3L4M2MO2N2N2N1O2N2N2N2N1O2N2O1000O10O100000O1000O1000O1000O1000O100000O1000O01L4L4L4L3D=O1O1O1O001N2O1O1O1O1O1O1O1O100O1O1O00100O100O100O100O2O1O1O10O01O100O10O100000O0101O00001NS1nNO1000O01000000O01000O1000O010O10O0100O010O10001N101N2O0O2O0O2O1NO2O0O2O001N101N2O0O2O1N101N2O001N101N2O0O3N3L5L3L5L3L5L3M3L5Lli_2"}, "label": "umbrella the lady with red shoes holding"}]}
{"id": 52192, "image": "COCO_train2014_000000052192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black umbrella the woman in red boots in holding\"."}], "task": "refering", "answer_template": "The \"the black umbrella the woman in red boots in holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 78, 79, 80, 81, 83, 84, 93, 94, 95, 96, 98, 99, 100, 109, 110, 114, 115], "bbox": [0.2340566037735849, 0.17978125, 0.69875, 0.33482812500000003], "iscrowd": 0, "area": 9467.057349999995, "rle": {"size": [640, 424], "counts": "jQn11lc04L3N2M4L3M4L3M4M2M4L3M4L3M3N3L3M4L3O2O001O01O01O001O0O2O0O3N1N2O1N2O1N3N1NO2N2N1O21N4M3L4M2MO2N2N2N1O2N2N2N2N1O2N2O1000O10O100000O1000O1000O1000O1000O100000O1000O01L4L4L4L3D=O1O1O1O001N2O1O1O1O1O1O1O1O100O1O1O00100O100O100O100O2O1O1O10O01O100O10O100000O0101O00001NS1nNO1000O01000000O01000O1000O010O10O0100O010O10001N101N2O0O2O0O2O1NO2O0O2O001N101N2O0O2O1N101N2O001N101N2O0O3N3L5L3L5L3L5L3M3L5Lli_2"}, "label": "the black umbrella the woman in red boots in holding"}]}
{"id": 461799, "image": "COCO_train2014_000000461799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black colored car in front of another black car\"."}], "task": "refering", "answer_template": "The \"black colored car in front of another black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156], "bbox": [0.48343749999999996, 0.5578969957081545, 0.82496875, 0.8203862660944207], "iscrowd": 0, "area": 8166.935400000001, "rle": {"size": [233, 640], "counts": "^fV21X71O1M3K5K5N1O2O00001O001O0O101O001O00001O0O1000000000000O101O000000000O100000001O0O2N2N1O2M3N1O2N2N1O2N1O1N2O1O101O00000000000O2O0000000000N2O1O001O1O1O1000000001N10000000000000000O1000001O000000000O100000000001O01O0001O00001N1000001O00000O2O00001M2J7L30001O001O00001O001O00001O001O0O2O0006K8G01O0O1O2N1O20O00100O00100O1O00100O00O10001O000000001O000000001O0O10001O000000001O00000O1O2N1O1O1O101O0000001O000001O01O001O1O001O1O1O0O5J5L5JRZi0"}, "label": "black colored car in front of another black car"}]}
{"id": 461799, "image": "COCO_train2014_000000461799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car in front of another black car\"."}], "task": "refering", "answer_template": "The \"a black car in front of another black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156], "bbox": [0.48343749999999996, 0.5578969957081545, 0.82496875, 0.8203862660944207], "iscrowd": 0, "area": 8166.935400000001, "rle": {"size": [233, 640], "counts": "^fV21X71O1M3K5K5N1O2O00001O001O0O101O001O00001O0O1000000000000O101O000000000O100000001O0O2N2N1O2M3N1O2N2N1O2N1O1N2O1O101O00000000000O2O0000000000N2O1O001O1O1O1000000001N10000000000000000O1000001O000000000O100000000001O01O0001O00001N1000001O00000O2O00001M2J7L30001O001O00001O001O00001O001O0O2O0006K8G01O0O1O2N1O20O00100O00100O1O00100O00O10001O000000001O000000001O0O10001O000000001O00000O1O2N1O1O1O101O0000001O000001O01O001O1O001O1O1O0O5J5L5JRZi0"}, "label": "a black car in front of another black car"}]}
{"id": 461799, "image": "COCO_train2014_000000461799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"1st car from the left\"."}], "task": "refering", "answer_template": "The \"1st car from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146], "bbox": [0.108453125, 0.5813733905579399, 0.394140625, 0.8348497854077254], "iscrowd": 0, "area": 7744.375449999999, "rle": {"size": [233, 640], "counts": "Qk?4V7U1jN4K1O100O1O100O1O100O1O100O1O100O10000000000000001O0000O1N2N20000000O101O0000000000000000000000000000000000000O100000000000000000000000000000000000000O100001O001O001O001O001O001O001O001O010O001O001CWJXOi5g0>N101N101N1O2O0O2O001O0010O001O00003M2O2M2N3M2N2N0001O0000000000001O00000000001O01O0000000001O00000O1M3M3O2O01O1O1O1OO101O00000000001O0O10001O001O1O001O1N1]O]I`0h6M2N2NhSh2"}, "label": "1st car from the left"}]}
{"id": 461799, "image": "COCO_train2014_000000461799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car stopped in the traffic signal , behind a car\"."}], "task": "refering", "answer_template": "The \"a car stopped in the traffic signal , behind a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146], "bbox": [0.108453125, 0.5813733905579399, 0.394140625, 0.8348497854077254], "iscrowd": 0, "area": 7744.375449999999, "rle": {"size": [233, 640], "counts": "Qk?4V7U1jN4K1O100O1O100O1O100O1O100O1O100O10000000000000001O0000O1N2N20000000O101O0000000000000000000000000000000000000O100000000000000000000000000000000000000O100001O001O001O001O001O001O001O001O010O001O001CWJXOi5g0>N101N101N1O2O0O2O001O0010O001O00003M2O2M2N3M2N2N0001O0000000000001O00000000001O01O0000000001O00000O1M3M3O2O01O1O1O1OO101O00000000001O0O10001O001O1O001O1N1]O]I`0h6M2N2NhSh2"}, "label": "a car stopped in the traffic signal , behind a car"}]}
{"id": 35817, "image": "COCO_train2014_000000035817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty chair directly in front of the cake near the woman standing in the green dress\"."}], "task": "refering", "answer_template": "The \"the empty chair directly in front of the cake near the woman standing in the green dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330], "bbox": [0.125125, 0.6451635514018691, 0.373921875, 0.9871028037383178], "iscrowd": 0, "area": 12732.821549999999, "rle": {"size": [428, 640], "counts": "PjQ11\\=5K4L4L4M3L4L00O2N1O1O101N1O1O2N1O1O001O00010O00001O001O00001OSMN`H3_72\\HNd77XHHh7=SHCm7b0nG^OR8g0jGXOV8m0eGSO[8R1`GnN`8W1\\GhNe8\\1VGdNl8`1oF`NR9c1iF]NW9h1dFXN\\9m1`FRN`9S2[FmMe9X2VFhMj9]2QFcMo9b2mE]MS:h2hEXMX:n201O000000000000000000000000000O10000000000000000O100000000O1000000O10000O01000O10000O1000000O10000O10000O1000000O10000O10000O10O00001O00001O00001O0000001O00001O0000001O00001O03M5K6K4K6J5L1M2J6K500000UNaGiN_8l0WHiNi7l0nHhNR7m0dIhN\\6m0ZJhNf5m0b3E;ES[W5"}, "label": "the empty chair directly in front of the cake near the woman standing in the green dress"}]}
{"id": 35817, "image": "COCO_train2014_000000035817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that the birthday cake is on\"."}], "task": "refering", "answer_template": "The \"the table that the birthday cake is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 145, 146, 147, 163, 164, 165, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 254, 255, 259], "bbox": [0.0, 0.42733644859813086, 0.487375, 0.7639953271028037], "iscrowd": 0, "area": 22052.790999999997, "rle": {"size": [428, 640], "counts": "c8`0k<2O1O001N2O001O1N101O1O0O2O001O1O0O2O1O001N2O001O1N101O1O0O2O1O001N2O001O1N101O1O0O2O1O001N2A^NREb1m:`0N2O0O2N2N1O2O0O0E_MlEb2S:<0O1SMgEg2c:LK401000O\\OZMcFg2[9cM]F]2b9g0O0C`LoFa3P9aLnF`3Q9=1000O001000O01000O010000O010000O10000O101O00O010000O2O0000O010O1000O3N1OO0101O1O0OO2000O1000=C0O10000O2O004Lc0\\O3N00000O10000O10001O0O10000O1000000O102N4K5L1OO10O10000O0100000O14K?B4L00000O100000000O0100VO[EmNe:R1kE_NU:a1j010O10000O0100000O10O10O10000O01000O1000O01000000O0100XO]NgEb1Z:_N[EMOd1f:aNRE28]1f:kNYET1_:bN\\E:4U1_:VO`Ej0`:VO`Ei0]:]N`Ej03i0\\:\\OdEd0\\:@_E`0b:_O_Ea0a:^O`Eb0`:]OaEc0_:]O`Ec0a:]O_Ec0a:^O^Eb0b:S110000O10000O10001N10M300O010000O10001N01N2O02O4L0O1000000O10000O01000O10004K2O1O1N101O1QNiDg1X;WNkDf1b;L1O0O2O1O1N10000O1000O01000000O2O000O`0A8H0O10001N10000O101O0O10001O0O10000O2O000O10001N10001N_kX4"}, "label": "the table that the birthday cake is on"}]}
{"id": 35817, "image": "COCO_train2014_000000035817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a birthday cake with lighted candles and a bouquet of flowers\"."}], "task": "refering", "answer_template": "The \"a birthday cake with lighted candles and a bouquet of flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 145, 146, 147, 163, 164, 165, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 254, 255, 259], "bbox": [0.0, 0.42733644859813086, 0.487375, 0.7639953271028037], "iscrowd": 0, "area": 22052.790999999997, "rle": {"size": [428, 640], "counts": "c8`0k<2O1O001N2O001O1N101O1O0O2O001O1O0O2O1O001N2O001O1N101O1O0O2O1O001N2O001O1N101O1O0O2O1O001N2A^NREb1m:`0N2O0O2N2N1O2O0O0E_MlEb2S:<0O1SMgEg2c:LK401000O\\OZMcFg2[9cM]F]2b9g0O0C`LoFa3P9aLnF`3Q9=1000O001000O01000O010000O010000O10000O101O00O010000O2O0000O010O1000O3N1OO0101O1O0OO2000O1000=C0O10000O2O004Lc0\\O3N00000O10000O10001O0O10000O1000000O102N4K5L1OO10O10000O0100000O14K?B4L00000O100000000O0100VO[EmNe:R1kE_NU:a1j010O10000O0100000O10O10O10000O01000O1000O01000000O0100XO]NgEb1Z:_N[EMOd1f:aNRE28]1f:kNYET1_:bN\\E:4U1_:VO`Ej0`:VO`Ei0]:]N`Ej03i0\\:\\OdEd0\\:@_E`0b:_O_Ea0a:^O`Eb0`:]OaEc0_:]O`Ec0a:]O_Ec0a:^O^Eb0b:S110000O10000O10001N10M300O010000O10001N01N2O02O4L0O1000000O10000O01000O10004K2O1O1N101O1QNiDg1X;WNkDf1b;L1O0O2O1O1N10000O1000O01000000O2O000O`0A8H0O10001N10000O101O0O10001O0O10000O2O000O10001N10001N_kX4"}, "label": "a birthday cake with lighted candles and a bouquet of flowers"}]}
{"id": 35817, "image": "COCO_train2014_000000035817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair with grey padding , on the right side of a table , with a woman wearing a green dress behind it\"."}], "task": "refering", "answer_template": "The \"a wooden chair with grey padding , on the right side of a table , with a woman wearing a green dress behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 217, 218, 219, 220, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.37570312499999997, 0.5685280373831776, 0.580078125, 0.9842757009345794], "iscrowd": 0, "area": 16046.989049999996, "rle": {"size": [428, 640], "counts": "QdT33Y=o0POP1POn0SO1N3N1N2O1N100O100O1O10L3O20O10O010O0100O00100O010O010O10O0100O0010O0100O0L5H8H7I8O010000O1003M3M4L3M3M00000000O1F9F;00O10000O10000O10000O12N2N1Od0\\O1N1000000O2O000OYOeG_L[8b3PHRLo7o3g01O1N2O0O2O1O1N2O001N2O1N2O001N2O1N2O001N2O1NZMfG=Y8BjG<U8CPH;o7DTH;j7D[H9e7F^H8a7GcH7\\7IfH6Z7HiH7V7IkH7U7DPI<o6_O\\I<d6^OgI=X6]OTJ>k5]OaJ=_5]OlJ>S5]OXK>h4[OeK?Z4\\OYL7f3DmLOS3Kb[`3"}, "label": "a wooden chair with grey padding , on the right side of a table , with a woman wearing a green dress behind it"}]}
{"id": 11244, "image": "COCO_train2014_000000011244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bottled partially covered in plastic\"."}], "task": "refering", "answer_template": "The \"a green bottled partially covered in plastic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 115, 116, 132, 133, 149, 150, 165, 166, 167, 168, 182, 183, 184, 185, 199, 200, 201, 202, 216, 217, 218, 219, 233, 234, 235, 236, 250, 251, 252, 253, 267, 268, 269, 270], "bbox": [0.7125581395348838, 0.32022, 0.9230866807610993, 0.8633200000000001], "iscrowd": 0, "area": 17215.943799999997, "rle": {"size": [500, 473], "counts": "hnT58f>b0^Of0ZOf0ZOf0C=K5L4L4L5K4K5L4L4L4L4K5M3M3N2M3M3O1O100O1O1O1O1O1M3M3M3M3M3M4L3M3M3M3M3ZNf1O1O10000O1000000000000000000000000000001O0000000O1000000000TIlHX5T7aJjIh4W6PKTJl4m5gJcJU5^5`JQK[5P5aJWK[5e7L4L4L2N2N3M2N2N6iLYEU1Q;QNfEh1Y<K5J6Jl0TO4L4L3M4L4L3M4L4Llda0"}, "label": "a green bottled partially covered in plastic"}]}
{"id": 11244, "image": "COCO_train2014_000000011244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a front green bottle\"."}], "task": "refering", "answer_template": "The \"a front green bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 115, 116, 132, 133, 149, 150, 165, 166, 167, 168, 182, 183, 184, 185, 199, 200, 201, 202, 216, 217, 218, 219, 233, 234, 235, 236, 250, 251, 252, 253, 267, 268, 269, 270], "bbox": [0.7125581395348838, 0.32022, 0.9230866807610993, 0.8633200000000001], "iscrowd": 0, "area": 17215.943799999997, "rle": {"size": [500, 473], "counts": "hnT58f>b0^Of0ZOf0ZOf0C=K5L4L4L5K4K5L4L4L4L4K5M3M3N2M3M3O1O100O1O1O1O1O1M3M3M3M3M3M4L3M3M3M3M3ZNf1O1O10000O1000000000000000000000000000001O0000000O1000000000TIlHX5T7aJjIh4W6PKTJl4m5gJcJU5^5`JQK[5P5aJWK[5e7L4L4L2N2N3M2N2N6iLYEU1Q;QNfEh1Y<K5J6Jl0TO4L4L3M4L4L3M4L4Llda0"}, "label": "a front green bottle"}]}
{"id": 11244, "image": "COCO_train2014_000000011244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bottle in the middle of 3\"."}], "task": "refering", "answer_template": "The \"a green bottle in the middle of 3\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 114, 115, 131, 132, 148, 149, 165, 182, 199, 216], "bbox": [0.7132346723044398, 0.31454000000000004, 0.8564482029598309, 0.78868], "iscrowd": 0, "area": 2847.2798000000007, "rle": {"size": [500, 473], "counts": "P]U57m>6J6XNg13N2N2M3O21N21O2N1O1O1O20O3M4K4K6I6H[M]Cl2d<1000001O0001O1O1M3L4L4K5L4L4L4K5C=POP1F:0000O1000001O000000000O10000000000000O1000O10000000cQQ1"}, "label": "a green bottle in the middle of 3"}]}
{"id": 429040, "image": "COCO_train2014_000000429040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand next to the remote\"."}], "task": "refering", "answer_template": "The \"the hand next to the remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 33, 34, 35, 51, 52, 53, 69, 70, 71, 87, 88], "bbox": [0.824, 0.00448, 1.0, 0.3296], "iscrowd": 0, "area": 7948.16885, "rle": {"size": [375, 500], "counts": "Wlf44c;7I7I8H7I7I4L2N2N2N2N2N=C4L3M3M3M3M4L5K6J6J6J5K6JO100000000000000O1000000000000O10O100000000000O100O1H8I7I7I7I700O1000000O10000O10000O10000O100002N9G9G8H8H1O000000001O0`0@P1kNV1gNeM"}, "label": "the hand next to the remote"}]}
{"id": 429040, "image": "COCO_train2014_000000429040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hand closest to the remote\"."}], "task": "refering", "answer_template": "The \"hand closest to the remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 33, 34, 35, 51, 52, 53, 69, 70, 71, 87, 88], "bbox": [0.824, 0.00448, 1.0, 0.3296], "iscrowd": 0, "area": 7948.16885, "rle": {"size": [375, 500], "counts": "Wlf44c;7I7I8H7I7I4L2N2N2N2N2N=C4L3M3M3M3M4L5K6J6J6J5K6JO100000000000000O1000000000000O10O100000000000O100O1H8I7I7I7I700O1000000O10000O10000O10000O100002N9G9G8H8H1O000000001O0`0@P1kNV1gNeM"}, "label": "hand closest to the remote"}]}
{"id": 224241, "image": "COCO_train2014_000000224241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rollerslate of a boy wearing a grey and black t - shirt\"."}], "task": "refering", "answer_template": "The \"the rollerslate of a boy wearing a grey and black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [295, 315, 316, 336, 337, 338, 339, 358, 359], "bbox": [0.557921875, 0.69408203125, 0.8747812500000001, 0.8883984375], "iscrowd": 0, "area": 1904.3399500000003, "rle": {"size": [512, 640], "counts": "`mb51n?2N2N2N2O001O1O000O10000000000O1000000000000O10000000000O10000000000000000000001O2N1O2N1O1O2N1O2N1O2N00O1000000O10000J6_Oa0N11000O010O01000O010O10O10O10O010O10O10O010EMi@3V?Oh@1Y?0f@0Z?2b@0]?910O01000O010O010O10O10O010O10O10O010O10O010O10O10O010O10O10O010O0O2N2N1O2N1O2N2N1O`_g01``XO1O1O10O0100O100O100O100O100O100O100O01O001O1N3N2NkTX1"}, "label": "the rollerslate of a boy wearing a grey and black t - shirt"}]}
{"id": 224241, "image": "COCO_train2014_000000224241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard with a man in black and red shoes\"."}], "task": "refering", "answer_template": "The \"a skateboard with a man in black and red shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [295, 315, 316, 336, 337, 338, 339, 358, 359], "bbox": [0.557921875, 0.69408203125, 0.8747812500000001, 0.8883984375], "iscrowd": 0, "area": 1904.3399500000003, "rle": {"size": [512, 640], "counts": "`mb51n?2N2N2N2O001O1O000O10000000000O1000000000000O10000000000O10000000000000000000001O2N1O2N1O1O2N1O2N1O2N00O1000000O10000J6_Oa0N11000O010O01000O010O10O10O10O010O10O10O010EMi@3V?Oh@1Y?0f@0Z?2b@0]?910O01000O010O010O10O10O010O10O10O010O10O010O10O10O010O10O10O010O0O2N2N1O2N1O2N2N1O`_g01``XO1O1O10O0100O100O100O100O100O100O100O01O001O1N3N2NkTX1"}, "label": "a skateboard with a man in black and red shoes"}]}
{"id": 412657, "image": "COCO_train2014_000000412657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black flat - screen computer monitor\"."}], "task": "refering", "answer_template": "The \"a black flat - screen computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.58921875, 0.662927400468384, 0.99703125, 0.9842622950819672], "iscrowd": 0, "area": 30112.403999999995, "rle": {"size": [427, 640], "counts": "ecm42S=6J6J6J6J6I7J6J6J6J6J6J6J6I7J6J6K50000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O1000000000O1000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000O10000000O1000000000000O1000\\Od0lMoc0"}, "label": "a black flat - screen computer monitor"}]}
{"id": 412657, "image": "COCO_train2014_000000412657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer screen on a table being watched by a boy\"."}], "task": "refering", "answer_template": "The \"a computer screen on a table being watched by a boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.58921875, 0.662927400468384, 0.99703125, 0.9842622950819672], "iscrowd": 0, "area": 30112.403999999995, "rle": {"size": [427, 640], "counts": "ecm42S=6J6J6J6J6I7J6J6J6J6J6J6J6I7J6J6K50000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O1000000000O1000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000O10000000O1000000000000O1000\\Od0lMoc0"}, "label": "a computer screen on a table being watched by a boy"}]}
{"id": 412657, "image": "COCO_train2014_000000412657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rectangular computer monitor\"."}], "task": "refering", "answer_template": "The \"a rectangular computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.5937187500000001, 0.662552693208431, 1.0, 0.990632318501171], "iscrowd": 0, "area": 28123.14984999999, "rle": {"size": [427, 640], "counts": "ekn44o<8I7H8H8I7H8H8I7H8I7H8H8I700000001O000O10000000000O1000000000000O1000000000000O1000000000000O10000000001O0O1000000000000O1000000000000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O100000O100000O100000O100000O100000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O1000O10000000O1000O10000000O1000O10000000O1000O100000O100000O100000O100000Oi1XN]E"}, "label": "a rectangular computer monitor"}]}
{"id": 412657, "image": "COCO_train2014_000000412657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"thin lcd computer monitor with black casing\"."}], "task": "refering", "answer_template": "The \"thin lcd computer monitor with black casing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.5937187500000001, 0.662552693208431, 1.0, 0.990632318501171], "iscrowd": 0, "area": 28123.14984999999, "rle": {"size": [427, 640], "counts": "ekn44o<8I7H8H8I7H8H8I7H8I7H8H8I700000001O000O10000000000O1000000000000O1000000000000O1000000000000O10000000001O0O1000000000000O1000000000000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O100000O100000O100000O100000O100000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O10000000O1000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O100000O1000O10000000O1000O10000000O1000O10000000O1000O100000O100000O100000O100000Oi1XN]E"}, "label": "thin lcd computer monitor with black casing"}]}
{"id": 109553, "image": "COCO_train2014_000000109553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a salami and mushroom pizza\"."}], "task": "refering", "answer_template": "The \"a salami and mushroom pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337], "bbox": [0.32039062500000004, 0.1595961995249406, 0.8537968750000001, 0.9860095011876485], "iscrowd": 0, "area": 92646.3619, "rle": {"size": [421, 640], "counts": "^`d2c0Q<R1SO?J6J6J5K6J7I6J7I6J6J7J5K6J5K3M3L4M2N3L4M3L4M3L3N3K\\JeHg5Y74M3N2M3N1N3N2M3M3N2M4M2N3N1N3N1O2M2O2M2O1N3N1N3N1N3N0O2O1N101N2N2N101N2N1O2N2O1N1O2N2O0O2N2N2O0O2N2N2N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2O1N2O1N2O0O101N100O100O2O0O100O2O0O100O101N100O101N100O10001N1000001O000O10001O00000O2O0000001O0O1000001O000O101O0000000O2O000000000O100000010O000000000001O0000000001O0001O000000000000001O01O0000000001O0000000O10000O2O0O10000O10001N100O10000O1O101N100O1O101N1O100O2N100O1O2O0O101N1O100O2N100O1O2O0000001O000O101O00001O00000O2O001O1O1N2O0O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1N101N2O1N2N1N3N2N2M3N2N1O2M3N2N2M2O2N2N2N3M2N2N3M2N3L3N2N3M2N3M2N2N3M2M4M2K5L5J5L5K4K5L5J6M4L3M4L3M4L4L3M4L3M4L5K5K6G9E:E<E:EY^V1"}, "label": "a salami and mushroom pizza"}]}
{"id": 109553, "image": "COCO_train2014_000000109553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked pizza with sauce , mushrooms , pepperoni and cheese\"."}], "task": "refering", "answer_template": "The \"an uncooked pizza with sauce , mushrooms , pepperoni and cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337], "bbox": [0.32039062500000004, 0.1595961995249406, 0.8537968750000001, 0.9860095011876485], "iscrowd": 0, "area": 92646.3619, "rle": {"size": [421, 640], "counts": "^`d2c0Q<R1SO?J6J6J5K6J7I6J7I6J6J7J5K6J5K3M3L4M2N3L4M3L4M3L3N3K\\JeHg5Y74M3N2M3N1N3N2M3M3N2M4M2N3N1N3N1O2M2O2M2O1N3N1N3N1N3N0O2O1N101N2N2N101N2N1O2N2O1N1O2N2O0O2N2N2O0O2N2N2N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2O1N2O1N2O0O101N100O100O2O0O100O2O0O100O101N100O101N100O10001N1000001O000O10001O00000O2O0000001O0O1000001O000O101O0000000O2O000000000O100000010O000000000001O0000000001O0001O000000000000001O01O0000000001O0000000O10000O2O0O10000O10001N100O10000O1O101N100O1O101N1O100O2N100O1O2O0O101N1O100O2N100O1O2O0000001O000O101O00001O00000O2O001O1O1N2O0O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1N101N2O1N2N1N3N2N2M3N2N1O2M3N2N2M2O2N2N2N3M2N2N3M2N3L3N2N3M2N3M2N2N3M2M4M2K5L5J5L5K4K5L5J6M4L3M4L3M4L4L3M4L3M4L5K5K6G9E:E<E:EY^V1"}, "label": "an uncooked pizza with sauce , mushrooms , pepperoni and cheese"}]}
{"id": 109553, "image": "COCO_train2014_000000109553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked pizza with no pepperoni\"."}], "task": "refering", "answer_template": "The \"an uncooked pizza with no pepperoni\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232], "bbox": [0.0017031250000000002, 0.0, 0.29046875, 0.6935154394299288], "iscrowd": 0, "area": 46882.8765, "rle": {"size": [421, 640], "counts": "V=Q9T400000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000001O0000000000000000000O1000000000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O3N1N2O0O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N2N2N3L3N2M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M3M4L3M3MT1lNc1]N[`j5"}, "label": "an uncooked pizza with no pepperoni"}]}
{"id": 109553, "image": "COCO_train2014_000000109553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left most pizza\"."}], "task": "refering", "answer_template": "The \"the left most pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232], "bbox": [0.0017031250000000002, 0.0, 0.29046875, 0.6935154394299288], "iscrowd": 0, "area": 46882.8765, "rle": {"size": [421, 640], "counts": "V=Q9T400000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000001O0000000000000000000O1000000000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O3N1N2O0O2N2M3N2N2N1O2N2N2M3N2N1O2N2N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N2N2N3L3N2M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M3M4L3M3MT1lNc1]N[`j5"}, "label": "the left most pizza"}]}
{"id": 470004, "image": "COCO_train2014_000000470004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate of goodies next to the cake\"."}], "task": "refering", "answer_template": "The \"the plate of goodies next to the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334, 335, 336], "bbox": [0.317828125, 0.7897423887587822, 0.700984375, 0.9647540983606558], "iscrowd": 0, "area": 12506.669799999998, "rle": {"size": [427, 640], "counts": "hPe21T=7M3L4L4M3L4M3M3N2O0O2N1O11O01O0000000010O000001O000001O000001O0001O0000000001O0000000O100O1O2N1O1O1O1O100O2O0000001O00000O101O000O101N100O101O000000001O00000010O001O010O001O00001O001O0001O0O100000000O10000000000O2N100O1O100O1O1O1O1O1O1O2N100000000O10000000000O101O0002N1O1O1O1O1O1O1O1O00001O00001O000000000000001O000000000O1000001O0000000000000O1000O1000000O010O1O1O12N3M3L101O00001O0O2O00000O100O100O2O000O2O0O100O100O0100O010MgNSDY1m;3O1O0010O10O11N101O0O2O0O101N1O2N1O2N101M2O2M3M2O2M3N1M4L4L^Y_2"}, "label": "the plate of goodies next to the cake"}]}
{"id": 470004, "image": "COCO_train2014_000000470004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate full of icing cake pieces topped with cherries\"."}], "task": "refering", "answer_template": "The \"a plate full of icing cake pieces topped with cherries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334, 335, 336], "bbox": [0.317828125, 0.7897423887587822, 0.700984375, 0.9647540983606558], "iscrowd": 0, "area": 12506.669799999998, "rle": {"size": [427, 640], "counts": "hPe21T=7M3L4L4M3L4M3M3N2O0O2N1O11O01O0000000010O000001O000001O000001O0001O0000000001O0000000O100O1O2N1O1O1O1O100O2O0000001O00000O101O000O101N100O101O000000001O00000010O001O010O001O00001O001O0001O0O100000000O10000000000O2N100O1O100O1O1O1O1O1O1O2N100000000O10000000000O101O0002N1O1O1O1O1O1O1O1O00001O00001O000000000000001O000000000O1000001O0000000000000O1000O1000000O010O1O1O12N3M3L101O00001O0O2O00000O100O100O2O000O2O0O100O100O0100O010MgNSDY1m;3O1O0010O10O11N101O0O2O0O101N1O2N1O2N101M2O2M3M2O2M3N1M4L4L^Y_2"}, "label": "a plate full of icing cake pieces topped with cherries"}]}
{"id": 470004, "image": "COCO_train2014_000000470004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cake with white frosting and many candles , sitting on a table\"."}], "task": "refering", "answer_template": "The \"a cake with white frosting and many candles , sitting on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306], "bbox": [0.09957812499999999, 0.5359953161592506, 0.41771875, 0.8925526932084309], "iscrowd": 0, "area": 21854.5356, "rle": {"size": [427, 640], "counts": "ZQk06n<7H8H8K5M4K4M3L4M4^Oa0K5O101N10000O2O0O100O2O000O100O2O0O10001N100O101N10000O2O000001OA?F;M2O1N2N2N2O101N100O100O10001N1000000O10001N100O10000O100O1N2O1O1N2O1O1O1N200001O5bKYGn3S9O00O100O1O10000O1000000O1000000O1O010O1O1001O1O001O1O1O1O1O2N00000000O100O100000000001O00O1O1O1O1OM301N101O2M4M1O2M3N2M3M2O1N2O1O100O0000000101N1O1001O100N2O3M4K3N2M2O1O010O001O001O010O001O001N2nM`EW1b:bNfE[1U;01O001O001O001O001O1O0O2O1O1TOcDG_;GTE7m:HVE4k:KXE2`bk4"}, "label": "a cake with white frosting and many candles , sitting on a table"}]}
{"id": 470004, "image": "COCO_train2014_000000470004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a birthday cake with many candles and squirrel on top of it\"."}], "task": "refering", "answer_template": "The \"a birthday cake with many candles and squirrel on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306], "bbox": [0.09957812499999999, 0.5359953161592506, 0.41771875, 0.8925526932084309], "iscrowd": 0, "area": 21854.5356, "rle": {"size": [427, 640], "counts": "ZQk06n<7H8H8K5M4K4M3L4M4^Oa0K5O101N10000O2O0O100O2O000O100O2O0O10001N100O101N10000O2O000001OA?F;M2O1N2N2N2O101N100O100O10001N1000000O10001N100O10000O100O1N2O1O1N2O1O1O1N200001O5bKYGn3S9O00O100O1O10000O1000000O1000000O1O010O1O1001O1O001O1O1O1O1O2N00000000O100O100000000001O00O1O1O1O1OM301N101O2M4M1O2M3N2M3M2O1N2O1O100O0000000101N1O1001O100N2O3M4K3N2M2O1O010O001O001O010O001O001N2nM`EW1b:bNfE[1U;01O001O001O001O001O1O0O2O1O1TOcDG_;GTE7m:HVE4k:KXE2`bk4"}, "label": "a birthday cake with many candles and squirrel on top of it"}]}
{"id": 470002, "image": "COCO_train2014_000000470002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating from a feeder in a tree\"."}], "task": "refering", "answer_template": "The \"a giraffe eating from a feeder in a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 88, 89, 90, 105, 106, 107, 124, 125, 141, 142, 158, 159, 160, 175, 176, 177, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 282, 283, 284, 296, 297, 300, 301, 313, 314, 317, 318, 330, 331, 334, 335, 347, 348, 350, 351, 352, 364, 365, 367, 368, 369], "bbox": [0.18577083333333333, 0.19775, 0.7610416666666667, 0.964046875], "iscrowd": 0, "area": 37921.511450000005, "rle": {"size": [640, 480], "counts": "\\hg11g00`b03W]O4hb00n\\O6Rc0=O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O00001O000K6G81O2O0O2N101N1O2O0O1O2O02L8I7I7H8I7I6IeMk^OS2o`0lMT_OX2f`0gM]_O\\2]`0cMg_Oa2S`0^MP@f2j?YMZ@j2`?UMc@n2W?RMl@R3n>lLVAX3d>gL_A\\3[>dLhA`3R>_LQBe3i=ZL[Be3b=ZLbBf3\\=YLgBg3j>O10000O2OkAcLe;]3PDmLQ<S3dCXM\\<g2_C_Ma<a2ZCdMf<\\2VCgMk<Y2PClMQ=R2lBRNT=n1hBUNY=k1bBZN^=e1_B_Na=a1ZBdNf=\\1VBgNk=Y1QBkNP>c30O1000000O101O0O100101N2N3N1N3M2O1N3N1N2N3N1N3RF[IY6g6bIZIa6g6[IYIg6j6SIWIo6k6lHWIV7l6dHTI]7Q7ZHRIg7R7nGTIS8Q7aGUI`8P7TGVIm8o6hFVIX9P7\\FVIe9o6oEWIR:[81O1O001O1N2aJcEY1^:fNhET1Y:kNnEn0R:SOSFg0n9XOXFb0i9]O^F<c9CcF6_9IhF0X90nFJS95VGBk8>^GXOc8g0gGoNZ8P1PHfNP8Z1ZH\\Ng7c1cHQN`7n1jH\\Mc7c2fHiLf7W3b4O001O1O1O10O01O1O1O1O010O1O1O001O01O00000000001O01O00000001O00000001O0O1L4K5L4L4K5002N2O1N2N2N1YDTMX6n2dIVMZ6l2aIYM]6i2_I[M_6g2]I\\Mb6f2YI]Mh6d2TI_Mk6b2QIaMo6a2mHaMS7a2iHaMW7a2dHcM[7_2aHcM_7_2]HcMc7_2XHeMg7\\2VHfMk7[2QHgMo7[2lGiMS8Y2iGiMW8Y2dGjM\\8X2`GkM_8W2]GkMc8V2YGmMh8T2TGoMl8R2PGiMW9Y2cFcMf9^2TF^MU:c2eEZMb:h2XETMQ;l2jDQM^;P3[DQMi;Q3QDPMS<Q3gCPM]<Q3]CoLg<S3SCnLQ=S3hBoL\\=g43L5L4L3M4K5L3M4K5L3M4L4K4M4N2NhC]KU9b4jFaKV9]4iFeKX9Y4gFjKX9V4fFmKZ9U4aFnK`9R4\\FQLg9n3TFTLo9l3lEWLV:i3eEZL]:f3^E]Ld:b3XE`Lk:`3QEbLR;a3eDbL];f3VD]Ll;k3gCXL[<P4XCRLk<a52N3lKkC3Z<hNQEm0T;fM_Fo1T>D;D<E;E<DdYW2"}, "label": "a giraffe eating from a feeder in a tree"}]}
{"id": 470002, "image": "COCO_train2014_000000470002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating from a tree\"."}], "task": "refering", "answer_template": "The \"a giraffe eating from a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 88, 89, 90, 105, 106, 107, 124, 125, 141, 142, 158, 159, 160, 175, 176, 177, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 282, 283, 284, 296, 297, 300, 301, 313, 314, 317, 318, 330, 331, 334, 335, 347, 348, 350, 351, 352, 364, 365, 367, 368, 369], "bbox": [0.18577083333333333, 0.19775, 0.7610416666666667, 0.964046875], "iscrowd": 0, "area": 37921.511450000005, "rle": {"size": [640, 480], "counts": "\\hg11g00`b03W]O4hb00n\\O6Rc0=O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O00001O000K6G81O2O0O2N101N1O2O0O1O2O02L8I7I7H8I7I6IeMk^OS2o`0lMT_OX2f`0gM]_O\\2]`0cMg_Oa2S`0^MP@f2j?YMZ@j2`?UMc@n2W?RMl@R3n>lLVAX3d>gL_A\\3[>dLhA`3R>_LQBe3i=ZL[Be3b=ZLbBf3\\=YLgBg3j>O10000O2OkAcLe;]3PDmLQ<S3dCXM\\<g2_C_Ma<a2ZCdMf<\\2VCgMk<Y2PClMQ=R2lBRNT=n1hBUNY=k1bBZN^=e1_B_Na=a1ZBdNf=\\1VBgNk=Y1QBkNP>c30O1000000O101O0O100101N2N3N1N3M2O1N3N1N2N3N1N3RF[IY6g6bIZIa6g6[IYIg6j6SIWIo6k6lHWIV7l6dHTI]7Q7ZHRIg7R7nGTIS8Q7aGUI`8P7TGVIm8o6hFVIX9P7\\FVIe9o6oEWIR:[81O1O001O1N2aJcEY1^:fNhET1Y:kNnEn0R:SOSFg0n9XOXFb0i9]O^F<c9CcF6_9IhF0X90nFJS95VGBk8>^GXOc8g0gGoNZ8P1PHfNP8Z1ZH\\Ng7c1cHQN`7n1jH\\Mc7c2fHiLf7W3b4O001O1O1O10O01O1O1O1O010O1O1O001O01O00000000001O01O00000001O00000001O0O1L4K5L4L4K5002N2O1N2N2N1YDTMX6n2dIVMZ6l2aIYM]6i2_I[M_6g2]I\\Mb6f2YI]Mh6d2TI_Mk6b2QIaMo6a2mHaMS7a2iHaMW7a2dHcM[7_2aHcM_7_2]HcMc7_2XHeMg7\\2VHfMk7[2QHgMo7[2lGiMS8Y2iGiMW8Y2dGjM\\8X2`GkM_8W2]GkMc8V2YGmMh8T2TGoMl8R2PGiMW9Y2cFcMf9^2TF^MU:c2eEZMb:h2XETMQ;l2jDQM^;P3[DQMi;Q3QDPMS<Q3gCPM]<Q3]CoLg<S3SCnLQ=S3hBoL\\=g43L5L4L3M4K5L3M4K5L3M4L4K4M4N2NhC]KU9b4jFaKV9]4iFeKX9Y4gFjKX9V4fFmKZ9U4aFnK`9R4\\FQLg9n3TFTLo9l3lEWLV:i3eEZL]:f3^E]Ld:b3XE`Lk:`3QEbLR;a3eDbL];f3VD]Ll;k3gCXL[<P4XCRLk<a52N3lKkC3Z<hNQEm0T;fM_Fo1T>D;D<E;E<DdYW2"}, "label": "a giraffe eating from a tree"}]}
{"id": 93176, "image": "COCO_train2014_000000093176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green be with red pillow near a woman opening a coke\"."}], "task": "refering", "answer_template": "The \"a green be with red pillow near a woman opening a coke\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 252, 253, 254, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 279, 280, 295, 296, 297, 298, 299, 300, 301, 302, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473], "bbox": [0.0, 0.4574019607843137, 0.7690522875816994, 0.9887908496732026], "iscrowd": 0, "area": 90943.99729999999, "rle": {"size": [612, 612], "counts": "fU14T>M`EZ1`:fNXDb2i;^MoBi3Q=U1000000001O01O000001O00000000001O000000001O00000000001O000000001O00000000010O00000000001O000000001O00000000001O000000001O00000000001O0001O000001O000000001O00000000001O000000001O00000000001O0000000010O0000000001O00000000001O000000001O00000000001O000000001O000001O000000000O1000000O1000000O1000000O1000000O1000000O1000000O101O000O1000000O1000000O1000000O1000000O100O100O1O100O100O1O100O2N100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O1oNQ1K5K5K5K5K5J6K5K5K5N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3O10000000000001O000000000000000000000POQ1VNi100O1000000O1000000O10001O0O1000000O1000000O10001O0O100001O00001O0000001O01O01O00001O00001O00001O00010O00001O00001O00001O00010O00001O00001O00001O000010O0001O00001O000000000001O0000000000000000000000001O00000000000O10000O10000O2O0O10000O10000O1O1O1O1O2N100O1O1O1O1O1O1O1O1O1O1L5J5K5K5K5K5K5K5K5C=ROn0QOo0ROo0POo0RO\\Pd2"}, "label": "a green be with red pillow near a woman opening a coke"}]}
{"id": 93176, "image": "COCO_train2014_000000093176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a twin bed that is nearest to the woman standing next to the window\"."}], "task": "refering", "answer_template": "The \"a twin bed that is nearest to the woman standing next to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 252, 253, 254, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 279, 280, 295, 296, 297, 298, 299, 300, 301, 302, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473], "bbox": [0.0, 0.4574019607843137, 0.7690522875816994, 0.9887908496732026], "iscrowd": 0, "area": 90943.99729999999, "rle": {"size": [612, 612], "counts": "fU14T>M`EZ1`:fNXDb2i;^MoBi3Q=U1000000001O01O000001O00000000001O000000001O00000000001O000000001O00000000010O00000000001O000000001O00000000001O000000001O00000000001O0001O000001O000000001O00000000001O000000001O00000000001O0000000010O0000000001O00000000001O000000001O00000000001O000000001O000001O000000000O1000000O1000000O1000000O1000000O1000000O1000000O101O000O1000000O1000000O1000000O1000000O100O100O1O100O100O1O100O2N100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O1oNQ1K5K5K5K5K5J6K5K5K5N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3O10000000000001O000000000000000000000POQ1VNi100O1000000O1000000O10001O0O1000000O1000000O10001O0O100001O00001O0000001O01O01O00001O00001O00001O00010O00001O00001O00001O00010O00001O00001O00001O000010O0001O00001O000000000001O0000000000000000000000001O00000000000O10000O10000O2O0O10000O10000O1O1O1O1O2N100O1O1O1O1O1O1O1O1O1O1L5J5K5K5K5K5K5K5K5C=ROn0QOo0ROo0POo0RO\\Pd2"}, "label": "a twin bed that is nearest to the woman standing next to the window"}]}
{"id": 93176, "image": "COCO_train2014_000000093176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed furthest from the person\"."}], "task": "refering", "answer_template": "The \"the bed furthest from the person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 260, 261, 262, 263, 282, 283, 284, 285, 304, 305, 306, 307, 326, 327, 328, 329, 348, 349, 350, 351, 369, 370, 371, 372, 373, 390, 391, 392, 393, 394, 395, 407, 408, 409, 410, 412, 413, 414, 415, 416, 417, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.5191013071895425, 0.4898856209150327, 1.0, 1.0], "iscrowd": 0, "area": 48769.56399999999, "rle": {"size": [612, 612], "counts": "ian5S3Q`0000O100000000000000O1000000000000O100000O10000000O1000000000000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001N1000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O000O2O0000001O0000001O0000001O00000000N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N101O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2M3H8I6I8H8I7H8H8I7H8I7H8H8I7H8H8I7H8K50000000000001O0000000000000000001O0000000000000000001O0O10000000000000001O0000000000000000001O0000000000000000001O00000000000000001O000000000O100000001O0000000hFgGe7R<jMV2jMW2iMV^O"}, "label": "the bed furthest from the person"}]}
{"id": 93176, "image": "COCO_train2014_000000093176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed farthest from the lady\"."}], "task": "refering", "answer_template": "The \"the bed farthest from the lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 260, 261, 262, 263, 282, 283, 284, 285, 304, 305, 306, 307, 326, 327, 328, 329, 348, 349, 350, 351, 369, 370, 371, 372, 373, 390, 391, 392, 393, 394, 395, 407, 408, 409, 410, 412, 413, 414, 415, 416, 417, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.5191013071895425, 0.4898856209150327, 1.0, 1.0], "iscrowd": 0, "area": 48769.56399999999, "rle": {"size": [612, 612], "counts": "ian5S3Q`0000O100000000000000O1000000000000O100000O10000000O1000000000000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001N1000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O000O2O0000001O0000001O0000001O00000000N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N101O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2M3H8I6I8H8I7H8H8I7H8I7H8H8I7H8H8I7H8K50000000000001O0000000000000000001O0000000000000000001O0O10000000000000001O0000000000000000001O0000000000000000001O00000000000000001O000000000O100000001O0000000hFgGe7R<jMV2jMW2iMV^O"}, "label": "the bed farthest from the lady"}]}
{"id": 322553, "image": "COCO_train2014_000000322553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two brown bears playing or fighting with each other in a body of water\"."}], "task": "refering", "answer_template": "The \"two brown bears playing or fighting with each other in a body of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241], "bbox": [0.18546875000000002, 0.24947483588621444, 0.5133593750000001, 0.6796717724288841], "iscrowd": 0, "area": 20067.34615000001, "rle": {"size": [457, 640], "counts": "oXe12U>3M3M3O2N1O1O1O1O1O1[B^Oa=c0^BA]=a0cB@[=i0N1O2N2N2M3M2N3M3N2M2N3M2N2N3M2N3N1N2N3M2N3M2N3M3M3N2M2N3M3L4M3QEbMn9a3J4M2N2F:J5O1O1O1O1O1N2O1O1O2O0000001O0O10001O02N100O1O2N100O1O2N100O1O2O0O1O1O2O1N3M3M3N2M3M3M3M3M3M1O1O2N1O1O1O1O2N100O1O2N1O100O2N1O2N101N1O2N1O2O1N1O2N1O2O0O001O0010O01O001O010O001O0010O01O000010O0001O0010O0001O0010O0001O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N1O2N2N2M3M3M3M3N2M3M3M3M3N2MS`Z4"}, "label": "two brown bears playing or fighting with each other in a body of water"}]}
{"id": 322553, "image": "COCO_train2014_000000322553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear getting bit\"."}], "task": "refering", "answer_template": "The \"a bear getting bit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241], "bbox": [0.18546875000000002, 0.24947483588621444, 0.5133593750000001, 0.6796717724288841], "iscrowd": 0, "area": 20067.34615000001, "rle": {"size": [457, 640], "counts": "oXe12U>3M3M3O2N1O1O1O1O1O1[B^Oa=c0^BA]=a0cB@[=i0N1O2N2N2M3M2N3M3N2M2N3M2N2N3M2N3N1N2N3M2N3M2N3M3M3N2M2N3M3L4M3QEbMn9a3J4M2N2F:J5O1O1O1O1O1N2O1O1O2O0000001O0O10001O02N100O1O2N100O1O2N100O1O2O0O1O1O2O1N3M3M3N2M3M3M3M3M3M1O1O2N1O1O1O1O2N100O1O2N1O100O2N1O2N101N1O2N1O2O1N1O2N1O2O0O001O0010O01O001O010O001O0010O01O000010O0001O0010O0001O0010O0001O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N1O2N2N2M3M3M3M3N2M3M3M3M3N2MS`Z4"}, "label": "a bear getting bit"}]}
{"id": 322553, "image": "COCO_train2014_000000322553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear trying to take a bite out of another bear\"."}], "task": "refering", "answer_template": "The \"a bear trying to take a bite out of another bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 146, 147, 148, 149, 150, 170, 171, 172, 173, 174, 194, 195, 196, 197, 198, 218, 219, 220, 221], "bbox": [0.34178125000000004, 0.3033698030634573, 0.630625, 0.6134792122538294], "iscrowd": 0, "area": 12834.8026, "rle": {"size": [457, 640], "counts": "UlQ32W>2N3M2N3N1N2N2N2N2N7I6J2O1N2N2N2N2N2N2N10O00000000O100O100O100O100O1O100O2O0O2O1N2O001O1O001O1O1N2O1O2N1O1O1O2N1O010O1O1O010O1O010O1O1O010O10O01O101N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1O00O1O1OO1O1O1O1O1O1O100001N2eLgEd2Z:ZMhEe2Y:XMjEg2Z:SMjEk2[:oLgEP3k:N1N3N1O1O1O1O0O2O001O1O001O1N101O1O001O0O2O1O001O1O001N2O1O1O1O1O1N2O2N1O2N2N2N1N3N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2NX`Y3"}, "label": "a bear trying to take a bite out of another bear"}]}
{"id": 322553, "image": "COCO_train2014_000000322553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear playing with his colleague like biting\"."}], "task": "refering", "answer_template": "The \"a teddy bear playing with his colleague like biting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 146, 147, 148, 149, 150, 170, 171, 172, 173, 174, 194, 195, 196, 197, 198, 218, 219, 220, 221], "bbox": [0.34178125000000004, 0.3033698030634573, 0.630625, 0.6134792122538294], "iscrowd": 0, "area": 12834.8026, "rle": {"size": [457, 640], "counts": "UlQ32W>2N3M2N3N1N2N2N2N2N7I6J2O1N2N2N2N2N2N2N10O00000000O100O100O100O100O1O100O2O0O2O1N2O001O1O001O1O1N2O1O2N1O1O1O2N1O010O1O1O010O1O010O1O1O010O10O01O101N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1O00O1O1OO1O1O1O1O1O1O100001N2eLgEd2Z:ZMhEe2Y:XMjEg2Z:SMjEk2[:oLgEP3k:N1N3N1O1O1O1O0O2O001O1O001O1N101O1O001O0O2O1O001O1O001N2O1O1O1O1O1N2O2N1O2N2N2N1N3N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2NX`Y3"}, "label": "a teddy bear playing with his colleague like biting"}]}
{"id": 52219, "image": "COCO_train2014_000000052219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table in an asian inspired restaurant\"."}], "task": "refering", "answer_template": "The \"a wooden table in an asian inspired restaurant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.29584375, 0.6314598540145985, 1.0, 0.7483211678832116], "iscrowd": 0, "area": 17274.8272, "rle": {"size": [411, 640], "counts": "fS\\2Q1j;2N2N1N3N1O1O0O100000000OkNdDm0\\;QOfDo0Z;oNhDQ1X;mNjDS1_;0000010O00000000000000000000000000000000000000001O000000000000000000000000000000000000001O000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O10000000000000000000O1000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000001O00000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000001O0000000000=CUG"}, "label": "a wooden table in an asian inspired restaurant"}]}
{"id": 52219, "image": "COCO_train2014_000000052219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"edge of the brown table closest to the camera\"."}], "task": "refering", "answer_template": "The \"edge of the brown table closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.29584375, 0.6314598540145985, 1.0, 0.7483211678832116], "iscrowd": 0, "area": 17274.8272, "rle": {"size": [411, 640], "counts": "fS\\2Q1j;2N2N1N3N1O1O0O100000000OkNdDm0\\;QOfDo0Z;oNhDQ1X;mNjDS1_;0000010O00000000000000000000000000000000000000001O000000000000000000000000000000000000001O000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O10000000000000000000O1000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000001O00000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000001O0000000000=CUG"}, "label": "edge of the brown table closest to the camera"}]}
{"id": 52219, "image": "COCO_train2014_000000052219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in black shirt eating\"."}], "task": "refering", "answer_template": "The \"girl in black shirt eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 84, 85, 86, 107, 108, 109, 130, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201], "bbox": [0.6653749999999999, 0.19043795620437956, 0.8088906249999999, 0.5764476885644769], "iscrowd": 0, "area": 8666.873750000004, "rle": {"size": [411, 640], "counts": "bS[52g<2N2M3N3]OEbD=\\;b0N3M3M4L3M3M4L5L6K5J3N2M4M2M2OO010O00aFQNT8o1eGaNR8^1hGkNU8U1eGSOZ8h2N2O1N3N1N2O1N2O2O0O100O2O0O01O001O001O001O001O001O010O001O001O001O001O4L6[LgG\\2_8nLQH^2a9N1O2M2O2M2O1O2M2O2N1N3N1N3N1N3M3M2N3M2N3M3N1N3M2NZj`1"}, "label": "girl in black shirt eating"}]}
{"id": 52219, "image": "COCO_train2014_000000052219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an attractive woman in black\"."}], "task": "refering", "answer_template": "The \"an attractive woman in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 84, 85, 86, 107, 108, 109, 130, 131, 132, 133, 154, 155, 156, 177, 178, 179, 200, 201], "bbox": [0.6653749999999999, 0.19043795620437956, 0.8088906249999999, 0.5764476885644769], "iscrowd": 0, "area": 8666.873750000004, "rle": {"size": [411, 640], "counts": "bS[52g<2N2M3N3]OEbD=\\;b0N3M3M4L3M3M4L5L6K5J3N2M4M2M2OO010O00aFQNT8o1eGaNR8^1hGkNU8U1eGSOZ8h2N2O1N3N1N2O1N2O2O0O100O2O0O01O001O001O001O001O001O010O001O001O001O001O4L6[LgG\\2_8nLQH^2a9N1O2M2O2M2O1O2M2O2N1N3N1N3N1N3M3M2N3M2N3M3N1N3M2NZj`1"}, "label": "an attractive woman in black"}]}
{"id": 52219, "image": "COCO_train2014_000000052219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with the pink shirt\"."}], "task": "refering", "answer_template": "The \"the girl with the pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 234, 235, 236], "bbox": [0.20590625, 0.177007299270073, 0.4924375, 0.7330656934306569], "iscrowd": 0, "area": 21097.764649999997, "rle": {"size": [411, 640], "counts": "bWe1:]<=B?B6J4K6K5J5L5K5L3M4M3L3M4L4L3M3M3M3N2M3M3M3M3M3M3M3M3N2M3M3N2O0O2O1O1O001O1O1O1N101O1O1O1O001O1N2O1O001O1O1O1N101O1000L4K5K5L4K5M3O1O100O1O100O1O100O00100O1O100O1O100O1O100O1O100^Ob000O1000000O100000SGhLW8X3gGkLX8T3gGnLY8R3eGQMX8P3gGSMW8n2gGTMX8l2gGUMY8k2gGVMX8k2fGVMZ8j2eGXMZ8h2eGZMX8i2gGWMU8m2jGUMQ8P3nGPMm7U3SHlLS7XO_In3]OlLT7[OXIk3CjLV7@PIe4P7b01PMkH:U7FmH8S7GQI5P7KRI3n6LVI0l6OVILm64VIGl68WIEj6;XIBi6?YI]Oh6c0ZIZOg6g0[IUOf6l0[IROe6n0]IoNd6R1]IkNd6V1^IfNc6Z1^IdNc6]1[IcNf6]1ZIbNh6^1VIaNm6^1SI`NQ7^1nHbNU7\\1kHbNY7\\1fHdN[7\\1eHbN^7\\1bHdN_7\\1`HbNd7]1\\H]Nj7c1UH]Nl7c1TH]Nm7a1SH_Nn7a1RH^NP8a1oG`NQ8`1oG_NR8a1nG_NS8`1mG_NT8`1nG^NT8`1mG`NS8_1nG`NS8_1nGcNQ8[1PHgNn7X1SHiNm7U1UHlNi7S1XHmNh7R1ZHkNi7S1[HhNg7W1\\HdNh7Z1[H`Ni7_1j1N3M2O2M3M4L4M3L4L4M4K4L4M3L4LeZR4"}, "label": "the girl with the pink shirt"}]}
{"id": 52219, "image": "COCO_train2014_000000052219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a pink shirt sitting at a table while eating\"."}], "task": "refering", "answer_template": "The \"a woman wearing a pink shirt sitting at a table while eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 234, 235, 236], "bbox": [0.20590625, 0.177007299270073, 0.4924375, 0.7330656934306569], "iscrowd": 0, "area": 21097.764649999997, "rle": {"size": [411, 640], "counts": "bWe1:]<=B?B6J4K6K5J5L5K5L3M4M3L3M4L4L3M3M3M3N2M3M3M3M3M3M3M3M3N2M3M3N2O0O2O1O1O001O1O1O1N101O1O1O1O001O1N2O1O001O1O1O1N101O1000L4K5K5L4K5M3O1O100O1O100O1O100O00100O1O100O1O100O1O100O1O100^Ob000O1000000O100000SGhLW8X3gGkLX8T3gGnLY8R3eGQMX8P3gGSMW8n2gGTMX8l2gGUMY8k2gGVMX8k2fGVMZ8j2eGXMZ8h2eGZMX8i2gGWMU8m2jGUMQ8P3nGPMm7U3SHlLS7XO_In3]OlLT7[OXIk3CjLV7@PIe4P7b01PMkH:U7FmH8S7GQI5P7KRI3n6LVI0l6OVILm64VIGl68WIEj6;XIBi6?YI]Oh6c0ZIZOg6g0[IUOf6l0[IROe6n0]IoNd6R1]IkNd6V1^IfNc6Z1^IdNc6]1[IcNf6]1ZIbNh6^1VIaNm6^1SI`NQ7^1nHbNU7\\1kHbNY7\\1fHdN[7\\1eHbN^7\\1bHdN_7\\1`HbNd7]1\\H]Nj7c1UH]Nl7c1TH]Nm7a1SH_Nn7a1RH^NP8a1oG`NQ8`1oG_NR8a1nG_NS8`1mG_NT8`1nG^NT8`1mG`NS8_1nG`NS8_1nGcNQ8[1PHgNn7X1SHiNm7U1UHlNi7S1XHmNh7R1ZHkNi7S1[HhNg7W1\\HdNh7Z1[H`Ni7_1j1N3M2O2M3M4L4M3L4L4M4K4L4M3L4LeZR4"}, "label": "a woman wearing a pink shirt sitting at a table while eating"}]}
{"id": 470012, "image": "COCO_train2014_000000470012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with a man in blue horts on top\"."}], "task": "refering", "answer_template": "The \"elephant with a man in blue horts on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 295, 296, 297], "bbox": [0.20090625, 0.3417330210772833, 1.0, 0.8181264637002341], "iscrowd": 0, "area": 72653.94434999999, "rle": {"size": [427, 640], "counts": "PSf14T=5L4K6K4K5L4L5J4M1N3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M2O2M2O2N1N3N1N2O2M2O2M2O2M2O2N1N3N1N3N1N3N1N3N1N3N1O2M201N1O101N101N101`GkKh7V4WHlKg7V4WHlKh7T4VHnKi7T4UHnKj7R4UHPLj7Q4SHQLn7n3QHTLo7l3oGVLQ8i3nGXLR8i3kGZLU8e3jG]LV8\\40O01000O1000O010000O10O1000O10O10O10000O01000O10O1000O1000O010000O10000O1000000O10001N10000O2O000O101O0O10001O0O10001N10000O2O000O10a0_O000O10000O11O1O1O1O1O1O1O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O001O1O001O000O10000000000000000O2O000000000000000O100000001O0000000O100000000000001O000O1000000000[KSHl3m7mKZHS4f7gK`HY4`7`KhH_4R810O001O00001O001O001O000010O01O00001O00YOfK^HY4a7hK_HX4_7kKaHT4^7mKbHS4]7oKbHQ4\\7RLdHn3Z7SLfHm3X7VLhHi3X7XLgHh3X7YLiHf3W7[LhHe3X7[LhHe3X7\\LhHc3X7^LgHb3Y7^LhHa3W7aLhH_3X7aLiH^3W7cLhH]3X7dLgH]3X7cLiH1@^2g7bMhHOB_2f7cMhHLCa2d7dMiHIEc2b7eMhHGGd2a7eMiHEGf2`7fMhHBJh2^7gMhH_OKj2]7gMhH]OMl2[7hMhHZONn2Y7jMhHWO0o2X7jMhHUO2R3U7jMiHRO3T3T7jMiHQO4U3S7kMiHmN6X3Q7lMhHkN8Y3o6mM`IS2`6nM_IR2a6nM_IR2a6oM^IQ2b6PN]IP2c6PN]IP2c6QN\\Io1c6SN\\In1c6RN]In1c6SNYIP2g6PNVIS2j6nMRIU2n6lMnHW2R7iMkHZ2T7hMhH[2X7eMeH^2[7cMaH`2_7f1000001N10000000000O10000000001O0O1000000001O0000001O00010O0000001O00001O00000010O0001O0000001O00001O0001O01O00001O0000001O00001O010O001O001O001O001O001O010O00001O001O001O001O001O010O001O001O001O001O001O010O00001O001O001O001O0010O01O001O001O001O001O001O010O001O00001O001O001O1O010O1O001O001O1O001O001O1O010O1O001O001O1O001O001O10O01O001O1O001O1O001O1O100O2N1O1O2N1O1WN`FEb97nF[OS9b0]GPOd8m0lGeNV8W1Q2L4M4L3M3N3M2O1O2N1OO001N2O1N2O0O21O1O1O1N3N1O1O1O1M3L4K6J5KXH"}, "label": "elephant with a man in blue horts on top"}]}
{"id": 470012, "image": "COCO_train2014_000000470012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with a man riding him on the back\"."}], "task": "refering", "answer_template": "The \"elephant with a man riding him on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 295, 296, 297], "bbox": [0.20090625, 0.3417330210772833, 1.0, 0.8181264637002341], "iscrowd": 0, "area": 72653.94434999999, "rle": {"size": [427, 640], "counts": "PSf14T=5L4K6K4K5L4L5J4M1N3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M2O2M2O2N1N3N1N2O2M2O2M2O2M2O2N1N3N1N3N1N3N1N3N1N3N1O2M201N1O101N101N101`GkKh7V4WHlKg7V4WHlKh7T4VHnKi7T4UHnKj7R4UHPLj7Q4SHQLn7n3QHTLo7l3oGVLQ8i3nGXLR8i3kGZLU8e3jG]LV8\\40O01000O1000O010000O10O1000O10O10O10000O01000O10O1000O1000O010000O10000O1000000O10001N10000O2O000O101O0O10001O0O10001N10000O2O000O10a0_O000O10000O11O1O1O1O1O1O1O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O001O1O001O000O10000000000000000O2O000000000000000O100000001O0000000O100000000000001O000O1000000000[KSHl3m7mKZHS4f7gK`HY4`7`KhH_4R810O001O00001O001O001O000010O01O00001O00YOfK^HY4a7hK_HX4_7kKaHT4^7mKbHS4]7oKbHQ4\\7RLdHn3Z7SLfHm3X7VLhHi3X7XLgHh3X7YLiHf3W7[LhHe3X7[LhHe3X7\\LhHc3X7^LgHb3Y7^LhHa3W7aLhH_3X7aLiH^3W7cLhH]3X7dLgH]3X7cLiH1@^2g7bMhHOB_2f7cMhHLCa2d7dMiHIEc2b7eMhHGGd2a7eMiHEGf2`7fMhHBJh2^7gMhH_OKj2]7gMhH]OMl2[7hMhHZONn2Y7jMhHWO0o2X7jMhHUO2R3U7jMiHRO3T3T7jMiHQO4U3S7kMiHmN6X3Q7lMhHkN8Y3o6mM`IS2`6nM_IR2a6nM_IR2a6oM^IQ2b6PN]IP2c6PN]IP2c6QN\\Io1c6SN\\In1c6RN]In1c6SNYIP2g6PNVIS2j6nMRIU2n6lMnHW2R7iMkHZ2T7hMhH[2X7eMeH^2[7cMaH`2_7f1000001N10000000000O10000000001O0O1000000001O0000001O00010O0000001O00001O00000010O0001O0000001O00001O0001O01O00001O0000001O00001O010O001O001O001O001O001O010O00001O001O001O001O001O010O001O001O001O001O001O010O00001O001O001O001O0010O01O001O001O001O001O001O010O001O00001O001O001O1O010O1O001O001O1O001O001O1O010O1O001O001O1O001O001O10O01O001O1O001O1O001O1O100O2N1O1O2N1O1WN`FEb97nF[OS9b0]GPOd8m0lGeNV8W1Q2L4M4L3M3N3M2O1O2N1OO001N2O1N2O0O21O1O1O1N3N1O1O1O1M3L4K6J5KXH"}, "label": "elephant with a man riding him on the back"}]}
{"id": 470012, "image": "COCO_train2014_000000470012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with blue shorts on elephant\"."}], "task": "refering", "answer_template": "The \"boy with blue shorts on elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 55, 56, 57, 58, 79, 80, 81, 82, 102, 103, 104, 105, 106, 124, 125, 127, 128, 150, 151, 152, 175, 176], "bbox": [0.41345312500000003, 0.08299765807962528, 0.66121875, 0.5326463700234192], "iscrowd": 0, "area": 10182.440100000002, "rle": {"size": [427, 640], "counts": "Te^32Y=1O1O1O2M2O1O1O2hLF[I;a6OWI3e62YIOd64[IKd69ZIGd6=ZIBf6`0XIAf6c0WI]Oi6e0TI]Ok6f0RIZOo6f0PI[Oo6g0oHYOR7g0mHZOS7f0kH[OU7f0jH[OV7e0iHgNPNa0X9i0gHZOY7f0dH\\O\\7e0aH^O_7b0^H@c7a0YHBf7`0WHAj7`0RHCn7>nGCS8>jGBY8?cGB_8?]GAe8a0WG_Om8a0PG_OR9a0ZFeN3j0g9V23M3N3N1O2N1O4L5K4M6I:F2N1O1O2N101N1O2N1O1O2N1O2O0O2N1O2N1O101O0010O0lEhNV8X1aGPO_8P1ZGXOe8i0RG_Oo8`0iFHW98aF1_9OXF9h9GQFa0o9]101O010O01O010O010O01O010O01O010O01O010O01O010O010O01O010OVOSFVNn9f1UFZNk9c1YF]Ng9_1\\FaNd9]1_FcN`9]1`FcNa9\\1`FdNb9Y1^FgNd9W1]FiNe9T1[FlNg9R1ZFmNi9n0YFSOh9g0]FXOe9c0^F^Od9<`FCb97bFJ`90dFO^9LfF4[9FiF:Z9@jF?Q;1mC]O];b0`DA`;`0\\DCe;<WDHi;9SDIo;6nCMR<f01O001O0O2O001O1O0O2N1O2N1N3N1O2N1O2N1O4L7IVYj2"}, "label": "boy with blue shorts on elephant"}]}
{"id": 470012, "image": "COCO_train2014_000000470012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in blue shorts riding an elephant\"."}], "task": "refering", "answer_template": "The \"the man in blue shorts riding an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 55, 56, 57, 58, 79, 80, 81, 82, 102, 103, 104, 105, 106, 124, 125, 127, 128, 150, 151, 152, 175, 176], "bbox": [0.41345312500000003, 0.08299765807962528, 0.66121875, 0.5326463700234192], "iscrowd": 0, "area": 10182.440100000002, "rle": {"size": [427, 640], "counts": "Te^32Y=1O1O1O2M2O1O1O2hLF[I;a6OWI3e62YIOd64[IKd69ZIGd6=ZIBf6`0XIAf6c0WI]Oi6e0TI]Ok6f0RIZOo6f0PI[Oo6g0oHYOR7g0mHZOS7f0kH[OU7f0jH[OV7e0iHgNPNa0X9i0gHZOY7f0dH\\O\\7e0aH^O_7b0^H@c7a0YHBf7`0WHAj7`0RHCn7>nGCS8>jGBY8?cGB_8?]GAe8a0WG_Om8a0PG_OR9a0ZFeN3j0g9V23M3N3N1O2N1O4L5K4M6I:F2N1O1O2N101N1O2N1O1O2N1O2O0O2N1O2N1O101O0010O0lEhNV8X1aGPO_8P1ZGXOe8i0RG_Oo8`0iFHW98aF1_9OXF9h9GQFa0o9]101O010O01O010O010O01O010O01O010O01O010O01O010O010O01O010OVOSFVNn9f1UFZNk9c1YF]Ng9_1\\FaNd9]1_FcN`9]1`FcNa9\\1`FdNb9Y1^FgNd9W1]FiNe9T1[FlNg9R1ZFmNi9n0YFSOh9g0]FXOe9c0^F^Od9<`FCb97bFJ`90dFO^9LfF4[9FiF:Z9@jF?Q;1mC]O];b0`DA`;`0\\DCe;<WDHi;9SDIo;6nCMR<f01O001O0O2O001O1O0O2N1O2N1N3N1O2N1O2N1O4L7IVYj2"}, "label": "the man in blue shorts riding an elephant"}]}
{"id": 404475, "image": "COCO_train2014_000000404475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bag the woman in blue is carrying on her shoulder\"."}], "task": "refering", "answer_template": "The \"bag the woman in blue is carrying on her shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 123, 124, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197], "bbox": [0.8629199999999999, 0.3884, 0.99606, 0.8176000000000001], "iscrowd": 0, "area": 5283.267199999998, "rle": {"size": [375, 500], "counts": "T[n43f;2iFMd66UIOj64oH1FI^58fJ4BMi5;TJNHOU6a0aIEI9g6`0nHo0T7i12N2N2M4M2N2N2M3N2N2N3L3N2N2M3N2N2N3L3N2N2N2M3N2M100O10000O100O010O10000O101N3N2M3N2N2M3N2M3N2M4M2N2M3N2M3N2M3F:E;Eh?"}, "label": "bag the woman in blue is carrying on her shoulder"}]}
{"id": 404475, "image": "COCO_train2014_000000404475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown piece of carry on luggage held by a woman in a blue jumpsuit\"."}], "task": "refering", "answer_template": "The \"a brown piece of carry on luggage held by a woman in a blue jumpsuit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 123, 124, 141, 142, 143, 159, 160, 161, 177, 178, 179, 195, 196, 197], "bbox": [0.8629199999999999, 0.3884, 0.99606, 0.8176000000000001], "iscrowd": 0, "area": 5283.267199999998, "rle": {"size": [375, 500], "counts": "T[n43f;2iFMd66UIOj64oH1FI^58fJ4BMi5;TJNHOU6a0aIEI9g6`0nHo0T7i12N2N2M4M2N2N2M3N2N2N3L3N2N2M3N2N2N3L3N2N2N2M3N2M100O10000O100O010O10000O101N3N2M3N2N2M3N2M3N2M4M2N2M3N2M3N2M3F:E;Eh?"}, "label": "a brown piece of carry on luggage held by a woman in a blue jumpsuit"}]}
{"id": 330752, "image": "COCO_train2014_000000330752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a desk chair holding a woman with brown hair in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a desk chair holding a woman with brown hair in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307], "bbox": [0.172046875, 0.477425, 0.526875, 0.9505500000000001], "iscrowd": 0, "area": 25753.397999999994, "rle": {"size": [400, 640], "counts": "iY[14n;>B?A>B>B>B>C=B?A>L4N2N2N2M3N3M2N200O100O100O100O2O0O100O10000O100O2Oj0VO00000000000O10000000O1000O100000000000000001O00000O100000000000001O0000000000000O1000001O0000001O001O00001O00001N101O00001O1O1O1O1O1O1O1O1O1O2N2N1O2N2N2M2O2N2N2N2N3M2N3M2N2N2N1O2N1O2N2N1O1O1O1N101O001O1O0O2N1PMSG\\2n8`MgGn1Z8oMmGk1U8QNRHk1o7RNVHk1Y9N3M2N2O1O0O10000O10001N10000O10000O10000O10000O10000O10000O10000O10000O01000O10000O010O10000O10O0100O10000O00100O1O100O1O10O01O1O100O1O10O02N101N1O2O0O2N1O2N2N1N3NiQf3"}, "label": "a desk chair holding a woman with brown hair in a blue shirt"}]}
{"id": 330752, "image": "COCO_train2014_000000330752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a woman in a blue shirt sitting in it\"."}], "task": "refering", "answer_template": "The \"a chair with a woman in a blue shirt sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307], "bbox": [0.172046875, 0.477425, 0.526875, 0.9505500000000001], "iscrowd": 0, "area": 25753.397999999994, "rle": {"size": [400, 640], "counts": "iY[14n;>B?A>B>B>B>C=B?A>L4N2N2N2M3N3M2N200O100O100O100O2O0O100O10000O100O2Oj0VO00000000000O10000000O1000O100000000000000001O00000O100000000000001O0000000000000O1000001O0000001O001O00001O00001N101O00001O1O1O1O1O1O1O1O1O1O2N2N1O2N2N2M2O2N2N2N2N3M2N3M2N2N2N1O2N1O2N2N1O1O1O1N101O001O1O0O2N1PMSG\\2n8`MgGn1Z8oMmGk1U8QNRHk1o7RNVHk1Y9N3M2N2O1O0O10000O10001N10000O10000O10000O10000O10000O10000O10000O10000O01000O10000O010O10000O10O0100O10000O00100O1O100O1O10O01O1O100O1O10O02N101N1O2O0O2N1O2N2N1N3NiQf3"}, "label": "a chair with a woman in a blue shirt sitting in it"}]}
{"id": 330752, "image": "COCO_train2014_000000330752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing blue long shirt working in front of computer\"."}], "task": "refering", "answer_template": "The \"a woman wearing blue long shirt working in front of computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 169, 170, 171, 172, 173, 193, 194, 195], "bbox": [0.28512499999999996, 0.26344999999999996, 0.53028125, 0.602875], "iscrowd": 0, "area": 7014.821499999999, "rle": {"size": [400, 640], "counts": "jXW21\\<:F3J6K5N2O1O1O1O1O1O1ZOSOiEm0R:_OfEc0V:FcE;[:n0O1N3N1N3M2O1N2N3M2N2M3O10001N1000000O101O000O101O000000000010O000001O1O1O100O1O0010000O2O0010O0mLhFm2`9N2[N`F5a9HbF7_9EdF;]9BfF=Z9@jF`0^9TOgFi0\\9nNmFP1\\:N1N1000101N2N1O1O2N1O1O2N1O1O2M2N2N00000000000000O1000O1000000O0100000O10000O0100000O10O0100O1O10O01O100000O10000O10000O100000O01000000000000O1001O0000001O000O101O0O1IkC3V<Ll[e3"}, "label": "a woman wearing blue long shirt working in front of computer"}]}
{"id": 330752, "image": "COCO_train2014_000000330752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with brown hair sitting in front of a computer\"."}], "task": "refering", "answer_template": "The \"a woman with brown hair sitting in front of a computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 121, 122, 123, 124, 144, 145, 146, 147, 169, 170, 171, 172, 173, 193, 194, 195], "bbox": [0.28512499999999996, 0.26344999999999996, 0.53028125, 0.602875], "iscrowd": 0, "area": 7014.821499999999, "rle": {"size": [400, 640], "counts": "jXW21\\<:F3J6K5N2O1O1O1O1O1O1ZOSOiEm0R:_OfEc0V:FcE;[:n0O1N3N1N3M2O1N2N3M2N2M3O10001N1000000O101O000O101O000000000010O000001O1O1O100O1O0010000O2O0010O0mLhFm2`9N2[N`F5a9HbF7_9EdF;]9BfF=Z9@jF`0^9TOgFi0\\9nNmFP1\\:N1N1000101N2N1O1O2N1O1O2N1O1O2M2N2N00000000000000O1000O1000000O0100000O10000O0100000O10O0100O1O10O01O100000O10000O10000O100000O01000000000000O1001O0000001O000O101O0O1IkC3V<Ll[e3"}, "label": "a woman with brown hair sitting in front of a computer"}]}
{"id": 330752, "image": "COCO_train2014_000000330752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stations emblem\"."}], "task": "refering", "answer_template": "The \"the stations emblem\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 245, 246, 247, 248, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319], "bbox": [0.6191718749999999, 0.676825, 0.912421875, 0.95415], "iscrowd": 0, "area": 13185.044650000002, "rle": {"size": [400, 640], "counts": "lQk42[<3M2M4M3M3L4M3M3L4M3M3L4N2M3M3M3M3M2N3M3M3M3M3M3M3N2M3M3M3N2O1N2O1N2O1O1O1O100O100O1O100O100O100O100O100O1000000001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O001O1O001O001O001O1O1O1O1O1O1O1O1O1O1O1O2N3M2N2N2N2N2N3M2N2N2N3M4L3MU`e0"}, "label": "the stations emblem"}]}
{"id": 330752, "image": "COCO_train2014_000000330752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the k5 symbol and chair in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the k5 symbol and chair in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 245, 246, 247, 248, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319], "bbox": [0.6191718749999999, 0.676825, 0.912421875, 0.95415], "iscrowd": 0, "area": 13185.044650000002, "rle": {"size": [400, 640], "counts": "lQk42[<3M2M4M3M3L4M3M3L4M3M3L4N2M3M3M3M3M2N3M3M3M3M3M3M3N2M3M3M3N2O1N2O1N2O1O1O1O100O100O1O100O100O100O100O100O1000000001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O001O1O001O001O001O1O1O1O1O1O1O1O1O1O1O1O2N3M2N2N2N2N2N3M2N2N2N3M4L3MU`e0"}, "label": "the k5 symbol and chair in the right hand picture"}]}
{"id": 248835, "image": "COCO_train2014_000000248835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a striped shirt taking a plate\"."}], "task": "refering", "answer_template": "The \"a man with a striped shirt taking a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 43, 44, 45, 46, 61, 62, 63, 77, 78, 79, 80, 95, 96, 112, 113], "bbox": [0.35914583333333333, 0.001359375, 0.7407083333333333, 0.295984375], "iscrowd": 0, "area": 12545.93105, "rle": {"size": [640, 480], "counts": "SU\\31oc00O2O0O101N100O101N10001N100M4K40000O2O000O2O000O101O0O10000000000001O01O0001O0000001O0000001O01O0001O00001O0001O01O0001O01O01O0000010O00000010O0001O0001O01O0000010O00001O0000001O000O101O0000001N10001O00000O2a]OPOX10j>P1l_OSOT13o>j0k_OVOP17S?c0k_OYOk0;Y?=h_O\\Oi0>]?6h_O_Oe0a0W?;R@XO?e0o>Z1k@lNS?V1g@QOW?m2O1O2N1O2N1O2N1O1O001O00001O001O002N2N1O2N2N1O2N1O2O1N1O2N2N1O2N1OK5K5K5K5K5N2N2N2N2N2O1N2N2N2N2N2N2M3N2N2M3N2N2YOP@[MS`0i1o_O\\Nf`0^1d_O]N]`0]1m_O^NT`0\\1V@_Nk?[1_@`Nb?Z1e1J6J6J7H7J6JST^2"}, "label": "a man with a striped shirt taking a plate"}]}
{"id": 248835, "image": "COCO_train2014_000000248835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the stripped shirt taking the plate\"."}], "task": "refering", "answer_template": "The \"the man in the stripped shirt taking the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 43, 44, 45, 46, 61, 62, 63, 77, 78, 79, 80, 95, 96, 112, 113], "bbox": [0.35914583333333333, 0.001359375, 0.7407083333333333, 0.295984375], "iscrowd": 0, "area": 12545.93105, "rle": {"size": [640, 480], "counts": "SU\\31oc00O2O0O101N100O101N10001N100M4K40000O2O000O2O000O101O0O10000000000001O01O0001O0000001O0000001O01O0001O00001O0001O01O0001O01O01O0000010O00000010O0001O0001O01O0000010O00001O0000001O000O101O0000001N10001O00000O2a]OPOX10j>P1l_OSOT13o>j0k_OVOP17S?c0k_OYOk0;Y?=h_O\\Oi0>]?6h_O_Oe0a0W?;R@XO?e0o>Z1k@lNS?V1g@QOW?m2O1O2N1O2N1O2N1O1O001O00001O001O002N2N1O2N2N1O2N1O2O1N1O2N2N1O2N1OK5K5K5K5K5N2N2N2N2N2O1N2N2N2N2N2N2M3N2N2M3N2N2YOP@[MS`0i1o_O\\Nf`0^1d_O]N]`0]1m_O^NT`0\\1V@_Nk?[1_@`Nb?Z1e1J6J6J7H7J6JST^2"}, "label": "the man in the stripped shirt taking the plate"}]}
{"id": 117770, "image": "COCO_train2014_000000117770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in a blue shirt and grey pants pitching a baseball\"."}], "task": "refering", "answer_template": "The \"boy in a blue shirt and grey pants pitching a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 75, 76, 77, 92, 93, 94, 95, 110, 111, 112, 126, 127, 128, 129, 143, 144, 145, 160, 161, 162, 176, 177, 178, 179], "bbox": [0.38677083333333334, 0.22555555555555556, 0.6154791666666667, 0.8556111111111111], "iscrowd": 0, "area": 9759.5806, "rle": {"size": [360, 480], "counts": "aeQ24R;3O1O1O001O1O0O2O1O0eKDjM=U2CkM>S2DkM=U2CkM=T2EjM<V2EiM;V2GhM:X2FhM:W2GiM8X2IfM8Y2HhM8W2HiM9V2GkM9T2HkM:S2FnM;P2EPN=n1CSN=l1CTN?j1ATNb0k1_OPNg0m1ZOoMk0P2UOfLW2X3iMhLX2W3hMiLZ2U3gMiL[2X2XMXL=`1\\2W2ZMVL:b1]2Y2[MRL8e1^2X2[MSL7d1_2X2[MSL7d1_2X2[MTL6c1l1oMPN[4NSL6b1j1ZNkMP46SL6c1f1R3UN[K5b1d1U3bNnJKl1a1X3iNgJGQ2]1Z3DeL;\\3FcL9]3JbL3`30]LOd33ZLLh35WLM^OVNn2o1aMO]OSNR3o1_M1\\OQNU3o1^M2[OoMW3Q2[M3[OmMZ3Q2YM5YOlM]3S2UM5ZOjMa3T2QM6YOhMf3U2lL9XOcMm3W2gL:WOaMS4W2aL=WO^MX4Y2\\Lj0e3YOULeNDa0W4m0PLcNM=T4S1jK_N6<P4i2TLTMm3j2VLTMj3k2ZLRMg3l2\\LQMg3l2\\LRMf3k2^LRMd3k2_LSMb3k2bLRM`3k2eLQM]3l2gLQMh3_2]L\\Mf3a2^L\\Mc3b2bLZM_3d2cL[M_3b2cL]M^3a2eL]M]3`2eL^M]3a2dL^M^3_2eL_M]3^2eLaM]3\\2gLaMZ3]2kL_MW3_2k2M3M3M2N3N2M3M3M2N3N3L3L5L4K4M4K5L3L5KmkP2"}, "label": "boy in a blue shirt and grey pants pitching a baseball"}]}
{"id": 117770, "image": "COCO_train2014_000000117770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy was playing\"."}], "task": "refering", "answer_template": "The \"a boy was playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 75, 76, 77, 92, 93, 94, 95, 110, 111, 112, 126, 127, 128, 129, 143, 144, 145, 160, 161, 162, 176, 177, 178, 179], "bbox": [0.38677083333333334, 0.22555555555555556, 0.6154791666666667, 0.8556111111111111], "iscrowd": 0, "area": 9759.5806, "rle": {"size": [360, 480], "counts": "aeQ24R;3O1O1O001O1O0O2O1O0eKDjM=U2CkM>S2DkM=U2CkM=T2EjM<V2EiM;V2GhM:X2FhM:W2GiM8X2IfM8Y2HhM8W2HiM9V2GkM9T2HkM:S2FnM;P2EPN=n1CSN=l1CTN?j1ATNb0k1_OPNg0m1ZOoMk0P2UOfLW2X3iMhLX2W3hMiLZ2U3gMiL[2X2XMXL=`1\\2W2ZMVL:b1]2Y2[MRL8e1^2X2[MSL7d1_2X2[MSL7d1_2X2[MTL6c1l1oMPN[4NSL6b1j1ZNkMP46SL6c1f1R3UN[K5b1d1U3bNnJKl1a1X3iNgJGQ2]1Z3DeL;\\3FcL9]3JbL3`30]LOd33ZLLh35WLM^OVNn2o1aMO]OSNR3o1_M1\\OQNU3o1^M2[OoMW3Q2[M3[OmMZ3Q2YM5YOlM]3S2UM5ZOjMa3T2QM6YOhMf3U2lL9XOcMm3W2gL:WOaMS4W2aL=WO^MX4Y2\\Lj0e3YOULeNDa0W4m0PLcNM=T4S1jK_N6<P4i2TLTMm3j2VLTMj3k2ZLRMg3l2\\LQMg3l2\\LRMf3k2^LRMd3k2_LSMb3k2bLRM`3k2eLQM]3l2gLQMh3_2]L\\Mf3a2^L\\Mc3b2bLZM_3d2cL[M_3b2cL]M^3a2eL]M]3`2eL^M]3a2dL^M^3_2eL_M]3^2eLaM]3\\2gLaMZ3]2kL_MW3_2k2M3M3M2N3N2M3M3M2N3N3L3L5L4K4M4K5L3L5KmkP2"}, "label": "a boy was playing"}]}
{"id": 297995, "image": "COCO_train2014_000000297995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man putting on a neck tie\"."}], "task": "refering", "answer_template": "The \"a man putting on a neck tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 333, 334], "bbox": [0.387921875, 0.2308470588235294, 0.711421875, 0.9887529411764706], "iscrowd": 0, "area": 32020.48755000001, "rle": {"size": [425, 640], "counts": "\\SW34U=6I3M4L2O1N1O2N101N1O101N1O2N101N1O2N101N1N2M3010O0000O2O0O100O2O000O2O0O100O2O0O100O2O0O100O2O01O10N100O101O0O1O1N3QKRN_NP2[1VNbNm1X1YNfNj1T1[NjNg17fMoKi0g3d12kMQLg0k3`13kMkKk0P4]13iMhKo0R4[15YOHi06YOGj08WOFl08UOFm08UOEn0:SODP10kL`LV2^3Q11lL^LT2`3Q10[OOf0YOVMXMV2_3e0[Nn0d1SOcMf1\\2\\NVMQ2i2PNUMR2k2oMQMT2n2mMoLV2P3mMkLV2U3o3O100O01N1N3N1O2M201O0010O0101N1O2O0O2N1N3M2N3M2N3M2N2O2M2O2N1N3N1O2M2O2M2O2N2N3M2N2N2N2O1N3M2N2O1N2N5K4M4K5K4M3L2O1N2O1N2O1N100O100O100O10O100001O001O00001O001O00001O000XJnHY5S7aJSI_5[7ZO[HhKf7V4]HhKd7W4^HgKb7X4bHfK^7Y4dHeK]7Y4fHeK[7Z4gHeKX7Z4lHcKU7\\4mHbKT7\\4oHcKQ7\\4RIaKn6^4P1L4oNQ1H7M4L4L3L5L4L4L3L5CmD_NW;_1;N101K5H8H7K6L4M2M4L4L4Mmb\\2"}, "label": "a man putting on a neck tie"}]}
{"id": 297995, "image": "COCO_train2014_000000297995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a standing man putting on a tie\"."}], "task": "refering", "answer_template": "The \"a standing man putting on a tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 333, 334], "bbox": [0.387921875, 0.2308470588235294, 0.711421875, 0.9887529411764706], "iscrowd": 0, "area": 32020.48755000001, "rle": {"size": [425, 640], "counts": "\\SW34U=6I3M4L2O1N1O2N101N1O101N1O2N101N1O2N101N1N2M3010O0000O2O0O100O2O000O2O0O100O2O0O100O2O0O100O2O01O10N100O101O0O1O1N3QKRN_NP2[1VNbNm1X1YNfNj1T1[NjNg17fMoKi0g3d12kMQLg0k3`13kMkKk0P4]13iMhKo0R4[15YOHi06YOGj08WOFl08UOFm08UOEn0:SODP10kL`LV2^3Q11lL^LT2`3Q10[OOf0YOVMXMV2_3e0[Nn0d1SOcMf1\\2\\NVMQ2i2PNUMR2k2oMQMT2n2mMoLV2P3mMkLV2U3o3O100O01N1N3N1O2M201O0010O0101N1O2O0O2N1N3M2N3M2N3M2N2O2M2O2N1N3N1O2M2O2M2O2N2N3M2N2N2N2O1N3M2N2O1N2N5K4M4K5K4M3L2O1N2O1N2O1N100O100O100O10O100001O001O00001O001O00001O000XJnHY5S7aJSI_5[7ZO[HhKf7V4]HhKd7W4^HgKb7X4bHfK^7Y4dHeK]7Y4fHeK[7Z4gHeKX7Z4lHcKU7\\4mHbKT7\\4oHcKQ7\\4RIaKn6^4P1L4oNQ1H7M4L4L3L5L4L4L3L5CmD_NW;_1;N101K5H8H7K6L4M2M4L4L4Mmb\\2"}, "label": "a standing man putting on a tie"}]}
{"id": 117772, "image": "COCO_train2014_000000117772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark haired man holding a small child\"."}], "task": "refering", "answer_template": "The \"a dark haired man holding a small child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 343, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 410, 411, 412, 427, 428, 429, 430, 431, 432, 433, 434, 435, 452, 453, 454, 455, 456, 457, 474, 475, 476, 477, 478, 479, 480, 495, 496, 497, 498, 499, 500, 501, 502, 503, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526], "bbox": [0.396390625, 0.03715625, 0.957203125, 0.9964999999999999], "iscrowd": 0, "area": 107443.74699999999, "rle": {"size": [640, 640], "counts": "Qkn45ic05K5K6J5K5L5J5K5K6h^ObNd?c1V@cNe?d1T@aNg?h1o_O^Nl?k1j_O[NQ`0n2K5K4M4L4M3L4L4L2N1O2O1N2N2N2N2O1N2N2N1O2OO01O001O001O00001O1O4L3M4L4L4L4L4L4L4L4L4L4L4cCmHg:0eEW7]OnHm:KeE1TOV74TIP;FgEOVOj7R;WHgEOXOl7n:VHiENZOm7k:VHPFl7n9THRFm7l9THTFm7j9THVFn7g9SHYFn7f9RHZFP8c9QH]FP8a9PH`FR8^9nGaFT8]9mGcFT8[9mGeFU8X9lGhFU8W9kGiFW8T9jGlFW8R9jGnFX8P9hGPGY8n8hGRGY8l8hGTGZ8i8gGWGZ8g8gGYG[8d8fG\\G[8c8eG]G]8`8dG`G]8^8dGbG]8\\8dGdG^8Y8cGgG^8W8cGiG_8U8aGkG`8S8aGmGa8P8`GPHa8n7`GRHa8l7`GTHb8i7_GWHe8d7\\G\\Hh8`7XGaHl8Y7UGgHP9S7QGmHT9m6mFSIW9h6jFXI[9b6fF^I^9^6bFbIc9X6^FhIg9R6ZFnIi9n5XFRJj9k5WFUJk9h5VFYJj9e5WF[Ji9e5WF[Ji9d5XF\\Jh9c5YF]Jg9b5ZF^Jf9h1PEo0[1YMe9g1]Ec0n0gMe9e1jE7b0TNd9d1VFM6_Nd9c1WFM6`Nd9a1WFN6aNc9`1WFO7bNa9_1XFN8cN`9^1XF08bN`9]1YFfMnNl1Z1QO_9\\1hFCIQO_9[1iFDHQO_9Z1jFDHSO]9Y1kFDHSO]9X1lFDHUO[9V1nFEGVOZ9U1PGCGXOY9T1QGDFYOX9S1RGCG[OW9Q1RGDG[OW9P1SGDG]OU9o0TGDG^OT9m0WGCF@S9m0WGCFAR9k0YGDEBQ9j0ZGCFCP9i0[GDEDo8h0\\GCFFm8f0^GDEFm8f0_GBEIk8d0aGBEKi8c0bGAFLh8c0bG@GNf8a0dGAFOf8?dGAG2c8<hG@F6`8:jG_OG8_87kG@G;]84lG@H=[82nG_OIa0X8OoG^OKd0V8LQH^OJh0T8IRH]OLl0P8GUHZOMP1n7DVHZONT1k7AXHXOOX1h7_OZHXON\\1g7[O\\HVOO`1e7XO]HVO0d1a7VO`HSO1h1_7SOaHTO1k1]7POcHRO2o1Z7nNeHQO3S2W7kNgHoN4W2`3bMTMV1YOoN5[2U3hM\\Mn0[OiN9b2h2nMeMf0a2^1ASNoM=a2a1YOXNVN7a2c1SO[N[N2b2d1nN^N`NMc2g1gN`NfNIc2h1bNcNkNEc2j1\\NfNPO@d2k1WNiNUO\\Od2m1QNkN[OWOe2o1kMnN@SOe2Z2\\MgNOoNe2j2[KTLJV2W2lNd2Y3VJgL?X1g2hNd2T5VJXLV3dNd2P8\\MoGe2Q8[MoGe2Q8ZMPHf2P8ZMPHf2o7[MQHe2o7ZMQHg2o7YMQHg2o7YMQHg2o7XMRHh2n7XMRHh2m7XMSHi2m7WMSHi2m7WMSHi2m7VMTHj2l7VMTHj2l7VMSHk2l7UMUHk2k7UMUHk2k7TMVHl2j7TMVHk2k7UMTHl2l7SMUHm2j7TMVHl2j7TMVHl2j7SMWHm2i7SMWHm2i7RMWHo2i7QMWHo2i7QMWHo2h7QMYHo2g7QMYHo2g7QMXHP3h7oLYHQ3g7oLYHQ3g7nLZHR3e7oL[HQ3e7oLZHR3f7nLZHR3f7nLZHR3f7nLZHR3f7nLZHR3f7nLYHS3f7nLZHR3f7nLZHR3f7nLZHR3f7nLZHR3f7nLYHS3g7lLZHT3e7mL[HS3e7mL[HS3e7mL[HS3e7mLZHT3f7lLZHT3f7lLZHT3e7mL[HS3e7mL[HS3e7mL[HS3e7mLZHT3f7lLZHT3e7mL[HS3e7mL[HS3e7lL\\HT3c7mL]HS3c7mL]HS3b7nL^HQ3c7oL]HQ3c7oL]HQ3b7PM^HP3b7PM^HP3b7PM^HP3a7QM_Ho2a7QM_Ho2`7RM`Hn2`7RM`Hn2`7RM`Hn2_7RMbHn2^7RMbHn2^7RMbHn2]7SMcHm2]7SMcHm2\\7TMdHl2\\7TMdHl2\\7TMdHl2[7UMeHk2[7UMeHk2[7UMeHk2Z7VMfHj2Z7VMfHj2Y7WMgHi2Y7VMhHj2X7VMhHj2W7WMiHi2W7WMiHi2W7WMiHi2V7XMjHh2V7XMjHh2V7XMjHh2U7YMkHg2U7YMkHg2U7YMkHg2T7ZMlHf2T7ZMlHf2T7ZMlHf2T7ZMlHf2S7[MmHe2S7[MmHe2S7[MmHe2R7\\MnHd2R7\\MnHd2R7\\MnHc2R7^MnHb2R7^MnHb2R7^MnHb2Q7_MoHa2Q7_MoHa2Q7_MoHa2Q7_MoHa2Q7_MPI`2P7`MPI`2P7`MQI_2o6ZJfG`2[1V3o6ZJfG`2\\1U3n6[JfG`2\\1U3n6ZJgGa2\\1T3m6[JgGb2\\1R3m6\\JgGb2\\1R3m6[JhGc2\\1Q3l6\\JhGc2\\1Q3l6[JiGd2\\1P3k6\\JiGd2]1o2j6]JiGd2]1o2a7QM_Hn2b7RM]Hi2i7WMWHc2S8YMlGa2k8oLUGk2S9SMlFh2[9WMeFb2c9]M]F]2i9cMVFX2Q:gMoER2Y:nMeEm1a:SN_Ef1h:YNXEb1o:]NQE]1U;bNkDX1\\;hNdDR1c;lN]Do0i;QOWDh0P<ZOnC`0Y<@eC;a<G\\C3k<OQCMV=4gBG_=;]B@j=b0SBYOT>h0hATO^>n0_AlNh>V1TAfNT?[1h@`N`?`1]@ZNl?f28L4M3M3M3L4M3M3M3L4M3Md0\\Ol1TNn^`0"}, "label": "a dark haired man holding a small child"}]}
{"id": 117772, "image": "COCO_train2014_000000117772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man carring his baby\"."}], "task": "refering", "answer_template": "The \"a man carring his baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 343, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 410, 411, 412, 427, 428, 429, 430, 431, 432, 433, 434, 435, 452, 453, 454, 455, 456, 457, 474, 475, 476, 477, 478, 479, 480, 495, 496, 497, 498, 499, 500, 501, 502, 503, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526], "bbox": [0.396390625, 0.03715625, 0.957203125, 0.9964999999999999], "iscrowd": 0, "area": 107443.74699999999, "rle": {"size": [640, 640], "counts": "Qkn45ic05K5K6J5K5L5J5K5K6h^ObNd?c1V@cNe?d1T@aNg?h1o_O^Nl?k1j_O[NQ`0n2K5K4M4L4M3L4L4L2N1O2O1N2N2N2N2O1N2N2N1O2OO01O001O001O00001O1O4L3M4L4L4L4L4L4L4L4L4L4L4cCmHg:0eEW7]OnHm:KeE1TOV74TIP;FgEOVOj7R;WHgEOXOl7n:VHiENZOm7k:VHPFl7n9THRFm7l9THTFm7j9THVFn7g9SHYFn7f9RHZFP8c9QH]FP8a9PH`FR8^9nGaFT8]9mGcFT8[9mGeFU8X9lGhFU8W9kGiFW8T9jGlFW8R9jGnFX8P9hGPGY8n8hGRGY8l8hGTGZ8i8gGWGZ8g8gGYG[8d8fG\\G[8c8eG]G]8`8dG`G]8^8dGbG]8\\8dGdG^8Y8cGgG^8W8cGiG_8U8aGkG`8S8aGmGa8P8`GPHa8n7`GRHa8l7`GTHb8i7_GWHe8d7\\G\\Hh8`7XGaHl8Y7UGgHP9S7QGmHT9m6mFSIW9h6jFXI[9b6fF^I^9^6bFbIc9X6^FhIg9R6ZFnIi9n5XFRJj9k5WFUJk9h5VFYJj9e5WF[Ji9e5WF[Ji9d5XF\\Jh9c5YF]Jg9b5ZF^Jf9h1PEo0[1YMe9g1]Ec0n0gMe9e1jE7b0TNd9d1VFM6_Nd9c1WFM6`Nd9a1WFN6aNc9`1WFO7bNa9_1XFN8cN`9^1XF08bN`9]1YFfMnNl1Z1QO_9\\1hFCIQO_9[1iFDHQO_9Z1jFDHSO]9Y1kFDHSO]9X1lFDHUO[9V1nFEGVOZ9U1PGCGXOY9T1QGDFYOX9S1RGCG[OW9Q1RGDG[OW9P1SGDG]OU9o0TGDG^OT9m0WGCF@S9m0WGCFAR9k0YGDEBQ9j0ZGCFCP9i0[GDEDo8h0\\GCFFm8f0^GDEFm8f0_GBEIk8d0aGBEKi8c0bGAFLh8c0bG@GNf8a0dGAFOf8?dGAG2c8<hG@F6`8:jG_OG8_87kG@G;]84lG@H=[82nG_OIa0X8OoG^OKd0V8LQH^OJh0T8IRH]OLl0P8GUHZOMP1n7DVHZONT1k7AXHXOOX1h7_OZHXON\\1g7[O\\HVOO`1e7XO]HVO0d1a7VO`HSO1h1_7SOaHTO1k1]7POcHRO2o1Z7nNeHQO3S2W7kNgHoN4W2`3bMTMV1YOoN5[2U3hM\\Mn0[OiN9b2h2nMeMf0a2^1ASNoM=a2a1YOXNVN7a2c1SO[N[N2b2d1nN^N`NMc2g1gN`NfNIc2h1bNcNkNEc2j1\\NfNPO@d2k1WNiNUO\\Od2m1QNkN[OWOe2o1kMnN@SOe2Z2\\MgNOoNe2j2[KTLJV2W2lNd2Y3VJgL?X1g2hNd2T5VJXLV3dNd2P8\\MoGe2Q8[MoGe2Q8ZMPHf2P8ZMPHf2o7[MQHe2o7ZMQHg2o7YMQHg2o7YMQHg2o7XMRHh2n7XMRHh2m7XMSHi2m7WMSHi2m7WMSHi2m7VMTHj2l7VMTHj2l7VMSHk2l7UMUHk2k7UMUHk2k7TMVHl2j7TMVHk2k7UMTHl2l7SMUHm2j7TMVHl2j7TMVHl2j7SMWHm2i7SMWHm2i7RMWHo2i7QMWHo2i7QMWHo2h7QMYHo2g7QMYHo2g7QMXHP3h7oLYHQ3g7oLYHQ3g7nLZHR3e7oL[HQ3e7oLZHR3f7nLZHR3f7nLZHR3f7nLZHR3f7nLZHR3f7nLYHS3f7nLZHR3f7nLZHR3f7nLZHR3f7nLZHR3f7nLYHS3g7lLZHT3e7mL[HS3e7mL[HS3e7mL[HS3e7mLZHT3f7lLZHT3f7lLZHT3e7mL[HS3e7mL[HS3e7mL[HS3e7mLZHT3f7lLZHT3e7mL[HS3e7mL[HS3e7lL\\HT3c7mL]HS3c7mL]HS3b7nL^HQ3c7oL]HQ3c7oL]HQ3b7PM^HP3b7PM^HP3b7PM^HP3a7QM_Ho2a7QM_Ho2`7RM`Hn2`7RM`Hn2`7RM`Hn2_7RMbHn2^7RMbHn2^7RMbHn2]7SMcHm2]7SMcHm2\\7TMdHl2\\7TMdHl2\\7TMdHl2[7UMeHk2[7UMeHk2[7UMeHk2Z7VMfHj2Z7VMfHj2Y7WMgHi2Y7VMhHj2X7VMhHj2W7WMiHi2W7WMiHi2W7WMiHi2V7XMjHh2V7XMjHh2V7XMjHh2U7YMkHg2U7YMkHg2U7YMkHg2T7ZMlHf2T7ZMlHf2T7ZMlHf2T7ZMlHf2S7[MmHe2S7[MmHe2S7[MmHe2R7\\MnHd2R7\\MnHd2R7\\MnHc2R7^MnHb2R7^MnHb2R7^MnHb2Q7_MoHa2Q7_MoHa2Q7_MoHa2Q7_MoHa2Q7_MPI`2P7`MPI`2P7`MQI_2o6ZJfG`2[1V3o6ZJfG`2\\1U3n6[JfG`2\\1U3n6ZJgGa2\\1T3m6[JgGb2\\1R3m6\\JgGb2\\1R3m6[JhGc2\\1Q3l6\\JhGc2\\1Q3l6[JiGd2\\1P3k6\\JiGd2]1o2j6]JiGd2]1o2a7QM_Hn2b7RM]Hi2i7WMWHc2S8YMlGa2k8oLUGk2S9SMlFh2[9WMeFb2c9]M]F]2i9cMVFX2Q:gMoER2Y:nMeEm1a:SN_Ef1h:YNXEb1o:]NQE]1U;bNkDX1\\;hNdDR1c;lN]Do0i;QOWDh0P<ZOnC`0Y<@eC;a<G\\C3k<OQCMV=4gBG_=;]B@j=b0SBYOT>h0hATO^>n0_AlNh>V1TAfNT?[1h@`N`?`1]@ZNl?f28L4M3M3M3L4M3M3M3L4M3Md0\\Ol1TNn^`0"}, "label": "a man carring his baby"}]}
{"id": 117772, "image": "COCO_train2014_000000117772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby with the dress on is being held\"."}], "task": "refering", "answer_template": "The \"the baby with the dress on is being held\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 311, 312, 316, 317, 318, 319, 320, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518], "bbox": [0.062921875, 0.4, 0.939328125, 0.986515625], "iscrowd": 0, "area": 100508.39084999994, "rle": {"size": [640, 640], "counts": "eWj02lc02O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2O1N2cNPOh_OR1o?YOm_Oi0j?BS@?f?JV@8h?MS@5k?OQ@3m?1P@0n?4n_ONo?7m_OKQ`09k_OIS`09l_OHR`0:l_OHR`0:l_OHQ`0U2N2N2N2N2N2M3N2N2N2N2M3O100O100O100O100O1O100O100O10O0100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1N2O1dM]JZFd5e9^JYFc5f9`JWFa5g9bJWF_5h9cJVF^5i9dJUF]5i9fJUF[5j9gJTFZ5j9jJSFW5l9kJRFV5m9lJPFV5n9mJPFT5o9nJoES5P:oJnER5P:RKmEo4R:SKlEn4Q:VKmEk4Q:XKmEi4Q:ZKmEg4Q:]KlEd4Q:`KmEa4Q:bKmE^4R:eKkE]4S:fKkE[4S:hKkEY4R:lKkEU4S:nKkES4S:PLkEQ4S:RLkEo3R:ULlEl3R:WLlEj3R:ZLkEg3S:T3M3N2N2N2M3N2N2N2N2M3N2N2N2M3O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O10O101O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O00001O00001O00001O00001N1000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O2N1N3M3N1N3M3M2O2M3M2N3N2M2N3M3N1N3M3M2O2M3M2N3N2M2N3M3N1N5K9G9H8G9G9kKTCe0Q=VORCh0n<XOTCe0m<ZOVCc0k<]OXC?j<_OYC?g<A[C<f<C]C:d<F^C7d<H^C6b<IaC4`<LbC1_<NdCO]<1eCM\\<1hCKY<5iCHX<7kCFV<:lCDU<:mCDT<<mCBT<=mCBT<>lCBT<=nCAT<>lCAU<>lCAU<?kCAU<>mC@U<?kC@V<?kC@V<`0jC@V<?lC_OU<a0kC^OW<`0jC_OW<a0iC_OW<`0kC^OV<b0jC]OX<a0iC^OX<b0iC]OW<b0jC]OW<c0iC\\OY<b0hC]OY<c0hC\\OY<b0hC]OZ<b0fC]O[<b0fC]O\\<b0eC\\O\\<c0eC]O\\<b0dC]O]<b0dC]O^<b0cCWOd<g0]CTOh<l0XCnNo<Q1RCiNS=V1nBeNX=Z1hB`N_=^1bB]Nc=c1^2N101N2O0O2O001N10O1000O1000O10O100000O01000000O0100000O10O1000O1000O10O100000O0100000O0100000O0100000O010000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O01000000O101O000O10001O0O10001O0001O0001O000000001O0000001O0001O0001O000000001O0000001O01O000001O000000001O0000001O01O000001OO1O1O1O1O1O1O1O1O2N1O4L3M3M3M4L3M3M4L3M3M4L3MVog0"}, "label": "the baby with the dress on is being held"}]}
{"id": 117772, "image": "COCO_train2014_000000117772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young toddler dressed in white , in the arms of an adult male\"."}], "task": "refering", "answer_template": "The \"a young toddler dressed in white , in the arms of an adult male\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 311, 312, 316, 317, 318, 319, 320, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518], "bbox": [0.062921875, 0.4, 0.939328125, 0.986515625], "iscrowd": 0, "area": 100508.39084999994, "rle": {"size": [640, 640], "counts": "eWj02lc02O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2O1N2cNPOh_OR1o?YOm_Oi0j?BS@?f?JV@8h?MS@5k?OQ@3m?1P@0n?4n_ONo?7m_OKQ`09k_OIS`09l_OHR`0:l_OHR`0:l_OHQ`0U2N2N2N2N2N2M3N2N2N2N2M3O100O100O100O100O1O100O100O10O0100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1N2O1dM]JZFd5e9^JYFc5f9`JWFa5g9bJWF_5h9cJVF^5i9dJUF]5i9fJUF[5j9gJTFZ5j9jJSFW5l9kJRFV5m9lJPFV5n9mJPFT5o9nJoES5P:oJnER5P:RKmEo4R:SKlEn4Q:VKmEk4Q:XKmEi4Q:ZKmEg4Q:]KlEd4Q:`KmEa4Q:bKmE^4R:eKkE]4S:fKkE[4S:hKkEY4R:lKkEU4S:nKkES4S:PLkEQ4S:RLkEo3R:ULlEl3R:WLlEj3R:ZLkEg3S:T3M3N2N2N2M3N2N2N2N2M3N2N2N2M3O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O10O101O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O00001O00001O00001O00001N1000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O2N1N3M3N1N3M3M2O2M3M2N3N2M2N3M3N1N3M3M2O2M3M2N3N2M2N3M3N1N5K9G9H8G9G9kKTCe0Q=VORCh0n<XOTCe0m<ZOVCc0k<]OXC?j<_OYC?g<A[C<f<C]C:d<F^C7d<H^C6b<IaC4`<LbC1_<NdCO]<1eCM\\<1hCKY<5iCHX<7kCFV<:lCDU<:mCDT<<mCBT<=mCBT<>lCBT<=nCAT<>lCAU<>lCAU<?kCAU<>mC@U<?kC@V<?kC@V<`0jC@V<?lC_OU<a0kC^OW<`0jC_OW<a0iC_OW<`0kC^OV<b0jC]OX<a0iC^OX<b0iC]OW<b0jC]OW<c0iC\\OY<b0hC]OY<c0hC\\OY<b0hC]OZ<b0fC]O[<b0fC]O\\<b0eC\\O\\<c0eC]O\\<b0dC]O]<b0dC]O^<b0cCWOd<g0]CTOh<l0XCnNo<Q1RCiNS=V1nBeNX=Z1hB`N_=^1bB]Nc=c1^2N101N2O0O2O001N10O1000O1000O10O100000O01000000O0100000O10O1000O1000O10O100000O0100000O0100000O0100000O010000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O01000000O101O000O10001O0O10001O0001O0001O000000001O0000001O0001O0001O000000001O0000001O01O000001O000000001O0000001O01O000001OO1O1O1O1O1O1O1O1O2N1O4L3M3M3M4L3M3M4L3M3M4L3MVog0"}, "label": "a young toddler dressed in white , in the arms of an adult male"}]}
{"id": 470028, "image": "COCO_train2014_000000470028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with long blonde hair walks onto a red bus\"."}], "task": "refering", "answer_template": "The \"a woman with long blonde hair walks onto a red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 290, 313], "bbox": [0.49651562499999996, 0.12918032786885245, 0.692421875, 0.9084543325526931], "iscrowd": 0, "area": 21830.773300000008, "rle": {"size": [427, 640], "counts": "`hT4b0e<8H7I6J6J5K4J7N1O2N1O01N100O100000O10O1M3N2N3N1000UFiNd7W1[HmNa7T1^HPO^7Q1aHTO[7l0dHXOY7h0fH\\OZ6]O^HW1W1AR6D^Hm0k0ZNBY1n6b1]IYMBY1i6g1`IRMD]1b6U2[IaL0^1`6h2\\I^M`6k4J6J5L5K4M1N10001N1000000O10001O1O1N3hJ\\He4e7YK]Hf4e7WK\\Hi4V8N1N1O2N1O1O1O1O010O10O1N2O2N1D<O1O2O1O1O10SLVHW2T8YMRHg2`8bLdG_3S92N2O1N2N1O2N2N1O2M3\\OhKQHZ4k7kKRHX4h7oKSHT4i7QLRHS4i7k0_JbJo2b5oL_Jm2f5QM[Jb1M_Mm5k0YJc1N]Mm5n0VJc12XMm5S1SJb17QMl5[1nIc1W7\\NkH_1Z7_NhH\\1\\7cNfHX1_7fNcHU1b7hN`HS1f7kN[HQ1j7mNXHn0m7POUHh0S8UOQH\\OHFa8l0jGiN<NS8X1d2N2M7J5K6J5J4M2N3M2MXoa2"}, "label": "a woman with long blonde hair walks onto a red bus"}]}
{"id": 470028, "image": "COCO_train2014_000000470028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hippie girl in a halter getting on a bus\"."}], "task": "refering", "answer_template": "The \"a hippie girl in a halter getting on a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 290, 313], "bbox": [0.49651562499999996, 0.12918032786885245, 0.692421875, 0.9084543325526931], "iscrowd": 0, "area": 21830.773300000008, "rle": {"size": [427, 640], "counts": "`hT4b0e<8H7I6J6J5K4J7N1O2N1O01N100O100000O10O1M3N2N3N1000UFiNd7W1[HmNa7T1^HPO^7Q1aHTO[7l0dHXOY7h0fH\\OZ6]O^HW1W1AR6D^Hm0k0ZNBY1n6b1]IYMBY1i6g1`IRMD]1b6U2[IaL0^1`6h2\\I^M`6k4J6J5L5K4M1N10001N1000000O10001O1O1N3hJ\\He4e7YK]Hf4e7WK\\Hi4V8N1N1O2N1O1O1O1O010O10O1N2O2N1D<O1O2O1O1O10SLVHW2T8YMRHg2`8bLdG_3S92N2O1N2N1O2N2N1O2M3\\OhKQHZ4k7kKRHX4h7oKSHT4i7QLRHS4i7k0_JbJo2b5oL_Jm2f5QM[Jb1M_Mm5k0YJc1N]Mm5n0VJc12XMm5S1SJb17QMl5[1nIc1W7\\NkH_1Z7_NhH\\1\\7cNfHX1_7fNcHU1b7hN`HS1f7kN[HQ1j7mNXHn0m7POUHh0S8UOQH\\OHFa8l0jGiN<NS8X1d2N2M7J5K6J5J4M2N3M2MXoa2"}, "label": "a hippie girl in a halter getting on a bus"}]}
{"id": 158737, "image": "COCO_train2014_000000158737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in yellow\"."}], "task": "refering", "answer_template": "The \"the man in yellow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 98, 114, 115, 116, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 209, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 294, 295, 296, 297, 299, 313, 314, 315, 332, 333, 348, 349, 350, 351, 366, 367], "bbox": [0.18366795366795366, 0.1869375, 0.845945945945946, 0.876125], "iscrowd": 0, "area": 48054.85025000005, "rle": {"size": [640, 518], "counts": "[Wl12nc01O1O2N1O2N1O1O1O001O1O1O1O0O2O1N2N2N1O2N>B2O1N2N2O1O1O1O1O1O0000O100O10000O103MO001O1O001N101O1O1O001O1N2O001O1O1O1N101O1O1N110O01O0010O01O00100O00Y^OcNm`0\\1k^OnNSa0P2Y_OUMY`0l2f_OUMX`0l2h_OUMW`0l2h_OUMW`0k2h_OWMW`0j2h_OWMV`0j2j_OWMU`0j2j_OWMQ?KgAn2VOZMn>8_AR3]>VM^Ak2^>\\M^Af2]>`M_Ac2[>eMaA]2[>h1K4M3L4WIQJ\\OS6?oI_OU6=mIBV6:kIEZ6eJcKY2TNQ3]6_JfK]2mMS3b6YJgK`2hMW3d6SJjKc2bMZ3h6lImKf2[M]3l6gIoKi2UM`3W:]LiEb3[:ZLfEf3]:WLcEh3_:WL`Ej3`:VL`Ej3`:ULaEj3`:VL`Ej3`:VL`Ei3a:WL_Ei3a:VL`Ej3`:VL`Ei3a:WL_Ei3a:WL_Ei3a:WL_Ej3`:UL`El3`:TL`Em3_:SLaEn3^:RLbEo3]:PLdEP4\\:oKeER4Z:mKgET4X:kKiEU4W:jKjEW4U:gI_Ee1=e4S:bIiEc14l4S:^IRF_1LT5n:eJSE\\5Q;^JPEc5R;YJoDg5T;UJmDl5U;PJlDQ6W;jIjDW6?\\I]99UF\\6:`Ib90UFa65eIf9FVFe61iIk9^OTFQ7GeIQ;f7O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O0PNYGfHi8l6RHfHo7l6`HRIa7i6eHUI\\7j6fHTI[7k6gHSIZ7l6hHRIY7m6iHQIW7n6kHQIV7n6lHPIU7o6mHoHT7P7nHnHS7Q7oHmHS7j6VIRIo6b6\\IYIj6[6aIaIe6R6l3G9M3N2N2N=C8G10000000001N10000000001N100000001N1000000O101N100O10000O2O0O10001N6K4K;F6J1N2O1N2O0O2O1O1N2N1O2M3N2M3N5J8H9H3L2O1N2O1N3N1N2OO0100O1O1O1O1O1O100O1O1O001O100O1O1O1O001O10O01O010O10O0100O0010O0100O010O10O010O01O10O010O010O010N1J7M20001O1O1O101O01O0001O1O2N101O010O010O01O000O2O00001O00000O2N1O1O1O1O2N1K6J:E<E[ca1"}, "label": "the man in yellow"}]}
{"id": 158737, "image": "COCO_train2014_000000158737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with spectacles playing with a ring\"."}], "task": "refering", "answer_template": "The \"a man with spectacles playing with a ring\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 98, 114, 115, 116, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 209, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 294, 295, 296, 297, 299, 313, 314, 315, 332, 333, 348, 349, 350, 351, 366, 367], "bbox": [0.18366795366795366, 0.1869375, 0.845945945945946, 0.876125], "iscrowd": 0, "area": 48054.85025000005, "rle": {"size": [640, 518], "counts": "[Wl12nc01O1O2N1O2N1O1O1O001O1O1O1O0O2O1N2N2N1O2N>B2O1N2N2O1O1O1O1O1O0000O100O10000O103MO001O1O001N101O1O1O001O1N2O001O1O1O1N101O1O1N110O01O0010O01O00100O00Y^OcNm`0\\1k^OnNSa0P2Y_OUMY`0l2f_OUMX`0l2h_OUMW`0l2h_OUMW`0k2h_OWMW`0j2h_OWMV`0j2j_OWMU`0j2j_OWMQ?KgAn2VOZMn>8_AR3]>VM^Ak2^>\\M^Af2]>`M_Ac2[>eMaA]2[>h1K4M3L4WIQJ\\OS6?oI_OU6=mIBV6:kIEZ6eJcKY2TNQ3]6_JfK]2mMS3b6YJgK`2hMW3d6SJjKc2bMZ3h6lImKf2[M]3l6gIoKi2UM`3W:]LiEb3[:ZLfEf3]:WLcEh3_:WL`Ej3`:VL`Ej3`:ULaEj3`:VL`Ej3`:VL`Ei3a:WL_Ei3a:VL`Ej3`:VL`Ei3a:WL_Ei3a:WL_Ei3a:WL_Ej3`:UL`El3`:TL`Em3_:SLaEn3^:RLbEo3]:PLdEP4\\:oKeER4Z:mKgET4X:kKiEU4W:jKjEW4U:gI_Ee1=e4S:bIiEc14l4S:^IRF_1LT5n:eJSE\\5Q;^JPEc5R;YJoDg5T;UJmDl5U;PJlDQ6W;jIjDW6?\\I]99UF\\6:`Ib90UFa65eIf9FVFe61iIk9^OTFQ7GeIQ;f7O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O0PNYGfHi8l6RHfHo7l6`HRIa7i6eHUI\\7j6fHTI[7k6gHSIZ7l6hHRIY7m6iHQIW7n6kHQIV7n6lHPIU7o6mHoHT7P7nHnHS7Q7oHmHS7j6VIRIo6b6\\IYIj6[6aIaIe6R6l3G9M3N2N2N=C8G10000000001N10000000001N100000001N1000000O101N100O10000O2O0O10001N6K4K;F6J1N2O1N2O0O2O1O1N2N1O2M3N2M3N5J8H9H3L2O1N2O1N3N1N2OO0100O1O1O1O1O1O100O1O1O001O100O1O1O1O001O10O01O010O10O0100O0010O0100O010O10O010O01O10O010O010O010N1J7M20001O1O1O101O01O0001O1O2N101O010O010O01O000O2O00001O00000O2N1O1O1O1O2N1K6J:E<E[ca1"}, "label": "a man with spectacles playing with a ring"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bench with a man on it\"."}], "task": "refering", "answer_template": "The \"a white bench with a man on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 176, 190, 195, 196, 197, 198, 213, 218, 219, 220, 221, 222, 223, 236], "bbox": [0.272765625, 0.30845833333333333, 0.733421875, 0.6270416666666667], "iscrowd": 0, "area": 9801.232800000002, "rle": {"size": [480, 640], "counts": "Tgb2;f>>A?A9GUObB7\\=BkB?j=3M2O1100O2O0O1]O]Qk1KonTN9O0VB@V=`0jBAj:0jF`0\\NBh:1iF=^NDh:3eF:cNEf:3eF8eNFe:3dF8gNGc:3dF6iNHb:3cF6kNGb:4bF5lNFc:6_F5nNEc:8]F3oNFd:8[F3QOEd:9ZF2ROEd:;WF1UODd:<UF1WOCd:=TF0XOBe:?QF0ZOAe:a0nEO\\OAf:c0hE1@\\Oh:f0bE7@SOm:i0^E=_OjNS;l0XEl0h:VOSEl0n:WOlDl0T;`1000000000O1000000000000O1`LkDU3U;kLQFo1o9QNVGj0j8VO`3000[A5U>KkAh08NQ<ZOgCP3Z;PMfDP3Z;PMfDP3Z;PMfDo2[;QMeDo2[;QMeDo2[;PMfDP3Z;PMfDP3Z;PMfDP3Z;PMfDo2[;QMeDo2[;QMeDo2[;QMeDo2[;QMfDn2Z;RMfDm2[;SMeDm2[;SMeDm2[;SMeDm2[;SMeDm2[;SMeDm2[;SMeDl2\\;TMdDl2\\;TMdDl2\\;TMdDl2\\;TMdDl2\\;TMdDl2[;UMeDj2\\;VMdDj2\\;VMW30000000000X3hL1O1O001O1O1O1O001O0000WOhDjMX;U2kDiMU;V2nDhMR;W2QEgMo:X2TEfMS;R2PElM[;g1iDWNa;^1bD`Nh;U1ZDiNR<k0oCUO[<`0fC@d<5]CKn<ISC7k=0000000000000000000001O001O11N101N101N101O0O2O0O2O1N01O1O0O2O001O1N101O001N101O1N101O001N101OYl_2"}, "label": "a white bench with a man on it"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bench between the man and woman\"."}], "task": "refering", "answer_template": "The \"the white bench between the man and woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 176, 190, 195, 196, 197, 198, 213, 218, 219, 220, 221, 222, 223, 236], "bbox": [0.272765625, 0.30845833333333333, 0.733421875, 0.6270416666666667], "iscrowd": 0, "area": 9801.232800000002, "rle": {"size": [480, 640], "counts": "Tgb2;f>>A?A9GUObB7\\=BkB?j=3M2O1100O2O0O1]O]Qk1KonTN9O0VB@V=`0jBAj:0jF`0\\NBh:1iF=^NDh:3eF:cNEf:3eF8eNFe:3dF8gNGc:3dF6iNHb:3cF6kNGb:4bF5lNFc:6_F5nNEc:8]F3oNFd:8[F3QOEd:9ZF2ROEd:;WF1UODd:<UF1WOCd:=TF0XOBe:?QF0ZOAe:a0nEO\\OAf:c0hE1@\\Oh:f0bE7@SOm:i0^E=_OjNS;l0XEl0h:VOSEl0n:WOlDl0T;`1000000000O1000000000000O1`LkDU3U;kLQFo1o9QNVGj0j8VO`3000[A5U>KkAh08NQ<ZOgCP3Z;PMfDP3Z;PMfDP3Z;PMfDo2[;QMeDo2[;QMeDo2[;PMfDP3Z;PMfDP3Z;PMfDP3Z;PMfDo2[;QMeDo2[;QMeDo2[;QMeDo2[;QMfDn2Z;RMfDm2[;SMeDm2[;SMeDm2[;SMeDm2[;SMeDm2[;SMeDm2[;SMeDl2\\;TMdDl2\\;TMdDl2\\;TMdDl2\\;TMdDl2\\;TMdDl2[;UMeDj2\\;VMdDj2\\;VMW30000000000X3hL1O1O001O1O1O1O001O0000WOhDjMX;U2kDiMU;V2nDhMR;W2QEgMo:X2TEfMS;R2PElM[;g1iDWNa;^1bD`Nh;U1ZDiNR<k0oCUO[<`0fC@d<5]CKn<ISC7k=0000000000000000000001O001O11N101N101N101O0O2O0O2O1N01O1O0O2O001O1N101O001N101O1N101O001N101OYl_2"}, "label": "the white bench between the man and woman"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the jacket\"."}], "task": "refering", "answer_template": "The \"the man in the jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 283, 284, 286, 287, 306, 309, 310, 332, 333], "bbox": [0.23771874999999998, 0.11570833333333333, 0.536234375, 0.8577291666666668], "iscrowd": 0, "area": 36675.75589999999, "rle": {"size": [480, 640], "counts": "Q^W22l>5K5J4M3M2N3M2N2N3L3N3M2N3M4L4L4L3L5Lj0WO3L3M3N3L3N2M4L3N3L5K5L5J8H7J5J6K5JT1lN<E3L4M2N3N2N1O2WIRI[5Q7]JYI_5h6ZJbIb5Q8L10O1O10O01O10OO2O1N2N101N101N101N1O101N10L3J7F9G:K4I72M3N2N2M3N2N2M3N2N3L3N2M3N2N2M4N100O101O0O10000O2O000O1000001O00000001O000001O0001O0000XI^IR5b6lJaIS5_6jJdIV5\\6gJhIY5X6cJlI\\5U6aJnI^5U6XJSJg5c70100O1O1O03N3M4L4L4K?Ba0_O=C5K5K5K00O1O1O1O1O1O001O1O1O1O1O2N1XM\\IRMe6i2cISM_6g2jITMW6g2QJUMQ6d2XJYMh5b2aJYMa5a2gJ[MZ5h0ZIkNc1:U5b0fIiN[1c0P5a0mIbNZ1j0k4a0RJ[NY1R1f4a0YJRNX1X1b4c0UMYOo2d0TMXOo2f0UMUOn2i0UMQOQ3k0TMjNU3T1nLbN[3\\1W5M2N2N3M2M3N3M1O2N2N2N2M2O2L4K5LTQ[4"}, "label": "the man in the jacket"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a jacket on his laptop\"."}], "task": "refering", "answer_template": "The \"a man in a jacket on his laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 283, 284, 286, 287, 306, 309, 310, 332, 333], "bbox": [0.23771874999999998, 0.11570833333333333, 0.536234375, 0.8577291666666668], "iscrowd": 0, "area": 36675.75589999999, "rle": {"size": [480, 640], "counts": "Q^W22l>5K5J4M3M2N3M2N2N3L3N3M2N3M4L4L4L3L5Lj0WO3L3M3N3L3N2M4L3N3L5K5L5J8H7J5J6K5JT1lN<E3L4M2N3N2N1O2WIRI[5Q7]JYI_5h6ZJbIb5Q8L10O1O10O01O10OO2O1N2N101N101N101N1O101N10L3J7F9G:K4I72M3N2N2M3N2N2M3N2N3L3N2M3N2N2M4N100O101O0O10000O2O000O1000001O00000001O000001O0001O0000XI^IR5b6lJaIS5_6jJdIV5\\6gJhIY5X6cJlI\\5U6aJnI^5U6XJSJg5c70100O1O1O03N3M4L4L4K?Ba0_O=C5K5K5K00O1O1O1O1O1O001O1O1O1O1O2N1XM\\IRMe6i2cISM_6g2jITMW6g2QJUMQ6d2XJYMh5b2aJYMa5a2gJ[MZ5h0ZIkNc1:U5b0fIiN[1c0P5a0mIbNZ1j0k4a0RJ[NY1R1f4a0YJRNX1X1b4c0UMYOo2d0TMXOo2f0UMUOn2i0UMQOQ3k0TMjNU3T1nLbN[3\\1W5M2N2N3M2M3N3M1O2N2N2N2M2O2L4K5LTQ[4"}, "label": "a man in a jacket on his laptop"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red luggage bag\"."}], "task": "refering", "answer_template": "The \"a red luggage bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.006734374999999999, 0.7415833333333333, 0.262921875, 0.9887708333333334], "iscrowd": 0, "area": 16343.740900000003, "rle": {"size": [480, 640], "counts": "Zg2:c>8G9H8G9H8G9J6K5K5K6J5J6K5K5K00000000000O100000000000000O010000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000001O1O2N1O2N4L6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J[il6"}, "label": "a red luggage bag"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"redluggage bottom left\"."}], "task": "refering", "answer_template": "The \"redluggage bottom left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.006734374999999999, 0.7415833333333333, 0.262921875, 0.9887708333333334], "iscrowd": 0, "area": 16343.740900000003, "rle": {"size": [480, 640], "counts": "Zg2:c>8G9H8G9H8G9J6K5K5K6J5J6K5K5K00000000000O100000000000000O010000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000001O1O2N1O2N4L6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J[il6"}, "label": "redluggage bottom left"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bench where a woman is sitting\"."}], "task": "refering", "answer_template": "The \"a white bench where a woman is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 134, 135, 136, 137, 153, 154, 158, 159, 160, 176, 177, 182, 183, 199, 200, 205, 206, 222, 223, 224, 225, 228, 229, 246, 247, 248, 251, 274, 275], "bbox": [0.656734375, 0.2929375, 1.0, 0.697375], "iscrowd": 0, "area": 11229.437500000004, "rle": {"size": [480, 640], "counts": "[RU62n>3YA5e0H_<f0`CZO`<m0XCUOf<m0XCTOh<n0VCSOR;MSFQ1iNSOT;2lEn0mNROV;6gEi0ROQOW;=_Ed0XOQOX;a0ZE`0[OPO[;f0SE<AoN[;l0mDf0R;@hD`0X;e1O001O00SNjD1U;LPE2P;LTE3k:LYE1g:N[E2d:M]E3c:L_E4`:KaE5_:LaE4^:LbE4^:LbE5]:KdE4\\:LdE5[:KfEXO@Kj:n0eEVOCLg:n0gEUOCMe:n0hETOEMc:o0iESODOb:n0jEROF0_:n0kEQOH0]:o0lEPOH1[:P1lEnNK2X:P1nE`N7`0k9P1SGPOl8o0VGPO\\NFi9U1QHUOSNJj9m0WHZOkMNl9d0^H]OdM3l9>bH@^M4Q::dH6^7HeH5]7IeH5]7IfH4\\7KeH3]7KfH2\\7LfH2\\7LgH2Z7MfH2\\7LeH3]7KdH4^7JcH5_7JbH4`7JaH7_7GbH9_7EbH<^7CbH>^7@cH`0n:10003N2M3N2M2O2M3N2M2O2M100O010ON3K4K6K4L5K4L5K4LVSc0NPm\\O5K5K5K5K5K4L5K5K3M00O10O10O100O100O100O1001O2N1O1O2N1O1O1O2N1O00001cDmNk8S1UGROf8n0ZGUOc8k0]GXO`8i0_GZO]8g0cG\\OZ8d0eG_OY8a0gGBV8?iGDT8=kGFR8<lGGQ8;mGHP8:nG_N\\Nj0d9i0nG]N`Nk0_9j0oG[NeNk0Y9l0oGZNjNk0T9m0PHXNoNk0n8n0RHWNROk0j8n0THWNUOk0d8n0WHWNWOl0_8m0ZHWNZOl0Y8l0^HXN[Oj0W8n0UHaNG>T8Q1eGQO9LR8S1bGTO>GP8U1aGUOb0Cm7X1`GVOe0@k7Z1^GXOj0[On7W1WG_Om0XOU8P1mFIQ1TO[8i0dF4S1QOb8c0XF>Y1dNP9e0fEh0[<oNdCR1f<dNYC]1n<]NPCd1R=42N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2NTH"}, "label": "a white bench where a woman is sitting"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench with a woman sitting on it\"."}], "task": "refering", "answer_template": "The \"the bench with a woman sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 134, 135, 136, 137, 153, 154, 158, 159, 160, 176, 177, 182, 183, 199, 200, 205, 206, 222, 223, 224, 225, 228, 229, 246, 247, 248, 251, 274, 275], "bbox": [0.656734375, 0.2929375, 1.0, 0.697375], "iscrowd": 0, "area": 11229.437500000004, "rle": {"size": [480, 640], "counts": "[RU62n>3YA5e0H_<f0`CZO`<m0XCUOf<m0XCTOh<n0VCSOR;MSFQ1iNSOT;2lEn0mNROV;6gEi0ROQOW;=_Ed0XOQOX;a0ZE`0[OPO[;f0SE<AoN[;l0mDf0R;@hD`0X;e1O001O00SNjD1U;LPE2P;LTE3k:LYE1g:N[E2d:M]E3c:L_E4`:KaE5_:LaE4^:LbE4^:LbE5]:KdE4\\:LdE5[:KfEXO@Kj:n0eEVOCLg:n0gEUOCMe:n0hETOEMc:o0iESODOb:n0jEROF0_:n0kEQOH0]:o0lEPOH1[:P1lEnNK2X:P1nE`N7`0k9P1SGPOl8o0VGPO\\NFi9U1QHUOSNJj9m0WHZOkMNl9d0^H]OdM3l9>bH@^M4Q::dH6^7HeH5]7IeH5]7IfH4\\7KeH3]7KfH2\\7LfH2\\7LgH2Z7MfH2\\7LeH3]7KdH4^7JcH5_7JbH4`7JaH7_7GbH9_7EbH<^7CbH>^7@cH`0n:10003N2M3N2M2O2M3N2M2O2M100O010ON3K4K6K4L5K4L5K4LVSc0NPm\\O5K5K5K5K5K4L5K5K3M00O10O10O100O100O100O1001O2N1O1O2N1O1O1O2N1O00001cDmNk8S1UGROf8n0ZGUOc8k0]GXO`8i0_GZO]8g0cG\\OZ8d0eG_OY8a0gGBV8?iGDT8=kGFR8<lGGQ8;mGHP8:nG_N\\Nj0d9i0nG]N`Nk0_9j0oG[NeNk0Y9l0oGZNjNk0T9m0PHXNoNk0n8n0RHWNROk0j8n0THWNUOk0d8n0WHWNWOl0_8m0ZHWNZOl0Y8l0^HXN[Oj0W8n0UHaNG>T8Q1eGQO9LR8S1bGTO>GP8U1aGUOb0Cm7X1`GVOe0@k7Z1^GXOj0[On7W1WG_Om0XOU8P1mFIQ1TO[8i0dF4S1QOb8c0XF>Y1dNP9e0fEh0[<oNdCR1f<dNYC]1n<]NPCd1R=42N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2NTH"}, "label": "the bench with a woman sitting on it"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the laptop\"."}], "task": "refering", "answer_template": "The \"the woman on the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 109, 110, 111, 112, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 248, 249, 250, 270, 271, 272, 273, 293, 294, 296], "bbox": [0.688734375, 0.211625, 0.91521875, 0.7512291666666666], "iscrowd": 0, "area": 17539.756900000004, "rle": {"size": [480, 640], "counts": "Xm^66d>8I6O2M3N1N3N100iBSOY<n0dCUO[<l0cCVO]<i0aC[O]<f0aC\\O^<e0_C^Oa<b0\\CBb<?\\CCc<>ZCEf<:YCHf<9WCIi<8TCKl<U1N1N3M3L3N3M3M2N3L3N3M3M2N3M3N1OO0010O01O0012M2O1N3N100O2N1O1fF`LS7b3jHdLQ7]3lHkLn6V3PIPMk6R3RITMi6m2TIWMj6j2TIYMj6i2RI[Ml6f2RI\\Mm6e2PI_Mn6c2nHaMP7`2nHdMo6^2mHgMP7U1THQNj0n0o6R1XHPNe0Q1R7P1ZHnMb0P1X7R1WHnM=o0_7T1UHlM:n0`7[1WHgM5m0c7a1YHaM2l0d7h1\\H[MLl0j7k1[HXMHn0n7k1[HUOe7l0[HTOd7l0]HSOb7n0_HQOa7o0`HPO`7P1bHnN]7S1dHlN\\7T1eHkN[7U1fHeN^7\\1cH^Nb7b1_HYNd7h1^HQNg7o1ZHkMk7U2VHfMn7Z2RHdMP8\\2PHbMQ8`2nG]MU8c2kG[MW8e2hGZMZ8f2fGXM\\8h2dGVM^8j2bGTM_8m2aGQMa8n2`GPMb8P3^GnLd8Q3]GmLe8R3\\GlLe8T3\\GjLf8V3ZGgLi8X3XGfLj8Y3WGeLk8[3UGbLn8]3SG`LP9_3QG^LS9a3lFaLS9^3nFcLZ1H]5c3ZIfLV1J`5_3ZIhLR1Ld5[3[IjLn0Ng5V3\\ImLi00k5S3\\InLf02n5`0]HZ1P1UN?6U69_HZ1m0XN8<[60dHY1j0\\N0a0c6HfHY1g0_NIf0j6@jHX1d0i0b6lMVIQ18S1c6jM`Ih0N_1a6fMnIa0Am1^6_M[J<WO\\2X6TMiJ:oNg2T6lLVK7fNl2V6jL\\K3_NS3U6gLmLX3^7O1J6H8L4L4L4L4M3L4L4L4L3L5J6J6J7I;E<DoQi0"}, "label": "the woman on the laptop"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman wearing a black leather jacket is sitting on a white bench while using her laptop computer\"."}], "task": "refering", "answer_template": "The \"a blonde woman wearing a black leather jacket is sitting on a white bench while using her laptop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 109, 110, 111, 112, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 248, 249, 250, 270, 271, 272, 273, 293, 294, 296], "bbox": [0.688734375, 0.211625, 0.91521875, 0.7512291666666666], "iscrowd": 0, "area": 17539.756900000004, "rle": {"size": [480, 640], "counts": "Xm^66d>8I6O2M3N1N3N100iBSOY<n0dCUO[<l0cCVO]<i0aC[O]<f0aC\\O^<e0_C^Oa<b0\\CBb<?\\CCc<>ZCEf<:YCHf<9WCIi<8TCKl<U1N1N3M3L3N3M3M2N3L3N3M3M2N3M3N1OO0010O01O0012M2O1N3N100O2N1O1fF`LS7b3jHdLQ7]3lHkLn6V3PIPMk6R3RITMi6m2TIWMj6j2TIYMj6i2RI[Ml6f2RI\\Mm6e2PI_Mn6c2nHaMP7`2nHdMo6^2mHgMP7U1THQNj0n0o6R1XHPNe0Q1R7P1ZHnMb0P1X7R1WHnM=o0_7T1UHlM:n0`7[1WHgM5m0c7a1YHaM2l0d7h1\\H[MLl0j7k1[HXMHn0n7k1[HUOe7l0[HTOd7l0]HSOb7n0_HQOa7o0`HPO`7P1bHnN]7S1dHlN\\7T1eHkN[7U1fHeN^7\\1cH^Nb7b1_HYNd7h1^HQNg7o1ZHkMk7U2VHfMn7Z2RHdMP8\\2PHbMQ8`2nG]MU8c2kG[MW8e2hGZMZ8f2fGXM\\8h2dGVM^8j2bGTM_8m2aGQMa8n2`GPMb8P3^GnLd8Q3]GmLe8R3\\GlLe8T3\\GjLf8V3ZGgLi8X3XGfLj8Y3WGeLk8[3UGbLn8]3SG`LP9_3QG^LS9a3lFaLS9^3nFcLZ1H]5c3ZIfLV1J`5_3ZIhLR1Ld5[3[IjLn0Ng5V3\\ImLi00k5S3\\InLf02n5`0]HZ1P1UN?6U69_HZ1m0XN8<[60dHY1j0\\N0a0c6HfHY1g0_NIf0j6@jHX1d0i0b6lMVIQ18S1c6jM`Ih0N_1a6fMnIa0Am1^6_M[J<WO\\2X6TMiJ:oNg2T6lLVK7fNl2V6jL\\K3_NS3U6gLmLX3^7O1J6H8L4L4L4L4M3L4L4L4L3L5J6J6J7I;E<DoQi0"}, "label": "a blonde woman wearing a black leather jacket is sitting on a white bench while using her laptop computer"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of black luggage sitting on a man ' s lap\"."}], "task": "refering", "answer_template": "The \"a piece of black luggage sitting on a man ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 284], "bbox": [0.28967187499999997, 0.42737499999999995, 0.5292656250000001, 0.718], "iscrowd": 0, "area": 15471.079600000005, "rle": {"size": [480, 640], "counts": "onf27c>:F`0@T1QO`0@2N2O1N2N3N1N2O1N2O1N2O1O1N2O1N2O1O000001O01O0001O00101N1O0000O101ZOe0O10000O2O000000001O0000000000000000O0100000O10000O10000O10000O1000000O1000O010000O10000O10000O1000000O10000O100O010O1O100O1O100O100O1O100O1O100O100O1O10O01O1O1O100O2N1jLmD^2T<F101N1O1N3M2N2N2O8H0lNgBb0Z=^OlB:V=FPC3Q=MUCLl<4o001000000000O010000000O10O106I7JQk\\4"}, "label": "a piece of black luggage sitting on a man ' s lap"}]}
{"id": 445462, "image": "COCO_train2014_000000445462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop bag on the lap of a man in a red coat\"."}], "task": "refering", "answer_template": "The \"a laptop bag on the lap of a man in a red coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 284], "bbox": [0.28967187499999997, 0.42737499999999995, 0.5292656250000001, 0.718], "iscrowd": 0, "area": 15471.079600000005, "rle": {"size": [480, 640], "counts": "onf27c>:F`0@T1QO`0@2N2O1N2N3N1N2O1N2O1N2O1O1N2O1N2O1O000001O01O0001O00101N1O0000O101ZOe0O10000O2O000000001O0000000000000000O0100000O10000O10000O10000O1000000O1000O010000O10000O10000O1000000O10000O100O010O1O100O1O100O100O1O100O1O100O100O1O10O01O1O1O100O2N1jLmD^2T<F101N1O1N3M2N2N2O8H0lNgBb0Z=^OlB:V=FPC3Q=MUCLl<4o001000000000O010000000O10O106I7JQk\\4"}, "label": "a laptop bag on the lap of a man in a red coat"}]}
{"id": 576543, "image": "COCO_train2014_000000576543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train 4522\"."}], "task": "refering", "answer_template": "The \"train 4522\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 269, 270, 271, 272], "bbox": [0.67640625, 0.4, 0.8606093749999999, 0.7573023255813953], "iscrowd": 0, "area": 15409.442250000002, "rle": {"size": [430, 640], "counts": "kSf54S=9[MFlGa0V8D[Ge0e8AmFf0`8QOcF>>j0m8nNbF<3o0X9lNbF:HQ1d9lNaFQ2\\9VNaFh1]9W1N1N3M3M1O100O1O1O1O100O1O1O1O100O1O100O1000000O1000000O100000000O1000000O1000000O1000000O100000000O100000000000O100000000000000000000O10001O1O001O1O1O001N2O1O001O1O1O001O1O1O0O2O1O1O001O1O1O2M9F9G:E=D;E;D=D;E<D;D=DkRU1"}, "label": "train 4522"}]}
{"id": 576543, "image": "COCO_train2014_000000576543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train who ' s front end is showing in the picture\"."}], "task": "refering", "answer_template": "The \"the train who ' s front end is showing in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 269, 270, 271, 272], "bbox": [0.67640625, 0.4, 0.8606093749999999, 0.7573023255813953], "iscrowd": 0, "area": 15409.442250000002, "rle": {"size": [430, 640], "counts": "kSf54S=9[MFlGa0V8D[Ge0e8AmFf0`8QOcF>>j0m8nNbF<3o0X9lNbF:HQ1d9lNaFQ2\\9VNaFh1]9W1N1N3M3M1O100O1O1O1O100O1O1O1O100O1O100O1000000O1000000O100000000O1000000O1000000O1000000O100000000O100000000000O100000000000000000000O10001O1O001O1O1O001N2O1O001O1O1O001O1O1O0O2O1O1O001O1O1O2M9F9G:E=D;E;D=D;E<D;D=DkRU1"}, "label": "the train who ' s front end is showing in the picture"}]}
{"id": 576543, "image": "COCO_train2014_000000576543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large brown train engine pulling a shorter car with a tarp over the top\"."}], "task": "refering", "answer_template": "The \"a large brown train engine pulling a shorter car with a tarp over the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 285, 299, 300, 301, 304, 322, 323], "bbox": [0.0, 0.24183720930232558, 0.635796875, 0.959953488372093], "iscrowd": 0, "area": 91778.0978, "rle": {"size": [430, 640], "counts": "b3c1d;o7XHO1000001O0O100000001N100000001N1000000O2O0000000O2O0000000O2O00000O101O00000O101O0O10000O2O000O10000O2O000O10001N10000O1K6aN`KhIf4X6\\KaIi4`6XKYIm4g6UKRIP5n6i0O00000000001O000000001O0000000100O001O1O1O10O01O1O1O010O00O10001O0O10001O1N101O1O1N2O1O1O0O100000000000000000000000001O000O101OEZInIf6R6[ImIe6R6\\InId6l5bIUJ]6e5j0I7N2O1O001O1O1O1O6J5K3M3M2N00000000000000000000000000O10O01O1O1O1O100O1O1O1O1N2O1O1N2O1O1O1N2O100O1O1O1O1O100O1O1O1O100000000000000000000000000000000000000000000000000000000O1000000000000O100000000O100O100O10000O100O101N1010O2O0O100O2O0O100O101N01N101O0O2O0O2O001N2O0O2O001N101N101O0O2O0O2O001N101O0O2O001N2O001N101O0O2O001N101O0O2O0O2O001N101O0O2O001N2O001N101O001N2O001O1O001N101O1O001O1N101O1O001O1N101O1O001O1N101ON2N2O10O100000000O101O00000O100000000O3N2N1O1N2O001O1O0O2O001O1O0O2O1O001N2O0O2O1N101N1O2O5J2O0O100N3M2O1O2`MVG9m8WOaGi0b:VOXDMi;2]DHe;7_DDb;<h0O1D\\cQ3"}, "label": "a large brown train engine pulling a shorter car with a tarp over the top"}]}
{"id": 134176, "image": "COCO_train2014_000000134176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small pink bike on the dirt\"."}], "task": "refering", "answer_template": "The \"a small pink bike on the dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 144, 145, 146, 162, 163, 181, 182, 198, 199, 200, 201, 217, 218], "bbox": [0.0027400000000000002, 0.5582666666666667, 0.1749, 0.9945866666666667], "iscrowd": 0, "area": 4888.319900000002, "rle": {"size": [375, 500], "counts": "To04a;3M2001O0000010O00^GJc56[JLf53XJOh51VJ1j50SJ2n5MQJ4o5LPJ5P6KPJ5P6KoI6Q6JoI6Q6KnI5Q6LnI4S6LkI5U6LiI5W6LdG0c14j6`0QIAo6c0lH^OU7i0RHTO\\O4b8P1dGF]8j1O0010O0010O010N1OaMaGi1_8VNeGh1Y8YNjGd1U8\\NoGb1P8\\NUHa15SNg6<XI_1^OkNV7E`Ib2\\6^MWJR2f5nM_Jn1a5QNdJl1Z5UNjJg1V5XNoJe1o4[NVKa1j4^NYKa1e4`N]K^1c4aN`K]1`4cNbK[1]4]NTI0b2a1Z4_NTI0d2_1W4aNVI0c2_1W4`NWI0a2b1W4^NYI0_2c1X4\\NZI1\\2e1Z4YN[I1Z2h1Z4WN]I1X2i1[4VN]I1V2k1\\4UN^IOV2m1\\4TN^IOU2n1]4SN^IOU2n1\\4TN_INU2n1\\4TN_IOS2n1^4SN_IOS2n1]4TN`IOR2n1]4SNaIOR2n1\\4TNbINQ2a0lMc0a6nNbIOP2`0QN?]6ROcINo1b0UN9X6XOeIMo1`0YN6S6]OgIKo1`0\\N3n5BhIKP2=\\N4l5DjIIP2=]N3k5EiIKQ2:]N4k5EYL8mM1k5FXL9PNNj5GVL;RNLk84ZGGh87b1Jgmf4"}, "label": "a small pink bike on the dirt"}]}
{"id": 134176, "image": "COCO_train2014_000000134176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl ' s pink and orange bike\"."}], "task": "refering", "answer_template": "The \"little girl ' s pink and orange bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 144, 145, 146, 162, 163, 181, 182, 198, 199, 200, 201, 217, 218], "bbox": [0.0027400000000000002, 0.5582666666666667, 0.1749, 0.9945866666666667], "iscrowd": 0, "area": 4888.319900000002, "rle": {"size": [375, 500], "counts": "To04a;3M2001O0000010O00^GJc56[JLf53XJOh51VJ1j50SJ2n5MQJ4o5LPJ5P6KPJ5P6KoI6Q6JoI6Q6KnI5Q6LnI4S6LkI5U6LiI5W6LdG0c14j6`0QIAo6c0lH^OU7i0RHTO\\O4b8P1dGF]8j1O0010O0010O010N1OaMaGi1_8VNeGh1Y8YNjGd1U8\\NoGb1P8\\NUHa15SNg6<XI_1^OkNV7E`Ib2\\6^MWJR2f5nM_Jn1a5QNdJl1Z5UNjJg1V5XNoJe1o4[NVKa1j4^NYKa1e4`N]K^1c4aN`K]1`4cNbK[1]4]NTI0b2a1Z4_NTI0d2_1W4aNVI0c2_1W4`NWI0a2b1W4^NYI0_2c1X4\\NZI1\\2e1Z4YN[I1Z2h1Z4WN]I1X2i1[4VN]I1V2k1\\4UN^IOV2m1\\4TN^IOU2n1]4SN^IOU2n1\\4TN_INU2n1\\4TN_IOS2n1^4SN_IOS2n1]4TN`IOR2n1]4SNaIOR2n1\\4TNbINQ2a0lMc0a6nNbIOP2`0QN?]6ROcINo1b0UN9X6XOeIMo1`0YN6S6]OgIKo1`0\\N3n5BhIKP2=\\N4l5DjIIP2=]N3k5EiIKQ2:]N4k5EYL8mM1k5FXL9PNNj5GVL;RNLk84ZGGh87b1Jgmf4"}, "label": "little girl ' s pink and orange bike"}]}
{"id": 134176, "image": "COCO_train2014_000000134176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike holding a tabletop and cooler\"."}], "task": "refering", "answer_template": "The \"a bike holding a tabletop and cooler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 58, 59, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 130, 131, 132, 148, 149, 150], "bbox": [0.19128, 0.20664, 0.40953999999999996, 0.6976533333333333], "iscrowd": 0, "area": 12910.519800000002, "rle": {"size": [375, 500], "counts": "_ZS1<Z;1O1N3M2N3M2N3M2N3M2N3M3M3M3N2N3M2N3M2N3M=POhMiGZ2Q8mMkGW2n7PNnGS2l7UNoGn1l7R1L4L3M4L3N3M2O2N1O1O1N3N1O1O1O02ON2K5I601O00001O0O2O1O1O1O1O2O0O100O01N1O101N1O2N1O101N1O1O1OO2O0OK`IVK_6j45101O02N2O1N3N1N3M2O1N3N1N3M2O1N3N1O1O2\\L[HR3g7nLZHo2i7oLYHn2i7QMYHl2i7TMXHh2l7VMVHg2l7XMVHe2m7YMTHe2o7ZM[HX2j7fMiHf0h9I6J:Gdl[3"}, "label": "a bike holding a tabletop and cooler"}]}
{"id": 134176, "image": "COCO_train2014_000000134176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red cooler on the back of a bike\"."}], "task": "refering", "answer_template": "The \"a red cooler on the back of a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 58, 59, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 130, 131, 132, 148, 149, 150], "bbox": [0.19128, 0.20664, 0.40953999999999996, 0.6976533333333333], "iscrowd": 0, "area": 12910.519800000002, "rle": {"size": [375, 500], "counts": "_ZS1<Z;1O1N3M2N3M2N3M2N3M2N3M3M3M3N2N3M2N3M2N3M=POhMiGZ2Q8mMkGW2n7PNnGS2l7UNoGn1l7R1L4L3M4L3N3M2O2N1O1O1N3N1O1O1O02ON2K5I601O00001O0O2O1O1O1O1O2O0O100O01N1O101N1O2N1O101N1O1O1OO2O0OK`IVK_6j45101O02N2O1N3N1N3M2O1N3N1N3M2O1N3N1O1O2\\L[HR3g7nLZHo2i7oLYHn2i7QMYHl2i7TMXHh2l7VMVHg2l7XMVHe2m7YMTHe2o7ZM[HX2j7fMiHf0h9I6J:Gdl[3"}, "label": "a red cooler on the back of a bike"}]}
{"id": 85028, "image": "COCO_train2014_000000085028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man throwing the ball from the picther ' s mound\"."}], "task": "refering", "answer_template": "The \"the man throwing the ball from the picther ' s mound\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 128, 147, 148, 149, 150, 151, 171, 172, 173, 174, 195, 196, 197, 198, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288], "bbox": [0.38915625, 0.3425586854460094, 0.61303125, 0.8526995305164319], "iscrowd": 0, "area": 12622.734100000003, "rle": {"size": [426, 640], "counts": "agW36S=3N2M2O2N1N2O1O1O2M4M3M3M4K3N1O100O100O01O010O010O10O10O0100O10O01000O1O1O1O1O001O10O01O1O01000O10O010000O01O1O010O1O001\\HlNP3U1mLmNT3S1jLoNU3Q1`HPOQ41_3d1SL^Nl3h1lK[NS4j1gKWNZ4n1^KUNa4Q2WKRNh4W2nJjMR5Y2iJjMV5X2gJiMY5Y2gIoLHi0a6Z2cIRMHe0e6[2`ISMHc0h6[2]IWMG?m6Z2YI\\MF;Q7Z2WI^MA=X7W2UIeNk6]1RIdNo6]1nHdNR7j3001O001O00001L3M3N3L3M4L3M3N3L3N2L4J6C=F:001O1N1N3N1N3M2O2M3M201N101O1O1N2O_NcGVN\\8j1eGVNY8k1hGUNW8k1jGTNW8k1kGSNV8l1kGSNU8m1lGRNU8m1lGRNU8m1lGQNV8n1lGPNT8o1oGoMR8o1QHoMP8n1UHnMl7P2[HkMf7S2`HhMa7U2fHeM[7Y2m1N2M2N3^NaE6a:EeE7^:CiE9Z:AmE<oYW3"}, "label": "the man throwing the ball from the picther ' s mound"}]}
{"id": 85028, "image": "COCO_train2014_000000085028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player pitching the ball\"."}], "task": "refering", "answer_template": "The \"the player pitching the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 128, 147, 148, 149, 150, 151, 171, 172, 173, 174, 195, 196, 197, 198, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288], "bbox": [0.38915625, 0.3425586854460094, 0.61303125, 0.8526995305164319], "iscrowd": 0, "area": 12622.734100000003, "rle": {"size": [426, 640], "counts": "agW36S=3N2M2O2N1N2O1O1O2M4M3M3M4K3N1O100O100O01O010O010O10O10O0100O10O01000O1O1O1O1O001O10O01O1O01000O10O010000O01O1O010O1O001\\HlNP3U1mLmNT3S1jLoNU3Q1`HPOQ41_3d1SL^Nl3h1lK[NS4j1gKWNZ4n1^KUNa4Q2WKRNh4W2nJjMR5Y2iJjMV5X2gJiMY5Y2gIoLHi0a6Z2cIRMHe0e6[2`ISMHc0h6[2]IWMG?m6Z2YI\\MF;Q7Z2WI^MA=X7W2UIeNk6]1RIdNo6]1nHdNR7j3001O001O00001L3M3N3L3M4L3M3N3L3N2L4J6C=F:001O1N1N3N1N3M2O2M3M201N101O1O1N2O_NcGVN\\8j1eGVNY8k1hGUNW8k1jGTNW8k1kGSNV8l1kGSNU8m1lGRNU8m1lGRNU8m1lGQNV8n1lGPNT8o1oGoMR8o1QHoMP8n1UHnMl7P2[HkMf7S2`HhMa7U2fHeM[7Y2m1N2M2N3^NaE6a:EeE7^:CiE9Z:AmE<oYW3"}, "label": "the player pitching the ball"}]}
{"id": 85028, "image": "COCO_train2014_000000085028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player running the bases directly behind the pitcher\"."}], "task": "refering", "answer_template": "The \"the player running the bases directly behind the pitcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 31, 32, 33, 54, 55, 56, 77, 78, 79, 100, 101, 102, 103, 122, 123, 125, 126], "bbox": [0.31246874999999996, 0.04854460093896713, 0.515015625, 0.4056338028169014], "iscrowd": 0, "area": 7348.7422, "rle": {"size": [426, 640], "counts": "h[c21Y=0O1O1O2O0O1O1O100O2N1O100O1O1O2O00000M3M3M4L3M2M4M3M2N3ZNVOiFm0m8@mFc0R9AgFb0Y9BaF?`9EZF<h9GRF:o9_11N2O1O2M2O1O1N3N1O1O1N2O2N1N2O1OM202N1O2N2000O2O1O001O010O10O010O010O011O1N3N1N3N1d0]ON101O000O2O001N101O0O2O001O001O00001O00VN\\En0c:PO`Eo0`:oNbER1]:lNfES1Z:kNjET1W:hNQFR1S:gNVFT1l9gN\\FS1S;K010N2O1O1O001O1N2O1O001O1O1N101O1N2M3M2O2M3M3MTWQ4"}, "label": "the player running the bases directly behind the pitcher"}]}
{"id": 199721, "image": "COCO_train2014_000000199721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a boy in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 154, 155, 156, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 245, 246, 247, 248, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 382, 383, 384, 385, 386], "bbox": [0.589890625, 0.17977083333333335, 0.8427031249999999, 0.9887708333333334], "iscrowd": 0, "area": 37324.73880000001, "rle": {"size": [480, 640], "counts": "_aa5c0Z>5L1O2M3O0010O10O0100O010O01000O010O10@UOjBk0V=XOhBh0W=[OfBe0Z=_OcBa0\\=BbB=^=c0O1O001O001\\EhN^5HgI`1=WOj5\\OgI^11DV6POhI\\1E2a6eNhIZ1YO?l6ZNjIW1iNQ1[7jMkI]4S6fKkI[4S6hKkIY4R6SLdIn3Z6]L\\Ic3c6gLTIZ3j6QMlHP3Q7\\MeHe2Y7b2N2N2`L]HAe7>dHWO_7g0kHoNW7P1QIgNQ7X1QIeNQ7Z1QIbNR7]1QI_NQ7`1QI]NQ7_1TI]No6b1SI[No6d1TIXNn6g1TIVNn6j1SIRNP7m1SIPNn6P2SIoMm6P2VInMj6Q2YIlMh6T2YIkMg6T2\\IjMd6U2_IiMa6W2aIgM_6X2dIeM]6[2dIdM\\6\\2fIbMZ6]2iI`MX6`2jI\\MX6d2iIYMY6f2jIVMX6j2jIRMX6n2m2O10000O100O10000O10000O100001O000000000000001O0000000000001O000oFRMP6n2PJRMP6n2oISMQ6m2oITMP6l2oIUMQ6l2mIUMS6k2mIVMR6k2lIVMT6k2kIUMU6k2jIWMU6j2iIWMW6j2hIWMW6j2gIWMY6j2eIWM[6i2dIYM[6h2dIXM\\6i2bIXM^6i2`IYM_6g2aIYM_6h2_IYMa6h2]IZM`6i2_IWM`6k2^IWM_6k2`IVM_6l2`IUMY6R3eIQMb2WOaNZ4kNbL]2^OTN`4^OTLW2ESN[4ESLP2LSNV4KQLk13QNQ43nKW1h0^N_39lKd0\\1kN[4QO]J]1P2ZO`5>kJYO[5e0iJSO]5k0gJmN_5P1[4N2N2N2N2N3M2N2N2M3N2N2N2N2N3M2NRS_1"}, "label": "a boy in a blue shirt"}]}
{"id": 199721, "image": "COCO_train2014_000000199721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy in a blue indians shirt holding a wii controller\"."}], "task": "refering", "answer_template": "The \"young boy in a blue indians shirt holding a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 154, 155, 156, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 245, 246, 247, 248, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 382, 383, 384, 385, 386], "bbox": [0.589890625, 0.17977083333333335, 0.8427031249999999, 0.9887708333333334], "iscrowd": 0, "area": 37324.73880000001, "rle": {"size": [480, 640], "counts": "_aa5c0Z>5L1O2M3O0010O10O0100O010O01000O010O10@UOjBk0V=XOhBh0W=[OfBe0Z=_OcBa0\\=BbB=^=c0O1O001O001\\EhN^5HgI`1=WOj5\\OgI^11DV6POhI\\1E2a6eNhIZ1YO?l6ZNjIW1iNQ1[7jMkI]4S6fKkI[4S6hKkIY4R6SLdIn3Z6]L\\Ic3c6gLTIZ3j6QMlHP3Q7\\MeHe2Y7b2N2N2`L]HAe7>dHWO_7g0kHoNW7P1QIgNQ7X1QIeNQ7Z1QIbNR7]1QI_NQ7`1QI]NQ7_1TI]No6b1SI[No6d1TIXNn6g1TIVNn6j1SIRNP7m1SIPNn6P2SIoMm6P2VInMj6Q2YIlMh6T2YIkMg6T2\\IjMd6U2_IiMa6W2aIgM_6X2dIeM]6[2dIdM\\6\\2fIbMZ6]2iI`MX6`2jI\\MX6d2iIYMY6f2jIVMX6j2jIRMX6n2m2O10000O100O10000O10000O100001O000000000000001O0000000000001O000oFRMP6n2PJRMP6n2oISMQ6m2oITMP6l2oIUMQ6l2mIUMS6k2mIVMR6k2lIVMT6k2kIUMU6k2jIWMU6j2iIWMW6j2hIWMW6j2gIWMY6j2eIWM[6i2dIYM[6h2dIXM\\6i2bIXM^6i2`IYM_6g2aIYM_6h2_IYMa6h2]IZM`6i2_IWM`6k2^IWM_6k2`IVM_6l2`IUMY6R3eIQMb2WOaNZ4kNbL]2^OTN`4^OTLW2ESN[4ESLP2LSNV4KQLk13QNQ43nKW1h0^N_39lKd0\\1kN[4QO]J]1P2ZO`5>kJYO[5e0iJSO]5k0gJmN_5P1[4N2N2N2N2N3M2N2N2M3N2N2N2N2N3M2NRS_1"}, "label": "young boy in a blue indians shirt holding a wii controller"}]}
{"id": 199721, "image": "COCO_train2014_000000199721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with beige shirt on facing towards televsion and away from the camera\"."}], "task": "refering", "answer_template": "The \"man with beige shirt on facing towards televsion and away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 28, 29, 30, 31, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375], "bbox": [0.0, 0.0015625, 0.37415625, 0.9835833333333334], "iscrowd": 0, "area": 73914.08435, "rle": {"size": [480, 640], "counts": "]66^>=F:M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N3M2M3N2M3N2N2M3N2M3N2N2lFRMR6Q3cIZM[6h2ZIbMd6a2QIjMm6X2hHSNV7o1^H\\N`7g1UHdNi7^1lGlNR8W1cGTO[8i3N3M2O2M3M2O2M2N3N1N3M7J6I7I7J6I7J7I6J5K4M3L4L4L3M4L4M3L4L4L4TLVF[2W;0100O1O1O1O100O1O1O1O1O100O1O1O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000iNbMlE^2U:iMcEV2_:RNXEn1j:YNmDg1T;aNcDL3a0\\;>hD^O[;?jD]OX;?mD^OU;?oD^OS;?QE]OS;?QE^OR;=TE_Oo:>TE_Oo:?SE]OQ;a0QE\\OR;b0QEZOR;d0PEYOS;e0f1N3M2N2N2N2N1O000000000000000000001O0000000000000000000000000000000000000000001O000000nCIg1i0]2^OjKLm1d1V1`NlLMR2V2;mMbMOW2`2FaMQN2\\2k2QOSMbN4a2U3\\NgLRO6f2`3fMZLB9k2j3RMmK2;P3U4\\L`Kc0=T3`4hKTKQ1?[3j4RKgJb1a0^3V5_JYJR2b0b3^5^LcJd3Z5[LhJf3V5ZLkJh3c0YJh2o1fLj3>[Ji2k1jLl3MSJ[O9l3h1mLn3CWJB6m3d1PMQ4YOYJJ4l3b1RMS4PO[J13l3_1SMi4POjIl3]1UMg4nNoIl3Y1XMf4lNRJl3R1_Mk4cNUJn3j0fMo4\\NXJn3h0hMn4YN\\Jo3e0jMn4VN^JP4c0lMn5T2QJnMn5R2QJoMo5Q2PJQNn5P2QJRNn5n1QJTNl5n1SJTNk5n1SJTNX2jMLW4hMQNR2UN1o3hMnMn1`N3g3kMkMh1kN8_3kMiMb1UO=W3nMeMc1YO:W3nMbMf1]O5V3QN_Mi1_O1W3QN\\Mm1GGR3WNYMP22[Of5d0gJnN`5l0gJkNf5i0^JQOQ6c0RJWO]6<fI_Oh66[mk5"}, "label": "man with beige shirt on facing towards televsion and away from the camera"}]}
{"id": 199721, "image": "COCO_train2014_000000199721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a tan shirt and jeans playing the wii entertainment system\"."}], "task": "refering", "answer_template": "The \"a man in a tan shirt and jeans playing the wii entertainment system\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 28, 29, 30, 31, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375], "bbox": [0.0, 0.0015625, 0.37415625, 0.9835833333333334], "iscrowd": 0, "area": 73914.08435, "rle": {"size": [480, 640], "counts": "]66^>=F:M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N3M2M3N2M3N2N2M3N2M3N2N2lFRMR6Q3cIZM[6h2ZIbMd6a2QIjMm6X2hHSNV7o1^H\\N`7g1UHdNi7^1lGlNR8W1cGTO[8i3N3M2O2M3M2O2M2N3N1N3M7J6I7I7J6I7J7I6J5K4M3L4L4L3M4L4M3L4L4L4TLVF[2W;0100O1O1O1O100O1O1O1O1O100O1O1O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000iNbMlE^2U:iMcEV2_:RNXEn1j:YNmDg1T;aNcDL3a0\\;>hD^O[;?jD]OX;?mD^OU;?oD^OS;?QE]OS;?QE^OR;=TE_Oo:>TE_Oo:?SE]OQ;a0QE\\OR;b0QEZOR;d0PEYOS;e0f1N3M2N2N2N2N1O000000000000000000001O0000000000000000000000000000000000000000001O000000nCIg1i0]2^OjKLm1d1V1`NlLMR2V2;mMbMOW2`2FaMQN2\\2k2QOSMbN4a2U3\\NgLRO6f2`3fMZLB9k2j3RMmK2;P3U4\\L`Kc0=T3`4hKTKQ1?[3j4RKgJb1a0^3V5_JYJR2b0b3^5^LcJd3Z5[LhJf3V5ZLkJh3c0YJh2o1fLj3>[Ji2k1jLl3MSJ[O9l3h1mLn3CWJB6m3d1PMQ4YOYJJ4l3b1RMS4PO[J13l3_1SMi4POjIl3]1UMg4nNoIl3Y1XMf4lNRJl3R1_Mk4cNUJn3j0fMo4\\NXJn3h0hMn4YN\\Jo3e0jMn4VN^JP4c0lMn5T2QJnMn5R2QJoMo5Q2PJQNn5P2QJRNn5n1QJTNl5n1SJTNk5n1SJTNX2jMLW4hMQNR2UN1o3hMnMn1`N3g3kMkMh1kN8_3kMiMb1UO=W3nMeMc1YO:W3nMbMf1]O5V3QN_Mi1_O1W3QN\\Mm1GGR3WNYMP22[Of5d0gJnN`5l0gJkNf5i0^JQOQ6c0RJWO]6<fI_Oh66[mk5"}, "label": "a man in a tan shirt and jeans playing the wii entertainment system"}]}
{"id": 240684, "image": "COCO_train2014_000000240684.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a white hooded jacket\"."}], "task": "refering", "answer_template": "The \"man in a white hooded jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 113, 114, 115, 132, 133, 150, 151, 168, 169, 186], "bbox": [0.33033999999999997, 0.30562666666666666, 0.4382, 0.7954933333333334], "iscrowd": 0, "area": 6087.936900000001, "rle": {"size": [375, 500], "counts": "\\cl1?`:j0M4L4L4L4L4M3XHSN_5Q2YJUNe5o1RJWNm5m1kIYNS6j1eI\\NZ6h1^I^N`6f1WI`Nh6d1PIbNn6T3O2N1L5J5K6K4M4K4M4LCRJWKn5h4TJWKl5i4VJTKl5k4UJTKk5k4XJSKi5l4XJRKi5m4ZJQKg5n4ZJPKg5o4b0O0O2O0O101O0O2hKTIc3`7O2dLbH:Oc0LcNc7<TIDMW1^OhNc77XJk0XNmNa72`Jk0QNRO`7OgJi0kMXOT:c0oE[OV:?nE_OV:;gEI^:1eENS;O1NddV3"}, "label": "man in a white hooded jacket"}]}
{"id": 240684, "image": "COCO_train2014_000000240684.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue pant and white jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing blue pant and white jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 113, 114, 115, 132, 133, 150, 151, 168, 169, 186], "bbox": [0.33033999999999997, 0.30562666666666666, 0.4382, 0.7954933333333334], "iscrowd": 0, "area": 6087.936900000001, "rle": {"size": [375, 500], "counts": "\\cl1?`:j0M4L4L4L4L4M3XHSN_5Q2YJUNe5o1RJWNm5m1kIYNS6j1eI\\NZ6h1^I^N`6f1WI`Nh6d1PIbNn6T3O2N1L5J5K6K4M4K4M4LCRJWKn5h4TJWKl5i4VJTKl5k4UJTKk5k4XJSKi5l4XJRKi5m4ZJQKg5n4ZJPKg5o4b0O0O2O0O101O0O2hKTIc3`7O2dLbH:Oc0LcNc7<TIDMW1^OhNc77XJk0XNmNa72`Jk0QNRO`7OgJi0kMXOT:c0oE[OV:?nE_OV:;gEI^:1eENS;O1NddV3"}, "label": "a man wearing blue pant and white jacket"}]}
{"id": 298034, "image": "COCO_train2014_000000298034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bathroom sink on the end with a duck - like object next to it\"."}], "task": "refering", "answer_template": "The \"a bathroom sink on the end with a duck - like object next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374], "bbox": [0.09128125000000001, 0.7216041666666667, 0.4615, 0.9563333333333334], "iscrowd": 0, "area": 20157.2442, "rle": {"size": [480, 640], "counts": "cbk01m>:F4L4L4L4L4L4L4M3M3M2N2N3N1N3M2N3M2O2M2N3M2N2O0O2O0O101N101N101N100O2O0O2O0O2O0O10001N100O2O000O101O0O100O2O000O10000O100O01000O10000O100O13M5K4L5KO10O100000000000O100000000000O01000000O1N2N2C=00O1000000000000000000000000000000O100000000000O100000000000000000000O1000000O0100000O1000000O100000000O1000000O1000000O1000000O100000O01000001N1000000O2O00000O101O000O10001N100O10000O2O0O100O101N100O100O2O0O100O101N100O1O1O1O2N1N3N1O2N1O2M3N1O2N2N1N3N1O2N2M2O2M3N1N3N1N3M4M7H9HR[Q5"}, "label": "a bathroom sink on the end with a duck - like object next to it"}]}
{"id": 298034, "image": "COCO_train2014_000000298034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sink to the left of the yellow soap dispenser\"."}], "task": "refering", "answer_template": "The \"sink to the left of the yellow soap dispenser\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374], "bbox": [0.09128125000000001, 0.7216041666666667, 0.4615, 0.9563333333333334], "iscrowd": 0, "area": 20157.2442, "rle": {"size": [480, 640], "counts": "cbk01m>:F4L4L4L4L4L4L4M3M3M2N2N3N1N3M2N3M2O2M2N3M2N2O0O2O0O101N101N101N100O2O0O2O0O2O0O10001N100O2O000O101O0O100O2O000O10000O100O01000O10000O100O13M5K4L5KO10O100000000000O100000000000O01000000O1N2N2C=00O1000000000000000000000000000000O100000000000O100000000000000000000O1000000O0100000O1000000O100000000O1000000O1000000O1000000O100000O01000001N1000000O2O00000O101O000O10001N100O10000O2O0O100O101N100O100O2O0O100O101N100O1O1O1O2N1N3N1O2N1O2M3N1O2N2N1N3N1O2N2M2O2M3N1N3N1N3M4M7H9HR[Q5"}, "label": "sink to the left of the yellow soap dispenser"}]}
{"id": 330806, "image": "COCO_train2014_000000330806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire in black\"."}], "task": "refering", "answer_template": "The \"the umpire in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 238, 239, 240, 241, 260, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 352, 354, 355, 356], "bbox": [0.2963125, 0.5283333333333333, 0.512046875, 0.9303333333333333], "iscrowd": 0, "area": 15157.655249999996, "rle": {"size": [480, 640], "counts": "m^i2=a>4L3L3O1N101O1N1nNYOkCi0R<ZOlCh0Q<\\OlCe0Q<_OmCc0P<@nCb0n;BPD?n;DPD>m;EQD=m;ERD;m;FRD<l;FSD;k;FTD<i;GVD9i;IUD9i;JUD7i;`1O2N2N1O2O1O010O1O001O00M3N2M3N2N2M3N3L3N2M3N2N21O2N3M3M2N3M3M2O0O1O1O1O1O2N10000000000O1000001O00000oEhLZ8X3dGjL\\8V3aGmL^8T3^GQM`8P3[GUMd8l2WGYMf8j2VGZMf8j2XGXMd8l2[GVM`8o2^GRM`8P3_GQM`8o2`GRM_8o2_GSM`8n2_GSM`8n2_GSM_8o2`GRM^8o2bGRM^8n2`GTM`8l2_GUM`8l2_GUMa8k2]GWMb8i2^GXMb8h2\\GZMd8f2[G[Me8e2YG]Mg8c2XG^Mh8a2XG`Mh8`2VGbMj8^2UGcMk8^2SGcMn8W40000001O00001YOoF]KR9?nFm22cLP9`0oFk23dLo8?oFk26dLk8a0PGi27eLj8`0PGi29fLh8`0PGh29hLg8?QGg2;iLe8?QGf2<iLe8?PGf2>jLb8?RGQ1FHh0gNa8?QGQ1HFj0hN_8>RGR1FFl0gN_8<TGT1BGc:POPFW1^OGi;7XDFl;7VDGm;7SDGQ<6QDHR<6oCG\\<NfC1c<EaC7f<B_C:g<_O^C<h=K4K5LPla4"}, "label": "the umpire in black"}]}
{"id": 330806, "image": "COCO_train2014_000000330806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire crouching down\"."}], "task": "refering", "answer_template": "The \"an umpire crouching down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 238, 239, 240, 241, 260, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 352, 354, 355, 356], "bbox": [0.2963125, 0.5283333333333333, 0.512046875, 0.9303333333333333], "iscrowd": 0, "area": 15157.655249999996, "rle": {"size": [480, 640], "counts": "m^i2=a>4L3L3O1N101O1N1nNYOkCi0R<ZOlCh0Q<\\OlCe0Q<_OmCc0P<@nCb0n;BPD?n;DPD>m;EQD=m;ERD;m;FRD<l;FSD;k;FTD<i;GVD9i;IUD9i;JUD7i;`1O2N2N1O2O1O010O1O001O00M3N2M3N2N2M3N3L3N2M3N2N21O2N3M3M2N3M3M2O0O1O1O1O1O2N10000000000O1000001O00000oEhLZ8X3dGjL\\8V3aGmL^8T3^GQM`8P3[GUMd8l2WGYMf8j2VGZMf8j2XGXMd8l2[GVM`8o2^GRM`8P3_GQM`8o2`GRM_8o2_GSM`8n2_GSM`8n2_GSM_8o2`GRM^8o2bGRM^8n2`GTM`8l2_GUM`8l2_GUMa8k2]GWMb8i2^GXMb8h2\\GZMd8f2[G[Me8e2YG]Mg8c2XG^Mh8a2XG`Mh8`2VGbMj8^2UGcMk8^2SGcMn8W40000001O00001YOoF]KR9?nFm22cLP9`0oFk23dLo8?oFk26dLk8a0PGi27eLj8`0PGi29fLh8`0PGh29hLg8?QGg2;iLe8?QGf2<iLe8?PGf2>jLb8?RGQ1FHh0gNa8?QGQ1HFj0hN_8>RGR1FFl0gN_8<TGT1BGc:POPFW1^OGi;7XDFl;7VDGm;7SDGQ<6QDHR<6oCG\\<NfC1c<EaC7f<B_C:g<_O^C<h=K4K5LPla4"}, "label": "an umpire crouching down"}]}
{"id": 330806, "image": "COCO_train2014_000000330806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 216, 217], "bbox": [0.380953125, 0.26981249999999996, 0.5456406250000001, 0.5899166666666666], "iscrowd": 0, "area": 8981.783499999998, "rle": {"size": [480, 640], "counts": "ebb3>_>d0\\O6J5H7I8XNUNQFR2m9QNmET2Q:nMlEV2Q:lMkEY2Q:jMkE]2P:eMlEb2n9aMnEe2o9[MoEk2m9VMoEQ3m9oLPFW3m9jLoE\\3n9dLoE_3Q:bLlE`3T:f0000000000N2N2O2H7I7O1N3N1O1O1O101N1O100O101O00001O1O1O1O010O1O1O100gNfDjN[;S1kDiNV;V1oDeNR;Z1REcNm:^1VE^Nj:b1ZEZNe:g1`EUN_:k1eEQNZ:P2jElMV:T2Z110O00O010000O101N10000O2O0O10001N2N2[OnB[OS=d0PCXOR=h0PCTOR=k0QCQOQ=o0QCmNQ=R1?O2O0O1O101K41O0010O101N100O100FlAFU>7PBFT>3QBK^fX4"}, "label": "a baseball player swinging a bat"}]}
{"id": 330806, "image": "COCO_train2014_000000330806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player wearing blue and the number twenty eight\"."}], "task": "refering", "answer_template": "The \"a baseball player wearing blue and the number twenty eight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 216, 217], "bbox": [0.380953125, 0.26981249999999996, 0.5456406250000001, 0.5899166666666666], "iscrowd": 0, "area": 8981.783499999998, "rle": {"size": [480, 640], "counts": "ebb3>_>d0\\O6J5H7I8XNUNQFR2m9QNmET2Q:nMlEV2Q:lMkEY2Q:jMkE]2P:eMlEb2n9aMnEe2o9[MoEk2m9VMoEQ3m9oLPFW3m9jLoE\\3n9dLoE_3Q:bLlE`3T:f0000000000N2N2O2H7I7O1N3N1O1O1O101N1O100O101O00001O1O1O1O010O1O1O100gNfDjN[;S1kDiNV;V1oDeNR;Z1REcNm:^1VE^Nj:b1ZEZNe:g1`EUN_:k1eEQNZ:P2jElMV:T2Z110O00O010000O101N10000O2O0O10001N2N2[OnB[OS=d0PCXOR=h0PCTOR=k0QCQOQ=o0QCmNQ=R1?O2O0O1O101K41O0010O101N100O100FlAFU>7PBFT>3QBK^fX4"}, "label": "a baseball player wearing blue and the number twenty eight"}]}
{"id": 330806, "image": "COCO_train2014_000000330806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player preparing to catch the ball\"."}], "task": "refering", "answer_template": "The \"a player preparing to catch the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 264, 265, 287, 288, 289, 309, 310, 311, 312, 331, 332, 333, 334, 335, 336, 357, 358], "bbox": [0.4145625, 0.6021875, 0.6203125, 0.8873958333333334], "iscrowd": 0, "area": 6515.052100000001, "rle": {"size": [480, 640], "counts": "gcl32e>:O1N2O1O1N2O1N01O001O001N2O001O001O001O1O1N2O001O1N2O1O1O0Ol`70^]H1]D1c;0ZD2f;NXD4P:KVF3F5P:NXFNE7n91[FID8o93[FED:Q:3XFED:T:4UFCD;V:8QFAB:]:=hE]OF9a:c0`EWOJ8f:l1TEUNl:o1oDRNR;m21O1O001O100O2N1O2N1O2N13N2M3N2M3?BN1ON1O2N1O2N1O2N2N3M2N2N2N2N3M2N3L5L0000001O000000001O0000001O00001]OYCkNi<o0\\CPOe<j0`CVOb<c0dC\\O]<?hC@Y<:lCFV<3PDLQ<OSD1o;HWD7Z=O001N101N1O2O0O2N10m_a3"}, "label": "a player preparing to catch the ball"}]}
{"id": 330806, "image": "COCO_train2014_000000330806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player wearing white\"."}], "task": "refering", "answer_template": "The \"a baseball player wearing white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 264, 265, 287, 288, 289, 309, 310, 311, 312, 331, 332, 333, 334, 335, 336, 357, 358], "bbox": [0.4145625, 0.6021875, 0.6203125, 0.8873958333333334], "iscrowd": 0, "area": 6515.052100000001, "rle": {"size": [480, 640], "counts": "gcl32e>:O1N2O1O1N2O1N01O001O001N2O001O001O001O1O1N2O001O1N2O1O1O0Ol`70^]H1]D1c;0ZD2f;NXD4P:KVF3F5P:NXFNE7n91[FID8o93[FED:Q:3XFED:T:4UFCD;V:8QFAB:]:=hE]OF9a:c0`EWOJ8f:l1TEUNl:o1oDRNR;m21O1O001O100O2N1O2N1O2N13N2M3N2M3?BN1ON1O2N1O2N1O2N2N3M2N2N2N2N3M2N3L5L0000001O000000001O0000001O00001]OYCkNi<o0\\CPOe<j0`CVOb<c0dC\\O]<?hC@Y<:lCFV<3PDLQ<OSD1o;HWD7Z=O001N101N1O2O0O2N10m_a3"}, "label": "a baseball player wearing white"}]}
{"id": 502838, "image": "COCO_train2014_000000502838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big black leather couch\"."}], "task": "refering", "answer_template": "The \"a big black leather couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 150, 151, 152, 153, 164, 165, 166, 167, 168, 174, 175, 187, 188, 189, 190, 191, 211, 212, 213, 214], "bbox": [0.08440625, 0.16541666666666668, 0.6897187499999999, 0.59775], "iscrowd": 0, "area": 39089.82725000001, "rle": {"size": [480, 640], "counts": "g]i06i><C<D5L4L4M3M3L4M2N3M2N2N2M3N1O2N1O2N1N3N1O2N1O1O2M2O2N1O2N1O1N3N1O2N1O2N1N3N2N1O2N2N1N3N2N2N1O2N2M2O2N2N1O2N2N2N101N2N1O2N2N101N2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N00O10000O10000O10000O1000000O10000O10000O10000O10000O10001N10000O10000O1000000O10000O101O0O10000O10000O10000O10000O101O0O10000O10000O10000O10000O01000O100YNTFZNl9Z1jF[NV9[1]G]Nc8\\1hG`NX8_1lG]NU8b1PHZNP8e1UHWNj7i1[HRNf7m1_HoMa7o1i2O1O1O001O1N2O1O1O001O1O1O10O0100O100O10O0100O100O10O0100O100O010O100O1000O1000O100000O10000O10O1000O10000O1000O01000000O10000O10O1000O10000O1000O01000000O10000O1000O01000000O10000O10O11O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O2N1O2N1O1O2N1O2N1O2N1N2O2N1O2N1O2N1N2O2N1O2N1O1O2M2O2N1O1O2N1O2N1N2O2N1O1O2N1O2M2O101N000000O10001O0O100000000O2O00000O101O00000O101O000O100O1K6J5J6K5L4M4M2M3M3M3L5J5L5J5L5J6K5K5J6K5J6K5J6K4KTVm2"}, "label": "a big black leather couch"}]}
{"id": 502838, "image": "COCO_train2014_000000502838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black leather couch behind child\"."}], "task": "refering", "answer_template": "The \"black leather couch behind child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 150, 151, 152, 153, 164, 165, 166, 167, 168, 174, 175, 187, 188, 189, 190, 191, 211, 212, 213, 214], "bbox": [0.08440625, 0.16541666666666668, 0.6897187499999999, 0.59775], "iscrowd": 0, "area": 39089.82725000001, "rle": {"size": [480, 640], "counts": "g]i06i><C<D5L4L4M3M3L4M2N3M2N2N2M3N1O2N1O2N1N3N1O2N1O1O2M2O2N1O2N1O1N3N1O2N1O2N1N3N2N1O2N2N1N3N2N2N1O2N2M2O2N2N1O2N2N2N101N2N1O2N2N101N2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N00O10000O10000O10000O1000000O10000O10000O10000O10000O10001N10000O10000O1000000O10000O101O0O10000O10000O10000O10000O101O0O10000O10000O10000O10000O01000O100YNTFZNl9Z1jF[NV9[1]G]Nc8\\1hG`NX8_1lG]NU8b1PHZNP8e1UHWNj7i1[HRNf7m1_HoMa7o1i2O1O1O001O1N2O1O1O001O1O1O10O0100O100O10O0100O100O10O0100O100O010O100O1000O1000O100000O10000O10O1000O10000O1000O01000000O10000O10O1000O10000O1000O01000000O10000O1000O01000000O10000O10O11O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O2N1O2N1O1O2N1O2N1O2N1N2O2N1O2N1O2N1N2O2N1O2N1O1O2M2O2N1O1O2N1O2N1N2O2N1O1O2N1O2M2O101N000000O10001O0O100000000O2O00000O101O00000O101O000O100O1K6J5J6K5L4M4M2M3M3M3L5J5L5J5L5J6K5K5J6K5J6K5J6K4KTVm2"}, "label": "black leather couch behind child"}]}
{"id": 502838, "image": "COCO_train2014_000000502838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair on which a baby is sitting\"."}], "task": "refering", "answer_template": "The \"a chair on which a baby is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 123, 124, 125, 126, 127, 145, 146, 147, 148, 150, 151, 168, 169, 170, 171, 174, 191, 192, 193, 194, 197, 198, 215, 216, 237, 238, 260, 261, 269, 283, 284, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.31640625, 0.2825625, 0.732421875, 0.8919374999999999], "iscrowd": 0, "area": 23313.16875, "rle": {"size": [480, 640], "counts": "i^o23k>4L4L4L4M3L4L4L4L4M2mMTOnEP1a9A[Fc0T9NhF6h8:TGJg8<UGGh8<TGHh8<TGHi8;SGIj8:RGJk89RGaNDOW9d1PG\\NN0n8h1PGWN71f8k1oFSNa01]8P3hGPMU8n2PHRMl7n2YHQMd7o2_HPM_7P3dHPMY7P3kHoLQ7R3RInLk6R3XInLe6R3^InL_6R3cIoLY6R3iIoLT6Q3nIPMP6n2SJSMj5m2YJSMe5l2]JUM`5k2bJVM\\5j2eJWMY5i2hJXMU5h2mJYMQ5g2PKZMn4f2SK[Mj4f2WK[Mg4d2\\KYLaMi0o6o2aKWLbMj0k6o2dKVLdMk0e6o2hKULfMl0^6P3mKSLhMl0Y6P3QLSLiMm0S6P3ULRLkMn0l5Q3ZLPLmMn0g5R3]LoKoMo0a5Q3^MjLc2W3`MdLb2\\3aM_L`2b3cMXL`2g3dMTL^2l3dMPL]2Q4Y40WHnKd4Q4]KQL`4P4`KRL^4n3bKTL\\4Y3aHQMS3HY4W3gHnLQ3LW4U3jHnLo2OT4S3PIlLl23R4o2UImLi26P4l2ZIkLg2;m3i2^IkLe2>j3g2dIiLb2b0h3c2iIjL_2e0f3`2VMaMi2^2XMdMe2\\2\\MfMb2X2`MjM^2U2cMmM[2R2fMPNW2P2jMRNT2l1nMVNP2i1QNXNm1h1TNZNj1e1WN[Ni1FlIi0]4Ag1ERJd0X4Ge1DYJ`0S4Ld1C^J<o31c1AeJ7j38a1@jJ3f3<`1@PKOa3a0_1^OWKJ\\3h0]1]O\\KFX3m0\\1[ObK_OW3V1V1[Og0d0ZOZOg0g0YOXOh0h0XOVOj0j0UOVOl0j0TOVOl0j0TOVOl0j0TOVOl0j0TOVOl0j0TOVOl0j0UOUOk0k0UOUOk0k0UOUOk0k0UOUOk0k0VOUOi0k0WOUOi0k0WOUOi0k0XOTOh0l0XOTOh0l0XOTOh0l0YOSOg0m0YOSOg0m0YOSOg0m0YOSOg0m0ZOROf0n0ZOSOe0m0[OSOe0m0\\OROd0n0\\OROd0n0\\OROd0n0]OQOc0o0]OQOc0o0]OQOc0o0]OQOc0o0^OPOb0P1^OQOa0o0_OQOa0o0@PO?Q1AoN?Q1AoN?Q1AoN>R1CmN=S1CmN=S1CmN<T1EkN;U1FlN8T1IlN5U1MkN1U10lNNT14lNIU18lNFT1<kNCU1>lN_OT1d0lNZOT1g0mNWOS1k0mNSOS1o0lNoNU1R1lNlNT1V1lNhNT1Y1oNbNR1`1oN]NQ1e1QOWNo0j1SORNn0P2TOlMl0V2VORJ[Oo1_1P4WOmI\\OR2^1S4VOgI^OV2\\1U4VOaI@Z2Z1V4WO\\IA]2Y1Y4UOVIEa2V1[4UOPIGe2T1\\4UOmHHf2T1_4SOjHIg2T1a4ROgHJh2T1b4QOfHJh2V1d4oNcHKi2V1f4nN`HLj2V1g4mN_HMi2W1j4kN\\HNj2W1l4iNZH0j2W1m4iNXH0k2W1o4eNXH4h2X1Q5`NZH8e2X1S5ZN\\H>a2X1P6hNoIY1Q6fNPJZ1P6fNPJZ1P6fNoI[1Q6eNoI[1Q6eNoI[1Q6eNnI\\1S6cNmI]1S6cNmI]1S6cNlI^1T6]NQJc1o5XNVJh1`910000O10000O1000000O100O100O2O0O10000O100O100O101N100O10000O100O100O2O0O100O101O0O100O101N100O10001N100O101N100O100OL5J6J5K6I6K6J5K6J5K6J5K6J6J5J7JPk_2"}, "label": "a chair on which a baby is sitting"}]}
{"id": 502838, "image": "COCO_train2014_000000502838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a high chair\"."}], "task": "refering", "answer_template": "The \"a high chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 123, 124, 125, 126, 127, 145, 146, 147, 148, 150, 151, 168, 169, 170, 171, 174, 191, 192, 193, 194, 197, 198, 215, 216, 237, 238, 260, 261, 269, 283, 284, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.31640625, 0.2825625, 0.732421875, 0.8919374999999999], "iscrowd": 0, "area": 23313.16875, "rle": {"size": [480, 640], "counts": "i^o23k>4L4L4L4M3L4L4L4L4M2mMTOnEP1a9A[Fc0T9NhF6h8:TGJg8<UGGh8<TGHh8<TGHi8;SGIj8:RGJk89RGaNDOW9d1PG\\NN0n8h1PGWN71f8k1oFSNa01]8P3hGPMU8n2PHRMl7n2YHQMd7o2_HPM_7P3dHPMY7P3kHoLQ7R3RInLk6R3XInLe6R3^InL_6R3cIoLY6R3iIoLT6Q3nIPMP6n2SJSMj5m2YJSMe5l2]JUM`5k2bJVM\\5j2eJWMY5i2hJXMU5h2mJYMQ5g2PKZMn4f2SK[Mj4f2WK[Mg4d2\\KYLaMi0o6o2aKWLbMj0k6o2dKVLdMk0e6o2hKULfMl0^6P3mKSLhMl0Y6P3QLSLiMm0S6P3ULRLkMn0l5Q3ZLPLmMn0g5R3]LoKoMo0a5Q3^MjLc2W3`MdLb2\\3aM_L`2b3cMXL`2g3dMTL^2l3dMPL]2Q4Y40WHnKd4Q4]KQL`4P4`KRL^4n3bKTL\\4Y3aHQMS3HY4W3gHnLQ3LW4U3jHnLo2OT4S3PIlLl23R4o2UImLi26P4l2ZIkLg2;m3i2^IkLe2>j3g2dIiLb2b0h3c2iIjL_2e0f3`2VMaMi2^2XMdMe2\\2\\MfMb2X2`MjM^2U2cMmM[2R2fMPNW2P2jMRNT2l1nMVNP2i1QNXNm1h1TNZNj1e1WN[Ni1FlIi0]4Ag1ERJd0X4Ge1DYJ`0S4Ld1C^J<o31c1AeJ7j38a1@jJ3f3<`1@PKOa3a0_1^OWKJ\\3h0]1]O\\KFX3m0\\1[ObK_OW3V1V1[Og0d0ZOZOg0g0YOXOh0h0XOVOj0j0UOVOl0j0TOVOl0j0TOVOl0j0TOVOl0j0TOVOl0j0TOVOl0j0UOUOk0k0UOUOk0k0UOUOk0k0UOUOk0k0VOUOi0k0WOUOi0k0WOUOi0k0XOTOh0l0XOTOh0l0XOTOh0l0YOSOg0m0YOSOg0m0YOSOg0m0YOSOg0m0ZOROf0n0ZOSOe0m0[OSOe0m0\\OROd0n0\\OROd0n0\\OROd0n0]OQOc0o0]OQOc0o0]OQOc0o0]OQOc0o0^OPOb0P1^OQOa0o0_OQOa0o0@PO?Q1AoN?Q1AoN?Q1AoN>R1CmN=S1CmN=S1CmN<T1EkN;U1FlN8T1IlN5U1MkN1U10lNNT14lNIU18lNFT1<kNCU1>lN_OT1d0lNZOT1g0mNWOS1k0mNSOS1o0lNoNU1R1lNlNT1V1lNhNT1Y1oNbNR1`1oN]NQ1e1QOWNo0j1SORNn0P2TOlMl0V2VORJ[Oo1_1P4WOmI\\OR2^1S4VOgI^OV2\\1U4VOaI@Z2Z1V4WO\\IA]2Y1Y4UOVIEa2V1[4UOPIGe2T1\\4UOmHHf2T1_4SOjHIg2T1a4ROgHJh2T1b4QOfHJh2V1d4oNcHKi2V1f4nN`HLj2V1g4mN_HMi2W1j4kN\\HNj2W1l4iNZH0j2W1m4iNXH0k2W1o4eNXH4h2X1Q5`NZH8e2X1S5ZN\\H>a2X1P6hNoIY1Q6fNPJZ1P6fNPJZ1P6fNoI[1Q6eNoI[1Q6eNoI[1Q6eNnI\\1S6cNmI]1S6cNmI]1S6cNlI^1T6]NQJc1o5XNVJh1`910000O10000O1000000O100O100O2O0O10000O100O100O101N100O10000O100O100O2O0O100O101O0O100O101N100O10001N100O101N100O100OL5J6J5K6I6K6J5K6J5K6J5K6J6J5J7JPk_2"}, "label": "a high chair"}]}
{"id": 470072, "image": "COCO_train2014_000000470072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person using their right hand to flip burgers on a grill\"."}], "task": "refering", "answer_template": "The \"a person using their right hand to flip burgers on a grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 115, 116, 118, 119, 120, 121, 138, 139, 141, 142, 143, 144, 161, 162, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 302, 324, 325], "bbox": [0.0043749999999999995, 0.008797169811320754, 0.29703125, 0.9868160377358491], "iscrowd": 0, "area": 42756.11345, "rle": {"size": [424, 640], "counts": "lW1c4e8b6^Ie0[O1O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001OfHSOY2m0eMBn1>PN1d1NZNb0X1^OeNR1n0nNPOa1d0^NZOl1>TN@Q2=oMAV2=iMA\\2<dMBa2;_MCf2;YMCk2;UMCP3;oLCV3:jLD[39eLE`39_LEf38ZLFk37ULFQ48nKFW47iKG\\46dKHa46^KHg45YKIl44TKJP55oJIW53iJK`5N`J0i5JVJ9o5_OQJd0S6VOlIl0Y6mNgIV1]6cNcI_1c6ZN\\Ii1h6PNXIR2m6gMSI\\2R7]MmHf2S7WMmHk2S7SMmHP3R7oLmHS3T7jLlHY3S7eLmH^3R7aLmHa3S7]LmHf3R7YLmHi3S7ULmHn3o1_Kb2a0_KR4f1dKm29\\KV4]1gK[30XK\\4S1kKg3GVK`4i0QLS4^OSKd4`0UL_4UOQKh47YLk4mNnJm4M]LW5eNkJk6W5SIiJm6Z5PIfJo6]5PIbJP7`5nH`JR7k500O101O000O1000000O1000000O1000000O101O3M3L3N3M3M2M4M3M2N3L4M2N3M3L3N3M3M3L3N3M3L3M4L4L3M4L4L3M4K5L3M4L4L3M4L4L3M4L4K4M4L4L3M4L4L3M4L4K4M4L4L3M4L4L4L3M4K5L3M4L4L3MUUj5"}, "label": "a person using their right hand to flip burgers on a grill"}]}
{"id": 470072, "image": "COCO_train2014_000000470072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person whose left arm is visible\"."}], "task": "refering", "answer_template": "The \"the person whose left arm is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.719078125, 0.005047169811320755, 1.0, 0.9983254716981131], "iscrowd": 0, "area": 45704.12820000001, "rle": {"size": [424, 640], "counts": "h`n51S=5K4K6K4L5K4K6K4L5K4L5J501O001O00001O001O001O001O001O001O001O00001O001O001O001cGRN^5n1jIlNT6U1SIFi6;_H`0^7g2N3M3M2N3M3M2N3M3M3M3M2N3M3M001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O00000000001O00000000000000001O0000000FoH]JQ7]5UIcJk6X5ZIhJf6S5_ImJb6m4cISK\\6i4iIVKX6e4mI[KR6a4SJ_Km5\\4XJdKg5X4^JhKb5S4cJmK\\5o3iJQLV5j3PKVLP5e3UK[Lj4`3\\K`Ld4[3aKeL^4V3hKjLX4Q3mKoLR4l2TLTMl3f2ZLZMe3b2`L^M_3]2gLcMY3X2lLhMS3S2SMmMm2n1XMRNg2i1_MWNa2c1eM]NZ2_1kMaNU2Y1QNgNn1U1WNkNh1P1^NPOb1j0dNVO[1f0iN[OW1?oNAP1;UOEk05[OKd01AOl600000000000000000000000000000000000000000000000000000000O10000000000000000000000000O10000000N"}, "label": "the person whose left arm is visible"}]}
{"id": 470072, "image": "COCO_train2014_000000470072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person standing on the right side of a grill with an arm up\"."}], "task": "refering", "answer_template": "The \"a person standing on the right side of a grill with an arm up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.719078125, 0.005047169811320755, 1.0, 0.9983254716981131], "iscrowd": 0, "area": 45704.12820000001, "rle": {"size": [424, 640], "counts": "h`n51S=5K4K6K4L5K4K6K4L5K4L5J501O001O00001O001O001O001O001O001O001O00001O001O001O001cGRN^5n1jIlNT6U1SIFi6;_H`0^7g2N3M3M2N3M3M2N3M3M3M3M2N3M3M001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O00000000001O00000000000000001O0000000FoH]JQ7]5UIcJk6X5ZIhJf6S5_ImJb6m4cISK\\6i4iIVKX6e4mI[KR6a4SJ_Km5\\4XJdKg5X4^JhKb5S4cJmK\\5o3iJQLV5j3PKVLP5e3UK[Lj4`3\\K`Ld4[3aKeL^4V3hKjLX4Q3mKoLR4l2TLTMl3f2ZLZMe3b2`L^M_3]2gLcMY3X2lLhMS3S2SMmMm2n1XMRNg2i1_MWNa2c1eM]NZ2_1kMaNU2Y1QNgNn1U1WNkNh1P1^NPOb1j0dNVO[1f0iN[OW1?oNAP1;UOEk05[OKd01AOl600000000000000000000000000000000000000000000000000000000O10000000000000000000000000O10000000N"}, "label": "a person standing on the right side of a grill with an arm up"}]}
{"id": 11324, "image": "COCO_train2014_000000011324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe not facing the camera\"."}], "task": "refering", "answer_template": "The \"giraffe not facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 93, 94, 96, 97, 98, 99, 116, 117, 118, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 308, 309, 310, 322, 323, 324, 325, 326, 345, 346, 347, 348], "bbox": [0.0, 0.21973214285714285, 0.56659375, 0.9887946428571429], "iscrowd": 0, "area": 65189.49029999998, "rle": {"size": [448, 640], "counts": "Z<6h=>B=C=C8H2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2iJmK?U4]ORL>P4^OVL>l3AWL;k3CZL8h3G[L5g3J\\L2f3L_LOc31_LKd35]LGe3:]LBd3?]L]Oe3d0\\LXOf3i0\\LSOe3n0\\LnNf3S1[LiNg3X1[LdNf3]1[L_Ng3b1[LZNf3g1[LUNg3l1ZLPNh3Q2ZLkMg3U2[LgMg3Z2ZLbMh3_2YL^Mh3c2XLZMj3g2WLUMk3l2VLQMk3P3VLlLl3U3ULhLl3Y3TLdLn3]3SL_Lo3b3RL[Lo3f3RLVLP4k3QLQLQ4P4PLmKQ4T4oKiKS4X4nKeKS4\\4nK`KT4a4mK[KU4f4lKWKV4i4jKTKX4m4iKoJY4R5Z21O1O1O1O1O1O1O1O1O1O1O1O1O1M4L3M3M3M4L3N2M3M4L101N1O2N100O2N101N1O2N100O2N1O2O0O2N100O2NnM_IhL`6Y3fIbLY6^3nIYLS6i3RJoKo5R4XJdKk5\\4[J\\Kf5f4aJPK`5Q5iJdJX5]5e11N1O102M2N2N2O1N3M2N2O1N2N4M3L3M4L4M3L4N101N2O1N101O1N2O1N101N2O[O^IlIa6R6dIkI[6S6kIkIT6S6QJkIm5T6XJiIh5U6\\JhIe5V6`JfI`5Y6U1N1O2N2O001O010000O10O0100O1O1O1O1O001O1O1O1O1O1O001O1O1O1O2N1O2N2N1O2N1O2N101N1O2N2N1O2N1O2M2O2N2N1O2N1O2N1O2N1O0000000000000000000000000000001O000000000000000000000000000000000002N2N2N2N2O1N2NjJ`Hn3^7PLgHo3W7nKPIP4n6mKXIo3i6nK]Io3d6kKeIQ4\\6hKlIV4T6eKSJY4n5`KYJ_4e7O2M2O1O2N1O1O2N1O1N3N1O1O2N1O1O2M2O1O2N1O1O2N1N2O2N1O1O2N1O1O2M2O1O1O001O1N1N3M2N3M2N3M3M2N3M2N3M3M201O001O001O2N2N2N3M2N2N2N2M3N2N3L3N2N2M3N2N3M2M3N2Ndkh3"}, "label": "giraffe not facing the camera"}]}
{"id": 11324, "image": "COCO_train2014_000000011324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking off in the distance\"."}], "task": "refering", "answer_template": "The \"a giraffe looking off in the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 93, 94, 96, 97, 98, 99, 116, 117, 118, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 308, 309, 310, 322, 323, 324, 325, 326, 345, 346, 347, 348], "bbox": [0.0, 0.21973214285714285, 0.56659375, 0.9887946428571429], "iscrowd": 0, "area": 65189.49029999998, "rle": {"size": [448, 640], "counts": "Z<6h=>B=C=C8H2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2iJmK?U4]ORL>P4^OVL>l3AWL;k3CZL8h3G[L5g3J\\L2f3L_LOc31_LKd35]LGe3:]LBd3?]L]Oe3d0\\LXOf3i0\\LSOe3n0\\LnNf3S1[LiNg3X1[LdNf3]1[L_Ng3b1[LZNf3g1[LUNg3l1ZLPNh3Q2ZLkMg3U2[LgMg3Z2ZLbMh3_2YL^Mh3c2XLZMj3g2WLUMk3l2VLQMk3P3VLlLl3U3ULhLl3Y3TLdLn3]3SL_Lo3b3RL[Lo3f3RLVLP4k3QLQLQ4P4PLmKQ4T4oKiKS4X4nKeKS4\\4nK`KT4a4mK[KU4f4lKWKV4i4jKTKX4m4iKoJY4R5Z21O1O1O1O1O1O1O1O1O1O1O1O1O1M4L3M3M3M4L3N2M3M4L101N1O2N100O2N101N1O2N100O2N1O2O0O2N100O2NnM_IhL`6Y3fIbLY6^3nIYLS6i3RJoKo5R4XJdKk5\\4[J\\Kf5f4aJPK`5Q5iJdJX5]5e11N1O102M2N2N2O1N3M2N2O1N2N4M3L3M4L4M3L4N101N2O1N101O1N2O1N101N2O[O^IlIa6R6dIkI[6S6kIkIT6S6QJkIm5T6XJiIh5U6\\JhIe5V6`JfI`5Y6U1N1O2N2O001O010000O10O0100O1O1O1O1O001O1O1O1O1O1O001O1O1O1O2N1O2N2N1O2N1O2N101N1O2N2N1O2N1O2M2O2N2N1O2N1O2N1O2N1O0000000000000000000000000000001O000000000000000000000000000000000002N2N2N2N2O1N2NjJ`Hn3^7PLgHo3W7nKPIP4n6mKXIo3i6nK]Io3d6kKeIQ4\\6hKlIV4T6eKSJY4n5`KYJ_4e7O2M2O1O2N1O1O2N1O1N3N1O1O2N1O1O2M2O1O2N1O1O2N1N2O2N1O1O2N1O1O2M2O1O1O001O1N1N3M2N3M2N3M3M2N3M2N3M3M201O001O001O2N2N2N3M2N2N2N2M3N2N3L3N2N2M3N2N3M2M3N2Ndkh3"}, "label": "a giraffe looking off in the distance"}]}
{"id": 11324, "image": "COCO_train2014_000000011324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with one ear cocked back , almost appearing to smirk at the camera\"."}], "task": "refering", "answer_template": "The \"a giraffe with one ear cocked back , almost appearing to smirk at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 16, 17, 36, 37, 39, 40, 58, 59, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361], "bbox": [0.3908125, 0.004486607142857142, 0.8538125000000001, 0.9865401785714285], "iscrowd": 0, "area": 64397.53995, "rle": {"size": [448, 640], "counts": "j^]32o=5J5K6J6J4L2O0O2N1O2N2N1O2O0O2N1O2N2N101N1O2N2N1O2O0O2N2N1O2N1O2O1N1O2N1O2N101N2N1O2N000010O0001O000010O000001O00010O00001O0000010O00001O0002O0O1O2N10001N10001N10001N10001O0O10gKlMfLS2V3PNkLS2m2QNTMR2d2RN\\MS2[2QNeMT2R2PNmMU2l1nMSNX2d1lM\\NY2[1kMdNZ2T1jMlN[2k0iMTO\\2e0gM[O^2<fMC`24dMLj2BZM=V3POnLP1U3jNnL6T4DPL8V4BnK:W4AmK;Y4_OkK=[4^OhK>^4\\OfK<d4^O`K:j4@ZK8P5BTK6U5FnJ3[5GiJ1a5IcJOg5K]J2k5EYJ:m5^OVJ`0f9O1N2N2O1N20000001O00000000001O00000000001O0000101N1O2N1O2N1O2N2N1O2N9G=C=C4L1O001O00001O001O00001O001O00001O001O00001O0000O1O1O1O1O1O1O1O1O1O1O1O100O1oNQ1O1N2O1OF:\\Od0[Oe0[Oe0D<K5K5K5L4K5K5N21O1O0O2O10O2O00001N101O0O2O000O2O001O0O101O0O2Oh0XO7H7J7I6J7H7J7I6I8I6J7H6K001O1N2N2M3N2N2M3N1O2M3N2N2N2M3PNUGhNm8Q1lG[NU8`1cHkM`7n1]2I7I7I7L4K5L2N3O10000O10O010000O10OG^CUOc<i0bCTO^<j0=N2N2N101N2N2N2NQPY1"}, "label": "a giraffe with one ear cocked back , almost appearing to smirk at the camera"}]}
{"id": 11324, "image": "COCO_train2014_000000011324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe totally facing forward\"."}], "task": "refering", "answer_template": "The \"giraffe totally facing forward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 16, 17, 36, 37, 39, 40, 58, 59, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361], "bbox": [0.3908125, 0.004486607142857142, 0.8538125000000001, 0.9865401785714285], "iscrowd": 0, "area": 64397.53995, "rle": {"size": [448, 640], "counts": "j^]32o=5J5K6J6J4L2O0O2N1O2N2N1O2O0O2N1O2N2N101N1O2N2N1O2O0O2N2N1O2N1O2O1N1O2N1O2N101N2N1O2N000010O0001O000010O000001O00010O00001O0000010O00001O0002O0O1O2N10001N10001N10001N10001O0O10gKlMfLS2V3PNkLS2m2QNTMR2d2RN\\MS2[2QNeMT2R2PNmMU2l1nMSNX2d1lM\\NY2[1kMdNZ2T1jMlN[2k0iMTO\\2e0gM[O^2<fMC`24dMLj2BZM=V3POnLP1U3jNnL6T4DPL8V4BnK:W4AmK;Y4_OkK=[4^OhK>^4\\OfK<d4^O`K:j4@ZK8P5BTK6U5FnJ3[5GiJ1a5IcJOg5K]J2k5EYJ:m5^OVJ`0f9O1N2N2O1N20000001O00000000001O00000000001O0000101N1O2N1O2N1O2N2N1O2N9G=C=C4L1O001O00001O001O00001O001O00001O001O00001O0000O1O1O1O1O1O1O1O1O1O1O1O100O1oNQ1O1N2O1OF:\\Od0[Oe0[Oe0D<K5K5K5L4K5K5N21O1O0O2O10O2O00001N101O0O2O000O2O001O0O101O0O2Oh0XO7H7J7I6J7H7J7I6I8I6J7H6K001O1N2N2M3N2N2M3N1O2M3N2N2N2M3PNUGhNm8Q1lG[NU8`1cHkM`7n1]2I7I7I7L4K5L2N3O10000O10O010000O10OG^CUOc<i0bCTO^<j0=N2N2N101N2N2N2NQPY1"}, "label": "giraffe totally facing forward"}]}
{"id": 519228, "image": "COCO_train2014_000000519228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in tan shirt and khaki shorts walking along the beach\"."}], "task": "refering", "answer_template": "The \"a man in tan shirt and khaki shorts walking along the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 99, 100, 101, 122, 123, 124, 145, 146, 147, 168, 169, 170, 191, 192, 193, 214, 215, 216, 236, 237, 239], "bbox": [0.2703125, 0.24184397163120566, 0.4218125, 0.7353900709219858], "iscrowd": 0, "area": 10068.956200000004, "rle": {"size": [423, 640], "counts": "ShW26Q=1O001O1O1O2N00010O001O0001OO1O1O1O2N1H8C=G9]KdNPMg1m2`NeLj1Y3YNaLl1^3VN\\KI_NT2T6UNYKM\\NR2Z6SNWKN[NQ2^6TNRK0ZNP2c6SNoJ0YNQ2g6SNjJl2V5VMiIWO4e3R6UMcI^O7_3W6SM]IC:X3]6UMTIH=i2j6\\NRIb1T7]NhHa1^7_N^H`1g7j16J6K5J7M300O010O100O1O2M2O2M4M2N1N1N3M23M2N2N2O1N2O2N2M01O1O1N101O0O2O1O0102M2O2UHeKe6]4VIeKn6Z4lHiKX7X4aHjKY8c3XGeLn8g37I7I2N3M2N2N2UMgFd1^9RNlFh1_:UNjD[1c;N3M2Nfbh4"}, "label": "a man in tan shirt and khaki shorts walking along the beach"}]}
{"id": 519228, "image": "COCO_train2014_000000519228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the khaki shorts\"."}], "task": "refering", "answer_template": "The \"the man in the khaki shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 99, 100, 101, 122, 123, 124, 145, 146, 147, 168, 169, 170, 191, 192, 193, 214, 215, 216, 236, 237, 239], "bbox": [0.2703125, 0.24184397163120566, 0.4218125, 0.7353900709219858], "iscrowd": 0, "area": 10068.956200000004, "rle": {"size": [423, 640], "counts": "ShW26Q=1O001O1O1O2N00010O001O0001OO1O1O1O2N1H8C=G9]KdNPMg1m2`NeLj1Y3YNaLl1^3VN\\KI_NT2T6UNYKM\\NR2Z6SNWKN[NQ2^6TNRK0ZNP2c6SNoJ0YNQ2g6SNjJl2V5VMiIWO4e3R6UMcI^O7_3W6SM]IC:X3]6UMTIH=i2j6\\NRIb1T7]NhHa1^7_N^H`1g7j16J6K5J7M300O010O100O1O2M2O2M4M2N1N1N3M23M2N2N2O1N2O2N2M01O1O1N101O0O2O1O0102M2O2UHeKe6]4VIeKn6Z4lHiKX7X4aHjKY8c3XGeLn8g37I7I2N3M2N2N2UMgFd1^9RNlFh1_:UNjD[1c;N3M2Nfbh4"}, "label": "the man in the khaki shorts"}]}
{"id": 240709, "image": "COCO_train2014_000000240709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dining chair closest to ' cpml '.\"."}], "task": "refering", "answer_template": "The \"the dining chair closest to ' cpml '.\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 180, 181, 203, 204, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 316, 317], "bbox": [0.718453125, 0.46683417085427137, 0.900078125, 1.0], "iscrowd": 0, "area": 8645.520899999992, "rle": {"size": [398, 640], "counts": "dTc56Q<7H8I7H8I7H9H7H8I7K5J6G:O000000001O00000000001O000000010O00000001O01O0001O0001O01O00000010O000001O01O0001O000c0]O01O001O00001ONROgGkMV8W2n01O2N1N3Nb0^O3M3M3M2MG:fNZ1[Oe03M2N2N2M3N2N2N2N3M1O03M2NKVNkEh1X:43N2M3M4^N\\EU1k:dNVE[1n:O2N100O1O1O1O1O1O100O1O8POcDb0f;M2N2N2O2M5Koeh0"}, "label": "the dining chair closest to ' cpml '."}]}
{"id": 240709, "image": "COCO_train2014_000000240709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair closest to cpml\"."}], "task": "refering", "answer_template": "The \"the chair closest to cpml\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 180, 181, 203, 204, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 316, 317], "bbox": [0.718453125, 0.46683417085427137, 0.900078125, 1.0], "iscrowd": 0, "area": 8645.520899999992, "rle": {"size": [398, 640], "counts": "dTc56Q<7H8I7H8I7H9H7H8I7K5J6G:O000000001O00000000001O000000010O00000001O01O0001O0001O01O00000010O000001O01O0001O000c0]O01O001O00001ONROgGkMV8W2n01O2N1N3Nb0^O3M3M3M2MG:fNZ1[Oe03M2N2N2M3N2N2N2N3M1O03M2NKVNkEh1X:43N2M3M4^N\\EU1k:dNVE[1n:O2N100O1O1O1O1O1O100O1O8POcDb0f;M2N2N2O2M5Koeh0"}, "label": "the chair closest to cpml"}]}
{"id": 240709, "image": "COCO_train2014_000000240709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair closest to the refrigerator\"."}], "task": "refering", "answer_template": "The \"a chair closest to the refrigerator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314], "bbox": [0.49484374999999997, 0.4671608040201005, 0.7028749999999999, 0.9982914572864321], "iscrowd": 0, "area": 16772.034450000006, "rle": {"size": [398, 640], "counts": "`\\k34Z<2O0kC9a;?K41oNeDj0];5J6K5J5K6J6J6K4K6K5QIjM\\3[2QLSNo3Z2UKUNj4X2]JVNb5W2eIUN[6n31C=G82O001O010O000010O01O010O001O010O000010O01O010O001O0100O1O010O100O100O100O010O1O100O100O100OVIjJ]6V5YISKg6Z51N01N2M3C[JnIg5S663M4N102O001^JbIQ5_6lJoIi4V7iK^Hl2l7bLkHn2[80000O2O0O10001N10001N10001N10001N10001N10001N10001N101O0O101O0O101O0O10]FeMT9Z2eFnMZ9Q2`FWN_9]20N100O2E:[Oe0[Og0YOg0BTbY2"}, "label": "a chair closest to the refrigerator"}]}
{"id": 240709, "image": "COCO_train2014_000000240709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest chair\"."}], "task": "refering", "answer_template": "The \"the closest chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314], "bbox": [0.49484374999999997, 0.4671608040201005, 0.7028749999999999, 0.9982914572864321], "iscrowd": 0, "area": 16772.034450000006, "rle": {"size": [398, 640], "counts": "`\\k34Z<2O0kC9a;?K41oNeDj0];5J6K5J5K6J6J6K4K6K5QIjM\\3[2QLSNo3Z2UKUNj4X2]JVNb5W2eIUN[6n31C=G82O001O010O000010O01O010O001O010O000010O01O010O001O0100O1O010O100O100O100O010O1O100O100O100OVIjJ]6V5YISKg6Z51N01N2M3C[JnIg5S663M4N102O001^JbIQ5_6lJoIi4V7iK^Hl2l7bLkHn2[80000O2O0O10001N10001N10001N10001N10001N10001N10001N101O0O101O0O101O0O10]FeMT9Z2eFnMZ9Q2`FWN_9]20N100O2E:[Oe0[Og0YOg0BTbY2"}, "label": "the closest chair"}]}
{"id": 322630, "image": "COCO_train2014_000000322630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult elephant standing with a baby elephant\"."}], "task": "refering", "answer_template": "The \"an adult elephant standing with a baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310], "bbox": [0.12770491803278688, 0.002125, 1.0, 0.881046875], "iscrowd": 0, "area": 151075.8652500001, "rle": {"size": [640, 427], "counts": "WfR1e0Qc0`0A9F:F9G:G9F:G8I8H8H8I6I8H8H8H7I8I7H7I7K5N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2M3N2O1O2O0O100O1O100O100O100O2O0O100O100O2O2M2O2M2O2M2O2M2O2M2O2M2O2N101O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1O1O1O1O1N5L3M4L4L3M4L4L3M4L4K4M4L4L4L3M4dJaDo1c;iKRGf3U=_O8H1N10000000000000001O0O1000000000000O101O0O10000O10000O101O000O10000O1000001O0000000000000000001O00000000000000001O0000000000000O101O0000000000000000001O00000000000000001O00000000000000001O000000000000000O101O00N2L4L4N2000000000000O100000000000000O100000000000O10O101O000000000O1000001O000000000000001O00000000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001gJaLoI`3l5iLoIX3m5QMmIo2o5ZMlIf2Q6cMiI^2R6kMiIU2T6SNgIn1T6\\NfId1W6dNdI]1W6mNcIS1Z6UOaIl0\\6\\O^Id0`6CZI?c6IWI7g60TI0j68oHJn6=mHCR7d0hH]OU7j0fHVOX7R1aHPO\\7W1_HiN_7^1\\HcNa7e1XH\\Nf7k1UHUNj7R2PHoMm7Y2lGiMQ8`2hG`MV8h2dGYMY8P3`GPM^8[7N2M2N3M2N3M3M2O2M2N4M2M3M3N2M3M3N2M4L3N2M3N2M3M3N3L3M3N2M3M3Nf0YOo0`NQM"}, "label": "an adult elephant standing with a baby elephant"}]}
{"id": 322630, "image": "COCO_train2014_000000322630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult elephant\"."}], "task": "refering", "answer_template": "The \"the adult elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310], "bbox": [0.12770491803278688, 0.002125, 1.0, 0.881046875], "iscrowd": 0, "area": 151075.8652500001, "rle": {"size": [640, 427], "counts": "WfR1e0Qc0`0A9F:F9G:G9F:G8I8H8H8I6I8H8H8H7I8I7H7I7K5N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2M3N2O1O2O0O100O1O100O100O100O2O0O100O100O2O2M2O2M2O2M2O2M2O2M2O2M2O2N101O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1O1O1O1O1N5L3M4L4L3M4L4L3M4L4K4M4L4L4L3M4dJaDo1c;iKRGf3U=_O8H1N10000000000000001O0O1000000000000O101O0O10000O10000O101O000O10000O1000001O0000000000000000001O00000000000000001O0000000000000O101O0000000000000000001O00000000000000001O00000000000000001O000000000000000O101O00N2L4L4N2000000000000O100000000000000O100000000000O10O101O000000000O1000001O000000000000001O00000000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001gJaLoI`3l5iLoIX3m5QMmIo2o5ZMlIf2Q6cMiI^2R6kMiIU2T6SNgIn1T6\\NfId1W6dNdI]1W6mNcIS1Z6UOaIl0\\6\\O^Id0`6CZI?c6IWI7g60TI0j68oHJn6=mHCR7d0hH]OU7j0fHVOX7R1aHPO\\7W1_HiN_7^1\\HcNa7e1XH\\Nf7k1UHUNj7R2PHoMm7Y2lGiMQ8`2hG`MV8h2dGYMY8P3`GPM^8[7N2M2N3M2N3M3M2O2M2N4M2M3M3N2M3M3N2M4L3N2M3N2M3M3N3L3M3N2M3M3Nf0YOo0`NQM"}, "label": "the adult elephant"}]}
{"id": 322630, "image": "COCO_train2014_000000322630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant with trunk currled and a peice of grass\"."}], "task": "refering", "answer_template": "The \"a baby elephant with trunk currled and a peice of grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 257, 258, 259, 260, 272, 273, 274, 275, 287, 288, 289, 290, 302, 303, 304, 305], "bbox": [0.05039812646370023, 0.5267968749999999, 0.3931850117096019, 0.9079531249999999], "iscrowd": 0, "area": 25539.491450000005, "rle": {"size": [640, 427], "counts": "cT>m0gb0d0F:F3M3N2M3M3N2O1N2O1O1O2N1O1O1N2O1N01O01O0010O01O0010O01O0010O0001O010O0010O01O0010O0001O010bB[M`9e2VFfMi9[2mEPNQ:P2eE[N[:e1ZEfNe:Z1QEQOn:P1gD\\OW;d0_DGa;8VD2i;OSD6l;IQD;n;FnC>R<AkCc0U<]OgCh0X<WOeCm0[<SOaCQ1_<nN]CW1b<jNZCZ1f<P300000001N10000000000O10000000000O10000000000O100000001O0O10000000000O10000000000O10000000000O1000001O000O10000000000O1000M3M3M3L5L3L4M3L4M3M3L4M3L4M3L4M3^Ob0\\Oe0TOk0_MncQ5"}, "label": "a baby elephant with trunk currled and a peice of grass"}]}
{"id": 322630, "image": "COCO_train2014_000000322630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby elephant standing next to the larger elephant\"."}], "task": "refering", "answer_template": "The \"the baby elephant standing next to the larger elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 257, 258, 259, 260, 272, 273, 274, 275, 287, 288, 289, 290, 302, 303, 304, 305], "bbox": [0.05039812646370023, 0.5267968749999999, 0.3931850117096019, 0.9079531249999999], "iscrowd": 0, "area": 25539.491450000005, "rle": {"size": [640, 427], "counts": "cT>m0gb0d0F:F3M3N2M3M3N2O1N2O1O1O2N1O1O1N2O1N01O01O0010O01O0010O01O0010O0001O010O0010O01O0010O0001O010bB[M`9e2VFfMi9[2mEPNQ:P2eE[N[:e1ZEfNe:Z1QEQOn:P1gD\\OW;d0_DGa;8VD2i;OSD6l;IQD;n;FnC>R<AkCc0U<]OgCh0X<WOeCm0[<SOaCQ1_<nN]CW1b<jNZCZ1f<P300000001N10000000000O10000000000O10000000000O100000001O0O10000000000O10000000000O10000000000O1000001O000O10000000000O1000M3M3M3L5L3L4M3L4M3M3L4M3L4M3L4M3^Ob0\\Oe0TOk0_MncQ5"}, "label": "the baby elephant standing next to the larger elephant"}]}
{"id": 453704, "image": "COCO_train2014_000000453704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a fire truck that has red and white stripes with the number 28 on the back\"."}], "task": "refering", "answer_template": "The \"the back of a fire truck that has red and white stripes with the number 28 on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.24025, 0.20014117647058824, 0.722265625, 0.8563294117647059], "iscrowd": 0, "area": 62383.71120000001, "rle": {"size": [425, 640], "counts": "XVP2P1l:]1bN^1L4000000000000O10000000000000001O0000000000000O10000000000000000000000000000O1000000BZLYGf3d8a0N2N5J5L5K5J6K5K5_Oa0QOo0L4M2M4L4M1N100O100O2O0O100O100O100O100O100O100O100O2N100O100O100O100O100O100000O0100000O1000000O10O10O1000000O100000O0100O1O1O100O1O100O001O100O1O100O1O1O010O1O100O1001O000O100000001O01O001O000010O01O001O001O01O01O001O001O010O00001O001O010O001O00001O010O001O001O00101N1O2O1N101O0O2O1O0O2O001N2O001N101O1N1000O0100O1000O010000O10O01O1O1O1O001O1O1O0O2O1O1O1O002N1O1O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O2N1H8H8G9G:G8H8O0000001O00000000001O00000000001O0000000000001O00000000001O000O1000001O00000000001O00T1lNhdY2"}, "label": "the back of a fire truck that has red and white stripes with the number 28 on the back"}]}
{"id": 453704, "image": "COCO_train2014_000000453704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back end of a fire truck with the number 28\"."}], "task": "refering", "answer_template": "The \"back end of a fire truck with the number 28\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.24025, 0.20014117647058824, 0.722265625, 0.8563294117647059], "iscrowd": 0, "area": 62383.71120000001, "rle": {"size": [425, 640], "counts": "XVP2P1l:]1bN^1L4000000000000O10000000000000001O0000000000000O10000000000000000000000000000O1000000BZLYGf3d8a0N2N5J5L5K5J6K5K5_Oa0QOo0L4M2M4L4M1N100O100O2O0O100O100O100O100O100O100O100O2N100O100O100O100O100O100000O0100000O1000000O10O10O1000000O100000O0100O1O1O100O1O100O001O100O1O100O1O1O010O1O100O1001O000O100000001O01O001O000010O01O001O001O01O01O001O001O010O00001O001O010O001O00001O010O001O001O00101N1O2O1N101O0O2O1O0O2O001N2O001N101O1N1000O0100O1000O010000O10O01O1O1O1O001O1O1O0O2O1O1O1O002N1O1O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O2N1H8H8G9G:G8H8O0000001O00000000001O00000000001O0000000000001O00000000001O000O1000001O00000000001O00T1lNhdY2"}, "label": "back end of a fire truck with the number 28"}]}
{"id": 453704, "image": "COCO_train2014_000000453704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fire truck on the right\"."}], "task": "refering", "answer_template": "The \"the fire truck on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252], "bbox": [0.7073437499999999, 0.45435294117647057, 1.0, 0.721764705882353], "iscrowd": 0, "area": 15002.360399999998, "rle": {"size": [425, 640], "counts": "_Vl51X=2N2N1O2N2iCKV;<aDH];?YDFe;Q1N2N2N001O1O001O00001O00001O0000001O00001O00000000000ZEXNk9h1TF\\Ni9e1TF_Nj9a1TFcNj9]1UFgNh9Y1VFmNf9S1YFPOe9Q1YFQOf9Q1WFPOi9Q1TFROk9o0SFSOl9o0QFROo9U2O1O1O001O1O001O1O1O001O0000000000000000O1RNnFDS97YG_Oj8>aGWOb8h0P2O1O1O1O100O1G9M3N200001O000000001O0TD\\OS;d0lD_OR;b0kDCQ;>nDGn:;nDKn:[1M4L4L3M4L4L3M4L3M:F1O001O001O001O00001O001O00001O0000001O0000000000000000000000O100O10000O100O10000O100O101O0O2O0O101O0O1O2O0O1O2O0O1O2O0O2O00000O1000000O100000000O1000001N2O1O009Ed0[Oe0[Oe0\\OhG"}, "label": "the fire truck on the right"}]}
{"id": 453704, "image": "COCO_train2014_000000453704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ladder on a firetruck\"."}], "task": "refering", "answer_template": "The \"ladder on a firetruck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252], "bbox": [0.7073437499999999, 0.45435294117647057, 1.0, 0.721764705882353], "iscrowd": 0, "area": 15002.360399999998, "rle": {"size": [425, 640], "counts": "_Vl51X=2N2N1O2N2iCKV;<aDH];?YDFe;Q1N2N2N001O1O001O00001O00001O0000001O00001O00000000000ZEXNk9h1TF\\Ni9e1TF_Nj9a1TFcNj9]1UFgNh9Y1VFmNf9S1YFPOe9Q1YFQOf9Q1WFPOi9Q1TFROk9o0SFSOl9o0QFROo9U2O1O1O001O1O001O1O1O001O0000000000000000O1RNnFDS97YG_Oj8>aGWOb8h0P2O1O1O1O100O1G9M3N200001O000000001O0TD\\OS;d0lD_OR;b0kDCQ;>nDGn:;nDKn:[1M4L4L3M4L4L3M4L3M:F1O001O001O001O00001O001O00001O0000001O0000000000000000000000O100O10000O100O10000O100O101O0O2O0O101O0O1O2O0O1O2O0O1O2O0O2O00000O1000000O100000000O1000001N2O1O009Ed0[Oe0[Oe0\\OhG"}, "label": "ladder on a firetruck"}]}
{"id": 371786, "image": "COCO_train2014_000000371786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the balding male cook\"."}], "task": "refering", "answer_template": "The \"the balding male cook\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 59, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 313, 314, 315, 316, 335, 336, 337, 338], "bbox": [0.366234375, 0.06898584905660378, 0.79053125, 0.999316037735849], "iscrowd": 0, "area": 47020.04870000001, "rle": {"size": [424, 640], "counts": "cTQ31U=1O1O1O1O2N1N2O1O1O2N1O1N2O1O2N1O1001O010O1O0012N7H6K0O01000O010O1000O01000O0100O10O1O001O1O0O2O001O1O001NmKZOgKf0S4CkK<o3MnK3m35QLJn3:oKER4>lKAT4b0iK^OW4e0fK[OZ4h0dKWO\\4l0aKTO_4o0_KoNb4T1[KlNe4\\1TKcNl4h1iJYNV5h1iJYNV5h1jJXNU5i1jJWNV5j1jJVNU5k1jJVNU5k1kJTNU5n1iJSNV5o1hJRNW5P2hJoMX5S2fJnMY5T2fJlMY5V2eJjM[5X2cJiM\\5Z2bJfM]5\\2aJdM_5^2`JbM`5_2^JbMc5^2\\JbMd5_2ZJaMg5`2WJaMi5_2WJaMj5_2TJaMm5`2RJ`Mn5`2QJaMo5`2PJ_MQ6b2mI_MT6_2lIbMT6^2lIaMU6_2jIbMV6]2kIcMU6]2jIcMX6[2iIeMW6[2hIfMX6Z2hIfMX6Y2hIgMY6Y2fIhM[6V2fIjMZ6V2eIkM[6T2fIlMZ6T2eIlM\\6T2dIlM]6R2cIoM]6Q2VI\\Nj6c1_HUOa7X32OdJfJj2Z5TMiJk2W5RMnJl2Q5RMSKm2o4nLUKR3l4jLXKT3j4gLZKY3f4fL[K[3e4cL\\K^3d4`L^K_3c4_L^Kb3b4\\L_Kd3b4YL`Kh3_4WLcKi3]4ULdKk3]4SLdKn3\\4PLeKQ4[4mKfKS4[4kKfKV4Y4iKgKX4Z4eKgK]4Y4aKhK`4X4^KhKc4Y4[KgKg4X4XKiKh4X4VKhKl4X4RKhKP5X4nJiKR5W4lJjKV5V4hJjKZ5V4dJkK]5U4aJkK`5U4_JkKc5U4[JlKf5T4XJlKj5T4TJlKm5T4QJmKQ6e51003K5L4TLUIQ1o6RNlIm1X6UMaJk2n71O001O00001O001O001O001O00001O001O001O001O0O101O001O001O00001O001O001O001O001O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1L4J6K5K6I7J6I7J5K6I7J6J5J7TLdHd1b7ZN_Hb1f7\\N[Ha1i7\\NZH_1j7^NYH_1k7]NXH`1l7\\NWHa1m7[NVHb1n7ZNVHb1m7[NVHb1n7ZNUHc1o7YNTHd1P8YNRHc1S8YNPHd1S8YNRHb1R8ZN`HT1d7hNjHi0[7SOjHh0Z7TOkHf0o9L5J6J6KmUg1"}, "label": "the balding male cook"}]}
{"id": 371786, "image": "COCO_train2014_000000371786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was cooking\"."}], "task": "refering", "answer_template": "The \"a man was cooking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 59, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 313, 314, 315, 316, 335, 336, 337, 338], "bbox": [0.366234375, 0.06898584905660378, 0.79053125, 0.999316037735849], "iscrowd": 0, "area": 47020.04870000001, "rle": {"size": [424, 640], "counts": "cTQ31U=1O1O1O1O2N1N2O1O1O2N1O1N2O1O2N1O1001O010O1O0012N7H6K0O01000O010O1000O01000O0100O10O1O001O1O0O2O001O1O001NmKZOgKf0S4CkK<o3MnK3m35QLJn3:oKER4>lKAT4b0iK^OW4e0fK[OZ4h0dKWO\\4l0aKTO_4o0_KoNb4T1[KlNe4\\1TKcNl4h1iJYNV5h1iJYNV5h1jJXNU5i1jJWNV5j1jJVNU5k1jJVNU5k1kJTNU5n1iJSNV5o1hJRNW5P2hJoMX5S2fJnMY5T2fJlMY5V2eJjM[5X2cJiM\\5Z2bJfM]5\\2aJdM_5^2`JbM`5_2^JbMc5^2\\JbMd5_2ZJaMg5`2WJaMi5_2WJaMj5_2TJaMm5`2RJ`Mn5`2QJaMo5`2PJ_MQ6b2mI_MT6_2lIbMT6^2lIaMU6_2jIbMV6]2kIcMU6]2jIcMX6[2iIeMW6[2hIfMX6Z2hIfMX6Y2hIgMY6Y2fIhM[6V2fIjMZ6V2eIkM[6T2fIlMZ6T2eIlM\\6T2dIlM]6R2cIoM]6Q2VI\\Nj6c1_HUOa7X32OdJfJj2Z5TMiJk2W5RMnJl2Q5RMSKm2o4nLUKR3l4jLXKT3j4gLZKY3f4fL[K[3e4cL\\K^3d4`L^K_3c4_L^Kb3b4\\L_Kd3b4YL`Kh3_4WLcKi3]4ULdKk3]4SLdKn3\\4PLeKQ4[4mKfKS4[4kKfKV4Y4iKgKX4Z4eKgK]4Y4aKhK`4X4^KhKc4Y4[KgKg4X4XKiKh4X4VKhKl4X4RKhKP5X4nJiKR5W4lJjKV5V4hJjKZ5V4dJkK]5U4aJkK`5U4_JkKc5U4[JlKf5T4XJlKj5T4TJlKm5T4QJmKQ6e51003K5L4TLUIQ1o6RNlIm1X6UMaJk2n71O001O00001O001O001O001O00001O001O001O001O0O101O001O001O00001O001O001O001O001O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1L4J6K5K6I7J6I7J5K6I7J6J5J7TLdHd1b7ZN_Hb1f7\\N[Ha1i7\\NZH_1j7^NYH_1k7]NXH`1l7\\NWHa1m7[NVHb1n7ZNVHb1m7[NVHb1n7ZNUHc1o7YNTHd1P8YNRHc1S8YNPHd1S8YNRHb1R8ZN`HT1d7hNjHi0[7SOjHh0Z7TOkHf0o9L5J6J6KmUg1"}, "label": "a man was cooking"}]}
{"id": 371786, "image": "COCO_train2014_000000371786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue shirt assists a chef at a grill\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue shirt assists a chef at a grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 61, 62, 63, 84, 85, 86, 87, 108, 109, 110, 111, 131, 132, 133, 134, 155, 156, 157, 178, 179, 180, 201, 202, 203, 224, 225, 226, 248, 249, 270, 271, 272, 293, 294, 316, 317, 338, 339, 340], "bbox": [0.6779375, 0.1168867924528302, 0.8665156249999999, 1.0], "iscrowd": 0, "area": 20547.012399999996, "rle": {"size": [424, 640], "counts": "Qic55n<5J7L4N101N1O2O1N1kCYO80l:i0hD\\O61P;R1iDTOU;a1N2O1N3M2O1N2N2OOXEcMf:b2M2O0O3M3N1N3N2M2O2N1O2N2N1O2N2QNeLmI\\3m5kLQJW3h5oLWJR3d5SM[Jo2^5XM`Jj2Y5\\MfJe2T5`MlJb2m4cMSK_2g4eMYK\\2a4iM_KY2U4QNkKP2n3VNRLl1j3VNVLl1>nKW1W2[Nl1<nKW1X2^Nk17PLX1W2cNi11RL[1V2eNh1MUL[1U2iNg1HVL]1U2mNd1DYL]1T2POd1_O[L^1S2TOc1ZO^L_1P2YOa1UObL_1o1]O`1POdL`1n1A_1lNfL`1l1F]1gNjL`1k1J\\1bNmL`1i1O[1]NoLb1g13Y1XNSMb1f18W1SN\\M[1_1g0Q1kMjMU1V1T1n0cMVN8e1Z2e64L5K5K2N2N2N2N2N1O1O1O001O1O1O001O1O001O1DaLmF`3c7YLdIV4X6WLZIo3b6]LPIi3k6cLiHb3R7\\1M5K5J6K5YLZIe0k6YOVI<V7AmH1a7M`HLh72ZHIk74WHGo77SHDS89nGCX89kGB[8;fGA`8<bG^Od8?^G]Oh8`0ZG[Ol8b0ZGUOl8g0n1M2M6K9G9F:G:FcoR1"}, "label": "a woman wearing a blue shirt assists a chef at a grill"}]}
{"id": 371786, "image": "COCO_train2014_000000371786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman working at a grill directly behind a man working on the same grill who is holding a spatula\"."}], "task": "refering", "answer_template": "The \"a woman working at a grill directly behind a man working on the same grill who is holding a spatula\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 61, 62, 63, 84, 85, 86, 87, 108, 109, 110, 111, 131, 132, 133, 134, 155, 156, 157, 178, 179, 180, 201, 202, 203, 224, 225, 226, 248, 249, 270, 271, 272, 293, 294, 316, 317, 338, 339, 340], "bbox": [0.6779375, 0.1168867924528302, 0.8665156249999999, 1.0], "iscrowd": 0, "area": 20547.012399999996, "rle": {"size": [424, 640], "counts": "Qic55n<5J7L4N101N1O2O1N1kCYO80l:i0hD\\O61P;R1iDTOU;a1N2O1N3M2O1N2N2OOXEcMf:b2M2O0O3M3N1N3N2M2O2N1O2N2N1O2N2QNeLmI\\3m5kLQJW3h5oLWJR3d5SM[Jo2^5XM`Jj2Y5\\MfJe2T5`MlJb2m4cMSK_2g4eMYK\\2a4iM_KY2U4QNkKP2n3VNRLl1j3VNVLl1>nKW1W2[Nl1<nKW1X2^Nk17PLX1W2cNi11RL[1V2eNh1MUL[1U2iNg1HVL]1U2mNd1DYL]1T2POd1_O[L^1S2TOc1ZO^L_1P2YOa1UObL_1o1]O`1POdL`1n1A_1lNfL`1l1F]1gNjL`1k1J\\1bNmL`1i1O[1]NoLb1g13Y1XNSMb1f18W1SN\\M[1_1g0Q1kMjMU1V1T1n0cMVN8e1Z2e64L5K5K2N2N2N2N2N1O1O1O001O1O1O001O1O001O1DaLmF`3c7YLdIV4X6WLZIo3b6]LPIi3k6cLiHb3R7\\1M5K5J6K5YLZIe0k6YOVI<V7AmH1a7M`HLh72ZHIk74WHGo77SHDS89nGCX89kGB[8;fGA`8<bG^Od8?^G]Oh8`0ZG[Ol8b0ZGUOl8g0n1M2M6K9G9F:G:FcoR1"}, "label": "a woman working at a grill directly behind a man working on the same grill who is holding a spatula"}]}
{"id": 371786, "image": "COCO_train2014_000000371786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person who is cut in half wearing a bright purple shirt\"."}], "task": "refering", "answer_template": "The \"the person who is cut in half wearing a bright purple shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229, 251, 252, 274, 275, 297, 298, 320, 321, 343], "bbox": [0.9021874999999999, 0.21797169811320755, 1.0, 0.9550471698113208], "iscrowd": 0, "area": 12149.116850000002, "rle": {"size": [424, 640], "counts": "^R_71U=3M3M3cDHf9i0hE\\OT:l0dEYOW:o0`EVO\\:R1ZEUOa:l1K4L4[J_M?c2nNQNn0R2POPNm0S2QOoMl0T2iNXNS1k1^NdN`1^1QNROk1Q1fM^OW2e0dM@[2[OcKeNP2U2[2bN]LTOV1^2[2_NdLmNo0i2[2ZNjLfNk0T3Z2VNROR2m0oM`Ne2]1\\MRNV3m1jLgMa3Z2`LcMb3]2_L_Md3`2^L\\Me3d2\\LXMg3h2QM^LQ3b3YMQLj2n3X300010O00100O001B>K4M4K5H8@?A`0G9M2O2M3M2O2M3M2O2M3M2O2M3M3NUM"}, "label": "the person who is cut in half wearing a bright purple shirt"}]}
{"id": 371786, "image": "COCO_train2014_000000371786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing purple shirt with back to camera\"."}], "task": "refering", "answer_template": "The \"man wearing purple shirt with back to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229, 251, 252, 274, 275, 297, 298, 320, 321, 343], "bbox": [0.9021874999999999, 0.21797169811320755, 1.0, 0.9550471698113208], "iscrowd": 0, "area": 12149.116850000002, "rle": {"size": [424, 640], "counts": "^R_71U=3M3M3cDHf9i0hE\\OT:l0dEYOW:o0`EVO\\:R1ZEUOa:l1K4L4[J_M?c2nNQNn0R2POPNm0S2QOoMl0T2iNXNS1k1^NdN`1^1QNROk1Q1fM^OW2e0dM@[2[OcKeNP2U2[2bN]LTOV1^2[2_NdLmNo0i2[2ZNjLfNk0T3Z2VNROR2m0oM`Ne2]1\\MRNV3m1jLgMa3Z2`LcMb3]2_L_Md3`2^L\\Me3d2\\LXMg3h2QM^LQ3b3YMQLj2n3X300010O00100O001B>K4M4K5H8@?A`0G9M2O2M3M2O2M3M2O2M3M2O2M3M3NUM"}, "label": "man wearing purple shirt with back to camera"}]}
{"id": 388171, "image": "COCO_train2014_000000388171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ship in the near distance sailing to the shore\"."}], "task": "refering", "answer_template": "The \"a ship in the near distance sailing to the shore\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153], "bbox": [0.43134375, 0.24820930232558142, 0.688390625, 0.45246511627906977], "iscrowd": 0, "area": 7498.639850000003, "rle": {"size": [430, 640], "counts": "hQd31Z=3N20000SCO]<1bC1]<OcC2\\<NcC4\\<LdC4\\<`0O1O1O3M2N3M00000000000000000001O00000000000000000000K5K5J6J6K5J6K5J61O1O1O1O1O1O1O:F2N3M3M4L5K2N3M2N2N2N2N00000000XOAaD?^;H\\D8d;i00000000000O1000000000000O1000000000000O11O2N2N1O2N4L3M3M3M000000000000001O00000000000000O1J6J6J7I6G9E;00O100001O5K4L5K5K4L5K4M1N2N2N3M2N2N2N1O0000000000000001O01O000000000000002N2N2N2N1O2Oe0ZOT\\c2"}, "label": "a ship in the near distance sailing to the shore"}]}
{"id": 388171, "image": "COCO_train2014_000000388171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat letting off steam\"."}], "task": "refering", "answer_template": "The \"a boat letting off steam\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153], "bbox": [0.43134375, 0.24820930232558142, 0.688390625, 0.45246511627906977], "iscrowd": 0, "area": 7498.639850000003, "rle": {"size": [430, 640], "counts": "hQd31Z=3N20000SCO]<1bC1]<OcC2\\<NcC4\\<LdC4\\<`0O1O1O3M2N3M00000000000000000001O00000000000000000000K5K5J6J6K5J6K5J61O1O1O1O1O1O1O:F2N3M3M4L5K2N3M2N2N2N2N00000000XOAaD?^;H\\D8d;i00000000000O1000000000000O1000000000000O11O2N2N1O2N4L3M3M3M000000000000001O00000000000000O1J6J6J7I6G9E;00O100001O5K4L5K5K4L5K4M1N2N2N3M2N2N2N1O0000000000000001O01O000000000000002N2N2N2N1O2Oe0ZOT\\c2"}, "label": "a boat letting off steam"}]}
{"id": 76882, "image": "COCO_train2014_000000076882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl\"."}], "task": "refering", "answer_template": "The \"the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 327, 328, 329, 330, 331, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.24318749999999997, 0.3024532710280374, 0.9858750000000001, 0.9953271028037384], "iscrowd": 0, "area": 64321.35420000001, "rle": {"size": [428, 640], "counts": "hcQ22X=2E;K5001O000000001O000000001O0000000000001O00000000000000000000001O000000000000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O001O001O1O001O001OWl?0hS@100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1ROcNlE^1n9hNQFY1h9oNVFR1c9VO[Fk0a9ZOUFo0i9[1N2O1N2N2O1N2N2N2O1N2N2N2O1N2O1O100O1O100O1O100O1O100O1O100O1O100O1O1O1O1O1O1SObKkH_4i6oKTIR4e6WLXIj3g6YLVIh3h6\\LUIe3j6^LSIc3k6aLRI`3l6a1O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N201N100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O10000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O11O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1`LmJeL89l4o2TK\\L:a0c4Q3ZKSL<h0[4R3aKkK<P1T4R3hKcK=W1l3S3[MiLf2T3_MiLb2T3dMhL\\2V3jMfLW2W3oMeLR2X3SNeLn1Y3T4M4L4L5K4L4L5K4L5K4L5K4L4L5K4L5K4L5K4L4[OUb3"}, "label": "the girl"}]}
{"id": 76882, "image": "COCO_train2014_000000076882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady wearing grey jacket looking to grab an item\"."}], "task": "refering", "answer_template": "The \"lady wearing grey jacket looking to grab an item\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 327, 328, 329, 330, 331, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.24318749999999997, 0.3024532710280374, 0.9858750000000001, 0.9953271028037384], "iscrowd": 0, "area": 64321.35420000001, "rle": {"size": [428, 640], "counts": "hcQ22X=2E;K5001O000000001O000000001O0000000000001O00000000000000000000001O000000000000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O001O001O1O001O001OWl?0hS@100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1ROcNlE^1n9hNQFY1h9oNVFR1c9VO[Fk0a9ZOUFo0i9[1N2O1N2N2O1N2N2N2O1N2N2N2O1N2O1O100O1O100O1O100O1O100O1O100O1O100O1O1O1O1O1O1SObKkH_4i6oKTIR4e6WLXIj3g6YLVIh3h6\\LUIe3j6^LSIc3k6aLRI`3l6a1O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N201N100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O10000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O11O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1`LmJeL89l4o2TK\\L:a0c4Q3ZKSL<h0[4R3aKkK<P1T4R3hKcK=W1l3S3[MiLf2T3_MiLb2T3dMhL\\2V3jMfLW2W3oMeLR2X3SNeLn1Y3T4M4L4L5K4L4L5K4L5K4L5K4L4L5K4L5K4L5K4L4[OUb3"}, "label": "lady wearing grey jacket looking to grab an item"}]}
{"id": 76882, "image": "COCO_train2014_000000076882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a north face jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing a north face jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 277, 278, 279, 280, 302], "bbox": [0.0, 0.004485981308411215, 0.3636875, 0.8719158878504673], "iscrowd": 0, "area": 48193.779899999994, "rle": {"size": [428, 640], "counts": "V8i1R8b3_L`3nNS1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O001O0000001O0000001O0000001O0000001O0000001O0000001O001O1O2N1O1O1O1O1O2N1O1O1O1O2N00000000000000001O000O10000000ZLjNVKV1i4mNUKS1i4POVKP1i4ROVKn0i4UOUKk0i4XOVKh0i4ZOVKf0h4^OVKb0h4AWK?f4EYK;e4HZK8d4LRK<k4HoJ=o4FmJ=Q5GkJ;R5IlJ8R5KnJ4P5OoJ1n44QKKm48RKHl4<RKDk4b0SK]Ok4h0RKXOl4l0SKSOj4S1SKmNk4W1SKiNk4\\1SKcNj4c1SK]Nk4g1TKXNj4m1SKSNj4R2UKmMi4X2TKhMj4\\2TKdMi4b2SK_Mj4g2PK\\Mm4i2mJ[Mo4l2kJWMR5n2iJUMT5c5M3M3M3M3L4M3M3M3M3M3M3J6TOl0\\Od0O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O10000001O0000001O00000000O1O1O1O1O1O100O1O1O1O1O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100N2N2N2N2N2O1N2N2N2N2O1N2N2N2NPaZ5"}, "label": "a man wearing a north face jacket"}]}
{"id": 76882, "image": "COCO_train2014_000000076882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man looks down at the grown while wearing a north face jacket\"."}], "task": "refering", "answer_template": "The \"a man looks down at the grown while wearing a north face jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 277, 278, 279, 280, 302], "bbox": [0.0, 0.004485981308411215, 0.3636875, 0.8719158878504673], "iscrowd": 0, "area": 48193.779899999994, "rle": {"size": [428, 640], "counts": "V8i1R8b3_L`3nNS1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O001O0000001O0000001O0000001O0000001O0000001O0000001O001O1O2N1O1O1O1O1O2N1O1O1O1O2N00000000000000001O000O10000000ZLjNVKV1i4mNUKS1i4POVKP1i4ROVKn0i4UOUKk0i4XOVKh0i4ZOVKf0h4^OVKb0h4AWK?f4EYK;e4HZK8d4LRK<k4HoJ=o4FmJ=Q5GkJ;R5IlJ8R5KnJ4P5OoJ1n44QKKm48RKHl4<RKDk4b0SK]Ok4h0RKXOl4l0SKSOj4S1SKmNk4W1SKiNk4\\1SKcNj4c1SK]Nk4g1TKXNj4m1SKSNj4R2UKmMi4X2TKhMj4\\2TKdMi4b2SK_Mj4g2PK\\Mm4i2mJ[Mo4l2kJWMR5n2iJUMT5c5M3M3M3M3L4M3M3M3M3M3M3J6TOl0\\Od0O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O10000001O0000001O00000000O1O1O1O1O1O100O1O1O1O1O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100N2N2N2N2N2O1N2N2N2N2O1N2N2N2NPaZ5"}, "label": "a man looks down at the grown while wearing a north face jacket"}]}
{"id": 412756, "image": "COCO_train2014_000000412756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid pitching a baseball\"."}], "task": "refering", "answer_template": "The \"a kid pitching a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 64, 65, 66, 80, 81, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 111, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 203, 204, 205, 218, 219, 220, 233, 234, 235, 249, 250, 264, 265], "bbox": [0.3044496487119438, 0.163890625, 0.9211475409836065, 0.76025], "iscrowd": 0, "area": 34089.70560000001, "rle": {"size": [640, 427], "counts": "d[a26ic0;E3M1O2N2N2N1O2N2O0O2N2O1O001O1O001O1O00100000O10000O10001O0100O1O2N1O100O100O10000O2O000O10O100O100O1O2O0O100O100O0010O001O0010O00000O1O1O2N1O100O1O2N1O1O1O1O2O0O1O1O1O00O101O001O2N2N2N2M3N2N1O1O1O1O1O1N2O1O1i_OmNOBX;b1UDJ7gNb;`1hCb0;QNj;^1\\CZ1>[MT<\\1oBR2b0dL\\<e5XC`Jd<k5QCYJk<c6K4L00001O0O101O0XEbHn8^7QGcHo8^7oFcHQ9]7nFcHS9]7lFdHT9]7jFdHW9[7iFeHW9[7hFfHX9[7fFfHZ9Z7eFgH\\9X7cFiH]9X7aFiH_9W7`FjH`9V7_FkHb9U7\\FlHd9T7[FmHe9S7ZFnHf9S7YFmHh9R7VFPIj9P7TFRIl9o6PFTIQ:k6mEWIS:i6jEZIV:h6fEZIZ:P7YESIh:R80000001O000O101O0000001O0000001O001O001O00nNdGaF[8^9gGaFY8_9iG]FY8c9kGXFV8g9RHPFP8P:XMUFPMU:P3QF^L\\:a3jEnKa:R4eEcK`:]4fE^KX:c4nEWKR:h4VFRKi9n4]FmJb9S5dFgJ[9Z5kFaJT9_5RG\\Jm8c5ZGWJe8j5eGmIZ8S6i201O1YOWBUKj=j4dBiJ]=V5g0O1O1O1N2O1O001N2O1O1O1O1O24KQMTAo0g>PO[AT1`>kNbAY1Y>eNiA`1R>_NPB`1P>`NPB`1P>`NQB`1n=`NRB`1n=`NSB`1l=`NUB_1k=aNUB`1j=`NWB_1i=aNXB_1h=_NZB`1f=`N\\B^1d=aN^B_1a=`NaB_1_=UNmAkNf0o2]=TNoAkNf0P3\\=RNQBlNe0R3Z=PNSBmNe0Q3Y=PNTBmNe0R3X=oMUBmNe0T3V=lM\\CR2d<mM^CS2b<kM`CT2a<jMbCU2^<hMeCW2S?01N1O1O2M3N1O2N2M2O2N2M2O2N2N1N3N2N2N1N3N2N1N3N2N1POQ1L4L3M4Lknd0"}, "label": "a kid pitching a baseball"}]}
{"id": 412756, "image": "COCO_train2014_000000412756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy throwing ball\"."}], "task": "refering", "answer_template": "The \"boy throwing ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 64, 65, 66, 80, 81, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 111, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 203, 204, 205, 218, 219, 220, 233, 234, 235, 249, 250, 264, 265], "bbox": [0.3044496487119438, 0.163890625, 0.9211475409836065, 0.76025], "iscrowd": 0, "area": 34089.70560000001, "rle": {"size": [640, 427], "counts": "d[a26ic0;E3M1O2N2N2N1O2N2O0O2N2O1O001O1O001O1O00100000O10000O10001O0100O1O2N1O100O100O10000O2O000O10O100O100O1O2O0O100O100O0010O001O0010O00000O1O1O2N1O100O1O2N1O1O1O1O2O0O1O1O1O00O101O001O2N2N2N2M3N2N1O1O1O1O1O1N2O1O1i_OmNOBX;b1UDJ7gNb;`1hCb0;QNj;^1\\CZ1>[MT<\\1oBR2b0dL\\<e5XC`Jd<k5QCYJk<c6K4L00001O0O101O0XEbHn8^7QGcHo8^7oFcHQ9]7nFcHS9]7lFdHT9]7jFdHW9[7iFeHW9[7hFfHX9[7fFfHZ9Z7eFgH\\9X7cFiH]9X7aFiH_9W7`FjH`9V7_FkHb9U7\\FlHd9T7[FmHe9S7ZFnHf9S7YFmHh9R7VFPIj9P7TFRIl9o6PFTIQ:k6mEWIS:i6jEZIV:h6fEZIZ:P7YESIh:R80000001O000O101O0000001O0000001O001O001O00nNdGaF[8^9gGaFY8_9iG]FY8c9kGXFV8g9RHPFP8P:XMUFPMU:P3QF^L\\:a3jEnKa:R4eEcK`:]4fE^KX:c4nEWKR:h4VFRKi9n4]FmJb9S5dFgJ[9Z5kFaJT9_5RG\\Jm8c5ZGWJe8j5eGmIZ8S6i201O1YOWBUKj=j4dBiJ]=V5g0O1O1O1N2O1O001N2O1O1O1O1O24KQMTAo0g>PO[AT1`>kNbAY1Y>eNiA`1R>_NPB`1P>`NPB`1P>`NQB`1n=`NRB`1n=`NSB`1l=`NUB_1k=aNUB`1j=`NWB_1i=aNXB_1h=_NZB`1f=`N\\B^1d=aN^B_1a=`NaB_1_=UNmAkNf0o2]=TNoAkNf0P3\\=RNQBlNe0R3Z=PNSBmNe0Q3Y=PNTBmNe0R3X=oMUBmNe0T3V=lM\\CR2d<mM^CS2b<kM`CT2a<jMbCU2^<hMeCW2S?01N1O1O2M3N1O2N2M2O2N2M2O2N2N1N3N2N2N1N3N2N1N3N2N1POQ1L4L3M4Lknd0"}, "label": "boy throwing ball"}]}
{"id": 412756, "image": "COCO_train2014_000000412756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in back ,\"."}], "task": "refering", "answer_template": "The \"the boy in back ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 20, 21, 22, 35, 36, 37, 38, 50, 51, 52, 53, 65, 66, 67, 68, 81, 82, 83, 84, 97, 98, 99], "bbox": [0.3611241217798595, 0.030156250000000002, 0.6300468384074941, 0.305546875], "iscrowd": 0, "area": 10851.115850000002, "rle": {"size": [640, 427], "counts": "X[P31lc05K4L4K5L4L5ZNVOR@k0l?[On_Og0P`0Ai_O`0U`0K`_O8]`0Ja_O8]`0Jb_O7]`0Ja_O8]`0Jb_O8[`0Jc_O9Z`0Hf_O:W`0Hg_O:W`0Hg_O;V`0Gi_O;T`0Hi_O;U`0Gi_O;U`0Hg_O;W`0Gf_O<X`0h1M2O2N2N2N2M3O1N101N2O1N2N101N1O100O1O100O1O101N1O10000O10ZLk@`2T?_MPA_2P?`MRA_2n>_MVA_2i>`MZA`2e>^M\\Ac2d>[M]Af2b>YM_Ah2a>WM_Aj2`>jLlAV3_?0000O100000000000000O10000000000]Od_ObM\\`0^2g_O_MY`0a2k_OZMV`0k1a_OhN;ZOV`0n1l_OYN2ET`0R2o_ORNd`0n1^_OnMd`0Q2g00000001O1O1O1O5K;E00N2N2iMc^Oh1]a0UNk^Oe1Wa0XNo^Oc1Sa0ZNT_O`1ka0L4K6J5K6K5J7I5K2O1N2NoUS3"}, "label": "the boy in back ,"}]}
{"id": 412756, "image": "COCO_train2014_000000412756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the background\"."}], "task": "refering", "answer_template": "The \"the boy in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 20, 21, 22, 35, 36, 37, 38, 50, 51, 52, 53, 65, 66, 67, 68, 81, 82, 83, 84, 97, 98, 99], "bbox": [0.3611241217798595, 0.030156250000000002, 0.6300468384074941, 0.305546875], "iscrowd": 0, "area": 10851.115850000002, "rle": {"size": [640, 427], "counts": "X[P31lc05K4L4K5L4L5ZNVOR@k0l?[On_Og0P`0Ai_O`0U`0K`_O8]`0Ja_O8]`0Jb_O7]`0Ja_O8]`0Jb_O8[`0Jc_O9Z`0Hf_O:W`0Hg_O:W`0Hg_O;V`0Gi_O;T`0Hi_O;U`0Gi_O;U`0Hg_O;W`0Gf_O<X`0h1M2O2N2N2N2M3O1N101N2O1N2N101N1O100O1O100O1O101N1O10000O10ZLk@`2T?_MPA_2P?`MRA_2n>_MVA_2i>`MZA`2e>^M\\Ac2d>[M]Af2b>YM_Ah2a>WM_Aj2`>jLlAV3_?0000O100000000000000O10000000000]Od_ObM\\`0^2g_O_MY`0a2k_OZMV`0k1a_OhN;ZOV`0n1l_OYN2ET`0R2o_ORNd`0n1^_OnMd`0Q2g00000001O1O1O1O5K;E00N2N2iMc^Oh1]a0UNk^Oe1Wa0XNo^Oc1Sa0ZNT_O`1ka0L4K6J5K6K5J7I5K2O1N2NoUS3"}, "label": "the boy in the background"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop computer in front of the man seated to the left in the picture\"."}], "task": "refering", "answer_template": "The \"the laptop computer in front of the man seated to the left in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.19340625, 0.7348243559718969, 0.554734375, 0.9842622950819672], "iscrowd": 0, "area": 19148.8605, "rle": {"size": [427, 640], "counts": "eSd12V=3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3N2N2O1O1000000O2O000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000000000000000000000O1000000000000000000000000O10000000000000000000000000000000000000000000I7H8H8H8H8H8H8H8H8H8I7H8H]af3"}, "label": "the laptop computer in front of the man seated to the left in the picture"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop on left side\"."}], "task": "refering", "answer_template": "The \"laptop on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.19340625, 0.7348243559718969, 0.554734375, 0.9842622950819672], "iscrowd": 0, "area": 19148.8605, "rle": {"size": [427, 640], "counts": "eSd12V=3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3N2N2O1O1000000O2O000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000000000000000000000O1000000000000000000000000O10000000000000000000000000000000000000000000I7H8H8H8H8H8H8H8H8H8I7H8H]af3"}, "label": "laptop on left side"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white computer being used\"."}], "task": "refering", "answer_template": "The \"a black and white computer being used\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.6191093750000001, 0.6317564402810304, 0.993265625, 0.9898594847775175], "iscrowd": 0, "area": 26733.553600000007, "rle": {"size": [427, 640], "counts": "WaU51V=4L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4L4M3L4L4L4M3L4L4L4M3L4L4M3O1000000O100000000O100000000O100000000O10001O00000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O2O000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O10000000000O100000000O1000001O0O100000000OO2M3L4L4L4L3M4M3L4L4L3M4L4M3L4L3M4L4L4M3L3M4L4L4L4M3L3M4L4L4L4M2M4L4L4L7J9Fi\\1"}, "label": "a black and white computer being used"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop sitting in front of the guy wearing the white shirt\"."}], "task": "refering", "answer_template": "The \"the laptop sitting in front of the guy wearing the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.6191093750000001, 0.6317564402810304, 0.993265625, 0.9898594847775175], "iscrowd": 0, "area": 26733.553600000007, "rle": {"size": [427, 640], "counts": "WaU51V=4L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4L4M3L4L4L4M3L4L4L4M3L4L4M3O1000000O100000000O100000000O100000000O10001O00000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O2O000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O10000000000O100000000O1000001O0O100000000OO2M3L4L4L4L3M4M3L4L4L3M4L4M3L4L3M4L4L4M3L3M4L4L4L4M3L3M4L4L4L4M2M4L4L4L7J9Fi\\1"}, "label": "the laptop sitting in front of the guy wearing the white shirt"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with an american flag on his sleeve\"."}], "task": "refering", "answer_template": "The \"a man with an american flag on his sleeve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.003, 0.013489461358313817, 0.280375, 0.9977517564402809], "iscrowd": 0, "area": 51965.435650000014, "rle": {"size": [427, 640], "counts": "XP1h7m0f4O1N2O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O1000000000000000000000000000000000000000000000O100hJI\\M7c2L[M4e2MZM3e20YM0f23XMMh25VMKi28UMHj2;TMEl2=RMCm2`0QMAn2a0PM_OP3c0nL]OR3e0lL[OT3h0iLYOU3j0iLVOW3l0gLTOY3n0eLRO[3P1cLQO\\3R1aLnN_3T1_LlN`3W1^LiNb3Y1\\LhNc3[1ZLeNf3]1XLcNh3_1VLaNj3a1TL`Nj3d1SL\\Nm3j1mKVNS4o1hKQNX4U3bJkL]5\\3]JeLb5^3[JbLe5^3[JbLe5_3ZJaLe5a3ZJ`Le5a3ZJ_Lf5b3YJ^Lg5b3YJ^Lf5d3YJ]Lf5d3YJ\\Lg5e3XJ[Lh5e3XJ[Lh5f3WJ[Lg5g3XJZLg5g3XJZLg5g3XJYLh5g3XJZLf5h3YJYLf5h3YJXLg5i3XJXLf5i3ZJXLd5j3[JVLd5l3[JULc5m3\\JTLa5o3^JRL`5o3`JQL_5Q4`JPL^5R4aJoK]5S4bJmK\\5U4dJlKZ5V4cJmK[5U4bJnK\\5T4`JoK^5T4^JPL`5Q4^JRL`5P4\\JSLc5o3ZJTLc5o3YJULe5l3XJWLg5h5M3N2N1O2E;E:01N2O1N101N2O0O2O1N101N2O1O0eNXGRNj8l1[GPNf8m1aGnM`8`0SG0g0WOW8h0WGJn0SOn7P1YGHX:7mEBV:<PF]OS:b0QFXOR:f0TFSOo9k0Z1N2N3M2N2N2M3N2N2N2M4M2NWUP6"}, "label": "a man with an american flag on his sleeve"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a police man looking at laptop\"."}], "task": "refering", "answer_template": "The \"a police man looking at laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.003, 0.013489461358313817, 0.280375, 0.9977517564402809], "iscrowd": 0, "area": 51965.435650000014, "rle": {"size": [427, 640], "counts": "XP1h7m0f4O1N2O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O1000000000000000000000000000000000000000000000O100hJI\\M7c2L[M4e2MZM3e20YM0f23XMMh25VMKi28UMHj2;TMEl2=RMCm2`0QMAn2a0PM_OP3c0nL]OR3e0lL[OT3h0iLYOU3j0iLVOW3l0gLTOY3n0eLRO[3P1cLQO\\3R1aLnN_3T1_LlN`3W1^LiNb3Y1\\LhNc3[1ZLeNf3]1XLcNh3_1VLaNj3a1TL`Nj3d1SL\\Nm3j1mKVNS4o1hKQNX4U3bJkL]5\\3]JeLb5^3[JbLe5^3[JbLe5_3ZJaLe5a3ZJ`Le5a3ZJ_Lf5b3YJ^Lg5b3YJ^Lf5d3YJ]Lf5d3YJ\\Lg5e3XJ[Lh5e3XJ[Lh5f3WJ[Lg5g3XJZLg5g3XJZLg5g3XJYLh5g3XJZLf5h3YJYLf5h3YJXLg5i3XJXLf5i3ZJXLd5j3[JVLd5l3[JULc5m3\\JTLa5o3^JRL`5o3`JQL_5Q4`JPL^5R4aJoK]5S4bJmK\\5U4dJlKZ5V4cJmK[5U4bJnK\\5T4`JoK^5T4^JPL`5Q4^JRL`5P4\\JSLc5o3ZJTLc5o3YJULe5l3XJWLg5h5M3N2N1O2E;E:01N2O1N101N2O0O2O1N101N2O1O0eNXGRNj8l1[GPNf8m1aGnM`8`0SG0g0WOW8h0WGJn0SOn7P1YGHX:7mEBV:<PF]OS:b0QFXOR:f0TFSOo9k0Z1N2N3M2N2N2M3N2N2N2M4M2NWUP6"}, "label": "a police man looking at laptop"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white man with a gray mustache and a white shirt\"."}], "task": "refering", "answer_template": "The \"a white man with a gray mustache and a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 288, 289, 290, 311, 312, 313, 334, 335, 336], "bbox": [0.26237499999999997, 0.022482435597189696, 0.88009375, 0.9842622950819673], "iscrowd": 0, "area": 74433.89949999998, "rle": {"size": [427, 640], "counts": "_\\V21W=3N2M2N3M3N1N3M3N1N3M3N1N3M3N2M2N3M3N1N3M3N1N3M3N1N3M3M2O2M3M3N1N3M3N1N3M3M2O2M3O0O2N2N1O2O1N2N2N1O2O1N2N2N2N101N2N2N2N2O0O2N2N200O1O001O1O100O1O001O1O1O1000O010000O1000O010000O10000O0100000O10000O01000O10000O10O10O10000O1000O010000O10000O01000O10000O10O10O100`NYK\\Jg4_5gKWJZ4d5ULRJk3h5dLoI[3m5TMhIm2S6X2L4L4L4L4L3M4L4L4M3L4L4L3M4L4O1O1O100O1O100O1O1O100O1O1O2O0O1O1O100O1O100ROlFdMU9X2oFgMR9T2SGkMn8Q2VGoMj8m1ZGSNf8h1_GXNb8c1bG\\N_8_1fGaNZ8[1jGeNV8V1oGjNQ8R1SHnNm7n0WHQOj7j0[HVOe7f0_HZOa7a0dH_O\\7=hHBY7:kHFU75PIKP71TIOl6LYI3i6H[I8e6D_I<g910O1000000O100000000000000001O0000001O000000001O0000001O000000001N10000O2L3L4M3L4M4K4M3L4M4K4M3L4M3L5L3L4M3L5L3L4M3L4M4K4L4M3L5L3L4N200O2O0O2N3N1N3N2M2N3N2M2^KVKa0l4YOYKe0i4WOZKh0i4RO\\Km0f4nN]KQ1e4jN`KU1b4fNaKY1b4YNkKf1V4[MhLe2Z3[LdMe3R6O0O2O001O001O0O101O001O001O0O2O00001O001O0O2O001O00001N101O001O001O0O101O001O001O0O2O00001O001N101O001O00001N101O001O001O0O101O001O001N2O1O001O1O1O1N2O1O1O1O1O0O2O1O1O1O1O1O1N2O001O1O1O1O1N2O1O1O1O001N2O1O1O3M4L3M4K4M4L3L4@iko0"}, "label": "a white man with a gray mustache and a white shirt"}]}
{"id": 76885, "image": "COCO_train2014_000000076885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in white looking at the monitor\"."}], "task": "refering", "answer_template": "The \"the man in white looking at the monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 288, 289, 290, 311, 312, 313, 334, 335, 336], "bbox": [0.26237499999999997, 0.022482435597189696, 0.88009375, 0.9842622950819673], "iscrowd": 0, "area": 74433.89949999998, "rle": {"size": [427, 640], "counts": "_\\V21W=3N2M2N3M3N1N3M3N1N3M3N1N3M3N2M2N3M3N1N3M3N1N3M3N1N3M3M2O2M3M3N1N3M3N1N3M3M2O2M3O0O2N2N1O2O1N2N2N1O2O1N2N2N2N101N2N2N2N2O0O2N2N200O1O001O1O100O1O001O1O1O1000O010000O1000O010000O10000O0100000O10000O01000O10000O10O10O10000O1000O010000O10000O01000O10000O10O10O100`NYK\\Jg4_5gKWJZ4d5ULRJk3h5dLoI[3m5TMhIm2S6X2L4L4L4L4L3M4L4L4M3L4L4L3M4L4O1O1O100O1O100O1O1O100O1O1O2O0O1O1O100O1O100ROlFdMU9X2oFgMR9T2SGkMn8Q2VGoMj8m1ZGSNf8h1_GXNb8c1bG\\N_8_1fGaNZ8[1jGeNV8V1oGjNQ8R1SHnNm7n0WHQOj7j0[HVOe7f0_HZOa7a0dH_O\\7=hHBY7:kHFU75PIKP71TIOl6LYI3i6H[I8e6D_I<g910O1000000O100000000000000001O0000001O000000001O0000001O000000001N10000O2L3L4M3L4M4K4M3L4M4K4M3L4M3L5L3L4M3L5L3L4M3L4M4K4L4M3L5L3L4N200O2O0O2N3N1N3N2M2N3N2M2^KVKa0l4YOYKe0i4WOZKh0i4RO\\Km0f4nN]KQ1e4jN`KU1b4fNaKY1b4YNkKf1V4[MhLe2Z3[LdMe3R6O0O2O001O001O0O101O001O001O0O2O00001O001O0O2O001O00001N101O001O001O0O101O001O001O0O2O00001O001N101O001O00001N101O001O001O0O101O001O001N2O1O001O1O1O1N2O1O1O1O1O0O2O1O1O1O1O1O1N2O001O1O1O1O1N2O1O1O1O001N2O1O1O3M4L3M4K4M4L3L4@iko0"}, "label": "the man in white looking at the monitor"}]}
{"id": 44123, "image": "COCO_train2014_000000044123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white swan swimming on the left side of another swan\"."}], "task": "refering", "answer_template": "The \"a white swan swimming on the left side of another swan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 123, 145, 146, 168, 169, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240], "bbox": [0.26234375, 0.2781041666666667, 0.7403437500000001, 0.6064375000000001], "iscrowd": 0, "area": 17069.937650000003, "rle": {"size": [480, 640], "counts": "^m^21m>1O2O0O2N101N2N1O2O0O2N1O2O1N1O2N101N2N1O2N1O20O10000O100O10SEXOb7g0[H@b7`0YHHd78XHOe70XH7e7IVH?g7@VHg0g7@mGg0Q8G]G?a8OlF6T97[FMe9g20000001N2O001O1O1O1O1O1O1O1O1O1O1O2N3L4M2N3M6J7gNjDZN];X1iDiN^;i0hDXO_;3nDNY;[OWEe0a<1O1O100O1O010O1O1O100O1O1O100O1O1O100O001O100O1O1O100O1O100O100O100O01000O100O10000O100O10000O10O010000O100O10000O100O10000O10O0100O10000O100O10000O100O10O10O1O100O100O100O1O100O100O010O1O100O10000001O1O001O001O0O100000000000001O00000000000O1000001O0000000000000000001N10002N5K2N1O1O1O001O1O001O1O1N101O00001O00000000000000001N10000000000000001O0000000O100O101N4M3L3M010O100O10O0100O1O100O010O100O100O00100O100O10O0100O100O001O1O1O1O1N101O1O1O1O1O001O1O1O1O001O1O1O1O1O00jb]2"}, "label": "a white swan swimming on the left side of another swan"}]}
{"id": 44123, "image": "COCO_train2014_000000044123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the swan closest to photographer\"."}], "task": "refering", "answer_template": "The \"the swan closest to photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 123, 145, 146, 168, 169, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240], "bbox": [0.26234375, 0.2781041666666667, 0.7403437500000001, 0.6064375000000001], "iscrowd": 0, "area": 17069.937650000003, "rle": {"size": [480, 640], "counts": "^m^21m>1O2O0O2N101N2N1O2O0O2N1O2O1N1O2N101N2N1O2N1O20O10000O100O10SEXOb7g0[H@b7`0YHHd78XHOe70XH7e7IVH?g7@VHg0g7@mGg0Q8G]G?a8OlF6T97[FMe9g20000001N2O001O1O1O1O1O1O1O1O1O1O1O2N3L4M2N3M6J7gNjDZN];X1iDiN^;i0hDXO_;3nDNY;[OWEe0a<1O1O100O1O010O1O1O100O1O1O100O1O1O100O001O100O1O1O100O1O100O100O100O01000O100O10000O100O10000O10O010000O100O10000O100O10000O10O0100O10000O100O10000O100O10O10O1O100O100O100O1O100O100O010O1O100O10000001O1O001O001O0O100000000000001O00000000000O1000001O0000000000000000001N10002N5K2N1O1O1O001O1O001O1O1N101O00001O00000000000000001N10000000000000001O0000000O100O101N4M3L3M010O100O10O0100O1O100O010O100O100O00100O100O10O0100O100O001O1O1O1O1N101O1O1O1O1O001O1O1O1O001O1O1O1O1O00jb]2"}, "label": "the swan closest to photographer"}]}
{"id": 44123, "image": "COCO_train2014_000000044123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"swan in back of other swan\"."}], "task": "refering", "answer_template": "The \"swan in back of other swan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 53, 54, 55, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128], "bbox": [0.300640625, 0.049562499999999995, 0.670796875, 0.3491458333333333], "iscrowd": 0, "area": 13036.694749999999, "rle": {"size": [480, 640], "counts": "iQj26i>1O1O1O0O2O1O1O1O001O1O1O001O001O001N101O0O2O1N101N101O100VDSOa9m0]FYO_9g0_F_O]9a0aFE[9;cFLX95eF1W9OgF7V9HhF?S96RF1m96fE1Y:6ZE1f:U20000000000001O0O1O1O1O1O2N1O1O1O1O1O2N5fMeDe0b;kNoDm0j<cNhBb0_=ROdBP1e=1N2O0O2O1N2O1N101N2O1N2O10O01O1O1O10O01O1O100O001O1O100O1O0010000O10O10O100O10000O10000O10O10O100O100O100O100O1O010O100O1O100O100O1000O100000000000000000000O1000000000001O0000000O2O0000001O0O10001N10000O101O0O10001N10001N2O1O1N101O1N2O1O1N2O1O0O2O1O001N101O0O2O1O1N2O0O100O1O00100O001O0O101O0O101O00010O000010O00000100O1O1O1O1O2N1O2N1O2N1O2N1N3LmiR3"}, "label": "swan in back of other swan"}]}
{"id": 44123, "image": "COCO_train2014_000000044123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goose behind another goose\"."}], "task": "refering", "answer_template": "The \"a goose behind another goose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 53, 54, 55, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128], "bbox": [0.300640625, 0.049562499999999995, 0.670796875, 0.3491458333333333], "iscrowd": 0, "area": 13036.694749999999, "rle": {"size": [480, 640], "counts": "iQj26i>1O1O1O0O2O1O1O1O001O1O1O001O001O001N101O0O2O1N101N101O100VDSOa9m0]FYO_9g0_F_O]9a0aFE[9;cFLX95eF1W9OgF7V9HhF?S96RF1m96fE1Y:6ZE1f:U20000000000001O0O1O1O1O1O2N1O1O1O1O1O2N5fMeDe0b;kNoDm0j<cNhBb0_=ROdBP1e=1N2O0O2O1N2O1N101N2O1N2O10O01O1O1O10O01O1O100O001O1O100O1O0010000O10O10O100O10000O10000O10O10O100O100O100O100O1O010O100O1O100O100O1000O100000000000000000000O1000000000001O0000000O2O0000001O0O10001N10000O101O0O10001N10001N2O1O1N101O1N2O1O1N2O1O0O2O1O001N101O0O2O1O1N2O0O100O1O00100O001O0O101O0O101O00010O000010O00000100O1O1O1O1O2N1O2N1O2N1O2N1N3LmiR3"}, "label": "a goose behind another goose"}]}
{"id": 248932, "image": "COCO_train2014_000000248932.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a white shirt\"."}], "task": "refering", "answer_template": "The \"a girl in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 90, 91, 106, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 207, 208, 209, 210, 224, 225, 226, 241, 242, 243, 257, 258, 259, 274, 275, 276, 277], "bbox": [0.13670833333333335, 0.17321875, 0.431125, 0.7407812500000001], "iscrowd": 0, "area": 25332.352949999997, "rle": {"size": [640, 480], "counts": "]fY13lc03M2N3L4M2N2N1O1O2M2O1N2N1O1O1N2gMQOaAP1_>QOTA[1l>fN^@o1a?SN\\@o1d?RNZ@o1e?TNX@m1h?TNU@n1j?TNT@l1j?XNT@f1m?]NP@c1o?bNn_O\\1S`0gNk_OW1U`0mNi_OR1V`0ROg_On0X`0VOf_Oi0Y`0[Oe_Od0Z`0`1O100L6M2N3M2N2N3eNZ1B=F:K5001O1O001O1OPDoIT:S6kEPJR:P6mETJP:m5oEVJP:i5oEZJo9f5oE_Jn9a5dEoJZ:P5WEaKh:_4TEgKk:X4SEkKl:U4REnKm:R4QEQLn:o3PEULo:j3oDYLP;g3nD]LQ;b3mDbLR;]3lDgLR;T601O1OZORE[Hl:c7ZE[He:b7`E]H_:`7eE_HZ:_7jE`H`9GZFf7:bHZ9MXF_7b0cHT93UFZ7j0bHo8:QFS7T1bHi8T8[GkGc8T8aGkG\\8U8jGhGT8W8RHfGj7[8j13N1N12N4K4L5L4N20O01O1O1O010O1O1O010O1O1O2N1O1O2N1O1O2K4L4L5K4L4L5L3PKlBS3X=hLjBX3X=fLiBY3Z=dLfB[3^=bLbB^3a=`L_B^3e=_L[Ba3h=]LXBb3k=[LUBd3o=mKhAL:V4Q>gKnAM2Z4U>bK_B[4d>N3M3M3N2M3N2H9G8H8HX^Z5"}, "label": "a girl in a white shirt"}]}
{"id": 248932, "image": "COCO_train2014_000000248932.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white female in a white shirt and dark patterned board shorts\"."}], "task": "refering", "answer_template": "The \"a white female in a white shirt and dark patterned board shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 90, 91, 106, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 207, 208, 209, 210, 224, 225, 226, 241, 242, 243, 257, 258, 259, 274, 275, 276, 277], "bbox": [0.13670833333333335, 0.17321875, 0.431125, 0.7407812500000001], "iscrowd": 0, "area": 25332.352949999997, "rle": {"size": [640, 480], "counts": "]fY13lc03M2N3L4M2N2N1O1O2M2O1N2N1O1O1N2gMQOaAP1_>QOTA[1l>fN^@o1a?SN\\@o1d?RNZ@o1e?TNX@m1h?TNU@n1j?TNT@l1j?XNT@f1m?]NP@c1o?bNn_O\\1S`0gNk_OW1U`0mNi_OR1V`0ROg_On0X`0VOf_Oi0Y`0[Oe_Od0Z`0`1O100L6M2N3M2N2N3eNZ1B=F:K5001O1O001O1OPDoIT:S6kEPJR:P6mETJP:m5oEVJP:i5oEZJo9f5oE_Jn9a5dEoJZ:P5WEaKh:_4TEgKk:X4SEkKl:U4REnKm:R4QEQLn:o3PEULo:j3oDYLP;g3nD]LQ;b3mDbLR;]3lDgLR;T601O1OZORE[Hl:c7ZE[He:b7`E]H_:`7eE_HZ:_7jE`H`9GZFf7:bHZ9MXF_7b0cHT93UFZ7j0bHo8:QFS7T1bHi8T8[GkGc8T8aGkG\\8U8jGhGT8W8RHfGj7[8j13N1N12N4K4L5L4N20O01O1O1O010O1O1O010O1O1O2N1O1O2N1O1O2K4L4L5K4L4L5L3PKlBS3X=hLjBX3X=fLiBY3Z=dLfB[3^=bLbB^3a=`L_B^3e=_L[Ba3h=]LXBb3k=[LUBd3o=mKhAL:V4Q>gKnAM2Z4U>bK_B[4d>N3M3M3N2M3N2H9G8H8HX^Z5"}, "label": "a white female in a white shirt and dark patterned board shorts"}]}
{"id": 248932, "image": "COCO_train2014_000000248932.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man next to the surfboard\"."}], "task": "refering", "answer_template": "The \"man next to the surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 78, 79, 94, 95, 96, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 230, 231, 232, 233, 247, 248, 249, 250, 264, 265, 266, 267, 281, 282, 283, 284, 300, 301], "bbox": [0.4404583333333333, 0.15562499999999999, 0.8119791666666667, 0.7758437499999999], "iscrowd": 0, "area": 33826.99220000001, "rle": {"size": [640, 480], "counts": "hiT48cc07J5L5K4M2M3M4L3L5K4L4L2NO010O0021N3N1N3N1N3C<K6N1N3M2O2M101000O01000O01000Of_O^N[>a1bAjNW>U1fAUOT>i0iACP><mAOl=0PB=h=AVBj0c=b3XO3M3L4M3M4K6K4L4K6K4L;D4M3M1O1O1O1O1O1O1O1O1O1O1O1N2K5WOlF`FY9W9m0G9L4NlNfGeFW8T9VHiFg7U9aHhF]7U9kHhFR7W9k1O2M3N3M3O1010O000010O01O001O01O01O003N3L5L3N2N01O0O10002O3M3Ld0]OO0O2N3M4K5L8G8I2M2O0O2O0O2O0O2O0OYJWHXOi7`0hHWOW7b0lJbMU5^1RL[Nm3_1^L\\Nc3]1hL_NW3[1TM`Nm2Y1]MdNb2V1gMgNZ2W1jMgNU2\\1jMbNW2`1iM]NX2e1hMYNX2l1eMQN]2Y2YMfMg2_2hGcM\\3Ln4e2ZGfMo=^2gAhMZ>\\2ZAlMe>Y2n@oMS?_301O01N101O0O2O2M4M2N3L7J9G9E7J3M3L4M3M3L4M4I8F:F:D=Be0\\Olng1"}, "label": "man next to the surfboard"}]}
{"id": 248932, "image": "COCO_train2014_000000248932.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the white shirt with a red emblem\"."}], "task": "refering", "answer_template": "The \"the man with the white shirt with a red emblem\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 78, 79, 94, 95, 96, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 230, 231, 232, 233, 247, 248, 249, 250, 264, 265, 266, 267, 281, 282, 283, 284, 300, 301], "bbox": [0.4404583333333333, 0.15562499999999999, 0.8119791666666667, 0.7758437499999999], "iscrowd": 0, "area": 33826.99220000001, "rle": {"size": [640, 480], "counts": "hiT48cc07J5L5K4M2M3M4L3L5K4L4L2NO010O0021N3N1N3N1N3C<K6N1N3M2O2M101000O01000O01000Of_O^N[>a1bAjNW>U1fAUOT>i0iACP><mAOl=0PB=h=AVBj0c=b3XO3M3L4M3M4K6K4L4K6K4L;D4M3M1O1O1O1O1O1O1O1O1O1O1O1N2K5WOlF`FY9W9m0G9L4NlNfGeFW8T9VHiFg7U9aHhF]7U9kHhFR7W9k1O2M3N3M3O1010O000010O01O001O01O01O003N3L5L3N2N01O0O10002O3M3Ld0]OO0O2N3M4K5L8G8I2M2O0O2O0O2O0O2O0OYJWHXOi7`0hHWOW7b0lJbMU5^1RL[Nm3_1^L\\Nc3]1hL_NW3[1TM`Nm2Y1]MdNb2V1gMgNZ2W1jMgNU2\\1jMbNW2`1iM]NX2e1hMYNX2l1eMQN]2Y2YMfMg2_2hGcM\\3Ln4e2ZGfMo=^2gAhMZ>\\2ZAlMe>Y2n@oMS?_301O01N101O0O2O2M4M2N3L7J9G9E7J3M3L4M3M3L4M4I8F:F:D=Be0\\Olng1"}, "label": "the man with the white shirt with a red emblem"}]}
{"id": 248932, "image": "COCO_train2014_000000248932.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue surf board lying on the ground between two people\"."}], "task": "refering", "answer_template": "The \"a blue surf board lying on the ground between two people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 243, 244, 245, 246, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 294, 295, 296, 297, 298, 299, 311, 312, 313, 314, 315, 316, 317, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 362, 363, 364, 365, 366, 367, 368, 369, 370, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.3020833333333333, 0.592375, 0.8141666666666667, 1.0], "iscrowd": 0, "area": 42228.268500000006, "rle": {"size": [640, 480], "counts": "\\Sk2n0ib0m1SNl1SNa0@;E;E:E<E:H7O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O100001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O1O1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1N3J6J^Pg1"}, "label": "a blue surf board lying on the ground between two people"}]}
{"id": 248932, "image": "COCO_train2014_000000248932.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue surfboard laying on the grass\"."}], "task": "refering", "answer_template": "The \"a blue surfboard laying on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 243, 244, 245, 246, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 294, 295, 296, 297, 298, 299, 311, 312, 313, 314, 315, 316, 317, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 362, 363, 364, 365, 366, 367, 368, 369, 370, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.3020833333333333, 0.592375, 0.8141666666666667, 1.0], "iscrowd": 0, "area": 42228.268500000006, "rle": {"size": [640, 480], "counts": "\\Sk2n0ib0m1SNl1SNa0@;E;E:E<E:H7O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O100001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O1O1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N1N3J6J^Pg1"}, "label": "a blue surfboard laying on the grass"}]}
{"id": 240741, "image": "COCO_train2014_000000240741.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white blanket on a bed\"."}], "task": "refering", "answer_template": "The \"a white blanket on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.01046875, 0.6971830985915493, 0.6027968749999999, 0.996056338028169], "iscrowd": 0, "area": 32915.89635000001, "rle": {"size": [426, 640], "counts": "oV3_3[9`000001O000000000000001O0000000000001O000000000000001O00000000000O2O000000000000001O0000000000001O000000000000001O0000000000001O00000000000O2O000000000000001O0000000000001O000000000000001O0000000000001O00000000000O101O0000000000001O000000000000001O0000000000001O000000000000001O00000000000O2O0000000000001O000000000000001O0000000000001O000000000000001O00000000000O2O000000000000001O0000000000001O000000000000001O0000000000001O00000000000O2O000000000000O100O100O100O100O100O10000O10000001O1O001O1O1O001O1O001O1O0O2O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001N2O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1N101O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001N2O001O1O001O1O1O00ieY3"}, "label": "a white blanket on a bed"}]}
{"id": 240741, "image": "COCO_train2014_000000240741.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bed in the corner with white comforter\"."}], "task": "refering", "answer_template": "The \"bed in the corner with white comforter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.01046875, 0.6971830985915493, 0.6027968749999999, 0.996056338028169], "iscrowd": 0, "area": 32915.89635000001, "rle": {"size": [426, 640], "counts": "oV3_3[9`000001O000000000000001O0000000000001O000000000000001O00000000000O2O000000000000001O0000000000001O000000000000001O0000000000001O00000000000O2O000000000000001O0000000000001O000000000000001O0000000000001O00000000000O101O0000000000001O000000000000001O0000000000001O000000000000001O00000000000O2O0000000000001O000000000000001O0000000000001O000000000000001O00000000000O2O000000000000001O0000000000001O000000000000001O0000000000001O00000000000O2O000000000000O100O100O100O100O100O10000O10000001O1O001O1O1O001O1O001O1O0O2O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001N2O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1N101O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001N2O001O1O001O1O1O00ieY3"}, "label": "bed in the corner with white comforter"}]}
{"id": 27750, "image": "COCO_train2014_000000027750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman at a picnic licking her thumb\"."}], "task": "refering", "answer_template": "The \"a woman at a picnic licking her thumb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 20, 21, 22, 23, 24, 25, 37, 38, 39, 40, 41, 42, 53, 54, 55, 56, 57, 58, 59, 60, 71, 72, 73, 74, 75, 76, 77, 78, 88, 89, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 145, 146, 156, 157, 158, 159, 160, 174, 175, 176, 177, 191, 192, 193, 200, 201, 202, 208, 209, 217, 218, 219, 235, 236], "bbox": [0.16779166666666667, 0.0, 0.9408333333333334, 0.611234375], "iscrowd": 0, "area": 52573.7555, "rle": {"size": [640, 480], "counts": "afb17gc09H8G9H8G7J4K5L3L5L3L5L3L5L3L5L5K5J6K5K5J7J7I7I7I8G<E;E;E;Eh0WOg0ZO5K4L5J6K4L5K4L5J5L5K4L5K5J5L5K3M1O2N2M2O2O001N2O001O1O0O2O001O1O0O2O00O1O100O010O1O100O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O00100O100O00100O1O10O0100O1O10O01O100O2O1N2N101N2N101N2N2O0O2N2O1N101N2N2O0O2N2O1N2fNjDgIX;R6[EaIf:X6mEZIU:`6i1J6N3L3N2N2M3N2N2M4N3N1N3M2N3M3M2O2M2N3M4L4M3L4L4L4L3N1N1O1O1N200O2O000O10000O2O0O10000O2O000O10001N1000O10O10000O10O10O100O1O10O01O1O5K4L4M3L4L5K4L4L4L5L3L<Dd0\\O;C=CdbZ1Ie]eN6J6J3M3M3M2N1O2N2N2N2N102M3M3N0O1O1O101N1O2O0O2N101N2N2N2O1N3M2O1N2N2O0O0001O000O101O00001N10001O0O1O1O2N2N1O2N2N2N1O2N2N2N1N3N2N1O2N2M3N3L4L4L4L4LXUa0"}, "label": "a woman at a picnic licking her thumb"}]}
{"id": 27750, "image": "COCO_train2014_000000027750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady licking her thumb while someone is handing her a banana in a hotdog bun\"."}], "task": "refering", "answer_template": "The \"a lady licking her thumb while someone is handing her a banana in a hotdog bun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 20, 21, 22, 23, 24, 25, 37, 38, 39, 40, 41, 42, 53, 54, 55, 56, 57, 58, 59, 60, 71, 72, 73, 74, 75, 76, 77, 78, 88, 89, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 145, 146, 156, 157, 158, 159, 160, 174, 175, 176, 177, 191, 192, 193, 200, 201, 202, 208, 209, 217, 218, 219, 235, 236], "bbox": [0.16779166666666667, 0.0, 0.9408333333333334, 0.611234375], "iscrowd": 0, "area": 52573.7555, "rle": {"size": [640, 480], "counts": "afb17gc09H8G9H8G7J4K5L3L5L3L5L3L5L3L5L5K5J6K5K5J7J7I7I7I8G<E;E;E;Eh0WOg0ZO5K4L5J6K4L5K4L5J5L5K4L5K5J5L5K3M1O2N2M2O2O001N2O001O1O0O2O001O1O0O2O00O1O100O010O1O100O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O00100O100O00100O1O10O0100O1O10O01O100O2O1N2N101N2N101N2N2O0O2N2O1N101N2N2O0O2N2O1N2fNjDgIX;R6[EaIf:X6mEZIU:`6i1J6N3L3N2N2M3N2N2M4N3N1N3M2N3M3M2O2M2N3M4L4M3L4L4L4L3N1N1O1O1N200O2O000O10000O2O0O10000O2O000O10001N1000O10O10000O10O10O100O1O10O01O1O5K4L4M3L4L5K4L4L4L5L3L<Dd0\\O;C=CdbZ1Ie]eN6J6J3M3M3M2N1O2N2N2N2N102M3M3N0O1O1O101N1O2O0O2N101N2N2N2O1N3M2O1N2N2O0O0001O000O101O00001N10001O0O1O1O2N2N1O2N2N2N1O2N2N2N1N3N2N1O2N2M3N3L4L4L4L4LXUa0"}, "label": "a lady licking her thumb while someone is handing her a banana in a hotdog bun"}]}
{"id": 27750, "image": "COCO_train2014_000000027750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand that is holding a bun with a banana in it\"."}], "task": "refering", "answer_template": "The \"a hand that is holding a bun with a banana in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 192, 193, 194, 195, 208, 209, 210, 211, 217, 225, 226, 227, 228, 234, 235, 242, 243, 244, 245, 251, 252, 258, 259, 260, 261, 262, 268, 269, 275, 276, 277, 278, 285, 286, 291, 292, 293, 294, 295, 302, 303, 308, 309, 310, 311, 312, 313, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 334, 335, 336, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.006, 0.45618749999999997, 0.8569375, 0.986515625], "iscrowd": 0, "area": 50913.347149999965, "rle": {"size": [640, 480], "counts": "d_21ic02O1O4L5K6M2O1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O2N1O1N3M2N2M3N3M2M3N2N2N2M3N2N2N2M3N2M3N2M3N2M3M3N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2M3N2N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2M3N2N2O1O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O10_KkFWNT9k0nGROR8j0UHSOj7k0\\HROd7j0cHSO\\7k0jHROV7j0QISOn6k0XIROh6j0_ISO`6k0fIROZ6j0mISOR6k0TJROk5k0\\JROd5j0cJSO\\5j0kJSOU5;_KC`47hKHX45mKIU42oKMS4NRL0o3LUL3m3HXL6j3EZL:g3B^L;e3@_L?c3\\ObLb0`3YOdLf0]3VOhLh0Z3SOjLl0X3oNmLo0T3mNPMQ1S3jNRMT1P3gNSMY1n2cNUM]1m2^NVMb1l2YNWMg1a:0000O1000001O0000000O1000000000000O100000000000000O2O0000000000000O1000000000000O10000000001O000O100O1O001O1O001O001O0010O01O001O1O001O001O001O010O001O001O1O001O001O0010O010O01O010O00100O0010O010O01O010O0010O01O010O010O0010O01O010O1O010O010O0010O01O01M2N2O2M2O2M2O0O2N2O1N2O1N2O1N1O2O1N2O1N2O1N2O0O2N2O1N2O1J5dMiMjB_2P=kMdB\\2V=PN]BW2^=TNVBS2d=WNPBP2R>TNaAS2b>d13L3nMYAhNi>T1]AgNf>T1bAfNa>V1fAeN\\>V1lAdNW>X1PBbNS>Y1TBcNn=Y1YBaNj=[1]B_Ng=[1aB`Ne=X1aBcNg=S1l2F9F:GfgZ1"}, "label": "a hand that is holding a bun with a banana in it"}]}
{"id": 27750, "image": "COCO_train2014_000000027750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hand holding banana in hot dog bun\"."}], "task": "refering", "answer_template": "The \"hand holding banana in hot dog bun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 192, 193, 194, 195, 208, 209, 210, 211, 217, 225, 226, 227, 228, 234, 235, 242, 243, 244, 245, 251, 252, 258, 259, 260, 261, 262, 268, 269, 275, 276, 277, 278, 285, 286, 291, 292, 293, 294, 295, 302, 303, 308, 309, 310, 311, 312, 313, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 334, 335, 336, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.006, 0.45618749999999997, 0.8569375, 0.986515625], "iscrowd": 0, "area": 50913.347149999965, "rle": {"size": [640, 480], "counts": "d_21ic02O1O4L5K6M2O1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O2N1O1N3M2N2M3N3M2M3N2N2N2M3N2N2N2M3N2M3N2M3N2M3M3N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2M3N2N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2M3N2N2O1O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O10_KkFWNT9k0nGROR8j0UHSOj7k0\\HROd7j0cHSO\\7k0jHROV7j0QISOn6k0XIROh6j0_ISO`6k0fIROZ6j0mISOR6k0TJROk5k0\\JROd5j0cJSO\\5j0kJSOU5;_KC`47hKHX45mKIU42oKMS4NRL0o3LUL3m3HXL6j3EZL:g3B^L;e3@_L?c3\\ObLb0`3YOdLf0]3VOhLh0Z3SOjLl0X3oNmLo0T3mNPMQ1S3jNRMT1P3gNSMY1n2cNUM]1m2^NVMb1l2YNWMg1a:0000O1000001O0000000O1000000000000O100000000000000O2O0000000000000O1000000000000O10000000001O000O100O1O001O1O001O001O0010O01O001O1O001O001O001O010O001O001O1O001O001O0010O010O01O010O00100O0010O010O01O010O0010O01O010O010O0010O01O010O1O010O010O0010O01O01M2N2O2M2O2M2O0O2N2O1N2O1N2O1N1O2O1N2O1N2O1N2O0O2N2O1N2O1J5dMiMjB_2P=kMdB\\2V=PN]BW2^=TNVBS2d=WNPBP2R>TNaAS2b>d13L3nMYAhNi>T1]AgNf>T1bAfNa>V1fAeN\\>V1lAdNW>X1PBbNS>Y1TBcNn=Y1YBaNj=[1]B_Ng=[1aB`Ne=X1aBcNg=S1l2F9F:GfgZ1"}, "label": "hand holding banana in hot dog bun"}]}
{"id": 142439, "image": "COCO_train2014_000000142439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby playing tennis\"."}], "task": "refering", "answer_template": "The \"baby playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 65, 66, 67, 77, 78, 79, 88, 89, 90, 91, 100, 101, 102, 103, 112, 113, 114, 115, 124, 125, 126, 127, 136, 137, 138, 139, 148, 149, 150, 151, 160, 161, 162, 172, 173, 174, 184, 185, 186, 196, 197, 198, 199], "bbox": [0.3462388059701492, 0.24774000000000002, 0.6554925373134327, 0.93918], "iscrowd": 0, "area": 25337.145299999993, "rle": {"size": [500, 335], "counts": "klh1o0Y>f0PFiNf4h1QKZNl4l1hIeMaNa0e7P2aInM_N4m7T2\\IUN\\NIV8X2VI]NXN]O`8[2QIXOl6n0lHVOR7g4N3M2N2N3M2N3N1N2O2N1O1]NUH\\Km7^4\\H]Ke7^4eH[K]7a4lHYKV7a4TIYKm6c4YIZKi6b4\\I[Ke6d4]IZKd6e4^IXKd6h4^IUKb6k4`ISKa6l4bIQK`6m4bIQK_6o4bInJ`6Q5R2O1O100O1O1O10000000WOTKjFl4V9VKhFj4X9XKgFg4Z9ZKcFf4^9]K^Fd4b9^KZFd4f9^KVFd4k9^KQFc4o9a0001O00001O000O101O00WObJ\\G^5a8fJ_GY5^8kJaGU5]8nJbGR5[8RKdGo4X8UKhGk4U8WKkGj4Q8ZKnGh4m7\\KSHd4c1_Jm3R1_J`4a1aJl3S1bJ^4_1aJk3U1eJ[4]1cJj3V1iJW4[1dJ\\3f1XKh3X1eJn2U2hKX3W1fJj2Y2nKR3V1gJh2[2QLP3T1gJh2\\2SLn2U1fJd2`2VLk2U1fJa2c2YLh2U1fJ_2e2ZLh2U1dJ]2h2[Lg2X1aJZ2k2[Li2X1]JY2\\9dMdFZ2a9bMaFZ2e9_M`F_2d9[M`Fe2c9SMbFl2a9nLcFQ3`9iLdFW3m:O1N2M:Ga0^Oa0_Ob0_Oa0^Oa0_OXkg1"}, "label": "baby playing tennis"}]}
{"id": 142439, "image": "COCO_train2014_000000142439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian toddler wearing a floral top , pink capris , and white sandals while holding an orange tennis raquet . .\"."}], "task": "refering", "answer_template": "The \"an asian toddler wearing a floral top , pink capris , and white sandals while holding an orange tennis raquet . .\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 65, 66, 67, 77, 78, 79, 88, 89, 90, 91, 100, 101, 102, 103, 112, 113, 114, 115, 124, 125, 126, 127, 136, 137, 138, 139, 148, 149, 150, 151, 160, 161, 162, 172, 173, 174, 184, 185, 186, 196, 197, 198, 199], "bbox": [0.3462388059701492, 0.24774000000000002, 0.6554925373134327, 0.93918], "iscrowd": 0, "area": 25337.145299999993, "rle": {"size": [500, 335], "counts": "klh1o0Y>f0PFiNf4h1QKZNl4l1hIeMaNa0e7P2aInM_N4m7T2\\IUN\\NIV8X2VI]NXN]O`8[2QIXOl6n0lHVOR7g4N3M2N2N3M2N3N1N2O2N1O1]NUH\\Km7^4\\H]Ke7^4eH[K]7a4lHYKV7a4TIYKm6c4YIZKi6b4\\I[Ke6d4]IZKd6e4^IXKd6h4^IUKb6k4`ISKa6l4bIQK`6m4bIQK_6o4bInJ`6Q5R2O1O100O1O1O10000000WOTKjFl4V9VKhFj4X9XKgFg4Z9ZKcFf4^9]K^Fd4b9^KZFd4f9^KVFd4k9^KQFc4o9a0001O00001O000O101O00WObJ\\G^5a8fJ_GY5^8kJaGU5]8nJbGR5[8RKdGo4X8UKhGk4U8WKkGj4Q8ZKnGh4m7\\KSHd4c1_Jm3R1_J`4a1aJl3S1bJ^4_1aJk3U1eJ[4]1cJj3V1iJW4[1dJ\\3f1XKh3X1eJn2U2hKX3W1fJj2Y2nKR3V1gJh2[2QLP3T1gJh2\\2SLn2U1fJd2`2VLk2U1fJa2c2YLh2U1fJ_2e2ZLh2U1dJ]2h2[Lg2X1aJZ2k2[Li2X1]JY2\\9dMdFZ2a9bMaFZ2e9_M`F_2d9[M`Fe2c9SMbFl2a9nLcFQ3`9iLdFW3m:O1N2M:Ga0^Oa0_Ob0_Oa0^Oa0_OXkg1"}, "label": "an asian toddler wearing a floral top , pink capris , and white sandals while holding an orange tennis raquet . ."}]}
{"id": 117871, "image": "COCO_train2014_000000117871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting on something has on a blue shirt whit white hair and glasses\"."}], "task": "refering", "answer_template": "The \"man sitting on something has on a blue shirt whit white hair and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 134, 135, 136, 157, 158, 159, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.73165625, 0.2966276346604215, 1.0, 1.0], "iscrowd": 0, "area": 29005.274549999995, "rle": {"size": [427, 640], "counts": "SbS62S=6K5L4O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1ROnNbES1Z:ROeEn0X:VOfEk0V:ZOhEg0W:[OhEe0W:]OgEd0Y:]OeEd0Z:^OeEb0Z:@dEa0\\:@bEa0]:AbE?]:CaE>_:C_E?_:C`E=`:D^E=a:E]E<b:E^E;b:W1O1O100O100O1O1O1O1O100O1O1O1O1O1mJWMCj2;]M_Od2?aM]O`2a0fMZO[2d0kMWOV2g0RNROo1l0XNnNi1P1YNoNh1o0[NoNf1o0\\NPOe1o0]NoNd1P1^NnNc1Q1_NUNZ2j1gMhMg2V2\\M^Mo2a2SMVMU3i2mLQMX3n2jLnLY3Q3iLlLY3S3hLjL[3V3fLfL]3Y3eLcL^3]3cL_L`3a3aL\\La3c3`LZLc3f3^LVLe3j3\\LRLg3m3[LoKh3Q4i2O10000O10000O100O10000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1aHaKP6a4gIgKX6\\4`IjK_6Z4ZIiKf6\\4QIiKo6Z4jHjKU7V501O1O001O1O001O1O001O001O1O001O001O001O1O001O001O1O001O1O2N1O1O1O2N1O1O1O2N1O1O1O1mLmK"}, "label": "man sitting on something has on a blue shirt whit white hair and glasses"}]}
{"id": 117871, "image": "COCO_train2014_000000117871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with gray hair wearing a short sleeve shirt\"."}], "task": "refering", "answer_template": "The \"a man with gray hair wearing a short sleeve shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 134, 135, 136, 157, 158, 159, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.73165625, 0.2966276346604215, 1.0, 1.0], "iscrowd": 0, "area": 29005.274549999995, "rle": {"size": [427, 640], "counts": "SbS62S=6K5L4O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1ROnNbES1Z:ROeEn0X:VOfEk0V:ZOhEg0W:[OhEe0W:]OgEd0Y:]OeEd0Z:^OeEb0Z:@dEa0\\:@bEa0]:AbE?]:CaE>_:C_E?_:C`E=`:D^E=a:E]E<b:E^E;b:W1O1O100O100O1O1O1O1O100O1O1O1O1O1mJWMCj2;]M_Od2?aM]O`2a0fMZO[2d0kMWOV2g0RNROo1l0XNnNi1P1YNoNh1o0[NoNf1o0\\NPOe1o0]NoNd1P1^NnNc1Q1_NUNZ2j1gMhMg2V2\\M^Mo2a2SMVMU3i2mLQMX3n2jLnLY3Q3iLlLY3S3hLjL[3V3fLfL]3Y3eLcL^3]3cL_L`3a3aL\\La3c3`LZLc3f3^LVLe3j3\\LRLg3m3[LoKh3Q4i2O10000O10000O100O10000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1aHaKP6a4gIgKX6\\4`IjK_6Z4ZIiKf6\\4QIiKo6Z4jHjKU7V501O1O001O1O001O1O001O001O1O001O001O001O1O001O001O1O001O1O2N1O1O1O2N1O1O1O2N1O1O1O1mLmK"}, "label": "a man with gray hair wearing a short sleeve shirt"}]}
{"id": 117871, "image": "COCO_train2014_000000117871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing on the side of a mountain in a purple jacket\"."}], "task": "refering", "answer_template": "The \"a woman standing on the side of a mountain in a purple jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 151, 152, 153, 154, 155, 177, 178, 200, 201, 222, 223, 224, 245, 246, 247, 268, 269, 270, 293], "bbox": [0.59384375, 0.332248243559719, 0.7787656249999999, 0.8291803278688524], "iscrowd": 0, "area": 9055.03715, "rle": {"size": [427, 640], "counts": "_dn42Y=001hBOT=4000O100OmBNn<3QCOm<80O1O1O100O0100000000O1000000O100000O10O1000000O100000000TGJb47]KKb45^KLa44_KLa44_KLa44_KM`43`KM`43_KNa41`KO`41`KO_42_K0a40[K4e4LUK:j4GRK=n4CfJi0Y5XOdJk0\\5UO`Jo0_5RO[JT1e5lNRJ]1n5cNoI`1X5RNQI?d1c1Z5PNPI<e1f1[5PNmH;e1g1^5oMmH:b1j1a5mMkH:a1l1d5jM]HK:?b1n1h5nMcH5c1o1j5RN]HOh1o1k5iNRJY1l5jNQJW1o5mNmIc0dNQNe7e1]I4T71fHK`77]HDi7>SH[OU8k0dGWO\\8m0^GTOc8o0XGQOj8e23N2O1N2O1N2O1N2O2M2O2N3N1O11O4L4M3N2N3[HQLd5U4SJmKo5T4lInKW6Q4eIPL]6Q4eHRL5OV7j4bHWK_7Z52O2M1O1O1O0O2O1O1N5K9F9H6bMUG1Q9QNTGS19e0V9mN\\Gj0P9_N`G[1\\:nN^D1VYk1"}, "label": "a woman standing on the side of a mountain in a purple jacket"}]}
{"id": 117871, "image": "COCO_train2014_000000117871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a purple jacket and khaki pants just threw a frisbee\"."}], "task": "refering", "answer_template": "The \"a woman in a purple jacket and khaki pants just threw a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 151, 152, 153, 154, 155, 177, 178, 200, 201, 222, 223, 224, 245, 246, 247, 268, 269, 270, 293], "bbox": [0.59384375, 0.332248243559719, 0.7787656249999999, 0.8291803278688524], "iscrowd": 0, "area": 9055.03715, "rle": {"size": [427, 640], "counts": "_dn42Y=001hBOT=4000O100OmBNn<3QCOm<80O1O1O100O0100000000O1000000O100000O10O1000000O100000000TGJb47]KKb45^KLa44_KLa44_KLa44_KM`43`KM`43_KNa41`KO`41`KO_42_K0a40[K4e4LUK:j4GRK=n4CfJi0Y5XOdJk0\\5UO`Jo0_5RO[JT1e5lNRJ]1n5cNoI`1X5RNQI?d1c1Z5PNPI<e1f1[5PNmH;e1g1^5oMmH:b1j1a5mMkH:a1l1d5jM]HK:?b1n1h5nMcH5c1o1j5RN]HOh1o1k5iNRJY1l5jNQJW1o5mNmIc0dNQNe7e1]I4T71fHK`77]HDi7>SH[OU8k0dGWO\\8m0^GTOc8o0XGQOj8e23N2O1N2O1N2O1N2O2M2O2N3N1O11O4L4M3N2N3[HQLd5U4SJmKo5T4lInKW6Q4eIPL]6Q4eHRL5OV7j4bHWK_7Z52O2M1O1O1O0O2O1O1N5K9F9H6bMUG1Q9QNTGS19e0V9mN\\Gj0P9_N`G[1\\:nN^D1VYk1"}, "label": "a woman in a purple jacket and khaki pants just threw a frisbee"}]}
{"id": 371824, "image": "COCO_train2014_000000371824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blond haired , blue eyed young boy in a blue jacket\"."}], "task": "refering", "answer_template": "The \"a blond haired , blue eyed young boy in a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505], "bbox": [0.44420312500000003, 0.21834375, 0.97321875, 0.957171875], "iscrowd": 0, "area": 107822.56530000003, "rle": {"size": [640, 640], "counts": "]ma52lc06J5K5K3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N3M3N2M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N3L3N2N2N2M3N2N23MO1N2O1O1N2O2M2O2M2O2M2O2WM]LPEe3l:_LSEc3h:aLVEb3f:bLXE_3d:eLZE^3a:fL^E[3^:jL_EY3\\:kLcEV3Y:oLeES3V:QMhER3S:RMlEQ3n9TMoEP3l9SMSFP3g9TMWFP3c9UMZFo2`9UM\\FQ3^9TM]FR3]9RM`FS3Z9QMbFU3X9m3L5J6J6J6L4K5K5K5L4K5K4L5K5L4K5K5K5K5L4K5K4L2O1N2O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001PHZJd1f5ZN]Jf1b5YN`Jf1`5YNbJf1_5WNdJh1\\5WNeJi1[5UNhJj1X5UNjJj1W5TNkJk1U5SNnJl1R5SNPKl1P5SNQKm1P5PNSKo1n4oMTKP2n4lMUKS2n4iMTKV2n4gMSKY2o4cMTK\\2n4aMTK^2o4^MSKa2o4[MTKd2n4YMTKf2n4VMTKj2o4RMSKm2o4PMSKo2o4mLTKR3m4kLVKT3k4iLXKV3i4gLYKY3i4cLZK\\3g4aL\\K^3e4_L^K`3c4]L`Kb3a4[LbKd3_4YLcKg3^4VLeKi3\\4TLgKk3Z4RLiKm3X4PLkKo3V4nKlKR4U4kKnKT4S4iKPLV4R4fKQLY4]4WKeKi4n4bJUK]5\\5PJgJo5i5_IZJ`6c901O1O001O1O1QKeBW3[=hLgBW3Z=hLhBV3Y=iLhBV3Y=hLiBW3W=iLkBU3V=jLkBU3V=iLmBU3S=kLnBT3S=kLnBT3S=jLPCT3Q=kLPCT3P=lLRCR3o<lLSCS3n<lLSCS3n<lLTCR3l<mLVCR3k<mLWCQ3j<nLWCQ3i<nLYCQ3h<nLZCP3g<oLZCP3g<nL\\CP3d<PM]Co2d<PM]Co2d<oL_Co2a<QM`Cn2a<QMaCm2`<QMbCn2_<QMcCn2]<QMdCn2]<PMfCn2[<QMfCn2[<QMgCm2Z<QMhCn2Y<QMiCm2X<RMiCm2X<QMkCa2b<^M_CP2S=oMoB^1c=`N_Bn0S>QOoA=b>B_ALS?2_2O1O[j:"}, "label": "a blond haired , blue eyed young boy in a blue jacket"}]}
{"id": 371824, "image": "COCO_train2014_000000371824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with blue eyes\"."}], "task": "refering", "answer_template": "The \"a boy with blue eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505], "bbox": [0.44420312500000003, 0.21834375, 0.97321875, 0.957171875], "iscrowd": 0, "area": 107822.56530000003, "rle": {"size": [640, 640], "counts": "]ma52lc06J5K5K3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N3M3N2M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N3L3N2N2N2M3N2N23MO1N2O1O1N2O2M2O2M2O2M2O2WM]LPEe3l:_LSEc3h:aLVEb3f:bLXE_3d:eLZE^3a:fL^E[3^:jL_EY3\\:kLcEV3Y:oLeES3V:QMhER3S:RMlEQ3n9TMoEP3l9SMSFP3g9TMWFP3c9UMZFo2`9UM\\FQ3^9TM]FR3]9RM`FS3Z9QMbFU3X9m3L5J6J6J6L4K5K5K5L4K5K4L5K5L4K5K5K5K5L4K5K4L2O1N2O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001PHZJd1f5ZN]Jf1b5YN`Jf1`5YNbJf1_5WNdJh1\\5WNeJi1[5UNhJj1X5UNjJj1W5TNkJk1U5SNnJl1R5SNPKl1P5SNQKm1P5PNSKo1n4oMTKP2n4lMUKS2n4iMTKV2n4gMSKY2o4cMTK\\2n4aMTK^2o4^MSKa2o4[MTKd2n4YMTKf2n4VMTKj2o4RMSKm2o4PMSKo2o4mLTKR3m4kLVKT3k4iLXKV3i4gLYKY3i4cLZK\\3g4aL\\K^3e4_L^K`3c4]L`Kb3a4[LbKd3_4YLcKg3^4VLeKi3\\4TLgKk3Z4RLiKm3X4PLkKo3V4nKlKR4U4kKnKT4S4iKPLV4R4fKQLY4]4WKeKi4n4bJUK]5\\5PJgJo5i5_IZJ`6c901O1O001O1O1QKeBW3[=hLgBW3Z=hLhBV3Y=iLhBV3Y=hLiBW3W=iLkBU3V=jLkBU3V=iLmBU3S=kLnBT3S=kLnBT3S=jLPCT3Q=kLPCT3P=lLRCR3o<lLSCS3n<lLSCS3n<lLTCR3l<mLVCR3k<mLWCQ3j<nLWCQ3i<nLYCQ3h<nLZCP3g<oLZCP3g<nL\\CP3d<PM]Co2d<PM]Co2d<oL_Co2a<QM`Cn2a<QMaCm2`<QMbCn2_<QMcCn2]<QMdCn2]<PMfCn2[<QMfCn2[<QMgCm2Z<QMhCn2Y<QMiCm2X<RMiCm2X<QMkCa2b<^M_CP2S=oMoB^1c=`N_Bn0S>QOoA=b>B_ALS?2_2O1O[j:"}, "label": "a boy with blue eyes"}]}
{"id": 371824, "image": "COCO_train2014_000000371824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smiling boy\"."}], "task": "refering", "answer_template": "The \"the smiling boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 355, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496], "bbox": [0.0515625, 0.206546875, 0.605375, 0.95765625], "iscrowd": 0, "area": 120948.14275000001, "rle": {"size": [640, 640], "counts": "lVe09Vc0a0^Ob0_Oa0_Oa0^Ob0_Oa0D<N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1PKbK]J_4X5QLbJP4S5_LhJb3n4mLkJU3j4ZMQKg2e4hMUKY2`4VN[Kk1a4^NXKd1f4cNTK^1j4hNQKY1m4nNlJT1R5POkJQ1S5ROkJo0S5TOkJm0S5VOjJl0T5WOjJj0T5YOjJh0U5ZOhJh0V5ZOiJg0U5\\OiJe0U5^OhJd0V5_OhJb0V5AhJ`0V5CgJ?W5DgJ=W5FgJ;X5GfJ:Y5GfJ:Y5HfJ8Z5IdJ8[5JcJ7\\5KbJ6]5LbJ4]5NaJ3^5O`J2_5O`J2_51_JO`53^JNa55\\JLc56[JKd58YJIf5:WJGi5;TJFk5=RJDm5X7O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100000000O10000000000O100000000O10000000000O10000000000O10000000000O10000000000O1001O001O00001O001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O001O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1XJPJ^MP6`2SJ_Mn5_2UJ_Ml5^2WJaMi5^2YJaMh5\\2\\JbMe5[2^JdMd5Y2^JfMc5W2`JhMa5V2bJhM`5T2cJkM^5S2dJlM]5Q2fJnM\\5o1gJoMZ5n1iJQNX5m1jJRNX5j1kJUNV5i1lJVNU5g1nJXNT5d1nJ\\NS5b1oJ]NR5`1QK_NQ5^1PKbNQ5[1RKdNo4Z1SKeNo4W1SKiNn4U1TKjNn4R1UKmNl4Q1UKoNl4n0WKQOk4l0WKSOj4k0WKUOj4i0XKVOj4g0XKXOi4f0XKZOi4d0YK[Oi4a0ZK^Og4`0ZK@g4>[KAg4<[KCf4;[KEf49\\KFf47\\KHe45]KKd43^KLd41^KNc40^K0c4N_K1c4L_K3b4K_K5b4I`K6b4FaK9`4EaK;a4BaK=`4AbK>_4@bK`0`4]ObKb0`4[ObKd0`4XObKh0`4UObKj0a4ROaKm0a4lNdKT1^4eNhKZ1Z4_NlK`1V4XNQLg1Q4RNVLl1l3mMZLR2h3gM_LW2d3aMcL]2_3\\MgLc2[3VMlLh2V3QMPMn2R3kLUMS3m2fLYMY3i2`L^M^3c9O2N2N1O2N1O2N1O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3MWam4"}, "label": "the smiling boy"}]}
{"id": 371824, "image": "COCO_train2014_000000371824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smiling child on the left\"."}], "task": "refering", "answer_template": "The \"smiling child on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 355, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496], "bbox": [0.0515625, 0.206546875, 0.605375, 0.95765625], "iscrowd": 0, "area": 120948.14275000001, "rle": {"size": [640, 640], "counts": "lVe09Vc0a0^Ob0_Oa0_Oa0^Ob0_Oa0D<N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1PKbK]J_4X5QLbJP4S5_LhJb3n4mLkJU3j4ZMQKg2e4hMUKY2`4VN[Kk1a4^NXKd1f4cNTK^1j4hNQKY1m4nNlJT1R5POkJQ1S5ROkJo0S5TOkJm0S5VOjJl0T5WOjJj0T5YOjJh0U5ZOhJh0V5ZOiJg0U5\\OiJe0U5^OhJd0V5_OhJb0V5AhJ`0V5CgJ?W5DgJ=W5FgJ;X5GfJ:Y5GfJ:Y5HfJ8Z5IdJ8[5JcJ7\\5KbJ6]5LbJ4]5NaJ3^5O`J2_5O`J2_51_JO`53^JNa55\\JLc56[JKd58YJIf5:WJGi5;TJFk5=RJDm5X7O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100000000O10000000000O100000000O10000000000O10000000000O10000000000O10000000000O1001O001O00001O001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O001O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1XJPJ^MP6`2SJ_Mn5_2UJ_Ml5^2WJaMi5^2YJaMh5\\2\\JbMe5[2^JdMd5Y2^JfMc5W2`JhMa5V2bJhM`5T2cJkM^5S2dJlM]5Q2fJnM\\5o1gJoMZ5n1iJQNX5m1jJRNX5j1kJUNV5i1lJVNU5g1nJXNT5d1nJ\\NS5b1oJ]NR5`1QK_NQ5^1PKbNQ5[1RKdNo4Z1SKeNo4W1SKiNn4U1TKjNn4R1UKmNl4Q1UKoNl4n0WKQOk4l0WKSOj4k0WKUOj4i0XKVOj4g0XKXOi4f0XKZOi4d0YK[Oi4a0ZK^Og4`0ZK@g4>[KAg4<[KCf4;[KEf49\\KFf47\\KHe45]KKd43^KLd41^KNc40^K0c4N_K1c4L_K3b4K_K5b4I`K6b4FaK9`4EaK;a4BaK=`4AbK>_4@bK`0`4]ObKb0`4[ObKd0`4XObKh0`4UObKj0a4ROaKm0a4lNdKT1^4eNhKZ1Z4_NlK`1V4XNQLg1Q4RNVLl1l3mMZLR2h3gM_LW2d3aMcL]2_3\\MgLc2[3VMlLh2V3QMPMn2R3kLUMS3m2fLYMY3i2`L^M^3c9O2N2N1O2N1O2N1O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3MWam4"}, "label": "smiling child on the left"}]}
{"id": 27763, "image": "COCO_train2014_000000027763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike covered in yellow stickers\"."}], "task": "refering", "answer_template": "The \"a bike covered in yellow stickers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 29, 30, 31, 54, 55, 56, 59, 60, 61, 62, 77, 78, 79, 80, 83, 84, 85, 86, 87, 88, 89, 90, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 194, 196, 197, 198, 199, 201, 202, 203, 204, 205, 206, 211, 212, 213, 214, 215, 216, 219, 220, 221, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 242, 243, 249, 250, 251, 252, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 282, 283, 294], "bbox": [0.16703125000000002, 0.020708333333333332, 1.0, 0.7209166666666668], "iscrowd": 0, "area": 88243.65434999997, "rle": {"size": [480, 640], "counts": "m[b12l>5K4M4K5K5K5L4K6J6K4K6J5L5J5eDkMc9[2PFoMn9V2eESNY:R2[EXNb:T3N3N1N2N3N1O2N1O1O2N1O2N1O2N1O1N3N1O2N1O1O2N1O2N1O2O0O101N101O0O101N101N101O0O101N101N10001N101N101N10001N101N1O1O2N1O2N1O1O001jLdIgM]6V2fIiM[6T2hIlMW6R2mImMS6R2nInMR6R2nInMR6S2mImMR6j1fI\\K8i2R6j1hI]K5i2S6i1iI^K5h2Q6i1mI_K2g2Q6i1oI`K0f2Q6i1QJaKNe2Q6h1SJcKMd2o5i1VJcKKc2o5k1VJbKLa2n5n1UJbKM_2n5Q2TJ_KO_2l5S2VJ^KN^2l5U2VJ]KN]2l5W2UJ]KO[2l5X2VJ\\KN\\2k5Z2WJZKM\\2l5Z2XJZKL[2l5[2XJ[KKZ2m5[2YJZKKZ2k5]2[JYKIZ2l5^2ZJYKIY2m5_2ZJXKIX2l5a2\\JVKHY2k5c2]JRKJZ2h5f2]JoJL[2f5h2^JkJM]2d5j2_JhJN]2b5l2`JfJO^2`5n2SKaKPO>l5S4SKYKYOb0c5W4SKQKBf0Z5[4TKjJHh0T5`4SKgJLf0Q5d4TKeJBHDl0f5i4SKcJ]O0Ja0f5n4RKaJXO8O7g5R5RK^JTO?3Ng5W5QK[JVOb02Ig5\\5QKXJVOe02Dg5`5QKWJVOg02_Og5e5PKUJWOi02ZOh5h5PKTJVOl02ROk5P6lJRJXOT1n5k4kJoIXOS1o5P5hJmIYOo0R6U5eJlIYOl0T6Z5cJiIYOi0W6_5`JhIYOf0Y6d5^JeIYOd0[6h5fIdI;0D`0^6n5`IeI>MD<`6S6\\IgIi02n5e6VJSIl5o6Q12N2N2mJkHX2W7dMnHY2T7fMmHX2V7gMjHW2X7iMhHU2Z7kMfHS2\\7nMcHP2^7QNcHl1_7TNaHj1`7WN`Hg1a7ZN_Hd1b7]N^Ha1d7_N\\H_1e7bN\\HZ1f7gNZHV1h7lNWHQ1k7POUHm0n7SORHj0P8WOQHe0R8[OnGb0T8_OlG>W8BiG<X8EhG8[8HeG5d8F\\G6g8JYG3i8NWGOl81TGLo84QGIR97nFFT9;mFAV9?jF^OY9c0fFZO]9f0cFWO`9i0`FTOd9k0_FPOc9P1aFiNb9W1aFcNb9]1Q2000O2O0000000O10[MmC\\2S<aMoC`2Q<]MQDd2U<2N100O1O1O1O100O1O2N1O100O1O1O1O1O2O0O1O1O1O100O1O1O2N100O1O1O1O100O1O2N1O100O1O1O1O101N1O1O1O1O100O5K5K5[FbK^8c4YGaKh8c4nFbKR9c4eFaK[9R501O01N1O1O1O1O2N1O100O1O2N1001O001O100O1O1OO101O00_NiJYIV5g6oJUIQ5k6oJUIQ5k6oJVIP5j6PKVIP5j6QKUIl4n6TKRIj4P7VKPIh4R7XKoHd4T7\\KlHb4V7^KjH_4Y7bKfH\\4\\7dKdHZ4]7gKdHU4_7kKaHS4a7mK_HQ4c7PL\\Hm3g7SLYHk3i7ULWHi3k7WLVHf3l7ZLTHc3o7^LoGa3S8_LjGb3X8^LfGa3]8_L`Gb3a8`L\\G`3f8aLVG`3l8`LRGa3o8`LmFb3T9_LhFd3X9\\LfFe3[9\\LaFf3`9[L]Ff3d9[LXFh3h9g01O1hLWFX1i9gNZFX1f9fN\\FY1e9dN^F\\1b9bN`F]1a9`NbF`1^9^NeFa1[9]NgFb1Z9[NiFe1W9YNkFg1U9VNnFi1S9UNoFk1Q9RNSGm1l8RNVGn1j8QNWGo1i8QNWGn1j8RNVGn1j8RNWGm1i8SNWGm1i8SNWGm1i8SNWGm1i8SNWGm1i8SNWGm1i8SNUGn1l8RNSGo1l8RNRGP2n8PNQGQ2o8oMoFS2Q9mMnFT2R9lMmFU2S9kMkFV2V9jMiFW2W9iMgFY2Y9gMfFZ2Z9fMeF[2[9dMdF^2\\9aMdF`2[9aMcFa2^9]MbFc2e9VMZFl2l9mLTFT3R:eLnE\\3i:2O1N2O0O2N2M31O001O1O010O010O01O00001O001O010O00001O001O0\\HPLX4Q4fKPLZ4P4fKQLY4P4eKQL[4o3eKQL[4o3dKRL\\4o3bKSL]4m3cKSL]4n3aKSL_4m3aKSL_4n3_KTL`4l3`KTL`4l3_KULa4l3]KULc4k3]KULc4l3[KULe4k3[KULe4l3YKULg4k3XKVLh4j3XKVLh4k3VKULk4k3UKULk4l3SKULm4k3SKULm4l3QKULo4k3PKVLP5k3oJULQ5m3lJTLT5o3iJQLW5Q4fJPLZ5R4cJoK]5T4`JlK`5V4]JkKc5W4[JiKe5Y4XJhKh5[4UJeKk5]4PJeKQ6]4gIiKY6Y4_ImKa6V4VIPLj6R4nHTLR7n3fHXLZ7g50000000000000000000O10000000001O000000000000000000000O10000000000000000000000O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1SJjGo4W8oJlGo4U8nJoGQ5Q8mJSHP5n7mJVHQ5k7lJYHR5h7lJ[HS5e7jJ_HT5b7jJaHT5`7iJdHU5]7iJfHV5c8N2N2N2O1N2N2N2O1N2M3M3M3M3M3M3L4M3M3M3M3eLhDj2[;TMiDg2[;UMiDg2Z;WMiDe2Z;XMjDd2l;L5J5L4L4L5H7@`0A?@oI"}, "label": "a bike covered in yellow stickers"}]}
{"id": 27763, "image": "COCO_train2014_000000027763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bicycle with a small white dog in front of it\"."}], "task": "refering", "answer_template": "The \"a yellow bicycle with a small white dog in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 29, 30, 31, 54, 55, 56, 59, 60, 61, 62, 77, 78, 79, 80, 83, 84, 85, 86, 87, 88, 89, 90, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 194, 196, 197, 198, 199, 201, 202, 203, 204, 205, 206, 211, 212, 213, 214, 215, 216, 219, 220, 221, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 242, 243, 249, 250, 251, 252, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 282, 283, 294], "bbox": [0.16703125000000002, 0.020708333333333332, 1.0, 0.7209166666666668], "iscrowd": 0, "area": 88243.65434999997, "rle": {"size": [480, 640], "counts": "m[b12l>5K4M4K5K5K5L4K6J6K4K6J5L5J5eDkMc9[2PFoMn9V2eESNY:R2[EXNb:T3N3N1N2N3N1O2N1O1O2N1O2N1O2N1O1N3N1O2N1O1O2N1O2N1O2O0O101N101O0O101N101N101O0O101N101N10001N101N101N10001N101N1O1O2N1O2N1O1O001jLdIgM]6V2fIiM[6T2hIlMW6R2mImMS6R2nInMR6R2nInMR6S2mImMR6j1fI\\K8i2R6j1hI]K5i2S6i1iI^K5h2Q6i1mI_K2g2Q6i1oI`K0f2Q6i1QJaKNe2Q6h1SJcKMd2o5i1VJcKKc2o5k1VJbKLa2n5n1UJbKM_2n5Q2TJ_KO_2l5S2VJ^KN^2l5U2VJ]KN]2l5W2UJ]KO[2l5X2VJ\\KN\\2k5Z2WJZKM\\2l5Z2XJZKL[2l5[2XJ[KKZ2m5[2YJZKKZ2k5]2[JYKIZ2l5^2ZJYKIY2m5_2ZJXKIX2l5a2\\JVKHY2k5c2]JRKJZ2h5f2]JoJL[2f5h2^JkJM]2d5j2_JhJN]2b5l2`JfJO^2`5n2SKaKPO>l5S4SKYKYOb0c5W4SKQKBf0Z5[4TKjJHh0T5`4SKgJLf0Q5d4TKeJBHDl0f5i4SKcJ]O0Ja0f5n4RKaJXO8O7g5R5RK^JTO?3Ng5W5QK[JVOb02Ig5\\5QKXJVOe02Dg5`5QKWJVOg02_Og5e5PKUJWOi02ZOh5h5PKTJVOl02ROk5P6lJRJXOT1n5k4kJoIXOS1o5P5hJmIYOo0R6U5eJlIYOl0T6Z5cJiIYOi0W6_5`JhIYOf0Y6d5^JeIYOd0[6h5fIdI;0D`0^6n5`IeI>MD<`6S6\\IgIi02n5e6VJSIl5o6Q12N2N2mJkHX2W7dMnHY2T7fMmHX2V7gMjHW2X7iMhHU2Z7kMfHS2\\7nMcHP2^7QNcHl1_7TNaHj1`7WN`Hg1a7ZN_Hd1b7]N^Ha1d7_N\\H_1e7bN\\HZ1f7gNZHV1h7lNWHQ1k7POUHm0n7SORHj0P8WOQHe0R8[OnGb0T8_OlG>W8BiG<X8EhG8[8HeG5d8F\\G6g8JYG3i8NWGOl81TGLo84QGIR97nFFT9;mFAV9?jF^OY9c0fFZO]9f0cFWO`9i0`FTOd9k0_FPOc9P1aFiNb9W1aFcNb9]1Q2000O2O0000000O10[MmC\\2S<aMoC`2Q<]MQDd2U<2N100O1O1O1O100O1O2N1O100O1O1O1O1O2O0O1O1O1O100O1O1O2N100O1O1O1O100O1O2N1O100O1O1O1O101N1O1O1O1O100O5K5K5[FbK^8c4YGaKh8c4nFbKR9c4eFaK[9R501O01N1O1O1O1O2N1O100O1O2N1001O001O100O1O1OO101O00_NiJYIV5g6oJUIQ5k6oJUIQ5k6oJVIP5j6PKVIP5j6QKUIl4n6TKRIj4P7VKPIh4R7XKoHd4T7\\KlHb4V7^KjH_4Y7bKfH\\4\\7dKdHZ4]7gKdHU4_7kKaHS4a7mK_HQ4c7PL\\Hm3g7SLYHk3i7ULWHi3k7WLVHf3l7ZLTHc3o7^LoGa3S8_LjGb3X8^LfGa3]8_L`Gb3a8`L\\G`3f8aLVG`3l8`LRGa3o8`LmFb3T9_LhFd3X9\\LfFe3[9\\LaFf3`9[L]Ff3d9[LXFh3h9g01O1hLWFX1i9gNZFX1f9fN\\FY1e9dN^F\\1b9bN`F]1a9`NbF`1^9^NeFa1[9]NgFb1Z9[NiFe1W9YNkFg1U9VNnFi1S9UNoFk1Q9RNSGm1l8RNVGn1j8QNWGo1i8QNWGn1j8RNVGn1j8RNWGm1i8SNWGm1i8SNWGm1i8SNWGm1i8SNWGm1i8SNWGm1i8SNUGn1l8RNSGo1l8RNRGP2n8PNQGQ2o8oMoFS2Q9mMnFT2R9lMmFU2S9kMkFV2V9jMiFW2W9iMgFY2Y9gMfFZ2Z9fMeF[2[9dMdF^2\\9aMdF`2[9aMcFa2^9]MbFc2e9VMZFl2l9mLTFT3R:eLnE\\3i:2O1N2O0O2N2M31O001O1O010O010O01O00001O001O010O00001O001O0\\HPLX4Q4fKPLZ4P4fKQLY4P4eKQL[4o3eKQL[4o3dKRL\\4o3bKSL]4m3cKSL]4n3aKSL_4m3aKSL_4n3_KTL`4l3`KTL`4l3_KULa4l3]KULc4k3]KULc4l3[KULe4k3[KULe4l3YKULg4k3XKVLh4j3XKVLh4k3VKULk4k3UKULk4l3SKULm4k3SKULm4l3QKULo4k3PKVLP5k3oJULQ5m3lJTLT5o3iJQLW5Q4fJPLZ5R4cJoK]5T4`JlK`5V4]JkKc5W4[JiKe5Y4XJhKh5[4UJeKk5]4PJeKQ6]4gIiKY6Y4_ImKa6V4VIPLj6R4nHTLR7n3fHXLZ7g50000000000000000000O10000000001O000000000000000000000O10000000000000000000000O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1SJjGo4W8oJlGo4U8nJoGQ5Q8mJSHP5n7mJVHQ5k7lJYHR5h7lJ[HS5e7jJ_HT5b7jJaHT5`7iJdHU5]7iJfHV5c8N2N2N2O1N2N2N2O1N2M3M3M3M3M3M3L4M3M3M3M3eLhDj2[;TMiDg2[;UMiDg2Z;WMiDe2Z;XMjDd2l;L5J5L4L4L5H7@`0A?@oI"}, "label": "a yellow bicycle with a small white dog in front of it"}]}
{"id": 27763, "image": "COCO_train2014_000000027763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rear tire of a black bike\"."}], "task": "refering", "answer_template": "The \"the rear tire of a black bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 277, 278, 279], "bbox": [0.0064218750000000005, 0.21872916666666667, 0.1895, 0.7390625], "iscrowd": 0, "area": 18694.79935, "rle": {"size": [480, 640], "counts": "Yo1;g>Q1nNS1mNS1nNR1mNS1mNh0YOO000O101O00001N10001O0O101O0O101O0O101N10000O100O10000O100O1O100O100O100O100O100O100O100O1O100O100O100O100O2O0O100O2N1O1O1O2N1O1O1O2O0O1O1O2N1O1O2N1O1O1O2N1O1O101N1O1O2M2O2M2N3N1N3N1N3M2O2M2N2M4M2N3M2N3M4L6I7J6J6J6J6G9AePc7"}, "label": "the rear tire of a black bike"}]}
{"id": 27763, "image": "COCO_train2014_000000027763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color cycle front wheel near by yellow colored cycle \\ and puppy dog\"."}], "task": "refering", "answer_template": "The \"a black color cycle front wheel near by yellow colored cycle \\ and puppy dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 277, 278, 279], "bbox": [0.0064218750000000005, 0.21872916666666667, 0.1895, 0.7390625], "iscrowd": 0, "area": 18694.79935, "rle": {"size": [480, 640], "counts": "Yo1;g>Q1nNS1mNS1nNR1mNS1mNh0YOO000O101O00001N10001O0O101O0O101O0O101N10000O100O10000O100O1O100O100O100O100O100O100O100O1O100O100O100O100O2O0O100O2N1O1O1O2N1O1O1O2O0O1O1O2N1O1O2N1O1O1O2N1O1O101N1O1O2M2O2M2N3N1N3N1N3M2O2M2N2M4M2N3M2N3M4L6I7J6J6J6J6G9AePc7"}, "label": "a black color cycle front wheel near by yellow colored cycle \\ and puppy dog"}]}
{"id": 35964, "image": "COCO_train2014_000000035964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a plaid shirt to the right of two other women\"."}], "task": "refering", "answer_template": "The \"a woman in a plaid shirt to the right of two other women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 134, 135, 136, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 249, 252], "bbox": [0.757421875, 0.25843457943925235, 0.999375, 0.705607476635514], "iscrowd": 0, "area": 18346.201800000003, "rle": {"size": [428, 640], "counts": "PnZ6e0d<o0RO5J3N2M3N3L3N2O100O2O0O100O100O2O0O100O101N100O100O101N100O100O2O0O100O100O2O0O100O100O1_NTMbHl2W7`McHa2Y7gMcHY2[7nM_HS2_7SN]Hm1a7YN[Hg1c7_NXHb1f7dNVH\\1h7jNSHW1m7mNoGS1P8U2O100O100O00100O100000000001O000000001O000000001O0000001O000000001O0000001O0000001VLWHo1i7PN\\Hl1d7SN]Hm1c7QN_Ho1b7oM`HP2`7oMaHQ2_7mMcHS2]7lMeHS2[7kMgHU2Z7iMgHW2Y7QM_Io2`8O0000001O0000001O00001O0000001O0000001O0001O01O01O10O01O1O100O1O1O100O1O1O100O1OO2N100OW1iNQ1POn0QOkF"}, "label": "a woman in a plaid shirt to the right of two other women"}]}
{"id": 35964, "image": "COCO_train2014_000000035964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a red plaid shirt\"."}], "task": "refering", "answer_template": "The \"woman in a red plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 134, 135, 136, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 249, 252], "bbox": [0.757421875, 0.25843457943925235, 0.999375, 0.705607476635514], "iscrowd": 0, "area": 18346.201800000003, "rle": {"size": [428, 640], "counts": "PnZ6e0d<o0RO5J3N2M3N3L3N2O100O2O0O100O100O2O0O100O101N100O100O101N100O100O2O0O100O100O2O0O100O100O1_NTMbHl2W7`McHa2Y7gMcHY2[7nM_HS2_7SN]Hm1a7YN[Hg1c7_NXHb1f7dNVH\\1h7jNSHW1m7mNoGS1P8U2O100O100O00100O100000000001O000000001O000000001O0000001O000000001O0000001O0000001VLWHo1i7PN\\Hl1d7SN]Hm1c7QN_Ho1b7oM`HP2`7oMaHQ2_7mMcHS2]7lMeHS2[7kMgHU2Z7iMgHW2Y7QM_Io2`8O0000001O0000001O00001O0000001O0000001O0001O01O01O10O01O1O100O1O1O100O1O1O100O1OO2N100OW1iNQ1POn0QOkF"}, "label": "woman in a red plaid shirt"}]}
{"id": 35964, "image": "COCO_train2014_000000035964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in center\"."}], "task": "refering", "answer_template": "The \"woman in center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.40554687500000003, 0.2636214953271028, 0.66709375, 0.662126168224299], "iscrowd": 0, "area": 17673.5551, "rle": {"size": [428, 640], "counts": "mm\\3<m<3M3L5L3M3M3M3L4N2N2@fNlD\\1S;gNhD\\1W;gNeD[1Z;<O1O00001O00010O00001O00001O100O1O1N2O1O1N2O1O1[Oe0kNU1N2O1N2N2O1N2N2O1N2N2O2M3O1N2O1N2O1O0O100O100O10000O1000000000000000000000000000000000000010O001O001O001O0010O01O001O001O001O001O002N2N2N1O2N2N1O2N2M2O]1cN2N3M2N2N3M00001O0O2O001O00001O001O00001N101O001O0O101N101O0O3N2M4M2N2M4M2M4M2M3N3M2M4M2M3N3M2M2O2J5J7ITih2"}, "label": "woman in center"}]}
{"id": 35964, "image": "COCO_train2014_000000035964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black and pink jacket reaching for an item on a table\"."}], "task": "refering", "answer_template": "The \"a woman in a black and pink jacket reaching for an item on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.40554687500000003, 0.2636214953271028, 0.66709375, 0.662126168224299], "iscrowd": 0, "area": 17673.5551, "rle": {"size": [428, 640], "counts": "mm\\3<m<3M3L5L3M3M3M3L4N2N2@fNlD\\1S;gNhD\\1W;gNeD[1Z;<O1O00001O00010O00001O00001O100O1O1N2O1O1N2O1O1[Oe0kNU1N2O1N2N2O1N2N2O1N2N2O2M3O1N2O1N2O1O0O100O100O10000O1000000000000000000000000000000000000010O001O001O001O0010O01O001O001O001O001O002N2N2N1O2N2N1O2N2M2O]1cN2N3M2N2N3M00001O0O2O001O00001O001O00001N101O001O0O101N101O0O3N2M4M2N2M4M2M4M2M3N3M2M4M2M3N3M2M2O2J5J7ITih2"}, "label": "a woman in a black and pink jacket reaching for an item on a table"}]}
{"id": 35964, "image": "COCO_train2014_000000035964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady is putting chili on the meat ball sandwich\"."}], "task": "refering", "answer_template": "The \"the lady is putting chili on the meat ball sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 71, 72, 73, 74, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 213, 214], "bbox": [0.0090625, 0.18679906542056077, 0.33059375, 0.6179439252336449], "iscrowd": 0, "area": 24313.278599999998, "rle": {"size": [428, 640], "counts": "ce24U=5L5J6M3L4M2N3L4M0O2O000O2O001N10001N100O104L6I8I6J7H?B1N101O0O2O001N101N101N101N10001N101N100O100O100O1\\ObLYG_3f8cLYG]3e8fLYG[3f8gLXGZ3g8hLWGZ3g8hLWGY3h8iLVGX3i8jLVGV3h8nLUGS3j8h0O1O1O1O1O1N2O1O1O1O1O2M2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O2N100001O0000001O0000001O000000001N1000001O0000001O0000001O001O1O2N1O2N1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O100O101N100O100O1O101N100O1O1O2N1O2N101N1O2N1O2N1O2N101N1O2N1O2N1O2M2M4L3M4L3M4L3M4L3M4L3M4jN[gb5"}, "label": "the lady is putting chili on the meat ball sandwich"}]}
{"id": 35964, "image": "COCO_train2014_000000035964.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl was cooking the food and serving\"."}], "task": "refering", "answer_template": "The \"a girl was cooking the food and serving\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 71, 72, 73, 74, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 213, 214], "bbox": [0.0090625, 0.18679906542056077, 0.33059375, 0.6179439252336449], "iscrowd": 0, "area": 24313.278599999998, "rle": {"size": [428, 640], "counts": "ce24U=5L5J6M3L4M2N3L4M0O2O000O2O001N10001N100O104L6I8I6J7H?B1N101O0O2O001N101N101N101N10001N101N100O100O100O1\\ObLYG_3f8cLYG]3e8fLYG[3f8gLXGZ3g8hLWGZ3g8hLWGY3h8iLVGX3i8jLVGV3h8nLUGS3j8h0O1O1O1O1O1N2O1O1O1O1O2M2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O2N100001O0000001O0000001O000000001N1000001O0000001O0000001O001O1O2N1O2N1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O100O101N100O100O1O101N100O1O1O2N1O2N101N1O2N1O2N1O2N101N1O2N1O2N1O2M2M4L3M4L3M4L3M4L3M4L3M4jN[gb5"}, "label": "a girl was cooking the food and serving"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white and red motorcycle\"."}], "task": "refering", "answer_template": "The \"the white and red motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 292, 293, 294], "bbox": [0.458015625, 0.4923743016759777, 0.9022656250000001, 0.9599441340782122], "iscrowd": 0, "area": 27717.258599999997, "rle": {"size": [358, 640], "counts": "UhV31T;1O1O1O2N1O1O1O2N1O1O1hLDeK>Z4CdK>\\4CaK?_4B_K?a4B]K`0b4BXKb0g4@UKc0k4^OQKe0o4]OlJf0T5\\OfJh0Z5ZOaJi0_5YO\\Jj0d5XOWJk0i5WORJl0n5VOlIm0U6R20000000O100O1O100O100O2N100O100O1O1O2O0O1O100O1OVOlLeIR3\\6PMbIP3]6TM`Ik2a6VM_Ih2a6YM_If2b6[M]Ic2d6^M]I]2g6dMXIX2k6iMUIV2l6kMSI[2f6fMXIa2c6`MZIc2d6^M[Ic2e6^MXId2g6]MXId2h6]MUIe2k6[MTIf2l6j0O10000O10000O1001O00001O0000001O1O001O1O1O001O1O001O1O1O2N1O2N1O01O000000000O10000000000O100000000000000000000001O00000000000000000000000000000000kHWLm6i3QIYLo6g3oHZLR7m30O100O100O100O101N1O2O0O2N101N2N101N3M34L1O2N1O2N2NO1O1N2O1N10O001O001O001O0010O010O001O1O001O1O0O100000000O10000000000O100000O10O100000000O100000000O010000O10O10O10000O10O10O10cM[HY1e7gN[HY1d7gN^HW1c7iN^HU1b7kN`HS1a7mN`HQ1a7nN`HP1a7POaHn0`7ROaHl0`7SObHk0^7VObHi0_7VOcHh0]7YOdHd0^7[OgH?[7AkH6X7IoHOS70TIHn68WI@l6?[IYOg6g0l10000000000_OhE4X:=0^OhE6X:JhE6X:IhE8X:GiE9W:GhE:X:EiE;W:EiE;W:DjE<V:DkE;U:DlE<T:CmE<T:DlE<T:CmE=S:CmE=S:BnE=S:BnE>S:AnE>]:00O3N3M3M3MXie0"}, "label": "the white and red motorcycle"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bike\"."}], "task": "refering", "answer_template": "The \"a white bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 292, 293, 294], "bbox": [0.458015625, 0.4923743016759777, 0.9022656250000001, 0.9599441340782122], "iscrowd": 0, "area": 27717.258599999997, "rle": {"size": [358, 640], "counts": "UhV31T;1O1O1O2N1O1O1O2N1O1O1hLDeK>Z4CdK>\\4CaK?_4B_K?a4B]K`0b4BXKb0g4@UKc0k4^OQKe0o4]OlJf0T5\\OfJh0Z5ZOaJi0_5YO\\Jj0d5XOWJk0i5WORJl0n5VOlIm0U6R20000000O100O1O100O100O2N100O100O1O1O2O0O1O100O1OVOlLeIR3\\6PMbIP3]6TM`Ik2a6VM_Ih2a6YM_If2b6[M]Ic2d6^M]I]2g6dMXIX2k6iMUIV2l6kMSI[2f6fMXIa2c6`MZIc2d6^M[Ic2e6^MXId2g6]MXId2h6]MUIe2k6[MTIf2l6j0O10000O10000O1001O00001O0000001O1O001O1O1O001O1O001O1O1O2N1O2N1O01O000000000O10000000000O100000000000000000000001O00000000000000000000000000000000kHWLm6i3QIYLo6g3oHZLR7m30O100O100O100O101N1O2O0O2N101N2N101N3M34L1O2N1O2N2NO1O1N2O1N10O001O001O001O0010O010O001O1O001O1O0O100000000O10000000000O100000O10O100000000O100000000O010000O10O10O10000O10O10O10cM[HY1e7gN[HY1d7gN^HW1c7iN^HU1b7kN`HS1a7mN`HQ1a7nN`HP1a7POaHn0`7ROaHl0`7SObHk0^7VObHi0_7VOcHh0]7YOdHd0^7[OgH?[7AkH6X7IoHOS70TIHn68WI@l6?[IYOg6g0l10000000000_OhE4X:=0^OhE6X:JhE6X:IhE8X:GiE9W:GhE:X:EiE;W:EiE;W:DjE<V:DkE;U:DlE<T:CmE<T:DlE<T:CmE=S:CmE=S:BnE=S:BnE>S:AnE>]:00O3N3M3M3MXie0"}, "label": "a white bike"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in the red , whte and blue outfit leaning towards the white bus\"."}], "task": "refering", "answer_template": "The \"man in the red , whte and blue outfit leaning towards the white bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 149, 151, 152], "bbox": [0.484953125, 0.16762569832402235, 0.691421875, 0.5345810055865922], "iscrowd": 0, "area": 8339.613000000001, "rle": {"size": [358, 640], "counts": "na\\33Z:k0I5L2000O100O1O100O11O0100O010O11O001O1O0O2O1O001O1O001N1O2N2N2N1O2N2N2N2O1O100O010O100O100O100O010O100O100O100O100O100O10000000000O10000000000eGoMa7Q2XHWNg7j1UHYNj7j1QHZNn7_20O100O1000O0100O1O1N2O1O000O10000O10O10O01000O01000O010O100O2O1N2N3N1N2O1N2O1N2N3N1N2N2N2YOkF[OW9d0jF[OX9c0jFZOY9d0iFYOZ9e0hFXO\\9e0eFYO^9d0eFZO^9b0eF[O^9c0d0M7I:F[iT2"}, "label": "man in the red , whte and blue outfit leaning towards the white bus"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow race bike under repair\"."}], "task": "refering", "answer_template": "The \"a yellow race bike under repair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 259, 260, 261, 262, 263, 264, 279, 281, 282, 283, 284, 285, 286], "bbox": [0.11821875, 0.2289385474860335, 0.50471875, 0.9786871508379889], "iscrowd": 0, "area": 40340.48094999999, "rle": {"size": [358, 640], "counts": "Ukj0?Y:b0F:I6J7I7I7H5L4L4M3N2N3L3N2N2M3N2N1N2O1iNgL`JZ3^5hL`JZ3_5hL_JY3_5iL`JX3_5jL`JV3^5lLaJU3^5mLaJS3]5oLcJQ3]5oLbJR3]5oLcJQ3\\5PMcJQ3\\5PMdJP3\\5PMcJQ3\\5PMdJP3[5PMfJP3Z5PMeJP3[5QMeJn2[5RMeJn2[5SMeJm2[5RMcJP3S3\\L_Ne0\\NP3V3[L]Ne0\\NQ3W3[L\\Ne0ZNR3[3ZLZNc0ZNT3]3YLXNd0XNV3`3WLVNc0YNW3a3WLUNc0WNX3e3ULTNb0VNZ3g3ULQNb0UN[3k3SLoMb0UN]3l3QLoMb0SN^3n3RLnMg4S2YKmMf4T2ZKmMd4T2\\KlMc4T2^KlM_4W2aK[MIkNd4l3cKUMOmN[4Q4gKmL4POS4T4jKiL9lNo3]4hKcLn4`3RK\\LP5f3PKWLP5l3PKPLR5Q4oJmKP5V4PKhKP5Z4QKdKm4X4gJ^K<8m4\\4kJWK9<j4`4\\K^Kd4d4\\K[Kb4h4^KVKa4m4_KRK`4P5`KnJ_4j1RKW1`0mL\\4o1TKU1?kL\\4S2TKS1^5PObJQ1]5oNdJQ1[5oNeJR1Z5oNeJR1Z5nNfJS1Y5mNgJT1Y5lNfJU1Y5kNgJV1X5jNhJW1W5jNhJV1X5iNiJW1W5eNmJ\\1S5aNoJ_1Q5aNPK^1P5bNPK^1P5bNPK^1P5_NSKb1m4VNZKj1f4nMbKR2^4gMiKY2W4`MPL`2P4XMXLi2g3PM`LP3a3gLgLY3Y3`LoL_3\\5001O00000000001O0004L0000001O000000oHdLX6\\3gIeLY6\\3eIfLZ6Z3fIfLZ6Z3eIgL\\6X3cIjL\\6V3dIjL\\6V3cIlL\\6T3cImL]6S3bInL^6R3bIoL]6Q3bIPM^6P3aIRM^6n2aISM`6m2_ISMa6m2_ITM`6l2`IUM_6k2aIVM^6j2bIWM]6i2cIXMi5ETJS33YMi5CTJT33ZM_5\\O[J64S33[MY5@aJ22S34\\MT5DfJL2T34]Mo4GkJH2T34^Mj4KPKB2V33^Mi4LQK@3V33_Ml4GoJD1V34`Mn4DmJE1W34aMP5@kJH2V34bMR5]OhJJ2W34cMS5ZOgJL2X33aMV5[OeJL1X34aMW5ZOeJL0Y34aMX5ZOcJL1Y34aMY5YOcJL0[33`M[5YOaJL1[34_M[5YOaJb34UM\\5YO_Jb35UM]5XO^Jd34TM_5XO]Jc34UM`5WO\\Jd34UMa5WOZJd35TMc5WOYJd34UMd5VOXJf34SMd5XOWJe35SMd5XOXJd34TMe5XOVJd35TMe5XOVJe34SMf5YOVJc34TMf5YOVJc34TMg5YOTJc35TMV1WOo22gKc34RMU1\\Oo2OhKc34RMR1_OR3MgKb35RMP1BT3IhKb34SMn0EU3GhKa35SMl0GW3EhKb34RMm0HV3`3mKhLl0JW3]3nKgLk0MW3\\3jKkLn0K[OBU3i4_MfKZOCW3g4^MhKYOAZ3g4\\MjKWO@_3e4YMlKWO@`3c4YMoKTO_Od3b4WMQLSO]Og3b4VM_Lk2a3TM`Ll2`3SM`Ln2`3QM`LP3_3PMbLP3^3oLbLQ3_3nLbLR3^3mLcLS3]3lLcLU3]3jLdLV3[3kLeLT3\\3kLdLV3]3hLdLX3\\3gLeLY3\\3eLdL\\3]3bLdL]3]3bLdL^3]3`LdL_3^3_LcLa3^3]LcLb3^3_LaLa3`3^L`La3b3_L]La3b3`L^L`3b3aL]L^3d3cL[L]3e3cL[L\\3f3eLYL[3g3eLYLZ3g3hLXLX3h3hLXLW3i3jLVLT3l3lLULR3l3nLTLQ3m3nLULo2m3PMTLo2l3RMULl2l3SMULk2m3UMTLi2m3VMTLi2m3VMULg2m3YMSLf2m3ZMULd2l3\\MTLb2n3]MTLa2m3^MUL`2l3`MVL\\2l3cMVL[2j3eMYLV2j3jMWLQ2m3nMVLl1n3TNSLg1Q4XNRLb1R4]NPL^1T4bNnKX1U4hNnKR1V4mNmKm0a7C>Bbb^3"}, "label": "a yellow race bike under repair"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow and blue motorcycle\"."}], "task": "refering", "answer_template": "The \"yellow and blue motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 259, 260, 261, 262, 263, 264, 279, 281, 282, 283, 284, 285, 286], "bbox": [0.11821875, 0.2289385474860335, 0.50471875, 0.9786871508379889], "iscrowd": 0, "area": 40340.48094999999, "rle": {"size": [358, 640], "counts": "Ukj0?Y:b0F:I6J7I7I7H5L4L4M3N2N3L3N2N2M3N2N1N2O1iNgL`JZ3^5hL`JZ3_5hL_JY3_5iL`JX3_5jL`JV3^5lLaJU3^5mLaJS3]5oLcJQ3]5oLbJR3]5oLcJQ3\\5PMcJQ3\\5PMdJP3\\5PMcJQ3\\5PMdJP3[5PMfJP3Z5PMeJP3[5QMeJn2[5RMeJn2[5SMeJm2[5RMcJP3S3\\L_Ne0\\NP3V3[L]Ne0\\NQ3W3[L\\Ne0ZNR3[3ZLZNc0ZNT3]3YLXNd0XNV3`3WLVNc0YNW3a3WLUNc0WNX3e3ULTNb0VNZ3g3ULQNb0UN[3k3SLoMb0UN]3l3QLoMb0SN^3n3RLnMg4S2YKmMf4T2ZKmMd4T2\\KlMc4T2^KlM_4W2aK[MIkNd4l3cKUMOmN[4Q4gKmL4POS4T4jKiL9lNo3]4hKcLn4`3RK\\LP5f3PKWLP5l3PKPLR5Q4oJmKP5V4PKhKP5Z4QKdKm4X4gJ^K<8m4\\4kJWK9<j4`4\\K^Kd4d4\\K[Kb4h4^KVKa4m4_KRK`4P5`KnJ_4j1RKW1`0mL\\4o1TKU1?kL\\4S2TKS1^5PObJQ1]5oNdJQ1[5oNeJR1Z5oNeJR1Z5nNfJS1Y5mNgJT1Y5lNfJU1Y5kNgJV1X5jNhJW1W5jNhJV1X5iNiJW1W5eNmJ\\1S5aNoJ_1Q5aNPK^1P5bNPK^1P5bNPK^1P5_NSKb1m4VNZKj1f4nMbKR2^4gMiKY2W4`MPL`2P4XMXLi2g3PM`LP3a3gLgLY3Y3`LoL_3\\5001O00000000001O0004L0000001O000000oHdLX6\\3gIeLY6\\3eIfLZ6Z3fIfLZ6Z3eIgL\\6X3cIjL\\6V3dIjL\\6V3cIlL\\6T3cImL]6S3bInL^6R3bIoL]6Q3bIPM^6P3aIRM^6n2aISM`6m2_ISMa6m2_ITM`6l2`IUM_6k2aIVM^6j2bIWM]6i2cIXMi5ETJS33YMi5CTJT33ZM_5\\O[J64S33[MY5@aJ22S34\\MT5DfJL2T34]Mo4GkJH2T34^Mj4KPKB2V33^Mi4LQK@3V33_Ml4GoJD1V34`Mn4DmJE1W34aMP5@kJH2V34bMR5]OhJJ2W34cMS5ZOgJL2X33aMV5[OeJL1X34aMW5ZOeJL0Y34aMX5ZOcJL1Y34aMY5YOcJL0[33`M[5YOaJL1[34_M[5YOaJb34UM\\5YO_Jb35UM]5XO^Jd34TM_5XO]Jc34UM`5WO\\Jd34UMa5WOZJd35TMc5WOYJd34UMd5VOXJf34SMd5XOWJe35SMd5XOXJd34TMe5XOVJd35TMe5XOVJe34SMf5YOVJc34TMf5YOVJc34TMg5YOTJc35TMV1WOo22gKc34RMU1\\Oo2OhKc34RMR1_OR3MgKb35RMP1BT3IhKb34SMn0EU3GhKa35SMl0GW3EhKb34RMm0HV3`3mKhLl0JW3]3nKgLk0MW3\\3jKkLn0K[OBU3i4_MfKZOCW3g4^MhKYOAZ3g4\\MjKWO@_3e4YMlKWO@`3c4YMoKTO_Od3b4WMQLSO]Og3b4VM_Lk2a3TM`Ll2`3SM`Ln2`3QM`LP3_3PMbLP3^3oLbLQ3_3nLbLR3^3mLcLS3]3lLcLU3]3jLdLV3[3kLeLT3\\3kLdLV3]3hLdLX3\\3gLeLY3\\3eLdL\\3]3bLdL]3]3bLdL^3]3`LdL_3^3_LcLa3^3]LcLb3^3_LaLa3`3^L`La3b3_L]La3b3`L^L`3b3aL]L^3d3cL[L]3e3cL[L\\3f3eLYL[3g3eLYLZ3g3hLXLX3h3hLXLW3i3jLVLT3l3lLULR3l3nLTLQ3m3nLULo2m3PMTLo2l3RMULl2l3SMULk2m3UMTLi2m3VMTLi2m3VMULg2m3YMSLf2m3ZMULd2l3\\MTLb2n3]MTLa2m3^MUL`2l3`MVL\\2l3cMVL[2j3eMYLV2j3jMWLQ2m3nMVLl1n3TNSLg1Q4XNRLb1R4]NPL^1T4bNnKX1U4hNnKR1V4mNmKm0a7C>Bbb^3"}, "label": "yellow and blue motorcycle"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the black sweater and jeans\"."}], "task": "refering", "answer_template": "The \"the man in the black sweater and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 296, 297], "bbox": [0.8847343750000001, 0.1970949720670391, 1.0, 0.9756145251396647], "iscrowd": 0, "area": 13883.311499999998, "rle": {"size": [358, 640], "counts": "jnU62S;6J5J4^HFS4=iKEV4>UK^OPN6j6>mJGUNLm6`0UJ=jNUOP7a0oIQ1i0TNX3l0iKT1i0^NPOEY3k0hLW1g0hNdNFi3=iLW1c0Ia2POjLY1>1e2gNjLZ1:9\\OhMg2g0aMX19W1Q2aMeMY16Z1U2^McMY15\\1W2]MbMX12`1\\2YM`MW1Nf1b2UM^Mc5a2]J^Me5a2\\J]Mf5b2`1O1N10O1O1M3N2M3N2N3M2N3M2M4M2N3M4L5L4K4L5L4K4M4K4L5H7J6K4M3L4M201N1O100O1M3M4K4M3M3L4I8H7K6L3L4M4K4L6IbL"}, "label": "the man in the black sweater and jeans"}]}
{"id": 298110, "image": "COCO_train2014_000000298110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a hooded sweatshirt and glasses\"."}], "task": "refering", "answer_template": "The \"a boy in a hooded sweatshirt and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 296, 297], "bbox": [0.8847343750000001, 0.1970949720670391, 1.0, 0.9756145251396647], "iscrowd": 0, "area": 13883.311499999998, "rle": {"size": [358, 640], "counts": "jnU62S;6J5J4^HFS4=iKEV4>UK^OPN6j6>mJGUNLm6`0UJ=jNUOP7a0oIQ1i0TNX3l0iKT1i0^NPOEY3k0hLW1g0hNdNFi3=iLW1c0Ia2POjLY1>1e2gNjLZ1:9\\OhMg2g0aMX19W1Q2aMeMY16Z1U2^McMY15\\1W2]MbMX12`1\\2YM`MW1Nf1b2UM^Mc5a2]J^Me5a2\\J]Mf5b2`1O1N10O1O1M3N2M3N2N3M2N3M2M4M2N3M4L5L4K4L5L4K4M4K4L5H7J6K4M3L4M201N1O100O1M3M4K4M3M3L4I8H7K6L3L4M4K4L6IbL"}, "label": "a boy in a hooded sweatshirt and glasses"}]}
{"id": 322695, "image": "COCO_train2014_000000322695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a second horse on which a boy is sitting wearing black t - shirt and black cap\"."}], "task": "refering", "answer_template": "The \"a second horse on which a boy is sitting wearing black t - shirt and black cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 240, 241, 243, 264, 265, 266], "bbox": [0.412578125, 0.499131455399061, 0.6087187500000001, 0.7911032863849765], "iscrowd": 0, "area": 5716.098599999998, "rle": {"size": [426, 640], "counts": "_Q^31X=4L3jCJQ;<^D2_;o0O01N1dNTDY1m;1101N1O1O01O001O01003N2N1O3M2M2O1O0O2O1N2O1N2O2N1N2O1N2O1O0O2O1oEhMl8Y2PGjMQ9X2jFjMW9Y2bFiM_9Z2XFlMi9j20000000[ORFmMo9n1XFoMi9m1]FPNf9h1bFUNa9b1hF[N[9]1mF`NW9Z1`1N2M3N3M2N2N2N3M2N2N3M5K5K01N10000O10000M3M3M4N1000O010O1O010O10PO\\OYEd0h:_OUE`0l:CQE=i0_Ob87aF:j0Ec84_F8k0Jc80_F7k0Lf80[F<f0Fo81XF>a0CX92SF9f0GW9`0RFXOf0:Y9<TFVOd08`9a0nETOb0;b9?oESOb0:b96gEB7Jc0;b96kE_OR18V98U21O001O0010O01O0mE\\O`7e0[H@e7?WGJWOLb9:VGg0i8XOWGj0h8VOWGk0i8UOVGk0k8UOTGk0m8UORGj0Q9UOnFj0T9UOlFk0X9ROgFm0_9oN`FP1e9mNoF>V9^O^2M\\hW3"}, "label": "a second horse on which a boy is sitting wearing black t - shirt and black cap"}]}
{"id": 322695, "image": "COCO_train2014_000000322695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second horse from the right\"."}], "task": "refering", "answer_template": "The \"second horse from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 240, 241, 243, 264, 265, 266], "bbox": [0.412578125, 0.499131455399061, 0.6087187500000001, 0.7911032863849765], "iscrowd": 0, "area": 5716.098599999998, "rle": {"size": [426, 640], "counts": "_Q^31X=4L3jCJQ;<^D2_;o0O01N1dNTDY1m;1101N1O1O01O001O01003N2N1O3M2M2O1O0O2O1N2O1N2O2N1N2O1N2O1O0O2O1oEhMl8Y2PGjMQ9X2jFjMW9Y2bFiM_9Z2XFlMi9j20000000[ORFmMo9n1XFoMi9m1]FPNf9h1bFUNa9b1hF[N[9]1mF`NW9Z1`1N2M3N3M2N2N2N3M2N2N3M5K5K01N10000O10000M3M3M4N1000O010O1O010O10PO\\OYEd0h:_OUE`0l:CQE=i0_Ob87aF:j0Ec84_F8k0Jc80_F7k0Lf80[F<f0Fo81XF>a0CX92SF9f0GW9`0RFXOf0:Y9<TFVOd08`9a0nETOb0;b9?oESOb0:b96gEB7Jc0;b96kE_OR18V98U21O001O0010O01O0mE\\O`7e0[H@e7?WGJWOLb9:VGg0i8XOWGj0h8VOWGk0i8UOVGk0k8UOTGk0m8UORGj0Q9UOnFj0T9UOlFk0X9ROgFm0_9oN`FP1e9mNoF>V9^O^2M\\hW3"}, "label": "second horse from the right"}]}
{"id": 322695, "image": "COCO_train2014_000000322695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse with the rider furthest to the right\"."}], "task": "refering", "answer_template": "The \"the horse with the rider furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293], "bbox": [0.5587500000000001, 0.5589671361502347, 0.7788437500000001, 0.8718309859154929], "iscrowd": 0, "area": 8362.707449999994, "rle": {"size": [426, 640], "counts": "PWe44T=4K5ZOFVD=h;DVD`0g;AWDb0f;AWDa0W;^OjDU1V;?001N1N3L5K3N2O1O1O1O1O1O1O1O1O1O001O100O1000O10000000001N2O1O1O001O2N2N2N2N2N2N3M5K3VFVNn7m1nGUNQ8Q2gGQNY8Y2[GiMe8^2RGdMm8]30O10000fNfF`N]9Y1VGVNP9e1]100ObFQNb7P2\\HRNd7o1YHSNf7n1ZGoM74^8S2SGnM:3c8e2oFgMR9X32N2N2E<J6ZOg0E;ROm0M3M2N3L3N2N1O000O10000O100B_C3a<KbC4^<JeC4\\<JeC7[<GgC9e<0\\N0nE0R:6hEIY:=aECc0JW8h0RG_Oe0LW8i0PG\\Of0MY8k0mFZOh0K\\8U1`FSOP1K`8j1[GYNe8l1SGWNn8Q3100VNZFJ0aNg9d1^FEi9;XFAk9>WFUOU:k0mElNZ:S1jEaN_:^1h0O0OREnNf9Q1VFXOf9Y2M3M3M1O01O001O1O1O1J6[Oe0I71O1hERNQ9o1kFRNW9[2VFjMn9g26J7bMUFn0U;N2\\OTDGo;AjD5VOOeTk1"}, "label": "the horse with the rider furthest to the right"}]}
{"id": 322695, "image": "COCO_train2014_000000322695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse furthest to the right in the picture\"."}], "task": "refering", "answer_template": "The \"the horse furthest to the right in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293], "bbox": [0.5587500000000001, 0.5589671361502347, 0.7788437500000001, 0.8718309859154929], "iscrowd": 0, "area": 8362.707449999994, "rle": {"size": [426, 640], "counts": "PWe44T=4K5ZOFVD=h;DVD`0g;AWDb0f;AWDa0W;^OjDU1V;?001N1N3L5K3N2O1O1O1O1O1O1O1O1O1O001O100O1000O10000000001N2O1O1O001O2N2N2N2N2N2N3M5K3VFVNn7m1nGUNQ8Q2gGQNY8Y2[GiMe8^2RGdMm8]30O10000fNfF`N]9Y1VGVNP9e1]100ObFQNb7P2\\HRNd7o1YHSNf7n1ZGoM74^8S2SGnM:3c8e2oFgMR9X32N2N2E<J6ZOg0E;ROm0M3M2N3L3N2N1O000O10000O100B_C3a<KbC4^<JeC4\\<JeC7[<GgC9e<0\\N0nE0R:6hEIY:=aECc0JW8h0RG_Oe0LW8i0PG\\Of0MY8k0mFZOh0K\\8U1`FSOP1K`8j1[GYNe8l1SGWNn8Q3100VNZFJ0aNg9d1^FEi9;XFAk9>WFUOU:k0mElNZ:S1jEaN_:^1h0O0OREnNf9Q1VFXOf9Y2M3M3M1O01O001O1O1O1J6[Oe0I71O1hERNQ9o1kFRNW9[2VFjMn9g26J7bMUFn0U;N2\\OTDGo;AjD5VOOeTk1"}, "label": "the horse furthest to the right in the picture"}]}
{"id": 543881, "image": "COCO_train2014_000000543881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skateboarding\"."}], "task": "refering", "answer_template": "The \"a man skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 30, 31, 32, 51, 52, 53, 54, 55, 56, 58, 59, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 216, 238, 239, 261, 262, 283, 284, 285, 286], "bbox": [0.191296875, 0.03303030303030303, 0.620609375, 0.85997668997669], "iscrowd": 0, "area": 33835.574900000014, "rle": {"size": [429, 640], "counts": "gWc11W=5J7J6J5J7L4N101N101N2O0O2O1N1O3XE_O`8c0WGEg8?oFIo8:hFNW95_F3_9o1N2O1N2N2O1N2N2O1N1O2O1N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O0O2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O001O1O1O1N20001N100O2N1O1OO2N]OXIcJg6\\5]IbJa6_5cI^J\\6b5hI[JW6e5lIYJS6g5QJVJn5i5VJUJi5k5[JRJc5o5m000_LgHo0\\7kNiHS1\\7fNhHW1^7aNgH]1^7ZNhHd1]7RNjHl1[7kMkHS2Z7cMlH]2X7ZMmHe2X7QMnHo2i8N100O101N100O100O2O0O100O2OUK\\MmNc2Q1bMmN^2P1fMoNZ2n0jMROU2l0nMSOR2j0RNUOn1h0VNPNYOhMa2U4ZNPNXOkM]2S4^NPNWOlM[2Q4bNPNVOnMX2o3fNPNTOQNV2k3kNQNROTNR2g3ROQNoNWNo1e3VORNmNXNm1b3[OSNkNZNi1`3ASNiN]Ne1\\3GUNfN^Nc1Y3LWNbN`Nb1U31YN_NaN`1S35ZN]NcN]1Q39YN]NeNZ1o2>YNZNgNY1m2b0XNWNjNX1k2f0WNSNnNX1l0cM1W31PNROe19ZM?V32lMVOi11YMd0W31iMYOl1IYMj0W30eM]OP2AXMn0X30cMAQ2YOXMS1Y3O_MEo29GN[MIo26KMXMLm250KTM1k226IPM5i21<FmL8g22?BkL<f23a0\\OkLa0d24c0WOkLd0b26e2J[M6e2J[M6f2IZM8e2IZM7f2IZM7g2HYM8g2HYM8g2HYM8h2GXM9h2GXM9i2A\\M`0h7000000000000001O0001O001O1O1O010O1O001O1O001O10O01O1O1O00O100O10O10O10O010000O010O100O10O10O100O10O010000O10O0100O10O001O001O001O1O0010O01O001O0010O100O100O100O1000O0100O100O100O11O1O1N2O1O1O1N2O1O1N2O1N2N2O3LRgU3"}, "label": "a man skateboarding"}]}
{"id": 543881, "image": "COCO_train2014_000000543881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green shirt riding a skateboard\"."}], "task": "refering", "answer_template": "The \"a man wearing a green shirt riding a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 30, 31, 32, 51, 52, 53, 54, 55, 56, 58, 59, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 216, 238, 239, 261, 262, 283, 284, 285, 286], "bbox": [0.191296875, 0.03303030303030303, 0.620609375, 0.85997668997669], "iscrowd": 0, "area": 33835.574900000014, "rle": {"size": [429, 640], "counts": "gWc11W=5J7J6J5J7L4N101N101N2O0O2O1N1O3XE_O`8c0WGEg8?oFIo8:hFNW95_F3_9o1N2O1N2N2O1N2N2O1N1O2O1N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O0O2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O001O1O1O1N20001N100O2N1O1OO2N]OXIcJg6\\5]IbJa6_5cI^J\\6b5hI[JW6e5lIYJS6g5QJVJn5i5VJUJi5k5[JRJc5o5m000_LgHo0\\7kNiHS1\\7fNhHW1^7aNgH]1^7ZNhHd1]7RNjHl1[7kMkHS2Z7cMlH]2X7ZMmHe2X7QMnHo2i8N100O101N100O100O2O0O100O2OUK\\MmNc2Q1bMmN^2P1fMoNZ2n0jMROU2l0nMSOR2j0RNUOn1h0VNPNYOhMa2U4ZNPNXOkM]2S4^NPNWOlM[2Q4bNPNVOnMX2o3fNPNTOQNV2k3kNQNROTNR2g3ROQNoNWNo1e3VORNmNXNm1b3[OSNkNZNi1`3ASNiN]Ne1\\3GUNfN^Nc1Y3LWNbN`Nb1U31YN_NaN`1S35ZN]NcN]1Q39YN]NeNZ1o2>YNZNgNY1m2b0XNWNjNX1k2f0WNSNnNX1l0cM1W31PNROe19ZM?V32lMVOi11YMd0W31iMYOl1IYMj0W30eM]OP2AXMn0X30cMAQ2YOXMS1Y3O_MEo29GN[MIo26KMXMLm250KTM1k226IPM5i21<FmL8g22?BkL<f23a0\\OkLa0d24c0WOkLd0b26e2J[M6e2J[M6f2IZM8e2IZM7f2IZM7g2HYM8g2HYM8g2HYM8h2GXM9h2GXM9i2A\\M`0h7000000000000001O0001O001O1O1O010O1O001O1O001O10O01O1O1O00O100O10O10O10O010000O010O100O10O10O100O10O010000O10O0100O10O001O001O001O1O0010O01O001O0010O100O100O100O1000O0100O100O100O11O1O1N2O1O1O1N2O1O1N2O1N2N2O3LRgU3"}, "label": "a man wearing a green shirt riding a skateboard"}]}
{"id": 543882, "image": "COCO_train2014_000000543882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table withe different dishess\"."}], "task": "refering", "answer_template": "The \"a table withe different dishess\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 54, 55, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.123703125, 0.1001152073732719, 0.9676562499999999, 0.9865207373271889], "iscrowd": 0, "area": 131242.9021999999, "rle": {"size": [434, 640], "counts": "WQR11`=100OY2hM2N1N2O2N1O1O2N1O1O1O2N1O1O2N1^O[MUFf2j9c0N1O1N3N2N1O2N2N2N1O2N2N2N2N1O2N2N1O2N2N2N1O2N2N1O1O100O0101OO1O1N101O1O1O001O3M5K4L5J2O1O2N1O1O101N1O1O1O2N1O1O1FoIQIS6k6<O1N2O2N1O1O1I8M2O1O1N3N1O1O1O2M2O1O1O2N1O1O1O2O0O1O101N1O1O1@VHVKl7g4VHXKk7g4VHXKk7f4VHZKl7d4UH[Kn7_4VH`Kl7^4UHaKm7\\4THdKn7Z4SHeKo7Y4g0O1M4M2M11O00L4O1O1C[FmLe9o2bFmL_9R3dFlL[9U3g0I6K6J1K400O1O2O01O001O0001N3L9H7J6J1O1N21O00000O1O001O0000O1N200001O000lD\\NZ:e1eE\\NZ:d1eE]N[:h1`EWNa:i1^E\\N^:e1`E[Na:e1^E\\Nb:e1[E\\Nf:e1WE]Ni:W2O00002N1ON2N2000000O1001O0SFaMUO1V9_2dGbMoN5\\9]2aGiM_8W2aGiM_8U2cGmM[8R2eGoM[8P2fGPNZ8P2fGPNZ8o1fGRNZ8o1eGQN[8o1eGQN[8o1dGRN\\8n1dGRN\\8o1cGQN]8o1aGRN`8R2[GoMe8U2VGlMj8T2TGnMl8R2SGoMm8Q2RGPNn8P2nFTNR9m1mFSNS9n1kFSNU9Q2eFQN[9S300000000000O1O1O10000O10000O100000000O1000000000000000000O1O1O1O100O100O100O100O1000000O1000000O1O1O1O1O1N21O000000001O002N2N2N00001O001O001O001O001O001O2N3M00001O00000000000000001O00001O0UGkKY8U4fGlKZ8U4eGkK[8U4dGlK\\8U4bGlK^8U4`GlK`8U4_GkKa8U4^GlKb8U4\\GmKc8T4\\GlKd8T4\\GlKd8a40000000001O00000000001O0000001O0000001O00001O00001O1O1O1O1O1O00001O000]HlJh6T5XIlJh6T5WImJi6S5WImJi6T5UInJj6R5UIoJj6R5VInJj6S5TInJl6R5TInJl6S5RInJn6R5RInJn6S5PInJP7R5oHoJQ7R5nHoJCG\\7[5oHoJEF\\7[5nHRKR7o4lHRKT7P5iHQKW7S5dHmJ]7d501O1O1O00O1001O0000O1000000002N0000FiHYJW7f5kHYJU7f5mHYJS7S5gHZK7BS7S5hHZK5CS7R5iHZK6CQ7R5lHYK3EINT7U5RIUK2GHOT7W5RIQK2IHOT7W5SIPK1JH0S7V5VInJOMGOT7V5WIlJ0NE0T7V5hIoJW6P5jIPKV6P5jIQKTOKk6T5RJVKn5j4RJVKn5j4RJUKo5k4RJoJS6Q5W100O1001O00000iHoJm5Q5SJQKk5S5QJoJm5T5PJoJm5U5oImJo5S5QJmJo5T5PJlJP6T5PJmJo5T5oImJQ6S5oInJP6R5PJnJP6S5oImJQ6S5oInJP6S5oImJQ6S5oInJP6S5oImJQ6S5nInJR6S5lIoJS6R5kIoJU6S5hIoJW6R5gIoJY6R5eIoJ[6R5cIPK\\6Q5bIPK^6S6O00001O001J500010O01O01O01O010O01O01O010O000O2N101N1O101N1O2N100I8O001O00001N101O0O2O00001N101O0O101O001N10001N101O000O2O001N10001N101O0M3I8I6J7I6I7J7I6J7H7J6J7H7J7I6J6I8I6J7I6I7J7I6O2N10001N101N10hd8"}, "label": "a table withe different dishess"}]}
{"id": 543882, "image": "COCO_train2014_000000543882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long table covered in food\"."}], "task": "refering", "answer_template": "The \"a long table covered in food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 54, 55, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.123703125, 0.1001152073732719, 0.9676562499999999, 0.9865207373271889], "iscrowd": 0, "area": 131242.9021999999, "rle": {"size": [434, 640], "counts": "WQR11`=100OY2hM2N1N2O2N1O1O2N1O1O1O2N1O1O2N1^O[MUFf2j9c0N1O1N3N2N1O2N2N2N1O2N2N2N2N1O2N2N1O2N2N2N1O2N2N1O1O100O0101OO1O1N101O1O1O001O3M5K4L5J2O1O2N1O1O101N1O1O1O2N1O1O1FoIQIS6k6<O1N2O2N1O1O1I8M2O1O1N3N1O1O1O2M2O1O1O2N1O1O1O2O0O1O101N1O1O1@VHVKl7g4VHXKk7g4VHXKk7f4VHZKl7d4UH[Kn7_4VH`Kl7^4UHaKm7\\4THdKn7Z4SHeKo7Y4g0O1M4M2M11O00L4O1O1C[FmLe9o2bFmL_9R3dFlL[9U3g0I6K6J1K400O1O2O01O001O0001N3L9H7J6J1O1N21O00000O1O001O0000O1N200001O000lD\\NZ:e1eE\\NZ:d1eE]N[:h1`EWNa:i1^E\\N^:e1`E[Na:e1^E\\Nb:e1[E\\Nf:e1WE]Ni:W2O00002N1ON2N2000000O1001O0SFaMUO1V9_2dGbMoN5\\9]2aGiM_8W2aGiM_8U2cGmM[8R2eGoM[8P2fGPNZ8P2fGPNZ8o1fGRNZ8o1eGQN[8o1eGQN[8o1dGRN\\8n1dGRN\\8o1cGQN]8o1aGRN`8R2[GoMe8U2VGlMj8T2TGnMl8R2SGoMm8Q2RGPNn8P2nFTNR9m1mFSNS9n1kFSNU9Q2eFQN[9S300000000000O1O1O10000O10000O100000000O1000000000000000000O1O1O1O100O100O100O100O1000000O1000000O1O1O1O1O1N21O000000001O002N2N2N00001O001O001O001O001O001O2N3M00001O00000000000000001O00001O0UGkKY8U4fGlKZ8U4eGkK[8U4dGlK\\8U4bGlK^8U4`GlK`8U4_GkKa8U4^GlKb8U4\\GmKc8T4\\GlKd8T4\\GlKd8a40000000001O00000000001O0000001O0000001O00001O00001O1O1O1O1O1O00001O000]HlJh6T5XIlJh6T5WImJi6S5WImJi6T5UInJj6R5UIoJj6R5VInJj6S5TInJl6R5TInJl6S5RInJn6R5RInJn6S5PInJP7R5oHoJQ7R5nHoJCG\\7[5oHoJEF\\7[5nHRKR7o4lHRKT7P5iHQKW7S5dHmJ]7d501O1O1O00O1001O0000O1000000002N0000FiHYJW7f5kHYJU7f5mHYJS7S5gHZK7BS7S5hHZK5CS7R5iHZK6CQ7R5lHYK3EINT7U5RIUK2GHOT7W5RIQK2IHOT7W5SIPK1JH0S7V5VInJOMGOT7V5WIlJ0NE0T7V5hIoJW6P5jIPKV6P5jIQKTOKk6T5RJVKn5j4RJVKn5j4RJUKo5k4RJoJS6Q5W100O1001O00000iHoJm5Q5SJQKk5S5QJoJm5T5PJoJm5U5oImJo5S5QJmJo5T5PJlJP6T5PJmJo5T5oImJQ6S5oInJP6R5PJnJP6S5oImJQ6S5oInJP6S5oImJQ6S5oInJP6S5oImJQ6S5nInJR6S5lIoJS6R5kIoJU6S5hIoJW6R5gIoJY6R5eIoJ[6R5cIPK\\6Q5bIPK^6S6O00001O001J500010O01O01O01O010O01O01O010O000O2N101N1O101N1O2N100I8O001O00001N101O0O2O00001N101O0O101O001N10001N101O000O2O001N10001N101O0M3I8I6J7I6I7J7I6J7H7J6J7H7J7I6J6I8I6J7I6I7J7I6O2N10001N101N10hd8"}, "label": "a long table covered in food"}]}
{"id": 543882, "image": "COCO_train2014_000000543882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl of food in table\"."}], "task": "refering", "answer_template": "The \"bowl of food in table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 228, 229, 252], "bbox": [0.6559062499999999, 0.31412442396313367, 1.0, 0.6447926267281107], "iscrowd": 0, "area": 14315.203449999995, "rle": {"size": [434, 640], "counts": "oTb51`=101N101N101N101O0O2O0O2O0O2O001N101O000O2O000O2O0lC^OX;c0aDC^;>_DEa;;^DFb;;\\DFd;:ZDHe;9ZDHf;9XDHh;P11000O100000O1O1O1O00100O1O1O1O1O1O1O10OBoC\\OR<c0QDZOQ<d0?0001O0O10001O0O10001N1000000O100000000000WDFZ::hDAb06e:9hDCb06d:7iDDc06d:5gDHd04d:4gDJc03g:2eDLd02g:2dDNd01h:0bD2e0Ni:a0WE@i:`0VE@k:`0TE@l:`0TEAl:?SEAm:`0REAn:>QECo:>PECP;<PEDP;=oDCo:`0PEAm:a0SE_Ok:d0TE]Ok:d0TE\\Ol:e0SE[Om:g0QEZOn:i0nDXOR;j0lDWOS;d10001O001O001O00001O001O00001O001O0O101O001O001O00001O001O00001O001O00001O001O001O00001N101O00001O001O001O00001O001O00001O0011N3M2O2M3NN101O00001O00000010O001O010O001O010O001O010O001OlE^MX9c2^FeMd9Z2[FgMe9Y2ZFhMg9W2WFjMj9W2TFjMm9U2RFkMo9l21O0O2O001N101N101O0O2O0O2O1O3L3K5^M[ER1h;hNYD;S7"}, "label": "bowl of food in table"}]}
{"id": 421005, "image": "COCO_train2014_000000421005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in jeans , a goonies shirt , and brown shoes catching a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in jeans , a goonies shirt , and brown shoes catching a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 142, 143, 144, 157, 158, 159, 160, 172, 173, 174, 175, 187, 188, 189, 202, 203, 204, 217, 218, 219, 232, 233, 234, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 292, 293], "bbox": [0.4480796252927401, 0.38106249999999997, 0.6977283372365339, 0.845140625], "iscrowd": 0, "area": 15408.952650000003, "rle": {"size": [640, 427], "counts": "Xlg31gc09G8J6O2N1N2O2N1J6H9SJgNkH_1P7jNgH[1W7jNbH[1[7kN]HY1b7kNWHY1g7lNSHX1j7mNQHg0lK]NR<n0PHc0_8AXEhN4g1e:DnDZONP1V;IdD3_O2o;M[Dg2e;]MTDe2n;e21N[DPIU:3jDk6P1WIQ:U7mEQIl9e0_El4e0dJj9>cEl4b0jJj98gEj4`0PKh95kEh4>UKf92mEf4>ZKe9NPFd4=`Kb9KSFb4;eKb9HTF`4<jK^9EYF]4;oK\\9CZF[49VL]9^O\\FX45_La9jN[E1Q1c41gLm9dNUFb4MmLn9_NWFb4ISMP:[NXF_4GYMQ:WNYF^4D_Ma;_2^DdMV;oMSD^4f0fMP;TNVDW4h0iMi:R6N2N2O1O1N2O1O010000]OiD]HW;_7mDaHS;\\7PEcHQ;Y7SEfHn:V7VEiHk:S7YElHh:P7]EnHd:o6`EoHa:m6dEQI_:i6fETI]:g6a1L4XJkB`4Z=[KjB`4[=^KfB_4a=\\KaB`4a=RKTB8<b4d=WKSB2:c4g=\\KeBX4f=hKZBm2XOiMj>ZOnAl2BVMk>0aAW1FE_?TOj@T1?ROg>Li@o0f`0[OY_Oc0e`0BZ_O;i`0FV_O7l`0JT_O5n`0LP_O3Ra0Ok^O0Wa02g^OL\\a06`^OJaa0;Y^ODia0Y11O=B3N2M3N2N2M3N2M3M3H8M3M4K6K5KVZ`2"}, "label": "a man in jeans , a goonies shirt , and brown shoes catching a frisbee"}]}
{"id": 421005, "image": "COCO_train2014_000000421005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man getting ready to catch a frisbee\"."}], "task": "refering", "answer_template": "The \"a man getting ready to catch a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 142, 143, 144, 157, 158, 159, 160, 172, 173, 174, 175, 187, 188, 189, 202, 203, 204, 217, 218, 219, 232, 233, 234, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 292, 293], "bbox": [0.4480796252927401, 0.38106249999999997, 0.6977283372365339, 0.845140625], "iscrowd": 0, "area": 15408.952650000003, "rle": {"size": [640, 427], "counts": "Xlg31gc09G8J6O2N1N2O2N1J6H9SJgNkH_1P7jNgH[1W7jNbH[1[7kN]HY1b7kNWHY1g7lNSHX1j7mNQHg0lK]NR<n0PHc0_8AXEhN4g1e:DnDZONP1V;IdD3_O2o;M[Dg2e;]MTDe2n;e21N[DPIU:3jDk6P1WIQ:U7mEQIl9e0_El4e0dJj9>cEl4b0jJj98gEj4`0PKh95kEh4>UKf92mEf4>ZKe9NPFd4=`Kb9KSFb4;eKb9HTF`4<jK^9EYF]4;oK\\9CZF[49VL]9^O\\FX45_La9jN[E1Q1c41gLm9dNUFb4MmLn9_NWFb4ISMP:[NXF_4GYMQ:WNYF^4D_Ma;_2^DdMV;oMSD^4f0fMP;TNVDW4h0iMi:R6N2N2O1O1N2O1O010000]OiD]HW;_7mDaHS;\\7PEcHQ;Y7SEfHn:V7VEiHk:S7YElHh:P7]EnHd:o6`EoHa:m6dEQI_:i6fETI]:g6a1L4XJkB`4Z=[KjB`4[=^KfB_4a=\\KaB`4a=RKTB8<b4d=WKSB2:c4g=\\KeBX4f=hKZBm2XOiMj>ZOnAl2BVMk>0aAW1FE_?TOj@T1?ROg>Li@o0f`0[OY_Oc0e`0BZ_O;i`0FV_O7l`0JT_O5n`0LP_O3Ra0Ok^O0Wa02g^OL\\a06`^OJaa0;Y^ODia0Y11O=B3N2M3N2N2M3N2M3M3H8M3M4K6K5KVZ`2"}, "label": "a man getting ready to catch a frisbee"}]}
{"id": 101522, "image": "COCO_train2014_000000101522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoli is mixed into a dish that contains carrots and rice\"."}], "task": "refering", "answer_template": "The \"a piece of broccoli is mixed into a dish that contains carrots and rice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 28, 29, 30, 31, 50, 51, 52, 53, 54, 74, 75, 76, 77, 98, 99], "bbox": [0.201328125, 0.0014035087719298245, 0.40642187500000004, 0.250233918128655], "iscrowd": 0, "area": 8708.43, "rle": {"size": [513, 640], "counts": "VfP23l?5L4K4L5K4L5L4K4L5K3N2M2N3M3M3N2M3N2O1N2O001O1N2O1O1N2O1O001N2O001O0O2O001O0O2O001O001N101O001O001N101OJoB\\MQ=d2PC\\Mo<d2SCZMm<f2TCZMj<g2WCXMi<h2YCWMf<i2900001N1O100O1O1O1O10O01O1O0010O01O010O10O010O010O0oMWCl0i<iNcCU1]<kNdCU1\\<kNdCU1\\<jNeCU1[<lNfCS1Z<lNgCS1Y<nNgCR1Y<mNhCS1W<nNjCQ1V<oNjCQ1U<oNlCQ1T<oNlCQ1S<oNoCo0R<QOnCo0R<POoCP1P<QOPD4XONh<NPD2]OMc<0RD1_OL_<3RDOCL[<4SDOFJV<8TDMJGS<;UDKLHo;=UDJ0Fk;`0UDI4Dg;b0b210O100O010000000O0100000000O01000004L6I6K\\[n5"}, "label": "a piece of broccoli is mixed into a dish that contains carrots and rice"}]}
{"id": 101522, "image": "COCO_train2014_000000101522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of broccoli that is right in the center\"."}], "task": "refering", "answer_template": "The \"the piece of broccoli that is right in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 171, 172], "bbox": [0.28521874999999997, 0.22413255360623782, 0.516890625, 0.4436062378167641], "iscrowd": 0, "area": 8197.840999999995, "rle": {"size": [513, 640], "counts": "eik26j?5K4M3L4L5K7J4L4L4IcNbA`1[>`NdAd1Y>5M3M00000O2O00001O000O2O0000001O0O101O0010O010O001O010O1O101N1O2O0O2N101N2N101O1O1O2M2NO100O100O100O10000O100O100O010O100O010O100O0O2N1N3O0100O100O100O100O2O1O1O1O1O1O1O1O1O010O001O000010O100O1O100O010O1O100O2O1N2O1N2O1N3N3L3N1O1N2O001N02N1O1O1N3N1O1M5K5D<L4L3M3L3N3M2J7ITdj4"}, "label": "the piece of broccoli that is right in the center"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch which is behind the woman who is wearing black dress\"."}], "task": "refering", "answer_template": "The \"a couch which is behind the woman who is wearing black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 216, 217, 218, 219, 220], "bbox": [0.00186, 0.46781333333333336, 0.23020000000000002, 0.9876266666666667], "iscrowd": 0, "area": 18199.501800000002, "rle": {"size": [375, 500], "counts": "db0Z2]9^1aN4M3M3M3L3N4L4K5L3M000O1000000O10000O10000O10000O1000000000000L4M3L4L4M3L4M3C=00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000SJ`Jc5`5[JmJZ5S5cJ[KR5k5J3M3M3M3M3M3M4L3M3M3M3M3M3L4Mc0]Oi0WOi0WOg0YO5K4L5K4LYd\\4"}, "label": "a couch which is behind the woman who is wearing black dress"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey ottoman sitting in a matching chair\"."}], "task": "refering", "answer_template": "The \"a grey ottoman sitting in a matching chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 216, 217, 218, 219, 220], "bbox": [0.00186, 0.46781333333333336, 0.23020000000000002, 0.9876266666666667], "iscrowd": 0, "area": 18199.501800000002, "rle": {"size": [375, 500], "counts": "db0Z2]9^1aN4M3M3M3L3N4L4K5L3M000O1000000O10000O10000O10000O1000000000000L4M3L4L4M3L4M3C=00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000SJ`Jc5`5[JmJZ5S5cJ[KR5k5J3M3M3M3M3M3M4L3M3M3M3M3M3L4Mc0]Oi0WOi0WOg0YO5K4L5K4LYd\\4"}, "label": "a grey ottoman sitting in a matching chair"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue leather chair behind lady in red\"."}], "task": "refering", "answer_template": "The \"blue leather chair behind lady in red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 158, 159, 160, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.6648200000000001, 0.5333333333333333, 0.96296, 1.0], "iscrowd": 0, "area": 13444.583, "rle": {"size": [375, 500], "counts": "Tni31[;;E<D;E;E;E;L4N2N2N2O1N2N3N1N2N2N2O1N2O1O1O1O100O2N1O1O1O1O1O1O1:F>B>B>C>A=C4L4L4L5K4L4KjX71SgH=D<D<C=D<_O:D9G9lNT1I7J6N2N2N1O2N2N2N2N2N2N2N2O1N2N2N2O1N21O3L5L3M4L3M4L3M3M1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O00001O001O001O001N101O00001O001O001O001O0_MnG]1S8_NRH_1n7_NVH_1k7]NZHa1f7\\N_Ha1c7[NbHc1_7XNfH:g8B_G3k8H[GM_o6"}, "label": "blue leather chair behind lady in red"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black color couch in the back side of the lady with red top\"."}], "task": "refering", "answer_template": "The \"black color couch in the back side of the lady with red top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 137, 140, 141, 154, 155, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.6039, 0.4597333333333333, 0.97012, 0.9896], "iscrowd": 0, "area": 13627.866150000005, "rle": {"size": [375, 500], "counts": "nj^39];3M3\\Nk1G<D1OL4K6K6I9H7H9H5J6K5J5L5E\\c10f\\N:G9L2N3N1N2N3N1N2N2O2M2N2N2N2N3M2M3N2N2M3N2M3N2N2N2M3N2N3M2M3N2N2N21O1O1O1O3M;E;E:C=\\Om0WO`Z;_1SdD4L4K5VMhMQL\\2k3fMRL_2j3cMTLa2W3]MfJ5Q2b2X3nMdLW2[3hMXKZO]O8=j2m4cMZK\\OXOU4^5]L[K_OSOX4c5VL[KP4g4mKXKU4n4cKSK^4R61O1O1N101O1O1O1N2O1O1O1O1N2O1O1O100?A>B>B00001O00001O0O10001O00001O00001O00001O00001O00001O00001O0000001N10001O00001bMkGX1U8dNPH[1P8bNTHW1S8fNPHT1U8kNmGn0Y8POjGi0]8VOcG4ADQ97`GMY91[1K5L4K\\W5"}, "label": "black color couch in the back side of the lady with red top"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair behind the woman with the red top\"."}], "task": "refering", "answer_template": "The \"the chair behind the woman with the red top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 137, 140, 141, 154, 155, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.6039, 0.4597333333333333, 0.97012, 0.9896], "iscrowd": 0, "area": 13627.866150000005, "rle": {"size": [375, 500], "counts": "nj^39];3M3\\Nk1G<D1OL4K6K6I9H7H9H5J6K5J5L5E\\c10f\\N:G9L2N3N1N2N3N1N2N2O2M2N2N2N2N3M2M3N2N2M3N2M3N2N2N2M3N2N3M2M3N2N2N21O1O1O1O3M;E;E:C=\\Om0WO`Z;_1SdD4L4K5VMhMQL\\2k3fMRL_2j3cMTLa2W3]MfJ5Q2b2X3nMdLW2[3hMXKZO]O8=j2m4cMZK\\OXOU4^5]L[K_OSOX4c5VL[KP4g4mKXKU4n4cKSK^4R61O1O1N101O1O1O1N2O1O1O1O1N2O1O1O100?A>B>B00001O00001O0O10001O00001O00001O00001O00001O00001O00001O0000001N10001O00001bMkGX1U8dNPH[1P8bNTHW1S8fNPHT1U8kNmGn0Y8POjGi0]8VOcG4ADQ97`GMY91[1K5L4K\\W5"}, "label": "the chair behind the woman with the red top"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing at suitcase along with two other girls\"."}], "task": "refering", "answer_template": "The \"a man standing at suitcase along with two other girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 72, 73, 90, 91, 108, 109, 126, 127, 198, 216], "bbox": [0.0, 0.2809333333333333, 0.07964, 1.0], "iscrowd": 0, "area": 2867.4291499999995, "rle": {"size": [375, 500], "counts": "Z3n0\\64PM>ZNaNo1^1H1WNbN[2T1]O:WNcNc2l0WOa0UNcNk2e0QOg0TNdNn2b0oNi0RNfNR3>mNl0oMgNW3:kNP1kMgNjNK\\4;POS1iMgNjNLc42mN[1dMhNkNL^8\\1eHiNlNL_8]1bHhNoNL_8]1`HgNPOMa8_1ZHeNTOMc8`1UHkNm7X1nGiNS8Z1hGgNY8Z1dGgN]8[1^GgNc8[1WGgNj8S21N2O1N2O1N2O1N3N3L4L4L4K5L4K5L4K5L4LbYX5"}, "label": "a man standing at suitcase along with two other girls"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of the person in the gray sweater on the far left that is mostly cut out of the picture\"."}], "task": "refering", "answer_template": "The \"the arm of the person in the gray sweater on the far left that is mostly cut out of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 72, 73, 90, 91, 108, 109, 126, 127, 198, 216], "bbox": [0.0, 0.2809333333333333, 0.07964, 1.0], "iscrowd": 0, "area": 2867.4291499999995, "rle": {"size": [375, 500], "counts": "Z3n0\\64PM>ZNaNo1^1H1WNbN[2T1]O:WNcNc2l0WOa0UNcNk2e0QOg0TNdNn2b0oNi0RNfNR3>mNl0oMgNW3:kNP1kMgNjNK\\4;POS1iMgNjNLc42mN[1dMhNkNL^8\\1eHiNlNL_8]1bHhNoNL_8]1`HgNPOMa8_1ZHeNTOMc8`1UHkNm7X1nGiNS8Z1hGgNY8Z1dGgN]8[1^GgNc8[1WGgNj8S21N2O1N2O1N2O1N3N3L4L4L4K5L4K5L4K5L4LbYX5"}, "label": "the arm of the person in the gray sweater on the far left that is mostly cut out of the picture"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two women playing wii while smiling and laughing\"."}], "task": "refering", "answer_template": "The \"two women playing wii while smiling and laughing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 30, 31, 40, 41, 42, 48, 49, 58, 59, 60, 61, 66, 67, 68, 75, 76, 77, 78, 79, 83, 84, 85, 86, 87, 93, 94, 95, 96, 97, 101, 102, 103, 104, 105, 111, 112, 113, 114, 115, 119, 120, 121, 122, 123, 129, 130, 131, 132, 133, 137, 138, 139, 140, 141, 147, 148, 149, 150, 151, 155, 156, 157, 158, 165, 166, 167, 168, 169, 173, 174, 175, 176, 183, 184, 185, 186, 187, 190, 191, 192, 193, 194, 202, 203, 204, 205, 208, 209, 211, 212, 220, 221, 222, 223, 226, 227, 229, 230], "bbox": [0.1736, 0.10786666666666668, 0.88146, 1.0], "iscrowd": 0, "area": 58501.18125000001, "rle": {"size": [375, 500], "counts": "[QP18Y;9F;F9F;E:G:E:K6J5K6J5K6J6I6K6J5K6J>B>C=C=D=C>A>C>B=C4K10000O1000000O10000O101iNnGgNR8X1W1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O11O001O001O1O001O001O001O001O1O001O1O2N1[GWNZ7j1cHZN[7h1aHZN_7g1_HZNa7h1[HZNe7g1XH\\Ng7f1UH\\Nk7e1RH]Nn7e1nG]NR8d1kG_NT8c1hG_NX8b1eG`N[8^201O000000001O0000001O0XJhLQ3X3lLnLQ3R3lLTMQ3m2jLZMS3f2jL`MS3`2jLfMS3Z2jLlMS3V2gLQNV3T2bLRN[3T2\\LRNa3S2WLSN;WNg1k3fMTN9aNh1a3eMUN:jNg1V3gMVN9TOf1d5PNlJg1l7F9G:F9UOenl06_QSO9G8H8H8H8K5K5L4K5K5^LnM[MW2^2PN\\MU2^2RN\\MS2_2SN\\MQ2^2VN\\Mo1^2XN\\Mm1_2YN[Ml1_2ZN]Mj1]2]N]Mh1^2^N\\Mg1^2`N]Me1\\2bN^Mc1_2aN[Md1e2]NVMg1i2[NQMj1o2WNkLk1W3WNdLj1_3WN[Lk1g3VNTLl1o3k22M4M2N3M2N2N3M2N3M2M4MaN[IQNc6R2_IlM^6V2eIgMY6[2jIcMS6_2PJ^Mn5d2VJXMh5j2[JlLi5W3ZJ_Lk5c3XJRLn5Q4i02IlKjHV4W7400000001O000000000000001O1O1O18I7H8H9H7Hn1RN01O000010O0001O00001O001O2N2N2N2N2N3M2N2bFWNj8k1RG^Ng8d1UGeNd8]2J7I2N0000O100000lMT2L4M3L3M4oMQJWNR6h1YJjMk5V2_J\\Me5c2fJPM]5o2oJbLT5^3c1O1O1O1O2N2N3M2N3M2N3M2N2N3M2N3M2O2M2N3M2N3M2N3M2N3M2H8G:E:D=CT_e0"}, "label": "two women playing wii while smiling and laughing"}]}
{"id": 101530, "image": "COCO_train2014_000000101530.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing all black and another wearing a red shirt and black jeans\"."}], "task": "refering", "answer_template": "The \"a woman wearing all black and another wearing a red shirt and black jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 30, 31, 40, 41, 42, 48, 49, 58, 59, 60, 61, 66, 67, 68, 75, 76, 77, 78, 79, 83, 84, 85, 86, 87, 93, 94, 95, 96, 97, 101, 102, 103, 104, 105, 111, 112, 113, 114, 115, 119, 120, 121, 122, 123, 129, 130, 131, 132, 133, 137, 138, 139, 140, 141, 147, 148, 149, 150, 151, 155, 156, 157, 158, 165, 166, 167, 168, 169, 173, 174, 175, 176, 183, 184, 185, 186, 187, 190, 191, 192, 193, 194, 202, 203, 204, 205, 208, 209, 211, 212, 220, 221, 222, 223, 226, 227, 229, 230], "bbox": [0.1736, 0.10786666666666668, 0.88146, 1.0], "iscrowd": 0, "area": 58501.18125000001, "rle": {"size": [375, 500], "counts": "[QP18Y;9F;F9F;E:G:E:K6J5K6J5K6J6I6K6J5K6J>B>C=C=D=C>A>C>B=C4K10000O1000000O10000O101iNnGgNR8X1W1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O11O001O001O1O001O001O001O001O1O001O1O2N1[GWNZ7j1cHZN[7h1aHZN_7g1_HZNa7h1[HZNe7g1XH\\Ng7f1UH\\Nk7e1RH]Nn7e1nG]NR8d1kG_NT8c1hG_NX8b1eG`N[8^201O000000001O0000001O0XJhLQ3X3lLnLQ3R3lLTMQ3m2jLZMS3f2jL`MS3`2jLfMS3Z2jLlMS3V2gLQNV3T2bLRN[3T2\\LRNa3S2WLSN;WNg1k3fMTN9aNh1a3eMUN:jNg1V3gMVN9TOf1d5PNlJg1l7F9G:F9UOenl06_QSO9G8H8H8H8K5K5L4K5K5^LnM[MW2^2PN\\MU2^2RN\\MS2_2SN\\MQ2^2VN\\Mo1^2XN\\Mm1_2YN[Ml1_2ZN]Mj1]2]N]Mh1^2^N\\Mg1^2`N]Me1\\2bN^Mc1_2aN[Md1e2]NVMg1i2[NQMj1o2WNkLk1W3WNdLj1_3WN[Lk1g3VNTLl1o3k22M4M2N3M2N2N3M2N3M2M4MaN[IQNc6R2_IlM^6V2eIgMY6[2jIcMS6_2PJ^Mn5d2VJXMh5j2[JlLi5W3ZJ_Lk5c3XJRLn5Q4i02IlKjHV4W7400000001O000000000000001O1O1O18I7H8H9H7Hn1RN01O000010O0001O00001O001O2N2N2N2N2N3M2N2bFWNj8k1RG^Ng8d1UGeNd8]2J7I2N0000O100000lMT2L4M3L3M4oMQJWNR6h1YJjMk5V2_J\\Me5c2fJPM]5o2oJbLT5^3c1O1O1O1O2N2N3M2N3M2N3M2N2N3M2N3M2O2M2N3M2N3M2N3M2N3M2H8G:E:D=CT_e0"}, "label": "a woman wearing all black and another wearing a red shirt and black jeans"}]}
{"id": 35995, "image": "COCO_train2014_000000035995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck in which \" afg \" is written on its front\"."}], "task": "refering", "answer_template": "The \"the truck in which \" afg \" is written on its front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 221, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289], "bbox": [0.0416875, 0.0, 0.757859375, 0.8654823529411765], "iscrowd": 0, "area": 68858.85865, "rle": {"size": [425, 640], "counts": "iV;1X=1O2N1O1O2N1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O1OO100001O1O001O001O1O001O0O2O001O1O001O001O1O001O001O001O1O001O001O001O001O001O00001O001O001O001O001O001O001N101O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O6J1O2N1N3N1O2N1O1O1O1O1O1OZOf0M3O1O1N2O1O1O1O1N2O100O100O10000O10000O100O10000O100000000O10O1000000000O1000000000000O100000000000000O1001O001O001O001O001O001O001O001O001O0O2O0VImL^3T3`LRM[3n2dLWMX3j2fL\\MU3d2jL_MT3b2jL`MU3`2iLcMV3^2hLdMW3\\2hLfMW3[2gLgMX3Z2fLhMY3Y2eLjMY3V2fLmMX3T2eLPNY3Q2eLSNX3n1fLUNX3k1gL]NR3d1lLgNj2Z1TMQOb2P1\\MVO_2k0_MVOa2l0[MVOe2o0UMROk2R1PMoNP3V1nKGS4<`KM`4^40000000000000O2O00000000000O100O2O0O100O100O100OWMeGf0[8YOiGe0U8[OoGb0Q8^ORH?m7AXH;g7E]H8c7H`H5_7KeH2[7NhHOW71lHMT72lH0R71nHOR70oH0P71oH0Q7OPI1o6ORI1n6OQI2n6NSI2m6MTI3k6NUI3j6LVI5i6LWI4i6KXI5g6KYI6f6KZI5f6J[I6d6J\\I7d6I\\I7c6I^I8a6H^I9a6G`I9`6FaI:^6GbI9^6FbI;]6EdI;\\6EdI;[6EeI<[6DeI=Y6ChI=X6BhI?W6BiI>W6AjI?U6AlI?T6AkI`0T6@mI`0S6@mI`0S6_OmIc0Q6]OPJc0P6]OPJc0o5]OQJd0o5\\OQJd0o5[ORJe0m5[OSJf0m5ZOSJf0m5YOTJg0k5YOVJh0i5XOVJi0i5WOXJi0h5WOXJi0h5VOXJk0g5UOZJk0f5UOZJk0e5UO[Jl0e5SO\\Jm0d5SO\\Jn0b5RO^Jo0b5QO^Jo0b5PO_JP1`5POaJP1_5PO`JQ1_5oNbJQ1^5nNcJR1]5nNbJS1]5mNdJT1[5lNeJT1Z5lNfJU1Z5jNgJV1Y5jNgJV1X5jNhJW1X5iNhJW1X5hNiJX1V5hNkJX1U5hNjJY1U5gNlJZ1S5eNnJ[1R5eNmJ\\1R5dNoJ\\1Q5dNoJ\\1Q5cNoJ^1P5bNQK^1o4bNQK^1n4bNSK^1l4bNTK`1j4aNVK_1i4aNXK_1h4aNWK`1h4`NYK`1f4`N[K`1d4aN[K`1d4`N]K`1b4`N_K`1`4aN_Ka1_4_NbKa1^4_N\\Kg1c4YNYKl1f4UNUKP2i4SNPKS2o4PNjJU2T5nMfJW2Y5lM_JZ2`5hMZJ^2c5fMVJ_2i5cMPJc2n5]2O1N2O1O1N2O2M2O1L4L4L4L4M3L4L5K4L4M3L4L4L4L4M3L5K4O100O100O100O101N1O100O100O100O101N100O100O3N3L4M3L3N3L4M3L3N3L4M3L3N3L4M3L3N3L4M3L4M2M4M3L4M2M4M3L4M2M]WP2"}, "label": "the truck in which \" afg \" is written on its front"}]}
{"id": 35995, "image": "COCO_train2014_000000035995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"van cabin which is having full than other and letters are printed with afc\"."}], "task": "refering", "answer_template": "The \"van cabin which is having full than other and letters are printed with afc\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 221, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289], "bbox": [0.0416875, 0.0, 0.757859375, 0.8654823529411765], "iscrowd": 0, "area": 68858.85865, "rle": {"size": [425, 640], "counts": "iV;1X=1O2N1O1O2N1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O1OO100001O1O001O001O1O001O0O2O001O1O001O001O1O001O001O001O1O001O001O001O001O001O00001O001O001O001O001O001O001N101O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O6J1O2N1N3N1O2N1O1O1O1O1O1OZOf0M3O1O1N2O1O1O1O1N2O100O100O10000O10000O100O10000O100000000O10O1000000000O1000000000000O100000000000000O1001O001O001O001O001O001O001O001O001O0O2O0VImL^3T3`LRM[3n2dLWMX3j2fL\\MU3d2jL_MT3b2jL`MU3`2iLcMV3^2hLdMW3\\2hLfMW3[2gLgMX3Z2fLhMY3Y2eLjMY3V2fLmMX3T2eLPNY3Q2eLSNX3n1fLUNX3k1gL]NR3d1lLgNj2Z1TMQOb2P1\\MVO_2k0_MVOa2l0[MVOe2o0UMROk2R1PMoNP3V1nKGS4<`KM`4^40000000000000O2O00000000000O100O2O0O100O100O100OWMeGf0[8YOiGe0U8[OoGb0Q8^ORH?m7AXH;g7E]H8c7H`H5_7KeH2[7NhHOW71lHMT72lH0R71nHOR70oH0P71oH0Q7OPI1o6ORI1n6OQI2n6NSI2m6MTI3k6NUI3j6LVI5i6LWI4i6KXI5g6KYI6f6KZI5f6J[I6d6J\\I7d6I\\I7c6I^I8a6H^I9a6G`I9`6FaI:^6GbI9^6FbI;]6EdI;\\6EdI;[6EeI<[6DeI=Y6ChI=X6BhI?W6BiI>W6AjI?U6AlI?T6AkI`0T6@mI`0S6@mI`0S6_OmIc0Q6]OPJc0P6]OPJc0o5]OQJd0o5\\OQJd0o5[ORJe0m5[OSJf0m5ZOSJf0m5YOTJg0k5YOVJh0i5XOVJi0i5WOXJi0h5WOXJi0h5VOXJk0g5UOZJk0f5UOZJk0e5UO[Jl0e5SO\\Jm0d5SO\\Jn0b5RO^Jo0b5QO^Jo0b5PO_JP1`5POaJP1_5PO`JQ1_5oNbJQ1^5nNcJR1]5nNbJS1]5mNdJT1[5lNeJT1Z5lNfJU1Z5jNgJV1Y5jNgJV1X5jNhJW1X5iNhJW1X5hNiJX1V5hNkJX1U5hNjJY1U5gNlJZ1S5eNnJ[1R5eNmJ\\1R5dNoJ\\1Q5dNoJ\\1Q5cNoJ^1P5bNQK^1o4bNQK^1n4bNSK^1l4bNTK`1j4aNVK_1i4aNXK_1h4aNWK`1h4`NYK`1f4`N[K`1d4aN[K`1d4`N]K`1b4`N_K`1`4aN_Ka1_4_NbKa1^4_N\\Kg1c4YNYKl1f4UNUKP2i4SNPKS2o4PNjJU2T5nMfJW2Y5lM_JZ2`5hMZJ^2c5fMVJ_2i5cMPJc2n5]2O1N2O1O1N2O2M2O1L4L4L4L4M3L4L5K4L4M3L4L4L4L4M3L5K4O100O100O100O101N1O100O100O100O101N100O100O3N3L4M3L3N3L4M3L3N3L4M3L3N3L4M3L3N3L4M3L4M2M4M3L4M2M4M3L4M2M]WP2"}, "label": "van cabin which is having full than other and letters are printed with afc"}]}
{"id": 35995, "image": "COCO_train2014_000000035995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front corner of a white bus with man in front of it\"."}], "task": "refering", "answer_template": "The \"front corner of a white bus with man in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0015, 0.0033882352941176468, 0.20340625, 0.971835294117647], "iscrowd": 0, "area": 41416.956, "rle": {"size": [425, 640], "counts": "jc0\\6`0]600000000O100000000O1000000000000001O1O1O001O1O1O001O1O1O1O001O1O3L4M2N3M3M3M3M2N3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M2N2N001O00ROnFfMQ9Z2TGaMl8_2YG]Mf8c2_GXMa8h2dGTM[8l2iGPMW8P3Q10000O1000000000000001O001O1O1O001O1O1O002N4K5L4L4XIYLf3k3kKdLQ4a3iKcLT4d3dK`LY4g3`K\\L]4k3[KYLb4k3YKZLc4g3\\K\\La4e3]K_L`4b3^KbL_4_3`KeL\\4\\3bKhL[4Y3cKkLZ4V3eKmLX4T3fKQMV4Q3gKSMc3QNlKm4`0UMW3[NVLh7\\3[1C>A>C=B>\\Oe0ZOX]c6"}, "label": "front corner of a white bus with man in front of it"}]}
{"id": 35995, "image": "COCO_train2014_000000035995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"leftmost truck\"."}], "task": "refering", "answer_template": "The \"leftmost truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0015, 0.0033882352941176468, 0.20340625, 0.971835294117647], "iscrowd": 0, "area": 41416.956, "rle": {"size": [425, 640], "counts": "jc0\\6`0]600000000O100000000O1000000000000001O1O1O001O1O1O001O1O1O1O001O1O3L4M2N3M3M3M3M2N3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M2N2N001O00ROnFfMQ9Z2TGaMl8_2YG]Mf8c2_GXMa8h2dGTM[8l2iGPMW8P3Q10000O1000000000000001O001O1O1O001O1O1O002N4K5L4L4XIYLf3k3kKdLQ4a3iKcLT4d3dK`LY4g3`K\\L]4k3[KYLb4k3YKZLc4g3\\K\\La4e3]K_L`4b3^KbL_4_3`KeL\\4\\3bKhL[4Y3cKkLZ4V3eKmLX4T3fKQMV4Q3gKSMc3QNlKm4`0UMW3[NVLh7\\3[1C>A>C=B>\\Oe0ZOX]c6"}, "label": "leftmost truck"}]}
{"id": 322720, "image": "COCO_train2014_000000322720.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chocolate cake decorated with toy bulldozers\"."}], "task": "refering", "answer_template": "The \"chocolate cake decorated with toy bulldozers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 190, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.5807259953161592, 0.511609375, 0.9865573770491803], "iscrowd": 0, "area": 43556.145249999994, "rle": {"size": [427, 640], "counts": "n8b0i<S1lNU1lNS1mN;D100O10000O10000O1O1O1O100O1O1O1O1O1O1O1O1K5E;O10000O10000O10000O10000O10000000000000000000000O100000000000mJWHd4i7XK[Hh4e7UK_Hj4Q801O001O001O0000001O0000001O002N2N2N2NO100O10000O1000O010000O100O100O10000O100O10000O100000000000000O10000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000O1K5J60000000000O1000000000000000000000000000000001O000000001O002N2N3M?A2N3M0000000000001N100000000000001O000000000O10O000O2O000010O0001O01O01O00010O0100O1O010O1O100O1O10O01O100O100O00100O1O10O01O2O0O2M2E;G:L3M4L4L4N3L3N2N1O001O1O1000O10000OUNSEZ1m:fNTEY1l:gNTEY1k:hNVEW1j:hNWEX1i:hNXEW1g:jNZEU1f:kNZEU1f:kN[ET1d:mN]ER1c:nN]ER1c:nN^ER1`:nNaER1_:nNaER1_:nNaER1^:oNbER1_:lNaET1a:jN_EV1c:hN^EW1d:fN]EZ1e:dN[E\\1m:\\NTEc1Z;01O1O1O1O1O1O1O001N2O1N8I7H3N1N101N2O0O5K4M4KRVR4"}, "label": "chocolate cake decorated with toy bulldozers"}]}
{"id": 322720, "image": "COCO_train2014_000000322720.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheet cake including its decorations\"."}], "task": "refering", "answer_template": "The \"the sheet cake including its decorations\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 190, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.5807259953161592, 0.511609375, 0.9865573770491803], "iscrowd": 0, "area": 43556.145249999994, "rle": {"size": [427, 640], "counts": "n8b0i<S1lNU1lNS1mN;D100O10000O10000O1O1O1O100O1O1O1O1O1O1O1O1K5E;O10000O10000O10000O10000O10000000000000000000000O100000000000mJWHd4i7XK[Hh4e7UK_Hj4Q801O001O001O0000001O0000001O002N2N2N2NO100O10000O1000O010000O100O100O10000O100O10000O100000000000000O10000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000O1K5J60000000000O1000000000000000000000000000000001O000000001O002N2N3M?A2N3M0000000000001N100000000000001O000000000O10O000O2O000010O0001O01O01O00010O0100O1O010O1O100O1O10O01O100O100O00100O1O10O01O2O0O2M2E;G:L3M4L4L4N3L3N2N1O001O1O1000O10000OUNSEZ1m:fNTEY1l:gNTEY1k:hNVEW1j:hNWEX1i:hNXEW1g:jNZEU1f:kNZEU1f:kN[ET1d:mN]ER1c:nN]ER1c:nN^ER1`:nNaER1_:nNaER1_:nNaER1^:oNbER1_:lNaET1a:jN_EV1c:hN^EW1d:fN]EZ1e:dN[E\\1m:\\NTEc1Z;01O1O1O1O1O1O1O001N2O1N8I7H3N1N101N2O0O5K4M4KRVR4"}, "label": "the sheet cake including its decorations"}]}
{"id": 117922, "image": "COCO_train2014_000000117922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sign with stop on it facing the viewer\"."}], "task": "refering", "answer_template": "The \"a sign with stop on it facing the viewer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221], "bbox": [0.437109375, 0.1955075187969925, 0.700484375, 0.5101127819548872], "iscrowd": 0, "area": 22798.739700000002, "rle": {"size": [532, 640], "counts": "lca4U2]>3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2O1O1O1O1N2O1O00000000000000000000000O10000000000000000000000000000O01000000000000000000000000000O10000000000000000000000000000O10O1000000000001O1O1O1O1O001O1O1O1N2N1N3N2N2N2N2N2N1O2M3N2N2N2N1O2N2N2M3N2N1O2N2N2N2N2N2M2O2N2N2N2N2N1O2M3N2N2N2N2N1O2N2M?Bc0]Od0\\OjaS3"}, "label": "a sign with stop on it facing the viewer"}]}
{"id": 117922, "image": "COCO_train2014_000000117922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stop sign\"."}], "task": "refering", "answer_template": "The \"a stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221], "bbox": [0.437109375, 0.1955075187969925, 0.700484375, 0.5101127819548872], "iscrowd": 0, "area": 22798.739700000002, "rle": {"size": [532, 640], "counts": "lca4U2]>3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2O1O1O1O1N2O1O00000000000000000000000O10000000000000000000000000000O01000000000000000000000000000O10000000000000000000000000000O10O1000000000001O1O1O1O1O001O1O1O1N2N1N3N2N2N2N2N2N1O2M3N2N2N2N1O2N2N2M3N2N1O2N2N2N2N2N2M2O2N2N2N2N2N1O2M3N2N2N2N2N1O2N2M?Bc0]Od0\\OjaS3"}, "label": "a stop sign"}]}
{"id": 134309, "image": "COCO_train2014_000000134309.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the traffic light facing west\"."}], "task": "refering", "answer_template": "The \"the traffic light facing west\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 140, 141, 142, 143, 155, 156, 157, 158, 170, 171, 172, 173, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 231, 232, 233, 247, 248], "bbox": [0.38897911832946636, 0.38371875, 0.60169373549884, 0.704328125], "iscrowd": 0, "area": 16350.132499999992, "rle": {"size": [640, 431], "counts": "UXY3b1^b0h1WNX1iN2N2M3N3M2M3N2N2N1N101O1N2O001N2O001N2O1O0O2O1O1N101O1N101O1N2O001N2O1O001N2O001N2O1O0O2O1O00000001O0000001O0000001O00001O0000001O0000001O00001O01O0001O0000001O003Ml0TOl0TOl0TOl0TOl0TOl0TOfR[3"}, "label": "the traffic light facing west"}]}
{"id": 134309, "image": "COCO_train2014_000000134309.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the three bulb traffic light\"."}], "task": "refering", "answer_template": "The \"the three bulb traffic light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 140, 141, 142, 143, 155, 156, 157, 158, 170, 171, 172, 173, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 231, 232, 233, 247, 248], "bbox": [0.38897911832946636, 0.38371875, 0.60169373549884, 0.704328125], "iscrowd": 0, "area": 16350.132499999992, "rle": {"size": [640, 431], "counts": "UXY3b1^b0h1WNX1iN2N2M3N3M2M3N2N2N1N101O1N2O001N2O001N2O1O0O2O1O1N101O1N101O1N2O001N2O1O001N2O001N2O1O0O2O1O00000001O0000001O0000001O00001O0000001O0000001O00001O01O0001O0000001O003Ml0TOl0TOl0TOl0TOl0TOl0TOfR[3"}, "label": "the three bulb traffic light"}]}
{"id": 437416, "image": "COCO_train2014_000000437416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair near the window\"."}], "task": "refering", "answer_template": "The \"the chair near the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 196, 197, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 312, 332], "bbox": [0.42281250000000004, 0.4096041666666667, 0.599, 0.8346666666666667], "iscrowd": 0, "area": 8497.56555, "rle": {"size": [480, 640], "counts": "`Zo34R18\\<MgBo0U=b0nBfNV<S2N2O00N2ZNeCe0]<ZOgCFV=8j0O10O2O001O00001O001N101O001gCQO_:o0ZEZOd:g0TEAk:l0aD]O];j0TD_Ok;m1M2O00XOYDWNg;e1`DWNa;f1eDWN[;h1hDVNX;h1Q1QOTCIl<8ZC@h<`0]CYOe<g0]CSOg<m0f00000000O10000000000000000O100000000000000O1_Od0ZOi0F:F;E:F9G5J5L4LQOdEbMX:b2oEXMl9l2[FmL\\9\\3jF^LS9e3TGTLl8l3T10000000I7O010007I1O1O000000000009Go0QOe0[O\\McGeN]8W1PH`NP8]1\\HYNe7c1hHTNS7d1cISNX6j1YJlMg5l1kJkMU5m1S4H8H8H8G9H8H8HSYh3"}, "label": "the chair near the window"}]}
{"id": 437416, "image": "COCO_train2014_000000437416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red covered chair\"."}], "task": "refering", "answer_template": "The \"red covered chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 196, 197, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 312, 332], "bbox": [0.42281250000000004, 0.4096041666666667, 0.599, 0.8346666666666667], "iscrowd": 0, "area": 8497.56555, "rle": {"size": [480, 640], "counts": "`Zo34R18\\<MgBo0U=b0nBfNV<S2N2O00N2ZNeCe0]<ZOgCFV=8j0O10O2O001O00001O001N101O001gCQO_:o0ZEZOd:g0TEAk:l0aD]O];j0TD_Ok;m1M2O00XOYDWNg;e1`DWNa;f1eDWN[;h1hDVNX;h1Q1QOTCIl<8ZC@h<`0]CYOe<g0]CSOg<m0f00000000O10000000000000000O100000000000000O1_Od0ZOi0F:F;E:F9G5J5L4LQOdEbMX:b2oEXMl9l2[FmL\\9\\3jF^LS9e3TGTLl8l3T10000000I7O010007I1O1O000000000009Go0QOe0[O\\McGeN]8W1PH`NP8]1\\HYNe7c1hHTNS7d1cISNX6j1YJlMg5l1kJkMU5m1S4H8H8H8G9H8H8HSYh3"}, "label": "red covered chair"}]}
{"id": 511146, "image": "COCO_train2014_000000511146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane right behind the blue plane\"."}], "task": "refering", "answer_template": "The \"the plane right behind the blue plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 196, 197, 207, 208, 209, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 281, 282], "bbox": [0.001453125, 0.4991062801932367, 0.6251249999999999, 0.8107729468599033], "iscrowd": 0, "area": 10300.072000000002, "rle": {"size": [414, 640], "counts": "[e0P1n;d0\\O0000000000000O1000000000000000000000000000000000000000001O0000000000000000000O100000000000000000000000000000000001O0000001O0000001O00000010O000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O000000010O00000011N2N2N01O0000001O0O10001O0000001O0000001O000O10001O08HO10001O000000001O000000001O0O1000001O000000001O000000001O0O1000001O000000001O00000EXD@h;`0YD_Oh;`0YD^Oh;b0YD]Og;c0ZD\\Of;d0[D[Oe;e0\\DZOe;e0\\DZOd;f0]DYOc;g0:O101O000000001O0O1000001N1000000O2O00000O1000000O2O00000O10fXc1Edg\\N1O1O1O1O1O1O1O1O001O1O1O1N2O1O1O1O1O10000000000000000001N10000000000O1O1O2M2O1N2O1O1N2O1N2O`jP3"}, "label": "the plane right behind the blue plane"}]}
{"id": 511146, "image": "COCO_train2014_000000511146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plane with white body\"."}], "task": "refering", "answer_template": "The \"plane with white body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 196, 197, 207, 208, 209, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 281, 282], "bbox": [0.001453125, 0.4991062801932367, 0.6251249999999999, 0.8107729468599033], "iscrowd": 0, "area": 10300.072000000002, "rle": {"size": [414, 640], "counts": "[e0P1n;d0\\O0000000000000O1000000000000000000000000000000000000000001O0000000000000000000O100000000000000000000000000000000001O0000001O0000001O00000010O000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O000000010O00000011N2N2N01O0000001O0O10001O0000001O0000001O000O10001O08HO10001O000000001O000000001O0O1000001O000000001O000000001O0O1000001O000000001O00000EXD@h;`0YD_Oh;`0YD^Oh;b0YD]Og;c0ZD\\Of;d0[D[Oe;e0\\DZOe;e0\\DZOd;f0]DYOc;g0:O101O000000001O0O1000001N1000000O2O00000O1000000O2O00000O10fXc1Edg\\N1O1O1O1O1O1O1O1O001O1O1O1N2O1O1O1O1O10000000000000000001N10000000000O1O1O2M2O1N2O1O1N2O1N2O`jP3"}, "label": "plane with white body"}]}
{"id": 511146, "image": "COCO_train2014_000000511146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large airplane with the word american on the side\"."}], "task": "refering", "answer_template": "The \"a large airplane with the word american on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 116, 117, 118, 139, 140, 141, 142, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289], "bbox": [0.0073281250000000004, 0.23432367149758454, 0.98278125, 0.8193478260869564], "iscrowd": 0, "area": 54709.782049999994, "rle": {"size": [414, 640], "counts": "iS22k<3N3M3M2N3M2N3L3N3M2N3M2N3M2N3N1N3M2N3M2mF]Ng6f1VI\\Nh6f1WI\\Nf6g1WI[Ng6h1VIZNh6h1WIYNg6j1VIYNh6h1VIZNh6i1VIXNh6j1VIYNg6j1VIXNh6j1WIWNg6l1VIVNh6l1VIWNg6l1WIUNg6m1WIUNg6n1VITNi6m1VIRNj6Q2SIPNl6R2RInMn6U2PIjMQ7W2mHjMS7X2jHhMW7Y2hHfMY7\\2dHdM]7]2aHdM_7^2_HaMb7a2[H_Mf7b2XH^Mi7d2UH\\Mj7f2THZMm7h2PHXMQ8i2nGWMR8j31O1O100O1O1O100O1O10O01O1O100O1O1O100O1O100O1O1O010O1O1O100O1O100O1O1O100O1O00100O1O100O1O1O100O1O1O10O01O100O1O1O100O1O1O100O1O10O01O1O100O1O1O100O1O100O1O1O010O1O1O100O1O100O1O1O100O1O001O100O1O1O1O100O1O1O1O100O001O1O100O1O1O1O100O1O1O1O10O01O1O1O100O1O1O1O10O000000001O000000001O000000001O00000O101O000000001O000000001O000000001O000000001O000000001O000000001O000000001N100000001O000000001O0000001O000000001O000000001O000000001O000000001O001O1N2O1O1O1O1O1O1O1O1O1O1O1O001O1O1O00000000001O00000000001O0000000000001O00000O1000001O0000000000001O000000000000000000O10000000000000000O10000000000000O10O100000000000000O10000000000000000O100000000000000O10000O1000000O10000O010O1N2N2N2O1N2N2N2N2N2N20000000000000000001O00000000000000000000000000000O100000000000000000000000001O000000000000000000000000000000000000000000000O1000000000001O00000000000000000000000000000000000000000000000000000000001O0O10000000000000O1O100O1O100O100O1O100O100000000O10000000000O10000000000O10O10000000O10000000000O10000000000O10000000000O100000001O1N2O1O2N1O1O1N2O1O1O1O1O1N2O1O1O1O001N2O1O001O1O0O2O1O1O001O1N2O001O1O001N2O1O001O1O0O2O1M3N1O2N2M2O2N2N2M2O2N2N1N3NkU4"}, "label": "a large airplane with the word american on the side"}]}
{"id": 511146, "image": "COCO_train2014_000000511146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an american airlines jet\"."}], "task": "refering", "answer_template": "The \"an american airlines jet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 116, 117, 118, 139, 140, 141, 142, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289], "bbox": [0.0073281250000000004, 0.23432367149758454, 0.98278125, 0.8193478260869564], "iscrowd": 0, "area": 54709.782049999994, "rle": {"size": [414, 640], "counts": "iS22k<3N3M3M2N3M2N3L3N3M2N3M2N3M2N3N1N3M2N3M2mF]Ng6f1VI\\Nh6f1WI\\Nf6g1WI[Ng6h1VIZNh6h1WIYNg6j1VIYNh6h1VIZNh6i1VIXNh6j1VIYNg6j1VIXNh6j1WIWNg6l1VIVNh6l1VIWNg6l1WIUNg6m1WIUNg6n1VITNi6m1VIRNj6Q2SIPNl6R2RInMn6U2PIjMQ7W2mHjMS7X2jHhMW7Y2hHfMY7\\2dHdM]7]2aHdM_7^2_HaMb7a2[H_Mf7b2XH^Mi7d2UH\\Mj7f2THZMm7h2PHXMQ8i2nGWMR8j31O1O100O1O1O100O1O10O01O1O100O1O1O100O1O100O1O1O010O1O1O100O1O100O1O1O100O1O00100O1O100O1O1O100O1O1O10O01O100O1O1O100O1O1O100O1O10O01O1O100O1O1O100O1O100O1O1O010O1O1O100O1O100O1O1O100O1O001O100O1O1O1O100O1O1O1O100O001O1O100O1O1O1O100O1O1O1O10O01O1O1O100O1O1O1O10O000000001O000000001O000000001O00000O101O000000001O000000001O000000001O000000001O000000001O000000001O000000001N100000001O000000001O0000001O000000001O000000001O000000001O000000001O001O1N2O1O1O1O1O1O1O1O1O1O1O1O001O1O1O00000000001O00000000001O0000000000001O00000O1000001O0000000000001O000000000000000000O10000000000000000O10000000000000O10O100000000000000O10000000000000000O100000000000000O10000O1000000O10000O010O1N2N2N2O1N2N2N2N2N2N20000000000000000001O00000000000000000000000000000O100000000000000000000000001O000000000000000000000000000000000000000000000O1000000000001O00000000000000000000000000000000000000000000000000000000001O0O10000000000000O1O100O1O100O100O1O100O100000000O10000000000O10000000000O10O10000000O10000000000O10000000000O10000000000O100000001O1N2O1O2N1O1O1N2O1O1O1O1O1N2O1O1O1O001N2O1O001O1O0O2O1O1O001O1N2O001O1O001N2O1O001O1O0O2O1M3N1O2N2M2O2N2N2M2O2N2N1N3NkU4"}, "label": "an american airlines jet"}]}
{"id": 298160, "image": "COCO_train2014_000000298160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white colored bus\"."}], "task": "refering", "answer_template": "The \"white colored bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 55, 56, 57, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 100, 101, 102, 103, 117, 118, 119, 120, 121, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 149, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210], "bbox": [0.05175, 0.12358934169278996, 0.5265625, 0.40800940438871475], "iscrowd": 0, "area": 29700.43955, "rle": {"size": [638, 640], "counts": "Whd01lc02N3M3N2M3M3M2N3N2M3d]OWO^a0l0]^OZO_a0i0]^O\\O_a0f0]^OA_a0a0\\^OE`a0>[^OEba0]1K5K5K5K4L5K5L4K5K4L5K5L4K5K5L3L5K5L4K5K4M4K5N200000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O1000000000000O1000000000000000000O1000000000000000000O10000mLSAU1m>hNYAU1g>iN^AT1b>iNcAU1]>hNiAT1X>jNlAj0^>SOhAa0a>\\OdA8f>F_AMk>0ZAEo>8WA[OS?c0QAQOY?n0l1O100O1O100O1O100O100O1O100O1O100O1O10Qbd0B]^[O;E9G9G7I2N2N3M2N2N3M2N2N2N3M2M3N3M2N2N3M2N0000000000000000000000000000000000000000000000O1000000000000000000000000001O001O00001O00001O001O0O101O001O00001O001O00001O001O0000N9F=D<C<D=D<C=Dbdl5"}, "label": "white colored bus"}]}
{"id": 298160, "image": "COCO_train2014_000000298160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a bus with atlantic on the side\"."}], "task": "refering", "answer_template": "The \"this is a bus with atlantic on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 55, 56, 57, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 100, 101, 102, 103, 117, 118, 119, 120, 121, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 149, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210], "bbox": [0.05175, 0.12358934169278996, 0.5265625, 0.40800940438871475], "iscrowd": 0, "area": 29700.43955, "rle": {"size": [638, 640], "counts": "Whd01lc02N3M3N2M3M3M2N3N2M3d]OWO^a0l0]^OZO_a0i0]^O\\O_a0f0]^OA_a0a0\\^OE`a0>[^OEba0]1K5K5K5K4L5K5L4K5K4L5K5L4K5K5L3L5K5L4K5K4M4K5N200000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O1000000000000O1000000000000000000O1000000000000000000O10000mLSAU1m>hNYAU1g>iN^AT1b>iNcAU1]>hNiAT1X>jNlAj0^>SOhAa0a>\\OdA8f>F_AMk>0ZAEo>8WA[OS?c0QAQOY?n0l1O100O1O100O1O100O100O1O100O1O100O1O10Qbd0B]^[O;E9G9G7I2N2N3M2N2N3M2N2N2N3M2M3N3M2N2N3M2N0000000000000000000000000000000000000000000000O1000000000000000000000000001O001O00001O00001O001O0O101O001O00001O001O00001O001O0000N9F=D<C<D=D<C=Dbdl5"}, "label": "this is a bus with atlantic on the side"}]}
{"id": 298160, "image": "COCO_train2014_000000298160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black vehicle to the left of a man and a horse\"."}], "task": "refering", "answer_template": "The \"a black vehicle to the left of a man and a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 71, 92, 93, 94, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 300], "bbox": [0.0, 0.020313479623824453, 0.11546875000000001, 0.6046551724137931], "iscrowd": 0, "area": 13057.478449999997, "rle": {"size": [638, 640], "counts": "e0f0?o0a2hNg:6oBc20RNj<\\OYCb4b<_K`CX9e7iF_HS9`7nFcHo8]7QGgHj8Y7WGjHf8U7[GnHb8R7^GRI]8n6dGUIY8j6hGZIS8g6nG[Io7d6RH`Ij7`6VHcIf7_6YHeIc7[6]HiI_7X6`HlI[7U6eHoIW7R6hHQJT7Q6kHSJQ7m5PIVJl6k5SIYJh6h5XI[Je6f5ZI^Jb6b5^IbJ]6`5bIdJZ6\\5fIgJV6[5iIiJS6W5nIlJn5U5QJoJj5S5VJoJg5Q5YJSKc5n4]JUK^5l4cJWKY5j4gJYKT5h4mJ[Kn4g4RK[Kj4f4WK]Kd4e4[K_K`4b4aKaKZ4a4fKbKU4_4lKcKP4_4PLdKk3^4ULfKe3[4\\LhK_3Z4`LhIYOeNR4d7fLaI[OROQ3W8dMaH]O_OP2k8dN_G_Oc07k8:\\FAc=`0WBBm=>lAFV>;dAG`>9ZAIi>7QALS?4g@O^?O[@4k?Jo_O8X`0Ea_O>e`0@U_Oc0Pa0[Ob^Oo0da0a06J6K5J6J6J6K6I6J6J6K5JnaP;"}, "label": "a black vehicle to the left of a man and a horse"}]}
{"id": 298160, "image": "COCO_train2014_000000298160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black bus with a large mirror\"."}], "task": "refering", "answer_template": "The \"black bus with a large mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 71, 92, 93, 94, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 300], "bbox": [0.0, 0.020313479623824453, 0.11546875000000001, 0.6046551724137931], "iscrowd": 0, "area": 13057.478449999997, "rle": {"size": [638, 640], "counts": "e0f0?o0a2hNg:6oBc20RNj<\\OYCb4b<_K`CX9e7iF_HS9`7nFcHo8]7QGgHj8Y7WGjHf8U7[GnHb8R7^GRI]8n6dGUIY8j6hGZIS8g6nG[Io7d6RH`Ij7`6VHcIf7_6YHeIc7[6]HiI_7X6`HlI[7U6eHoIW7R6hHQJT7Q6kHSJQ7m5PIVJl6k5SIYJh6h5XI[Je6f5ZI^Jb6b5^IbJ]6`5bIdJZ6\\5fIgJV6[5iIiJS6W5nIlJn5U5QJoJj5S5VJoJg5Q5YJSKc5n4]JUK^5l4cJWKY5j4gJYKT5h4mJ[Kn4g4RK[Kj4f4WK]Kd4e4[K_K`4b4aKaKZ4a4fKbKU4_4lKcKP4_4PLdKk3^4ULfKe3[4\\LhK_3Z4`LhIYOeNR4d7fLaI[OROQ3W8dMaH]O_OP2k8dN_G_Oc07k8:\\FAc=`0WBBm=>lAFV>;dAG`>9ZAIi>7QALS?4g@O^?O[@4k?Jo_O8X`0Ea_O>e`0@U_Oc0Pa0[Ob^Oo0da0a06J6K5J6J6J6K6I6J6J6K5JnaP;"}, "label": "black bus with a large mirror"}]}
{"id": 27826, "image": "COCO_train2014_000000027826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traveling lady\"."}], "task": "refering", "answer_template": "The \"traveling lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 193, 194, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 348, 349, 350, 351, 365, 366, 367, 368, 382, 383, 384, 385], "bbox": [0.3702083333333333, 0.1529375, 0.8307083333333334, 0.98815625], "iscrowd": 0, "area": 63575.541349999985, "rle": {"size": [640, 480], "counts": "Pb_37cc07O1N1O1O1O100O1O1O1O100O1N2K5J6K5UNk1M3N2M2O2N2M3N2M01O1N3N2M3N2N2NiBQMb9n2RFbMk9]2iESNT:m1`EbN^:\\1kD_OR;`0nDDo:;QEJk:6TENi:5SEOk:3QE1l:2PE2m:0PE4m:OnD6P;LlD8Q;KkD9R;JjD;R;HjD<T;FhD`0S;CiDg0f9cKSGi3SOn0a9dKRGh8d8dGQGe8e8_GYGj8]8XGaGQ9U8RGjGV9l7nFQH[9f7gFYHn9P7d2mNR1nNc0]O1O1N2O1O2N1DU@bLl?^3V@_Lk?`3Y@\\Lh?c3<N100O1O00N3M2N2N2N3L3N2N2N3M3M3M3L4M3TKdLQI_3o6nLaHU3`7WMPHl2Q8`M_Gc2a8kMcFc2^9o31O1O001O1O001O100O00100O001Oc3^L`0_Ob0^Oa0@<C1O1O010O1O1001O1O1N2N3M2N2N3M2aAnKl<T4lBTLR=o3eBXLZ=j3]B_La=c3WBdLh=_3oAhLP>f4N1O1O2N1O1O1O1O1O2N1nLoISHR6`7aJYH`5Z7RK`Ho4S7cKgH^4k6ULoHm3c6fLVI[3d6PMVIQ3h6VMRIk2k6]MoHd2o6dMjH]2S7kMgHW2V7PNdHQ2Y7WNaHj1]7]N]Hd1`7`5N2M3N2M3N1N3N2M3N2M2N3N2M3O0O2N2O1N2N101N2N2O0O2N2O1N2N4L4M2M4L4M2M4L4M3L3M4M3L3M2O2M2SMf@Q1]?lNl@j0W?UOQA=W?@RA2U?MSAFT?8TAZOS?d0VAnNR?P1Q2N1N3N1O2M2O2M2OiZb1"}, "label": "traveling lady"}]}
{"id": 27826, "image": "COCO_train2014_000000027826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a blue shirt holding a lot of luggage smiling\"."}], "task": "refering", "answer_template": "The \"a girl wearing a blue shirt holding a lot of luggage smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 193, 194, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 348, 349, 350, 351, 365, 366, 367, 368, 382, 383, 384, 385], "bbox": [0.3702083333333333, 0.1529375, 0.8307083333333334, 0.98815625], "iscrowd": 0, "area": 63575.541349999985, "rle": {"size": [640, 480], "counts": "Pb_37cc07O1N1O1O1O100O1O1O1O100O1N2K5J6K5UNk1M3N2M2O2N2M3N2M01O1N3N2M3N2N2NiBQMb9n2RFbMk9]2iESNT:m1`EbN^:\\1kD_OR;`0nDDo:;QEJk:6TENi:5SEOk:3QE1l:2PE2m:0PE4m:OnD6P;LlD8Q;KkD9R;JjD;R;HjD<T;FhD`0S;CiDg0f9cKSGi3SOn0a9dKRGh8d8dGQGe8e8_GYGj8]8XGaGQ9U8RGjGV9l7nFQH[9f7gFYHn9P7d2mNR1nNc0]O1O1N2O1O2N1DU@bLl?^3V@_Lk?`3Y@\\Lh?c3<N100O1O00N3M2N2N2N3L3N2N2N3M3M3M3L4M3TKdLQI_3o6nLaHU3`7WMPHl2Q8`M_Gc2a8kMcFc2^9o31O1O001O1O001O100O00100O001Oc3^L`0_Ob0^Oa0@<C1O1O010O1O1001O1O1N2N3M2N2N3M2aAnKl<T4lBTLR=o3eBXLZ=j3]B_La=c3WBdLh=_3oAhLP>f4N1O1O2N1O1O1O1O1O2N1nLoISHR6`7aJYH`5Z7RK`Ho4S7cKgH^4k6ULoHm3c6fLVI[3d6PMVIQ3h6VMRIk2k6]MoHd2o6dMjH]2S7kMgHW2V7PNdHQ2Y7WNaHj1]7]N]Hd1`7`5N2M3N2M3N1N3N2M3N2M2N3N2M3O0O2N2O1N2N101N2N2O0O2N2O1N2N4L4M2M4L4M2M4L4M3L3M4M3L3M2O2M2SMf@Q1]?lNl@j0W?UOQA=W?@RA2U?MSAFT?8TAZOS?d0VAnNR?P1Q2N1N3N1O2M2O2M2OiZb1"}, "label": "a girl wearing a blue shirt holding a lot of luggage smiling"}]}
{"id": 175284, "image": "COCO_train2014_000000175284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child in a red coat\"."}], "task": "refering", "answer_template": "The \"a small child in a red coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 63, 64, 65, 81, 82, 83, 98, 99, 100, 101, 117, 118, 119, 135, 136, 137, 153, 154, 155, 171, 172, 173], "bbox": [0.48625999999999997, 0.15909333333333334, 0.66444, 0.7497866666666666], "iscrowd": 0, "area": 12360.414, "rle": {"size": [375, 500], "counts": "fTi22a;8H7I7I8H7I7I7J7H7I7I8K4L4L5K6I6K5K6J5K5K6J5K5K5J7J5XORK^JS5[5RKbJT5W5PKfJU5S5PKjJS5R5PKlJS5o4RKnJP5o4n0N1O1000O0100O100O01lMTK]Mk4k2XKkLh4\\3[KZLe4m3^KiKb4^4aKWK`4P5dKeJ]4`5gKVJY4Q6d06I7J4K01O000O2O001O0SJRKQ5o4^JdK_5\\5N3L4M2N3M1O1O0fN[1VOj0O1N101O1N2O0O2O1N6K4K6SOjFoN[9i0o0H9H8G8In[m1"}, "label": "a small child in a red coat"}]}
{"id": 175284, "image": "COCO_train2014_000000175284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a red jacket learning to ski\"."}], "task": "refering", "answer_template": "The \"a little boy in a red jacket learning to ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 63, 64, 65, 81, 82, 83, 98, 99, 100, 101, 117, 118, 119, 135, 136, 137, 153, 154, 155, 171, 172, 173], "bbox": [0.48625999999999997, 0.15909333333333334, 0.66444, 0.7497866666666666], "iscrowd": 0, "area": 12360.414, "rle": {"size": [375, 500], "counts": "fTi22a;8H7I7I8H7I7I7J7H7I7I8K4L4L5K6I6K5K6J5K5K6J5K5K5J7J5XORK^JS5[5RKbJT5W5PKfJU5S5PKjJS5R5PKlJS5o4RKnJP5o4n0N1O1000O0100O100O01lMTK]Mk4k2XKkLh4\\3[KZLe4m3^KiKb4^4aKWK`4P5dKeJ]4`5gKVJY4Q6d06I7J4K01O000O2O001O0SJRKQ5o4^JdK_5\\5N3L4M2N3M1O1O0fN[1VOj0O1N101O1N2O0O2O1N6K4K6SOjFoN[9i0o0H9H8G8In[m1"}, "label": "a little boy in a red jacket learning to ski"}]}
{"id": 175284, "image": "COCO_train2014_000000175284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"taller child wearing mint green pants holding skii sticks with skiis on their feet\"."}], "task": "refering", "answer_template": "The \"taller child wearing mint green pants holding skii sticks with skiis on their feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 26, 40, 41, 42, 43, 44, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 96, 97, 98, 114, 115, 116, 132, 133, 134, 150, 151, 152, 168, 169, 170], "bbox": [0.26006, 0.10088, 0.5050399999999999, 0.7697866666666667], "iscrowd": 0, "area": 17023.497649999998, "rle": {"size": [375, 500], "counts": "Wf_1n0X:c0M4M2N2M3N2M3N00O010O0010O01O10O01O010O101N3M2OGYNcFe1\\9`NbF^1\\9gNcFX1[9lNdFR1[9ROdFl0Z9ZOdFe0Z9_OfF>Y9FfF9X9S100100O1i0WOb0]O6K4L5K5K4K6K<D=C=C>C2N101N1O101N1O2N100O2N4M3L5K4M0ON2O2N1O1O2N1O001O001N1N2O2N1N2TNoJQMS5S3QKbLQ5c3TKRLn4Q4WKeKj4`4ZKUKh4P5]KeJe4^58d0J2N3M3M3M2O2001O0O1hKiJo1W5PNQKi1o4VNXKd1h4ZNaK^1`4aNhKX1Y4fNoKS1Q4kNXLn0h3QO_Lj0`3UOhLe0W3YOPMb0P3]OTMa0k2^OYM`0g2^O^M?a2_OdM?]2^OhM=\\2_OhM=\\2^OiM>\\2]OiM>[2^OiM>[2]OjM>[2^OjM=Z2_OjM=[2^OS5K5L4LmZj2"}, "label": "taller child wearing mint green pants holding skii sticks with skiis on their feet"}]}
{"id": 175284, "image": "COCO_train2014_000000175284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid in teal snow pants holding ski poles\"."}], "task": "refering", "answer_template": "The \"a kid in teal snow pants holding ski poles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 26, 40, 41, 42, 43, 44, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 96, 97, 98, 114, 115, 116, 132, 133, 134, 150, 151, 152, 168, 169, 170], "bbox": [0.26006, 0.10088, 0.5050399999999999, 0.7697866666666667], "iscrowd": 0, "area": 17023.497649999998, "rle": {"size": [375, 500], "counts": "Wf_1n0X:c0M4M2N2M3N2M3N00O010O0010O01O10O01O010O101N3M2OGYNcFe1\\9`NbF^1\\9gNcFX1[9lNdFR1[9ROdFl0Z9ZOdFe0Z9_OfF>Y9FfF9X9S100100O1i0WOb0]O6K4L5K5K4K6K<D=C=C>C2N101N1O101N1O2N100O2N4M3L5K4M0ON2O2N1O1O2N1O001O001N1N2O2N1N2TNoJQMS5S3QKbLQ5c3TKRLn4Q4WKeKj4`4ZKUKh4P5]KeJe4^58d0J2N3M3M3M2O2001O0O1hKiJo1W5PNQKi1o4VNXKd1h4ZNaK^1`4aNhKX1Y4fNoKS1Q4kNXLn0h3QO_Lj0`3UOhLe0W3YOPMb0P3]OTMa0k2^OYM`0g2^O^M?a2_OdM?]2^OhM=\\2_OhM=\\2^OiM>\\2]OiM>[2^OiM>[2]OjM>[2^OjM=Z2_OjM=[2^OS5K5L4LmZj2"}, "label": "a kid in teal snow pants holding ski poles"}]}
{"id": 502971, "image": "COCO_train2014_000000502971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike in the front ridden by a person doing a stunt\"."}], "task": "refering", "answer_template": "The \"the bike in the front ridden by a person doing a stunt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 291, 292], "bbox": [0.6193125, 0.5237704918032787, 0.84071875, 0.8309367681498829], "iscrowd": 0, "area": 7263.29825, "rle": {"size": [427, 640], "counts": "PkU51Y=1O001O1O1O10O01O1O1O1O001O1O1O1O1SDJd:6VE3h:MPE<n:EQE>l:DRE?k:BTE?k:BSEa0l:_OSEb0l:@REb0m:^OREc0m:^OREc0n:]OPEf0n:\\OQEd0o:\\OQEd0o:\\OQEe0m:\\OSEd0m:]OREd0l:]OSEd0m:\\OSEd0l:]OTEb0m:_OREa0m:@SE`0m:@SE`0l:ATE?b9cNfG]2Z8cMeG]2\\8dMcG\\2]8dMbG]2_8cM`G]2`8cM_G]2b8dM[G]2f8cMXG^2i8cMTG^2m8bMQG_2P9bMmF_2T9aMjF`2W9aMfF_2\\9bMaF_2`9aM^Fm1F[Nn9IYFj1NZNk9LUFk11XNl9MPFl15UNm9OlEm19QNm9Z2SFdMo9[2SFcMn9\\2a0O100O1O2O0O1O1O2O0O1O00001O00001O0000001O0mNTEHl:9XEAj:`0XE\\Oh:0gD7d0Ee:4iD7d0Ad:7iD9Q<GQD8n;HTD8k;GWD9g;GZD9f;F[D;d;D\\D`0a;_O_Dd0_;[OaDh0];WOdDk0Z;TOfDo0X;POhDS1V;lNkDV1c;2N2O1O1O1N101O1O001O001N2OO100000001N1000001O000O2O1O1O1N2M3N2N3M2M3N4K5I7I7IPYZ1"}, "label": "the bike in the front ridden by a person doing a stunt"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden toilet seat in the stall on the right\"."}], "task": "refering", "answer_template": "The \"wooden toilet seat in the stall on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 363, 364, 365, 366, 387], "bbox": [0.709546875, 0.2831458333333333, 0.9758437499999999, 0.9505625], "iscrowd": 0, "area": 37061.39705000001, "rle": {"size": [480, 640], "counts": "]Re61o>3nA:c<X1J6I7I7J7H4M3L5K4M3L5L3L3M100O1SOm0K5K5K5K5K5K5K4L5K5K5N2N2N2M3N5K4L5J6K4M4M3L5K5L4K5L4K4M3L3N2M3M4M2M3N2M4M2M3M3N3L3N2M3N3L100O101O0O10000O100O101O0O100O10000O2O0O10000O1000001O01O00000000001N100000000O101O000O101O001O1N2O1O001N2O1O1O001^J_Jb1c5\\N^Jd1c5YN^Jh1c5VN^Ji1d5UN]Jj1d5TN^Jk1d5SN\\Jm1f5QN[Jn1g5oM]Jn1e5PNPK[1Q5cNTKY1n4eNUKW1o4fNSKX1Q5cNSKZ1P5cNSKZ1Q5bNRK[1Q5aNRK]1R5_NQK^1R5_NQK^1R5^NQK`1S5\\NPKa1S5\\NPKa1T5ZNoJc1U5ZNnJc1d6kL_IR3Y9N3H7G:K4N3M2N3M2N3M2N3M2O2M2N3M2N3M2N3M4L7I7I8H7I7Iod6"}, "label": "wooden toilet seat in the stall on the right"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toilet at right\"."}], "task": "refering", "answer_template": "The \"toilet at right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 363, 364, 365, 366, 387], "bbox": [0.709546875, 0.2831458333333333, 0.9758437499999999, 0.9505625], "iscrowd": 0, "area": 37061.39705000001, "rle": {"size": [480, 640], "counts": "]Re61o>3nA:c<X1J6I7I7J7H4M3L5K4M3L5L3L3M100O1SOm0K5K5K5K5K5K5K4L5K5K5N2N2N2M3N5K4L5J6K4M4M3L5K5L4K5L4K4M3L3N2M3M4M2M3N2M4M2M3M3N3L3N2M3N3L100O101O0O10000O100O101O0O100O10000O2O0O10000O1000001O01O00000000001N100000000O101O000O101O001O1N2O1O001N2O1O1O001^J_Jb1c5\\N^Jd1c5YN^Jh1c5VN^Ji1d5UN]Jj1d5TN^Jk1d5SN\\Jm1f5QN[Jn1g5oM]Jn1e5PNPK[1Q5cNTKY1n4eNUKW1o4fNSKX1Q5cNSKZ1P5cNSKZ1Q5bNRK[1Q5aNRK]1R5_NQK^1R5_NQK^1R5^NQK`1S5\\NPKa1S5\\NPKa1T5ZNoJc1U5ZNnJc1d6kL_IR3Y9N3H7G:K4N3M2N3M2N3M2N3M2O2M2N3M2N3M2N3M4L7I7I8H7I7Iod6"}, "label": "toilet at right"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden toilet to the left of another wooden toilet\"."}], "task": "refering", "answer_template": "The \"a wooden toilet to the left of another wooden toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 117, 118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 323, 324, 325, 326, 347, 348, 349, 371], "bbox": [0.040453125, 0.28020833333333334, 0.32190624999999995, 0.949875], "iscrowd": 0, "area": 34722.259900000005, "rle": {"size": [480, 640], "counts": "aa<3j>5K4M4K5K4L5L4K4L5K5L3L5K5M2N3M3M2O2M3M2O2M3M2O1N2N1O2O0O2N2O0O2N101N1O2N2O0fLRMaJP3S5^MhJd2m4jMoJX2e4WNVKj1d4_NXKc1f4bNUK`1i4dNRK^1m4fNoJ\\1n4jNmJY1P5kNlJX1Q5mNjJV1R5oNiJU1T5oNhJT1U5QOfJR1W5ROeJQ1W5_4M3M3L4M3M3L4M3N2M10O010O0100O01000O010O10O010O0100O010O10O10000O100O10000O10000O10000O1N3L3M3M3L4M2N2M3N3N3L4M3L4M3L4M3M3L4I7J6J6N2N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1N2O1O1O2N1O1O1YLjFa1X9kM\\GT2e8YMmGg2T8eL_H[3d9O0000001N1O1N2N3M2N2M4M2N2N2N3M;Da0@?ROWU[6"}, "label": "a wooden toilet to the left of another wooden toilet"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the commode on the left\"."}], "task": "refering", "answer_template": "The \"the commode on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 117, 118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 323, 324, 325, 326, 347, 348, 349, 371], "bbox": [0.040453125, 0.28020833333333334, 0.32190624999999995, 0.949875], "iscrowd": 0, "area": 34722.259900000005, "rle": {"size": [480, 640], "counts": "aa<3j>5K4M4K5K4L5L4K4L5K5L3L5K5M2N3M3M2O2M3M2O2M3M2O1N2N1O2O0O2N2O0O2N101N1O2N2O0fLRMaJP3S5^MhJd2m4jMoJX2e4WNVKj1d4_NXKc1f4bNUK`1i4dNRK^1m4fNoJ\\1n4jNmJY1P5kNlJX1Q5mNjJV1R5oNiJU1T5oNhJT1U5QOfJR1W5ROeJQ1W5_4M3M3L4M3M3L4M3N2M10O010O0100O01000O010O10O010O0100O010O10O10000O100O10000O10000O10000O1N3L3M3M3L4M2N2M3N3N3L4M3L4M3L4M3M3L4I7J6J6N2N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1N2O1O1O2N1O1O1YLjFa1X9kM\\GT2e8YMmGg2T8eL_H[3d9O0000001N1O1N2N3M2N2M4M2N2N2N3M;Da0@?ROWU[6"}, "label": "the commode on the left"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a trash container is shown to the right of a commode in a crowded stall\"."}], "task": "refering", "answer_template": "The \"a trash container is shown to the right of a commode in a crowded stall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.17696875, 0.7681666666666668, 0.45168749999999996, 0.9906458333333334], "iscrowd": 0, "area": 15421.046899999998, "rle": {"size": [480, 640], "counts": "k]e11l>3M3M2N3N2M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3N2M3M3M3M3M3N1N3M3M3O100000000000000000000000000000000001O00000000000000000O100000000000000000000000000000001O000000000000000000000000000O100000000000000000001O0000000000000000000000000000000000000O1000000000001O00000000000000000000000000000000000000000000000O2O000000000000000O1F:F:F:F:F:F:F:F:FdUT5"}, "label": "a trash container is shown to the right of a commode in a crowded stall"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink on the left\"."}], "task": "refering", "answer_template": "The \"the sink on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.17696875, 0.7681666666666668, 0.45168749999999996, 0.9906458333333334], "iscrowd": 0, "area": 15421.046899999998, "rle": {"size": [480, 640], "counts": "k]e11l>3M3M2N3N2M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3N2M3M3M3M3M3N1N3M3M3O100000000000000000000000000000000001O00000000000000000O100000000000000000000000000000001O000000000000000000000000000O100000000000000000001O0000000000000000000000000000000000000O1000000000001O00000000000000000000000000000000000000000000000O2O000000000000000O1F:F:F:F:F:F:F:F:FdUT5"}, "label": "the sink on the left"}]}
{"id": 314556, "image": "COCO_train2014_000000314556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a square trough in a bathroom\"."}], "task": "refering", "answer_template": "The \"a square trough in a bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.52753125, 0.79325, 0.7955, 0.9842708333333333], "iscrowd": 0, "area": 13893.5686, "rle": {"size": [480, 640], "counts": "oin49g>;E:F:F;E:F;E:F7I00000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000004L3M4K4M4L4L3M4L4L3M4L3M4L4L3M4L3M4L4L3M4L3M4L4L\\nl1"}, "label": "a square trough in a bathroom"}]}
{"id": 3259, "image": "COCO_train2014_000000003259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop , open and turned on\"."}], "task": "refering", "answer_template": "The \"laptop , open and turned on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.5314687499999999, 0.31872651356993736, 0.9670781250000001, 0.8018789144050105], "iscrowd": 0, "area": 34861.2038, "rle": {"size": [479, 640], "counts": "ZZo41m>3M2N3M3M3M3N2M3N2N2N2N2N1O2N2N2N2M3N2N2N2N2N2N1O2N2M3N2N2N2N2N2N2N2N1O2M3N2N2N2N2N2N2N2N1N3N2N1O1O1O001O1O0O2O1O1O0000000000000O10O10000000000000O10000000O1000000000O1000000000O100000O1000000000000000O10O10000000000000000O10O100000000000001O00001N01O1O1G9M3N2M3M3M3N2M3M2N3M3N2M3M3M3M3N2M3M3M2N3N2M3M3M3M3N2M3M3M3M3N1N3M3M3N2M3N2O1O1O100000O2O00001O00001O001O000O2O00001O00001O00001O001N10001O00001O00001O000O2O00001N1M4L3M3M4L30001O00001O00001M2N2M4L3M4L3M3N3L3M3M4L3M3N3L3M3M4L3M3M4M2M4L3M3M4L3N2M4L3M3M4L3N2M4L3M3M4L3M4M2M3M4L3M3M4M2M3M4L3M3M4L3N2M4L3M_d9"}, "label": "laptop , open and turned on"}]}
{"id": 3259, "image": "COCO_train2014_000000003259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and silver laptop on a wooden desk\"."}], "task": "refering", "answer_template": "The \"a black and silver laptop on a wooden desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.5314687499999999, 0.31872651356993736, 0.9670781250000001, 0.8018789144050105], "iscrowd": 0, "area": 34861.2038, "rle": {"size": [479, 640], "counts": "ZZo41m>3M2N3M3M3M3N2M3N2N2N2N2N1O2N2N2N2M3N2N2N2N2N2N1O2N2M3N2N2N2N2N2N2N2N1O2M3N2N2N2N2N2N2N2N1N3N2N1O1O1O001O1O0O2O1O1O0000000000000O10O10000000000000O10000000O1000000000O1000000000O100000O1000000000000000O10O10000000000000000O10O100000000000001O00001N01O1O1G9M3N2M3M3M3N2M3M2N3M3N2M3M3M3M3N2M3M3M2N3N2M3M3M3M3N2M3M3M3M3N1N3M3M3N2M3N2O1O1O100000O2O00001O00001O001O000O2O00001O00001O00001O001N10001O00001O00001O000O2O00001N1M4L3M3M4L30001O00001O00001M2N2M4L3M4L3M3N3L3M3M4L3M3N3L3M3M4L3M3M4M2M4L3M3M4L3N2M4L3M3M4L3N2M4L3M3M4L3M4M2M3M4L3M3M4M2M3M4L3M3M4L3N2M4L3M_d9"}, "label": "a black and silver laptop on a wooden desk"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear vase in the back left of the group of vases\"."}], "task": "refering", "answer_template": "The \"clear vase in the back left of the group of vases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 189, 190, 191, 206, 207, 208, 223, 224, 225, 239, 240, 241, 242, 256, 257, 258, 259, 273, 274, 275, 290, 291, 292, 307, 308, 309, 324, 325, 326, 341, 342, 343, 358, 359, 360, 376, 377], "bbox": [0.07820833333333334, 0.45721875, 0.28775, 0.9659062500000001], "iscrowd": 0, "area": 20021.649649999996, "rle": {"size": [640, 480], "counts": "^Zh0c0gb0f0[Oe0[Of0ZOe0ZOf0[Oe0[Oe0ZOf0I7N2N3M2N2O1N2N2N2N3N1N2N2N2N2O1N3M2N2N2^M]IiFa7U9f1000001O0000000000001O0000000000001O000000000000001O000000kK[FoMe9i1TGfMl8S2lG\\MU8\\2cHTM\\7g2ZIhLf6X3kIWLU6h3]JgKc5X4oJWKQ5h4T5O1O1O1O1O1O1O1O1O1O1O1O2O5J5K5K6J5Ke1\\Nd0[O4L4M3L4L4M3L4L4M3L4L4M3LhZe6"}, "label": "clear vase in the back left of the group of vases"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase to the left of the one with an open rose contains a white unopened flower\"."}], "task": "refering", "answer_template": "The \"a vase to the left of the one with an open rose contains a white unopened flower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 189, 190, 191, 206, 207, 208, 223, 224, 225, 239, 240, 241, 242, 256, 257, 258, 259, 273, 274, 275, 290, 291, 292, 307, 308, 309, 324, 325, 326, 341, 342, 343, 358, 359, 360, 376, 377], "bbox": [0.07820833333333334, 0.45721875, 0.28775, 0.9659062500000001], "iscrowd": 0, "area": 20021.649649999996, "rle": {"size": [640, 480], "counts": "^Zh0c0gb0f0[Oe0[Of0ZOe0ZOf0[Oe0[Oe0ZOf0I7N2N3M2N2O1N2N2N2N3N1N2N2N2N2O1N3M2N2N2^M]IiFa7U9f1000001O0000000000001O0000000000001O000000000000001O000000kK[FoMe9i1TGfMl8S2lG\\MU8\\2cHTM\\7g2ZIhLf6X3kIWLU6h3]JgKc5X4oJWKQ5h4T5O1O1O1O1O1O1O1O1O1O1O1O2O5J5K5K6J5Ke1\\Nd0[O4L4M3L4L4M3L4L4M3L4L4M3LhZe6"}, "label": "a vase to the left of the one with an open rose contains a white unopened flower"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with blonde hair\"."}], "task": "refering", "answer_template": "The \"woman with blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 110, 111, 127, 128, 143, 144, 146, 159, 160, 161, 162, 163, 180, 181, 194, 197, 198, 199, 211, 212, 214, 215, 216, 228, 229, 231, 232, 245, 246, 248, 249, 251, 262, 265, 266, 268, 282, 283, 285, 299, 316], "bbox": [0.39839583333333334, 0.248484375, 0.8156041666666667, 0.8108281249999999], "iscrowd": 0, "area": 17863.912149999993, "rle": {"size": [640, 480], "counts": "_dg31jc05L4K5L4K5L4K5L4L4R_OOk=3PBa0^=B]BR1Q=QOnB^1a<eN^C_1_<aNaCa1^<^NaCe1]<[NbCg1]<YNcCi1[<WNdCk1[<UNdCn1Z<RNfCP2X<PNhCQ2W<oMjCR2T<nMlCT2S<kMmCV2R<jMnCW2R<hMoCW2Q<iMoCW2Q<iMoCW2R<hMnCX2R<hMnCX2S<gMnCX2R<hMnCX2S<fMnCZ2R<fMnCZ2R<fMoCY2R<fMT501O00UOgNo^OY1Pa0iNo^OW1o`0mNo^OS1o`0POP_Oo0Pa0TOn^Ol0Pa0WOo^Oi0o`0[Oo^Oe0o`0^OP_Ob0o`0Ao^O?o`0DP_O<n`0HP_O7Pa0JP_O6n`0LR_O4l`0OS_O1k`01U_OOj`02V_ONh`05W_OKg`07Y_OHg`09Y_OGe`0;[_OEc`0>\\_OBb`0`0^_O@a`0a0g_OWOW`0l0P@lNn?V1]1O3[Of0[Od0E^W80jhG3N2N2M4M2N2M3N4K8l_O_O\\<g0RCc0P<C^Cf1d;aNjCi2X;]MVDm3l:YLbDW4Y;Y22N3M2N3M2N2N3M2N3M2N3M2J7H7I7I8G8I8H7I7H7K5M2O2N2M3N1O2M3N1O2M3N1O2M3N1O2M3N2M2K6K5J5L5J6K4K6K5G8F;E;F9F;E;F:E:Fck=KhTB5K5K5K4M4K5L:M=CM3L4M3M3M3L4M3M3L4M3M3L4M3M3MeUg1"}, "label": "woman with blonde hair"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blond woman behind the flowers in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the blond woman behind the flowers in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 110, 111, 127, 128, 143, 144, 146, 159, 160, 161, 162, 163, 180, 181, 194, 197, 198, 199, 211, 212, 214, 215, 216, 228, 229, 231, 232, 245, 246, 248, 249, 251, 262, 265, 266, 268, 282, 283, 285, 299, 316], "bbox": [0.39839583333333334, 0.248484375, 0.8156041666666667, 0.8108281249999999], "iscrowd": 0, "area": 17863.912149999993, "rle": {"size": [640, 480], "counts": "_dg31jc05L4K5L4K5L4K5L4L4R_OOk=3PBa0^=B]BR1Q=QOnB^1a<eN^C_1_<aNaCa1^<^NaCe1]<[NbCg1]<YNcCi1[<WNdCk1[<UNdCn1Z<RNfCP2X<PNhCQ2W<oMjCR2T<nMlCT2S<kMmCV2R<jMnCW2R<hMoCW2Q<iMoCW2Q<iMoCW2R<hMnCX2R<hMnCX2S<gMnCX2R<hMnCX2S<fMnCZ2R<fMnCZ2R<fMoCY2R<fMT501O00UOgNo^OY1Pa0iNo^OW1o`0mNo^OS1o`0POP_Oo0Pa0TOn^Ol0Pa0WOo^Oi0o`0[Oo^Oe0o`0^OP_Ob0o`0Ao^O?o`0DP_O<n`0HP_O7Pa0JP_O6n`0LR_O4l`0OS_O1k`01U_OOj`02V_ONh`05W_OKg`07Y_OHg`09Y_OGe`0;[_OEc`0>\\_OBb`0`0^_O@a`0a0g_OWOW`0l0P@lNn?V1]1O3[Of0[Od0E^W80jhG3N2N2M4M2N2M3N4K8l_O_O\\<g0RCc0P<C^Cf1d;aNjCi2X;]MVDm3l:YLbDW4Y;Y22N3M2N3M2N2N3M2N3M2N3M2J7H7I7I8G8I8H7I7H7K5M2O2N2M3N1O2M3N1O2M3N1O2M3N1O2M3N2M2K6K5J5L5J6K4K6K5G8F;E;F9F;E;F:E:Fck=KhTB5K5K5K4M4K5L:M=CM3L4M3M3M3L4M3M3L4M3M3L4M3M3MeUg1"}, "label": "the blond woman behind the flowers in the right hand picture"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a black , having his hands beside on the plant bottle\"."}], "task": "refering", "answer_template": "The \"a person wearing a black , having his hands beside on the plant bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 170, 171, 172, 187, 188, 189, 204, 205, 206, 221, 222, 223, 238, 239, 240, 255, 256, 272, 273, 289, 290, 306, 307], "bbox": [0.0029791666666666664, 0.404171875, 0.13720833333333335, 0.8068593750000002], "iscrowd": 0, "area": 13402.759949999998, "rle": {"size": [640, 480], "counts": "Sl02_7>`4<hEb4X:c2001O000000001O0000001O0O10001O00001O0000001O00001O000O101O0000001O00001O0000001aM^2N2N3M2M3N2N3M2M3N2N3M2M3N3M2N2M4M4L3M4K5L4LV_R8"}, "label": "a person wearing a black , having his hands beside on the plant bottle"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of a man sitting down\"."}], "task": "refering", "answer_template": "The \"the arm of a man sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 170, 171, 172, 187, 188, 189, 204, 205, 206, 221, 222, 223, 238, 239, 240, 255, 256, 272, 273, 289, 290, 306, 307], "bbox": [0.0029791666666666664, 0.404171875, 0.13720833333333335, 0.8068593750000002], "iscrowd": 0, "area": 13402.759949999998, "rle": {"size": [640, 480], "counts": "Sl02_7>`4<hEb4X:c2001O000000001O0000001O0O10001O00001O0000001O00001O000O101O0000001O00001O0000001aM^2N2N3M2M3N2N3M2M3N2N3M2M3N3M2N2M4M4L3M4K5L4LV_R8"}, "label": "the arm of a man sitting down"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front most vase\"."}], "task": "refering", "answer_template": "The \"the front most vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 209, 210, 211, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 275, 276, 277, 278, 279, 280, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382], "bbox": [0.20675, 0.48410937499999995, 0.5273333333333333, 0.9842656250000001], "iscrowd": 0, "area": 35410.933899999996, "rle": {"size": [640, 480], "counts": "iZn1;_c0W1iNV1lNl0WO5K4L4L5K4L4M4K4L4L5K4L4L5K4L5K4L4L1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100H8A?A?B>A?A?B>C=O100O10000O10000O10000O10000O1000000000000000000000000000000000000001O00000000000000000000001O0000000000001O00000gFVF\\8j9\\GfH\\6Z7`IkH_6U7]IQIa6X:O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1E;N100O2O0O2O000O2O0O2O1N2O1O1N101N5L6kKmBP1Z=_N^CT1U`0A>CXY]4"}, "label": "the front most vase"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"widest glass vase\"."}], "task": "refering", "answer_template": "The \"widest glass vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 209, 210, 211, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 275, 276, 277, 278, 279, 280, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382], "bbox": [0.20675, 0.48410937499999995, 0.5273333333333333, 0.9842656250000001], "iscrowd": 0, "area": 35410.933899999996, "rle": {"size": [640, 480], "counts": "iZn1;_c0W1iNV1lNl0WO5K4L4L5K4L4M4K4L4L5K4L4L5K4L5K4L4L1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100H8A?A?B>A?A?B>C=O100O10000O10000O10000O10000O1000000000000000000000000000000000000001O00000000000000000000001O0000000000001O00000gFVF\\8j9\\GfH\\6Z7`IkH_6U7]IQIa6X:O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1E;N100O2O0O2O000O2O0O2O1N2O1O1N101N5L6kKmBP1Z=_N^CT1U`0A>CXY]4"}, "label": "widest glass vase"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest flower vase\"."}], "task": "refering", "answer_template": "The \"the smallest flower vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 282, 283, 284, 285, 299, 300, 301, 302, 303, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 350, 351, 352, 353, 354, 367, 368, 369, 370, 371, 384, 385, 386, 387], "bbox": [0.6202291666666666, 0.656734375, 0.8329583333333334, 0.987078125], "iscrowd": 0, "area": 19042.403250000007, "rle": {"size": [640, 480], "counts": "Qij5\\2X?]2N3M2N2N3M2O1N3M1O1O1O1O1O1O1O1O100O1O1O1O1RO]J^Cd5b<aJXC`5i<dJQC]5o<dJoB]5Q=eJlB\\5T=eJjB\\5V=fJgB[5Y=fJfBZ5Z=f000000000000000000000001O00000000000000000000000000000000001O000000000000000000000nIjBZ5W=cJPCX5Q=fJQCY5P=dJRC\\5o<bJSC]5n<aJSC_5h=O1O001O001N1O2O1N1O2N1O2N1O2N2N1O2N2N2N3M3M3N1NWaa1"}, "label": "the smallest flower vase"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray haired woman standing in the background\"."}], "task": "refering", "answer_template": "The \"a gray haired woman standing in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 17, 18, 19, 20, 34, 35, 36, 37, 51, 52, 53, 68, 69, 85, 86, 102, 103, 119, 120, 121, 136, 137, 138, 153, 154, 155], "bbox": [0.0, 0.014421875, 0.20222916666666665, 0.43165624999999996], "iscrowd": 0, "area": 15933.665049999998, "rle": {"size": [640, 480], "counts": "`0d1e1b2j:aMUE_2k:cMRE_2m:dMPE\\2o:lMiDU2W;TN`Dl1`;XN[Di1e;YNYDg1f;\\NWDe1i;]NSDe1m;]NPDe1o;`NkCa1T<W3O1O1O1N2O1O0000001O001O1O1O001N10000000001O00001OgIXDZ4h;fKYDY4g;hKZDW4e;iKTE^3l:cLYEW3g:kLZER3f:oL\\En2d:TM\\Ej2d:WM]Eh2c:TMcEi2]:WMgEf2X:[MmE_2T:bMPFY2o9hM]Fj1d9XNmFU1S9mN]G`0e8AaG7_8KdGO]83eGI\\87gGEY8=iG_OW8c0kGYOU8i0mGROU8o0nGmNQ8T1THfNl7\\1VH`Nk7a1VH]Ni7e1YHWNg7k1^HoMa7R2dHhM]7Y2gHbMX7`2lHZMU7g2mHVMR7h2TIUMl6f2Z5L5K6J4M3L4M2N3M3O1N1O2N2N1O3M2M4K4K6J5L5M3M2N4L6J[Z_7"}, "label": "a gray haired woman standing in the background"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old woman in black\"."}], "task": "refering", "answer_template": "The \"an old woman in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 17, 18, 19, 20, 34, 35, 36, 37, 51, 52, 53, 68, 69, 85, 86, 102, 103, 119, 120, 121, 136, 137, 138, 153, 154, 155], "bbox": [0.0, 0.014421875, 0.20222916666666665, 0.43165624999999996], "iscrowd": 0, "area": 15933.665049999998, "rle": {"size": [640, 480], "counts": "`0d1e1b2j:aMUE_2k:cMRE_2m:dMPE\\2o:lMiDU2W;TN`Dl1`;XN[Di1e;YNYDg1f;\\NWDe1i;]NSDe1m;]NPDe1o;`NkCa1T<W3O1O1O1N2O1O0000001O001O1O1O001N10000000001O00001OgIXDZ4h;fKYDY4g;hKZDW4e;iKTE^3l:cLYEW3g:kLZER3f:oL\\En2d:TM\\Ej2d:WM]Eh2c:TMcEi2]:WMgEf2X:[MmE_2T:bMPFY2o9hM]Fj1d9XNmFU1S9mN]G`0e8AaG7_8KdGO]83eGI\\87gGEY8=iG_OW8c0kGYOU8i0mGROU8o0nGmNQ8T1THfNl7\\1VH`Nk7a1VH]Ni7e1YHWNg7k1^HoMa7R2dHhM]7Y2gHbMX7`2lHZMU7g2mHVMR7h2TIUMl6f2Z5L5K6J4M3L4M2N3M3O1N1O2N2N1O3M2M4K4K6J5L5M3M2N4L6J[Z_7"}, "label": "an old woman in black"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower vase that is very narrow\"."}], "task": "refering", "answer_template": "The \"a flower vase that is very narrow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 178, 179, 180, 195, 196, 197, 212, 213, 229, 230, 246, 247, 263, 264, 280, 281, 282, 297, 298, 299, 314, 315, 316, 331, 332, 333], "bbox": [0.4683541666666667, 0.42609375, 0.6119791666666666, 0.8432031249999999], "iscrowd": 0, "area": 12628.55815, "rle": {"size": [640, 480], "counts": "Rn\\47ac0;k_OAb<b0\\B?a=F[A[1d>iNY@X2d?S1O2N1O2O0O2N2N2O0O2N2M2M4L4L3@mJPBW5o=kJlAY5o1QKa9G]D[5b1]KQ:Q5ZE^Kf:P5\\D`Kd;^6001O00000000000000000000001O000000000000000000001O00000lKeDC[;O[EIe:IQFOo9CgF5Z9]OZG<f8VOPHb0Q8oNeHi0[7jNYIo0j6`NlIX1f<H8I6I8H8H7I8HWgc3"}, "label": "a flower vase that is very narrow"}]}
{"id": 421059, "image": "COCO_train2014_000000421059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a narrow vase among all\"."}], "task": "refering", "answer_template": "The \"a narrow vase among all\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 178, 179, 180, 195, 196, 197, 212, 213, 229, 230, 246, 247, 263, 264, 280, 281, 282, 297, 298, 299, 314, 315, 316, 331, 332, 333], "bbox": [0.4683541666666667, 0.42609375, 0.6119791666666666, 0.8432031249999999], "iscrowd": 0, "area": 12628.55815, "rle": {"size": [640, 480], "counts": "Rn\\47ac0;k_OAb<b0\\B?a=F[A[1d>iNY@X2d?S1O2N1O2O0O2N2N2O0O2N2M2M4L4L3@mJPBW5o=kJlAY5o1QKa9G]D[5b1]KQ:Q5ZE^Kf:P5\\D`Kd;^6001O00000000000000000000001O000000000000000000001O00000lKeDC[;O[EIe:IQFOo9CgF5Z9]OZG<f8VOPHb0Q8oNeHi0[7jNYIo0j6`NlIX1f<H8I6I8H8H7I8HWgc3"}, "label": "a narrow vase among all"}]}
{"id": 412868, "image": "COCO_train2014_000000412868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tall giraffe near the top of the tree\"."}], "task": "refering", "answer_template": "The \"the tall giraffe near the top of the tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 43, 44, 45, 59, 60, 61, 76, 77, 93, 94, 109, 110, 111, 126, 127, 128, 142, 143, 144, 145, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 229, 245, 246, 262, 263, 279, 280], "bbox": [0.14316666666666666, 0.058374999999999996, 0.6592083333333333, 0.7228125000000001], "iscrowd": 0, "area": 31928.580300000005, "rle": {"size": [640, 480], "counts": "c^[17Z1KPa0?d^OFYa0d0[^O@ca0j0Q^OZOna0\\1N2N2N3M300O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O00100O1O100O10000O100O10000O100O100O10000O100O10000O100O10000O10O010O001000O100O10000O100O10000O2O2M2O1O2M2N2M2O0O01000O01000O10001N101N101O0O2O001N2O000O2O0O10000000001O01O0006J8H9G8H9G7H3L3N2N3L3N2N2M4M2N2M4M2N2M3N4L3L5L3[DlHZ:X7_EnH]:V7\\EPIb:T7WERIf:Q7SEVIj:R8L4M3M3L3NM3L3M4L4L3M4L4K4M4L4L3M4L4L3M4L4L3fMeCQM]<o2fCmL[<S3hCjLW<W3mCdLT<\\3oCaLQ<_3RD]Ln;d3VDWLk;i3XDTL]:fNSFV5CPL]:mNjET5MkKZ:VOcEo46gKZ:]O[El4>cKY:FREh4i0^KV:OkDb4S1[KU:6cD_4[1WKT:h5oEUJS:h5RFSJP:j5UFSJl9j5W2L3M4L3M4L3M3M4L3M4L3M3D=YOf0[Oe0[Of0A>1O01O001O00003M2N2N3N1N2N3M2N3M2N2N3M2N2N2O2M2O1N2O2M2O1O1N3N1N2O1N3N1Nf]V3"}, "label": "the tall giraffe near the top of the tree"}]}
{"id": 412868, "image": "COCO_train2014_000000412868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest zebra\"."}], "task": "refering", "answer_template": "The \"the tallest zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 43, 44, 45, 59, 60, 61, 76, 77, 93, 94, 109, 110, 111, 126, 127, 128, 142, 143, 144, 145, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 229, 245, 246, 262, 263, 279, 280], "bbox": [0.14316666666666666, 0.058374999999999996, 0.6592083333333333, 0.7228125000000001], "iscrowd": 0, "area": 31928.580300000005, "rle": {"size": [640, 480], "counts": "c^[17Z1KPa0?d^OFYa0d0[^O@ca0j0Q^OZOna0\\1N2N2N3M300O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O00100O1O100O10000O100O10000O100O100O10000O100O10000O100O10000O10O010O001000O100O10000O100O10000O2O2M2O1O2M2N2M2O0O01000O01000O10001N101N101O0O2O001N2O000O2O0O10000000001O01O0006J8H9G8H9G7H3L3N2N3L3N2N2M4M2N2M4M2N2M3N4L3L5L3[DlHZ:X7_EnH]:V7\\EPIb:T7WERIf:Q7SEVIj:R8L4M3M3L3NM3L3M4L4L3M4L4K4M4L4L3M4L4L3M4L4L3fMeCQM]<o2fCmL[<S3hCjLW<W3mCdLT<\\3oCaLQ<_3RD]Ln;d3VDWLk;i3XDTL]:fNSFV5CPL]:mNjET5MkKZ:VOcEo46gKZ:]O[El4>cKY:FREh4i0^KV:OkDb4S1[KU:6cD_4[1WKT:h5oEUJS:h5RFSJP:j5UFSJl9j5W2L3M4L3M4L3M3M4L3M4L3M3D=YOf0[Oe0[Of0A>1O01O001O00003M2N2N3N1N2N3M2N3M2N2N3M2N2N2O2M2O1N2O2M2O1O1N3N1N2O1N3N1Nf]V3"}, "label": "the tallest zebra"}]}
{"id": 543947, "image": "COCO_train2014_000000543947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair to the left side of the table\"."}], "task": "refering", "answer_template": "The \"a chair to the left side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 357, 358, 359, 374, 375, 376, 377, 378], "bbox": [0.0036666666666666666, 0.717890625, 0.26595833333333335, 0.9982656249999999], "iscrowd": 0, "area": 17276.670299999998, "rle": {"size": [640, 480], "counts": "Tg1k4T?1000000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O10mM`AfN`>Z1bAdN^>[1dAdN[>]1gAaNY>_1iA_NW>`1lA^NS>c1oA[NQ>e1QBYNo=f1TBXNk=i1WBUNi=k1YBSNg=l1\\BRNd=n1]BQNb=P2`BnM`=Q2cBmM]=S2eBkMZ=U2iBiMW=W2kBgMU=Y2mBeMS=Z2PCdMP=\\2RCbMn<^2SCaMn<]2UCaMk<_2WC_Mj<`2XC^Mh<a2[C]Mf<b2\\C\\Md<f2[CYMf<g2YCYMg<h2XCXMi<i2UCWMk<j2SCWMn<j2PCVMP=k2oBUMS=k2kBUMV=k2hBVMZ=h2fBWM\\=h2dBXM^=e2cBZM`=d2_B]Mb=b2^B]Me=a2[B_Mg=_2YBaMh=]2XBcMk=[2UBeMl=Z2TBeMo=W1o@QOR1HQ>T1QClNQ=S1oBmNS=P1nBoNT=P1lBPOV=m0jBSOY=j0hBVOY=i0gBVO\\=g0eBYO]=e0bB[O`=c0aB]Oa=`0`B_Ob=`0^B@d=9bBEa=OkBOf`0L3N_\\k6"}, "label": "a chair to the left side of the table"}]}
{"id": 543947, "image": "COCO_train2014_000000543947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty chair facing to the right\"."}], "task": "refering", "answer_template": "The \"an empty chair facing to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 357, 358, 359, 374, 375, 376, 377, 378], "bbox": [0.0036666666666666666, 0.717890625, 0.26595833333333335, 0.9982656249999999], "iscrowd": 0, "area": 17276.670299999998, "rle": {"size": [640, 480], "counts": "Tg1k4T?1000000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000O10mM`AfN`>Z1bAdN^>[1dAdN[>]1gAaNY>_1iA_NW>`1lA^NS>c1oA[NQ>e1QBYNo=f1TBXNk=i1WBUNi=k1YBSNg=l1\\BRNd=n1]BQNb=P2`BnM`=Q2cBmM]=S2eBkMZ=U2iBiMW=W2kBgMU=Y2mBeMS=Z2PCdMP=\\2RCbMn<^2SCaMn<]2UCaMk<_2WC_Mj<`2XC^Mh<a2[C]Mf<b2\\C\\Md<f2[CYMf<g2YCYMg<h2XCXMi<i2UCWMk<j2SCWMn<j2PCVMP=k2oBUMS=k2kBUMV=k2hBVMZ=h2fBWM\\=h2dBXM^=e2cBZM`=d2_B]Mb=b2^B]Me=a2[B_Mg=_2YBaMh=]2XBcMk=[2UBeMl=Z2TBeMo=W1o@QOR1HQ>T1QClNQ=S1oBmNS=P1nBoNT=P1lBPOV=m0jBSOY=j0hBVOY=i0gBVO\\=g0eBYO]=e0bB[O`=c0aB]Oa=`0`B_Ob=`0^B@d=9bBEa=OkBOf`0L3N_\\k6"}, "label": "an empty chair facing to the right"}]}
{"id": 543947, "image": "COCO_train2014_000000543947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a black chair with its back towards the camera\"."}], "task": "refering", "answer_template": "The \"this is a black chair with its back towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 295, 296, 297, 298, 299, 300, 301, 302, 312, 313, 314, 315, 316, 317, 318, 319, 329, 330, 331, 332, 333, 334, 335, 336, 346, 348, 349, 350, 351, 352, 353, 363, 366, 367, 369, 370, 380, 383, 384, 386, 387], "bbox": [0.3595416666666667, 0.60590625, 0.7959583333333333, 1.0], "iscrowd": 0, "area": 29269.060899999993, "rle": {"size": [640, 480], "counts": "lc\\3^2aa0[1dN3N1N2O1O1N2O1000000O10001N1000000O10`N`1aN^1101O000O101O0O101O000O2O1O0O2O001O0O2O001N2O001O0O2O001N101O1N101O001N101O0O2O001O1N101O0O2O001O0O2O1O0O2O001O0O2O001N2O0C>lNS1O2O0O2N1O8H;E;E;E;E;F1N1O101N100O1O100O1O2O0O100O1O101O0O1000E:D=D<C<K60O01000O10O10O10O01000O010000O01000O010000O01000O010000O010000O01000O0100000O01000O01000O10O1000O01000O10O10O100000O0100005K8H7I8H8H8H7I8H8H7I8H1O01O1O001O001N1[KkE_OV:4gF[OY99`GZOa89QHAo73cHG^7LTINn6DeI5_6[OSJ?o<H]dl1"}, "label": "this is a black chair with its back towards the camera"}]}
{"id": 519376, "image": "COCO_train2014_000000519376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange that is most fully visible\"."}], "task": "refering", "answer_template": "The \"the orange that is most fully visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191], "bbox": [0.36982, 0.5435435435435435, 0.6377999999999999, 0.8988288288288289], "iscrowd": 0, "area": 12917.420750000001, "rle": {"size": [333, 500], "counts": "P]l14R::G9G9F:G9H8L4K4M2N2M3N2N2M3N2N2M3N2N2N2O1N2O1N2O1N100O101N100O101N100O100O2O0O10001O00000O10000000000O100000000O1000000000000O10000000000000000000000000000000000O10000000O10000000001O1O1O1O1O001O1O1N2O1N101O1N2O1N2O001N2O1N2O1O0O2N2N3M3M3M3M3M3M3N2M3M3M3M3K5L:E>C=Bacj1"}, "label": "the orange that is most fully visible"}]}
{"id": 519376, "image": "COCO_train2014_000000519376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange next to a rubber fish\"."}], "task": "refering", "answer_template": "The \"an orange next to a rubber fish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191], "bbox": [0.36982, 0.5435435435435435, 0.6377999999999999, 0.8988288288288289], "iscrowd": 0, "area": 12917.420750000001, "rle": {"size": [333, 500], "counts": "P]l14R::G9G9F:G9H8L4K4M2N2M3N2N2M3N2N2M3N2N2N2O1N2O1N2O1N100O101N100O101N100O100O2O0O10001O00000O10000000000O100000000O1000000000000O10000000000000000000000000000000000O10000000O10000000001O1O1O1O1O001O1O1N2O1N101O1N2O1N2O001N2O1N2O1O0O2N2N3M3M3M3M3M3M3N2M3M3M3M3K5L:E>C=Bacj1"}, "label": "an orange next to a rubber fish"}]}
{"id": 60624, "image": "COCO_train2014_000000060624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the screen\"."}], "task": "refering", "answer_template": "The \"the man on the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 147, 148, 149, 150, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.297515625, 0.2022663551401869, 0.617375, 0.8277570093457943], "iscrowd": 0, "area": 33253.8068, "rle": {"size": [428, 640], "counts": "eg_21o<=I6K6UOk0J6I6K6I7J6J5L5N2N2N2N1O101N1O1O2N1O100O2N1O1O2N1O101N2N2N2N2N100O1O1O1O1O1O100O1O1N2O1O1O1N2O1O1O1N2O1O1O10000O10000O10000O100O10000O10000O10000O1000000O10000O1000000O1000000O1ZNRKnJn4i4^KTKb4j4cKRK^4l4gKQKY4m4lKPKT4n4QLoJo3o4VLmJk3Q5[LkJe3S5`LhJb3V5dLdJ^3Z5gLaJ[3]5kL]JW3b5Z2N1O2N2N200O100O1O100O100O1O100O10000O100O10000O100O100O10000O100O10000000000001gJoJU2Q5kMUKo1k4PNZKl1f4SN^Kj1b4VNaKg1`4WNcKg1]4XNgKe1Y4[NjKa1W4^NlK`1U4^NnK`1R4`NPL^1P4aNSL]1m3cNUL[1k3dNXLZ1i3dNZLZ1f3fN\\LX1d3gN`LV1`3iNcLU1]3kNeLS1\\3kNgLS1Y3mNiLQ1W3nNlLP1T3oNoLo0R3oNQMo0o2POTMn0m2QOVMl0l2QOWMm0k2POXMn0i2POZMn0h2oN[Mo0g2nN\\MP1e2lN`MR1b2jNcMS1_2jNdMT1^2iNeMU1]2hNfMV1\\2gNgMW1[2fNiMW1Y2fNjMX1\\2aNgM]1Y2bNjM\\1W2bNlM\\1U2aNoM]1R2]NVN`1j1\\N\\Nb1i6N2N2N2N3M2N2N2N2N2M4M2N2N3_Oa0FoRV3"}, "label": "the man on the screen"}]}
{"id": 60624, "image": "COCO_train2014_000000060624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheldon from the big bang theory holding a package on a television\"."}], "task": "refering", "answer_template": "The \"sheldon from the big bang theory holding a package on a television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 147, 148, 149, 150, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.297515625, 0.2022663551401869, 0.617375, 0.8277570093457943], "iscrowd": 0, "area": 33253.8068, "rle": {"size": [428, 640], "counts": "eg_21o<=I6K6UOk0J6I6K6I7J6J5L5N2N2N2N1O101N1O1O2N1O100O2N1O1O2N1O101N2N2N2N2N100O1O1O1O1O1O100O1O1N2O1O1O1N2O1O1O1N2O1O1O10000O10000O10000O100O10000O10000O10000O1000000O10000O1000000O1000000O1ZNRKnJn4i4^KTKb4j4cKRK^4l4gKQKY4m4lKPKT4n4QLoJo3o4VLmJk3Q5[LkJe3S5`LhJb3V5dLdJ^3Z5gLaJ[3]5kL]JW3b5Z2N1O2N2N200O100O1O100O100O1O100O10000O100O10000O100O100O10000O100O10000000000001gJoJU2Q5kMUKo1k4PNZKl1f4SN^Kj1b4VNaKg1`4WNcKg1]4XNgKe1Y4[NjKa1W4^NlK`1U4^NnK`1R4`NPL^1P4aNSL]1m3cNUL[1k3dNXLZ1i3dNZLZ1f3fN\\LX1d3gN`LV1`3iNcLU1]3kNeLS1\\3kNgLS1Y3mNiLQ1W3nNlLP1T3oNoLo0R3oNQMo0o2POTMn0m2QOVMl0l2QOWMm0k2POXMn0i2POZMn0h2oN[Mo0g2nN\\MP1e2lN`MR1b2jNcMS1_2jNdMT1^2iNeMU1]2hNfMV1\\2gNgMW1[2fNiMW1Y2fNjMX1\\2aNgM]1Y2bNjM\\1W2bNlM\\1U2aNoM]1R2]NVN`1j1\\N\\Nb1i6N2N2N2N3M2N2N2N2N2M4M2N2N3_Oa0FoRV3"}, "label": "sheldon from the big bang theory holding a package on a television"}]}
{"id": 60624, "image": "COCO_train2014_000000060624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hand holding remote\"."}], "task": "refering", "answer_template": "The \"hand holding remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 267, 271, 272, 273, 274, 275, 290, 296, 297, 298, 312, 313, 314, 319, 320, 321, 336, 337, 342, 343, 344], "bbox": [0.601421875, 0.38724299065420564, 1.0, 0.9890420560747664], "iscrowd": 0, "area": 36420.6494, "rle": {"size": [428, 640], "counts": "oXQ5a0j<5K4L5L4K2N2N1O1ZOhNXEY1b:nN\\ER1^:VO_Ek0]:ZObEg0_:YO_Eg0b:ZO\\Ef0e:[OYEf0h:XOXEh0i:XOUEi0l:WOSEj0n:UOQEk0P;UOoDk0R;UOmDl0T;SOkDT1Q;iNnDZ1S;cNmD^1U;_NkDc1V;83N1N3M2N2N3M2N3M2N3M2N2N3M2N3M2N3M2N2N2N2N2N2N2N1O2NiIc0aN[OZ1T1^NjN\\1c1_N[N\\1R2^NlMa1\\2ZNbMd1_2d4002M3N2M3N2N3M2O1N2O00000O10O1000O1000O01000O10O01O1O1O00100O1O001001O1N2O1O001N2O1O001O1O1N2O001O1O1N3N1O2N1O2N2M2O2N1O2N2N1N3N1O1O1O1O0O2O1O001O1O1O001N2O001O1O1O001O1O0O2O1O1O001O1O001N2O001O1O001O1O001O1O1O0O2O1O001O1O001O1O002N3N2M2N3N2M6J8H8XIlIi5S7H7I2N01O00001O00000000001O0000000000001O00000000001O000O1000001O01O0000000001O000000001O000000001O00000001O01O000000001O0000000_MjJlLW5h2lK`LU4Z3[L[Le3b3fLVL[3g3kLULU3h3PMVLQ3g3SMWLm2f3fM"}, "label": "hand holding remote"}]}
{"id": 60624, "image": "COCO_train2014_000000060624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding a remote\"."}], "task": "refering", "answer_template": "The \"a hand holding a remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 267, 271, 272, 273, 274, 275, 290, 296, 297, 298, 312, 313, 314, 319, 320, 321, 336, 337, 342, 343, 344], "bbox": [0.601421875, 0.38724299065420564, 1.0, 0.9890420560747664], "iscrowd": 0, "area": 36420.6494, "rle": {"size": [428, 640], "counts": "oXQ5a0j<5K4L5L4K2N2N1O1ZOhNXEY1b:nN\\ER1^:VO_Ek0]:ZObEg0_:YO_Eg0b:ZO\\Ef0e:[OYEf0h:XOXEh0i:XOUEi0l:WOSEj0n:UOQEk0P;UOoDk0R;UOmDl0T;SOkDT1Q;iNnDZ1S;cNmD^1U;_NkDc1V;83N1N3M2N2N3M2N3M2N3M2N2N3M2N3M2N3M2N2N2N2N2N2N2N1O2NiIc0aN[OZ1T1^NjN\\1c1_N[N\\1R2^NlMa1\\2ZNbMd1_2d4002M3N2M3N2N3M2O1N2O00000O10O1000O1000O01000O10O01O1O1O00100O1O001001O1N2O1O001N2O1O001O1O1N2O001O1O1N3N1O2N1O2N2M2O2N1O2N2N1N3N1O1O1O1O0O2O1O001O1O1O001N2O001O1O1O001O1O0O2O1O1O001O1O001N2O001O1O001O1O001O1O1O0O2O1O001O1O001O1O002N3N2M2N3N2M6J8H8XIlIi5S7H7I2N01O00001O00000000001O0000000000001O00000000001O000O1000001O01O0000000001O000000001O000000001O00000001O01O000000001O0000000_MjJlLW5h2lK`LU4Z3[L[Le3b3fLVL[3g3kLULU3h3PMVLQ3g3SMWLm2f3fM"}, "label": "a hand holding a remote"}]}
{"id": 371923, "image": "COCO_train2014_000000371923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skiing wearing a red ski outfit\"."}], "task": "refering", "answer_template": "The \"a man skiing wearing a red ski outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 245, 246, 249, 250, 260, 261, 262, 264, 265, 266, 276, 277, 280, 281, 282, 291, 292, 296, 297, 307, 308, 312, 313, 322, 323, 327, 328, 329, 343, 344], "bbox": [0.25379391100702575, 0.36646875, 0.9593911007025762, 0.9678281249999999], "iscrowd": 0, "area": 35642.610399999976, "rle": {"size": [640, 427], "counts": "[]T28cc08K5K4M4L3M3M4L3M3N3L3N1O0O2O001O0QO`N^_Oa1``0hNW_OY1g`0oNP_OT1o`0oNl^OT1Ra0n0N2O1O1O100O100O100O101N2O1N101O1O1O001O2N1O1O1O1gAmLc;T3TDXMi;i2RD^Mj;c2RDcMk;_2kCXNf;i1lChNP<Z1iCmNU<U1fCPOX<R1cCSO[<n0aCWO]<k0_CYO_<X4L5K6J5L4L9G0O10001O0001O0nGbDa7_;ZHeDg7Z;XHfDj7Z;SHhDm7Z;oGjDP8b;2N2N1O2N1O2M2O2N1OkMeDgKY;n3SERLl:l3WEULEmMi:k5jETL[OTNi:f5oEVLVOWNP1Ma8d5]GcLOlMb8^5cGeLJnMb8\\5hGdLEQNc8Y5kGeLASNd8W5nGcL_OVNc8U5RHcLZOXNe8S5THdLVOZNf8P5XHkLjNUNo8n4ZHSM^NPNY9j4^HlMa7R2cHmM]7P2jHlMV7Q2TIhMk6V2^IdMb6Z2cIdM\\6[2iIaMV6^2oI`MP6^2UJ_Mk5`2YJ^Me5a2`J\\Ma5a2cJ]M]5b2gJ[MZ5d2hJZMX5f2kJWMV5h2lJVMT5k2mJSMT5m2nJPMR5Q3oJmLR5S3oJkLQ5U3RKhLo4X3QKgLo4Z3RKdLn4]3SKaLm4`3SK^Ln4c3SKZLn4f3TKXLl4i3UKTLl4m3UKPLl4Q4UKlKl4U4UKiKk4Y4VKcKk4^4ZK[Kg4f4Y5]Ln@Y2Q?dMWAX2h>gM^AU2a>kMaAU2]>jMfAU2Y>jMjAU2U>jMnAU2Q>jMRBU2m=kMVBS2i=mMYBS2e=mM]BR2b=nMaBP2^=PNdBP2Z=PNgBP2X=oMkBQ2S=oMnBQ2Q=oMPCQ2o<oMRCR2l<nMVBSO9o2`=nMVBTO=m2[=oMWBUO`0k2X=oMYBVOb0i2U=PNXBXOe0h2Q=PNYBYOh0f2n<QNZBYOk0e2i<RN[BZOn0c2g<RN[B[OP1b2d<RN\\B^OR1_2`<UN\\B\\OV1^2^<WNYB]OZ1[2\\<ZNXB[O_1Z2W<]NXBZOb1X2U<`NWBYOe1W2S<QOXB\\N`1c2W<ROXB\\Nb1a2U<SOYB]Nc1`2R<SOZB_Ne1]2Q<[OQDe0m;[OTDe0k;[OWDd0i;[OYDe0e;[O]Dd0c;ZOaDd0_;\\ObDd0\\;]OeDb0[;]OgDb0X;^OkD`0U;_OmD`0V;[OmDd0V;XOlDg0W;UOlDi0W;QOmDo0V;lNmDQ1X;jNkDT1Y;hNhDW1\\;dNgDZ1Z;dNhD[1Y;cNjDZ1W;dNmD[1R;dNPE[1o:aNYE\\1g:cN\\EZ1e:eN_EX1`:hNdEU1\\:jNgEU1W:jNkEW1T:gNoEY1o9gNRFZ1m9eNUF[1i9eNYF[1e9dN]F]1a9cNaF]1^9bNcF^1\\9bNfF]1o=010O000O10001O0000000O2O00001O00001O0000010O01O010O0010O01O010O1O2O1N1O2O001O001O0000O100O2O0N2N2N2O10O3L3M3M1O1O100O1O1O1O1O1O1000000001O001O1O1O001O1OO1O1O1O1O2N1O1[Oo\\O:Qc0DR]O:Pc0CR]O<ob0@U]O?Wc0O1O001N2N2M3L3NQQ:"}, "label": "a man skiing wearing a red ski outfit"}]}
{"id": 371923, "image": "COCO_train2014_000000371923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skiing wearing a red ski suit\"."}], "task": "refering", "answer_template": "The \"a man skiing wearing a red ski suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 245, 246, 249, 250, 260, 261, 262, 264, 265, 266, 276, 277, 280, 281, 282, 291, 292, 296, 297, 307, 308, 312, 313, 322, 323, 327, 328, 329, 343, 344], "bbox": [0.25379391100702575, 0.36646875, 0.9593911007025762, 0.9678281249999999], "iscrowd": 0, "area": 35642.610399999976, "rle": {"size": [640, 427], "counts": "[]T28cc08K5K4M4L3M3M4L3M3N3L3N1O0O2O001O0QO`N^_Oa1``0hNW_OY1g`0oNP_OT1o`0oNl^OT1Ra0n0N2O1O1O100O100O100O101N2O1N101O1O1O001O2N1O1O1O1gAmLc;T3TDXMi;i2RD^Mj;c2RDcMk;_2kCXNf;i1lChNP<Z1iCmNU<U1fCPOX<R1cCSO[<n0aCWO]<k0_CYO_<X4L5K6J5L4L9G0O10001O0001O0nGbDa7_;ZHeDg7Z;XHfDj7Z;SHhDm7Z;oGjDP8b;2N2N1O2N1O2M2O2N1OkMeDgKY;n3SERLl:l3WEULEmMi:k5jETL[OTNi:f5oEVLVOWNP1Ma8d5]GcLOlMb8^5cGeLJnMb8\\5hGdLEQNc8Y5kGeLASNd8W5nGcL_OVNc8U5RHcLZOXNe8S5THdLVOZNf8P5XHkLjNUNo8n4ZHSM^NPNY9j4^HlMa7R2cHmM]7P2jHlMV7Q2TIhMk6V2^IdMb6Z2cIdM\\6[2iIaMV6^2oI`MP6^2UJ_Mk5`2YJ^Me5a2`J\\Ma5a2cJ]M]5b2gJ[MZ5d2hJZMX5f2kJWMV5h2lJVMT5k2mJSMT5m2nJPMR5Q3oJmLR5S3oJkLQ5U3RKhLo4X3QKgLo4Z3RKdLn4]3SKaLm4`3SK^Ln4c3SKZLn4f3TKXLl4i3UKTLl4m3UKPLl4Q4UKlKl4U4UKiKk4Y4VKcKk4^4ZK[Kg4f4Y5]Ln@Y2Q?dMWAX2h>gM^AU2a>kMaAU2]>jMfAU2Y>jMjAU2U>jMnAU2Q>jMRBU2m=kMVBS2i=mMYBS2e=mM]BR2b=nMaBP2^=PNdBP2Z=PNgBP2X=oMkBQ2S=oMnBQ2Q=oMPCQ2o<oMRCR2l<nMVBSO9o2`=nMVBTO=m2[=oMWBUO`0k2X=oMYBVOb0i2U=PNXBXOe0h2Q=PNYBYOh0f2n<QNZBYOk0e2i<RN[BZOn0c2g<RN[B[OP1b2d<RN\\B^OR1_2`<UN\\B\\OV1^2^<WNYB]OZ1[2\\<ZNXB[O_1Z2W<]NXBZOb1X2U<`NWBYOe1W2S<QOXB\\N`1c2W<ROXB\\Nb1a2U<SOYB]Nc1`2R<SOZB_Ne1]2Q<[OQDe0m;[OTDe0k;[OWDd0i;[OYDe0e;[O]Dd0c;ZOaDd0_;\\ObDd0\\;]OeDb0[;]OgDb0X;^OkD`0U;_OmD`0V;[OmDd0V;XOlDg0W;UOlDi0W;QOmDo0V;lNmDQ1X;jNkDT1Y;hNhDW1\\;dNgDZ1Z;dNhD[1Y;cNjDZ1W;dNmD[1R;dNPE[1o:aNYE\\1g:cN\\EZ1e:eN_EX1`:hNdEU1\\:jNgEU1W:jNkEW1T:gNoEY1o9gNRFZ1m9eNUF[1i9eNYF[1e9dN]F]1a9cNaF]1^9bNcF^1\\9bNfF]1o=010O000O10001O0000000O2O00001O00001O0000010O01O010O0010O01O010O1O2O1N1O2O001O001O0000O100O2O0N2N2N2O10O3L3M3M1O1O100O1O1O1O1O1O1000000001O001O1O1O001O1OO1O1O1O1O2N1O1[Oo\\O:Qc0DR]O:Pc0CR]O<ob0@U]O?Wc0O1O001N2N2M3L3NQQ:"}, "label": "a man skiing wearing a red ski suit"}]}
{"id": 371923, "image": "COCO_train2014_000000371923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white pants skiing\"."}], "task": "refering", "answer_template": "The \"a man in white pants skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 80, 81, 82, 95, 96, 97, 110, 111, 112, 113, 126, 127, 128, 141, 143], "bbox": [0.19695550351288055, 0.04990625, 0.6039812646370023, 0.419359375], "iscrowd": 0, "area": 14621.689550000001, "rle": {"size": [640, 427], "counts": "fcd12ic09I5M1N2N2N2O1N2O1N200O1O101N10O0001N10000O2O000001O1O102M2N1@P]OORc00?01N2Og\\OJhb06R]OOPc0`000LYOo\\Og0lb0_OS]O`0lb0BT]O>lb0CS]O=lb0<0O12M3iNmN]_OT1a`0SOY_Oo0c`0VOi^OA8\\1l`0VOj^OB6[1l`0UOl^OC5[1l`0_OS_Oc0k`0^OS_Oe0j`0]OT_Of0j`0[OU_Og0i`0YOV_Oi0i`0Y1M3N1O2N2O1N2N2N2O1N=C;E6J4M3L4L4N1O00001O1O1OO2N:FO1N2O2N2N1O2N1]AkJX>a5N1O1O1O2N1O2N1WBUJ[=m5bBVJ]=j5`BYJ_=i5]B\\J`=R61O0`MkBUNV=k1mBQNS=Q2oBkMQ=V2RCfMo<Z2UCaMk<`2XC\\Mi<e2YCWMg<j2^CYLSO4`=c3SDYLm;i3UDRLm;S1ZB^1l1[Mj;X1bBU1`?kNi@m0U?TOSAd0m>\\OSAd0l>]OUAc0j>[O[Ab0e>]O^Aa0a>^OcAa0\\>]OhAa0W>^OmA`0R>_OQBa0n=]OTBc0k=\\OXBc0g=]O[Bb0d=^O^Ba0b=]OaBb0^=^OdBa0[=^OhBa0W=_OkB`0U=_OmBd0n<[OVCd0h<\\OZCc0f<[O]Cd0b<\\O`Cc0`<[OdCd0Z<\\OhCc0X<[OkCd0T<\\OoCb0Q<\\OTDb0j;^OYD`0g;^O^D`0a;_OcD?\\;_OhDb0U;\\OoDe0n:ZOTEe0l:YOWEf0i:YOXEh0m:POVEo0n>01O0010O000000O2N1O2N3M2N2N3M2N3M3M2O2M3MeoY3"}, "label": "a man in white pants skiing"}]}
{"id": 371923, "image": "COCO_train2014_000000371923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skier in all white\"."}], "task": "refering", "answer_template": "The \"skier in all white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 80, 81, 82, 95, 96, 97, 110, 111, 112, 113, 126, 127, 128, 141, 143], "bbox": [0.19695550351288055, 0.04990625, 0.6039812646370023, 0.419359375], "iscrowd": 0, "area": 14621.689550000001, "rle": {"size": [640, 427], "counts": "fcd12ic09I5M1N2N2N2O1N2O1N200O1O101N10O0001N10000O2O000001O1O102M2N1@P]OORc00?01N2Og\\OJhb06R]OOPc0`000LYOo\\Og0lb0_OS]O`0lb0BT]O>lb0CS]O=lb0<0O12M3iNmN]_OT1a`0SOY_Oo0c`0VOi^OA8\\1l`0VOj^OB6[1l`0UOl^OC5[1l`0_OS_Oc0k`0^OS_Oe0j`0]OT_Of0j`0[OU_Og0i`0YOV_Oi0i`0Y1M3N1O2N2O1N2N2N2O1N=C;E6J4M3L4L4N1O00001O1O1OO2N:FO1N2O2N2N1O2N1]AkJX>a5N1O1O1O2N1O2N1WBUJ[=m5bBVJ]=j5`BYJ_=i5]B\\J`=R61O0`MkBUNV=k1mBQNS=Q2oBkMQ=V2RCfMo<Z2UCaMk<`2XC\\Mi<e2YCWMg<j2^CYLSO4`=c3SDYLm;i3UDRLm;S1ZB^1l1[Mj;X1bBU1`?kNi@m0U?TOSAd0m>\\OSAd0l>]OUAc0j>[O[Ab0e>]O^Aa0a>^OcAa0\\>]OhAa0W>^OmA`0R>_OQBa0n=]OTBc0k=\\OXBc0g=]O[Bb0d=^O^Ba0b=]OaBb0^=^OdBa0[=^OhBa0W=_OkB`0U=_OmBd0n<[OVCd0h<\\OZCc0f<[O]Cd0b<\\O`Cc0`<[OdCd0Z<\\OhCc0X<[OkCd0T<\\OoCb0Q<\\OTDb0j;^OYD`0g;^O^D`0a;_OcD?\\;_OhDb0U;\\OoDe0n:ZOTEe0l:YOWEf0i:YOXEh0m:POVEo0n>01O0010O000000O2N1O2N3M2N2N3M2N3M3M2O2M3MeoY3"}, "label": "skier in all white"}]}
{"id": 339159, "image": "COCO_train2014_000000339159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a empty brown chair sitting at a table\"."}], "task": "refering", "answer_template": "The \"a empty brown chair sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 184, 185, 207, 230, 231, 253, 254, 255, 276, 277, 278], "bbox": [0.001921875, 0.15233333333333335, 0.20821874999999998, 0.7667291666666667], "iscrowd": 0, "area": 19851.089, "rle": {"size": [480, 640], "counts": "Sj0=V=]1cN\\1eN\\1cN\\1eN\\1hNX1O010000O1000O010000O10000O01000O100O1`MbI^L]6[3mIcLS6U3VJiLk5P3^JnLb5j2gJUMY5h2kJVMV5h2nJVMR5i2PKVMo4i2UKTMl4k2VKTMj4j2ZKTMf4k2\\KSMe4k2^KTMb4k2aKSM_4k2dKSM\\4m2gKQMY4m2jKRMV4m2lKQMU4m2oKQMQ4n2QLPMP4n2SLQMl3o2WLoLi3o2ZLoLg3P3[LoLe3o2^LPMb3o2aLnL`3P3cLoL]3P3eLoLZ3P3jLmLW3R3kLmLU3Q3nLnLR3Q3PMmLQ3Q3SMmLm2R3UMmLj2R3YMlLh2T3YMkLg2T3\\MjLd2V3]MhLd2W3^MhLb2X3`MeLa2Z3aMeL^2\\3cMcL]2\\3eMbL\\2^3fM`LZ2`3gM_LY2`3d40O0100O100O100O010O100O100O010O100O100O10O010000O100O010O100O100O10O0100O100O10O0100O100O1000O0100O1lNSDUOm;j0WDROj;m0[DoNd;Q1`DkNa;T1dDhN_;T1eDhNd;o0aDlNi;j0[DROo;d0UDXOT<?`1FZa]7"}, "label": "a empty brown chair sitting at a table"}]}
{"id": 339159, "image": "COCO_train2014_000000339159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair to the left of pizza\"."}], "task": "refering", "answer_template": "The \"a chair to the left of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 184, 185, 207, 230, 231, 253, 254, 255, 276, 277, 278], "bbox": [0.001921875, 0.15233333333333335, 0.20821874999999998, 0.7667291666666667], "iscrowd": 0, "area": 19851.089, "rle": {"size": [480, 640], "counts": "Sj0=V=]1cN\\1eN\\1cN\\1eN\\1hNX1O010000O1000O010000O10000O01000O100O1`MbI^L]6[3mIcLS6U3VJiLk5P3^JnLb5j2gJUMY5h2kJVMV5h2nJVMR5i2PKVMo4i2UKTMl4k2VKTMj4j2ZKTMf4k2\\KSMe4k2^KTMb4k2aKSM_4k2dKSM\\4m2gKQMY4m2jKRMV4m2lKQMU4m2oKQMQ4n2QLPMP4n2SLQMl3o2WLoLi3o2ZLoLg3P3[LoLe3o2^LPMb3o2aLnL`3P3cLoL]3P3eLoLZ3P3jLmLW3R3kLmLU3Q3nLnLR3Q3PMmLQ3Q3SMmLm2R3UMmLj2R3YMlLh2T3YMkLg2T3\\MjLd2V3]MhLd2W3^MhLb2X3`MeLa2Z3aMeL^2\\3cMcL]2\\3eMbL\\2^3fM`LZ2`3gM_LY2`3d40O0100O100O100O010O100O100O010O100O100O10O010000O100O010O100O100O10O0100O100O10O0100O100O1000O0100O1lNSDUOm;j0WDROj;m0[DoNd;Q1`DkNa;T1dDhN_;T1eDhNd;o0aDlNi;j0[DROo;d0UDXOT<?`1FZa]7"}, "label": "a chair to the left of pizza"}]}
{"id": 281815, "image": "COCO_train2014_000000281815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white keyboard\"."}], "task": "refering", "answer_template": "The \"white keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.30965625, 0.8049414519906323, 0.8137968749999999, 0.9865573770491803], "iscrowd": 0, "area": 20035.8852, "rle": {"size": [427, 640], "counts": "_mb25V=?A>B?@8I00000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O10000000O1000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000000000000000001O2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2M3N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2NWda1"}, "label": "white keyboard"}]}
{"id": 281815, "image": "COCO_train2014_000000281815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white keyboard\"."}], "task": "refering", "answer_template": "The \"a white keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.30965625, 0.8049414519906323, 0.8137968749999999, 0.9865573770491803], "iscrowd": 0, "area": 20035.8852, "rle": {"size": [427, 640], "counts": "_mb25V=?A>B?@8I00000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O10000000O1000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000000000000000001O2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2M3N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2NWda1"}, "label": "a white keyboard"}]}
{"id": 52448, "image": "COCO_train2014_000000052448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an oven baked pizza with cheese , spinach , and tomatoes\"."}], "task": "refering", "answer_template": "The \"an oven baked pizza with cheese , spinach , and tomatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 218, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 327, 328, 329, 330, 352, 353], "bbox": [0.13482812500000002, 0.0468125, 0.812359375, 0.9187291666666667], "iscrowd": 0, "area": 102052.64255000002, "rle": {"size": [480, 640], "counts": "l`X12j>9G8H9G8H9G9G8G:TDjM`:^2[EeM_:d2[E_M_:i2\\EZM^:o2\\ETM^:T3]EoL]:n3J6J5K6J7I6J6J6J7K4M3M3M4L3M3M4L3M3M3M4L3M3M3M4L3M3M3M4M2M3M4L3M3M2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2O0O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2O0O2N2N1O2N1O2N2O001N2O0O2O1O0O2O001N2O001N101N101O1N101O0O2O0O2O1O0O2O001N10001N100O2O000O2O001N100O2O000O2O000O2O000O2O0O2O000O2O1ON101N1O2N2M2O2M3N1O2M3N1N3N1O2M3N1N3N2N1N3N1N3N2N2M3N1N3N2N2M2O2M3N2M2O2N2M3N2M2O2N2M3N1N3N2N2M2O2M3N2N1N3N2M3N2N1N3N2M3N1O2M3N2M2O2M3N2N2M2O2M3N2N1N3N2M3N1O3L3N2M3N3M2M3N2M4M2N2N3N1O1O1O2N1O1O1N3N1O1O2N1O1O1O2N1N2O1O2O00010O0001O01O01O00010O00010O000010O00010O0001O01O01O01O01O00010O000010O00010O01O00010O00010O000010O00010O0001O01O01O01O01O00010O00010O000010O00100O00100O010000O010000O010000O01000O01000O10O10O1000O010O1O101N1O1O1O100O1O2N100O1O1O1O2O0O1O1O1O101N1O1O1O1O2N1O1O2O0O1O2N1O1O2N1O1O2O0O1O2N1O1O2N1O2O0O1O2N1O2N2N2N2N2O1N2N2N2N2N2N1O2N2O1N2N2L4L4M3L4M3L4UN\\D4R^h1"}, "label": "an oven baked pizza with cheese , spinach , and tomatoes"}]}
{"id": 52448, "image": "COCO_train2014_000000052448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two slices of pizza , slightly pulled out from the rest of the pizza\"."}], "task": "refering", "answer_template": "The \"two slices of pizza , slightly pulled out from the rest of the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 330, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380], "bbox": [0.367421875, 0.42695833333333333, 0.841015625, 0.9595416666666666], "iscrowd": 0, "area": 46101.7085, "rle": {"size": [480, 640], "counts": "kb^31l>3N3M3L4M2N3L4M3M2M4M3M2M4M3M3L3N3M2M3N2N2M3N3M2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N3L3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N3L3N2N2M3N2N2M3N2N2M3N2M3N2N3L3N2N2M3N2N2M3N2N2M3N2N2M3N3M2M2O2N2M3O1O0O2O1O1O1O1N101O1O1O1N101O1O1O10O2O00001O0O101O00001O0O2O000O2O000O2O0O101O0O101O0O2O0O2O000O2O001N100O2O001N101O0O101N101N100O2O0O2N101N100O2O0O2N100O2O0O2N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2N101N101N101N101N101N101N101N101N1O2O0O2N1O2O0O2N1O2O0O1O1O2O0O1O1O100O1O1O100O1O1O1O2M2O1O1N2O1O1N2O1O1N2O1N2O1O2M2O1O1N2O1O1N2O1O1N2O2N1N2O1O1O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1N2O2M2N2N2N3M4K4M3M3M3M4L3M3L4M3M4L3M3M3M3L4L\\a_1"}, "label": "two slices of pizza , slightly pulled out from the rest of the pizza"}]}
{"id": 52448, "image": "COCO_train2014_000000052448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom right side of a pizza\"."}], "task": "refering", "answer_template": "The \"the bottom right side of a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 330, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380], "bbox": [0.367421875, 0.42695833333333333, 0.841015625, 0.9595416666666666], "iscrowd": 0, "area": 46101.7085, "rle": {"size": [480, 640], "counts": "kb^31l>3N3M3L4M2N3L4M3M2M4M3M2M4M3M3L3N3M2M3N2N2M3N3M2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N3L3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N3L3N2N2M3N2N2M3N2N2M3N2M3N2N3L3N2N2M3N2N2M3N2N2M3N2N2M3N3M2M2O2N2M3O1O0O2O1O1O1O1N101O1O1O1N101O1O1O10O2O00001O0O101O00001O0O2O000O2O000O2O0O101O0O101O0O2O0O2O000O2O001N100O2O001N101O0O101N101N100O2O0O2N101N100O2O0O2N100O2O0O2N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2N101N101N101N101N101N101N101N101N1O2O0O2N1O2O0O2N1O2O0O1O1O2O0O1O1O100O1O1O100O1O1O1O2M2O1O1N2O1O1N2O1O1N2O1N2O1O2M2O1O1N2O1O1N2O1O1N2O2N1N2O1O1O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1N2O2M2N2N2N3M4K4M3M3M3M4L3M3L4M3M4L3M3M3M3L4L\\a_1"}, "label": "the bottom right side of a pizza"}]}
{"id": 322790, "image": "COCO_train2014_000000322790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing yellow colour jerkin\"."}], "task": "refering", "answer_template": "The \"a women wearing yellow colour jerkin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268], "bbox": [0.516328125, 0.23145882352941177, 0.7356874999999999, 0.7617882352941177], "iscrowd": 0, "area": 13431.42085, "rle": {"size": [425, 640], "counts": "maY46R=4K2N1O2O1N2N101N2O1O1N101O1N2O001N2OO10O01O10O01O10O010000010O010O01O10O01O010O010O1O010O10O0mGWOh3h0UL_Oh3a0TLEj3:SLMi34SL2k3MRL:k3ERLa0l3_OPLh0l3`0iJFU5h0YJ]Of5i0QJZOo5m0gIUOZ6Q1]IRO^6Y1YIjNb6a1UIbNg6R4K5L4K5L400O00100O100O100O11O00000000000000001O00000000001O2N2N2N2N2N2N3M2N2E:C>jNV1H7L5K5K4L3M4L4L4M3RNoE?U:@WF1m9N[FGh99]FVOP:j0X100O1000O01000000O2M2N2N3M2N3M2N3M2N3M3M2N3M2N[oU2"}, "label": "a women wearing yellow colour jerkin"}]}
{"id": 322790, "image": "COCO_train2014_000000322790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier , in a yellow jacket , towing a small skier\"."}], "task": "refering", "answer_template": "The \"a skier , in a yellow jacket , towing a small skier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268], "bbox": [0.516328125, 0.23145882352941177, 0.7356874999999999, 0.7617882352941177], "iscrowd": 0, "area": 13431.42085, "rle": {"size": [425, 640], "counts": "maY46R=4K2N1O2O1N2N101N2O1O1N101O1N2O001N2OO10O01O10O01O10O010000010O010O01O10O01O010O010O1O010O10O0mGWOh3h0UL_Oh3a0TLEj3:SLMi34SL2k3MRL:k3ERLa0l3_OPLh0l3`0iJFU5h0YJ]Of5i0QJZOo5m0gIUOZ6Q1]IRO^6Y1YIjNb6a1UIbNg6R4K5L4K5L400O00100O100O100O11O00000000000000001O00000000001O2N2N2N2N2N2N3M2N2E:C>jNV1H7L5K5K4L3M4L4L4M3RNoE?U:@WF1m9N[FGh99]FVOP:j0X100O1000O01000000O2M2N2N3M2N3M2N3M2N3M3M2N3M2N[oU2"}, "label": "a skier , in a yellow jacket , towing a small skier"}]}
{"id": 322790, "image": "COCO_train2014_000000322790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl in pink jacket being pulled on skis\"."}], "task": "refering", "answer_template": "The \"little girl in pink jacket being pulled on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 143, 144, 145, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260], "bbox": [0.21787499999999999, 0.3191058823529412, 0.340234375, 0.7550588235294117], "iscrowd": 0, "area": 7278.761699999998, "rle": {"size": [425, 640], "counts": "^jj12h:OfG1T88gGJT8>dGGV8b0bGDW8e0`GA\\8f0\\G_Oc8e0UG@i8e0QG^On8f0mF]OP9h0kFZOS9k0hFWOW9m0dFUOZ9P1bFRO\\9^2N2O1N3M2O1N2O1N3N1N2O1O2N1OUOk0M3M3M3M2N23L6K5J6K4K6K5K5J6K5J5LYNkGaNo7`1UHaNg7[1aHeN_7R1iHoNW7n0kHROV7k0mHTOT7i0nHUOU7h0nHUOU7ZOPH8n0<U7XOUH5h0b0U7UOWH6g0b0U7TOXH7e0c0U7SOYH8d0c0V7QOZH8d0d0T7QO\\H8b0e0T7PO[H;c0c0T7nN\\H=d0b0W8^OkGa0U8^OnG?R8APH=P8BSH<n7CUH:k7FXH7h7H\\H5e7I^H5b7JbH3_7KeH3Z7LjH1V7NnHOS7OQINo61Z^_5"}, "label": "little girl in pink jacket being pulled on skis"}]}
{"id": 322790, "image": "COCO_train2014_000000322790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a pink snowsuit and wearing a white hat being towed on skis\"."}], "task": "refering", "answer_template": "The \"a girl in a pink snowsuit and wearing a white hat being towed on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 143, 144, 145, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260], "bbox": [0.21787499999999999, 0.3191058823529412, 0.340234375, 0.7550588235294117], "iscrowd": 0, "area": 7278.761699999998, "rle": {"size": [425, 640], "counts": "^jj12h:OfG1T88gGJT8>dGGV8b0bGDW8e0`GA\\8f0\\G_Oc8e0UG@i8e0QG^On8f0mF]OP9h0kFZOS9k0hFWOW9m0dFUOZ9P1bFRO\\9^2N2O1N3M2O1N2O1N3N1N2O1O2N1OUOk0M3M3M3M2N23L6K5J6K4K6K5K5J6K5J5LYNkGaNo7`1UHaNg7[1aHeN_7R1iHoNW7n0kHROV7k0mHTOT7i0nHUOU7h0nHUOU7ZOPH8n0<U7XOUH5h0b0U7UOWH6g0b0U7TOXH7e0c0U7SOYH8d0c0V7QOZH8d0d0T7QO\\H8b0e0T7PO[H;c0c0T7nN\\H=d0b0W8^OkGa0U8^OnG?R8APH=P8BSH<n7CUH:k7FXH7h7H\\H5e7I^H5b7JbH3_7KeH3Z7LjH1V7NnHOS7OQINo61Z^_5"}, "label": "a girl in a pink snowsuit and wearing a white hat being towed on skis"}]}
{"id": 257255, "image": "COCO_train2014_000000257255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe bending over behind another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe bending over behind another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 133, 134, 135, 136, 151, 152, 153, 155, 186, 187, 204, 205], "bbox": [0.35428, 0.5075466666666667, 0.69158, 0.9371733333333335], "iscrowd": 0, "area": 4768.0632000000005, "rle": {"size": [375, 500], "counts": "XTQ23b;8H3N3L3M3N2M3M2N3N1N30O01O00100O001O1O1O\\OfELY:2mEK\\8HjH;oNKm73TI0TOKe77WILYOK^79ZIICGQ7a0\\IFj7:WHDj7=g1001O000O2O001O001O00001O001O001O0O2N1O1O2N101O001O00002N1O100O1000001O00000O10001O00000O10001O0000000000001O000000000O2OImEjNS:U1PFiNP:U1:O1O100\\O`E1`:ObEN_:1cEN]:2dEM\\:2eEN[:2eEN[:1fEOZ:1fEOY:2gENY:3fEMZ:3fEL[:4eEL[:4eEL[:4eEL[:4eELZ:4gELY:4hEKX:5hEKX:0_EI96Y:1^EH:7X:1_EG:7W:1`EH97W:1nEOR:1oENP:3PFNo92QFOn90TF0k90UF0k90UF1j9OWF1h9OYF0g90ZF0e90\\F0c90]F0c90^F0a90`FO`91aFO^91cFO\\91eFN[92eFOZ90hFOX91iFOV91jF0U90lFOT91lF0S9OoF0Q90oF1P9OQG0o80QG1n8NTG2k8NUG2j8OWG1h8OXG1h8OYG1f8N[G3d8M]G2c8N]G3b8M_G3`8MbG2\\80dG0[80gGOY80hGOX82iGMV83kGMT84g10000000001O1M3LgVh1"}, "label": "a giraffe bending over behind another giraffe"}]}
{"id": 257255, "image": "COCO_train2014_000000257255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shorter giraffe in back\"."}], "task": "refering", "answer_template": "The \"shorter giraffe in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 133, 134, 135, 136, 151, 152, 153, 155, 186, 187, 204, 205], "bbox": [0.35428, 0.5075466666666667, 0.69158, 0.9371733333333335], "iscrowd": 0, "area": 4768.0632000000005, "rle": {"size": [375, 500], "counts": "XTQ23b;8H3N3L3M3N2M3M2N3N1N30O01O00100O001O1O1O\\OfELY:2mEK\\8HjH;oNKm73TI0TOKe77WILYOK^79ZIICGQ7a0\\IFj7:WHDj7=g1001O000O2O001O001O00001O001O001O0O2N1O1O2N101O001O00002N1O100O1000001O00000O10001O00000O10001O0000000000001O000000000O2OImEjNS:U1PFiNP:U1:O1O100\\O`E1`:ObEN_:1cEN]:2dEM\\:2eEN[:2eEN[:1fEOZ:1fEOY:2gENY:3fEMZ:3fEL[:4eEL[:4eEL[:4eEL[:4eELZ:4gELY:4hEKX:5hEKX:0_EI96Y:1^EH:7X:1_EG:7W:1`EH97W:1nEOR:1oENP:3PFNo92QFOn90TF0k90UF0k90UF1j9OWF1h9OYF0g90ZF0e90\\F0c90]F0c90^F0a90`FO`91aFO^91cFO\\91eFN[92eFOZ90hFOX91iFOV91jF0U90lFOT91lF0S9OoF0Q90oF1P9OQG0o80QG1n8NTG2k8NUG2j8OWG1h8OXG1h8OYG1f8N[G3d8M]G2c8N]G3b8M_G3`8MbG2\\80dG0[80gGOY80hGOX82iGMV83kGMT84g10000000001O1M3LgVh1"}, "label": "shorter giraffe in back"}]}
{"id": 257255, "image": "COCO_train2014_000000257255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing in front of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe standing in front of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211], "bbox": [0.18034, 0.6292266666666667, 0.73484, 0.9348266666666667], "iscrowd": 0, "area": 11026.982550000002, "rle": {"size": [375, 500], "counts": "ZoQ12e;000COTE2k:OSE2m:ORE1n:0QE1n:1PEOP;2nD0Q;1nDOR;2mDNR;4mDMR;<0001O4]OAfEa0Z:=1O1O2M2O1O2N1N2O2N1O1N3N1O1O0O1000O10O10000O1000000O1000000O1000O10O1000000O11O000000000000001N100000000000001O0000000000000O2O000O1O1O100O1O2N100O1O1O100O1O2N100O1O1O1O1O1O2N1O1O1N2O1O2N1O2N1O2M2O2N1O2N1O1O2M2O2O001O0O2O001O001O001O0O2O00000000000O10O100000000000O1000000000000O1000000001[OiFhNW9W1jFiNW9U1jFkNV9T1lFkNU9S1lFmNT9R1mFnNT9P1nFoNS9o0nFPOS9o0nFQOS9m0oFROQ9m0PGSOQ9k0PGUOP9j0QGVOP9i0QGVOo8i0RGWOn8h0SGXOn8g0SGYOl8g0TGYOl8g0TGYOl8g0UGYOj8g0VGYOj8g0VGYOk8f0VGZOi8f0WGZOi8e0XG[Oh8e0XG[Oh8e0YG[Of8e0ZG[Og8d0YG\\Og8d0ZG\\Oe8d0[G\\Oe8d0[G\\Oe8d0\\G\\Oc8d0]G\\Od8b0]G^Oc8b0]G_Ob8a0_G^Oa8b0_G^Oa8b0_G_O`8a0aG^O`8a0`G_O`8a0`G_O`8a0aG_O^8a0bG_O^8`0cG@^8?bGB]8>dGA\\8?dGA\\8?dGB[8>fGA[8>eGB[8>eGCZ8=gGBY8>gGBY8=hGDX8;hGEX8;iGDW8<iGDW8<iGEV8;kGDV8;jGEV8;jGFU8:lGET8;lGET8:mGGS88nGGR89nGGR89nGHQ88oGHR87oGHQ88oGIP87PHIP87QHIo76QHJo75RHLm74THKl76SHKm74THKl75THLk74UHLl73UHMj73VHMj73VHNi72XHMi72WHNi73VHNi72XHMh73XHNh71XHOh71YHOf71ZHOf71ZH0f7O[H0e70[H1d70\\HOd71\\H0d7O\\H1d7O]H1b7O^H1c7N]H3b7M_H2b7M^H4a7L_H4a7M_H3a7L_H4a7L_H5`7KaH4`7K`H6_7JaHOg70ZHGn79h101OW^`1"}, "label": "a giraffe standing in front of another giraffe"}]}
{"id": 257255, "image": "COCO_train2014_000000257255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe staring at the camera\"."}], "task": "refering", "answer_template": "The \"the giraffe staring at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211], "bbox": [0.18034, 0.6292266666666667, 0.73484, 0.9348266666666667], "iscrowd": 0, "area": 11026.982550000002, "rle": {"size": [375, 500], "counts": "ZoQ12e;000COTE2k:OSE2m:ORE1n:0QE1n:1PEOP;2nD0Q;1nDOR;2mDNR;4mDMR;<0001O4]OAfEa0Z:=1O1O2M2O1O2N1N2O2N1O1N3N1O1O0O1000O10O10000O1000000O1000000O1000O10O1000000O11O000000000000001N100000000000001O0000000000000O2O000O1O1O100O1O2N100O1O1O100O1O2N100O1O1O1O1O1O2N1O1O1N2O1O2N1O2N1O2M2O2N1O2N1O1O2M2O2O001O0O2O001O001O001O0O2O00000000000O10O100000000000O1000000000000O1000000001[OiFhNW9W1jFiNW9U1jFkNV9T1lFkNU9S1lFmNT9R1mFnNT9P1nFoNS9o0nFPOS9o0nFQOS9m0oFROQ9m0PGSOQ9k0PGUOP9j0QGVOP9i0QGVOo8i0RGWOn8h0SGXOn8g0SGYOl8g0TGYOl8g0TGYOl8g0UGYOj8g0VGYOj8g0VGYOk8f0VGZOi8f0WGZOi8e0XG[Oh8e0XG[Oh8e0YG[Of8e0ZG[Og8d0YG\\Og8d0ZG\\Oe8d0[G\\Oe8d0[G\\Oe8d0\\G\\Oc8d0]G\\Od8b0]G^Oc8b0]G_Ob8a0_G^Oa8b0_G^Oa8b0_G_O`8a0aG^O`8a0`G_O`8a0`G_O`8a0aG_O^8a0bG_O^8`0cG@^8?bGB]8>dGA\\8?dGA\\8?dGB[8>fGA[8>eGB[8>eGCZ8=gGBY8>gGBY8=hGDX8;hGEX8;iGDW8<iGDW8<iGEV8;kGDV8;jGEV8;jGFU8:lGET8;lGET8:mGGS88nGGR89nGGR89nGHQ88oGHR87oGHQ88oGIP87PHIP87QHIo76QHJo75RHLm74THKl76SHKm74THKl75THLk74UHLl73UHMj73VHMj73VHNi72XHMi72WHNi73VHNi72XHMh73XHNh71XHOh71YHOf71ZHOf71ZH0f7O[H0e70[H1d70\\HOd71\\H0d7O\\H1d7O]H1b7O^H1c7N]H3b7M_H2b7M^H4a7L_H4a7M_H3a7L_H4a7L_H5`7KaH4`7K`H6_7JaHOg70ZHGn79h101OW^`1"}, "label": "the giraffe staring at the camera"}]}
{"id": 478445, "image": "COCO_train2014_000000478445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy who is sitting up high\"."}], "task": "refering", "answer_template": "The \"guy who is sitting up high\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 185, 186, 187, 208, 209, 210, 211, 231, 232, 233, 234, 256, 257, 279, 280, 281, 302, 303, 304], "bbox": [0.072046875, 0.4806367924528302, 0.25407812500000004, 0.9206132075471699], "iscrowd": 0, "area": 6820.862049999996, "rle": {"size": [424, 640], "counts": "nYc0252\\<:cC0k;9VDGg;;YDEe;>ZDCc;?]DB`;`0aD@\\;b0dD_O[;a0eD@Y;a0gD@W;b0iD^OV;b0jD_OT;b0lD_OR;b0nD_OP;b0QE\\OP;d0PE\\Oo:f0PEYOP;h0PEWOQ;i0PEVOo:k0QETOo:m0QESOn:o0QEQOn:P1SEPOk:Q1UEoNj:S1UEnNg:U1YEkNe:W1\\EiN`:[1_EeN^:^1bEbN\\:`1dEaN4_Ok9S2QF]NNFo9o1SFeNl9]1SFcNl9^1UFaNk9_1UFaNj9a1VF^Ni9c1WF]Ni9c1WF]Nh9e1XFYNi9>TF23@i9=WF11Ah9=[F0LDh9;_FOIFi99aFOGGh99eFNBIi99lFG\\O0g98nFIZOOk1[Ol5k0PILZOMi1]On5h0PINYOMh1_On5d0SI1VOLh1Ao5?UI4UOKe1DR6;UI7SOJd1GT65WI;QOHa1LX6LYI`0oNG^1O_6CVIh0mNE]12l7:fFD^13k79hFC\\16k76jFD[17j74mFE3Bf0g0Y81oFF0Df0h0Z8MRGGJGh0g0\\8JSGJEGj0g0^8GUG687c8BVG73;g8]OXG7N?j8YOYG8Ka0l8VO[G:Dc0P9SO]G=^Oc0S9POaGU2\\8kMeG[1@cNk80hGZ1@eNg81jGY1AeNd82mGV1BgNa82nGV1DfN]8IPG5Q1Z1DgNZ8JUGNR1^1BiNW8J`H\\1ZOjNU8JcH\\1XOiNT8KfH\\1mNbNE7g8KiHn1`NWNf8KlHm1^NXNe8JPIn1ZNWNf:h1[EVNg:i1YEVNi:i1XEUNj:j1;gNkD=V;_OPE>Q;^OUE?o:ZOWEc0k;M1N1O0N2O2O001O010O0O10001O0000000O100000000O100O106IhTU6"}, "label": "guy who is sitting up high"}]}
{"id": 478445, "image": "COCO_train2014_000000478445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in a chair\"."}], "task": "refering", "answer_template": "The \"a man sitting in a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 185, 186, 187, 208, 209, 210, 211, 231, 232, 233, 234, 256, 257, 279, 280, 281, 302, 303, 304], "bbox": [0.072046875, 0.4806367924528302, 0.25407812500000004, 0.9206132075471699], "iscrowd": 0, "area": 6820.862049999996, "rle": {"size": [424, 640], "counts": "nYc0252\\<:cC0k;9VDGg;;YDEe;>ZDCc;?]DB`;`0aD@\\;b0dD_O[;a0eD@Y;a0gD@W;b0iD^OV;b0jD_OT;b0lD_OR;b0nD_OP;b0QE\\OP;d0PE\\Oo:f0PEYOP;h0PEWOQ;i0PEVOo:k0QETOo:m0QESOn:o0QEQOn:P1SEPOk:Q1UEoNj:S1UEnNg:U1YEkNe:W1\\EiN`:[1_EeN^:^1bEbN\\:`1dEaN4_Ok9S2QF]NNFo9o1SFeNl9]1SFcNl9^1UFaNk9_1UFaNj9a1VF^Ni9c1WF]Ni9c1WF]Nh9e1XFYNi9>TF23@i9=WF11Ah9=[F0LDh9;_FOIFi99aFOGGh99eFNBIi99lFG\\O0g98nFIZOOk1[Ol5k0PILZOMi1]On5h0PINYOMh1_On5d0SI1VOLh1Ao5?UI4UOKe1DR6;UI7SOJd1GT65WI;QOHa1LX6LYI`0oNG^1O_6CVIh0mNE]12l7:fFD^13k79hFC\\16k76jFD[17j74mFE3Bf0g0Y81oFF0Df0h0Z8MRGGJGh0g0\\8JSGJEGj0g0^8GUG687c8BVG73;g8]OXG7N?j8YOYG8Ka0l8VO[G:Dc0P9SO]G=^Oc0S9POaGU2\\8kMeG[1@cNk80hGZ1@eNg81jGY1AeNd82mGV1BgNa82nGV1DfN]8IPG5Q1Z1DgNZ8JUGNR1^1BiNW8J`H\\1ZOjNU8JcH\\1XOiNT8KfH\\1mNbNE7g8KiHn1`NWNf8KlHm1^NXNe8JPIn1ZNWNf:h1[EVNg:i1YEVNi:i1XEUNj:j1;gNkD=V;_OPE>Q;^OUE?o:ZOWEc0k;M1N1O0N2O2O001O010O0O10001O0000000O100000000O100O106IhTU6"}, "label": "a man sitting in a chair"}]}
{"id": 478445, "image": "COCO_train2014_000000478445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player cheering posing and wearing blue and red color dress\"."}], "task": "refering", "answer_template": "The \"a tennis player cheering posing and wearing blue and red color dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 146, 147, 148, 169, 170, 171, 172, 193, 194, 195, 216, 217, 218, 238, 239, 240, 241, 263, 264], "bbox": [0.37959375, 0.3671698113207547, 0.5143125, 0.750495283018868], "iscrowd": 0, "area": 6923.150349999997, "rle": {"size": [424, 640], "counts": "kiT37k<9L4L3jE^Oa7e0[H_Ob7d0YHAd7a0WHEf7?WHAh7c0TH_Oi7e0UH[Oi7P1mGROQ8U1iGkNU8]1eGcNP8j1nGVNQ8n1lGRNS8S2iGlMV8[2eGeMZ8b2`G\\Ma8h2\\GXMd8k2_GoL`8T3`GkL_8X3^GhLa8\\3\\GeLc8]3[GcLe8_3YGbLf8P40000001N2O2N1O1O1O1O0PLPGg3R9XLoFg3Z9N2M3N2\\O_F^Mg9\\2^F^Me9a2_FWMf9h2>jMbEn0Q;QOREm0o:QOUEk0l:UO[Ec0f:\\O]Eb0d:]O[Ec0g:\\OXEe0h:[OWEe0i:[OWEf0i:YOXEg0o0QOT89lFe0o0UOU85lFDDn0Y1[OX82PGb0e0_O[81oF?e0A]80mF?e0C^8MnF?:Ni8CmF>K]OEg0c9]OmF?Ee0_9lNlF>Ai0e9iNiFV2V9jMkFU2V9kMjFT2W9mMhFR2Y9nMgFR2Y9nMgF`1@bNl9LdF_1CdNo9H^F`1GhNk9G_F\\1KlNg9G`FW1MROe9D`FU10VOZ:c0kE]OV:>oEAR:7VFHe;01O1O1O1O1N102Mn^P4"}, "label": "a tennis player cheering posing and wearing blue and red color dress"}]}
{"id": 478445, "image": "COCO_train2014_000000478445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player\"."}], "task": "refering", "answer_template": "The \"tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 146, 147, 148, 169, 170, 171, 172, 193, 194, 195, 216, 217, 218, 238, 239, 240, 241, 263, 264], "bbox": [0.37959375, 0.3671698113207547, 0.5143125, 0.750495283018868], "iscrowd": 0, "area": 6923.150349999997, "rle": {"size": [424, 640], "counts": "kiT37k<9L4L3jE^Oa7e0[H_Ob7d0YHAd7a0WHEf7?WHAh7c0TH_Oi7e0UH[Oi7P1mGROQ8U1iGkNU8]1eGcNP8j1nGVNQ8n1lGRNS8S2iGlMV8[2eGeMZ8b2`G\\Ma8h2\\GXMd8k2_GoL`8T3`GkL_8X3^GhLa8\\3\\GeLc8]3[GcLe8_3YGbLf8P40000001N2O2N1O1O1O1O0PLPGg3R9XLoFg3Z9N2M3N2\\O_F^Mg9\\2^F^Me9a2_FWMf9h2>jMbEn0Q;QOREm0o:QOUEk0l:UO[Ec0f:\\O]Eb0d:]O[Ec0g:\\OXEe0h:[OWEe0i:[OWEf0i:YOXEg0o0QOT89lFe0o0UOU85lFDDn0Y1[OX82PGb0e0_O[81oF?e0A]80mF?e0C^8MnF?:Ni8CmF>K]OEg0c9]OmF?Ee0_9lNlF>Ai0e9iNiFV2V9jMkFU2V9kMjFT2W9mMhFR2Y9nMgFR2Y9nMgF`1@bNl9LdF_1CdNo9H^F`1GhNk9G_F\\1KlNg9G`FW1MROe9D`FU10VOZ:c0kE]OV:>oEAR:7VFHe;01O1O1O1O1N102Mn^P4"}, "label": "tennis player"}]}
{"id": 412910, "image": "COCO_train2014_000000412910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and yellow surfboard , in memory of rebecca ehehalt\"."}], "task": "refering", "answer_template": "The \"a green and yellow surfboard , in memory of rebecca ehehalt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 125, 126, 127, 142, 143, 144, 145, 158, 159, 160, 161, 162, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314, 315, 316, 329, 330, 331, 332, 333, 346, 347, 348, 349, 350, 364, 365, 366, 367], "bbox": [0.3454054054054054, 0.26579687500000004, 0.6187318087318088, 0.9475156250000001], "iscrowd": 0, "area": 43281.45104999999, "rle": {"size": [640, 481], "counts": "ZPX39Q22X?i0m^O;n`0^1h_OVMV?n2h@VMR?n2k@VMo>n2PAUMk>o2RAGV<=iCHP<;oCJj;8UDMd;6\\DO];4bD1V;3hD3Q;0nD4l:0QE6g:MYE8`:j4H7J7I7H8I6I8I7I7H7J7H8I3M2M3N3M2M3N2M4M2N2M4M2M31O100O1O010O1O100O1O1O010O1O100O1O1O10O01O100O2N101N2N2N2O0O2N2O1N2N2N101N2N2O1N1O2O1N2N2N2O0O2N2O1N2N101N2N2N2O0O2N2O1N2N2N4M4K6J5L5J5K6K4K6J5K6K4K6J5K5J7I6J7I6J7I6K6I6J7I6mNT1YOf0ZOf0ZOg0ZO^oa3"}, "label": "a green and yellow surfboard , in memory of rebecca ehehalt"}]}
{"id": 412910, "image": "COCO_train2014_000000412910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and yellow surfboard , standing next to blue and orange surfboards\"."}], "task": "refering", "answer_template": "The \"a green and yellow surfboard , standing next to blue and orange surfboards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 125, 126, 127, 142, 143, 144, 145, 158, 159, 160, 161, 162, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314, 315, 316, 329, 330, 331, 332, 333, 346, 347, 348, 349, 350, 364, 365, 366, 367], "bbox": [0.3454054054054054, 0.26579687500000004, 0.6187318087318088, 0.9475156250000001], "iscrowd": 0, "area": 43281.45104999999, "rle": {"size": [640, 481], "counts": "ZPX39Q22X?i0m^O;n`0^1h_OVMV?n2h@VMR?n2k@VMo>n2PAUMk>o2RAGV<=iCHP<;oCJj;8UDMd;6\\DO];4bD1V;3hD3Q;0nD4l:0QE6g:MYE8`:j4H7J7I7H8I6I8I7I7H7J7H8I3M2M3N3M2M3N2M4M2N2M4M2M31O100O1O010O1O100O1O1O010O1O100O1O1O10O01O100O2N101N2N2N2O0O2N2O1N2N2N101N2N2O1N1O2O1N2N2N2O0O2N2O1N2N101N2N2N2O0O2N2O1N2N2N4M4K6J5L5J5K6K4K6J5K6K4K6J5K5J7I6J7I6J7I6K6I6J7I6mNT1YOf0ZOf0ZOg0ZO^oa3"}, "label": "a green and yellow surfboard , standing next to blue and orange surfboards"}]}
{"id": 412910, "image": "COCO_train2014_000000412910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange and pink surfboard\"."}], "task": "refering", "answer_template": "The \"an orange and pink surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 42, 43, 44, 45, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 127, 128, 129, 145, 146, 162, 163, 179], "bbox": [0.4286070686070686, 0.042734375000000005, 0.705135135135135, 0.45689062500000005], "iscrowd": 0, "area": 19741.01465, "rle": {"size": [640, 481], "counts": "UmP42gc07G9F:G9G9L5M3L4L4L4L4L5K4M3L4L4L4L4L4M3M3M4L3N2M4L3M4L3N2M4L3M4M2M3M4M2M3M4L3N3M3M3M3M3M3M4L3M3M3M3M3M4M2N2M3N2N3M3L4M3M2M4M3M3M3M2N3M3M3M3L4N1N1OM4K4M4K4M3L5L3L4L5K4M4K4L4M4K4L5L4K6J5L4K5M3M4L3M3M3M3N2M4L3M3M3M3L5L3M5J6K4L5J5L5J6K4M4K5L3L5L3M4K5K6K4K5L5J5K6K7H<Dkch2"}, "label": "an orange and pink surfboard"}]}
{"id": 412910, "image": "COCO_train2014_000000412910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange memorial surfboard\"."}], "task": "refering", "answer_template": "The \"an orange memorial surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 42, 43, 44, 45, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 127, 128, 129, 145, 146, 162, 163, 179], "bbox": [0.4286070686070686, 0.042734375000000005, 0.705135135135135, 0.45689062500000005], "iscrowd": 0, "area": 19741.01465, "rle": {"size": [640, 481], "counts": "UmP42gc07G9F:G9G9L5M3L4L4L4L4L5K4M3L4L4L4L4L4M3M3M4L3N2M4L3M4L3N2M4L3M4M2M3M4M2M3M4L3N3M3M3M3M3M3M4L3M3M3M3M3M4M2N2M3N2N3M3L4M3M2M4M3M3M3M2N3M3M3M3L4N1N1OM4K4M4K4M3L5L3L4L5K4M4K4L4M4K4L5L4K6J5L4K5M3M4L3M3M3M3N2M4L3M3M3M3L5L3M5J6K4L5J5L5J6K4M4K5L3L5L3M4K5K6K4K5L5J5K6K7H<Dkch2"}, "label": "an orange memorial surfboard"}]}
{"id": 412910, "image": "COCO_train2014_000000412910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue surfboard\"."}], "task": "refering", "answer_template": "The \"a blue surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 96, 97, 98, 113, 114, 115, 116, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 184, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 282, 283, 284, 285, 299, 300, 301, 302, 316, 317, 318, 319, 333, 334, 335, 336, 350, 351, 352, 367, 368, 369, 385], "bbox": [0.5646569646569647, 0.1901875, 0.8439916839916841, 0.9667031250000001], "iscrowd": 0, "area": 46161.462450000006, "rle": {"size": [640, 481], "counts": "aXZ54jc04K5L4L5K4L4K5L5K4L4L4K6K4L4L4^FnMP1V2fMSOV2S1_L6^3X8M3M4L3L4M3M3M2N3M3L3O2M3N1O2M3N2M2O2M3N1N3N2N1N3N2M3N1N3N2N1N3N2M2O2M3N1N3N2N2M2O2M3N1N3N2N1N3N0O01O001O1O00000O2O00001O0000001N10001O00002N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M5L5K5K5K5H8G9G9F9H9G9G9G9G9G9F:G8H9G9G9G9F:G9Gg3YLm0SOl0TOm0ROf`^1"}, "label": "a blue surfboard"}]}
{"id": 412910, "image": "COCO_train2014_000000412910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue surfboard\"."}], "task": "refering", "answer_template": "The \"blue surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 96, 97, 98, 113, 114, 115, 116, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 184, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 282, 283, 284, 285, 299, 300, 301, 302, 316, 317, 318, 319, 333, 334, 335, 336, 350, 351, 352, 367, 368, 369, 385], "bbox": [0.5646569646569647, 0.1901875, 0.8439916839916841, 0.9667031250000001], "iscrowd": 0, "area": 46161.462450000006, "rle": {"size": [640, 481], "counts": "aXZ54jc04K5L4L5K4L4K5L5K4L4L4K6K4L4L4^FnMP1V2fMSOV2S1_L6^3X8M3M4L3L4M3M3M2N3M3L3O2M3N1O2M3N2M2O2M3N1N3N2N1N3N2M3N1N3N2N1N3N2M2O2M3N1N3N2N2M2O2M3N1N3N2N1N3N0O01O001O1O00000O2O00001O0000001N10001O00002N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M5L5K5K5K5H8G9G9F9H9G9G9G9G9G9F:G8H9G9G9G9F:G9Gg3YLm0SOl0TOm0ROf`^1"}, "label": "blue surfboard"}]}
{"id": 330991, "image": "COCO_train2014_000000330991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the women is holding her cellphone\"."}], "task": "refering", "answer_template": "The \"the women is holding her cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.24851562500000002, 0.19030373831775702, 0.9127500000000002, 0.9898598130841121], "iscrowd": 0, "area": 69415.2561, "rle": {"size": [428, 640], "counts": "gkR22X=3M2M4M2M4M2N2M3N2M3N2M3N2M3M3N2O1N2O1N101O1N2O1N2O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O100O100O100O10000O100O100O1000000O10000O1000000O10000O1000000O10000O1O100O1O100O1O100O1O100O1O1O100O1`KoLdNR3X1VMdNj2W1`McNa2Y1gMaN[2[1mM`NT2\\1TNTLhNn0V3j2YNoKjNR1o2k2^NiKlNY1g2k2cN_KSOa1\\2m22oLOn26nLLP39lLHQ3?iLCT3e0eL]OX3k0[L]Ob3j0oKA_3_5L4M3L4M3O1O100O1O1O100O1O1O100O1O1O100O1000000000000000000O100000000000000000000O100000000000000000000O100000O10000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000>Bc0]Oc0]Od0\\Oc0]Oc0]Od0\\Oc0]Oc0]O2N00000000000jLWHi0i7VOXHj0h7TOZHl0f7SO[Hm0e7RO\\Hn0d7PO^HP1b7oN^HR1b7lN`HT1`7kNaHU1_7jNbHV1^7hNdHX1\\7gNeHY1[7fNfHZ1Z7eNgH[1Y7dNgH]1Y7bNhH^1W7bNjH^1V7bNjH^1V7aNkH_1U7`NlH`1T7_NmHa1S7^NnHb1R7^NmHc1S7\\NnHd1R7[NnHf1R7YNoHg1Q7YNnHh1R7WNoHi1Q7VNoHk1Q7TNPIl1P7TNoHm1Q7RNPIn1P7QNPIP2P7oMQIQ2o6nMQIS2o6mMQIS2o6lMQIU2o6kMQIU2o6kMPIV2P7jMPIV2P7jMoHW2Q7hMPIX2P7hMoHY2Q7gMoHY2Q7gMoHY2Q7gMnHZ2Q7gMoHY2Q7gMoHY2Q7gMnHZ2R7fMnHZ2R7fMnHZ2R7fMnHZ2R7fMmH[2S7eMmH[2S7dMnH\\2R7dMmH]2S7cMmH]2S7cMmH]2S7cMmH]2S7cMlH^2T7bMlH^2T7bMlH^2T7bMkH_2U7aMkH_2U7aMkH_2U7aMkH_2U7`MkHa2U7_MkHa2U7_MkHa2U7_MjHb2V7^MjHb2U7_MkHa2U7_MkHa2U7_MjHb2V7^MjHb2V7^MjHb2V7^MiHc2W7]MiHc2X7\\MhHd2X7[MiHe2W7[MhHf2Y7YMgHg2Y7YMgHg2Z7XMeHi2[7WMeHi2[7WMeHh2]7WMcHi2]7VMcHk2]7UMcHk2^7SMcHm2]7SMdHl2]7RMdHn2\\7RMdHn2\\7QMeHo2\\7PMdHP3]7nLdHR3]7mLcHS3^7kLdHT3]7kLcHU3^7iLcHW3^7bLhH^3c800000001O0000000000001O0000000000001O0000001O1O1O2N<D7I1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N2N2N2N3M2N2N3M2N2N3M2N2N3M2Nmof0"}, "label": "the women is holding her cellphone"}]}
{"id": 330991, "image": "COCO_train2014_000000330991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl clicking on her mobile phone\"."}], "task": "refering", "answer_template": "The \"a girl clicking on her mobile phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.24851562500000002, 0.19030373831775702, 0.9127500000000002, 0.9898598130841121], "iscrowd": 0, "area": 69415.2561, "rle": {"size": [428, 640], "counts": "gkR22X=3M2M4M2M4M2N2M3N2M3N2M3N2M3M3N2O1N2O1N101O1N2O1N2O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O100O100O100O10000O100O100O1000000O10000O1000000O10000O1000000O10000O1O100O1O100O1O100O1O100O1O1O100O1`KoLdNR3X1VMdNj2W1`McNa2Y1gMaN[2[1mM`NT2\\1TNTLhNn0V3j2YNoKjNR1o2k2^NiKlNY1g2k2cN_KSOa1\\2m22oLOn26nLLP39lLHQ3?iLCT3e0eL]OX3k0[L]Ob3j0oKA_3_5L4M3L4M3O1O100O1O1O100O1O1O100O1O1O100O1000000000000000000O100000000000000000000O100000000000000000000O100000O10000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000>Bc0]Oc0]Od0\\Oc0]Oc0]Od0\\Oc0]Oc0]O2N00000000000jLWHi0i7VOXHj0h7TOZHl0f7SO[Hm0e7RO\\Hn0d7PO^HP1b7oN^HR1b7lN`HT1`7kNaHU1_7jNbHV1^7hNdHX1\\7gNeHY1[7fNfHZ1Z7eNgH[1Y7dNgH]1Y7bNhH^1W7bNjH^1V7bNjH^1V7aNkH_1U7`NlH`1T7_NmHa1S7^NnHb1R7^NmHc1S7\\NnHd1R7[NnHf1R7YNoHg1Q7YNnHh1R7WNoHi1Q7VNoHk1Q7TNPIl1P7TNoHm1Q7RNPIn1P7QNPIP2P7oMQIQ2o6nMQIS2o6mMQIS2o6lMQIU2o6kMQIU2o6kMPIV2P7jMPIV2P7jMoHW2Q7hMPIX2P7hMoHY2Q7gMoHY2Q7gMoHY2Q7gMnHZ2Q7gMoHY2Q7gMoHY2Q7gMnHZ2R7fMnHZ2R7fMnHZ2R7fMnHZ2R7fMmH[2S7eMmH[2S7dMnH\\2R7dMmH]2S7cMmH]2S7cMmH]2S7cMmH]2S7cMlH^2T7bMlH^2T7bMlH^2T7bMkH_2U7aMkH_2U7aMkH_2U7aMkH_2U7`MkHa2U7_MkHa2U7_MkHa2U7_MjHb2V7^MjHb2U7_MkHa2U7_MkHa2U7_MjHb2V7^MjHb2V7^MjHb2V7^MiHc2W7]MiHc2X7\\MhHd2X7[MiHe2W7[MhHf2Y7YMgHg2Y7YMgHg2Z7XMeHi2[7WMeHi2[7WMeHh2]7WMcHi2]7VMcHk2]7UMcHk2^7SMcHm2]7SMdHl2]7RMdHn2\\7RMdHn2\\7QMeHo2\\7PMdHP3]7nLdHR3]7mLcHS3^7kLdHT3]7kLcHU3^7iLcHW3^7bLhH^3c800000001O0000000000001O0000000000001O0000001O1O1O2N<D7I1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N2N2N2N3M2N2N3M2N2N3M2N2N3M2Nmof0"}, "label": "a girl clicking on her mobile phone"}]}
{"id": 330991, "image": "COCO_train2014_000000330991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with a book in her arms behind the lady holding the phone up\"."}], "task": "refering", "answer_template": "The \"the person with a book in her arms behind the lady holding the phone up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 208, 209, 210, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 303, 323, 324, 325, 326], "bbox": [0.0463125, 0.5249766355140187, 0.189765625, 0.9866588785046728], "iscrowd": 0, "area": 13472.212850000004, "rle": {"size": [428, 640], "counts": "Pj<b0f<c0]Oc0]Od0\\O]2cM5K5L4L4L3L4M3M3O1O10000O100O10000O100O10000O10000O100O10000O100O10000O100001O1O2N1O1O2N1O1O2N1O1O2N100O2N1O1O9G?A7I2N3M3M2N3M2N3M2N3M3M2N3M5K7I7I7I6J7I7I7I7I6J7I7I7I6J7I7Id`h6"}, "label": "the person with a book in her arms behind the lady holding the phone up"}]}
{"id": 330991, "image": "COCO_train2014_000000330991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man looking right side wearing full coat\"."}], "task": "refering", "answer_template": "The \"a man looking right side wearing full coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 186, 187, 188, 210, 211, 233, 234, 256, 257, 280, 303, 326], "bbox": [0.11321874999999999, 0.4757242990654206, 0.217375, 0.9858878504672898], "iscrowd": 0, "area": 4602.352649999998, "rle": {"size": [428, 640], "counts": "mYn02o<a0I1O1N200O2O3M4L4L3M3M3M2M4M2N2N3M2N3M3M3M3M4L3M3N2M4L3M4L5K5L4K5KQNZF=a9_OhFa0S9ZOUGh0f8TOaGn0[8nNkGT1P8jNTHY1g7eN\\H_1_7`NdHc1W7[NlHi1a7cMbHa2\\7[MfHi2m84L4?A6JN2N2N2N2N2NN2M3M3M3M3M4M3Le0[Oe0[O?A3M3Mho`6"}, "label": "a man looking right side wearing full coat"}]}
{"id": 330991, "image": "COCO_train2014_000000330991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man behind the woman with the brown jacket\"."}], "task": "refering", "answer_template": "The \"the man behind the woman with the brown jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 186, 187, 188, 210, 211, 233, 234, 256, 257, 280, 303, 326], "bbox": [0.11321874999999999, 0.4757242990654206, 0.217375, 0.9858878504672898], "iscrowd": 0, "area": 4602.352649999998, "rle": {"size": [428, 640], "counts": "mYn02o<a0I1O1N200O2O3M4L4L3M3M3M2M4M2N2N3M2N3M3M3M3M4L3M3N2M4L3M4L5K5L4K5KQNZF=a9_OhFa0S9ZOUGh0f8TOaGn0[8nNkGT1P8jNTHY1g7eN\\H_1_7`NdHc1W7[NlHi1a7cMbHa2\\7[MfHi2m84L4?A6JN2N2N2N2N2NN2M3M3M3M3M4M3Le0[Oe0[O?A3M3Mho`6"}, "label": "the man behind the woman with the brown jacket"}]}
{"id": 576749, "image": "COCO_train2014_000000576749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right most bench in front of water and a boat\"."}], "task": "refering", "answer_template": "The \"the right most bench in front of water and a boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [321, 333, 334, 335, 336, 337, 338, 339, 340, 351, 352, 353, 354, 355, 356, 357, 358, 359, 370, 371, 372, 373, 374, 375, 376, 377, 378, 389, 390, 397], "bbox": [0.5089038461538461, 0.7213437500000001, 0.9293076923076923, 0.9011250000000001], "iscrowd": 0, "area": 9648.533000000003, "rle": {"size": [640, 520], "counts": "_UV5?lb0c0^Ob0O02O0O2O1N101N2O03N:F:F3MM3L4L4L4L5L4L400000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O1000000000000000000000000000O1000000000000000000000O1000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000Q^OlNRa0T1l^OlNVa0T1h^OmNYa0S1e^OnN\\a0R1a^OoNaa0P1^^OQOb`0Mk_OR1AQOf`09]_Of0KROi`0d0P_O:5SOl`0n1Q_OSNQa0m1m^OTNTa0l1j^OUNWa0k1g^OUN[a0W21O17I<D<ED;C=C=C=C=C=C=Cnef0"}, "label": "the right most bench in front of water and a boat"}]}
{"id": 576749, "image": "COCO_train2014_000000576749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large boat carrying many people , which has four masts and which is being watched by a couple on a bench\"."}], "task": "refering", "answer_template": "The \"a large boat carrying many people , which has four masts and which is being watched by a couple on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 262, 263, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 312, 313, 314, 315], "bbox": [0.034096153846153845, 0.584859375, 0.8726923076923078, 0.699171875], "iscrowd": 0, "area": 18332.776199999993, "rle": {"size": [640, 520], "counts": "Vd;7`c0;I6N3M201N2N101N101N1O2O0O2N2O0O2N101N1O2O1N1010O01O010O0010O0O1O001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O000000000000000000000000000000001O00000000000001O00000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000001O2N1O2N1O2N1O0000000000000000000000000000000000000000000000000000000001O00000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000000O1000000000000000O10000000000000O1N2O1O1N2N2N2N200000000000000000O100O010O100O1O1000000000000000O1000000000000000003M000O2OK5J6O100001O00000000000O101O0000000000001O1N2O1O1O1O1N1000O1000O010000O0100000O1N1O2N2N2O0O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N3M3MTlX1"}, "label": "a large boat carrying many people , which has four masts and which is being watched by a couple on a bench"}]}
{"id": 576749, "image": "COCO_train2014_000000576749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cruise boat\"."}], "task": "refering", "answer_template": "The \"cruise boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 262, 263, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 312, 313, 314, 315], "bbox": [0.034096153846153845, 0.584859375, 0.8726923076923078, 0.699171875], "iscrowd": 0, "area": 18332.776199999993, "rle": {"size": [640, 520], "counts": "Vd;7`c0;I6N3M201N2N101N101N1O2O0O2N2O0O2N101N1O2O1N1010O01O010O0010O0O1O001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O000000000000000000000000000000001O00000000000001O00000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000001O2N1O2N1O2N1O0000000000000000000000000000000000000000000000000000000001O00000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000000O1000000000000000O10000000000000O1N2O1O1N2N2N2N200000000000000000O100O010O100O1O1000000000000000O1000000000000000003M000O2OK5J6O100001O00000000000O101O0000000000001O1N2O1O1O1O1N1000O1000O010000O0100000O1N1O2N2N2O0O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N3M3MTlX1"}, "label": "cruise boat"}]}
{"id": 576749, "image": "COCO_train2014_000000576749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left most of two benches\"."}], "task": "refering", "answer_template": "The \"left most of two benches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [324, 325, 326, 327, 328, 329, 330, 331, 332, 343, 344, 345, 346, 347, 348, 349, 350, 351, 362, 363, 364, 365, 366, 367, 368, 369, 370, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 400], "bbox": [0.04890384615384615, 0.766140625, 0.5111923076923077, 0.9186249999999999], "iscrowd": 0, "area": 17135.46950000001, "rle": {"size": [640, 520], "counts": "TV`02ac0>A`0@?B=B?B=L40000000001O000000C=J6000000000O1000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O10O100000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O10000000O1000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000O1000000000000000O1000000000000000000000000000000O1000000001O2N1O2N2N1O2N1O2N2N1O2N2N000000000000O10000000O10J6]Oc0^Ob0]Oc0]OYYn4"}, "label": "left most of two benches"}]}
{"id": 576749, "image": "COCO_train2014_000000576749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banch\"."}], "task": "refering", "answer_template": "The \"a banch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [324, 325, 326, 327, 328, 329, 330, 331, 332, 343, 344, 345, 346, 347, 348, 349, 350, 351, 362, 363, 364, 365, 366, 367, 368, 369, 370, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 400], "bbox": [0.04890384615384615, 0.766140625, 0.5111923076923077, 0.9186249999999999], "iscrowd": 0, "area": 17135.46950000001, "rle": {"size": [640, 520], "counts": "TV`02ac0>A`0@?B=B?B=L40000000001O000000C=J6000000000O1000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O10O100000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O10000000O1000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000O1000000000000000O1000000000000000000000000000000O1000000001O2N1O2N2N1O2N1O2N2N1O2N2N000000000000O10000000O10J6]Oc0^Ob0]Oc0]OYYn4"}, "label": "a banch"}]}
{"id": 330993, "image": "COCO_train2014_000000330993.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green colored car is parked beside the road\"."}], "task": "refering", "answer_template": "The \"green colored car is parked beside the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.745265625, 0.0034813084112149534, 1.0, 0.9870560747663553], "iscrowd": 0, "area": 44622.6049, "rle": {"size": [428, 640], "counts": "d_W67h<a0@?G:K4M4K4M4K100O101N100O100O2O0O10001O0000001O0001O01O000000001O01O0001O0000001O01O01O00001O001O01O01O0ZISNm1n1gMeNR2Z1bMZOc29RM;R3YObLZ1Z3bNZLS2a3iMTLj2g3TMmK`3n3\\LfKX4U4fK_Km4]4oJXK`5`4dJTKi5f4b1K5J6J6K5J6J6J7H7I7H8I7K5M3M3M3L4M3M4L3M3L3N2N2N2N2N3L3N2N2N2N2M3N2N2N2N2M3N2N2O1000001O000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000O1000000O"}, "label": "green colored car is parked beside the road"}]}
{"id": 330993, "image": "COCO_train2014_000000330993.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green vehicle parked on side of road\"."}], "task": "refering", "answer_template": "The \"a green vehicle parked on side of road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.745265625, 0.0034813084112149534, 1.0, 0.9870560747663553], "iscrowd": 0, "area": 44622.6049, "rle": {"size": [428, 640], "counts": "d_W67h<a0@?G:K4M4K4M4K100O101N100O100O2O0O10001O0000001O0001O01O000000001O01O0001O0000001O01O01O00001O001O01O01O0ZISNm1n1gMeNR2Z1bMZOc29RM;R3YObLZ1Z3bNZLS2a3iMTLj2g3TMmK`3n3\\LfKX4U4fK_Km4]4oJXK`5`4dJTKi5f4b1K5J6J6K5J6J6J7H7I7H8I7K5M3M3M3L4M3M4L3M3L3N2N2N2N2N3L3N2N2N2N2M3N2N2N2N2M3N2N2O1000001O000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000O1000000O"}, "label": "a green vehicle parked on side of road"}]}
{"id": 183538, "image": "COCO_train2014_000000183538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair next to the table and behind another chair\"."}], "task": "refering", "answer_template": "The \"chair next to the table and behind another chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295], "bbox": [0.7073437499999999, 0.4981882352941176, 0.925203125, 0.8487529411764706], "iscrowd": 0, "area": 12384.970049999998, "rle": {"size": [425, 640], "counts": "U[l55l<8H7I8M3N1O2N2O0O2N2N3M6J6J6K3L2N2N2N2N2O1N2N2N2N2N2N101N2N2N2N2N1O2O1N2N2N2N2N1O2O1N2N2N2N1O200O1O100O001O100O1O1O010O1O1O100O00100O1O1O10O01O100O1O100O010O10000O10O10000000000O10000000000O100M3F:F9G:F:F:F:M300O100O100O100O100O1N2N2N1O2N2N2M3N2N2N2N2N2N2N2M3N2N4L4L3M4L4L4L4Kbfc0"}, "label": "chair next to the table and behind another chair"}]}
{"id": 183538, "image": "COCO_train2014_000000183538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table that a camera is sitting on\"."}], "task": "refering", "answer_template": "The \"table that a camera is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.7974117647058823, 1.0, 1.0], "iscrowd": 0, "area": 41973.7918, "rle": {"size": [425, 640], "counts": "f<c0l;j000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000000000l0TO"}, "label": "table that a camera is sitting on"}]}
{"id": 183538, "image": "COCO_train2014_000000183538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with sun on it that is brighter than the one behind it\"."}], "task": "refering", "answer_template": "The \"a chair with sun on it that is brighter than the one behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 266, 267, 268, 269, 290, 291, 314], "bbox": [0.53184375, 0.5169882352941176, 0.77696875, 0.8727058823529411], "iscrowd": 0, "area": 11983.180200000004, "rle": {"size": [425, 640], "counts": "mi]42V=2N1O2O1N1O2N2N2N1O2N2O0O2N2N2N1O2N2N1O01O0000000O10000000000000000O10001O000000000O11O001XEUOQ9l0kFYOS9g0jF^OT9c0hFBV9>gFGW9:eFKY96cFO[91bF4\\9M`F8^9H_F=_9D]Fa0b9_O[Fd0d9\\OZFh0d9ZOYFi0e9_1N2N2N2O1N2O1N2O1O0O2O1N2O1O1N2O1N2O1O1N101O1N2O1O1O1O1N2O1O1O000O10000000O0100000000000J6E;I7L4M3M3M3N2O1N2N2O1N2O1L4I7O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N3M2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O2M2N4L4L4MTTk1"}, "label": "a chair with sun on it that is brighter than the one behind it"}]}
{"id": 371955, "image": "COCO_train2014_000000371955.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batter\"."}], "task": "refering", "answer_template": "The \"a batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 209, 210, 211, 212, 232, 233, 234, 235, 257, 258, 280, 281, 304, 305], "bbox": [0.109546875, 0.303375, 0.315171875, 0.8089791666666667], "iscrowd": 0, "area": 12825.089599999996, "rle": {"size": [480, 640], "counts": "nRQ13k>8H9G7I3M3M3L4L4L4L4L3K6H6K6O1O001O001N101O1O0O2N2N\\OhCdNW<X1nCgNQ<W1SDiNk;Y1VDfNh;[1ZDdNd;]1]DbNb;`1_D_N_;b1cD]NZ;f1fDZNX;g1lDUNR;m1REmMo:U2k02O2N1N2O2N1O3M?@`0A5K4L5K5K5K5J6K5L6J5L5J5L3L4M3L4M3L4M3L4M2M4M3L4M3M3M00001O001OgKXHZN:]1^78aHdMg0n1g6=\\J\\Od5c0dJWO\\5g0lJROU5l0SKmNn4Q1[KhNf4U1bKdN_4Z1iK`NW4^1QL[NP4c1RL]No3`1SL_Nn3`1SL_Nm3a1TL_Nj3b1WL]Nh3d1YL\\Ne3f1\\LXNc3i1^LVNa3k1`LTN`3l1aLSN^3n1cLQN\\3o1fLoMZ3Q2hLnMW3R2lLlMS3T2oLkMQ3T2RMiMn2W2m400000001iNZC4e<M]C0c<0_CNa<2bCK^<6cCH]<8fCD\\<;fCB\\<>fC^O\\<a0fC]O[<c0gCYO\\<e0fCXO\\<g0gCVOZ<i0hCTOZ<k0o0M2N4M2M4L3M4M2M4LSU]6"}, "label": "a batter"}]}
{"id": 371955, "image": "COCO_train2014_000000371955.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player that has the number 15\"."}], "task": "refering", "answer_template": "The \"a baseball player that has the number 15\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 209, 210, 211, 212, 232, 233, 234, 235, 257, 258, 280, 281, 304, 305], "bbox": [0.109546875, 0.303375, 0.315171875, 0.8089791666666667], "iscrowd": 0, "area": 12825.089599999996, "rle": {"size": [480, 640], "counts": "nRQ13k>8H9G7I3M3M3L4L4L4L4L3K6H6K6O1O001O001N101O1O0O2N2N\\OhCdNW<X1nCgNQ<W1SDiNk;Y1VDfNh;[1ZDdNd;]1]DbNb;`1_D_N_;b1cD]NZ;f1fDZNX;g1lDUNR;m1REmMo:U2k02O2N1N2O2N1O3M?@`0A5K4L5K5K5K5J6K5L6J5L5J5L3L4M3L4M3L4M3L4M2M4M3L4M3M3M00001O001OgKXHZN:]1^78aHdMg0n1g6=\\J\\Od5c0dJWO\\5g0lJROU5l0SKmNn4Q1[KhNf4U1bKdN_4Z1iK`NW4^1QL[NP4c1RL]No3`1SL_Nn3`1SL_Nm3a1TL_Nj3b1WL]Nh3d1YL\\Ne3f1\\LXNc3i1^LVNa3k1`LTN`3l1aLSN^3n1cLQN\\3o1fLoMZ3Q2hLnMW3R2lLlMS3T2oLkMQ3T2RMiMn2W2m400000001iNZC4e<M]C0c<0_CNa<2bCK^<6cCH]<8fCD\\<;fCB\\<>fC^O\\<a0fC]O[<c0gCYO\\<e0fCXO\\<g0gCVOZ<i0hCTOZ<k0o0M2N4M2M4L3M4M2M4LSU]6"}, "label": "a baseball player that has the number 15"}]}
{"id": 371955, "image": "COCO_train2014_000000371955.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt no number\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt no number\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 354, 355, 356, 357, 377], "bbox": [0.38932812499999997, 0.5932499999999999, 0.608421875, 0.9865208333333333], "iscrowd": 0, "area": 13257.592849999995, "rle": {"size": [480, 640], "counts": "fUe32h>6J7I6J6J6J6K5K5K5K5K5K5K5K5K6K3M3N1O2NkN\\DROc;T1]DgNb;[1aDbN\\;`1iD[NV;f1oDRNS;n1n01O1ON2I6J711N2O1N2N101O`CfN];Y1cDoNV;P1kDVOo:i0QE[Ok:e0VE^Og:a0YEDc:<]EG`:a0WECf:^2L3N101O0O2O001N101N2O001N101O0O2O0O100O1O10OA`000O1O010O100O00100O100O00100O10O101O000O2O00001N100O2N1N2N3N1N2O2M2O2`LYEc2i:\\MXEb2j:\\MYEb2l:YMVEd2P;WMQEd2Y;UMjDb2o;H7L5K5K4M4K5L3L5L4L3102N2N2O1O1N2O1O1O1O1O118HN2I7J6J6J7H8I<DlXe3"}, "label": "a man in a black shirt no number"}]}
{"id": 371955, "image": "COCO_train2014_000000371955.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire\"."}], "task": "refering", "answer_template": "The \"the umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 354, 355, 356, 357, 377], "bbox": [0.38932812499999997, 0.5932499999999999, 0.608421875, 0.9865208333333333], "iscrowd": 0, "area": 13257.592849999995, "rle": {"size": [480, 640], "counts": "fUe32h>6J7I6J6J6J6K5K5K5K5K5K5K5K5K6K3M3N1O2NkN\\DROc;T1]DgNb;[1aDbN\\;`1iD[NV;f1oDRNS;n1n01O1ON2I6J711N2O1N2N101O`CfN];Y1cDoNV;P1kDVOo:i0QE[Ok:e0VE^Og:a0YEDc:<]EG`:a0WECf:^2L3N101O0O2O001N101N2O001N101O0O2O0O100O1O10OA`000O1O010O100O00100O100O00100O10O101O000O2O00001N100O2N1N2N3N1N2O2M2O2`LYEc2i:\\MXEb2j:\\MYEb2l:YMVEd2P;WMQEd2Y;UMjDb2o;H7L5K5K4M4K5L3L5L4L3102N2N2O1O1N2O1O1O1O1O118HN2I7J6J6J7H8I<DlXe3"}, "label": "the umpire"}]}
{"id": 527597, "image": "COCO_train2014_000000527597.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant standing next to its mother\"."}], "task": "refering", "answer_template": "The \"a baby elephant standing next to its mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 288, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351, 352, 353, 354, 355, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.12857812500000002, 0.19220833333333334, 0.56103125, 0.9870208333333333], "iscrowd": 0, "area": 69412.15209999999, "rle": {"size": [480, 640], "counts": "e[W12l45n4LPK<h4FUKb0d4@ZKh0_4YO_Kn0Z4TOcKS1X4oNfKU1U4mNjKW1R4jNmKZ1n3hNRL[1j3fNUL^1f3dNYL`1b3bN]Lb1_3_NaLd1Z3^NeLf1W3[NhLi1S3YNlLi1k2_NUMb1`2hN_MY1V2ROiMP1l1ZOSNh0b1B^N>W1MhN5o03PONj08UOJc0=]OC=c0B_O7g0H[O1k0NVOKQ15PODW1:jN@\\1?fN\\O^1c0dNYO_1g0aNWOa1h0aNTOb1k0_NROd1m0^NoNe1Q1\\NkNg1U1YNiNi1V1YNfNj1Z1VNcNm1]1TN^NP2b1QNYNS2f1PNTNT2l1TNhMP2X2XN\\Ml1c2]NhLP2X3l40000O10001O0O1000000O10K5dN[1J7N2N2N1N3NRLbF^2\\9bMgF]2W9cMkF]2S9cMPG\\2n8dMUG[2h8fMZGY2e8gM^GX2`8iMbGV2\\8jMfGV2X8jMkGU2R8lMQHS2m7mMVHR2h7nMZHQ2e7oM^HP2`7QNbHn1o7_MSHa2a8iLbGV3T:000O0100000000000000O1000000000000O1000000O1000000O1000000O10O1000O10000O100000eLZ3100000000000000000000000001O00000000000^LiG;W8YORHj0n7jN\\HX1d7\\NfHf1Z7nMPIT2l92N2M3N2N2001O1O001O1O100O00100O1O100O1O100O2O0O1O100O2N100O1OJ7A>B>010O2O001O00001O00000O1000O1000000000000O10O10000000SMm2WOi0WOi0YOg01O2M2O2N1O1O2N1O1O2N0001O1O1O100O1O1O00100O1O1O2iKTHU1m7cNUIa0n6WO^Ic0c6UOfIh0]:L4L4M2M3M100O1O1O100O001O010O001O10O0O1O1N3N1O1N2002L3N3L3N3L3N3L3N3Lg\\S4"}, "label": "a baby elephant standing next to its mother"}]}
{"id": 527597, "image": "COCO_train2014_000000527597.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant next to larger elephant\"."}], "task": "refering", "answer_template": "The \"baby elephant next to larger elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 288, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351, 352, 353, 354, 355, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.12857812500000002, 0.19220833333333334, 0.56103125, 0.9870208333333333], "iscrowd": 0, "area": 69412.15209999999, "rle": {"size": [480, 640], "counts": "e[W12l45n4LPK<h4FUKb0d4@ZKh0_4YO_Kn0Z4TOcKS1X4oNfKU1U4mNjKW1R4jNmKZ1n3hNRL[1j3fNUL^1f3dNYL`1b3bN]Lb1_3_NaLd1Z3^NeLf1W3[NhLi1S3YNlLi1k2_NUMb1`2hN_MY1V2ROiMP1l1ZOSNh0b1B^N>W1MhN5o03PONj08UOJc0=]OC=c0B_O7g0H[O1k0NVOKQ15PODW1:jN@\\1?fN\\O^1c0dNYO_1g0aNWOa1h0aNTOb1k0_NROd1m0^NoNe1Q1\\NkNg1U1YNiNi1V1YNfNj1Z1VNcNm1]1TN^NP2b1QNYNS2f1PNTNT2l1TNhMP2X2XN\\Ml1c2]NhLP2X3l40000O10001O0O1000000O10K5dN[1J7N2N2N1N3NRLbF^2\\9bMgF]2W9cMkF]2S9cMPG\\2n8dMUG[2h8fMZGY2e8gM^GX2`8iMbGV2\\8jMfGV2X8jMkGU2R8lMQHS2m7mMVHR2h7nMZHQ2e7oM^HP2`7QNbHn1o7_MSHa2a8iLbGV3T:000O0100000000000000O1000000000000O1000000O1000000O1000000O10O1000O10000O100000eLZ3100000000000000000000000001O00000000000^LiG;W8YORHj0n7jN\\HX1d7\\NfHf1Z7nMPIT2l92N2M3N2N2001O1O001O1O100O00100O1O100O1O100O2O0O1O100O2N100O1OJ7A>B>010O2O001O00001O00000O1000O1000000000000O10O10000000SMm2WOi0WOi0YOg01O2M2O2N1O1O2N1O1O2N0001O1O1O100O1O1O00100O1O1O2iKTHU1m7cNUIa0n6WO^Ic0c6UOfIh0]:L4L4M2M3M100O1O1O100O001O010O001O10O0O1O1N3N1O1N2002L3N3L3N3L3N3L3N3Lg\\S4"}, "label": "baby elephant next to larger elephant"}]}
{"id": 232691, "image": "COCO_train2014_000000232691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a striped shirt flying a kite\"."}], "task": "refering", "answer_template": "The \"a man in a striped shirt flying a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 260, 261, 262, 263, 277, 278, 279, 294, 295, 296, 311, 312, 313, 328, 329, 330, 345, 346, 347, 362, 363, 379, 380, 381], "bbox": [0.3310833333333333, 0.6103125, 0.5237083333333333, 0.989546875], "iscrowd": 0, "area": 9882.687749999996, "rle": {"size": [640, 480], "counts": "\\jS3g0]4Bo9k0jEVOo9V1kEkNm9a1nE`Nk9l1fDnMgN7[<X2ZDTNVOEf;IhCj2o0TN[O]Om;Z3\\DcM]OXOU<b3nCaMBRO_<d6O11O000000000000000000000000000000001O00000O101O0O1aI_C_5k<SJZCl5]=lN\\BkKf=Q4dBhK]=T4lBfKT=a3YB_Ll0Im<f3eBPLd08j<d3mCULZ<g3Q2N1TN^@YOd?b0d@YO^?b0i@ZOY?@i_OI7OQ1d0o>DT32M3N2O1O1OCl\\O0Rc00S]OMlb03U]OMjb04V]OLib04X]OLib03V]ONkb00U]O1lb0MT]O4lb0LS]O4ob0HS]O8Zc000001O0O10001O0000001O0000001O01O1O1O1O100O1O100O1O10gf^4"}, "label": "a man in a striped shirt flying a kite"}]}
{"id": 232691, "image": "COCO_train2014_000000232691.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 260, 261, 262, 263, 277, 278, 279, 294, 295, 296, 311, 312, 313, 328, 329, 330, 345, 346, 347, 362, 363, 379, 380, 381], "bbox": [0.3310833333333333, 0.6103125, 0.5237083333333333, 0.989546875], "iscrowd": 0, "area": 9882.687749999996, "rle": {"size": [640, 480], "counts": "\\jS3g0]4Bo9k0jEVOo9V1kEkNm9a1nE`Nk9l1fDnMgN7[<X2ZDTNVOEf;IhCj2o0TN[O]Om;Z3\\DcM]OXOU<b3nCaMBRO_<d6O11O000000000000000000000000000000001O00000O101O0O1aI_C_5k<SJZCl5]=lN\\BkKf=Q4dBhK]=T4lBfKT=a3YB_Ll0Im<f3eBPLd08j<d3mCULZ<g3Q2N1TN^@YOd?b0d@YO^?b0i@ZOY?@i_OI7OQ1d0o>DT32M3N2O1O1OCl\\O0Rc00S]OMlb03U]OMjb04V]OLib04X]OLib03V]ONkb00U]O1lb0MT]O4lb0LS]O4ob0HS]O8Zc000001O0O10001O0000001O0000001O01O1O1O1O100O1O100O1O10gf^4"}, "label": "man"}]}
{"id": 576758, "image": "COCO_train2014_000000576758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long white train\"."}], "task": "refering", "answer_template": "The \"a long white train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 24, 25, 46, 47, 48, 49, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324], "bbox": [0.0, 0.027523809523809523, 0.45421875, 0.9915714285714285], "iscrowd": 0, "area": 71182.2251, "rle": {"size": [420, 640], "counts": "g:Y2[6`4`K`4eN[11O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001N101N2O0O2N2O0O2N2O0O2O0O2N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N101N2N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101K5F9F;E:F;F:E:Fln^4"}, "label": "a long white train"}]}
{"id": 576758, "image": "COCO_train2014_000000576758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue train is on the tracks\"."}], "task": "refering", "answer_template": "The \"the blue train is on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 24, 25, 46, 47, 48, 49, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324], "bbox": [0.0, 0.027523809523809523, 0.45421875, 0.9915714285714285], "iscrowd": 0, "area": 71182.2251, "rle": {"size": [420, 640], "counts": "g:Y2[6`4`K`4eN[11O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001N101N2O0O2N2O0O2N2O0O2O0O2N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N101N2N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101N2N101N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O2O1N101K5F9F;E:F;F:E:Fln^4"}, "label": "the blue train is on the tracks"}]}
{"id": 371960, "image": "COCO_train2014_000000371960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cow with a bird on it ' s back\"."}], "task": "refering", "answer_template": "The \"a white cow with a bird on it ' s back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 78, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 291, 292, 293, 309, 310, 311, 312, 314, 315, 316, 332, 333, 334, 335], "bbox": [0.39392187500000003, 0.18672853828306266, 0.8211562499999999, 0.9781438515081208], "iscrowd": 0, "area": 52657.05464999998, "rle": {"size": [431, 640], "counts": "aUZ3;T=001N2O1O1O1N200O100O00100O010O00100O010O01O001O1O001OQJD[JLV3?Y2OYJG]3:T29XJAd35m1f0VJZOl30h1P1TJUOT4Jb1\\1RJnN[4F;FXLe2\\3eM<KPLd2c3`M>2fKb2k3\\M>8_K`2R4XM?e3@[L`0e3@ZLa0g3^OYLa0i3^OWLa0j3_OVL`0k3@TL?n3ARL=P4CPL<Q4CPL;R4EmK;T4ElK9V4GjK8W4HhK7Z4HgK6[4JdK4_4LaKOd41\\KJh46XKFl4;TK@P5a0PKZOS5h0lJWOU5i0lJUOU5l0jJSOW5n0iJPOX5Q1hJmNX5T1hJkNY5V1gJhNZ5Y1eJfN\\5[1dJcN\\5_1dJ_N]5a1cJ^N^5c1bJ[N_5f1aJXN_5i1aJVN_5l1aJRN^5Q2aJnM^5T2cJeMa5^2_J\\Md5f2\\JTMg5P3YJjLi5Y3WJbLk5b3UJYLl5k3TJRLl5P4TJnKl5U4TJiKk5Y4VJdKk5^4TJ`Km5a4TJ\\Kn5e4QJZKo5h4QJUKQ6k4PJSKP6T1iIf17TMP6W1jId16UMP6W1kId15SMP6Y1nIb13SMl5_1RJ^12RMe5g1[JV10QM`5n1bJP1OQM_5o1dJo0LQMb5o1cJP1KPMc5o1dJP1IoLe5o1dJR1GnLf5o1dJS1FlLh5P2cJT1EkLi5o1dJU1DkLi5o1dJV1BkLm5l1bJY1AjLP6i1aJ]1_OiLS6f1`Ja1]OhLW6b1^Jf1[OhLY6^1^Jj1YOgL]6Z1\\Jn1WOhL`6V1[JR2UOgLd6R1YJW2SOfLi6m0VJ]2QOeLV9[3jFdLX9[3hFeLX9[3hFdLX9\\3hFdLX9]3hFbLY9^3;1O100O1N2M3N200O1TOYLPHg3m7`LnGa3n7gLmGZ3P8mLlGS3Q8TMjGm2R8[MjGe2S8]1M3M3N2N2N2M3N2N2N2M3N3M2N2M3N2M2N3M2N3N1N3M2N3M2O2N101O001N101O001N101O001O0100O100O100O10000O1ZK_J^1b5oMPKP2Q5dM[K\\2^5TLYKk3V701N100O10001O000000001O0000[NWLdIi3Q6PM\\IU3Y6dMTIa2_6^2J64LL4M3N2N2N2N2N2O12N2N2N3M1O001O0O101O001`K[IU2d6eMfIX2Y6aMRJ[2S6ZMWJc2Z6`LTJ^3P8N01O1O001O1O1\\MlKkKT4S4WLdKj3X4cL^K^3_4mLXKS3f4WMRKj2j4\\MSKe2j4aMRK_2k4gMRKZ2k4lMQKU2k4QNRKo1l4WNPKj1l4\\NQKe1h4dNUK\\1e4nNVKT1b4ZOWKg0b4X4G8I7J7L3M4L3M3L5L3M3M4L3M3L5L3M3M7F>^Oa0_Ob0^Oa0@ej_1"}, "label": "a white cow with a bird on it ' s back"}]}
{"id": 371960, "image": "COCO_train2014_000000371960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white cow with black bird on it ' s back butting heads with a brown cow\"."}], "task": "refering", "answer_template": "The \"white cow with black bird on it ' s back butting heads with a brown cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 78, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 291, 292, 293, 309, 310, 311, 312, 314, 315, 316, 332, 333, 334, 335], "bbox": [0.39392187500000003, 0.18672853828306266, 0.8211562499999999, 0.9781438515081208], "iscrowd": 0, "area": 52657.05464999998, "rle": {"size": [431, 640], "counts": "aUZ3;T=001N2O1O1O1N200O100O00100O010O00100O010O01O001O1O001OQJD[JLV3?Y2OYJG]3:T29XJAd35m1f0VJZOl30h1P1TJUOT4Jb1\\1RJnN[4F;FXLe2\\3eM<KPLd2c3`M>2fKb2k3\\M>8_K`2R4XM?e3@[L`0e3@ZLa0g3^OYLa0i3^OWLa0j3_OVL`0k3@TL?n3ARL=P4CPL<Q4CPL;R4EmK;T4ElK9V4GjK8W4HhK7Z4HgK6[4JdK4_4LaKOd41\\KJh46XKFl4;TK@P5a0PKZOS5h0lJWOU5i0lJUOU5l0jJSOW5n0iJPOX5Q1hJmNX5T1hJkNY5V1gJhNZ5Y1eJfN\\5[1dJcN\\5_1dJ_N]5a1cJ^N^5c1bJ[N_5f1aJXN_5i1aJVN_5l1aJRN^5Q2aJnM^5T2cJeMa5^2_J\\Md5f2\\JTMg5P3YJjLi5Y3WJbLk5b3UJYLl5k3TJRLl5P4TJnKl5U4TJiKk5Y4VJdKk5^4TJ`Km5a4TJ\\Kn5e4QJZKo5h4QJUKQ6k4PJSKP6T1iIf17TMP6W1jId16UMP6W1kId15SMP6Y1nIb13SMl5_1RJ^12RMe5g1[JV10QM`5n1bJP1OQM_5o1dJo0LQMb5o1cJP1KPMc5o1dJP1IoLe5o1dJR1GnLf5o1dJS1FlLh5P2cJT1EkLi5o1dJU1DkLi5o1dJV1BkLm5l1bJY1AjLP6i1aJ]1_OiLS6f1`Ja1]OhLW6b1^Jf1[OhLY6^1^Jj1YOgL]6Z1\\Jn1WOhL`6V1[JR2UOgLd6R1YJW2SOfLi6m0VJ]2QOeLV9[3jFdLX9[3hFeLX9[3hFdLX9\\3hFdLX9]3hFbLY9^3;1O100O1N2M3N200O1TOYLPHg3m7`LnGa3n7gLmGZ3P8mLlGS3Q8TMjGm2R8[MjGe2S8]1M3M3N2N2N2M3N2N2N2M3N3M2N2M3N2M2N3M2N3N1N3M2N3M2O2N101O001N101O001N101O001O0100O100O100O10000O1ZK_J^1b5oMPKP2Q5dM[K\\2^5TLYKk3V701N100O10001O000000001O0000[NWLdIi3Q6PM\\IU3Y6dMTIa2_6^2J64LL4M3N2N2N2N2N2O12N2N2N3M1O001O0O101O001`K[IU2d6eMfIX2Y6aMRJ[2S6ZMWJc2Z6`LTJ^3P8N01O1O001O1O1\\MlKkKT4S4WLdKj3X4cL^K^3_4mLXKS3f4WMRKj2j4\\MSKe2j4aMRK_2k4gMRKZ2k4lMQKU2k4QNRKo1l4WNPKj1l4\\NQKe1h4dNUK\\1e4nNVKT1b4ZOWKg0b4X4G8I7J7L3M4L3M3L5L3M3M4L3M3L5L3M3M7F>^Oa0_Ob0^Oa0@ej_1"}, "label": "white cow with black bird on it ' s back butting heads with a brown cow"}]}
{"id": 371960, "image": "COCO_train2014_000000371960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow\"."}], "task": "refering", "answer_template": "The \"a brown cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 280, 281, 283, 284, 285, 300, 301, 304, 307, 308, 323, 327, 330, 331], "bbox": [0.001453125, 0.33763341067285385, 0.45764062499999997, 0.9892575406032482], "iscrowd": 0, "area": 53223.010149999995, "rle": {"size": [431, 640], "counts": "Sb0X6W7000001O00000O2N1000001O0000001O0000001O0000001O0000lJSJg2l5VMWJj2i5RM[Jn2e5nL_JR3a5kLbJV3\\5gLhJY3X5dLkJ\\3U5`LoJ`3Q5]LRKc3n4YLVKg3j4VLYKj3g4RL]Kn3c4oK`KQ4`4kKdKU4\\4hKgKX4Y4dKkK\\4U4aKnK_4R4]KRLc4n3ZKULf4k3VKYLj4g3SK\\Lm4d3oJ`LQ5o500000N2M3M3M3M3M3M3M3M3M3M3L4M4M2M3N2M3N1N3M3N2N2N2N101N2N2N2N2N1O2O1N2N2N2N1010000000O10000000000O10000000000O100000001O0O2O00001O001N10001O001O0O101O001O00001O001O00001O002N4L5J5L5K4L5K5K6J6J6J6J6J6J6J3M00O100O100O10000O1ZNYKdJg4Y5gK[JZ4c5RLSJn3j5`LjIb3R6kLdIU3Z6Q2M3M3N2M300000001O001O0010O01O001O001O002N3M2N3M3M3N2M3M3M3M3M3L4M3M3M3M2N1O1N3N1O1O2N1O16J3NO00010O0000001O00001O000010O000001O00001O000010O00003N;D5KO2N1O100O2N1O1O2N1O1O1O10O0001N100O2O001N100O2O0O101N100O1[OjFQMV9l2QGoLP9o2WGlLj8Q3k0N2N3L3N2G:@?H8L5K4iNXDb0m;]OZD9k;F[D0i;Om0O1N2O1N2Olha4"}, "label": "a brown cow"}]}
{"id": 371960, "image": "COCO_train2014_000000371960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bull\"."}], "task": "refering", "answer_template": "The \"a brown bull\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 280, 281, 283, 284, 285, 300, 301, 304, 307, 308, 323, 327, 330, 331], "bbox": [0.001453125, 0.33763341067285385, 0.45764062499999997, 0.9892575406032482], "iscrowd": 0, "area": 53223.010149999995, "rle": {"size": [431, 640], "counts": "Sb0X6W7000001O00000O2N1000001O0000001O0000001O0000001O0000lJSJg2l5VMWJj2i5RM[Jn2e5nL_JR3a5kLbJV3\\5gLhJY3X5dLkJ\\3U5`LoJ`3Q5]LRKc3n4YLVKg3j4VLYKj3g4RL]Kn3c4oK`KQ4`4kKdKU4\\4hKgKX4Y4dKkK\\4U4aKnK_4R4]KRLc4n3ZKULf4k3VKYLj4g3SK\\Lm4d3oJ`LQ5o500000N2M3M3M3M3M3M3M3M3M3M3L4M4M2M3N2M3N1N3M3N2N2N2N101N2N2N2N2N1O2O1N2N2N2N1010000000O10000000000O10000000000O100000001O0O2O00001O001N10001O001O0O101O001O00001O001O00001O002N4L5J5L5K4L5K5K6J6J6J6J6J6J6J3M00O100O100O10000O1ZNYKdJg4Y5gK[JZ4c5RLSJn3j5`LjIb3R6kLdIU3Z6Q2M3M3N2M300000001O001O0010O01O001O001O002N3M2N3M3M3N2M3M3M3M3M3L4M3M3M3M2N1O1N3N1O1O2N1O16J3NO00010O0000001O00001O000010O000001O00001O000010O00003N;D5KO2N1O100O2N1O1O2N1O1O1O10O0001N100O2O001N100O2O0O101N100O1[OjFQMV9l2QGoLP9o2WGlLj8Q3k0N2N3L3N2G:@?H8L5K4iNXDb0m;]OZD9k;F[D0i;Om0O1N2O1N2Olha4"}, "label": "a brown bull"}]}
{"id": 544001, "image": "COCO_train2014_000000544001.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player in white\"."}], "task": "refering", "answer_template": "The \"a tennis player in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 214, 215, 216, 237, 238, 259, 260, 261, 282, 283, 284, 285], "bbox": [0.27715625, 0.39617924528301884, 0.47279687499999995, 0.874929245283019], "iscrowd": 0, "area": 8072.824600000003, "rle": {"size": [424, 640], "counts": "ocY22n<>I8I5J101N101N101N101N2O0O2O0O2O0O1OK5K5L400O1ORNUObGk0^8XO_Gg0b8]OZGb0g8AVG?k8BSG=n8EPG:Q9HmF8S9JkF5W9LgF3Z9OdF1\\91aFO`93^FMc94[FMd95XFMi94SFOl93oE0\\9TObFn0K3\\9\\OcFb0O4Z9C_F<52[9m0cFSO]9P1`FQO_9Q1_FPO`9\\2O001O1O1O2N1C>O8I6I7J6I7J5J2O1N3N1N3N1N2O0O10]MZJdMe5Z2`JeM_5X2fJfM[5V2kJgMV5T2PKjMQ5R2UKkMl4R2XKlMi4S2ZKkMi4R2XKlMP5l1SKQNm4o1TKPNl4m0\\HHj2YOk4k0bHIc2[Ok4i0iHH^2]Oi4j0lHG]2]Og4k0PIEZ2_Of4l0RIDR:;RFBo9<TF^OP:a0TFWOQ:i0W1O1O101N1O100O1010O001O1O0010O01000O10000O2O0O10000O10001N100O10000O9H\\Z[4"}, "label": "a tennis player in white"}]}
{"id": 544001, "image": "COCO_train2014_000000544001.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a tennis racquet\"."}], "task": "refering", "answer_template": "The \"a woman with a tennis racquet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 214, 215, 216, 237, 238, 259, 260, 261, 282, 283, 284, 285], "bbox": [0.27715625, 0.39617924528301884, 0.47279687499999995, 0.874929245283019], "iscrowd": 0, "area": 8072.824600000003, "rle": {"size": [424, 640], "counts": "ocY22n<>I8I5J101N101N101N101N2O0O2O0O2O0O1OK5K5L400O1ORNUObGk0^8XO_Gg0b8]OZGb0g8AVG?k8BSG=n8EPG:Q9HmF8S9JkF5W9LgF3Z9OdF1\\91aFO`93^FMc94[FMd95XFMi94SFOl93oE0\\9TObFn0K3\\9\\OcFb0O4Z9C_F<52[9m0cFSO]9P1`FQO_9Q1_FPO`9\\2O001O1O1O2N1C>O8I6I7J6I7J5J2O1N3N1N3N1N2O0O10]MZJdMe5Z2`JeM_5X2fJfM[5V2kJgMV5T2PKjMQ5R2UKkMl4R2XKlMi4S2ZKkMi4R2XKlMP5l1SKQNm4o1TKPNl4m0\\HHj2YOk4k0bHIc2[Ok4i0iHH^2]Oi4j0lHG]2]Og4k0PIEZ2_Of4l0RIDR:;RFBo9<TF^OP:a0TFWOQ:i0W1O1O101N1O100O1010O001O1O0010O01000O10000O2O0O10000O10001N100O10000O9H\\Z[4"}, "label": "a woman with a tennis racquet"}]}
{"id": 52484, "image": "COCO_train2014_000000052484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of pizza nearest to the soda can\"."}], "task": "refering", "answer_template": "The \"slice of pizza nearest to the soda can\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 109, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241], "bbox": [0.3961875, 0.19633333333333333, 0.877125, 0.6115833333333334], "iscrowd": 0, "area": 25217.7409, "rle": {"size": [480, 640], "counts": "S[g31n>1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O001O1O001O1O1O001O1O10O01O1O001O1O1O001O1O1O010O1N101O1O1O001O1O1O001O1O001O1O1N101O1O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1O001O1O1O0O2O1O1O001O1O1O001O1O1O001O1O1O00100O1O00100O1O001O100O001O10O01O1O10O01O1O10O01O1O1O010O1O1O010O1O1O010O1O1O00100O001O100O001O100M2O2N2N2M2O2N2M3N1O2N2M3N1O2M3N2N1O2M31N101O001N2O001O0O2O001O1N101O0O2O001N2O0O2O001N2O001N101N2O001N2O001N2O001N101N2O001N2O001N2O0O2O1O0O2O1O0O2O1N101O1N101O1N2O1N2O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N101N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N101N2O1N2O2M3N2N2M2O2M3N2M3N2M3N2NfYU1"}, "label": "slice of pizza nearest to the soda can"}]}
{"id": 52484, "image": "COCO_train2014_000000052484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza slices on the blue and white plate\"."}], "task": "refering", "answer_template": "The \"pizza slices on the blue and white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.37940625, 0.454125, 0.991765625, 0.8094166666666667], "iscrowd": 0, "area": 36562.55530000001, "rle": {"size": [480, 640], "counts": "gVb39g>2M2O2M2O2M2N2O1N2O1O1O00100O001O1O1O001O10O01O1O001O001O10O01O001O1O001O010O001O1O001O00100O001O001O001O01O0001O0O101O0000001N10001O000O101O00001O0O10001O00000O2O00001O000O101O00001O0O10001O0O101O0O10001N10001O0O10000O10000O10000O10000O10000O10000O10000O10000O2O000O1000000O10000O10000O10000O10000O10000O10000O10000O100O100O100O100O101N100O1O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O101N100O100O100O100O100O100O10000O100O100O100O100O100O100O101N100O10000O100O100O100O2O0O100O100O100O1O100O1O101N100O1O100O1O100O1O100O2N100O1O100O1O100O1O10001O00000000000000000O1000001O00000002N2N1O2N1O2N1O2O1N1O2N1O2N1O2N2N1O2N1O2N2O0O1O1O2N1O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1O2N100O1O2N1O2N2N2N2N2N2N2N2O1M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2M3I7H8Ijo1"}, "label": "pizza slices on the blue and white plate"}]}
{"id": 52484, "image": "COCO_train2014_000000052484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of pizza nearest to the front of the picture\"."}], "task": "refering", "answer_template": "The \"the slice of pizza nearest to the front of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.37940625, 0.454125, 0.991765625, 0.8094166666666667], "iscrowd": 0, "area": 36562.55530000001, "rle": {"size": [480, 640], "counts": "gVb39g>2M2O2M2O2M2N2O1N2O1O1O00100O001O1O1O001O10O01O1O001O001O10O01O001O1O001O010O001O1O001O00100O001O001O001O01O0001O0O101O0000001N10001O000O101O00001O0O10001O00000O2O00001O000O101O00001O0O10001O0O101O0O10001N10001O0O10000O10000O10000O10000O10000O10000O10000O10000O2O000O1000000O10000O10000O10000O10000O10000O10000O10000O100O100O100O100O101N100O1O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O101N100O100O100O100O100O100O10000O100O100O100O100O100O100O101N100O10000O100O100O100O2O0O100O100O100O1O100O1O101N100O1O100O1O100O1O100O2N100O1O100O1O100O1O10001O00000000000000000O1000001O00000002N2N1O2N1O2N1O2O1N1O2N1O2N1O2N2N1O2N1O2N2O0O1O1O2N1O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1O2N100O1O2N1O2N2N2N2N2N2N2N2O1M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2M3I7H8Ijo1"}, "label": "the slice of pizza nearest to the front of the picture"}]}
{"id": 552199, "image": "COCO_train2014_000000552199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table with all the food on it\"."}], "task": "refering", "answer_template": "The \"the table with all the food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 248, 249, 250, 251, 260, 261, 270, 271, 272, 283, 284, 285, 286, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.307359375, 0.6243559718969556, 0.9415625000000001, 0.9883840749414521], "iscrowd": 0, "area": 18402.637200000005, "rle": {"size": [427, 640], "counts": "U_b2<n<4M2N1O1O1N3N4Ln0ROc0]O=B1000000001O00001O00001O001O00001O00001O001O00001O00001O00001O000000001O00000000001O000000001O000000001O00000000001O0000001O001O001O001O001O001O001O001O001O001O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O00001O0BZDWOf;i0?0O0010O0010O0010O01O01O01O010O00010O010O01O010O010O010O1O010O001O001O001O001O001O1O0000000000000000001O00000000000000000000000000O10000O10001N10000O1000000O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O01O01O0001O0000010O00001O001O100O1O001O1O100O001O1O1O1O010O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O010O1O1O1OI7N2N3M2N2N2O1N2O1100O001O1O1O10O01O1O1O001O10O0001O000010O0001O00001O01O01O00C=0001O0001O001O1O100O1O1O1O1O100O1O1O1O1O1OJPOSDP1m;TOoCk0Q<71O000010O01O00001O0010O01O1O001O1O010O1O1O0000O2N1O11O1O1O2N101N1O1O2N1O101N1O1O2N1O1O2O0O1O2N1O2N2O1N2N]b?"}, "label": "the table with all the food on it"}]}
{"id": 552199, "image": "COCO_train2014_000000552199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"selection of thanksgiving casseroles\"."}], "task": "refering", "answer_template": "The \"selection of thanksgiving casseroles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 335], "bbox": [0.3253125, 0.5299063231850117, 0.9036406249999999, 0.9429976580796252], "iscrowd": 0, "area": 32397.01345000001, "rle": {"size": [427, 640], "counts": "eQg23n<;F:J6O00001O001O001O00001O001N10001O001O001O00001O0000001O01O010O1O0010O01O1O010O0010O01O0001N10001N10001N101O0O2O001N101O0O2O0O101O0O2O001N101O0O2O001N101O0O2O0O2O000O2O001N101O0O2O001N101O0O2O001N100O2O001N101O0O2O001N101O0O2O001N100O2O001N101O0O2O001N10001N101O0O101O0O2O0O101O0O2O000O2O001N101O0O101O0O2O0O101O0O2O000O2O001N10001N101O0O101O0O2O001N01000O100O01000O1000O01000O010O1O010O1O00100O001O10O01O10O01O00100O00100O001O10O01O10O01O1O010O1O010O1O0010O01O10O01O10O01O10O01O10O01O10O01O10O0100O0100O01000O01000O01000O010000O1000O010000O10000O1N2O1N1O2O1N2O1N2N2O1N2O1N20O01O1O100O1O1O100O1O010O1O1O100O1O1O010O1O100O1O1M3N1N3N2M3N2M3O1000O1000000001O00000000001O00000O1O1O2M2O1O1O1O1O2N1O1O1O1O1O2N100O2N2N2N3M2O1N3M2N2N3M2O1N3L3Jfci0"}, "label": "selection of thanksgiving casseroles"}]}
{"id": 552199, "image": "COCO_train2014_000000552199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman reaching into a cupboard\"."}], "task": "refering", "answer_template": "The \"a woman reaching into a cupboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 213, 214, 215, 216, 237, 238, 239, 260], "bbox": [0.24923437499999998, 0.25981264637002344, 0.39826562499999996, 0.7709836065573772], "iscrowd": 0, "area": 12363.857600000001, "rle": {"size": [427, 640], "counts": "PnR29j<9N1O2M2N3K4L5K4L5K5L3M3M3L3O2O1O1O1O001O1O1fNnMZGS2d8TNVGm1h8ZNRGi1j8^NPGk1f8[NUGo1`8XNYGP2_8c1J6K4M3M4L3M6J7I8H9G5K5K1O1O1O001O100OO1O1N2H8M3N2N2N200O100O2O0O1eJSIW4n6bK[I\\4m7L3L3N2M3OYOXLlGh3Q8j0M3M30001O0001O0001O03]LXH^1l7YNdH[1`7\\NYIQ1i6gN_IV1e6cNaI[1g6ZN]Ie1e6UN`Ii1c6RNaIm1Q9M2N2N2YOhDWOZ;?oD[OX;:]ZP5"}, "label": "a woman reaching into a cupboard"}]}
{"id": 552199, "image": "COCO_train2014_000000552199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black top opening the cabinet\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black top opening the cabinet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 213, 214, 215, 216, 237, 238, 239, 260], "bbox": [0.24923437499999998, 0.25981264637002344, 0.39826562499999996, 0.7709836065573772], "iscrowd": 0, "area": 12363.857600000001, "rle": {"size": [427, 640], "counts": "PnR29j<9N1O2M2N3K4L5K4L5K5L3M3M3L3O2O1O1O1O001O1O1fNnMZGS2d8TNVGm1h8ZNRGi1j8^NPGk1f8[NUGo1`8XNYGP2_8c1J6K4M3M4L3M6J7I8H9G5K5K1O1O1O001O100OO1O1N2H8M3N2N2N200O100O2O0O1eJSIW4n6bK[I\\4m7L3L3N2M3OYOXLlGh3Q8j0M3M30001O0001O0001O03]LXH^1l7YNdH[1`7\\NYIQ1i6gN_IV1e6cNaI[1g6ZN]Ie1e6UN`Ii1c6RNaIm1Q9M2N2N2YOhDWOZ;?oD[OX;:]ZP5"}, "label": "a woman wearing a black top opening the cabinet"}]}
{"id": 44298, "image": "COCO_train2014_000000044298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with the red sweatshirt to the right\"."}], "task": "refering", "answer_template": "The \"the person with the red sweatshirt to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 100, 101, 118, 134, 135, 151, 152, 168, 169, 185, 186, 202, 203, 219, 220, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 305], "bbox": [0.8138124999999999, 0.167796875, 0.997, 0.7511249999999999], "iscrowd": 0, "area": 15409.506700000002, "rle": {"size": [640, 480], "counts": "Ujd72mc02M2O1O2M2O2M2O2M2O2M2O2M2O1N3N1O2M2O2M2O2M2O1N3L3N3M2M3N2M4M2M3N2M3N3L3N2M3`MeMYC^2c<nMlBZ2P=RN[B[2a=RNhA\\2U>j1K4M3L4M4K4M3K5I8SL`IQJg6m5dIdIb6Z6hIWI_6h6bIRId6l6_IlHi6Q7YIiHm6V7UIbHR7\\7QI`HR7_7oH`HS7]7PIaHQ7]7QIbHP7]7RIaHo6]7SIbHo6]7RIaHo6^7RIaHo6_7RI_Ho6a7RI]HP7b7RI[Ho6e7RIYHo6f7SIXHn6h7TIUHn6j7SITHn6l7SIQHo6o7RIoGo6Q8SIlGn6S8TIkGn6T8SIjGn6V8TIgGm6Y8TIeGm6[8TIcGn6[8\\2000001O`M_2fGPi0"}, "label": "the person with the red sweatshirt to the right"}]}
{"id": 44298, "image": "COCO_train2014_000000044298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man who is standing in front of the two girls\"."}], "task": "refering", "answer_template": "The \"a man who is standing in front of the two girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 100, 101, 118, 134, 135, 151, 152, 168, 169, 185, 186, 202, 203, 219, 220, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 305], "bbox": [0.8138124999999999, 0.167796875, 0.997, 0.7511249999999999], "iscrowd": 0, "area": 15409.506700000002, "rle": {"size": [640, 480], "counts": "Ujd72mc02M2O1O2M2O2M2O2M2O2M2O2M2O1N3N1O2M2O2M2O2M2O1N3L3N3M2M3N2M4M2M3N2M3N3L3N2M3`MeMYC^2c<nMlBZ2P=RN[B[2a=RNhA\\2U>j1K4M3L4M4K4M3K5I8SL`IQJg6m5dIdIb6Z6hIWI_6h6bIRId6l6_IlHi6Q7YIiHm6V7UIbHR7\\7QI`HR7_7oH`HS7]7PIaHQ7]7QIbHP7]7RIaHo6]7SIbHo6]7RIaHo6^7RIaHo6_7RI_Ho6a7RI]HP7b7RI[Ho6e7RIYHo6f7SIXHn6h7TIUHn6j7SITHn6l7SIQHo6o7RIoGo6Q8SIlGn6S8TIkGn6T8SIjGn6V8TIgGm6Y8TIeGm6[8TIcGn6[8\\2000001O`M_2fGPi0"}, "label": "a man who is standing in front of the two girls"}]}
{"id": 44298, "image": "COCO_train2014_000000044298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with her dress hiked up\"."}], "task": "refering", "answer_template": "The \"girl with her dress hiked up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 80, 81, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 382, 383, 384, 385], "bbox": [0.08729166666666667, 0.018625, 0.9932291666666667, 0.987015625], "iscrowd": 0, "area": 166457.77950000003, "rle": {"size": [640, 480], "counts": "Saj04hc07I7I6J7I7H8I7I7I6J7I7L4L4L4L3M3L3N2N3M2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2M3N2N2N3M2N2N2N2N2N3M2N3M4L3M4L3M4L3M4L3L4M4L4L4L4L4L4L3M3M2N3M3M3M3L3N3M3M3M3M2N3M3M3M3L3N3M3M3M3M3Ma1_N3M2M3N2N3M2N2N2O2M2O1N2O1N3N1N2O1O2M2O1N2O2M2O1N2O0O2O0O2O1O0O2O1N101N2O0O2O1N101N101N2O001N2O0O2O1N1C>A?@`0A>G:M3L3M4M3L3N3L4L4M2N3N2N1O2O1N2N1O2N2N1O2N2N101N2N2N1O2N2N1O2N2O0O2N2N2N3M4L5K4L4M4K4L5K4L4L5K4L4M3M3MO100000000000000O100000000000000O100000000000000O100000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000_Oa0E<M3M3M3M3O1O001O1O1O1O1O1O1O001O1O1O1O1X@hMf=Y2dA]N[>d1PAQOo>R3N1O2N1O1O2N1O1O2N101N1O1O2N1O2N1iHPK\\OQ53bKJ`4DSL:m3UOeLi0]3fNUMW1l2hNVMV1k2iNXMT1j2jNXMT1i2kNZMR1h2lN[MQ1f2mN]MQ1d2nN_Mo0c2oN_MP1a2oNbMn0`2PObMn0_2QOdMl0]2SOfMj0\\2SOgMk0\\2ROeMm0_2oNaMQ1b2lN_MS1e2iN\\MW1f2fN[MY1i2cNWM]1l2_NVM`1m2]NTMb1P3ZNPMf1S3WNnLh1V3TNkLk1X3RNiLm1Z3PNfLP2^3kMdLU2\\3jMeLU2\\3jMdLV2]3iMdLV2]3iMdLV2]3iMdLV2\\3iMeLW2\\3hMeLW2\\3hMeLW2\\3hMdLX2\\3hMeLW2\\3hMeLX2[3gMeLY2\\3fMeLY2[3fMgLY2Z3fMgLY2Z3fMfLZ2[3eMfLZ2Z3fMgLY2Z3fMfLZ2[3eMfLZ2[3eMfLZ2[3eMfL[2Y3eMgL[2Z3dMgL[2Z3dMgL[2Z3dMfL\\2Z3cMhL\\2Y3cMhL\\2Y3cMhL\\2Y3cMgL]2Y3cMhL\\2Y3cMhL\\2Y3cMgL^2Y3aMhL^2X3bMiL]2X3bMiL]2X3bMhL^2Y3aMhL^2Y3aMhL^2X3aMiL_2X3`MiL_2X3`MiL_2X3`MiL_2W3aMiL_2X3`MiL_2X3`MiL_2X3`MhL`2X3`MiL_2X3`MiL_2X3`MiL`2W3_MiLa2W3_MjL`2W3^MkLa2V3^MjLb2W3]MjLb2W3]MjLb2V3^MkLa2V3^MjLb2W3]MjLa2X3^MiL`2X3`MhL_2Z3`MgL^2[3aMfL]2\\3bMdL]2]3cMdL[2^3cMdL[2^3aMfL]2\\3cLcM\\3b9N2N1O7\\Oc0A`0G9F9G:G9F9H9F:F`[1"}, "label": "girl with her dress hiked up"}]}
{"id": 44298, "image": "COCO_train2014_000000044298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blond woman in a green velvet dress exposing a black garter\"."}], "task": "refering", "answer_template": "The \"blond woman in a green velvet dress exposing a black garter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 80, 81, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 382, 383, 384, 385], "bbox": [0.08729166666666667, 0.018625, 0.9932291666666667, 0.987015625], "iscrowd": 0, "area": 166457.77950000003, "rle": {"size": [640, 480], "counts": "Saj04hc07I7I6J7I7H8I7I7I6J7I7L4L4L4L3M3L3N2N3M2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2M3N2N2N3M2N2N2N2N2N3M2N3M4L3M4L3M4L3M4L3L4M4L4L4L4L4L4L3M3M2N3M3M3M3L3N3M3M3M3M2N3M3M3M3L3N3M3M3M3M3Ma1_N3M2M3N2N3M2N2N2O2M2O1N2O1N3N1N2O1O2M2O1N2O2M2O1N2O0O2O0O2O1O0O2O1N101N2O0O2O1N101N101N2O001N2O0O2O1N1C>A?@`0A>G:M3L3M4M3L3N3L4L4M2N3N2N1O2O1N2N1O2N2N1O2N2N101N2N2N1O2N2N1O2N2O0O2N2N2N3M4L5K4L4M4K4L5K4L4L5K4L4M3M3MO100000000000000O100000000000000O100000000000000O100000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000_Oa0E<M3M3M3M3O1O001O1O1O1O1O1O1O001O1O1O1O1X@hMf=Y2dA]N[>d1PAQOo>R3N1O2N1O1O2N1O1O2N101N1O1O2N1O2N1iHPK\\OQ53bKJ`4DSL:m3UOeLi0]3fNUMW1l2hNVMV1k2iNXMT1j2jNXMT1i2kNZMR1h2lN[MQ1f2mN]MQ1d2nN_Mo0c2oN_MP1a2oNbMn0`2PObMn0_2QOdMl0]2SOfMj0\\2SOgMk0\\2ROeMm0_2oNaMQ1b2lN_MS1e2iN\\MW1f2fN[MY1i2cNWM]1l2_NVM`1m2]NTMb1P3ZNPMf1S3WNnLh1V3TNkLk1X3RNiLm1Z3PNfLP2^3kMdLU2\\3jMeLU2\\3jMdLV2]3iMdLV2]3iMdLV2]3iMdLV2\\3iMeLW2\\3hMeLW2\\3hMeLW2\\3hMdLX2\\3hMeLW2\\3hMeLX2[3gMeLY2\\3fMeLY2[3fMgLY2Z3fMgLY2Z3fMfLZ2[3eMfLZ2Z3fMgLY2Z3fMfLZ2[3eMfLZ2[3eMfLZ2[3eMfL[2Y3eMgL[2Z3dMgL[2Z3dMgL[2Z3dMfL\\2Z3cMhL\\2Y3cMhL\\2Y3cMhL\\2Y3cMgL]2Y3cMhL\\2Y3cMhL\\2Y3cMgL^2Y3aMhL^2X3bMiL]2X3bMiL]2X3bMhL^2Y3aMhL^2Y3aMhL^2X3aMiL_2X3`MiL_2X3`MiL_2X3`MiL_2W3aMiL_2X3`MiL_2X3`MiL_2X3`MhL`2X3`MiL_2X3`MiL_2X3`MiL`2W3_MiLa2W3_MjL`2W3^MkLa2V3^MjLb2W3]MjLb2W3]MjLb2V3^MkLa2V3^MjLb2W3]MjLa2X3^MiL`2X3`MhL_2Z3`MgL^2[3aMfL]2\\3bMdL]2]3cMdL[2^3cMdL[2^3aMfL]2\\3cLcM\\3b9N2N1O7\\Oc0A`0G9F9G:G9F9H9F:F`[1"}, "label": "blond woman in a green velvet dress exposing a black garter"}]}
{"id": 44298, "image": "COCO_train2014_000000044298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women in a gold and black sequined dress standing behind a women with a huge green bow on her dress\"."}], "task": "refering", "answer_template": "The \"a women in a gold and black sequined dress standing behind a women with a huge green bow on her dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 23, 35, 36, 37, 38, 39, 40, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 120, 121, 136, 137, 138, 153, 154, 170, 171, 187, 188, 189, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 290, 291, 292, 307, 308, 309, 324, 325, 326, 341, 342, 343, 358, 359, 375, 376], "bbox": [0.008979166666666667, 0.0561875, 0.4015, 0.986515625], "iscrowd": 0, "area": 41188.319950000005, "rle": {"size": [640, 480], "counts": "[k21jc09F9G:G8G:G9F9gJPNZHV2b7jM\\H[2a7fM\\H`2`7`M^Hf2^7[M`Hi2]7WMaHo2Z7SMcHR3n6kLgD3Y4Z3i6fLlD1X4`3f6bLPENY4g3`6^LUELm21RNm3e9YLZEIi28UNn3b9[LQH0YNl3`9ULPH7]Nl3Z9QLRH;`Nh3X9QLQH?eNb3V9SLmGd0iN[3V9jLVG2AW3U9GgF;V9IgF9P8fJQIU5lN7R8dJPIf:P7[EmHh:R7YE]H0Jj:i7VEZH4Jm5KfNP8ZKWH7Ki51eNQ5XKeM4ZM9Lf57eNf4`KmMLXM]69WNX4jKVNU6eMPNl3TL_Ni5iMSN^3]LiN^5lMTNS3fLQOS5PNWNe2oL[Oh4RNYNZ2XMD\\4VN[Nm1bMLR4ZN\\Na1jM3i3_N]NU1SN9a3eN\\Nh0\\Na0i2_MbJ\\1X4<fNf0`2eM_J[1[41oNm0X2iMZJ]1^4EXOT1o1mMXJ]1`4YOB\\1f1HhNcNKd1\\1MgNVN6l1S10fNkM`0T2i04j3LUL6j3IUL;i3DoJmMeLb2[8@nJTNcL^2_8\\OmJ[N_L\\2c8YOmJ_N\\LZ2f8VOoJcNVL[2i8iNgF]N[4b0QLZ2l8hNjFXNZ4i0lKY2o8gNZKSOdKW2Q9fN[KVOaKU2S9eN\\KYO^KS2U9dN]Kb2b4]M^Ke2`4\\M`Kd2]4_McKb2Y4aMgK`2U4bMlK^2Q4eMoK\\2m3gMSLZ2i3hMXLY2d3jM\\LV2a3mM_LT2]3nMdLS2X3PNhLP2V3RNjLo1R3SNoLn1m2UNSMl1i2WNWMi1f2YNZMi1TL[K]5o2_Ng1_KQLl5[2eNi2V1ZMjNh2Q1ZMPOi2i0\\MVOf2e0]M[Of2>^MBd28aMGa24bMLa2MdM2^2HfM8]2BfM>\\2\\OiMc0Y2XOjMh0Y2QOkMo0W2kNlMV1W2cNmM\\1V2]NnMd1T2UNQNk1R2mMRNT2P2eMUN[2n1]MVNc2n1UMWNi2o1nLUNR3P2eLUNY3Q2^LSNb3R2ULSNi3S2nKQNR4T2eKPNZ4l8O0VOg@mL[?o2m@kLT?R3SAjLn>T3XAfLi>Y3P1O1001O001O001O00001O000000O1000000O1000000O1000000O100000000O10000O100O1O1O2N2N2N2N2N2N2N2O1N2M3N2N2N2N2N2N2M3M3M3L4L4L4L4L5K4L4L4L5K4M3L4L5KkYc5"}, "label": "a women in a gold and black sequined dress standing behind a women with a huge green bow on her dress"}]}
{"id": 44298, "image": "COCO_train2014_000000044298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a gold dress with brown hair\"."}], "task": "refering", "answer_template": "The \"a girl in a gold dress with brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 23, 35, 36, 37, 38, 39, 40, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 120, 121, 136, 137, 138, 153, 154, 170, 171, 187, 188, 189, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 290, 291, 292, 307, 308, 309, 324, 325, 326, 341, 342, 343, 358, 359, 375, 376], "bbox": [0.008979166666666667, 0.0561875, 0.4015, 0.986515625], "iscrowd": 0, "area": 41188.319950000005, "rle": {"size": [640, 480], "counts": "[k21jc09F9G:G8G:G9F9gJPNZHV2b7jM\\H[2a7fM\\H`2`7`M^Hf2^7[M`Hi2]7WMaHo2Z7SMcHR3n6kLgD3Y4Z3i6fLlD1X4`3f6bLPENY4g3`6^LUELm21RNm3e9YLZEIi28UNn3b9[LQH0YNl3`9ULPH7]Nl3Z9QLRH;`Nh3X9QLQH?eNb3V9SLmGd0iN[3V9jLVG2AW3U9GgF;V9IgF9P8fJQIU5lN7R8dJPIf:P7[EmHh:R7YE]H0Jj:i7VEZH4Jm5KfNP8ZKWH7Ki51eNQ5XKeM4ZM9Lf57eNf4`KmMLXM]69WNX4jKVNU6eMPNl3TL_Ni5iMSN^3]LiN^5lMTNS3fLQOS5PNWNe2oL[Oh4RNYNZ2XMD\\4VN[Nm1bMLR4ZN\\Na1jM3i3_N]NU1SN9a3eN\\Nh0\\Na0i2_MbJ\\1X4<fNf0`2eM_J[1[41oNm0X2iMZJ]1^4EXOT1o1mMXJ]1`4YOB\\1f1HhNcNKd1\\1MgNVN6l1S10fNkM`0T2i04j3LUL6j3IUL;i3DoJmMeLb2[8@nJTNcL^2_8\\OmJ[N_L\\2c8YOmJ_N\\LZ2f8VOoJcNVL[2i8iNgF]N[4b0QLZ2l8hNjFXNZ4i0lKY2o8gNZKSOdKW2Q9fN[KVOaKU2S9eN\\KYO^KS2U9dN]Kb2b4]M^Ke2`4\\M`Kd2]4_McKb2Y4aMgK`2U4bMlK^2Q4eMoK\\2m3gMSLZ2i3hMXLY2d3jM\\LV2a3mM_LT2]3nMdLS2X3PNhLP2V3RNjLo1R3SNoLn1m2UNSMl1i2WNWMi1f2YNZMi1TL[K]5o2_Ng1_KQLl5[2eNi2V1ZMjNh2Q1ZMPOi2i0\\MVOf2e0]M[Of2>^MBd28aMGa24bMLa2MdM2^2HfM8]2BfM>\\2\\OiMc0Y2XOjMh0Y2QOkMo0W2kNlMV1W2cNmM\\1V2]NnMd1T2UNQNk1R2mMRNT2P2eMUN[2n1]MVNc2n1UMWNi2o1nLUNR3P2eLUNY3Q2^LSNb3R2ULSNi3S2nKQNR4T2eKPNZ4l8O0VOg@mL[?o2m@kLT?R3SAjLn>T3XAfLi>Y3P1O1001O001O001O00001O000000O1000000O1000000O1000000O100000000O10000O100O1O1O2N2N2N2N2N2N2N2O1N2M3N2N2N2N2N2N2M3M3M3L4L4L4L4L5K4L4L4L5K4M3L4L5KkYc5"}, "label": "a girl in a gold dress with brown hair"}]}
{"id": 77067, "image": "COCO_train2014_000000077067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"winter truck\"."}], "task": "refering", "answer_template": "The \"winter truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 370, 371], "bbox": [0.0, 0.5737916666666667, 0.47220312499999995, 0.963875], "iscrowd": 0, "area": 40877.465800000005, "rle": {"size": [480, 640], "counts": "a9S3m;Y1gN00000O10O10000000O010000000000O010000000O1000O1000O1000000000O0100000000O10O10000000O1000O1000O1001O3M3L3N3M3M2N3M3M2M10000000001O000O100000001OO10O100000000000O01000000000000O01000000O100O1O001O1O1O1O100O001O1O1O1O1L4L3M4M3O1000000O0100000O100000O10O1000000O1000O1000O1000000O010000000O1000O1000000000000O10O1000000000O1000000000O1000O1000000000000O1000O10000000O1000000000O10O10000000O1000O100000O10000000O10O1000000000O02O001O1O001N101O1O001N101O1O001N01000000000O10000000O100000000000000001O001O0010O00010O01O001N101O2M2O1N101N2O0O2O1O1N101N2O0O2O1O1N1M4L4L4M3M4N1N2O2N1N2O001N2RN_C[1U=N3L3N3M2M4M2N3LmRn4"}, "label": "winter truck"}]}
{"id": 77067, "image": "COCO_train2014_000000077067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red and black truck driving on a snowy road\"."}], "task": "refering", "answer_template": "The \"red and black truck driving on a snowy road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 370, 371], "bbox": [0.0, 0.5737916666666667, 0.47220312499999995, 0.963875], "iscrowd": 0, "area": 40877.465800000005, "rle": {"size": [480, 640], "counts": "a9S3m;Y1gN00000O10O10000000O010000000000O010000000O1000O1000O1000000000O0100000000O10O10000000O1000O1000O1001O3M3L3N3M3M2N3M3M2M10000000001O000O100000001OO10O100000000000O01000000000000O01000000O100O1O001O1O1O1O100O001O1O1O1O1L4L3M4M3O1000000O0100000O100000O10O1000000O1000O1000O1000000O010000000O1000O1000000000000O10O1000000000O1000000000O1000O1000000000000O1000O10000000O1000000000O10O10000000O1000O100000O10000000O10O1000000000O02O001O1O001N101O1O001N101O1O001N01000000000O10000000O100000000000000001O001O0010O00010O01O001N101O2M2O1N101N2O0O2O1O1N101N2O0O2O1O1N1M4L4L4M3M4N1N2O2N1N2O001N2RN_C[1U=N3L3N3M2M4M2N3LmRn4"}, "label": "red and black truck driving on a snowy road"}]}
{"id": 380171, "image": "COCO_train2014_000000380171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan dog looking out the passenger window of the van\"."}], "task": "refering", "answer_template": "The \"a tan dog looking out the passenger window of the van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 140, 141, 142, 163, 164, 165, 185, 186, 187], "bbox": [0.049375, 0.17258333333333334, 0.20206249999999998, 0.5086041666666666], "iscrowd": 0, "area": 8870.416800000003, "rle": {"size": [480, 640], "counts": "_W?3m>2N00000000000O100000000000000eLLkD4_20e8>ZGBa8g0[GYO_8o0_GQO[8U1dGlNV8Z1jGfNR8^1mGcNP8a1nG`NQ8a1nG`NQ8b1lG`NS8a1gGeNX8]1YGQOe8S1SGSOk8h1[FYNd9j1ZFQMAP1T:Q2aFmM_9S2dFjM[9X2fFfMY9[2iFcMW9]2kFaMU9b2jF\\MV9e2lFXMT9i2mFUMR9m2oFQMQ9P3PGnLo8S3SGkLm8V3UGgLj8\\3cF_LM3`9_3cF`LL0_9b3eF_LLM_9e3eF_Lk9b3UF_Li9b3WF_Lg9b3YF_Le9b3[F`Lb9a3^F`La9`3_FaL_9a3_FaL_9`3aF`L^9a3bF^L^9c3bF\\L^9f3aFYL_9h3aFWL_9j3aFUL^9m3bFRL^9o3bFPL_9Q4_FoKa9R4_FmKa9T4_FkKb9U4^FjKb9X4]FgKd9Y4\\FfKe9Z4[FeKe9\\4[FcKd9_4>1O1O1N200O100O10K4J7J6J6C=VNfDG`;6fDZOg;FQD6=Ik;CiCK6g0;Bl;MnC?U=@kB?V=BiB=Y=BfB=^=BaB=Q>01N1O2KW]_7"}, "label": "a tan dog looking out the passenger window of the van"}]}
{"id": 380171, "image": "COCO_train2014_000000380171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a golden colored dog sitting on the passenger side\"."}], "task": "refering", "answer_template": "The \"a golden colored dog sitting on the passenger side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 140, 141, 142, 163, 164, 165, 185, 186, 187], "bbox": [0.049375, 0.17258333333333334, 0.20206249999999998, 0.5086041666666666], "iscrowd": 0, "area": 8870.416800000003, "rle": {"size": [480, 640], "counts": "_W?3m>2N00000000000O100000000000000eLLkD4_20e8>ZGBa8g0[GYO_8o0_GQO[8U1dGlNV8Z1jGfNR8^1mGcNP8a1nG`NQ8a1nG`NQ8b1lG`NS8a1gGeNX8]1YGQOe8S1SGSOk8h1[FYNd9j1ZFQMAP1T:Q2aFmM_9S2dFjM[9X2fFfMY9[2iFcMW9]2kFaMU9b2jF\\MV9e2lFXMT9i2mFUMR9m2oFQMQ9P3PGnLo8S3SGkLm8V3UGgLj8\\3cF_LM3`9_3cF`LL0_9b3eF_LLM_9e3eF_Lk9b3UF_Li9b3WF_Lg9b3YF_Le9b3[F`Lb9a3^F`La9`3_FaL_9a3_FaL_9`3aF`L^9a3bF^L^9c3bF\\L^9f3aFYL_9h3aFWL_9j3aFUL^9m3bFRL^9o3bFPL_9Q4_FoKa9R4_FmKa9T4_FkKb9U4^FjKb9X4]FgKd9Y4\\FfKe9Z4[FeKe9\\4[FcKd9_4>1O1O1N200O100O10K4J7J6J6C=VNfDG`;6fDZOg;FQD6=Ik;CiCK6g0;Bl;MnC?U=@kB?V=BiB=Y=BfB=^=BaB=Q>01N1O2KW]_7"}, "label": "a golden colored dog sitting on the passenger side"}]}
{"id": 232717, "image": "COCO_train2014_000000232717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with towel and seat\"."}], "task": "refering", "answer_template": "The \"a chair with towel and seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 221, 222, 223, 243, 244, 245, 246, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 333, 334, 335, 336, 355, 356, 357, 359], "bbox": [0.43862500000000004, 0.5110416666666667, 0.7254687500000001, 0.9403958333333333], "iscrowd": 0, "area": 16607.7607, "rle": {"size": [480, 640], "counts": "aRT46h>3K5J6K5K5K4M4K401N1O20O01O010O001O010O0010OQDnNP:T1hEQOY:S1cElN^:X1]EhNe:\\1VEcNk:_1REaNo:`1PE_NQ;a1nD_NT;`1lD`NT;a1jD_NW;a1iD_NW;b1gD^NY;c1gD]NY;c1fD]NZ;e1eDZN\\;f1cD[N\\;g1cDXN^;h1aDYN_;g1aDXN_;j1_DWNa;i1_DVNa;l1]DUNc;k1]DTNc;m1\\DSNe;n1ZDRNe;o1ZDQNg;o1YDQNf;Q2XDoMi;Q2WDoMh;S2VDmMk;S2UDmMj;T2UDlMl;U2SDjMm;e20O010O010O02K4L4L4L4M300O1O100000O2N100O101N1000001O0O10001O0000001C_DYMa;f2`DZM^;g2cDZMZ;g2gDYMV;i2kDWMS;j2nDVMo:l2RETMl:n2TERMi:P3XEPMf:Q3[EPMa:R3`EnL^:T3bElL[:V3fEjLX:W3iEiLT:Y3mEgLR:Z3nEfLQ:Z3QFfLl9[3UFeLj9\\3VFdLi9\\3XFdLg9\\3ZFdLe9\\3\\FdLd9\\3\\FeLb9[3_FeL`9[3aFeL_9Z3bFfL]9[3cFeL\\9[3eFeL[9Z3fFfLY9Z3hFgLV9Z3jFfLV9Y3kFgLU9X3lFhLS9X3oFgLQ9Y3oFgLQ9X3PGhLP9W3QGjLm8W3SGiLm8V3TGjLl8U3UGkLk8T3VGlLi8U3WGkLi8T3XGlLh8S3XGmLh8S3YGlLh8S3_GfLa8X3e1M2M3N3M2N3L3N3M2M3N3M2M4O001N10001O0O2O001N10001N101O001N2O1O1N3O0O101O003M3M3M3N2M3M3N2M3M3MZWb2"}, "label": "a chair with towel and seat"}]}
{"id": 232717, "image": "COCO_train2014_000000232717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer chair with a red towel on it\"."}], "task": "refering", "answer_template": "The \"a computer chair with a red towel on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 221, 222, 223, 243, 244, 245, 246, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 333, 334, 335, 336, 355, 356, 357, 359], "bbox": [0.43862500000000004, 0.5110416666666667, 0.7254687500000001, 0.9403958333333333], "iscrowd": 0, "area": 16607.7607, "rle": {"size": [480, 640], "counts": "aRT46h>3K5J6K5K5K4M4K401N1O20O01O010O001O010O0010OQDnNP:T1hEQOY:S1cElN^:X1]EhNe:\\1VEcNk:_1REaNo:`1PE_NQ;a1nD_NT;`1lD`NT;a1jD_NW;a1iD_NW;b1gD^NY;c1gD]NY;c1fD]NZ;e1eDZN\\;f1cD[N\\;g1cDXN^;h1aDYN_;g1aDXN_;j1_DWNa;i1_DVNa;l1]DUNc;k1]DTNc;m1\\DSNe;n1ZDRNe;o1ZDQNg;o1YDQNf;Q2XDoMi;Q2WDoMh;S2VDmMk;S2UDmMj;T2UDlMl;U2SDjMm;e20O010O010O02K4L4L4L4M300O1O100000O2N100O101N1000001O0O10001O0000001C_DYMa;f2`DZM^;g2cDZMZ;g2gDYMV;i2kDWMS;j2nDVMo:l2RETMl:n2TERMi:P3XEPMf:Q3[EPMa:R3`EnL^:T3bElL[:V3fEjLX:W3iEiLT:Y3mEgLR:Z3nEfLQ:Z3QFfLl9[3UFeLj9\\3VFdLi9\\3XFdLg9\\3ZFdLe9\\3\\FdLd9\\3\\FeLb9[3_FeL`9[3aFeL_9Z3bFfL]9[3cFeL\\9[3eFeL[9Z3fFfLY9Z3hFgLV9Z3jFfLV9Y3kFgLU9X3lFhLS9X3oFgLQ9Y3oFgLQ9X3PGhLP9W3QGjLm8W3SGiLm8V3TGjLl8U3UGkLk8T3VGlLi8U3WGkLi8T3XGlLh8S3XGmLh8S3YGlLh8S3_GfLa8X3e1M2M3N3M2N3L3N3M2M3N3M2M4O001N10001O0O2O001N10001N101O001N2O1O1N3O0O101O003M3M3M3N2M3M3N2M3M3MZWb2"}, "label": "a computer chair with a red towel on it"}]}
{"id": 232717, "image": "COCO_train2014_000000232717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an armchair without a towel\"."}], "task": "refering", "answer_template": "The \"an armchair without a towel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [317, 339, 340, 341, 357, 358, 359, 360, 361, 362, 363, 364, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.535546875, 0.8095208333333334, 0.83884375, 0.9932083333333334], "iscrowd": 0, "area": 8776.910600000003, "rle": {"size": [480, 640], "counts": "beQ51P]1f0YQOc0]O00000001O000O10000000001O000000000000001O0000000000001O0000000O1000001O000000000000001O000000000000001O0000000O1000001O0000000000001O000000000000001O000000000O10001O000000000000000000O1O1O1O1O100O1O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10O01O1O1O1O100O1O10000000000001O00000000001O0000000O10001O:F9G9G9G9G:F9G9G^j_1"}, "label": "an armchair without a towel"}]}
{"id": 232717, "image": "COCO_train2014_000000232717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair\"."}], "task": "refering", "answer_template": "The \"a black chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [317, 339, 340, 341, 357, 358, 359, 360, 361, 362, 363, 364, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.535546875, 0.8095208333333334, 0.83884375, 0.9932083333333334], "iscrowd": 0, "area": 8776.910600000003, "rle": {"size": [480, 640], "counts": "beQ51P]1f0YQOc0]O00000001O000O10000000001O000000000000001O0000000000001O0000000O1000001O000000000000001O000000000000001O0000000O1000001O0000000000001O000000000000001O000000000O10001O000000000000000000O1O1O1O1O100O1O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10O01O1O1O1O100O1O10000000000001O00000000001O0000000O10001O:F9G9G9G9G:F9G9G^j_1"}, "label": "a black chair"}]}
{"id": 347407, "image": "COCO_train2014_000000347407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with mobile on hand , typing text in mobile\"."}], "task": "refering", "answer_template": "The \"a girl with mobile on hand , typing text in mobile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 21, 22, 23, 24, 42, 43, 44, 45, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 252, 253, 254, 255, 256], "bbox": [0.001507537688442211, 0.0, 0.3297319932998325, 0.9190999999999999], "iscrowd": 0, "area": 47814.85340000001, "rle": {"size": [400, 597], "counts": "ma0P2Z5a5D<E;D=C<D<D7I4L3N3M3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M3M2N3M000000000000001O000000000001O0000001O00000010O000001O0000001O0000001O0000010O0000001O0000001O0000001O100O2M3N2N2M3N2M2O2N2M3N2N2MR1XFhLV8X3gGlLW8V3fGnLX8S3eGPMZ8Q3dGRMY8o2fGSMY8n2eGUMX8c3O2N1N3N1O2M3N1N3N000O2O000O2O0O101O0O101O0O101N10001N100O2O001N10001N100O2O2N6I7J6J7H7mLdIFb63`IKg6M]I1h6H\\I6j6BZI;l6_OVI`0o6YOUIe0U7oNoHn0a7aNbH^1V8iMmGT2Z9N1O1N2O1N101O1N2O001N2O1N101O1N2O001N2O1N101O1N2O002M3N2M3N3M2M3N2N2M3N2M3N2N2M4M^Zl4"}, "label": "a girl with mobile on hand , typing text in mobile"}]}
{"id": 347407, "image": "COCO_train2014_000000347407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian girl wearing stripes looks down at her phone\"."}], "task": "refering", "answer_template": "The \"an asian girl wearing stripes looks down at her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 21, 22, 23, 24, 42, 43, 44, 45, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 252, 253, 254, 255, 256], "bbox": [0.001507537688442211, 0.0, 0.3297319932998325, 0.9190999999999999], "iscrowd": 0, "area": 47814.85340000001, "rle": {"size": [400, 597], "counts": "ma0P2Z5a5D<E;D=C<D<D7I4L3N3M3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M3M2N3M000000000000001O000000000001O0000001O00000010O000001O0000001O0000001O0000010O0000001O0000001O0000001O100O2M3N2N2M3N2M2O2N2M3N2N2MR1XFhLV8X3gGlLW8V3fGnLX8S3eGPMZ8Q3dGRMY8o2fGSMY8n2eGUMX8c3O2N1N3N1O2M3N1N3N000O2O000O2O0O101O0O101O0O101N10001N100O2O001N10001N100O2O2N6I7J6J7H7mLdIFb63`IKg6M]I1h6H\\I6j6BZI;l6_OVI`0o6YOUIe0U7oNoHn0a7aNbH^1V8iMmGT2Z9N1O1N2O1N101O1N2O001N2O1N101O1N2O001N2O1N101O1N2O002M3N2M3N3M2M3N2N2M3N2M3N2N2M4M^Zl4"}, "label": "an asian girl wearing stripes looks down at her phone"}]}
{"id": 347407, "image": "COCO_train2014_000000347407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting at a table with women on both sides of him\"."}], "task": "refering", "answer_template": "The \"a man sitting at a table with women on both sides of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 34, 35, 36, 37, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 161, 162], "bbox": [0.5900335008375209, 0.0033750000000000004, 0.8435510887772195, 0.546175], "iscrowd": 0, "area": 19514.818250000008, "rle": {"size": [400, 597], "counts": "^bY45Y<3N2N2N2N2N2N2N3M1O1O1O001O001O1O00100O00100O00100O010O010000O1000O10<D;E;E?Ac0aF\\Mk7f3O1O1O1O1O1N2O1O4L3L4aN\\KlJh4P5^KjJe4R5bKgJb4U5eKeJ[4Z5kKaJV4^5lK_JU4`5mK_JS4a5nK]JT4b5nK\\JR4d5oK[JQ4e5QLXJQ4g5PLXJP4h5\\10001O01O01O00001O0000000000O100O10000O10000O10000O100O100O1O100O1O100O100O1O100O1O100O1N2M3N200000000000000dKfIb2Z6[MiIe2W6WMnIh2R6UMQJk2o5RMUJm2k5oLZJm0\\OMZ6SO]Jj0^O4U6nNbJg0_O;o5kNeJd0B`0j5hNiJa0Cg0d5eNmJ<EP1^5`NRK:EV1Y5\\NYK8A\\1V5YN_K5^Ob1S5UNfK2ZOj1\\7QNgHn1n8000O100000O10001N101N2O0O2O0O<Ef0YO=D7HhTT1"}, "label": "a man sitting at a table with women on both sides of him"}]}
{"id": 347407, "image": "COCO_train2014_000000347407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man who is sitting along with the girls\"."}], "task": "refering", "answer_template": "The \"a man who is sitting along with the girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 34, 35, 36, 37, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 161, 162], "bbox": [0.5900335008375209, 0.0033750000000000004, 0.8435510887772195, 0.546175], "iscrowd": 0, "area": 19514.818250000008, "rle": {"size": [400, 597], "counts": "^bY45Y<3N2N2N2N2N2N2N3M1O1O1O001O001O1O00100O00100O00100O010O010000O1000O10<D;E;E?Ac0aF\\Mk7f3O1O1O1O1O1N2O1O4L3L4aN\\KlJh4P5^KjJe4R5bKgJb4U5eKeJ[4Z5kKaJV4^5lK_JU4`5mK_JS4a5nK]JT4b5nK\\JR4d5oK[JQ4e5QLXJQ4g5PLXJP4h5\\10001O01O01O00001O0000000000O100O10000O10000O10000O100O100O1O100O1O100O100O1O100O1O100O1N2M3N200000000000000dKfIb2Z6[MiIe2W6WMnIh2R6UMQJk2o5RMUJm2k5oLZJm0\\OMZ6SO]Jj0^O4U6nNbJg0_O;o5kNeJd0B`0j5hNiJa0Cg0d5eNmJ<EP1^5`NRK:EV1Y5\\NYK8A\\1V5YN_K5^Ob1S5UNfK2ZOj1\\7QNgHn1n8000O100000O10001N101N2O0O2O0O<Ef0YO=D7HhTT1"}, "label": "a man who is sitting along with the girls"}]}
{"id": 347407, "image": "COCO_train2014_000000347407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with frizzy hair sits behind a briefcase with different cards inside of it\"."}], "task": "refering", "answer_template": "The \"a woman with frizzy hair sits behind a briefcase with different cards inside of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 50, 51, 52, 53, 70, 71, 72, 73, 74, 75, 90, 91, 92], "bbox": [0.3116750418760469, 0.0, 0.5962479061976549, 0.3236], "iscrowd": 0, "area": 12396.907299999997, "rle": {"size": [400, 597], "counts": "`UY21_<0O100O1O100O1O001N2N2N2N2N2N2O1N1O2N2O1O1O1O1O100O1O001O1O1O1TNoNXHR1a7YOZHh0a7A[H?d7FWH:i7KRH6m7OoG1P82mGOR83lGMT84kGMT85kGKT87jGJU87jGIV89iGGV8:iGGV8;hGFW8;iGDW8>gGCX8?gGAX8a0fG@Z8`0fG_O[8b0cG_O]8b0bG^O^8c0`G^O`8c0_G\\Ob8d0]G]Oc8d0[G]Oe8d0ZG\\Of8e0XG[Oi8f0UG[Ok8U20O1000000O1000000O0100000O1000000O1000000O1000000O1000000O100000O010000O1000000O1000000O1000000O1000000O100000O010000O1000000O10000O1000000O10000O1000000O10O10ROlF`NT9\\1QGbNP9Z1TGfNl8V1XGjNh8R1\\GmNe8o0`GPO`8l0dGTO\\8f0jGZOV8:WHDj71aHO_7EmH;d9O2O00001N10000O2O1O1O1N101O1N10W^n2"}, "label": "a woman with frizzy hair sits behind a briefcase with different cards inside of it"}]}
{"id": 347407, "image": "COCO_train2014_000000347407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a plaid shirt and curly hair\"."}], "task": "refering", "answer_template": "The \"a woman with a plaid shirt and curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 50, 51, 52, 53, 70, 71, 72, 73, 74, 75, 90, 91, 92], "bbox": [0.3116750418760469, 0.0, 0.5962479061976549, 0.3236], "iscrowd": 0, "area": 12396.907299999997, "rle": {"size": [400, 597], "counts": "`UY21_<0O100O1O100O1O001N2N2N2N2N2N2O1N1O2N2O1O1O1O1O100O1O001O1O1O1TNoNXHR1a7YOZHh0a7A[H?d7FWH:i7KRH6m7OoG1P82mGOR83lGMT84kGMT85kGKT87jGJU87jGIV89iGGV8:iGGV8;hGFW8;iGDW8>gGCX8?gGAX8a0fG@Z8`0fG_O[8b0cG_O]8b0bG^O^8c0`G^O`8c0_G\\Ob8d0]G]Oc8d0[G]Oe8d0ZG\\Of8e0XG[Oi8f0UG[Ok8U20O1000000O1000000O0100000O1000000O1000000O1000000O1000000O100000O010000O1000000O1000000O1000000O1000000O100000O010000O1000000O10000O1000000O10000O1000000O10O10ROlF`NT9\\1QGbNP9Z1TGfNl8V1XGjNh8R1\\GmNe8o0`GPO`8l0dGTO\\8f0jGZOV8:WHDj71aHO_7EmH;d9O2O00001N10000O2O1O1O1N101O1N10W^n2"}, "label": "a woman with a plaid shirt and curly hair"}]}
{"id": 199956, "image": "COCO_train2014_000000199956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stop sign for the streets charles ave and ridgewood\"."}], "task": "refering", "answer_template": "The \"a stop sign for the streets charles ave and ridgewood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 186, 187, 188], "bbox": [0.058828124999999995, 0.17927083333333332, 0.2715, 0.48056250000000006], "iscrowd": 0, "area": 16108.43795, "rle": {"size": [480, 640], "counts": "nna0T1Q=l0M3N2N2N2N2N2N2M3N2N2N2N2N3M2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N3M2M3N2N2N2N100O10000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O10000000000001N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2M2O2N2N2N>BY1gNaXj6"}, "label": "a stop sign for the streets charles ave and ridgewood"}]}
{"id": 199956, "image": "COCO_train2014_000000199956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red board\"."}], "task": "refering", "answer_template": "The \"red board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 186, 187, 188], "bbox": [0.058828124999999995, 0.17927083333333332, 0.2715, 0.48056250000000006], "iscrowd": 0, "area": 16108.43795, "rle": {"size": [480, 640], "counts": "nna0T1Q=l0M3N2N2N2N2N2N2M3N2N2N2N2N3M2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N3M2M3N2N2N2N100O10000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O10000000000001N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2M2O2N2N2N>BY1gNaXj6"}, "label": "red board"}]}
{"id": 306454, "image": "COCO_train2014_000000306454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vegetable in the middle\"."}], "task": "refering", "answer_template": "The \"the vegetable in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 141, 148, 149, 165, 170, 171, 172, 193, 194, 195, 212, 213, 214, 215, 216, 217], "bbox": [0.14384375, 0.34838407494145196, 0.51778125, 0.6548711943793911], "iscrowd": 0, "area": 4445.725000000001, "rle": {"size": [427, 640], "counts": "\\`V12X=4L3M4O10O101O0O100O2O000O101O0O1000O0010O000O101O000O2O00000Okj00O0WUO0O10O10O1000O0100O10O10O1000O01OO1O01O00010O000010O01O00010O000010O0001O01O01O01O01O0001O0000001O0001O01O0000001O0000010O0000000000000001O0O100000000000001O000000000000000O2O000000000000000000000000O1000O1000000000000000O10000000000000O1000O1000000000000001N100O100O2O0O100K5K5J7I6M3L4M3\\OhNQE\\1m:hNnD[1P;iNjD[1S;`0N1O2M2O2N101O10O01O010O10O010O010O1O1O1O0O2O1O1O1O1O001O1N2O1O1N3N2M2O2M3N1N3N2M2O2M3N2M4L3M4L4L4L3M4K5L4LddP4"}, "label": "the vegetable in the middle"}]}
{"id": 306454, "image": "COCO_train2014_000000306454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot in the middle\"."}], "task": "refering", "answer_template": "The \"the carrot in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 141, 148, 149, 165, 170, 171, 172, 193, 194, 195, 212, 213, 214, 215, 216, 217], "bbox": [0.14384375, 0.34838407494145196, 0.51778125, 0.6548711943793911], "iscrowd": 0, "area": 4445.725000000001, "rle": {"size": [427, 640], "counts": "\\`V12X=4L3M4O10O101O0O100O2O000O101O0O1000O0010O000O101O000O2O00000Okj00O0WUO0O10O10O1000O0100O10O10O1000O01OO1O01O00010O000010O01O00010O000010O0001O01O01O01O01O0001O0000001O0001O01O0000001O0000010O0000000000000001O0O100000000000001O000000000000000O2O000000000000000000000000O1000O1000000000000000O10000000000000O1000O1000000000000001N100O100O2O0O100K5K5J7I6M3L4M3\\OhNQE\\1m:hNnD[1P;iNjD[1S;`0N1O2M2O2N101O10O01O010O10O010O010O1O1O1O0O2O1O1O1O1O001O1N2O1O1N3N2M2O2M3N1N3N2M2O2M3N2M4L3M4L4L4L3M4K5L4LddP4"}, "label": "the carrot in the middle"}]}
{"id": 306454, "image": "COCO_train2014_000000306454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fat carrot closest to the knife\"."}], "task": "refering", "answer_template": "The \"the fat carrot closest to the knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 30, 53, 76, 99, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 192, 193], "bbox": [0.305, 0.02857142857142857, 0.448828125, 0.5584309133489461], "iscrowd": 0, "area": 7222.675749999999, "rle": {"size": [427, 640], "counts": "o_a23m<=C<E;D=D;D<D=D;DIYFaM[9_2RGaMb8_2iGaMk7`2aH_MT7`2YI`MZ6a2RJ^Mb5c2jJ\\Mk4c2bK]MX4]2TLbMj3U2bLjM\\3m1PMSNn2]1cMcN[2a0dN^OR6d1M2N2M4M2N2N3L3N2N3M2N2M2O1O2N1O100O10000O10000O10000O10000O10000O10O10O01O1O001O1O002N1O2N101M2O2N1N3N1N3N1O2M2N3M2N2N3L3N3M2N3L3N3M2N:Fa0^OSQc4"}, "label": "the fat carrot closest to the knife"}]}
{"id": 306454, "image": "COCO_train2014_000000306454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vegetable closest to the knife handle\"."}], "task": "refering", "answer_template": "The \"the vegetable closest to the knife handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 30, 53, 76, 99, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 192, 193], "bbox": [0.305, 0.02857142857142857, 0.448828125, 0.5584309133489461], "iscrowd": 0, "area": 7222.675749999999, "rle": {"size": [427, 640], "counts": "o_a23m<=C<E;D=D;D<D=D;DIYFaM[9_2RGaMb8_2iGaMk7`2aH_MT7`2YI`MZ6a2RJ^Mb5c2jJ\\Mk4c2bK]MX4]2TLbMj3U2bLjM\\3m1PMSNn2]1cMcN[2a0dN^OR6d1M2N2M4M2N2N3L3N2N3M2N2M2O1O2N1O100O10000O10000O10000O10000O10000O10O10O01O1O001O1O002N1O2N101M2O2N1N3N1N3N1O2M2N3M2N2N3L3N3M2N3L3N3M2N:Fa0^OSQc4"}, "label": "the vegetable closest to the knife handle"}]}
{"id": 544026, "image": "COCO_train2014_000000544026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue car\"."}], "task": "refering", "answer_template": "The \"blue car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 68, 69, 70, 71, 77, 78, 79, 80, 86, 87, 88, 89, 96], "bbox": [0.6023938223938224, 0.44255813953488377, 1.0, 0.7443669250645994], "iscrowd": 0, "area": 8617.4838, "rle": {"size": [387, 259], "counts": "QUk1k0R;a0D6I4M3L5L3M3L4M4L3L4M4K4M3M4K4N2N2N3L2O00000O0100O10O0100O1000O0100O10O0100O100O10O10O100O100O1000O01000001N101O00001N101O001N10001N101O0O2O0O101O0O2O1O3L4M2N:E;F1N101O1N2O1O0O2O1O1N2O1N4M4L3L4M3M2M4M3L]H"}, "label": "blue car"}]}
{"id": 544026, "image": "COCO_train2014_000000544026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of cake with blue icing\"."}], "task": "refering", "answer_template": "The \"a piece of cake with blue icing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 68, 69, 70, 71, 77, 78, 79, 80, 86, 87, 88, 89, 96], "bbox": [0.6023938223938224, 0.44255813953488377, 1.0, 0.7443669250645994], "iscrowd": 0, "area": 8617.4838, "rle": {"size": [387, 259], "counts": "QUk1k0R;a0D6I4M3L5L3M3L4M4L3L4M4K4M3M4K4N2N2N3L2O00000O0100O10O0100O1000O0100O10O0100O100O10O10O100O100O1000O01000001N101O00001N101O001N10001N101O0O2O0O101O0O2O1O3L4M2N:E;F1N101O1N2O1O0O2O1O1N2O1N4M4L3L4M3M2M4M3L]H"}, "label": "a piece of cake with blue icing"}]}
{"id": 470305, "image": "COCO_train2014_000000470305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women in jeans and green t - shirt\"."}], "task": "refering", "answer_template": "The \"a women in jeans and green t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 26, 27, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 142, 143, 165, 166, 188, 189], "bbox": [0.11279687499999999, 0.04058080808080808, 0.256515625, 0.6417676767676768], "iscrowd": 0, "area": 11423.708700000003, "rle": {"size": [396, 640], "counts": "Vok0:Q<4M3L4M2M3N000O2bNGgF9R9?\\FBb9d0ZF\\Od9g0ZFZOc9i1N2N200O100O100O1O100iNjMWHW2h7jMXHV2h7kMVHV2i7lMVHT2i7nMUHS2k7W1O100O1O100O100O10000001O000000000000001O0000000000001O002O3L\\JkKT3R4iLULV3h3hL]LY3_3dLgL_3U3\\LoLi3n2]JbLe0P1T5[2SJbLLT2T6Y1mIoNT6e30O001O001O001O001O0O1O2N101WLaIb1`6]NbIm0MdMc6]1cIl02_M\\6e1dIi07[MV6k1eIg0R7WO\\I;e6EU3K7I7^Oa0^OVPh5"}, "label": "a women in jeans and green t - shirt"}]}
{"id": 470305, "image": "COCO_train2014_000000470305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl sitting on end of pillar\"."}], "task": "refering", "answer_template": "The \"girl sitting on end of pillar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 26, 27, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 142, 143, 165, 166, 188, 189], "bbox": [0.11279687499999999, 0.04058080808080808, 0.256515625, 0.6417676767676768], "iscrowd": 0, "area": 11423.708700000003, "rle": {"size": [396, 640], "counts": "Vok0:Q<4M3L4M2M3N000O2bNGgF9R9?\\FBb9d0ZF\\Od9g0ZFZOc9i1N2N200O100O100O1O100iNjMWHW2h7jMXHV2h7kMVHV2i7lMVHT2i7nMUHS2k7W1O100O1O100O100O10000001O000000000000001O0000000000001O002O3L\\JkKT3R4iLULV3h3hL]LY3_3dLgL_3U3\\LoLi3n2]JbLe0P1T5[2SJbLLT2T6Y1mIoNT6e30O001O001O001O001O0O1O2N101WLaIb1`6]NbIm0MdMc6]1cIl02_M\\6e1dIi07[MV6k1eIg0R7WO\\I;e6EU3K7I7^Oa0^OVPh5"}, "label": "girl sitting on end of pillar"}]}
{"id": 208165, "image": "COCO_train2014_000000208165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman laying down next to a cat\"."}], "task": "refering", "answer_template": "The \"a woman laying down next to a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 32, 38, 39, 40, 41, 53, 54, 55, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 268, 269, 270], "bbox": [0.276828125, 0.0022500000000000003, 0.9, 0.9101111111111111], "iscrowd": 0, "area": 62945.862999999976, "rle": {"size": [360, 640], "counts": "R\\n1c0b:5J5L4K5L5J5L4K2O2M2OdNmFh0P9[OYG;f8HbGOZ85X14M3L5K5_Ob0E;VNkMVJ_2j5eMkIb2T6cMeI`2Z6fM^I]2a6PNPIS2o6T1O1O1O1O2N1O00000000O100000000000000O10000000000000000O100O100O100O100O101O2M4M2M3N2M2O2M3N1N3N3L5PNkGe0X8kNWHR1R91O4L4L3M101N100O2O0OO2M3N1N2O1N01O0010O01O001O0010O01O001O001O010O001O001O001O1O1O1O001O1O1O1O1O010O1O100O001O100O1O100O2N101N1O2O0O2N100O2N101N1O2O0O2N101O001O00001O001O0O101M2O1O2N1N2O2N1N3N1N2O2M2O1N3N1N3N1N2O2M2O1N3N0O2O1N1O2O1N101N2O0O2O1N2O0O2N2O0O2O1N1O2O1N1O2N2O1N1O2N2O0O2M3N1N3N2M2O2N2M2O2M3N2L3M4M3L3M4M3M2M4M3M2M4M3M3L3O2M3N1N3N2N1N3N2N1N3N2M2O2O00001O000O2O0000001O00001O0000001O00001O000O2O0000001O00001O0001O01O000000O1N2O1N2O1N2O1N2O1O1N2O2M2O1N2O1N2O2N1N3N2M2O2M2O2M3N1O2M3N1N3N1N3N2M2O2N2N102N2N2N3M2N2N2N3M2N1000O10000O10O0100O10000O100O1001O1N3gJlKY3V4VL[Lg3g3SLaLi3b3PLfLl3]3mKkLo3Z5M3M3M3M3M3M3M2N2N2N2N2N2N2N2N2N2N2N2N2N2M3L5K4M4K4L4L5L3L5K4M3L5K4L5CiXf0"}, "label": "a woman laying down next to a cat"}]}
{"id": 208165, "image": "COCO_train2014_000000208165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman laying next to her cat\"."}], "task": "refering", "answer_template": "The \"a woman laying next to her cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 32, 38, 39, 40, 41, 53, 54, 55, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 268, 269, 270], "bbox": [0.276828125, 0.0022500000000000003, 0.9, 0.9101111111111111], "iscrowd": 0, "area": 62945.862999999976, "rle": {"size": [360, 640], "counts": "R\\n1c0b:5J5L4K5L5J5L4K2O2M2OdNmFh0P9[OYG;f8HbGOZ85X14M3L5K5_Ob0E;VNkMVJ_2j5eMkIb2T6cMeI`2Z6fM^I]2a6PNPIS2o6T1O1O1O1O2N1O00000000O100000000000000O10000000000000000O100O100O100O100O101O2M4M2M3N2M2O2M3N1N3N3L5PNkGe0X8kNWHR1R91O4L4L3M101N100O2O0OO2M3N1N2O1N01O0010O01O001O0010O01O001O001O010O001O001O001O1O1O1O001O1O1O1O1O010O1O100O001O100O1O100O2N101N1O2O0O2N100O2N101N1O2O0O2N101O001O00001O001O0O101M2O1O2N1N2O2N1N3N1N2O2M2O1N3N1N3N1N2O2M2O1N3N0O2O1N1O2O1N101N2O0O2O1N2O0O2N2O0O2O1N1O2O1N1O2N2O1N1O2N2O0O2M3N1N3N2M2O2N2M2O2M3N2L3M4M3L3M4M3M2M4M3M2M4M3M3L3O2M3N1N3N2N1N3N2N1N3N2M2O2O00001O000O2O0000001O00001O0000001O00001O000O2O0000001O00001O0001O01O000000O1N2O1N2O1N2O1N2O1O1N2O2M2O1N2O1N2O2N1N3N2M2O2M2O2M3N1O2M3N1N3N1N3N2M2O2N2N102N2N2N3M2N2N2N3M2N1000O10000O10O0100O10000O100O1001O1N3gJlKY3V4VL[Lg3g3SLaLi3b3PLfLl3]3mKkLo3Z5M3M3M3M3M3M3M2N2N2N2N2N2N2N2N2N2N2N2N2N2M3L5K4M4K4L4L5L3L5K4M3L5K4L5CiXf0"}, "label": "a woman laying next to her cat"}]}
{"id": 208165, "image": "COCO_train2014_000000208165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange and gold bed spread\"."}], "task": "refering", "answer_template": "The \"orange and gold bed spread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 118, 119, 120, 121, 134, 135, 136, 138, 139, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 186, 187, 188, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.0, 0.3268333333333333, 1.0, 0.9979722222222221], "iscrowd": 0, "area": 73754.6944, "rle": {"size": [360, 640], "counts": "f3a1g9T3mLi2VM01O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O00001O000000001O0000001O0000001O000000000000000000001O00000XL`Kh0`4WObKh0^4XOdKf0\\4ZOeKe0[4[OfKd0Z4\\OgKc0Y4]OiKa0W4_OiKa0W4^OjKb0V4^OkKa0U4_OkKa0U4_OkKa0U4_OlK`0T4@lK`0T4@lK`0U4^OlKb0T4^OmKa0S4_OmKa0S4_OmKa0S4_OnK`0R4@nK`0R4@nK`0R4@oK?Q4AoK?Q4@PL`0P4@PL`0P4@QL?o3AQL?o3AQL?o3ARL>n3BRL>n3BRL>o3ARL>n3BSL=m3CSL=m3CTL<l3DUL;k3EUL;k3EVL:j3FWL9i3GXL8h3HXL8h3HXL8h3HXL8h3HXL8h3HYL7g3IYL7g3IYL7g3HZL8f3HZL8g3FZL:f3FZL:f3E[L;e3E[L;e3D\\L<d3D\\L<d3C]L=c3C^L<b3C_L>`3B`L>`3AaL?_3mN`KdNQ1_2_3jNdKfNm0`2_3hNgKfNk0b2^3iNgKdNk0c2^3iNgKcNl0d2]3iNhKbNk0e2]3iNiK`Nk0g2]3hNiK`Nk0g2\\3jNiK]Nl0i2[3jNjK\\Nk0j2[3jNkKZNk0l2Z3jNlKYNj0m2Z3jNXMV1h2kNWMU1i2kNWMU1i2kNWMU1j2jNVMV1j2jNVMV1j2kNUMU1k2kNUMU1k2kNUMU1k2kNUMU1l2jNTMV1l2jNTMV1l2jNTMV1l2jNTMV1l2jNTMV1l2jNTMV1m2hNTMX1l2hNTMX1l2gNUMY1k2gNUMY1k2gNUMY1k2gNUMY1l2fNTMZ1l2fNTMZ1l2fNTMZ1l2eNUM[1k2eNUM[1k2eNUM[1l2dNTM\\1l2dNTM\\1l2dNTM\\1l2cNUM]1k2cNUM]1k2cNUM]1l2bNTM^1l2_MXKc0l1n1l2XM`Ki0d1o1l2RMhKl0]1R2_3mMaLS2a3kM_LU2b3jM_LU2c3iM]LW2d3hM\\LX2f3eM\\LZ2e3eM[L[2g3cMYL]2h3bMYL]2h3_M[La2f3[M]Le2d3WM`Lh2P60001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O1O001O001O1O001O001O1O001O000000O1000000000000O1000000000000O10000000000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O2O000O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100000000001O00000000000000001O000000000000002N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O001O001O00001O001O001O001O001O0000001O0000001O0000001O0000001O0000001O0000001O00001O00001O00001O0000001O00001O00001O1O001O00000000O10000O100N2O1N2O1N2O1M3L4M3L4M3N2O1N2O1N2O1N2M3N2N2O1O100O1O100O1O1001O000000000000001O000O2O001O00001O001O00001O00001O001O000000001O00000000_KRNIm17TNHl19UNEk1;VNDj1<ZN@f1a0]N\\Ob1d0bNXO^1h0ROhNn0Y1ROfNn0Z1SOeNm0[1TOeNk0\\1UOcNk0]1VObNj0^1XO`Nh0`1YO_Ng0b1ZO]Ne0c1\\O\\Nd0d1^OZNb0g1^OXNb0h1@VN`0k1ATN>l1GoM9R2LhM4X2OeM1[20dM0]21bMN^24`MLa24^MLb26\\MJe27YMIg2:SMIn2:jLJV3;aLI`3:YLIg3<VLBj3b0SL]On3S1bKlN^4U1cKiN^4W1bKhN^4Y1bKfN_4Y1cKeN]4[1dKdN]4[1dKdN\\4\\1eKcN[4]1gKaNZ4^1gKaNY4_1hK`NY4_1iK_NW4a1jK^NW4a1jK^NV4b1lK\\NU4c1lK\\NT4d1nKZNS4e1oKYNQ4g1PLXNP4h1RLVNo3i1RLVNn3j1TLTNm3k1TLTNl3l1VLRNk3m1WLQNi3o1XLPNi3o1XLPNh3P2YLoMg3Q2ZLnMg3Q2ZLnMf3R2ZLnMg3Q2ZLnMf3R2ZLnMg3Q2YLoMg3Q2YLoMh3P2XLPNh3P2XLPNh3P2YLoMh3\\1dJSNe1a0g3[1fJSNc1b0h3Y1gJTNa1c0h3X1iJTN_1d0i3W1iJTN_1d0h3X1jJSN^1e0i3W1jJSN]1f0i3W1jJSN]1f0j3V1jJSN\\1g0j3U1lJSNZ1h0j3U1lJSNZ1h0k3T1lJSNZ1h0j3U1lJSNZ1h0k3T1lJSNY1i0k3T1lJSNY1i0l3S1lJSNX1j0l3R1nJSNV1k0m3Q1mJTNV1k0m3Q1nJSNU1l0m3Q1nJRNV1m0m3P1nJRNU1n0m3o0oJSNT1n0n3n0oJSNT1d0ZN@c5h1lL`NkMFZ5b0RN_OoLMXO"}, "label": "orange and gold bed spread"}]}
{"id": 208165, "image": "COCO_train2014_000000208165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed that a woman and cat are laying on\"."}], "task": "refering", "answer_template": "The \"the bed that a woman and cat are laying on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 118, 119, 120, 121, 134, 135, 136, 138, 139, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 186, 187, 188, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.0, 0.3268333333333333, 1.0, 0.9979722222222221], "iscrowd": 0, "area": 73754.6944, "rle": {"size": [360, 640], "counts": "f3a1g9T3mLi2VM01O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O00001O000000001O0000001O0000001O000000000000000000001O00000XL`Kh0`4WObKh0^4XOdKf0\\4ZOeKe0[4[OfKd0Z4\\OgKc0Y4]OiKa0W4_OiKa0W4^OjKb0V4^OkKa0U4_OkKa0U4_OkKa0U4_OlK`0T4@lK`0T4@lK`0U4^OlKb0T4^OmKa0S4_OmKa0S4_OmKa0S4_OnK`0R4@nK`0R4@nK`0R4@oK?Q4AoK?Q4@PL`0P4@PL`0P4@QL?o3AQL?o3AQL?o3ARL>n3BRL>n3BRL>o3ARL>n3BSL=m3CSL=m3CTL<l3DUL;k3EUL;k3EVL:j3FWL9i3GXL8h3HXL8h3HXL8h3HXL8h3HXL8h3HYL7g3IYL7g3IYL7g3HZL8f3HZL8g3FZL:f3FZL:f3E[L;e3E[L;e3D\\L<d3D\\L<d3C]L=c3C^L<b3C_L>`3B`L>`3AaL?_3mN`KdNQ1_2_3jNdKfNm0`2_3hNgKfNk0b2^3iNgKdNk0c2^3iNgKcNl0d2]3iNhKbNk0e2]3iNiK`Nk0g2]3hNiK`Nk0g2\\3jNiK]Nl0i2[3jNjK\\Nk0j2[3jNkKZNk0l2Z3jNlKYNj0m2Z3jNXMV1h2kNWMU1i2kNWMU1i2kNWMU1j2jNVMV1j2jNVMV1j2kNUMU1k2kNUMU1k2kNUMU1k2kNUMU1l2jNTMV1l2jNTMV1l2jNTMV1l2jNTMV1l2jNTMV1l2jNTMV1m2hNTMX1l2hNTMX1l2gNUMY1k2gNUMY1k2gNUMY1k2gNUMY1l2fNTMZ1l2fNTMZ1l2fNTMZ1l2eNUM[1k2eNUM[1k2eNUM[1l2dNTM\\1l2dNTM\\1l2dNTM\\1l2cNUM]1k2cNUM]1k2cNUM]1l2bNTM^1l2_MXKc0l1n1l2XM`Ki0d1o1l2RMhKl0]1R2_3mMaLS2a3kM_LU2b3jM_LU2c3iM]LW2d3hM\\LX2f3eM\\LZ2e3eM[L[2g3cMYL]2h3bMYL]2h3_M[La2f3[M]Le2d3WM`Lh2P60001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O1O001O001O1O001O001O1O001O000000O1000000000000O1000000000000O10000000000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O2O000O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100000000001O00000000000000001O000000000000002N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O001O001O00001O001O001O001O001O0000001O0000001O0000001O0000001O0000001O0000001O00001O00001O00001O0000001O00001O00001O1O001O00000000O10000O100N2O1N2O1N2O1M3L4M3L4M3N2O1N2O1N2O1N2M3N2N2O1O100O1O100O1O1001O000000000000001O000O2O001O00001O001O00001O00001O001O000000001O00000000_KRNIm17TNHl19UNEk1;VNDj1<ZN@f1a0]N\\Ob1d0bNXO^1h0ROhNn0Y1ROfNn0Z1SOeNm0[1TOeNk0\\1UOcNk0]1VObNj0^1XO`Nh0`1YO_Ng0b1ZO]Ne0c1\\O\\Nd0d1^OZNb0g1^OXNb0h1@VN`0k1ATN>l1GoM9R2LhM4X2OeM1[20dM0]21bMN^24`MLa24^MLb26\\MJe27YMIg2:SMIn2:jLJV3;aLI`3:YLIg3<VLBj3b0SL]On3S1bKlN^4U1cKiN^4W1bKhN^4Y1bKfN_4Y1cKeN]4[1dKdN]4[1dKdN\\4\\1eKcN[4]1gKaNZ4^1gKaNY4_1hK`NY4_1iK_NW4a1jK^NW4a1jK^NV4b1lK\\NU4c1lK\\NT4d1nKZNS4e1oKYNQ4g1PLXNP4h1RLVNo3i1RLVNn3j1TLTNm3k1TLTNl3l1VLRNk3m1WLQNi3o1XLPNi3o1XLPNh3P2YLoMg3Q2ZLnMg3Q2ZLnMf3R2ZLnMg3Q2ZLnMf3R2ZLnMg3Q2YLoMg3Q2YLoMh3P2XLPNh3P2XLPNh3P2YLoMh3\\1dJSNe1a0g3[1fJSNc1b0h3Y1gJTNa1c0h3X1iJTN_1d0i3W1iJTN_1d0h3X1jJSN^1e0i3W1jJSN]1f0i3W1jJSN]1f0j3V1jJSN\\1g0j3U1lJSNZ1h0j3U1lJSNZ1h0k3T1lJSNZ1h0j3U1lJSNZ1h0k3T1lJSNY1i0k3T1lJSNY1i0l3S1lJSNX1j0l3R1nJSNV1k0m3Q1mJTNV1k0m3Q1nJSNU1l0m3Q1nJRNV1m0m3P1nJRNU1n0m3o0oJSNT1n0n3n0oJSNT1d0ZN@c5h1lL`NkMFZ5b0RN_OoLMXO"}, "label": "the bed that a woman and cat are laying on"}]}
{"id": 355621, "image": "COCO_train2014_000000355621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing black t - shirt , white pant and holding the juice glass\"."}], "task": "refering", "answer_template": "The \"a person wearing black t - shirt , white pant and holding the juice glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 131, 132, 145, 146, 147, 159, 160, 161, 162, 163, 175, 176, 177, 178, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 219, 220, 221, 222, 223, 235, 236, 237, 238, 250, 251, 252, 253, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 313, 314, 329], "bbox": [0.6163700234192038, 0.31460937499999997, 1.0, 0.937078125], "iscrowd": 0, "area": 28311.194550000007, "rle": {"size": [640, 427], "counts": "QhT54_c0?K6K4K5L5L3M3M4M2M3M4M1N1O2]NWNl@j1o>[NPAf1l>]NSAf1h>^NWAc1e>`N[Ab1a>aN^A`1a>aN]Aa1b>_N^Ac1a>]N^Ad1h3hMh5d0`Fe1\\3RNT6:^Fg1Z3RNW67_Fh1W3SN[65]Fj1U3SN]63^Fk1R3UN_61]Fl1R3TNb60[Fn1P3TNd6N]FQ2i2SNj6M\\FT2X2^N\\7]O]FX2S2[N`7^O]FZ2o1YNc7]O^F^2l1TNe7@_Fa2f1PNj7_OaFk2Z1fMT8AaFU3<oKUO[1m9BbF^30gK@Y1m9CdFh3B^KLX1l9DeFQ4VOVK9T1l9EfFa4^OjKj9GhF_4]OkKj9FiF`4]OiKi9IjF^4\\OiKj9IkF^4[OhKj9JlF^4ZOgKi9LmF^4YOgKi9KoF^4XOfKi9LPG^4WOeKa96XGU4VOeK[9=_Go3ROgKX9a0gG\\5R8kJoGT5P8mJQHR5o7nJQHR5o7nJRHQ5o7oJQHP5o7PKQHP5o7PKRHo4o7PKRHP5m7PKTHn4m7RKSHl4o7UKQHb4X8]KiGZ4_8fKaG]3\\9dLdFU3c9jL^Fo2h9RMWFh2o9YMQF`2V:`MjEY2\\:gMeER2a:oM^Ek1i:UNWEi1j:WNWEg1j:dMiCQO\\1[3l:cMjCQOZ1b3f:[MSDROi0W4P;cLZDUOe0^4n:YL`DXOa0e4l:PLeDZO>m4h:fKmD\\O:T5f:\\KSE_O6[5d:RKXEBFE1l5m:jJ_EDBG1Q6k:`JeEG]OJ2V6h:UJkEKYOL3Z6i:hInE1UOO2^6Q<bIkC23c6m;ZIoC43f6l;UIoC74g6l;`ISDc6l;\\ISDg6l;XISDk6l;TITDo6i;f0O1O1O2N1O1O1O2N1O1O1O1O2N1O0000jHQEZ5o:fJZEQ5g:nJ_El4a:TKfEe4Z:[KoE\\4R:dKQFW4P:iKSFT4m9lKVFP4k9oKZFl3i9PLZFo3g9nK\\FP4g9mKZFR4i9jKZFT4i9iKXFW4i9fKZFY4f9fK[FY4g9dK[F\\4f9aK]F^4c<01N101O^1bN6I7J5Kj0VO?@8I3M3M4K3N001O1N101O1O1N101O1O0O2O1O1O0O0M3K4L4LS^O"}, "label": "a person wearing black t - shirt , white pant and holding the juice glass"}]}
{"id": 355621, "image": "COCO_train2014_000000355621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with juice glass\"."}], "task": "refering", "answer_template": "The \"man with juice glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 131, 132, 145, 146, 147, 159, 160, 161, 162, 163, 175, 176, 177, 178, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 219, 220, 221, 222, 223, 235, 236, 237, 238, 250, 251, 252, 253, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 313, 314, 329], "bbox": [0.6163700234192038, 0.31460937499999997, 1.0, 0.937078125], "iscrowd": 0, "area": 28311.194550000007, "rle": {"size": [640, 427], "counts": "QhT54_c0?K6K4K5L5L3M3M4M2M3M4M1N1O2]NWNl@j1o>[NPAf1l>]NSAf1h>^NWAc1e>`N[Ab1a>aN^A`1a>aN]Aa1b>_N^Ac1a>]N^Ad1h3hMh5d0`Fe1\\3RNT6:^Fg1Z3RNW67_Fh1W3SN[65]Fj1U3SN]63^Fk1R3UN_61]Fl1R3TNb60[Fn1P3TNd6N]FQ2i2SNj6M\\FT2X2^N\\7]O]FX2S2[N`7^O]FZ2o1YNc7]O^F^2l1TNe7@_Fa2f1PNj7_OaFk2Z1fMT8AaFU3<oKUO[1m9BbF^30gK@Y1m9CdFh3B^KLX1l9DeFQ4VOVK9T1l9EfFa4^OjKj9GhF_4]OkKj9FiF`4]OiKi9IjF^4\\OiKj9IkF^4[OhKj9JlF^4ZOgKi9LmF^4YOgKi9KoF^4XOfKi9LPG^4WOeKa96XGU4VOeK[9=_Go3ROgKX9a0gG\\5R8kJoGT5P8mJQHR5o7nJQHR5o7nJRHQ5o7oJQHP5o7PKQHP5o7PKRHo4o7PKRHP5m7PKTHn4m7RKSHl4o7UKQHb4X8]KiGZ4_8fKaG]3\\9dLdFU3c9jL^Fo2h9RMWFh2o9YMQF`2V:`MjEY2\\:gMeER2a:oM^Ek1i:UNWEi1j:WNWEg1j:dMiCQO\\1[3l:cMjCQOZ1b3f:[MSDROi0W4P;cLZDUOe0^4n:YL`DXOa0e4l:PLeDZO>m4h:fKmD\\O:T5f:\\KSE_O6[5d:RKXEBFE1l5m:jJ_EDBG1Q6k:`JeEG]OJ2V6h:UJkEKYOL3Z6i:hInE1UOO2^6Q<bIkC23c6m;ZIoC43f6l;UIoC74g6l;`ISDc6l;\\ISDg6l;XISDk6l;TITDo6i;f0O1O1O2N1O1O1O2N1O1O1O1O2N1O0000jHQEZ5o:fJZEQ5g:nJ_El4a:TKfEe4Z:[KoE\\4R:dKQFW4P:iKSFT4m9lKVFP4k9oKZFl3i9PLZFo3g9nK\\FP4g9mKZFR4i9jKZFT4i9iKXFW4i9fKZFY4f9fK[FY4g9dK[F\\4f9aK]F^4c<01N101O^1bN6I7J5Kj0VO?@8I3M3M4K3N001O1N101O1O1N101O1O0O2O1O1O0O0M3K4L4LS^O"}, "label": "man with juice glass"}]}
{"id": 355621, "image": "COCO_train2014_000000355621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting on a sofa behind two standing people\"."}], "task": "refering", "answer_template": "The \"a woman sitting on a sofa behind two standing people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 199, 213, 233, 234, 248, 249, 263, 264, 265, 278, 279, 280, 293, 294, 308], "bbox": [0.220327868852459, 0.527203125, 0.7410772833723653, 0.8805781250000001], "iscrowd": 0, "area": 5036.17215, "rle": {"size": [640, 427], "counts": "bTk12lc02N2M4M2O1O00O2O1N2N101N2N2O0O10lb0<h\\O7I3M3M3MN21O001O1O0010O001O1N2O1O1O1Jj\\OAWc0>6O1O010O2N2L4Igdi13\\[VN6K5J7J5J6K5J7J5J6K5J2`NVNd@m1W?\\N_@j1^?]NX@i1e?^NR@g1k?Y1M3NHR@`Lm?^3W@bLg?]3\\@cLc?[3`@eL`?X3c@fL^?X3e@fL\\?MU@U3e0kLX?0U@m2Pa0H7I8H8L40O10001O0O1000010O000000N2N2M3N2N2M3N2N2M3N2N2M3N3M2M3N2N2M3N2M3N1O1N101N2O1O1N10001O000O2O0000000O10000O1MokT2"}, "label": "a woman sitting on a sofa behind two standing people"}]}
{"id": 355621, "image": "COCO_train2014_000000355621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman sitting on the couch\"."}], "task": "refering", "answer_template": "The \"the woman sitting on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 199, 213, 233, 234, 248, 249, 263, 264, 265, 278, 279, 280, 293, 294, 308], "bbox": [0.220327868852459, 0.527203125, 0.7410772833723653, 0.8805781250000001], "iscrowd": 0, "area": 5036.17215, "rle": {"size": [640, 427], "counts": "bTk12lc02N2M4M2O1O00O2O1N2N101N2N2O0O10lb0<h\\O7I3M3M3MN21O001O1O0010O001O1N2O1O1O1Jj\\OAWc0>6O1O010O2N2L4Igdi13\\[VN6K5J7J5J6K5J7J5J6K5J2`NVNd@m1W?\\N_@j1^?]NX@i1e?^NR@g1k?Y1M3NHR@`Lm?^3W@bLg?]3\\@cLc?[3`@eL`?X3c@fL^?X3e@fL\\?MU@U3e0kLX?0U@m2Pa0H7I8H8L40O10001O0O1000010O000000N2N2M3N2N2M3N2N2M3N2N2M3N3M2M3N2N2M3N2M3N1O1N101N2O1O1N10001O000O2O0000000O10000O1MokT2"}, "label": "the woman sitting on the couch"}]}
{"id": 175405, "image": "COCO_train2014_000000175405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat sitting on the edge of the bathtub\"."}], "task": "refering", "answer_template": "The \"a black and white cat sitting on the edge of the bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 118, 119, 120, 121, 122, 123, 136, 137, 139, 140, 154, 155], "bbox": [0.5861799999999999, 0.25562666666666667, 0.8745599999999999, 0.6897866666666667], "iscrowd": 0, "area": 10940.6498, "rle": {"size": [375, 500], "counts": "R_[35a;;XEBb9h0[FYO]9Q1`FQOX9Y1dFhNT9b1iF_No8l1lFWNk8S2RGkMi8b2O1O101N00000000J6C>B=G9H8G9I72N2N101N2N1O2N2N1O2N1O2N2N1O2N2N100O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2O0O100O1O2O0O1O2O0O2N100O2N100O2O0O1O2O0O1I8O01O0100O2O0O2O1N2O1O1N101N2OO0010O01O01O00001O0O10O1O1O1O1O001O100O1O010ChGZMY8a2mG^MS8\\2THbMm7Z2XHeMh7Z2ZHdMf7\\2k00N100O2N100O001001O2N3M3M2M4M3M5Ki0WO6J6J6I7J6JZmf0"}, "label": "a black and white cat sitting on the edge of the bathtub"}]}
{"id": 175405, "image": "COCO_train2014_000000175405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray and white cat sitting on the edge of a bathtub\"."}], "task": "refering", "answer_template": "The \"a gray and white cat sitting on the edge of a bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 118, 119, 120, 121, 122, 123, 136, 137, 139, 140, 154, 155], "bbox": [0.5861799999999999, 0.25562666666666667, 0.8745599999999999, 0.6897866666666667], "iscrowd": 0, "area": 10940.6498, "rle": {"size": [375, 500], "counts": "R_[35a;;XEBb9h0[FYO]9Q1`FQOX9Y1dFhNT9b1iF_No8l1lFWNk8S2RGkMi8b2O1O101N00000000J6C>B=G9H8G9I72N2N101N2N1O2N2N1O2N1O2N2N1O2N2N100O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2O0O100O1O2O0O1O2O0O2N100O2N100O2O0O1O2O0O1I8O01O0100O2O0O2O1N2O1O1N101N2OO0010O01O01O00001O0O10O1O1O1O1O001O100O1O010ChGZMY8a2mG^MS8\\2THbMm7Z2XHeMh7Z2ZHdMf7\\2k00N100O2N100O001001O2N3M3M2M4M3M5Ki0WO6J6J6I7J6JZmf0"}, "label": "a gray and white cat sitting on the edge of a bathtub"}]}
{"id": 175405, "image": "COCO_train2014_000000175405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat sitting on a wooden toilet\"."}], "task": "refering", "answer_template": "The \"a black cat sitting on a wooden toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 130, 131, 132, 133, 148, 149, 150, 151, 166, 167, 168, 169, 184, 185, 186, 187, 202, 203, 204], "bbox": [0.15816, 0.38789333333333337, 0.42928, 0.9213866666666666], "iscrowd": 0, "area": 15417.147499999997, "rle": {"size": [375, 500], "counts": "lSm08^;1O1O100O1O1O1O10O0100O100O100O010O010O01O010O001000O010O0100000O1_OEfE;:Oh8JmF793g8GoF7;5;Cm62]H6<84L9]OZ5:mI6<e0CM;SOe56QJ5=e0_O82mNo51SJ5=f0[O<1lNR6MUJ7=n2[5lLXJ6=P3X5jL\\J7<o2W5jL]J7<o2V5kL^J7;n2V5lL^J7=l2U5oMkJP2T5RNlJm1T5VNiJi1X5[NeJd1[5_NbJ`1^5cN`J\\1a5hN[JV1g5kNYJS1h5\\22N102M3M2^OQJZKR6_4g000000O2N01O0001oI[Kn4e4nJaKP5_4lJhKQ5W4kJPLT5n3iJXLU5i3fJ[LZ5e3cJ^L]5b3`JaL`5_3^JcLb5]3\\JeLc5\\3\\JeLd5g40000000O100O100O1O100O1K5G9N2N2O1O1O1O2N1N3N2N1N2O2N1O1O1O1O1N3N1N2N2O1N2O1O1O2TOeHSMND^7V3lHoL\\7n2hHoLZ7n2kHnLX7o2i0M3N3L3lMUGZ1o8cNUGX1P9dN\\Gn0m9G6J5L6J<ESUX3"}, "label": "a black cat sitting on a wooden toilet"}]}
{"id": 175405, "image": "COCO_train2014_000000175405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat lying on a toilet seat\"."}], "task": "refering", "answer_template": "The \"a cat lying on a toilet seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 130, 131, 132, 133, 148, 149, 150, 151, 166, 167, 168, 169, 184, 185, 186, 187, 202, 203, 204], "bbox": [0.15816, 0.38789333333333337, 0.42928, 0.9213866666666666], "iscrowd": 0, "area": 15417.147499999997, "rle": {"size": [375, 500], "counts": "lSm08^;1O1O100O1O1O1O10O0100O100O100O010O010O01O010O001000O010O0100000O1_OEfE;:Oh8JmF793g8GoF7;5;Cm62]H6<84L9]OZ5:mI6<e0CM;SOe56QJ5=e0_O82mNo51SJ5=f0[O<1lNR6MUJ7=n2[5lLXJ6=P3X5jL\\J7<o2W5jL]J7<o2V5kL^J7;n2V5lL^J7=l2U5oMkJP2T5RNlJm1T5VNiJi1X5[NeJd1[5_NbJ`1^5cN`J\\1a5hN[JV1g5kNYJS1h5\\22N102M3M2^OQJZKR6_4g000000O2N01O0001oI[Kn4e4nJaKP5_4lJhKQ5W4kJPLT5n3iJXLU5i3fJ[LZ5e3cJ^L]5b3`JaL`5_3^JcLb5]3\\JeLc5\\3\\JeLd5g40000000O100O100O1O100O1K5G9N2N2O1O1O1O2N1N3N2N1N2O2N1O1O1O1O1N3N1N2N2O1N2O1O1O2TOeHSMND^7V3lHoL\\7n2hHoLZ7n2kHnLX7o2i0M3N3L3lMUGZ1o8cNUGX1P9dN\\Gn0m9G6J5L6J<ESUX3"}, "label": "a cat lying on a toilet seat"}]}
{"id": 60720, "image": "COCO_train2014_000000060720.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in a blocking position\"."}], "task": "refering", "answer_template": "The \"the man in a blocking position\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 169, 170, 171, 191, 192, 193, 214, 215, 216, 237, 238, 239, 260, 261, 262, 282, 283, 284, 285], "bbox": [0.28796875, 0.38329166666666664, 0.450359375, 0.7671458333333334], "iscrowd": 0, "area": 6281.255249999998, "rle": {"size": [480, 640], "counts": "ZRg24i>3N2M3M4L3H8N2O1OJTB[Oj=b0:010O0O2O1gNC[CF8i0\\<6XCMf<V1O2OOO2H8E;O2O2M2O2O1O1O1O1O100O1O1O1O001OL42CQMfDQ3V;?O0O2N2O0O2O1mEULo8j3PGXLP9g3nF\\LR9d3kF_LU9`3hFdLX9i1YFRO<YO[9X1TF_N9l02Ca9j0]FfN2k0KJg9<eFnNJk0E0m90kFTOEQ2a9hNmFWOCo1a9gNnF[OAm1a9gNPG[OAl1`9gNQG]O@j1b9fNoF_O@Y1DhNP:<nFC_OU1FkNn9;nFE@Q1FoNl9:oFG_Om0IQOi9;oFJ^Of0MTOf9;PGL]Ob0OWOc9;RGK_O?OZO_9<SGJB=O\\O\\9<TGLA91_OY9;WGIeNJl0?2BU9;ZGJD60CS9;ZGKG2U:3UFJIOS:7UFIKLQ:<U3O00001O0000001O0000001O000000O1O1O1O1O2N101O01O0GbAK21_>0gAN`XU5"}, "label": "the man in a blocking position"}]}
{"id": 60720, "image": "COCO_train2014_000000060720.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with his arm in the air playing sports\"."}], "task": "refering", "answer_template": "The \"a woman with his arm in the air playing sports\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 169, 170, 171, 191, 192, 193, 214, 215, 216, 237, 238, 239, 260, 261, 262, 282, 283, 284, 285], "bbox": [0.28796875, 0.38329166666666664, 0.450359375, 0.7671458333333334], "iscrowd": 0, "area": 6281.255249999998, "rle": {"size": [480, 640], "counts": "ZRg24i>3N2M3M4L3H8N2O1OJTB[Oj=b0:010O0O2O1gNC[CF8i0\\<6XCMf<V1O2OOO2H8E;O2O2M2O2O1O1O1O1O100O1O1O1O001OL42CQMfDQ3V;?O0O2N2O0O2O1mEULo8j3PGXLP9g3nF\\LR9d3kF_LU9`3hFdLX9i1YFRO<YO[9X1TF_N9l02Ca9j0]FfN2k0KJg9<eFnNJk0E0m90kFTOEQ2a9hNmFWOCo1a9gNnF[OAm1a9gNPG[OAl1`9gNQG]O@j1b9fNoF_O@Y1DhNP:<nFC_OU1FkNn9;nFE@Q1FoNl9:oFG_Om0IQOi9;oFJ^Of0MTOf9;PGL]Ob0OWOc9;RGK_O?OZO_9<SGJB=O\\O\\9<TGLA91_OY9;WGIeNJl0?2BU9;ZGJD60CS9;ZGKG2U:3UFJIOS:7UFIKLQ:<U3O00001O0000001O0000001O000000O1O1O1O1O2N101O01O0GbAK21_>0gAN`XU5"}, "label": "a woman with his arm in the air playing sports"}]}
{"id": 462129, "image": "COCO_train2014_000000462129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in red under umbrella\"."}], "task": "refering", "answer_template": "The \"person in red under umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 225, 226, 248, 249, 250, 271, 272, 273, 293, 294, 295, 296, 297, 316, 317, 318, 319, 339, 340, 362, 363, 364, 365, 385, 386, 387, 388, 409, 410, 411, 432, 433, 434, 456, 457, 458, 479, 480, 481, 502, 503, 504], "bbox": [0.7640156250000001, 0.37048437500000003, 0.9367968750000001, 0.9524531249999999], "iscrowd": 0, "area": 19626.840150000007, "rle": {"size": [640, 640], "counts": "fna99ec06J4M2M3N3b_OWO`=k0\\BXOd=j0bA]O_NLm?i0`A@]NKS`0f0d@YO@k0k?N`@l0j>QNUAT1FU1U?hMSAU1BX1[?dMRAh3n>YLPAi3m>[LQAe3n>]LPAd3o>]LRAb3m>_LSAa3l>`LTA`3a=ULZCd0jNZ3j=XLWCa4g<cKVC^4]<QL`CP4_<TL]CU21kM`<4[Ce1d0XNm;:XCY1S1_N`;:[CX1W1`NZ;9]CX1[1aNS;8aCX1\\1eNm:5eCW1_1gNi:1gCY1`1jNe:NiCZ1b1lN`:JnCZ1b1PO\\:GPDZ1e1SOW:BSD\\1f1VOS:[2nEhMS1mMn5[4oHkMn0nMR6V4QImMi0PNU6S4RIoMg0nMW6S4SIQNc0mMY6R4TISNa0kM\\6Q4SIVN>jM_6P4SIYN;hMa6o3UIZN8gMd6n3TI^N5dMg6n3TI`N3bMi6n3TIbN0bMk6l3UIeNM_Mo6k3UIgNJ_MP7j3VIjNG]MR7i3WImNC[MW7g3VIPOAZMX7f3WISO^OWM[7f3XIVOXOUM`7e3XI\\OQOQMg7b3XICjNlLn7a3XIGfNhLR8a3XIIcNgLV8_3WIL`NfLY8^3WIM\\NiL]8Y3WI0WNjLb8V3WIS1i6mNWIS1j6lNVIU1a7SN_Hm1a7SN_Hm1b7RN^Hm1d7QN]Ho1g7nJ^Ff2j1]2R8bMnG^2S8aMmG_2S8aMmG_2T8_MmGa2T8^MlGb2T8^MkGc2V8\\MjGd2W8ZMiGg2W8YMiGg2X8XMgGi2Z8VMfGj2[8TMeGm2\\8RMdGoM\\NY4Q:gMbGQN_NV4P:gM`GUNbN=Fe2Y:hN\\GXN]O@SO_3U:hNYGTN2R3f8iNVGmM?Y3\\8kN]IS1d6nN\\IP1e6jNcIS1`6hNfIT1^6gNgIU1\\6gNiIU1[6eNlIV1X6eNnIV1V6eNoIW1V6bNQJY1S6`NUJ[1P6oMfJl1l;K5K5K5K5J6K5J>C3M3L6KS]h0"}, "label": "person in red under umbrella"}]}
{"id": 462129, "image": "COCO_train2014_000000462129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a red jacket , black pants and white tennis shoes holding a black umbrella waiting to cross the street\"."}], "task": "refering", "answer_template": "The \"a lady in a red jacket , black pants and white tennis shoes holding a black umbrella waiting to cross the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 225, 226, 248, 249, 250, 271, 272, 273, 293, 294, 295, 296, 297, 316, 317, 318, 319, 339, 340, 362, 363, 364, 365, 385, 386, 387, 388, 409, 410, 411, 432, 433, 434, 456, 457, 458, 479, 480, 481, 502, 503, 504], "bbox": [0.7640156250000001, 0.37048437500000003, 0.9367968750000001, 0.9524531249999999], "iscrowd": 0, "area": 19626.840150000007, "rle": {"size": [640, 640], "counts": "fna99ec06J4M2M3N3b_OWO`=k0\\BXOd=j0bA]O_NLm?i0`A@]NKS`0f0d@YO@k0k?N`@l0j>QNUAT1FU1U?hMSAU1BX1[?dMRAh3n>YLPAi3m>[LQAe3n>]LPAd3o>]LRAb3m>_LSAa3l>`LTA`3a=ULZCd0jNZ3j=XLWCa4g<cKVC^4]<QL`CP4_<TL]CU21kM`<4[Ce1d0XNm;:XCY1S1_N`;:[CX1W1`NZ;9]CX1[1aNS;8aCX1\\1eNm:5eCW1_1gNi:1gCY1`1jNe:NiCZ1b1lN`:JnCZ1b1PO\\:GPDZ1e1SOW:BSD\\1f1VOS:[2nEhMS1mMn5[4oHkMn0nMR6V4QImMi0PNU6S4RIoMg0nMW6S4SIQNc0mMY6R4TISNa0kM\\6Q4SIVN>jM_6P4SIYN;hMa6o3UIZN8gMd6n3TI^N5dMg6n3TI`N3bMi6n3TIbN0bMk6l3UIeNM_Mo6k3UIgNJ_MP7j3VIjNG]MR7i3WImNC[MW7g3VIPOAZMX7f3WISO^OWM[7f3XIVOXOUM`7e3XI\\OQOQMg7b3XICjNlLn7a3XIGfNhLR8a3XIIcNgLV8_3WIL`NfLY8^3WIM\\NiL]8Y3WI0WNjLb8V3WIS1i6mNWIS1j6lNVIU1a7SN_Hm1a7SN_Hm1b7RN^Hm1d7QN]Ho1g7nJ^Ff2j1]2R8bMnG^2S8aMmG_2S8aMmG_2T8_MmGa2T8^MlGb2T8^MkGc2V8\\MjGd2W8ZMiGg2W8YMiGg2X8XMgGi2Z8VMfGj2[8TMeGm2\\8RMdGoM\\NY4Q:gMbGQN_NV4P:gM`GUNbN=Fe2Y:hN\\GXN]O@SO_3U:hNYGTN2R3f8iNVGmM?Y3\\8kN]IS1d6nN\\IP1e6jNcIS1`6hNfIT1^6gNgIU1\\6gNiIU1[6eNlIV1X6eNnIV1V6eNoIW1V6bNQJY1S6`NUJ[1P6oMfJl1l;K5K5K5K5J6K5J>C3M3L6KS]h0"}, "label": "a lady in a red jacket , black pants and white tennis shoes holding a black umbrella waiting to cross the street"}]}
{"id": 290098, "image": "COCO_train2014_000000290098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the picture of the pizza in the background is blurry\"."}], "task": "refering", "answer_template": "The \"the picture of the pizza in the background is blurry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56], "bbox": [0.134359375, 0.0175, 0.48910937499999996, 0.1777570093457944], "iscrowd": 0, "area": 10830.224199999999, "rle": {"size": [428, 640], "counts": "_oS12Z=4K4M4L3L5L3M2M100O10001N100O10000O2O000O100O101O000O10001O0O10000O101O000O1000001N10000O10001O0O10000O2O00000O10001O00000000001O00000000001O000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000N2O1N2O1N2O1N2N2O11O00000000001O00000000001O00000000002M3N2N2N2N0000000O100000000000000O100000001O00000O10001O0000000O10001O000000000O101O00000000000O2O0O100O1O100O2N1O100O1O101N1O100O1O1O101N1O3N8G:FVdX4"}, "label": "the picture of the pizza in the background is blurry"}]}
{"id": 290098, "image": "COCO_train2014_000000290098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry pizza pie on the far side of the table\"."}], "task": "refering", "answer_template": "The \"a blurry pizza pie on the far side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56], "bbox": [0.134359375, 0.0175, 0.48910937499999996, 0.1777570093457944], "iscrowd": 0, "area": 10830.224199999999, "rle": {"size": [428, 640], "counts": "_oS12Z=4K4M4L3L5L3M2M100O10001N100O10000O2O000O100O101O000O10001O0O10000O101O000O1000001N10000O10001O0O10000O2O00000O10001O00000000001O00000000001O000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000N2O1N2O1N2O1N2N2O11O00000000001O00000000001O00000000002M3N2N2N2N0000000O100000000000000O100000001O00000O10001O0000000O10001O000000000O101O00000000000O2O0O100O1O100O2N1O100O1O101N1O100O1O1O101N1O3N8G:FVdX4"}, "label": "a blurry pizza pie on the far side of the table"}]}
{"id": 290098, "image": "COCO_train2014_000000290098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small pizza that is in focus\"."}], "task": "refering", "answer_template": "The \"the small pizza that is in focus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.052593749999999995, 0.2764018691588785, 0.9107031250000001, 0.9730373831775702], "iscrowd": 0, "area": 125374.08690000001, "rle": {"size": [428, 640], "counts": "[^>6S=8I6I8I6I8I7H7J7H7J7H7J7H7J7H6K3L3N3L4M3L4M3L4M2O2O1N2O1N2O1N2O0O2O1N2N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N2N101N2O1N2O1N2O1N101N2O1N2O1N2O0O2O1N2N2O1N2O0O2O1N2O0O2O0O100O101N100O1O101N1000000O2O00000O2O0000000O2O000000001N100000001N100000001O0O1000001O0O10001O00000O101O0000000O2O0000000O2O000000001N100000001N1000001O000O10001O000O10001O00000O101O00000O101O0000000O2O0000000O2O0000001O0O1000001O0O1000001O000O10001O000O10001O000000001O000O101O000000001O000000001O0O1000001O000000001O000000000O10001O000000000000000000000O2O00000000000000000000001N100000000000000000001O00000O10000000000000001O0000000O1001O0000000001O01O0000000000001O000001O00000001O00000000000010O000000000001O00000001O0001O0000000000001O0001O00000001O000000001O0001O01O0000001O0000001O0001O01O0000001O00001O00000O2O0O100O2O000O101N100O101N100O101O0O100O2O0O101N100O101O0O100O2O0O100O2O0O10001N100O101N100O101N10001N100O2O0O2O1N2O1N2O1N101O1N2O1N2O1N101N2O1N2O1O1N2O0O2O1N2O1N2O1N101O1N2N2N2N2M2O2N2N2N2M3N2N1O2N2M3N2N2N2N2M2O2N2N2N2M3N2N1O2N2M3N2N2N2M2O2N2N2N2M3N2N3M2N3L3N2N3M2N3L3N2N3M2N3L3N2N3M2N3L3N2N3M2N3L3N2N3M2K6G8H8H9H7Hcbg0"}, "label": "the small pizza that is in focus"}]}
{"id": 535860, "image": "COCO_train2014_000000535860.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is behind another giraffe and not under a tree\"."}], "task": "refering", "answer_template": "The \"a giraffe that is behind another giraffe and not under a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 199, 200, 201, 222, 223, 224, 245, 246, 247, 248, 268, 269, 270, 271], "bbox": [0.59865625, 0.41174999999999995, 0.793171875, 0.6995625], "iscrowd": 0, "area": 4664.619549999995, "rle": {"size": [480, 640], "counts": "fgc54k>1O2N100O1K40001O010O1N1010O1000004L01O100O1O1O11O0000000000000000O1000000000000O100O10O010O10O10O0100VCBP;>nDFQ;:lDLP;4nD1o:0mD7P;ISD_O<m0];DVDA:7C5k;3UDE:3L4a;5WDG6275V;f0\\DXOb02Q;Z1SEdNl:Z1ZEcNe:[1cE^N_:`1cE^N^:`1dE_N]:`1eE^N\\:`1jE_NS:c1iE_NW:h1aEYN_:n1WEUNj:T2_DYNa;c20MXDnLh;P3[DoLf;\\2eDfMGNd;Z2oDgMQ;b1eE_N\\:Y1kEgNU:Y1jEhNV:X1iEiNX:V1fElN[:S1cEoN^:Q1_EmNf:R1WEQOj:n0TETOm:k0QEUOR;j0lDVOW;i0gDWO\\;i0`DXOc;g0[DYOh;f0VDZOm;e0lC@W<b12N3M2N3M2N2N3H7M4M2M4O010O0101O2N5K000]CdNa;\\1]DeNe;[1WDgNk;]1nCdNT<`1eCaNP=P1iBQOY=Z13M3M2IZBPOh=;mBCV=7_Rn1"}, "label": "a giraffe that is behind another giraffe and not under a tree"}]}
{"id": 535860, "image": "COCO_train2014_000000535860.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the far right that is walking\"."}], "task": "refering", "answer_template": "The \"the giraffe on the far right that is walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 199, 200, 201, 222, 223, 224, 245, 246, 247, 248, 268, 269, 270, 271], "bbox": [0.59865625, 0.41174999999999995, 0.793171875, 0.6995625], "iscrowd": 0, "area": 4664.619549999995, "rle": {"size": [480, 640], "counts": "fgc54k>1O2N100O1K40001O010O1N1010O1000004L01O100O1O1O11O0000000000000000O1000000000000O100O10O010O10O10O0100VCBP;>nDFQ;:lDLP;4nD1o:0mD7P;ISD_O<m0];DVDA:7C5k;3UDE:3L4a;5WDG6275V;f0\\DXOb02Q;Z1SEdNl:Z1ZEcNe:[1cE^N_:`1cE^N^:`1dE_N]:`1eE^N\\:`1jE_NS:c1iE_NW:h1aEYN_:n1WEUNj:T2_DYNa;c20MXDnLh;P3[DoLf;\\2eDfMGNd;Z2oDgMQ;b1eE_N\\:Y1kEgNU:Y1jEhNV:X1iEiNX:V1fElN[:S1cEoN^:Q1_EmNf:R1WEQOj:n0TETOm:k0QEUOR;j0lDVOW;i0gDWO\\;i0`DXOc;g0[DYOh;f0VDZOm;e0lC@W<b12N3M2N3M2N2N3H7M4M2M4O010O0101O2N5K000]CdNa;\\1]DeNe;[1WDgNk;]1nCdNT<`1eCaNP=P1iBQOY=Z13M3M2IZBPOh=;mBCV=7_Rn1"}, "label": "the giraffe on the far right that is walking"}]}
{"id": 519477, "image": "COCO_train2014_000000519477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby has a yellow object in her hand\"."}], "task": "refering", "answer_template": "The \"a baby has a yellow object in her hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.07478, 0.26742937853107346, 0.8384599999999999, 0.8247175141242937], "iscrowd": 0, "area": 50457.45185, "rle": {"size": [354, 500], "counts": "T]=4i91TG3j86lFMQ9>cFFZ9R1O1N2O1N2N2N2M3M3L4M4L3M3L4L4K5L4L4K5L4K5L4L4K5N2O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O101O0O100O10000O10000O100O10000O10000O100O10000O100O10000O1000000000000O10001O000000000O10000000000000000O100000000000000O100000000000000O1000000000000000000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O0001O01O00001O00001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N1O8Hf0ZOg0YO2N001O00001O0000001O0000001O0000001O00001O0000001O0000001O001O001O1O001O001O10O01O1O001O001O1O001O1O001O1O001O001O1O001O1O002N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O2N1O1O2Nofk0"}, "label": "a baby has a yellow object in her hand"}]}
{"id": 519477, "image": "COCO_train2014_000000519477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child holding the tooth brush\"."}], "task": "refering", "answer_template": "The \"the child holding the tooth brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.07478, 0.26742937853107346, 0.8384599999999999, 0.8247175141242937], "iscrowd": 0, "area": 50457.45185, "rle": {"size": [354, 500], "counts": "T]=4i91TG3j86lFMQ9>cFFZ9R1O1N2O1N2N2N2M3M3L4M4L3M3L4L4K5L4L4K5L4K5L4L4K5N2O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O101O0O100O10000O10000O100O10000O10000O100O10000O100O10000O1000000000000O10001O000000000O10000000000000000O100000000000000O100000000000000O1000000000000000000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O0001O01O00001O00001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N1O8Hf0ZOg0YO2N001O00001O0000001O0000001O0000001O00001O0000001O0000001O001O001O1O001O001O10O01O1O001O001O1O001O1O001O1O001O001O1O001O1O002N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O2N1O1O2Nofk0"}, "label": "the child holding the tooth brush"}]}
{"id": 519477, "image": "COCO_train2014_000000519477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child in the background\"."}], "task": "refering", "answer_template": "The \"the child in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 21, 22, 23, 24, 25, 26, 27, 28, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 74, 75, 76, 81, 82, 83, 92, 93], "bbox": [0.1278, 0.0, 0.6183, 0.47062146892655365], "iscrowd": 0, "area": 19912.055699999997, "rle": {"size": [354, 500], "counts": "\\Xf0j0n9:I6J6J6L4M3M3M4L3M3M300O10000O10000O10000O2O000O10000O10000O10000O10001N10000O10000O1000000O10001N1000000001O000100O10O01O100O10O2O1O1O1O1N2O2N3M3M3L5L3M3M3M3M2M3N2N3M2N2M3N2N2N3M2N3L4M3M2N1O1O2MWMHkJ7U52cJM]5<\\JBd5?\\J_Oe5b0[J]Oe5c0\\J[Oe5f0\\JWOe5j0[JTOf5n0YJQOg5R1S24L3M4L3M3M4L3M2N3M2N2N2N2N1O2N2N1O2N2N2N3M4L3M000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000001O001O00010O010O01O01O010O01O010O01O01O010O010O0010O0100O10O01O100O01001N100O101O0O101O0O10001N100O2O000O101O0O100O2O000O2O3M3I8F:GmnQ2"}, "label": "the child in the background"}]}
{"id": 519477, "image": "COCO_train2014_000000519477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of the head of another child in the bath tub\"."}], "task": "refering", "answer_template": "The \"the top of the head of another child in the bath tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 21, 22, 23, 24, 25, 26, 27, 28, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 74, 75, 76, 81, 82, 83, 92, 93], "bbox": [0.1278, 0.0, 0.6183, 0.47062146892655365], "iscrowd": 0, "area": 19912.055699999997, "rle": {"size": [354, 500], "counts": "\\Xf0j0n9:I6J6J6L4M3M3M4L3M3M300O10000O10000O10000O2O000O10000O10000O10000O10001N10000O10000O1000000O10001N1000000001O000100O10O01O100O10O2O1O1O1O1N2O2N3M3M3L5L3M3M3M3M2M3N2N3M2N2M3N2N2N3M2N3L4M3M2N1O1O2MWMHkJ7U52cJM]5<\\JBd5?\\J_Oe5b0[J]Oe5c0\\J[Oe5f0\\JWOe5j0[JTOf5n0YJQOg5R1S24L3M4L3M3M4L3M2N3M2N2N2N2N1O2N2N1O2N2N2N3M4L3M000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000001O001O00010O010O01O01O010O01O010O01O01O010O010O0010O0100O10O01O100O01001N100O101O0O101O0O10001N100O2O000O101O0O100O2O000O2O3M3I8F:GmnQ2"}, "label": "the top of the head of another child in the bath tub"}]}
{"id": 306485, "image": "COCO_train2014_000000306485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smallest laptop\"."}], "task": "refering", "answer_template": "The \"smallest laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 236, 256, 257, 258, 259, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.030859375, 0.5980624999999999, 0.275953125, 0.875875], "iscrowd": 0, "area": 12646.071100000001, "rle": {"size": [480, 640], "counts": "`h93l>2N101N1O1O1O2O0O1O1O1O101N1O1O1O100O1O2N100O1O1O1O2O0O1O1O1O100O2N1O100O1O1O2N100O1O1O1O101N1O1O100O1O1O2O000L4L4L4L5K4L4L4K5L4L4L5K4L4L4L4L4O2M2O1O1O1000001O01O0000000001O00000001O01O00000000000010O000000000O100O1O2N1O1O1O1O1O1O1O1O2O0O1O1O1O1O1O1O1O1O2N1O1O100O1O1O1O1O2N1O1O1O1O1O100O1O2N1O1O1O1N2N2L4M3Lo0QOS1mNSfh6"}, "label": "smallest laptop"}]}
{"id": 306485, "image": "COCO_train2014_000000306485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mini laptop on top of the computer table\"."}], "task": "refering", "answer_template": "The \"a mini laptop on top of the computer table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 236, 256, 257, 258, 259, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.030859375, 0.5980624999999999, 0.275953125, 0.875875], "iscrowd": 0, "area": 12646.071100000001, "rle": {"size": [480, 640], "counts": "`h93l>2N101N1O1O1O2O0O1O1O1O101N1O1O1O100O1O2N100O1O1O1O2O0O1O1O1O100O2N1O100O1O1O2N100O1O1O1O101N1O1O100O1O1O2O000L4L4L4L5K4L4L4K5L4L4L5K4L4L4L4L4O2M2O1O1O1000001O01O0000000001O00000001O01O00000000000010O000000000O100O1O2N1O1O1O1O1O1O1O1O2O0O1O1O1O1O1O1O1O1O2N1O1O100O1O1O1O1O2N1O1O1O1O1O100O1O2N1O1O1O1N2N2L4M3Lo0QOS1mNSfh6"}, "label": "a mini laptop on top of the computer table"}]}
{"id": 306485, "image": "COCO_train2014_000000306485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor sitting between two laptops\"."}], "task": "refering", "answer_template": "The \"a computer monitor sitting between two laptops\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.27353125, 0.36941666666666667, 0.594703125, 0.8847083333333333], "iscrowd": 0, "area": 42512.544, "rle": {"size": [480, 640], "counts": "aVb2[2f<W3hL00001O0000001O000010O000001O0000001O000010O000001O0000001O00010O0000001O0000001O0001O01O00001O0000001[G]Jk7c5THkJ_7U5`HmJ`7R5_HoJa7R5^HoJa7Q5^HQKa7o4^HSKa7n4]HSKc7m4\\HUKc7k4]HVKb7j4]HWKc7j4[HXKd7h4[HZKe7e4[H\\Kd7d4[H]Ke7P6O1O1O001O1O000010O0000000000000000000010O00000000000000000001O01O00O10000O10000O2O0O10000O10000O10000O2O000O10000O10000O100O2O000O10000O10000O101M2O1O1O1O1O1O1O1O1O1N3O000001O0000001O000010O0001O00001O0000001O01O01O00001O00001O0001O01O00001O00001O000010O0N2[Oe0[Oe0[Oe0[Oe0[Oe0[Oe0[Oe0[Oe0[Of0ZO`Wi3"}, "label": "a computer monitor sitting between two laptops"}]}
{"id": 306485, "image": "COCO_train2014_000000306485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle computer screen in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the middle computer screen in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.27353125, 0.36941666666666667, 0.594703125, 0.8847083333333333], "iscrowd": 0, "area": 42512.544, "rle": {"size": [480, 640], "counts": "aVb2[2f<W3hL00001O0000001O000010O000001O0000001O000010O000001O0000001O00010O0000001O0000001O0001O01O00001O0000001[G]Jk7c5THkJ_7U5`HmJ`7R5_HoJa7R5^HoJa7Q5^HQKa7o4^HSKa7n4]HSKc7m4\\HUKc7k4]HVKb7j4]HWKc7j4[HXKd7h4[HZKe7e4[H\\Kd7d4[H]Ke7P6O1O1O001O1O000010O0000000000000000000010O00000000000000000001O01O00O10000O10000O2O0O10000O10000O10000O2O000O10000O10000O100O2O000O10000O10000O101M2O1O1O1O1O1O1O1O1O1N3O000001O0000001O000010O0001O00001O0000001O01O01O00001O00001O0001O01O00001O00001O000010O0N2[Oe0[Oe0[Oe0[Oe0[Oe0[Oe0[Oe0[Oe0[Of0ZO`Wi3"}, "label": "the middle computer screen in the right hand picture"}]}
{"id": 306485, "image": "COCO_train2014_000000306485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dell laptop\"."}], "task": "refering", "answer_template": "The \"a black dell laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 246, 247, 248, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.42325, 0.55925, 0.998875, 0.9858958333333334], "iscrowd": 0, "area": 36534.54765000001, "rle": {"size": [480, 640], "counts": "Z_o32n>4K4M4L2M1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O100N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3O1001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001L4M2M4L4L3M4L3M4L4L3M4L4L3M4L3M4M3L3M4L4L3M4L3M4L4L3M4L4L3M4L3M4M\\3"}, "label": "a black dell laptop"}]}
{"id": 306485, "image": "COCO_train2014_000000306485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black laptop right of the monitor\"."}], "task": "refering", "answer_template": "The \"black laptop right of the monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 246, 247, 248, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.42325, 0.55925, 0.998875, 0.9858958333333334], "iscrowd": 0, "area": 36534.54765000001, "rle": {"size": [480, 640], "counts": "Z_o32n>4K4M4L2M1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O100N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3O1001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001L4M2M4L4L3M4L3M4L4L3M4L4L3M4L3M4M3L3M4L4L3M4L3M4L4L3M4L4L3M4L3M4M\\3"}, "label": "black laptop right of the monitor"}]}
{"id": 240951, "image": "COCO_train2014_000000240951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle closest to the motorcycle with skulls\"."}], "task": "refering", "answer_template": "The \"the motorcycle closest to the motorcycle with skulls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 24, 25, 26, 27, 28, 29, 39, 40, 41, 42, 43, 44, 54, 55, 56, 57, 58, 59], "bbox": [0.6206588235294117, 0.0, 1.0, 0.16764062500000002], "iscrowd": 0, "area": 12196.312100000003, "rle": {"size": [640, 425], "counts": "gQU5h0Uc05K4OWOV]O>ib0^OZ]Oc0]b0[Ob]O11f0Xb0\\OS^Od0ma0\\OR^Of0na0XOQ^Oj0na0WOg]O0Ko0]b0QOg]OV1Wb0jNj]OV1`b01O1O1L4M3\\O^Ne^Of1Za0ZNd^Oh1[a0YNd^Oi1[a0YNc^Oh1[a0YNe^Oh1Za0XNf^Oh1Ya0YNh^Og1Wa0YNi^Og1Wa0ZNh^Og1Va0ZNj^Of1Va0ZNj^Og1Ta0ZNl^Of1Ta0ZNl^Of1Ta0ZNl^Og1Ra0ZNo^Oe1Qa0[No^Of1o`0[NQ_Oe1o`0\\NP_Od1Pa0\\NP_Od1o`0]NQ_Oc1o`0]NQ_Oc1n`0_NR_O`1n`0`NR_O`1m`0aNS_O_1m`0aNS_O_1m`0bNR_O^1m`0cNS_O\\1l`0fNU_OY1k`0gNU_OY1k`0gNU_OY1k`0hNS_OY1m`0hNo^O[1Qa0j00000001O0O100O100O100O101N100O2O0O2O0O2O0O2O0O2O0O2O0O2N1O1N100O010UNW^O_1ja0aNX^O]1ha0cNZ^O[1fa0dN^^OX1ca0hN_^OV1`a0kNc^OQ1^a0oNd^Oo0\\a0POh^Ol0Xa0UOj^Od0[a0\\Oh^O<\\a0Dg^O5^a0Kd^OJda07R1O101Ob]OMYO9h`0Hg_O\\1V`0cNi_Ob1T`0]Nj_Og1U`0XNj_Oj1V`0UNj_Om1U`0WNg_Oi1Y`0ZNd_Og1[`0]N`_Od1``0`N\\_Oa1c`0cNY_O]1g`0fNP_Oa1o`0h001O000000000000001O0000000000000000001O00000000000000000000001O00000000000000000000000000000000001O00"}, "label": "the motorcycle closest to the motorcycle with skulls"}]}
{"id": 240951, "image": "COCO_train2014_000000240951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle to the right of the one decorated as a skeleton\"."}], "task": "refering", "answer_template": "The \"the motorcycle to the right of the one decorated as a skeleton\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 24, 25, 26, 27, 28, 29, 39, 40, 41, 42, 43, 44, 54, 55, 56, 57, 58, 59], "bbox": [0.6206588235294117, 0.0, 1.0, 0.16764062500000002], "iscrowd": 0, "area": 12196.312100000003, "rle": {"size": [640, 425], "counts": "gQU5h0Uc05K4OWOV]O>ib0^OZ]Oc0]b0[Ob]O11f0Xb0\\OS^Od0ma0\\OR^Of0na0XOQ^Oj0na0WOg]O0Ko0]b0QOg]OV1Wb0jNj]OV1`b01O1O1L4M3\\O^Ne^Of1Za0ZNd^Oh1[a0YNd^Oi1[a0YNc^Oh1[a0YNe^Oh1Za0XNf^Oh1Ya0YNh^Og1Wa0YNi^Og1Wa0ZNh^Og1Va0ZNj^Of1Va0ZNj^Og1Ta0ZNl^Of1Ta0ZNl^Of1Ta0ZNl^Og1Ra0ZNo^Oe1Qa0[No^Of1o`0[NQ_Oe1o`0\\NP_Od1Pa0\\NP_Od1o`0]NQ_Oc1o`0]NQ_Oc1n`0_NR_O`1n`0`NR_O`1m`0aNS_O_1m`0aNS_O_1m`0bNR_O^1m`0cNS_O\\1l`0fNU_OY1k`0gNU_OY1k`0gNU_OY1k`0hNS_OY1m`0hNo^O[1Qa0j00000001O0O100O100O100O101N100O2O0O2O0O2O0O2O0O2O0O2O0O2N1O1N100O010UNW^O_1ja0aNX^O]1ha0cNZ^O[1fa0dN^^OX1ca0hN_^OV1`a0kNc^OQ1^a0oNd^Oo0\\a0POh^Ol0Xa0UOj^Od0[a0\\Oh^O<\\a0Dg^O5^a0Kd^OJda07R1O101Ob]OMYO9h`0Hg_O\\1V`0cNi_Ob1T`0]Nj_Og1U`0XNj_Oj1V`0UNj_Om1U`0WNg_Oi1Y`0ZNd_Og1[`0]N`_Od1``0`N\\_Oa1c`0cNY_O]1g`0fNP_Oa1o`0h001O000000000000001O0000000000000000001O00000000000000000000001O00000000000000000000000000000000001O00"}, "label": "the motorcycle to the right of the one decorated as a skeleton"}]}
{"id": 240951, "image": "COCO_train2014_000000240951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a personalized motorcycle featuring skulls\"."}], "task": "refering", "answer_template": "The \"a personalized motorcycle featuring skulls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 20, 21, 22, 23, 24, 35, 36, 37, 38, 50, 51, 52, 53, 65, 66, 67, 68, 69, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 201, 202, 203, 204, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 322, 323, 324, 338, 339], "bbox": [0.23896470588235294, 0.01846875, 0.7738117647058823, 0.9756562500000001], "iscrowd": 0, "area": 69630.70935, "rle": {"size": [640, 425], "counts": "\\mo13lc03L5L3M3M3L5L3O1N3M2N1O000001O0000000O101O00000000001O0000000000001O0000lNVOc^O2Hh0_a0[Oe^O2Lc0Ya0@f^O31=Sa0Dm^O01;T3Hf91VCL0;=WOl1a0[:6^CFN<;;l0]OW;;eCAM<:o0:iNd;?mC]OK<8T19dNc;b0SDYOI=7X17`Nb;e0XDVOH=5]16ZNb;i0\\DSOG<4c13VNb;j0cDPOE<2h11RNb;l0gDnND<0m10mMa;o0lD1AY1NgMa;Q1SEN]O^1LcMa;S1WELZOc1K^M`;V1\\EIWOZ2[;oM`EFTO\\2[;mMcEGPO^2[;lMfEh3Y:XLiEg3V:XLlEh3S:S3O001O2N8H8H7I8H8H8H8H8H8Ha4_Ke0[O7H2O1O1O1O2N1O1O1O1O2N1O1O1O20O00001O01O01O000N2O000O100O10000O100O3N2N2M3N3M2M3N2M2O1k1UN2N1O2N1O1O2N1O2M2O2N1O2Na@aN^<^1^CfNd<X1XClNi<S1SCQOn<n0nBVOT=h0iB[OX=d0dB@^=>^BEd=:XBJi=5SBOm=2nA2R>NjA6V>JgA9Y>GcA=]>C_Aa0a>_O[Ae0e>[OWAh0j>XORAl0n>TOn@o0S?QOj@R1V?oNi@P1X?QOg@n0Z?SOe@m0[?TOd@k0]?UOc@k0]?VOb@j0^?WOa@j0^?WOa@i0_?l101O000iMYLUDh3f;gLoCY3l;VMjCk2Q<dMdC\\2W<SN_Cm1\\<cNYCd1\\<f3WO=D8G9G8eLnFaKW9W4TGcKk8Y4`GbK^8Z4lGaKS8[4XH`Kf7]4cH^K[7`4oH[Ko6b4WI\\Kh6a4]I_Ka6^4dIaK[6V3YGiKa2Q1S6V3fG]K\\2\\1m5W3QHPKW2i1f5l0YG^OU4Db4n0[G\\OU4E_4o0_GYOT4F\\4b0\\GQN7d1T4GY4=cGXN2b1T4GW47lG_NKa1T4GU42THeNEe0BId4h0P4O\\HjNB>C1a4g0m3IeHPO\\O5J9[4g0^KkNk7i0]IWOWOH3f0R4e0UK_Og77iIVO1?\\3a1h2jNkIVO1>_3a1c2lNnISO2?_3a1`2mNoISO2>b3a1\\2nNQJRO1?d3`1Y2oNRJRO2>e3`1U2QOUJoN2?g3_1R2SOUJoN2?j3\\1h4TN^G?m3[1f4UN^G?n3Y1]6gNeIW1[6gNiIW1X6gNkIV1V6hNoIU1Q6jNQJU1o5iNUJT1m5iNVJV1j5iNXJT1j5jNYJT1[<N2N3M2N2N3M3M3L4M3M3M3M3MU_k1"}, "label": "a personalized motorcycle featuring skulls"}]}
{"id": 240951, "image": "COCO_train2014_000000240951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcycle with skulls\"."}], "task": "refering", "answer_template": "The \"motorcycle with skulls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 20, 21, 22, 23, 24, 35, 36, 37, 38, 50, 51, 52, 53, 65, 66, 67, 68, 69, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 201, 202, 203, 204, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 322, 323, 324, 338, 339], "bbox": [0.23896470588235294, 0.01846875, 0.7738117647058823, 0.9756562500000001], "iscrowd": 0, "area": 69630.70935, "rle": {"size": [640, 425], "counts": "\\mo13lc03L5L3M3M3L5L3O1N3M2N1O000001O0000000O101O00000000001O0000000000001O0000lNVOc^O2Hh0_a0[Oe^O2Lc0Ya0@f^O31=Sa0Dm^O01;T3Hf91VCL0;=WOl1a0[:6^CFN<;;l0]OW;;eCAM<:o0:iNd;?mC]OK<8T19dNc;b0SDYOI=7X17`Nb;e0XDVOH=5]16ZNb;i0\\DSOG<4c13VNb;j0cDPOE<2h11RNb;l0gDnND<0m10mMa;o0lD1AY1NgMa;Q1SEN]O^1LcMa;S1WELZOc1K^M`;V1\\EIWOZ2[;oM`EFTO\\2[;mMcEGPO^2[;lMfEh3Y:XLiEg3V:XLlEh3S:S3O001O2N8H8H7I8H8H8H8H8H8Ha4_Ke0[O7H2O1O1O1O2N1O1O1O1O2N1O1O1O20O00001O01O01O000N2O000O100O10000O100O3N2N2M3N3M2M3N2M2O1k1UN2N1O2N1O1O2N1O2M2O2N1O2Na@aN^<^1^CfNd<X1XClNi<S1SCQOn<n0nBVOT=h0iB[OX=d0dB@^=>^BEd=:XBJi=5SBOm=2nA2R>NjA6V>JgA9Y>GcA=]>C_Aa0a>_O[Ae0e>[OWAh0j>XORAl0n>TOn@o0S?QOj@R1V?oNi@P1X?QOg@n0Z?SOe@m0[?TOd@k0]?UOc@k0]?VOb@j0^?WOa@j0^?WOa@i0_?l101O000iMYLUDh3f;gLoCY3l;VMjCk2Q<dMdC\\2W<SN_Cm1\\<cNYCd1\\<f3WO=D8G9G8eLnFaKW9W4TGcKk8Y4`GbK^8Z4lGaKS8[4XH`Kf7]4cH^K[7`4oH[Ko6b4WI\\Kh6a4]I_Ka6^4dIaK[6V3YGiKa2Q1S6V3fG]K\\2\\1m5W3QHPKW2i1f5l0YG^OU4Db4n0[G\\OU4E_4o0_GYOT4F\\4b0\\GQN7d1T4GY4=cGXN2b1T4GW47lG_NKa1T4GU42THeNEe0BId4h0P4O\\HjNB>C1a4g0m3IeHPO\\O5J9[4g0^KkNk7i0]IWOWOH3f0R4e0UK_Og77iIVO1?\\3a1h2jNkIVO1>_3a1c2lNnISO2?_3a1`2mNoISO2>b3a1\\2nNQJRO1?d3`1Y2oNRJRO2>e3`1U2QOUJoN2?g3_1R2SOUJoN2?j3\\1h4TN^G?m3[1f4UN^G?n3Y1]6gNeIW1[6gNiIW1X6gNkIV1V6hNoIU1Q6jNQJU1o5iNUJT1m5iNVJV1j5iNXJT1j5jNYJT1[<N2N3M2N2N3M3M3L4M3M3M3M3MU_k1"}, "label": "motorcycle with skulls"}]}
{"id": 486713, "image": "COCO_train2014_000000486713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown horse carrying a person in white chaps and blue shirt\"."}], "task": "refering", "answer_template": "The \"a dark brown horse carrying a person in white chaps and blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 248, 249, 250, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387, 405, 406, 407, 408, 409, 410, 428, 429, 430, 431, 432], "bbox": [0.597671875, 0.5030152671755724, 0.9047812499999999, 1.0], "iscrowd": 0, "area": 21357.203849999998, "rle": {"size": [524, 640], "counts": "b]T61Y`02n_OOh?2V@2f?0X@4e?MY@O5L^?8i@IT?5QAJl>4YAMd>1aAO\\>OiA1S>NQB2k=OWB3c=0_B0\\=2gBNT=5nBJn<9TCFh<<[CCa<?`CCZ<`0fCBV<?kCCQ<?oCCl;`0SDJQ;_OiCi0V11h:?YEL]:3cE7T:HlEc0j9\\OUFk0f9TOZFn0d9SO[FP1c9oN_FQ1`9oNaFQ1^9nNeFR1Y9mNiFS1V9mNkFS1S9mNoF6SN@m::QG2cNXOZ:e0UGOQOPOi9P1XGMP:3QFIR:6PFGR:9oECS:=nE@U:?jEAX:?gE_O\\:`0dE^O_:a0`EB_:>`ED_:;aEG^:9`EJ_:5aEM^:2bE0]:0bE0`:N_EXOQNe0d<2ZE[ORN`0g<3WE_OSN9j<6REDTN3m<7oDHUNLP=:kDLUNGS=:hD2`;L`D6`;H`D:`;D_D?a;_O_Dd0`;ZO`Di0_;UO`Do0_;oNaDT1Y=3L4jCeNW9^1]DUOW1^OZ:`1]DVOU1[O[:c1_DUOR1YO]:e1`DTO<A^OHd;f1^DVO41[;m0^Db0`;A]Da0a;V2M3N2N2M3N2N2N2MaNSEeMk:Y2eE[MX:d2RFVMl9i2]FPMb9P3eFhL[9Z3jF_LX9a3mFYLS9h3h1001O1N20kNcC[N\\<g1dCWNi1POh8i2_EWNh1SOf8g2cEVNd1WOg8c2fEWN]1[Ok8`2gEVNW1@R9Z2hEVNT1BR9Y2kEYNn0@T9Y2nEZNi0^OY9X2nE[N<iNFe0o9X2PF[N7lNGb0R:X2PFSMLP17WOJ>R:Y2QFQMMR135o9h1RFoLNT1N7R:f1ZFBf9?ZFVMCn1S:l0[FTMDo1Q:m0\\FRMDQ2P:n0\\FnLFT2n9n0]FlLGU2l9o0gFQOY9P1fFPOZ9o0hFPOW9n0lFROT9n0lFROT9n0mFQOR9P1nFPOQ9R1nFnNQ9S1PGlNn8V1RGjNm8W1SGiNk8Y1VGfNi8\\1VGdNi8]1WGcNi8]1WGcNh8^1XGbNh8_1WGaNh8`1XG`Nh8`1XG`Nh8`1XG]Ne8j1ZGRNa8W2_GfM_8_2aG^M_8e2aGXMa8j2^GSMd8n2\\GoLQ8eNeG]4:kLU8jN^G[4=fLZ8ROUGZ4`0_La8WOnFZ4a0[Le8]OhFX4c0ZLg8_OdFW4e0XLi8C_FW4g0ULl8Z4TGdKn8\\4RGbKQ9]4oFbKR9^4nF`KU9`4iF`KX9`4hF^KZ9b4fF]K\\9b4dF\\K_9d4`F[Kb9d4^FZKe9e4[FYKf9h4ZFWKg9j4XFRKl9n4i000001O00RMaEO_:KkEMY:1mEHW:8kECW:>lE[OX:d0lETOX:m0QFfNR:[1WFWNm9i1[21O0I426J9H0O2NnMYBa1g=^N[Bb1d=\\N^Bd1f=VN]Bi1f=RN`Bj1S>M2O2M3N1_N\\A10m0X?L5K4YO[@a0l?Fnln0"}, "label": "a dark brown horse carrying a person in white chaps and blue shirt"}]}
{"id": 486713, "image": "COCO_train2014_000000486713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a darm brown horse ridden by a lady in blue\"."}], "task": "refering", "answer_template": "The \"a darm brown horse ridden by a lady in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 248, 249, 250, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387, 405, 406, 407, 408, 409, 410, 428, 429, 430, 431, 432], "bbox": [0.597671875, 0.5030152671755724, 0.9047812499999999, 1.0], "iscrowd": 0, "area": 21357.203849999998, "rle": {"size": [524, 640], "counts": "b]T61Y`02n_OOh?2V@2f?0X@4e?MY@O5L^?8i@IT?5QAJl>4YAMd>1aAO\\>OiA1S>NQB2k=OWB3c=0_B0\\=2gBNT=5nBJn<9TCFh<<[CCa<?`CCZ<`0fCBV<?kCCQ<?oCCl;`0SDJQ;_OiCi0V11h:?YEL]:3cE7T:HlEc0j9\\OUFk0f9TOZFn0d9SO[FP1c9oN_FQ1`9oNaFQ1^9nNeFR1Y9mNiFS1V9mNkFS1S9mNoF6SN@m::QG2cNXOZ:e0UGOQOPOi9P1XGMP:3QFIR:6PFGR:9oECS:=nE@U:?jEAX:?gE_O\\:`0dE^O_:a0`EB_:>`ED_:;aEG^:9`EJ_:5aEM^:2bE0]:0bE0`:N_EXOQNe0d<2ZE[ORN`0g<3WE_OSN9j<6REDTN3m<7oDHUNLP=:kDLUNGS=:hD2`;L`D6`;H`D:`;D_D?a;_O_Dd0`;ZO`Di0_;UO`Do0_;oNaDT1Y=3L4jCeNW9^1]DUOW1^OZ:`1]DVOU1[O[:c1_DUOR1YO]:e1`DTO<A^OHd;f1^DVO41[;m0^Db0`;A]Da0a;V2M3N2N2M3N2N2N2MaNSEeMk:Y2eE[MX:d2RFVMl9i2]FPMb9P3eFhL[9Z3jF_LX9a3mFYLS9h3h1001O1N20kNcC[N\\<g1dCWNi1POh8i2_EWNh1SOf8g2cEVNd1WOg8c2fEWN]1[Ok8`2gEVNW1@R9Z2hEVNT1BR9Y2kEYNn0@T9Y2nEZNi0^OY9X2nE[N<iNFe0o9X2PF[N7lNGb0R:X2PFSMLP17WOJ>R:Y2QFQMMR135o9h1RFoLNT1N7R:f1ZFBf9?ZFVMCn1S:l0[FTMDo1Q:m0\\FRMDQ2P:n0\\FnLFT2n9n0]FlLGU2l9o0gFQOY9P1fFPOZ9o0hFPOW9n0lFROT9n0lFROT9n0mFQOR9P1nFPOQ9R1nFnNQ9S1PGlNn8V1RGjNm8W1SGiNk8Y1VGfNi8\\1VGdNi8]1WGcNi8]1WGcNh8^1XGbNh8_1WGaNh8`1XG`Nh8`1XG`Nh8`1XG]Ne8j1ZGRNa8W2_GfM_8_2aG^M_8e2aGXMa8j2^GSMd8n2\\GoLQ8eNeG]4:kLU8jN^G[4=fLZ8ROUGZ4`0_La8WOnFZ4a0[Le8]OhFX4c0ZLg8_OdFW4e0XLi8C_FW4g0ULl8Z4TGdKn8\\4RGbKQ9]4oFbKR9^4nF`KU9`4iF`KX9`4hF^KZ9b4fF]K\\9b4dF\\K_9d4`F[Kb9d4^FZKe9e4[FYKf9h4ZFWKg9j4XFRKl9n4i000001O00RMaEO_:KkEMY:1mEHW:8kECW:>lE[OX:d0lETOX:m0QFfNR:[1WFWNm9i1[21O0I426J9H0O2NnMYBa1g=^N[Bb1d=\\N^Bd1f=VN]Bi1f=RN`Bj1S>M2O2M3N1_N\\A10m0X?L5K4YO[@a0l?Fnln0"}, "label": "a darm brown horse ridden by a lady in blue"}]}
{"id": 486713, "image": "COCO_train2014_000000486713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse with a blue sash\"."}], "task": "refering", "answer_template": "The \"a white horse with a blue sash\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 216, 217, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 355, 356, 359, 378, 379, 380, 401, 402, 403, 424], "bbox": [0.38684375000000004, 0.4257251908396947, 0.700921875, 0.9740648854961832], "iscrowd": 0, "area": 24090.745899999994, "rle": {"size": [524, 640], "counts": "kUo34V`04L6J9G9@<I5K5K5K6K4BmM^BW2_=mM[BX2b=:M3N2N21O001O1O001NZC[Mh;e2SDiMNAn:o2fD_NW;^3N3RHoJZ4S5aKoJ`4U5YKnJg4V5SKlJm4X5lJjJU5Y5dJkJ[5Y5_JjJ`5Y5[JiJe5X5XJjJg5X5VJjJi5Y5hITKX6m4oHkKP7h610O100O010OeGdHg7]7RHkHk7n7L4L3M4K5L3VMTGoMP9o1`G_Md8_2eGVM`8f2hGmL`8Q3jG_L^8`3dFXLc:h3XEcLc:[3YEoLb:Q3YEYMb:f2[EbMa:Y4M2L5L3M00001O00000O2O0OVNQFXMo9e2UFZMk9d2WF\\Mh9c2\\FZMe9c2`F[M_9d2eFYM\\9e2gFYMY9e2kFYMV9e2nFYMQ9e2SGYMn8e2TGZMl8e2VG[Mj8b2ZG\\Mf8c2\\G\\Me8b2^G]Ma8a2bG^M_8`2cG_M]8_2gG_MY8`2iG`MW8]2mGaMS8^2oGaMR8]2QHbMn7\\2VHbMj7]2XHbMh7]2[HaMe7_2]H`Mb7_2bHYMc7f2R3O100O1O11O00010O001O00001O00001O00001O0O2O001O001O1O001O001N101O1N101O1N2O2M2O1O1N2O1O2M2O1NoNlCnMR<o1RDoMP<n1SDQNm;n1UDPNm;n1UDQNk;n1VDRNk;l1WDQNk;m1YDPNh;o1ZDoMh;o1[DmMg;R2[DlMf;S2bDeM_;Z2Z1N3L3N2M3N2N2N2M3N2N2O2N2O0O2N2N1O2O1N1O2N2N2O1N2O1O1N101N101N101N101N1O3M3M3M3KWnQ3"}, "label": "a white horse with a blue sash"}]}
{"id": 486713, "image": "COCO_train2014_000000486713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light gray horse in the center\"."}], "task": "refering", "answer_template": "The \"a light gray horse in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 216, 217, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 355, 356, 359, 378, 379, 380, 401, 402, 403, 424], "bbox": [0.38684375000000004, 0.4257251908396947, 0.700921875, 0.9740648854961832], "iscrowd": 0, "area": 24090.745899999994, "rle": {"size": [524, 640], "counts": "kUo34V`04L6J9G9@<I5K5K5K6K4BmM^BW2_=mM[BX2b=:M3N2N21O001O1O001NZC[Mh;e2SDiMNAn:o2fD_NW;^3N3RHoJZ4S5aKoJ`4U5YKnJg4V5SKlJm4X5lJjJU5Y5dJkJ[5Y5_JjJ`5Y5[JiJe5X5XJjJg5X5VJjJi5Y5hITKX6m4oHkKP7h610O100O010OeGdHg7]7RHkHk7n7L4L3M4K5L3VMTGoMP9o1`G_Md8_2eGVM`8f2hGmL`8Q3jG_L^8`3dFXLc:h3XEcLc:[3YEoLb:Q3YEYMb:f2[EbMa:Y4M2L5L3M00001O00000O2O0OVNQFXMo9e2UFZMk9d2WF\\Mh9c2\\FZMe9c2`F[M_9d2eFYM\\9e2gFYMY9e2kFYMV9e2nFYMQ9e2SGYMn8e2TGZMl8e2VG[Mj8b2ZG\\Mf8c2\\G\\Me8b2^G]Ma8a2bG^M_8`2cG_M]8_2gG_MY8`2iG`MW8]2mGaMS8^2oGaMR8]2QHbMn7\\2VHbMj7]2XHbMh7]2[HaMe7_2]H`Mb7_2bHYMc7f2R3O100O1O11O00010O001O00001O00001O00001O0O2O001O001O1O001O001N101O1N101O1N2O2M2O1O1N2O1O2M2O1NoNlCnMR<o1RDoMP<n1SDQNm;n1UDPNm;n1UDQNk;n1VDRNk;l1WDQNk;m1YDPNh;o1ZDoMh;o1[DmMg;R2[DlMf;S2bDeM_;Z2Z1N3L3N2M3N2N2N2M3N2N2O2N2O0O2N2N1O2O1N1O2N2N2O1N2O1O1N101N101N101N101N1O3M3M3M3KWnQ3"}, "label": "a light gray horse in the center"}]}
{"id": 208189, "image": "COCO_train2014_000000208189.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the black coat with the red umbrella\"."}], "task": "refering", "answer_template": "The \"the woman in the black coat with the red umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 375, 376, 377, 378, 379, 398, 399, 400, 401, 420, 421, 422, 423, 443, 444, 446], "bbox": [0.206734375, 0.137078125, 0.5033749999999999, 0.8539375], "iscrowd": 0, "area": 49564.156850000014, "rle": {"size": [640, 640], "counts": "^kb21jc05K5K5K5K5K4K6K5K5K5K4L5K5K5K4K6K5J6K5J5K6K5J6K5J5K6K5VNRL\\CT4a<\\LmBh3R=gL[B_3h;eKaEk5]:aJWEa5g:dJVE\\5i:hJSEX5l:mJQES5m:RKoDo4o:VKnDj4R;YKjDh4V;[KfDe4[;^KbDb4^;_K`Db4`;o10O1001O2N1O2N2N2N2hIYGS1h8kNYGU1i8gNZGX1h8dNZG\\1h8`N[G_1g8]N[Gc1i8WNZGh1j8SNWGn1l8lMWGS2m8gMVGX2o8bMRG^2S9[MPGd2V9TMmFk2Z9mLgFS3`9dLcF[3c9]L`Fb3g9VLZFj3m9mKVFR4P:gKRFX4U:_KmEa4[:jJSFU5]<O000000000000O2O1O1O1O2N1N2O1O1O1N2L4J6K5J6\\Od0M3N2N2N2N2M4UDWHZ;Z8M2N3L3N3O0010O01O010O02O1N3N8G3N0O1O100O1O10O0O1O1N3N1O1O2M2O2N3kN^GfFe8n8eGPG^8m8dGSG^8j8cGUG`8g8bGYG_8e8aG[Gb8a8`G_Gb8^8_GaGc8]8^G_Gh8]8ZG_Gj8`8WG]Gm8a8TG[GQ9b8QGZGT9V1QFo5l0hHW9V1RFn5R;oITEl5n:QJWEk5k:RJ[Ei5f:UJ_Eg5c:VJcEe5_:XJhEb5Z:[JmE_5U:^JQF]5Q:`JVFZ5l9cJ[FW5g9fJ_FU5b9iJeFQ5]9lJjFn4X9oJnFl4Y<J6J5K6I6K6J6I6K6J3L5G9K4L5J5L5K4L5J5L5K4L5J6K4L5J5L:FmYV6"}, "label": "the woman in the black coat with the red umbrella"}]}
{"id": 208189, "image": "COCO_train2014_000000208189.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman walking and wearing a long black coat and standing under a pink umbrella\"."}], "task": "refering", "answer_template": "The \"a woman walking and wearing a long black coat and standing under a pink umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 375, 376, 377, 378, 379, 398, 399, 400, 401, 420, 421, 422, 423, 443, 444, 446], "bbox": [0.206734375, 0.137078125, 0.5033749999999999, 0.8539375], "iscrowd": 0, "area": 49564.156850000014, "rle": {"size": [640, 640], "counts": "^kb21jc05K5K5K5K5K4K6K5K5K5K4L5K5K5K4K6K5J6K5J5K6K5J6K5J5K6K5VNRL\\CT4a<\\LmBh3R=gL[B_3h;eKaEk5]:aJWEa5g:dJVE\\5i:hJSEX5l:mJQES5m:RKoDo4o:VKnDj4R;YKjDh4V;[KfDe4[;^KbDb4^;_K`Db4`;o10O1001O2N1O2N2N2N2hIYGS1h8kNYGU1i8gNZGX1h8dNZG\\1h8`N[G_1g8]N[Gc1i8WNZGh1j8SNWGn1l8lMWGS2m8gMVGX2o8bMRG^2S9[MPGd2V9TMmFk2Z9mLgFS3`9dLcF[3c9]L`Fb3g9VLZFj3m9mKVFR4P:gKRFX4U:_KmEa4[:jJSFU5]<O000000000000O2O1O1O1O2N1N2O1O1O1N2L4J6K5J6\\Od0M3N2N2N2N2M4UDWHZ;Z8M2N3L3N3O0010O01O010O02O1N3N8G3N0O1O100O1O10O0O1O1N3N1O1O2M2O2N3kN^GfFe8n8eGPG^8m8dGSG^8j8cGUG`8g8bGYG_8e8aG[Gb8a8`G_Gb8^8_GaGc8]8^G_Gh8]8ZG_Gj8`8WG]Gm8a8TG[GQ9b8QGZGT9V1QFo5l0hHW9V1RFn5R;oITEl5n:QJWEk5k:RJ[Ei5f:UJ_Eg5c:VJcEe5_:XJhEb5Z:[JmE_5U:^JQF]5Q:`JVFZ5l9cJ[FW5g9fJ_FU5b9iJeFQ5]9lJjFn4X9oJnFl4Y<J6J5K6I6K6J6I6K6J3L5G9K4L5J5L5K4L5J5L5K4L5J6K4L5J5L:FmYV6"}, "label": "a woman walking and wearing a long black coat and standing under a pink umbrella"}]}
{"id": 159038, "image": "COCO_train2014_000000159038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white platter without brown food on it\"."}], "task": "refering", "answer_template": "The \"white platter without brown food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220], "bbox": [0.0, 0.65432, 0.24630000000000002, 0.9851733333333333], "iscrowd": 0, "area": 12036.141599999999, "rle": {"size": [375, 500], "counts": "g7j3l7100O11O001O1O001O001O1O001O1_LVHS3W80000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000001O1O1O2N1O1O1O2N1O1O1O2N3M4L4L3M4L4L4L4L4L4L4L4L4L4L4L5K5K6J5K5K6JafY4"}, "label": "white platter without brown food on it"}]}
{"id": 159038, "image": "COCO_train2014_000000159038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty plate on left\"."}], "task": "refering", "answer_template": "The \"empty plate on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220], "bbox": [0.0, 0.65432, 0.24630000000000002, 0.9851733333333333], "iscrowd": 0, "area": 12036.141599999999, "rle": {"size": [375, 500], "counts": "g7j3l7100O11O001O1O001O001O1O001O1_LVHS3W80000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000001O1O1O2N1O1O1O2N1O1O1O2N3M4L4L3M4L4L4L4L4L4L4L4L4L4L4L5K5K6J5K5K6JafY4"}, "label": "empty plate on left"}]}
{"id": 159038, "image": "COCO_train2014_000000159038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"meat on middle plate\"."}], "task": "refering", "answer_template": "The \"meat on middle plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208, 209, 220, 221, 222, 223, 224, 225, 226, 227, 228], "bbox": [0.21622, 0.70384, 0.69932, 0.98312], "iscrowd": 0, "area": 18053.395250000005, "rle": {"size": [375, 500], "counts": "UjW11g;5L4L4K5L4K5I7J6J6J6J6J6I7J6J6J5K00000O10000000000O10000000O1000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O100000000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O001O2N1O1O1O1O1O1O1O2N1O1O1O1N2O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O001O001O00Snf1"}, "label": "meat on middle plate"}]}
{"id": 159038, "image": "COCO_train2014_000000159038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate in the middle\"."}], "task": "refering", "answer_template": "The \"the plate in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208, 209, 220, 221, 222, 223, 224, 225, 226, 227, 228], "bbox": [0.21622, 0.70384, 0.69932, 0.98312], "iscrowd": 0, "area": 18053.395250000005, "rle": {"size": [375, 500], "counts": "UjW11g;5L4L4K5L4K5I7J6J6J6J6J6I7J6J6J5K00000O10000000000O10000000O1000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O100000000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O001O2N1O1O1O1O1O1O1O2N1O1O1O1N2O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O001O001O00Snf1"}, "label": "the plate in the middle"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the racquet on the far right held up by the shortest person\"."}], "task": "refering", "answer_template": "The \"the racquet on the far right held up by the shortest person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 157, 158, 180, 203], "bbox": [0.832828125, 0.16430913348946136, 0.98921875, 0.5958782201405152], "iscrowd": 0, "area": 7153.9198, "rle": {"size": [427, 640], "counts": "S`n61W=4M2N2M3N0O10000O100O1PMDSI<k6KPI4P72jHOU77gHHX7>eH@Y7g0eHVOZ7P1dHmN[7Y1cHdN\\7b1aH\\N]7l1l16jDmNl9Z2O1N2O1O1O1O00O100O1O1O1O1O1O100O1000O2O0100O100O100O100O1O100O100O100O101N100O100O100O100O100O1O100O1O1N3N2M3M4M2M3N3L3M3N2M4M2M3M4M2M3M4M2M3M3K6J5K5J7J[j2"}, "label": "the racquet on the far right held up by the shortest person"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis racket being held by the girl on the righr\"."}], "task": "refering", "answer_template": "The \"a tennis racket being held by the girl on the righr\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 157, 158, 180, 203], "bbox": [0.832828125, 0.16430913348946136, 0.98921875, 0.5958782201405152], "iscrowd": 0, "area": 7153.9198, "rle": {"size": [427, 640], "counts": "S`n61W=4M2N2M3N0O10000O100O1PMDSI<k6KPI4P72jHOU77gHHX7>eH@Y7g0eHVOZ7P1dHmN[7Y1cHdN\\7b1aH\\N]7l1l16jDmNl9Z2O1N2O1O1O1O00O100O1O1O1O1O1O100O1000O2O0100O100O100O100O1O100O100O100O101N100O100O100O100O100O1O100O1O1N3N2M3M4M2M3N3L3M3N2M4M2M3M4M2M3M4M2M3M3K6J5K5J7J[j2"}, "label": "a tennis racket being held by the girl on the righr"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy who has his tennis racket held up next to two girls\"."}], "task": "refering", "answer_template": "The \"a young boy who has his tennis racket held up next to two girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 117, 118, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330], "bbox": [0.089953125, 0.30337236533957845, 0.42579687499999996, 0.9820140515222481], "iscrowd": 0, "area": 35269.462499999994, "rle": {"size": [427, 640], "counts": "eZh03V=3M4L3M4L3M4N1N3N1O2M2O2N1N3N1O2M2O2N2M2O2N1N3N2N1N3N1O2M3N:F9F7J6J5NO5L4K2OO00001O001O001ON100O010O100O100O100O100O00100O100O100O2O1N2O1N2O1N>C1NXERNS:n1fE\\NW:e1bEeNZ3\\O_3m2oKbMQ4d2ZKkMg4Q3gIdMY6g401O0000000000000001O000000000000000000001O0000fNhIaKX6]4QJ\\Ko5b4ZJWKf5h4^JUKb5i4aJVK_5h4dJWK\\5g4gJXKY5g4iJYKV5f4lJYKT5f4nJYKR5f4PKYKP5f4QKZKo4e4SKZKm4e4UKYKl4f4VKXKk4h4VKVKk4i4WKUKj4j4XKUKh4j4ZKTKh4j4YKUKh4j4ZKTKg4k4m100O1O1O1O1O100000000000000000000001O000000000000000000001O001O001O001O001O10O01O001O001O001O1O001O001O0mHbKY5_4dJcK\\5]4bJfK]5[4`JgK`5Y4^JjKa5W4\\JkKd5V4ZJkKf5U4XJnKg5S4VJoKj5Q4TJRLk5o3RJSLn4iNaKT5_OULh4QOgKk4^OWLb4YOmK`4_OYL]4@RLP6e3[JXLe5a3dJ]L\\5b3gJ\\LZ5b3jJ[LV5d3mJZLT5d3oJZLQ5e3QKZLo4e3TKYLm4e3^2N2N2N2N2N3M3M4K5L4L5K4L4L\\1dNa0_O>BRgh4"}, "label": "a young boy who has his tennis racket held up next to two girls"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the male tennis player who is holding his racket in the air\"."}], "task": "refering", "answer_template": "The \"the male tennis player who is holding his racket in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 117, 118, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330], "bbox": [0.089953125, 0.30337236533957845, 0.42579687499999996, 0.9820140515222481], "iscrowd": 0, "area": 35269.462499999994, "rle": {"size": [427, 640], "counts": "eZh03V=3M4L3M4L3M4N1N3N1O2M2O2N1N3N1O2M2O2N2M2O2N1N3N2N1N3N1O2M3N:F9F7J6J5NO5L4K2OO00001O001O001ON100O010O100O100O100O100O00100O100O100O2O1N2O1N2O1N>C1NXERNS:n1fE\\NW:e1bEeNZ3\\O_3m2oKbMQ4d2ZKkMg4Q3gIdMY6g401O0000000000000001O000000000000000000001O0000fNhIaKX6]4QJ\\Ko5b4ZJWKf5h4^JUKb5i4aJVK_5h4dJWK\\5g4gJXKY5g4iJYKV5f4lJYKT5f4nJYKR5f4PKYKP5f4QKZKo4e4SKZKm4e4UKYKl4f4VKXKk4h4VKVKk4i4WKUKj4j4XKUKh4j4ZKTKh4j4YKUKh4j4ZKTKg4k4m100O1O1O1O1O100000000000000000000001O000000000000000000001O001O001O001O001O10O01O001O001O001O1O001O001O0mHbKY5_4dJcK\\5]4bJfK]5[4`JgK`5Y4^JjKa5W4\\JkKd5V4ZJkKf5U4XJnKg5S4VJoKj5Q4TJRLk5o3RJSLn4iNaKT5_OULh4QOgKk4^OWLb4YOmK`4_OYL]4@RLP6e3[JXLe5a3dJ]L\\5b3gJ\\LZ5b3jJ[LV5d3mJZLT5d3oJZLQ5e3QKZLo4e3TKYLm4e3^2N2N2N2N2N3M3M4K5L4L5K4L4L\\1dNa0_O>BRgh4"}, "label": "the male tennis player who is holding his racket in the air"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player wearing glasses and holding a racket to her chest\"."}], "task": "refering", "answer_template": "The \"a tennis player wearing glasses and holding a racket to her chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 196, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.39642187500000003, 0.3677049180327869, 0.590890625, 0.9843091334894614], "iscrowd": 0, "area": 23336.24535, "rle": {"size": [427, 640], "counts": "WXZ32X=2M3N3L3N2[MBYHa0f7ESH=k7InG;P8KiG7U8OeG3Z83^G07TOa6S1QIK3\\Oa45nKi0WOJMCg43RLg0SOEHLQ5LQLi0oNBD3[5FoKk0kNP1U5XNmKo0fNk0\\5YNlKo3S4TLjKm3U4WLhKi3W4ZLfKg3Z4ZLeKf3Z4]LcKd3\\4_LaKb3^4aL`K_3`4lLTKU3k4nLSKR3m4PMPKQ3o4RMoJn2Q5TMlJm2S5UMlJk2S5g200O100000000000000000000000000001O000000001O001O001O001O001O1O001O1O001O1bIfJi4[5TKiJj4W5TKnJi4S5TKQKj4P5SKTKk4m4QKVKo4n4kJUKT5o4eJTK[5P5^JTKa5P5XJSKh5X601O00001O001O001O1O001O001O1O001O001O1O001O1O001O2N1O1O1O1O1O1O1O1O1SMSH:n7B]H6d7FfH3[7HQI0U7GVI1R7DXI5W:H8H8H7I8HSk\\3"}, "label": "a tennis player wearing glasses and holding a racket to her chest"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl holds a racket to her chest and stands between two other people\"."}], "task": "refering", "answer_template": "The \"a girl holds a racket to her chest and stands between two other people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 196, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.39642187500000003, 0.3677049180327869, 0.590890625, 0.9843091334894614], "iscrowd": 0, "area": 23336.24535, "rle": {"size": [427, 640], "counts": "WXZ32X=2M3N3L3N2[MBYHa0f7ESH=k7InG;P8KiG7U8OeG3Z83^G07TOa6S1QIK3\\Oa45nKi0WOJMCg43RLg0SOEHLQ5LQLi0oNBD3[5FoKk0kNP1U5XNmKo0fNk0\\5YNlKo3S4TLjKm3U4WLhKi3W4ZLfKg3Z4ZLeKf3Z4]LcKd3\\4_LaKb3^4aL`K_3`4lLTKU3k4nLSKR3m4PMPKQ3o4RMoJn2Q5TMlJm2S5UMlJk2S5g200O100000000000000000000000000001O000000001O001O001O001O001O1O001O1O001O1bIfJi4[5TKiJj4W5TKnJi4S5TKQKj4P5SKTKk4m4QKVKo4n4kJUKT5o4eJTK[5P5^JTKa5P5XJSKh5X601O00001O001O001O1O001O001O1O001O001O1O001O1O001O2N1O1O1O1O1O1O1O1O1SMSH:n7B]H6d7FfH3[7HQI0U7GVI1R7DXI5W:H8H8H7I8HSk\\3"}, "label": "a girl holds a racket to her chest and stands between two other people"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl pointing her hand towards right corner of the picture\"."}], "task": "refering", "answer_template": "The \"the girl pointing her hand towards right corner of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 173, 174, 175, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338], "bbox": [0.5289843750000001, 0.45882903981264633, 0.8829375, 0.9889227166276346], "iscrowd": 0, "area": 18973.91395, "rle": {"size": [427, 640], "counts": "fb]49Q=1N3M3M2N3M2N3N1N3N1O2L3N3L4M2N3O0O2N101N101O1O001O001O0ZG_Nb5b1hIZOQ6f0nHb0k6_OSHi1f7U2I6J7I7I7I7I5K01O001O1O100O001O1O1O1O1O1O1O2N1O1O2N2N1fJSIT4o6hKUIV4m6fKVIY4l6dKTI]4n6_KSIb4P7YKQIh4h7100O1O100O1000000000000000000000000000000000000000000000000000001O0000QLZHY2f7gMdHo1\\7PNoHf1Q7ZNYI\\1g6cNdIS1\\6mNgIP1Y6oNkIn0U6ROoIj0Q6UOSJh0m5XOVJe0j5ZOZJc0f5]O^J?b5@bJ=^5CeJ:[5EjJ7V5ImJ4S5KQK2o4NTKOl40YKLg44\\KIc48`KE`4:dKC\\4=hK_OW4a0mK\\OS4d0PLYOo3g0VLUOj3i0[LTOe3i0^LWOa3h0bLWO^3g0eLXO[3f0hLYOW3h0jLWOV3h0lLWOS3j0g4O10O0100O0010O0100O010O10O01O010O01O010O010O01O010O001O01O01O001O010O001O010O00001O010O001O010O001O00010O000001O000000001O000000001O000100O010O010O0010O0100O010O11N101O001OO1O2M2N3N1NdRo0"}, "label": "the girl pointing her hand towards right corner of the picture"}]}
{"id": 298306, "image": "COCO_train2014_000000298306.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the right\"."}], "task": "refering", "answer_template": "The \"the girl in the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 173, 174, 175, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338], "bbox": [0.5289843750000001, 0.45882903981264633, 0.8829375, 0.9889227166276346], "iscrowd": 0, "area": 18973.91395, "rle": {"size": [427, 640], "counts": "fb]49Q=1N3M3M2N3M2N3N1N3N1O2L3N3L4M2N3O0O2N101N101O1O001O001O0ZG_Nb5b1hIZOQ6f0nHb0k6_OSHi1f7U2I6J7I7I7I7I5K01O001O1O100O001O1O1O1O1O1O1O2N1O1O2N2N1fJSIT4o6hKUIV4m6fKVIY4l6dKTI]4n6_KSIb4P7YKQIh4h7100O1O100O1000000000000000000000000000000000000000000000000000001O0000QLZHY2f7gMdHo1\\7PNoHf1Q7ZNYI\\1g6cNdIS1\\6mNgIP1Y6oNkIn0U6ROoIj0Q6UOSJh0m5XOVJe0j5ZOZJc0f5]O^J?b5@bJ=^5CeJ:[5EjJ7V5ImJ4S5KQK2o4NTKOl40YKLg44\\KIc48`KE`4:dKC\\4=hK_OW4a0mK\\OS4d0PLYOo3g0VLUOj3i0[LTOe3i0^LWOa3h0bLWO^3g0eLXO[3f0hLYOW3h0jLWOV3h0lLWOS3j0g4O10O0100O0010O0100O010O10O01O010O01O010O010O01O010O001O01O01O001O010O001O010O00001O010O001O010O001O00010O000001O000000001O000000001O000100O010O010O0010O0100O010O11N101O001OO1O2M2N3N1NdRo0"}, "label": "the girl in the right"}]}
{"id": 232770, "image": "COCO_train2014_000000232770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a volunteer sticker\"."}], "task": "refering", "answer_template": "The \"a man with a volunteer sticker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389, 405, 406, 407, 408, 409, 410, 411, 412, 429, 430, 431, 432, 433, 451, 452, 453, 454, 455, 456, 474, 475, 476, 477, 478, 479, 480, 497, 498, 499, 500, 501, 502, 503, 520, 521, 522, 523, 524, 525, 526], "bbox": [0.42696875, 0.083140625, 0.9460625, 0.9887656250000001], "iscrowd": 0, "area": 104108.66984999996, "rle": {"size": [640, 640], "counts": "`P[52kc06I7J6J6I5L3L4N3M2N3L3N2N3L3N3bMVNbBl1]=XN^Bj1b=ZNWBj1g=[NSBg1l=]NoAe1P>`NiAd1U>T2O1O101N1O1O1O2N100O1O2N1O1cJXKaKi4^4nKkJT4S5hLRJX3m5kLPJV3o5mLnIU3P6nLmIS3R6PMkIQ3T6QMjIP3U6SMhIo2V6TMhIl2W6WMfIj2Y6YMdIh2[6ZMcIh2[6[MbIf2]6]M`Id2_6_M_Ia2`6bM]I`2a6bM]I_2b6dM[I]2d6fMYI[2f6hMWIZ2g6iMVIX2i6kMUIU2j6mMTIT2k6oMRIR2m6PNQIR2m6QNPIP2o6RNoHo1P7SNnHn1Q7UNmHk1R7WNlHj1S7XNkHi1T7ZNiHh1V7YNhHh1X7YNfHh1Z7ZNcHg1\\7[NcHe1]7\\NaHe1_7]N^Hd1b7]N\\He1b7]N[He1e7]NWHe1i7\\NTHf1l7[NQHg1o7[NmGg1R8[NkGg1U8S5O00000O10002N3L4M3M3M3M3L3N1O2N1O1N2O100O1O1O2O0O1O1O100O1O101N1O1O100O1O1O2O0O1O1iMgBfMY=T2lBlMU=o1oBQNR=i1SCWNc<o1aCRNo;X2VDhM[;c2iD]Mh:n2\\ERMY:T3lEmLo9S3UFmLh9Q3]FoL_9Q3eFoLX9o2mFRMS9i2QGWMo8b2XG^Mh8Z2`GgM_8Q2iGoMX8h1PHXNP8a1WH_Ni7Y1_HhN`7P1hHQOX7f0PI\\On6<ZIFd68^IJ_65cIMX64jINQ63QJOk52VJ0e51]J1^50dJ2X5NjJ4Q5NPK4k4MWK5g4I[K9b4F`K<^4CcK?[4_OgKc0V4\\OlKf0R4XOPLj0m3VOTLm0i3ROXLQ1hM^LQ3a2WOU1aM^LV3^2YOW1ZM_L]3Z2YO[1SM^Lc3X2ZO^1lL^Li3S2\\Ob1eL^Lo3P2\\Od1_LaLT4k1]Od1[LeLY4f1\\Of1VLiL^4`1\\Og1SLlLc4[1ZOi1oKPMh4V1YOj1kKUMl4o0ZOl1fKYMR5i0XOo1aK]MW5c0XOP2]KaM\\5>WOQ2YKfM`58WOR2UKjMf52UOT2RKmMj5NTOV2mJQNn5IUOc4l0[KUOf4k0YKUOg4k0YKUOh4k0WKUOi4l0VKTOk4k0UKUOk4l0TKTOl4m0RKTOo4k0QKUOo4l0PKTOQ5l0nJTOR5l0nJTOS5l0lJTOT5m0kJSOU5m0jJTOW5l0hJTOX5m0gJSOZ5m0eJSO[5n0dJRO]5o0aJQO_5P1`JPO`5Q1_JoNb5Q1\\JPOd5Q1[JoNf5R1XJnNh5S1WJmNj5S1UJmNk5T1TJlNl5U1SJkNn5U1PJlNP6V1nIjNR6W1mIiNT6X1eImN[6U1^IPOb6S1UISOk6o0nHVOS7k600001O00001O0000001O00001O00001O0000001O00001O00001O0000001O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O0_KmGUMT8j2cH_L^7`3YIiKh6V4iIYKX6g4kIUKV6j4nIRKR6n4QJoJP6Q5SJkJQ6Q5SJkJR6P5RJlJS6o4PJnJT6o4oImJV6n4nInJV6n4nInJW6m4lIPKY6l4jIPKZ6l4jIPK[6k4hIRK]6j4fIRK^6j4fIRK_6i4eISK`6h4cIUKb6g4aIUKd6f4`IVKe6e4_IWKe6e4`4L4K5K4M3L4L4L4M2M4L4L4M3L4L3N3L4K5J6J5J7J6J6I7J5K6I7J6J6I6Kd]e0"}, "label": "a man with a volunteer sticker"}]}
{"id": 232770, "image": "COCO_train2014_000000232770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bald headed man with a flowered shirt\"."}], "task": "refering", "answer_template": "The \"bald headed man with a flowered shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389, 405, 406, 407, 408, 409, 410, 411, 412, 429, 430, 431, 432, 433, 451, 452, 453, 454, 455, 456, 474, 475, 476, 477, 478, 479, 480, 497, 498, 499, 500, 501, 502, 503, 520, 521, 522, 523, 524, 525, 526], "bbox": [0.42696875, 0.083140625, 0.9460625, 0.9887656250000001], "iscrowd": 0, "area": 104108.66984999996, "rle": {"size": [640, 640], "counts": "`P[52kc06I7J6J6I5L3L4N3M2N3L3N2N3L3N3bMVNbBl1]=XN^Bj1b=ZNWBj1g=[NSBg1l=]NoAe1P>`NiAd1U>T2O1O101N1O1O1O2N100O1O2N1O1cJXKaKi4^4nKkJT4S5hLRJX3m5kLPJV3o5mLnIU3P6nLmIS3R6PMkIQ3T6QMjIP3U6SMhIo2V6TMhIl2W6WMfIj2Y6YMdIh2[6ZMcIh2[6[MbIf2]6]M`Id2_6_M_Ia2`6bM]I`2a6bM]I_2b6dM[I]2d6fMYI[2f6hMWIZ2g6iMVIX2i6kMUIU2j6mMTIT2k6oMRIR2m6PNQIR2m6QNPIP2o6RNoHo1P7SNnHn1Q7UNmHk1R7WNlHj1S7XNkHi1T7ZNiHh1V7YNhHh1X7YNfHh1Z7ZNcHg1\\7[NcHe1]7\\NaHe1_7]N^Hd1b7]N\\He1b7]N[He1e7]NWHe1i7\\NTHf1l7[NQHg1o7[NmGg1R8[NkGg1U8S5O00000O10002N3L4M3M3M3M3L3N1O2N1O1N2O100O1O1O2O0O1O1O100O1O101N1O1O100O1O1O2O0O1O1iMgBfMY=T2lBlMU=o1oBQNR=i1SCWNc<o1aCRNo;X2VDhM[;c2iD]Mh:n2\\ERMY:T3lEmLo9S3UFmLh9Q3]FoL_9Q3eFoLX9o2mFRMS9i2QGWMo8b2XG^Mh8Z2`GgM_8Q2iGoMX8h1PHXNP8a1WH_Ni7Y1_HhN`7P1hHQOX7f0PI\\On6<ZIFd68^IJ_65cIMX64jINQ63QJOk52VJ0e51]J1^50dJ2X5NjJ4Q5NPK4k4MWK5g4I[K9b4F`K<^4CcK?[4_OgKc0V4\\OlKf0R4XOPLj0m3VOTLm0i3ROXLQ1hM^LQ3a2WOU1aM^LV3^2YOW1ZM_L]3Z2YO[1SM^Lc3X2ZO^1lL^Li3S2\\Ob1eL^Lo3P2\\Od1_LaLT4k1]Od1[LeLY4f1\\Of1VLiL^4`1\\Og1SLlLc4[1ZOi1oKPMh4V1YOj1kKUMl4o0ZOl1fKYMR5i0XOo1aK]MW5c0XOP2]KaM\\5>WOQ2YKfM`58WOR2UKjMf52UOT2RKmMj5NTOV2mJQNn5IUOc4l0[KUOf4k0YKUOg4k0YKUOh4k0WKUOi4l0VKTOk4k0UKUOk4l0TKTOl4m0RKTOo4k0QKUOo4l0PKTOQ5l0nJTOR5l0nJTOS5l0lJTOT5m0kJSOU5m0jJTOW5l0hJTOX5m0gJSOZ5m0eJSO[5n0dJRO]5o0aJQO_5P1`JPO`5Q1_JoNb5Q1\\JPOd5Q1[JoNf5R1XJnNh5S1WJmNj5S1UJmNk5T1TJlNl5U1SJkNn5U1PJlNP6V1nIjNR6W1mIiNT6X1eImN[6U1^IPOb6S1UISOk6o0nHVOS7k600001O00001O0000001O00001O00001O0000001O00001O00001O0000001O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O0_KmGUMT8j2cH_L^7`3YIiKh6V4iIYKX6g4kIUKV6j4nIRKR6n4QJoJP6Q5SJkJQ6Q5SJkJR6P5RJlJS6o4PJnJT6o4oImJV6n4nInJV6n4nInJW6m4lIPKY6l4jIPKZ6l4jIPK[6k4hIRK]6j4fIRK^6j4fIRK_6i4eISK`6h4cIUKb6g4aIUKd6f4`IVKe6e4_IWKe6e4`4L4K5K4M3L4L4L4M2M4L4L4M3L4L3N3L4K5J6J5J7J6J6I7J5K6I7J6J6I6Kd]e0"}, "label": "bald headed man with a flowered shirt"}]}
{"id": 232770, "image": "COCO_train2014_000000232770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in a blue shirt having a man write on an orange\"."}], "task": "refering", "answer_template": "The \"a young boy in a blue shirt having a man write on an orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520], "bbox": [0.033, 0.32360937500000003, 0.758890625, 0.9860468749999999], "iscrowd": 0, "area": 102015.84600000003, "rle": {"size": [640, 640], "counts": "Qg=1lc04L3M4L4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3N200O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O100O100O100O100O10000O100O100O100O100O100O100O1_KdI^K\\6`4jI[KW6c4PJXKP6e4WJWKi5g4^JSKc5k4dJPK\\5n4oJhJR5V5eKUJ[4h5PLoIQ4o5RLPJn3n5ULPJl3n5WLQJi3m5[LPJf3m5^LRJb3l5aLRJ`3l5cLSJ]3k5fLSJ[3k5iLSJW3j5mLTJT3j5oLUJQ3i5RMUJo2i5TMVJl2h5WMWJi2f5\\MWJe2g5^MXJb2f5aMXJ`2h5aMWJ_2h5cMVJ^2j5dMTJ\\2l5eMRJ\\2m5fMRJZ2n5gMPJZ2P6k500O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O1O001O001QIaH[1_7cNdH\\1]7bNdH^1\\7`NfH`1[7]NhHb1X7\\NjHd1W7YNkHg1V7VNlHj1T7UNnHj1S7SNoHm1Q7QNQIo1P7nMSIQ2m6mMUIS2l6jMVIV2j6iMXIV2i6gMYIY2h6dMZI\\2f6bM\\I^2e6_M^I`2b6^M`Ib2c6ZM^If2d6VM_Ii2d6RM^In2d6nL^IR3e6iL^IV3d6gL]IY3f6bL\\I^3f6^L\\Ib3g6YKhGiNd1m5f6RKnGnN]1P6h6iJTHROV1U6h6aJ[HVOo0X6i6XJbH[Og0]6i6PJiH_O?a6k6gISKY6o4^IXKb6S901O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O0000001O2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N1O1O1O001O1O1O1O1O1O1O00VOj0N2NO1N1N3N2N2N1O2M3N2N1O2M3N2N1O2N2M3N4L4L4K6K4L4L5]Nb10001O1O1N2O1O1O1N2O1O2N1O1N2O1O1O1O1N100000O0100000O10O10O1000O10O10000O0100000O01000O1000O10O10000O01O1O2N2N2N2M3N3M2N2N2M4M6J6JR^P3"}, "label": "a young boy in a blue shirt having a man write on an orange"}]}
{"id": 232770, "image": "COCO_train2014_000000232770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy holding a lemon\"."}], "task": "refering", "answer_template": "The \"a boy holding a lemon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520], "bbox": [0.033, 0.32360937500000003, 0.758890625, 0.9860468749999999], "iscrowd": 0, "area": 102015.84600000003, "rle": {"size": [640, 640], "counts": "Qg=1lc04L3M4L4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L3M4L4L3M4L4L3M4L4L3M4L4L3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3N200O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O100O100O100O100O10000O100O100O100O100O100O100O1_KdI^K\\6`4jI[KW6c4PJXKP6e4WJWKi5g4^JSKc5k4dJPK\\5n4oJhJR5V5eKUJ[4h5PLoIQ4o5RLPJn3n5ULPJl3n5WLQJi3m5[LPJf3m5^LRJb3l5aLRJ`3l5cLSJ]3k5fLSJ[3k5iLSJW3j5mLTJT3j5oLUJQ3i5RMUJo2i5TMVJl2h5WMWJi2f5\\MWJe2g5^MXJb2f5aMXJ`2h5aMWJ_2h5cMVJ^2j5dMTJ\\2l5eMRJ\\2m5fMRJZ2n5gMPJZ2P6k500O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O1O001O001QIaH[1_7cNdH\\1]7bNdH^1\\7`NfH`1[7]NhHb1X7\\NjHd1W7YNkHg1V7VNlHj1T7UNnHj1S7SNoHm1Q7QNQIo1P7nMSIQ2m6mMUIS2l6jMVIV2j6iMXIV2i6gMYIY2h6dMZI\\2f6bM\\I^2e6_M^I`2b6^M`Ib2c6ZM^If2d6VM_Ii2d6RM^In2d6nL^IR3e6iL^IV3d6gL]IY3f6bL\\I^3f6^L\\Ib3g6YKhGiNd1m5f6RKnGnN]1P6h6iJTHROV1U6h6aJ[HVOo0X6i6XJbH[Og0]6i6PJiH_O?a6k6gISKY6o4^IXKb6S901O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O0000001O2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N1O1O1O001O1O1O1O1O1O1O00VOj0N2NO1N1N3N2N2N1O2M3N2N1O2M3N2N1O2N2M3N4L4L4K6K4L4L5]Nb10001O1O1N2O1O1O1N2O1O2N1O1N2O1O1O1O1N100000O0100000O10O10O1000O10O10000O0100000O01000O1000O10O10000O01O1O2N2N2N2M3N3M2N2N2M4M6J6JR^P3"}, "label": "a boy holding a lemon"}]}
{"id": 478532, "image": "COCO_train2014_000000478532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver car with a california license plate\"."}], "task": "refering", "answer_template": "The \"a silver car with a california license plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 275], "bbox": [0.7990781250000001, 0.4834272300469484, 0.9972031250000001, 0.753075117370892], "iscrowd": 0, "area": 10614.981950000001, "rle": {"size": [426, 640], "counts": "Ukd63T=h0XO4L3M3M3M3M4L3M3M3M3M4M2000000O2O0000000O2O00000O101O000O10001O0O1000001N1000001N1000000O2O00000O101O000O10001O000O10001O0O1000001N1000000O2O00000O101O000O10001O0O1000001N1000000O2O00000O101O00000O101O000O10001O0O1000001N1000000O2O00000O101O000OcLdFn2\\9PMkGk1n:mN\\a0"}, "label": "a silver car with a california license plate"}]}
{"id": 478532, "image": "COCO_train2014_000000478532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front end of the toyota car\"."}], "task": "refering", "answer_template": "The \"the front end of the toyota car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 275], "bbox": [0.7990781250000001, 0.4834272300469484, 0.9972031250000001, 0.753075117370892], "iscrowd": 0, "area": 10614.981950000001, "rle": {"size": [426, 640], "counts": "Ukd63T=h0XO4L3M3M3M3M4L3M3M3M3M4M2000000O2O0000000O2O00000O101O000O10001O0O1000001N1000001N1000000O2O00000O101O000O10001O000O10001O0O1000001N1000000O2O00000O101O000O10001O0O1000001N1000000O2O00000O101O00000O101O000O10001O0O1000001N1000000O2O00000O101O000OcLdFn2\\9PMkGk1n:mN\\a0"}, "label": "the front end of the toyota car"}]}
{"id": 298312, "image": "COCO_train2014_000000298312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses running on a beach\"."}], "task": "refering", "answer_template": "The \"a man with glasses running on a beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 131, 132, 133, 134, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 319, 320, 338, 339, 340, 342, 343], "bbox": [0.71315625, 0.24912499999999999, 0.9350781250000001, 0.8847083333333333], "iscrowd": 0, "area": 18698.157300000003, "rle": {"size": [480, 640], "counts": "hTf61m>5M2N1O1O2PJFVM;f2NTM3k2ORM3m2NRM39EcL9S329HbL7S32:JaL6S30;M`L4T3O<O^L2V3N<b0mK@g3N<c0lK_Oh3N;f0kK]Oi3M<g0jK\\Oj3M;j0iKYOk3N<j0hKXOl3N;m0hKUO_1BZO<`3n0fKTO\\1IZO6c3P1eKROZ1OYO0h3o0eKSOV14YOKl3o0eKROS19WOHP4n0fKROl0a0ZO@T4n0fKQOe0j0\\OYOY4m0eKQO>T1_OnN_4o0cKoN8i1VOZNo4o0bKnN2U2UOQNW5m0bKmNK]2XOkM\\5k0`KnNDm3l4VM`KmN]OS4T5PM^KnNWOW4\\5lL]KY4c4hK\\KX4e4hKeJaNHg5c5iK`JP5`5X21O0O2O1O001O1O1O1N2O1O0ZOYJSHh5\\6jJfIGK`5Z6PLdIQ4U6a2G9K5M3N2M3N2M3N2M4M2N2N2N2N2N1O2O12N4L9G1N2O1O0O2001O1O1O1O1N3N1N2N2N4L4L1O100O1N3N5K5K5K4K6K5J6L3L5L4L4K9H7I9F4M3M4K3N3M2M3N1O2M3N14L1N1O1N2iB`Ni<l11N2O1N3N1M4M1N2N2O1N2L5E;L4K5]OfA:Zhc0"}, "label": "a man with glasses running on a beach"}]}
{"id": 298312, "image": "COCO_train2014_000000298312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man jogging on the beach\"."}], "task": "refering", "answer_template": "The \"a man jogging on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 131, 132, 133, 134, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 319, 320, 338, 339, 340, 342, 343], "bbox": [0.71315625, 0.24912499999999999, 0.9350781250000001, 0.8847083333333333], "iscrowd": 0, "area": 18698.157300000003, "rle": {"size": [480, 640], "counts": "hTf61m>5M2N1O1O2PJFVM;f2NTM3k2ORM3m2NRM39EcL9S329HbL7S32:JaL6S30;M`L4T3O<O^L2V3N<b0mK@g3N<c0lK_Oh3N;f0kK]Oi3M<g0jK\\Oj3M;j0iKYOk3N<j0hKXOl3N;m0hKUO_1BZO<`3n0fKTO\\1IZO6c3P1eKROZ1OYO0h3o0eKSOV14YOKl3o0eKROS19WOHP4n0fKROl0a0ZO@T4n0fKQOe0j0\\OYOY4m0eKQO>T1_OnN_4o0cKoN8i1VOZNo4o0bKnN2U2UOQNW5m0bKmNK]2XOkM\\5k0`KnNDm3l4VM`KmN]OS4T5PM^KnNWOW4\\5lL]KY4c4hK\\KX4e4hKeJaNHg5c5iK`JP5`5X21O0O2O1O001O1O1O1N2O1O0ZOYJSHh5\\6jJfIGK`5Z6PLdIQ4U6a2G9K5M3N2M3N2M3N2M4M2N2N2N2N2N1O2O12N4L9G1N2O1O0O2001O1O1O1O1N3N1N2N2N4L4L1O100O1N3N5K5K5K4K6K5J6L3L5L4L4K9H7I9F4M3M4K3N3M2M3N1O2M3N14L1N1O1N2iB`Ni<l11N2O1N3N1M4M1N2N2O1N2L5E;L4K5]OfA:Zhc0"}, "label": "a man jogging on the beach"}]}
{"id": 298312, "image": "COCO_train2014_000000298312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a striped shirt\"."}], "task": "refering", "answer_template": "The \"a boy in a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 193, 194, 195, 216, 217, 218, 239, 240, 241, 242, 262, 263, 264, 285, 286, 287, 308, 309, 310, 332, 333, 334, 355, 356, 357], "bbox": [0.399390625, 0.4101666666666667, 0.53240625, 0.9190416666666666], "iscrowd": 0, "area": 13818.568550000004, "rle": {"size": [480, 640], "counts": "gYh3^1e<l0D=M3M3N10O200O0G:KjNnMhEl1\\:XNaEa1d:dNZET1h:TOVEk0e:^OXE`0j:EPE;P;JlD9>]Nd9^1iEg0LZNR:R1nEV2X9T2E7H6YHdI`6b6XI`Ih6d6SI]Im6f6nH\\IR7i6gHZIX7j6bHXI^7U70001O000000001O0000001O000001N100O1cNbHWK_7k4cHQK^7P5dHkJ]7W5fHcJ\\7]5PIVJP7l5R1QK^G`3b8\\L`Gf3`8VLcGk3^8PLdGR4\\8iKhGV4^8bKdG_4`8ZKcGe4\\910O010005K5K1O10O0O1O100O2QN^F\\Nc9_O]Fc06N]9^OcF879W9]OiFN8c0Q9\\OeHc0]7[OeHd0\\7YOfHf0\\7XOfHf0]7VOgHMXM5S:KaIOa6NfII_61\\4JTl[4"}, "label": "a boy in a striped shirt"}]}
{"id": 298312, "image": "COCO_train2014_000000298312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy running on a beach near the ocean\"."}], "task": "refering", "answer_template": "The \"a little boy running on a beach near the ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 193, 194, 195, 216, 217, 218, 239, 240, 241, 242, 262, 263, 264, 285, 286, 287, 308, 309, 310, 332, 333, 334, 355, 356, 357], "bbox": [0.399390625, 0.4101666666666667, 0.53240625, 0.9190416666666666], "iscrowd": 0, "area": 13818.568550000004, "rle": {"size": [480, 640], "counts": "gYh3^1e<l0D=M3M3N10O200O0G:KjNnMhEl1\\:XNaEa1d:dNZET1h:TOVEk0e:^OXE`0j:EPE;P;JlD9>]Nd9^1iEg0LZNR:R1nEV2X9T2E7H6YHdI`6b6XI`Ih6d6SI]Im6f6nH\\IR7i6gHZIX7j6bHXI^7U70001O000000001O0000001O000001N100O1cNbHWK_7k4cHQK^7P5dHkJ]7W5fHcJ\\7]5PIVJP7l5R1QK^G`3b8\\L`Gf3`8VLcGk3^8PLdGR4\\8iKhGV4^8bKdG_4`8ZKcGe4\\910O010005K5K1O10O0O1O100O2QN^F\\Nc9_O]Fc06N]9^OcF879W9]OiFN8c0Q9\\OeHc0]7[OeHd0\\7YOfHf0\\7XOfHf0]7VOgHMXM5S:KaIOa6NfII_61\\4JTl[4"}, "label": "a little boy running on a beach near the ocean"}]}
{"id": 298312, "image": "COCO_train2014_000000298312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a grey swimsuit\"."}], "task": "refering", "answer_template": "The \"a girl in a grey swimsuit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 191, 192, 213, 214, 215, 236, 237, 238, 260, 261, 262, 283, 284, 306, 307, 308, 330, 331], "bbox": [0.27160937500000004, 0.4441875, 0.42551562500000006, 0.8706666666666667], "iscrowd": 0, "area": 7769.044499999998, "rle": {"size": [480, 640], "counts": "cja24j>4M2M201000O100O10000O0100O1000O0100O100O1O1O1N2O10OTO@XCa0Q1B[9:cE4n0Hi8i0TFBQ1If8j0TFAR1Je8Q2XGSNe8P2XGSNf8n1WGUNi8k1UGXNj8i1SGZNl8f1SG\\Nl8e1QG^Nm8d1QG]No8e1mF^NR9d1kF]NU9e1gF^NX9e1dF\\N[9g1aF[N_9k1YFXNf9a301O0010O010O100O010O100O2OhFPK_8P5^GSKc8l4YGXKi8f4UG[Kn8c4oF^KV9_4hFbK\\9\\4aFeKa9k44M4K5QKgFn3]OoKm:d3VE]LR;W3PEjLR;R3PEnLR;k1gDPO8VOS;a1mDWO1XOS;Z1SE\\OKZOT;R1YEBB]OW;i0^EI[O^OY;d0`EOUO]O];`0aEc0a:YOcEe0_:WOdEh0[<O2N1O2N3L5L3M3M2N2N1O2N1N3N1O2N2MSS\\5"}, "label": "a girl in a grey swimsuit"}]}
{"id": 298312, "image": "COCO_train2014_000000298312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl running on the beach\"."}], "task": "refering", "answer_template": "The \"little girl running on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 191, 192, 213, 214, 215, 236, 237, 238, 260, 261, 262, 283, 284, 306, 307, 308, 330, 331], "bbox": [0.27160937500000004, 0.4441875, 0.42551562500000006, 0.8706666666666667], "iscrowd": 0, "area": 7769.044499999998, "rle": {"size": [480, 640], "counts": "cja24j>4M2M201000O100O10000O0100O1000O0100O100O1O1O1N2O10OTO@XCa0Q1B[9:cE4n0Hi8i0TFBQ1If8j0TFAR1Je8Q2XGSNe8P2XGSNf8n1WGUNi8k1UGXNj8i1SGZNl8f1SG\\Nl8e1QG^Nm8d1QG]No8e1mF^NR9d1kF]NU9e1gF^NX9e1dF\\N[9g1aF[N_9k1YFXNf9a301O0010O010O100O010O100O2OhFPK_8P5^GSKc8l4YGXKi8f4UG[Kn8c4oF^KV9_4hFbK\\9\\4aFeKa9k44M4K5QKgFn3]OoKm:d3VE]LR;W3PEjLR;R3PEnLR;k1gDPO8VOS;a1mDWO1XOS;Z1SE\\OKZOT;R1YEBB]OW;i0^EI[O^OY;d0`EOUO]O];`0aEc0a:YOcEe0_:WOdEh0[<O2N1O2N3L5L3M3M2N2N1O2N1N3N1O2N2MSS\\5"}, "label": "little girl running on the beach"}]}
{"id": 183626, "image": "COCO_train2014_000000183626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair with cat on it\"."}], "task": "refering", "answer_template": "The \"chair with cat on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 311, 312, 313, 334, 335, 336], "bbox": [0.1229375, 0.1488758782201405, 0.787125, 0.982576112412178], "iscrowd": 0, "area": 71963.71330000003, "rle": {"size": [427, 640], "counts": "_RQ14U=4L5L4K4L5K5L3N3L4M2M4M3M2M4M3M3L3N3M3L3N3M3L3N3L4M2N3L4M2N3L4M2N3L4M2N3L4M2M4M3M2M4M3M2M4M3M2M4M3L3N3M3L3N3M3L3N3M3L4M2N3L4M2M4M3M2M4M3M2M4M3M2M4M3L10000O10000O10000O10000O2O0O10000O10000O10000O100O10000O10000O10000O10001N100O10000O10000O10000O10000O100O10000O10000O101O0O100O10000O10000O10000O10000O100O10000O10000O101O0O100O10000O10000O10000O10000O100O100001O2N3M3N1N3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M2N3M3N1N3M2N3M3aLUI8m6XOcIf0`6jNnIU1T6XN^Jg1e5dMnJ[2R8O1O1^MaEV2`:gMbEY2_:eMcEZ2^:dMcE\\2g:O1O001O00O1000001O00000O1000000000000O10000001O001O001O001OPGkMk6T2kHVNU7j1eH\\N\\7d1bH]N^7c1aH^N`7a1`H_N`7b1^H_Nc7a1[H`Ne7b1XH_Ni7b1TH_Nm7a1QH`No7b1oG^NR8b1lG_NT8c1iG^NX8c1eG^N[8c1cG^N^8c1`G]N`8e1]G\\Nd8d1ZG]Nc5oNPMf2ZM\\Nc5RORMd2XM[Nc5VOTM_2WM\\Nd5WOTM_2VMZNf5XOSM_2UM[Ng5XORMg3n2ZLQMf3o2\\LoLd3R3\\LmLd3S3]LlLc3T3_LjLa3V3`LiL`3W3aLhL_3X3cLfL]3Z3dLeL\\3\\3dLcL\\3]3fLaLZ3_3gLaLX3_3iL`LW3`3kL^LU3b3mL\\LS3d3oLZLQ3f3QMXLo2i3RMULn2k3SMTLm2l3UMRLk2n3WMPLi2P4YMnKg2R4[MlKf2S4[MlKe2U4\\MhKe2X4\\MgKd2Y4]MfKc2Z4_MdKa2\\4`MbKa2^4aM`K_2`4gMZKY2g4oMoJR2Q5WNfJi1Z5`N]J`1c5hNTJY1l5S300001O000000000000001O01O0000N2M3M4L3N2M3M3N2PO^ITKe6k4o0N1O2N1O1O2N100O2N1O1O2N1O2N100O2N1O1O2N1O2N100O2N1O1O2N1O2N1O1QMTG_1m8`NSG_1n8aNSG]1n8bNSG]1o8bNRG\\1o8dNRGZ1P9eNPGZ1Q9fNPGX1Q9hNoFX1R9gNoFW1R9iNoFU1R9jNoFU1S9jNnFT1S9lNmFS1U9lNlFR1U9nNlFP1U9POkFo0W9POjFo0V9QOjFn0W9ROjFl0X9ROjFl0W9TOiFk0Y9TOhFj0Y9TOiFk0X9TOjFj0X9TOiFl0W9SOkFk0V9SOmFk0a9gN`FX1k:N1O2N2N101N1O2N2N1O2N1O2N101N2N1O2N1O3M3M3N2M4L3M3M3Mi]h1"}, "label": "chair with cat on it"}]}
{"id": 183626, "image": "COCO_train2014_000000183626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red chair that the cat is sitting in\"."}], "task": "refering", "answer_template": "The \"the red chair that the cat is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 311, 312, 313, 334, 335, 336], "bbox": [0.1229375, 0.1488758782201405, 0.787125, 0.982576112412178], "iscrowd": 0, "area": 71963.71330000003, "rle": {"size": [427, 640], "counts": "_RQ14U=4L5L4K4L5K5L3N3L4M2M4M3M2M4M3M3L3N3M3L3N3M3L3N3L4M2N3L4M2N3L4M2N3L4M2N3L4M2M4M3M2M4M3M2M4M3M2M4M3L3N3M3L3N3M3L3N3M3L4M2N3L4M2M4M3M2M4M3M2M4M3M2M4M3L10000O10000O10000O10000O2O0O10000O10000O10000O100O10000O10000O10000O10001N100O10000O10000O10000O10000O100O10000O10000O101O0O100O10000O10000O10000O10000O100O10000O10000O101O0O100O10000O10000O10000O10000O100O100001O2N3M3N1N3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N3M2N3M2N3M3N1N3M2N3M3aLUI8m6XOcIf0`6jNnIU1T6XN^Jg1e5dMnJ[2R8O1O1^MaEV2`:gMbEY2_:eMcEZ2^:dMcE\\2g:O1O001O00O1000001O00000O1000000000000O10000001O001O001O001OPGkMk6T2kHVNU7j1eH\\N\\7d1bH]N^7c1aH^N`7a1`H_N`7b1^H_Nc7a1[H`Ne7b1XH_Ni7b1TH_Nm7a1QH`No7b1oG^NR8b1lG_NT8c1iG^NX8c1eG^N[8c1cG^N^8c1`G]N`8e1]G\\Nd8d1ZG]Nc5oNPMf2ZM\\Nc5RORMd2XM[Nc5VOTM_2WM\\Nd5WOTM_2VMZNf5XOSM_2UM[Ng5XORMg3n2ZLQMf3o2\\LoLd3R3\\LmLd3S3]LlLc3T3_LjLa3V3`LiL`3W3aLhL_3X3cLfL]3Z3dLeL\\3\\3dLcL\\3]3fLaLZ3_3gLaLX3_3iL`LW3`3kL^LU3b3mL\\LS3d3oLZLQ3f3QMXLo2i3RMULn2k3SMTLm2l3UMRLk2n3WMPLi2P4YMnKg2R4[MlKf2S4[MlKe2U4\\MhKe2X4\\MgKd2Y4]MfKc2Z4_MdKa2\\4`MbKa2^4aM`K_2`4gMZKY2g4oMoJR2Q5WNfJi1Z5`N]J`1c5hNTJY1l5S300001O000000000000001O01O0000N2M3M4L3N2M3M3N2PO^ITKe6k4o0N1O2N1O1O2N100O2N1O1O2N1O2N100O2N1O1O2N1O2N100O2N1O1O2N1O2N1O1QMTG_1m8`NSG_1n8aNSG]1n8bNSG]1o8bNRG\\1o8dNRGZ1P9eNPGZ1Q9fNPGX1Q9hNoFX1R9gNoFW1R9iNoFU1R9jNoFU1S9jNnFT1S9lNmFS1U9lNlFR1U9nNlFP1U9POkFo0W9POjFo0V9QOjFn0W9ROjFl0X9ROjFl0W9TOiFk0Y9TOhFj0Y9TOiFk0X9TOjFj0X9TOiFl0W9SOkFk0V9SOmFk0a9gN`FX1k:N1O2N2N101N1O2N2N1O2N1O2N101N2N1O2N1O3M3M3N2M4L3M3M3Mi]h1"}, "label": "the red chair that the cat is sitting in"}]}
{"id": 552272, "image": "COCO_train2014_000000552272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chair in the right side of the image\"."}], "task": "refering", "answer_template": "The \"wooden chair in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 175, 198, 220, 221, 222, 225, 243, 244, 245, 248, 268, 271, 291, 294], "bbox": [0.60334375, 0.40739583333333335, 0.82421875, 0.7718958333333333], "iscrowd": 0, "area": 3756.2508, "rle": {"size": [480, 640], "counts": "aVe56e>d0[O9H8H9\\Oc0fNU12J6B=J7I71O1O001OK5N2O1F:I7H8L4L4L4L4Lga0XOR_O3M4M2O2N1O2N1O2N1SDCV9>]FOc92QF9P:FeEe0[:\\OYEn0i:QOjDZ1X;U13M3K6G8F;_Oe0[Ohmn0TOdSQO;E3L5L2N1N3N1O0O10O01YCUOX;j0`DA^;?UDNj;d11O10O0100O0I8H8H9G8H9F:GaWd1"}, "label": "wooden chair in the right side of the image"}]}
{"id": 552272, "image": "COCO_train2014_000000552272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow chair beneath a boy and a striped shirt\"."}], "task": "refering", "answer_template": "The \"a yellow chair beneath a boy and a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 175, 198, 220, 221, 222, 225, 243, 244, 245, 248, 268, 271, 291, 294], "bbox": [0.60334375, 0.40739583333333335, 0.82421875, 0.7718958333333333], "iscrowd": 0, "area": 3756.2508, "rle": {"size": [480, 640], "counts": "aVe56e>d0[O9H8H9\\Oc0fNU12J6B=J7I71O1O001OK5N2O1F:I7H8L4L4L4L4Lga0XOR_O3M4M2O2N1O2N1O2N1SDCV9>]FOc92QF9P:FeEe0[:\\OYEn0i:QOjDZ1X;U13M3K6G8F;_Oe0[Ohmn0TOdSQO;E3L5L2N1N3N1O0O10O01YCUOX;j0`DA^;?UDNj;d11O10O0100O0I8H8H9G8H9F:GaWd1"}, "label": "a yellow chair beneath a boy and a striped shirt"}]}
{"id": 552272, "image": "COCO_train2014_000000552272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair being used by a boy in orange\"."}], "task": "refering", "answer_template": "The \"a chair being used by a boy in orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 170, 187, 193, 210, 216, 233, 256, 279], "bbox": [0.127515625, 0.4140833333333333, 0.4245, 0.756375], "iscrowd": 0, "area": 3053.941650000001, "rle": {"size": [480, 640], "counts": "cdV1:f>4L4M3L4L4L4YETOS7o0aHZO^7i0VHAi7b0kGHT8;aGN_84UG5k8NjF;V9G^F`0e9[23L4M3M4L3M3L4M3mLmDk1`;oMSEj0h;nNaDm0l<L5J5L4L5J5L5J\\kl10gTSN100O001O1O2O4K5K6J6J5L5J7I7J5L5K2NN2N2N2O1N2N2N2O1HU1kNVX\\5"}, "label": "a chair being used by a boy in orange"}]}
{"id": 191828, "image": "COCO_train2014_000000191828.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing upright on a snowboard\"."}], "task": "refering", "answer_template": "The \"a man standing upright on a snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 102, 103, 125, 126, 127, 148, 149, 150, 171, 172, 173, 194, 195, 196, 217, 218, 219, 241, 242, 264, 265], "bbox": [0.4425625, 0.2463231850117096, 0.5469375000000001, 0.7547306791569086], "iscrowd": 0, "area": 8604.9317, "rle": {"size": [427, 640], "counts": "nSf3<m<3L3N2N3N102M2O1O1N2O6Ko0POe0\\Ob0]Od1]N2M3N2M3N3M001O0010OQIcJP6]5lIeJW6Z5fIgJZ6Z5dIgJ]6[5_IfJb6\\5YIeJh6^5QIfJP7j51O001O1O1O1O0O2O1O1O1O1O1O1O001N2O1O1O1O1O2N5[MiHkN\\7o0mHbN_7X1nHUN^7e1Y2I6K6J5L4K6K4L:Fegh3"}, "label": "a man standing upright on a snowboard"}]}
{"id": 191828, "image": "COCO_train2014_000000191828.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green coat on a snowboard\"."}], "task": "refering", "answer_template": "The \"a man in a green coat on a snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 102, 103, 125, 126, 127, 148, 149, 150, 171, 172, 173, 194, 195, 196, 217, 218, 219, 241, 242, 264, 265], "bbox": [0.4425625, 0.2463231850117096, 0.5469375000000001, 0.7547306791569086], "iscrowd": 0, "area": 8604.9317, "rle": {"size": [427, 640], "counts": "nSf3<m<3L3N2N3N102M2O1O1N2O6Ko0POe0\\Ob0]Od1]N2M3N2M3N3M001O0010OQIcJP6]5lIeJW6Z5fIgJZ6Z5dIgJ]6[5_IfJb6\\5YIeJh6^5QIfJP7j51O001O1O1O1O0O2O1O1O1O1O1O1O001N2O1O1O1O1O2N5[MiHkN\\7o0mHbN_7X1nHUN^7e1Y2I6K6J5L4K6K4L:Fegh3"}, "label": "a man in a green coat on a snowboard"}]}
{"id": 347483, "image": "COCO_train2014_000000347483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small potted plant below a hanging potted plant\"."}], "task": "refering", "answer_template": "The \"a small potted plant below a hanging potted plant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 337, 338, 339], "bbox": [0.67771875, 0.7476168224299066, 0.876703125, 1.0], "iscrowd": 0, "area": 5034.805450000001, "rle": {"size": [428, 640], "counts": "cie52X=2YMOPH4Q82fGO]85ZGNh86PGLR98fFJ\\98]FKe96VFLl96mEMT:5bE2HTO\\:l0aE72nN_:U2`ElMb:R2]EPNd:n1[ESNg:j1YEWNj:f1UE[Nm:c1QE`NQ;]1nDdNU;Y1jDhNY;e11N3M3N1N1OO1000O10O10000lNXNbFj1_9ZNZFh1f9^NRFc1o9bNkE_1V:eNcE]1]:i000001N1000001O00001N1^Ob0M3N3M2M4N5K5K6J5K4L001O000O20O100O1O100O100O1O100O1O^OiC1W<OiC1V<OjC2W<LjC4Z<HfC8]<DdC<`<_O`Cb0e<0000O10O1000000O010000000O1000O1000O100001N10000O1O2O0O1O1O101N1O100O1O2N100O1O1O101NcRQ1"}, "label": "a small potted plant below a hanging potted plant"}]}
{"id": 347483, "image": "COCO_train2014_000000347483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plants on the ground beneath a hanging potted plant\"."}], "task": "refering", "answer_template": "The \"plants on the ground beneath a hanging potted plant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 337, 338, 339], "bbox": [0.67771875, 0.7476168224299066, 0.876703125, 1.0], "iscrowd": 0, "area": 5034.805450000001, "rle": {"size": [428, 640], "counts": "cie52X=2YMOPH4Q82fGO]85ZGNh86PGLR98fFJ\\98]FKe96VFLl96mEMT:5bE2HTO\\:l0aE72nN_:U2`ElMb:R2]EPNd:n1[ESNg:j1YEWNj:f1UE[Nm:c1QE`NQ;]1nDdNU;Y1jDhNY;e11N3M3N1N1OO1000O10O10000lNXNbFj1_9ZNZFh1f9^NRFc1o9bNkE_1V:eNcE]1]:i000001N1000001O00001N1^Ob0M3N3M2M4N5K5K6J5K4L001O000O20O100O1O100O100O1O100O1O^OiC1W<OiC1V<OjC2W<LjC4Z<HfC8]<DdC<`<_O`Cb0e<0000O10O1000000O010000000O1000O1000O100001N10000O1O2O0O1O1O101N1O100O1O2N100O1O1O101NcRQ1"}, "label": "plants on the ground beneath a hanging potted plant"}]}
{"id": 183646, "image": "COCO_train2014_000000183646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a airplane that is in front of another airlplane\"."}], "task": "refering", "answer_template": "The \"a airplane that is in front of another airlplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 244, 245, 246, 247, 261, 262, 267, 268, 269, 270, 271, 291, 292, 293, 314, 315], "bbox": [0.291484375, 0.33039719626168224, 0.8209843750000001, 0.8904205607476636], "iscrowd": 0, "area": 33627.088350000005, "rle": {"size": [428, 640], "counts": "U]^21[=0O2O001N2O001XOJWD7i;JVD7i;KUD6i;LUD5k;LTD5k;LTD4k;NSD4l;MSD4l;MSD3m;NQD4m;ORD0n;1QD0n;1QD0n;1QDOn;3QDNn;3RDLn;5QDLm;6RDKm;j001O0O2O1O0O2O001O1N101O010O1O010O010O00100O100O1SOVNVFk1i9UNXFk1f9VNZFk1e9UN\\FQ2\\9QNcFY2S9gMmFc2h8^MYGd2d8\\M\\Gd2c8^M\\Ga2e8_M\\G`2c8aM]G^2d8bM]G]2b8eM]G[2c8eM]GZ2c8gM^GX2b8hM^GX2a8jM_GT2b8lM^GT2a8mM_GS2a8mM`GQ2`8QN_Go1a8QN`Gm1`8TN`Gl1`8TN`Gk1`8WN`Gg1a8YN_Gg1`8ZNaGd1`8\\N`Gc1`8_N_Ga1a8_N`G_1`8bN`G]1a8cN_G\\1a8fN_Gm0ROdN_9?_Gm0ROcN`9`0_Gk0ROdN`9a0^Gk0ROcNa9c0\\Gj0SOcNa9c0]Gh0SOdNa9d0\\Gh0SOcNb9e0\\Gg0SObNb9h0ZGe0WO`N`9k0YGe0o8[ORGc0o8]OQGc0o8^OQGa0o8_OQG`0P9@PG`0P9@QG>P9CoF=Q9CPG<P9DPG;Q9EoF;Q9FnF9S9GmF9S9GlF:T9FlF9U9HiF9W9GiF9W9G\\FQOJX1k9G[FQOJX1k9HYFQOKW1m9HWFROLV1m9HWFSOKU1n9HVFTOKT1P:ISFTOMS1P:ISFTOLS1R:IQFUOMR1R:IPFVOMR1S:IoEVOMP1U:JmEWOMP1V:ImEWOMo0W:JkEXONn0W:KiEXOOn0X:JiEXOOm0Y:KgEZONl0[:JfE[OOk0[:KeEd0\\:V10O10000O100O10000O100O101O0O10000O100O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O100000000O1000001O000000000000O3M4L5K4gFfMY7^2cHbM]7b2_H_M`7e2]H[Mb7j2YHWMg7l2UHUMj7o2RHRMm7R3PHmLP8W3lGjLS8[3hGfLW8^3eGcLZ8a3cG_L\\8e3`G[La8X4O1O1O1O001O5K5K6K4K6J5K5K6J5K5K6J3N1N1O2N2N2N2N2N2O1N101OO100O1O100O010O1O100O1O100O00100O100O1O100O00100O100O1O100TK`Ik2a6UMeId2\\6[MlI^2T6aMSJX2n5gMYJS2g5lM`Jm1b5RNdJg1]5XNjJb1V5]NQK\\1P5cNWKV1j4jNYKS1g4lNYKT1h4kNYKQOVNT1b6JXKRObNi0U66XKROmN=l5`0WKSOYO2`5j0XKSOYO2`5j0WKUOYO0a5k0UKUOZO0b5j0UKUOZO0b5j0TKWOYOOd5i0SKXOZONd5j0QKYOZOJdNAQ7\\1RKXOZOJdNAQ7\\1QKYOZOJi5l0mJ[OZOIi5k0nJ[OYOJj5k0mJZOZOJj5k0mJZOYOLj5i0nJZOYOLj5i0nJZOXOMk5h0nJZOXONi5h0oJZOXONj5g0oJZOXONj5g0oJZOWOOk5f0oJZOWO0j5f0oJZOVO0l5e0oJZOVO0l5e0PKYOTO3l5c0QKYOSO4l5c0SKVOSO6k5c0SKSOUO:i5b0nL^OQ3b0oL^OT3?mL@U3=mLCU37oLHS33QMLU3KoL5T8O1O1O100N[h_1"}, "label": "a airplane that is in front of another airlplane"}]}
{"id": 183646, "image": "COCO_train2014_000000183646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toy airplane is on the grass\"."}], "task": "refering", "answer_template": "The \"the toy airplane is on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 244, 245, 246, 247, 261, 262, 267, 268, 269, 270, 271, 291, 292, 293, 314, 315], "bbox": [0.291484375, 0.33039719626168224, 0.8209843750000001, 0.8904205607476636], "iscrowd": 0, "area": 33627.088350000005, "rle": {"size": [428, 640], "counts": "U]^21[=0O2O001N2O001XOJWD7i;JVD7i;KUD6i;LUD5k;LTD5k;LTD4k;NSD4l;MSD4l;MSD3m;NQD4m;ORD0n;1QD0n;1QD0n;1QDOn;3QDNn;3RDLn;5QDLm;6RDKm;j001O0O2O1O0O2O001O1N101O010O1O010O010O00100O100O1SOVNVFk1i9UNXFk1f9VNZFk1e9UN\\FQ2\\9QNcFY2S9gMmFc2h8^MYGd2d8\\M\\Gd2c8^M\\Ga2e8_M\\G`2c8aM]G^2d8bM]G]2b8eM]G[2c8eM]GZ2c8gM^GX2b8hM^GX2a8jM_GT2b8lM^GT2a8mM_GS2a8mM`GQ2`8QN_Go1a8QN`Gm1`8TN`Gl1`8TN`Gk1`8WN`Gg1a8YN_Gg1`8ZNaGd1`8\\N`Gc1`8_N_Ga1a8_N`G_1`8bN`G]1a8cN_G\\1a8fN_Gm0ROdN_9?_Gm0ROcN`9`0_Gk0ROdN`9a0^Gk0ROcNa9c0\\Gj0SOcNa9c0]Gh0SOdNa9d0\\Gh0SOcNb9e0\\Gg0SObNb9h0ZGe0WO`N`9k0YGe0o8[ORGc0o8]OQGc0o8^OQGa0o8_OQG`0P9@PG`0P9@QG>P9CoF=Q9CPG<P9DPG;Q9EoF;Q9FnF9S9GmF9S9GlF:T9FlF9U9HiF9W9GiF9W9G\\FQOJX1k9G[FQOJX1k9HYFQOKW1m9HWFROLV1m9HWFSOKU1n9HVFTOKT1P:ISFTOMS1P:ISFTOLS1R:IQFUOMR1R:IPFVOMR1S:IoEVOMP1U:JmEWOMP1V:ImEWOMo0W:JkEXONn0W:KiEXOOn0X:JiEXOOm0Y:KgEZONl0[:JfE[OOk0[:KeEd0\\:V10O10000O100O10000O100O101O0O10000O100O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O100000000O1000001O000000000000O3M4L5K4gFfMY7^2cHbM]7b2_H_M`7e2]H[Mb7j2YHWMg7l2UHUMj7o2RHRMm7R3PHmLP8W3lGjLS8[3hGfLW8^3eGcLZ8a3cG_L\\8e3`G[La8X4O1O1O1O001O5K5K6K4K6J5K5K6J5K5K6J3N1N1O2N2N2N2N2N2O1N101OO100O1O100O010O1O100O1O100O00100O100O1O100O00100O100O1O100TK`Ik2a6UMeId2\\6[MlI^2T6aMSJX2n5gMYJS2g5lM`Jm1b5RNdJg1]5XNjJb1V5]NQK\\1P5cNWKV1j4jNYKS1g4lNYKT1h4kNYKQOVNT1b6JXKRObNi0U66XKROmN=l5`0WKSOYO2`5j0XKSOYO2`5j0WKUOYO0a5k0UKUOZO0b5j0UKUOZO0b5j0TKWOYOOd5i0SKXOZONd5j0QKYOZOJdNAQ7\\1RKXOZOJdNAQ7\\1QKYOZOJi5l0mJ[OZOIi5k0nJ[OYOJj5k0mJZOZOJj5k0mJZOYOLj5i0nJZOYOLj5i0nJZOXOMk5h0nJZOXONi5h0oJZOXONj5g0oJZOXONj5g0oJZOWOOk5f0oJZOWO0j5f0oJZOVO0l5e0oJZOVO0l5e0PKYOTO3l5c0QKYOSO4l5c0SKVOSO6k5c0SKSOUO:i5b0nL^OQ3b0oL^OT3?mL@U3=mLCU37oLHS33QMLU3KoL5T8O1O1O100N[h_1"}, "label": "the toy airplane is on the grass"}]}
{"id": 183646, "image": "COCO_train2014_000000183646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toy airplane is blue and white and on the other side of the front plane\"."}], "task": "refering", "answer_template": "The \"a toy airplane is blue and white and on the other side of the front plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 141, 142, 143, 144, 145, 163, 164, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 193, 194, 209, 210, 211, 212, 213, 214, 232, 233, 236, 237, 238, 259, 260, 261, 282, 283, 284, 305, 306], "bbox": [0.091421875, 0.3827102803738318, 0.47462499999999996, 0.8788551401869158], "iscrowd": 0, "area": 16574.603800000004, "rle": {"size": [428, 640], "counts": "nkh06V=7I6J5K1O1O2N1O1O2M3N3M2N2N2Ng0YO:F4L1O0010O0010O0001O0O2O0O101N101O0O1O2K4L5O0O101O0O2O0O2O[NYNeGg1[8ZNeGd1[8^NdGb1\\8_NdG`1[8bNdG]1]8dNbG\\1]8fNcGX1^8iNaGW1_8jNaGU1^8mNaGR1`8oN`GP1_8SO_Gm0a8TO_Gj0a8XO^Gh0b8YO^Ge0b8]O]Gc0c8^O\\Gb0d8_O\\G?d8C[G=e8C\\G<c8F\\G9e8G\\G8c8J\\G6d8J]G4c8N\\G2d8N]G0d81[GOe81\\GNe82ZGMg84XGLi83WGMi84VGKl84TGLl85SGJo85QGKo86PGJQ95oFJR97lFJT96lFJU96jFIW98hFHY97gFIY98fFH[97eFI[98cFH_97aFI_98`FHa97_FIa98^FHb99]FGd99ZFHf9:XFFi9:VFEk9<TFDm9<RFDn9=QFCP:=oECQ:>mECT:=kECU:>jEBW:?gE@Z:a0eE_O[:b0dE^O]:b0aE_O_:b0`E^Oa:b0^E^Ob:c0]E]Od:c0[E]Oe:d0ZE[Oh:d1001O00001O001N101O00001O001O0FPEXN]2h0T5o0aHXNl1X1d5?aHXN\\1h1S6ObHYNk0X2c6_OcHXN;h2S7nNdHYNJX3b7_NeHg1\\7XNeHf1\\7YNfHf1Z7lM\\HPO:T3[7jM]HQO9T3Z7jM_HQO8T3Z7iM`HRO7T3Y7iMbHRO5U3Y7hMfHPO2W3Y7gM\\IX2d6gM]IY2c6fM_IX2c6fM_IY2a6fM`IZ2a6dMaI[2_6dMcI[2]6eMcI[2^6cMdI\\2\\6dMeI[2\\6cMeI]2[6cMfI\\2Z6dMgIZ2[6eMfIZ2Z6fMfIZ2Z6fMgIY2Z6fMgIY2Y6gMgIY2Z6fMgIY2Y6gMhIX2X6hMiIW2X6hMiIW2W6iMiIV2Y6iMhIV2X6jMiIT2X6lMiIR2Y6mMhIQ2Y6oMhIn1Z6RNgIl1[6SNfIk1[6TNfIk1\\6TNeIj1\\6VNmI_1U6aNRJW1o5iNVJQ1k5oNWJn0j5SOWJj0i5XOXJd0j5\\OXJ?j5CWJ6o5JSJNR62PJHU68lIAY6`0Y3010O10O01O10O0100O00100O001O0010O01O001O00100O001O001O010O001O0010O01O001O0010O0O10000N3N1E;O1O1O1O2N1O1O1O1Nbd\\4"}, "label": "a toy airplane is blue and white and on the other side of the front plane"}]}
{"id": 183646, "image": "COCO_train2014_000000183646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two toy airplanes sitting on brown grass\"."}], "task": "refering", "answer_template": "The \"two toy airplanes sitting on brown grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 141, 142, 143, 144, 145, 163, 164, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 193, 194, 209, 210, 211, 212, 213, 214, 232, 233, 236, 237, 238, 259, 260, 261, 282, 283, 284, 305, 306], "bbox": [0.091421875, 0.3827102803738318, 0.47462499999999996, 0.8788551401869158], "iscrowd": 0, "area": 16574.603800000004, "rle": {"size": [428, 640], "counts": "nkh06V=7I6J5K1O1O2N1O1O2M3N3M2N2N2Ng0YO:F4L1O0010O0010O0001O0O2O0O101N101O0O1O2K4L5O0O101O0O2O0O2O[NYNeGg1[8ZNeGd1[8^NdGb1\\8_NdG`1[8bNdG]1]8dNbG\\1]8fNcGX1^8iNaGW1_8jNaGU1^8mNaGR1`8oN`GP1_8SO_Gm0a8TO_Gj0a8XO^Gh0b8YO^Ge0b8]O]Gc0c8^O\\Gb0d8_O\\G?d8C[G=e8C\\G<c8F\\G9e8G\\G8c8J\\G6d8J]G4c8N\\G2d8N]G0d81[GOe81\\GNe82ZGMg84XGLi83WGMi84VGKl84TGLl85SGJo85QGKo86PGJQ95oFJR97lFJT96lFJU96jFIW98hFHY97gFIY98fFH[97eFI[98cFH_97aFI_98`FHa97_FIa98^FHb99]FGd99ZFHf9:XFFi9:VFEk9<TFDm9<RFDn9=QFCP:=oECQ:>mECT:=kECU:>jEBW:?gE@Z:a0eE_O[:b0dE^O]:b0aE_O_:b0`E^Oa:b0^E^Ob:c0]E]Od:c0[E]Oe:d0ZE[Oh:d1001O00001O001N101O00001O001O0FPEXN]2h0T5o0aHXNl1X1d5?aHXN\\1h1S6ObHYNk0X2c6_OcHXN;h2S7nNdHYNJX3b7_NeHg1\\7XNeHf1\\7YNfHf1Z7lM\\HPO:T3[7jM]HQO9T3Z7jM_HQO8T3Z7iM`HRO7T3Y7iMbHRO5U3Y7hMfHPO2W3Y7gM\\IX2d6gM]IY2c6fM_IX2c6fM_IY2a6fM`IZ2a6dMaI[2_6dMcI[2]6eMcI[2^6cMdI\\2\\6dMeI[2\\6cMeI]2[6cMfI\\2Z6dMgIZ2[6eMfIZ2Z6fMfIZ2Z6fMgIY2Z6fMgIY2Y6gMgIY2Z6fMgIY2Y6gMhIX2X6hMiIW2X6hMiIW2W6iMiIV2Y6iMhIV2X6jMiIT2X6lMiIR2Y6mMhIQ2Y6oMhIn1Z6RNgIl1[6SNfIk1[6TNfIk1\\6TNeIj1\\6VNmI_1U6aNRJW1o5iNVJQ1k5oNWJn0j5SOWJj0i5XOXJd0j5\\OXJ?j5CWJ6o5JSJNR62PJHU68lIAY6`0Y3010O10O01O10O0100O00100O001O0010O01O001O00100O001O001O010O001O0010O01O001O0010O0O10000N3N1E;O1O1O1O2N1O1O1O1Nbd\\4"}, "label": "two toy airplanes sitting on brown grass"}]}
{"id": 11618, "image": "COCO_train2014_000000011618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"delta aeroplain running the air port road\"."}], "task": "refering", "answer_template": "The \"delta aeroplain running the air port road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 98, 99, 100, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 115, 116, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 171, 172, 173, 174, 175, 176, 177], "bbox": [0.03565625, 0.12783529411764705, 0.9240468750000002, 0.5081411764705882], "iscrowd": 0, "area": 31007.717, "rle": {"size": [425, 640], "counts": "_f95R=30000000000O10000000000O1000O1O1O1O001O1O1O00000000000000000O1000O100000000000000000000000000000000000000000000O1000000000000000001O0000000000000000001O000UOKbD5^;KbD5^;KaD6_;JaD6_;JaD6_;JaD6`;J_D6a;J_D6a;J_D7`;I`D7`;I`D7`;I`D7`;I`D7`;J_D6a;J_D6a;J_D6a;J_D6a;J_D6a;J_D7a;I^D7b;I^D7b;I^D7b;I^D7b;I^D7b;I^D7b;J\\D7d;I\\D7d;I\\D7d;I\\D8c;H]D8c;H]D8c;H]D8d;H[D8e;H[D8e;H[D8e;H[D8e;H[D8e;H[D8e;IZD7f;IZD8e;H[D8e;H[D8e;H[D8e;H[D8e;IZD7g;HYD8g;HYD8g;HYD8g;HXD9h;GXD:g;FYD:g;GXD9h;GXD9h;GXD9h;GXD9h;GXD9h;GXD9h;HWD8j;GVD9j;GVD9j;GVD:i;FWD;h;EXD<g;EXD<g;DYD>e;B[D?d;A\\D`0c;@]Db0a;^O_Dc0f8XOfI5dMd0d8\\OdI1gMc0e8@`IMjMe0e8B^IHmMf0d8G[ICQNf0d8KWI_OUNg0b8OUIZOYNg0b83QIWO\\Nf0b88nHRO`Ng0a8;kHnNdNg0a8?gHjNhNg0`8d0dHeNlNh0_8g0^HdNSOe0^8l0YHaNYOc0^8P1TH_N]Ob0]8T1QH[NBa0]8X1lGYNF?^8\\1gGVNK?]8_1cGTNO=_8b1]GSN3;b8Z2]GgMd8W2\\GiMe8W2ZGiMh8V2WGkMj8T2UGlMl8S2SGnMo8U31O2N2N1O2N1O2N2N1O2N2N1O2N1M4L4K4I8K5O0O2N1O2N2N1O2N101N2N1O2N2N1O2N1O2O1N1O2N2N1O2N1O2O1N1O2N00000000O2O0000001O0000001O000000001O00000O2O0000001O000000001O00001O00001O0O10001O00001O00001O00001O0000001O0O101O00001O00001O0000001O0000001N100000001O00000000001O000000001O0O100000001O0000O100O100O100O10000O010O10000O100O10000O100O1000000000O100000000000000000O100000000000000000O1000000000000O01000000000O1000000000000O100000O100000O10000000000O1000000000000O01000000000000000000000000000O10O10000000000000000000000000O100000O100000000000000000000000O100000O1000000000000000000000O1000000000O10000000000000000000O01000000O100000000O100000000O010000000O100000000O100000000O10O100000O1000000O1000000O2O1O001O1N101O1O0O2O1O001O0O2O1O001N2O001O1N101O1O001N2O001O1N2O1O1N3M2N2N2N2N2N2M3N2N3M2JoUd0"}, "label": "delta aeroplain running the air port road"}]}
{"id": 11618, "image": "COCO_train2014_000000011618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white delta airplane\"."}], "task": "refering", "answer_template": "The \"a white delta airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 75, 76, 98, 99, 100, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 115, 116, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 171, 172, 173, 174, 175, 176, 177], "bbox": [0.03565625, 0.12783529411764705, 0.9240468750000002, 0.5081411764705882], "iscrowd": 0, "area": 31007.717, "rle": {"size": [425, 640], "counts": "_f95R=30000000000O10000000000O1000O1O1O1O001O1O1O00000000000000000O1000O100000000000000000000000000000000000000000000O1000000000000000001O0000000000000000001O000UOKbD5^;KbD5^;KaD6_;JaD6_;JaD6_;JaD6`;J_D6a;J_D6a;J_D7`;I`D7`;I`D7`;I`D7`;I`D7`;J_D6a;J_D6a;J_D6a;J_D6a;J_D6a;J_D7a;I^D7b;I^D7b;I^D7b;I^D7b;I^D7b;I^D7b;J\\D7d;I\\D7d;I\\D7d;I\\D8c;H]D8c;H]D8c;H]D8d;H[D8e;H[D8e;H[D8e;H[D8e;H[D8e;H[D8e;IZD7f;IZD8e;H[D8e;H[D8e;H[D8e;H[D8e;IZD7g;HYD8g;HYD8g;HYD8g;HXD9h;GXD:g;FYD:g;GXD9h;GXD9h;GXD9h;GXD9h;GXD9h;GXD9h;HWD8j;GVD9j;GVD9j;GVD:i;FWD;h;EXD<g;EXD<g;DYD>e;B[D?d;A\\D`0c;@]Db0a;^O_Dc0f8XOfI5dMd0d8\\OdI1gMc0e8@`IMjMe0e8B^IHmMf0d8G[ICQNf0d8KWI_OUNg0b8OUIZOYNg0b83QIWO\\Nf0b88nHRO`Ng0a8;kHnNdNg0a8?gHjNhNg0`8d0dHeNlNh0_8g0^HdNSOe0^8l0YHaNYOc0^8P1TH_N]Ob0]8T1QH[NBa0]8X1lGYNF?^8\\1gGVNK?]8_1cGTNO=_8b1]GSN3;b8Z2]GgMd8W2\\GiMe8W2ZGiMh8V2WGkMj8T2UGlMl8S2SGnMo8U31O2N2N1O2N1O2N2N1O2N2N1O2N1M4L4K4I8K5O0O2N1O2N2N1O2N101N2N1O2N2N1O2N1O2O1N1O2N2N1O2N1O2O1N1O2N00000000O2O0000001O0000001O000000001O00000O2O0000001O000000001O00001O00001O0O10001O00001O00001O00001O0000001O0O101O00001O00001O0000001O0000001N100000001O00000000001O000000001O0O100000001O0000O100O100O100O10000O010O10000O100O10000O100O1000000000O100000000000000000O100000000000000000O1000000000000O01000000000O1000000000000O100000O100000O10000000000O1000000000000O01000000000000000000000000000O10O10000000000000000000000000O100000O100000000000000000000000O100000O1000000000000000000000O1000000000O10000000000000000000O01000000O100000000O100000000O010000000O100000000O100000000O10O100000O1000000O1000000O2O1O001O1N101O1O0O2O1O001O0O2O1O001N2O001O1N101O1O001N2O001O1N2O1O1N3M2N2N2N2N2N2M3N2N3M2JoUd0"}, "label": "a white delta airplane"}]}
{"id": 552291, "image": "COCO_train2014_000000552291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table on the right\"."}], "task": "refering", "answer_template": "The \"the table on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 219, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 337, 338], "bbox": [0.54971875, 0.5636279069767443, 0.8830312499999999, 0.9870232558139536], "iscrowd": 0, "area": 17790.860699999994, "rle": {"size": [430, 640], "counts": "WRd42Z=2N3M2N3M2N2N3O010O002N5L4K5K5L3L1O010O0001O0O1O2NL5[EeNR9_1mFcNQ9^1oFcNQ9\\1oFeNQ9[1nFfNR9Y1nFhNR9W1nFjNR9U1nFlNR9T1mFmNS9R1mFoNS9P1mFQOS9n0mFSOS9m0lFTOT9k0kFWOU9i0iFYOW9h0cF]O]9c0_FA`9a0^F@b9a0\\FAd9?ZFBf9>YFDf9=XFDi9<UFFj9;TFFl9:SFHm98QFJn97PFJQ:6mELR:4mEMS:4kENU:2hE0X:d1O010O0010O01O10O016I4MO001O001QGRM\\7n2`H^MX7c2cHhMV7X2fHmMY7T2bHoM_7Q2]HSNc7n1YHTNh7m1SHWNm7i1oGZNR8g1iG]NW8c1eG`N\\8Y30O1000O0100O100O100O27H10O00010O0010O01O0UOl0F901O0O2O1O1O1N2O1O1N2O1O1N10O1000O10O10O1000O0100000O010000O10000O10001O0O10000O101O000O10001N10000O10001O0O10000O2O000O1000001N10000O10001N1000000Ok0VO8H8G5L1O1N101O1O0O2O001N2O001N2O001O0O2O1O0O2O001N2O001O1N2O3M2MnTo0"}, "label": "the table on the right"}]}
{"id": 552291, "image": "COCO_train2014_000000552291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the end of a table , with a pink tablecloth at which eight people are sitting\"."}], "task": "refering", "answer_template": "The \"the end of a table , with a pink tablecloth at which eight people are sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 219, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 337, 338], "bbox": [0.54971875, 0.5636279069767443, 0.8830312499999999, 0.9870232558139536], "iscrowd": 0, "area": 17790.860699999994, "rle": {"size": [430, 640], "counts": "WRd42Z=2N3M2N3M2N2N3O010O002N5L4K5K5L3L1O010O0001O0O1O2NL5[EeNR9_1mFcNQ9^1oFcNQ9\\1oFeNQ9[1nFfNR9Y1nFhNR9W1nFjNR9U1nFlNR9T1mFmNS9R1mFoNS9P1mFQOS9n0mFSOS9m0lFTOT9k0kFWOU9i0iFYOW9h0cF]O]9c0_FA`9a0^F@b9a0\\FAd9?ZFBf9>YFDf9=XFDi9<UFFj9;TFFl9:SFHm98QFJn97PFJQ:6mELR:4mEMS:4kENU:2hE0X:d1O010O0010O01O10O016I4MO001O001QGRM\\7n2`H^MX7c2cHhMV7X2fHmMY7T2bHoM_7Q2]HSNc7n1YHTNh7m1SHWNm7i1oGZNR8g1iG]NW8c1eG`N\\8Y30O1000O0100O100O100O27H10O00010O0010O01O0UOl0F901O0O2O1O1O1N2O1O1N2O1O1N10O1000O10O10O1000O0100000O010000O10000O10001O0O10000O101O000O10001N10000O10001O0O10000O2O000O1000001N10000O10001N1000000Ok0VO8H8G5L1O1N101O1O0O2O001N2O001N2O001O0O2O1O0O2O001N2O001O1N2O3M2MnTo0"}, "label": "the end of a table , with a pink tablecloth at which eight people are sitting"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"down side of the wooden table\"."}], "task": "refering", "answer_template": "The \"down side of the wooden table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 216, 217, 218, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 283, 285, 288, 289], "bbox": [0.24553124999999998, 0.48738161559888576, 0.6113125, 1.0], "iscrowd": 0, "area": 11745.319049999998, "rle": {"size": [359, 640], "counts": "_Wg12S;8G7J2M4M3O1N2OUOPF`0\\:00000O01000O100O10000O01O1O010O00010O00010O000010O00010O01O00010O00010O00010O000010bGAR6?mIHn57RJOh51XJ5b5K]J=]5BcJc0X5]OhJc0X5]OgJd0Y5\\OgJd0Y5\\OfJd0[5\\OeJd0[5\\OeJd0[5\\OdJ:mMA_75dJ9UNZOX7=cJ9h5GWJ9j5GVJ9j5GUJ9l5GTJ8m5HSJ8m5HRJ8o5HQJ7P6IPJ7P6IoI7R6InI7R6ImI8S6HmI7T6IlI7T6IkI8U6HkI7V6IjI7V6IiI8W6HiI7X6IgI8Y6HgI8Y6HgI7Z6IeI8[6HeI8[6HeI8[6HdI8]6HcI8]6HbI9^6GbI8_6HaI8_6NZI3e65TIKl6=lHCT7d0dH]O\\7j0]HVOc7Q1UHPOk7W1nGiNW1YO\\5m2_JXMa5h2[J\\Me5d2VJaMj5^2SJfMm5Z2nIkMR6U2jIoMV6Q2eITN[6l1aIXN_6h1\\I]Nd6S300000000000O1L4K5L4K5K5L4K5L4K5K500001O00001N010000000O01000000000O1000000O1000]MXHf0HHo7BeI<[6DgI:X6GiI8W6GjI:U6FlI9T6GlI9T6GmI9Q6HoI8Q6HoI8Q6GQJ9n5GRJ9n5GRJ9m5HTJ8k5HUJ8k5HUJ8k5HVJ8i5GXJ9g5HYJ9f5G[J8e5H[J8e5H[J9d5G]J8b5H_J9`5G`J9`5GaJ8_5HaJ9]5HcJ8]5GeJ8[5HeJ9Z5GfJ9Z5GgJ7Y5IhJ0_50aJJe56\\JBk5>_2O01000O1000O1iH_Oh3`0mKMR43bK:]4FbK<]4CdK?Y4BhKd0Q4\\OoKd0Q4[OQLd0n3]ORLc0n3\\OSLd0l3]OULb0j3^OXLa0h3_OYL`0f3@\\L?d3A\\L?c3A_L>a3B`L=_3CcL<]3DdL;[3EfL;Z3EgL:X3FjL9V3GlL7S3IoL6Q3JPM5o2KUM1l2OU4010O01O00001O001O00FEjE:W:JdE6_:76H8IfSg2"}, "label": "down side of the wooden table"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty plastic chair\"."}], "task": "refering", "answer_template": "The \"an empty plastic chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 233, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.002515625, 0.6966295264623955, 0.335296875, 0.9865181058495821], "iscrowd": 0, "area": 14028.129700000005, "rle": {"size": [359, 640], "counts": "Xn0R1V:R1mNi0WO8IO00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O001O010O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O0001O0O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O010O100O14L5K5K6J5K6J5J7J5K5K6J5K6J5KPed4"}, "label": "an empty plastic chair"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty plastic chair on a patio with two children sitting in other chairs\"."}], "task": "refering", "answer_template": "The \"an empty plastic chair on a patio with two children sitting in other chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 233, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.002515625, 0.6966295264623955, 0.335296875, 0.9865181058495821], "iscrowd": 0, "area": 14028.129700000005, "rle": {"size": [359, 640], "counts": "Xn0R1V:R1mNi0WO8IO00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O001O010O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O0001O0O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O010O100O14L5K5K6J5K6J5J7J5K5K6J5K6J5KPed4"}, "label": "an empty plastic chair on a patio with two children sitting in other chairs"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby sitting the chair and cutting the food\"."}], "task": "refering", "answer_template": "The \"a baby sitting the chair and cutting the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 143, 144, 166, 167, 189, 190, 191, 192, 212, 213, 214, 235, 236], "bbox": [0.21214062500000003, 0.2330640668523677, 0.4243125, 0.8263788300835655], "iscrowd": 0, "area": 12390.925900000004, "rle": {"size": [359, 640], "counts": "Pi_1b0c:o0QO=D3M4M2M4L3N0O2O0O2O0O2O001N101N101N100O2gJVMS2j2lMXMR2j2TKYMX2Od2Y3XMjLf2Y3VMjLi2W3WKiNh4o0QKnL4T2j4m0UKPMNU2l4i0XKDg4;\\KDc4:aKE^49eKGZ48gKIX45kKJU45mKKR43PLNo30TL0k3OWL1h3MZL3f3K]L4b3M_L2a3MaL1`3NaL2^3OdLN]31fLL[34fLKZ34kLGV39mLCT3<oLBR3<oLCR3=oLAR3>oLBQ3>oLAR3>PM@R3?nL_OT3a0mL\\OU3d0kL[OV3e0jLYOX3g0iLVOZ3i0fLUO\\3k0dLRO_3n0bLeNj3[1VLaNo3^1RL^NQ4b1oK[NT4e1lKYNW4f1jKWNX4i1hKVNY4j1hKSN[4l1eKTN[4l1eKSN\\4m1dKSN\\4m1eKRN\\4m1dKSN\\4m1dKSN\\4m1dKSN]4l1dKSN\\4m1dKSN\\4l1eKTN\\4k1dKTN]4l1dKSN\\4m1dKSN]4l1cKTN]4l1cKTN]4l1dKSN]4l1cKTN]4k1dKUN]4j1cKVN]4j1dKTN^4k1bKUN_4j1bKUN^4k1cKTN^4k1cKTN]4l1dKSN]4k1eKTN[4l1a2O1O1O001O1O1O1N2O001O2N1N2O1O1O1N2O1O1N2O1O2M9H1O1N2O1O1O1N2O1O1N2O1O1N3N2NmkP4"}, "label": "a baby sitting the chair and cutting the food"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy playing with his food at an outside table\"."}], "task": "refering", "answer_template": "The \"a young boy playing with his food at an outside table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 143, 144, 166, 167, 189, 190, 191, 192, 212, 213, 214, 235, 236], "bbox": [0.21214062500000003, 0.2330640668523677, 0.4243125, 0.8263788300835655], "iscrowd": 0, "area": 12390.925900000004, "rle": {"size": [359, 640], "counts": "Pi_1b0c:o0QO=D3M4M2M4L3N0O2O0O2O0O2O001N101N101N100O2gJVMS2j2lMXMR2j2TKYMX2Od2Y3XMjLf2Y3VMjLi2W3WKiNh4o0QKnL4T2j4m0UKPMNU2l4i0XKDg4;\\KDc4:aKE^49eKGZ48gKIX45kKJU45mKKR43PLNo30TL0k3OWL1h3MZL3f3K]L4b3M_L2a3MaL1`3NaL2^3OdLN]31fLL[34fLKZ34kLGV39mLCT3<oLBR3<oLCR3=oLAR3>oLBQ3>oLAR3>PM@R3?nL_OT3a0mL\\OU3d0kL[OV3e0jLYOX3g0iLVOZ3i0fLUO\\3k0dLRO_3n0bLeNj3[1VLaNo3^1RL^NQ4b1oK[NT4e1lKYNW4f1jKWNX4i1hKVNY4j1hKSN[4l1eKTN[4l1eKSN\\4m1dKSN\\4m1eKRN\\4m1dKSN\\4m1dKSN\\4m1dKSN]4l1dKSN\\4m1dKSN\\4l1eKTN\\4k1dKTN]4l1dKSN\\4m1dKSN]4l1cKTN]4l1cKTN]4l1dKSN]4l1cKTN]4k1dKUN]4j1cKVN]4j1dKTN^4k1bKUN_4j1bKUN^4k1cKTN^4k1cKTN]4l1dKSN]4k1eKTN[4l1a2O1O1O001O1O1O1N2O001O2N1N2O1O1O1N2O1O1N2O1O2M9H1O1N2O1O1O1N2O1O1N2O1O1N3N2NmkP4"}, "label": "a young boy playing with his food at an outside table"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the bigger girl is in\"."}], "task": "refering", "answer_template": "The \"the chair the bigger girl is in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 177, 178, 179, 196, 197, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.522375, 0.45158774373259053, 0.855828125, 1.0], "iscrowd": 0, "area": 20265.945850000007, "rle": {"size": [359, 640], "counts": "X^e31j:<I7J6J6I7ZOf0M3L4L3G:SOm0O100OM4A>D=I6J7J52O1N101N2O1O001O1O00001O00001O01O01O01O01UHjMc6V2lH\\NT7d1`HgN`7Y1]HjNc7]21O01O0001O00001O01O01O000000000000001O00000000000000000000000000000000000000000000O100O1O100O100O1O100O1O1O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1OiNXIQNh6P2V12N2N3M3M3M3M3M3M2O1N3L3N3M2N3M2M3N3M2N3M2M4M2N2N3N1N100O1O100O10^M`J]O_59mJFR58RKGn47UKHj46ZKIe46^KIb45aKJ^44fKKZ43iKLV43mKLR42RLMn31ULNj31YLDQ49TL[OU4d0nK\\OP4b0TL]Ol3a0WL_Og3`0\\L@b3>bLB]3<fLCY3<jLDT3:PMFn28VMHi25[MJd23aMM]21gMOW2OmM1R2LSN2l1LXN4f1J^N6`1IcN7[1GiN8V1FnN:P1ESO;l0CWO=g0C[O<e0C]O=b0B@>>BD>;AG>9@J?5AN=2B0=OC3<NB4=O_O3?\\5O1O1O1N2O1O1O1O1N2O2N1O1N[XP1"}, "label": "the chair the bigger girl is in"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair with red headed girl sitting in it\"."}], "task": "refering", "answer_template": "The \"chair with red headed girl sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 177, 178, 179, 196, 197, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.522375, 0.45158774373259053, 0.855828125, 1.0], "iscrowd": 0, "area": 20265.945850000007, "rle": {"size": [359, 640], "counts": "X^e31j:<I7J6J6I7ZOf0M3L4L3G:SOm0O100OM4A>D=I6J7J52O1N101N2O1O001O1O00001O00001O01O01O01O01UHjMc6V2lH\\NT7d1`HgN`7Y1]HjNc7]21O01O0001O00001O01O01O000000000000001O00000000000000000000000000000000000000000000O100O1O100O100O1O100O1O1O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1OiNXIQNh6P2V12N2N3M3M3M3M3M3M2O1N3L3N3M2N3M2M3N3M2N3M2M4M2N2N3N1N100O1O100O10^M`J]O_59mJFR58RKGn47UKHj46ZKIe46^KIb45aKJ^44fKKZ43iKLV43mKLR42RLMn31ULNj31YLDQ49TL[OU4d0nK\\OP4b0TL]Ol3a0WL_Og3`0\\L@b3>bLB]3<fLCY3<jLDT3:PMFn28VMHi25[MJd23aMM]21gMOW2OmM1R2LSN2l1LXN4f1J^N6`1IcN7[1GiN8V1FnN:P1ESO;l0CWO=g0C[O<e0C]O=b0B@>>BD>;AG>9@J?5AN=2B0=OC3<NB4=O_O3?\\5O1O1O1N2O1O1O1O1N2O2N1O1N[XP1"}, "label": "chair with red headed girl sitting in it"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with two plates on it\"."}], "task": "refering", "answer_template": "The \"a table with two plates on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173], "bbox": [0.2585625, 0.4244846796657381, 0.604765625, 0.6041782729805013], "iscrowd": 0, "area": 8131.171500000004, "rle": {"size": [359, 640], "counts": "UQj11h:?N2O0O2O1N101N101N2O0O2O1N101O0O2O0O101N100O2O0O2N1O110O00010O000010O01O01O01O00010O001O00010O00000000000001O0000000000001O0000000000001O000000O10000O1000000O1000000O10O1000O1000000O100FcNjF]1V9fNgFZ1Y9iNcFX1]99O100000000000000O100000000000000O10000000000000000O10O100000000O10O100000O10000000O01000000@jFoNV9j0QGUOP9c0XG]Oh89bGG]8OmG2d9O100000000O10O100000O100000fNJcG6]82ZGNg8:QGFo8>mFBS9>mFBS9>mFBS9=nFBS9>mFBS9>mFBS9>mFBR9>oFAR9?nFAR9?nFAR9>oFBQ9>nFCR9=nFBS9>mFBS9=nFCR9=nFCR9=nFCR9<oFCR9=nFCS9<mFDS9<mFDS9;nFDT9;lFET9;lFDV9:kFFU9:kFEV9;iFFX99hFFY9;fFEZ9;fFD\\9<cFD]9<bFD`9<^FEb9;]FFc9:[FGg99WFHi98VFHk99SFHn97PFJQ:6nEKS:5kEKV:5hEMX:4fEL\\:3cEN]:=2O0O101N100O2O0O5L3L4M3Le`h2"}, "label": "a table with two plates on it"}]}
{"id": 183653, "image": "COCO_train2014_000000183653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with food on it\"."}], "task": "refering", "answer_template": "The \"a table with food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173], "bbox": [0.2585625, 0.4244846796657381, 0.604765625, 0.6041782729805013], "iscrowd": 0, "area": 8131.171500000004, "rle": {"size": [359, 640], "counts": "UQj11h:?N2O0O2O1N101N101N2O0O2O1N101O0O2O0O101N100O2O0O2N1O110O00010O000010O01O01O01O00010O001O00010O00000000000001O0000000000001O0000000000001O000000O10000O1000000O1000000O10O1000O1000000O100FcNjF]1V9fNgFZ1Y9iNcFX1]99O100000000000000O100000000000000O10000000000000000O10O100000000O10O100000O10000000O01000000@jFoNV9j0QGUOP9c0XG]Oh89bGG]8OmG2d9O100000000O10O100000O100000fNJcG6]82ZGNg8:QGFo8>mFBS9>mFBS9>mFBS9=nFBS9>mFBS9>mFBS9>mFBR9>oFAR9?nFAR9?nFAR9>oFBQ9>nFCR9=nFBS9>mFBS9=nFCR9=nFCR9=nFCR9<oFCR9=nFCS9<mFDS9<mFDS9;nFDT9;lFET9;lFDV9:kFFU9:kFEV9;iFFX99hFFY9;fFEZ9;fFD\\9<cFD]9<bFD`9<^FEb9;]FFc9:[FGg99WFHi98VFHk99SFHn97PFJQ:6nEKS:5kEKV:5hEMX:4fEL\\:3cEN]:=2O0O101N100O2O0O5L3L4M3Le`h2"}, "label": "a table with food on it"}]}
{"id": 347495, "image": "COCO_train2014_000000347495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing grey shirt and red pant who skating in snow\"."}], "task": "refering", "answer_template": "The \"a man wearing grey shirt and red pant who skating in snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 67, 68, 69, 70, 71, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 125, 126, 127, 128, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 188, 189, 190, 203, 204, 205, 218, 219, 220, 221, 233, 234], "bbox": [0.33354460093896715, 0.13596875, 0.8225117370892019, 0.6643749999999999], "iscrowd": 0, "area": 33165.18435, "rle": {"size": [640, 426], "counts": "Zoh22jc04L4]\\OHSc02n\\O`0ob09O2N1N2^]O]Ofa0d0R^OGja0:T^OKha0Z1L3a^OfMSa0\\2l^OdMRa0`2k^OaMUa0`2j^O`MUa0e201O1O1O1N3M3N101N2N1O2O0O2N2O0O2N2N1N3K5K5M3N2N2N1O2N2M2OM31N3N1O2M2O1O2M2O1N2O2N1N2O1O100O2N100N2N2N2N2N3M2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2TEmIi7S6THRJk7n5lG\\JR8e5hGbJV8_5dGgJ[8Y5XGUKg8l4lFbKR9_4lFdKR9]4WFQJFc1j9SNQFZ66cL\\9aM\\Fm54gL]9`M\\Fj54iL^9m3_FUL`9V70[J_Fk0a9UO_Fk0a9dNbFTLNX5`9bNjFoKG_5_9`NUG_1l8^NVGb1j8]NXGb1h8\\NZGd1f8WN_Gh1d8TN_Gk1b8RN`Gn1a8PNaGo1_8PNbGo1^8QNdGn1\\8PNfGP2Y8PNhGP2X8oMiGQ2W8mMkGR2U8nMmGQ2S8mMoGS2P8mMQHS2o7nMPHQ2P8PNPHP2P8QNPHn1o7SNRHl1n7TNRHk1o7VNQHi1n7XNSHg1m7ZNRHf1m7[NTHd1l7\\NUHc1j7_NUHa1k7_NVH_1k7bNUH]1k7cNUH]1k7cNVH\\1j7eNVHZ1j7fNVHZ1k7fNUHY1k7gNVHW1k7iNUHW1k7jNUHU1k7\\M\\FWOj1\\3j7]M_FUOg1]3j7^McFQOc1a2WNUMc9Y1gFmN`1a2ZNXM`9X1hFoN^1]2^N[M\\9Y1iFnN`1X2_N`MY9Y1iFoN`1T2aNcMV9Y1kFoN`1P2bNhMT9X1jFPOb1m1aNkMS9W1lFQOa1e1kM]Mg0f0Q9T1oFTO`1b1jM^Mh0h0P9m0TG[O[1`1fNWNk8m0UG]OZ1\\1hNZNj8k0VG^OY1[1hN\\Ni8j0WG_OY1Y1iN]Nh8i0XGAV1X1kN^Ng8h0ZGAU1U1nNaNd8g0ZGCT1P1SOfN_8g0[GBS1m0WOjN\\8e0\\GDQ1g0]OoNW8e0[GEQ1b0BTOR8e0\\GDQ1=FZOn7d0[GFQ16K@j7c0[GFQ12NEf7d0[GEQ1L3Kc7b0YGGR1F71`7`0XGIR1_O;7^7?VGJY4Ee4<TG0V4Dh47VG4R4Dk43VG9o3CQ5MSG?l3C[5BlFk0i3B[7=eHB\\7>dHB\\7>dHA]7`0bH_O_7a0cH\\O^7d0eHXO[7i0n51O00001O001000O1O001O1O01N2O001001N100O1O1O2N1N2N3MdY_1"}, "label": "a man wearing grey shirt and red pant who skating in snow"}]}
{"id": 347495, "image": "COCO_train2014_000000347495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in red pants snowboarding\"."}], "task": "refering", "answer_template": "The \"a person in red pants snowboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 67, 68, 69, 70, 71, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 125, 126, 127, 128, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 188, 189, 190, 203, 204, 205, 218, 219, 220, 221, 233, 234], "bbox": [0.33354460093896715, 0.13596875, 0.8225117370892019, 0.6643749999999999], "iscrowd": 0, "area": 33165.18435, "rle": {"size": [640, 426], "counts": "Zoh22jc04L4]\\OHSc02n\\O`0ob09O2N1N2^]O]Ofa0d0R^OGja0:T^OKha0Z1L3a^OfMSa0\\2l^OdMRa0`2k^OaMUa0`2j^O`MUa0e201O1O1O1N3M3N101N2N1O2O0O2N2O0O2N2N1N3K5K5M3N2N2N1O2N2M2OM31N3N1O2M2O1O2M2O1N2O2N1N2O1O100O2N100N2N2N2N2N3M2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2TEmIi7S6THRJk7n5lG\\JR8e5hGbJV8_5dGgJ[8Y5XGUKg8l4lFbKR9_4lFdKR9]4WFQJFc1j9SNQFZ66cL\\9aM\\Fm54gL]9`M\\Fj54iL^9m3_FUL`9V70[J_Fk0a9UO_Fk0a9dNbFTLNX5`9bNjFoKG_5_9`NUG_1l8^NVGb1j8]NXGb1h8\\NZGd1f8WN_Gh1d8TN_Gk1b8RN`Gn1a8PNaGo1_8PNbGo1^8QNdGn1\\8PNfGP2Y8PNhGP2X8oMiGQ2W8mMkGR2U8nMmGQ2S8mMoGS2P8mMQHS2o7nMPHQ2P8PNPHP2P8QNPHn1o7SNRHl1n7TNRHk1o7VNQHi1n7XNSHg1m7ZNRHf1m7[NTHd1l7\\NUHc1j7_NUHa1k7_NVH_1k7bNUH]1k7cNUH]1k7cNVH\\1j7eNVHZ1j7fNVHZ1k7fNUHY1k7gNVHW1k7iNUHW1k7jNUHU1k7\\M\\FWOj1\\3j7]M_FUOg1]3j7^McFQOc1a2WNUMc9Y1gFmN`1a2ZNXM`9X1hFoN^1]2^N[M\\9Y1iFnN`1X2_N`MY9Y1iFoN`1T2aNcMV9Y1kFoN`1P2bNhMT9X1jFPOb1m1aNkMS9W1lFQOa1e1kM]Mg0f0Q9T1oFTO`1b1jM^Mh0h0P9m0TG[O[1`1fNWNk8m0UG]OZ1\\1hNZNj8k0VG^OY1[1hN\\Ni8j0WG_OY1Y1iN]Nh8i0XGAV1X1kN^Ng8h0ZGAU1U1nNaNd8g0ZGCT1P1SOfN_8g0[GBS1m0WOjN\\8e0\\GDQ1g0]OoNW8e0[GEQ1b0BTOR8e0\\GDQ1=FZOn7d0[GFQ16K@j7c0[GFQ12NEf7d0[GEQ1L3Kc7b0YGGR1F71`7`0XGIR1_O;7^7?VGJY4Ee4<TG0V4Dh47VG4R4Dk43VG9o3CQ5MSG?l3C[5BlFk0i3B[7=eHB\\7>dHB\\7>dHA]7`0bH_O_7a0cH\\O^7d0eHXO[7i0n51O00001O001000O1O001O1O01N2O001001N100O1O1O2N1N2N3MdY_1"}, "label": "a person in red pants snowboarding"}]}
{"id": 560489, "image": "COCO_train2014_000000560489.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 potted plants , the smallest and the tallest\"."}], "task": "refering", "answer_template": "The \"2 potted plants , the smallest and the tallest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 38, 39, 40, 41, 57, 58, 75, 76, 93, 94, 111], "bbox": [0.15338, 0.10786666666666668, 0.31012, 0.47192000000000006], "iscrowd": 0, "area": 5851.551550000001, "rle": {"size": [375, 500], "counts": "YXl01f;2N3M2N1O1`FOX71cH7Z7JaH<^7D^Ha0kNWOW88mHd0dN]O\\80oHV1m6kNQIY1l6hNRIY1^6[NfH=k0X1]6`NeH9l0X1]6dNdH5m0X1]6gNdH3m0W1_6hNaH2n0W1c6]O\\Ic0f6\\OXIe0i6[OUIg0k6YOTIg0m6YOQIh0P7YOnHg0R7ZOmHg0S7YOlHg0T7ZOkHf0V7ZOhHg0Y7XO\\HhN3P2a7YOZHoNNh1g7[OYHQ1h7POVHQ1i7_11O00000O100000001O00HZL_He3b780000001N1000000N2eMoHN]7FgHTO@;j7?bI^O_6`0dI^O]6`0gI]OZ6c0hIZOY6e0PJROl5R1Y240O1O1O1O1O001O1O1O102N9F5L3M2N6IiXn3"}, "label": "2 potted plants , the smallest and the tallest"}]}
{"id": 560489, "image": "COCO_train2014_000000560489.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall green plant and short green plant to the left of another small green plant\"."}], "task": "refering", "answer_template": "The \"a tall green plant and short green plant to the left of another small green plant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 38, 39, 40, 41, 57, 58, 75, 76, 93, 94, 111], "bbox": [0.15338, 0.10786666666666668, 0.31012, 0.47192000000000006], "iscrowd": 0, "area": 5851.551550000001, "rle": {"size": [375, 500], "counts": "YXl01f;2N3M2N1O1`FOX71cH7Z7JaH<^7D^Ha0kNWOW88mHd0dN]O\\80oHV1m6kNQIY1l6hNRIY1^6[NfH=k0X1]6`NeH9l0X1]6dNdH5m0X1]6gNdH3m0W1_6hNaH2n0W1c6]O\\Ic0f6\\OXIe0i6[OUIg0k6YOTIg0m6YOQIh0P7YOnHg0R7ZOmHg0S7YOlHg0T7ZOkHf0V7ZOhHg0Y7XO\\HhN3P2a7YOZHoNNh1g7[OYHQ1h7POVHQ1i7_11O00000O100000001O00HZL_He3b780000001N1000000N2eMoHN]7FgHTO@;j7?bI^O_6`0dI^O]6`0gI]OZ6c0hIZOY6e0PJROl5R1Y240O1O1O1O1O001O1O1O102N9F5L3M2N6IiXn3"}, "label": "a tall green plant and short green plant to the left of another small green plant"}]}
{"id": 28012, "image": "COCO_train2014_000000028012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the portion of fence to the left of the girl\"."}], "task": "refering", "answer_template": "The \"the portion of fence to the left of the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 123, 135, 136, 137, 138, 150, 151, 152, 165, 166, 169, 170, 180, 181, 183, 184, 185, 195, 196, 197, 198, 199, 200, 213], "bbox": [0.0, 0.27745312499999997, 0.36953051643192486, 0.620453125], "iscrowd": 0, "area": 17219.7396, "rle": {"size": [640, 426], "counts": "g6o4o=R1O1O1O10000000000000000000000000O100000000000000000000003M1O1O1O1O1O1O001O1O1O1O100OM3I7J6jMV2O100O100O100O100O100O10000O100O10000O101N1SOc_OWN]`0i1m0O10000000000O10SBmMm9S2jEWNU:i1jEXNV:h1iEYNW:g1iEZNV:e1jE\\NV:d1iE^NV:b1iE_NW:a1iE`NV:`1iEaNW:_1hEcNX:[1iEeNW:T1PFlNP:k0YFVOf9b0aF_O_99iFHV9OSG1m8G[G:d8A`G`0`8_OaG?a8@`G>b8A_G=c8B]G>d8A]G=e8A]G=f8AZG>h8AYG>h8AYG=i8AXG>j8AUGa0k8^OUGc0k8\\OUGe0k8YOVGh0j8WOVGj0j8UOUGm0k8ROUGo0k8oNVGR1k=1O1N2O1O1O1N3N2N1O2N2N1O2O1N2N1O2N2O0O2N2N1O2N2O1N1O2N2OO1D;D=C=K4N3N1N3N2M2N3N2M2O7HVjW5"}, "label": "the portion of fence to the left of the girl"}]}
{"id": 28012, "image": "COCO_train2014_000000028012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fence to the left behind the woman\"."}], "task": "refering", "answer_template": "The \"the fence to the left behind the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 123, 135, 136, 137, 138, 150, 151, 152, 165, 166, 169, 170, 180, 181, 183, 184, 185, 195, 196, 197, 198, 199, 200, 213], "bbox": [0.0, 0.27745312499999997, 0.36953051643192486, 0.620453125], "iscrowd": 0, "area": 17219.7396, "rle": {"size": [640, 426], "counts": "g6o4o=R1O1O1O10000000000000000000000000O100000000000000000000003M1O1O1O1O1O1O001O1O1O1O100OM3I7J6jMV2O100O100O100O100O100O10000O100O10000O101N1SOc_OWN]`0i1m0O10000000000O10SBmMm9S2jEWNU:i1jEXNV:h1iEYNW:g1iEZNV:e1jE\\NV:d1iE^NV:b1iE_NW:a1iE`NV:`1iEaNW:_1hEcNX:[1iEeNW:T1PFlNP:k0YFVOf9b0aF_O_99iFHV9OSG1m8G[G:d8A`G`0`8_OaG?a8@`G>b8A_G=c8B]G>d8A]G=e8A]G=f8AZG>h8AYG>h8AYG=i8AXG>j8AUGa0k8^OUGc0k8\\OUGe0k8YOVGh0j8WOVGj0j8UOUGm0k8ROUGo0k8oNVGR1k=1O1N2O1O1O1N3N2N1O2N2N1O2O1N2N1O2N2O0O2N2N1O2N2O1N1O2N2OO1D;D=C=K4N3N1N3N2M2N3N2M2O7HVjW5"}, "label": "the fence to the left behind the woman"}]}
{"id": 314734, "image": "COCO_train2014_000000314734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cup extending off the frame of the photo\"."}], "task": "refering", "answer_template": "The \"a brown cup extending off the frame of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55], "bbox": [0.17703125, 0.0011007025761124122, 0.455546875, 0.2], "iscrowd": 0, "area": 10783.0326, "rle": {"size": [427, 640], "counts": "kS_11Z=3M3M3M3M3M3M3M3M4L3M3N2M3M3M3M3M3M4L3M3M3M3M3M1OO1O100O1O100O1O1O100O1O100O1O1000000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000001O01O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O0000L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L4L4L4LQaa4"}, "label": "a brown cup extending off the frame of the photo"}]}
{"id": 314734, "image": "COCO_train2014_000000314734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red cup\"."}], "task": "refering", "answer_template": "The \"red cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55], "bbox": [0.17703125, 0.0011007025761124122, 0.455546875, 0.2], "iscrowd": 0, "area": 10783.0326, "rle": {"size": [427, 640], "counts": "kS_11Z=3M3M3M3M3M3M3M3M4L3M3N2M3M3M3M3M3M4L3M3M3M3M3M1OO1O100O1O100O1O1O100O1O100O1O1000000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000001O01O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O0000L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L4L4L4LQaa4"}, "label": "red cup"}]}
{"id": 314734, "image": "COCO_train2014_000000314734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lower portion of wrap\"."}], "task": "refering", "answer_template": "The \"lower portion of wrap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.100515625, 0.42077283372365337, 0.61, 0.7896018735362997], "iscrowd": 0, "area": 39456.019, "rle": {"size": [427, 640], "counts": "RZk04V=4M3L5K4L4L3M3N1N3M2N3M3M2O2M2N3M3M2N3N1N3M3M2N3N2M2N3M2N3M3N1N3M2N3M3M2O1N2N2N1O2N101O001O001O1O001O010O001O1O001O001O0010O01O1O001O001O0010O01O1O001O001O00100O001O001O001O001O10O01O001O001O001O10O01O001O001O1O0010O01O001O001O1O001O010O001O1O001O001O010O000000000000000001O00000000000000001O00000O10000000001O000000O10000000000O100000000O1000000000000000O10000000001O000000000000000000001O0000000000000000001O00000O100000000000001O0000000000000000000000000000001O000000000000000000000000000000000000000000001O00000000000000000000000000001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3M3M3M3N1N3M3M2N2N2O1N2N2N2N2N1O2O1N2N2N2H8G9G9H7H9G9GUQX3"}, "label": "lower portion of wrap"}]}
{"id": 314734, "image": "COCO_train2014_000000314734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a veg roll which lies on a cutting pad\"."}], "task": "refering", "answer_template": "The \"a veg roll which lies on a cutting pad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.100515625, 0.42077283372365337, 0.61, 0.7896018735362997], "iscrowd": 0, "area": 39456.019, "rle": {"size": [427, 640], "counts": "RZk04V=4M3L5K4L4L3M3N1N3M2N3M3M2O2M2N3M3M2N3N1N3M3M2N3N2M2N3M2N3M3N1N3M2N3M3M2O1N2N2N1O2N101O001O001O1O001O010O001O1O001O001O0010O01O1O001O001O0010O01O1O001O001O00100O001O001O001O001O10O01O001O001O001O10O01O001O001O1O0010O01O001O001O1O001O010O001O1O001O001O010O000000000000000001O00000000000000001O00000O10000000001O000000O10000000000O100000000O1000000000000000O10000000001O000000000000000000001O0000000000000000001O00000O100000000000001O0000000000000000000000000000001O000000000000000000000000000000000000000000001O00000000000000000000000000001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3M3M3M3N1N3M3M2N2N2O1N2N2N2N2N1O2O1N2N2N2H8G9G9H7H9G9GUQX3"}, "label": "a veg roll which lies on a cutting pad"}]}
{"id": 314734, "image": "COCO_train2014_000000314734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wrap resting on the other wrap\"."}], "task": "refering", "answer_template": "The \"the wrap resting on the other wrap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148], "bbox": [0.166421875, 0.14156908665105386, 0.48726562500000004, 0.44494145199063234], "iscrowd": 0, "area": 21146.726149999995, "rle": {"size": [427, 640], "counts": "nh\\1:l<8H4L5H8H8I7J6J6J6J6J6L4O1O0O2O1N2O1O1N2O1O1N2O1O1O001O1O1O1O100O1O1O1O1O1O1O100O101O00000000001O00000O101O00000000001O0000000O2O01O00000001O000000001O00000000010O000000001O00000000001O01O000000010O000000010O0000000010O000000010O000001O0001O0001O00000O101O000000001O0O1000001O0000000O2O000000001O000O10001O0000001N100000001O001O0O2O1O1O1O1O1O1O1N2O001O1O1O1O1O1N2O1N2N1O2N2N2N2N2N2N2N2O2M3M3M3M3M3M4L3M9G?A?AlcX4"}, "label": "the wrap resting on the other wrap"}]}
{"id": 298350, "image": "COCO_train2014_000000298350.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboard under the boy with blue shorts\"."}], "task": "refering", "answer_template": "The \"the skateboard under the boy with blue shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 256, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 324, 325, 335, 336], "bbox": [0.0, 0.74981220657277, 0.783921875, 1.0], "iscrowd": 0, "area": 16985.44810000001, "rle": {"size": [426, 640], "counts": "Q:j0_<6K1O1O1O1O]OmCOS<0oC0P<ORD1m;NUD1k;MXD3g;L[D3d;M^D3a;KbD4_;JdD4];JeD6[;HgD7Y;GjD8U<00O100O2N01O001O00001O001O01O01O00001O000000001O000001O0000000000000001O000000000O100000000lC9e:HZE9e:GZE:e:G[E:d:F[E<d:D\\E=c:C\\E?c:A\\E`0c:A]E`0`:B_E`0]:CcE>Z:DeE=X:FhE:V:IhE8V:JjE6S:MlE4S:MmE3S:MlE4T:LlE4T:LlE4T:LlE4T:LlE4T:LkE5U:KkE5U:KjE5W:LhE2Z:OdE1]:3^ELd:8XEFj:\\11N2N2O1O1N2O1O1N2O1O1N2O100O1O1O100O1O100O1O1O2O00000000000000000O1O100O100N2CWDWOk;i0WDROl;n080000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000001O00000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000001O000000000000000000000000000000000O100000000000000O1000000000000001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00010O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001N100O2O0O2N100OTC3^<i0C101N2O0N3N5K<D00000000O100000000000000000O1000000000000000000000000000000^Ob0000VOZEPOf:i0hEPOX:j0UFoNk9k0\\110000000O100000O1000O10000000000O010000000O100000O1000O1000000000O0100000000O100000O10O100O100O100O010O100O100O1O010O10O0100O010O100O010O1O010O10O0100O010O100O00100O010O100O100O2O0O100O1O101N100OhRi1"}, "label": "the skateboard under the boy with blue shorts"}]}
{"id": 298350, "image": "COCO_train2014_000000298350.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard being riden by a man\"."}], "task": "refering", "answer_template": "The \"a skateboard being riden by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 256, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 324, 325, 335, 336], "bbox": [0.0, 0.74981220657277, 0.783921875, 1.0], "iscrowd": 0, "area": 16985.44810000001, "rle": {"size": [426, 640], "counts": "Q:j0_<6K1O1O1O1O]OmCOS<0oC0P<ORD1m;NUD1k;MXD3g;L[D3d;M^D3a;KbD4_;JdD4];JeD6[;HgD7Y;GjD8U<00O100O2N01O001O00001O001O01O01O00001O000000001O000001O0000000000000001O000000000O100000000lC9e:HZE9e:GZE:e:G[E:d:F[E<d:D\\E=c:C\\E?c:A\\E`0c:A]E`0`:B_E`0]:CcE>Z:DeE=X:FhE:V:IhE8V:JjE6S:MlE4S:MmE3S:MlE4T:LlE4T:LlE4T:LlE4T:LlE4T:LkE5U:KkE5U:KjE5W:LhE2Z:OdE1]:3^ELd:8XEFj:\\11N2N2O1O1N2O1O1N2O1O1N2O100O1O1O100O1O100O1O1O2O00000000000000000O1O100O100N2CWDWOk;i0WDROl;n080000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000001O00000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000001O000000000000000000000000000000000O100000000000000O1000000000000001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00010O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001N100O2O0O2N100OTC3^<i0C101N2O0N3N5K<D00000000O100000000000000000O1000000000000000000000000000000^Ob0000VOZEPOf:i0hEPOX:j0UFoNk9k0\\110000000O100000O1000O10000000000O010000000O100000O1000O1000000000O0100000000O100000O10O100O100O100O010O100O100O1O010O10O0100O010O100O010O1O010O10O0100O010O100O00100O010O100O100O2O0O100O1O101N100OhRi1"}, "label": "a skateboard being riden by a man"}]}
{"id": 355697, "image": "COCO_train2014_000000355697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with a drink in hand\"."}], "task": "refering", "answer_template": "The \"girl with a drink in hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 167, 168, 169, 170, 171, 191, 192, 193, 194, 195, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333, 334, 350, 351, 352, 353, 354, 372, 373, 374], "bbox": [0.19259375, 0.3734937238493724, 0.563296875, 0.9995606694560669], "iscrowd": 0, "area": 32461.53920000001, "rle": {"size": [478, 640], "counts": "Uli11k>2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2O1N3M2N2N2O1N2N2N2O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1N2O1O2O0O1O1O100O1O0001N10000O2O0O100O2O0O^JmM\\NQ2e1RNYNm1f1WNYNg1g1[NXNd1h1_NVN_1k1cNTN]1k1fNTNY1j1jNUNW1j1jNUNV1l1kNSNT1n1lNQNU1n1mNPNS1Q2mNoMR1R2oNlMR1S2POkMP1U2ROjMn0U2SOjMm0W2TOgMm0X2UOgMk0X2WOfMi0Z2YOdMh0[2ZOdMf0[2\\OcMd0\\2_OcMa0[2BcM>]2DbM<\\2GbM:^2H`M7a2K]M4d2NYM3g2OWM0j21TMOm23QMLP36mLJT38jLGW3;fLE[3=cLC]3?`LC_3?_LC_3?^LD`3>^LE^3?^LF^3<`LF^3<`JdM>R2Q5<_IdN^1Q1R5n1kJTNT5m1kJUNS5l1kJVNT5k1kJVNT5k1jJWNU5S2]JmMg5[2gImM[6Y52N3M2N3M2N3M200O101N100O100O10O0100O101N101N101N101N105J3N0O10001O00001N1000O10VIUHa6m7\\ITHd6m7ZIUHd6T8O001O002N2M4M2N3M3M1O0O1000O1000O100000000O2O0000000O2O001O1O0O2O1O001N2O1O1O1O0O2O1O1O1N2O1O1O1O2N2N2N2N2mNXFlLj9Q3ZFlLh9Q3[FmLg9P3]FmLe9Q3\\FnLf9P3[FoLg9n2[FQMg9m2ZFRMh9l2ZFRMh9l2ZFQMi9m2XFRMj9l2XFRMj9k2W1M3M3M3M2O2M3M3M4`N_C=c<@bC<a<AeC9]<DkC5X<CYD0j;GfD0\\YR4"}, "label": "girl with a drink in hand"}]}
{"id": 355697, "image": "COCO_train2014_000000355697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women holding a beer bottle sitting between two pwesons playing vilie\"."}], "task": "refering", "answer_template": "The \"a women holding a beer bottle sitting between two pwesons playing vilie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 167, 168, 169, 170, 171, 191, 192, 193, 194, 195, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333, 334, 350, 351, 352, 353, 354, 372, 373, 374], "bbox": [0.19259375, 0.3734937238493724, 0.563296875, 0.9995606694560669], "iscrowd": 0, "area": 32461.53920000001, "rle": {"size": [478, 640], "counts": "Uli11k>2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2O1N3M2N2N2O1N2N2N2O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1N2O1O2O0O1O1O100O1O0001N10000O2O0O100O2O0O^JmM\\NQ2e1RNYNm1f1WNYNg1g1[NXNd1h1_NVN_1k1cNTN]1k1fNTNY1j1jNUNW1j1jNUNV1l1kNSNT1n1lNQNU1n1mNPNS1Q2mNoMR1R2oNlMR1S2POkMP1U2ROjMn0U2SOjMm0W2TOgMm0X2UOgMk0X2WOfMi0Z2YOdMh0[2ZOdMf0[2\\OcMd0\\2_OcMa0[2BcM>]2DbM<\\2GbM:^2H`M7a2K]M4d2NYM3g2OWM0j21TMOm23QMLP36mLJT38jLGW3;fLE[3=cLC]3?`LC_3?_LC_3?^LD`3>^LE^3?^LF^3<`LF^3<`JdM>R2Q5<_IdN^1Q1R5n1kJTNT5m1kJUNS5l1kJVNT5k1kJVNT5k1jJWNU5S2]JmMg5[2gImM[6Y52N3M2N3M2N3M200O101N100O100O10O0100O101N101N101N101N105J3N0O10001O00001N1000O10VIUHa6m7\\ITHd6m7ZIUHd6T8O001O002N2M4M2N3M3M1O0O1000O1000O100000000O2O0000000O2O001O1O0O2O1O001N2O1O1O1O0O2O1O1O1N2O1O1O1O2N2N2N2N2mNXFlLj9Q3ZFlLh9Q3[FmLg9P3]FmLe9Q3\\FnLf9P3[FoLg9n2[FQMg9m2ZFRMh9l2ZFRMh9l2ZFQMi9m2XFRMj9l2XFRMj9k2W1M3M3M3M2O2M3M3M4`N_C=c<@bC<a<AeC9]<DkC5X<CYD0j;GfD0\\YR4"}, "label": "a women holding a beer bottle sitting between two pwesons playing vilie"}]}
{"id": 454002, "image": "COCO_train2014_000000454002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white table with a cake on top\"."}], "task": "refering", "answer_template": "The \"a white table with a cake on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.11721698113207547, 0.537703125, 0.9500707547169811, 1.0], "iscrowd": 0, "area": 92553.04985, "rle": {"size": [640, 424], "counts": "efo0h0lb0=L4L4K5L5K4L4L4K5L6b_OWN\\>R2RAkNZ>]1VAROb>V3O2N1O1Og0XO7I7G9G;E=C=D=B:F6M3000000O100000000O100000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1N2O001O1O1O001O1O1O1O001O1O1N6K4L4L5K4L5K4L4L5K4L5kNaB_Kc=U4QC_KS=U4e1C<I8H7J6J7I6J7H7J6J7I6J7H7J6J7I6I8I6JWc<"}, "label": "a white table with a cake on top"}]}
{"id": 454002, "image": "COCO_train2014_000000454002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table covered with a white table cloth\"."}], "task": "refering", "answer_template": "The \"a table covered with a white table cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.11721698113207547, 0.537703125, 0.9500707547169811, 1.0], "iscrowd": 0, "area": 92553.04985, "rle": {"size": [640, 424], "counts": "efo0h0lb0=L4L4K5L5K4L4L4K5L6b_OWN\\>R2RAkNZ>]1VAROb>V3O2N1O1Og0XO7I7G9G;E=C=D=B:F6M3000000O100000000O100000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1N2O001O1O1O001O1O1O1O001O1O1N6K4L4L5K4L5K4L4L5K4L5kNaB_Kc=U4QC_KS=U4e1C<I8H7J6J7I6J7H7J6J7I6J7H7J6J7I6I8I6JWc<"}, "label": "a table covered with a white table cloth"}]}
{"id": 454002, "image": "COCO_train2014_000000454002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"upper portion of the table in which the flowers are kept\"."}], "task": "refering", "answer_template": "The \"upper portion of the table in which the flowers are kept\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 201, 202, 203, 204, 205, 206, 207, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.20778301886792452, 0.540265625, 0.8938915094339622, 0.68571875], "iscrowd": 0, "area": 13156.2582, "rle": {"size": [640, 424], "counts": "R\\g11oc06J5L4K6K4K6J5L5J5L2M0001O001O0O101O001O00001O00001O001O00001O001O000O2O001O00000000000000000000001O0000000O100000000000001O000000000000000000000O2O0000000000O100O1O1O1O1O1O1O1O1O1O1O001O100O1O1O1O100O100000000000000O1000O1000000000O100000000000000O1000000000O1000O100000000000000O100000000000000O010000000000000O100000000000000O1000O1000000000O100000000000000O1000000000O1000O100000000000000O100000000000000O010N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N21O1N2O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1N3N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1N2O1O1O2N1O1O1O1O1Oifk0"}, "label": "upper portion of the table in which the flowers are kept"}]}
{"id": 454002, "image": "COCO_train2014_000000454002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a decorated table in which wedding cake is arranged and placed\"."}], "task": "refering", "answer_template": "The \"a decorated table in which wedding cake is arranged and placed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 201, 202, 203, 204, 205, 206, 207, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.20778301886792452, 0.540265625, 0.8938915094339622, 0.68571875], "iscrowd": 0, "area": 13156.2582, "rle": {"size": [640, 424], "counts": "R\\g11oc06J5L4K6K4K6J5L5J5L2M0001O001O0O101O001O00001O00001O001O00001O001O000O2O001O00000000000000000000001O0000000O100000000000001O000000000000000000000O2O0000000000O100O1O1O1O1O1O1O1O1O1O1O001O100O1O1O1O100O100000000000000O1000O1000000000O100000000000000O1000000000O1000O100000000000000O100000000000000O010000000000000O100000000000000O1000O1000000000O100000000000000O1000000000O1000O100000000000000O100000000000000O010N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N21O1N2O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1N3N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1N2O1O1O2N1O1O1O1O1Oifk0"}, "label": "a decorated table in which wedding cake is arranged and placed"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toothbrush that is blue and white\"."}], "task": "refering", "answer_template": "The \"the toothbrush that is blue and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 19, 20, 21, 22, 23, 24, 36, 37, 38, 39, 40, 54, 55, 56, 57, 72, 73, 74, 75, 90, 91, 108], "bbox": [0.007, 0.011904761904761904, 0.427, 0.8166666666666667], "iscrowd": 0, "area": 12629.125, "rle": {"size": [210, 500], "counts": "ml0f2`3;00010O000001O01O0001O01O0001O01O0001O0001O01O0001O01O0001O01O000010O0001O010O000010O00010O00100O10O01O10O01O100O00100O010O1O10O01O10O01O10O0100O1O010N2O001O1O1N101O1O001N2O1O001O1O0O2O1O001O1N2O001O1O0O2O1O0O2O1N1O2O1N101N2N2O0O2N2O010O1O010O1O010O1O010O1O10O01O10O01O10O01O10O01O10O0100O1O010O1O010O1O010O1O100O1O1O100O1O010O1O100O1O100O1O100O1O100O1O10O01O100O1O100O1O100O1O1O100O1O10O01O100O1O100O1O100O1O100OnQk1"}, "label": "the toothbrush that is blue and white"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue color brush is side of other yellow color brush\"."}], "task": "refering", "answer_template": "The \"a blue color brush is side of other yellow color brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 19, 20, 21, 22, 23, 24, 36, 37, 38, 39, 40, 54, 55, 56, 57, 72, 73, 74, 75, 90, 91, 108], "bbox": [0.007, 0.011904761904761904, 0.427, 0.8166666666666667], "iscrowd": 0, "area": 12629.125, "rle": {"size": [210, 500], "counts": "ml0f2`3;00010O000001O01O0001O01O0001O01O0001O0001O01O0001O01O0001O01O000010O0001O010O000010O00010O00100O10O01O10O01O100O00100O010O1O10O01O10O01O10O0100O1O010N2O001O1O1N101O1O001N2O1O001O1O0O2O1O001O1N2O001O1O0O2O1O0O2O1N1O2O1N101N2N2O0O2N2O010O1O010O1O010O1O010O1O10O01O10O01O10O01O10O01O10O0100O1O010O1O010O1O010O1O100O1O1O100O1O010O1O100O1O100O1O100O1O100O1O10O01O100O1O100O1O100O1O1O100O1O10O01O100O1O100O1O100O1O100OnQk1"}, "label": "a blue color brush is side of other yellow color brush"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red handle\"."}], "task": "refering", "answer_template": "The \"the red handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141], "bbox": [0.539, 0.24285714285714285, 0.991, 0.9880952380952381], "iscrowd": 0, "area": 17400.0, "rle": {"size": [210, 500], "counts": "Ubg11`62N101N1O2O0O2N1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O2N100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O2O0O1O1O1O1O10O0001O0001O01O00001O01O01O0000010O00001O00010O00000010O00100O00100O001O10O01O10O01O010O1O00100O00100O00100O001O010O1O010O1O010O1O00100O0010O01O10O01Oo0QOS1nNR1mNde0"}, "label": "the red handle"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red toothbrush\"."}], "task": "refering", "answer_template": "The \"the red toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141], "bbox": [0.539, 0.24285714285714285, 0.991, 0.9880952380952381], "iscrowd": 0, "area": 17400.0, "rle": {"size": [210, 500], "counts": "Ubg11`62N101N1O2O0O2N1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O2N100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O2O0O1O1O1O1O10O0001O0001O01O00001O01O01O0000010O00001O00010O00000010O00100O00100O001O10O01O10O01O010O1O00100O00100O00100O001O010O1O010O1O010O1O00100O0010O01O10O01Oo0QOS1nNR1mNde0"}, "label": "the red toothbrush"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toothbrush handle with green stripes\"."}], "task": "refering", "answer_template": "The \"a toothbrush handle with green stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 61, 62, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.147, 0.006428571428571429, 1.0, 0.9897619047619047], "iscrowd": 0, "area": 32518.25, "rle": {"size": [210, 500], "counts": "o[?2_62O0O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O2O0O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O2N1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O00010O01O00010O001O01O01O00010O0010O0001O01O01O0010O0001O010O000010O0001O010O00010O001O01O01O00010O001O01O01O00010O001O01O01O010O000010O01O00010O001O01O01O0010O01O00010O0010O0001O010O0000100O1O00100O1O1O100O001O100O1O100O00100O1O100O1O10O01O100O1O100O00100O1O100O100O1O100O1O100O1O100O00100O1O010O010O00100O0010O01O010O1O010O0010O0100O0010O01O010O1O010O0010O01O10O010O01O010O1O010O0010O01O10O01O2O0O100O1O100O1O100O1O100O1O100O1O100O1O2O0O100O1O100O1O100O1O100O1O100O1O100O100O2N100O1O100O1N2O1N2O1O1N2O1O1N2O1O1N6K8GC"}, "label": "a toothbrush handle with green stripes"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the handle of a white and yellow toothbrush\"."}], "task": "refering", "answer_template": "The \"the handle of a white and yellow toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 126, 127, 128, 129], "bbox": [0.0039, 0.015047619047619048, 0.65828, 0.9865714285714287], "iscrowd": 0, "area": 24380.1239, "rle": {"size": [210, 500], "counts": "eb07Z6`0@8I0O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O101N1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O100O100O1O100O1O100O1O01O01O01O01O010O00010O0O2O00001O00001O00001O001O00001O000O2O000O2O00001N101O000O2O000O2O00001N10001O0O2O000O2M2N2010O00010O0001O010O00010O00010O00010O00010O0010O00010O00010O00010O00010O010O00001O01O01O001O000010O0001O000010O01O00001O00010O00001O0010O0001O00001O010O00001O00010O00001O001O01O01O1O1O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O010O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O1O1O1O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1OTRS1"}, "label": "the handle of a white and yellow toothbrush"}]}
{"id": 388469, "image": "COCO_train2014_000000388469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the handle of a yellow toothbrush\"."}], "task": "refering", "answer_template": "The \"the handle of a yellow toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 126, 127, 128, 129], "bbox": [0.0039, 0.015047619047619048, 0.65828, 0.9865714285714287], "iscrowd": 0, "area": 24380.1239, "rle": {"size": [210, 500], "counts": "eb07Z6`0@8I0O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O101N1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O100O100O1O100O1O100O1O01O01O01O01O010O00010O0O2O00001O00001O00001O001O00001O000O2O000O2O00001N101O000O2O000O2O00001N10001O0O2O000O2M2N2010O00010O0001O010O00010O00010O00010O00010O0010O00010O00010O00010O00010O010O00001O01O01O001O000010O0001O000010O01O00001O00010O00001O0010O0001O00001O010O00001O00010O00001O001O01O01O1O1O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O010O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O1O1O1O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1OTRS1"}, "label": "the handle of a yellow toothbrush"}]}
{"id": 396663, "image": "COCO_train2014_000000396663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red bus on the right\"."}], "task": "refering", "answer_template": "The \"the red bus on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297, 298, 320, 321], "bbox": [0.85759375, 0.15056206088992977, 1.0, 0.8898829039812647], "iscrowd": 0, "area": 20610.735649999995, "rle": {"size": [427, 640], "counts": "YQU7a1f;8K4L4L2N101O1O001N1000000O100000000O10001O0O2O2N1dGlM_5U2\\JZNS5n1fJaNn4b1lJ\\OY4^2jIkMP6l4K?YO?J5L4L5K5J2O1O1O1N2O1O001O0O2O001O0O2O00000O10000000O0101O000O10001O0O1000001N100000000O101O001N1000000O100000001N2k4UKSI"}, "label": "the red bus on the right"}]}
{"id": 396663, "image": "COCO_train2014_000000396663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus to the right of another red bus\"."}], "task": "refering", "answer_template": "The \"a red bus to the right of another red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297, 298, 320, 321], "bbox": [0.85759375, 0.15056206088992977, 1.0, 0.8898829039812647], "iscrowd": 0, "area": 20610.735649999995, "rle": {"size": [427, 640], "counts": "YQU7a1f;8K4L4L2N101O1O001N1000000O100000000O10001O0O2O2N1dGlM_5U2\\JZNS5n1fJaNn4b1lJ\\OY4^2jIkMP6l4K?YO?J5L4L5K5J2O1O1O1N2O1O001O0O2O001O0O2O00000O10000000O0101O000O10001O0O1000001N100000000O101O001N1000000O100000001N2k4UKSI"}, "label": "a red bus to the right of another red bus"}]}
{"id": 396663, "image": "COCO_train2014_000000396663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the scissor doors of a bus type vehcile\"."}], "task": "refering", "answer_template": "The \"the scissor doors of a bus type vehcile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276], "bbox": [0.0015, 0.16355971896955504, 0.248328125, 0.809297423887588], "iscrowd": 0, "area": 34792.030399999996, "rle": {"size": [427, 640], "counts": "a?]2n:i4WK^1bNO2O0000001O00001O0O10001O00001O00000O2O000000001O0000000O101O000000001O000O10001O00000000001N100000001O000000000O2O000000001O00000O10001O000000001O0O100000001O000000001N10000000001O00000O101O00000000001O0O1000001O000000000O2O000000001O0000000O101O00000jNXJiJh5j4jJQKW5`4\\K[Kd4X4mKdKS4n3`LlKa3g3QMTLo2^3dM]L\\2V3T4A>C=B>J6M4L3M3M3M3M3M4L3M3M3M3M4L3M3M3M3MZ_X6"}, "label": "the scissor doors of a bus type vehcile"}]}
{"id": 396663, "image": "COCO_train2014_000000396663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bus parked to the left of private bus number 1235\"."}], "task": "refering", "answer_template": "The \"a yellow bus parked to the left of private bus number 1235\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276], "bbox": [0.0015, 0.16355971896955504, 0.248328125, 0.809297423887588], "iscrowd": 0, "area": 34792.030399999996, "rle": {"size": [427, 640], "counts": "a?]2n:i4WK^1bNO2O0000001O00001O0O10001O00001O00000O2O000000001O0000000O101O000000001O000O10001O00000000001N100000001O000000000O2O000000001O00000O10001O000000001O0O100000001O000000001N10000000001O00000O101O00000000001O0O1000001O000000000O2O000000001O0000000O101O00000jNXJiJh5j4jJQKW5`4\\K[Kd4X4mKdKS4n3`LlKa3g3QMTLo2^3dM]L\\2V3T4A>C=B>J6M4L3M3M3M3M3M4L3M3M3M3M4L3M3M3M3MZ_X6"}, "label": "a yellow bus parked to the left of private bus number 1235"}]}
{"id": 396663, "image": "COCO_train2014_000000396663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus number 1235\"."}], "task": "refering", "answer_template": "The \"bus number 1235\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.20990625000000002, 0.16405152224824354, 0.898078125, 0.8651756440281031], "iscrowd": 0, "area": 104560.47244999999, "rle": {"size": [427, 640], "counts": "gTh14\\<V1mNS1VOk0UOj0ZOf0@;E5K5K5K5K6K4K5K5K5K6J5K5M3N2M4L3N2M3M3N3L3N2O100O2O0O1O100O2O0O100O101N1O100O10001N10000O101O0O10000O2O000O10001N10000O101O0O10000O2O00000O101O000000001N10000000001O0O1000001O00000O101O0000000O1000000000001O0000000000000000000000001O000000000000000000000000001O00000O100000000000000000001O0000000000000000000000001O000000000000000000000001O01O0001O000000001O0000000010O00000001O000000001O000000010O000000001O000000001O00000O101O0O10001N100O10000O100O2O000O100O10000O2O0O101O1N2O001N2O0O2O1O0O2O0O101O0O2O0O101O0O101O0O101N101O000010O1O10O01O10O01O01O0001O0001O011N0O2O0O101O0O2O0O2O0O2O0O101N101O0O2M2O2M2O1N3N1N3M2O2M2O1N3N1N3M2O1N2O1O100O2O0O10000O100O10000O101N10000O100O101O0O100O2O000O2O0O101O0O101N10001N2O1N3N1O2M3N1N10001N100O2O000O101N10001N100O2O000O2O0O101N10001N100010O002O1N1O2N10O001O00001N101O00001N1M4G8M3O5K6K4K6J6J2N1O2N1O2N1O2N1O1O2O0O2N1O6J9G:]Ob0YOh0YOmlj0"}, "label": "bus number 1235"}]}
{"id": 396663, "image": "COCO_train2014_000000396663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red an white bus in between two other buses\"."}], "task": "refering", "answer_template": "The \"the red an white bus in between two other buses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.20990625000000002, 0.16405152224824354, 0.898078125, 0.8651756440281031], "iscrowd": 0, "area": 104560.47244999999, "rle": {"size": [427, 640], "counts": "gTh14\\<V1mNS1VOk0UOj0ZOf0@;E5K5K5K5K6K4K5K5K5K6J5K5M3N2M4L3N2M3M3N3L3N2O100O2O0O1O100O2O0O100O101N1O100O10001N10000O101O0O10000O2O000O10001N10000O101O0O10000O2O00000O101O000000001N10000000001O0O1000001O00000O101O0000000O1000000000001O0000000000000000000000001O000000000000000000000000001O00000O100000000000000000001O0000000000000000000000001O000000000000000000000001O01O0001O000000001O0000000010O00000001O000000001O000000010O000000001O000000001O00000O101O0O10001N100O10000O100O2O000O100O10000O2O0O101O1N2O001N2O0O2O1O0O2O0O101O0O2O0O101O0O101O0O101N101O000010O1O10O01O10O01O01O0001O0001O011N0O2O0O101O0O2O0O2O0O2O0O101N101O0O2M2O2M2O1N3N1N3M2O2M2O1N3N1N3M2O1N2O1O100O2O0O10000O100O10000O101N10000O100O101O0O100O2O000O2O0O101O0O101N10001N2O1N3N1O2M3N1N10001N100O2O000O101N10001N100O2O000O2O0O101N10001N100010O002O1N1O2N10O001O00001N101O00001N1M4G8M3O5K6K4K6J6J2N1O2N1O2N1O2N1O1O2O0O2N1O6J9G:]Ob0YOh0YOmlj0"}, "label": "the red an white bus in between two other buses"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third horse in is holding a rider with a blue flag\"."}], "task": "refering", "answer_template": "The \"the third horse in is holding a rider with a blue flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.48637499999999995, 0.4036556603773585, 0.665546875, 1.0], "iscrowd": 0, "area": 14354.099599999996, "rle": {"size": [424, 640], "counts": "[RQ41Z=4L5K7I4MN100O10O0cCGT1IQ9]1bFhN^9_1WFdNj9c1kEaNU:e1aE_N_:U2O1000000O10O1SOXFZNh9g1]O`1M4L3M3M3M3M3oFXL^8[4L4M3M4L2N1O101N2N2O2M2N2O1N2N2O1N2OO0UOUIZKm6e4j0WOkGbLU8W3UHeLk7W3]HfLc7X3bHdL^7[3gHbLY7\\3Y1011N101N2O1O1N101N2O1N2O0O2N2_IQLf3R4nK\\Lm3g3kJbMR5_2lInKi0h1W5d2eJ`MY5b2`JdM_5^2XJjMh5k4O1O100OGXJTIh5d6a0TOiI`J^6\\5g0M4L3O2N3N3M2N2N7I;E5H6G9M3N2M2N3N2ROhFiMZ9V2SG\\Mn8e2j00O10fMfEg1Z:ZNhEc1Y:\\NkE5n0_OX9;mE1a;NbDMe;N^DMb<KR]h2"}, "label": "the third horse in is holding a rider with a blue flag"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown horse with a white forehead and stripe down its nose between another dark brown horse and a light brown horse\"."}], "task": "refering", "answer_template": "The \"a dark brown horse with a white forehead and stripe down its nose between another dark brown horse and a light brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.48637499999999995, 0.4036556603773585, 0.665546875, 1.0], "iscrowd": 0, "area": 14354.099599999996, "rle": {"size": [424, 640], "counts": "[RQ41Z=4L5K7I4MN100O10O0cCGT1IQ9]1bFhN^9_1WFdNj9c1kEaNU:e1aE_N_:U2O1000000O10O1SOXFZNh9g1]O`1M4L3M3M3M3M3oFXL^8[4L4M3M4L2N1O101N2N2O2M2N2O1N2N2O1N2OO0UOUIZKm6e4j0WOkGbLU8W3UHeLk7W3]HfLc7X3bHdL^7[3gHbLY7\\3Y1011N101N2O1O1N101N2O1N2O0O2N2_IQLf3R4nK\\Lm3g3kJbMR5_2lInKi0h1W5d2eJ`MY5b2`JdM_5^2XJjMh5k4O1O100OGXJTIh5d6a0TOiI`J^6\\5g0M4L3O2N3N3M2N2N7I;E5H6G9M3N2M2N3N2ROhFiMZ9V2SG\\Mn8e2j00O10fMfEg1Z:ZNhEc1Y:\\NkE5n0_OX9;mE1a;NbDMe;N^DMb<KR]h2"}, "label": "a dark brown horse with a white forehead and stripe down its nose between another dark brown horse and a light brown horse"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right most horse\"."}], "task": "refering", "answer_template": "The \"the right most horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 315, 318, 337, 338, 341], "bbox": [0.680375, 0.4701179245283019, 0.9786406249999999, 1.0], "iscrowd": 0, "area": 17076.369150000006, "rle": {"size": [424, 640], "counts": "^\\d51U=9G9F:hDGQ9a0eFDX9c0_FBn0TOh6a1RH@a0RO\\O0o7e1kG^Oe0QO]O0R8Q3[HTM_OOV8n2VHhMi7Z2oGmMQ8e3O100O2O0O10[OPHPLQ8P4THjKm7V4c0O0=D2M3iGYKi7R5010XNUHSNj7m1eHeM[7Z2nH]MR7c2XITMh6j2_IQMa6n2bIPM^6o2eInL\\6P3gIoLY6P3jInLV6P3nInLR6P3RJmLo5Q3UJmLj5S3XJlLh5R3X2N2L4K5K5N2O1O0O`0gNjDCY;8mDDU;7QEFR;4UEGo:4WEHl:2[EJh:1]ELXm40e^K6J5IEZCa0b<6L4K4M4H9I:E8I3L4L4L4L4L4O10000O1000000O2O:F9G9G9F9H6J6Jm0RO1000000O1000000O100000CbHPK]7l4jHSKT7i4RIVKm6g4YIWKg6d4_I[K`6a4U1K5K6J5K5N200O2O000O101O0O10000O10000O2O000O100bNbFSO^9l0TGZNU9e1X1N1O1O1O1O100O1O1O11O1O1O1O1O1N2OXNnD\\1^;101O2N2N2N2N2M3N2M2M2O1O001ORNWE_1h:^N]Ea1b:\\NdEb1[:[NjEd1U:XNRFf1i:O2M2O1hN_Dg0c;SOdDj0];ROhDl0m;L3M4K4M4L3M3MTa5"}, "label": "the right most horse"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"furthest right horse , don ' t include extra part of rider\"."}], "task": "refering", "answer_template": "The \"furthest right horse , don ' t include extra part of rider\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 315, 318, 337, 338, 341], "bbox": [0.680375, 0.4701179245283019, 0.9786406249999999, 1.0], "iscrowd": 0, "area": 17076.369150000006, "rle": {"size": [424, 640], "counts": "^\\d51U=9G9F:hDGQ9a0eFDX9c0_FBn0TOh6a1RH@a0RO\\O0o7e1kG^Oe0QO]O0R8Q3[HTM_OOV8n2VHhMi7Z2oGmMQ8e3O100O2O0O10[OPHPLQ8P4THjKm7V4c0O0=D2M3iGYKi7R5010XNUHSNj7m1eHeM[7Z2nH]MR7c2XITMh6j2_IQMa6n2bIPM^6o2eInL\\6P3gIoLY6P3jInLV6P3nInLR6P3RJmLo5Q3UJmLj5S3XJlLh5R3X2N2L4K5K5N2O1O0O`0gNjDCY;8mDDU;7QEFR;4UEGo:4WEHl:2[EJh:1]ELXm40e^K6J5IEZCa0b<6L4K4M4H9I:E8I3L4L4L4L4L4O10000O1000000O2O:F9G9G9F9H6J6Jm0RO1000000O1000000O100000CbHPK]7l4jHSKT7i4RIVKm6g4YIWKg6d4_I[K`6a4U1K5K6J5K5N200O2O000O101O0O10000O10000O2O000O100bNbFSO^9l0TGZNU9e1X1N1O1O1O1O100O1O1O11O1O1O1O1O1N2OXNnD\\1^;101O2N2N2N2N2M3N2M2M2O1O001ORNWE_1h:^N]Ea1b:\\NdEb1[:[NjEd1U:XNRFf1i:O2M2O1hN_Dg0c;SOdDj0];ROhDl0m;L3M4K4M4L3M3MTa5"}, "label": "furthest right horse , don ' t include extra part of rider"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown horse with white lower legs and white mane and nose\"."}], "task": "refering", "answer_template": "The \"a light brown horse with white lower legs and white mane and nose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 170, 171, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332], "bbox": [0.295703125, 0.4308962264150943, 0.479109375, 1.0], "iscrowd": 0, "area": 17025.95460000001, "rle": {"size": [424, 640], "counts": "Ua^26e<`0J6I7J6I7L3M2N2N1O2O3L9G7I4L4L4L4K5L4WHlLW5X3nHmL6e0k6Y3fHTMX7^401O1O100001O1O001^NaHYM_7g2gHSMZ7l2lHmLU7S3QIgLP7X3VIbLk6^3ZI\\Lf6d3bITL^6m3^11O1O1O1O1O1jGQLQ7P4iHULW7l3dHXL\\7j401O00000010O0000O2N1O1O1O1O2L3I7H8L4L5L3M3LjMVJSL@<W6W3gJYL]O8i5b3kJRLG5U5Q4UKfK1Nl4]4TKaK=K\\4d4YK\\K<2Z4c4hL`KW3a4fLbKZ3^4WLQLh3P4RLVLm3k3mK[LS4e3hK`LY4_3`KhL`4X3YKoLh4P3RKVMo4a50M4L3M3N2N2N2M2O2N2I7E;I7K5K4_L_Hf0X8XOkGa0[8\\OiG=]8AfG8`8EdG7_8GeG3_8JgG0\\8NiGKZ8OQHIS8MZHJl7N_HGi78[H^Ol7`0_2O1O1O1O1H8EkfY4"}, "label": "a light brown horse with white lower legs and white mane and nose"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown horse with a blond mane and a mine riding on it\"."}], "task": "refering", "answer_template": "The \"a light brown horse with a blond mane and a mine riding on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 170, 171, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332], "bbox": [0.295703125, 0.4308962264150943, 0.479109375, 1.0], "iscrowd": 0, "area": 17025.95460000001, "rle": {"size": [424, 640], "counts": "Ua^26e<`0J6I7J6I7L3M2N2N1O2O3L9G7I4L4L4L4K5L4WHlLW5X3nHmL6e0k6Y3fHTMX7^401O1O100001O1O001^NaHYM_7g2gHSMZ7l2lHmLU7S3QIgLP7X3VIbLk6^3ZI\\Lf6d3bITL^6m3^11O1O1O1O1O1jGQLQ7P4iHULW7l3dHXL\\7j401O00000010O0000O2N1O1O1O1O2L3I7H8L4L5L3M3LjMVJSL@<W6W3gJYL]O8i5b3kJRLG5U5Q4UKfK1Nl4]4TKaK=K\\4d4YK\\K<2Z4c4hL`KW3a4fLbKZ3^4WLQLh3P4RLVLm3k3mK[LS4e3hK`LY4_3`KhL`4X3YKoLh4P3RKVMo4a50M4L3M3N2N2N2M2O2N2I7E;I7K5K4_L_Hf0X8XOkGa0[8\\OiG=]8AfG8`8EdG7_8GeG3_8JgG0\\8NiGKZ8OQHIS8MZHJl7N_HGi78[H^Ol7`0_2O1O1O1O1H8EkfY4"}, "label": "a light brown horse with a blond mane and a mine riding on it"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown horse that is all the way to the left of three other horses\"."}], "task": "refering", "answer_template": "The \"a dark brown horse that is all the way to the left of three other horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 164, 165, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 255, 256, 257, 278, 279, 280, 301, 302, 324, 325], "bbox": [0.077234375, 0.41705188679245286, 0.218359375, 0.9887971698113207], "iscrowd": 0, "area": 12657.597749999997, "rle": {"size": [424, 640], "counts": "`od03S=2E<L3XG]Og4l0mG[O]2Kb5P2lGQNg12Z6`2bIbM\\6b2^IbMa6a2XIdMf6f2nH^MQ7o2`HTM^7V3WHmLh7Z3PHhLo7Y4N2O1N2ODZHXKf7c4`H\\K_7h4_HWK`7m4;6J7I2O0O1O100O101O0000000LE?TOgHdK^7T4o0M2ON2O2ON1M40O1lHRLo4o3mJYLo4h3gJdLV5\\3TJ]Mj5b2UJaMi5_2eIZLKZ1n5`NWJl3I^Nn5kMTJ\\6l5fIQJ[6P6>001O1F9J7J5POQ1E:I8M2N2O1N2N2N2N2O1N2N2K4K6I7J6M2O1N3L2M31OWMSF[2k9fMXFf016f9TO]F?68]9YO`F8<;V9ZOfFM`0c0\\;K6I6Ieh^6"}, "label": "a dark brown horse that is all the way to the left of three other horses"}]}
{"id": 347511, "image": "COCO_train2014_000000347511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse to the left of the american flag\"."}], "task": "refering", "answer_template": "The \"the horse to the left of the american flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 164, 165, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 255, 256, 257, 278, 279, 280, 301, 302, 324, 325], "bbox": [0.077234375, 0.41705188679245286, 0.218359375, 0.9887971698113207], "iscrowd": 0, "area": 12657.597749999997, "rle": {"size": [424, 640], "counts": "`od03S=2E<L3XG]Og4l0mG[O]2Kb5P2lGQNg12Z6`2bIbM\\6b2^IbMa6a2XIdMf6f2nH^MQ7o2`HTM^7V3WHmLh7Z3PHhLo7Y4N2O1N2ODZHXKf7c4`H\\K_7h4_HWK`7m4;6J7I2O0O1O100O101O0000000LE?TOgHdK^7T4o0M2ON2O2ON1M40O1lHRLo4o3mJYLo4h3gJdLV5\\3TJ]Mj5b2UJaMi5_2eIZLKZ1n5`NWJl3I^Nn5kMTJ\\6l5fIQJ[6P6>001O1F9J7J5POQ1E:I8M2N2O1N2N2N2N2O1N2N2K4K6I7J6M2O1N3L2M31OWMSF[2k9fMXFf016f9TO]F?68]9YO`F8<;V9ZOfFM`0c0\\;K6I6Ieh^6"}, "label": "the horse to the left of the american flag"}]}
{"id": 511358, "image": "COCO_train2014_000000511358.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"philips sonicare brush image on the mirror\"."}], "task": "refering", "answer_template": "The \"philips sonicare brush image on the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 43, 44, 66, 89, 90, 111, 112, 113, 134, 135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 249, 250, 251], "bbox": [0.850671875, 0.00020833333333333335, 0.9597343749999998, 0.6445208333333333], "iscrowd": 0, "area": 13801.390000000009, "rle": {"size": [480, 640], "counts": "cTo75h>a1_Na1_Nb1^N8H3M3M3M3N3L3M3M3M4L3N2M4M2N2M4M2M4M2N3L3N3I6C>aN`HoJm7d4_1O1O00001O000000000OfGZL]5e3bJ_L[5a3eJaLY5_3gJdLV5\\3jJgLS5Y3mJiLQ5W3oJlLn4T3QKPMl4P3TKRMk4m2UKVMh4j2XKYMe4g2[K[Mc4e2\\K_Ma4a2_KaM_4^2bKdM\\4\\2dKfMZ4Z2fKhMX4X2gKkMW4V2hKlMV4_2_KcM_4h2VKZMg4R3nJPMc4o6A?A?A?A`0@?A`0@?A`0@VQ<"}, "label": "philips sonicare brush image on the mirror"}]}
{"id": 511358, "image": "COCO_train2014_000000511358.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection of a blue electric toothbrush\"."}], "task": "refering", "answer_template": "The \"a reflection of a blue electric toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 43, 44, 66, 89, 90, 111, 112, 113, 134, 135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 249, 250, 251], "bbox": [0.850671875, 0.00020833333333333335, 0.9597343749999998, 0.6445208333333333], "iscrowd": 0, "area": 13801.390000000009, "rle": {"size": [480, 640], "counts": "cTo75h>a1_Na1_Nb1^N8H3M3M3M3N3L3M3M3M4L3N2M4M2N2M4M2M4M2N3L3N3I6C>aN`HoJm7d4_1O1O00001O000000000OfGZL]5e3bJ_L[5a3eJaLY5_3gJdLV5\\3jJgLS5Y3mJiLQ5W3oJlLn4T3QKPMl4P3TKRMk4m2UKVMh4j2XKYMe4g2[K[Mc4e2\\K_Ma4a2_KaM_4^2bKdM\\4\\2dKfMZ4Z2fKhMX4X2gKkMW4V2hKlMV4_2_KcM_4h2VKZMg4R3nJPMc4o6A?A?A?A`0@?A`0@?A`0@VQ<"}, "label": "a reflection of a blue electric toothbrush"}]}
{"id": 544127, "image": "COCO_train2014_000000544127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black baggage with a note\"."}], "task": "refering", "answer_template": "The \"a black baggage with a note\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 141, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 283, 296, 297, 298, 299, 300, 316, 317], "bbox": [0.27670833333333333, 0.334921875, 0.8005625, 0.81528125], "iscrowd": 0, "area": 54897.3974, "rle": {"size": [640, 480], "counts": "X\\c2a5^>2N1O2N1O1O2N1O2O0O2N1O1O2N1O2N1O1O2N1O2O0O1O2N1O2N1O1O2N1O2N1O1O2O0O2N1O1O2N1O2N1O2N1O1O2N101N10001O001O00001O001O000O2O001O0O101O0O2O000O2O001O0O2O000O20O01O01O01O0010O00010O01O01O01O0010O00010O01O010O000010O01O01O01O010O000010O01O01O01O010O000010O01O01O01O010O0010O00010O01O01O010O01O01O01O010O00010O0010O0010O005L;D0001N10001O00001O001O00001N10001O00001O0O2L3N2M4L3N2M4L3N2M4L3N2M4L3N3L3M3M4L3L4L5L3L4L5L3L4M4K4L5L3L4M4K4L4M4K4L4M4K4M3L5K4M4K4L4M4K4M3L5K4M3L5K4M3L5L3L5K4M3L5K4M3L5L3L4L5L3L4L5L4Kbgk1"}, "label": "a black baggage with a note"}]}
{"id": 544127, "image": "COCO_train2014_000000544127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black suitcase with the tag\"."}], "task": "refering", "answer_template": "The \"the black suitcase with the tag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 141, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 283, 296, 297, 298, 299, 300, 316, 317], "bbox": [0.27670833333333333, 0.334921875, 0.8005625, 0.81528125], "iscrowd": 0, "area": 54897.3974, "rle": {"size": [640, 480], "counts": "X\\c2a5^>2N1O2N1O1O2N1O2O0O2N1O1O2N1O2N1O1O2N1O2O0O1O2N1O2N1O1O2N1O2N1O1O2O0O2N1O1O2N1O2N1O2N1O1O2N101N10001O001O00001O001O000O2O001O0O101O0O2O000O2O001O0O2O000O20O01O01O01O0010O00010O01O01O01O0010O00010O01O010O000010O01O01O01O010O000010O01O01O01O010O000010O01O01O01O010O0010O00010O01O01O010O01O01O01O010O00010O0010O0010O005L;D0001N10001O00001O001O00001N10001O00001O0O2L3N2M4L3N2M4L3N2M4L3N2M4L3N3L3M3M4L3L4L5L3L4L5L3L4M4K4L5L3L4M4K4L4M4K4L4M4K4M3L5K4M4K4L4M4K4M3L5K4M3L5K4M3L5L3L5K4M3L5K4M3L5L3L4L5L3L4L5L4Kbgk1"}, "label": "the black suitcase with the tag"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s black tie under all the other ties he is wearing\"."}], "task": "refering", "answer_template": "The \"a man ' s black tie under all the other ties he is wearing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 145, 146, 162, 179, 196, 229, 245], "bbox": [0.43945833333333334, 0.30628125, 0.6008958333333333, 0.62690625], "iscrowd": 0, "area": 1863.9071499999995, "rle": {"size": [640, 480], "counts": "_XT41nc01O2O0O1O001O00001O00001O01O01O001O00001O00001O01O01O00001O0O2O0O01N1O1O2N1O2N1O2N1OS[42bdK9F:F:F:E;]Oc01O1O1O1O1O1O1O1O1O1OA?3M4L3M4L3M4L3J7J5Kmig3"}, "label": "a man ' s black tie under all the other ties he is wearing"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the solid black tie around the mans neck\"."}], "task": "refering", "answer_template": "The \"the solid black tie around the mans neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 145, 146, 162, 179, 196, 229, 245], "bbox": [0.43945833333333334, 0.30628125, 0.6008958333333333, 0.62690625], "iscrowd": 0, "area": 1863.9071499999995, "rle": {"size": [640, 480], "counts": "_XT41nc01O2O0O1O001O00001O00001O01O01O001O00001O00001O01O01O00001O0O2O0O01N1O1O2N1O2N1O2N1OS[42bdK9F:F:F:E;]Oc01O1O1O1O1O1O1O1O1O1OA?3M4L3M4L3M4L3J7J5Kmig3"}, "label": "the solid black tie around the mans neck"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady sitting with her husband and wearing black color dress\"."}], "task": "refering", "answer_template": "The \"a lady sitting with her husband and wearing black color dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 207, 208], "bbox": [0.0, 0.190640625, 0.34158333333333335, 0.565921875], "iscrowd": 0, "area": 24500.096699999995, "rle": {"size": [640, 480], "counts": "]73lc07J7H7I7J7H7n]OdNTa0b1f^O_NXa0i1a^OWN_a0o1[^OQNda0W2O0O101N100K5J7I6kNYMk@m2Q?]Mc@j2W?`M^@f2^?P1K6K4K5L5J5L4N201N100O2O0O101N100O2O000O101N100O2O0O101N100O2O0000001O00010O00001O001O001O001O0010O01O001O001O001O00001O010O001O0010O10O10O01000O0100O10O01000O0100O01SLbBW1]=aNkB`1U=WNRCj1o<QNUCP2j<lMYCV2g<fM\\CZ2d<bM_C`2a<\\MbCe2^<VMfCj2Z<SMhCo2X<lLlCU3S<hLoCY3R<bLRD_3m;^LUDd3k;XLXDh3i;SLZDo3e;nK^DS4b;hKaDY4_;dKdD]4^=2O0O2O1O0O2O1N2O0OM4M3M5K4K6K5K5K4K6K5K5J5L5K5K4K6K5K5J5L5K5K5J5L5K5K5J5L5K5J6KQXU6"}, "label": "a lady sitting with her husband and wearing black color dress"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding cup\"."}], "task": "refering", "answer_template": "The \"woman holding cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 207, 208], "bbox": [0.0, 0.190640625, 0.34158333333333335, 0.565921875], "iscrowd": 0, "area": 24500.096699999995, "rle": {"size": [640, 480], "counts": "]73lc07J7H7I7J7H7n]OdNTa0b1f^O_NXa0i1a^OWN_a0o1[^OQNda0W2O0O101N100K5J7I6kNYMk@m2Q?]Mc@j2W?`M^@f2^?P1K6K4K5L5J5L4N201N100O2O0O101N100O2O000O101N100O2O0O101N100O2O0000001O00010O00001O001O001O001O0010O01O001O001O001O00001O010O001O0010O10O10O01000O0100O10O01000O0100O01SLbBW1]=aNkB`1U=WNRCj1o<QNUCP2j<lMYCV2g<fM\\CZ2d<bM_C`2a<\\MbCe2^<VMfCj2Z<SMhCo2X<lLlCU3S<hLoCY3R<bLRD_3m;^LUDd3k;XLXDh3i;SLZDo3e;nK^DS4b;hKaDY4_;dKdD]4^=2O0O2O1O0O2O1N2O0OM4M3M5K4K6K5K5K4K6K5K5J5L5K5K4K6K5K5J5L5K5K5J5L5K5K5J5L5K5J6KQXU6"}, "label": "woman holding cup"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"striped tie askew on the man in the white shirt\"."}], "task": "refering", "answer_template": "The \"striped tie askew on the man in the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 129, 130, 146, 147, 163, 164, 165, 179, 180, 181, 182, 196, 197, 198, 199, 213, 214, 216, 229, 230, 231, 246, 247, 248, 263, 264], "bbox": [0.4837916666666667, 0.278125, 0.7275416666666668, 0.692859375], "iscrowd": 0, "area": 11728.1411, "rle": {"size": [640, 480], "counts": "j\\a42lc03M4K4M4K5L3M4K4M4K5L3L5L3M4K4M4K5L3M4K4M2M2O1O1N3N1N2O1O2M2O1N2O2[KjL^HW3c7jLYHX3g7jLVHW3j7kLRHX3n7iLoGX3Q8jLkGX3U8jLhGW3Y8jLdGW3^8hL_GZ3b8fL\\G[3f8eLVG^3j8bLTG_3n8_LQGb3Q9\\LmF]3iL`L[<2kF`3_9^L_Fd3b9[L\\Fh3l<1N3M2N1O00100O1O1O0PNoKTDQ4m;RLPDm3R<TLkCl3W<VLeCj3^<WL^Ci3d<YLXCf3l<ZLQCf3R=[LiBf3Y=\\LcBd3`=\\L]Bd3e=^LWBb3l=_LPBa3R>aLjA^3Z>bLcA^3`>cL\\A]3f>eLVA[3m>eLRAY3P?iLn@U3U?g04K4M4K4M4K4M4K5L3kLg_O`2^`0_Me_OZ2_`0eMf_OW2[`0hMi_O\\2o?dMT@a2e?]M`@g2Z?WMj@m2P?QMTAS3e>kL`AY3Z>eLjA]3R>aLRB]3[?M3N2M3N2M3M5L6I7J6I7J6I7I7G9H9F9G9Hkaa2"}, "label": "striped tie askew on the man in the white shirt"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"striped tie of the man\"."}], "task": "refering", "answer_template": "The \"striped tie of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 129, 130, 146, 147, 163, 164, 165, 179, 180, 181, 182, 196, 197, 198, 199, 213, 214, 216, 229, 230, 231, 246, 247, 248, 263, 264], "bbox": [0.4837916666666667, 0.278125, 0.7275416666666668, 0.692859375], "iscrowd": 0, "area": 11728.1411, "rle": {"size": [640, 480], "counts": "j\\a42lc03M4K4M4K5L3M4K4M4K5L3L5L3M4K4M4K5L3M4K4M2M2O1O1N3N1N2O1O2M2O1N2O2[KjL^HW3c7jLYHX3g7jLVHW3j7kLRHX3n7iLoGX3Q8jLkGX3U8jLhGW3Y8jLdGW3^8hL_GZ3b8fL\\G[3f8eLVG^3j8bLTG_3n8_LQGb3Q9\\LmF]3iL`L[<2kF`3_9^L_Fd3b9[L\\Fh3l<1N3M2N1O00100O1O1O0PNoKTDQ4m;RLPDm3R<TLkCl3W<VLeCj3^<WL^Ci3d<YLXCf3l<ZLQCf3R=[LiBf3Y=\\LcBd3`=\\L]Bd3e=^LWBb3l=_LPBa3R>aLjA^3Z>bLcA^3`>cL\\A]3f>eLVA[3m>eLRAY3P?iLn@U3U?g04K4M4K4M4K4M4K5L3kLg_O`2^`0_Me_OZ2_`0eMf_OW2[`0hMi_O\\2o?dMT@a2e?]M`@g2Z?WMj@m2P?QMTAS3e>kL`AY3Z>eLjA]3R>aLRB]3[?M3N2M3N2M3M5L6I7J6I7J6I7I7G9H9F9G9Hkaa2"}, "label": "striped tie of the man"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing three ties sitting down\"."}], "task": "refering", "answer_template": "The \"a man wearing three ties sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 289, 290, 291, 292, 293, 294, 295, 298, 299, 300, 301, 302, 303, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 323, 324, 325, 326, 327, 332, 333, 334, 335, 336, 337, 340, 341, 342, 343, 344, 349, 350, 351, 352, 353, 357, 358, 359, 360, 361, 366, 367, 368, 369, 370, 374, 375, 376, 377, 378, 379, 383, 384, 385, 386, 387], "bbox": [0.0, 0.044578125, 0.9330416666666667, 0.98815625], "iscrowd": 0, "area": 141082.2991, "rle": {"size": [640, 480], "counts": "k<7hc0?A?B>A?A?A?A?B>A?A?A?B>A?A?A?B4K1O1O1O100O1O1O100O100000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000O10O100000O100000000O100000000O100000SLVDDj;<XDBg;?[D_Oe;`0_D]Oa;c0PElNP;S1UEiNk:W1YEeNf:[1_EaNa:_1bE^N^:b1fEZNZ:e1kEWNU:i1oESNP:m1UFoMk9Q2YFkMg9T2]FiMc9W2aFeM_9[2eFaMZ9_2kF]MU9c2oFYMQ9f2TGVMl8j2WGSMi8l2\\GPMd8P3`GlL_8U3eGgL[8X3jGdLV8\\3nG`LR8_3RH^Ln7b3VHZLi7g3[HULe7j3`HRL`7n3Y4O10O010000O10O0100O10N2N1O2N1O2N1O2N2M2M4L3L5L4L3M4K4M4L3M4L4K4M4L3E<D;D=D<I601O0O2O001N2O001O0O2O1O001N101O001N2O001N110O010O10O010O10O010O10O010O010O0100O010O010O10O010O010O0100O010O2O1N3N1N2O2M2iLjHbIW7Y6WI\\Il6^6ZI`Ig6[6_IbIc6Z6bIdI`6Y6cIdI_6Y6eIeI\\6Y6nI]IU6[6jFVIm<h6SCXIo<f6QCZIQ=d6PCZIT=c6lB]IV=a6jB_IX=`4jBZMNUN[=[4kB_MJVN_=S4lBfMEWNZ>T3VBdNh>>\\AA_?@f@?U`0bNP@]1Za0O1O1O1O1O1O1O100O100O10000O10000O10000O100O10000O10000O10000O100O1000000001O1O001O1O001O1O001O001O1O001O1O1O1O2N1O1O1b@[N_<f1^C^N`<c1]CaNa<`1\\CdNb<]1\\CfNb<[1[CiNc<X1ZCmNc<T1ZCPOd<Q1YCSOe<o0WCTOh<m0VCTOj<m0SCVOl<k0QCWOo<k0mBWOS=l0hBWOW=l0eBUO[=n0]BXOb=k0TB\\Ol=g0kA_OU>d0aAD^>o201O00001O001O00001O001O001O00001O001O001O001N1M4L3M4L3M4L3M4L3M4M2M4L3M4L3M4L3M4L3M4L3M4L3M4M2M4L3M4L3M4L3M4L3M4L3M4hLY3M2N2N3M2N3M2N3L3N3M2N3M2N3M2N3M2N3L3N3M3M3M3M3M3M3M3M3M3L4M3N2M3M3M3N2M3M3M3M3N2Ma0_On0POP1YNYfc0"}, "label": "a man wearing three ties sitting down"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde man with glasses wearing three ties\"."}], "task": "refering", "answer_template": "The \"a blonde man with glasses wearing three ties\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 289, 290, 291, 292, 293, 294, 295, 298, 299, 300, 301, 302, 303, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 323, 324, 325, 326, 327, 332, 333, 334, 335, 336, 337, 340, 341, 342, 343, 344, 349, 350, 351, 352, 353, 357, 358, 359, 360, 361, 366, 367, 368, 369, 370, 374, 375, 376, 377, 378, 379, 383, 384, 385, 386, 387], "bbox": [0.0, 0.044578125, 0.9330416666666667, 0.98815625], "iscrowd": 0, "area": 141082.2991, "rle": {"size": [640, 480], "counts": "k<7hc0?A?B>A?A?A?A?B>A?A?A?B>A?A?A?B4K1O1O1O100O1O1O100O100000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000O10O100000O100000000O100000000O100000SLVDDj;<XDBg;?[D_Oe;`0_D]Oa;c0PElNP;S1UEiNk:W1YEeNf:[1_EaNa:_1bE^N^:b1fEZNZ:e1kEWNU:i1oESNP:m1UFoMk9Q2YFkMg9T2]FiMc9W2aFeM_9[2eFaMZ9_2kF]MU9c2oFYMQ9f2TGVMl8j2WGSMi8l2\\GPMd8P3`GlL_8U3eGgL[8X3jGdLV8\\3nG`LR8_3RH^Ln7b3VHZLi7g3[HULe7j3`HRL`7n3Y4O10O010000O10O0100O10N2N1O2N1O2N1O2N2M2M4L3L5L4L3M4K4M4L3M4L4K4M4L3E<D;D=D<I601O0O2O001N2O001O0O2O1O001N101O001N2O001N110O010O10O010O10O010O10O010O010O0100O010O010O10O010O010O0100O010O2O1N3N1N2O2M2iLjHbIW7Y6WI\\Il6^6ZI`Ig6[6_IbIc6Z6bIdI`6Y6cIdI_6Y6eIeI\\6Y6nI]IU6[6jFVIm<h6SCXIo<f6QCZIQ=d6PCZIT=c6lB]IV=a6jB_IX=`4jBZMNUN[=[4kB_MJVN_=S4lBfMEWNZ>T3VBdNh>>\\AA_?@f@?U`0bNP@]1Za0O1O1O1O1O1O1O100O100O10000O10000O10000O100O10000O10000O10000O100O1000000001O1O001O1O001O1O001O001O1O001O1O1O1O2N1O1O1b@[N_<f1^C^N`<c1]CaNa<`1\\CdNb<]1\\CfNb<[1[CiNc<X1ZCmNc<T1ZCPOd<Q1YCSOe<o0WCTOh<m0VCTOj<m0SCVOl<k0QCWOo<k0mBWOS=l0hBWOW=l0eBUO[=n0]BXOb=k0TB\\Ol=g0kA_OU>d0aAD^>o201O00001O001O00001O001O001O00001O001O001O001N1M4L3M4L3M4L3M4L3M4M2M4L3M4L3M4L3M4L3M4L3M4L3M4M2M4L3M4L3M4L3M4L3M4L3M4hLY3M2N2N3M2N3M2N3L3N3M2N3M2N3M2N3M2N3L3N3M3M3M3M3M3M3M3M3M3L4M3N2M3M3M3N2M3M3M3M3N2Ma0_On0POP1YNYfc0"}, "label": "a blonde man with glasses wearing three ties"}]}
{"id": 437632, "image": "COCO_train2014_000000437632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple velvet couch\"."}], "task": "refering", "answer_template": "The \"a purple velvet couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 168, 169, 185, 202, 219, 220, 236, 237, 251, 252, 253, 254, 268, 269, 270, 271, 279, 280, 285, 286, 287, 288, 296, 297, 298, 302, 303, 304, 305, 313, 314, 315, 319, 320, 321, 322, 330, 331, 332, 336, 337, 338, 339, 346, 347, 348, 349, 353, 354, 355, 356, 363, 364, 365, 366, 370, 371, 372, 373, 380, 381, 382, 383, 388, 389, 390], "bbox": [0.40625, 0.358234375, 1.0, 1.0], "iscrowd": 0, "area": 35433.7579, "rle": {"size": [640, 480], "counts": "o^j3n0Va0l1UNk1_Oa01O1O1O000000O100000000O1000001O000O100000000O100000000O10000000000000001O00000000000001O1O1O2N1O2N1O1O2N1O2N1O101N1O2N1O1O2N1O2N1O1L5J5K5K6J6J`0@=C<D=Ci0WOXmR2TOaSmM7I8G8I7I8H7I7_K[N`Fm1_9_NnEh1Q:eN[Ec1d:jN]Dj1b;bNaBo2^=j101N1O101N100O2O0O1O2O0O101N100O2N100O2O0O101N1O101N100O101PKnHoLS7Q3RIiLo6W3UIeLl6Z3YI`Lh6`3]I[Lc6d3cIVL_6i3eISL[6m3jImKW6S4RJcKP6\\4dJPK\\5P5WK\\Jj4d5iKiIX4V6VL[Ik3e6WLYIi3g6XLWIj3h6XLUIi3k6YLSIg3m6_LlHc3S7bLhH^3X7hLaHY3_7e4O00WEdHn8\\7QGmHh8R7WGXI`8h6`G`IX8`6gGjIQ8U6nGUJi7k5VH^Jb7b5^HgJZ7X5eHRKR7n4mH[Kk6f4SIdKe6[4ZIjKb6V4^IlK`6T4_IPL_6o3`ITL^6l3aIWL]6i3cIYL\\6g3bI\\L\\6d3cI`LZ6`3eIcLZ6\\3eIgLY6Y3gIiLW6W3hImLV6R3iIQMU6o2jITMT6l2hIZMW6e2eIaMY6_2bIiMW6[2eIkMT5Z3gJmLR4X4iKoKo2W;hNW1iNlC"}, "label": "a purple velvet couch"}]}
{"id": 159105, "image": "COCO_train2014_000000159105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden padded chair\"."}], "task": "refering", "answer_template": "The \"the wooden padded chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 162, 163, 179, 180, 193, 194, 195, 210, 211, 212, 227, 228, 229, 244], "bbox": [0.36372916666666666, 0.3786875, 0.6241458333333334, 0.6625156249999999], "iscrowd": 0, "area": 5296.360649999997, "rle": {"size": [640, 480], "counts": "jf]37>Mib06S]OLkb08Q]OJnb0;m\\OGPc0f0O3Lc0]Oc0^O9F1ZO`Ml_O`2S`0aMl_O`2S`0bMk_Om0ga0SOX^Ob0J[Ona04U^Oc0MXOPb05P^Oe0[b0]Oa]Oe0_b0\\O^]Of0bb0<01O000000000010O0000000001O00000001O01O00000000001O0001O000001O00000000001O01O00000001O0000000D<N3O000000M2F;E;E;E;EQX1MYhN6K3L3M4M3L3M4M2M4L4L3N3L3M4O10O2N102M4L4M2MN2O2M2O2M2100O100O100O1000104K1OO2N1O1N2O2d0[Og0YOQW`3"}, "label": "the wooden padded chair"}]}
{"id": 576896, "image": "COCO_train2014_000000576896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green bag on woman ' s shoulder\"."}], "task": "refering", "answer_template": "The \"green bag on woman ' s shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 188, 190, 191, 192, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284], "bbox": [0.184140625, 0.5358611111111111, 0.41871875000000003, 0.9887777777777779], "iscrowd": 0, "area": 10259.485050000003, "rle": {"size": [360, 640], "counts": "cjY11U;2N2N2N2O1N2XMG_J;Y5MeJ5Y5NdJ4\\5MbJ4_5J`J8`5H^J:b5E]J=c5C[J?e5AYJa0g5_OXJb0h5]OWJe0i5LdI6\\6=oHER7l0ZHVOf7V200000000000O101O000000000000001O000000000000000000001O0000000000000000000000000000001O00000000O1N2M3O1O1O1N2O1O1OdL`HR3^7oLfHn2Y7RMgHo2X7PMiHP3W7PMjHP3T7QMlHP3S7oLnHR3Q7nLoHS3P7mLPIT3o6kLRIV3l6kLUIU3j6kLUI7J\\2P7\\MWI6M]2k6]MXI32_2e6]MZI17`2^6_MZI2:^2Z6aM\\I1<]2W6aM^I1>]2S6bM_I1a0[2o5dM_I2d0Y2l5dMaI3e0X2h5fMcI1i0W2c5hMdI1m0T2^5jMeI3R1o1X5nMfI3V1l1S5QNgI2]1h1k4UNiI3b1c1c4[NkI2i1]1[4`NlI4o1W1T4eNmI3V2S1l3jNnI3\\2n0e3nNPJ4a2i0]3TOQJ4h2c0V3YORJ3n2`0o2\\OTJ4o2?m2\\OTJ5Q3>j2]OTJ6T3<g2_OTJ5W3;e2@SJ4Z3<b2@UJ3[3<`2ATJ3^3;]2CTJ2a3:[2DTJ1c3:Z2DRJ2f39X2ERJ1h39W2EPJ2j39V2EoI2m38T2EPJ2n38[2HgM7Y2HiM7V2IlM6T2InM6Q2Jf4100O2O0O101O010O0100O010O010O010O2O2M^bR4"}, "label": "green bag on woman ' s shoulder"}]}
{"id": 576896, "image": "COCO_train2014_000000576896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green color hand bag holding by a girl\"."}], "task": "refering", "answer_template": "The \"green color hand bag holding by a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 188, 190, 191, 192, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284], "bbox": [0.184140625, 0.5358611111111111, 0.41871875000000003, 0.9887777777777779], "iscrowd": 0, "area": 10259.485050000003, "rle": {"size": [360, 640], "counts": "cjY11U;2N2N2N2O1N2XMG_J;Y5MeJ5Y5NdJ4\\5MbJ4_5J`J8`5H^J:b5E]J=c5C[J?e5AYJa0g5_OXJb0h5]OWJe0i5LdI6\\6=oHER7l0ZHVOf7V200000000000O101O000000000000001O000000000000000000001O0000000000000000000000000000001O00000000O1N2M3O1O1O1N2O1O1OdL`HR3^7oLfHn2Y7RMgHo2X7PMiHP3W7PMjHP3T7QMlHP3S7oLnHR3Q7nLoHS3P7mLPIT3o6kLRIV3l6kLUIU3j6kLUI7J\\2P7\\MWI6M]2k6]MXI32_2e6]MZI17`2^6_MZI2:^2Z6aM\\I1<]2W6aM^I1>]2S6bM_I1a0[2o5dM_I2d0Y2l5dMaI3e0X2h5fMcI1i0W2c5hMdI1m0T2^5jMeI3R1o1X5nMfI3V1l1S5QNgI2]1h1k4UNiI3b1c1c4[NkI2i1]1[4`NlI4o1W1T4eNmI3V2S1l3jNnI3\\2n0e3nNPJ4a2i0]3TOQJ4h2c0V3YORJ3n2`0o2\\OTJ4o2?m2\\OTJ5Q3>j2]OTJ6T3<g2_OTJ5W3;e2@SJ4Z3<b2@UJ3[3<`2ATJ3^3;]2CTJ2a3:[2DTJ1c3:Z2DRJ2f39X2ERJ1h39W2EPJ2j39V2EoI2m38T2EPJ2n38[2HgM7Y2HiM7V2IlM6T2InM6Q2Jf4100O2O0O101O010O0100O010O010O010O2O2M^bR4"}, "label": "green color hand bag holding by a girl"}]}
{"id": 576896, "image": "COCO_train2014_000000576896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing specs and green tote enjoying kite fly\"."}], "task": "refering", "answer_template": "The \"a girl wearing specs and green tote enjoying kite fly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285], "bbox": [0.125421875, 0.3209444444444445, 0.49154687499999994, 0.9898611111111112], "iscrowd": 0, "area": 32046.110500000006, "rle": {"size": [360, 640], "counts": "P_l01V;2N3N0O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1N2O1O1O1N2F:L4_ObNTGb1k8>N2N2O1N2O1ROhMjHZ2U7hMiHY2V7hMiHY2U7jMiHW2V7jMiHW2W7iMhHX2W7jMgHW2Y7iMfHX2Z7hMfHX2Z7iMdHX2\\7hMcHY2]7hMaHY2^7hMaHY2_7gM_H[2a7fM]H[2T7YMRI?EZ2U7\\MUIW3f6PMXIP3f6SMYIm2f6VMXIj2f6YMYIg2e6]MYIc2f6_MYIa2e6cMYI]2f6S1N2N2O1N2N2O1N2O1O100O1O100O1O100O100O1O100O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1N2O1N2N2N2O1N2N2N2O1N2N2N200O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O100O100O1O100O1000000O10000000000O100000000000000000000001O000000000000001O0ZOWLmIi3R6XLnIh3Q6YLoIg3P6ZLPJf3o5ZLRJf3`4]LPLN@e3o1ZL[O`0UOBAe3n1\\LROc0_O]OAd3m1aLmMO1e0c0XOBc3l1mNbNaMBc3j1mNcNaMCd3f1lNgNaMCd3d1kNiNbMCd3b1kNjNbMBf3b1hNlNdM]Oi3e1cNnNeMXOn3h1^NnN_2o0bMQO`2m0aMROb2j0_MVOc2g0_MWOd2f0]MZOe2c0]M[Of2b0[M]Oh2>\\M@g2<[MDh27[MHk20XM0o2GRM;S3\\OPMd0i62O0O2O0O2OO010O00100O2N2O1N2N2O1N1O2N2O1N3M2O2MYoa3"}, "label": "a girl wearing specs and green tote enjoying kite fly"}]}
{"id": 576896, "image": "COCO_train2014_000000576896.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with black hair with bag on her shoulder\"."}], "task": "refering", "answer_template": "The \"a woman with black hair with bag on her shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285], "bbox": [0.125421875, 0.3209444444444445, 0.49154687499999994, 0.9898611111111112], "iscrowd": 0, "area": 32046.110500000006, "rle": {"size": [360, 640], "counts": "P_l01V;2N3N0O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1N2O1O1O1N2F:L4_ObNTGb1k8>N2N2O1N2O1ROhMjHZ2U7hMiHY2V7hMiHY2U7jMiHW2V7jMiHW2W7iMhHX2W7jMgHW2Y7iMfHX2Z7hMfHX2Z7iMdHX2\\7hMcHY2]7hMaHY2^7hMaHY2_7gM_H[2a7fM]H[2T7YMRI?EZ2U7\\MUIW3f6PMXIP3f6SMYIm2f6VMXIj2f6YMYIg2e6]MYIc2f6_MYIa2e6cMYI]2f6S1N2N2O1N2N2O1N2O1O100O1O100O1O100O100O1O100O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1N2O1N2N2N2O1N2N2N2O1N2N2N200O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O100O100O1O100O1000000O10000000000O100000000000000000000001O000000000000001O0ZOWLmIi3R6XLnIh3Q6YLoIg3P6ZLPJf3o5ZLRJf3`4]LPLN@e3o1ZL[O`0UOBAe3n1\\LROc0_O]OAd3m1aLmMO1e0c0XOBc3l1mNbNaMBc3j1mNcNaMCd3f1lNgNaMCd3d1kNiNbMCd3b1kNjNbMBf3b1hNlNdM]Oi3e1cNnNeMXOn3h1^NnN_2o0bMQO`2m0aMROb2j0_MVOc2g0_MWOd2f0]MZOe2c0]M[Of2b0[M]Oh2>\\M@g2<[MDh27[MHk20XM0o2GRM;S3\\OPMd0i62O0O2O0O2OO010O00100O2N2O1N2N2O1N1O2N2O1N3M2O2MYoa3"}, "label": "a woman with black hair with bag on her shoulder"}]}
{"id": 560517, "image": "COCO_train2014_000000560517.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bowl with the word nnizhigu on it\"."}], "task": "refering", "answer_template": "The \"a wooden bowl with the word nnizhigu on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.6897968750000001, 0.04044392523364486, 1.0, 0.8022429906542057], "iscrowd": 0, "area": 52747.946099999994, "rle": {"size": [428, 640], "counts": "__h51Q=;E;E:F;D;F;H7N3M3M2N3M2oE_MV9c2eFgMS9\\2hFoMP9R2lFXNm8j1mFaNl8V3I8H5K5K5K5K5K5J6K5K5K5K5K5K5K5K101N1O2O0O2N1O2O0O2N101N1O2N101N1O2O0O2N1O2O0O2N101N2N1O2O0O2N101N1O2N101N101O001N101O001N101O001N101O1N101O000O2O00000O2O00000O2O00000O101O0O10001O0O10001O0O1000001N1000001N1000001N1000000O2O00000O2O00000O2O00000O101O000O1010O00000001O000000001O000000010O000000001O0000001O000000010O000000001O000000001O0001O0001O0000001O000002N1cId5"}, "label": "a wooden bowl with the word nnizhigu on it"}]}
{"id": 560517, "image": "COCO_train2014_000000560517.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bamboo steamer basket\"."}], "task": "refering", "answer_template": "The \"the bamboo steamer basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.6897968750000001, 0.04044392523364486, 1.0, 0.8022429906542057], "iscrowd": 0, "area": 52747.946099999994, "rle": {"size": [428, 640], "counts": "__h51Q=;E;E:F;D;F;H7N3M3M2N3M2oE_MV9c2eFgMS9\\2hFoMP9R2lFXNm8j1mFaNl8V3I8H5K5K5K5K5K5J6K5K5K5K5K5K5K5K101N1O2O0O2N1O2O0O2N101N1O2N101N1O2O0O2N1O2O0O2N101N2N1O2O0O2N101N1O2N101N101O001N101O001N101O001N101O1N101O000O2O00000O2O00000O2O00000O101O0O10001O0O10001O0O1000001N1000001N1000001N1000000O2O00000O2O00000O2O00000O101O000O1010O00000001O000000001O000000010O000000001O0000001O000000010O000000001O000000001O0001O0001O0000001O000002N1cId5"}, "label": "the bamboo steamer basket"}]}
{"id": 560517, "image": "COCO_train2014_000000560517.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate of green vegetables\"."}], "task": "refering", "answer_template": "The \"a plate of green vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309], "bbox": [0.16459375, 0.1155373831775701, 0.77096875, 0.8927570093457943], "iscrowd": 0, "area": 107015.22945, "rle": {"size": [428, 640], "counts": "SS\\1j4T8>B?A>B>B?A>B>B?A>C>O00001O00001O001O00001O001O000000000000000O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000O100O100O100O100O100O100O1000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000001N101N10001N101N10001N101N10001N101N10001N10O001O1O001O001O1O010O001O1O0100O10000O100O11O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O1O1N2O1O1O001N2O1O1O1O0O1000001N100000001N1000001O0O1000010O0000000010O000001O01O000001O00101N3M2N3M2O2M2N3M2N3M2O2M3M2N3M2O2M2N3M2N3M2O2M2N3M2N3N1N3M2N3M3M2O0O1O1O1O1O1O1N2O1O1O001N2O1O1O:Fb0]Oc0^Ob0^Oa0_O3M3L4M3M3M3M2M4M3M3M3M3L4M3M3M3M3L4M3M3M3M3L4K5J6J6J6J5L5JcSm1"}, "label": "a plate of green vegetables"}]}
{"id": 560517, "image": "COCO_train2014_000000560517.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green leafy vegetables\"."}], "task": "refering", "answer_template": "The \"green leafy vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309], "bbox": [0.16459375, 0.1155373831775701, 0.77096875, 0.8927570093457943], "iscrowd": 0, "area": 107015.22945, "rle": {"size": [428, 640], "counts": "SS\\1j4T8>B?A>B>B?A>B>B?A>C>O00001O00001O001O00001O001O000000000000000O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000O100O100O100O100O100O100O1000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000001N101N10001N101N10001N101N10001N101N10001N10O001O1O001O001O1O010O001O1O0100O10000O100O11O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O1O1N2O1O1O001N2O1O1O1O0O1000001N100000001N1000001O0O1000010O0000000010O000001O01O000001O00101N3M2N3M2O2M2N3M2N3M2O2M3M2N3M2O2M2N3M2N3M2O2M2N3M2N3N1N3M2N3M3M2O0O1O1O1O1O1O1N2O1O1O001N2O1O1O:Fb0]Oc0^Ob0^Oa0_O3M3L4M3M3M3M2M4M3M3M3M3L4M3M3M3M3L4M3M3M3M3L4K5J6J6J6J5L5JcSm1"}, "label": "green leafy vegetables"}]}
{"id": 28038, "image": "COCO_train2014_000000028038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch behind the person wearing headphones\"."}], "task": "refering", "answer_template": "The \"couch behind the person wearing headphones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 35, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 133, 150, 151], "bbox": [0.5271250000000001, 0.00025, 0.88565625, 0.38527083333333334], "iscrowd": 0, "area": 31407.311100000003, "rle": {"size": [480, 640], "counts": "lom4:c>R3nL?A5L4N2M3N2N2M4M2N2N2M3N2N2N3M2N1O1N101O1O1O1O0O2O1OO100O100O010O100O1O10O0100O100O100O00100O100O10O01O100O100O10O0100O1O100O010O100O100O1O100O100O100O100O1O100O01000O100O100O10000O100O100O100O10000O100O10O010000O100O100O100O10000O100O100O100O100O1O100O10O0100O100O100O1O100O100O100O100O1O100O100O100O100O010O1O100O100O100O100O1O100O1O1N2O1O1OgD_MU:f30001O0000001N1000001O000000O10000O10000O10000O10000O10000O1000O010000O10000O106J8G9H8H8G9H9G8G9H8H8G9H8G:G8H8GQSR1"}, "label": "couch behind the person wearing headphones"}]}
{"id": 28038, "image": "COCO_train2014_000000028038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ugly couch with someone sitting on it\"."}], "task": "refering", "answer_template": "The \"an ugly couch with someone sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 35, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 133, 150, 151], "bbox": [0.5271250000000001, 0.00025, 0.88565625, 0.38527083333333334], "iscrowd": 0, "area": 31407.311100000003, "rle": {"size": [480, 640], "counts": "lom4:c>R3nL?A5L4N2M3N2N2M4M2N2N2M3N2N2N3M2N1O1N101O1O1O1O0O2O1OO100O100O010O100O1O10O0100O100O100O00100O100O10O01O100O100O10O0100O1O100O010O100O100O1O100O100O100O100O1O100O01000O100O100O10000O100O100O100O10000O100O10O010000O100O100O100O10000O100O100O100O100O1O100O10O0100O100O100O1O100O100O100O100O1O100O100O100O100O010O1O100O100O100O100O1O100O1O1N2O1O1OgD_MU:f30001O0000001N1000001O000000O10000O10000O10000O10000O10000O1000O010000O10000O106J8G9H8H8G9H9G8G9H8H8G9H8G:G8H8GQSR1"}, "label": "an ugly couch with someone sitting on it"}]}
{"id": 28038, "image": "COCO_train2014_000000028038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black and yellow shirt sitting in front of black laptop\"."}], "task": "refering", "answer_template": "The \"man in black and yellow shirt sitting in front of black laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 328, 329, 330, 352, 353, 375], "bbox": [0.11825000000000001, 0.0720625, 0.5675625, 0.9729791666666666], "iscrowd": 0, "area": 50190.608649999995, "rle": {"size": [480, 640], "counts": "TfS1:d>=D5K3L3N3M3L3N3L4M3M2M4M3M2M2O2M2O2N101N101O0O2O001N101O0O100O2O000O10001N100000001O000000001O1O001VIVM\\1k2aNZM\\1g2aN]M]1c2bNaM[1`2bNeM[1\\2bNiM[1W2cNnMZ1S2cNRN[1n1cNVNZ1j1eNZNX1g1fN]NX1c1gN`NV1a1hNcNU1]1jNgNT1Y1jNjNT1W1kNlNS1S1lNQOQ1P1mNTOP1l0oNXOo0g0QO\\Ol0d0SO_Ol0`0SOCk0=TOFj0:UOIj06VOLh04WOOg01XO2g0MXO6f0JZO7e0IZO:e0EZO>d0B[Oa0c0_O\\Od0c0[O\\Oh0b0XO]Ok0b0TO]On0b0RO]OQ1a0oN^OT1`0lN_OW1`0hNXOb1f0^NYOe1e0[NZOh1e0VN\\Ol1b0TN]On1b0RN^OP2a0PN]OS2a0mM^OV2`0kM_OW2`0hM_O[2?fM@\\2>dMA_2>aM@a2?_MAc2=^MAd2?[MAf2>[M@g2?YM@h2a0XM\\Ok2c0VM[Ol2d0TM[On2e0RMYOo2g0QMWOR3i0nLUOT3j0lLUOV3j0kLSOX3m0gLRO\\3m0fLoN\\3Q1eLkN^3T1eLhN]3X1dLeN^3[1dLaN^3_1cL]Na3a1bL[N`3e1bLWN`3i1aLSNb3l1aLPNa3P2`LmMb3S2`LiMc3U2_LgMd3Y2^LcMd3\\2^LaMd3^2`L\\Mc3d2`LWMb3h2bLSMa3k2cLoL`3Q3Y4N101N1O2O0O2O0O2O001NM4K4L5L3L5K5K5L5L3lLTLoJP4n4TLnJo3P5TLlJo3Q5YLfJj3X5]LaJe3^5bLYJb3d5eLUJ]3j5jLnIX3P6oLhIT3V6TMbIn2]6k2N2N200O1O1O1O1O1O100O1O1O1O1O1O1O100O2N1O2N101N2N1O2N1O2O0O2N101N100O2N101N100O2N101N101N1O1L5I6K6J5K5M4N101N100O2O1N3N1N3N1N3fNfEjM[:S2jEiMX:U2kEiMW:T2mEhMU:V2nEhMT:T2QFgMR:W2RFcMS:[2Z1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N2N2N2N2N2N3M2N3L4M3M2N3M3KPeQ4"}, "label": "man in black and yellow shirt sitting in front of black laptop"}]}
{"id": 28038, "image": "COCO_train2014_000000028038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing headphones and a blue and yellow t - shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing headphones and a blue and yellow t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 328, 329, 330, 352, 353, 375], "bbox": [0.11825000000000001, 0.0720625, 0.5675625, 0.9729791666666666], "iscrowd": 0, "area": 50190.608649999995, "rle": {"size": [480, 640], "counts": "TfS1:d>=D5K3L3N3M3L3N3L4M3M2M4M3M2M2O2M2O2N101N101O0O2O001N101O0O100O2O000O10001N100000001O000000001O1O001VIVM\\1k2aNZM\\1g2aN]M]1c2bNaM[1`2bNeM[1\\2bNiM[1W2cNnMZ1S2cNRN[1n1cNVNZ1j1eNZNX1g1fN]NX1c1gN`NV1a1hNcNU1]1jNgNT1Y1jNjNT1W1kNlNS1S1lNQOQ1P1mNTOP1l0oNXOo0g0QO\\Ol0d0SO_Ol0`0SOCk0=TOFj0:UOIj06VOLh04WOOg01XO2g0MXO6f0JZO7e0IZO:e0EZO>d0B[Oa0c0_O\\Od0c0[O\\Oh0b0XO]Ok0b0TO]On0b0RO]OQ1a0oN^OT1`0lN_OW1`0hNXOb1f0^NYOe1e0[NZOh1e0VN\\Ol1b0TN]On1b0RN^OP2a0PN]OS2a0mM^OV2`0kM_OW2`0hM_O[2?fM@\\2>dMA_2>aM@a2?_MAc2=^MAd2?[MAf2>[M@g2?YM@h2a0XM\\Ok2c0VM[Ol2d0TM[On2e0RMYOo2g0QMWOR3i0nLUOT3j0lLUOV3j0kLSOX3m0gLRO\\3m0fLoN\\3Q1eLkN^3T1eLhN]3X1dLeN^3[1dLaN^3_1cL]Na3a1bL[N`3e1bLWN`3i1aLSNb3l1aLPNa3P2`LmMb3S2`LiMc3U2_LgMd3Y2^LcMd3\\2^LaMd3^2`L\\Mc3d2`LWMb3h2bLSMa3k2cLoL`3Q3Y4N101N1O2O0O2O0O2O001NM4K4L5L3L5K5K5L5L3lLTLoJP4n4TLnJo3P5TLlJo3Q5YLfJj3X5]LaJe3^5bLYJb3d5eLUJ]3j5jLnIX3P6oLhIT3V6TMbIn2]6k2N2N200O1O1O1O1O1O100O1O1O1O1O1O1O100O2N1O2N101N2N1O2N1O2O0O2N101N100O2N101N100O2N101N101N1O1L5I6K6J5K5M4N101N100O2O1N3N1N3N1N3fNfEjM[:S2jEiMX:U2kEiMW:T2mEhMU:V2nEhMT:T2QFgMR:W2RFcMS:[2Z1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N2N2N2N2N2N3M2N3L4M3M2N3M3KPeQ4"}, "label": "a man wearing headphones and a blue and yellow t - shirt"}]}
{"id": 28038, "image": "COCO_train2014_000000028038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with blue eyes and glasses looking to the left\"."}], "task": "refering", "answer_template": "The \"a person with blue eyes and glasses looking to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.481375, 0.37712500000000004, 1.0, 1.0], "iscrowd": 0, "area": 88070.68884999999, "rle": {"size": [480, 640], "counts": "Ye`4;b>e0[Oc0]O4K4M4L4L4cDTNa9o1_F`No8d1PGgNa8^1^GcN^8a1`GaN_8`1`GaN^8`1bGaN\\8a1bGaN\\8a1cG`N[8b1dG_NZ8b1fG_NX8c1fG_NX8c1gG^NX8c1gG^NW8d1hG]NV8d1iG^NU8d1jG]NT8e1kG\\NS8f1kGiNf7X1ZHlNa7V1^HmN^7U1aHoN[7R1cHSOX7o0gHTOU7m0kHWOP7^4N1O1O1O100O1O1O1O1O1O1O100O1O1O10000O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O100O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000002N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2A>cN^1aN^1cN^1bN]1cN]J"}, "label": "a person with blue eyes and glasses looking to the left"}]}
{"id": 28038, "image": "COCO_train2014_000000028038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with glasses looking up\"."}], "task": "refering", "answer_template": "The \"woman with glasses looking up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.481375, 0.37712500000000004, 1.0, 1.0], "iscrowd": 0, "area": 88070.68884999999, "rle": {"size": [480, 640], "counts": "Ye`4;b>e0[Oc0]O4K4M4L4L4cDTNa9o1_F`No8d1PGgNa8^1^GcN^8a1`GaN_8`1`GaN^8`1bGaN\\8a1bGaN\\8a1cG`N[8b1dG_NZ8b1fG_NX8c1fG_NX8c1gG^NX8c1gG^NW8d1hG]NV8d1iG^NU8d1jG]NT8e1kG\\NS8f1kGiNf7X1ZHlNa7V1^HmN^7U1aHoN[7R1cHSOX7o0gHTOU7m0kHWOP7^4N1O1O1O100O1O1O1O1O1O1O100O1O1O10000O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O100O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000002N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2A>cN^1aN^1cN^1bN]1cN]J"}, "label": "woman with glasses looking up"}]}
{"id": 511373, "image": "COCO_train2014_000000511373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red and yellow bus with white ribbon and bow\"."}], "task": "refering", "answer_template": "The \"red and yellow bus with white ribbon and bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329], "bbox": [0.266296875, 0.17079166666666667, 0.9016874999999999, 0.8269583333333334], "iscrowd": 0, "area": 107338.8769, "rle": {"size": [480, 640], "counts": "bQ`22W>i0WOh0XOh0E<F9G9G9G:E:G9G:F9G9F:G9G9G8H9G9F9H9G8H8H9F9I7O100O2N1O100O1O100O1O1O100O1O1000000O1000000O10000O100000O0100000O01000O10O10O1M3M2O2M3M3N1N3O1O1O010O1O01000000O1000000O1000000O1000000O10000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000001O00001O00001O00001O01O01O00001O001O0000100O001O1O1O1O1O1O00100O1O1O1O00001O0000010O00001O00001O00010O00001O00001OO10001N100O10000O100O100O2N100O1O1O1M3L4M3M3M4L3N2O100000000O10001O00000O100000000O101O0000000O100000000O2O000000000O100000001N100000001O0O101O001O0O101O00001O0O101O00001N10001O00001N10001O00001O000010O00010O0010O01O01O00001O000O2N1N2N3N1N2O2M2C=N3N1O2O0O101N100O2O0O101N100O2O000O2O7H>C>A?B>A?B=B?B>A?B=B?B>A?B^Wm0"}, "label": "red and yellow bus with white ribbon and bow"}]}
{"id": 511373, "image": "COCO_train2014_000000511373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an english bus with a plate beginning with an 8\"."}], "task": "refering", "answer_template": "The \"an english bus with a plate beginning with an 8\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329], "bbox": [0.266296875, 0.17079166666666667, 0.9016874999999999, 0.8269583333333334], "iscrowd": 0, "area": 107338.8769, "rle": {"size": [480, 640], "counts": "bQ`22W>i0WOh0XOh0E<F9G9G9G:E:G9G:F9G9F:G9G9G8H9G9F9H9G8H8H9F9I7O100O2N1O100O1O100O1O1O100O1O1000000O1000000O10000O100000O0100000O01000O10O10O1M3M2O2M3M3N1N3O1O1O010O1O01000000O1000000O1000000O1000000O10000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000001O00001O00001O00001O01O01O00001O001O0000100O001O1O1O1O1O1O00100O1O1O1O00001O0000010O00001O00001O00010O00001O00001OO10001N100O10000O100O100O2N100O1O1O1M3L4M3M3M4L3N2O100000000O10001O00000O100000000O101O0000000O100000000O2O000000000O100000001N100000001O0O101O001O0O101O00001O0O101O00001N10001O00001N10001O00001O000010O00010O0010O01O01O00001O000O2N1N2N3N1N2O2M2C=N3N1O2O0O101N100O2O0O101N100O2O000O2O7H>C>A?B>A?B=B?B>A?B=B?B>A?B^Wm0"}, "label": "an english bus with a plate beginning with an 8"}]}
{"id": 11661, "image": "COCO_train2014_000000011661.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man eating pizza\"."}], "task": "refering", "answer_template": "The \"a man eating pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.16115625, 0.125, 0.70490625, 0.9898598130841122], "iscrowd": 0, "area": 74122.46615, "rle": {"size": [428, 640], "counts": "j^[12V=4K5K5L4K5K5L4K5L4K5K5L4K5L4K5K5L3L5L4M3M3M3N2M3M3M3N2M3M3M3N2M3M3M3O1O100O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1O1O1O1O1O1O001O1O1fKZJm0g5ROfJa0[5^OQK6P5HULTOl3k0VLSOk3l0XLQOi3n0YLPOh3n0[LoNg3P1\\LmNe3R1]LlNd3S1^LjNd3U1^LhNd3V1`LfNb3Y1`LeNa3Z1aLcNa3\\1bL`N`3^1X4O1O1O1N2O1O1O1O100O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O10O100000000000000000000000000000000000000000000000000000001O001O001O1O001O001O001O001O1O001O001O001O001O1dHXNP3j1oLWNP3k1mLXNQ3i1nLXNP3k1mLXNQ3i1nLXNP3k1mLWNR3l1iLXNU3k1bL\\N\\3g1RLkNl3X1QLiNm3[1PLgNn3[1oKgNP4\\1nJoLEf1[5_1lJCR5?kJCS5a0hJCV5`0fJBY5Z4M2O2M2O1N3N1O2M2O2N1O1N3N1O2N1N3N1O1O2N1N3N1O2N1O2M2O1O2N2M3N2N2M3N2M3N2N2M3N2M3N3L3N2N2M3N2M3N2M3N2N2M3N2M3M4L3M3M3N3L4L4L5K4L4L5K4L4M3L5K4L4LPg^2"}, "label": "a man eating pizza"}]}
{"id": 11661, "image": "COCO_train2014_000000011661.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man eatting pizza\"."}], "task": "refering", "answer_template": "The \"man eatting pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.16115625, 0.125, 0.70490625, 0.9898598130841122], "iscrowd": 0, "area": 74122.46615, "rle": {"size": [428, 640], "counts": "j^[12V=4K5K5L4K5K5L4K5L4K5K5L4K5L4K5K5L3L5L4M3M3M3N2M3M3M3N2M3M3M3N2M3M3M3O1O100O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1O1O1O1O1O1O001O1O1fKZJm0g5ROfJa0[5^OQK6P5HULTOl3k0VLSOk3l0XLQOi3n0YLPOh3n0[LoNg3P1\\LmNe3R1]LlNd3S1^LjNd3U1^LhNd3V1`LfNb3Y1`LeNa3Z1aLcNa3\\1bL`N`3^1X4O1O1O1N2O1O1O1O100O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O10O100000000000000000000000000000000000000000000000000000001O001O001O1O001O001O001O001O1O001O001O001O001O1dHXNP3j1oLWNP3k1mLXNQ3i1nLXNP3k1mLXNQ3i1nLXNP3k1mLWNR3l1iLXNU3k1bL\\N\\3g1RLkNl3X1QLiNm3[1PLgNn3[1oKgNP4\\1nJoLEf1[5_1lJCR5?kJCS5a0hJCV5`0fJBY5Z4M2O2M2O1N3N1O2M2O2N1O1N3N1O2N1N3N1O1O2N1N3N1O2N1O2M2O1O2N2M3N2N2M3N2M3N2N2M3N2M3N3L3N2N2M3N2M3N2M3N2N2M3N2M3M4L3M3M3N3L4L4L5K4L4L5K4L4M3L5K4L4LPg^2"}, "label": "man eatting pizza"}]}
{"id": 52626, "image": "COCO_train2014_000000052626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partial view of a white hand holding a burger\"."}], "task": "refering", "answer_template": "The \"a partial view of a white hand holding a burger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 256, 278, 279, 300, 301, 302, 303, 323, 324, 325, 326, 346, 347, 348, 349, 369, 370, 371, 372], "bbox": [0.065734375, 0.6337083333333333, 0.21909375, 0.9887708333333334], "iscrowd": 0, "area": 10076.956499999998, "rle": {"size": [480, 640], "counts": "lSd04g>8H8G9H8H8H8H7I8E9F9F:G9G:J5N2N2N2N2N2N2N2N2N2N2N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2002N3M2N3M2N3M2N3N1N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N5K9G:F8H3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3N1N3M3M3L4L4L\\mY7"}, "label": "a partial view of a white hand holding a burger"}]}
{"id": 52626, "image": "COCO_train2014_000000052626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a thumb\"."}], "task": "refering", "answer_template": "The \"a thumb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 256, 278, 279, 300, 301, 302, 303, 323, 324, 325, 326, 346, 347, 348, 349, 369, 370, 371, 372], "bbox": [0.065734375, 0.6337083333333333, 0.21909375, 0.9887708333333334], "iscrowd": 0, "area": 10076.956499999998, "rle": {"size": [480, 640], "counts": "lSd04g>8H8G9H8H8H8H7I8E9F9F:G9G:J5N2N2N2N2N2N2N2N2N2N2N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2002N3M2N3M2N3M2N3N1N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N5K9G:F8H3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3N1N3M3M3L4L4L\\mY7"}, "label": "a thumb"}]}
{"id": 52626, "image": "COCO_train2014_000000052626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray pair of american eagle jeans\"."}], "task": "refering", "answer_template": "The \"a gray pair of american eagle jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300], "bbox": [0.0, 0.0, 0.151, 0.8113333333333334], "iscrowd": 0, "area": 27691.1196, "rle": {"size": [480, 640], "counts": "P3\\6c7m2PMT1UOl0O1O001O00100O001O001O1O001O1O010O001O1O001O1O001O00100O001O001O1O001O1O010O001O1O001O1O001O00L5L3L4M3L4L5`J]L^Oh3`0]LWOi3f0]LQOh3n0]LjNh3S1]LeNi3X1]L`Nh3^1]LWMj4g2\\KTMf4i2`KSMa4k2cKQM^4m2hKoLY4o2lKlLV4Q3PLkLQ4R3TLjLm3U3XLfLj3W3\\LeLe3X3aLcLa3[3dLaL\\3]3iL_LY3^3mL^LT3`3QM[LQ3b3UMZLk2e3YMWLi2f3]MULe2h3aMTL_2k3]4M3M3N1N3N2M3M2O2I7_Oa0_O`0@a0_Oa0_Oa0_OS_n7"}, "label": "a gray pair of american eagle jeans"}]}
{"id": 52626, "image": "COCO_train2014_000000052626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing khaki shorts with a blue eagle on the back side\"."}], "task": "refering", "answer_template": "The \"a person wearing khaki shorts with a blue eagle on the back side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300], "bbox": [0.0, 0.0, 0.151, 0.8113333333333334], "iscrowd": 0, "area": 27691.1196, "rle": {"size": [480, 640], "counts": "P3\\6c7m2PMT1UOl0O1O001O00100O001O001O1O001O1O010O001O1O001O1O001O00100O001O001O1O001O1O010O001O1O001O1O001O00L5L3L4M3L4L5`J]L^Oh3`0]LWOi3f0]LQOh3n0]LjNh3S1]LeNi3X1]L`Nh3^1]LWMj4g2\\KTMf4i2`KSMa4k2cKQM^4m2hKoLY4o2lKlLV4Q3PLkLQ4R3TLjLm3U3XLfLj3W3\\LeLe3X3aLcLa3[3dLaL\\3]3iL_LY3^3mL^LT3`3QM[LQ3b3UMZLk2e3YMWLi2f3]MULe2h3aMTL_2k3]4M3M3N1N3N2M3M2O2I7_Oa0_O`0@a0_Oa0_Oa0_OS_n7"}, "label": "a person wearing khaki shorts with a blue eagle on the back side"}]}
{"id": 126355, "image": "COCO_train2014_000000126355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black laptop in the left of the man\"."}], "task": "refering", "answer_template": "The \"black laptop in the left of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 204, 205, 206, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273], "bbox": [0.714078125, 0.5120093457943925, 1.0, 0.7927570093457943], "iscrowd": 0, "area": 6166.09835, "rle": {"size": [428, 640], "counts": "VZo54X=4L1O0O2O1O001N10001N1000000O1000001N10000O100000010O00000001O4LVX1BRhN=I6TCB^<f001O000O10N2M3O1N2O0N3L4M3M3L3MZc43e\\K4mBLh<`0M2N3M6J000000O10O10000000O1000000O1000000O1000O0N3N2N1N3N1N3N1101NIiN\\DV1a;nN_DQ1^;ROcDl0\\;WOdDh0Y;\\OgDb0W;AjD>S;FmD8Q;LoDMJ[OT;f0ZEIF_On:h0dE@e:c0_EUOb:m0m01N3M2N2N3M3N200000000O10000000M3M3L3N31O1lCQOi;Y10000O2O00000000001N10001O1O6J=C?@1000000000O10O1000N2L4L4M2M4M3M3M3M3M3L3M4K5K5L4K5K5K4L5K5K5LYI"}, "label": "black laptop in the left of the man"}]}
{"id": 126355, "image": "COCO_train2014_000000126355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop with stickers on it and a men with spects using the laptop\"."}], "task": "refering", "answer_template": "The \"a laptop with stickers on it and a men with spects using the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288], "bbox": [0.28103125, 0.3209813084112149, 0.6852812500000001, 0.8603037383177571], "iscrowd": 0, "area": 37204.32689999999, "rle": {"size": [428, 640], "counts": "o`[23W=4M3N2M3N2N2N2N2M101O001N101O001O0O2O001O0O2O001O001N101O001O0O2O001O0O2O001O001N101O001O0O2O001O0O2O00001O0O2O001O0O2O001O001N101O001O0O2O001O0O2O001O001N101O001N101O001O0O2O001O001N101O0C>WOh0WOj0WOh0WOj0WOh0H9O0001O0O10000O10000O1000000O10000O10000O10000O10000O1000001N10000O10000O10000O10000O1000000O10000O10000O10000O101O0O1000000O10000O10000O10000O10000O1000000O10000O10001N10000O10000O1000000O10000O10000O10000O10000O1000000O2O000O10000O10000O10000O1000000O10000O10000O10000O10000O2O00000O10000O10000O100@`0YOg0YOg0YOg0YOg0YOg0YOg0YOdlc2"}, "label": "a laptop with stickers on it and a men with spects using the laptop"}]}
{"id": 126355, "image": "COCO_train2014_000000126355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop that the man is holding\"."}], "task": "refering", "answer_template": "The \"the laptop that the man is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288], "bbox": [0.28103125, 0.3209813084112149, 0.6852812500000001, 0.8603037383177571], "iscrowd": 0, "area": 37204.32689999999, "rle": {"size": [428, 640], "counts": "o`[23W=4M3N2M3N2N2N2N2M101O001N101O001O0O2O001O0O2O001O001N101O001O0O2O001O0O2O001O001N101O001O0O2O001O0O2O00001O0O2O001O0O2O001O001N101O001O0O2O001O0O2O001O001N101O001N101O001O0O2O001O001N101O0C>WOh0WOj0WOh0WOj0WOh0H9O0001O0O10000O10000O1000000O10000O10000O10000O10000O1000001N10000O10000O10000O10000O1000000O10000O10000O10000O101O0O1000000O10000O10000O10000O10000O1000000O10000O10001N10000O10000O1000000O10000O10000O10000O10000O1000000O2O000O10000O10000O10000O1000000O10000O10000O10000O10000O2O00000O10000O10000O100@`0YOg0YOg0YOg0YOg0YOg0YOg0YOdlc2"}, "label": "the laptop that the man is holding"}]}
{"id": 142741, "image": "COCO_train2014_000000142741.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train cars\"."}], "task": "refering", "answer_template": "The \"train cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386, 403, 404, 405, 406, 407, 408, 409, 426, 427, 428, 429, 430, 431, 432, 449, 450, 451, 452, 453, 454, 473, 474, 475, 476, 477], "bbox": [0.53828125, 0.325453125, 0.790546875, 0.8952656250000001], "iscrowd": 0, "area": 41400.92045, "rle": {"size": [640, 640], "counts": "TSh6?Uc0a0^Ob0_O`0@a0\\Od0XOg0ZOg0XOh0XOg0YOh0YOe0ZOc0]Od0\\Od0]Oc0\\Od0I8N100O2N1O1O2N100O2N1O1O2N100O0000000001O000000000O2O00000000001O00002N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2N2N2L4M3M3L4M3L4M3M3L4M3L4M3M3L4M3L4M3M3L4M3M3L4M3L5L4L4K5L4K5L4L4K4M4K5L4L4ZNh[c2"}, "label": "train cars"}]}
{"id": 142741, "image": "COCO_train2014_000000142741.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gray commuter train that has a white door at the rear\"."}], "task": "refering", "answer_template": "The \"gray commuter train that has a white door at the rear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386, 403, 404, 405, 406, 407, 408, 409, 426, 427, 428, 429, 430, 431, 432, 449, 450, 451, 452, 453, 454, 473, 474, 475, 476, 477], "bbox": [0.53828125, 0.325453125, 0.790546875, 0.8952656250000001], "iscrowd": 0, "area": 41400.92045, "rle": {"size": [640, 640], "counts": "TSh6?Uc0a0^Ob0_O`0@a0\\Od0XOg0ZOg0XOh0XOg0YOh0YOe0ZOc0]Od0\\Od0]Oc0\\Od0I8N100O2N1O1O2N100O2N1O1O2N100O0000000001O000000000O2O00000000001O00002N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2N2N2L4M3M3L4M3L4M3M3L4M3L4M3M3L4M3L4M3M3L4M3M3L4M3L5L4L4K5L4K5L4L4K4M4K5L4L4ZNh[c2"}, "label": "gray commuter train that has a white door at the rear"}]}
{"id": 118169, "image": "COCO_train2014_000000118169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green shirt is playing wii , holding the controllers to her head\"."}], "task": "refering", "answer_template": "The \"a woman in a green shirt is playing wii , holding the controllers to her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 40, 58, 59, 60, 61, 62, 63, 64, 80, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390, 406, 407, 408, 409, 410, 411, 412, 413, 430, 431, 432, 433, 434, 435, 436, 453, 454, 455, 456, 457, 458, 459], "bbox": [0.4413125, 0.0242831541218638, 0.994578125, 0.9847670250896058], "iscrowd": 0, "area": 127922.67029999998, "rle": {"size": [558, 640], "counts": "_bj43Xa04M2M4L3N2M4L3N2M4L3N3L3M3N3L3N2M4L3N2M3M3N2M3M4M2M3M3N2M3M3N2M3M4M2M3M3N2bJeLfL^3W3hLcL[3[3kL_LW3_3oLZLU3c3QMWLP3g3WMRLl2l3ZMmKh2R4^MhKd2V4aMdKa2Z4bMcKa2[4bMbK`2\\4bMaK`2^4cM^K`2_4cM_K^2`4eM\\K^2b4dM[K^2d4eMYK]2d4fMYK\\2f4gMTK^2j4dMoJb2o4bMiJe2U5]MeJh2[5ZM]Jm2b5TMWJR3i5PMoIW3P6jLlIY3T6iLiIY3V6hLgIZ3Y6hLdIZ3[6hLaIZ3_6gL_I[3`6gL\\I[3d6fLZI\\3e6fLXI[3i6eLTI^3k6dLRI]3n6dLoH_3P7_301O1O001O001O001O0010O01O001O001O001O0010O100O10O0100O100O10000O100O010O10000000000000O1000000000000000000O100000005K9G8H9G9F:G4L4L4L3M4L4L3M4L4L3N3L4M2M4M3L3N3L4M2M4L4M2M4M3L4M2M4M3L3N3L4M2M4M3L3N3L4L3N2M01O001O001O001O001O001O00001O010O001O001O001O001O001O00001O001O001O001O002N1O2N1O2N1\\CnMj9T2RFnMn9S2PFmMQ:T2lEnMT:T2hEnMX:S2fEnM[:S2aEoM_:R2_EnMb:T2ZEnMf:S2WEoMi:S2TEnMl:S2QEoMo:S2nDmMS;T2jDnMV;T2_DUNa;l1TD^Nl;d1hCfNX<[1]CnNd<n2000O1O1O1O1O100O1O1O1O1O100O1O1O1O1O2N100O1O1O1O1O100O1O1O1O1O100O1O1O1001O00001O0000001O00001O000010O0001O00001O00001O0000001O00001O00001O00001O00001O0000001O00010O001O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2Nb3_Ln3QLo3QLaS1"}, "label": "a woman in a green shirt is playing wii , holding the controllers to her head"}]}
{"id": 118169, "image": "COCO_train2014_000000118169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling woman in a green shirt with her hands on her head\"."}], "task": "refering", "answer_template": "The \"a smiling woman in a green shirt with her hands on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 40, 58, 59, 60, 61, 62, 63, 64, 80, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390, 406, 407, 408, 409, 410, 411, 412, 413, 430, 431, 432, 433, 434, 435, 436, 453, 454, 455, 456, 457, 458, 459], "bbox": [0.4413125, 0.0242831541218638, 0.994578125, 0.9847670250896058], "iscrowd": 0, "area": 127922.67029999998, "rle": {"size": [558, 640], "counts": "_bj43Xa04M2M4L3N2M4L3N2M4L3N3L3M3N3L3N2M4L3N2M3M3N2M3M4M2M3M3N2M3M3N2M3M4M2M3M3N2bJeLfL^3W3hLcL[3[3kL_LW3_3oLZLU3c3QMWLP3g3WMRLl2l3ZMmKh2R4^MhKd2V4aMdKa2Z4bMcKa2[4bMbK`2\\4bMaK`2^4cM^K`2_4cM_K^2`4eM\\K^2b4dM[K^2d4eMYK]2d4fMYK\\2f4gMTK^2j4dMoJb2o4bMiJe2U5]MeJh2[5ZM]Jm2b5TMWJR3i5PMoIW3P6jLlIY3T6iLiIY3V6hLgIZ3Y6hLdIZ3[6hLaIZ3_6gL_I[3`6gL\\I[3d6fLZI\\3e6fLXI[3i6eLTI^3k6dLRI]3n6dLoH_3P7_301O1O001O001O001O0010O01O001O001O001O0010O100O10O0100O100O10000O100O010O10000000000000O1000000000000000000O100000005K9G8H9G9F:G4L4L4L3M4L4L3M4L4L3N3L4M2M4M3L3N3L4M2M4L4M2M4M3L4M2M4M3L3N3L4M2M4M3L3N3L4L3N2M01O001O001O001O001O001O00001O010O001O001O001O001O001O00001O001O001O001O002N1O2N1O2N1\\CnMj9T2RFnMn9S2PFmMQ:T2lEnMT:T2hEnMX:S2fEnM[:S2aEoM_:R2_EnMb:T2ZEnMf:S2WEoMi:S2TEnMl:S2QEoMo:S2nDmMS;T2jDnMV;T2_DUNa;l1TD^Nl;d1hCfNX<[1]CnNd<n2000O1O1O1O1O100O1O1O1O1O100O1O1O1O1O2N100O1O1O1O1O100O1O1O1O1O100O1O1O1001O00001O0000001O00001O000010O0001O00001O00001O0000001O00001O00001O00001O00001O0000001O00010O001O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2Nb3_Ln3QLo3QLaS1"}, "label": "a smiling woman in a green shirt with her hands on her head"}]}
{"id": 118169, "image": "COCO_train2014_000000118169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on the left\"."}], "task": "refering", "answer_template": "The \"woman on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 210, 211, 212, 233, 234, 235, 236, 255, 256, 257, 258, 259, 260, 277, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 352, 353, 354, 369, 370, 371, 372, 373, 374, 375, 376, 377, 392, 393, 394, 395, 396, 397, 398, 399, 400, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 439, 440, 441, 442, 443, 444, 445, 446, 447], "bbox": [0.0391875, 0.44044802867383515, 0.4604218749999999, 0.9865232974910395], "iscrowd": 0, "area": 50330.75730000002, "rle": {"size": [558, 640], "counts": "kP>5Wa0:F:F;E7cN]1J6I6K5J6K5J6J6K4K6K5J6K3L4O1O100O2O0O1O100O100O101N100O100OO101Ok0UO6IO2N2N2N1O2N2N2N1O2O1N2N1O2N2N2N1OO\\MmKPHS4g7^LQHb3g7hLWHW3k6jMSIV2n6jMQIU2P7mMmHT2T7mMjHR2W7PNfHQ2[7PNcHo1^7RNaHn1`7SN]Hm1d7UNZHk1g7VNWHk1h7WNUHk1k7VNRHm1l7TNRHo1m7RNPHP2o7RNnGQ2Q8PNkGT2S8nMjGU2U8kMiGW2W8jMfGY2Y8_3O00100O001O1O0001O00000000001O00000000001O1O1O1O001O1O1O1O1O1O001O1O001O1O001O1O1O001O1O001N2O001O1O001O1O001O1O001O1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N3M2N2N2N2N2N1O2N1O1O2N1O2N1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O002N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N5K8H8H8H4L4L3M3M4L3M4L3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M2N3M3M3M3M3M3L3N2N2N2N2N1O2N2N2N2N1O2NbPl5"}, "label": "woman on the left"}]}
{"id": 118169, "image": "COCO_train2014_000000118169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with long brown hair and a silver watch\"."}], "task": "refering", "answer_template": "The \"a woman with long brown hair and a silver watch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 210, 211, 212, 233, 234, 235, 236, 255, 256, 257, 258, 259, 260, 277, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 352, 353, 354, 369, 370, 371, 372, 373, 374, 375, 376, 377, 392, 393, 394, 395, 396, 397, 398, 399, 400, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 439, 440, 441, 442, 443, 444, 445, 446, 447], "bbox": [0.0391875, 0.44044802867383515, 0.4604218749999999, 0.9865232974910395], "iscrowd": 0, "area": 50330.75730000002, "rle": {"size": [558, 640], "counts": "kP>5Wa0:F:F;E7cN]1J6I6K5J6K5J6J6K4K6K5J6K3L4O1O100O2O0O1O100O100O101N100O100OO101Ok0UO6IO2N2N2N1O2N2N2N1O2O1N2N1O2N2N2N1OO\\MmKPHS4g7^LQHb3g7hLWHW3k6jMSIV2n6jMQIU2P7mMmHT2T7mMjHR2W7PNfHQ2[7PNcHo1^7RNaHn1`7SN]Hm1d7UNZHk1g7VNWHk1h7WNUHk1k7VNRHm1l7TNRHo1m7RNPHP2o7RNnGQ2Q8PNkGT2S8nMjGU2U8kMiGW2W8jMfGY2Y8_3O00100O001O1O0001O00000000001O00000000001O1O1O1O001O1O1O1O1O1O001O1O001O1O001O1O1O001O1O001N2O001O1O001O1O001O1O001O1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N3M2N2N2N2N2N1O2N1O1O2N1O2N1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O002N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N5K8H8H8H4L4L3M3M4L3M4L3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M2N3M3M3M3M3M3L3N2N2N2N2N1O2N2N2N2N1O2NbPl5"}, "label": "a woman with long brown hair and a silver watch"}]}
{"id": 372121, "image": "COCO_train2014_000000372121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall glass of beer on a table\"."}], "task": "refering", "answer_template": "The \"a tall glass of beer on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173], "bbox": [0.41784375, 0.13799163179916316, 0.594125, 0.4679497907949791], "iscrowd": 0, "area": 14096.654549999997, "rle": {"size": [478, 640], "counts": "ejl32W>g0YOf0ZOf0[Oe0ZOg0YOf0H8O2M2O1N2O000000001N100000001N1000000O101O0000000010O00000001OO10001O000O100000000O101O000O10000O100O101N100O1O100O2O0O101N100O2M2N2N3C<0002O1N;F<C0001N10001O00001O0O101O00001O01O000N2N2N2O2M2N2N2N3L8H7H9H9F9H9GeWi3"}, "label": "a tall glass of beer on a table"}]}
{"id": 372121, "image": "COCO_train2014_000000372121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of beer\"."}], "task": "refering", "answer_template": "The \"a glass of beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173], "bbox": [0.41784375, 0.13799163179916316, 0.594125, 0.4679497907949791], "iscrowd": 0, "area": 14096.654549999997, "rle": {"size": [478, 640], "counts": "ejl32W>g0YOf0ZOf0[Oe0ZOg0YOf0H8O2M2O1N2O000000001N100000001N1000000O101O0000000010O00000001OO10001O000O100000000O101O000O10000O100O101N100O1O100O2O0O101N100O2M2N2N3C<0002O1N;F<C0001N10001O00001O0O101O00001O01O000N2N2N2O2M2N2N2N3L8H7H9H9F9H9GeWi3"}, "label": "a glass of beer"}]}
{"id": 372121, "image": "COCO_train2014_000000372121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with green toppings\"."}], "task": "refering", "answer_template": "The \"a pizza with green toppings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305], "bbox": [0.026859375, 0.46223849372384934, 0.483375, 0.7790794979079497], "iscrowd": 0, "area": 35380.661349999995, "rle": {"size": [478, 640], "counts": "hV8;l=j0K5L4K5L4K5L4K5L2M3M3N2M3N2M3N2M3N2N2O1O1N2O1O1N2O1O1N2O1N2O001N2O001N101O1N101O1N101O0O2O1N101O1N101O001O001O001O001O001O001O001O001O001O001O00001O001O00001O001O00001O001O00001O000000001O0000000000000000001O000000000000000000001O0000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000O10000O100O100O10000O100O101O0O101N100O2O000O2O0O101O0O101N10001N101N101N101O0O2O0O2O001N101N101O001O001O001O0O2O001O00001O001O0O2O001O0O2O001N101O001N101O0O2O00001N101O0O2O1O001N101O0O2O001O0O2N2N1O2M2O2N1N3N1O2M4M2N3L4M2N3L4M2N3M3L3K6AiWj4"}, "label": "a pizza with green toppings"}]}
{"id": 372121, "image": "COCO_train2014_000000372121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with green spinach on it loaded with cheese\"."}], "task": "refering", "answer_template": "The \"a pizza with green spinach on it loaded with cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305], "bbox": [0.026859375, 0.46223849372384934, 0.483375, 0.7790794979079497], "iscrowd": 0, "area": 35380.661349999995, "rle": {"size": [478, 640], "counts": "hV8;l=j0K5L4K5L4K5L4K5L2M3M3N2M3N2M3N2M3N2N2O1O1N2O1O1N2O1O1N2O1N2O001N2O001N101O1N101O1N101O0O2O1N101O1N101O001O001O001O001O001O001O001O001O001O001O00001O001O00001O001O00001O001O00001O000000001O0000000000000000001O000000000000000000001O0000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000O10000O100O100O10000O100O101O0O101N100O2O000O2O0O101O0O101N10001N101N101N101O0O2O0O2O001N101N101O001O001O001O0O2O001O00001O001O0O2O001O0O2O001N101O001N101O0O2O00001N101O0O2O1O001N101O0O2O001O0O2N2N1O2M2O2N1N3N1O2M4M2N3L4M2N3L4M2N3M3L3K6AiWj4"}, "label": "a pizza with green spinach on it loaded with cheese"}]}
{"id": 372121, "image": "COCO_train2014_000000372121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat that is directly behind the table\"."}], "task": "refering", "answer_template": "The \"the boat that is directly behind the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 30, 31, 32, 33, 34, 35, 36, 37, 53, 55, 56, 57, 58, 59, 60, 78, 81, 82, 83], "bbox": [0.30984375000000003, 0.016589958158995815, 0.6573437500000001, 0.21855648535564853], "iscrowd": 0, "area": 7894.896099999999, "rle": {"size": [478, 640], "counts": "dnl21l>1O001O1H71000000O1000000O10000O1000000OZO0`B0^=:ZBEe=d0TB\\Oj=l0O2N2N1O2N1O2N1O2O10O2O0O101O0O100O2O0O100H]BUOd=d0_B@`=;aBJ^=0gB2Y=GmB:R=ASC`0e=1O1O1O1O1O1N2O1O3Me0[O5K0000O10000O100O10000ROhB5Y=VOkB[1j<a0E<0J6H8H8O1001O01O00000000000000010O00000000000000010O00000000000000010O000000b0^O8H8H8H4MN100O10000O100O10001N100O10000O100O2O000O100O10000O101N10000O105J6K4L1N2O0O2O1O0O2O001O1O001O1O001O1O001N101O1O001O0000001O0000OlBYNS=f1300O100O10O0100O100O2O1N2O1N101N2O1N2O1N2O1N2O1N2N2O1N2N1O2N3N2M2N3M3N2M3M2NnTV3"}, "label": "the boat that is directly behind the table"}]}
{"id": 372121, "image": "COCO_train2014_000000372121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat in the water\"."}], "task": "refering", "answer_template": "The \"a boat in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 30, 31, 32, 33, 34, 35, 36, 37, 53, 55, 56, 57, 58, 59, 60, 78, 81, 82, 83], "bbox": [0.30984375000000003, 0.016589958158995815, 0.6573437500000001, 0.21855648535564853], "iscrowd": 0, "area": 7894.896099999999, "rle": {"size": [478, 640], "counts": "dnl21l>1O001O1H71000000O1000000O10000O1000000OZO0`B0^=:ZBEe=d0TB\\Oj=l0O2N2N1O2N1O2N1O2O10O2O0O101O0O100O2O0O100H]BUOd=d0_B@`=;aBJ^=0gB2Y=GmB:R=ASC`0e=1O1O1O1O1O1N2O1O3Me0[O5K0000O10000O100O10000ROhB5Y=VOkB[1j<a0E<0J6H8H8O1001O01O00000000000000010O00000000000000010O00000000000000010O000000b0^O8H8H8H4MN100O10000O100O10001N100O10000O100O2O000O100O10000O101N10000O105J6K4L1N2O0O2O1O0O2O001O1O001O1O001O1O001N101O1O001O0000001O0000OlBYNS=f1300O100O10O0100O100O2O1N2O1N101N2O1N2O1N2O1N2O1N2N2O1N2N1O2N3N2M2N3M3N2M3M2NnTV3"}, "label": "a boat in the water"}]}
{"id": 224667, "image": "COCO_train2014_000000224667.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young main a blue tee shirt and jeans doing a trick on a skateboard\"."}], "task": "refering", "answer_template": "The \"a young main a blue tee shirt and jeans doing a trick on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 59, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 148, 149, 150, 172], "bbox": [0.43676562499999994, 0.10440281030444964, 0.5723125, 0.5129742388758782], "iscrowd": 0, "area": 9760.936799999998, "rle": {"size": [427, 640], "counts": "Qkd36P=4K6J6N1N3N2M3M2O2N3O4M2M4L4hE]N`8f1\\G]Nb8g1YG\\Nf8g1UG]Nh8f1TG]Nj8Y2_FjM_9S3O1N2N2Nl0TO100O1O1O1O100O1O1M3N2N2N1N101O1O1O0O2O1O01000000XOVHnKk7P4[HlKe7R4aHjK_7U4fHfK\\7W4m0N2N2N2O13M]OmFRMo8m2RGVMl8g2VG]Mf8b2[GaMb8^2_GeM^8Z2cGeM^8Z2dGdM]8Z2eGfM[8Y2fGfM\\8X2eGgM]8W2dGiM]8V2dGhM^8Y2`GfMb8Z2]GfMc8Y2bGbM`8Z2jG^MX8^2QH\\MP8a2XHXMj7d2a1M3L4M3L4L4M4K?BlUb3"}, "label": "a young main a blue tee shirt and jeans doing a trick on a skateboard"}]}
{"id": 224667, "image": "COCO_train2014_000000224667.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a blue t shirt doing a trick on a skateboard\"."}], "task": "refering", "answer_template": "The \"a guy in a blue t shirt doing a trick on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 59, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 148, 149, 150, 172], "bbox": [0.43676562499999994, 0.10440281030444964, 0.5723125, 0.5129742388758782], "iscrowd": 0, "area": 9760.936799999998, "rle": {"size": [427, 640], "counts": "Qkd36P=4K6J6N1N3N2M3M2O2N3O4M2M4L4hE]N`8f1\\G]Nb8g1YG\\Nf8g1UG]Nh8f1TG]Nj8Y2_FjM_9S3O1N2N2Nl0TO100O1O1O1O100O1O1M3N2N2N1N101O1O1O0O2O1O01000000XOVHnKk7P4[HlKe7R4aHjK_7U4fHfK\\7W4m0N2N2N2O13M]OmFRMo8m2RGVMl8g2VG]Mf8b2[GaMb8^2_GeM^8Z2cGeM^8Z2dGdM]8Z2eGfM[8Y2fGfM\\8X2eGgM]8W2dGiM]8V2dGhM^8Y2`GfMb8Z2]GfMc8Y2bGbM`8Z2jG^MX8^2QH\\MP8a2XHXMj7d2a1M3L4M3L4L4M4K?BlUb3"}, "label": "a guy in a blue t shirt doing a trick on a skateboard"}]}
{"id": 175523, "image": "COCO_train2014_000000175523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white food truck\"."}], "task": "refering", "answer_template": "The \"a blue and white food truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211], "bbox": [0.0, 0.28733333333333333, 0.208890625, 0.5888749999999999], "iscrowd": 0, "area": 10563.356000000003, "rle": {"size": [480, 640], "counts": "a4X3h;l0TO000000000000000000000000OJPL^EP4b:QL]Eo3c:RL\\En3d:SL[Em3e:700001O000lMjE^OV:a0nE\\OR:c0RFZOn9e0WFVOj9j0YFoNk9c0XEeNR1LoN;j:Q1ZEeNY15a9T1ZEdNW14b9W1]2N1O1O11O00000000001O1O100O1O`0@1O1O006J7I5K1kC^Mj;c2SD_Mm;a2QDbMo;h2O1O1VDRM`;\\3K5K5K5K3NO010O001O010O00001_NSEnNl:i0dEPO]:l0nElNR:S1VFfNk9W1YFhNf9W1^FfNc9X1aFeN_9Z1dFdN\\9U1nDcNj16X9U1REaNj10[9]1Y2N1O2M201O0O2O000O2O001N101O0100O1]BPOT=Q1gBVOW=j0dB^Og0e0b:c0VEGf:;ZEBf:a0ZE\\Oh:e0WEYOj:i0VESOl:n0TEPOm:R1PEnNQ;S1lDnNV;R1hDnN];o0aDPOh;j0VDVOU<a0hC@X=d0f0YO2O2N2N1N3N1O2N2M2O2N1O2N2M3N3M2N]m\\7"}, "label": "a blue and white food truck"}]}
{"id": 175523, "image": "COCO_train2014_000000175523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red food truck\"."}], "task": "refering", "answer_template": "The \"a red food truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248], "bbox": [0.176875, 0.2494375, 0.8588437499999999, 0.6326458333333334], "iscrowd": 0, "area": 61646.042850000005, "rle": {"size": [480, 640], "counts": "TSe19f>o0ROm0SOm0ROn0SO5K1O2N2N2N1O2N1O001O1O00000000000000O1000000O1000000O10000000O1000O1000000O10000O100000000000000000000000000001O00000000000O10000000000000000000000000000000000000001O0000000000000000000O100000000000000000000000000000001O00000000000000000000000000000O10000000001O0000000000001O00001O00001O00001O0000001O00001O000O2O00001O00000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O100000001O000000000000001O00000000000000001O00000000000000000O1000000000001O00000000000000000000000000000000000000000000000O100000000000000000001O00000000000000000000000000000000000000000O10000000000000000000000000000000O1O100O100O2O101O1N101O0O2O001N101O1N101O0O2O001N2O001N101O0O2O1O0O2O0O2O0O2O1N1O2M2O2N1N3N2N1N3N1O2M2N3M2N3M3M1N2O2N1M3N2N2M4M2N2N2O2M2O1N3N1N2O2O00000O2O000O2O00001O00001O0000001O1O001O1O1O1O1O1O1O2K6KS1lNUnY1"}, "label": "a red food truck"}]}
{"id": 175523, "image": "COCO_train2014_000000175523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange food truck\"."}], "task": "refering", "answer_template": "The \"an orange food truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248], "bbox": [0.176875, 0.2494375, 0.8588437499999999, 0.6326458333333334], "iscrowd": 0, "area": 61646.042850000005, "rle": {"size": [480, 640], "counts": "TSe19f>o0ROm0SOm0ROn0SO5K1O2N2N2N1O2N1O001O1O00000000000000O1000000O1000000O10000000O1000O1000000O10000O100000000000000000000000000001O00000000000O10000000000000000000000000000000000000001O0000000000000000000O100000000000000000000000000000001O00000000000000000000000000000O10000000001O0000000000001O00001O00001O00001O0000001O00001O000O2O00001O00000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O100000001O000000000000001O00000000000000001O00000000000000000O1000000000001O00000000000000000000000000000000000000000000000O100000000000000000001O00000000000000000000000000000000000000000O10000000000000000000000000000000O1O100O100O2O101O1N101O0O2O001N101O1N101O0O2O001N2O001N101O0O2O1O0O2O0O2O0O2O1N1O2M2O2N1N3N2N1N3N1O2M2N3M2N3M3M1N2O2N1M3N2N2M4M2N2N2O2M2O1N3N1N2O2O00000O2O000O2O00001O00001O0000001O1O001O1O1O1O1O1O1O2K6KS1lNUnY1"}, "label": "an orange food truck"}]}
{"id": 544169, "image": "COCO_train2014_000000544169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a black hair and a white game remote in his hand\"."}], "task": "refering", "answer_template": "The \"a boy with a black hair and a white game remote in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 362, 363, 364, 365, 366, 379, 380, 381, 382, 383], "bbox": [0.32585416666666667, 0.434984375, 0.8490208333333332, 0.9843125], "iscrowd": 0, "area": 37094.50244999999, "rle": {"size": [640, 480], "counts": "dSR31mc02M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2N2N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N3N1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2N2N2N2O1N2N2N2cMiKoDY4n:jKQEW4m:kKQEW4l:lKREV4k:nKRET4k:oKSES4j:PLUEQ4h:RLVEP4d:VLYEm3c:XLZEh3f:ZLXEf3i:[LTEf3m:[LQEe3P;]LmDb3U;_LhDb3Y;_LeDa3U;gLiDY3Q;PMlDP3U;QMjDn2W;RMhDn2Z;QMeDo2\\;RMcDn2];RMbDo2^;QMaDP3^;QMbDn2_;RM`Do2_;SM_Dn2`;SM_Dn2`;SM_Dn2a;RM^Do2a;SM]Dn2b;SM]Dn2c;RM\\Do2d;QM\\Dn2d;TMZDm2f;SMYDn2g;RMXDo2h;d2000000000O1000O1000000000cIZDa4f;XKaDh4_;WKbDi4^;VKdDi4\\;VKeDj4[;UKfDk4Z;TKgDl4Y;SKiDl4W;SKjDm4V;RKkDn4U;QKlDo4U;PKlDo4V;oJjDo4Y;PKgDo4[;PKeDn4_;oJcDn4`;PKaDo4a;nJaDP5b;nJ_DP5e;mJ[DS5g;kJZDS5i;jJXDU5l;hJUDV5n;hJRDX5P<eJRDZ5Y=01O2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N01N1O1[Of00O1O1O010O1O10O0L4M4F92N2O1N1O2O1N2O1O0O2O1N2O1N101O2M2O1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2J6RNZ@\\Ol?a0^@QOj?k0_@hNh?V1a@]Ne?`1_1O1N2N2N2O2M3M4L3M3N3L3K5K6J5K5KmZ]1"}, "label": "a boy with a black hair and a white game remote in his hand"}]}
{"id": 544169, "image": "COCO_train2014_000000544169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with more hair playing a video game\"."}], "task": "refering", "answer_template": "The \"a man with more hair playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 362, 363, 364, 365, 366, 379, 380, 381, 382, 383], "bbox": [0.32585416666666667, 0.434984375, 0.8490208333333332, 0.9843125], "iscrowd": 0, "area": 37094.50244999999, "rle": {"size": [640, 480], "counts": "dSR31mc02M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2N2N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N3N1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2N2N2N2O1N2N2N2cMiKoDY4n:jKQEW4m:kKQEW4l:lKREV4k:nKRET4k:oKSES4j:PLUEQ4h:RLVEP4d:VLYEm3c:XLZEh3f:ZLXEf3i:[LTEf3m:[LQEe3P;]LmDb3U;_LhDb3Y;_LeDa3U;gLiDY3Q;PMlDP3U;QMjDn2W;RMhDn2Z;QMeDo2\\;RMcDn2];RMbDo2^;QMaDP3^;QMbDn2_;RM`Do2_;SM_Dn2`;SM_Dn2`;SM_Dn2a;RM^Do2a;SM]Dn2b;SM]Dn2c;RM\\Do2d;QM\\Dn2d;TMZDm2f;SMYDn2g;RMXDo2h;d2000000000O1000O1000000000cIZDa4f;XKaDh4_;WKbDi4^;VKdDi4\\;VKeDj4[;UKfDk4Z;TKgDl4Y;SKiDl4W;SKjDm4V;RKkDn4U;QKlDo4U;PKlDo4V;oJjDo4Y;PKgDo4[;PKeDn4_;oJcDn4`;PKaDo4a;nJaDP5b;nJ_DP5e;mJ[DS5g;kJZDS5i;jJXDU5l;hJUDV5n;hJRDX5P<eJRDZ5Y=01O2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N01N1O1[Of00O1O1O010O1O10O0L4M4F92N2O1N1O2O1N2O1O0O2O1N2O1N101O2M2O1N2O1N2O1O1N2O1N2O2M2O1O1N2O1N2O1N2J6RNZ@\\Ol?a0^@QOj?k0_@hNh?V1a@]Ne?`1_1O1N2N2N2O2M3M4L3M3N3L3K5K6J5K5KmZ]1"}, "label": "a man with more hair playing a video game"}]}
{"id": 544169, "image": "COCO_train2014_000000544169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with crew cut and white socks\"."}], "task": "refering", "answer_template": "The \"a man with crew cut and white socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 234, 235, 236, 251, 252, 253, 254, 268, 269, 270, 271, 285, 286, 287, 288, 301, 302, 303, 304, 318, 319, 320, 321, 334, 335, 336, 337, 338, 351, 352, 353, 354, 367, 368, 369, 370, 371, 384, 385, 386, 387, 388], "bbox": [0.6066041666666667, 0.52815625, 0.9519583333333335, 0.9876093749999999], "iscrowd": 0, "area": 22613.54535, "rle": {"size": [640, 480], "counts": "Q_f51nc05L3L3N0O101O0O100O100O100N3N1N2O2M2N2N3M2M4M2N2N3L3N3M2M3N3L3M3N3L2O2N2M3N1O2M3N1O2O1N10000O2O000O101O0O100O1O00001O0O100001O1O1N2N2M3N1N3M3N2M3M3N2M3M2O11c^OlNl>UOcAb3S>hLiA\\3P>lLjAY3Q>lLjAY3Q>lLjAX3Q>`1K4L4K4K5L5L3N2M3ZNcIfE`6S:gIiE\\6R:jIkEX6S:kIiEW6V:kIhEW6V:lIfEV6Y:lIeEU6Z:mIcEU6\\:mIbEU6\\:mIaEU6^:nI_ES6`:oI]ES6c:nI[ES6d:oIYET6f:c1O100O100O10O1000000000000000O01VNcEQJ]:g5QFSJo9k5XFPJh9n5`FlI`9S6gFgIZ9V6nFdIR9Z6VG`Ik8^6YG_Ih8^6[G`If8^6]GaId8]6^GbIc8\\6a2N3N1N3N6J6Ib0_O2N1N2O2M2O1O1N3M3L5K4L5K5K5L3L5K5K5G9E:I8H`0@h0XOg0XO^n="}, "label": "a man with crew cut and white socks"}]}
{"id": 544169, "image": "COCO_train2014_000000544169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in jeans playing the wii\"."}], "task": "refering", "answer_template": "The \"a man in jeans playing the wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 234, 235, 236, 251, 252, 253, 254, 268, 269, 270, 271, 285, 286, 287, 288, 301, 302, 303, 304, 318, 319, 320, 321, 334, 335, 336, 337, 338, 351, 352, 353, 354, 367, 368, 369, 370, 371, 384, 385, 386, 387, 388], "bbox": [0.6066041666666667, 0.52815625, 0.9519583333333335, 0.9876093749999999], "iscrowd": 0, "area": 22613.54535, "rle": {"size": [640, 480], "counts": "Q_f51nc05L3L3N0O101O0O100O100O100N3N1N2O2M2N2N3M2M4M2N2N3L3N3M2M3N3L3M3N3L2O2N2M3N1O2M3N1O2O1N10000O2O000O101O0O100O1O00001O0O100001O1O1N2N2M3N1N3M3N2M3M3N2M3M2O11c^OlNl>UOcAb3S>hLiA\\3P>lLjAY3Q>lLjAY3Q>lLjAX3Q>`1K4L4K4K5L5L3N2M3ZNcIfE`6S:gIiE\\6R:jIkEX6S:kIiEW6V:kIhEW6V:lIfEV6Y:lIeEU6Z:mIcEU6\\:mIbEU6\\:mIaEU6^:nI_ES6`:oI]ES6c:nI[ES6d:oIYET6f:c1O100O100O10O1000000000000000O01VNcEQJ]:g5QFSJo9k5XFPJh9n5`FlI`9S6gFgIZ9V6nFdIR9Z6VG`Ik8^6YG_Ih8^6[G`If8^6]GaId8]6^GbIc8\\6a2N3N1N3N6J6Ib0_O2N1N2O2M2O1O1N3M3L5K4L5K5K5L3L5K5K5G9E:I8H`0@h0XOg0XO^n="}, "label": "a man in jeans playing the wii"}]}
{"id": 257451, "image": "COCO_train2014_000000257451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 80, 81, 82, 98, 99, 100, 116, 117, 118, 119, 123, 124, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 225, 226, 227, 231, 232, 233, 243, 244, 245, 250, 251], "bbox": [0.4582, 0.205065963060686, 0.99476, 0.9848284960422163], "iscrowd": 0, "area": 43349.1023, "rle": {"size": [379, 500], "counts": "Rld22e;5K4L5UOKjE:T:h0O0O2WFDP8Z2J5J6K6eIULY4Q4YKZLd4m3mJ^LQ5g3aJdL\\5o4O20000000O010000N1M4L4L4L3M6L5M2N3M3M3M3M2N3M3L4I7J6I6K;E2M2O0OiGUMb7j2]HZMa7d2`H_M^7`2aHeM[7[2eHhMY7V2gHnMW7R2hHRNU7n1jHUNS7l1kHXNS7h1lH\\NQ7d1mH`NQ7`1nHcNo6X3M2N2N3L3J6@a0_O?B?L3N2N3N1N2]1dN3LO2N100O2N1O1O2N2N2N101N2M3K5K4M3L3M3N2M3J6H9G8I7L40000000001O0000000000010O000000000000001O00000000000010O00000000000001O0000000000000001O000000000O100000000000000O10000000O100000O0100000000O1000O10O1000000O1000O10O1000000O1000O10O1000000O1000O3N1O1O2N1O1O2N1O2N1O1O2M2O1O2N1O1O2N1O1O2N3M4K4M4L3M3M4L3M2N000000000O1000000001O00010O0000001O0000001O0000001O00001O0000010O0L4J6L5M2M3M3NZ1eNl1TNcj0"}, "label": "the zebra on the right"}]}
{"id": 257451, "image": "COCO_train2014_000000257451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra that is slightly taller\"."}], "task": "refering", "answer_template": "The \"the zebra that is slightly taller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 80, 81, 82, 98, 99, 100, 116, 117, 118, 119, 123, 124, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 225, 226, 227, 231, 232, 233, 243, 244, 245, 250, 251], "bbox": [0.4582, 0.205065963060686, 0.99476, 0.9848284960422163], "iscrowd": 0, "area": 43349.1023, "rle": {"size": [379, 500], "counts": "Rld22e;5K4L5UOKjE:T:h0O0O2WFDP8Z2J5J6K6eIULY4Q4YKZLd4m3mJ^LQ5g3aJdL\\5o4O20000000O010000N1M4L4L4L3M6L5M2N3M3M3M3M2N3M3L4I7J6I6K;E2M2O0OiGUMb7j2]HZMa7d2`H_M^7`2aHeM[7[2eHhMY7V2gHnMW7R2hHRNU7n1jHUNS7l1kHXNS7h1lH\\NQ7d1mH`NQ7`1nHcNo6X3M2N2N3L3J6@a0_O?B?L3N2N3N1N2]1dN3LO2N100O2N1O1O2N2N2N101N2M3K5K4M3L3M3N2M3J6H9G8I7L40000000001O0000000000010O000000000000001O00000000000010O00000000000001O0000000000000001O000000000O100000000000000O10000000O100000O0100000000O1000O10O1000000O1000O10O1000000O1000O10O1000000O1000O3N1O1O2N1O1O2N1O2N1O1O2M2O1O2N1O1O2N1O1O2N3M4K4M4L3M3M4L3M2N000000000O1000000001O00010O0000001O0000001O0000001O00001O0000010O0L4J6L5M2M3M3NZ1eNl1TNcj0"}, "label": "the zebra that is slightly taller"}]}
{"id": 257451, "image": "COCO_train2014_000000257451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra furthest from bush\"."}], "task": "refering", "answer_template": "The \"zebra furthest from bush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 96, 97, 98, 108, 109, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 221, 222, 234, 235, 239, 240], "bbox": [0.01534, 0.2808970976253298, 0.47353999999999996, 0.9797625329815303], "iscrowd": 0, "area": 33197.5245, "rle": {"size": [379, 500], "counts": "oT3c0V;f0ZOf0ZOf0[Od0UHoLd5g3VIiLj6R4101N2O2N2M3N3L3N2N[IlKg5Q4WJWLf5P4nIXLo5m4M3M3M2O0O00O10000O10001N10000O10000O10001N10000O10000N2M3N2N3L3N2M3N2M3N2N2M3N2M3N2M3N2N3L3N2M3N2M3N200000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000001N1000000000000O2Ob0^O>B3M4L3M3L5L3M3M4L3M00000O100000000N2N2N2M3N2M3NO001000100O2N1O1O2N1O2O0O1O2N1O1O2O0O2N1N2O2NWOhJRKW5m4TKjJk4V5_KaJ`4^5n0O01O0O5L4L5J6K5J6K4M4K5K3N2M3M3N3I6K5J6K5J6K1OO3015J5L4K5L4K5L3L5L4K5K9D=Ce0[OUVQ3"}, "label": "zebra furthest from bush"}]}
{"id": 257451, "image": "COCO_train2014_000000257451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra farthest from the tree\"."}], "task": "refering", "answer_template": "The \"the zebra farthest from the tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 96, 97, 98, 108, 109, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 221, 222, 234, 235, 239, 240], "bbox": [0.01534, 0.2808970976253298, 0.47353999999999996, 0.9797625329815303], "iscrowd": 0, "area": 33197.5245, "rle": {"size": [379, 500], "counts": "oT3c0V;f0ZOf0ZOf0[Od0UHoLd5g3VIiLj6R4101N2O2N2M3N3L3N2N[IlKg5Q4WJWLf5P4nIXLo5m4M3M3M2O0O00O10000O10001N10000O10000O10001N10000O10000N2M3N2N3L3N2M3N2M3N2N2M3N2M3N2M3N2N3L3N2M3N2M3N200000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000001N1000000000000O2Ob0^O>B3M4L3M3L5L3M3M4L3M00000O100000000N2N2N2M3N2M3NO001000100O2N1O1O2N1O2O0O1O2N1O1O2O0O2N1N2O2NWOhJRKW5m4TKjJk4V5_KaJ`4^5n0O01O0O5L4L5J6K5J6K4M4K5K3N2M3M3N3I6K5J6K5J6K1OO3015J5L4K5L4K5L3L5L4K5K9D=Ce0[OUVQ3"}, "label": "the zebra farthest from the tree"}]}
{"id": 290224, "image": "COCO_train2014_000000290224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sitting bench of a picnic table\"."}], "task": "refering", "answer_template": "The \"the sitting bench of a picnic table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 283, 306], "bbox": [0.0, 0.7330679156908665, 0.3544375, 0.8762997658079624], "iscrowd": 0, "area": 4896.210500000004, "rle": {"size": [427, 640], "counts": "]:3R=7J5oCBQ;b0oD_Ol:e0SE^Ok:b0UE_Oj:a0VEAh:?WEBi:?VE@k:`0UE@k:`0UE@k:`0TEAl:?TE@m:`0SE@m:`0REAn:`0QE@o:`0QE_OP;a0oD@Q;`0oD@Q;`0oD@Q;`0nD@S;`0mD@S;`0mD@S;a0kD@U;`0kD_OV;a0jD_OV;a0iD@W;`0gDBY;>dDE\\;;aDG`;:\\DId;7ZDKf;m00O1ZO[DFe;7`DG`;7dDG\\;7hDFY;9jDEV;;lD@W;`0j000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O1000000000000000O1000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000002N3M2N2N2N2N2N2N2N4L7I8H6J000000RO[EVOe:9lEGT:8mEHS:8mEHS:7oEHQ:7PFIP:6QFJo96QFJo95RFKn94TFKl95TFJP:2QFNf;01O00000000Qf\\5"}, "label": "the sitting bench of a picnic table"}]}
{"id": 290224, "image": "COCO_train2014_000000290224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the seat of a blue picnic table\"."}], "task": "refering", "answer_template": "The \"the seat of a blue picnic table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 283, 306], "bbox": [0.0, 0.7330679156908665, 0.3544375, 0.8762997658079624], "iscrowd": 0, "area": 4896.210500000004, "rle": {"size": [427, 640], "counts": "]:3R=7J5oCBQ;b0oD_Ol:e0SE^Ok:b0UE_Oj:a0VEAh:?WEBi:?VE@k:`0UE@k:`0UE@k:`0TEAl:?TE@m:`0SE@m:`0REAn:`0QE@o:`0QE_OP;a0oD@Q;`0oD@Q;`0oD@Q;`0nD@S;`0mD@S;`0mD@S;a0kD@U;`0kD_OV;a0jD_OV;a0iD@W;`0gDBY;>dDE\\;;aDG`;:\\DId;7ZDKf;m00O1ZO[DFe;7`DG`;7dDG\\;7hDFY;9jDEV;;lD@W;`0j000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O1000000000000000O1000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000002N3M2N2N2N2N2N2N2N4L7I8H6J000000RO[EVOe:9lEGT:8mEHS:8mEHS:7oEHQ:7PFIP:6QFJo96QFJo95RFKn94TFKl95TFJP:2QFNf;01O00000000Qf\\5"}, "label": "the seat of a blue picnic table"}]}
{"id": 290224, "image": "COCO_train2014_000000290224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark colored car parked behind the trees\"."}], "task": "refering", "answer_template": "The \"dark colored car parked behind the trees\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 107, 108, 124, 125, 126, 127, 128, 130], "bbox": [0.3958125, 0.18426229508196723, 0.709171875, 0.39775175644028105], "iscrowd": 0, "area": 7204.1343, "rle": {"size": [427, 640], "counts": "jdY31m<>D<M2M4L;G2N101N01O0100O010OO2L3N3L3N3O0O2K5J5L5K4K6KY\\31fcL4L4K5L1G82N3M201O1O1O3N`\\8AjcG7I7H8I7I7L4L4K5L4L4N2M3N2N2M4M3L4M4L3O001O00001O00001O001O00O1000000O1000000O10000H8O10000000000O100000003M3M2M4K5I6I8I6I7YOQDJU<NYe1O_[N2N2N00O1MRY6;cfI2N1O2XDCb:=[EH^:=_EIb:4]ENd:O\\E4d:J\\E;a:B_Ea0a:\\O_Eg0a:WO^Ek0c:RO^EP1b:nN]ET1Z;4L3M4L3M3O00100O10OO2000O100M3L4M3M3O2N3M4L4L3N3L4L5K7I6Ji]]2"}, "label": "dark colored car parked behind the trees"}]}
{"id": 290224, "image": "COCO_train2014_000000290224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car\"."}], "task": "refering", "answer_template": "The \"a blue car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 107, 108, 124, 125, 126, 127, 128, 130], "bbox": [0.3958125, 0.18426229508196723, 0.709171875, 0.39775175644028105], "iscrowd": 0, "area": 7204.1343, "rle": {"size": [427, 640], "counts": "jdY31m<>D<M2M4L;G2N101N01O0100O010OO2L3N3L3N3O0O2K5J5L5K4K6KY\\31fcL4L4K5L1G82N3M201O1O1O3N`\\8AjcG7I7H8I7I7L4L4K5L4L4N2M3N2N2M4M3L4M4L3O001O00001O00001O001O00O1000000O1000000O10000H8O10000000000O100000003M3M2M4K5I6I8I6I7YOQDJU<NYe1O_[N2N2N00O1MRY6;cfI2N1O2XDCb:=[EH^:=_EIb:4]ENd:O\\E4d:J\\E;a:B_Ea0a:\\O_Eg0a:WO^Ek0c:RO^EP1b:nN]ET1Z;4L3M4L3M3O00100O10OO2000O100M3L4M3M3O2N3M4L4L3N3L4L5K7I6Ji]]2"}, "label": "a blue car"}]}
{"id": 290224, "image": "COCO_train2014_000000290224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench part of a picnic table to the back of another bench\"."}], "task": "refering", "answer_template": "The \"the bench part of a picnic table to the back of another bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 256, 262, 263, 277, 278, 279, 285], "bbox": [0.072125, 0.6621779859484778, 0.47125000000000006, 0.8400702576112412], "iscrowd": 0, "area": 4626.514450000001, "rle": {"size": [427, 640], "counts": "_`c01Y=2N3M2O2M2N3N1100O100O1O1M3NO01000O01000O01000_NMiE3U:0kE0T:1kE0U:0kEOV:1jEOU:2jEOV:2iEMX:3hEMW:4hEMX:3hELY:4gELX:5gELY:5bDJk00d:?\\EAc:`0\\EAd:?[EAf:?UEFj:;TEGl:9SEGn:9REGm::SEFm::REFo::QEFo:O\\EKj:5UEER;;o00O1000000000000000O100000000000000000000000000000000000000000000000000000000000000000O10000000000000O10000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000000000000000000000000000000O1000O100000000000000000000000000000000000000000000000XDE`:;`EJ[:6eENW:2iE0U:0kE1T:OlE2S:MnE3R:MPF2o9NSF0m90UFNk92WFLi94WFKj95UFLk94UFLk94UFLk94TFMk94UFLk94TFLm94SFLm94SFLm94RFMn93RFMn93QFNo92QFMP:3oEMR:3nELS:4mELS:6jEJW:=bEB_:e0YE[Oh:l0QETOo:S1jDlNW;c11O100aNcDo0^;QOiDg0X;YOoD?R;@VE9j:G]E1d:NdEJ]:6\\1O10000O101N2O1N2O1OVe\\4"}, "label": "the bench part of a picnic table to the back of another bench"}]}
{"id": 11697, "image": "COCO_train2014_000000011697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black vehicle is parked in the corner by the road\"."}], "task": "refering", "answer_template": "The \"a black vehicle is parked in the corner by the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 353, 354, 355, 356, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.7670833333333333, 0.713578125, 1.0, 1.0], "iscrowd": 0, "area": 26406.072500000002, "rle": {"size": [640, 480], "counts": "aaV7a0fa0i1_Oa000O10000000000O0100000000<D>A>C>B9GN2N2N13N5K000000000O10000000000O1000000000000O10000000000O10000O100J6C=O11OO1O100001O0000001O0000001O00001O0000001O0000001O0000001O00000000001O000000000000O100O1O100O100O100O100O1O100"}, "label": "a black vehicle is parked in the corner by the road"}]}
{"id": 11697, "image": "COCO_train2014_000000011697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left front fender of a black truck\"."}], "task": "refering", "answer_template": "The \"the left front fender of a black truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 353, 354, 355, 356, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.7670833333333333, 0.713578125, 1.0, 1.0], "iscrowd": 0, "area": 26406.072500000002, "rle": {"size": [640, 480], "counts": "aaV7a0fa0i1_Oa000O10000000000O0100000000<D>A>C>B9GN2N2N13N5K000000000O10000000000O1000000000000O10000000000O10000O100J6C=O11OO1O100001O0000001O0000001O00001O0000001O0000001O0000001O00000000001O000000000000O100O1O100O100O100O100O1O100"}, "label": "the left front fender of a black truck"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a orange tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a orange tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 153, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 245, 246, 268, 269, 291, 292, 293, 314, 315, 316, 337, 338, 339], "bbox": [0.62528125, 0.34831249999999997, 0.766859375, 0.8674166666666667], "iscrowd": 0, "area": 13021.797050000003, "rle": {"size": [480, 640], "counts": "[gk59e>c0\\O5L4K5K5M3M3N2M3M3M3M2N3M2N3N1N2N2N3eNXNYEj1f:YNVEi1S3\\MW4n0bHh2T2\\LY5P1_Hj2R2WL_5T5ZJnJf5X5RJkJm5Z5mIgJR6_5hIbJX6b5bIaJ]6c5^I^Jb6f5XI\\Jh6h5SIYJm6k5mHXJR7m5hHTJX7P6bHQJ_7h6000O2O001N2O1O1O1UIYHZ6i7_I`H^6T8M3L4gMaGjMa8X2fG\\M\\8h2jGlLX8W3mG_LU8e3mGTLT8o3oGiKS8[4oG]KS8f4R14L3M4L3N3L3N3L3N:F6I2O1N3N1N[LoG<Q8AWH:i7C^H8b7FfH5Z7HmH3S7KTI1m6K[IOg6M`INb6MgIMZ60mIKU61SJIo53XJHj54^JFd56cJE_57iJC[57l4J6J7I6J6JVcU2"}, "label": "a man wearing a orange tie"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a red jacket and an orange tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a red jacket and an orange tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 153, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 245, 246, 268, 269, 291, 292, 293, 314, 315, 316, 337, 338, 339], "bbox": [0.62528125, 0.34831249999999997, 0.766859375, 0.8674166666666667], "iscrowd": 0, "area": 13021.797050000003, "rle": {"size": [480, 640], "counts": "[gk59e>c0\\O5L4K5K5M3M3N2M3M3M3M2N3M2N3N1N2N2N3eNXNYEj1f:YNVEi1S3\\MW4n0bHh2T2\\LY5P1_Hj2R2WL_5T5ZJnJf5X5RJkJm5Z5mIgJR6_5hIbJX6b5bIaJ]6c5^I^Jb6f5XI\\Jh6h5SIYJm6k5mHXJR7m5hHTJX7P6bHQJ_7h6000O2O001N2O1O1O1UIYHZ6i7_I`H^6T8M3L4gMaGjMa8X2fG\\M\\8h2jGlLX8W3mG_LU8e3mGTLT8o3oGiKS8[4oG]KS8f4R14L3M4L3N3L3N3L3N:F6I2O1N3N1N[LoG<Q8AWH:i7C^H8b7FfH5Z7HmH3S7KTI1m6K[IOg6M`INb6MgIMZ60mIKU61SJIo53XJHj54^JFd56cJE_57iJC[57l4J6J7I6J6JVcU2"}, "label": "a man wearing a red jacket and an orange tie"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing wearing green coat\"."}], "task": "refering", "answer_template": "The \"woman standing wearing green coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 303, 304, 326, 327, 328, 350], "bbox": [0.15914062499999998, 0.36714583333333334, 0.287015625, 0.8964374999999999], "iscrowd": 0, "area": 14129.275200000002, "rle": {"size": [480, 640], "counts": "nQ`1b0]>4UO]OQCg0k<@oBc0m<DmB?o<l0L4L4L4aDiMo9[2eEoMZ:T2ZEWNc:m1QE]Nm:l2N3M2N3M2N2O2M2N2N9Gb0^O=C<D<E5J6J5L5`HoHi6T7PIPIP7T7hHPIX7]7O0O2O00001O00001O00001N10O1O1O1O1O10000O10O10O10002M3N2N2M3QNdHVL]7b3PIXLR7^3[I]Lg6_3`I\\Lb6a3cIZL`6c3fIXL\\6d3jIXLX6e3mIWLU6e3oIYLS6d3oI[LS6b3oI\\LT6c2lJ\\MV5^2PK`MT5W2RKhMS5P2QKoMS5k1PKTNU5e1oJXNV5a0hG_Ob<6[1Ge`e6"}, "label": "woman standing wearing green coat"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in green coat\"."}], "task": "refering", "answer_template": "The \"woman in green coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 303, 304, 326, 327, 328, 350], "bbox": [0.15914062499999998, 0.36714583333333334, 0.287015625, 0.8964374999999999], "iscrowd": 0, "area": 14129.275200000002, "rle": {"size": [480, 640], "counts": "nQ`1b0]>4UO]OQCg0k<@oBc0m<DmB?o<l0L4L4L4aDiMo9[2eEoMZ:T2ZEWNc:m1QE]Nm:l2N3M2N3M2N2O2M2N2N9Gb0^O=C<D<E5J6J5L5`HoHi6T7PIPIP7T7hHPIX7]7O0O2O00001O00001O00001N10O1O1O1O1O10000O10O10O10002M3N2N2M3QNdHVL]7b3PIXLR7^3[I]Lg6_3`I\\Lb6a3cIZL`6c3fIXL\\6d3jIXLX6e3mIWLU6e3oIYLS6d3oI[LS6b3oI\\LT6c2lJ\\MV5^2PK`MT5W2RKhMS5P2QKoMS5k1PKTNU5e1oJXNV5a0hG_Ob<6[1Ge`e6"}, "label": "woman in green coat"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a panda next to a man in a red jacket\"."}], "task": "refering", "answer_template": "The \"a panda next to a man in a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 195, 196, 197, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.474296875, 0.43552083333333336, 0.65321875, 0.853875], "iscrowd": 0, "area": 17530.020700000005, "rle": {"size": [480, 640], "counts": "Qh^4X1g=9lNaNUDh1`;_N\\Dj1X;^NaDm1W;WNbDS2W;S1H9G9G8G:H;E:F8HO001N2O0000O1O100O1O1O1O100O02O0000000000000000000O10O1N2N2O1N200O10000O0101O001O1O001O1O0UJXG^5X9M2M4M3M3M2N1O001O0D=O02N2O1N2O2M2ON2O1O2M3N2N2N1N3N2N2N1N3N2N2M3N1O1O1N2O1O1O2M2O1O1N2O1O1O2M2O1J6F:F<Dk0UOj0VOlgW3"}, "label": "a panda next to a man in a red jacket"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cut out figure of panda bear chewing on bamboo beside a man with a red winter coat\"."}], "task": "refering", "answer_template": "The \"a cut out figure of panda bear chewing on bamboo beside a man with a red winter coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 195, 196, 197, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.474296875, 0.43552083333333336, 0.65321875, 0.853875], "iscrowd": 0, "area": 17530.020700000005, "rle": {"size": [480, 640], "counts": "Qh^4X1g=9lNaNUDh1`;_N\\Dj1X;^NaDm1W;WNbDS2W;S1H9G9G8G:H;E:F8HO001N2O0000O1O100O1O1O1O100O02O0000000000000000000O10O1N2N2O1N200O10000O0101O001O1O001O1O0UJXG^5X9M2M4M3M3M2N1O001O0D=O02N2O1N2O2M2ON2O1O2M3N2N2N1N3N2N2N1N3N2N2M3N1O1O1N2O1O1O2M2O1O1N2O1O1O2M2O1J6F:F<Dk0UOj0VOlgW3"}, "label": "a cut out figure of panda bear chewing on bamboo beside a man with a red winter coat"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the panda on the left\"."}], "task": "refering", "answer_template": "The \"the panda on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 191, 192, 193, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.2986875, 0.4341666666666667, 0.482703125, 0.8713958333333333], "iscrowd": 0, "area": 17871.199299999997, "rle": {"size": [480, 640], "counts": "dji25]>`0A>TNQO[E[1a:iN\\EY1`:kN^EX1a:hN]EZ1b:hN[E[1c:fN[E\\1d:fNYE]1e:dNYE_1e:bNYE`1g:aNVEb1h:_NVEc1i:^NUEe1i:[1N2N2O1N1O2N2N2N2N2M3N2Md0\\O2O0O2O0O2N2O4K5L3L5K5M3L4L4L3M3M0001O0000000O101O00000000001O000O2O00O1N2N2N2001O1N2O1O1O001O1SK[G]3f8[LbGd3_8SLiGm3X8kKPHT4Q8eKUHZ4m7^K[Ha4R901O001O001O0000O010000000O100O1N2N101N2N2O11O001O1O001N2O001gLmEc1S:QNeFc1\\9oM]Ge1[;D=C=Cb0^Oehj4"}, "label": "the panda on the left"}]}
{"id": 224692, "image": "COCO_train2014_000000224692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the panda closest to the women in the green jacket\"."}], "task": "refering", "answer_template": "The \"the panda closest to the women in the green jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 191, 192, 193, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.2986875, 0.4341666666666667, 0.482703125, 0.8713958333333333], "iscrowd": 0, "area": 17871.199299999997, "rle": {"size": [480, 640], "counts": "dji25]>`0A>TNQO[E[1a:iN\\EY1`:kN^EX1a:hN]EZ1b:hN[E[1c:fN[E\\1d:fNYE]1e:dNYE_1e:bNYE`1g:aNVEb1h:_NVEc1i:^NUEe1i:[1N2N2O1N1O2N2N2N2N2M3N2Md0\\O2O0O2O0O2N2O4K5L3L5K5M3L4L4L3M3M0001O0000000O101O00000000001O000O2O00O1N2N2N2001O1N2O1O1O001O1SK[G]3f8[LbGd3_8SLiGm3X8kKPHT4Q8eKUHZ4m7^K[Ha4R901O001O001O0000O010000000O100O1N2N101N2N2O11O001O1O001N2O001gLmEc1S:QNeFc1\\9oM]Ge1[;D=C=Cb0^Oehj4"}, "label": "the panda closest to the women in the green jacket"}]}
{"id": 167348, "image": "COCO_train2014_000000167348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl in a green shirt holding a tennis racket while balancing a ball\"."}], "task": "refering", "answer_template": "The \"a young girl in a green shirt holding a tennis racket while balancing a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 77, 78, 79, 80, 98, 99, 100, 101, 102, 121, 122, 123, 124, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 229, 230, 231, 232, 233, 234, 235, 250, 251, 252, 253, 254, 255, 256, 272, 273, 274, 275, 276, 277, 278, 294, 295, 296, 297, 298, 299, 300, 316, 317, 318, 319, 320, 321, 322, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 405, 406, 407, 408, 409, 410, 427, 428, 429, 430, 431, 432, 449, 450, 451, 452, 453, 454, 471, 472, 473, 474, 475, 476], "bbox": [0.37176470588235294, 0.13176470588235295, 0.6870588235294118, 0.988235294117647], "iscrowd": 0, "area": 70141.5936, "rle": {"size": [612, 612], "counts": "WdX45hb08I6M3001O100OO1L4K5`^OVOa`0o0\\_OTO_`0R1\\_ORO``0R1__OPO[`0U1d_OlNW`0Z1g_OgNT`0_1i_OcNS`0c1j_O^NQ`0h1m_OZNm?a2K5K5K5J5L5J5K6@?_Ob0O001O0O2O001N2O0_HfJY1[5QKjJh02V4T5oJ_K2Em4n4lJTL^OUOe5g4jJZMXNVNm6a4gJiNW5X1fJlNY5U1cJoN\\5Q1aJSO^5n0]JWOc5h0ZJ\\Oe5e0TJBk5>lILS65dI4\\6K[I?d6Q6O100O1O1O100O1O1O010O1O1O100O10000iMeCeM[<W2jChMV<V2mChMT<V2nCjMR<U2PDjMP<T2SDkMm;S2VDlMi;S2YDmMg;R2[DlMe;S2^DlMb;R2aDoLXOKV<U3dDeLC3j;W3eDYLO<];Z3oEaLS:^3PF^LQ:b3i2N101N101O0O2O00N2M3N2M3M4M2M3M4M2M3L4K6J5K5012N3M2N`0@3M3M3M3M3M3M4L4L4L3M4L4L3M001O0000000000000000001O000001O000000001O0001O0001O000000001O0000100^CSMk8n2RGVMl8j2QGZMn8g2oF]Mo8d2nF_MQ9b2kFcMS9^2jFeMU9\\2hFhMV9X2hFkMW9W2eFlMZ9V2aFnM_9U2[FnMd9a2kEbMT:f2bE\\M[:l2^EWM]:R3\\EQM^:Y3ZEkLa:a3UEbLc3WNk2i5RISLn3_Ne2c7XMhH[2^7aMlHT2Y7gMSIm1R7PNWIe1n6WN]I]1g6`NcICT89WHZOo7b0[HQOl7l0^HfNi7V1aH^Ne7^1fHTN`7i1jHkM\\7Q2oHaMX7[2d5TOk_b3"}, "label": "a young girl in a green shirt holding a tennis racket while balancing a ball"}]}
{"id": 167348, "image": "COCO_train2014_000000167348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl bouncing a giant tennis ball on a racket\"."}], "task": "refering", "answer_template": "The \"a girl bouncing a giant tennis ball on a racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 77, 78, 79, 80, 98, 99, 100, 101, 102, 121, 122, 123, 124, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 229, 230, 231, 232, 233, 234, 235, 250, 251, 252, 253, 254, 255, 256, 272, 273, 274, 275, 276, 277, 278, 294, 295, 296, 297, 298, 299, 300, 316, 317, 318, 319, 320, 321, 322, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 405, 406, 407, 408, 409, 410, 427, 428, 429, 430, 431, 432, 449, 450, 451, 452, 453, 454, 471, 472, 473, 474, 475, 476], "bbox": [0.37176470588235294, 0.13176470588235295, 0.6870588235294118, 0.988235294117647], "iscrowd": 0, "area": 70141.5936, "rle": {"size": [612, 612], "counts": "WdX45hb08I6M3001O100OO1L4K5`^OVOa`0o0\\_OTO_`0R1\\_ORO``0R1__OPO[`0U1d_OlNW`0Z1g_OgNT`0_1i_OcNS`0c1j_O^NQ`0h1m_OZNm?a2K5K5K5J5L5J5K6@?_Ob0O001O0O2O001N2O0_HfJY1[5QKjJh02V4T5oJ_K2Em4n4lJTL^OUOe5g4jJZMXNVNm6a4gJiNW5X1fJlNY5U1cJoN\\5Q1aJSO^5n0]JWOc5h0ZJ\\Oe5e0TJBk5>lILS65dI4\\6K[I?d6Q6O100O1O1O100O1O1O010O1O1O100O10000iMeCeM[<W2jChMV<V2mChMT<V2nCjMR<U2PDjMP<T2SDkMm;S2VDlMi;S2YDmMg;R2[DlMe;S2^DlMb;R2aDoLXOKV<U3dDeLC3j;W3eDYLO<];Z3oEaLS:^3PF^LQ:b3i2N101N101O0O2O00N2M3N2M3M4M2M3M4M2M3L4K6J5K5012N3M2N`0@3M3M3M3M3M3M4L4L4L3M4L4L3M001O0000000000000000001O000001O000000001O0001O0001O000000001O0000100^CSMk8n2RGVMl8j2QGZMn8g2oF]Mo8d2nF_MQ9b2kFcMS9^2jFeMU9\\2hFhMV9X2hFkMW9W2eFlMZ9V2aFnM_9U2[FnMd9a2kEbMT:f2bE\\M[:l2^EWM]:R3\\EQM^:Y3ZEkLa:a3UEbLc3WNk2i5RISLn3_Ne2c7XMhH[2^7aMlHT2Y7gMSIm1R7PNWIe1n6WN]I]1g6`NcICT89WHZOo7b0[HQOl7l0^HfNi7V1aH^Ne7^1fHTN`7i1jHkM\\7Q2oHaMX7[2d5TOk_b3"}, "label": "a girl bouncing a giant tennis ball on a racket"}]}
{"id": 167348, "image": "COCO_train2014_000000167348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing all white\"."}], "task": "refering", "answer_template": "The \"a woman wearing all white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 50, 51, 72, 73, 93, 94, 95, 116, 117, 138, 139, 140, 160, 161, 182, 183, 184, 205, 206, 227, 228], "bbox": [0.2622222222222222, 0.07040849673202615, 0.3938398692810457, 0.4634967320261437], "iscrowd": 0, "area": 9450.0111, "rle": {"size": [612, 612], "counts": "hgo2<hb04K6J3N2M2N3M3c@VO[<m0SB^OA3S>o0iAWOH2^>m0`A>^>U2_NbK[D_4c;gKXDZ4f;mKTDS4k;QLRDP4l;TLQDl3n;XLoCi3o;[LnCe3R<^LkC`3X<aLfC\\3^<eL_CX3e<jLYCS3k<nLRCP3R=SMiBj2[=]13M3M3N1O2000O100001000O2O1O1O1O2N1OO1000001YK]Bj3e=QL`Bm3c=nK`BQ4\\>ObBeKX<S4fCQLQ=W3kBmLU=S3hBnL[=P3bBRM`=n2]BSMf=l2WBUMk=m1^A_Ne0EP>]1hAoN5EV>V1jATONG[>o0kA[O[O2n>=lA`0W>\\OkAb0Y>ZOjAd0Y>YOhAf0[>VOhAg0\\>VOeAi0]>TOfAj0\\>SOgAj0\\>SOiAi0Y>TOlAh0V>UOoAg0S>VOQBg0U>ROnAl0X`0L4M2M4M2M4M5Jm^m6"}, "label": "a woman wearing all white"}]}
{"id": 167348, "image": "COCO_train2014_000000167348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with blonde hair and a blue purse\"."}], "task": "refering", "answer_template": "The \"a woman with blonde hair and a blue purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 50, 51, 72, 73, 93, 94, 95, 116, 117, 138, 139, 140, 160, 161, 182, 183, 184, 205, 206, 227, 228], "bbox": [0.2622222222222222, 0.07040849673202615, 0.3938398692810457, 0.4634967320261437], "iscrowd": 0, "area": 9450.0111, "rle": {"size": [612, 612], "counts": "hgo2<hb04K6J3N2M2N3M3c@VO[<m0SB^OA3S>o0iAWOH2^>m0`A>^>U2_NbK[D_4c;gKXDZ4f;mKTDS4k;QLRDP4l;TLQDl3n;XLoCi3o;[LnCe3R<^LkC`3X<aLfC\\3^<eL_CX3e<jLYCS3k<nLRCP3R=SMiBj2[=]13M3M3N1O2000O100001000O2O1O1O1O2N1OO1000001YK]Bj3e=QL`Bm3c=nK`BQ4\\>ObBeKX<S4fCQLQ=W3kBmLU=S3hBnL[=P3bBRM`=n2]BSMf=l2WBUMk=m1^A_Ne0EP>]1hAoN5EV>V1jATONG[>o0kA[O[O2n>=lA`0W>\\OkAb0Y>ZOjAd0Y>YOhAf0[>VOhAg0\\>VOeAi0]>TOfAj0\\>SOgAj0\\>SOiAi0Y>TOlAh0V>UOoAg0S>VOQBg0U>ROnAl0X`0L4M2M4M2M4M5Jm^m6"}, "label": "a woman with blonde hair and a blue purse"}]}
{"id": 298427, "image": "COCO_train2014_000000298427.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green piece of broccoli surrounded by a lot of different types of food on a white plate\"."}], "task": "refering", "answer_template": "The \"a green piece of broccoli surrounded by a lot of different types of food on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 34, 35, 36, 37, 38, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 129, 130, 131], "bbox": [0.477234375, 0.02841121495327103, 0.7236874999999999, 0.38051401869158874], "iscrowd": 0, "area": 15504.64895, "rle": {"size": [428, 640], "counts": "abo31X=4M3L4L4L5K4L4L4M3L5K4L4M3N2M4L3M3N2M4L3N2M3M3M4M2M3M3M3N3L2N2O1N2N2O2M2N2N2O1N2N2O2M2N2001O00001O01O0001O0000001O01O0001O00001O01O0001O0000001O01O01O00001O0010O01O1O001O001O10O01O10O100O100O10O010O010O001O010O001O010O0010O01O0010O00010O01O00010O00010O001O01O01O01O01O001POUF^Nj9`1ZF^Nf9a1]F\\Ne9a1^F^Nb9a1aF\\N`9b1cF]N^9a1eF\\N\\9b1hF\\NY9b1iF\\NZ9`1jF^NX9_1jF_NY9]1kFaNW9\\1kFbNX9Y1mFeNU9V1PGgNS9T1f1J7J5K5J6K5KolY2"}, "label": "a green piece of broccoli surrounded by a lot of different types of food on a white plate"}]}
{"id": 372156, "image": "COCO_train2014_000000372156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train cake on the bottom\"."}], "task": "refering", "answer_template": "The \"the train cake on the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 381, 382, 383, 384], "bbox": [0.261234375, 0.5910208333333333, 0.77190625, 0.9595416666666666], "iscrowd": 0, "area": 46861.699299999986, "rle": {"size": [480, 640], "counts": "af^22k>6J6J3M3M4L3O1N2O1N2O1N3N1N2O1N2O1N2N3N1\\OcNcC_1Y<hNcCY1Z<lNbCV1Z<POcCQ1Y<m0M3L5K4M3L4M3001OL5L3L4L4L4L4M3L4M4L3N2N2N2M3N2N3M2O10000O10001O0000000O1000001O000O100000001O0000000000000010O000000000000000000000000000000001O000000000000001O001O001O001O001O0010O01O10O01O010O0010N1000001O00001O00001O0O101O00001OO10000O10000O10000O10000O10O10O10000O10000O10000O10000O10000O1000O010000O101O0O10000O10000O10000001O1O001O1O1O1O1O000000001O00000000000000001O000000000000001O00000000000000001O000000000000000000000O10000000010O00000001O00000000001O000000001O00000000001O00000000001O00001O001O001O00001O001O0O2O00001O001O001O000O2O001N1N3M2M4M2N2N3M2M4M3M3M3M3L4M3M3M3M3M3L4M3M3M4L3L4M3M3MeST2"}, "label": "the train cake on the bottom"}]}
{"id": 372156, "image": "COCO_train2014_000000372156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom train cake in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the bottom train cake in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 381, 382, 383, 384], "bbox": [0.261234375, 0.5910208333333333, 0.77190625, 0.9595416666666666], "iscrowd": 0, "area": 46861.699299999986, "rle": {"size": [480, 640], "counts": "af^22k>6J6J3M3M4L3O1N2O1N2O1N3N1N2O1N2O1N2N3N1\\OcNcC_1Y<hNcCY1Z<lNbCV1Z<POcCQ1Y<m0M3L5K4M3L4M3001OL5L3L4L4L4L4M3L4M4L3N2N2N2M3N2N3M2O10000O10001O0000000O1000001O000O100000001O0000000000000010O000000000000000000000000000000001O000000000000001O001O001O001O001O0010O01O10O01O010O0010N1000001O00001O00001O0O101O00001OO10000O10000O10000O10000O10O10O10000O10000O10000O10000O10000O1000O010000O101O0O10000O10000O10000001O1O001O1O1O1O1O000000001O00000000000000001O000000000000001O00000000000000001O000000000000000000000O10000000010O00000001O00000000001O000000001O00000000001O00000000001O00001O001O001O00001O001O0O2O00001O001O001O000O2O001N1N3M2M4M2N2N3M2M4M3M3M3M3L4M3M3M3M3M3L4M3M3M4L3L4M3M3MeST2"}, "label": "the bottom train cake in the right hand picture"}]}
{"id": 413121, "image": "COCO_train2014_000000413121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe standing in front of other two giraffes is looking on to the left\"."}], "task": "refering", "answer_template": "The \"a tall giraffe standing in front of other two giraffes is looking on to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 240, 241, 242, 265, 266, 267, 288, 289, 290, 291, 292, 312, 313, 314, 315, 335, 337, 338], "bbox": [0.452453125, 0.5988179669030733, 0.7038125, 0.9895508274231679], "iscrowd": 0, "area": 7461.326350000003, "rle": {"size": [423, 640], "counts": "[Rh32T=1O2M2O2N1O0000001O01O0001O1N2L4O1O1O100N2N2N2O1O10O2O0O2O6I01O00103L5KXOgCg0U<WOoCl0U<1O001O1O010O100O00100O010O100O00100O0010O01O1O010O00100O001O010O1O001O001\\DfNX;j1K5K4L00O100000000O1000nDeNW:U2010O1O010O001K4E<M3L3M4L3O2O001O1O001O001O1O010O001O1O00100O001O0010O0001O000001O003M:F8H01O00000000000[Oe0103L4L4L4L4MO001O100O001O1O10OL5K5F9Gc0]O^T^2"}, "label": "a tall giraffe standing in front of other two giraffes is looking on to the left"}]}
{"id": 413121, "image": "COCO_train2014_000000413121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing in the shade\"."}], "task": "refering", "answer_template": "The \"a giraffe standing in the shade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 240, 241, 242, 265, 266, 267, 288, 289, 290, 291, 292, 312, 313, 314, 315, 335, 337, 338], "bbox": [0.452453125, 0.5988179669030733, 0.7038125, 0.9895508274231679], "iscrowd": 0, "area": 7461.326350000003, "rle": {"size": [423, 640], "counts": "[Rh32T=1O2M2O2N1O0000001O01O0001O1N2L4O1O1O100N2N2N2O1O10O2O0O2O6I01O00103L5KXOgCg0U<WOoCl0U<1O001O1O010O100O00100O010O100O00100O0010O01O1O010O00100O001O010O1O001O001\\DfNX;j1K5K4L00O100000000O1000nDeNW:U2010O1O010O001K4E<M3L3M4L3O2O001O1O001O001O1O010O001O1O00100O001O0010O0001O000001O003M:F8H01O00000000000[Oe0103L4L4L4L4MO001O100O001O1O10OL5K5F9Gc0]O^T^2"}, "label": "a giraffe standing in the shade"}]}
{"id": 470471, "image": "COCO_train2014_000000470471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"airplane seat next to the window\"."}], "task": "refering", "answer_template": "The \"airplane seat next to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 50, 51, 52, 53, 68, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 122, 123, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 196, 197, 214, 215, 232, 233], "bbox": [0.81404, 0.0028, 1.0, 0.9915733333333334], "iscrowd": 0, "area": 25702.996700000003, "rle": {"size": [375, 500], "counts": "jUe46b0Mo9`0nEEc9h0ZF^OV9o0hFVOg8X1VGnN^8[1_GjN[8Z1bGjNX8_2J6J6J7I6J6J6J6J6J6K5J6J6J6J7I6J6J6L4L4K5L4L4L4K6K4L4M3L4M3L4M3M3L4M3L5L3L4M3M3M5J;F:Fe0ZO4M1O1O0O2O1O1O0O2O1O0O2O1O000O10000O0100000O1000000O10000O1000000O10000O0100000O100O"}, "label": "airplane seat next to the window"}]}
{"id": 470471, "image": "COCO_train2014_000000470471.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a blue airplane seat with a grey plastic tray on it\"."}], "task": "refering", "answer_template": "The \"the back of a blue airplane seat with a grey plastic tray on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 50, 51, 52, 53, 68, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 122, 123, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 196, 197, 214, 215, 232, 233], "bbox": [0.81404, 0.0028, 1.0, 0.9915733333333334], "iscrowd": 0, "area": 25702.996700000003, "rle": {"size": [375, 500], "counts": "jUe46b0Mo9`0nEEc9h0ZF^OV9o0hFVOg8X1VGnN^8[1_GjN[8Z1bGjNX8_2J6J6J7I6J6J6J6J6J6K5J6J6J6J7I6J6J6L4L4K5L4L4L4K6K4L4M3L4M3L4M3M3L4M3L5L3L4M3M3M5J;F:Fe0ZO4M1O1O0O2O1O1O0O2O1O0O2O1O000O10000O0100000O1000000O10000O1000000O10000O0100000O100O"}, "label": "the back of a blue airplane seat with a grey plastic tray on it"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman peeling an onion\"."}], "task": "refering", "answer_template": "The \"a woman peeling an onion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 293, 294, 309, 310, 311, 312, 313, 317, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 377, 379, 380], "bbox": [0.39675, 0.207625, 0.8427187500000001, 0.9621458333333334], "iscrowd": 0, "area": 48530.50175, "rle": {"size": [480, 640], "counts": "V_g3:d>4N2M2O2N2M3N2M3N2M3N1N2O1N2O1N2O000O100O1OB>O1O1O10O01O1O1O100O1O0101O1O1O2N1O1N2O1O1O2N1O2N1N2O2M4L3I8I6J6I7J6J7H7J6I8I7I7H`0A:F7H7J5K5J5N3O100O1O1O100O00100O1O1O100O1O1O010O1O1O100O1O00100O1O0010O01O0010O01O010O001O010O00O101O0O10001NlMlI_KS6a4ZJSKd5n4hJgJU5Z5YKQJk3]OlKb6e0^Ih31dK\\6W6fIiIW6Y6jIgIR6[6PJRJ`5o5aJPJ^5R6cJlI[5V6gJgIY5[6iJbIU5`6nJ]IQ5d6QKYIo4i6SKTIk4n6XKnHh4T7ZKiHd4Y7^KdHb4^7`K`H]4b7fK[HY4f7iKYHU4i7mKUHR4k7QLRHo3o7SLoGm3P8ULoGk3Q8WLlGi3T8ZLjGf3V8a110O10O100O01000O10O01000M2N3M3M2N3N101N2O0O2O0O2O10O2O0O10001N10001N102M3YIVHT6m7bIYH_6R8101N101O1N2O1_Ob0UOj0O2O00001O001O001O001O00001O001O0O2O00001O001O001O0010O0010O010O0010O0010O0010O010O2N2O1N2O2M2OfLiGLT82oG1n7KWH6X1bLl4U3PJ<n0aLR5P3UJf0?ZL_5k2VJR11ULk5d2YJ^1BoKX6]2ZJk1TOjKd6V2]JY2e5aM`J^2b5]MbJa2b5YMcJe2_5VMeJi2k5dLZJY3e6fK_IX4g8M2N3M2N3M2N3M2M4M2N2N3M2M4H7I8H8H8J6J6I7J6J7H7J6J6I5L4L4KjQ_1"}, "label": "a woman peeling an onion"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a grey hoodie and black shirt peeling an egg\"."}], "task": "refering", "answer_template": "The \"a woman with a grey hoodie and black shirt peeling an egg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 293, 294, 309, 310, 311, 312, 313, 317, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 377, 379, 380], "bbox": [0.39675, 0.207625, 0.8427187500000001, 0.9621458333333334], "iscrowd": 0, "area": 48530.50175, "rle": {"size": [480, 640], "counts": "V_g3:d>4N2M2O2N2M3N2M3N2M3N1N2O1N2O1N2O000O100O1OB>O1O1O10O01O1O1O100O1O0101O1O1O2N1O1N2O1O1O2N1O2N1N2O2M4L3I8I6J6I7J6J7H7J6I8I7I7H`0A:F7H7J5K5J5N3O100O1O1O100O00100O1O1O100O1O1O010O1O1O100O1O00100O1O0010O01O0010O01O010O001O010O00O101O0O10001NlMlI_KS6a4ZJSKd5n4hJgJU5Z5YKQJk3]OlKb6e0^Ih31dK\\6W6fIiIW6Y6jIgIR6[6PJRJ`5o5aJPJ^5R6cJlI[5V6gJgIY5[6iJbIU5`6nJ]IQ5d6QKYIo4i6SKTIk4n6XKnHh4T7ZKiHd4Y7^KdHb4^7`K`H]4b7fK[HY4f7iKYHU4i7mKUHR4k7QLRHo3o7SLoGm3P8ULoGk3Q8WLlGi3T8ZLjGf3V8a110O10O100O01000O10O01000M2N3M3M2N3N101N2O0O2O0O2O10O2O0O10001N10001N102M3YIVHT6m7bIYH_6R8101N101O1N2O1_Ob0UOj0O2O00001O001O001O001O00001O001O0O2O00001O001O001O0010O0010O010O0010O0010O0010O010O2N2O1N2O2M2OfLiGLT82oG1n7KWH6X1bLl4U3PJ<n0aLR5P3UJf0?ZL_5k2VJR11ULk5d2YJ^1BoKX6]2ZJk1TOjKd6V2]JY2e5aM`J^2b5]MbJa2b5YMcJe2_5VMeJi2k5dLZJY3e6fK_IX4g8M2N3M2N3M2N3M2M4M2N2N3M2M4H7I8H8H8J6J6I7J6J7H7J6J6I5L4L4KjQ_1"}, "label": "a woman with a grey hoodie and black shirt peeling an egg"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in black and pink\"."}], "task": "refering", "answer_template": "The \"a girl in black and pink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 162, 163, 164, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 351, 352, 353, 354, 376], "bbox": [0.07753125, 0.11685416666666668, 0.483703125, 0.9550624999999999], "iscrowd": 0, "area": 54615.3121, "rle": {"size": [480, 640], "counts": "odg08b>7M3O1O2N100O1O1O101N1O1O100O2N1O010O000010O00WOQOaCo0`<VOZCk0f<ZOUCe0l<_OoBa0R=DhB=X=e01O100O1O1O100O1O100O1O1O100ObDUOh8j0WG]Od8c0ZG]Oh8R1hFnNY9R1eFnN]9R1aFnNa9R1]FoNd9Q1ZFoNh9Q1VFoNl9Q1RFoNP:Q1nEPOS:P1kEPOV:Q1iEnNY:R1eEoN\\:Q1bEoN`:Q1^EoNd:P1[EPOg:P1WEQOj:o0TEQOn:o0PEQOR;o0lDROU;n0iDROY;n0fDQO[;P1cDQO^;o0`DQO;VO^:i1UEQO=XO^:g1SEQO?YO_:f1PERO`0ZO`:d1nDROb0\\O`:b1lDROd0]Oa:`1jDTOd0^Ob:[2^EgMb:W2^EjMc:T2]EnMb:R2]EPNc:n1]ESNd:k1\\EWNd:h1\\EYNd:e1\\E\\Ne:b1[E`Na:c1^E^N`:c1`E_N\\:c1dE^NY:d1gE]NV:f1iE\\NS:g1lEZNQ:i1nEYNn9k1PFVNm9n1QFSNl9P2SFRNi9R2UFoMh9U2VFlMg9X2WFjMe9Y2ZFhMc9\\2[FfMa9n3ZNfJdI\\5Y6iJbI[5Z6jJaIY5\\6lJ_IW5o4eJPK8MW5R5ZLlJf3S5]LjJe3T5^LiJc3V5`LgJa3X5aLfJa3Y5aLeJ_3Z5dLcJ]3\\5fLaJ\\3]5gL`JZ3_5hL_JZ3`5hL^JX3a5T3O2N100O2N1O2N100O2N1O1O2O0O2N1O1dLTFd1m9YNUFf1l9XNVFg1k9VNXFi1j9TNYFj1h9SN[Fm1f9PN\\Fo1e9oM]FP2d9mM_FR2c9kM`FT2`9jMbFU2_9hMdFW2S;O1O2O0O100010O00000001O01O000001O01O0001O000001O01O000001O01O00000010O000000010O00000001O01O2N2O0O2N2N2N2O1N2N2N2I7^Oa0M4L4M3L4L4M3L4M3L3M4M3L4M3L4M3L4M2N3L<E=C=B?B8H4K5L4L4K5L4L4K5L2N2UNWG[Mk8c2hGkLZ8S3WH\\Lj7b3l1M2N3N2M2N3N1N3M2O2M2L5K7I;E:F:E;F:F:F:F:Fn]j4"}, "label": "a girl in black and pink"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in pink top and black sweater\"."}], "task": "refering", "answer_template": "The \"woman in pink top and black sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 162, 163, 164, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 351, 352, 353, 354, 376], "bbox": [0.07753125, 0.11685416666666668, 0.483703125, 0.9550624999999999], "iscrowd": 0, "area": 54615.3121, "rle": {"size": [480, 640], "counts": "odg08b>7M3O1O2N100O1O1O101N1O1O100O2N1O010O000010O00WOQOaCo0`<VOZCk0f<ZOUCe0l<_OoBa0R=DhB=X=e01O100O1O1O100O1O100O1O1O100ObDUOh8j0WG]Od8c0ZG]Oh8R1hFnNY9R1eFnN]9R1aFnNa9R1]FoNd9Q1ZFoNh9Q1VFoNl9Q1RFoNP:Q1nEPOS:P1kEPOV:Q1iEnNY:R1eEoN\\:Q1bEoN`:Q1^EoNd:P1[EPOg:P1WEQOj:o0TEQOn:o0PEQOR;o0lDROU;n0iDROY;n0fDQO[;P1cDQO^;o0`DQO;VO^:i1UEQO=XO^:g1SEQO?YO_:f1PERO`0ZO`:d1nDROb0\\O`:b1lDROd0]Oa:`1jDTOd0^Ob:[2^EgMb:W2^EjMc:T2]EnMb:R2]EPNc:n1]ESNd:k1\\EWNd:h1\\EYNd:e1\\E\\Ne:b1[E`Na:c1^E^N`:c1`E_N\\:c1dE^NY:d1gE]NV:f1iE\\NS:g1lEZNQ:i1nEYNn9k1PFVNm9n1QFSNl9P2SFRNi9R2UFoMh9U2VFlMg9X2WFjMe9Y2ZFhMc9\\2[FfMa9n3ZNfJdI\\5Y6iJbI[5Z6jJaIY5\\6lJ_IW5o4eJPK8MW5R5ZLlJf3S5]LjJe3T5^LiJc3V5`LgJa3X5aLfJa3Y5aLeJ_3Z5dLcJ]3\\5fLaJ\\3]5gL`JZ3_5hL_JZ3`5hL^JX3a5T3O2N100O2N1O2N100O2N1O1O2O0O2N1O1dLTFd1m9YNUFf1l9XNVFg1k9VNXFi1j9TNYFj1h9SN[Fm1f9PN\\Fo1e9oM]FP2d9mM_FR2c9kM`FT2`9jMbFU2_9hMdFW2S;O1O2O0O100010O00000001O01O000001O01O0001O000001O01O000001O01O00000010O000000010O00000001O01O2N2O0O2N2N2N2O1N2N2N2I7^Oa0M4L4M3L4L4M3L4M3L3M4M3L4M3L4M3L4M2N3L<E=C=B?B8H4K5L4L4K5L4L4K5L2N2UNWG[Mk8c2hGkLZ8S3WH\\Lj7b3l1M2N3N2M2N3N1N3M2O2M2L5K7I;E:F:E;F:F:F:F:Fn]j4"}, "label": "woman in pink top and black sweater"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of a brown haired girl wearing brown jacket\"."}], "task": "refering", "answer_template": "The \"back of a brown haired girl wearing brown jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 133, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275, 298, 321], "bbox": [0.78715625, 0.2691458333333333, 1.0, 0.8254583333333333], "iscrowd": 0, "area": 15628.512200000001, "rle": {"size": [480, 640], "counts": "P_\\79c>=C9G5K5L4K5K5I7J5K6I7J6L4M3M3N1N3M3M3M3N1N3M3M2O2M3M2N3N2M2N3O1O0O2O1N2O0O2O4K5L5J6K5J6K3N0O00010O00010O0001O01O01O01O01O00010O00010O1O1O1O2N101N1O1O2N1O1O2N101N1O1O2N1O1O3N<C>B8H8Hb0^O7I2O0O2O1N2O0O2N2O1N101N2O2M3N3L3M3N3L3N0O1010Oc0^O=B>C3L10O01O10O2O1O1M3E;E:G:E:F:F;E=CRF"}, "label": "back of a brown haired girl wearing brown jacket"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in black nearest pink towel hanging on wall\"."}], "task": "refering", "answer_template": "The \"girl in black nearest pink towel hanging on wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 133, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 205, 225, 226, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275, 298, 321], "bbox": [0.78715625, 0.2691458333333333, 1.0, 0.8254583333333333], "iscrowd": 0, "area": 15628.512200000001, "rle": {"size": [480, 640], "counts": "P_\\79c>=C9G5K5L4K5K5I7J5K6I7J6L4M3M3N1N3M3M3M3N1N3M3M2O2M3M2N3N2M2N3O1O0O2O1N2O0O2O4K5L5J6K5J6K3N0O00010O00010O0001O01O01O01O01O00010O00010O1O1O1O2N101N1O1O2N1O1O2N101N1O1O2N1O1O3N<C>B8H8Hb0^O7I2O0O2O1N2O0O2N2O1N101N2O2M3N3L3M3N3L3N0O1010Oc0^O=B>C3L10O01O10O2O1O1M3E;E:G:E:F:F;E=CRF"}, "label": "girl in black nearest pink towel hanging on wall"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl standing behind and in - between two other girls\"."}], "task": "refering", "answer_template": "The \"a girl standing behind and in - between two other girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 105, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 195, 196, 218, 240, 241, 263, 264, 286], "bbox": [0.43792187499999996, 0.22841666666666666, 0.59528125, 0.729375], "iscrowd": 0, "area": 10503.29365, "rle": {"size": [480, 640], "counts": "l]S42m>3M4L4gEL_57\\JLc56YJMf54XJOf53VJ0i52SJ1m5OQJ3n5OnI4Q6NkI5T6MiI5V6LgI7X6KdI8\\6IaI:]6G`I<_6F]I=b6E[I=d6EXI>g6CVI`0i6BTI`0j6FoH<Q7MdH5[74\\HKf7>oG_OU8m0^GQOf8`1hF\\N\\9b32N2N1O0O1000O100000O100O3N2M3M3N2M3O1000000000O1001O1O1O001O1M3L4L301N2M3M3M3L4J6I8H7O1100O1O010O0001O010O001O100O1O1L4M3M3L4]Ob0J7L4L4M3L5K5L4D=DPYi3"}, "label": "a girl standing behind and in - between two other girls"}]}
{"id": 282067, "image": "COCO_train2014_000000282067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark haired girl who is wearing a grey shirt with a red collar\"."}], "task": "refering", "answer_template": "The \"the dark haired girl who is wearing a grey shirt with a red collar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 105, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 195, 196, 218, 240, 241, 263, 264, 286], "bbox": [0.43792187499999996, 0.22841666666666666, 0.59528125, 0.729375], "iscrowd": 0, "area": 10503.29365, "rle": {"size": [480, 640], "counts": "l]S42m>3M4L4gEL_57\\JLc56YJMf54XJOf53VJ0i52SJ1m5OQJ3n5OnI4Q6NkI5T6MiI5V6LgI7X6KdI8\\6IaI:]6G`I<_6F]I=b6E[I=d6EXI>g6CVI`0i6BTI`0j6FoH<Q7MdH5[74\\HKf7>oG_OU8m0^GQOf8`1hF\\N\\9b32N2N1O0O1000O100000O100O3N2M3M3N2M3O1000000000O1001O1O1O001O1M3L4L301N2M3M3M3L4J6I8H7O1100O1O010O0001O010O001O100O1O1L4M3M3L4]Ob0J7L4L4M3L5K5L4D=DPYi3"}, "label": "the dark haired girl who is wearing a grey shirt with a red collar"}]}
{"id": 331222, "image": "COCO_train2014_000000331222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue couch that has three people sitting on it\"."}], "task": "refering", "answer_template": "The \"a blue couch that has three people sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 187, 188, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283], "bbox": [0.04678125, 0.5161250000000001, 0.340328125, 0.7354791666666667], "iscrowd": 0, "area": 14843.498349999996, "rle": {"size": [480, 640], "counts": "`[>`0a=P1_O`0M3N3L3O2O01O01O010O0010O00010O01O01O01O010O0010O00010O01O01O01N101N101O0O101N101O00001O001O001O00001N101O00001O001O001O00001O001O00001O001O001O000O2O001O000000000000O1000000000001O1O2N1N2O2N1O2N1O2N1O1O2N1O2M2O2N1O1O2N000000O10O10000000000000000000O0100000000000000000000O0100000000000000000O10O100000000000000000O1000O1000000000000000O100000O100000000000O10000000`0@g0YOh0XO`_U6"}, "label": "a blue couch that has three people sitting on it"}]}
{"id": 331222, "image": "COCO_train2014_000000331222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue sofa on top of red and yellow crates\"."}], "task": "refering", "answer_template": "The \"a blue sofa on top of red and yellow crates\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 187, 188, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283], "bbox": [0.04678125, 0.5161250000000001, 0.340328125, 0.7354791666666667], "iscrowd": 0, "area": 14843.498349999996, "rle": {"size": [480, 640], "counts": "`[>`0a=P1_O`0M3N3L3O2O01O01O010O0010O00010O01O01O01O010O0010O00010O01O01O01N101N101O0O101N101O00001O001O001O00001N101O00001O001O001O00001O001O00001O001O001O000O2O001O000000000000O1000000000001O1O2N1N2O2N1O2N1O2N1O1O2N1O2M2O2N1O1O2N000000O10O10000000000000000000O0100000000000000000000O0100000000000000000O10O100000000000000000O1000O1000000000000000O100000O100000000000O10000000`0@g0YOh0XO`_U6"}, "label": "a blue sofa on top of red and yellow crates"}]}
{"id": 331222, "image": "COCO_train2014_000000331222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black couch\"."}], "task": "refering", "answer_template": "The \"black couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290], "bbox": [0.374703125, 0.5123333333333333, 0.618265625, 0.75], "iscrowd": 0, "area": 10616.474900000003, "rle": {"size": [480, 640], "counts": "hg`33m>1O5K5J5L5K4L5J5L4L3lBhNZ<Z1`ChNb<Y1[ChNg<Y1UCiNk<Y1QCiNo<c1101O0O10001O00000O101O01O0010O01O00100O2N1O1O2O0O1O0001O01O0001O000001O0dCQNh;b2100O101O00000010O000000010O0001000O20O2N7K3MO0001M2M3N1O2N1N2O00000O10O10O1000O1O100O1O1O1O101N1O100O1O1O1O010O1O1O1O100O01000O1000000O01000O1O100O10O1000000000000O1000000000000000000000001O000000000I7Ca0TOl0UOmSb3"}, "label": "black couch"}]}
{"id": 331222, "image": "COCO_train2014_000000331222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leather couch\"."}], "task": "refering", "answer_template": "The \"the leather couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290], "bbox": [0.374703125, 0.5123333333333333, 0.618265625, 0.75], "iscrowd": 0, "area": 10616.474900000003, "rle": {"size": [480, 640], "counts": "hg`33m>1O5K5J5L5K4L5J5L4L3lBhNZ<Z1`ChNb<Y1[ChNg<Y1UCiNk<Y1QCiNo<c1101O0O10001O00000O101O01O0010O01O00100O2N1O1O2O0O1O0001O01O0001O000001O0dCQNh;b2100O101O00000010O000000010O0001000O20O2N7K3MO0001M2M3N1O2N1N2O00000O10O10O1000O1O100O1O1O1O101N1O100O1O1O1O010O1O1O1O100O01000O1000000O01000O1O100O10O1000000000000O1000000000000000000000001O000000000I7Ca0TOl0UOmSb3"}, "label": "the leather couch"}]}
{"id": 544215, "image": "COCO_train2014_000000544215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a side view of an elephant with tusks , standing next to a body of water\"."}], "task": "refering", "answer_template": "The \"a side view of an elephant with tusks , standing next to a body of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 188, 189, 190, 191, 192, 193, 208, 209, 211, 212, 213, 214, 215, 216, 231, 232, 235, 236, 237, 238, 239, 254, 255, 258, 259, 260, 261, 262], "bbox": [0.04375, 0.27166666666666667, 0.4329375000000001, 0.6838333333333334], "iscrowd": 0, "area": 27847.86654999999, "rle": {"size": [480, 640], "counts": "Sj=2m>1O001O1N2O001O1O1N101O1N2M2O2M9H`0_O`0A?@`0A?@?B6I7J6I7J5K2M2O2NSNlFQNS9h1kGcMR8W2aH^M^7\\2e2I7I7I6M42N2N2N2M3N2N2N2N2000000000O1O1O1O1O00100001O00001O00001O00001O00001O001O3M3M3MO10000000000000000O100000O10000000002N4L4L4L4L4L4L4L4L4L4L4L4kERLP9R4dFYLY9k3\\F^Lb9b4N2N2N3M6J5K4MO0000000O2O0000001O000000001N10000^NSGQMm8m2WGQMj8m2YGQMg8n2\\GPMd8P3^GmLc8R3aGjL`8U3cGiL^8U3eGhL\\8W3gGgLY8Y3iGdLX8[3kGbLV8]3mGaLT8]3nGaLS8_3oG_LQ8`3QH^LP8a3SH\\Lo7b3SH]Lm7c3UHZLl7e3h1O101N100O100O1O10O01000000001O000O101O00001O0\\F]Lk7c3TH^Ll7b3RHaLl7a3QHaLo7_3PHbLP8^3nGdLR8]3lGdLT8\\3jGfLV8Z3hGhLX8X3gGiLY8X3dGkL[8U3dGlL[8U3cGmL]8T3`GnL`8R3^GPMb8U3QGSMo8R3cFWM]9W4O000O10000000010O01L4G9G8HkF[K`8\\4gGeKZ8R4mGoKT8h3RHZLn7P4gGQLZ8[4XGgKh8e4jF\\KW9V5001O1O10O01O1O001O100O001O1WOaFoK`9g3nFTLR9d3\\GWLd8_3d1H8K5K6J7H7J7I6J7I6J7H7J7I6J7I6I8IRoY5"}, "label": "a side view of an elephant with tusks , standing next to a body of water"}]}
{"id": 544215, "image": "COCO_train2014_000000544215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with long white tusks standing next to a pool of water\"."}], "task": "refering", "answer_template": "The \"an elephant with long white tusks standing next to a pool of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 188, 189, 190, 191, 192, 193, 208, 209, 211, 212, 213, 214, 215, 216, 231, 232, 235, 236, 237, 238, 239, 254, 255, 258, 259, 260, 261, 262], "bbox": [0.04375, 0.27166666666666667, 0.4329375000000001, 0.6838333333333334], "iscrowd": 0, "area": 27847.86654999999, "rle": {"size": [480, 640], "counts": "Sj=2m>1O001O1N2O001O1O1N101O1N2M2O2M9H`0_O`0A?@`0A?@?B6I7J6I7J5K2M2O2NSNlFQNS9h1kGcMR8W2aH^M^7\\2e2I7I7I6M42N2N2N2M3N2N2N2N2000000000O1O1O1O1O00100001O00001O00001O00001O00001O001O3M3M3MO10000000000000000O100000O10000000002N4L4L4L4L4L4L4L4L4L4L4L4kERLP9R4dFYLY9k3\\F^Lb9b4N2N2N3M6J5K4MO0000000O2O0000001O000000001N10000^NSGQMm8m2WGQMj8m2YGQMg8n2\\GPMd8P3^GmLc8R3aGjL`8U3cGiL^8U3eGhL\\8W3gGgLY8Y3iGdLX8[3kGbLV8]3mGaLT8]3nGaLS8_3oG_LQ8`3QH^LP8a3SH\\Lo7b3SH]Lm7c3UHZLl7e3h1O101N100O100O1O10O01000000001O000O101O00001O0\\F]Lk7c3TH^Ll7b3RHaLl7a3QHaLo7_3PHbLP8^3nGdLR8]3lGdLT8\\3jGfLV8Z3hGhLX8X3gGiLY8X3dGkL[8U3dGlL[8U3cGmL]8T3`GnL`8R3^GPMb8U3QGSMo8R3cFWM]9W4O000O10000000010O01L4G9G8HkF[K`8\\4gGeKZ8R4mGoKT8h3RHZLn7P4gGQLZ8[4XGgKh8e4jF\\KW9V5001O1O10O01O1O001O100O001O1WOaFoK`9g3nFTLR9d3\\GWLd8_3d1H8K5K6J7H7J7I6J7I6J7H7J7I6J7I6I8IRoY5"}, "label": "an elephant with long white tusks standing next to a pool of water"}]}
{"id": 36318, "image": "COCO_train2014_000000036318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small glass tumbler\"."}], "task": "refering", "answer_template": "The \"a small glass tumbler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 26, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 143], "bbox": [0.15371875, 0.0351288056206089, 0.354484375, 0.4100234192037471], "iscrowd": 0, "area": 16049.632149999998, "rle": {"size": [427, 640], "counts": "XZY16T=5J6K4K6K5J6K4K6K5L3M4L4K5L3M4L4K4M3M3M4K4M3M3M3L4M3M3M3L5L3M3M3L4M3M3M3L3N1O001N2O001O1O0000000000001N10000000000000000000O100O01000O10000O100O10000O10O10O100O10000O10000O2O0O10000O10000O2O0O10000O10001O0O10001O0O10001O0O10001O0O10001O0O1XOmFXMS9]2kGPMW8c2_1D<D=D;D<D\\V\\5"}, "label": "a small glass tumbler"}]}
{"id": 36318, "image": "COCO_train2014_000000036318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small cup\"."}], "task": "refering", "answer_template": "The \"a small cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 26, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 143], "bbox": [0.15371875, 0.0351288056206089, 0.354484375, 0.4100234192037471], "iscrowd": 0, "area": 16049.632149999998, "rle": {"size": [427, 640], "counts": "XZY16T=5J6K4K6K5J6K4K6K5L3M4L4K5L3M4L4K4M3M3M4K4M3M3M3L4M3M3M3L5L3M3M3L4M3M3M3L3N1O001N2O001O1O0000000000001N10000000000000000000O100O01000O10000O100O10000O10O10O100O10000O10000O2O0O10000O10000O2O0O10000O10001O0O10001O0O10001O0O10001O0O10001O0O1XOmFXMS9]2kGPMW8c2_1D<D=D;D<D\\V\\5"}, "label": "a small cup"}]}
{"id": 36318, "image": "COCO_train2014_000000036318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza on plate with food\"."}], "task": "refering", "answer_template": "The \"pizza on plate with food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.1199375, 0.34156908665105384, 0.9490624999999999, 0.9618032786885244], "iscrowd": 0, "area": 102918.87639999998, "rle": {"size": [427, 640], "counts": "c]P1=]<b0]Oc0L4N2M2O2M3N2M3M3N1N3N2M3N2M3M2O2O1N2N2O1N1O2N2O1N2N2O0O2N2N2O1N2N101N2N2N2O1N1O2O1N2N2N2O0O2N2O1N2N2O0O2N2NOZG`Ke8a411O2O0O2N1O2O0O1O2O0O2N1O2O0O2N100O2N1O2O0O2N101N1O1O2O0O2N101N1O2N101N1O101N1O2O0O2N1O2O0O1O2O001N101N101O0O100O10000O10000O100O10000O100O10000O2O000O100O10000O100O10000O10000O100O10000O100O10000O101N10000O10000O100O10000O100O10000O10000O100O10000O100O101O0O10000O100O10000O100O10000O10000O100O10000O100O10001N100000000O10000000000O1000000000000O10000000000O10001O0000O010000000O10O1000O10000000O0100000000O0100000O1000O1000O100000O0100000000O10O100000O1000O10O10000000O0100000000O10O1000O100000000O2O0000000O10001O0O100000001N100000000O2O00000O1000001O0O1000000O2O0000000O101O00000O10001O0O100000000O2O0000000O101O000O1000001O0O100000000O2O00000O10001O000O1000001N100000000O2O0000000O10001O0O100000001N100000000O2O0O100O2N101N1O101N1O2O0O2N101N1O2O0O2N101N1O2O0O1O2O0O2N101N1O2O0O2N101N1O2O0O2N101N1O101N1O2O0O2O0O2N101N1O2O0O2N101N1O101N1O2O0O2N101N1O2O0O2N101N1O2O0O2N100O2N102M2N2O2M2N3N1N3M2O2M2N3N1N3M2N3K4K5L5K4L5J5L5K4K6K4L5K4K6K4L5K4Kg`="}, "label": "pizza on plate with food"}]}
{"id": 36318, "image": "COCO_train2014_000000036318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gourmet chicken pizza\"."}], "task": "refering", "answer_template": "The \"a gourmet chicken pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.1199375, 0.34156908665105384, 0.9490624999999999, 0.9618032786885244], "iscrowd": 0, "area": 102918.87639999998, "rle": {"size": [427, 640], "counts": "c]P1=]<b0]Oc0L4N2M2O2M3N2M3M3N1N3N2M3N2M3M2O2O1N2N2O1N1O2N2O1N2N2O0O2N2N2O1N2N101N2N2N2O1N1O2O1N2N2N2O0O2N2O1N2N2O0O2N2NOZG`Ke8a411O2O0O2N1O2O0O1O2O0O2N1O2O0O2N100O2N1O2O0O2N101N1O1O2O0O2N101N1O2N101N1O101N1O2O0O2N1O2O0O1O2O001N101N101O0O100O10000O10000O100O10000O100O10000O2O000O100O10000O100O10000O10000O100O10000O100O10000O101N10000O10000O100O10000O100O10000O10000O100O10000O100O101O0O10000O100O10000O100O10000O10000O100O10000O100O10001N100000000O10000000000O1000000000000O10000000000O10001O0000O010000000O10O1000O10000000O0100000000O0100000O1000O1000O100000O0100000000O10O100000O1000O10O10000000O0100000000O10O1000O100000000O2O0000000O10001O0O100000001N100000000O2O00000O1000001O0O1000000O2O0000000O101O00000O10001O0O100000000O2O0000000O101O000O1000001O0O100000000O2O00000O10001O000O1000001N100000000O2O0000000O10001O0O100000001N100000000O2O0O100O2N101N1O101N1O2O0O2N101N1O2O0O2N101N1O2O0O1O2O0O2N101N1O2O0O2N101N1O2O0O2N101N1O101N1O2O0O2O0O2N101N1O2O0O2N101N1O101N1O2O0O2N101N1O2O0O2N101N1O2O0O2N100O2N102M2N2O2M2N3N1N3M2O2M2N3N1N3M2N3K4K5L5K4L5J5L5K4K6K4L5K4K6K4L5K4Kg`="}, "label": "a gourmet chicken pizza"}]}
{"id": 429536, "image": "COCO_train2014_000000429536.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman dressed in black sitting with a man in front of a red table with a cake on it\"."}], "task": "refering", "answer_template": "The \"a woman dressed in black sitting with a man in front of a red table with a cake on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 48, 49, 50, 65, 66, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.53764, 0.13258666666666666, 0.8427, 0.67864], "iscrowd": 0, "area": 20921.324349999995, "rle": {"size": [375, 500], "counts": "TgR35_;6J7I7I6J7I7I6J7J4K3M3M3M3M3M3N2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2N2N2N2O100O100O10000O1000O010O10O0100O010O10O0TOULnIk3P6_LgIa3X6iL`IW3]6R1O2N2N2N1O2N2N2N2N2N1O200O100O10O01O100O100O010O100O100O010O100O1O1000000000000000000000O100000000000000001O000001O01O0003M2N3N1N4L5RLgIT2_6cMeI\\2`6]McIb2c6VM`Ij2g7O001O1O100O1O001N2O1N5L4K5K4M4K5L4K5K`1aN<Cifl0"}, "label": "a woman dressed in black sitting with a man in front of a red table with a cake on it"}]}
{"id": 429536, "image": "COCO_train2014_000000429536.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman sitting at a red table taking a slice of a small cake\"."}], "task": "refering", "answer_template": "The \"a young woman sitting at a red table taking a slice of a small cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 48, 49, 50, 65, 66, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.53764, 0.13258666666666666, 0.8427, 0.67864], "iscrowd": 0, "area": 20921.324349999995, "rle": {"size": [375, 500], "counts": "TgR35_;6J7I7I6J7I7I6J7J4K3M3M3M3M3M3N2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2N2N2N2O100O100O10000O1000O010O10O0100O010O10O0TOULnIk3P6_LgIa3X6iL`IW3]6R1O2N2N2N1O2N2N2N2N2N1O200O100O10O01O100O100O010O100O100O010O100O1O1000000000000000000000O100000000000000001O000001O01O0003M2N3N1N4L5RLgIT2_6cMeI\\2`6]McIb2c6VM`Ij2g7O001O1O100O1O001N2O1N5L4K5K4M4K5L4K5K`1aN<Cifl0"}, "label": "a young woman sitting at a red table taking a slice of a small cake"}]}
{"id": 429536, "image": "COCO_train2014_000000429536.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt eating with a woman\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt eating with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 25, 39, 40, 41, 42, 43, 57, 58, 59, 60, 61, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 191, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223], "bbox": [0.00186, 0.07053333333333334, 0.6441800000000001, 0.9888533333333334], "iscrowd": 0, "area": 53352.5491, "rle": {"size": [375, 500], "counts": "Tb0P5f61O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1O100O1O1O100O1O100O1O100O100O100O100O100O100O100O100O101N100000000000000000000000000O1K5M3000000O1000000dNeH\\N[7]1PI_NP7Y1\\IbNe6W1fIdN[6U1PJVNfN2]7c1_JXNe5e1`JWNc5f1X2M3M300O1O1O100O1O100O1O100O1O100O1O100O1O1O1001O1O001OO1O1O2N1O1O11O00001O00001bIQOR2o0mMUOP2k0oMYOn1h0QN[Ol1e0SN_Oj1a0UNCh1>VNFg1:XNJe16[NNa13]N2_1N_N7^1IaN<[1EbNa0Z1_OdNg0X1YOfNl0W1UOfNQ1V1oNjNR1U1nNkNS1T1nNkNS1T1mNlNT1S1lNmNT1S1mNlNU1S1jNmNW1R1iNnNX1Q1iNnNX1Q1hNoNY1P1hNoNZ1o0fNQO[1n0eNSO[1l0fNSO[1l0eNTO\\1k0eNTO]1i0dNWO]1h0cNYO]1e0eNZO\\1e0dNZO^1d0cN[O`1c0aN[Oa1c0`N\\Ob1b0_NXOi1f0XNoNS2o0nMPOU2n0kMROV2l0lMROV2m0jMSOX2k0hMTOZ2k0gMTOZ2k0fMTO]2i0dMWO]2h0dMVO^2i0bMWO`2g0_MYOb2g0^McNZNkNX4a2^MdN^NgNU4X1kNnNPM45CP4Z1nNlNUMN1Im3]1nNkN[MGNMj3`1oNjNXNDi2b1QOgNh2X1ZMcNj2]1XM^Nk2a1f300O100O100O1000000O100000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O010N2N2N2N2N2N2N2N2M3C=00000O100O2N1O1O1O100O1O101O2N8G6K000O101O0O10001O0O10001N101O1N110O00100O010O0010O0100O0010O0jNiEm0W:POlEo0T:POmEk0X:SOkEf0g:I7I6JimP2"}, "label": "a man in a black shirt eating with a woman"}]}
{"id": 429536, "image": "COCO_train2014_000000429536.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in a dark polo shirt and blue jeans seen in profile at a restaurant table\"."}], "task": "refering", "answer_template": "The \"a young man in a dark polo shirt and blue jeans seen in profile at a restaurant table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 25, 39, 40, 41, 42, 43, 57, 58, 59, 60, 61, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 191, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223], "bbox": [0.00186, 0.07053333333333334, 0.6441800000000001, 0.9888533333333334], "iscrowd": 0, "area": 53352.5491, "rle": {"size": [375, 500], "counts": "Tb0P5f61O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1O100O1O1O100O1O100O1O100O100O100O100O100O100O100O100O101N100000000000000000000000000O1K5M3000000O1000000dNeH\\N[7]1PI_NP7Y1\\IbNe6W1fIdN[6U1PJVNfN2]7c1_JXNe5e1`JWNc5f1X2M3M300O1O1O100O1O100O1O100O1O100O1O100O1O1O1001O1O001OO1O1O2N1O1O11O00001O00001bIQOR2o0mMUOP2k0oMYOn1h0QN[Ol1e0SN_Oj1a0UNCh1>VNFg1:XNJe16[NNa13]N2_1N_N7^1IaN<[1EbNa0Z1_OdNg0X1YOfNl0W1UOfNQ1V1oNjNR1U1nNkNS1T1nNkNS1T1mNlNT1S1lNmNT1S1mNlNU1S1jNmNW1R1iNnNX1Q1iNnNX1Q1hNoNY1P1hNoNZ1o0fNQO[1n0eNSO[1l0fNSO[1l0eNTO\\1k0eNTO]1i0dNWO]1h0cNYO]1e0eNZO\\1e0dNZO^1d0cN[O`1c0aN[Oa1c0`N\\Ob1b0_NXOi1f0XNoNS2o0nMPOU2n0kMROV2l0lMROV2m0jMSOX2k0hMTOZ2k0gMTOZ2k0fMTO]2i0dMWO]2h0dMVO^2i0bMWO`2g0_MYOb2g0^McNZNkNX4a2^MdN^NgNU4X1kNnNPM45CP4Z1nNlNUMN1Im3]1nNkN[MGNMj3`1oNjNXNDi2b1QOgNh2X1ZMcNj2]1XM^Nk2a1f300O100O100O1000000O100000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O010N2N2N2N2N2N2N2N2M3C=00000O100O2N1O1O1O100O1O101O2N8G6K000O101O0O10001O0O10001N101O1N110O00100O010O0010O0100O0010O0jNiEm0W:POlEo0T:POmEk0X:SOkEf0g:I7I6JimP2"}, "label": "a young man in a dark polo shirt and blue jeans seen in profile at a restaurant table"}]}
{"id": 298479, "image": "COCO_train2014_000000298479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat head on right\"."}], "task": "refering", "answer_template": "The \"cat head on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 129, 130, 131], "bbox": [0.5166425120772946, 0.19935251798561152, 0.8133333333333332, 0.8308273381294965], "iscrowd": 0, "area": 14677.41005, "rle": {"size": [278, 414], "counts": "lTj1?g0[18ROl20XL[1b0nNR3W2cLmM]3g31N2N3N1N3M2N3M00000010O0000000001O00000O1001O00001O00001O00001O00001O0000001O01O0001O000001O01O000000010O000000010O01O1O010O001O0010O01O0010O01O001O001O001O000000O0100000000O100O1N2M33MO1fNeKXN\\4e1mKTNT4HPLj0^5TOgJb0`5]OdJ<`5CcJ6b5IfJK_54c1O0O100O100O100O100O1O100O1O2O0O100MUjd0"}, "label": "cat head on right"}]}
{"id": 298479, "image": "COCO_train2014_000000298479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat lying in a bed with her baby\"."}], "task": "refering", "answer_template": "The \"a cat lying in a bed with her baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 129, 130, 131], "bbox": [0.5166425120772946, 0.19935251798561152, 0.8133333333333332, 0.8308273381294965], "iscrowd": 0, "area": 14677.41005, "rle": {"size": [278, 414], "counts": "lTj1?g0[18ROl20XL[1b0nNR3W2cLmM]3g31N2N3N1N3M2N3M00000010O0000000001O00000O1001O00001O00001O00001O00001O0000001O01O0001O000001O01O000000010O000000010O01O1O010O001O0010O01O0010O01O001O001O001O000000O0100000000O100O1N2M33MO1fNeKXN\\4e1mKTNT4HPLj0^5TOgJb0`5]OdJ<`5CcJ6b5IfJK_54c1O0O100O100O100O100O1O100O1O2O0O100MUjd0"}, "label": "a cat lying in a bed with her baby"}]}
{"id": 265713, "image": "COCO_train2014_000000265713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bunch of two bananas near oranges\"."}], "task": "refering", "answer_template": "The \"bunch of two bananas near oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 207, 208, 220, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 294, 295, 296], "bbox": [0.23512, 0.6194400000000001, 0.59796, 0.90028], "iscrowd": 0, "area": 12919.420450000001, "rle": {"size": [500, 500], "counts": "Rni17]?7I7I3M2N2N3M2N2N4N00O2O1N3N3L01O1O1O100O1O1O1O101N8H2O1N2M2N100O2O0O1000001O000O101N1O1O1O2N00001O001O00001O00001O00010O001O00001O000O2O00001O00001O001O0O1000000000000000000O01000O100O100O10000O100O100O10O10O100O100O010O010O1O010O00100O0010O01O10O01O001O001O1O001O001O00001O0000000O101O00000O10001O000001O000000000O10000O1000000O10001O1N2N2M3M3N2M3iNmBJV=4mBIU=4PCGW=2mBJ\\=@UC;V>EfiQ3"}, "label": "bunch of two bananas near oranges"}]}
{"id": 265713, "image": "COCO_train2014_000000265713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two bananas connected at the stalk and covered with brown blemishes sit in a bowl\"."}], "task": "refering", "answer_template": "The \"two bananas connected at the stalk and covered with brown blemishes sit in a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 207, 208, 220, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 294, 295, 296], "bbox": [0.23512, 0.6194400000000001, 0.59796, 0.90028], "iscrowd": 0, "area": 12919.420450000001, "rle": {"size": [500, 500], "counts": "Rni17]?7I7I3M2N2N3M2N2N4N00O2O1N3N3L01O1O1O100O1O1O1O101N8H2O1N2M2N100O2O0O1000001O000O101N1O1O1O2N00001O001O00001O00001O00010O001O00001O000O2O00001O00001O001O0O1000000000000000000O01000O100O100O10000O100O100O10O10O100O100O010O010O1O010O00100O0010O01O10O01O001O001O1O001O001O00001O0000000O101O00000O10001O000001O000000000O10000O1000000O10001O1N2N2M3M3N2M3iNmBJV=4mBIU=4PCGW=2mBJ\\=@UC;V>EfiQ3"}, "label": "two bananas connected at the stalk and covered with brown blemishes sit in a bowl"}]}
{"id": 265713, "image": "COCO_train2014_000000265713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bananas above the clementines\"."}], "task": "refering", "answer_template": "The \"bananas above the clementines\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132], "bbox": [0.2281, 0.1354, 0.64742, 0.42328], "iscrowd": 0, "area": 18462.088549999997, "rle": {"size": [500, 500], "counts": "_kg15]?6J5K2O1O1N2O1O1O1N101O100O1O100O1O10O01O100O1O10000O10O0100O100O10000O010O100N2K5L3L5N101N2N1O2XOYNaCh1]<[NbCf1[<]NcCe1Z<^NdCc1Z<`NeCa1W<cNhC]1X<dNgC]1W<eNhC\\1W<eNhC[1W<gNhCZ1W<gNhCY1X<hNhCX1V<jNiCV1W<kNhCV1W<kNhCV1W<kNiCT1W<U1O010O1O00100O001O010O10O010O0100O0100O0100O010O10O010O10O10O10O10O1000O010000O010000O1000000O01000000000000000000000000O10000000000000000000000000000000000000001O00001O00000010O00010O00010O000010O00010O00010O0001O01O010O1O00100O00100O1O00100O00100O1O1O100O2N100O1O101fN\\C@e<=aC_O`<?gCZO\\<b0lCXOV<e0RDSOQ<i0`1M3M4K4K5^Oije2"}, "label": "bananas above the clementines"}]}
{"id": 265713, "image": "COCO_train2014_000000265713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3 slightly bruised bananas lay next to some oranges followed by more bananas\"."}], "task": "refering", "answer_template": "The \"3 slightly bruised bananas lay next to some oranges followed by more bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132], "bbox": [0.2281, 0.1354, 0.64742, 0.42328], "iscrowd": 0, "area": 18462.088549999997, "rle": {"size": [500, 500], "counts": "_kg15]?6J5K2O1O1N2O1O1O1N101O100O1O100O1O10O01O100O1O10000O10O0100O100O10000O010O100N2K5L3L5N101N2N1O2XOYNaCh1]<[NbCf1[<]NcCe1Z<^NdCc1Z<`NeCa1W<cNhC]1X<dNgC]1W<eNhC\\1W<eNhC[1W<gNhCZ1W<gNhCY1X<hNhCX1V<jNiCV1W<kNhCV1W<kNhCV1W<kNiCT1W<U1O010O1O00100O001O010O10O010O0100O0100O0100O010O10O010O10O10O10O10O1000O010000O010000O1000000O01000000000000000000000000O10000000000000000000000000000000000000001O00001O00000010O00010O00010O000010O00010O00010O0001O01O010O1O00100O00100O1O00100O00100O1O1O100O2N100O1O101fN\\C@e<=aC_O`<?gCZO\\<b0lCXOV<e0RDSOQ<i0`1M3M4K4K5^Oije2"}, "label": "3 slightly bruised bananas lay next to some oranges followed by more bananas"}]}
{"id": 19959, "image": "COCO_train2014_000000019959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"triangle piece of pizza\"."}], "task": "refering", "answer_template": "The \"triangle piece of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 72, 73, 74, 75, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 271, 272, 273, 274, 275, 276, 294, 295], "bbox": [0.11090322580645162, 0.1438222222222222, 0.8709677419354839, 0.8382000000000002], "iscrowd": 0, "area": 73705.73490000001, "rle": {"size": [450, 620], "counts": "Vcn02o=1N3N1N3N2M2O2M3N1N3N1O2M3N1N3N2M2O2M2O2M3N1N3N2M2O2N1N3N2M2O2M3N1N3N1N3N2M2O2N1N3N2M2O2M3N1N3N1N3N2M2O2N2M2O1N2O2M2O1N3N1N2O2M2O1N2O2N1N2O2M2O1N2O2M2O1N3N1N2O2N1N2O1N3N1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O1N3N1N2O2M2O1N3N1O1N2O2M2O1N3N1N2O1N3N1N2O2N1N2O2M2O1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O2M2O1N2O2M2O1N3N1O1100O01000000000000O1000000O1000000O10000O1000000O1000000O100000O0O2O1O1N101N2O0O2O1N101O001O1N101O001O1N101O001O1O0O2O001O1O0O2O001O1N101N101N2O001N101N2O0O2O0O2O1N101N101N2O001N101N2O0O2O0O2O1N101N2O0O2O001N2O0O2O0O2O1N101N101N2O0O2O001N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101O1N101N101N2O0O2O0O2O1N101N101N2O001N101O1O0O2O001O1N101O001O1N101O001N2O001O001N2O001O1N101O001O1N101O001N2O001O001N2O001O0O2O1O001O0O2O1O001N101O1O001N101O1O0O2O001O1O0O2O001O1N101O001N2O001N101N2O0O101N101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101N10001N2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O0OilR1"}, "label": "triangle piece of pizza"}]}
{"id": 19959, "image": "COCO_train2014_000000019959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pizza with strawberries\"."}], "task": "refering", "answer_template": "The \"a slice of pizza with strawberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 72, 73, 74, 75, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 271, 272, 273, 274, 275, 276, 294, 295], "bbox": [0.11090322580645162, 0.1438222222222222, 0.8709677419354839, 0.8382000000000002], "iscrowd": 0, "area": 73705.73490000001, "rle": {"size": [450, 620], "counts": "Vcn02o=1N3N1N3N2M2O2M3N1N3N1O2M3N1N3N2M2O2M2O2M3N1N3N2M2O2N1N3N2M2O2M3N1N3N1N3N2M2O2N1N3N2M2O2M3N1N3N1N3N2M2O2N2M2O1N2O2M2O1N3N1N2O2M2O1N2O2N1N2O2M2O1N2O2M2O1N3N1N2O2N1N2O1N3N1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O1N3N1N2O2M2O1N3N1O1N2O2M2O1N3N1N2O1N3N1N2O2N1N2O2M2O1N2O2M2O1N3N1N2O1O2M2O1N3N1N2O2M2O1N2O2M2O1N3N1O1100O01000000000000O1000000O1000000O10000O1000000O1000000O100000O0O2O1O1N101N2O0O2O1N101O001O1N101O001O1N101O001O1O0O2O001O1O0O2O001O1N101N101N2O001N101N2O0O2O0O2O1N101N101N2O001N101N2O0O2O0O2O1N101N2O0O2O001N2O0O2O0O2O1N101N101N2O0O2O001N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101O1N101N101N2O0O2O0O2O1N101N101N2O001N101O1O0O2O001O1N101O001O1N101O001N2O001O001N2O001O1N101O001O1N101O001N2O001O001N2O001O0O2O1O001O0O2O1O001N101O1O001N101O1O0O2O001O1O0O2O001O1N101O001N2O001N101N2O0O101N101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O101N101N101N101N100O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101N10001N2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O0OilR1"}, "label": "a slice of pizza with strawberries"}]}
{"id": 19959, "image": "COCO_train2014_000000019959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the largest slice of pizza\"."}], "task": "refering", "answer_template": "The \"the largest slice of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 194, 195], "bbox": [0.3611129032258064, 0.00022222222222222223, 1.0, 0.5233777777777777], "iscrowd": 0, "area": 58582.128000000004, "rle": {"size": [450, 620], "counts": "P^R31Q>1O1O1O1O1O1O1O1O1O001O001O001O1O001O1O1O1O1O1O2N2N2N2N3M3M1O001O001O00001O00001O001O00001O001O001O1O001O001O1O001O1O001O1O1O001O1O8H1O001O001O001O001O001O001O001O00001O001O001O00001O001O2N1O2N1O1O2N2N2N2N2N2N2N2N2N2N2N1O1O1O00001O001O00001O001O001O001O1O00001O00001O00001O001O001O001O2N3M2N2N1O1O1O1O1O1O1O1O0000001O0000001O0000001O1O1O1O1O1O1O1O1O1O2N1O1O1O001O1O001O2N1O2N2N001O001O1O001O002N1O1O1O1O1O1O2N1O2N1O001O1O1O1O001O1O1O1O1O001O001O001O0000001O00001O0000001O00001O00001O001O1O1O001O1O001O1O1O001O0000001O00001O00001O001O001O1O1O1O001O1O1O001O001O001O001O1O001O1O001O1O001O00001O001O00001O00001O00001O0000001O0000001O0000000000000000000000O1O1O1O1O1O1O1O100O1O10000O10000O10000O10000O10000O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O`0@a1_Na1]Nc1]NTK"}, "label": "the largest slice of pizza"}]}
{"id": 52729, "image": "COCO_train2014_000000052729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skiies of the italian skiier in the blue outfit\"."}], "task": "refering", "answer_template": "The \"the skiies of the italian skiier in the blue outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [281, 282, 283, 293, 294, 297, 298, 299, 309, 310, 311, 325, 326, 327, 341, 342], "bbox": [0.5624824355971897, 0.8, 0.9801405152224824, 0.9887656249999999], "iscrowd": 0, "area": 3352.029849999992, "rle": {"size": [640, 427], "counts": "_`f47ic01000O10O100000000O100000O0100000000O100000O0100000000O1000000O010000000O10000O1O010O1O100O1O100O1O10O01O100O1O100O1O10O01O100O1O100O1O100O00100O10000000000001N10000000000000001N10000000eMZO[Af0`>^OaAb0^>_ObAa0^>^OcAb0\\>^OeAb0Z>_OgA?Y>AhA?X>@iA`0V>AjA?U>AlA?T>AlA?S>AnA?Q>APB?o=BQB>o=ASB=m=CTB=k=DUB<j=DWB<i=DWB;i=EXB9i=GXB8i=HWB7i=IXB6h=JZB4f=MZB2g=MZB2f=OZBOg=1ZBNf=2[BMf=3ZBLf=4[BKe=5\\BJe=6[BHf=8\\BFd=;\\BDd=<Q3O01O00010O001O01O01O01O01O00010O001O01O01O01O01O0010O0001O01O01O010O02N101N1O2O0O2O0O2N100O2N10iR5"}, "label": "the skiies of the italian skiier in the blue outfit"}]}
{"id": 52729, "image": "COCO_train2014_000000052729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and yellow skis on the feet of a person in blue\"."}], "task": "refering", "answer_template": "The \"black and yellow skis on the feet of a person in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [281, 282, 283, 293, 294, 297, 298, 299, 309, 310, 311, 325, 326, 327, 341, 342], "bbox": [0.5624824355971897, 0.8, 0.9801405152224824, 0.9887656249999999], "iscrowd": 0, "area": 3352.029849999992, "rle": {"size": [640, 427], "counts": "_`f47ic01000O10O100000000O100000O0100000000O100000O0100000000O1000000O010000000O10000O1O010O1O100O1O100O1O10O01O100O1O100O1O10O01O100O1O100O1O100O00100O10000000000001N10000000000000001N10000000eMZO[Af0`>^OaAb0^>_ObAa0^>^OcAb0\\>^OeAb0Z>_OgA?Y>AhA?X>@iA`0V>AjA?U>AlA?T>AlA?S>AnA?Q>APB?o=BQB>o=ASB=m=CTB=k=DUB<j=DWB<i=DWB;i=EXB9i=GXB8i=HWB7i=IXB6h=JZB4f=MZB2g=MZB2f=OZBOg=1ZBNf=2[BMf=3ZBLf=4[BKe=5\\BJe=6[BHf=8\\BFd=;\\BDd=<Q3O01O00010O001O01O01O01O01O00010O001O01O01O01O01O0010O0001O01O01O010O02N101N1O2O0O2O0O2N100O2N10iR5"}, "label": "black and yellow skis on the feet of a person in blue"}]}
{"id": 52729, "image": "COCO_train2014_000000052729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis that man in red is wearing\"."}], "task": "refering", "answer_template": "The \"skis that man in red is wearing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 199, 200, 201, 211, 212, 217, 218, 219, 227], "bbox": [0.045011709601873534, 0.5757656250000001, 0.6479156908665105, 0.681390625], "iscrowd": 0, "area": 1912.5026999999993, "rle": {"size": [640, 427], "counts": "mW<1PX10o[O1O1O1O10O01O100O100O10O10O100O1000O0100O10000O010O100O10000O02O001OO1000O10O10000O1000O10O10000O1000000O100000000001O000N2O1N2O1N2O1NWjj0OjUUO1S\\O0ic04O1O101O000000001O000000001O0000000010O00000010O000001O01O0001O00010O00000010O000001O01O0001O00010O00000010O01O1O100O1O010O1O100O1O100OPT80PlG2O1N2N1O2N2O0O00000001O0000001O01O01O0001O01O00010O0001O0000001O00001O0000001O00001O01O0010O00010O0001O010O00010O001O01O0\\[m2"}, "label": "skis that man in red is wearing"}]}
{"id": 52729, "image": "COCO_train2014_000000052729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis of the man in the middle\"."}], "task": "refering", "answer_template": "The \"the skis of the man in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 246, 247, 248, 249, 259, 264, 265, 266, 274, 275, 281], "bbox": [0.2612177985948478, 0.684828125, 0.8189695550351288, 0.817609375], "iscrowd": 0, "area": 1613.5129500000012, "rle": {"size": [640, 427], "counts": "h^V21oc02N2N2N2N2O2M2N2N2M3N2O11O0000001O00001O0000001N1N2O2M2N2N2N100O1O1Obeg0LaZXO2O1O1O1N3N10O010O00010O01O01O010O01O01O010O01O01O01O010O01O01O010O01O010O010O0010O010O01O010O010O0010O010O01O01O010O01O010O010O0010O010O01O01N101O000O2O001N1011N6J6KW\\c03bc\\O3N30O0010O010O010O01O010O010O00010O010O010O0010O010O01O01O010O010O01O010N10001O0O10000O100O1O10]X`1"}, "label": "the skis of the man in the middle"}]}
{"id": 52729, "image": "COCO_train2014_000000052729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier in the middle skis\"."}], "task": "refering", "answer_template": "The \"the skier in the middle skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 246, 247, 248, 249, 259, 264, 265, 266, 274, 275, 281], "bbox": [0.2612177985948478, 0.684828125, 0.8189695550351288, 0.817609375], "iscrowd": 0, "area": 1613.5129500000012, "rle": {"size": [640, 427], "counts": "h^V21oc02N2N2N2N2O2M2N2N2M3N2O11O0000001O00001O0000001N1N2O2M2N2N2N100O1O1Obeg0LaZXO2O1O1O1N3N10O010O00010O01O01O010O01O01O010O01O01O01O010O01O01O010O01O010O010O0010O010O01O010O010O0010O010O01O01O010O01O010O010O0010O010O01O01N101O000O2O001N1011N6J6KW\\c03bc\\O3N30O0010O010O010O01O010O010O00010O010O010O0010O010O01O01O010O010O01O010N10001O0O10000O100O1O10]X`1"}, "label": "the skier in the middle skis"}]}
{"id": 101882, "image": "COCO_train2014_000000101882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike being pushed in the garage\"."}], "task": "refering", "answer_template": "The \"the bike being pushed in the garage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 260, 261, 262, 263, 264, 283, 284, 285, 286, 287, 307, 308, 309, 330, 331, 332, 353, 354], "bbox": [0.33934375, 0.555296442687747, 0.50103125, 0.874407114624506], "iscrowd": 0, "area": 10594.487099999998, "rle": {"size": [506, 640], "counts": "Xb[35Y?`0A?@`0K5L4L4K5L4L4L4L4K3N1O2N2M3K4L5L4K4M4K5K5M2O2N2M3N1O2M3N1O2N2M3N1O2N0O100000O01000000O01000O1000O10O10L4L4L3M4M3L4L3M4L4N3M3N1O2M2O2N1N3N2N1N3N1O2M4M3M3M3L3N3M3L4M3M3N2M3N2M2O2M3M3N2M3N2K5J5L5J6K5J6Kaim4"}, "label": "the bike being pushed in the garage"}]}
{"id": 101882, "image": "COCO_train2014_000000101882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white color scooty , pushed by a girl\"."}], "task": "refering", "answer_template": "The \"a white color scooty , pushed by a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 260, 261, 262, 263, 264, 283, 284, 285, 286, 287, 307, 308, 309, 330, 331, 332, 353, 354], "bbox": [0.33934375, 0.555296442687747, 0.50103125, 0.874407114624506], "iscrowd": 0, "area": 10594.487099999998, "rle": {"size": [506, 640], "counts": "Xb[35Y?`0A?@`0K5L4L4K5L4L4L4L4K3N1O2N2M3K4L5L4K4M4K5K5M2O2N2M3N1O2M3N1O2N2M3N1O2N0O100000O01000000O01000O1000O10O10L4L4L3M4M3L4L3M4L4N3M3N1O2M2O2N1N3N2N1N3N1O2M4M3M3M3L3N3M3L4M3M3N2M3N2M2O2M3M3N2M3N2K5J5L5J6K5J6Kaim4"}, "label": "a white color scooty , pushed by a girl"}]}
{"id": 101882, "image": "COCO_train2014_000000101882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wheeling their bicycle up a ramp into a small garage ,\"."}], "task": "refering", "answer_template": "The \"a person wheeling their bicycle up a ramp into a small garage ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 282, 283, 304, 305, 306, 327, 328, 329, 350, 351, 373, 374], "bbox": [0.230453125, 0.4442094861660079, 0.462671875, 0.917292490118577], "iscrowd": 0, "area": 10731.821199999997, "rle": {"size": [506, 640], "counts": "cbY24b?4L4L4L4M3L5L3L4ZOf0L4XObNPCb1g<m0O1N2O1O1N2O1N2O1N2KZNQDKa;a0fDlN_O@h;f1oD\\NJFT;P2XElMe;W2m01N3M2M4K4YO[MVDj2d;hMkC\\2P<j0J6M3N2O1N2O1O1N2O1N2O1N2O1O001O1N200000000O100000O10O0100O010O1TNlDQNT<m1SDhMT<U2SD`MP<a2c0J72M3N20O0100O100O0100O101N100O100O2N1O1fMgB;Om0m=oNUBf0X>WOjAh0X>TOjAl0X>POkAo0`>0001O001O001O00001O000000000000000000000000000000000001O01O00010O0001O01O01O01O0O1O1O1O1O1O1N2O1N3N3L3M4KXgY5"}, "label": "a person wheeling their bicycle up a ramp into a small garage ,"}]}
{"id": 101882, "image": "COCO_train2014_000000101882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a scarf and skirt , rolling a scooter up a narrow ramp into a small storage facility\"."}], "task": "refering", "answer_template": "The \"a person wearing a scarf and skirt , rolling a scooter up a narrow ramp into a small storage facility\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 282, 283, 304, 305, 306, 327, 328, 329, 350, 351, 373, 374], "bbox": [0.230453125, 0.4442094861660079, 0.462671875, 0.917292490118577], "iscrowd": 0, "area": 10731.821199999997, "rle": {"size": [506, 640], "counts": "cbY24b?4L4L4L4M3L5L3L4ZOf0L4XObNPCb1g<m0O1N2O1O1N2O1N2O1N2KZNQDKa;a0fDlN_O@h;f1oD\\NJFT;P2XElMe;W2m01N3M2M4K4YO[MVDj2d;hMkC\\2P<j0J6M3N2O1N2O1O1N2O1N2O1N2O1O001O1N200000000O100000O10O0100O010O1TNlDQNT<m1SDhMT<U2SD`MP<a2c0J72M3N20O0100O100O0100O101N100O100O2N1O1fMgB;Om0m=oNUBf0X>WOjAh0X>TOjAl0X>POkAo0`>0001O001O001O00001O000000000000000000000000000000000001O01O00010O0001O01O01O01O0O1O1O1O1O1O1N2O1N3N3L3M4KXgY5"}, "label": "a person wearing a scarf and skirt , rolling a scooter up a narrow ramp into a small storage facility"}]}
{"id": 101882, "image": "COCO_train2014_000000101882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown vest standing outside of a garage\"."}], "task": "refering", "answer_template": "The \"a man in a brown vest standing outside of a garage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325, 346, 347, 348, 369, 370, 371, 392, 393, 394], "bbox": [0.058499999999999996, 0.45055335968379445, 0.189671875, 0.9886758893280632], "iscrowd": 0, "area": 12969.644, "rle": {"size": [506, 640], "counts": "Vcb01f?>A?B=C:UDjNa8]1XGWOS8Q1eG_Ol7i0kG\\OQ8k0gGZOU8n0bGWOY8Q1`GSO\\8T1\\GQO`8W1WGnNe8Y1SGlNh8\\1SGeNj8X4L4M3M3L3N2N2N2N2N1O2O1O1O1N2O1O1O1N101O1N2O1O1O1N2O001O1N200000L4H8I7H8I7H8]L_HfNh7U1cI]Me6_2]3K5L4K4M4K5L4K5K5L4K5M3N2N3M3M3M4L3M3M3M3M4L3N2M3M4L3M3MRWP8"}, "label": "a man in a brown vest standing outside of a garage"}]}
{"id": 101882, "image": "COCO_train2014_000000101882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the orange vest\"."}], "task": "refering", "answer_template": "The \"the man in the orange vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325, 346, 347, 348, 369, 370, 371, 392, 393, 394], "bbox": [0.058499999999999996, 0.45055335968379445, 0.189671875, 0.9886758893280632], "iscrowd": 0, "area": 12969.644, "rle": {"size": [506, 640], "counts": "Vcb01f?>A?B=C:UDjNa8]1XGWOS8Q1eG_Ol7i0kG\\OQ8k0gGZOU8n0bGWOY8Q1`GSO\\8T1\\GQO`8W1WGnNe8Y1SGlNh8\\1SGeNj8X4L4M3M3L3N2N2N2N2N1O2O1O1O1N2O1O1O1N101O1N2O1O1O1N2O001O1N200000L4H8I7H8I7H8]L_HfNh7U1cI]Me6_2]3K5L4K4M4K5L4K5K5L4K5M3N2N3M3M3M4L3M3M3M3M4L3N2M3M4L3M3MRWP8"}, "label": "the man in the orange vest"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair behind the donuts\"."}], "task": "refering", "answer_template": "The \"chair behind the donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 22, 23, 24, 25, 26, 39, 40, 41, 42, 43, 44, 45, 57, 58, 59, 60, 75, 76, 77, 78, 93, 94, 95, 96, 112, 113, 114], "bbox": [0.18414, 0.0, 0.52144, 0.5842342342342343], "iscrowd": 0, "area": 16499.019650000002, "rle": {"size": [333, 500], "counts": "[om08T:a0_O8I4K4L4L4M3L4L5L3L8H7J7H8H7J7H8H7J7H8I2M100`NeKnL[4R3gKkLZ4U3iKhLW4X3kKeLV4[3lKcLT4]3nK`LS4`3oK^LR4a3PL\\LQ4d3QLZLo3f3TLWLl3i3VLTLk3l3WLQLj3o3XLnKi3R4YLkKh3U4ZLhKg3X4\\LdKe3]4V1O0000000000000000001O00000000000000001O0000000000000000001O000001O000000000WOgK^KY4a4mK[KR4d4SLXKm3X2jKE<oMk3\\2ULXO2[Ni3\\2aLmNIeNf3^2WMVNVOZOc3_2[MTNSO\\Ob3`2^MPNSO^O_3a2bMmM^3S2eLjM[3U2iLgMX3Y2iLdMY3[2]200O100\\OSHgNn7X1THgNl7X1UHgNl7X1VHgNj7X1WHgNj7X1XHgNh7X1YHgNh7X1ZHgNf7X1[HgNf7X1\\HgNd7X1]HgNd7W1_HhNa7W1`HhNa7W1aHhN_7W1bHhN_7W1cHhN]7W1cHjN]7V1bHkN^7T1cHlN]7T1bHmN^7R1bHoN^75nG<c0@_71QH>`0B_7NTH?<D`7LVH>:G`7KVH>9Ha7JWH<8Ka7IWH<8Ka7IXH:7Na7HYH950b7GYH852b7FZH743b7F[H536b7E[H536b7E[H447a7E\\H329b7D\\H23:a7D\\H23:a7D\\H14;`7D]H02=a7C]HO3>`7C]HO3>`7C]HN3`0`7B^HM2a0`7B^HL3b0_7B^HL2c0`7A^HK3d0_7A^HK3d0_7AVI?j6AUI`0k6@UI`0k6@UI`0k6ATI?l6ATI?l6ATI?m6@SI`0o6_OPI`0R7_OnHa0S7^OmHa0V7]OjHc0`8O2O000O100O2N1O2N2N1O2N3M2NYe]2"}, "label": "chair behind the donuts"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the headrest seen behind the donuts\"."}], "task": "refering", "answer_template": "The \"the headrest seen behind the donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 22, 23, 24, 25, 26, 39, 40, 41, 42, 43, 44, 45, 57, 58, 59, 60, 75, 76, 77, 78, 93, 94, 95, 96, 112, 113, 114], "bbox": [0.18414, 0.0, 0.52144, 0.5842342342342343], "iscrowd": 0, "area": 16499.019650000002, "rle": {"size": [333, 500], "counts": "[om08T:a0_O8I4K4L4L4M3L4L5L3L8H7J7H8H7J7H8H7J7H8I2M100`NeKnL[4R3gKkLZ4U3iKhLW4X3kKeLV4[3lKcLT4]3nK`LS4`3oK^LR4a3PL\\LQ4d3QLZLo3f3TLWLl3i3VLTLk3l3WLQLj3o3XLnKi3R4YLkKh3U4ZLhKg3X4\\LdKe3]4V1O0000000000000000001O00000000000000001O0000000000000000001O000001O000000000WOgK^KY4a4mK[KR4d4SLXKm3X2jKE<oMk3\\2ULXO2[Ni3\\2aLmNIeNf3^2WMVNVOZOc3_2[MTNSO\\Ob3`2^MPNSO^O_3a2bMmM^3S2eLjM[3U2iLgMX3Y2iLdMY3[2]200O100\\OSHgNn7X1THgNl7X1UHgNl7X1VHgNj7X1WHgNj7X1XHgNh7X1YHgNh7X1ZHgNf7X1[HgNf7X1\\HgNd7X1]HgNd7W1_HhNa7W1`HhNa7W1aHhN_7W1bHhN_7W1cHhN]7W1cHjN]7V1bHkN^7T1cHlN]7T1bHmN^7R1bHoN^75nG<c0@_71QH>`0B_7NTH?<D`7LVH>:G`7KVH>9Ha7JWH<8Ka7IWH<8Ka7IXH:7Na7HYH950b7GYH852b7FZH743b7F[H536b7E[H536b7E[H447a7E\\H329b7D\\H23:a7D\\H23:a7D\\H14;`7D]H02=a7C]HO3>`7C]HO3>`7C]HN3`0`7B^HM2a0`7B^HL3b0_7B^HL2c0`7A^HK3d0_7A^HK3d0_7AVI?j6AUI`0k6@UI`0k6@UI`0k6ATI?l6ATI?l6ATI?m6@SI`0o6_OPI`0R7_OnHa0S7^OmHa0V7]OjHc0`8O2O000O100O2N1O2N2N1O2N3M2NYe]2"}, "label": "the headrest seen behind the donuts"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom of the two pastries\"."}], "task": "refering", "answer_template": "The \"the bottom of the two pastries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 169, 170, 171], "bbox": [0.33418000000000003, 0.43711711711711715, 0.5915199999999999, 0.7802102102102102], "iscrowd": 0, "area": 10711.488349999998, "rle": {"size": [333, 500], "counts": "njf12[:1kNOlG3P81oGOo73oGOn74PHNn74PHOk75THNh74mG_OH?X85oG^OF>Z87mG^OF=[88mG0R8W1O0O2O0O2O0O1O1000001OO01000O_HiMm6X2QIjMo6X2kHmMS7g2O2O1N1O2O0O2O000O2O000O10001N10000O2O000O10001N1000000000000000000000000000O1000001O00000000001N1000000O2O0000000O2O001O1O0O2O000O2O001N100lNQIiNQ7V1PIiNQ7V1oHjNR7U1oHiNS7V1mHjNT7U1mHiNU7V1kHjNV7U1kHjNV7U1kHiNW7V1iHjNX7U1iHiNY7V1gHjNZ7U1fHjN]7T1cHkN_7T1aHjNb7U1^HiNd7W1j000O1O1O1O002N2L4M4L3M:@oTR2"}, "label": "the bottom of the two pastries"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partially eaten donut\"."}], "task": "refering", "answer_template": "The \"a partially eaten donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 169, 170, 171], "bbox": [0.33418000000000003, 0.43711711711711715, 0.5915199999999999, 0.7802102102102102], "iscrowd": 0, "area": 10711.488349999998, "rle": {"size": [333, 500], "counts": "njf12[:1kNOlG3P81oGOo73oGOn74PHNn74PHOk75THNh74mG_OH?X85oG^OF>Z87mG^OF=[88mG0R8W1O0O2O0O2O0O1O1000001OO01000O_HiMm6X2QIjMo6X2kHmMS7g2O2O1N1O2O0O2O000O2O000O10001N10000O2O000O10001N1000000000000000000000000000O1000001O00000000001N1000000O2O0000000O2O001O1O0O2O000O2O001N100lNQIiNQ7V1PIiNQ7V1oHjNR7U1oHiNS7V1mHjNT7U1mHiNU7V1kHjNV7U1kHjNV7U1kHiNW7V1iHjNX7U1iHiNY7V1gHjNZ7U1fHjN]7T1cHkN_7T1aHjNb7U1^HiNd7W1j000O1O1O1O002N2L4M4L3M:@oTR2"}, "label": "a partially eaten donut"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in blue pants grabbing donut\"."}], "task": "refering", "answer_template": "The \"person in blue pants grabbing donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 25, 26, 27, 28, 29, 33, 34, 35, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 87, 88, 89, 99, 100, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215], "bbox": [0.38186000000000003, 0.0016816816816816818, 1.0, 0.9858858858858859], "iscrowd": 0, "area": 40147.970149999994, "rle": {"size": [333, 500], "counts": "_Un11\\:1N2N3M2N2O1N2N1O01OmNDVH=g7GWH:g7JVH6j7LTH5k7MTH2m7OQH2o7NPH2Q8MoG4Q8LoG3R8MmG4S8LlG4T8MkG4U8LjG4W8LiG4W8LhG5X8KgG6Y8JfG6[8JeG6[8JdG7\\8JbG7^8IaG7Z1^O\\5;ZI7X1@^59YI7X1A`58WI7X1Ca57UI7W1Ec55UI6W1Fe54TI6U1Hg52SI6U1Ii51QI7S1Kl5OoH6T1Ln5NnH6R1NQ6KlH7R1OS6d0kI^OU6a0iIAY6=fIEZ6;dIF]69bII^67_IKc63\\IOd61\\IOc61]IOd61\\IO[6POQIQ1e0N[6;cIF]6:cIE^6<aID_6<`ICb6=^IAd6`0[I_Of6a0YI^Oi6c0VI\\Ok6d0UIZOm6g0RIWOP7i0oHWOR7i0nHUOT7l0kHROW7n0hHROQ1D`4[1^JQOo0Hc4V1^JROn0Id4V1\\JROo0Je4S1\\JSOn0Kf4S1[JROo0Kg4R1YJTOP1Kf4Q1ZJTOP1Kg4Q1TJYOT1Gg4o1ZKQNf4o1ZKQNg4n1YKSNf4m1YKTNh4j1XKXNg4h1XKYNi4f1VK[Nj4d1WK[Nk4d1TK]Nm4b1RK^NP5a1oJ`NS5]1mJcNU5\\1jJeNW5Z1hJfNZ5Y1eJhN\\5V1dJjN^5U1aJlNa5R1_JmNc5Q1]JPOd5o0[JQOg5n0XJSOi5l0WJSOl5k0SJUOP6i0oIXOR6g0mIYOV6e0jI[OW6e0gI[O\\6c0cI^O^6a0aI_Oa6`0]IBd6=WIGj69PIMQ73hH2Z7`11N2O0O2O1N2O0O2O1N101N100O2O0O2O00001O001O00001O001O00001O001O00001O00eJiM`2W2^MkMc2T2[MnMe2R2YMPNh2o1VMTNi2l1UMVNk2i1SMZNn2e1PM]NP3b1oL`NR3^1mLdNS3\\1kLfNV3X1iLjNW3V1gLlNY3S1fLoN[3o0dLSO\\3m0bLUO_3i0`LYO`3g0^L[Ob3d0]L^Od3`0[LBiNhNV4f1oLDjNhNW4b1oLGjNgNW4a1PMHhNiNW4_1QMHhNjNV4]1RMJgNkNU4\\1TMIgNPOP4V1ZMJeNVOk3P1`MLdNYOg3k0eMLdNE[3>RNMbN9h2KeNMcN<d2FjNO`N>d2BmN0_N`0a2AoN0_Nb0a2\\OPO3`Nb0_2ZOQO5_Nc0^2XORO6`Nd0]2UORO8aNd0\\2TORO9bNd0Z2TOTO8bNe0Y2SOTO:bNd0Y2ROVO9aNf0X2QOWO9bNg0U2POYO9bNh0S2PO[O8bNi0R2oN]O6bNl0P2nN^O6bNm0o1mN_O6bNn0n1lN@6bNo0m1kNB5aNQ1k1kND4bNQ1i1kNE4bNR1h1jNF4bNS1g1iNH5_NS1h1hNI6^NT1g1fNK8\\NR1h1gNL8[NQ1i1gNM8YNQ1j1gNM:WNo0l1gNM;SNQ1P2dNMZ33fLNY31hLOX31hLOX31hLOX31hL0W30iL0W3OjL1V3OjL1V3OkL1T3OlL1T3OlL1T3OlL1T3NmL2S3NmL3R3MoL2Q3NoL2Q3NoL2Q3NoL3P3LQM4o2LRM3n2MRM3n2MRM4m2LSM4m2LSM4m2KTM5l2KUM5j2KVM5j2KVM5j2KVM5j2JWM7h2IXM7h2IYM6g2JYM6g2JYM7f2IZM7f2H[M8e2H[M8e2H\\M8c2H]M8c2H]M8c2H]M8c2G^M:a2F_M:a2F`M9`2G`M9`2G`M9`2G`M:_2EbM;^2EbM;^2EcM:]2FcM;\\2EdM;\\2DeM<[2DeM<[2DeM=Z2CgM<Y2DgM<X2EhM;X2DiM=V2CiM>W2BiM>W2BiM>W2BiM?V2AjM?V2@kM`0U2@kM`0U2@kMa0T2_OlMa0T2_OlMa0T2_OlMa0T2^OmMc0R2]OnMc0R2]OnMc0R2]OnMc0R2]OnMc0R2\\OoMe0P2[OPNe0P2[OPNe0P2[OPNe0P2[OPNf0o1ZOQNf0o1YORNg0n1YORNg0n1YORNh0m1XOSNh0m1XOSNh0m1WOSNj0m1VOSNk0l1UOSNl0m1TOSNl0m1TORNm0n1RORNP1m1POSNP1m1PORNQ1n1oNRNQ1n1oNQNS1n1lNSNT1m1kNSNV1m1iNSNX1m1gNTNZ1k1eNUN\\1k1cNVN]1j1bNVN_1j1aNVN`1i1\\M"}, "label": "person in blue pants grabbing donut"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands belonging to someone in jeans who is holding a donut with blue sprinkles\"."}], "task": "refering", "answer_template": "The \"hands belonging to someone in jeans who is holding a donut with blue sprinkles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 25, 26, 27, 28, 29, 33, 34, 35, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 87, 88, 89, 99, 100, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215], "bbox": [0.38186000000000003, 0.0016816816816816818, 1.0, 0.9858858858858859], "iscrowd": 0, "area": 40147.970149999994, "rle": {"size": [333, 500], "counts": "_Un11\\:1N2N3M2N2O1N2N1O01OmNDVH=g7GWH:g7JVH6j7LTH5k7MTH2m7OQH2o7NPH2Q8MoG4Q8LoG3R8MmG4S8LlG4T8MkG4U8LjG4W8LiG4W8LhG5X8KgG6Y8JfG6[8JeG6[8JdG7\\8JbG7^8IaG7Z1^O\\5;ZI7X1@^59YI7X1A`58WI7X1Ca57UI7W1Ec55UI6W1Fe54TI6U1Hg52SI6U1Ii51QI7S1Kl5OoH6T1Ln5NnH6R1NQ6KlH7R1OS6d0kI^OU6a0iIAY6=fIEZ6;dIF]69bII^67_IKc63\\IOd61\\IOc61]IOd61\\IO[6POQIQ1e0N[6;cIF]6:cIE^6<aID_6<`ICb6=^IAd6`0[I_Of6a0YI^Oi6c0VI\\Ok6d0UIZOm6g0RIWOP7i0oHWOR7i0nHUOT7l0kHROW7n0hHROQ1D`4[1^JQOo0Hc4V1^JROn0Id4V1\\JROo0Je4S1\\JSOn0Kf4S1[JROo0Kg4R1YJTOP1Kf4Q1ZJTOP1Kg4Q1TJYOT1Gg4o1ZKQNf4o1ZKQNg4n1YKSNf4m1YKTNh4j1XKXNg4h1XKYNi4f1VK[Nj4d1WK[Nk4d1TK]Nm4b1RK^NP5a1oJ`NS5]1mJcNU5\\1jJeNW5Z1hJfNZ5Y1eJhN\\5V1dJjN^5U1aJlNa5R1_JmNc5Q1]JPOd5o0[JQOg5n0XJSOi5l0WJSOl5k0SJUOP6i0oIXOR6g0mIYOV6e0jI[OW6e0gI[O\\6c0cI^O^6a0aI_Oa6`0]IBd6=WIGj69PIMQ73hH2Z7`11N2O0O2O1N2O0O2O1N101N100O2O0O2O00001O001O00001O001O00001O001O00001O00eJiM`2W2^MkMc2T2[MnMe2R2YMPNh2o1VMTNi2l1UMVNk2i1SMZNn2e1PM]NP3b1oL`NR3^1mLdNS3\\1kLfNV3X1iLjNW3V1gLlNY3S1fLoN[3o0dLSO\\3m0bLUO_3i0`LYO`3g0^L[Ob3d0]L^Od3`0[LBiNhNV4f1oLDjNhNW4b1oLGjNgNW4a1PMHhNiNW4_1QMHhNjNV4]1RMJgNkNU4\\1TMIgNPOP4V1ZMJeNVOk3P1`MLdNYOg3k0eMLdNE[3>RNMbN9h2KeNMcN<d2FjNO`N>d2BmN0_N`0a2AoN0_Nb0a2\\OPO3`Nb0_2ZOQO5_Nc0^2XORO6`Nd0]2UORO8aNd0\\2TORO9bNd0Z2TOTO8bNe0Y2SOTO:bNd0Y2ROVO9aNf0X2QOWO9bNg0U2POYO9bNh0S2PO[O8bNi0R2oN]O6bNl0P2nN^O6bNm0o1mN_O6bNn0n1lN@6bNo0m1kNB5aNQ1k1kND4bNQ1i1kNE4bNR1h1jNF4bNS1g1iNH5_NS1h1hNI6^NT1g1fNK8\\NR1h1gNL8[NQ1i1gNM8YNQ1j1gNM:WNo0l1gNM;SNQ1P2dNMZ33fLNY31hLOX31hLOX31hLOX31hL0W30iL0W3OjL1V3OjL1V3OkL1T3OlL1T3OlL1T3OlL1T3NmL2S3NmL3R3MoL2Q3NoL2Q3NoL2Q3NoL3P3LQM4o2LRM3n2MRM3n2MRM4m2LSM4m2LSM4m2KTM5l2KUM5j2KVM5j2KVM5j2KVM5j2JWM7h2IXM7h2IYM6g2JYM6g2JYM7f2IZM7f2H[M8e2H[M8e2H\\M8c2H]M8c2H]M8c2H]M8c2G^M:a2F_M:a2F`M9`2G`M9`2G`M9`2G`M:_2EbM;^2EbM;^2EcM:]2FcM;\\2EdM;\\2DeM<[2DeM<[2DeM=Z2CgM<Y2DgM<X2EhM;X2DiM=V2CiM>W2BiM>W2BiM>W2BiM?V2AjM?V2@kM`0U2@kM`0U2@kMa0T2_OlMa0T2_OlMa0T2_OlMa0T2^OmMc0R2]OnMc0R2]OnMc0R2]OnMc0R2]OnMc0R2\\OoMe0P2[OPNe0P2[OPNe0P2[OPNe0P2[OPNf0o1ZOQNf0o1YORNg0n1YORNg0n1YORNh0m1XOSNh0m1XOSNh0m1WOSNj0m1VOSNk0l1UOSNl0m1TOSNl0m1TORNm0n1RORNP1m1POSNP1m1PORNQ1n1oNRNQ1n1oNQNS1n1lNSNT1m1kNSNV1m1iNSNX1m1gNTNZ1k1eNUN\\1k1cNVN]1j1bNVN_1j1aNVN`1i1\\M"}, "label": "hands belonging to someone in jeans who is holding a donut with blue sprinkles"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding an eaten donut\"."}], "task": "refering", "answer_template": "The \"a hand holding an eaten donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 153, 154, 155, 156, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211], "bbox": [0.42354, 0.6248648648648649, 0.77076, 0.9889189189189189], "iscrowd": 0, "area": 11918.186799999996, "rle": {"size": [333, 500], "counts": "ZVU21Y:3O2O1O2N100O1O1O100O2N100O1O1O101N1O100O1O1O101N1O1N2O1N2O2M2N2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O1O2M2O1N2N101N101N101YOnMPIS2i6TNWIm1h6SNWIn1i6RNWIo1g6SNWIn1i6RNVIP2h6QNXIo1h6QNWIP2i6PNWIP2h6QNXIo1h6RNVIo1i6RNWIn1i6RNWIn1i6k0O10000O1000O10O10000O10000O1000000O0101O1O1O001O1O1O1O1O1O1O1N2O1O001O1O001O001O1O001O0O2O001O1O001O001O001O1O001O001O1O1O010O1O1O1O1O001O1O1O1O1O00101N7bNkG6]8ZOoGb0Q9M3M4L3M001O1O1O1O100O1O001O001O0N3N1N3N1OiRU1"}, "label": "a hand holding an eaten donut"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding a bitten donut\"."}], "task": "refering", "answer_template": "The \"a hand holding a bitten donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 153, 154, 155, 156, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211], "bbox": [0.42354, 0.6248648648648649, 0.77076, 0.9889189189189189], "iscrowd": 0, "area": 11918.186799999996, "rle": {"size": [333, 500], "counts": "ZVU21Y:3O2O1O2N100O1O1O100O2N100O1O1O101N1O100O1O1O101N1O1N2O1N2O2M2N2O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O1O2M2O1N2N101N101N101YOnMPIS2i6TNWIm1h6SNWIn1i6RNWIo1g6SNWIn1i6RNVIP2h6QNXIo1h6QNWIP2i6PNWIP2h6QNXIo1h6RNVIo1i6RNWIn1i6RNWIn1i6k0O10000O1000O10O10000O10000O1000000O0101O1O1O001O1O1O1O1O1O1O1N2O1O001O1O001O001O1O001O0O2O001O1O001O001O001O1O001O001O1O1O010O1O1O1O1O001O1O1O1O1O00101N7bNkG6]8ZOoGb0Q9M3M4L3M001O1O1O1O100O1O001O001O0N3N1N3N1OiRU1"}, "label": "a hand holding a bitten donut"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the legs of a person in the background\"."}], "task": "refering", "answer_template": "The \"the legs of a person in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 104, 105, 106, 107, 122, 123, 124, 125, 143], "bbox": [0.7847000000000001, 0.3432732732732733, 1.0, 0.5954954954954955], "iscrowd": 0, "area": 6753.815799999999, "rle": {"size": [333, 500], "counts": "Rno35V:5J6K5K3M3M3N3M2O1N2O1N2N2O1N2O1N2O1N2N2O0O101N1O2O0O101O00001O00100O1O001O1O00001O0000001O0000001O0000010O000000001O0000001O0000001O0000001O01O000001O0000001O0000001O0000001O0000001O0000001O0001O01O0000001O0000WNPI5Y3"}, "label": "the legs of a person in the background"}]}
{"id": 19967, "image": "COCO_train2014_000000019967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in background blue jeans\"."}], "task": "refering", "answer_template": "The \"person in background blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 104, 105, 106, 107, 122, 123, 124, 125, 143], "bbox": [0.7847000000000001, 0.3432732732732733, 1.0, 0.5954954954954955], "iscrowd": 0, "area": 6753.815799999999, "rle": {"size": [333, 500], "counts": "Rno35V:5J6K5K3M3M3N3M2O1N2O1N2N2O1N2O1N2O1N2N2O0O101N1O2O0O101O00001O00100O1O001O1O00001O0000001O0000001O0000010O000000001O0000001O0000001O0000001O01O000001O0000001O0000001O0000001O0000001O0000001O0001O01O0000001O0000WNPI5Y3"}, "label": "person in background blue jeans"}]}
{"id": 101891, "image": "COCO_train2014_000000101891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue bike behind the red car\"."}], "task": "refering", "answer_template": "The \"the blue bike behind the red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 124, 125, 126, 148, 149, 171, 172, 193, 194, 195, 196, 216, 217, 218, 219, 240, 241, 242, 243, 265, 266, 287, 288, 289, 310, 311, 312, 313, 335, 336, 359, 382, 383], "bbox": [0.424421875, 0.2657291666666667, 0.659390625, 1.0], "iscrowd": 0, "area": 13070.5583, "rle": {"size": [480, 640], "counts": "Vgo38g>7eLd0eFDT9<jFLP94oF4j8MTG:g8EWGb0d8_OZGc0e8]OZGf0d8ZOZGi0e8XOYGj0g8UOXGm0g8TOVGo0i8QOVGQ1i8oNVGS1j8mNSGV1l8jNTGW1k8jNSGX1l8hNTGY1k8gNTG[1l8eNRG]1m8cNSG^1l8cNRG_1m8aNSG`1m8_NRGc1m8^NRGd1l8\\NSGf1l8[NSGf1m8YNRGi1m8WNSGj1l8WNRGk1m8UNSGl1j0nLZ6W1jHm1k0nLZ6T1kHo1j0oLY6R1lHQ2j0oLX6g0VH\\3`1QLW6b0]H[3[1ULW6>bH[3V1WLX6=fHZ3Q1XLZ6=hHZ3l0ZL\\6;lHY3g0\\L^69oHY3b0]LCDo5e0PJW3>`LAFQ6b0TJU3:cL@HQ6?YJS35fLAIQ6=]JQ31iL_OLS68aJP3MlL^OLU67dJn2InL]ONW65gJl2DRM]OLZ65hJk2ATM\\OL]63jJj2]OVM[OM`62lJh2YOYMZOLc62nJf2TO]MYOLf60QKd2PO_MYOLh60SKb2lNbMWOMl6MUKa2hNeMVOLo6MWK^2dNiMVOKQ7MXK]2aNkMTOLT7KZK\\2^NmMSOKX7JYK]2\\NmMROMX:T2gFnMPOM\\:T2dFoMnNMb:U2^FlMPOOe:Z2TFgMVONi:a2jEaM[OOn:f2`EYMB0Q;m2VESMG0W;S3kDRMW;Y34000O100O2O000O01N1O3M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3L4M3M3L4L4O1O1000000000000O1000000000000O1000000000000O10000000000002N2N2N1O00M3M3M3M3M3L4M3M3M3M3M3M3M3M3MSgU3"}, "label": "the blue bike behind the red car"}]}
{"id": 101891, "image": "COCO_train2014_000000101891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue low rider bicycle\"."}], "task": "refering", "answer_template": "The \"the blue low rider bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 158, 159, 160, 179, 180, 181, 182, 203, 204, 226, 227, 228, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 367], "bbox": [0.7854062500000001, 0.34120833333333334, 0.9962500000000001, 0.9570833333333333], "iscrowd": 0, "area": 19753.7212, "rle": {"size": [480, 640], "counts": "hn[7;e>4M2M4M3L2OO0100TDYO]9g0bF^OZ9c0eFBW9>gFHT99jFMR92nF2n8OPG7l8IRG<j8EUG?h8@WGd0f8]OXGg0f8YOYGk0c8UO\\Go0b8QO]GS1_8nN_GV1_8jN_GZ1^8fNbG^1Z8cNdGa1Z8_NdGe1Y8[NgGh1W8XNgGl1V8UNhGo1V8QNiGR2T8nMkGV2S8jMhG]2U8dMfGa2Y8_McGe2\\8]M_Gh2`8YM[Gl2d8UMWGP3h8PMTGU3k8lLPGX3P9iLkF\\3T9Y1N2O1O001O1O1O1N101O1O1O1O001N2O1O1O001O1O1N2O00000000000O10000O2OZJhGd4X8[KnG`4Q8`KTH\\4l7cKYHY4g7eK_HW4a7hKdHT4[7lKkHo3U7PLPIl3P7SLUIi3j6WL[Ie3e6ZLaIa3_6^LfI^3Y6bLlI[3S6dLRJX3n5gLWJU3h5kL\\JR3d5mL^JR3b5lL`JT3`5kLaJU3^5kLcJV3\\5iLeJW3[5hLfJX3Y5hLhJX3X5gLjJY3U5fLlJZ3S5fLnJ[3Q5dLPK^3n4aLSK`3k4`LVKb3h4]LYKe3d4\\L]Kd3a4\\L`Kf3^4ZLbKg3\\4YLeKi3X4XLhKi3W4VLjKl3S4ULmKl3Q4TLQLm3m3SLSLo3j3QLWLP4g3QLYLQ4e3nK\\LS4b3nK^LT4_3lKbLU4]3kKdLV4Y3jKhLW4V3jKjLX4T3gKmL[4P3fKPM[4n2eKSM]4k2cKUM^4i2bKXM`4e2aK[M`4d2_K]Mc4`2^K`Mc4^2]KcMe4Z2\\KfMf4X2YKiMh4W6N2L4`K[i0"}, "label": "the blue low rider bicycle"}]}
{"id": 101891, "image": "COCO_train2014_000000101891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter blue bike on the right\"."}], "task": "refering", "answer_template": "The \"the shorter blue bike on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 158, 159, 160, 179, 180, 181, 182, 203, 204, 226, 227, 228, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 367], "bbox": [0.7854062500000001, 0.34120833333333334, 0.9962500000000001, 0.9570833333333333], "iscrowd": 0, "area": 19753.7212, "rle": {"size": [480, 640], "counts": "hn[7;e>4M2M4M3L2OO0100TDYO]9g0bF^OZ9c0eFBW9>gFHT99jFMR92nF2n8OPG7l8IRG<j8EUG?h8@WGd0f8]OXGg0f8YOYGk0c8UO\\Go0b8QO]GS1_8nN_GV1_8jN_GZ1^8fNbG^1Z8cNdGa1Z8_NdGe1Y8[NgGh1W8XNgGl1V8UNhGo1V8QNiGR2T8nMkGV2S8jMhG]2U8dMfGa2Y8_McGe2\\8]M_Gh2`8YM[Gl2d8UMWGP3h8PMTGU3k8lLPGX3P9iLkF\\3T9Y1N2O1O001O1O1O1N101O1O1O1O001N2O1O1O001O1O1N2O00000000000O10000O2OZJhGd4X8[KnG`4Q8`KTH\\4l7cKYHY4g7eK_HW4a7hKdHT4[7lKkHo3U7PLPIl3P7SLUIi3j6WL[Ie3e6ZLaIa3_6^LfI^3Y6bLlI[3S6dLRJX3n5gLWJU3h5kL\\JR3d5mL^JR3b5lL`JT3`5kLaJU3^5kLcJV3\\5iLeJW3[5hLfJX3Y5hLhJX3X5gLjJY3U5fLlJZ3S5fLnJ[3Q5dLPK^3n4aLSK`3k4`LVKb3h4]LYKe3d4\\L]Kd3a4\\L`Kf3^4ZLbKg3\\4YLeKi3X4XLhKi3W4VLjKl3S4ULmKl3Q4TLQLm3m3SLSLo3j3QLWLP4g3QLYLQ4e3nK\\LS4b3nK^LT4_3lKbLU4]3kKdLV4Y3jKhLW4V3jKjLX4T3gKmL[4P3fKPM[4n2eKSM]4k2cKUM^4i2bKXM`4e2aK[M`4d2_K]Mc4`2^K`Mc4^2]KcMe4Z2\\KfMf4X2YKiMh4W6N2L4`K[i0"}, "label": "the shorter blue bike on the right"}]}
{"id": 118277, "image": "COCO_train2014_000000118277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"puddle reflection of a light brown dog\"."}], "task": "refering", "answer_template": "The \"puddle reflection of a light brown dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [288, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381, 401, 402, 403, 404, 424, 425, 426, 427, 447, 448, 449, 470, 471], "bbox": [0.43029687499999997, 0.5515156250000001, 0.612125, 0.8848437500000002], "iscrowd": 0, "area": 13503.384099999996, "rle": {"size": [640, 640], "counts": "bQ]54hc04L4L4K4M4L4L4L4L3M4L4L4J5K6J6J6K5J5K6K5eM[2N2O001N2O1N2F_JTBa5m=_JRBl0Nd3o=bKPBj0=Z3c=lKoAi0b0Y3^=PLnAe0h0Z3Z=jLiBT3W=jLmBU3S=iLQCU3P=gLRCZ3P=bLQC_3P=^LRCa3Q=ZLQCg3Q=ULPCl3Q=PLQCQ4_>0O001O100O1O00100N2M2O2M3M3M3M2N3N2M3kNQKVCR5j<XKiBj4W=bK[B`4f=k02N2O1N1O2N2N2O0O2N2M3B>O001000e0[O=B>C=C5K5J7J5K5K0O0100O10O01O100O001@`0VOj0_O`0I52N4K6K4K5L4SMl_Od1]`0UNo_OY1^`0_Nn_OP1``0gNm_Oo0ha0H8H8I7H`bj4"}, "label": "puddle reflection of a light brown dog"}]}
{"id": 118277, "image": "COCO_train2014_000000118277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of a dog\"."}], "task": "refering", "answer_template": "The \"the reflection of a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [288, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381, 401, 402, 403, 404, 424, 425, 426, 427, 447, 448, 449, 470, 471], "bbox": [0.43029687499999997, 0.5515156250000001, 0.612125, 0.8848437500000002], "iscrowd": 0, "area": 13503.384099999996, "rle": {"size": [640, 640], "counts": "bQ]54hc04L4L4K4M4L4L4L4L3M4L4L4J5K6J6J6K5J5K6K5eM[2N2O001N2O1N2F_JTBa5m=_JRBl0Nd3o=bKPBj0=Z3c=lKoAi0b0Y3^=PLnAe0h0Z3Z=jLiBT3W=jLmBU3S=iLQCU3P=gLRCZ3P=bLQC_3P=^LRCa3Q=ZLQCg3Q=ULPCl3Q=PLQCQ4_>0O001O100O1O00100N2M2O2M3M3M3M2N3N2M3kNQKVCR5j<XKiBj4W=bK[B`4f=k02N2O1N1O2N2N2O0O2N2M3B>O001000e0[O=B>C=C5K5J7J5K5K0O0100O10O01O100O001@`0VOj0_O`0I52N4K6K4K5L4SMl_Od1]`0UNo_OY1^`0_Nn_OP1``0gNm_Oo0ha0H8H8I7H`bj4"}, "label": "the reflection of a dog"}]}
{"id": 527879, "image": "COCO_train2014_000000527879.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in a field laying down , surrounded by rocks\"."}], "task": "refering", "answer_template": "The \"a giraffe in a field laying down , surrounded by rocks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 129, 130, 131, 151, 152, 153, 173, 174, 175, 195, 196, 197, 198, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.3998125, 0.3014519906323185, 0.73803125, 0.7760889929742388], "iscrowd": 0, "area": 14144.6903, "rle": {"size": [427, 640], "counts": "\\Q[3<l<8H3M4L4N2N1O2N2N1O2N1O2N2N1O1O101N100O100O100O100O10000O1000000O10O100O01000O010O0100O1000O0100O100O100O100O10000O100dDUNW;o110O1N2O1O1O1N2O1O1O0O2O100O1O100O1O100O1O100O100O100O100O100O100O1O1N2O0O2N200O100O1O100O1O100O1N2O1N2O1N2N2O1N2O2M2N2O1N2NZNkFZOS9f0QGYOl8g0WGXOh8g0[GXOc8h0aGVO]8j0fGUOX8k0lGTOR8k0QHTOm7l0VHSOh7m0\\HQOb7o0aHPO]7P1TIbNi6^1YIbNf6]1]IbNa6_1`IaN^6_1dIaNZ6_1iIaNT6`1mI`NQ6`1QJ`Nm5`1VJ_Ni5a1XJ`Nf5_1\\JaNc5^1_JbNa5]1aJcN]5\\1eJdNZ5[1hJfNV5Z1lJeNS5Y1PKhNg4YOfHm1e2jNe4[OcHi1k2lNb4]O`Hf1Q3lN_4W1dKiN\\4U1fKkNY4S1kKlNU4R1mKnNR4P1QLPOm3P1ULoNl3o0WLPOh3n0[LROe3l0]LTOc3i0`LWO`3g0cLXO\\3f0gLZOY3d0iL[Oo2BPIQ1T4\\Ol2k0WMSOj2j0YMVOg2h0[MWOg2h0YMWOi2h0XMWOi2h0XMVOi2j0ZMROh2m0[MPOf2o0l4O000O0100O100O1001O1O1O1O2N2N2N1N3N1O001O001O010O1O001O00100O001O001O10O01O3L3KilU2"}, "label": "a giraffe in a field laying down , surrounded by rocks"}]}
{"id": 527879, "image": "COCO_train2014_000000527879.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe laying\"."}], "task": "refering", "answer_template": "The \"giraffe laying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 129, 130, 131, 151, 152, 153, 173, 174, 175, 195, 196, 197, 198, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.3998125, 0.3014519906323185, 0.73803125, 0.7760889929742388], "iscrowd": 0, "area": 14144.6903, "rle": {"size": [427, 640], "counts": "\\Q[3<l<8H3M4L4N2N1O2N2N1O2N1O2N2N1O1O101N100O100O100O100O10000O1000000O10O100O01000O010O0100O1000O0100O100O100O100O10000O100dDUNW;o110O1N2O1O1O1N2O1O1O0O2O100O1O100O1O100O1O100O100O100O100O100O100O1O1N2O0O2N200O100O1O100O1O100O1N2O1N2O1N2N2O1N2O2M2N2O1N2NZNkFZOS9f0QGYOl8g0WGXOh8g0[GXOc8h0aGVO]8j0fGUOX8k0lGTOR8k0QHTOm7l0VHSOh7m0\\HQOb7o0aHPO]7P1TIbNi6^1YIbNf6]1]IbNa6_1`IaN^6_1dIaNZ6_1iIaNT6`1mI`NQ6`1QJ`Nm5`1VJ_Ni5a1XJ`Nf5_1\\JaNc5^1_JbNa5]1aJcN]5\\1eJdNZ5[1hJfNV5Z1lJeNS5Y1PKhNg4YOfHm1e2jNe4[OcHi1k2lNb4]O`Hf1Q3lN_4W1dKiN\\4U1fKkNY4S1kKlNU4R1mKnNR4P1QLPOm3P1ULoNl3o0WLPOh3n0[LROe3l0]LTOc3i0`LWO`3g0cLXO\\3f0gLZOY3d0iL[Oo2BPIQ1T4\\Ol2k0WMSOj2j0YMVOg2h0[MWOg2h0YMWOi2h0XMWOi2h0XMVOi2j0ZMROh2m0[MPOf2o0l4O000O0100O100O1001O1O1O1O2N2N2N1N3N1O001O001O010O1O001O00100O001O001O10O01O3L3KilU2"}, "label": "giraffe laying"}]}
{"id": 208396, "image": "COCO_train2014_000000208396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with long sideburns wearing a hat and jacket smiling\"."}], "task": "refering", "answer_template": "The \"a man with long sideburns wearing a hat and jacket smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 118, 119, 120, 121, 135, 136, 137, 138, 139, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 177, 188, 189, 190, 191, 192, 193, 194, 195, 205, 206, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.41952, 0.24549848942598188, 0.8598599999999998, 0.9915709969788519], "iscrowd": 0, "area": 25941.939350000004, "rle": {"size": [331, 500], "counts": "h`T21Y:1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1N2M3L4L4M3L4L4^KbN3b1M`NNc11_NJe16\\NEh1;YN@k1?VN]On1c0SNXOQ2g0PNUOT2k0mMoNX2Q1hMkN\\2T1fMgN^2Y1bMbNc2]1^M_Nf2a1[MZNi2e1XM[Nh2e1YMYNh2g1XMYNh2f1ZMXNg2h1YMXNg2g1ZMXNg2f1[MZNe2e1\\MZNe2d1\\M]Nd2b1]M]Nd2a1^M_Nb2_1`M`Na2_1`M`Na2^1`McN`2\\1cMaN^2]1kM\\NU2b1TNWNl1g1]NQNd1n1eNjM[1T2iNjMW1T2lNkMT1S2POjMQ1T2SOjMm0T2VOkMj0T2YOiMh0U2\\OiMd0U2_OjM`0U2DhM=V2GhM9W2IhM7V2MhM3V2OjM1V20iM0X20gM0Y21fMOZ22eMN[23dMM\\24cML^24aML_24aML_25`MK`26_MJa27^MIb28]MHd28[MHe29ZMGf2:YMFg2:YMFg2o31O000000001O00000000001O000000001O000000001O00000000001O00000O101O00001O001O1O1O001O1O1O:F9G2TKWLR3j3lLWLT3i3jLZLU3g3hL[LX3f3eL]LZ3c3dL_L\\3b3aLaL^3`3_LbLa3^3^LdLa3]3\\LeLd3\\3[LeLd3[3[LfLX3h3fLZLg2X4YMhKT2l4jMWKb1\\5]NUK`0]5_OVK\\O^5c0V2L3J7IkU;LUjD:G9F;E:F:F:F:M3O1O1005K4L5K5K5K4L5KX[f0"}, "label": "a man with long sideburns wearing a hat and jacket smiling"}]}
{"id": 208396, "image": "COCO_train2014_000000208396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with hat on\"."}], "task": "refering", "answer_template": "The \"man with hat on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 118, 119, 120, 121, 135, 136, 137, 138, 139, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 177, 188, 189, 190, 191, 192, 193, 194, 195, 205, 206, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.41952, 0.24549848942598188, 0.8598599999999998, 0.9915709969788519], "iscrowd": 0, "area": 25941.939350000004, "rle": {"size": [331, 500], "counts": "h`T21Y:1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1N2M3L4L4M3L4L4^KbN3b1M`NNc11_NJe16\\NEh1;YN@k1?VN]On1c0SNXOQ2g0PNUOT2k0mMoNX2Q1hMkN\\2T1fMgN^2Y1bMbNc2]1^M_Nf2a1[MZNi2e1XM[Nh2e1YMYNh2g1XMYNh2f1ZMXNg2h1YMXNg2g1ZMXNg2f1[MZNe2e1\\MZNe2d1\\M]Nd2b1]M]Nd2a1^M_Nb2_1`M`Na2_1`M`Na2^1`McN`2\\1cMaN^2]1kM\\NU2b1TNWNl1g1]NQNd1n1eNjM[1T2iNjMW1T2lNkMT1S2POjMQ1T2SOjMm0T2VOkMj0T2YOiMh0U2\\OiMd0U2_OjM`0U2DhM=V2GhM9W2IhM7V2MhM3V2OjM1V20iM0X20gM0Y21fMOZ22eMN[23dMM\\24cML^24aML_24aML_25`MK`26_MJa27^MIb28]MHd28[MHe29ZMGf2:YMFg2:YMFg2o31O000000001O00000000001O000000001O000000001O00000000001O00000O101O00001O001O1O1O001O1O1O:F9G2TKWLR3j3lLWLT3i3jLZLU3g3hL[LX3f3eL]LZ3c3dL_L\\3b3aLaL^3`3_LbLa3^3^LdLa3]3\\LeLd3\\3[LeLd3[3[LfLX3h3fLZLg2X4YMhKT2l4jMWKb1\\5]NUK`0]5_OVK\\O^5c0V2L3J7IkU;LUjD:G9F;E:F:F:F:M3O1O1005K4L5K5K5K4L5KX[f0"}, "label": "man with hat on"}]}
{"id": 208396, "image": "COCO_train2014_000000208396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on left\"."}], "task": "refering", "answer_template": "The \"woman on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 112, 113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 183, 184, 185, 186, 187, 188, 201, 202, 203, 204, 205, 206], "bbox": [0.17406, 0.25099697885196376, 0.58614, 0.983595166163142], "iscrowd": 0, "area": 28509.37555, "rle": {"size": [331, 500], "counts": "P^l02W:2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3K5L4K5L4L4K5L4K5N2O1O1O100O1O100O1O1O100O1O100UNcKgM^4X2eKeM\\4Y2hKeMX4Z2kKcMV4\\2mKaMT4]2PL`MQ4_2SL^Mm3`2WL]Mj3b2YL[Mh3d2[LZMe3d2_LYMb3f2aLWM`3h2cLTM_3j2fLhLe3W3h1N2O1O1N2O1O1N20000O1000000O10000O1000000O10000O1000000O10000O101O000O10000O1000000O10000001O0O101N1O101N100O2O0O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N2N2N2N2M3M3M3N2M3L5J5J6J6J6J6K6J6SLZJo2k5mLYJm2m5oLWJl2n5QMTJk2P6RMTJR1A9[7BgH<^7_OdH?f7WO\\Hf0Z8eNiGX1f8M3N2N2M3N2N2N2O2O000O10000O1O1O2N100O1O1O1O1O101N1O1O1O1O100O2N1O1O1O10nfR2"}, "label": "woman on left"}]}
{"id": 208396, "image": "COCO_train2014_000000208396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with white hair\"."}], "task": "refering", "answer_template": "The \"a woman with white hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 112, 113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 183, 184, 185, 186, 187, 188, 201, 202, 203, 204, 205, 206], "bbox": [0.17406, 0.25099697885196376, 0.58614, 0.983595166163142], "iscrowd": 0, "area": 28509.37555, "rle": {"size": [331, 500], "counts": "P^l02W:2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3K5L4K5L4L4K5L4K5N2O1O1O100O1O100O1O1O100O1O100UNcKgM^4X2eKeM\\4Y2hKeMX4Z2kKcMV4\\2mKaMT4]2PL`MQ4_2SL^Mm3`2WL]Mj3b2YL[Mh3d2[LZMe3d2_LYMb3f2aLWM`3h2cLTM_3j2fLhLe3W3h1N2O1O1N2O1O1N20000O1000000O10000O1000000O10000O1000000O10000O101O000O10000O1000000O10000001O0O101N1O101N100O2O0O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N2N2N2N2M3M3M3N2M3L5J5J6J6J6J6K6J6SLZJo2k5mLYJm2m5oLWJl2n5QMTJk2P6RMTJR1A9[7BgH<^7_OdH?f7WO\\Hf0Z8eNiGX1f8M3N2N2M3N2N2N2O2O000O10000O1O1O2N100O1O1O1O1O101N1O1O1O1O100O2N1O1O1O10nfR2"}, "label": "a woman with white hair"}]}
{"id": 52751, "image": "COCO_train2014_000000052751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player that is wearing a catcher ' s mask , and a white uniform . he is facing towards the right\"."}], "task": "refering", "answer_template": "The \"the baseball player that is wearing a catcher ' s mask , and a white uniform . he is facing towards the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 40, 41, 42, 43, 63, 64, 65, 86, 87, 88, 109, 110, 111, 132, 133, 134, 156, 157, 178, 179, 180, 201, 202, 203, 224, 225, 247, 248, 270, 271, 272, 293, 294, 295], "bbox": [0.741421875, 0.001571753986332574, 0.900203125, 0.8060592255125285], "iscrowd": 0, "area": 19042.902200000004, "rle": {"size": [439, 640], "counts": "Yf[63c=4L4M3L4L4L4L3N3[JUOWNo0:i0YO[O`0R1VOROc0k0CYO5i0NZOK`1^OcNRO`NjMR3S3bNeLWN]1;cNP3Z3bN\\L\\Nd18[NP3d3`NSLaNk14UNP3m3WOkMh1T2[NiMf1V2^NfMc1Y2`NdMa1[2bNbM_1^2cN_M^1`2eN]M\\1b2gN[MZ1d2jNWMX1i2kNSMV1l2nNPMS1o2POnLQ1Q3SOkLn0T3VOhLk0X3XOdLi0Z3[OcLf0[3_OaLb0]3C_L>_3n4M3N2M3L4M3M3M4N1O10_NaMjG_2j7mMVHS2]7ZNcHf1P7gNPIY1k6mNTIS1k6nNUIR1j6oNVIQ1i6QOVIo0i6ROWIn0e6VO[Ii0`6^O^Ic0]6BWIfMDh2o6H]IaMBh2o6I_I`MAg2P7J^I`M@g2Q7J_I`M_Of2R7J_I`M^Og2R7K^I`M_Oe2S7K^IaM]Oe2T7L^I=b6D]I;c6F]I:c6G\\I9Y5QMZKg2]O8S4QMVL79a2\\O8T4WMjK76k2IIT1jL`1b0QM71T39ZOU1lL\\1^5\\MhMY1kLW1_5^MhM[1lLS1^5`MgM]1nLo0]5aMgM`1oLj0l8XOVGd0k8\\OXG>k8BXGNU92nF]O`9c0cFlNk9U1[100000001O100O1O1N2K6J5J6K5K5K5J6Ke]k0"}, "label": "the baseball player that is wearing a catcher ' s mask , and a white uniform . he is facing towards the right"}]}
{"id": 52751, "image": "COCO_train2014_000000052751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player in catcher ' s uniform\"."}], "task": "refering", "answer_template": "The \"baseball player in catcher ' s uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 40, 41, 42, 43, 63, 64, 65, 86, 87, 88, 109, 110, 111, 132, 133, 134, 156, 157, 178, 179, 180, 201, 202, 203, 224, 225, 247, 248, 270, 271, 272, 293, 294, 295], "bbox": [0.741421875, 0.001571753986332574, 0.900203125, 0.8060592255125285], "iscrowd": 0, "area": 19042.902200000004, "rle": {"size": [439, 640], "counts": "Yf[63c=4L4M3L4L4L4L3N3[JUOWNo0:i0YO[O`0R1VOROc0k0CYO5i0NZOK`1^OcNRO`NjMR3S3bNeLWN]1;cNP3Z3bN\\L\\Nd18[NP3d3`NSLaNk14UNP3m3WOkMh1T2[NiMf1V2^NfMc1Y2`NdMa1[2bNbM_1^2cN_M^1`2eN]M\\1b2gN[MZ1d2jNWMX1i2kNSMV1l2nNPMS1o2POnLQ1Q3SOkLn0T3VOhLk0X3XOdLi0Z3[OcLf0[3_OaLb0]3C_L>_3n4M3N2M3L4M3M3M4N1O10_NaMjG_2j7mMVHS2]7ZNcHf1P7gNPIY1k6mNTIS1k6nNUIR1j6oNVIQ1i6QOVIo0i6ROWIn0e6VO[Ii0`6^O^Ic0]6BWIfMDh2o6H]IaMBh2o6I_I`MAg2P7J^I`M@g2Q7J_I`M_Of2R7J_I`M^Og2R7K^I`M_Oe2S7K^IaM]Oe2T7L^I=b6D]I;c6F]I:c6G\\I9Y5QMZKg2]O8S4QMVL79a2\\O8T4WMjK76k2IIT1jL`1b0QM71T39ZOU1lL\\1^5\\MhMY1kLW1_5^MhM[1lLS1^5`MgM]1nLo0]5aMgM`1oLj0l8XOVGd0k8\\OXG>k8BXGNU92nF]O`9c0cFlNk9U1[100000001O100O1O1N2K6J5J6K5K5K5J6Ke]k0"}, "label": "baseball player in catcher ' s uniform"}]}
{"id": 52751, "image": "COCO_train2014_000000052751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cleveland indians player wearing white batting gloves\"."}], "task": "refering", "answer_template": "The \"a cleveland indians player wearing white batting gloves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 61, 62, 81, 82, 84, 85, 105, 106, 107, 108, 109, 129, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 222, 223, 224, 245, 246, 268, 269, 270, 291, 292, 293, 314, 315, 316, 317, 337, 338, 339, 340], "bbox": [0.55140625, 0.07302961275626424, 0.8063749999999998, 0.9396810933940775], "iscrowd": 0, "area": 23256.61919999999, "rle": {"size": [439, 640], "counts": "d\\g48V=`0A?F:G:F:G8H9G3N1N2N2001O0O1YN]E9b:DdE:_:_OgE?]:ZOhEd0[:VOiEi0V:kNXFS1X;N2O1N2N2O1N100O1O100O0001O001O1O100O000000000000000001OO10000000000O10002N1O2POoNVET1g:lNXE[1b:fN]E`1\\:aNdEe1V:\\NiEj1P:XN^E03n1Y:RNbEg2h6TMaL6dLT3[6kLoL2dLc3o5\\L\\MZ4V2gKiM[4V2eKhM]4W2dKhM^4W2cKgM_4W2bKhM`4oN]Kl034b4iNcKP1L6k5[MaJ\\2E8o5SMcJb2_O:U7ClH<W7AkH=X7@iH?]7[OdHd0b7VO^Hk0g7oNZHQ1k7iNVHX1n7bNSH^1R8\\NoGd1V8VNkGk1Y8hMoGX2U8`MoG`2U8XMoGi2_900001O00001O00001O00001O0\\FSM^8m2]GXMc8i2YG[Me8i2TG\\Mk8g2oF\\MQ9h2hF\\MW9c3N101O1O001N101K5O001TL]KaNe4[1SLPNm3n1WLPNj3l1[LRNf3j1_LSNb3j1cLTN^3i1fLUN[3g1jLVNW3g1nLWNS3f1QMXNP3d1UMYNl2d1YMZNh2c1]MZNd2b1aM[Nb2`1cM^N_2]1fMaN\\2c0cJWNU3S1[2b0eJUNU3W1X2a0gJRNV3[1V2?hJQNV3]1V2>aN@b1<aNBb19bNDb18bNE^1:eND\\1:gNCZ1<iNBX1;lNCU1;mNDS1;POCQ1;ROBo0<TOCm0;VOCl0:VOEl08WOEk09XOEj07YOHi05ZOIg06ZOHi05ZOIT1InN5[7O2MPid1"}, "label": "a cleveland indians player wearing white batting gloves"}]}
{"id": 52751, "image": "COCO_train2014_000000052751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player on the right with white batting gloves and no name showing on his jersey\"."}], "task": "refering", "answer_template": "The \"the player on the right with white batting gloves and no name showing on his jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 61, 62, 81, 82, 84, 85, 105, 106, 107, 108, 109, 129, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 222, 223, 224, 245, 246, 268, 269, 270, 291, 292, 293, 314, 315, 316, 317, 337, 338, 339, 340], "bbox": [0.55140625, 0.07302961275626424, 0.8063749999999998, 0.9396810933940775], "iscrowd": 0, "area": 23256.61919999999, "rle": {"size": [439, 640], "counts": "d\\g48V=`0A?F:G:F:G8H9G3N1N2N2001O0O1YN]E9b:DdE:_:_OgE?]:ZOhEd0[:VOiEi0V:kNXFS1X;N2O1N2N2O1N100O1O100O0001O001O1O100O000000000000000001OO10000000000O10002N1O2POoNVET1g:lNXE[1b:fN]E`1\\:aNdEe1V:\\NiEj1P:XN^E03n1Y:RNbEg2h6TMaL6dLT3[6kLoL2dLc3o5\\L\\MZ4V2gKiM[4V2eKhM]4W2dKhM^4W2cKgM_4W2bKhM`4oN]Kl034b4iNcKP1L6k5[MaJ\\2E8o5SMcJb2_O:U7ClH<W7AkH=X7@iH?]7[OdHd0b7VO^Hk0g7oNZHQ1k7iNVHX1n7bNSH^1R8\\NoGd1V8VNkGk1Y8hMoGX2U8`MoG`2U8XMoGi2_900001O00001O00001O00001O0\\FSM^8m2]GXMc8i2YG[Me8i2TG\\Mk8g2oF\\MQ9h2hF\\MW9c3N101O1O001N101K5O001TL]KaNe4[1SLPNm3n1WLPNj3l1[LRNf3j1_LSNb3j1cLTN^3i1fLUN[3g1jLVNW3g1nLWNS3f1QMXNP3d1UMYNl2d1YMZNh2c1]MZNd2b1aM[Nb2`1cM^N_2]1fMaN\\2c0cJWNU3S1[2b0eJUNU3W1X2a0gJRNV3[1V2?hJQNV3]1V2>aN@b1<aNBb19bNDb18bNE^1:eND\\1:gNCZ1<iNBX1;lNCU1;mNDS1;POCQ1;ROBo0<TOCm0;VOCl0:VOEl08WOEk09XOEj07YOHi05ZOIg06ZOHi05ZOIT1InN5[7O2MPid1"}, "label": "the player on the right with white batting gloves and no name showing on his jersey"}]}
{"id": 52751, "image": "COCO_train2014_000000052751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player with the number 7 on his back\"."}], "task": "refering", "answer_template": "The \"player with the number 7 on his back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 74, 75, 76, 77, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 258, 259, 260, 281, 282, 283, 304, 305, 306, 327, 328, 329, 352], "bbox": [0.09771875, 0.11615034168564921, 0.405375, 0.97250569476082], "iscrowd": 0, "area": 26099.730549999986, "rle": {"size": [439, 640], "counts": "aWk02_=1cB0[=7N2N2O3L3N1N101N101M3O1OO001O100O010O02O1N2O1N3N2M3N]OfCKX<5QDEm;;YDAd;`0_D^O`;b0i0101N1O100O1LVO\\Cj0d<4000000000001O00001O0O2O00001O1N2O2N1O2N1O2N1N2O2N0000O1006dEgNo7`1ZG[O[8j0^GBW8c0dGDW8`0`GH\\8l2N3L3N3M2dJdJX2_5cMgJZ2[5TMaJRN<c4V5PL[KYO1e4f4jKhLT4[3gKiLX4[3_KjL`4X6N2O0O1O100O00100O10O10mNVGYMi8f2^GTMc8j2aGTM^8k2fGSMY8m2iGQMX8o2jGnLV8R3mGkLS8U3W11]OfLgF[3Y9oL[FS3e9`0N101SNbL]Ia3U6XM[In2f6\\MiHj2W7UNfHn0\\7mNjHQ1W7nNkHP1V7nNmHP1S7oNPIo0Q7oNRIo0o6oNTIo0m6oNUIP1m6mNUIR1n6jNTIU1m6fNXIZ1i6`N\\I_1e6]N_Ib1c6PNjIo1W6]M]Jb2d5YMaJf2`5UMfJj2iNSMl7l2RHUMP8j2nGXMS8g2lGYMV8f2iGZMX8f2fG[M[8f2cGZM^8g2`GYMa8h2]GXMQ21[3g2cJWMo1>W3Z2iJXMl1e0Y3S2TJVM3ON2_2l0Z3m1VJWMM4ML1Oa2T1\\3f1WJYMM;]2m0\\3_1YJnMn1R1g3P1[JmMS1Q2e3WMlKl2ZOlMU1U2`3ZMlKf2]OkMW1X2Z3\\MQLb2]OiMX1\\2o2dMZLo2g0Ah2eM`Li2g0Gc2dMdLe2h0L]2fMeL`2n0NW2hMeL\\2S12P2b1oMcNj1_1WNgNn0m1RO[N?m1AZN1l10[NBk1?ZNUOm1j0_51O100M3M3NO0O21M2OM4L1hNjCT1]<O5O1N2O1N3N`0^O2N1O200O100000O10O1000000O100O1NfXS5"}, "label": "player with the number 7 on his back"}]}
{"id": 52751, "image": "COCO_train2014_000000052751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player in the left of the picture\"."}], "task": "refering", "answer_template": "The \"the baseball player in the left of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 74, 75, 76, 77, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 258, 259, 260, 281, 282, 283, 304, 305, 306, 327, 328, 329, 352], "bbox": [0.09771875, 0.11615034168564921, 0.405375, 0.97250569476082], "iscrowd": 0, "area": 26099.730549999986, "rle": {"size": [439, 640], "counts": "aWk02_=1cB0[=7N2N2O3L3N1N101N101M3O1OO001O100O010O02O1N2O1N3N2M3N]OfCKX<5QDEm;;YDAd;`0_D^O`;b0i0101N1O100O1LVO\\Cj0d<4000000000001O00001O0O2O00001O1N2O2N1O2N1O2N1N2O2N0000O1006dEgNo7`1ZG[O[8j0^GBW8c0dGDW8`0`GH\\8l2N3L3N3M2dJdJX2_5cMgJZ2[5TMaJRN<c4V5PL[KYO1e4f4jKhLT4[3gKiLX4[3_KjL`4X6N2O0O1O100O00100O10O10mNVGYMi8f2^GTMc8j2aGTM^8k2fGSMY8m2iGQMX8o2jGnLV8R3mGkLS8U3W11]OfLgF[3Y9oL[FS3e9`0N101SNbL]Ia3U6XM[In2f6\\MiHj2W7UNfHn0\\7mNjHQ1W7nNkHP1V7nNmHP1S7oNPIo0Q7oNRIo0o6oNTIo0m6oNUIP1m6mNUIR1n6jNTIU1m6fNXIZ1i6`N\\I_1e6]N_Ib1c6PNjIo1W6]M]Jb2d5YMaJf2`5UMfJj2iNSMl7l2RHUMP8j2nGXMS8g2lGYMV8f2iGZMX8f2fG[M[8f2cGZM^8g2`GYMa8h2]GXMQ21[3g2cJWMo1>W3Z2iJXMl1e0Y3S2TJVM3ON2_2l0Z3m1VJWMM4ML1Oa2T1\\3f1WJYMM;]2m0\\3_1YJnMn1R1g3P1[JmMS1Q2e3WMlKl2ZOlMU1U2`3ZMlKf2]OkMW1X2Z3\\MQLb2]OiMX1\\2o2dMZLo2g0Ah2eM`Li2g0Gc2dMdLe2h0L]2fMeL`2n0NW2hMeL\\2S12P2b1oMcNj1_1WNgNn0m1RO[N?m1AZN1l10[NBk1?ZNUOm1j0_51O100M3M3NO0O21M2OM4L1hNjCT1]<O5O1N2O1N3N`0^O2N1O200O100000O10O1000000O100O1NfXS5"}, "label": "the baseball player in the left of the picture"}]}
{"id": 454162, "image": "COCO_train2014_000000454162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a pink and white jacket\"."}], "task": "refering", "answer_template": "The \"a person wearing a pink and white jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 143, 144, 145, 146, 166, 167, 168, 190, 191, 213, 214, 236, 237, 238, 259, 260, 261], "bbox": [0.228375, 0.29862500000000003, 0.360015625, 0.6952083333333334], "iscrowd": 0, "area": 7995.282549999999, "rle": {"size": [480, 640], "counts": "YdT21k>5K5L4O000001O0000001O000000N2`AFX>c0I7L4N1O2N2fBnNc<U1XCXO]<k0_C@W<c0eCIo;n1F:F7G:E:E<B>H7I5K2O0O2O1N2N2O0O11O01OgN_GSLa8f2kHUMU7o2lHlLT7X3lHdLU7_3kH]LU7g3iHXLW7k3hHRL[7o3k12N2N2N30OO1M3M3M2N100O1OiLfEi1X:TNPFg1o9WNXFe1h9ZN^Fa1b9]NeFoN[OQ2P:nNZGm0f8QObGh0_8WOcGh0^8mNmGR1S8iNSHU1n7jNSHV1k7kNXHR1h7nN]Hm0b7TObHg0_7YOiH?W7AoH8T7FSI3o6KWINl60T4NSPP6"}, "label": "a person wearing a pink and white jacket"}]}
{"id": 568851, "image": "COCO_train2014_000000568851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an awkwardly eaten oyster pizza\"."}], "task": "refering", "answer_template": "The \"an awkwardly eaten oyster pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 255, 256], "bbox": [0.07584375, 0.24493749999999997, 0.434828125, 0.6808958333333333], "iscrowd": 0, "area": 27206.47365, "rle": {"size": [480, 640], "counts": "oUg0:c>;E<D;YBPOk<l1E;E8H5K4L5K4L5K4K6K4L4K5N2M3N2M3N2M3N2N2M3N2M3N1O2M3N2M3N2N1N2O001N101O0O2O0O2O001N1O2O0O2N101N1O2O010001N101N101O0O2O001N101N101O0O2O001N100O2O001N101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O2O0O2O001N2O0O2O001N101O0O2O0O2O1O0O2O0O2O001N1J7J5K6M3O0O2O0O2N2O0O2N2O0O2N200O010O1O010O10O0100O1O100O100O1O100O100O100O1O100O010O1O100O100O1O1O1N2N2O1N2O1N2O1N2O1N2N101N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O0O2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N101N2N2O1N2N\\^Y5"}, "label": "an awkwardly eaten oyster pizza"}]}
{"id": 568851, "image": "COCO_train2014_000000568851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an already bitten slice of pizza\"."}], "task": "refering", "answer_template": "The \"an already bitten slice of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 255, 256], "bbox": [0.07584375, 0.24493749999999997, 0.434828125, 0.6808958333333333], "iscrowd": 0, "area": 27206.47365, "rle": {"size": [480, 640], "counts": "oUg0:c>;E<D;YBPOk<l1E;E8H5K4L5K4L5K4K6K4L4K5N2M3N2M3N2M3N2N2M3N2M3N1O2M3N2M3N2N1N2O001N101O0O2O0O2O001N1O2O0O2N101N1O2O010001N101N101O0O2O001N101N101O0O2O001N100O2O001N101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O2O0O2O001N2O0O2O001N101O0O2O0O2O1O0O2O0O2O001N1J7J5K6M3O0O2O0O2N2O0O2N2O0O2N200O010O1O010O10O0100O1O100O100O1O100O100O100O1O100O010O1O100O100O1O1O1N2N2O1N2O1N2O1N2O1N2N101N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O0O2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N101N2N2O1N2N\\^Y5"}, "label": "an already bitten slice of pizza"}]}
{"id": 568851, "image": "COCO_train2014_000000568851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced pizza that ' s not beeen eaten yet\"."}], "task": "refering", "answer_template": "The \"sliced pizza that ' s not beeen eaten yet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383], "bbox": [0.532578125, 0.2426875, 0.9825781250000001, 0.9820208333333333], "iscrowd": 0, "area": 69219.02369999999, "rle": {"size": [480, 640], "counts": "iQP55k>6J6I8I6I7J6I7J7H7J6I7J6J6I8I6I7J6I7J7H7J6I7J6J6I8I6I7J6I7J7H7J6J6I7J7H7J6I7J6I7J2M10O10O10O0100O010O10O01000O0100O010O10O01000O0100O010O10O0100O01000O010O010O10O0100O10000O10O0100O100O10000O010O100O100O10O010000O100O10O0100O100O10O10O100O010O100O10O10O100O010O100O010O1000O0100O100O010O100O10O10O100O10O0100O100O01000O100O010O10O01000O010O010O0100O010O01000O010O010O010O10O010O10O010O01O001O001O1O010O001O001O001O010O1O001N1O2O0O2N101N1O2O1N101N2N3N2M3M3N2M3N2M4L3N2M4M2M3N3L3N2M4M2M3N3L3N3L4L4M2M4M3L4M3L4M3L4M3L4M3L3N3L3N2M4L3M4L3L4M4L3L5L3M3M4K4L5I6J6K6I8I6I8H7I8H7I8G8YOkn4"}, "label": "sliced pizza that ' s not beeen eaten yet"}]}
{"id": 568851, "image": "COCO_train2014_000000568851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza slice with 5 shellfish on it\"."}], "task": "refering", "answer_template": "The \"a pizza slice with 5 shellfish on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 35, 36, 37, 38, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 146, 147, 148, 149, 150, 151, 152, 153, 171, 172, 173], "bbox": [0.190453125, 0.056187499999999994, 0.864609375, 0.4516875], "iscrowd": 0, "area": 47496.30224999999, "rle": {"size": [480, 640], "counts": "^Yi12m>1N3N1N3N1O2M3N1N3N1O2M2O2N1O2M3N1O2N1O2N1N3N2N1O2N1O2N1N3N2N1O2N101N101O1N101O0O2O001N2O001N2O001O1N2O1O0O2O1O1N101O1N2O1O0O2O1O1N101O1O1N2O001N2O1O0O2O001N101O0O2O001N101O0O2O001O0O2O001N101O0O2O001N101O0O2O001O001N101O001O001N101O001O0O2O001O001O0O2O001O001N101O001O0O2O001O001O0O2O001O001N101O001O0O2O0O101N101N10001N101N100O2O001N100O2O0O2O000O2O001O000010O0001O00001O001O000010O0001O00001O00001OO2O000O100O10000O101N10000O101N100O10001N100O10001N100O10001N100O10001N100O10001N100O101O0O101N100O2O000O101N10001N100O2O000O101N10001N100O2O000O2O0O10001N100O2O0O101O0O100O2O000O101N10000O2O0O10001N100O101O0O100O2O000O2O0O100O2O000O101N2O001N101N101O0O2O0O2O1O0O2O0O2O001N101N101N101O0O2O0O2O001N101N101O0O2O0O2O001N101N10001N101N101O1N101N2O0O2O001N2O0O2O1O0O2O0O2O1O0O2O0O2O1O0O2O1N101O0O2O1N101O0O2O1N101N2O001N101N2O001N2O0O2O001N2O0O2O001N2O0O2O1O0O2O0OWeX1"}, "label": "a pizza slice with 5 shellfish on it"}]}
{"id": 183827, "image": "COCO_train2014_000000183827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small bed with an orange throw thrown across it\"."}], "task": "refering", "answer_template": "The \"a small bed with an orange throw thrown across it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 204, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 293], "bbox": [0.6650625, 0.5831850117096019, 0.9320937499999999, 0.822248243559719], "iscrowd": 0, "area": 7948.228500000003, "rle": {"size": [427, 640], "counts": "\\ma5_1l;1O00000000001O00000O101O00000000001O000000001O0O10000O101O0O100O101O0O100O10001N10000O101N10000O100O2O000O10000O2O0O10000O2O0O10000O101O0O100O101O0HaMcE`2\\:cMbE]2]:eMaE\\2_:52F:G:E:F:G9F;F9Flm?;ko_OT2N17K400O2N1O2N1O101N1O2N1O1O2O0O2N1O1O2N1O101N1O2N1O1O2O0O2N1O1O2N1O101N2N2N2N3M2H8G9H8H`ea0"}, "label": "a small bed with an orange throw thrown across it"}]}
{"id": 183827, "image": "COCO_train2014_000000183827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bed on the right\"."}], "task": "refering", "answer_template": "The \"a red bed on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 204, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 293], "bbox": [0.6650625, 0.5831850117096019, 0.9320937499999999, 0.822248243559719], "iscrowd": 0, "area": 7948.228500000003, "rle": {"size": [427, 640], "counts": "\\ma5_1l;1O00000000001O00000O101O00000000001O000000001O0O10000O101O0O100O101O0O100O10001N10000O101N10000O100O2O000O10000O2O0O10000O2O0O10000O101O0O100O101O0HaMcE`2\\:cMbE]2]:eMaE\\2_:52F:G:E:F:G9F;F9Flm?;ko_OT2N17K400O2N1O2N1O101N1O2N1O1O2O0O2N1O1O2N1O101N1O2N1O1O2O0O2N1O1O2N1O101N2N2N2N3M2H8G9H8H`ea0"}, "label": "a red bed on the right"}]}
{"id": 405013, "image": "COCO_train2014_000000405013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the baby is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the baby is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 45, 57, 65, 66, 67, 68, 79, 80, 87, 88, 89, 90, 91, 101, 102, 103, 110, 111, 112, 113, 114, 123, 124, 125, 133, 134, 135, 136, 137, 146, 147, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229, 252, 274, 275, 298], "bbox": [0.37584375, 0.06966666666666667, 0.9977499999999999, 0.7258333333333333], "iscrowd": 0, "area": 24369.569549999997, "rle": {"size": [480, 640], "counts": "gTa34g>4L5K4L5K4L4L5L3M4O00001O0O101O001O00O1N2N2M3N2O100O1O1O0101O001O00001O00001O001O1O001N2O1O1O1O1O1O1O1O001O1O1O001O00100O0010O0100O0010O0102M3M100O100O1O010O100O1O1:E;F;nNXhf2EhXYM2M3M4L3N2M3M4M2[B[Ol<j0nBZOo<l0jBXOS=\\1N2M3M3M3M3M3M3L4M3SOm000O2O3L4N1N3M3M1O001O1O1O001O100O001O1O1O001O1O1O012M3M3M3M2N2O1N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2N2O0O2N2N10O0O101N1O2O0O2O0O10mGUJR7i5lH\\JR7d5lH_JT7`5jHcJU7^5gHfJY7Y5eHkJY7U5eHnJZ7R5cHRK]7m4aHVKP1WOk4d5RJYKQ1YOh4]5UJ]KR1]Ob4V5ZJ`KS1@^4_6aK`I`4b6^K]Ic4d6\\K[If4e6YKZIh4R7lJmHV5R7jJmHW5R7jJmHX5R7hJmHY5S7gJlHZ5T7fJkH\\5T7dJkH]5U7cJjH_5U7aJjH`5W7_JhHc5W7]JhHd5X7\\JgHf5X7ZJgHi5W7WJhHk5X7TJgHn5Y7QJfHR6[7kIdHW6^7fIaH]6`7`I_Hb6b7\\I]Hg6P7n0cIUc0"}, "label": "the chair the baby is sitting in"}]}
{"id": 405013, "image": "COCO_train2014_000000405013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a highchair\"."}], "task": "refering", "answer_template": "The \"the back of a highchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 45, 57, 65, 66, 67, 68, 79, 80, 87, 88, 89, 90, 91, 101, 102, 103, 110, 111, 112, 113, 114, 123, 124, 125, 133, 134, 135, 136, 137, 146, 147, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229, 252, 274, 275, 298], "bbox": [0.37584375, 0.06966666666666667, 0.9977499999999999, 0.7258333333333333], "iscrowd": 0, "area": 24369.569549999997, "rle": {"size": [480, 640], "counts": "gTa34g>4L5K4L5K4L4L5L3M4O00001O0O101O001O00O1N2N2M3N2O100O1O1O0101O001O00001O00001O001O1O001N2O1O1O1O1O1O1O1O001O1O1O001O00100O0010O0100O0010O0102M3M100O100O1O010O100O1O1:E;F;nNXhf2EhXYM2M3M4L3N2M3M4M2[B[Ol<j0nBZOo<l0jBXOS=\\1N2M3M3M3M3M3M3L4M3SOm000O2O3L4N1N3M3M1O001O1O1O001O100O001O1O1O001O1O1O012M3M3M3M2N2O1N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2N2O0O2N2N10O0O101N1O2O0O2O0O10mGUJR7i5lH\\JR7d5lH_JT7`5jHcJU7^5gHfJY7Y5eHkJY7U5eHnJZ7R5cHRK]7m4aHVKP1WOk4d5RJYKQ1YOh4]5UJ]KR1]Ob4V5ZJ`KS1@^4_6aK`I`4b6^K]Ic4d6\\K[If4e6YKZIh4R7lJmHV5R7jJmHW5R7jJmHX5R7hJmHY5S7gJlHZ5T7fJkH\\5T7dJkH]5U7cJjH_5U7aJjH`5W7_JhHc5W7]JhHd5X7\\JgHf5X7ZJgHi5W7WJhHk5X7TJgHn5Y7QJfHR6[7kIdHW6^7fIaH]6`7`I_Hb6b7\\I]Hg6P7n0cIUc0"}, "label": "the back of a highchair"}]}
{"id": 405013, "image": "COCO_train2014_000000405013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a brown shirt to the right of the baby\"."}], "task": "refering", "answer_template": "The \"a person in a brown shirt to the right of the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 70, 71, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234], "bbox": [0.0, 0.11179166666666666, 0.24015625, 0.6308124999999999], "iscrowd": 0, "area": 26165.137000000002, "rle": {"size": [480, 640], "counts": "f1U2k<Z4fKZ1fN0000O2OO100001N2O1O2N1O1N3N1O1O2N1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O1N101O1O001N2O001O1O0O2O1O1O001N2O001O1O0O2O1O001O0O10000000001N1000000000000O2O0000000000001O000O100000001O00000000001O1N2O1O1O1O1O1O1O1O1O1O1O2M4M2N2N2N3M2N2N2N3M2N2N2M4M2N2N2N3M2N2N2N3M2N2N2N3L3N2N2N3M2N4L4L5K4L5K4L4K6K:F<D=C<DZaS7"}, "label": "a person in a brown shirt to the right of the baby"}]}
{"id": 405013, "image": "COCO_train2014_000000405013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in orange sweater sitting near baby\"."}], "task": "refering", "answer_template": "The \"person in orange sweater sitting near baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 70, 71, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234], "bbox": [0.0, 0.11179166666666666, 0.24015625, 0.6308124999999999], "iscrowd": 0, "area": 26165.137000000002, "rle": {"size": [480, 640], "counts": "f1U2k<Z4fKZ1fN0000O2OO100001N2O1O2N1O1N3N1O1O2N1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O1N101O1O001N2O001O1O0O2O1O1O001N2O001O1O0O2O1O001O0O10000000001N1000000000000O2O0000000000001O000O100000001O00000000001O1N2O1O1O1O1O1O1O1O1O1O1O2M4M2N2N2N3M2N2N2N3M2N2N2M4M2N2N2N3M2N2N2N3M2N2N2N3L3N2N2N3M2N4L4L5K4L5K4L4K6K:F<D=C<DZaS7"}, "label": "person in orange sweater sitting near baby"}]}
{"id": 405013, "image": "COCO_train2014_000000405013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy baby eating own and wearing yellow color dress\"."}], "task": "refering", "answer_template": "The \"a boy baby eating own and wearing yellow color dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.22246875, 0.006750000000000001, 1.0, 0.8494375000000001], "iscrowd": 0, "area": 121140.85035000002, "rle": {"size": [480, 640], "counts": "^WS22n>3L3N2N2N2N3L3N2N2N2N2N3L3N2N2N2N1N3N1O2N1O2N1N3N1O2N1O2M2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1O2M2O2N1O2N2M3N3M3M3M3M2M4M3M3M3M2M4M3M3M3M2N3L4M3M3M2N1N010000000O1000O100000O1000O10O1O1O1O1O1O001O1O1O1O1O001O1O1O1N2O001O1N101O001N101O001N101O001N2O001O001O010O001O01O000010O0001O00001O01O0001O00001O01O000001N1O1O1O2O0O1O1O2N1001O2N2O0TNZJ]Jh5^5aJ\\J`5a5gJZJ[5a5nJ\\IP6a6Z1K5L3M4K4M4K4M4K5A>01O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O00001O001O00001O001O00010O001O00001O002N0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000000000000000000000000000000000O1001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O0002gFUMZ6l2^I\\M`6f2WIcMg6_2QIiMn6X2iHQNU7Q2bHXN]7i1[H_Nc7c1THfNj7\\1mGmNR8T1eGVOX8k3O1N2N2O1N2N2O2M2O1N1O101N100O2N100O2N1O2O0O1O2N1O1O2N3N2M3M2N1O2N101N1O2N1O2N1O2O0O2N1O1O2N1O2O0O2N1O2N1O2N101N1O2N1O2N1O2O0O2N3M2N3M2O2M3M2N2N2N2N2O1N1O2M3N2M3M3M3L4L3N3L4L4M2M2N3N1N2N3M2C=B?TOk0TOWH"}, "label": "a boy baby eating own and wearing yellow color dress"}]}
{"id": 405013, "image": "COCO_train2014_000000405013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby sitting in a highchair eating a snack\"."}], "task": "refering", "answer_template": "The \"a baby sitting in a highchair eating a snack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.22246875, 0.006750000000000001, 1.0, 0.8494375000000001], "iscrowd": 0, "area": 121140.85035000002, "rle": {"size": [480, 640], "counts": "^WS22n>3L3N2N2N2N3L3N2N2N2N2N3L3N2N2N2N1N3N1O2N1O2N1N3N1O2N1O2M2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1O2M2O2N1O2N2M3N3M3M3M3M2M4M3M3M3M2M4M3M3M3M2N3L4M3M3M2N1N010000000O1000O100000O1000O10O1O1O1O1O1O001O1O1O1O1O001O1O1O1N2O001O1N101O001N101O001N101O001N2O001O001O010O001O01O000010O0001O00001O01O0001O00001O01O000001N1O1O1O2O0O1O1O2N1001O2N2O0TNZJ]Jh5^5aJ\\J`5a5gJZJ[5a5nJ\\IP6a6Z1K5L3M4K4M4K4M4K5A>01O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O00001O001O00001O001O00010O001O00001O002N0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000000000000000000000000000000000O1001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O0002gFUMZ6l2^I\\M`6f2WIcMg6_2QIiMn6X2iHQNU7Q2bHXN]7i1[H_Nc7c1THfNj7\\1mGmNR8T1eGVOX8k3O1N2N2O1N2N2O2M2O1N1O101N100O2N100O2N1O2O0O1O2N1O1O2N3N2M3M2N1O2N101N1O2N1O2N1O2O0O2N1O1O2N1O2O0O2N1O2N1O2N101N1O2N1O2N1O2O0O2N3M2N3M2O2M3M2N2N2N2N2O1N1O2M3N2M3M3M3L4L3N3L4L4M2M2N3N1N2N3M2C=B?TOk0TOWH"}, "label": "a baby sitting in a highchair eating a snack"}]}
{"id": 126483, "image": "COCO_train2014_000000126483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white motorcycle down a mountain road\"."}], "task": "refering", "answer_template": "The \"a green and white motorcycle down a mountain road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 236, 237, 248, 249, 250, 251, 263, 264, 265, 266, 278, 279, 280, 281, 294, 295], "bbox": [0.5768396226415095, 0.647828125, 0.8192216981132076, 0.8719374999999999], "iscrowd": 0, "area": 8468.304499999997, "rle": {"size": [640, 424], "counts": "o`i46jc01O00001O0i^O2\\>O\\@NA>n?D_@6]O:Q`0@_@=ZO7T`0\\Oa@a0XO4W`0ZO`@f0VO2X`0YO_@P2`?PN^@T2a?kM]@X2b?iMZ@[2AaMh?3c@n2\\?VM\\@P3b?h000O0001O001O010O00000001N100O2O0O1O101O0000001N100O101OO1O001N2M3M3M3L4O1O2N100O1O2N100O2N1O2O0N3N2M2O2N2H9F9H8K5J5L4M4K4bNj]Ol0\\b0ROh]Oc0ab0]Ob]O>_b0Be]O9]b0Gm0010O010O1O1O1O1000000000001OO100O100Nkf_1"}, "label": "a green and white motorcycle down a mountain road"}]}
{"id": 126483, "image": "COCO_train2014_000000126483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green motorcycle in front\"."}], "task": "refering", "answer_template": "The \"the green motorcycle in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 236, 237, 248, 249, 250, 251, 263, 264, 265, 266, 278, 279, 280, 281, 294, 295], "bbox": [0.5768396226415095, 0.647828125, 0.8192216981132076, 0.8719374999999999], "iscrowd": 0, "area": 8468.304499999997, "rle": {"size": [640, 424], "counts": "o`i46jc01O00001O0i^O2\\>O\\@NA>n?D_@6]O:Q`0@_@=ZO7T`0\\Oa@a0XO4W`0ZO`@f0VO2X`0YO_@P2`?PN^@T2a?kM]@X2b?iMZ@[2AaMh?3c@n2\\?VM\\@P3b?h000O0001O001O010O00000001N100O2O0O1O101O0000001N100O101OO1O001N2M3M3M3L4O1O2N100O1O2N100O2N1O2O0N3N2M2O2N2H9F9H8K5J5L4M4K4bNj]Ol0\\b0ROh]Oc0ab0]Ob]O>_b0Be]O9]b0Gm0010O010O1O1O1O1000000000001OO100O100Nkf_1"}, "label": "the green motorcycle in front"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in yellow coat\"."}], "task": "refering", "answer_template": "The \"the man in yellow coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 83, 84, 85, 105, 106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 198, 201, 202, 225], "bbox": [0.584859375, 0.17303191489361702, 0.7974218750000001, 0.7101063829787234], "iscrowd": 0, "area": 10315.838250000003, "rle": {"size": [376, 640], "counts": "djY41f;1O101N101O0O101O1N101lFGg6:SILk65PI0P71iH6U7KfH:Z7G`H?_7BZHd0e7]OVHi0i7XOQHm0o7TOjGS1T8nNgGW1Y8jN`G]1_8dN^G^1a8cN_G^1_8dN`G]1^8eNaG[1^8fNbG[1\\8gNcGY1\\8iNcGX1[8iNeGW1T7kMmIZ3P6iLnIl0ROW1m6RNnIf0WOY1h6UNnIb0[OY1e6YNmI=@[1_6]NnI8D[1^6_NkI5I]1Z6aNiI3N\\1X6dNgIO3^1U6dNfIM7_1R6gNdIJ;`1o5iNcIF`0a1l5IUJ8j5FXJ:g5E[JgNb0@Q5i1^JgNW7X1jHhNU7Z1jHfNX7Z1fHgN[7Y1cHgN^7Z1`HfNb7Y1]HgNe7Y1YHgNi7Y1UHhNk7Y1SHgNo7X1PHhNR8X1lGhNV8X1hGhNY8Y1eGhN\\8W1cGiN_8W1_GiNc8U22N101N2N2N1O0001O002N2N2M3N1O2N2N2M3N1O2N2N2M3N1103M4L3M3M2M4M3M3L3M4M3L4L3M4M3L5K5K5K5J3N1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1Nc__1"}, "label": "the man in yellow coat"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skiier in red pants\"."}], "task": "refering", "answer_template": "The \"skiier in red pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 83, 84, 85, 105, 106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 198, 201, 202, 225], "bbox": [0.584859375, 0.17303191489361702, 0.7974218750000001, 0.7101063829787234], "iscrowd": 0, "area": 10315.838250000003, "rle": {"size": [376, 640], "counts": "djY41f;1O101N101O0O101O1N101lFGg6:SILk65PI0P71iH6U7KfH:Z7G`H?_7BZHd0e7]OVHi0i7XOQHm0o7TOjGS1T8nNgGW1Y8jN`G]1_8dN^G^1a8cN_G^1_8dN`G]1^8eNaG[1^8fNbG[1\\8gNcGY1\\8iNcGX1[8iNeGW1T7kMmIZ3P6iLnIl0ROW1m6RNnIf0WOY1h6UNnIb0[OY1e6YNmI=@[1_6]NnI8D[1^6_NkI5I]1Z6aNiI3N\\1X6dNgIO3^1U6dNfIM7_1R6gNdIJ;`1o5iNcIF`0a1l5IUJ8j5FXJ:g5E[JgNb0@Q5i1^JgNW7X1jHhNU7Z1jHfNX7Z1fHgN[7Y1cHgN^7Z1`HfNb7Y1]HgNe7Y1YHgNi7Y1UHhNk7Y1SHgNo7X1PHhNR8X1lGhNV8X1hGhNY8Y1eGhN\\8W1cGiN_8W1_GiNc8U22N101N2N2N1O0001O002N2N2M3N1O2N2N2M3N1O2N2N2M3N1103M4L3M3M2M4M3M3L3M4M3L4L3M4M3L5K5K5K5J3N1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1Nc__1"}, "label": "skiier in red pants"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis of the person in blue\"."}], "task": "refering", "answer_template": "The \"the skis of the person in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 215, 216, 220, 238, 239, 242, 264, 265], "bbox": [0.381578125, 0.6683510638297873, 0.602, 0.9093882978723404], "iscrowd": 0, "area": 1658.1348500000036, "rle": {"size": [376, 640], "counts": "Yli25R;a0A?H9I61O00010O00001O01O010O001h0WO=D8H\\Ya0Kgf^O1O2N1O1O1N101O001O001O001O001N2O001O001O001OO100O00100O1O100O1O100O100O01000000000O10000000000000000O100000000000000O10000000000000000O1000000000S\\m2"}, "label": "the skis of the person in blue"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis worn by a person in a light blue coat\"."}], "task": "refering", "answer_template": "The \"the skis worn by a person in a light blue coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 215, 216, 220, 238, 239, 242, 264, 265], "bbox": [0.381578125, 0.6683510638297873, 0.602, 0.9093882978723404], "iscrowd": 0, "area": 1658.1348500000036, "rle": {"size": [376, 640], "counts": "Yli25R;a0A?H9I61O00010O00001O01O010O001h0WO=D8H\\Ya0Kgf^O1O2N1O1O1N101O001O001O001O001N2O001O001O001OO100O00100O1O100O1O100O100O01000000000O10000000000000000O100000000000000O10000000000000000O1000000000S\\m2"}, "label": "the skis worn by a person in a light blue coat"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in a blue jacket\"."}], "task": "refering", "answer_template": "The \"a skier in a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 78, 79, 80, 81, 100, 101, 102, 103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 195, 196, 216, 218, 219, 242], "bbox": [0.35778125, 0.1101063829787234, 0.5795781249999999, 0.8292021276595744], "iscrowd": 0, "area": 15877.136, "rle": {"size": [376, 640], "counts": "eVd22e;4M4K4M4K3NO10O010O010O010O10OoG_Ol4b0oJFn49mJ0P50jJ9R5GjJb0Q5^OkJj0Q5WOmJn0o4ROoJT1m4mNQKX1k4hNTKZ1k4fNTK\\1k4eNRK^1m4bNRK_1n4aNQKa1n4`NPKb1n4_NPKd1o4\\NPKf1o4[NoJf1Q5ZNnJh1Q5XNnJf1U5[NhJc1\\5]NcJ`1a5`N^J^1e5cNYJZ1k5fNSJX1Q6hNnIW1S6kNjIV1W6jNhIV1Y6jNeIW1\\6jNaIX1_6hN`IX1a6hN]IY1d6hNYIY1h6gNVIZ1k6fNTIZ1m6gNPIZ1P7gNnHZ1S7fNkH[1V7fNhHZ1Y7fNeH[1\\7eNbH\\1_7eN_H[1c7dN[H]1g7bNWH_1k7aNRH`1P8_NoGa1S8P12N2N2N2N2N20O0M4L4L4L306J7I6J7J6I6J7I7J4K3M3M3N2M3M3M3N2M3M3M3N2M3M3O1N2N2N2N2N2N2NoMbKjL\\4i2TLVMk3[2eLeMY3o1UMRNh2j1^MVN`2f1fMZNX2a1PN_Nn1\\1XNdNi1Z1XNfNh1Z1XNZOT1f0lNXOV1h0kNUOX1j0hNSO[1[N\\Lg1Y2L]1RNdLS2o1HZ28fMF[2:gMB\\2>dM@^2`0bM^O_2c0aMZOf2b0ZM\\Ol2`0UM\\OR3`0nL^OX3>hL_O_3;cLCc37_LCk37XLAS49oK@\\4:fK_Oe4;]K^Oj4`0X3O2N1O2M2M3M4L3MccR3"}, "label": "a skier in a blue jacket"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a blue jacket skiing\"."}], "task": "refering", "answer_template": "The \"woman in a blue jacket skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 78, 79, 80, 81, 100, 101, 102, 103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 195, 196, 216, 218, 219, 242], "bbox": [0.35778125, 0.1101063829787234, 0.5795781249999999, 0.8292021276595744], "iscrowd": 0, "area": 15877.136, "rle": {"size": [376, 640], "counts": "eVd22e;4M4K4M4K3NO10O010O010O010O10OoG_Ol4b0oJFn49mJ0P50jJ9R5GjJb0Q5^OkJj0Q5WOmJn0o4ROoJT1m4mNQKX1k4hNTKZ1k4fNTK\\1k4eNRK^1m4bNRK_1n4aNQKa1n4`NPKb1n4_NPKd1o4\\NPKf1o4[NoJf1Q5ZNnJh1Q5XNnJf1U5[NhJc1\\5]NcJ`1a5`N^J^1e5cNYJZ1k5fNSJX1Q6hNnIW1S6kNjIV1W6jNhIV1Y6jNeIW1\\6jNaIX1_6hN`IX1a6hN]IY1d6hNYIY1h6gNVIZ1k6fNTIZ1m6gNPIZ1P7gNnHZ1S7fNkH[1V7fNhHZ1Y7fNeH[1\\7eNbH\\1_7eN_H[1c7dN[H]1g7bNWH_1k7aNRH`1P8_NoGa1S8P12N2N2N2N2N20O0M4L4L4L306J7I6J7J6I6J7I7J4K3M3M3N2M3M3M3N2M3M3M3N2M3M3O1N2N2N2N2N2N2NoMbKjL\\4i2TLVMk3[2eLeMY3o1UMRNh2j1^MVN`2f1fMZNX2a1PN_Nn1\\1XNdNi1Z1XNfNh1Z1XNZOT1f0lNXOV1h0kNUOX1j0hNSO[1[N\\Lg1Y2L]1RNdLS2o1HZ28fMF[2:gMB\\2>dM@^2`0bM^O_2c0aMZOf2b0ZM\\Ol2`0UM\\OR3`0nL^OX3>hL_O_3;cLCc37_LCk37XLAS49oK@\\4:fK_Oe4;]K^Oj4`0X3O2N1O2M2M3M4L3MccR3"}, "label": "woman in a blue jacket skiing"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ski under the right foot of a skier wearing yellow and red\"."}], "task": "refering", "answer_template": "The \"a ski under the right foot of a skier wearing yellow and red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 173, 174, 198, 199, 200], "bbox": [0.45528124999999997, 0.5055053191489361, 0.744640625, 0.6870744680851063], "iscrowd": 0, "area": 614.9719999999993, "rle": {"size": [376, 640], "counts": "o\\[31h;2M1000O10O1000O10O1000O3N]m90cRF0O2O0O2N01O01O01O01O01O01O01O010O00010O00010O00010O00010O0010O00010O00010O00010O00010O00010O01O01O01O01O01O01O01O010O00011N1O10\\g00cXO1O2O0O2N2N0001O01O01O00001O01O01O001O01O01O01O01O00010O000010O0001O01O01O00010O000010O00010O0001O01O01O00010O000010O0001O010O000010O00010O0001O010O00onk1"}, "label": "a ski under the right foot of a skier wearing yellow and red"}]}
{"id": 380440, "image": "COCO_train2014_000000380440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right ski of a skier wearing red pants\"."}], "task": "refering", "answer_template": "The \"the right ski of a skier wearing red pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 173, 174, 198, 199, 200], "bbox": [0.45528124999999997, 0.5055053191489361, 0.744640625, 0.6870744680851063], "iscrowd": 0, "area": 614.9719999999993, "rle": {"size": [376, 640], "counts": "o\\[31h;2M1000O10O1000O10O1000O3N]m90cRF0O2O0O2N01O01O01O01O01O01O01O010O00010O00010O00010O00010O0010O00010O00010O00010O00010O00010O01O01O01O01O01O01O01O010O00011N1O10\\g00cXO1O2O0O2N2N0001O01O01O00001O01O01O001O01O01O01O01O00010O000010O0001O01O01O00010O000010O00010O0001O01O01O00010O000010O0001O010O000010O00010O0001O010O00onk1"}, "label": "the right ski of a skier wearing red pants"}]}
{"id": 429594, "image": "COCO_train2014_000000429594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse on the left\"."}], "task": "refering", "answer_template": "The \"the horse on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 70, 71, 91, 92, 93, 111, 112, 113, 114, 133, 134, 135, 154, 155, 156, 176, 177, 197, 198, 218, 219, 240], "bbox": [0.3219198664440735, 0.16183770883054893, 0.47085141903171956, 0.7446300715990454], "iscrowd": 0, "area": 10879.962600000004, "rle": {"size": [419, 599], "counts": "iR_21Q=0O1QD4g:MUE9i:HRE>k:CREc0k:_OoDg0n:@hDg0U;e0M4M2M4oN]NRFQ2n9oMRFR2m9nMRFS2o9b0O1O0N3M3M2O2M2N2N2O1NXNfFIX97iFJU96mFJQ96QGJl87UGJi87XGIf87[GJb87`GJ]86eGJY86hGKU86lGKS84mGOR8OoG2P8MPH5P8JPH7l7LTH6g7MXH5d7nMoGf1==`7mMVHe1;?[7lM]He17a0X7jMeHd13d0S7hMmHd10h0l6dMWId1Ln0g6\\MaIe1HT1`6VMkIe1FY1Y6QMTJf1B_18jLT6g1dId1InLe6\\1cIY2]6gMbIY2_6gMaIY2_6gM`IY2a6fM`IY2`6hM`IX2`6hM_IX2b6hM^IW2b6iM_IW2a6iM_IV2b6jM^IU2b6jM_IW2a6gMaIX2`6eMcIZ2]6eMeI[2[6bMhI]2Y6aMhI_2X6_MkIa2U6]MmIb2S6\\MPJc2Q6ZMQJg2n5XMSJh2n5UMTJk2m5RMTJP3k5nLWJR3j5jLXJW3h5fLZJX1kNOl6gNZJU1]OGX6SO[JR1FFQ6VOZJo0MFk5ZOXJk06Fe5]OVJg0>F`5BQJd0g0E\\5DQJ>m0HV5GPJ9S1KQ5IPJ3X1O_gQ4"}, "label": "the horse on the left"}]}
{"id": 429594, "image": "COCO_train2014_000000429594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse being led by the bridle by the right hand of a man in overalls and a hat\"."}], "task": "refering", "answer_template": "The \"a brown horse being led by the bridle by the right hand of a man in overalls and a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 70, 71, 91, 92, 93, 111, 112, 113, 114, 133, 134, 135, 154, 155, 156, 176, 177, 197, 198, 218, 219, 240], "bbox": [0.3219198664440735, 0.16183770883054893, 0.47085141903171956, 0.7446300715990454], "iscrowd": 0, "area": 10879.962600000004, "rle": {"size": [419, 599], "counts": "iR_21Q=0O1QD4g:MUE9i:HRE>k:CREc0k:_OoDg0n:@hDg0U;e0M4M2M4oN]NRFQ2n9oMRFR2m9nMRFS2o9b0O1O0N3M3M2O2M2N2N2O1NXNfFIX97iFJU96mFJQ96QGJl87UGJi87XGIf87[GJb87`GJ]86eGJY86hGKU86lGKS84mGOR8OoG2P8MPH5P8JPH7l7LTH6g7MXH5d7nMoGf1==`7mMVHe1;?[7lM]He17a0X7jMeHd13d0S7hMmHd10h0l6dMWId1Ln0g6\\MaIe1HT1`6VMkIe1FY1Y6QMTJf1B_18jLT6g1dId1InLe6\\1cIY2]6gMbIY2_6gMaIY2_6gM`IY2a6fM`IY2`6hM`IX2`6hM_IX2b6hM^IW2b6iM_IW2a6iM_IV2b6jM^IU2b6jM_IW2a6gMaIX2`6eMcIZ2]6eMeI[2[6bMhI]2Y6aMhI_2X6_MkIa2U6]MmIb2S6\\MPJc2Q6ZMQJg2n5XMSJh2n5UMTJk2m5RMTJP3k5nLWJR3j5jLXJW3h5fLZJX1kNOl6gNZJU1]OGX6SO[JR1FFQ6VOZJo0MFk5ZOXJk06Fe5]OVJg0>F`5BQJd0g0E\\5DQJ>m0HV5GPJ9S1KQ5IPJ3X1O_gQ4"}, "label": "a brown horse being led by the bridle by the right hand of a man in overalls and a hat"}]}
{"id": 429594, "image": "COCO_train2014_000000429594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse holding by a man by his left hand\"."}], "task": "refering", "answer_template": "The \"a horse holding by a man by his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 96, 97, 98, 99, 117, 118, 119, 120, 138, 139, 140, 141, 159, 160, 161, 162, 180, 181, 182, 201, 202, 203, 223, 224, 244, 245], "bbox": [0.582170283806344, 0.19408114558472553, 0.7551585976627714, 0.7747255369928401], "iscrowd": 0, "area": 14119.25165, "rle": {"size": [419, 599], "counts": "Wn^46i<7J7H7J4M3N2N2iD]Ok9f0RFH_9;`F1S90mF4n8MQG6l8KTG7i8IWG:e8H[G:6VNc7a1WH<OXNh7]1XH>KYNk6GiIb1A`0E\\NQ7DeIa1Ea0@^NV7BbI_1GV1K\\M[62oI]1KV1EaMa6NlIZ1Oj1T6mLjIZ11k1U6mLgIY13k1V6lLeI]12i1W6kLfI`1Ng1\\6iLdId1Md1_6iLbIf1L`1d6jL^I]5c6cJ\\I]5e6`0001ON2O1O1O1N01O10WJdIl4[6TKgIl4X6SKkIl4S6SKPJn4n5QKTJo4j5QKXJo4f5PK]JP5b5oJaJQ5\\5oJfJQ5Y5kJlJW5P5gJTK[5j4aJ[K`5f4XJ_Ki5m52N1O01OO1000O0100ZISJV6l5jIUJV6k5jIVJg4b0bKY5FVJe4h0aKR5IXJl3J\\LR1Mk4LZJf31YLP12f4NZJc38VLm07a4O\\J`3m1`Le31_J]3P2_La35`JY3S2`L\\37bJW3U2aLX38eJT3W2bLS38iJT3R8iLSHT3n7iLVHl2UOWMf8JZHl2T9M4L4M5J5N3O001O0O1000O00O2\\NVEj0l:UOWEh0j:VOZEg0h:VO\\Eg0e:WO_Ef0b:XObEe0`:XOcEf0^:XOfEe0[:YOiEd0^;L5L5K5KZnk1"}, "label": "a horse holding by a man by his left hand"}]}
{"id": 429594, "image": "COCO_train2014_000000429594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black horse with a white diamond on its head\"."}], "task": "refering", "answer_template": "The \"a black horse with a white diamond on its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 96, 97, 98, 99, 117, 118, 119, 120, 138, 139, 140, 141, 159, 160, 161, 162, 180, 181, 182, 201, 202, 203, 223, 224, 244, 245], "bbox": [0.582170283806344, 0.19408114558472553, 0.7551585976627714, 0.7747255369928401], "iscrowd": 0, "area": 14119.25165, "rle": {"size": [419, 599], "counts": "Wn^46i<7J7H7J4M3N2N2iD]Ok9f0RFH_9;`F1S90mF4n8MQG6l8KTG7i8IWG:e8H[G:6VNc7a1WH<OXNh7]1XH>KYNk6GiIb1A`0E\\NQ7DeIa1Ea0@^NV7BbI_1GV1K\\M[62oI]1KV1EaMa6NlIZ1Oj1T6mLjIZ11k1U6mLgIY13k1V6lLeI]12i1W6kLfI`1Ng1\\6iLdId1Md1_6iLbIf1L`1d6jL^I]5c6cJ\\I]5e6`0001ON2O1O1O1N01O10WJdIl4[6TKgIl4X6SKkIl4S6SKPJn4n5QKTJo4j5QKXJo4f5PK]JP5b5oJaJQ5\\5oJfJQ5Y5kJlJW5P5gJTK[5j4aJ[K`5f4XJ_Ki5m52N1O01OO1000O0100ZISJV6l5jIUJV6k5jIVJg4b0bKY5FVJe4h0aKR5IXJl3J\\LR1Mk4LZJf31YLP12f4NZJc38VLm07a4O\\J`3m1`Le31_J]3P2_La35`JY3S2`L\\37bJW3U2aLX38eJT3W2bLS38iJT3R8iLSHT3n7iLVHl2UOWMf8JZHl2T9M4L4M5J5N3O001O0O1000O00O2\\NVEj0l:UOWEh0j:VOZEg0h:VO\\Eg0e:WO_Ef0b:XObEe0`:XOcEf0^:XOfEe0[:YOiEd0^;L5L5K5KZnk1"}, "label": "a black horse with a white diamond on its head"}]}
{"id": 183835, "image": "COCO_train2014_000000183835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black motorcycle on left\"."}], "task": "refering", "answer_template": "The \"black motorcycle on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 142, 143, 165, 166, 167, 188, 189, 190, 191, 192, 193, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 394, 395, 396, 397, 400, 401, 402, 403, 417, 418, 419, 424, 425, 440, 441], "bbox": [0.139328125, 0.25225, 0.638203125, 0.85], "iscrowd": 0, "area": 63905.01035000002, "rle": {"size": [640, 640], "counts": "PTh13ic07H8I7H8I7H8I7H8H8I6I5F:E;F:E;H8M2O2N2M3N2M3dH_KWOc4j0^KROd4n0^KmH3i4b4[2\\KhH9f4]4b2]KeH:b4\\4i2\\KdH;]4[4P3[KaH>Y4X4W3\\K_H?S4X4_3ZK\\H?R4Y4b3ZKZH>Q4Z4e3ZKYH<o3\\4i3XKWH=m3]4l3WKVH=k3^4o3VKVH<h3`4S4TKTH=f3a4V4SKSH=d3b4Z4QKRH=a3d4]4PKQH<=oNZ2g5X5nJQH<=POW2h5\\5lJPH<<QOU2h5`5lJnG;=QOS2j5c5iJmG<=ROP2k5f5hJmG;<SOn1l5i5gJlG:=SOl1m5m5eJjG;=TOi1l5R6fJhG9=VOf1k5W6gJeG9=VOd1j5]6fJbG:>UOa1j5b6hJ_G8>WO^1i5h6hJ[G9>WO\\1h5m6hJYG9?VOX1j5R7hJWG7?XOS1j5[7gJRG7`0YOo0j5a7fJPG8`0XOk0k5h7eJmF7`0YOh0k5n7fJiF6a0ZOd0k5T8fJgF5`0[O`0l5\\8dJcF5b0[O;m5c8cJ_F5c0[O8m5j8cJ[F4c0]O4m5Q9bJWF5d0\\O0n5X9`JTF6d0\\OMP6]9^JRF5d0^OHR6d9[JoE6e0]ODT6j9ZKbFcN@U6P:WK`FdN]OW6U:TK_F\\5d9cJ[F\\5g9cJYF\\5i9cJXF[5j9eJUFZ5m9eJSFZ5o9eJQF[5P:dJQFZ5Q:fJnEY5T:fJmEX5U:gJkEX5X:fJhEY5Z:fJ[Ed5g:i12N2N2N2N2N2N2N3M2N2N2N2N2N2L4L4L4L5L3L4N23M3M1N1O0000001O000O101O0000001O000000001O0000000000001O000000000000001O0O1000000000001O0001O1O1O1O1O010O1O1O1O1O1O001O100O1O1O1O1O1O1O100O1O1O1O00000010O000000101N2N1O2N2O1N2N1O2O1N2N1O2N2O1N1O2N2N2O1N1O2N2N2O0O1O00O10O10O1K5K5L31000O1000O10O100000O010000O1000O0100O1000O010000O10O10O10000O01000O1bLnCnNR<n0WDlNj;P1_DjNa;T1gDfNZ;V1PEcNR;X1WEbNj:Z1_E`Nc:[1gE^NZ:^1PF[NR:`1YFXNh9f1aFRNa9l1fFmM[9R2mFfMU9Y2QG`MP9_2Z4N100O2N101N1O2O0O2N101N1O2O0O2N101N1O2O1N2N2O1N2O2M2N2O1N2N3N1N2N2O1N2N3N1M3N2M3N2M3N2M3N2N2M3N2M3N2M3Nif`4"}, "label": "black motorcycle on left"}]}
{"id": 183835, "image": "COCO_train2014_000000183835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike on the left\"."}], "task": "refering", "answer_template": "The \"the bike on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 142, 143, 165, 166, 167, 188, 189, 190, 191, 192, 193, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 394, 395, 396, 397, 400, 401, 402, 403, 417, 418, 419, 424, 425, 440, 441], "bbox": [0.139328125, 0.25225, 0.638203125, 0.85], "iscrowd": 0, "area": 63905.01035000002, "rle": {"size": [640, 640], "counts": "PTh13ic07H8I7H8I7H8I7H8H8I6I5F:E;F:E;H8M2O2N2M3N2M3dH_KWOc4j0^KROd4n0^KmH3i4b4[2\\KhH9f4]4b2]KeH:b4\\4i2\\KdH;]4[4P3[KaH>Y4X4W3\\K_H?S4X4_3ZK\\H?R4Y4b3ZKZH>Q4Z4e3ZKYH<o3\\4i3XKWH=m3]4l3WKVH=k3^4o3VKVH<h3`4S4TKTH=f3a4V4SKSH=d3b4Z4QKRH=a3d4]4PKQH<=oNZ2g5X5nJQH<=POW2h5\\5lJPH<<QOU2h5`5lJnG;=QOS2j5c5iJmG<=ROP2k5f5hJmG;<SOn1l5i5gJlG:=SOl1m5m5eJjG;=TOi1l5R6fJhG9=VOf1k5W6gJeG9=VOd1j5]6fJbG:>UOa1j5b6hJ_G8>WO^1i5h6hJ[G9>WO\\1h5m6hJYG9?VOX1j5R7hJWG7?XOS1j5[7gJRG7`0YOo0j5a7fJPG8`0XOk0k5h7eJmF7`0YOh0k5n7fJiF6a0ZOd0k5T8fJgF5`0[O`0l5\\8dJcF5b0[O;m5c8cJ_F5c0[O8m5j8cJ[F4c0]O4m5Q9bJWF5d0\\O0n5X9`JTF6d0\\OMP6]9^JRF5d0^OHR6d9[JoE6e0]ODT6j9ZKbFcN@U6P:WK`FdN]OW6U:TK_F\\5d9cJ[F\\5g9cJYF\\5i9cJXF[5j9eJUFZ5m9eJSFZ5o9eJQF[5P:dJQFZ5Q:fJnEY5T:fJmEX5U:gJkEX5X:fJhEY5Z:fJ[Ed5g:i12N2N2N2N2N2N2N3M2N2N2N2N2N2L4L4L4L5L3L4N23M3M1N1O0000001O000O101O0000001O000000001O0000000000001O000000000000001O0O1000000000001O0001O1O1O1O1O010O1O1O1O1O1O001O100O1O1O1O1O1O1O100O1O1O1O00000010O000000101N2N1O2N2O1N2N1O2O1N2N1O2N2O1N1O2N2N2O1N1O2N2N2O0O1O00O10O10O1K5K5L31000O1000O10O100000O010000O1000O0100O1000O010000O10O10O10000O01000O1bLnCnNR<n0WDlNj;P1_DjNa;T1gDfNZ;V1PEcNR;X1WEbNj:Z1_E`Nc:[1gE^NZ:^1PF[NR:`1YFXNh9f1aFRNa9l1fFmM[9R2mFfMU9Y2QG`MP9_2Z4N100O2N101N1O2O0O2N101N1O2O0O2N101N1O2O1N2N2O1N2O2M2N2O1N2N3N1N2N2O1N2N3N1M3N2M3N2M3N2M3N2N2M3N2M3N2M3Nif`4"}, "label": "the bike on the left"}]}
{"id": 183835, "image": "COCO_train2014_000000183835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red color motorcycle\"."}], "task": "refering", "answer_template": "The \"a red color motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 106, 107, 128, 129, 130, 131, 151, 152, 153, 154, 176, 177, 178, 179, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387, 388, 404, 405, 406, 407, 408, 409, 410, 427, 428, 429, 430, 431, 432, 433, 450, 451, 452, 453, 473, 474], "bbox": [0.53034375, 0.140078125, 0.97303125, 0.8764062499999999], "iscrowd": 0, "area": 66302.60549999999, "rle": {"size": [640, 640], "counts": "XYd62ac0>B>A?B>CVN]^OP2`a0QN`^OR2\\a0oMd^OS2Ya0nMg^OT2Va0mMj^OU2Sa0lMm^OV2Pa0kMP_OW2m`0jMS_OX2j`0iMV_OY2g`0hMY_O[2d`0eM\\_O]2a`0dM__O^2^`0dMa_O^2]`0bMc_O`2Z`0aMf_Oa2W`0aMh_Oa2V`0_Mj_Od2R`0]Mn_Oe2o?]MP@f2m?ZMS@i2i?XMW@l2e?UMZ@n2b?SM^@P3^?RMa@Q3\\?oLd@T3X?mLh@V3V?m0TKXKZJd0\\K`3n9SLeJ=^Ki3]9RLTK5^KS4R9kK`K1_K\\4i8dKgKOaKf4^8\\KPLOaKo4V8SKXLMcKY5k7lJaLKdKa5c7dJiLJdKl5Z7[JPMIgKU6o6SJZMHgK[6i6nI_MFiK\\6g6oI`MDiK^6g6oI_MCjK^6g6PJ]MBmK^6e6QJ^MAlK_6f6QJ]M_OnK`6d6RJ^M^OnK`6d6SJ\\M^OoK_6f6SJ[M^OoK_6U5QISLT1h2\\OPL_6U5TIPLR1j2[OPL`6V5WIkKo0o2ZOPL`6V5[IgKl0Q3ZORL_6V5^IdKi0T3ZOQL`6X5`I^Kh0X3XORL`6X5cI[Ke0[3WOTL`6V5hIVKc0^3VOVL_6V5lIRK`0a3UOWL_6V5oIoJ=d3UOWL_6W5RJiJ<h3SOYL^6V5WJeJ9j3SO[L]6V5jK_NiM[L]6V5kK^NhM\\L]6V5lK]NgM^L[6V5oKZNgM`LZ6W5oKXNgMaLZ6W5PLWNeMdLZ6U5RLUNeMfLY6U5RLUNeMfLY6U5SLTNdMgLY6V5SLRNdMiLX6U5ULPNdMkLW6U5VLoMcMlLW6U5WLnMbMmLW6U5XLmMoLnL31e6T5YL[NSMaLd6U5YLYNSMbLd6U5ZLWNRMeLd6T5[LVNQMgLc6S5]LTNRMhL`6U5_LPNSMkL^6V5_LkMVMoL[6V5_LiMXMRMX6U5aLfMYMUMV6U5bLaLiLiNb0AS6U5cL_LoLcN=JP6T5eL]LTM^N91n5U5eLZL[MWN69i5W5gLWL^NbNk4W5gLVL_NdNh4W5jLTL^NeNh4W5kLRL`NeNd4[5lLnKdNdN`4c5hLgKlNdNZ4m5dL]KWObNU4X6^LTK@bNQ4b6XLkJ[6]5^IaJ1eNl3R7mKWJi6P6QInIo6U6PIiIP7Y6PIfIP7\\6oHdIo6^6PIbIP7_6PI`Io6b6QI^In6d6QI[In6g6RIYIm6h6SIWIl6l6SITIl6m6SISIl6o6TIPIl6R7SInHk6T7UIkHj6W7VIiHi6Y7VIfHi6\\7WIdHh6]7WIcHh6`7WI_Hi6b7WI^Hg6d7YI[Hg6g7XIYHf6i7ZIVHf6k7ZIUHd6n7ZIRHf6o7ZIQHd6Q8\\InGc6U8\\IjGd6W8\\IiGb6Y8^IfGb6\\8^IcG`6_8bI^G^6c8dIZG[6i8eIUG[6l8gIPGY6R9iIkFW6W9jIfFU6\\9mIaFR6a9V22M2O1N3N1N3M2N2N2O0O001O1O100O001O1O1O010O1O1O1O00100O1O1O2N101N2N2N1O2N2O1O01000000O1000000O1000000O11O2M4M2M3M4mNfDYIZ;e6hDZIX;f6iDYIW;f6lDYIS;f6PEXIP;g6REXIo:f6TEYIk:f6WEYIi:g6YEWIg:h6\\EWIc:h6_EWIa:h6bEVI_:h6cEXI\\:g6gEVIZ:j6hETIX:k6jESIW:l6lERIU:l6mESIS:`3dDE\\1hLQ:^3iDIg<1_CNb<MdC1]<IiC6Y<DmC:T<_OSD`0n;XO[Dg0c;SOfDk0Z;mNoDR1Q;jNUET1j:oNUEP1k:QOVEm0j:UOVEj0h:XOXEg0h:[OXE`0k:BTE9o:ITE0o:2fETO\\:o0XFWNm9i1U41N1000000000000000000000O1N2M2N3N2M3M3N2M3N2M3M3O100O100O010O1O100O100O100O1O104K6K5J6K5J6J6Kl[:"}, "label": "a red color motorcycle"}]}
{"id": 183835, "image": "COCO_train2014_000000183835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front end of an orange motor cycle with chrome trim\"."}], "task": "refering", "answer_template": "The \"the front end of an orange motor cycle with chrome trim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 106, 107, 128, 129, 130, 131, 151, 152, 153, 154, 176, 177, 178, 179, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387, 388, 404, 405, 406, 407, 408, 409, 410, 427, 428, 429, 430, 431, 432, 433, 450, 451, 452, 453, 473, 474], "bbox": [0.53034375, 0.140078125, 0.97303125, 0.8764062499999999], "iscrowd": 0, "area": 66302.60549999999, "rle": {"size": [640, 640], "counts": "XYd62ac0>B>A?B>CVN]^OP2`a0QN`^OR2\\a0oMd^OS2Ya0nMg^OT2Va0mMj^OU2Sa0lMm^OV2Pa0kMP_OW2m`0jMS_OX2j`0iMV_OY2g`0hMY_O[2d`0eM\\_O]2a`0dM__O^2^`0dMa_O^2]`0bMc_O`2Z`0aMf_Oa2W`0aMh_Oa2V`0_Mj_Od2R`0]Mn_Oe2o?]MP@f2m?ZMS@i2i?XMW@l2e?UMZ@n2b?SM^@P3^?RMa@Q3\\?oLd@T3X?mLh@V3V?m0TKXKZJd0\\K`3n9SLeJ=^Ki3]9RLTK5^KS4R9kK`K1_K\\4i8dKgKOaKf4^8\\KPLOaKo4V8SKXLMcKY5k7lJaLKdKa5c7dJiLJdKl5Z7[JPMIgKU6o6SJZMHgK[6i6nI_MFiK\\6g6oI`MDiK^6g6oI_MCjK^6g6PJ]MBmK^6e6QJ^MAlK_6f6QJ]M_OnK`6d6RJ^M^OnK`6d6SJ\\M^OoK_6f6SJ[M^OoK_6U5QISLT1h2\\OPL_6U5TIPLR1j2[OPL`6V5WIkKo0o2ZOPL`6V5[IgKl0Q3ZORL_6V5^IdKi0T3ZOQL`6X5`I^Kh0X3XORL`6X5cI[Ke0[3WOTL`6V5hIVKc0^3VOVL_6V5lIRK`0a3UOWL_6V5oIoJ=d3UOWL_6W5RJiJ<h3SOYL^6V5WJeJ9j3SO[L]6V5jK_NiM[L]6V5kK^NhM\\L]6V5lK]NgM^L[6V5oKZNgM`LZ6W5oKXNgMaLZ6W5PLWNeMdLZ6U5RLUNeMfLY6U5RLUNeMfLY6U5SLTNdMgLY6V5SLRNdMiLX6U5ULPNdMkLW6U5VLoMcMlLW6U5WLnMbMmLW6U5XLmMoLnL31e6T5YL[NSMaLd6U5YLYNSMbLd6U5ZLWNRMeLd6T5[LVNQMgLc6S5]LTNRMhL`6U5_LPNSMkL^6V5_LkMVMoL[6V5_LiMXMRMX6U5aLfMYMUMV6U5bLaLiLiNb0AS6U5cL_LoLcN=JP6T5eL]LTM^N91n5U5eLZL[MWN69i5W5gLWL^NbNk4W5gLVL_NdNh4W5jLTL^NeNh4W5kLRL`NeNd4[5lLnKdNdN`4c5hLgKlNdNZ4m5dL]KWObNU4X6^LTK@bNQ4b6XLkJ[6]5^IaJ1eNl3R7mKWJi6P6QInIo6U6PIiIP7Y6PIfIP7\\6oHdIo6^6PIbIP7_6PI`Io6b6QI^In6d6QI[In6g6RIYIm6h6SIWIl6l6SITIl6m6SISIl6o6TIPIl6R7SInHk6T7UIkHj6W7VIiHi6Y7VIfHi6\\7WIdHh6]7WIcHh6`7WI_Hi6b7WI^Hg6d7YI[Hg6g7XIYHf6i7ZIVHf6k7ZIUHd6n7ZIRHf6o7ZIQHd6Q8\\InGc6U8\\IjGd6W8\\IiGb6Y8^IfGb6\\8^IcG`6_8bI^G^6c8dIZG[6i8eIUG[6l8gIPGY6R9iIkFW6W9jIfFU6\\9mIaFR6a9V22M2O1N3N1N3M2N2N2O0O001O1O100O001O1O1O010O1O1O1O00100O1O1O2N101N2N2N1O2N2O1O01000000O1000000O1000000O11O2M4M2M3M4mNfDYIZ;e6hDZIX;f6iDYIW;f6lDYIS;f6PEXIP;g6REXIo:f6TEYIk:f6WEYIi:g6YEWIg:h6\\EWIc:h6_EWIa:h6bEVI_:h6cEXI\\:g6gEVIZ:j6hETIX:k6jESIW:l6lERIU:l6mESIS:`3dDE\\1hLQ:^3iDIg<1_CNb<MdC1]<IiC6Y<DmC:T<_OSD`0n;XO[Dg0c;SOfDk0Z;mNoDR1Q;jNUET1j:oNUEP1k:QOVEm0j:UOVEj0h:XOXEg0h:[OXE`0k:BTE9o:ITE0o:2fETO\\:o0XFWNm9i1U41N1000000000000000000000O1N2M2N3N2M3M3N2M3N2M3M3O100O100O010O1O100O100O100O1O104K6K5J6K5J6J6Kl[:"}, "label": "the front end of an orange motor cycle with chrome trim"}]}
{"id": 183836, "image": "COCO_train2014_000000183836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"salad\"."}], "task": "refering", "answer_template": "The \"salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 134, 135, 136, 137, 138, 139, 140, 147, 148, 149, 150, 151, 152, 153, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 200, 201, 202, 203, 204], "bbox": [0.3389066666666667, 0.49956, 0.8485333333333335, 0.88912], "iscrowd": 0, "area": 26981.582550000006, "rle": {"size": [500, 375], "counts": "iZn17Y?`0A>A`0A>B>A4M4K5L4L4K5L4L4K5L3L5L4L4K5L4K5L4N20O2O001O00001O001O0O1000O1O1O1O1O1O001O00001O0O10001O00001O00001O000M4H7I7K501O00001O01O01O00001O00000010O0001O00001O000010O0001O0000001O00010OO100000000O10001O00000O1000001O001iNVFaLk9_3VF`Lj9_3WFaLj9]3WFcLi9]3WFcLj9[3WFdLj9[3XFdLh9\\3XFdLi9Z3XFfLh9Y3YFgLh9X3XFhLh9W3ZFgLg9X3ZFhLg9V3ZFjLf9V3ZFjLg9T3ZFlLf9S3[FmLe9S3\\FkLf9T3ZFlLf9S3[FmLf9R3ZFnLf9R3ZFnLf9R3[FlLg9R3ZFnLf9R3ZFmLh9R3XFmLi9S3WFlLj9S3XFkLj9T3VFkLk9U3UFjLm9T3TFkLm9U3SFjLn9V3SFhLo9W3QFhLP:W3QFhLQ:W3oEhLR:X3nEgLS:Y3nEfLS:X3nEgLS:Y3mEfLU:Y3kEfLV:Y3kEfLV:Z3jEeLX:Z3iEdLX:\\3hEcLZ:[3gEdLZ:\\3fEcL[:`0iD[2V<cMlC\\2T<cMnC[2T<cMnC\\2R<cMPD\\2Q<bMPD^2P<aMRD]2o;bMSD3Co1[<kMUD4BP2Y<kMWD3BP2Y<kMWD3AR2o<kMTCT2W=O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1I7E:E<Dc`k0"}, "label": "salad"}]}
{"id": 183836, "image": "COCO_train2014_000000183836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shrimp salad on a white plate next to a fork with a beer in front\"."}], "task": "refering", "answer_template": "The \"a shrimp salad on a white plate next to a fork with a beer in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 134, 135, 136, 137, 138, 139, 140, 147, 148, 149, 150, 151, 152, 153, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 200, 201, 202, 203, 204], "bbox": [0.3389066666666667, 0.49956, 0.8485333333333335, 0.88912], "iscrowd": 0, "area": 26981.582550000006, "rle": {"size": [500, 375], "counts": "iZn17Y?`0A>A`0A>B>A4M4K5L4L4K5L4L4K5L3L5L4L4K5L4K5L4N20O2O001O00001O001O0O1000O1O1O1O1O1O001O00001O0O10001O00001O00001O000M4H7I7K501O00001O01O01O00001O00000010O0001O00001O000010O0001O0000001O00010OO100000000O10001O00000O1000001O001iNVFaLk9_3VF`Lj9_3WFaLj9]3WFcLi9]3WFcLj9[3WFdLj9[3XFdLh9\\3XFdLi9Z3XFfLh9Y3YFgLh9X3XFhLh9W3ZFgLg9X3ZFhLg9V3ZFjLf9V3ZFjLg9T3ZFlLf9S3[FmLe9S3\\FkLf9T3ZFlLf9S3[FmLf9R3ZFnLf9R3ZFnLf9R3[FlLg9R3ZFnLf9R3ZFmLh9R3XFmLi9S3WFlLj9S3XFkLj9T3VFkLk9U3UFjLm9T3TFkLm9U3SFjLn9V3SFhLo9W3QFhLP:W3QFhLQ:W3oEhLR:X3nEgLS:Y3nEfLS:X3nEgLS:Y3mEfLU:Y3kEfLV:Y3kEfLV:Z3jEeLX:Z3iEdLX:\\3hEcLZ:[3gEdLZ:\\3fEcL[:`0iD[2V<cMlC\\2T<cMnC[2T<cMnC\\2R<cMPD\\2Q<bMPD^2P<aMRD]2o;bMSD3Co1[<kMUD4BP2Y<kMWD3BP2Y<kMWD3AR2o<kMTCT2W=O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1I7E:E<Dc`k0"}, "label": "a shrimp salad on a white plate next to a fork with a beer in front"}]}
{"id": 323108, "image": "COCO_train2014_000000323108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a boy petting the head of a giraffe\"."}], "task": "refering", "answer_template": "The \"this is a boy petting the head of a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 36, 37, 54, 55, 72, 73, 90, 91, 98, 99, 108, 109, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 180, 181, 198, 199, 200, 216, 217, 218], "bbox": [0.0, 0.10786666666666668, 0.54776, 0.9842666666666666], "iscrowd": 0, "area": 23128.705250000003, "rle": {"size": [375, 500], "counts": "b9o1j5n3SLm3C=1O00001O001O0oI^Nm1c1RN`Nk1`1TNdNi1]1VNeNh1[1WNiNf1X1YNhNg1X1XNiNh1X1WNiNh1W1WNjNi1W1VNiNj1W1VNiNj1X1TNjNk1V1UNjNk1W1TNiNl1W1TNiNl1X1SNiNl1W1TNiNl1X1SNhNm1X1SNhNm1Y1SNgNl1Y1TNgNl1Y1TNgNl1Z1SNfNm1Z1SNfNm1[1RNfNm1[1RNeNn1]1PNcNP2_1nMaNR2a1lM`NS2b1kM^NU2d1iM\\NW2f1gMZNY2i1eMWNZ2k1dMUN\\2m1bMSN^2o1`MQN`2Q2bLkN^3[1\\KVMNe1e4Z1RKXM7_1f4_1gJ[M`0W1i4U2VKkMj4W2SKkMl4V2RKkMR3UN[NR4`NjMQ3[N\\Nm3aNhMo2cN\\Nf3cNiMl2hN^Na3cNhMk2oN^NZ3fNgMh2VO_NV5]1QK`NP5[1XKaNh4[1_KbNa4Z1fKcN[4[1jKaNV4^1nK_NS4_1QL^No3a1VL[Nk3c1YLZNg3e1^LWNc3g1bLUN_3h1Y3O2N1O2N1O2N1O2N1O2N2N1N3O001N101O0O2O001O0O2O000O2O000O2O00001N10001N10001N101O000O2O000O2O000O2O00001N10001N10O10000000001O00000000001O00000000001O0000000000001N10000000001O00000000001O0O1O1O100O2N1O1O100O1O2N1O100O1O1O101N1O1O1O100O2N1O1O100O1O2N1O100O1O1O2N100O1O1O1O10O01O000010O0001O00010O00001O01O01O0000O2M2M3M3N3L3N2M3N2M4N1001O0010O01O0100O100O010O100O100O010O3N1N2O2M2O1N3N1N26K<C1OO2M2O4L6J5K6J6IQbb2"}, "label": "this is a boy petting the head of a giraffe"}]}
{"id": 323108, "image": "COCO_train2014_000000323108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy petting a graiffe who ' s behind a wired fence\"."}], "task": "refering", "answer_template": "The \"a boy petting a graiffe who ' s behind a wired fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 36, 37, 54, 55, 72, 73, 90, 91, 98, 99, 108, 109, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 180, 181, 198, 199, 200, 216, 217, 218], "bbox": [0.0, 0.10786666666666668, 0.54776, 0.9842666666666666], "iscrowd": 0, "area": 23128.705250000003, "rle": {"size": [375, 500], "counts": "b9o1j5n3SLm3C=1O00001O001O0oI^Nm1c1RN`Nk1`1TNdNi1]1VNeNh1[1WNiNf1X1YNhNg1X1XNiNh1X1WNiNh1W1WNjNi1W1VNiNj1W1VNiNj1X1TNjNk1V1UNjNk1W1TNiNl1W1TNiNl1X1SNiNl1W1TNiNl1X1SNhNm1X1SNhNm1Y1SNgNl1Y1TNgNl1Y1TNgNl1Z1SNfNm1Z1SNfNm1[1RNfNm1[1RNeNn1]1PNcNP2_1nMaNR2a1lM`NS2b1kM^NU2d1iM\\NW2f1gMZNY2i1eMWNZ2k1dMUN\\2m1bMSN^2o1`MQN`2Q2bLkN^3[1\\KVMNe1e4Z1RKXM7_1f4_1gJ[M`0W1i4U2VKkMj4W2SKkMl4V2RKkMR3UN[NR4`NjMQ3[N\\Nm3aNhMo2cN\\Nf3cNiMl2hN^Na3cNhMk2oN^NZ3fNgMh2VO_NV5]1QK`NP5[1XKaNh4[1_KbNa4Z1fKcN[4[1jKaNV4^1nK_NS4_1QL^No3a1VL[Nk3c1YLZNg3e1^LWNc3g1bLUN_3h1Y3O2N1O2N1O2N1O2N1O2N2N1N3O001N101O0O2O001O0O2O000O2O000O2O00001N10001N10001N101O000O2O000O2O000O2O00001N10001N10O10000000001O00000000001O00000000001O0000000000001N10000000001O00000000001O0O1O1O100O2N1O1O100O1O2N1O100O1O1O101N1O1O1O100O2N1O1O100O1O2N1O100O1O1O2N100O1O1O1O10O01O000010O0001O00010O00001O01O01O0000O2M2M3M3N3L3N2M3N2M4N1001O0010O01O0100O100O010O100O100O010O3N1N2O2M2O1N3N1N26K<C1OO2M2O4L6J5K6J6IQbb2"}, "label": "a boy petting a graiffe who ' s behind a wired fence"}]}
{"id": 323108, "image": "COCO_train2014_000000323108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hair of the girl who is watching a giraffe\"."}], "task": "refering", "answer_template": "The \"hair of the girl who is watching a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 223, 224], "bbox": [0.10118, 0.6325066666666667, 0.51578, 1.0], "iscrowd": 0, "area": 22294.598200000004, "rle": {"size": [375, 500], "counts": "Sob08\\;9G9G9G9G9G6K5J6J6J6J6L3N3M3L4M3L2M2O1N2O1O1N2O1N2O1N2O1N2O1000000O1000000O1000000O1000001N1000000O1000000O1000000O1000000O1000000O1000000O1000000O100001O0000000000001O0000000000001O0000000000001O000001O0001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O000O1000000O2O1N2N3M2O1N3M2N2N3N1N3M2N2O2M2N3M2O1N3M2N3M3N2M3M3M2O2M3M2N2N2O2M2N2N2O1N2N2N2O1N2N1O2N101N2N1O2O0O2M3N1N3N5JTZh2"}, "label": "hair of the girl who is watching a giraffe"}]}
{"id": 323108, "image": "COCO_train2014_000000323108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of young girl ' s head with light brown hair\"."}], "task": "refering", "answer_template": "The \"back of young girl ' s head with light brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 223, 224], "bbox": [0.10118, 0.6325066666666667, 0.51578, 1.0], "iscrowd": 0, "area": 22294.598200000004, "rle": {"size": [375, 500], "counts": "Sob08\\;9G9G9G9G9G6K5J6J6J6J6L3N3M3L4M3L2M2O1N2O1O1N2O1N2O1N2O1N2O1000000O1000000O1000000O1000001N1000000O1000000O1000000O1000000O1000000O1000000O1000000O100001O0000000000001O0000000000001O0000000000001O000001O0001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O000O1000000O2O1N2N3M2O1N3M2N2N3N1N3M2N2O2M2N3M2O1N3M2N3M3N2M3M3M2O2M3M2N2N2O2M2N2N2O1N2N2N2O1N2N1O2N101N2N1O2O0O2M3N1N3N5JTZh2"}, "label": "back of young girl ' s head with light brown hair"}]}
{"id": 323108, "image": "COCO_train2014_000000323108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child ' s arm on the far right resting on the fence\"."}], "task": "refering", "answer_template": "The \"a child ' s arm on the far right resting on the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 159, 160, 161], "bbox": [0.59708, 0.4804, 0.99802, 0.696], "iscrowd": 0, "area": 5285.132099999998, "rle": {"size": [375, 500], "counts": "oe]31f;1N10000O010O100O010O100O014K2O000O100O2O0O1000000O2O000000001N100000001O000O100000O1000O010000O100O10000O100O10000O100001O001O001O00001O001O001O001O010O001O010O001O010O001O0010O01O00100O001O010O001O010O1O0010O01O0010O01O001O010O001O010O000010O01O010O010O00010O010O0010O010O0010O0010O010O01O01O010O010O01O010O010O00010O010O0010O01O01O01O010O010O0010O00010O01O010O0010O00010O01O010O0010O0010O0010Od0]OU4"}, "label": "a child ' s arm on the far right resting on the fence"}]}
{"id": 323108, "image": "COCO_train2014_000000323108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the outstretched arm of an onlooker waiting to pet a docile giraffe\"."}], "task": "refering", "answer_template": "The \"the outstretched arm of an onlooker waiting to pet a docile giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 159, 160, 161], "bbox": [0.59708, 0.4804, 0.99802, 0.696], "iscrowd": 0, "area": 5285.132099999998, "rle": {"size": [375, 500], "counts": "oe]31f;1N10000O010O100O010O100O014K2O000O100O2O0O1000000O2O000000001N100000001O000O100000O1000O010000O100O10000O100O10000O100001O001O001O00001O001O001O001O010O001O010O001O010O001O0010O01O00100O001O010O001O010O1O0010O01O0010O01O001O010O001O010O000010O01O010O010O00010O010O0010O010O0010O0010O010O01O01O010O010O01O010O010O00010O010O0010O01O01O01O010O010O0010O00010O01O010O0010O00010O01O010O0010O0010O0010Od0]OU4"}, "label": "the outstretched arm of an onlooker waiting to pet a docile giraffe"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a pink and black dress\"."}], "task": "refering", "answer_template": "The \"a woman wearing a pink and black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 153, 154, 155, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 340], "bbox": [0.684859375, 0.35049411764705884, 0.95671875, 0.9545176470588235], "iscrowd": 0, "area": 28625.096299999997, "rle": {"size": [425, 640], "counts": "h[f54_4MU47hKJW49gKHV4<gKEX4?eKBX4b0eK_OZ4d0dK\\O[4f0cK\\OZ4g0YKEf4<jJ5S5N[Jc0d5_OTJh0k5ZOPJk0m5XOnIl0Q6VOkIn0R6UOiIo0V6SOfIQ1W6ROdIR1[6PO`IT1_6nN]IV1`6mN[IW1d6kNXIY1e6iNWI[1h6gNSI]1l6eNPI_1m6dNnH`1Q7bNjHc1S7[2O2N1N2O2M2O2M2O1O2M2O2M2O2N1N2O2M2O2M2O2O01O100O100O1O100O100O100O]1cNO2N01O00000O1000000O2O00000O1000000O2O00000O10000O101O000O10N2B>B=D=00001N1000000O10001O0O1000000O101O0O1000000O101O000O3N4L4L4K5L5K4L4K5L4L4L4K5L4L5K4K4M001N2O1O001N2O001O1N2O001O1N101O1O0O2O1O1O0O2O1L3K6J6J6J5K6K5J7Ie0[Of0ZO\\Z;"}, "label": "a woman wearing a pink and black dress"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a pink dress with polka dots sits on a sofa\"."}], "task": "refering", "answer_template": "The \"a woman in a pink dress with polka dots sits on a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 153, 154, 155, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 340], "bbox": [0.684859375, 0.35049411764705884, 0.95671875, 0.9545176470588235], "iscrowd": 0, "area": 28625.096299999997, "rle": {"size": [425, 640], "counts": "h[f54_4MU47hKJW49gKHV4<gKEX4?eKBX4b0eK_OZ4d0dK\\O[4f0cK\\OZ4g0YKEf4<jJ5S5N[Jc0d5_OTJh0k5ZOPJk0m5XOnIl0Q6VOkIn0R6UOiIo0V6SOfIQ1W6ROdIR1[6PO`IT1_6nN]IV1`6mN[IW1d6kNXIY1e6iNWI[1h6gNSI]1l6eNPI_1m6dNnH`1Q7bNjHc1S7[2O2N1N2O2M2O2M2O1O2M2O2M2O2N1N2O2M2O2M2O2O01O100O100O1O100O100O100O]1cNO2N01O00000O1000000O2O00000O1000000O2O00000O10000O101O000O10N2B>B=D=00001N1000000O10001O0O1000000O101O0O1000000O101O000O3N4L4L4K5L5K4L4K5L4L4L4K5L4L5K4K4M001N2O1O001N2O001O1N2O001O1N101O1O0O2O1O1O0O2O1L3K6J6J6J5K6K5J7Ie0[Of0ZO\\Z;"}, "label": "a woman in a pink dress with polka dots sits on a sofa"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a laptop with long curly hair\"."}], "task": "refering", "answer_template": "The \"a woman holding a laptop with long curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 150, 151, 152, 173, 174, 175, 264, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.51125, 0.38578823529411765, 0.6795625, 1.0], "iscrowd": 0, "area": 12878.081349999997, "rle": {"size": [425, 640], "counts": "[QX43o<8N1O2N2N1O2N2\\LMoI5e58XJKd59YJKb57]JMkNPOn4T1VLo0hMQNn5R1ZLo0dMTNm5o0_Lo0`MVNm5m0cLo0[MZNn5h0gLi1U3XNkLj1R3WNnLl1n2UNRMm1l2SNTMo1i2RNWMn1h2SNXMm1g2TNYMl1f2UNZMk1f2UNZMk1e2VN[Mj1d2WN\\Mi1c2XN]Mh1c2XN]Mh1c2XN]Mh1c2XN]Mh1b2YN^Mg1b2YN^Mg1b2YN^Mg1a2ZN_Mf1a2ZN_Mf1a2ZN_Mf1`2[N`Me1`2[N`Me1`2[N`Me1_2\\NaMd1_2\\NaMd1_2\\NaMd1_2\\NaMd1^2]NbMc1^2]NbMc1^2]NbMc1]2^NcMb1]2^NcMb1]2^NcMa1]2`NcM_1^2aNbMiNYMl1W5[O`MiNfM^1k4I_MiNbNa0Q4e0]MjNE\\OP3j1[MjNV5U1jJkNV5U1jJkNV5U1jJkNW5T1iJlNW5T1iJlNW5T1iJlNX5S1hJmNX5S1hJmNX5S1hJmNY5R1gJnNY5R1gJoNY5P1gJQOZ5m0fJTOZ5k0fJUOSNGm5R1PLXOaM?X68WLT1b3kN^Ld1U3ZNkLP2l2oMTMQ2n2mMRMS2o2lMQMT2Q3jMoLV2S3hMmLX2V3eMjL[2X3cMhL]2[3_MfLa2\\3ZMgLf2\\3TMgLk2^3nLeLR3c3cL`L]3f600000000000000000000000000M3H8H8H9G?A=C>UOmhd2"}, "label": "a woman holding a laptop with long curly hair"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing jeans holding a laptop sitting inbetween two other women\"."}], "task": "refering", "answer_template": "The \"a woman wearing jeans holding a laptop sitting inbetween two other women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 150, 151, 152, 173, 174, 175, 264, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.51125, 0.38578823529411765, 0.6795625, 1.0], "iscrowd": 0, "area": 12878.081349999997, "rle": {"size": [425, 640], "counts": "[QX43o<8N1O2N2N1O2N2\\LMoI5e58XJKd59YJKb57]JMkNPOn4T1VLo0hMQNn5R1ZLo0dMTNm5o0_Lo0`MVNm5m0cLo0[MZNn5h0gLi1U3XNkLj1R3WNnLl1n2UNRMm1l2SNTMo1i2RNWMn1h2SNXMm1g2TNYMl1f2UNZMk1f2UNZMk1e2VN[Mj1d2WN\\Mi1c2XN]Mh1c2XN]Mh1c2XN]Mh1c2XN]Mh1b2YN^Mg1b2YN^Mg1b2YN^Mg1a2ZN_Mf1a2ZN_Mf1a2ZN_Mf1`2[N`Me1`2[N`Me1`2[N`Me1_2\\NaMd1_2\\NaMd1_2\\NaMd1_2\\NaMd1^2]NbMc1^2]NbMc1^2]NbMc1]2^NcMb1]2^NcMb1]2^NcMa1]2`NcM_1^2aNbMiNYMl1W5[O`MiNfM^1k4I_MiNbNa0Q4e0]MjNE\\OP3j1[MjNV5U1jJkNV5U1jJkNV5U1jJkNW5T1iJlNW5T1iJlNW5T1iJlNX5S1hJmNX5S1hJmNX5S1hJmNY5R1gJnNY5R1gJoNY5P1gJQOZ5m0fJTOZ5k0fJUOSNGm5R1PLXOaM?X68WLT1b3kN^Ld1U3ZNkLP2l2oMTMQ2n2mMRMS2o2lMQMT2Q3jMoLV2S3hMmLX2V3eMjL[2X3cMhL]2[3_MfLa2\\3ZMgLf2\\3TMgLk2^3nLeLR3c3cL`L]3f600000000000000000000000000M3H8H8H9G?A=C>UOmhd2"}, "label": "a woman wearing jeans holding a laptop sitting inbetween two other women"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a red bath robe with a mask on her face and towel on her head filing her nails\"."}], "task": "refering", "answer_template": "The \"a woman in a red bath robe with a mask on her face and towel on her head filing her nails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 126, 147, 148, 149, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 284, 285, 286, 287, 288, 307, 308, 309, 310, 311, 330, 331, 332, 333, 334], "bbox": [0.347921875, 0.31868235294117647, 0.5367656249999999, 1.0], "iscrowd": 0, "area": 25601.9042, "rle": {"size": [425, 640], "counts": "nil2c0e<3N3aGZOT4i0fK]OW4g0cK]O\\4e0`K]O_4f0\\K]Oc4d0[K^Od4c0TIoNk0`0d5_OdHS1h0VOi0;j5m0ZIlNi09l5m0YImN1n0d68XImN0n0g67WImNOn0i66WIoNMl0k67VIoNLl0m67TIQOLi0P77QISOMg0Q78nHVOMd0U76lHZOLa0X7n3O100N2M3N2O1N2GiI[I\\6^6;J6I7POP1N1O2N2N2N2N2N2O100O1O100O1O1O100O1O100O1O1O10000O1000000O1000000000000O1001O1O001O001O1O1O001O1O2N1O1O2N1O1O2\\IYKb4h4\\K[Kb4f4\\K_K`4b4oJWKSO;k5`4nJXKTO>i5]4YJXKN4I7V6^4mIeLV6S54J6M3O10000001O1O1O1O0oLcIVO]6g0iIVOW6h0mIVOU6\\N]IW1a0;X6VN[I\\1?=V6jMiIh12=V6fMnIj1N?g8oMZGQ2n8gMSGX2o8eM[GR2f8mMkGb1U8]NSH\\1m7dNYHV1R:YOjC1Z<GQD2Q<Gh_j3"}, "label": "a woman in a red bath robe with a mask on her face and towel on her head filing her nails"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a red robe and has a towel on her head\"."}], "task": "refering", "answer_template": "The \"a woman wearing a red robe and has a towel on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 126, 147, 148, 149, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 284, 285, 286, 287, 288, 307, 308, 309, 310, 311, 330, 331, 332, 333, 334], "bbox": [0.347921875, 0.31868235294117647, 0.5367656249999999, 1.0], "iscrowd": 0, "area": 25601.9042, "rle": {"size": [425, 640], "counts": "nil2c0e<3N3aGZOT4i0fK]OW4g0cK]O\\4e0`K]O_4f0\\K]Oc4d0[K^Od4c0TIoNk0`0d5_OdHS1h0VOi0;j5m0ZIlNi09l5m0YImN1n0d68XImN0n0g67WImNOn0i66WIoNMl0k67VIoNLl0m67TIQOLi0P77QISOMg0Q78nHVOMd0U76lHZOLa0X7n3O100N2M3N2O1N2GiI[I\\6^6;J6I7POP1N1O2N2N2N2N2N2O100O1O100O1O1O100O1O100O1O1O10000O1000000O1000000000000O1001O1O001O001O1O1O001O1O2N1O1O2N1O1O2\\IYKb4h4\\K[Kb4f4\\K_K`4b4oJWKSO;k5`4nJXKTO>i5]4YJXKN4I7V6^4mIeLV6S54J6M3O10000001O1O1O1O0oLcIVO]6g0iIVOW6h0mIVOU6\\N]IW1a0;X6VN[I\\1?=V6jMiIh12=V6fMnIj1N?g8oMZGQ2n8gMSGX2o8eM[GR2f8mMkGb1U8]NSH\\1m7dNYHV1R:YOjC1Z<GQD2Q<Gh_j3"}, "label": "a woman wearing a red robe and has a towel on her head"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding a bottle with one foot on the couch and wearing black\"."}], "task": "refering", "answer_template": "The \"the woman holding a bottle with one foot on the couch and wearing black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 303, 304, 305, 327], "bbox": [0.12251562499999999, 0.4196470588235294, 0.39, 1.0], "iscrowd": 0, "area": 21120.555800000002, "rle": {"size": [425, 640], "counts": "aQQ17P=6J6J;F3M3M3_OlNgDX1V;kNgDW1W;mNeDV1X;`0M3M4M2M3N3M2M3N3M2RFYMV9i2gFZMW9h2fF[MX9h2cFYM^9i2_FWMb9k2[FVMe9V32N2N21O2N2M3N1OeL\\FP3d9mL^FT3a9kLaFU3^9kLcFU3\\9kLeFU3Z9kLgFU3Y9jLhFV3W9jLjFV3U9iLlFY3R9dLPG^3]91N2N1O1O2N1O1O10O01O001O001O0O2O1YGgL]7R5@1O1O1O2O00000O10000O1000000O1000000000000000N3L3M3M]OeJ^IX5c6PKWIm4k6TKUIi4m6YKRId4P7]KPI`4Q7cKnHZ4k6RKUIe01U4i6^KQI?5Q4i6^LWI_3k6]13M2N2N3M2N2N2N3M3M3M4M2O100O01O001O0010O100O101N10000O2O0O4M3L3N2M2O1N3N1N2O1N2O1fLTGT2m8jMVGU2k8hMWGW2k8gMWGX2j8eMXGZ2i8dMYG\\2i8`MZGg0B<V9kNYG>Ne0k8iNZG95l0b8hN[G<5i0c8gN[G>6h0`8gN\\Ga06e0\\9ZOgFd0Z9[OhFb0Y9]OiFb0X9VOoFi0R9POUGP1b:O1N2O1N2O1N2O1N2O1N2N2M3N2M3N2M3N^bQ5"}, "label": "the woman holding a bottle with one foot on the couch and wearing black"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in headphones\"."}], "task": "refering", "answer_template": "The \"girl in headphones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 303, 304, 305, 327], "bbox": [0.12251562499999999, 0.4196470588235294, 0.39, 1.0], "iscrowd": 0, "area": 21120.555800000002, "rle": {"size": [425, 640], "counts": "aQQ17P=6J6J;F3M3M3_OlNgDX1V;kNgDW1W;mNeDV1X;`0M3M4M2M3N3M2M3N3M2RFYMV9i2gFZMW9h2fF[MX9h2cFYM^9i2_FWMb9k2[FVMe9V32N2N21O2N2M3N1OeL\\FP3d9mL^FT3a9kLaFU3^9kLcFU3\\9kLeFU3Z9kLgFU3Y9jLhFV3W9jLjFV3U9iLlFY3R9dLPG^3]91N2N1O1O2N1O1O10O01O001O001O0O2O1YGgL]7R5@1O1O1O2O00000O10000O1000000O1000000000000000N3L3M3M]OeJ^IX5c6PKWIm4k6TKUIi4m6YKRId4P7]KPI`4Q7cKnHZ4k6RKUIe01U4i6^KQI?5Q4i6^LWI_3k6]13M2N2N3M2N2N2N3M3M3M4M2O100O01O001O0010O100O101N10000O2O0O4M3L3N2M2O1N3N1N2O1N2O1fLTGT2m8jMVGU2k8hMWGW2k8gMWGX2j8eMXGZ2i8dMYG\\2i8`MZGg0B<V9kNYG>Ne0k8iNZG95l0b8hN[G<5i0c8gN[G>6h0`8gN\\Ga06e0\\9ZOgFd0Z9[OhFb0Y9]OiFb0X9VOoFi0R9POUGP1b:O1N2O1N2O1N2O1N2O1N2N2M3N2M3N2M3N^bQ5"}, "label": "girl in headphones"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"arm of a sofa\"."}], "task": "refering", "answer_template": "The \"arm of a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 253, 254, 255, 256, 277, 278, 279, 280, 300, 301, 302, 303, 323, 324, 325, 326], "bbox": [0.018484375, 0.6640235294117647, 0.21329687500000002, 0.9895058823529411], "iscrowd": 0, "area": 9568.449900000001, "rle": {"size": [425, 640], "counts": "TY53U=5K5L5J4L2O0O1O2O0O100O2N100O2N100O1O2O0O1O2O0O100O2O0O1O106I9H9F9H8G9[FTM]8U3VGoLi8Z3lFiLT9g3N1O100O10000O101O00000O100000001N100000000O101O00O101O00000000000Oj0WO8H5K4L1O2N1O1O0O2O1O1O1O001O1O1N101O1O1O1O002N7H3N2N3M1O2N1O2N1O1N2O1O1O001O1O1O1N101O001O001O001O7H<E;Ej[`6"}, "label": "arm of a sofa"}]}
{"id": 544294, "image": "COCO_train2014_000000544294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"end of couch on left side\"."}], "task": "refering", "answer_template": "The \"end of couch on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 253, 254, 255, 256, 277, 278, 279, 280, 300, 301, 302, 303, 323, 324, 325, 326], "bbox": [0.018484375, 0.6640235294117647, 0.21329687500000002, 0.9895058823529411], "iscrowd": 0, "area": 9568.449900000001, "rle": {"size": [425, 640], "counts": "TY53U=5K5L5J4L2O0O1O2O0O100O2N100O2N100O1O2O0O1O2O0O100O2O0O1O106I9H9F9H8G9[FTM]8U3VGoLi8Z3lFiLT9g3N1O100O10000O101O00000O100000001N100000000O101O00O101O00000000000Oj0WO8H5K4L1O2N1O1O0O2O1O1O1O001O1O1N101O1O1O1O002N7H3N2N3M1O2N1O2N1O1N2O1O1O001O1O1O1N101O001O001O001O7H<E;Ej[`6"}, "label": "end of couch on left side"}]}
{"id": 265766, "image": "COCO_train2014_000000265766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cow grazing in the grass\"."}], "task": "refering", "answer_template": "The \"a black cow grazing in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 264, 268, 270], "bbox": [0.45225, 0.32102083333333337, 0.79184375, 0.6729583333333333], "iscrowd": 0, "area": 22277.674399999996, "rle": {"size": [480, 640], "counts": "\\dW43l>n1RN4K5L5K4L0O10O10O0100NWDbMU;[2mDiMo:X2PElMn:S2QEQNm:o1QEUNm:Q2kDSNS;P3M8H<D7I2N1O2N1O2N1O2O0O2N00VOUF_Lk9_3ZF^Lg9^3_F`L`9W3lFfLT9]3lF`LT9c3lFZLU9h3kFVLT9m3lFPLT9S4k03M3N2M5K6J2N01O000000001O000dNPFhMP:Q2[FkMe9S2bFhM^9V2gFgMZ9W2kFeMU9Y2QGcMo8[2VGbMj8\\2\\GaMc8[2eGaM[8_2hG_MW8b2S201O0001O00001O00010O00001O0000001O01O00000001O00000]EYMn8f2PG\\MP9e2mF^MR9b2lF`MT9`2jFbMV9^2hFeMW9[2gFgMY9Z2dFiM[9W2cFkM^9T2aFmM_9R2`FQN_9n1`FTN`9l1^FWNa9h1^FZNb9e1WFdNh9\\1UFdNn9[1QFdNS:^1gEbN\\:a1_E^Nd:l24L3N2N2N3M2M301O000015J9G6K5J4L0001O00001O001O010O001O0N2O2lNoEVMR:i2RFSMP:l2SFPMn9Q3Q1O0010O010O100O01N101O001O000O2O001O0012N8I001O001O001OO1O1O1O1O1N2N2O1O001O1O1O1O1O1O1N2O1O001N2N2M3N2N2M3N2N1O2M3N1O000O2O0N2O1N3oNiB6X=FQC4P=IXCNm<OZCIi<6Q1O2M2O1OgRn1"}, "label": "a black cow grazing in the grass"}]}
{"id": 265766, "image": "COCO_train2014_000000265766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow on the left\"."}], "task": "refering", "answer_template": "The \"the cow on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 279, 280, 281], "bbox": [0.14264062500000002, 0.3544791666666667, 0.34140625, 0.7380625], "iscrowd": 0, "area": 14593.714500000004, "rle": {"size": [480, 640], "counts": "ioZ12S=3aD0];4`CM73V<4]C051[<1^C4O0b<L]CQ1d<oNXCU1h<jNWCW1j<?010O0000001O00000000000O10N2N2O0O2N2ROTNfDn1R;\\NiDf1P;eNkD]1R;Y1O1O001O002M2O1O1O10O1O00\\ERLY:n3`EZL_:e3_E`L^:R4O1O1O001O1O1O1O00N2L4I8MhElKj9R4UFQLj9m3VFVLj9f3XF\\Lh9`3[FaLd9\\3^FfLb9[3ZFgLg9]3QFgLn9]3kEgLU:R4O010O10O01O010O010O010O00O0O2I7F:N2N2N2O1N3M2O3N100O2N100O2OoE]Lf8a3YGcLf8\\3XGgLh8X3VGjLl8T3SGmLQ9o2mFSMT9l2jFVMV9k2fFXM[9g2ZFdMf9k3001O001O1O1O1O1@jETLW:d3h0H7K6L4K4L5J6J7G9F`0@>B;Cf0[OfSU6"}, "label": "the cow on the left"}]}
{"id": 265766, "image": "COCO_train2014_000000265766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the cow on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 279, 280, 281], "bbox": [0.14264062500000002, 0.3544791666666667, 0.34140625, 0.7380625], "iscrowd": 0, "area": 14593.714500000004, "rle": {"size": [480, 640], "counts": "ioZ12S=3aD0];4`CM73V<4]C051[<1^C4O0b<L]CQ1d<oNXCU1h<jNWCW1j<?010O0000001O00000000000O10N2N2O0O2N2ROTNfDn1R;\\NiDf1P;eNkD]1R;Y1O1O001O002M2O1O1O10O1O00\\ERLY:n3`EZL_:e3_E`L^:R4O1O1O001O1O1O1O00N2L4I8MhElKj9R4UFQLj9m3VFVLj9f3XF\\Lh9`3[FaLd9\\3^FfLb9[3ZFgLg9]3QFgLn9]3kEgLU:R4O010O10O01O010O010O010O00O0O2I7F:N2N2N2O1N3M2O3N100O2N100O2OoE]Lf8a3YGcLf8\\3XGgLh8X3VGjLl8T3SGmLQ9o2mFSMT9l2jFVMV9k2fFXM[9g2ZFdMf9k3001O001O1O1O1O1@jETLW:d3h0H7K6L4K4L5J6J7G9F`0@>B;Cf0[OfSU6"}, "label": "the cow on the left in the right hand picture"}]}
{"id": 257576, "image": "COCO_train2014_000000257576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray cat has its head in the sink\"."}], "task": "refering", "answer_template": "The \"a gray cat has its head in the sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 276, 277, 299, 300, 322, 323, 345, 346], "bbox": [0.006734374999999999, 0.3550625, 0.46854687500000003, 0.9370833333333334], "iscrowd": 0, "area": 33715.54064999998, "rle": {"size": [480, 640], "counts": "jX21k>6J5K6I6K6`KXO`Jn0\\5YO[Jl0a5\\OUJk0f5]OQJh0l5_OkIg0P6AhId0S6DdIa0Y6F^I`0]6HZI=b6MTI8h60nH4o66fHOX7;[HJc7`0PHDo7h3N1O2N2M2O2N1O2N2N1O2N1O2NXNkHgKT7X4TIcKj6[4^IaK`6^4eI_KZ6`4jI_KT6_4PJ`Ko5_4UJ_Kj5_4ZJ`Ke5_4_J_Ka5^4cJbK[5\\4jJbKV5\\4mJcKR5[4SKcKm4Z4WKeKh4Y4]KeKc4X4bKgK\\4V4jKhKV4U4oKgKR4V4TLdKo3Z4VL`Kn3]4[3N3M2M4[Od0010O010O0010O10O0100000O1000000O10001N1000000O10000O2O00000O1000000O2O000O0100O10O10O10O010001N10000O1000000O10000O1001O000001O01O0000001O000000001O0000000010O0000010N101O0O2O0O101O0O2O001N11O01O0010O0001O010O001O01O01O01O01O00010O00010O00010O00010O00010O00001O00001O010O00001O00001O00001O00010O00001O001O00001O00001O001O00001O010O00001O00001O001O00001O00001O001O00001O00001O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1N3N1O1O1O2N1O1O2N1O1O1O2M2O1O1O2N1O1N3N1O1O2M2M4M3L3M4M3L3M4M3L3M4L4MYSo4"}, "label": "a gray cat has its head in the sink"}]}
{"id": 257576, "image": "COCO_train2014_000000257576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat with its head in a sink\"."}], "task": "refering", "answer_template": "The \"a black cat with its head in a sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 276, 277, 299, 300, 322, 323, 345, 346], "bbox": [0.006734374999999999, 0.3550625, 0.46854687500000003, 0.9370833333333334], "iscrowd": 0, "area": 33715.54064999998, "rle": {"size": [480, 640], "counts": "jX21k>6J5K6I6K6`KXO`Jn0\\5YO[Jl0a5\\OUJk0f5]OQJh0l5_OkIg0P6AhId0S6DdIa0Y6F^I`0]6HZI=b6MTI8h60nH4o66fHOX7;[HJc7`0PHDo7h3N1O2N2M2O2N1O2N2N1O2N1O2NXNkHgKT7X4TIcKj6[4^IaK`6^4eI_KZ6`4jI_KT6_4PJ`Ko5_4UJ_Kj5_4ZJ`Ke5_4_J_Ka5^4cJbK[5\\4jJbKV5\\4mJcKR5[4SKcKm4Z4WKeKh4Y4]KeKc4X4bKgK\\4V4jKhKV4U4oKgKR4V4TLdKo3Z4VL`Kn3]4[3N3M2M4[Od0010O010O0010O10O0100000O1000000O10001N1000000O10000O2O00000O1000000O2O000O0100O10O10O10O010001N10000O1000000O10000O1001O000001O01O0000001O000000001O0000000010O0000010N101O0O2O0O101O0O2O001N11O01O0010O0001O010O001O01O01O01O01O00010O00010O00010O00010O00010O00001O00001O010O00001O00001O00001O00010O00001O001O00001O00001O001O00001O010O00001O00001O001O00001O00001O001O00001O00001O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1N3N1O1O1O2N1O1O2N1O1O1O2M2O1O1O2N1O1N3N1O1O2M2M4M3L3M4M3L3M4M3L3M4L4MYSo4"}, "label": "a black cat with its head in a sink"}]}
{"id": 257576, "image": "COCO_train2014_000000257576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange cat\"."}], "task": "refering", "answer_template": "The \"the orange cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 247, 248], "bbox": [0.46814062500000003, 0.33639583333333334, 0.8147968750000001, 0.630125], "iscrowd": 0, "area": 22909.569700000007, "rle": {"size": [480, 640], "counts": "Yl\\48g>8H2N1O1O1O1M3K6J5K5K7I6J6K5J7J5M4K6K5K4L3M3M2N2N2O1N2O1O1O1O1O1O001O001O1O001O001O1O0O2O01000O01N2M3N2N110000000000O100000000O10000000000O100000000000001N10001O0000001O000O2O0000001O0000001N10000000O100O100O010O100O100O100O010O10000O100O1000O0101N100O10001N10000000001O000O1000001O000000001O00000000001O000000001O00000001O01O00000001O01O000001O01O01O11N1O1O1O100O1O1O1O100O1O1O1O2N100O1O1O100O1O1O1O100O1O1O101N1N2N2M3M3N2M3N2K5G9G9G9G9H9F]bg1"}, "label": "the orange cat"}]}
{"id": 257576, "image": "COCO_train2014_000000257576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown cat\"."}], "task": "refering", "answer_template": "The \"brown cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 247, 248], "bbox": [0.46814062500000003, 0.33639583333333334, 0.8147968750000001, 0.630125], "iscrowd": 0, "area": 22909.569700000007, "rle": {"size": [480, 640], "counts": "Yl\\48g>8H2N1O1O1O1M3K6J5K5K7I6J6K5J7J5M4K6K5K4L3M3M2N2N2O1N2O1O1O1O1O1O001O001O1O001O001O1O0O2O01000O01N2M3N2N110000000000O100000000O10000000000O100000000000001N10001O0000001O000O2O0000001O0000001N10000000O100O100O010O100O100O100O010O10000O100O1000O0101N100O10001N10000000001O000O1000001O000000001O00000000001O000000001O00000001O01O00000001O01O000001O01O01O11N1O1O1O100O1O1O1O100O1O1O1O2N100O1O1O100O1O1O1O100O1O1O101N1N2N2M3M3N2M3N2K5G9G9G9G9H9F]bg1"}, "label": "brown cat"}]}
{"id": 44600, "image": "COCO_train2014_000000044600.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skiis on a black person\"."}], "task": "refering", "answer_template": "The \"skiis on a black person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 314, 315, 316, 317, 318, 329, 330, 331, 333, 334, 344, 345, 346, 348, 349, 350, 361, 363, 364, 365, 379, 380, 381], "bbox": [0.24379166666666666, 0.7714375, 0.7795416666666667, 0.985890625], "iscrowd": 0, "area": 5175.15575000002, "rle": {"size": [640, 480], "counts": "TfY21oc00O2O0O2O0O2O0O2O0O2O001N101N101N10O010O010O010O010O10O010O010O01O010O010O010OT^OBX`0>h_OBW`0>i_OCW`0=h_ODW`0<j_ODV`0<i_OEV`0;k_OEU`0;j_OFU`0:l_OFT`0:k_OGT`09l_OHT`07m_OIR`08m_OIS`06n_OJQ`07n_OKQ`04o_OMQ`03o_OMP`03P@NP`02P@No?2Q@Oo?1P@0o?0R@0n?0Q@0o?0R@0n?0Q@0o?0Q@1o?OQ@0o?0Q@0P`0OQ@1n?0Q@0P`0OQ@0P`00o_O1P`0OP@1Q`0Oo_O1P`0OP@1Q`0Oo_O0Q`00o_O1Q`0On_O1R`0Oo_O0R`00m_O1R`0Oo_O0R`00m_O0S`00m_O1S`0Om_O0S`00m_O1S`0Om_O0S`00m_O0T`0Ol_O2S`0Om_O0T`0Ol_O1T`00l_O0S`00m_O0T`00l_O0S`00m_O0T`00k_O0U`00l_O0S`01l_OOU`00l_OOT`02k_OOU`00k_O0U`01k_ONU`02k_OOU`01k_ONU`02k_OOU`00k_O0U`01k_ONU`02k_OOU`01k_ONU`02k_ONV`02i_OOV`01k_ONU`03j_ONV`01k_ONU`03j_OMW`02i_OOV`02j_OMV`03j_OMW`03i_OMV`03j_OMW`03i_OLW`04i_OMV`03j_OMW`03i_OMV`03j_OMW`03i_OLW`04i_OMV`04i_OLX`03i_OLW`05h_OLX`03i_OLW`05h_OKX`05h_OLX`04h_OKX`05h_OLX`04h_OKX`05h_OKX`06g_OKY`04h_OKX`06g_OJZ`05f_OLY`04h_OKX`06g_OKY`04g_OLY`05f_OK[`04f_OLY`05f_OKZ`05f_OK[`05e_OKZ`05f_OK[`05d_OK\\`05e_OKZ`06e_OJ\\`05d_OL[`05d_OK]`04d_OK\\`06c_OK\\`05d_OK]`04d_OK\\`06c_OK]`04c_OL]`05c_OK\\`05d_OK]`05b_OK^`05b_OL^`04b_OKa`02__ONd`00[_O1g`0LZ_O3i`0KV_O5m`0HS_O9o`0EQ_O:Zb01O00010O00010O0010O0001O01O01O01O01O010O000010O00010O00010O01O00010O00010O0010O0001O01O01O01O01O010O000010O00010O01O01O01O00010O0002O0O100O1O2O0O1O100O2O0O1O100O2N100O1O10V\\R2"}, "label": "skiis on a black person"}]}
{"id": 323129, "image": "COCO_train2014_000000323129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the area of the table directly in front of the man ' s hand that is holding a spoon\"."}], "task": "refering", "answer_template": "The \"the area of the table directly in front of the man ' s hand that is holding a spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.017979166666666668, 0.9026875000000001, 0.6591666666666666, 1.0], "iscrowd": 0, "area": 14005.851350000004, "rle": {"size": [640, 480], "counts": "Rf81oc0000c]O6d`0JV_O<Pb07J6QOWO`^OP1`a0POY^OW1ga0iNS^O]1ma0cNl]Od1Tb05O0000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O0000000000000001O000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000001O0001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000010O00000000000000000000001O000000000000000000000000000000000000000000000000000000VmU3"}, "label": "the area of the table directly in front of the man ' s hand that is holding a spoon"}]}
{"id": 323129, "image": "COCO_train2014_000000323129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown table with multiple food crumbs on the surface\"."}], "task": "refering", "answer_template": "The \"a brown table with multiple food crumbs on the surface\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.017979166666666668, 0.9026875000000001, 0.6591666666666666, 1.0], "iscrowd": 0, "area": 14005.851350000004, "rle": {"size": [640, 480], "counts": "Rf81oc0000c]O6d`0JV_O<Pb07J6QOWO`^OP1`a0POY^OW1ga0iNS^O]1ma0cNl]Od1Tb05O0000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O0000000000000001O000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000001O0001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000010O00000000000000000000001O000000000000000000000000000000000000000000000000000000VmU3"}, "label": "a brown table with multiple food crumbs on the surface"}]}
{"id": 462398, "image": "COCO_train2014_000000462398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front headlights of both motorcycles\"."}], "task": "refering", "answer_template": "The \"the front headlights of both motorcycles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 238, 254, 255, 256, 261, 262, 277, 278], "bbox": [0.041890625, 0.36234192037470725, 0.39792187500000004, 0.8309367681498829], "iscrowd": 0, "area": 14457.97135, "rle": {"size": [427, 640], "counts": "Sb;4R=6K6J6J5]Od0M3N1N3N2M2O2M3N1N3N2M1O2O1N101N2O1N101N2O0O2O000O101O00001O0O101O00001O0O10000000N2N1O2N2N2N2N2N1O2N2N2N2N2N1XOhD]OZ;c0iDXOY;g0kDTOW;l0kDPOW;o0mDlNU;T1`010O10J5H9]Oc0N1O2N1O2N2N1O2N1O2N2N1O2N1O2O1O0O2O0O2O1O0O2O0O2O1O0O2O0O2O1O0O2O0O2O0O2O1O0O2O0O2O1O1N2O100001kLcFY2^9fMcFZ2]9eMdF[2]9cMdF50Z1\\9_NfF51[1Y9_NgF52[1X9^NgF55[1T9_NgF58[1R9]NhF6:[1n8^NiF5<\\1k8^NjF5<]1k8\\NjF5=_1i8[NkF5=_1i8ZNlF6;a1j8WNlF8:a1j8VNlF::`1j8UNmF::a1\\9_NdFa1\\9_NdFa1\\9^NeFb1[9^NeFb1[9^NeFb1[9]NfFc1Z9]NfFc1Y9]NhFb1Y9^NgFb1Y9^NgFb1Y9]NhFc1X9]NgFd1Y9\\NgFd1Y9[NhFe1X9[NhFe1W9\\NiFd1W9[NjFe1V9[NjFe1V9[NjFe1V9ZNkFe1V9[NjFe1V9ZNkFf1U9ZNkFf1V9YNjFf1^:O001O1O1O1O1O1N2O5K6J5K5K5K5K6J5K5JQg;j0TXD6K2M4M2M3M3N2M3N3L3006J6J6J6K5J6J5K6J6K5JPWP5"}, "label": "the front headlights of both motorcycles"}]}
{"id": 462398, "image": "COCO_train2014_000000462398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle second in line\"."}], "task": "refering", "answer_template": "The \"the motorcycle second in line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 238, 254, 255, 256, 261, 262, 277, 278], "bbox": [0.041890625, 0.36234192037470725, 0.39792187500000004, 0.8309367681498829], "iscrowd": 0, "area": 14457.97135, "rle": {"size": [427, 640], "counts": "Sb;4R=6K6J6J5]Od0M3N1N3N2M2O2M3N1N3N2M1O2O1N101N2O1N101N2O0O2O000O101O00001O0O101O00001O0O10000000N2N1O2N2N2N2N2N1O2N2N2N2N2N1XOhD]OZ;c0iDXOY;g0kDTOW;l0kDPOW;o0mDlNU;T1`010O10J5H9]Oc0N1O2N1O2N2N1O2N1O2N2N1O2N1O2O1O0O2O0O2O1O0O2O0O2O1O0O2O0O2O1O0O2O0O2O0O2O1O0O2O0O2O1O1N2O100001kLcFY2^9fMcFZ2]9eMdF[2]9cMdF50Z1\\9_NfF51[1Y9_NgF52[1X9^NgF55[1T9_NgF58[1R9]NhF6:[1n8^NiF5<\\1k8^NjF5<]1k8\\NjF5=_1i8[NkF5=_1i8ZNlF6;a1j8WNlF8:a1j8VNlF::`1j8UNmF::a1\\9_NdFa1\\9_NdFa1\\9^NeFb1[9^NeFb1[9^NeFb1[9]NfFc1Z9]NfFc1Y9]NhFb1Y9^NgFb1Y9^NgFb1Y9]NhFc1X9]NgFd1Y9\\NgFd1Y9[NhFe1X9[NhFe1W9\\NiFd1W9[NjFe1V9[NjFe1V9[NjFe1V9ZNkFe1V9[NjFe1V9ZNkFf1U9ZNkFf1V9YNjFf1^:O001O1O1O1O1O1N2O5K6J5K5K5K5K6J5K5JQg;j0TXD6K2M4M2M3M3N2M3N3L3006J6J6J6K5J6J5K6J6K5JPWP5"}, "label": "the motorcycle second in line"}]}
{"id": 462398, "image": "COCO_train2014_000000462398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle the nearest woman is on\"."}], "task": "refering", "answer_template": "The \"the motorcycle the nearest woman is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 127, 128, 133, 134, 145, 146, 147, 148, 149, 150, 151, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 301, 302, 303, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 335], "bbox": [0.11871875000000001, 0.26014051522248244, 0.9451875000000001, 0.9898594847775175], "iscrowd": 0, "area": 75662.64475000002, "rle": {"size": [427, 640], "counts": "QQP19P=`0_O;F3M4K4M4K4M3M4L3N3M2N2N3^OkMiEW2S:nMjEU2S:nMiEU2V:b0M2N2N3N0O2N1O2O0O2N1O2O0O2N1O2O0O2N2N101N1O2O001N101O001O0O2OO1000O010000O01000O1000O1000000O10O100000O1000000O1000000O0101O000000000000001O000O1000000000001O01O001O00001O01O01O00001OOO101N100O2O0O101N100O1^NbKXJ_4c5PLQJQ4j5`LjI`3T6kLbIW3]6PM\\IP3d6k101O000010N2O1O1O1N2O1O2N1N2O00O1O00O1O001O1O1O1N21N3N2NbNTIaLj6Y3aIdL]6W3mIfLP6V3[JfLc5T3hJiLV5R3TKkLj4U3[KhLc4Y3aKdL\\4^3iK^LU4b3PL[Ln3f3VLVLi3k3[LQLd3P4i22O2N8H8H8G9H8H8H8H8I7I7H8IQMfIUOU6h0lI_Oo5>SJEj5:VJJg54[JNd5[OWJcM6V3`5WO[J`M7b3V5mNdJ_M7i3P5gNlJ\\M5S4j4aNRKYM6[4e4ZNVKYM5c4h4lMTK^M6f4Q5`M[Ka2S5PMlJQ3_5dLaJ]3^5cLbJ]3_5aLaJa3^5_LbJa3_5^L`Jd3_5\\LaJd3`5ZLaJf3_5ZL`Jg3`5YL`Jh3d4PK^KW1Ni3d4RK\\KU1Oj3e4SK[KQ11l3d4VKVKP16j3c4YKTKn08k3c4YKQKn0<i3c4[KmJ>J@f0h4c4\\KiJ<0@d0h4b4\\LlJlNb0h4a4UKaK2^O1?j4_4PKmKOWO6>k4]4lJcL9POk4X6UKgIl4Y6TKgIm4X6SKhIm4X6RKhIo4X6QKhIo4X6QKhIo4X6QKhIP5W6oJjIQ5V6oJjIQ5V6oJjIQ5V6oJjIQ5V6oJjIR5U6nJkIR5U6oJjIQ5V6oJjIQ5V6PKjIo4V6QKjIP5U6QKjIo4V6QKjIo4V6RKiIn4W6RKiIn4W6SKhIm4X6SKhIm4X6TKgIl4Y6TKgIl4Y6UKfIk4Z6UKgIj4Z6VKeIj4[6VKeIk4Z6VKeIj4\\6UKdIk4\\6VKcIj4]6VKcIj4^6VKaIj4_6VKaIj4_6WK`Ii4a6VK_Ij4_6YK`Ig4]6\\KdIc4X6bKgI^4V6eKjI\\4R6hKmIX4P6kKPJU4o5mKPJS4Q6lKoIT4Q6PKiI;6e4Q6kJnI`01e4Q6jJoIa00e4Q6kJnIa01c4Q6lJmIb03a4P6mJmIb03a4P6mJmIb04`4o5oJlIa05a4n5nJmIa06_4n5PKlIb06\\4o5RKkIb06\\4m5UKmI>7[4e5_KTJ76Y4c5dKXJ26X4c5gKVJ27U4d5jKUJHIG=g4f5jKTJFKH<f4f5mKRJDNI9f4g5nKRJBOI9f4f5oKRJBOJ8e4g5oKSJBNI9e4g5oKRJCOI8c4h5QLQJCOI8c4h5QLRJBOJ7b4h5RLRJB0I6c4h5RLSJA0J5a4j5TLPJA1J6`4i5ULPJB1N0Z5P5VKPKA1N0Y5P5XKoJA2L0[5o4XKoJA2L1Z5o4XKoJA2J3[5m4ZKnJA3H3]5l4ZKnJAg0b1TO`1W5]MoJAf0]1ZOc1Q5_MoJAf0Z1@b1m4bMmJBg0V1Ee0IXNn4;nJAg0R1Kg0CZNm4<nJAg0o01f0^O]Nm4=mJAh0k06g0YO_Nl4?mJ@h0g0=g0ROcNm4>lJAh0c0d0h0lNeNR59fJGi0?i0h0gNhN`5LXJ4i0<n0i0aNkNR60_J9U1h0[NnNQ61`J4[1j0TNPOS61_J1`1k0oMROR63^JMg1j0i4YOhLc0Y3\\OjLa0V3_OmL>S8L4M4L3M3LdX31[gL5J4L5L3L4M4K4M3L5K4M3LP1QO4K2N2O1N2O0O2O1N1O2O1N2O0O2O1O1N101N2O0O2O1O1N101N2O1O2ZOf0K5O001O1O001O1O1O001O1O001O1O00000000O10000000000O010000000O1000cN\\KSJd4l5cKnI]4Q6dKoI\\4P6eKPJZ4P6gKPJY4o5hKQJX4o5hKQJX4n5jKQJV4o5jKQJV4o5jKQJU4P6kKPJU4P6kKPJT4P6mKPJS4P6mKPJR4Q6nKoIQ4R6oKnIQ4R6PLmIo3T6QLlIn3T6SLlIm3T6SLlIl3U6TLjIm3V6SLjIl3W6TLiIk3X6ULgIl3X6VLeIk3\\6ULaIm3`6SL]Io3d6QLYIP4i6PLTIR4l6oKRIR4n6oKoHR4R7oKkHS4U7nKhHS4Z7n0000O100O10001N10001N2O2N1000001O0O10001ZMVHIj75XHKh74ZHKf74[HLf72\\HMd71_HNa71`HO`70bHN`70aH0_7OcH0^7MeH2\\7LeH4[7KgH4Z7KfH5[7IgH6Z7IfH7[7GgH8Z7FhH8Z7GfH9[7EgH:Y7FgH:Z7DhH;Y7CiH<X7ChH=Y7BhH=Y7BgH>Z7AgH>Y7CgH;Z7EfH;Z7EgH:Z7EfH;Z7EgH:Y7FgH:Y7FhH9Y7FhH9X7GhH9X7GiH8W7HiH7Y7HhH7X7IiH6W7KhH5X7KiH4X7KhH5X7KiH4W7LkH2U7NPIMQ72TIIl67YICh6=]I^Oc6b0bIYO_6f0fIUOZ6k0kIQOT6P1oImNP6S1SJjNn5U1UJiNj5W1XJhNg5X1\\JeNd5[1_JcNa5\\1W3O1N3N1O1N2O2N1N2O1N3N2N4K5L4L3L5L4L`\\>"}, "label": "the motorcycle the nearest woman is on"}]}
{"id": 462398, "image": "COCO_train2014_000000462398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black motorcycle the woman is sitting on\"."}], "task": "refering", "answer_template": "The \"the black motorcycle the woman is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 127, 128, 133, 134, 145, 146, 147, 148, 149, 150, 151, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 301, 302, 303, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 335], "bbox": [0.11871875000000001, 0.26014051522248244, 0.9451875000000001, 0.9898594847775175], "iscrowd": 0, "area": 75662.64475000002, "rle": {"size": [427, 640], "counts": "QQP19P=`0_O;F3M4K4M4K4M3M4L3N3M2N2N3^OkMiEW2S:nMjEU2S:nMiEU2V:b0M2N2N3N0O2N1O2O0O2N1O2O0O2N1O2O0O2N2N101N1O2O001N101O001O0O2OO1000O010000O01000O1000O1000000O10O100000O1000000O1000000O0101O000000000000001O000O1000000000001O01O001O00001O01O01O00001OOO101N100O2O0O101N100O1^NbKXJ_4c5PLQJQ4j5`LjI`3T6kLbIW3]6PM\\IP3d6k101O000010N2O1O1O1N2O1O2N1N2O00O1O00O1O001O1O1O1N21N3N2NbNTIaLj6Y3aIdL]6W3mIfLP6V3[JfLc5T3hJiLV5R3TKkLj4U3[KhLc4Y3aKdL\\4^3iK^LU4b3PL[Ln3f3VLVLi3k3[LQLd3P4i22O2N8H8H8G9H8H8H8H8I7I7H8IQMfIUOU6h0lI_Oo5>SJEj5:VJJg54[JNd5[OWJcM6V3`5WO[J`M7b3V5mNdJ_M7i3P5gNlJ\\M5S4j4aNRKYM6[4e4ZNVKYM5c4h4lMTK^M6f4Q5`M[Ka2S5PMlJQ3_5dLaJ]3^5cLbJ]3_5aLaJa3^5_LbJa3_5^L`Jd3_5\\LaJd3`5ZLaJf3_5ZL`Jg3`5YL`Jh3d4PK^KW1Ni3d4RK\\KU1Oj3e4SK[KQ11l3d4VKVKP16j3c4YKTKn08k3c4YKQKn0<i3c4[KmJ>J@f0h4c4\\KiJ<0@d0h4b4\\LlJlNb0h4a4UKaK2^O1?j4_4PKmKOWO6>k4]4lJcL9POk4X6UKgIl4Y6TKgIm4X6SKhIm4X6RKhIo4X6QKhIo4X6QKhIo4X6QKhIP5W6oJjIQ5V6oJjIQ5V6oJjIQ5V6oJjIQ5V6oJjIR5U6nJkIR5U6oJjIQ5V6oJjIQ5V6PKjIo4V6QKjIP5U6QKjIo4V6QKjIo4V6RKiIn4W6RKiIn4W6SKhIm4X6SKhIm4X6TKgIl4Y6TKgIl4Y6UKfIk4Z6UKgIj4Z6VKeIj4[6VKeIk4Z6VKeIj4\\6UKdIk4\\6VKcIj4]6VKcIj4^6VKaIj4_6VKaIj4_6WK`Ii4a6VK_Ij4_6YK`Ig4]6\\KdIc4X6bKgI^4V6eKjI\\4R6hKmIX4P6kKPJU4o5mKPJS4Q6lKoIT4Q6PKiI;6e4Q6kJnI`01e4Q6jJoIa00e4Q6kJnIa01c4Q6lJmIb03a4P6mJmIb03a4P6mJmIb04`4o5oJlIa05a4n5nJmIa06_4n5PKlIb06\\4o5RKkIb06\\4m5UKmI>7[4e5_KTJ76Y4c5dKXJ26X4c5gKVJ27U4d5jKUJHIG=g4f5jKTJFKH<f4f5mKRJDNI9f4g5nKRJBOI9f4f5oKRJBOJ8e4g5oKSJBNI9e4g5oKRJCOI8c4h5QLQJCOI8c4h5QLRJBOJ7b4h5RLRJB0I6c4h5RLSJA0J5a4j5TLPJA1J6`4i5ULPJB1N0Z5P5VKPKA1N0Y5P5XKoJA2L0[5o4XKoJA2L1Z5o4XKoJA2J3[5m4ZKnJA3H3]5l4ZKnJAg0b1TO`1W5]MoJAf0]1ZOc1Q5_MoJAf0Z1@b1m4bMmJBg0V1Ee0IXNn4;nJAg0R1Kg0CZNm4<nJAg0o01f0^O]Nm4=mJAh0k06g0YO_Nl4?mJ@h0g0=g0ROcNm4>lJAh0c0d0h0lNeNR59fJGi0?i0h0gNhN`5LXJ4i0<n0i0aNkNR60_J9U1h0[NnNQ61`J4[1j0TNPOS61_J1`1k0oMROR63^JMg1j0i4YOhLc0Y3\\OjLa0V3_OmL>S8L4M4L3M3LdX31[gL5J4L5L3L4M4K4M3L5K4M3LP1QO4K2N2O1N2O0O2O1N1O2O1N2O0O2O1O1N101N2O0O2O1O1N101N2O1O2ZOf0K5O001O1O001O1O1O001O1O001O1O00000000O10000000000O010000000O1000cN\\KSJd4l5cKnI]4Q6dKoI\\4P6eKPJZ4P6gKPJY4o5hKQJX4o5hKQJX4n5jKQJV4o5jKQJV4o5jKQJU4P6kKPJU4P6kKPJT4P6mKPJS4P6mKPJR4Q6nKoIQ4R6oKnIQ4R6PLmIo3T6QLlIn3T6SLlIm3T6SLlIl3U6TLjIm3V6SLjIl3W6TLiIk3X6ULgIl3X6VLeIk3\\6ULaIm3`6SL]Io3d6QLYIP4i6PLTIR4l6oKRIR4n6oKoHR4R7oKkHS4U7nKhHS4Z7n0000O100O10001N10001N2O2N1000001O0O10001ZMVHIj75XHKh74ZHKf74[HLf72\\HMd71_HNa71`HO`70bHN`70aH0_7OcH0^7MeH2\\7LeH4[7KgH4Z7KfH5[7IgH6Z7IfH7[7GgH8Z7FhH8Z7GfH9[7EgH:Y7FgH:Z7DhH;Y7CiH<X7ChH=Y7BhH=Y7BgH>Z7AgH>Y7CgH;Z7EfH;Z7EgH:Z7EfH;Z7EgH:Y7FgH:Y7FhH9Y7FhH9X7GhH9X7GiH8W7HiH7Y7HhH7X7IiH6W7KhH5X7KiH4X7KhH5X7KiH4W7LkH2U7NPIMQ72TIIl67YICh6=]I^Oc6b0bIYO_6f0fIUOZ6k0kIQOT6P1oImNP6S1SJjNn5U1UJiNj5W1XJhNg5X1\\JeNd5[1_JcNa5\\1W3O1N3N1O1N2O2N1N2O1N3N2N4K5L4L3L5L4L`\\>"}, "label": "the black motorcycle the woman is sitting on"}]}
{"id": 331326, "image": "COCO_train2014_000000331326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a 4 burner gas stove with tea pot on it\"."}], "task": "refering", "answer_template": "The \"a 4 burner gas stove with tea pot on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.329015625, 0.6337028301886792, 0.75034375, 0.8943867924528303], "iscrowd": 0, "area": 24461.662800000006, "rle": {"size": [424, 640], "counts": "iTh21W=0O10000O100O1N2N2O1N2N2O1N2N2UMEgH=Y7M\\H4d76PHLP8?cGC]8g0XGYOi8Q1kFQOU9Y1_FiNa9_200000O10000001O0000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000hMWFQ1i9mNZFR1f9kN_FR1b9kNcFS1]9jNgFU1Y9iNkFU1U9hNPGV1P9gNTGX1l8eNYGY1g8eN\\GZ1d8cNaG[1_8bNfG\\1Z8aNjG^1V8`NnG^1o9O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2NWlQ2"}, "label": "a 4 burner gas stove with tea pot on it"}]}
{"id": 331326, "image": "COCO_train2014_000000331326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stove top\"."}], "task": "refering", "answer_template": "The \"a stove top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.329015625, 0.6337028301886792, 0.75034375, 0.8943867924528303], "iscrowd": 0, "area": 24461.662800000006, "rle": {"size": [424, 640], "counts": "iTh21W=0O10000O100O1N2N2O1N2N2O1N2N2UMEgH=Y7M\\H4d76PHLP8?cGC]8g0XGYOi8Q1kFQOU9Y1_FiNa9_200000O10000001O0000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000hMWFQ1i9mNZFR1f9kN_FR1b9kNcFS1]9jNgFU1Y9iNkFU1U9hNPGV1P9gNTGX1l8eNYGY1g8eN\\GZ1d8cNaG[1_8bNfG\\1Z8aNjG^1V8`NnG^1o9O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2NWlQ2"}, "label": "a stove top"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with blue plaid shirt and glasses\"."}], "task": "refering", "answer_template": "The \"boy with blue plaid shirt and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336, 356, 357, 358, 359, 379, 380, 381, 382], "bbox": [0.448046875, 0.32208333333333333, 0.695453125, 0.9818125], "iscrowd": 0, "area": 31429.455850000002, "rle": {"size": [480, 640], "counts": "[kV41m>3N1O2N2M2O2N2M2O2N2N1N3N2N1O2M101O1N2N2iNW1N2VEhMi8[2PG[N_8g1[GmNT8V1eG@j7b0oG3`70ZH`0Y7C`Hd0^7^O[Hi0b7YOYHm0e7X3M3N2M3M3O1O1O1]NcHbK^7]4iH]KW7b4PIWKQ7g4WIRKj6m4]IlJd6S5cIgJ]6W5kIbJV6]5g1O1N2O10000O1000000O1000000O1000000O1000000O1000000O1000000O10O100000001O00001O001O00001O00001O001O001O1O1O1O1O001O1O1O1cGaJW7`5cHfJ\\7[5^HkJa7U5ZHQKe7P5UHWKi7j4QH\\Kn7l5K6G8H8K5K6J5K5K5K5E:E<E;C<D=C=G8J7K5M2O2N2N2N1O4L4K5L3M4L4L4L4L4L4K4M3M4L3L5L3M3L5L3M3L;F9GbSk2"}, "label": "boy with blue plaid shirt and glasses"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fair haired guy in dark blue / electric blue checkered shirt playing wii\"."}], "task": "refering", "answer_template": "The \"fair haired guy in dark blue / electric blue checkered shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336, 356, 357, 358, 359, 379, 380, 381, 382], "bbox": [0.448046875, 0.32208333333333333, 0.695453125, 0.9818125], "iscrowd": 0, "area": 31429.455850000002, "rle": {"size": [480, 640], "counts": "[kV41m>3N1O2N2M2O2N2M2O2N2N1N3N2N1O2M101O1N2N2iNW1N2VEhMi8[2PG[N_8g1[GmNT8V1eG@j7b0oG3`70ZH`0Y7C`Hd0^7^O[Hi0b7YOYHm0e7X3M3N2M3M3O1O1O1]NcHbK^7]4iH]KW7b4PIWKQ7g4WIRKj6m4]IlJd6S5cIgJ]6W5kIbJV6]5g1O1N2O10000O1000000O1000000O1000000O1000000O1000000O1000000O10O100000001O00001O001O00001O00001O001O001O1O1O1O1O001O1O1O1cGaJW7`5cHfJ\\7[5^HkJa7U5ZHQKe7P5UHWKi7j4QH\\Kn7l5K6G8H8K5K6J5K5K5K5E:E<E;C<D=C=G8J7K5M2O2N2N2N1O4L4K5L3M4L4L4L4L4L4K4M3M4L3L5L3M3L5L3M3L;F9GbSk2"}, "label": "fair haired guy in dark blue / electric blue checkered shirt playing wii"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with long blonde hair\"."}], "task": "refering", "answer_template": "The \"a women with long blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 369, 370], "bbox": [0.0016875000000000002, 0.3236041666666667, 0.20393750000000002, 0.9865208333333333], "iscrowd": 0, "area": 25264.6906, "rle": {"size": [480, 640], "counts": "ch07_>:F:F:F:M3M3M4L3M4C<O1WDoMb:S2TE]Nd:d1SEmNd:o2G:F9F;E9H9F:G8H7H7I8H7I7I8H7I8H7I7I3N2N2N3M2M3M3M3M3M3M3N20001N10000O100O10000O101O0000000000000000001O0001O001O001O1O1O2N101H8H9F:G:F:E=D;E;E;D:G9F:F:G8G8H8I8G8I7J7J5L4K6K3M4K5L4O1O1O001O1O1O100O001O1O1O1O012J7H7H9H7I8H7I8H7I7I8H\\Z^7"}, "label": "a women with long blonde hair"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl holding wii remote who is not wearing a scarf\"."}], "task": "refering", "answer_template": "The \"girl holding wii remote who is not wearing a scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 369, 370], "bbox": [0.0016875000000000002, 0.3236041666666667, 0.20393750000000002, 0.9865208333333333], "iscrowd": 0, "area": 25264.6906, "rle": {"size": [480, 640], "counts": "ch07_>:F:F:F:M3M3M4L3M4C<O1WDoMb:S2TE]Nd:d1SEmNd:o2G:F9F;E9H9F:G8H7H7I8H7I7I8H7I8H7I7I3N2N2N3M2M3M3M3M3M3M3N20001N10000O100O10000O101O0000000000000000001O0001O001O001O1O1O2N101H8H9F:G:F:E=D;E;E;D:G9F:F:G8G8H8I8G8I7J7J5L4K6K3M4K5L4O1O1O001O1O1O100O001O1O1O1O012J7H7H9H7I8H7I8H7I7I8H\\Z^7"}, "label": "girl holding wii remote who is not wearing a scarf"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in gray checkered polo playing wii\"."}], "task": "refering", "answer_template": "The \"a man in gray checkered polo playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 122, 123, 124, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330, 331, 351, 352, 353, 354, 374, 375, 376, 377], "bbox": [0.264609375, 0.2876458333333333, 0.45168749999999996, 0.9865208333333333], "iscrowd": 0, "area": 23959.918950000003, "rle": {"size": [480, 640], "counts": "Z__22a>=K5L5J5L4K6J5YHiNb1Z1XNnNe1T1ZKgNL<g4o0[KVOYO3Y5i0[KFeNJl5b0]K6QN_O`6<^KT2^4oM_KS2`4nM^KT2`4QN[KQ2e4RNZKl1f4WNYKg1g4\\NXKb1h4aNWK]1i4fNVKX1i4kNWKS1i4POVKn0j4TOVKj0j4YO[JWME^3P6]OPJaMOP3P6CeIhM;c2P6l0PJROP6Q1oImNQ6U1oIiNQ6Z1nIdNR6^1nI`NQ6c1oI[NQ6d1RJZNn5c1WJ[Ni5b1\\J\\Nd5`1bJ^N]5_1iJ_NW5^1nJ`NR5\\1SKcNn3]MYKl3m0gNg3`M[Kg3Q1iNa3dM\\K`32[L4a2\\4gM]Kd2WOhMl0_O3T2d4UNTKa2[OcMj0F2h1a4lNVKS2@_Mg0NO[1m4VOlJS2B[Mf04No0V5AbJR2FVMd0;Ld0`5JYJR2HRMc0a0K7j56nIQ2LmLb0g0ILS6`0dIR2k0fM[OGV6d0`IR2m0fM\\OJP6?dIT2n0fM]OMj5;gIT2Q1gM\\O0f5g3lJ\\L]O3`5b3QK_L]O5[5]3VKaL]O9W5V3[KcL]O<S5Q3^KfL^O?n4k2bKiL^Oc0k4d2dKmL_Oe0g4`2fKnLBg0c4]2gKoLDh0b4[2fKPMGf0b4\\2cKK^46_KKa4Z5O11O1O00O11O1OO1010O0\\G]Kk6c4SI`Kl6`4QIeKm6\\4PIgKo6Y4nHlKP7T4nHoKQ7R4lHRLR7Q4iHRLV7Q4eHRLZ7Q4`HRL`7Q4[HQLe7R4VHQLj7b5F:B?A>C>D;F:F;D;F:F;VNkF\\M`9e2^1O000010O0001O0N3M2N5K6K4K5L5J5L4L4WOQZT5"}, "label": "a man in gray checkered polo playing wii"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with a gray scarf\"."}], "task": "refering", "answer_template": "The \"the woman with a gray scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.78415625, 0.42310416666666667, 1.0, 1.0], "iscrowd": 0, "area": 28008.070700000004, "rle": {"size": [480, 640], "counts": "Pe[72?3g=2YBOb=5]BN]=5dBLW=8hBHU=<kBDQ=?oBAn<c0RC\\Ok<g0VCYOf<k0ZCTOc<P1\\CPOa<S1_CnN]<V1cCiNZ<Z1fCfNV<_1iCbNS<a1mC_NQ<d1nC\\NQ<f1nC[No;g1QDYNm;j1RDVNm;l1RDTNl;o1SDPNl;T2RDlMl;W2SDiMl;X2TDgMk;[2UDeMk;Z2VDeM3d0c:h1ZEWOQ1[Ni7_2UGF=QN]8Z2UGc0k8^OSGc0l8i2O1O1O100O1O1O100O1O1O10000000000O100000000O10000000000N2O1N2]Oc0A?I6N3N2M3N2N2N2O1O1O1O100O1O1O1O10001N10000000000000000000000O100000000000001O000000000000000001O0000001O001O001O1O100O1RI^IU5j6fJ^IU5e6eJaIW5b6eJcIV5a6iJ`IS5d6lJ^IQ5e6nJ[IP5i6mJYIP5j6nJWIP5l6cJ`IZ5U8N2L4L>AZ3gLkC"}, "label": "the woman with a gray scarf"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with a gray scarf\"."}], "task": "refering", "answer_template": "The \"the woman with a gray scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.78415625, 0.42310416666666667, 1.0, 1.0], "iscrowd": 0, "area": 28008.070700000004, "rle": {"size": [480, 640], "counts": "Pe[72?3g=2YBOb=5]BN]=5dBLW=8hBHU=<kBDQ=?oBAn<c0RC\\Ok<g0VCYOf<k0ZCTOc<P1\\CPOa<S1_CnN]<V1cCiNZ<Z1fCfNV<_1iCbNS<a1mC_NQ<d1nC\\NQ<f1nC[No;g1QDYNm;j1RDVNm;l1RDTNl;o1SDPNl;T2RDlMl;W2SDiMl;X2TDgMk;[2UDeMk;Z2VDeM3d0c:h1ZEWOQ1[Ni7_2UGF=QN]8Z2UGc0k8^OSGc0l8i2O1O1O100O1O1O100O1O1O10000000000O100000000O10000000000N2O1N2]Oc0A?I6N3N2M3N2N2N2O1O1O1O100O1O1O1O10001N10000000000000000000000O100000000000001O000000000000000001O0000001O001O001O1O100O1RI^IU5j6fJ^IU5e6eJaIW5b6eJcIV5a6iJ`IS5d6lJ^IQ5e6nJ[IP5i6mJYIP5j6nJWIP5l6cJ`IZ5U8N2L4L>AZ3gLkC"}, "label": "the woman with a gray scarf"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch behind the blonde girl\"."}], "task": "refering", "answer_template": "The \"the couch behind the blonde girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 256, 257, 258, 276, 279, 280, 281, 299, 302, 303, 304, 322, 325, 326, 327, 345, 348, 349, 350, 368, 369, 371, 372, 373], "bbox": [0.0, 0.6580625, 0.248390625, 1.0], "iscrowd": 0, "area": 10595.521999999999, "rle": {"size": [480, 640], "counts": "R:1o>1N`2bA[Oc9f0\\FZOd9g0\\FWOe9j0ZFVOf9[ORF57`0f9XOcFNEj0h9UORGFVOS1i9SObG^OdN_1j9oNYHQ1f7lN^HT1b7iNaHV1`7fNeHY1[7cNiH]1V7aNmH_1S7]NTI`1X:M4M2N3M2M4M3M2M4L3M4L3N3L3M4L3M`_f05X`YO8I7H8I7H8H9H4K3N3L3N3L4M2M4K5J5J7RN^MXGg2i8h1O10SKPFi4Q:VKPFj4T:N10000O2O0O10001N100O2O000O101N10000O2O0O101O0O100O2O000O101N10001O02N2N01N1O1O101N1O1O2E:E;F:E<D;E;E;E<D;E;EmTQ7"}, "label": "the couch behind the blonde girl"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair behind a girl\"."}], "task": "refering", "answer_template": "The \"a brown chair behind a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 256, 257, 258, 276, 279, 280, 281, 299, 302, 303, 304, 322, 325, 326, 327, 345, 348, 349, 350, 368, 369, 371, 372, 373], "bbox": [0.0, 0.6580625, 0.248390625, 1.0], "iscrowd": 0, "area": 10595.521999999999, "rle": {"size": [480, 640], "counts": "R:1o>1N`2bA[Oc9f0\\FZOd9g0\\FWOe9j0ZFVOf9[ORF57`0f9XOcFNEj0h9UORGFVOS1i9SObG^OdN_1j9oNYHQ1f7lN^HT1b7iNaHV1`7fNeHY1[7cNiH]1V7aNmH_1S7]NTI`1X:M4M2N3M2M4M3M2M4L3M4L3N3L3M4L3M`_f05X`YO8I7H8I7H8H9H4K3N3L3N3L4M2M4K5J5J7RN^MXGg2i8h1O10SKPFi4Q:VKPFj4T:N10000O2O0O10001N100O2O000O101N10000O2O0O101O0O100O2O000O101N10001O02N2N01N1O1O101N1O1O2E:E;F:E<D;E;E;E<D;E;EmTQ7"}, "label": "a brown chair behind a girl"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch behind two men\"."}], "task": "refering", "answer_template": "The \"couch behind two men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 285, 286, 287, 290, 291, 292, 293, 308, 309, 310, 313, 314, 315, 316, 317, 331, 337, 338, 339, 340, 354, 355, 356, 359, 360, 361, 362, 363, 377, 378, 379, 382, 383, 384, 385, 386], "bbox": [0.40075000000000005, 0.6862708333333334, 0.8047187499999999, 0.9974166666666666], "iscrowd": 0, "area": 15879.551200000005, "rle": {"size": [480, 640], "counts": "n^h31f>9G9F:G9TOnNaC[1Z<iNfCX1T<mNmCS1n;QOTDn0g;VOZDj0`;[OaDe0Z;_OgDa0f9cNQGQ1YO<f9mNaFl0J7e9[OnEb0>3d9o0]FQOc9n0^FROc9l0^FTOb9k0_FUOa9j0`FVO`9i0aFWO`9g0aFYO_9f0bFZO^9f0bFZO^9e0cF[O^9c0cF]O]9b0eF]O[9c0eF]O[9b0fF^OZ9a0gF_OZ9?gFAY9?gFAY9>hFBX9=iFCX9;iFEW9;iFEW9:jFFV99kFGU99kFGU98lFHU96lFJT95mFKS95mFKS94nFLR93PGLP93QGMo83QGMP91QGOo80RG0n8NTG2l8LVG4j8JXG6h8HZG8g8D\\G<d8B^G>b8@`G`0`8]OcGc0]8[OeGe0[8YOgGg0Z8VOhGj0X8SOmGk0S8SOWHe0i7YO`H`0`7^OiH;W7BSI7m6G\\I2e6KYjT18\\YkN<E;M300000000001O00kL_OkGa0U8:PGFP9h0bFWO_9i0aFWO_9j0`FVOa9i0`FVO`9k0_FUOa9k0_FUOa9l0^FTOb9l0^FTOb9m0]FSOc9m0^FROb9o0]FQOd9n0\\FROd9o0[FQOe9P1ZFPOf9P1ZFPOf9Q1YFoNg9Q1ZFnNf9S1YFmNg9S1YFmNh9S1WFmNi9S1WFmNi9T1VFlNj9T1WFkNi9V1VFjNj9V1VFjNj9W1UFiNk9W1UFiNl9W1SFiNm9W1SFiNm9X1SFgNm9Z1RFfNn9Z1RFfNn9[1QFeNo9[1QFeNP:[1oEeNQ:[1oEeNQ:\\1oEcNQ:]1oEcNQ:^1nEbNR:^1nEbNR:_1mEaNS:_1mEaNT:_1lE`NT:`1lE`NT:a1kE^NV:b1jE^NV:c1iE]NW:d1hE\\NX:d1hE\\NX:e1hEZNY:e1gE[NY:f1fEZNZ:f1fEZNZ:g1eEYN[:g1eEYN[:h1dEXN\\:h1eEWN\\:i1cEWN]:i1cEWN]:j1aEWN_:i1`EXN`:i1]EYNc:g1\\EZNd:g1ZEZNf:g1XEZNi:e1UE]Nk:d1SE]Nm:c1RE^Nn:c1PE^NP;b1nD`NR;g200001O0000003M3M4L2N2N1O1O1O1L4K6K4L4L4K5L4L4L5G8G9G9G9G9G:FTWj1"}, "label": "couch behind two men"}]}
{"id": 495169, "image": "COCO_train2014_000000495169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch that is behind all of the people playing wii ; it has a person ' s jacket on it\"."}], "task": "refering", "answer_template": "The \"a brown couch that is behind all of the people playing wii ; it has a person ' s jacket on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 285, 286, 287, 290, 291, 292, 293, 308, 309, 310, 313, 314, 315, 316, 317, 331, 337, 338, 339, 340, 354, 355, 356, 359, 360, 361, 362, 363, 377, 378, 379, 382, 383, 384, 385, 386], "bbox": [0.40075000000000005, 0.6862708333333334, 0.8047187499999999, 0.9974166666666666], "iscrowd": 0, "area": 15879.551200000005, "rle": {"size": [480, 640], "counts": "n^h31f>9G9F:G9TOnNaC[1Z<iNfCX1T<mNmCS1n;QOTDn0g;VOZDj0`;[OaDe0Z;_OgDa0f9cNQGQ1YO<f9mNaFl0J7e9[OnEb0>3d9o0]FQOc9n0^FROc9l0^FTOb9k0_FUOa9j0`FVO`9i0aFWO`9g0aFYO_9f0bFZO^9f0bFZO^9e0cF[O^9c0cF]O]9b0eF]O[9c0eF]O[9b0fF^OZ9a0gF_OZ9?gFAY9?gFAY9>hFBX9=iFCX9;iFEW9;iFEW9:jFFV99kFGU99kFGU98lFHU96lFJT95mFKS95mFKS94nFLR93PGLP93QGMo83QGMP91QGOo80RG0n8NTG2l8LVG4j8JXG6h8HZG8g8D\\G<d8B^G>b8@`G`0`8]OcGc0]8[OeGe0[8YOgGg0Z8VOhGj0X8SOmGk0S8SOWHe0i7YO`H`0`7^OiH;W7BSI7m6G\\I2e6KYjT18\\YkN<E;M300000000001O00kL_OkGa0U8:PGFP9h0bFWO_9i0aFWO_9j0`FVOa9i0`FVO`9k0_FUOa9k0_FUOa9l0^FTOb9l0^FTOb9m0]FSOc9m0^FROb9o0]FQOd9n0\\FROd9o0[FQOe9P1ZFPOf9P1ZFPOf9Q1YFoNg9Q1ZFnNf9S1YFmNg9S1YFmNh9S1WFmNi9S1WFmNi9T1VFlNj9T1WFkNi9V1VFjNj9V1VFjNj9W1UFiNk9W1UFiNl9W1SFiNm9W1SFiNm9X1SFgNm9Z1RFfNn9Z1RFfNn9[1QFeNo9[1QFeNP:[1oEeNQ:[1oEeNQ:\\1oEcNQ:]1oEcNQ:^1nEbNR:^1nEbNR:_1mEaNS:_1mEaNT:_1lE`NT:`1lE`NT:a1kE^NV:b1jE^NV:c1iE]NW:d1hE\\NX:d1hE\\NX:e1hEZNY:e1gE[NY:f1fEZNZ:f1fEZNZ:g1eEYN[:g1eEYN[:h1dEXN\\:h1eEWN\\:i1cEWN]:i1cEWN]:j1aEWN_:i1`EXN`:i1]EYNc:g1\\EZNd:g1ZEZNf:g1XEZNi:e1UE]Nk:d1SE]Nm:c1RE^Nn:c1PE^NP;b1nD`NR;g200001O0000003M3M4L2N2N1O1O1O1L4K6K4L4L4K5L4L4L5G8G9G9G9G9G:FTWj1"}, "label": "a brown couch that is behind all of the people playing wii ; it has a person ' s jacket on it"}]}
{"id": 290370, "image": "COCO_train2014_000000290370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a woman holding a thing of bananas\"."}], "task": "refering", "answer_template": "The \"this is a woman holding a thing of bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 296, 297, 298, 299, 300, 301, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 364, 365, 366, 367, 368, 369, 381, 382, 383, 384, 385, 386], "bbox": [0.2756458333333333, 0.082765625, 0.7910208333333333, 0.9906406249999999], "iscrowd": 0, "area": 90919.73329999999, "rle": {"size": [640, 480], "counts": "^gb21nc04K4L5L4K5L3L5K5L4K4L5L3L4M3L4L4M2M4L3N2M4M2M3O1N3M2OO0010O01O00010O00101N3ROj^OmNXa0R1l^OjNUa0U1o^OfNSa0Z1Q_O`NQa0_1S_O]Nn`0b1i0O1011N2N1O2N1O2N101N2jM\\NkAe1R>TOWAn0d>UOZAo0a>UO[AP1`>SO^AQ1]>SO`AQ1\\>QObAS1Y>ROcAR1X>TObAn0\\>XO_Ai0^6\\MS1R5cN_KS1c4eNmKR1T4eN\\LR1g3jNdLl0^3VOhL`0Z3@mL7V3GSM0n20ZMFh2:_M]Od2b0dMTO^2k0iMmNY2S1iMiNZ2V1hMgNY2[LfI]4S4TOY2^LgI[4S4SOY2`LfI\\4S4POY2cLgI[4R4nNY2fLgI[4R4lNY2gLhIZ4P4mNS3Q1mLmNV3R1jLmNW3S1iLlNX3S1iLmNW3R1jLmNW3R1kLlNV3S1kLlNW3R1jLmNW3R1jLmNW3]NZHl0`4f0W3ZN]HO0XO]4^2W3VNbHI6@R4`2W3SNeHB`0Hd3b2X3oMjHD>Ha3d2X3lMmHG>E^3h2W3gMSIJ<C[3k2X3cMUIN;AY3m2X3`MXI1V5^2c1`MYI0U5_2c1_MZI1T5_2c1^M\\I1S5_2b1^M]I2R5_2b1]M_I2P5`2b1\\M`I3o4`2b1\\MaI1o4c2_3\\MbLc2`3[MaLd2^:O1O1O1O1O1O1O2N100O1O1O1O1O1O1O1O1O100O10001O000000000000000000000000001O000000000000000000000000010O01O1O001O001O1O001O001O001O10O01O001O1O001O002N1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O2N1O101N1O1O2N1O1O2N1O2N1O1`B]La:e3YEaLe:`3WEdLh:]3SEhLm:Y3mDlLR;U3jDoLU;R3fDSMY;o2aDVMU4oNY1l3^JYMT4TOY1d3^J]MU4XOW1]3^J`MV4\\OQ1[3eJ^MT4A=f3ZKnLT4EJR4lK^LV4ITO^4bLmKU4NaNi4UM^KU43lMU5iMmJV47YM`5]N]JU4=dLk5ROmIV4`0PLQ9k3e3K5K6K4K5K5K5L4K5Kg0YOo0QOo0RO_Vn1"}, "label": "this is a woman holding a thing of bananas"}]}
{"id": 290370, "image": "COCO_train2014_000000290370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long - haired woman displaying bananas wearing paisley pants and black top\"."}], "task": "refering", "answer_template": "The \"the long - haired woman displaying bananas wearing paisley pants and black top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 296, 297, 298, 299, 300, 301, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 364, 365, 366, 367, 368, 369, 381, 382, 383, 384, 385, 386], "bbox": [0.2756458333333333, 0.082765625, 0.7910208333333333, 0.9906406249999999], "iscrowd": 0, "area": 90919.73329999999, "rle": {"size": [640, 480], "counts": "^gb21nc04K4L5L4K5L3L5K5L4K4L5L3L4M3L4L4M2M4L3N2M4M2M3O1N3M2OO0010O01O00010O00101N3ROj^OmNXa0R1l^OjNUa0U1o^OfNSa0Z1Q_O`NQa0_1S_O]Nn`0b1i0O1011N2N1O2N1O2N101N2jM\\NkAe1R>TOWAn0d>UOZAo0a>UO[AP1`>SO^AQ1]>SO`AQ1\\>QObAS1Y>ROcAR1X>TObAn0\\>XO_Ai0^6\\MS1R5cN_KS1c4eNmKR1T4eN\\LR1g3jNdLl0^3VOhL`0Z3@mL7V3GSM0n20ZMFh2:_M]Od2b0dMTO^2k0iMmNY2S1iMiNZ2V1hMgNY2[LfI]4S4TOY2^LgI[4S4SOY2`LfI\\4S4POY2cLgI[4R4nNY2fLgI[4R4lNY2gLhIZ4P4mNS3Q1mLmNV3R1jLmNW3S1iLlNX3S1iLmNW3R1jLmNW3R1kLlNV3S1kLlNW3R1jLmNW3R1jLmNW3]NZHl0`4f0W3ZN]HO0XO]4^2W3VNbHI6@R4`2W3SNeHB`0Hd3b2X3oMjHD>Ha3d2X3lMmHG>E^3h2W3gMSIJ<C[3k2X3cMUIN;AY3m2X3`MXI1V5^2c1`MYI0U5_2c1_MZI1T5_2c1^M\\I1S5_2b1^M]I2R5_2b1]M_I2P5`2b1\\M`I3o4`2b1\\MaI1o4c2_3\\MbLc2`3[MaLd2^:O1O1O1O1O1O1O2N100O1O1O1O1O1O1O1O1O100O10001O000000000000000000000000001O000000000000000000000000010O01O1O001O001O1O001O001O001O10O01O001O1O001O002N1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O2N1O101N1O1O2N1O1O2N1O2N1O1`B]La:e3YEaLe:`3WEdLh:]3SEhLm:Y3mDlLR;U3jDoLU;R3fDSMY;o2aDVMU4oNY1l3^JYMT4TOY1d3^J]MU4XOW1]3^J`MV4\\OQ1[3eJ^MT4A=f3ZKnLT4EJR4lK^LV4ITO^4bLmKU4NaNi4UM^KU43lMU5iMmJV47YM`5]N]JU4=dLk5ROmIV4`0PLQ9k3e3K5K6K4K5K5K5L4K5Kg0YOo0QOo0RO_Vn1"}, "label": "the long - haired woman displaying bananas wearing paisley pants and black top"}]}
{"id": 290370, "image": "COCO_train2014_000000290370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man facing the camera wearing long jean shorts and an orange shirt with white , black and red details\"."}], "task": "refering", "answer_template": "The \"a man facing the camera wearing long jean shorts and an orange shirt with white , black and red details\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 104, 105, 120, 121, 122, 123, 137, 138, 139, 140, 154, 155, 156, 157, 172, 173, 174, 189, 190, 206, 207, 223, 224], "bbox": [0.0635, 0.2501875, 0.27210416666666665, 0.585796875], "iscrowd": 0, "area": 12308.11655, "rle": {"size": [640, 480], "counts": "Scc03lc03M3M3M2N3M3M3M2N3M3M3M2N4L3M4L4L3M4L4L3M4L3f@jMe<Z2WChMg<\\2UCeMj<^2RCdMm<`2oBbMo<h4O1ROo0H700O101O00001O00001O0000001O00001O0O101O00001O01O1O10O01O4M5J7I7J5J00100O1O1O100O00100N2C=C=C=D<C<H9O1O1O1N2O2N6J6J6J6J6J6J6J6J6J6J6I9H8H8H`ji6"}, "label": "a man facing the camera wearing long jean shorts and an orange shirt with white , black and red details"}]}
{"id": 290370, "image": "COCO_train2014_000000290370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt stands behind a curly haired dog\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt stands behind a curly haired dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 104, 105, 120, 121, 122, 123, 137, 138, 139, 140, 154, 155, 156, 157, 172, 173, 174, 189, 190, 206, 207, 223, 224], "bbox": [0.0635, 0.2501875, 0.27210416666666665, 0.585796875], "iscrowd": 0, "area": 12308.11655, "rle": {"size": [640, 480], "counts": "Scc03lc03M3M3M2N3M3M3M2N3M3M3M2N4L3M4L4L3M4L4L3M4L3f@jMe<Z2WChMg<\\2UCeMj<^2RCdMm<`2oBbMo<h4O1ROo0H700O101O00001O00001O0000001O00001O0O101O00001O01O1O10O01O4M5J7I7J5J00100O1O1O100O00100N2C=C=C=D<C<H9O1O1O1N2O2N6J6J6J6J6J6J6J6J6J6J6I9H8H8H`ji6"}, "label": "a man in an orange shirt stands behind a curly haired dog"}]}
{"id": 290370, "image": "COCO_train2014_000000290370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown color dog in the front side\"."}], "task": "refering", "answer_template": "The \"brown color dog in the front side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [292, 308, 309, 310, 325, 326, 327, 342, 343, 344, 358, 359, 360, 361, 374, 375, 376, 377, 378, 379], "bbox": [0.007854166666666667, 0.7760625, 0.3079791666666667, 1.0], "iscrowd": 0, "area": 10009.032550000002, "rle": {"size": [640, 480], "counts": "nS31oc01N1O11ON2OLW\\O3jc01N21OO1NJZ\\O6gc0100MI\\\\O8ec0100MG^\\O:dc00MF_\\O;cc00O1MDb\\O<`c01O1MBd\\O>^c0100LAf\\O`0]c00O1O100O100O100O010O1O100O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2M3L3N3L4M3L4M3L4K5L4L4K5L4L4K5H8I7I7O10000O10000O010O10000O10000O10000000000000000001O00000000000001O01O2N2N3M2N3M2N3M2SNj_OLX`03j_OJ\\`00f_ON``0Lb_O3c`0H__O5b`0J`_O4``0Lc_O1\\`00f_ONY`04h_OJW`07l_OFS`0;o_OCV`09k_OE]`03f_OJa`0Oa_OOf`0J]_O3i`0HX_O7n`0BT_O<Sa0]OP_O`0Zb0N2N3M8HSl^6"}, "label": "brown color dog in the front side"}]}
{"id": 290370, "image": "COCO_train2014_000000290370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog closest to the bananas that is light brown\"."}], "task": "refering", "answer_template": "The \"the dog closest to the bananas that is light brown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [292, 308, 309, 310, 325, 326, 327, 342, 343, 344, 358, 359, 360, 361, 374, 375, 376, 377, 378, 379], "bbox": [0.007854166666666667, 0.7760625, 0.3079791666666667, 1.0], "iscrowd": 0, "area": 10009.032550000002, "rle": {"size": [640, 480], "counts": "nS31oc01N1O11ON2OLW\\O3jc01N21OO1NJZ\\O6gc0100MI\\\\O8ec0100MG^\\O:dc00MF_\\O;cc00O1MDb\\O<`c01O1MBd\\O>^c0100LAf\\O`0]c00O1O100O100O100O010O1O100O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2M3L3N3L4M3L4M3L4K5L4L4K5L4L4K5H8I7I7O10000O10000O010O10000O10000O10000000000000000001O00000000000001O01O2N2N3M2N3M2N3M2SNj_OLX`03j_OJ\\`00f_ON``0Lb_O3c`0H__O5b`0J`_O4``0Lc_O1\\`00f_ONY`04h_OJW`07l_OFS`0;o_OCV`09k_OE]`03f_OJa`0Oa_OOf`0J]_O3i`0HX_O7n`0BT_O<Sa0]OP_O`0Zb0N2N3M8HSl^6"}, "label": "the dog closest to the bananas that is light brown"}]}
{"id": 265796, "image": "COCO_train2014_000000265796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis of the man in the red jacket\"."}], "task": "refering", "answer_template": "The \"the skis of the man in the red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 241, 262, 263, 283, 284, 285, 304, 305, 306, 307, 325, 326, 327], "bbox": [0.136546875, 0.6539110070257612, 0.537078125, 1.0], "iscrowd": 0, "area": 2202.7766, "rle": {"size": [427, 640], "counts": "keT14X=0O10O010O010O10O1O0010O01O00001N0010O010O010O010O01O010O010O010O01O010O01O010O010O010O010O0010O0010O010O010O0010O010O010O010O010O00010O010O010O010O01O010O01O010O010O010O01O01O010O010O010O01O01O010O010O010O010O00010O100O10O01O100O10O0100O100O00100O10O0100O10O0O1O2N1010O100OWCM\\<3cC0[<OfC2X<OgC3W<MjC3T<OlC0U<OmC0R<1nCOR<0oC0P<1PDOP<0RDNn;2SDNm;2SDNl;2TDOl;1SDOm;1SD0m;0RD1m;OSD2m;NRD3n;LRD4n;MQD4o;KQD5o;LPD4Q<KoC6P<KoC5R<JnC6R<KmC5T<JlC6T<KkC5V<JjC6V<JjC6W<JhC6X<JhC6Y<JfC6Z<JfC6[<JdC6\\<JdC6]<JbC6_<IaC7_<J`C6a<I_C8`<I_C7b<H^C8b<I]C7d<H\\C8d<I[C7g<GYC9l<000000000000000000001O000001O0000001O0000001O0000001O01O01O000000O1O1O1O11O000000001O0001O01O000000001O000000001O0000001O01O000001O0000002N1O2N[bk3"}, "label": "the skis of the man in the red jacket"}]}
{"id": 265796, "image": "COCO_train2014_000000265796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"long red skis on snow\"."}], "task": "refering", "answer_template": "The \"long red skis on snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 241, 262, 263, 283, 284, 285, 304, 305, 306, 307, 325, 326, 327], "bbox": [0.136546875, 0.6539110070257612, 0.537078125, 1.0], "iscrowd": 0, "area": 2202.7766, "rle": {"size": [427, 640], "counts": "keT14X=0O10O010O010O10O1O0010O01O00001N0010O010O010O010O01O010O010O010O01O010O01O010O010O010O010O0010O0010O010O010O0010O010O010O010O010O00010O010O010O010O01O010O01O010O010O010O01O01O010O010O010O01O01O010O010O010O010O00010O100O10O01O100O10O0100O100O00100O10O0100O10O0O1O2N1010O100OWCM\\<3cC0[<OfC2X<OgC3W<MjC3T<OlC0U<OmC0R<1nCOR<0oC0P<1PDOP<0RDNn;2SDNm;2SDNl;2TDOl;1SDOm;1SD0m;0RD1m;OSD2m;NRD3n;LRD4n;MQD4o;KQD5o;LPD4Q<KoC6P<KoC5R<JnC6R<KmC5T<JlC6T<KkC5V<JjC6V<JjC6W<JhC6X<JhC6Y<JfC6Z<JfC6[<JdC6\\<JdC6]<JbC6_<IaC7_<J`C6a<I_C8`<I_C7b<H^C8b<I]C7d<H\\C8d<I[C7g<GYC9l<000000000000000000001O000001O0000001O0000001O0000001O01O01O000000O1O1O1O11O000000001O0001O01O000000001O000000001O0000001O01O000001O0000002N1O2N[bk3"}, "label": "long red skis on snow"}]}
{"id": 265796, "image": "COCO_train2014_000000265796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier with a blue suite between two other skiers\"."}], "task": "refering", "answer_template": "The \"a skier with a blue suite between two other skiers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 104, 105, 106, 127, 128, 129, 130, 151, 152, 153, 174, 175, 176, 197, 198, 221, 244], "bbox": [0.53353125, 0.25370023419203747, 0.669265625, 0.6805854800936768], "iscrowd": 0, "area": 6038.123099999999, "rle": {"size": [427, 640], "counts": "Q[^42j<?O1O1O1O1O2N11O2N1O2N1O2N00O100O1O100O1O1O100O1O7I7J7H7I7J3[OYN`Ej1\\:[N`Eh1[:^NaEe1[:i0L3N3M3M3M3M3M3M3M4L5K4L5L3fGVLk6o3oHRLQ7R4iHPLX7T4bHmK^7W4\\HjKl7g48G5L0O2N101N101UNcGeN^8Q1oGlNR8k0XHROi7h0`HUOa7d0gHYOZ7a0oHiNeNE^8[1VIfNgNKT8i0e3100O2O0O12O0O1OoC[O];]1MO1N3M2N3N1N5J6J6I7JWWh2"}, "label": "a skier with a blue suite between two other skiers"}]}
{"id": 265796, "image": "COCO_train2014_000000265796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cross - country skier situated in between two other skiers\"."}], "task": "refering", "answer_template": "The \"a cross - country skier situated in between two other skiers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 104, 105, 106, 127, 128, 129, 130, 151, 152, 153, 174, 175, 176, 197, 198, 221, 244], "bbox": [0.53353125, 0.25370023419203747, 0.669265625, 0.6805854800936768], "iscrowd": 0, "area": 6038.123099999999, "rle": {"size": [427, 640], "counts": "Q[^42j<?O1O1O1O1O2N11O2N1O2N1O2N00O100O1O100O1O1O100O1O7I7J7H7I7J3[OYN`Ej1\\:[N`Eh1[:^NaEe1[:i0L3N3M3M3M3M3M3M3M4L5K4L5L3fGVLk6o3oHRLQ7R4iHPLX7T4bHmK^7W4\\HjKl7g48G5L0O2N101N101UNcGeN^8Q1oGlNR8k0XHROi7h0`HUOa7d0gHYOZ7a0oHiNeNE^8[1VIfNgNKT8i0e3100O2O0O12O0O1OoC[O];]1MO1N3M2N3N1N5J6J6I7JWWh2"}, "label": "a cross - country skier situated in between two other skiers"}]}
{"id": 126534, "image": "COCO_train2014_000000126534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small pile of carrots on a pile of beans on a plate\"."}], "task": "refering", "answer_template": "The \"a small pile of carrots on a pile of beans on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 340], "bbox": [0.059379084967320264, 0.5130718954248366, 0.8313562091503267, 0.6912091503267973], "iscrowd": 0, "area": 29923.798399999996, "rle": {"size": [612, 612], "counts": "eje01Sc05J5K6K5J5K6K4K2N2O0O1O2O0O1O2O0O101O00001O00001O001O00001O00010O00001O00001O00001O00001O00001O00001O001O00001O00001O2N3N2M2N1O1O0000001O00001O0000001O00001O0000001O0000001O00010O0000001O00001O0000001O00000010O00010O000010O00010O0001O01O01O00010O00010O000010O00010O0001O01O01O00010O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001OO1N2O1O1O1O1N2O1O1O1O1N2O1O1O11O0000000000001O01O001O000010O01O00001O00010O001O00001O010O00001O001O01O01O00001O00O2O000O100O10000O2O000O100O10001N100O10000O101O0O100O10000O2O000O100O10001N100O10000O101O0O100O10000O2O000O100O10001N1000000000001O0000000000001O000000000000001O000000O1000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O10000000000000O01000000000O10O100000O10000000O010000000000O011O1O2N1N2oNd^O8]a0Eg^O9[a0Ci^O;Xa0Bl^O<Ua0Ao^O<Sa0AQ_O=Qa0_OS_O?ka0N1O0O2O1O001O1O0O2O1O001O1N101O1O00gUm1"}, "label": "a small pile of carrots on a pile of beans on a plate"}]}
{"id": 126534, "image": "COCO_train2014_000000126534.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"several slices of carrot\"."}], "task": "refering", "answer_template": "The \"several slices of carrot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 340], "bbox": [0.059379084967320264, 0.5130718954248366, 0.8313562091503267, 0.6912091503267973], "iscrowd": 0, "area": 29923.798399999996, "rle": {"size": [612, 612], "counts": "eje01Sc05J5K6K5J5K6K4K2N2O0O1O2O0O1O2O0O101O00001O00001O001O00001O00010O00001O00001O00001O00001O00001O00001O001O00001O00001O2N3N2M2N1O1O0000001O00001O0000001O00001O0000001O0000001O00010O0000001O00001O0000001O00000010O00010O000010O00010O0001O01O01O00010O00010O000010O00010O0001O01O01O00010O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001OO1N2O1O1O1O1N2O1O1O1O1N2O1O1O11O0000000000001O01O001O000010O01O00001O00010O001O00001O010O00001O001O01O01O00001O00O2O000O100O10000O2O000O100O10001N100O10000O101O0O100O10000O2O000O100O10001N100O10000O101O0O100O10000O2O000O100O10001N1000000000001O0000000000001O000000000000001O000000O1000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O10000000000000O01000000000O10O100000O10000000O010000000000O011O1O2N1N2oNd^O8]a0Eg^O9[a0Ci^O;Xa0Bl^O<Ua0Ao^O<Sa0AQ_O=Qa0_OS_O?ka0N1O0O2O1O001O1O0O2O1O001O1N101O1O00gUm1"}, "label": "several slices of carrot"}]}
{"id": 151112, "image": "COCO_train2014_000000151112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lights attached to the pole\"."}], "task": "refering", "answer_template": "The \"the lights attached to the pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 20, 21, 22, 23, 38, 39, 40, 41, 58, 59, 60, 77, 78, 79, 96, 97, 98, 116, 117, 135, 136], "bbox": [0.04441947565543071, 0.03728125, 0.214438202247191, 0.333546875], "iscrowd": 0, "area": 10084.665299999999, "rle": {"size": [640, 534], "counts": "gQ??_c04L4L5K4L3N1O2N1h^OZOV?g0e@_OY?a0d@B\\??b@C]?=c@D\\?=c@C]?>b@C]?=c@C]?>a@D^?<b@E]?;c@E\\?<d@D\\?=c@C]?=c@D\\?<d@D\\?<d@D\\?=c@C]?=c@C]?=c@C]?=b@E]?<b@D^?<b@D^?<b@D^?<b@D^?=`@D`?<`@E_?;a@E_?;a@E_?<`@D`?<`@D`?<_@Ea?;_@F_?<`@D_?=a@C_?=a@C^?>b@B]?`0a@B^?>b@B]?b0`@^O_?b2O100O2N100O1O105Ja0@?A>A10001N10000O10000O2O000O100000001O0001O00001B=^Oc0]Ob0]Oc0^Oc0]Ob0]Oc0^Od0\\Oe0ZObkU8"}, "label": "the lights attached to the pole"}]}
{"id": 151112, "image": "COCO_train2014_000000151112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the backside of a traffic light\"."}], "task": "refering", "answer_template": "The \"the backside of a traffic light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 20, 21, 22, 23, 38, 39, 40, 41, 58, 59, 60, 77, 78, 79, 96, 97, 98, 116, 117, 135, 136], "bbox": [0.04441947565543071, 0.03728125, 0.214438202247191, 0.333546875], "iscrowd": 0, "area": 10084.665299999999, "rle": {"size": [640, 534], "counts": "gQ??_c04L4L5K4L3N1O2N1h^OZOV?g0e@_OY?a0d@B\\??b@C]?=c@D\\?=c@C]?>b@C]?=c@C]?>a@D^?<b@E]?;c@E\\?<d@D\\?=c@C]?=c@D\\?<d@D\\?<d@D\\?=c@C]?=c@C]?=c@C]?=b@E]?<b@D^?<b@D^?<b@D^?<b@D^?=`@D`?<`@E_?;a@E_?;a@E_?<`@D`?<`@D`?<_@Ea?;_@F_?<`@D_?=a@C_?=a@C^?>b@B]?`0a@B^?>b@B]?b0`@^O_?b2O100O2N100O1O105Ja0@?A>A10001N10000O10000O2O000O100000001O0001O00001B=^Oc0]Ob0]Oc0^Oc0]Ob0]Oc0^Od0\\Oe0ZObkU8"}, "label": "the backside of a traffic light"}]}
{"id": 323147, "image": "COCO_train2014_000000323147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black jacket and red tie\"."}], "task": "refering", "answer_template": "The \"a man in a black jacket and red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.055796875, 0.12096916299559472, 0.46867187499999996, 0.9883920704845816], "iscrowd": 0, "area": 72277.57574999996, "rle": {"size": [454, 640], "counts": "^Z`0:T<g0WDH[;j1G8H9H8G8H9I7I6K6I7J6I8I8G8H8I7H7J3M4N101N1O2O0O2N101N2N101N2N100O1O1O100O1O100O1O100O1O100O1O100O1O10O01O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O100O1O100O1PNSHgMm7U2ZHgMg7X2^HcMc7\\2bH`M^7e1RHaMe0e0Z7g1[H\\M`0h0V7i1dHWM8o0T7g1nIWNS6e1QJ[No5b1TJ\\Nn5c1TJ[Nm5d1TJ[Nm5d1UJZNl5e1UJYNm5f1UJXNl5g1^JoMc5P2Z3O1O1O1O1O1O1O1O1O00100O1O1000000000000000000000000O100000000000000000000000000001O00001O00001O00001O00001O00001O00001O00001O00001O001O1kGiMY4Y2VKVNj4k1SKWNm4j1oJYNQ5i1kJZNT5g1iJ[NW5g1eJ[N[5f1bJ]N]5e1_J]Na5d1[J_Ne5c1WJ_Nh5c1UJ`Nj5b1SJ_Nm5b1QJ_No5c1oI^NP6c1nI^NR6`2oHaMQ7`2nH`MR7`2iHfMV7[2cHkM]7V2\\HPNd7Q2VHUNi7T4000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O1O001O001O001O001O1aN^IXKb6c4gIYKZ6a4nI\\KR6_4WJ]Kj5]4_J_Ka5\\4gJaKZ5Y4oJcKR5V4WKgKi4T4`KhKa4R4hKjKX4Q4PLlKQ4n3V3K6J8H:F9\\Ne1E:F;F9FW`f4"}, "label": "a man in a black jacket and red tie"}]}
{"id": 323147, "image": "COCO_train2014_000000323147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black suit with a red , white , and blue tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a black suit with a red , white , and blue tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.055796875, 0.12096916299559472, 0.46867187499999996, 0.9883920704845816], "iscrowd": 0, "area": 72277.57574999996, "rle": {"size": [454, 640], "counts": "^Z`0:T<g0WDH[;j1G8H9H8G8H9I7I6K6I7J6I8I8G8H8I7H7J3M4N101N1O2O0O2N101N2N101N2N100O1O1O100O1O100O1O100O1O100O1O100O1O10O01O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O100O1O100O1PNSHgMm7U2ZHgMg7X2^HcMc7\\2bH`M^7e1RHaMe0e0Z7g1[H\\M`0h0V7i1dHWM8o0T7g1nIWNS6e1QJ[No5b1TJ\\Nn5c1TJ[Nm5d1TJ[Nm5d1UJZNl5e1UJYNm5f1UJXNl5g1^JoMc5P2Z3O1O1O1O1O1O1O1O1O00100O1O1000000000000000000000000O100000000000000000000000000001O00001O00001O00001O00001O00001O00001O00001O00001O001O1kGiMY4Y2VKVNj4k1SKWNm4j1oJYNQ5i1kJZNT5g1iJ[NW5g1eJ[N[5f1bJ]N]5e1_J]Na5d1[J_Ne5c1WJ_Nh5c1UJ`Nj5b1SJ_Nm5b1QJ_No5c1oI^NP6c1nI^NR6`2oHaMQ7`2nH`MR7`2iHfMV7[2cHkM]7V2\\HPNd7Q2VHUNi7T4000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O1O001O001O001O001O1aN^IXKb6c4gIYKZ6a4nI\\KR6_4WJ]Kj5]4_J_Ka5\\4gJaKZ5Y4oJcKR5V4WKgKi4T4`KhKa4R4hKjKX4Q4PLlKQ4n3V3K6J8H:F9\\Ne1E:F;F9FW`f4"}, "label": "a man wearing a black suit with a red , white , and blue tie"}]}
{"id": 323147, "image": "COCO_train2014_000000323147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the grey shirt\"."}], "task": "refering", "answer_template": "The \"the man in the grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366], "bbox": [0.5232812499999999, 0.0013436123348017622, 0.9447656249999999, 1.0], "iscrowd": 0, "area": 87818.50670000001, "rle": {"size": [454, 640], "counts": "mnd42R>3M2M3J6J6K5F:A?B>A?B>WOi0^Nb1eN[1UOk0]Oc0N2O1N2N2N2N2N2N2N2N2O1O100O100O100O100O100O100O100aLjGm0W8mNQHQ1o7jNXHS1i7iN]HU1c7gNdHU1]7fNjHX1V7gNmHV1T7iNoHU1Q7jNRIS1o6lNUIQ1k6nNWIP1j6oNXIP1h6oN[In0f6QO\\In0d6RO]Il0d6SO^Il0b6SO`Ik0a6TObIj0^6UOdIj0\\6UOfIi0[6VOhIh0X6WOkIf0V6XOnIf0R6XORJe0o5ZOSJd0n5ZOVJd0j5[OYJb0h5\\O\\Jb0d5]O^Ja0c5^O`J`0`5_OcJ>^5BdJ<\\5DfJ5_5J_40000O1000000O10000000000O1000000000000001O00001O001O00001O00001O001O00001O00001O001O001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O2nDYOa8P1TGTOj8U1kFoNS9Z1aFkN]9_1WFeNg9h1jE\\NT:m2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O4L4L4L4L4L4L4L4L4K5E;D<D;F;D<E;D<SOm0ROV1iNX1hNRY?"}, "label": "the man in the grey shirt"}]}
{"id": 323147, "image": "COCO_train2014_000000323147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing grey colored suit standing beside with black colored suit man\"."}], "task": "refering", "answer_template": "The \"a man wearing grey colored suit standing beside with black colored suit man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366], "bbox": [0.5232812499999999, 0.0013436123348017622, 0.9447656249999999, 1.0], "iscrowd": 0, "area": 87818.50670000001, "rle": {"size": [454, 640], "counts": "mnd42R>3M2M3J6J6K5F:A?B>A?B>WOi0^Nb1eN[1UOk0]Oc0N2O1N2N2N2N2N2N2N2N2O1O100O100O100O100O100O100O100aLjGm0W8mNQHQ1o7jNXHS1i7iN]HU1c7gNdHU1]7fNjHX1V7gNmHV1T7iNoHU1Q7jNRIS1o6lNUIQ1k6nNWIP1j6oNXIP1h6oN[In0f6QO\\In0d6RO]Il0d6SO^Il0b6SO`Ik0a6TObIj0^6UOdIj0\\6UOfIi0[6VOhIh0X6WOkIf0V6XOnIf0R6XORJe0o5ZOSJd0n5ZOVJd0j5[OYJb0h5\\O\\Jb0d5]O^Ja0c5^O`J`0`5_OcJ>^5BdJ<\\5DfJ5_5J_40000O1000000O10000000000O1000000000000001O00001O001O00001O00001O001O00001O00001O001O001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O2nDYOa8P1TGTOj8U1kFoNS9Z1aFkN]9_1WFeNg9h1jE\\NT:m2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O4L4L4L4L4L4L4L4L4K5E;D<D;F;D<E;D<SOm0ROV1iNX1hNRY?"}, "label": "a man wearing grey colored suit standing beside with black colored suit man"}]}
{"id": 454219, "image": "COCO_train2014_000000454219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl flying a kite\"."}], "task": "refering", "answer_template": "The \"a little girl flying a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 234, 235, 236, 249, 250, 251, 264, 265, 266, 280, 281, 294, 295, 296, 309, 310, 311, 325, 326, 340, 341], "bbox": [0.601364705882353, 0.616609375, 0.7994117647058824, 1.0], "iscrowd": 0, "area": 11551.248300000001, "rle": {"size": [640, 425], "counts": "c^P57ic01O1O001O01O01O001O00001O00001O00001O00001d\\ODl10Q?<R_OEf1;S?0U_OH`1d0=VOg<i0fB:?nNn:l0eCVOLh0m0=b0jNo:o0`CQ15\\OZ1eNP;Q1YCR3e1nKR;X5mDiJR;Y5lDhJS;Y5lDhJR;[5mDeJR;]5mDcJR;^5nDbJQ;`5nD`JR;a5mD_JR;c5mD]JS;c5mD]JR;e5mD[JS;f5kD[JU;e5jD\\JV;e5hD\\JX;e5eD]J[;e5bD\\J^;n5WDSJi;R704L000000000001O000O2O00001O001N101O1O1O2N2N2M3N4PNoC^KUOCT=g4jCcKf<V4^CgKg<T4[CiKk<Q4XCmKY=a3kB[LY=b3e1^Lo_OP3d`0K5YMP@W1ba0J2M3N3Jecd1"}, "label": "a little girl flying a kite"}]}
{"id": 454219, "image": "COCO_train2014_000000454219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl with kite\"."}], "task": "refering", "answer_template": "The \"little girl with kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 234, 235, 236, 249, 250, 251, 264, 265, 266, 280, 281, 294, 295, 296, 309, 310, 311, 325, 326, 340, 341], "bbox": [0.601364705882353, 0.616609375, 0.7994117647058824, 1.0], "iscrowd": 0, "area": 11551.248300000001, "rle": {"size": [640, 425], "counts": "c^P57ic01O1O001O01O01O001O00001O00001O00001O00001d\\ODl10Q?<R_OEf1;S?0U_OH`1d0=VOg<i0fB:?nNn:l0eCVOLh0m0=b0jNo:o0`CQ15\\OZ1eNP;Q1YCR3e1nKR;X5mDiJR;Y5lDhJS;Y5lDhJR;[5mDeJR;]5mDcJR;^5nDbJQ;`5nD`JR;a5mD_JR;c5mD]JS;c5mD]JR;e5mD[JS;f5kD[JU;e5jD\\JV;e5hD\\JX;e5eD]J[;e5bD\\J^;n5WDSJi;R704L000000000001O000O2O00001O001N101O1O1O2N2N2M3N4PNoC^KUOCT=g4jCcKf<V4^CgKg<T4[CiKk<Q4XCmKY=a3kB[LY=b3e1^Lo_OP3d`0K5YMP@W1ba0J2M3N3Jecd1"}, "label": "little girl with kite"}]}
{"id": 93786, "image": "COCO_train2014_000000093786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in between 2 other giraffes eating from a tree limb\"."}], "task": "refering", "answer_template": "The \"a giraffe in between 2 other giraffes eating from a tree limb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 125, 126, 142, 143, 160, 177, 178, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 252, 253, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 299, 300, 301, 303, 304, 316, 317, 318, 320, 333, 334, 335, 337, 350, 351, 352, 354, 355, 367, 368, 369, 371], "bbox": [0.381375, 0.28490625, 0.9082916666666667, 0.9335625000000001], "iscrowd": 0, "area": 28989.91895, "rle": {"size": [640, 480], "counts": "Xbb36cc0:F;I7L3M4L3N3L3M3N2M3M3N2M3M3N2N2M4M2N2M3N2N2N1N2O2N1N3N1O1010O01O00010O00O2eNT_O]Om`0:\\_ODk`0N]_O2h`0Aa_O=f`0VOa_Oj0e`0PO\\_On0ma0O3L3N2M3N2N2M4Mm]c0JUb\\O5J5N3M2mN^Of^Od0Pa0HZ^OC6g0[a0O[^OZO9i0Za01Z^OVO:k0[a01Z^OTO:m0Ya05d^ONZa03d^OO[a02d^O0Za01d^O1[a01c^O1[a0[1O^1aN2O1O1O100O1O1O1O1O1dDfKQ7Z4mHgKS7Z4jHhKV7Y4hHhKX7[4iGaLW8e3[GcLf8b3lFfLT9_3_FjL`9\\3mEQMT:T3nDdMR;e500001O00001O0eJgDF:mNo:]1nD[OP1_NR:V2VEPOX1]Nc9e2^E_N_1_NS9V3VF[Ma<i2oC`LU<c3R24L4L4M3L4L4L4L4M3L4LP1PO1O1O2N100O2N1O1O=Ca0@4KkKYCV1b<jNdCU1W<kNoCT1l;lNZDS1a;mNeDR1W;mNoDR1l:nN[ElM1k0c:Y1gGbNX8^1oG\\NP8d1WHUNj7j1YHTNf7l1]HRNb7n1aHoM_7Q2cHnM\\7R2gHlMX7T2kHiMV7V2mHhMR7X2PIgMo6Y2\\50001O01O2N2N101N2N2N1O2N2N1O2O1N1O2N2N2N1O2N2N2O1N1O2N2N2N1O2N1O100O1O1O1O1O1O1QD`K\\8a4aGaK_8`4^GbKc8^4ZGdKf8^4VGeKi8]4VFbLj9`3kEiLU:Y3aEoL_:T3mD]MS;h50000001O000000000aMWDgLi;Q3cDkL];l2oDQMQ;R2PFiMQ:X2QFeMP:[2SFaMm9`2VF\\Mj9e2XFXMh9Z3kE^LX:c3lEVLV:l1mCoNQ2oNT:m1RDSOn1iNW:j1RD\\Oj1dN[:f1RDDh1_N\\:d1SDLh>I_AFW?NPAG`?Ki@Ob?Cg@6fa0J6JnQk0"}, "label": "a giraffe in between 2 other giraffes eating from a tree limb"}]}
{"id": 93786, "image": "COCO_train2014_000000093786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe whose head cannot be seen\"."}], "task": "refering", "answer_template": "The \"the giraffe whose head cannot be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 125, 126, 142, 143, 160, 177, 178, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 252, 253, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 299, 300, 301, 303, 304, 316, 317, 318, 320, 333, 334, 335, 337, 350, 351, 352, 354, 355, 367, 368, 369, 371], "bbox": [0.381375, 0.28490625, 0.9082916666666667, 0.9335625000000001], "iscrowd": 0, "area": 28989.91895, "rle": {"size": [640, 480], "counts": "Xbb36cc0:F;I7L3M4L3N3L3M3N2M3M3N2M3M3N2N2M4M2N2M3N2N2N1N2O2N1N3N1O1010O01O00010O00O2eNT_O]Om`0:\\_ODk`0N]_O2h`0Aa_O=f`0VOa_Oj0e`0PO\\_On0ma0O3L3N2M3N2N2M4Mm]c0JUb\\O5J5N3M2mN^Of^Od0Pa0HZ^OC6g0[a0O[^OZO9i0Za01Z^OVO:k0[a01Z^OTO:m0Ya05d^ONZa03d^OO[a02d^O0Za01d^O1[a01c^O1[a0[1O^1aN2O1O1O100O1O1O1O1O1dDfKQ7Z4mHgKS7Z4jHhKV7Y4hHhKX7[4iGaLW8e3[GcLf8b3lFfLT9_3_FjL`9\\3mEQMT:T3nDdMR;e500001O00001O0eJgDF:mNo:]1nD[OP1_NR:V2VEPOX1]Nc9e2^E_N_1_NS9V3VF[Ma<i2oC`LU<c3R24L4L4M3L4L4L4L4M3L4LP1PO1O1O2N100O2N1O1O=Ca0@4KkKYCV1b<jNdCU1W<kNoCT1l;lNZDS1a;mNeDR1W;mNoDR1l:nN[ElM1k0c:Y1gGbNX8^1oG\\NP8d1WHUNj7j1YHTNf7l1]HRNb7n1aHoM_7Q2cHnM\\7R2gHlMX7T2kHiMV7V2mHhMR7X2PIgMo6Y2\\50001O01O2N2N101N2N2N1O2N2N1O2O1N1O2N2N2N1O2N2N2O1N1O2N2N2N1O2N1O100O1O1O1O1O1O1QD`K\\8a4aGaK_8`4^GbKc8^4ZGdKf8^4VGeKi8]4VFbLj9`3kEiLU:Y3aEoL_:T3mD]MS;h50000001O000000000aMWDgLi;Q3cDkL];l2oDQMQ;R2PFiMQ:X2QFeMP:[2SFaMm9`2VF\\Mj9e2XFXMh9Z3kE^LX:c3lEVLV:l1mCoNQ2oNT:m1RDSOn1iNW:j1RD\\Oj1dN[:f1RDDh1_N\\:d1SDLh>I_AFW?NPAG`?Ki@Ob?Cg@6fa0J6JnQk0"}, "label": "the giraffe whose head cannot be seen"}]}
{"id": 93786, "image": "COCO_train2014_000000093786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small giraffe to the left of three others\"."}], "task": "refering", "answer_template": "The \"a small giraffe to the left of three others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 121, 122, 123, 138, 139, 140, 155, 156, 172, 173, 189, 190, 191, 205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 243, 257, 258, 259, 274, 275, 276, 291, 292, 293, 308, 309, 310, 325, 326, 327, 342, 343, 344, 359, 360, 361, 362, 377, 378, 379], "bbox": [0.08416666666666667, 0.267875, 0.31304166666666666, 0.97190625], "iscrowd": 0, "area": 23544.013600000006, "rle": {"size": [640, 480], "counts": "U[i01mc0;F9F;E:G4K5K5L4K5K5K5gKfMeF_2W9aMgFd2U9^MhFg2U9YMjFk2R9VMlFo2P9SMmFR3P9nLoFV3m8kLQGZ3l8gLQG_3j8bLTGc3i8]LVGg3f8[LWGj3e8WLYGn3R1gKT3;iKR4o0dKW3<gKU4n0`K[3;eKZ4m0[K]3<eK]4j0YK`3;cKb4h0TKc3<cKe4f0PKc3e0_K`4j0lJc3Q1YKW4P1iJd3]1PKo3Y1eJb3i1jJg3`1aJc3T2cJ_3f1^Jc3`2\\JW3m1ZJh3g2PJU3S2UJm3o2fIP3S2XJW4Q3[Il26YL^6`:0O2K4L4L4M3L5KdIoCQ5m;]JiDc5j<00001O2\\JRJRMo5R3UJeLm5^3XJXLj5k3[JkKg5Y4_J[Kc5h4gJhJ\\5[5nJVJS5b4gI^Ia1`1j4R5jI\\If1S1a4a5oIZIj1f0X4EeKf2_NmLo18m34hKg2aNkLT2Kd3b0hKh2fNjLV9=UHj2jNgLl8?[Hj2oNeLb8`0`Hk2TOcLW8a0gHm2WO`Ln7b0lHn2]O]Lb7e0RIo2AZLY7f0WI[3[OmKZ7f0]Ig3UObKX7g0dIQ4POVKX7h0iI[4lNkJV7j0PJ_4kNeJQ7k0UJd4lN_Jj6l0\\Jj4kNXJe6m0aJo4V7PKkHU5o6kJRIZ5i6eJXIb5`6]JbIi5X6VJiIP6P6PJQJW6h5hIYJ^6c5_I^Jg6b5QI`JU7a5cHaJc7k8bLjCWOP<b0`D[OZ;>UEAd:8lEEn93bFE^94QGBQ97_G_Ob8:mG]OT8<\\HZOe7?jHXO]7;RI[O[73QIG[7GoH5]7YOmHc0\\=L4K5L4L4L4Lkd]6"}, "label": "a small giraffe to the left of three others"}]}
{"id": 93786, "image": "COCO_train2014_000000093786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe bends it ' s head down towards other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe bends it ' s head down towards other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 182, 183, 184, 185, 186, 195, 196, 203, 213], "bbox": [0.4640833333333333, 0.340984375, 1.0, 0.5494999999999999], "iscrowd": 0, "area": 18438.177300000007, "rle": {"size": [640, 480], "counts": "^d[48`c0:E;K5N2M3N2M3N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M2O1N2O1N2O1O1O1O00N2O001O1O1O1O1O1O1]Oc_ObM]`0]2m_OZMT`0d2e0N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O20000001O00000O1000000000000O2O00000000000O100000001O00000O1000000000000O2O0000000000000O1000001O00000O1000000000000O2O0000000000000O1000001O00001N10001O00010O0000001O00001O01O01O00001O00001O01O0001O00001O00010O00001O00001O00010O0000001O00001O01O01O00001O00001O01O0001O00001O00010O00001O00001O0001O01O00001O00001O01O01O00001O00000010O0001O00001O002O1N2N3M2N2N2N2N2hH"}, "label": "a giraffe bends it ' s head down towards other giraffes"}]}
{"id": 93786, "image": "COCO_train2014_000000093786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foremost giraffe that is necking with another giraffe\"."}], "task": "refering", "answer_template": "The \"the foremost giraffe that is necking with another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 182, 183, 184, 185, 186, 195, 196, 203, 213], "bbox": [0.4640833333333333, 0.340984375, 1.0, 0.5494999999999999], "iscrowd": 0, "area": 18438.177300000007, "rle": {"size": [640, 480], "counts": "^d[48`c0:E;K5N2M3N2M3N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M2O1N2O1N2O1O1O1O00N2O001O1O1O1O1O1O1]Oc_ObM]`0]2m_OZMT`0d2e0N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O20000001O00000O1000000000000O2O00000000000O100000001O00000O1000000000000O2O0000000000000O1000001O00000O1000000000000O2O0000000000000O1000001O00001N10001O00010O0000001O00001O01O01O00001O00001O01O0001O00001O00010O00001O00001O00010O0000001O00001O01O01O00001O00001O01O0001O00001O00010O00001O00001O0001O01O00001O00001O01O01O00001O00000010O0001O00001O002O1N2N3M2N2N2N2N2hH"}, "label": "the foremost giraffe that is necking with another giraffe"}]}
{"id": 36445, "image": "COCO_train2014_000000036445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut far from the camera\"."}], "task": "refering", "answer_template": "The \"the donut far from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 19, 20, 21, 22, 23, 24, 35, 36, 37, 38, 39, 50, 51, 52, 53, 66], "bbox": [0.3191294117647059, 0.019858657243816254, 0.6106823529411763, 0.4084805653710248], "iscrowd": 0, "area": 10335.865700000002, "rle": {"size": [283, 425], "counts": "jcU14f85K5L4K5L4K5K5L4K4L3N2M3N3L3M3N2M3M4M2M3N2M3M3N3L3N1N2N2O0O2N100O101O0O10001O000O101O000O101O00000O1000000O1000000O10O100O1O010O100O10000O101N10000O101O0O100O101O0O100O101O0O10000O2O0O10000O2O0O10000O2O0O10000O2O1O1N2O1N2O1O1M3N2N2N2M3_NSIQ1o6kNXIP1j6mN\\Im0g6oN`Il0\\7Kf0ZOk`]1"}, "label": "the donut far from the camera"}]}
{"id": 36445, "image": "COCO_train2014_000000036445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cookie in the back side of the image\"."}], "task": "refering", "answer_template": "The \"cookie in the back side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 19, 20, 21, 22, 23, 24, 35, 36, 37, 38, 39, 50, 51, 52, 53, 66], "bbox": [0.3191294117647059, 0.019858657243816254, 0.6106823529411763, 0.4084805653710248], "iscrowd": 0, "area": 10335.865700000002, "rle": {"size": [283, 425], "counts": "jcU14f85K5L4K5L4K5K5L4K4L3N2M3N3L3M3N2M3M4M2M3N2M3M3N3L3N1N2N2O0O2N100O101O0O10001O000O101O000O101O00000O1000000O1000000O10O100O1O010O100O10000O101N10000O101O0O100O101O0O100O101O0O10000O2O0O10000O2O0O10000O2O0O10000O2O1O1N2O1N2O1O1M3N2N2N2M3_NSIQ1o6kNXIP1j6mN\\Im0g6oN`Il0\\7Kf0ZOk`]1"}, "label": "cookie in the back side of the image"}]}
{"id": 36445, "image": "COCO_train2014_000000036445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pastry that is to the left of the other two pastries\"."}], "task": "refering", "answer_template": "The \"the pastry that is to the left of the other two pastries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 46, 47, 48, 49, 50, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 92, 93, 94, 95, 108, 109], "bbox": [0.08529411764705883, 0.28275618374558303, 0.39503529411764704, 0.7142402826855123], "iscrowd": 0, "area": 12093.328249999997, "rle": {"size": [283, 425], "counts": "^R:2e88J5K6K4K5K4L5K4M3L5K4L3M3N2M4N1N2N2N3M2N2N2N3N0O2N2N2N2N2N2N2N101O1O1N101O001N101O1O0O2O001O0O2O1O0O2O001O0O2O0000001O0000001O0000001O00000000000000O1000000000000O101O000000001O0O1000000O2N100O1O2O1N3M2N2O2M2N3N1N2N3N1N2N3M2N2O2M2N2N2N3N1N2N3M4L4L3N4J5I7I8H:E;FZlV2"}, "label": "the pastry that is to the left of the other two pastries"}]}
{"id": 36445, "image": "COCO_train2014_000000036445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pastry on the far left\"."}], "task": "refering", "answer_template": "The \"the pastry on the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 46, 47, 48, 49, 50, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 92, 93, 94, 95, 108, 109], "bbox": [0.08529411764705883, 0.28275618374558303, 0.39503529411764704, 0.7142402826855123], "iscrowd": 0, "area": 12093.328249999997, "rle": {"size": [283, 425], "counts": "^R:2e88J5K6K4K5K4L5K4M3L5K4L3M3N2M4N1N2N2N3M2N2N2N3N0O2N2N2N2N2N2N2N101O1O1N101O001N101O1O0O2O001O0O2O1O0O2O001O0O2O0000001O0000001O0000001O00000000000000O1000000000000O101O000000001O0O1000000O2N100O1O2O1N3M2N2O2M2N3N1N2N3N1N2N3M2N2O2M2N2N2N3N1N2N3M4L4L3N4J5I7I8H:E;FZlV2"}, "label": "the pastry on the far left"}]}
{"id": 36445, "image": "COCO_train2014_000000036445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cupcake on the bottom right\"."}], "task": "refering", "answer_template": "The \"cupcake on the bottom right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115], "bbox": [0.4454117647058824, 0.33219081272084805, 0.7438588235294117, 0.8029328621908128], "iscrowd": 0, "area": 13231.34475, "rle": {"size": [283, 425], "counts": "h\\d11e74ZI3`6O^I2a6N_I3`6ORILF8W7OoHKG;U75jH7R7m0J6K5J5K6K5L3N1N2N2N101N2N2O1O1N2O1N2O1O1N101N2O1O0O2O1N2O1N2OO10O100O101O0O10000O2O00000O2O000O101O0O10O100O100O10000000O0100000000O10001O00001N10001O000O2O001N2O0O2O1N10000O103L2O0O101NO2O0100O101N5K1O1O1O1O1O101N2WMSJa2V6M2N3M2N4L4M3M3M8H6POYI@j66^IGkUn0"}, "label": "cupcake on the bottom right"}]}
{"id": 36445, "image": "COCO_train2014_000000036445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottom most pastry to the right of other pastry\"."}], "task": "refering", "answer_template": "The \"bottom most pastry to the right of other pastry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115], "bbox": [0.4454117647058824, 0.33219081272084805, 0.7438588235294117, 0.8029328621908128], "iscrowd": 0, "area": 13231.34475, "rle": {"size": [283, 425], "counts": "h\\d11e74ZI3`6O^I2a6N_I3`6ORILF8W7OoHKG;U75jH7R7m0J6K5J5K6K5L3N1N2N2N101N2N2O1O1N2O1N2O1O1N101N2O1O0O2O1N2O1N2OO10O100O101O0O10000O2O00000O2O000O101O0O10O100O100O10000000O0100000000O10001O00001N10001O000O2O001N2O0O2O1N10000O103L2O0O101NO2O0100O101N5K1O1O1O1O1O101N2WMSJa2V6M2N3M2N4L4M3M3M8H6POYI@j66^IGkUn0"}, "label": "bottom most pastry to the right of other pastry"}]}
{"id": 372319, "image": "COCO_train2014_000000372319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus bench with a blonde lady sitting in it\"."}], "task": "refering", "answer_template": "The \"a bus bench with a blonde lady sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 132, 133, 134, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 286, 288, 289, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.427140625, 0.27320833333333333, 1.0, 1.0], "iscrowd": 0, "area": 66636.13420000001, "rle": {"size": [480, 640], "counts": "o]P41i>6J6J6J6I7J6J6I7H8I7H8I7H8I7H8I7H8I7M3000000O100000000O100000QLXEf3g:ZL^Eb3b:\\LcEa3]:^LhE^3o:N00000000O10000000000O100000000000000000000000000O10000000000000000000000000000O10000O100O100O100O100O100O100O100O100000000000000000000000000001O1O1O2N1O1O1O2N1O1O2N1O1O2N001O1O001O1O000000O1000000O1000000O100O1K5L4L4L4N200O100O100O100O100O100O10000O100O100O100O100O100O100O1O1N2O1N2O1O1N2O1N2O1N2N2M3L4M3L4L4M3L4N2O1O1O1O100K5I7H8J6M3M3M3M3N2M3M3M3M3M3M3jLaHhNb7V1bHeNa7Y1dHbN^7]1eH^N^7c1cHXN`7h1cHRN`7o1aHlMb7T2`HgMc7Z2_H`Md7`2_HZMd7g2_HRMd7n2bHjL`7W3eH`L^7`3eHULc7l3h1001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1M2M4L3N3L4M2M4M2M4M3L3M4M2M4L4M2M4L3N3L4L3N3L3M4M3L3M4L3N3L4L3N3L3M4M2M4L4M2M4L3N3L4L3M4M2M4L4M2M4L3N3L3M4M2M3M4M2M3M4M2M3M4L3N2M4L3N2M4cNhJ"}, "label": "a bus bench with a blonde lady sitting in it"}]}
{"id": 372319, "image": "COCO_train2014_000000372319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty seat next to the woman in purple\"."}], "task": "refering", "answer_template": "The \"the empty seat next to the woman in purple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 132, 133, 134, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 286, 288, 289, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.427140625, 0.27320833333333333, 1.0, 1.0], "iscrowd": 0, "area": 66636.13420000001, "rle": {"size": [480, 640], "counts": "o]P41i>6J6J6J6I7J6J6I7H8I7H8I7H8I7H8I7H8I7M3000000O100000000O100000QLXEf3g:ZL^Eb3b:\\LcEa3]:^LhE^3o:N00000000O10000000000O100000000000000000000000000O10000000000000000000000000000O10000O100O100O100O100O100O100O100O100000000000000000000000000001O1O1O2N1O1O1O2N1O1O2N1O1O2N001O1O001O1O000000O1000000O1000000O100O1K5L4L4L4N200O100O100O100O100O100O10000O100O100O100O100O100O100O1O1N2O1N2O1O1N2O1N2O1N2N2M3L4M3L4L4M3L4N2O1O1O1O100K5I7H8J6M3M3M3M3N2M3M3M3M3M3M3jLaHhNb7V1bHeNa7Y1dHbN^7]1eH^N^7c1cHXN`7h1cHRN`7o1aHlMb7T2`HgMc7Z2_H`Md7`2_HZMd7g2_HRMd7n2bHjL`7W3eH`L^7`3eHULc7l3h1001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1M2M4L3N3L4M2M4M2M4M3L3M4M2M4L4M2M4L3N3L4L3N3L3M4M3L3M4L3N3L4L3N3L3M4M2M4L4M2M4L3N3L4L3M4M2M4L4M2M4L3N3L3M4M2M3M4M2M3M4M2M3M4L3N2M4L3N2M4cNhJ"}, "label": "the empty seat next to the woman in purple"}]}
{"id": 372319, "image": "COCO_train2014_000000372319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus seat that a woman is setting her toes on , not which she is sitting on\"."}], "task": "refering", "answer_template": "The \"a bus seat that a woman is setting her toes on , not which she is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.179, 0.33221875, 1.0], "iscrowd": 0, "area": 46961.52545, "rle": {"size": [480, 640], "counts": "b:^4a:1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000000000000000000000N2B>B>B>B>C=O100O1O100O100O1O100O100O1O100O100O1O10O001N1O1O2O0O2N1OM4K4L4K5L5K4K5L4L5L3M3M3M3M4L3M3002O0O2N2N101N2N1O2O1N1O2N2O0O2N2N2O0O2N2N101N2N1O2O1N1O2N2N2O0O2N2N101N2N1O2O1N1O2N2O1N1O2N2O0O2N2N101N2N1O2O1N2N1O2O1N1O2N2O0O2N2N1O2O1N2N1O2O1N1O2N2O0O2N2N101N2N1O2L4J6J5J7J6I6K6J6I6K6J6I6K6J6I7QNbD6c;FbD3e;HaD1f;I_D1h;I]D0l;H[D1l;IYD1n;IWD0Q<JTD0U^X6"}, "label": "a bus seat that a woman is setting her toes on , not which she is sitting on"}]}
{"id": 372319, "image": "COCO_train2014_000000372319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the metal back of a bus seat row in front of a woman in a pink shirt\"."}], "task": "refering", "answer_template": "The \"the metal back of a bus seat row in front of a woman in a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.179, 0.33221875, 1.0], "iscrowd": 0, "area": 46961.52545, "rle": {"size": [480, 640], "counts": "b:^4a:1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000000000000000000000N2B>B>B>B>C=O100O1O100O100O1O100O100O1O100O100O1O10O001N1O1O2O0O2N1OM4K4L4K5L5K4K5L4L5L3M3M3M3M4L3M3002O0O2N2N101N2N1O2O1N1O2N2O0O2N2N2O0O2N2N101N2N1O2O1N1O2N2N2O0O2N2N101N2N1O2O1N1O2N2O1N1O2N2O0O2N2N101N2N1O2O1N2N1O2O1N1O2N2O0O2N2N1O2O1N2N1O2O1N1O2N2O0O2N2N101N2N1O2L4J6J5J7J6I6K6J6I6K6J6I6K6J6I7QNbD6c;FbD3e;HaD1f;I_D1h;I]D0l;H[D1l;IYD1n;IWD0Q<JTD0U^X6"}, "label": "the metal back of a bus seat row in front of a woman in a pink shirt"}]}
{"id": 372319, "image": "COCO_train2014_000000372319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a barefoot woman holding a bottle of water\"."}], "task": "refering", "answer_template": "The \"a barefoot woman holding a bottle of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 63, 83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 199, 200, 201, 215, 216, 217, 218, 223, 224, 238, 239, 240, 245, 246, 260, 261, 262, 263, 264, 265, 268, 269, 281, 283, 284, 285, 286, 287, 291, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 352, 353, 354, 375, 376, 377], "bbox": [0.233203125, 0.07345833333333333, 0.783375, 0.9800625000000001], "iscrowd": 0, "area": 42105.48705, "rle": {"size": [480, 640], "counts": "WVV27d>>C7L3N1N3M3N2O1N2NlN]Bm0d=mN`BS1h=0O10O000O0100O00100O00100O001O1O001O001O1O001N2O00001O000O2O00001N1O1gCcNo:^1oDgNl:Z1SEhNk:Y1TEhNl:Y1SEhNk:Y1UEhNj:X1VEhNi:Z1VEgNg:[1XEgNe:Z1\\EfNb:\\1^EdN`:^1`EcN]:^1dEbNZ:`1fE`NY:a1fEaNX:_1iEaNV:_1kEaNT:_1mEbNP:_1QFaNn9_1SFaNl9_1TFbNh9`1ZF`N`9e1aF[NZ9h1hFWNS9m1oFSNk8Q2VGPNe8S2]GmM]8V2fGiMU8[2lGfMo7]2SHcMg7a2ZH`M`7d2aH\\MZ7h2hHXMS7k2nHVMl6n2UISMg6n2\\IQMc6o2^IRM`6n2aISM]6m2eISMY6m2hITMV6l2lITMR6l2oIUMo5m2oITMP6U3gImLW6\\3`IfL_6c3TIbLj6k5O1O10O1000O100000O10O1000000O10O10000`KYHg1g7XN\\Hf1c7YNaHe1_7ZNbHe1_7YNcHg1]7XNeHg1[7WNgHi1Y7TNjHl1V7QNlHT1GSM]7g1nHl03VMR7k1mHk0\\8QOfGl0^8QOeGl0^8QObGSOSOm0_9M_GXOUOd0a91YG^OYO:b97SGB^O0c9>mFDCHc9d0gFGI^Od9k0`FJOTOf9Q1XFN6iNf9Y1RF0;aNf9^1oE2n:NSE1n:NRE2n:NSE1m:OSE1_:aN\\E^160[:fN^EZ170Y:iN`EV180W:lN`ET190V:nN`ER1;OT:RO_Eo0=OQ:VObEi0>hN[OT1c:^OcEf0?fN]OV1^:CdEa0a0eN^OW1Z:GfE=g0Lc9HeE<h0Lc9IeE9j0M`9KfE8j0M`9KfE7k0N`9KdE5n00^9KeE3n02^9JdE3o03]9KcE0R15[9KcEOS16[9JbEOT17Z9JcELU1;W9IdEKV1<V9JcEIY1=S9JdEG[1?Q9JeEE[1a0Q9IdEE\\1b0P9IdED]1b0P9JcED^1a0o8LbEB`1a0o8MbEA_1b0o8LcEA`1a0n8McEB`1`0m8MdEBa1?l8NeEB`1?l8MeECa1>l8MdEEa1=k8NdEDc1<k8NdEEb1<k8MdEFc1;j8NdEGc1:j8NcEGe1:i8OaEGf19j80aEEg1:i80`EFh19h81`EEi19i82]EEk18S8AZFa0IDk1:R8EVF>LCm18Q8LSF9NBP28[7J]FL1=433BP28[74ZF0921AS27Y7i0cF@1_OT27X7j0dF_O0_OU28W7j0dF@O^OW26W7m0nFSODIX27X7k0aFnN16a20m6l0PGSOR2OP7n0nFSOS2MQ7o0lFUOR2JT7R1iFTOU<k0kCUOU<l0kCSOU<m0kCSOV<l0jCTOV<l0jCTOW<k0iCUOW<k0iCTOY<k0gCTOZ<m0eCRO\\<n0dCQO^<n0bCQO_<o0i001O000010O0001O0000_FTOm4n0QKSOn4n0RKROn4P1oJQOR5P1lJPOT5Q1lJnNU5S1jJmNV5T1hJlNY5T1gJkNZ5V1Y40O1O00O101N100NMnN\\Bn0b=UOiB?U=Fn05K4L5K4L5K5L3RNoNlES1o9A_Eb0[:EbE<k99RFIh9=UFEf9a0XFAf9`0YFBd9`0[FAd9`0ZFCd9>[FCd9>[FDc9=[FFc9;\\FFc9;\\FGa9;^FFb9:]FHb98^FI`99^FHb98^FI`98`FH`98`FI_98`FH_9:`FG_9;^FG`9<^FEa9=]FC]4dMd0[5WOhJj0Y5mNoJS1Q5dNWK\\1k4`NWKa1j4[NYKe1g4XN[Ki1e4TN]Km1d4nM`KR2`4kMbKU2_4gMdKZ2]4bMfK^2[4_MfKb2]4YMdKh2^4TMdKl2^4oLdKR3^4jLcKW3_4eLcK[3_4`LcKa3`4XLdKg3_4TLcKm3_4nKcKS4_4gKeKY4]4bKeK_4c4WK^Kj4d4RK]Ko4e4nJ[KS5g4jJXKX5j4dJWK\\5l4aJTK`5n4]JRKd5S72O2N2N3L3M3K5K5K5L4K5K5L4K5K5M3M3N2M3M3M3M4K5K5K5J6K5K5K5J7J6J5K6J6J6K5J6mL]E`1i:ZN[Eb1j:ZNZEa1l:ZNXEb1m:XNXEc1U<L7I7H9G9H7H=CaoP2"}, "label": "a barefoot woman holding a bottle of water"}]}
{"id": 372319, "image": "COCO_train2014_000000372319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a barefoot female in a pink tank top holding a water bottle\"."}], "task": "refering", "answer_template": "The \"a barefoot female in a pink tank top holding a water bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 63, 83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 199, 200, 201, 215, 216, 217, 218, 223, 224, 238, 239, 240, 245, 246, 260, 261, 262, 263, 264, 265, 268, 269, 281, 283, 284, 285, 286, 287, 291, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 352, 353, 354, 375, 376, 377], "bbox": [0.233203125, 0.07345833333333333, 0.783375, 0.9800625000000001], "iscrowd": 0, "area": 42105.48705, "rle": {"size": [480, 640], "counts": "WVV27d>>C7L3N1N3M3N2O1N2NlN]Bm0d=mN`BS1h=0O10O000O0100O00100O00100O001O1O001O001O1O001N2O00001O000O2O00001N1O1gCcNo:^1oDgNl:Z1SEhNk:Y1TEhNl:Y1SEhNk:Y1UEhNj:X1VEhNi:Z1VEgNg:[1XEgNe:Z1\\EfNb:\\1^EdN`:^1`EcN]:^1dEbNZ:`1fE`NY:a1fEaNX:_1iEaNV:_1kEaNT:_1mEbNP:_1QFaNn9_1SFaNl9_1TFbNh9`1ZF`N`9e1aF[NZ9h1hFWNS9m1oFSNk8Q2VGPNe8S2]GmM]8V2fGiMU8[2lGfMo7]2SHcMg7a2ZH`M`7d2aH\\MZ7h2hHXMS7k2nHVMl6n2UISMg6n2\\IQMc6o2^IRM`6n2aISM]6m2eISMY6m2hITMV6l2lITMR6l2oIUMo5m2oITMP6U3gImLW6\\3`IfL_6c3TIbLj6k5O1O10O1000O100000O10O1000000O10O10000`KYHg1g7XN\\Hf1c7YNaHe1_7ZNbHe1_7YNcHg1]7XNeHg1[7WNgHi1Y7TNjHl1V7QNlHT1GSM]7g1nHl03VMR7k1mHk0\\8QOfGl0^8QOeGl0^8QObGSOSOm0_9M_GXOUOd0a91YG^OYO:b97SGB^O0c9>mFDCHc9d0gFGI^Od9k0`FJOTOf9Q1XFN6iNf9Y1RF0;aNf9^1oE2n:NSE1n:NRE2n:NSE1m:OSE1_:aN\\E^160[:fN^EZ170Y:iN`EV180W:lN`ET190V:nN`ER1;OT:RO_Eo0=OQ:VObEi0>hN[OT1c:^OcEf0?fN]OV1^:CdEa0a0eN^OW1Z:GfE=g0Lc9HeE<h0Lc9IeE9j0M`9KfE8j0M`9KfE7k0N`9KdE5n00^9KeE3n02^9JdE3o03]9KcE0R15[9KcEOS16[9JbEOT17Z9JcELU1;W9IdEKV1<V9JcEIY1=S9JdEG[1?Q9JeEE[1a0Q9IdEE\\1b0P9IdED]1b0P9JcED^1a0o8LbEB`1a0o8MbEA_1b0o8LcEA`1a0n8McEB`1`0m8MdEBa1?l8NeEB`1?l8MeECa1>l8MdEEa1=k8NdEDc1<k8NdEEb1<k8MdEFc1;j8NdEGc1:j8NcEGe1:i8OaEGf19j80aEEg1:i80`EFh19h81`EEi19i82]EEk18S8AZFa0IDk1:R8EVF>LCm18Q8LSF9NBP28[7J]FL1=433BP28[74ZF0921AS27Y7i0cF@1_OT27X7j0dF_O0_OU28W7j0dF@O^OW26W7m0nFSODIX27X7k0aFnN16a20m6l0PGSOR2OP7n0nFSOS2MQ7o0lFUOR2JT7R1iFTOU<k0kCUOU<l0kCSOU<m0kCSOV<l0jCTOV<l0jCTOW<k0iCUOW<k0iCTOY<k0gCTOZ<m0eCRO\\<n0dCQO^<n0bCQO_<o0i001O000010O0001O0000_FTOm4n0QKSOn4n0RKROn4P1oJQOR5P1lJPOT5Q1lJnNU5S1jJmNV5T1hJlNY5T1gJkNZ5V1Y40O1O00O101N100NMnN\\Bn0b=UOiB?U=Fn05K4L5K4L5K5L3RNoNlES1o9A_Eb0[:EbE<k99RFIh9=UFEf9a0XFAf9`0YFBd9`0[FAd9`0ZFCd9>[FCd9>[FDc9=[FFc9;\\FFc9;\\FGa9;^FFb9:]FHb98^FI`99^FHb98^FI`98`FH`98`FI_98`FH_9:`FG_9;^FG`9<^FEa9=]FC]4dMd0[5WOhJj0Y5mNoJS1Q5dNWK\\1k4`NWKa1j4[NYKe1g4XN[Ki1e4TN]Km1d4nM`KR2`4kMbKU2_4gMdKZ2]4bMfK^2[4_MfKb2]4YMdKh2^4TMdKl2^4oLdKR3^4jLcKW3_4eLcK[3_4`LcKa3`4XLdKg3_4TLcKm3_4nKcKS4_4gKeKY4]4bKeK_4c4WK^Kj4d4RK]Ko4e4nJ[KS5g4jJXKX5j4dJWK\\5l4aJTK`5n4]JRKd5S72O2N2N3L3M3K5K5K5L4K5K5L4K5K5M3M3N2M3M3M3M4K5K5K5J6K5K5K5J7J6J5K6J6J6K5J6mL]E`1i:ZN[Eb1j:ZNZEa1l:ZNXEb1m:XNXEc1U<L7I7H9G9H7H=CaoP2"}, "label": "a barefoot female in a pink tank top holding a water bottle"}]}
{"id": 77408, "image": "COCO_train2014_000000077408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow that is in the middle on the pasture\"."}], "task": "refering", "answer_template": "The \"the cow that is in the middle on the pasture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 38, 39, 55, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 304, 305, 306, 307, 308, 309, 310, 311, 312, 327, 330, 331, 334, 335, 350, 353, 354, 357, 358, 373, 374, 376, 377, 380, 381], "bbox": [0.18103125, 0.05604166666666666, 0.85021875, 1.0], "iscrowd": 0, "area": 87406.45465000004, "rle": {"size": [480, 640], "counts": "Xcf13l>5K5K5J6J;F?@c0^Oo0PO3M4M2M4M2N2O2M2N3N1N2N3N1TEcLR:]4H9H7G9H6J4K6K4L5lH_I_5f6YJ`Ie5e6SJ`Ij5e6gIhIW6b7N2O1O2M2O1O1O1N2O1O1NgNnIgIQ6R6YJlIf5n5cJQJZ5i5PKUJo4c5\\K[Ja4`5iK^JT4\\5WLaJi3X5aLeJ^3T5mLiJS3U5QMiJo2U5UMiJj2W5ZMfJf2X5^MfJa2Z5bMdJ^2Z5gMcJY2\\5jMbJV2]5mMaJS2]5QNaJo1^5UN_Jk1_5l3O1N2O1N2000000O100001O2N2N2N1O2N2M2M3N2M3N3L3N2N2N1O1O1O1O1O1O1O1O0000001O000O2O5K6J6J6J6J6J5K6J6J6J6J6J5K5K5K5K5K5K5K5K5K5K5K4M4L3M3M000O10000000000OmKkL]LU3c3oLYLQ3g3RMVLn2j3VMiKS3X4oLZK\\3f4hLlJR1XN0m6QOcJS1dNIi6XO[JR1POCe6^OUJR1YO]Oc6FjIQ1EWOa6O`In01PO_6:UIj0>hN^6g3dIVL\\6k3eISL\\6l3gIQLY6P4hInKX6S4hIlKX6U4gIjKZ6W4fIhK[6X4eIgK\\6Y4cIgK`6W4`IhKc6V4]IiKf6T4ZIkKj6S4VIlKm6R4SImKQ7P4nHoKV7o3jHoKZ7o3fHPL^7m3aHRLc7k3^HSLg7j3YHTLl7i3SHVLn7k3RHSLo7n3QHQLo7o3QHPLP8Q4PHmKR8R4oGlKR8U4mGjKT8V4mGhKT8X4mGgKS8Z4lGeKU8[4lGcKU8]4lGaKU8`4kG^KV8b4jG]KW8c4jG[KW8e4jGXKX8h4iGVKY8h4iGVKX8j4o000O10000O10000O1000000O01000O10000O01000O10000O01000O2O4L5K5J5L5K5J5L5K5K4L5K5K4L5J6K2N1O1O2N1O1O2N1O1O2N1O1N2O00000000jJ<PKDo4a1mI_NS6d1jI]NT6h1hIXNW6l1fITNY6Q2cIoM]6T2`ImM^6X2^IhMa6\\2\\IdMa6c2[I^Ma6o2UIQMh6]3mHcLS7h3bHXL^7f5O00001O0000001O01O01O0000001N100O2O0O1O101N100O101O000O101O0O1000O01aMUHiMj7R2^HfMCTNn7Q4gHiM[OVNm7n2`H\\Mb0]1ROWNm7k2UJlNnMYNl7j2ZJkNjM[Nk7i2_JjNgM[Nk7j2bJhNcM^Nj7j2fJfN`M`Ni7j0WHR1c2A[5\\OUHP1d2AV5@XHo0c2^OT5C\\Hn0c2[OP5GaHm0`2YOn4JfHk0`2VOh40kHj0h2iN\\4=QIg0i:ZOaE?[:CeE>Y:BhE`0T:AmE`0P:AQF`0m9@TFb0h9_OYFb0d9_O]Fb0a9^O`Fc0^9^ObFc0\\9\\OfFe0W9ZOlFg0R9XOPGh0o8WOSGj0j8UOYGl0e8SO]Gn0a8QOaGP1]8nNfGS1Y8kNiGU1X8hNjGX1X8cNkG]1V8`NlGa1i:00000000001O000O1000001O000000000010O001O001O010O00001O001O0010O01O001O001O0010O01O001O001O00010O001O001O001O010O001O001O1O00100O001O001O1O001O1O0O2O1N2O1O1N2O1O1N2O1N2O1N2N2N2N2Onj\\1"}, "label": "the cow that is in the middle on the pasture"}]}
{"id": 77408, "image": "COCO_train2014_000000077408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow facing the camera\"."}], "task": "refering", "answer_template": "The \"cow facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 38, 39, 55, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 304, 305, 306, 307, 308, 309, 310, 311, 312, 327, 330, 331, 334, 335, 350, 353, 354, 357, 358, 373, 374, 376, 377, 380, 381], "bbox": [0.18103125, 0.05604166666666666, 0.85021875, 1.0], "iscrowd": 0, "area": 87406.45465000004, "rle": {"size": [480, 640], "counts": "Xcf13l>5K5K5J6J;F?@c0^Oo0PO3M4M2M4M2N2O2M2N3N1N2N3N1TEcLR:]4H9H7G9H6J4K6K4L5lH_I_5f6YJ`Ie5e6SJ`Ij5e6gIhIW6b7N2O1O2M2O1O1O1N2O1O1NgNnIgIQ6R6YJlIf5n5cJQJZ5i5PKUJo4c5\\K[Ja4`5iK^JT4\\5WLaJi3X5aLeJ^3T5mLiJS3U5QMiJo2U5UMiJj2W5ZMfJf2X5^MfJa2Z5bMdJ^2Z5gMcJY2\\5jMbJV2]5mMaJS2]5QNaJo1^5UN_Jk1_5l3O1N2O1N2000000O100001O2N2N2N1O2N2M2M3N2M3N3L3N2N2N1O1O1O1O1O1O1O1O0000001O000O2O5K6J6J6J6J6J5K6J6J6J6J6J5K5K5K5K5K5K5K5K5K5K5K4M4L3M3M000O10000000000OmKkL]LU3c3oLYLQ3g3RMVLn2j3VMiKS3X4oLZK\\3f4hLlJR1XN0m6QOcJS1dNIi6XO[JR1POCe6^OUJR1YO]Oc6FjIQ1EWOa6O`In01PO_6:UIj0>hN^6g3dIVL\\6k3eISL\\6l3gIQLY6P4hInKX6S4hIlKX6U4gIjKZ6W4fIhK[6X4eIgK\\6Y4cIgK`6W4`IhKc6V4]IiKf6T4ZIkKj6S4VIlKm6R4SImKQ7P4nHoKV7o3jHoKZ7o3fHPL^7m3aHRLc7k3^HSLg7j3YHTLl7i3SHVLn7k3RHSLo7n3QHQLo7o3QHPLP8Q4PHmKR8R4oGlKR8U4mGjKT8V4mGhKT8X4mGgKS8Z4lGeKU8[4lGcKU8]4lGaKU8`4kG^KV8b4jG]KW8c4jG[KW8e4jGXKX8h4iGVKY8h4iGVKX8j4o000O10000O10000O1000000O01000O10000O01000O10000O01000O2O4L5K5J5L5K5J5L5K5K4L5K5K4L5J6K2N1O1O2N1O1O2N1O1O2N1O1N2O00000000jJ<PKDo4a1mI_NS6d1jI]NT6h1hIXNW6l1fITNY6Q2cIoM]6T2`ImM^6X2^IhMa6\\2\\IdMa6c2[I^Ma6o2UIQMh6]3mHcLS7h3bHXL^7f5O00001O0000001O01O01O0000001N100O2O0O1O101N100O101O000O101O0O1000O01aMUHiMj7R2^HfMCTNn7Q4gHiM[OVNm7n2`H\\Mb0]1ROWNm7k2UJlNnMYNl7j2ZJkNjM[Nk7i2_JjNgM[Nk7j2bJhNcM^Nj7j2fJfN`M`Ni7j0WHR1c2A[5\\OUHP1d2AV5@XHo0c2^OT5C\\Hn0c2[OP5GaHm0`2YOn4JfHk0`2VOh40kHj0h2iN\\4=QIg0i:ZOaE?[:CeE>Y:BhE`0T:AmE`0P:AQF`0m9@TFb0h9_OYFb0d9_O]Fb0a9^O`Fc0^9^ObFc0\\9\\OfFe0W9ZOlFg0R9XOPGh0o8WOSGj0j8UOYGl0e8SO]Gn0a8QOaGP1]8nNfGS1Y8kNiGU1X8hNjGX1X8cNkG]1V8`NlGa1i:00000000001O000O1000001O000000000010O001O001O010O00001O001O0010O01O001O001O0010O01O001O001O00010O001O001O001O010O001O001O1O00100O001O001O1O001O1O0O2O1N2O1O1N2O1O1N2O1N2O1N2N2N2N2Onj\\1"}, "label": "cow facing the camera"}]}
{"id": 487009, "image": "COCO_train2014_000000487009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an animal smelling another animal\"."}], "task": "refering", "answer_template": "The \"an animal smelling another animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 283, 284], "bbox": [0.156390625, 0.2855, 0.43385937500000005, 0.7272083333333333], "iscrowd": 0, "area": 26705.744650000008, "rle": {"size": [480, 640], "counts": "kS_15i><E;E;E;E7I6I8I7I6J6J6J2M1000O10000O01O001O1O001O001O1O001O1O001O0011001N101N101N101N101O0O2O0O2O0O2O0J7G8H9F9L5N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2O001N101O001N2O1O001N2O1O001N2O1O001N3N2N2N2M3N1O0O10000000000000O10000000000000000000001O2N1O2N1O2N1O2N1O1O2N1O2N1N3N1O2N1O1O2N1O1O0000O1000000O1000O1N101O1N2O0O2O1N4^K\\Fc3g9ZL[Fd3g9YL]Fc3g9ZL\\Fc3g9YL\\Fb3j9[LYF^3m9_LVF\\3P:aLSF=Hl1W<I7I7H9H7I7I7H8I8H7Ig]Y5"}, "label": "an animal smelling another animal"}]}
{"id": 93793, "image": "COCO_train2014_000000093793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine glass sitting on red tablecloth next to plate of food\"."}], "task": "refering", "answer_template": "The \"wine glass sitting on red tablecloth next to plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 28, 29, 30, 31, 52, 53, 75, 76, 97, 98, 99, 100], "bbox": [0.21440625, 0.0001405152224824356, 0.4168125, 0.3125058548009368], "iscrowd": 0, "area": 7471.302249999998, "rle": {"size": [427, 640], "counts": "STi14W=2N2N2N2N2N2N2N2N2N2N001O1O1O1O1_F]OX6d0gI\\OY6e0eI\\O[6e0dI[O\\6f0bI[O^6f0aIZO_6g0`IYO`6h0^IYOb6h0]IXOc6i0[IVOg6k0XITOi6l0VISOl6n0SIROm6o0RIPOo6Q1PIoNP7R1oHnNQ7S1mHnNS7S1lHmNT7T1kHlNU7U1jHjNW7W1hHiNX7X1gHhNY7Y1fHgNZ7Z1eHfN[7[1cHfN]7[1bHdN_7]1`HcN`7^1_HbNa7_1^HaNb7`1]H`Nc7a1\\H_Nd7b1[H]Nf7d1XH]Nh7d1WH\\Ni7e1VH[Nj7f1UHYNl7h1RHYNn7h1QHXNo7i1oGXNQ8U2_GmMb8g2eF^M[9Y3000O1000000O1000000O10000N2N2M3N2N2O11O1O1O1O1O1O2N1OfNWGSNh8h1^GXNa8c1fG\\NY8^1nGaNR8[1THbNm7]1UHbNk7]1XHaNh7^1ZHaNf7^1]H_Nd7`1T2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1N2N2N2N2N2N2N2NYak4"}, "label": "wine glass sitting on red tablecloth next to plate of food"}]}
{"id": 93793, "image": "COCO_train2014_000000093793.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an obscured wine glass\"."}], "task": "refering", "answer_template": "The \"an obscured wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 28, 29, 30, 31, 52, 53, 75, 76, 97, 98, 99, 100], "bbox": [0.21440625, 0.0001405152224824356, 0.4168125, 0.3125058548009368], "iscrowd": 0, "area": 7471.302249999998, "rle": {"size": [427, 640], "counts": "STi14W=2N2N2N2N2N2N2N2N2N2N001O1O1O1O1_F]OX6d0gI\\OY6e0eI\\O[6e0dI[O\\6f0bI[O^6f0aIZO_6g0`IYO`6h0^IYOb6h0]IXOc6i0[IVOg6k0XITOi6l0VISOl6n0SIROm6o0RIPOo6Q1PIoNP7R1oHnNQ7S1mHnNS7S1lHmNT7T1kHlNU7U1jHjNW7W1hHiNX7X1gHhNY7Y1fHgNZ7Z1eHfN[7[1cHfN]7[1bHdN_7]1`HcN`7^1_HbNa7_1^HaNb7`1]H`Nc7a1\\H_Nd7b1[H]Nf7d1XH]Nh7d1WH\\Ni7e1VH[Nj7f1UHYNl7h1RHYNn7h1QHXNo7i1oGXNQ8U2_GmMb8g2eF^M[9Y3000O1000000O1000000O10000N2N2M3N2N2O11O1O1O1O1O1O2N1OfNWGSNh8h1^GXNa8c1fG\\NY8^1nGaNR8[1THbNm7]1UHbNk7]1XHaNh7^1ZHaNf7^1]H_Nd7`1T2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1N2N2N2N2N2N2N2NYak4"}, "label": "an obscured wine glass"}]}
{"id": 216676, "image": "COCO_train2014_000000216676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white surfboard leaning on a grey brick wall\"."}], "task": "refering", "answer_template": "The \"a white surfboard leaning on a grey brick wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 299, 300], "bbox": [0.0, 0.20260416666666667, 0.095453125, 0.8207708333333333], "iscrowd": 0, "area": 11836.6928, "rle": {"size": [480, 640], "counts": "R3f0[>]1dN\\1cN]1dN\\1cN]1dNo0PO0010O001O010O001O0O1O2N1O2M2O2N3M4L3M3M4L3L5K4L4L5K4L4L5K4L5K4L4L5L3L6J5K6J5J6I8I6I8H7J6I8I6I8H7J6IW1jN8G9G9]OcS_8"}, "label": "a white surfboard leaning on a grey brick wall"}]}
{"id": 216676, "image": "COCO_train2014_000000216676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a standing surf board\"."}], "task": "refering", "answer_template": "The \"a standing surf board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 299, 300], "bbox": [0.0, 0.20260416666666667, 0.095453125, 0.8207708333333333], "iscrowd": 0, "area": 11836.6928, "rle": {"size": [480, 640], "counts": "R3f0[>]1dN\\1cN]1dN\\1cN]1dNo0PO0010O001O010O001O0O1O2N1O2M2O2N3M4L3M3M4L3L5K4L4L5K4L4L5K4L5K4L4L5L3L6J5K6J5J6I8I6I8H7J6I8I6I8H7J6IW1jN8G9G9]OcS_8"}, "label": "a standing surf board"}]}
{"id": 577126, "image": "COCO_train2014_000000577126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy on the right\"."}], "task": "refering", "answer_template": "The \"the boy on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342], "bbox": [0.58078125, 0.31208920187793426, 0.895296875, 0.9911971830985915], "iscrowd": 0, "area": 36502.2283, "rle": {"size": [426, 640], "counts": "UQk43U=7J7I6J6J7H7J6J4L5K4L4M4K4L5L3L4L5L3M4L3M4L3M00100O1O001O1O001O1O10O011O0N3N1O5J8I3M2M3N\\L\\MdLb2Z3aMgL\\2X3gMiLW2T3mMmLQ2Q3RNnLn1P3dN`L[1^3VOdI_NW2[2S4JdK5[4OaK2]43_KN`46[KMd45YKLg46VKKj47SKJn48mJKR57kJJU58gJJZ57bJK;WMf3P3kKLCPNa4V2hK3_OiMh4U2hKZ1X4gNeK[1Z4fNeK[1Z4gNcK[1\\4fNcK[1]4fNaK[1^4fNaK[1^4gN`KZ1`4fN`KZ1`4gN^KZ1a4gN^KZ1b4fN^KZ1a4gN^KZ1b4fN]K[1c4eN]K[1b4fN]K[1c4eN]K[1c4f3O100O10000O2O000O1000000000000O10000000000O1000000000000O10000001O1O010O001O1O001O1O001O1O001O001O1WJZKl2f4SM\\Kl2e4RM]Km2d4PM_Ko2a4PMbKn2_4PMcKo2^4PMcKo2]4PMeKo2\\4oLgKo2Y4PMiKo2X4PMiKo2X4oLjKP3V4oLlKP3V4mLmKR3S4mLnKR3T4kLnKT3S4jLoKU3S4hLPLV3R4hLoKW3R4gLQLW3Q4fLQLY3Q4dLRLZ3P4bLSL]3o3`LTL^3n3_LUL_3n3]LTLb3n3ZLVLd3l3YLVLf3l3WLWLg3R4nKQLQ4\\4aKfK^4e6N1O1O1O2N1O1O1O2N1O1O2O0O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O8H5K5K4L5K8]OnD^NZ;U1oDgN[;j0mDTOa;8h`k0"}, "label": "the boy on the right"}]}
{"id": 577126, "image": "COCO_train2014_000000577126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a superman t - shirt\"."}], "task": "refering", "answer_template": "The \"a boy in a superman t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 264, 265, 266, 287], "bbox": [0.48440625, 0.2631455399061033, 0.6623749999999999, 0.8256103286384977], "iscrowd": 0, "area": 15265.159850000004, "rle": {"size": [426, 640], "counts": "jYQ42T=4K5L4oIEoN?o0IkN7S12gNMX1:bNE]1b0QJoNg33hK2_6P1jIUOj3HoK2\\6T1gIVOm3BRL2Z6X1dIYOg2\\OcN2Q5W2WLkMgNNQ5Z2WKWNbNEU1IQ5^2oJnNOcNR5b2lJnN_6U1]InNb6S1\\IoNb6T1[InNc6U1YInNe6W1UIkNj6X1QIlNm6V1PIlNo6V1mHmNR7T1nHmNP7U1nHlNP7U1PIlNo6U1PIlNo6U1QIlNm6U1RIlNm6U1RIlNn6T1QImNo6S1QImNo6S1PIoNo6Q1QIoNn6Q1RIPOO`Mj6`3WIPOMcMk6\\3XIXOh6h0XIVOj6i0VIVOk6j0VIVOj6j0VIVOj6i0WIXOh6e3O10001O00001O0000001O000010O01O0000SOYIWKg6c4dIXK\\6e4PJRKQ6k4YJmJg5Q5Z1O2N100O2N101N2O3L4L5L3L4M2M3N2M2O3L<E1N1[NVFJk94WFKi94XFLi92XFNh91ZFMh91ZFNf91]FMc92_FMa92bFL_92cFM]93eFKZ95hFJX92nFLR92QGMo81TGOoNoNg9Q1]G5a8IbG8\\8GgG?Q8_OSHd0h7POfHS1V7iNoHX1n6fNVI[1g6cN]Ih0V7TOPIe0o9J6J6I7Jdci2"}, "label": "a boy in a superman t - shirt"}]}
{"id": 577126, "image": "COCO_train2014_000000577126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy wearing a white super - man t - shirt\"."}], "task": "refering", "answer_template": "The \"a little boy wearing a white super - man t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 264, 265, 266, 287], "bbox": [0.48440625, 0.2631455399061033, 0.6623749999999999, 0.8256103286384977], "iscrowd": 0, "area": 15265.159850000004, "rle": {"size": [426, 640], "counts": "jYQ42T=4K5L4oIEoN?o0IkN7S12gNMX1:bNE]1b0QJoNg33hK2_6P1jIUOj3HoK2\\6T1gIVOm3BRL2Z6X1dIYOg2\\OcN2Q5W2WLkMgNNQ5Z2WKWNbNEU1IQ5^2oJnNOcNR5b2lJnN_6U1]InNb6S1\\IoNb6T1[InNc6U1YInNe6W1UIkNj6X1QIlNm6V1PIlNo6V1mHmNR7T1nHmNP7U1nHlNP7U1PIlNo6U1PIlNo6U1QIlNm6U1RIlNm6U1RIlNn6T1QImNo6S1QImNo6S1PIoNo6Q1QIoNn6Q1RIPOO`Mj6`3WIPOMcMk6\\3XIXOh6h0XIVOj6i0VIVOk6j0VIVOj6j0VIVOj6i0WIXOh6e3O10001O00001O0000001O000010O01O0000SOYIWKg6c4dIXK\\6e4PJRKQ6k4YJmJg5Q5Z1O2N100O2N101N2O3L4L5L3L4M2M3N2M2O3L<E1N1[NVFJk94WFKi94XFLi92XFNh91ZFMh91ZFNf91]FMc92_FMa92bFL_92cFM]93eFKZ95hFJX92nFLR92QGMo81TGOoNoNg9Q1]G5a8IbG8\\8GgG?Q8_OSHd0h7POfHS1V7iNoHX1n6fNVI[1g6cN]Ih0V7TOPIe0o9J6J6I7Jdci2"}, "label": "a little boy wearing a white super - man t - shirt"}]}
{"id": 577126, "image": "COCO_train2014_000000577126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in glasses sits with two young boys\"."}], "task": "refering", "answer_template": "The \"the man in glasses sits with two young boys\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110, 129, 130, 132, 133, 134, 135, 156, 157, 158, 179, 180, 181, 193, 194, 202, 203, 204, 216, 217, 218, 225, 226, 227, 228, 239, 240, 241, 242, 249, 250, 251, 263, 264, 265, 266, 273, 274, 285, 286, 287, 288, 289, 296, 297, 307, 308, 309, 310, 311, 312, 313, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.347015625, 0.029366197183098592, 0.94234375, 0.9754225352112675], "iscrowd": 0, "area": 46319.76094999998, "rle": {"size": [426, 640], "counts": "[Um21X=2N1O100O1O2N100O1O1O2N100O1O2N100O1O1O2N100O1O2N1O100O1hK@eKb0X42UKOi4e0cJ]O[5g0bJZO\\5i0aJYO]5j0`JYO^5h0aJYO_5h0^JZOc5f0[J[Oe5e0YJ^Og5b0WJ_Oi5a0VJ@k5`0SJAm5`0QJAP6>PJCo5>oICR6=lIDT6<kIEV6;hIFX6;gIEZ6:eIG\\69bIH^68aII`67_IIa68]IId66[IKe66YIKh65VILj64VILk64SIMm64RILo63PINP73nHNS71mHOS72kHOV71iHOW71hH0Y70eH1[70dH0_7M`H4c7J\\H6h7GVH:l7DTH<m7DQH=o7CQH=P8CoG=R8CnG<S8CmG=S8DlG<U8DjG<W8CjG<W8DhG<X8EgG;Z8DfG<[8DeG;\\8DdG<\\8EcG;^8EbG:_8EaG;`8E_G;a8F^G:c8E^G:c8F\\G:e8FZG:f8F[G9f8GYG9h8GWG9i8GWG9j8GVG8k8GUG9l8GSG9m8HSG7n8HRG8o8HPG8Q9HnF8R9HnF8S9HkF9V9GiF9X9FgF;Y9FfF:[9EdF<]9DbF<_9g1001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O00cHZMW4e2aKfM]4Y2\\KQNb4n1VK]Ng4c1RKgNl4X1nJQOQ5m0lJXOU5e0hJAX5<fJH\\54aJ1`5L]J:d5BYJc0h5ZOUJk0m5QOQJS1P6jNoIZ1Q6cNnI`1T6\\NlIf1U6WNjIm1W6oMiIS2X6jMgIY2X6fMhI\\2W6cMiI`2U6_MjId2U6[MkIg2S6YMlIk2R6TMnIm2R6RMmIP3S6oLmIQ3T6nLkIT3T6lLlIT3U6kLkIU3V6jLiIX3W6gLiIY3X6fLgI\\3X6dLhI\\3Y6cLfI^3[6aLeI`3[6_LdIb3\\6^LdIc3\\6\\LdId3]6[LcIe3^6ZLbIg3^6XLbIh3^6XLbIi3^6VLbIj3_6ULaIk3`6TL`Im3`6RL`In3`6RL`Io3`6PL`IP4a6oK_IQ4b6nK^IS4b6lK^IT4f6hKZIY4h6dKXI\\4k6aKUI`4m6]KRId4R7XKnHh4j7N2M3M3M3N2M3M3M3N2000001O000000001O0000001O0O10001O0000001O0000001O00001O00001O001O1O1O1O001O1O1O1O001O1O1O100O0100000O2O4L5J6K3M1O1N3N1OkG^Mc5`2[JcMe5[2ZJgMh5W2VJlMj5S2TJoMm5P2QJRNQ6k1oIWNQ6h1mIZNV6c1hI`NZ6]1dIeN_6W1`IkNc61aG1j11f6EfG;b12k6ZOiGd0Z14f7LZH6d7KZH7f7IXH:f7GXH;h7DWH>h7CVH`0i7@UHb0j7_OTHd0k7[OUHf0j7[OTHg0l7YORHj0l7WORHk0m7UORHn0m7XOkGj0T8^OcGd0]8B[Ga0c8GUG:k8S2O2O0O2N100O2O0O2O0O101N100O2N101N1O101N1O2N1O1O2N1O1O2N1O2N1lHbKZ5W61O1O2N1O1O1O2N1O1N2O2N1O1O2N4L3M4L4L4L4L3M4L4UKkHY3e8N2M3N2N2N2N1O2N2N2N2N2N2N2N2N2N2M3I7]OUT?"}, "label": "the man in glasses sits with two young boys"}]}
{"id": 577126, "image": "COCO_train2014_000000577126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man seated between two children holding a hotdog\"."}], "task": "refering", "answer_template": "The \"a man seated between two children holding a hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110, 129, 130, 132, 133, 134, 135, 156, 157, 158, 179, 180, 181, 193, 194, 202, 203, 204, 216, 217, 218, 225, 226, 227, 228, 239, 240, 241, 242, 249, 250, 251, 263, 264, 265, 266, 273, 274, 285, 286, 287, 288, 289, 296, 297, 307, 308, 309, 310, 311, 312, 313, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.347015625, 0.029366197183098592, 0.94234375, 0.9754225352112675], "iscrowd": 0, "area": 46319.76094999998, "rle": {"size": [426, 640], "counts": "[Um21X=2N1O100O1O2N100O1O1O2N100O1O2N100O1O1O2N100O1O2N1O100O1hK@eKb0X42UKOi4e0cJ]O[5g0bJZO\\5i0aJYO]5j0`JYO^5h0aJYO_5h0^JZOc5f0[J[Oe5e0YJ^Og5b0WJ_Oi5a0VJ@k5`0SJAm5`0QJAP6>PJCo5>oICR6=lIDT6<kIEV6;hIFX6;gIEZ6:eIG\\69bIH^68aII`67_IIa68]IId66[IKe66YIKh65VILj64VILk64SIMm64RILo63PINP73nHNS71mHOS72kHOV71iHOW71hH0Y70eH1[70dH0_7M`H4c7J\\H6h7GVH:l7DTH<m7DQH=o7CQH=P8CoG=R8CnG<S8CmG=S8DlG<U8DjG<W8CjG<W8DhG<X8EgG;Z8DfG<[8DeG;\\8DdG<\\8EcG;^8EbG:_8EaG;`8E_G;a8F^G:c8E^G:c8F\\G:e8FZG:f8F[G9f8GYG9h8GWG9i8GWG9j8GVG8k8GUG9l8GSG9m8HSG7n8HRG8o8HPG8Q9HnF8R9HnF8S9HkF9V9GiF9X9FgF;Y9FfF:[9EdF<]9DbF<_9g1001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O00cHZMW4e2aKfM]4Y2\\KQNb4n1VK]Ng4c1RKgNl4X1nJQOQ5m0lJXOU5e0hJAX5<fJH\\54aJ1`5L]J:d5BYJc0h5ZOUJk0m5QOQJS1P6jNoIZ1Q6cNnI`1T6\\NlIf1U6WNjIm1W6oMiIS2X6jMgIY2X6fMhI\\2W6cMiI`2U6_MjId2U6[MkIg2S6YMlIk2R6TMnIm2R6RMmIP3S6oLmIQ3T6nLkIT3T6lLlIT3U6kLkIU3V6jLiIX3W6gLiIY3X6fLgI\\3X6dLhI\\3Y6cLfI^3[6aLeI`3[6_LdIb3\\6^LdIc3\\6\\LdId3]6[LcIe3^6ZLbIg3^6XLbIh3^6XLbIi3^6VLbIj3_6ULaIk3`6TL`Im3`6RL`In3`6RL`Io3`6PL`IP4a6oK_IQ4b6nK^IS4b6lK^IT4f6hKZIY4h6dKXI\\4k6aKUI`4m6]KRId4R7XKnHh4j7N2M3M3M3N2M3M3M3N2000001O000000001O0000001O0O10001O0000001O0000001O00001O00001O001O1O1O1O001O1O1O1O001O1O1O100O0100000O2O4L5J6K3M1O1N3N1OkG^Mc5`2[JcMe5[2ZJgMh5W2VJlMj5S2TJoMm5P2QJRNQ6k1oIWNQ6h1mIZNV6c1hI`NZ6]1dIeN_6W1`IkNc61aG1j11f6EfG;b12k6ZOiGd0Z14f7LZH6d7KZH7f7IXH:f7GXH;h7DWH>h7CVH`0i7@UHb0j7_OTHd0k7[OUHf0j7[OTHg0l7YORHj0l7WORHk0m7UORHn0m7XOkGj0T8^OcGd0]8B[Ga0c8GUG:k8S2O2O0O2N100O2O0O2O0O101N100O2N101N1O101N1O2N1O1O2N1O1O2N1O2N1lHbKZ5W61O1O2N1O1O1O2N1O1N2O2N1O1O2N4L3M4L4L4L4L3M4L4UKkHY3e8N2M3N2N2N2N1O2N2N2N2N2N2N2N2N2N2M3I7]OUT?"}, "label": "a man seated between two children holding a hotdog"}]}
{"id": 192104, "image": "COCO_train2014_000000192104.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown horse in the foreground\"."}], "task": "refering", "answer_template": "The \"a light brown horse in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 354], "bbox": [0.16990624999999998, 0.5426249999999999, 0.44565625, 0.9014166666666665], "iscrowd": 0, "area": 16421.7961, "rle": {"size": [480, 640], "counts": "e\\c13l>3L5L3L2N2N2TOCSC>m<j0O2O00001O010O000010O002O0O1O101N1O1O100O2N100001O010O2N1O2O0O2N1O5L7H2O1N2N2O1N3M2O2M2N2O2M2O1N2N3N2M2N6J6I8lDcLW:]4F0000001O00000000O011O1O001O00000000000O1kM`EJ`:[20lM`EGa:9hE\\OZ:d0SFgNU:Z1h1000000001O0000001O0000001O0000001O000000001O01O0002N1O1O1O1O1O2VEXMY9h2^FbM`9`2]FbMb9`2[FcMc9_2ZFcMe9^2ZFdMd9^2YFdMf9]2XFeMg9\\2WFdMk9]2SFcMm9^2QFcMo9a2lE`MT:d2gE]MY:g2cEYM]:e31O0N2NaEhKZ:V4dEoK[:n3eEVLZ:h3eE\\L[:`3eEeLY:R400O100OL4K6K4N3M2O2N1O2M2O2O02O2M2O1N2OO2L3N3M2M3N3M4WOj0F:G9VOj0B>^O^SV5"}, "label": "a light brown horse in the foreground"}]}
{"id": 192104, "image": "COCO_train2014_000000192104.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chestnut colored horse with a khaki saddle and maroon colored blankets\"."}], "task": "refering", "answer_template": "The \"a chestnut colored horse with a khaki saddle and maroon colored blankets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 354], "bbox": [0.16990624999999998, 0.5426249999999999, 0.44565625, 0.9014166666666665], "iscrowd": 0, "area": 16421.7961, "rle": {"size": [480, 640], "counts": "e\\c13l>3L5L3L2N2N2TOCSC>m<j0O2O00001O010O000010O002O0O1O101N1O1O100O2N100001O010O2N1O2O0O2N1O5L7H2O1N2N2O1N3M2O2M2N2O2M2O1N2N3N2M2N6J6I8lDcLW:]4F0000001O00000000O011O1O001O00000000000O1kM`EJ`:[20lM`EGa:9hE\\OZ:d0SFgNU:Z1h1000000001O0000001O0000001O0000001O000000001O01O0002N1O1O1O1O1O2VEXMY9h2^FbM`9`2]FbMb9`2[FcMc9_2ZFcMe9^2ZFdMd9^2YFdMf9]2XFeMg9\\2WFdMk9]2SFcMm9^2QFcMo9a2lE`MT:d2gE]MY:g2cEYM]:e31O0N2NaEhKZ:V4dEoK[:n3eEVLZ:h3eE\\L[:`3eEeLY:R400O100OL4K6K4N3M2O2N1O2M2O2O02O2M2O1N2OO2L3N3M2M3N3M4WOj0F:G9VOj0B>^O^SV5"}, "label": "a chestnut colored horse with a khaki saddle and maroon colored blankets"}]}
{"id": 233071, "image": "COCO_train2014_000000233071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow union pacific freight train engine with lights on pulling blue cargo box car\"."}], "task": "refering", "answer_template": "The \"yellow union pacific freight train engine with lights on pulling blue cargo box car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 29, 30, 31, 32, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355], "bbox": [0.228859375, 0.0, 0.533453125, 0.9460566448801743], "iscrowd": 0, "area": 63405.13234999999, "rle": {"size": [459, 640], "counts": "SdQ21Y>;E<D6J5K4L5K5K5K1O00O100O10YG8l2GaJo2]5R3M2N2_Oa0UOk0XOh0^Ob0K5L4L5J5K5F:G9G9G9G8G9H8H8O100000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000h1XN2N1O1O2M2O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O:F<D5K2N1O2N1O1O2N1O2N1O2N1O1O2mF[K\\8]5M4L4L4L3MO1O10O0000001O0000001O00002N2N2N3M2N2jHgKW4\\4hKlMn1V2RNoM^1]2`NhMg0m2YOWMN^32_5N10001O0O2O00002MnjU4"}, "label": "yellow union pacific freight train engine with lights on pulling blue cargo box car"}]}
{"id": 233071, "image": "COCO_train2014_000000233071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow train with bird on the front\"."}], "task": "refering", "answer_template": "The \"yellow train with bird on the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 29, 30, 31, 32, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355], "bbox": [0.228859375, 0.0, 0.533453125, 0.9460566448801743], "iscrowd": 0, "area": 63405.13234999999, "rle": {"size": [459, 640], "counts": "SdQ21Y>;E<D6J5K4L5K5K5K1O00O100O10YG8l2GaJo2]5R3M2N2_Oa0UOk0XOh0^Ob0K5L4L5J5K5F:G9G9G9G8G9H8H8O100000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000h1XN2N1O1O2M2O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O:F<D5K2N1O2N1O1O2N1O2N1O2N1O1O2mF[K\\8]5M4L4L4L3MO1O10O0000001O0000001O00002N2N2N3M2N2jHgKW4\\4hKlMn1V2RNoM^1]2`NhMg0m2YOWMN^32_5N10001O0O2O00002MnjU4"}, "label": "yellow train with bird on the front"}]}
{"id": 233071, "image": "COCO_train2014_000000233071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train that is not in the lead\"."}], "task": "refering", "answer_template": "The \"the train that is not in the lead\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 293], "bbox": [0.423859375, 0.0022440087145969498, 0.82678125, 0.7550544662309367], "iscrowd": 0, "area": 54858.22974999999, "rle": {"size": [459, 640], "counts": "V_i33X>8H9G9G8H9G2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M3M3M3M3M3M3M4L3M3M3M3M3M3M2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3kGRJg7_6J6J6J6J6J6J2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2000001O1O1O2N1O1N2N2N2N2N3M2N2N2N2N2N0O01O100O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O100O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O100O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O010O1O0L5L9F<E;D<D<`MiFYOc90nFI`9@PG:]9oNSGj0\\;J6I7I7J7HX_a1"}, "label": "the train that is not in the lead"}]}
{"id": 233071, "image": "COCO_train2014_000000233071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train trailing behind the other one\"."}], "task": "refering", "answer_template": "The \"the train trailing behind the other one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 293], "bbox": [0.423859375, 0.0022440087145969498, 0.82678125, 0.7550544662309367], "iscrowd": 0, "area": 54858.22974999999, "rle": {"size": [459, 640], "counts": "V_i33X>8H9G9G8H9G2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M3M3M3M3M3M3M4L3M3M3M3M3M3M2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3kGRJg7_6J6J6J6J6J6J2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2000001O1O1O2N1O1N2N2N2N2N3M2N2N2N2N2N0O01O100O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O100O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O100O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O010O1O0L5L9F<E;D<D<`MiFYOc90nFI`9@PG:]9oNSGj0\\;J6I7I7J7HX_a1"}, "label": "the train trailing behind the other one"}]}
{"id": 478833, "image": "COCO_train2014_000000478833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half of an egg salad sandwich which has a long piece of onion sticking out from it\"."}], "task": "refering", "answer_template": "The \"a half of an egg salad sandwich which has a long piece of onion sticking out from it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 72, 73, 74, 75, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.0019375, 0.15975, 0.40996875, 0.8514583333333333], "iscrowd": 0, "area": 53581.256349999996, "rle": {"size": [480, 640], "counts": "[e07h>h0XOh0XO<B9F:F:E:G:G9I7J5J1O1O101N1O1O100O1O101N1O1O100O1O101N1O1O100O1O1O2O0O1O100O1O1O2N100O1O1O1O101N1O1O1O100O1O2N1O100O1O1O101N2N2N3N1N2N2N2O1N2N2O1N2N2N2O1N1O2O0O2N1O2O0O2N100O2N1O2O0O2N101N1O2N100O2N101N1O2N101N1O101N1O2N101N1O2O1N2N2N2O1N1O2O1N2N2N2O0O2N2O1N2N2N101N2N2O1N2N1O2O1N2N2O1N10100O100O1000O010000O1000O010000O1001N101O1O1O001N2O1O2N2N2N2M3M3M3N2M3M3M3M3M3M3N2M3M2N3M2N2N3N1N2N3M2N2N2N3N1N2N3L3N2N3L3N2N3L3N2N3L3N2M4M2N2M3N3M2M3N3M2M3N3M2M3N3M2M3N3M4K4M3M3L4M4L3L4M3M4K4M3M3L4M4L3L4M3L4M4L3L4M4L5J6K6J5J6Koj`5"}, "label": "a half of an egg salad sandwich which has a long piece of onion sticking out from it"}]}
{"id": 478833, "image": "COCO_train2014_000000478833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toasted sandwich slice with white sprouts and purple onion sticking out of it\"."}], "task": "refering", "answer_template": "The \"a toasted sandwich slice with white sprouts and purple onion sticking out of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 72, 73, 74, 75, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.0019375, 0.15975, 0.40996875, 0.8514583333333333], "iscrowd": 0, "area": 53581.256349999996, "rle": {"size": [480, 640], "counts": "[e07h>h0XOh0XO<B9F:F:E:G:G9I7J5J1O1O101N1O1O100O1O101N1O1O100O1O101N1O1O100O1O1O2O0O1O100O1O1O2N100O1O1O1O101N1O1O1O100O1O2N1O100O1O1O101N2N2N3N1N2N2N2O1N2N2O1N2N2N2O1N1O2O0O2N1O2O0O2N100O2N1O2O0O2N101N1O2N100O2N101N1O2N101N1O101N1O2N101N1O2O1N2N2N2O1N1O2O1N2N2N2O0O2N2O1N2N2N101N2N2O1N2N1O2O1N2N2O1N10100O100O1000O010000O1000O010000O1001N101O1O1O001N2O1O2N2N2N2M3M3M3N2M3M3M3M3M3M3N2M3M2N3M2N2N3N1N2N3M2N2N2N3N1N2N3L3N2N3L3N2N3L3N2N3L3N2M4M2N2M3N3M2M3N3M2M3N3M2M3N3M2M3N3M4K4M3M3L4M4L3L4M3M4K4M3M3L4M4L3L4M3L4M4L3L4M4L5J6K6J5J6Koj`5"}, "label": "a toasted sandwich slice with white sprouts and purple onion sticking out of it"}]}
{"id": 478833, "image": "COCO_train2014_000000478833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a panini with no purple onion\"."}], "task": "refering", "answer_template": "The \"a panini with no purple onion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 333, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 383], "bbox": [0.342140625, 0.22658333333333333, 0.9353906249999999, 0.9501875], "iscrowd": 0, "area": 83807.43419999999, "rle": {"size": [480, 640], "counts": "hjV34g>8I8H7M4L3M4K4M4L3M4K4M3M4PNYN_Fj1^9hNoE]1m9VO`Em0^:n1L3M4L3N2M4L3M4M2M4L3M4L3M3N3L3M4L3M2N2N2N2O1N2N2N2N2N2N2N101N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N1O2N2O1N2N2O1N2O1N2O1N2O0O01O01O01O010O010O0010O01O010O01O01O010O0010O010O01O01O00001O001O01O01O00001O0010O0001O00001O0010O0001O00001O010O00001O0XJ`GP5`8lJeGS5\\8hJhGX5X8dJmG\\5S8_JQHa5h8O1O001O10O0000001N100000001O00000O2O0000001O01O0000010O0000010O00000010O0001O10O01O00100O001O00100O001O010O1O001O01O0O10001O000O101O000O10001O00000O2O000000001O00000O2O0000001O000O100O2O0O100O2N100O100O2O0O1O1\\OkI]HV6_7nI_HS6]7SJaHm5\\7XJaHi5[7]JcHd5\\7^JaHc5^7`J_Ha5a7o0O101N1O100O2O0O100O2N101N101N100O2N101O0O101O001O000O2O001O001O0O101O001O00001N101O00001O0O2O001O00001O0O2O0nLcGA]89iGGX82nGNR80PH0Q8OoG0R80nG0R80nG0S8OmG1S8OmG1T8NlG2T8NlG1U8OkG1V8NjG2V8NjG1X8MiG2X8NhG2Y8MgGdNIGa8e1fGdNMB_8i1dGeN1]O\\8n1cGeN5YOY8Q2bGfN8UOW8U2aGfN8TOY8U2`GfN8SOZ8V2^GgN8SOZ8U2_GhN8QO[8V2]GhN9QO[8W2\\GgN:QO\\8W2ZGgN<PO[8X2ZGfN=ROZ8W2YGfN?QOY8Y2XGeN`0QOZ8Y2VGdNV:[1kEdNW:[1iEdNX:\\1hEbN[:]1eEbN\\:]1eEaN^:^1bEaN`:^1`EaNa:_1_E_Nd:_1]E`Nd:`1]E^Ne:a1]E[Nf:c1\\E[Ne:e1]EXNe:g1]EUNf:j1[ETNg:j1Y1O1O1O1N2O1O1O1N2N2N2N2N2N1O3M2N3M3M3M2M4M3_ORPc0"}, "label": "a panini with no purple onion"}]}
{"id": 478833, "image": "COCO_train2014_000000478833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich placed in the plate in middle\"."}], "task": "refering", "answer_template": "The \"a sandwich placed in the plate in middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 333, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 383], "bbox": [0.342140625, 0.22658333333333333, 0.9353906249999999, 0.9501875], "iscrowd": 0, "area": 83807.43419999999, "rle": {"size": [480, 640], "counts": "hjV34g>8I8H7M4L3M4K4M4L3M4K4M3M4PNYN_Fj1^9hNoE]1m9VO`Em0^:n1L3M4L3N2M4L3M4M2M4L3M4L3M3N3L3M4L3M2N2N2N2O1N2N2N2N2N2N2N101N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N1O2N2O1N2N2O1N2O1N2O1N2O0O01O01O01O010O010O0010O01O010O01O01O010O0010O010O01O01O00001O001O01O01O00001O0010O0001O00001O0010O0001O00001O010O00001O0XJ`GP5`8lJeGS5\\8hJhGX5X8dJmG\\5S8_JQHa5h8O1O001O10O0000001N100000001O00000O2O0000001O01O0000010O0000010O00000010O0001O10O01O00100O001O00100O001O010O1O001O01O0O10001O000O101O000O10001O00000O2O000000001O00000O2O0000001O000O100O2O0O100O2N100O100O2O0O1O1\\OkI]HV6_7nI_HS6]7SJaHm5\\7XJaHi5[7]JcHd5\\7^JaHc5^7`J_Ha5a7o0O101N1O100O2O0O100O2N101N101N100O2N101O0O101O001O000O2O001O001O0O101O001O00001N101O00001O0O2O001O00001O0O2O0nLcGA]89iGGX82nGNR80PH0Q8OoG0R80nG0R80nG0S8OmG1S8OmG1T8NlG2T8NlG1U8OkG1V8NjG2V8NjG1X8MiG2X8NhG2Y8MgGdNIGa8e1fGdNMB_8i1dGeN1]O\\8n1cGeN5YOY8Q2bGfN8UOW8U2aGfN8TOY8U2`GfN8SOZ8V2^GgN8SOZ8U2_GhN8QO[8V2]GhN9QO[8W2\\GgN:QO\\8W2ZGgN<PO[8X2ZGfN=ROZ8W2YGfN?QOY8Y2XGeN`0QOZ8Y2VGdNV:[1kEdNW:[1iEdNX:\\1hEbN[:]1eEbN\\:]1eEaN^:^1bEaN`:^1`EaNa:_1_E_Nd:_1]E`Nd:`1]E^Ne:a1]E[Nf:c1\\E[Ne:e1]EXNe:g1]EUNf:j1[ETNg:j1Y1O1O1O1N2O1O1O1N2N2N2N2N2N1O3M2N3M3M3M2M4M3_ORPc0"}, "label": "a sandwich placed in the plate in middle"}]}
{"id": 478833, "image": "COCO_train2014_000000478833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry panini sandwich sitting in the background of the photo\"."}], "task": "refering", "answer_template": "The \"a blurry panini sandwich sitting in the background of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84], "bbox": [0.401375, 0.0, 0.700484375, 0.2375], "iscrowd": 0, "area": 19793.387400000003, "rle": {"size": [480, 640], "counts": "__h3n0Q>;F8G8I6I6K5J6K2N0O2O1O0O2O001N101O0O2O001O0O2O1O0O2O001N101O0O2O001N101O1N101O0O2O001O001O001O00000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O10000000000000000O1O1O1O2N1O1O1O3M4L4L4L3M4L4L4L8H8H7I8E;]Ocni2"}, "label": "a blurry panini sandwich sitting in the background of the photo"}]}
{"id": 478833, "image": "COCO_train2014_000000478833.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry sandwich in the distance\"."}], "task": "refering", "answer_template": "The \"a blurry sandwich in the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84], "bbox": [0.401375, 0.0, 0.700484375, 0.2375], "iscrowd": 0, "area": 19793.387400000003, "rle": {"size": [480, 640], "counts": "__h3n0Q>;F8G8I6I6K5J6K2N0O2O1O0O2O001N101O0O2O001O0O2O1O0O2O001N101O0O2O001N101O1N101O0O2O001O001O001O00000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O10000000000000000O1O1O1O2N1O1O1O3M4L4L4L3M4L4L4L8H8H7I8E;]Ocni2"}, "label": "a blurry sandwich in the distance"}]}
{"id": 577140, "image": "COCO_train2014_000000577140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two apples below the bowl on the table\"."}], "task": "refering", "answer_template": "The \"the two apples below the bowl on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 115, 116, 117, 118, 134, 135, 136, 137, 153, 154, 155, 156, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 210, 211, 212, 213], "bbox": [0.06697718631178706, 0.22471875, 0.2761216730038023, 0.521734375], "iscrowd": 0, "area": 16533.407550000004, "rle": {"size": [640, 526], "counts": "\\Uf0`0la0e1[Ne1`N_1O2O1N101O1N101N2O001N2O001O1O001O1O001O001O000000000000000000001O0000000001O00000000100O001O001O001O1O001N101O001O1O001O001O0O2O1O1O1O2N1O1O1O2N1N2O2M2N2N2O4K6J5K6fLY@P2m?hMX@U2n?bMW@]2n?[MV@c2e`0N3M2N2N3N1N2N2N3M2O1N3M2N2M3M3M3M5K<D<D<DSk]7"}, "label": "the two apples below the bowl on the table"}]}
{"id": 577140, "image": "COCO_train2014_000000577140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green apple\"."}], "task": "refering", "answer_template": "The \"a green apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 257, 272, 274, 275], "bbox": [0.2779277566539924, 0.43353125, 0.5448859315589354, 0.62953125], "iscrowd": 0, "area": 13730.933150000003, "rle": {"size": [640, 526], "counts": "fbk23hc09G:E:H8I8I4K5L4L5K4M3L4M3M3M3M2N3M3M3M3M2N3N2M2N0010O0001O010O01O01O010O010O1O010O10O10O10O1000O010000O0100000O1000O10000000000O1001O00000001O0000000000000001O0000001O0000001O01O0001O00001O00001O0000010O01O100O00100O1O010O10O1000O01N2N2N2M3N1O2N2N2N2N3M3L5L3M3K5L4K5L5K4K5L4K5L4RObRe4"}, "label": "a green apple"}]}
{"id": 577140, "image": "COCO_train2014_000000577140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow apple\"."}], "task": "refering", "answer_template": "The \"a yellow apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 257, 272, 274, 275], "bbox": [0.2779277566539924, 0.43353125, 0.5448859315589354, 0.62953125], "iscrowd": 0, "area": 13730.933150000003, "rle": {"size": [640, 526], "counts": "fbk23hc09G:E:H8I8I4K5L4L5K4M3L4M3M3M3M2N3M3M3M3M2N3N2M2N0010O0001O010O01O01O010O010O1O010O10O10O10O1000O010000O0100000O1000O10000000000O1001O00000001O0000000000000001O0000001O0000001O01O0001O00001O00001O0000010O01O100O00100O1O010O10O1000O01N2N2N2M3N1O2N2N2N2N3M3L5L3M3K5L4K5L5K4K5L4K5L4RObRe4"}, "label": "a yellow apple"}]}
{"id": 577140, "image": "COCO_train2014_000000577140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red apple with two holes on top\"."}], "task": "refering", "answer_template": "The \"red apple with two holes on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351], "bbox": [0.27058935361216735, 0.60909375, 0.5561026615969582, 0.793796875], "iscrowd": 0, "area": 14789.112050000002, "rle": {"size": [640, 526], "counts": "aVi21ic0;E;F;@?@`0K3L2N2N2N2N2O1N2N2N2N2N2O1N2N1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O10000O10000O100O1000O010000O100O10000O10000O10000000000000000000000000000000000000000000000000000000000001O00000000001O0000000000001O001O1O001N2O001O001O1O001O1O001O001O1O001O001N2M2O2M3N1N4L3N3L4L4M3L4M3L4K4L5L4K7I8H9G9ATVa4"}, "label": "red apple with two holes on top"}]}
{"id": 577140, "image": "COCO_train2014_000000577140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red apple with two black spots near the top\"."}], "task": "refering", "answer_template": "The \"a red apple with two black spots near the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351], "bbox": [0.27058935361216735, 0.60909375, 0.5561026615969582, 0.793796875], "iscrowd": 0, "area": 14789.112050000002, "rle": {"size": [640, 526], "counts": "aVi21ic0;E;F;@?@`0K3L2N2N2N2N2O1N2N2N2N2N2O1N2N1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O10000O10000O100O1000O010000O100O10000O10000O10000000000000000000000000000000000000000000000000000000000001O00000000001O0000000000001O001O1O001N2O001O001O1O001O1O001O001O1O001O001N2M2O2M3N1N4L3N3L4L4M3L4M3L4K4L5L4K7I8H9G9ATVa4"}, "label": "a red apple with two black spots near the top"}]}
{"id": 134772, "image": "COCO_train2014_000000134772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown wooden chain\"."}], "task": "refering", "answer_template": "The \"a brown wooden chain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 272, 273, 292], "bbox": [0.705875, 0.19852083333333334, 0.92446875, 0.7225833333333334], "iscrowd": 0, "area": 22577.499500000005, "rle": {"size": [480, 640], "counts": "bVd66e>6J5K5K5J6K4L2N1O2aNmNRET1g:XOREk0P;WOjDj0X;YObDi0`;YOZDh0i;[OoCg0]8^NRKc2Q5RNYJk1k5kN]IV1d6_OfHc0Y72QHOo7g0YG[Og8_30000009G2NO1O1O1O1O2O0O1O1O1O1O100O1O1O1O101N1O1O1O100000000000000O2O00000000000000000O1000001O0000000000000O100000001O00000000000O100001O001O010O001O001O0m0TO0O010O100O100O01E:B?@?A`0G9J6QN]FZNi9a1_FVNf9e1cFRNc9i1eFoM`9l1gFoM\\9k1mFnMW9m1QGmMQ9n1XGlMk8o1U2K4M3M3L5L3L4M3L4M5K4K5L5J5L4Lfkf0"}, "label": "a brown wooden chain"}]}
{"id": 134772, "image": "COCO_train2014_000000134772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair is behind a lady\"."}], "task": "refering", "answer_template": "The \"a brown chair is behind a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 272, 273, 292], "bbox": [0.705875, 0.19852083333333334, 0.92446875, 0.7225833333333334], "iscrowd": 0, "area": 22577.499500000005, "rle": {"size": [480, 640], "counts": "bVd66e>6J5K5K5J6K4L2N1O2aNmNRET1g:XOREk0P;WOjDj0X;YObDi0`;YOZDh0i;[OoCg0]8^NRKc2Q5RNYJk1k5kN]IV1d6_OfHc0Y72QHOo7g0YG[Og8_30000009G2NO1O1O1O1O2O0O1O1O1O1O100O1O1O1O101N1O1O1O100000000000000O2O00000000000000000O1000001O0000000000000O100000001O00000000000O100001O001O010O001O001O0m0TO0O010O100O100O01E:B?@?A`0G9J6QN]FZNi9a1_FVNf9e1cFRNc9i1eFoM`9l1gFoM\\9k1mFnMW9m1QGmMQ9n1XGlMk8o1U2K4M3M3L5L3L4M3L4M5K4K5L5J5L4Lfkf0"}, "label": "a brown chair is behind a lady"}]}
{"id": 142969, "image": "COCO_train2014_000000142969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left side train is capitalas utd express\"."}], "task": "refering", "answer_template": "The \"left side train is capitalas utd express\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238], "bbox": [0.21832812499999998, 0.2808983451536643, 0.43071874999999993, 0.6898817966903074], "iscrowd": 0, "area": 19433.0213, "rle": {"size": [423, 640], "counts": "RQj1W1i;<C8gDWNm:Q2RERNd:U2[EoM]:V2cElMW:e2J6J6J6L4N2M3N1O2N2N1O2N2M2O2N1O2N2N1O2M2O2N2N101O0O2O1N101O0O2O1O0O2O00O010O010O10O01000O0100O010O100O100O100000010O01O001O1O001O001O1O00100O001O001O1O001O001O01O00001O001N2O001O001O1O001O001O001O1O001O001M3K4L5L3N3N2N1O2N1O2N1O2N2N1O2^OjFPMW9P3nFjLT9T3a0O3L5L3L4L4M3L5POo0dNbWf4"}, "label": "left side train is capitalas utd express"}]}
{"id": 142969, "image": "COCO_train2014_000000142969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one yellow color old train standing on the track along with other train\"."}], "task": "refering", "answer_template": "The \"there is one yellow color old train standing on the track along with other train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238], "bbox": [0.21832812499999998, 0.2808983451536643, 0.43071874999999993, 0.6898817966903074], "iscrowd": 0, "area": 19433.0213, "rle": {"size": [423, 640], "counts": "RQj1W1i;<C8gDWNm:Q2RERNd:U2[EoM]:V2cElMW:e2J6J6J6L4N2M3N1O2N2N1O2N2M2O2N1O2N2N1O2M2O2N2N101O0O2O1N101O0O2O1O0O2O00O010O010O10O01000O0100O010O100O100O100000010O01O001O1O001O001O1O00100O001O001O1O001O001O01O00001O001N2O001O001O1O001O001O001O1O001O001M3K4L5L3N3N2N1O2N1O2N1O2N2N1O2^OjFPMW9P3nFjLT9T3a0O3L5L3L4L4M3L5POo0dNbWf4"}, "label": "there is one yellow color old train standing on the track along with other train"}]}
{"id": 142969, "image": "COCO_train2014_000000142969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train that has a windshield wiper on the window\"."}], "task": "refering", "answer_template": "The \"a train that has a windshield wiper on the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245], "bbox": [0.411421875, 0.28763593380614655, 0.735203125, 0.7325768321513002], "iscrowd": 0, "area": 30329.335450000002, "rle": {"size": [423, 640], "counts": "^V]32U=1N3N1O1N3N1O2M2O1O2RDV1Z9]1O1O1O1N2O1O1N2O2N1N2O1O1O1N2O1O1N2O1O1N3N1O1N2O1O1O1N2O1O1N2O1O1N3N1O1O1N2O1O1N2O1O1N2O1O1O2M2O001N2O001O1N2O001N2O001O1N10000O1000001N10000O1000000O10000O10001O0O1000000000000000000000000000000000000000000000000000000000000000000001O0000000000001O0000001O00001O0000001O0000001O0000001N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O2N1O2N1O2N2N1O2M2O2M2N3M2O4K6J6K5J6J6K5J6J6J6K5J6J6K5J6J6J6K5J6J6K5J6J6K5JfbU2"}, "label": "a train that has a windshield wiper on the window"}]}
{"id": 224891, "image": "COCO_train2014_000000224891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller giraffe in the front walking in the zoo\"."}], "task": "refering", "answer_template": "The \"the smaller giraffe in the front walking in the zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 51, 52, 68, 69, 70, 85, 86, 87, 103, 104, 105, 121, 122, 123, 138, 139, 140, 141, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 224, 225, 226, 227, 228, 229, 230, 241, 242, 244, 245, 246, 258, 259, 260, 262, 263, 274, 275, 276, 277, 279, 280, 281, 291, 292, 294, 296, 297, 298, 308, 311, 313, 315, 325, 328, 330, 332, 341, 342, 345, 346, 347, 349, 362, 363], "bbox": [0.011758691206543968, 0.11909375, 0.5911656441717791, 0.9303437500000001], "iscrowd": 0, "area": 41635.70495000001, "rle": {"size": [640, 489], "counts": "ej35e1M_`05\\_ONe`03V_O0j`02S_OOm`03P_ONQa02m^OOSa03j^ONWa03f^ONZa05b^OL^a07^^OJca08Y^OIga0;T^OFla0b0k]O_OVb0U1001ON2N2N2N100O1O11O1O1N21OYNo]O_1Zb00O001O0010OD=O001O001O3\\KjMXGZ2c8hM]GZ2_8gMoGo1l7RN[Hj1a7WN`Hl1[7UNgHl1T7VNmHm1n6TNSIo1h6RNYIP2c6RN]IQ2]6QNdIQ2V6RNkIQ2o5QNRJR2g5QN[JP2`5RNaJQ2X5RNiJP2R5RNPKP2j4RNYKn1a4VN`Kk1[4WNhKi1R4ZNQLf1j3\\NXLe1b3^NaLb1Z3`NiL`1Q3WN]Mj1]2UNjMk1P2TNVNm1d1RNbNo1X1PNnNQ2l0nMZOS2`0kMGV23hM4Z2FcMa0^2WNmL]Jd0b7_2mMTMYJ;Q8a2bM[MUJ2`8b2XMcMPJIo8d2mLYNY3f1cLXNd3g1YLYNk3g1ZHnLe1[1U6f1PHVMg1T1]6f1eG\\Mk1n0d6e1[GcMn1h0j6f1QGhMR2b0Q7e1gFoMU2<W7f1\\FUNZ25^7e1RF]N\\2Ne7f1hEbN_2Im7d1^EiNb2CS8d3jG\\LY8d3dG\\L`8d3\\G\\Lg8d3VG\\Ln8c3oF]LT9c3iF]LZ9h3^FXLf9k3oEXLU:m3XE^Ll:f3bDdLa;k53L5L3M4L3M4K4M3L4L4L4K5L4L4L4L4L4L4K52N2N2N2N2N2N2N3M2N2N2N2N2N2M2O2NgMbCPM\\<P3gCPMW<o2lCQMR<o2oCQMo;o2TDQMj;n2YDQMf;n2\\DSM_2^N\\5`4WHRMY2_Na5^4YHRMS2aNe5\\4ZHTMm1aNi5[4]HSMg1cNm5Y4_HTM`1dNR6X4_HUM\\1cNU6X4bHTMV1fNX6U4eHWMm0eN_6S4gH\\Mb0bNg6S4mHZM5dNo6Q4TIZMEgNW7n3[IZMWOiN_7l3`I\\MiNiNh7k3cI]M[NlNS8f3hI_MiMPO`8`3mI`MYMUOj8Z3RJcMhLXOW9U3VJdNk5[1ZJ`Nh5^1^J[Nd5d1aJVNa5j1dJPN]5o1hJkMZ5T2lJeMV5Z2oJSM[KXOh9e3RKlLZK[Of9h3UKgLZK\\Oc9l3\\K]LU5a3f500O1O1O100O1N3MnE`Lf3_3YLcLh3\\3ULgLk3Z3QLiLP4W3mKkLS4V3iKmLX4T3cKPM\\4Q3aKQM`4o2]KSMc4o2YKSMh4m2UKUMk4l2RKVMo4k2mJWMS5j2jJYMV5g2gJ[MY5g2cJ[M^5e2_J]Ma5d2\\J^Me5b2XJ`Mh5b2TJ`Mm5`2PJ_MS6b2jI[M\\6f2`IXMd6i2UIYMP7g2hH\\M\\7f2[H^Mh7c2QH_MT8a2dGbM`8`2WGdMm8\\2kFgMY9Z2mE\\NW:c53L5L3M4L3L5L3M4L3L5L3M4K4M4M2M4M2M4M3L5L3L4N2N2N2O2M2N2N2N2O2M2N2N3M4L3<D001O00100O001O00iLPB=P>ROmBb0R=mNaCQ1j?O0O2O1O0O2O0O200O010O01000O010O10O10O10OL5J5J7J6I6K6_OX_l3"}, "label": "the smaller giraffe in the front walking in the zoo"}]}
{"id": 224891, "image": "COCO_train2014_000000224891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the left of the others\"."}], "task": "refering", "answer_template": "The \"a giraffe to the left of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 51, 52, 68, 69, 70, 85, 86, 87, 103, 104, 105, 121, 122, 123, 138, 139, 140, 141, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 224, 225, 226, 227, 228, 229, 230, 241, 242, 244, 245, 246, 258, 259, 260, 262, 263, 274, 275, 276, 277, 279, 280, 281, 291, 292, 294, 296, 297, 298, 308, 311, 313, 315, 325, 328, 330, 332, 341, 342, 345, 346, 347, 349, 362, 363], "bbox": [0.011758691206543968, 0.11909375, 0.5911656441717791, 0.9303437500000001], "iscrowd": 0, "area": 41635.70495000001, "rle": {"size": [640, 489], "counts": "ej35e1M_`05\\_ONe`03V_O0j`02S_OOm`03P_ONQa02m^OOSa03j^ONWa03f^ONZa05b^OL^a07^^OJca08Y^OIga0;T^OFla0b0k]O_OVb0U1001ON2N2N2N100O1O11O1O1N21OYNo]O_1Zb00O001O0010OD=O001O001O3\\KjMXGZ2c8hM]GZ2_8gMoGo1l7RN[Hj1a7WN`Hl1[7UNgHl1T7VNmHm1n6TNSIo1h6RNYIP2c6RN]IQ2]6QNdIQ2V6RNkIQ2o5QNRJR2g5QN[JP2`5RNaJQ2X5RNiJP2R5RNPKP2j4RNYKn1a4VN`Kk1[4WNhKi1R4ZNQLf1j3\\NXLe1b3^NaLb1Z3`NiL`1Q3WN]Mj1]2UNjMk1P2TNVNm1d1RNbNo1X1PNnNQ2l0nMZOS2`0kMGV23hM4Z2FcMa0^2WNmL]Jd0b7_2mMTMYJ;Q8a2bM[MUJ2`8b2XMcMPJIo8d2mLYNY3f1cLXNd3g1YLYNk3g1ZHnLe1[1U6f1PHVMg1T1]6f1eG\\Mk1n0d6e1[GcMn1h0j6f1QGhMR2b0Q7e1gFoMU2<W7f1\\FUNZ25^7e1RF]N\\2Ne7f1hEbN_2Im7d1^EiNb2CS8d3jG\\LY8d3dG\\L`8d3\\G\\Lg8d3VG\\Ln8c3oF]LT9c3iF]LZ9h3^FXLf9k3oEXLU:m3XE^Ll:f3bDdLa;k53L5L3M4L3M4K4M3L4L4L4K5L4L4L4L4L4L4K52N2N2N2N2N2N2N3M2N2N2N2N2N2M2O2NgMbCPM\\<P3gCPMW<o2lCQMR<o2oCQMo;o2TDQMj;n2YDQMf;n2\\DSM_2^N\\5`4WHRMY2_Na5^4YHRMS2aNe5\\4ZHTMm1aNi5[4]HSMg1cNm5Y4_HTM`1dNR6X4_HUM\\1cNU6X4bHTMV1fNX6U4eHWMm0eN_6S4gH\\Mb0bNg6S4mHZM5dNo6Q4TIZMEgNW7n3[IZMWOiN_7l3`I\\MiNiNh7k3cI]M[NlNS8f3hI_MiMPO`8`3mI`MYMUOj8Z3RJcMhLXOW9U3VJdNk5[1ZJ`Nh5^1^J[Nd5d1aJVNa5j1dJPN]5o1hJkMZ5T2lJeMV5Z2oJSM[KXOh9e3RKlLZK[Of9h3UKgLZK\\Oc9l3\\K]LU5a3f500O1O1O100O1N3MnE`Lf3_3YLcLh3\\3ULgLk3Z3QLiLP4W3mKkLS4V3iKmLX4T3cKPM\\4Q3aKQM`4o2]KSMc4o2YKSMh4m2UKUMk4l2RKVMo4k2mJWMS5j2jJYMV5g2gJ[MY5g2cJ[M^5e2_J]Ma5d2\\J^Me5b2XJ`Mh5b2TJ`Mm5`2PJ_MS6b2jI[M\\6f2`IXMd6i2UIYMP7g2hH\\M\\7f2[H^Mh7c2QH_MT8a2dGbM`8`2WGdMm8\\2kFgMY9Z2mE\\NW:c53L5L3M4L3L5L3M4L3L5L3M4K4M4M2M4M2M4M3L5L3L4N2N2N2O2M2N2N2N2O2M2N2N3M4L3<D001O00100O001O00iLPB=P>ROmBb0R=mNaCQ1j?O0O2O1O0O2O0O200O010O01000O010O10O10O10OL5J5J7J6I6K6_OX_l3"}, "label": "a giraffe to the left of the others"}]}
{"id": 224891, "image": "COCO_train2014_000000224891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle giraffe with a patch of sunlight hitting its back\"."}], "task": "refering", "answer_template": "The \"the middle giraffe with a patch of sunlight hitting its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 21, 22, 23, 38, 39, 40, 41, 55, 56, 57, 58, 74, 75, 76, 92, 93, 94, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 254, 265, 266, 267, 270, 271, 282, 283, 284, 288, 299, 300, 301, 302, 305, 316, 317, 318, 319, 322, 333, 335, 336, 339, 349, 350, 353, 356, 366, 367, 370, 371, 373, 387, 388, 389, 390], "bbox": [0.2346830265848671, 0.007718750000000001, 1.0, 0.9937656250000002], "iscrowd": 0, "area": 65133.55664999996, "rle": {"size": [640, 489], "counts": "fnW25jc06I7I4M0O1O2O0O2N101N1O2O0O1O2O0O2]Oc0\\Oc00100O001O1O0010O01O1O001000O010O10000O01000O10000O010O10000O01000O101N1011N2N2O1N2O1N3M2O1N2NiMo^Oa1Pa0]NR_Od1l`0ZNX_Of1g`0WN\\_Oj1b`0TNb_Ol1\\`0RNg_Oo1X`0nMl_OR2R`0mMP@T2o?iMU@W2i?gMZ@Z2g`00O100O100O100O100O100O100O100O100O100O100O101N1O100O100O100O100O100O100O100O100O100O100O100O100O100O10O010O0010O001O001N10001O0O2O000O2O001O0O101O001N101O0O101O0QJ^L_Kc3^4_LaKb3]4`LcK`3Z4cLfK^3W4dLhK]3V4fLiK[3S4hLmKX3hK]La6=gMW3`KdLf67iMV3ZKkLi61nMT3RKRMn6KQNT3iJYMQ7FWNQ3aJaMT7@[NQ3XJgMY7[O`Nh4[1[KfNf4U1]KkNd4Q1_KPOb4k0aKVO`4e0cK\\O]4`0fK@\\4<fKE[46gKLY40jK0X4KkK6V4EmK<T4_OoKb0S4XOPLh0S4ROPLo0R4kNQLV1Q4eLPN]3R2WLWNi3l1kKYNZ4i1ZK]Nj4e1jJaNY5b1[JdNi5f74L4ZOe0G:G9F9K6K5K5L3L5K5L4K4L5L4K4L5L4K5L3L5K5L4F9[Of0]Oc02N1O2N2N1N3N2N2N1O2N2N1N3N2NZMUDWMj;h2YDXMe;g2^DYM`;f2cDZM[;e2hD\\MT;e2mD\\MQ;c2RE]Ml:b2WE^Mg:a2\\E_Mb:`2aE`M]:`2eEaMX:^2jEcMS:]2PFcMn9\\2UFeMh9[2ZFeMd9Z2_FfM_9Y2dFhMY9W2iFjMU9U2nFlMn8U2UGjMi8U2\\GiMb8V2bGhM]8W2gGhMW8W2mGgMR8Y2QHfMl7Z2XHeMf7Z2^HdMa7[2cHdM[7\\2hHcMV7\\2oHaMP7^2TIaMj6^2ZI`Md6`2`I_M^6a2eI^MX6b2lI\\MS6c2QJ\\Ml5d2XJZMg5e2]JZM`5g2dJWMdMQNQ6h4_LUM[MVNW6d4bLUMQMZN]6a4fLTMhL]Nc6^4iLSM_LbNh6\\4lLQMVLfNn6Y4PMoLnKjNS7V4SMoLdKnNY7S4WMnL[KQO^7R4ZMkLSKVOd7n3[MgMe2Y2^MdMc2[2_McMa2]2bM`M^2a2cM\\M_2c2dMZM\\2f2fMWM\\2h2fMVMZ2j2iMRMX2o2iMnLY2Q3jMlLV2T3lMiLU2W3nMfLS2Z3nMcLS2]3PN`LQ2_3QN^LP2c3aNjK`1W4cNeK^1Z4fNaK[1`4hN\\KY1c4R71O10O01O010O10O01O010O10O01O10O010O01O10O02N101N100O2N101N100O2N101N1O2O0O101N1O2lFdJa3\\5YLmJc3U5WLSKf3P5QLYKk3j4nK]Ko3g4hK^KY4e4_K_Ka4d4WKaKi4c4oJbKQ5`4gJeKY5_4^JfKc5]4UJgKk5\\4mIiKS6[4eIiK\\6Z4[IkKe6X4TImKl6c800001O00010O001O0O1O2N1N3N1O^1bNk2TMWC"}, "label": "the middle giraffe with a patch of sunlight hitting its back"}]}
{"id": 224891, "image": "COCO_train2014_000000224891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tall giraffe in the center\"."}], "task": "refering", "answer_template": "The \"the tall giraffe in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 21, 22, 23, 38, 39, 40, 41, 55, 56, 57, 58, 74, 75, 76, 92, 93, 94, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 254, 265, 266, 267, 270, 271, 282, 283, 284, 288, 299, 300, 301, 302, 305, 316, 317, 318, 319, 322, 333, 335, 336, 339, 349, 350, 353, 356, 366, 367, 370, 371, 373, 387, 388, 389, 390], "bbox": [0.2346830265848671, 0.007718750000000001, 1.0, 0.9937656250000002], "iscrowd": 0, "area": 65133.55664999996, "rle": {"size": [640, 489], "counts": "fnW25jc06I7I4M0O1O2O0O2N101N1O2O0O1O2O0O2]Oc0\\Oc00100O001O1O0010O01O1O001000O010O10000O01000O10000O010O10000O01000O101N1011N2N2O1N2O1N3M2O1N2NiMo^Oa1Pa0]NR_Od1l`0ZNX_Of1g`0WN\\_Oj1b`0TNb_Ol1\\`0RNg_Oo1X`0nMl_OR2R`0mMP@T2o?iMU@W2i?gMZ@Z2g`00O100O100O100O100O100O100O100O100O100O100O101N1O100O100O100O100O100O100O100O100O100O100O100O100O100O10O010O0010O001O001N10001O0O2O000O2O001O0O101O001N101O0O101O0QJ^L_Kc3^4_LaKb3]4`LcK`3Z4cLfK^3W4dLhK]3V4fLiK[3S4hLmKX3hK]La6=gMW3`KdLf67iMV3ZKkLi61nMT3RKRMn6KQNT3iJYMQ7FWNQ3aJaMT7@[NQ3XJgMY7[O`Nh4[1[KfNf4U1]KkNd4Q1_KPOb4k0aKVO`4e0cK\\O]4`0fK@\\4<fKE[46gKLY40jK0X4KkK6V4EmK<T4_OoKb0S4XOPLh0S4ROPLo0R4kNQLV1Q4eLPN]3R2WLWNi3l1kKYNZ4i1ZK]Nj4e1jJaNY5b1[JdNi5f74L4ZOe0G:G9F9K6K5K5L3L5K5L4K4L5L4K4L5L4K5L3L5K5L4F9[Of0]Oc02N1O2N2N1N3N2N2N1O2N2N1N3N2NZMUDWMj;h2YDXMe;g2^DYM`;f2cDZM[;e2hD\\MT;e2mD\\MQ;c2RE]Ml:b2WE^Mg:a2\\E_Mb:`2aE`M]:`2eEaMX:^2jEcMS:]2PFcMn9\\2UFeMh9[2ZFeMd9Z2_FfM_9Y2dFhMY9W2iFjMU9U2nFlMn8U2UGjMi8U2\\GiMb8V2bGhM]8W2gGhMW8W2mGgMR8Y2QHfMl7Z2XHeMf7Z2^HdMa7[2cHdM[7\\2hHcMV7\\2oHaMP7^2TIaMj6^2ZI`Md6`2`I_M^6a2eI^MX6b2lI\\MS6c2QJ\\Ml5d2XJZMg5e2]JZM`5g2dJWMdMQNQ6h4_LUM[MVNW6d4bLUMQMZN]6a4fLTMhL]Nc6^4iLSM_LbNh6\\4lLQMVLfNn6Y4PMoLnKjNS7V4SMoLdKnNY7S4WMnL[KQO^7R4ZMkLSKVOd7n3[MgMe2Y2^MdMc2[2_McMa2]2bM`M^2a2cM\\M_2c2dMZM\\2f2fMWM\\2h2fMVMZ2j2iMRMX2o2iMnLY2Q3jMlLV2T3lMiLU2W3nMfLS2Z3nMcLS2]3PN`LQ2_3QN^LP2c3aNjK`1W4cNeK^1Z4fNaK[1`4hN\\KY1c4R71O10O01O010O10O01O010O10O01O10O010O01O10O02N101N100O2N101N100O2N101N1O2O0O101N1O2lFdJa3\\5YLmJc3U5WLSKf3P5QLYKk3j4nK]Ko3g4hK^KY4e4_K_Ka4d4WKaKi4c4oJbKQ5`4gJeKY5_4^JfKc5]4UJgKk5\\4mIiKS6[4eIiK\\6Z4[IkKe6X4TImKl6c800001O00010O001O0O1O2N1N3N1O^1bNk2TMWC"}, "label": "the tall giraffe in the center"}]}
{"id": 224895, "image": "COCO_train2014_000000224895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pedestal sink between the toilet and the tub\"."}], "task": "refering", "answer_template": "The \"a pedestal sink between the toilet and the tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 246, 247, 248, 249, 263, 264, 265, 280, 281, 297, 298, 314, 315, 332], "bbox": [0.4311666666666667, 0.540140625, 0.6900208333333334, 0.835125], "iscrowd": 0, "area": 10323.3751, "rle": {"size": [640, 480], "counts": "ofQ48gc05L3M4L3M2N1O2N1O1O2M2O1O1O2N1O1O1O001N2O001O1O001O001O001N101O001UA_NU;a1aDkN];V1XDVOf;j0QDAm;?iCMU<4`C8^<HXCc0f<_OhBS1W=mNWBf1h=Z2O1O1O00001O0000001O0000000000O10000O10000O100O1N2H8N2E;E;E;E;E;F:E;E;E;E;J6000000O2O0000000O101O00000O10001O000O1000001N100000000O2O0O1O1O2N3M2O1N3M2N3M2M4M2M4I6Jehl2"}, "label": "a pedestal sink between the toilet and the tub"}]}
{"id": 224895, "image": "COCO_train2014_000000224895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white pedastal sink\"."}], "task": "refering", "answer_template": "The \"the white pedastal sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 246, 247, 248, 249, 263, 264, 265, 280, 281, 297, 298, 314, 315, 332], "bbox": [0.4311666666666667, 0.540140625, 0.6900208333333334, 0.835125], "iscrowd": 0, "area": 10323.3751, "rle": {"size": [640, 480], "counts": "ofQ48gc05L3M4L3M2N1O2N1O1O2M2O1O1O2N1O1O1O001N2O001O1O001O001O001N101O001UA_NU;a1aDkN];V1XDVOf;j0QDAm;?iCMU<4`C8^<HXCc0f<_OhBS1W=mNWBf1h=Z2O1O1O00001O0000001O0000000000O10000O10000O100O1N2H8N2E;E;E;E;E;F:E;E;E;E;J6000000O2O0000000O101O00000O10001O000O1000001N100000000O2O0O1O1O2N3M2O1N3M2N3M2M4M2M4I6Jehl2"}, "label": "the white pedastal sink"}]}
{"id": 388740, "image": "COCO_train2014_000000388740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty wooden chair on the small wall to the right\"."}], "task": "refering", "answer_template": "The \"the empty wooden chair on the small wall to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 247, 248, 249, 270, 271, 272, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340], "bbox": [0.663203125, 0.6498349056603773, 0.8460625, 0.9966745283018867], "iscrowd": 0, "area": 6879.081050000006, "rle": {"size": [424, 640], "counts": "nn_51U=5L4SOIaD9\\;L`D6^;O]D3c;OZD2e;1WD1i;g000O100O100O10000O100O10000000O1N1O2N2N200000O10000000000O101O00000O1000000O100000000O100L4M3M3L4M3M4K4M3M3L4M3M6I7J6J3L3N3M3L4M3M3L4M3M3L4MBXGbLe8]3bG`L]8^3iG_LW8^3oG_LP8_3VH]Lk7<ZGe2h9XMZFf2h9YMZFd2i9ZMXFd2k9ZMWFc2k9[MWFb2m9\\MTFb2[:M2N3L3N3L3N3L3M4M2M4L3N3L3M4M2M4M2M4L2O2M3M2O2M2O4K4L]fX1"}, "label": "the empty wooden chair on the small wall to the right"}]}
{"id": 388740, "image": "COCO_train2014_000000388740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown wooden chair with a white and red seat cover\"."}], "task": "refering", "answer_template": "The \"a brown wooden chair with a white and red seat cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 247, 248, 249, 270, 271, 272, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340], "bbox": [0.663203125, 0.6498349056603773, 0.8460625, 0.9966745283018867], "iscrowd": 0, "area": 6879.081050000006, "rle": {"size": [424, 640], "counts": "nn_51U=5L4SOIaD9\\;L`D6^;O]D3c;OZD2e;1WD1i;g000O100O100O10000O100O10000000O1N1O2N2N200000O10000000000O101O00000O1000000O100000000O100L4M3M3L4M3M4K4M3M3L4M3M6I7J6J3L3N3M3L4M3M3L4M3M3L4MBXGbLe8]3bG`L]8^3iG_LW8^3oG_LP8_3VH]Lk7<ZGe2h9XMZFf2h9YMZFd2i9ZMXFd2k9ZMWFc2k9[MWFb2m9\\MTFb2[:M2N3L3N3L3N3L3M4M2M4L3N3L3M4M2M4M2M4L2O2M3M2O2M2O4K4L]fX1"}, "label": "a brown wooden chair with a white and red seat cover"}]}
{"id": 396933, "image": "COCO_train2014_000000396933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bed being laid on by a dog\"."}], "task": "refering", "answer_template": "The \"a white bed being laid on by a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 275, 287, 288, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 379, 380, 381, 382, 383, 384, 385, 386, 387, 402, 403, 404, 405, 406], "bbox": [0.505109375, 0.5383204633204633, 0.9963124999999999, 0.9904054054054053], "iscrowd": 0, "area": 38083.716949999995, "rle": {"size": [518, 640], "counts": "SlS5=^>[1eN[1QLnL_I]4a6cKTHi5k7Y100000000000000000\\KnFT2R9iMRGV2n8hMUGW2k8fMYGY2g8dM]G[2c8bMaG]2_8aMdG^2\\8aMfG^2Z8bMjGZ2V8fMaHc1_7]NbHb1^7^NbHb1^7^NcHa1]7_NcHa1]7_NdH`1\\7`NdH`1[7aNeH_1[7aNfH^1Z7cNeH]1[7cNfH\\1Z7dNfH\\1Z7dNgH[1Y7eNgH[1Y7eNgH[1Y7eNhHZ1X7fNgH\\1X7eNgH[1Y7cNiH]1W7aNjH`1V7]NmHc1S7[NnHf1R7WNQIi1o6VNRIj1n6UNRIl1n6TNRIl1n6SNSIm1m6SNRIn1n6QNSIo1m6QNSIo1m6PNSIQ2m6oMSIQ2m6nMSIS2m6mMSIS2m6mMSIS2m6mMRIT2n6lMRIT2n6lMRIT2n6lMQIU2o6lMPIT2P7lMoHU2Q7kMoHU2Q7kMoHU2Q7kMnHV2R7kMmHU2S7lMlHT2T7mMjHT2V7lMjHT2V7mMhHT2X7mMgHS2Y7nMfHR2Z7oMdHS2[7nMdHR2\\7oMcHQ2]7PNaHQ2_7PN`HP2`7QN^HP2b7RN\\Hn1d7SN[Hm1e7TNYHm1f7TNZHl1f7TNZHl1f7SNZHn1f7QN[Ho1e7QNZHP2f7oM[HQ2e7oM[HQ2e7nM\\HR2d7mM]HS2c7mM]Hl1j7SNWHm1i7SNXHk1i7TNXHl1h7SNYHl1h7TNXHl1h7SNYHl1h7SNYHm1g7SNZHk1g7TNZHk1g7UNYHk1g7TNZHk1g7TNZHl1f7TNZHk1g7UNYHk1g7UNZHi1g7WNYHi1g7WNYHh1h7XNXHh1h7XNXHg1i7YNWHf1j7ZNWHe1i7[NWHd1j7\\NUHe1k7[NUHd1l7\\NTHd1l7\\NTHc1m7]NSHc1m7]NSHb1n7^NRHb1n7^NRHa1o7_NQH_1Q8aNoG^1R8bNnG\\1T8dNlGZ1V8fNjGX1X8hNhGW1Y8iNgGU1[8kNeGS1]8mNcGR1^8nNaGQ1a8nN`GP1b8PO^Go0c8PO^Gn0d8RO\\Gl0e8TO\\Gk0e8UO[Gi0g8WOYGh0h8WOYGi0g8WOYGi0g8VOZGj0f8VOZGj0f8UO[Gk0e8UO[Gk0e8UO[Gk0e8TO\\Gl0d8TO\\Gl0d8SO\\Gn0d8RO\\Gn0d8QO]Go0c8QO]Go0c8QO]Go0c8PO^GP1b8PO^GP1b8oN_GQ1a8oN_GQ1a8nN`GR1`8nN`GR1`8mNaGS1_8mNaGS1_8mNaGS1_8lNbGT1^8lNbGT1^8kNcGU1]8kNbGV1^8iNcGX1\\8hNdGX1\\8hNdGX1\\8gNeGY1[8gNeGY1[8fNfGZ1Z8fNfGZ1Z8eNgG[1Y8eNgG[1Y8eNgG[1Y8dNhG\\1X8dNhG\\1X8cNiG^1V8bNkG]1U8cNkG]1U8bNmG]1S8cNmG]1S8cNnG\\1R8cNoG]1Q8cNoG]1Q8cNPH\\1P8dNPH\\1P8cNRH\\1n7dNRH]1l7dNUH[1k7dNVH\\1j7dNWH[1i7eNWH[1i7eNWH[1i7dNYH[1g7eNYH[1g7eNZHZ1f7eN[H[1e7eN\\HZ1d7fN\\HZ1d7eN^H[1a7eN_H[1a7eN`HZ1`7fN`HZ1`7eNaH[1_7eNcHY1]7gNdHX1\\7gNfHW1[7iNfHV1Z7jNgHU1Y7jNiHU1W7kNjHS1W7mNjHR1V7nNlHP1T7oNnHo0S7QOnHn0R7ROoHm0Q7ROQIl0P7TOQIk0o6UORIj0n6VOTIg0m6XOUIg0k6YOVIf0j6ZOWIe0i6ZOYId0h6\\OYIc0g6]OZIb0f6]O]I`0d6@`G@Ho0h8A`GBGl0i8B`GDFh0k8C`GGDe0l8D`GICb0m8E`GJC`0m8EaGMA<o8G`GOA8o8I`G1@5P9IaG4^O1R9K`G6]ONS9KaG8\\OLS9LaG:[OHU9N`G<ZOEV9NaG?YOAU91bG`0XO]OW93aGb0WOZOX93bGd0VOXOX94bGf0UOUOY94cGi0SOQO[96bGk0ROnN\\97bGm0ROjN\\98cGP1POfN^9:bGQ1i9nNXFT1f9lNZFV1d9jN\\FX1b9gN_F[1_9eN`F^1^9bNbF_1]9`NdFb1Z9^NfFd1X9[NiFg1U9YNkFi1S9WNmFk1Q9TNPGm1o8SNQGo1m8QNSGQ2k8nMUGR2l8nMSGR2n8mMRGS2o8mMPGR2R9nMmFR2T9mMlFS2U9mMjFS2W9mMhFR2Z9mMfFS2[9mMdFS2]9lMcFT2^9lMaFS2a9mM^FS2j;0O10O1O1O1O1O1001O1O1O1O1O1O100000000000000000000001O3M7I7E;D<D<D<Ddd0"}, "label": "a white bed being laid on by a dog"}]}
{"id": 396933, "image": "COCO_train2014_000000396933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed that the dog is laying on\"."}], "task": "refering", "answer_template": "The \"the bed that the dog is laying on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 275, 287, 288, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 379, 380, 381, 382, 383, 384, 385, 386, 387, 402, 403, 404, 405, 406], "bbox": [0.505109375, 0.5383204633204633, 0.9963124999999999, 0.9904054054054053], "iscrowd": 0, "area": 38083.716949999995, "rle": {"size": [518, 640], "counts": "SlS5=^>[1eN[1QLnL_I]4a6cKTHi5k7Y100000000000000000\\KnFT2R9iMRGV2n8hMUGW2k8fMYGY2g8dM]G[2c8bMaG]2_8aMdG^2\\8aMfG^2Z8bMjGZ2V8fMaHc1_7]NbHb1^7^NbHb1^7^NcHa1]7_NcHa1]7_NdH`1\\7`NdH`1[7aNeH_1[7aNfH^1Z7cNeH]1[7cNfH\\1Z7dNfH\\1Z7dNgH[1Y7eNgH[1Y7eNgH[1Y7eNhHZ1X7fNgH\\1X7eNgH[1Y7cNiH]1W7aNjH`1V7]NmHc1S7[NnHf1R7WNQIi1o6VNRIj1n6UNRIl1n6TNRIl1n6SNSIm1m6SNRIn1n6QNSIo1m6QNSIo1m6PNSIQ2m6oMSIQ2m6nMSIS2m6mMSIS2m6mMSIS2m6mMRIT2n6lMRIT2n6lMRIT2n6lMQIU2o6lMPIT2P7lMoHU2Q7kMoHU2Q7kMoHU2Q7kMnHV2R7kMmHU2S7lMlHT2T7mMjHT2V7lMjHT2V7mMhHT2X7mMgHS2Y7nMfHR2Z7oMdHS2[7nMdHR2\\7oMcHQ2]7PNaHQ2_7PN`HP2`7QN^HP2b7RN\\Hn1d7SN[Hm1e7TNYHm1f7TNZHl1f7TNZHl1f7SNZHn1f7QN[Ho1e7QNZHP2f7oM[HQ2e7oM[HQ2e7nM\\HR2d7mM]HS2c7mM]Hl1j7SNWHm1i7SNXHk1i7TNXHl1h7SNYHl1h7TNXHl1h7SNYHl1h7SNYHm1g7SNZHk1g7TNZHk1g7UNYHk1g7TNZHk1g7TNZHl1f7TNZHk1g7UNYHk1g7UNZHi1g7WNYHi1g7WNYHh1h7XNXHh1h7XNXHg1i7YNWHf1j7ZNWHe1i7[NWHd1j7\\NUHe1k7[NUHd1l7\\NTHd1l7\\NTHc1m7]NSHc1m7]NSHb1n7^NRHb1n7^NRHa1o7_NQH_1Q8aNoG^1R8bNnG\\1T8dNlGZ1V8fNjGX1X8hNhGW1Y8iNgGU1[8kNeGS1]8mNcGR1^8nNaGQ1a8nN`GP1b8PO^Go0c8PO^Gn0d8RO\\Gl0e8TO\\Gk0e8UO[Gi0g8WOYGh0h8WOYGi0g8WOYGi0g8VOZGj0f8VOZGj0f8UO[Gk0e8UO[Gk0e8UO[Gk0e8TO\\Gl0d8TO\\Gl0d8SO\\Gn0d8RO\\Gn0d8QO]Go0c8QO]Go0c8QO]Go0c8PO^GP1b8PO^GP1b8oN_GQ1a8oN_GQ1a8nN`GR1`8nN`GR1`8mNaGS1_8mNaGS1_8mNaGS1_8lNbGT1^8lNbGT1^8kNcGU1]8kNbGV1^8iNcGX1\\8hNdGX1\\8hNdGX1\\8gNeGY1[8gNeGY1[8fNfGZ1Z8fNfGZ1Z8eNgG[1Y8eNgG[1Y8eNgG[1Y8dNhG\\1X8dNhG\\1X8cNiG^1V8bNkG]1U8cNkG]1U8bNmG]1S8cNmG]1S8cNnG\\1R8cNoG]1Q8cNoG]1Q8cNPH\\1P8dNPH\\1P8cNRH\\1n7dNRH]1l7dNUH[1k7dNVH\\1j7dNWH[1i7eNWH[1i7eNWH[1i7dNYH[1g7eNYH[1g7eNZHZ1f7eN[H[1e7eN\\HZ1d7fN\\HZ1d7eN^H[1a7eN_H[1a7eN`HZ1`7fN`HZ1`7eNaH[1_7eNcHY1]7gNdHX1\\7gNfHW1[7iNfHV1Z7jNgHU1Y7jNiHU1W7kNjHS1W7mNjHR1V7nNlHP1T7oNnHo0S7QOnHn0R7ROoHm0Q7ROQIl0P7TOQIk0o6UORIj0n6VOTIg0m6XOUIg0k6YOVIf0j6ZOWIe0i6ZOYId0h6\\OYIc0g6]OZIb0f6]O]I`0d6@`G@Ho0h8A`GBGl0i8B`GDFh0k8C`GGDe0l8D`GICb0m8E`GJC`0m8EaGMA<o8G`GOA8o8I`G1@5P9IaG4^O1R9K`G6]ONS9KaG8\\OLS9LaG:[OHU9N`G<ZOEV9NaG?YOAU91bG`0XO]OW93aGb0WOZOX93bGd0VOXOX94bGf0UOUOY94cGi0SOQO[96bGk0ROnN\\97bGm0ROjN\\98cGP1POfN^9:bGQ1i9nNXFT1f9lNZFV1d9jN\\FX1b9gN_F[1_9eN`F^1^9bNbF_1]9`NdFb1Z9^NfFd1X9[NiFg1U9YNkFi1S9WNmFk1Q9TNPGm1o8SNQGo1m8QNSGQ2k8nMUGR2l8nMSGR2n8mMRGS2o8mMPGR2R9nMmFR2T9mMlFS2U9mMjFS2W9mMhFR2Z9mMfFS2[9mMdFS2]9lMcFT2^9lMaFS2a9mM^FS2j;0O10O1O1O1O1O1001O1O1O1O1O1O100000000000000000000001O3M7I7E;D<D<D<Ddd0"}, "label": "the bed that the dog is laying on"}]}
{"id": 396933, "image": "COCO_train2014_000000396933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black cat lying down next to pink towel\"."}], "task": "refering", "answer_template": "The \"black cat lying down next to pink towel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 25, 26, 27, 28, 33, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 232, 233, 234, 235, 256, 257, 258, 259], "bbox": [0.031265625000000005, 0.03003861003861004, 0.49153125, 0.6630888030888031], "iscrowd": 0, "area": 41063.972, "rle": {"size": [518, 640], "counts": "hY:3Q`04L5K5K5K4L5K5K5K5K4L5K5K5K4L5K5K5K4L5K5K5K4L5K5K5K4L5K5I6H7I7I7J6I7M3M3M3M3M4K4M3M3M3O1QOcIRH_6m7aISH`6l7`ISHb6l7_ISHc6k7]ITHe6j7\\IVHf6h7[IVHg6i7YIWHi6g7WIXHk6g7VIXHl6f7TIYHn6e7SIZHP7d7QI[HP7d7PI[HS7c7mH]HT7b7lH]HW7a7jH^HW7`7jH_HY7_7gHaHZ7^7gH`H[7_7eH_H_7_7aH^Hc7a7^H\\Hf7m6QHaI:_Oi7n6nGcIa8Z6aGeI`8X6cGgI_8V6bGjI_8T6cGkI_8R6cGmI^8P6dGoI_8n5cGQJ^8m5dGRJ]8k5fGUJY8j5jGUJU8k5lGUJS8k5oGTJP8m5QHRJn7o5RHQJm7o5UHPJj7Q6WHnIh7R6\\HkIc7V6`HhI^7X6X11O1O001O001O00100O001O001O001O1O001O001O0M4M3O0010O01O10O01O010O010O1O010O0010O0100O00100O00100O1O1_LQIjMo6Q2WImMj6o1[IRME]NP7^4`InLGbNj6[4fIjLGjNd6X4jIgLIoN^6U4mIgLLROX6S4oIfL0UOR6P4SJdL3ZOj5o3VJbL7]Od5l3ZJdL6^Oa5j3\\JhL5\\O`5h3^JkL5[O^5e3bJoL2ZO]5c3dJSM1WO\\5b3hJUM^7g2eHYM\\7b2iH\\MY7`2jH`MW7[2mHdMU7X2oHgMR7U2QIkMo6Q2VImMl6o1WIQNj6j1ZIUNh6g1\\IXNe6c1_I\\Nc6`1aI_N_6]1eIcN\\6Y1hIeNZ6Y1gIgNZ6W1gIhN[6W1jIdNW6Z1\\4O001N2O0O10000O100O10000O100O10O010000O100O10O10O01O010O10O010O0100O010O010O010O10O01O010O10O010O0100O010O010O010O10O010O1000O0100000O010000000O010000000O01000000O010000000O010000000O010000000O010000000O01002N2N2M3N2N2N1N3N2N2N2N2M3N2N2N2N2M3N]jT5"}, "label": "black cat lying down next to pink towel"}]}
{"id": 396933, "image": "COCO_train2014_000000396933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat on a bed\"."}], "task": "refering", "answer_template": "The \"a black cat on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 25, 26, 27, 28, 33, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 232, 233, 234, 235, 256, 257, 258, 259], "bbox": [0.031265625000000005, 0.03003861003861004, 0.49153125, 0.6630888030888031], "iscrowd": 0, "area": 41063.972, "rle": {"size": [518, 640], "counts": "hY:3Q`04L5K5K5K4L5K5K5K5K4L5K5K5K4L5K5K5K4L5K5K5K4L5K5K5K4L5K5I6H7I7I7J6I7M3M3M3M3M4K4M3M3M3O1QOcIRH_6m7aISH`6l7`ISHb6l7_ISHc6k7]ITHe6j7\\IVHf6h7[IVHg6i7YIWHi6g7WIXHk6g7VIXHl6f7TIYHn6e7SIZHP7d7QI[HP7d7PI[HS7c7mH]HT7b7lH]HW7a7jH^HW7`7jH_HY7_7gHaHZ7^7gH`H[7_7eH_H_7_7aH^Hc7a7^H\\Hf7m6QHaI:_Oi7n6nGcIa8Z6aGeI`8X6cGgI_8V6bGjI_8T6cGkI_8R6cGmI^8P6dGoI_8n5cGQJ^8m5dGRJ]8k5fGUJY8j5jGUJU8k5lGUJS8k5oGTJP8m5QHRJn7o5RHQJm7o5UHPJj7Q6WHnIh7R6\\HkIc7V6`HhI^7X6X11O1O001O001O00100O001O001O001O1O001O001O0M4M3O0010O01O10O01O010O010O1O010O0010O0100O00100O00100O1O1_LQIjMo6Q2WImMj6o1[IRME]NP7^4`InLGbNj6[4fIjLGjNd6X4jIgLIoN^6U4mIgLLROX6S4oIfL0UOR6P4SJdL3ZOj5o3VJbL7]Od5l3ZJdL6^Oa5j3\\JhL5\\O`5h3^JkL5[O^5e3bJoL2ZO]5c3dJSM1WO\\5b3hJUM^7g2eHYM\\7b2iH\\MY7`2jH`MW7[2mHdMU7X2oHgMR7U2QIkMo6Q2VImMl6o1WIQNj6j1ZIUNh6g1\\IXNe6c1_I\\Nc6`1aI_N_6]1eIcN\\6Y1hIeNZ6Y1gIgNZ6W1gIhN[6W1jIdNW6Z1\\4O001N2O0O10000O100O10000O100O10O010000O100O10O10O01O010O10O010O0100O010O010O010O10O01O010O10O010O0100O010O010O010O10O010O1000O0100000O010000000O010000000O01000000O010000000O010000000O010000000O010000000O01002N2N2M3N2N2N1N3N2N2N2N2M3N2N2N2N2M3N]jT5"}, "label": "a black cat on a bed"}]}
{"id": 36488, "image": "COCO_train2014_000000036488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person closest to the wall\"."}], "task": "refering", "answer_template": "The \"the person closest to the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 250, 251, 252], "bbox": [0.41859375, 0.003953488372093023, 1.0, 0.6901860465116278], "iscrowd": 0, "area": 57779.99640000001, "rle": {"size": [430, 640], "counts": "kc`32Z=4M2N100O1O100O1O100O100O101O000O100000000O1000000O10000O100O10000O100O100O10000O10000O100000000O100000000O1000000O101O00000001O000000001O000000001O000000001O0000O1000000000000O10002N1O2N1N3N1O:PHROl2o0RMROn2n0RMROn2n0QMSOn2n0QMSOo2m0QMSOo2m0PMUOo2k0PMVOP3j0PMVOP3j0oLWOP3j0PMVOP3j0oLWOQ3j0mLWOS3i0mLWOS3i0lLYOS3g0lLZOT3f0lLZOS3g0lLZOT3f0kL[OU3e0kL[OU3e0jL\\OV3d0iL]OW3c0iL^OU3d0iL]OW3c0hL^OX3b0hL^OX3b0gL_OY3a0fL@Z3`0fL@Y3a0fL@Z3`0fLAY3?fLBZ3>eLC[3>dLB\\3>cLC]3=bLD]3=cLC]3=bLD^3<aLF^3:bLF^3:aLG_39`LH_39aLG_39`LH`39^LHb38^LHb38]LJb36^LJa37^LJb36]LKc35]LKc35\\LLd34[LMe33WL1i3OSL6k3LPL8P4HlK<T4DiK?W4AfKb0Z4^OeKc0[4]OeKc0Z4^OfKb0Z4^OeKd0Z4\\OfKd0Z4\\OeKe0[4\\OdKd0\\4\\OcKe0\\4\\OdKd0\\4\\OdKd0\\4\\OcKe0]4[OcKf0\\4ZOcKg0]4YOcKg0]4YOcKg0\\4ZOcKg0]4YOcKg0]4ZOaKg0_4YOaKg0_4YO`Ki0_4WOaKi0_4WOaKi0^4XOaKi0_4WOaKi0_4WO`Kj0`4VO`Kj0`4VO_Kk0a4UO_Kk0a4VO^Kk0`4VO_Kk0a4UO_Kk0a4UO^Kl0b4TO^Kl0b4TO^Kl0b4TO]Km0c4SO]Km0c4SO\\Ko0b4RO^Kn0b4RO]Ko0c4QO]Ko0c4QO]Ko0c4QO\\KP1d4PO\\KP1d4PO[KQ1d4PO\\KQ1c4oN\\KR1d4nN\\KR1d4nN\\KR1d4nN[KS1e4mN[KS1e4mNZKT1f4lNZKT1e4mN[KT1d4lN[KU1e4kN[KU1e4kNZKV1f4jNZKV1f4jNYKW1g4iNYKW1f4jNZKV1f4jNYKX1f4hNZKX1f4hNYKY1g4gNYKY1g4gNXKZ1h4eNYK[1g4eNXK\\1g4eNYK[1g4eNXK]1g4cNXK^1h4bNWK_1i4aNVK`1j4`NVK`1j4`NUKa1j4`NUKa1k4^NUKc1k4]NTKe1k4[NUKe1k4[NTKf1l4ZNSKg1m4YNRKh1n4XNQKi1o4WNPKj1P5VNPKj1P5VNoJl1P5TNoJm1Q5RNoJo1Q5QNnJP2R5PNnJP2R5PNmJQ2S5oMlJR2T5nMkJS2U5mMjJU2T5lMkJU2U5kMkJU2T5lMkJU2U5kMjJV2U5jMkJW2T5jMkJW2U5iMkJW2T5jMkJX2S5iMlJX2T5gMlJZ2S5gMmJY2R5hMnJX2R5gMoJY2P5hMoJY2Q5gMoJY2P5hMPKY2n4gMSKY2m4gMSKY2l4hMTKX2k4hMVKX2j4hMVKX2i4iMWKW2i4hMWKY2h4hMXKY2f4hMZKX2f4gM[KY2d4hM\\KX2c4iM]KW2c4hM^KX2a4iM_KW2`4jM`KV2`4iM`KY2^4hMbKX2^4hMbKX2]4hMdKX2[4iMeKW2[4iMeKW2Z4jMfKV2Y4jMhKV2X4jMhKV2W4kMiKV2U4kMkKU2U4kMkKU2T4lMlKT2T4lMlKT2S4mMmKS2R4nMlKT2T4PNeKS2Z4VNZKo1d4^300O100O1O100O2N100O1O100000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0001O0000000001O000M"}, "label": "the person closest to the wall"}]}
{"id": 36488, "image": "COCO_train2014_000000036488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on far end of couch wearing white tshirt\"."}], "task": "refering", "answer_template": "The \"person on far end of couch wearing white tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 250, 251, 252], "bbox": [0.41859375, 0.003953488372093023, 1.0, 0.6901860465116278], "iscrowd": 0, "area": 57779.99640000001, "rle": {"size": [430, 640], "counts": "kc`32Z=4M2N100O1O100O1O100O100O101O000O100000000O1000000O10000O100O10000O100O100O10000O10000O100000000O100000000O1000000O101O00000001O000000001O000000001O000000001O0000O1000000000000O10002N1O2N1N3N1O:PHROl2o0RMROn2n0RMROn2n0QMSOn2n0QMSOo2m0QMSOo2m0PMUOo2k0PMVOP3j0PMVOP3j0oLWOP3j0PMVOP3j0oLWOQ3j0mLWOS3i0mLWOS3i0lLYOS3g0lLZOT3f0lLZOS3g0lLZOT3f0kL[OU3e0kL[OU3e0jL\\OV3d0iL]OW3c0iL^OU3d0iL]OW3c0hL^OX3b0hL^OX3b0gL_OY3a0fL@Z3`0fL@Y3a0fL@Z3`0fLAY3?fLBZ3>eLC[3>dLB\\3>cLC]3=bLD]3=cLC]3=bLD^3<aLF^3:bLF^3:aLG_39`LH_39aLG_39`LH`39^LHb38^LHb38]LJb36^LJa37^LJb36]LKc35]LKc35\\LLd34[LMe33WL1i3OSL6k3LPL8P4HlK<T4DiK?W4AfKb0Z4^OeKc0[4]OeKc0Z4^OfKb0Z4^OeKd0Z4\\OfKd0Z4\\OeKe0[4\\OdKd0\\4\\OcKe0\\4\\OdKd0\\4\\OdKd0\\4\\OcKe0]4[OcKf0\\4ZOcKg0]4YOcKg0]4YOcKg0\\4ZOcKg0]4YOcKg0]4ZOaKg0_4YOaKg0_4YO`Ki0_4WOaKi0_4WOaKi0^4XOaKi0_4WOaKi0_4WO`Kj0`4VO`Kj0`4VO_Kk0a4UO_Kk0a4VO^Kk0`4VO_Kk0a4UO_Kk0a4UO^Kl0b4TO^Kl0b4TO^Kl0b4TO]Km0c4SO]Km0c4SO\\Ko0b4RO^Kn0b4RO]Ko0c4QO]Ko0c4QO]Ko0c4QO\\KP1d4PO\\KP1d4PO[KQ1d4PO\\KQ1c4oN\\KR1d4nN\\KR1d4nN\\KR1d4nN[KS1e4mN[KS1e4mNZKT1f4lNZKT1e4mN[KT1d4lN[KU1e4kN[KU1e4kNZKV1f4jNZKV1f4jNYKW1g4iNYKW1f4jNZKV1f4jNYKX1f4hNZKX1f4hNYKY1g4gNYKY1g4gNXKZ1h4eNYK[1g4eNXK\\1g4eNYK[1g4eNXK]1g4cNXK^1h4bNWK_1i4aNVK`1j4`NVK`1j4`NUKa1j4`NUKa1k4^NUKc1k4]NTKe1k4[NUKe1k4[NTKf1l4ZNSKg1m4YNRKh1n4XNQKi1o4WNPKj1P5VNPKj1P5VNoJl1P5TNoJm1Q5RNoJo1Q5QNnJP2R5PNnJP2R5PNmJQ2S5oMlJR2T5nMkJS2U5mMjJU2T5lMkJU2U5kMkJU2T5lMkJU2U5kMjJV2U5jMkJW2T5jMkJW2U5iMkJW2T5jMkJX2S5iMlJX2T5gMlJZ2S5gMmJY2R5hMnJX2R5gMoJY2P5hMoJY2Q5gMoJY2P5hMPKY2n4gMSKY2m4gMSKY2l4hMTKX2k4hMVKX2j4hMVKX2i4iMWKW2i4hMWKY2h4hMXKY2f4hMZKX2f4gM[KY2d4hM\\KX2c4iM]KW2c4hM^KX2a4iM_KW2`4jM`KV2`4iM`KY2^4hMbKX2^4hMbKX2]4hMdKX2[4iMeKW2[4iMeKW2Z4jMfKV2Y4jMhKV2X4jMhKV2W4kMiKV2U4kMkKU2U4kMkKU2T4lMlKT2T4lMlKT2S4mMmKS2R4nMlKT2T4PNeKS2Z4VNZKo1d4^300O100O1O100O2N100O1O100000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0001O0000000001O000M"}, "label": "person on far end of couch wearing white tshirt"}]}
{"id": 36488, "image": "COCO_train2014_000000036488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foremost arm holding a wii controller\"."}], "task": "refering", "answer_template": "The \"the foremost arm holding a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.08621875, 0.5293488372093024, 1.0, 0.9595348837209303], "iscrowd": 0, "area": 51208.44204999999, "rle": {"size": [430, 640], "counts": "dWh01]=1O001O1O001O10O01O1O001O1O1O001O10O01O1O001O1O001O1O00100O001O1O1O1O001N2N2O1N1O2O1N2N2N101N2N2O0cNiNbFX1[9kNdFV1Y9nNgFR1V9QOjFn0U9SOlFl0T9UOlFj0T9VOmFj0R9WOnFh0R9XOnFh0R9XOoFg0Q9YOPGg0n8ZOSGe0m8[OTGd0k8]OVGb0j8^OZG?e8A_G;`8F`G:`8F`G:`8F`G;^8FbG:^8FbG:^8FbG:]8GcG:\\8FdG:\\8FdG:[8GeG9[8GeG:Z8FeG;Z8FeG;[8EdG<\\8DcG>[8CdG>\\8BdG>\\8BcG?\\8BcG`0\\8@cGa0]8_ObGb0]8_ObGb0^8^OaGd0^8\\OaGd0_8]O`Gd0`8\\O_Ge0a8[O^Gf0a8[O^Gf0b8ZO]Gf0d8ZO[Gg0d8ZO[Gg0e8YOZGh0f8XOYGi0g8WOXGj0g8WOXGi0i8WOVGj0j8VOUGk0k8UOTGl0k8UOTGl0l8TOSGm0m8SORGm0o8SOPGn0o8SOPGn0P9ROoFo0Q9QOnFP1R9POmF0^OMf93kFNBOb94lFKE0_95mFHG2\\96nFDI6Y96PG@J9U98RG\\OK;T99RGXON>P9:SGUOOa0n8:UGQO0d0j8<THFl7:SHHl78RHJn76PHLo75oGMQ83nGOQ81mG1S8OkG3T8NjG4V8LiG6V8JhG8X8HfG:Y8GfG:Z8FdG=[8CcG?]8AaGa0_8_O`Gb0`8^O^Ge0a8[O]Gg0c8YO[Gi0e8WO[Gi0e8WO[Gi0f8VOZGk0e8UO[Gk0e8UO[Gk0e8UO\\Gj0d8VO\\Gk0c8UO]GHXO3[95]G_OA<S94\\GXOId0k84\\GoN3m0`84^HLb74^HLb74^HLb74^HLb74^HLc73]HLd74\\HLd74]HJd76\\HJd76\\HJd76\\HIf76ZHJf76ZHJf76ZHIg77YHIg77YHHh78YHGg79YHGh78XHGi79WHGi79WHGi79WHFj7:VHFj7:VHEk7;VHDk7;UHEk7;UHDl7<THDl7<THCm7=SHCm7=SHCm7=SHBo7=QHCo7=RHBn7>RHAo7?QHAo7?QH@Q8?oGAQ8?oGAQ8?oG@R8`0nG@S8?nG_OS8a0lG@T8`0kGAU8?kG@V8`0iGAX8>gGCY8=fGC[8=dGD\\8<cGD_8;`GF`8:_GGa89_GFb8;\\GFe8<WGEi8=TGCm8`0oFAQ9b0eFD]9>SFOm93dE:\\:[1100O2O000O101O0O11O01O0001O01O000001O01O0000010O00000010O00000010O00000010O000001O01O0001O0001O0001O01O0001O0001O01O000001O0001O0001O0001O01O000000010O0000000010O000001O01O000001O0001O0001O000001O01O0000010O0000000010O000000010O00000001O01O0001O000001O01O001O01O01O001O00010O001O00010O001O000010O01O00001O010O00001O010O00001O0010O0001O001O00001O001O00001O001O00001O001O00001O001O00001O0000001O0000001O00000010O0001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O01O01O0000010O00000010O000001O0001O01O000010O000001O01O0001O0001O01O0000010O00001O01O0001O0001O01O0000010O0000001O01O000000001O000000001O000000001O000000001O000000001O000000001O0000000000001O000000000000001O000000000000001O000000000000001O000000000000001OfF"}, "label": "the foremost arm holding a wii controller"}]}
{"id": 36488, "image": "COCO_train2014_000000036488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"arm holding remote with bracelet\"."}], "task": "refering", "answer_template": "The \"arm holding remote with bracelet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.08621875, 0.5293488372093024, 1.0, 0.9595348837209303], "iscrowd": 0, "area": 51208.44204999999, "rle": {"size": [430, 640], "counts": "dWh01]=1O001O1O001O10O01O1O001O1O1O001O10O01O1O001O1O001O1O00100O001O1O1O1O001N2N2O1N1O2O1N2N2N101N2N2O0cNiNbFX1[9kNdFV1Y9nNgFR1V9QOjFn0U9SOlFl0T9UOlFj0T9VOmFj0R9WOnFh0R9XOnFh0R9XOoFg0Q9YOPGg0n8ZOSGe0m8[OTGd0k8]OVGb0j8^OZG?e8A_G;`8F`G:`8F`G:`8F`G;^8FbG:^8FbG:^8FbG:]8GcG:\\8FdG:\\8FdG:[8GeG9[8GeG:Z8FeG;Z8FeG;[8EdG<\\8DcG>[8CdG>\\8BdG>\\8BcG?\\8BcG`0\\8@cGa0]8_ObGb0]8_ObGb0^8^OaGd0^8\\OaGd0_8]O`Gd0`8\\O_Ge0a8[O^Gf0a8[O^Gf0b8ZO]Gf0d8ZO[Gg0d8ZO[Gg0e8YOZGh0f8XOYGi0g8WOXGj0g8WOXGi0i8WOVGj0j8VOUGk0k8UOTGl0k8UOTGl0l8TOSGm0m8SORGm0o8SOPGn0o8SOPGn0P9ROoFo0Q9QOnFP1R9POmF0^OMf93kFNBOb94lFKE0_95mFHG2\\96nFDI6Y96PG@J9U98RG\\OK;T99RGXON>P9:SGUOOa0n8:UGQO0d0j8<THFl7:SHHl78RHJn76PHLo75oGMQ83nGOQ81mG1S8OkG3T8NjG4V8LiG6V8JhG8X8HfG:Y8GfG:Z8FdG=[8CcG?]8AaGa0_8_O`Gb0`8^O^Ge0a8[O]Gg0c8YO[Gi0e8WO[Gi0e8WO[Gi0f8VOZGk0e8UO[Gk0e8UO[Gk0e8UO\\Gj0d8VO\\Gk0c8UO]GHXO3[95]G_OA<S94\\GXOId0k84\\GoN3m0`84^HLb74^HLb74^HLb74^HLb74^HLc73]HLd74\\HLd74]HJd76\\HJd76\\HJd76\\HIf76ZHJf76ZHJf76ZHIg77YHIg77YHHh78YHGg79YHGh78XHGi79WHGi79WHGi79WHFj7:VHFj7:VHEk7;VHDk7;UHEk7;UHDl7<THDl7<THCm7=SHCm7=SHCm7=SHBo7=QHCo7=RHBn7>RHAo7?QHAo7?QH@Q8?oGAQ8?oGAQ8?oG@R8`0nG@S8?nG_OS8a0lG@T8`0kGAU8?kG@V8`0iGAX8>gGCY8=fGC[8=dGD\\8<cGD_8;`GF`8:_GGa89_GFb8;\\GFe8<WGEi8=TGCm8`0oFAQ9b0eFD]9>SFOm93dE:\\:[1100O2O000O101O0O11O01O0001O01O000001O01O0000010O00000010O00000010O00000010O000001O01O0001O0001O0001O01O0001O0001O01O000001O0001O0001O0001O01O000000010O0000000010O000001O01O000001O0001O0001O000001O01O0000010O0000000010O000000010O00000001O01O0001O000001O01O001O01O01O001O00010O001O00010O001O000010O01O00001O010O00001O010O00001O0010O0001O001O00001O001O00001O001O00001O001O00001O001O00001O0000001O0000001O00000010O0001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O01O01O0000010O00000010O000001O0001O01O000010O000001O01O0001O0001O01O0000010O00001O01O0001O0001O01O0000010O0000001O01O000000001O000000001O000000001O000000001O000000001O000000001O0000000000001O000000000000001O000000000000001O000000000000001O000000000000001OfF"}, "label": "arm holding remote with bracelet"}]}
{"id": 36488, "image": "COCO_train2014_000000036488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wiimote held by a woman\"."}], "task": "refering", "answer_template": "The \"a wiimote held by a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 257, 259, 260, 261, 262], "bbox": [0.055453125000000006, 0.615046511627907, 0.42778124999999995, 0.768046511627907], "iscrowd": 0, "area": 11241.893099999994, "rle": {"size": [430, 640], "counts": "e\\?9T=9G7I2N2N1O2N1O1O1O100O1O1O1O1O1O1O2N1O1O1O1000000000000000001O0001O000000000000000001O00000000000000000000001O000000N2O1N2O1N2O1000000000000000000O10000000000003M3ZDcNU;_1dDfN\\;f10001O0000000000001O000000O1N2M3N2N2N2N2M3N2O1O1N2O1O1N2O1O1O1O1O1O1O1O10000000000O100000_DWO]:i0aEYO_:g0_E[Oa:d0]E@b:`0\\EBd:>YEEg:;WEGi:9TEKk:6RELn:4nD0R;0kD3U;NdD7];j03M3N2N2O1N2N2O1N2O2M2O1N2O100O1O1O1O1O100O1O100OaDTO^:l0`EVO`:j0^EXOb:h0\\EZOd:f0ZE\\Of:e0WE]Oi:d0TE^Ol:b0RE@n:a0oDBP;?mDCS;>jDDV;=gDEY;<dDF\\;V1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1N2N2O1N2O1N2N8I8G8H9H8G8Id\\i4"}, "label": "a wiimote held by a woman"}]}
{"id": 151178, "image": "COCO_train2014_000000151178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult giraffe staring into the eyes of a young giraffe\"."}], "task": "refering", "answer_template": "The \"an adult giraffe staring into the eyes of a young giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 106, 107, 108, 109, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 178, 179, 180, 181, 182, 183, 204, 205, 206], "bbox": [0.586515625, 0.22022916666666664, 1.0, 0.5101041666666666], "iscrowd": 0, "area": 18843.638349999997, "rle": {"size": [480, 640], "counts": "Zo_51k>5K6I6K6M2N3M2O2M2N3N1N001O01OmNjN]DW1c;jN[DV1e;kN[DU1e;lNYDT1h;mNVDS1k;nNTDR1k;oNTDQ1m;POQDQ1o;POPDo0P<ROoCo0Q<ROmCn0T<SOjCm0V<TOjCl0V<UOhCk0Y<VOcCm0\\<UO`Cm0a<SO\\Co0e<ROWCQ1i<b0010O00O2N1N3M2N3M2O2M3M3M3O1001N1000001O00001O0O10001O00001O0O10001O00001O0O10001O000dMiCo1d<01N2O1O1O1O001O001N1000010O01O01O01O01O01O010O00010O000010O01O000010O0001O001O01O01O000010O01O00001O01O01O001O01O01O00001O010O00001O01O01O001O00010O00001O010O00001O00001O001O00010O00001O001O00001O000001O01O000000001O000000001O01O000001O0000001O000000010O000000001O000000001O0001O0001O000000001O000001O01O0000000010O00000001O000001O01O0000000010O00000001O0001O0001O000000010oJ"}, "label": "an adult giraffe staring into the eyes of a young giraffe"}]}
{"id": 151178, "image": "COCO_train2014_000000151178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the right\"."}], "task": "refering", "answer_template": "The \"the giraffe on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 106, 107, 108, 109, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 178, 179, 180, 181, 182, 183, 204, 205, 206], "bbox": [0.586515625, 0.22022916666666664, 1.0, 0.5101041666666666], "iscrowd": 0, "area": 18843.638349999997, "rle": {"size": [480, 640], "counts": "Zo_51k>5K6I6K6M2N3M2O2M2N3N1N001O01OmNjN]DW1c;jN[DV1e;kN[DU1e;lNYDT1h;mNVDS1k;nNTDR1k;oNTDQ1m;POQDQ1o;POPDo0P<ROoCo0Q<ROmCn0T<SOjCm0V<TOjCl0V<UOhCk0Y<VOcCm0\\<UO`Cm0a<SO\\Co0e<ROWCQ1i<b0010O00O2N1N3M2N3M2O2M3M3M3O1001N1000001O00001O0O10001O00001O0O10001O00001O0O10001O000dMiCo1d<01N2O1O1O1O001O001N1000010O01O01O01O01O01O010O00010O000010O01O000010O0001O001O01O01O000010O01O00001O01O01O001O01O01O00001O010O00001O01O01O001O00010O00001O010O00001O00001O001O00010O00001O001O00001O000001O01O000000001O000000001O01O000001O0000001O000000010O000000001O000000001O0001O0001O000000001O000001O01O0000000010O00000001O000001O01O0000000010O00000001O0001O0001O000000010oJ"}, "label": "the giraffe on the right"}]}
{"id": 151178, "image": "COCO_train2014_000000151178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe ' s reflection\"."}], "task": "refering", "answer_template": "The \"giraffe ' s reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 126, 146, 147, 148, 149, 150, 168, 169, 170, 189, 190, 191, 192, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 323, 324, 325, 327, 328, 346, 347, 348, 350, 351, 370, 371, 373, 374], "bbox": [0.07246875, 0.2584166666666667, 0.5309062499999999, 0.9865208333333334], "iscrowd": 0, "area": 29209.029449999998, "rle": {"size": [480, 640], "counts": "[^f05h>5K5K5K4L2N1O1O00100VNUOaEo0[:XO_EX1o9oNlE`1d9gNUFh1\\9_N^Fb1`9dN[F]1b9kNWFW1g9oNTFR1j9UOPFl0n9[OlEf0Q:V2N2N0OK6K5J6N3n0SO3L1O1O100O1O100O1O100O^OgF_KX9\\4SG_Km8Z4`G`K_8Z4mGaKR8Z4]1I6J6N3N1O1O2N1O1O2N1O2N101O001O1N101O001O1N101O001O1N12O0O10000O100O100O01000O10O01000O0100O10O11N2O2M2N3L4M2N3L3N3L4M2M4M2N3L4M2M4M3M2]GVKl6l4PI[Kl6g4oHaKk6a4QIfKk6[4RIlKi6U4SIPLl6P4PIULn6l3nHXLR7h3kH[LU7e3gH_LX7b3dHbL\\7^3aHeL_7[3aHeL^7\\3bHeL]7[3cHeL]7[3cHeL\\7\\3dHdL\\7\\3dHdL[7^3dHbL\\7[3gHeLY7V3lHjLS7g2^ISMg6m2YI_L[7b3dHjKo7W4]11O0O2O0O2O000O2N1O00010O00001O001O00kN\\FoLe9d1UFiNb0XOX9n1ZFgNj0oNl8Z2]FcNj:]1YE`Ng:_1]E^Nc:b1`E[N`:e1cEXN]:g1gEUN[:j1gETNY:l1jEQNV:n1nEoMR:Q2QFkMP:U2SFhMm9W2WFfMi9Z2^100O2O000O10000O10001N10000O10000O100O100O100O1O010O100O1O100O100O00100O100O1O100O010O1O100O100O10O100000000000O010000000000O1D]N\\Cc1e<`NWC`1j<cNRC]1o<910000001O00001O000011N2N2O1N2N2O1N2N100O01O00010O00001O01O01OhNhBh0X=SOnBl0R=POSCP1_=O1O101N1O100O1N2O1O1O1O1O1O2N1O1O1O1O1N2O1G^^\\4"}, "label": "giraffe ' s reflection"}]}
{"id": 151178, "image": "COCO_train2014_000000151178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe that you can see their legs\"."}], "task": "refering", "answer_template": "The \"the giraffe that you can see their legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 126, 146, 147, 148, 149, 150, 168, 169, 170, 189, 190, 191, 192, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 323, 324, 325, 327, 328, 346, 347, 348, 350, 351, 370, 371, 373, 374], "bbox": [0.07246875, 0.2584166666666667, 0.5309062499999999, 0.9865208333333334], "iscrowd": 0, "area": 29209.029449999998, "rle": {"size": [480, 640], "counts": "[^f05h>5K5K5K4L2N1O1O00100VNUOaEo0[:XO_EX1o9oNlE`1d9gNUFh1\\9_N^Fb1`9dN[F]1b9kNWFW1g9oNTFR1j9UOPFl0n9[OlEf0Q:V2N2N0OK6K5J6N3n0SO3L1O1O100O1O100O1O100O^OgF_KX9\\4SG_Km8Z4`G`K_8Z4mGaKR8Z4]1I6J6N3N1O1O2N1O1O2N1O2N101O001O1N101O001O1N101O001O1N12O0O10000O100O100O01000O10O01000O0100O10O11N2O2M2N3L4M2N3L3N3L4M2M4M2N3L4M2M4M3M2]GVKl6l4PI[Kl6g4oHaKk6a4QIfKk6[4RIlKi6U4SIPLl6P4PIULn6l3nHXLR7h3kH[LU7e3gH_LX7b3dHbL\\7^3aHeL_7[3aHeL^7\\3bHeL]7[3cHeL]7[3cHeL\\7\\3dHdL\\7\\3dHdL[7^3dHbL\\7[3gHeLY7V3lHjLS7g2^ISMg6m2YI_L[7b3dHjKo7W4]11O0O2O0O2O000O2N1O00010O00001O001O00kN\\FoLe9d1UFiNb0XOX9n1ZFgNj0oNl8Z2]FcNj:]1YE`Ng:_1]E^Nc:b1`E[N`:e1cEXN]:g1gEUN[:j1gETNY:l1jEQNV:n1nEoMR:Q2QFkMP:U2SFhMm9W2WFfMi9Z2^100O2O000O10000O10001N10000O10000O100O100O100O1O010O100O1O100O100O00100O100O1O100O010O1O100O100O10O100000000000O010000000000O1D]N\\Cc1e<`NWC`1j<cNRC]1o<910000001O00001O000011N2N2O1N2N2O1N2N100O01O00010O00001O01O01OhNhBh0X=SOnBl0R=POSCP1_=O1O101N1O100O1N2O1O1O1O1O1O2N1O1O1O1O1N2O1G^^\\4"}, "label": "the giraffe that you can see their legs"}]}
{"id": 347787, "image": "COCO_train2014_000000347787.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red mug with a spoon in it\"."}], "task": "refering", "answer_template": "The \"the red mug with a spoon in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 165, 166, 167, 168, 182, 183, 184, 185, 186, 201, 202, 203, 204], "bbox": [0.15954, 0.63512, 0.38117999999999996, 0.91736], "iscrowd": 0, "area": 8712.2582, "rle": {"size": [375, 500], "counts": "nbm07Z;:I8J3M3M2N2N3N1N3M2N2N1O100O1AbNeF^1W9hNgFX1V9mNgFT1X9oNfFQ1Y9ROeFn0[9g0O1O1O1O1O1O1O1O010O1O101N101N2O1N101N2O0O2N101N101N101O001O0O2O001O001O0000000000000O100000000000000001N1000000O100QNdGb0]8\\OeGc0\\8\\OeGd0[8[OeGe0\\8ZOdGg0\\8XOeGg0\\8YOcGg0^8XObGi0_8UOaGk0`8TOaGk0`8TO`Gm0a8QO_Go0b8QO^Gn0c8QO]GP1d8nN\\GR1e8mN\\GR1e8mN[GT1f8jNZGV1g8iNYGX1h8gNXGX1i8gNWGY1j8fNVGY1m8eNTGX1o8gNQGX1Q9gNoFX1U9fNkFW1Y9gNgFX1\\9fNdFY1`9dN`F>`:@aE2Y;OcXa3"}, "label": "the red mug with a spoon in it"}]}
{"id": 347787, "image": "COCO_train2014_000000347787.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee\"."}], "task": "refering", "answer_template": "The \"a cup of coffee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 165, 166, 167, 168, 182, 183, 184, 185, 186, 201, 202, 203, 204], "bbox": [0.15954, 0.63512, 0.38117999999999996, 0.91736], "iscrowd": 0, "area": 8712.2582, "rle": {"size": [375, 500], "counts": "nbm07Z;:I8J3M3M2N2N3N1N3M2N2N1O100O1AbNeF^1W9hNgFX1V9mNgFT1X9oNfFQ1Y9ROeFn0[9g0O1O1O1O1O1O1O1O010O1O101N101N2O1N101N2O0O2N101N101N101O001O0O2O001O001O0000000000000O100000000000000001N1000000O100QNdGb0]8\\OeGc0\\8\\OeGd0[8[OeGe0\\8ZOdGg0\\8XOeGg0\\8YOcGg0^8XObGi0_8UOaGk0`8TOaGk0`8TO`Gm0a8QO_Go0b8QO^Gn0c8QO]GP1d8nN\\GR1e8mN\\GR1e8mN[GT1f8jNZGV1g8iNYGX1h8gNXGX1i8gNWGY1j8fNVGY1m8eNTGX1o8gNQGX1Q9gNoFX1U9fNkFW1Y9gNgFX1\\9fNdFY1`9dN`F>`:@aE2Y;OcXa3"}, "label": "a cup of coffee"}]}
{"id": 568974, "image": "COCO_train2014_000000568974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person wearing green cargo pants standing in front of a bench\"."}], "task": "refering", "answer_template": "The \"person wearing green cargo pants standing in front of a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 34, 35, 52, 53, 70, 71, 88, 89, 106, 107, 124, 125, 142, 143, 159, 160, 161, 177, 178, 179, 194, 195, 196, 197, 213, 214, 215], "bbox": [0.80034, 0.00021333333333333333, 1.0, 0.8973066666666667], "iscrowd": 0, "area": 18322.850350000004, "rle": {"size": [375, 500], "counts": "lhb44[;:L4N2N2N2N2N2N1O001O000O101O0000001O00001O0000001O00001O000O101QOmNXGS1f8POYGP1e8TOYGm0d8WOZGi0V1lN_5?YIe0V1nN_5a0YIa0V1PO_5c0YI>U1QOa5b0ZI=S1TOb5`0ZI<R1VOd5?YI;R1WOd5?ZI;Q1VOe1a0LO]M:R1WOY1j1YOeN[N;S1VOo0V2@[N]N:S1UOf0b2HPN]N:U1TO<n20dM_N:U1TO3Y36[M`N9W1TOIc3>PMbN9W1TO_Oo3e0fLcN8Y1SOVOZ4l0\\LdN8Y1ROQOb4o0ULfN8Z1ROmNf4P1RLhN6[1ROjNk4Q1mKiN7\\1QOhNP5o0[M9eMeNU5o0WM<`3@aL`0b3]O^Lc0e3YO]Lf0f3VO[Lj0h3SOXLm0k3oNVLQ1m3kNTLU1o3gNRLY1V701O00000O100000001O00000000001O00000000001O0000000000M3M3M3M3A?TLl3ULk3"}, "label": "person wearing green cargo pants standing in front of a bench"}]}
{"id": 568974, "image": "COCO_train2014_000000568974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid with green pants and brown shoes is standing up next to a bench\"."}], "task": "refering", "answer_template": "The \"a kid with green pants and brown shoes is standing up next to a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 34, 35, 52, 53, 70, 71, 88, 89, 106, 107, 124, 125, 142, 143, 159, 160, 161, 177, 178, 179, 194, 195, 196, 197, 213, 214, 215], "bbox": [0.80034, 0.00021333333333333333, 1.0, 0.8973066666666667], "iscrowd": 0, "area": 18322.850350000004, "rle": {"size": [375, 500], "counts": "lhb44[;:L4N2N2N2N2N2N1O001O000O101O0000001O00001O0000001O00001O000O101QOmNXGS1f8POYGP1e8TOYGm0d8WOZGi0V1lN_5?YIe0V1nN_5a0YIa0V1PO_5c0YI>U1QOa5b0ZI=S1TOb5`0ZI<R1VOd5?YI;R1WOd5?ZI;Q1VOe1a0LO]M:R1WOY1j1YOeN[N;S1VOo0V2@[N]N:S1UOf0b2HPN]N:U1TO<n20dM_N:U1TO3Y36[M`N9W1TOIc3>PMbN9W1TO_Oo3e0fLcN8Y1SOVOZ4l0\\LdN8Y1ROQOb4o0ULfN8Z1ROmNf4P1RLhN6[1ROjNk4Q1mKiN7\\1QOhNP5o0[M9eMeNU5o0WM<`3@aL`0b3]O^Lc0e3YO]Lf0f3VO[Lj0h3SOXLm0k3oNVLQ1m3kNTLU1o3gNRLY1V701O00000O100000001O00000000001O00000000001O0000000000M3M3M3M3A?TLl3ULk3"}, "label": "a kid with green pants and brown shoes is standing up next to a bench"}]}
{"id": 568974, "image": "COCO_train2014_000000568974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person behind the bench by the coffee cup\"."}], "task": "refering", "answer_template": "The \"the person behind the bench by the coffee cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 36, 37, 54, 55, 91, 92, 108, 109, 110, 126, 127, 128, 144, 162], "bbox": [0.0, 0.00112, 0.16386, 0.7083466666666667], "iscrowd": 0, "area": 9403.1538, "rle": {"size": [375, 500], "counts": "2Q1h5=hMd1ZOYOf0m0TOROm0n0ROSOn0m0ROROo0m0QOTOo0l0QOSOP1m0oNTOQ1k0POTOQ1l0oNTOQ1l0nNTOS1k0nNUOR1k0mNUOT1k0lNTOU1k0kNVOU1j0kNUOV1j0jNWOV1i0jNVOW1j0iNVOW1i0iNWOX1i0hNoN`1Q1_NiNh1V1YNcNn1]1QNaNR2_1nM`NS2_1mMaNT2_1lM`NU2`1kM`NU2_1kMaNV2_1jMaNV2_1iMbNW2]1jMbNV2_1iMbNW2]1jMcNV2]1jMbNW2^1hMcNX2\\1iMdNW2\\1hMeNX2[1hMdNY2[1gMfNY2Z1gMfNY2Z1fMfN[2Y1fMgNZ2Y1fMgNZ29XKSO]2d0\\28[KPOZ2h0[26`KoNT2l0\\24dKlNQ2P1[2cNWKW1a0TOl1S1\\2cNZKQ1b0VOh1W1[2bN`Kj0b0YOc1\\1[2bNdKc0a0]O`1^1[2bNiK<a0A[1a1[2cNmK5a0CW1g1Z2aNULLl2c1o0bN2_1NaN2_1NbN0`1O`N1a1N`N1a1N_N1c1N^N1b1O_NOc10]N0d0cKG\\4FO=lKNT4FO6TL4l3FO0]L9d3HNHeL`0]3ILBnLc0W3K[14eN]O_K<k57fN@`K6j59fND`K1i5:hNGaKKh5=gNKaKFh5>gNNbKBf5`0hN1]1McN2`1KbN1b1LaN0d1L_N0f1K^N1TYi4"}, "label": "the person behind the bench by the coffee cup"}]}
{"id": 568974, "image": "COCO_train2014_000000568974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing behind a coffee cup\"."}], "task": "refering", "answer_template": "The \"a man standing behind a coffee cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 36, 37, 54, 55, 91, 92, 108, 109, 110, 126, 127, 128, 144, 162], "bbox": [0.0, 0.00112, 0.16386, 0.7083466666666667], "iscrowd": 0, "area": 9403.1538, "rle": {"size": [375, 500], "counts": "2Q1h5=hMd1ZOYOf0m0TOROm0n0ROSOn0m0ROROo0m0QOTOo0l0QOSOP1m0oNTOQ1k0POTOQ1l0oNTOQ1l0nNTOS1k0nNUOR1k0mNUOT1k0lNTOU1k0kNVOU1j0kNUOV1j0jNWOV1i0jNVOW1j0iNVOW1i0iNWOX1i0hNoN`1Q1_NiNh1V1YNcNn1]1QNaNR2_1nM`NS2_1mMaNT2_1lM`NU2`1kM`NU2_1kMaNV2_1jMaNV2_1iMbNW2]1jMbNV2_1iMbNW2]1jMcNV2]1jMbNW2^1hMcNX2\\1iMdNW2\\1hMeNX2[1hMdNY2[1gMfNY2Z1gMfNY2Z1fMfN[2Y1fMgNZ2Y1fMgNZ29XKSO]2d0\\28[KPOZ2h0[26`KoNT2l0\\24dKlNQ2P1[2cNWKW1a0TOl1S1\\2cNZKQ1b0VOh1W1[2bN`Kj0b0YOc1\\1[2bNdKc0a0]O`1^1[2bNiK<a0A[1a1[2cNmK5a0CW1g1Z2aNULLl2c1o0bN2_1NaN2_1NbN0`1O`N1a1N`N1a1N_N1c1N^N1b1O_NOc10]N0d0cKG\\4FO=lKNT4FO6TL4l3FO0]L9d3HNHeL`0]3ILBnLc0W3K[14eN]O_K<k57fN@`K6j59fND`K1i5:hNGaKKh5=gNKaKFh5>gNNbKBf5`0hN1]1McN2`1KbN1b1LaN0d1L_N0f1K^N1TYi4"}, "label": "a man standing behind a coffee cup"}]}
{"id": 134799, "image": "COCO_train2014_000000134799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a silver ring is holding a phone\"."}], "task": "refering", "answer_template": "The \"a man with a silver ring is holding a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 85, 86, 87, 88, 89, 91, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 248, 249, 250, 251, 252, 259, 260, 261, 262, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 294, 295, 296, 297, 298, 304, 305, 306, 307, 308, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344], "bbox": [0.200984375, 0.15280660377358493, 1.0, 0.9865094339622642], "iscrowd": 0, "area": 91075.48795000001, "rle": {"size": [424, 640], "counts": "Vje11V=1N2O1O2M2O1O1N2O1O1N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N3L3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2H8]Oc0^Ob0]Oc0]Oc0]Oc0]Oc0]Oc0I7O1N2O2N1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1O100O100O1O100OO1N3N1O2M3N2N2M3N1O2M3N2N2M3N2N2M2O2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N1O2M3N2N2M3N2N2M3N1O2M3N2O1N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2N101N2N2N2N2N101N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2L5K5K5J6K5K5J6K5K5J6K4Li1VN^PR1SO`PnN6J5L4K5K5K5K6K4K5K5K5L5J5K5K5L4K6J5K5J6K5K6J5K5K5K5K5K6J5K5K5L4O2N100O1O1O2N100O1O1O2O0O1O1O1O101N1O1O100O2N1O1O100O2N1O100O1O2N1O100O1O1O2O0O1O1O1O2O0O1O1O101N1O1O1O101N1O1O100O1O2N1O10;DU1kNU1kNc0^OO0000001O0000010O0000001O0000001O0001O01O0000001O0000001O101N1O2N1O2N2N1O2N1O2O1N1O2N1O2N1O2N1O1O2N100O2N1O1O2N1O0000O10001O000000000000000000001O0000000000000000000O2N100O1O1O100O1O1O100O2N1O100O1O1O1O100O1O1O010O000010eJkMB[2HXOUOQ11d0lNn7mNhC"}, "label": "a man with a silver ring is holding a phone"}]}
{"id": 134799, "image": "COCO_train2014_000000134799.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark skinned person holding a cell phone recording a speach\"."}], "task": "refering", "answer_template": "The \"dark skinned person holding a cell phone recording a speach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 85, 86, 87, 88, 89, 91, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 248, 249, 250, 251, 252, 259, 260, 261, 262, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 294, 295, 296, 297, 298, 304, 305, 306, 307, 308, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344], "bbox": [0.200984375, 0.15280660377358493, 1.0, 0.9865094339622642], "iscrowd": 0, "area": 91075.48795000001, "rle": {"size": [424, 640], "counts": "Vje11V=1N2O1O2M2O1O1N2O1O1N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N3L3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2H8]Oc0^Ob0]Oc0]Oc0]Oc0]Oc0]Oc0I7O1N2O2N1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1O100O100O1O100OO1N3N1O2M3N2N2M3N1O2M3N2N2M3N2N2M2O2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N1O2M3N2N2M3N2N2M3N1O2M3N2O1N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2N101N2N2N2N2N101N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2L5K5K5J6K5K5J6K5K5J6K4Li1VN^PR1SO`PnN6J5L4K5K5K5K6K4K5K5K5L5J5K5K5L4K6J5K5J6K5K6J5K5K5K5K5K6J5K5K5L4O2N100O1O1O2N100O1O1O2O0O1O1O1O101N1O1O100O2N1O1O100O2N1O100O1O2N1O100O1O1O2O0O1O1O1O2O0O1O1O101N1O1O1O101N1O1O100O1O2N1O10;DU1kNU1kNc0^OO0000001O0000010O0000001O0000001O0001O01O0000001O0000001O101N1O2N1O2N2N1O2N1O2O1N1O2N1O2N1O2N1O1O2N100O2N1O1O2N1O0000O10001O000000000000000000001O0000000000000000000O2N100O1O1O100O1O1O100O2N1O100O1O1O1O100O1O1O010O000010eJkMB[2HXOUOQ11d0lNn7mNhC"}, "label": "dark skinned person holding a cell phone recording a speach"}]}
{"id": 183953, "image": "COCO_train2014_000000183953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver lifeboat\"."}], "task": "refering", "answer_template": "The \"a silver lifeboat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [333, 334, 335, 336, 337, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 386, 387, 388, 389, 390], "bbox": [0.627375, 0.8480624999999999, 1.0, 0.99809375], "iscrowd": 0, "area": 10514.025699999993, "rle": {"size": [640, 480], "counts": "SYm51nc03N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N1O2M2O1O2N1O2N1O2N1O00O1N2M3N2M3L4L4J6M3O2N1N2O1N2O100O100O1O10000010[^OPOX`0P1d_OTO\\`0l0c_OUO]`0k0b_OVO^`0k0__OWOa`0i0^_OXOb`0i0\\_OXOd`0i0Y_OYOg`0g0X_OYOi`0h0S_O[Om`0e0Q_O]Oo`0d0m^O_OSa0b0i^OAXa0>f^ODZa0=b^OF^a0^100000O2O0000001O00001O0000010O01O102M3M3M01O000010O000001N10001O0000001O0O101N1N2N3L4M3O2N1O1O1O1O00N2N2O1000000000000000O1000000000001O00000000001O1O1O2N1O1O1O2N1O1O2N1O1Oc0]O[]O"}, "label": "a silver lifeboat"}]}
{"id": 183953, "image": "COCO_train2014_000000183953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small white boat in the water next to the large one the woman is standing on\"."}], "task": "refering", "answer_template": "The \"the small white boat in the water next to the large one the woman is standing on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [333, 334, 335, 336, 337, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 386, 387, 388, 389, 390], "bbox": [0.627375, 0.8480624999999999, 1.0, 0.99809375], "iscrowd": 0, "area": 10514.025699999993, "rle": {"size": [640, 480], "counts": "SYm51nc03N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N1O2M2O1O2N1O2N1O2N1O00O1N2M3N2M3L4L4J6M3O2N1N2O1N2O100O100O1O10000010[^OPOX`0P1d_OTO\\`0l0c_OUO]`0k0b_OVO^`0k0__OWOa`0i0^_OXOb`0i0\\_OXOd`0i0Y_OYOg`0g0X_OYOi`0h0S_O[Om`0e0Q_O]Oo`0d0m^O_OSa0b0i^OAXa0>f^ODZa0=b^OF^a0^100000O2O0000001O00001O0000010O01O102M3M3M01O000010O000001N10001O0000001O0O101N1N2N3L4M3O2N1O1O1O1O00N2N2O1000000000000000O1000000000001O00000000001O1O1O2N1O1O1O2N1O1O2N1O1Oc0]O[]O"}, "label": "the small white boat in the water next to the large one the woman is standing on"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a houseplant to the right of a television\"."}], "task": "refering", "answer_template": "The \"a houseplant to the right of a television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 44, 45, 46, 47, 48, 49, 50, 61, 62, 63, 64, 66, 67, 80, 81, 82, 83, 84, 99, 100, 101, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 235, 236], "bbox": [0.617542735042735, 0.002234375, 1.0, 0.57128125], "iscrowd": 0, "area": 35147.8162, "rle": {"size": [640, 468], "counts": "`ed54ic09G9G9G9G9G9G5K4L3M4M2M4L3N3L3M4N101O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O00OTD[L_7b3cHcLY7[3hHkLS7S3oHPMn6m2TITMnK^O^:\\3eIWMkKBW1Ka7Z3_KYMgKGQ1Lh7Q3bK]McKKl0Ln7j2dK`M`KOf0MU8b2fKcM]K1c0OY8Z2jKfMXK3`03]8R2lKiMUK5=5a8k1nKjN^K_Oc8d1RLmNWKFd8[1VLPORKIi8S1]K]M`L]3R8TO_KgMXLU3Y8RO_KmMULS3[8mNbKRNQLR3]8jNcKVNnKR3^8fNdK[NlKP3`8bNfK`NiKo2`8_NgKeNgKm2b8\\NhKiNdKm2c8WNkKnN`Kl2e8TNkKn2T4PMmKQ3S4lLnKW3Q4gLPLZ3P4fLoK\\3P4eLmK\\3T4fLiK[3W4fLfK\\3Z4eLdK\\3\\4fLaK[3_4fL^K\\3b4eLXK`3h4aLQKe3o4]LjJh3V5YLcJm3]5TL]JQ4c5QLUJU4k5lKnIZ4R6gKhI^4X6dK`Ib4`6_KZIf4f6[KSIk4m6VKlHo4U7SKdHR5\\7oJ]HW5c7jJWH[5i7gJoG_5Q8bJhGd5X8]JbGh5^8m20000000000001O00000O100`FZGe7g8XH\\Gg7e8WH]Gh7d8UH_Gi7c8UH_Gj7b8SHaGk7a8SHbGk7_8RHdGl7^8RHdGm7^8oGeGP8\\8nGfGP8\\8mGgGR8Z8lGhGR8Z8kGiGT8X8jGjGU8W8hGmGU8U8iGmGV8T8gGoGW8S8gGoGX8S8dGPH[8Q8cGQH[8Q8bGRH]8i9N2O2N1N2O1N2O1O1N2O1N3N1O1N2O1N2O1\\Od0jNUO"}, "label": "a houseplant to the right of a television"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plant with the most leafs on the right\"."}], "task": "refering", "answer_template": "The \"the plant with the most leafs on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 44, 45, 46, 47, 48, 49, 50, 61, 62, 63, 64, 66, 67, 80, 81, 82, 83, 84, 99, 100, 101, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 235, 236], "bbox": [0.617542735042735, 0.002234375, 1.0, 0.57128125], "iscrowd": 0, "area": 35147.8162, "rle": {"size": [640, 468], "counts": "`ed54ic09G9G9G9G9G9G5K4L3M4M2M4L3N3L3M4N101O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O00OTD[L_7b3cHcLY7[3hHkLS7S3oHPMn6m2TITMnK^O^:\\3eIWMkKBW1Ka7Z3_KYMgKGQ1Lh7Q3bK]McKKl0Ln7j2dK`M`KOf0MU8b2fKcM]K1c0OY8Z2jKfMXK3`03]8R2lKiMUK5=5a8k1nKjN^K_Oc8d1RLmNWKFd8[1VLPORKIi8S1]K]M`L]3R8TO_KgMXLU3Y8RO_KmMULS3[8mNbKRNQLR3]8jNcKVNnKR3^8fNdK[NlKP3`8bNfK`NiKo2`8_NgKeNgKm2b8\\NhKiNdKm2c8WNkKnN`Kl2e8TNkKn2T4PMmKQ3S4lLnKW3Q4gLPLZ3P4fLoK\\3P4eLmK\\3T4fLiK[3W4fLfK\\3Z4eLdK\\3\\4fLaK[3_4fL^K\\3b4eLXK`3h4aLQKe3o4]LjJh3V5YLcJm3]5TL]JQ4c5QLUJU4k5lKnIZ4R6gKhI^4X6dK`Ib4`6_KZIf4f6[KSIk4m6VKlHo4U7SKdHR5\\7oJ]HW5c7jJWH[5i7gJoG_5Q8bJhGd5X8]JbGh5^8m20000000000001O00000O100`FZGe7g8XH\\Gg7e8WH]Gh7d8UH_Gi7c8UH_Gj7b8SHaGk7a8SHbGk7_8RHdGl7^8RHdGm7^8oGeGP8\\8nGfGP8\\8mGgGR8Z8lGhGR8Z8kGiGT8X8jGjGU8W8hGmGU8U8iGmGV8T8gGoGW8S8gGoGX8S8dGPH[8Q8cGQH[8Q8bGRH]8i9N2O2N1N2O1N2O1O1N2O1N3N1O1N2O1N2O1\\Od0jNUO"}, "label": "the plant with the most leafs on the right"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall plant without leaves\"."}], "task": "refering", "answer_template": "The \"a tall plant without leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 2, 17, 18, 19, 35, 36, 52, 53, 69, 70, 87, 104, 121, 138, 154, 155, 171, 172, 173, 188, 189, 190, 205, 206], "bbox": [0.02252136752136752, 0.00940625, 0.20474358974358972, 0.5599999999999999], "iscrowd": 0, "area": 6360.924499999998, "rle": {"size": [640, 468], "counts": "Wl62oc02O1N3M2O1N2N3N1N2M3N3M2N2N2M4M2YDb0h2_OUMa0m2@PMa0P3AmL?U3AhLa0Y3@dLa0\\3@bL`0`3A]L`0f3^OXLb0l3]OQLd0Q4[OlKg0W4XOfKh0i4POPKQ1X5VOXJk0Q6nNlIR1\\6gNbI[1_6eN_I[1d6cN[I^1g6aNWI`1k6^NTIb1o6]NoHc1S7\\NkHe1X7ZNfHf1]7XNbHh1a7WN]Hi1f7UNYHk1i7UNeG[2^8cMjFT3Y9kLnEl3U:RLTEn1UN;h<gM[D`2VOE[<QNgCk28oNn;ZNaCb0[1A]Ob1c;aNaC[ONg0]3]1Q9eNaCWO=4Z3P2d8kNQHU1l7lNVHT1f7nN\\HR1`7QOaHo0\\7ROfHn0V7TOlHl0P7WORIh0h6fNPJZ1k5bN^J]1^5]NmJb1Q3QNdHa0]4]1P3ZNVH<l4Y1n2dNiG5\\5V1l2<WMBi2=YMBg2:^MEc25cMJ]21iMNa2BfM=o:N3M2O1N2N2O1fN`]OR1cb0mNg]Og0[b0YOn]O=Sb0CW^O1ka0NT100000kjX7"}, "label": "a tall plant without leaves"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant on the left side which is near the door\"."}], "task": "refering", "answer_template": "The \"plant on the left side which is near the door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 2, 17, 18, 19, 35, 36, 52, 53, 69, 70, 87, 104, 121, 138, 154, 155, 171, 172, 173, 188, 189, 190, 205, 206], "bbox": [0.02252136752136752, 0.00940625, 0.20474358974358972, 0.5599999999999999], "iscrowd": 0, "area": 6360.924499999998, "rle": {"size": [640, 468], "counts": "Wl62oc02O1N3M2O1N2N3N1N2M3N3M2N2N2M4M2YDb0h2_OUMa0m2@PMa0P3AmL?U3AhLa0Y3@dLa0\\3@bL`0`3A]L`0f3^OXLb0l3]OQLd0Q4[OlKg0W4XOfKh0i4POPKQ1X5VOXJk0Q6nNlIR1\\6gNbI[1_6eN_I[1d6cN[I^1g6aNWI`1k6^NTIb1o6]NoHc1S7\\NkHe1X7ZNfHf1]7XNbHh1a7WN]Hi1f7UNYHk1i7UNeG[2^8cMjFT3Y9kLnEl3U:RLTEn1UN;h<gM[D`2VOE[<QNgCk28oNn;ZNaCb0[1A]Ob1c;aNaC[ONg0]3]1Q9eNaCWO=4Z3P2d8kNQHU1l7lNVHT1f7nN\\HR1`7QOaHo0\\7ROfHn0V7TOlHl0P7WORIh0h6fNPJZ1k5bN^J]1^5]NmJb1Q3QNdHa0]4]1P3ZNVH<l4Y1n2dNiG5\\5V1l2<WMBi2=YMBg2:^MEc25cMJ]21iMNa2BfM=o:N3M2O1N2N2O1fN`]OR1cb0mNg]Og0[b0YOn]O=Sb0CW^O1ka0NT100000kjX7"}, "label": "plant on the left side which is near the door"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a living room chair\"."}], "task": "refering", "answer_template": "The \"a living room chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.7117094017094017, 0.61290625, 1.0, 0.98709375], "iscrowd": 0, "area": 25817.885200000004, "rle": {"size": [640, 468], "counts": "\\e`63lc07I7I8I5J3M2O2M3QMlNmBW1m<mNSCU1g<POWCS1c<QO\\CR1^<RObCP1X<TOgCo0S<UOlCn0n;WOPDl0j;XOVDi0e;[OZDh0`;\\O_Dg0[;]OeDd0X;^OgDd0X;]OfDd0Y;]OgDd0X;\\OgDe0X;\\OgDf0X;ZOgDh0X;YOgDg0X;ZOhDg0W;YOiDh0U;YOkDg0U;\\OiDd0U;_OiDb0V;AgD?X;DfD=Y;FeD:Y;IeD7[;KcD6[;NbD3];ObD0^;3_DN_;5_DL`;7]DIb;:]DFb;<\\DDc;`0ZDAe;a0YD@e;d0YD[Og;g0WDZOg;j0VDWOi;k0cCG]<<]CHa<l300O100O100O100O10000O1000000001O0000000000000000000000000000001O0000000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O2N2N1\\KQH"}, "label": "a living room chair"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty , modern , futuristic - designed chair\"."}], "task": "refering", "answer_template": "The \"an empty , modern , futuristic - designed chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.7117094017094017, 0.61290625, 1.0, 0.98709375], "iscrowd": 0, "area": 25817.885200000004, "rle": {"size": [640, 468], "counts": "\\e`63lc07I7I8I5J3M2O2M3QMlNmBW1m<mNSCU1g<POWCS1c<QO\\CR1^<RObCP1X<TOgCo0S<UOlCn0n;WOPDl0j;XOVDi0e;[OZDh0`;\\O_Dg0[;]OeDd0X;^OgDd0X;]OfDd0Y;]OgDd0X;\\OgDe0X;\\OgDf0X;ZOgDh0X;YOgDg0X;ZOhDg0W;YOiDh0U;YOkDg0U;\\OiDd0U;_OiDb0V;AgD?X;DfD=Y;FeD:Y;IeD7[;KcD6[;NbD3];ObD0^;3_DN_;5_DL`;7]DIb;:]DFb;<\\DDc;`0ZDAe;a0YD@e;d0YD[Og;g0WDZOg;j0VDWOi;k0cCG]<<]CHa<l300O100O100O100O10000O1000000001O0000000000000000000000000000001O0000000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O2N2N1\\KQH"}, "label": "an empty , modern , futuristic - designed chair"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small black desk chair a man is sitting in\"."}], "task": "refering", "answer_template": "The \"small black desk chair a man is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 323, 324, 340, 341, 357, 358, 359, 360, 374, 375, 376, 377, 378], "bbox": [0.0005982905982905984, 0.8054218750000001, 0.281517094017094, 1.0], "iscrowd": 0, "area": 7033.732649999996, "rle": {"size": [640, 468], "counts": "^b0<cb0T1O2H8I8G:G8G:G5FPM\\_OT3c`05O1O100O1N2O1O1O1O1O1O1O1O100001O1O1O001O1O1O3M4L4L4L4M3L4L4L4L8Hc0]Ob0^O5K00000000O2O0000000000001O00000000000000000O1M3N2N2N2M3N2N2M3N2O11O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O0O2O1O1O001O1O1O1O001O1O1O001O1O1O00R\\a6"}, "label": "small black desk chair a man is sitting in"}]}
{"id": 323218, "image": "COCO_train2014_000000323218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair with a man in white shirt sitting on it\"."}], "task": "refering", "answer_template": "The \"a black chair with a man in white shirt sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 323, 324, 340, 341, 357, 358, 359, 360, 374, 375, 376, 377, 378], "bbox": [0.0005982905982905984, 0.8054218750000001, 0.281517094017094, 1.0], "iscrowd": 0, "area": 7033.732649999996, "rle": {"size": [640, 468], "counts": "^b0<cb0T1O2H8I8G:G8G:G5FPM\\_OT3c`05O1O100O1N2O1O1O1O1O1O1O1O100001O1O1O001O1O1O3M4L4L4L4M3L4L4L4L8Hc0]Ob0^O5K00000000O2O0000000000001O00000000000000000O1M3N2N2N2M3N2N2M3N2O11O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O0O2O1O1O001O1O1O1O001O1O1O001O1O1O00R\\a6"}, "label": "a black chair with a man in white shirt sitting on it"}]}
{"id": 347796, "image": "COCO_train2014_000000347796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the macbook pro next to the cup\"."}], "task": "refering", "answer_template": "The \"the back of the macbook pro next to the cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 139, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216], "bbox": [0.07719038817005545, 0.34471875, 0.40094269870609983, 0.528015625], "iscrowd": 0, "area": 16273.185000000005, "rle": {"size": [640, 541], "counts": "n^j0=cc0c0]Ob0^Oc0]Ob0^Oa0_O0000000000000000000000001O0000000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000O1000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O0000000000000000000000000000000000001O0000000000YMa_Om1_`0SNb_Ol1^`0SNe_Ok1[`0UNg_Oi1Y`0WNi_Og1W`0YNk_Oe1U`0[Nl_Od1T`0\\Nn_Ob1R`0^NP@`1P`0`NR@^1n?bNS@]1m?cNU@[1k?eNW@Z1h?fNZ@X1k?cNV@\\1o?_NS@_1R`0\\NP@b1U`0YNT@^1aa0F9G:Fg\\O]O_c0<8IM2N3M3N1100O10000O100O10000O100O5LmUZ6"}, "label": "the back of the macbook pro next to the cup"}]}
{"id": 347796, "image": "COCO_train2014_000000347796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop whose screen we can see\"."}], "task": "refering", "answer_template": "The \"laptop whose screen we can see\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 197, 198, 199, 200, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 255, 256, 257], "bbox": [0.36502772643253234, 0.41215624999999995, 0.6660998151571165, 0.583328125], "iscrowd": 0, "area": 10279.832, "rle": {"size": [640, 541], "counts": "`Pl32nc03M2N3M3M3M3M2N3L4M3M2N3M3M3M3M2N3M3L4M2N3M3M3M3M2N3M3M3L3N3M3M3M001O0000000O101O00000000001O000O1000001O00000000001O0O100000001O000000000O2O00000000001O00000O10001O0000000002O0O2N2N2N2N1O2M3N2N1N3N2M3N1O2M3N2M3N1O2M3N2N1N3N2M3N2N1N3N2M3N1O2M3N2M3N1O2M101N10001N101N10001N100O2O000O2O0O2O000O2O0O101O0O101N10001N101O0O10W]a3"}, "label": "laptop whose screen we can see"}]}
{"id": 347796, "image": "COCO_train2014_000000347796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small black laptop\"."}], "task": "refering", "answer_template": "The \"small black laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 197, 198, 199, 200, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 255, 256, 257], "bbox": [0.36502772643253234, 0.41215624999999995, 0.6660998151571165, 0.583328125], "iscrowd": 0, "area": 10279.832, "rle": {"size": [640, 541], "counts": "`Pl32nc03M2N3M3M3M3M2N3L4M3M2N3M3M3M3M2N3M3L4M2N3M3M3M3M2N3M3M3L3N3M3M3M001O0000000O101O00000000001O000O1000001O00000000001O0O100000001O000000000O2O00000000001O00000O10001O0000000002O0O2N2N2N2N1O2M3N2N1N3N2M3N1O2M3N2M3N1O2M3N2N1N3N2M3N2N1N3N2M3N1O2M3N2M3N1O2M101N10001N101N10001N100O2O000O2O0O2O000O2O0O101O0O101N10001N101O0O10W]a3"}, "label": "small black laptop"}]}
{"id": 347796, "image": "COCO_train2014_000000347796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a closeup of a apple laptop\"."}], "task": "refering", "answer_template": "The \"a closeup of a apple laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 253, 254, 255, 256, 257, 258, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427], "bbox": [0.0, 0.582015625, 0.6194085027726433, 0.9842656250000001], "iscrowd": 0, "area": 73350.74919999999, "rle": {"size": [640, 541], "counts": "^<c2]a0e4[K00000000000O10000NhH\\CX7d<2000000000O100000000000O1000000000000000000000000O100000000000000000000000O01000000000000000000000000O1000000000000000000000000O1000000000O10000000000000O1000000000000000000000000O1000000000000000000000O10O1000000000000000000000000O1000000000000000000000000O100000O100000000000000000O1000000000000000000000000O100000000000000000O100000O1000000000000000000000000O1000000000000000000000000O1000O10000000000000000000O10000000000000000000000O100000000000000000O100000O1000000000000000000000000O1000000000000000000000000O1000O10000N2L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L4L4L4L3N3L4L4L4L4L4L4L4L4L4L4L4L3M4L5K4L5K4L4L5K4L5L3L4L5K4L5K4L5K4L4L5K4L5K4L4L5K4LV\\P4"}, "label": "a closeup of a apple laptop"}]}
{"id": 347796, "image": "COCO_train2014_000000347796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop that is very blurry\"."}], "task": "refering", "answer_template": "The \"a laptop that is very blurry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 253, 254, 255, 256, 257, 258, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427], "bbox": [0.0, 0.582015625, 0.6194085027726433, 0.9842656250000001], "iscrowd": 0, "area": 73350.74919999999, "rle": {"size": [640, 541], "counts": "^<c2]a0e4[K00000000000O10000NhH\\CX7d<2000000000O100000000000O1000000000000000000000000O100000000000000000000000O01000000000000000000000000O1000000000000000000000000O1000000000O10000000000000O1000000000000000000000000O1000000000000000000000O10O1000000000000000000000000O1000000000000000000000000O100000O100000000000000000O1000000000000000000000000O100000000000000000O100000O1000000000000000000000000O1000000000000000000000000O1000O10000000000000000000O10000000000000000000000O100000000000000000O100000O1000000000000000000000000O1000000000000000000000000O1000O10000N2L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L4L4L4L3N3L4L4L4L4L4L4L4L4L4L4L4L3M4L5K4L5K4L4L5K4L5L3L4L5K4L5K4L5K4L4L5K4L5K4L4L5K4LV\\P4"}, "label": "a laptop that is very blurry"}]}
{"id": 519836, "image": "COCO_train2014_000000519836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vintage white car\"."}], "task": "refering", "answer_template": "The \"a vintage white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 126, 127, 128, 129, 130, 140, 141, 142, 143, 154], "bbox": [0.0, 0.44301999999999997, 0.303775, 0.65592], "iscrowd": 0, "area": 6972.377850000002, "rle": {"size": [500, 400], "counts": "T7=V?k0VOk0UOf0YO2O1O1N2O1O1N2O1MmLcCT3\\<201O1O00O1N3_OdCfM_<Y2bCdMa<[2aCaMb<^2;O2N1O1O1O1O1O1N2O2N1O1O0000000000000O1000000000000000001O000O100O100O100O1O100O100O1O100O100O1O100O101N1O100O100O100O1O100O100O1O100O100O1O100O1000000O10000O10000O10000O2O7I9G3L101O0O2O1M2O2N1O2NRoW4"}, "label": "a vintage white car"}]}
{"id": 519836, "image": "COCO_train2014_000000519836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car on the near side of the street\"."}], "task": "refering", "answer_template": "The \"the car on the near side of the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 126, 127, 128, 129, 130, 140, 141, 142, 143, 154], "bbox": [0.0, 0.44301999999999997, 0.303775, 0.65592], "iscrowd": 0, "area": 6972.377850000002, "rle": {"size": [500, 400], "counts": "T7=V?k0VOk0UOf0YO2O1O1N2O1O1N2O1MmLcCT3\\<201O1O00O1N3_OdCfM_<Y2bCdMa<[2aCaMb<^2;O2N1O1O1O1O1O1N2O2N1O1O0000000000000O1000000000000000001O000O100O100O100O1O100O100O1O100O100O1O100O101N1O100O100O100O1O100O100O1O100O100O1O100O1000000O10000O10000O10000O2O7I9G3L101O0O2O1M2O2N1O2NRoW4"}, "label": "the car on the near side of the street"}]}
{"id": 151200, "image": "COCO_train2014_000000151200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"batter holding a bat on a baseball field\"."}], "task": "refering", "answer_template": "The \"batter holding a bat on a baseball field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 23, 24, 25, 26, 39, 40, 41, 54, 55, 56, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 127, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 204, 205, 206, 219, 220, 221, 234, 235, 236, 249, 250, 251, 264, 265, 266, 279, 280, 281, 294, 295, 296, 309, 310], "bbox": [0.49512880562060885, 0.033703125, 0.916135831381733, 0.8741562500000001], "iscrowd": 0, "area": 59061.404800000004, "rle": {"size": [640, 427], "counts": "VTT43gc0l0TOl0TOf0ZO7I7H8I7I6QNP2O1N2N2N2N2N1O2N2O1N2N2N1O2N1OD<D=C<D<E;I76J6J6I7I7I8H;E:F;E:F;bMnHZG\\7]8hH^Gc7Y8`HbGj7U8YHgGR8o7RHlGm:V5WEeJ[;j4hDRKj;^4YD]KZ<Q4jCjKg<f3]CULh<g3[CULj<g3ZCTLk<i3XCRLm<j3VCRLo<j3g1L4L4M1N1O1O1O100O1O1O1O100O10000000000000000O10000000000000O100000000000001O1O001O1O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O001O1O1N2O1Oh1XN1O0O2O1O1O1bMeJUF\\5n8VL[Fk3h8hMbFY2`8S5SOm0ROn0G8O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1N101N2N2N2N1O2N2N2N2N2N5J6K5K5K5K4L5K5K6J9G8E;F;D;E;E<D;F:\\OnXf0"}, "label": "batter holding a bat on a baseball field"}]}
{"id": 151200, "image": "COCO_train2014_000000151200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an out of focus baseball player beyond the fence with the number 19 on his jersey\"."}], "task": "refering", "answer_template": "The \"an out of focus baseball player beyond the fence with the number 19 on his jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 23, 24, 25, 26, 39, 40, 41, 54, 55, 56, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 127, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 204, 205, 206, 219, 220, 221, 234, 235, 236, 249, 250, 251, 264, 265, 266, 279, 280, 281, 294, 295, 296, 309, 310], "bbox": [0.49512880562060885, 0.033703125, 0.916135831381733, 0.8741562500000001], "iscrowd": 0, "area": 59061.404800000004, "rle": {"size": [640, 427], "counts": "VTT43gc0l0TOl0TOf0ZO7I7H8I7I6QNP2O1N2N2N2N2N1O2N2O1N2N2N1O2N1OD<D=C<D<E;I76J6J6I7I7I8H;E:F;E:F;bMnHZG\\7]8hH^Gc7Y8`HbGj7U8YHgGR8o7RHlGm:V5WEeJ[;j4hDRKj;^4YD]KZ<Q4jCjKg<f3]CULh<g3[CULj<g3ZCTLk<i3XCRLm<j3VCRLo<j3g1L4L4M1N1O1O1O100O1O1O1O100O10000000000000000O10000000000000O100000000000001O1O001O1O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O001O1O1N2O1Oh1XN1O0O2O1O1O1bMeJUF\\5n8VL[Fk3h8hMbFY2`8S5SOm0ROn0G8O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1N101N2N2N2N1O2N2N2N2N2N5J6K5K5K5K4L5K5K6J9G8E;F;D;E;E<D;F:\\OnXf0"}, "label": "an out of focus baseball player beyond the fence with the number 19 on his jersey"}]}
{"id": 126625, "image": "COCO_train2014_000000126625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair behind a woman drinking wine\"."}], "task": "refering", "answer_template": "The \"a chair behind a woman drinking wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 201, 202, 224, 225, 247, 248, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343], "bbox": [0.7556093749999999, 0.4706807511737089, 0.969875, 0.9878403755868544], "iscrowd": 0, "area": 13922.637200000006, "rle": {"size": [426, 640], "counts": "jaY69Q==Ba1_N7J6I8kHUM]3Q3]LPMa3W3XLkLg3[3SLeLl3h3gKZLX4U4YKlKe4c4ZJQLd5l5O1N101N010000O100O100O10000O11O0000001O0000010O01Od0\\OX1iNa0^O8H8H7J3L4L00010O000000001N10000000001O000000001O0000000O2O000000001O5K5F;G8H9H7H9Ga0_Oe0[Olj7U2PSHb0@00O100O1O100O101N1O101N100O1O2O0O100O2N1O1O1O2N1N6K4K6E:F;C<E;E\\b7"}, "label": "a chair behind a woman drinking wine"}]}
{"id": 126625, "image": "COCO_train2014_000000126625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair behind a woman who is drinking wine\"."}], "task": "refering", "answer_template": "The \"a brown chair behind a woman who is drinking wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 201, 202, 224, 225, 247, 248, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343], "bbox": [0.7556093749999999, 0.4706807511737089, 0.969875, 0.9878403755868544], "iscrowd": 0, "area": 13922.637200000006, "rle": {"size": [426, 640], "counts": "jaY69Q==Ba1_N7J6I8kHUM]3Q3]LPMa3W3XLkLg3[3SLeLl3h3gKZLX4U4YKlKe4c4ZJQLd5l5O1N101N010000O100O100O10000O11O0000001O0000010O01Od0\\OX1iNa0^O8H8H7J3L4L00010O000000001N10000000001O000000001O0000000O2O000000001O5K5F;G8H9H7H9Ga0_Oe0[Olj7U2PSHb0@00O100O1O100O101N1O101N100O1O2O0O100O2N1O1O1O2N1N6K4K6E:F;C<E;E\\b7"}, "label": "a brown chair behind a woman who is drinking wine"}]}
{"id": 478885, "image": "COCO_train2014_000000478885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a apple desktop computer\"."}], "task": "refering", "answer_template": "The \"a apple desktop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 199, 200], "bbox": [0.529375, 0.19293427230046947, 0.75959375, 0.5623474178403756], "iscrowd": 0, "area": 17672.502699999997, "rle": {"size": [426, 640], "counts": "RW]48g<;E;E;E<C<E;E;E<D;E;M4O01O0001O0000010O00001O0000010O0000001O01O01O0000001O01O01O00000010O000001O000010O000001O0001O01O00001O0001O01O00000010O0001O00000010O000001O00010O0000001O0001O01O00001O01O0001O00000010O0001O0000010O0000001O00010O0000001O01O0001O00001O01O000N2G9G9F;F9G9G9G9G9G9G:F9G9G9FWoo1"}, "label": "a apple desktop computer"}]}
{"id": 478885, "image": "COCO_train2014_000000478885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white imac computer that is also turned on\"."}], "task": "refering", "answer_template": "The \"the white imac computer that is also turned on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 199, 200], "bbox": [0.529375, 0.19293427230046947, 0.75959375, 0.5623474178403756], "iscrowd": 0, "area": 17672.502699999997, "rle": {"size": [426, 640], "counts": "RW]48g<;E;E;E<C<E;E;E<D;E;M4O01O0001O0000010O00001O0000010O0000001O01O01O0000001O01O01O00000010O000001O000010O000001O0001O01O00001O0001O01O00000010O0001O00000010O000001O00010O0000001O0001O01O00001O01O0001O00000010O0001O0000010O0000001O00010O0000001O01O0001O00001O01O000N2G9G9F;F9G9G9G9G9G9G:F9G9G9FWoo1"}, "label": "the white imac computer that is also turned on"}]}
{"id": 478885, "image": "COCO_train2014_000000478885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a college student ' s pc keyboard\"."}], "task": "refering", "answer_template": "The \"a college student ' s pc keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 217, 218, 219, 220, 221, 222, 242, 243, 244, 245, 268], "bbox": [0.438890625, 0.5471126760563381, 0.6960625, 0.7456807511737088], "iscrowd": 0, "area": 5470.482649999998, "rle": {"size": [426, 640], "counts": "iTe34S=5N0O2N1O101N1O2O0O1O2N100O2N1O101N1O2O0O1O2N100O2N1O2O00010O0010O0010O00010O0010O00010O00010O01O01O010Ob0^O01O0BDnC=Q<EmC;S<FlC:T<GkC:U<=1XOjC;V<CkC=V<AkC`0T<@lC`0U<^OlCb0U<]OkCd0U<ZOlCf0U<XOlCi0Y<101O000010O2O0O2NG]OmCb0Q<@oC?Q<BoC>P<CPD<P<EPD;o;FQD9n;IRD6n;KRD5m;LSD3m;MSD4l;KTD6k;KTD7l;HSD9n;ERD<o;CPD?\\<1N3N1O1N3N1O2M2O2O000010O01O01O01O010O010O00010O0010O01O01O010O01O01O01O010O0010O00010O010O01O01O000O2N1N2O2N1O2N1O1O2M2O1O2N1O1O2M2O2N1O1O2NVk`2"}, "label": "a college student ' s pc keyboard"}]}
{"id": 478885, "image": "COCO_train2014_000000478885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a keyboard of the apple brand computer kept on a executive table\"."}], "task": "refering", "answer_template": "The \"a keyboard of the apple brand computer kept on a executive table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 217, 218, 219, 220, 221, 222, 242, 243, 244, 245, 268], "bbox": [0.438890625, 0.5471126760563381, 0.6960625, 0.7456807511737088], "iscrowd": 0, "area": 5470.482649999998, "rle": {"size": [426, 640], "counts": "iTe34S=5N0O2N1O101N1O2O0O1O2N100O2N1O101N1O2O0O1O2N100O2N1O2O00010O0010O0010O00010O0010O00010O00010O01O01O010Ob0^O01O0BDnC=Q<EmC;S<FlC:T<GkC:U<=1XOjC;V<CkC=V<AkC`0T<@lC`0U<^OlCb0U<]OkCd0U<ZOlCf0U<XOlCi0Y<101O000010O2O0O2NG]OmCb0Q<@oC?Q<BoC>P<CPD<P<EPD;o;FQD9n;IRD6n;KRD5m;LSD3m;MSD4l;KTD6k;KTD7l;HSD9n;ERD<o;CPD?\\<1N3N1O1N3N1O2M2O2O000010O01O01O01O010O010O00010O0010O01O01O010O01O01O01O010O0010O00010O010O01O01O000O2N1N2O2N1O2N1O1O2M2O1O2N1O1O2M2O2N1O1O2NVk`2"}, "label": "a keyboard of the apple brand computer kept on a executive table"}]}
{"id": 478885, "image": "COCO_train2014_000000478885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor in the middle of the other two\"."}], "task": "refering", "answer_template": "The \"the monitor in the middle of the other two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 147, 148, 149, 150], "bbox": [0.360953125, 0.15842723004694834, 0.54759375, 0.4450234741784037], "iscrowd": 0, "area": 11626.7157, "rle": {"size": [426, 640], "counts": "YUP3>m<k0TOm0SOl0TO3M0000001O0001O0001O000000001O01O000001O000000001O01O000001O000000001O01O000001O0000000010O00000001O0000000010O00000001O000000010O000000001O000000010O00000000001O000001O01O000000001O0001O0001O000000001O0001O000E;ZOf0YOg0ZOf0[Oe0N2O1N3M2NVbh3"}, "label": "the monitor in the middle of the other two"}]}
{"id": 52902, "image": "COCO_train2014_000000052902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a flip phone\"."}], "task": "refering", "answer_template": "The \"a woman holding a flip phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 29, 30, 31, 32, 33, 34, 35, 52, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.047984375, 0.0594847775175644, 0.784125, 0.9920608899297423], "iscrowd": 0, "area": 116215.8258, "rle": {"size": [427, 640], "counts": "gZ=2U=4L3M4L4L4N2N2N2N2M3L4M3L4D<N2M3N2M3N2M3M3O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1M3M3M3M3M3M3M3H8F:N2N2N2N2O1N200O10000O1000000000000000000O2N1N2O1N2WOi0WOi0M3M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3O100O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O1N2N2N2M3N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2O1N2N2N2N2O1N2O1N2N2O1N2N2O1N2O10000O100O2O000O100O10000O100O10000O1000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O11O00001O00010O001O00001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O009G9G1O1O1O2N1O1O1O2N2N2N1O2N2N2N2N2N2N2N2N2N?A4L3M3M4L3M3M3M3M1O2N1O1O2N1O7I=C=CV1jN7I5K3M3M3M4L1O1O001O100O1O1O1O1O1O001O1O1O1O1O1O1O1O1O2N1O1O1O2N2N1O2N1O2N2N1O2N2N2N2N2N2N2N2N2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N2N2N2N2N1O2N3M2N2N2N2N3M2N2N2N2N3M2O1N2N2N3M2N2N2N2N9GZTi1"}, "label": "a woman holding a flip phone"}]}
{"id": 52902, "image": "COCO_train2014_000000052902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman smiling and holding up a cell phone ; she is wearing a checkered scarf\"."}], "task": "refering", "answer_template": "The \"a young woman smiling and holding up a cell phone ; she is wearing a checkered scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 29, 30, 31, 32, 33, 34, 35, 52, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.047984375, 0.0594847775175644, 0.784125, 0.9920608899297423], "iscrowd": 0, "area": 116215.8258, "rle": {"size": [427, 640], "counts": "gZ=2U=4L3M4L4L4N2N2N2N2M3L4M3L4D<N2M3N2M3N2M3M3O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1M3M3M3M3M3M3M3H8F:N2N2N2N2O1N200O10000O1000000000000000000O2N1N2O1N2WOi0WOi0M3M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3O100O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O1N2N2N2M3N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2O1N2N2N2N2O1N2O1N2N2O1N2N2O1N2O10000O100O2O000O100O10000O100O10000O1000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O11O00001O00010O001O00001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O009G9G1O1O1O2N1O1O1O2N2N2N1O2N2N2N2N2N2N2N2N2N?A4L3M3M4L3M3M3M3M1O2N1O1O2N1O7I=C=CV1jN7I5K3M3M3M4L1O1O001O100O1O1O1O1O1O001O1O1O1O1O1O1O1O1O2N1O1O1O2N2N1O2N1O2N2N1O2N2N2N2N2N2N2N2N2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N2N2N2N2N1O2N3M2N2N2N2N3M2N2N2N2N3M2O1N2N2N3M2N2N2N2N9GZTi1"}, "label": "a young woman smiling and holding up a cell phone ; she is wearing a checkered scarf"}]}
{"id": 323240, "image": "COCO_train2014_000000323240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing glasses and a pink shirt\"."}], "task": "refering", "answer_template": "The \"a girl wearing glasses and a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 280, 281], "bbox": [0.07062500000000001, 0.34741666666666665, 0.30101562499999995, 0.7218541666666667], "iscrowd": 0, "area": 17297.7651, "rle": {"size": [480, 640], "counts": "Q]e02j>5J6K5J5L4K5K5L4K5L4K5N2O1O1O100O1O1O1O10O01O1O100O1O1O100O1O1O010O1O1O100O1O100O1O1O010O1O1O100O1O100O1O1J5F;E;G9N2N2M3N2N<D=B6K2N2M3N2N2N2000000O100000000O10000000000O100000000O100000000O10O100000000O101N1O100O1O2O0O1O100O2N100O3M2O2M3M3N1N3M3N1N3M3N1N3M3N1N3M3N1N3M3N1N3M3N6I:F;F:E;E:G:D<\\Od0]OTXa6"}, "label": "a girl wearing glasses and a pink shirt"}]}
{"id": 323240, "image": "COCO_train2014_000000323240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian girl with a pink shirt eating at the table\"."}], "task": "refering", "answer_template": "The \"an asian girl with a pink shirt eating at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 280, 281], "bbox": [0.07062500000000001, 0.34741666666666665, 0.30101562499999995, 0.7218541666666667], "iscrowd": 0, "area": 17297.7651, "rle": {"size": [480, 640], "counts": "Q]e02j>5J6K5J5L4K5K5L4K5L4K5N2O1O1O100O1O1O1O10O01O1O100O1O1O100O1O1O010O1O1O100O1O100O1O1O010O1O1O100O1O100O1O1J5F;E;G9N2N2M3N2N<D=B6K2N2M3N2N2N2000000O100000000O10000000000O100000000O100000000O10O100000000O101N1O100O1O2O0O1O100O2N100O3M2O2M3M3N1N3M3N1N3M3N1N3M3N1N3M3N1N3M3N1N3M3N6I:F;F:E;E:G:D<\\Od0]OTXa6"}, "label": "an asian girl with a pink shirt eating at the table"}]}
{"id": 323240, "image": "COCO_train2014_000000323240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown table with people eating christmas food\"."}], "task": "refering", "answer_template": "The \"a brown table with people eating christmas food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 264, 265, 287, 288, 361, 362, 363, 364, 384, 385, 386, 387], "bbox": [0.342953125, 0.5988958333333334, 0.84765625, 0.9966458333333333], "iscrowd": 0, "area": 5449.458899999998, "rle": {"size": [480, 640], "counts": "hnV31l>3M3L310O100O100O100O1G81010O0000010O00K5O1N2O1O1O1N2O10000O10000000000000000001O001O001O8H0000001O03M3M2N3M3M3Me[96ncF8L310O1O00001O000000001O0001O01O00000gB_OS<a0kCKl;5QDMP<2nC1R<NlC4U<KiC8W<GhC:Y<EdC>]<A^Ce0b<ZOXCl0i<e0100O1[NoBY1R=;O01YNnB^1S=aNmB_1T=_NlBc1Y=O1O109eNfB<R>M3NO0O10001O06J<DQha10oW^N1O100O1O1O1O1O1[BI_<8_CIa<8]CIc<8[CIe<8YCIg<8WCIi<8UCIk<8SCIm<8oBKQ=6hB0X=1bB4^=g00000000000000000001O00000000000000001O1O001O1O1O1O001O1O1O001O1O1OO100O1O1O1O1O1O1O1O1O100001O0000001O0000001O0000001O0000001O0000001O0G9C=DbQ]1"}, "label": "a brown table with people eating christmas food"}]}
{"id": 560809, "image": "COCO_train2014_000000560809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed which is to the right of another bed\"."}], "task": "refering", "answer_template": "The \"a bed which is to the right of another bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.18053125, 0.3591866028708134, 0.99803125, 0.9883971291866028], "iscrowd": 0, "area": 73950.97050000001, "rle": {"size": [418, 640], "counts": "Sh_12n<2N2M3N2N2N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O100O1O1O1O1O1O100O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000O01000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1O1O1O100O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1N2N2N2N2N2N2N2N2O1N2N2^Ob0XNh1SOm000O1O100O1_OdHZK^7f4?001O001O0]HSKQ7n4iHYKU7g4fH`KX7a4dHdKZ7R50000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000O100001O000000000000000000000000000000000000001O0000000000000000000O10000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O0000001O0000001O0000^7"}, "label": "a bed which is to the right of another bed"}]}
{"id": 560809, "image": "COCO_train2014_000000560809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed on the right\"."}], "task": "refering", "answer_template": "The \"the bed on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.18053125, 0.3591866028708134, 0.99803125, 0.9883971291866028], "iscrowd": 0, "area": 73950.97050000001, "rle": {"size": [418, 640], "counts": "Sh_12n<2N2M3N2N2N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O100O1O1O1O1O1O100O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000O01000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1O1O1O100O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1N2N2N2N2N2N2N2N2O1N2N2^Ob0XNh1SOm000O1O100O1_OdHZK^7f4?001O001O0]HSKQ7n4iHYKU7g4fH`KX7a4dHdKZ7R50000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000O100001O000000000000000000000000000000000000001O0000000000000000000O10000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O0000001O0000001O0000^7"}, "label": "the bed on the right"}]}
{"id": 478892, "image": "COCO_train2014_000000478892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green chair to the right of another chair\"."}], "task": "refering", "answer_template": "The \"a green chair to the right of another chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170], "bbox": [0.21153125, 0.2696573875802998, 0.516515625, 0.4674089935760171], "iscrowd": 0, "area": 10521.6647, "rle": {"size": [467, 640], "counts": "dUn14^>2M3M4M2M3M4L3N2M4L3N101O001N101O0O2O001N101O001N101O0O2O001O0O2O001N101O1O001O0O2O001O001O1O1N101O1O1O1O1O1N101O1O1O1O1N101O001O001O1N101O001O1O1O001O1O1O010O1O1O00O100O100O100O100O10001N100O100O100O10000O100O100O2O0O10000O100O100O10000O101N100O100O10000O100O2O0O101N101O0O101N101N101N101O1N2O1N2O0O2O1O1N2O0O2O1N2O1N2O1O1N2O1N2O1N2O1N2O000O2O0O2O0O101N10001N101N100O2O000O2O0O101N10^e]4"}, "label": "a green chair to the right of another chair"}]}
{"id": 478892, "image": "COCO_train2014_000000478892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green plastic chair closest to the potted tree\"."}], "task": "refering", "answer_template": "The \"a green plastic chair closest to the potted tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170], "bbox": [0.21153125, 0.2696573875802998, 0.516515625, 0.4674089935760171], "iscrowd": 0, "area": 10521.6647, "rle": {"size": [467, 640], "counts": "dUn14^>2M3M4M2M3M4L3N2M4L3N101O001N101O0O2O001N101O001N101O0O2O001O0O2O001N101O1O001O0O2O001O001O1O1N101O1O1O1O1O1N101O1O1O1O1N101O001O001O1N101O001O1O1O001O1O1O010O1O1O00O100O100O100O100O10001N100O100O100O10000O100O100O2O0O10000O100O100O10000O101N100O100O10000O100O2O0O101N101O0O101N101N101N101O1N2O1N2O0O2O1O1N2O0O2O1N2O1N2O1O1N2O1N2O1N2O1N2O000O2O0O2O0O101N10001N101N100O2O000O2O0O101N10^e]4"}, "label": "a green plastic chair closest to the potted tree"}]}
{"id": 478892, "image": "COCO_train2014_000000478892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the left in the middle\"."}], "task": "refering", "answer_template": "The \"the chair on the left in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 232, 233], "bbox": [0.034515625, 0.38064239828693786, 0.338953125, 0.6107494646680941], "iscrowd": 0, "area": 12841.612100000006, "rle": {"size": [467, 640], "counts": "`W:3`>5J6J6J6K4K4L5K4L5L3L4L2N2O0O2N2N2O0O2N2N101N2N2N101N2N2N101O1O001N2O1O001O1O1O001O1O001N2O1O001O1O1O001O1O001O0O100000O10000O100O100O10O0100O1O100O100O100O100O010O100O100O100O100O10O0100O100O100O100O100O101N100O100O101N100O100O2O0O100O101N100O1O101N100O100O2O0O100O100O2O0O100O101N100O100O2O0O100O101N100O2O1N2O1N2O1N2O1N2O0O2O1N2O1O1N2O1O1N2O1O1N101N2O001N2O001N2O001N2O001N101O1N101NmTQ6"}, "label": "the chair on the left in the middle"}]}
{"id": 478892, "image": "COCO_train2014_000000478892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green chair in the middle of green chairs\"."}], "task": "refering", "answer_template": "The \"a green chair in the middle of green chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 232, 233], "bbox": [0.034515625, 0.38064239828693786, 0.338953125, 0.6107494646680941], "iscrowd": 0, "area": 12841.612100000006, "rle": {"size": [467, 640], "counts": "`W:3`>5J6J6J6K4K4L5K4L5L3L4L2N2O0O2N2N2O0O2N2N101N2N2N101N2N2N101O1O001N2O1O001O1O1O001O1O001N2O1O001O1O1O001O1O001O0O100000O10000O100O100O10O0100O1O100O100O100O100O010O100O100O100O100O10O0100O100O100O100O100O101N100O100O101N100O100O2O0O100O101N100O1O101N100O100O2O0O100O100O2O0O100O101N100O100O2O0O100O101N100O2O1N2O1N2O1N2O1N2O0O2O1N2O1O1N2O1O1N2O1O1N101N2O001N2O001N2O001N2O001N101O1N101NmTQ6"}, "label": "a green chair in the middle of green chairs"}]}
{"id": 478892, "image": "COCO_train2014_000000478892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair on the bottom right\"."}], "task": "refering", "answer_template": "The \"a chair on the bottom right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387], "bbox": [0.7814687499999999, 0.6973875802997859, 1.0, 0.9769593147751606], "iscrowd": 0, "area": 11778.831000000002, "rle": {"size": [467, 640], "counts": "n^T71`>2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2O1O10000O2O000O10000O101N10000O1O1I6I8I6I8I6I7I8G80N2O1O1N2O02O1O1O1O1O1000000000000000000001O00000000000000000000001O00000O100000000000001O0O10001O0000001O0O10001O0000001O000O101O0000001N1000001O0O10001O0O10001O000O101O00000O2O00000O2`MlE?_O"}, "label": "a chair on the bottom right"}]}
{"id": 478892, "image": "COCO_train2014_000000478892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the left side of the table furthest from the red vase\"."}], "task": "refering", "answer_template": "The \"the chair on the left side of the table furthest from the red vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 207, 208, 209, 231, 232, 345, 346, 347, 348, 349, 368, 369, 370], "bbox": [0.000671875, 0.5041970021413277, 0.19260937500000003, 1.0], "iscrowd": 0, "area": 5725.469149999997, "rle": {"size": [467, 640], "counts": "[7T1`=OiGHS28lMIT26mMJS26mMJT25kMLU24kMLU24kMLU24jMMW22iMNW22hMOX21hMOX21hMOX21gM0Z2OfM1Z2OfM1Z2OeM2[2NeM2\\2McM4]2MbM3^2MbM3^2N`M2b2M^M3b2N]M2c2N\\M3d2N[M2f2NYM2g2NXM3h2NWM2i2NVM3k2MTM3l2MTM3l2NRM3n2NQM2P3NoL2R3NlL3T3NkL2V3NhL3Y3MfL3[3MdL3]3MaL4`3L_L4b3L]L3e3MYL4h3LWL3k3MTL3m3NPL2R4NmK1U4OiK2W4NiK1Y4NgK2Z4LfK4\\4KdK4^4JcK6^4IaK7a4H_K8b4F_K9c4F\\K;e4D[K;g4CYK=i4BWK>i4AXK>j4AUK`0l4_OTK`0n4^ORKb0P5]OPKc0Q5[OPKd0R5[OmJf0T5YOlJf0V5XOkJg0W5XOhJi0Y5VOgJi0[5UOeJl0\\5SOdJl0]5SOdJl0^5SOaJn0`5QO`Jn0b5PO^JQ1b9O1O010O101N1O100O101N100O1O100O2O0O1O100O101N1O100O100O2N100O100O2N100O100O101N1O100O100Oo[[7"}, "label": "the chair on the left side of the table furthest from the red vase"}]}
{"id": 323249, "image": "COCO_train2014_000000323249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red tram driving through the city at night\"."}], "task": "refering", "answer_template": "The \"a red tram driving through the city at night\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314], "bbox": [0.515015625, 0.443204081632653, 0.700015625, 0.7513673469387754], "iscrowd": 0, "area": 16422.65285, "rle": {"size": [490, 640], "counts": "eWn4e0o=g0YOh0YOf0YOg0YOf0M3O1O1O1O1O1O1O1O101O00000000000000000000000000000000000000001O0001O000000000000000000000000000000000001O0000000000000000000000000000000000000000001O00000001O00000000001O00000000001O00000001O0001O00000000009G:F;E:F:F:F:B>Al0UOoak2"}, "label": "a red tram driving through the city at night"}]}
{"id": 323249, "image": "COCO_train2014_000000323249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train with a red front\"."}], "task": "refering", "answer_template": "The \"the train with a red front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314], "bbox": [0.515015625, 0.443204081632653, 0.700015625, 0.7513673469387754], "iscrowd": 0, "area": 16422.65285, "rle": {"size": [490, 640], "counts": "eWn4e0o=g0YOh0YOf0YOg0YOf0M3O1O1O1O1O1O1O1O101O00000000000000000000000000000000000000001O0001O000000000000000000000000000000000001O0000000000000000000000000000000000000000001O00000001O00000000001O00000000001O00000001O0001O00000000009G:F;E:F:F:F:B>Al0UOoak2"}, "label": "the train with a red front"}]}
{"id": 323249, "image": "COCO_train2014_000000323249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an oncoming public bus in front of another bus\"."}], "task": "refering", "answer_template": "The \"an oncoming public bus in front of another bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315], "bbox": [0.51609375, 0.4419795918367347, 0.7089062500000001, 0.7580204081632653], "iscrowd": 0, "area": 17141.338249999993, "rle": {"size": [490, 640], "counts": "aXn42f>b0]Oc0dM\\2F:M3L4M3M3O1O100O1O1O100JbK_E_4a:bK]E_4c:400000001O00000O10000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O00000000000000000000000000hKoEW3V;L3M4L4L3M4L3M4KV1iN:F:G9F;E:GPeh2"}, "label": "an oncoming public bus in front of another bus"}]}
{"id": 323249, "image": "COCO_train2014_000000323249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white trolley car\"."}], "task": "refering", "answer_template": "The \"a red and white trolley car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315], "bbox": [0.51609375, 0.4419795918367347, 0.7089062500000001, 0.7580204081632653], "iscrowd": 0, "area": 17141.338249999993, "rle": {"size": [490, 640], "counts": "aXn42f>b0]Oc0dM\\2F:M3L4M3M3O1O100O1O1O100JbK_E_4a:bK]E_4c:400000001O00000O10000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O00000000000000000000000000hKoEW3V;L3M4L4L3M4L3M4KV1iN:F:G9F;E:GPeh2"}, "label": "a red and white trolley car"}]}
{"id": 134837, "image": "COCO_train2014_000000134837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an plane which is going to move\"."}], "task": "refering", "answer_template": "The \"an plane which is going to move\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 282, 283, 288, 289, 293, 294, 304, 305, 306, 311, 312, 313, 316, 317, 327, 328, 329, 334, 335, 336, 339], "bbox": [0.0015, 0.40637647058823534, 1.0, 0.9906352941176471], "iscrowd": 0, "area": 66445.41674999997, "rle": {"size": [425, 640], "counts": "Xf0=`<<0O100000O10000000000000000O1000000000000000000O10000000000000000O1000000000000000000O01000000000000000O100000O100O1O100O10O0100O1O100O100O10O0100O1O100O10000000O10O100000000O100000O10000000000000000O1000000000000000000O10000000000000000O100000000\\O^OaDb0^;D]D<b;JYD6f;0UD0j;h0O1O1O1O1O1O2N100O2O0O10001N100jGXNj4j1QK\\Nm4d1nJcNP5]1mJgNQ5Z1nJhNQ5X1mJjNR5W1mJjNS5V1lJkNS5V1kJmNS5T1lJmNS5T1lJmNS5T1mJmNQ5T1oJlNP5U1oJlNP5U1PKlNo4T1QKlNn4U1QKlNn4U1RKkNm4W1RKjNl4W1SKjNl4W1TKiNk4X1UKiNi4X1WKhNi4X1VKiNi4X1WKiNg4X1YKhNf4Y1YKiNe4Y1ZKgNe4[1ZKfNd4\\1ZKeNe4]1ZKeNd4]1ZKeNd4\\1ZKfNd4[1\\KgNb4Z1]KfNc4Z1\\KhNc4Y1\\KgNc4Z1]KfNc4Z1\\KgNd4Z1[KfNd4[1[KfNe4Z1[KfNe4[1YKfNf4[1ZKdNg4]1XKcNh4]1WKcNj4]1VKcNi4_1UKbNk4^1SKcNn4^1nJeNQ5\\1lJfNU5Z1hJiNX5Y1dJiN\\5Y1`JhN`5[1]JdNe5_1VJmLXOR1c6S2QJjL@Q1_6X2mIgLHn0\\6]2hIdL1l0X6c2bI`L;i0U6i2\\I^Lb0g0R6n2XIZLi0g0P6[3SJcLn5]3SJaLm5a3SJ\\Lo5d3RJZLo5f3RJXLo5i3QJULo5l3RJRLo5n3RJPLo5Q4QJmKo5T4RJjKo5V4RJhKo5Y4`10O10000O1000000O10000O10000O010000O01000O01000O10O1000O0100O100O010O1O100O100O1O010O100O1O1000000000000000001O0000000000001O000000000000000000000000000000000000001O0000000000000000001O000000000000000000001O000000001O1O1O1O001O1O1O1O1O1O1O1O1O_OZH`Ke7`4\\H`Kc7`4^H`Ka7a4_H_K`7a4aH_Kl0HY5i4mI^Kg0OX5c4RJ]Ke01Y5b4SJ]Ka04[5_4UJ]K>5]5^4VJ]K;6_5]4WJ]K79a5Z4ZJ\\K3;c5Y4[J\\K0<e5X4\\J\\KL?g5V4]J[KJ`0i5U4^J[KCe0o5P4_J`La5`3`J`L_5`3aJ`L_5`3bJ_L^5a3bJ`L]5`3dJ_L]5`3cJ`L]5`3dJ_L\\5a3eJ`LY5`3gJaLX5`3hJ`LW5`3iJbLU5^3lJbLS5^3mJcLR5]3oJcLP5]3PKcLP5]3QKcLn4]3RKdLn4[3RKfLm4Z3TKfLk4Z3UKfLk4[3TKfLk4Z3UKfLk4Z3UKfLk4Z3VKeLj4[3VKeLj4[3VKeLj4[3VKfLi4Z3XKeLi4Y3XKgLh4Y3XKgLh4Y3XKgLi4X3XKgLh4Y3XKgLi4W3XKiLh4W3XKhLj4W3WKhLi4X3WKgLk4X3UKhLk4X3UKgLm4W3UKhLk4X3UKhLl4W3TKhLm4X3TKgLm4X3SKgLn4X3SKhLm4X3SKgLo4X3RKgLn4Y3RKfLP5Y3PKgLP5Y3QKeLQ5Y3PKfLQ5Z3oJcLU5\\3lJ`LW5`3iJ_LX5a3hJ]L[5a3hJ\\LY5d3hJYLZ5g3hJUL[5j3fJSL\\5m3fJiKb5V4j1O001O000O2O00001N10001O0O101O0O2O00001N101O001N101O001N101O001N101O001N101O001N101O001N101O0O2O1O001N101O001N2O001O0O2O1O001N10bGVM^6i2aI[M]6d2bI`M[6_2dIfMZ6Y2eIkMX6U2gIoMW6P2hITNV6j1jIYNU6f1kIZNV6e1iI\\NY6a1gI`NY6`1fIaN[6^1dIcN]6\\1aIfN`6X1`IiNa6V1^IlNb6T1TIUOm6j0jH_OV7b0dHC]7=^HGc78XHMi73RH1o7OnG3S8LkG5V8\\24K5L4K5L4K5ZLmFo2X9QMiFh2]9XMdF_2e9`M]FV2k9jMVFn1Q:RNPFe1X:[NiE\\1_:dNcER1e:nN\\Ei0l:WOUEg0m:XOTEg0l:YOVEf0i:ZOYEd0g:\\OYEd0g:\\OYEd0g:\\OYEd0g:\\OZEc0f:]OZEd0e:\\O[Ed0e:\\O[Ed0e:\\O\\Ec0d:]O\\Ec0d:]O\\Ec0d:]O\\Ed0c:\\O^Ec0b:]O^Ec0b:]O^Ec0b:]O^Ec0b:]O^Ec0b:]O_Eb0a:^O_Ec0`:]O`Ec0`:]O`Ec0`:]OaEb0_:^OaEb0_:^OaEb0_:^OaEc0^:]OcEb0]:^OcEb0]:^OcEb0]:^OcEb0]:]OdEc0\\:]OeEc0Z:]OfEc0Z:]OfEc0Z:]OfEc0Z:]OgEb0Y:^OgEb0Y:^OgEb0Y:^OgEc0X:]OhEc0X:]OiEb0W:^OiEb0W:^OiEb0W:^OiEb0W:^OjEb0U:^OkEb0U:^OkEb0U:^OkEb0U:^OlEa0T:_OlEa0T:_OlEb0S:^OmEb0S:^OmEb0S:^OnEa0R:_OnEa0R:_OnEa0R:_OnEa0R:_OoEa0P:_OPFa0P:_OPFa0P:_OPFa0P:_OPFa0P:_OQF`0o9@QFa0n9_ORFa0n9_ORFa0n9_OSF`0m9@SF`0m9@SF`0m9@SFa0l9_OUF`0k9@UF`0k9@UF`0k9@UF`0k9@UF`0k9@VF?j9@WFa0h9_OXFa0h9_OXFa0h9_OYF`0g9@YF`0g9@YF`0g9@YFa0f9_O[F`0d9A\\F?d9A\\F?d9A\\F?d9A\\F?d9A]F?b9A^F?b9A^F?b9A^F?b9A_F>a9B_F>a9B_F>a9B_F?`9A`F?`9AaF>_9BaF>_9BaF>_9BaF>_9BbF>]9BcF>]9BcF>]9BcF>]9BdF=\\9CdF=\\9CdF>[9BeF>[9BeF>[9BfF=Z9CfF=Z9CfF=Z9CfF=Z9CgF=X9ChF=X9ChF=X9ChF=a1"}, "label": "an plane which is going to move"}]}
{"id": 134837, "image": "COCO_train2014_000000134837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plane with two large windows on top , with a spinning propeller that has red around the edges\"."}], "task": "refering", "answer_template": "The \"a white plane with two large windows on top , with a spinning propeller that has red around the edges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 282, 283, 288, 289, 293, 294, 304, 305, 306, 311, 312, 313, 316, 317, 327, 328, 329, 334, 335, 336, 339], "bbox": [0.0015, 0.40637647058823534, 1.0, 0.9906352941176471], "iscrowd": 0, "area": 66445.41674999997, "rle": {"size": [425, 640], "counts": "Xf0=`<<0O100000O10000000000000000O1000000000000000000O10000000000000000O1000000000000000000O01000000000000000O100000O100O1O100O10O0100O1O100O100O10O0100O1O100O10000000O10O100000000O100000O10000000000000000O1000000000000000000O10000000000000000O100000000\\O^OaDb0^;D]D<b;JYD6f;0UD0j;h0O1O1O1O1O1O2N100O2O0O10001N100jGXNj4j1QK\\Nm4d1nJcNP5]1mJgNQ5Z1nJhNQ5X1mJjNR5W1mJjNS5V1lJkNS5V1kJmNS5T1lJmNS5T1lJmNS5T1mJmNQ5T1oJlNP5U1oJlNP5U1PKlNo4T1QKlNn4U1QKlNn4U1RKkNm4W1RKjNl4W1SKjNl4W1TKiNk4X1UKiNi4X1WKhNi4X1VKiNi4X1WKiNg4X1YKhNf4Y1YKiNe4Y1ZKgNe4[1ZKfNd4\\1ZKeNe4]1ZKeNd4]1ZKeNd4\\1ZKfNd4[1\\KgNb4Z1]KfNc4Z1\\KhNc4Y1\\KgNc4Z1]KfNc4Z1\\KgNd4Z1[KfNd4[1[KfNe4Z1[KfNe4[1YKfNf4[1ZKdNg4]1XKcNh4]1WKcNj4]1VKcNi4_1UKbNk4^1SKcNn4^1nJeNQ5\\1lJfNU5Z1hJiNX5Y1dJiN\\5Y1`JhN`5[1]JdNe5_1VJmLXOR1c6S2QJjL@Q1_6X2mIgLHn0\\6]2hIdL1l0X6c2bI`L;i0U6i2\\I^Lb0g0R6n2XIZLi0g0P6[3SJcLn5]3SJaLm5a3SJ\\Lo5d3RJZLo5f3RJXLo5i3QJULo5l3RJRLo5n3RJPLo5Q4QJmKo5T4RJjKo5V4RJhKo5Y4`10O10000O1000000O10000O10000O010000O01000O01000O10O1000O0100O100O010O1O100O100O1O010O100O1O1000000000000000001O0000000000001O000000000000000000000000000000000000001O0000000000000000001O000000000000000000001O000000001O1O1O1O001O1O1O1O1O1O1O1O1O_OZH`Ke7`4\\H`Kc7`4^H`Ka7a4_H_K`7a4aH_Kl0HY5i4mI^Kg0OX5c4RJ]Ke01Y5b4SJ]Ka04[5_4UJ]K>5]5^4VJ]K;6_5]4WJ]K79a5Z4ZJ\\K3;c5Y4[J\\K0<e5X4\\J\\KL?g5V4]J[KJ`0i5U4^J[KCe0o5P4_J`La5`3`J`L_5`3aJ`L_5`3bJ_L^5a3bJ`L]5`3dJ_L]5`3cJ`L]5`3dJ_L\\5a3eJ`LY5`3gJaLX5`3hJ`LW5`3iJbLU5^3lJbLS5^3mJcLR5]3oJcLP5]3PKcLP5]3QKcLn4]3RKdLn4[3RKfLm4Z3TKfLk4Z3UKfLk4[3TKfLk4Z3UKfLk4Z3UKfLk4Z3VKeLj4[3VKeLj4[3VKeLj4[3VKfLi4Z3XKeLi4Y3XKgLh4Y3XKgLh4Y3XKgLi4X3XKgLh4Y3XKgLi4W3XKiLh4W3XKhLj4W3WKhLi4X3WKgLk4X3UKhLk4X3UKgLm4W3UKhLk4X3UKhLl4W3TKhLm4X3TKgLm4X3SKgLn4X3SKhLm4X3SKgLo4X3RKgLn4Y3RKfLP5Y3PKgLP5Y3QKeLQ5Y3PKfLQ5Z3oJcLU5\\3lJ`LW5`3iJ_LX5a3hJ]L[5a3hJ\\LY5d3hJYLZ5g3hJUL[5j3fJSL\\5m3fJiKb5V4j1O001O000O2O00001N10001O0O101O0O2O00001N101O001N101O001N101O001N101O001N101O001N101O001N101O0O2O1O001N101O001N2O001O0O2O1O001N10bGVM^6i2aI[M]6d2bI`M[6_2dIfMZ6Y2eIkMX6U2gIoMW6P2hITNV6j1jIYNU6f1kIZNV6e1iI\\NY6a1gI`NY6`1fIaN[6^1dIcN]6\\1aIfN`6X1`IiNa6V1^IlNb6T1TIUOm6j0jH_OV7b0dHC]7=^HGc78XHMi73RH1o7OnG3S8LkG5V8\\24K5L4K5L4K5ZLmFo2X9QMiFh2]9XMdF_2e9`M]FV2k9jMVFn1Q:RNPFe1X:[NiE\\1_:dNcER1e:nN\\Ei0l:WOUEg0m:XOTEg0l:YOVEf0i:ZOYEd0g:\\OYEd0g:\\OYEd0g:\\OYEd0g:\\OZEc0f:]OZEd0e:\\O[Ed0e:\\O[Ed0e:\\O\\Ec0d:]O\\Ec0d:]O\\Ec0d:]O\\Ed0c:\\O^Ec0b:]O^Ec0b:]O^Ec0b:]O^Ec0b:]O^Ec0b:]O_Eb0a:^O_Ec0`:]O`Ec0`:]O`Ec0`:]OaEb0_:^OaEb0_:^OaEb0_:^OaEc0^:]OcEb0]:^OcEb0]:^OcEb0]:^OcEb0]:]OdEc0\\:]OeEc0Z:]OfEc0Z:]OfEc0Z:]OfEc0Z:]OgEb0Y:^OgEb0Y:^OgEb0Y:^OgEc0X:]OhEc0X:]OiEb0W:^OiEb0W:^OiEb0W:^OiEb0W:^OjEb0U:^OkEb0U:^OkEb0U:^OkEb0U:^OlEa0T:_OlEa0T:_OlEb0S:^OmEb0S:^OmEb0S:^OnEa0R:_OnEa0R:_OnEa0R:_OnEa0R:_OoEa0P:_OPFa0P:_OPFa0P:_OPFa0P:_OPFa0P:_OQF`0o9@QFa0n9_ORFa0n9_ORFa0n9_OSF`0m9@SF`0m9@SF`0m9@SFa0l9_OUF`0k9@UF`0k9@UF`0k9@UF`0k9@UF`0k9@VF?j9@WFa0h9_OXFa0h9_OXFa0h9_OYF`0g9@YF`0g9@YF`0g9@YFa0f9_O[F`0d9A\\F?d9A\\F?d9A\\F?d9A\\F?d9A]F?b9A^F?b9A^F?b9A^F?b9A_F>a9B_F>a9B_F>a9B_F?`9A`F?`9AaF>_9BaF>_9BaF>_9BaF>_9BbF>]9BcF>]9BcF>]9BcF>]9BdF=\\9CdF=\\9CdF>[9BeF>[9BeF>[9BfF=Z9CfF=Z9CfF=Z9CfF=Z9CgF=X9ChF=X9ChF=X9ChF=a1"}, "label": "a white plane with two large windows on top , with a spinning propeller that has red around the edges"}]}
{"id": 421560, "image": "COCO_train2014_000000421560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver oven with a towel on it\"."}], "task": "refering", "answer_template": "The \"a silver oven with a towel on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 329, 330], "bbox": [0.18716666666666668, 0.5878281249999999, 0.43643750000000003, 0.8334999999999999], "iscrowd": 0, "area": 16570.776550000002, "rle": {"size": [640, 480], "counts": "ich1c0\\c0h0YOf0ZOg0YOg0YOg0YO7I0001O000001O0000000000001O0000000000001O00000000001O00000001O0001O0000000000001O0000000000001O00000000001O0000000001O01O0000000000001O0000000000001O00000000001O0000000000010O0000000000001O0000000000001O000000000d0]Of0YOf0ZOg0YOf0ZOg0YOokX5"}, "label": "a silver oven with a towel on it"}]}
{"id": 421560, "image": "COCO_train2014_000000421560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver oven range between two white cabinets\"."}], "task": "refering", "answer_template": "The \"silver oven range between two white cabinets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 329, 330], "bbox": [0.18716666666666668, 0.5878281249999999, 0.43643750000000003, 0.8334999999999999], "iscrowd": 0, "area": 16570.776550000002, "rle": {"size": [640, 480], "counts": "ich1c0\\c0h0YOf0ZOg0YOg0YOg0YO7I0001O000001O0000000000001O0000000000001O00000000001O00000001O0001O0000000000001O0000000000001O00000000001O0000000001O01O0000000000001O0000000000001O00000000001O0000000000010O0000000000001O0000000000001O000000000d0]Of0YOf0ZOg0YOf0ZOg0YOokX5"}, "label": "silver oven range between two white cabinets"}]}
{"id": 421562, "image": "COCO_train2014_000000421562.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unmade bed\"."}], "task": "refering", "answer_template": "The \"an unmade bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 191, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.299390625, 0.4834433962264151, 1.0, 0.9890330188679246], "iscrowd": 0, "area": 64246.53055, "rle": {"size": [424, 640], "counts": "kg_2=e<>B>B>H9G8H8H9G8H8H8H9G8H8H9G8G9H8K6IO1N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N1O1O001O1O001O001O001O001O1O001O0000001O0000000000000000000000000000001O000000001O00000000001O0001O0001O00000000001O000000000000000000001O000000000000000000000000000000000000000000000000000001O00000O1000000O1000000O100000000O100O100O1O100O100O1O100O1O100O1O100O1O100O1O10000O10001N1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O100O10000O10001N100O10000O10000O10000O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100000000O10001O0000000000000000000O10000000000000000000000O10000000000000000O1000000O1000000O1000O10O10O1000O010000O10O10O1000O10O1000O010000O01000O10O10O10000O01000O10O10O1000O10O1000O010000O01000O10O10O10O10O1000O01O1N2N1O2N2N2N101N2N2N1O2N2N1O2N2N2N101N2N2N1O2O1000O0100O10O010000O010O10O10O100O01000O10O0100O10O10O100O01000O010O100O01000O10O010000O010O1000O0100O10O010j0VOl0SOn0SOm0ROTF"}, "label": "an unmade bed"}]}
{"id": 421562, "image": "COCO_train2014_000000421562.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a a bed with a white sheet in front of a window\"."}], "task": "refering", "answer_template": "The \"a a bed with a white sheet in front of a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 191, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.299390625, 0.4834433962264151, 1.0, 0.9890330188679246], "iscrowd": 0, "area": 64246.53055, "rle": {"size": [424, 640], "counts": "kg_2=e<>B>B>H9G8H8H9G8H8H8H9G8H8H9G8G9H8K6IO1N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N1O1O001O1O001O001O001O001O1O001O0000001O0000000000000000000000000000001O000000001O00000000001O0001O0001O00000000001O000000000000000000001O000000000000000000000000000000000000000000000000000001O00000O1000000O1000000O100000000O100O100O1O100O100O1O100O1O100O1O100O1O100O1O10000O10001N1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O100O10000O10001N100O10000O10000O10000O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100000000O10001O0000000000000000000O10000000000000000000000O10000000000000000O1000000O1000000O1000O10O10O1000O010000O10O10O1000O10O1000O010000O01000O10O10O10000O01000O10O10O1000O10O1000O010000O01000O10O10O10O10O1000O01O1N2N1O2N2N2N101N2N2N1O2N2N1O2N2N2N101N2N2N1O2O1000O0100O10O010000O010O10O10O100O01000O10O0100O10O10O100O01000O010O100O01000O10O010000O010O1000O0100O10O010j0VOl0SOn0SOm0ROTF"}, "label": "a a bed with a white sheet in front of a window"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black suit with a red tie\"."}], "task": "refering", "answer_template": "The \"a man in a black suit with a red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.22540624999999997, 0.3271561771561771, 0.4269375, 1.0], "iscrowd": 0, "area": 18984.059950000006, "rle": {"size": [429, 640], "counts": "hcl18_<f0L5K4L5K5K4L5K5K4L5K4M4L3L4M3L4M3M4K4M3M3M3O1O2M2O1O1O1N3N10001O00001^IaLT3_3kLbLU3_3jLaLV3_3jLaLV3_3iLbLV3`3iL`LW3`3gLbLY3_3eLbLZ3_3eLbL[3^3dLcL]26RKX3_2cL]27TKV3\\2fL]26WKU3X2jL^2\\4]MhKb2Y4[MjKe2W4WMlKh2U4UMnKk2R4RMQLm2Q4PMQLP3o3mLTLS3l3jLWLU3k3gLXLY3h3dL[L[3g3aL\\L_3d3_L^La3b3\\LaLc3`3ZLcLf3^3VLeLj3[3SLhLm3X3PLlLn3V3oKlLQ4T3lKoLT4Q3iKRMW4o2eKTM[4l2bKWM]4m51O00001O00001iHaK^5_4bJbK]5^4bJeK\\5\\4cJfK\\5Y4dJhK[5X4dJjKW1FY1a4_MkKT10e0DRMa4U1lKR11f0DQM_4W1mKP13f0BRM_4U1PLP12h0@QM^4T1ULQ1Ni0APM\\4S1[LQ1Jj0@QM\\4P1`LQ1Fm0@PMZ4R1cLn0En0^ORM[4P1fLm0Co0\\OTM\\4;]K2\\1]1Bo0YOWM\\44gMb1iNn0TO\\M]41`Nl0TNc1PO_M]4OeNk0SNb1kNdMo7f0`H]OK]1\\OUO:\\OR8a0]HELT1_OVO6@V8<ZHOJk0CVO4DX89WH6Kb0EXO1GY88VH<J9IYONJZ87UHa0K0J\\OMKZ87THg0JGN\\OKOZ85THm0J^O>@e74SHS1IUO?Df73SHW1ImN[9KlF]1HdN]9NkFb1l9\\NUFi1e:4L5L3L5K1O1O01M2L4K5L4L4L5K4L4L4L4L5G8D_^i4"}, "label": "a man in a black suit with a red tie"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a red and silver power tie\"."}], "task": "refering", "answer_template": "The \"a man with a red and silver power tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.22540624999999997, 0.3271561771561771, 0.4269375, 1.0], "iscrowd": 0, "area": 18984.059950000006, "rle": {"size": [429, 640], "counts": "hcl18_<f0L5K4L5K5K4L5K5K4L5K4M4L3L4M3L4M3M4K4M3M3M3O1O2M2O1O1O1N3N10001O00001^IaLT3_3kLbLU3_3jLaLV3_3jLaLV3_3iLbLV3`3iL`LW3`3gLbLY3_3eLbLZ3_3eLbL[3^3dLcL]26RKX3_2cL]27TKV3\\2fL]26WKU3X2jL^2\\4]MhKb2Y4[MjKe2W4WMlKh2U4UMnKk2R4RMQLm2Q4PMQLP3o3mLTLS3l3jLWLU3k3gLXLY3h3dL[L[3g3aL\\L_3d3_L^La3b3\\LaLc3`3ZLcLf3^3VLeLj3[3SLhLm3X3PLlLn3V3oKlLQ4T3lKoLT4Q3iKRMW4o2eKTM[4l2bKWM]4m51O00001O00001iHaK^5_4bJbK]5^4bJeK\\5\\4cJfK\\5Y4dJhK[5X4dJjKW1FY1a4_MkKT10e0DRMa4U1lKR11f0DQM_4W1mKP13f0BRM_4U1PLP12h0@QM^4T1ULQ1Ni0APM\\4S1[LQ1Jj0@QM\\4P1`LQ1Fm0@PMZ4R1cLn0En0^ORM[4P1fLm0Co0\\OTM\\4;]K2\\1]1Bo0YOWM\\44gMb1iNn0TO\\M]41`Nl0TNc1PO_M]4OeNk0SNb1kNdMo7f0`H]OK]1\\OUO:\\OR8a0]HELT1_OVO6@V8<ZHOJk0CVO4DX89WH6Kb0EXO1GY88VH<J9IYONJZ87UHa0K0J\\OMKZ87THg0JGN\\OKOZ85THm0J^O>@e74SHS1IUO?Df73SHW1ImN[9KlF]1HdN]9NkFb1l9\\NUFi1e:4L5L3L5K1O1O01M2L4K5L4L4L5K4L4L4L4L5G8D_^i4"}, "label": "a man with a red and silver power tie"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a metal and marble end table with a tea cup , near a lamp\"."}], "task": "refering", "answer_template": "The \"a metal and marble end table with a tea cup , near a lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 252, 270, 273, 274, 293, 296, 297, 316, 319, 340, 341, 342], "bbox": [0.733875, 0.6185081585081584, 0.9754687499999999, 1.0], "iscrowd": 0, "area": 7099.680950000012, "rle": {"size": [429, 640], "counts": "nUU66V=7I2O1N2O004K5L4L5K4K5LZORD1j;OZD1b;OcD0X;2kDNQ;2SENi:2\\EM_:4eEMV:3oELm95VFLe94_FL\\95iFKR95RGKj85ZGKb85cGKX85lGKP86THJg76]HJ_76eHJW76nHJm66WIJe66`IJ[66iIJS66QJKj56ZJFe5:_JBa5>cJ^O]5b0hJYOj4KoG0Hl0c3UOf43kGk0l;;0000000001O0O100001O00000000000lNRDi0n;UOTDk0l;SOWDl0W<N0000001O00000000000000000000000000000000001OlF\\O_5d0XJEh5;nIOS60cI;\\6E[Id0e6\\OQIn0o6ROgHX1Y7hN^Ha1b7_NVHh1k7XNmGl1W8TNbGk1f8UNRGj1W9VNbFh1i0_MP8V3THoLl7i2XH[Mh7\\2\\HiMd7n1aHVN_7a1fHcNZ7T1jHQOV7f0oH^OQ79SIIP7OTI0R7HPI9T7GhH9\\7G`H8e7HVH9n7HmG7X8IcG8a8HZG8k8HQG8S9HiF7]9H_F8e9HWF7n9InE7V:IfE7^:I^E7f:IVE7n:InD8U;HgD9];h03L4M3M3M3L4M3M4L3L4M3M30001O0O100000001N10000O101O0O2O001N2J5N3M[]6"}, "label": "a metal and marble end table with a tea cup , near a lamp"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table directly right of man with yellow / black tie\"."}], "task": "refering", "answer_template": "The \"table directly right of man with yellow / black tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 252, 270, 273, 274, 293, 296, 297, 316, 319, 340, 341, 342], "bbox": [0.733875, 0.6185081585081584, 0.9754687499999999, 1.0], "iscrowd": 0, "area": 7099.680950000012, "rle": {"size": [429, 640], "counts": "nUU66V=7I2O1N2O004K5L4L5K4K5LZORD1j;OZD1b;OcD0X;2kDNQ;2SENi:2\\EM_:4eEMV:3oELm95VFLe94_FL\\95iFKR95RGKj85ZGKb85cGKX85lGKP86THJg76]HJ_76eHJW76nHJm66WIJe66`IJ[66iIJS66QJKj56ZJFe5:_JBa5>cJ^O]5b0hJYOj4KoG0Hl0c3UOf43kGk0l;;0000000001O0O100001O00000000000lNRDi0n;UOTDk0l;SOWDl0W<N0000001O00000000000000000000000000000000001OlF\\O_5d0XJEh5;nIOS60cI;\\6E[Id0e6\\OQIn0o6ROgHX1Y7hN^Ha1b7_NVHh1k7XNmGl1W8TNbGk1f8UNRGj1W9VNbFh1i0_MP8V3THoLl7i2XH[Mh7\\2\\HiMd7n1aHVN_7a1fHcNZ7T1jHQOV7f0oH^OQ79SIIP7OTI0R7HPI9T7GhH9\\7G`H8e7HVH9n7HmG7X8IcG8a8HZG8k8HQG8S9HiF7]9H_F8e9HWF7n9InE7V:IfE7^:I^E7f:IVE7n:InD8U;HgD9];h03L4M3M3M3L4M3M4L3L4M3M30001O0O100000001N10000O101O0O2O001N2J5N3M[]6"}, "label": "table directly right of man with yellow / black tie"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a brown and gold tie talks to a man wearing a red and blue tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a brown and gold tie talks to a man wearing a red and blue tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316], "bbox": [0.61425, 0.34543123543123544, 0.785640625, 0.9442424242424242], "iscrowd": 0, "area": 16832.379599999993, "rle": {"size": [429, 640], "counts": "imT51[=2M3N1N3jNISE9g:NUE5e:2XE0b:8YEKa:<\\EF^:a0^EB[:f0bE[OY:l0dEVOV:Q1fEROT:U1iEmNQ:[2K5O1O1N101O001N2O001N2O001N101O1N101O1N101O4WOj0H9H8G8H9I7J5J7I7I6J7I6K3L3N3M3M3MO1O1O0O2O1iM_JaLb5`3aJ[L`5f3cJUL^5l3eJoK]5R4eJhK]5Z4eJaK\\5c4dJXK]5l4cJoJ^5U5bJfJ_5]5aJ_J`5e5`JUJb5o5^JlId5V6]JeIf5]6ZJ^Ih5e6`0UJ]Ik4b6lJeIW5Y6`JoIb5Q6WJUJi5g62N1O2O1N1O2N101N1OO101iLXIGh67jIYOV6f0ZJkNg5S1bJeN^5Z1gJbNY5]1lJ_NT5`1QK[NQ5c1TKYNl4e1ZKWNf4h1_KTNa4k1dKQN]4m1hKoMX4P2mKlMS4S2RLiMn3U2XLfMQ4Q2U4@>B<Da0_OdTi1"}, "label": "a man wearing a brown and gold tie talks to a man wearing a red and blue tie"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian man in a black suit and tie sitting on a couch\"."}], "task": "refering", "answer_template": "The \"an asian man in a black suit and tie sitting on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316], "bbox": [0.61425, 0.34543123543123544, 0.785640625, 0.9442424242424242], "iscrowd": 0, "area": 16832.379599999993, "rle": {"size": [429, 640], "counts": "imT51[=2M3N1N3jNISE9g:NUE5e:2XE0b:8YEKa:<\\EF^:a0^EB[:f0bE[OY:l0dEVOV:Q1fEROT:U1iEmNQ:[2K5O1O1N101O001N2O001N2O001N101O1N101O1N101O4WOj0H9H8G8H9I7J5J7I7I6J7I6K3L3N3M3M3MO1O1O0O2O1iM_JaLb5`3aJ[L`5f3cJUL^5l3eJoK]5R4eJhK]5Z4eJaK\\5c4dJXK]5l4cJoJ^5U5bJfJ_5]5aJ_J`5e5`JUJb5o5^JlId5V6]JeIf5]6ZJ^Ih5e6`0UJ]Ik4b6lJeIW5Y6`JoIb5Q6WJUJi5g62N1O2O1N1O2N101N1OO101iLXIGh67jIYOV6f0ZJkNg5S1bJeN^5Z1gJbNY5]1lJ_NT5`1QK[NQ5c1TKYNl4e1ZKWNf4h1_KTNa4k1dKQN]4m1hKoMX4P2mKlMS4S2RLiMn3U2XLfMQ4Q2U4@>B<Da0_OdTi1"}, "label": "an asian man in a black suit and tie sitting on a couch"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coffee table to the left of the other coffee table\"."}], "task": "refering", "answer_template": "The \"a coffee table to the left of the other coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 284, 287, 288, 308, 310, 311, 331, 333, 334], "bbox": [0.353875, 0.6626573426573426, 0.5939687499999999, 0.9993006993006991], "iscrowd": 0, "area": 7278.84435, "rle": {"size": [429, 640], "counts": "Son22Z=7I3M2M4M2N2N5L4L4L4K5L^OSDHi;9YDHc;8aDH[;8iDHR;9REGj:9ZEGb:9aEH[:8iEHS:8QFHk98YFHc98`FI[98iFHS98QGHk88YGHc88`GI\\88gGHU88oGHm78WHHe78_HH]78fHIV77nHIn67VIIf67^II^67eIHX69lICT6=PJ_OP6a0TJ\\Ok5d0YJXOg5i0[JTOe5l0^3000O1000000001O000000000[OdC:\\<CfC=[<@gC`0Z<^OgCb0Y<\\OiCd0`<0O10000O100000001O0000000000000HZOiCf0V<]OhCc0W<:00000001O0fFXOo5h0eID[6<YI1f6OnH=R7CbHi0^7WOUHV1k7jNiGb1W8^N]Gn1c8^101O01M2D<D<D<D<C=D<D<DZGgNP6P1RJWOn5i0jI_OV6a0bIG^69ZIOf61RI8m6HkH?V7AbHc0b7]OWHe0n7[OjGi0Z8WO^Gl0g8TOQGo0T9QOdFS1a9lNWFW1n9P14K5J6H8H8G9H8H8H8G8M400000000O101N100O101N100O101N100O100O2M2O1O1O2N1O3M4L4LZd\\3"}, "label": "a coffee table to the left of the other coffee table"}]}
{"id": 20156, "image": "COCO_train2014_000000020156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small round table on the left\"."}], "task": "refering", "answer_template": "The \"the small round table on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 284, 287, 288, 308, 310, 311, 331, 333, 334], "bbox": [0.353875, 0.6626573426573426, 0.5939687499999999, 0.9993006993006991], "iscrowd": 0, "area": 7278.84435, "rle": {"size": [429, 640], "counts": "Son22Z=7I3M2M4M2N2N5L4L4L4K5L^OSDHi;9YDHc;8aDH[;8iDHR;9REGj:9ZEGb:9aEH[:8iEHS:8QFHk98YFHc98`FI[98iFHS98QGHk88YGHc88`GI\\88gGHU88oGHm78WHHe78_HH]78fHIV77nHIn67VIIf67^II^67eIHX69lICT6=PJ_OP6a0TJ\\Ok5d0YJXOg5i0[JTOe5l0^3000O1000000001O000000000[OdC:\\<CfC=[<@gC`0Z<^OgCb0Y<\\OiCd0`<0O10000O100000001O0000000000000HZOiCf0V<]OhCc0W<:00000001O0fFXOo5h0eID[6<YI1f6OnH=R7CbHi0^7WOUHV1k7jNiGb1W8^N]Gn1c8^101O01M2D<D<D<D<C=D<D<DZGgNP6P1RJWOn5i0jI_OV6a0bIG^69ZIOf61RI8m6HkH?V7AbHc0b7]OWHe0n7[OjGi0Z8WO^Gl0g8TOQGo0T9QOdFS1a9lNWFW1n9P14K5J6H8H8G9H8H8H8G8M400000000O101N100O101N100O101N100O100O2M2O1O1O2N1O3M4L4LZd\\3"}, "label": "the small round table on the left"}]}
{"id": 405181, "image": "COCO_train2014_000000405181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car speeding through an intersection\"."}], "task": "refering", "answer_template": "The \"a black car speeding through an intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 158, 159, 160, 161, 176, 177, 178, 179, 197, 215], "bbox": [0.81236, 0.5668533333333333, 0.99606, 0.87248], "iscrowd": 0, "area": 6436.229599999997, "rle": {"size": [375, 500], "counts": "Pnd42a;5K6J5K6J5K6J5K5N101N1O1O1O2N100O1O2N1O1O101N1O1O1O2O0O10001O0000001O0000001O0000001O0000001O0000001O0000001O0O10001O0000001O001O1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1Y1gN^?"}, "label": "a black car speeding through an intersection"}]}
{"id": 405181, "image": "COCO_train2014_000000405181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurred image of front of car\"."}], "task": "refering", "answer_template": "The \"a blurred image of front of car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 158, 159, 160, 161, 176, 177, 178, 179, 197, 215], "bbox": [0.81236, 0.5668533333333333, 0.99606, 0.87248], "iscrowd": 0, "area": 6436.229599999997, "rle": {"size": [375, 500], "counts": "Pnd42a;5K6J5K6J5K6J5K5N101N1O1O1O2N100O1O2N1O1O101N1O1O1O2O0O10001O0000001O0000001O0000001O0000001O0000001O0000001O0O10001O0000001O001O1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1Y1gN^?"}, "label": "a blurred image of front of car"}]}
{"id": 77504, "image": "COCO_train2014_000000077504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane with people getting on board\"."}], "task": "refering", "answer_template": "The \"an airplane with people getting on board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 96, 97, 98, 99, 100, 101, 102, 103, 111, 112, 117, 118, 119, 120, 121, 122, 123, 126, 127, 128, 129, 130, 131, 132, 134, 135, 139, 140, 141, 142, 143, 144, 145, 146, 147, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 198, 199, 200, 201, 202, 203, 213, 214, 222, 223, 224, 225, 236, 237], "bbox": [0.076390625, 0.241288056206089, 1.0, 0.68903981264637], "iscrowd": 0, "area": 41712.05884999999, "rle": {"size": [427, 640], "counts": "ccd08o<7K4M2N2N3L3N3M2N2N2O0O2N2N2O1N1O2N101N101O0O2O000O2O001N100O2O000O2O000O2O000O2O000O101N10001N1O101N1O1O2N1O1O2O0O1O2N1O1O1O2O0O1O1O2N1O1O1O2O0O1O100O2O0O1O101N100O101O0O10001O0O1000000O101O0O1000000O10001O000000000000001O000000000000001O000O100000001O000000000000001O0gGZLm6f3QI_Ll6a3SIcLj6]3TIgLj6Y3TIiLl6W3SIkLl6V3SIjLm6X3PIiLP7Z3mHgLR7[3lHeLT7]3iHdLW7^3[HmK3e0b7`3YHlK3e0d7`3XHlK3d0e7b3VHlK3b0g7c3UHhLk7\\4O100000O100O10000\\OTHhKl7X4VHfKj7m3THPL21k7o3UHnK21i7Q4WHmK10g7S4[HkK0Of7V4\\HjKR8V4a0O<E2N1O1KdFYL^9i1oFET9BhG;X8DjG;W8CkG<U8CmG<S8CnG=R8BoG>Q8BnG?R8@oG`0R8_OnGa0R8_OnGa0R8^OoGb0Q8^OoGb0Q8]OoGd0Q8[OPHe0Q8XOQHh0o7WORHi0n7UOTHk0l7TOUHm0j7SOUHn0k7oNXHQ1i7mNXHR1i7nNWHR1i7nNWHR1i7nNWHR1i7nNVHS1j7nNUHR1l7ROoGm0R8UOlGk0T8VOkGj0U8VOkGj0U8ROnGo0R8nNQHR1P8lNQHS1P8kNRHU1n7kNRHU1n7lNQHT1o7lNPHU1P8mNnGS1S8mNlGR1U8oNjGQ1V8oNjGQ1V8jNUFMf1X1U8jNWFLg1X1S8iNUHV1k7iNWHV1i7jNXHU1h7kNYHT1g7lNZHS1f7mN\\HQ1e7oN[HP1e7QO[Hn0e7VOXHi0h7WOYHh0g7YOYHc0j7]OWH?m7@UH;n7AWH<k7CWH9l7GTH6o7Jj2O[=0eB4L1O100O1O000010O01OO10001O000001O01O000002N3N1N1O2N1O100O1O2N1O100ON2N2N2N30O1O1O00101N1O2N1\\C]O^<m0L4KN2N2N2N3M200O10001O01O9G1O100O1O1O1O:G0O1O1O1O10O01O1O1OK6N1O100O1100O001O1XDbNa;d110O001OO1O2N1O1001O010O2N2N2N3N4K5K5K1O100O1O1O1O100O1O1O1O100O1O1O1O1O100OI`EdM_:\\2dE`M]:_28O2N100000001O0001O2N2N2O1N1O1O1O2O1N2N3M9G100O1O1O001O100O1O1O0010O000001O01O01OB>M3N2N3O0O100N2O2M2O1O1O1O2O000000001O0001O001O101N3M2N5K6K2M1O0000102M2N1O1O2N3N0O001O0010O000000001O000O1O1O1O2]OUF_MP:`2RFZMS:e2:O1O2N1O1O1O101N1^OVEbNk:^1VEaNj:^1XE`Ni:`1XE_Nh:`1ZEXNn:g1;00O100O01N2O001O0O2O1M2O2N1N3N2N1N3N1O2M3N1O2M2O2N1N10000O2O000O2O3M3L2O1N10112O1N1OPO\\FXNf9^1dF`N]9V1mFiNT9m0VGROk8e0_GYOc8<gGCZ84PHJQ8LYH4h7BaH>W:O001O001O000001O0000001O001O00001O0O2O001O00001O001O001O000O2OO10000O10000O1G\\CKd<5^CIb<7_CGb<9`CE_<<8O1000000000000000000000O1000000000GXJ"}, "label": "an airplane with people getting on board"}]}
{"id": 77504, "image": "COCO_train2014_000000077504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane being boarded\"."}], "task": "refering", "answer_template": "The \"an airplane being boarded\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 96, 97, 98, 99, 100, 101, 102, 103, 111, 112, 117, 118, 119, 120, 121, 122, 123, 126, 127, 128, 129, 130, 131, 132, 134, 135, 139, 140, 141, 142, 143, 144, 145, 146, 147, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 198, 199, 200, 201, 202, 203, 213, 214, 222, 223, 224, 225, 236, 237], "bbox": [0.076390625, 0.241288056206089, 1.0, 0.68903981264637], "iscrowd": 0, "area": 41712.05884999999, "rle": {"size": [427, 640], "counts": "ccd08o<7K4M2N2N3L3N3M2N2N2O0O2N2N2O1N1O2N101N101O0O2O000O2O001N100O2O000O2O000O2O000O2O000O101N10001N1O101N1O1O2N1O1O2O0O1O2N1O1O1O2O0O1O1O2N1O1O1O2O0O1O100O2O0O1O101N100O101O0O10001O0O1000000O101O0O1000000O10001O000000000000001O000000000000001O000O100000001O000000000000001O0gGZLm6f3QI_Ll6a3SIcLj6]3TIgLj6Y3TIiLl6W3SIkLl6V3SIjLm6X3PIiLP7Z3mHgLR7[3lHeLT7]3iHdLW7^3[HmK3e0b7`3YHlK3e0d7`3XHlK3d0e7b3VHlK3b0g7c3UHhLk7\\4O100000O100O10000\\OTHhKl7X4VHfKj7m3THPL21k7o3UHnK21i7Q4WHmK10g7S4[HkK0Of7V4\\HjKR8V4a0O<E2N1O1KdFYL^9i1oFET9BhG;X8DjG;W8CkG<U8CmG<S8CnG=R8BoG>Q8BnG?R8@oG`0R8_OnGa0R8_OnGa0R8^OoGb0Q8^OoGb0Q8]OoGd0Q8[OPHe0Q8XOQHh0o7WORHi0n7UOTHk0l7TOUHm0j7SOUHn0k7oNXHQ1i7mNXHR1i7nNWHR1i7nNWHR1i7nNWHR1i7nNVHS1j7nNUHR1l7ROoGm0R8UOlGk0T8VOkGj0U8VOkGj0U8ROnGo0R8nNQHR1P8lNQHS1P8kNRHU1n7kNRHU1n7lNQHT1o7lNPHU1P8mNnGS1S8mNlGR1U8oNjGQ1V8oNjGQ1V8jNUFMf1X1U8jNWFLg1X1S8iNUHV1k7iNWHV1i7jNXHU1h7kNYHT1g7lNZHS1f7mN\\HQ1e7oN[HP1e7QO[Hn0e7VOXHi0h7WOYHh0g7YOYHc0j7]OWH?m7@UH;n7AWH<k7CWH9l7GTH6o7Jj2O[=0eB4L1O100O1O000010O01OO10001O000001O01O000002N3N1N1O2N1O100O1O2N1O100ON2N2N2N30O1O1O00101N1O2N1\\C]O^<m0L4KN2N2N2N3M200O10001O01O9G1O100O1O1O1O:G0O1O1O1O10O01O1O1OK6N1O100O1100O001O1XDbNa;d110O001OO1O2N1O1001O010O2N2N2N3N4K5K5K1O100O1O1O1O100O1O1O1O100O1O1O1O1O100OI`EdM_:\\2dE`M]:_28O2N100000001O0001O2N2N2O1N1O1O1O2O1N2N3M9G100O1O1O001O100O1O1O0010O000001O01O01OB>M3N2N3O0O100N2O2M2O1O1O1O2O000000001O0001O001O101N3M2N5K6K2M1O0000102M2N1O1O2N3N0O001O0010O000000001O000O1O1O1O2]OUF_MP:`2RFZMS:e2:O1O2N1O1O1O101N1^OVEbNk:^1VEaNj:^1XE`Ni:`1XE_Nh:`1ZEXNn:g1;00O100O01N2O001O0O2O1M2O2N1N3N2N1N3N1O2M3N1O2M2O2N1N10000O2O000O2O3M3L2O1N10112O1N1OPO\\FXNf9^1dF`N]9V1mFiNT9m0VGROk8e0_GYOc8<gGCZ84PHJQ8LYH4h7BaH>W:O001O001O000001O0000001O001O00001O0O2O001O00001O001O001O000O2OO10000O10000O1G\\CKd<5^CIb<7_CGb<9`CE_<<8O1000000000000000000000O1000000000GXJ"}, "label": "an airplane being boarded"}]}
{"id": 282310, "image": "COCO_train2014_000000282310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player base - man tagging a runner\"."}], "task": "refering", "answer_template": "The \"baseball player base - man tagging a runner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 130, 131, 132, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 216, 217, 218, 219, 220, 228, 229, 230, 234, 235, 236, 237, 246, 251, 252, 253, 254, 268, 269, 270, 271, 284, 285, 286, 301, 302], "bbox": [0.40054166666666663, 0.25621875, 1.0, 0.7532500000000001], "iscrowd": 0, "area": 32556.726199999997, "rle": {"size": [640, 480], "counts": "V[h31nc03L3M4M2M3M4M2M4N1O2O0O01O00001O00001O001N10001O01O101N100O1O2O0O1O101N100O1O2O0O1O10O0000001O0000000O101O0000O1O1N2N2O1NF:L51N1O100O1O2N100O1O101N11O0010O01O1O010O001O010O0010O01O010O001O01N1N2O2M2N2N2N3N1N2N3M2N2O1N3M2N2O1O1O1O1O1O1XNkMeAV2Z>nMbAS2]>QN_AP2a>SN\\Am1c>VN[Aj1d>ZNYAf1e>_NYA`1f>dNWA]1f>hNXAW1f>mNXAS1g>QOVAo0h>VOVAi0i>[OTAe0j>@TA?j>FSA:m>HQA8P?Jn@5R?Nk@2V?0h@OY?3d@M\\?[21O10O100O10000O101N10000O100O2O2O000`EWKg5j4WJWKj5j4RJYKn5g4oI[KR6e4lI]KT6c4iI_KX6a4eIbK[6_4bIcK]6^4`IdKa6\\4\\IgKd6Y4ZIhKg6Y4UIjKk6V4TIkKl6U4RIlKo6T4PImKo6U4oHkKR7U4lHmKT7S4kHmKV7S4iHnKW7R4gHnK[7S4cHmK^7S4aHlK`7U4^HlKc7T4\\HkKf74VEo2R3mLi72VER3o2lLP8LTEX3k2lLW8EPE_3h2kL`8]OjDj3c2jLi8UOfDQ4IULZ2d0f9g3WFYLj9h3SFXLn9k3[E`KkNe0k;m3REhKnN:Q<P4iDoLV;m501O00001O001O001O0101N2O0O2O1N101N2O1N101N2O0O2O1N2N101N2O0O2O1N101N2O1N101N2N1I8H8J5K6J6J6J5O2N2O001N2O001N2O1O0O2O1O010O1O001O100dKn@h3R?TLUAi3l>QL[Am3_?M2N3N1N3M2O2N1N3N1PM__O^2b`0XMh_Og2j`0O0O2O1O001O1N101O1O001N2O001O1O00oD"}, "label": "baseball player base - man tagging a runner"}]}
{"id": 282310, "image": "COCO_train2014_000000282310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player on right\"."}], "task": "refering", "answer_template": "The \"player on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 130, 131, 132, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 216, 217, 218, 219, 220, 228, 229, 230, 234, 235, 236, 237, 246, 251, 252, 253, 254, 268, 269, 270, 271, 284, 285, 286, 301, 302], "bbox": [0.40054166666666663, 0.25621875, 1.0, 0.7532500000000001], "iscrowd": 0, "area": 32556.726199999997, "rle": {"size": [640, 480], "counts": "V[h31nc03L3M4M2M3M4M2M4N1O2O0O01O00001O00001O001N10001O01O101N100O1O2O0O1O101N100O1O2O0O1O10O0000001O0000000O101O0000O1O1N2N2O1NF:L51N1O100O1O2N100O1O101N11O0010O01O1O010O001O010O0010O01O010O001O01N1N2O2M2N2N2N3N1N2N3M2N2O1N3M2N2O1O1O1O1O1O1XNkMeAV2Z>nMbAS2]>QN_AP2a>SN\\Am1c>VN[Aj1d>ZNYAf1e>_NYA`1f>dNWA]1f>hNXAW1f>mNXAS1g>QOVAo0h>VOVAi0i>[OTAe0j>@TA?j>FSA:m>HQA8P?Jn@5R?Nk@2V?0h@OY?3d@M\\?[21O10O100O10000O101N10000O100O2O2O000`EWKg5j4WJWKj5j4RJYKn5g4oI[KR6e4lI]KT6c4iI_KX6a4eIbK[6_4bIcK]6^4`IdKa6\\4\\IgKd6Y4ZIhKg6Y4UIjKk6V4TIkKl6U4RIlKo6T4PImKo6U4oHkKR7U4lHmKT7S4kHmKV7S4iHnKW7R4gHnK[7S4cHmK^7S4aHlK`7U4^HlKc7T4\\HkKf74VEo2R3mLi72VER3o2lLP8LTEX3k2lLW8EPE_3h2kL`8]OjDj3c2jLi8UOfDQ4IULZ2d0f9g3WFYLj9h3SFXLn9k3[E`KkNe0k;m3REhKnN:Q<P4iDoLV;m501O00001O001O001O0101N2O0O2O1N101N2O1N101N2O0O2O1N2N101N2O0O2O1N101N2O1N101N2N1I8H8J5K6J6J6J5O2N2O001N2O001N2O1O0O2O1O010O1O001O100dKn@h3R?TLUAi3l>QL[Am3_?M2N3N1N3M2O2N1N3N1PM__O^2b`0XMh_Og2j`0O0O2O1O001O1N101O1O001N2O001O1O00oD"}, "label": "player on right"}]}
{"id": 282310, "image": "COCO_train2014_000000282310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"that is a baseball player sliding into home plate\"."}], "task": "refering", "answer_template": "The \"that is a baseball player sliding into home plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 86, 87, 102, 103, 119, 120, 122, 123, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 227, 228, 238, 239, 240, 255, 256, 257, 273], "bbox": [0.003, 0.213484375, 0.4704166666666667, 0.701125], "iscrowd": 0, "area": 27623.6557, "rle": {"size": [640, 480], "counts": "fn01lc0a0_Oa0SLnNmCc1`;dN[Dm1Q;\\NiDT2e:SNVE]2W:kMdEf2i9aMRFc2g9fMSF]2h9jMSFZ2f9nMUFU2h9QNRFR2m9QNnER2P:RNkER2S:RNgEQ2X:RNcEQ2\\:RN_EQ2`:RNZES2c:RNWEQ2h:RNSEQ2l:RNoDQ2P;SNjDP2U;_3M2O1O1O2N1N2O1O^HUER6i:mI[EQ6d:nI_EQ6_:oIeEo5[:oIhEQ6V:nInEP6Q:oIRFP6n9nIVFP6i9oIZFP6e9PJ^Fo5a9PJaFo5^9PJeFo5[9PJhFn5W9QJlFo5R9QJQGl5P9SJRGl5m8SJWGj5i8VJYGh5h8VJ\\Gh5c8XJ^Gg5c8XJ^Gg5b8XJaGf5_8ZJbGf5^8YJdGe5\\8[JeGd5[8\\JgGc5Y8\\JhGc5Y8\\JhGc5Y8]JhGa5Y8^JhGb5X8]JjGa5W8^JjGa5W8^JkGa5U8^JlGa5U8^JlGa5U8^JmG`5U8_JkGa5U8^JmG`5T8_JmG`5T8_JlG\\5Z8cJfGY5_8fJ`GV5h8gJXGU5o8iJPGR5X9kJhFP5`9mJ`Fo4Y<L5N2001O1O1N2O001O1O1N2O1O001O1N2O1O001O1N2O100O0010O1O1O2O0O2N2N1O2N2N1O2N1O2N2N1XLk@d2W?YMm@d2T?ZMQAc2P?ZMUA_2P?^MUAU2`?^Md@V2k`0F:M2O2M3N1N101N101N2O0O2O1N101N2O0O2O1N101N101N2O00001O00000O2O00001O0000001N10001O0000001O0O101O0000001O0O100000000000O10O1000000000O01O100O100O1O100O1O2O0O100O1O2O0O1O100O2O0O1O100O2N100O100O2N100Oion4"}, "label": "that is a baseball player sliding into home plate"}]}
{"id": 282310, "image": "COCO_train2014_000000282310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player with number 13 on uniform\"."}], "task": "refering", "answer_template": "The \"baseball player with number 13 on uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 86, 87, 102, 103, 119, 120, 122, 123, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 227, 228, 238, 239, 240, 255, 256, 257, 273], "bbox": [0.003, 0.213484375, 0.4704166666666667, 0.701125], "iscrowd": 0, "area": 27623.6557, "rle": {"size": [640, 480], "counts": "fn01lc0a0_Oa0SLnNmCc1`;dN[Dm1Q;\\NiDT2e:SNVE]2W:kMdEf2i9aMRFc2g9fMSF]2h9jMSFZ2f9nMUFU2h9QNRFR2m9QNnER2P:RNkER2S:RNgEQ2X:RNcEQ2\\:RN_EQ2`:RNZES2c:RNWEQ2h:RNSEQ2l:RNoDQ2P;SNjDP2U;_3M2O1O1O2N1N2O1O^HUER6i:mI[EQ6d:nI_EQ6_:oIeEo5[:oIhEQ6V:nInEP6Q:oIRFP6n9nIVFP6i9oIZFP6e9PJ^Fo5a9PJaFo5^9PJeFo5[9PJhFn5W9QJlFo5R9QJQGl5P9SJRGl5m8SJWGj5i8VJYGh5h8VJ\\Gh5c8XJ^Gg5c8XJ^Gg5b8XJaGf5_8ZJbGf5^8YJdGe5\\8[JeGd5[8\\JgGc5Y8\\JhGc5Y8\\JhGc5Y8]JhGa5Y8^JhGb5X8]JjGa5W8^JjGa5W8^JkGa5U8^JlGa5U8^JlGa5U8^JmG`5U8_JkGa5U8^JmG`5T8_JmG`5T8_JlG\\5Z8cJfGY5_8fJ`GV5h8gJXGU5o8iJPGR5X9kJhFP5`9mJ`Fo4Y<L5N2001O1O1N2O001O1O1N2O1O001O1N2O1O001O1N2O100O0010O1O1O2O0O2N2N1O2N2N1O2N1O2N2N1XLk@d2W?YMm@d2T?ZMQAc2P?ZMUA_2P?^MUAU2`?^Md@V2k`0F:M2O2M3N1N101N101N2O0O2O1N101N2O0O2O1N101N101N2O00001O00000O2O00001O0000001N10001O0000001O0O101O0000001O0O100000000000O10O1000000000O01O100O100O1O100O1O2O0O100O1O2O0O1O100O2O0O1O100O2N100O100O2N100Oion4"}, "label": "baseball player with number 13 on uniform"}]}
{"id": 528071, "image": "COCO_train2014_000000528071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant who is facing another elephant in some shrubbery\"."}], "task": "refering", "answer_template": "The \"an elephant who is facing another elephant in some shrubbery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 340, 341, 342, 343, 344], "bbox": [0.54575, 0.0022482435597189696, 1.0, 0.986510538641686], "iscrowd": 0, "area": 100427.136, "rle": {"size": [427, 640], "counts": "ima42V=3L4L5K4M3N3L3N2N3L3N3WGlNV5W1hJmNT5U1jJoNR5T1lJoNo4T1oJPOm4R1RKQOi4S1TKQOh4Q1PKYOl4i0kJBP5b0fJIV59bJ1e2QNnMQ2[O2T2ROkMl025`13iMIg07m0U1eMeN_199V4HmKEb4<aKQOo4o0e31O001O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O2O0O1OTOaEPO]:g0nEXOR:>YFBe94gFKX9LSG4k8C`G=^8ZOmGe0]:00O10000O11O001O001O001O001O001O001O001O00001O001O000000O100000000000000O1000000000000O100000000OD=A?A?B=B?B>L4K4L5L4K4L5L4K5L3L5K5L4K4L5L4K5L34M3M4K5L4L4L3L5L4L4L4K7J>B>B>A?B>B>A?B>B4L0O1000000O1000000O10000O1000000O1000000O1000000O1000O0100001O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O0O2O001O1O001O001O001O1O001O001OO1000000O10000O1000000O10000O1000000O10000O10O1000O1000M2O2M3L4L3M4L4L5L3L4L4L4L5K4L4M3L4L4L5K4L4M3L4L5K4\\Nd1kJ]4"}, "label": "an elephant who is facing another elephant in some shrubbery"}]}
{"id": 528071, "image": "COCO_train2014_000000528071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big elephant is sleeping\"."}], "task": "refering", "answer_template": "The \"the big elephant is sleeping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 340, 341, 342, 343, 344], "bbox": [0.54575, 0.0022482435597189696, 1.0, 0.986510538641686], "iscrowd": 0, "area": 100427.136, "rle": {"size": [427, 640], "counts": "ima42V=3L4L5K4M3N3L3N2N3L3N3WGlNV5W1hJmNT5U1jJoNR5T1lJoNo4T1oJPOm4R1RKQOi4S1TKQOh4Q1PKYOl4i0kJBP5b0fJIV59bJ1e2QNnMQ2[O2T2ROkMl025`13iMIg07m0U1eMeN_199V4HmKEb4<aKQOo4o0e31O001O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O2O0O1OTOaEPO]:g0nEXOR:>YFBe94gFKX9LSG4k8C`G=^8ZOmGe0]:00O10000O11O001O001O001O001O001O001O001O00001O001O000000O100000000000000O1000000000000O100000000OD=A?A?B=B?B>L4K4L5L4K4L5L4K5L3L5K5L4K4L5L4K5L34M3M4K5L4L4L3L5L4L4L4K7J>B>B>A?B>B>A?B>B4L0O1000000O1000000O10000O1000000O1000000O1000000O1000O0100001O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O0O2O001O1O001O001O001O1O001O001OO1000000O10000O1000000O10000O1000000O10000O10O1000O1000M2O2M3L4L3M4L4L5L3L4L4L4L5K4L4M3L4L4L5K4L4M3L4L5K4\\Nd1kJ]4"}, "label": "the big elephant is sleeping"}]}
{"id": 528071, "image": "COCO_train2014_000000528071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant on the left\"."}], "task": "refering", "answer_template": "The \"the elephant on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 241, 242, 243, 253, 254, 255, 257, 258, 259, 260, 261, 264, 265, 276, 277, 280, 281, 282, 283, 287, 288, 299, 303, 304, 305, 306, 310, 311, 327, 328, 329, 333, 334], "bbox": [0.0, 0.10697892271662764, 0.589046875, 0.9898594847775176], "iscrowd": 0, "area": 99140.95520000001, "rle": {"size": [427, 640], "counts": "e1V:U3O10O0100O10O0100O00100O10O0100O1O100O100O100O1O100O100O1O1O1N2O1N2O1O1N2O1O1N2O1N3I6J6J6L4O1000000O101O00000O1000000O101O00000O100000000O2O00000O2O0000000O2O00000O101O00000O2O0000000O2O00000O2O0000000O2O00000O101O0O100O2N100002N2N3M2N2N2N3M2N2N3M2N2N3M2N2N1O2N2N5K7I6J7I6J7H8I6J7I6I8I6J000O100000000O1000000O1000000O010000000O1000000O100000000O1000000O1000000O10000000000000000000000000000000ZOf0M3N2M3M3N2M2O2M3N3L3N2M4L3N2M3N3L3N2M4M2M3M3N2M3N3L3N2L4K5L4K5K5K5L4K5jNV1M3N2N2M3N2N2M3N2N2N2M3N2N2M3N2N1N3N101O1O010O010O0010O0100O0010O011N101O1O0010O010O01O010O1O010O010O0010O01O010O00010O010O00010O0010O0100O1O10OmMoJlKR5n3SKRLn4i3VKXLj4b3[K^Lf4\\3_KdLb4V3cKkL]4o2hKQMX4k2lKUMU4e2PL[MQ4_2TLbMl3X2YLhMh3T2[LlMe3P2_LQNa3k1bLUN^3g1fLYNZ3d1iL]NV3_1nLaNR3\\1QMdNo2X1UMhNk2U1XMlNg2P1]MPOe2k0^MUOf2c0^M^Od28cMHa2IjM9X2XOQNk0Q2gNXN]1j1TN_NP2c1aMeNc2o54L3M4K4M4L3M4L3L5L3M4L3L5I7I7J5J7I7I7J6I6J7J6I7J6K5`IkIb5W7F9G:F9G6J6J6J6J6J6K4K6J6J6Jb0^OX1hNX^]3"}, "label": "the elephant on the left"}]}
{"id": 528071, "image": "COCO_train2014_000000528071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is the elephant on the left\"."}], "task": "refering", "answer_template": "The \"this is the elephant on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 241, 242, 243, 253, 254, 255, 257, 258, 259, 260, 261, 264, 265, 276, 277, 280, 281, 282, 283, 287, 288, 299, 303, 304, 305, 306, 310, 311, 327, 328, 329, 333, 334], "bbox": [0.0, 0.10697892271662764, 0.589046875, 0.9898594847775176], "iscrowd": 0, "area": 99140.95520000001, "rle": {"size": [427, 640], "counts": "e1V:U3O10O0100O10O0100O00100O10O0100O1O100O100O100O1O100O100O1O1O1N2O1N2O1O1N2O1O1N2O1N3I6J6J6L4O1000000O101O00000O1000000O101O00000O100000000O2O00000O2O0000000O2O00000O101O00000O2O0000000O2O00000O2O0000000O2O00000O101O0O100O2N100002N2N3M2N2N2N3M2N2N3M2N2N3M2N2N1O2N2N5K7I6J7I6J7H8I6J7I6I8I6J000O100000000O1000000O1000000O010000000O1000000O100000000O1000000O1000000O10000000000000000000000000000000ZOf0M3N2M3M3N2M2O2M3N3L3N2M4L3N2M3N3L3N2M4M2M3M3N2M3N3L3N2L4K5L4K5K5K5L4K5jNV1M3N2N2M3N2N2M3N2N2N2M3N2N2M3N2N1N3N101O1O010O010O0010O0100O0010O011N101O1O0010O010O01O010O1O010O010O0010O01O010O00010O010O00010O0010O0100O1O10OmMoJlKR5n3SKRLn4i3VKXLj4b3[K^Lf4\\3_KdLb4V3cKkL]4o2hKQMX4k2lKUMU4e2PL[MQ4_2TLbMl3X2YLhMh3T2[LlMe3P2_LQNa3k1bLUN^3g1fLYNZ3d1iL]NV3_1nLaNR3\\1QMdNo2X1UMhNk2U1XMlNg2P1]MPOe2k0^MUOf2c0^M^Od28cMHa2IjM9X2XOQNk0Q2gNXN]1j1TN_NP2c1aMeNc2o54L3M4K4M4L3M4L3L5L3M4L3L5I7I7J5J7I7I7J6I6J7J6I7J6K5`IkIb5W7F9G:F9G6J6J6J6J6J6K4K6J6J6Jb0^OX1hNX^]3"}, "label": "this is the elephant on the left"}]}
{"id": 503500, "image": "COCO_train2014_000000503500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle on the far left that is behind two older gentlemen\"."}], "task": "refering", "answer_template": "The \"a motorcycle on the far left that is behind two older gentlemen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 138, 141, 142, 161, 164, 165, 184, 187, 188, 207, 210, 211], "bbox": [0.0, 0.2847702407002188, 0.197171875, 0.6094967177242887], "iscrowd": 0, "area": 4232.59115, "rle": {"size": [457, 640], "counts": "U4<\\2R1e7EiFX1X9l14L3M4L3M3M2N2N3M2N2N2M4bMiEc0S<L5J5K6J5L_bl0^OX]SO6UCLg<m0L4K5WOSORDQ1m;f0O1O1O1O1O200O001O10O01O0001N1N2N3N12N2N2O1N2N2N06K7I8H7H8I8H7H7J5K6I7JoSU7"}, "label": "a motorcycle on the far left that is behind two older gentlemen"}]}
{"id": 503500, "image": "COCO_train2014_000000503500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike in the left to the back\"."}], "task": "refering", "answer_template": "The \"bike in the left to the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 138, 141, 142, 161, 164, 165, 184, 187, 188, 207, 210, 211], "bbox": [0.0, 0.2847702407002188, 0.197171875, 0.6094967177242887], "iscrowd": 0, "area": 4232.59115, "rle": {"size": [457, 640], "counts": "U4<\\2R1e7EiFX1X9l14L3M4L3M3M2N2N3M2N2N2M4bMiEc0S<L5J5K6J5L_bl0^OX]SO6UCLg<m0L4K5WOSORDQ1m;f0O1O1O1O1O200O001O10O01O0001N1N2N3N12N2N2O1N2N2N06K7I8H7H8I8H7H7J5K6I7JoSU7"}, "label": "bike in the left to the back"}]}
{"id": 503500, "image": "COCO_train2014_000000503500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcycle\"."}], "task": "refering", "answer_template": "The \"motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 85, 86, 87, 88, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 365, 366], "bbox": [0.49582812499999995, 0.17940919037199124, 0.9836406249999999, 0.9883807439824945], "iscrowd": 0, "area": 72368.72124999999, "rle": {"size": [457, 640], "counts": "me]41V>6J7H7J7H7J6J7H5L4L4K6K4N2N2N2M3N1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O001O1O1N2O1N2N2M3M3M3L4M3M3M3L4M3M3M3M3L301O4M4K6J5K3N1N1O101N1O2N101N1O2O0O2N1O2O0O00000000000O10000000000BRJaHn5_7SJ`Hm5`7SJ`Hm5_7UJ`Hk5`7UJ`Hj5`7XJ^Hi5b7WJ^Hh5c7YJ\\Hf5d7[J\\He5d7\\J[Hc5f7]JYHc5h7^JWHb5i7^JVHb5k7_JSHb5m7^JRHd5m7>N1O1UIdIh5\\6lIPJS6Q6aI[J^6h60M3N2N201N10000000000O101OO1O1O1O001O1O1O1O0N3N2M3M3M2O2M3M3M3N1N3M3M3N2M3M4M3O1O1O2M2O1O2N2N1N3N2N2N2N1N3N2N2N2N2M2O2N2N2N2M2O2N2N2N2N2N1O2N2M3N1O1O001O001O001O00001N101O001O001O001O00001O001N101O001O0010O100O1O100O2O0O100O100O1N3N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N1O2N1O2N2N2M4M2N3L3N2cJmJ`1V5\\NlJb1W5[NkJd1W5XNlJg1T5WNoJh1Q5TMfJhN:S4Q5TMhJeN:V4n4TMkJbN:Y4k4TMnJ_N:[4i4UMPK\\N:^4f4UMSKYN:a4c4TMWKWN8d4b4TMXKVN8e4^4VM]KQN8h4X4YMgLe2X3]MiLb2T3aMmL^2S3bMmL]2T3dMlL[2T3eMmLZ2S3fMnLX2S3hMnLW2S3hMnLW2R3jMnLT2S3lMnLR2S3nMnLP2S3PNnLm1T3SNlLl1U3UNkLh1W3XNjLe1Y3ZNhLd1Y3\\NhLa1Z3_NgL_1Z3bNfL[1\\3eNeLY1\\3gNeLV1]3jNcLT1_3lNbLR1_3oNaLn0b3POaLm0`3QOeLj0]3SOiLZNQMP2W6DmLUNSMQ2S6H^M1f2LaMKd23aMFc27dMAa2;eM]O`29nM_OV27WNBm13bNDW8I\\S4"}, "label": "motorcycle"}]}
{"id": 503500, "image": "COCO_train2014_000000503500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a byke was rod\"."}], "task": "refering", "answer_template": "The \"a byke was rod\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 85, 86, 87, 88, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 365, 366], "bbox": [0.49582812499999995, 0.17940919037199124, 0.9836406249999999, 0.9883807439824945], "iscrowd": 0, "area": 72368.72124999999, "rle": {"size": [457, 640], "counts": "me]41V>6J7H7J7H7J6J7H5L4L4K6K4N2N2N2M3N1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O001O1O1N2O1N2N2M3M3M3L4M3M3M3L4M3M3M3M3L301O4M4K6J5K3N1N1O101N1O2N101N1O2O0O2N1O2O0O00000000000O10000000000BRJaHn5_7SJ`Hm5`7SJ`Hm5_7UJ`Hk5`7UJ`Hj5`7XJ^Hi5b7WJ^Hh5c7YJ\\Hf5d7[J\\He5d7\\J[Hc5f7]JYHc5h7^JWHb5i7^JVHb5k7_JSHb5m7^JRHd5m7>N1O1UIdIh5\\6lIPJS6Q6aI[J^6h60M3N2N201N10000000000O101OO1O1O1O001O1O1O1O0N3N2M3M3M2O2M3M3M3N1N3M3M3N2M3M4M3O1O1O2M2O1O2N2N1N3N2N2N2N1N3N2N2N2N2M2O2N2N2N2M2O2N2N2N2N2N1O2N2M3N1O1O001O001O001O00001N101O001O001O001O00001O001N101O001O0010O100O1O100O2O0O100O100O1N3N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N1O2N1O2N2N2M4M2N3L3N2cJmJ`1V5\\NlJb1W5[NkJd1W5XNlJg1T5WNoJh1Q5TMfJhN:S4Q5TMhJeN:V4n4TMkJbN:Y4k4TMnJ_N:[4i4UMPK\\N:^4f4UMSKYN:a4c4TMWKWN8d4b4TMXKVN8e4^4VM]KQN8h4X4YMgLe2X3]MiLb2T3aMmL^2S3bMmL]2T3dMlL[2T3eMmLZ2S3fMnLX2S3hMnLW2S3hMnLW2R3jMnLT2S3lMnLR2S3nMnLP2S3PNnLm1T3SNlLl1U3UNkLh1W3XNjLe1Y3ZNhLd1Y3\\NhLa1Z3_NgL_1Z3bNfL[1\\3eNeLY1\\3gNeLV1]3jNcLT1_3lNbLR1_3oNaLn0b3POaLm0`3QOeLj0]3SOiLZNQMP2W6DmLUNSMQ2S6H^M1f2LaMKd23aMFc27dMAa2;eM]O`29nM_OV27WNBm13bNDW8I\\S4"}, "label": "a byke was rod"}]}
{"id": 200404, "image": "COCO_train2014_000000200404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a yellow shirt and orange cap\"."}], "task": "refering", "answer_template": "The \"a man wearing a yellow shirt and orange cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 225, 226, 248, 249, 271, 295, 296, 318, 319, 341, 342, 343, 364, 365, 366], "bbox": [0.80565625, 0.41289583333333335, 0.9413125000000001, 0.904625], "iscrowd": 0, "area": 6931.264099999996, "rle": {"size": [480, 640], "counts": "iSb7d0X><Do0SO9G5L4L4L2N3N1N3N1REYM`9j2\\FPNl8Q2QGhNe;k0IO0001O00001O00001O000010O0000SLnMlJS2m4SNSKl1m4UNSKi1n4XNRKg1n4ZNRKe1n4\\NRKc1n4^NRKa1n4`NRKa1l4`NUK`1j4aNUK`1i4aNWK`1g4aNYK`1e4aN[K`1d4`N\\Ka1b4`N^K`1a4aN_K_1a4aN_K^1b4bN^K^1b4bN^K^1a4cN`K\\1`4dN`K\\1`4dN`K\\1a4dN^K\\1c4YNZHSOW3d2_4XNhKh1Y4WNgKj1Y4TNhKl1Y4RNkKk1V4TNnKh1S4WNPLf1R4XNSLd1o3YNULc1n3YNhLR1\\3jNgLS1Z3kNkLQ1W3mNlLQ1T3lNQMQ1P3mNTMP1b8L2N1O2O0O1O1O2N1O1N2O2N1O1N2O1O2M4Md\\a0"}, "label": "a man wearing a yellow shirt and orange cap"}]}
{"id": 200404, "image": "COCO_train2014_000000200404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the yellow shirt and orange hat\"."}], "task": "refering", "answer_template": "The \"the man in the yellow shirt and orange hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 225, 226, 248, 249, 271, 295, 296, 318, 319, 341, 342, 343, 364, 365, 366], "bbox": [0.80565625, 0.41289583333333335, 0.9413125000000001, 0.904625], "iscrowd": 0, "area": 6931.264099999996, "rle": {"size": [480, 640], "counts": "iSb7d0X><Do0SO9G5L4L4L2N3N1N3N1REYM`9j2\\FPNl8Q2QGhNe;k0IO0001O00001O00001O000010O0000SLnMlJS2m4SNSKl1m4UNSKi1n4XNRKg1n4ZNRKe1n4\\NRKc1n4^NRKa1n4`NRKa1l4`NUK`1j4aNUK`1i4aNWK`1g4aNYK`1e4aN[K`1d4`N\\Ka1b4`N^K`1a4aN_K_1a4aN_K^1b4bN^K^1b4bN^K^1a4cN`K\\1`4dN`K\\1`4dN`K\\1a4dN^K\\1c4YNZHSOW3d2_4XNhKh1Y4WNgKj1Y4TNhKl1Y4RNkKk1V4TNnKh1S4WNPLf1R4XNSLd1o3YNULc1n3YNhLR1\\3jNgLS1Z3kNkLQ1W3mNlLQ1T3lNQMQ1P3mNTMP1b8L2N1O2O0O1O1O2N1O1N2O2N1O1N2O1O2M4Md\\a0"}, "label": "the man in the yellow shirt and orange hat"}]}
{"id": 200404, "image": "COCO_train2014_000000200404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow road emergency truck\"."}], "task": "refering", "answer_template": "The \"a yellow road emergency truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 284, 285], "bbox": [0.198875, 0.34568750000000004, 0.5528125, 0.7501875000000001], "iscrowd": 0, "area": 29095.590799999994, "rle": {"size": [480, 640], "counts": "Tjk17R>h0[Of0M2O2M2O2M2O1N3M2O2M2O1N2O1N101N2N2O0O2O1N2O0O2O1N2N101N2OON3M3M3M2N3M3M3M2N3M3M2O2M3M3M2N2N2N3M2N2N2N3N11O1O100O1001O001O0010O01O001O001O001O1O001O001O001O001O001O001O100O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O0010O01O001O00001O001O001O001O000000O10000000001O0O1000000000000O10000000000O2O3M2]OfF\\K]9a4lFVKW9f4d0N3M2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M00000000000O1000000000000000000000000000000000000000000O100]OnDXMR;d2g0L4M3L4L4M3L3M4L4M3N2N2N2M3N2N2N2N2N2N2N2N2M3N6J6J6J6JbjU4"}, "label": "a yellow road emergency truck"}]}
{"id": 200404, "image": "COCO_train2014_000000200404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow emergency truck\"."}], "task": "refering", "answer_template": "The \"the yellow emergency truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 284, 285], "bbox": [0.198875, 0.34568750000000004, 0.5528125, 0.7501875000000001], "iscrowd": 0, "area": 29095.590799999994, "rle": {"size": [480, 640], "counts": "Tjk17R>h0[Of0M2O2M2O2M2O1N3M2O2M2O1N2O1N101N2N2O0O2O1N2O0O2O1N2N101N2OON3M3M3M2N3M3M3M2N3M3M2O2M3M3M2N2N2N3M2N2N2N3N11O1O100O1001O001O0010O01O001O001O001O1O001O001O001O001O001O001O100O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O0010O01O001O00001O001O001O001O000000O10000000001O0O1000000000000O10000000000O2O3M2]OfF\\K]9a4lFVKW9f4d0N3M2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M00000000000O1000000000000000000000000000000000000000000O100]OnDXMR;d2g0L4M3L4L4M3L3M4L4M3N2N2N2M3N2N2N2N2N2N2N2N2M3N6J6J6J6JbjU4"}, "label": "the yellow emergency truck"}]}
{"id": 200404, "image": "COCO_train2014_000000200404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white van stopping on the side of road\"."}], "task": "refering", "answer_template": "The \"the white van stopping on the side of road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 320, 321], "bbox": [0.8219218749999999, 0.31224999999999997, 1.0, 0.8205208333333334], "iscrowd": 0, "area": 21077.926850000003, "rle": {"size": [480, 640], "counts": "dkf7P1n;S2C<J6J6I7J6J6J6I7J6J6J7M20000000000O100000001O000000000000000000001O0000000O1000000000001O000000000000000000001O000O10000000000000001O0000000000000XG\\JT8d5lGaJo7_5PHfJl7[5SHhJj7X5UHkJh7V5XHlJf7T5YHoJe7Q5[HQKc7o4WHYKg7R6N1O00001O0;E?A?A3MM3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M4L3N2M3V1jNX4hKjE"}, "label": "the white van stopping on the side of road"}]}
{"id": 200404, "image": "COCO_train2014_000000200404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white van to the right of the people\"."}], "task": "refering", "answer_template": "The \"the white van to the right of the people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 320, 321], "bbox": [0.8219218749999999, 0.31224999999999997, 1.0, 0.8205208333333334], "iscrowd": 0, "area": 21077.926850000003, "rle": {"size": [480, 640], "counts": "dkf7P1n;S2C<J6J6I7J6J6J6I7J6J6J7M20000000000O100000001O000000000000000000001O0000000O1000000000001O000000000000000000001O000O10000000000000001O0000000000000XG\\JT8d5lGaJo7_5PHfJl7[5SHhJj7X5UHkJh7V5XHlJf7T5YHoJe7Q5[HQKc7o4WHYKg7R6N1O00001O0;E?A?A3MM3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M4L3N2M3V1jNX4hKjE"}, "label": "the white van to the right of the people"}]}
{"id": 536278, "image": "COCO_train2014_000000536278.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teddy bear facing away\"."}], "task": "refering", "answer_template": "The \"the teddy bear facing away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 314, 315, 316], "bbox": [0.593265625, 0.307875, 0.817421875, 0.7910208333333333], "iscrowd": 0, "area": 21629.475049999997, "rle": {"size": [480, 640], "counts": "m]b5a0Z>c0]O:E7J5K6J6J5K6J6J5K6J6J5M4M3M2O2M3M2N3N2M2N3N1N2N3M2O1N3N1_NeK\\H\\4b7QLSHP4j7XLoGi3o7_LjGb3T8fLeG\\3X8g1N2N3M2N2M4M20001O0000001N10001O00001O00001O00001O0002O0N3N2N1O2M2O2N2N1N3N2N1O1N1000000O10000O1000000O1000000O1000000O100dK\\G\\2d8cM_G[2a8dMcGX2^8gMfGV2Z8iMiGU2V8kMnGR2R8mMRHo1o7oMUHo1k7PNYHm1h7QN[Hm1f7QN^Hk1c7TNaHi1`7UNcHi1]7VNgHg1Z7WNjHe1X7YNkH]1YN_Ml8S1oHZ1YN`Mi8S1RIZ1X7eNlHW1V7gNlHX1U7fNnHW1S7hNoHV1S7gNQIW1o6\\N`Ia1b6QNlIm1f9M2O2M3M2O2M2N3M2O2M3M2O2M2N3M2O4K4L5K5L4Kc`f1"}, "label": "the teddy bear facing away"}]}
{"id": 536278, "image": "COCO_train2014_000000536278.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear whose face cannot be seen\"."}], "task": "refering", "answer_template": "The \"teddy bear whose face cannot be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 314, 315, 316], "bbox": [0.593265625, 0.307875, 0.817421875, 0.7910208333333333], "iscrowd": 0, "area": 21629.475049999997, "rle": {"size": [480, 640], "counts": "m]b5a0Z>c0]O:E7J5K6J6J5K6J6J5K6J6J5M4M3M2O2M3M2N3N2M2N3N1N2N3M2O1N3N1_NeK\\H\\4b7QLSHP4j7XLoGi3o7_LjGb3T8fLeG\\3X8g1N2N3M2N2M4M20001O0000001N10001O00001O00001O00001O0002O0N3N2N1O2M2O2N2N1N3N2N1O1N1000000O10000O1000000O1000000O1000000O100dK\\G\\2d8cM_G[2a8dMcGX2^8gMfGV2Z8iMiGU2V8kMnGR2R8mMRHo1o7oMUHo1k7PNYHm1h7QN[Hm1f7QN^Hk1c7TNaHi1`7UNcHi1]7VNgHg1Z7WNjHe1X7YNkH]1YN_Ml8S1oHZ1YN`Mi8S1RIZ1X7eNlHW1V7gNlHX1U7fNnHW1S7hNoHV1S7gNQIW1o6\\N`Ia1b6QNlIm1f9M2O2M3M2O2M2N3M2O2M3M2O2M2N3M2O4K4L5K5L4Kc`f1"}, "label": "teddy bear whose face cannot be seen"}]}
{"id": 511706, "image": "COCO_train2014_000000511706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow construction vehicle\"."}], "task": "refering", "answer_template": "The \"the yellow construction vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.0074062500000000005, 0.47408333333333336, 0.401859375, 0.6962916666666668], "iscrowd": 0, "area": 13878.60475, "rle": {"size": [480, 640], "counts": "bd2b0]>4L1O1O1O2J5N2O1O100XOSO\\Cn0c<YOWCg0i<h00000O100000O010O100O01000O10O0100O1O00100O1O10O2O2N2N2M2O1O1O1O1N1N2N2O001O001K50fMdCo1h<O000OWOoMfDP2Z;VN`Di1a;\\N[Dc1d;aNYD_1g;eNVDY1k;jNRDV1n;mNPDQ1P<P10O100O100O01000O100O100001O1O001O1O1O001O1O000000000000000000O1O1O1O1O100O1O1O1SOjCoNW<Q1mCjNU<T1PDbNV<^1nCYNW<g1a000O10001N10]C^Nh;b1XDdNb;\\1]DhN`;X1_DjN`;>mCHb0Ka;9RDL;Ld;2WD14Ne;L]D6LOg;FdD9D3g;^OlD?[O4o<KPC6Q=MjB4S=2jBOR=P1N3LnNSC9i<TOWC=2<e<TO`C>L>T=_OmBa0T=]OmBc0T=\\OlBc0V=[OkBe0W=XOkBh0X=SOiBm0Z=POfB`0GOm=1TBNl=2TBOk=1VBNj=2WBMi=3XBLh=4YBIi=7XBFk=:UBCm==TB@n=`0900000000000001O0000000000001O00000001O00010O0N2O1N3M2N200O3M4L4M2M3M2N3N1N3M2N3N0O1000000001O00000000001O0001O000001O0O100O101N101N101N100O2O0O2O01O1N2O1O1O1N2O1O1O00O010O100O5L6I3N1N2O1N2N2O1N2N2K5J6J7JeWc5"}, "label": "the yellow construction vehicle"}]}
{"id": 511706, "image": "COCO_train2014_000000511706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large yellow machine\"."}], "task": "refering", "answer_template": "The \"the large yellow machine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.0074062500000000005, 0.47408333333333336, 0.401859375, 0.6962916666666668], "iscrowd": 0, "area": 13878.60475, "rle": {"size": [480, 640], "counts": "bd2b0]>4L1O1O1O2J5N2O1O100XOSO\\Cn0c<YOWCg0i<h00000O100000O010O100O01000O10O0100O1O00100O1O10O2O2N2N2M2O1O1O1O1N1N2N2O001O001K50fMdCo1h<O000OWOoMfDP2Z;VN`Di1a;\\N[Dc1d;aNYD_1g;eNVDY1k;jNRDV1n;mNPDQ1P<P10O100O100O01000O100O100001O1O001O1O1O001O1O000000000000000000O1O1O1O1O100O1O1O1SOjCoNW<Q1mCjNU<T1PDbNV<^1nCYNW<g1a000O10001N10]C^Nh;b1XDdNb;\\1]DhN`;X1_DjN`;>mCHb0Ka;9RDL;Ld;2WD14Ne;L]D6LOg;FdD9D3g;^OlD?[O4o<KPC6Q=MjB4S=2jBOR=P1N3LnNSC9i<TOWC=2<e<TO`C>L>T=_OmBa0T=]OmBc0T=\\OlBc0V=[OkBe0W=XOkBh0X=SOiBm0Z=POfB`0GOm=1TBNl=2TBOk=1VBNj=2WBMi=3XBLh=4YBIi=7XBFk=:UBCm==TB@n=`0900000000000001O0000000000001O00000001O00010O0N2O1N3M2N200O3M4L4M2M3M2N3N1N3M2N3N0O1000000001O00000000001O0001O000001O0O100O101N101N101N100O2O0O2O01O1N2O1O1O1N2O1O1O00O010O100O5L6I3N1N2O1N2N2O1N2N2K5J6J7JeWc5"}, "label": "the large yellow machine"}]}
{"id": 274139, "image": "COCO_train2014_000000274139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog resting on the ground holding a stick\"."}], "task": "refering", "answer_template": "The \"a dog resting on the ground holding a stick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 305, 306, 307, 320, 321, 322, 336, 337], "bbox": [0.026188235294117648, 0.133078125, 0.9897176470588235, 0.9959374999999999], "iscrowd": 0, "area": 112908.93580000002, "rle": {"size": [640, 425], "counts": "jd72nc01O001O1O1O1O1O001O1O1O001O100O1O001O1O1O1O001O100O1O001O100O0010O01O10O00001O001O01O01O00001O00001Om\\O^Ohb0b0T]OEib0k0O0O0001O00O10000000001N100000000O100\\_OXOP>h0mA^OP>b0mACQ>=lAHQ>9lAMQ>3mA1Q>OnA5o=KoA9o=GPB=m=CQBc0nNWN\\>V1eBj0hNRNb>T1fBW1Y=iNgBX1X=hNgBY1Y=gNgBZ1W=gNiBZ1V=fNiB\\1V=dNiB^1V=bNhBa1W=_NgBd1X=\\NfAjNf0l2d=YNfAnNb0l2g=VNfAQO?l2j=RNgAUO;l2m=oMgAXO8k2Q>lMgA\\O4k2T>iMgA^O3k2T>gMiA@1k2U>eMiAA0n2U>`MkADNo2U>]MlA_3S>`LmAa3S>_LlAc3S>U10WKmAf3R>T1000000000000000000000000O01O10O01O0010O01OlGWJR2i5nMYJP2h5oMZJo1f5RN[Jm1d5SN^Jk1b5UN`Ji1`5XNaJf1_5ZNcJd1\\5]NfJa1X5bNiJ\\1T5gNnJW1n4mNTKQ1l4nNWKP1j4nNYKP1h4oNZKo0g4oN\\Ko0e4oN^Ko0c4PO^Ko0c4oN_KP1a4oNaKP1`4nNbKQ1_4mNcKR1]4mNeKR1\\4lNfKS1[4lN`L9`3HfL2Z3MnLKS35_MXOa2h0bMUOc2f0`MWOe2d0^MZOe2b0^M[Og2`0\\M]Oh2?[M^Oj2=YM@k2;XMDl27WMFm25VMIn23TMKQ3OSMNQ3NQM1R3JQM4S3GPM7T3CQM:R3@SM?P3[OTMc0P3lN^MS1f2iN[MV1h2gNYMX1k2eNUM[1n2bNRM]1R3aNmL^1P4gMoKX2X4cMgK[2a4`M^K_2h4\\MXKc2k4\\MTKd2n4ZMRKf2o4ZMPKe2R5ZMnJf2R5ZMnJe2T5[MkJe2U5[MkJd2W5[MiJe2X5[MgJd2Z5\\MfJd2[5[MeJe2[5\\MdJc2^5\\MbJd2_5[M`Je2a5[M]Jg2d5YMZJg2g5YMWJi2j5VMUJj2m5UMQJm2o5TMoIm2R6RMoGDhN[3[9QMeGNlNR3_9PMeGc0UN_2W:nLbGi0QNY2^:nLaGk0mMY2c:lL^Gm0mMX2e:kL^Go0iMX2i:jL]GS5c8mJ\\GS5f8mJYGS5g8lJYGU5g8jJYGW5g8hJZGX5g8fJYGZ5h8eJXG\\5h8dJXG\\5h8dJWG]5i8cJWG]5j8bJUG^5l8bJSG_5m8aJSG_5m8aJRG`5o8_JPGb5P9]JPGc5Q9]JoFc5Q9]JnFd5R9\\JmFe5T9ZJkFg5U9YJkFf5V9ZJiFg5W9YJhFh5Y9WJfFj5Z9UJgFk5Y9UJfFk5[9UJdFl5\\9TJcFm5^9RJaFo5_9QJaFo5_9QJ`FP6`9oI`FQ6a9oI^FR6c9mI]FS6c9lI]FU6c9kI\\FV6e9iIZFX6g;O01O1O100O1O00100O1O100O00100O1O100O010O1O2O0O1O2O0O2N101N1O101N101N100O2O0O2N101N100O2O0O2O0O100O2O0O100O101N100O100O101N100O100O101N10000O101N101N101N2O0O2O0O2O0O2N101N2N101N1O2O0O2N101N1O101N1O2N1O1O2N101N1O1O2N1O2N2O1N2N2N2N2N2N2N2O1N4L5K6J:F8H2N2O0N3N2N2mLiDYMZ;ZMhDo43bMc;[2cD_M`;_2eD\\M\\;c2iDXMY;e2mDUMV;i2oDRMS;l2REnLQ;o2VEkLl:S3WEjLl:R3WEkLm:Q3WEjLn:R3VEjLn:R3VEjLn:R3VEjLn:Q3WEkLn:P3`EbLd:Z3T3L6I6K5K5I5H9dNb^OKfa02a^OBha0;Q1M3M4K5L4LhX2"}, "label": "a dog resting on the ground holding a stick"}]}
{"id": 274139, "image": "COCO_train2014_000000274139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown color dog stick in the hand\"."}], "task": "refering", "answer_template": "The \"a brown color dog stick in the hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 305, 306, 307, 320, 321, 322, 336, 337], "bbox": [0.026188235294117648, 0.133078125, 0.9897176470588235, 0.9959374999999999], "iscrowd": 0, "area": 112908.93580000002, "rle": {"size": [640, 425], "counts": "jd72nc01O001O1O1O1O1O001O1O1O001O100O1O001O1O1O1O001O100O1O001O100O0010O01O10O00001O001O01O01O00001O00001Om\\O^Ohb0b0T]OEib0k0O0O0001O00O10000000001N100000000O100\\_OXOP>h0mA^OP>b0mACQ>=lAHQ>9lAMQ>3mA1Q>OnA5o=KoA9o=GPB=m=CQBc0nNWN\\>V1eBj0hNRNb>T1fBW1Y=iNgBX1X=hNgBY1Y=gNgBZ1W=gNiBZ1V=fNiB\\1V=dNiB^1V=bNhBa1W=_NgBd1X=\\NfAjNf0l2d=YNfAnNb0l2g=VNfAQO?l2j=RNgAUO;l2m=oMgAXO8k2Q>lMgA\\O4k2T>iMgA^O3k2T>gMiA@1k2U>eMiAA0n2U>`MkADNo2U>]MlA_3S>`LmAa3S>_LlAc3S>U10WKmAf3R>T1000000000000000000000000O01O10O01O0010O01OlGWJR2i5nMYJP2h5oMZJo1f5RN[Jm1d5SN^Jk1b5UN`Ji1`5XNaJf1_5ZNcJd1\\5]NfJa1X5bNiJ\\1T5gNnJW1n4mNTKQ1l4nNWKP1j4nNYKP1h4oNZKo0g4oN\\Ko0e4oN^Ko0c4PO^Ko0c4oN_KP1a4oNaKP1`4nNbKQ1_4mNcKR1]4mNeKR1\\4lNfKS1[4lN`L9`3HfL2Z3MnLKS35_MXOa2h0bMUOc2f0`MWOe2d0^MZOe2b0^M[Og2`0\\M]Oh2?[M^Oj2=YM@k2;XMDl27WMFm25VMIn23TMKQ3OSMNQ3NQM1R3JQM4S3GPM7T3CQM:R3@SM?P3[OTMc0P3lN^MS1f2iN[MV1h2gNYMX1k2eNUM[1n2bNRM]1R3aNmL^1P4gMoKX2X4cMgK[2a4`M^K_2h4\\MXKc2k4\\MTKd2n4ZMRKf2o4ZMPKe2R5ZMnJf2R5ZMnJe2T5[MkJe2U5[MkJd2W5[MiJe2X5[MgJd2Z5\\MfJd2[5[MeJe2[5\\MdJc2^5\\MbJd2_5[M`Je2a5[M]Jg2d5YMZJg2g5YMWJi2j5VMUJj2m5UMQJm2o5TMoIm2R6RMoGDhN[3[9QMeGNlNR3_9PMeGc0UN_2W:nLbGi0QNY2^:nLaGk0mMY2c:lL^Gm0mMX2e:kL^Go0iMX2i:jL]GS5c8mJ\\GS5f8mJYGS5g8lJYGU5g8jJYGW5g8hJZGX5g8fJYGZ5h8eJXG\\5h8dJXG\\5h8dJWG]5i8cJWG]5j8bJUG^5l8bJSG_5m8aJSG_5m8aJRG`5o8_JPGb5P9]JPGc5Q9]JoFc5Q9]JnFd5R9\\JmFe5T9ZJkFg5U9YJkFf5V9ZJiFg5W9YJhFh5Y9WJfFj5Z9UJgFk5Y9UJfFk5[9UJdFl5\\9TJcFm5^9RJaFo5_9QJaFo5_9QJ`FP6`9oI`FQ6a9oI^FR6c9mI]FS6c9lI]FU6c9kI\\FV6e9iIZFX6g;O01O1O100O1O00100O1O100O00100O1O100O010O1O2O0O1O2O0O2N101N1O101N101N100O2O0O2N101N100O2O0O2O0O100O2O0O100O101N100O100O101N100O100O101N10000O101N101N101N2O0O2O0O2O0O2N101N2N101N1O2O0O2N101N1O101N1O2N1O1O2N101N1O1O2N1O2N2O1N2N2N2N2N2N2N2O1N4L5K6J:F8H2N2O0N3N2N2mLiDYMZ;ZMhDo43bMc;[2cD_M`;_2eD\\M\\;c2iDXMY;e2mDUMV;i2oDRMS;l2REnLQ;o2VEkLl:S3WEjLl:R3WEkLm:Q3WEjLn:R3VEjLn:R3VEjLn:R3VEjLn:Q3WEkLn:P3`EbLd:Z3T3L6I6K5K5I5H9dNb^OKfa02a^OBha0;Q1M3M4K5L4LhX2"}, "label": "a brown color dog stick in the hand"}]}
{"id": 274139, "image": "COCO_train2014_000000274139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog\"."}], "task": "refering", "answer_template": "The \"a brown dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 275, 276, 277, 278, 279, 280, 281, 290, 291, 292, 293, 294, 305, 306, 307, 320, 321, 322, 323, 335, 336, 337], "bbox": [0.0, 0.130640625, 1.0, 0.9995937500000001], "iscrowd": 0, "area": 122763.72804999999, "rle": {"size": [640, 425], "counts": "i43mc01O0O2O1O1O0O2O1O001O1N2O001O0O2O001O001N101O1O0O2O001O001O10O01O010O1O010O00100O001O10O01O010O1O0100O0100O10O10O10O01000O0100O010O1a_OPOm=o0oAVOQ>j0jA[OU>e0gA@Y>?dAE[>;aAJ^>6_AN`>2`AO_>2_A0_>1aA2\\>NcA6Z>JfA9W>GhA=U>CkA`0R>AmAb0P>^OoAf0m=[OSBh0j=XOUBj0j=VOVBl0h=TOWBn0h=SOWBo0g=QOYBP1f=POYBS1d=nN\\BS1c=mN\\BV1b=jN^BX1`=iN_BX1`=hN_B[1_=eNaB\\1^=dNaB_1]=aNcB`1[=aNdBb1Z=_NeBb1Z=^NeBe1Y=\\NeBi1W=WNhBm1U=TNiBP2T=QNjBT2R=lMlBY2P=iMnBY2Q=gMnBZ2R=gMlBZ2T=fMkB[2U=fMiB[2W=iMcBY2\\=lM_BU2a=oMZBR2f=RNUBo1j=T2000000O10000O10000O1000000O0100O010O10O010O001O001O00O1O1OhHkIh0T6WOPJg0o5WOUJh0j5WOZJg0e5WO_Jh0`5WOdJg0[5XOhJg0W5YOkJf0T5YOPKe0P5ZORKe0m4ZOWKd0i4[OYKd0g4ZO`Ka0`4^OfK=[4BkK8V4FPL5Q4IVL1k3N[LLf32`LIa35fLE]38iLB[39lLAX3:nLAV3;PM_OT3<SM^OQ3=VM]On2>YM\\Ok2?[M\\Oi2?YM@k2;WMDm27VMGn24TMKP3OSM0Q3ISM6Q3CSM<Q3]OTMa0P3ZORMe0R3VOPMi0T3QOoLn0U3mNmLR1X3hNjLW1Z3dNiL[1Z3_NiL`1\\3ZNfLf1[4WMgKh2]4RMfKn2^4lLeKR3_4iLcKW3c:O100O2O0O100O101N100O100O101N100O100O2O0O100O101N10O001O010O0cMZMaCg2_<^M[Cb2e<dMUC]2k<hMoBY2P=lMkBU2U=PNmAi2R>c100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O1O1O2N1N2O1O1O2N1O1N2O1O2N1O1O1O101O0O1000001N1000000O2O00000O101O000O10001O0O1000001N1000001N1000001O0O2O001O0O2O1O001N101O001N2N1O2O0O2N2N1O2O0O2N1O2O1N1O2N101N1O2N2O0O2M2O2M2O2M3N1N3M3N2M3N2M3N2M3N2M3M3N2M4M2jLlE\\LW:a3PFYLR:d3TFVLo9g3WFTLk9i3\\FPLg9m3_FmKd9P4bFcKi9Y4^FYKm9c4YFPKQ:m4UFfJV:V5QF\\JZ:a5U2L4L4L4L4K5J6J5L5J6J7J5J6J6K6H7I7I7J6I8H7ITJ"}, "label": "a brown dog"}]}
{"id": 274139, "image": "COCO_train2014_000000274139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a golden retriever laying down holding a stick\"."}], "task": "refering", "answer_template": "The \"a golden retriever laying down holding a stick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 275, 276, 277, 278, 279, 280, 281, 290, 291, 292, 293, 294, 305, 306, 307, 320, 321, 322, 323, 335, 336, 337], "bbox": [0.0, 0.130640625, 1.0, 0.9995937500000001], "iscrowd": 0, "area": 122763.72804999999, "rle": {"size": [640, 425], "counts": "i43mc01O0O2O1O1O0O2O1O001O1N2O001O0O2O001O001N101O1O0O2O001O001O10O01O010O1O010O00100O001O10O01O010O1O0100O0100O10O10O10O01000O0100O010O1a_OPOm=o0oAVOQ>j0jA[OU>e0gA@Y>?dAE[>;aAJ^>6_AN`>2`AO_>2_A0_>1aA2\\>NcA6Z>JfA9W>GhA=U>CkA`0R>AmAb0P>^OoAf0m=[OSBh0j=XOUBj0j=VOVBl0h=TOWBn0h=SOWBo0g=QOYBP1f=POYBS1d=nN\\BS1c=mN\\BV1b=jN^BX1`=iN_BX1`=hN_B[1_=eNaB\\1^=dNaB_1]=aNcB`1[=aNdBb1Z=_NeBb1Z=^NeBe1Y=\\NeBi1W=WNhBm1U=TNiBP2T=QNjBT2R=lMlBY2P=iMnBY2Q=gMnBZ2R=gMlBZ2T=fMkB[2U=fMiB[2W=iMcBY2\\=lM_BU2a=oMZBR2f=RNUBo1j=T2000000O10000O10000O1000000O0100O010O10O010O001O001O00O1O1OhHkIh0T6WOPJg0o5WOUJh0j5WOZJg0e5WO_Jh0`5WOdJg0[5XOhJg0W5YOkJf0T5YOPKe0P5ZORKe0m4ZOWKd0i4[OYKd0g4ZO`Ka0`4^OfK=[4BkK8V4FPL5Q4IVL1k3N[LLf32`LIa35fLE]38iLB[39lLAX3:nLAV3;PM_OT3<SM^OQ3=VM]On2>YM\\Ok2?[M\\Oi2?YM@k2;WMDm27VMGn24TMKP3OSM0Q3ISM6Q3CSM<Q3]OTMa0P3ZORMe0R3VOPMi0T3QOoLn0U3mNmLR1X3hNjLW1Z3dNiL[1Z3_NiL`1\\3ZNfLf1[4WMgKh2]4RMfKn2^4lLeKR3_4iLcKW3c:O100O2O0O100O101N100O100O101N100O100O2O0O100O101N10O001O010O0cMZMaCg2_<^M[Cb2e<dMUC]2k<hMoBY2P=lMkBU2U=PNmAi2R>c100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O1O1O2N1N2O1O1O2N1O1N2O1O2N1O1O1O101O0O1000001N1000000O2O00000O101O000O10001O0O1000001N1000001N1000001O0O2O001O0O2O1O001N101O001N2N1O2O0O2N2N1O2O0O2N1O2O1N1O2N101N1O2N2O0O2M2O2M2O2M3N1N3M3N2M3N2M3N2M3N2M3M3N2M4M2jLlE\\LW:a3PFYLR:d3TFVLo9g3WFTLk9i3\\FPLg9m3_FmKd9P4bFcKi9Y4^FYKm9c4YFPKQ:m4UFfJV:V5QF\\JZ:a5U2L4L4L4L4K5J6J5L5J6J7J5J6J6K6H7I7I7J6I8H7ITJ"}, "label": "a golden retriever laying down holding a stick"}]}
{"id": 421596, "image": "COCO_train2014_000000421596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is bending over to eat the grass\"."}], "task": "refering", "answer_template": "The \"a giraffe that is bending over to eat the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 313, 314, 315, 316, 333, 334, 336, 337], "bbox": [0.467265625, 0.4463958333333334, 0.7974062500000001, 0.8843125000000001], "iscrowd": 0, "area": 23918.34669999999, "rle": {"size": [480, 640], "counts": "h]\\46d>:F9G6J604K5K5L5J5K5L4K6J5K5L4]D_M^:m3D;F:E8H8H7J1N0001O000QOo0]Od0^Oa0N3M2M4M3L4M2N3;E=B:G8H8G:G8H8G:G8H1N10000O1000O0M4L4M2M4L4L3M4L4L4L3M4L4L3F;YOg00O10000O1000O010000O10000O10O10O103L5L3M4K4M4K4M4L3L5L3L5L3M4K5M3L4M3L4M3M3L4M3M0O10WMUH\\Nk7Y1eHbNZ7S1UIjNk6S1PHfM7T1h7W1jGQN=d0i7[1bG]Nd05j7^1[GiNi0El7Z3RH`LQ8e3mGVLU8o3iGkKY8Z4fGaK\\8_4gG[K[8e4iGVKY8j4o00O01J6I6I8H8H7I8H8J68H7I7I8H7I2NO1O1N2O001O1O1O1O1N2O1gMkEEV:<WFUOj9k0fFdN[9\\1U21O001O00001O001O00001SCYN^<g1`C\\N_<e1]C_Nb<a1[CcNd<m101O1001O001O0O101O001O0O1O2O0O2N100OO1O2M2O1N2N100O100O11UO]CWOc<a0jCZOV<e0PDUOQ<k0SDPOn;>\\CHl0Di;b0`CGb=7bBE_=9eBE[=:iBBX=<g0O1N^dl1"}, "label": "a giraffe that is bending over to eat the grass"}]}
{"id": 421596, "image": "COCO_train2014_000000421596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe bending down\"."}], "task": "refering", "answer_template": "The \"giraffe bending down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 313, 314, 315, 316, 333, 334, 336, 337], "bbox": [0.467265625, 0.4463958333333334, 0.7974062500000001, 0.8843125000000001], "iscrowd": 0, "area": 23918.34669999999, "rle": {"size": [480, 640], "counts": "h]\\46d>:F9G6J604K5K5L5J5K5L4K6J5K5L4]D_M^:m3D;F:E8H8H7J1N0001O000QOo0]Od0^Oa0N3M2M4M3L4M2N3;E=B:G8H8G:G8H8G:G8H1N10000O1000O0M4L4M2M4L4L3M4L4L4L3M4L4L3F;YOg00O10000O1000O010000O10000O10O10O103L5L3M4K4M4K4M4L3L5L3L5L3M4K5M3L4M3L4M3M3L4M3M0O10WMUH\\Nk7Y1eHbNZ7S1UIjNk6S1PHfM7T1h7W1jGQN=d0i7[1bG]Nd05j7^1[GiNi0El7Z3RH`LQ8e3mGVLU8o3iGkKY8Z4fGaK\\8_4gG[K[8e4iGVKY8j4o00O01J6I6I8H8H7I8H8J68H7I7I8H7I2NO1O1N2O001O1O1O1O1N2O1gMkEEV:<WFUOj9k0fFdN[9\\1U21O001O00001O001O00001SCYN^<g1`C\\N_<e1]C_Nb<a1[CcNd<m101O1001O001O0O101O001O0O1O2O0O2N100OO1O2M2O1N2N100O100O11UO]CWOc<a0jCZOV<e0PDUOQ<k0SDPOn;>\\CHl0Di;b0`CGb=7bBE_=9eBE[=:iBBX=<g0O1N^dl1"}, "label": "giraffe bending down"}]}
{"id": 421596, "image": "COCO_train2014_000000421596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest giraffe who ' s head is upright\"."}], "task": "refering", "answer_template": "The \"the tallest giraffe who ' s head is upright\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 81, 101, 102, 103, 104, 105, 106, 127, 128, 129, 130, 131, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 247, 248, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319], "bbox": [0.387578125, 0.1178125, 0.9161093749999999, 0.8180416666666667], "iscrowd": 0, "area": 28916.922350000004, "rle": {"size": [480, 640], "counts": "i[d32i>6K4O1O2M2O2N1O1O2N1O2N1O1O1OO10000O10IVOYBi0h=\\OSBd0m=701O1O1O1N2O1O1N2O1N2M3N2M3N2O1N1011N2O2M3`NeBW1_=3N1N3M2OjNhBh0X=TOlBk0V=oNoBQ1`=00O11O001O00100O100O10000O01000O10000O10000O10000O10000O100O10000O01000O10000O100O10000O01000O100O10000O1000O010000O100O10000O1000O0100O10000O1000O010000O100O1000O010000O100O1000O010000O100O1O010O1O101N100O1O2O0^CkMW<V2dCoM[<[210O1O10O0100O1O010O10O0100O00100O10O01O10O01O10O01O1O10O01O100O1O1O100O1O100O1O1O100O00100O1O2N100O1O100O1000000O10000000001O001O001O1N2O1O1OO1O1O1O1O001O1O001O1O001O1N101O1O001O1O001N2O0O10001N100O10000O100O2OUGTMb5l2\\JVMd5i2[JYMe5g2YJ[Mg5d2XJ^Mh5b2VJ`Mj5`2TJbMl5c2mI_MT6g2cI[M]6k2ZIXMf6o2QISMo6T3hHnLX7X3`HjL`7Z51N2O2M2N2N2O1N2N2N2O1N2N3N1N2N2N2O1N2N2N2O1N3M201O1O001O1n0QO7J5K4LO010O10OA`0J6I6K6eMXFXOm9b0\\FUO1cNo8R2YGROJUOi8c1gGkNDBi8]1PHdNYO0j8Y1UH_NPO8o8V1VH\\NjN`0T9P1mHPOW7m0hHTO[7i0dHWO`7f0mF@BLe9`0]Fb0[OoN[:<nEk1V:[15K6I6K5XNmE\\N[:S1T2_Oa0_OTSi0"}, "label": "the tallest giraffe who ' s head is upright"}]}
{"id": 421596, "image": "COCO_train2014_000000421596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest giraffe\"."}], "task": "refering", "answer_template": "The \"the tallest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 81, 101, 102, 103, 104, 105, 106, 127, 128, 129, 130, 131, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 247, 248, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319], "bbox": [0.387578125, 0.1178125, 0.9161093749999999, 0.8180416666666667], "iscrowd": 0, "area": 28916.922350000004, "rle": {"size": [480, 640], "counts": "i[d32i>6K4O1O2M2O2N1O1O2N1O2N1O1O1OO10000O10IVOYBi0h=\\OSBd0m=701O1O1O1N2O1O1N2O1N2M3N2M3N2O1N1011N2O2M3`NeBW1_=3N1N3M2OjNhBh0X=TOlBk0V=oNoBQ1`=00O11O001O00100O100O10000O01000O10000O10000O10000O10000O100O10000O01000O10000O100O10000O01000O100O10000O1000O010000O100O10000O1000O0100O10000O1000O010000O100O1000O010000O100O1000O010000O100O1O010O1O101N100O1O2O0^CkMW<V2dCoM[<[210O1O10O0100O1O010O10O0100O00100O10O01O10O01O10O01O1O10O01O100O1O1O100O1O100O1O1O100O00100O1O2N100O1O100O1000000O10000000001O001O001O1N2O1O1OO1O1O1O1O001O1O001O1O001O1N101O1O001O1O001N2O0O10001N100O10000O100O2OUGTMb5l2\\JVMd5i2[JYMe5g2YJ[Mg5d2XJ^Mh5b2VJ`Mj5`2TJbMl5c2mI_MT6g2cI[M]6k2ZIXMf6o2QISMo6T3hHnLX7X3`HjL`7Z51N2O2M2N2N2O1N2N2N2O1N2N3N1N2N2N2O1N2N2N2O1N3M201O1O001O1n0QO7J5K4LO010O10OA`0J6I6K6eMXFXOm9b0\\FUO1cNo8R2YGROJUOi8c1gGkNDBi8]1PHdNYO0j8Y1UH_NPO8o8V1VH\\NjN`0T9P1mHPOW7m0hHTO[7i0dHWO`7f0mF@BLe9`0]Fb0[OoN[:<nEk1V:[15K6I6K5XNmE\\N[:S1T2_Oa0_OTSi0"}, "label": "the tallest giraffe"}]}
{"id": 102111, "image": "COCO_train2014_000000102111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with red sleeves and a red pony tail\"."}], "task": "refering", "answer_template": "The \"a woman with red sleeves and a red pony tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 25, 26, 27, 48, 49, 50, 71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 140, 141, 163, 164], "bbox": [0.07575, 0.0031875000000000002, 0.28145312499999997, 0.4746666666666667], "iscrowd": 0, "area": 12236.484050000001, "rle": {"size": [480, 640], "counts": "Pef01f>;E;E;H8H8H8H8K6J5J6K5K6J7H8I6J7H8K5J6J5L5J6J4M1N1O1OMWFRKi9n4WFRKh9o4XFQKh9Q4\\FeLLZOg9R4cF^LG_Oe9U4fFYLEBe9U4hFWLCDd9W4jFSLBFd9W4lFQLAHa9X4PGnK_OJ`9Z4RGjK^OL`9Z4TGhK\\ON_9\\4VGdK\\OO]9^4]GbKb8`4]G`Kb8a4^G_Ka8b4m02N1O2N1O2N1L4G:N1001O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1N2O1O1N2YLnE_2S:`MnE_2S:aMmE]2U:bMmE[2U:dMlE[2U:dMlEZ2V:fMkEX2V:hMjEV2X:iMiEU2Y:kMhE930Z:GcE6b0@R:9]E5W<KiC3Y<MhCO[<0gCM[<3hCG[<9hCBZ<=U10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10002Mmcg6"}, "label": "a woman with red sleeves and a red pony tail"}]}
{"id": 102111, "image": "COCO_train2014_000000102111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a red shirt and braid standing behind a vegetable sculpture\"."}], "task": "refering", "answer_template": "The \"a woman with a red shirt and braid standing behind a vegetable sculpture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 25, 26, 27, 48, 49, 50, 71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 140, 141, 163, 164], "bbox": [0.07575, 0.0031875000000000002, 0.28145312499999997, 0.4746666666666667], "iscrowd": 0, "area": 12236.484050000001, "rle": {"size": [480, 640], "counts": "Pef01f>;E;E;H8H8H8H8K6J5J6K5K6J7H8I6J7H8K5J6J5L5J6J4M1N1O1OMWFRKi9n4WFRKh9o4XFQKh9Q4\\FeLLZOg9R4cF^LG_Oe9U4fFYLEBe9U4hFWLCDd9W4jFSLBFd9W4lFQLAHa9X4PGnK_OJ`9Z4RGjK^OL`9Z4TGhK\\ON_9\\4VGdK\\OO]9^4]GbKb8`4]G`Kb8a4^G_Ka8b4m02N1O2N1O2N1L4G:N1001O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1N2O1O1N2YLnE_2S:`MnE_2S:aMmE]2U:bMmE[2U:dMlE[2U:dMlEZ2V:fMkEX2V:hMjEV2X:iMiEU2Y:kMhE930Z:GcE6b0@R:9]E5W<KiC3Y<MhCO[<0gCM[<3hCG[<9hCBZ<=U10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10002Mmcg6"}, "label": "a woman with a red shirt and braid standing behind a vegetable sculpture"}]}
{"id": 208612, "image": "COCO_train2014_000000208612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green chair between two other chairs\"."}], "task": "refering", "answer_template": "The \"a green chair between two other chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 280], "bbox": [0.00075, 0.18083333333333332, 0.39659374999999997, 0.719], "iscrowd": 0, "area": 36651.6595, "rle": {"size": [480, 640], "counts": "Yg08f>7J6I8H4M2M3M3N2N2N1O2N2N2mNeN]D^1\\;oN\\DS1^;ZOYDh0e;[1N2O1N2N2O0O2O0O2N100O2O0O1O2O0O2O0O1O2O0O101N1O2O0O101N100O2O0O101N101N100O2O0O101N101N100O2O0O101N101N100O2O0O101N101N100O2O0O101O01O010O01O01O010O00010O01L3D<E<F9N3M2N2M4M2N3M2N2M4M2M3N3L3N3L3N2N3L3N3M20001O0O10O100O1O010O100O10O0100O100M3L4L4M3L4M4K4L4M3L4N200O1O1O1O2dN`GVL`8h3dGUL]8i3gGTLZ8j3iGTLX8j3lGTLT8j3PHSLQ8k3RHSLo7k3UHRLl7l3WHRLk7k3YHRLh7n3ZHPLf7o3\\HoKe7Q4]HlKd7T4^HiKc7W4^HgKc7Y4_HdKb7[4aHbK`7^4a101O000O100000000O101O00000000000000001O00000000000000000O2N1O1O1N2O1O1O1O1N3N1O1O1O1N2O1O1O1O2M2O1O1O1N2O1O1O10001O000000000000000M4D;WOi0WOi0C=O1O1O2N2M3N3M2N2N2N2N2Nejd5"}, "label": "a green chair between two other chairs"}]}
{"id": 208612, "image": "COCO_train2014_000000208612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty green bench\"."}], "task": "refering", "answer_template": "The \"empty green bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 280], "bbox": [0.00075, 0.18083333333333332, 0.39659374999999997, 0.719], "iscrowd": 0, "area": 36651.6595, "rle": {"size": [480, 640], "counts": "Yg08f>7J6I8H4M2M3M3N2N2N1O2N2N2mNeN]D^1\\;oN\\DS1^;ZOYDh0e;[1N2O1N2N2O0O2O0O2N100O2O0O1O2O0O2O0O1O2O0O101N1O2O0O101N100O2O0O101N101N100O2O0O101N101N100O2O0O101N101N100O2O0O101N101N100O2O0O101O01O010O01O01O010O00010O01L3D<E<F9N3M2N2M4M2N3M2N2M4M2M3N3L3N3L3N2N3L3N3M20001O0O10O100O1O010O100O10O0100O100M3L4L4M3L4M4K4L4M3L4N200O1O1O1O2dN`GVL`8h3dGUL]8i3gGTLZ8j3iGTLX8j3lGTLT8j3PHSLQ8k3RHSLo7k3UHRLl7l3WHRLk7k3YHRLh7n3ZHPLf7o3\\HoKe7Q4]HlKd7T4^HiKc7W4^HgKc7Y4_HdKb7[4aHbK`7^4a101O000O100000000O101O00000000000000001O00000000000000000O2N1O1O1N2O1O1O1O1N3N1O1O1O1N2O1O1O1O2M2O1O1O1N2O1O1O10001O000000000000000M4D;WOi0WOi0C=O1O1O2N2M3N3M2N2N2N2N2Nejd5"}, "label": "empty green bench"}]}
{"id": 208612, "image": "COCO_train2014_000000208612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue metal chair with a woman sitting in it\"."}], "task": "refering", "answer_template": "The \"a blue metal chair with a woman sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 235, 236, 237, 238, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334], "bbox": [0.23932812499999997, 0.24045833333333333, 0.6960625, 0.8696666666666667], "iscrowd": 0, "area": 33427.15899999999, "rle": {"size": [480, 640], "counts": "fQX23a>=D<I7I7I7O1N2O1N2O1N2O1O1N101N101O0O2O1N101N101O0O2O0O2O001N110O1O010O010O010O0010O010O01O10O010O010O010O1O100O00100O100O1O010O0001N10001N100O2O000O2O010O000010O0001O00010O000010O0001O0001TOeCROZ<l0iCSOW<j0lCWOS<g0PDXOQ<d0SD[Om;b0VD_Oi;?ZD@f;>\\DBe;<\\DEc;:^DFb;9_DGa;8`DI_;6bDJ_;4bDL^;2dDO[;0fD0Z;NhD2Y;KiD6V;HlD8T;GmD9S;EoD<P;BRE>o:_OSEb0l:UO]Ek0\\<0O2O00001O00001N10001O0N2O2M2N3M2O1N3M2O2M2N2N3M2^OQNQDQ2P<QNkCR2U<oMgCS2[<mMbCU2^<61O100mM^Cf1`;PNaE6TOi1Z;QNdE2VOk1V;SNSFk1l9WNUFh1j9XNXFf1h9ZNXFf1h9[NVFg1i9YNWFg1h9ZNWFg1i9ZNUFg1k9YNUFg1k9YNTFh1l9YNSFg1m9YNRFh1m9ZNRFf1n9ZNQFg1o9ZNPFf1P:ZNnEh1R:YNkEi1T:XNkEi1U:XNhEj1X:VNfEl1Z:UNdEl1\\:dNSE]1m:eNoD]1P;fNmDZ1T;hNhDX1Z;V11O0000001O00000O2O0000001N1OZNfLiGX3Z8iLdGV3]8lLbGQ3`8PM_Go2c8RM[Gl2g8VMWGh2k8ZMTGd2n8]MoFb2S9_MkFa2W9`MfF_2\\9gM\\FZ2e9QNoEP2R:WNeEj1[:]100O10000O1lLjK_KV4b4lK[KT4f4nKVKS4k4oKRKQ4o4QL]J`4e5aKQJf4P6\\KfIj4\\6XKVIS5k6oJgH\\5[7S11]1cN0001O000000001O0000001O0O1O1O1O2N100O1O2N1O1O101N1O1O1O1O2O00000O2O000000001O0000001O0000001O000kNjH`JV7_2kHZO3UNS7[2oH_O0UNQ7X2SIAOVNo6T2VIDNRNR7U2TIGMSNo6Q2YIKJRNo6n1[INIRNn6j1^I28aM[6X2bI5l7EYH:i7@]H=e7^OcH4i7E_HKP8OWH@X8;oGUO`8e0Q3J6J;EbXk2"}, "label": "a blue metal chair with a woman sitting in it"}]}
{"id": 208612, "image": "COCO_train2014_000000208612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"metal seat with a woman sitting on it\"."}], "task": "refering", "answer_template": "The \"metal seat with a woman sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 235, 236, 237, 238, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334], "bbox": [0.23932812499999997, 0.24045833333333333, 0.6960625, 0.8696666666666667], "iscrowd": 0, "area": 33427.15899999999, "rle": {"size": [480, 640], "counts": "fQX23a>=D<I7I7I7O1N2O1N2O1N2O1O1N101N101O0O2O1N101N101O0O2O0O2O001N110O1O010O010O010O0010O010O01O10O010O010O010O1O100O00100O100O1O010O0001N10001N100O2O000O2O010O000010O0001O00010O000010O0001O0001TOeCROZ<l0iCSOW<j0lCWOS<g0PDXOQ<d0SD[Om;b0VD_Oi;?ZD@f;>\\DBe;<\\DEc;:^DFb;9_DGa;8`DI_;6bDJ_;4bDL^;2dDO[;0fD0Z;NhD2Y;KiD6V;HlD8T;GmD9S;EoD<P;BRE>o:_OSEb0l:UO]Ek0\\<0O2O00001O00001N10001O0N2O2M2N3M2O1N3M2O2M2N2N3M2^OQNQDQ2P<QNkCR2U<oMgCS2[<mMbCU2^<61O100mM^Cf1`;PNaE6TOi1Z;QNdE2VOk1V;SNSFk1l9WNUFh1j9XNXFf1h9ZNXFf1h9[NVFg1i9YNWFg1h9ZNWFg1i9ZNUFg1k9YNUFg1k9YNTFh1l9YNSFg1m9YNRFh1m9ZNRFf1n9ZNQFg1o9ZNPFf1P:ZNnEh1R:YNkEi1T:XNkEi1U:XNhEj1X:VNfEl1Z:UNdEl1\\:dNSE]1m:eNoD]1P;fNmDZ1T;hNhDX1Z;V11O0000001O00000O2O0000001N1OZNfLiGX3Z8iLdGV3]8lLbGQ3`8PM_Go2c8RM[Gl2g8VMWGh2k8ZMTGd2n8]MoFb2S9_MkFa2W9`MfF_2\\9gM\\FZ2e9QNoEP2R:WNeEj1[:]100O10000O1lLjK_KV4b4lK[KT4f4nKVKS4k4oKRKQ4o4QL]J`4e5aKQJf4P6\\KfIj4\\6XKVIS5k6oJgH\\5[7S11]1cN0001O000000001O0000001O0O1O1O1O2N100O1O2N1O1O101N1O1O1O1O2O00000O2O000000001O0000001O0000001O000kNjH`JV7_2kHZO3UNS7[2oH_O0UNQ7X2SIAOVNo6T2VIDNRNR7U2TIGMSNo6Q2YIKJRNo6n1[INIRNn6j1^I28aM[6X2bI5l7EYH:i7@]H=e7^OcH4i7E_HKP8OWH@X8;oGUO`8e0Q3J6J;EbXk2"}, "label": "metal seat with a woman sitting on it"}]}
{"id": 569062, "image": "COCO_train2014_000000569062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt and jeans skateboarding\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt and jeans skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 47, 61, 62, 63, 64, 80, 81, 82, 98, 99, 100, 116, 117, 118, 134, 135, 136], "bbox": [0.38626, 0.1648048048048048, 0.67062, 0.6632732732732732], "iscrowd": 0, "area": 7619.687199999998, "rle": {"size": [333, 500], "counts": "aUo12[:0O100O2N100O101N1O100O2O0O1O10O010O010O1O100O100O010O10000001N10001O0O10001O000O101kGBY6>fIIT67jI0P61oI5l5LQJ;j5ETJb0g5^OXJh0mNlNS6=nJ\\1Q5eNmJ]1R5cNlJ`1S5`NlJ`1T5bNiJ`1W5`NhJa1X5_NfJc1Z5]NeJd1Z5^NcJc1^5]NaJ3UOd0Z6\\O\\J2\\O`0W6FUJJH=m55SJ_O3:i5^1ZJ^Ng5a1\\J]Nd5b1`J[N`5e1cJXN]5g1fJWNZ5h1jJTNW5l1kJRNU5m1oJPNQ5o1RKoMn4Q2UKlMk4S2YKiMh4V2[KhMe4X2aKbM_4]2i1O14M4L4M3L4M2M4L4K5K5K5PJ[Ln4i3jJ[LW5i3_J\\La5X400000001O0000000000PN`JSO`5oN\\Jm0?OU5SO`Je0f03k4WOaJ?n05a4[OeJ7U1:U4^OiJ1\\1<k3BmJIc1`0`3GPKAj1c0V3KkMNW21oMHS27SNBo1>P4O1O1O10O0100O10O10O1000O10O1000O0100000O01000E:1000000K5Iace1"}, "label": "a man wearing a black shirt and jeans skateboarding"}]}
{"id": 569062, "image": "COCO_train2014_000000569062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man in black on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 47, 61, 62, 63, 64, 80, 81, 82, 98, 99, 100, 116, 117, 118, 134, 135, 136], "bbox": [0.38626, 0.1648048048048048, 0.67062, 0.6632732732732732], "iscrowd": 0, "area": 7619.687199999998, "rle": {"size": [333, 500], "counts": "aUo12[:0O100O2N100O101N1O100O2O0O1O10O010O010O1O100O100O010O10000001N10001O0O10001O000O101kGBY6>fIIT67jI0P61oI5l5LQJ;j5ETJb0g5^OXJh0mNlNS6=nJ\\1Q5eNmJ]1R5cNlJ`1S5`NlJ`1T5bNiJ`1W5`NhJa1X5_NfJc1Z5]NeJd1Z5^NcJc1^5]NaJ3UOd0Z6\\O\\J2\\O`0W6FUJJH=m55SJ_O3:i5^1ZJ^Ng5a1\\J]Nd5b1`J[N`5e1cJXN]5g1fJWNZ5h1jJTNW5l1kJRNU5m1oJPNQ5o1RKoMn4Q2UKlMk4S2YKiMh4V2[KhMe4X2aKbM_4]2i1O14M4L4M3L4M2M4L4K5K5K5PJ[Ln4i3jJ[LW5i3_J\\La5X400000001O0000000000PN`JSO`5oN\\Jm0?OU5SO`Je0f03k4WOaJ?n05a4[OeJ7U1:U4^OiJ1\\1<k3BmJIc1`0`3GPKAj1c0V3KkMNW21oMHS27SNBo1>P4O1O1O10O0100O10O10O1000O10O1000O0100000O01000E:1000000K5Iace1"}, "label": "a man in black on a skateboard"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a group of red and green apples\"."}], "task": "refering", "answer_template": "The \"a group of red and green apples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 86, 96, 97, 98, 101, 102, 103, 104, 105, 106, 107, 108, 109, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 284, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.069234375, 0.14116822429906542, 0.783625, 0.9882242990654206], "iscrowd": 0, "area": 105651.02304999999, "rle": {"size": [428, 640], "counts": "icb02X=a0_O5K5K4L5K5K5J6K6J5K6J5K5L3L3M4L3M4M3M2N3M2N2N2N2N2N2N2N2N101N2N2N2N2N2N2N2O1N2N1O2N2N2N2O1N2N2N2N2N2N1O2O1N2N2N2N2N2N2O1N2N1O2N2N2N2N2O0O2N1O1O100O1O2O0O100O100O2O0O1O100O101N1000000O101O000000000O100000000000O0100000000000O01000000000000O0100000000000000O101O000000000O10000O101N100O100O101N101N101N101N101N101N2O0O2O1N101N2O1N101N2O0O2O1N101N2O1N101N2O0O2O1N[IcK^4[4XK\\L]4c3_KbLa4\\3\\KiLd4U3ZKoLf4o2XKUMh4i2VKZMl4c2RKaMn4^2oJeMR5Y2kJkMV5S2gJRNY5l1eJWN\\5g1aJ]N`5a1]JdNc5[1YJiNh5U1XJmNh5Q1XJPOi5n0VJTOk5j0UJXOk5f0UJ[Ol5d0SJ^On5?QJCP6;PJFQ68oIJQ64oIMR61mI1T6NkI4U6JkI7V6GjI:a5^LoJo6m4XIRKg6i4`IUK_6g4hIWKW6d4QJZKn5b4YJ]Kf5_4`J_K_5^4gJ`KX5]4nJaKR5\\4RKdKl4Y4ZKeKe4X4Y2N3L3O1N2N3M2N2N2N2O1N2O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O100O100O100O100O1O100O100O1000000000000O100000000000O10O1000000000000O100000000000000O1000000000000O10000000000O1000000000000O1000lKmMoLS2Q3mMoLS2Q3mMoLS2P3oMoLQ2Q3oMoLP2R3PNnLP2R3PNnLP2R3QNmLo1S3QNmLo1S3QNmLo1S3RNlLn1T3RNlLm1U3SNkLm1T3TNlLm1S3TNlLl1T3TNlLl1T3TNlLl1T3TNlLl1T3UNkLk1U3UNkLl1T3TNlLl1T3UNkLk1U3UNkLk1T3VNlLj1T3VNlLj1T3WNkLi1U3WNkLj1T3VNlLj1T3WNkLi1T3XNlLh1T3XNlLh1S3YNbLR2^3oMVL]2i3cMPLd2o3]MkKi2U4WMhKm2V4UMfKn2Z4RMdKQ3[4oLbKT3]4nL_KU3a4kL\\KY3b4iLZKZ3f4fLWK]3i4dLSK`3k4aLRKb3n4_LnJe3P5\\LnJf3R5[LjJh3U5YLhJk3V5WLfJl3Z5TLcJP4[5RLaJQ4_5oK^JU4`5mK\\JV4c5kKZJY4e5hKWJ[4h5lKPJV4P6d1N101N1O2O0O2O0O2O0O2N1O2O0O2N2O1N1O2N2O1N1O2O1N1O2N2M3L3N3L4M3L5L3L4M4L3L5L3L4M4K5L4L4K6K4K5L5J5L5K4K6K4K6K5J5L6J6I6K6I6K5J5Ommi1"}, "label": "a group of red and green apples"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red and green apple in a fruit bowl\"."}], "task": "refering", "answer_template": "The \"red and green apple in a fruit bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 86, 96, 97, 98, 101, 102, 103, 104, 105, 106, 107, 108, 109, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 284, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.069234375, 0.14116822429906542, 0.783625, 0.9882242990654206], "iscrowd": 0, "area": 105651.02304999999, "rle": {"size": [428, 640], "counts": "icb02X=a0_O5K5K4L5K5K5J6K6J5K6J5K5L3L3M4L3M4M3M2N3M2N2N2N2N2N2N2N2N101N2N2N2N2N2N2N2O1N2N1O2N2N2N2O1N2N2N2N2N2N1O2O1N2N2N2N2N2N2O1N2N1O2N2N2N2N2O0O2N1O1O100O1O2O0O100O100O2O0O1O100O101N1000000O101O000000000O100000000000O0100000000000O01000000000000O0100000000000000O101O000000000O10000O101N100O100O101N101N101N101N101N101N2O0O2O1N101N2O1N101N2O0O2O1N101N2O1N101N2O0O2O1N[IcK^4[4XK\\L]4c3_KbLa4\\3\\KiLd4U3ZKoLf4o2XKUMh4i2VKZMl4c2RKaMn4^2oJeMR5Y2kJkMV5S2gJRNY5l1eJWN\\5g1aJ]N`5a1]JdNc5[1YJiNh5U1XJmNh5Q1XJPOi5n0VJTOk5j0UJXOk5f0UJ[Ol5d0SJ^On5?QJCP6;PJFQ68oIJQ64oIMR61mI1T6NkI4U6JkI7V6GjI:a5^LoJo6m4XIRKg6i4`IUK_6g4hIWKW6d4QJZKn5b4YJ]Kf5_4`J_K_5^4gJ`KX5]4nJaKR5\\4RKdKl4Y4ZKeKe4X4Y2N3L3O1N2N3M2N2N2N2O1N2O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O100O100O100O100O1O100O100O1000000000000O100000000000O10O1000000000000O100000000000000O1000000000000O10000000000O1000000000000O1000lKmMoLS2Q3mMoLS2Q3mMoLS2P3oMoLQ2Q3oMoLP2R3PNnLP2R3PNnLP2R3QNmLo1S3QNmLo1S3QNmLo1S3RNlLn1T3RNlLm1U3SNkLm1T3TNlLm1S3TNlLl1T3TNlLl1T3TNlLl1T3TNlLl1T3UNkLk1U3UNkLl1T3TNlLl1T3UNkLk1U3UNkLk1T3VNlLj1T3VNlLj1T3WNkLi1U3WNkLj1T3VNlLj1T3WNkLi1T3XNlLh1T3XNlLh1S3YNbLR2^3oMVL]2i3cMPLd2o3]MkKi2U4WMhKm2V4UMfKn2Z4RMdKQ3[4oLbKT3]4nL_KU3a4kL\\KY3b4iLZKZ3f4fLWK]3i4dLSK`3k4aLRKb3n4_LnJe3P5\\LnJf3R5[LjJh3U5YLhJk3V5WLfJl3Z5TLcJP4[5RLaJQ4_5oK^JU4`5mK\\JV4c5kKZJY4e5hKWJ[4h5lKPJV4P6d1N101N1O2O0O2O0O2O0O2N1O2O0O2N2O1N1O2N2O1N1O2O1N1O2N2M3L3N3L4M3L5L3L4M4L3L5L3L4M4K5L4L4K6K4K5L5J5L5K4K6K4K6K5J5L6J6I6K6I6K5J5Ommi1"}, "label": "red and green apple in a fruit bowl"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana in fruit bowl\"."}], "task": "refering", "answer_template": "The \"banana in fruit bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340], "bbox": [0.648203125, 0.0017757009345794393, 1.0, 0.9875934579439253], "iscrowd": 0, "area": 51786.346199999985, "rle": {"size": [428, 640], "counts": "dj]5V1S<4L3M3N2M3M3M3M3M3M3N1N3M3M3M3M3M3N2M3M3M3M3M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3N2M3M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N0000O1O100O1O100O1O100O1000001N100000000O2O00000O1000001O00000001O0001O0001O000001O01O00000000010O1O1O1O1O1O10O01O1O1O1O100O1O001O1O11O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N1"}, "label": "banana in fruit bowl"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana with other fruits in a bowl\"."}], "task": "refering", "answer_template": "The \"banana with other fruits in a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340], "bbox": [0.648203125, 0.0017757009345794393, 1.0, 0.9875934579439253], "iscrowd": 0, "area": 51786.346199999985, "rle": {"size": [428, 640], "counts": "dj]5V1S<4L3M3N2M3M3M3M3M3M3N1N3M3M3M3M3M3N2M3M3M3M3M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3N2M3M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N0000O1O100O1O100O1O100O1000001N100000000O2O00000O1000001O00000001O0001O0001O000001O01O00000000010O1O1O1O1O1O10O01O1O1O1O100O1O001O1O11O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N1"}, "label": "banana with other fruits in a bowl"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partially obscured banana\"."}], "task": "refering", "answer_template": "The \"a partially obscured banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 228, 229, 250, 251, 252, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.8110312499999999, 0.5407710280373832, 1.0, 0.985233644859813], "iscrowd": 0, "area": 13229.660200000008, "rle": {"size": [428, 640], "counts": "iZi61[=000O1M3N2M3M3N2M3N2M3M3N2M3N2M3N2M3M3N2M3N2N2M3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2N2M3N2N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1hH"}, "label": "a partially obscured banana"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana that is barely visible to the far right of the fruit\"."}], "task": "refering", "answer_template": "The \"banana that is barely visible to the far right of the fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 228, 229, 250, 251, 252, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.8110312499999999, 0.5407710280373832, 1.0, 0.985233644859813], "iscrowd": 0, "area": 13229.660200000008, "rle": {"size": [428, 640], "counts": "iZi61[=000O1M3N2M3M3N2M3N2M3M3N2M3N2M3N2M3M3N2M3N2N2M3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2N2M3N2N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1hH"}, "label": "banana that is barely visible to the far right of the fruit"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blurred image of yellow skinned apple in background\"."}], "task": "refering", "answer_template": "The \"blurred image of yellow skinned apple in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 98, 99, 100, 101, 122, 123, 124, 146, 147], "bbox": [0.28504687500000003, 0.13605140186915887, 0.48864062500000005, 0.45689252336448594], "iscrowd": 0, "area": 8835.272149999999, "rle": {"size": [428, 640], "counts": "RV\\29P=5L3L4L5L3L4N3M2N2N3L3N2N2N2N2N2M2O2O1N2N1O2O1N1O2O1N2N1O2O1N2N2O1N1O2O1N2O1N2O0O2O1N2O1N101N2O1N2O1N101N2O0O2O001N2O0O2O0O2O1O0O2O001N101O1N101OiNmFUNR9j1ZGmMe8R2aGjM^8T2gGiMX8W2mGfMR8X2THdMl7[2XHcMf7\\2f1O1N101N2O1N101N2N2N2M3N2M3N2N2N2O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1O2N2N2N2N2N3M2N2N2N2N2N2N2N3MgPY4"}, "label": "blurred image of yellow skinned apple in background"}]}
{"id": 331505, "image": "COCO_train2014_000000331505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the part of the orange in the background that is almost completely covered by other fruit\"."}], "task": "refering", "answer_template": "The \"the part of the orange in the background that is almost completely covered by other fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 57, 58, 59, 60, 61, 62, 85, 86], "bbox": [0.47112499999999996, 0.06532710280373832, 0.747921875, 0.24658878504672896], "iscrowd": 0, "area": 3641.954900000002, "rle": {"size": [428, 640], "counts": "nXn33W=4M2M4L3N2N2N2O1N1O3M3N1N3M2N3N2M1O0001O01O0010O01O010O0010O100O10000O100O100O100O2O2M3N1O2N001N2O1O001O1O001N2O001O1O001O0O2O001O001O1N101O001O0010O000001O0000000000010O0000000000001O01O0000000000000000001O00000000001O0000000000001O00000000001N10001O1O001O001N2O001O1O001O001N20O10O10O1000O0100O10O10O1000O010000O010000O010000O010000001O1O1O1O1O1O001CmVS2"}, "label": "the part of the orange in the background that is almost completely covered by other fruit"}]}
{"id": 503541, "image": "COCO_train2014_000000503541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suitcase that isn ' t red\"."}], "task": "refering", "answer_template": "The \"the suitcase that isn ' t red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303], "bbox": [0.063953125, 0.4169555035128805, 0.437890625, 0.9023887587822014], "iscrowd": 0, "area": 37954.6982, "rle": {"size": [427, 640], "counts": "WZa0c2g:c1^N1N2N200N2O1O1LbKXG`4i81O1O1N101O1O0O101O0O101O001N10001O0O2O000O2O00001N101O000O2O001N10001O0O100000O0100000O10O1000O10000O1000O10O1000000O10O10O1000000O10O10O1000000O10O1000OO2M3M2N3M3N2M2N3N2000O10000O0100000O10O10O100000O01000000O01000O1000O10O100000O010001N4M2N3M1N2O001N2O1O001N2O1O1O0O1000O01000000O10O1000O10000O0100000O10O1000O10000O0100000O1000O01000000O10O1000O100000O02O001O0O2O001O000O2O001O0O2O001O0O101O001N101O001N10001O0O2O001O0O2O00001N101O001N1G:B=B>C>B=B?B=C>Acme4"}, "label": "the suitcase that isn ' t red"}]}
{"id": 503541, "image": "COCO_train2014_000000503541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and red beaded suitcase sitting to the left of other red luggage\"."}], "task": "refering", "answer_template": "The \"a white and red beaded suitcase sitting to the left of other red luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303], "bbox": [0.063953125, 0.4169555035128805, 0.437890625, 0.9023887587822014], "iscrowd": 0, "area": 37954.6982, "rle": {"size": [427, 640], "counts": "WZa0c2g:c1^N1N2N200N2O1O1LbKXG`4i81O1O1N101O1O0O101O0O101O001N10001O0O2O000O2O00001N101O000O2O001N10001O0O100000O0100000O10O1000O10000O1000O10O1000000O10O10O1000000O10O10O1000000O10O1000OO2M3M2N3M3N2M2N3N2000O10000O0100000O10O10O100000O01000000O01000O1000O10O100000O010001N4M2N3M1N2O001N2O1O001N2O1O1O0O1000O01000000O10O1000O10000O0100000O10O1000O10000O0100000O1000O01000000O10O1000O100000O02O001O0O2O001O000O2O001O0O2O001O0O101O001N101O001N10001O0O2O001O0O2O00001N101O001N1G:B=B>C>B=B?B=C>Acme4"}, "label": "a white and red beaded suitcase sitting to the left of other red luggage"}]}
{"id": 503541, "image": "COCO_train2014_000000503541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a medium size suitcase in the middle\"."}], "task": "refering", "answer_template": "The \"there is a medium size suitcase in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 199, 200, 222], "bbox": [0.377828125, 0.30337236533957845, 0.73465625, 0.6269555035128805], "iscrowd": 0, "area": 20451.9265, "rle": {"size": [427, 640], "counts": "mRU3:d<=C>K4N3L3N3M2M4M20001O001N101O00001O001N101O00001O001N101O00001O0O2O0^EkMn9V2nEmMR:S2kEPNU:P2hESNX:n1dEUN\\:]2O10001O0000000O1000001O00000O10000000000000000000000O10000000000000000000000O10000000000000000000000000000000000000O1000001O0000000000000000000000001O0000000000000000000000001O00000000000000000000001O00000000000000000000001O000000000O2O000000000010O0000000001O00000000001O01O0hEVMo9j2PFZMm9f2RF_Mj9a2UFcMi9\\2VFhMg9X2XFmMd9S2[FQNb9o1]FVN_9n2M1OO2N1O1O1O1O1O2N1O1N2O1O1O2N1O1O1N2M3M4L3M3N2N2O1N3M2O1N2O1N2N3N1N9G9H9F9G:G8G:FZfV2"}, "label": "there is a medium size suitcase in the middle"}]}
{"id": 503541, "image": "COCO_train2014_000000503541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"medium red leather suticase\"."}], "task": "refering", "answer_template": "The \"medium red leather suticase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 199, 200, 222], "bbox": [0.377828125, 0.30337236533957845, 0.73465625, 0.6269555035128805], "iscrowd": 0, "area": 20451.9265, "rle": {"size": [427, 640], "counts": "mRU3:d<=C>K4N3L3N3M2M4M20001O001N101O00001O001N101O00001O001N101O00001O0O2O0^EkMn9V2nEmMR:S2kEPNU:P2hESNX:n1dEUN\\:]2O10001O0000000O1000001O00000O10000000000000000000000O10000000000000000000000O10000000000000000000000000000000000000O1000001O0000000000000000000000001O0000000000000000000000001O00000000000000000000001O00000000000000000000001O000000000O2O000000000010O0000000001O00000000001O01O0hEVMo9j2PFZMm9f2RF_Mj9a2UFcMi9\\2VFhMg9X2XFmMd9S2[FQNb9o1]FVN_9n2M1OO2N1O1O1O1O1O2N1O1N2O1O1O2N1O1O1N2M3M4L3M3N2N2O1N3M2O1N2O1N2N3N1N9G9H9F9G:G8G:FZfV2"}, "label": "medium red leather suticase"}]}
{"id": 503541, "image": "COCO_train2014_000000503541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small red carry on suitcase with metal decorations\"."}], "task": "refering", "answer_template": "The \"a small red carry on suitcase with metal decorations\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267], "bbox": [0.42589062499999997, 0.5443559718969555, 0.6635, 0.782903981264637], "iscrowd": 0, "area": 14373.1145, "rle": {"size": [427, 640], "counts": "jTb3>^<?A`0A>A?A?C=000000000000000000000000001O000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000002M3M2N3L4C=C>fN_D:P<@SD3[<IhCJg<2S]i2"}, "label": "a small red carry on suitcase with metal decorations"}]}
{"id": 503541, "image": "COCO_train2014_000000503541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red small box\"."}], "task": "refering", "answer_template": "The \"red small box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267], "bbox": [0.42589062499999997, 0.5443559718969555, 0.6635, 0.782903981264637], "iscrowd": 0, "area": 14373.1145, "rle": {"size": [427, 640], "counts": "jTb3>^<?A`0A>A?A?C=000000000000000000000000001O000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000002M3M2N3L4C=C>fN_D:P<@SD3[<IhCJg<2S]i2"}, "label": "red small box"}]}
{"id": 143096, "image": "COCO_train2014_000000143096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"some pieces of apple or something next to the carrots and meat\"."}], "task": "refering", "answer_template": "The \"some pieces of apple or something next to the carrots and meat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289], "bbox": [0.30407812500000003, 0.5416470588235294, 0.6121875, 0.8672], "iscrowd": 0, "area": 15775.692350000001, "rle": {"size": [425, 640], "counts": "RWa23U=5L3M4K4M4K4M4L3L10000O100O10000O2O000O100O10000O10000O2O0O100O1O1O1O1O1O1O1O3M2N3M2N2M4M2N2O2O001N10001N10001N1jNjMXGW2e8lMZGT2e8nMYGS2f8nMZGS2d8oM[GP2e8RNZGn1d8TN[Gl1e8VNZGj1e8XNZGg1f8[NXGf1g8[NYGd1f8_NYGa1f8aNXG_1h8cNWG]1h8dNVG]1j8eNTG\\1j8gNSGZ1m8hNPGZ1o8gNPGY1P9a1O0O2O0000N2O1N1O3M3M4L4L4L4M2M4M3N2N0000O1000000O101O001N101O001N101O001N2O010O0010O01O010O0001O00001O0000001O001O1O1O2N1O1O1O1O1O1O1O00001O001O00001O001O001N10001O001O002N1O2N2N1O2M3N1O2N2N1O2N2N1O2N1O2N2M2O<D9G2N2M3N2M4M2M3N2N2M3N2M3N2M3NlbV3"}, "label": "some pieces of apple or something next to the carrots and meat"}]}
{"id": 143096, "image": "COCO_train2014_000000143096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a serving of cooked potatoes in a white plate next to some roast and carrots\"."}], "task": "refering", "answer_template": "The \"a serving of cooked potatoes in a white plate next to some roast and carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289], "bbox": [0.30407812500000003, 0.5416470588235294, 0.6121875, 0.8672], "iscrowd": 0, "area": 15775.692350000001, "rle": {"size": [425, 640], "counts": "RWa23U=5L3M4K4M4K4M4L3L10000O100O10000O2O000O100O10000O10000O2O0O100O1O1O1O1O1O1O1O3M2N3M2N2M4M2N2O2O001N10001N10001N1jNjMXGW2e8lMZGT2e8nMYGS2f8nMZGS2d8oM[GP2e8RNZGn1d8TN[Gl1e8VNZGj1e8XNZGg1f8[NXGf1g8[NYGd1f8_NYGa1f8aNXG_1h8cNWG]1h8dNVG]1j8eNTG\\1j8gNSGZ1m8hNPGZ1o8gNPGY1P9a1O0O2O0000N2O1N1O3M3M4L4L4L4M2M4M3N2N0000O1000000O101O001N101O001N101O001N2O010O0010O01O010O0001O00001O0000001O001O1O1O2N1O1O1O1O1O1O1O00001O001O00001O001O001N10001O001O002N1O2N2N1O2M3N1O2N2N1O2N2N1O2N1O2N2M2O<D9G2N2M3N2M4M2M3N2N2M3N2M3N2M3NlbV3"}, "label": "a serving of cooked potatoes in a white plate next to some roast and carrots"}]}
{"id": 519929, "image": "COCO_train2014_000000519929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden drumstick to the right of another drumstick\"."}], "task": "refering", "answer_template": "The \"a wooden drumstick to the right of another drumstick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 40, 41, 42, 59, 60, 61, 78, 79, 80, 98, 99, 118, 119, 137, 138], "bbox": [0.21230000000000002, 0.12178666666666667, 0.7199800000000001, 0.6118933333333333], "iscrowd": 0, "area": 4794.2099499999995, "rle": {"size": [375, 500], "counts": "WlV11e;1O2M3N1O2M3N1O2M3N1O2N1N3N2N101000O01000O10O10O100O01000O10O10O1000O01000O0100O10O10O1000O010000O01000O10O010000O01000O10O10O10O10O1000O0100O10O10O1000O010000O01000O10O10O1000O10O100000O0100000O0100000O01000000O0100000O0100000O10O1000O10O1000O1000O10O1000O10O1000O10O100000O0100000O01000000O0100000O018H00O010N2O1O001N01O1O1O1O2N2M3N1O2N2N1OlZ72SeH5L3M4K1000O010000O10O10O1000O010000O10O10O10000O01000O10O1000O1000O010000O10O10O1000O010000O10O1N101N2N1O2N2N101N2N1O2Nc]c1"}, "label": "a wooden drumstick to the right of another drumstick"}]}
{"id": 519929, "image": "COCO_train2014_000000519929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a drum stick the kid is holding in his left hand\"."}], "task": "refering", "answer_template": "The \"a drum stick the kid is holding in his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 93, 94, 95, 112, 113, 114, 131, 132, 133, 134, 151, 152, 153, 154, 171, 172], "bbox": [0.11956, 0.31869333333333333, 0.5925, 0.7246933333333333], "iscrowd": 0, "area": 5427.0003, "rle": {"size": [375, 500], "counts": "WSf02c;3M2N3M3L301O1O01000O01000O010O1000O01000O0100O01000O01000O10O01000O01000O0101N10000O2O0O10001N10000O2O0O1000O010O10O10O010O10O01000O010O0100O01000O010O010O10O10O10O010O01000O0100O010O01000O010O10O01000O010O0100O01000O010O10O010O10O10O0100O010O10O10O010O10O01000O0100O010O01000O010O10O01000O010O10O10O01000O0100O0100O01000O0100O01000O010O01000O01000O010O10O10O010O10O10O10O10O10O010O10O10O01000O010O10O10O01000O0100O0100O0100O001O1N101O1O001O001O1O001O1O001O1O001O2N2N3MW^Z2"}, "label": "a drum stick the kid is holding in his left hand"}]}
{"id": 519929, "image": "COCO_train2014_000000519929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"drum stick which is facing front\"."}], "task": "refering", "answer_template": "The \"drum stick which is facing front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 93, 94, 95, 112, 113, 114, 131, 132, 133, 134, 151, 152, 153, 154, 171, 172], "bbox": [0.11956, 0.31869333333333333, 0.5925, 0.7246933333333333], "iscrowd": 0, "area": 5427.0003, "rle": {"size": [375, 500], "counts": "WSf02c;3M2N3M3L301O1O01000O01000O010O1000O01000O0100O01000O01000O10O01000O01000O0101N10000O2O0O10001N10000O2O0O1000O010O10O10O010O10O01000O010O0100O01000O010O010O10O10O10O010O01000O0100O010O01000O010O10O01000O010O0100O01000O010O10O010O10O10O0100O010O10O10O010O10O01000O0100O010O01000O010O10O01000O010O10O10O01000O0100O0100O01000O0100O01000O010O01000O01000O010O10O10O010O10O10O10O10O10O010O10O10O01000O010O10O10O01000O0100O0100O0100O001O1N101O1O001O001O1O001O1O001O1O001O2N2N3MW^Z2"}, "label": "drum stick which is facing front"}]}
{"id": 265980, "image": "COCO_train2014_000000265980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair with brown seat\"."}], "task": "refering", "answer_template": "The \"a white chair with brown seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 223, 224, 225, 226, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 333, 334, 336, 337, 339, 356, 360, 362, 383, 385], "bbox": [0.497765625, 0.502125, 0.835765625, 0.9864375], "iscrowd": 0, "area": 16574.568250000008, "rle": {"size": [480, 640], "counts": "hne4<a>4L3M2L3K6K5J5I8C<F;O0100O010I6J7I6F;C516010O01O010O00100101O0O101O0O2O000O2O0O2O000O2O001N10001N101O0O101N101O0O101O0O2O000O2O001N100O2O001N100hCUN[;l1`DWN`;j1\\DXNf;i1VDYNk;g1QD\\No;e1nC\\NT<e1gC^NY<T21N2O0O2N2N1O2000O101O0O101N10001N10000010OeEhMn7X2iGQNW8P2^GZNb8f1TGdNl8\\1jFnNW9Q1`FXO`9i0UFBj9j20000001G8I7I7H8I8G8I7I7H8K6O0O1O100OM3L5L3M4L3M3L5L3M4L3L4N3O00010O0100O1O100O100O5K100O10O01O100O1O100O010O1O1006Jd0\\Oe0ZOf0[Oe0[O2N00O1000OnLS3N2M3M3M3M3M2O2M3M2N3M3M2O2M3M2N3M2N3N2M2N3M3M2O2M3M2N3M2N100O1O2N1O1O012M3M5K6J6JZo`1"}, "label": "a white chair with brown seat"}]}
{"id": 265980, "image": "COCO_train2014_000000265980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white chair with brown seat cushion\"."}], "task": "refering", "answer_template": "The \"white chair with brown seat cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 223, 224, 225, 226, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 333, 334, 336, 337, 339, 356, 360, 362, 383, 385], "bbox": [0.497765625, 0.502125, 0.835765625, 0.9864375], "iscrowd": 0, "area": 16574.568250000008, "rle": {"size": [480, 640], "counts": "hne4<a>4L3M2L3K6K5J5I8C<F;O0100O010I6J7I6F;C516010O01O010O00100101O0O101O0O2O000O2O0O2O000O2O001N10001N101O0O101N101O0O101O0O2O000O2O001N100O2O001N100hCUN[;l1`DWN`;j1\\DXNf;i1VDYNk;g1QD\\No;e1nC\\NT<e1gC^NY<T21N2O0O2N2N1O2000O101O0O101N10001N10000010OeEhMn7X2iGQNW8P2^GZNb8f1TGdNl8\\1jFnNW9Q1`FXO`9i0UFBj9j20000001G8I7I7H8I8G8I7I7H8K6O0O1O100OM3L5L3M4L3M3L5L3M4L3L4N3O00010O0100O1O100O100O5K100O10O01O100O1O100O010O1O1006Jd0\\Oe0ZOf0[Oe0[O2N00O1000OnLS3N2M3M3M3M3M2O2M3M2N3M3M2O2M3M2N3M2N3N2M2N3M3M2O2M3M2N3M2N100O1O2N1O1O012M3M5K6J6JZo`1"}, "label": "white chair with brown seat cushion"}]}
{"id": 265980, "image": "COCO_train2014_000000265980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown colored net chair near the table lamp\"."}], "task": "refering", "answer_template": "The \"a brown colored net chair near the table lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 187, 188, 210, 211, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 326, 327, 330, 350, 353, 373], "bbox": [0.161796875, 0.4330208333333333, 0.37753125, 0.9701041666666667], "iscrowd": 0, "area": 11918.70825, "rle": {"size": [480, 640], "counts": "`n`1<d>=C=C=C=D<C>B=CjLXDa3_;;D<D=D;D<D<D<E2MO1J7G8H8H9F9H8H8H9F90010O:G>A?B>A7JN1O2iERNX7o1WHaNi7a1eGoN[8R1TG^Ok8e0dFJ\\97SF9m9]20YMnG`NR8a1f2000000000000000O10000000001O00000000000O1000000000001O0000000000000000000001O01O00000000000000000001O0001O000001O1O1O2N1O2N1O1O2N1O2O0O1O2N1O1O2N1O2N1O1O2N1O2O0O1O2N15L9F;J63M4J6@?@a0@`0_OSTj5"}, "label": "a brown colored net chair near the table lamp"}]}
{"id": 265980, "image": "COCO_train2014_000000265980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair with a cane seat pushed under the side of a table\"."}], "task": "refering", "answer_template": "The \"a wooden chair with a cane seat pushed under the side of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 187, 188, 210, 211, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 326, 327, 330, 350, 353, 373], "bbox": [0.161796875, 0.4330208333333333, 0.37753125, 0.9701041666666667], "iscrowd": 0, "area": 11918.70825, "rle": {"size": [480, 640], "counts": "`n`1<d>=C=C=C=D<C>B=CjLXDa3_;;D<D=D;D<D<D<E2MO1J7G8H8H9F9H8H8H9F90010O:G>A?B>A7JN1O2iERNX7o1WHaNi7a1eGoN[8R1TG^Ok8e0dFJ\\97SF9m9]20YMnG`NR8a1f2000000000000000O10000000001O00000000000O1000000000001O0000000000000000000001O01O00000000000000000001O0001O000001O1O1O2N1O2N1O1O2N1O2O0O1O2N1O1O2N1O2N1O1O2N1O2O0O1O2N15L9F;J63M4J6@?@a0@`0_OSTj5"}, "label": "a wooden chair with a cane seat pushed under the side of a table"}]}
{"id": 462589, "image": "COCO_train2014_000000462589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl with cat food in it\"."}], "task": "refering", "answer_template": "The \"a bowl with cat food in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294], "bbox": [0.58646875, 0.4547083333333333, 0.8255312499999998, 0.7660625], "iscrowd": 0, "area": 17648.2095, "rle": {"size": [480, 640], "counts": "oQ`51k>;F:E<E:E6K5J6K5M3M3L3N3M3M3M3L4M2N3M3M3L4M2N1O1O2N100O2N101N1O2O0O2O0O2N101N1O2O0O2N101N100O2O00000O2O000O2O000O2O001O0O101O0O2O000O2O00000O0101O0000001O0000001O0O10001O0000001O0000001O0O10001O0000001O00000O2O0000001N2O1O1O001O1N2O1O1O1O1N2N2N2O0O2N2N2N2N2N2N2N2N2N1O2N2N2N2M3N2N2N3M5K5K5K5J6K5K4L5F:D<DnVd1"}, "label": "a bowl with cat food in it"}]}
{"id": 462589, "image": "COCO_train2014_000000462589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green boawl with dry cat food in it\"."}], "task": "refering", "answer_template": "The \"a green boawl with dry cat food in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294], "bbox": [0.58646875, 0.4547083333333333, 0.8255312499999998, 0.7660625], "iscrowd": 0, "area": 17648.2095, "rle": {"size": [480, 640], "counts": "oQ`51k>;F:E<E:E6K5J6K5M3M3L3N3M3M3M3L4M2N3M3M3L4M2N1O1O2N100O2N101N1O2O0O2O0O2N101N1O2O0O2N101N100O2O00000O2O000O2O000O2O001O0O101O0O2O000O2O00000O0101O0000001O0000001O0O10001O0000001O0000001O0O10001O0000001O00000O2O0000001N2O1O1O001O1N2O1O1O1O1N2N2N2O0O2N2N2N2N2N2N2N2N2N1O2N2N2N2M3N2N2N3M5K5K5K5J6K5K4L5F:D<DnVd1"}, "label": "a green boawl with dry cat food in it"}]}
{"id": 462589, "image": "COCO_train2014_000000462589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue color cat drawing plate\"."}], "task": "refering", "answer_template": "The \"a blue color cat drawing plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 308, 309], "bbox": [0.27303125, 0.3861458333333333, 0.57978125, 0.7793958333333334], "iscrowd": 0, "area": 28780.436000000005, "rle": {"size": [480, 640], "counts": "jYb27a><E:F;E:E<G8K6K4K5K6K2M4M2M3M3N3M2N2N2N3N1N2N2N2O2M2N2N2N3N1N2N2N3N1N1O2N100O2N1O2N100O2N1O2O0O2N1O101N1O2O001N10001O001N101O00001N101O001O0O101O001O0O101O001O0O2O0000000O100000000O10000000000O010000000O11O0000001O0000001O0000001O00000O2O000000001O0000001O0O100O2N100O101N1O100O2O0O1O2O0O2O1N1O2O1N101N2N101N2O0O2N2O0O2O1N1O2O1N101M3N2N2M3N3M2N2M3N3M2M3N2N2M4M2N2M3N3M2N2M7J5F:G9G:F9F:G9G:Fcjm3"}, "label": "a blue color cat drawing plate"}]}
{"id": 462589, "image": "COCO_train2014_000000462589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue dish with cats painted on it\"."}], "task": "refering", "answer_template": "The \"blue dish with cats painted on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 308, 309], "bbox": [0.27303125, 0.3861458333333333, 0.57978125, 0.7793958333333334], "iscrowd": 0, "area": 28780.436000000005, "rle": {"size": [480, 640], "counts": "jYb27a><E:F;E:E<G8K6K4K5K6K2M4M2M3M3N3M2N2N2N3N1N2N2N2O2M2N2N2N3N1N2N2N3N1N1O2N100O2N1O2N100O2N1O2O0O2N1O101N1O2O001N10001O001N101O00001N101O001O0O101O001O0O101O001O0O2O0000000O100000000O10000000000O010000000O11O0000001O0000001O0000001O00000O2O000000001O0000001O0O100O2N100O101N1O100O2O0O1O2O0O2O1N1O2O1N101N2N101N2O0O2N2O0O2O1N1O2O1N101M3N2N2M3N3M2N2M3N3M2M3N2N2M4M2N2M3N3M2N2M7J5F:G9G:F9F:G9G:Fcjm3"}, "label": "blue dish with cats painted on it"}]}
{"id": 462589, "image": "COCO_train2014_000000462589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black cats head and ears\"."}], "task": "refering", "answer_template": "The \"black cats head and ears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [280, 302, 303, 304, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 360, 361, 362, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.15462499999999998, 0.7416874999999999, 0.763625, 0.9989166666666667], "iscrowd": 0, "area": 27244.439150000006, "rle": {"size": [480, 640], "counts": "ej^18\\>?A?A?F;H7I7J6I8I4K5L4K5L4N2N2O000O10000O101O0O100O100001O1O1O001O1O001O1O1O001O1O1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N00001O001O00001O00001O001O00001O001O00001O001O0O101O001O00001O000000O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O10O01O100000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001N1000001O0000001O0000001O0000001O0000001O0000001O001O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1N2O1O2N1O1O1O1O1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O0000O100000000O100000000O010000000O100000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000O10000000O100000000000000000000000000000000000000000000000000000000001O2N3M2N3M2L4L5K4MbZV2"}, "label": "black cats head and ears"}]}
{"id": 462589, "image": "COCO_train2014_000000462589.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat ' s black head\"."}], "task": "refering", "answer_template": "The \"cat ' s black head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [280, 302, 303, 304, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 360, 361, 362, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.15462499999999998, 0.7416874999999999, 0.763625, 0.9989166666666667], "iscrowd": 0, "area": 27244.439150000006, "rle": {"size": [480, 640], "counts": "ej^18\\>?A?A?F;H7I7J6I8I4K5L4K5L4N2N2O000O10000O101O0O100O100001O1O1O001O1O001O1O1O001O1O1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N00001O001O00001O00001O001O00001O001O00001O001O0O101O001O00001O000000O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O10O01O100000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001N1000001O0000001O0000001O0000001O0000001O0000001O001O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1N2O1O2N1O1O1O1O1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O0000O100000000O100000000O010000000O100000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000O10000000O100000000000000000000000000000000000000000000000000000000001O2N3M2N3M2L4L5K4MbZV2"}, "label": "cat ' s black head"}]}
{"id": 85759, "image": "COCO_train2014_000000085759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy that is skateboarding\"."}], "task": "refering", "answer_template": "The \"the boy that is skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 54, 55, 56, 66, 67, 68, 69, 70, 71, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 203, 204, 205, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 246, 247, 248, 249, 261, 262, 263, 276, 277, 278, 279], "bbox": [0.2228169014084507, 0.096625, 0.7765023474178404, 0.8202187500000001], "iscrowd": 0, "area": 48824.05394999998, "rle": {"size": [640, 426], "counts": "Udk12lc03N3L4M2M4M3L3M4M2M2O2M2O2M2N3N1N3N1N3N1N4L3N2L4M3L5L4M7H5K4M3L5K4M3L4L3N2M2N3N1N3N1N3N1N3N1O2M2O2N1O2N1O2O0O2N1O2O0O2N1O2O0O2N1O2O0O1O101O0O10001O0O1O2M2N2N2N3M2N2N2N3M2N2N2N3N1^GlIX3T6fLoIX3R6fLRJX3o5eLUJY3k5eLXJZ3h5dL\\JZ3e5cL_JZ3b5dLaJ[3_5cLeJ[3[5cLiJ[3X5bLkJ\\3V5bLlJ^3T5]KQLc4o3\\KRLd4o3YKSLg4m3WKULh4l3VKVLj4j3TKXLl4i3QKYLo4g3PKZLP5f3nJ\\LQ5e3oJ[LQ5f3mJ[LS5e3mJ\\LR5d3nJ\\LQ5f3mJ[LS5e3mJ[LS5e3lJ\\LT5d3lJ\\LT5e3jJ\\LU5e3kJ[LU5f3hJ\\LX5e3eJ]L[5c3cJ_L]5b3`J`L`5a3^J`Lb5a3[JaLe5`3XJbLh5_3UJcLk5]3eISM[6n2_IWMa6j2XI\\Mh6e2TI^Ml6c2PI`MP7a2mHaMS7_2jHdMV7]2fHfMZ7[2bHhM^7]70001O00000000000000[NbHhF]7T9jHjFV7V9kHiFU7W9mHgFS7Y9nHfFR7Z9PIdFP7\\9QIcFo6]9SIaFm6_9TI`Fk6a9WI]Fi6c9XI\\F\\6TORI`:d0ZFX6YOSI]:f0YFT6^OUIY:i0WFP6CVIU:l0WFk5HXIQ:n0VFh5LYIn9o0VFe5O\\Ik9P1UFb53\\Ik8MYGW1I\\57_Ie81ZGU1IY5;`I_86\\GR1Io4f0hIo7>aGk0Jf4h9oJ]F=J`4k9VKZF;J\\4o9ZKVF;JY4Q:^KTF:JU4T:cKPF9LR4V:fKmE9LQ4V:hKmE8LP4V:jKmE7Lo3W:kKkE8Mm3W:mKkE6Mn3W:mKlE6Kn3b3TKU2l0ZJ3Lo3e3fK`1=jJ0On3f3ZLl0`7TOTI8k6I`IKa64bII_67cIG]68fIZOf6f0\\IhNT7W1oHiNo6W1SIiNk6W1WIiNg6V1\\IiNc6W1_IiN_6V1dIjN\\6T1fIlNY6R1jImNW6Q1kIoNT6P1nIPOR6m0QJSOn5l0TJTOl5i0WJ]MhKT1P:^1ZJVMoK\\1e9\\1_JQMULc1Y9[1^KfN^4Z1eKgNX4W1lKkNo3T1VLnNe3P1_LSO[3l0iLQOX3l0mLQOS3n0RMnNo2P1UMmNl2P1YMmNg2R1^MjNc2T1aMjN^2U1fMhN[2U1jMiNU2U1PNhNQ2V1TNgNl1V1YNgNh1W1\\NgNe1U1`NhNa1V1cNhN]1U1iNgNX1W1lNgNT1W1QOeNP1X1UOfNk0X1ZOdNg0Y1^OeNb0Y1AeN`0Y1BgN>V1EiN=T1ElN;R1[9MlXk1"}, "label": "the boy that is skateboarding"}]}
{"id": 85759, "image": "COCO_train2014_000000085759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in jeans skateboarding\"."}], "task": "refering", "answer_template": "The \"a man in jeans skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 54, 55, 56, 66, 67, 68, 69, 70, 71, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 203, 204, 205, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 246, 247, 248, 249, 261, 262, 263, 276, 277, 278, 279], "bbox": [0.2228169014084507, 0.096625, 0.7765023474178404, 0.8202187500000001], "iscrowd": 0, "area": 48824.05394999998, "rle": {"size": [640, 426], "counts": "Udk12lc03N3L4M2M4M3L3M4M2M2O2M2O2M2N3N1N3N1N3N1N4L3N2L4M3L5L4M7H5K4M3L5K4M3L4L3N2M2N3N1N3N1N3N1N3N1O2M2O2N1O2N1O2O0O2N1O2O0O2N1O2O0O2N1O2O0O1O101O0O10001O0O1O2M2N2N2N3M2N2N2N3M2N2N2N3N1^GlIX3T6fLoIX3R6fLRJX3o5eLUJY3k5eLXJZ3h5dL\\JZ3e5cL_JZ3b5dLaJ[3_5cLeJ[3[5cLiJ[3X5bLkJ\\3V5bLlJ^3T5]KQLc4o3\\KRLd4o3YKSLg4m3WKULh4l3VKVLj4j3TKXLl4i3QKYLo4g3PKZLP5f3nJ\\LQ5e3oJ[LQ5f3mJ[LS5e3mJ\\LR5d3nJ\\LQ5f3mJ[LS5e3mJ[LS5e3lJ\\LT5d3lJ\\LT5e3jJ\\LU5e3kJ[LU5f3hJ\\LX5e3eJ]L[5c3cJ_L]5b3`J`L`5a3^J`Lb5a3[JaLe5`3XJbLh5_3UJcLk5]3eISM[6n2_IWMa6j2XI\\Mh6e2TI^Ml6c2PI`MP7a2mHaMS7_2jHdMV7]2fHfMZ7[2bHhM^7]70001O00000000000000[NbHhF]7T9jHjFV7V9kHiFU7W9mHgFS7Y9nHfFR7Z9PIdFP7\\9QIcFo6]9SIaFm6_9TI`Fk6a9WI]Fi6c9XI\\F\\6TORI`:d0ZFX6YOSI]:f0YFT6^OUIY:i0WFP6CVIU:l0WFk5HXIQ:n0VFh5LYIn9o0VFe5O\\Ik9P1UFb53\\Ik8MYGW1I\\57_Ie81ZGU1IY5;`I_86\\GR1Io4f0hIo7>aGk0Jf4h9oJ]F=J`4k9VKZF;J\\4o9ZKVF;JY4Q:^KTF:JU4T:cKPF9LR4V:fKmE9LQ4V:hKmE8LP4V:jKmE7Lo3W:kKkE8Mm3W:mKkE6Mn3W:mKlE6Kn3b3TKU2l0ZJ3Lo3e3fK`1=jJ0On3f3ZLl0`7TOTI8k6I`IKa64bII_67cIG]68fIZOf6f0\\IhNT7W1oHiNo6W1SIiNk6W1WIiNg6V1\\IiNc6W1_IiN_6V1dIjN\\6T1fIlNY6R1jImNW6Q1kIoNT6P1nIPOR6m0QJSOn5l0TJTOl5i0WJ]MhKT1P:^1ZJVMoK\\1e9\\1_JQMULc1Y9[1^KfN^4Z1eKgNX4W1lKkNo3T1VLnNe3P1_LSO[3l0iLQOX3l0mLQOS3n0RMnNo2P1UMmNl2P1YMmNg2R1^MjNc2T1aMjN^2U1fMhN[2U1jMiNU2U1PNhNQ2V1TNgNl1V1YNgNh1W1\\NgNe1U1`NhNa1V1cNhN]1U1iNgNX1W1lNgNT1W1QOeNP1X1UOfNk0X1ZOdNg0Y1^OeNb0Y1AeN`0Y1BgN>V1EiN=T1ElN;R1[9MlXk1"}, "label": "a man in jeans skateboarding"}]}
{"id": 356097, "image": "COCO_train2014_000000356097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two middle planes\"."}], "task": "refering", "answer_template": "The \"two middle planes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 169, 191, 192, 193, 215, 216, 239], "bbox": [0.249171875, 0.3640486725663717, 0.43962500000000004, 0.6494469026548673], "iscrowd": 0, "area": 3118.8805999999986, "rle": {"size": [452, 640], "counts": "ZWW21T>1O001O000O2O001O001O000O2O001O00001N101O00001O0O2O00001O000O1000000000000000000O101M2O1N2_OYObCi0^<YO_Ch0`<ZO]Ch0c<ZOYCh0g<YOWCh0i<:O100001O00100O001O001O1O001OM3N2M3M3M3M3O1OfNfCl0W<SOlCm0S<QOPDo0o;POSDP1l;nNWDR1h;lN[DT1d;jN_DV1_;jNcDV1\\;hNgDX1S<00CVC_Oj<`0WC@i<?XCAh<=ZCCf<<[CDe<;\\CFc<9^CGb<8_CHa<6bCI^<6cCJ]<5dCK\\<4eCL[<3fCMZ<2gCNX<1jCOU<1lCOS<1nCOP<2QDNn;2TDMk;4UDLi;7VDIi;8WDIg;8YDHe;:[DFd;NgC1e01c;LmC2?2b;KUD094b;I\\D035`;JbDON6`;I^E6Y<O100O1O1O1Omcn4"}, "label": "two middle planes"}]}
{"id": 356097, "image": "COCO_train2014_000000356097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two identical blue planes between two other identical planes /\"."}], "task": "refering", "answer_template": "The \"two identical blue planes between two other identical planes /\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 169, 191, 192, 193, 215, 216, 239], "bbox": [0.249171875, 0.3640486725663717, 0.43962500000000004, 0.6494469026548673], "iscrowd": 0, "area": 3118.8805999999986, "rle": {"size": [452, 640], "counts": "ZWW21T>1O001O000O2O001O001O000O2O001O00001N101O00001O0O2O00001O000O1000000000000000000O101M2O1N2_OYObCi0^<YO_Ch0`<ZO]Ch0c<ZOYCh0g<YOWCh0i<:O100001O00100O001O001O1O001OM3N2M3M3M3M3O1OfNfCl0W<SOlCm0S<QOPDo0o;POSDP1l;nNWDR1h;lN[DT1d;jN_DV1_;jNcDV1\\;hNgDX1S<00CVC_Oj<`0WC@i<?XCAh<=ZCCf<<[CDe<;\\CFc<9^CGb<8_CHa<6bCI^<6cCJ]<5dCK\\<4eCL[<3fCMZ<2gCNX<1jCOU<1lCOS<1nCOP<2QDNn;2TDMk;4UDLi;7VDIi;8WDIg;8YDHe;:[DFd;NgC1e01c;LmC2?2b;KUD094b;I\\D035`;JbDON6`;I^E6Y<O100O1O1O1Omcn4"}, "label": "two identical blue planes between two other identical planes /"}]}
{"id": 397058, "image": "COCO_train2014_000000397058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a phone with square and stripes on the case\"."}], "task": "refering", "answer_template": "The \"a phone with square and stripes on the case\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 351, 352, 353, 354, 375], "bbox": [0.021078125, 0.23510775862068967, 0.48820312499999996, 0.9644181034482758], "iscrowd": 0, "area": 54572.95075, "rle": {"size": [464, 640], "counts": "ia66V>6J5K5K6K4_OQO\\CS1m<4DiNZC\\1c<eNZC_1d<bNYCb1f<5N3M2O1N2N3M2O1N2N3N1N2N2N3N1N2N2O2M2N2N2O2M2N2O2M2N3M3N2M3M3N2M3M3N2M2N3M3N2M3M3N1N2N2N2O1N2N2O1N2N2N2O1N3M2O1N2N2N2O1N2N2O1N2N2N2O1N3M2O2M2N3M2O1N3M2O2M2N2O2M2N2N10O01O10O01O1O10O01O10O01O1O101N3O1N2O001N2O1Oe0[O0O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O101O000O10000O0100000O10000O10000O1000000O10O2O001O00001O001O00001O001O00001N100O2N1O2N1O1O2N1N3M2O1N3M2O2M2N2N3N1N2N3N1N3M2O1N3M2O2M2N2N3N1N3M2O1N3M2O1N3M2O2M2N2O2M2N3M2O1N3M2O2M2N2O2M2N2O2M2N3M2O1N3M2O2M2N2O2M2N3N1N2N3M2O1N3M2O2M2N2O2M2N3N1N2N3M2O2M2N2O2M2N2O2M2N3N1N2N3M2O2M2N2O2M2N3N1N2N3N1L4@a0^OWZd4"}, "label": "a phone with square and stripes on the case"}]}
{"id": 397058, "image": "COCO_train2014_000000397058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"phone on left\"."}], "task": "refering", "answer_template": "The \"phone on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 351, 352, 353, 354, 375], "bbox": [0.021078125, 0.23510775862068967, 0.48820312499999996, 0.9644181034482758], "iscrowd": 0, "area": 54572.95075, "rle": {"size": [464, 640], "counts": "ia66V>6J5K5K6K4_OQO\\CS1m<4DiNZC\\1c<eNZC_1d<bNYCb1f<5N3M2O1N2N3M2O1N2N3N1N2N2N3N1N2N2O2M2N2N2O2M2N2O2M2N3M3N2M3M3N2M3M3N2M2N3M3N2M3M3N1N2N2N2O1N2N2O1N2N2N2O1N3M2O1N2N2N2O1N2N2O1N2N2N2O1N3M2O2M2N3M2O1N3M2O2M2N2O2M2N2N10O01O10O01O1O10O01O10O01O1O101N3O1N2O001N2O1Oe0[O0O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O101O000O10000O0100000O10000O10000O1000000O10O2O001O00001O001O00001O001O00001N100O2N1O2N1O1O2N1N3M2O1N3M2O2M2N2N3N1N2N3N1N3M2O1N3M2O2M2N2N3N1N3M2O1N3M2O1N3M2O2M2N2O2M2N3M2O1N3M2O2M2N2O2M2N2O2M2N3M2O1N3M2O2M2N2O2M2N3N1N2N3M2O1N3M2O2M2N2O2M2N3N1N2N3M2O2M2N2O2M2N2O2M2N3N1N2N3M2O2M2N2O2M2N3N1N2N3N1L4@a0^OWZd4"}, "label": "phone on left"}]}
{"id": 53004, "image": "COCO_train2014_000000053004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female in blue jeans\"."}], "task": "refering", "answer_template": "The \"female in blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 116, 117, 131, 132, 143, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 190, 191, 192, 193, 205, 206, 207, 208, 220, 221, 222, 223, 235, 236, 237, 238, 250, 251, 252, 266, 267, 281, 282, 296, 297, 310, 311, 312, 325], "bbox": [0.5667447306791569, 0.298375, 0.9306791569086651, 0.9190312500000001], "iscrowd": 0, "area": 26853.798249999996, "rle": {"size": [640, 427], "counts": "i`g42nc03M3M1O1N2L3M4O1O1N3O1N1O2O0O1O2N100O2O001O000O11O00012M5K00000000000000001O001NO1N2N2N2N2iDoN]3S1bLoN[3S1cLoN[3S1eLnNX3S1hLoNT3R1mLPOo2R1kGXOT2Fo5R1[G0a2POP6R1]G4_2kNS6Q1]G:\\2fNT6P1_G`0Y2aNV6n0bGg0S2]NX6l0dGn0P2VNZ6k0gGV1i1PN_6i0gG^1e1jMb6g0QG^2Y2kLd6g0RGd2V2fLh6h0mFh2W2bLj6i0jFj2Z2]Ll6k0fFk2\\2[Lm6m0bFl2_2XLo6n0]Fn2c2TLo6P1WFS3e2QLT7i5dH^J[7d5_HaJ`7a5ZHdJf7]5]GYIB\\1P9^5ZG[IAX1h7lNgHd6M[I@W1j7VO_HZ64ZIBV1i7[O^HW66YI@V1k7^O\\HV66WI@X1k7A\\HR66WI@W1n7CYHR66oJ`7SOXHP66mJ`7WOWHo56jJc7XOWHX6L`Jk7ZOXHc8i7^GVHa8j7`GUHa8j7`GVH_8k7k1O000100O1O100O100O1O100O1O100O1GmGeDS8X;<L4L4M3L5L3M4N2M3]NcF`H`9Y7lF^HW9\\7TG\\HP9]7\\G[Hf8_7fGXH^8c7hGYHZ8d7jGXHZ8c7kGYHX8d7U2K6K8G9eIoBY5Z=`JfB]5g=XJZBf5Q>N2N3L3N2N2N2N3G8dN\\AoLQ?o2XA]LU?a3g0M2N2N3M3M3M3M3M5K5K4K6K5J6J5K6J6E]_b0"}, "label": "female in blue jeans"}]}
{"id": 53004, "image": "COCO_train2014_000000053004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in blue dress using remote control\"."}], "task": "refering", "answer_template": "The \"a girl in blue dress using remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 116, 117, 131, 132, 143, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 190, 191, 192, 193, 205, 206, 207, 208, 220, 221, 222, 223, 235, 236, 237, 238, 250, 251, 252, 266, 267, 281, 282, 296, 297, 310, 311, 312, 325], "bbox": [0.5667447306791569, 0.298375, 0.9306791569086651, 0.9190312500000001], "iscrowd": 0, "area": 26853.798249999996, "rle": {"size": [640, 427], "counts": "i`g42nc03M3M1O1N2L3M4O1O1N3O1N1O2O0O1O2N100O2O001O000O11O00012M5K00000000000000001O001NO1N2N2N2N2iDoN]3S1bLoN[3S1cLoN[3S1eLnNX3S1hLoNT3R1mLPOo2R1kGXOT2Fo5R1[G0a2POP6R1]G4_2kNS6Q1]G:\\2fNT6P1_G`0Y2aNV6n0bGg0S2]NX6l0dGn0P2VNZ6k0gGV1i1PN_6i0gG^1e1jMb6g0QG^2Y2kLd6g0RGd2V2fLh6h0mFh2W2bLj6i0jFj2Z2]Ll6k0fFk2\\2[Lm6m0bFl2_2XLo6n0]Fn2c2TLo6P1WFS3e2QLT7i5dH^J[7d5_HaJ`7a5ZHdJf7]5]GYIB\\1P9^5ZG[IAX1h7lNgHd6M[I@W1j7VO_HZ64ZIBV1i7[O^HW66YI@V1k7^O\\HV66WI@X1k7A\\HR66WI@W1n7CYHR66oJ`7SOXHP66mJ`7WOWHo56jJc7XOWHX6L`Jk7ZOXHc8i7^GVHa8j7`GUHa8j7`GVH_8k7k1O000100O1O100O100O1O100O1O100O1GmGeDS8X;<L4L4M3L5L3M4N2M3]NcF`H`9Y7lF^HW9\\7TG\\HP9]7\\G[Hf8_7fGXH^8c7hGYHZ8d7jGXHZ8c7kGYHX8d7U2K6K8G9eIoBY5Z=`JfB]5g=XJZBf5Q>N2N3L3N2N2N2N3G8dN\\AoLQ?o2XA]LU?a3g0M2N2N3M3M3M3M3M5K5K4K6K5J6J5K6J6E]_b0"}, "label": "a girl in blue dress using remote control"}]}
{"id": 53004, "image": "COCO_train2014_000000053004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in baby blue hoodie and pajamas playing wii\"."}], "task": "refering", "answer_template": "The \"woman in baby blue hoodie and pajamas playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 110, 111, 125, 126, 127, 139, 140, 141, 142, 154, 155, 156, 157, 169, 170, 171, 172, 184, 185, 186, 187, 199, 200, 201, 202, 215, 216, 217, 230, 231, 232, 244, 245, 246, 247, 262], "bbox": [0.2911709601873536, 0.28137500000000004, 0.5057611241217798, 0.7474375], "iscrowd": 0, "area": 17954.651950000003, "rle": {"size": [640, 427], "counts": "\\]^24kc03TOKg]O9Sb08^]OL]b0o0J6L2O2M2RBZN`9i1_F[N\\9f1dF^NV9e1hF\\NV9f1YDnMl0<j:j1lC[NT1Kn:V2ZC_Na1]OR;o3iDQLU;W4eDiKY;_4`DbK^;f4[D\\Kc;j4WDWKY:^OQFb5_OQK_:HnDF=h5OkJd:7ZET5LfJi:8ZEX5FbJm:9[E\\5A\\JS;;ZET6f:mIXET6g:oIVER6i:QJUEn5l:UJPEl5P;WJmDi5S;b100000O10O10000000O1O1O2N2N2N2N1O2N2N3M2N3M36a0^OI1O0O10001O00000001O01O1O2O1N1O2ON2O1O1O1O:F2N2ZKcCe1a<UNfCYN4P2j<mNoDi0Y;iNWEn0P;eN`EP1R?F:F:gNU]Oj0Sc0M3M3M3K5Ge]S4"}, "label": "woman in baby blue hoodie and pajamas playing wii"}]}
{"id": 53004, "image": "COCO_train2014_000000053004.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady standing in the left side of the image\"."}], "task": "refering", "answer_template": "The \"lady standing in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 110, 111, 125, 126, 127, 139, 140, 141, 142, 154, 155, 156, 157, 169, 170, 171, 172, 184, 185, 186, 187, 199, 200, 201, 202, 215, 216, 217, 230, 231, 232, 244, 245, 246, 247, 262], "bbox": [0.2911709601873536, 0.28137500000000004, 0.5057611241217798, 0.7474375], "iscrowd": 0, "area": 17954.651950000003, "rle": {"size": [640, 427], "counts": "\\]^24kc03TOKg]O9Sb08^]OL]b0o0J6L2O2M2RBZN`9i1_F[N\\9f1dF^NV9e1hF\\NV9f1YDnMl0<j:j1lC[NT1Kn:V2ZC_Na1]OR;o3iDQLU;W4eDiKY;_4`DbK^;f4[D\\Kc;j4WDWKY:^OQFb5_OQK_:HnDF=h5OkJd:7ZET5LfJi:8ZEX5FbJm:9[E\\5A\\JS;;ZET6f:mIXET6g:oIVER6i:QJUEn5l:UJPEl5P;WJmDi5S;b100000O10O10000000O1O1O2N2N2N2N1O2N2N3M2N3M36a0^OI1O0O10001O00000001O01O1O2O1N1O2ON2O1O1O1O:F2N2ZKcCe1a<UNfCYN4P2j<mNoDi0Y;iNWEn0P;eN`EP1R?F:F:gNU]Oj0Sc0M3M3M3K5Ge]S4"}, "label": "lady standing in the left side of the image"}]}
{"id": 560909, "image": "COCO_train2014_000000560909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large black purse with gold emblem that a woman and a dog is looking through\"."}], "task": "refering", "answer_template": "The \"a large black purse with gold emblem that a woman and a dog is looking through\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 306, 307, 308, 309], "bbox": [0.22567187500000002, 0.5431310679611651, 0.5236875, 0.9251456310679611], "iscrowd": 0, "area": 20717.6514, "rle": {"size": [412, 640], "counts": "Qcj12j<1O1N2O2N1N2O1N2O1O1N2O2N2M4M2N3L3N3L3N3M2M4M2N3L3N3M2M4M2N3M2M4M2N3M2M4M2N2M4M2N2N2M4M2N2M3N2N0nFbLf8]3XGfLg8Z3VGkLi8e3O0O2O1O001N2O001N2O00000O1000000O10000O10000000000O100000000O100000000O10000000000O100000000O0100O100O010O100O10O0100O100O10O0100O100O010O10000O0100000O01000O1000O10O1000O010000O10O1000O10O10O10000O2O1O0O2O1O0O2O1O001N2O0O3N2M3M3M2O2M3M3N2M3M3N1N3M3N2M3M3N1N6J9H8G8H9H8G9G8I8Ga\\j3"}, "label": "a large black purse with gold emblem that a woman and a dog is looking through"}]}
{"id": 560909, "image": "COCO_train2014_000000560909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black purse on a bed with a woman and two dogs\"."}], "task": "refering", "answer_template": "The \"a black purse on a bed with a woman and two dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 306, 307, 308, 309], "bbox": [0.22567187500000002, 0.5431310679611651, 0.5236875, 0.9251456310679611], "iscrowd": 0, "area": 20717.6514, "rle": {"size": [412, 640], "counts": "Qcj12j<1O1N2O2N1N2O1N2O1O1N2O2N2M4M2N3L3N3L3N3M2M4M2N3L3N3M2M4M2N3M2M4M2N3M2M4M2N2M4M2N2N2M4M2N2M3N2N0nFbLf8]3XGfLg8Z3VGkLi8e3O0O2O1O001N2O001N2O00000O1000000O10000O10000000000O100000000O100000000O10000000000O100000000O0100O100O010O100O10O0100O100O10O0100O100O010O10000O0100000O01000O1000O10O1000O010000O10O1000O10O10O10000O2O1O0O2O1O0O2O1O001N2O0O3N2M3M3M2O2M3M3N2M3M3N1N3M3N2M3M3N1N6J9H8G8H9H8G9G8I8Ga\\j3"}, "label": "a black purse on a bed with a woman and two dogs"}]}
{"id": 560909, "image": "COCO_train2014_000000560909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown dog with his nose in the purse\"."}], "task": "refering", "answer_template": "The \"a light brown dog with his nose in the purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 218, 219, 220, 221, 241, 242, 243, 244, 265, 266, 267, 288, 289, 290], "bbox": [0.34403125, 0.39169902912621357, 0.6463593750000001, 0.8572815533980582], "iscrowd": 0, "area": 18101.195200000006, "rle": {"size": [412, 640], "counts": "Vfh21j<1O0O2O1O1N2O1NO`CK^<7bCHf1Oo8;\\EE\\18W94^ED8948U:NbE6K4c:J]Eo0c:e0O1O1O10O01000000000O100000000O10000000000O100000000O1000000O100000000O10000000000000O100000000000000000000000000000001O1O001O001O1O001O001O1O001O010O1O001O100O2N1O2O1N1O2N101N1O4L4L4M4K4L4L3M3M3M001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2O1N4L3N3L4M1N2N2O1N2O1N1O010O11N100O2N100O2N100O2N2O1N3M2O2\\LmGY2U8dMmG[2V8aMlG_2V8]MmGa2U8\\MmGc2V8YMmGf2U9O2N101N1O2N1O1O2N1O2N1O2N1N3N1O2M6K5K6bNdEOh;M3M3L4M3M4Lkoj2"}, "label": "a light brown dog with his nose in the purse"}]}
{"id": 560909, "image": "COCO_train2014_000000560909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog has mouth in woman ' s bag\"."}], "task": "refering", "answer_template": "The \"dog has mouth in woman ' s bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 218, 219, 220, 221, 241, 242, 243, 244, 265, 266, 267, 288, 289, 290], "bbox": [0.34403125, 0.39169902912621357, 0.6463593750000001, 0.8572815533980582], "iscrowd": 0, "area": 18101.195200000006, "rle": {"size": [412, 640], "counts": "Vfh21j<1O0O2O1O1N2O1NO`CK^<7bCHf1Oo8;\\EE\\18W94^ED8948U:NbE6K4c:J]Eo0c:e0O1O1O10O01000000000O100000000O10000000000O100000000O1000000O100000000O10000000000000O100000000000000000000000000000001O1O001O001O1O001O001O1O001O010O1O001O100O2N1O2O1N1O2N101N1O4L4L4M4K4L4L3M3M3M001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2O1N4L3N3L4M1N2N2O1N2O1N1O010O11N100O2N100O2N100O2N2O1N3M2O2\\LmGY2U8dMmG[2V8aMlG_2V8]MmGa2U8\\MmGc2V8YMmGf2U9O2N101N1O2N1O1O2N1O2N1O2N1N3N1O2M6K5K6bNdEOh;M3M3L4M3M4Lkoj2"}, "label": "dog has mouth in woman ' s bag"}]}
{"id": 126737, "image": "COCO_train2014_000000126737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food item between others\"."}], "task": "refering", "answer_template": "The \"a food item between others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 188, 189, 190, 191, 206, 207, 208, 209, 210, 211, 212, 213, 228, 229, 230, 231, 232, 233, 234, 235, 250, 251, 252, 253, 254, 255, 256, 273, 274, 275, 276], "bbox": [0.3400653594771242, 0.25300653594771244, 0.7172549019607843, 0.58359477124183], "iscrowd": 0, "area": 31727.321900000006, "rle": {"size": [612, 612], "counts": "c`l31Sc03L4M3M2M4M3M3L4M3M3M3L4M3M3L4M3M3L3N3M3M3L4M3M3L4M3M3L4M3M3M3L3N3M1N3N1O1N3N1O1O2M2O1O2M2O1O2M2O1O2N1N2O2N1N2O2N1N2O2N1O0O2O001N101O0O101O001N101O0O2O001N101O001N101O0O2O001N100000O01000O01000O0100000O01000O01000O01000O10O1000O01000O01000O10O1000O01000O01000O0100000O01000O01000O010000O01000O01000O01000O10O10O10O10O10O10O1000O01000O10O10O10O10O10O1000O01000O01000O2O1O1O1N101O1N2O1O1N2O1O1O0O2O1O1N2O1O1N2O1O003K6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6JXTW3"}, "label": "a food item between others"}]}
{"id": 126737, "image": "COCO_train2014_000000126737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dirty butter knife on a plate , being used to eat food\"."}], "task": "refering", "answer_template": "The \"a dirty butter knife on a plate , being used to eat food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 194, 195, 196, 197, 218, 219], "bbox": [0.6741503267973856, 0.22022875816993465, 0.9955065359477124, 0.4337091503267974], "iscrowd": 0, "area": 8294.6399, "rle": {"size": [612, 612], "counts": "knf7l0Yb0?@010O0010O010O010O010O01O010O010O010O010O01O01O010O010O010O010O0010O010O010O010O010O0010O0010O010O010O010O010O010O010O010O01O01O010O010O010O010O010O010O010O01O010O010O01O010O010O010O010O010O00010O010O010O0010O010O010O01O010O010O0010O010O010O01O010O010O010O0010O010O010O01O010O010O010O0010O010O010O01O010O010O010O0010O010O010O01O010O010O010O00100O010O0010O010O010O01O010O010O010O0010O01C<Beb1"}, "label": "a dirty butter knife on a plate , being used to eat food"}]}
{"id": 126737, "image": "COCO_train2014_000000126737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of food on the left - hand side of the plate\"."}], "task": "refering", "answer_template": "The \"the piece of food on the left - hand side of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 355, 356, 357, 358, 359, 360], "bbox": [0.08052287581699347, 0.3688235294117647, 0.532467320261438, 0.7636437908496733], "iscrowd": 0, "area": 40398.2054, "rle": {"size": [612, 612], "counts": "lcm02nb06K6I7J5J7J6I7J5J7J6I6K6I7J5J7J6I7J5J7J6I6K6I7J5N3L4L4L3M4M3L3M4L4L3M4M3L4L3M4L4M2M4L3M1O1O2O0O1O1O1O101N1O1O1O1O101N1O1O1O100O2N1O1000000010O00000000010O1O1O1O001O1O1O1O001N2O1O1O1O001O1N2O1O001O1O1O1N2O001O1O1O1O0O2O1O1O1O1O001N2O1O1O001O1N2O1O1O001O1O1N2O1O001O1O1O1N101O1O1O1O1O0O2O1O1O1O001O1N2O1O1O001O1N2O1O001O1N2O1O1N101O1O1N2O001N2O1O1O1N101O1N2O1O0O2O1O1O1N2O001N2O1O1O1N101O1N2O1O0O2O1O1O1N2O001N2O1O1O0O2O1O1N2O1O0O2O1O1O1N101O1N2O1O1O0O2O1O1N2O001N2O1O1O1N101O1N2O1O001N2O1O1N2O001N2O1O1O1N2O2N2M3N2N2N2M3N2N3L3N2N2M3N2N2N2M3N2N2M3N2N2N2Me`Z5"}, "label": "the piece of food on the left - hand side of the plate"}]}
{"id": 257815, "image": "COCO_train2014_000000257815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red bag\"."}], "task": "refering", "answer_template": "The \"red bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 79, 80, 81, 97, 98, 99, 115, 116, 117, 118, 133, 134, 135, 151, 152, 168, 169, 186, 203], "bbox": [0.5929791666666666, 0.15137499999999998, 0.9998958333333333, 0.50303125], "iscrowd": 0, "area": 9000.10005000001, "rle": {"size": [640, 480], "counts": "bWb51mc02N1O2N101O10O10O010O10O010O01O10001N1000001O00001N10001O001O1O001O001O010O1000O0100O10000O100O10O10O10000O100O10O10O10000O100O1000O0100000000000000O10000000O100000O100000O100000O1000O10000000000O10O100000O10000O1000000O10000O10000O1000000O10000O10000O10O10001M3N2N2N3M2N3L3N2N3M2N2M2O1O1O1O3L4M3M3M3L4M3M2M4M2N2N3L3N2O1N3M2N2N3M2N6J7I6J4L3N3L3N0O100O100O1000O101O000000000M4L3N2N3M2N2N3hLl@g0m9"}, "label": "red bag"}]}
{"id": 257815, "image": "COCO_train2014_000000257815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bag on the back of a man on a bike\"."}], "task": "refering", "answer_template": "The \"a red bag on the back of a man on a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 79, 80, 81, 97, 98, 99, 115, 116, 117, 118, 133, 134, 135, 151, 152, 168, 169, 186, 203], "bbox": [0.5929791666666666, 0.15137499999999998, 0.9998958333333333, 0.50303125], "iscrowd": 0, "area": 9000.10005000001, "rle": {"size": [640, 480], "counts": "bWb51mc02N1O2N101O10O10O010O10O010O01O10001N1000001O00001N10001O001O1O001O001O010O1000O0100O10000O100O10O10O10000O100O10O10O10000O100O1000O0100000000000000O10000000O100000O100000O100000O1000O10000000000O10O100000O10000O1000000O10000O10000O1000000O10000O10000O10O10001M3N2N2N3M2N3L3N2N3M2N2M2O1O1O1O3L4M3M3M3L4M3M2M4M2N2N3L3N2O1N3M2N2N3M2N6J7I6J4L3N3L3N0O100O100O1000O101O000000000M4L3N2N3M2N2N3hLl@g0m9"}, "label": "a red bag on the back of a man on a bike"}]}
{"id": 331544, "image": "COCO_train2014_000000331544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop screen in front of the person in the brown jacket furthest to the right\"."}], "task": "refering", "answer_template": "The \"the laptop screen in front of the person in the brown jacket furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 337, 338, 339], "bbox": [0.6134999999999999, 0.7384741784037558, 0.89196875, 0.989037558685446], "iscrowd": 0, "area": 12041.88645, "rle": {"size": [426, 640], "counts": "hiS53W=4K4M4K4M3M4L3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M1O5K4L5K4L5K3M0000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2N2M3N2N2N2N2N2M3N2N2O1O1N2O1O1O1N2O1O1O1N2O100O1O1O100O1O100O1O100O1O100O100O100O100O1000000000000000002N2O1N2N1O2N2OY\\l0"}, "label": "the laptop screen in front of the person in the brown jacket furthest to the right"}]}
{"id": 495387, "image": "COCO_train2014_000000495387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair at the end of the table closest to the wall with the square painting\"."}], "task": "refering", "answer_template": "The \"the chair at the end of the table closest to the wall with the square painting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 180, 181, 182, 195, 196, 197, 198, 210, 211, 225, 240, 255], "bbox": [0.0032318501170960183, 0.51828125, 0.23208430913348946, 0.759140625], "iscrowd": 0, "area": 4626.0614000000005, "rle": {"size": [640, 427], "counts": "kc13mc03M3[]O6m`0Ml^O<Pa0Gh^Ob0Ta0@f^Oi0Ua0ZOc^Oo0Xa0TOa^OU1[a0m0T_OZMS`0l2d_OWM\\`0j1^_O\\Oa`0e0o_OjNR`0V1]@ZNc?g1Z1100A`0L3M3M3M3NmNa_OiN^`0W1h_OdNX`0[1n_O`NR`0`1R@\\Nm?d1W@YNi?f1[@WNe?h1Y101O010O10O01O10O0100O010O10O1000O0100000O010000O01000000O1000000O10000O1000000O10001N3N1O2M3N1Ob0]O4M2N2N2M3N2N00001O001N10001O00001O00Xl]6"}, "label": "the chair at the end of the table closest to the wall with the square painting"}]}
{"id": 495387, "image": "COCO_train2014_000000495387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair closest to the window\"."}], "task": "refering", "answer_template": "The \"the chair closest to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 180, 181, 182, 195, 196, 197, 198, 210, 211, 225, 240, 255], "bbox": [0.0032318501170960183, 0.51828125, 0.23208430913348946, 0.759140625], "iscrowd": 0, "area": 4626.0614000000005, "rle": {"size": [640, 427], "counts": "kc13mc03M3[]O6m`0Ml^O<Pa0Gh^Ob0Ta0@f^Oi0Ua0ZOc^Oo0Xa0TOa^OU1[a0m0T_OZMS`0l2d_OWM\\`0j1^_O\\Oa`0e0o_OjNR`0V1]@ZNc?g1Z1100A`0L3M3M3M3NmNa_OiN^`0W1h_OdNX`0[1n_O`NR`0`1R@\\Nm?d1W@YNi?f1[@WNe?h1Y101O010O10O01O10O0100O010O10O1000O0100000O010000O01000000O1000000O10000O1000000O10001N3N1O2M3N1Ob0]O4M2N2N2M3N2N00001O001N10001O00001O00Xl]6"}, "label": "the chair closest to the window"}]}
{"id": 495387, "image": "COCO_train2014_000000495387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair entirely visible\"."}], "task": "refering", "answer_template": "The \"the chair entirely visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 294, 295, 296, 297, 298, 310, 313, 325], "bbox": [0.5948009367681498, 0.569328125, 0.9456440281030444, 0.961875], "iscrowd": 0, "area": 18813.789950000013, "rle": {"size": [640, 427], "counts": "_Vo43lc06n]OOi?Q1W_OSOf`0P1W_OQOi`0R1S_OoNl`0T1P_OnNPa0U1l^OlNTa0W1g^OjNZa0m11M3N2N2N2KX^OQNia0P231H8N1O2N1O2M3N1N3XOYMT@h2j?^MP@d2o?`Ml_Oc2S`0e01OO10000O101O1O0O2O3M1N2O4K:F5L3mAhKX<^4aCcK^<e4XC^Kh<g4PC]Kn<i4kBXKV=h5100FcBPJ]=o5;N2hN\\B_KY>\\4h0K5N2M3O1O1O010O10001O0000002N1O0O1000000000000000001O000001OO10O1J6@?F;O1N1010O01O001O010O100O10O10O10O001O010O1O1O1O1O1003M9G9G9G4L3M2N3M3MVNbBTM\\=R2lBaL5U1m<V2YDaMe;[2iD]MU;_2YEWMg:e2eETMZ:h2lEWMS:d2TFZMl9b2YF^Mf9^2_F_Mc9]2bF_Ma9]2eF`M[9]2V4L5K6J6J6J6I5L5K5K5K7I9GTP>"}, "label": "the chair entirely visible"}]}
{"id": 495387, "image": "COCO_train2014_000000495387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair to the right of the table when looking at the picture\"."}], "task": "refering", "answer_template": "The \"the chair to the right of the table when looking at the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 249, 250, 251, 252, 253, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 294, 295, 296, 297, 298, 310, 313, 325], "bbox": [0.5948009367681498, 0.569328125, 0.9456440281030444, 0.961875], "iscrowd": 0, "area": 18813.789950000013, "rle": {"size": [640, 427], "counts": "_Vo43lc06n]OOi?Q1W_OSOf`0P1W_OQOi`0R1S_OoNl`0T1P_OnNPa0U1l^OlNTa0W1g^OjNZa0m11M3N2N2N2KX^OQNia0P231H8N1O2N1O2M3N1N3XOYMT@h2j?^MP@d2o?`Ml_Oc2S`0e01OO10000O101O1O0O2O3M1N2O4K:F5L3mAhKX<^4aCcK^<e4XC^Kh<g4PC]Kn<i4kBXKV=h5100FcBPJ]=o5;N2hN\\B_KY>\\4h0K5N2M3O1O1O010O10001O0000002N1O0O1000000000000000001O000001OO10O1J6@?F;O1N1010O01O001O010O100O10O10O10O001O010O1O1O1O1O1003M9G9G9G4L3M2N3M3MVNbBTM\\=R2lBaL5U1m<V2YDaMe;[2iD]MU;_2YEWMg:e2eETMZ:h2lEWMS:d2TFZMl9b2YF^Mf9^2_F_Mc9]2bF_Ma9]2eF`M[9]2V4L5K6J6J6J6I5L5K5K5K7I9GTP>"}, "label": "the chair to the right of the table when looking at the picture"}]}
{"id": 388894, "image": "COCO_train2014_000000388894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man whose reflection is showing in the mirror\"."}], "task": "refering", "answer_template": "The \"a man whose reflection is showing in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 293, 294, 295, 316, 317, 318, 319, 340, 341, 342], "bbox": [0.7189687499999999, 0.0001405152224824356, 0.999984375, 1.0], "iscrowd": 0, "area": 47738.080449999994, "rle": {"size": [427, 640], "counts": "dko54R=<D=C<D<D<C9H8H7H9L3N0eKQMZNP3Q1QMTK5h3k2R1VMmJ7m3d2T1nMiNS2T1RNiNo1U1UNhNl1k0fLjKc1W3h1m0gLjKe1V3e1n0hLiKh1V3a1o0iLiKk1T3]1R1iLhKn1S3Y1T1TOiNm0U1WOhNi0W1NSN3k11RNOm14QNMn15PNKo19nMHP2;nMEQ2>mMCQ2`0mM@S2b0kM_OS2d0kM\\OT2g0jMZOT2i0jMWOU2l0iMUOV2l0iMTOV2o0hMROV2Q1hMoNW2T1gMmNW2V1gMjNY2X1eMiNY2Z1eMgN=WL<U5UOdN=ZL<T5VOcN<ZL=V5UO`N<]L>U5TO_N<^L>V5TO]N;`L?V5TOZN<aL?X5SO_N4\\Lg0X5SO`N0[Ll0W5RO_N0[Lm0X5RO^NN]Ln0X5RO\\NN^Ln0Y5RO[NL_LP1Y5ROYNL_LQ1[5QOWNKaLS1Z5POZOn0i0POWOo0l0oNUOo0l0QOUOm0l0SOUOk0l0UOTOj0m0VOTOh0m0XOTOf0m0ZOTOd0m0\\OSOc0n0]OSOa0n0_OSO`0m0@SO`0m0@ROb0m0^OROc0n0]OROc0n0]OQOd0o0\\OPOe0P1[OoNf0Q1ZOnNg0R1YOmNi0R1WOmNj0S1VOlNk0T1UOkNl0\\NYOcNKQ3l0[N^O`NFT3m0[NA^NBV3n0[NE[N]OY3o0ZNJYNWO\\3Q1YNLXNSO^3R1YN0UNnNa3S1YN4RNiNd3T1XN8QNdNf3U1XN<nM_Ni3V1XN>mM\\Nk3V1VNb0mMXNl3X1UNb0nMVNl3Y1TNe0nMRNo3X1RNh0nMPN]4k0cMY1nMlM_4k0aM\\1g3dNXL_1g3`NWLc1h3]NVLg1h3ZNULi1j3WNULl1j3SNTLQ2j3oMTLT2k3lMSLX2l3gMSL[2l3eMRL_2l3aMRLb2n3^MPLe2o3ZMoKi2Q4VMmKn2R4QMmIAZ1a3i4mLlIDX1c3k4hLlIHQ1g3S5`LlIKi0m3Z5XLkINf0n3_5SLkI0c0Q4a5nKkI3b0Q4b5lKkI3c0T4a5hKlI4a0W4c5dKkI6a0Y4c5`KlI6`0^4c5[KlI8`0_4d5SL[Jn3f5PLYJR4h5mKWJT4j5UKeI6a0f4k5RKeI7?i4m5nJeI8>k4n5lJdI8>m4o5iJdI9<P5Q6eJdI;:Q5S6bJdI<9S5T6_JdI=7V5]6iJcIX5]6gJbI[5^6dJbI]5^6cJaI^5_6eJ\\I]5d6fJXI[5h6d0000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1Z1cMG"}, "label": "a man whose reflection is showing in the mirror"}]}
{"id": 388894, "image": "COCO_train2014_000000388894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in front of mirror , and a tooth brush in mouth\"."}], "task": "refering", "answer_template": "The \"a man in front of mirror , and a tooth brush in mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 293, 294, 295, 316, 317, 318, 319, 340, 341, 342], "bbox": [0.7189687499999999, 0.0001405152224824356, 0.999984375, 1.0], "iscrowd": 0, "area": 47738.080449999994, "rle": {"size": [427, 640], "counts": "dko54R=<D=C<D<D<C9H8H7H9L3N0eKQMZNP3Q1QMTK5h3k2R1VMmJ7m3d2T1nMiNS2T1RNiNo1U1UNhNl1k0fLjKc1W3h1m0gLjKe1V3e1n0hLiKh1V3a1o0iLiKk1T3]1R1iLhKn1S3Y1T1TOiNm0U1WOhNi0W1NSN3k11RNOm14QNMn15PNKo19nMHP2;nMEQ2>mMCQ2`0mM@S2b0kM_OS2d0kM\\OT2g0jMZOT2i0jMWOU2l0iMUOV2l0iMTOV2o0hMROV2Q1hMoNW2T1gMmNW2V1gMjNY2X1eMiNY2Z1eMgN=WL<U5UOdN=ZL<T5VOcN<ZL=V5UO`N<]L>U5TO_N<^L>V5TO]N;`L?V5TOZN<aL?X5SO_N4\\Lg0X5SO`N0[Ll0W5RO_N0[Lm0X5RO^NN]Ln0X5RO\\NN^Ln0Y5RO[NL_LP1Y5ROYNL_LQ1[5QOWNKaLS1Z5POZOn0i0POWOo0l0oNUOo0l0QOUOm0l0SOUOk0l0UOTOj0m0VOTOh0m0XOTOf0m0ZOTOd0m0\\OSOc0n0]OSOa0n0_OSO`0m0@SO`0m0@ROb0m0^OROc0n0]OROc0n0]OQOd0o0\\OPOe0P1[OoNf0Q1ZOnNg0R1YOmNi0R1WOmNj0S1VOlNk0T1UOkNl0\\NYOcNKQ3l0[N^O`NFT3m0[NA^NBV3n0[NE[N]OY3o0ZNJYNWO\\3Q1YNLXNSO^3R1YN0UNnNa3S1YN4RNiNd3T1XN8QNdNf3U1XN<nM_Ni3V1XN>mM\\Nk3V1VNb0mMXNl3X1UNb0nMVNl3Y1TNe0nMRNo3X1RNh0nMPN]4k0cMY1nMlM_4k0aM\\1g3dNXL_1g3`NWLc1h3]NVLg1h3ZNULi1j3WNULl1j3SNTLQ2j3oMTLT2k3lMSLX2l3gMSL[2l3eMRL_2l3aMRLb2n3^MPLe2o3ZMoKi2Q4VMmKn2R4QMmIAZ1a3i4mLlIDX1c3k4hLlIHQ1g3S5`LlIKi0m3Z5XLkINf0n3_5SLkI0c0Q4a5nKkI3b0Q4b5lKkI3c0T4a5hKlI4a0W4c5dKkI6a0Y4c5`KlI6`0^4c5[KlI8`0_4d5SL[Jn3f5PLYJR4h5mKWJT4j5UKeI6a0f4k5RKeI7?i4m5nJeI8>k4n5lJdI8>m4o5iJdI9<P5Q6eJdI;:Q5S6bJdI<9S5T6_JdI=7V5]6iJcIX5]6gJbI[5^6dJbI]5^6cJaI^5_6eJ\\I]5d6fJXI[5h6d0000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1Z1cMG"}, "label": "a man in front of mirror , and a tooth brush in mouth"}]}
{"id": 388894, "image": "COCO_train2014_000000388894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s reflection in the mirror\"."}], "task": "refering", "answer_template": "The \"a man ' s reflection in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.218890625, 0.40156908665105384, 0.5577343750000001, 0.7948243559718969], "iscrowd": 0, "area": 18899.547749999994, "rle": {"size": [427, 640], "counts": "^fj13X=4[LOgI4T69bIG\\6e0YI\\Od6Q1RIoNk6^1jHcNT7c1hH]NU7g1iHZNT7j1kHVNS7l1lHUNQ7o1nHQNo6S2oHnMo6U2PIkMm6Y2QIhMl6\\2SIdMk6^2TIcMl6^2SIbMl6`2RIaMn6`2QI`Mo6a2oH`MP7b2nH_MR7a2nH_MQ7c2mH^MS7c2lH]MT7d2jH]MU7e2jH[MV7f2hH[MW7f2iHZMW7g2gHZMY7g2fHYMY7i2eHXM[7i2dHWM[7k2cHVM]7j2cHVM]7k2aHVM^7l2aHTM_7m2_HTM`7n2_HRMb7n2\\HSMd7m2\\HSMd7n2ZHSMf7n2XHSMi7m2UHTMk7m2RHUMn7l2oGVMQ8j2mGXMT8o300001O0000001O0000001O0000001O0000001O00001O0000001O001O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O3M3M4L3M3M3M4L3M3M4L2N2N2N2N3M2N2N2N2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O0000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O00001O1O001O1O1O001O1O1O001O1O1O001O1O1O1O3M3M3M3M3Mgee3"}, "label": "a man ' s reflection in the mirror"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white umbrella is next to a red and white umbrella\"."}], "task": "refering", "answer_template": "The \"a white umbrella is next to a red and white umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 48, 49, 50, 51, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84], "bbox": [0.014553571428571428, 0.22425595238095236, 0.2722098214285714, 0.49642857142857144], "iscrowd": 0, "area": 7715.047199999999, "rle": {"size": [336, 448], "counts": "i\\23[:5L4L3L5L4L3L5L4L4K4M4L4K4M4L4K3K5K401O1O001O001O1O001O001O1O010O00001O00010O0000001O01O01O000O2O00001O000O2O0001O0O2O00001O001O001N10001O001N101O000O2O001N10001N101O001O00001O001O1O2N1O2N2N1O2N1O2N2N1N3M3M2M4M2N3M3L3L5K4K6K5K4LRkZ3"}, "label": "a white umbrella is next to a red and white umbrella"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white , opened umbrella with designs sitting on a table\"."}], "task": "refering", "answer_template": "The \"white , opened umbrella with designs sitting on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 48, 49, 50, 51, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84], "bbox": [0.014553571428571428, 0.22425595238095236, 0.2722098214285714, 0.49642857142857144], "iscrowd": 0, "area": 7715.047199999999, "rle": {"size": [336, 448], "counts": "i\\23[:5L4L3L5L4L3L5L4L4K4M4L4K4M4L4K3K5K401O1O001O001O1O001O001O1O010O00001O00010O0000001O01O01O000O2O00001O000O2O0001O0O2O00001O001O001N10001O001N101O000O2O001N10001N101O001O00001O001O1O2N1O2N2N1O2N1O2N2N1N3M3M2M4M2N3M3L3L5K4K6K5K4LRkZ3"}, "label": "white , opened umbrella with designs sitting on a table"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black sweatshirt with a girl and lots of tables with colorful umbrellas\"."}], "task": "refering", "answer_template": "The \"a man in a black sweatshirt with a girl and lots of tables with colorful umbrellas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 45, 46, 47, 61, 62, 63, 77, 78, 79, 93, 94, 95, 109, 110, 111, 125, 126, 127, 141, 142, 143], "bbox": [0.8238169642857143, 0.07851190476190475, 1.0, 0.7407142857142857], "iscrowd": 0, "area": 11177.073449999998, "rle": {"size": [336, 448], "counts": "`Yi33Y:5K6I6ZFGo8n0M4K5L4L3L5nNXNXIl1P6oMkI<Ni1T6oMiI`0Je1o4lMoK3SOc0Hb1U5QNdKOYOR3R5XMZKZ3f4PMoJQ3R5V10O1O100O1O1O1M3M2O2M3M3M4L3M3M4L3MAnKgJo3\\5]LXJa3j5?jKgK]2U4aMRL]2m3^M[L`2d3[MdLd2Z3XMlLg2S3TMUMj2k2PM]Mn2T5O0O2N100O2O010O00010O010O0001N1O2N1lIWL`5_4M3M3M3QNdKeM_4]1mLRNU3d1dMkM]2T2P3O2N1O1O2N1000001OoM"}, "label": "a man in a black sweatshirt with a girl and lots of tables with colorful umbrellas"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing position and holding the our two hand blue color shirt and jeans pant\"."}], "task": "refering", "answer_template": "The \"a man standing position and holding the our two hand blue color shirt and jeans pant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 45, 46, 47, 61, 62, 63, 77, 78, 79, 93, 94, 95, 109, 110, 111, 125, 126, 127, 141, 142, 143], "bbox": [0.8238169642857143, 0.07851190476190475, 1.0, 0.7407142857142857], "iscrowd": 0, "area": 11177.073449999998, "rle": {"size": [336, 448], "counts": "`Yi33Y:5K6I6ZFGo8n0M4K5L4L3L5nNXNXIl1P6oMkI<Ni1T6oMiI`0Je1o4lMoK3SOc0Hb1U5QNdKOYOR3R5XMZKZ3f4PMoJQ3R5V10O1O100O1O1O1M3M2O2M3M3M4L3M3M4L3MAnKgJo3\\5]LXJa3j5?jKgK]2U4aMRL]2m3^M[L`2d3[MdLd2Z3XMlLg2S3TMUMj2k2PM]Mn2T5O0O2N100O2O010O00010O010O0001N1O2N1lIWL`5_4M3M3M3QNdKeM_4]1mLRNU3d1dMkM]2T2P3O2N1O1O2N1000001OoM"}, "label": "a man standing position and holding the our two hand blue color shirt and jeans pant"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"multi color umbrella sitting the chair\"."}], "task": "refering", "answer_template": "The \"multi color umbrella sitting the chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 98, 99, 100, 101, 102, 114, 115, 116, 117, 118, 132, 133], "bbox": [0.1331473214285714, 0.47654761904761905, 0.41122767857142856, 0.6922916666666667], "iscrowd": 0, "area": 6336.7736, "rle": {"size": [336, 448], "counts": "mkc03[:4K4O1CFaF=Y9E_Ff0_97N1O2N2O001O1O001O1O2M2O1O2N1O2N1O1O2M10001O000000001N100000001O000O10000000000000001O0000000000010O000010O000001O01O01O00010O0000010O00001O00001O00001O0000010O000001O0000001O0O101O00001O1N101O1O1O0QO`GNa80aGNa81`GMb81`GMa82bGL_82cGL_82cGL^83dGK^84dGI^85eGI[87gGF[88P1O1O0O2O1N1O2NV^f2"}, "label": "multi color umbrella sitting the chair"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella with red and yellow segments , as well as green stripes , resting over a chair\"."}], "task": "refering", "answer_template": "The \"an umbrella with red and yellow segments , as well as green stripes , resting over a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 98, 99, 100, 101, 102, 114, 115, 116, 117, 118, 132, 133], "bbox": [0.1331473214285714, 0.47654761904761905, 0.41122767857142856, 0.6922916666666667], "iscrowd": 0, "area": 6336.7736, "rle": {"size": [336, 448], "counts": "mkc03[:4K4O1CFaF=Y9E_Ff0_97N1O2N2O001O1O001O1O2M2O1O2N1O2N1O1O2M10001O000000001N100000001O000O10000000000000001O0000000000010O000010O000001O01O01O00010O0000010O00001O00001O00001O0000010O000001O0000001O0O101O00001O1N101O1O1O0QO`GNa80aGNa81`GMb81`GMa82bGL_82cGL_82cGL^83dGK^84dGI^85eGI[87gGF[88P1O1O0O2O1N1O2NV^f2"}, "label": "an umbrella with red and yellow segments , as well as green stripes , resting over a chair"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman stands holding paper\"."}], "task": "refering", "answer_template": "The \"a woman stands holding paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 39, 40, 41, 55, 56, 57, 58, 72, 73, 74, 88, 89, 90, 91, 104, 105, 106, 107, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 152, 153, 154, 155, 168, 169, 170, 171, 184, 185, 186], "bbox": [0.4803348214285714, 0.14119047619047617, 0.7365178571428571, 0.9883928571428572], "iscrowd": 0, "area": 20748.632599999997, "rle": {"size": [336, 448], "counts": "bcV25g4Lk0e0eN^OW1f0eN]OW1h0eN\\OW1g0fN\\OV1g0iN[OT1h0iNYOV1j0gNVOY1m0eNRO\\1P1dKcNW2>T2R1`KiNU25[2U1ZKQOT2Jb2W1TKYOS2@i2a2PM_MP3h2jLWMW3P3WLZMi3g40O001OkMcJXO]5i0eJTO[5l0gJSOY5l0jJQOV5o0mJPOR5o0QKoNS5l0oJSOa5<aJCg54[JLk5KXJ4l5FVJ:n5ASJ?Q6[OQJf0R6TOoIm0j700001O0O10001O0O102N2N2N2M3N2N23LO2N2N2N1N3N2N1O2N1O2N2N1O2M2O2N1O2N2N1O2N1O2M3N1O2N1O2N2N1O2M2O2N2N1O2N1O2N2J5I8I6I8I7H8I7H8I7I7H8I7H8I7H8I7I7H8I7I8G8I7I7H8I7I7IaaV1"}, "label": "a woman stands holding paper"}]}
{"id": 315168, "image": "COCO_train2014_000000315168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 39, 40, 41, 55, 56, 57, 58, 72, 73, 74, 88, 89, 90, 91, 104, 105, 106, 107, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 152, 153, 154, 155, 168, 169, 170, 171, 184, 185, 186], "bbox": [0.4803348214285714, 0.14119047619047617, 0.7365178571428571, 0.9883928571428572], "iscrowd": 0, "area": 20748.632599999997, "rle": {"size": [336, 448], "counts": "bcV25g4Lk0e0eN^OW1f0eN]OW1h0eN\\OW1g0fN\\OV1g0iN[OT1h0iNYOV1j0gNVOY1m0eNRO\\1P1dKcNW2>T2R1`KiNU25[2U1ZKQOT2Jb2W1TKYOS2@i2a2PM_MP3h2jLWMW3P3WLZMi3g40O001OkMcJXO]5i0eJTO[5l0gJSOY5l0jJQOV5o0mJPOR5o0QKoNS5l0oJSOa5<aJCg54[JLk5KXJ4l5FVJ:n5ASJ?Q6[OQJf0R6TOoIm0j700001O0O10001O0O102N2N2N2M3N2N23LO2N2N2N1N3N2N1O2N1O2N2N1O2M2O2N1O2N2N1O2N1O2M3N1O2N1O2N2N1O2M2O2N2N1O2N1O2N2J5I8I6I8I7H8I7H8I7I7H8I7H8I7H8I7I7H8I7I8G8I7I7H8I7I7IaaV1"}, "label": "a woman in a black shirt"}]}
{"id": 225060, "image": "COCO_train2014_000000225060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"screen with three windows up\"."}], "task": "refering", "answer_template": "The \"screen with three windows up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 228, 229], "bbox": [0.7349218750000001, 0.2005625, 1.0, 0.5868125000000001], "iscrowd": 0, "area": 22200.310849999998, "rle": {"size": [480, 640], "counts": "_ol65f>6K4K5K6K4K6J5L4K6J5L5J5K6J5L4K6J5L5J5K5L5J5K6L30010O001O001O01O01O001O01O01O001O00010O001O0010O0001O001O01O01O001O01O01O001O0010O0001O001O01O01O001O01O01O001O0010O0001O0010O0001O001O01O01O001O010O00001O0010O0001O001O01O01O001O010O00001O0010O0001O0010O0001O001O010O00001O010O00001O0010O0001O0010O01O00001O010O00001O0010O0001O0010O01O000VL"}, "label": "screen with three windows up"}]}
{"id": 225060, "image": "COCO_train2014_000000225060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coumputer screen\"."}], "task": "refering", "answer_template": "The \"a coumputer screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 228, 229], "bbox": [0.7349218750000001, 0.2005625, 1.0, 0.5868125000000001], "iscrowd": 0, "area": 22200.310849999998, "rle": {"size": [480, 640], "counts": "_ol65f>6K4K5K6K4K6J5L4K6J5L5J5K6J5L4K6J5L5J5K5L5J5K6L30010O001O001O01O01O001O01O01O001O00010O001O0010O0001O001O01O01O001O01O01O001O0010O0001O001O01O01O001O01O01O001O0010O0001O0010O0001O001O01O01O001O010O00001O0010O0001O001O01O01O001O010O00001O0010O0001O0010O0001O001O010O00001O010O00001O0010O0001O0010O01O00001O010O00001O0010O0001O0010O01O000VL"}, "label": "a coumputer screen"}]}
{"id": 184106, "image": "COCO_train2014_000000184106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very large pizza , half of which has pepperoni and cheese and the other half has many different toppings\"."}], "task": "refering", "answer_template": "The \"a very large pizza , half of which has pepperoni and cheese and the other half has many different toppings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 108, 109, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 173, 174, 175, 176, 177, 178, 179, 180, 181, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.010609375, 0.23939583333333334, 0.9587187499999998, 0.9893958333333333], "iscrowd": 0, "area": 150676.63005, "rle": {"size": [480, 640], "counts": "na33k>2N3L4M3L4M3M2M4M3M3YDPOY9T1fFmNW9U1iFlNS9X1lFiNQ9Z1nFgNo8\\1PGeNl8_1SGbNj8`1VGaNf8c1XG_Ne8d1ZG]Nc8f1\\G[Na8h1^GXN`8j1`GWN^8k1aGVN\\8m1cGTNZ8o1eGRNY8P2eGRNX8P2hGQNU8R2jGoMT8S2kGnMR8U2mGlMP8W2oGjMo7W2QHjMl7Y2SHhMk7Z2SHhMj7Z2VHgMg7[2YHfMe7[2[HfMb7\\2^HdM`7]2aHdM]7]2cHdMZ7^2fHcMW7_2hHcMV7]2kHdMR7^2nHcMP7^2PIcMn6^2RIcMl6^2TIcMj6]2WIdMh6\\2XIeMf6\\2YIeMf6\\2ZIeMd6[2]IfMa6[2_IfM_6[2aIfM]6[2bIgM\\6Z2dIgMZ6Y2fIiMX6X2hIhMW6Y2hIiMV6X2jIiMT6W2mIjMQ6W2nIkMP6V2PJkMn5V2QJlMn5T2RJmMl5S2TJoMj5R2VJnMj5R2UJPNi5Q2WJPNg5P2YJRNf5n1ZJSNd5n1[JTNc5m1]JTNa5m1^JUNa5j1`JVN_5k1aJVN]5k1bJWN]5i1cJXN[5h1eJZNY5g1gJZNW5g1hJ[NW5e1iJ\\NU5e1jJ\\NV5c1kJ^NS5c1lJ_NS5a1mJ_NS5a1lJaNS5^1nJcNP5^1PKbNP5^1oJdNP5\\1PKeNn4\\1QKeNo4Z1RKgNm4Y1RKiNm4W1SKiNl4X1SKjNl4U1UKlNj4T1UKmNj4T1VKmNi4S1VKnNj4R1VKoNi4P1WKROg4o0YKQOg4o0YKROf4n0YKTOe4l0\\KTOd4l0[KVOd4j0\\KWOc4i0\\KXOc4h0^KYOa4g0^K[O`4f0`KZO`4f0_K\\O_4e0aK[O^4e0bK]O\\4d0dK]O[4c0dK^O[4c0eK^OY4b0hK^OX4b0gK@W4a0iK_OV4b0iK@U4a0kK@T4?lKBS4?mKBQ4?nKBQ4?oKBP4=PLDo3=QLDn3<QLEn3<RLEm3;RLGm38TLHk39ULHj38ULIj38VLIi36WLKh36XLKg35XLLh34XLMf34YLNf31[LOd32[LOe31[LOd32[LOe30\\L0d30[L1d30\\L0d30\\L0c31\\L0d3O]L1b30]L1c3O]L1c3O\\L2c3N^L2b3N]L3b3N^L2b3N]L3b3N^L2b3M^L4b3L^L4a3M_L3a3M^L4a3L`L4`3L_L5`3L`L4`3L_L5a3K_L5`3K`L6`3J`L6_3K`L6`3J`L6_3JaL7_3IaL7_3I`L8_3IaL7_3IaL7^3IbL8^3HbL8]3IbL8^3HbL8^3GbL:]3GcL9]3GbL:]3GcL9]3GbL:]3FdL:\\3FcL;\\3FdL:\\3FcL;]3DdL<[3EeL;[3EdL<[3EeL;[3EdL<[3DfL<Z3DeL=[3CeL=Z3DeL=[3BfL>Y3CfL>Z3BfL>[3AdL`0]3_OcLa0]3^OcLc0]3]OcLc0]3]OcLc0]3]ObLd0^3[OcLe0^3ZOaLg0`3XO`Lh0`3XO_Li0b3VO^Lj0b3UO^Ll0c3hJQLj3<^1d3eJSLl38`1e3cJULl36a1f3aJVLl34d1f3_JXLl32e1g3\\JZLn3Ng1i3YJ[Lo3Lh1i3XJ^Ln3Ij1j3VJ_Ln3Gm1a4QN_Ko1b4PN]KQ2d4nM\\KR2f4kMZKV2g4iMYKW2i4gMVKZ2k4eMUK[2m4cMRK^2o4`MRK`2P5^MoJc2R5\\MnJd2T5ZMkJg2V5WMkJi2V5VMjJj2X5TMgJm2Z5RMfJn2\\5PMcJQ3^5mLcJS3_5kL`JV3a5iL_JW3c5gL\\JZ3e5dL\\J\\3f5bLYJ_3h5`LXJ`3j5^LUJc3l5\\LTJd3n5YLSJg3n5XLQJi3Q6ULoIk3R6TLmIm3U6PLlIP4U6oKjIR4X6lKhIT4Y6kKfIV4\\6gKeIY4\\6fKcI[4_6cKaI^4_6aK`I`4b6^K^Ib4c6\\K]Ie4e6YK[Ig4f6XKZIh4g6h12\\IWIm4j6d12N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2NO100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O101N1O100O100O1O100O100O1O100O100O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O0001O01O01O2N2N101N2N2N101N2N2N1O2O1N2N1O2N2O1N1O2N2O0O2N2N2N101N2N2N101N2N2N1O2O1N1O2N2N101N2N1O2O1N2N1O2N2O0O2N2N101N2N2N1O2O3L2N1O2N2O1N2N2N2O1N2N2N2N2O1M3N2N2N2N2N2N2N1N3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N1O2N3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3L4M2N3M3M3M3M3M2N5K5K6J5K6I7I6I8I6J7I6J7H8I6Jkk;"}, "label": "a very large pizza , half of which has pepperoni and cheese and the other half has many different toppings"}]}
{"id": 184106, "image": "COCO_train2014_000000184106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giant pizza with many different toppings\"."}], "task": "refering", "answer_template": "The \"a giant pizza with many different toppings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 108, 109, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 173, 174, 175, 176, 177, 178, 179, 180, 181, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.010609375, 0.23939583333333334, 0.9587187499999998, 0.9893958333333333], "iscrowd": 0, "area": 150676.63005, "rle": {"size": [480, 640], "counts": "na33k>2N3L4M3L4M3M2M4M3M3YDPOY9T1fFmNW9U1iFlNS9X1lFiNQ9Z1nFgNo8\\1PGeNl8_1SGbNj8`1VGaNf8c1XG_Ne8d1ZG]Nc8f1\\G[Na8h1^GXN`8j1`GWN^8k1aGVN\\8m1cGTNZ8o1eGRNY8P2eGRNX8P2hGQNU8R2jGoMT8S2kGnMR8U2mGlMP8W2oGjMo7W2QHjMl7Y2SHhMk7Z2SHhMj7Z2VHgMg7[2YHfMe7[2[HfMb7\\2^HdM`7]2aHdM]7]2cHdMZ7^2fHcMW7_2hHcMV7]2kHdMR7^2nHcMP7^2PIcMn6^2RIcMl6^2TIcMj6]2WIdMh6\\2XIeMf6\\2YIeMf6\\2ZIeMd6[2]IfMa6[2_IfM_6[2aIfM]6[2bIgM\\6Z2dIgMZ6Y2fIiMX6X2hIhMW6Y2hIiMV6X2jIiMT6W2mIjMQ6W2nIkMP6V2PJkMn5V2QJlMn5T2RJmMl5S2TJoMj5R2VJnMj5R2UJPNi5Q2WJPNg5P2YJRNf5n1ZJSNd5n1[JTNc5m1]JTNa5m1^JUNa5j1`JVN_5k1aJVN]5k1bJWN]5i1cJXN[5h1eJZNY5g1gJZNW5g1hJ[NW5e1iJ\\NU5e1jJ\\NV5c1kJ^NS5c1lJ_NS5a1mJ_NS5a1lJaNS5^1nJcNP5^1PKbNP5^1oJdNP5\\1PKeNn4\\1QKeNo4Z1RKgNm4Y1RKiNm4W1SKiNl4X1SKjNl4U1UKlNj4T1UKmNj4T1VKmNi4S1VKnNj4R1VKoNi4P1WKROg4o0YKQOg4o0YKROf4n0YKTOe4l0\\KTOd4l0[KVOd4j0\\KWOc4i0\\KXOc4h0^KYOa4g0^K[O`4f0`KZO`4f0_K\\O_4e0aK[O^4e0bK]O\\4d0dK]O[4c0dK^O[4c0eK^OY4b0hK^OX4b0gK@W4a0iK_OV4b0iK@U4a0kK@T4?lKBS4?mKBQ4?nKBQ4?oKBP4=PLDo3=QLDn3<QLEn3<RLEm3;RLGm38TLHk39ULHj38ULIj38VLIi36WLKh36XLKg35XLLh34XLMf34YLNf31[LOd32[LOe31[LOd32[LOe30\\L0d30[L1d30\\L0d30\\L0c31\\L0d3O]L1b30]L1c3O]L1c3O\\L2c3N^L2b3N]L3b3N^L2b3N]L3b3N^L2b3M^L4b3L^L4a3M_L3a3M^L4a3L`L4`3L_L5`3L`L4`3L_L5a3K_L5`3K`L6`3J`L6_3K`L6`3J`L6_3JaL7_3IaL7_3I`L8_3IaL7_3IaL7^3IbL8^3HbL8]3IbL8^3HbL8^3GbL:]3GcL9]3GbL:]3GcL9]3GbL:]3FdL:\\3FcL;\\3FdL:\\3FcL;]3DdL<[3EeL;[3EdL<[3EeL;[3EdL<[3DfL<Z3DeL=[3CeL=Z3DeL=[3BfL>Y3CfL>Z3BfL>[3AdL`0]3_OcLa0]3^OcLc0]3]OcLc0]3]OcLc0]3]ObLd0^3[OcLe0^3ZOaLg0`3XO`Lh0`3XO_Li0b3VO^Lj0b3UO^Ll0c3hJQLj3<^1d3eJSLl38`1e3cJULl36a1f3aJVLl34d1f3_JXLl32e1g3\\JZLn3Ng1i3YJ[Lo3Lh1i3XJ^Ln3Ij1j3VJ_Ln3Gm1a4QN_Ko1b4PN]KQ2d4nM\\KR2f4kMZKV2g4iMYKW2i4gMVKZ2k4eMUK[2m4cMRK^2o4`MRK`2P5^MoJc2R5\\MnJd2T5ZMkJg2V5WMkJi2V5VMjJj2X5TMgJm2Z5RMfJn2\\5PMcJQ3^5mLcJS3_5kL`JV3a5iL_JW3c5gL\\JZ3e5dL\\J\\3f5bLYJ_3h5`LXJ`3j5^LUJc3l5\\LTJd3n5YLSJg3n5XLQJi3Q6ULoIk3R6TLmIm3U6PLlIP4U6oKjIR4X6lKhIT4Y6kKfIV4\\6gKeIY4\\6fKcI[4_6cKaI^4_6aK`I`4b6^K^Ib4c6\\K]Ie4e6YK[Ig4f6XKZIh4g6h12\\IWIm4j6d12N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2NO100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O101N1O100O100O1O100O100O1O100O100O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O0001O01O01O2N2N101N2N2N101N2N2N1O2O1N2N1O2N2O1N1O2N2O0O2N2N2N101N2N2N101N2N2N1O2O1N1O2N2N101N2N1O2O1N2N1O2N2O0O2N2N101N2N2N1O2O3L2N1O2N2O1N2N2N2O1N2N2N2N2O1M3N2N2N2N2N2N2N1N3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N1O2N3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3L4M2N3M3M3M3M3M2N5K5K6J5K6I7I6I8I6J7I6J7H8I6Jkk;"}, "label": "a giant pizza with many different toppings"}]}
{"id": 184106, "image": "COCO_train2014_000000184106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big slice of pizza in the hands of a person\"."}], "task": "refering", "answer_template": "The \"a big slice of pizza in the hands of a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 104, 105, 106, 107, 126, 127, 128, 149], "bbox": [0.485953125, 0.013833333333333333, 0.76775, 0.36627083333333327], "iscrowd": 0, "area": 10455.990349999995, "rle": {"size": [480, 640], "counts": "[na42l>3N3M001O001O001O1O0010O01O001O001O1O001O001O0O2O0O2N2O0O2O0O2O0O2O0O101N101N1O2O0O2O0O2O00010O001O0N3N1O2N1O2N1O2N1O1O2N1O2N1O2N1N2O1O2N1N2O1O2N1N2O2N1O2M3N1O2N2N1O2N1O2N2N1O2N2N1O2O1N101O1O0O2O1O001N101O1O010000O2O0O1O101N1O100Oe0[O1O2O0O2N100O2N100O2N101N2N2N2O1N1O2O1N1O010O1O0010J5M33N1N2N2O1N2N3M2N2O1N2N2N3M2N2O1N2N2N2N2N2N2O0O2N2N2N2N2N2O1N3M2N2N3M2O2MlhU2"}, "label": "a big slice of pizza in the hands of a person"}]}
{"id": 184106, "image": "COCO_train2014_000000184106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of pizza in the man ' s hand\"."}], "task": "refering", "answer_template": "The \"the piece of pizza in the man ' s hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 104, 105, 106, 107, 126, 127, 128, 149], "bbox": [0.485953125, 0.013833333333333333, 0.76775, 0.36627083333333327], "iscrowd": 0, "area": 10455.990349999995, "rle": {"size": [480, 640], "counts": "[na42l>3N3M001O001O001O1O0010O01O001O001O1O001O001O0O2O0O2N2O0O2O0O2O0O2O0O101N101N1O2O0O2O0O2O00010O001O0N3N1O2N1O2N1O2N1O1O2N1O2N1O2N1N2O1O2N1N2O1O2N1N2O2N1O2M3N1O2N2N1O2N1O2N2N1O2N2N1O2O1N101O1O0O2O1O001N101O1O010000O2O0O1O101N1O100Oe0[O1O2O0O2N100O2N100O2N101N2N2N2O1N1O2O1N1O010O1O0010J5M33N1N2N2O1N2N3M2N2O1N2N2N3M2N2O1N2N2N2N2N2N2O0O2N2N2N2N2N2O1N3M2N2N3M2O2MlhU2"}, "label": "the piece of pizza in the man ' s hand"}]}
{"id": 347948, "image": "COCO_train2014_000000347948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a london style bus behind another one\"."}], "task": "refering", "answer_template": "The \"a london style bus behind another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 232, 233, 256], "bbox": [0.037828125000000004, 0.2869791666666667, 0.1743125, 0.6625624999999999], "iscrowd": 0, "area": 10779.450850000001, "rle": {"size": [480, 640], "counts": "^];X1h=e1[N00000000O100000mNS10B>00000000N2L4N2O1O1O1O0101O00001O001O1hDiLe:Y3WEiLi:h3O1O002N4L001O002N2N1O2N2N2N2N101N2N2N1O0000000nEXKl9h4TF^KN2\\9`4fFjK[9U4eFkK[9U4dFlK\\9T4bFnK^9j401O0000000001O01O000001@?]Oc0E<F9F:G9G:D;C=D=D<F:F:E?E][g7"}, "label": "a london style bus behind another one"}]}
{"id": 347948, "image": "COCO_train2014_000000347948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus that is following another bus\"."}], "task": "refering", "answer_template": "The \"a bus that is following another bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 232, 233, 256], "bbox": [0.037828125000000004, 0.2869791666666667, 0.1743125, 0.6625624999999999], "iscrowd": 0, "area": 10779.450850000001, "rle": {"size": [480, 640], "counts": "^];X1h=e1[N00000000O100000mNS10B>00000000N2L4N2O1O1O1O0101O00001O001O1hDiLe:Y3WEiLi:h3O1O002N4L001O002N2N1O2N2N2N2N101N2N2N1O0000000nEXKl9h4TF^KN2\\9`4fFjK[9U4eFkK[9U4dFlK\\9T4bFnK^9j401O0000000001O01O000001@?]Oc0E<F9F:G9G:D;C=D=D<F:F:E?E][g7"}, "label": "a bus that is following another bus"}]}
{"id": 347948, "image": "COCO_train2014_000000347948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red double decked bus on a street with another red bus behind\"."}], "task": "refering", "answer_template": "The \"red double decked bus on a street with another red bus behind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332], "bbox": [0.1449375, 0.12808333333333333, 0.756734375, 0.8471875000000001], "iscrowd": 0, "area": 105005.21495, "rle": {"size": [480, 640], "counts": "dm[19Y>?B=B>B>C=B>C=B>B>C>A>K5L4L4K5L4L4L4L5J5L4L4L4L4L4K5N3M2N2N4K7J6J6K5L3L10000O2O00000O1000000O10000O10O1N2O001N2O0O2O1N2O00100000O101O0O1000000O10000O1000000O101O000O10000O1000000O10000O10001O0O1000000O10000O1000000O101O0O1000000O1000000O10000O10001O0O10000O1000000O1000000O101O0O1000000O10000O1000000O10001O0O10000O4M2N3M2M4M2N3L3N3M2N2M4M2N2N0O2O000O1000001N10000O1000001N1000000O101O0O1000000O2O000O1000001NO2N2N2N101N2N2N2N2N2N2N2N2N2O1N2N2TOlJjGV5Y8nJ^GT5e8PKRGR5Q9=2N3M3M2N3M3M00000000000000000000000000001O000001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O01O00000001O00000000001O00000000001O00000000001O00QKeFY4[9dKhF\\4X9bKjF^4V9_KmFa4S9]KoFd4P9YKSGg4m8WKUGi4k8TKXGl4^900001O00000000001O00000000001O00000000001O00000000001O1O8H9E;D<E;E;E;D<E;E;E:F;D3gMQG_NQ9oNcIm0_6kNkIQ1X6nNjIm0Y6SOiIi0Y6WOiIe0Z6ZOhIb0Z6]OhI?[6AgI;\\6DfI8\\6H^40001O000O2O00001O00000O2O2N5KhmX2"}, "label": "red double decked bus on a street with another red bus behind"}]}
{"id": 347948, "image": "COCO_train2014_000000347948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red doubledecker bus in the front\"."}], "task": "refering", "answer_template": "The \"a red doubledecker bus in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332], "bbox": [0.1449375, 0.12808333333333333, 0.756734375, 0.8471875000000001], "iscrowd": 0, "area": 105005.21495, "rle": {"size": [480, 640], "counts": "dm[19Y>?B=B>B>C=B>C=B>B>C>A>K5L4L4K5L4L4L4L5J5L4L4L4L4L4K5N3M2N2N4K7J6J6K5L3L10000O2O00000O1000000O10000O10O1N2O001N2O0O2O1N2O00100000O101O0O1000000O10000O1000000O101O000O10000O1000000O10000O10001O0O1000000O10000O1000000O101O0O1000000O1000000O10000O10001O0O10000O1000000O1000000O101O0O1000000O10000O1000000O10001O0O10000O4M2N3M2M4M2N3L3N3M2N2M4M2N2N0O2O000O1000001N10000O1000001N1000000O101O0O1000000O2O000O1000001NO2N2N2N101N2N2N2N2N2N2N2N2N2O1N2N2TOlJjGV5Y8nJ^GT5e8PKRGR5Q9=2N3M3M2N3M3M00000000000000000000000000001O000001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O00000000001O01O00000001O00000000001O00000000001O00000000001O00QKeFY4[9dKhF\\4X9bKjF^4V9_KmFa4S9]KoFd4P9YKSGg4m8WKUGi4k8TKXGl4^900001O00000000001O00000000001O00000000001O00000000001O1O8H9E;D<E;E;E;D<E;E;E:F;D3gMQG_NQ9oNcIm0_6kNkIQ1X6nNjIm0Y6SOiIi0Y6WOiIe0Z6ZOhIb0Z6]OhI?[6AgI;\\6DfI8\\6H^40001O000O2O00001O00000O2O2N5KhmX2"}, "label": "a red doubledecker bus in the front"}]}
{"id": 192301, "image": "COCO_train2014_000000192301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pile of purple plums\"."}], "task": "refering", "answer_template": "The \"a pile of purple plums\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 261, 266, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.006937500000000001, 0.7350585480093677, 0.745171875, 0.9870023419203747], "iscrowd": 0, "area": 32187.18595, "rle": {"size": [427, 640], "counts": "RR2=n<00000000000000000000000000000000000O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O11O000000000000000000001O0000000000000000001O00000001O000000000001O0000\\Od000O1O1O100O1O100O1O100O1O100O1O1O100O1O100N2N2N2N2000000O10000O1000000O10000O10000O1000000O1001O001O001O1O001O001O1O001O1O001O001O[OVNjEi1U:g0O100O100O1O100O1O100O100O1O100O1O100O1O100O100O10000O1000000O10000O1000000O10000O10000O100000000000000001O0000000000000000001O000000001O1O001O1O1O001O1O001O001O001O001O001O00001O001O001O001O001O00001O001O1O001O001O001O1O001O001O001O001O1O0000O1000000O10000O1000000O1000000O101O00000001O1O001O1O001O1O001O1O1O00O100O1O1O1O1O1O1O1O1O1O1001O00001O00001O00001O00001O00001O00001O001O00001O0000001O0000001O0000001O000000O1M3M3N2M3N20000000000O1000000000000000000O100000000000000001O1O001O001O1O001O1O001O001O1O001O1O001O0000O100O10000O100O10000O100O10000O10000O100O11O1O1O1O1O1O1O1O1O1O1O1O2N2N2N1O2N2N5K6J3M2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N2N2N2N2N2N1OmaS2"}, "label": "a pile of purple plums"}]}
{"id": 192301, "image": "COCO_train2014_000000192301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple plums\"."}], "task": "refering", "answer_template": "The \"purple plums\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 261, 266, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.006937500000000001, 0.7350585480093677, 0.745171875, 0.9870023419203747], "iscrowd": 0, "area": 32187.18595, "rle": {"size": [427, 640], "counts": "RR2=n<00000000000000000000000000000000000O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O11O000000000000000000001O0000000000000000001O00000001O000000000001O0000\\Od000O1O1O100O1O100O1O100O1O100O1O1O100O1O100N2N2N2N2000000O10000O1000000O10000O10000O1000000O1001O001O001O1O001O001O1O001O1O001O001O[OVNjEi1U:g0O100O100O1O100O1O100O100O1O100O1O100O1O100O100O10000O1000000O10000O1000000O10000O10000O100000000000000001O0000000000000000001O000000001O1O001O1O1O001O1O001O001O001O001O001O00001O001O001O001O001O00001O001O1O001O001O001O1O001O001O001O001O1O0000O1000000O10000O1000000O1000000O101O00000001O1O001O1O001O1O001O1O1O00O100O1O1O1O1O1O1O1O1O1O1001O00001O00001O00001O00001O00001O00001O001O00001O0000001O0000001O0000001O000000O1M3M3N2M3N20000000000O1000000000000000000O100000000000000001O1O001O001O1O001O1O001O001O1O001O1O001O0000O100O10000O100O10000O100O10000O10000O100O11O1O1O1O1O1O1O1O1O1O1O1O2N2N2N1O2N2N5K6J3M2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N2N2N2N2N2N1OmaS2"}, "label": "purple plums"}]}
{"id": 192301, "image": "COCO_train2014_000000192301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green apples are in a white bin next to a bin or oranges and plums\"."}], "task": "refering", "answer_template": "The \"green apples are in a white bin next to a bin or oranges and plums\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 253, 254, 255], "bbox": [0.0, 0.19550351288056206, 0.5568906250000001, 0.7662997658079624], "iscrowd": 0, "area": 50981.993, "rle": {"size": [427, 640], "counts": "a4T5U8?A3N100000000001O00000001O000N2M3N2N2N2M3M3M4L3O1000000000000O10001O000001O00001O0000001O000000O2O00000O1O100O1O1O1O1O1O100O1O100O1O100O0100O100O100O2N100O100O100N2N00O100100O1O1O1O11O1O100O1O10000O100O100O10000O010O10O011N2O1N2O1N101O1N2O1N2O1N2O2M2O1O1N3N1O1N2O001N10000O2O000O10001N1000001N10000O2O000O101O0O10001N10000O2O000O101O0O101O0O10001N10001N10001O0O10001O0O101O00001N1000001O0O101O00001O01O01O00010O000001O1O1O2N1O1O1N3N1OM3M2N2O2M2N200000010O0000010O1O1O100O1O1O1O100O1TOVGVMk8c2^G[Mb8e2`GXMa8h2aGVM_8i2eGSM\\8m2fGPM\\8n2gGPMY8P3iGmLX8S3P1O10000O2O0O10001N1000001O0000001O1O00100lM`E[1a:aNcE^1]:^NgEb1Z:ZNiEf1X:VNlEi1U:RNoEn1h:OO101N10[Oe00O10O01000O0100000O1000O01000000O1000000O10000O1000000O1000001N1000000O2O3M2NO001O10O10O1000O10000000O010000000000O10001O1O001N2O1O001O1O1O001O1N2EQYf3"}, "label": "green apples are in a white bin next to a bin or oranges and plums"}]}
{"id": 192301, "image": "COCO_train2014_000000192301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bunch of green apples on a fruit stand also containing plums and clementines\"."}], "task": "refering", "answer_template": "The \"a bunch of green apples on a fruit stand also containing plums and clementines\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 253, 254, 255], "bbox": [0.0, 0.19550351288056206, 0.5568906250000001, 0.7662997658079624], "iscrowd": 0, "area": 50981.993, "rle": {"size": [427, 640], "counts": "a4T5U8?A3N100000000001O00000001O000N2M3N2N2N2M3M3M4L3O1000000000000O10001O000001O00001O0000001O000000O2O00000O1O100O1O1O1O1O1O100O1O100O1O100O0100O100O100O2N100O100O100N2N00O100100O1O1O1O11O1O100O1O10000O100O100O10000O010O10O011N2O1N2O1N101O1N2O1N2O1N2O2M2O1O1N3N1O1N2O001N10000O2O000O10001N1000001N10000O2O000O101O0O10001N10000O2O000O101O0O101O0O10001N10001N10001O0O10001O0O101O00001N1000001O0O101O00001O01O01O00010O000001O1O1O2N1O1O1N3N1OM3M2N2O2M2N200000010O0000010O1O1O100O1O1O1O100O1TOVGVMk8c2^G[Mb8e2`GXMa8h2aGVM_8i2eGSM\\8m2fGPM\\8n2gGPMY8P3iGmLX8S3P1O10000O2O0O10001N1000001O0000001O1O00100lM`E[1a:aNcE^1]:^NgEb1Z:ZNiEf1X:VNlEi1U:RNoEn1h:OO101N10[Oe00O10O01000O0100000O1000O01000000O1000000O10000O1000000O1000001N1000000O2O3M2NO001O10O10O1000O10000000O010000000000O10001O1O001N2O1O001O1O1O001O1N2EQYf3"}, "label": "a bunch of green apples on a fruit stand also containing plums and clementines"}]}
{"id": 421677, "image": "COCO_train2014_000000421677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red kite\"."}], "task": "refering", "answer_template": "The \"red kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 27, 28, 49, 50, 51, 72, 74, 95, 97, 118, 120, 121, 141, 144, 164, 165, 167, 188, 189, 190, 191, 212, 213, 214, 236, 237, 260, 261, 283, 284, 285], "bbox": [0.15006250000000002, 0.0005056179775280899, 0.408875, 0.6977902621722847], "iscrowd": 0, "area": 4351.422750000002, "rle": {"size": [534, 640], "counts": "mVb1`0P`0=C<D<C>CQOnALf=0lBOe<NnCMg;4gDFP;;_E]OZ:c0UFVOe9i0dFUOW9h0PGXOl8f0[GZO_8d0hG]OS8`0UH_Of7>bHCY7<mHDm6=WIDd6<aICZ6>jICP6>UJBe5>`JC^4LVG`0b4DW4T1nKmNQ4m0VLSOl3d0[L\\OQ42ULNZ:1O0O1000001O01O0010O01O1O001O\\HKfN5Y1OdN1[16^NKb1>UNBj1j0kMVOW2V1[MjNj2e1bL[Ng3e1PL[N]4[1`KfNb5:[JFk57RJIQ68kIHY67dII_67^IIe68XIHj69RIGQ7:kHGW7:eHG]7:_HFe7;VHFl7;PHES8;jGFX89fGG]89`GHb87\\GJg85VGKm85PGLQ94mFLV94gFL\\93bFLc93ZFKk95SFIQ:6mEHX:8eEG_:8_EDh:<k2O1O100O00O00001N11O100O00100O1O1O010O000001O0000001mAGe;9YDJe;7XDLh;4VDOi;2TD0k;1SD0n;0oC2R<NlC3T<OiC3W<MgC4Z<LcC6]<KaC6`<J^C8b<G\\C;e<C[C>e<BZCb0d<[O]Cj0`<TO_Co0`<oNaCR1`<kNaCU1a<iN_Co0hNSOV?n0n@oNQ?Q1:O10O10002M2OKc@SO[?m0i@POV?o0810O100O02O2N1OL_@UO_?j0e@TO[?k0i@SOV?k0;01O1N1011O00KV@@f?`0^@^Oa?`0<O1O001N2O1O11000Hi_O2_`0JkhT6"}, "label": "red kite"}]}
{"id": 421677, "image": "COCO_train2014_000000421677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red , grey , white and black kite\"."}], "task": "refering", "answer_template": "The \"a red , grey , white and black kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 27, 28, 49, 50, 51, 72, 74, 95, 97, 118, 120, 121, 141, 144, 164, 165, 167, 188, 189, 190, 191, 212, 213, 214, 236, 237, 260, 261, 283, 284, 285], "bbox": [0.15006250000000002, 0.0005056179775280899, 0.408875, 0.6977902621722847], "iscrowd": 0, "area": 4351.422750000002, "rle": {"size": [534, 640], "counts": "mVb1`0P`0=C<D<C>CQOnALf=0lBOe<NnCMg;4gDFP;;_E]OZ:c0UFVOe9i0dFUOW9h0PGXOl8f0[GZO_8d0hG]OS8`0UH_Of7>bHCY7<mHDm6=WIDd6<aICZ6>jICP6>UJBe5>`JC^4LVG`0b4DW4T1nKmNQ4m0VLSOl3d0[L\\OQ42ULNZ:1O0O1000001O01O0010O01O1O001O\\HKfN5Y1OdN1[16^NKb1>UNBj1j0kMVOW2V1[MjNj2e1bL[Ng3e1PL[N]4[1`KfNb5:[JFk57RJIQ68kIHY67dII_67^IIe68XIHj69RIGQ7:kHGW7:eHG]7:_HFe7;VHFl7;PHES8;jGFX89fGG]89`GHb87\\GJg85VGKm85PGLQ94mFLV94gFL\\93bFLc93ZFKk95SFIQ:6mEHX:8eEG_:8_EDh:<k2O1O100O00O00001N11O100O00100O1O1O010O000001O0000001mAGe;9YDJe;7XDLh;4VDOi;2TD0k;1SD0n;0oC2R<NlC3T<OiC3W<MgC4Z<LcC6]<KaC6`<J^C8b<G\\C;e<C[C>e<BZCb0d<[O]Cj0`<TO_Co0`<oNaCR1`<kNaCU1a<iN_Co0hNSOV?n0n@oNQ?Q1:O10O10002M2OKc@SO[?m0i@POV?o0810O100O02O2N1OL_@UO_?j0e@TO[?k0i@SOV?k0;01O1N1011O00KV@@f?`0^@^Oa?`0<O1O001N2O1O11000Hi_O2_`0JkhT6"}, "label": "a red , grey , white and black kite"}]}
{"id": 36658, "image": "COCO_train2014_000000036658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with dark blue party dress\"."}], "task": "refering", "answer_template": "The \"a girl with dark blue party dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 295, 316, 317, 318, 338, 339, 340, 360, 361, 362, 363, 364, 384, 385, 386, 387], "bbox": [0.687953125, 0.5327291666666667, 0.8480624999999999, 1.0], "iscrowd": 0, "area": 12694.653649999997, "rle": {"size": [480, 640], "counts": "Xe^61n>3L4M3M3M2M4mB]Od;f0XD\\Og;e0WD]Oh;d0WD]Oh;d0VD^Oh;d0VD^Oi;c0UD_Oi;c0VD^Oh;d0VD^Oi;c0UD_Oj;b0UD_Oj;c0SD_Om;b0QD_On;c0oC_OP<c0nC^OQ<d0mC]OR<e0lC\\OS<f0kC[OW:VO]G_1[N]OU:YO]G[1]N]OS:]O]GW1_N]OQ:B\\GQ1cN^Om9G^Gk0dN_Oo8ROfGi0d0f0gN@e8RO[G3b0g0e0e0gNAf8ROZG;:?m0d0hNAe8FfG7j0b0kNBd8GeG7j0h1_7TNeG7i0f1`7UNgG6g0f1a7VNfG6g0e1a7WNgG7e0c1b7XNhG7d0b1b7ZNiG5c0b1b7[NkG5`0a1c7]NlG3`0`1c7^NmG98Z1i7_NPH>NT1P8aNQHa3n7`LSH_3m7aLSH_3m7bLRH^3n7bLSH]3m7cLSH]3m7dLSH[3m7eLSH[3m7eLTHZ3l7fLTHZ3l7gLSHW3o7iLRHS3Q8mLoGP3T8QMlGk2W8UMiGi2Z8VMgGg2[8YMeGf2]8ZMcGc2_8]MaGa2b8^M_G_2c8aM]G]2f8cMZG[2g8eMYG[2h8dMYGZ2h8aMRGbN5l3k8_MTGcN2m3k8_LgGF_ONNl3Q9YLgGI\\O3Ll3R9RLUH1hNo3a:RL]EP4b:QL\\EQ4d:8M2N3N2M2N[MPF=n9]OYFd0e9[O^Fe0_9[OcFd0]9[OfFb0[9]OgFa0Y9_OiF?X9@kF=V9BlF<Z9]OjF`0\\9YOjFb0X9ZOPG`0R9\\OVG>m8]OYG?k8\\OYGa0l8XOXGf0l8lNaGQ1Z;M2N3L4M`0@6JRP]1"}, "label": "a girl with dark blue party dress"}]}
{"id": 36658, "image": "COCO_train2014_000000036658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with umbrella standing\"."}], "task": "refering", "answer_template": "The \"a woman with umbrella standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 295, 316, 317, 318, 338, 339, 340, 360, 361, 362, 363, 364, 384, 385, 386, 387], "bbox": [0.687953125, 0.5327291666666667, 0.8480624999999999, 1.0], "iscrowd": 0, "area": 12694.653649999997, "rle": {"size": [480, 640], "counts": "Xe^61n>3L4M3M3M2M4mB]Od;f0XD\\Og;e0WD]Oh;d0WD]Oh;d0VD^Oh;d0VD^Oi;c0UD_Oi;c0VD^Oh;d0VD^Oi;c0UD_Oj;b0UD_Oj;c0SD_Om;b0QD_On;c0oC_OP<c0nC^OQ<d0mC]OR<e0lC\\OS<f0kC[OW:VO]G_1[N]OU:YO]G[1]N]OS:]O]GW1_N]OQ:B\\GQ1cN^Om9G^Gk0dN_Oo8ROfGi0d0f0gN@e8RO[G3b0g0e0e0gNAf8ROZG;:?m0d0hNAe8FfG7j0b0kNBd8GeG7j0h1_7TNeG7i0f1`7UNgG6g0f1a7VNfG6g0e1a7WNgG7e0c1b7XNhG7d0b1b7ZNiG5c0b1b7[NkG5`0a1c7]NlG3`0`1c7^NmG98Z1i7_NPH>NT1P8aNQHa3n7`LSH_3m7aLSH_3m7bLRH^3n7bLSH]3m7cLSH]3m7dLSH[3m7eLSH[3m7eLTHZ3l7fLTHZ3l7gLSHW3o7iLRHS3Q8mLoGP3T8QMlGk2W8UMiGi2Z8VMgGg2[8YMeGf2]8ZMcGc2_8]MaGa2b8^M_G_2c8aM]G]2f8cMZG[2g8eMYG[2h8dMYGZ2h8aMRGbN5l3k8_MTGcN2m3k8_LgGF_ONNl3Q9YLgGI\\O3Ll3R9RLUH1hNo3a:RL]EP4b:QL\\EQ4d:8M2N3N2M2N[MPF=n9]OYFd0e9[O^Fe0_9[OcFd0]9[OfFb0[9]OgFa0Y9_OiF?X9@kF=V9BlF<Z9]OjF`0\\9YOjFb0X9ZOPG`0R9\\OVG>m8]OYG?k8\\OYGa0l8XOXGf0l8lNaGQ1Z;M2N3L4M`0@6JRP]1"}, "label": "a woman with umbrella standing"}]}
{"id": 36658, "image": "COCO_train2014_000000036658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male model wearing top hat\"."}], "task": "refering", "answer_template": "The \"male model wearing top hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 29, 30, 31, 52, 53, 54, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376], "bbox": [0.114265625, 0.029812500000000002, 0.4796875, 0.9998958333333333], "iscrowd": 0, "area": 64330.88119999999, "rle": {"size": [480, 640], "counts": "i`R13l>7J1N3M2N3N1N2M3N2N2M2O2N1O0O1O102M3N0O0100001O1N2O1O1O000O010G9M4O1N2O1L4K4J7O1O0AbM\\D`2c;?O1N2O1N2OL4L4MaDVMW12k7e2jFbMj0h0b7n1fG\\M7e1Y7W1aHZMKV2P7b0VI^M@]2P76_Ih0\\6ZOeIh0V6[OkIe0Q6]OQJd0j5^OXJc0_5CcJ=U5IlJ9k4MUK5a44_KN_41aK2]4NcK3\\4McK6\\4HdK;Z4EfK<Y4CgK`0W4@iKa0W4^OhKe0V4ZOgKl0NXKY3h3hLV1A`Kf3V3gLa2m2bIhLj3>e2e2fM_M[2a2aMbM`2]2\\MhMd2X2SMPNn2o1QMRNP3n1oLRNR3n1mLPNV3o1jLoMY3Q2fLlM^3S2bLdMh3\\2WLmLa4R3g30O10O100O100O10000oNTDmNm;R1UDmNk;S1UDlNl;U1TDiNm;W1SDiNm;KQDe03^Om;MTDa0OBm;MYD<KEm;O\\D8GIm;O`D5CJn;1hDLZO2o;2REAoN=o;2_ELb:4`EJ`:6cEF]:;eEC[:<gEBZ:>gE@Z:`0gE_OY:a0hE]OY:c0R20N2L4K5J6J6J6O0101O00000000007Ig0YO2N00000O1000000000000000000000O01000000000000000000000cCAY:>iDGBNd;;iDJ_OMh;g1UD\\Nj;d1QDbNn;^1oCfNP<Y1nCjNR<V1mClNR<T1nCmNQ<R1oCPOP<P1oCQOQ<n0PDSOo;m0PDUOo;k0QDVOn;i0RDYOm;g0QD[Oe;n0ZDUO[;^2F:O001O2N1O2N1O1N2O1O1dIfKP2[4nMiKo1X4PNlKl1U4SNnKj1T4TNoKh1S4WNQLe1P4YNULc1l3\\NYL_1h3`N^LZ1c3eNbLV1_3hNgLR1[3mNjLn0W3QOnLj0S3UORMf0n2YOXMb0i2]O\\M>d2C`M8a2GdM3]2MhMNY21lMJT27PNDQ2;TN@l1`0aNSO`1l0mNgNS1Z1ZOWNh0h1EkM<T26ZMMc2g0iL\\OS3\\6M3L4L4M3L8I7H8I7H8I7H8G9G8G5K5I]lk4"}, "label": "male model wearing top hat"}]}
{"id": 36658, "image": "COCO_train2014_000000036658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black top hat adorned with a purple ribbon is found walking with a cane\"."}], "task": "refering", "answer_template": "The \"a man wearing a black top hat adorned with a purple ribbon is found walking with a cane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 29, 30, 31, 52, 53, 54, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376], "bbox": [0.114265625, 0.029812500000000002, 0.4796875, 0.9998958333333333], "iscrowd": 0, "area": 64330.88119999999, "rle": {"size": [480, 640], "counts": "i`R13l>7J1N3M2N3N1N2M3N2N2M2O2N1O0O1O102M3N0O0100001O1N2O1O1O000O010G9M4O1N2O1L4K4J7O1O0AbM\\D`2c;?O1N2O1N2OL4L4MaDVMW12k7e2jFbMj0h0b7n1fG\\M7e1Y7W1aHZMKV2P7b0VI^M@]2P76_Ih0\\6ZOeIh0V6[OkIe0Q6]OQJd0j5^OXJc0_5CcJ=U5IlJ9k4MUK5a44_KN_41aK2]4NcK3\\4McK6\\4HdK;Z4EfK<Y4CgK`0W4@iKa0W4^OhKe0V4ZOgKl0NXKY3h3hLV1A`Kf3V3gLa2m2bIhLj3>e2e2fM_M[2a2aMbM`2]2\\MhMd2X2SMPNn2o1QMRNP3n1oLRNR3n1mLPNV3o1jLoMY3Q2fLlM^3S2bLdMh3\\2WLmLa4R3g30O10O100O100O10000oNTDmNm;R1UDmNk;S1UDlNl;U1TDiNm;W1SDiNm;KQDe03^Om;MTDa0OBm;MYD<KEm;O\\D8GIm;O`D5CJn;1hDLZO2o;2REAoN=o;2_ELb:4`EJ`:6cEF]:;eEC[:<gEBZ:>gE@Z:`0gE_OY:a0hE]OY:c0R20N2L4K5J6J6J6O0101O00000000007Ig0YO2N00000O1000000000000000000000O01000000000000000000000cCAY:>iDGBNd;;iDJ_OMh;g1UD\\Nj;d1QDbNn;^1oCfNP<Y1nCjNR<V1mClNR<T1nCmNQ<R1oCPOP<P1oCQOQ<n0PDSOo;m0PDUOo;k0QDVOn;i0RDYOm;g0QD[Oe;n0ZDUO[;^2F:O001O2N1O2N1O1N2O1O1dIfKP2[4nMiKo1X4PNlKl1U4SNnKj1T4TNoKh1S4WNQLe1P4YNULc1l3\\NYL_1h3`N^LZ1c3eNbLV1_3hNgLR1[3mNjLn0W3QOnLj0S3UORMf0n2YOXMb0i2]O\\M>d2C`M8a2GdM3]2MhMNY21lMJT27PNDQ2;TN@l1`0aNSO`1l0mNgNS1Z1ZOWNh0h1EkM<T26ZMMc2g0iL\\OS3\\6M3L4L4M3L8I7H8I7H8I7H8G9G8G5K5I]lk4"}, "label": "a man wearing a black top hat adorned with a purple ribbon is found walking with a cane"}]}
{"id": 20279, "image": "COCO_train2014_000000020279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing the purple color tshirt and front side brown color cat\"."}], "task": "refering", "answer_template": "The \"a man wearing the purple color tshirt and front side brown color cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 157, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229], "bbox": [0.7331875, 0.0, 1.0, 0.5493333333333333], "iscrowd": 0, "area": 30066.3301, "rle": {"size": [480, 640], "counts": "_ok62i>5K6I6K5K5M4M2M3N2M4M2M3N2N2M4M2M3N2M4M2N2M3N3L3N2N2M4M2M3N2N2N3M2O1N2N3N1N2O1N3M2O1N2N2O2M2N2O1N3N1N2N2O2M2N2O1N3N3L3M3N2M3M3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N3M2N3M2N3M2N3M2N3M2N3L3N3M3M2N3M2N3M2N3M2N0000000000000000000000000000000000000000000000000000000000000000000000000000000000SOm0RMn2RMn2"}, "label": "a man wearing the purple color tshirt and front side brown color cat"}]}
{"id": 20279, "image": "COCO_train2014_000000020279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"faceless person wearing red\"."}], "task": "refering", "answer_template": "The \"faceless person wearing red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 157, 158, 159, 160, 181, 182, 183, 205, 206, 228, 229], "bbox": [0.7331875, 0.0, 1.0, 0.5493333333333333], "iscrowd": 0, "area": 30066.3301, "rle": {"size": [480, 640], "counts": "_ok62i>5K6I6K5K5M4M2M3N2M4M2M3N2N2M4M2M3N2M4M2N2M3N3L3N2N2M4M2M3N2N2N3M2O1N2N3N1N2O1N3M2O1N2N2O2M2N2O1N3N1N2N2O2M2N2O1N3N3L3M3N2M3M3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N3M2N3M2N3M2N3M2N3M2N3L3N3M3M2N3M2N3M2N3M2N0000000000000000000000000000000000000000000000000000000000000000000000000000000000SOm0RMn2RMn2"}, "label": "faceless person wearing red"}]}
{"id": 20279, "image": "COCO_train2014_000000020279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teenage boy in an orange jacket , petting a cat\"."}], "task": "refering", "answer_template": "The \"a teenage boy in an orange jacket , petting a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 185, 186, 187, 188], "bbox": [0.0, 0.0023333333333333335, 0.6875625000000001, 0.5319583333333333], "iscrowd": 0, "area": 77808.9431, "rle": {"size": [480, 640], "counts": "Y1m1o;U1O1O1O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O3M5K5K5K5K5K6J7I8H8H8H8H7I8H8H1O0O1000O100000000000000000000000000O1000000000000001O001O1O001O001O1O001N2O001O001O1O001O1O001O1O1O1O1O1O2N1O1O1O1O1O1OO1O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1M3N2N2N2M3N2N2N2N2M3O1000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O0000000000O10000000000O100000000O10000000000O10000000000O10000000000O100000000O10000000000O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O10O0100O10000O100O100O100O100O100O10000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O1000001O3M4L3M4L3L4M4L3M3M4L3M3M4K4L5J5K5K6J5J6K6J5K5K6J5J6K6J5K6J5K5K6Iedm2"}, "label": "a teenage boy in an orange jacket , petting a cat"}]}
{"id": 20279, "image": "COCO_train2014_000000020279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange jacket petting a cat\"."}], "task": "refering", "answer_template": "The \"a man in an orange jacket petting a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 185, 186, 187, 188], "bbox": [0.0, 0.0023333333333333335, 0.6875625000000001, 0.5319583333333333], "iscrowd": 0, "area": 77808.9431, "rle": {"size": [480, 640], "counts": "Y1m1o;U1O1O1O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O3M5K5K5K5K5K6J7I8H8H8H8H7I8H8H1O0O1000O100000000000000000000000000O1000000000000001O001O1O001O001O1O001N2O001O001O1O001O1O001O1O1O1O1O1O2N1O1O1O1O1O1OO1O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1M3N2N2N2M3N2N2N2N2M3O1000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O0000000000O10000000000O100000000O10000000000O10000000000O10000000000O100000000O10000000000O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O10O0100O10000O100O100O100O100O100O10000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O1000001O3M4L3M4L3L4M4L3M3M4L3M3M4K4L5J5K5K6J5J6K6J5K5K6J5J6K6J5K6J5K5K6Iedm2"}, "label": "a man in an orange jacket petting a cat"}]}
{"id": 323389, "image": "COCO_train2014_000000323389.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball catcher waiting for the pitch\"."}], "task": "refering", "answer_template": "The \"baseball catcher waiting for the pitch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 208, 209, 210, 211, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.00253125, 0.531824480369515, 0.353484375, 0.9868822170900693], "iscrowd": 0, "area": 23263.971550000002, "rle": {"size": [433, 640], "counts": "`V1W1\\;o0M3M3M3N2M4L3M3M3M3N6I5K2N2N2N2O1O1O1N2O1O1O1N2O2N1O1O1N2O1O1O1O100O1O100O1O100O100O1O100O1O100O1O2O0O1O100O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1BmJ]HT5b7?O0O1O100O100O1O100O100O1O100O1O100O10000000000OO2O1O1O1O1O1O1O1]M[HYOf7c0^H\\Ob7a0bH^O_7>dHA^7;fHD\\7;dHD^7;bHD`7;`HDb7;^HDd7<[HCg7<YHCi7<VHDk7<UHCm7<SHCo7=PHAS8>mGAT8?lG@V8?jG@W8`0iG_OY8F^GRO7X1\\8BbGUO2X1^8^OdGZONW1_8[OgG^OJV1a8XOhGBGU1c8TOjGHBS1e8ROlGK_OQ1h8ROjGM^OP1i8QOlGNZOQ1l8nNlG1XOP1m8mNmG4UOm0Y9eNcG>TOk0h:UOXEh0k:XOUEf0m:ZOSEd0P;\\OPE`0S;@nD=T;CnD9T;GnD4U;LmD0U;0lDLX;3jDIX;7jDEX;<hD@[;`0gD\\O[;d0i00001O00000000001O0001O00000001O00000000001OO1O101N1O100O1O1O100O1O1O100O2N100O2N1O100O2N10001O00001O00010O00001O00001O00001O000000TOoC7R<HSD3n;KWD1k;MYDOj;NZDNi;N]DNe;Oci^5"}, "label": "baseball catcher waiting for the pitch"}]}
{"id": 323389, "image": "COCO_train2014_000000323389.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher squatting in behind home plate\"."}], "task": "refering", "answer_template": "The \"the catcher squatting in behind home plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 208, 209, 210, 211, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.00253125, 0.531824480369515, 0.353484375, 0.9868822170900693], "iscrowd": 0, "area": 23263.971550000002, "rle": {"size": [433, 640], "counts": "`V1W1\\;o0M3M3M3N2M4L3M3M3M3N6I5K2N2N2N2O1O1O1N2O1O1O1N2O2N1O1O1N2O1O1O1O100O1O100O1O100O100O1O100O1O100O1O2O0O1O100O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1BmJ]HT5b7?O0O1O100O100O1O100O100O1O100O1O100O10000000000OO2O1O1O1O1O1O1O1]M[HYOf7c0^H\\Ob7a0bH^O_7>dHA^7;fHD\\7;dHD^7;bHD`7;`HDb7;^HDd7<[HCg7<YHCi7<VHDk7<UHCm7<SHCo7=PHAS8>mGAT8?lG@V8?jG@W8`0iG_OY8F^GRO7X1\\8BbGUO2X1^8^OdGZONW1_8[OgG^OJV1a8XOhGBGU1c8TOjGHBS1e8ROlGK_OQ1h8ROjGM^OP1i8QOlGNZOQ1l8nNlG1XOP1m8mNmG4UOm0Y9eNcG>TOk0h:UOXEh0k:XOUEf0m:ZOSEd0P;\\OPE`0S;@nD=T;CnD9T;GnD4U;LmD0U;0lDLX;3jDIX;7jDEX;<hD@[;`0gD\\O[;d0i00001O00000000001O0001O00000001O00000000001OO1O101N1O100O1O1O100O1O1O100O2N100O2N1O100O2N10001O00001O00010O00001O00001O00001O000000TOoC7R<HSD3n;KWD1k;MYDOj;NZDNi;N]DNe;Oci^5"}, "label": "the catcher squatting in behind home plate"}]}
{"id": 356159, "image": "COCO_train2014_000000356159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black wet suit on a white surf board surfing alongside another man who is closer to the shore\"."}], "task": "refering", "answer_template": "The \"a man wearing a black wet suit on a white surf board surfing alongside another man who is closer to the shore\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 207, 208, 209, 210, 211, 231, 232], "bbox": [0.023953125, 0.49054945054945054, 0.24756250000000005, 0.8582142857142857], "iscrowd": 0, "area": 5993.390499999998, "rle": {"size": [364, 640], "counts": "Xn52Y;1O2O1N2N1O2N2O1O001N2O001O1O1O0O2OI7N20000O10O1000O10000O01kEEY9;fFGY98gFIY97fFKY95fFMY93fFNY92gF0X90gF2X9NhF3W9LiF5W9KiF6V9JiF8i7GeH1C8g7LbHLF:f7O`HGI<f71]HCM=d75[H^O0>d79XHYO4?b7=UHUO8`0a7?THPO<a0`7c0PHlN?6@Ko7a1aHaNDNi7b1aH`NJMc7d1bH_NMLa7e1aH_N1K\\7g1bH_N3JY7h1cH^N7HU7k1bH^N;FS7m1`H^N?CP7Q2_H\\Nc0An6T2^H[N\\8g1bGZN^8Y20000010O1O100O1O10O01O100ZN\\Gc0e8\\O`G?a8AbG<^8DfG7[8IhG3X8NlGMU83oGIQ87RH@T8`0PHSOY8m0S10000000VGQOV7P1jHPOV7P1iHROV7n0iHTOV7l0iHUOW7j0jHVOV7j0cH\\O^7d0^H@b7`0[HBf7>_G<b8D\\G>d8BZG?h8@WG`0j8Q10E<K4L5K4L4L5K4L4L5L3O0000O100O1000N1111O1N100O010000O2N101N101N2N1N3M3M3N3M4K4MhT[5"}, "label": "a man wearing a black wet suit on a white surf board surfing alongside another man who is closer to the shore"}]}
{"id": 429887, "image": "COCO_train2014_000000429887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clockface with multiple hands\"."}], "task": "refering", "answer_template": "The \"the clockface with multiple hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 67, 68, 69, 70, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 141, 142, 143, 144, 156, 157], "bbox": [0.4024590163934426, 0.138359375, 0.733512880562061, 0.457], "iscrowd": 0, "area": 20171.2871, "rle": {"size": [640, 427], "counts": "ff[3l0kb0>G7J6I6J7J6I7J4L5K4L5K5K4K6K4L5L3N3L4M2N2M3N2N2M3N2N2M3NO11N2O1O1O1O1O001O001O0010O01O001O001O001O00001O001N101O001O001O00001N101O01O100O010O1O010O10O01O10O0100O00100O010O1O010O10O0100O00100O1O1O1O2O0O1O1O2O0O100O2M2O1N3O0000001O00001M2O1N2O2N1N2N3M2N2N2N3Ic0@5L5K3L5K4K6K4K5L4VN`^On0fa0nNo^O<UeW2"}, "label": "the clockface with multiple hands"}]}
{"id": 429887, "image": "COCO_train2014_000000429887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top clock face\"."}], "task": "refering", "answer_template": "The \"the top clock face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 67, 68, 69, 70, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 141, 142, 143, 144, 156, 157], "bbox": [0.4024590163934426, 0.138359375, 0.733512880562061, 0.457], "iscrowd": 0, "area": 20171.2871, "rle": {"size": [640, 427], "counts": "ff[3l0kb0>G7J6I6J7J6I7J4L5K4L5K5K4K6K4L5L3N3L4M2N2M3N2N2M3N2N2M3NO11N2O1O1O1O1O001O001O0010O01O001O001O001O00001O001N101O001O001O00001N101O01O100O010O1O010O10O01O10O0100O00100O010O1O010O10O0100O00100O1O1O1O2O0O1O1O2O0O100O2M2O1N3O0000001O00001M2O1N2O2N1N2N3M2N2N2N3Ic0@5L5K3L5K4K6K4K5L4VN`^On0fa0nNo^O<UeW2"}, "label": "the top clock face"}]}
{"id": 257858, "image": "COCO_train2014_000000257858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black metal chair\"."}], "task": "refering", "answer_template": "The \"black metal chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 178, 179, 180, 201], "bbox": [0.6773125, 0.20945, 0.8717812500000001, 0.5957], "iscrowd": 0, "area": 6803.014850000001, "rle": {"size": [400, 640], "counts": "YgY52\\<2O2M2O2M2O000O0100O10O10O10O01OO1O1O1N2O1O1O1N2O2N1O1VN_OeGc0]8GVG:l81fF0\\9`12N2N2N20010O0000000001O000000001O01O001O010O00kFmMn7R2mGSN4Ij6T2mHXN7Fm6Q2hH\\N:Fm6o1cH`N?Bo6m1^HbNd0Cn6k1YHcNk0El6g1UHdNQ1Gj6\\2XIfMi6U2YImMg6o1\\ISNd6h1^IZNb6b1`I^Nc6_1]I`Nf6`1XI`Nk6_1SI`NP7`1oH_NT7`1jH_NY7a1eH_N^7`1`H_Nc7a1[H_Nh7`1VH_Nm7a1QH_NR8`1jGaNY8_1dGbN_8]1^GcNe8]1XGdNk8[1RGeNQ9[1kFgNX9X1eFhN^9X1_FiNd9V1YFjNj9l13L3N3N1O2N1N3N1O2N1N3N1VO^EEd:8^EG`1[O^7l0TGG^1_O_7g0UGIZ1Cb7a0VGKU1Hf7:WGMQ1Lh75YGOk00m7NZG1g04o7I\\G3a08S8D]G4<<W8_O^G58<]8]O]G63=f8ZOYG8N=P9WOSG<I=S:CiE=\\:AbE>b:BZE>j:BSE=R;BjD?Z;75K5K5K5K6J5KWlo0"}, "label": "black metal chair"}]}
{"id": 257858, "image": "COCO_train2014_000000257858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair with the man in the white shirt and blue short ' s sitting in it\"."}], "task": "refering", "answer_template": "The \"the chair with the man in the white shirt and blue short ' s sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 178, 179, 180, 201], "bbox": [0.6773125, 0.20945, 0.8717812500000001, 0.5957], "iscrowd": 0, "area": 6803.014850000001, "rle": {"size": [400, 640], "counts": "YgY52\\<2O2M2O2M2O000O0100O10O10O10O01OO1O1O1N2O1O1O1N2O2N1O1VN_OeGc0]8GVG:l81fF0\\9`12N2N2N20010O0000000001O000000001O01O001O010O00kFmMn7R2mGSN4Ij6T2mHXN7Fm6Q2hH\\N:Fm6o1cH`N?Bo6m1^HbNd0Cn6k1YHcNk0El6g1UHdNQ1Gj6\\2XIfMi6U2YImMg6o1\\ISNd6h1^IZNb6b1`I^Nc6_1]I`Nf6`1XI`Nk6_1SI`NP7`1oH_NT7`1jH_NY7a1eH_N^7`1`H_Nc7a1[H_Nh7`1VH_Nm7a1QH_NR8`1jGaNY8_1dGbN_8]1^GcNe8]1XGdNk8[1RGeNQ9[1kFgNX9X1eFhN^9X1_FiNd9V1YFjNj9l13L3N3N1O2N1N3N1O2N1N3N1VO^EEd:8^EG`1[O^7l0TGG^1_O_7g0UGIZ1Cb7a0VGKU1Hf7:WGMQ1Lh75YGOk00m7NZG1g04o7I\\G3a08S8D]G4<<W8_O^G58<]8]O]G63=f8ZOYG8N=P9WOSG<I=S:CiE=\\:AbE>b:BZE>j:BSE=R;BjD?Z;75K5K5K5K6J5KWlo0"}, "label": "the chair with the man in the white shirt and blue short ' s sitting in it"}]}
{"id": 528198, "image": "COCO_train2014_000000528198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck that is behind pulled by a black pick up truck\"."}], "task": "refering", "answer_template": "The \"a white truck that is behind pulled by a black pick up truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 246, 247, 248], "bbox": [0.459265625, 0.38518749999999996, 0.9222187500000001, 0.6271666666666665], "iscrowd": 0, "area": 26873.57875, "rle": {"size": [480, 640], "counts": "`QZ41m>4L4M3L4L4M3L3N3L4O1N2J6H7I8H8H8H8H8M3N1O2N000000000000000000000000000000000000000000000001O00000000000O100000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O1O001O1O2N1O1O1O2N1O1O2N100O1O2N00000000000000001O00000000000000001O000000000000O1000000O100000000O10001O0O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O10000000000000000000000000000000001O0000000000000000000000000000000000000000000001O0000000000000000000000000001O0000000000000000000000000N2A?@`0A?@`0A_Xg0"}, "label": "a white truck that is behind pulled by a black pick up truck"}]}
{"id": 528198, "image": "COCO_train2014_000000528198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white travel trailer parked infront of a house\"."}], "task": "refering", "answer_template": "The \"a white travel trailer parked infront of a house\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 246, 247, 248], "bbox": [0.459265625, 0.38518749999999996, 0.9222187500000001, 0.6271666666666665], "iscrowd": 0, "area": 26873.57875, "rle": {"size": [480, 640], "counts": "`QZ41m>4L4M3L4L4M3L3N3L4O1N2J6H7I8H8H8H8H8M3N1O2N000000000000000000000000000000000000000000000001O00000000000O100000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O1O001O1O2N1O1O1O2N1O1O2N100O1O2N00000000000000001O00000000000000001O000000000000O1000000O100000000O10001O0O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O10000000000000000000000000000000001O0000000000000000000000000000000000000000000001O0000000000000000000000000001O0000000000000000000000000N2A?@`0A?@`0A_Xg0"}, "label": "a white travel trailer parked infront of a house"}]}
{"id": 528198, "image": "COCO_train2014_000000528198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large black pick - up truck with a white trailer hitched to it\"."}], "task": "refering", "answer_template": "The \"a large black pick - up truck with a white trailer hitched to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 232, 237, 238], "bbox": [0.056921875, 0.44666666666666666, 0.39778125000000003, 0.6061458333333333], "iscrowd": 0, "area": 10889.4053, "rle": {"size": [480, 640], "counts": "UTa01k>7K3O1M3C<O1O101O0O10000O5L9G1N2O001O1O00001O000000001OO10O1000000000O100O1O1O1O1O100N2N2O11O1N2O2M2N2N2N1O1O1O1O10O0100O1O010O010O010O010O01O0100O100O100O100000000000000O1000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000O10e0[O00000000000000001O00000O10000000000002N2N3M2N2N2N1O001O00001O000000000001O000000000O10000O100O1O100O100O100O100O100O100O1N2O10001O000000000O100000000000000B>[OUXd5"}, "label": "a large black pick - up truck with a white trailer hitched to it"}]}
{"id": 528198, "image": "COCO_train2014_000000528198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car in the road\"."}], "task": "refering", "answer_template": "The \"a car in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 232, 237, 238], "bbox": [0.056921875, 0.44666666666666666, 0.39778125000000003, 0.6061458333333333], "iscrowd": 0, "area": 10889.4053, "rle": {"size": [480, 640], "counts": "UTa01k>7K3O1M3C<O1O101O0O10000O5L9G1N2O001O1O00001O000000001OO10O1000000000O100O1O1O1O1O100N2N2O11O1N2O2M2N2N2N1O1O1O1O10O0100O1O010O010O010O010O01O0100O100O100O100000000000000O1000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000O10e0[O00000000000000001O00000O10000000000002N2N3M2N2N2N1O001O00001O000000000001O000000000O10000O100O1O100O100O100O100O100O100O1N2O10001O000000000O100000000000000B>[OUXd5"}, "label": "a car in the road"}]}
{"id": 347976, "image": "COCO_train2014_000000347976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car parked near some bikes\"."}], "task": "refering", "answer_template": "The \"a white car parked near some bikes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 120, 121, 122, 132, 133, 134, 144, 145, 146, 156, 157, 168, 169], "bbox": [0.0, 0.49892000000000003, 0.22651651651651653, 0.79634], "iscrowd": 0, "area": 7922.892550000001, "rle": {"size": [500, 333], "counts": "\\:=^>S1dN]1nNQ1F;F8G01O001O010O001O1O001O001O0010O01O1O001O001N1O2N1O2N1N3N2N1O2N1N3N1O2M2O2M3N1O1N2O0EcC_M_<`2;O1N200O010O100O100N101O1O1O1O2N3M3L4M2N3M3M3M7Ia0_O;E^em3"}, "label": "a white car parked near some bikes"}]}
{"id": 347976, "image": "COCO_train2014_000000347976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"side end of a white sedan\"."}], "task": "refering", "answer_template": "The \"side end of a white sedan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 120, 121, 122, 132, 133, 134, 144, 145, 146, 156, 157, 168, 169], "bbox": [0.0, 0.49892000000000003, 0.22651651651651653, 0.79634], "iscrowd": 0, "area": 7922.892550000001, "rle": {"size": [500, 333], "counts": "\\:=^>S1dN]1nNQ1F;F8G01O001O010O001O1O001O001O0010O01O1O001O001N1O2N1O2N1N3N2N1O2N1N3N1O2M2O2M3N1O1N2O0EcC_M_<`2;O1N200O010O100O100N101O1O1O1O2N3M3L4M2N3M3M3M7Ia0_O;E^em3"}, "label": "side end of a white sedan"}]}
{"id": 347976, "image": "COCO_train2014_000000347976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bike in between two other bikes\"."}], "task": "refering", "answer_template": "The \"a black bike in between two other bikes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 125, 126, 127, 128, 136, 137, 139, 140, 141, 147, 148, 149, 150, 151, 152, 153, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174], "bbox": [0.31666666666666665, 0.5093799999999999, 0.8251951951951952, 0.8264], "iscrowd": 0, "area": 12410.465500000008, "rle": {"size": [500, 333], "counts": "`cc11]?7J6N2n@Id>k0G9H5K4L4L5K4L4L4M1N2N2N2N2N2O1N2N2O1N100O101O0O10000mNQM`EQ3_:RM^En2b:TM\\El2d:TM\\El2d:TM[Em2e:SM[Em2e:SM[Em2e:SM[Em2d:TM\\El2d:TM\\El2d:TM[Em2e:SM[Em2e:SM[Em2e:SM[Em2e:SM[Em2e:SM[El2e:UMZEl2f:TMZEl2f:TMZEk2g:WMTEl2l:VMnDHGl2\\;UMaD45I4l2W;VMeD3<f2o:UMbEi2_:VMcEh2g;N2M3M3N2M2N3L4M3L3NkMWNbFg1^9YNdFd1^9\\N[EM9e1]:^NZEM:a1^:bNXEN;\\1`:fNUEN<X1b:iNRE0=S1d:mNoD0>Q1c:POZDM:3j0P1a:oNZD1:On0o0^:QOYD48MS1m0Z:RO[D58LT1k0Z:TOYD4;MR1k0Z:TOXD3>NP1k0Z:SOXD3`0Oo0j0X:UOXD1c00m0i0Y:UOWD2d00l0i0Y:UOVD2f00k0i0X:\\OnDJk0j0W:[OoDKj0j0W:ZOoDMi0i0Y:YOoDNh0i0X:ZOPEMg0j0Y:XOQENf0j0Y:WOREOd0k0Z:UOSE0a0l0]:TOREO`0o0^:ROREO>P1b:POPE0=Q1d:oNmD1=Q1h:nNjD1<S1k:lNhD1;:mN9R<[OfD29;SO5o;^OdD2LND>10l;BcD2KND?7Kh;EbD2K0B?;Hg;GaD2J0C?<Hf;GaD2I0C`0?Gc;H`D2I1D>a0Gb;J^D0J1E>c0Db;OZDNK2D>g0_Oc;7nCO3MEb0S=7WCXOEm0h<oN[Cl0OYOMU1a<fNgCf2V<YMkCk2P<UMPDo2l;RMTDP3j;oLWDR3g;oLYDS3d;nL\\DT3^;PMbDQ3];oLcDR3\\;mLeDT3Z;lLfDT3Z;lLfDU3Y;kLgDV3X;jLhDW3W;hLjDX3V;jLhDW3W;iLjDV3W;hLlDW3T;hLnDV3S;hLPEV3i;O1N2O2M4M3L1O10O01O01O01O001O00001O1O0010OO2N2`NoB9S=FoB7T=GoB3V=KlB1W=NkBK\\=4fB@e=>\\BAf==[BAh==YBBh==YBAj==WBAl==UBBm==SBAQ><f0M3M2NUPl0"}, "label": "a black bike in between two other bikes"}]}
{"id": 347976, "image": "COCO_train2014_000000347976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black bicycle directly below the sign\"."}], "task": "refering", "answer_template": "The \"the black bicycle directly below the sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 125, 126, 127, 128, 136, 137, 139, 140, 141, 147, 148, 149, 150, 151, 152, 153, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174], "bbox": [0.31666666666666665, 0.5093799999999999, 0.8251951951951952, 0.8264], "iscrowd": 0, "area": 12410.465500000008, "rle": {"size": [500, 333], "counts": "`cc11]?7J6N2n@Id>k0G9H5K4L4L5K4L4L4M1N2N2N2N2N2O1N2N2O1N100O101O0O10000mNQM`EQ3_:RM^En2b:TM\\El2d:TM\\El2d:TM[Em2e:SM[Em2e:SM[Em2e:SM[Em2d:TM\\El2d:TM\\El2d:TM[Em2e:SM[Em2e:SM[Em2e:SM[Em2e:SM[Em2e:SM[El2e:UMZEl2f:TMZEl2f:TMZEk2g:WMTEl2l:VMnDHGl2\\;UMaD45I4l2W;VMeD3<f2o:UMbEi2_:VMcEh2g;N2M3M3N2M2N3L4M3L3NkMWNbFg1^9YNdFd1^9\\N[EM9e1]:^NZEM:a1^:bNXEN;\\1`:fNUEN<X1b:iNRE0=S1d:mNoD0>Q1c:POZDM:3j0P1a:oNZD1:On0o0^:QOYD48MS1m0Z:RO[D58LT1k0Z:TOYD4;MR1k0Z:TOXD3>NP1k0Z:SOXD3`0Oo0j0X:UOXD1c00m0i0Y:UOWD2d00l0i0Y:UOVD2f00k0i0X:\\OnDJk0j0W:[OoDKj0j0W:ZOoDMi0i0Y:YOoDNh0i0X:ZOPEMg0j0Y:XOQENf0j0Y:WOREOd0k0Z:UOSE0a0l0]:TOREO`0o0^:ROREO>P1b:POPE0=Q1d:oNmD1=Q1h:nNjD1<S1k:lNhD1;:mN9R<[OfD29;SO5o;^OdD2LND>10l;BcD2KND?7Kh;EbD2K0B?;Hg;GaD2J0C?<Hf;GaD2I0C`0?Gc;H`D2I1D>a0Gb;J^D0J1E>c0Db;OZDNK2D>g0_Oc;7nCO3MEb0S=7WCXOEm0h<oN[Cl0OYOMU1a<fNgCf2V<YMkCk2P<UMPDo2l;RMTDP3j;oLWDR3g;oLYDS3d;nL\\DT3^;PMbDQ3];oLcDR3\\;mLeDT3Z;lLfDT3Z;lLfDU3Y;kLgDV3X;jLhDW3W;hLjDX3V;jLhDW3W;iLjDV3W;hLlDW3T;hLnDV3S;hLPEV3i;O1N2O2M4M3L1O10O01O01O01O001O00001O1O0010OO2N2`NoB9S=FoB7T=GoB3V=KlB1W=NkBK\\=4fB@e=>\\BAf==[BAh==YBBh==YBAj==WBAl==UBBm==SBAQ><f0M3M2NUPl0"}, "label": "the black bicycle directly below the sign"}]}
{"id": 347976, "image": "COCO_train2014_000000347976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back tire and seat of black bike partially out of frame\"."}], "task": "refering", "answer_template": "The \"back tire and seat of black bike partially out of frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 141, 142, 143, 153, 154, 155, 165, 166, 167, 177, 178, 179, 189, 190, 191], "bbox": [0.79993993993994, 0.55102, 0.9957357357357357, 0.87748], "iscrowd": 0, "area": 7146.480650000001, "rle": {"size": [500, 333], "counts": "kWR4b0j>>iNUOQC[1e<V1C:F5K5K4M4K4L2M3N3L3N3K4L4M4L3M4M2N101O1O1O001O1O10O1O001O1O2N2N5LNfKYEj3f:oKgEl3S;M2N2N2N3M2N2N3L4M3L4M2M4L4J6J6J6J6J6J6J5K6J6K5K6I6K5CR5"}, "label": "back tire and seat of black bike partially out of frame"}]}
{"id": 167755, "image": "COCO_train2014_000000167755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small brown cow standing outside\"."}], "task": "refering", "answer_template": "The \"a small brown cow standing outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 259, 276, 277, 282, 283, 299, 300, 301], "bbox": [0.0, 0.4383411214953271, 0.47171874999999996, 0.9286682242990655], "iscrowd": 0, "area": 31055.453099999995, "rle": {"size": [428, 640], "counts": "m69R=a0_Og0YO5K4L5J5L4_FlMc7Y2UHkMj7Y2oGkMP8Y2jGjMV8Y2cGkM\\8Y2ZGXN\\8k1XGgN^8W3L4M3L4M1N3M3N2M2OoN\\H\\La7_3hH`LU7Z3UIcLk6W3]IhLa6Z3aIcL_6_3bI^L]6e3dIXL\\6j3eISLZ6o3hInKX6T4iIiKW6Y4iIeKW6]4hIbKX6`4gI_KX6d4hIZKX6j4S14L3M4L3M3L4M2N3M2N3M]NeHWMY7h2QIPMn6n2XIPMf6o2]IPMb6`2PJ_Mo5\\2YJcMd5Y2cJfM\\5V2jJjMT5U2oJjMP5T2SKmMk4R2XKlMh4R2\\KlMd4S2_KkMa4S2cKkM]4U2eKiM[4X2fKfMY4[2iKcMW4^2iKaMW4_2kK^MV4c2kK[MU4e2mKYMS4h2nKVMR4j2PLTMP4m2QLQMo3P3QLoLo3Q3SLmLm3T3U3001O1O001O001O1O001O01O01O001O00001O001O000000O100O100O10001N100000000O2O000000000O2O00000000001N10000000001O0000O010000000000000000O1000000000000O10000000000O010000O1000O02O000O2O001N10001N`0A2N3L4M2N2M4M2N2M4Mb0^O3M3L4M2N000O100000O101XNRHPNo7n1YHkMh7T2_HeMb7Y2RITMo6k2UIQMl6m2ZIgLn6W3e1O1N2N2O1N2O1N1O2O0O2N101N1O2O0O2O0O2N101N1O2N1O2N1O1N2O1O1O1O1O1N200000000000001OO10000O10000O10000O10O010000O10000O12N2M3NO1O1O1O1O1O1O1O0O21O0100O101N1O100002N1N3N1O2N1O001O0N3N1cNcDh0_;SOgDj0\\;QOjDl0W;POnDm0l;M2M4M2M4M1N3K4M3L[P]4"}, "label": "a small brown cow standing outside"}]}
{"id": 167755, "image": "COCO_train2014_000000167755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow standing up\"."}], "task": "refering", "answer_template": "The \"a brown cow standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 259, 276, 277, 282, 283, 299, 300, 301], "bbox": [0.0, 0.4383411214953271, 0.47171874999999996, 0.9286682242990655], "iscrowd": 0, "area": 31055.453099999995, "rle": {"size": [428, 640], "counts": "m69R=a0_Og0YO5K4L5J5L4_FlMc7Y2UHkMj7Y2oGkMP8Y2jGjMV8Y2cGkM\\8Y2ZGXN\\8k1XGgN^8W3L4M3L4M1N3M3N2M2OoN\\H\\La7_3hH`LU7Z3UIcLk6W3]IhLa6Z3aIcL_6_3bI^L]6e3dIXL\\6j3eISLZ6o3hInKX6T4iIiKW6Y4iIeKW6]4hIbKX6`4gI_KX6d4hIZKX6j4S14L3M4L3M3L4M2N3M2N3M]NeHWMY7h2QIPMn6n2XIPMf6o2]IPMb6`2PJ_Mo5\\2YJcMd5Y2cJfM\\5V2jJjMT5U2oJjMP5T2SKmMk4R2XKlMh4R2\\KlMd4S2_KkMa4S2cKkM]4U2eKiM[4X2fKfMY4[2iKcMW4^2iKaMW4_2kK^MV4c2kK[MU4e2mKYMS4h2nKVMR4j2PLTMP4m2QLQMo3P3QLoLo3Q3SLmLm3T3U3001O1O001O001O1O001O01O01O001O00001O001O000000O100O100O10001N100000000O2O000000000O2O00000000001N10000000001O0000O010000000000000000O1000000000000O10000000000O010000O1000O02O000O2O001N10001N`0A2N3L4M2N2M4M2N2M4Mb0^O3M3L4M2N000O100000O101XNRHPNo7n1YHkMh7T2_HeMb7Y2RITMo6k2UIQMl6m2ZIgLn6W3e1O1N2N2O1N2O1N1O2O0O2N101N1O2O0O2O0O2N101N1O2N1O2N1O1N2O1O1O1O1O1N200000000000001OO10000O10000O10000O10O010000O10000O12N2M3NO1O1O1O1O1O1O1O0O21O0100O101N1O100002N1N3N1O2N1O001O0N3N1cNcDh0_;SOgDj0\\;QOjDl0W;POnDm0l;M2M4M2M4M1N3K4M3L[P]4"}, "label": "a brown cow standing up"}]}
{"id": 167755, "image": "COCO_train2014_000000167755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow in the middle\"."}], "task": "refering", "answer_template": "The \"the cow in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 194], "bbox": [0.28665625, 0.316822429906542, 0.60728125, 0.5552102803738318], "iscrowd": 0, "area": 10457.06895, "rle": {"size": [428, 640], "counts": "hd\\22Y=1O2N100000000O10000000000000000000000000000000000000000\\CLP<4mC1Q<0kC4T<LhC:V<FgC?W<<O2O0I8J5N3L3N3L3N3L3N2N2M20O010000000O010000000O1000001O0O101O00000O2O00001O0O10001O02N1O10O0N2N2O1N3M2N2100O1O010O1O10O01O100O001O2O1N2N2N2N2N2N1O2N2N2N3M3M1O00000010N1O2O1N3N1N3M2O2M2N2O2M101N2O1N2O1O0O10000O01000O0100O010O1000O010O1O10O01O1O1O01O00001O000O20O1O2O0O1M4O0O2O00001O001O0O101O001O2hNPDP1V<00O10001O0000001N2O1O1N2N2O1N3N2M3N2L4L4L3MdTY3"}, "label": "the cow in the middle"}]}
{"id": 167755, "image": "COCO_train2014_000000167755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cow eats hay from a wooden trough\"."}], "task": "refering", "answer_template": "The \"a white cow eats hay from a wooden trough\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 194], "bbox": [0.28665625, 0.316822429906542, 0.60728125, 0.5552102803738318], "iscrowd": 0, "area": 10457.06895, "rle": {"size": [428, 640], "counts": "hd\\22Y=1O2N100000000O10000000000000000000000000000000000000000\\CLP<4mC1Q<0kC4T<LhC:V<FgC?W<<O2O0I8J5N3L3N3L3N3L3N2N2M20O010000000O010000000O1000001O0O101O00000O2O00001O0O10001O02N1O10O0N2N2O1N3M2N2100O1O010O1O10O01O100O001O2O1N2N2N2N2N2N1O2N2N2N3M3M1O00000010N1O2O1N3N1N3M2O2M2N2O2M101N2O1N2O1O0O10000O01000O0100O010O1000O010O1O10O01O1O1O01O00001O000O20O1O2O0O1M4O0O2O00001O001O0O101O001O2hNPDP1V<00O10001O0000001N2O1O1N2N2O1N3N2M3N2L4L4L3MdTY3"}, "label": "a white cow eats hay from a wooden trough"}]}
{"id": 249675, "image": "COCO_train2014_000000249675.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bus behind a police man\"."}], "task": "refering", "answer_template": "The \"a blue bus behind a police man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 379, 380, 381, 382, 383], "bbox": [0.44542187499999997, 0.22397496087636934, 0.9535625, 0.7130359937402191], "iscrowd": 0, "area": 88095.29100000001, "rle": {"size": [639, 640], "counts": "]Tb57\\c0<E<C=C<D=C=ZB_M`9m2YFZM[9S3bFoLR9^3kFdLi8h3UGZL_8S4^GoKV8^4gGdKV8_4iGbKW8_4fGcKZ8^4cGdK]8\\4aGfK_8[4^GgKa8[4\\GgKd8_4UGbKk8R5_FPKa9d5iE^JW:V6SElIm:c70000000O01000000000000000000O100000002N5^ERGi9R9RFnFo9W9kEjFU:Z9fEgFZ:_9O10000000000000O1000O100000000000000000O01000000000000000000O010000000000000000000O0100000000000000000O10O100000001O00000000000O10000000001O00000000000O10000000001O000000000O10000000001O000000000O1000000000001O0000000O1000000000001O0000000O100000000000001O0000000O1000000000001O0000000O100000000000001O00000O10000000000000001O000O10000000000000001O00000000000000000000001O00000O100000000000001O0O10000O2O0O2O0O2O0O2O001N101N101N101N2O0O2O001N101N101N101N101O0O2O02O0O2O1N101OO001O001O001O001O001O001O001O001O001O0TOUDdIl;Z6_D\\Ib;a6o0O2N1N3N1O2M2O2N1N3N1N3YOf0VOk0VOi0WOj0VOi0VOk0VOgab0"}, "label": "a blue bus behind a police man"}]}
{"id": 249675, "image": "COCO_train2014_000000249675.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bus in front of a traffic officer\"."}], "task": "refering", "answer_template": "The \"a blue bus in front of a traffic officer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 379, 380, 381, 382, 383], "bbox": [0.44542187499999997, 0.22397496087636934, 0.9535625, 0.7130359937402191], "iscrowd": 0, "area": 88095.29100000001, "rle": {"size": [639, 640], "counts": "]Tb57\\c0<E<C=C<D=C=ZB_M`9m2YFZM[9S3bFoLR9^3kFdLi8h3UGZL_8S4^GoKV8^4gGdKV8_4iGbKW8_4fGcKZ8^4cGdK]8\\4aGfK_8[4^GgKa8[4\\GgKd8_4UGbKk8R5_FPKa9d5iE^JW:V6SElIm:c70000000O01000000000000000000O100000002N5^ERGi9R9RFnFo9W9kEjFU:Z9fEgFZ:_9O10000000000000O1000O100000000000000000O01000000000000000000O010000000000000000000O0100000000000000000O10O100000001O00000000000O10000000001O00000000000O10000000001O000000000O10000000001O000000000O1000000000001O0000000O1000000000001O0000000O100000000000001O0000000O1000000000001O0000000O100000000000001O00000O10000000000000001O000O10000000000000001O00000000000000000000001O00000O100000000000001O0O10000O2O0O2O0O2O0O2O001N101N101N101N2O0O2O001N101N101N101N101O0O2O02O0O2O1N101OO001O001O001O001O001O001O001O001O001O0TOUDdIl;Z6_D\\Ib;a6o0O2N1N3N1O2M2O2N1N3N1N3YOf0VOk0VOi0WOj0VOi0VOk0VOgab0"}, "label": "a blue bus in front of a traffic officer"}]}
{"id": 249675, "image": "COCO_train2014_000000249675.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was standing on the road\"."}], "task": "refering", "answer_template": "The \"a man was standing on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 238, 239, 240, 261, 262, 263, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 332, 333, 353, 354, 355, 376, 377, 378, 399, 400, 401, 422, 423, 424, 445, 446], "bbox": [0.344125, 0.4172300469483568, 0.502609375, 0.8412676056338028], "iscrowd": 0, "area": 16183.38545, "rle": {"size": [639, 640], "counts": "^cY42lc08H8I6I8H8H6J3N2L5L4K5L4K5K5L4K5L4K5_BUMe9Q3mCoLQ16k:P3nCWMl0Oo:P3oC^NEfNY<Q3kC`NFbN]<S3gC\\O];kLQEn3[O[Od;iLmD\\7S;gHgD]7X;fHcD\\7];gH^D[7b;a000000001O0000010O01O0010O0000010O01O00100N2O1N2O0O2O1O1N101N2N?A1N3N2N1O2M3N2N2N2QN_C]Lc<_3Q2M2M4M3L5]Od0K4K6K4K6K3L5L4K5L3L5L4K4M4K5L4L3I8F:H8K4M4L4L3M4LPSV6"}, "label": "a man was standing on the road"}]}
{"id": 249675, "image": "COCO_train2014_000000249675.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in uniform wearing yellow vest\"."}], "task": "refering", "answer_template": "The \"man in uniform wearing yellow vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 238, 239, 240, 261, 262, 263, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 332, 333, 353, 354, 355, 376, 377, 378, 399, 400, 401, 422, 423, 424, 445, 446], "bbox": [0.344125, 0.4172300469483568, 0.502609375, 0.8412676056338028], "iscrowd": 0, "area": 16183.38545, "rle": {"size": [639, 640], "counts": "^cY42lc08H8I6I8H8H6J3N2L5L4K5L4K5K5L4K5L4K5_BUMe9Q3mCoLQ16k:P3nCWMl0Oo:P3oC^NEfNY<Q3kC`NFbN]<S3gC\\O];kLQEn3[O[Od;iLmD\\7S;gHgD]7X;fHcD\\7];gH^D[7b;a000000001O0000010O01O0010O0000010O01O00100N2O1N2O0O2O1O1N101N2N?A1N3N2N1O2M3N2N2N2QN_C]Lc<_3Q2M2M4M3L5]Od0K4K6K4K6K3L5L4K5L3L5L4K4M4K5L4L3I8F:H8K4M4L4L3M4LPSV6"}, "label": "man in uniform wearing yellow vest"}]}
{"id": 405324, "image": "COCO_train2014_000000405324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass to the left of a plate of food\"."}], "task": "refering", "answer_template": "The \"a wine glass to the left of a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 132, 133, 134, 144, 145, 146, 156, 157, 168, 169, 180, 181, 182, 192, 193], "bbox": [0.0, 0.56992, 0.18678678678678678, 0.9171999999999999], "iscrowd": 0, "area": 7288.554900000001, "rle": {"size": [500, 333], "counts": "X9[2X=2O1N1jDcMT9_2jFdMT9]2jFgMR9[2mFhMo8Z2oFjMn8V2QGnMk8T2SGPNi8R2VGPNh8Q2VGPNi8R2UGoMj8R2UGoMk8R2TGnMk8T2SGmMl8U2RGlMn8T2RGlMn8U2PGlMQ9T2mFmMS9Y2fFhMZ9^2_FcMa9c2WF`Mi9e2oE]MQ:S400001O00001O0000001O01OcMgF_NY9_1iF`NY9]1iFcNW9\\1jFcNW9[1lFdNU9Z1lFeNU9Y1mFgNS9X1nFgNS9W1oFiNR9U1oFjNR9T1QGkNo8T1RGkNo8S1SGmNn8Q1SGnNn8P1TGPOl8o0UGPOl8n0WGQOj8l0XGTOh8k0YGTOk8g0WGYOP9?QG@V98lFH[90gFN`9JbF6e9B^3GloS4"}, "label": "a wine glass to the left of a plate of food"}]}
{"id": 405324, "image": "COCO_train2014_000000405324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass on tha tabule\"."}], "task": "refering", "answer_template": "The \"a glass on tha tabule\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 132, 133, 134, 144, 145, 146, 156, 157, 168, 169, 180, 181, 182, 192, 193], "bbox": [0.0, 0.56992, 0.18678678678678678, 0.9171999999999999], "iscrowd": 0, "area": 7288.554900000001, "rle": {"size": [500, 333], "counts": "X9[2X=2O1N1jDcMT9_2jFdMT9]2jFgMR9[2mFhMo8Z2oFjMn8V2QGnMk8T2SGPNi8R2VGPNh8Q2VGPNi8R2UGoMj8R2UGoMk8R2TGnMk8T2SGmMl8U2RGlMn8T2RGlMn8U2PGlMQ9T2mFmMS9Y2fFhMZ9^2_FcMa9c2WF`Mi9e2oE]MQ:S400001O00001O0000001O01OcMgF_NY9_1iF`NY9]1iFcNW9\\1jFcNW9[1lFdNU9Z1lFeNU9Y1mFgNS9X1nFgNS9W1oFiNR9U1oFjNR9T1QGkNo8T1RGkNo8S1SGmNn8Q1SGnNn8P1TGPOl8o0UGPOl8n0WGQOj8l0XGTOh8k0YGTOk8g0WGYOP9?QG@V98lFH[90gFN`9JbF6e9B^3GloS4"}, "label": "a glass on tha tabule"}]}
{"id": 405324, "image": "COCO_train2014_000000405324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl smiling at the camera with silverware in hands\"."}], "task": "refering", "answer_template": "The \"a girl smiling at the camera with silverware in hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 40, 41, 42, 43, 44, 45, 52, 53, 54, 55, 56, 57, 63, 64, 65, 66, 67, 68, 69, 74, 75, 76, 77, 78, 79, 80, 81, 82, 86, 87, 88, 89, 90, 91, 92, 93, 94, 98, 99, 100, 101, 102, 103, 104, 105, 106, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 134, 135, 137, 139, 140, 141, 142, 153, 154], "bbox": [0.16195195195195194, 0.0022400000000000002, 0.9009009009009009, 0.70562], "iscrowd": 0, "area": 54133.19105, "rle": {"size": [500, 333], "counts": "ldj05_?8H9G9G8G10000000O100000O1000O100000O1gMlNQFU1e9YOUFh0a9HYF9]95]FM`9:[FFd9`0VF@j9f0QFZOn9l0lETOT:S1fElNZ:\\1_EdN_:e1YE\\Nf:Z3N1O200O10O1000O01000O10O1000O10O10O10O10OGZEfKd:Z4aEcK]:^4fE_KX:c4:2N3QOP1H8H7J7H7J7J6K4M4K5L3M4K4M2M2O3L5M3M4M1O1O1O1N2O1O1O2M2O1OO10O10O10000002N3N1N2cIoFT6T9iImFW6T9gImFX6X9O010O100O1O100O00100O1O100O01O100O1O100O11O0000001O000000001O000000001O001O001O001O001O001O001O001O001O1N2O1O1O2N1O1O1O2N0000O1000000000001N2O1O2N1O2N1O2M3N1O2N1O2N1O2N1N3O0O2N101N1O2O2M6J7I6K6N1O1eHgIQ5^800O1O100O0010O01O10O010O01O1O001O00100O001O2N1O2N2O0WMTFHn91]FIe90fFJ[90`GZOb8?eG@\\89jGHV81QHNV8EQH:X8WOoGh0Y8jNnGU1Y;O3L3M3M3L4M6J2N2K6H_i?"}, "label": "a girl smiling at the camera with silverware in hands"}]}
{"id": 405324, "image": "COCO_train2014_000000405324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women was smilling\"."}], "task": "refering", "answer_template": "The \"a women was smilling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 40, 41, 42, 43, 44, 45, 52, 53, 54, 55, 56, 57, 63, 64, 65, 66, 67, 68, 69, 74, 75, 76, 77, 78, 79, 80, 81, 82, 86, 87, 88, 89, 90, 91, 92, 93, 94, 98, 99, 100, 101, 102, 103, 104, 105, 106, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 134, 135, 137, 139, 140, 141, 142, 153, 154], "bbox": [0.16195195195195194, 0.0022400000000000002, 0.9009009009009009, 0.70562], "iscrowd": 0, "area": 54133.19105, "rle": {"size": [500, 333], "counts": "ldj05_?8H9G9G8G10000000O100000O1000O100000O1gMlNQFU1e9YOUFh0a9HYF9]95]FM`9:[FFd9`0VF@j9f0QFZOn9l0lETOT:S1fElNZ:\\1_EdN_:e1YE\\Nf:Z3N1O200O10O1000O01000O10O1000O10O10O10O10OGZEfKd:Z4aEcK]:^4fE_KX:c4:2N3QOP1H8H7J7H7J7J6K4M4K5L3M4K4M2M2O3L5M3M4M1O1O1O1N2O1O1O2M2O1OO10O10O10000002N3N1N2cIoFT6T9iImFW6T9gImFX6X9O010O100O1O100O00100O1O100O01O100O1O100O11O0000001O000000001O000000001O001O001O001O001O001O001O001O001O1N2O1O1O2N1O1O1O2N0000O1000000000001N2O1O2N1O2N1O2M3N1O2N1O2N1O2N1N3O0O2N101N1O2O2M6J7I6K6N1O1eHgIQ5^800O1O100O0010O01O10O010O01O1O001O00100O001O2N1O2N2O0WMTFHn91]FIe90fFJ[90`GZOb8?eG@\\89jGHV81QHNV8EQH:X8WOoGh0Y8jNnGU1Y;O3L3M3M3L4M6J2N2K6H_i?"}, "label": "a women was smilling"}]}
{"id": 405324, "image": "COCO_train2014_000000405324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty wine tumbler with out wine side position of the eating one lady\"."}], "task": "refering", "answer_template": "The \"empty wine tumbler with out wine side position of the eating one lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 155, 166, 167, 178, 179, 191, 202, 203, 214, 215], "bbox": [0.8478078078078077, 0.64812, 1.0, 1.0], "iscrowd": 0, "area": 4694.483950000001, "rle": {"size": [500, 333], "counts": "SUZ41a?2N2N2O2N1O2N1O2N100O2O1O1N2O1RMEaF<[95WFMe98WFJe9;YFGb9?[FC`9b0^FA]9d0`FAX9d0fF\\OW9i0gFWOW9l0hFTOU9Q1iFoNV9S1iFmNU9W1iFiNV9Y1jFfNU9\\1jFdNT9`1jF`NU9b1jF^NT9f1fF^NY9d1aFaN^9c1[FaNe9b1RFdNm9b30000O1000jNlElLS:R3QFmLo9P3TFPMl9m2XFRMg9l2]FSMc9j2aFUM_9h2dFXM[9g2hFXMX9g2[GgL"}, "label": "empty wine tumbler with out wine side position of the eating one lady"}]}
{"id": 151371, "image": "COCO_train2014_000000151371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unknown blue car in front of a truck\"."}], "task": "refering", "answer_template": "The \"an unknown blue car in front of a truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.002875, 0.6494613583138173, 0.671484375, 1.0], "iscrowd": 0, "area": 38241.927899999995, "rle": {"size": [427, 640], "counts": "^S1b4h8100O100O11O001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000000000001O0000000000000000000000001O00000000000000000000001O00000000000000000000001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O00001O001O00001O001O00001O001O00001O001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O001O00001O001O00001O00001O00001O00001O001O00001O00000010O000000000001O00000000001O000000001O0000001O0000001O0000001O0000001O0000000000001O000000000000001O0000000000001O0000000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O0000001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O2N3M3M2N3M2N3M2NlTg2"}, "label": "an unknown blue car in front of a truck"}]}
{"id": 151371, "image": "COCO_train2014_000000151371.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car\"."}], "task": "refering", "answer_template": "The \"a blue car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.002875, 0.6494613583138173, 0.671484375, 1.0], "iscrowd": 0, "area": 38241.927899999995, "rle": {"size": [427, 640], "counts": "^S1b4h8100O100O11O001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000000000001O0000000000000000000000001O00000000000000000000001O00000000000000000000001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O00001O001O00001O001O00001O001O00001O001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O001O00001O001O00001O00001O00001O00001O001O00001O00000010O000000000001O00000000001O000000001O0000001O0000001O0000001O0000001O0000000000001O000000000000001O0000000000001O0000000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O000000001O0000001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O2N3M3M2N3M2N3M2NlTg2"}, "label": "a blue car"}]}
{"id": 192337, "image": "COCO_train2014_000000192337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing with some grass on its mouth\"."}], "task": "refering", "answer_template": "The \"a zebra standing with some grass on its mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 105, 106, 108, 109, 110, 111, 112, 113, 128, 131, 132, 133, 135, 136, 153, 154, 155, 156, 158, 159, 176, 177, 178, 179, 181, 182], "bbox": [0.573765625, 0.09199170124481329, 0.961375, 0.9136514522821577], "iscrowd": 0, "area": 26260.21780000002, "rle": {"size": [241, 640], "counts": "g_f22\\77I7J5J7]N\\OnKj0k3_OmKe0n3a1N2O2M2N2O2M2N1O1001O10O01O1O1O010O1N101O1N2O1O1N2O1O1N2O1N2O1O1O100O1O100O1O1O2O0O1O1000000O100000000000000O1000000000000O10cMoMZOQ2c0RN]On1`0UN@k1=XNCh1:[NFg15]NJd11`NOb1JcN6^1EfN;\\1^OiNb0X1YOlNg0V1ROoNn0S1kNROV1o0cNWO\\1W1PNoNP2Y3000O1O1N2N2N2O1N2N2N2M3M3M3L4M3M3M3L4N2N2N1O2N2M3E;@`01O001O0O2O001O001O6J5K6jL\\MS1i2cN^M[1h2ZN_Mc1g2SN`Mk1V4M2N2NO1O1O1O1O1UMUNOk10YNMh12[NHi18YN[OR2e0PNTNU3l1mLjM[3V2gL`Ma3`2h000000O10000000O1000O10000000000O100000000O100000000O010000000O10000000000O100000000O1000O1000O10O1O100O1O010O100O1O010OfM[MJf23]MLg2N\\M1h2JZM5i2FZM9j2AYM>j2]OXMc0l2WOUMj0o2PORMQ1Q3jNPMW1T3cNmL^1V3ZNnLg1V3PNnLQ2X41O1N101O1O1O1O1O1N3N2N2N2N2N2H8E;J6=C=D;D9H1NN2O2M2O2WLhNT1Z1iNnNQ1T1kNUOm0m0PO232G>KILa0KGLb0Mei5"}, "label": "a zebra standing with some grass on its mouth"}]}
{"id": 192337, "image": "COCO_train2014_000000192337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra facing left\"."}], "task": "refering", "answer_template": "The \"a zebra facing left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 105, 106, 108, 109, 110, 111, 112, 113, 128, 131, 132, 133, 135, 136, 153, 154, 155, 156, 158, 159, 176, 177, 178, 179, 181, 182], "bbox": [0.573765625, 0.09199170124481329, 0.961375, 0.9136514522821577], "iscrowd": 0, "area": 26260.21780000002, "rle": {"size": [241, 640], "counts": "g_f22\\77I7J5J7]N\\OnKj0k3_OmKe0n3a1N2O2M2N2O2M2N1O1001O10O01O1O1O010O1N101O1N2O1O1N2O1O1N2O1N2O1O1O100O1O100O1O1O2O0O1O1000000O100000000000000O1000000000000O10cMoMZOQ2c0RN]On1`0UN@k1=XNCh1:[NFg15]NJd11`NOb1JcN6^1EfN;\\1^OiNb0X1YOlNg0V1ROoNn0S1kNROV1o0cNWO\\1W1PNoNP2Y3000O1O1N2N2N2O1N2N2N2M3M3M3L4M3M3M3L4N2N2N1O2N2M3E;@`01O001O0O2O001O001O6J5K6jL\\MS1i2cN^M[1h2ZN_Mc1g2SN`Mk1V4M2N2NO1O1O1O1O1UMUNOk10YNMh12[NHi18YN[OR2e0PNTNU3l1mLjM[3V2gL`Ma3`2h000000O10000000O1000O10000000000O100000000O100000000O010000000O10000000000O100000000O1000O1000O10O1O100O1O010O100O1O010OfM[MJf23]MLg2N\\M1h2JZM5i2FZM9j2AYM>j2]OXMc0l2WOUMj0o2PORMQ1Q3jNPMW1T3cNmL^1V3ZNnLg1V3PNnLQ2X41O1N101O1O1O1O1O1N3N2N2N2N2N2H8E;J6=C=D;D9H1NN2O2M2O2WLhNT1Z1iNnNQ1T1kNUOm0m0PO232G>KILa0KGLb0Mei5"}, "label": "a zebra facing left"}]}
{"id": 192337, "image": "COCO_train2014_000000192337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra eating grass with another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra eating grass with another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 28, 29, 30, 31, 32, 33, 47, 48, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 97, 98, 99, 100, 101, 102, 103, 104, 116, 117, 120, 121, 122, 123, 126, 127, 139, 144, 145, 146, 147, 149, 150, 162, 167, 168, 169, 170, 172, 173], "bbox": [0.03825, 0.11004149377593361, 0.561875, 0.8888381742738589], "iscrowd": 0, "area": 26556.808100000002, "rle": {"size": [241, 640], "counts": "hn58V79H7ZIAn5`0SJJa58^J4U5NkJ6n4LSK4k4NUK1j40VK1i40WKOh43XKMg43ZKLe46ZKJe47\\KIc48]KEf0[Oe2R1eLA>FGNW2k0dM_O90HNU2e0jM[O?MB5S2d0lMWOf0L[O:P2e0oMSOi0LXO=o1e0oMPOk00VO=n1e0PNlNl03WO=k1e0RNjNk06WO=j1d0TNhNk01^Oc0b1e0TNgNk0OBg0\\1f0VNcNd08K?Z1g0WNbNb09M]OJa0`1W1WNaN`0<f0Gb0\\1XNaN?=m0A;b1YN`N:a0V1ZO6f1ZN^N8d0]1TO0j1\\N]N6f0[3n0^L\\N7f0[3o0^L[N5g0^3m0^L[N3i0_3m0^LYN2k0a3l0]LYN2j0b3m0\\LXN3j0b3m0\\LXN3j0a3o0\\LWN4h0a3Q1[LVN5g0b3S1YLUN6f0c3T1XLUN6f0b3V1XLTN6d0d3X1VLSN7c0e3Z1eLeN\\3Z1eL_Nl3W1TLaNjg??]]@l0WOi0N2N2N2N2N2N2M3N2N1O2O1O1O1O1N2O1O1O1O100000000000O1000O10000000000000O100O10000O100O1000ZM\\M?c2]OdMa0\\2[OgMf0X2WOlMi0T2SOoMn0Q2oNQNR1n1lNUNT1k1iNWNX1h1eN[N\\1e1aN^N_1b1^N`Nc1_1YNeNh1[1RNjNo1U1lMQOT2o0fMhNh2i28L4M3N2N2N2M3N2N2M3N2M3M3N2M3N2N101N2N2O1N2O1002N3M2M4M2N2M4M2N3L5L5K4K6K5K5J6K4L5K5K5K2N001O1O1O^MaNRO_1m0[O[Nd0e1ESN:l10lMOS25iMLW27eMJZ29cMH]2;`ME_2>^MCb2?[MCc2`0ZMAd2h2N1O1N3N110O010O01O01O010O010O01O10O010O010O010O100O010O100O010O10O1000001N10001O0O101O00003M5K3M4L1O1O001O1O1O001O1O0O2O00001O00001O00001O00001O0000O1O1O1O1O1O1O1O1O1O1O1O2M2gMcK\\1^4dNeKX1\\4gNkKR1V4mNULg0n3WOULf0l3ZOVLc0k3\\OXL=m3BVL7n5D4K5LknQ2"}, "label": "a zebra eating grass with another zebra"}]}
{"id": 192337, "image": "COCO_train2014_000000192337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra , facing right is partially hidden behind a tree\"."}], "task": "refering", "answer_template": "The \"a zebra , facing right is partially hidden behind a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 28, 29, 30, 31, 32, 33, 47, 48, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 97, 98, 99, 100, 101, 102, 103, 104, 116, 117, 120, 121, 122, 123, 126, 127, 139, 144, 145, 146, 147, 149, 150, 162, 167, 168, 169, 170, 172, 173], "bbox": [0.03825, 0.11004149377593361, 0.561875, 0.8888381742738589], "iscrowd": 0, "area": 26556.808100000002, "rle": {"size": [241, 640], "counts": "hn58V79H7ZIAn5`0SJJa58^J4U5NkJ6n4LSK4k4NUK1j40VK1i40WKOh43XKMg43ZKLe46ZKJe47\\KIc48]KEf0[Oe2R1eLA>FGNW2k0dM_O90HNU2e0jM[O?MB5S2d0lMWOf0L[O:P2e0oMSOi0LXO=o1e0oMPOk00VO=n1e0PNlNl03WO=k1e0RNjNk06WO=j1d0TNhNk01^Oc0b1e0TNgNk0OBg0\\1f0VNcNd08K?Z1g0WNbNb09M]OJa0`1W1WNaN`0<f0Gb0\\1XNaN?=m0A;b1YN`N:a0V1ZO6f1ZN^N8d0]1TO0j1\\N]N6f0[3n0^L\\N7f0[3o0^L[N5g0^3m0^L[N3i0_3m0^LYN2k0a3l0]LYN2j0b3m0\\LXN3j0b3m0\\LXN3j0a3o0\\LWN4h0a3Q1[LVN5g0b3S1YLUN6f0c3T1XLUN6f0b3V1XLTN6d0d3X1VLSN7c0e3Z1eLeN\\3Z1eL_Nl3W1TLaNjg??]]@l0WOi0N2N2N2N2N2N2M3N2N1O2O1O1O1O1N2O1O1O1O100000000000O1000O10000000000000O100O10000O100O1000ZM\\M?c2]OdMa0\\2[OgMf0X2WOlMi0T2SOoMn0Q2oNQNR1n1lNUNT1k1iNWNX1h1eN[N\\1e1aN^N_1b1^N`Nc1_1YNeNh1[1RNjNo1U1lMQOT2o0fMhNh2i28L4M3N2N2N2M3N2N2M3N2M3M3N2M3N2N101N2N2O1N2O1002N3M2M4M2N2M4M2N3L5L5K4K6K5K5J6K4L5K5K5K2N001O1O1O^MaNRO_1m0[O[Nd0e1ESN:l10lMOS25iMLW27eMJZ29cMH]2;`ME_2>^MCb2?[MCc2`0ZMAd2h2N1O1N3N110O010O01O01O010O010O01O10O010O010O010O100O010O100O010O10O1000001N10001O0O101O00003M5K3M4L1O1O001O1O1O001O1O0O2O00001O00001O00001O00001O0000O1O1O1O1O1O1O1O1O1O1O1O2M2gMcK\\1^4dNeKX1\\4gNkKR1V4mNULg0n3WOULf0l3ZOVLc0k3\\OXL=m3BVL7n5D4K5LknQ2"}, "label": "a zebra , facing right is partially hidden behind a tree"}]}
{"id": 257874, "image": "COCO_train2014_000000257874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is the left kite , in the hands of a woman\"."}], "task": "refering", "answer_template": "The \"this is the left kite , in the hands of a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 97, 98, 99, 115, 116, 117, 133, 134, 152, 169, 170, 187], "bbox": [0.40696, 0.38006006006006005, 0.54162, 0.8759159159159159], "iscrowd": 0, "area": 3189.582349999999, "rle": {"size": [333, 500], "counts": "lcR22Y:2M3N2VMH]K9a4L[K6\\3FUK8\\14^3FRK;[12b3EoJ>[1Oe3EnJ?Y1Nh3DoJ`0U1Oj3CPK`0Q10n3BPK`0n01P4APK`0m0LW4FkJ`0j0G`4JeJ?i0Ee4MbJ=f0Fk4N_J=41a5CZJ>C<V6WOWJo1k5RNUJl1l5UNTJi1m5XNSJe1o5]NPJa1Q6`NnI_1S6bNmI\\1T6eNlI>Q7BoH7X7IiH5W7LiH2Y7NhH0Y70gH0Z7OfH0\\7OeHO^7ObH0`7O`H1b7M^H2d7M[H3g7LYH3j7KVH4l7KTH5n7JQH5Q8JnG6U8HkG7W8IhG7Z8GfG8\\8GdG8^8HaG7a8H^G9b8IZG8h8HUG9m8GPG;Q9ElF<V9<2N2O1N3M4M2M3M4L3N3L3M4M2M]YZ2"}, "label": "this is the left kite , in the hands of a woman"}]}
{"id": 257874, "image": "COCO_train2014_000000257874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cellophane kite and hand holding a rope\"."}], "task": "refering", "answer_template": "The \"an orange cellophane kite and hand holding a rope\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 97, 98, 99, 115, 116, 117, 133, 134, 152, 169, 170, 187], "bbox": [0.40696, 0.38006006006006005, 0.54162, 0.8759159159159159], "iscrowd": 0, "area": 3189.582349999999, "rle": {"size": [333, 500], "counts": "lcR22Y:2M3N2VMH]K9a4L[K6\\3FUK8\\14^3FRK;[12b3EoJ>[1Oe3EnJ?Y1Nh3DoJ`0U1Oj3CPK`0Q10n3BPK`0n01P4APK`0m0LW4FkJ`0j0G`4JeJ?i0Ee4MbJ=f0Fk4N_J=41a5CZJ>C<V6WOWJo1k5RNUJl1l5UNTJi1m5XNSJe1o5]NPJa1Q6`NnI_1S6bNmI\\1T6eNlI>Q7BoH7X7IiH5W7LiH2Y7NhH0Y70gH0Z7OfH0\\7OeHO^7ObH0`7O`H1b7M^H2d7M[H3g7LYH3j7KVH4l7KTH5n7JQH5Q8JnG6U8HkG7W8IhG7Z8GfG8\\8GdG8^8HaG7a8H^G9b8IZG8h8HUG9m8GPG;Q9ElF<V9<2N2O1N3M4M2M3M4L3N3L3M4M2M]YZ2"}, "label": "an orange cellophane kite and hand holding a rope"}]}
{"id": 208724, "image": "COCO_train2014_000000208724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white zebra with his head near the ground\"."}], "task": "refering", "answer_template": "The \"a black and white zebra with his head near the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 195, 196, 197, 198, 213, 214, 215, 219, 220, 236, 238, 241, 242, 243, 244, 266, 267], "bbox": [0.262203125, 0.2548125, 0.62803125, 0.6625208333333333], "iscrowd": 0, "area": 25403.797499999993, "rle": {"size": [480, 640], "counts": "go^2[1k<hNfCe1Q<eN`Ch1W<h0G3N1L5G9G9G9L4O001O1N2O1O1O0O2O1OL4J5K6J6O10O0100O1O010O100O010O1TOPEhMQ;X2XE_Mg:c2j0001O0O2O001O1No0RO5L4L4K5L4L4K6K4K9H000O10000O10000O100O11O2SO_FWLb9e3eFWL\\9e3kFVLW9e3PGXLP9e3XGVLi8f3Z1M3L4N2N3M2N2N2N3M2O10001N10000O101O0O10001O0O10001N10000O101O000O101O000O10001O0O10001O000O1000000O0100000O0100O0100O00100O010O10O01O10O010O0100O03N1O1N2O2N1N2aFUMg6m2UIVMj6k2TIVMm6j2QIXMn6j2oHWMQ7j2mHXMS7h2jHZMV7h2fH[MZ7i2^HZMb7k2VHYMi7l2oGXMQ8m2fGVMZ8o2^GUMa8Q3VGPMm8W3hFkLY9\\42O2N1N2N3L3M3M4L3M4L3M3M4M2M3N3ROVEaMm:`2UE[Mm:e2WETMm:l2[EiLg:X3`010O01O100O00nNaDdN_;U1hDkNW;P1oDPOQ;j0UEVOj:e0\\E[Oc:a0cE^O]:>gECW:9mEHS:4QFMo9MVF4i9G[F:f9@_FX1j8eNXG^1e8dNXG_1g8cNVG_1h8dNUG_1j8bNRGa1n8`NoFb1Q9_NlFd1S9\\NjFg1V9kNVFU1l9lNQFT1Q:lNlET1W:mNPFh0S:B]FAf9d0gFgN\\9W1cFiN`9U1_FkNd9S1[FmNh9Q1VFPOm9n0RFQOS:l0lESOZ:i0eEVO`:f0`EROl:g0WEQOU;h0g1I7I7IhZ_3"}, "label": "a black and white zebra with his head near the ground"}]}
{"id": 208724, "image": "COCO_train2014_000000208724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foremost zebra , in front of the others\"."}], "task": "refering", "answer_template": "The \"the foremost zebra , in front of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 195, 196, 197, 198, 213, 214, 215, 219, 220, 236, 238, 241, 242, 243, 244, 266, 267], "bbox": [0.262203125, 0.2548125, 0.62803125, 0.6625208333333333], "iscrowd": 0, "area": 25403.797499999993, "rle": {"size": [480, 640], "counts": "go^2[1k<hNfCe1Q<eN`Ch1W<h0G3N1L5G9G9G9L4O001O1N2O1O1O0O2O1OL4J5K6J6O10O0100O1O010O100O010O1TOPEhMQ;X2XE_Mg:c2j0001O0O2O001O1No0RO5L4L4K5L4L4K6K4K9H000O10000O10000O100O11O2SO_FWLb9e3eFWL\\9e3kFVLW9e3PGXLP9e3XGVLi8f3Z1M3L4N2N3M2N2N2N3M2O10001N10000O101O0O10001O0O10001N10000O101O000O101O000O10001O0O10001O000O1000000O0100000O0100O0100O00100O010O10O01O10O010O0100O03N1O1N2O2N1N2aFUMg6m2UIVMj6k2TIVMm6j2QIXMn6j2oHWMQ7j2mHXMS7h2jHZMV7h2fH[MZ7i2^HZMb7k2VHYMi7l2oGXMQ8m2fGVMZ8o2^GUMa8Q3VGPMm8W3hFkLY9\\42O2N1N2N3L3M3M4L3M4L3M3M4M2M3N3ROVEaMm:`2UE[Mm:e2WETMm:l2[EiLg:X3`010O01O100O00nNaDdN_;U1hDkNW;P1oDPOQ;j0UEVOj:e0\\E[Oc:a0cE^O]:>gECW:9mEHS:4QFMo9MVF4i9G[F:f9@_FX1j8eNXG^1e8dNXG_1g8cNVG_1h8dNUG_1j8bNRGa1n8`NoFb1Q9_NlFd1S9\\NjFg1V9kNVFU1l9lNQFT1Q:lNlET1W:mNPFh0S:B]FAf9d0gFgN\\9W1cFiN`9U1_FkNd9S1[FmNh9Q1VFPOm9n0RFQOS:l0lESOZ:i0eEVO`:f0`EROl:g0WEQOU;h0g1I7I7IhZ_3"}, "label": "the foremost zebra , in front of the others"}]}
{"id": 208724, "image": "COCO_train2014_000000208724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra that ' s farthest to the right with only its back showing\"."}], "task": "refering", "answer_template": "The \"the zebra that ' s farthest to the right with only its back showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 199, 200, 201, 222, 223, 224, 244, 245, 247, 270], "bbox": [0.5704687500000001, 0.2865625, 0.763421875, 0.6556875000000001], "iscrowd": 0, "area": 11463.265149999997, "rle": {"size": [480, 640], "counts": "mY[52e>:F9G:F9N3hNhNcD[1];gN_D[1a;hNYD\\1e;hNVDZ1j;hNQD\\1n;gNWDQ1i;QO`Dd0`;_OhD7W;LQEIo:9ZE]Oe:f0h14M2N3M2N3M3M2N3L3N3M3M2N3M3M2N3M2M4M3M2N0UF\\MZ7c2eH`MZ7_2fHbMZ7^2eHdMZ7[2eHgMZ7Y2fHhMZ7X2dHkM[7T2eHmM[7R2eHoM[7P2dHSN[7m1dHTN\\7k1cHWN]7h1cHWN^7i1aHWNa7i1^HVNd7i1\\HVNf7l1VHTNl7o1PHPNR8T2hGlMZ8W2bGhM`8R3eFlL^9Y42N2N3M2N2N2O1N3M2O1N2O1N3N1N2N2ROnDmMT;Q2XEbMj:]2m0O1O1O2O0O1O11O101N1O2N101N2N1O2N101N1O2N2N101N1O2N101N1O2N6PFZL\\8j3YG[Li8i3lF\\LV9h3_F]Lc9g3SF]Ln9Z42N2N2N2N2N1O7Ia0_O`0@a0_Oa0_OW2iMW_V2"}, "label": "the zebra that ' s farthest to the right with only its back showing"}]}
{"id": 208724, "image": "COCO_train2014_000000208724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back end of a zebra to the far right\"."}], "task": "refering", "answer_template": "The \"back end of a zebra to the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 199, 200, 201, 222, 223, 224, 244, 245, 247, 270], "bbox": [0.5704687500000001, 0.2865625, 0.763421875, 0.6556875000000001], "iscrowd": 0, "area": 11463.265149999997, "rle": {"size": [480, 640], "counts": "mY[52e>:F9G:F9N3hNhNcD[1];gN_D[1a;hNYD\\1e;hNVDZ1j;hNQD\\1n;gNWDQ1i;QO`Dd0`;_OhD7W;LQEIo:9ZE]Oe:f0h14M2N3M2N3M3M2N3L3N3M3M2N3M3M2N3M2M4M3M2N0UF\\MZ7c2eH`MZ7_2fHbMZ7^2eHdMZ7[2eHgMZ7Y2fHhMZ7X2dHkM[7T2eHmM[7R2eHoM[7P2dHSN[7m1dHTN\\7k1cHWN]7h1cHWN^7i1aHWNa7i1^HVNd7i1\\HVNf7l1VHTNl7o1PHPNR8T2hGlMZ8W2bGhM`8R3eFlL^9Y42N2N3M2N2N2O1N3M2O1N2O1N3N1N2N2ROnDmMT;Q2XEbMj:]2m0O1O1O2O0O1O11O101N1O2N101N2N1O2N101N1O2N2N101N1O2N101N1O2N6PFZL\\8j3YG[Li8i3lF\\LV9h3_F]Lc9g3SF]Ln9Z42N2N2N2N2N1O7Ia0_O`0@a0_Oa0_OW2iMW_V2"}, "label": "back end of a zebra to the far right"}]}
{"id": 208724, "image": "COCO_train2014_000000208724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra behind another zebra , nearest a wood structure\"."}], "task": "refering", "answer_template": "The \"zebra behind another zebra , nearest a wood structure\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 120, 121, 122, 143, 144, 166, 167, 189, 193, 212, 216, 235, 239, 258], "bbox": [0.21959374999999998, 0.2826666666666667, 0.4324375, 0.6722916666666667], "iscrowd": 0, "area": 4831.282, "rle": {"size": [480, 640], "counts": "iWR22m>4M3L4kAj0Z<o0N2O1iEjMa7X2\\HlMb7U2ZHQNb7P2\\HUNa7l1[HYNb7i1[H\\Nb7e1[H_Nb7c1ZH^Nh7c1UH[Nn7g1nGYNU8g1iGXNZ8g1eGXN^8g1`GXNd8g1[GXNh8g1VGYNm8f1RGYNQ9e1PGZNR9e1QGWNQ9h1QGTNR9k1QGQNQ9n1RGnMP9Q2SGkMo8S2UGiMm8T2Q2N2M3M3N2M3N2M3N2M3N2M3M2O2M3M2O2M3M2N3N2M2N3N2M201O001N2O001O1O001O1O001OWkc0=gS\\Oe0F:K5K5J5L5K5L41N101O1O0O2O1L4J5K6J6J5K6J6J6J5K6J6JRnY5"}, "label": "zebra behind another zebra , nearest a wood structure"}]}
{"id": 528213, "image": "COCO_train2014_000000528213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skating and making video\"."}], "task": "refering", "answer_template": "The \"a man skating and making video\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 138, 139, 140, 141, 156, 157, 158, 159, 175, 176, 193, 194], "bbox": [0.7042999999999999, 0.5325816023738872, 0.8481799999999999, 0.8854005934718101], "iscrowd": 0, "area": 4124.234600000001, "rle": {"size": [337, 500], "counts": "nQd32Y:60000UFKZ95eFMZ93eFOZ91fF0V82WHNB2V83UHKE3U85SHHG5T87RHDJ5T8:oGAL6U89oGAK6W8:mG@L5X8;lG@K6Y8:lG@K5Y8=kG^OK5\\8<iG_OJ6]8;iG_OI6_8<gG^OJ5a8<eG_OI6b8<dG^OI6d8=bG]OJ5`0Dc7k0QH[OL7>Eg0Nf5l0XIZOL7?Ge0Mh5l0WIYOM6>Kd0Kk5l0TIF;Fd0Im5l0SID<Hc0Io5^1]IlNb0FR6^1ZIoNb0DT6\\1ZIRO`0CW6Y1XIYO>^OZ6Y1WI^O:ZO_6W1VID7VOd6T1TIL3QOi6S1SI00mNn6R1QI6MiNS7o0QId0o6\\OPIe0P7ZOQIf0P7XOQIh0o6YOPIf0R7ZOlHg0T7[OjHe0W75kHnNV7P1mHnNS7R1PIkNQ7T1QIjNP7T1SIiNn6W1UIfNl6X1WIfNj6Y1YIdNg6\\1[IbNf6\\1^IaNc6^1_I_Nb6`1aI^N`6`1dI]N]6`1gI^NZ6`1\\1O2N2N2N3M3L5L3M3M3M3M4L3Feih0"}, "label": "a man skating and making video"}]}
{"id": 528213, "image": "COCO_train2014_000000528213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in shorts riding a skateboard in the street\"."}], "task": "refering", "answer_template": "The \"a person in shorts riding a skateboard in the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 138, 139, 140, 141, 156, 157, 158, 159, 175, 176, 193, 194], "bbox": [0.7042999999999999, 0.5325816023738872, 0.8481799999999999, 0.8854005934718101], "iscrowd": 0, "area": 4124.234600000001, "rle": {"size": [337, 500], "counts": "nQd32Y:60000UFKZ95eFMZ93eFOZ91fF0V82WHNB2V83UHKE3U85SHHG5T87RHDJ5T8:oGAL6U89oGAK6W8:mG@L5X8;lG@K6Y8:lG@K5Y8=kG^OK5\\8<iG_OJ6]8;iG_OI6_8<gG^OJ5a8<eG_OI6b8<dG^OI6d8=bG]OJ5`0Dc7k0QH[OL7>Eg0Nf5l0XIZOL7?Ge0Mh5l0WIYOM6>Kd0Kk5l0TIF;Fd0Im5l0SID<Hc0Io5^1]IlNb0FR6^1ZIoNb0DT6\\1ZIRO`0CW6Y1XIYO>^OZ6Y1WI^O:ZO_6W1VID7VOd6T1TIL3QOi6S1SI00mNn6R1QI6MiNS7o0QId0o6\\OPIe0P7ZOQIf0P7XOQIh0o6YOPIf0R7ZOlHg0T7[OjHe0W75kHnNV7P1mHnNS7R1PIkNQ7T1QIjNP7T1SIiNn6W1UIfNl6X1WIfNj6Y1YIdNg6\\1[IbNf6\\1^IaNc6^1_I_Nb6`1aI^N`6`1dI]N]6`1gI^NZ6`1\\1O2N2N2N3M3L5L3M3M3M3M4L3Feih0"}, "label": "a person in shorts riding a skateboard in the street"}]}
{"id": 274266, "image": "COCO_train2014_000000274266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing next to a young girl on a grassy hillside\"."}], "task": "refering", "answer_template": "The \"a man standing next to a young girl on a grassy hillside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 139, 140, 141, 157, 158, 175, 176, 177, 193, 194, 195, 212, 213, 230, 231], "bbox": [0.7519600000000001, 0.5247466666666667, 0.87538, 1.0], "iscrowd": 0, "area": 5875.3918, "rle": {"size": [375, 500], "counts": "PnY4>R;=dEXOX9P1`FWO]9n0[FXOb9b1K4M4L3M4K3N3C[McGg2[8[MaGi2]8YM`Gj2^89O1N3N1O2M2O1O2K4M4L8I9FQ1PO1N20000000000000^KVJW3k5gLZJU3f5iL`JS3a5kLcJR3]5lLiJP3X5nLmJn2S5QMPKm2Q5mLUK>TOU1g5\\NWK:[OU1_5^NXK:AS1Y5\\N\\K?@Q1U5_N\\K>BQ1n6nNVIo0k6QO[Ih0f6WOcI`0]6AiI8X6GlI5T6LmI2T6MoI0R6MSJ0m50WJIm56k2N101O006Jk]f0"}, "label": "a man standing next to a young girl on a grassy hillside"}]}
{"id": 274266, "image": "COCO_train2014_000000274266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black jacket\"."}], "task": "refering", "answer_template": "The \"a man in a black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 139, 140, 141, 157, 158, 175, 176, 177, 193, 194, 195, 212, 213, 230, 231], "bbox": [0.7519600000000001, 0.5247466666666667, 0.87538, 1.0], "iscrowd": 0, "area": 5875.3918, "rle": {"size": [375, 500], "counts": "PnY4>R;=dEXOX9P1`FWO]9n0[FXOb9b1K4M4L3M4K3N3C[McGg2[8[MaGi2]8YM`Gj2^89O1N3N1O2M2O1O2K4M4L8I9FQ1PO1N20000000000000^KVJW3k5gLZJU3f5iL`JS3a5kLcJR3]5lLiJP3X5nLmJn2S5QMPKm2Q5mLUK>TOU1g5\\NWK:[OU1_5^NXK:AS1Y5\\N\\K?@Q1U5_N\\K>BQ1n6nNVIo0k6QO[Ih0f6WOcI`0]6AiI8X6GlI5T6LmI2T6MoI0R6MSJ0m50WJIm56k2N101O006Jk]f0"}, "label": "a man in a black jacket"}]}
{"id": 487260, "image": "COCO_train2014_000000487260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple and green bus\"."}], "task": "refering", "answer_template": "The \"a purple and green bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 184, 185, 186, 207, 208, 209], "bbox": [0.0008125000000000001, 0.4388333333333333, 0.18924999999999997, 0.7634444444444444], "iscrowd": 0, "area": 9080.15235, "rle": {"size": [360, 640], "counts": "Z`0X3o72O0O2O1O2N3M1O0000000000000000000000000O10O100000000O1N2N2N2000000000000000000000000O1000000000000000000000O1000000000000000000000000000000001O7I1O001O1O6J6UOYb4TNX`K4L5K4L1O00000000O100000000000000O01000N2N2N2M3N2N2MkYf5"}, "label": "a purple and green bus"}]}
{"id": 487260, "image": "COCO_train2014_000000487260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bus with a \" buses \" logo on it\"."}], "task": "refering", "answer_template": "The \"a green bus with a \" buses \" logo on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 184, 185, 186, 207, 208, 209], "bbox": [0.0008125000000000001, 0.4388333333333333, 0.18924999999999997, 0.7634444444444444], "iscrowd": 0, "area": 9080.15235, "rle": {"size": [360, 640], "counts": "Z`0X3o72O0O2O1O2N3M1O0000000000000000000000000O10O100000000O1N2N2N2000000000000000000000000O1000000000000000000000O1000000000000000000000000000000001O7I1O001O1O6J6UOYb4TNX`K4L5K4L1O00000000O100000000000000O01000N2N2N2M3N2N2MkYf5"}, "label": "a green bus with a \" buses \" logo on it"}]}
{"id": 487260, "image": "COCO_train2014_000000487260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white double decker bus with green diagonal stripe on bottom following behind blue truck\"."}], "task": "refering", "answer_template": "The \"white double decker bus with green diagonal stripe on bottom following behind blue truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247], "bbox": [0.507, 0.18297222222222223, 0.8765468750000001, 0.7950555555555556], "iscrowd": 0, "area": 43576.67784999999, "rle": {"size": [360, 640], "counts": "hSb32d:b0^Ob0]Oc0^Ob0^Ob0^Ob0D=H7H8I7M30001O0000000000001O00000O1000001O2N2N1O0000000000000000001O0000001O00001O00001N10001O2N2N2N2N2N6J6J7I6J6J4L1O00000000000000001O00000O100000001O000000000000001O00000000000000001O000000000O10001O00000000000000001O000000000000001O01N10001O00001O001O00001O000O2O00001O001O00010O01O10O0O2O0O1O2O0O1O2O0O1O2O0O2O00001O000O2O00001O001O0O101O00001O000O2O001O00001O0O101O00001O001N10001O00001O00001O001O00001O00001O000O2N1O2O0O1O2O0O101N100O2O0O2O0O108G?@a0_O`0^NPek0"}, "label": "white double decker bus with green diagonal stripe on bottom following behind blue truck"}]}
{"id": 487260, "image": "COCO_train2014_000000487260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white two story bus with a blue tow truck in front of it\"."}], "task": "refering", "answer_template": "The \"a green and white two story bus with a blue tow truck in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247], "bbox": [0.507, 0.18297222222222223, 0.8765468750000001, 0.7950555555555556], "iscrowd": 0, "area": 43576.67784999999, "rle": {"size": [360, 640], "counts": "hSb32d:b0^Ob0]Oc0^Ob0^Ob0^Ob0D=H7H8I7M30001O0000000000001O00000O1000001O2N2N1O0000000000000000001O0000001O00001O00001N10001O2N2N2N2N2N6J6J7I6J6J4L1O00000000000000001O00000O100000001O000000000000001O00000000000000001O000000000O10001O00000000000000001O000000000000001O01N10001O00001O001O00001O000O2O00001O001O00010O01O10O0O2O0O1O2O0O1O2O0O1O2O0O2O00001O000O2O00001O001O0O101O00001O000O2O001O00001O0O101O00001O001N10001O00001O00001O001O00001O00001O000O2N1O2O0O1O2O0O101N100O2O0O2O0O108G?@a0_O`0^NPek0"}, "label": "a green and white two story bus with a blue tow truck in front of it"}]}
{"id": 315229, "image": "COCO_train2014_000000315229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue bike in front of the umbrella\"."}], "task": "refering", "answer_template": "The \"the blue bike in front of the umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 102, 103, 115, 116, 128, 129, 141, 142, 154, 155, 167, 168, 181], "bbox": [0.8412, 0.33606, 0.9613866666666666, 0.74592], "iscrowd": 0, "area": 6581.47675, "rle": {"size": [500, 375], "counts": "iRj41^?;D<D<D<E;E9G8H7H9E;E:G:E;K4L5K4L3M4L4L3M4L4L3M4L4L3M4N0O2O0O2O1N101N2O0O2O0O2O0O2O0O10oIXGW5f8`JeL?kh6"}, "label": "the blue bike in front of the umbrella"}]}
{"id": 315229, "image": "COCO_train2014_000000315229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike wheel second from camera\"."}], "task": "refering", "answer_template": "The \"bike wheel second from camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 76, 77, 88, 89, 90, 101, 102, 114, 115, 127, 128], "bbox": [0.8143466666666667, 0.2199, 0.9720533333333333, 0.54892], "iscrowd": 0, "area": 4356.620550000002, "rle": {"size": [500, 375], "counts": "dTe43Z?;E:F;E:F;H7I8H7I8H7I?A0ROgLeEX3`:RMREn2Q;g04K5L3M4L3L5L4L3L5L4L3L5L3M4L4O001O1O001O1O001O001O1O001O1O001O001O1O001O1N1O2N1N3J6@?A`0@Zi4"}, "label": "bike wheel second from camera"}]}
{"id": 315229, "image": "COCO_train2014_000000315229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike tire with silver and black bumper that is second from the bottom of the row\"."}], "task": "refering", "answer_template": "The \"bike tire with silver and black bumper that is second from the bottom of the row\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 76, 77, 88, 89, 90, 101, 102, 114, 115, 127, 128], "bbox": [0.8143466666666667, 0.2199, 0.9720533333333333, 0.54892], "iscrowd": 0, "area": 4356.620550000002, "rle": {"size": [500, 375], "counts": "dTe43Z?;E:F;E:F;H7I8H7I8H7I?A0ROgLeEX3`:RMREn2Q;g04K5L3M4L3L5L4L3L5L4L3L5L3M4L4O001O1O001O1O001O001O1O001O1O001O001O1O001O1N1O2N1N3J6@?A`0@Zi4"}, "label": "bike tire with silver and black bumper that is second from the bottom of the row"}]}
{"id": 94045, "image": "COCO_train2014_000000094045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rear end of the giraffe that is partially blocked by giraffe in foreground\"."}], "task": "refering", "answer_template": "The \"the rear end of the giraffe that is partially blocked by giraffe in foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 111, 112, 113, 125, 126, 127, 138, 139, 140, 151, 152, 153, 164, 165, 166, 177, 178, 179, 190, 191, 203, 204, 217], "bbox": [0.5573066666666667, 0.40188, 0.82696, 0.8985], "iscrowd": 0, "area": 9026.249549999995, "rle": {"size": [500, 375], "counts": "cgV32c?2O1N2N2O1N2N2O1N2O1N2N10O01O010O001N2O0O2O0O2O0O2O1N101N2O:E;E:G:WFgMa5o2gITMY6P3_ITM`6P3YITMg6P3QITMn6P3kHSMV7Q3bHSM]7Q3\\HRMe7R3lGWMU8m2\\G^Me8f2kFdMV9]4001N101N101O0O2O001J5G9G:F9G:F9H9OcFPLV7U4bHnK_7U4[HmKe7X4SHkKn7X4kGkKU8Z4cGjK]8Y4]GiKc8\\4UGgKk8]4nFfKS9a1iF<LVN[9Y1RG\\1o8^NZG^1f8]NcGBSO<[9LkGZO]Oa0h80UHPOFg0V83]HhN0l0d76iJ]O\\5=QKWOU5b0R5J7L310OoBGf:9VE1d:OXE:d:FVEe0d:[OXEo0c:POYEY1b:gNaDX2];o02O0O2N101Nk1VN7D<B>B^\\o0"}, "label": "the rear end of the giraffe that is partially blocked by giraffe in foreground"}]}
{"id": 94045, "image": "COCO_train2014_000000094045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture showing backside of a giraffe\"."}], "task": "refering", "answer_template": "The \"a picture showing backside of a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 111, 112, 113, 125, 126, 127, 138, 139, 140, 151, 152, 153, 164, 165, 166, 177, 178, 179, 190, 191, 203, 204, 217], "bbox": [0.5573066666666667, 0.40188, 0.82696, 0.8985], "iscrowd": 0, "area": 9026.249549999995, "rle": {"size": [500, 375], "counts": "cgV32c?2O1N2N2O1N2N2O1N2O1N2N10O01O010O001N2O0O2O0O2O0O2O1N101N2O:E;E:G:WFgMa5o2gITMY6P3_ITM`6P3YITMg6P3QITMn6P3kHSMV7Q3bHSM]7Q3\\HRMe7R3lGWMU8m2\\G^Me8f2kFdMV9]4001N101N101O0O2O001J5G9G:F9G:F9H9OcFPLV7U4bHnK_7U4[HmKe7X4SHkKn7X4kGkKU8Z4cGjK]8Y4]GiKc8\\4UGgKk8]4nFfKS9a1iF<LVN[9Y1RG\\1o8^NZG^1f8]NcGBSO<[9LkGZO]Oa0h80UHPOFg0V83]HhN0l0d76iJ]O\\5=QKWOU5b0R5J7L310OoBGf:9VE1d:OXE:d:FVEe0d:[OXEo0c:POYEY1b:gNaDX2];o02O0O2N101Nk1VN7D<B>B^\\o0"}, "label": "a picture showing backside of a giraffe"}]}
{"id": 94045, "image": "COCO_train2014_000000094045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shorter giraffe\"."}], "task": "refering", "answer_template": "The \"a shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 54, 55, 56, 68, 69, 82, 95], "bbox": [0.17037333333333332, 0.16848, 0.37687999999999994, 0.42524], "iscrowd": 0, "area": 3464.3369500000003, "rle": {"size": [500, 375], "counts": "V\\o01b?3L3N2N1O1N2O001O1O0O2O1O001N2O1O001N2FXOgAi0Y>[ObAe0`>7000O1001H7K5O2O000O2O001O02dNmAU1S>hNmA[1S>4M3M3M3McNYBQ1c=oN`BR1]=nNdBT1Y=kNiBW1U=hNlBZ1T=bNnB`1R=]NoBe1Q=WNPCl1\\=1NXN]B`1b=^NaBc1]=[NfBf1d=00100O1O001O1O001O100O001O1O1O001O1O010O1O1O0iSb3"}, "label": "a shorter giraffe"}]}
{"id": 94045, "image": "COCO_train2014_000000094045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe behind other giraffe\"."}], "task": "refering", "answer_template": "The \"giraffe behind other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 54, 55, 56, 68, 69, 82, 95], "bbox": [0.17037333333333332, 0.16848, 0.37687999999999994, 0.42524], "iscrowd": 0, "area": 3464.3369500000003, "rle": {"size": [500, 375], "counts": "V\\o01b?3L3N2N1O1N2O001O1O0O2O1O001N2O1O001N2FXOgAi0Y>[ObAe0`>7000O1001H7K5O2O000O2O001O02dNmAU1S>hNmA[1S>4M3M3M3McNYBQ1c=oN`BR1]=nNdBT1Y=kNiBW1U=hNlBZ1T=bNnB`1R=]NoBe1Q=WNPCl1\\=1NXN]B`1b=^NaBc1]=[NfBf1d=00100O1O001O1O001O100O001O1O1O001O1O010O1O1O0iSb3"}, "label": "giraffe behind other giraffe"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a doughnut sitting above others\"."}], "task": "refering", "answer_template": "The \"a doughnut sitting above others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 35, 36, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219], "bbox": [0.192953125, 0.024833333333333332, 0.6543593750000001, 0.5930625], "iscrowd": 0, "area": 63164.10039999998, "rle": {"size": [480, 640], "counts": "Wgi13l>n0QOP1QOn0RO;E3L4M3M3M3L4M3M3M3M3L4M2N3M3L4M3M3M3M3L4M3M3M3L4M3M3M3L4N2N2M3N1O2M3N2N2M101O001N10001N101O0O101O0O2O00001N101O0O101O0O2O000O2O001O0O101O0O2O001N10001N101O000O2O001N1000001O0000001O0000001O0000001O000000001N1000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001N1000001O000000001O000000001O00000000001O01O00001O01O0001O00010O00001O00010O00001O01O01O00001O01O01O00001O00001N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O100O2N100O3M3N2M3M2O2M3M3N2M3M3M3N2M3M3M3M2N3M3N2M3M3M3M3M3M3N2M3M3M4L3M3M4M2M3M3M4L3M3M3N3L3M3M3M4L3M3N3I6H8I7H9G8H8I7H9G8H8IX^W3"}, "label": "a doughnut sitting above others"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a seasoned bagel is piled atop other bagels on a plate\"."}], "task": "refering", "answer_template": "The \"a seasoned bagel is piled atop other bagels on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 35, 36, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219], "bbox": [0.192953125, 0.024833333333333332, 0.6543593750000001, 0.5930625], "iscrowd": 0, "area": 63164.10039999998, "rle": {"size": [480, 640], "counts": "Wgi13l>n0QOP1QOn0RO;E3L4M3M3M3L4M3M3M3M3L4M2N3M3L4M3M3M3M3L4M3M3M3L4M3M3M3L4N2N2M3N1O2M3N2N2M101O001N10001N101O0O101O0O2O00001N101O0O101O0O2O000O2O001O0O101O0O2O001N10001N101O000O2O001N1000001O0000001O0000001O0000001O000000001N1000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001N1000001O000000001O000000001O00000000001O01O00001O01O0001O00010O00001O00010O00001O01O01O00001O01O01O00001O00001N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O100O2N100O3M3N2M3M2O2M3M3N2M3M3M3N2M3M3M3M2N3M3N2M3M3M3M3M3M3N2M3M3M4L3M3M4M2M3M3M4L3M3M3N3L3M3M3M4L3M3N3I6H8I7H9G8H8I7H9G8H8IX^W3"}, "label": "a seasoned bagel is piled atop other bagels on a plate"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"coffee cup with floral design\"."}], "task": "refering", "answer_template": "The \"coffee cup with floral design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90], "bbox": [0.7213437500000001, 0.0022500000000000003, 0.97921875, 0.22247916666666664], "iscrowd": 0, "area": 14489.382350000003, "rle": {"size": [480, 640], "counts": "`bh6b1o<`0O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O101N1O00001O0000001O000000001O0000001O0000001O000000001O0000001O00000000000000001O01O0000000000000001O00000000000000001O0000O10000O100O10000O10000O100O10000O100O10000O2O0O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1000O0100O013M3L4M3L4M3M3L4M3M3L4M3L4M3L4G9F:GaQ6"}, "label": "coffee cup with floral design"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white cup with the bamboo painted on it\"."}], "task": "refering", "answer_template": "The \"the white cup with the bamboo painted on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90], "bbox": [0.7213437500000001, 0.0022500000000000003, 0.97921875, 0.22247916666666664], "iscrowd": 0, "area": 14489.382350000003, "rle": {"size": [480, 640], "counts": "`bh6b1o<`0O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O101N1O00001O0000001O000000001O0000001O0000001O000000001O0000001O00000000000000001O01O0000000000000001O00000000000000001O0000O10000O100O10000O10000O100O10000O100O10000O2O0O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1000O0100O013M3L4M3L4M3M3L4M3M3L4M3L4M3L4G9F:GaQ6"}, "label": "the white cup with the bamboo painted on it"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toasted bread on the bottom\"."}], "task": "refering", "answer_template": "The \"toasted bread on the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260], "bbox": [0.0009062499999999999, 0.16912500000000003, 0.358140625, 0.6853958333333333], "iscrowd": 0, "area": 36500.095449999986, "rle": {"size": [480, 640], "counts": "_b0R5m92M3N3L3N2M4M2M3O2M2O1O2M2O2N1N2O2M101N2N2O1N2O1N2N2O1N101N4L3N0O2O001N10001O0O2O001N101O000O2O001N101O00001O001O001O001O00001O001O001O01O1O100O1O010O1O1O1O1O1OO1O1O1O1O100O1000000000000000000001O001O001O001O001O001O001O001O1O001O001O1O001O1O001O4L7bJnG`3V8nK\\Hn3e7nKbHn3_7PLfHl3[7RLjHj3W7TLnHh3S7VLQIg3P7WLSIg3n6WLVIf3o6SLUIk3P7oKSIo3Q9N2N2N3M2N2N2M2O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O1O1O1O1O001O1O1O1O001O1O001O00001O1O1O1O001O1O1N2N2N2N2O0O2N2N1O2N2O1N1O1O2N100O1O1O1O1O1O001N1000000O101O003M3L4N2M3M4L5K4L2N2N2O2M2O1NT\\P6"}, "label": "toasted bread on the bottom"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bowl next to a plate of bagels\"."}], "task": "refering", "answer_template": "The \"a green bowl next to a plate of bagels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.67596875, 0.6727291666666667, 0.9993437500000001, 1.0], "iscrowd": 0, "area": 23887.62110000001, "rle": {"size": [480, 640], "counts": "Z][6;_><F9H3M3N2M3N2N2M3O1O1N2O1O1O1N2O1O1N2O1O1O1N2N20000000000O1000000000000O10000000000O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O10000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O10000000000000000O10000000000000000000000O1000000O1000000O1000000O10000"}, "label": "a green bowl next to a plate of bagels"}]}
{"id": 446303, "image": "COCO_train2014_000000446303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green cup next to the plate\"."}], "task": "refering", "answer_template": "The \"green cup next to the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.67596875, 0.6727291666666667, 0.9993437500000001, 1.0], "iscrowd": 0, "area": 23887.62110000001, "rle": {"size": [480, 640], "counts": "Z][6;_><F9H3M3N2M3N2N2M3O1O1N2O1O1O1N2O1O1N2O1O1O1N2N20000000000O1000000000000O10000000000O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O10000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O10000000000000000O10000000000000000000000O1000000O1000000O1000000O10000"}, "label": "green cup next to the plate"}]}
{"id": 487264, "image": "COCO_train2014_000000487264.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"larger bird on left\"."}], "task": "refering", "answer_template": "The \"larger bird on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.0, 0.05991666666666667, 0.6448125, 0.8728958333333333], "iscrowd": 0, "area": 103836.9553, "rle": {"size": [480, 640], "counts": "k6h1V=2O2M2O2M2O2M2O2M3N1N3N1N3N1N3N1N3M3N1N3N1N3N1N3N1N3N2M2O2M2O2M2O2MO2O0O2N101N1O2O0O21N2O1N2O0O2O1N2O1N1O2O1N2O1N101N2O1N2O0O2O1N2O1N101N2O1N2O0O2O1N2O0O2O1N2O1N101N2O1O100O101O0O101O0O101O0O2O000O2O000O2O001N10001N10001N101O0O101O0UHRJc6o5\\IRJd6o5[IQJe6o5ZIRJe6P6YIQJg6o5XISJg6m5XITJg6n5XIRJh6n5WISJi6m5VITJi6n5UISJk6m5TIUJk6j5UIWJj6j5VIVJj6j5UIWJk6i5TIXJk6i5TIXJl6h5SIZJl6f5TIZJk6g5TIZJl6f5SI[Jm6e5RI\\Jm6e5RI\\Jn6d5RI]Jm6c5RI^Jm6c5RI^Jn6b5QI_Jn6b5QI_Jo6a5QI_Jo6a5PIaJn6`5QIaJo6_5PIbJP7^5oHcJP7^5PIbJP7^5oHcJQ7\\5oHfJo6[5PIfJP7e600O2O000O2O000O2O000O101O0O101O0O10001N10001N10001N10000O2O000O2O000O101O0O101O0O101O0O10001N10001N1000001N10001N10001N10000O2O000O2O000O101O0O101O0O2O1O2M2O2N1O2M3N1O2M2O2N2M2O2N1N3N2N1N3N1O2M2O2N1N101O0O2O001N101O000O2O001N101O0O2O001N10001O0010O01O010O0010O0001O010O0010O01O0010O01O01O01O010O001O010O0010O0001O01O00001O00010O0000001O00001O01O0001O001O1O001O10O01O1O1O001O1O1O010O1O001O100O00100O001O100O00100O001O100O001O100O00100O001O100O00100O001O100O001O10OjG^Kn5c4nI`KR6c4hI_KZ6d4`I^K`6e4ZI^Kb6i4XIYK`6S5ZIPK]6\\5]IgJZ6e5`I]JY6n5aIUJV6W6eIkIR6a6hIaIP6k6jIXIn5P8H9G9G8I8G:F9D=A?@`0@`0@`0@`0@`0@?A`XZ3"}, "label": "larger bird on left"}]}
{"id": 446307, "image": "COCO_train2014_000000446307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow train with black trim\"."}], "task": "refering", "answer_template": "The \"a yellow train with black trim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 194, 195, 196, 197, 218, 219, 220], "bbox": [0.44768749999999996, 0.3474680306905371, 0.6635624999999999, 0.6844501278772379], "iscrowd": 0, "area": 13367.555999999997, "rle": {"size": [391, 640], "counts": "lh]3m0S;;E8H8H9G8G9J7M2O2M2O1N3N1O1N3N1N2O2M3N2M3N2M3N1O2M3N2M100O10000000O100000000000000000000000000000000000000000000000000000000000000000000O10000O2O001N10001N101O0O101O0O2O000O2N1O2M2O2M2O1N3M2O2M3M2N3M2N3N2M2C>N2N1O2M3N1O2N2N1N3N2N1O2N1N3N2N1O2N2M2O2N2N1O2M3N1O2N2N1O2M2O2N9G^ma2"}, "label": "a yellow train with black trim"}]}
{"id": 446307, "image": "COCO_train2014_000000446307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow train pulling into station\"."}], "task": "refering", "answer_template": "The \"a yellow train pulling into station\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 194, 195, 196, 197, 218, 219, 220], "bbox": [0.44768749999999996, 0.3474680306905371, 0.6635624999999999, 0.6844501278772379], "iscrowd": 0, "area": 13367.555999999997, "rle": {"size": [391, 640], "counts": "lh]3m0S;;E8H8H9G8G9J7M2O2M2O1N3N1O1N3N1N2O2M3N2M3N2M3N1O2M3N2M100O10000000O100000000000000000000000000000000000000000000000000000000000000000000O10000O2O001N10001N101O0O101O0O2O000O2N1O2M2O2M2O1N3M2O2M3M2N3M2N3N2M2C>N2N1O2M3N1O2N2N1N3N2N1O2N1N3N2N1O2N2M2O2N2N1O2M3N1O2N2N1O2M2O2N9G^ma2"}, "label": "a yellow train pulling into station"}]}
{"id": 225124, "image": "COCO_train2014_000000225124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child is skating on road and looking somewhere else\"."}], "task": "refering", "answer_template": "The \"the child is skating on road and looking somewhere else\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 134, 135, 146, 147, 148, 160, 161, 173, 174, 185, 186, 187, 198, 199], "bbox": [0.27136, 0.5447000000000001, 0.42973333333333336, 0.86182], "iscrowd": 0, "area": 4127.688900000001, "rle": {"size": [500, 375], "counts": "hVb12a?2N3M2O1O2N1O1WMFVF<a9N\\F3[99`FGY9e0bF\\Od8@gEZ1d1WOe8AeEX1e1XOe8CdEU15jNh0?_9CcET12QOh09c9BcET1OWOh03f9BcET1NZOf01h9BdES1N[Od01j9AdES1N]Oa0CYOLd:1dES1N^O?C_OG`:5dES1N_O=BED[:9eER1MB;@a:LWER1MC9_Od:LVER1MD7^Og:LUER1MF4]Oj:JVES1LG2\\Ol:KVER1LIOZOQ;JTES1LJMZOT;HSET1LKKYOX;GQEU1LMHXO[;FQEU1L?9VNX:5cEV1L>7[NZ:0cEW1L>4^N^:LbEX1L=2bN`:IbEX1L=0dNc:FaEY1L<NhNf:B`EZ1L;LlNi:^O_E[1K<JnNl:[O_E[1K;HSO\\;b0lDV1T;jNlDV1T;jNlDV1T;jNlDSOIZ1c;CdDnN1X1_;AZDSO644R1_;nN_E3ROh0c;TO\\E6RO`0e;YOZE:RO6g;_OYE<QONj;EWE?oNFm;JWE=POBm;0UE=c;B`D;a;DbD:^;EdD9^=N1N3L^TX3"}, "label": "the child is skating on road and looking somewhere else"}]}
{"id": 225124, "image": "COCO_train2014_000000225124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the blue shirt behind the rail\"."}], "task": "refering", "answer_template": "The \"the girl in the blue shirt behind the rail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 134, 135, 146, 147, 148, 160, 161, 173, 174, 185, 186, 187, 198, 199], "bbox": [0.27136, 0.5447000000000001, 0.42973333333333336, 0.86182], "iscrowd": 0, "area": 4127.688900000001, "rle": {"size": [500, 375], "counts": "hVb12a?2N3M2O1O2N1O1WMFVF<a9N\\F3[99`FGY9e0bF\\Od8@gEZ1d1WOe8AeEX1e1XOe8CdEU15jNh0?_9CcET12QOh09c9BcET1OWOh03f9BcET1NZOf01h9BdES1N[Od01j9AdES1N]Oa0CYOLd:1dES1N^O?C_OG`:5dES1N_O=BED[:9eER1MB;@a:LWER1MC9_Od:LVER1MD7^Og:LUER1MF4]Oj:JVES1LG2\\Ol:KVER1LIOZOQ;JTES1LJMZOT;HSET1LKKYOX;GQEU1LMHXO[;FQEU1L?9VNX:5cEV1L>7[NZ:0cEW1L>4^N^:LbEX1L=2bN`:IbEX1L=0dNc:FaEY1L<NhNf:B`EZ1L;LlNi:^O_E[1K<JnNl:[O_E[1K;HSO\\;b0lDV1T;jNlDV1T;jNlDV1T;jNlDSOIZ1c;CdDnN1X1_;AZDSO644R1_;nN_E3ROh0c;TO\\E6RO`0e;YOZE:RO6g;_OYE<QONj;EWE?oNFm;JWE=POBm;0UE=c;B`D;a;DbD:^;EdD9^=N1N3L^TX3"}, "label": "the girl in the blue shirt behind the rail"}]}
{"id": 225124, "image": "COCO_train2014_000000225124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy has a blue helmet on his head\"."}], "task": "refering", "answer_template": "The \"the boy has a blue helmet on his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 85, 86, 98, 99, 100, 111, 112, 113, 124, 125, 126, 136, 137, 138, 139, 140, 150, 151, 152, 153, 163, 164, 176, 177, 189, 190, 202, 203, 215, 216], "bbox": [0.50328, 0.31164, 0.8058933333333332, 0.92512], "iscrowd": 0, "area": 14960.9736, "rle": {"size": [500, 375], "counts": "Qbl22`?3M2N3O1N2N2OO0100O010O1O010O1VC@`:?`EGY:;fELS:4lE37\\OX7b0`H55ZOY7b0aH74YOY7`0bH94XOY7`0`H<6TO^4LTLe0UO`07QO\\42RL=YOc08oNZ4:nK5]Of09lNY4a0jKNVNH`0R1U1hNZ4c0hKLXNO7m0^1gNZ4b0QK[OSO`0M8Ki0h1cN[4d0kJDQO56X2a1\\M]4e0cJf0^OZ1a1\\M]4f0^Ji0CW1`1\\M^4f0XJo0FQ1b1[M`4f0WJY4X1RK`4h0VJW4X1SKa4g0UJX4X1QKc4j0SJU4X1SKe4i0RJU4V1TKg4j0PJT4V1TKj4j0nIb4f0fJ\\5d8O010001O00001N10001O001O0O103mLhIdL\\6S3jIlLZ6l2lIRMX6f2mIXMW6a2nI^MV6gMQIX4l0oMW6]MVIa4f0QN_6n1dIPN]6m1fIQN\\6m1fIRNZ6l1jIRNW6l1mI^LPO=S7S3\\KfLe4X3bKaL_4]3jK\\LW4b3PLXLQ4e3XLTLh3k3R4N1N3N1N3N2M2O2N102N3YOgCmM\\<P2jCjMX<T2mCgMV<V2f0N3M3M2N3M2N4P1POM3L4L4M3L4M3L4L4L4L4Ki0WO3N3L4M3L3M4M3L^ZS1"}, "label": "the boy has a blue helmet on his head"}]}
{"id": 225124, "image": "COCO_train2014_000000225124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skater wearing a blue helmet\"."}], "task": "refering", "answer_template": "The \"skater wearing a blue helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 85, 86, 98, 99, 100, 111, 112, 113, 124, 125, 126, 136, 137, 138, 139, 140, 150, 151, 152, 153, 163, 164, 176, 177, 189, 190, 202, 203, 215, 216], "bbox": [0.50328, 0.31164, 0.8058933333333332, 0.92512], "iscrowd": 0, "area": 14960.9736, "rle": {"size": [500, 375], "counts": "Qbl22`?3M2N3O1N2N2OO0100O010O1O010O1VC@`:?`EGY:;fELS:4lE37\\OX7b0`H55ZOY7b0aH74YOY7`0bH94XOY7`0`H<6TO^4LTLe0UO`07QO\\42RL=YOc08oNZ4:nK5]Of09lNY4a0jKNVNH`0R1U1hNZ4c0hKLXNO7m0^1gNZ4b0QK[OSO`0M8Ki0h1cN[4d0kJDQO56X2a1\\M]4e0cJf0^OZ1a1\\M]4f0^Ji0CW1`1\\M^4f0XJo0FQ1b1[M`4f0WJY4X1RK`4h0VJW4X1SKa4g0UJX4X1QKc4j0SJU4X1SKe4i0RJU4V1TKg4j0PJT4V1TKj4j0nIb4f0fJ\\5d8O010001O00001N10001O001O0O103mLhIdL\\6S3jIlLZ6l2lIRMX6f2mIXMW6a2nI^MV6gMQIX4l0oMW6]MVIa4f0QN_6n1dIPN]6m1fIQN\\6m1fIRNZ6l1jIRNW6l1mI^LPO=S7S3\\KfLe4X3bKaL_4]3jK\\LW4b3PLXLQ4e3XLTLh3k3R4N1N3N1N3N2M2O2N102N3YOgCmM\\<P2jCjMX<T2mCgMV<V2f0N3M3M2N3M2N4P1POM3L4L4M3L4M3L4L4L4L4Ki0WO3N3L4M3L3M4M3L^ZS1"}, "label": "skater wearing a blue helmet"}]}
{"id": 495460, "image": "COCO_train2014_000000495460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwich with a large bubble hole in the bread\"."}], "task": "refering", "answer_template": "The \"the half of the sandwich with a large bubble hole in the bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190], "bbox": [0.001265625, 0.0024444444444444444, 0.6103125, 0.6713611111111111], "iscrowd": 0, "area": 68170.8156, "rle": {"size": [360, 640], "counts": "\\<X1n9j1WNj1VNl0TO1O001O001O001N101O001O0010O010O01O010O3N1N3M2O2M2O2M3N1N3M2O2M2O2M2N3N0O100O100O1O100000000000000000O10000000000000000000000001O0000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O000000O10000O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O10000O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O101N2N2O1N2N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2N2O1N3M2O1N2N2O1M4I6J6J6J6L4N3M2O1N2O1N2O2M2N_Ob00O100O2N100O100O100O2N100O100O101N1O100O10mNhMVIW2k6XNgHg1Y7gNYHY1g7T1000000000000000000000000O100000000000DVHZMj7b2\\H[Me7d2^HZMb7f2`000O100010001O0O2cNaGN_8OhGMX80nGMS8NUHNk7OZHOd70`HO_7NgH0W7NoH0P7F]I8a6FeI7[6FkI8Y6@mI>`8O2N2N2N2N001O0O101O001O001O00chg2"}, "label": "the half of the sandwich with a large bubble hole in the bread"}]}
{"id": 495460, "image": "COCO_train2014_000000495460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left half of the sandwhich\"."}], "task": "refering", "answer_template": "The \"the left half of the sandwhich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190], "bbox": [0.001265625, 0.0024444444444444444, 0.6103125, 0.6713611111111111], "iscrowd": 0, "area": 68170.8156, "rle": {"size": [360, 640], "counts": "\\<X1n9j1WNj1VNl0TO1O001O001O001N101O001O0010O010O01O010O3N1N3M2O2M2O2M3N1N3M2O2M2O2M2N3N0O100O100O1O100000000000000000O10000000000000000000000001O0000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O000000O10000O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O10000O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O101N2N2O1N2N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2N2O1N2N3N1N2N2O1N3M2O1N2N2O1M4I6J6J6J6L4N3M2O1N2O1N2O2M2N_Ob00O100O2N100O100O100O2N100O100O101N1O100O10mNhMVIW2k6XNgHg1Y7gNYHY1g7T1000000000000000000000000O100000000000DVHZMj7b2\\H[Me7d2^HZMb7f2`000O100010001O0O2cNaGN_8OhGMX80nGMS8NUHNk7OZHOd70`HO_7NgH0W7NoH0P7F]I8a6FeI7[6FkI8Y6@mI>`8O2N2N2N2N001O0O101O001O001O00chg2"}, "label": "the left half of the sandwhich"}]}
{"id": 495460, "image": "COCO_train2014_000000495460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the portion of the sandwich on the right\"."}], "task": "refering", "answer_template": "The \"the portion of the sandwich on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.5876718750000001, 0.006916666666666667, 1.0, 0.9823055555555555], "iscrowd": 0, "area": 67104.78225, "rle": {"size": [360, 640], "counts": "X\\T41W;1SO2]F0\\97bFKX9;fFFU9`0jFAS9b0kF@S9b0kF@T9a0jFAU9`0jFA\\8EmGl0EA]8HiGh0HB^8JeGf0KB^8KeGd0LB^8LdGc0LC_8LbG]1]8eNaG[1^8i0O001O0O2O00001N10001O0O101O000O2N1O2N1N2O2N1O1O2N1O1N3N1O1O2N1O3M2N2M4M2N2N3M2N2N3L3N2N1O2N1O2N1N3N1O2N1O2N1O2M2N3K4K6K4L5K4L5K4L5L301N3N2M3M4M2M3M3N3L3L4L5L3L4L5K4L2N3M3N101N2O001O1N101O1O001N2O001O1O0O2O1O1O0O2O1O1O001O1N2O1O001O1N2O001O1O1O1N101O1O1O0O2O1O1O1O001O1N2O001O1O1O1O001O1N2O001O1O1O1O001N2O100O001O1O1O001O1O1O10O01O1O1O0000000001O0000000000000000001O00000O10000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000001O00000001O00000000000000001O000001O00cL]3UK_4"}, "label": "the portion of the sandwich on the right"}]}
{"id": 495460, "image": "COCO_train2014_000000495460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich portion to the right of another sandwich portion\"."}], "task": "refering", "answer_template": "The \"a sandwich portion to the right of another sandwich portion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.5876718750000001, 0.006916666666666667, 1.0, 0.9823055555555555], "iscrowd": 0, "area": 67104.78225, "rle": {"size": [360, 640], "counts": "X\\T41W;1SO2]F0\\97bFKX9;fFFU9`0jFAS9b0kF@S9b0kF@T9a0jFAU9`0jFA\\8EmGl0EA]8HiGh0HB^8JeGf0KB^8KeGd0LB^8LdGc0LC_8LbG]1]8eNaG[1^8i0O001O0O2O00001N10001O0O101O000O2N1O2N1N2O2N1O1O2N1O1N3N1O1O2N1O3M2N2M4M2N2N3M2N2N3L3N2N1O2N1O2N1N3N1O2N1O2N1O2M2N3K4K6K4L5K4L5K4L5L301N3N2M3M4M2M3M3N3L3L4L5L3L4L5K4L2N3M3N101N2O001O1N101O1O001N2O001O1O0O2O1O1O0O2O1O1O001O1N2O1O001O1N2O001O1O1O1N101O1O1O0O2O1O1O1O001O1N2O001O1O1O1O001O1N2O001O1O1O1O001N2O100O001O1O1O001O1O1O10O01O1O1O0000000001O0000000000000000001O00000O10000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000001O00000001O00000000000000001O000001O00cL]3UK_4"}, "label": "a sandwich portion to the right of another sandwich portion"}]}
{"id": 339816, "image": "COCO_train2014_000000339816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the painted white surface of the table without bananas\"."}], "task": "refering", "answer_template": "The \"the painted white surface of the table without bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 3, 4, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 71, 89, 90, 93, 94], "bbox": [0.003359375, 0.015720250521920668, 0.987265625, 0.27189979123173275], "iscrowd": 0, "area": 25972.313050000008, "rle": {"size": [479, 640], "counts": "co03f>8M3M3M3M3M3M3M2M4M3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3L4M3M3M3M2N3M3M3M3N2M3N1N3N21N5I8I4K101N100O100O100O2O0O1O1ZNQD:Q<FWD1j;OXDMj;2YDKi;4YDHi;8YDEh;;ZDAh;?[D]Of;c0\\DYOf;g0\\DVOe;j0]DROe;n0]DoNd;Q1T11000O1000O1000000O1000000O10O1000O1000000O1000000O10O100000O1000000O100000O0100000000O10000000hN[BV1Q>F8G10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O10000000000000001O00000000000000000000000O100000000000000000000000000000000000000O1000000000000000_OIZB7e=OVB1h=6SBJl=e0N2O1N2O1N2O1N2O0O200000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O0000000000000000000000000000000000O10000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000OYCjNg;V1XDkNh;W1TDlNk;[1lCgNT<`1dCbN[<e1\\C]Nd<P2O001O001O001O001O001eCaMT<_2kCbMU<_2iCbMW<c20001O0000000000001O00000000000000000000000001N1000000O1000000O1000000O1000000O2O001O0O1M4K4L5L3L5K4L5K4M4K4L4L5K4L5K4L5K4LXg3"}, "label": "the painted white surface of the table without bananas"}]}
{"id": 339816, "image": "COCO_train2014_000000339816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table with white top being used to hold fruits\"."}], "task": "refering", "answer_template": "The \"table with white top being used to hold fruits\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 3, 4, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 71, 89, 90, 93, 94], "bbox": [0.003359375, 0.015720250521920668, 0.987265625, 0.27189979123173275], "iscrowd": 0, "area": 25972.313050000008, "rle": {"size": [479, 640], "counts": "co03f>8M3M3M3M3M3M3M2M4M3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3L4M3M3M3M2N3M3M3M3N2M3N1N3N21N5I8I4K101N100O100O100O2O0O1O1ZNQD:Q<FWD1j;OXDMj;2YDKi;4YDHi;8YDEh;;ZDAh;?[D]Of;c0\\DYOf;g0\\DVOe;j0]DROe;n0]DoNd;Q1T11000O1000O1000000O1000000O10O1000O1000000O1000000O10O100000O1000000O100000O0100000000O10000000hN[BV1Q>F8G10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O10000000000000001O00000000000000000000000O100000000000000000000000000000000000000O1000000000000000_OIZB7e=OVB1h=6SBJl=e0N2O1N2O1N2O1N2O0O200000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O0000000000000000000000000000000000O10000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000OYCjNg;V1XDkNh;W1TDlNk;[1lCgNT<`1dCbN[<e1\\C]Nd<P2O001O001O001O001O001eCaMT<_2kCbMU<_2iCbMW<c20001O0000000000001O00000000000000000000000001N1000000O1000000O1000000O1000000O2O001O0O1M4K4L5L3L5K4L5K4M4K4L4L5K4L5K4L5K4LXg3"}, "label": "table with white top being used to hold fruits"}]}
{"id": 339816, "image": "COCO_train2014_000000339816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bunches of bananas containing a lot of bananas per bundle\"."}], "task": "refering", "answer_template": "The \"bunches of bananas containing a lot of bananas per bundle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.14260960334029227, 1.0, 1.0], "iscrowd": 0, "area": 252696.41685, "rle": {"size": [479, 640], "counts": "g4S:l40O10001O0O1000000O100000000O10001O0O1000000O1000000O10001O000O1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1O1N2O1O1N2O1O2M2O1O1N2ZO]DoMd;Q2]DmMd;R2^DkMd;T2^DjMc;V2^DhMc;W2_DfMc;Y2_DeMb;[2`DbMa;]2a0O100O1O100O100000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O1002N3M2N2N3M2N1O00O10000O100O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000001O001O001O00001O001O003M4iC[Ml;l20001O0000000000000000001O000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000000000001O0000000000000000001O"}, "label": "bunches of bananas containing a lot of bananas per bundle"}]}
{"id": 339816, "image": "COCO_train2014_000000339816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table covered with newspaper and bananas\"."}], "task": "refering", "answer_template": "The \"a table covered with newspaper and bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.14260960334029227, 1.0, 1.0], "iscrowd": 0, "area": 252696.41685, "rle": {"size": [479, 640], "counts": "g4S:l40O10001O0O1000000O100000000O10001O0O1000000O1000000O10001O000O1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1O1N2O1O1N2O1O2M2O1O1N2ZO]DoMd;Q2]DmMd;R2^DkMd;T2^DjMc;V2^DhMc;W2_DfMc;Y2_DeMb;[2`DbMa;]2a0O100O1O100O100000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O1002N3M2N2N3M2N1O00O10000O100O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000001O001O001O00001O001O003M4iC[Ml;l20001O0000000000000000001O000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000000000001O0000000000000000001O"}, "label": "a table covered with newspaper and bananas"}]}
{"id": 69480, "image": "COCO_train2014_000000069480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown goat with its ears up\"."}], "task": "refering", "answer_template": "The \"a brown goat with its ears up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 174, 175, 197, 198, 221, 222, 244, 245, 267, 268], "bbox": [0.569140625, 0.20931640625, 0.753328125, 0.634765625], "iscrowd": 0, "area": 9109.042599999999, "rle": {"size": [512, 640], "counts": "XUf52n?6I7J3M1O1001O0010WA^OQ>b0lAAT>?`AMa>2]A0d>O[A2f>MXA5i>JVA7k>`01O00hNX1K5K^OgBeNe0]Og;m1gChN>]Oh;k1mCjNM]ON1U<k1PD\\Om;f0TDZOj;g0WDZOf;h0ZDXOe;h0\\DXO]1hNW8P2]FZOX1iNY8n1`F^Oo0gN_8k1dFCc0fNh8g1gFH5hNR9P1`FaN:l1HiN]9h0`Gf1_8[NbGd1]8\\NfGb1X8`NiG_1S8eNgHa0Y7_OgH`0Y7@hH?Y7AhH=Y7CgH]NiMd1a9NhH]NjM`1_94gH\\NnM6O>^9P1eH\\NUNH2f0W9V1cH[NU9d1lF\\NT9d1lF\\NS9e1nFZNR9f1nFYNBM_7i1QIYN\\O2c7e1QIYNXO7e7a1SIXNUO:h7_1SIVNRO>k7\\1SIUNPOb0m7Z1SISNmNf0P8W1SISNiNj0S8U1SIQNfNn0W8Q1TI4l6LTI4l6MSI3l6NUI1k60TI0l60UIOk61UIOk60VI0i61XINh61YIOg61ZINf62ZINf61[IJj66WIDm6=SI@P7?QI_OQ7a0PI\\OR7d0nHYOU7f0mH]N`MKm9h1eHXNo7h1U300O2O0O2O0O101O0O101N101N100O2O0O101N100O2M2O2M2O1N3N3N2M3N2M3N2M3N2M3N2M5Lck^2"}, "label": "a brown goat with its ears up"}]}
{"id": 69480, "image": "COCO_train2014_000000069480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow whose head is hidden behind the white cow\"."}], "task": "refering", "answer_template": "The \"a brown cow whose head is hidden behind the white cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 174, 175, 197, 198, 221, 222, 244, 245, 267, 268], "bbox": [0.569140625, 0.20931640625, 0.753328125, 0.634765625], "iscrowd": 0, "area": 9109.042599999999, "rle": {"size": [512, 640], "counts": "XUf52n?6I7J3M1O1001O0010WA^OQ>b0lAAT>?`AMa>2]A0d>O[A2f>MXA5i>JVA7k>`01O00hNX1K5K^OgBeNe0]Og;m1gChN>]Oh;k1mCjNM]ON1U<k1PD\\Om;f0TDZOj;g0WDZOf;h0ZDXOe;h0\\DXO]1hNW8P2]FZOX1iNY8n1`F^Oo0gN_8k1dFCc0fNh8g1gFH5hNR9P1`FaN:l1HiN]9h0`Gf1_8[NbGd1]8\\NfGb1X8`NiG_1S8eNgHa0Y7_OgH`0Y7@hH?Y7AhH=Y7CgH]NiMd1a9NhH]NjM`1_94gH\\NnM6O>^9P1eH\\NUNH2f0W9V1cH[NU9d1lF\\NT9d1lF\\NS9e1nFZNR9f1nFYNBM_7i1QIYN\\O2c7e1QIYNXO7e7a1SIXNUO:h7_1SIVNRO>k7\\1SIUNPOb0m7Z1SISNmNf0P8W1SISNiNj0S8U1SIQNfNn0W8Q1TI4l6LTI4l6MSI3l6NUI1k60TI0l60UIOk61UIOk60VI0i61XINh61YIOg61ZINf62ZINf61[IJj66WIDm6=SI@P7?QI_OQ7a0PI\\OR7d0nHYOU7f0mH]N`MKm9h1eHXNo7h1U300O2O0O2O0O101O0O101N101N100O2O0O101N100O2M2O2M2O1N3N3N2M3N2M3N2M3N2M3N2M5Lck^2"}, "label": "a brown cow whose head is hidden behind the white cow"}]}
{"id": 372588, "image": "COCO_train2014_000000372588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the colorful bedspread on a bed farthest from the window\"."}], "task": "refering", "answer_template": "The \"the colorful bedspread on a bed farthest from the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.0016875000000000002, 0.6422175732217573, 0.64740625, 1.0], "iscrowd": 0, "area": 53230.5112, "rle": {"size": [478, 640], "counts": "gh0S5k90O1O100LlJ]FU5c9kJ]FU5d920000001O000000001O000000001O01O0001O000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O000001O000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O001O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N3M4I7F:F:GPRY3"}, "label": "the colorful bedspread on a bed farthest from the window"}]}
{"id": 372588, "image": "COCO_train2014_000000372588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hotel room bed , closest to the room door\"."}], "task": "refering", "answer_template": "The \"the hotel room bed , closest to the room door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.0016875000000000002, 0.6422175732217573, 0.64740625, 1.0], "iscrowd": 0, "area": 53230.5112, "rle": {"size": [478, 640], "counts": "gh0S5k90O1O100LlJ]FU5c9kJ]FU5d920000001O000000001O000000001O01O0001O000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O000001O000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O001O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N3M4I7F:F:GPRY3"}, "label": "the hotel room bed , closest to the room door"}]}
{"id": 372588, "image": "COCO_train2014_000000372588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a bed exactly in front of the tv\"."}], "task": "refering", "answer_template": "The \"there is a bed exactly in front of the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 336, 337, 338, 359, 360], "bbox": [0.146015625, 0.46531380753138074, 0.750234375, 0.9057740585774058], "iscrowd": 0, "area": 41241.88665, "rle": {"size": [478, 640], "counts": "^T\\11l>101N1O2O0O2O0O100O1O101N100O100O1O2O0O100O100O2N100O100O101N1O100O100O2O000O100O100O2O000O101N2O2N1N3N1N2O2M3N<D;D<E<C2O001N100O100O10000O2O0O10000O100O2O0O10000O100O101O0O1000000000001O000000000O101O000000000000001O000O1001O00010O0000001O00010O00000010O0001O0000010O00001O00010O0000001O01O01O00000010O000000010O00000001O01O0001O000001O01O000000010O000000010O0000001O01O000001O0001O0001O000001O000000010O00000000001O0001O000001O000000000010O00000001O00000001O01O00000000001O0001O000001O0000000010O0000000001O00000001O01O00000000001O01O000001O000000000010O01O2N1O2N2N2N2N101N2N2N1O2N2N2N1O2N2O1N2N1O2N2N2N1O2N2O1N1O2]FhJ\\9Z5aFhJ^9\\501O1O001O01O00000000000O10001O0000000000000O1L5N1O100O1O100O100O2N100O1O1O1O1O1O1O2N1O1O1O1L4J6TOl0M4K4M3M3M3M6L4L5K4L5L3L5K4L4M4K4L5K4M4K4L5K4M3L5K4L5Lc[Z2"}, "label": "there is a bed exactly in front of the tv"}]}
{"id": 372588, "image": "COCO_train2014_000000372588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"double - sized bed with multi - colored blanket\"."}], "task": "refering", "answer_template": "The \"double - sized bed with multi - colored blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 336, 337, 338, 359, 360], "bbox": [0.146015625, 0.46531380753138074, 0.750234375, 0.9057740585774058], "iscrowd": 0, "area": 41241.88665, "rle": {"size": [478, 640], "counts": "^T\\11l>101N1O2O0O2O0O100O1O101N100O100O1O2O0O100O100O2N100O100O101N1O100O100O2O000O100O100O2O000O101N2O2N1N3N1N2O2M3N<D;D<E<C2O001N100O100O10000O2O0O10000O100O2O0O10000O100O101O0O1000000000001O000000000O101O000000000000001O000O1001O00010O0000001O00010O00000010O0001O0000010O00001O00010O0000001O01O01O00000010O000000010O00000001O01O0001O000001O01O000000010O000000010O0000001O01O000001O0001O0001O000001O000000010O00000000001O0001O000001O000000000010O00000001O00000001O01O00000000001O0001O000001O0000000010O0000000001O00000001O01O00000000001O01O000001O000000000010O01O2N1O2N2N2N2N101N2N2N1O2N2N2N1O2N2O1N2N1O2N2N2N1O2N2O1N1O2]FhJ\\9Z5aFhJ^9\\501O1O001O01O00000000000O10001O0000000000000O1L5N1O100O1O100O100O2N100O1O1O1O1O1O1O2N1O1O1O1L4J6TOl0M4K4M3M3M3M6L4L5K4L5L3L5K4L4M4K4L5K4M4K4L5K4M3L5K4L5Lc[Z2"}, "label": "double - sized bed with multi - colored blanket"}]}
{"id": 110447, "image": "COCO_train2014_000000110447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red teddy bear sitting on the grass near a brown teddy bear\"."}], "task": "refering", "answer_template": "The \"a red teddy bear sitting on the grass near a brown teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 209, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268], "bbox": [0.23070833333333332, 0.310109375, 0.797, 0.701125], "iscrowd": 0, "area": 30908.25179999999, "rle": {"size": [640, 480], "counts": "VdU22mc03M2N3M3M2N3M2O2N2N1O2N101N1O001O01O0001O0001O0001O0X_O@P>`0iAIV>6jALT>5kAMS>3nANP>2PB0n=0RB2m=MSB5k=LTB6j=JVB8h=HXB:f=FZB<e=DZB>d=B\\B`0b=@_Ba0_=_OaBc0]=]OcBe0[=\\OeBe0Z=ZOfBh0X=XOiBf0X=ZOiBe0W=\\OhBd0X=\\OiBc0W=]OjBa0X=^OhBb0X=^OiBa0W=@hB?Y=AhB>X=BhB>X=CgB<[=CdB>\\=BdB>\\=BdB>\\=CcB<^=DbB<^=DbB<_=D`B;a=E_B;a=E_B;a=F^B9c=G]B9c=G\\B:e=FZB:f=FZB9g=GYB9g=HXB8h=HXB7i=IWB7j=HVB8j=IUB7k=IUB6l=JTB6l=JSB7m=ISB6o=IQB7o=IQB7o=IQB6P>KoA5Q>KoA5Q>KoA5R>JnA5S>KmA5S>KmA5S>KlA5U>KkA5U>KkA5V>JjA5W>KiA5W>KiA5W>KiA5W>LhA3Y>MgA3V>0jA0R>4mALQ>7oAIm=;RBFk==UBBi=a0VB@i=a0WB_Oh=b0WB_Oh=b0XB]Oh=d0WB]Oh=d0XB\\Og=e0XB\\Of=g0YBYOf=h0YBYOf=h0ZBXOe=i0ZBXOe=i0XBYOh=h0UB[Oj=f0SB]Ol=f0oA]OP>f0kA]OT>f0fA^OY>e0^ABa>a0WAEh>j2O1O1O1O1O1O1O1O2M2O101N100O101N100O101N100O101N100O2O000O101N100O101N100O100O100O101O2M3N2M2O2M3N2M2O2M3N2M22O1O1O1N2O1O1O3L4M3K6K4K6J5L2M100O2O0O100O1O100O100O2O0O010O1O100N2O0O2O1N2O1N2O0O2O1N2O1N4M2L4M4K4M3L5L3L4M4L3L5L3L4M3L5L3L4L5J5K5L4O20O1O10O0N2N3N1N2N2N3M2O1N3M2N2N3N1N2N2N3M2O1N3M2N2H9F9HbWl1"}, "label": "a red teddy bear sitting on the grass near a brown teddy bear"}]}
{"id": 110447, "image": "COCO_train2014_000000110447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red teddy bear sitting on the ground next to another teddy bear\"."}], "task": "refering", "answer_template": "The \"a red teddy bear sitting on the ground next to another teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 209, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268], "bbox": [0.23070833333333332, 0.310109375, 0.797, 0.701125], "iscrowd": 0, "area": 30908.25179999999, "rle": {"size": [640, 480], "counts": "VdU22mc03M2N3M3M2N3M2O2N2N1O2N101N1O001O01O0001O0001O0001O0X_O@P>`0iAIV>6jALT>5kAMS>3nANP>2PB0n=0RB2m=MSB5k=LTB6j=JVB8h=HXB:f=FZB<e=DZB>d=B\\B`0b=@_Ba0_=_OaBc0]=]OcBe0[=\\OeBe0Z=ZOfBh0X=XOiBf0X=ZOiBe0W=\\OhBd0X=\\OiBc0W=]OjBa0X=^OhBb0X=^OiBa0W=@hB?Y=AhB>X=BhB>X=CgB<[=CdB>\\=BdB>\\=BdB>\\=CcB<^=DbB<^=DbB<_=D`B;a=E_B;a=E_B;a=F^B9c=G]B9c=G\\B:e=FZB:f=FZB9g=GYB9g=HXB8h=HXB7i=IWB7j=HVB8j=IUB7k=IUB6l=JTB6l=JSB7m=ISB6o=IQB7o=IQB7o=IQB6P>KoA5Q>KoA5Q>KoA5R>JnA5S>KmA5S>KmA5S>KlA5U>KkA5U>KkA5V>JjA5W>KiA5W>KiA5W>KiA5W>LhA3Y>MgA3V>0jA0R>4mALQ>7oAIm=;RBFk==UBBi=a0VB@i=a0WB_Oh=b0WB_Oh=b0XB]Oh=d0WB]Oh=d0XB\\Og=e0XB\\Of=g0YBYOf=h0YBYOf=h0ZBXOe=i0ZBXOe=i0XBYOh=h0UB[Oj=f0SB]Ol=f0oA]OP>f0kA]OT>f0fA^OY>e0^ABa>a0WAEh>j2O1O1O1O1O1O1O1O2M2O101N100O101N100O101N100O101N100O2O000O101N100O101N100O100O100O101O2M3N2M2O2M3N2M2O2M3N2M22O1O1O1N2O1O1O3L4M3K6K4K6J5L2M100O2O0O100O1O100O100O2O0O010O1O100N2O0O2O1N2O1N2O0O2O1N2O1N4M2L4M4K4M3L5L3L4M4L3L5L3L4M3L5L3L4L5J5K5L4O20O1O10O0N2N3N1N2N2N3M2O1N3M2N2N3N1N2N2N3M2O1N3M2N2H9F9HbWl1"}, "label": "a red teddy bear sitting on the ground next to another teddy bear"}]}
{"id": 110447, "image": "COCO_train2014_000000110447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown stuffed bear\"."}], "task": "refering", "answer_template": "The \"brown stuffed bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 170, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 222, 223], "bbox": [0.02675, 0.19140625, 0.4352916666666667, 0.595875], "iscrowd": 0, "area": 27623.29595000001, "rle": {"size": [640, 480], "counts": "h]8g0Wc08I7I5J3N2M3N2O1O1O1]MaNdB_1R=QOgBP1P=YOoBh0k<_OSCb0k<@UCa0i<AUC`0j<BTC?k<BUC>i<EUC=i<DVC=i<DWC<h<EWC<g<FXC;g<EZC<d<E[C<d<E[C<c<E^C;a<F^C;a<E`C<^<EaC;_<F`C9b<F_C9a<H^C7c<I]C7c<H^C7c<I]C7T;oMbDj1:6T;SNdCN:h1o06S;WN\\C2a0a1P16Q;jNoCP1o06R;lNnCm0Q17o:oNoCj0R16o:QOPDh0P18o:ROQDd0Q19m:VORD`0Q1:m:VORD`0P1:o:WOQD=Q1<o:WOPD<P1>P;VOQD;o0>Q;XOPD8P1`0P;YOoC7Q1?Q;ZOoC6P1`0R;ZOnC5P1`0S;\\OnC1P1c0R;\\OoC0o0d0R;]OPDMn0e0S;_OoCJo0g0R;AnCGg0P1[;\\OmCCf0S1];]OkC^Oh0U1^;^OjC\\Of0X1_;^OkC^O?U1g;_OiC_O;S1l;_OiCB4Q1R<_OjCCOn0X<@iCc2V<`MiC^2X<cMgC]2X<eMhCY2Y<hMfCX2Y<kMeCT2\\<mMdCR2[<QNcCn1^<SNaCm1^<UNbCj1^<XN`Cg1`<[N_Ce1a<l21O1L4L4L4L4N2N2N2N2O010O100fMbBQN_=o1eBmM[=R2fBnMZ=R2gBlMZ=T2fBlMZ=T2gBkMY=T2iBjMX=V2hBjMX=V2iBiMV=W2kBhMV=X2kBgMU=Y2lBfMT=Z2lBfMT=Y2nBgMR=X2nBiMQ=W2PCiMo<W2QCjMn<U2TCkMl<T2UCnMh<R2ZCoMc<Q2^CRN^<n1cCTN[<k1gCWNU<i1lCXNR<h1PDWNo;i1RDWNn;h1SDWNm;i1TDWNk;j1UDVNk;i1UDWNk;j1UDSNn;l1RDPNQ<R2oCjMS<W2mCiMR<Y2nCfMQ<[2oCeMo;^2PDbMo;_2QDaMn;a2QD_Mn;b2RD_Mk;d2TD\\Ml;d2TD\\Ml;e2SD[Mm;j0TCZOn0Ln;i0VCZOl0Mo;f0YC\\Og0NP<d0\\C]Od0OP<c0^C^Oa00P<`0aC@?0Q<=cCC;0R<;eCE91Q<9gCG72Q<4jCJ53P<0oCM04P<MRDON5o;IVD3I6Q<CYD7F7P<@]D8C:n;[ObD;@;n;nNmDg0UO<n;kNnDi0TO=n;iNnDj0TO>n;fNoDl0SO?m;dNREl0QO`0n;cNQEm0QO`0^<RNbD^1PO`0`>@`A?a>A_A?a>A_A?a>A_A?a>A_A?b>@^A?d>@]A?d>@]A>d>B^A:f>D[A:g>EZA9i>EXA8k>GVA7m>GUA[ORONl?f0SAUOo?i0R@POU`0o0l_OkN[`0S1Z1O2N0000O10000O100O100O10000O100O10L6J8HjTY5"}, "label": "brown stuffed bear"}]}
{"id": 110447, "image": "COCO_train2014_000000110447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear with a brown color sitting on a grass\"."}], "task": "refering", "answer_template": "The \"a teddy bear with a brown color sitting on a grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 170, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 222, 223], "bbox": [0.02675, 0.19140625, 0.4352916666666667, 0.595875], "iscrowd": 0, "area": 27623.29595000001, "rle": {"size": [640, 480], "counts": "h]8g0Wc08I7I5J3N2M3N2O1O1O1]MaNdB_1R=QOgBP1P=YOoBh0k<_OSCb0k<@UCa0i<AUC`0j<BTC?k<BUC>i<EUC=i<DVC=i<DWC<h<EWC<g<FXC;g<EZC<d<E[C<d<E[C<c<E^C;a<F^C;a<E`C<^<EaC;_<F`C9b<F_C9a<H^C7c<I]C7c<H^C7c<I]C7T;oMbDj1:6T;SNdCN:h1o06S;WN\\C2a0a1P16Q;jNoCP1o06R;lNnCm0Q17o:oNoCj0R16o:QOPDh0P18o:ROQDd0Q19m:VORD`0Q1:m:VORD`0P1:o:WOQD=Q1<o:WOPD<P1>P;VOQD;o0>Q;XOPD8P1`0P;YOoC7Q1?Q;ZOoC6P1`0R;ZOnC5P1`0S;\\OnC1P1c0R;\\OoC0o0d0R;]OPDMn0e0S;_OoCJo0g0R;AnCGg0P1[;\\OmCCf0S1];]OkC^Oh0U1^;^OjC\\Of0X1_;^OkC^O?U1g;_OiC_O;S1l;_OiCB4Q1R<_OjCCOn0X<@iCc2V<`MiC^2X<cMgC]2X<eMhCY2Y<hMfCX2Y<kMeCT2\\<mMdCR2[<QNcCn1^<SNaCm1^<UNbCj1^<XN`Cg1`<[N_Ce1a<l21O1L4L4L4L4N2N2N2N2O010O100fMbBQN_=o1eBmM[=R2fBnMZ=R2gBlMZ=T2fBlMZ=T2gBkMY=T2iBjMX=V2hBjMX=V2iBiMV=W2kBhMV=X2kBgMU=Y2lBfMT=Z2lBfMT=Y2nBgMR=X2nBiMQ=W2PCiMo<W2QCjMn<U2TCkMl<T2UCnMh<R2ZCoMc<Q2^CRN^<n1cCTN[<k1gCWNU<i1lCXNR<h1PDWNo;i1RDWNn;h1SDWNm;i1TDWNk;j1UDVNk;i1UDWNk;j1UDSNn;l1RDPNQ<R2oCjMS<W2mCiMR<Y2nCfMQ<[2oCeMo;^2PDbMo;_2QDaMn;a2QD_Mn;b2RD_Mk;d2TD\\Ml;d2TD\\Ml;e2SD[Mm;j0TCZOn0Ln;i0VCZOl0Mo;f0YC\\Og0NP<d0\\C]Od0OP<c0^C^Oa00P<`0aC@?0Q<=cCC;0R<;eCE91Q<9gCG72Q<4jCJ53P<0oCM04P<MRDON5o;IVD3I6Q<CYD7F7P<@]D8C:n;[ObD;@;n;nNmDg0UO<n;kNnDi0TO=n;iNnDj0TO>n;fNoDl0SO?m;dNREl0QO`0n;cNQEm0QO`0^<RNbD^1PO`0`>@`A?a>A_A?a>A_A?a>A_A?a>A_A?b>@^A?d>@]A?d>@]A>d>B^A:f>D[A:g>EZA9i>EXA8k>GVA7m>GUA[ORONl?f0SAUOo?i0R@POU`0o0l_OkN[`0S1Z1O2N0000O10000O100O100O10000O100O10L6J8HjTY5"}, "label": "a teddy bear with a brown color sitting on a grass"}]}
{"id": 12145, "image": "COCO_train2014_000000012145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front edge of a tan scooter with a carrying container on it\"."}], "task": "refering", "answer_template": "The \"the front edge of a tan scooter with a carrying container on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 253, 254, 276, 277, 299, 300, 322], "bbox": [0.0016875000000000002, 0.42472916666666666, 0.12978125000000001, 0.8763958333333334], "iscrowd": 0, "area": 8818.192799999999, "rle": {"size": [480, 640], "counts": "cj0a1]:Q3lMT21O1O1O2N1O1lKaGf1`8ZNaGd1a8[N_Gd1b8\\N_Ga1c8^N^G`1d8`N]G]1e8cN[G[1h8dNYGY1i8fNXGX1j8hNWGU1k8kNUGS1m8mNTGP1o8nNRGP1P9POQGn0P9ROPGm0Q9SOPGk0Q9TOPGk0R9TOoFj0R9VOnFi0S9WOnFg0S9XOnFg0T9XOmFf0T9ZOlFe0U9[OlFc0V9[OlFc0U9]OkFb0V9^OkF`0W9_OiFa0W9^OkF`0V9@kF>W9AiF>X9BiF<X9CiFeN[OQ1n99hF`NDQ1e9?hFYNMS1\\9d0hGWOZ8g0kGTOV8l0oGnNS8Q1QHjNP8V1THeNm7Z1YH`Ni7_1[H\\Nf7d1k201N10001O0O101O001N2O2N1O1O1N2O2N1O1N2O2N2N3L4M3M2N3L4M2N3M3LkjT8"}, "label": "the front edge of a tan scooter with a carrying container on it"}]}
{"id": 12145, "image": "COCO_train2014_000000012145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back end of a brown motorcycle\"."}], "task": "refering", "answer_template": "The \"the back end of a brown motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 253, 254, 276, 277, 299, 300, 322], "bbox": [0.0016875000000000002, 0.42472916666666666, 0.12978125000000001, 0.8763958333333334], "iscrowd": 0, "area": 8818.192799999999, "rle": {"size": [480, 640], "counts": "cj0a1]:Q3lMT21O1O1O2N1O1lKaGf1`8ZNaGd1a8[N_Gd1b8\\N_Ga1c8^N^G`1d8`N]G]1e8cN[G[1h8dNYGY1i8fNXGX1j8hNWGU1k8kNUGS1m8mNTGP1o8nNRGP1P9POQGn0P9ROPGm0Q9SOPGk0Q9TOPGk0R9TOoFj0R9VOnFi0S9WOnFg0S9XOnFg0T9XOmFf0T9ZOlFe0U9[OlFc0V9[OlFc0U9]OkFb0V9^OkF`0W9_OiFa0W9^OkF`0V9@kF>W9AiF>X9BiF<X9CiFeN[OQ1n99hF`NDQ1e9?hFYNMS1\\9d0hGWOZ8g0kGTOV8l0oGnNS8Q1QHjNP8V1THeNm7Z1YH`Ni7_1[H\\Nf7d1k201N10001O0O101O001N2O2N1O1O1N2O2N1O1N2O2N2N3L4M3M2N3L4M2N3M3LkjT8"}, "label": "the back end of a brown motorcycle"}]}
{"id": 241527, "image": "COCO_train2014_000000241527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"piece of fruit behind cake\"."}], "task": "refering", "answer_template": "The \"piece of fruit behind cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 144, 145, 146, 168, 169], "bbox": [0.27690625, 0.18641666666666667, 0.484859375, 0.4414166666666667], "iscrowd": 0, "area": 6313.928349999999, "rle": {"size": [480, 640], "counts": "STc21n>4M3L4L4M3L4L4L4M2M4L4L4M1N2N101N1O2N101N1O2O0O2N1O2O0O2N101N1O2N101N1O2OO01O1O100O00100O1O00100O1O00100OI7I8G8H8I7H9O01O100O001O100O00100O1O1O010O1O10O01O1O10O01O1O10O01O100O1O00100O1O010O1O1O010O1O10O01001O2N1N3N1O1O2M2O2N1O2M2O2N1O1N3N1O2N1O2M2O2N1O2M2O1O2N1N3Nlaj4"}, "label": "piece of fruit behind cake"}]}
{"id": 241527, "image": "COCO_train2014_000000241527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced pieces of fruit on top of a dessert\"."}], "task": "refering", "answer_template": "The \"sliced pieces of fruit on top of a dessert\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 171, 172, 173], "bbox": [0.3799375, 0.23139583333333333, 0.7095312500000001, 0.43683333333333335], "iscrowd": 0, "area": 9573.156799999999, "rle": {"size": [480, 640], "counts": "h`b31o>2M2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O0O2O1O1O1N2O001O1N2O1O1O0O2O1O1O1N2O001N2O1O1O1N101O1O1N2O001O1N2O1O1O0O2O1O1N2O1OO10O1000O10O100000O010000O10O1000O10O1000O1000O10O100000O01000O0M4M3L4M2M4M3M3L3N3L4M30O10001O0O10000O1000000O10001O0O1000000O10000O10001O0O1000000O1000000O101O000O10000O1000000O101O01O1O1O001O1O1O1O001O1O1N2O1O0O2O1O1O1N101O1N2O1O001N2O1O1N101O1O1N2O1O0O2O1O1O1N101O1N2O1O001N2O1O1N2O001N2OUQg2"}, "label": "sliced pieces of fruit on top of a dessert"}]}
{"id": 241527, "image": "COCO_train2014_000000241527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pieces of yellow fruit on top of cake\"."}], "task": "refering", "answer_template": "The \"pieces of yellow fruit on top of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 171, 172, 173], "bbox": [0.3799375, 0.23139583333333333, 0.7095312500000001, 0.43683333333333335], "iscrowd": 0, "area": 9573.156799999999, "rle": {"size": [480, 640], "counts": "h`b31o>2M2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O0O2O1O1O1N2O001O1N2O1O1O0O2O1O1O1N2O001N2O1O1O1N101O1O1N2O001O1N2O1O1O0O2O1O1N2O1OO10O1000O10O100000O010000O10O1000O10O1000O1000O10O100000O01000O0M4M3L4M2M4M3M3L3N3L4M30O10001O0O10000O1000000O10001O0O1000000O10000O10001O0O1000000O1000000O101O000O10000O1000000O101O01O1O1O001O1O1O1O001O1O1N2O1O0O2O1O1O1N101O1N2O1O001N2O1O1N101O1O1N2O1O0O2O1O1O1N101O1N2O1O001N2O1O1N2O001N2OUQg2"}, "label": "pieces of yellow fruit on top of cake"}]}
{"id": 577405, "image": "COCO_train2014_000000577405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange in front of bowl\"."}], "task": "refering", "answer_template": "The \"orange in front of bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 331], "bbox": [0.25940625, 0.4494166666666667, 0.5700000000000001, 0.8282291666666667], "iscrowd": 0, "area": 28126.108900000007, "rle": {"size": [480, 640], "counts": "jS^28b>?A`0@>B9J5L4L5L3L4L5K4L4L5K4L3M3M2N3M201N1O2N101N1O2N1O2O0O1O2N1O2O0O2N1O100O2N1O1O101N1O1O101N1O1O101O0O10001O000O101O0O100O2O0O100O2O0O100O2O0O100O100O2O0O100O100O100O100O100O100O2O0O100O1000000000000000001OO1000000O1000000O1000001N1000000O1000000O101O000O2O001O0O101O000O2O0O2O0O101O0O2O0O101N100O2O001N101N2O2M2O1N3N1O1N3N1N2O2M2O1N3N1O1N3N1N3N1N3N1N2N3M2N3L4M3M2M4M3M3M2M4M3K4L5J6J6K4K6Ja0@b0]OUcP4"}, "label": "orange in front of bowl"}]}
{"id": 577405, "image": "COCO_train2014_000000577405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange that is next to an apple and another orange and it is in front\"."}], "task": "refering", "answer_template": "The \"an orange that is next to an apple and another orange and it is in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 331], "bbox": [0.25940625, 0.4494166666666667, 0.5700000000000001, 0.8282291666666667], "iscrowd": 0, "area": 28126.108900000007, "rle": {"size": [480, 640], "counts": "jS^28b>?A`0@>B9J5L4L5L3L4L5K4L4L5K4L3M3M2N3M201N1O2N101N1O2N1O2O0O1O2N1O2O0O2N1O100O2N1O1O101N1O1O101N1O1O101O0O10001O000O101O0O100O2O0O100O2O0O100O2O0O100O100O2O0O100O100O100O100O100O100O2O0O100O1000000000000000001OO1000000O1000000O1000001N1000000O1000000O101O000O2O001O0O101O000O2O0O2O0O101O0O2O0O101N100O2O001N101N2O2M2O1N3N1O1N3N1N2O2M2O1N3N1O1N3N1N3N1N3N1N2N3M2N3L4M3M2M4M3M3M2M4M3K4L5J6J6K4K6Ja0@b0]OUcP4"}, "label": "an orange that is next to an apple and another orange and it is in front"}]}
{"id": 577405, "image": "COCO_train2014_000000577405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leftmost orange in the middle of a white plate with 3 other oranges , an apple , and a muffin setting on top of the fruit\"."}], "task": "refering", "answer_template": "The \"the leftmost orange in the middle of a white plate with 3 other oranges , an apple , and a muffin setting on top of the fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 258, 259, 280, 281, 282], "bbox": [0.12662500000000002, 0.41460416666666666, 0.34550000000000003, 0.7545833333333333], "iscrowd": 0, "area": 14196.564, "rle": {"size": [480, 640], "counts": "jWV16f>;D<E8G7J5J6K6I6K6I6J7J5M3O2M2O2M2N3N1N2O2M2O1N101N2O0O2N2O0O2O1N101N2O0O2O0O2O1N101N2O001N2O0O2O0O2O1O0O2O1N101O1N101N101O1O001N2O001O0O100000000000O01000000000000O100000O1000000000000mMdGbM\\8\\2iGaMX8[2nGbMR8\\2SHaMm7]2XH_Mi7_2ZH`Mf7]2`2N2N3M3L3N3M2O2N2N1O2O1N1O2N2N101N2N1O2N2O1N2N2N2N101N2N3M3M2O2M3M2N3M2N3N1N2N2NdUT6"}, "label": "the leftmost orange in the middle of a white plate with 3 other oranges , an apple , and a muffin setting on top of the fruit"}]}
{"id": 298878, "image": "COCO_train2014_000000298878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white tent that appears to be falling down and an orange flag\"."}], "task": "refering", "answer_template": "The \"a green and white tent that appears to be falling down and an orange flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125], "bbox": [0.0, 0.252, 0.41304449648711944, 0.39525], "iscrowd": 0, "area": 9633.325650000002, "rle": {"size": [640, 427], "counts": "c61nc03M2M3M3M3M3M3M3M3N10001N10001N10001O0O1000001N10000O101O000O10001O0O10000O2O00000O101O0O1000001N1000000O2O000O10001O0O1000001N10000O101O000O10001O0O10000O2O00000O101O0O1000001N1000000O2O000O10001O0O1000001N10000O101O000O10001N1000000O2O00000O101O0O1000001N1002N1O100O1O1O1O100O2N1O1O100O1O1O1O2O0O1O1O1O100O1O1O2N100O1O1M3L4M3L5K4M3L4M3L4M3L4M3L5L3L_el4"}, "label": "a green and white tent that appears to be falling down and an orange flag"}]}
{"id": 298878, "image": "COCO_train2014_000000298878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tilted green tent behind a girl\"."}], "task": "refering", "answer_template": "The \"a tilted green tent behind a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125], "bbox": [0.0, 0.252, 0.41304449648711944, 0.39525], "iscrowd": 0, "area": 9633.325650000002, "rle": {"size": [640, 427], "counts": "c61nc03M2M3M3M3M3M3M3M3N10001N10001N10001O0O1000001N10000O101O000O10001O0O10000O2O00000O101O0O1000001N1000000O2O000O10001O0O1000001N10000O101O000O10001O0O10000O2O00000O101O0O1000001N1000000O2O000O10001O0O1000001N10000O101O000O10001N1000000O2O00000O101O0O1000001N1002N1O100O1O1O1O100O2N1O1O100O1O1O1O2O0O1O1O1O100O1O1O2N100O1O1M3L4M3L5K4M3L4M3L4M3L4M3L5L3L_el4"}, "label": "a tilted green tent behind a girl"}]}
{"id": 388991, "image": "COCO_train2014_000000388991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s black pants\"."}], "task": "refering", "answer_template": "The \"a man ' s black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 203, 220, 236, 237, 252, 253, 254, 269, 270, 271, 285, 286, 287, 288, 303, 304, 305, 321, 322, 337, 338, 339, 354, 355, 356, 371, 372, 373, 388, 389, 390], "bbox": [0.79975, 0.44520312500000003, 1.0, 1.0], "iscrowd": 0, "area": 17166.34875, "rle": {"size": [640, 480], "counts": "b^`71mc03M2N3M2N3M2N3M2O2M2N3M2N3M2N2N3M2O2M2N3M2N3M2N3M2N3N1N3M2N2SCjMn7Y2YG_Ne8c1bFVO]9l0jELU:5SEd0k:^OkDk0U;UOiDm0V;UOfDn0Y;SOeDo0Z;TOaDo0];WO]Dk0b;ZOXDh0f;_ORDd0m;AmCa0Q<EiC=V<g3N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1M3M3N2M4L3N2M3M3N2M3M3N2dN`FUHc9j7lFeGW9Z8QG]GQ9b8WGTGl8m8S12N1O1]Nd1_N`1QOo0nNS1J5L5J5L4L5J5L5K4KaF"}, "label": "a man ' s black pants"}]}
{"id": 388991, "image": "COCO_train2014_000000388991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mens black slacks or pants\"."}], "task": "refering", "answer_template": "The \"mens black slacks or pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 203, 220, 236, 237, 252, 253, 254, 269, 270, 271, 285, 286, 287, 288, 303, 304, 305, 321, 322, 337, 338, 339, 354, 355, 356, 371, 372, 373, 388, 389, 390], "bbox": [0.79975, 0.44520312500000003, 1.0, 1.0], "iscrowd": 0, "area": 17166.34875, "rle": {"size": [640, 480], "counts": "b^`71mc03M2N3M2N3M2N3M2O2M2N3M2N3M2N2N3M2O2M2N3M2N3M2N3M2N3N1N3M2N2SCjMn7Y2YG_Ne8c1bFVO]9l0jELU:5SEd0k:^OkDk0U;UOiDm0V;UOfDn0Y;SOeDo0Z;TOaDo0];WO]Dk0b;ZOXDh0f;_ORDd0m;AmCa0Q<EiC=V<g3N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1M3M3N2M4L3N2M3M3N2M3M3N2dN`FUHc9j7lFeGW9Z8QG]GQ9b8WGTGl8m8S12N1O1]Nd1_N`1QOo0nNS1J5L5J5L4L5J5L5K4KaF"}, "label": "mens black slacks or pants"}]}
{"id": 388991, "image": "COCO_train2014_000000388991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a beard and glasses\"."}], "task": "refering", "answer_template": "The \"a man with a beard and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 320, 321, 322, 337, 338, 339, 354, 355, 356, 371, 372, 373, 388, 389, 390], "bbox": [0.2298125, 0.125609375, 1.0, 1.0], "iscrowd": 0, "area": 116033.11525000002, "rle": {"size": [640, 480], "counts": "oSU2;bc05K5K5K5K4L5H8G9K4K6K4M4L3M4L3M3L4M3M2N3L3N3M2N3L4M2N3L3N3M2N3L4M2N3M2M4L3M4K5L3M4L3M4L3M4L3M4L4M2N3M2N3M0000001O2O0O2O0O3N2ZO`IhCc6V<_IhCc6D_IY<0PDl6n;VIQDk6n;WIPDk6n;WIoCk6P<WInCk6P<WImCk6Q<c0O2N1O2N1jNWHTFj7h9eHkE^7P:QIdEP7W:_1L5J5L5L3N2N3M2M4M2N2O2N100O1O100O2N1O100O1O2O0O1O1O100O2N100O1O100O2N100O1O100O2N100O10000O2O00000O10000O1000001N10001N101O000O2O01O01O00001O00001O00001O001O01O01O00001O00010O00001O010O000010O0001O00010O00001O0001O01O01O00001O00010O00001O002O0O2N2N1O2N2O0O2N2N1O2N2N2N2N3M2N2N2N3M2N3M3M3Md0\\O00000000001N100000001O00001O0000001O00001O0000001O0O1O2N1O100O2N10001N10000O2O000O2O0O10010O000010O000010O00010O0000010O00010O00010O1O10O01O1O010O1O010O1O010OhFiFa1:k2l8cKlFl0m0b3W8`KmF9a1W4b7_KoFCU2o4k6\\KUGmNe2h5U6ZK^Jf4c5XK^Ji4a5VK`Jk4_5TKbJn4]5PKdJR5Z5mJgJT5X5kJiJW5V5gJkJ[5S5dJnJ^5R5_JoJb5g92N2N2N2N1O2N2N2N2N1O2N2N2N1O3M4L4L4L4L3M4L4L6J9G2N1O1O001O1O1O1O001O1O1O1O002N1O2N1O2N1O2N2N1O2N1O9GX2jLmD"}, "label": "a man with a beard and glasses"}]}
{"id": 388991, "image": "COCO_train2014_000000388991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a middle aged man that just blew out his candles on top of a cake\"."}], "task": "refering", "answer_template": "The \"a middle aged man that just blew out his candles on top of a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 320, 321, 322, 337, 338, 339, 354, 355, 356, 371, 372, 373, 388, 389, 390], "bbox": [0.2298125, 0.125609375, 1.0, 1.0], "iscrowd": 0, "area": 116033.11525000002, "rle": {"size": [640, 480], "counts": "oSU2;bc05K5K5K5K4L5H8G9K4K6K4M4L3M4L3M3L4M3M2N3L3N3M2N3L4M2N3L3N3M2N3L4M2N3M2M4L3M4K5L3M4L3M4L3M4L3M4L4M2N3M2N3M0000001O2O0O2O0O3N2ZO`IhCc6V<_IhCc6D_IY<0PDl6n;VIQDk6n;WIPDk6n;WIoCk6P<WInCk6P<WImCk6Q<c0O2N1O2N1jNWHTFj7h9eHkE^7P:QIdEP7W:_1L5J5L5L3N2N3M2M4M2N2O2N100O1O100O2N1O100O1O2O0O1O1O100O2N100O1O100O2N100O1O100O2N100O10000O2O00000O10000O1000001N10001N101O000O2O01O01O00001O00001O00001O001O01O01O00001O00010O00001O010O000010O0001O00010O00001O0001O01O01O00001O00010O00001O002O0O2N2N1O2N2O0O2N2N1O2N2N2N2N3M2N2N2N3M2N3M3M3Md0\\O00000000001N100000001O00001O0000001O00001O0000001O0O1O2N1O100O2N10001N10000O2O000O2O0O10010O000010O000010O00010O0000010O00010O00010O1O10O01O1O010O1O010O1O010OhFiFa1:k2l8cKlFl0m0b3W8`KmF9a1W4b7_KoFCU2o4k6\\KUGmNe2h5U6ZK^Jf4c5XK^Ji4a5VK`Jk4_5TKbJn4]5PKdJR5Z5mJgJT5X5kJiJW5V5gJkJ[5S5dJnJ^5R5_JoJb5g92N2N2N2N1O2N2N2N2N1O2N2N2N1O3M4L4L4L4L3M4L4L6J9G2N1O1O001O1O1O1O001O1O1O1O002N1O2N1O2N1O2N2N1O2N1O9GX2jLmD"}, "label": "a middle aged man that just blew out his candles on top of a cake"}]}
{"id": 85893, "image": "COCO_train2014_000000085893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a made bed with pink flip flops on it\"."}], "task": "refering", "answer_template": "The \"a made bed with pink flip flops on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328, 349, 350], "bbox": [0.0, 0.18133891213389122, 0.48168749999999994, 0.9521129707112971], "iscrowd": 0, "area": 58397.665199999996, "rle": {"size": [478, 640], "counts": "a9e0n<\\1oNP1M3L5L3O100O2O0O101N100O101N10001N100O2O0O100O2O0O101N100O101N100O2O0O101O0O100O2O0O101N100O2O0O100O2O0O101N100O101O0O101N100O2O0O100O2O0O101N100O2O0O100O2N100O2O0O1O100O2fGdJP7\\5nHhJo6Y5oHjJo6W5PIkJo6U5oHnJo6S5oHPKP7Q5mHSKP7n4oHTKo6m4oHVKP7j4nHYKP7h4nH[KQ7e4mH_KP7c4nH_KP7b4nHaKQ7_4mHdKQ7]4mHfKR7Z4mHiKP7Y4mHjKQ7W4mHlKR7T4lHoKR7R4mHPLR7P6M2N2O1N2O2M2N2O1N3N1N2O1N2N3N1N2O1N2N3N1N2O1N2N3N1N4M3LN2N3N1N3M2N2O2M2N3M2O1O2O0010O001N1N2O2M2N3M20001O001O00001OmM[JnJe5R5gJcJY5\\5SKXJm4h5^KoI`4Q6lKeIR4[6YL\\Ie3d6gLRIW3n6\\21O0001O0000000000000000001O0000000000000000001O00010O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N100O1O1O1O2N1O?POTlj4"}, "label": "a made bed with pink flip flops on it"}]}
{"id": 85893, "image": "COCO_train2014_000000085893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a first bed with pink colour sandals\"."}], "task": "refering", "answer_template": "The \"a first bed with pink colour sandals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328, 349, 350], "bbox": [0.0, 0.18133891213389122, 0.48168749999999994, 0.9521129707112971], "iscrowd": 0, "area": 58397.665199999996, "rle": {"size": [478, 640], "counts": "a9e0n<\\1oNP1M3L5L3O100O2O0O101N100O101N10001N100O2O0O100O2O0O101N100O101N100O2O0O101O0O100O2O0O101N100O2O0O100O2O0O101N100O101O0O101N100O2O0O100O2O0O101N100O2O0O100O2N100O2O0O1O100O2fGdJP7\\5nHhJo6Y5oHjJo6W5PIkJo6U5oHnJo6S5oHPKP7Q5mHSKP7n4oHTKo6m4oHVKP7j4nHYKP7h4nH[KQ7e4mH_KP7c4nH_KP7b4nHaKQ7_4mHdKQ7]4mHfKR7Z4mHiKP7Y4mHjKQ7W4mHlKR7T4lHoKR7R4mHPLR7P6M2N2O1N2O2M2N2O1N3N1N2O1N2N3N1N2O1N2N3N1N2O1N2N3N1N4M3LN2N3N1N3M2N2O2M2N3M2O1O2O0010O001N1N2O2M2N3M20001O001O00001OmM[JnJe5R5gJcJY5\\5SKXJm4h5^KoI`4Q6lKeIR4[6YL\\Ie3d6gLRIW3n6\\21O0001O0000000000000000001O0000000000000000001O00010O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N100O1O1O1O2N1O?POTlj4"}, "label": "a first bed with pink colour sandals"}]}
{"id": 85893, "image": "COCO_train2014_000000085893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed on the floor with a purple blanket and a black and white patterned pillow\"."}], "task": "refering", "answer_template": "The \"a bed on the floor with a purple blanket and a black and white patterned pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.377640625, 0.37684100418410044, 0.9012812499999999, 0.9768410041841005], "iscrowd": 0, "area": 64796.58655000001, "rle": {"size": [478, 640], "counts": "V]a31m>0O2O1N101N1O2L4K4M4K3N3L3N2M4M2N3L3N3L3M4L5K4M3001O001O1O001O001O001O010O00O1N2O1O1N2O1N2O1O1N2O1O2M2O1N2O1O1N2O1N2O1O1N2O1O1N2O2M2O1O1N2O1N2O1O1N2N2N2M3N2M3N2M3N2N2M3N2M3N2M3L4L4L4L4L5K4L4L4M34L5K4L5K1ON2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N3oNP1M3M3M3M3M3L4M3M3M3M3M3M3N2M3M3M3M3M3N2M3M3O11O001O001O010O001O00001O001O001O001O00001O0000001O000000001O000000001O01O000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O0000000O2O000000001O000000001O00000000001O0000000000000000000000000000O1000000000000000000000000000000O10000000000000YIWIY5i6eJYI[5o6\\JSIc5U7TJmHk5\\7jIfHV6c7`I_H_6P801O1O001O001O1O001M3N1O2Nf0ZOn0nJXFS3o:J4L3M3M3L5K4M3L4L5K4L4L4LZ1fNQPm0"}, "label": "a bed on the floor with a purple blanket and a black and white patterned pillow"}]}
{"id": 85893, "image": "COCO_train2014_000000085893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mattress on the ground with a pillow and a purple blanket\"."}], "task": "refering", "answer_template": "The \"a mattress on the ground with a pillow and a purple blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.377640625, 0.37684100418410044, 0.9012812499999999, 0.9768410041841005], "iscrowd": 0, "area": 64796.58655000001, "rle": {"size": [478, 640], "counts": "V]a31m>0O2O1N101N1O2L4K4M4K3N3L3N2M4M2N3L3N3L3M4L5K4M3001O001O1O001O001O001O010O00O1N2O1O1N2O1N2O1O1N2O1O2M2O1N2O1O1N2O1N2O1O1N2O1O1N2O2M2O1O1N2O1N2O1O1N2N2N2M3N2M3N2M3N2N2M3N2M3N2M3L4L4L4L4L5K4L4L4M34L5K4L5K1ON2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N3oNP1M3M3M3M3M3L4M3M3M3M3M3M3N2M3M3M3M3M3N2M3M3O11O001O001O010O001O00001O001O001O001O00001O0000001O000000001O000000001O01O000001O000000001O000000001O000000001O00000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O0000000O2O000000001O000000001O00000000001O0000000000000000000000000000O1000000000000000000000000000000O10000000000000YIWIY5i6eJYI[5o6\\JSIc5U7TJmHk5\\7jIfHV6c7`I_H_6P801O1O001O001O1O001M3N1O2Nf0ZOn0nJXFS3o:J4L3M3M3L5K4M3L4L5K4L4L4LZ1fNQPm0"}, "label": "a mattress on the ground with a pillow and a purple blanket"}]}
{"id": 429959, "image": "COCO_train2014_000000429959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"in a laptop movie playing a man talking phone another man near by him\"."}], "task": "refering", "answer_template": "The \"in a laptop movie playing a man talking phone another man near by him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.2703541666666667, 0.37434375000000003, 0.9098125, 0.69803125], "iscrowd": 0, "area": 59893.5249, "rle": {"size": [640, 480], "counts": "l_a2k0Tc0W1jN\\1dNQ1oNg0YOg0XO8I000000000O100000O10O10000000000O1000000000000000000000000000000000000O1000000000O10000000000000000000000000O100000000000000000000000000000000000O01000000000000000000000000000000000000O100000000000000000000000O10000000000000000000O100000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O100000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000001O3M3M3M4Ld0\\Od0[Oe0WOW1jNo0POf0[Od0[O]_j0"}, "label": "in a laptop movie playing a man talking phone another man near by him"}]}
{"id": 429959, "image": "COCO_train2014_000000429959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a video playing on the screen of a mac book pro , which is resting on a man ' s lap\"."}], "task": "refering", "answer_template": "The \"a video playing on the screen of a mac book pro , which is resting on a man ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.2703541666666667, 0.37434375000000003, 0.9098125, 0.69803125], "iscrowd": 0, "area": 59893.5249, "rle": {"size": [640, 480], "counts": "l_a2k0Tc0W1jN\\1dNQ1oNg0YOg0XO8I000000000O100000O10O10000000000O1000000000000000000000000000000000000O1000000000O10000000000000000000000000O100000000000000000000000000000000000O01000000000000000000000000000000000000O100000000000000000000000O10000000000000000000O100000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O100000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000001O3M3M3M4Ld0\\Od0[Oe0WOW1jNo0POf0[Od0[O]_j0"}, "label": "a video playing on the screen of a mac book pro , which is resting on a man ' s lap"}]}
{"id": 176008, "image": "COCO_train2014_000000176008.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottom bunk\"."}], "task": "refering", "answer_template": "The \"bottom bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.08689583333333334, 0.5754374999999999, 0.9468125, 0.9731874999999999], "iscrowd": 0, "area": 80577.29910000003, "rle": {"size": [640, 480], "counts": "hej0a0kb0h0XOg0L5L3N3L4M2M3M3M3L4M3M3L4M3L4M3M3L4M3M3L4M3M3L3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N1O1N101N101O1N101O1N101O1N101O0O2O1N101O1N101O0O2O1O0O2O1N101O0O2O1O0O2O1O001O1O001N101O00000000000000000000O01000000000000000000000000000000000000O10000000000001O00001O001O00001O0O2O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O0O101O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O0O2O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O0O101O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O000O2O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001N10001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O000O2O001O00001O001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1OY1gNTg?"}, "label": "bottom bunk"}]}
{"id": 176008, "image": "COCO_train2014_000000176008.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottom bunk\"."}], "task": "refering", "answer_template": "The \"a bottom bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.08689583333333334, 0.5754374999999999, 0.9468125, 0.9731874999999999], "iscrowd": 0, "area": 80577.29910000003, "rle": {"size": [640, 480], "counts": "hej0a0kb0h0XOg0L5L3N3L4M2M3M3M3L4M3M3L4M3L4M3M3L4M3M3L4M3M3L3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N1O1N101N101O1N101O1N101O1N101O0O2O1N101O1N101O0O2O1O0O2O1N101O0O2O1O0O2O1O001O1O001N101O00000000000000000000O01000000000000000000000000000000000000O10000000000001O00001O001O00001O0O2O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O0O101O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O0O2O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O0O101O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O000O2O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001N10001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O000O2O001O00001O001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1OY1gNTg?"}, "label": "a bottom bunk"}]}
{"id": 176008, "image": "COCO_train2014_000000176008.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a upper deck bed with rails\"."}], "task": "refering", "answer_template": "The \"a upper deck bed with rails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 137, 138, 139], "bbox": [0.0555625, 0.173171875, 1.0, 0.360671875], "iscrowd": 0, "area": 28599.700000000004, "rle": {"size": [640, 480], "counts": "dRa0c0]c0000O01000O100O10000O1000O010000O10000O10000O10O10O100O10000O10000O01000O1O1M3M3M3M3M3M3N1N3M3M3M3M3N2O10000O01000O10000\\Od000001O001N10001O001O001O00001N101O00001O001O00001O0O2O00000000000000000O100000000000000000000O100000O100000000000O100000000000000000000O1000000000000000000O10000000O100000000000O1000000000000000000O100000000000000000000O10000000O1000000000O100000000000000000000O10000000001O2N3M2N2M2O0000000000O10000000O0100000000O10000000O10O100000000O10000000O010000000000O100000O10O10000000000O100000O10O100000000O100000O1000O100000000O100000O10O10000000000O1000O1000O100000000O100000O1000O100000000O1000O1000O10000000000O1000O1000O100000000O1000O100000O100000000O1000O1000O10000000000O10O100000O100000000O10O10000000O100000000O10O100000O10000000000O010000000O100000000O10O10000000O100000000O010000000O10000000000O010000000O100000000O01000000000O100000000O010000000O1000000000O0100000000O1000000aL"}, "label": "a upper deck bed with rails"}]}
{"id": 102281, "image": "COCO_train2014_000000102281.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy wearing a black shirt with a maroon tie\"."}], "task": "refering", "answer_template": "The \"the boy wearing a black shirt with a maroon tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.5005625, 0.0022708333333333335, 0.917984375, 0.9705833333333334], "iscrowd": 0, "area": 79269.602, "rle": {"size": [480, 640], "counts": "URf41n>2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2N2N2M3N3M2N2M6K>B>B=C1N2O2N1O2M2O2N1O1N3WOcKlF^4o8gKPG[4k8iKTGX4g8mKWGU4d8PL[GR4_8SL`Gn3[8WLeGj3V8ZLjGf3Q8_LoGa3l7dLTH]3f7hLZHX3a7mL_HT3\\7PMdHP3W7UMjHk2m6]MSIc2\\6nMdIR2k5_NUJc1Y5mNgJT1f4^OZKd0S4MmK5f36ZLK[3?eLCn2h0RMYOc2Q1]MQOW2Y1iMhNk1c1UN_Na1i1_NYN^1h1bN\\NX1f1iN\\NS1e1mN_Nm0c1SOaNh0`1XOdNb0^1^OfN=[1ChN8Z1HjN3W1MmNMU13oNHR18ROBP1>i600000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O001O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1^DRNP:P2YEgNe:l2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2eJVHQ31\\Kk7_1`Hh2IhKj7[1jH`2_OTLi7X1UIW2UO`Li7T1_Io1kNlLh7Q1jIe1aNZMh7l0TJ]1WNfMg7i0_JT1mMRNg7e0hJm0dM]Nf7b0SKc0P6XO]J;g5AfJ2_5Io4LdSh0"}, "label": "the boy wearing a black shirt with a maroon tie"}]}
{"id": 102281, "image": "COCO_train2014_000000102281.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man dressed in black getting help with his red tie\"."}], "task": "refering", "answer_template": "The \"a young man dressed in black getting help with his red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.5005625, 0.0022708333333333335, 0.917984375, 0.9705833333333334], "iscrowd": 0, "area": 79269.602, "rle": {"size": [480, 640], "counts": "URf41n>2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2N2N2M3N3M2N2M6K>B>B=C1N2O2N1O2M2O2N1O1N3WOcKlF^4o8gKPG[4k8iKTGX4g8mKWGU4d8PL[GR4_8SL`Gn3[8WLeGj3V8ZLjGf3Q8_LoGa3l7dLTH]3f7hLZHX3a7mL_HT3\\7PMdHP3W7UMjHk2m6]MSIc2\\6nMdIR2k5_NUJc1Y5mNgJT1f4^OZKd0S4MmK5f36ZLK[3?eLCn2h0RMYOc2Q1]MQOW2Y1iMhNk1c1UN_Na1i1_NYN^1h1bN\\NX1f1iN\\NS1e1mN_Nm0c1SOaNh0`1XOdNb0^1^OfN=[1ChN8Z1HjN3W1MmNMU13oNHR18ROBP1>i600000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O001O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1^DRNP:P2YEgNe:l2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2eJVHQ31\\Kk7_1`Hh2IhKj7[1jH`2_OTLi7X1UIW2UO`Li7T1_Io1kNlLh7Q1jIe1aNZMh7l0TJ]1WNfMg7i0_JT1mMRNg7e0hJm0dM]Nf7b0SKc0P6XO]J;g5AfJ2_5Io4LdSh0"}, "label": "a young man dressed in black getting help with his red tie"}]}
{"id": 102281, "image": "COCO_train2014_000000102281.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in striped shirt adjusting a boy ' s tie\"."}], "task": "refering", "answer_template": "The \"a woman in striped shirt adjusting a boy ' s tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 27, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.03370833333333333, 0.611796875, 0.9887708333333334], "iscrowd": 0, "area": 102053.40379999996, "rle": {"size": [480, 640], "counts": "U>f0X>2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2O1N2N2L4L4ZN[MmFi2h8mMjFV2l8]NgFg1P9eNiF_1o8kNkFY1m8ROmFQ1m7d3K5L4L4M3M3N2M3N2O1N2N2O1N2O1O1O1O1M3K5O1N2VOfGlJ\\8S5lGeJU8Z5nGbJT8\\5oGaJS8^5h0O1\\Od000O10000000000000000O10000000000000000O1000000000000000000000000000000000000000000001O00O1000000lMjE^OV:`0nE^OR:`0QF_Oo9?UF_Ok9?XF@h9>[FAe9=_FAa9=bFA_9=eFA[9=hFBX9<lFBT9<oFCQ9;SGCm8;VGDj8:YGEg89]GEc89`GF`88cGG]87fGG[86iGIW85lGIU84nGLR81RHMo71THNl71VHMk73UHMk72WHLj73XHLh74YHJh75ZHJf75[HJf76[HHf77\\HFf7:[HDf7;\\HCe7<\\HBf7>[H@f7?\\H_Oe7a0\\H]Oe7b0\\H\\Of7c0\\H[Oe7e0[3O10000000000000000000000000000001O2N1O1O1O2N1O1O2N1O2N00000000000000000000O10000O10000O100O1000000000000000000001O00000000001O000001O01O00000000000000001O001O00001O001O00001iFWOU4i0kKYOS4g0lK]Oo3f0oK]Ol3f0SL^Og3f0WL^Oc3e0\\L^O_3f0_L^O;kNZNk1Y1^O:TOmMa1h1^O9^O]MZ1X2\\O8AUMX1a2[O6EnLT1k2[O4HeLS1U3XO4L\\LQ1]3XO4NSLo0h3WO2m2LVM1l2MVM2j2MWM3j2KXM4i2IZM5g2G^M8c2CbM<`2_OdM`0]2[OfMe0\\2WOfMj0[2SOgMm0Z2POgMQ1Z2kNhMV1Y2gNhMY1Z2cNiM]1X2`NiMa1X2\\NjMc1X2YNjMh1W2UNjMl1W2QNkMn1W2oMjMR2W2kMkMT2W2iMjMX2W2eMjM\\2W2aMkM^2W2_MjMb2W2[LjNe3Y1YLgNg3X610O01O010O1O010O000001O000000001OO1000000O10O1000O10O10000O101O1N2O001N2O1N101O1N2N101N2N2O0O2O1N2N101N2N2N1O3M2N2O1O1O100O1O1O1O2N11O1O2N1O2N1O2N00O2O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N1O2N2N1M4L4L3L5L4L4L4L=B?Booc3"}, "label": "a woman in striped shirt adjusting a boy ' s tie"}]}
{"id": 102281, "image": "COCO_train2014_000000102281.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a striped shirt\"."}], "task": "refering", "answer_template": "The \"a girl wearing a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 27, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.03370833333333333, 0.611796875, 0.9887708333333334], "iscrowd": 0, "area": 102053.40379999996, "rle": {"size": [480, 640], "counts": "U>f0X>2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2O1N2N2L4L4ZN[MmFi2h8mMjFV2l8]NgFg1P9eNiF_1o8kNkFY1m8ROmFQ1m7d3K5L4L4M3M3N2M3N2O1N2N2O1N2O1O1O1O1M3K5O1N2VOfGlJ\\8S5lGeJU8Z5nGbJT8\\5oGaJS8^5h0O1\\Od000O10000000000000000O10000000000000000O1000000000000000000000000000000000000000000001O00O1000000lMjE^OV:`0nE^OR:`0QF_Oo9?UF_Ok9?XF@h9>[FAe9=_FAa9=bFA_9=eFA[9=hFBX9<lFBT9<oFCQ9;SGCm8;VGDj8:YGEg89]GEc89`GF`88cGG]87fGG[86iGIW85lGIU84nGLR81RHMo71THNl71VHMk73UHMk72WHLj73XHLh74YHJh75ZHJf75[HJf76[HHf77\\HFf7:[HDf7;\\HCe7<\\HBf7>[H@f7?\\H_Oe7a0\\H]Oe7b0\\H\\Of7c0\\H[Oe7e0[3O10000000000000000000000000000001O2N1O1O1O2N1O1O2N1O2N00000000000000000000O10000O10000O100O1000000000000000000001O00000000001O000001O01O00000000000000001O001O00001O001O00001iFWOU4i0kKYOS4g0lK]Oo3f0oK]Ol3f0SL^Og3f0WL^Oc3e0\\L^O_3f0_L^O;kNZNk1Y1^O:TOmMa1h1^O9^O]MZ1X2\\O8AUMX1a2[O6EnLT1k2[O4HeLS1U3XO4L\\LQ1]3XO4NSLo0h3WO2m2LVM1l2MVM2j2MWM3j2KXM4i2IZM5g2G^M8c2CbM<`2_OdM`0]2[OfMe0\\2WOfMj0[2SOgMm0Z2POgMQ1Z2kNhMV1Y2gNhMY1Z2cNiM]1X2`NiMa1X2\\NjMc1X2YNjMh1W2UNjMl1W2QNkMn1W2oMjMR2W2kMkMT2W2iMjMX2W2eMjM\\2W2aMkM^2W2_MjMb2W2[LjNe3Y1YLgNg3X610O01O010O1O010O000001O000000001OO1000000O10O1000O10O10000O101O1N2O001N2O1N101O1N2N101N2N2O0O2O1N2N101N2N2N1O3M2N2O1O1O100O1O1O1O2N11O1O2N1O2N1O2N00O2O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N1O2N2N1M4L4L3L5L4L4L4L=B?Booc3"}, "label": "a girl wearing a striped shirt"}]}
{"id": 323475, "image": "COCO_train2014_000000323475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple on the top left\"."}], "task": "refering", "answer_template": "The \"the apple on the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 119, 120, 121, 122, 136, 137, 138, 154, 155, 156], "bbox": [0.028687499999999998, 0.179640625, 0.3579166666666667, 0.423734375], "iscrowd": 0, "area": 12049.459550000001, "rle": {"size": [640, 480], "counts": "km8g0Vc06K5J6K5J6K5K5K5L4K5K3M3M3M3N2M2N3M3N2N2N2O1N1O2N2O1N2N2N2O0O2N2N2N2O1N1O2N2O1N2N1L5O1N2N2N1O2O0O001O10O01O1O00100O10O10000O10000O10O10_Nk@VNV?g1VAPNj>m1bAjM^>T2gAiMY>T2mAiMS>T2TBhMl=V2YBgMg=V2_BgMa=X2cBeM]=Y2iBcMW=\\2W2O1O1O1O1O1O1O1O1O1O1O2N1N200O100O2O0O2O001N100O2O0O2O0O101N101N10001N101N101N101O0O2O1N101N2O0O2O001N2O0O2O1O0O2O001N2O0O2O1O0O2O1N101O1N101O0O2O1N101O1N10ZoP6"}, "label": "the apple on the top left"}]}
{"id": 323475, "image": "COCO_train2014_000000323475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple behind the large orange\"."}], "task": "refering", "answer_template": "The \"the apple behind the large orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 238, 239, 240, 241, 255, 256, 257], "bbox": [0.0, 0.41176562499999997, 0.3733333333333333, 0.698828125], "iscrowd": 0, "area": 20320.15545, "rle": {"size": [640, 480], "counts": "d;o0Va0l1G9N2M3M4L3M3M3M3M4L3N2M3N3N1N2O1N2O2N1N2O1N3N1N2O1O1N101N100O101O0O100O2O0O10001N100O10001O0O10001O000O101O00000O2O000000001N1000001O0O10001O000O10001O0000K4M4L4K5L4L4K5L4L4K5N2O2N1O1O1N3N1O2N2N2N2M3N2N2N2N2N2N1N3N2N2N2N2N2N2M2O2N2N1O2N2N2N101N2N1O2O1N1O1O100O1O1O101N1O1O100O1O1O100O1000000O1000000O1000000O2O00002M2O2N1O2M2O2N1O2M3N1O2N1N3N1O2N4K6K4L5K_ik5"}, "label": "the apple behind the large orange"}]}
{"id": 323475, "image": "COCO_train2014_000000323475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple in a bowl with other fruit\"."}], "task": "refering", "answer_template": "The \"an apple in a bowl with other fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 238, 239, 240, 241, 255, 256, 257], "bbox": [0.0, 0.41176562499999997, 0.3733333333333333, 0.698828125], "iscrowd": 0, "area": 20320.15545, "rle": {"size": [640, 480], "counts": "d;o0Va0l1G9N2M3M4L3M3M3M3M4L3N2M3N3N1N2O1N2O2N1N2O1N3N1N2O1O1N101N100O101O0O100O2O0O10001N100O10001O0O10001O000O101O00000O2O000000001N1000001O0O10001O000O10001O0000K4M4L4K5L4L4K5L4L4K5N2O2N1O1O1N3N1O2N2N2N2M3N2N2N2N2N2N1N3N2N2N2N2N2N2M2O2N2N1O2N2N2N101N2N1O2O1N1O1O100O1O1O101N1O1O100O1O1O100O1000000O1000000O1000000O2O00002M2O2N1O2M2O2N1O2M3N1O2N1N3N1O2N4K6K4L5K_ik5"}, "label": "an apple in a bowl with other fruit"}]}
{"id": 520087, "image": "COCO_train2014_000000520087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sliced carrots at the top right of the dish\"."}], "task": "refering", "answer_template": "The \"the sliced carrots at the top right of the dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 83, 105, 106, 107, 128, 129, 130], "bbox": [0.46359374999999997, 0.11770833333333333, 0.67484375, 0.36281250000000004], "iscrowd": 0, "area": 4857.890649999999, "rle": {"size": [480, 640], "counts": "`h[41o>001O001O0O2O001O001O001O001O0O2O001O001O001O001O0O2M2N3M2O2M2N3N1N3N1N3N1N3O0010O01O010O0010O010O01O010O0010O010O01O101N2N2O1N2O2M2N2O1N201N1O1O1O1O1OO2N100O1O100O101N1O100O101N1O101N100O20O7I7J1NM3M3M4L3M3M4L3N2M4L3M3M4N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O\\jQ3"}, "label": "the sliced carrots at the top right of the dish"}]}
{"id": 520087, "image": "COCO_train2014_000000520087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced carrots in a blue dish with other vegetables\"."}], "task": "refering", "answer_template": "The \"sliced carrots in a blue dish with other vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 83, 105, 106, 107, 128, 129, 130], "bbox": [0.46359374999999997, 0.11770833333333333, 0.67484375, 0.36281250000000004], "iscrowd": 0, "area": 4857.890649999999, "rle": {"size": [480, 640], "counts": "`h[41o>001O001O0O2O001O001O001O001O0O2O001O001O001O001O0O2M2N3M2O2M2N3N1N3N1N3N1N3O0010O01O010O0010O010O01O010O0010O010O01O101N2N2O1N2O2M2N2O1N201N1O1O1O1O1OO2N100O1O100O101N1O100O101N1O101N100O20O7I7J1NM3M3M4L3M3M4L3N2M4L3M3M4N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O\\jQ3"}, "label": "sliced carrots in a blue dish with other vegetables"}]}
{"id": 520092, "image": "COCO_train2014_000000520092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a note on a vase\"."}], "task": "refering", "answer_template": "The \"a note on a vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 83, 84, 85, 86, 87, 88, 89, 90, 98, 99, 100, 101, 102, 103, 104, 105, 106, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 146, 147, 148, 149, 150, 151, 152, 153, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 273, 274, 275, 276, 277, 278, 279, 280, 281], "bbox": [0.08286036036036036, 0.13976, 0.6497072072072073, 1.0], "iscrowd": 0, "area": 82255.95200000002, "rle": {"size": [500, 444], "counts": "Q`b0:X?;E5K6K5J5K6J3M3N2M3M3M3K5K5L3`IbNfNc1W1bNcNb1Z1cN`Na1^1cN]N`1`1eNZN_1c1gNWN\\1f1iNUNZ1h1kNSNX1k1lNoMX1n1nNlMU1Q2POjMR1U2ROfMP1X2UOdMl0[2XOaMi0^2\\O\\Mf0c2^OYMc0f2AUMa0j2ATM`0j2DSM=l2EQM=n2FoL;P3HmL9R3IkL9T3JiL7V3KhL6W3MeL5Z3MdL4\\3N_L5`3M]L5c3MXL6g3LVL6j3LRL6m3LoK7P4LlK6T4KhK8W4KeK7[4JbK8]4K[K;e4HRK>m4U5000000O100000000000000000000O100000000000000000000O100000000000000O1000000000000O1000000000000O10000N2O1N2N2O1N2O1N2N2O1N2O1N2N2N2N2M3M3N2M3M3N2N2N2N2M3N2N2N2M3M3N2M3M3M3N2O1O1O1O1O1O100000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O100O1O100O1O1O1N2O1O001O1N2O1O1hMX2E;E;`N`1_Oa0D=C=QOo0J6J5K6G9hNX1E:F;dMcD6g;FaDKj;1^D@m;<g1N4JQS\\2"}, "label": "a note on a vase"}]}
{"id": 520092, "image": "COCO_train2014_000000520092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase to the left of the 3 smaller pictures\"."}], "task": "refering", "answer_template": "The \"the vase to the left of the 3 smaller pictures\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 83, 84, 85, 86, 87, 88, 89, 90, 98, 99, 100, 101, 102, 103, 104, 105, 106, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 146, 147, 148, 149, 150, 151, 152, 153, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 273, 274, 275, 276, 277, 278, 279, 280, 281], "bbox": [0.08286036036036036, 0.13976, 0.6497072072072073, 1.0], "iscrowd": 0, "area": 82255.95200000002, "rle": {"size": [500, 444], "counts": "Q`b0:X?;E5K6K5J5K6J3M3N2M3M3M3K5K5L3`IbNfNc1W1bNcNb1Z1cN`Na1^1cN]N`1`1eNZN_1c1gNWN\\1f1iNUNZ1h1kNSNX1k1lNoMX1n1nNlMU1Q2POjMR1U2ROfMP1X2UOdMl0[2XOaMi0^2\\O\\Mf0c2^OYMc0f2AUMa0j2ATM`0j2DSM=l2EQM=n2FoL;P3HmL9R3IkL9T3JiL7V3KhL6W3MeL5Z3MdL4\\3N_L5`3M]L5c3MXL6g3LVL6j3LRL6m3LoK7P4LlK6T4KhK8W4KeK7[4JbK8]4K[K;e4HRK>m4U5000000O100000000000000000000O100000000000000000000O100000000000000O1000000000000O1000000000000O10000N2O1N2N2O1N2O1N2N2O1N2O1N2N2N2N2M3M3N2M3M3N2N2N2N2M3N2N2N2M3M3N2M3M3M3N2O1O1O1O1O1O100000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O100O1O100O1O1O1N2O1O001O1N2O1O1hMX2E;E;`N`1_Oa0D=C=QOo0J6J5K6G9hNX1E:F;dMcD6g;FaDKj;1^D@m;<g1N4JQS\\2"}, "label": "the vase to the left of the 3 smaller pictures"}]}
{"id": 397212, "image": "COCO_train2014_000000397212.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat on the inside looking at a cat on the outside\"."}], "task": "refering", "answer_template": "The \"a cat on the inside looking at a cat on the outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 44, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.0016200000000000001, 0.0, 0.51724, 1.0], "iscrowd": 0, "area": 66806.17705, "rle": {"size": [375, 500], "counts": "V=W:2^10000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000001O00001O1O001O001O1O001O00100O001O001O1O001O001O1O001O000000001O00000000000000001O000000000000001O000000000eMC\\I>[6LcI4X63fIMY66eIJZ69dIGZ6<eIDZ6?dIBY6b0eI^OZ6e0dI[OZ6i0dIWO\\6k0bIUO]6n0aISO]6o0bIQO]6R1aInN^6U1`IkN_6X1_IhN_6\\1_IdN_6`1_IaN^6c1`I]N]6g1bIYN\\6k1bIUN[6P2cIPN[6T2cIlM[6X2cIiMh5m2VJSM^5[3`JeL_5]3`JcL_5`3_J`L`5c3^J^L_5f3_JZL`5i3^JWLa5l3]JTLb5o3\\JQLb5R4]JnKb5U5N1O1N2O1N2N3M2N2N2N2N3M2N2N2N2N2N3cMRIMP7LbIG`61SJBo5;h2N2M4M2M3O1O1O100O100O100O1O100O100O1O101N1O100O100O1O100O1O100O100O1O100O1O100O100O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2Njbh2"}, "label": "a cat on the inside looking at a cat on the outside"}]}
{"id": 397212, "image": "COCO_train2014_000000397212.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat looks at another cat outside the window\"."}], "task": "refering", "answer_template": "The \"a cat looks at another cat outside the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 44, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.0016200000000000001, 0.0, 0.51724, 1.0], "iscrowd": 0, "area": 66806.17705, "rle": {"size": [375, 500], "counts": "V=W:2^10000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000001O00001O1O001O001O1O001O00100O001O001O1O001O001O1O001O000000001O00000000000000001O000000000000001O000000000eMC\\I>[6LcI4X63fIMY66eIJZ69dIGZ6<eIDZ6?dIBY6b0eI^OZ6e0dI[OZ6i0dIWO\\6k0bIUO]6n0aISO]6o0bIQO]6R1aInN^6U1`IkN_6X1_IhN_6\\1_IdN_6`1_IaN^6c1`I]N]6g1bIYN\\6k1bIUN[6P2cIPN[6T2cIlM[6X2cIiMh5m2VJSM^5[3`JeL_5]3`JcL_5`3_J`L`5c3^J^L_5f3_JZL`5i3^JWLa5l3]JTLb5o3\\JQLb5R4]JnKb5U5N1O1N2O1N2N3M2N2N2N2N3M2N2N2N2N2N3cMRIMP7LbIG`61SJBo5;h2N2M4M2M3O1O1O100O100O100O1O100O100O1O101N1O100O100O1O100O1O100O100O1O100O1O100O100O1O2O0O1O1O1O1O2N1O1O1O1O1O1O2Njbh2"}, "label": "a cat looks at another cat outside the window"}]}
{"id": 397212, "image": "COCO_train2014_000000397212.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and black striped cat with green eyes\"."}], "task": "refering", "answer_template": "The \"a brown and black striped cat with green eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 31, 32, 33, 35, 49, 50, 51, 52, 53, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 232], "bbox": [0.72096, 0.05376, 1.0, 0.9419466666666667], "iscrowd": 0, "area": 35470.53555, "rle": {"size": [375, 500], "counts": "SST4d0a:d0E;G8J7I7SLUNeMR2Q2`NbMd1[2eN[M`1`2kNTM[1h2mNoLW1m2ROhLU1S3TOaLS1\\3UOlK^1P4W3M3L5gN^HWNe7g1_HSNe7j1`HQNc7m1aHmMd7o1fHfM^7X2Q1O2O0O2N11O10O0100O010O1O10O01O10O01O100O2N2O1N2N2O1N2N0010O01O10O01O010O0O2O001O001O001O001O1O0000001O0000000O101O00000000001O01N4M4L3M2N1O1O1N2O1O1O001O000O2O00001N100O2O0O101N10O0O2N101N1O2O0O2N1O1SI^Mf4c2PKkMj4V2mJXNm4j1iJcNR5^1lJiNo4W1oJPOk4R1SKTOg4m0WKXOe4j0XKZOe4i0WK\\Oc4i0YK[Ob4i0[K\\O_4i0]K[O]4k0_KZO[4k0aKYOZ4V4TNk1mNT1POPL"}, "label": "a brown and black striped cat with green eyes"}]}
{"id": 397212, "image": "COCO_train2014_000000397212.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and black striped cat sitting on an outside deck looking through a window\"."}], "task": "refering", "answer_template": "The \"a brown and black striped cat sitting on an outside deck looking through a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 31, 32, 33, 35, 49, 50, 51, 52, 53, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 232], "bbox": [0.72096, 0.05376, 1.0, 0.9419466666666667], "iscrowd": 0, "area": 35470.53555, "rle": {"size": [375, 500], "counts": "SST4d0a:d0E;G8J7I7SLUNeMR2Q2`NbMd1[2eN[M`1`2kNTM[1h2mNoLW1m2ROhLU1S3TOaLS1\\3UOlK^1P4W3M3L5gN^HWNe7g1_HSNe7j1`HQNc7m1aHmMd7o1fHfM^7X2Q1O2O0O2N11O10O0100O010O1O10O01O10O01O100O2N2O1N2N2O1N2N0010O01O10O01O010O0O2O001O001O001O001O1O0000001O0000000O101O00000000001O01N4M4L3M2N1O1O1N2O1O1O001O000O2O00001N100O2O0O101N10O0O2N101N1O2O0O2N1O1SI^Mf4c2PKkMj4V2mJXNm4j1iJcNR5^1lJiNo4W1oJPOk4R1SKTOg4m0WKXOe4j0XKZOe4i0WK\\Oc4i0YK[Ob4i0[K\\O_4i0]K[O]4k0_KZO[4k0aKYOZ4V4TNk1mNT1POPL"}, "label": "a brown and black striped cat sitting on an outside deck looking through a window"}]}
{"id": 53150, "image": "COCO_train2014_000000053150.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car directly behind the man in the black shirt\"."}], "task": "refering", "answer_template": "The \"a blue car directly behind the man in the black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 223, 224, 225, 226, 241, 242, 246, 247, 248, 249, 265, 269, 270, 271, 272], "bbox": [0.5164375, 0.5483138173302108, 0.8686875000000001, 0.8011007025761124], "iscrowd": 0, "area": 7510.7676, "rle": {"size": [427, 640], "counts": "ZZZ4=l<3M3M3N20O100O1O2O2M2O1N2N2O1N2OgXl0I_gSO8H3M4L7IO010O1O1O2O1N1O2O0O2O0O2O2M2O2M3Njm72SRH3M4K5`DAm9b0iEOm94jE<l9GnE>o9ElE`0R:ChEc0V:_OdEf0Z:X1M3N0O01O010O1O2N1G9N2O1O2N100O2O0001O01O000010O0000010O00010O2N2O0O001O010OO2O001O000O2O001O1N101O1O0O2O000O100O010N2N2O1N2O100O01000O1000O1000000O2O00000O1000WNhD_1X;`NjD_1U;aNnD]1R;bNPE]1P;bNSE\\1m:dNUEZ1j:dNZE[1a;J8DRhR1"}, "label": "a blue car directly behind the man in the black shirt"}]}
{"id": 53150, "image": "COCO_train2014_000000053150.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car behind the man walking an elephant\"."}], "task": "refering", "answer_template": "The \"a blue car behind the man walking an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 223, 224, 225, 226, 241, 242, 246, 247, 248, 249, 265, 269, 270, 271, 272], "bbox": [0.5164375, 0.5483138173302108, 0.8686875000000001, 0.8011007025761124], "iscrowd": 0, "area": 7510.7676, "rle": {"size": [427, 640], "counts": "ZZZ4=l<3M3M3N20O100O1O2O2M2O1N2N2O1N2OgXl0I_gSO8H3M4L7IO010O1O1O2O1N1O2O0O2O0O2O2M2O2M3Njm72SRH3M4K5`DAm9b0iEOm94jE<l9GnE>o9ElE`0R:ChEc0V:_OdEf0Z:X1M3N0O01O010O1O2N1G9N2O1O2N100O2O0001O01O000010O0000010O00010O2N2O0O001O010OO2O001O000O2O001O1N101O1O0O2O000O100O010N2N2O1N2O100O01000O1000O1000000O2O00000O1000WNhD_1X;`NjD_1U;aNnD]1R;bNPE]1P;bNSE\\1m:dNUEZ1j:dNZE[1a;J8DRhR1"}, "label": "a blue car behind the man walking an elephant"}]}
{"id": 53150, "image": "COCO_train2014_000000053150.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was on black t - shirt\"."}], "task": "refering", "answer_template": "The \"a man was on black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 199, 200, 222, 223, 245, 246, 268, 269, 291, 292, 293, 314, 315, 316, 338, 339], "bbox": [0.6656875, 0.5066042154566744, 0.7738750000000001, 0.9885948477751756], "iscrowd": 0, "area": 6133.440300000003, "rle": {"size": [427, 640], "counts": "Qma56Q=9F;E;G9I7H9G8]EQN\\9W2YFRNb9U2RFUNh9l2J6ROl0M3K5N2O0O2N2N2N2N1O1OO1N2O1O1N2O2M2ZLkGU2V8hMQH]2f7aMaH_2[7]MlHa2R7\\MUIVO2T2g6cNlJU1R5iNSKV1l4fNYKZ1e4dN_K\\1Z80O0000O2N1O1O1O1O1O1O1O1O1O1N2N2N2M3N2N3M2M3N3M2N3L5L5K4L5JmQl1"}, "label": "a man was on black t - shirt"}]}
{"id": 53150, "image": "COCO_train2014_000000053150.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant trainer standing beside an elephant walking down the street\"."}], "task": "refering", "answer_template": "The \"an elephant trainer standing beside an elephant walking down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 199, 200, 222, 223, 245, 246, 268, 269, 291, 292, 293, 314, 315, 316, 338, 339], "bbox": [0.6656875, 0.5066042154566744, 0.7738750000000001, 0.9885948477751756], "iscrowd": 0, "area": 6133.440300000003, "rle": {"size": [427, 640], "counts": "Qma56Q=9F;E;G9I7H9G8]EQN\\9W2YFRNb9U2RFUNh9l2J6ROl0M3K5N2O0O2N2N2N2N1O1OO1N2O1O1N2O2M2ZLkGU2V8hMQH]2f7aMaH_2[7]MlHa2R7\\MUIVO2T2g6cNlJU1R5iNSKV1l4fNYKZ1e4dN_K\\1Z80O0000O2N1O1O1O1O1O1O1O1O1O1N2N2N2M3N2N3M2M3N3M2N3L5L5K4L5JmQl1"}, "label": "an elephant trainer standing beside an elephant walking down the street"}]}
{"id": 184223, "image": "COCO_train2014_000000184223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult zebra lying on its side on the ground\"."}], "task": "refering", "answer_template": "The \"an adult zebra lying on its side on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.169171875, 0.4352594339622642, 0.885390625, 0.7322405660377359], "iscrowd": 0, "area": 34466.350900000005, "rle": {"size": [424, 640], "counts": "mo\\11W=1N100O101N100O1O1O1O2M2O1O00000000001O00000O100000O10000000000010O0001O00010O000000EEkC;S<LgC4X<>N2O4L8G8I7J5K2N1O2N2O0O2N1O2N2O0O2N2N1O2O1N1O2N2N101N2O0O2O002M2O1N3N000O2O0O10000O101N10000O100O1000O0O2O11N100O10000O100O2O000000001N01000000O100000000O100O1O001O1O1O1O1O100O10000O10000O010001O00001O001O001O001O00001O001O001O00000000000000001O000000000000000000000000000000000000000000000000001O000000000000000001O00000O2O0000001O0000001O0000001O0000001O0000001O0000001O00001O00000O2O0O100O2N100O2O0O1O101O0O101O0O10UFXMW9f2fF^MZ9b2aFcM_9]2]FgMd9W2\\FjMd9V2[FkMe9U2YFmMh9Q2XFPNh9P2WFQNi9o1VFRNk9l1UFUNk9k1TFVNl9c21O00001N10001O0O10001N10001N10001N10000O2O00001N10001O0O101O0O10001O01O01O001O001O1O001O001O1O001O001O1O001O001O1O001O1O1O2N1O101N1O1O1O1O001O001N2O0O2O1O001N2O001O1O0O1010O01O00001O001O00001O010O00001O001O00001O010O00001N1000001O00001O0000001O00O1000001O0O100000000O10001O000000000000001O00000O1O1O1O1O1O100O100O1001O1O2N1O2N1N5L4L5K4L3M00000000000000000000000000001O0000000F_CLb<2aCM_<2dCL\\<2gCMk<Momm0"}, "label": "an adult zebra lying on its side on the ground"}]}
{"id": 184223, "image": "COCO_train2014_000000184223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra laying on the dirt\"."}], "task": "refering", "answer_template": "The \"a zebra laying on the dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.169171875, 0.4352594339622642, 0.885390625, 0.7322405660377359], "iscrowd": 0, "area": 34466.350900000005, "rle": {"size": [424, 640], "counts": "mo\\11W=1N100O101N100O1O1O1O2M2O1O00000000001O00000O100000O10000000000010O0001O00010O000000EEkC;S<LgC4X<>N2O4L8G8I7J5K2N1O2N2O0O2N1O2N2O0O2N2N1O2O1N1O2N2N101N2O0O2O002M2O1N3N000O2O0O10000O101N10000O100O1000O0O2O11N100O10000O100O2O000000001N01000000O100000000O100O1O001O1O1O1O1O100O10000O10000O010001O00001O001O001O001O00001O001O001O00000000000000001O000000000000000000000000000000000000000000000000001O000000000000000001O00000O2O0000001O0000001O0000001O0000001O0000001O0000001O00001O00000O2O0O100O2N100O2O0O1O101O0O101O0O10UFXMW9f2fF^MZ9b2aFcM_9]2]FgMd9W2\\FjMd9V2[FkMe9U2YFmMh9Q2XFPNh9P2WFQNi9o1VFRNk9l1UFUNk9k1TFVNl9c21O00001N10001O0O10001N10001N10001N10000O2O00001N10001O0O101O0O10001O01O01O001O001O1O001O001O1O001O001O1O001O001O1O001O1O1O2N1O101N1O1O1O1O001O001N2O0O2O1O001N2O001O1O0O1010O01O00001O001O00001O010O00001O001O00001O010O00001N1000001O00001O0000001O00O1000001O0O100000000O10001O000000000000001O00000O1O1O1O1O1O100O100O1001O1O2N1O2N1N5L4L5K4L3M00000000000000000000000000001O0000000F_CLb<2aCM_<2dCL\\<2gCMk<Momm0"}, "label": "a zebra laying on the dirt"}]}
{"id": 397217, "image": "COCO_train2014_000000397217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit in front of a cake\"."}], "task": "refering", "answer_template": "The \"a man in a suit in front of a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 315, 316, 317, 318, 319, 321, 322, 323, 324, 325, 326, 327, 328, 329, 336, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 358, 359, 360, 361, 369, 370, 371], "bbox": [0.27558823529411763, 0.10145424836601308, 0.9946895424836601, 0.765408496732026], "iscrowd": 0, "area": 83023.4038, "rle": {"size": [612, 612], "counts": "m]U32Rc01N2N3M2O1N2N2N2O1N2N3M2O1N2N2N2O2M2N3MM4L3L4N210000O100O100O10000O100O10001N100O100O1S^OZO\\a0f0c^O[O]a0e0b^O\\O]a0U110O0100O0100O010O01000O010O01000O001O0O2O1O001O001N101O1O001O0O2O001O00001N10001O001O0O101O00001O0O2O00001O001N10001O00001N101O00001O0O2O00001O000O2O001O00001O0O2O00001O000O2O001O00001N101O00001O0O101O000000000O10000000000O1000000O1N2M3N2M4M2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2M3N2001O1O00100O001O001O1O001O1O001O1O001O0013M5J6K5K5J7J5K5J6K5J6K2N0O2O000O100O10001N10000O10000O2O0O10000[MYIlHh6Q7_IkHb6Q7eIkH[6R7lIiHU6S7SJiHm5T7ZJhHf5U7aJfH`5W7gJeHY5W7oJdHS5X7TKdHl4Z7ZKbHf4]7_K^Hb4b7bKZH^4e7gKWHY4i7kKRHV4m7oKoGR4P8RLkGo3T8VLhGj3X8ZLdGf3[8b300O101N100O100O101N100O100O100O2O0O100O100O101N100O100000001O00000000000000001O000000000000001O000000000000000O2O000000000000001O000cHeDQ7[;lHjDR7V;lHnDR7R;kHSES7a;N2N22N2O1N2N2N2N2M2M4L4L4L4WF`Hf7d7^GUIa8g80001O0000001O00001O00001O00000O2N1O101N1O1O2N1O1O1O2N1O1O2O0O1O2N1O1O1O2N1O1O2N100O2N100O101O00001O00001N1000001O00001O000O2O0000001O00001O0O101O0000001O00001N10001O00006Jc0]Ob0^Oc0\\Oc0^Oc0]Oc0]Ob0^Oc0]Oc0]Ob0^Oc0]Oc0\\OQ\\1"}, "label": "a man in a suit in front of a cake"}]}
{"id": 397217, "image": "COCO_train2014_000000397217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit cutting a cake\"."}], "task": "refering", "answer_template": "The \"a man in a suit cutting a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 315, 316, 317, 318, 319, 321, 322, 323, 324, 325, 326, 327, 328, 329, 336, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 358, 359, 360, 361, 369, 370, 371], "bbox": [0.27558823529411763, 0.10145424836601308, 0.9946895424836601, 0.765408496732026], "iscrowd": 0, "area": 83023.4038, "rle": {"size": [612, 612], "counts": "m]U32Rc01N2N3M2O1N2N2N2O1N2N3M2O1N2N2N2O2M2N3MM4L3L4N210000O100O100O10000O100O10001N100O100O1S^OZO\\a0f0c^O[O]a0e0b^O\\O]a0U110O0100O0100O010O01000O010O01000O001O0O2O1O001O001N101O1O001O0O2O001O00001N10001O001O0O101O00001O0O2O00001O001N10001O00001N101O00001O0O2O00001O000O2O001O00001O0O2O00001O000O2O001O00001N101O00001O0O101O000000000O10000000000O1000000O1N2M3N2M4M2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2M3N2001O1O00100O001O001O1O001O1O001O1O001O0013M5J6K5K5J7J5K5J6K5J6K2N0O2O000O100O10001N10000O10000O2O0O10000[MYIlHh6Q7_IkHb6Q7eIkH[6R7lIiHU6S7SJiHm5T7ZJhHf5U7aJfH`5W7gJeHY5W7oJdHS5X7TKdHl4Z7ZKbHf4]7_K^Hb4b7bKZH^4e7gKWHY4i7kKRHV4m7oKoGR4P8RLkGo3T8VLhGj3X8ZLdGf3[8b300O101N100O100O101N100O100O100O2O0O100O100O101N100O100000001O00000000000000001O000000000000001O000000000000000O2O000000000000001O000cHeDQ7[;lHjDR7V;lHnDR7R;kHSES7a;N2N22N2O1N2N2N2N2M2M4L4L4L4WF`Hf7d7^GUIa8g80001O0000001O00001O00001O00000O2N1O101N1O1O2N1O1O1O2N1O1O2O0O1O2N1O1O1O2N1O1O2N100O2N100O101O00001O00001N1000001O00001O000O2O0000001O00001O0O101O0000001O00001N10001O00006Jc0]Ob0^Oc0\\Oc0^Oc0]Oc0]Ob0^Oc0]Oc0]Ob0^Oc0]Oc0\\OQ\\1"}, "label": "a man in a suit cutting a cake"}]}
{"id": 397217, "image": "COCO_train2014_000000397217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing an off yellow blouse\"."}], "task": "refering", "answer_template": "The \"a woman wearing an off yellow blouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 111, 112, 113, 114, 115, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 314, 315], "bbox": [0.030441176470588235, 0.1865686274509804, 0.3606535947712418, 0.6479248366013073], "iscrowd": 0, "area": 37523.64199999999, "rle": {"size": [612, 612], "counts": "Sc;m3V?1O2N1O1O1O1O1O100O2N1O1O1O1O1O1O101N1O1O1O1O1[NnKWDS4f;g1M4L3L4M3M3M3M3M3M4L3M3M3M3O1O1O1O2N1O1O10[Od0F:3M4L3M3M3M4M2N2M3N3M2N0000001O00000000001O00000O101O0000000100O001O1O00100O001O1O010O1O001O10O01O1O001O10O01O1O0010O01O1O00100O001O1O3M2O2M2N3M3M2O2M2N3M2N3N1N3M2N3M3M2O2M2N3M2N3N1N3M2N3M3N1N3M2N3M2N3N1N2N1O1O10O01O1O1O010O1O1O001O100O002N7I7J7H7I7I8H7J6I6J3M4L3M4M2M4L3M3M4M2M4L3M4L3N3L3M3M4L3M4L3M4L3M4K4MbYY7"}, "label": "a woman wearing an off yellow blouse"}]}
{"id": 397217, "image": "COCO_train2014_000000397217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the left\"."}], "task": "refering", "answer_template": "The \"the woman on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 111, 112, 113, 114, 115, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 314, 315], "bbox": [0.030441176470588235, 0.1865686274509804, 0.3606535947712418, 0.6479248366013073], "iscrowd": 0, "area": 37523.64199999999, "rle": {"size": [612, 612], "counts": "Sc;m3V?1O2N1O1O1O1O1O100O2N1O1O1O1O1O1O101N1O1O1O1O1[NnKWDS4f;g1M4L3L4M3M3M3M3M3M4L3M3M3M3O1O1O1O2N1O1O10[Od0F:3M4L3M3M3M4M2N2M3N3M2N0000001O00000000001O00000O101O0000000100O001O1O00100O001O1O010O1O001O10O01O1O001O10O01O1O0010O01O1O00100O001O1O3M2O2M2N3M3M2O2M2N3M2N3N1N3M2N3M3M2O2M2N3M2N3N1N3M2N3M3N1N3M2N3M2N3N1N2N1O1O10O01O1O1O010O1O1O001O100O002N7I7J7H7I7I8H7J6I6J3M4L3M4M2M4L3M3M4M2M4L3M4L3N3L3M3M4L3M4L3M4L3M4K4MbYY7"}, "label": "the woman on the left"}]}
{"id": 151466, "image": "COCO_train2014_000000151466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"serving dish for fruit\"."}], "task": "refering", "answer_template": "The \"serving dish for fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 130, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204], "bbox": [0.59540625, 0.27039583333333334, 0.9779843749999999, 0.5055208333333333], "iscrowd": 0, "area": 10207.293499999994, "rle": {"size": [480, 640], "counts": "agb54h>8I6J7L1N2O1O1O0O2O1O1N2O1O1O1N2O001N2O1O1OkNdBg0l=O10O2O000O1N2O0O2O00O010O10O010O0100O010O010O10O10O10O010O0100O010O01O001O0010O01O001O1O010O001O001O010O001O0010O01O001O2N100O1O1O1O1O1O100O1O1O1O00100O1O1O00000010O000000000010O0000000001O01O00000000010O00O1000001O000000000000000O100000000000000000000O1000000000000O1000000O1000000O100000000O10000O1O1O100O1O1O1O100O1O1O1O100O1O001O100O1O1O1O100O1O1O1O1O100O1O1O1O10O01O1O1O1O100O1O1O1O10O001O0O2O1O001O001J5O20O01O010O0101N3N1O2M2O2M3M2L5K4Lc\\6"}, "label": "serving dish for fruit"}]}
{"id": 151466, "image": "COCO_train2014_000000151466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl holding the fruit\"."}], "task": "refering", "answer_template": "The \"the bowl holding the fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 130, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204], "bbox": [0.59540625, 0.27039583333333334, 0.9779843749999999, 0.5055208333333333], "iscrowd": 0, "area": 10207.293499999994, "rle": {"size": [480, 640], "counts": "agb54h>8I6J7L1N2O1O1O0O2O1O1N2O1O1O1N2O001N2O1O1OkNdBg0l=O10O2O000O1N2O0O2O00O010O10O010O0100O010O010O10O10O10O010O0100O010O01O001O0010O01O001O1O010O001O001O010O001O0010O01O001O2N100O1O1O1O1O1O100O1O1O1O00100O1O1O00000010O000000000010O0000000001O01O00000000010O00O1000001O000000000000000O100000000000000000000O1000000000000O1000000O1000000O100000000O10000O1O1O100O1O1O1O100O1O1O1O100O1O001O100O1O1O1O100O1O1O1O1O100O1O1O1O10O01O1O1O1O100O1O1O1O10O001O0O2O1O001O001J5O20O01O010O0101N3N1O2M2O2M3M2L5K4Lc\\6"}, "label": "the bowl holding the fruit"}]}
{"id": 151466, "image": "COCO_train2014_000000151466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark blue cup\"."}], "task": "refering", "answer_template": "The \"the dark blue cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 55, 56, 57, 58, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 127], "bbox": [0.38932812499999997, 0.004125, 0.57303125, 0.30972916666666667], "iscrowd": 0, "area": 14538.680650000002, "rle": {"size": [480, 640], "counts": "Tgd33m>;E;E;E:F;E;E;E:F;E5oDiLo9X3mElLR:V3iEnLV:T3eEPMZ:Q3bERM^:P3]ETMb:f3O1O001O1O1O001O1O000000O100000000000000000O1000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000O100000000000M3I7I7I7M3L4M3I7I7H8H8H8H8H8H8I7H8H8H8H8HVoo3"}, "label": "the dark blue cup"}]}
{"id": 151466, "image": "COCO_train2014_000000151466.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue colored glass\"."}], "task": "refering", "answer_template": "The \"a blue colored glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 55, 56, 57, 58, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 127], "bbox": [0.38932812499999997, 0.004125, 0.57303125, 0.30972916666666667], "iscrowd": 0, "area": 14538.680650000002, "rle": {"size": [480, 640], "counts": "Tgd33m>;E;E;E:F;E;E;E:F;E5oDiLo9X3mElLR:V3iEnLV:T3eEPMZ:Q3bERM^:P3]ETMb:f3O1O001O1O1O001O1O000000O100000000000000000O1000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000O100000000000M3I7I7I7M3L4M3I7I7H8H8H8H8H8H8I7H8H8H8H8HVoo3"}, "label": "a blue colored glass"}]}
{"id": 454570, "image": "COCO_train2014_000000454570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2nd to theleft brown horse drinking\"."}], "task": "refering", "answer_template": "The \"2nd to theleft brown horse drinking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 144, 145, 146, 147, 148, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 235, 236, 237], "bbox": [0.23675000000000002, 0.35283372365339577, 0.45053125000000005, 0.750679156908665], "iscrowd": 0, "area": 11032.559799999999, "rle": {"size": [427, 640], "counts": "kco11[=0O1O2N1O1[D5h9LaEk0^:VOYES1f:nNWEU1h:lNTEX1k:iNREZ1f:nNVEU1c:n0G9G9I7L4L2N0000000O10002N1O2N101O00001N10000000O2O2N2L4L4003M2N2N4L=Cg0YO000002N1O1O2N1O1O2^O^GWLc8b3lGULV8b3n0G9XOh0M2N2N2N2000000000O1000O10000000000000000000O011O1O2N1N3N1O2N1O1NO2N2N2N1O2N2I7I6J7K5N2O0O2O1N2O1N3M2O1N2O2N1N2O1O2M2O1O2M2O2N2M3N2O1N1O2N2N2N2N2N2Nbcb4"}, "label": "2nd to theleft brown horse drinking"}]}
{"id": 454570, "image": "COCO_train2014_000000454570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse with the light brown mane and the big splash of water in it ' s face\"."}], "task": "refering", "answer_template": "The \"the horse with the light brown mane and the big splash of water in it ' s face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 164, 165, 166, 187, 188, 189, 209, 210, 211, 212, 232, 233, 235], "bbox": [0.12354687499999999, 0.3990632318501171, 0.272296875, 0.7455269320843092], "iscrowd": 0, "area": 6504.61465, "rle": {"size": [427, 640], "counts": "cVQ14U=9oCDS:0YEd03Fd:HUEl0EEX;_OQEb1o:`NmDa1T;`NjDa1W;8101N101O0OO2N1O2O0O2N1O1O100O1O1O1O100O1O1O1L4K5L2M2O11]OQFlMR:T2b001O00001O0000001O00001O00001O00001O00001O001O03N3L3N3fFgM\\7[2^HhMb7\\2VHhMk7Z2nGiMS8[2fGgM\\8[2]GhMe8\\32O2N2N1`N`FTOc9l0_FPOb9R1_FjNd9U1]FfNg9[1ZF`Ni9`1YFZNk9g1VFSNn9m1TFnMP:Q2QFjMS:W2`0@XE]Nh:\\1h0H7J6J6L4L5K4L4L4L4LlTR6"}, "label": "the horse with the light brown mane and the big splash of water in it ' s face"}]}
{"id": 454570, "image": "COCO_train2014_000000454570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse on the far left side of image drinking water\"."}], "task": "refering", "answer_template": "The \"horse on the far left side of image drinking water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 164, 165, 166, 187, 188, 189, 209, 210, 211, 212, 232, 233, 235], "bbox": [0.12354687499999999, 0.3990632318501171, 0.272296875, 0.7455269320843092], "iscrowd": 0, "area": 6504.61465, "rle": {"size": [427, 640], "counts": "cVQ14U=9oCDS:0YEd03Fd:HUEl0EEX;_OQEb1o:`NmDa1T;`NjDa1W;8101N101O0OO2N1O2O0O2N1O1O100O1O1O1O100O1O1O1L4K5L2M2O11]OQFlMR:T2b001O00001O0000001O00001O00001O00001O00001O001O03N3L3N3fFgM\\7[2^HhMb7\\2VHhMk7Z2nGiMS8[2fGgM\\8[2]GhMe8\\32O2N2N1`N`FTOc9l0_FPOb9R1_FjNd9U1]FfNg9[1ZF`Ni9`1YFZNk9g1VFSNn9m1TFnMP:Q2QFjMS:W2`0@XE]Nh:\\1h0H7J6J6L4L5K4L4L4L4LlTR6"}, "label": "horse on the far left side of image drinking water"}]}
{"id": 454570, "image": "COCO_train2014_000000454570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second horse from the right\"."}], "task": "refering", "answer_template": "The \"the second horse from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264], "bbox": [0.361859375, 0.38569086651053863, 0.5929062500000001, 0.7788758782201405], "iscrowd": 0, "area": 14632.98664999999, "rle": {"size": [427, 640], "counts": "noP31[=0O2N1O1SC4Y<MaC;]<:O2VD_Of:b0oDIP;\\1L2M2N3M2N2M2M4K5J6K5M3N2L4L3O1O00L4O20O001O00100OO100O110O0SOcFnM^9P2gFlMY9T2kFfMX9Y2m0O1O100O001O100OXGjM^6W2aIjM^6Z2^IhM`6]2[IdMd6_2YIbMf6b2UI_Mk6Q3dHQM\\7[3SHjLl7e3aG`L_8V40O0N3O001N1VOnGYLU8d2SH]Mg8b2R1O1O1O100000000000001000000000O1O4L4Lf0ZO2N2M3N2N2N3L6K\\NcGcNT8]1oGcNXOoN^8_2\\HdMXOh0[8g1kHTNU7o1lHmMT7U2lHiMS7Y2nHeMR7]2mHbMS7`2hHSMjN113]8R3\\HoL]OMV8X3VHoLFEV8_3eGYMf8f30001O1OoM\\GTOLaNh8[2_G\\N4C]8Q2\\GdMLb0:M^8Y2ZGXM5MOj0b8Q2eGZN[8Q2WGRNi8X30O1N2O1O110M2N3L3K6MTGUL\\8a3lGbLT8\\3iGiLV8V40O2O001O1QOhGiLY8f2QG`Mk0FV8d2XG`Mk9]2c0N2M3J6DoDVNX;f1:L4K6J6CRe\\3"}, "label": "the second horse from the right"}]}
{"id": 454570, "image": "COCO_train2014_000000454570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse with some white on his neck and a white front leg drinks from a river\"."}], "task": "refering", "answer_template": "The \"horse with some white on his neck and a white front leg drinks from a river\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264], "bbox": [0.361859375, 0.38569086651053863, 0.5929062500000001, 0.7788758782201405], "iscrowd": 0, "area": 14632.98664999999, "rle": {"size": [427, 640], "counts": "noP31[=0O2N1O1SC4Y<MaC;]<:O2VD_Of:b0oDIP;\\1L2M2N3M2N2M2M4K5J6K5M3N2L4L3O1O00L4O20O001O00100OO100O110O0SOcFnM^9P2gFlMY9T2kFfMX9Y2m0O1O100O001O100OXGjM^6W2aIjM^6Z2^IhM`6]2[IdMd6_2YIbMf6b2UI_Mk6Q3dHQM\\7[3SHjLl7e3aG`L_8V40O0N3O001N1VOnGYLU8d2SH]Mg8b2R1O1O1O100000000000001000000000O1O4L4Lf0ZO2N2M3N2N2N3L6K\\NcGcNT8]1oGcNXOoN^8_2\\HdMXOh0[8g1kHTNU7o1lHmMT7U2lHiMS7Y2nHeMR7]2mHbMS7`2hHSMjN113]8R3\\HoL]OMV8X3VHoLFEV8_3eGYMf8f30001O1OoM\\GTOLaNh8[2_G\\N4C]8Q2\\GdMLb0:M^8Y2ZGXM5MOj0b8Q2eGZN[8Q2WGRNi8X30O1N2O1O110M2N3L3K6MTGUL\\8a3lGbLT8\\3iGiLV8V40O2O001O1QOhGiLY8f2QG`Mk0FV8d2XG`Mk9]2c0N2M3J6DoDVNX;f1:L4K6J6CRe\\3"}, "label": "horse with some white on his neck and a white front leg drinks from a river"}]}
{"id": 569261, "image": "COCO_train2014_000000569261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra lying down in brown dirt\"."}], "task": "refering", "answer_template": "The \"a zebra lying down in brown dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 193, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 266, 267, 268, 273, 274, 275, 276, 277, 278, 279, 280, 282, 283, 291, 292], "bbox": [0.12126463700234193, 0.496625, 0.9296018735362999, 0.838203125], "iscrowd": 0, "area": 51916.0776, "rle": {"size": [640, 427], "counts": "nlP18gc09F;E;E:G:E4L4M3L4L4L3N3L4L4L3N3L4L3N1N3M3M3N2M3M3M3N2O1O1O1O1O1O1O1O1N2O1O001O1O1O1O1O1O1O1OM30O1000000000000000000000000O2O01O0000001O0001O010O001O00010O001O010O001O010O00001O010O001O010O000010O00000000001O0000000000001O00000000001O00O1M3M4M2M4L3M3N3N10001O0O101O000O2O001O0O101O000O2O00001N101O000O10O1O1O00100O1O001O1O10O01O1O1O000100O100O100O2N100O100O100O1O100O100O1O2O0O100O2O0O101N101O0O101O000O2O000O10000O[O_BmJa=Q5aBoJ_=n4dBRK[=m4gBSKY=j4jBVKV=h4lBXKT=e4PCZKo<f4RCZKn<d4TC\\Kl<b4VC^Ki<a4YC_Kg<`4ZC`Kf<^4]CaKb<^4`CbK`<\\4bCdK^<[4cCeK]<Y4eCfK[<Y4gCgKY<W4jChKV<W4lChKS<W4oCiKQ<U4RDjKn;T4TDkKl;U4VDjKj;T4YDkKg;S4[DmKe;Q4^DnKa;R4aDmK_;Q4cDnK^;Q4Y2010O01O010O001O010O0010O01O010O1O8H2OO0O1O2N100O2N2O1N2N2N2O1N2N2N2O100O2N1O1O1O1O2N1O1O1O1O2N1O101N2N2N2N2N2N2O1O0O2O1O000O10O100000O0100000O1000O10O1000O10O12M2N2N2N2M4M3M2N3M3M3M3L4L3QLZA_2]`0F9G=C:F8G8I4L3M3M2N3M2N3MfXb0"}, "label": "a zebra lying down in brown dirt"}]}
{"id": 569261, "image": "COCO_train2014_000000569261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the foreground that is lying down\"."}], "task": "refering", "answer_template": "The \"the zebra in the foreground that is lying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 193, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 266, 267, 268, 273, 274, 275, 276, 277, 278, 279, 280, 282, 283, 291, 292], "bbox": [0.12126463700234193, 0.496625, 0.9296018735362999, 0.838203125], "iscrowd": 0, "area": 51916.0776, "rle": {"size": [640, 427], "counts": "nlP18gc09F;E;E:G:E4L4M3L4L4L3N3L4L4L3N3L4L3N1N3M3M3N2M3M3M3N2O1O1O1O1O1O1O1O1N2O1O001O1O1O1O1O1O1O1OM30O1000000000000000000000000O2O01O0000001O0001O010O001O00010O001O010O001O010O00001O010O001O010O000010O00000000001O0000000000001O00000000001O00O1M3M4M2M4L3M3N3N10001O0O101O000O2O001O0O101O000O2O00001N101O000O10O1O1O00100O1O001O1O10O01O1O1O000100O100O100O2N100O100O100O1O100O100O1O2O0O100O2O0O101N101O0O101O000O2O000O10000O[O_BmJa=Q5aBoJ_=n4dBRK[=m4gBSKY=j4jBVKV=h4lBXKT=e4PCZKo<f4RCZKn<d4TC\\Kl<b4VC^Ki<a4YC_Kg<`4ZC`Kf<^4]CaKb<^4`CbK`<\\4bCdK^<[4cCeK]<Y4eCfK[<Y4gCgKY<W4jChKV<W4lChKS<W4oCiKQ<U4RDjKn;T4TDkKl;U4VDjKj;T4YDkKg;S4[DmKe;Q4^DnKa;R4aDmK_;Q4cDnK^;Q4Y2010O01O010O001O010O0010O01O010O1O8H2OO0O1O2N100O2N2O1N2N2N2O1N2N2N2O100O2N1O1O1O1O2N1O1O1O1O2N1O101N2N2N2N2N2N2O1O0O2O1O000O10O100000O0100000O1000O10O1000O10O12M2N2N2N2M4M3M2N3M3M3M3L4L3QLZA_2]`0F9G=C:F8G8I4L3M3M2N3M2N3MfXb0"}, "label": "the zebra in the foreground that is lying down"}]}
{"id": 569261, "image": "COCO_train2014_000000569261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing in the zoo\"."}], "task": "refering", "answer_template": "The \"a zebra standing in the zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 135, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 156, 157, 167, 168, 182, 183, 197], "bbox": [0.0012412177985948478, 0.194484375, 0.5274941451990632, 0.615], "iscrowd": 0, "area": 33977.7124, "rle": {"size": [640, 427], "counts": "Xh0P4P`0001O01O00000001O01O00000001O01O00000000010O00000000010O00000000010O0000000000O2O000O1000001N103M5J6K4L5J6K5K5K4K6K5K9Fc0^Ob0^Oh0WO01N2N2N101N2N101N2N1O2N2I7H7iMkC^L\\<b3Q20O1000O02O8H8H8G9H8H8G9H8H8G8I00000O2O0000000001O00000M3G9F:F:F:F:F:G9F:F:G9O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O12N1O2N1O1O2N1O2N1O1O2N101N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O10100O3M4L3M4L3MN2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O2M2O1N101_L]@b2e?]M^@f0O6c?TOa@`073Z?[Ob@>?No>De@9f0Jg>Lf@5n0F]>4g@1X1AR>=j@M^1^Oi=d0l@I[a07g^OE\\a0:g^OA\\a0=T101O1OkRn3"}, "label": "a zebra standing in the zoo"}]}
{"id": 569261, "image": "COCO_train2014_000000569261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that is standing\"."}], "task": "refering", "answer_template": "The \"a zebra that is standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 135, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 156, 157, 167, 168, 182, 183, 197], "bbox": [0.0012412177985948478, 0.194484375, 0.5274941451990632, 0.615], "iscrowd": 0, "area": 33977.7124, "rle": {"size": [640, 427], "counts": "Xh0P4P`0001O01O00000001O01O00000001O01O00000000010O00000000010O00000000010O0000000000O2O000O1000001N103M5J6K4L5J6K5K5K4K6K5K9Fc0^Ob0^Oh0WO01N2N2N101N2N101N2N1O2N2I7H7iMkC^L\\<b3Q20O1000O02O8H8H8G9H8H8G9H8H8G8I00000O2O0000000001O00000M3G9F:F:F:F:F:G9F:F:G9O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O12N1O2N1O1O2N1O2N1O1O2N101N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O10100O3M4L3M4L3MN2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O2M2O1N101_L]@b2e?]M^@f0O6c?TOa@`073Z?[Ob@>?No>De@9f0Jg>Lf@5n0F]>4g@1X1AR>=j@M^1^Oi=d0l@I[a07g^OE\\a0:g^OA\\a0=T101O1OkRn3"}, "label": "a zebra that is standing"}]}
{"id": 446383, "image": "COCO_train2014_000000446383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse that has its head on the back of another zebra\"."}], "task": "refering", "answer_template": "The \"horse that has its head on the back of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 284, 285, 306, 307, 308, 329, 330, 331], "bbox": [0.312125, 0.26728337236533956, 0.45682812500000003, 0.9959484777517564], "iscrowd": 0, "area": 9815.993049999997, "rle": {"size": [427, 640], "counts": "ccc2:k<9L3M3M1O1O2M2O1_IROd0o0YOTOf0n0mM>R2CnM=P2EoM;Q2FoM:P2HnM9Q2HnM9Q2HoM8o1KoM6P2KPN4P2MoM4P2MPN3MaNeMf1\\2JKgN^Mg1g2BKiNXMi1l2_OLkNTMg1P3^OLlNRMh1P3\\OOmNoLh1R3[OOnNmLh1S3[O1mNjLi1U3ZO2mNgLk1U3YO5lNdLl1W3XO5mNbLl1X3XO7lN_Ll1[3VO9nNZLl1^3gNh0]OfKn1c3VNW1MPKP2j3fMd1R3]NjLh1T3YNgLm1W3U4N2N1N200O1OPJ`MfLB`3m2JcN5]1LdN3[1MfN3Z1MfN3Y1MiN2V1OjN0W10iN0V10lNNT13lNMT12nNLR15nNIT17lNHU17lNGU1:kNEV1;jNDW1;jNDV1=jNAX1>iNBW1>gNCY1>eNC\\1<cNE^1;`NF`1:`NFa1:YNMe14QN6o1IkM?T2AiMb0V2^OhMf0W2ZOfMk0X2V5M3N2N2M4M2M6Jb0^O8H6J4L5K5K5K5K5K5Kdm`4"}, "label": "horse that has its head on the back of another zebra"}]}
{"id": 446383, "image": "COCO_train2014_000000446383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra that is laying its head on another zebra\"."}], "task": "refering", "answer_template": "The \"the zebra that is laying its head on another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 284, 285, 306, 307, 308, 329, 330, 331], "bbox": [0.312125, 0.26728337236533956, 0.45682812500000003, 0.9959484777517564], "iscrowd": 0, "area": 9815.993049999997, "rle": {"size": [427, 640], "counts": "ccc2:k<9L3M3M1O1O2M2O1_IROd0o0YOTOf0n0mM>R2CnM=P2EoM;Q2FoM:P2HnM9Q2HnM9Q2HoM8o1KoM6P2KPN4P2MoM4P2MPN3MaNeMf1\\2JKgN^Mg1g2BKiNXMi1l2_OLkNTMg1P3^OLlNRMh1P3\\OOmNoLh1R3[OOnNmLh1S3[O1mNjLi1U3ZO2mNgLk1U3YO5lNdLl1W3XO5mNbLl1X3XO7lN_Ll1[3VO9nNZLl1^3gNh0]OfKn1c3VNW1MPKP2j3fMd1R3]NjLh1T3YNgLm1W3U4N2N1N200O1OPJ`MfLB`3m2JcN5]1LdN3[1MfN3Z1MfN3Y1MiN2V1OjN0W10iN0V10lNNT13lNMT12nNLR15nNIT17lNHU17lNGU1:kNEV1;jNDW1;jNDV1=jNAX1>iNBW1>gNCY1>eNC\\1<cNE^1;`NF`1:`NFa1:YNMe14QN6o1IkM?T2AiMb0V2^OhMf0W2ZOfMk0X2V5M3N2N2M4M2M6Jb0^O8H6J4L5K5K5K5K5K5Kdm`4"}, "label": "the zebra that is laying its head on another zebra"}]}
{"id": 446383, "image": "COCO_train2014_000000446383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra which is standing from the right side\"."}], "task": "refering", "answer_template": "The \"a zebra which is standing from the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 129, 130, 131, 132, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 279, 280, 281, 287, 288, 289, 290, 291, 302, 303, 304, 311, 312, 313, 314, 325, 326, 327, 335, 336], "bbox": [0.14809375, 0.27878220140515225, 0.74646875, 1.0], "iscrowd": 0, "area": 25755.373400000004, "rle": {"size": [427, 640], "counts": "^oW1j0`<8H8H9G1O1O1O1O1O100O001O1O1O1O1O1O1O00L4I7J6I7M4N1O2M3N2N3M2MkD\\Oi9b0SFDn99nEMR:1jE6U:GhE?X:_OeEf0\\:WOdEl0[:ROeEP1[:mNgET1V;3M2N2N3]ObNUE`1k:bNQE`1o:bNnD^1T;cNhD\\1[;83M2N2M4M2N3L3N2O2M2O2M2O1O2M2O2M2O1N3N1O2M2O1N3N1N2O1O1N2OlSn11SlQN1N2N2N1O2N2N2N2N2O001O1N2O1O1O1O001O1O1O1N2O001O1O1O1O1O1O0O2O1O1O1M3F9eDeN_:]2B?D;G:G9G8G:G9F9H4K4M3L4N2N2M3N2N2N2M3N2N2M3N2N2LcNaIVLZ6k3PJoKk5R4`JgK[5[4nJ_Kh4j4b1:4MRJ[JU4f5fK_JY4a5cKdJ\\4m5oJYJo4R7O1O2N1N2O1O2N1cNQH]MP8c2UHWMm7h2YHQMh7P3\\HjLe7V3aHcL`7]3fH\\L\\7c3iHVLY7j3T101OI[LhFf3T9:N2N2M3M3J7H7I8H9G8H9G8H9J5L5L3M4K4M3M3M3L2O00001N10001N101O001N6K5K4L5J5K6J3L3M3N2M3M2O2M3L4K5L3L5L4K5L4K4M4K5L4K5L3L8H9G:F9G:C<Bg]S2"}, "label": "a zebra which is standing from the right side"}]}
{"id": 446383, "image": "COCO_train2014_000000446383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing front of two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra standing front of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 126, 127, 128, 129, 148, 149, 150, 151, 152, 164, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 306, 308, 309, 310, 322, 323, 324, 331, 332, 333], "bbox": [0.029640624999999997, 0.22988290398126462, 0.613859375, 1.0], "iscrowd": 0, "area": 59823.0023, "rle": {"size": [427, 640], "counts": "kZ81V91kJ0o4>gJDS5j0cJXOZ5S1]JnNa5[1VJgNi5g1iIZNU6]2TIeMj6c2nH_MP7j2dH[M[7l2WH\\Mg7i2PH]Mn7i2hG]MV8m3N2O1N2O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O100O100O1O10O0100OO101O001O00001N101O00001O1N2O0O2O1N2O0O2O1N2O0O2O1O1O001N2O1O001O1N3N1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O2N1N2O1O1O1O10000001O0000001O0001O01O0000001O0000001O0000001O000O101O00000O2O0000001O0001O0001O00000000010O000000000010O00000001O0001O000001O00000001O01O0000000010O1O1O001O100O1O1O0001O01O0O100O10000O2O001N100O2O000O10000O100O10O10O10O1O001O010O1O001O10O010O0100O010O10O010O0100O010O10O010O0100O010O10O01O100PIPLd4Q4SKXLm4h3kJ`LT5b3cJfL]5Z3ZJoLe5R3SJVMm5j2kI^MT6c2cIfM\\6\\2[IlMe6]4O100O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1O1O1O1O100O1O1O1O10O0H8I7K5M4KRNbIXM[6j2hIUMT6n2nIPMo5T3RJkLj5Y3WJeLe5`3\\J^La5f3`JYL\\5l3eJQLX5S4iJlKU5V4lJhKS5Z4nJeKo4_4QK_Kn4c4SK[Km4f4TKYKk4h4WKjJYOB_5g5WKeJU5\\5lJbJT5`5kJ^JV5d5jJYJW5i5hJVJW5l5jJQJW5Q6hJnIX5T6n01N3N2N2N1O2N2N2N2N2N2NO2N1A?00000O1N3M2XOdIaJ^6]5gI]J\\6`5jIZJY6c5i0N2M3M3N2N2M3N3L3N2M3N2N3L4M4K4M4K4L5K4L5K7I7I7H9H9G8F:F;E:Fk0UO`jV3"}, "label": "a zebra standing front of two other zebras"}]}
{"id": 446383, "image": "COCO_train2014_000000446383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra whose head is in the middle of group\"."}], "task": "refering", "answer_template": "The \"zebra whose head is in the middle of group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 126, 127, 128, 129, 148, 149, 150, 151, 152, 164, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 306, 308, 309, 310, 322, 323, 324, 331, 332, 333], "bbox": [0.029640624999999997, 0.22988290398126462, 0.613859375, 1.0], "iscrowd": 0, "area": 59823.0023, "rle": {"size": [427, 640], "counts": "kZ81V91kJ0o4>gJDS5j0cJXOZ5S1]JnNa5[1VJgNi5g1iIZNU6]2TIeMj6c2nH_MP7j2dH[M[7l2WH\\Mg7i2PH]Mn7i2hG]MV8m3N2O1N2O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O100O100O1O10O0100OO101O001O00001N101O00001O1N2O0O2O1N2O0O2O1N2O0O2O1O1O001N2O1O001O1N3N1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O2N1N2O1O1O1O10000001O0000001O0001O01O0000001O0000001O0000001O000O101O00000O2O0000001O0001O0001O00000000010O000000000010O00000001O0001O000001O00000001O01O0000000010O1O1O001O100O1O1O0001O01O0O100O10000O2O001N100O2O000O10000O100O10O10O10O1O001O010O1O001O10O010O0100O010O10O010O0100O010O10O010O0100O010O10O01O100PIPLd4Q4SKXLm4h3kJ`LT5b3cJfL]5Z3ZJoLe5R3SJVMm5j2kI^MT6c2cIfM\\6\\2[IlMe6]4O100O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1O1O1O1O100O1O1O1O10O0H8I7K5M4KRNbIXM[6j2hIUMT6n2nIPMo5T3RJkLj5Y3WJeLe5`3\\J^La5f3`JYL\\5l3eJQLX5S4iJlKU5V4lJhKS5Z4nJeKo4_4QK_Kn4c4SK[Km4f4TKYKk4h4WKjJYOB_5g5WKeJU5\\5lJbJT5`5kJ^JV5d5jJYJW5i5hJVJW5l5jJQJW5Q6hJnIX5T6n01N3N2N2N1O2N2N2N2N2N2NO2N1A?00000O1N3M2XOdIaJ^6]5gI]J\\6`5jIZJY6c5i0N2M3M3N2N2M3N3L3N2M3N2N3L4M4K4M4K4L5K4L5K7I7I7H9H9G8F:F;E:Fk0UO`jV3"}, "label": "zebra whose head is in the middle of group"}]}
{"id": 298931, "image": "COCO_train2014_000000298931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy with blonde hair brushing his teeth\"."}], "task": "refering", "answer_template": "The \"a young boy with blonde hair brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 116, 128, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.0, 0.09213213213213213, 0.46994, 0.9954954954954955], "iscrowd": 0, "area": 53090.73975, "rle": {"size": [333, 500], "counts": "X33V:7J7H7J6I7J7J5L4L5J5L3M3M2N2N2N3M2N2N2N3L3N2N2N2N3M2N2mMnLhLS3S2PMkL0Q1R3n1UNQNm1l1UNRNm1m1TNSNm1k1TNTNm1l1SNSNo1k1RNUNo1i1RNVNo1j1QNUNQ2i1PNWNQ2g1PNXNQ2h1oMWNS2g1nMYNR2f1oMYNS2f1mMYNU2e1lM[NT2e1lMZNV2d1kM[NV2d1kM\\NV2c1jM\\NX2b1iM^NW2a1jM^NX2_1jMaNX2\\1iMcNY2Z1iMfNX2X1iMhNX2U1jMjNY2R1iMnNX2P1iMoNY2n0iMROX2l0iMSOZ2i0hMWOY2g0hMYOY2d0iM[OY2c0hM]OY2`0iM_OZ2=hMCY2:iMFX26kMIW24kMLW20kMOW2NkM2V2KlM4V2IlM7W2DkM<X2_OjMa0Z2YOhMg0[2UOfMk0]2POeMP1l50000000000000000000000000000O100000000000000000000000000O100000000000000000000000000001O00001O001O00001O001O00001O001O00001O001O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O2N2N2N3M2N2N2N2N3M1O001O1O1O1O1O1O100O1O1O001O001O001O1O001O001O1O1O1O1O1O1O1RJ`Lm4a3RKaLl4`3RKcLl4_3RKbLk4b3SK`Li0H\\2o3hLZLj0IX2T4lLULi0JT2Y4oLoKk0JQ2^4QMiKl0Km1e4SMaKn0Mi1j4UMZKP1Ne1c5YN_Jb1f5\\N\\J^1j5[N^J_1[7J6K6I6J6\\O]oe2"}, "label": "a young boy with blonde hair brushing his teeth"}]}
{"id": 298931, "image": "COCO_train2014_000000298931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little blond boy\"."}], "task": "refering", "answer_template": "The \"little blond boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 116, 128, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.0, 0.09213213213213213, 0.46994, 0.9954954954954955], "iscrowd": 0, "area": 53090.73975, "rle": {"size": [333, 500], "counts": "X33V:7J7H7J6I7J7J5L4L5J5L3M3M2N2N2N3M2N2N2N3L3N2N2N2N3M2N2mMnLhLS3S2PMkL0Q1R3n1UNQNm1l1UNRNm1m1TNSNm1k1TNTNm1l1SNSNo1k1RNUNo1i1RNVNo1j1QNUNQ2i1PNWNQ2g1PNXNQ2h1oMWNS2g1nMYNR2f1oMYNS2f1mMYNU2e1lM[NT2e1lMZNV2d1kM[NV2d1kM\\NV2c1jM\\NX2b1iM^NW2a1jM^NX2_1jMaNX2\\1iMcNY2Z1iMfNX2X1iMhNX2U1jMjNY2R1iMnNX2P1iMoNY2n0iMROX2l0iMSOZ2i0hMWOY2g0hMYOY2d0iM[OY2c0hM]OY2`0iM_OZ2=hMCY2:iMFX26kMIW24kMLW20kMOW2NkM2V2KlM4V2IlM7W2DkM<X2_OjMa0Z2YOhMg0[2UOfMk0]2POeMP1l50000000000000000000000000000O100000000000000000000000000O100000000000000000000000000001O00001O001O00001O001O00001O001O00001O001O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O2N2N2N3M2N2N2N2N3M1O001O1O1O1O1O1O100O1O1O001O001O001O1O001O001O1O1O1O1O1O1O1RJ`Lm4a3RKaLl4`3RKcLl4_3RKbLk4b3SK`Li0H\\2o3hLZLj0IX2T4lLULi0JT2Y4oLoKk0JQ2^4QMiKl0Km1e4SMaKn0Mi1j4UMZKP1Ne1c5YN_Jb1f5\\N\\J^1j5[N^J_1[7J6K6I6J6\\O]oe2"}, "label": "little blond boy"}]}
{"id": 298931, "image": "COCO_train2014_000000298931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mom brushing her baby ' s teeth\"."}], "task": "refering", "answer_template": "The \"a mom brushing her baby ' s teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.41656, 0.0, 1.0, 0.9865165165165165], "iscrowd": 0, "area": 77734.50274999999, "rle": {"size": [333, 500], "counts": "^eS24n9`0A`0A>K5K5K5K6J5K5K5K5K5K1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1oJXL[3i3dLYLZ3i3dLXL[3i3dLYLZ3h3fLYLX3h3gLZLW3h3SLQLI8S4h3SLRLI6S4i3RLSLJ6R4h3RLTLK5R4i3QLTLL3R4j3PLULM2R4j3PLULM2R4j3oKWLM0S4j3nKXLOOQ4j3oKXLOOQ4j3nKZLOMR4j3mK[L1KQ4k3mK[L1KQ4k3lK[L3KP4k3lKYL5Nm3d4TL\\Kk3d4UL]Kj3b4WL_Kh3a4XL`Kg3`4ZL`Ke3_4\\LcKb3]4^LdKa3\\4_LeK`3Z4bLhK[3X4eLkKX3U4hLnKU3Q4lLRLQ3n3PMTLm2l3SMWLj2h3XMZLe2f3\\M\\La2d3`M_L\\2b3dM`LY2`3hMbLU2^3lMdLQ2\\3PNfLm1Z3TNgLj1Y3WNgLh1Y3YNgLf1Z3ZNeLf1[3[NeLd1[3]NeLb1[3_NdLa1\\3`NdL_1\\3bNdL]1\\3dNcL\\1^3dNbL[1^3fNbLY1^3hNbLW1^3jNaLV1_3kNaLT1_3mNaLR1_3oN`LQ1a3nN`LQ1`3oNaLP1`3nNaLR1`3mNaLR1`3lNbLS1_3kNbLU1`3iNaLV1`3hNbLW1_3gNbLY1`3eNaLZ1`3dNbL[1_3dNaL\\1`3bNbL]1`3`NbL_1_3`NbL_1_3_NbLa1_3]NcLb1_3\\NbLc1_3[NcLd1^3ZNdLe1S6O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1HhFYOV9MiFg03\\OU9MhFc06@S9?PGAP9<SGDn88UGHl84WGLi81ZGOg8M[G4_90000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"}, "label": "a mom brushing her baby ' s teeth"}]}
{"id": 298931, "image": "COCO_train2014_000000298931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in white t - shirt\"."}], "task": "refering", "answer_template": "The \"the woman in white t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.41656, 0.0, 1.0, 0.9865165165165165], "iscrowd": 0, "area": 77734.50274999999, "rle": {"size": [333, 500], "counts": "^eS24n9`0A`0A>K5K5K5K6J5K5K5K5K5K1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1oJXL[3i3dLYLZ3i3dLXL[3i3dLYLZ3h3fLYLX3h3gLZLW3h3SLQLI8S4h3SLRLI6S4i3RLSLJ6R4h3RLTLK5R4i3QLTLL3R4j3PLULM2R4j3PLULM2R4j3oKWLM0S4j3nKXLOOQ4j3oKXLOOQ4j3nKZLOMR4j3mK[L1KQ4k3mK[L1KQ4k3lK[L3KP4k3lKYL5Nm3d4TL\\Kk3d4UL]Kj3b4WL_Kh3a4XL`Kg3`4ZL`Ke3_4\\LcKb3]4^LdKa3\\4_LeK`3Z4bLhK[3X4eLkKX3U4hLnKU3Q4lLRLQ3n3PMTLm2l3SMWLj2h3XMZLe2f3\\M\\La2d3`M_L\\2b3dM`LY2`3hMbLU2^3lMdLQ2\\3PNfLm1Z3TNgLj1Y3WNgLh1Y3YNgLf1Z3ZNeLf1[3[NeLd1[3]NeLb1[3_NdLa1\\3`NdL_1\\3bNdL]1\\3dNcL\\1^3dNbL[1^3fNbLY1^3hNbLW1^3jNaLV1_3kNaLT1_3mNaLR1_3oN`LQ1a3nN`LQ1`3oNaLP1`3nNaLR1`3mNaLR1`3lNbLS1_3kNbLU1`3iNaLV1`3hNbLW1_3gNbLY1`3eNaLZ1`3dNbL[1_3dNaL\\1`3bNbL]1`3`NbL_1_3`NbL_1_3_NbLa1_3]NcLb1_3\\NbLc1_3[NcLd1^3ZNdLe1S6O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1HhFYOV9MiFg03\\OU9MhFc06@S9?PGAP9<SGDn88UGHl84WGLi81ZGOg8M[G4_90000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"}, "label": "the woman in white t - shirt"}]}
{"id": 85939, "image": "COCO_train2014_000000085939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white pot\"."}], "task": "refering", "answer_template": "The \"a white pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 63, 64, 65, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251], "bbox": [0.596625, 0.07920833333333334, 0.9977499999999999, 0.6095416666666666], "iscrowd": 0, "area": 43962.88979999999, "rle": {"size": [480, 640], "counts": "ZWc5i0V>6I7I6K6I7I6J7J5J3M3M4M2M3O1N2N2N2N3M2N2N2N2N2O1N2N1O2N2N1O2N1O2N2N101N2N1O2N2N1O2N2N101N1O2O1N1O2O1N1O2O1N1O2O1N1O2O1N1O2N101N2N101N2N101N2N101N1O101N1O100O1O2O0O1O1O101N1O100O2N100O1O1O2O0O1O101N1O100O1O2N100O1O2hN]IPJd6P6_IkIc6U6_IhIb6X6`IeIa6[6aIaIb6^6bI]I_6c6jISIW6m6n00010O0000001O000000001O0000000000000001O00000000000000000000000000000000001O00000000000nHQIS6o6iIVIV6j6fI[IY6e6cI`I\\6`6`IeI_6[6]IjIb6V6ZInIf6W7O1O1O1N2O1O1O001O1O1O1N2O1O1O001O1O1O1N2O1O1O001O1O1O1N101O1O1O1O001O1N2O001O1N2N2O0O2N2O1N1O2O1N2N2O0O2N2O1N1O2M3N2N2N3M3M3M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3M3K5K5K5K6I9H8H8H8H8H8H8H8H_8"}, "label": "a white pot"}]}
{"id": 85939, "image": "COCO_train2014_000000085939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white vase with dents in it\"."}], "task": "refering", "answer_template": "The \"a white vase with dents in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 63, 64, 65, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251], "bbox": [0.596625, 0.07920833333333334, 0.9977499999999999, 0.6095416666666666], "iscrowd": 0, "area": 43962.88979999999, "rle": {"size": [480, 640], "counts": "ZWc5i0V>6I7I6K6I7I6J7J5J3M3M4M2M3O1N2N2N2N3M2N2N2N2N2O1N2N1O2N2N1O2N1O2N2N101N2N1O2N2N1O2N2N101N1O2O1N1O2O1N1O2O1N1O2O1N1O2O1N1O2N101N2N101N2N101N2N101N1O101N1O100O1O2O0O1O1O101N1O100O2N100O1O1O2O0O1O101N1O100O1O2N100O1O2hN]IPJd6P6_IkIc6U6_IhIb6X6`IeIa6[6aIaIb6^6bI]I_6c6jISIW6m6n00010O0000001O000000001O0000000000000001O00000000000000000000000000000000001O00000000000nHQIS6o6iIVIV6j6fI[IY6e6cI`I\\6`6`IeI_6[6]IjIb6V6ZInIf6W7O1O1O1N2O1O1O001O1O1O1N2O1O1O001O1O1O1N2O1O1O001O1O1O1N101O1O1O1O001O1N2O001O1N2N2O0O2N2O1N1O2O1N2N2O0O2N2O1N1O2M3N2N2N3M3M3M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3M3K5K5K5K6I9H8H8H8H8H8H8H8H_8"}, "label": "a white vase with dents in it"}]}
{"id": 85939, "image": "COCO_train2014_000000085939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black vase on a table\"."}], "task": "refering", "answer_template": "The \"black vase on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 25, 26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 280, 281, 282], "bbox": [0.017499999999999998, 0.05, 0.47087500000000004, 0.7234583333333333], "iscrowd": 0, "area": 61483.807949999995, "rle": {"size": [480, 640], "counts": "k\\55[>a0_Ob0^Ob0C<M4M2N3L3N3L3N3M2M4M2N3L3N3L3N3M2M4M2N3L3N3M2N3M2N3M3L3N2N2N2N1O2M3N2N2N2M2O2N2M3N2N2N2N101N2N2N2N2N2N101N2N2N2N2N1O2O0YMdIiL]6V3dIiL^6V3cIhL^6X3bIgL_6X3cIfL_6Y3aIfL`6Z3aIdLa6Z3`IeLa6[3`IcLa6]3`IaLb6]3aI`L`6`3bI]L`6b3bI[L_6d3dIYL]6g3eIVL]6i3eITL\\6k3fISL\\6l3fIQL[6o3gInKZ6R4hIkKZ6S4iIjKX6V4jIgKX6X4lIbKV6^4QJWKS6i4TJmJP6R5o10001O001N10001O001O00000000000O1000001O000000000000000O10000000000000001O000O1000000000000000000O100000001O000000000O1000000001O2N3lFoJS8S5fGUKW8n4bGYK[8i4^G`K^8d5N3L3M3M4L3M001O1O001O1O1O0O2O1N1O2O1N2O0O2O1N2O0O2O0O1O2O0O2O0O101N100O2O0O1O2O0O101N101N100O2O0O1O2O0O2O0O2O1N101N2O0O2N2O0O2O1N101N2O0O2O1N1O2O1N101N2O0O2O1N101N2N101N2O0O2O1N101N2N1N3N2N1N3N2N1O2M3N1O2M3N1O2M2O2N2N1N3N2N2M3N2N3L3N2N2N3L3N2N3L3K5I7J7H7I7J6I8_Ongn4"}, "label": "black vase on a table"}]}
{"id": 85939, "image": "COCO_train2014_000000085939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black vase on white table , to the left of white vase\"."}], "task": "refering", "answer_template": "The \"black vase on white table , to the left of white vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 25, 26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 280, 281, 282], "bbox": [0.017499999999999998, 0.05, 0.47087500000000004, 0.7234583333333333], "iscrowd": 0, "area": 61483.807949999995, "rle": {"size": [480, 640], "counts": "k\\55[>a0_Ob0^Ob0C<M4M2N3L3N3L3N3M2M4M2N3L3N3L3N3M2M4M2N3L3N3M2N3M2N3M3L3N2N2N2N1O2M3N2N2N2M2O2N2M3N2N2N2N101N2N2N2N2N2N101N2N2N2N2N1O2O0YMdIiL]6V3dIiL^6V3cIhL^6X3bIgL_6X3cIfL_6Y3aIfL`6Z3aIdLa6Z3`IeLa6[3`IcLa6]3`IaLb6]3aI`L`6`3bI]L`6b3bI[L_6d3dIYL]6g3eIVL]6i3eITL\\6k3fISL\\6l3fIQL[6o3gInKZ6R4hIkKZ6S4iIjKX6V4jIgKX6X4lIbKV6^4QJWKS6i4TJmJP6R5o10001O001N10001O001O00000000000O1000001O000000000000000O10000000000000001O000O1000000000000000000O100000001O000000000O1000000001O2N3lFoJS8S5fGUKW8n4bGYK[8i4^G`K^8d5N3L3M3M4L3M001O1O001O1O1O0O2O1N1O2O1N2O0O2O1N2O0O2O0O1O2O0O2O0O101N100O2O0O1O2O0O101N101N100O2O0O1O2O0O2O0O2O1N101N2O0O2N2O0O2O1N101N2O0O2O1N1O2O1N101N2O0O2O1N101N2N101N2O0O2O1N101N2N1N3N2N1N3N2N1O2M3N1O2M3N1O2M2O2N2N1N3N2N2M3N2N3L3N2N2N3L3N2N3L3K5I7J7H7I7J6I8_Ongn4"}, "label": "black vase on white table , to the left of white vase"}]}
{"id": 85939, "image": "COCO_train2014_000000085939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown and black vase behind and to the left of the all black vase\"."}], "task": "refering", "answer_template": "The \"the brown and black vase behind and to the left of the all black vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 115, 116, 138, 161, 184], "bbox": [0.0, 0.0, 0.10232812499999999, 0.4981875], "iscrowd": 0, "area": 8538.4042, "rle": {"size": [480, 640], "counts": "0e4\\:i2WMN1O2N2O0O2N2K5J5K6K5J5K6K5J6J5N3N2M2O2N2M2O2QOXFbLk9\\3XFaLj9]3XFaLi9]3ZF`Li9^3YF`Li9]3ZF`Lh9_3ZF_Lh9^3[F`Lg9^3\\F^Lf9`3]F^Le9`3o0M3N1O2M3N1N3N2M3N2M3N2M3N2M3N2N2M3N2M3M3L4L4L4L4L4M3L4L4L4Bi_]8"}, "label": "the brown and black vase behind and to the left of the all black vase"}]}
{"id": 85939, "image": "COCO_train2014_000000085939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and black vase with a design of diamonds and lines\"."}], "task": "refering", "answer_template": "The \"a brown and black vase with a design of diamonds and lines\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 115, 116, 138, 161, 184], "bbox": [0.0, 0.0, 0.10232812499999999, 0.4981875], "iscrowd": 0, "area": 8538.4042, "rle": {"size": [480, 640], "counts": "0e4\\:i2WMN1O2N2O0O2N2K5J5K6K5J5K6K5J6J5N3N2M2O2N2M2O2QOXFbLk9\\3XFaLj9]3XFaLi9]3ZF`Li9^3YF`Li9]3ZF`Lh9_3ZF_Lh9^3[F`Lg9^3\\F^Lf9`3]F^Le9`3o0M3N1O2M3N1N3N2M3N2M3N2M3N2M3N2N2M3N2M3M3L4L4L4L4L4M3L4L4L4Bi_]8"}, "label": "a brown and black vase with a design of diamonds and lines"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with his mouth open\"."}], "task": "refering", "answer_template": "The \"a man with his mouth open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 19, 20, 21, 22, 23, 33, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 53, 54, 55, 63, 64, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0033723653395784543, 0.026968750000000003, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 184660.09415, "rle": {"size": [640, 427], "counts": "_R1<`c0m3TLX1gN4L4M3L4M3L4L4M3L4M3L4L4M3L4M3L4L4M3L4M3L4L4M3L4O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1UJZF\\1f9bNeFT1\\9iNPGm0Q9QOYGg0g8WOdG?]8_OiG=W8@nG=S8AQH<P8BTH<l7BXH;i7B\\H<d7B`H;a7CcH:^7DfH:Z7CkH:V7DnH9S7EQI9o6EUI8l6EXI:h6D\\I9e6E_I8b6FbI8^6EgI8Z6FjI8V6FnI7S6FRJ7o5GUJ7k5GYJ6h5H\\J6d5G`J7a5GcJ6^5HfJ5[5IhJ5Y5HlJ5U5InJ4T5JPK3Q5KSK2n4KVK2l4LXK1i4MZK1g4M]KOe4N`KOa4ObKO_4OeKM]41gKLZ42iKLX44jKIW47jKFX4:jKCW4<lKAU4?lK^OV4b0lK[OU4d0nKYOS4g0nKVOT4j0nKSOS4l0oKROR4n0PLnNR4R1PLkNQ4U1]7O1000O01000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O001O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2cCkLk7V3SHlLl7V3QHlLn7U3QHmLm7U3PHmLo7T3PHmLo7T3oGoLo7T3mGnLR8U3kGlLT8V3iGmLU8V3gGlLX8W3eGjLZ8X3cGkL[8X3bGiL]8Z3_GhL`8Z3]GhLb8[3[GgLc8\\3YGfLf8]3VGeLi8]3UGeLi8^3SGdLl8k3eFVLZ9Z7N1O1O2N1O1O2N1O1O1O001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O000O101O00001O00001O001O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1^Nb1hMY2gMX2hMjG"}, "label": "a man with his mouth open"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black and white striped suit wearing a large amount of eye shadow all around his eyes\"."}], "task": "refering", "answer_template": "The \"a man in a black and white striped suit wearing a large amount of eye shadow all around his eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 19, 20, 21, 22, 23, 33, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 53, 54, 55, 63, 64, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0033723653395784543, 0.026968750000000003, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 184660.09415, "rle": {"size": [640, 427], "counts": "_R1<`c0m3TLX1gN4L4M3L4M3L4L4M3L4M3L4L4M3L4M3L4L4M3L4M3L4L4M3L4O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1UJZF\\1f9bNeFT1\\9iNPGm0Q9QOYGg0g8WOdG?]8_OiG=W8@nG=S8AQH<P8BTH<l7BXH;i7B\\H<d7B`H;a7CcH:^7DfH:Z7CkH:V7DnH9S7EQI9o6EUI8l6EXI:h6D\\I9e6E_I8b6FbI8^6EgI8Z6FjI8V6FnI7S6FRJ7o5GUJ7k5GYJ6h5H\\J6d5G`J7a5GcJ6^5HfJ5[5IhJ5Y5HlJ5U5InJ4T5JPK3Q5KSK2n4KVK2l4LXK1i4MZK1g4M]KOe4N`KOa4ObKO_4OeKM]41gKLZ42iKLX44jKIW47jKFX4:jKCW4<lKAU4?lK^OV4b0lK[OU4d0nKYOS4g0nKVOT4j0nKSOS4l0oKROR4n0PLnNR4R1PLkNQ4U1]7O1000O01000000O10000O10000O1000000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O001O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2cCkLk7V3SHlLl7V3QHlLn7U3QHmLm7U3PHmLo7T3PHmLo7T3oGoLo7T3mGnLR8U3kGlLT8V3iGmLU8V3gGlLX8W3eGjLZ8X3cGkL[8X3bGiL]8Z3_GhL`8Z3]GhLb8[3[GgLc8\\3YGfLf8]3VGeLi8]3UGeLi8^3SGdLl8k3eFVLZ9Z7N1O1O2N1O1O2N1O1O1O001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O000O101O00001O00001O001O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1^Nb1hMY2gMX2hMjG"}, "label": "a man in a black and white striped suit wearing a large amount of eye shadow all around his eyes"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing the tie\"."}], "task": "refering", "answer_template": "The \"a man wearing the tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 185, 186, 187, 200, 201, 202, 215, 216, 217, 230, 231, 245, 246, 260, 261, 275, 276, 290, 291, 305, 306, 320], "bbox": [0.3469086651053864, 0.47246875, 0.572576112412178, 0.9353906249999999], "iscrowd": 0, "area": 11880.724550000003, "rle": {"size": [640, 427], "counts": "mkl2j0Tc0o0ROn0kL[MTCc3o;_LmC^4W;eKfDm4i:UKSE^5[:eJbEk5o9WJnEn5m9UJoEQ6m9QJPFT6k9oIQFW6l9jIQF[6m9eIoEa6n9`IoEe6n9\\IoEb6V:^IfEa6^:`I_E^6e:dIUE[6Q;eIjD[6Z;fIbDY6a;iIZDV6i;n03M3L4M3M3L4M3M2M4M3M3L4M3M3L4M2N3L4M3M3L4M3M3L3N3M3L4F:@`0\\Lj@n1f?RN^@Y1S`0fNQ@g0^`0YOg_O2j`0NZ_O^OVa0a0U1O100O1O10O01O100O00010O0010O01O101N1O2O0O1O2O0O2O0O1O2O0O2N101N1O101N1O2O0O1OXRb3"}, "label": "a man wearing the tie"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the necktie of a man wearing stripes\"."}], "task": "refering", "answer_template": "The \"the necktie of a man wearing stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 185, 186, 187, 200, 201, 202, 215, 216, 217, 230, 231, 245, 246, 260, 261, 275, 276, 290, 291, 305, 306, 320], "bbox": [0.3469086651053864, 0.47246875, 0.572576112412178, 0.9353906249999999], "iscrowd": 0, "area": 11880.724550000003, "rle": {"size": [640, 427], "counts": "mkl2j0Tc0o0ROn0kL[MTCc3o;_LmC^4W;eKfDm4i:UKSE^5[:eJbEk5o9WJnEn5m9UJoEQ6m9QJPFT6k9oIQFW6l9jIQF[6m9eIoEa6n9`IoEe6n9\\IoEb6V:^IfEa6^:`I_E^6e:dIUE[6Q;eIjD[6Z;fIbDY6a;iIZDV6i;n03M3L4M3M3L4M3M2M4M3M3L4M3M3L4M2N3L4M3M3L4M3M3L3N3M3L4F:@`0\\Lj@n1f?RN^@Y1S`0fNQ@g0^`0YOg_O2j`0NZ_O^OVa0a0U1O100O1O10O01O100O00010O0010O01O101N1O2O0O1O2O0O2O0O1O2O0O2N101N1O101N1O2O0O1OXRb3"}, "label": "the necktie of a man wearing stripes"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man hiding back at white and black jacket wearing man\"."}], "task": "refering", "answer_template": "The \"a man hiding back at white and black jacket wearing man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 122, 123, 124, 136, 137, 138, 139, 150, 151, 152, 153, 154, 165, 166, 167, 168, 180, 181, 182, 195, 196, 210, 225], "bbox": [0.0, 0.3086875, 0.31803278688524594, 0.70146875], "iscrowd": 0, "area": 12798.358350000002, "rle": {"size": [640, 427], "counts": "h9X4f?000O0100000O100000O01001O1O2N1O1O2M2O2N1O2N1O1O2M2O2N1O1O2N1O2N1N2O2N1O2N1O1O2M2O1O1O1001O000000001O00000000001O000000010O00000001O01O000001O01O000001O0001O0001O0001O0001O0000O2N100O1O101N1O100O1O101N1O100O1012M3M3L4M3L4M3L4M4K4M3L4M3L4M3L4M3L4M3L5L3L4M3L4M3L4M3L4M3L4Mmce5"}, "label": "a man hiding back at white and black jacket wearing man"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man behind and to the left of the man dressed like beetle juice\"."}], "task": "refering", "answer_template": "The \"the man behind and to the left of the man dressed like beetle juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 122, 123, 124, 136, 137, 138, 139, 150, 151, 152, 153, 154, 165, 166, 167, 168, 180, 181, 182, 195, 196, 210, 225], "bbox": [0.0, 0.3086875, 0.31803278688524594, 0.70146875], "iscrowd": 0, "area": 12798.358350000002, "rle": {"size": [640, 427], "counts": "h9X4f?000O0100000O100000O01001O1O2N1O1O2M2O2N1O2N1O1O2M2O2N1O1O2N1O2N1N2O2N1O2N1O1O2M2O1O1O1001O000000001O00000000001O000000010O00000001O01O000001O01O000001O0001O0001O0001O0001O0000O2N100O1O101N1O100O1O101N1O100O1012M3M3L4M3L4M3L4M4K4M3L4M3L4M3L4M3L4M3L5L3L4M3L4M3L4M3L4M3L4Mmce5"}, "label": "the man behind and to the left of the man dressed like beetle juice"}]}
{"id": 364468, "image": "COCO_train2014_000000364468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person head on right back\"."}], "task": "refering", "answer_template": "The \"person head on right back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 131, 132, 133, 134, 145, 146, 147, 148, 149, 161, 162, 163, 164, 178, 179], "bbox": [0.7185948477751756, 0.3294375, 1.0, 0.5228125], "iscrowd": 0, "area": 9257.69885, "rle": {"size": [640, 427], "counts": "oSP6`0]c04K6L3M3N2N2M3L5K4M2M3O2N1N2O2N1O1O2N100O2O0O2O000O2O0O101N10001N100O2O00001O0O101O00001O0O101O001O00001O00001O00001O00001O00001O00010O000010O01O0010O010O010O010O0100O0100O010O010O101N100O101N100O101N100000000O10000000000O10000000001N10k0UOUG"}, "label": "person head on right back"}]}
{"id": 28595, "image": "COCO_train2014_000000028595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child in pink and blue smiles gleefully at a kite\"."}], "task": "refering", "answer_template": "The \"a young child in pink and blue smiles gleefully at a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 204, 205, 206, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 321, 322, 323, 324, 325, 336, 337, 338, 339], "bbox": [0.30985948477751757, 0.37184375, 0.7511007025761125, 1.0], "iscrowd": 0, "area": 48218.73230000001, "rle": {"size": [640, 427], "counts": "W]c23mc02N1c^OOh>4h@>T?Df@d0V?^Of@h0U?[Og@k0U?WOg@l0X?VOd@m0Z?VOb@m0]?UO_@n0`?TO\\@n0d?h1N2001O1N101O1On_OYMj>g2i@XNf>f3LN2O2N1N3N2N5J;F5K4K4MO1Nb0_O4K4L5L4K4L4M3L4L4M2M4WMTJWGP6e8TJXGR7a7SIZHQ7b7SIZHQ7b7SI[HQ7`7SI\\HQ7`7SI]HP7_7TI]HQ7^7SI^HQ7^7SI_HP7]7TI_Hn6_7VI^Hk6`7ZI\\Hf6d7^IWHd6h7_IUHb6i7bITH^6l7cISH^6l7cISH^6_OjGk7i1eHR7X7oHfHT7W7nHhHU7U7lHjHW7S7jHlHX7R7iHmHW7R7kHmHU7S7lHlHT7T7mHjHT7V7PIfHP7Z7TIbHl6]7l2000000O100000000O1000O100000000000000000000000000000O0100000000000000000000000000000001O0000000000000000000_FbH`6^7]IfHb6Z7ZIkHe6Z7SIjHl6[7jHkHU7Z7aHlH^7X7XHoHg7[7cGQI]8Z9N2M3L4dHoDe5S;[JPE_5S;`JPE[5T;cJPEW5S;iJoDR5T;mJPEm4S;RKPEj4S;UKPEf4R;YKQEd4P;\\KSE`4n:`KTE?ROR2l;]MVE:UOT2d9EYGSNXOU2^9I\\GoMZOT2Z9N_GjM[OT2V93aGgM]OQ2S98cGcM^OR2n8;gG`M_OQ2j8`0iG\\MAP2f8e0mGVMo0d0S7W2THnLU1?h6c2YH`La1b0W6n2`601000O10000O01000O100001N1VMZ@Y1h?eN\\@V1f?jN^@Q1d?mN`@o0a?QOc@j0_?TOf@g0[?YOj@a0X?]Om@=U?Co@8R?GTAKW?3^2LXoQ2"}, "label": "a young child in pink and blue smiles gleefully at a kite"}]}
{"id": 28595, "image": "COCO_train2014_000000028595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl in a pink jacket\"."}], "task": "refering", "answer_template": "The \"a young girl in a pink jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 204, 205, 206, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 321, 322, 323, 324, 325, 336, 337, 338, 339], "bbox": [0.30985948477751757, 0.37184375, 0.7511007025761125, 1.0], "iscrowd": 0, "area": 48218.73230000001, "rle": {"size": [640, 427], "counts": "W]c23mc02N1c^OOh>4h@>T?Df@d0V?^Of@h0U?[Og@k0U?WOg@l0X?VOd@m0Z?VOb@m0]?UO_@n0`?TO\\@n0d?h1N2001O1N101O1On_OYMj>g2i@XNf>f3LN2O2N1N3N2N5J;F5K4K4MO1Nb0_O4K4L5L4K4L4M3L4L4M2M4WMTJWGP6e8TJXGR7a7SIZHQ7b7SIZHQ7b7SI[HQ7`7SI\\HQ7`7SI]HP7_7TI]HQ7^7SI^HQ7^7SI_HP7]7TI_Hn6_7VI^Hk6`7ZI\\Hf6d7^IWHd6h7_IUHb6i7bITH^6l7cISH^6l7cISH^6_OjGk7i1eHR7X7oHfHT7W7nHhHU7U7lHjHW7S7jHlHX7R7iHmHW7R7kHmHU7S7lHlHT7T7mHjHT7V7PIfHP7Z7TIbHl6]7l2000000O100000000O1000O100000000000000000000000000000O0100000000000000000000000000000001O0000000000000000000_FbH`6^7]IfHb6Z7ZIkHe6Z7SIjHl6[7jHkHU7Z7aHlH^7X7XHoHg7[7cGQI]8Z9N2M3L4dHoDe5S;[JPE_5S;`JPE[5T;cJPEW5S;iJoDR5T;mJPEm4S;RKPEj4S;UKPEf4R;YKQEd4P;\\KSE`4n:`KTE?ROR2l;]MVE:UOT2d9EYGSNXOU2^9I\\GoMZOT2Z9N_GjM[OT2V93aGgM]OQ2S98cGcM^OR2n8;gG`M_OQ2j8`0iG\\MAP2f8e0mGVMo0d0S7W2THnLU1?h6c2YH`La1b0W6n2`601000O10000O01000O100001N1VMZ@Y1h?eN\\@V1f?jN^@Q1d?mN`@o0a?QOc@j0_?TOf@g0[?YOj@a0X?]Om@=U?Co@8R?GTAKW?3^2LXoQ2"}, "label": "a young girl in a pink jacket"}]}
{"id": 225210, "image": "COCO_train2014_000000225210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on surfboard bottom right screen\"."}], "task": "refering", "answer_template": "The \"person on surfboard bottom right screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [314, 315, 316, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384, 404, 405, 406, 407, 426, 427, 428, 429, 448, 449, 450, 451, 470, 471, 472, 473, 474], "bbox": [0.17928104575163398, 0.6369281045751635, 0.5709477124183006, 0.9937581699346406], "iscrowd": 0, "area": 14779.062450000003, "rle": {"size": [612, 612], "counts": "RUR22Qc02N2O1O0O2O1O001N2O001O001O1O00100O001O1O1O010O1O1O01O00007I1O001O00001O00000000000001O000000001O000000001N100000001O000000001O000000001O0000001O00000O101^O]Oa^Oc0_a0A]^O`0ba0C\\^O>ba0E[^O<Xa0^Oh^O60>Ta01j^O2Ra00n^O1o`02P_O0m`02Q_O1k`0W1M2M4M3M2O2N2N1O2N2N1O1O001O1O1O1O001O100O10001N101N100O10O1fMZ@U1e?kMZ@=1K3l1h?PN]@0OP2U`0mMm_OS2b`010O01l@ZNh<f1jBPOo<o0gBCQ=>eB3S=NcB<\\=E[Bc0f=]201O01O00001O00001O0000001O1O1O1O1O1O1O1O1O1O1K5A?K5K6J6J5K6J6J5L5J6J6L3O2N2N101N2N2N001001O2Q_OoMf`0Z2NO2N1O1O1O2O04L4M3L2N10O01O100O010O1O11N2O2N1O2L3M3M4L3M4L3M3M4L3M7I4K1SOQ^Oe0Pb0YOQ^Og0oa0YOQ^Og0Pb0WOR^Oh0na0WOS^Oi0ma0VOU^O5J=Yb0Cg]O=Zb0Bg]O=Yb0@k]O?]b001O1O001O001O1O001O1O002N1O1O2N1O1Oobl4"}, "label": "person on surfboard bottom right screen"}]}
{"id": 225210, "image": "COCO_train2014_000000225210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfer with both arms outstreched\"."}], "task": "refering", "answer_template": "The \"a surfer with both arms outstreched\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [314, 315, 316, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384, 404, 405, 406, 407, 426, 427, 428, 429, 448, 449, 450, 451, 470, 471, 472, 473, 474], "bbox": [0.17928104575163398, 0.6369281045751635, 0.5709477124183006, 0.9937581699346406], "iscrowd": 0, "area": 14779.062450000003, "rle": {"size": [612, 612], "counts": "RUR22Qc02N2O1O0O2O1O001N2O001O001O1O00100O001O1O1O010O1O1O01O00007I1O001O00001O00000000000001O000000001O000000001N100000001O000000001O000000001O0000001O00000O101^O]Oa^Oc0_a0A]^O`0ba0C\\^O>ba0E[^O<Xa0^Oh^O60>Ta01j^O2Ra00n^O1o`02P_O0m`02Q_O1k`0W1M2M4M3M2O2N2N1O2N2N1O1O001O1O1O1O001O100O10001N101N100O10O1fMZ@U1e?kMZ@=1K3l1h?PN]@0OP2U`0mMm_OS2b`010O01l@ZNh<f1jBPOo<o0gBCQ=>eB3S=NcB<\\=E[Bc0f=]201O01O00001O00001O0000001O1O1O1O1O1O1O1O1O1O1K5A?K5K6J6J5K6J6J5L5J6J6L3O2N2N101N2N2N001001O2Q_OoMf`0Z2NO2N1O1O1O2O04L4M3L2N10O01O100O010O1O11N2O2N1O2L3M3M4L3M4L3M3M4L3M7I4K1SOQ^Oe0Pb0YOQ^Og0oa0YOQ^Og0Pb0WOR^Oh0na0WOS^Oi0ma0VOU^O5J=Yb0Cg]O=Zb0Bg]O=Yb0@k]O?]b001O1O001O001O1O001O1O002N1O1O2N1O1Oobl4"}, "label": "a surfer with both arms outstreched"}]}
{"id": 225210, "image": "COCO_train2014_000000225210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfer bend forward and touching his left leg\"."}], "task": "refering", "answer_template": "The \"a surfer bend forward and touching his left leg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 53, 54, 55, 56, 75, 76, 77, 78, 97, 98, 99, 121], "bbox": [0.31687908496732026, 0.0467483660130719, 0.5558496732026144, 0.27013071895424834], "iscrowd": 0, "area": 7202.01665, "rle": {"size": [612, 612], "counts": "\\oc36nb00000001O0O10001O00000O2O000000000O101O0000000O10001O00000001O000010O000001O000010O0001O000010O000010O00010O00010O0001O01O01O01O01O01O01O01O]^OOS`02f_Om0a?SOY@T1f?lNW@X1i?hNU@Z1j?fNU@\\1j?dNU@^1k?bNS@_1m?aNR@a1m?_NR@c1n?]NP@e1o?[No_Og1Q`0YNl_Oi1U`0XNg_Oi1\\`0>1O1O1O1N2N2L4L4L3M4M32N1O2N2N2N1O2N1O1N3N1O2N1O2M5L3K5G:I5M3L4L4M3N2NULl@h3R?WLo@k3o>ULQAl3n>TLRAl3o>RLRAo3m>QLSAo3m>QLSAo3Q?1O00001OhN^AnMc>P2dAkM[>T2lAfMT>Y2\\1O1bMe_OR2\\`0hMl_OU2d`0N2M3M4L3M4L4VOj0G9Go_R5"}, "label": "a surfer bend forward and touching his left leg"}]}
{"id": 225210, "image": "COCO_train2014_000000225210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfer on top of the way with his hand in the air and a hand on his knee\"."}], "task": "refering", "answer_template": "The \"a surfer on top of the way with his hand in the air and a hand on his knee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 53, 54, 55, 56, 75, 76, 77, 78, 97, 98, 99, 121], "bbox": [0.31687908496732026, 0.0467483660130719, 0.5558496732026144, 0.27013071895424834], "iscrowd": 0, "area": 7202.01665, "rle": {"size": [612, 612], "counts": "\\oc36nb00000001O0O10001O00000O2O000000000O101O0000000O10001O00000001O000010O000001O000010O0001O000010O000010O00010O00010O0001O01O01O01O01O01O01O01O]^OOS`02f_Om0a?SOY@T1f?lNW@X1i?hNU@Z1j?fNU@\\1j?dNU@^1k?bNS@_1m?aNR@a1m?_NR@c1n?]NP@e1o?[No_Og1Q`0YNl_Oi1U`0XNg_Oi1\\`0>1O1O1O1N2N2L4L4L3M4M32N1O2N2N2N1O2N1O1N3N1O2N1O2M5L3K5G:I5M3L4L4M3N2NULl@h3R?WLo@k3o>ULQAl3n>TLRAl3o>RLRAo3m>QLSAo3m>QLSAo3Q?1O00001OhN^AnMc>P2dAkM[>T2lAfMT>Y2\\1O1bMe_OR2\\`0hMl_OU2d`0N2M3M4L3M4L4VOj0G9Go_R5"}, "label": "a surfer on top of the way with his hand in the air and a hand on his knee"}]}
{"id": 561082, "image": "COCO_train2014_000000561082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark colored beverage in a glass next to a plate of food\"."}], "task": "refering", "answer_template": "The \"a dark colored beverage in a glass next to a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113], "bbox": [0.81696875, 0.0002810304449648712, 0.999015625, 0.3236065573770492], "iscrowd": 0, "area": 10280.838150000003, "rle": {"size": [427, 640], "counts": "YVj69m<4M4L3M4K5L4mCPOg;U1UDoNe;^1L5K4L5J6K4L5K4K6K4L5J5L5K4L5J5L4L4L4O2O0000000000001O0000000000000001O0000000O1000000000000000000O100N2O1N2N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N[="}, "label": "a dark colored beverage in a glass next to a plate of food"}]}
{"id": 561082, "image": "COCO_train2014_000000561082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass top right corner\"."}], "task": "refering", "answer_template": "The \"glass top right corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113], "bbox": [0.81696875, 0.0002810304449648712, 0.999015625, 0.3236065573770492], "iscrowd": 0, "area": 10280.838150000003, "rle": {"size": [427, 640], "counts": "YVj69m<4M4L3M4K5L4mCPOg;U1UDoNe;^1L5K4L5J6K4L5K4K6K4L5J5L5K4L5J5L4L4L4O2O0000000000001O0000000000000001O0000000O1000000000000000000O100N2O1N2N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N[="}, "label": "glass top right corner"}]}
{"id": 126909, "image": "COCO_train2014_000000126909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue tie\"."}], "task": "refering", "answer_template": "The \"the blue tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 16, 17, 28, 29, 39, 40, 41, 51, 52, 53, 63, 64, 65, 75, 76, 77, 87, 88, 89, 99, 100, 101, 110, 111, 112, 113, 122, 123, 124, 125, 134, 135, 136, 137, 146, 147, 148, 149, 158, 159, 160, 171, 172], "bbox": [0.20582582582582584, 0.013099999999999999, 0.49600600600600603, 0.81086], "iscrowd": 0, "area": 21270.9788, "rle": {"size": [500, 333], "counts": "SQR17V?9G8H9G8H9G8H8H9G8H9G8H9G8H8H9G8H9G8G:G8H9G8H8H9H7I8I6J7I6J6I8I6J6J4L5J5L4L4L4L5J5L4L4L5K4K5L4L4L5K4K5L4L5K4L4K5L4O20O1O1O100O1O1O1O101N1J6@`0_Oa0_Oa0_Oa0@`0_Oa0_Oa0_Oa0@a0^Oa0_Oa0@`0_Oa0_Oa0_Oa0@`0C=H8G9H8G:G8G9H8G9HnPb2"}, "label": "the blue tie"}]}
{"id": 126909, "image": "COCO_train2014_000000126909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark blue tie with white symbol and writing\"."}], "task": "refering", "answer_template": "The \"dark blue tie with white symbol and writing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 16, 17, 28, 29, 39, 40, 41, 51, 52, 53, 63, 64, 65, 75, 76, 77, 87, 88, 89, 99, 100, 101, 110, 111, 112, 113, 122, 123, 124, 125, 134, 135, 136, 137, 146, 147, 148, 149, 158, 159, 160, 171, 172], "bbox": [0.20582582582582584, 0.013099999999999999, 0.49600600600600603, 0.81086], "iscrowd": 0, "area": 21270.9788, "rle": {"size": [500, 333], "counts": "SQR17V?9G8H9G8H9G8H8H9G8H9G8H9G8H8H9G8H9G8G:G8H9G8H8H9H7I8I6J7I6J6I8I6J6J4L5J5L4L4L4L5J5L4L4L5K4K5L4L4L5K4K5L4L5K4L4K5L4O20O1O1O100O1O1O1O101N1J6@`0_Oa0_Oa0_Oa0@`0_Oa0_Oa0_Oa0@a0^Oa0_Oa0@`0_Oa0_Oa0_Oa0@`0C=H8G9H8G:G8G9H8G9HnPb2"}, "label": "dark blue tie with white symbol and writing"}]}
{"id": 126909, "image": "COCO_train2014_000000126909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white diagonally striped tie\"."}], "task": "refering", "answer_template": "The \"black and white diagonally striped tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 12, 13, 14, 15, 24, 25, 26, 27, 36, 37, 38, 39, 48, 49, 50, 51, 60, 61, 62, 63, 72, 73, 74, 75, 84, 85, 86, 87, 96, 98, 99, 108, 109, 110, 120, 121, 122, 132, 133, 134, 144, 145, 146, 156, 157, 158, 168, 169, 170, 180, 181, 182, 192, 193, 194, 204, 205, 206], "bbox": [0.006756756756756757, 0.00636, 0.2868168168168168, 0.98614], "iscrowd": 0, "area": 24688.24505, "rle": {"size": [500, 333], "counts": "a^14R?>C=B>I7J6J6\\J\\NRMj1Z2mN]MY1n1_OiMg0c10TN6X1a0_NEl0R1lNTO`0c1WOcN4U2CQNIn1UMXKi2P3^OY2lMjJ]2S3ROe2dN[JQ2V3gNP3[OmIe1Y3^NY30aIY1\\3fNQ34fIm0_3oNi26kIb0b3XO`29QJ6e3AW2<WJJh3Jn1?3AJb06^OGe09[OEg0;YOBj0>VO_Om0a0RO]OQ1c0oNZOT1f0lNVLJ^O]1\\4iNTL3WOV1e4gNRL<oNP1n4eNQLd0gNj0X5bNoKm0_Nd0b5_NnKU1VN?l5\\NlK^1nM9V6YNjKg1gM3^6UNjKQ2^MMh6RNhKZ2VMGR7oMkJg5U5YJUJ]6k5cIPJb6P6^IlIf6T6ZIgIk6Y6UIcIo6]6QI^IT7b6lHZIX7f6gHVI^7j6bHRIb7n6^HmHg7S7=0000000000000000000kJYI]1g6eMeJm1NSLY3R1UN]2^NkMW2jNg0n2nL@V1_M\\2d2ZLV15WLn3U2iKn2SOmJb5g1XKX4V5ZKfJj4h5hJUJ[5Y6WJdIl5j6fIRI^6o73M3L5L3L4M3M3L4M3M3L4M3M4]Oj0POU1kNU1kNT1mNT1kNT1lN]ec3"}, "label": "black and white diagonally striped tie"}]}
{"id": 126909, "image": "COCO_train2014_000000126909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tie with smcs on it\"."}], "task": "refering", "answer_template": "The \"tie with smcs on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 18, 19, 20, 29, 30, 31, 41, 42, 43, 53, 54, 55, 65, 66, 67, 77, 78, 79, 89, 90, 91, 101, 102, 103, 113, 114, 115, 125, 126, 127, 137, 138, 139, 149, 150, 151, 160, 161, 162, 163, 172, 173, 174, 175, 184, 185, 186, 187, 196, 197, 198, 199, 208, 209, 210], "bbox": [0.3703303303303303, 0.0018, 0.6968768768768769, 1.0], "iscrowd": 0, "area": 31644.555850000004, "rle": {"size": [500, 333], "counts": "m`l11Y?;E;E;E;E;E;E;E;F:E;E;N2O1O1O1I8D;D;F:E;F:E;F:E;F:E;F:E;F:E;F:E;F:nL[FJP:2YF]OS:`0UFPOU:m0TFcNW:Z1S200000000C=A?J6I7J6J6I7J6J6N20000000000000000000000000000000000000XOh0^Ob0O1O1O1O1O1O1O1O1O1O1XEOl52RJ=`5D_Jk0R5VOmJX1e4iNYKf1Y4[NfKa2^3`M`Ld3]2]LbMf6[O[Id0S7nNnHP1a7bN`H]1n7UNSHi1\\8iMeGV2h8]MYGb2m:N2J6H8I7H8H8H8H8H8H8H8LWZa1"}, "label": "tie with smcs on it"}]}
{"id": 126909, "image": "COCO_train2014_000000126909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tie that says sm cs\"."}], "task": "refering", "answer_template": "The \"tie that says sm cs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 18, 19, 20, 29, 30, 31, 41, 42, 43, 53, 54, 55, 65, 66, 67, 77, 78, 79, 89, 90, 91, 101, 102, 103, 113, 114, 115, 125, 126, 127, 137, 138, 139, 149, 150, 151, 160, 161, 162, 163, 172, 173, 174, 175, 184, 185, 186, 187, 196, 197, 198, 199, 208, 209, 210], "bbox": [0.3703303303303303, 0.0018, 0.6968768768768769, 1.0], "iscrowd": 0, "area": 31644.555850000004, "rle": {"size": [500, 333], "counts": "m`l11Y?;E;E;E;E;E;E;E;F:E;E;N2O1O1O1I8D;D;F:E;F:E;F:E;F:E;F:E;F:E;F:E;F:nL[FJP:2YF]OS:`0UFPOU:m0TFcNW:Z1S200000000C=A?J6I7J6J6I7J6J6N20000000000000000000000000000000000000XOh0^Ob0O1O1O1O1O1O1O1O1O1O1XEOl52RJ=`5D_Jk0R5VOmJX1e4iNYKf1Y4[NfKa2^3`M`Ld3]2]LbMf6[O[Id0S7nNnHP1a7bN`H]1n7UNSHi1\\8iMeGV2h8]MYGb2m:N2J6H8I7H8H8H8H8H8H8H8LWZa1"}, "label": "tie that says sm cs"}]}
{"id": 126910, "image": "COCO_train2014_000000126910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plate on the table\"."}], "task": "refering", "answer_template": "The \"a white plate on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [305, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.8947968749999999, 1.0, 1.0], "iscrowd": 0, "area": 22858.63995000002, "rle": {"size": [640, 427], "counts": "bc0>gb0k0M300000O100000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10001O000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000O1000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000TO^^O"}, "label": "a white plate on the table"}]}
{"id": 126910, "image": "COCO_train2014_000000126910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate in front of the woman holding the hamburger\"."}], "task": "refering", "answer_template": "The \"the plate in front of the woman holding the hamburger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [305, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.8947968749999999, 1.0, 1.0], "iscrowd": 0, "area": 22858.63995000002, "rle": {"size": [640, 427], "counts": "bc0>gb0k0M300000O100000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10001O000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000O1000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000TO^^O"}, "label": "the plate in front of the woman holding the hamburger"}]}
{"id": 126910, "image": "COCO_train2014_000000126910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue cloth to the right of woman\"."}], "task": "refering", "answer_template": "The \"blue cloth to the right of woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 327, 328, 329], "bbox": [0.8066978922716627, 0.62509375, 1.0, 0.929515625], "iscrowd": 0, "area": 12959.30455, "rle": {"size": [640, 427], "counts": "YVh68ac07I7I7^Ob0K5L4K6J5K5K5L4J6I8H7I7H8H8H8G9H8H8G9N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O10000000000O1000000000000O100000000000000O100000000000000O010000000000000O10O100PJ[Be5W>UMbE"}, "label": "blue cloth to the right of woman"}]}
{"id": 126910, "image": "COCO_train2014_000000126910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the partial blue chair back to the woman ' s right\"."}], "task": "refering", "answer_template": "The \"the partial blue chair back to the woman ' s right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 327, 328, 329], "bbox": [0.8066978922716627, 0.62509375, 1.0, 0.929515625], "iscrowd": 0, "area": 12959.30455, "rle": {"size": [640, 427], "counts": "YVh68ac07I7I7^Ob0K5L4K6J5K5K5L4J6I8H7I7H8H8H8G9H8H8G9N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O10000000000O1000000000000O100000000000000O100000000000000O010000000000000O10O100PJ[Be5W>UMbE"}, "label": "the partial blue chair back to the woman ' s right"}]}
{"id": 544703, "image": "COCO_train2014_000000544703.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dining chair with a cushion\"."}], "task": "refering", "answer_template": "The \"a dining chair with a cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 209], "bbox": [0.0, 0.2989240506329114, 0.0998125, 0.8666772151898734], "iscrowd": 0, "area": 6722.023800000001, "rle": {"size": [316, 640], "counts": "Q3T1h8P2PN:F9G:F8H00000000000O100000000000O100000000000000000O100000O1000000TN_J]Oa55mJKS5F\\K:d4WOkKi0U4jNXLV1h3jNXLV1h3kNWLU1i3kNWLU1h3lNYLS1g3nNXLR1i3mNWLR1j3nNVLR1k3mNULS1l3kNULU1k3kNULU1l3jNTLV1m3iNSLW1m3hNTL]1h3TNfLR2_55K6L3OhMjHh1n7J6H8D<D;E<Db_a5"}, "label": "a dining chair with a cushion"}]}
{"id": 544703, "image": "COCO_train2014_000000544703.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown chair next to the tv\"."}], "task": "refering", "answer_template": "The \"the brown chair next to the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 219], "bbox": [0.318640625, 0.28313291139240504, 0.552046875, 0.9157278481012658], "iscrowd": 0, "area": 19480.999499999998, "rle": {"size": [316, 640], "counts": "^So15c95O2^NImHL@=_7MlHHD<^7OjHHF<\\71jHEI:\\7k0cHTO^7m0\\HXOd7`1OU1lN`0@<C6K0iJdKm4^4oJgKo4Y4oJkKo4c4M3[KSKW4o4fKTKX4n4eKUK;I]3c5_L^J`3f5[L[Je3i5WLWJi3n50000O10O100TOWL^K2@g3o4[L\\K2Dc3o4^LXK4G_3n4aLUK5MZ3l4T10000001O0O1000000000K5G900O10000000O10000000000000000000000000000000000O100000000000000000O1000000000000O1000000000000O100003M2M4M3M3NO00001O00001O0O2N2AnJQLT5n3RKkKQ5S4h0D;D6K5K7I8HO10O10O10000O1000O01PI^Mf6b2VIcMi6j2N2cIkLo5W3lInLR6S3kIQMS6^3N2NoLRJV23SMb5d0_JX2MTMe5a0bJ[2FUMj6n21N2N4L4EkHgM]7f0eH73SO_7>`HJOM7Jc76YHM4L51e7M[I3mah2"}, "label": "the brown chair next to the tv"}]}
{"id": 544703, "image": "COCO_train2014_000000544703.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair with a pillow on it\"."}], "task": "refering", "answer_template": "The \"chair with a pillow on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 219], "bbox": [0.318640625, 0.28313291139240504, 0.552046875, 0.9157278481012658], "iscrowd": 0, "area": 19480.999499999998, "rle": {"size": [316, 640], "counts": "^So15c95O2^NImHL@=_7MlHHD<^7OjHHF<\\71jHEI:\\7k0cHTO^7m0\\HXOd7`1OU1lN`0@<C6K0iJdKm4^4oJgKo4Y4oJkKo4c4M3[KSKW4o4fKTKX4n4eKUK;I]3c5_L^J`3f5[L[Je3i5WLWJi3n50000O10O100TOWL^K2@g3o4[L\\K2Dc3o4^LXK4G_3n4aLUK5MZ3l4T10000001O0O1000000000K5G900O10000000O10000000000000000000000000000000000O100000000000000000O1000000000000O1000000000000O100003M2M4M3M3NO00001O00001O0O2N2AnJQLT5n3RKkKQ5S4h0D;D6K5K7I8HO10O10O10000O1000O01PI^Mf6b2VIcMi6j2N2cIkLo5W3lInLR6S3kIQMS6^3N2NoLRJV23SMb5d0_JX2MTMe5a0bJ[2FUMj6n21N2N4L4EkHgM]7f0eH73SO_7>`HJOM7Jc76YHM4L51e7M[I3mah2"}, "label": "chair with a pillow on it"}]}
{"id": 470976, "image": "COCO_train2014_000000470976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a blue dress on the left side of the picture\"."}], "task": "refering", "answer_template": "The \"a lady in a blue dress on the left side of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 75, 76, 77, 92, 93, 94, 108, 109, 110, 111, 125, 126, 127, 128, 142, 143, 144, 145, 159, 160, 161, 162, 176, 177, 178, 179, 193, 194, 195, 196], "bbox": [0.3707692307692308, 0.12584375, 0.5770686070686071, 0.5123593750000001], "iscrowd": 0, "area": 16705.6389, "rle": {"size": [640, 481], "counts": "SS`36gc05K5K5J6K6J5K5K5V^OhNe`0^1T_OQO^`0T1\\_O[OV`0j0d_OFm?`2E4L4L4L4L3M3L4M2N3M3M3O1O10O0100O1O100O010O1O100kNcJfC^5X<gJdCZ5Z<lJbCS5]<RK_Co4_<WK]Ci4a<\\K[Ce4c<aKXC`4f<eKWC[4g<_1O2N2N200O10O0100O100O100O010O100O100O0101O001O1O1N2O1O1O1O1O1N2O0[NSCWLn<b3YC]Lh<\\3^CdLc<V3cCiL^<P3iCoLX<j2oCTMT<d2RD\\MQ<[2VDdMm;S2ZDlMj;j1]DUNf;b1`D^Nc;Y1dDeNa;R1eDmN^;j0iDUOZ;b0lD^O^;2iDMd;@cD?i;PO]Dn0d?G=C=CVbn3"}, "label": "a lady in a blue dress on the left side of the picture"}]}
{"id": 470976, "image": "COCO_train2014_000000470976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding a bowl of pasta while smiling\"."}], "task": "refering", "answer_template": "The \"woman holding a bowl of pasta while smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 75, 76, 77, 92, 93, 94, 108, 109, 110, 111, 125, 126, 127, 128, 142, 143, 144, 145, 159, 160, 161, 162, 176, 177, 178, 179, 193, 194, 195, 196], "bbox": [0.3707692307692308, 0.12584375, 0.5770686070686071, 0.5123593750000001], "iscrowd": 0, "area": 16705.6389, "rle": {"size": [640, 481], "counts": "SS`36gc05K5K5J6K6J5K5K5V^OhNe`0^1T_OQO^`0T1\\_O[OV`0j0d_OFm?`2E4L4L4L4L3M3L4M2N3M3M3O1O10O0100O1O100O010O1O100kNcJfC^5X<gJdCZ5Z<lJbCS5]<RK_Co4_<WK]Ci4a<\\K[Ce4c<aKXC`4f<eKWC[4g<_1O2N2N200O10O0100O100O100O010O100O100O0101O001O1O1N2O1O1O1O1O1N2O0[NSCWLn<b3YC]Lh<\\3^CdLc<V3cCiL^<P3iCoLX<j2oCTMT<d2RD\\MQ<[2VDdMm;S2ZDlMj;j1]DUNf;b1`D^Nc;Y1dDeNa;R1eDmN^;j0iDUOZ;b0lD^O^;2iDMd;@cD?i;PO]Dn0d?G=C=CVbn3"}, "label": "woman holding a bowl of pasta while smiling"}]}
{"id": 470976, "image": "COCO_train2014_000000470976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing near a woman\"."}], "task": "refering", "answer_template": "The \"a man standing near a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 61, 62, 63, 78, 79, 80, 81, 96, 97, 98, 113, 114, 115, 129, 130, 131, 132, 146, 147, 148, 149, 164, 165, 166, 182, 183, 198, 199, 200, 215, 216, 217, 232, 233, 234, 249, 250, 251, 266, 267, 268, 284, 285, 300, 301, 302, 303, 317, 318, 319, 320], "bbox": [0.5795426195426195, 0.095375, 0.8628898128898128, 0.8187656249999999], "iscrowd": 0, "area": 26655.109600000014, "rle": {"size": [640, 481], "counts": "jb^56jc04K3N4L2MBh\\O7Wc0Hn\\O6Pc0JT]O6hb0J[]O4db0L`]ONbb02f001O0000000000000001O000iK=QDD_L5m>`0_DK];=_DA_;e0\\D[Ob;i0\\D_O[;d0eD]OX;e0hD\\OV;e0jD\\Oe4TOaNb1YLLT5fN_Ng1]K>]NiM]74gNe1QKV1T6UMjNf1oJ^1POeLl67UOg1kJe1P6dLUOg1gJi1T6`LTOh1eJk1V6^LUOh1aJm1Z6[LUOh1^JP2[2RL_M=f2JQ3i1ZJR2]2PM5UOT3k1TJT2_2SM6nNW3l1oIV2b2UM5kNX3Y6`LRK5eN[3Y6]LXK5_N^3Y6ZL^K4ZNb3X6WLcK5UNc3Y6ULhK5oMf3Y6RLnK4jMj3X6oKSL5eMk3X6nKXL4aMn3W6lK\\L4]MP4U6kKcL3XMQ4U6jKhL2TMT4R6iKoL0PMX4o5gKUMOlL[4n5dKZMNiL_4k5bK`MLfLc4j5_KcMKdLg4j5ZKhMJ_Lm4h5WKPNFYLT5g5SKWNCSL[5e5PK`N^OlKc5b5nJiNYOfKj5^5mJoNWOPLa5l4ZKWOROoKe5k4UKYOTOmKh5k4PK[OUOlKl5j4kJ\\OXOkKn5i4gJm0j5cNTJ^1`6oM]IS2e6lMYIV2g6hMYIY2g6gMWI[2j6dMUI]2k6fMQI[2P7fMlH\\2T7dMhH`2Z7R51O1O1O1N2O1O1O1O1N2WKYHWMh7b2cHYM^7`2mH[MT7_2VI[Ml6^2_I]Mb6\\2iI_MY6Y2RJbMo5X2ZJdMg5U2dJeM^5T2mJgMT5;lGRNY3`1l4;RHkMZ3g1e4;dHWMP3[2]4:YMBi2;\\MBe2;`MBb29dMDb24bMJd2NaMNf2J_M3g2D_M9g2_O^M>h2ZO]Mb0j2UO\\Mh0Q3iNTMT1o:M3M2M4M3M3M1O2M2O1O2N1O2M2O1O2N1N3N1M4K4K5LahX1"}, "label": "a man standing near a woman"}]}
{"id": 470976, "image": "COCO_train2014_000000470976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man han in glass\"."}], "task": "refering", "answer_template": "The \"a man han in glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 61, 62, 63, 78, 79, 80, 81, 96, 97, 98, 113, 114, 115, 129, 130, 131, 132, 146, 147, 148, 149, 164, 165, 166, 182, 183, 198, 199, 200, 215, 216, 217, 232, 233, 234, 249, 250, 251, 266, 267, 268, 284, 285, 300, 301, 302, 303, 317, 318, 319, 320], "bbox": [0.5795426195426195, 0.095375, 0.8628898128898128, 0.8187656249999999], "iscrowd": 0, "area": 26655.109600000014, "rle": {"size": [640, 481], "counts": "jb^56jc04K3N4L2MBh\\O7Wc0Hn\\O6Pc0JT]O6hb0J[]O4db0L`]ONbb02f001O0000000000000001O000iK=QDD_L5m>`0_DK];=_DA_;e0\\D[Ob;i0\\D_O[;d0eD]OX;e0hD\\OV;e0jD\\Oe4TOaNb1YLLT5fN_Ng1]K>]NiM]74gNe1QKV1T6UMjNf1oJ^1POeLl67UOg1kJe1P6dLUOg1gJi1T6`LTOh1eJk1V6^LUOh1aJm1Z6[LUOh1^JP2[2RL_M=f2JQ3i1ZJR2]2PM5UOT3k1TJT2_2SM6nNW3l1oIV2b2UM5kNX3Y6`LRK5eN[3Y6]LXK5_N^3Y6ZL^K4ZNb3X6WLcK5UNc3Y6ULhK5oMf3Y6RLnK4jMj3X6oKSL5eMk3X6nKXL4aMn3W6lK\\L4]MP4U6kKcL3XMQ4U6jKhL2TMT4R6iKoL0PMX4o5gKUMOlL[4n5dKZMNiL_4k5bK`MLfLc4j5_KcMKdLg4j5ZKhMJ_Lm4h5WKPNFYLT5g5SKWNCSL[5e5PK`N^OlKc5b5nJiNYOfKj5^5mJoNWOPLa5l4ZKWOROoKe5k4UKYOTOmKh5k4PK[OUOlKl5j4kJ\\OXOkKn5i4gJm0j5cNTJ^1`6oM]IS2e6lMYIV2g6hMYIY2g6gMWI[2j6dMUI]2k6fMQI[2P7fMlH\\2T7dMhH`2Z7R51O1O1O1N2O1O1O1O1N2WKYHWMh7b2cHYM^7`2mH[MT7_2VI[Ml6^2_I]Mb6\\2iI_MY6Y2RJbMo5X2ZJdMg5U2dJeM^5T2mJgMT5;lGRNY3`1l4;RHkMZ3g1e4;dHWMP3[2]4:YMBi2;\\MBe2;`MBb29dMDb24bMJd2NaMNf2J_M3g2D_M9g2_O^M>h2ZO]Mb0j2UO\\Mh0Q3iNTMT1o:M3M2M4M3M3M1O2M2O1O2N1O2M2O1O2N1N3N1M4K4K5LahX1"}, "label": "a man han in glass"}]}
{"id": 470976, "image": "COCO_train2014_000000470976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the backs of two chairs with a persons pants and feet standing behind them\"."}], "task": "refering", "answer_template": "The \"the backs of two chairs with a persons pants and feet standing behind them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 198, 199, 215, 231, 232, 248, 265, 282, 299, 316, 333], "bbox": [0.5846153846153845, 0.432546875, 0.72993762993763, 0.8531875], "iscrowd": 0, "area": 4474.8451000000005, "rle": {"size": [640, 481], "counts": "PS`55Xc0b0^Oc0E:N2N2N2N3M2N2N:WLhMZEb2\\:`MaEk2U:WMhET3m9oLPF\\3f9fLWFe3_9\\L_Fn3Z9RLcFY4T9hKiFa4P9_KnFd4R9\\KkFg4T9YKjFe2eMkNa;`NgF`2ZNfNn:jNfF[2TO[NV:ZOcFV2OPN]9KaFQ2h0eMg89`Fl1k<UNRCf1U=ZNiBa1_=^N_B^1g=cNVBX1R>gNlAT1\\>kNbAP1f>oNXAl0P?TOo@g0W?XOi@g0Y?XOf@h0\\?WOd@h0^?XO`@g0c?YO\\@f0f?ZOY@e0i?\\OT@d0n?\\OQ@c0Q`0^Ol_Ob0V`0^Oi_O`0Z`0Ad_O>^`0B`_O>b`0C\\_O<f`0DX_O<j`0ET_O9o`0Ho^O7Sa0Jj^O6Ya0Je^O5]a0L`^O4ca0L[^O2ha0OV^O0ma01o]OOSb02k]OMXb03e]OM]b04a]OKbb05\\]OIgb0f03M3M3M3L5L3M3M4JPo`2"}, "label": "the backs of two chairs with a persons pants and feet standing behind them"}]}
{"id": 470976, "image": "COCO_train2014_000000470976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair that the man ' s hand is resting on\"."}], "task": "refering", "answer_template": "The \"the chair that the man ' s hand is resting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 198, 199, 215, 231, 232, 248, 265, 282, 299, 316, 333], "bbox": [0.5846153846153845, 0.432546875, 0.72993762993763, 0.8531875], "iscrowd": 0, "area": 4474.8451000000005, "rle": {"size": [640, 481], "counts": "PS`55Xc0b0^Oc0E:N2N2N2N3M2N2N:WLhMZEb2\\:`MaEk2U:WMhET3m9oLPF\\3f9fLWFe3_9\\L_Fn3Z9RLcFY4T9hKiFa4P9_KnFd4R9\\KkFg4T9YKjFe2eMkNa;`NgF`2ZNfNn:jNfF[2TO[NV:ZOcFV2OPN]9KaFQ2h0eMg89`Fl1k<UNRCf1U=ZNiBa1_=^N_B^1g=cNVBX1R>gNlAT1\\>kNbAP1f>oNXAl0P?TOo@g0W?XOi@g0Y?XOf@h0\\?WOd@h0^?XO`@g0c?YO\\@f0f?ZOY@e0i?\\OT@d0n?\\OQ@c0Q`0^Ol_Ob0V`0^Oi_O`0Z`0Ad_O>^`0B`_O>b`0C\\_O<f`0DX_O<j`0ET_O9o`0Ho^O7Sa0Jj^O6Ya0Je^O5]a0L`^O4ca0L[^O2ha0OV^O0ma01o]OOSb02k]OMXb03e]OM]b04a]OKbb05\\]OIgb0f03M3M3M3L5L3M3M4JPo`2"}, "label": "the chair that the man ' s hand is resting on"}]}
{"id": 479168, "image": "COCO_train2014_000000479168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a black shirt and striped pants standing directly behind the batter\"."}], "task": "refering", "answer_template": "The \"a guy in a black shirt and striped pants standing directly behind the batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 94, 95, 96, 109, 110, 111, 124, 125, 126, 139, 140, 154, 155, 169, 170, 184, 185, 199, 201, 214, 215, 216, 217, 229, 230, 231, 232, 243, 244, 245, 246, 247, 258, 259, 260, 261, 273, 274, 275, 276, 287, 288, 289, 290, 291], "bbox": [0.16662763466042155, 0.22384374999999998, 0.4800234192037471, 0.847640625], "iscrowd": 0, "area": 18473.387600000002, "rle": {"size": [640, 427], "counts": "fl\\11oc01N3N1N2O0O10000O100O100O100O100O1O100O1O1O100O1O100O1G9O1O1O1O1O1O1O1oNmNT_OT1g`0^Ok^Oc0o`0Z1K500O1WMZMWDg2R8iMYHEDb2P8]NfGWO8]2P8iN[GlNd0\\2n7VOQGaNo0Z2n7AgFXNZ1V2n7L^FQNb1T2o74UFjM6ZOj0i2i8>lEbM7@T1`2h8c1kETL]1Z2g8Y2ZGhMd8Y2]GgMb8Y2^GhMa8X2_GiM_8W2cGiM\\8W2dGjM[8V2fGjMc6RMYIS55lM`6TMYIP53QNb6RMYIm41VNd6PMYIi4O]Ne6mL\\Ie4N`Nd6nL\\Ib4NbNf6^MkHn3`0`Ni6dMeHl3b0]Nl6iMaHi3c0QNY7XNSHe3e0_Ml7nN]Gc3g0lL_8BjFa3g0_Lm82ZF^3k0SLV9b0nEZ3l0mK^9j0dEY3o0hKb9o0_EW3P1fKf9T1YEc0J3X1QNi9h3\\EPNk0UNm9g3aEmMd0[Nl9g3gEgM=cNl9e3mEbM8hNk9f3RF\\M4oNj9d3WFXM0SOk9d3XFUMMXOT2Lj3f3ZJQMH]On13P4^3]JPME_Ol15S4Z3_JoLBBk17T4W3bJmL@Eh19W4S3cJmL^OGf1<Y4P3eJjL]OJc1>[4m2hJhLZOMb1`0]4i2jJhLWOO`1b0_4f2lJgLVO0[1h0c4`2oJeLSO4W1l0g4Y2RKdLQO6cN6f1l0d5T2kLkLaKc0h1b0l5o1lLmMoL9V6h1lLSNfL9^6c1lLZN\\L8i6\\1lL`NPL;T7T1lLgNbK>c7i0lLZ1T3cNoL]1R3dLWHi0h4d2Q3aL[Hf0g4i2o2^L^Hd0e4o2P3XL_Hc0e4U3n2RLcHc0b4\\3n2jKfHd0_4c3R4VLQLk3Q4oKSLQ4P4hKSLY4i91O10O0100O10O01O100OO2UOj0M4K4M3M4K5L4L5hMZ_Of0Ta0TOV_O;Xa0^OR_O0\\a0J^]Z4"}, "label": "a guy in a black shirt and striped pants standing directly behind the batter"}]}
{"id": 479168, "image": "COCO_train2014_000000479168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in glasses waiting to bat\"."}], "task": "refering", "answer_template": "The \"a man in glasses waiting to bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 94, 95, 96, 109, 110, 111, 124, 125, 126, 139, 140, 154, 155, 169, 170, 184, 185, 199, 201, 214, 215, 216, 217, 229, 230, 231, 232, 243, 244, 245, 246, 247, 258, 259, 260, 261, 273, 274, 275, 276, 287, 288, 289, 290, 291], "bbox": [0.16662763466042155, 0.22384374999999998, 0.4800234192037471, 0.847640625], "iscrowd": 0, "area": 18473.387600000002, "rle": {"size": [640, 427], "counts": "fl\\11oc01N3N1N2O0O10000O100O100O100O100O1O100O1O1O100O1O100O1G9O1O1O1O1O1O1O1oNmNT_OT1g`0^Ok^Oc0o`0Z1K500O1WMZMWDg2R8iMYHEDb2P8]NfGWO8]2P8iN[GlNd0\\2n7VOQGaNo0Z2n7AgFXNZ1V2n7L^FQNb1T2o74UFjM6ZOj0i2i8>lEbM7@T1`2h8c1kETL]1Z2g8Y2ZGhMd8Y2]GgMb8Y2^GhMa8X2_GiM_8W2cGiM\\8W2dGjM[8V2fGjMc6RMYIS55lM`6TMYIP53QNb6RMYIm41VNd6PMYIi4O]Ne6mL\\Ie4N`Nd6nL\\Ib4NbNf6^MkHn3`0`Ni6dMeHl3b0]Nl6iMaHi3c0QNY7XNSHe3e0_Ml7nN]Gc3g0lL_8BjFa3g0_Lm82ZF^3k0SLV9b0nEZ3l0mK^9j0dEY3o0hKb9o0_EW3P1fKf9T1YEc0J3X1QNi9h3\\EPNk0UNm9g3aEmMd0[Nl9g3gEgM=cNl9e3mEbM8hNk9f3RF\\M4oNj9d3WFXM0SOk9d3XFUMMXOT2Lj3f3ZJQMH]On13P4^3]JPME_Ol15S4Z3_JoLBBk17T4W3bJmL@Eh19W4S3cJmL^OGf1<Y4P3eJjL]OJc1>[4m2hJhLZOMb1`0]4i2jJhLWOO`1b0_4f2lJgLVO0[1h0c4`2oJeLSO4W1l0g4Y2RKdLQO6cN6f1l0d5T2kLkLaKc0h1b0l5o1lLmMoL9V6h1lLSNfL9^6c1lLZN\\L8i6\\1lL`NPL;T7T1lLgNbK>c7i0lLZ1T3cNoL]1R3dLWHi0h4d2Q3aL[Hf0g4i2o2^L^Hd0e4o2P3XL_Hc0e4U3n2RLcHc0b4\\3n2jKfHd0_4c3R4VLQLk3Q4oKSLQ4P4hKSLY4i91O10O0100O10O01O100OO2UOj0M4K4M3M4K5L4L5hMZ_Of0Ta0TOV_O;Xa0^OR_O0\\a0J^]Z4"}, "label": "a man in glasses waiting to bat"}]}
{"id": 479168, "image": "COCO_train2014_000000479168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a baseball bat\"."}], "task": "refering", "answer_template": "The \"a man holding a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 36, 37, 38, 39, 40, 52, 53, 54, 55, 66, 67, 68, 69, 70, 71, 72, 81, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 218, 219, 220, 227, 228, 229, 233, 234, 235, 242, 243, 244, 249, 250, 251, 257, 258, 264, 265, 266, 279, 280, 281, 295, 296, 310, 311, 325, 326, 340, 341], "bbox": [0.12918032786885245, 0.048265625, 0.9329508196721311, 0.98028125], "iscrowd": 0, "area": 66083.07294999996, "rle": {"size": [640, 427], "counts": "njR11lc05K5K4L5K4L3M4L3L5L3O101N1O2N1O2N1O2N1O001O001O0010O01O0010O01O00000O1O11O000000000O1O101N100O2O0O100O2O001O000O1000O1000000O101O1O1O2M2O1O1O2M2O1ODT^OgNja0Y1X^OfNfa0Z1^^OdNaa0\\1b^ObN]a0]1g^ObNWa0T1V_OjNh`0X1Y_OgNf`0Y1]_OfNa`0[1Q11N101O1N1KbNh]O`1Wb04jMcNjA]1n=QOlAo0S>WOhAi0W>^OdAa0Z>FaA:^>J_A6`>N]A2b>2[ANd>6ZAJe>9XAGh><UADk>>TABk>a0RA_On>d0o@\\OQ?h0l@WOT?k0j@VOT?n0i@ROW?P1h@POV?S1h@nNW?m2O2N2N2O1O1O001O1M3L4M3M2M4M3L3N3J6D;jNW1C=N101N1O2N1O2O0mMnGXHS8i7mGUHT8k7mGSHT8m7mGRHS8n7oGoGR8R8nGmGS8R8oGmGP8T8PHkGP8U8RHiGn7X8RHgGo7X8SHfGo7Y8QHgGo7X8SHfGo7Y8RHeGo7Z8RHfGn7Z8SHbFHa0V8k8oHjFR7T9[I_Fg6_9n1K4M4O0000001O00002N2N3N2N1O2N2N1O2N1N2O2N1O2N1O1O2N2N1N3N1O2N1O2N1O2N2N2N3N2M3M3M3M2O1N1O1O2N1O3N1N2N=D3L3M2O1N1O100O1O100O3M3N2M2O1OiIZJUNd5cK]JX462\\5cKaJX4:3S5cKeJW4>6j4aKjJV4c0:^4]KRKU4h0>R4ZKXKT4n0a0g3YK^KQ4R1f0^3UKcKP4X1j0S3SKgKo3]1n0n2lJgKR4b1R1i2ZJRL`4]1U1P4hNULW1i3hN\\LV1b3hNcLX1[3eNlLX1R3gNSMW1l2fN[MW1c2gNcMX1[2fNkM[1PMUKT3^33]1aLYK\\3W3;_1QL]Kd3S3a0n2_OPMh0j2XOVMn0e2QOZMU1a2kN_M[1\\2eNcMa1W2_NiMf1S2YNmMm1m1SNRNR2k1mMUNU2i1lMVNU2i1kMWNW2g1iMYNX2g1gMXN[2g1fMXN\\2f1dMZN]2e1cM[N_2c1aM\\Na2c1`M\\Nb2b1^M^Nc2a1]M^Nf2`1[M_Nf2`1[M^Nj2^1WMaNP3X1PMgNY3RMPIc094_3V2_3kLmJn0e1V2d3eLiJT1d1U2j3`LdJZ1c1U2b6jM^IU2e6iM\\IV2e6iM[IV2h6hMYIW2j6fMVIY2m6eMSI[2P7aMRI]2Q7aMoH_2T7^MmH`2W7]MiHc2Y7[MhHc2\\7ZMdHf2_7WMaHh2c7UM^Hj2\\<O100O2O0O101O00000000000000000O2O1O001O1O001O1O1O001O1O1O1O2M2O1O2M2O3L4M4L2N2N2N2N5K4L2O1N1O1O1O1N2O1N2N5L7F9Gema0"}, "label": "a man holding a baseball bat"}]}
{"id": 479168, "image": "COCO_train2014_000000479168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a uniformed young man preparing to hit a baseball\"."}], "task": "refering", "answer_template": "The \"a uniformed young man preparing to hit a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 36, 37, 38, 39, 40, 52, 53, 54, 55, 66, 67, 68, 69, 70, 71, 72, 81, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 218, 219, 220, 227, 228, 229, 233, 234, 235, 242, 243, 244, 249, 250, 251, 257, 258, 264, 265, 266, 279, 280, 281, 295, 296, 310, 311, 325, 326, 340, 341], "bbox": [0.12918032786885245, 0.048265625, 0.9329508196721311, 0.98028125], "iscrowd": 0, "area": 66083.07294999996, "rle": {"size": [640, 427], "counts": "njR11lc05K5K4L5K4L3M4L3L5L3O101N1O2N1O2N1O2N1O001O001O0010O01O0010O01O00000O1O11O000000000O1O101N100O2O0O100O2O001O000O1000O1000000O101O1O1O2M2O1O1O2M2O1ODT^OgNja0Y1X^OfNfa0Z1^^OdNaa0\\1b^ObN]a0]1g^ObNWa0T1V_OjNh`0X1Y_OgNf`0Y1]_OfNa`0[1Q11N101O1N1KbNh]O`1Wb04jMcNjA]1n=QOlAo0S>WOhAi0W>^OdAa0Z>FaA:^>J_A6`>N]A2b>2[ANd>6ZAJe>9XAGh><UADk>>TABk>a0RA_On>d0o@\\OQ?h0l@WOT?k0j@VOT?n0i@ROW?P1h@POV?S1h@nNW?m2O2N2N2O1O1O001O1M3L4M3M2M4M3L3N3J6D;jNW1C=N101N1O2N1O2O0mMnGXHS8i7mGUHT8k7mGSHT8m7mGRHS8n7oGoGR8R8nGmGS8R8oGmGP8T8PHkGP8U8RHiGn7X8RHgGo7X8SHfGo7Y8QHgGo7X8SHfGo7Y8RHeGo7Z8RHfGn7Z8SHbFHa0V8k8oHjFR7T9[I_Fg6_9n1K4M4O0000001O00002N2N3N2N1O2N2N1O2N1N2O2N1O2N1O1O2N2N1N3N1O2N1O2N1O2N2N2N3N2M3M3M3M2O1N1O1O2N1O3N1N2N=D3L3M2O1N1O100O1O100O3M3N2M2O1OiIZJUNd5cK]JX462\\5cKaJX4:3S5cKeJW4>6j4aKjJV4c0:^4]KRKU4h0>R4ZKXKT4n0a0g3YK^KQ4R1f0^3UKcKP4X1j0S3SKgKo3]1n0n2lJgKR4b1R1i2ZJRL`4]1U1P4hNULW1i3hN\\LV1b3hNcLX1[3eNlLX1R3gNSMW1l2fN[MW1c2gNcMX1[2fNkM[1PMUKT3^33]1aLYK\\3W3;_1QL]Kd3S3a0n2_OPMh0j2XOVMn0e2QOZMU1a2kN_M[1\\2eNcMa1W2_NiMf1S2YNmMm1m1SNRNR2k1mMUNU2i1lMVNU2i1kMWNW2g1iMYNX2g1gMXN[2g1fMXN\\2f1dMZN]2e1cM[N_2c1aM\\Na2c1`M\\Nb2b1^M^Nc2a1]M^Nf2`1[M_Nf2`1[M^Nj2^1WMaNP3X1PMgNY3RMPIc094_3V2_3kLmJn0e1V2d3eLiJT1d1U2j3`LdJZ1c1U2b6jM^IU2e6iM\\IV2e6iM[IV2h6hMYIW2j6fMVIY2m6eMSI[2P7aMRI]2Q7aMoH_2T7^MmH`2W7]MiHc2Y7[MhHc2\\7ZMdHf2_7WMaHh2c7UM^Hj2\\<O100O2O0O101O00000000000000000O2O1O001O1O001O1O1O001O1O1O1O2M2O1O2M2O3L4M4L2N2N2N2N5K4L2O1N1O1O1O1N2O1N2N5L7F9Gema0"}, "label": "a uniformed young man preparing to hit a baseball"}]}
{"id": 225213, "image": "COCO_train2014_000000225213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on sand\"."}], "task": "refering", "answer_template": "The \"zebra on sand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 122, 123, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 154, 156, 157, 171, 172, 173, 176, 177, 180, 195, 196, 199, 203, 218], "bbox": [0.323796875, 0.07166276346604215, 0.85996875, 0.6134426229508196], "iscrowd": 0, "area": 36456.11095, "rle": {"size": [427, 640], "counts": "n[f21W=5J6K5M2cEDj7=nG0k72mG8P8HjG?T8CdGd0\\8\\O]Gk0c8VOUGQ1l8oNlFX1T9hNeF`1[9`N]Fg1d9XNUFo1k9h02N1O1O1O1O00O1O1ON2N2M3N2N2N2N2N2M3N2O11O0O2O001O001N1010O010O010O010O1D;2O0O20O01O1O001O0010O01O001O001O0010O01O001O001O010O001O001O010O010O2O001N101N101N1oFRMd7n2YH[M`7g2\\HaM_7_2^HjM[7X2aHPNZ7P2bHZNW7h1eH`NV7`1gHiNR7Y1jHoNP7R1mHZMJo0X7h1kHYM4k0P7f2VIWMi6f2^IVMGgN\\6S4UJoL]OoN^6Q4eJ`LmN@^6P4fJ^LkND_6n3gJ\\LjNG_6l3iJ[LgNK`6m3gJVLiNN`6Q4cJoKmN1a6T4^JiKPO5b6W4[JbKRO9b6[4WJZKVO=c6^4TJRKYOa0c6b4SJ`Km5e4mI\\KS6j4fIWKZ6n4_ITKa6Q5YIPKh6i5000C^IQJb6m5eInI[6`3[IlNU7S1PIgNR7Y1QIcNP7\\1[IYNg6f1_2O100O1O100O100O2N100O010O100O00100O100O01000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O100000O01000000O100000000O1000000O1000000O100000000O01O10nHbLW4]3iKdLV4\\3iKfLW4Y3iKgLX4X3hKiLX4V3gKlLX4U3gKlLY4S3fKoLZ4P3fKQMZ4n2eKTMZ4l2fKUMZ4k2eKVM[4i2dKYM\\4f2dKZM]4e2bK]M]4c2cK^M]4a2bKaM^4_2aKbM_4]2`KeM_4[2aKfM_4Y2aK^Mi4a2VK`Mk4`2TKaMk4_2TKcMl4\\2TKdMm4[2RKgMn4X2RKiMn4V2RKjMo4W2nJiMT5Z2hJeMZ5^2aJbMa5b2YJ^Mi5e2oI^MS6e2fI]M\\6g2[I\\Mh6f2QI\\MQ7g2gH\\M[7g2^H[Md7i2SHZMo7o32N2N2N3M2N2N2N2N2N2N2N3M2N2N3M2M3L5K4L4L5J5O1010O1O010O1O010O00100O00100O00100O00100O0010iNXFjNh9S1\\FkNe9R1^FnNc9m0aFRO`9k0cFUO^9g0eFYO\\9b0hF]OY9`0jF@W9<lFDT98QGFQ96RGJo87oFIQ9;kFDZ9<bFDc9;YFEl9:PFEU:;gEG\\:3eE0\\:JgE7\\:CeE`0]:YOeEi0]:POfER1Y;3M2N3N101O0PE]NU:c1aEcNc:^1SEgNQ;g13N3L4L4M3L4L4M3L4L4L4ZOfC5\\<FiC7h<M4M2MRZU1"}, "label": "zebra on sand"}]}
{"id": 225213, "image": "COCO_train2014_000000225213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra is near a hippopotamus\"."}], "task": "refering", "answer_template": "The \"a zebra is near a hippopotamus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 122, 123, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 154, 156, 157, 171, 172, 173, 176, 177, 180, 195, 196, 199, 203, 218], "bbox": [0.323796875, 0.07166276346604215, 0.85996875, 0.6134426229508196], "iscrowd": 0, "area": 36456.11095, "rle": {"size": [427, 640], "counts": "n[f21W=5J6K5M2cEDj7=nG0k72mG8P8HjG?T8CdGd0\\8\\O]Gk0c8VOUGQ1l8oNlFX1T9hNeF`1[9`N]Fg1d9XNUFo1k9h02N1O1O1O1O00O1O1ON2N2M3N2N2N2N2N2M3N2O11O0O2O001O001N1010O010O010O010O1D;2O0O20O01O1O001O0010O01O001O001O0010O01O001O001O010O001O001O010O010O2O001N101N101N1oFRMd7n2YH[M`7g2\\HaM_7_2^HjM[7X2aHPNZ7P2bHZNW7h1eH`NV7`1gHiNR7Y1jHoNP7R1mHZMJo0X7h1kHYM4k0P7f2VIWMi6f2^IVMGgN\\6S4UJoL]OoN^6Q4eJ`LmN@^6P4fJ^LkND_6n3gJ\\LjNG_6l3iJ[LgNK`6m3gJVLiNN`6Q4cJoKmN1a6T4^JiKPO5b6W4[JbKRO9b6[4WJZKVO=c6^4TJRKYOa0c6b4SJ`Km5e4mI\\KS6j4fIWKZ6n4_ITKa6Q5YIPKh6i5000C^IQJb6m5eInI[6`3[IlNU7S1PIgNR7Y1QIcNP7\\1[IYNg6f1_2O100O1O100O100O2N100O010O100O00100O100O01000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O100000O01000000O100000000O1000000O1000000O100000000O01O10nHbLW4]3iKdLV4\\3iKfLW4Y3iKgLX4X3hKiLX4V3gKlLX4U3gKlLY4S3fKoLZ4P3fKQMZ4n2eKTMZ4l2fKUMZ4k2eKVM[4i2dKYM\\4f2dKZM]4e2bK]M]4c2cK^M]4a2bKaM^4_2aKbM_4]2`KeM_4[2aKfM_4Y2aK^Mi4a2VK`Mk4`2TKaMk4_2TKcMl4\\2TKdMm4[2RKgMn4X2RKiMn4V2RKjMo4W2nJiMT5Z2hJeMZ5^2aJbMa5b2YJ^Mi5e2oI^MS6e2fI]M\\6g2[I\\Mh6f2QI\\MQ7g2gH\\M[7g2^H[Md7i2SHZMo7o32N2N2N3M2N2N2N2N2N2N2N3M2N2N3M2M3L5K4L4L5J5O1010O1O010O1O010O00100O00100O00100O00100O0010iNXFjNh9S1\\FkNe9R1^FnNc9m0aFRO`9k0cFUO^9g0eFYO\\9b0hF]OY9`0jF@W9<lFDT98QGFQ96RGJo87oFIQ9;kFDZ9<bFDc9;YFEl9:PFEU:;gEG\\:3eE0\\:JgE7\\:CeE`0]:YOeEi0]:POfER1Y;3M2N3N101O0PE]NU:c1aEcNc:^1SEgNQ;g13N3L4L4M3L4L4M3L4L4L4ZOfC5\\<FiC7h<M4M2MRZU1"}, "label": "a zebra is near a hippopotamus"}]}
{"id": 225213, "image": "COCO_train2014_000000225213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one zebra faces away from the camera and is nearly obscured by the other zebra\"."}], "task": "refering", "answer_template": "The \"one zebra faces away from the camera and is nearly obscured by the other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 38, 39, 40, 41, 64, 107, 108, 130, 131, 132, 153, 154, 155, 156], "bbox": [0.6671875, 0.0002107728337236534, 0.8224218750000001, 0.472248243559719], "iscrowd": 0, "area": 4474.146650000002, "rle": {"size": [427, 640], "counts": "U`b51Z=0O10OO2O1N2N2O1N1O2UFh1b4YN]Kg1b4[N^Ke1`4]N`Kb1`4^N`Kc1_4^NaKb1^4_NbK`1_4`NaK`1_4`N`K`1a4`N_K=VMK[7H_K3fMNl6O]KJWN1[66XMEh2;^M^Oc2b0S500000O2O0000000O10000000001N1SHCl2=RMFm2:QMIn27PMLo24oLNQ32nL0k2XOmHi0V42l2VOmHh0U45m2TOmHg0T48n2ROmHf0W46l20VMMj23YMIh27\\MDe2<^M@c2`0`M\\Oa2d0cMXO\\2i0gMSOZ2m0R500001O001O001O0[HQOn2P1hL[OV3e0aLE^3<XLOf31XL2g3OYL2e3N\\L2c3O]L2a3NbLO^32dLK\\35hLFY3;iLBW3>mL^OR3d0PMYOP3g0TMUOl2l0VMQOj2o0ZMmNf2T1\\MiNd2X1_McNb2`1c44L3M3M3O20O1O10O01_NnDd0S;oN\\Eo0c;H9H7HR__1"}, "label": "one zebra faces away from the camera and is nearly obscured by the other zebra"}]}
{"id": 225213, "image": "COCO_train2014_000000225213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra facing away from camera\"."}], "task": "refering", "answer_template": "The \"a zebra facing away from camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 38, 39, 40, 41, 64, 107, 108, 130, 131, 132, 153, 154, 155, 156], "bbox": [0.6671875, 0.0002107728337236534, 0.8224218750000001, 0.472248243559719], "iscrowd": 0, "area": 4474.146650000002, "rle": {"size": [427, 640], "counts": "U`b51Z=0O10OO2O1N2N2O1N1O2UFh1b4YN]Kg1b4[N^Ke1`4]N`Kb1`4^N`Kc1_4^NaKb1^4_NbK`1_4`NaK`1_4`N`K`1a4`N_K=VMK[7H_K3fMNl6O]KJWN1[66XMEh2;^M^Oc2b0S500000O2O0000000O10000000001N1SHCl2=RMFm2:QMIn27PMLo24oLNQ32nL0k2XOmHi0V42l2VOmHh0U45m2TOmHg0T48n2ROmHf0W46l20VMMj23YMIh27\\MDe2<^M@c2`0`M\\Oa2d0cMXO\\2i0gMSOZ2m0R500001O001O001O0[HQOn2P1hL[OV3e0aLE^3<XLOf31XL2g3OYL2e3N\\L2c3O]L2a3NbLO^32dLK\\35hLFY3;iLBW3>mL^OR3d0PMYOP3g0TMUOl2l0VMQOj2o0ZMmNf2T1\\MiNd2X1_McNb2`1c44L3M3M3O20O1O10O01_NnDd0S;oN\\Eo0c;H9H7HR__1"}, "label": "a zebra facing away from camera"}]}
{"id": 12224, "image": "COCO_train2014_000000012224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with a cigarette smiling while man next to her is having trouble with an umbrella\"."}], "task": "refering", "answer_template": "The \"woman with a cigarette smiling while man next to her is having trouble with an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 297, 298, 299, 300, 301, 302, 303, 304, 305, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 330, 332, 336, 337, 338, 339, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 386, 387, 388, 389, 390], "bbox": [0.4404583333333333, 0.137078125, 1.0, 0.986515625], "iscrowd": 0, "area": 71659.50514999997, "rle": {"size": [640, 480], "counts": "X\\T41oc01N3M2N2N3M2O2M2N3M2N1O00100O1O00000001O00000O10001O00000002N1O101N1O2N1O1O2O0O2N1O1O2N1O101N1O2N1O1O2O0O1O1O2N10001N10001O0O101O000O10000O1000O10O1000000O0100O1O010O1O100O00100O1O10O01O100O12N2M2O2N1O0000001O0000001O000O101O0000001O0000O100000000O1000000000kJXNRHg1c7jNXHV1\\7[O_He0U7KgH5j6?QIAY6Z1bIeNb5\\2ZJdMa5f2ZJZMb5o2jFQKY3P2h5Y3mFjJU3m1i5b3QGdJP3j1i5W4QJhKj5d4PJ\\Kj5n4RJRKh5V5VJjJd5^5ZJbJ`5f5^JYJ]5o5aJQJY5X6dJhIV5`6hJ`IT5f6jJZIU5j6hJVIT1SNl0l8mMPIf0fN[1^8mMlH6XOl1P8kMhHHI[2d7jMcHYO;k2V7jM_HiNm0[3i6iMaKV2c4gM]KW2g4gMYKX2k4dMVK\\2l4bMTK]2o4aMQK_2Q5_MoJ`2S5^MnJb2T5\\MmJc2T5YMoJf2T5TMPKl2R5nLRKR3o4gLWKX3i:00O1O100O1O100O100O1O100O1O10000000000000001O0000000O10000000000000000000000000000001O0000001O01O01O0000001O00001O0000001O0000001O001O1O1O001O1O1O1O010O1O1O1O1O001O1O1O1O001O1]CZLj8g3RG^Ll8c3QGaLm8b3mFdLP9_3jFfLT9^3fFfLX9^3bFfL\\9^3^FUNQ8P2hGR1X5SOaJo3_2VLZMl6F`D"}, "label": "woman with a cigarette smiling while man next to her is having trouble with an umbrella"}]}
{"id": 12224, "image": "COCO_train2014_000000012224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman with a cigarette\"."}], "task": "refering", "answer_template": "The \"a young woman with a cigarette\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 297, 298, 299, 300, 301, 302, 303, 304, 305, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 330, 332, 336, 337, 338, 339, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 386, 387, 388, 389, 390], "bbox": [0.4404583333333333, 0.137078125, 1.0, 0.986515625], "iscrowd": 0, "area": 71659.50514999997, "rle": {"size": [640, 480], "counts": "X\\T41oc01N3M2N2N3M2O2M2N3M2N1O00100O1O00000001O00000O10001O00000002N1O101N1O2N1O1O2O0O2N1O1O2N1O101N1O2N1O1O2O0O1O1O2N10001N10001O0O101O000O10000O1000O10O1000000O0100O1O010O1O100O00100O1O10O01O100O12N2M2O2N1O0000001O0000001O000O101O0000001O0000O100000000O1000000000kJXNRHg1c7jNXHV1\\7[O_He0U7KgH5j6?QIAY6Z1bIeNb5\\2ZJdMa5f2ZJZMb5o2jFQKY3P2h5Y3mFjJU3m1i5b3QGdJP3j1i5W4QJhKj5d4PJ\\Kj5n4RJRKh5V5VJjJd5^5ZJbJ`5f5^JYJ]5o5aJQJY5X6dJhIV5`6hJ`IT5f6jJZIU5j6hJVIT1SNl0l8mMPIf0fN[1^8mMlH6XOl1P8kMhHHI[2d7jMcHYO;k2V7jM_HiNm0[3i6iMaKV2c4gM]KW2g4gMYKX2k4dMVK\\2l4bMTK]2o4aMQK_2Q5_MoJ`2S5^MnJb2T5\\MmJc2T5YMoJf2T5TMPKl2R5nLRKR3o4gLWKX3i:00O1O100O1O100O100O1O100O1O10000000000000001O0000000O10000000000000000000000000000001O0000001O01O01O0000001O00001O0000001O0000001O001O1O1O001O1O1O1O010O1O1O1O1O001O1O1O1O001O1]CZLj8g3RG^Ll8c3QGaLm8b3mFdLP9_3jFfLT9^3fFfLX9^3bFfL\\9^3^FUNQ8P2hGR1X5SOaJo3_2VLZMl6F`D"}, "label": "a young woman with a cigarette"}]}
{"id": 12224, "image": "COCO_train2014_000000012224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing in white tee shirt next to woman smoking a cig\"."}], "task": "refering", "answer_template": "The \"man standing in white tee shirt next to woman smoking a cig\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 106, 107, 110, 111, 112, 113, 123, 124, 125, 127, 128, 129, 130, 141, 142, 144, 145, 146, 147, 158, 159, 161, 162, 163, 164, 165, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 227, 229, 230, 231, 232, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 279, 280, 281, 282, 296, 297, 298, 313, 314, 316, 317, 330, 331, 332, 333, 334, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 381, 382, 383, 384], "bbox": [0.2846458333333333, 0.240453125, 0.7250833333333333, 0.9887656249999999], "iscrowd": 0, "area": 54188.516050000006, "rle": {"size": [640, 480], "counts": "hie2g0ob0e0C<D<D?Ac0]Oc0]O;D8I8J5J6K5K8H7I7H2O001O001N101O02O4K4M4K4M4K5L4K6K5M3N3M2K2M100O10000O100O100O100O4M4K4M3K6Kh0WO`0A4L0O2O0OnAUNQ:k1kE_No9`1oEhNl9X1QFQOi9n0TF\\Of9d0WFEd99ZF4\\9L`Fb0V9]OhFP1n8POoF\\1h8cNVGi1a8WN\\GU2\\8iMaGd2V8\\MgGo2Q8PMmGZ3l7fLQHg0bM;U:mNVHl0fM>k9gN\\Hm0kM>f9eN\\HP1oM:e9eNZHS1TN6b9gNWHV1XN2`9hNUHY1]NN^9iNRH\\1aNJ]9jNoG_1eNF\\9kNkGb1jNCZ9lNhGe1mN@[9kNdGh1RO]OZ9kNaGk1TO[O[9jN]Gn1YOXOY9kNZGQ2\\OVOY9jNVGS2BSOj7RMQIh1_OW2EPOe7\\MRI]1AY2InN]7fMUIT1A\\2LkNX7nMXIk0@_21hNS7UNYId0_Oc24eNP7^NVIT5J_Lm6dNUIm4N_Li6iNXIh4N`Lf6mN[Ib40aLb6QO\\I_41aL`6UO]IS4:hLW6XO^Io3;jLT6[O`Ij3=lLP6]OaIg3?mLm5AbIb3a0mLj5EdI]3b0oLh5GeIY3d0PMf5U6ZJlIe5T6]JkIb5U6_JkI_5V6cJiI\\5W6eJiIZ5W6gJiIX5W6jJhIT5Y6mJgIR5Z6oJfIo4Z6RKfIn4Y6TKfIl4Y6UKgIk4X6WKgIi4X6XKhIh4W6ZKhIf4W6[KiIe4V6]KiIc4V6^KjIb4U6_KkI`4V6aKiI_4V6cKiI]4V6dKjI\\4U6fKkIY4U6gKkIY4T6iKkIW4U6jKjIV4V6jKjIV4U6kKkIU4U6jKlIV4S6kKmIU4S6kKmIU4S6kKmIU4R6kKoIU4Q6kKoIU4P6lKPJT4P6kKRJT4m5mKSJS4m5mKSJS4m5lKTJT4k5mKUJS4k5mKUJS4j5mKWJS4i5mKWJS4i5mKWJS4h5mKYJS4g5mKYJS4f5nKZJR4f5mK[JS4e5mK[JS4d5nK]JQ4c5nK^JR4a5oK_JQ4a5oK_JQ4a5nK`JR4_5oKaJQ4_5oKaJQ4^5oKcJQ4]5oKcJQ4]5oKcJQ4\\5PLdJP4\\5oKeJQ4Z5PLfJP4Z5PLfJQ4Y5oKhJQ4V5PLjJQ4U5oKkJR4S5PLlJQ4l4UJgGj1R3^4S5kKaJb4\\5`KYJl4d5VKQJV5l5mJhI_5T6dJaIh5\\6ZJYIR6X6\\JWIV6f6e3K4M3L4M3L4M3L5L4K7I7J6I7J5J7J6I7J6D>Aa0@`0_Oa0@>C<E:G:F9G:FiVb2"}, "label": "man standing in white tee shirt next to woman smoking a cig"}]}
{"id": 12224, "image": "COCO_train2014_000000012224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person fiddling with an umbrella\"."}], "task": "refering", "answer_template": "The \"person fiddling with an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 106, 107, 110, 111, 112, 113, 123, 124, 125, 127, 128, 129, 130, 141, 142, 144, 145, 146, 147, 158, 159, 161, 162, 163, 164, 165, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 227, 229, 230, 231, 232, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 279, 280, 281, 282, 296, 297, 298, 313, 314, 316, 317, 330, 331, 332, 333, 334, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 381, 382, 383, 384], "bbox": [0.2846458333333333, 0.240453125, 0.7250833333333333, 0.9887656249999999], "iscrowd": 0, "area": 54188.516050000006, "rle": {"size": [640, 480], "counts": "hie2g0ob0e0C<D<D?Ac0]Oc0]O;D8I8J5J6K5K8H7I7H2O001O001N101O02O4K4M4K4M4K5L4K6K5M3N3M2K2M100O10000O100O100O100O4M4K4M3K6Kh0WO`0A4L0O2O0OnAUNQ:k1kE_No9`1oEhNl9X1QFQOi9n0TF\\Of9d0WFEd99ZF4\\9L`Fb0V9]OhFP1n8POoF\\1h8cNVGi1a8WN\\GU2\\8iMaGd2V8\\MgGo2Q8PMmGZ3l7fLQHg0bM;U:mNVHl0fM>k9gN\\Hm0kM>f9eN\\HP1oM:e9eNZHS1TN6b9gNWHV1XN2`9hNUHY1]NN^9iNRH\\1aNJ]9jNoG_1eNF\\9kNkGb1jNCZ9lNhGe1mN@[9kNdGh1RO]OZ9kNaGk1TO[O[9jN]Gn1YOXOY9kNZGQ2\\OVOY9jNVGS2BSOj7RMQIh1_OW2EPOe7\\MRI]1AY2InN]7fMUIT1A\\2LkNX7nMXIk0@_21hNS7UNYId0_Oc24eNP7^NVIT5J_Lm6dNUIm4N_Li6iNXIh4N`Lf6mN[Ib40aLb6QO\\I_41aL`6UO]IS4:hLW6XO^Io3;jLT6[O`Ij3=lLP6]OaIg3?mLm5AbIb3a0mLj5EdI]3b0oLh5GeIY3d0PMf5U6ZJlIe5T6]JkIb5U6_JkI_5V6cJiI\\5W6eJiIZ5W6gJiIX5W6jJhIT5Y6mJgIR5Z6oJfIo4Z6RKfIn4Y6TKfIl4Y6UKgIk4X6WKgIi4X6XKhIh4W6ZKhIf4W6[KiIe4V6]KiIc4V6^KjIb4U6_KkI`4V6aKiI_4V6cKiI]4V6dKjI\\4U6fKkIY4U6gKkIY4T6iKkIW4U6jKjIV4V6jKjIV4U6kKkIU4U6jKlIV4S6kKmIU4S6kKmIU4S6kKmIU4R6kKoIU4Q6kKoIU4P6lKPJT4P6kKRJT4m5mKSJS4m5mKSJS4m5lKTJT4k5mKUJS4k5mKUJS4j5mKWJS4i5mKWJS4i5mKWJS4h5mKYJS4g5mKYJS4f5nKZJR4f5mK[JS4e5mK[JS4d5nK]JQ4c5nK^JR4a5oK_JQ4a5oK_JQ4a5nK`JR4_5oKaJQ4_5oKaJQ4^5oKcJQ4]5oKcJQ4]5oKcJQ4\\5PLdJP4\\5oKeJQ4Z5PLfJP4Z5PLfJQ4Y5oKhJQ4V5PLjJQ4U5oKkJR4S5PLlJQ4l4UJgGj1R3^4S5kKaJb4\\5`KYJl4d5VKQJV5l5mJhI_5T6dJaIh5\\6ZJYIR6X6\\JWIV6f6e3K4M3L4M3L4M3L5L4K7I7J6I7J5J7J6I7J6D>Aa0@`0_Oa0@>C<E:G:F9G:FiVb2"}, "label": "person fiddling with an umbrella"}]}
{"id": 118724, "image": "COCO_train2014_000000118724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe not eating grass from a pole\"."}], "task": "refering", "answer_template": "The \"the giraffe not eating grass from a pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 197, 198, 218, 219, 220, 240, 241, 242, 243, 264, 265, 266, 288, 289, 311, 312], "bbox": [0.46568750000000003, 0.48325526932084306, 0.645515625, 0.8891100702576112], "iscrowd": 0, "area": 4926.490549999997, "rle": {"size": [427, 640], "counts": "kal3:P=O101N1O2N1O110O010O0100oC0f:0WE7e:JWE9j:FPE>R;CiD?Z;@bDb0`;_OZDd0i;[OPDj0R<62M1O100O1001O100001N10[DSOn:n0oDYOk:i0nDBm:`0lDJo:^1J6K5J6K4K5L2M100O1O100O1OXOUFRNk9d1SGhMl8P2[1K6I6M301O0O100O1013L5K5K5K5K6J5KYN[Eo0`:POcER1W:oNmES1m9mNVFU1d9lN`FV10QNf8i0^GH1e0_8DQI6n6JXI2e6ObIK]65iIGT69SJBk5?YJ_Oe5`0`J]O^5c0hJYOV5f0X3VO`Di0b;YOZDf0j;\\OQDd0Q<83N3M2M3N3M2OO1O1O1O11O2O1O100O2N00100O1O1O100O1O1O010O1O1NR^n2"}, "label": "the giraffe not eating grass from a pole"}]}
{"id": 118724, "image": "COCO_train2014_000000118724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby giraffe in the middle\"."}], "task": "refering", "answer_template": "The \"the baby giraffe in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 197, 198, 218, 219, 220, 240, 241, 242, 243, 264, 265, 266, 288, 289, 311, 312], "bbox": [0.46568750000000003, 0.48325526932084306, 0.645515625, 0.8891100702576112], "iscrowd": 0, "area": 4926.490549999997, "rle": {"size": [427, 640], "counts": "kal3:P=O101N1O2N1O110O010O0100oC0f:0WE7e:JWE9j:FPE>R;CiD?Z;@bDb0`;_OZDd0i;[OPDj0R<62M1O100O1001O100001N10[DSOn:n0oDYOk:i0nDBm:`0lDJo:^1J6K5J6K4K5L2M100O1O100O1OXOUFRNk9d1SGhMl8P2[1K6I6M301O0O100O1013L5K5K5K5K6J5KYN[Eo0`:POcER1W:oNmES1m9mNVFU1d9lN`FV10QNf8i0^GH1e0_8DQI6n6JXI2e6ObIK]65iIGT69SJBk5?YJ_Oe5`0`J]O^5c0hJYOV5f0X3VO`Di0b;YOZDf0j;\\OQDd0Q<83N3M2M3N3M2OO1O1O1O11O2O1O100O2N00100O1O1O100O1O1O010O1O1NR^n2"}, "label": "the baby giraffe in the middle"}]}
{"id": 118724, "image": "COCO_train2014_000000118724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second giraffe on the left\"."}], "task": "refering", "answer_template": "The \"second giraffe on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 145, 146, 169, 192, 193, 215, 216, 217, 238, 239, 240, 261, 262, 263, 264, 284, 285, 286, 287], "bbox": [0.325953125, 0.38868852459016395, 0.493, 0.8369086651053864], "iscrowd": 0, "area": 6104.25985, "rle": {"size": [427, 640], "counts": "_Zg21Y=2N1N3N1O1N3M2N2N3O00001O00001O013L7I6J6J6K4N3TEaNd1I[5j1gHjNd1^Og5j1\\HTOd1TOQ6l1RHZOd1nN[6Y3]IiLe6`3nHbLT7m3YHSLk7e43M2N3M3M3L4M3M3M3M3M3M3M2N3lMiFNY91TG@o8>_GTOc8j0kGgNX8X1o1N2O1N2N2O1N2N10O0100O01000O01000O0100O0100O03N2N2M2O2N2M3N2N1O2N2O4Kc0]OlNYEAi:`0WE^Oj:c0WEZOn:d0REXOT;e0mDXOY;d0gDZO^;c0cDYOn;9SDDQ<:PDCR<=nC@U<`0lC]OV<c0::FN2N2N2N3M4L4L5K4LfUW4"}, "label": "second giraffe on the left"}]}
{"id": 118724, "image": "COCO_train2014_000000118724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second giraffe from the left\"."}], "task": "refering", "answer_template": "The \"the second giraffe from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 145, 146, 169, 192, 193, 215, 216, 217, 238, 239, 240, 261, 262, 263, 264, 284, 285, 286, 287], "bbox": [0.325953125, 0.38868852459016395, 0.493, 0.8369086651053864], "iscrowd": 0, "area": 6104.25985, "rle": {"size": [427, 640], "counts": "_Zg21Y=2N1N3N1O1N3M2N2N3O00001O00001O013L7I6J6J6K4N3TEaNd1I[5j1gHjNd1^Og5j1\\HTOd1TOQ6l1RHZOd1nN[6Y3]IiLe6`3nHbLT7m3YHSLk7e43M2N3M3M3L4M3M3M3M3M3M3M2N3lMiFNY91TG@o8>_GTOc8j0kGgNX8X1o1N2O1N2N2O1N2N10O0100O01000O01000O0100O0100O03N2N2M2O2N2M3N2N1O2N2O4Kc0]OlNYEAi:`0WE^Oj:c0WEZOn:d0REXOT;e0mDXOY;d0gDZO^;c0cDYOn;9SDDQ<:PDCR<=nC@U<`0lC]OV<c0::FN2N2N2N3M4L4L5K4LfUW4"}, "label": "the second giraffe from the left"}]}
{"id": 118724, "image": "COCO_train2014_000000118724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left giraffe in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the left giraffe in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 166, 167, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282], "bbox": [0.1928125, 0.36107728337236533, 0.342578125, 0.8646135831381733], "iscrowd": 0, "area": 6722.298750000001, "rle": {"size": [427, 640], "counts": "nac11W=:F9G9G:ZEUO[8o2M3O100O1O1O1OH8@`0@a0F98H9G9G8H9G9H1N1O1OQO`FUN_9f1jFXNT9c1VGYNi8a1bG[N\\8a1mG]Nn7a1\\H[N`7g1fHUNV7n1oHnMo6Z2nHdMP7U3V1i0WO9G1O1O1O2N1O1O1O2N1N01\\Od0\\Od0J6K3M4M2M4cNaEMa:3jE_OZ:`0PFTOR:k0ZFgNj9X1S101O0O2O001O1N2O1O1N2O1O1O1N4N2000O10000001O10O01O1O1O1O001N2O1N2O2M4M2M4MY\\_5"}, "label": "the left giraffe in the right hand picture"}]}
{"id": 118724, "image": "COCO_train2014_000000118724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left giraffe in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the left giraffe in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 144, 145, 166, 167, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282], "bbox": [0.1928125, 0.36107728337236533, 0.342578125, 0.8646135831381733], "iscrowd": 0, "area": 6722.298750000001, "rle": {"size": [427, 640], "counts": "nac11W=:F9G9G:ZEUO[8o2M3O100O1O1O1OH8@`0@a0F98H9G9G8H9G9H1N1O1OQO`FUN_9f1jFXNT9c1VGYNi8a1bG[N\\8a1mG]Nn7a1\\H[N`7g1fHUNV7n1oHnMo6Z2nHdMP7U3V1i0WO9G1O1O1O2N1O1O1O2N1N01\\Od0\\Od0J6K3M4M2M4cNaEMa:3jE_OZ:`0PFTOR:k0ZFgNj9X1S101O0O2O001O1N2O1O1N2O1O1O1N4N2000O10000001O10O01O1O1O1O001N2O1N2O2M4M2M4MY\\_5"}, "label": "the left giraffe in the right hand picture"}]}
{"id": 479172, "image": "COCO_train2014_000000479172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orioles baseball player waiting on pitch\"."}], "task": "refering", "answer_template": "The \"orioles baseball player waiting on pitch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 83, 84, 85, 105, 106, 107, 108, 109, 129, 130, 131, 132, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 219, 220, 222, 223, 242, 246, 247, 265, 269, 270], "bbox": [0.5192187500000001, 0.1859389671361502, 0.76071875, 0.7891079812206572], "iscrowd": 0, "area": 15251.945049999995, "rle": {"size": [426, 640], "counts": "h]Z41U=4L4L4L4L4M3O1O100O1O1O1N2N2N2N2N2N00001O000O10001N1O1O000000000000000O3N2N1O2M2O2M2O2N2N101O001N101O1jMZNhHg1X7]NbHe1^7`N\\Ha1d7cNWH^1i7fNRH[1n7jNlGW1T8mNfGT1[8PO]GT1c8i11O00A?L4001O00000000000O101O1O1O1O1O1O11O1O2N1O1O1N3N1O1O2N1002N2O1Na0_OO1N3N2M4M3L4M2M4M3M3L4M3L4M2M4M3L4LjLmH9P7FUI8i6G\\I8a6FeI9X6BQJ<l5_ObJ<[5_O_KO_4M_LZO`3d0bL^O]3`0eLAY3=iLDW39kLIS35PMMn20TM3i2KYM8e2G\\Md0X2[OiMe0X2RNWJk0a3S1X2nM\\Jn0]3T1e2jN^MT1c2kN`MS1`2lNcMQ1^2mNeMQ1\\2nNgMP1X2POkMm0V2ROlMl0U2TOmMh0U2WOnMf0S2YOoMd0S2\\OoMa0R2^OQN>Q2AQN=P2BSN:o1FSN7n1HTN5n1JUN3l1LWN0k10WNMj11ZNKj11ZNLj1OZNMl1LYN1Tbo1"}, "label": "orioles baseball player waiting on pitch"}]}
{"id": 479172, "image": "COCO_train2014_000000479172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a base ball player playing\"."}], "task": "refering", "answer_template": "The \"a base ball player playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 83, 84, 85, 105, 106, 107, 108, 109, 129, 130, 131, 132, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 219, 220, 222, 223, 242, 246, 247, 265, 269, 270], "bbox": [0.5192187500000001, 0.1859389671361502, 0.76071875, 0.7891079812206572], "iscrowd": 0, "area": 15251.945049999995, "rle": {"size": [426, 640], "counts": "h]Z41U=4L4L4L4L4M3O1O100O1O1O1N2N2N2N2N2N00001O000O10001N1O1O000000000000000O3N2N1O2M2O2M2O2N2N101O001N101O1jMZNhHg1X7]NbHe1^7`N\\Ha1d7cNWH^1i7fNRH[1n7jNlGW1T8mNfGT1[8PO]GT1c8i11O00A?L4001O00000000000O101O1O1O1O1O1O11O1O2N1O1O1N3N1O1O2N1002N2O1Na0_OO1N3N2M4M3L4M2M4M3M3L4M3L4M2M4M3L4LjLmH9P7FUI8i6G\\I8a6FeI9X6BQJ<l5_ObJ<[5_O_KO_4M_LZO`3d0bL^O]3`0eLAY3=iLDW39kLIS35PMMn20TM3i2KYM8e2G\\Md0X2[OiMe0X2RNWJk0a3S1X2nM\\Jn0]3T1e2jN^MT1c2kN`MS1`2lNcMQ1^2mNeMQ1\\2nNgMP1X2POkMm0V2ROlMl0U2TOmMh0U2WOnMf0S2YOoMd0S2\\OoMa0R2^OQN>Q2AQN=P2BSN:o1FSN7n1HTN5n1JUN3l1LWN0k10WNMj11ZNKj11ZNLj1OZNMl1LYN1Tbo1"}, "label": "a base ball player playing"}]}
{"id": 479172, "image": "COCO_train2014_000000479172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher\"."}], "task": "refering", "answer_template": "The \"the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 180, 181, 182, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 273, 274, 275, 296, 297, 298], "bbox": [0.78196875, 0.4501408450704225, 1.0, 0.8348826291079812], "iscrowd": 0, "area": 11867.287349999997, "rle": {"size": [426, 640], "counts": "mY`6;k<9H4O0O101N100O100O01000O10N1I8N101O0011O0O1N2O1N2O1O1N2O000O010O01JeC[O^<c042O1N1O2WNWOoFk0P9VOoFm0m8UOQGo0k8ROTGR1i8oNSFKc0Y1W9nNTFLc0X1V9nNUFMb0Y1V9kNVFO?[1X9[OaFl0\\9WO]FP1b9QOZFS1d9nN[FU1c9lN[FV1d9Y1O1O1O2M3N1ZGULk7m3QHXLl7i3RH[Lk7g3RHZLm7h3RHYLm7i3QHWLo7j3oGWLQ8k3mGVLR8l3lGTLS8n3kGTLT8n3jGRLV8a401O01O100N2O1O001O1O1N2O1O001O1bK^GR4n8O0O2O001O1cNnF`NR9a1oFdMN9T9S2oFaM0:Q9W2oF\\M3;o8Y2oFYM5<m8[2oFVM6>k8]2PGSM7>j8`2^GUMl8k2i0101N1000<C101O001OO1O010O1O100O10O01N2O1O1N2oNoEgNR:X1RFdNo9Z1VFbNl9\\1XF_Nj9_1\\F\\Nf9b1R1N2N3N1N2N3M2N3M4M2M4L4LbG"}, "label": "the catcher"}]}
{"id": 479172, "image": "COCO_train2014_000000479172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher is behind the plate\"."}], "task": "refering", "answer_template": "The \"the catcher is behind the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 180, 181, 182, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 273, 274, 275, 296, 297, 298], "bbox": [0.78196875, 0.4501408450704225, 1.0, 0.8348826291079812], "iscrowd": 0, "area": 11867.287349999997, "rle": {"size": [426, 640], "counts": "mY`6;k<9H4O0O101N100O100O01000O10N1I8N101O0011O0O1N2O1N2O1O1N2O000O010O01JeC[O^<c042O1N1O2WNWOoFk0P9VOoFm0m8UOQGo0k8ROTGR1i8oNSFKc0Y1W9nNTFLc0X1V9nNUFMb0Y1V9kNVFO?[1X9[OaFl0\\9WO]FP1b9QOZFS1d9nN[FU1c9lN[FV1d9Y1O1O1O2M3N1ZGULk7m3QHXLl7i3RH[Lk7g3RHZLm7h3RHYLm7i3QHWLo7j3oGWLQ8k3mGVLR8l3lGTLS8n3kGTLT8n3jGRLV8a401O01O100N2O1O001O1O1N2O1O001O1bK^GR4n8O0O2O001O1cNnF`NR9a1oFdMN9T9S2oFaM0:Q9W2oF\\M3;o8Y2oFYM5<m8[2oFVM6>k8]2PGSM7>j8`2^GUMl8k2i0101N1000<C101O001OO1O010O1O100O10O01N2O1O1N2oNoEgNR:X1RFdNo9Z1VFbNl9\\1XF_Nj9_1\\F\\Nf9b1R1N2N3N1N2N3M2N3M4M2M4L4LbG"}, "label": "the catcher is behind the plate"}]}
{"id": 569286, "image": "COCO_train2014_000000569286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza right side with the works\"."}], "task": "refering", "answer_template": "The \"pizza right side with the works\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 180, 181, 182, 183, 198, 199, 200], "bbox": [0.507875, 0.19325, 0.8089791666666667, 0.701125], "iscrowd": 0, "area": 25511.595699999998, "rle": {"size": [480, 480], "counts": "d_b3T1b=m0\\Od0]Ob0]Od0\\Od0\\O>B2O1N2N2N2N2O1N2N2N2N2O1O10O01O100O100O1O100O1O100O010O1O100O100O1O100O1O010O100O1O100O100O1O100O00100O100O1O100O10O01O10O01O010O10O01O010O10O01O010O1O010O10O01O010O10O01O010O1O0100O101O0O2O0O2O0O2N1N3N1N3N1N3N1N3N1N3N1N3N1N3N1N3M2O2M2O2M2O2L3K6K4L5J5L5K4K6K4L4K6K4L5J`0Aj0VOj0UO^lZ1"}, "label": "pizza right side with the works"}]}
{"id": 569286, "image": "COCO_train2014_000000569286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half a pizza with a variety of toppings\"."}], "task": "refering", "answer_template": "The \"a half a pizza with a variety of toppings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 180, 181, 182, 183, 198, 199, 200], "bbox": [0.507875, 0.19325, 0.8089791666666667, 0.701125], "iscrowd": 0, "area": 25511.595699999998, "rle": {"size": [480, 480], "counts": "d_b3T1b=m0\\Od0]Ob0]Od0\\Od0\\O>B2O1N2N2N2N2O1N2N2N2N2O1O10O01O100O100O1O100O1O100O010O1O100O100O1O100O1O010O100O1O100O100O1O100O00100O100O1O100O10O01O10O01O010O10O01O010O10O01O010O1O010O10O01O010O10O01O010O1O0100O101O0O2O0O2O0O2N1N3N1N3N1N3N1N3N1N3N1N3N1N3N1N3M2O2M2O2M2O2L3K6K4L5J5L5K4K6K4L4K6K4L5J`0Aj0VOj0UO^lZ1"}, "label": "a half a pizza with a variety of toppings"}]}
{"id": 569286, "image": "COCO_train2014_000000569286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hawaiin pizza\"."}], "task": "refering", "answer_template": "The \"the hawaiin pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 209, 210, 211, 212, 228, 229], "bbox": [0.13527083333333334, 0.243625, 0.4979166666666667, 0.77875], "iscrowd": 0, "area": 33821.3634, "rle": {"size": [480, 480], "counts": "gen0c0W>9H7I8G8I8G9H7I8M2M4M2M4M3L3M4M2M4M2M3N1N2O2M2N2O1N2O2M2O1N2O1N3M2O1N2O1N3N1N2O1N2N3N1N2O1N2O2M2O1N2O1N3M2O1N2O1N101N2O0O2N2O1N101N2O1N101N2O0O2O1O1N101N2O1N101N2O0O2O1N2O0O2O1N2O0O2O0O101O0O100O2O0O101N100O2O0O100O2O0O101O00001O0000001O00001O00001O0O10001O00001O00001O00001O0000001O00001O00001O0000001O00001O0O101O0000001O00001O00001OO1M3NV1jNh1WNm1XM]D]N\\Ra3"}, "label": "the hawaiin pizza"}]}
{"id": 569286, "image": "COCO_train2014_000000569286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a pizza with ham and pinapple\"."}], "task": "refering", "answer_template": "The \"half of a pizza with ham and pinapple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 209, 210, 211, 212, 228, 229], "bbox": [0.13527083333333334, 0.243625, 0.4979166666666667, 0.77875], "iscrowd": 0, "area": 33821.3634, "rle": {"size": [480, 480], "counts": "gen0c0W>9H7I8G8I8G9H7I8M2M4M2M4M3L3M4M2M4M2M3N1N2O2M2N2O1N2O2M2O1N2O1N3M2O1N2O1N3N1N2O1N2N3N1N2O1N2O2M2O1N2O1N3M2O1N2O1N101N2O0O2N2O1N101N2O1N101N2O0O2O1O1N101N2O1N101N2O0O2O1N2O0O2O1N2O0O2O0O101O0O100O2O0O101N100O2O0O100O2O0O101O00001O0000001O00001O00001O0O10001O00001O00001O00001O0000001O00001O00001O0000001O00001O0O101O0000001O00001O00001OO1M3NV1jNh1WNm1XM]D]N\\Ra3"}, "label": "half of a pizza with ham and pinapple"}]}
{"id": 282568, "image": "COCO_train2014_000000282568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soccer player in a red jersey with the number five on it\"."}], "task": "refering", "answer_template": "The \"a soccer player in a red jersey with the number five on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 50, 51, 65, 66, 80, 81, 93, 94, 95, 96, 108, 109, 110, 111, 123, 124, 125, 126, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 165, 166, 169, 170, 171, 180, 181, 184, 185, 186, 187, 200, 201, 202, 215, 216, 217, 230, 231, 232, 245, 246, 247, 260, 261, 262, 275, 276, 277, 290, 291, 292, 305], "bbox": [0.03713615023474178, 0.084703125, 0.5165258215962442, 0.884703125], "iscrowd": 0, "area": 27091.00734999999, "rle": {"size": [640, 426], "counts": "_Z:1mc03N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N000000O10000O1O1N2O1N2O1O1N2O1O1N2O11N101O00010O001O001O00001O0010O0001O001O0B?O01O010O10O010O01O001O0010O01O001O001OXNKf_O5Z`00b_OO^`05^_OKc`09Y_OGf`0=U_ODk`0a0f^OI[a0]100001O00001O010O000\\1eN1N2N2O0O2N2O1N2N101N2N2O1N2N2O2N2O010O100O010O101N2N3M2N2N3N1N2N2N3M2N2N3M2N2N2N3J5I7J7I6J6J6iFQIT5V7ZJ[I`5k6oIfI^4UOmI\\7T1QJh4CTIc6b1[JT5W7[JZI^5m6QJeIh5a6gIPJS6S:J5K2N2N2N2N2K5K5K5K5K5J[LQMfFj2b9_MRF`2Q:jM\\E_2d:_301O01O01bDfHg2d0g2f6^JZJX1UOZ4_6[JdJR1TOc4U6WJPKn0POl4V9mJoFT5S9dJQG]5Q9\\JSGf5n8TJVGl5`;1010O0hJPB_4m=WK\\BR5\\=cJmBg5c=9]BQJ_<g63M2N2kKiB]1Z=TN\\Cb1f<PNlCi1X<hMWDU2k;]MeD_2^;RMQEk2V>L3M4L3M4K4M4L3M4L<D=C6J6IiWP4"}, "label": "a soccer player in a red jersey with the number five on it"}]}
{"id": 282568, "image": "COCO_train2014_000000282568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a frisbee player in a red shirt and a tattoo on his calf jumping up\"."}], "task": "refering", "answer_template": "The \"a frisbee player in a red shirt and a tattoo on his calf jumping up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 50, 51, 65, 66, 80, 81, 93, 94, 95, 96, 108, 109, 110, 111, 123, 124, 125, 126, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 165, 166, 169, 170, 171, 180, 181, 184, 185, 186, 187, 200, 201, 202, 215, 216, 217, 230, 231, 232, 245, 246, 247, 260, 261, 262, 275, 276, 277, 290, 291, 292, 305], "bbox": [0.03713615023474178, 0.084703125, 0.5165258215962442, 0.884703125], "iscrowd": 0, "area": 27091.00734999999, "rle": {"size": [640, 426], "counts": "_Z:1mc03N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N000000O10000O1O1N2O1N2O1O1N2O1O1N2O11N101O00010O001O001O00001O0010O0001O001O0B?O01O010O10O010O01O001O0010O01O001O001OXNKf_O5Z`00b_OO^`05^_OKc`09Y_OGf`0=U_ODk`0a0f^OI[a0]100001O00001O010O000\\1eN1N2N2O0O2N2O1N2N101N2N2O1N2N2O2N2O010O100O010O101N2N3M2N2N3N1N2N2N3M2N2N3M2N2N2N3J5I7J7I6J6J6iFQIT5V7ZJ[I`5k6oIfI^4UOmI\\7T1QJh4CTIc6b1[JT5W7[JZI^5m6QJeIh5a6gIPJS6S:J5K2N2N2N2N2K5K5K5K5K5J[LQMfFj2b9_MRF`2Q:jM\\E_2d:_301O01O01bDfHg2d0g2f6^JZJX1UOZ4_6[JdJR1TOc4U6WJPKn0POl4V9mJoFT5S9dJQG]5Q9\\JSGf5n8TJVGl5`;1010O0hJPB_4m=WK\\BR5\\=cJmBg5c=9]BQJ_<g63M2N2kKiB]1Z=TN\\Cb1f<PNlCi1X<hMWDU2k;]MeD_2^;RMQEk2V>L3M4L3M4K4M4L3M4L<D=C6J6IiWP4"}, "label": "a frisbee player in a red shirt and a tattoo on his calf jumping up"}]}
{"id": 282568, "image": "COCO_train2014_000000282568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black man who is trying to grab the soccer ball\"."}], "task": "refering", "answer_template": "The \"the black man who is trying to grab the soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 68, 69, 70, 84, 85, 100, 101, 114, 115, 116, 129, 130, 131, 132, 144, 145, 146, 147, 148, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 205, 206, 207, 208, 220, 221, 222, 223, 235, 236, 237, 238, 250, 251, 252, 265, 266, 267, 280, 282], "bbox": [0.5557981220657278, 0.1303125, 0.9263380281690141, 0.811921875], "iscrowd": 0, "area": 31987.014599999988, "rle": {"size": [640, 426], "counts": "\\Wd43kc05K6J5K5K5K5L4K5K3M3N1N3M3O0010O01000O01RBQOj8o0RGWOk8j0QG\\Om8d0oFBo8=nFHP99lFLS94kFNT93iF1V9OhF3X9LfF8X9IcEXNgNQ2f;G^E^NhNo1h;D\\EcNhNk1l;BXEhNiNh1o;@UEmNgNg1R<\\OUEROeNd1W<YOREVOcNd1\\<UOoD[OaNb1a<ROlDY2U;fMhD^2W;aMhDa2Y;^MeDd2\\;[MbDg2_;YM^Di2c;WM[Dj2f;WMWDk2i;UMTDm2m;TMPDn2o;SMoCP3P<QMmCS3Q<mLmCX3P<iLmC\\3P<dLjCf3R<[LhCm3Z3hJl4Z7jJnHV5T7aJTI]5o6YJYIg5i6mIaIS6a6aIiI_6c6kHgIU7c6VHgIk7V92N2N1O2N2M3M3K5L4M3L3N3L4M3L4M3L4K5K4L5K5L4M3M3M3M3N1N3M3N200O2O1O1O1N2O2NO1O001N2O1O001N2O1OO1O001O1O1O001O1O1O001O1O1N2O1O1N2N2N3L3N2gIPDX4Y<eKjCl3f<oK_Cg3j<VLYC`3Q=\\LTCZ3U=bLoBT3[=hLjBh2f=TM]Bc2l=YMYBd2i=XM[Bf2g=UM_Bh2d=QMbBm2`=mLfBQ3W?N3L3M3M3M3M3N3L3M3M7G:D<E;D<E;ESPc0"}, "label": "the black man who is trying to grab the soccer ball"}]}
{"id": 282568, "image": "COCO_train2014_000000282568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a frisbee player in a black shirt\"."}], "task": "refering", "answer_template": "The \"a frisbee player in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 68, 69, 70, 84, 85, 100, 101, 114, 115, 116, 129, 130, 131, 132, 144, 145, 146, 147, 148, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 205, 206, 207, 208, 220, 221, 222, 223, 235, 236, 237, 238, 250, 251, 252, 265, 266, 267, 280, 282], "bbox": [0.5557981220657278, 0.1303125, 0.9263380281690141, 0.811921875], "iscrowd": 0, "area": 31987.014599999988, "rle": {"size": [640, 426], "counts": "\\Wd43kc05K6J5K5K5K5L4K5K3M3N1N3M3O0010O01000O01RBQOj8o0RGWOk8j0QG\\Om8d0oFBo8=nFHP99lFLS94kFNT93iF1V9OhF3X9LfF8X9IcEXNgNQ2f;G^E^NhNo1h;D\\EcNhNk1l;BXEhNiNh1o;@UEmNgNg1R<\\OUEROeNd1W<YOREVOcNd1\\<UOoD[OaNb1a<ROlDY2U;fMhD^2W;aMhDa2Y;^MeDd2\\;[MbDg2_;YM^Di2c;WM[Dj2f;WMWDk2i;UMTDm2m;TMPDn2o;SMoCP3P<QMmCS3Q<mLmCX3P<iLmC\\3P<dLjCf3R<[LhCm3Z3hJl4Z7jJnHV5T7aJTI]5o6YJYIg5i6mIaIS6a6aIiI_6c6kHgIU7c6VHgIk7V92N2N1O2N2M3M3K5L4M3L3N3L4M3L4M3L4K5K4L5K5L4M3M3M3M3N1N3M3N200O2O1O1O1N2O2NO1O001N2O1O001N2O1OO1O001O1O1O001O1O1O001O1O1N2O1O1N2N2N3L3N2gIPDX4Y<eKjCl3f<oK_Cg3j<VLYC`3Q=\\LTCZ3U=bLoBT3[=hLjBh2f=TM]Bc2l=YMYBd2i=XM[Bf2g=UM_Bh2d=QMbBm2`=mLfBQ3W?N3L3M3M3M3M3N3L3M3M7G:D<E;D<E;ESPc0"}, "label": "a frisbee player in a black shirt"}]}
{"id": 331724, "image": "COCO_train2014_000000331724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apples sitting between oranges and melons\"."}], "task": "refering", "answer_template": "The \"apples sitting between oranges and melons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.3096875, 0.6384074941451992, 0.704625, 0.8717798594847775], "iscrowd": 0, "area": 16683.480199999998, "rle": {"size": [427, 640], "counts": "emb21W=3M3N2M3M3N2K5B>O100O100O100O100O100O100O100O10000001O001O001O1O001O001O1O001O001ON2N2N2N2M3N2N2M3N200O1O1O100O1O1O100O11O00001O00H800O100O100O1O100O100O1O100O100N2O1O1O1O1O1O11O00000000000000001O00000000000000L4K5J6N20000000000000000000000000000001O1O1O001O1O1O1O1O1O1O1O1O001O0000O100O10000O100O100001O001O001O001O1O1O1O1O1O1O000000000000000000000000O100O1O100O100O1O11O2N1O1O2N1O1O2N1O1O1O00000000000000001O001O00001O00001O001O007I8H1OO10000O1000000O10000O13M4L2N1O00001O001]OQDHo;7SDHn;6TDIl;6VDIj;6XDIh;6ZDIf;6]DHd;7]DHc;7_DHa;5cDJ];3gDLZ;0kDNU;OoD0X<N\\^^2"}, "label": "apples sitting between oranges and melons"}]}
{"id": 331724, "image": "COCO_train2014_000000331724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stack of apples on a cart\"."}], "task": "refering", "answer_template": "The \"a stack of apples on a cart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.3096875, 0.6384074941451992, 0.704625, 0.8717798594847775], "iscrowd": 0, "area": 16683.480199999998, "rle": {"size": [427, 640], "counts": "emb21W=3M3N2M3M3N2K5B>O100O100O100O100O100O100O100O10000001O001O001O1O001O001O1O001O001ON2N2N2N2M3N2N2M3N200O1O1O100O1O1O100O11O00001O00H800O100O100O1O100O100O1O100O100N2O1O1O1O1O1O11O00000000000000001O00000000000000L4K5J6N20000000000000000000000000000001O1O1O001O1O1O1O1O1O1O1O1O001O0000O100O10000O100O100001O001O001O001O1O1O1O1O1O1O000000000000000000000000O100O1O100O100O1O11O2N1O1O2N1O1O2N1O1O1O00000000000000001O001O00001O00001O001O007I8H1OO10000O1000000O10000O13M4L2N1O00001O001]OQDHo;7SDHn;6TDIl;6VDIj;6XDIh;6ZDIf;6]DHd;7]DHc;7_DHa;5cDJ];3gDLZ;0kDNU;OoD0X<N\\^^2"}, "label": "a stack of apples on a cart"}]}
{"id": 61395, "image": "COCO_train2014_000000061395.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the middle\"."}], "task": "refering", "answer_template": "The \"the zebra in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 241, 242, 243, 244], "bbox": [0.37328125, 0.4347619047619048, 0.620484375, 0.6418840579710146], "iscrowd": 0, "area": 8093.106399999999, "rle": {"size": [483, 640], "counts": "gn`31P?3M3M2N3M3M3M2N3N2O0O2O1O1N101O1N101O1N2O001N101O001N101O0O2O001O001O1O100O1O1O1O001O1O1O1O1O00100O00001O000000001O0001O0001O0000O100O10000O10000O10000OhCZNU;f1iD^NU;b1iDaNV;^1gDgNX;Y1bDnN];R1\\DVOc;R23N2M3N2M3N2M3L4K5L4K5L4M300O10000003M3M3M2N3N2M3M3M3M3M2N2NkNlC\\OS<d0VDSOj;l0`DlN_;T1dDiN];U1gDhNY;W1kDgNT;Y1[1O11O001O10O1000O0100000000001O001O0O101O007I1O1O000000O010000000O1oNjB4V=HQC0F_OY==WC0CBV=:^COk<M\\CNf<MbCNbhb3"}, "label": "the zebra in the middle"}]}
{"id": 61395, "image": "COCO_train2014_000000061395.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra grazing in grass standing between two other zebras\"."}], "task": "refering", "answer_template": "The \"zebra grazing in grass standing between two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 241, 242, 243, 244], "bbox": [0.37328125, 0.4347619047619048, 0.620484375, 0.6418840579710146], "iscrowd": 0, "area": 8093.106399999999, "rle": {"size": [483, 640], "counts": "gn`31P?3M3M2N3M3M3M2N3N2O0O2O1O1N101O1N101O1N2O001N101O001N101O0O2O001O001O1O100O1O1O1O001O1O1O1O1O00100O00001O000000001O0001O0001O0000O100O10000O10000O10000OhCZNU;f1iD^NU;b1iDaNV;^1gDgNX;Y1bDnN];R1\\DVOc;R23N2M3N2M3N2M3L4K5L4K5L4M300O10000003M3M3M2N3N2M3M3M3M3M2N2NkNlC\\OS<d0VDSOj;l0`DlN_;T1dDiN];U1gDhNY;W1kDgNT;Y1[1O11O001O10O1000O0100000000001O001O0O101O007I1O1O000000O010000000O1oNjB4V=HQC0F_OY==WC0CBV=:^COk<M\\CNf<MbCNbhb3"}, "label": "zebra grazing in grass standing between two other zebras"}]}
{"id": 380889, "image": "COCO_train2014_000000380889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"projection of a white screen with green boxes and text\"."}], "task": "refering", "answer_template": "The \"projection of a white screen with green boxes and text\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321], "bbox": [0.8046875, 0.004305882352941177, 1.0, 0.9172470588235294], "iscrowd": 0, "area": 44743.5, "rle": {"size": [425, 640], "counts": "mge6=l<i0WOj0VOj0VOi0WOj0VOj0VOi0WOj0VOe0[O?A`0@`0@?A`0@`0@?A`0@`0@8H0000000000O10000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000001O000O10000000000000000000000000000000000000000O10000000000000000000000000o5QJnI"}, "label": "projection of a white screen with green boxes and text"}]}
{"id": 380889, "image": "COCO_train2014_000000380889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large projection screen that you can only see a small portion of\"."}], "task": "refering", "answer_template": "The \"large projection screen that you can only see a small portion of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321], "bbox": [0.8046875, 0.004305882352941177, 1.0, 0.9172470588235294], "iscrowd": 0, "area": 44743.5, "rle": {"size": [425, 640], "counts": "mge6=l<i0WOj0VOj0VOi0WOj0VOj0VOi0WOj0VOe0[O?A`0@`0@?A`0@`0@?A`0@`0@8H0000000000O10000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000001O000O10000000000000000000000000000000000000000O10000000000000000000000000o5QJnI"}, "label": "large projection screen that you can only see a small portion of"}]}
{"id": 380889, "image": "COCO_train2014_000000380889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor\"."}], "task": "refering", "answer_template": "The \"a computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236], "bbox": [0.055218750000000004, 0.3286588235294118, 0.292484375, 0.7061882352941177], "iscrowd": 0, "area": 21606.951499999996, "rle": {"size": [425, 640], "counts": "Xe>1X=;E:F;E;E;E;E;E;E;E;E;E;E:F4K10001O0000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O0000000000000000000`0Ab0]Oc0]Ob0^Oc0]Oc0]Ob0^Oc0]O^gk5"}, "label": "a computer monitor"}]}
{"id": 380889, "image": "COCO_train2014_000000380889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a monitor with a white cable plugged into it\"."}], "task": "refering", "answer_template": "The \"the back of a monitor with a white cable plugged into it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236], "bbox": [0.055218750000000004, 0.3286588235294118, 0.292484375, 0.7061882352941177], "iscrowd": 0, "area": 21606.951499999996, "rle": {"size": [425, 640], "counts": "Xe>1X=;E:F;E;E;E;E;E;E;E;E;E;E:F4K10001O0000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O0000000000000000000`0Ab0]Oc0]Ob0^Oc0]Oc0]Ob0^Oc0]O^gk5"}, "label": "the back of a monitor with a white cable plugged into it"}]}
{"id": 503777, "image": "COCO_train2014_000000503777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown couch against the blue wall\"."}], "task": "refering", "answer_template": "The \"brown couch against the blue wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329], "bbox": [0.12334375, 0.5264470588235294, 0.585171875, 0.9864941176470589], "iscrowd": 0, "area": 37328.90145, "rle": {"size": [425, 640], "counts": "QRQ1<k<4L3M3M4N1O2N1O1O2N1O2N1I7F;E:N3N1N2O2M2O2N1N2O2Mb1_N1O1N2O2M2O1O1N2O1N3N1O1N2O1O1O1O2N1O1N2O1O1O2N1O1O1O1O1O2M2O1O1O1O1O1O1O1O0O10O1000000O10000O10O10O10000O1000000O01000O10000O10000O10O10O1000000O10000O01000O10000O1000000O01000O10000O10000O0100000O10000O10000O01000O1000000O1000O010000O10000O10000O0100000O10000O1000O010000O1000000O1000O010000O10000O100000O0100000000SOUH\\Lk7b3o0M3O1O10000000O100000000000O10000000000000000000000000000000000000000O1O100O100O1O100O1O100O1O100O1N2O1O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O1N3L3L4O1N2O1N3N1]OjDQOY;m0kDoNW;P1jDmNX;R1a0N101O0O2O001O001O1O001O001O001O001O0010O100O01M2N3M2N3M2H]U^3"}, "label": "brown couch against the blue wall"}]}
{"id": 503777, "image": "COCO_train2014_000000503777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown color couch in front of blue wall\"."}], "task": "refering", "answer_template": "The \"brown color couch in front of blue wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329], "bbox": [0.12334375, 0.5264470588235294, 0.585171875, 0.9864941176470589], "iscrowd": 0, "area": 37328.90145, "rle": {"size": [425, 640], "counts": "QRQ1<k<4L3M3M4N1O2N1O1O2N1O2N1I7F;E:N3N1N2O2M2O2N1N2O2Mb1_N1O1N2O2M2O1O1N2O1N3N1O1N2O1O1O1O2N1O1N2O1O1O2N1O1O1O1O1O2M2O1O1O1O1O1O1O1O0O10O1000000O10000O10O10O10000O1000000O01000O10000O10000O10O10O1000000O10000O01000O10000O1000000O01000O10000O10000O0100000O10000O10000O01000O1000000O1000O010000O10000O10000O0100000O10000O1000O010000O1000000O1000O010000O10000O100000O0100000000SOUH\\Lk7b3o0M3O1O10000000O100000000000O10000000000000000000000000000000000000000O1O100O100O1O100O1O100O1O100O1N2O1O1N2O1N2O2M2O1N2O1N2O2M2O1N2O1N2O1N3L3L4O1N2O1N3N1]OjDQOY;m0kDoNW;P1jDmNX;R1a0N101O0O2O001O001O1O001O001O001O001O0010O100O01M2N3M2N3M2H]U^3"}, "label": "brown color couch in front of blue wall"}]}
{"id": 503777, "image": "COCO_train2014_000000503777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a drooping plant sits atop a stone coffee table in front of a brown sofa\"."}], "task": "refering", "answer_template": "The \"a drooping plant sits atop a stone coffee table in front of a brown sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 242, 243, 244, 245, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 314], "bbox": [0.46559375000000003, 0.6022588235294117, 0.734203125, 0.8831529411764707], "iscrowd": 0, "area": 10699.239899999999, "rle": {"size": [425, 640], "counts": "WQl33P=5K6M3O1N2N101N2N2O1N1O2O1N2O001O1O1O1O001O1O1O001O1O000000000O1000000C=O10O10O10000B>0O2O000O2O0O2O0O101N13N1O1O1O1O1O1O2N4L4L2NO1000O1O2O1N1O2O0O1O100O1O100O1O11O3M4L1O00O10000O100O10O00100O0010O01O001O001002M21O5K4L1O00O1N3M2O1N2O1N2N2O1N1O2O1N1O2O10O3N3ZMiES2Y:dMnE\\2`:O01000O101O1N2O1O2nMSEc1X;101NO2G9J5010O010O01O100O100O012M3M3N1N3N2M6K2M3M3N2M3N2M3M3NTWV2"}, "label": "a drooping plant sits atop a stone coffee table in front of a brown sofa"}]}
{"id": 503777, "image": "COCO_train2014_000000503777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower pot on the table\"."}], "task": "refering", "answer_template": "The \"a flower pot on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 242, 243, 244, 245, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 314], "bbox": [0.46559375000000003, 0.6022588235294117, 0.734203125, 0.8831529411764707], "iscrowd": 0, "area": 10699.239899999999, "rle": {"size": [425, 640], "counts": "WQl33P=5K6M3O1N2N101N2N2O1N1O2O1N2O001O1O1O1O001O1O1O001O1O000000000O1000000C=O10O10O10000B>0O2O000O2O0O2O0O101N13N1O1O1O1O1O1O2N4L4L2NO1000O1O2O1N1O2O0O1O100O1O100O1O11O3M4L1O00O10000O100O10O00100O0010O01O001O001002M21O5K4L1O00O1N3M2O1N2O1N2N2O1N1O2O1N1O2O10O3N3ZMiES2Y:dMnE\\2`:O01000O101O1N2O1O2nMSEc1X;101NO2G9J5010O010O01O100O100O012M3M3N1N3N2M6K2M3M3N2M3N2M3M3NTWV2"}, "label": "a flower pot on the table"}]}
{"id": 503777, "image": "COCO_train2014_000000503777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch along the brick wall\"."}], "task": "refering", "answer_template": "The \"the couch along the brick wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 344], "bbox": [0.7794531250000001, 0.5368941176470589, 1.0, 0.9712235294117647], "iscrowd": 0, "area": 18953.723650000007, "rle": {"size": [425, 640], "counts": "S\\_6:n<3M3M3M3M3N0jDD^9>VF4b9MZF<`9F\\F`0a9@[Fd0d9^OWFf0i9ZOQFl0o9X1N10000O2O00001N10000O2O000O101O000O2O000O101O0O1000000O10000O101O000O101O0O101O0O101O000O1N3N1N2K6K4O1N3N1O1N3N1O1O101N100O2O0O1O2O0O10001O00001O00001O00001O00001O000010O01O00001O000010O0001O001O000010O0001O00001O001O01O01O00001O000010O01O00001O001O10fKeIe1Z6XMgLk0nK"}, "label": "the couch along the brick wall"}]}
{"id": 503777, "image": "COCO_train2014_000000503777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair underneath hearth\"."}], "task": "refering", "answer_template": "The \"the chair underneath hearth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 344], "bbox": [0.7794531250000001, 0.5368941176470589, 1.0, 0.9712235294117647], "iscrowd": 0, "area": 18953.723650000007, "rle": {"size": [425, 640], "counts": "S\\_6:n<3M3M3M3M3N0jDD^9>VF4b9MZF<`9F\\F`0a9@[Fd0d9^OWFf0i9ZOQFl0o9X1N10000O2O00001N10000O2O000O101O000O2O000O101O0O1000000O10000O101O000O101O0O101O0O101O000O1N3N1N2K6K4O1N3N1O1N3N1O1O101N100O2O0O1O2O0O10001O00001O00001O00001O00001O000010O01O00001O000010O0001O001O000010O0001O00001O001O01O01O00001O000010O01O00001O001O10fKeIe1Z6XMgLk0nK"}, "label": "the chair underneath hearth"}]}
{"id": 151523, "image": "COCO_train2014_000000151523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup\"."}], "task": "refering", "answer_template": "The \"a white cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 140, 141, 156, 157, 158, 159, 160, 161, 162], "bbox": [0.3151, 0.12890000000000001, 0.8515666666666667, 0.34245], "iscrowd": 0, "area": 27139.867199999997, "rle": {"size": [600, 600], "counts": "Uk^31eb03M3L3N3M2N3O1O001O001O1O001O001O2N1O1O1O1O1O1O1O1O1O1O1O1O100O2O0O100ON32M2N3N1N3N1N3N1N3N1N3N1N2N3M2M4L3N3L3N3L3N3L3N3L3M4M00000O10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000O1O00N2N2N3M2N2N2N2N2N2N2N2M3N3M2N2N2N3M3N2N2O1N2O1N101N2O1N2O1N2O1N2O1N2N2O1N2N101N2O1N2N2O10000O100O10001N2O001N2O1N101O1N2O1O0O2O1N4M3M3LhQd1"}, "label": "a white cup"}]}
{"id": 151523, "image": "COCO_train2014_000000151523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small rimmed bowl behind a delicious pizza\"."}], "task": "refering", "answer_template": "The \"a small rimmed bowl behind a delicious pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 140, 141, 156, 157, 158, 159, 160, 161, 162], "bbox": [0.3151, 0.12890000000000001, 0.8515666666666667, 0.34245], "iscrowd": 0, "area": 27139.867199999997, "rle": {"size": [600, 600], "counts": "Uk^31eb03M3L3N3M2N3O1O001O001O1O001O001O2N1O1O1O1O1O1O1O1O1O1O1O1O100O2O0O100ON32M2N3N1N3N1N3N1N3N1N3N1N2N3M2M4L3N3L3N3L3N3L3N3L3M4M00000O10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000O1O00N2N2N3M2N2N2N2N2N2N2N2M3N3M2N2N2N3M3N2N2O1N2O1N101N2O1N2O1N2O1N2O1N2N2O1N2N101N2O1N2N2O10000O100O10001N2O001N2O1N101O1N2O1O0O2O1N4M3M3LhQd1"}, "label": "a small rimmed bowl behind a delicious pizza"}]}
{"id": 151523, "image": "COCO_train2014_000000151523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white photo of a pitcher of liquid\"."}], "task": "refering", "answer_template": "The \"a black and white photo of a pitcher of liquid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 21, 22, 23, 24, 25, 26, 27, 42, 43, 44, 45, 46, 47, 48, 64, 65, 66, 67, 68, 69, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 170], "bbox": [0.024716666666666668, 0.04045, 0.3348333333333333, 0.39775], "iscrowd": 0, "area": 29691.9405, "rle": {"size": [600, 600], "counts": "Pj81gb01`]O0Vb01g]O6Tb0Ki]O<Rb0:K5K5K5K4L5K5K5K5K5K5K5K5K5K4L5K5K5K5K5K5K5K5K5K4L5K5K5K5K5K5L4K5K5K4L5K5K5K5K2N000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000001N10000O1000000O100`KRCV3n<[LaCe3_<lKQDR4[=K6E:F;E:J7L3N3M2N3UNk_Oo0W`0POn_Oj0U`0TOP@f0R`0XOT@a0P`0]OU@=m?BU@;n?BV@:m?DV@8l?GV@6m?HV@4l?JW@2m?LV@0l?O`1NhfY7"}, "label": "a black and white photo of a pitcher of liquid"}]}
{"id": 151523, "image": "COCO_train2014_000000151523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pitcher full of liquid\"."}], "task": "refering", "answer_template": "The \"a pitcher full of liquid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 21, 22, 23, 24, 25, 26, 27, 42, 43, 44, 45, 46, 47, 48, 64, 65, 66, 67, 68, 69, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 170], "bbox": [0.024716666666666668, 0.04045, 0.3348333333333333, 0.39775], "iscrowd": 0, "area": 29691.9405, "rle": {"size": [600, 600], "counts": "Pj81gb01`]O0Vb01g]O6Tb0Ki]O<Rb0:K5K5K5K4L5K5K5K5K5K5K5K5K5K4L5K5K5K5K5K5K5K5K5K4L5K5K5K5K5K5L4K5K5K4L5K5K5K5K2N000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000001N10000O1000000O100`KRCV3n<[LaCe3_<lKQDR4[=K6E:F;E:J7L3N3M2N3UNk_Oo0W`0POn_Oj0U`0TOP@f0R`0XOT@a0P`0]OU@=m?BU@;n?BV@:m?DV@8l?GV@6m?HV@4l?JW@2m?LV@0l?O`1NhfY7"}, "label": "a pitcher full of liquid"}]}
{"id": 200678, "image": "COCO_train2014_000000200678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hey man watching over an active grill\"."}], "task": "refering", "answer_template": "The \"hey man watching over an active grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 147, 148, 149, 150, 152, 153, 171, 172, 173, 175, 176, 198, 221, 244, 267, 290, 313, 334], "bbox": [0.40878125, 0.3074375, 0.672296875, 0.87725], "iscrowd": 0, "area": 5469.595150000002, "rle": {"size": [480, 640], "counts": "RPk31n>2N1O101O00001O00001O00001O00001O00010O00001O00001O0000001O0000000010O00000001O000000001O0000001O000000001O00000000001O000000001O0000000O101O0000000O1O1O2O0O1hG@o1a0PNAo1?QNAn1EWHh0k5Dl1D\\Hf0h5GS28kMIW25iMLX21iMOY2OgM2Z2LeM5^2HbM8`2F`M;a2C_M=d2@^M>d2@_M=e8M3N3L3Mh_;OX`D2N2M3N2RCHQ;:lDKP;7lDNU:HQE<f02Z:DiD<j05^:B`D;n09b:@XD9R1=g:NVE6k:HQE>P;_OmDf0o1_NZ6n0_Gi0W1]NPOMY8V1TGe0Z1cNUOD\\8Z3TH\\M\\O[O_8_3iG`MDROb8e3]GcMMjNe8_5WGbJh8m5O0O101N1O2O001OL4dM[IULl6S1TGR1_;nNSE2Z;Ol1O10000O103M3L5L3M3M4K4KUWR3"}, "label": "hey man watching over an active grill"}]}
{"id": 200678, "image": "COCO_train2014_000000200678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with hand on grill\"."}], "task": "refering", "answer_template": "The \"person with hand on grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 147, 148, 149, 150, 152, 153, 171, 172, 173, 175, 176, 198, 221, 244, 267, 290, 313, 334], "bbox": [0.40878125, 0.3074375, 0.672296875, 0.87725], "iscrowd": 0, "area": 5469.595150000002, "rle": {"size": [480, 640], "counts": "RPk31n>2N1O101O00001O00001O00001O00001O00010O00001O00001O0000001O0000000010O00000001O000000001O0000001O000000001O00000000001O000000001O0000000O101O0000000O1O1O2O0O1hG@o1a0PNAo1?QNAn1EWHh0k5Dl1D\\Hf0h5GS28kMIW25iMLX21iMOY2OgM2Z2LeM5^2HbM8`2F`M;a2C_M=d2@^M>d2@_M=e8M3N3L3Mh_;OX`D2N2M3N2RCHQ;:lDKP;7lDNU:HQE<f02Z:DiD<j05^:B`D;n09b:@XD9R1=g:NVE6k:HQE>P;_OmDf0o1_NZ6n0_Gi0W1]NPOMY8V1TGe0Z1cNUOD\\8Z3TH\\M\\O[O_8_3iG`MDROb8e3]GcMMjNe8_5WGbJh8m5O0O101N1O2O001OL4dM[IULl6S1TGR1_;nNSE2Z;Ol1O10000O103M3L5L3M3M4K4KUWR3"}, "label": "person with hand on grill"}]}
{"id": 200678, "image": "COCO_train2014_000000200678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a white shirt and blue jeans standing outside holding a blue umbrella\"."}], "task": "refering", "answer_template": "The \"a woman wearing a white shirt and blue jeans standing outside holding a blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 311, 312, 313, 334, 335, 336, 357, 358, 359, 381], "bbox": [0.5137968749999999, 0.2977708333333334, 0.635765625, 0.9521458333333334], "iscrowd": 0, "area": 15856.593000000003, "rle": {"size": [480, 640], "counts": "U^j4;Z>=K5K4O2N2N101OROWBd0i=ZOYBf0j=UOXBl0_5TOi2m0SI]OT2Jg4k0nHEV2Ak4l0[HROCl0Q3WOP5m0UHc0f2aNS5o0PHk0f2XNX5]3cJdL[5b3aJ^LY4LPJk3b1YL_3JjJ7JR4h1nKa31VJI793X4l1eKa3?]JE3\\4l1aKc3`0\\JE2^4l1]Kf3a0ZJF2`4j1ZKk3a6QL`In3g6kKYIT4Q7aKRI]4V7ZKlHe4Y7UKiHk4\\7nJgHP5^7jJdHV5d8O001O00001O0O2O00001O0000001N1000O1O001O1O0hGfJi6]5UIeJk6[5RIiJm6W5QIlJn6S5RIoJm6Q5RIPKo6S5lHoJS7R5lHoJS7Q5mHoJS7R5jHQKV1^Om3b5jJRKo0HU4W5VJbJ;b0U1LW4Z5aJkJT12W4U5cJkJR17W4R6dKTJT1iNl0T7lMYJl0SOQ1e6nM^J<Cc1P6kMdJ9Ej1h5eMmJ9Co1a5]MYK=]OR2S8aMZHS2a:I8\\Od0UOk0EVk\\3"}, "label": "a woman wearing a white shirt and blue jeans standing outside holding a blue umbrella"}]}
{"id": 200678, "image": "COCO_train2014_000000200678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black women , dressed in jeans and a white top , holding an umbrella for another person whom is operating a grill\"."}], "task": "refering", "answer_template": "The \"a black women , dressed in jeans and a white top , holding an umbrella for another person whom is operating a grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 311, 312, 313, 334, 335, 336, 357, 358, 359, 381], "bbox": [0.5137968749999999, 0.2977708333333334, 0.635765625, 0.9521458333333334], "iscrowd": 0, "area": 15856.593000000003, "rle": {"size": [480, 640], "counts": "U^j4;Z>=K5K4O2N2N101OROWBd0i=ZOYBf0j=UOXBl0_5TOi2m0SI]OT2Jg4k0nHEV2Ak4l0[HROCl0Q3WOP5m0UHc0f2aNS5o0PHk0f2XNX5]3cJdL[5b3aJ^LY4LPJk3b1YL_3JjJ7JR4h1nKa31VJI793X4l1eKa3?]JE3\\4l1aKc3`0\\JE2^4l1]Kf3a0ZJF2`4j1ZKk3a6QL`In3g6kKYIT4Q7aKRI]4V7ZKlHe4Y7UKiHk4\\7nJgHP5^7jJdHV5d8O001O00001O0O2O00001O0000001N1000O1O001O1O0hGfJi6]5UIeJk6[5RIiJm6W5QIlJn6S5RIoJm6Q5RIPKo6S5lHoJS7R5lHoJS7Q5mHoJS7R5jHQKV1^Om3b5jJRKo0HU4W5VJbJ;b0U1LW4Z5aJkJT12W4U5cJkJR17W4R6dKTJT1iNl0T7lMYJl0SOQ1e6nM^J<Cc1P6kMdJ9Ej1h5eMmJ9Co1a5]MYK=]OR2S8aMZHS2a:I8\\Od0UOk0EVk\\3"}, "label": "a black women , dressed in jeans and a white top , holding an umbrella for another person whom is operating a grill"}]}
{"id": 536555, "image": "COCO_train2014_000000536555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat sitting on a luggage and staring at the camera\"."}], "task": "refering", "answer_template": "The \"a black and white cat sitting on a luggage and staring at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 311, 312, 313, 334, 335, 336], "bbox": [0.06398437500000001, 0.06016470588235294, 0.635625, 1.0], "iscrowd": 0, "area": 69710.3765, "rle": {"size": [425, 640], "counts": "`Qa02X=2M4M2M4M2M3N3aD\\OQ:g0hE^OX:e0_EB_:a0ZEDf:?REGl:<lDJS;X1O0O2O0O2O0O2O0O2O0O2O0O2O0O2M3N1O2N1O2N2N2N3L3N2N2N2N3M2N2M3N2N2N3M2N2N2M3N3M2N2N2N2N3L3N2N2N2N2N2N2M3N2N2N2N1O2N2M3N2N2N1O2N2N2UKhJ`1[5^NeJb1]5\\NcJd1_5ZNaJe1a5ZN_Jf1c5YN\\Jf1g5XNYJh1i5VNWJi1l5UNTJk1n5TNQJk1Q6TNoIl1S6RNmIm1V6QNjIo1X6oMhIP2e6eMZI[2\\7oLdHQ3R8YLnGf3k800O100O100O1001O1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1N3M2N2N2N2N3M[KoGU4o7kKRHV4k7jKWHW4h7hKYHY4f7eK\\H\\4b7cK`H_4^7`KcHa4\\7]KgHc4X7[KjHf4U7YKlHh4R7WKPIj4o6UKRIm4l6QKVIP5i6nJYIS5e6mJ\\IT5\\71O0000000001O000000000000000000000000001O01O00001O001O00001O001O01O01O001O00001O001O00001O00010O001O00001O001O00001O001O01O01O001O00001O00001O001O00010O001O00001O001O00001O0010O0001O00001O001O00001O0O2O000O2O001N10001N101O0O101O0O101O0O2O000010O01O01O01O011N2N3M3N2M3M2O2M3M3N1N3M3M3N1N3M3N2M3M2N3N3L4L5L4K410001N101O0O2O001NdLoG]1Q8PN^HU2a7XMmHm2e84L01O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O003M3M2N3M3M3M2N3M3F:E;F9F;E;F:EX[P3"}, "label": "a black and white cat sitting on a luggage and staring at the camera"}]}
{"id": 536555, "image": "COCO_train2014_000000536555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat sitting up to the left of another\"."}], "task": "refering", "answer_template": "The \"a black and white cat sitting up to the left of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 311, 312, 313, 334, 335, 336], "bbox": [0.06398437500000001, 0.06016470588235294, 0.635625, 1.0], "iscrowd": 0, "area": 69710.3765, "rle": {"size": [425, 640], "counts": "`Qa02X=2M4M2M4M2M3N3aD\\OQ:g0hE^OX:e0_EB_:a0ZEDf:?REGl:<lDJS;X1O0O2O0O2O0O2O0O2O0O2O0O2O0O2M3N1O2N1O2N2N2N3L3N2N2N2N3M2N2M3N2N2N3M2N2N2M3N3M2N2N2N2N3L3N2N2N2N2N2N2M3N2N2N2N1O2N2M3N2N2N1O2N2N2UKhJ`1[5^NeJb1]5\\NcJd1_5ZNaJe1a5ZN_Jf1c5YN\\Jf1g5XNYJh1i5VNWJi1l5UNTJk1n5TNQJk1Q6TNoIl1S6RNmIm1V6QNjIo1X6oMhIP2e6eMZI[2\\7oLdHQ3R8YLnGf3k800O100O100O1001O1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1N3M2N2N2N2N3M[KoGU4o7kKRHV4k7jKWHW4h7hKYHY4f7eK\\H\\4b7cK`H_4^7`KcHa4\\7]KgHc4X7[KjHf4U7YKlHh4R7WKPIj4o6UKRIm4l6QKVIP5i6nJYIS5e6mJ\\IT5\\71O0000000001O000000000000000000000000001O01O00001O001O00001O001O01O01O001O00001O001O00001O00010O001O00001O001O00001O001O01O01O001O00001O00001O001O00010O001O00001O001O00001O0010O0001O00001O001O00001O0O2O000O2O001N10001N101O0O101O0O101O0O2O000010O01O01O01O011N2N3M3N2M3M2O2M3M3N1N3M3M3N1N3M3N2M3M2N3N3L4L5L4K410001N101O0O2O001NdLoG]1Q8PN^HU2a7XMmHm2e84L01O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O003M3M2N3M3M3M2N3M3F:E;F9F;E;F:EX[P3"}, "label": "a black and white cat sitting up to the left of another"}]}
{"id": 536555, "image": "COCO_train2014_000000536555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and white colour cat sitting on a bag\"."}], "task": "refering", "answer_template": "The \"a brown and white colour cat sitting on a bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.432328125, 0.16536470588235294, 0.94940625, 0.6497411764705882], "iscrowd": 0, "area": 46987.58605000001, "rle": {"size": [425, 640], "counts": "XQc32V=2O1N3M2N2O1N2O2N1RDBP;>jDHU;9fDKZ;7_DO`;2[D2e;j0O001O1O010O001O001O001O100O1O1O1O1O1O00100O1O1O1O1O1O1O1O1O2N4L4L4L4L4L4L4L4L4L3N3L4L4L3M4L4L4L1O00100O1O00100O1O00100O1O1O100O1O1O101N1O100O2O0O100O2O2M3N2M2O2N2M3N1N100O100O10000O100O100O100000O001O1O1O001O1O1O001O1O1O00O10000O1000000O10000O10O1000O1000000O10000O1000000O10000O1000000O1000O01000000O10000O1000000O01000O1000000O1000000O10O10O1000000O1000000000O10001O00000000000000001N10000000000000001O00000O10000000001O000000000O10001O00000000000000001N10000000000O1000001N10000O1000000O101O000O10000O1000000O2O00000O10000O10001O0O10002M2O2N1O1N3N1O2N1N2O2N1N3N1O1O2M2O2N1N2N3M2N3M2M3N3M2N3M2N3M2N2N3M5J5L5K4L5K4L5K4J6H9F9H9G8H9GTS="}, "label": "a brown and white colour cat sitting on a bag"}]}
{"id": 536555, "image": "COCO_train2014_000000536555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat laying down behind sitting cat\"."}], "task": "refering", "answer_template": "The \"cat laying down behind sitting cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.432328125, 0.16536470588235294, 0.94940625, 0.6497411764705882], "iscrowd": 0, "area": 46987.58605000001, "rle": {"size": [425, 640], "counts": "XQc32V=2O1N3M2N2O1N2O2N1RDBP;>jDHU;9fDKZ;7_DO`;2[D2e;j0O001O1O010O001O001O001O100O1O1O1O1O1O00100O1O1O1O1O1O1O1O1O2N4L4L4L4L4L4L4L4L4L3N3L4L4L3M4L4L4L1O00100O1O00100O1O00100O1O1O100O1O1O101N1O100O2O0O100O2O2M3N2M2O2N2M3N1N100O100O10000O100O100O100000O001O1O1O001O1O1O001O1O1O00O10000O1000000O10000O10O1000O1000000O10000O1000000O10000O1000000O1000O01000000O10000O1000000O01000O1000000O1000000O10O10O1000000O1000000000O10001O00000000000000001N10000000000000001O00000O10000000001O000000000O10001O00000000000000001N10000000000O1000001N10000O1000000O101O000O10000O1000000O2O00000O10000O10001O0O10002M2O2N1O1N3N1O2N1N2O2N1N3N1O1O2M2O2N1N2N3M2N3M2M3N3M2N3M2N3M2N2N3M5J5L5K4L5K4L5K4J6H9F9H9G8H9GTS="}, "label": "cat laying down behind sitting cat"}]}
{"id": 307190, "image": "COCO_train2014_000000307190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is sitting besides a child wearing red t shirt\"."}], "task": "refering", "answer_template": "The \"a man is sitting besides a child wearing red t shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.08953124999999999, 0.12725, 0.6300625, 0.989875], "iscrowd": 0, "area": 85767.03709999997, "rle": {"size": [480, 640], "counts": "RTk01h>:G9F:F:G9F:F9G:G9F:F:F:G9F:F:F:G8G:F:J4N1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1J6K5K5J6K5J6K5K5J6K5J6K5K5O1O100O1O1O1O1O1O100O1O1O1O00100O1O1O1O1O100O1O1O1O100000000O10000000000O100000000O100000000O10000000000O100000O2O0O2O0O2O0O2N100O2O0O2O0O2N101N101[HZMm2g2QM]Ml2d2RM_Mm2b2PMaMn2`2oLdMP3]2mLfMQ3[2mLiMQ3X2lLkMR3V2lLmMR3U2kLnMT3R2jLQNT3Q2iLSNU3m1iLVNU3k1hLYNW3h1fL[NX3f1fL]NX3e1eL^NZ3b1dLbNY3`1dLcN[3]1cLfN[3\\1bLgN]3Y1aLjN]3X1`LlN]3U1`LoN_3Q1_LRO_3P1^LSOa3o0[LTOc3n0ZLUOd3n0XLVOf3k0WLXOg3j0VLYOi3i0SLZOk3h0QL\\On3e0oK_On3d0nK_OP4c0lKAS4`0jKCT4`0gKDX4=eKEY4?cKD[4>aKF\\4=aKE]4?_KD^4?^KE`4=]KF`4>[KEc4>ZKEd4=YKFd4]5M2O2M3M3M2O2M3M2O2M3M3N1N3M3M3N1N3M3N2M2N4L3N2M3M3N2M3M3N2M4L3M3N2N2N2N2O1N010O0001O010O000010O01O00001N1O2N1O101N1O2N11O2O1N3N1N2N3N1N2O1N3M2O1N2O2M2Ngi^3"}, "label": "a man is sitting besides a child wearing red t shirt"}]}
{"id": 307190, "image": "COCO_train2014_000000307190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a red shirt helps a baby brush their teeth\"."}], "task": "refering", "answer_template": "The \"a man wearing a red shirt helps a baby brush their teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.08953124999999999, 0.12725, 0.6300625, 0.989875], "iscrowd": 0, "area": 85767.03709999997, "rle": {"size": [480, 640], "counts": "RTk01h>:G9F:F:G9F:F9G:G9F:F:F:G9F:F:F:G8G:F:J4N1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1J6K5K5J6K5J6K5K5J6K5J6K5K5O1O100O1O1O1O1O1O100O1O1O1O00100O1O1O1O1O100O1O1O1O100000000O10000000000O100000000O100000000O10000000000O100000O2O0O2O0O2O0O2N100O2O0O2O0O2N101N101[HZMm2g2QM]Ml2d2RM_Mm2b2PMaMn2`2oLdMP3]2mLfMQ3[2mLiMQ3X2lLkMR3V2lLmMR3U2kLnMT3R2jLQNT3Q2iLSNU3m1iLVNU3k1hLYNW3h1fL[NX3f1fL]NX3e1eL^NZ3b1dLbNY3`1dLcN[3]1cLfN[3\\1bLgN]3Y1aLjN]3X1`LlN]3U1`LoN_3Q1_LRO_3P1^LSOa3o0[LTOc3n0ZLUOd3n0XLVOf3k0WLXOg3j0VLYOi3i0SLZOk3h0QL\\On3e0oK_On3d0nK_OP4c0lKAS4`0jKCT4`0gKDX4=eKEY4?cKD[4>aKF\\4=aKE]4?_KD^4?^KE`4=]KF`4>[KEc4>ZKEd4=YKFd4]5M2O2M3M3M2O2M3M2O2M3M3N1N3M3M3N1N3M3N2M2N4L3N2M3M3N2M3M3N2M4L3M3N2N2N2N2O1N010O0001O010O000010O01O00001N1O2N1O101N1O2N11O2O1N3N1N2N3N1N2O1N3M2O1N2O2M2Ngi^3"}, "label": "a man wearing a red shirt helps a baby brush their teeth"}]}
{"id": 307190, "image": "COCO_train2014_000000307190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child sitting on a toilet and brushing his or her teeth\"."}], "task": "refering", "answer_template": "The \"a child sitting on a toilet and brushing his or her teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 128, 129, 130, 131, 148, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 289, 290, 291, 292, 293, 313, 314, 315, 336, 337, 338, 359, 360, 361, 382, 383, 384], "bbox": [0.44731249999999995, 0.23991666666666667, 0.7533593749999999, 0.9843125], "iscrowd": 0, "area": 30657.83014999999, "rle": {"size": [480, 640], "counts": "YXV41n>2O1N3N1N3N1N2O2M2N3N1N3N1N2O2M4M4L3M3N3L3N2M2O1N100O100O2O0O100O10O001O001O010O001O1O2O1N2N00O101N100O100O101N10000O101N100O101N100O2O0O101N100O2N100O2O0O1L5J5L4M4L3M3M4L3M3M4L3N2N3L3ZGULU6m3kITLQ6P4mIQLQ6Q4nIQLo5Q4QJoKl5U4RJmKk5U4TJlKj5V4UJlKi5V4VJkKh5V4WJkKh5V4WJlKf5W4YJiKf5X4YJjKe5X4YJiKf5X4YJjKe5X4ZJhKe5Y4ZJiKc5Z4[JgKd5Z4\\JgKb5[4\\JfKc5[4\\JgKb5[4\\JhKa5Y4_JiK]5Z4aJjK[5W4dJlKZ5U4eJmKY5S4fJPLX5Q4aJWL]5i3]J_La5b3XJfLf5Z3TJnLj5V6N2O1N2N3lK[Gg1j8eMeG[2a8oLoGQ3V8UL_Hk3\\9O00000000000000000000000000000000000000000001O00000000000000000000000O1O1N2_Oa0^Nb1N2cJRJg1o5RNZJj1h5PN_Jm1c5mMeJQ1F]Lg5\\2kJT1T6gNRJX1n5dNXJY1l5_N[J^1k5YNZJf1l5PNZJm1l5iM[JT2l5bMZJ\\2k5ZM[Jd2k5SM[Jj2P9M2O2M3hMZDQ1h;`NjD\\1a<hNbBe0S>I7J6J6JWgY2"}, "label": "a child sitting on a toilet and brushing his or her teeth"}]}
{"id": 307190, "image": "COCO_train2014_000000307190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child brushing their teeth\"."}], "task": "refering", "answer_template": "The \"the child brushing their teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 128, 129, 130, 131, 148, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 289, 290, 291, 292, 293, 313, 314, 315, 336, 337, 338, 359, 360, 361, 382, 383, 384], "bbox": [0.44731249999999995, 0.23991666666666667, 0.7533593749999999, 0.9843125], "iscrowd": 0, "area": 30657.83014999999, "rle": {"size": [480, 640], "counts": "YXV41n>2O1N3N1N3N1N2O2M2N3N1N3N1N2O2M4M4L3M3N3L3N2M2O1N100O100O2O0O100O10O001O001O010O001O1O2O1N2N00O101N100O100O101N10000O101N100O101N100O2O0O101N100O2N100O2O0O1L5J5L4M4L3M3M4L3M3M4L3N2N3L3ZGULU6m3kITLQ6P4mIQLQ6Q4nIQLo5Q4QJoKl5U4RJmKk5U4TJlKj5V4UJlKi5V4VJkKh5V4WJkKh5V4WJlKf5W4YJiKf5X4YJjKe5X4YJiKf5X4YJjKe5X4ZJhKe5Y4ZJiKc5Z4[JgKd5Z4\\JgKb5[4\\JfKc5[4\\JgKb5[4\\JhKa5Y4_JiK]5Z4aJjK[5W4dJlKZ5U4eJmKY5S4fJPLX5Q4aJWL]5i3]J_La5b3XJfLf5Z3TJnLj5V6N2O1N2N3lK[Gg1j8eMeG[2a8oLoGQ3V8UL_Hk3\\9O00000000000000000000000000000000000000000001O00000000000000000000000O1O1N2_Oa0^Nb1N2cJRJg1o5RNZJj1h5PN_Jm1c5mMeJQ1F]Lg5\\2kJT1T6gNRJX1n5dNXJY1l5_N[J^1k5YNZJf1l5PNZJm1l5iM[JT2l5bMZJ\\2k5ZM[Jd2k5SM[Jj2P9M2O2M3hMZDQ1h;`NjD\\1a<hNbBe0S>I7J6J6JWgY2"}, "label": "the child brushing their teeth"}]}
{"id": 315383, "image": "COCO_train2014_000000315383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown color chair\"."}], "task": "refering", "answer_template": "The \"a brown color chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 316, 317, 318, 319, 339, 340, 341, 342, 362, 363, 364, 365, 385, 386, 387, 388, 408, 409], "bbox": [0.74753125, 0.5771539961013645, 0.963671875, 0.988382066276803], "iscrowd": 0, "area": 14814.27775, "rle": {"size": [513, 640], "counts": "en_71j?6I7J6J6I7J6J6I8I6J6J6I7J6J6I7J6J7H7J6J6I7J6J6J6I7J7I6M3000O1K5K5K5K5K5L4K5K5K5K4O21N10001N100O2O000O101N100O2O000O2O0O1M3M4L3O101O00001O00001O0000001O00001N1O1O2N1O00001OYNUEYNk:g1YEVNf:i1_ESNa:m1cEPN\\:P2hElMX:S2mEiMS:W2QFfMo9X2UFeMl9Z2^F]Mb9b2iFSMX9k2P2O1N2O1N2N2N2N2N2O1N1O2N2N2N2N2N2O1N2N2N2N3M3M2N3N2M3M2N3M3M2M4M3M2M4M3L3N3L4M2N3L4MRW;"}, "label": "a brown color chair"}]}
{"id": 315383, "image": "COCO_train2014_000000315383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown chair the person is sitting on\"."}], "task": "refering", "answer_template": "The \"the brown chair the person is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 316, 317, 318, 319, 339, 340, 341, 342, 362, 363, 364, 365, 385, 386, 387, 388, 408, 409], "bbox": [0.74753125, 0.5771539961013645, 0.963671875, 0.988382066276803], "iscrowd": 0, "area": 14814.27775, "rle": {"size": [513, 640], "counts": "en_71j?6I7J6J6I7J6J6I8I6J6J6I7J6J6I7J6J7H7J6J6I7J6J6J6I7J7I6M3000O1K5K5K5K5K5L4K5K5K5K4O21N10001N100O2O000O101N100O2O000O2O0O1M3M4L3O101O00001O00001O0000001O00001N1O1O2N1O00001OYNUEYNk:g1YEVNf:i1_ESNa:m1cEPN\\:P2hElMX:S2mEiMS:W2QFfMo9X2UFeMl9Z2^F]Mb9b2iFSMX9k2P2O1N2O1N2N2N2N2N2O1N1O2N2N2N2N2N2O1N2N2N2N3M3M2N3N2M3M2N3M3M2M4M3M2M4M3L3N3L4M2N3L4MRW;"}, "label": "the brown chair the person is sitting on"}]}
{"id": 225274, "image": "COCO_train2014_000000225274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark brown pony\"."}], "task": "refering", "answer_template": "The \"dark brown pony\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 214, 215, 216, 217, 229, 230, 231, 232, 244, 245, 246, 247, 259, 260, 261, 262, 274, 275, 276, 277, 289, 290, 291, 292, 304, 305, 308, 320], "bbox": [0.26946135831381735, 0.563359375, 0.549016393442623, 0.938640625], "iscrowd": 0, "area": 12975.975799999997, "rle": {"size": [640, 427], "counts": "cXX2;dc0e0ZO7J6J5J7J5J7J5K]1bNf1[N001N101O001O0010O01O010O0010O0SMeA7Z>4iA^OX>l0kAgNT>`0iA]N7e0P>T1PBUN49k=h1WBnM1Li=Z2^BgMMBd=m2eB`MQ>e2WBYM_=l2gBmLT=Y3SC`Lg<e3h15J6J7J5J7I:Gb0]Oc0]Oc0^Ob0^OO1M3D[CZIi<_6`C[Ic<^6c0J6J7I6J6I7J7I6]MTALR?2[A]Ok>a0cAmNd>Q1hA]N_>a1oAmMW>Q2m1O2M2N22N4M3L4L4L4M3L4L4L4M3L5K4O13M2N3M2N3M2N3M2N3N5Ja0_O`0@8G1M2O2N1N2O2N1N3N1O2M2O1N3N1O2M2O2N1N2O2N1N3N1OQRh3"}, "label": "dark brown pony"}]}
{"id": 225274, "image": "COCO_train2014_000000225274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small and underweight brown horse\"."}], "task": "refering", "answer_template": "The \"a small and underweight brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 214, 215, 216, 217, 229, 230, 231, 232, 244, 245, 246, 247, 259, 260, 261, 262, 274, 275, 276, 277, 289, 290, 291, 292, 304, 305, 308, 320], "bbox": [0.26946135831381735, 0.563359375, 0.549016393442623, 0.938640625], "iscrowd": 0, "area": 12975.975799999997, "rle": {"size": [640, 427], "counts": "cXX2;dc0e0ZO7J6J5J7J5J7J5K]1bNf1[N001N101O001O0010O01O010O0010O0SMeA7Z>4iA^OX>l0kAgNT>`0iA]N7e0P>T1PBUN49k=h1WBnM1Li=Z2^BgMMBd=m2eB`MQ>e2WBYM_=l2gBmLT=Y3SC`Lg<e3h15J6J7J5J7I:Gb0]Oc0]Oc0^Ob0^OO1M3D[CZIi<_6`C[Ic<^6c0J6J7I6J6I7J7I6]MTALR?2[A]Ok>a0cAmNd>Q1hA]N_>a1oAmMW>Q2m1O2M2N22N4M3L4L4L4M3L4L4L4M3L5K4O13M2N3M2N3M2N3M2N3N5Ja0_O`0@8G1M2O2N1N2O2N1N3N1O2M2O1N3N1O2M2O2N1N2O2N1N3N1OQRh3"}, "label": "a small and underweight brown horse"}]}
{"id": 28674, "image": "COCO_train2014_000000028674.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of an empty plastic office chair at the table opposite where a man sits using a laptop\"."}], "task": "refering", "answer_template": "The \"the back of an empty plastic office chair at the table opposite where a man sits using a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322, 323, 345, 346, 368, 369], "bbox": [0.0, 0.6316041666666667, 0.164515625, 1.0], "iscrowd": 0, "area": 10191.327050000002, "rle": {"size": [480, 640], "counts": "R:k4U:1N100O100O100O100O100O100O101N1O100O100000000O100000001N1000000O1000L4L4L3M4M3L4L4L3M4M3E;O1M3M3L3AfDWM];h2fDRM^;m2;O1N2M3_OlCSNV<n1jCoMY<P2iCkM[<T29XOYCWOg<i0ZCUOg<k0ZCROh<n0YCnNj<R1WChNm<X1UCaNQ=_1800000000O0100000N2L4K5L4K6K4O10000O1000000O100000000O2O00001N10001O001N3N3MW_j7"}, "label": "the back of an empty plastic office chair at the table opposite where a man sits using a laptop"}]}
{"id": 176138, "image": "COCO_train2014_000000176138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich cut in wedges to the left of the fries\"."}], "task": "refering", "answer_template": "The \"the sandwich cut in wedges to the left of the fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 288, 289], "bbox": [0.15192488262910797, 0.62021875, 0.48953051643192486, 0.8292187500000001], "iscrowd": 0, "area": 12524.6424, "rle": {"size": [640, 426], "counts": "gSY13ic05J5L4C>nNQ100O2N100O2N100O2O0O1O2O0m^OfM_`0[2`_OfM_`0\\2__OeMa`0[2^_OfMa`0[2^_OfMb`0[2\\_OfMc`0[2Y_OaMJ5l`0\\2Y_OgMg`0Y2X_OhMg`0Y2X_OhMg`0Z2W_OgMi`0Y2V_OhMi`0Y2W_OgMh`0h210O10_OX_OjMg`0h2O100O1O\\O\\_OnMd`0Q2\\_OPNc`0P2]_OQNb`0g20XO^_ORNa`0n1__OSNa`0l1__OUN``0k1`_OVN_`0j1a_OWN_`0i1a_OWN^`0i1b_OXN]`0h1c_OYN]`0f1c_O[N\\`0e1e_O[NZ`0e1f_O\\NZ`0d1e_O]NZ`0c1f_O^NZ`0a1f_O`NY`0`1h_O`NW`0`1i_OaNW`0^1i_OcNV`0]1j_OdNU`0]1k_OcNU`0\\1k_OeNT`0[1l_OfNT`0Y1l_OhNS`0X1m_OiNR`0W1o_OiNQ`0W1n_OjNQ`0V1o_OkNP`0U1P@lNP`0S1Q@mNn?S1R@nNm?R1S@oNm?Q1R@oNn?Q1R@POn?o0S@QOl?o0T@ROk?n0U@SOk?l0U@UOj?k0W@UOh?l0W@UOi?j0W@WOi?h0W@YOj?e0V@\\Oj?c0W@]Oi?b0W@_Oi?a0V@@k?>U@Ck?<V@Dj?;V@Fk?8U@Ik?6U@Kk?5T@Lm?2T@Nl?3R@Nn?R2001O00001O004L4L5K4L4L4L4L4L6J7I7I7WNS^OQ1Tb0hNn]OV1Yb0cNh]O[1ab0N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2NbgW4"}, "label": "the sandwich cut in wedges to the left of the fries"}]}
{"id": 176138, "image": "COCO_train2014_000000176138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich on the plate\"."}], "task": "refering", "answer_template": "The \"the sandwich on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 288, 289], "bbox": [0.15192488262910797, 0.62021875, 0.48953051643192486, 0.8292187500000001], "iscrowd": 0, "area": 12524.6424, "rle": {"size": [640, 426], "counts": "gSY13ic05J5L4C>nNQ100O2N100O2N100O2O0O1O2O0m^OfM_`0[2`_OfM_`0\\2__OeMa`0[2^_OfMa`0[2^_OfMb`0[2\\_OfMc`0[2Y_OaMJ5l`0\\2Y_OgMg`0Y2X_OhMg`0Y2X_OhMg`0Z2W_OgMi`0Y2V_OhMi`0Y2W_OgMh`0h210O10_OX_OjMg`0h2O100O1O\\O\\_OnMd`0Q2\\_OPNc`0P2]_OQNb`0g20XO^_ORNa`0n1__OSNa`0l1__OUN``0k1`_OVN_`0j1a_OWN_`0i1a_OWN^`0i1b_OXN]`0h1c_OYN]`0f1c_O[N\\`0e1e_O[NZ`0e1f_O\\NZ`0d1e_O]NZ`0c1f_O^NZ`0a1f_O`NY`0`1h_O`NW`0`1i_OaNW`0^1i_OcNV`0]1j_OdNU`0]1k_OcNU`0\\1k_OeNT`0[1l_OfNT`0Y1l_OhNS`0X1m_OiNR`0W1o_OiNQ`0W1n_OjNQ`0V1o_OkNP`0U1P@lNP`0S1Q@mNn?S1R@nNm?R1S@oNm?Q1R@oNn?Q1R@POn?o0S@QOl?o0T@ROk?n0U@SOk?l0U@UOj?k0W@UOh?l0W@UOi?j0W@WOi?h0W@YOj?e0V@\\Oj?c0W@]Oi?b0W@_Oi?a0V@@k?>U@Ck?<V@Dj?;V@Fk?8U@Ik?6U@Kk?5T@Lm?2T@Nl?3R@Nn?R2001O00001O004L4L5K4L4L4L4L4L6J7I7I7WNS^OQ1Tb0hNn]OV1Yb0cNh]O[1ab0N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2NbgW4"}, "label": "the sandwich on the plate"}]}
{"id": 561163, "image": "COCO_train2014_000000561163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass of wine between fork and plate\"."}], "task": "refering", "answer_template": "The \"glass of wine between fork and plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 142, 143, 165, 166, 167, 188, 189, 190], "bbox": [0.13309375, 0.1879625292740047, 0.291671875, 0.5926932084309133], "iscrowd": 0, "area": 9583.026900000003, "rle": {"size": [427, 640], "counts": "RbS16c<d0E;M3N2M3N2M3M3N2M3N2M3N2M2O1N3N1N2N2O2M2O1O1O2N1O1O1O1N1000001O0000000O2O0ZG^Mc6b2UIgMj6Y2UIiMj6X2TIjMk6V2UIjMk6V2TIlMk6T2TInMj6S2UInMk6S2SIoMl6Q2SIQNl6o1SIRNm6n1SIRNm6o1QISNn6P2nHQNR7R2jHoMU7V2gHjMY7Z2bHhM]7[2_HfMa7^2[HbMe7b2VH_Mj7e2QH]Mn7P400O100010O001O0aMSH@n7>WH]Oj7c0YHZOh7d0]HXOc7g0bHUO_7i0eHTO[7k0jHQOW7m0mHoNT7Q1lHoNU7o0mHPOS7o0nHQOS7m0nHSOR7l0PIROR7l0oHROW7g0lHWOY7d0jHZOZ7`0iH^O\\7<gHB_78dHF`74cHJb70bHMd7M^H2i7EZH9^:M2N3N1N3M2O1N3M2O2M2N`Sm5"}, "label": "glass of wine between fork and plate"}]}
{"id": 561163, "image": "COCO_train2014_000000561163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of red wine on a rush placemat behind a plate of food and to the left of another glass of wine\"."}], "task": "refering", "answer_template": "The \"a glass of red wine on a rush placemat behind a plate of food and to the left of another glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 142, 143, 165, 166, 167, 188, 189, 190], "bbox": [0.13309375, 0.1879625292740047, 0.291671875, 0.5926932084309133], "iscrowd": 0, "area": 9583.026900000003, "rle": {"size": [427, 640], "counts": "RbS16c<d0E;M3N2M3N2M3M3N2M3N2M3N2M2O1N3N1N2N2O2M2O1O1O2N1O1O1O1N1000001O0000000O2O0ZG^Mc6b2UIgMj6Y2UIiMj6X2TIjMk6V2UIjMk6V2TIlMk6T2TInMj6S2UInMk6S2SIoMl6Q2SIQNl6o1SIRNm6n1SIRNm6o1QISNn6P2nHQNR7R2jHoMU7V2gHjMY7Z2bHhM]7[2_HfMa7^2[HbMe7b2VH_Mj7e2QH]Mn7P400O100010O001O0aMSH@n7>WH]Oj7c0YHZOh7d0]HXOc7g0bHUO_7i0eHTO[7k0jHQOW7m0mHoNT7Q1lHoNU7o0mHPOS7o0nHQOS7m0nHSOR7l0PIROR7l0oHROW7g0lHWOY7d0jHZOZ7`0iH^O\\7<gHB_78dHF`74cHJb70bHMd7M^H2i7EZH9^:M2N3N1N3M2O1N3M2O2M2N`Sm5"}, "label": "a glass of red wine on a rush placemat behind a plate of food and to the left of another glass of wine"}]}
{"id": 389137, "image": "COCO_train2014_000000389137.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black color shirt with cigarette walking on the pavement\"."}], "task": "refering", "answer_template": "The \"a man with black color shirt with cigarette walking on the pavement\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 345, 346, 347, 368, 369, 370], "bbox": [0.0017031250000000002, 0.42022774327122153, 0.1390625, 0.9865217391304348], "iscrowd": 0, "area": 15888.6035, "rle": {"size": [483, 640], "counts": "Qj0^1o:j2oMQ2L5L3L4L4M3L5K4M3L4M3L5K4M3L4L4M4K4L0j0VO<D001O001O1O001O001O001O001O1O001O001N101O001O1O001O1O1O1O1O1O1O001O1O1O1O1lN_GdKb8V4gGfKZ8T4oGhKR8R4WHjKj7P4_HlKb7n3gHmK[7m3nHoKS7k3V2I7K401O1O1O1O1O1O5K9H8G9G9G9G9G9G9G9G:F9G]`S8"}, "label": "a man with black color shirt with cigarette walking on the pavement"}]}
{"id": 315411, "image": "COCO_train2014_000000315411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown wooden spoon with its handle carved like and animal face looking upwards\"."}], "task": "refering", "answer_template": "The \"a brown wooden spoon with its handle carved like and animal face looking upwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 234, 235, 250, 251, 266, 267, 282, 283, 284, 298, 299, 300, 314, 315, 316, 331, 332, 333, 347, 348, 349, 364, 365, 366, 381], "bbox": [0.43335416666666665, 0.549421875, 0.8768125, 0.980890625], "iscrowd": 0, "area": 10528.173400000003, "rle": {"size": [640, 480], "counts": "`cR43fc07I6I8I6J7L3O2N1N2O2N1N2O1O2M2O1O2N1N2O2N1N2O2N1O1O2N1O1O1O2N1O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2N1O100001N1000001O000O2O0000001O000O2O000O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1M3N2N2M3N2M3N2N2M3N2N2N20001O0O1000000O10000O1000000O1000000O10000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000kN]On^Oc0Sa0_Oj^O`0Xa0Be^O>\\a0Cb^O=_a0E]^O<da0EY^O<ha0FT^O:na0Go]O:Sb0Gi]O:Xb0Ge]O:\\b0c02O1O1O1O1N3N1O1O1N2O10O01O00001O1O1O001O100O001O1N2O001O1N2O1O1N101O1M3N2N1O2N2N2MSUU1"}, "label": "a brown wooden spoon with its handle carved like and animal face looking upwards"}]}
{"id": 315411, "image": "COCO_train2014_000000315411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top spoon in the salad bowl\"."}], "task": "refering", "answer_template": "The \"the top spoon in the salad bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 248, 249, 250, 265, 266, 281, 282, 297, 298, 312, 313, 314, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 360, 361, 362, 363, 364, 377, 378, 379, 380, 381], "bbox": [0.19908333333333333, 0.590953125, 0.7367708333333333, 1.0], "iscrowd": 0, "area": 15237.423700000003, "rle": {"size": [640, 480], "counts": "ebl1k0Qc09J7I6I2O1N2O1O1N2O1O1O1O1N2O1O1O1O1O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O100O10O0100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O010O100O100O100O100O10O001O001O00001O00001N10001O00001M2N2M3N3L3N2M4M2N2M4M2M3L5K4M3N2N3N1O1O2N1N101O1O1O1O1O1O1O100O1O1O1O10O01N2O1N2O1N2O1N2O10000O10O100000O1000000O1000000O1000O1000O1000000O10000000O01000000O0100000O10O100000O10O1000O1000O1000O1000O10O100000O010000000O01000O10O0100O100O010O1000O2O1O1O2M2O10000O10000O10000O101N100O100O2O000F]\\^2"}, "label": "the top spoon in the salad bowl"}]}
{"id": 315411, "image": "COCO_train2014_000000315411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden spoon turned upside down in a bowl\"."}], "task": "refering", "answer_template": "The \"a wooden spoon turned upside down in a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 248, 249, 250, 265, 266, 281, 282, 297, 298, 312, 313, 314, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 360, 361, 362, 363, 364, 377, 378, 379, 380, 381], "bbox": [0.19908333333333333, 0.590953125, 0.7367708333333333, 1.0], "iscrowd": 0, "area": 15237.423700000003, "rle": {"size": [640, 480], "counts": "ebl1k0Qc09J7I6I2O1N2O1O1N2O1O1O1O1N2O1O1O1O1O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O100O10O0100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O010O100O100O100O100O10O001O001O00001O00001N10001O00001M2N2M3N3L3N2M4M2N2M4M2M3L5K4M3N2N3N1O1O2N1N101O1O1O1O1O1O1O100O1O1O1O10O01N2O1N2O1N2O1N2O10000O10O100000O1000000O1000000O1000O1000O1000000O10000000O01000000O0100000O10O100000O10O1000O1000O1000O1000O10O100000O010000000O01000O10O0100O100O010O1000O2O1O1O2M2O10000O10000O10000O101N100O100O2O000F]\\^2"}, "label": "a wooden spoon turned upside down in a bowl"}]}
{"id": 438292, "image": "COCO_train2014_000000438292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a car tie\"."}], "task": "refering", "answer_template": "The \"a man with a car tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 352, 353, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.068125, 0.393953125, 0.9978125], "iscrowd": 0, "area": 76197.70225000003, "rle": {"size": [480, 640], "counts": "g99U>c0B=M4L3M4L4L3M4L3M3N2N2O1N2N2O2M2N2O1N2N2O1N2N2O1M3M3N2M3N3L3N2M3M3N2N2M3N2M3N2N2M3nIhKZ1Z4_NmK^1U4[NSLd1n3UNYLj1h3oM_LP2b3lMbLR2`3jMdLU2]3hMfLW2[3eMiLY2Y3cMkL\\2V3bMlL]2U3`MnL_2S3_MoL_2S3^MPMa2Q3]MQMc2o2[MSMd2m2ZMVMe2k2XMXMg2i2VMZMi2g2TM\\Mk2e2jKPK6^2P4b2fKUK6\\2S4`2bKYK9Y2T4c3iK_LV4c3gK_LX4c3fK^LZ4c3cK_L]4b3`K`L_4a3_KaLa4`3\\KbLd4_3YKcLg4^3VKdLi4^3TKdLl4]3QKeLo4f6O10000O1000000O10000O10000O10000O10000O10000O10000O10000O101O00000O1000000O100000000O100000000O10000O100O10000O100O1oMSFoNm9Q1TFmNm9R1UFlNl9S1VFkNk9T1WFjNj9U1WFkNi9T1YFjNh9Q1^FmNc9o0bFnN`9m0fFPO\\9k0iFTOX9i0lFTOV9j0]GcNe8[1a2N2N2O1O10000O10000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000010O001O1O1O1O001O1O1O1O0eEfNP7[1mHhNR7Y1kHjNT7W1jHkNU7U1jHmNU7T1hHPOV7Q1hHQOW7P1gHROX7o0fHSOY7m0fHUOY7l0cHXO\\7i0`H[O_7f0]H^Ob7c0ZHAe7d0SH@l7e0lG_OS8g0cG^O\\8R1QGROn8\\1`FiN_9]3O1O1O1O1O1O001O1O1O001O1O1VMdKnJ\\4k4\\LdJe3^1bJR1e8hNcGEn95ZF_NP;[1YEZNo:`1[1G:D<_Oefe5"}, "label": "a man with a car tie"}]}
{"id": 438292, "image": "COCO_train2014_000000438292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in glasses\"."}], "task": "refering", "answer_template": "The \"a man in glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 352, 353, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.068125, 0.393953125, 0.9978125], "iscrowd": 0, "area": 76197.70225000003, "rle": {"size": [480, 640], "counts": "g99U>c0B=M4L3M4L4L3M4L3M3N2N2O1N2N2O2M2N2O1N2N2O1N2N2O1M3M3N2M3N3L3N2M3M3N2N2M3N2M3N2N2M3nIhKZ1Z4_NmK^1U4[NSLd1n3UNYLj1h3oM_LP2b3lMbLR2`3jMdLU2]3hMfLW2[3eMiLY2Y3cMkL\\2V3bMlL]2U3`MnL_2S3_MoL_2S3^MPMa2Q3]MQMc2o2[MSMd2m2ZMVMe2k2XMXMg2i2VMZMi2g2TM\\Mk2e2jKPK6^2P4b2fKUK6\\2S4`2bKYK9Y2T4c3iK_LV4c3gK_LX4c3fK^LZ4c3cK_L]4b3`K`L_4a3_KaLa4`3\\KbLd4_3YKcLg4^3VKdLi4^3TKdLl4]3QKeLo4f6O10000O1000000O10000O10000O10000O10000O10000O10000O10000O101O00000O1000000O100000000O100000000O10000O100O10000O100O1oMSFoNm9Q1TFmNm9R1UFlNl9S1VFkNk9T1WFjNj9U1WFkNi9T1YFjNh9Q1^FmNc9o0bFnN`9m0fFPO\\9k0iFTOX9i0lFTOV9j0]GcNe8[1a2N2N2O1O10000O10000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000010O001O1O1O1O001O1O1O1O0eEfNP7[1mHhNR7Y1kHjNT7W1jHkNU7U1jHmNU7T1hHPOV7Q1hHQOW7P1gHROX7o0fHSOY7m0fHUOY7l0cHXO\\7i0`H[O_7f0]H^Ob7c0ZHAe7d0SH@l7e0lG_OS8g0cG^O\\8R1QGROn8\\1`FiN_9]3O1O1O1O1O1O001O1O1O001O1O1VMdKnJ\\4k4\\LdJe3^1bJR1e8hNcGEn95ZF_NP;[1YEZNo:`1[1G:D<_Oefe5"}, "label": "a man in glasses"}]}
{"id": 438292, "image": "COCO_train2014_000000438292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman stands next to a man and smiles\"."}], "task": "refering", "answer_template": "The \"a woman stands next to a man and smiles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.3733125, 0.09910416666666667, 0.6638437500000001, 1.0], "iscrowd": 0, "area": 63500.23474999999, "rle": {"size": [480, 640], "counts": "\\^`36b>=D<D=jJYO`KS17YN_OQ1mNT2Z1VMZOR1lMZ3a2nKUOT1mMY3g2mKTNTOXOk1A`3l2gKVNROZOm1WOe3R3bKZNPOYOh7W2]I]NnNXOk7U2\\I`NlNWOl7S2]IeNiNTOm7S2^IgNgNSOm7R2aIjNdNoNn7S2cIlNX7R1lHlNT7S1nHkNS7T1nHlNR7T1nHkNS7T1nHlNR7S1PIkNQ7T1PIlNP7T1PIkNQ7T1PIlNP7S1RIkNo6U1QIkNo6T1RIkNo6T1RIlNn6S1TIkNm6U1SIkNm6T1TIkNm6T1TIkNm6T1TIkNm6U1TIiNm6V1TIhNn6W1SIhNn6X1UIdNl6[1YI_Ni6`1\\3O100O10000000000000000000000000000001O001O00001O001O00001O001O001O1O001O1O1O001O1O1O1O001O1O1O1O001O1O4L6J7I7I8oDnLd9X4O2N1O2N2N1O2N00001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K5K4L5K5K5K6J5K5bL`KRLe4X3RLbLT4R3WLiLn3k2^LPMg3e2eLUMa3_2jL\\M[3Z2n4E:\\OQ1mNj`T3"}, "label": "a woman stands next to a man and smiles"}]}
{"id": 438292, "image": "COCO_train2014_000000438292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in a sweater standing next to a man\"."}], "task": "refering", "answer_template": "The \"the woman in a sweater standing next to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.3733125, 0.09910416666666667, 0.6638437500000001, 1.0], "iscrowd": 0, "area": 63500.23474999999, "rle": {"size": [480, 640], "counts": "\\^`36b>=D<D=jJYO`KS17YN_OQ1mNT2Z1VMZOR1lMZ3a2nKUOT1mMY3g2mKTNTOXOk1A`3l2gKVNROZOm1WOe3R3bKZNPOYOh7W2]I]NnNXOk7U2\\I`NlNWOl7S2]IeNiNTOm7S2^IgNgNSOm7R2aIjNdNoNn7S2cIlNX7R1lHlNT7S1nHkNS7T1nHlNR7T1nHkNS7T1nHlNR7S1PIkNQ7T1PIlNP7T1PIkNQ7T1PIlNP7S1RIkNo6U1QIkNo6T1RIkNo6T1RIlNn6S1TIkNm6U1SIkNm6T1TIkNm6T1TIkNm6T1TIkNm6U1TIiNm6V1TIhNn6W1SIhNn6X1UIdNl6[1YI_Ni6`1\\3O100O10000000000000000000000000000001O001O00001O001O00001O001O001O1O001O1O1O001O1O1O1O001O1O1O1O001O1O4L6J7I7I8oDnLd9X4O2N1O2N2N1O2N00001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K5K4L5K5K5K6J5K5bL`KRLe4X3RLbLT4R3WLiLn3k2^LPMg3e2eLUMa3_2jL\\M[3Z2n4E:\\OQ1mNj`T3"}, "label": "the woman in a sweater standing next to a man"}]}
{"id": 380949, "image": "COCO_train2014_000000380949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant lying on it ' s side\"."}], "task": "refering", "answer_template": "The \"a baby elephant lying on it ' s side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 336, 337], "bbox": [0.287015625, 0.13625292740046838, 0.888078125, 0.9696018735362997], "iscrowd": 0, "area": 65461.37465000002, "rle": {"size": [427, 640], "counts": "]o\\2f0e<6J=C1O1O2N1O1O2N1O2N2N2N001O001O001O001O01O001O00001O0016I9G8I0O010O1OO2O001O000O2O001O000O2O001O0O101N101N100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O10O0100O10XEbNc9]1YFjNd9W1VFQOh9n0TFYOj9g0PFBm9>PFGn98SFJj97WFJg96YFLe94[FNc92]F1`9O`F3^9McF4Z9MfF4Y9KhF5X9KhF5X9KiF4W9LiF4V9MjF3V9MjF3V9MjF3V9NjF1V9OjF1U91jFOV91jFOV91jFOV92jFMV93jFMU95jFJW96iFJW97hFIW98jFGV99jFGV9:iFFW9:iFFV9<iFDW9<jFCV9>iFBW9>iFBV9?jFAV9`0iF@W9`0jF_OU9c0jF]OV9c0jF]OV9d0iF\\OW9d0iF\\OV9e0kFZOU9g0jFYOV9g0jFYOU9i0jFWOV9i0jFWOV9j0jFUOV9k0jFUOU9l0kFTOU9m0jFSOV9m0jFSOU9o0kFoNV9Q1jFoNV9R1iFnNW9R1iFnNV9S1jFmNV9T1jFkNV9U1jFkNU9W1jFiNV9W1jFiNV9X1iFhNV9Y1kFfNU9[1jFeNV9[1jFeNV9[1jFeNU9]1jFcNV9]1kFbNU9^1kFbNT9_1lFaNT9`1kF`NU9`1kF`NT9a1mF^NS9b1mF^Ni4nNXOd2oK^Ne4[OROY2XL\\Na4HoNl1`L\\N]42mNb1gL[NZ48lN]1iL\\NZ4;iNZ1mL[NZ4>fNX1oLZNZ4b0dNT1RMYNZ4g0`NQ1VMXNY4k0^Nm0YMXNY4n0[Nj0\\MXNX4Q1YNP2f1SNXNm1h1TNWNl1h1VNWNj1h1XNVNi1j1XNUNi1i1YNVNh1h1ZNVNh1h1ZNWNf1i1[NVNf1h1\\NVNf1h1\\NWNe1h1\\NWNe1g1]NWNd1h1^NWNc1h1^NWNc1g1`NVNb1h1`NWN`1h1bNVN`1i1aNVN`1h1bNVN_1i1cNVN^1i1cNUN^1j1dNUN\\1j1fNUN[1j1fNTN[1k1VLhL[1\\1_2k1VLjL\\1Y1`2k1TLmL]1W1_2l1SLnL_1T1_2m1RLPM`1R1_2l1QLSMa1P1^2m1PLTMc1m0_2m1nKWMe1j0]2n1nKYMf1g0]2P2lKZMh1e0]2o1kK^Mh1a0^2P2jK`Mi1?]2P2jKbMk1;]2R2gKdMm19\\2R2gKfMn17[2R2gKhMo14\\2S2dKjMQ22[2S2eKkMR2O[2T2cKnMS2MZ2U2bKoMU2JZ2W2`KPNW2G[2W2^KTNX2B[2Z2\\KUNZ2_O[2[2[KWN[2\\O\\2\\2XKYN]2YO\\2]2WK[N^2VO\\2_2UK\\NX8b1hG_NX8a1gG`NZ8^1fGcNZ8]1eGdN[8[1eGgN[8X1dGiN\\8V1dGkN\\8U1cGlN^8T1`GmN`8S1_GnNa8S1]GnNd8R1ZGoNf8Q1YGPOf8R1XGoNg8R1YGoNf8R1XGoNg8S1WGnNh8T1VGmNj8T1TGmNk8U1SGlNm8U1QGlNn8i2O100O2N100O2N100O2N1N2O2M2N2O2O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00010O00001O002N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1N2O1N2O2N1N2O1N2O2M2O1N2O2N1N2O1N2M4K4L4M3L5L3L4L4M3L5L3L4L4M3L5@?iLPGW1k9hNgF6j9IgFB^9=j1M2O1O2N2N2N2M2O2N2N2N2Neim0"}, "label": "a baby elephant lying on it ' s side"}]}
{"id": 380949, "image": "COCO_train2014_000000380949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant that is on its side and has its legs in the air\"."}], "task": "refering", "answer_template": "The \"a baby elephant that is on its side and has its legs in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 336, 337], "bbox": [0.287015625, 0.13625292740046838, 0.888078125, 0.9696018735362997], "iscrowd": 0, "area": 65461.37465000002, "rle": {"size": [427, 640], "counts": "]o\\2f0e<6J=C1O1O2N1O1O2N1O2N2N2N001O001O001O001O01O001O00001O0016I9G8I0O010O1OO2O001O000O2O001O000O2O001O0O101N101N100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O10O0100O10XEbNc9]1YFjNd9W1VFQOh9n0TFYOj9g0PFBm9>PFGn98SFJj97WFJg96YFLe94[FNc92]F1`9O`F3^9McF4Z9MfF4Y9KhF5X9KhF5X9KiF4W9LiF4V9MjF3V9MjF3V9MjF3V9NjF1V9OjF1U91jFOV91jFOV91jFOV92jFMV93jFMU95jFJW96iFJW97hFIW98jFGV99jFGV9:iFFW9:iFFV9<iFDW9<jFCV9>iFBW9>iFBV9?jFAV9`0iF@W9`0jF_OU9c0jF]OV9c0jF]OV9d0iF\\OW9d0iF\\OV9e0kFZOU9g0jFYOV9g0jFYOU9i0jFWOV9i0jFWOV9j0jFUOV9k0jFUOU9l0kFTOU9m0jFSOV9m0jFSOU9o0kFoNV9Q1jFoNV9R1iFnNW9R1iFnNV9S1jFmNV9T1jFkNV9U1jFkNU9W1jFiNV9W1jFiNV9X1iFhNV9Y1kFfNU9[1jFeNV9[1jFeNV9[1jFeNU9]1jFcNV9]1kFbNU9^1kFbNT9_1lFaNT9`1kF`NU9`1kF`NT9a1mF^NS9b1mF^Ni4nNXOd2oK^Ne4[OROY2XL\\Na4HoNl1`L\\N]42mNb1gL[NZ48lN]1iL\\NZ4;iNZ1mL[NZ4>fNX1oLZNZ4b0dNT1RMYNZ4g0`NQ1VMXNY4k0^Nm0YMXNY4n0[Nj0\\MXNX4Q1YNP2f1SNXNm1h1TNWNl1h1VNWNj1h1XNVNi1j1XNUNi1i1YNVNh1h1ZNVNh1h1ZNWNf1i1[NVNf1h1\\NVNf1h1\\NWNe1h1\\NWNe1g1]NWNd1h1^NWNc1h1^NWNc1g1`NVNb1h1`NWN`1h1bNVN`1i1aNVN`1h1bNVN_1i1cNVN^1i1cNUN^1j1dNUN\\1j1fNUN[1j1fNTN[1k1VLhL[1\\1_2k1VLjL\\1Y1`2k1TLmL]1W1_2l1SLnL_1T1_2m1RLPM`1R1_2l1QLSMa1P1^2m1PLTMc1m0_2m1nKWMe1j0]2n1nKYMf1g0]2P2lKZMh1e0]2o1kK^Mh1a0^2P2jK`Mi1?]2P2jKbMk1;]2R2gKdMm19\\2R2gKfMn17[2R2gKhMo14\\2S2dKjMQ22[2S2eKkMR2O[2T2cKnMS2MZ2U2bKoMU2JZ2W2`KPNW2G[2W2^KTNX2B[2Z2\\KUNZ2_O[2[2[KWN[2\\O\\2\\2XKYN]2YO\\2]2WK[N^2VO\\2_2UK\\NX8b1hG_NX8a1gG`NZ8^1fGcNZ8]1eGdN[8[1eGgN[8X1dGiN\\8V1dGkN\\8U1cGlN^8T1`GmN`8S1_GnNa8S1]GnNd8R1ZGoNf8Q1YGPOf8R1XGoNg8R1YGoNf8R1XGoNg8S1WGnNh8T1VGmNj8T1TGmNk8U1SGlNm8U1QGlNn8i2O100O2N100O2N100O2N1N2O2M2N2O2O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00010O00001O002N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1N2O1N2O2N1N2O1N2O2M2O1N2O2N1N2O1N2M4K4L4M3L5L3L4L4M3L5L3L4L4M3L5@?iLPGW1k9hNgF6j9IgFB^9=j1M2O1O2N2N2N2M2O2N2N2N2Neim0"}, "label": "a baby elephant that is on its side and has its legs in the air"}]}
{"id": 380949, "image": "COCO_train2014_000000380949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large elephant in the corner\"."}], "task": "refering", "answer_template": "The \"a large elephant in the corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 257, 258, 259, 276, 277, 278, 280, 281, 282, 299, 300, 301, 303, 304, 305, 327, 328], "bbox": [0.003, 0.0, 0.31784375, 0.9546838407494145], "iscrowd": 0, "area": 64703.623900000006, "rle": {"size": [427, 640], "counts": "hj0_7l5`4`K0000001O0000001O000000001O0000001O0000001O000000001O0000001O0O1000001O0000001O0000001O000000001O0000001O000000001OfKSOVLm0h3EhK;W45ZKKe4f0lJYOS5W1^JiNa5h1mIZNR6W2VIQNi6_2aHhM]7T4O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1:F<D<D;E<D<D<D;E<D<D;E<D<D<D;E6J1O000000001O000000001O0000001O000000001O0000001N100000001O000000001O0000001O000000001O0000001O000000001O000000]OFUD:_;3aDLT;`0kD@d:R1[EnNT:d1kE\\Nd9V2\\FiMT9h2kFXMd8Z3[GfLT8l3kGTLd7R5YOg0mNS1mNS1mNS1RNn1N2M3N2N2M3N2N2M3L6K5K5J6K5JQWf5"}, "label": "a large elephant in the corner"}]}
{"id": 380949, "image": "COCO_train2014_000000380949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"medium sized elephant\"."}], "task": "refering", "answer_template": "The \"medium sized elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 257, 258, 259, 276, 277, 278, 280, 281, 282, 299, 300, 301, 303, 304, 305, 327, 328], "bbox": [0.003, 0.0, 0.31784375, 0.9546838407494145], "iscrowd": 0, "area": 64703.623900000006, "rle": {"size": [427, 640], "counts": "hj0_7l5`4`K0000001O0000001O000000001O0000001O0000001O000000001O0000001O0O1000001O0000001O0000001O000000001O0000001O000000001OfKSOVLm0h3EhK;W45ZKKe4f0lJYOS5W1^JiNa5h1mIZNR6W2VIQNi6_2aHhM]7T4O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1:F<D<D;E<D<D<D;E<D<D;E<D<D<D;E6J1O000000001O000000001O0000001O000000001O0000001N100000001O000000001O0000001O000000001O0000001O000000001O000000]OFUD:_;3aDLT;`0kD@d:R1[EnNT:d1kE\\Nd9V2\\FiMT9h2kFXMd8Z3[GfLT8l3kGTLd7R5YOg0mNS1mNS1mNS1RNn1N2M3N2N2M3N2N2M3L6K5K5J6K5JQWf5"}, "label": "medium sized elephant"}]}
{"id": 299029, "image": "COCO_train2014_000000299029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"long blond hair wearing all black and a knee brace on left knee\"."}], "task": "refering", "answer_template": "The \"long blond hair wearing all black and a knee brace on left knee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 194, 195, 196, 217, 218, 219, 240, 241, 263, 264, 286, 287, 308, 309, 310, 331, 332, 333], "bbox": [0.40393749999999995, 0.3838958333333334, 0.565046875, 0.8810625], "iscrowd": 0, "area": 9863.6906, "rle": {"size": [480, 640], "counts": "_ii38f>8I7I2M2O1O2N1O1N2O2N1O1O1N2O1O1O2M2fMYOZFh0S9LkF5S9MlFLZ97dFF]9=aFB`9a0^F]Oc9f0[FWOg9k0XFSOS9RO\\Fo1?nNU9WOXFn1a0kNW9YOUFP2a0gNZ9[ORFR2a0bN]9_OnES2c0^N_9AkEU2c0ZNb9Z2[FfMd9]2ZFbMg9a2VF_Mj9d2SF\\Ml9g2RFXMo9k2nEUMR:n2kERMT:Q3jEoLV:l310O1000O010^1bN1N2O1O1N2O1O1O1000000000000000UOgGoJX8S1cGP2>jLo7S1gGm1c0lLf7U1jGh1j0nL]7X1kGGM>R1bNV7W1nGF44m0oNQ7U1QHE:Jk0[O]61SHR1?E?@i0GT69PHi0f0XODCR10e03m5`1lHiNT3Fo3`1PIhNS3Gk3`1VIhNo2Hi3_1[IiNm2Gg3^1`IiNk2He3]1cIjNb:V1_EiNb:V1aEgN_:Z1d1O10O010O01O1N101N2O0O2O1N2O1N2M3L4K5K6K4KhbR4"}, "label": "long blond hair wearing all black and a knee brace on left knee"}]}
{"id": 299029, "image": "COCO_train2014_000000299029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in dark clothing and white socks\"."}], "task": "refering", "answer_template": "The \"a woman in dark clothing and white socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 194, 195, 196, 217, 218, 219, 240, 241, 263, 264, 286, 287, 308, 309, 310, 331, 332, 333], "bbox": [0.40393749999999995, 0.3838958333333334, 0.565046875, 0.8810625], "iscrowd": 0, "area": 9863.6906, "rle": {"size": [480, 640], "counts": "_ii38f>8I7I2M2O1O2N1O1N2O2N1O1O1N2O1O1O2M2fMYOZFh0S9LkF5S9MlFLZ97dFF]9=aFB`9a0^F]Oc9f0[FWOg9k0XFSOS9RO\\Fo1?nNU9WOXFn1a0kNW9YOUFP2a0gNZ9[ORFR2a0bN]9_OnES2c0^N_9AkEU2c0ZNb9Z2[FfMd9]2ZFbMg9a2VF_Mj9d2SF\\Ml9g2RFXMo9k2nEUMR:n2kERMT:Q3jEoLV:l310O1000O010^1bN1N2O1O1N2O1O1O1000000000000000UOgGoJX8S1cGP2>jLo7S1gGm1c0lLf7U1jGh1j0nL]7X1kGGM>R1bNV7W1nGF44m0oNQ7U1QHE:Jk0[O]61SHR1?E?@i0GT69PHi0f0XODCR10e03m5`1lHiNT3Fo3`1PIhNS3Gk3`1VIhNo2Hi3_1[IiNm2Gg3^1`IiNk2He3]1cIjNb:V1_EiNb:V1aEgN_:Z1d1O10O010O01O1N101N2O0O2O1N2O1N2M3L4K5K6K4KhbR4"}, "label": "a woman in dark clothing and white socks"}]}
{"id": 299029, "image": "COCO_train2014_000000299029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a frisbee in his hand\"."}], "task": "refering", "answer_template": "The \"a man with a frisbee in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 191, 211, 212, 234, 235, 257, 258, 280, 281, 303, 304, 326, 327], "bbox": [0.15203125, 0.32093750000000004, 0.34459375, 0.87275], "iscrowd": 0, "area": 10552.203650000001, "rle": {"size": [480, 640], "counts": "fe]19g>2N2N2N2N1O2N2M3M3M3M2N3M2mNkNXDW1d;QOVDQ1e;YOQDl0j;]OoCf0m;_OPDc0n;_OPDc0n;^2iN=C<Dm0SO4M3gHPJ\\5S6aJPJ\\5S6`JPJ^5S6^JPJ`5R6]JPJb5R6[JoIe5S6XJnIi5S6TJnIl5T6QJmIo5U6nImIQ6U6lIlIT6V6iIkIW6W6gIiIY6_701O01O2N1O1O100O1O8H7I5K3N0O000_LcJ_M]5\\2jJbMV5Y2QKeMo4W2WKhMi43kH:a2Ae45mH4b2Ec46mHOe2I`47mHKg2L^48mHFi21[49UIZOd2<X49XMFj29WMEk2:UMBP3=j501O1N2O1O1N101O1N1000O10O01O10O0010O01O010O00010O0010O01O00010O001O01O01O0010O0001O01O01O001O00001O001O1M2N3M2NaUT6"}, "label": "a man with a frisbee in his hand"}]}
{"id": 299029, "image": "COCO_train2014_000000299029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black man holding a fribee\"."}], "task": "refering", "answer_template": "The \"a black man holding a fribee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 191, 211, 212, 234, 235, 257, 258, 280, 281, 303, 304, 326, 327], "bbox": [0.15203125, 0.32093750000000004, 0.34459375, 0.87275], "iscrowd": 0, "area": 10552.203650000001, "rle": {"size": [480, 640], "counts": "fe]19g>2N2N2N2N1O2N2M3M3M3M2N3M2mNkNXDW1d;QOVDQ1e;YOQDl0j;]OoCf0m;_OPDc0n;_OPDc0n;^2iN=C<Dm0SO4M3gHPJ\\5S6aJPJ\\5S6`JPJ^5S6^JPJ`5R6]JPJb5R6[JoIe5S6XJnIi5S6TJnIl5T6QJmIo5U6nImIQ6U6lIlIT6V6iIkIW6W6gIiIY6_701O01O2N1O1O100O1O8H7I5K3N0O000_LcJ_M]5\\2jJbMV5Y2QKeMo4W2WKhMi43kH:a2Ae45mH4b2Ec46mHOe2I`47mHKg2L^48mHFi21[49UIZOd2<X49XMFj29WMEk2:UMBP3=j501O1N2O1O1N101O1N1000O10O01O10O0010O01O010O00010O0010O01O00010O001O01O01O0010O0001O01O01O001O00001O001O1M2N3M2NaUT6"}, "label": "a black man holding a fribee"}]}
{"id": 127006, "image": "COCO_train2014_000000127006.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green metal bench sitting in water in front of a large patch of green plants apart from other benches\"."}], "task": "refering", "answer_template": "The \"green metal bench sitting in water in front of a large patch of green plants apart from other benches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297], "bbox": [0.61690625, 0.6555269320843092, 0.932609375, 0.8560187353629978], "iscrowd": 0, "area": 9491.153049999995, "rle": {"size": [427, 640], "counts": "aoT52Y=4L3M3N7H<D<D<D7I3M2O0OC=L400000001O0000000001O0001O0000000000001O000000000001O01O00000001O0001O000001O000000010O000000000010O0000000001O0001O000001O000001O0001O00000001O01O000000001O01O00000001O0001O000001O00000001O01O00000000010O000000000010O0000000001O01O000001O000001O0001O00000001O01O000000000010O0000000001O01O00000001O0001O01O1O1O16J6J6K3LO1O0001O0O10000O10O1O00100O001O10O100O2O0O1O2O1N2N3N1Nb0^Onba0"}, "label": "green metal bench sitting in water in front of a large patch of green plants apart from other benches"}]}
{"id": 127006, "image": "COCO_train2014_000000127006.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bench sitting to the right of two other benches sitting in water\"."}], "task": "refering", "answer_template": "The \"a green bench sitting to the right of two other benches sitting in water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297], "bbox": [0.61690625, 0.6555269320843092, 0.932609375, 0.8560187353629978], "iscrowd": 0, "area": 9491.153049999995, "rle": {"size": [427, 640], "counts": "aoT52Y=4L3M3N7H<D<D<D7I3M2O0OC=L400000001O0000000001O0001O0000000000001O000000000001O01O00000001O0001O000001O000000010O000000000010O0000000001O0001O000001O000001O0001O00000001O01O000000001O01O00000001O0001O000001O00000001O01O00000000010O000000000010O0000000001O01O000001O000001O0001O00000001O01O000000000010O0000000001O01O00000001O0001O01O1O1O16J6J6K3LO1O0001O0O10000O10O1O00100O001O10O100O2O0O1O2O1N2N3N1Nb0^Onba0"}, "label": "a green bench sitting to the right of two other benches sitting in water"}]}
{"id": 225313, "image": "COCO_train2014_000000225313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"semi truck parked on street\"."}], "task": "refering", "answer_template": "The \"semi truck parked on street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.7678281250000001, 0.7437793427230047, 0.9967500000000001, 1.0], "iscrowd": 0, "area": 11240.5068, "rle": {"size": [426, 640], "counts": "lR]64T=2N2N2N2N2N2N2N2N3N1N2N2N2N2N2N2N2O10000000000O1000000000000O1000000000000O10000oNQ1[Oe00000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000O10000000000O100000001O0O10000000000O10000000000O1000000002N3M2N2N2N2N2N2N2N2N3M3M4L3M4L3M2N1O1O1O1O1O1O1O1O1O1F:QOo?"}, "label": "semi truck parked on street"}]}
{"id": 225313, "image": "COCO_train2014_000000225313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a truck partially taken in a photo\"."}], "task": "refering", "answer_template": "The \"the back of a truck partially taken in a photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.7678281250000001, 0.7437793427230047, 0.9967500000000001, 1.0], "iscrowd": 0, "area": 11240.5068, "rle": {"size": [426, 640], "counts": "lR]64T=2N2N2N2N2N2N2N2N3N1N2N2N2N2N2N2N2O10000000000O1000000000000O1000000000000O10000oNQ1[Oe00000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000O10000000000O100000001O0O10000000000O10000000000O1000000002N3M2N2N2N2N2N2N2N2N3M3M4L3M4L3M2N1O1O1O1O1O1O1O1O1O1F:QOo?"}, "label": "the back of a truck partially taken in a photo"}]}
{"id": 528418, "image": "COCO_train2014_000000528418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goose\"."}], "task": "refering", "answer_template": "The \"a goose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 293, 294, 295], "bbox": [0.09044496487119437, 0.415953125, 0.9819906323185011, 0.853453125], "iscrowd": 0, "area": 73518.8975, "rle": {"size": [640, 427], "counts": "chh0g0Uc05YOh0F9F:I8K4L5L3L4M4K4M2M4M3L3N3L4N1O2M3N2N1O2N2N1O2N2N2M2O2N2N1O2N2N1O2N2M2O2N1O2N101O001O001N101N10001N101N101O0O2O0O2O0O2O001N101N101N101O0O101N101N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O001N101O001O001O0O2O001O001O001O1O001O001O001O00001O0010O01O0010O0001O010O001O00010O001O010O001O00010O0010O01O010O00010O001O010O001O010O001O000010O01O001O001O001O010O00001O001O001O01O0001O00001O00001O01O0001O00001O00001O00010O0000001O00001O000000000001O000000000001O000000000001O0000000000000O100O100O101N1O100O100O100aNPDmJQ<R5RDiJQ<V5RDdJR<[5PD_JV<_5mCXJZ<g5iCoI_<Q6h0O1O2N1O1O101N1O1O101N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O001O1N2O1QKUAg4l>VKWAi4Q?O001N10001O001N10001O00001N10001O001N10001O00001cN_@XNb?g1_@YNa?e1a@[N`?c1a@]N_?a1c@^N_?_1c@aN]?^1d@aN]?]1e@cN\\?[1e@dN\\?[1e@eN\\?X1g@fNZ?Y1g@gNY?X1h@hNY?V1h@iNY?U1i@kNX?S1i@lNX?S1i@mNX?P1k@nNV?Q1k@oNU?P1l@oNV?n0m@QOS?o0m@QOT?m0m@ROT?m0n@ROR?n0n@QOT?m0m@SOS?l0o@ROS?l0n@TOR?l0n@SOS?l0n@SOT?k0n@ROT?m0m@ROU?m0k@ROV?m0l@QOU?n0l@QOV?n0j@QOW?m0l@QOV?m0l@QOU?n0o@nNR?P1SAkNP?S1o1N3N1J;DXe4"}, "label": "a goose"}]}
{"id": 528418, "image": "COCO_train2014_000000528418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white duck or goose with its bill tucked under its wing\"."}], "task": "refering", "answer_template": "The \"a white duck or goose with its bill tucked under its wing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 293, 294, 295], "bbox": [0.09044496487119437, 0.415953125, 0.9819906323185011, 0.853453125], "iscrowd": 0, "area": 73518.8975, "rle": {"size": [640, 427], "counts": "chh0g0Uc05YOh0F9F:I8K4L5L3L4M4K4M2M4M3L3N3L4N1O2M3N2N1O2N2N1O2N2N2M2O2N2N1O2N2N1O2N2M2O2N1O2N101O001O001N101N10001N101N101O0O2O0O2O0O2O001N101N101N101O0O101N101N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O001N101O001O001O0O2O001O001O001O1O001O001O001O00001O0010O01O0010O0001O010O001O00010O001O010O001O00010O0010O01O010O00010O001O010O001O010O001O000010O01O001O001O001O010O00001O001O001O01O0001O00001O00001O01O0001O00001O00001O00010O0000001O00001O000000000001O000000000001O000000000001O0000000000000O100O100O101N1O100O100O100aNPDmJQ<R5RDiJQ<V5RDdJR<[5PD_JV<_5mCXJZ<g5iCoI_<Q6h0O1O2N1O1O101N1O1O101N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O001O1N2O1QKUAg4l>VKWAi4Q?O001N10001O001N10001O00001N10001O001N10001O00001cN_@XNb?g1_@YNa?e1a@[N`?c1a@]N_?a1c@^N_?_1c@aN]?^1d@aN]?]1e@cN\\?[1e@dN\\?[1e@eN\\?X1g@fNZ?Y1g@gNY?X1h@hNY?V1h@iNY?U1i@kNX?S1i@lNX?S1i@mNX?P1k@nNV?Q1k@oNU?P1l@oNV?n0m@QOS?o0m@QOT?m0m@ROT?m0n@ROR?n0n@QOT?m0m@SOS?l0o@ROS?l0n@TOR?l0n@SOS?l0n@SOT?k0n@ROT?m0m@ROU?m0k@ROV?m0l@QOU?n0l@QOV?n0j@QOW?m0l@QOV?m0l@QOU?n0o@nNR?P1SAkNP?S1o1N3N1J;DXe4"}, "label": "a white duck or goose with its bill tucked under its wing"}]}
{"id": 487464, "image": "COCO_train2014_000000487464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gold car moving on the street in front of a truck and other vehicles\"."}], "task": "refering", "answer_template": "The \"gold car moving on the street in front of a truck and other vehicles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 168, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212, 213, 225], "bbox": [0.0, 0.315546875, 0.2240941176470588, 0.676546875], "iscrowd": 0, "area": 17691.8493, "rle": {"size": [640, 425], "counts": "Z6h5X>o0QO6J5K5K0000O10000000000O10000000000O10000000000O2O0N2N2O1N2N2N2O1N2N2N2O1N2O1N2O100O1O1O100O1O1O100O1O2N2O0O2N2O11O1N2O2N10O1N3M2N3L3N2N3M2N3M2N3M2N2M4M2N3M5K4L5K4L5Ki0VO3N2N2M3K5K5J6POP1Hf0ZOjm]6"}, "label": "gold car moving on the street in front of a truck and other vehicles"}]}
{"id": 487464, "image": "COCO_train2014_000000487464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow car\"."}], "task": "refering", "answer_template": "The \"the yellow car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 168, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212, 213, 225], "bbox": [0.0, 0.315546875, 0.2240941176470588, 0.676546875], "iscrowd": 0, "area": 17691.8493, "rle": {"size": [640, 425], "counts": "Z6h5X>o0QO6J5K5K0000O10000000000O10000000000O10000000000O2O0N2N2O1N2N2N2O1N2N2N2O1N2O1N2O100O1O1O100O1O1O100O1O2N2O0O2N2O11O1N2O2N10O1N3M2N3L3N2N3M2N3M2N3M2N2M4M2N3M5K4L5K4L5Ki0VO3N2N2M3K5K5J6POP1Hf0ZOjm]6"}, "label": "the yellow car"}]}
{"id": 487464, "image": "COCO_train2014_000000487464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a beard and cigarette riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man with a beard and cigarette riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 129, 130, 144, 145, 146, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 177, 190, 191, 206, 221, 236], "bbox": [0.5107058823529412, 0.342375, 0.849294117647059, 0.6988749999999999], "iscrowd": 0, "area": 9254.425550000005, "rle": {"size": [640, 425], "counts": "lmW41nc02O1N1O2N2O1N2N2N1O2N2N2N01O01O00000O2O0000001N10001O0O10001O1O1N2O1O1O1N2O1O2O0O1O1O1O100O1O1O1O100O1O1O100O100VOhNk^OY1l`0UOo^Ok0n`0[On^Og0Pa0]Om^Oc0Ra0V101N1O101O0O101O2N2M3N2N2N1O1O2N1T@cLS?^3c@kL]?k3100O10000O4M2N3M2N3M2O1XLX@T3Z`001O100O1O00100O1O00100QOR_OjNo`0S1V_OnNg`0P1^_OYOU`0e0P@Cf?:`@IZ?4k@0o>2QA3h>NXA7a>IaA<Y>CiAf0m=YOUBQ1`=mNcBR1^=lNdBR1]=mNhBn0Y=QOnBh0S=WOSCc0o<ZOXCROSNm0f>0]CkNhN;m=i0m3N1O1N2O1O1O1O1N3N1O1O1O1N2O1N2O1N3M2O4K4M3LXfW1"}, "label": "a man with a beard and cigarette riding a motorcycle"}]}
{"id": 487464, "image": "COCO_train2014_000000487464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smoking man in overalls riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a smoking man in overalls riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 129, 130, 144, 145, 146, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 177, 190, 191, 206, 221, 236], "bbox": [0.5107058823529412, 0.342375, 0.849294117647059, 0.6988749999999999], "iscrowd": 0, "area": 9254.425550000005, "rle": {"size": [640, 425], "counts": "lmW41nc02O1N1O2N2O1N2N2N1O2N2N2N01O01O00000O2O0000001N10001O0O10001O1O1N2O1O1O1N2O1O2O0O1O1O1O100O1O1O1O100O1O1O100O100VOhNk^OY1l`0UOo^Ok0n`0[On^Og0Pa0]Om^Oc0Ra0V101N1O101O0O101O2N2M3N2N2N1O1O2N1T@cLS?^3c@kL]?k3100O10000O4M2N3M2N3M2O1XLX@T3Z`001O100O1O00100O1O00100QOR_OjNo`0S1V_OnNg`0P1^_OYOU`0e0P@Cf?:`@IZ?4k@0o>2QA3h>NXA7a>IaA<Y>CiAf0m=YOUBQ1`=mNcBR1^=lNdBR1]=mNhBn0Y=QOnBh0S=WOSCc0o<ZOXCROSNm0f>0]CkNhN;m=i0m3N1O1N2O1O1O1O1N3N1O1O1O1N2O1N2O1N3M2O4K4M3LXfW1"}, "label": "a smoking man in overalls riding a motorcycle"}]}
{"id": 487464, "image": "COCO_train2014_000000487464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white four door vehicle in traffic\"."}], "task": "refering", "answer_template": "The \"a white four door vehicle in traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 146, 147, 148, 149, 161, 162, 163, 164, 177, 178], "bbox": [0.6377882352941177, 0.237640625, 0.9992, 0.50353125], "iscrowd": 0, "area": 17763.848399999995, "rle": {"size": [640, 425], "counts": "SaY51nc03N2N2N2N2N3M2N2N2N2N2N3M2M3N2N2N2N2N3M2N2N2N2N2N3M2M3N2N3M3M3M3M3M3M3M3M3M3M3L4M1O1O2N1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1O1O1O1O2M2O1O1O1O2N1O1O1O1O1O2N1O1N2O1OO1O1O1O100O1O1O1O1O1O00100O1O100O1O100O1O100O1O010O1O100O1O100O100O1000O2O001O001O1O0101N3N1N3N1N2O2M2O2M2O1N3N1O1Nl0UOf2YMcF"}, "label": "a white four door vehicle in traffic"}]}
{"id": 487464, "image": "COCO_train2014_000000487464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver suv behind pick up truck\"."}], "task": "refering", "answer_template": "The \"silver suv behind pick up truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 146, 147, 148, 149, 161, 162, 163, 164, 177, 178], "bbox": [0.6377882352941177, 0.237640625, 0.9992, 0.50353125], "iscrowd": 0, "area": 17763.848399999995, "rle": {"size": [640, 425], "counts": "SaY51nc03N2N2N2N2N3M2N2N2N2N2N3M2M3N2N2N2N2N3M2N2N2N2N2N3M2M3N2N3M3M3M3M3M3M3M3M3M3M3L4M1O1O2N1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1O1O1O1O2M2O1O1O1O2N1O1O1O1O1O2N1O1N2O1OO1O1O1O100O1O1O1O1O1O00100O1O100O1O100O1O100O1O010O1O100O1O100O100O1000O2O001O001O1O0101N3N1N3N1N2O2M2O2M2O1N3N1O1Nl0UOf2YMcF"}, "label": "silver suv behind pick up truck"}]}
{"id": 118827, "image": "COCO_train2014_000000118827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soccer player in green and white , dribbling the ball\"."}], "task": "refering", "answer_template": "The \"a soccer player in green and white , dribbling the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 132, 133, 134, 155, 156, 157, 158, 179, 180, 181, 182, 202, 203, 204, 205, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 318, 319, 337, 338, 341, 342, 361, 364, 365, 387, 388], "bbox": [0.68596875, 0.23750511247443762, 0.9320625, 0.9624539877300613], "iscrowd": 0, "area": 26512.41700000001, "rle": {"size": [489, 640], "counts": "Xna6b1f=e1\\N1O1N2O2M2O1N3N1N2N3nD\\LU:k3`E_L`:U4O010O1000O0100O01000OjMSGRNm8l1YGoMg8o1_GnMa8P2dGmM[8[1aHaN`7_1U31N2O1N101N2O0\\K^N`Kd1^4^NaKc1]4`N`Ka1_4aN_Ka1^4cN`K]1_4eN_K]1_4eN`K\\1]4hNfGWOe3R2d4iNdGYOe3P2d4QOZKP1d4SOYKn0e4UOYKl0f4WOXKi0f4[OWKe0i4^OUKb0i4BTK?l4DQK<n4GPK9P5InJ6R5MlJ3T5MmJ2R5NoJ2Q5NZHnMY2S2^5OWHRNY2o1_50VHVNX2j1b5OUH[NV2g1d5OTH^NT2e1h5f0TJZOm5i0nIYOQ6j0kIXOU6k0gIVOZ6k0bIWO^6k0^IVOc6k0[IVOe6l0XIUOh6l0VIUOk6k0SIUOn6m0nHUOR7l0kHVOU7k0gHXOY7j0bHYO_7g0]H[Od7U400001N12O1O5J=D3M3N2N9G3M3N2M3;EL2XJiGY4Y8_KlGb4W8XKlGh4V8RKnGn4T8lJPHT5T8eJoG[5o81O001O1N101O1O1N101O1N101QMRJaLo5^3SJ`Ln5^3VJ^Ll5a3VJ]Lj5b3ZJ[Lg5d3aJTL`5j3YK]Ki4b4lKiJU4U5VLaJk3^5ZL\\Jg3c5]LZJd3e5_LXJb3f5bLWJ_3h5R3N2E:M4M3L4M4L3L5L3L5L4K4M4L3M4L3L5L3M6J;E:E<E;E:F:D<C=D<C\\Xd0"}, "label": "a soccer player in green and white , dribbling the ball"}]}
{"id": 118827, "image": "COCO_train2014_000000118827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soccer player wearing a green and white uniform\"."}], "task": "refering", "answer_template": "The \"a soccer player wearing a green and white uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 132, 133, 134, 155, 156, 157, 158, 179, 180, 181, 182, 202, 203, 204, 205, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 318, 319, 337, 338, 341, 342, 361, 364, 365, 387, 388], "bbox": [0.68596875, 0.23750511247443762, 0.9320625, 0.9624539877300613], "iscrowd": 0, "area": 26512.41700000001, "rle": {"size": [489, 640], "counts": "Xna6b1f=e1\\N1O1N2O2M2O1N3N1N2N3nD\\LU:k3`E_L`:U4O010O1000O0100O01000OjMSGRNm8l1YGoMg8o1_GnMa8P2dGmM[8[1aHaN`7_1U31N2O1N101N2O0\\K^N`Kd1^4^NaKc1]4`N`Ka1_4aN_Ka1^4cN`K]1_4eN_K]1_4eN`K\\1]4hNfGWOe3R2d4iNdGYOe3P2d4QOZKP1d4SOYKn0e4UOYKl0f4WOXKi0f4[OWKe0i4^OUKb0i4BTK?l4DQK<n4GPK9P5InJ6R5MlJ3T5MmJ2R5NoJ2Q5NZHnMY2S2^5OWHRNY2o1_50VHVNX2j1b5OUH[NV2g1d5OTH^NT2e1h5f0TJZOm5i0nIYOQ6j0kIXOU6k0gIVOZ6k0bIWO^6k0^IVOc6k0[IVOe6l0XIUOh6l0VIUOk6k0SIUOn6m0nHUOR7l0kHVOU7k0gHXOY7j0bHYO_7g0]H[Od7U400001N12O1O5J=D3M3N2N9G3M3N2M3;EL2XJiGY4Y8_KlGb4W8XKlGh4V8RKnGn4T8lJPHT5T8eJoG[5o81O001O1N101O1O1N101O1N101QMRJaLo5^3SJ`Ln5^3VJ^Ll5a3VJ]Lj5b3ZJ[Lg5d3aJTL`5j3YK]Ki4b4lKiJU4U5VLaJk3^5ZL\\Jg3c5]LZJd3e5_LXJb3f5bLWJ_3h5R3N2E:M4M3L4M4L3L5L3L5L4K4M4L3M4L3L5L3M6J;E:E<E;E:F:D<C=D<C\\Xd0"}, "label": "a soccer player wearing a green and white uniform"}]}
{"id": 118827, "image": "COCO_train2014_000000118827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire in the yellow shirt\"."}], "task": "refering", "answer_template": "The \"the umpire in the yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336, 337, 358], "bbox": [0.542078125, 0.07206543967280164, 0.73309375, 0.8941513292433538], "iscrowd": 0, "area": 31862.328100000006, "rle": {"size": [489, 640], "counts": "^kU5:[28`9=_E?R:`2H8H8F9G:M3N2M3N2N>B3L4M4L3M4QKfIi0Z6VOgIi0[6TOhIj0Y6TOjIk0W6ROlIl0V6ROlIm0T6[MhI^O7U3S6oLXJEH[3Q6aLjJMWOa3e8TLaGk3o90O0100O001O1O1N1O2O1WOQEVMQ;h2h0O2M3N1O2NnN[MdEc2[:U1O3M3M2N3M3M2[OkKUFX4e9h0M4M3N1N3UOVJUHl5T8a07H5K5K5K6J6J6Kd1[N0000000002N2N3M2N2N2N2Nn0QO1O2M2O1O2M2L5J5K6D;_Oa0A`0G8H9G8\\Oe0J5M4N1N3N1N3M2O2M2O2M2O2L3M4K4M5I6I7I8nNQ1K5K8H9G8TOl0XOo0^OVaa2"}, "label": "the umpire in the yellow shirt"}]}
{"id": 118827, "image": "COCO_train2014_000000118827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"refree in yellow and black uniform\"."}], "task": "refering", "answer_template": "The \"refree in yellow and black uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336, 337, 358], "bbox": [0.542078125, 0.07206543967280164, 0.73309375, 0.8941513292433538], "iscrowd": 0, "area": 31862.328100000006, "rle": {"size": [489, 640], "counts": "^kU5:[28`9=_E?R:`2H8H8F9G:M3N2M3N2N>B3L4M4L3M4QKfIi0Z6VOgIi0[6TOhIj0Y6TOjIk0W6ROlIl0V6ROlIm0T6[MhI^O7U3S6oLXJEH[3Q6aLjJMWOa3e8TLaGk3o90O0100O001O1O1N1O2O1WOQEVMQ;h2h0O2M3N1O2NnN[MdEc2[:U1O3M3M2N3M3M2[OkKUFX4e9h0M4M3N1N3UOVJUHl5T8a07H5K5K5K6J6J6Kd1[N0000000002N2N3M2N2N2N2Nn0QO1O2M2O1O2M2L5J5K6D;_Oa0A`0G8H9G8\\Oe0J5M4N1N3N1N3M2O2M2O2M2O2L3M4K4M5I6I7I8nNQ1K5K8H9G8TOl0XOo0^OVaa2"}, "label": "refree in yellow and black uniform"}]}
{"id": 118827, "image": "COCO_train2014_000000118827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"number nineteen on the white team running\"."}], "task": "refering", "answer_template": "The \"number nineteen on the white team running\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 328, 329, 351, 352], "bbox": [0.26956250000000004, 0.11235173824130879, 0.5168125, 0.9168507157464212], "iscrowd": 0, "area": 30351.566899999994, "rle": {"size": [489, 640], "counts": "Tib29T2?l9KmEP1U9ZOdFa1^8hN\\G_2k0]LR4\\1nJ\\2j0[LS4_1PKZ2i0YLQ4c1TKY2e0WLR4e1WKY2b0TLR4i1YKW2`0SLR4k1]KV2<PLS4P2^KT2:nKT4Y2ZKn1=jKT4e2VKe1?jKZ4j2nJa1`0iKb4o7VKUHi4o7PKUHP5R90000O1000000O10000O1000ASKbFm4Y9YKfFf4U9bKiF^4V7nK_ILYOV4l6RNSIn1b6_N\\I`1Y6mNfIS1o5Y4L31000000O100hNWJPIh5P7_JjHa5U7eJeH\\5Z7jJaHV5S1^JZ4b0^JP5V1eJU4?cJl4V1hJU4?bJi4X1lJS4>cJg4W1oJS4<eJf4U1QKT4<dJe4U1SKT4^6hKfIV4\\6dKkIY4W6`KPJ^4R6\\KUJa4m5XKYJg4i5SK[Jm4P801O000011ERKXFo4g9;O001N101O0O2O001N101O0O2O0O2N1M4M2M4M2lGVJg6m5WITJf6o5YISJd6P6[IQJa6R6eHhJY7f6M2N3N2K4H9F9G:F9G:F:jNU1cMnD>X;BmD7T;JQE0P;0UEIm:7XEBi:?\\EZOf:e0_EVOa:k0aEQOa:o0aEmN`:T1bEhN`:X1aEfN`:Z1bEbN_:_1cE]N_:c1cEYN^:h1b100O01_CXN`;h1ZD_Nd;b1VDdNj;\\1PDkNn;V1lCQOS<o0hCVOX<o1O1O00100O001O10M2O1O2M2O1O2I6gNeCZOd<d0R1N3M2N2N3M2N2N3L3N2N2N3M2N2N3M2N2N`\\c4"}, "label": "number nineteen on the white team running"}]}
{"id": 118827, "image": "COCO_train2014_000000118827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing soccer in a white uniform with the number 19 on it\"."}], "task": "refering", "answer_template": "The \"a man playing soccer in a white uniform with the number 19 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 328, 329, 351, 352], "bbox": [0.26956250000000004, 0.11235173824130879, 0.5168125, 0.9168507157464212], "iscrowd": 0, "area": 30351.566899999994, "rle": {"size": [489, 640], "counts": "Tib29T2?l9KmEP1U9ZOdFa1^8hN\\G_2k0]LR4\\1nJ\\2j0[LS4_1PKZ2i0YLQ4c1TKY2e0WLR4e1WKY2b0TLR4i1YKW2`0SLR4k1]KV2<PLS4P2^KT2:nKT4Y2ZKn1=jKT4e2VKe1?jKZ4j2nJa1`0iKb4o7VKUHi4o7PKUHP5R90000O1000000O10000O1000ASKbFm4Y9YKfFf4U9bKiF^4V7nK_ILYOV4l6RNSIn1b6_N\\I`1Y6mNfIS1o5Y4L31000000O100hNWJPIh5P7_JjHa5U7eJeH\\5Z7jJaHV5S1^JZ4b0^JP5V1eJU4?cJl4V1hJU4?bJi4X1lJS4>cJg4W1oJS4<eJf4U1QKT4<dJe4U1SKT4^6hKfIV4\\6dKkIY4W6`KPJ^4R6\\KUJa4m5XKYJg4i5SK[Jm4P801O000011ERKXFo4g9;O001N101O0O2O001N101O0O2O0O2N1M4M2M4M2lGVJg6m5WITJf6o5YISJd6P6[IQJa6R6eHhJY7f6M2N3N2K4H9F9G:F9G:F:jNU1cMnD>X;BmD7T;JQE0P;0UEIm:7XEBi:?\\EZOf:e0_EVOa:k0aEQOa:o0aEmN`:T1bEhN`:X1aEfN`:Z1bEbN_:_1cE]N_:c1cEYN^:h1b100O01_CXN`;h1ZD_Nd;b1VDdNj;\\1PDkNn;V1lCQOS<o0hCVOX<o1O1O00100O001O10M2O1O2M2O1O2I6gNeCZOd<d0R1N3M2N2N3M2N2N3L3N2N2N3M2N2N3M2N2N`\\c4"}, "label": "a man playing soccer in a white uniform with the number 19 on it"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white car with its hood open\"."}], "task": "refering", "answer_template": "The \"the white car with its hood open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.6446875000000001, 0.23370892018779343, 1.0, 0.6696713615023474], "iscrowd": 0, "area": 24090.664299999997, "rle": {"size": [426, 640], "counts": "Sn[53T=2M3N2N2N2M4N10001O00010O01O10O01O1O00100O1001O001O1O001dCAk;`0PDJj;6SD7c;T1D001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O100O1O1O1O1O001O1O1O001O1O001O1O010O1O001O1O001O1O001O1O001O1O1O00100O1O001O1O001O1O1O001O1O2N1O101N1O2N1O2O0O1O2N1O100O1O1O1O100O1O1O1O100O1O1O2N101N3M3M3M3N2M4L3M3M3N2M3M2N00O2O00001N101O0O2O00001N101O0O;E1O1O1O001O1O1O1O010001O0OZL[Gi2e8WM[Gj2d8VM]Gi2c8VM^Gj2a8WM_Gi2a8WM`Gi2_8WMaGi2_8WMbGh2^8XMbGh2]8YMcGg2]8XMeGh2Z8XMfGh2Z8XMgGg2X8ZMhGf2X8ZMhGf2X8ZMiGe2W8ZMjGg2T8ZMmGe2S8[MmGe2S8[MmGe2S8[MnGd2R8\\MnGd2Q8]MPHb2P8]MQHd2n7\\MSHc2m7]MSHc2m7]MSHc2m7]MTHb2l7^MTHb2l7^MUHSOKQ3P8kMdHT2\\7lMdHU2[7jMfHV2Y7kMhHT2X7kMiHU2W7kMiHU2W7jMkHV2T7jMlHV2T7iMnHV2R7jMnHV2R7iMoHW2Q7iMPIV2P7iMQIX2n6hMRIX2n6gMTIX2l6hMTIX2l6gMUIY2k6gMVIW2k6hMWIV2j6jMWIS2k6cMaIZ2`6\\MkIa2b8MPJ"}, "label": "the white car with its hood open"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck with the hood up\"."}], "task": "refering", "answer_template": "The \"a white truck with the hood up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.6446875000000001, 0.23370892018779343, 1.0, 0.6696713615023474], "iscrowd": 0, "area": 24090.664299999997, "rle": {"size": [426, 640], "counts": "Sn[53T=2M3N2N2N2M4N10001O00010O01O10O01O1O00100O1001O001O1O001dCAk;`0PDJj;6SD7c;T1D001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O100O1O1O1O1O001O1O1O001O1O001O1O010O1O001O1O001O1O001O1O001O1O1O00100O1O001O1O001O1O1O001O1O2N1O101N1O2N1O2O0O1O2N1O100O1O1O1O100O1O1O1O100O1O1O2N101N3M3M3M3N2M4L3M3M3N2M3M2N00O2O00001N101O0O2O00001N101O0O;E1O1O1O001O1O1O1O010001O0OZL[Gi2e8WM[Gj2d8VM]Gi2c8VM^Gj2a8WM_Gi2a8WM`Gi2_8WMaGi2_8WMbGh2^8XMbGh2]8YMcGg2]8XMeGh2Z8XMfGh2Z8XMgGg2X8ZMhGf2X8ZMhGf2X8ZMiGe2W8ZMjGg2T8ZMmGe2S8[MmGe2S8[MmGe2S8[MnGd2R8\\MnGd2Q8]MPHb2P8]MQHd2n7\\MSHc2m7]MSHc2m7]MSHc2m7]MTHb2l7^MTHb2l7^MUHSOKQ3P8kMdHT2\\7lMdHU2[7jMfHV2Y7kMhHT2X7kMiHU2W7kMiHU2W7jMkHV2T7jMlHV2T7iMnHV2R7jMnHV2R7iMoHW2Q7iMPIV2P7iMQIX2n6hMRIX2n6gMTIX2l6hMTIX2l6gMUIY2k6gMVIW2k6hMWIV2j6jMWIS2k6cMaIZ2`6\\MkIa2b8MPJ"}, "label": "a white truck with the hood up"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue color old model car numbered 7h31 - 71\"."}], "task": "refering", "answer_template": "The \"blue color old model car numbered 7h31 - 71\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.2438125, 0.1348356807511737, 0.901953125, 0.9348356807511737], "iscrowd": 0, "area": 94695.79055, "rle": {"size": [426, 640], "counts": "oRQ25m<;E:I8K4L5KK5004M2N3M3M3M2M4M3M2N3M3M3N1O2N2N1O2N2N2N1O2N2N2N1O2O1N1O2N2N2N1O2N2YNgLYIZ3b6oLYIS3b6UMYIm2a6\\MZIf2`6dMZI]2a6lMVIZ2d6oMoH[2l6U2J5K6M3M2O2N2M3N1O2M3N2M2O2N2M2O2N2M3N1N3N2N1N3N2N2M2O2N2M3N1N3N2N1N3N2N1O2N101N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O0000001O0000001O01O0001O0000001O0000001O000000001O0000001O0000001O0000001O00000000000000O2O000000001O000O101O0000001O000O10001O0000001O0O1000001O0000001N1000001O0000000O2O0000001O00000O2O000000001O000O101O000000001O0O10001O0000001O0O1000001O0000001N1000001O0000000O2O000000k1UN2N1O2M2O2N1O2N1O1O1O0O2O00001O001O001N101O00001O001O0O2O001O00001O001N101O001O00001O1N2O1O001O1O1O1O1N2O001O1O1O1O1O0O2O1O1O1O1O001O1N2O1O1O1O001O001N2O001O001O001O0O2O1O001O1O001O0O2O1O001O1O001N2O001O1O001O1O0O2O1O001O001O1O0O2O1O2N1O1O2N1O1N2O2N1O1O2N1O1O2M2O1O2N1O4L4L5K4J6J7I6J6cNZEKn:LUE3Q;DTE:R;]OREb0T;VOPEh0i;O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1ORmi0"}, "label": "blue color old model car numbered 7h31 - 71"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black classic chevrolet car\"."}], "task": "refering", "answer_template": "The \"black classic chevrolet car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.2438125, 0.1348356807511737, 0.901953125, 0.9348356807511737], "iscrowd": 0, "area": 94695.79055, "rle": {"size": [426, 640], "counts": "oRQ25m<;E:I8K4L5KK5004M2N3M3M3M2M4M3M2N3M3M3N1O2N2N1O2N2N2N1O2N2N2N1O2O1N1O2N2N2N1O2N2YNgLYIZ3b6oLYIS3b6UMYIm2a6\\MZIf2`6dMZI]2a6lMVIZ2d6oMoH[2l6U2J5K6M3M2O2N2M3N1O2M3N2M2O2N2M2O2N2M3N1N3N2N1N3N2N2M2O2N2M3N1N3N2N1N3N2N1O2N101N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O0000001O0000001O01O0001O0000001O0000001O000000001O0000001O0000001O0000001O00000000000000O2O000000001O000O101O0000001O000O10001O0000001O0O1000001O0000001N1000001O0000000O2O0000001O00000O2O000000001O000O101O000000001O0O10001O0000001O0O1000001O0000001N1000001O0000000O2O000000k1UN2N1O2M2O2N1O2N1O1O1O0O2O00001O001O001N101O00001O001O0O2O001O00001O001N101O001O00001O1N2O1O001O1O1O1O1N2O001O1O1O1O1O0O2O1O1O1O1O001O1N2O1O1O1O001O001N2O001O001O001O0O2O1O001O1O001O0O2O1O001O1O001N2O001O1O001O1O0O2O1O001O001O1O0O2O1O2N1O1O2N1O1N2O2N1O1O2N1O1O2M2O1O2N1O4L4L5K4J6J7I6J6cNZEKn:LUE3Q;DTE:R;]OREb0T;VOPEh0i;O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1ORmi0"}, "label": "black classic chevrolet car"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vintage black car\"."}], "task": "refering", "answer_template": "The \"vintage black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333], "bbox": [0.248296875, 0.1528169014084507, 0.8989687500000001, 0.9797652582159624], "iscrowd": 0, "area": 93033.70449999996, "rle": {"size": [426, 640], "counts": "gZR22U=5K5K5K5L4K5lCHo:=iDJS;X1MAPEkNo:W1SEfNk:]1VEaNh:b1YE[Nf:h1\\EUNc:m1^EQN`:R2bEjM]:X293N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2E:]Od0\\Oc0D<H9F9G:L3M3M4L3M4M2O1N3N1O1O1O000O101O00001O0O101O0000001N10001O000O101O2N8Hi0WO3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M3M2N3M3M3N1N00001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O000000O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N20000000000000000000000000000000001O000000000O1000000000000O10000000000O100000000O100000000O10000000000O100000001N101O1O001O1N101O001O1N101O1O001N101O1O001O1N101O3M6J6I7J6J6J5K6I7^H`Jn6f5lHZJT7h5jHYJU7h5jHXJV7j5hHVJX7k5gHVJW7o501O00001N101O001O0O101O001O000O2O001O001N101O1O001N2O1O1O001O1N2O1O1O001N2O1O1O001O1N2O1O001O1N2O1O001O1N2O1O002N3M3L3N3M2N3M2M4MO010000O010O1000O0100O10O10O1000O010000O01000O2O001N2O1O0O2O1O1N101O1N2O001N8I9G2M2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1N2O1O3M3L4M2N3M3M3L4M2N3M3L4L4L`0@3M2N3M2N3M2N3M2N3M2N3M2N3MTgj0"}, "label": "vintage black car"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark antique pickup truck\"."}], "task": "refering", "answer_template": "The \"dark antique pickup truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333], "bbox": [0.248296875, 0.1528169014084507, 0.8989687500000001, 0.9797652582159624], "iscrowd": 0, "area": 93033.70449999996, "rle": {"size": [426, 640], "counts": "gZR22U=5K5K5K5L4K5lCHo:=iDJS;X1MAPEkNo:W1SEfNk:]1VEaNh:b1YE[Nf:h1\\EUNc:m1^EQN`:R2bEjM]:X293N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2E:]Od0\\Oc0D<H9F9G:L3M3M4L3M4M2O1N3N1O1O1O000O101O00001O0O101O0000001N10001O000O101O2N8Hi0WO3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M3M2N3M3M3N1N00001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O000000O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N20000000000000000000000000000000001O000000000O1000000000000O10000000000O100000000O100000000O10000000000O100000001N101O1O001O1N101O001O1N101O1O001N101O1O001O1N101O3M6J6I7J6J6J5K6I7^H`Jn6f5lHZJT7h5jHYJU7h5jHXJV7j5hHVJX7k5gHVJW7o501O00001N101O001O0O101O001O000O2O001O001N101O1O001N2O1O1O001O1N2O1O1O001N2O1O1O001O1N2O1O001O1N2O1O001O1N2O1O002N3M3L3N3M2N3M2M4MO010000O010O1000O0100O10O10O1000O010000O01000O2O001N2O1O0O2O1O1N101O1N2O001N8I9G2M2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1N2O1O3M3L4M2N3M3M3L4M2N3M3L4L4L`0@3M2N3M2N3M2N3M2N3M2N3M2N3MTgj0"}, "label": "dark antique pickup truck"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white , antique vehicle , with the hood open\"."}], "task": "refering", "answer_template": "The \"a white , antique vehicle , with the hood open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.64315625, 0.23629107981220657, 1.0, 0.6721830985915493], "iscrowd": 0, "area": 25989.662, "rle": {"size": [426, 640], "counts": "Ya[54R=4L5K4L5K4M4K4L5K4L401N101N1O2O001O00001O001O001O001O001O0O101O001O001O001O001O1O001O001O1O001O1O001N101O1O001O001O1O001O1O001O2N2N3M3M3M3M3M3M3M1OO10001N100O10000O10000O2O0O10000O1001O1O10O01O1O1O1O001O1O100O001O1O1XFgL]9d3O1O1O10O01O1O1O1O1O001O100O1O2N2N1O2N2N2N2O0O2N2N2N2N2N1O2N2O1N2N2N1O2N2N1O0001O001O001O001O1O001N101O001N2M4M5J5K6K1O0101N101O1N101O1N101N101O1N101O0O2O1N101O1O0O2O001O0O10000O1000001N1000`LlGj1T8VNoGg1P8YNSHe1m7[NUHc1k7\\NVHd1j7\\NVHe1h7[NYHe1g7[NZHd1f7[N[He1d7\\N\\Hd1d7[N]He1c7[N]He1c7ZN^Hf1a7[N_Hf1`7YNaHg1_7XNbHh1^7XNbHh1]7XNdHh1\\7XNdHh1\\7WNeHi1[7WNfHh1Y7XNm300O`L"}, "label": "a white , antique vehicle , with the hood open"}]}
{"id": 299051, "image": "COCO_train2014_000000299051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"very old fashioned white car sitting beside the blue one of similar kind\"."}], "task": "refering", "answer_template": "The \"very old fashioned white car sitting beside the blue one of similar kind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.64315625, 0.23629107981220657, 1.0, 0.6721830985915493], "iscrowd": 0, "area": 25989.662, "rle": {"size": [426, 640], "counts": "Ya[54R=4L5K4L5K4M4K4L5K4L401N101N1O2O001O00001O001O001O001O001O0O101O001O001O001O001O1O001O001O1O001O1O001N101O1O001O001O1O001O1O001O2N2N3M3M3M3M3M3M3M1OO10001N100O10000O10000O2O0O10000O1001O1O10O01O1O1O1O001O1O100O001O1O1XFgL]9d3O1O1O10O01O1O1O1O1O001O100O1O2N2N1O2N2N2N2O0O2N2N2N2N2N1O2N2O1N2N2N1O2N2N1O0001O001O001O001O1O001N101O001N2M4M5J5K6K1O0101N101O1N101O1N101N101O1N101O0O2O1N101O1O0O2O001O0O10000O1000001N1000`LlGj1T8VNoGg1P8YNSHe1m7[NUHc1k7\\NVHd1j7\\NVHe1h7[NYHe1g7[NZHd1f7[N[He1d7\\N\\Hd1d7[N]He1c7[N]He1c7ZN^Hf1a7[N_Hf1`7YNaHg1_7XNbHh1^7XNbHh1]7XNdHh1\\7XNdHh1\\7WNeHi1[7WNfHh1Y7XNm300O`L"}, "label": "very old fashioned white car sitting beside the blue one of similar kind"}]}
{"id": 348203, "image": "COCO_train2014_000000348203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second and third pans from the eft on the stove\"."}], "task": "refering", "answer_template": "The \"second and third pans from the eft on the stove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 316, 317, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 356, 357, 358, 359, 360, 361, 362, 363, 364, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.451109375, 0.740375, 0.8614062499999999, 0.953875], "iscrowd": 0, "area": 20033.461699999993, "rle": {"size": [480, 640], "counts": "ejW43m>3M4L3M4K3N1O2N2N2N2N1O2M3N2N2N1O2N2N2N1N3N2N2N2N1O2N2N2M2O2N2N2N2N1O2N2M3N1O2N2N1O00000O100000000000000O100000000000000O100000000000000O1000000000000O10001O000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O11O001O001O001O001O001O010O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00010O001O001O001O001O001O001O001O001O001O001O000O2O0O2O0O2O0O2O005J:G9F:GUYY1"}, "label": "second and third pans from the eft on the stove"}]}
{"id": 348203, "image": "COCO_train2014_000000348203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pan with a brown substance in it\"."}], "task": "refering", "answer_template": "The \"the pan with a brown substance in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 316, 317, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 356, 357, 358, 359, 360, 361, 362, 363, 364, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.451109375, 0.740375, 0.8614062499999999, 0.953875], "iscrowd": 0, "area": 20033.461699999993, "rle": {"size": [480, 640], "counts": "ejW43m>3M4L3M4K3N1O2N2N2N2N1O2M3N2N2N1O2N2N2N1N3N2N2N2N1O2N2N2M2O2N2N2N2N1O2N2M3N1O2N2N1O00000O100000000000000O100000000000000O100000000000000O1000000000000O10001O000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O11O001O001O001O001O001O010O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00010O001O001O001O001O001O001O001O001O001O001O000O2O0O2O0O2O0O2O005J:G9F:GUYY1"}, "label": "the pan with a brown substance in it"}]}
{"id": 348203, "image": "COCO_train2014_000000348203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chef in the back washing his hands in the sink\"."}], "task": "refering", "answer_template": "The \"the chef in the back washing his hands in the sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292], "bbox": [0.49265625, 0.2494375, 0.7262968750000001, 0.7437916666666666], "iscrowd": 0, "area": 20549.496, "rle": {"size": [480, 640], "counts": "Tnc41l>9H8G3M4O001O001O0cKWOcJj0Z5YOeJg0Y5\\OeJf0X5^OfJb0W5BhJ>X5CgJ>W5GeJ9Z5JdJ7[5JcJ7\\5LbJ5\\5NaJ3_5N`J1`52]JNd53[JKf58WJGk5;RJDo5a0lI_OU6d0gI[OZ6g0cIZO^6h0_IXOa6j0]IVOd6k0ZITOg6m0WITOj6m0SITOm6m0QITOP7m0lHUOU7l0hHUOY7k0bHZO_7o0nGYOS8l0dGWO^8n0XGVOh8P1nFSOS9R1cFSO^9V3101N1O2O0O1O2N2O0O2N2N1OdG`K[6_4bIgK\\6W4bIoK[6P4cITL]6j3aIZL_6f3]I^Lc6b3YIbLf6`3SIfLm6[3lHjLU7[3`HiLb7Z51O1O001O1N2O1O001O1N3N1O2N1O2M2O2N1O1O0O100000000O100000000O10001O00000O10001O1O1N101O1O1O1N2O1O1O1O0O2O1O1O1O1N4M3M3M4L4K5L4L4L4L4L4K5L4L4L4L4K5L6J5K5K5J5L5K4L4D=B=C>CTha2"}, "label": "the chef in the back washing his hands in the sink"}]}
{"id": 348203, "image": "COCO_train2014_000000348203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a chef washing his hands\"."}], "task": "refering", "answer_template": "The \"this is a chef washing his hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292], "bbox": [0.49265625, 0.2494375, 0.7262968750000001, 0.7437916666666666], "iscrowd": 0, "area": 20549.496, "rle": {"size": [480, 640], "counts": "Tnc41l>9H8G3M4O001O001O0cKWOcJj0Z5YOeJg0Y5\\OeJf0X5^OfJb0W5BhJ>X5CgJ>W5GeJ9Z5JdJ7[5JcJ7\\5LbJ5\\5NaJ3_5N`J1`52]JNd53[JKf58WJGk5;RJDo5a0lI_OU6d0gI[OZ6g0cIZO^6h0_IXOa6j0]IVOd6k0ZITOg6m0WITOj6m0SITOm6m0QITOP7m0lHUOU7l0hHUOY7k0bHZO_7o0nGYOS8l0dGWO^8n0XGVOh8P1nFSOS9R1cFSO^9V3101N1O2O0O1O2N2O0O2N2N1OdG`K[6_4bIgK\\6W4bIoK[6P4cITL]6j3aIZL_6f3]I^Lc6b3YIbLf6`3SIfLm6[3lHjLU7[3`HiLb7Z51O1O001O1N2O1O001O1N3N1O2N1O2M2O2N1O1O0O100000000O100000000O10001O00000O10001O1O1N101O1O1O1N2O1O1O1O0O2O1O1O1O1N4M3M3M4L4K5L4L4L4L4L4K5L4L4L4L4K5L6J5K5K5J5L5K4L4D=B=C>CTha2"}, "label": "this is a chef washing his hands"}]}
{"id": 348203, "image": "COCO_train2014_000000348203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female chef wearing a white chefs hat looking into the pan\"."}], "task": "refering", "answer_template": "The \"a female chef wearing a white chefs hat looking into the pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238], "bbox": [0.17875000000000002, 0.25608333333333333, 0.365140625, 0.6079583333333333], "iscrowd": 0, "area": 13631.742399999996, "rle": {"size": [480, 640], "counts": "`fe11l>8H8H8H7I3M2O2N2M3N1O2N2N2N2N1O2O1N2O1N2N2O1N2M2O2N2M3N2N2M3N2oNlMSEW2g:QNTER2g:VNREn1i:W1N101N2O1O1O1O1O1O1N2N2O1N2O0O2N2O1O1N2O1O1O10000O10000000000000000000000000O10000000001O001O001O01O01O001O001O001O010O1O001O001O0lMTFTOm9e0^FVOc9g0aFXO^9g0dFXO^9d0fFZO\\9b0gF]O[9?hF@Y9=kFAZ98jFE^93eFK_9NfF0^9IhF4]9CiF:\\9_OiF?m;O2N1N3N1H9CQRn5"}, "label": "a female chef wearing a white chefs hat looking into the pan"}]}
{"id": 348203, "image": "COCO_train2014_000000348203.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the women is cooking in the kitchen\"."}], "task": "refering", "answer_template": "The \"the women is cooking in the kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238], "bbox": [0.17875000000000002, 0.25608333333333333, 0.365140625, 0.6079583333333333], "iscrowd": 0, "area": 13631.742399999996, "rle": {"size": [480, 640], "counts": "`fe11l>8H8H8H7I3M2O2N2M3N1O2N2N2N2N1O2O1N2O1N2N2O1N2M2O2N2M3N2N2M3N2oNlMSEW2g:QNTER2g:VNREn1i:W1N101N2O1O1O1O1O1O1N2N2O1N2O0O2N2O1O1N2O1O1O10000O10000000000000000000000000O10000000001O001O001O01O01O001O001O001O010O1O001O001O0lMTFTOm9e0^FVOc9g0aFXO^9g0dFXO^9d0fFZO\\9b0gF]O[9?hF@Y9=kFAZ98jFE^93eFK_9NfF0^9IhF4]9CiF:\\9_OiF?m;O2N1N3N1H9CQRn5"}, "label": "the women is cooking in the kitchen"}]}
{"id": 53294, "image": "COCO_train2014_000000053294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle is parking in the road along with another vehicles\"."}], "task": "refering", "answer_template": "The \"a bicycle is parking in the road along with another vehicles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 256], "bbox": [0.0, 0.42742203742203744, 0.20784375000000002, 0.6935135135135135], "iscrowd": 0, "area": 9340.84735, "rle": {"size": [481, 640], "counts": "Q8\\2W;^1L310O0100O100O;F4K1]MbDc1^;^NbDa1^;`NbD^1_;cNaD\\1e;_NZD`1g;aNYD^1h;bNXD\\1i;eNWDZ1i;gNWDW1k;iNUDV1k;jNVDT1l;lNSDT1m;mNSDQ1o;oNRDo0n;RORDl0P<TOPDh0S<YOmCd0V<\\OjCa0X<@hC=[<CeC9^<HcC4`<LbCO`<1cCI`<8bCC_<?bC[Ob<f0l01O1O1O1O1O1O001O1O1O001O1O0^BlNW=c1L1O1O001\\OSNQDn1o;RNQDo1n;RNQDn1o;RNQDn1o;SNPDm1P<SNoCn1Q<RNoCn1Q<RNoCn1Q<RNoCm1R<SNnCm1Q<TNnCm1R<TNiCP2W<PNeCT2[<90000OnMiC\\1W<_NRD]1n;cNTD[1l;eNUDZ1k;fNVDY1i;hNWDX1i;hNXDX1g;hNYDY1f;gNZD[1d;eN\\D\\1c;dN]D]1b;cN^D^1a;bN_D_1a;`N_Db1_;^NaDc1_;[NbDf1_;VNcDk1^;oMfDS2S<001O00001O00001O00001O00010O000010O001O0O2O0O2N2O1N1O2O1N2N2O1N2N2O1M3N2N4L3L5L4L4L4KWl]7"}, "label": "a bicycle is parking in the road along with another vehicles"}]}
{"id": 53294, "image": "COCO_train2014_000000053294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle leaning on a bench\"."}], "task": "refering", "answer_template": "The \"a bicycle leaning on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 256], "bbox": [0.0, 0.42742203742203744, 0.20784375000000002, 0.6935135135135135], "iscrowd": 0, "area": 9340.84735, "rle": {"size": [481, 640], "counts": "Q8\\2W;^1L310O0100O100O;F4K1]MbDc1^;^NbDa1^;`NbD^1_;cNaD\\1e;_NZD`1g;aNYD^1h;bNXD\\1i;eNWDZ1i;gNWDW1k;iNUDV1k;jNVDT1l;lNSDT1m;mNSDQ1o;oNRDo0n;RORDl0P<TOPDh0S<YOmCd0V<\\OjCa0X<@hC=[<CeC9^<HcC4`<LbCO`<1cCI`<8bCC_<?bC[Ob<f0l01O1O1O1O1O1O001O1O1O001O1O0^BlNW=c1L1O1O001\\OSNQDn1o;RNQDo1n;RNQDn1o;RNQDn1o;SNPDm1P<SNoCn1Q<RNoCn1Q<RNoCn1Q<RNoCm1R<SNnCm1Q<TNnCm1R<TNiCP2W<PNeCT2[<90000OnMiC\\1W<_NRD]1n;cNTD[1l;eNUDZ1k;fNVDY1i;hNWDX1i;hNXDX1g;hNYDY1f;gNZD[1d;eN\\D\\1c;dN]D]1b;cN^D^1a;bN_D_1a;`N_Db1_;^NaDc1_;[NbDf1_;VNcDk1^;oMfDS2S<001O00001O00001O00001O00010O000010O001O0O2O0O2N2O1N1O2O1N2N2O1N2N2O1M3N2N4L3L5L4L4L4KWl]7"}, "label": "a bicycle leaning on a bench"}]}
{"id": 53294, "image": "COCO_train2014_000000053294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle resting on a kickstand\"."}], "task": "refering", "answer_template": "The \"a black motorcycle resting on a kickstand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 332, 334, 335, 348, 349, 350, 351, 352, 353, 354, 358, 359, 373, 374, 375, 376], "bbox": [0.15132812499999998, 0.06062370062370062, 0.771734375, 0.9823284823284824], "iscrowd": 0, "area": 94625.78495000002, "rle": {"size": [481, 640], "counts": "[m]1g0W>d0\\O6J6J6J6J7I6J6J6J6J4L4M3N2M2O2M3N2N1N3N2N2M3N1N3N2N2O1O0O2O1N2O0O2O1N2O1N101N2O1O1N101N2O1N101N2O1O1N101N2O1N100O101O0O100O100B_J[Gb5e8_JZGb5e8^JZGc5f8^JYGb5g8_JXGa5h8_JWGb5i8_JVGa5j8_JUGb5k8_JTGb5k8^JTGc5l8^JSGb5l8`JSG`5m8:000001O000000000O101O0YNkIbJU6]5mIbJS6\\5PJbJQ6Y5UJfJk5X5XJgJh5X5ZJgJg5V5\\JiJd5U5_JiJb5V5_JjJa5T5aJlJ_5R5cJnJ]5Q5eJmJ\\5Q5fJoJ[5n4gJRKX5n4iJRKW5l4kJTKU5l4lJRKU5m4lJSKS5m4nJSKR5l4oJTKQ5l4oJTKP5l4QKTKo4k4SKTKm4k4TKUKk4l4UKTKk4k4VKUKj4j4WKVKi4[4lIgJ_1k0d4Z4SJhJb1g0[4\\4XJlJc1;]4`4ZJRK\\1=Y4_4hL_KX3_4lL_KT3^4PMaKP3]4TMaKk2`4VM`Ki2`4YM^Kg2b4[M\\Kd2d4^M[Kb2e4`MYK`2f4m310O100O001O1O001O1O0F:K5K5L5M2001O001O01O01O001O0010O01O00001O001O00001O001O00001O0O10000000000000000M3L4L3M31O3L3N3I6F;N2N1O2NO100O101L3L4M3F:I81O001O1O0O2O001O001O0O2O001O001O000O1000O101O0000000000001O0000000O1000001N10000O10000O2O000001O2N2N102M3M2N3M3M>B1O10O01OZOjLbEV3\\:mLbES3\\:PMdEP3Z:SMeEl2Z:VMeEj2[:XMdEh2]:WMcEh2^:YMaEg2^:ZMaEf2_:[MaEd2_:\\MbEd2U1kLd7b0VGc2T1mLf7a0VGb2R1nLh7a0VG`2P1QMj7`0WG^2l0TMm7>XG^2i0UMo7>YG]2d0XMR8<[G[2a0ZMT8<[GZ2>\\MW8;\\GX2;^MY8:^GW26aMn0Ac5h0YIV24bMn0H_5a0]IX22bMP1K\\5<aIX21aMR10W58fIX2O`MS17S52jIX2M`MV1<m4NnIX2M]MZ1b0f4JRJW2M^M[1e0a4HVJV2L]M]1g0_4GXJU2J_M_1e0_4HWJU2I^Mb1e0]4JVJT2J^Mb1f0\\4JWJn2^1XM[4JWJm2_1YMY4KXJk2`1[MV4KYJj2a1\\MU4KZJg2c1^MR4L[Je2d1_MP4M\\Jc2d1`MP4OZJ`2h1aMm30[J^2i1aMl32[J[2k1cMh34]JX27TLW1a1T44]JV29VLV1_1T46]Jk1b0aLm0^1S47^Jg1c0fLl0\\1Q49`Jc1d0iLj0\\1Q49aJ`1e0lLi0[1P4:aJ^1g0mLi0[1m3<cJZ1i0oLf0\\1m3=cJV1k0RMe0[1k3?eJR1l0UMd0Z1j3`0eJR1m0UMc0Y1i3b0gJo0m0XMa0X1i3c0iJm0l0ZMa0V1i3d0jJl0m0ZM?V1h3j3jKQK=U1g3l3lKPK;U1g3l3oKWK2m0n3m3PLXK0j0n3P4TLVKLk0n3Q4VLUKKj0m3S4XLTKJi0X3\\N_Lj5>QKJi0Z3\\N\\Lm5?oJJh0Z3]NZLo5b0mJIf0\\3^NWLR6c0kJHe0_3^NSLU6e0jJHb0`3`NQLW6f0hJGa0c3`NnKY6g0hJG>d3bNkK\\6i0eJF=g3U5aL`JG:i3V5`L`JF:j3X5^L`JF8l3Y5^L`JE6m3\\5\\L`JE6n3[5[LaJE5o3[5[LbJE3o3]5ZLbJE2P4]5ZLdJC0g3\\NaLl:]3XE`Li:_3[E]Lf:b3=O2N1O1O1O1O1O1XN]D1d;O\\D0e;0^DLd;4`DFb;:bD@b;>_D_Of;=\\D@h;>YD^Ok;6hCWO=a0n;5gCZO;?P<6eC\\O;<S<5cC_O<9S<7bC@;7V<7_CC;4X<7^CF:O]<9ZCH:Hb<>UCIc=6]BHg=6ZBFk=7VBGn=7=010O010O010O010O010O010O010O010O02O1N3N1N2OT\\T2"}, "label": "a black motorcycle resting on a kickstand"}]}
{"id": 53294, "image": "COCO_train2014_000000053294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle with its kickstand down\"."}], "task": "refering", "answer_template": "The \"a black motorcycle with its kickstand down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 332, 334, 335, 348, 349, 350, 351, 352, 353, 354, 358, 359, 373, 374, 375, 376], "bbox": [0.15132812499999998, 0.06062370062370062, 0.771734375, 0.9823284823284824], "iscrowd": 0, "area": 94625.78495000002, "rle": {"size": [481, 640], "counts": "[m]1g0W>d0\\O6J6J6J6J7I6J6J6J6J4L4M3N2M2O2M3N2N1N3N2N2M3N1N3N2N2O1O0O2O1N2O0O2O1N2O1N101N2O1O1N101N2O1N101N2O1O1N101N2O1N100O101O0O100O100B_J[Gb5e8_JZGb5e8^JZGc5f8^JYGb5g8_JXGa5h8_JWGb5i8_JVGa5j8_JUGb5k8_JTGb5k8^JTGc5l8^JSGb5l8`JSG`5m8:000001O000000000O101O0YNkIbJU6]5mIbJS6\\5PJbJQ6Y5UJfJk5X5XJgJh5X5ZJgJg5V5\\JiJd5U5_JiJb5V5_JjJa5T5aJlJ_5R5cJnJ]5Q5eJmJ\\5Q5fJoJ[5n4gJRKX5n4iJRKW5l4kJTKU5l4lJRKU5m4lJSKS5m4nJSKR5l4oJTKQ5l4oJTKP5l4QKTKo4k4SKTKm4k4TKUKk4l4UKTKk4k4VKUKj4j4WKVKi4[4lIgJ_1k0d4Z4SJhJb1g0[4\\4XJlJc1;]4`4ZJRK\\1=Y4_4hL_KX3_4lL_KT3^4PMaKP3]4TMaKk2`4VM`Ki2`4YM^Kg2b4[M\\Kd2d4^M[Kb2e4`MYK`2f4m310O100O001O1O001O1O0F:K5K5L5M2001O001O01O01O001O0010O01O00001O001O00001O001O00001O0O10000000000000000M3L4L3M31O3L3N3I6F;N2N1O2NO100O101L3L4M3F:I81O001O1O0O2O001O001O0O2O001O001O000O1000O101O0000000000001O0000000O1000001N10000O10000O2O000001O2N2N102M3M2N3M3M>B1O10O01OZOjLbEV3\\:mLbES3\\:PMdEP3Z:SMeEl2Z:VMeEj2[:XMdEh2]:WMcEh2^:YMaEg2^:ZMaEf2_:[MaEd2_:\\MbEd2U1kLd7b0VGc2T1mLf7a0VGb2R1nLh7a0VG`2P1QMj7`0WG^2l0TMm7>XG^2i0UMo7>YG]2d0XMR8<[G[2a0ZMT8<[GZ2>\\MW8;\\GX2;^MY8:^GW26aMn0Ac5h0YIV24bMn0H_5a0]IX22bMP1K\\5<aIX21aMR10W58fIX2O`MS17S52jIX2M`MV1<m4NnIX2M]MZ1b0f4JRJW2M^M[1e0a4HVJV2L]M]1g0_4GXJU2J_M_1e0_4HWJU2I^Mb1e0]4JVJT2J^Mb1f0\\4JWJn2^1XM[4JWJm2_1YMY4KXJk2`1[MV4KYJj2a1\\MU4KZJg2c1^MR4L[Je2d1_MP4M\\Jc2d1`MP4OZJ`2h1aMm30[J^2i1aMl32[J[2k1cMh34]JX27TLW1a1T44]JV29VLV1_1T46]Jk1b0aLm0^1S47^Jg1c0fLl0\\1Q49`Jc1d0iLj0\\1Q49aJ`1e0lLi0[1P4:aJ^1g0mLi0[1m3<cJZ1i0oLf0\\1m3=cJV1k0RMe0[1k3?eJR1l0UMd0Z1j3`0eJR1m0UMc0Y1i3b0gJo0m0XMa0X1i3c0iJm0l0ZMa0V1i3d0jJl0m0ZM?V1h3j3jKQK=U1g3l3lKPK;U1g3l3oKWK2m0n3m3PLXK0j0n3P4TLVKLk0n3Q4VLUKKj0m3S4XLTKJi0X3\\N_Lj5>QKJi0Z3\\N\\Lm5?oJJh0Z3]NZLo5b0mJIf0\\3^NWLR6c0kJHe0_3^NSLU6e0jJHb0`3`NQLW6f0hJGa0c3`NnKY6g0hJG>d3bNkK\\6i0eJF=g3U5aL`JG:i3V5`L`JF:j3X5^L`JF8l3Y5^L`JE6m3\\5\\L`JE6n3[5[LaJE5o3[5[LbJE3o3]5ZLbJE2P4]5ZLdJC0g3\\NaLl:]3XE`Li:_3[E]Lf:b3=O2N1O1O1O1O1O1XN]D1d;O\\D0e;0^DLd;4`DFb;:bD@b;>_D_Of;=\\D@h;>YD^Ok;6hCWO=a0n;5gCZO;?P<6eC\\O;<S<5cC_O<9S<7bC@;7V<7_CC;4X<7^CF:O]<9ZCH:Hb<>UCIc=6]BHg=6ZBFk=7VBGn=7=010O010O010O010O010O010O010O010O02O1N3N1N2OT\\T2"}, "label": "a black motorcycle with its kickstand down"}]}
{"id": 282671, "image": "COCO_train2014_000000282671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in brown coat calling\"."}], "task": "refering", "answer_template": "The \"a man in brown coat calling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 194, 195, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.42126562500000003, 0.44933333333333336, 0.612453125, 0.9998125000000001], "iscrowd": 0, "area": 18912.878399999994, "rle": {"size": [480, 640], "counts": "_kn3:f>3L3M2N3L3M2N2N2B\\O^Bg0_=>O10000O10000O100O1O100iDlNa8U1XGKk6bNgGO1d1S15n6aNjGMO^1g0g0\\7RNlGNMZ1?Q1e7nMiGLOe4V8c1L4M4K5L4K5L4L`0_O101O0000000000000000001O01O001O001O1O1O2N1O100O6_HkHk6W7RIoHi6T7SInHl6c7O2H8_O`0\\Oe0L37J3L4M2N9G5J3M7I1OQMkGTOT8k0nGTOR8k0oGUOR8j0nGVOR8i0oGWOQ8i0oGWOQ8h0PHXOP8g0PHZOP8f0PHZOP8e0QH[Oo7e0QH[OP8c0QH]Oo7b0QH_Oo7a0QH_Oo7`0RH@n7`0RHAm7>XH^Oh7b0fHPOZ7o0TIdNn6Z1`IXNa6g1eIcMVN^OV8m2kIoLeNNb7Q3_3A[D_Mg;^2[DaMg;\\2\\D`Mg;^2`0M3M3M3M3M3N2L4M3M3M3M3M3M3L4L4L4M3L4I[nc3"}, "label": "a man in brown coat calling"}]}
{"id": 282671, "image": "COCO_train2014_000000282671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in glasses on the phone\"."}], "task": "refering", "answer_template": "The \"a man in glasses on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 194, 195, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.42126562500000003, 0.44933333333333336, 0.612453125, 0.9998125000000001], "iscrowd": 0, "area": 18912.878399999994, "rle": {"size": [480, 640], "counts": "_kn3:f>3L3M2N3L3M2N2N2B\\O^Bg0_=>O10000O10000O100O1O100iDlNa8U1XGKk6bNgGO1d1S15n6aNjGMO^1g0g0\\7RNlGNMZ1?Q1e7nMiGLOe4V8c1L4M4K5L4K5L4L`0_O101O0000000000000000001O01O001O001O1O1O2N1O100O6_HkHk6W7RIoHi6T7SInHl6c7O2H8_O`0\\Oe0L37J3L4M2N9G5J3M7I1OQMkGTOT8k0nGTOR8k0oGUOR8j0nGVOR8i0oGWOQ8i0oGWOQ8h0PHXOP8g0PHZOP8f0PHZOP8e0QH[Oo7e0QH[OP8c0QH]Oo7b0QH_Oo7a0QH_Oo7`0RH@n7`0RHAm7>XH^Oh7b0fHPOZ7o0TIdNn6Z1`IXNa6g1eIcMVN^OV8m2kIoLeNNb7Q3_3A[D_Mg;^2[DaMg;\\2\\D`Mg;^2`0M3M3M3M3M3N2L4M3M3M3M3M3M3L4L4L4M3L4I[nc3"}, "label": "a man in glasses on the phone"}]}
{"id": 233526, "image": "COCO_train2014_000000233526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small boy on left\"."}], "task": "refering", "answer_template": "The \"small boy on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 146, 147, 148, 149, 169, 170, 171, 192, 193, 194, 216, 217, 239, 262], "bbox": [0.347984375, 0.3284487534626039, 0.48793749999999997, 0.8800277008310249], "iscrowd": 0, "area": 8453.053949999996, "rle": {"size": [361, 640], "counts": "Zi^25R;8I7H6K4dN^OiGg0R8\\OlGh0P8ZOmGl0m7WOQHm0j7VOUHn0f7UOWHP1c7SO[HQ1`7RO_HR1]7PO`HU1\\7mNaHW1\\7kNaHZ1[7hNcH\\1Y7[1O1O1O1O001N2O1O1O1O1O1iJgK\\3Z4^LkKb3V4YLnKg3T4RLRLm3P4PLQLP4P4oKPLQ4R4bKcK@:o4T4VKkKL1m4V4mJPL8Il4m4TKSKl4h0SKf22aLl4g0SKh22`Lk4h0TKf23bLi4f0VKg21bLj4g0UKg22aLj4f0VKg22aLi4h0VKf22aLj4g0TKh23_Ln4d0QK^2JjL<0n4b0mJd2^6UMeIj2`6oLbIQ3\\7O001O001O001O001N2O002N1N2O1N2O2eNaGjN1a0`8c0cGhN0c0^8b0RHZOP8;\\GXOg0;n7;_HA]7c0fH[Ol6R1XIjNg6W1_IdN_6]1f11O0000J7M3L5L3M4L5L4K4M4L4Lj^c3"}, "label": "small boy on left"}]}
{"id": 233526, "image": "COCO_train2014_000000233526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small dark - skinned boy\"."}], "task": "refering", "answer_template": "The \"a small dark - skinned boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 146, 147, 148, 149, 169, 170, 171, 192, 193, 194, 216, 217, 239, 262], "bbox": [0.347984375, 0.3284487534626039, 0.48793749999999997, 0.8800277008310249], "iscrowd": 0, "area": 8453.053949999996, "rle": {"size": [361, 640], "counts": "Zi^25R;8I7H6K4dN^OiGg0R8\\OlGh0P8ZOmGl0m7WOQHm0j7VOUHn0f7UOWHP1c7SO[HQ1`7RO_HR1]7PO`HU1\\7mNaHW1\\7kNaHZ1[7hNcH\\1Y7[1O1O1O1O001N2O1O1O1O1O1iJgK\\3Z4^LkKb3V4YLnKg3T4RLRLm3P4PLQLP4P4oKPLQ4R4bKcK@:o4T4VKkKL1m4V4mJPL8Il4m4TKSKl4h0SKf22aLl4g0SKh22`Lk4h0TKf23bLi4f0VKg21bLj4g0UKg22aLj4f0VKg22aLi4h0VKf22aLj4g0TKh23_Ln4d0QK^2JjL<0n4b0mJd2^6UMeIj2`6oLbIQ3\\7O001O001O001O001N2O002N1N2O1N2O2eNaGjN1a0`8c0cGhN0c0^8b0RHZOP8;\\GXOg0;n7;_HA]7c0fH[Ol6R1XIjNg6W1_IdN_6]1f11O0000J7M3L5L3M4L5L4K4M4L4Lj^c3"}, "label": "a small dark - skinned boy"}]}
{"id": 86075, "image": "COCO_train2014_000000086075.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in green shirt\"."}], "task": "refering", "answer_template": "The \"the woman in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378], "bbox": [0.0, 0.22246875, 0.8419375, 0.986515625], "iscrowd": 0, "area": 87206.28499999997, "rle": {"size": [640, 480], "counts": "\\:Z9d:2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1]LQFTMP:k2UFPMl9n2YFnLh9P3^FkLc9T3aFhL`9V3eFfL\\9Y3iFaLY9]3lF_LU9`3oF\\LR9b3TGYLm8f3UGXLl8f3XGWLi8h3YGVLh8h3[GVLf8i3\\GULe8i3_GTLb8k3`GSLa8k3bGSL_8l3dGRL\\8l3gGSLY8l3iGRLX8l3lGRLT8m3nGRLR8l3QHRLP8m3RHRLo7k3UHRLl7m3VHRLj7l3YHSLg7l3\\HQLe7m3^HRLb7l3_HULa7j3_HVLb7h3_HYLa7g3^HZLb7g3\\HYLe7g3ZHZLf7f3YHZLh7f3WH[Li7e3VH\\Lj7d3UH\\Ll7d3SH]Lm7c3RH^Ln7b3QH^LP8b3oG_LQ8a3PH^LP8b3RH[Lo7f3SHWLm7i3UHTLl7l3VHRLj7n3XHlKl7T4WHdKn7\\4TH^KP8b4RHXKR8h4c30000001O00000000000000001O2N1O2N1O1O2N000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O10000O1001M2N2N3M2N2N3M2N3M2M3N3L3K6I6J6K6I6K5]Md2N101N100_J^Fc0d9\\O\\Fc0e9]O\\Fb0e9]O[Fb0f9^O[Fa0e9^O\\Fa0f9^O[FaNMdNi9k2ZFaNObNg9m2[FaNO_Nh9o2YFbN1]Nf9P3[FbN1[Nf9R3YFdN2XNe9T3ZFcN3VNd9W3YFcN5oMi9W3XFjN0gMQ:R3YFWOH_MX:U3SF[OGXM_:Y3kE@GoLg:\\3eEDZ;6hDKZ;1gDO\\;LgD3\\;HfD9\\;CfD<];_OeDa0^;ZOeDf0];TOfDl0];oNfDP1];kMPCOc1W2`;PMfCe0l0Z2b<dM^C\\2e<`M]C^2f<_M[C`2i<]MXCa2k<\\MVCd2m<XMTCg2P=VMPCi2S=TMnBk2U?O0O1O101N1O1O100N3M2O1N2O1N3M2O1N2O1N2O1001O00001O00001O00001O00001O00001OO1O100O1C=00O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000001O00000O100000000O100000000O100000000O100000000O100000000O1000000O10001O000O10001O0001O01O01O000010O0000010O0001O01O01O000010O0001O01O01O000010O0001O01O01O00010O00001O01O000O101O0O100O2O0O100O104K;F1O0O100O100O2O0O10000O100O100O100O10002M2O2M2O2M_T_1"}, "label": "the woman in green shirt"}]}
{"id": 86075, "image": "COCO_train2014_000000086075.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in green\"."}], "task": "refering", "answer_template": "The \"the woman in green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378], "bbox": [0.0, 0.22246875, 0.8419375, 0.986515625], "iscrowd": 0, "area": 87206.28499999997, "rle": {"size": [640, 480], "counts": "\\:Z9d:2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1]LQFTMP:k2UFPMl9n2YFnLh9P3^FkLc9T3aFhL`9V3eFfL\\9Y3iFaLY9]3lF_LU9`3oF\\LR9b3TGYLm8f3UGXLl8f3XGWLi8h3YGVLh8h3[GVLf8i3\\GULe8i3_GTLb8k3`GSLa8k3bGSL_8l3dGRL\\8l3gGSLY8l3iGRLX8l3lGRLT8m3nGRLR8l3QHRLP8m3RHRLo7k3UHRLl7m3VHRLj7l3YHSLg7l3\\HQLe7m3^HRLb7l3_HULa7j3_HVLb7h3_HYLa7g3^HZLb7g3\\HYLe7g3ZHZLf7f3YHZLh7f3WH[Li7e3VH\\Lj7d3UH\\Ll7d3SH]Lm7c3RH^Ln7b3QH^LP8b3oG_LQ8a3PH^LP8b3RH[Lo7f3SHWLm7i3UHTLl7l3VHRLj7n3XHlKl7T4WHdKn7\\4TH^KP8b4RHXKR8h4c30000001O00000000000000001O2N1O2N1O1O2N000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O10000O1001M2N2N3M2N2N3M2N3M2M3N3L3K6I6J6K6I6K5]Md2N101N100_J^Fc0d9\\O\\Fc0e9]O\\Fb0e9]O[Fb0f9^O[Fa0e9^O\\Fa0f9^O[FaNMdNi9k2ZFaNObNg9m2[FaNO_Nh9o2YFbN1]Nf9P3[FbN1[Nf9R3YFdN2XNe9T3ZFcN3VNd9W3YFcN5oMi9W3XFjN0gMQ:R3YFWOH_MX:U3SF[OGXM_:Y3kE@GoLg:\\3eEDZ;6hDKZ;1gDO\\;LgD3\\;HfD9\\;CfD<];_OeDa0^;ZOeDf0];TOfDl0];oNfDP1];kMPCOc1W2`;PMfCe0l0Z2b<dM^C\\2e<`M]C^2f<_M[C`2i<]MXCa2k<\\MVCd2m<XMTCg2P=VMPCi2S=TMnBk2U?O0O1O101N1O1O100N3M2O1N2O1N3M2O1N2O1N2O1001O00001O00001O00001O00001O00001OO1O100O1C=00O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000001O00000O100000000O100000000O100000000O100000000O100000000O1000000O10001O000O10001O0001O01O01O000010O0000010O0001O01O01O000010O0001O01O01O000010O0001O01O01O00010O00001O01O000O101O0O100O2O0O100O104K;F1O0O100O100O2O0O10000O100O100O100O10002M2O2M2O2M_T_1"}, "label": "the woman in green"}]}
{"id": 86075, "image": "COCO_train2014_000000086075.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a white shirt and black shorts\"."}], "task": "refering", "answer_template": "The \"a girl in a white shirt and black shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 177, 178, 179, 180, 193, 194, 195, 196, 197, 229, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 294, 295, 296, 297, 298, 299, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 345, 346, 347, 348, 362, 363, 364, 365, 379, 380, 381, 382], "bbox": [0.3086041666666667, 0.413484375, 0.8898958333333333, 0.9887656249999999], "iscrowd": 0, "area": 38235.16125, "rle": {"size": [640, 480], "counts": "eSm22gc07I7oKGTD`0l;O^C8a<8iBOW=?TBHk=h0_A@`>o0j@XOU?j20000O100O10000O100O100O10000O1VN]KWDc4i;fKmC[4T<h100001O001O00001O001O0000kLaIjH^6U7cIkH]6T7eIkH[6S7gInHX6Q7jInHV6Q7lIoHS6o6oIQIQ6n6QJQIo5n6RJSIm5k6VJTIj5k6WJUIi5j6XJUIi5i6YJUIi5j6XJUIi5j6XJUIi5i6YJUIi5j6XJUIi5j6XJTIj5j6YJTIh5k6YJSIi5l6XJSIi5k6YJTIh5k6YJSIi5l6XJSIi5k6YJSIi5l6XJSIi5l6XJRIj5l6XJSIi5l6XJRIj5m6WJRIj5l6XJSIi5l6XJRIj5m6WJRIj5l6XJRIj5m6WJRIj5m6WJQIk5m6WJRIj5m6WJSIi5l6XJTIi5j6XJVIh5i6YJVIh5j6XJVIh5i6YJWIg5i6YJWIg5i6YJWIg5h6ZJXIf5h6ZJWIg5i6YJWIg5h6ZJXIf5h6ZJXIf5g6[JYIe5g6[JYIe5g6[JXIf5g6[JYIe5g6[JYIe5g6[JYI_4\\NaJZ8Q1ZI[4gN\\Jo7Y1YIZ4ROUJe7a1YIW4]OPJZ7i1YIT4X8lKhGQ4\\8nKdGo3_8QLaGl3c8SL\\Gl3f8TLZGi3i8WLWGg3l8XLTGf3n8ZLRGc3Q9]LoFa3T9^LkFa3W9_LiF^3Z9bLfF\\3]9cLcF[3_9eLaFY3a9gL^FW3f9hLZFV3h9jLXFT3k9kLUFR3n9nLRFP3P:PMPFn2S:QMlEm2W:SMiEk2Y:UMgEi2\\:VMdEg2_:YMaEe2a:[M_Eb2e:]MZEb2h:^MXE`2j:`MVE]2n:bMRE\\2P;dMPEY2T;fMkDY2W;gMiDV2Z;kMeDS2^;lMbDQ2a;oM_Do1c;QN]Dl1g;SNXDf0QOZNi<P1VDf0P=ZOPCf0Q=YOoBg0Q=YOoBg0Q=YOoBg0R=XOmBi0S=WOmBi0T=VOlBi0U=WOkBi0U=WOkBi0V=VOjBj0V=VOiBk0W=UOiBk0X=TOhBk0Y=UOgBk0Y=UOgBj0[=UOdBl0\\=SOeBm0[=nNjBQ1X=iNmBW1U=dNnB[1e=oM_BQ2h?00O100000001N10000000000O100000001N100000000O100000000O2O001O001O0O2O1O001O0O2O1O001O1N101O001O1N100000O10O0O2O0O2N101N1010O010O00100O010O010O10O010O0100O010O010O1O010O010O10O0100O100O10\\O`]OH_b08f]ODZb0;g]OEXb0:i]OGWb08i]OIVb06k]OKUb04d]O]O2`0\\b04e]OM]b0Ne]O3\\b0Je]O7]b0Ed]O;^b0Bc]O=ab0_O`]Oa0ob0O1O1O1O100O1O1O2N3M2O2M2NPhP1"}, "label": "a girl in a white shirt and black shorts"}]}
{"id": 249921, "image": "COCO_train2014_000000249921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby zebra walking in front of another zebra\"."}], "task": "refering", "answer_template": "The \"a baby zebra walking in front of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 116, 130, 131, 132, 133, 134, 149, 150, 152], "bbox": [0.24328, 0.47002666666666665, 0.4849, 0.6959733333333333], "iscrowd": 0, "area": 5277.770200000002, "rle": {"size": [375, 500], "counts": "Rl\\1=R;=F6L4N3M2OO0O2N101N2O1O1N2O1O1N2O1O1O1O001O1O10O2O001N2O001N2O1O1N2O1O1N2UFA\\8`0TGLo8[12O2M3N3L3N2N2N2N2N2N2N2N2N2N2O12N2N2N3M2N2N2N2N6J5KoNdFG\\98gFFY9:hFEX9;jFCV9<oFAP9?UG\\Ok8c0ZGYOf8g0^GUOb8j0cGRO]8n0gGnNX8S1V1O100O10000O1000O1000O12N2N1O1O2N1O1O2N1O1O2N103M4L3LPOXF5e9I^F:^9DeF>X9@kF>V9AlF=T9BcF_O3m0]9KcF4_9KaF3b9L^F3d9K]F4f9JaFMb92lF_OV9?S1O3M3L5L4LcXn2"}, "label": "a baby zebra walking in front of another zebra"}]}
{"id": 249921, "image": "COCO_train2014_000000249921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young zebra walking\"."}], "task": "refering", "answer_template": "The \"young zebra walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 116, 130, 131, 132, 133, 134, 149, 150, 152], "bbox": [0.24328, 0.47002666666666665, 0.4849, 0.6959733333333333], "iscrowd": 0, "area": 5277.770200000002, "rle": {"size": [375, 500], "counts": "Rl\\1=R;=F6L4N3M2OO0O2N101N2O1O1N2O1O1N2O1O1O1O001O1O10O2O001N2O001N2O1O1N2O1O1N2UFA\\8`0TGLo8[12O2M3N3L3N2N2N2N2N2N2N2N2N2N2O12N2N2N3M2N2N2N2N6J5KoNdFG\\98gFFY9:hFEX9;jFCV9<oFAP9?UG\\Ok8c0ZGYOf8g0^GUOb8j0cGRO]8n0gGnNX8S1V1O100O10000O1000O1000O12N2N1O1O2N1O1O2N1O1O2N103M4L3LPOXF5e9I^F:^9DeF>X9@kF>V9AlF=T9BcF_O3m0]9KcF4_9KaF3b9L^F3d9K]F4f9JaFMb92lF_OV9?S1O3M3L5L4LcXn2"}, "label": "young zebra walking"}]}
{"id": 249921, "image": "COCO_train2014_000000249921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing next to a baby zebra on a grassy plain\"."}], "task": "refering", "answer_template": "The \"a zebra standing next to a baby zebra on a grassy plain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158], "bbox": [0.44032, 0.37213333333333337, 0.8295, 0.6997866666666667], "iscrowd": 0, "area": 11113.938, "rle": {"size": [375, 500], "counts": "ng`22a;5K5K5L5J5\\OVOWFm0h9UOTFm0k9UORFl0o9VOlEl0T:VOiEk0X:9001O001O001O001O001O1O001O1O1O1010O0000010O000010O000010O0000010O1O00100O001O100O001O10O01O102M3M3M4cGZNb6l1VIUNj6T2kHmMV7\\2_HfMa7d2TH]Mm7k2hGWMX8T300001O0000001J5C=H8I7I8G8O1000000O1011N2N1O2N1O2N2O0O2N1O2N2N1OZOdFSO[9l0gFSOX9m0jFSOT9m0oFROo8o0RGPOm8P1UGPOi8P1ZGnNe8R1]GnNa8R1aGmN^8S1eGlNY8T1iGlNU8T1mGkNQ8W1PHiNn7W1UHgNj7Y1XHfNg7Z1[HeNd7[1_HbN`7_1bH`N]7`1fH]NZ7c1hH\\NW7e1kHYNT7g1^10O100000001N1000O1O100O010O1O100O00100O1O100O00100O1O1O1O0011O1O2N1N3O0O1O2RGZNh7f1TH`Ni7b1SHbNl7^1PHgNn7Z1oGkNo7V1lGkNV8V1gGjN[8W1aGjNb8W1ZGjNg8R22N1007J9FN3M2O1N3M2N4L4SOm0L4M3M2M4N21O13M3L3N3J5I8G8I8HZmn0"}, "label": "a zebra standing next to a baby zebra on a grassy plain"}]}
{"id": 249921, "image": "COCO_train2014_000000249921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large black and white zebra standing in a field beside a smaller zebra\"."}], "task": "refering", "answer_template": "The \"large black and white zebra standing in a field beside a smaller zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158], "bbox": [0.44032, 0.37213333333333337, 0.8295, 0.6997866666666667], "iscrowd": 0, "area": 11113.938, "rle": {"size": [375, 500], "counts": "ng`22a;5K5K5L5J5\\OVOWFm0h9UOTFm0k9UORFl0o9VOlEl0T:VOiEk0X:9001O001O001O001O001O1O001O1O1O1010O0000010O000010O000010O0000010O1O00100O001O100O001O10O01O102M3M3M4cGZNb6l1VIUNj6T2kHmMV7\\2_HfMa7d2TH]Mm7k2hGWMX8T300001O0000001J5C=H8I7I8G8O1000000O1011N2N1O2N1O2N2O0O2N1O2N2N1OZOdFSO[9l0gFSOX9m0jFSOT9m0oFROo8o0RGPOm8P1UGPOi8P1ZGnNe8R1]GnNa8R1aGmN^8S1eGlNY8T1iGlNU8T1mGkNQ8W1PHiNn7W1UHgNj7Y1XHfNg7Z1[HeNd7[1_HbN`7_1bH`N]7`1fH]NZ7c1hH\\NW7e1kHYNT7g1^10O100000001N1000O1O100O010O1O100O00100O1O100O00100O1O1O1O0011O1O2N1N3O0O1O2RGZNh7f1TH`Ni7b1SHbNl7^1PHgNn7Z1oGkNo7V1lGkNV8V1gGjN[8W1aGjNb8W1ZGjNg8R22N1007J9FN3M2O1N3M2N4L4SOm0L4M3M2M4N21O13M3L3N3J5I8G8I8HZmn0"}, "label": "large black and white zebra standing in a field beside a smaller zebra"}]}
{"id": 274499, "image": "COCO_train2014_000000274499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in uniform sitting back and looking up smiling at someone else\"."}], "task": "refering", "answer_template": "The \"woman in uniform sitting back and looking up smiling at someone else\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.142203125, 0.3066875, 0.52403125, 0.796625], "iscrowd": 0, "area": 34957.922300000006, "rle": {"size": [480, 640], "counts": "kmZ1?`>h0XO8H9F:G7I6J6J6I7J3M1O2N1N2O1O1O1O1N2O1O1O1O1N2PN[LdHf3Z7aL`H`3^7fL]H[3b7jLYHW3e7oLUHS3j7RMQHP3l7UMoGm2o7YMhGl2W8XMaGm2]8XM[Gm2d8h1N2N2O1N2O1000000O10000O10001N100O1O100O100O100O1O100O100O1O100O2O0O1O100O100O1O100O1000000000000000001O0000000000000000000001O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O2N3M3M3M3M1mKWGm1j8QNWGo1j8oMXGQ2i8lMXGT2i8jMXGV2i8hMXGX2j8eMXGZ2k8cMUG]2m8`MTG`2n8]MTGb2n8WLWG=L\\3e9bL\\F^3e9`L\\F`3f9]L\\Fb3e9\\L\\Fd3e9[L[Fe3f9YL[Fg3f9WL\\Fh3e9VL\\Fj3e9TL\\Fl3]:0O000001O0000001O001O001O001O001O001O010O001O001O001O001O001O001O001O000010O01O001O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1N3N1O1O1O2M2O1O1O1O2M2O1O1O2N1N2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O3M5J5L5K5K5K`c^4"}, "label": "woman in uniform sitting back and looking up smiling at someone else"}]}
{"id": 274499, "image": "COCO_train2014_000000274499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a seated scout\"."}], "task": "refering", "answer_template": "The \"a seated scout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.142203125, 0.3066875, 0.52403125, 0.796625], "iscrowd": 0, "area": 34957.922300000006, "rle": {"size": [480, 640], "counts": "kmZ1?`>h0XO8H9F:G7I6J6J6I7J3M1O2N1N2O1O1O1O1N2O1O1O1O1N2PN[LdHf3Z7aL`H`3^7fL]H[3b7jLYHW3e7oLUHS3j7RMQHP3l7UMoGm2o7YMhGl2W8XMaGm2]8XM[Gm2d8h1N2N2O1N2O1000000O10000O10001N100O1O100O100O100O1O100O100O1O100O2O0O1O100O100O1O100O1000000000000000001O0000000000000000000001O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O2N3M3M3M3M1mKWGm1j8QNWGo1j8oMXGQ2i8lMXGT2i8jMXGV2i8hMXGX2j8eMXGZ2k8cMUG]2m8`MTG`2n8]MTGb2n8WLWG=L\\3e9bL\\F^3e9`L\\F`3f9]L\\Fb3e9\\L\\Fd3e9[L[Fe3f9YL[Fg3f9WL\\Fh3e9VL\\Fj3e9TL\\Fl3]:0O000001O0000001O001O001O001O001O001O010O001O001O001O001O001O001O001O000010O01O001O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1N3N1O1O1O2M2O1O1O1O2M2O1O1O2N1N2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O3M5J5L5K5K5K`c^4"}, "label": "a seated scout"}]}
{"id": 274499, "image": "COCO_train2014_000000274499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"scout man standing up\"."}], "task": "refering", "answer_template": "The \"scout man standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 36, 37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 226, 227, 241, 242, 243, 249, 250, 264, 265, 266, 272, 273, 287, 288, 289, 295, 296, 311, 312], "bbox": [0.456734375, 0.0471875, 0.891578125, 0.7820208333333334], "iscrowd": 0, "area": 47477.17535000001, "rle": {"size": [480, 640], "counts": "`SY41l>6I8I6I7J7H7J7H7J6J7H7J7H7J6I8I6I7J7H7J7H7J7I7H7J7H8I6I8I6M4K5L3L5L3L5L4K4M2M100O100O2O0O100O100O100O10000O100O100O100O101N100O100O100O100O100O100O100O100O2O0O10000O100O100O100O1oM]HeLc7Z3ZIiKg6V4R2O1OH8ROn0ROn0ROn0UOk0100O1O1O1O1O1O1O1O1O1O1O100O1O1O1000000001O0000001O000000001N1000001O0000001O000000001O0000001O0000001O0000002N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2NU1kN00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0O101O00001O0000001O00001O00001O2N3M3M2N3M3M3M2N3M3M2M4M3M3M2N3M3M3M2N3M3M2N3M2N2N3M2N3M2N3M2N3M2N2^3cLL3N3M2M4M2N3M2M4M2N2M4M2N3M2M4M2N3H7fNZ1eNWTP1"}, "label": "scout man standing up"}]}
{"id": 274499, "image": "COCO_train2014_000000274499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man to the right\"."}], "task": "refering", "answer_template": "The \"man to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 36, 37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 226, 227, 241, 242, 243, 249, 250, 264, 265, 266, 272, 273, 287, 288, 289, 295, 296, 311, 312], "bbox": [0.456734375, 0.0471875, 0.891578125, 0.7820208333333334], "iscrowd": 0, "area": 47477.17535000001, "rle": {"size": [480, 640], "counts": "`SY41l>6I8I6I7J7H7J7H7J6J7H7J7H7J6I8I6I7J7H7J7H7J7I7H7J7H8I6I8I6M4K5L3L5L3L5L4K4M2M100O100O2O0O100O100O100O10000O100O100O100O101N100O100O100O100O100O100O100O100O2O0O10000O100O100O100O1oM]HeLc7Z3ZIiKg6V4R2O1OH8ROn0ROn0ROn0UOk0100O1O1O1O1O1O1O1O1O1O1O100O1O1O1000000001O0000001O000000001N1000001O0000001O000000001O0000001O0000001O0000002N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2NU1kN00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0O101O00001O0000001O00001O00001O2N3M3M2N3M3M3M2N3M3M2M4M3M3M2N3M3M3M2N3M3M2N3M2N2N3M2N3M2N3M2N3M2N2^3cLL3N3M2M4M2N3M2M4M2N2M4M2N3M2M4M2N3H7fNZ1eNWTP1"}, "label": "man to the right"}]}
{"id": 430148, "image": "COCO_train2014_000000430148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a red leather sofa\"."}], "task": "refering", "answer_template": "The \"this is a red leather sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 385, 386, 387, 388, 389], "bbox": [0.743546875, 0.6152916666666666, 0.9725781250000001, 0.9787291666666665], "iscrowd": 0, "area": 19691.937900000004, "rle": {"size": [480, 640], "counts": "d_o6b1[=U1kN4L3M3M3M3M3M3M3M3M3M4L3M3M3M3000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001N1O1O1O1O100O1O1O1O2N1O100O1O1O1O1O1O1O2O0O1O1O1O1O1O1O100O2N1O1001O1O2N1O1O2O0O1O2N1O2N1O1O2N1O101N1O1O2N1O1O2N1O1O2O0O1O2I6G9F;E:F:F;F9F;E:F:G:E:FZS8"}, "label": "this is a red leather sofa"}]}
{"id": 430148, "image": "COCO_train2014_000000430148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red couch facing the fireplace\"."}], "task": "refering", "answer_template": "The \"the red couch facing the fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 385, 386, 387, 388, 389], "bbox": [0.743546875, 0.6152916666666666, 0.9725781250000001, 0.9787291666666665], "iscrowd": 0, "area": 19691.937900000004, "rle": {"size": [480, 640], "counts": "d_o6b1[=U1kN4L3M3M3M3M3M3M3M3M3M4L3M3M3M3000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001N1O1O1O1O100O1O1O1O2N1O100O1O1O1O1O1O1O2O0O1O1O1O1O1O1O100O2N1O1001O1O2N1O1O2O0O1O2N1O2N1O1O2N1O101N1O1O2N1O1O2N1O1O2O0O1O2I6G9F;E:F:F;F9F;E:F:G:E:FZS8"}, "label": "the red couch facing the fireplace"}]}
{"id": 405579, "image": "COCO_train2014_000000405579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a green shirt that is using a fork and knife to eat\"."}], "task": "refering", "answer_template": "The \"a person in a green shirt that is using a fork and knife to eat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 159, 160, 181, 182, 183, 205, 206, 229], "bbox": [0.764640625, 0.006744730679156909, 1.0, 0.6449414519906322], "iscrowd": 0, "area": 21802.368949999996, "rle": {"size": [427, 640], "counts": "RQ\\61X=2N3M2M4M2N3M3M2M4M2N3M2N3L4M2N3M2N3L3N3M3M2N3L3N3M2N3M3L3N3M2N3M2M4M3M2N3L3N3M2N3M3L3N3M2N3M2M4M3M2N3M2M4M2N3M3N101O001O001O0000000000O10000000000O1000000000000O1000000000000O10000000gH]Kj5c4UJ`Ki5`4WJaKh5^4YJcKf5]4ZJdKe5\\4ZJfKe5Z4[JgKd5Y4\\JiKb5W4^JjKa5V4^JlKa5S4`JnK_5R4aJoK^5Q4bJQL\\5o3cJSL\\5m3dJTL[5l3eJULZ5k3fJVLY5i3gJYLX5g3hJ[LV5e3jJ\\LU5d3kJ]LT5c3kJ_LT5a3lJ`LS5`3mJaLR5^3oJdLo4\\3PKfLo4Z3QKgLn4Y3RKhLm4X3SKiLl4W3SKlLk4T3UKmLj4R3WKoLh4Q3XKPMg4P3XKRMg4n2YKSMf4m2ZKUMd4k2\\KVMc4j2\\KXMc4g2^KZMa4f2_K[M`4e2`K]M^4c2aK_M^4a2bK`M]4`2cKaM\\4_2dKbM[4]2eKeMZ4[2fKgMX4Y2hKhMW4X2iKiMV4W2iKkMV4U2jKlMU4T2kKmMT4R2mKPNQ4P2nKRNQ4n1oKSNP4m1PLTNo3l1QLUNn3k1aLhM]3N`0`N\\O"}, "label": "a person in a green shirt that is using a fork and knife to eat"}]}
{"id": 405579, "image": "COCO_train2014_000000405579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pants of the man cutting the food\"."}], "task": "refering", "answer_template": "The \"the pants of the man cutting the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 159, 160, 181, 182, 183, 205, 206, 229], "bbox": [0.764640625, 0.006744730679156909, 1.0, 0.6449414519906322], "iscrowd": 0, "area": 21802.368949999996, "rle": {"size": [427, 640], "counts": "RQ\\61X=2N3M2M4M2N3M3M2M4M2N3M2N3L4M2N3M2N3L3N3M3M2N3L3N3M2N3M3L3N3M2N3M2M4M3M2N3L3N3M2N3M3L3N3M2N3M2M4M3M2N3M2M4M2N3M3N101O001O001O0000000000O10000000000O1000000000000O1000000000000O10000000gH]Kj5c4UJ`Ki5`4WJaKh5^4YJcKf5]4ZJdKe5\\4ZJfKe5Z4[JgKd5Y4\\JiKb5W4^JjKa5V4^JlKa5S4`JnK_5R4aJoK^5Q4bJQL\\5o3cJSL\\5m3dJTL[5l3eJULZ5k3fJVLY5i3gJYLX5g3hJ[LV5e3jJ\\LU5d3kJ]LT5c3kJ_LT5a3lJ`LS5`3mJaLR5^3oJdLo4\\3PKfLo4Z3QKgLn4Y3RKhLm4X3SKiLl4W3SKlLk4T3UKmLj4R3WKoLh4Q3XKPMg4P3XKRMg4n2YKSMf4m2ZKUMd4k2\\KVMc4j2\\KXMc4g2^KZMa4f2_K[M`4e2`K]M^4c2aK_M^4a2bK`M]4`2cKaM\\4_2dKbM[4]2eKeMZ4[2fKgMX4Y2hKhMW4X2iKiMV4W2iKkMV4U2jKlMU4T2kKmMT4R2mKPNQ4P2nKRNQ4n1oKSNP4m1PLTNo3l1QLUNn3k1aLhM]3N`0`N\\O"}, "label": "the pants of the man cutting the food"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 94, 95, 96, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234], "bbox": [0.0, 0.21278688524590164, 0.316828125, 0.7187353629976581], "iscrowd": 0, "area": 25212.284050000006, "rle": {"size": [427, 640], "counts": "Z8f0o;f0N2O1N2O2M2N3N2M2O2M3M2N3N2M2O2M2N2O1N3N1N2N2O2M2O1N2O1N3N1N2O2M3N2M4L2O1O1O2N00100O1O1O10O0100O100lNTLiHk3W7WLgHj3X7XLgHh3X7ZLgHf3W7]LgHd3m6hLSIX3l6jL\\H\\O>i3V7lLZH_O>e3W7TMmHl2Q7TMQIl2n6WMPIi2o6[MnHe2P7_MmHb2R7fMgHY2X7jMgHV2X7kMhHU2W7mMhHS2P7VLoHh10R2P7YLmHf12R2Q7X2O010000000N2L4L3101O00001O0O101O0O10001N1000LiITIV6m6kIRIU6n6lIRIS6P7lIoHS6S722N0O1000O012N1O2M2O1O1O1O100001O001N101N2N1O2`KlHe2V7TLiHd01X3X7QLmHb0K^3Y7mKQI`0Gc3Z7jKTI<Ei3[7gKVI8BP4X8hKjGW4f8N2O1N2N2N2O1O1O10000O1000O01O100O1O100O1O1L5G9N101O1O0O2O1O2N1O2N2N101N21N4M00M3N1O2M3N1O2M3N1N3M3N1N3M3N2N2M3N2N3M2M3N3M2M3N2N3L3M3M4L3M3N3L3M3M4M2M3NRoe5"}, "label": "a woman in a white shirt"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt with curly hair\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt with curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 94, 95, 96, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234], "bbox": [0.0, 0.21278688524590164, 0.316828125, 0.7187353629976581], "iscrowd": 0, "area": 25212.284050000006, "rle": {"size": [427, 640], "counts": "Z8f0o;f0N2O1N2O2M2N3N2M2O2M3M2N3N2M2O2M2N2O1N3N1N2N2O2M2O1N2O1N3N1N2O2M3N2M4L2O1O1O2N00100O1O1O10O0100O100lNTLiHk3W7WLgHj3X7XLgHh3X7ZLgHf3W7]LgHd3m6hLSIX3l6jL\\H\\O>i3V7lLZH_O>e3W7TMmHl2Q7TMQIl2n6WMPIi2o6[MnHe2P7_MmHb2R7fMgHY2X7jMgHV2X7kMhHU2W7mMhHS2P7VLoHh10R2P7YLmHf12R2Q7X2O010000000N2L4L3101O00001O0O101O0O10001N1000LiITIV6m6kIRIU6n6lIRIS6P7lIoHS6S722N0O1000O012N1O2M2O1O1O1O100001O001N101N2N1O2`KlHe2V7TLiHd01X3X7QLmHb0K^3Y7mKQI`0Gc3Z7jKTI<Ei3[7gKVI8BP4X8hKjGW4f8N2O1N2N2N2O1O1O10000O1000O01O100O1O100O1O1L5G9N101O1O0O2O1O2N1O2N2N101N21N4M00M3N1O2M3N1O2M3N1N3M3N1N3M3N2N2M3N2N3M2M3N3M2M3N2N3L3M3M4L3M3N3L3M3M4M2M3NRoe5"}, "label": "a woman in a white shirt with curly hair"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in the background behind the guy in the green polo shirt\"."}], "task": "refering", "answer_template": "The \"a person in the background behind the guy in the green polo shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 136, 137, 159, 160, 182, 183, 205, 206, 229], "bbox": [0.91346875, 0.00468384074941452, 1.0, 0.6243793911007026], "iscrowd": 0, "area": 9955.35215000001, "rle": {"size": [427, 640], "counts": "joc75T=6J5K6I6K6`GoN\\4V1bKkNZ4Z1cKhNX4_1eKbNW4c1fK_NV4g1gKYNV4l1hKUNT4Q2hKQNS4U2kKlMQ4Z2kKgMR4^2lKcMP4b2cKjMX4d2oJlMm4d2WJkMf5e2^IkM^6d4M3N1O2N2M3N2N1O2M3N2N1O2M3N2N1N3N2N2N2M2OO1N2O0O2O1N2N101N2O1N1O2Oj3ULWL"}, "label": "a person in the background behind the guy in the green polo shirt"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of a man in a black tshirt\"."}], "task": "refering", "answer_template": "The \"the arm of a man in a black tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 136, 137, 159, 160, 182, 183, 205, 206, 229], "bbox": [0.91346875, 0.00468384074941452, 1.0, 0.6243793911007026], "iscrowd": 0, "area": 9955.35215000001, "rle": {"size": [427, 640], "counts": "joc75T=6J5K6I6K6`GoN\\4V1bKkNZ4Z1cKhNX4_1eKbNW4c1fK_NV4g1gKYNV4l1hKUNT4Q2hKQNS4U2kKlMQ4Z2kKgMR4^2lKcMP4b2cKjMX4d2oJlMm4d2WJkMf5e2^IkM^6d4M3N1O2N2M3N2N1O2M3N2N1O2M3N2N1N3N2N2N2M2OO1N2O0O2O1N2N101N2O1N1O2Oj3ULWL"}, "label": "the arm of a man in a black tshirt"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in a red and black plaid shirt stares straight ahead\"."}], "task": "refering", "answer_template": "The \"a young man in a red and black plaid shirt stares straight ahead\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 124, 125, 126, 147, 148, 149, 170, 171, 172, 173, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 263, 264, 265, 266], "bbox": [0.356828125, 0.2876346604215457, 0.60271875, 0.7707962529274005], "iscrowd": 0, "area": 19944.54675, "rle": {"size": [427, 640], "counts": "^[o21o<>D=G8H8I6ZOf0L4M4L2M101O1O1O10O10O1000O010000O01000O100O2O0O100O100SNkMcHW2W7SNdHm1X7[NcHf1Y7ROSHn0j7[2M4L3M3O1O1O101N1O101N2N101N2N100O101O0000001O000000000001O00000000001O00001O001O00001O001O00001O001ULdHe1]7XNfHg1]7TNeHm1\\7oMgHP2[7mMgHR2[7kMfHV2[7gMgHX2[7eMgHZ2[7cMfH]2\\7aMeH_2]7]MeHb2]7[MdHf2]7hLTIW3a80O0100O010O001O010O001O001O10O01O001O001O1O001O001O1O1O1O001O1O1O1O1O2N2N2N2N2M3N4L4L3M4K5L4L3M4K5L3M6J6J7I7H8H8H[cY3"}, "label": "a young man in a red and black plaid shirt stares straight ahead"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.439453125, 0.07318501170960187, 1.0, 1.0], "iscrowd": 0, "area": 71829.36764999997, "rle": {"size": [427, 640], "counts": "fae31X=4M3L4M3L3N3L4M3M2N2M2O2N1O2N2N1O2N1O2N101NDXDWOh;i0YDVOg;j0[DTOd;m0]DPOe;P1]DlNf;S191O0000000000000000000010O001O001O0010O01O001O_OUObDl0[;WOeDh0Z;ZOeDf0Y;]OfDc0X;@gD`0W;ChD>U;DkD<T;ElD;R;FoD:o:GRE9l:HUE9h:HYE8f:G\\E9d:E^E;a:CbE>\\:AfE?Z:_OiE`0W:_OjEb0U:^OlEa0T:^OmEc0Q:]OjEA@Q1f:^OgEDCn0f:]OeEHFk0d:]OeEIGj0d:]OdEJIh0c:^OcEKJh0b:]OcELLf0a:^ObEMMe0a:^OaENOd0^:_OcEMOd0^:@aEL3c0\\:A`EL5d0Z:@`EM6c0Z:@_EM8c0Y:@_EL:d0V:@_EL<d0U:@^EM=c0T:A^EL?d0R:1nEOR:0oE0Q:0oE1P:NQF2o9NQF2o9NQF3n9LSF4l9MTF3l9LUF5j9KVF5j9JWF6i9JWF7h9IXF7h9HYF8g9HYF9e9G\\F9d9G]F8c9H]F9b9F_F:a9F_F:a9E`F<_9DaF<_9CbF=]9DcF=\\9BeF>[9BeF>[9AfF`0X9@iF`0W9@iF`0W9_OjFb0U9]OlFc0h1ROT4<TJb0a1ZOZ43VJc0_1\\OZ40XJe0\\1^OY4N[Jd0Z1AZ4J]Je0X1CZ4G_Jg0U1D[4EaJf0S1GZ4CdJf0P1I\\4AcJh0o0I]4^ObJl0P1H]4\\OaJn0P1I]4YOaJR1P1J[4TOcJT1Q1NW4mNfJW1R1>g3\\NUKX1S1>g3ZNSK[1U1=g3XNQK^1W1<g3VNPK`1Y1;e3VNoJb1[1:fNdMX4`0eKd1\\1:dNeMZ4=cKg1_19_NhM]48cKi1`19\\NjM_45bKk1b1n0l2WM`Km1d1l0l2OSM2l2OTM1k20TM2j2OUM3i2NWM3g2NXM4`NYLQ4d3_M8_2IaM8\\2KcM6Y2NgM2aNUL[3m3TN8f1MYN5a1O`N2Z13eNNV17jNJP1:POGk0>TODh0>YOBd0a0[O@b0a0_O@?a0B@9b0HFJ?8BC`0?A]O`0d0AYOa0h0_OUOb0m0^OPOd0Q1\\OmNe0T1[OjNf0V1[OhNf0Y1ZOeNg0\\1XOcNi0^1WO`Ni0a1WO_Ni0b1WO]Ni0d1WO[Ni0f1VO[Ni0e1XOZNh0g1XOYNg0h1XOXNh0i1WOXNh0h1YOWNi0h1VOYNl0e1TOZNn0e1QO\\NP1c1oN]NS1a1nN_NS1`1nN^NS1b1mN^NS1b1nN\\NS1c1oN\\NQ1d1oN[NR1e1POYNP1g1SOUNn0j1Z500001O00nNcD6];IdD7\\;HeD8\\;GdD:\\;DeD<\\;BeD>\\;@eD`0];]OdDc0i;oNXDR1P<000000001O000000001O000000001O000000001O0000001O0000001O1O2N1O1O2N1O1O1O2N2N3M2N3M3M3M4fFQNP7R2_H\\Na7h3O0000010O001O1O2N2N1O2N3M2N3M1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N2N2N1O1O2N1O2N1O2N1O2N1O2N1O2N2N2N2N2N2N"}, "label": "a man wearing a green shirt"}]}
{"id": 299085, "image": "COCO_train2014_000000299085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with dark hair wearing a green polo shirt and a watch on his left hand\"."}], "task": "refering", "answer_template": "The \"a man with dark hair wearing a green polo shirt and a watch on his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.439453125, 0.07318501170960187, 1.0, 1.0], "iscrowd": 0, "area": 71829.36764999997, "rle": {"size": [427, 640], "counts": "fae31X=4M3L4M3L3N3L4M3M2N2M2O2N1O2N2N1O2N1O2N101NDXDWOh;i0YDVOg;j0[DTOd;m0]DPOe;P1]DlNf;S191O0000000000000000000010O001O001O0010O01O001O_OUObDl0[;WOeDh0Z;ZOeDf0Y;]OfDc0X;@gD`0W;ChD>U;DkD<T;ElD;R;FoD:o:GRE9l:HUE9h:HYE8f:G\\E9d:E^E;a:CbE>\\:AfE?Z:_OiE`0W:_OjEb0U:^OlEa0T:^OmEc0Q:]OjEA@Q1f:^OgEDCn0f:]OeEHFk0d:]OeEIGj0d:]OdEJIh0c:^OcEKJh0b:]OcELLf0a:^ObEMMe0a:^OaENOd0^:_OcEMOd0^:@aEL3c0\\:A`EL5d0Z:@`EM6c0Z:@_EM8c0Y:@_EL:d0V:@_EL<d0U:@^EM=c0T:A^EL?d0R:1nEOR:0oE0Q:0oE1P:NQF2o9NQF2o9NQF3n9LSF4l9MTF3l9LUF5j9KVF5j9JWF6i9JWF7h9IXF7h9HYF8g9HYF9e9G\\F9d9G]F8c9H]F9b9F_F:a9F_F:a9E`F<_9DaF<_9CbF=]9DcF=\\9BeF>[9BeF>[9AfF`0X9@iF`0W9@iF`0W9_OjFb0U9]OlFc0h1ROT4<TJb0a1ZOZ43VJc0_1\\OZ40XJe0\\1^OY4N[Jd0Z1AZ4J]Je0X1CZ4G_Jg0U1D[4EaJf0S1GZ4CdJf0P1I\\4AcJh0o0I]4^ObJl0P1H]4\\OaJn0P1I]4YOaJR1P1J[4TOcJT1Q1NW4mNfJW1R1>g3\\NUKX1S1>g3ZNSK[1U1=g3XNQK^1W1<g3VNPK`1Y1;e3VNoJb1[1:fNdMX4`0eKd1\\1:dNeMZ4=cKg1_19_NhM]48cKi1`19\\NjM_45bKk1b1n0l2WM`Km1d1l0l2OSM2l2OTM1k20TM2j2OUM3i2NWM3g2NXM4`NYLQ4d3_M8_2IaM8\\2KcM6Y2NgM2aNUL[3m3TN8f1MYN5a1O`N2Z13eNNV17jNJP1:POGk0>TODh0>YOBd0a0[O@b0a0_O@?a0B@9b0HFJ?8BC`0?A]O`0d0AYOa0h0_OUOb0m0^OPOd0Q1\\OmNe0T1[OjNf0V1[OhNf0Y1ZOeNg0\\1XOcNi0^1WO`Ni0a1WO_Ni0b1WO]Ni0d1WO[Ni0f1VO[Ni0e1XOZNh0g1XOYNg0h1XOXNh0i1WOXNh0h1YOWNi0h1VOYNl0e1TOZNn0e1QO\\NP1c1oN]NS1a1nN_NS1`1nN^NS1b1mN^NS1b1nN\\NS1c1oN\\NQ1d1oN[NR1e1POYNP1g1SOUNn0j1Z500001O00nNcD6];IdD7\\;HeD8\\;GdD:\\;DeD<\\;BeD>\\;@eD`0];]OdDc0i;oNXDR1P<000000001O000000001O000000001O000000001O0000001O0000001O1O2N1O1O2N1O1O1O2N2N3M2N3M3M3M4fFQNP7R2_H\\Na7h3O0000010O001O1O2N2N1O2N3M2N3M1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N2N2N1O1O2N1O2N1O2N1O2N1O2N1O2N2N2N2N2N2N"}, "label": "a man with dark hair wearing a green polo shirt and a watch on his left hand"}]}
{"id": 430159, "image": "COCO_train2014_000000430159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dirty open toilet\"."}], "task": "refering", "answer_template": "The \"a dirty open toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 105, 106, 117, 118, 125, 126, 127, 128, 129, 130, 137, 138, 139, 140, 141, 142, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177], "bbox": [0.442012012012012, 0.40224, 0.9278978978978979, 0.8292200000000001], "iscrowd": 0, "area": 17837.304400000005, "rle": {"size": [500, 333], "counts": "iRX27j>d0K5K5K5mCgNh9[1VFgNi9Z1UFgNk9Z1SFgNm9Z1PFiNo9Y1nEhNR:Y1lEhNT:Y1jEiNU:X1iEiNW:X1fEjNZ:W1dEkN[:V1cEkN]:V1aEkN_:V1^ElNb:U1ZEoNe:R1VEROj:o0RETOn:n0mDVOR;k0jDXOV;i0fDZOZ;g0aD^O^;c0^D@b;a0ZDBf;V201O00001O01O01O000010O000001O01O01O00010OO1000001O0O10001O000O10001O00000O101O0O10001N10000O101N10000O2O000O10001N101N10001O0O2O001O001N101O00N1K6K5K5L3M4M3L4L3M4L4M3L3M4L4L4M2N3O1O1O001O1N2O001O1O001OjMkEnNU:n0PFROo9k0VFSOi9j0]FUOb9h0cFVO]9f0hFZOW9b0oF\\OQ9b0SG]Ol8a0WG^Oi8`0[G_Od8?`G_O_8`0dG@[8>iG@W8>mGAR8=RHAn7=UHCj7;ZHCf7;^HDc78`HGb75bHJ_71fHM\\7OhH0Y7LjH3X7IlH4V7InH3V7InH3V7ImH4W7HmH4W7HmH4W7HmH4W7HmH4V7ImH5V7GnH5[;LW\\;"}, "label": "a dirty open toilet"}]}
{"id": 430159, "image": "COCO_train2014_000000430159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white toilet with the lid up between two other toilets\"."}], "task": "refering", "answer_template": "The \"a white toilet with the lid up between two other toilets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 105, 106, 117, 118, 125, 126, 127, 128, 129, 130, 137, 138, 139, 140, 141, 142, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177], "bbox": [0.442012012012012, 0.40224, 0.9278978978978979, 0.8292200000000001], "iscrowd": 0, "area": 17837.304400000005, "rle": {"size": [500, 333], "counts": "iRX27j>d0K5K5K5mCgNh9[1VFgNi9Z1UFgNk9Z1SFgNm9Z1PFiNo9Y1nEhNR:Y1lEhNT:Y1jEiNU:X1iEiNW:X1fEjNZ:W1dEkN[:V1cEkN]:V1aEkN_:V1^ElNb:U1ZEoNe:R1VEROj:o0RETOn:n0mDVOR;k0jDXOV;i0fDZOZ;g0aD^O^;c0^D@b;a0ZDBf;V201O00001O01O01O000010O000001O01O01O00010OO1000001O0O10001O000O10001O00000O101O0O10001N10000O101N10000O2O000O10001N101N10001O0O2O001O001N101O00N1K6K5K5L3M4M3L4L3M4L4M3L3M4L4L4M2N3O1O1O001O1N2O001O1O001OjMkEnNU:n0PFROo9k0VFSOi9j0]FUOb9h0cFVO]9f0hFZOW9b0oF\\OQ9b0SG]Ol8a0WG^Oi8`0[G_Od8?`G_O_8`0dG@[8>iG@W8>mGAR8=RHAn7=UHCj7;ZHCf7;^HDc78`HGb75bHJ_71fHM\\7OhH0Y7LjH3X7IlH4V7InH3V7InH3V7ImH4W7HmH4W7HmH4W7HmH4W7HmH4V7ImH5V7GnH5[;LW\\;"}, "label": "a white toilet with the lid up between two other toilets"}]}
{"id": 356435, "image": "COCO_train2014_000000356435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"big burger with fries in the plate\"."}], "task": "refering", "answer_template": "The \"big burger with fries in the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 79, 80, 81, 82, 83, 84, 85, 92, 93, 94, 95, 96, 97, 98, 105, 106, 107, 108, 109, 110, 111, 118, 119, 120, 121, 122, 123, 131, 132, 133, 134, 135, 145, 146, 147], "bbox": [0.10141333333333334, 0.27762, 0.61744, 0.62886], "iscrowd": 0, "area": 24629.0357, "rle": {"size": [500, 375], "counts": "Vib0a0c>Q1SOh0B4L3M3M3M3M4L3M3M3M3M3M4L3M4M2M4L2N3L3L4L4L4L4L5K4M3000000O2O000O1000000O2O000O10000O2O00000O10001O0000000000001O00000000001O00000000001O000000000000000000O1000000O101O000O1000000O2O00000O1000001N1000000O10001O0O1000001N101O001N101O001N1O2M2N3N1N3M2O2M2N3M3N1N3M2N3N1N3M2O2N1N3N1N3N1O2M2O2N1N3N1O2M3N1O2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2M3N2M4L3N3L3N3L3M4M2M4M2M4L3NRdU2"}, "label": "big burger with fries in the plate"}]}
{"id": 356435, "image": "COCO_train2014_000000356435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cheesburger piled high with lettuce\"."}], "task": "refering", "answer_template": "The \"a cheesburger piled high with lettuce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 79, 80, 81, 82, 83, 84, 85, 92, 93, 94, 95, 96, 97, 98, 105, 106, 107, 108, 109, 110, 111, 118, 119, 120, 121, 122, 123, 131, 132, 133, 134, 135, 145, 146, 147], "bbox": [0.10141333333333334, 0.27762, 0.61744, 0.62886], "iscrowd": 0, "area": 24629.0357, "rle": {"size": [500, 375], "counts": "Vib0a0c>Q1SOh0B4L3M3M3M3M4L3M3M3M3M3M4L3M4M2M4L2N3L3L4L4L4L4L5K4M3000000O2O000O1000000O2O000O10000O2O00000O10001O0000000000001O00000000001O00000000001O000000000000000000O1000000O101O000O1000000O2O00000O1000001N1000000O10001O0O1000001N101O001N101O001N1O2M2N3N1N3M2O2M2N3M3N1N3M2N3N1N3M2O2N1N3N1N3N1O2M2O2N1N3N1O2M3N1O2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2M3N2M4L3N3L3N3L3M4M2M4M2M4L3NRdU2"}, "label": "a cheesburger piled high with lettuce"}]}
{"id": 258134, "image": "COCO_train2014_000000258134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a brown ponytail and a strap over her shoulder holding a pink umbrella\"."}], "task": "refering", "answer_template": "The \"a woman with a brown ponytail and a strap over her shoulder holding a pink umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.524796875, 0.5040322580645161, 0.9037499999999999, 0.9730875576036866], "iscrowd": 0, "area": 25477.55714999999, "rle": {"size": [434, 640], "counts": "Pj^46U=7I7L4O1O1N2M3J6N2N2N2N2N2N2N2O1O1O1O1O1O1O1N2O1O1O1O100O1O1O1O1O1O100O10000O10000O100O100O100O10000O100O1000000O100000000O1000000O100O100O1O100O100O1O100O100^NYMXHh2d7eMSH[2i7jMUHW2i7mMTHT2j7oMUHQ2i7SNTHn1i7XNTHh1i7_NRHb1`7PO\\HP1b7SO\\Hn0b7UO]Hk0a7XO]Hi0a7ZO^Hf0`7]O_Hc0`7_O^Hb0b7_O\\Hb0c7g2O1O1L4N2N2O1N2O1O100O100O100O100000000000000O1000000000000000000001O00000000001O0000001O000000001O0000001O00001O1UMcH_O]77nHHS70UIOl6MXI2i6KYI5g6KYI5h6JWI7k6FVI:k6CWI=k6_OWIa0j6\\OYIc0h6ZO[Ie0g6WO\\Ih0e6VO]Ii0d6_NTHCZ1m1j6VNPHIY1P2j6SNPHJW1S2l6nMTHHS1Y2k6kMnIT2U6gMnIX2g8O1O1O001O1O2N2N00001O00001O0000001O00001O00001O001O00001O001O001O00001O001O00001O001O1O1O2N1O1O1O2N2N1O2N1O2N2N1O2N1O2N1O2N7I4L3M3M3M3M2Niki0"}, "label": "a woman with a brown ponytail and a strap over her shoulder holding a pink umbrella"}]}
{"id": 258134, "image": "COCO_train2014_000000258134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding an orange umbrella\"."}], "task": "refering", "answer_template": "The \"man holding an orange umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.524796875, 0.5040322580645161, 0.9037499999999999, 0.9730875576036866], "iscrowd": 0, "area": 25477.55714999999, "rle": {"size": [434, 640], "counts": "Pj^46U=7I7L4O1O1N2M3J6N2N2N2N2N2N2N2O1O1O1O1O1O1O1N2O1O1O1O100O1O1O1O1O1O100O10000O10000O100O100O100O10000O100O1000000O100000000O1000000O100O100O1O100O100O1O100O100^NYMXHh2d7eMSH[2i7jMUHW2i7mMTHT2j7oMUHQ2i7SNTHn1i7XNTHh1i7_NRHb1`7PO\\HP1b7SO\\Hn0b7UO]Hk0a7XO]Hi0a7ZO^Hf0`7]O_Hc0`7_O^Hb0b7_O\\Hb0c7g2O1O1L4N2N2O1N2O1O100O100O100O100000000000000O1000000000000000000001O00000000001O0000001O000000001O0000001O00001O1UMcH_O]77nHHS70UIOl6MXI2i6KYI5g6KYI5h6JWI7k6FVI:k6CWI=k6_OWIa0j6\\OYIc0h6ZO[Ie0g6WO\\Ih0e6VO]Ii0d6_NTHCZ1m1j6VNPHIY1P2j6SNPHJW1S2l6nMTHHS1Y2k6kMnIT2U6gMnIX2g8O1O1O001O1O2N2N00001O00001O0000001O00001O00001O001O00001O001O001O00001O001O00001O001O1O1O2N1O1O1O2N2N1O2N1O2N2N1O2N1O2N1O2N7I4L3M3M3M3M2Niki0"}, "label": "man holding an orange umbrella"}]}
{"id": 12377, "image": "COCO_train2014_000000012377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottom bunk bed with a bag next to it\"."}], "task": "refering", "answer_template": "The \"a bottom bunk bed with a bag next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 120, 121, 122, 123, 124, 125, 126, 132, 133, 134, 135, 136, 137, 138, 139, 145, 146, 147, 148, 149, 150, 151, 152, 157, 158, 159, 160, 161, 162, 163, 164, 165, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228], "bbox": [0.012026666666666666, 0.48928, 0.7494399999999999, 0.98814], "iscrowd": 0, "area": 49670.73285, "rle": {"size": [500, 375], "counts": "ol2a0a>b0^Ob0^Ob0^Ob0C=N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O100000000000000O10000000000000000000000000000O1000000000000000001O0000000O1000000000000000000000000000000001O00001O0000001O00001O0000001O00001O00001O0000001N100O2O000O2O0M3M3M4L3L4M4L3M3L5L3M3M3L5L3M3M4K4M3M4L3L4M3M4M2M3N3L3N2M4M2M3N2M4M2M3N3L3N2M3N3L3N2M4M2N2M4M2N2M3N3M2N2M4M2N2M4M2N2N2M4M2N2M4M2N2N3L3N2N2N3L3Bbd]1"}, "label": "a bottom bunk bed with a bag next to it"}]}
{"id": 12377, "image": "COCO_train2014_000000012377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottom bed on bunk bed\"."}], "task": "refering", "answer_template": "The \"bottom bed on bunk bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 120, 121, 122, 123, 124, 125, 126, 132, 133, 134, 135, 136, 137, 138, 139, 145, 146, 147, 148, 149, 150, 151, 152, 157, 158, 159, 160, 161, 162, 163, 164, 165, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228], "bbox": [0.012026666666666666, 0.48928, 0.7494399999999999, 0.98814], "iscrowd": 0, "area": 49670.73285, "rle": {"size": [500, 375], "counts": "ol2a0a>b0^Ob0^Ob0^Ob0C=N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O100000000000000O10000000000000000000000000000O1000000000000000001O0000000O1000000000000000000000000000000001O00001O0000001O00001O0000001O00001O00001O0000001N100O2O000O2O0M3M3M4L3L4M4L3M3L5L3M3M3L5L3M3M4K4M3M4L3L4M3M4M2M3N3L3N2M4M2M3N2M4M2M3N3L3N2M3N3L3N2M4M2N2M4M2N2M3N3M2N2M4M2N2M4M2N2N2M4M2N2M4M2N2N3L3N2N2N3L3Bbd]1"}, "label": "bottom bed on bunk bed"}]}
{"id": 12377, "image": "COCO_train2014_000000012377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a top bunk with a blanket neatly folded under a pillow\"."}], "task": "refering", "answer_template": "The \"a top bunk with a blanket neatly folded under a pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 15, 16, 17, 18, 19, 20, 21, 27, 28, 29, 30, 31, 32, 33, 34, 35, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 91, 92], "bbox": [0.002986666666666667, 0.0427, 0.7400800000000001, 0.4337], "iscrowd": 0, "area": 31984.51445, "rle": {"size": [500, 375], "counts": "fb0f3n;00001O00O10000O100O10000O10000O100O10000O10000O100O1000O001O010O1O00100O001O010O1O0010O01O1O010O1O0010O01O1O001O001N2O001O1O0O2O001O1O0O2O001O1O0O2O1O001O001O10O010O0100O010O10O010O0100O0010O0100O010O10O010O0100O0010O0100O010O10O010O0100O010O1O010O010O10O010O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O1000O0100O10000O101N101O1N101O0O2O1N101O0O2O1O0O2O0O2O1O0O2O0O2O001N2O001N101N2O001N101N2O001N101O0O2O1N101O0O2O1N101O0O2O1O0O2O0O2O1O0O2O0O2O001N2O001N101N3N1O1N2O2M2O1O5J8I7I7H8IPh_1"}, "label": "a top bunk with a blanket neatly folded under a pillow"}]}
{"id": 12377, "image": "COCO_train2014_000000012377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top bunk with pillow and blanket\"."}], "task": "refering", "answer_template": "The \"the top bunk with pillow and blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 15, 16, 17, 18, 19, 20, 21, 27, 28, 29, 30, 31, 32, 33, 34, 35, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 65, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 91, 92], "bbox": [0.002986666666666667, 0.0427, 0.7400800000000001, 0.4337], "iscrowd": 0, "area": 31984.51445, "rle": {"size": [500, 375], "counts": "fb0f3n;00001O00O10000O100O10000O10000O100O10000O10000O100O1000O001O010O1O00100O001O010O1O0010O01O1O010O1O0010O01O1O001O001N2O001O1O0O2O001O1O0O2O001O1O0O2O1O001O001O10O010O0100O010O10O010O0100O0010O0100O010O10O010O0100O0010O0100O010O10O010O0100O010O1O010O010O10O010O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O1000O0100O10000O101N101O1N101O0O2O1N101O0O2O1O0O2O0O2O1O0O2O0O2O001N2O001N101N2O001N101N2O001N101O0O2O1N101O0O2O1N101O0O2O1O0O2O0O2O1O0O2O0O2O001N2O001N101N3N1O1N2O2M2O1O5J8I7I7H8IPh_1"}, "label": "the top bunk with pillow and blanket"}]}
{"id": 503906, "image": "COCO_train2014_000000503906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry individual with a baseball glove on\"."}], "task": "refering", "answer_template": "The \"a blurry individual with a baseball glove on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 104, 105, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 199, 200, 218, 219, 220, 221, 222, 223, 245, 246, 269, 270, 292, 293, 315, 316, 317, 339, 340], "bbox": [0.444421875, 0.1385480093676815, 0.799015625, 0.9586651053864169], "iscrowd": 0, "area": 23315.71195, "rle": {"size": [427, 640], "counts": "ncf31X=8H3M2O2M3L3N3L4M2N3L4N2N2O2N2M3N1N3N2M3N1N4M2M3N2M3M3M3N6I9G1O101N1O1O1O2O0O1O1O2O0O100]NlLnHV3o6oLnHQ3P7SMmHn2Q7VMmHj2`6hLhHa0f0i2_6PMaH;n0e2^6PN`IT2[6PNcIP2\\6TNaIm1^6WN]Ij1b6[NZIe1e6`NVIa1j6cNQI_1m6`2O100O2O0O1O2O0O101N101N10001N100O2O000O101N1bIVI[6m61O001O001O001O001N2O001O001O001N101O0O2cKYIT2i6gMbIQ2`6kMcIS2`6hMcI9j0aNe5Q1eI<i0aNd5o0eI?j0_Nd5?SJQ1<^Nc59XJV18\\No5OmId1m9O101N100O2O0O1O2O001O001O001O0O2O[FiNV7X1gHjNZ7U1dHmN]7R1aHPO_7P1`HQOa7n0]HTOc7l0[HVOf7j0WHXOi7h0VHYOk7f0SH]Ol7d0QH^OP8b0mGAS8`0fGFY8<`GK_85[G3c8NVG9i8HPG?o8AgFi0Y9e101O01O01O1O10O2O2N1OO001O1N1105J6K5J6K6J5J7J7I5MO001O10O01O001OO1N2OSElMb:h2H1OO1N2O000O2N1O101N1O2N3L3M3L4M4L3M3L101N2N2N2N2O1N2N2N2N101O1O1O015J2N2O1N2O1N2O1N3M2N2O1N3MP]e1"}, "label": "a blurry individual with a baseball glove on"}]}
{"id": 503906, "image": "COCO_train2014_000000503906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurred man\"."}], "task": "refering", "answer_template": "The \"a blurred man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 104, 105, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 195, 196, 197, 199, 200, 218, 219, 220, 221, 222, 223, 245, 246, 269, 270, 292, 293, 315, 316, 317, 339, 340], "bbox": [0.444421875, 0.1385480093676815, 0.799015625, 0.9586651053864169], "iscrowd": 0, "area": 23315.71195, "rle": {"size": [427, 640], "counts": "ncf31X=8H3M2O2M3L3N3L4M2N3L4N2N2O2N2M3N1N3N2M3N1N4M2M3N2M3M3M3N6I9G1O101N1O1O1O2O0O1O1O2O0O100]NlLnHV3o6oLnHQ3P7SMmHn2Q7VMmHj2`6hLhHa0f0i2_6PMaH;n0e2^6PN`IT2[6PNcIP2\\6TNaIm1^6WN]Ij1b6[NZIe1e6`NVIa1j6cNQI_1m6`2O100O2O0O1O2O0O101N101N10001N100O2O000O101N1bIVI[6m61O001O001O001O001N2O001O001O001N101O0O2cKYIT2i6gMbIQ2`6kMcIS2`6hMcI9j0aNe5Q1eI<i0aNd5o0eI?j0_Nd5?SJQ1<^Nc59XJV18\\No5OmId1m9O101N100O2O0O1O2O001O001O001O0O2O[FiNV7X1gHjNZ7U1dHmN]7R1aHPO_7P1`HQOa7n0]HTOc7l0[HVOf7j0WHXOi7h0VHYOk7f0SH]Ol7d0QH^OP8b0mGAS8`0fGFY8<`GK_85[G3c8NVG9i8HPG?o8AgFi0Y9e101O01O01O1O10O2O2N1OO001O1N1105J6K5J6K6J5J7J7I5MO001O10O01O001OO1N2OSElMb:h2H1OO1N2O000O2N1O101N1O2N3L3M3L4M4L3M3L101N2N2N2N2O1N2N2N2N101O1O1O015J2N2O1N2O1N2O1N3M2N2O1N3MP]e1"}, "label": "a blurred man"}]}
{"id": 503906, "image": "COCO_train2014_000000503906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player holding bat with number 18 on his back\"."}], "task": "refering", "answer_template": "The \"player holding bat with number 18 on his back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 29, 30, 31, 32, 33, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.0015, 0.013442622950819673, 0.664203125, 0.9887822014051523], "iscrowd": 0, "area": 106404.0187, "rle": {"size": [427, 640], "counts": "Pg0`3i92L4M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3M3L4M3M3M3L4D<E;N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N3M2N2N2N200O100O1O100O100O1O100O100O1O100O1O100O100O100000000O10000000000O10000000000O10000000000O1000000000000000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O00000000001O000000O1O1O1O1O100O1[O\\FbMe9Z2h0L4M3L4L4L4L4L4N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N200O100O0100O001O010O001O10O01O0010O01O1O0010O01O03N2N2M4M2M3N001N100O100O1000000000000001O0000000001O00000001WHDa2<^MG`2:]MIb28\\MJc27[MKd26ZMLe25XMNg22XM1f20XM2g2OWM3h2NVM4i2MTM6k2KSM8l2HRM:m2FRM<m2EPM>o2CoL?P3BnL`0Q3AmLb0Q3_OlLd0S3\\OlLf0S3[OkLg0T3ZOjLh0U3YOiLi0V3XOgLl0W3UOdLP1Z3QOcLS1\\3nN_LW1_3kN\\LZ1c3gNYL^1d3dNWLa1g3aNTLd1k3]NPLh1n3[NmKi1Q4YNjKl1U4UNfKQ2W4QNeKS2Y4oMeKS2Z4nMeKT2X4nMhKR2W4PNhKQ2U4QNjKQ2S4YOeJg0Z5[OdJg0Y5[OgJe0W5]OhJe0V5]OiJc0U5_OjJc0S5_OmJb0Q5@mJa0Q5AoJ`0o4APK`0n4CmJc0o4R4J5K6J4L1O1O2N1O2N1O1O2N1O2N1O1O2N1O2M2O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1N3N1O1O2N1O2N1O1O2N001O1O1O1O1O1O1O00100O100O1O100O010O100O100O100O10O0100O100O107H9H8G101N101N101N100O00O2N1N3L3L5L3010O2O0O101N101N1O1J7J5J6J7J5JX\\i2"}, "label": "player holding bat with number 18 on his back"}]}
{"id": 503906, "image": "COCO_train2014_000000503906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a black cap\"."}], "task": "refering", "answer_template": "The \"a person wearing a black cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 29, 30, 31, 32, 33, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.0015, 0.013442622950819673, 0.664203125, 0.9887822014051523], "iscrowd": 0, "area": 106404.0187, "rle": {"size": [427, 640], "counts": "Pg0`3i92L4M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3M3L4M3M3M3L4D<E;N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N3M2N2N2N200O100O1O100O100O1O100O100O1O100O1O100O100O100000000O10000000000O10000000000O10000000000O1000000000000000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O00000000001O000000O1O1O1O1O100O1[O\\FbMe9Z2h0L4M3L4L4L4L4L4N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N200O100O0100O001O010O001O10O01O0010O01O1O0010O01O03N2N2M4M2M3N001N100O100O1000000000000001O0000000001O00000001WHDa2<^MG`2:]MIb28\\MJc27[MKd26ZMLe25XMNg22XM1f20XM2g2OWM3h2NVM4i2MTM6k2KSM8l2HRM:m2FRM<m2EPM>o2CoL?P3BnL`0Q3AmLb0Q3_OlLd0S3\\OlLf0S3[OkLg0T3ZOjLh0U3YOiLi0V3XOgLl0W3UOdLP1Z3QOcLS1\\3nN_LW1_3kN\\LZ1c3gNYL^1d3dNWLa1g3aNTLd1k3]NPLh1n3[NmKi1Q4YNjKl1U4UNfKQ2W4QNeKS2Y4oMeKS2Z4nMeKT2X4nMhKR2W4PNhKQ2U4QNjKQ2S4YOeJg0Z5[OdJg0Y5[OgJe0W5]OhJe0V5]OiJc0U5_OjJc0S5_OmJb0Q5@mJa0Q5AoJ`0o4APK`0n4CmJc0o4R4J5K6J4L1O1O2N1O2N1O1O2N1O2N1O1O2N1O2M2O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1N3N1O1O2N1O2N1O1O2N001O1O1O1O1O1O1O00100O100O1O100O010O100O100O100O10O0100O100O107H9H8G101N101N101N100O00O2N1N3L3L5L3010O2O0O101N101N1O1J7J5J6J7J5JX\\i2"}, "label": "a person wearing a black cap"}]}
{"id": 405604, "image": "COCO_train2014_000000405604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple wireless keyboard\"."}], "task": "refering", "answer_template": "The \"apple wireless keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5545, 0.6045, 0.9977499999999999, 0.9842708333333334], "iscrowd": 0, "area": 43688.67839999999, "rle": {"size": [480, 640], "counts": "`jV58a>:E9G9G8I7L4K5K5L4K6K4K5K5L4K5K5L4K6J5L4K5L4K5K5L5J5K5N200001O01O000000000000001O01O0000000001O0000000001O0001O0000000000001O01O0000000001O00000001O000001O00000000000010O00000000000001O000001O000001O00000000000010O00000000000001O000001O000001O000000000001O01O0000000000001O0001O00000001O000000000001O01O0000000000001O0001O0000000001O00000001O0001O0000000000001O01O000000000001O00000001O0001O0000000000000010O000000000000000000001O0000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O0000000000000000001O000000000000000O10001O0000=CS4mKf0"}, "label": "apple wireless keyboard"}]}
{"id": 405604, "image": "COCO_train2014_000000405604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the keyboard not attached to the laptop\"."}], "task": "refering", "answer_template": "The \"the keyboard not attached to the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5545, 0.6045, 0.9977499999999999, 0.9842708333333334], "iscrowd": 0, "area": 43688.67839999999, "rle": {"size": [480, 640], "counts": "`jV58a>:E9G9G8I7L4K5K5L4K6K4K5K5L4K5K5L4K6J5L4K5L4K5K5L5J5K5N200001O01O000000000000001O01O0000000001O0000000001O0001O0000000000001O01O0000000001O00000001O000001O00000000000010O00000000000001O000001O000001O00000000000010O00000000000001O000001O000001O000000000001O01O0000000000001O0001O00000001O000000000001O01O0000000000001O0001O0000000001O00000001O0001O0000000000001O01O000000000001O00000001O0001O0000000000000010O000000000000000000001O0000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O0000000000000000001O000000000000000O10001O0000=CS4mKf0"}, "label": "the keyboard not attached to the laptop"}]}
{"id": 405604, "image": "COCO_train2014_000000405604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop keyboard with laptop screen\"."}], "task": "refering", "answer_template": "The \"laptop keyboard with laptop screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287], "bbox": [0.0, 0.42883333333333334, 0.567984375, 0.7277083333333334], "iscrowd": 0, "area": 46894.365300000005, "rle": {"size": [480, 640], "counts": "h6n0R>m1SNZ1eN10000000000000000000000000000000000000000000000000000000000000000000O01000000000000000000000000000000000000000000000000000000000000000000O10000000O1000000000000000000000000000000000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000000O10000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000O10000000000000000000000000000000O10000000000000000000000000000000000000000000O1000000000000000000000O1000000000000000000000000000000000000000000000000000O10000000000000O10000000000000000000000000000000000000000000000000000000000000O1000O10000000000000000000000000000000000000000000000004L5K4L4L4L5K4L4K5K6K4K5K5L5J5K5K5L5J5K5K5L5J5K5L4K6J5K5LkQQ4"}, "label": "laptop keyboard with laptop screen"}]}
{"id": 446565, "image": "COCO_train2014_000000446565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right most vulture\"."}], "task": "refering", "answer_template": "The \"right most vulture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 177, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 318], "bbox": [0.63046875, 0.36020833333333335, 0.8684531249999999, 0.7699791666666668], "iscrowd": 0, "area": 13734.456949999993, "rle": {"size": [480, 640], "counts": "jam51m>1UB1e<1WC8b<0SC;FGX<5iCQ1U<TObCQ1\\<l0M2O2M2N3N1N3O001N101O001O001N101O0100O101N1N2N3M2N2N2O0OfE\\MX8d2YF[Mm03i8a2QFjMP1Fo8\\3lFcLV9`3fF_L[9f3_FZLc9OhE\\4X:dKhE[4Z:dKgEZ4Z:fKfEZ4[:eKfEY4[:gKeEX4]:gKdEX4\\:hKdEW4^:hKcEU4_:kKaET4g:M2O2M2O2M2N2O2M2O2M2O2M2O1N3N1N3O001O0010hDPM^:P3\\EVMd:j2VE\\Mk:c2PEcMo:W31O010M2M3L5L3L5L3N3O1O001N2O001O1O001O1O0O2O1O001O1O1O001N20O101O0O101O0O101O0O100001O001O001O001N2O001O001O1SNaCV1_<iNdCU1]<iNfCU1Z<jNiCT1\\<fNhCW1]<bNgC]1P=N2N2O1N2N2L5L3M3M3M3M3M3M3M4L3M3M`PW1"}, "label": "right most vulture"}]}
{"id": 446565, "image": "COCO_train2014_000000446565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rightmost and lowest vulture\"."}], "task": "refering", "answer_template": "The \"the rightmost and lowest vulture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 177, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 318], "bbox": [0.63046875, 0.36020833333333335, 0.8684531249999999, 0.7699791666666668], "iscrowd": 0, "area": 13734.456949999993, "rle": {"size": [480, 640], "counts": "jam51m>1UB1e<1WC8b<0SC;FGX<5iCQ1U<TObCQ1\\<l0M2O2M2N3N1N3O001N101O001O001N101O0100O101N1N2N3M2N2N2O0OfE\\MX8d2YF[Mm03i8a2QFjMP1Fo8\\3lFcLV9`3fF_L[9f3_FZLc9OhE\\4X:dKhE[4Z:dKgEZ4Z:fKfEZ4[:eKfEY4[:gKeEX4]:gKdEX4\\:hKdEW4^:hKcEU4_:kKaET4g:M2O2M2O2M2N2O2M2O2M2O2M2O1N3N1N3O001O0010hDPM^:P3\\EVMd:j2VE\\Mk:c2PEcMo:W31O010M2M3L5L3L5L3N3O1O001N2O001O1O001O1O0O2O1O001O1O1O001N20O101O0O101O0O101O0O100001O001O001O001N2O001O001O1SNaCV1_<iNdCU1]<iNfCU1Z<jNiCT1\\<fNhCW1]<bNgC]1P=N2N2O1N2N2L5L3M3M3M3M3M3M3M4L3M3M`PW1"}, "label": "the rightmost and lowest vulture"}]}
{"id": 446565, "image": "COCO_train2014_000000446565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hawk in middle in front eating carcass\"."}], "task": "refering", "answer_template": "The \"hawk in middle in front eating carcass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308], "bbox": [0.2919375, 0.4715, 0.49839062500000003, 0.8005416666666666], "iscrowd": 0, "area": 14074.62645, "rle": {"size": [480, 640], "counts": "VPh2:Y>`0@?A`0@?A?E<J4K4L4L3N3L4L4L3N3L4M3O001O1O10O01O1O00010O00001O01O01O000010O00011N1O100O1O101N1O1O1O`EnKT:Q4jERLV:m3hEVLY:h3eE[L[:d3cE_L]:R40000001O000001O0001O00000000001O0ZOaEkL_:S3dElL\\:Q3gEoLY:n2jERMV:l2lESMV:i2mEWMS:g2PFXMP:e2SFZMn9d2TF\\Ml9a2WF^Mj9a2WF_Mj9`2WF_Mj9_2WF`Mk9^2VFbMk9\\2VFcMl9[2UFeMl9Y2VFfMk9X2VFgMm9V2TFjMm9T2TFkMn9S2SFmMn9Q2SFoMn9o1SFPNP:n1PFRNP:m1QFRNo9n1RFRNn9m1SFSNm9l1TFSNm9l1TFTNk9l1VFSNk9l1VFTNj9k1WFUNh9k1YFTNh9k1YFUNg9j1ZFUNg9k1YFTNg9l1ZFSNg9l1ZFSNg9l1g1O1O1O001O1O1O1O100O1O1O1O1O001O1O1O1O1O1Of0[OiVf4"}, "label": "hawk in middle in front eating carcass"}]}
{"id": 446565, "image": "COCO_train2014_000000446565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vulture in the middle with its back facing the camera\"."}], "task": "refering", "answer_template": "The \"the vulture in the middle with its back facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308], "bbox": [0.2919375, 0.4715, 0.49839062500000003, 0.8005416666666666], "iscrowd": 0, "area": 14074.62645, "rle": {"size": [480, 640], "counts": "VPh2:Y>`0@?A`0@?A?E<J4K4L4L3N3L4L4L3N3L4M3O001O1O10O01O1O00010O00001O01O01O000010O00011N1O100O1O101N1O1O1O`EnKT:Q4jERLV:m3hEVLY:h3eE[L[:d3cE_L]:R40000001O000001O0001O00000000001O0ZOaEkL_:S3dElL\\:Q3gEoLY:n2jERMV:l2lESMV:i2mEWMS:g2PFXMP:e2SFZMn9d2TF\\Ml9a2WF^Mj9a2WF_Mj9`2WF_Mj9_2WF`Mk9^2VFbMk9\\2VFcMl9[2UFeMl9Y2VFfMk9X2VFgMm9V2TFjMm9T2TFkMn9S2SFmMn9Q2SFoMn9o1SFPNP:n1PFRNP:m1QFRNo9n1RFRNn9m1SFSNm9l1TFSNm9l1TFTNk9l1VFSNk9l1VFTNj9k1WFUNh9k1YFTNh9k1YFUNg9j1ZFUNg9k1YFTNg9l1ZFSNg9l1ZFSNg9l1g1O1O1O001O1O1O1O100O1O1O1O1O001O1O1O1O1O1Of0[OiVf4"}, "label": "the vulture in the middle with its back facing the camera"}]}
{"id": 544875, "image": "COCO_train2014_000000544875.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet to the right\"."}], "task": "refering", "answer_template": "The \"the toilet to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 147, 148, 149, 150, 164, 165, 166, 167, 181, 182, 183, 184, 198, 199, 200, 201, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 250, 251, 252, 253, 267, 268, 269, 284, 285, 286, 301, 302, 303], "bbox": [0.6695833333333333, 0.32687499999999997, 0.9476875, 0.77040625], "iscrowd": 0, "area": 22823.927150000007, "rle": {"size": [640, 480], "counts": "ibY67fc07I7I4lMZOj@i0P?@k@b0o>Hl@9o>Ol@3n>7m@Jm>?n@Bo>e0m@\\OP?k0j@WOT?j0l@WOR?k0l@XOP?k0o@XOm>j0QAZOk>g0UA]Oe>f0YA^Oc>d0\\A_O`>c0^AA]>a0cACX>?fADW>>dAHY>:eAIX>8hAJU>8jAI_1aMk:h2fCHZ1lMf:^2oCHU1VNc:S2XDIo0^Nb:k1^DIk0`Ne:i1_DHg0cNi:g1_DHc0dNm:f1_DH>gNQ;c1aDG8jNV;a1aDG4kNZ;d6O2N10001O0000001N10001O0000001O00001O000O101O000001N2O001O1O001N2O001O1O001N2O0gJ`DR2a;V31eJ^DV2b;iM_DW2b;gM_DX2c;gM]DY2e;dM\\D\\2e;bM]D]2d;aM]D]11eLc;m1]DW19iL[;n1]DS1a0kLT;P2]Do0g0nLm:Q2]Dk0P1PMe:S2[Dh0X1RM^:T2\\Dh0Z1QM[:U2\\Di0]1oLX:V2]Di0d=TO^Bk0d=RO^Bl0e=QO^Bm0e=PO\\Bo0h=lN[BR1i=iNYBV1k=eNXBY1l=bNVB]1n=_NSB`1Q>[NRBc1V`0O2N1O2N1O2N1O2N1N3N1O2M2N3M2O2M2N3M2O2M3M2N3M2N3M2M4M5K6J7I6JjX?"}, "label": "the toilet to the right"}]}
{"id": 544875, "image": "COCO_train2014_000000544875.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toilet to the right of a similar toilet\"."}], "task": "refering", "answer_template": "The \"a toilet to the right of a similar toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 147, 148, 149, 150, 164, 165, 166, 167, 181, 182, 183, 184, 198, 199, 200, 201, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 250, 251, 252, 253, 267, 268, 269, 284, 285, 286, 301, 302, 303], "bbox": [0.6695833333333333, 0.32687499999999997, 0.9476875, 0.77040625], "iscrowd": 0, "area": 22823.927150000007, "rle": {"size": [640, 480], "counts": "ibY67fc07I7I4lMZOj@i0P?@k@b0o>Hl@9o>Ol@3n>7m@Jm>?n@Bo>e0m@\\OP?k0j@WOT?j0l@WOR?k0l@XOP?k0o@XOm>j0QAZOk>g0UA]Oe>f0YA^Oc>d0\\A_O`>c0^AA]>a0cACX>?fADW>>dAHY>:eAIX>8hAJU>8jAI_1aMk:h2fCHZ1lMf:^2oCHU1VNc:S2XDIo0^Nb:k1^DIk0`Ne:i1_DHg0cNi:g1_DHc0dNm:f1_DH>gNQ;c1aDG8jNV;a1aDG4kNZ;d6O2N10001O0000001N10001O0000001O00001O000O101O000001N2O001O1O001N2O001O1O001N2O0gJ`DR2a;V31eJ^DV2b;iM_DW2b;gM_DX2c;gM]DY2e;dM\\D\\2e;bM]D]2d;aM]D]11eLc;m1]DW19iL[;n1]DS1a0kLT;P2]Do0g0nLm:Q2]Dk0P1PMe:S2[Dh0X1RM^:T2\\Dh0Z1QM[:U2\\Di0]1oLX:V2]Di0d=TO^Bk0d=RO^Bl0e=QO^Bm0e=PO\\Bo0h=lN[BR1i=iNYBV1k=eNXBY1l=bNVB]1n=_NSB`1Q>[NRBc1V`0O2N1O2N1O2N1O2N1N3N1O2M2N3M2O2M2N3M2O2M3M2N3M2N3M2M4M5K6J7I6JjX?"}, "label": "a toilet to the right of a similar toilet"}]}
{"id": 544875, "image": "COCO_train2014_000000544875.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet to the left of another toilet\"."}], "task": "refering", "answer_template": "The \"the toilet to the left of another toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 139, 140, 141, 155, 156, 157, 158, 159, 172, 173, 174, 175, 189, 190, 191, 192, 205, 206, 207, 208, 209, 222, 223, 224, 225, 226, 240, 241, 242, 243, 257, 258, 259, 260, 274, 275, 276, 277, 291, 292, 293], "bbox": [0.08108333333333334, 0.33896875, 0.3633541666666667, 0.78040625], "iscrowd": 0, "area": 22354.82744999999, "rle": {"size": [640, 480], "counts": "Pgh01nc03M3N2M2N3M2N2N3M2N2N3M2O1N3N1O1N3N1O2M2O1N3N1O1N2O1O1N2O1O1N3N1SNUNcAl1W>\\NeAf1U>bNhAQ2\\=ZNbB\\2f<iMVCk2U<ZMhCk2R<XMlCl2POnKb<Z1[DV3c;mL[DT3c;nL[DT3c;oL[DS3b;oL]DR3a;PM]DR3a;PM^DQ3`;QM_DP3_;RM_Do2`;TM^Dm2`;UM^Dm2`;VM^Dk2`;XM^Di2`;ZM]Dh2a;[M]Df2a;\\M]Df2a;]M]Dd2a;^M^Dc2^;aM`Da2\\;cMcD]2\\;U3N1O00100O010O010O10O100O100O10O10O100O10000O100O100O100O1O1O1O1O1N2O1O1O1O1N2O2M2jN\\DnIg;n5^DnId;o5bDlIa;P6gDjIZ;S6\\1L4H7I8H8M2N3L4M2N3M3L4M3M2N3oMj@TOY?c0QAYOR??WA]Ol>a0WA\\Ok>b0XA[Ok>a0WA^Ok>`0WA^Ok>`0ZAXOl>d0[2M3L5L6I6K6Ilon5"}, "label": "the toilet to the left of another toilet"}]}
{"id": 544875, "image": "COCO_train2014_000000544875.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dirty porcelian toilet with the lid up\"."}], "task": "refering", "answer_template": "The \"a dirty porcelian toilet with the lid up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 139, 140, 141, 155, 156, 157, 158, 159, 172, 173, 174, 175, 189, 190, 191, 192, 205, 206, 207, 208, 209, 222, 223, 224, 225, 226, 240, 241, 242, 243, 257, 258, 259, 260, 274, 275, 276, 277, 291, 292, 293], "bbox": [0.08108333333333334, 0.33896875, 0.3633541666666667, 0.78040625], "iscrowd": 0, "area": 22354.82744999999, "rle": {"size": [640, 480], "counts": "Pgh01nc03M3N2M2N3M2N2N3M2N2N3M2O1N3N1O1N3N1O2M2O1N3N1O1N2O1O1N2O1O1N3N1SNUNcAl1W>\\NeAf1U>bNhAQ2\\=ZNbB\\2f<iMVCk2U<ZMhCk2R<XMlCl2POnKb<Z1[DV3c;mL[DT3c;nL[DT3c;oL[DS3b;oL]DR3a;PM]DR3a;PM^DQ3`;QM_DP3_;RM_Do2`;TM^Dm2`;UM^Dm2`;VM^Dk2`;XM^Di2`;ZM]Dh2a;[M]Df2a;\\M]Df2a;]M]Dd2a;^M^Dc2^;aM`Da2\\;cMcD]2\\;U3N1O00100O010O010O10O100O100O10O10O100O10000O100O100O100O1O1O1O1O1N2O1O1O1O1N2O2M2jN\\DnIg;n5^DnId;o5bDlIa;P6gDjIZ;S6\\1L4H7I8H8M2N3L4M2N3M3L4M3M2N3oMj@TOY?c0QAYOR??WA]Ol>a0WA\\Ok>b0XA[Ok>a0WA^Ok>`0WA^Ok>`0ZAXOl>d0[2M3L5L6I6K6Ilon5"}, "label": "a dirty porcelian toilet with the lid up"}]}
{"id": 528493, "image": "COCO_train2014_000000528493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in orange shirt and blue pants enjoying ice skating\"."}], "task": "refering", "answer_template": "The \"a man in orange shirt and blue pants enjoying ice skating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 114, 115, 129, 130, 131, 143, 144, 145, 146, 158, 159, 160, 161, 173, 174, 175, 176, 188, 189, 190, 191, 203, 204, 205, 206, 219, 220, 221, 234, 235, 236], "bbox": [0.5767056074766356, 0.2944375, 0.8047663551401869, 0.6878437500000001], "iscrowd": 0, "area": 14629.938750000003, "rle": {"size": [640, 428], "counts": "cgj46]`0NfB9X=N\\B:a=MSB;j=LjA<S>KaA<]>KWA=g>Im@?b0UNP=b1SBa0k0mMQ=i1hAa0U1gMR=Z4kBgKT=\\4iBeKV=^4gBbKY=`4eBaK[=b4aB_K^=h4\\BXKc=P5UBQKj=W5nAjJQ>c5O100O1O1aNWJfDj5Z;[J`Df5`;^J[Dc5d;cJVD^5j;gJQDY5o;lJkCT5U<QKfCP5Z<UK`Cl4`<UOaCiL_<W3dCfL[<Z3iCcLW<]3lC`LT<_3PD^Lo;c3TDZLl;e3XDXLh;h3[DULd;n3]DoKc;S4^DjKb;X4_DeKa;]4`D`K_;c4bDZK^;h4cDUK];m4eDnJ[;U5fDhJZ;Z5gDcJY;_5hD^JW;e5jDXJV;j5kDSJU;o5lDnIS;U6nDhIS;Y6Z11O100O1`CeIX;\\6cDgI^;Z6^DiIb;:dCZ5e0_Jh;3fC`5>_Jn;LhCf55aJT<EjCl5NaJj<`5QCcJo<_5mBcJT=T6001O001O0]KTDR1m;gNaDQ1_;iNoDo0R;cMkBh0a2]1d:hMoBd0k2\\1W:kMTCa0S3\\1[:]NnE`1T:XNUFe1V>M3M3M4M2M3M3M3M3M3M3M4L3M3M3M3M3M3M[Rd1"}, "label": "a man in orange shirt and blue pants enjoying ice skating"}]}
{"id": 528493, "image": "COCO_train2014_000000528493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in red jacket ready for skiing\"."}], "task": "refering", "answer_template": "The \"a person in red jacket ready for skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 114, 115, 129, 130, 131, 143, 144, 145, 146, 158, 159, 160, 161, 173, 174, 175, 176, 188, 189, 190, 191, 203, 204, 205, 206, 219, 220, 221, 234, 235, 236], "bbox": [0.5767056074766356, 0.2944375, 0.8047663551401869, 0.6878437500000001], "iscrowd": 0, "area": 14629.938750000003, "rle": {"size": [640, 428], "counts": "cgj46]`0NfB9X=N\\B:a=MSB;j=LjA<S>KaA<]>KWA=g>Im@?b0UNP=b1SBa0k0mMQ=i1hAa0U1gMR=Z4kBgKT=\\4iBeKV=^4gBbKY=`4eBaK[=b4aB_K^=h4\\BXKc=P5UBQKj=W5nAjJQ>c5O100O1O1aNWJfDj5Z;[J`Df5`;^J[Dc5d;cJVD^5j;gJQDY5o;lJkCT5U<QKfCP5Z<UK`Cl4`<UOaCiL_<W3dCfL[<Z3iCcLW<]3lC`LT<_3PD^Lo;c3TDZLl;e3XDXLh;h3[DULd;n3]DoKc;S4^DjKb;X4_DeKa;]4`D`K_;c4bDZK^;h4cDUK];m4eDnJ[;U5fDhJZ;Z5gDcJY;_5hD^JW;e5jDXJV;j5kDSJU;o5lDnIS;U6nDhIS;Y6Z11O100O1`CeIX;\\6cDgI^;Z6^DiIb;:dCZ5e0_Jh;3fC`5>_Jn;LhCf55aJT<EjCl5NaJj<`5QCcJo<_5mBcJT=T6001O001O0]KTDR1m;gNaDQ1_;iNoDo0R;cMkBh0a2]1d:hMoBd0k2\\1W:kMTCa0S3\\1[:]NnE`1T:XNUFe1V>M3M3M4M2M3M3M3M3M3M3M4L3M3M3M3M3M3M[Rd1"}, "label": "a person in red jacket ready for skiing"}]}
{"id": 528493, "image": "COCO_train2014_000000528493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person skiing , wear plaid and blue pants\"."}], "task": "refering", "answer_template": "The \"a person skiing , wear plaid and blue pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 141, 142, 143, 155, 156, 157, 158, 170, 171, 172, 173, 187, 188, 202, 203, 217, 218, 219, 232, 233], "bbox": [0.36614485981308414, 0.34725, 0.6152102803738317, 0.67675], "iscrowd": 0, "area": 10156.826100000002, "rle": {"size": [640, 428], "counts": "[]R32lc05L5K4L2N2N101O00001O0001O1O001O1O001O10O01O1O00O100O1O100O001J6J5L5L4O001O001O1O4L:F>B8fNmMX@[2`?hM\\@c2Z?aMa@j2U?ZMf@Q3P?QMm@X3k>kLQA[3j>o0N2O1O1O1O1O1O1O1N2YBbJj<_5UCcJi<^5UCdJj<]5UCcJk<_5RCbJo<^5PCbJP=_5nBcJQ=_5lBbJU=^5iBcJW=_5gB`JZ=c5bB^J`=b5]B_Je=m54L>B001O001O1O0010O10O10O01O001O001fMfAkN\\>Q1lAiNW>2QCH\\=KiB1[=JjB1Z=KkB0Z=KiB2[=JiB3l=TOZBi0b`0L4L3M4L3M3L4KYgV3"}, "label": "a person skiing , wear plaid and blue pants"}]}
{"id": 528493, "image": "COCO_train2014_000000528493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in plaid jacket and blue pants on skis\"."}], "task": "refering", "answer_template": "The \"woman in plaid jacket and blue pants on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 141, 142, 143, 155, 156, 157, 158, 170, 171, 172, 173, 187, 188, 202, 203, 217, 218, 219, 232, 233], "bbox": [0.36614485981308414, 0.34725, 0.6152102803738317, 0.67675], "iscrowd": 0, "area": 10156.826100000002, "rle": {"size": [640, 428], "counts": "[]R32lc05L5K4L2N2N101O00001O0001O1O001O1O001O10O01O1O00O100O1O100O001J6J5L5L4O001O001O1O4L:F>B8fNmMX@[2`?hM\\@c2Z?aMa@j2U?ZMf@Q3P?QMm@X3k>kLQA[3j>o0N2O1O1O1O1O1O1O1N2YBbJj<_5UCcJi<^5UCdJj<]5UCcJk<_5RCbJo<^5PCbJP=_5nBcJQ=_5lBbJU=^5iBcJW=_5gB`JZ=c5bB^J`=b5]B_Je=m54L>B001O001O1O0010O10O10O01O001O001fMfAkN\\>Q1lAiNW>2QCH\\=KiB1[=JjB1Z=KkB0Z=KiB2[=JiB3l=TOZBi0b`0L4L3M4L3M3L4KYgV3"}, "label": "woman in plaid jacket and blue pants on skis"}]}
{"id": 364653, "image": "COCO_train2014_000000364653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby lamb next to the mother lamb\"."}], "task": "refering", "answer_template": "The \"a baby lamb next to the mother lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 179, 180, 181, 182, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.8130625, 0.4071360381861575, 0.9671875, 0.8137231503579954], "iscrowd": 0, "area": 10747.551500000001, "rle": {"size": [419, 640], "counts": "Ynd69i<5K5L4L4K3N2M3O1O6iDgNX:X1bEoN]:R1[EVOd:j0UE^Oj:_1O10O1007G4L_1bN2M2N3M2OO01O010O001O001O001O00O10000000EhKkGX4S8<O2N2N200000O100001O1O001O001O001003L3I7C>QOU1E6K0O2N1O2O0O2N1O2O0O22M3N3M2M4M2M4M2M3N1N101N21O0O10001O000O2O6I7I7I7IgZ8"}, "label": "a baby lamb next to the mother lamb"}]}
{"id": 364653, "image": "COCO_train2014_000000364653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a lamb standing beside a larger sheep\"."}], "task": "refering", "answer_template": "The \"this is a lamb standing beside a larger sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 179, 180, 181, 182, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.8130625, 0.4071360381861575, 0.9671875, 0.8137231503579954], "iscrowd": 0, "area": 10747.551500000001, "rle": {"size": [419, 640], "counts": "Ynd69i<5K5L4L4K3N2M3O1O6iDgNX:X1bEoN]:R1[EVOd:j0UE^Oj:_1O10O1007G4L_1bN2M2N3M2OO01O010O001O001O001O00O10000000EhKkGX4S8<O2N2N200000O100001O1O001O001O001003L3I7C>QOU1E6K0O2N1O2O0O2N1O2O0O22M3N3M2M4M2M4M2M3N1N101N21O0O10001O000O2O6I7I7I7IgZ8"}, "label": "this is a lamb standing beside a larger sheep"}]}
{"id": 364653, "image": "COCO_train2014_000000364653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is big sheep sitting with small sheep\"."}], "task": "refering", "answer_template": "The \"there is big sheep sitting with small sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272], "bbox": [0.370109375, 0.304272076372315, 0.842546875, 0.7860859188544153], "iscrowd": 0, "area": 37122.04075, "rle": {"size": [419, 640], "counts": "_UQ34n<5K5K5K5K5K3M3M3M4L2M2O1O1O1O2oF^N^6c1aIbNY6`1aIiN[6X1^InNb6T1VISOi6n0TIUOk6l0RIVOm6k0QIXOn6g0QI[Oo6e0PI]Oo6c0QI^On6b0QI_Oo6b0PI_Oo6b0oH@P7a0oH@Q7?nHCQ7>nHCQ7>mHDS7<lHES7<kHFU79kHHU78iHJV77hHKX75fHMY73gHNY72eH0Z71dH1\\7ObH3]7MbH5^7KaH6^7K`H7`7I^H8c7G\\H;c7F[H<e7DYH>f7BYH`0g7@WHb0h7_OVHc0j7]OTHe0k7[OTHg0l7YORHi0m7XOQHj0o7U200O10000000000000000000000O1O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1O10000004L4L4M3M3N3LN3M3L3N2N1O1O1O1O1O1O1O1O100O1O1O10000001WKnH_3R7_LPIa3P7\\LTIc3m6YLVIg3j6WLYIh3g6UL\\Ik3d6RL`Im3a6oKbIQ4^6lKeIT4[6iKiIV4X6cKnI]4]7O001O001O000000001O0000000000001O0000000000001O0000000000001O000000001O0000001N10000O2O000O100O2O0O10001N100O101N100O100O2O0O101O000O2O00001O001O00001O0O2O00001O00001O001N10001O001O00001O00001O001O00001O00001O001O001O00001O001O001O001O001O1O001O1O001O001O1O001O1O001O1O00100O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O1O1O10O01O1fNSE<n:^O[E>f:\\OcE`0^:ZOjEc0a;JTQY1"}, "label": "there is big sheep sitting with small sheep"}]}
{"id": 364653, "image": "COCO_train2014_000000364653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother sheep sitting in the grass\"."}], "task": "refering", "answer_template": "The \"a mother sheep sitting in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272], "bbox": [0.370109375, 0.304272076372315, 0.842546875, 0.7860859188544153], "iscrowd": 0, "area": 37122.04075, "rle": {"size": [419, 640], "counts": "_UQ34n<5K5K5K5K5K3M3M3M4L2M2O1O1O1O2oF^N^6c1aIbNY6`1aIiN[6X1^InNb6T1VISOi6n0TIUOk6l0RIVOm6k0QIXOn6g0QI[Oo6e0PI]Oo6c0QI^On6b0QI_Oo6b0PI_Oo6b0oH@P7a0oH@Q7?nHCQ7>nHCQ7>mHDS7<lHES7<kHFU79kHHU78iHJV77hHKX75fHMY73gHNY72eH0Z71dH1\\7ObH3]7MbH5^7KaH6^7K`H7`7I^H8c7G\\H;c7F[H<e7DYH>f7BYH`0g7@WHb0h7_OVHc0j7]OTHe0k7[OTHg0l7YORHi0m7XOQHj0o7U200O10000000000000000000000O1O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1O10000004L4L4M3M3N3LN3M3L3N2N1O1O1O1O1O1O1O1O100O1O1O10000001WKnH_3R7_LPIa3P7\\LTIc3m6YLVIg3j6WLYIh3g6UL\\Ik3d6RL`Im3a6oKbIQ4^6lKeIT4[6iKiIV4X6cKnI]4]7O001O001O000000001O0000000000001O0000000000001O0000000000001O000000001O0000001N10000O2O000O100O2O0O10001N100O101N100O100O2O0O101O000O2O00001O001O00001O0O2O00001O00001O001N10001O001O00001O00001O001O00001O00001O001O001O00001O001O001O001O001O1O001O1O001O001O1O001O1O001O1O00100O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O1O1O10O01O1fNSE<n:^O[E>f:\\OcE`0^:ZOjEc0a;JTQY1"}, "label": "a mother sheep sitting in the grass"}]}
{"id": 143470, "image": "COCO_train2014_000000143470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lonely piece of broccoli separated from the rest of the broccoli\"."}], "task": "refering", "answer_template": "The \"a lonely piece of broccoli separated from the rest of the broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274], "bbox": [0.74065625, 0.5078688524590165, 0.9400625, 0.8], "iscrowd": 0, "area": 9756.784499999998, "rle": {"size": [427, 640], "counts": "VnU61X=4L3M3M4L3M4M2M3M4L3M4L3M3N3M2M3N3M2N3L3N2N3L3N2N1O1N2O1O0O2O1O1N2O1O1O1N2O001N2O1O1O1N2O1O1O001O1O100O1O1O1O1O1O00100O100000000001N10000L4I71O1O2O0O1O1O1O1O1O2N100O1O1O2N1O1O100O1OWNZFOd91]FOb91`FN_92bFM^93dFI^96eFF^99cFF]9:eFC]9<dFC^9;dFB^9=dF@^9?cF@^9?dF^O^9`0dF_O^9?dF^O^9a0dF]O]9b0dF\\O^9c0bF\\O`9c0_F^Ob9a0]F_Of9>[FBf9;[FEg91bFMd9GcF8Y;O0O2N1O2O0O2N2O1N3M3N1NRo?"}, "label": "a lonely piece of broccoli separated from the rest of the broccoli"}]}
{"id": 143470, "image": "COCO_train2014_000000143470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of broccoli on the front right\"."}], "task": "refering", "answer_template": "The \"the piece of broccoli on the front right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274], "bbox": [0.74065625, 0.5078688524590165, 0.9400625, 0.8], "iscrowd": 0, "area": 9756.784499999998, "rle": {"size": [427, 640], "counts": "VnU61X=4L3M3M4L3M4M2M3M4L3M4L3M3N3M2M3N3M2N3L3N2N3L3N2N1O1N2O1O0O2O1O1N2O1O1O1N2O001N2O1O1O1N2O1O1O001O1O100O1O1O1O1O1O00100O100000000001N10000L4I71O1O2O0O1O1O1O1O1O2N100O1O1O2N1O1O100O1OWNZFOd91]FOb91`FN_92bFM^93dFI^96eFF^99cFF]9:eFC]9<dFC^9;dFB^9=dF@^9?cF@^9?dF^O^9`0dF_O^9?dF^O^9a0dF]O]9b0dF\\O^9c0bF\\O`9c0_F^Ob9a0]F_Of9>[FBf9;[FEg91bFMd9GcF8Y;O0O2N1O2O0O2N2O1N3M3N1NRo?"}, "label": "the piece of broccoli on the front right"}]}
{"id": 143470, "image": "COCO_train2014_000000143470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"piece of broccoli within a broccoli and pasta dish\"."}], "task": "refering", "answer_template": "The \"piece of broccoli within a broccoli and pasta dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194], "bbox": [0.29685937500000004, 0.3322014051522248, 0.509765625, 0.5883840749414521], "iscrowd": 0, "area": 9502.902900000003, "rle": {"size": [427, 640], "counts": "W]_26m<:F:K6M2M3N2M4M2M3N3L3N2M3N2M2O1N2O2M2O1N3N1N2O1N3N1N2O1N3N1N2N2O2M2N00O2N1O1O2O0O1O2N100O101O01O10O0100000O10000O10000O10000O10000001O2N2N1O2N2N1O2N1O2N2N1O00O100O100O1O100O100O1O2M2N3M2M4M3M2N3M2N3L4M2N3O001O1N101O1O001O0O2O1O001O2N3L4M2N3M3M3M2M3N1O1O2N1O2M2O1O2K4L5KngR4"}, "label": "piece of broccoli within a broccoli and pasta dish"}]}
{"id": 143470, "image": "COCO_train2014_000000143470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two pieces of broccoli on the left\"."}], "task": "refering", "answer_template": "The \"the two pieces of broccoli on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 257, 258], "bbox": [0.063671875, 0.26990632318501173, 0.322640625, 0.7671428571428571], "iscrowd": 0, "area": 21695.305, "rle": {"size": [427, 640], "counts": "[[a0>k<2N3L3N2000001O0000000O101O00000O101O00000000001O0000000O2O00000000001O00000dML]G8]8_2I6L4M4M2N3L3N3L3N3M2M4M2O2N1O2N3M3M3M3M3M3M3M3M2N3M3M2N3M1O1O1O2N10001O0000001O00001O0000001O00000014K6J5L4K5K4L01O0000001O00001O000O1O1O1O1O1O001O1O00100O001O001O0000000100O2N1O1O2O0O1O2N100O2N1O1O2O0O1O2N1O100fLQHT1T8gNTHR1o7iN[Ho0h7lNaHl0c7POeHi0^7ROlHf0X7UOQIb0S7ZOUI?n6@XI7l6G[IOj60[IHj66\\I@i6>^IYOf6f0Q3N4L4K5L4Lbjd5"}, "label": "the two pieces of broccoli on the left"}]}
{"id": 143470, "image": "COCO_train2014_000000143470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the three pieces of broccoli on the far left of the plate\"."}], "task": "refering", "answer_template": "The \"the three pieces of broccoli on the far left of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 233, 234, 235, 236, 257, 258], "bbox": [0.063671875, 0.26990632318501173, 0.322640625, 0.7671428571428571], "iscrowd": 0, "area": 21695.305, "rle": {"size": [427, 640], "counts": "[[a0>k<2N3L3N2000001O0000000O101O00000O101O00000000001O0000000O2O00000000001O00000dML]G8]8_2I6L4M4M2N3L3N3L3N3M2M4M2O2N1O2N3M3M3M3M3M3M3M3M2N3M3M2N3M1O1O1O2N10001O0000001O00001O0000001O00000014K6J5L4K5K4L01O0000001O00001O000O1O1O1O1O1O001O1O00100O001O001O0000000100O2N1O1O2O0O1O2N100O2N1O1O2O0O1O2N1O100fLQHT1T8gNTHR1o7iN[Ho0h7lNaHl0c7POeHi0^7ROlHf0X7UOQIb0S7ZOUI?n6@XI7l6G[IOj60[IHj66\\I@i6>^IYOf6f0Q3N4L4K5L4Lbjd5"}, "label": "the three pieces of broccoli on the far left of the plate"}]}
{"id": 299122, "image": "COCO_train2014_000000299122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cute little girl with long hair\"."}], "task": "refering", "answer_template": "The \"a cute little girl with long hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 368], "bbox": [0.003359375, 0.2670833333333333, 0.40862499999999996, 0.9509375], "iscrowd": 0, "area": 36939.703649999996, "rle": {"size": [480, 640], "counts": "kZ16h>g0ZOc0]O1O1N2O001O1N2O001N2O1O1O0O2O1O1N101O1O1N101O1O1N101O1N2O1O001N2O1O0O2O1O1O0O2O1O000O1000O0100000O101O0O101O000O2O00001N10001O0O101O000O101O000O2O0000O001O1O1O001O1O001O1PE^Mb9c2[FaMc9_2[FdMd9]2XFhMf9Y2WFkMg9V2UFnMj9R2TFRNj9o1RFVNl9k1QFXNn9i1nE\\NP:d1nE`NP:a1lEdNR:]1kEfNT:[1iEiNU:R3N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2M2O2N2N1O2N2N1O2H8A?A>B?A?C<0100O010O10O0100O010O100O00100O010O10O0100O010O100O010O10O0100O010O10aMgJVKZ5i4]KaJc4]5_KcJa4\\5aKbJ`4]5aKcJ^4]5dKbJ\\4\\5fKdJZ4[5hKcJY4\\5hKdJX4[5jKdJV4Z5lKeJT4[5nKdJR4[5oKeJQ4Z5QLdJP4Z5RLfJn3Y5TLfJk3Z5VLeJk3Z5WLeJi3Y5YLgJg3X5[LgJe3X5\\LgJe3X5]LgJb3Y5`LfJ`3X5bLgJ`3W5bLhJ_3V5bLjJ_3T5cLjJ_3S5cLmJ]3R5eLmJ\\3Q5eLnJ]3P5eLoJ\\3n4fLRK[3l4^3N2N2N1O2O1O001O1O1O0O2O1O1O001O1O1O0O2O1O001N2O1O001N2O1O0O2O1O1O0O2O1O001N2O1O001N2O1O0O2O1O001N2O1O0N3L8I:E;E;F:E<Db0_Oa0^OXn`5"}, "label": "a cute little girl with long hair"}]}
{"id": 299122, "image": "COCO_train2014_000000299122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl in a rainbow sweater smiling in front of a pizza\"."}], "task": "refering", "answer_template": "The \"a young girl in a rainbow sweater smiling in front of a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 368], "bbox": [0.003359375, 0.2670833333333333, 0.40862499999999996, 0.9509375], "iscrowd": 0, "area": 36939.703649999996, "rle": {"size": [480, 640], "counts": "kZ16h>g0ZOc0]O1O1N2O001O1N2O001N2O1O1O0O2O1O1N101O1O1N101O1O1N101O1N2O1O001N2O1O0O2O1O1O0O2O1O000O1000O0100000O101O0O101O000O2O00001N10001O0O101O000O101O000O2O0000O001O1O1O001O1O001O1PE^Mb9c2[FaMc9_2[FdMd9]2XFhMf9Y2WFkMg9V2UFnMj9R2TFRNj9o1RFVNl9k1QFXNn9i1nE\\NP:d1nE`NP:a1lEdNR:]1kEfNT:[1iEiNU:R3N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2M2O2N2N1O2N2N1O2H8A?A>B?A?C<0100O010O10O0100O010O100O00100O010O10O0100O010O100O010O10O0100O010O10aMgJVKZ5i4]KaJc4]5_KcJa4\\5aKbJ`4]5aKcJ^4]5dKbJ\\4\\5fKdJZ4[5hKcJY4\\5hKdJX4[5jKdJV4Z5lKeJT4[5nKdJR4[5oKeJQ4Z5QLdJP4Z5RLfJn3Y5TLfJk3Z5VLeJk3Z5WLeJi3Y5YLgJg3X5[LgJe3X5\\LgJe3X5]LgJb3Y5`LfJ`3X5bLgJ`3W5bLhJ_3V5bLjJ_3T5cLjJ_3S5cLmJ]3R5eLmJ\\3Q5eLnJ]3P5eLoJ\\3n4fLRK[3l4^3N2N2N1O2O1O001O1O1O0O2O1O1O001O1O1O0O2O1O001N2O1O001N2O1O0O2O1O1O0O2O1O001N2O1O001N2O1O0O2O1O001N2O1O0N3L8I:E;E;F:E<Db0_Oa0^OXn`5"}, "label": "a young girl in a rainbow sweater smiling in front of a pizza"}]}
{"id": 348277, "image": "COCO_train2014_000000348277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle with one smaller wheel and a larger wheel\"."}], "task": "refering", "answer_template": "The \"a bicycle with one smaller wheel and a larger wheel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 51, 52, 53, 62, 63, 64, 66, 67, 68, 69, 70, 80, 81, 84, 85, 86, 87, 98], "bbox": [0.45582, 0.014, 0.95204, 0.3976266666666667], "iscrowd": 0, "area": 21698.161000000007, "rle": {"size": [375, 500], "counts": "lcc24^;:G9G5M3N2O0O2O1O1N101O0O100O10000O100O10O10O1O1O001O001O0O1O1O1O1[O`NRGc1h8`NYGa1c8aN^G`1^8gN^GZ1`8nNYGS1d8P1N2O11O1O001N2K5L4M3M3M3M3N1N3K5O11O0O2O1N101N2N2N110O10O11N10000000000000000000000O101O00000000000000001O00001O0000001O00101N3M2N3M4L5K5L2M4L3M3M3M3M2O0O2N1O00000000001O00000000001O000000000000O10000000000000O010000000000000O0100000000000O10O10000000000000O0100000001O00000O1000001O0O1000000O10001O0O10000O101O000O2O00000O2O00001N10001N1000001N100O2N100O2N1O101N1O2O1N2N2N2O1N2N2N2N2N2O2M3M3M3M2N3M3M3M3M3K4L5K5K5K;D^g8"}, "label": "a bicycle with one smaller wheel and a larger wheel"}]}
{"id": 348277, "image": "COCO_train2014_000000348277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large bicycle with small back wheel and larger front wheel hanging from the ceiling next to a dort mall clock\"."}], "task": "refering", "answer_template": "The \"large bicycle with small back wheel and larger front wheel hanging from the ceiling next to a dort mall clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 51, 52, 53, 62, 63, 64, 66, 67, 68, 69, 70, 80, 81, 84, 85, 86, 87, 98], "bbox": [0.45582, 0.014, 0.95204, 0.3976266666666667], "iscrowd": 0, "area": 21698.161000000007, "rle": {"size": [375, 500], "counts": "lcc24^;:G9G5M3N2O0O2O1O1N101O0O100O10000O100O10O10O1O1O001O001O0O1O1O1O1[O`NRGc1h8`NYGa1c8aN^G`1^8gN^GZ1`8nNYGS1d8P1N2O11O1O001N2K5L4M3M3M3M3N1N3K5O11O0O2O1N101N2N2N110O10O11N10000000000000000000000O101O00000000000000001O00001O0000001O00101N3M2N3M4L5K5L2M4L3M3M3M3M2O0O2N1O00000000001O00000000001O000000000000O10000000000000O010000000000000O0100000000000O10O10000000000000O0100000001O00000O1000001O0O1000000O10001O0O10000O101O000O2O00000O2O00001N10001N1000001N100O2N100O2N1O101N1O2O1N2N2N2O1N2N2N2N2N2O2M3M3M3M2N3M3M3M3M3K4L5K5K5K;D^g8"}, "label": "large bicycle with small back wheel and larger front wheel hanging from the ceiling next to a dort mall clock"}]}
{"id": 348277, "image": "COCO_train2014_000000348277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a suspended green bike\"."}], "task": "refering", "answer_template": "The \"a suspended green bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 18, 19, 20, 21, 22, 23, 24, 38, 39, 40, 41, 42, 43, 57, 58, 59, 60, 61, 76, 77, 78, 79, 95, 96], "bbox": [0.0, 0.00288, 0.41974, 0.4215466666666667], "iscrowd": 0, "area": 16190.50785, "rle": {"size": [375, 500], "counts": "[18o:?B?E:10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1L4L4L4K5L4L4L401O000001O000000001O0000001O000000001eEGV9:iF0m82QG7f8KXG?^8C`G>_8C`G=`8E]G=b8E\\G;d8GZG9f8HYG9f8IXG7h8KUG6k8LSG5l8LSG4m8NQG2o80oF1P91nFOR92lFOT93jFNU9Z10001O00001O00001O00001O00001O00001O00001O00O1000000000000O1000000000000O13M3M3M3N2M3M4L3M3M3M3M4L3M2N1O1O1O1O2N1O1O1O1O2N1O1O13N3L4L3M4L4L2NO0000001O0001O0001O0000001O01O000001O0000010O0001O1N3N1O1O1N2O2N1N2O1O2N1N2O1O2M2O1O1O2M2M3N2N1N3N2N2N2M3N1O2M3N2N2K5I7J5KWTZ3"}, "label": "a suspended green bike"}]}
{"id": 348277, "image": "COCO_train2014_000000348277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green bike hanging from ceiling\"."}], "task": "refering", "answer_template": "The \"green bike hanging from ceiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 18, 19, 20, 21, 22, 23, 24, 38, 39, 40, 41, 42, 43, 57, 58, 59, 60, 61, 76, 77, 78, 79, 95, 96], "bbox": [0.0, 0.00288, 0.41974, 0.4215466666666667], "iscrowd": 0, "area": 16190.50785, "rle": {"size": [375, 500], "counts": "[18o:?B?E:10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1L4L4L4K5L4L4L401O000001O000000001O0000001O000000001eEGV9:iF0m82QG7f8KXG?^8C`G>_8C`G=`8E]G=b8E\\G;d8GZG9f8HYG9f8IXG7h8KUG6k8LSG5l8LSG4m8NQG2o80oF1P91nFOR92lFOT93jFNU9Z10001O00001O00001O00001O00001O00001O00001O00O1000000000000O1000000000000O13M3M3M3N2M3M4L3M3M3M3M4L3M2N1O1O1O1O2N1O1O1O1O2N1O1O13N3L4L3M4L4L2NO0000001O0001O0001O0000001O01O000001O0000010O0001O1N3N1O1O1N2O2N1N2O1O2N1N2O1O2M2O1O1O2M2M3N2N1N3N2N2N2M3N1O2M3N2N2K5I7J5KWTZ3"}, "label": "green bike hanging from ceiling"}]}
{"id": 348277, "image": "COCO_train2014_000000348277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike tire obscured by light\"."}], "task": "refering", "answer_template": "The \"bike tire obscured by light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 36, 37, 38, 54, 55, 56, 57, 58, 73, 74, 75, 76, 91, 92, 93, 94, 109, 110, 111], "bbox": [0.00176, 0.09176, 0.25236, 0.5270666666666667], "iscrowd": 0, "area": 9835.6051, "rle": {"size": [375, 500], "counts": "j=8^;10O10SE4o9LQF4n9LRF4o9LQF4o9LQF3o9NPF3P:LQF3`9NUFO:4a9MUFO:3a9NUF093b9LVF093a9NVFO93`9NXFO73b9MWF172a9NXFO73b9MXF063a9MYF161a9OYFO63a9MZF052a9N[F033a9N[F041b9N[F033a9M]F022b9N[F131b0Lh72dG111b01e7MiG012?7c7GmG002`0<^7BRH101>a0]7\\OVH2N1>g0Y7WO[H0O2;l0W7TO]HO10;R1T7nN`H001:V1S7iNcH0009]1Q7cNeHO127b1o6]NhH0207f1m6ZNjH0205i1o6WNiHO314k1P7UNhH0402n1R7QNhH1303P2Q7PNhH0401S2R7mMiHO411U2R7kMgH060OX2S7hMhH06OO[2S7eMhH150O\\2S7dMgH08OM_2T7bMgH070M`2U7_MgH17OMc2T7^MgH08OKf2V7ZMfH19OKh2U7XMgH19OJj2U7WMgH09OJm2V7\\MQIGHn2W7[MPIGHQ3W7YMPIFHS3X7fMhH\\2V7T1M3M3LZKgIU4_7^O2O00O10000O10O001O001O1O001O0O2O1O001O001O0O2O1O001O1O1O0O2O1O1O001O1O0O2O1O1O001O1O0O2N2L4M2N3L4M3L3N3M3L3N4L6I8I6I7J6J6I7J]jX4"}, "label": "bike tire obscured by light"}]}
{"id": 348277, "image": "COCO_train2014_000000348277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large white light glare over the bicycle\"."}], "task": "refering", "answer_template": "The \"the large white light glare over the bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 36, 37, 38, 54, 55, 56, 57, 58, 73, 74, 75, 76, 91, 92, 93, 94, 109, 110, 111], "bbox": [0.00176, 0.09176, 0.25236, 0.5270666666666667], "iscrowd": 0, "area": 9835.6051, "rle": {"size": [375, 500], "counts": "j=8^;10O10SE4o9LQF4n9LRF4o9LQF4o9LQF3o9NPF3P:LQF3`9NUFO:4a9MUFO:3a9NUF093b9LVF093a9NVFO93`9NXFO73b9MWF172a9NXFO73b9MXF063a9MYF161a9OYFO63a9MZF052a9N[F033a9N[F041b9N[F033a9M]F022b9N[F131b0Lh72dG111b01e7MiG012?7c7GmG002`0<^7BRH101>a0]7\\OVH2N1>g0Y7WO[H0O2;l0W7TO]HO10;R1T7nN`H001:V1S7iNcH0009]1Q7cNeHO127b1o6]NhH0207f1m6ZNjH0205i1o6WNiHO314k1P7UNhH0402n1R7QNhH1303P2Q7PNhH0401S2R7mMiHO411U2R7kMgH060OX2S7hMhH06OO[2S7eMhH150O\\2S7dMgH08OM_2T7bMgH070M`2U7_MgH17OMc2T7^MgH08OKf2V7ZMfH19OKh2U7XMgH19OJj2U7WMgH09OJm2V7\\MQIGHn2W7[MPIGHQ3W7YMPIFHS3X7fMhH\\2V7T1M3M3LZKgIU4_7^O2O00O10000O10O001O001O1O001O0O2O1O001O001O0O2O1O001O1O1O0O2O1O1O001O1O0O2O1O1O001O1O0O2N2L4M2N3L4M3L3N3M3L3N4L6I8I6I7J6J6I7J]jX4"}, "label": "the large white light glare over the bicycle"}]}
{"id": 36981, "image": "COCO_train2014_000000036981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in the middle who is sitting on the motorcycle\"."}], "task": "refering", "answer_template": "The \"a girl in the middle who is sitting on the motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 128, 129, 130, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 290, 291, 292, 313, 314, 335, 336, 337, 359, 360, 361, 383, 384], "bbox": [0.40203125, 0.28941666666666666, 0.7179062500000001, 0.989875], "iscrowd": 0, "area": 22245.246649999994, "rle": {"size": [480, 640], "counts": "UUi33m>3M1N2O1O1O001O00001O001O1O1O000000000O1O010O1O100O1O100O1000000O1000000000000O10000000000000000000000000000000000000000000000O0100000O10O1000O10O1000O100O0100O0100O0100O010O010O00100O00100O1O10000O0100O10000O1XBHh<8SCOi0Ea:=aD4m0A`:;_D9P1]O^:a1`EbN]:_1cEaN\\:a1cE_N\\:c1cE]NZ:f1fEZNW:i1iEWNT:l1kEVNo9Q2oEoMg9]2WFcMf9c2WF]M`9n2^FRM`9R3]FoL`9W3]FiLa9\\3\\FdLc9`3ZF`Ld9f3UF]Li9^4O1N2hHXKV4j4eK]KX4d4fK_KX4a4gKcKV4^4hKeKV4\\4hKhKU4Y4iKjKU4V4jKnKS4S4kKPLT4P4jKTLT4l3hKYLW4f3gK^LX4b3eKcLY4_3aKfL^4^3[KgLc4]3UKhLj4^3mJgLQ5e6O2N1O2N001O1O001O100O001O1O0100000O1002M2ZNnJ[IS5`6QK`Io4\\6UKdIl4V6XKlIg4Q6\\KoIe4n5\\KSJd4k5^KUJb4j5_KVJb4g5_KZJXOfNn4n6lK]JROgNT5i6mK`JkNiNY5e6oKgJS4V5PLhJS4T5PLjJT4Q5nKnJV4m4lKRKW4i4mKUKU4h4nKVKT4f4PLXKR4e4QLYKR4b4QL]KR4_4PL`KT4Z4PLdKT4W4nKhKW4P4mKoKW4j3mKULV4e3oKYLT4b3oK]LT4?[IZ1e2VNR48hIV1Z2aNQ40TJT1m1lNR4GaJN\\Na0T3HS4@VKe0j0Ij64YIJi64YIIQ;Ni\\d2"}, "label": "a girl in the middle who is sitting on the motorcycle"}]}
{"id": 36981, "image": "COCO_train2014_000000036981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sits on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a woman sits on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 128, 129, 130, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 290, 291, 292, 313, 314, 335, 336, 337, 359, 360, 361, 383, 384], "bbox": [0.40203125, 0.28941666666666666, 0.7179062500000001, 0.989875], "iscrowd": 0, "area": 22245.246649999994, "rle": {"size": [480, 640], "counts": "UUi33m>3M1N2O1O1O001O00001O001O1O1O000000000O1O010O1O100O1O100O1000000O1000000000000O10000000000000000000000000000000000000000000000O0100000O10O1000O10O1000O100O0100O0100O0100O010O010O00100O00100O1O10000O0100O10000O1XBHh<8SCOi0Ea:=aD4m0A`:;_D9P1]O^:a1`EbN]:_1cEaN\\:a1cE_N\\:c1cE]NZ:f1fEZNW:i1iEWNT:l1kEVNo9Q2oEoMg9]2WFcMf9c2WF]M`9n2^FRM`9R3]FoL`9W3]FiLa9\\3\\FdLc9`3ZF`Ld9f3UF]Li9^4O1N2hHXKV4j4eK]KX4d4fK_KX4a4gKcKV4^4hKeKV4\\4hKhKU4Y4iKjKU4V4jKnKS4S4kKPLT4P4jKTLT4l3hKYLW4f3gK^LX4b3eKcLY4_3aKfL^4^3[KgLc4]3UKhLj4^3mJgLQ5e6O2N1O2N001O1O001O100O001O1O0100000O1002M2ZNnJ[IS5`6QK`Io4\\6UKdIl4V6XKlIg4Q6\\KoIe4n5\\KSJd4k5^KUJb4j5_KVJb4g5_KZJXOfNn4n6lK]JROgNT5i6mK`JkNiNY5e6oKgJS4V5PLhJS4T5PLjJT4Q5nKnJV4m4lKRKW4i4mKUKU4h4nKVKT4f4PLXKR4e4QLYKR4b4QL]KR4_4PL`KT4Z4PLdKT4W4nKhKW4P4mKoKW4j3mKULV4e3oKYLT4b3oK]LT4?[IZ1e2VNR48hIV1Z2aNQ40TJT1m1lNR4GaJN\\Na0T3HS4@VKe0j0Ij64YIJi64YIIQ;Ni\\d2"}, "label": "a woman sits on a motorcycle"}]}
{"id": 36981, "image": "COCO_train2014_000000036981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman whose head is blocking the o in rock\"."}], "task": "refering", "answer_template": "The \"the woman whose head is blocking the o in rock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 236, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 326, 327, 328, 349, 350, 372, 373], "bbox": [0.051578124999999996, 0.267, 0.33559375, 0.9889791666666667], "iscrowd": 0, "area": 27898.54369999999, "rle": {"size": [480, 640], "counts": "kf?5j>:E5K3M4M2M3M3N2M4L3N2M3N3L3N2M3N2N3L3N2N2M3N2M4M2N2N2N2N3N1N2N2N2N2N3M2N2O0O2N101N100O2O0O2O0O101N101N1O101N2N3N2M2O2M3M2O2MP1PO2O1N2N2J6F:L4M2N3L4K5K5K5K8I7K5K3M3M2N2N3M2N2N3N1N2O2N2N2iJeGe3]8TLmGg3V8QLUHh3n7QL[Hk3g7QL]Hm3e7oK_Ho3c7mKaHQ4a7kKbHT4U9O10000000000O1000L4J6K6K4N2O1O1N2O1O1N2O1O1N2O2N1O1N2O9G8H8H8YJkIS2^6bMoIT2W6dMUJR2m5jM]Jn1d5lMgJn1[5mMnJ>iNWN\\6W1SK;iNWNU6\\1]K3eNZNP6_1gKLk50bJE_58jJ@Y5=jJAX5=kJ^OX5b0jJXO[5g0gJSO^5l0cJPOa5o0aJkNc5U1U4O1O00100O00100O0010O01O001O001O00001O001O001O001O001O001O001L7I<D`PW6"}, "label": "the woman whose head is blocking the o in rock"}]}
{"id": 36981, "image": "COCO_train2014_000000036981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blond girl in red dress posing with left arm behind head\"."}], "task": "refering", "answer_template": "The \"blond girl in red dress posing with left arm behind head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 236, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 326, 327, 328, 349, 350, 372, 373], "bbox": [0.051578124999999996, 0.267, 0.33559375, 0.9889791666666667], "iscrowd": 0, "area": 27898.54369999999, "rle": {"size": [480, 640], "counts": "kf?5j>:E5K3M4M2M3M3N2M4L3N2M3N3L3N2M3N2N3L3N2N2M3N2M4M2N2N2N2N3N1N2N2N2N2N3M2N2O0O2N101N100O2O0O2O0O101N101N1O101N2N3N2M2O2M3M2O2MP1PO2O1N2N2J6F:L4M2N3L4K5K5K5K8I7K5K3M3M2N2N3M2N2N3N1N2O2N2N2iJeGe3]8TLmGg3V8QLUHh3n7QL[Hk3g7QL]Hm3e7oK_Ho3c7mKaHQ4a7kKbHT4U9O10000000000O1000L4J6K6K4N2O1O1N2O1O1N2O1O1N2O2N1O1N2O9G8H8H8YJkIS2^6bMoIT2W6dMUJR2m5jM]Jn1d5lMgJn1[5mMnJ>iNWN\\6W1SK;iNWNU6\\1]K3eNZNP6_1gKLk50bJE_58jJ@Y5=jJAX5=kJ^OX5b0jJXO[5g0gJSO^5l0cJPOa5o0aJkNc5U1U4O1O00100O00100O0010O01O001O001O00001O001O001O001O001O001O001L7I<D`PW6"}, "label": "blond girl in red dress posing with left arm behind head"}]}
{"id": 36981, "image": "COCO_train2014_000000036981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing with both hands on her hips\"."}], "task": "refering", "answer_template": "The \"a woman standing with both hands on her hips\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 340, 341, 362, 363, 364, 385, 386, 387], "bbox": [0.70090625, 0.14616666666666667, 0.9616250000000001, 0.983625], "iscrowd": 0, "area": 34931.6348, "rle": {"size": [480, 640], "counts": "cdb64h>6K4kD7m6OkH9P7LiGZ1S8jNdG`1V8fN_Ge1\\8`NZGi1a8\\NVGn1e8Z2J6K7I7H9G8H8H8H8H8H8G9I6gJcGf3f8jKfGR4g9K5L4K6K5K5L4K6K4K5L5J5L4K>D0000000O010000O1[N[NoEd1o9dNlE\\1Q:lNiEU1T:SOgEm0V:P2M22O2N2N2N1O2N2N1OP1PO3M3M3M3M3M3M3M1O000001O00100O1O1O1O1O1O003ZLdMbJ_2n4VNjJm1j4aNRKa1f4eNYK^1_4iN_KZ1Y4mNeKU1S4TOjKn0b3HZL9c3LYL6d30XL2d34WL5_3NlKgLjNd3o4IbK_L_Od4h4QO`Ki1[4[N]Kn1^4^4K5K4M4L4M1N2O1N2O2M2O1O1N2N2O1N2N2O1N3N1N2N2O1N2N2O1N2N2O2M2O1N2N2O1N2N2O1N2N3N1N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2N2N2N2N3M2N2M3M3M3L4M3M3M4MX`;"}, "label": "a woman standing with both hands on her hips"}]}
{"id": 36981, "image": "COCO_train2014_000000036981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl standing on the right hand side with long dark curly hair with her hands on her hips\"."}], "task": "refering", "answer_template": "The \"a girl standing on the right hand side with long dark curly hair with her hands on her hips\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 340, 341, 362, 363, 364, 385, 386, 387], "bbox": [0.70090625, 0.14616666666666667, 0.9616250000000001, 0.983625], "iscrowd": 0, "area": 34931.6348, "rle": {"size": [480, 640], "counts": "cdb64h>6K4kD7m6OkH9P7LiGZ1S8jNdG`1V8fN_Ge1\\8`NZGi1a8\\NVGn1e8Z2J6K7I7H9G8H8H8H8H8H8G9I6gJcGf3f8jKfGR4g9K5L4K6K5K5L4K6K4K5L5J5L4K>D0000000O010000O1[N[NoEd1o9dNlE\\1Q:lNiEU1T:SOgEm0V:P2M22O2N2N2N1O2N2N1OP1PO3M3M3M3M3M3M3M1O000001O00100O1O1O1O1O1O003ZLdMbJ_2n4VNjJm1j4aNRKa1f4eNYK^1_4iN_KZ1Y4mNeKU1S4TOjKn0b3HZL9c3LYL6d30XL2d34WL5_3NlKgLjNd3o4IbK_L_Od4h4QO`Ki1[4[N]Kn1^4^4K5K4M4L4M1N2O1N2O2M2O1O1N2N2O1N2N2O1N3N1N2N2O1N2N2O1N2N2O2M2O1N2N2O1N2N2O1N2N3N1N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2N2N2N2N3M2N2M3M3M3L4M3M3M4MX`;"}, "label": "a girl standing on the right hand side with long dark curly hair with her hands on her hips"}]}
{"id": 53370, "image": "COCO_train2014_000000053370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second motorcycle with vsu996\"."}], "task": "refering", "answer_template": "The \"the second motorcycle with vsu996\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 251, 272, 273, 274, 295, 296, 297], "bbox": [0.6623593750000001, 0.2584166666666667, 0.9471875000000001, 0.7528125], "iscrowd": 0, "area": 19106.4183, "rle": {"size": [480, 640], "counts": "gmV66b><H8L5J4M2M3N2M3N2M3N2N1N2O1O001N2O1O001N2O00000O2O000O2O000O101N1O1O2ROkMoDU2Q;nMkDS2U;QNfDP2Y;UNbDm1];WN^Dj1b;YN[Dg1d;l0O00000O2O00001N10000O2O00010O010O0100O010O010O100O10O001O2N1O1O1O1O2N2N2N^MoDW1o:iNUEU1i:kN\\ER1b:nNcEn0\\:SOfEl0X:TOkEk0S:UOoEk0o9UOSFk0k9UOXFj0f9WO[Fi0c9WO`Fh0_9WOcFi0[9WO[FQOWOg1^:WO[FW1c9jN\\FX1c9gN]F[1b9bN`F`1^9^NdFd1[9XNhFk1U9SNmFP2Q9mMQGV2o8dMTG_2R9VMPGl2a:3M3M3N2M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3N1N1O00000000O1000000O100O10001N100O100O10000O100O10000O100O100O1O100nKeE^3R;M3L4L4M3L3N1N3N2M2N3N2M2N3K5K5K4K6K5K4L5K5K4L5K:F>BRe?"}, "label": "the second motorcycle with vsu996"}]}
{"id": 53370, "image": "COCO_train2014_000000053370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle with the license place ysu996\"."}], "task": "refering", "answer_template": "The \"a motorcycle with the license place ysu996\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 251, 272, 273, 274, 295, 296, 297], "bbox": [0.6623593750000001, 0.2584166666666667, 0.9471875000000001, 0.7528125], "iscrowd": 0, "area": 19106.4183, "rle": {"size": [480, 640], "counts": "gmV66b><H8L5J4M2M3N2M3N2M3N2N1N2O1O001N2O1O001N2O00000O2O000O2O000O101N1O1O2ROkMoDU2Q;nMkDS2U;QNfDP2Y;UNbDm1];WN^Dj1b;YN[Dg1d;l0O00000O2O00001N10000O2O00010O010O0100O010O010O100O10O001O2N1O1O1O1O2N2N2N^MoDW1o:iNUEU1i:kN\\ER1b:nNcEn0\\:SOfEl0X:TOkEk0S:UOoEk0o9UOSFk0k9UOXFj0f9WO[Fi0c9WO`Fh0_9WOcFi0[9WO[FQOWOg1^:WO[FW1c9jN\\FX1c9gN]F[1b9bN`F`1^9^NdFd1[9XNhFk1U9SNmFP2Q9mMQGV2o8dMTG_2R9VMPGl2a:3M3M3N2M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3N1N1O00000000O1000000O100O10001N100O100O10000O100O10000O100O100O1O100nKeE^3R;M3L4L4M3L3N1N3N2M2N3N2M2N3K5K5K4K6K5K4L5K5K4L5K:F>BRe?"}, "label": "a motorcycle with the license place ysu996"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older daughter in stripe shirt\"."}], "task": "refering", "answer_template": "The \"older daughter in stripe shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 26, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 278, 279, 280, 281, 282, 302, 303, 304, 305, 326], "bbox": [0.107296875, 0.0005, 0.446109375, 0.8386250000000001], "iscrowd": 0, "area": 51127.33984999999, "rle": {"size": [480, 640], "counts": "gdP1:_>9F;E:F;^Ob0G8H9H7[NeM\\Fc2R9aNTFg1d9o1J7H7J7I6J7I6I8I6K6I6K7I6I8I7I4L4K4M4L3M4M2N2O2O0O100aNkFRMW9X2aGeM_8X2gGdMZ8Y2mGcMT8Y2RHdMn7Y2XHdMi7X2]HdMd7Z2aHcM`7[2dHbM\\7\\2iH`MY7]2mH_MS7`2h2M2O2O1O00000O0100M3M3N2M2O200O1O1O100O1O100000000O01000000000O100O100O100O100O10000000000000O10O100000000000000000000O1YOkLbEV3Z:PMcEQ3Z:UMbEl2[:[M^Eh2`:m0N2N2N2N2[J_Kh0d4SOcKi0_4UOfKf0\\4ZOfKb0\\4^OgK=[4BlK6V4JYLEi3;cLWO_3i0kLkNW3U1lLfNV3Z1mLaNU3^1oL]NS3c1oLYNS3g1PMTNR3l1QMmMS3S2PMfMT3Z2nL`MV3_2nL\\MT3d2oLWMS3i2oLSMS3m2PMnLR3R3TMfLo2X3^MZLd2f3dMkKc2U4T400O10000O1000000O100O100O1O100O100O1O1000000000000O1000000O10000O1@WEiLi:W3`000O100O100O1000000O10000O10000O1O1N2O1N2N2O2M5K4M3L4L4M3L4M4K4L4M3L4L4J7B=C=CVmU5"}, "label": "older daughter in stripe shirt"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby is watching plate by wearing white tea shirt with red , pink and light green lines\"."}], "task": "refering", "answer_template": "The \"a baby is watching plate by wearing white tea shirt with red , pink and light green lines\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 26, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 278, 279, 280, 281, 282, 302, 303, 304, 305, 326], "bbox": [0.107296875, 0.0005, 0.446109375, 0.8386250000000001], "iscrowd": 0, "area": 51127.33984999999, "rle": {"size": [480, 640], "counts": "gdP1:_>9F;E:F;^Ob0G8H9H7[NeM\\Fc2R9aNTFg1d9o1J7H7J7I6J7I6I8I6K6I6K7I6I8I7I4L4K4M4L3M4M2N2O2O0O100aNkFRMW9X2aGeM_8X2gGdMZ8Y2mGcMT8Y2RHdMn7Y2XHdMi7X2]HdMd7Z2aHcM`7[2dHbM\\7\\2iH`MY7]2mH_MS7`2h2M2O2O1O00000O0100M3M3N2M2O200O1O1O100O1O100000000O01000000000O100O100O100O100O10000000000000O10O100000000000000000000O1YOkLbEV3Z:PMcEQ3Z:UMbEl2[:[M^Eh2`:m0N2N2N2N2[J_Kh0d4SOcKi0_4UOfKf0\\4ZOfKb0\\4^OgK=[4BlK6V4JYLEi3;cLWO_3i0kLkNW3U1lLfNV3Z1mLaNU3^1oL]NS3c1oLYNS3g1PMTNR3l1QMmMS3S2PMfMT3Z2nL`MV3_2nL\\MT3d2oLWMS3i2oLSMS3m2PMnLR3R3TMfLo2X3^MZLd2f3dMkKc2U4T400O10000O1000000O100O100O1O100O100O1O1000000000000O1000000O10000O1@WEiLi:W3`000O100O100O1000000O10000O10000O1O1N2O1N2N2O2M5K4M3L4L4M3L4M4K4L4M3L4L4J7B=C=CVmU5"}, "label": "a baby is watching plate by wearing white tea shirt with red , pink and light green lines"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wood chair behind he girl with the striped shirt\"."}], "task": "refering", "answer_template": "The \"a wood chair behind he girl with the striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 232, 253, 254, 255, 276, 277, 278, 279, 300, 301, 302, 323, 324, 325, 347], "bbox": [0.001953125, 0.6181875, 0.181171875, 0.9090833333333334], "iscrowd": 0, "area": 7167.086649999998, "rle": {"size": [480, 640], "counts": "lh01m>4L4L4L4L4L4L4O1O1O1O100O1O100O1O1O100O1O100O1O1O1M3N2M3N2M3M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3M3N2M3N2M3M3N2M3N1N0010O00010O010O006K;D;E4M2M4L3N2M4M2M3M4M2M3M4M2M3M3N3L7J8G9G9H1N1O2O0O2N100O2O0O2N101N1O2O0O1O2O0O2O0O2N10_We7"}, "label": "a wood chair behind he girl with the striped shirt"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chair with children on it\"."}], "task": "refering", "answer_template": "The \"wooden chair with children on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 232, 253, 254, 255, 276, 277, 278, 279, 300, 301, 302, 323, 324, 325, 347], "bbox": [0.001953125, 0.6181875, 0.181171875, 0.9090833333333334], "iscrowd": 0, "area": 7167.086649999998, "rle": {"size": [480, 640], "counts": "lh01m>4L4L4L4L4L4L4O1O1O1O100O1O100O1O1O100O1O100O1O1O1M3N2M3N2M3M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3M3N2M3N2M3M3N2M3N1N0010O00010O010O006K;D;E4M2M4L3N2M4M2M3M4M2M3M4M2M3M3N3L7J8G9G9H1N1O2O0O2N100O2O0O2N101N1O2O0O1O2O0O2O0O2N10_We7"}, "label": "wooden chair with children on it"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kitchen cabniet full of plates of various food\"."}], "task": "refering", "answer_template": "The \"a kitchen cabniet full of plates of various food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.4413958333333333, 1.0, 1.0], "iscrowd": 0, "area": 107745.92535000002, "rle": {"size": [480, 640], "counts": "om01o>000O100O10000O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O100001O00001O00001O001O00O100N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000000000000000000000O1000000000000000000000000O100O100O100VOPMcEQ3]:oLcEQ3\\:QMbEP3]:QMbEP3]:QMbEP3^:PMaEQ3^:QMZEDN^3g:g0O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100DWK`Fj4_9XK_Fi4a9YK\\Fh4c9;00000000000000N2O1O1O1001O1O0\\K`Fh3a9TLdFj3\\9TLhFj3Y9RLlFl3T9QLQGm3P9oKTGP4l8nKWGQ4j8kKZGT4j9O1O1O1O001O00001O00oMQLQIo3n6ULoHk3P7WLoHi3Q7WLoHi3P7YLoHg3Q7YLoHg3P7[LoHe3Q7\\LnHd3Q7]LVH\\4i7fKSH]4m7cKRH^4m7dKRH\\4n7dKSH[4l7gKTHX4l7iKSHW4l7jKUHU4j7mKUHS4k7nKUHQ4j7QLVHn3j7TLTHl3k7XLQHi3o7ZLnGf3Q8]LmGc3R8aLjG`3V8bLgG_3Y8dLbG^3^8b10000001O000000000000000000000000000000O10000000000O100000000N2O1O1O1O1N2O1O1L4F:O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O10000O10000000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000O100O100O1O100O1O100O1001O0000001O00001O00001O00001O00001O0000001O00001O00001OO100O100O100O100O100O1O100O1000000O100000000O100000000000000000000000000000SKUIe1k6dMWISN0X4i6dMXITN1V4g6eMZITN0V4f6fMZITN1U4e6fM\\ISN0W4d6eM]ITN0V4c6fM^ISNOW4c6eM_ITNMX4d6cMaITNKY4d6cMaITNJZ4e6bMcIRNG]4g6_MeIRND_4g6_MgInMDc4e6_MUJa2k5_MUJa2k5^MVJb2j5^MUJc2k5]MVJb2j5]MXJb2h5^MYJa2h5^MYJa2g5_MXJb2h5]MYJc2g5UMSI[NU1a4i5SMRIZNV1d4[6^LaHkNU1h4[6\\LiIe3X6YLhIh3S6WKbHQ1\\1h3R6\\LmIe3S6ZLmIg3S6XLnIh3T6`KbH3\\1]4T6\\KdH3Z1a4e6ZK^If4d6VK^I?\\NL2W3U8[L_I>^NM5V3g9iLXFO7T3W:hLnEV3R:hLQFW3o9fLTFZ3h:1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1N2M3N2K5M3O1O1O1O1O1N2O1O1O1N2O100O1O100O1O10000O100000dKVFb3j9]LYFa3g9]L^F`3c9\\LbFb3^9YLjFd3W9VLPGh3P9TLVGj3S:M2N3M2N3M2N2N"}, "label": "a kitchen cabniet full of plates of various food"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing glasses and a black shirt\"."}], "task": "refering", "answer_template": "The \"a lady wearing glasses and a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 33, 34, 35, 36, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 171, 172, 173, 174, 195, 196], "bbox": [0.431828125, 0.0, 0.710234375, 0.5112708333333333], "iscrowd": 0, "area": 24143.49659999999, "rle": {"size": [480, 640], "counts": "l^Q41i><nBMe:?VDb0c;b1H8H8H8I7J6L4M3L4M3L4M3M3M2N3M3L4M9TOf0K5K5J6N2N2N2N2M2O2N2N1O2N1O2N2N1O2N2N1O2N1O00O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O00100O1O100O1bJmGn3T8QLnGn3R8PLPHo3Q8oKRHo3o7PLRHP4n7nKUHP4l7nKWHP4j7nKXHQ4i7nKYHQ4g7mK[HR4f7kK]HT4d7jK^HV4b7gKaHX4`7dKdH[4]7_KjH`4V7XKRIo2aNcNS:^1UFXNl9i1\\FnMd9S2dFcM]9^2kFUMY9l2]1101N1O100O100O100O12N3M2N2O1N2L4M3L4L4M3L4M4K=De0ZO8H10O0100O100O1O10O010A>O2N1O1O2N1N3N1O1O20O100O1N3NO001O1O02O2M2O2M2O2M2O2M2O2N1N3N1N3N1N3[OfA`0_>M2N3M2N2N3MfRg2"}, "label": "a lady wearing glasses and a black shirt"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in glasses stands by the children in the kitchen\"."}], "task": "refering", "answer_template": "The \"a woman in glasses stands by the children in the kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 33, 34, 35, 36, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 171, 172, 173, 174, 195, 196], "bbox": [0.431828125, 0.0, 0.710234375, 0.5112708333333333], "iscrowd": 0, "area": 24143.49659999999, "rle": {"size": [480, 640], "counts": "l^Q41i><nBMe:?VDb0c;b1H8H8H8I7J6L4M3L4M3L4M3M3M2N3M3L4M9TOf0K5K5J6N2N2N2N2M2O2N2N1O2N1O2N2N1O2N2N1O2N1O00O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O00100O1O100O1bJmGn3T8QLnGn3R8PLPHo3Q8oKRHo3o7PLRHP4n7nKUHP4l7nKWHP4j7nKXHQ4i7nKYHQ4g7mK[HR4f7kK]HT4d7jK^HV4b7gKaHX4`7dKdH[4]7_KjH`4V7XKRIo2aNcNS:^1UFXNl9i1\\FnMd9S2dFcM]9^2kFUMY9l2]1101N1O100O100O100O12N3M2N2O1N2L4M3L4L4M3L4M4K=De0ZO8H10O0100O100O1O10O010A>O2N1O1O2N1N3N1O1O20O100O1N3NO001O1O02O2M2O2M2O2M2O2M2O2N1N3N1N3N1N3[OfA`0_>M2N3M2N2N3MfRg2"}, "label": "a woman in glasses stands by the children in the kitchen"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small child between the other picking up the plate\"."}], "task": "refering", "answer_template": "The \"the small child between the other picking up the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 99, 100, 101, 121, 122, 123, 124, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 261, 262, 284, 285], "bbox": [0.297640625, 0.1829583333333333, 0.49103125, 0.7345208333333333], "iscrowd": 0, "area": 15903.286199999999, "rle": {"size": [480, 640], "counts": "eVi21o>5J7J6I:G:E<E:E8I3L3N3L4M3L3N3L4M3Mo0PO>C1N2O1N2O1N101N2O1N2O1N2O1O1N101O1N2O1O1N2O1O1N101O1^GfJ[7\\5^HkJa7V5XHPKh7T6O2O001N10001N101O0O1N3MRNVHjLh7X3YHhLc7Z3_HeL_7]3bHcL[7^3hH`LV7b3kH]LS7d3oHSLFfNY7W5SIoKKfNQ7\\5UI_KC[Of7V5jH^K@XOh7[5iH\\K^7c4dH\\K[7e4fH[KX7e4kHYKT7h4mHXKQ7h4QIWKn6j4TIUKk6j4WIUKh6l4YIRKg6n4[IQKb6Q5aIlJZ6Z5gIeJ[6X5iIdJ[6Y5gIeJ[6X5iIeJY6Y5iIdJZ6U1hHS2U1fLU6f0]IW2e8WMTH[2_:G7I7L4O1N2O1N2O1O1N2O1N2:FN2N2N2M3N1O2N1OfBgNo<W1oBjNT=S1jBPOW=o0fBSO]=k0aBVOa=T12M3M2O2M2O3L4L3N3G`bh4"}, "label": "the small child between the other picking up the plate"}]}
{"id": 323722, "image": "COCO_train2014_000000323722.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girls with whales and fish on her shirt\"."}], "task": "refering", "answer_template": "The \"the little girls with whales and fish on her shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 99, 100, 101, 121, 122, 123, 124, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 261, 262, 284, 285], "bbox": [0.297640625, 0.1829583333333333, 0.49103125, 0.7345208333333333], "iscrowd": 0, "area": 15903.286199999999, "rle": {"size": [480, 640], "counts": "eVi21o>5J7J6I:G:E<E:E8I3L3N3L4M3L3N3L4M3Mo0PO>C1N2O1N2O1N101N2O1N2O1N2O1O1N101O1N2O1O1N2O1O1N101O1^GfJ[7\\5^HkJa7V5XHPKh7T6O2O001N10001N101O0O1N3MRNVHjLh7X3YHhLc7Z3_HeL_7]3bHcL[7^3hH`LV7b3kH]LS7d3oHSLFfNY7W5SIoKKfNQ7\\5UI_KC[Of7V5jH^K@XOh7[5iH\\K^7c4dH\\K[7e4fH[KX7e4kHYKT7h4mHXKQ7h4QIWKn6j4TIUKk6j4WIUKh6l4YIRKg6n4[IQKb6Q5aIlJZ6Z5gIeJ[6X5iIdJ[6Y5gIeJ[6X5iIeJY6Y5iIdJZ6U1hHS2U1fLU6f0]IW2e8WMTH[2_:G7I7L4O1N2O1N2O1O1N2O1N2:FN2N2N2M3N1O2N1OfBgNo<W1oBjNT=S1jBPOW=o0fBSO]=k0aBVOa=T12M3M2O2M2O3L4L3N3G`bh4"}, "label": "the little girls with whales and fish on her shirt"}]}
{"id": 364683, "image": "COCO_train2014_000000364683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant facing towards a red cement wall\"."}], "task": "refering", "answer_template": "The \"elephant facing towards a red cement wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 287, 288, 289, 291, 292, 293, 309, 310, 311, 312, 314, 315, 316, 332, 333, 334, 335, 337, 338, 339, 360, 361, 362], "bbox": [0.45949999999999996, 0.3710625, 0.7954062499999999, 0.9053958333333334], "iscrowd": 0, "area": 33035.4288, "rle": {"size": [480, 640], "counts": "TVZ4;d>1O1O2N1O1N3H7N201N1O2N010O1O1O100O001O100O1O0O2O1_JiNXMX1h2jNTMX1l2jNQMW1o2kNmLV1T3lNiLU1W3lNfLV1Z3lNcLU1]3mN_LU1`3nN]LS1c3oNYLR1h3POUL`0hKVOT8<PLOaL@`7c0lKGPM@U7j0iKE_5=]JCe5?XJ@j5b0RJ^OP6d0mI[OU6f0hIZOZ6h0cIWO_6k0]IUOe6l0YISOi6R1PInNR7W1hHhNFHo4d1UKdNJLQ5c1QK`NN0P5d1mJ\\N31P5f1gJZN:1o4i1_JXNc00n4k1YJVNj00n4m1PJUNQ11o4n1iIRNR18U5i1cIQNR1=[5f1gHeNg1Lb5R3WJUMi5k2PJ\\MP6d2iIcMW6]2aIkM_6U2ZIRNf6n1SIYNm6m400000000O10000000000O10000000000O101OkNiHaJW7\\5YIWJh6d5aIWJ_6e5hIXJX6d5nIZJR6c5SJYJo5b5j1H8J7K4L4L4M300000001O0000000000001O000002N1O1O100O1O1O10O01O1O100O1^HTKn4m4PKXKl4i4RKXKo4g4PKZKP5g4nJZKR5g4lJZKT5g4jJZKW5f4gJ\\KX5e4]JeKc5\\4QJoKo5S4cIYL]6U4jHVLW7i50000001O0000001O0000001O0000001O0O101O00001N10001O1O0O2O1O001N101O1O001N2O001O1O001O1O0O2O1O1O1mN^HfJc7R5fHlJ[7m4nHPKS7h4WITKk6e4l1L4M3MbFkKV8R4^G\\Lc8Q50O1O2M2F:E<E:E<E:E;E<D;I8J5K6J6J;C=B>B>AiTm1"}, "label": "elephant facing towards a red cement wall"}]}
{"id": 364683, "image": "COCO_train2014_000000364683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant facing towards a red wall to the right of another elephant facing away from the wall\"."}], "task": "refering", "answer_template": "The \"an elephant facing towards a red wall to the right of another elephant facing away from the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 287, 288, 289, 291, 292, 293, 309, 310, 311, 312, 314, 315, 316, 332, 333, 334, 335, 337, 338, 339, 360, 361, 362], "bbox": [0.45949999999999996, 0.3710625, 0.7954062499999999, 0.9053958333333334], "iscrowd": 0, "area": 33035.4288, "rle": {"size": [480, 640], "counts": "TVZ4;d>1O1O2N1O1N3H7N201N1O2N010O1O1O100O001O100O1O0O2O1_JiNXMX1h2jNTMX1l2jNQMW1o2kNmLV1T3lNiLU1W3lNfLV1Z3lNcLU1]3mN_LU1`3nN]LS1c3oNYLR1h3POUL`0hKVOT8<PLOaL@`7c0lKGPM@U7j0iKE_5=]JCe5?XJ@j5b0RJ^OP6d0mI[OU6f0hIZOZ6h0cIWO_6k0]IUOe6l0YISOi6R1PInNR7W1hHhNFHo4d1UKdNJLQ5c1QK`NN0P5d1mJ\\N31P5f1gJZN:1o4i1_JXNc00n4k1YJVNj00n4m1PJUNQ11o4n1iIRNR18U5i1cIQNR1=[5f1gHeNg1Lb5R3WJUMi5k2PJ\\MP6d2iIcMW6]2aIkM_6U2ZIRNf6n1SIYNm6m400000000O10000000000O10000000000O101OkNiHaJW7\\5YIWJh6d5aIWJ_6e5hIXJX6d5nIZJR6c5SJYJo5b5j1H8J7K4L4L4M300000001O0000000000001O000002N1O1O100O1O1O10O01O1O100O1^HTKn4m4PKXKl4i4RKXKo4g4PKZKP5g4nJZKR5g4lJZKT5g4jJZKW5f4gJ\\KX5e4]JeKc5\\4QJoKo5S4cIYL]6U4jHVLW7i50000001O0000001O0000001O0000001O0O101O00001N10001O1O0O2O1O001N101O1O001N2O001O1O001O1O0O2O1O1O1mN^HfJc7R5fHlJ[7m4nHPKS7h4WITKk6e4l1L4M3MbFkKV8R4^G\\Lc8Q50O1O2M2F:E<E:E<E:E;E<D;I8J5K6J6J;C=B>B>AiTm1"}, "label": "an elephant facing towards a red wall to the right of another elephant facing away from the wall"}]}
{"id": 364683, "image": "COCO_train2014_000000364683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with its trunk and face visible\"."}], "task": "refering", "answer_template": "The \"an elephant with its trunk and face visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 263, 264, 265, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 305, 306, 307, 309, 324, 325, 326, 328, 329, 330], "bbox": [0.050562499999999996, 0.3484583333333333, 0.5528125, 0.8585833333333334], "iscrowd": 0, "area": 53362.948099999994, "rle": {"size": [480, 640], "counts": "U\\?1`><D=D;D<D<D<D<D=D;J61O21N6J7I7I7J5J7I7I7I7J5J5L1O1O1O100O1O101N1O1O100O1O1O100O1O1O101N1O1O100O1OG:D;M4L3M\\GbJm7[5THhJl7U5SHoJl7S5PHPKP8Q5kGSKT8P5gGSKY8n4bGVK]8l4_GWK`8e500O100O100O10O01000000O10000O1000000O10000O1000O010[Oe0XOh0H7100000000O100000000O100000000O010000000O10000000000O100000000O100000000O100000000O01O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1OVHbKT5]4gJiKY5\\4]JiKc5`4oIeKQ6d4`IbKa6f4QI_Ko6j4bH\\K^7T6O0000000001O00000001O0001O0000000000000010O000000000001O0O1O100O1O1O101N1O1O100O1O1O2O0O1O1O100O1dM\\I[Le6c3^2M2M3N2N2N2N2M3N2N2N10100O10O0100O10O0100O1OVFeLP8\\3oGfLP8Y3PHiLo7W3PHkLo7T3QHnLm7S3RHoLm7P3SHSMk7m2THUMk7j2VHWMh7j2YHVMf7j2ZHWMe7j2[HVMd7j2]HWMa7i2_HXM_7i2bHWM]7i2dHWM[7j2dHWM[7i2fHXMW7i2jHWMU7i2kHXMT7h2mHXMR7i2nHWMP7j2PIWMo6i2RIXMl6h2TIYMk6h2RI[Ml6f2RI]Mm6c2QI`Mn6`2QIbMn6^2PIfMm6\\2PIgMo6Y2oHhMQ7Y2mHhMT7X2kHhMU7Y2iHhMX7X2hHgMY7X2gHhMY7Y2fHgM[7X2eHhM[7Y2dHgM]7X2cHhM^7X2aHhM_7Y2`HgMb7W2^HiMe7U2ZHkMh7S2XHmMj7R2UHnMm7P2THoMo7o1PHQNR8n1mGRNU8l1kGTNW8k1hGUNZ8j1eGVN^8k1^GTNe8n1WGRNk8o1RGPNQ9R2kFnMX9S2dFlM_9U2^FkMd9W2YFgMj9[2RFeMP:c34M2M3N2M4M2M3N2M3N3L3N2M3N2M6K7H7J7H8I6I8H8C<C>C=B=CgiU4"}, "label": "an elephant with its trunk and face visible"}]}
{"id": 364683, "image": "COCO_train2014_000000364683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant whose face is visible\"."}], "task": "refering", "answer_template": "The \"the elephant whose face is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 263, 264, 265, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 305, 306, 307, 309, 324, 325, 326, 328, 329, 330], "bbox": [0.050562499999999996, 0.3484583333333333, 0.5528125, 0.8585833333333334], "iscrowd": 0, "area": 53362.948099999994, "rle": {"size": [480, 640], "counts": "U\\?1`><D=D;D<D<D<D<D=D;J61O21N6J7I7I7J5J7I7I7I7J5J5L1O1O1O100O1O101N1O1O100O1O1O100O1O1O101N1O1O100O1OG:D;M4L3M\\GbJm7[5THhJl7U5SHoJl7S5PHPKP8Q5kGSKT8P5gGSKY8n4bGVK]8l4_GWK`8e500O100O100O10O01000000O10000O1000000O10000O1000O010[Oe0XOh0H7100000000O100000000O100000000O010000000O10000000000O100000000O100000000O100000000O01O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1OVHbKT5]4gJiKY5\\4]JiKc5`4oIeKQ6d4`IbKa6f4QI_Ko6j4bH\\K^7T6O0000000001O00000001O0001O0000000000000010O000000000001O0O1O100O1O1O101N1O1O100O1O1O2O0O1O1O100O1dM\\I[Le6c3^2M2M3N2N2N2N2M3N2N2N10100O10O0100O10O0100O1OVFeLP8\\3oGfLP8Y3PHiLo7W3PHkLo7T3QHnLm7S3RHoLm7P3SHSMk7m2THUMk7j2VHWMh7j2YHVMf7j2ZHWMe7j2[HVMd7j2]HWMa7i2_HXM_7i2bHWM]7i2dHWM[7j2dHWM[7i2fHXMW7i2jHWMU7i2kHXMT7h2mHXMR7i2nHWMP7j2PIWMo6i2RIXMl6h2TIYMk6h2RI[Ml6f2RI]Mm6c2QI`Mn6`2QIbMn6^2PIfMm6\\2PIgMo6Y2oHhMQ7Y2mHhMT7X2kHhMU7Y2iHhMX7X2hHgMY7X2gHhMY7Y2fHgM[7X2eHhM[7Y2dHgM]7X2cHhM^7X2aHhM_7Y2`HgMb7W2^HiMe7U2ZHkMh7S2XHmMj7R2UHnMm7P2THoMo7o1PHQNR8n1mGRNU8l1kGTNW8k1hGUNZ8j1eGVN^8k1^GTNe8n1WGRNk8o1RGPNQ9R2kFnMX9S2dFlM_9U2^FkMd9W2YFgMj9[2RFeMP:c34M2M3N2M4M2M3N2M3N3L3N2M3N2M6K7H7J7H8I6I8H8C<C>C=B=CgiU4"}, "label": "the elephant whose face is visible"}]}
{"id": 20619, "image": "COCO_train2014_000000020619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a jar containing purple candies\"."}], "task": "refering", "answer_template": "The \"a jar containing purple candies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154], "bbox": [0.5415625000000001, 0.0, 0.7246718750000001, 0.4], "iscrowd": 0, "area": 19632.072450000003, "rle": {"size": [480, 640], "counts": "ifR5a3^;2N2N101N2N1O2N2O1N1O2N2UOmKgFT4f8_LXGb3f8aLXGa3g8`LWGb3h8^LWGc3i8^LVGc3i8^LUGd3i8^LUGd3j8\\LUGe3k8\\LSGf3l8[LRGg3m8YLQGi3o8XLoFj3P9WLmFl3R9R1O001O1O00000000000000O100000000000000000000000000O10000000001O000000000000000O100000000000000000000000000O10000000000O100O100O1O100O1VKnFe3S9[LnFd3R9\\LoFb3R9^LoF`3R9`LoF^3R9bLnF^3R9bLoF\\3R9dLoFZ3R9eLPGZ3P9eLRGY3o8fLSGX3n8gLSGY3o8dLSGZ3o8dLSGZ3P9cLRG\\3P9aLRG\\3Q9bLPGP3`9mLbFe2m9XMUFZ2Z:cMhEn1m;Bf0ZOe0ZOW^b2"}, "label": "a jar containing purple candies"}]}
{"id": 20619, "image": "COCO_train2014_000000020619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle with purple pieces in it\"."}], "task": "refering", "answer_template": "The \"a bottle with purple pieces in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154], "bbox": [0.5415625000000001, 0.0, 0.7246718750000001, 0.4], "iscrowd": 0, "area": 19632.072450000003, "rle": {"size": [480, 640], "counts": "ifR5a3^;2N2N101N2N1O2N2O1N1O2N2UOmKgFT4f8_LXGb3f8aLXGa3g8`LWGb3h8^LWGc3i8^LVGc3i8^LUGd3i8^LUGd3j8\\LUGe3k8\\LSGf3l8[LRGg3m8YLQGi3o8XLoFj3P9WLmFl3R9R1O001O1O00000000000000O100000000000000000000000000O10000000001O000000000000000O100000000000000000000000000O10000000000O100O100O1O100O1VKnFe3S9[LnFd3R9\\LoFb3R9^LoF`3R9`LoF^3R9bLnF^3R9bLoF\\3R9dLoFZ3R9eLPGZ3P9eLRGY3o8fLSGX3n8gLSGY3o8dLSGZ3o8dLSGZ3P9cLRG\\3P9aLRG\\3Q9bLPGP3`9mLbFe2m9XMUFZ2Z:cMhEn1m;Bf0ZOe0ZOW^b2"}, "label": "a bottle with purple pieces in it"}]}
{"id": 20619, "image": "COCO_train2014_000000020619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tiny slice of white cake with pink icing and sprinkles\"."}], "task": "refering", "answer_template": "The \"a tiny slice of white cake with pink icing and sprinkles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279], "bbox": [0.048875, 0.4494375, 0.208984375, 0.7617916666666666], "iscrowd": 0, "area": 10816.211700000002, "rle": {"size": [480, 640], "counts": "ni>a0P=a1M3L5L3L5L3L4N3L3N3M2N3L3N2N3M2N3L3N2N3M1N2O1O1O1O1N2O1O1O1N2O1O1O1N10O1O1O1O001O1O1O001O1O1O1O001O1O10O01O1O1O001O1O1O1O001O1O1O001O1O1O2N1O2N2N1O2N2O0O2N2N2N1O2N2N1O2N2N2N1O2K5J5J7J6J5J7J6J6I6K6J6J5J7Jfn\\7"}, "label": "a tiny slice of white cake with pink icing and sprinkles"}]}
{"id": 20619, "image": "COCO_train2014_000000020619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"piece of cake on a plate with a candy sucker\"."}], "task": "refering", "answer_template": "The \"piece of cake on a plate with a candy sucker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279], "bbox": [0.048875, 0.4494375, 0.208984375, 0.7617916666666666], "iscrowd": 0, "area": 10816.211700000002, "rle": {"size": [480, 640], "counts": "ni>a0P=a1M3L5L3L5L3L4N3L3N3M2N3L3N2N3M2N3L3N2N3M1N2O1O1O1O1N2O1O1O1N2O1O1O1N10O1O1O1O001O1O1O001O1O1O1O001O1O10O01O1O1O001O1O1O1O001O1O1O001O1O1O2N1O2N2N1O2N2O0O2N2N2N1O2N2N1O2N2N2N1O2K5J5J7J6J5J7J6J6I6K6J6J5J7Jfn\\7"}, "label": "piece of cake on a plate with a candy sucker"}]}
{"id": 20619, "image": "COCO_train2014_000000020619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pie on end of table\"."}], "task": "refering", "answer_template": "The \"pie on end of table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 158, 159, 160], "bbox": [0.7264062499999999, 0.17527083333333332, 1.0, 0.37304166666666666], "iscrowd": 0, "area": 12205.395849999999, "rle": {"size": [480, 640], "counts": "VRj6f0Y>;E2M3N2N2N2N3M2O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O0O2O00000O101O000000001O00000000001O000O10O10000O10000O10000O10000O10000O100O10O0100O100O100O100O101N100O100O10000O100O101N100O100O100001O3N2M3M4L5L:E001O00100O001O0001O0000000O2O0000001O000000001O0000000O101O000000001O0000001O000O2O00001O00001O000O2N1O2N1O2N2N2M2O2N2N1N3M3N2M2M4J6J6J7I6K5J6J`L"}, "label": "pie on end of table"}]}
{"id": 20619, "image": "COCO_train2014_000000020619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"almost full pie\"."}], "task": "refering", "answer_template": "The \"almost full pie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 158, 159, 160], "bbox": [0.7264062499999999, 0.17527083333333332, 1.0, 0.37304166666666666], "iscrowd": 0, "area": 12205.395849999999, "rle": {"size": [480, 640], "counts": "VRj6f0Y>;E2M3N2N2N2N3M2O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O0O2O00000O101O000000001O00000000001O000O10O10000O10000O10000O10000O10000O100O10O0100O100O100O100O101N100O100O10000O100O101N100O100O100001O3N2M3M4L5L:E001O00100O001O0001O0000000O2O0000001O000000001O0000000O101O000000001O0000001O000O2O00001O00001O000O2N1O2N1O2N2N2M2O2N2N1N3M3N2M2M4J6J6J7I6K5J6J`L"}, "label": "almost full pie"}]}
{"id": 422029, "image": "COCO_train2014_000000422029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench closest to the camera\"."}], "task": "refering", "answer_template": "The \"the bench closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 307, 308, 309, 310, 311, 333], "bbox": [0.316734375, 0.4608899297423888, 0.61546875, 0.9753629976580797], "iscrowd": 0, "area": 21841.8066, "rle": {"size": [427, 640], "counts": "hkd2d0f<8I6I8I6I8H3N02N3M2O1N2N0000O10001N102N3M4K5L8H:E:G:F:E:G3M0O1[G_L`7a3`H_L`7a3`H_L_7b3bH]L^7c3bH]L^7c3bH]L^7c3bH]L]7d3cH\\L]7d3cH\\L]7d3cH\\L\\7e3eH[LZ7e3fH[LZ7e3fH[LY7f3gHZLY7f3gHZLY7f3gHZLX7g3hHYLX7g3iHXLW7h3iHXLW7h3iHXLV7i3jHWLV7i3jHXLU7h3kHXLT7h3mHXLS7h3nHWLR7i3nHWLQ7j3oHVLQ7j3oHVLQ7j3oHVLP7k3PIULP7k3PIULP7k3QIULm6l3SITLm6l3SITLm6l3SITLm6l3SITLl6m3TISLl6m3TISLl6m3UIRLj6o3VIQLj6o3VIQLj6o3VIQLi6P4WIQLh6o3XIQLh6o3XIQLg6P4ZIoKf6Q4ZIoKf6Q4ZIoKe6R4[InKe6R4[InKe6R4[InKe6R4[InKd6S4]IlKc6T4]ImKb6S4^ImKa6T4_IlKa6T4_IlKa6T4_IlK`6U4`IkK`6U4aIjK_6V4aIjK^6W4bIiK^6W4bIiK[6Z4eIfKV6_4jIbKU6^4kIbKU6^4lIaKT6_4lIaKT6_4lIaKT6h4cI\\KY6R5YISKd6n52N1K6I7I4L3M4L3N1N0000001O000YOg0oNQ1E;O2N1O1O1O1N2O2N1O1O1O1O1O1N2O1O16J9G9G9G9G3M00000000000000J6^Ob0]Oc0C=O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1W1iN0000O1000000O101O0O1`MfGN[82_2O100O100O1M6JlZV3"}, "label": "the bench closest to the camera"}]}
{"id": 422029, "image": "COCO_train2014_000000422029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"1st bench\"."}], "task": "refering", "answer_template": "The \"1st bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 307, 308, 309, 310, 311, 333], "bbox": [0.316734375, 0.4608899297423888, 0.61546875, 0.9753629976580797], "iscrowd": 0, "area": 21841.8066, "rle": {"size": [427, 640], "counts": "hkd2d0f<8I6I8I6I8H3N02N3M2O1N2N0000O10001N102N3M4K5L8H:E:G:F:E:G3M0O1[G_L`7a3`H_L`7a3`H_L_7b3bH]L^7c3bH]L^7c3bH]L^7c3bH]L]7d3cH\\L]7d3cH\\L]7d3cH\\L\\7e3eH[LZ7e3fH[LZ7e3fH[LY7f3gHZLY7f3gHZLY7f3gHZLX7g3hHYLX7g3iHXLW7h3iHXLW7h3iHXLV7i3jHWLV7i3jHXLU7h3kHXLT7h3mHXLS7h3nHWLR7i3nHWLQ7j3oHVLQ7j3oHVLQ7j3oHVLP7k3PIULP7k3PIULP7k3QIULm6l3SITLm6l3SITLm6l3SITLm6l3SITLl6m3TISLl6m3TISLl6m3UIRLj6o3VIQLj6o3VIQLj6o3VIQLi6P4WIQLh6o3XIQLh6o3XIQLg6P4ZIoKf6Q4ZIoKf6Q4ZIoKe6R4[InKe6R4[InKe6R4[InKe6R4[InKd6S4]IlKc6T4]ImKb6S4^ImKa6T4_IlKa6T4_IlKa6T4_IlK`6U4`IkK`6U4aIjK_6V4aIjK^6W4bIiK^6W4bIiK[6Z4eIfKV6_4jIbKU6^4kIbKU6^4lIaKT6_4lIaKT6_4lIaKT6h4cI\\KY6R5YISKd6n52N1K6I7I4L3M4L3N1N0000001O000YOg0oNQ1E;O2N1O1O1O1N2O2N1O1O1O1O1O1N2O1O16J9G9G9G9G3M00000000000000J6^Ob0]Oc0C=O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1W1iN0000O1000000O101O0O1`MfGN[82_2O100O100O1M6JlZV3"}, "label": "1st bench"}]}
{"id": 422029, "image": "COCO_train2014_000000422029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench closer to the garbage can\"."}], "task": "refering", "answer_template": "The \"the bench closer to the garbage can\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267], "bbox": [0.500765625, 0.4157377049180328, 0.778140625, 0.7730444964871195], "iscrowd": 0, "area": 13158.823650000004, "rle": {"size": [427, 640], "counts": "\\QV45U=5L4K5K4L5L4N2N2M3N2M3L4SDaNf;b1O0100000000000000000O10000000000000000O1000O10000000000000O1000000000000000000O100000O100000000000O100mD`Na0Le8d1iFiN:Ek8b1kFRO2^OP9`1mFTO2^Oo8^1oFUO1_On8\\1PGVO2_Om8[1QGWO1^On8[1PGXO2]OKFg8e1\\GYOM@JHl8`1[GZOKNj8U2RGPNm8W3000000000000O101O000000I7F:F:F:N110O100O1O100O100O10FoLYFQ3h9RMUFm2l9VMQFj2o98100O100O1O7J9F9L4MM3K6J6K4K010O01O0010OO2O0O1O1O1\\Od000000000000O100000000000000000000000001O000000000O1000002N3M0000000000O101O000\\Od0cN^1O00000M4IWPk1"}, "label": "the bench closer to the garbage can"}]}
{"id": 422029, "image": "COCO_train2014_000000422029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two benches that are close to a trash can\"."}], "task": "refering", "answer_template": "The \"two benches that are close to a trash can\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267], "bbox": [0.500765625, 0.4157377049180328, 0.778140625, 0.7730444964871195], "iscrowd": 0, "area": 13158.823650000004, "rle": {"size": [427, 640], "counts": "\\QV45U=5L4K5K4L5L4N2N2M3N2M3L4SDaNf;b1O0100000000000000000O10000000000000000O1000O10000000000000O1000000000000000000O100000O100000000000O100mD`Na0Le8d1iFiN:Ek8b1kFRO2^OP9`1mFTO2^Oo8^1oFUO1_On8\\1PGVO2_Om8[1QGWO1^On8[1PGXO2]OKFg8e1\\GYOM@JHl8`1[GZOKNj8U2RGPNm8W3000000000000O101O000000I7F:F:F:N110O100O1O100O100O10FoLYFQ3h9RMUFm2l9VMQFj2o98100O100O1O7J9F9L4MM3K6J6K4K010O01O0010OO2O0O1O1O1\\Od000000000000O100000000000000000000000001O000000000O1000002N3M0000000000O101O000\\Od0cN^1O00000M4IWPk1"}, "label": "two benches that are close to a trash can"}]}
{"id": 151699, "image": "COCO_train2014_000000151699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holds an apple in front of a man with a bow\"."}], "task": "refering", "answer_template": "The \"a woman holds an apple in front of a man with a bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351, 352], "bbox": [0.045203125, 0.13401805869074493, 0.544078125, 0.9898645598194131], "iscrowd": 0, "area": 51072.849749999994, "rle": {"size": [443, 640], "counts": "Uk<:\\=e0[O=D5J7I6J6J6J6J7I6J6J6I7J7H7I7J6I7I8I6I7I7J6I8H7J6\\MkIZM\\6b2iIVMf6^2`IYMR7X2SIaMV7Y2oHbMS7[2RIbMo6[2WIbMj6Z2[IcMf6Z2`IbMa6]2bIaMZ6c2iIYMQ6m2RJQMn5o2UJmLm5R3UJkLm5S3VJkLl5S3WJiLk5U3XJiLj5U3YJgLi5X3YJeLj5X3^2O2N1nLbEd2l:O0O2O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O0010O01O1O001O001OmFPN[6P2dIRN[6n1dISN\\6m1bIVN]6j1bIWN^6i1aIXNm0Ol1i1UM[Ni01R2d1UM[Ng04S2a1VM\\Ne04U2b1TMZNe06W2b1RMXNe09X2b1QMUNe0:Z2c1oLSNe0<\\2c1mLQNf0=]2e1jLoMf0?_2d1iLmMf0a0a2d1gLlMf0a0c2f1dLiMg0d0d2e1cLgMh0e0e2f1aLfMg0f0h2g1XLiMn0b0j2g1oKoMU1<k2g1gKTN\\17m2h1cKRN_17n2i1`KRN_18P3j1\\KoMn0m0f3Y1WKjMS1m0f3]1RKhMW1k0g3n2YLSMf3n2ZLQMf3o2ZLQMf3o2ZLQMf3P3YLQMf3o2ZLQMf3P3YLPMg3P3YLQMf3o2ZLQMf3P3YLPMg3P3YLPMg3P3YLQMf3P3YLPMg3P3YLPMg3Q3XLoLh3Q3XLoLh3Q3XLoLh3R3WLnLi3R3WLnLi3S3VLmLj3S3VLmLj3S3VLmLj3T3ULlLk3T3ULlLk3T3ULlLk3U3TLkLl3U3TLjLm3W3RLiLn3W3RLiLn3W3RLiLn3X3QLhLo3X3QLhLo3X3QLhLo3X3QLhLo3Y3PLgLP4Y3PLgLP4Y3PLgLP4Y3PLgLP4Y3PLgLP4Y3PLgLP4Z3oKfLR4Y3nKgLR4Y3nKgLR4Y3nKgLR4Y3nKgLR4Y3nKgLR4Z3mKfLS4Z3mKfLS4Z3mKfLR4[3nKeLQ4\\3oKdLP4]3PLcLo3_3PLaLo3`3QL`Ln3a3RL_Lm3b3SL^Ll3c3TL]Lk3d3UL\\Lj3f3ULZLk3f3ULZLj3g3VLZLh3g3XLZLf3g3ZL[Lc3f3]L[La3f3_L[L_3f3aL\\L\\3e3dL\\LZ3e3fL\\LX3e3hL]LU3d3kL]LS3d3mL]LR3c3mL`LP3a3PM`Lm2c3RM^Lk2d3UM^Lg2d3YM]Ld2e3\\M\\La2f3_M\\L]2f3cM]LY2d3gM_LT2c3lM`Lo1b3PNbLk1`3UNdLe1^3[NeL`1]3`NfL[1\\3eNgLX1Y3hNkLS1V3mNmLo0T3QOoLl0Q3TORMh0o2XOUMc0l2\\OXMa0h2_O[M=f2C]M9d2G`M5`2KcMFTKKZ7?eMBg2>\\M_Od2a0`MZOa2f0Z500O10000000001O0000000O1000000000000000000O1000000000000000000O10000000000000000000000000000000000000000000000001O00000001O00000000000000000000000000001O00000O1000000000000O100000001O000O100000001O3M2M4M3M[Tn3"}, "label": "a woman holds an apple in front of a man with a bow"}]}
{"id": 151699, "image": "COCO_train2014_000000151699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black dress holds an apple with an arrow in it\"."}], "task": "refering", "answer_template": "The \"a woman in a black dress holds an apple with an arrow in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351, 352], "bbox": [0.045203125, 0.13401805869074493, 0.544078125, 0.9898645598194131], "iscrowd": 0, "area": 51072.849749999994, "rle": {"size": [443, 640], "counts": "Uk<:\\=e0[O=D5J7I6J6J6J6J7I6J6J6I7J7H7I7J6I7I8I6I7I7J6I8H7J6\\MkIZM\\6b2iIVMf6^2`IYMR7X2SIaMV7Y2oHbMS7[2RIbMo6[2WIbMj6Z2[IcMf6Z2`IbMa6]2bIaMZ6c2iIYMQ6m2RJQMn5o2UJmLm5R3UJkLm5S3VJkLl5S3WJiLk5U3XJiLj5U3YJgLi5X3YJeLj5X3^2O2N1nLbEd2l:O0O2O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O0010O01O1O001O001OmFPN[6P2dIRN[6n1dISN\\6m1bIVN]6j1bIWN^6i1aIXNm0Ol1i1UM[Ni01R2d1UM[Ng04S2a1VM\\Ne04U2b1TMZNe06W2b1RMXNe09X2b1QMUNe0:Z2c1oLSNe0<\\2c1mLQNf0=]2e1jLoMf0?_2d1iLmMf0a0a2d1gLlMf0a0c2f1dLiMg0d0d2e1cLgMh0e0e2f1aLfMg0f0h2g1XLiMn0b0j2g1oKoMU1<k2g1gKTN\\17m2h1cKRN_17n2i1`KRN_18P3j1\\KoMn0m0f3Y1WKjMS1m0f3]1RKhMW1k0g3n2YLSMf3n2ZLQMf3o2ZLQMf3o2ZLQMf3P3YLQMf3o2ZLQMf3P3YLPMg3P3YLQMf3o2ZLQMf3P3YLPMg3P3YLPMg3P3YLQMf3P3YLPMg3P3YLPMg3Q3XLoLh3Q3XLoLh3Q3XLoLh3R3WLnLi3R3WLnLi3S3VLmLj3S3VLmLj3S3VLmLj3T3ULlLk3T3ULlLk3T3ULlLk3U3TLkLl3U3TLjLm3W3RLiLn3W3RLiLn3W3RLiLn3X3QLhLo3X3QLhLo3X3QLhLo3X3QLhLo3Y3PLgLP4Y3PLgLP4Y3PLgLP4Y3PLgLP4Y3PLgLP4Y3PLgLP4Z3oKfLR4Y3nKgLR4Y3nKgLR4Y3nKgLR4Y3nKgLR4Y3nKgLR4Z3mKfLS4Z3mKfLS4Z3mKfLR4[3nKeLQ4\\3oKdLP4]3PLcLo3_3PLaLo3`3QL`Ln3a3RL_Lm3b3SL^Ll3c3TL]Lk3d3UL\\Lj3f3ULZLk3f3ULZLj3g3VLZLh3g3XLZLf3g3ZL[Lc3f3]L[La3f3_L[L_3f3aL\\L\\3e3dL\\LZ3e3fL\\LX3e3hL]LU3d3kL]LS3d3mL]LR3c3mL`LP3a3PM`Lm2c3RM^Lk2d3UM^Lg2d3YM]Ld2e3\\M\\La2f3_M\\L]2f3cM]LY2d3gM_LT2c3lM`Lo1b3PNbLk1`3UNdLe1^3[NeL`1]3`NfL[1\\3eNgLX1Y3hNkLS1V3mNmLo0T3QOoLl0Q3TORMh0o2XOUMc0l2\\OXMa0h2_O[M=f2C]M9d2G`M5`2KcMFTKKZ7?eMBg2>\\M_Od2a0`MZOa2f0Z500O10000000001O0000000O1000000000000000000O1000000000000000000O10000000000000000000000000000000000000000000000001O00000001O00000000000000000000000000001O00000O1000000000000O100000001O000O100000001O3M2M4M3M[Tn3"}, "label": "a woman in a black dress holds an apple with an arrow in it"}]}
{"id": 151699, "image": "COCO_train2014_000000151699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little blond haired girl wearing a long blue dress , white shoes , and a gray holster over her neck\"."}], "task": "refering", "answer_template": "The \"a little blond haired girl wearing a long blue dress , white shoes , and a gray holster over her neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 87, 88, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296], "bbox": [0.68740625, 0.16503386004514672, 0.9341093749999999, 0.8113092550790069], "iscrowd": 0, "area": 24263.78935000001, "rle": {"size": [443, 640], "counts": "Uen53f=2N2N3M2N2N3M2M4M2N2N3M2N2N3M2N3M2N2N3M2O1N3M2N2N3`DSNn:o1QETNk:n1TESNT:f2jE\\Mm9l2SFTMf9S3ZFmLc9V3]FjLa9Y3^FhL_9Z3aFfL]9\\3cFdL[9^3eFbLY9`3fFaLX9b3gF_LV9c3jF]LT9e3lF[L`8JhGl3HZL\\8NjGj3JXLZ81jGi3KVLX84kGh3MTLV87kGf3NTLU88kGf30RLT8:jGe32PLT8=hGe33nKT8k4lGUKS8l4lGUKS8l4kGVKT8Y5O1N2N2M3N2N2^HoIX7S6gHoIW7R6gHPJW7R6hHoIW7Q6hHQJ]5NoJ2`0[6]4jIhJ7e0P6b4cJ[K^5d4fJXK[5g4hJVKZ5h4hJVKY5i4iJUKX5j4jJUKV5j4lJRKW5m4kJoJX5Q5i100O10000O10000O100O10000O1000O10000O1000001N10000O2@nJUHS5j7`0O001O1O1O2N1O2N1O2M5L4L4L3M4L4L4_KWIh1n6lM]IP2g6gMbIU2b6dMeIX2_6bMgIZ2\\6aMiI\\2Y6`MmI\\2V6^MQJ]2R6_MSJ^2P6]MVJ_2l5]MZJ_2i5nLkJn2T8M4K5L3L4M2M4M2M4M2M4M2N3L3N2M2O2M2O2M3N1N3N1N3N1N4M:F4K5L4K\\ka0"}, "label": "a little blond haired girl wearing a long blue dress , white shoes , and a gray holster over her neck"}]}
{"id": 151699, "image": "COCO_train2014_000000151699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl with blonde hair is wearing a blue princess style dress while a man attempts to shoot an apple on a woman ' s hand with a bow and arrow\"."}], "task": "refering", "answer_template": "The \"a young girl with blonde hair is wearing a blue princess style dress while a man attempts to shoot an apple on a woman ' s hand with a bow and arrow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 87, 88, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296], "bbox": [0.68740625, 0.16503386004514672, 0.9341093749999999, 0.8113092550790069], "iscrowd": 0, "area": 24263.78935000001, "rle": {"size": [443, 640], "counts": "Uen53f=2N2N3M2N2N3M2M4M2N2N3M2N2N3M2N3M2N2N3M2O1N3M2N2N3`DSNn:o1QETNk:n1TESNT:f2jE\\Mm9l2SFTMf9S3ZFmLc9V3]FjLa9Y3^FhL_9Z3aFfL]9\\3cFdL[9^3eFbLY9`3fFaLX9b3gF_LV9c3jF]LT9e3lF[L`8JhGl3HZL\\8NjGj3JXLZ81jGi3KVLX84kGh3MTLV87kGf3NTLU88kGf30RLT8:jGe32PLT8=hGe33nKT8k4lGUKS8l4lGUKS8l4kGVKT8Y5O1N2N2M3N2N2^HoIX7S6gHoIW7R6gHPJW7R6hHoIW7Q6hHQJ]5NoJ2`0[6]4jIhJ7e0P6b4cJ[K^5d4fJXK[5g4hJVKZ5h4hJVKY5i4iJUKX5j4jJUKV5j4lJRKW5m4kJoJX5Q5i100O10000O10000O100O10000O1000O10000O1000001N10000O2@nJUHS5j7`0O001O1O1O2N1O2N1O2M5L4L4L3M4L4L4_KWIh1n6lM]IP2g6gMbIU2b6dMeIX2_6bMgIZ2\\6aMiI\\2Y6`MmI\\2V6^MQJ]2R6_MSJ^2P6]MVJ_2l5]MZJ_2i5nLkJn2T8M4K5L3L4M2M4M2M4M2M4M2N3L3N2M2O2M2O2M3N1N3N1N3N1N4M:F4K5L4K\\ka0"}, "label": "a young girl with blonde hair is wearing a blue princess style dress while a man attempts to shoot an apple on a woman ' s hand with a bow and arrow"}]}
{"id": 151699, "image": "COCO_train2014_000000151699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man squatting holding a bow\"."}], "task": "refering", "answer_template": "The \"a man squatting holding a bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 288, 289, 290], "bbox": [0.41375, 0.13257336343115123, 0.732625, 0.7707900677200904], "iscrowd": 0, "area": 36329.879299999986, "rle": {"size": [443, 640], "counts": "`nb31j=00000O101cL5hHKU7?dHAX7i0aHXO^7P1[HPOe7W1THiNk7_1nGbNQ8e1UFeNl0Go8^2mFbMR9c2jF]MV9g2fFYMZ9k2aFVM^9n2]FUMb9n2ZFSMf9o2VFSMj9P3RFQMo9]30001O000000001O00000000001O00000000001OO100000000O100000000O100000000O100000O10O100000000O10WNcLTI]3k6eLTI[3k6gLTIi3[6YLdIi3Y6YLfIh3X6ZLgIg3W6[LgIg3W6[LhIf3V6[LjIe3U6]LiId3V6^LiIb3V6`LhIa3W6`LiI`3V6aLjI_3U6bLjI_3U6cLhI_3W6bLfIa3Y6`LeIb3Z6_LdIc3\\6]LbIe3]6\\LbIe3]6\\LaIf3^6[L`Ig3_6ZL`Ig3_6k100O1O1O100O1O1O1O1N2N2N2N2O1N2N2N2J6K5N2O1N2N2N1O200O1O100O010O1O010O1O10O01O10O0100O01000O010O1000O01000O10001N10000O101O000O101O0O10001M2L4L4M3L5L3L4L4N3O3L8I8G8I4K2N2O1N2O1N2O2M2O1N2N2O1N2bM\\HcNe7Y1SIQNo6k1\\ImMe6o1eIiM]6T2mIdMT6X2VJ`Ml5\\2l2L5L3L4L4L6K8G8_Ob0YO^kY2"}, "label": "a man squatting holding a bow"}]}
{"id": 151699, "image": "COCO_train2014_000000151699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with bow\"."}], "task": "refering", "answer_template": "The \"man with bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 288, 289, 290], "bbox": [0.41375, 0.13257336343115123, 0.732625, 0.7707900677200904], "iscrowd": 0, "area": 36329.879299999986, "rle": {"size": [443, 640], "counts": "`nb31j=00000O101cL5hHKU7?dHAX7i0aHXO^7P1[HPOe7W1THiNk7_1nGbNQ8e1UFeNl0Go8^2mFbMR9c2jF]MV9g2fFYMZ9k2aFVM^9n2]FUMb9n2ZFSMf9o2VFSMj9P3RFQMo9]30001O000000001O00000000001O00000000001OO100000000O100000000O100000000O100000O10O100000000O10WNcLTI]3k6eLTI[3k6gLTIi3[6YLdIi3Y6YLfIh3X6ZLgIg3W6[LgIg3W6[LhIf3V6[LjIe3U6]LiId3V6^LiIb3V6`LhIa3W6`LiI`3V6aLjI_3U6bLjI_3U6cLhI_3W6bLfIa3Y6`LeIb3Z6_LdIc3\\6]LbIe3]6\\LbIe3]6\\LaIf3^6[L`Ig3_6ZL`Ig3_6k100O1O1O100O1O1O1O1N2N2N2N2O1N2N2N2J6K5N2O1N2N2N1O200O1O100O010O1O010O1O10O01O10O0100O01000O010O1000O01000O10001N10000O101O000O101O0O10001M2L4L4M3L5L3L4L4N3O3L8I8G8I4K2N2O1N2O1N2O2M2O1N2N2O1N2bM\\HcNe7Y1SIQNo6k1\\ImMe6o1eIiM]6T2mIdMT6X2VJ`Ml5\\2l2L5L3L4L4L6K8G8_Ob0YO^kY2"}, "label": "man with bow"}]}
{"id": 4244, "image": "COCO_train2014_000000004244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a counter with white paper taped to it\"."}], "task": "refering", "answer_template": "The \"a counter with white paper taped to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [316, 317, 336, 337, 338, 339, 340, 341, 342, 343, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 428, 429, 430, 431, 432, 433, 434, 435, 453, 454, 455], "bbox": [0.2214705882352941, 0.664640522875817, 0.9463071895424836, 0.9465196078431373], "iscrowd": 0, "area": 38995.39505, "rle": {"size": [612, 612], "counts": "iga21Rc01O2O0O2N1O1O2O0O2N1O1O2O0O1O2N1O2N100O2N1O2N100O2N1O2N100O2N1O1O2O0O2N1O1O2O0O2N1O1O2N100O2N1O2N100O2N1O2N100O2O0O101N101N100O2O0O2O0O101N100O2O0O2O0O101N101N100O2O0O101N101N100O2O0O2O0O101N100O2O001O01O01O001O01O01O00010O001O01O01O0010O0001O01O01O001O01O01O0010O0001O01O01O0010O0001O0010O0001O01O01O0010O0001O010O000010O0001O0010O0001O010O000010O0001O010O000010O01O000010O0001O010O000010O01O00010O000010O01O000010O01O00010O000010O01O00010O001O00010O000010O01O00010O001O01O01O00010O001O00010O001O01O01O0010O0001O01O01O0010O0001O0010O0001O01O01N10O010O0010O010O01O010O1O010O0010O01O010O00100O0010O01O010O00100O100O2N100O1O2O0O1O2O0O1O101N1O100O2N100O1O2O0O100O2N100O1O2O0O1O101N1O100O2N100O2N100O1O2O0O100O2N100O1O2O0O1O101N1O100O2N100O1O2O0O1O101N100O1O2O0O1O2O0O1O101N1O100O2N100O1O2O0O1O101N100O1O2O0O1O101N1O100O2N100O1O2O0O1O2O0O1O101N1O100O2O0O1O101N1O100O2N100O1O2O0O1O101N1O10`Xc0"}, "label": "a counter with white paper taped to it"}]}
{"id": 4244, "image": "COCO_train2014_000000004244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white kitchen prep table with lime colored tape on it\"."}], "task": "refering", "answer_template": "The \"a white kitchen prep table with lime colored tape on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [316, 317, 336, 337, 338, 339, 340, 341, 342, 343, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 428, 429, 430, 431, 432, 433, 434, 435, 453, 454, 455], "bbox": [0.2214705882352941, 0.664640522875817, 0.9463071895424836, 0.9465196078431373], "iscrowd": 0, "area": 38995.39505, "rle": {"size": [612, 612], "counts": "iga21Rc01O2O0O2N1O1O2O0O2N1O1O2O0O1O2N1O2N100O2N1O2N100O2N1O2N100O2N1O1O2O0O2N1O1O2O0O2N1O1O2N100O2N1O2N100O2N1O2N100O2O0O101N101N100O2O0O2O0O101N100O2O0O2O0O101N101N100O2O0O101N101N100O2O0O2O0O101N100O2O001O01O01O001O01O01O00010O001O01O01O0010O0001O01O01O001O01O01O0010O0001O01O01O0010O0001O0010O0001O01O01O0010O0001O010O000010O0001O0010O0001O010O000010O0001O010O000010O01O000010O0001O010O000010O01O00010O000010O01O000010O01O00010O000010O01O00010O001O00010O000010O01O00010O001O01O01O00010O001O00010O001O01O01O0010O0001O01O01O0010O0001O0010O0001O01O01N10O010O0010O010O01O010O1O010O0010O01O010O00100O0010O01O010O00100O100O2N100O1O2O0O1O2O0O1O101N1O100O2N100O1O2O0O100O2N100O1O2O0O1O101N1O100O2N100O2N100O1O2O0O100O2N100O1O2O0O1O101N1O100O2N100O1O2O0O1O101N100O1O2O0O1O2O0O1O101N1O100O2N100O1O2O0O1O101N100O1O2O0O1O101N1O100O2N100O1O2O0O1O2O0O1O101N1O100O2O0O1O101N1O100O2N100O1O2O0O1O101N1O10`Xc0"}, "label": "a white kitchen prep table with lime colored tape on it"}]}
{"id": 37012, "image": "COCO_train2014_000000037012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white van that says roxy ' s on the side\"."}], "task": "refering", "answer_template": "The \"a white van that says roxy ' s on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 343, 344, 345, 346, 347, 348, 349, 350, 351, 366, 367, 368, 369, 370, 371, 372, 373, 392, 393, 394, 395], "bbox": [0.44495098039215686, 0.3033660130718954, 1.0, 0.784264705882353], "iscrowd": 0, "area": 79400.06240000001, "rle": {"size": [612, 612], "counts": "XlR51ob05K5J5L5K5K5J6K5K5J6K5K5K5J6K5K5K5J6K5K5K5J6K5K5K5J6K4L5K5J6K5M3O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O001O1O1O00001O0000001O000000001O0O10001O000000001O0000001O000000001O00000O2O000000001O0000001O0000001O000000001N1000001O000000001O0000001O000000001O0000010O0001O0010O0001O01O01O0010O0001O00010O001O01O01O000010O01O00010O00001O010O000010O0001O001O00001O000010O01O00001O00001O001O00001O01O01O001O00001O00001O001O00010O00001O00001O001O00001O00010O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00k4UKgC"}, "label": "a white van that says roxy ' s on the side"}]}
{"id": 37012, "image": "COCO_train2014_000000037012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green truck with the word roxy ' s gourmet grilled chi on it on the grass near other trucks and people walking\"."}], "task": "refering", "answer_template": "The \"a green truck with the word roxy ' s gourmet grilled chi on it on the grass near other trucks and people walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 343, 344, 345, 346, 347, 348, 349, 350, 351, 366, 367, 368, 369, 370, 371, 372, 373, 392, 393, 394, 395], "bbox": [0.44495098039215686, 0.3033660130718954, 1.0, 0.784264705882353], "iscrowd": 0, "area": 79400.06240000001, "rle": {"size": [612, 612], "counts": "XlR51ob05K5J5L5K5K5J6K5K5J6K5K5K5J6K5K5K5J6K5K5K5J6K5K5K5J6K4L5K5J6K5M3O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O001O1O1O00001O0000001O000000001O0O10001O000000001O0000001O000000001O00000O2O000000001O0000001O0000001O000000001N1000001O000000001O0000001O000000001O0000010O0001O0010O0001O01O01O0010O0001O00010O001O01O01O000010O01O00010O00001O010O000010O0001O001O00001O000010O01O00001O00001O001O00001O01O01O001O00001O00001O001O00010O00001O00001O001O00001O00010O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00k4UKgC"}, "label": "a green truck with the word roxy ' s gourmet grilled chi on it on the grass near other trucks and people walking"}]}
{"id": 37012, "image": "COCO_train2014_000000037012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in all black\"."}], "task": "refering", "answer_template": "The \"the woman in all black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 266, 267, 268, 269, 288, 289, 290, 310, 311, 312, 331, 332, 333, 353, 354, 355, 375, 376, 397, 398], "bbox": [0.06320261437908496, 0.3177124183006536, 0.28893790849673207, 0.8369117647058825], "iscrowd": 0, "area": 18668.77180000001, "rle": {"size": [612, 612], "counts": "nhg0=fb02N1O2J6H8H8G9G8H9G9K5M3M2N3M3M3M3M2N3M3M3M2N1O1O2N1dL^LVGd3i8^LTGc3k8`LRGa3n8bLnF`3P9cLkF`3U9cLWFn3h9ULeE]4Z:eKYEf4f:^KWEc4g:`KWE`4i:cKTE]4k:fKTEd3SOPLh;>SE`3\\OmK`;g0QEZ3FjKY;n0oDW3MhKR;U1oDP35gKl:[1mDl2<gKf:_1lDh2V<\\MgCc2Z<_MdC_2^<dM`CY2b<iM\\CU2e<nMYCQ2h<RNUCm1l<UNSCi1n<U2010O1O00100O1O2N2N101N2N1O2O1N2N1O2N2O0O2N2N2O0O2N2N2N101N2N2N10O0001O00010O00O101O0O100O20O1O1O0O2O1O1OOZKgAd4\\>02N1O2N2M3N2O1O1O1O1O1O1O0UMVA]1l>aNWA\\1l>`NYA]1i>_N\\A]1f>`N^A]1e>`N_A]1c>_NaA^1a>_NdA]1_>`NdA]1U`0M2M3M4M2M3M4L3N2M3MbfS8"}, "label": "the woman in all black"}]}
{"id": 37012, "image": "COCO_train2014_000000037012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in black with a pink hat standing to the left of a yellow roxy ' s truck\"."}], "task": "refering", "answer_template": "The \"person in black with a pink hat standing to the left of a yellow roxy ' s truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 266, 267, 268, 269, 288, 289, 290, 310, 311, 312, 331, 332, 333, 353, 354, 355, 375, 376, 397, 398], "bbox": [0.06320261437908496, 0.3177124183006536, 0.28893790849673207, 0.8369117647058825], "iscrowd": 0, "area": 18668.77180000001, "rle": {"size": [612, 612], "counts": "nhg0=fb02N1O2J6H8H8G9G8H9G9K5M3M2N3M3M3M3M2N3M3M3M2N1O1O2N1dL^LVGd3i8^LTGc3k8`LRGa3n8bLnF`3P9cLkF`3U9cLWFn3h9ULeE]4Z:eKYEf4f:^KWEc4g:`KWE`4i:cKTE]4k:fKTEd3SOPLh;>SE`3\\OmK`;g0QEZ3FjKY;n0oDW3MhKR;U1oDP35gKl:[1mDl2<gKf:_1lDh2V<\\MgCc2Z<_MdC_2^<dM`CY2b<iM\\CU2e<nMYCQ2h<RNUCm1l<UNSCi1n<U2010O1O00100O1O2N2N101N2N1O2O1N2N1O2N2O0O2N2N2O0O2N2N2N101N2N2N10O0001O00010O00O101O0O100O20O1O1O0O2O1O1OOZKgAd4\\>02N1O2N2M3N2O1O1O1O1O1O1O0UMVA]1l>aNWA\\1l>`NYA]1i>_N\\A]1f>`N^A]1e>`N_A]1c>_NaA^1a>_NdA]1_>`NdA]1U`0M2M3M4M2M3M4L3N2M3MbfS8"}, "label": "person in black with a pink hat standing to the left of a yellow roxy ' s truck"}]}
{"id": 323734, "image": "COCO_train2014_000000323734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of bread on the left of the photo\"."}], "task": "refering", "answer_template": "The \"the slice of bread on the left of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 46, 47, 69, 70, 92, 93, 94, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233], "bbox": [0.00140625, 0.01688279301745636, 0.19615625, 0.7376059850374064], "iscrowd": 0, "area": 19077.2956, "rle": {"size": [401, 640], "counts": "T`0e5a3[32N2N2M4M2N2N2N2N2N2N2N2N2N3M2M3N2N2N2N2N2N2N2N3M2N2N2N2M3N2N2N2N2N3M2N2N2N2N2N2M3N2N2N2N3M2N2N2N2N2N2N2M3N2N3M2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2N3M2N2M3N2N2N2N2N2N2N2N3M2N2N2N2M3N2N2N2N2N3M2N2N2N2N2N2M3N2N2N3M2N2N2N3M3M4L3M4L3L5L3M4L3M4L3M4L3MUPY6"}, "label": "the slice of bread on the left of the photo"}]}
{"id": 323734, "image": "COCO_train2014_000000323734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"corner of white bread on far left of sandwich\"."}], "task": "refering", "answer_template": "The \"corner of white bread on far left of sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 46, 47, 69, 70, 92, 93, 94, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233], "bbox": [0.00140625, 0.01688279301745636, 0.19615625, 0.7376059850374064], "iscrowd": 0, "area": 19077.2956, "rle": {"size": [401, 640], "counts": "T`0e5a3[32N2N2M4M2N2N2N2N2N2N2N2N2N3M2M3N2N2N2N2N2N2N2N3M2N2N2N2M3N2N2N2N2N3M2N2N2N2N2N2M3N2N2N2N3M2N2N2N2N2N2N2M3N2N3M2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2N3M2N2M3N2N2N2N2N2N2N2N3M2N2N2N2M3N2N2N2N2N3M2N2N2N2N2N2M3N2N2N3M2N2N2N3M3M4L3M4L3L5L3M4L3M4L3M4L3MUPY6"}, "label": "corner of white bread on far left of sandwich"}]}
{"id": 323734, "image": "COCO_train2014_000000323734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich on a whit plate next to some fries\"."}], "task": "refering", "answer_template": "The \"a sandwich on a whit plate next to some fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 70, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311], "bbox": [0.0, 0.0472069825436409, 0.63496875, 0.9867830423940149], "iscrowd": 0, "area": 102294.84155000004, "rle": {"size": [401, 640], "counts": "^;n0h9k1_Oa0000O100000000000iGhLU7X3kHjLS7V3mHlLQ7T3oHnLo6R3QIPMm6P3SIRMk6n2UITMi6l2WIVMg6j2XIYMf6g2ZI[Md6e2\\I^Ma6b2_I`M_6`2aIbM]6^2cIdM[6\\2eIfMY6Z2gIhMW6X2iIjMU6V2kIlMS6T2mInMQ6R2oIPNo5P2QJRNm5n1SJTNk5l1UJVNi5j1WJXNg5h1YJZNe5f1[J\\Nc5d1\\J_Nb5a1^JaN`5_1`JcN^5]1bJeN\\5[1dJgNZ5Y1fJiNX5W1hJkNV5U1jJmNT5S1lJoNR5Q1nJQOP5o0PKSOn4m0RKUOl4k0TKWOj4i0VKYOh4g0XK[Of4e0ZK]Od4c0\\K_Ob4a0]KBa4>_KD_4<aKF]4:cKH[48eKJY46gKLW44iKNU42kK0S40mK2Q4NoK4o3LQL6m3JSL8k3HUL:i3FWL<g3DYL>e3B[La0b3_O^Lc0`3]O_Lg0^3YObLi0\\3WOdLl0Y3TOgLn0W3ROiLQ1T3oNlLS1R3mNnLV1o2jNQMX1m2hNSM[1j2eNVM]1h2cNXM`1e2`N[Mc1b2]N^Me1`2[N`Mh1]2XNcMj1[2VNeMm1X2SNhMo1V2QNjMR2S2nMmMT2Q2lMoMW2n1iMRNY2l1gMTN\\2i1eMVN]2h1cMXN`2e1`M[Nb2c1^M]Ne2`1[M`Ng2^1YMbNj2[1VMeNl2Y1TMgNo2V1QMjNQ3T1oLlNT3Q1lLoNV3o0jLQOY3l0gLTO[3j0eLWO]3f0cLZO_3d0aL\\Ob3a0^L_Oe3>[LBg3<YLDj39VLGl37TLIo34QLLQ42oKNT4OlK1V4MjK3Y4JgK6[4HeK8^4DcK<_4BaK>b4_O^Ka0d4\\O]Kd0f4YOZKg0h4WOXKi0X8001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O1000000000000000000000000O100000000000000000000000000O10000000000002N2N2N2N2N2N2RHnNY4T1cKoN\\4S1_KQO`4Q1\\KROc4P1YKSOf4o0VKUOh4m0SKWOl4k0PKXOo4j0mJYOR5i0iJ[OV5g0[Ig0d6[OUIl0i6VOoHQ1P7QOiHU1V7T2O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1^OaIPK`6o4cImJ`6Q5cIkJ_6T5dIhJ^6W5?O1O2N1O2N1O1O2N1O2O0O2N1O1O2N1O2N1O1O2N1O2N1O2N1O101N000000000000000000000000O10O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N4M2M4M3L3N3L4M2M4M3L3L5K5K4L5K5L3L5K4L5K5K4L5K5K4M^[k2"}, "label": "a sandwich on a whit plate next to some fries"}]}
{"id": 323734, "image": "COCO_train2014_000000323734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich with red meat for breakfast\"."}], "task": "refering", "answer_template": "The \"sandwich with red meat for breakfast\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 70, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311], "bbox": [0.0, 0.0472069825436409, 0.63496875, 0.9867830423940149], "iscrowd": 0, "area": 102294.84155000004, "rle": {"size": [401, 640], "counts": "^;n0h9k1_Oa0000O100000000000iGhLU7X3kHjLS7V3mHlLQ7T3oHnLo6R3QIPMm6P3SIRMk6n2UITMi6l2WIVMg6j2XIYMf6g2ZI[Md6e2\\I^Ma6b2_I`M_6`2aIbM]6^2cIdM[6\\2eIfMY6Z2gIhMW6X2iIjMU6V2kIlMS6T2mInMQ6R2oIPNo5P2QJRNm5n1SJTNk5l1UJVNi5j1WJXNg5h1YJZNe5f1[J\\Nc5d1\\J_Nb5a1^JaN`5_1`JcN^5]1bJeN\\5[1dJgNZ5Y1fJiNX5W1hJkNV5U1jJmNT5S1lJoNR5Q1nJQOP5o0PKSOn4m0RKUOl4k0TKWOj4i0VKYOh4g0XK[Of4e0ZK]Od4c0\\K_Ob4a0]KBa4>_KD_4<aKF]4:cKH[48eKJY46gKLW44iKNU42kK0S40mK2Q4NoK4o3LQL6m3JSL8k3HUL:i3FWL<g3DYL>e3B[La0b3_O^Lc0`3]O_Lg0^3YObLi0\\3WOdLl0Y3TOgLn0W3ROiLQ1T3oNlLS1R3mNnLV1o2jNQMX1m2hNSM[1j2eNVM]1h2cNXM`1e2`N[Mc1b2]N^Me1`2[N`Mh1]2XNcMj1[2VNeMm1X2SNhMo1V2QNjMR2S2nMmMT2Q2lMoMW2n1iMRNY2l1gMTN\\2i1eMVN]2h1cMXN`2e1`M[Nb2c1^M]Ne2`1[M`Ng2^1YMbNj2[1VMeNl2Y1TMgNo2V1QMjNQ3T1oLlNT3Q1lLoNV3o0jLQOY3l0gLTO[3j0eLWO]3f0cLZO_3d0aL\\Ob3a0^L_Oe3>[LBg3<YLDj39VLGl37TLIo34QLLQ42oKNT4OlK1V4MjK3Y4JgK6[4HeK8^4DcK<_4BaK>b4_O^Ka0d4\\O]Kd0f4YOZKg0h4WOXKi0X8001O00000000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O1000000000000000000000000O100000000000000000000000000O10000000000002N2N2N2N2N2N2RHnNY4T1cKoN\\4S1_KQO`4Q1\\KROc4P1YKSOf4o0VKUOh4m0SKWOl4k0PKXOo4j0mJYOR5i0iJ[OV5g0[Ig0d6[OUIl0i6VOoHQ1P7QOiHU1V7T2O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1^OaIPK`6o4cImJ`6Q5cIkJ_6T5dIhJ^6W5?O1O2N1O2N1O1O2N1O2O0O2N1O1O2N1O2N1O1O2N1O2N1O2N1O101N000000000000000000000000O10O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N4M2M4M3L3N3L4M2M4M3L3L5K5K4L5K5L3L5K4L5K5K4L5K5K4M^[k2"}, "label": "sandwich with red meat for breakfast"}]}
{"id": 151702, "image": "COCO_train2014_000000151702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black male\"."}], "task": "refering", "answer_template": "The \"black male\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 73, 74, 85, 86, 96, 97, 108, 109, 120, 121, 132, 133], "bbox": [0.0, 0.38797260273972606, 0.23720496894409937, 0.9038356164383561], "iscrowd": 0, "area": 5544.156400000001, "rle": {"size": [365, 322], "counts": "n99R;2O1N3mNEQG<i7m0SHTOa7[1ZHhNe7\\1VHeNj7^1QHdNn7`1nGbNQ8a1kG`NU8\\2O1000000O10O10O10000O1000000O1N2N2N2O1N2N10gM^Hi0`7WOmH>Q7B]I2a6NlIGS68YJ^Oe5b0cJXO[5h0mJPOT5o0d2100O100O100O100O100O101N100O100]O_NZGb1e8cNWG^1f8gNVGZ1i8jNTGV1j8oNSGQ1m8h00O10O01O10O0100O3N6I7J6I7Jf`g2"}, "label": "black male"}]}
{"id": 151702, "image": "COCO_train2014_000000151702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black man\"."}], "task": "refering", "answer_template": "The \"black man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 73, 74, 85, 86, 96, 97, 108, 109, 120, 121, 132, 133], "bbox": [0.0, 0.38797260273972606, 0.23720496894409937, 0.9038356164383561], "iscrowd": 0, "area": 5544.156400000001, "rle": {"size": [365, 322], "counts": "n99R;2O1N3mNEQG<i7m0SHTOa7[1ZHhNe7\\1VHeNj7^1QHdNn7`1nGbNQ8a1kG`NU8\\2O1000000O10O10O10000O1000000O1N2N2N2O1N2N10gM^Hi0`7WOmH>Q7B]I2a6NlIGS68YJ^Oe5b0cJXO[5h0mJPOT5o0d2100O100O100O100O100O101N100O100]O_NZGb1e8cNWG^1f8gNVGZ1i8jNTGV1j8oNSGQ1m8h00O10O01O10O0100O3N6I7J6I7Jf`g2"}, "label": "black man"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white dress in the rain\"."}], "task": "refering", "answer_template": "The \"a woman in a white dress in the rain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 129, 130, 131, 132, 146, 147, 148, 149, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353, 366, 367, 368, 369, 370, 383, 384, 385, 386, 387], "bbox": [0.48133333333333334, 0.288671875, 0.8938124999999999, 0.9887656249999999], "iscrowd": 0, "area": 49668.874699999986, "rle": {"size": [640, 480], "counts": "Zi`43jc07J001O10000O0100O010O01O0P]OIZb07^]O2ab0f00010ON2N2O10U@DS<<ZC:c<FjBQ1R=nNiB[1S=fNgBc1V=]NdBl1Y=TNbBT2[=lM_B]2^=dM[Be2b=[MYBm2`;WLhEm0eNR3k:VLhD:^1>nNV3h:VMQFEUOZ3f:ZMlE]O\\O]3e:^MgEVOB`3d:aMbET4[:nKcEU4Z:lKeEW4k9fImEU26X4j9WLTFl3h9ULXFn3e9QL]FQ4_9oKdFS4X9lKkFU4NeIP8V2SHY5l7gJVHX5i7hJYHW5g7iJYHW5f7iJ\\HV5c7jJ_HU5`7kJbHT5]7lJdHT5[7lJgHS5X7mJjHR5V7PJkGnNP1R7U7lIQHoNl0T7S7kIRHQOm0S7Q7jISHTOc0[OYOf7P8kITHTOd0]OVOd7Q8jIUHVOd0^OUOb7Q8jIVHVOd0ASO_7U8fIUH[Od0APO^7S7\\HRJe0oN@8a0d0Y7Q6UHoH3l05@C1i0c0V7P6YIXIhNn0k0HV7P6`I\\J\\OAV7Q6dIYJXOCV7R6fIXJUODW7S6gIVJTOEV7S6iIUJTOEU7U6hIUJTOEU7U6iITJTOFS7T6kIUJTOER7U6lITJSOGQ7T6mITJTOFP7U6nISJTOEP7W6mITJSODQ7V6oITJROCQ7X6nITJSOBP7Y6nITJSOBP7X6oIUJk7j5VHVJj7i5WHVJj7j5VHUJk7k5UHTJl7k5UHUJk7k5UHTJl7k5UHTJl7l5SHTJn7l5RHSJo7l5RHTJn7k5VHQJk7m5ZHoIg7Q6T3O1O1O10000000000000000000000000000000000004L1O1O1O1O1O1O001O1O1O1O001O1O1O1O1O0SDaIY:c6bEcIY:`6dEeIR3Dh2i6TJiIm2Ei2d6PJVIXOa0o3Kc2`6mITJ`3A^2]6hIYJj3@a1I`JR8o3ZHo03PKb7R4bH9;dKS7R4lHkNl9V1YFZMR;f2VEbLZ;]3e20O0100O2O1N101N2O0O2N3N3L2O1N2O1N2O1N2O1N2O1O1N2O1O1O2M5L3M2N3L4M2N3L3N2N2N2M4M2N2N3L3N4L5K4K6K4L5J:E``o0"}, "label": "a woman in a white dress in the rain"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a light colored dress in the rain\"."}], "task": "refering", "answer_template": "The \"a woman in a light colored dress in the rain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 129, 130, 131, 132, 146, 147, 148, 149, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 246, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 331, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353, 366, 367, 368, 369, 370, 383, 384, 385, 386, 387], "bbox": [0.48133333333333334, 0.288671875, 0.8938124999999999, 0.9887656249999999], "iscrowd": 0, "area": 49668.874699999986, "rle": {"size": [640, 480], "counts": "Zi`43jc07J001O10000O0100O010O01O0P]OIZb07^]O2ab0f00010ON2N2O10U@DS<<ZC:c<FjBQ1R=nNiB[1S=fNgBc1V=]NdBl1Y=TNbBT2[=lM_B]2^=dM[Be2b=[MYBm2`;WLhEm0eNR3k:VLhD:^1>nNV3h:VMQFEUOZ3f:ZMlE]O\\O]3e:^MgEVOB`3d:aMbET4[:nKcEU4Z:lKeEW4k9fImEU26X4j9WLTFl3h9ULXFn3e9QL]FQ4_9oKdFS4X9lKkFU4NeIP8V2SHY5l7gJVHX5i7hJYHW5g7iJYHW5f7iJ\\HV5c7jJ_HU5`7kJbHT5]7lJdHT5[7lJgHS5X7mJjHR5V7PJkGnNP1R7U7lIQHoNl0T7S7kIRHQOm0S7Q7jISHTOc0[OYOf7P8kITHTOd0]OVOd7Q8jIUHVOd0^OUOb7Q8jIVHVOd0ASO_7U8fIUH[Od0APO^7S7\\HRJe0oN@8a0d0Y7Q6UHoH3l05@C1i0c0V7P6YIXIhNn0k0HV7P6`I\\J\\OAV7Q6dIYJXOCV7R6fIXJUODW7S6gIVJTOEV7S6iIUJTOEU7U6hIUJTOEU7U6iITJTOFS7T6kIUJTOER7U6lITJSOGQ7T6mITJTOFP7U6nISJTOEP7W6mITJSODQ7V6oITJROCQ7X6nITJSOBP7Y6nITJSOBP7X6oIUJk7j5VHVJj7i5WHVJj7j5VHUJk7k5UHTJl7k5UHUJk7k5UHTJl7k5UHTJl7l5SHTJn7l5RHSJo7l5RHTJn7k5VHQJk7m5ZHoIg7Q6T3O1O1O10000000000000000000000000000000000004L1O1O1O1O1O1O001O1O1O1O001O1O1O1O1O0SDaIY:c6bEcIY:`6dEeIR3Dh2i6TJiIm2Ei2d6PJVIXOa0o3Kc2`6mITJ`3A^2]6hIYJj3@a1I`JR8o3ZHo03PKb7R4bH9;dKS7R4lHkNl9V1YFZMR;f2VEbLZ;]3e20O0100O2O1N101N2O0O2N3N3L2O1N2O1N2O1N2O1N2O1O1N2O1O1O2M5L3M2N3L4M2N3L3N2N2N2M4M2N2N3L3N4L5K4K6K4L5J:E``o0"}, "label": "a woman in a light colored dress in the rain"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow jacket holding a piece of plastic\"."}], "task": "refering", "answer_template": "The \"man in yellow jacket holding a piece of plastic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 126, 127, 142, 143, 144, 145, 157, 158, 159, 160, 161, 162, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 377, 378, 379, 380, 381, 382], "bbox": [0.2106875, 0.27709375, 0.5838958333333333, 0.98815625], "iscrowd": 0, "area": 51748.39295, "rle": {"size": [640, 480], "counts": "dgo12hc06bFKkN;S1HfN>W1FcN?[1D^Nb0`1AYNe0e1^OTNh0j1[OPNj0n1YOkMm0S2VOfMP1W2TObMR1\\2QO^MT1`2oNYMW1e2lNTMZ1j2iNoL]1o2fNjL`1T3cNfLb1W3bNbLd1\\3_N]Lg1a3\\NXLj1f3YNTLl1k3VNnKP2Q4RNhKT2X4mMaKY2^4iM[K]2e4dMUKa2j4`MPKf2P5[MiJk2V5WMdJn2\\5RM`JR3_5SMYJQ3g5RMQJS3n5QMiIU3W6oLoHg3P7[LdHP4\\7QL_HT4_7nK[HW4e7jKUH[4j7gKQH]4o7dKkGa4T8aKfGd4Z8]KaGg4^8^300O100O100O1O100O100O10M3J6J5K6J6J6J5K6J6K5J5K6J6J6J5K6J6J6J5K]G^HT5\\7XJ^Ih5\\6eI]J[6]5QI]Kn6^4_H[La7_3kG[MU8X700O100O100O001O1O1O1O1O1O1O1O1O1O100O1E;QOn0N3N2M3M3M3O100O100O10000O10O010000O100O11O001O00001O001N10001O001O001J8C=C<PJnHdN_7X1eH[Ng7b1]HQNP8k1SHiMZ8S2jG`Mb8]2bGVMj8g2WGoLV9m2kFiLa9S3aFcLk9Z3UF]LX:_3iEWLc:f3^EPLn:m3SEiKZ;S4gDcKe;Z4[D`Kn;]4SDbKQ<Z4PDeKS<W4oCgKU<V4lCiKX<S4hCmK[<P4fCoK]<n3dCQL`<k3aCTLb<i3_CVLd<V1]C@OZOh<P1ZDgNoN8i<n0lEQOV:k0oEROS:k0XFjNk9R1d4M3L4M3L4M3M3L4M3L4Mhgl3"}, "label": "man in yellow jacket holding a piece of plastic"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow coat\"."}], "task": "refering", "answer_template": "The \"man in yellow coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 126, 127, 142, 143, 144, 145, 157, 158, 159, 160, 161, 162, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 377, 378, 379, 380, 381, 382], "bbox": [0.2106875, 0.27709375, 0.5838958333333333, 0.98815625], "iscrowd": 0, "area": 51748.39295, "rle": {"size": [640, 480], "counts": "dgo12hc06bFKkN;S1HfN>W1FcN?[1D^Nb0`1AYNe0e1^OTNh0j1[OPNj0n1YOkMm0S2VOfMP1W2TObMR1\\2QO^MT1`2oNYMW1e2lNTMZ1j2iNoL]1o2fNjL`1T3cNfLb1W3bNbLd1\\3_N]Lg1a3\\NXLj1f3YNTLl1k3VNnKP2Q4RNhKT2X4mMaKY2^4iM[K]2e4dMUKa2j4`MPKf2P5[MiJk2V5WMdJn2\\5RM`JR3_5SMYJQ3g5RMQJS3n5QMiIU3W6oLoHg3P7[LdHP4\\7QL_HT4_7nK[HW4e7jKUH[4j7gKQH]4o7dKkGa4T8aKfGd4Z8]KaGg4^8^300O100O100O1O100O100O10M3J6J5K6J6J6J5K6J6K5J5K6J6J6J5K6J6J6J5K]G^HT5\\7XJ^Ih5\\6eI]J[6]5QI]Kn6^4_H[La7_3kG[MU8X700O100O100O001O1O1O1O1O1O1O1O1O1O100O1E;QOn0N3N2M3M3M3O100O100O10000O10O010000O100O11O001O00001O001N10001O001O001J8C=C<PJnHdN_7X1eH[Ng7b1]HQNP8k1SHiMZ8S2jG`Mb8]2bGVMj8g2WGoLV9m2kFiLa9S3aFcLk9Z3UF]LX:_3iEWLc:f3^EPLn:m3SEiKZ;S4gDcKe;Z4[D`Kn;]4SDbKQ<Z4PDeKS<W4oCgKU<V4lCiKX<S4hCmK[<P4fCoK]<n3dCQL`<k3aCTLb<i3_CVLd<V1]C@OZOh<P1ZDgNoN8i<n0lEQOV:k0oEROS:k0XFjNk9R1d4M3L4M3L4M3M3L4M3L4Mhgl3"}, "label": "man in yellow coat"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear plastic umbrella\"."}], "task": "refering", "answer_template": "The \"clear plastic umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 70, 71, 72, 87, 88, 89, 104, 105, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 189, 190, 191, 193], "bbox": [0.10266666666666667, 0.14451562499999998, 0.39525, 0.5344375], "iscrowd": 0, "area": 19188.773250000002, "rle": {"size": [640, 480], "counts": "_mn01jc06F:G8H9F:G8H9F:G9G8G:ZMoLSDZ3k;hLkCb3S<`LdCh3[<ZL`Ck3^<WL^Cm3`<UL\\Cn3d<SLWCR4g<PLUCT4i<nKSCV4k<lKPCX4o<jKmBZ4Q=gKlB\\4T=eKgB`4W=bKeBa4Z=aKbBb4]=`K^Be4`=n0O2O0O2N1O2N1O4M0O00001O01O0001O0[CXIS<h6cCaI^<n63M2N2N2N2N210O01O0102M3N3L3N2N2M4M2M3N2M3N2M3N1N3N2M3N2N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2M4M2N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2N2M3J6I8H8M3N1O2No]OiNZa0T1e^OnN]a0o0b^OTO^a0`0o]OAb01aa06m^OLUa0KS_O4]b0001O0000O1N2M2O2N^be5"}, "label": "clear plastic umbrella"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear umbrella bent in the wind\"."}], "task": "refering", "answer_template": "The \"clear umbrella bent in the wind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 70, 71, 72, 87, 88, 89, 104, 105, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 189, 190, 191, 193], "bbox": [0.10266666666666667, 0.14451562499999998, 0.39525, 0.5344375], "iscrowd": 0, "area": 19188.773250000002, "rle": {"size": [640, 480], "counts": "_mn01jc06F:G8H9F:G8H9F:G9G8G:ZMoLSDZ3k;hLkCb3S<`LdCh3[<ZL`Ck3^<WL^Cm3`<UL\\Cn3d<SLWCR4g<PLUCT4i<nKSCV4k<lKPCX4o<jKmBZ4Q=gKlB\\4T=eKgB`4W=bKeBa4Z=aKbBb4]=`K^Be4`=n0O2O0O2N1O2N1O4M0O00001O01O0001O0[CXIS<h6cCaI^<n63M2N2N2N2N210O01O0102M3N3L3N2N2M4M2M3N2M3N2M3N1N3N2M3N2N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2M4M2N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2N2M3J6I8H8M3N1O2No]OiNZa0T1e^OnN]a0o0b^OTO^a0`0o]OAb01aa06m^OLUa0KS_O4]b0001O0000O1N2M2O2N^be5"}, "label": "clear umbrella bent in the wind"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear umbrella covered with black flowers\"."}], "task": "refering", "answer_template": "The \"a clear umbrella covered with black flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 127, 128, 129, 131, 132, 133, 145, 146, 149, 150, 166, 167, 179, 196, 197], "bbox": [0.4016458333333333, 0.13909375, 0.864625, 0.504359375], "iscrowd": 0, "area": 22216.14515, "rle": {"size": [640, 480], "counts": "khh31nc01O100O1O1O2N2N2N3M3M2N3M3M2O2M3M3M2N3M3M2N1O1O1O2O0O1O1O1O100O1O1O100O1O2N100O1O1O2N101N1O2N2O0O3M4L4M3L5K4M3L4M2M2O2M3N2M2O1N2O0O2O0O2O1N1O2O0O2O1N101N101N2O0O2jBhKb:X4VEQLh:Q4QEULo:k3mDZLQ;h3jD]LU;d3eDTL_NIj<T4cDSLhNIc<V4eDkKnNO\\<V4gDeKSO4U<Y4gD^KYO:o;Y4aEgK^:Z4bEeK_:\\4`EdK_:^4`EbK`:^4lCZKo08U;_4hCZKT18T;^4dC[K[17Q;k4PEUKQ;i4QEWKP;f4QE[Ko:m3hERLZ:l3iEPLZ:n3hEoK[:o3l2N2N2O1N2N2O2M2O1N2O1N2O1N2O1N3N1N2O1N200O10000O2O0O100O100000001N1000000O2O0000000O20O1O1O1O100O1O1O100O1O1O1O010O1O1O100O2O001N101O0d@lLk=U3RBlLo=T3oAmLR>S3kAoLV>R3gAoLZ>Q3dAPM]>P3`ARMa>o2\\ARMd>o2ZAQMh>o2VARMk>o2QASMP?m2n@TMS?l2k@UMV?j31O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1J5K6I9Hc0\\Oe0\\Oc0\\Oe0\\Oh]X1"}, "label": "a clear umbrella covered with black flowers"}]}
{"id": 12440, "image": "COCO_train2014_000000012440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear and flower umbrella the lady is holding\"."}], "task": "refering", "answer_template": "The \"a clear and flower umbrella the lady is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 127, 128, 129, 131, 132, 133, 145, 146, 149, 150, 166, 167, 179, 196, 197], "bbox": [0.4016458333333333, 0.13909375, 0.864625, 0.504359375], "iscrowd": 0, "area": 22216.14515, "rle": {"size": [640, 480], "counts": "khh31nc01O100O1O1O2N2N2N3M3M2N3M3M2O2M3M3M2N3M3M2N1O1O1O2O0O1O1O1O100O1O1O100O1O2N100O1O1O2N101N1O2N2O0O3M4L4M3L5K4M3L4M2M2O2M3N2M2O1N2O0O2O0O2O1N1O2O0O2O1N101N101N2O0O2jBhKb:X4VEQLh:Q4QEULo:k3mDZLQ;h3jD]LU;d3eDTL_NIj<T4cDSLhNIc<V4eDkKnNO\\<V4gDeKSO4U<Y4gD^KYO:o;Y4aEgK^:Z4bEeK_:\\4`EdK_:^4`EbK`:^4lCZKo08U;_4hCZKT18T;^4dC[K[17Q;k4PEUKQ;i4QEWKP;f4QE[Ko:m3hERLZ:l3iEPLZ:n3hEoK[:o3l2N2N2O1N2N2O2M2O1N2O1N2O1N2O1N3N1N2O1N200O10000O2O0O100O100000001N1000000O2O0000000O20O1O1O1O100O1O1O100O1O1O1O010O1O1O100O2O001N101O0d@lLk=U3RBlLo=T3oAmLR>S3kAoLV>R3gAoLZ>Q3dAPM]>P3`ARMa>o2\\ARMd>o2ZAQMh>o2VARMk>o2QASMP?m2n@TMS?l2k@UMV?j31O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1J5K6I9Hc0\\Oe0\\Oc0\\Oe0\\Oh]X1"}, "label": "a clear and flower umbrella the lady is holding"}]}
{"id": 315555, "image": "COCO_train2014_000000315555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman bottom right corner\"."}], "task": "refering", "answer_template": "The \"woman bottom right corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 229, 251, 252, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 341, 342, 343, 344], "bbox": [0.5937187500000001, 0.5996252927400468, 1.0, 0.9883840749414521], "iscrowd": 0, "area": 15261.358900000003, "rle": {"size": [427, 640], "counts": "hkn42U=4L4O10O1000O10000000O01000000O1000O0100O1000O010000O010O10O10O100O01000O10O010000O01000O10O0100O10O0100O00100O10O0100O10O0100O10O0100O10O0100O1O010O100O010O10O0100O10O0100O10O0100O10O01O100O01000000O10O10O1000000O10O1000O1000000O0100O10O01O100O001O10O01O100O00100O1O00100O1O010O1O10O01O100O001O100O00100O1O01000nDjNn9V1mEPOS:P1hEUOX:k0`E]O_:h10000O10000O10000O100O01000O10000O1000000O100000000O100000000O100000O10O10000N2N2N2N2N2N2N2N2N2I7B>N2N2M3N2N2N2N2N1O2M3N2N2N2O10000O10000O10000O10000O100O1000O2O1O1O1M2O2N2N2N2N2]MUGc0l8YO[Gc0g8XO`Gd0b8VOfGg0[8WOiGf0X8XOmGd0S8[OQHb0P8\\OTHa0m7]OWHa0i7]O`H;Q8SOdH1cN"}, "label": "woman bottom right corner"}]}
{"id": 315555, "image": "COCO_train2014_000000315555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in the water with a ponytail\"."}], "task": "refering", "answer_template": "The \"a woman in the water with a ponytail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 229, 251, 252, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 341, 342, 343, 344], "bbox": [0.5937187500000001, 0.5996252927400468, 1.0, 0.9883840749414521], "iscrowd": 0, "area": 15261.358900000003, "rle": {"size": [427, 640], "counts": "hkn42U=4L4O10O1000O10000000O01000000O1000O0100O1000O010000O010O10O10O100O01000O10O010000O01000O10O0100O10O0100O00100O10O0100O10O0100O10O0100O10O0100O1O010O100O010O10O0100O10O0100O10O0100O10O01O100O01000000O10O10O1000000O10O1000O1000000O0100O10O01O100O001O10O01O100O00100O1O00100O1O010O1O10O01O100O001O100O00100O1O01000nDjNn9V1mEPOS:P1hEUOX:k0`E]O_:h10000O10000O10000O100O01000O10000O1000000O100000000O100000000O100000O10O10000N2N2N2N2N2N2N2N2N2I7B>N2N2M3N2N2N2N2N1O2M3N2N2N2O10000O10000O10000O10000O100O1000O2O1O1O1M2O2N2N2N2N2]MUGc0l8YO[Gc0g8XO`Gd0b8VOfGg0[8WOiGf0X8XOmGd0S8[OQHb0P8\\OTHa0m7]OWHa0i7]O`H;Q8SOdH1cN"}, "label": "a woman in the water with a ponytail"}]}
{"id": 315555, "image": "COCO_train2014_000000315555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the water between the two girls\"."}], "task": "refering", "answer_template": "The \"the boy in the water between the two girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 243, 244, 245, 246, 247, 266, 267, 268, 270, 271, 294], "bbox": [0.4423125, 0.4366042154566745, 0.8292187500000001, 0.8108899297423887], "iscrowd": 0, "area": 11406.224499999998, "rle": {"size": [427, 640], "counts": "ZXf31Z=2N2N2N2N2N2N2N0O101O000000000000000O101O000000O1000000O0100000O1000000O1000O10O1000000O1000000O0100000O1000O1O10O01O100O010O1O10O0100O1O010O100O00100O10O01O102M2O2eESOS8o0kGROU8P1hGQOX8Q1dGQO\\8Q1aGPO^8^1TGcNl8o1`FSN`9Q2[FPNd9U2WFlMi9W2RFkMn9Y2mEhMR:i20000O101O0O1UOPM[GP3a8YMYGh2f8\\MUGf2j8n0O1O1O1O1O100O100O100O1O10000000000000000000001O00000O1000fL]Gl1d8RN`Gk1`8TNbGk1_8SNcGl1^8RNdGm1\\8RNeGn1\\8oMgGP2Z8mMiGR2X8jMlGU2Z8cMhG]2`8TMhGk2]9O00001O0nNkEnNU:l0UFPOl9i0_FROa9h0iFTOX9e0SGVOm8h0YGTOg8k0n1O0100O1O100O1O1001O1O2O0O2N1OO10000O1000000O1001O1O001O1O1O1O1O00000000001O0000000000001O0001O000001O00000000001O00000000001O0000000O10001O0O010O001O001J5O20O100O10O0O2O001O001N101O001OUa]1"}, "label": "the boy in the water between the two girls"}]}
{"id": 315555, "image": "COCO_train2014_000000315555.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with black hair training dogs\"."}], "task": "refering", "answer_template": "The \"a women with black hair training dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 243, 244, 245, 246, 247, 266, 267, 268, 270, 271, 294], "bbox": [0.4423125, 0.4366042154566745, 0.8292187500000001, 0.8108899297423887], "iscrowd": 0, "area": 11406.224499999998, "rle": {"size": [427, 640], "counts": "ZXf31Z=2N2N2N2N2N2N2N0O101O000000000000000O101O000000O1000000O0100000O1000000O1000O10O1000000O1000000O0100000O1000O1O10O01O100O010O1O10O0100O1O010O100O00100O10O01O102M2O2eESOS8o0kGROU8P1hGQOX8Q1dGQO\\8Q1aGPO^8^1TGcNl8o1`FSN`9Q2[FPNd9U2WFlMi9W2RFkMn9Y2mEhMR:i20000O101O0O1UOPM[GP3a8YMYGh2f8\\MUGf2j8n0O1O1O1O1O100O100O100O1O10000000000000000000001O00000O1000fL]Gl1d8RN`Gk1`8TNbGk1_8SNcGl1^8RNdGm1\\8RNeGn1\\8oMgGP2Z8mMiGR2X8jMlGU2Z8cMhG]2`8TMhGk2]9O00001O0nNkEnNU:l0UFPOl9i0_FROa9h0iFTOX9e0SGVOm8h0YGTOg8k0n1O0100O1O100O1O1001O1O2O0O2N1OO10000O1000000O1001O1O001O1O1O1O1O00000000001O0000000000001O0001O000001O00000000001O00000000001O0000000O10001O0O010O001O001J5O20O100O10O0O2O001O001N101O001OUa]1"}, "label": "a women with black hair training dogs"}]}
{"id": 135332, "image": "COCO_train2014_000000135332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wrought iron chair with a beige striped seat faces a sliding glass window\"."}], "task": "refering", "answer_template": "The \"a wrought iron chair with a beige striped seat faces a sliding glass window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380, 381, 382], "bbox": [0.4023125, 0.5333333333333333, 0.6427343750000001, 0.9871069182389938], "iscrowd": 0, "area": 20230.447999999997, "rle": {"size": [477, 640], "counts": "WPh31g>6N2N8H9G8H4M3L3M4L3M3M2M4hFZNS5j1cJ]N[5h1[J`Nb5e1TJcNh5c1nIeNP6_1eIiN[6Z1[ImNd6]1lHjNT7_4O101N2O010O01000O010O11O0O1000001O000000001O001O001O00001O001O0010O0001O001O010O001O0[JdGi4_8QKdGo4V90O010O0100O001O010O0010O0100O01O1O000000000000001O0000RLQGn1P9jMWGV2l8cMYG\\2i8^M[Gb2g8XM^Gg2\\:01O001O001O001O001O001O001O001O001O001O001O001O001O001O0000000N3M2O1N2N2N3M2N3M2N31O4K6K3L02M4M4K3N1N1O2O0O2Ia0^OjXZ3"}, "label": "a wrought iron chair with a beige striped seat faces a sliding glass window"}]}
{"id": 135332, "image": "COCO_train2014_000000135332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lovely metal backed chair with a white cushion faces out a sliding glass door\"."}], "task": "refering", "answer_template": "The \"a lovely metal backed chair with a white cushion faces out a sliding glass door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380, 381, 382], "bbox": [0.4023125, 0.5333333333333333, 0.6427343750000001, 0.9871069182389938], "iscrowd": 0, "area": 20230.447999999997, "rle": {"size": [477, 640], "counts": "WPh31g>6N2N8H9G8H4M3L3M4L3M3M2M4hFZNS5j1cJ]N[5h1[J`Nb5e1TJcNh5c1nIeNP6_1eIiN[6Z1[ImNd6]1lHjNT7_4O101N2O010O01000O010O11O0O1000001O000000001O001O001O00001O001O0010O0001O001O010O001O0[JdGi4_8QKdGo4V90O010O0100O001O010O0010O0100O01O1O000000000000001O0000RLQGn1P9jMWGV2l8cMYG\\2i8^M[Gb2g8XM^Gg2\\:01O001O001O001O001O001O001O001O001O001O001O001O001O001O0000000N3M2O1N2N2N3M2N3M2N31O4K6K3L02M4M4K3N1N1O2O0O2Ia0^OjXZ3"}, "label": "a lovely metal backed chair with a white cushion faces out a sliding glass door"}]}
{"id": 135332, "image": "COCO_train2014_000000135332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair facing the kitchen\"."}], "task": "refering", "answer_template": "The \"chair facing the kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 249, 250, 272, 273, 293, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387, 388], "bbox": [0.6687031250000001, 0.5548008385744234, 0.90490625, 1.0], "iscrowd": 0, "area": 16058.335799999992, "rle": {"size": [477, 640], "counts": "bhW6d0P>X1iNh0[O5O1O100O1O1O100O1O100O1O1O100O1O1O2O0O1O100O1O1O100O1O100O10000000000000000000000000000000000000000000001O00000000000000000001O00001O00001O00001O00001O00001O00001O00001O00001O01O01O0000001O00001O00001O00001OO1L4L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L4L5K4L]MTG]Oh8b0bGYO]8c0mGXOR8d0XHWOf7f0dHUO[7g0oHTOP7h0ZISOe6i0eIROY6k0QJPOn5l0\\JoNc5m0gJnNX5n0RKmNl4P1^KkNa4Q1iKjNV4R1TLiNk3S1_LhN_3U1V5L4L4L3M6J<DmTl0"}, "label": "chair facing the kitchen"}]}
{"id": 135332, "image": "COCO_train2014_000000135332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair furthest from the counter\"."}], "task": "refering", "answer_template": "The \"chair furthest from the counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 249, 250, 272, 273, 293, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387, 388], "bbox": [0.6687031250000001, 0.5548008385744234, 0.90490625, 1.0], "iscrowd": 0, "area": 16058.335799999992, "rle": {"size": [477, 640], "counts": "bhW6d0P>X1iNh0[O5O1O100O1O1O100O1O100O1O1O100O1O1O2O0O1O100O1O1O100O1O100O10000000000000000000000000000000000000000000001O00000000000000000001O00001O00001O00001O00001O00001O00001O00001O00001O01O01O0000001O00001O00001O00001OO1L4L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L4L5K4L]MTG]Oh8b0bGYO]8c0mGXOR8d0XHWOf7f0dHUO[7g0oHTOP7h0ZISOe6i0eIROY6k0QJPOn5l0\\JoNc5m0gJnNX5n0RKmNl4P1^KkNa4Q1iKjNV4R1TLiNk3S1_LhN_3U1V5L4L4L3M6J<DmTl0"}, "label": "chair furthest from the counter"}]}
{"id": 430244, "image": "COCO_train2014_000000430244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue , striped shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue , striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 88, 89, 90, 111, 112, 113, 134, 135, 136, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319], "bbox": [0.756734375, 0.16629166666666664, 0.967421875, 0.8179791666666667], "iscrowd": 0, "area": 25118.0905, "rle": {"size": [480, 640], "counts": "acS74j>3L3M4M3L4M2M4L4M2M4M3L4L3N3L4M3L3M4M3L3N3L4M3L3N3N2fFYMW6i2hIWMX6k2bIZM]6h2YHaNf7a1PGGP9U3O1O1O0O2O1O001ZOSJXHn5e7g0O2M3N1N3N2M2O2N2M2POeH[J]7b5gHZJ\\7d5gHYJZ7e5R1M3N1N3N2M2N3O10O10O10O01000O0100O01000O01000O010O10O10O10O10O10O101O1N101O0O2O1O0O2O1O0O2O001E;A>A`0@`0G8N3N1N3N2M2O2]JZGh4i8UK[Gh4f8VK^Gf4e8WK_Gf4b8XKcGc4`8ZKdGc4^8ZKfGb4\\8\\KhGa4Z8\\KjG`4_9L3L5H9^Ob0^Oc0F9G9F:G:E:G9G:E:GUe9"}, "label": "a man wearing a blue , striped shirt"}]}
{"id": 430244, "image": "COCO_train2014_000000430244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue stripe shirt watching another person playing wii sports\"."}], "task": "refering", "answer_template": "The \"a man in a blue stripe shirt watching another person playing wii sports\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 88, 89, 90, 111, 112, 113, 134, 135, 136, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319], "bbox": [0.756734375, 0.16629166666666664, 0.967421875, 0.8179791666666667], "iscrowd": 0, "area": 25118.0905, "rle": {"size": [480, 640], "counts": "acS74j>3L3M4M3L4M2M4L4M2M4M3L4L3N3L4M3L3M4M3L3N3L4M3L3N3N2fFYMW6i2hIWMX6k2bIZM]6h2YHaNf7a1PGGP9U3O1O1O0O2O1O001ZOSJXHn5e7g0O2M3N1N3N2M2O2N2M2POeH[J]7b5gHZJ\\7d5gHYJZ7e5R1M3N1N3N2M2N3O10O10O10O01000O0100O01000O01000O010O10O10O10O10O10O101O1N101O0O2O1O0O2O1O0O2O001E;A>A`0@`0G8N3N1N3N2M2O2]JZGh4i8UK[Gh4f8VK^Gf4e8WK_Gf4b8XKcGc4`8ZKdGc4^8ZKfGb4\\8\\KhGa4Z8\\KjG`4_9L3L5H9^Ob0^Oc0F9G9F:G:E:G9G:E:GUe9"}, "label": "a man in a blue stripe shirt watching another person playing wii sports"}]}
{"id": 553126, "image": "COCO_train2014_000000553126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cell phone sits on a key board\"."}], "task": "refering", "answer_template": "The \"a black cell phone sits on a key board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 133, 134, 135, 136, 137, 138, 139, 140, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.38733999999999996, 0.4005555555555556, 0.9729800000000001, 0.9713450292397662], "iscrowd": 0, "area": 32819.3725, "rle": {"size": [342, 500], "counts": "mnP21d:3N1N3N1N3N1N3N2M2O2M2O2M2O2N1N3N1N3N1N3N1N3N2M2O2M2O2M2O2N1N3N1N3N1N3N1N3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M2O2M2O2M2O2M2O1O0O2O00001N101O001O00001O001O0O2O001O00001O001O001N10001O001O001O001O000O2O001O001O0100O1O100O010O100O1O010O100O10O0100O1O100O010O100O1O010O100O10O0100O1O10O0100O100O1O010O100O10O0100O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O10O01O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001N2O1N2N101N2O2M4M2M3N3L3N3K4L5K4K6K4K6K4LX\\4"}, "label": "a black cell phone sits on a key board"}]}
{"id": 553126, "image": "COCO_train2014_000000553126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cell phone that is turned off ( nokia )\"."}], "task": "refering", "answer_template": "The \"cell phone that is turned off ( nokia )\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 133, 134, 135, 136, 137, 138, 139, 140, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.38733999999999996, 0.4005555555555556, 0.9729800000000001, 0.9713450292397662], "iscrowd": 0, "area": 32819.3725, "rle": {"size": [342, 500], "counts": "mnP21d:3N1N3N1N3N1N3N2M2O2M2O2M2O2N1N3N1N3N1N3N1N3N2M2O2M2O2M2O2N1N3N1N3N1N3N1N3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M2O2M2O2M2O2M2O1O0O2O00001N101O001O00001O001O0O2O001O00001O001O001N10001O001O001O001O000O2O001O001O0100O1O100O010O100O1O010O100O10O0100O1O100O010O100O1O010O100O10O0100O1O10O0100O100O1O010O100O10O0100O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O10O01O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001N2O1N2N101N2O2M4M2M3N3L3N3K4L5K4K6K4K6K4LX\\4"}, "label": "cell phone that is turned off ( nokia )"}]}
{"id": 553126, "image": "COCO_train2014_000000553126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"nokia phone which is completely kept on the laptop keyboard\"."}], "task": "refering", "answer_template": "The \"nokia phone which is completely kept on the laptop keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 96, 97, 98, 99, 114, 115, 131, 132, 133, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 203, 204, 205, 206], "bbox": [0.26838, 0.37353801169590645, 0.60732, 0.949766081871345], "iscrowd": 0, "area": 15122.937550000002, "rle": {"size": [342, 500], "counts": "dP]12b:8H8H8H6J3M4L3M3M3L4M4L3M3M3M4L3M3M3M3M4L2N3M2N2N2N2N2N2N2N3M2N2N2M3N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2M3N2N2N2NdKiJ]3U5dLnJZ3P5hLQKX3l4iLWKU3g4lL[KS3c4oL_Ko2`4QMbKn2]4SMeKk2[4TMgKk2X4VMjKh2V4WMlKi2S4WMoKg2P4YMRLf2n3YMULe2j3\\MWLc2i3\\MZLb2e3_M\\L`2d3_M_L_2`3bMaL^2^3aMeL]2Z3cMhL\\2X3dMjLZ2U3fMmLY2S3gMoLW2Q3hMQMW2o2iMSMU2m2jMUMU2k2kMWMT2h2kMZMT2f2kM]MS2c2mM^MQ2c2nM`MP2`2PNaMo1_2PNdMm1]2SNeMk1[2TNgMk1Y2TNjMj1V2VNkMh1V2WNmMg1S2YNnMf1R2YNQNd1P2\\NQNc1o1\\NTNb1l1]NVNb1j1^NXN_1i1`NYN_1g1aN[N]1e1bN]N\\1d1dN^NZ1b1eN`NZ1`1fNbNX1^1gNdNW1]1hNfNV1Z1jNgNU1Y1jNjNS1W1mNjNR1V1mNmNQ1S1oNnNo0S1QOnNn0R1ROoNm0Q1ROROl0n0TOSOj0n0VOSOi0m0WOTOh0l0XOUOf0l0YOWOe0i0[OXOc0i0]OXOa0i0_OXO?i0@YO>h0BZO;g0EZO9g0GZO7g0H[O6f0J\\O3e0M\\O1e0O\\OOe01\\OMe02]OLd04^OIc07^OGc09^ODd0;l4000000O2O0000000O1000000O1000000O100000000O102N2NYeQ2"}, "label": "nokia phone which is completely kept on the laptop keyboard"}]}
{"id": 553126, "image": "COCO_train2014_000000553126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the phone that is stacked on the bottom\"."}], "task": "refering", "answer_template": "The \"the phone that is stacked on the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 96, 97, 98, 99, 114, 115, 131, 132, 133, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 203, 204, 205, 206], "bbox": [0.26838, 0.37353801169590645, 0.60732, 0.949766081871345], "iscrowd": 0, "area": 15122.937550000002, "rle": {"size": [342, 500], "counts": "dP]12b:8H8H8H6J3M4L3M3M3L4M4L3M3M3M4L3M3M3M3M4L2N3M2N2N2N2N2N2N2N3M2N2N2M3N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2M3N2N2N2NdKiJ]3U5dLnJZ3P5hLQKX3l4iLWKU3g4lL[KS3c4oL_Ko2`4QMbKn2]4SMeKk2[4TMgKk2X4VMjKh2V4WMlKi2S4WMoKg2P4YMRLf2n3YMULe2j3\\MWLc2i3\\MZLb2e3_M\\L`2d3_M_L_2`3bMaL^2^3aMeL]2Z3cMhL\\2X3dMjLZ2U3fMmLY2S3gMoLW2Q3hMQMW2o2iMSMU2m2jMUMU2k2kMWMT2h2kMZMT2f2kM]MS2c2mM^MQ2c2nM`MP2`2PNaMo1_2PNdMm1]2SNeMk1[2TNgMk1Y2TNjMj1V2VNkMh1V2WNmMg1S2YNnMf1R2YNQNd1P2\\NQNc1o1\\NTNb1l1]NVNb1j1^NXN_1i1`NYN_1g1aN[N]1e1bN]N\\1d1dN^NZ1b1eN`NZ1`1fNbNX1^1gNdNW1]1hNfNV1Z1jNgNU1Y1jNjNS1W1mNjNR1V1mNmNQ1S1oNnNo0S1QOnNn0R1ROoNm0Q1ROROl0n0TOSOj0n0VOSOi0m0WOTOh0l0XOUOf0l0YOWOe0i0[OXOc0i0]OXOa0i0_OXO?i0@YO>h0BZO;g0EZO9g0GZO7g0H[O6f0J\\O3e0M\\O1e0O\\OOe01\\OMe02]OLd04^OIc07^OGc09^ODd0;l4000000O2O0000000O1000000O1000000O100000000O102N2NYeQ2"}, "label": "the phone that is stacked on the bottom"}]}
{"id": 233642, "image": "COCO_train2014_000000233642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the oven the grandma and little girl are buttering bread in\"."}], "task": "refering", "answer_template": "The \"the oven the grandma and little girl are buttering bread in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.49887499999999996, 0.496625, 0.77190625, 0.9955], "iscrowd": 0, "area": 35715.473150000005, "rle": {"size": [480, 640], "counts": "nme4P2d7\\500000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000O100000000000000000000000000000000000000000000000000000000000000QM_H_Na7_1aHaN_7]1dHbN\\7\\1fHdNZ7Z1hHfNX7Y1iHgNW7Y1iHgNW7Y1iHgNW7X1kHgNU7Y1kHgNT7Z1lHfNT7Z1lHfNT7Z1lHfNT7Z1kHgNU7X1lHhNT7X1kHiNU7W1kHiNU7W1jHjNV7V1jHjNV7U1jHlNV7T1jHlNV7T1iHmNW7S1iHmNW7S1hHnNX7Q1iHoNW7Q1iHoNW7Q1iHoNW7Q1iHoNW7Q1iHoNW7Q1iHoNW7P1jHPOV7P1jHPOV7P1jHPOV7P1jHPOV7P1jHPOV7o0kHQOU7o0kHQOU7o0kHQOU7o0kHQOU7o0kHQOU7n0lHROT7n0lHROT7n0lHROT7n0lHROT7n0lHROT7n0lHROT7m0mHSOS7m0mHSOS7m0mHSOS7m0mHSOS7m0mHSOS7l0nHTOR7l0nHTOR7l0nHTOQ7m0oHSOQ7m0oHSOQ7l0PITOP7l0PIZLAa2_7U1PIYLD`2i7j0cHfLG]2Y8:PHXMJ\\2h8J^GiML\\2P9@UGTNL[2P9@TGTNN[2o8@SGUN0Y2n8ARGUN2Y2m8AQGVN6U2j8CQGWNd0h1\\80PGXNQ1[1P8<QIWOP7h0RIVOo6i0SIUOn6i0TIVOm6i0TIVOm6i0TIVOl6j0UIUOl6j0UIUOl6j0UIUOl6i0VIVOk6f0YIYOh6=bIB_64kIKV6JUJ5l5A^J>S:O1O1O1O2N1O[oS2"}, "label": "the oven the grandma and little girl are buttering bread in"}]}
{"id": 233642, "image": "COCO_train2014_000000233642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open oven with food inside\"."}], "task": "refering", "answer_template": "The \"an open oven with food inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.49887499999999996, 0.496625, 0.77190625, 0.9955], "iscrowd": 0, "area": 35715.473150000005, "rle": {"size": [480, 640], "counts": "nme4P2d7\\500000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000O100000000000000000000000000000000000000000000000000000000000000QM_H_Na7_1aHaN_7]1dHbN\\7\\1fHdNZ7Z1hHfNX7Y1iHgNW7Y1iHgNW7Y1iHgNW7X1kHgNU7Y1kHgNT7Z1lHfNT7Z1lHfNT7Z1lHfNT7Z1kHgNU7X1lHhNT7X1kHiNU7W1kHiNU7W1jHjNV7V1jHjNV7U1jHlNV7T1jHlNV7T1iHmNW7S1iHmNW7S1hHnNX7Q1iHoNW7Q1iHoNW7Q1iHoNW7Q1iHoNW7Q1iHoNW7Q1iHoNW7P1jHPOV7P1jHPOV7P1jHPOV7P1jHPOV7P1jHPOV7o0kHQOU7o0kHQOU7o0kHQOU7o0kHQOU7o0kHQOU7n0lHROT7n0lHROT7n0lHROT7n0lHROT7n0lHROT7n0lHROT7m0mHSOS7m0mHSOS7m0mHSOS7m0mHSOS7m0mHSOS7l0nHTOR7l0nHTOR7l0nHTOQ7m0oHSOQ7m0oHSOQ7l0PITOP7l0PIZLAa2_7U1PIYLD`2i7j0cHfLG]2Y8:PHXMJ\\2h8J^GiML\\2P9@UGTNL[2P9@TGTNN[2o8@SGUN0Y2n8ARGUN2Y2m8AQGVN6U2j8CQGWNd0h1\\80PGXNQ1[1P8<QIWOP7h0RIVOo6i0SIUOn6i0TIVOm6i0TIVOm6i0TIVOl6j0UIUOl6j0UIUOl6j0UIUOl6i0VIVOk6f0YIYOh6=bIB_64kIKV6JUJ5l5A^J>S:O1O1O1O2N1O[oS2"}, "label": "an open oven with food inside"}]}
{"id": 340139, "image": "COCO_train2014_000000340139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man dressed in white tennis gear with his arm behind him\"."}], "task": "refering", "answer_template": "The \"a man dressed in white tennis gear with his arm behind him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 67, 68, 69, 84, 85, 86, 101, 102, 103, 104, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 191, 192, 193, 194, 209, 210, 212, 213, 227, 228, 230], "bbox": [0.53518, 0.16681462140992168, 0.85876, 0.905822454308094], "iscrowd": 0, "area": 19770.218400000005, "rle": {"size": [383, 500], "counts": "WnU31jk10S`N1O2N1O1O1O1O1O001O1O1N101O1O1O001N2O1O0O2O1O0O2O001N10001O001O000WITO`2m0^MUOa2k0^MWOa2j0]MXOb2h0]MZOc2e0\\M^Ob2c0\\M_Od2`0[MBd2>ZMEf2:YMHf28YMJg25XMMg23XMOh20VM3h2NWM4h2LWM6g2KUM:i2GoLb0o2_OiLj0U3WOZKETO^1`5nNXK_2f4bMXKa2f4`MYK`2g4`MYK_2i4aMVK_2j4bMTK_2l4bMhJi2X5XMfJj2Y5WMeJk2[5UMdJk2\\5VMRJRO0n3n5f1O10O01000N2O2M2O1N3N1O2N1O2N1N3N1O2N1O2N1O110O0100O10OSOnJnJS5P5SKlJm4R5XKjJi4T5\\KhJd4V5aKgJ`4W5cKgJ_4W5cKgJ`4V5bKhJa4U5bKgJb4U5V1N3M3M2M4M2N3N1O2N1N2O2N1iL^Ho1d7nM]HR2d7kM^HU2c7hM_HX2c7eM^H[2c7bM_H^2c7]M`Hd2`7XMcHh2^7TMfHk2g4mLTNS3i1RMUNn2h1WMVNi2g1\\MWNd2f1aMXN_2f1fMWNZ2eNnL\\1i0KY2hNPM_1h0EX2jNTMa1f0AW2kNWMe1c0\\OW2lNYMj1a0VOV2nN^Mk1LnLKU2l2oNaM=\\OF6nN2m1l2oNdM7BNGRO;e1i2QOQNI]OP1L1k2SO`NQOSOh1FNk2VODi0dMLl2XODi0dMIl2[ODi0cMGl2^OEh0bMEm2@Eg0cMCk2BIg0_MBR3YOGR1\\M^OZ7?iH\\O\\7`0iHYO\\7d0S2M2K6J5JZZj0"}, "label": "a man dressed in white tennis gear with his arm behind him"}]}
{"id": 340139, "image": "COCO_train2014_000000340139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 67, 68, 69, 84, 85, 86, 101, 102, 103, 104, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 191, 192, 193, 194, 209, 210, 212, 213, 227, 228, 230], "bbox": [0.53518, 0.16681462140992168, 0.85876, 0.905822454308094], "iscrowd": 0, "area": 19770.218400000005, "rle": {"size": [383, 500], "counts": "WnU31jk10S`N1O2N1O1O1O1O1O001O1O1N101O1O1O001N2O1O0O2O1O0O2O001N10001O001O000WITO`2m0^MUOa2k0^MWOa2j0]MXOb2h0]MZOc2e0\\M^Ob2c0\\M_Od2`0[MBd2>ZMEf2:YMHf28YMJg25XMMg23XMOh20VM3h2NWM4h2LWM6g2KUM:i2GoLb0o2_OiLj0U3WOZKETO^1`5nNXK_2f4bMXKa2f4`MYK`2g4`MYK_2i4aMVK_2j4bMTK_2l4bMhJi2X5XMfJj2Y5WMeJk2[5UMdJk2\\5VMRJRO0n3n5f1O10O01000N2O2M2O1N3N1O2N1O2N1N3N1O2N1O2N1O110O0100O10OSOnJnJS5P5SKlJm4R5XKjJi4T5\\KhJd4V5aKgJ`4W5cKgJ_4W5cKgJ`4V5bKhJa4U5bKgJb4U5V1N3M3M2M4M2N3N1O2N1N2O2N1iL^Ho1d7nM]HR2d7kM^HU2c7hM_HX2c7eM^H[2c7bM_H^2c7]M`Hd2`7XMcHh2^7TMfHk2g4mLTNS3i1RMUNn2h1WMVNi2g1\\MWNd2f1aMXN_2f1fMWNZ2eNnL\\1i0KY2hNPM_1h0EX2jNTMa1f0AW2kNWMe1c0\\OW2lNYMj1a0VOV2nN^Mk1LnLKU2l2oNaM=\\OF6nN2m1l2oNdM7BNGRO;e1i2QOQNI]OP1L1k2SO`NQOSOh1FNk2VODi0dMLl2XODi0dMIl2[ODi0cMGl2^OEh0bMEm2@Eg0cMCk2BIg0_MBR3YOGR1\\M^OZ7?iH\\O\\7`0iHYO\\7d0S2M2K6J5JZZj0"}, "label": "man"}]}
{"id": 340139, "image": "COCO_train2014_000000340139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman dressed in white , wearing a visor , and swinging a tennis racket\"."}], "task": "refering", "answer_template": "The \"a woman dressed in white , wearing a visor , and swinging a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 21, 22, 23, 24, 38, 39, 40, 41, 42, 56, 57, 58, 59, 60, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 170, 171, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.0, 0.031462140992167105, 0.5164, 0.9123498694516972], "iscrowd": 0, "area": 42607.327999999994, "rle": {"size": [383, 500], "counts": "n9d0V;`0@5K5K5O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2N2O1N2N2O1N2N2O1N2N2N2O1N2K5L4K5L4K5L4K5G9H8H8G9H8cNnJdK[5Z4V1M3M3N2M3M3N2O1O1O1O1N2O1O1O1O1O1O1ZObHjL_7U3lH_LV7`3e0O1O1O1O1O1O1O1O1O1O100N2O1O1N2O1N2O1QOVGgNl8Y1XGbNi8]1]G\\Ne8c1`GXNa8h1dGRN]8m1k0O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O1000000000000000000000000O100000000000000000000001O001O1mHbNc3_1[LcNd3^1ZLdNe3]1YLeNf3\\1WLgNh3Z1VLhNi3Y1ULiNj3X1SLkNl3V1oKPOo3R1lKROQ3YOYKg1a1TOm2_O`K_1]1WOk2CfKX1Z1YOd2KQLn0V1[OX29`L>T1\\Oo1d0kL2T1[Oi1k0RMKT1[Ob1S1XMBV1\\O]1X1[M\\O<PNJ]1\\2[1\\MYO=PNK\\1e0dN1k2QOUO=QNL]1E@n0R2ROPO?QNLn2?6TOkN`0RNMm2<:VOhN`0QNNm29?WOcNa0RNOl27c0WO_Nc0RNOl27d0VO^Nc0SN0k28o2GWJ0k28o2FWJ3i27Q3EVJ4i28P3CZJ3f2:Q3BSN?m1_OTNa0m1^OSNb0n1\\OTNd0l1YOVNg0k1VOWNk0h1TOYNl0]60010O01O00001O010O00001O001O01O01O001O00010O001O000010O01O00001O001O00001O001O00001O001O00001O000001O01O000000001O0000000002O1N2N2N2N1O2O1N2N2M3@jYj2"}, "label": "a woman dressed in white , wearing a visor , and swinging a tennis racket"}]}
{"id": 340139, "image": "COCO_train2014_000000340139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white playing tennis\"."}], "task": "refering", "answer_template": "The \"a woman in white playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 21, 22, 23, 24, 38, 39, 40, 41, 42, 56, 57, 58, 59, 60, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 170, 171, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.0, 0.031462140992167105, 0.5164, 0.9123498694516972], "iscrowd": 0, "area": 42607.327999999994, "rle": {"size": [383, 500], "counts": "n9d0V;`0@5K5K5O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2N2O1N2N2O1N2N2O1N2N2N2O1N2K5L4K5L4K5L4K5G9H8H8G9H8cNnJdK[5Z4V1M3M3N2M3M3N2O1O1O1O1N2O1O1O1O1O1O1ZObHjL_7U3lH_LV7`3e0O1O1O1O1O1O1O1O1O1O100N2O1O1N2O1N2O1QOVGgNl8Y1XGbNi8]1]G\\Ne8c1`GXNa8h1dGRN]8m1k0O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O1000000000000000000000000O100000000000000000000001O001O1mHbNc3_1[LcNd3^1ZLdNe3]1YLeNf3\\1WLgNh3Z1VLhNi3Y1ULiNj3X1SLkNl3V1oKPOo3R1lKROQ3YOYKg1a1TOm2_O`K_1]1WOk2CfKX1Z1YOd2KQLn0V1[OX29`L>T1\\Oo1d0kL2T1[Oi1k0RMKT1[Ob1S1XMBV1\\O]1X1[M\\O<PNJ]1\\2[1\\MYO=PNK\\1e0dN1k2QOUO=QNL]1E@n0R2ROPO?QNLn2?6TOkN`0RNMm2<:VOhN`0QNNm29?WOcNa0RNOl27c0WO_Nc0RNOl27d0VO^Nc0SN0k28o2GWJ0k28o2FWJ3i27Q3EVJ4i28P3CZJ3f2:Q3BSN?m1_OTNa0m1^OSNb0n1\\OTNd0l1YOVNg0k1VOWNk0h1TOYNl0]60010O01O00001O010O00001O001O01O01O001O00010O001O000010O01O00001O001O00001O001O00001O001O00001O000001O01O000000001O0000000002O1N2N2N2N1O2O1N2N2M3@jYj2"}, "label": "a woman in white playing tennis"}]}
{"id": 413869, "image": "COCO_train2014_000000413869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and green airplane hovers above the runway\"."}], "task": "refering", "answer_template": "The \"a white and green airplane hovers above the runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 48, 49, 50, 71, 72, 73, 74, 84, 85, 86, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 180], "bbox": [0.012921875, 0.1191166077738516, 0.9350468749999999, 0.7303533568904593], "iscrowd": 0, "area": 33955.187350000044, "rle": {"size": [283, 640], "counts": "XZ21i81010O000001O0001O0001O0001O0001O0000010O0000000010O000001O0001O0001O0001O0001O00000\\MN`L2_32]LNc35ZLKf39VLGj3<SLDm3`0oKAP4b0nK]OTNMj5j0nKYOWNOj5k0lKVOZN0i5n0iKRO]N3h5n0dK\\O[4h0lJkNL?W5i0kJhNL`0Y5l0gJeNN`0\\5m0cJdNOa0]5o0aJ`N0b0_5Q1^J^N2a0`5R1]J_N1?b5`1]JaNb5`1]J`Nc5a1\\J_Nd5b1[J^Ne5]2O01O1O100O1O1O100O1O1O100O1O1O10O01O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O10O01O1O10O01O0001O0001O01O000000010O0000001O01O0001O0000001O000000001O0000001O0000010O000000001O0000001O000000001O00000010O000001O0000000000001O0000000000001O01O000000000001O000000000000001O00000001O000001OO1000000O100000000O100000000O2O00N2N11000UIZNb6h1YIZNg6m10O101O00000O10000000000O100000000O10^IXNQ6h1lI^NQ6b1oI`No5`1QJbNm5]1TJdNk5\\1UJcNl5^1SJaNn5_1SJ_Nn5b1QJ]NP6d1oI\\NQ6e1nI\\NQ6d1oI]NP6X2O001O1O1O1O00O101N1O100O100000000O1@mIUNV6k1=O100O1O1001O0000001O00001O0000001O000gIQNf5o1YJTNe5m1XJUNh5k1WJVNi5j1VJWNj5]2000001O00H9O000O10OJfImMZ6R2hImMX6R2jIlMW6S2jImMV6Q2mInMR6R2<N2O10O1000O10O100000000O1000000O1000O10000000000000000000000000000000O10000000000000000000000000000000000000000O10000000000000000000000000000000O10000000O1000000000000000000000000000000000000M3O12N2N0000000000000000001O0000000000000000000O100000000000000000000000000000000000001O000000000000000000000000000000000000000000000OnIQNZ5o1eJTNY5l1fJVNY5j1fJWNZ5l1bJUN^5R2ZJnMg5Y2QJhMP6`2100F:L4N2M3M3N2M30001N10001N10001N10001N2O1O1N2O1O1N2O1O1N2O1O2M2O001O0O2O001N101O0O101O0O2O001N101O0O2O001N10001N101O0O2O001O0O2Ok^;"}, "label": "a white and green airplane hovers above the runway"}]}
{"id": 413869, "image": "COCO_train2014_000000413869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"eva air cargo airplane taking off on the runway\"."}], "task": "refering", "answer_template": "The \"eva air cargo airplane taking off on the runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 48, 49, 50, 71, 72, 73, 74, 84, 85, 86, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 180], "bbox": [0.012921875, 0.1191166077738516, 0.9350468749999999, 0.7303533568904593], "iscrowd": 0, "area": 33955.187350000044, "rle": {"size": [283, 640], "counts": "XZ21i81010O000001O0001O0001O0001O0001O0000010O0000000010O000001O0001O0001O0001O0001O00000\\MN`L2_32]LNc35ZLKf39VLGj3<SLDm3`0oKAP4b0nK]OTNMj5j0nKYOWNOj5k0lKVOZN0i5n0iKRO]N3h5n0dK\\O[4h0lJkNL?W5i0kJhNL`0Y5l0gJeNN`0\\5m0cJdNOa0]5o0aJ`N0b0_5Q1^J^N2a0`5R1]J_N1?b5`1]JaNb5`1]J`Nc5a1\\J_Nd5b1[J^Ne5]2O01O1O100O1O1O100O1O1O100O1O1O10O01O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O10O01O1O10O01O0001O0001O01O000000010O0000001O01O0001O0000001O000000001O0000001O0000010O000000001O0000001O000000001O00000010O000001O0000000000001O0000000000001O01O000000000001O000000000000001O00000001O000001OO1000000O100000000O100000000O2O00N2N11000UIZNb6h1YIZNg6m10O101O00000O10000000000O100000000O10^IXNQ6h1lI^NQ6b1oI`No5`1QJbNm5]1TJdNk5\\1UJcNl5^1SJaNn5_1SJ_Nn5b1QJ]NP6d1oI\\NQ6e1nI\\NQ6d1oI]NP6X2O001O1O1O1O00O101N1O100O100000000O1@mIUNV6k1=O100O1O1001O0000001O00001O0000001O000gIQNf5o1YJTNe5m1XJUNh5k1WJVNi5j1VJWNj5]2000001O00H9O000O10OJfImMZ6R2hImMX6R2jIlMW6S2jImMV6Q2mInMR6R2<N2O10O1000O10O100000000O1000000O1000O10000000000000000000000000000000O10000000000000000000000000000000000000000O10000000000000000000000000000000O10000000O1000000000000000000000000000000000000M3O12N2N0000000000000000001O0000000000000000000O100000000000000000000000000000000000001O000000000000000000000000000000000000000000000OnIQNZ5o1eJTNY5l1fJVNY5j1fJWNZ5l1bJUN^5R2ZJnMg5Y2QJhMP6`2100F:L4N2M3M3N2M30001N10001N10001N10001N2O1O1N2O1O1N2O1O1N2O1O2M2O001O0O2O001N101O0O101O0O2O001N101O0O2O001N10001N101O0O2O001O0O2Ok^;"}, "label": "eva air cargo airplane taking off on the runway"}]}
{"id": 495790, "image": "COCO_train2014_000000495790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small giraffe\"."}], "task": "refering", "answer_template": "The \"the small giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 214, 215, 216, 236, 237, 238, 259, 260, 261, 282, 283, 305, 306, 329], "bbox": [0.263421875, 0.4812083333333333, 0.436234375, 0.8346666666666667], "iscrowd": 0, "area": 4489.804500000002, "rle": {"size": [480, 640], "counts": "h`_2Q1Q>V1kN8GM3M3M2N3M3E;ZOe00bNQOkDn0S;TOPEi0o:XOUEd0j:]OYE`0f:A]EOo:2n13N1O1O1N2O1O1O1O1N2O2N1O1O1O1N4M4SCQOa;S1ZDQOc;T1nCZOn;o1LT1lN00100O1O1O1O1O10dN[1N2O1N2O1N3N1N2O1N2O1jNU12O0O100O101N10000001O01O0001O00000O2N1O1O101N1O1O2O0O2N1O2N101N20O100O100O2O0O1002O1N2N2N3N1N2N2N2N2O1NRoX5"}, "label": "the small giraffe"}]}
{"id": 495790, "image": "COCO_train2014_000000495790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller giraffe walking in the back of the enclosure\"."}], "task": "refering", "answer_template": "The \"the smaller giraffe walking in the back of the enclosure\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 214, 215, 216, 236, 237, 238, 259, 260, 261, 282, 283, 305, 306, 329], "bbox": [0.263421875, 0.4812083333333333, 0.436234375, 0.8346666666666667], "iscrowd": 0, "area": 4489.804500000002, "rle": {"size": [480, 640], "counts": "h`_2Q1Q>V1kN8GM3M3M2N3M3E;ZOe00bNQOkDn0S;TOPEi0o:XOUEd0j:]OYE`0f:A]EOo:2n13N1O1O1N2O1O1O1O1N2O2N1O1O1O1N4M4SCQOa;S1ZDQOc;T1nCZOn;o1LT1lN00100O1O1O1O1O10dN[1N2O1N2O1N3N1N2O1N2O1jNU12O0O100O101N10000001O01O0001O00000O2N1O1O101N1O1O2O0O2N1O2N101N20O100O100O2O0O1002O1N2N2N3N1N2N2N2N2O1NRoX5"}, "label": "the smaller giraffe walking in the back of the enclosure"}]}
{"id": 495790, "image": "COCO_train2014_000000495790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe walking this way\"."}], "task": "refering", "answer_template": "The \"a tall giraffe walking this way\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 80, 99, 100, 101, 102, 103, 120, 121, 122, 123, 142, 143, 144, 145, 164, 165, 166, 167, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 305, 323, 325, 326, 328, 349, 351], "bbox": [0.060812500000000005, 0.14189583333333333, 0.5050625, 0.9031458333333333], "iscrowd": 0, "area": 29600.361599999997, "rle": {"size": [480, 640], "counts": "Ycb0=W>c0G7nCnNi9X1lEPOQ:V1dEROY:T1\\EUO`:Q1WEUOf:Q1REUOk:a2J7J5K6IGgEWLS:m3RFQLg9R4`FjK]9W4hFfKV9Y4h000O10000O100O2aN^100O1O2O1N2O1N2N2O1N2001O00001O00001O001O1O2N3gDfMg9d2iEeMT:f2[EbMc:_3N2N3M2N2N2O3L6J6J7J5J3M2N3M3N1N3M3M3N1N3M3M2O2M3M3M3N3L3M4L3N3L3M3N1N1O2N101N1O2N100O1O100O100OaNeI[J[6W5ZJbJe5V5kJcJT5U5XKhJh4h4kKTKU4h4RLUKo3h4U3M4M2K6G8J6M4K4N30O100O2O0O1O2O0O3N1N3N1N3N2M3PIeKW3]4dLfK[3]4_LhK_3[4\\LhKd3[4ULjKi3Y4RLkKm3X4mKkKR4Y4gKlKX4V4cKnK[4T4bKnK^4T4^KnKa4U4[KmKd4k0gI@IMh1Lg4c0PJZOI7\\1Nj4a0UJQOL`0R10m4=bJaNHR1d03Q5:kKDPO4U56oKEhN9W52SLFaN;\\5NVLHYN<`5KZLMlM=i5F^LO`M1`6OSLf1T4[NoK[1X4eNkKV1W4iNmKQ1V4oNmKl0U4SOnKi0T4WOoKd0S4[OQL`0Q4@RL;P4DTL2R4OQLFW4:kK^O[4a0iKXO[4h0hKRO[4m0iKoNX4Q1lKiNV4V1PLcNR4]1h4O100O2O000O101O00001O0000001O000010O000001O00001O0000001O00001O00001O0000001O00000000000O101O0001O001O1O1O00100O001O1O10O01C=00O011O000O11O10O01O1O10O01O1O1O101N2N2N1O100O1O001O100O1O1O100O001O1O1O1O101N1O1O1O2O0O2O1N2O1N2N3N1N2M4K4LQ_d4"}, "label": "a tall giraffe walking this way"}]}
{"id": 495790, "image": "COCO_train2014_000000495790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe in front of another giraffe , who seems to be in motion\"."}], "task": "refering", "answer_template": "The \"giraffe in front of another giraffe , who seems to be in motion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 80, 99, 100, 101, 102, 103, 120, 121, 122, 123, 142, 143, 144, 145, 164, 165, 166, 167, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 305, 323, 325, 326, 328, 349, 351], "bbox": [0.060812500000000005, 0.14189583333333333, 0.5050625, 0.9031458333333333], "iscrowd": 0, "area": 29600.361599999997, "rle": {"size": [480, 640], "counts": "Ycb0=W>c0G7nCnNi9X1lEPOQ:V1dEROY:T1\\EUO`:Q1WEUOf:Q1REUOk:a2J7J5K6IGgEWLS:m3RFQLg9R4`FjK]9W4hFfKV9Y4h000O10000O100O2aN^100O1O2O1N2O1N2N2O1N2001O00001O00001O001O1O2N3gDfMg9d2iEeMT:f2[EbMc:_3N2N3M2N2N2O3L6J6J7J5J3M2N3M3N1N3M3M3N1N3M3M2O2M3M3M3N3L3M4L3N3L3M3N1N1O2N101N1O2N100O1O100O100OaNeI[J[6W5ZJbJe5V5kJcJT5U5XKhJh4h4kKTKU4h4RLUKo3h4U3M4M2K6G8J6M4K4N30O100O2O0O1O2O0O3N1N3N1N3N2M3PIeKW3]4dLfK[3]4_LhK_3[4\\LhKd3[4ULjKi3Y4RLkKm3X4mKkKR4Y4gKlKX4V4cKnK[4T4bKnK^4T4^KnKa4U4[KmKd4k0gI@IMh1Lg4c0PJZOI7\\1Nj4a0UJQOL`0R10m4=bJaNHR1d03Q5:kKDPO4U56oKEhN9W52SLFaN;\\5NVLHYN<`5KZLMlM=i5F^LO`M1`6OSLf1T4[NoK[1X4eNkKV1W4iNmKQ1V4oNmKl0U4SOnKi0T4WOoKd0S4[OQL`0Q4@RL;P4DTL2R4OQLFW4:kK^O[4a0iKXO[4h0hKRO[4m0iKoNX4Q1lKiNV4V1PLcNR4]1h4O100O2O000O101O00001O0000001O000010O000001O00001O0000001O00001O00001O0000001O00000000000O101O0001O001O1O1O00100O001O1O10O01C=00O011O000O11O10O01O1O10O01O1O1O101N2N2N1O100O1O001O100O1O1O100O001O1O1O1O101N1O1O1O2O0O2O1N2O1N2N3N1N2M4K4LQ_d4"}, "label": "giraffe in front of another giraffe , who seems to be in motion"}]}
{"id": 561339, "image": "COCO_train2014_000000561339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large giraffee with head over railing\"."}], "task": "refering", "answer_template": "The \"large giraffee with head over railing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 221, 222, 223, 224, 238, 239, 240, 241, 245, 246, 247, 248, 261, 262, 263, 269, 270, 271, 272, 293, 294, 295, 317, 318], "bbox": [0.38775, 0.2866458333333333, 0.844390625, 0.8038541666666668], "iscrowd": 0, "area": 31340.240949999992, "rle": {"size": [480, 640], "counts": "\\ad3W1b=6J7J6I8H7J6J6K6J5K5K5M4N1O1O1O01O1O001O1O001O1O1O001O1O001O1O001O1O001O1OO010O100O1O100O100O1O100O100O100O1O100O100M3M3M4L3001O1O010O1O00100O001O1O010O1O00100O1O00100O001O1O010O1O00100O001O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O2N100000000000001O010O00001O001O00001O001O00001O010O001OO1O2N3L4M2N3N2N2N3M2N3M2N2N3M100O1O1YOfLfE[3U:oLfEQ3Y:TMbEm2\\:YM_Eh2`:m0O1O1O1O1O1O1O00100O100O100O100O1O1O1O1O1O1000000000000001N3N2N2N2N0000O100000000001O0000001O1O1001O2O11N10N0O0O100O2O0O100O100O1001O0`LjEU2W:gMmEX2S:eMRFY2n9eMVFY2k9cMYF]2f9]MbFa2g9nLbFQ3i:N1O2N2N1O2N1O2N2N1N3N1O001N10001O0O101N1O2M2N2N3M2N2N3M2N2N3M2N3M2N3N5J5J6J7I6J6Jo`^1"}, "label": "large giraffee with head over railing"}]}
{"id": 561339, "image": "COCO_train2014_000000561339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cute giraffe resting his head on a post\"."}], "task": "refering", "answer_template": "The \"cute giraffe resting his head on a post\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 221, 222, 223, 224, 238, 239, 240, 241, 245, 246, 247, 248, 261, 262, 263, 269, 270, 271, 272, 293, 294, 295, 317, 318], "bbox": [0.38775, 0.2866458333333333, 0.844390625, 0.8038541666666668], "iscrowd": 0, "area": 31340.240949999992, "rle": {"size": [480, 640], "counts": "\\ad3W1b=6J7J6I8H7J6J6K6J5K5K5M4N1O1O1O01O1O001O1O001O1O1O001O1O001O1O001O1O001O1OO010O100O1O100O100O1O100O100O100O1O100O100M3M3M4L3001O1O010O1O00100O001O1O010O1O00100O1O00100O001O1O010O1O00100O001O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O2N100000000000001O010O00001O001O00001O001O00001O010O001OO1O2N3L4M2N3N2N2N3M2N3M2N2N3M100O1O1YOfLfE[3U:oLfEQ3Y:TMbEm2\\:YM_Eh2`:m0O1O1O1O1O1O1O00100O100O100O100O1O1O1O1O1O1000000000000001N3N2N2N2N0000O100000000001O0000001O1O1001O2O11N10N0O0O100O2O0O100O100O1001O0`LjEU2W:gMmEX2S:eMRFY2n9eMVFY2k9cMYF]2f9]MbFa2g9nLbFQ3i:N1O2N2N1O2N1O2N2N1N3N1O001N10001O0O101N1O2M2N2N3M2N2N3M2N2N3M2N3M2N3N5J5J6J7I6J6Jo`^1"}, "label": "cute giraffe resting his head on a post"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a black jacket\"."}], "task": "refering", "answer_template": "The \"a boy in a black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337], "bbox": [0.367328125, 0.06967213114754098, 0.701671875, 0.9865105386416863], "iscrowd": 0, "area": 59508.4497, "rle": {"size": [427, 640], "counts": "]VR37l<=C=D<C>B=C;J5L4M2M4L4M3L4L4M3L4L3N3N2M3N2N2N2M3N2N1N3N2N2N2M4M4L4eHVKm5n4dI`KX6d4YIkKd6a5M3L3N3M3M3L4M4L3L5L4L4L4K5L4L3M4K5L1O0O2O0O100O100O1O100O100O100O1O100O1SOoF`MQ9]2SGaMn8[2WGcMj8X2]GeMd8V2bGgM`8T2fGjM[8P2mGmMT8n1RHoMP8l1VHRNl7h1[HUNf7f1R2O1O100O1O1O1O100O1O1O100O1O1O1000000O1000001N1000000O10000O1000000O1000000000000000000000001O0001O0000000000000000000000000000001O00001O001O010O1O001O1O001O001O1O1O5K4L4L5K4L4^EcNn8a1dFjN[9d2O1O1O1O00100O1O1O1O1O1O1O1O1O1O1O1O1[IPLm3Q4aKgLX4Z3VK_Mb4b2TKoMd4S2YKVN]4m1`K\\NV4i1gK`Nn3d1oKfNf3_1WLjN^3[1^LoNW3U1gLSOo2R1mLXOh2l0VMXOb2n5G9G8I8G9G9H7J7H8I6IYZ_2"}, "label": "a boy in a black jacket"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in fila shirt\"."}], "task": "refering", "answer_template": "The \"boy in fila shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337], "bbox": [0.367328125, 0.06967213114754098, 0.701671875, 0.9865105386416863], "iscrowd": 0, "area": 59508.4497, "rle": {"size": [427, 640], "counts": "]VR37l<=C=D<C>B=C;J5L4M2M4L4M3L4L4M3L4L3N3N2M3N2N2N2M3N2N1N3N2N2N2M4M4L4eHVKm5n4dI`KX6d4YIkKd6a5M3L3N3M3M3L4M4L3L5L4L4L4K5L4L3M4K5L1O0O2O0O100O100O1O100O100O100O1O100O1SOoF`MQ9]2SGaMn8[2WGcMj8X2]GeMd8V2bGgM`8T2fGjM[8P2mGmMT8n1RHoMP8l1VHRNl7h1[HUNf7f1R2O1O100O1O1O1O100O1O1O100O1O1O1000000O1000001N1000000O10000O1000000O1000000000000000000000001O0001O0000000000000000000000000000001O00001O001O010O1O001O1O001O001O1O1O5K4L4L5K4L4^EcNn8a1dFjN[9d2O1O1O1O00100O1O1O1O1O1O1O1O1O1O1O1O1[IPLm3Q4aKgLX4Z3VK_Mb4b2TKoMd4S2YKVN]4m1`K\\NV4i1gK`Nn3d1oKfNf3_1WLjN^3[1^LoNW3U1gLSOo2R1mLXOh2l0VMXOb2n5G9G8I8G9G9H7J7H8I6IYZ_2"}, "label": "boy in fila shirt"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a green zip up sweatshirt\"."}], "task": "refering", "answer_template": "The \"a boy in a green zip up sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.023984375, 0.23557377049180328, 0.41829687500000007, 0.999625292740047], "iscrowd": 0, "area": 58043.04784999999, "rle": {"size": [427, 640], "counts": "eb61W=6I8H8I6I8H8I5J4L5L3L5K4L5L3L5K4M4K4L5K4M4K4L5L3L4L5K4N2]GiKP8Y4oGhKP8Y4oGiKo7Y4nGiKP8Y4oGiKo7X4PHjKa0N[6Y4SIjK1>k6i3RImLl6V3RIkLm6V3RIkLm6V3QIlLn6i4N2O1O1O1N2O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100bMcHmN^7S1dHkN\\7U1fHiNZ7W1hHfNY7Z1iHdNW7\\1kHbNU7^1mH_NT7a1nH]NR7c1PI[NP7e1SIWNo6h1SIVNm6i1VITNk6l1WIRNi6n1YIPNg6P2[ImMf6S2\\IkMd6U2^IiMb6W2`IfMa6Z2bIbM_6^2dI^M]6b2fI[MZ6e2iIWMX6i2kISMV6m2mIoLT6Q3nIlLS6T3nIjLS6U3oIjLQ6V3oIiLR6W3oIgLR6Y3oIeLR6[3o100000000000000000000O101O000000000000000O1000000000000000000O10000000000000000O1000000000000000000O10000000000000000O100000000000dGnL_6R3`IQM^6o2aITM]6m2aIVM]6P3\\ISMb6S3WIPMg6V3PInLo6X3hHlLW7Y3`HmL^7Y3XHlLg7]4O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O002N3M2N3M3M2N3M3M2N3M3VNaHbMa7[2cHbM`7Z2cHdM`7Y2cHdM_7Y2dHeM_7X2dHeM_7W2dHgM^7W2eHfM^7V2eHhM^7T2fHiM\\7U2fHiM]7S2gHjM\\7S2fHkM\\7R2hHkM[7Q2hHmM[7P2hHmM[7n1iHPN[7h1lHUNX7d1mHZNX7^1`2H8_Oa0ZOijj4"}, "label": "a boy in a green zip up sweatshirt"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a green jacket looking at a cell phone\"."}], "task": "refering", "answer_template": "The \"a boy wearing a green jacket looking at a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.023984375, 0.23557377049180328, 0.41829687500000007, 0.999625292740047], "iscrowd": 0, "area": 58043.04784999999, "rle": {"size": [427, 640], "counts": "eb61W=6I8H8I6I8H8I5J4L5L3L5K4L5L3L5K4M4K4L5K4M4K4L5L3L4L5K4N2]GiKP8Y4oGhKP8Y4oGiKo7Y4nGiKP8Y4oGiKo7X4PHjKa0N[6Y4SIjK1>k6i3RImLl6V3RIkLm6V3RIkLm6V3QIlLn6i4N2O1O1O1N2O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100bMcHmN^7S1dHkN\\7U1fHiNZ7W1hHfNY7Z1iHdNW7\\1kHbNU7^1mH_NT7a1nH]NR7c1PI[NP7e1SIWNo6h1SIVNm6i1VITNk6l1WIRNi6n1YIPNg6P2[ImMf6S2\\IkMd6U2^IiMb6W2`IfMa6Z2bIbM_6^2dI^M]6b2fI[MZ6e2iIWMX6i2kISMV6m2mIoLT6Q3nIlLS6T3nIjLS6U3oIjLQ6V3oIiLR6W3oIgLR6Y3oIeLR6[3o100000000000000000000O101O000000000000000O1000000000000000000O10000000000000000O1000000000000000000O10000000000000000O100000000000dGnL_6R3`IQM^6o2aITM]6m2aIVM]6P3\\ISMb6S3WIPMg6V3PInLo6X3hHlLW7Y3`HmL^7Y3XHlLg7]4O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O002N3M2N3M3M2N3M3M2N3M3VNaHbMa7[2cHbM`7Z2cHdM`7Y2cHdM_7Y2dHeM_7X2dHeM_7W2dHgM^7W2eHfM^7V2eHhM^7T2fHiM\\7U2fHiM]7S2gHjM\\7S2fHkM\\7R2hHkM[7Q2hHmM[7P2hHmM[7n1iHPN[7h1lHUNX7d1mHZNX7^1`2H8_Oa0ZOijj4"}, "label": "a boy wearing a green jacket looking at a cell phone"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue umbrella shielding a boy in a black sweatshirt and a boy in a blue jacket\"."}], "task": "refering", "answer_template": "The \"a blue umbrella shielding a boy in a black sweatshirt and a boy in a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 34, 37, 38, 39, 40, 43, 44, 51, 52, 53, 54, 55, 56, 57, 60, 61, 62, 66, 67, 68, 79, 80, 83, 84, 85, 90, 91, 106, 107, 108, 113, 114, 130], "bbox": [0.223390625, 0.0022482435597189696, 1.0, 0.34831381733021083], "iscrowd": 0, "area": 28087.84569999999, "rle": {"size": [427, 640], "counts": "ZSl11Z=0000000000000J6J6K6I6J6K5O100O1000nNlCm0S<600O100OlNoCQ1P<5OjNQDR1o;5OiNRDS1n;lNRDU1m;50fNSDV1l;50eNTDW1k;iNUDX1k;hNUDX1k;gNUD[1i;fNWDZ1i;eNWD\\1h;eNXD[1h;dNXD]1g;cNYD^1g;bNYD^1f;bNZD_1f;aNZD_1e;aN[D`1e;_N\\Da1c;`N\\Da1d;^N]Db1b;_N]Dc1b;\\N_Dd1`;\\N`De1`;[N_Df1c;0MZN`Dg1b;0O100O100O101N1O100O1000000000000000O100000000O100000000O100000000O11O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1OZO]FdMc9R2kFkMT9P2TGnMk8o1\\GnMc8n1eGoMZ8m1nGPNQ8m1h1L4M3N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000003M3M3M3M2N3SFeNd7^1WHeNh7^1RHeNn7^1lGeNT8b1cGaN\\8i1WGZNi8P2jFTNU9W2^FlMa9P3O1O1O1O1O2N1O1O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1OO10000O10000O10000O10000O100O1M3N2M3N2M3N2J6J6J6J6I7J6J6J6K5M3M3M3L4M3M3M3M3M3M3M3M3M3O1O100O1O100O100O1O100O1O100O100O100O100O100O1O100O100O100O1000000O1000000O1000000O100001O00000000000000000000000000000000001O00000000001O000000001O00000000100000000000O11O1O2N1O1O2N1O1O2N1O1O1O4L5K4L4L4L1O001O1O001O1O001O1O1O001O1O001UE[Nl9g1nE`NP:`1jEeNW:\\1cEhN^:R21O1O1O1O1O1O1O1M3M3M3M3M3N2M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3N2M3M3M3M3M3M3McM"}, "label": "a blue umbrella shielding a boy in a black sweatshirt and a boy in a blue jacket"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue umbrella above boy in black\"."}], "task": "refering", "answer_template": "The \"blue umbrella above boy in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 34, 37, 38, 39, 40, 43, 44, 51, 52, 53, 54, 55, 56, 57, 60, 61, 62, 66, 67, 68, 79, 80, 83, 84, 85, 90, 91, 106, 107, 108, 113, 114, 130], "bbox": [0.223390625, 0.0022482435597189696, 1.0, 0.34831381733021083], "iscrowd": 0, "area": 28087.84569999999, "rle": {"size": [427, 640], "counts": "ZSl11Z=0000000000000J6J6K6I6J6K5O100O1000nNlCm0S<600O100OlNoCQ1P<5OjNQDR1o;5OiNRDS1n;lNRDU1m;50fNSDV1l;50eNTDW1k;iNUDX1k;hNUDX1k;gNUD[1i;fNWDZ1i;eNWD\\1h;eNXD[1h;dNXD]1g;cNYD^1g;bNYD^1f;bNZD_1f;aNZD_1e;aN[D`1e;_N\\Da1c;`N\\Da1d;^N]Db1b;_N]Dc1b;\\N_Dd1`;\\N`De1`;[N_Df1c;0MZN`Dg1b;0O100O100O101N1O100O1000000000000000O100000000O100000000O100000000O11O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1OZO]FdMc9R2kFkMT9P2TGnMk8o1\\GnMc8n1eGoMZ8m1nGPNQ8m1h1L4M3N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000003M3M3M3M2N3SFeNd7^1WHeNh7^1RHeNn7^1lGeNT8b1cGaN\\8i1WGZNi8P2jFTNU9W2^FlMa9P3O1O1O1O1O2N1O1O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1OO10000O10000O10000O10000O100O1M3N2M3N2M3N2J6J6J6J6I7J6J6J6K5M3M3M3L4M3M3M3M3M3M3M3M3M3O1O100O1O100O100O1O100O1O100O100O100O100O100O1O100O100O100O1000000O1000000O1000000O100001O00000000000000000000000000000000001O00000000001O000000001O00000000100000000000O11O1O2N1O1O2N1O1O2N1O1O1O4L5K4L4L4L1O001O1O001O1O001O1O1O001O1O001UE[Nl9g1nE`NP:`1jEeNW:\\1cEhN^:R21O1O1O1O1O1O1O1M3M3M3M3M3N2M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3N2M3M3M3M3M3M3McM"}, "label": "blue umbrella above boy in black"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a blue raincoat sitting two other boys\"."}], "task": "refering", "answer_template": "The \"a boy wearing a blue raincoat sitting two other boys\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 40, 41, 42, 43, 62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.664203125, 0.024824355971896955, 1.0, 0.9823185011709603], "iscrowd": 0, "area": 70486.15825000001, "rle": {"size": [427, 640], "counts": "n\\a5;m<d0[Od0]Od0[O?B7H8I7I6I8K5K5K5K5K5K5K5L7H?A?A<D4L5K4L5K4L5L4K4L5E:D=C=C<I8I6I4M4K4M3L4M4K4M3L5L3L4M3L5L3L4M4K4M3M3L5L300O2O00000O2O00000O10000O1000000O1000000O10000O1000001N1000000O10000O1000000O100000000O10000000001O000000000000000O100000000000000000000000001O01O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O10O02N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M3M2N4L5K5L4K5K5K2N2N2N2N2N2N2N2N2N2N2N2N2N>B1O001O001O010O001O001O001O001O001O001O003M6J6J^1bNW4iKQE"}, "label": "a boy wearing a blue raincoat sitting two other boys"}]}
{"id": 217276, "image": "COCO_train2014_000000217276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in blue jacket\"."}], "task": "refering", "answer_template": "The \"girl in blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 40, 41, 42, 43, 62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.664203125, 0.024824355971896955, 1.0, 0.9823185011709603], "iscrowd": 0, "area": 70486.15825000001, "rle": {"size": [427, 640], "counts": "n\\a5;m<d0[Od0]Od0[O?B7H8I7I6I8K5K5K5K5K5K5K5L7H?A?A<D4L5K4L5K4L5L4K4L5E:D=C=C<I8I6I4M4K4M3L4M4K4M3L5L3L4M3L5L3L4M4K4M3M3L5L300O2O00000O2O00000O10000O1000000O1000000O10000O1000001N1000000O10000O1000000O100000000O10000000001O000000000000000O100000000000000000000000001O01O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O10O02N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M3M2N4L5K5L4K5K5K2N2N2N2N2N2N2N2N2N2N2N2N2N>B1O001O001O010O001O001O001O001O001O001O003M6J6J^1bNW4iKQE"}, "label": "girl in blue jacket"}]}
{"id": 315581, "image": "COCO_train2014_000000315581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of a dog in the mirror\"."}], "task": "refering", "answer_template": "The \"the reflection of a dog in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 286, 287], "bbox": [0.407578125, 0.35852459016393445, 0.685484375, 0.8281733021077283], "iscrowd": 0, "area": 15779.430799999998, "rle": {"size": [427, 640], "counts": "PQ]33V=5L4L4L4L5K4L4L4L5K4L4L4K5L5K4L4L4L4L5K4L4L4L5K4K5L4L4L5K4L4L4L5VHeK[6_4_IdKa6_4ZIcKf6`4TIcKl6`4nHcKR7a4hHbKV7b4dHaK\\7W5000fM`HiN_7X1bHgN^7Y1dHfN[7Z1gHdNY7\\1iHbNW7^1kH`NT7b1nH[NR7e1QIXNo6h1TIVNk6j1YIRNg6n1^ImMa6T2dIgM\\6Z2hIaMX6_2mI\\MS6d2RJXMl5i2YJRMg5n2^JmLb5R3dJiL\\5W3iJdLW5[3oJ`LP5`3^200O1O1O010O1O10O01O1O12N5J6K5K4L5SHoKV6V4]IRLb6T4QITLo6P4eHXL[7l400O100000000O100004L3M3M3L5L3M3M4XOjGSLY8b1eG45WNY8`1gG55WNV8`1kG52XNT8a1nG41XNS8`1QH5e8H`G4b8IcG4_8IeG4\\8JiG1Y8OiGNY80jGMW83kGJV86lGFW89kGDV8<mG@T8?SHYOP8f0XHQOi7o0Z201O0O101O00001O0O101O00001N10001O000O2O001O001N2O001O001O11N4M4L3L5L1NO2M3N1O2N1O2N2N1O2N2M2O2N1O2N1O1O1O1O0O2O1O1O1O1Oo`c2"}, "label": "the reflection of a dog in the mirror"}]}
{"id": 315581, "image": "COCO_train2014_000000315581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog in the mirror\"."}], "task": "refering", "answer_template": "The \"the dog in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 286, 287], "bbox": [0.407578125, 0.35852459016393445, 0.685484375, 0.8281733021077283], "iscrowd": 0, "area": 15779.430799999998, "rle": {"size": [427, 640], "counts": "PQ]33V=5L4L4L4L5K4L4L4L5K4L4L4K5L5K4L4L4L4L5K4L4L4L5K4K5L4L4L5K4L4L4L5VHeK[6_4_IdKa6_4ZIcKf6`4TIcKl6`4nHcKR7a4hHbKV7b4dHaK\\7W5000fM`HiN_7X1bHgN^7Y1dHfN[7Z1gHdNY7\\1iHbNW7^1kH`NT7b1nH[NR7e1QIXNo6h1TIVNk6j1YIRNg6n1^ImMa6T2dIgM\\6Z2hIaMX6_2mI\\MS6d2RJXMl5i2YJRMg5n2^JmLb5R3dJiL\\5W3iJdLW5[3oJ`LP5`3^200O1O1O010O1O10O01O1O12N5J6K5K4L5SHoKV6V4]IRLb6T4QITLo6P4eHXL[7l400O100000000O100004L3M3M3L5L3M3M4XOjGSLY8b1eG45WNY8`1gG55WNV8`1kG52XNT8a1nG41XNS8`1QH5e8H`G4b8IcG4_8IeG4\\8JiG1Y8OiGNY80jGMW83kGJV86lGFW89kGDV8<mG@T8?SHYOP8f0XHQOi7o0Z201O0O101O00001O0O101O00001N10001O000O2O001O001N2O001O001O11N4M4L3L5L1NO2M3N1O2N1O2N2N1O2N2M2O2N1O2N1O1O1O1O0O2O1O1O1O1Oo`c2"}, "label": "the dog in the mirror"}]}
{"id": 438462, "image": "COCO_train2014_000000438462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with short blonde hair\"."}], "task": "refering", "answer_template": "The \"a woman with short blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 78, 79, 92, 93, 94, 95, 107, 108, 109, 110, 122, 123, 124, 125, 126, 137, 138, 139, 140], "bbox": [0.13842723004694835, 0.1988125, 0.4118779342723004, 0.416796875], "iscrowd": 0, "area": 10069.330699999999, "rle": {"size": [640, 426], "counts": "SSU1k0Rc05K5L2M4L3N3L3M4H7L5O0O2N101N1O101N101O001O00O10000000000O1000O100O1O0_Ob0H8N2O1N2N2N2L4M3N1N3N2N2N2N2N2O10000000000O10000000000000O101O00000000009G2N1O1O2N2N1O1O1bMf_OT1d`0_N__Oa1c`0ZN`_Oe1b`0WNa_Oi1``0SNd_Ol1^`0oMe_OQ2\\`0kMg_OU2Ra0O101O000O2O00001N2O1O001N3N2N2N7H6J3N2M3N2M3N2M3N2M4K5Kfdl4"}, "label": "a woman with short blonde hair"}]}
{"id": 438462, "image": "COCO_train2014_000000438462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman opn left\"."}], "task": "refering", "answer_template": "The \"woman opn left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 78, 79, 92, 93, 94, 95, 107, 108, 109, 110, 122, 123, 124, 125, 126, 137, 138, 139, 140], "bbox": [0.13842723004694835, 0.1988125, 0.4118779342723004, 0.416796875], "iscrowd": 0, "area": 10069.330699999999, "rle": {"size": [640, 426], "counts": "SSU1k0Rc05K5L2M4L3N3L3M4H7L5O0O2N101N1O101N101O001O00O10000000000O1000O100O1O0_Ob0H8N2O1N2N2N2L4M3N1N3N2N2N2N2N2O10000000000O10000000000000O101O00000000009G2N1O1O2N2N1O1O1bMf_OT1d`0_N__Oa1c`0ZN`_Oe1b`0WNa_Oi1``0SNd_Ol1^`0oMe_OQ2\\`0kMg_OU2Ra0O101O000O2O00001N2O1O001N3N2N2N7H6J3N2M3N2M3N2M3N2M4K5Kfdl4"}, "label": "woman opn left"}]}
{"id": 438462, "image": "COCO_train2014_000000438462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black dark blue shirt at a bar\"."}], "task": "refering", "answer_template": "The \"man in black dark blue shirt at a bar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 69, 70, 71, 83, 84, 85, 86, 97, 98, 99, 100, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 159, 160], "bbox": [0.48220657276995305, 0.15340625000000002, 0.752605633802817, 0.44293750000000004], "iscrowd": 0, "area": 13423.817300000006, "rle": {"size": [640, 426], "counts": "P[P42hc0?@?A`0@=D6I7O1O2N1O1O1O2N1O1O1O101N1O1O1O1O2N1O1O101N1O1O1O1O2N1O1O100O3N7I3L100O2O\\OS@UMm?e2Z@ZMe?a2b@^M^?\\2i@cMV?Z2o@fMP?Z2QAdMP?]2PAaMP?`2QA^MP?b2QA]Mn>d2W11N10000O2O0n_O[Mh>f2WA[Mi>e2WA[Mi>f2UA[Mj=OoAg27ZMi=<dAZ2b0\\Mi=g0WAn1o0\\Mi=f3VBZLi=g3VBZLi=g3WBYLh=h3WBYLh=h3XBXLg=j3WBWLh=j3RB\\Lm=e3hAfLW>b4O2N1O1O1O100O1O1001O0001O000001O00000000001O0[LfAe1Y>[NiAc1W>\\NkAc1U>\\NnAb1R>^NoAa1R>]NPBb1R>[NQBc1P>[NRBd1P>ZNQBe1P>XNTBf1n=UNZBb1l=ZN^BZ1g=bNdBR1b=jNiBi0_=QOkBc0]=XOmB;Z=AoB3Y=HPCLW=0a3L3Le_Q2"}, "label": "man in black dark blue shirt at a bar"}]}
{"id": 438462, "image": "COCO_train2014_000000438462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a dark blue shirt\"."}], "task": "refering", "answer_template": "The \"a man with a dark blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 69, 70, 71, 83, 84, 85, 86, 97, 98, 99, 100, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 159, 160], "bbox": [0.48220657276995305, 0.15340625000000002, 0.752605633802817, 0.44293750000000004], "iscrowd": 0, "area": 13423.817300000006, "rle": {"size": [640, 426], "counts": "P[P42hc0?@?A`0@=D6I7O1O2N1O1O1O2N1O1O1O101N1O1O1O1O2N1O1O101N1O1O1O1O2N1O1O100O3N7I3L100O2O\\OS@UMm?e2Z@ZMe?a2b@^M^?\\2i@cMV?Z2o@fMP?Z2QAdMP?]2PAaMP?`2QA^MP?b2QA]Mn>d2W11N10000O2O0n_O[Mh>f2WA[Mi>e2WA[Mi>f2UA[Mj=OoAg27ZMi=<dAZ2b0\\Mi=g0WAn1o0\\Mi=f3VBZLi=g3VBZLi=g3WBYLh=h3WBYLh=h3XBXLg=j3WBWLh=j3RB\\Lm=e3hAfLW>b4O2N1O1O1O100O1O1001O0001O000001O00000000001O0[LfAe1Y>[NiAc1W>\\NkAc1U>\\NnAb1R>^NoAa1R>]NPBb1R>[NQBc1P>[NRBd1P>ZNQBe1P>XNTBf1n=UNZBb1l=ZN^BZ1g=bNdBR1b=jNiBi0_=QOkBc0]=XOmB;Z=AoB3Y=HPCLW=0a3L3Le_Q2"}, "label": "a man with a dark blue shirt"}]}
{"id": 184513, "image": "COCO_train2014_000000184513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow lights that face away and the bulbs can ' t be seen\"."}], "task": "refering", "answer_template": "The \"the yellow lights that face away and the bulbs can ' t be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 80, 81, 82, 90, 91, 92, 100, 101, 102, 111, 112, 120, 121, 122, 130, 131, 132, 140, 141, 142, 151, 152, 160, 161, 162, 163, 170, 171, 172, 173, 174], "bbox": [0.003992932862190812, 0.4054, 0.42579505300353354, 0.9887400000000001], "iscrowd": 0, "area": 16133.877750000001, "rle": {"size": [500, 283], "counts": "Uj08d;l0PGH\\8:bGH8TOX5U1_JI7Eg4d0PKJ6Ff4b0QKK7Fe4a0RKK7Fe4a0RKK7Gc4a0TKJ7Gc4a0TKJ7Gc4a0TKJ7Ee4b0SKK6Cg4d0QKK6Ai4f0oJK6_Ok4h0mJK5^Om4j0lJJ5\\Oo4l0jJJ5ZOQ5n0hJJ5XOS5P1fJJ5VOU5R1dJK4SOX5S1cJL3QOZ5U1`JM4nN\\5W1^JM4lN]5Z1]JL4jN_5\\1[JL4hNa5^1YJL4fNc5`1WJL4dNe5b1UJL4bNg5c1TJM3`Ni5c1TJO1^Nj5d1UJ0O\\Nl5d1UJ1M\\Nn5c1UJ1M\\Nn5c1UJ1M\\Nn5c1UJ1M\\Nn5d1TJ0N\\Nn5d1TJ0N\\Nn5d1SJ1O[Nm5e1TJ0O[Nm5e1TJ0O[Nm5e1TJ1NZNn5e1TJ1NZNn5f1SJ0OZNn5h1QJN1ZNn5i1PJN1YNn5l1oIM1WNP6m1nIM1VNQ6o1lIL2UNR6P2kIL1UNT6P2jIL1TNU6R2hIK2SNV6S2gIK2RNV6U2gIJ2QNV6W2gII2PNV6Z2eIH4nMV6\\2eIG4mMV6^2cIH6jMV6`2aII8gMV6b2_IJ:dMW6X6hIhIX6Y6dIjI\\6V6aImI^6e700000000000kMgIoJ[6h4nIVKU6a4VJZK\\8n1RHjMS=I7I6J001O00000000000000000000001O00000000000000000000001O0000000000000000000000001O00002N6J7IeS_2"}, "label": "the yellow lights that face away and the bulbs can ' t be seen"}]}
{"id": 184513, "image": "COCO_train2014_000000184513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a traffic light , with the actual lights not visible\"."}], "task": "refering", "answer_template": "The \"the side of a traffic light , with the actual lights not visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 80, 81, 82, 90, 91, 92, 100, 101, 102, 111, 112, 120, 121, 122, 130, 131, 132, 140, 141, 142, 151, 152, 160, 161, 162, 163, 170, 171, 172, 173, 174], "bbox": [0.003992932862190812, 0.4054, 0.42579505300353354, 0.9887400000000001], "iscrowd": 0, "area": 16133.877750000001, "rle": {"size": [500, 283], "counts": "Uj08d;l0PGH\\8:bGH8TOX5U1_JI7Eg4d0PKJ6Ff4b0QKK7Fe4a0RKK7Fe4a0RKK7Gc4a0TKJ7Gc4a0TKJ7Gc4a0TKJ7Ee4b0SKK6Cg4d0QKK6Ai4f0oJK6_Ok4h0mJK5^Om4j0lJJ5\\Oo4l0jJJ5ZOQ5n0hJJ5XOS5P1fJJ5VOU5R1dJK4SOX5S1cJL3QOZ5U1`JM4nN\\5W1^JM4lN]5Z1]JL4jN_5\\1[JL4hNa5^1YJL4fNc5`1WJL4dNe5b1UJL4bNg5c1TJM3`Ni5c1TJO1^Nj5d1UJ0O\\Nl5d1UJ1M\\Nn5c1UJ1M\\Nn5c1UJ1M\\Nn5c1UJ1M\\Nn5d1TJ0N\\Nn5d1TJ0N\\Nn5d1SJ1O[Nm5e1TJ0O[Nm5e1TJ0O[Nm5e1TJ1NZNn5e1TJ1NZNn5f1SJ0OZNn5h1QJN1ZNn5i1PJN1YNn5l1oIM1WNP6m1nIM1VNQ6o1lIL2UNR6P2kIL1UNT6P2jIL1TNU6R2hIK2SNV6S2gIK2RNV6U2gIJ2QNV6W2gII2PNV6Z2eIH4nMV6\\2eIG4mMV6^2cIH6jMV6`2aII8gMV6b2_IJ:dMW6X6hIhIX6Y6dIjI\\6V6aImI^6e700000000000kMgIoJ[6h4nIVKU6a4VJZK\\8n1RHjMS=I7I6J001O00000000000000000000001O00000000000000000000001O0000000000000000000000001O00002N6J7IeS_2"}, "label": "the side of a traffic light , with the actual lights not visible"}]}
{"id": 184513, "image": "COCO_train2014_000000184513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow traffic light that has a sad face drawn on the red light\"."}], "task": "refering", "answer_template": "The \"a yellow traffic light that has a sad face drawn on the red light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 42, 43, 44, 45, 46, 47, 52, 53, 54, 55, 56, 57, 62, 63, 64, 65, 66, 67, 72, 73, 74, 75, 76, 77, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 102, 103, 104, 105, 106, 107, 112, 113, 114, 115, 116, 117, 122, 123, 124, 125, 126, 127, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 176, 177], "bbox": [0.24614840989399292, 0.16854, 0.7543462897526502, 0.95056], "iscrowd": 0, "area": 62505.57430000001, "rle": {"size": [500, 283], "counts": "d]R1m1U<R2jKW4_O`0@a0_O`0@a0_O`0@a0_O`0A4K0000001O00000000000000000000000000001O000000000001O00000000000001O00000000000000000000000000001O0001O0000000000O1000000000001O000000000000000000000000001O000000000000000000000O1000001O000000000000000000000000001O000000000000000000000000000O2O0000000000000000\\FS3a0m6QOP1kN]hQ1"}, "label": "a yellow traffic light that has a sad face drawn on the red light"}]}
{"id": 184513, "image": "COCO_train2014_000000184513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"signal\"."}], "task": "refering", "answer_template": "The \"signal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 42, 43, 44, 45, 46, 47, 52, 53, 54, 55, 56, 57, 62, 63, 64, 65, 66, 67, 72, 73, 74, 75, 76, 77, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 102, 103, 104, 105, 106, 107, 112, 113, 114, 115, 116, 117, 122, 123, 124, 125, 126, 127, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 176, 177], "bbox": [0.24614840989399292, 0.16854, 0.7543462897526502, 0.95056], "iscrowd": 0, "area": 62505.57430000001, "rle": {"size": [500, 283], "counts": "d]R1m1U<R2jKW4_O`0@a0_O`0@a0_O`0@a0_O`0A4K0000001O00000000000000000000000000001O000000000001O00000000000001O00000000000000000000000000001O0001O0000000000O1000000000001O000000000000000000000000001O000000000000000000000O1000001O000000000000000000000000001O000000000000000000000000000O2O0000000000000000\\FS3a0m6QOP1kN]hQ1"}, "label": "signal"}]}
{"id": 446661, "image": "COCO_train2014_000000446661.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"virgin atlantic aeroplane is ready to take - off\"."}], "task": "refering", "answer_template": "The \"virgin atlantic aeroplane is ready to take - off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 279, 282, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.061281249999999995, 0.49191588785046725, 0.9633750000000001, 0.8458878504672896], "iscrowd": 0, "area": 36315.040349999996, "rle": {"size": [428, 640], "counts": "Sc`02X=4L4L3O1N101N100O2O0O2O0O2O0O101N1O2O0O2O0O101N101N100O2O0O101N100O2O0O101N101N100O2O0O101N100O2O0O2O0O1jD`N`:a1^E`Na:d1ZE]Ng:T2O100O1O10000000A\\EVNd:e1bEZN^:f1dEXN\\:h1eEWN[:j1eEUN[:k1gESNY:m1hERNX:n1c001O00000001O00000000000000001O00000001O00000001O000000000000001O0000000001O000001O0000O1000000000000000000000001O000000000O1000000000000000000000010:E00001O0000000000000000001O00000001O0000000001O00000000000000001O0000O2M2O100000000000000000000000000000001O0000000001O01O000000001O0001O000001O00000001O01OPElMh:T2VEoMi:Q2UEQNk:o1SESNm:m1SESNn:l1RETNn:l1RETNn:U20000001O000000001O01O00000001O00000000001O00000001O0000000000000000000000000000001O000000000000000000000000000000@QEdNn:\\1SEcNm:]1TEbNl:^1VE`Nj:`1WE_Ni:a1YE]Ng:c1[E[Ne:e1a0000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000O100000000000000000000000000O1000000000000000000000000O100000000000000000000000000O100000000000000O1000001O0O10000000000O100000000O10000000000O1000001O0O1000000000O010O1O100O00100O1O100O010O1O100O00100O1O100O010O1O100O1O010O1O100O00100O1O1O10O01O1O1O1O010O1O1O1O1O010O1O1O1O1O10O01O1O1O1O10O01O1O1O1O100O001O1O1O100O1O001O1O100O1O1O001O100O1O1O1O00100O1O1O1O1O01WOQFUNn9j1SFWNm9g1UFYNk9e1VF\\Nj9a1YF_Ng9_1[F`Nf9^1[FcNe9[1]FeNc9Y1_FgNa9V1cFiN]9U1fFjNZ9T1iFkNW9S1lFiNW9T1mFiNU9U1nFhNT9V1nFgNU9W1nFfNT9W1PGfNR9X1b1N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2M3N2N_j9"}, "label": "virgin atlantic aeroplane is ready to take - off"}]}
{"id": 446661, "image": "COCO_train2014_000000446661.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"virgin atlantic plane front center\"."}], "task": "refering", "answer_template": "The \"virgin atlantic plane front center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 279, 282, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.061281249999999995, 0.49191588785046725, 0.9633750000000001, 0.8458878504672896], "iscrowd": 0, "area": 36315.040349999996, "rle": {"size": [428, 640], "counts": "Sc`02X=4L4L3O1N101N100O2O0O2O0O2O0O101N1O2O0O2O0O101N101N100O2O0O101N100O2O0O101N101N100O2O0O101N100O2O0O2O0O1jD`N`:a1^E`Na:d1ZE]Ng:T2O100O1O10000000A\\EVNd:e1bEZN^:f1dEXN\\:h1eEWN[:j1eEUN[:k1gESNY:m1hERNX:n1c001O00000001O00000000000000001O00000001O00000001O000000000000001O0000000001O000001O0000O1000000000000000000000001O000000000O1000000000000000000000010:E00001O0000000000000000001O00000001O0000000001O00000000000000001O0000O2M2O100000000000000000000000000000001O0000000001O01O000000001O0001O000001O00000001O01OPElMh:T2VEoMi:Q2UEQNk:o1SESNm:m1SESNn:l1RETNn:l1RETNn:U20000001O000000001O01O00000001O00000000001O00000001O0000000000000000000000000000001O000000000000000000000000000000@QEdNn:\\1SEcNm:]1TEbNl:^1VE`Nj:`1WE_Ni:a1YE]Ng:c1[E[Ne:e1a0000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000O100000000000000000000000000O1000000000000000000000000O100000000000000000000000000O100000000000000O1000001O0O10000000000O100000000O10000000000O1000001O0O1000000000O010O1O100O00100O1O100O010O1O100O00100O1O100O010O1O100O1O010O1O100O00100O1O1O10O01O1O1O1O010O1O1O1O1O010O1O1O1O1O10O01O1O1O1O10O01O1O1O1O100O001O1O1O100O1O001O1O100O1O1O001O100O1O1O1O00100O1O1O1O1O01WOQFUNn9j1SFWNm9g1UFYNk9e1VF\\Nj9a1YF_Ng9_1[F`Nf9^1[FcNe9[1]FeNc9Y1_FgNa9V1cFiN]9U1fFjNZ9T1iFkNW9S1lFiNW9T1mFiNU9U1nFhNT9V1nFgNU9W1nFfNT9W1PGfNR9X1b1N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2M3N2N_j9"}, "label": "virgin atlantic plane front center"}]}
{"id": 225477, "image": "COCO_train2014_000000225477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a dark suit and tie sitting at the table\"."}], "task": "refering", "answer_template": "The \"a man in a dark suit and tie sitting at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 28, 29, 30, 31, 32, 51, 52, 53, 54, 75, 76, 77], "bbox": [0.2245625, 0.009133489461358314, 0.43846875, 0.27147540983606555], "iscrowd": 0, "area": 9872.3262, "rle": {"size": [427, 640], "counts": "PRl17S=<D;E9G8H8H9G8H9HO0O1O1N2O2M3M3N200000000000000000000001O00000000002N6J7I7I6J5K1O001O001O001O001O00001O0000001O00001O0000001O00001O000000000000001O00000000000000O100O1WOWFoMj9Q2XFlMi9T2ZFhMg9X2[FeMf9[2]FbMc9^2_F_Mb9`2aF]M`9c2d0C=YOg0O100O100O100O2O0000000000000000000O1000000O1O1N2O2M2O200O1000000O1O1O100O1O001O1N4_O`CNh<Lnee4"}, "label": "a man in a dark suit and tie sitting at the table"}]}
{"id": 225477, "image": "COCO_train2014_000000225477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the body of a man in a suit and tie reaching for something at the dinner table\"."}], "task": "refering", "answer_template": "The \"the body of a man in a suit and tie reaching for something at the dinner table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 28, 29, 30, 31, 32, 51, 52, 53, 54, 75, 76, 77], "bbox": [0.2245625, 0.009133489461358314, 0.43846875, 0.27147540983606555], "iscrowd": 0, "area": 9872.3262, "rle": {"size": [427, 640], "counts": "PRl17S=<D;E9G8H8H9G8H9HO0O1O1N2O2M3M3N200000000000000000000001O00000000002N6J7I7I6J5K1O001O001O001O001O00001O0000001O00001O0000001O00001O000000000000001O00000000000000O100O1WOWFoMj9Q2XFlMi9T2ZFhMg9X2[FeMf9[2]FbMc9^2_F_Mb9`2aF]M`9c2d0C=YOg0O100O100O100O2O0000000000000000000O1000000O1O1N2O2M2O200O1000000O1O1O100O1O001O1N4_O`CNh<Lnee4"}, "label": "the body of a man in a suit and tie reaching for something at the dinner table"}]}
{"id": 225477, "image": "COCO_train2014_000000225477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stack of cupcakes\"."}], "task": "refering", "answer_template": "The \"stack of cupcakes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 33, 34, 55, 56, 57, 58, 78, 79, 80, 81, 100, 101, 102, 103, 104], "bbox": [0.373328125, 0.05168618266978923, 0.566734375, 0.31911007025761123], "iscrowd": 0, "area": 7935.265599999998, "rle": {"size": [427, 640], "counts": "lhS31Y=1O2N1O2N1N2O2N1O1O2N1N2O2N1O1O2N1O1M4K4K5L5L3O2N100O2N1O1O2O0O1O2N1O100O1O1O1O101N1O1O1O100O1L5H7I7N2O1O100O1O2O0O1O100O1O101N1O100O1O100O1O10O01O100O1O100O03N2N2N2N2N2M3N2N2N2N2N2N1N2O2N1O2N1O1N3N1O1O2N1O1O2hNhD8Z;AmD>T;[OREe0P;SOWEl0e;O001N2O001O1O1O003L5L4L5K4L4LY\\c3"}, "label": "stack of cupcakes"}]}
{"id": 225477, "image": "COCO_train2014_000000225477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tower of white cupcakes with red accents\"."}], "task": "refering", "answer_template": "The \"a tower of white cupcakes with red accents\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 33, 34, 55, 56, 57, 58, 78, 79, 80, 81, 100, 101, 102, 103, 104], "bbox": [0.373328125, 0.05168618266978923, 0.566734375, 0.31911007025761123], "iscrowd": 0, "area": 7935.265599999998, "rle": {"size": [427, 640], "counts": "lhS31Y=1O2N1O2N1N2O2N1O1O2N1N2O2N1O1O2N1O1M4K4K5L5L3O2N100O2N1O1O2O0O1O2N1O100O1O1O1O101N1O1O1O100O1L5H7I7N2O1O100O1O2O0O1O100O1O101N1O100O1O100O1O10O01O100O1O100O03N2N2N2N2N2M3N2N2N2N2N2N1N2O2N1O2N1O1N3N1O1O2N1O1O2hNhD8Z;AmD>T;[OREe0P;SOWEl0e;O001N2O001O1O1O003L5L4L5K4L4LY\\c3"}, "label": "a tower of white cupcakes with red accents"}]}
{"id": 86216, "image": "COCO_train2014_000000086216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second horse from the front\"."}], "task": "refering", "answer_template": "The \"the second horse from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [351, 372, 373, 393, 394, 395, 415, 416, 417, 424, 425, 426, 427, 437, 438, 439, 446, 447, 448, 449, 450, 451, 452, 453, 459, 460, 461, 467, 468, 469, 470, 471, 472, 481, 482, 483, 489, 490, 491, 492, 493, 494, 503, 504], "bbox": [0.23512698412698413, 0.6561874999999999, 1.0, 1.0], "iscrowd": 0, "area": 23419.317899999995, "rle": {"size": [640, 630], "counts": "lSm21lc03M3N3L3M3M3M3M3M3N2N2N2N2O2M2N2N2O1N2N2N2N2O1N3M2N2N2O1N1O2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2001O00001O00001O00001O01O01O00001O00001O00001O00001O00001O00001O00001O00000O2O0000001N1000001N1000001N1000000O2O00000O2O00000O2O0000001N1000001N1000001N100O1M3N3L3M3N2M4L3N2M3M4M2M3M3N3L3M3N2M4N10000O2O0000001O0000001O0000001O0000001O0000001O00000000000000000000000O1000000000000000000000000000O0100000000000000O100000000000000O1000000000000O101O001O001O001N2O001O001O001OiQT37UmkLk0eMFV@T1e?g1I6K5J6J6K5J6L4O1O1N3N1O1O1N2O1O1O1N2O1O2N1N2O1O1O00O01000000O10O1000O100000O01000000001N1000001O0J6F;D;E;G9K6J5K5J7J5K5J6K6J5K5J7J5K5K5J7J5K5JSC"}, "label": "the second horse from the front"}]}
{"id": 86216, "image": "COCO_train2014_000000086216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front horse in the right hand picture\"."}], "task": "refering", "answer_template": "The \"front horse in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 214, 215, 216, 217, 218, 219, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 440, 441, 442, 443, 444, 445, 446, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 462, 463, 465, 466, 467, 475, 476, 477, 478, 479, 480, 481, 484, 485, 487, 488, 498, 499, 500, 501, 502, 503], "bbox": [0.0021746031746031746, 0.37493750000000003, 1.0, 0.9894999999999999], "iscrowd": 0, "area": 166694.7392999999, "rle": {"size": [640, 630], "counts": "cV11lc0>B8H5K6K4K6J5L5K4HWNS^Om1ia0TNV^Om1ha07N2N2N2M3N2N2M3N1O2N2M3N2N2M3N2N1O2M3N2N2M3N2N1_LgL[FX3e9`NcD]1_;EaC8a<b3O3N1N3M2O2M21O100O2O0O100O100O100O2O000000000000001O0O1000000000001O000000000000001O00000O10000000cEPIl7n6SHUIl7k6RHXIm7h6QH[In7e6PH_In7`6QHcIn7]6oGgIP8Y6nGjIQ8V6mGmIS8Q6lGRJS8n5lGTJS8l5lGVJT8i5kGYJU8e5lG\\JS8d5lG^JT8a5kGaJU8^5jGdJV8Z5jGhJU8X5kGiJU8V5jGlJV8S5iGoJW8Q5gGQKX8o4gGSKY8l4gGUKY8j4fGXKZ8h4dGZK[8f4eG[K[8d4dG^K\\8a4dG`K\\8`4bGbK]8_4bGbK^8`4_GaKa8b4ZG`Kf8b4WG_Kh8e4SG]Km8e4PG\\KP9f4lF[KU9h4gFYKX9j4dFXK\\9k4`FVK`9l4\\FVKd9m4XFTKg9c70O1O001O1N2O001O1O1O001O1N20O01O100O00100O1O010O1O10O01O100O00100O001N2N1N3N1O2M3N1N3N2M2O2M2O2M3N1N3N2M2O2N1L5[Oe0I620O0001O00001O00010O00001O00001O01O01O00001O00001O01O01O00001O001O01O01O00001O0000010O00001O00000010O0001O0000001O01O0001O00001O0001O01O00001O0000010O00001O00001O000010O0001O00001O00001O01O01O00001O00001O000010O0001O00001O00001O01O01O00001O00001O00010O00001O00001O00001O00001O00001O0O101O00001O00001O0O101OO100000000000O100000O10000000O1000000000O100000O10000000000000O10O100000000000000O01000000000000000O10O100003Mm0SO4L5K4K6K4L5K3M001O1N101O0O2O001N10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O1000000O100000000O100000000O1000000000O0O2N2O1O1O001O1O1O1N101O1O1O1O0O2007I3M1N2O2N1O1O1N3N1O1O000O100000000O10000000000O100000000O100000000O1000000000cMiGQIV8e6YHWIg7^6gH_IY7U6VIhIj6o5cImI]6n5kIoIT6m5UJoIk5m5\\JoIe5m5cJnI^5P6gJlIZ5R6kJjIU5U6QKcIS5[6RKZIV5d6PKoHY5o6m3N2N1O2N2N2N2O0O2N2O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O010O1O1O1O10O01O1O1O1O010O1O001O10O01O1O0O2O1N101O1N101N2O0O2O1O0O2O0O2O1N101`LPL"}, "label": "front horse in the right hand picture"}]}
{"id": 86216, "image": "COCO_train2014_000000086216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark brown horse being rode by a black officer\"."}], "task": "refering", "answer_template": "The \"dark brown horse being rode by a black officer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 214, 215, 216, 217, 218, 219, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 440, 441, 442, 443, 444, 445, 446, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 462, 463, 465, 466, 467, 475, 476, 477, 478, 479, 480, 481, 484, 485, 487, 488, 498, 499, 500, 501, 502, 503], "bbox": [0.0021746031746031746, 0.37493750000000003, 1.0, 0.9894999999999999], "iscrowd": 0, "area": 166694.7392999999, "rle": {"size": [640, 630], "counts": "cV11lc0>B8H5K6K4K6J5L5K4HWNS^Om1ia0TNV^Om1ha07N2N2N2M3N2N2M3N1O2N2M3N2N2M3N2N1O2M3N2N2M3N2N1_LgL[FX3e9`NcD]1_;EaC8a<b3O3N1N3M2O2M21O100O2O0O100O100O100O2O000000000000001O0O1000000000001O000000000000001O00000O10000000cEPIl7n6SHUIl7k6RHXIm7h6QH[In7e6PH_In7`6QHcIn7]6oGgIP8Y6nGjIQ8V6mGmIS8Q6lGRJS8n5lGTJS8l5lGVJT8i5kGYJU8e5lG\\JS8d5lG^JT8a5kGaJU8^5jGdJV8Z5jGhJU8X5kGiJU8V5jGlJV8S5iGoJW8Q5gGQKX8o4gGSKY8l4gGUKY8j4fGXKZ8h4dGZK[8f4eG[K[8d4dG^K\\8a4dG`K\\8`4bGbK]8_4bGbK^8`4_GaKa8b4ZG`Kf8b4WG_Kh8e4SG]Km8e4PG\\KP9f4lF[KU9h4gFYKX9j4dFXK\\9k4`FVK`9l4\\FVKd9m4XFTKg9c70O1O001O1N2O001O1O1O001O1N20O01O100O00100O1O010O1O10O01O100O00100O001N2N1N3N1O2M3N1N3N2M2O2M2O2M3N1N3N2M2O2N1L5[Oe0I620O0001O00001O00010O00001O00001O01O01O00001O00001O01O01O00001O001O01O01O00001O0000010O00001O00000010O0001O0000001O01O0001O00001O0001O01O00001O0000010O00001O00001O000010O0001O00001O00001O01O01O00001O00001O000010O0001O00001O00001O01O01O00001O00001O00010O00001O00001O00001O00001O00001O0O101O00001O00001O0O101OO100000000000O100000O10000000O1000000000O100000O10000000000000O10O100000000000000O01000000000000000O10O100003Mm0SO4L5K4K6K4L5K3M001O1N101O0O2O001N10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O1000000O100000000O100000000O1000000000O0O2N2O1O1O001O1O1O1N101O1O1O1O0O2007I3M1N2O2N1O1O1N3N1O1O000O100000000O10000000000O100000000O100000000O1000000000cMiGQIV8e6YHWIg7^6gH_IY7U6VIhIj6o5cImI]6n5kIoIT6m5UJoIk5m5\\JoIe5m5cJnI^5P6gJlIZ5R6kJjIU5U6QKcIS5[6RKZIV5d6PKoHY5o6m3N2N1O2N2N2N2O0O2N2O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O010O1O1O1O10O01O1O1O1O010O1O001O10O01O1O0O2O1N101O1N101N2O0O2O1O0O2O0O2O1N101`LPL"}, "label": "dark brown horse being rode by a black officer"}]}
{"id": 561354, "image": "COCO_train2014_000000561354.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a school bus with the number 34\"."}], "task": "refering", "answer_template": "The \"a school bus with the number 34\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 268], "bbox": [0.2044375, 0.1052470588235294, 0.99684375, 0.7591764705882353], "iscrowd": 0, "area": 94926.58725, "rle": {"size": [425, 640], "counts": "n`f16S=8G:G8H8G:G8H8G:G8H8G9H6J1N10000O10000O101O0O10000O10000O2O000O10000O10000O2O000O10000O10001N10000O4M3M3L5L3M2M2O001N101O0O2O001N10ZOdG`L[8`3jG[LU8f3PHVLo7j3VHQLj7P4ZHlKd7U4h0000O10000O10001N1000000O10000O101O0O10000O10000O101N10000O10000O100O2O000O10000O10001N100O10000O10000O2O000O100O10000O101O0O10000O100O10001N10000O100O10000O2O000O10000O100O101O0O10000O10000O101N10000O10000O101N10000O10000O101O1N2O1N2O1O1N5L7I2M3N1O1N2O1N2O2N1N2O1O1N3N1O1N2OO010000O10000O10O010000O10000O01000O100O10M3K4K6O10O10000O101N10000O10000O2O000O100O10001N10000O100O101O0O10000O10001N100O10000O101O0O10000O100O2O000O10000O2O000O10000O101O0O10000O10001N10000O1000O010000O10O10O10000O10O10O10000O01000O1000O010000O1000O010000000O10001O000O100000001N10000000001N100000000O101O00000O1000001O000O1000001O0O100000000O2O0000000O10001O00000O1000001O0O100000001N10000000000O2O0000000O10001O000O1000001O0O10000000001N100000000O101O00000O1000001O000O1000001O0O10001O1^IiIP6X6oI`JX5b5fJ`JY5a5eJbJY5_5fJbJY5_5eJcJZ5_5dJbJZ5`5dJbJ[5_5dJcJZ5g6O1O1QK_Hd0E]2S9lLWGS3`9oLSFf2m9ZMVFc2j9]MXFa2h9`MZF:K[1k9[N]F7NW1f9bN^F52S1`9hNaF24Q1[9mNcFO9n0T9TOdFL=k0o8YOfFJa0f0j8@gFHe0b0d8F\\H5d7KbHO^71hHIX78l20000001O00010O01O010O00CGkC9T<KhC5W<0fCOZ<a0O00000O2O02N9H7H8Had0"}, "label": "a school bus with the number 34"}]}
{"id": 561354, "image": "COCO_train2014_000000561354.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus numbered 34\"."}], "task": "refering", "answer_template": "The \"bus numbered 34\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 268], "bbox": [0.2044375, 0.1052470588235294, 0.99684375, 0.7591764705882353], "iscrowd": 0, "area": 94926.58725, "rle": {"size": [425, 640], "counts": "n`f16S=8G:G8H8G:G8H8G:G8H8G9H6J1N10000O10000O101O0O10000O10000O2O000O10000O10000O2O000O10000O10001N10000O4M3M3L5L3M2M2O001N101O0O2O001N10ZOdG`L[8`3jG[LU8f3PHVLo7j3VHQLj7P4ZHlKd7U4h0000O10000O10001N1000000O10000O101O0O10000O10000O101N10000O10000O100O2O000O10000O10001N100O10000O10000O2O000O100O10000O101O0O10000O100O10001N10000O100O10000O2O000O10000O100O101O0O10000O10000O101N10000O10000O101N10000O10000O101O1N2O1N2O1O1N5L7I2M3N1O1N2O1N2O2N1N2O1O1N3N1O1N2OO010000O10000O10O010000O10000O01000O100O10M3K4K6O10O10000O101N10000O10000O2O000O100O10001N10000O100O101O0O10000O10001N100O10000O101O0O10000O100O2O000O10000O2O000O10000O101O0O10000O10001N10000O1000O010000O10O10O10000O10O10O10000O01000O1000O010000O1000O010000000O10001O000O100000001N10000000001N100000000O101O00000O1000001O000O1000001O0O100000000O2O0000000O10001O00000O1000001O0O100000001N10000000000O2O0000000O10001O000O1000001O0O10000000001N100000000O101O00000O1000001O000O1000001O0O10001O1^IiIP6X6oI`JX5b5fJ`JY5a5eJbJY5_5fJbJY5_5eJcJZ5_5dJbJZ5`5dJbJ[5_5dJcJZ5g6O1O1QK_Hd0E]2S9lLWGS3`9oLSFf2m9ZMVFc2j9]MXFa2h9`MZF:K[1k9[N]F7NW1f9bN^F52S1`9hNaF24Q1[9mNcFO9n0T9TOdFL=k0o8YOfFJa0f0j8@gFHe0b0d8F\\H5d7KbHO^71hHIX78l20000001O00010O01O010O00CGkC9T<KhC5W<0fCOZ<a0O00000O2O02N9H7H8Had0"}, "label": "bus numbered 34"}]}
{"id": 151756, "image": "COCO_train2014_000000151756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young baseball player wearing a helmet\"."}], "task": "refering", "answer_template": "The \"a young baseball player wearing a helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 74, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 259, 260, 279, 280, 282, 283, 301, 302, 303, 306], "bbox": [0.11578124999999999, 0.18644028103044497, 0.35120312499999995, 0.9261592505854802], "iscrowd": 0, "area": 20987.961700000007, "rle": {"size": [427, 640], "counts": "kdo02Y=0O101N101N1O101N1O100O1O1N2N2N2O0100000N2L4ROn0eHnN]2R1aMROR2Y1lMlNg1^1WNfN\\1e1bN`NR1i1lN[Nm0j1QOZNcLmNY3n23XN`LQOR3m2<VN^LSOn2k2c0UN[LWOh2i2m0YOjNk0V1WOcNn0]1TO]NP1c1ROWNQ1i1ROPNS1P2oNjMU1V2mNdMW1\\2kN`MW1`2lNZMV1f2nNUMT1k2oNoLT1Q3nNkLT1U3oNfLl0a3Z44L5K5J6J5L5J6J6K5J6K6K5M3M3L3N3L4N2M3N2M3N1O2M3N2N3L7J6K4N2N2N2N1O00N2O0O2O1N2O1N101N101N101N1011N5L3M0O100O100001O00PL]HV2c7hM_HY2a7cMbH]2_7`McH`2]7]MfHd2Z7YMiHf2X7VMkHj2T7TMoHm2o6RMbJ^1^5`NkJY1T5gNUKR1j4lN[KQ1e4mN]KS1b4mN_KS1a4kNaKV1^4iNdKV1[4iNgKW1Y4gNiK]1S4bNoKi1d3VN^Lk1a3TN`Ll1`3RNbLm1^3RNeLm1[3RNhLk1Y3SNkLj1U3UNPMg1Q3XNRMe1o2YNUMd1k2[NYMb1h2]N[M`1f2^N^M_1b2`NbM]1_2bNdM[1Y7M3M4L3L4M4L3M4K4L5K4Kce\\5"}, "label": "a young baseball player wearing a helmet"}]}
{"id": 151756, "image": "COCO_train2014_000000151756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in a blue shirt and black baseball helmet taking a swing\"."}], "task": "refering", "answer_template": "The \"a young boy in a blue shirt and black baseball helmet taking a swing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 74, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 259, 260, 279, 280, 282, 283, 301, 302, 303, 306], "bbox": [0.11578124999999999, 0.18644028103044497, 0.35120312499999995, 0.9261592505854802], "iscrowd": 0, "area": 20987.961700000007, "rle": {"size": [427, 640], "counts": "kdo02Y=0O101N101N1O101N1O100O1O1N2N2N2O0100000N2L4ROn0eHnN]2R1aMROR2Y1lMlNg1^1WNfN\\1e1bN`NR1i1lN[Nm0j1QOZNcLmNY3n23XN`LQOR3m2<VN^LSOn2k2c0UN[LWOh2i2m0YOjNk0V1WOcNn0]1TO]NP1c1ROWNQ1i1ROPNS1P2oNjMU1V2mNdMW1\\2kN`MW1`2lNZMV1f2nNUMT1k2oNoLT1Q3nNkLT1U3oNfLl0a3Z44L5K5J6J5L5J6J6K5J6K6K5M3M3L3N3L4N2M3N2M3N1O2M3N2N3L7J6K4N2N2N2N1O00N2O0O2O1N2O1N101N101N101N1011N5L3M0O100O100001O00PL]HV2c7hM_HY2a7cMbH]2_7`McH`2]7]MfHd2Z7YMiHf2X7VMkHj2T7TMoHm2o6RMbJ^1^5`NkJY1T5gNUKR1j4lN[KQ1e4mN]KS1b4mN_KS1a4kNaKV1^4iNdKV1[4iNgKW1Y4gNiK]1S4bNoKi1d3VN^Lk1a3TN`Ll1`3RNbLm1^3RNeLm1[3RNhLk1Y3SNkLj1U3UNPMg1Q3XNRMe1o2YNUMd1k2[NYMb1h2]N[M`1f2^N^M_1b2`NbM]1_2bNdM[1Y7M3M4L3L4M4L3M4K4L5K4Kce\\5"}, "label": "a young boy in a blue shirt and black baseball helmet taking a swing"}]}
{"id": 217293, "image": "COCO_train2014_000000217293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white shirt and gray pants and striped hat holding two baseball bats\"."}], "task": "refering", "answer_template": "The \"a man wearing a white shirt and gray pants and striped hat holding two baseball bats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 109, 110, 111, 124, 125, 126, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 273, 274, 275, 276, 277, 278, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 318, 319, 320, 321, 322, 323, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.14875878220140515, 0.29775, 0.614847775175644, 1.0], "iscrowd": 0, "area": 61628.216550000005, "rle": {"size": [640, 427], "counts": "n\\X15fc06J6J6K6I6J6J6K5J6J6J6J6K6I6J6J6J6K5J6J6J6K6I6M3N2O1N5K6K5J7I6K5J6J7J5J6J6K6I6J^2cMZ1eN2N100O1O100O1O100O1O1O2O0O1O100O1O100O1O101N1O100O1O100O1PN_E`Jb:`5aE]J_:b5fEYJ[:f5iEVJY:h5lETJT:k5PFQJQ:n5TFmIm9R6WFkIi9T6\\FgIe9Y6Q2O1O1O1O2N1O1O100O1000000000001O0000000OK6H8G9H8G9H8G9H8H8G9003M4L3M3M4L3M3M4L3M3M4M2M4L3N2M4N1O10cDVIb9i6WF\\Ik9d6oE`IS:`6fEdI^:Z6]EjIe:V6TEoIm:R6mDQJS;R6fDSJY;o5bDUJ];X7O001O1N101O1O1O001O1N101O1O1O001O1N101O1O1O001O1N101O1O1O001O1N3N1O2N1O2N1O2M2O2N1O1XLZFTMh9i2aFoL`9n2iFkLY9R3PGeLR9X3WGaLk8\\3^G\\Lc8a3fGXL\\8e3lGTLU8i3UHoKm7n3[HkKl7l3^HlKP8o2TE^LT3;W8V2YE\\Mj26\\8]1]E[N_21b8c0dEXOT2Mf8KiE6i1HT;5]4M3NlPV3"}, "label": "a man wearing a white shirt and gray pants and striped hat holding two baseball bats"}]}
{"id": 217293, "image": "COCO_train2014_000000217293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white long sleeved shirt holds two baseball bats\"."}], "task": "refering", "answer_template": "The \"a man in a white long sleeved shirt holds two baseball bats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 109, 110, 111, 124, 125, 126, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 273, 274, 275, 276, 277, 278, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 318, 319, 320, 321, 322, 323, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.14875878220140515, 0.29775, 0.614847775175644, 1.0], "iscrowd": 0, "area": 61628.216550000005, "rle": {"size": [640, 427], "counts": "n\\X15fc06J6J6K6I6J6J6K5J6J6J6J6K6I6J6J6J6K5J6J6J6K6I6M3N2O1N5K6K5J7I6K5J6J7J5J6J6K6I6J^2cMZ1eN2N100O1O100O1O100O1O1O2O0O1O100O1O100O1O101N1O100O1O100O1PN_E`Jb:`5aE]J_:b5fEYJ[:f5iEVJY:h5lETJT:k5PFQJQ:n5TFmIm9R6WFkIi9T6\\FgIe9Y6Q2O1O1O1O2N1O1O100O1000000000001O0000000OK6H8G9H8G9H8G9H8H8G9003M4L3M3M4L3M3M4L3M3M4M2M4L3N2M4N1O10cDVIb9i6WF\\Ik9d6oE`IS:`6fEdI^:Z6]EjIe:V6TEoIm:R6mDQJS;R6fDSJY;o5bDUJ];X7O001O1N101O1O1O001O1N101O1O1O001O1N101O1O1O001O1N101O1O1O001O1N3N1O2N1O2N1O2M2O2N1O1XLZFTMh9i2aFoL`9n2iFkLY9R3PGeLR9X3WGaLk8\\3^G\\Lc8a3fGXL\\8e3lGTLU8i3UHoKm7n3[HkKl7l3^HlKP8o2TE^LT3;W8V2YE\\Mj26\\8]1]E[N_21b8c0dEXOT2Mf8KiE6i1HT;5]4M3NlPV3"}, "label": "a man in a white long sleeved shirt holds two baseball bats"}]}
{"id": 12495, "image": "COCO_train2014_000000012495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with its head not visible but much of its body able to be seen\"."}], "task": "refering", "answer_template": "The \"a zebra with its head not visible but much of its body able to be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 254, 255, 256, 257, 258, 259, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 302, 303, 304, 305, 308, 309, 310, 311, 325, 326, 327, 328], "bbox": [0.0015, 0.0011294117647058823, 0.659578125, 0.9831529411764707], "iscrowd": 0, "area": 83423.39964999999, "rle": {"size": [425, 640], "counts": "Y=]6l64L4L4L4M2M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N101N1O1O001O001O1O001O001O1O001O001O1O001O001O1O1O5K4L5K4L5K4L5K4L5K4L4L5K4L5K4L1O000000000001O001O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O2O0O2N1O2N1O2N1O2N1O2N2N1O2N1O2N001O001O001O001O001O001O1O001O001O001ULkM`LU2]3WNZLj1b3cNTL]1h3oNPLR1l3ROSLn0i3WOULk0g3YOXLh0d3\\OZLe0b3@]La0_3C_L>`3D^L>_3E`L;^3H`L:^3HaL8]3L`L6]3MeKP1Y4SOYK[1e4gNYKZ1e4iNZKX1c4kN[KV1d4lN[KT1c4oN\\KQ1b4RO\\Ko0c4SO\\Km0b4VO\\Kl0b4VO]Kj0a4YO]Kh0b4ZO]Kf0a4]O^Kc0a4_O]Kb0b4A\\K?b4V4N1N2O1N2O1O2M2O1N2O2M2N2O2M2O1N2N3N1N2N2O2M2O1NZKiL_OT3b0mL^OR3a0oL@o2a0QM@n2`0RMBk2>WMBg2?ZMAe2>]MBa2?aMA\\2`0eM@Y2`0jM_OT2b0nM^Oo1b0SN^Ok1c0WN\\OLWKZ1^5kN[OJZKV1^5QOXOG^KS1[5WOWOEaKP1Y5]OUOBfKl0V5CUO_OiKi0T5IRO]OmKf0R5NQO\\OPLa0P55oNYOTL>n4:oNVOWL;l4`0lNTO\\L7i4f0kNRO_L3h4m0hNoNdLOe4S1U60O1O1O100O1O1O100O1O2N100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O2O0O1O1O100O1O1O100O10000O100O100O10O10O100O100O10000O100O100O10O10000000001O000000001O0000000O2O00000000001O000000001O000O1000O100O1O1O1O100O1O1O1O10O01O1O1O100O001O1O010O11N2O1O2M2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N5L3M3L5L3M3L5L3M4K_Tj2"}, "label": "a zebra with its head not visible but much of its body able to be seen"}]}
{"id": 12495, "image": "COCO_train2014_000000012495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra ' s head\"."}], "task": "refering", "answer_template": "The \"a zebra ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.28054687500000003, 0.0584235294117647, 1.0, 0.9820235294117646], "iscrowd": 0, "area": 118321.28239999998, "rle": {"size": [425, 640], "counts": "]oZ28o<9H9F9G5K5L4K5K5L4K5K5K5L4K5K3M3N1N3M2O2M2N3M2O2M3M2O2M2O2M2O2M2O2M2O2M3N1N3M2O2M2O2M2O2M2O2M3N1N100O100O101N100O100O1O2O0O100O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1[NWJjKj5Q4]JlKe5n3bJoK_5m3hJoKZ5k3mJRLT5j3SKRLn4n3TKoKn4P4UKlKl4T4VKiKk4X4WKdKk4[4WKbKj4^4YK^Kh4b4ZK[Kh4d4[KXKf4i4m1O00001O00010O00001O010O00001O01O01O001O00010O00001O01O010O10O01O10O01O10O01O100O010O1O010O1O10O0100O00100O00100O11N1000001O00001O00001O01O01O0000001O00001O00001O01O01O00O1O1O100O1O100O1O1O2O0O1O100O1O100O1O1O100O2N100O11O1O1O1O2N1O1O1O2N100O2N1O100O2N100O1O2O0O100O2N100O1O2O0O1O2O0O100O2N100O1O2O0O1O101N1O2O1N2O1O1O1N2O1O1O1N2O1O001N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O2N1N2O1O1O2M2O1O1O0O01O1O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O002N2N2N2N2N2N2N1O2N2N2N2N2N2N2N0O2O000000000000000000001O000000000WNi1]LiM"}, "label": "a zebra ' s head"}]}
{"id": 12495, "image": "COCO_train2014_000000012495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the front\"."}], "task": "refering", "answer_template": "The \"the zebra in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.28054687500000003, 0.0584235294117647, 1.0, 0.9820235294117646], "iscrowd": 0, "area": 118321.28239999998, "rle": {"size": [425, 640], "counts": "]oZ28o<9H9F9G5K5L4K5K5L4K5K5K5L4K5K3M3N1N3M2O2M2N3M2O2M3M2O2M2O2M2O2M2O2M2O2M3N1N3M2O2M2O2M2O2M2O2M3N1N100O100O101N100O100O1O2O0O100O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1[NWJjKj5Q4]JlKe5n3bJoK_5m3hJoKZ5k3mJRLT5j3SKRLn4n3TKoKn4P4UKlKl4T4VKiKk4X4WKdKk4[4WKbKj4^4YK^Kh4b4ZK[Kh4d4[KXKf4i4m1O00001O00010O00001O010O00001O01O01O001O00010O00001O01O010O10O01O10O01O10O01O100O010O1O010O1O10O0100O00100O00100O11N1000001O00001O00001O01O01O0000001O00001O00001O01O01O00O1O1O100O1O100O1O1O2O0O1O100O1O100O1O1O100O2N100O11O1O1O1O2N1O1O1O2N100O2N1O100O2N100O1O2O0O100O2N100O1O2O0O1O2O0O100O2N100O1O2O0O1O101N1O2O1N2O1O1O1N2O1O1O1N2O1O001N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O2N1N2O1O1O2M2O1O1O0O01O1O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O002N2N2N2N2N2N2N1O2N2N2N2N2N2N2N0O2O000000000000000000001O000000000WNi1]LiM"}, "label": "the zebra in the front"}]}
{"id": 12495, "image": "COCO_train2014_000000012495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra behind and to the right of the zebra in the foreground\"."}], "task": "refering", "answer_template": "The \"zebra behind and to the right of the zebra in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 38, 39, 40, 41, 62, 63, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 111, 112, 113, 114, 136, 137, 160], "bbox": [0.6640625, 0.003811764705882353, 1.0, 0.4464941176470588], "iscrowd": 0, "area": 12348.8907, "rle": {"size": [425, 640], "counts": "h]`51X=1N2N2N101N2N2N2O0O2N2N2O1N101O1N2O1N2O001N2O1N2O1O0O2O1N2O1O1N101N2O1O1N101N101O0O2O001N100O2O001N101N101O0O101N101O0O2O0O101O0O2O0O2O001N100O2O001N101O0O10YN_Ef0`:YOdEf0[:XOhEg0X:WOlEg0S:XOPFh0o9lN^FS1n:N1O2O0O1O2OO01O0010O01O010O000010O01O0010O01O01O010O010O010O2O1N1O2O1N2N2O1N2O1N2N100O10O01O10O0100O1O010O1O010O10O01O100O010O1O010O1O001O100O001O1O001O1O00010O001O00001O00010O00001O00001O00001O00001O001O2N3M2N3M1O1O2N1O1O1O1O2N1O1O1O1O2N1O10000O2O00M3G9F:F:F`M"}, "label": "zebra behind and to the right of the zebra in the foreground"}]}
{"id": 12495, "image": "COCO_train2014_000000012495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra behind two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 38, 39, 40, 41, 62, 63, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 111, 112, 113, 114, 136, 137, 160], "bbox": [0.6640625, 0.003811764705882353, 1.0, 0.4464941176470588], "iscrowd": 0, "area": 12348.8907, "rle": {"size": [425, 640], "counts": "h]`51X=1N2N2N101N2N2N2O0O2N2N2O1N101O1N2O1N2O001N2O1N2O1O0O2O1N2O1O1N101N2O1O1N101N101O0O2O001N100O2O001N101N101O0O101N101O0O2O0O101O0O2O0O2O001N100O2O001N101O0O10YN_Ef0`:YOdEf0[:XOhEg0X:WOlEg0S:XOPFh0o9lN^FS1n:N1O2O0O1O2OO01O0010O01O010O000010O01O0010O01O01O010O010O010O2O1N1O2O1N2N2O1N2O1N2N100O10O01O10O0100O1O010O1O010O10O01O100O010O1O010O1O001O100O001O1O001O1O00010O001O00001O00010O00001O00001O00001O00001O001O2N3M2N3M1O1O2N1O1O1O1O2N1O1O1O1O2N1O10000O2O00M3G9F:F:F`M"}, "label": "a zebra behind two other zebras"}]}
{"id": 274642, "image": "COCO_train2014_000000274642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with brown hair wearing a pink shirt\"."}], "task": "refering", "answer_template": "The \"a woman with brown hair wearing a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 23, 24, 25, 26, 36, 37, 38, 39, 40, 41, 42, 43, 44, 50, 51, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 94, 95, 96, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 201, 202, 203, 204, 216], "bbox": [0.22180327868852456, 0.025343749999999998, 0.9437704918032785, 0.6202343750000001], "iscrowd": 0, "area": 58915.79170000001, "rle": {"size": [640, 427], "counts": "ddk13ic03M4K4M4L3M4L3L5L3M4L3M4L4M2N3M2N3M2O2M2N3M2N3M2N3M2N3M3M4M2O2M2O1O2M2O2N1N3N1O1N3N1O2M2O1O2M2O2N1N2O2N1N3N1O2M2O1O2M2O2hN^KoBc4P=^KnBd4R=\\KmBf4Q=[KmBg4R=ZKmBh4Q=YKnBh4Q=XKnBj4Q=WKnBk4Q=UKnBl4Q=UKmBn4Q=SKoBm4P=TKoBn4o<SKQCm4o<SKQCm4n<SKRCo4l<RKTCn4e2iJh69bFP5b2iJm67aFP5^2mJP74aFP5\\2nJT72`FQ5X2PKX7O_FR5U2RK]7K_FV5n1RKc7H^FY5i1RKj7E]F[5d1SKo7B]F^5^1RKV8@[Fa5Y1SK\\8\\O[Fd5S1RKb8[OZFf5l0UKj8TO[Fi5d0WKR9POYFm5<YK[9jNYFo55[Kc9fNWFR6N^Kk9`NWFU6EaKT:ZNWFV6_OdK[:VNUF]8j9cGWF\\8j9dGUF]8k9cGUF\\8l9dGSF]8m9cGSF\\8n9dGQF]8o9cGQF]8o9bGQF^8P:bGPF^8o9cGQF\\8P:dGoE]8Q:cGoE]8Q:cGgEd8Z:a00O1000O010000O100O10000O010OO2O1O0O2O1O0O2O001O1N101O1N2O1O2M2O1O2M2O1O1O2M2O1O2M2O1O2M2O1O2M101O1N2O1O1O1O10000O0100000O10000O10000O01000O10000O10000OM4L4K4M4L4L5K4L4K5L4L4L5K4L4L4K5L4L4L5K4L4L4L4L4L5K4L4M3N2N2M4M2M3N2N2M3N3M2M4M3M3L4M3L4M3M2M4M3M3L4M3M3L4M2N3SKWA^4Y?L4L4K5L4L4K5L4K5L4L3L3O2O000O2O0O2O0hMo^O`1Sa0^NP_O`1Pa0`NR_O]1o`0bNT_O\\1m`0cNT_O[1m`0eNU_OY1l`0eNW_OX1j`0hNX_OV1h`0iNZ_OV1g`0iN[_OT1f`0lN\\_OR1e`0lN^_OQ1c`0oN^_OP1c`0nN`_Oo0a`0QOa_Om0_`0SOc_Oj0^`0UOd_Oj0[`0WOg_Of0Y`0ZOj_Od0V`0\\Ol_O`0U`0Al_O<U`0Dn_O9R`0HP@4Q`0LR@1o?OR@No?3S@Jm?6V@Fk?;l1010O1O0010O01O1O0010O01O1O0010O01O1O01N1Cin>"}, "label": "a woman with brown hair wearing a pink shirt"}]}
{"id": 274642, "image": "COCO_train2014_000000274642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing over a baby in a high chair\"."}], "task": "refering", "answer_template": "The \"a woman standing over a baby in a high chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 23, 24, 25, 26, 36, 37, 38, 39, 40, 41, 42, 43, 44, 50, 51, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 94, 95, 96, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 201, 202, 203, 204, 216], "bbox": [0.22180327868852456, 0.025343749999999998, 0.9437704918032785, 0.6202343750000001], "iscrowd": 0, "area": 58915.79170000001, "rle": {"size": [640, 427], "counts": "ddk13ic03M4K4M4L3M4L3L5L3M4L3M4L4M2N3M2N3M2O2M2N3M2N3M2N3M2N3M3M4M2O2M2O1O2M2O2N1N3N1O1N3N1O2M2O1O2M2O2N1N2O2N1N3N1O2M2O1O2M2O2hN^KoBc4P=^KnBd4R=\\KmBf4Q=[KmBg4R=ZKmBh4Q=YKnBh4Q=XKnBj4Q=WKnBk4Q=UKnBl4Q=UKmBn4Q=SKoBm4P=TKoBn4o<SKQCm4o<SKQCm4n<SKRCo4l<RKTCn4e2iJh69bFP5b2iJm67aFP5^2mJP74aFP5\\2nJT72`FQ5X2PKX7O_FR5U2RK]7K_FV5n1RKc7H^FY5i1RKj7E]F[5d1SKo7B]F^5^1RKV8@[Fa5Y1SK\\8\\O[Fd5S1RKb8[OZFf5l0UKj8TO[Fi5d0WKR9POYFm5<YK[9jNYFo55[Kc9fNWFR6N^Kk9`NWFU6EaKT:ZNWFV6_OdK[:VNUF]8j9cGWF\\8j9dGUF]8k9cGUF\\8l9dGSF]8m9cGSF\\8n9dGQF]8o9cGQF]8o9bGQF^8P:bGPF^8o9cGQF\\8P:dGoE]8Q:cGoE]8Q:cGgEd8Z:a00O1000O010000O100O10000O010OO2O1O0O2O1O0O2O001O1N101O1N2O1O2M2O1O2M2O1O1O2M2O1O2M2O1O2M2O1O2M101O1N2O1O1O1O10000O0100000O10000O10000O01000O10000O10000OM4L4K4M4L4L5K4L4K5L4L4L5K4L4L4K5L4L4L5K4L4L4L4L4L5K4L4M3N2N2M4M2M3N2N2M3N3M2M4M3M3L4M3L4M3M2M4M3M3L4M3M3L4M2N3SKWA^4Y?L4L4K5L4L4K5L4K5L4L3L3O2O000O2O0O2O0hMo^O`1Sa0^NP_O`1Pa0`NR_O]1o`0bNT_O\\1m`0cNT_O[1m`0eNU_OY1l`0eNW_OX1j`0hNX_OV1h`0iNZ_OV1g`0iN[_OT1f`0lN\\_OR1e`0lN^_OQ1c`0oN^_OP1c`0nN`_Oo0a`0QOa_Om0_`0SOc_Oj0^`0UOd_Oj0[`0WOg_Of0Y`0ZOj_Od0V`0\\Ol_O`0U`0Al_O<U`0Dn_O9R`0HP@4Q`0LR@1o?OR@No?3S@Jm?6V@Fk?;l1010O1O0010O01O1O0010O01O1O0010O01O1O01N1Cin>"}, "label": "a woman standing over a baby in a high chair"}]}
{"id": 274642, "image": "COCO_train2014_000000274642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy is touching his first birthday cake while his mother watches\"."}], "task": "refering", "answer_template": "The \"a young boy is touching his first birthday cake while his mother watches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 153, 154, 155, 156, 167, 168, 169, 170, 171, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 243, 244, 245, 246, 258, 259, 260, 274, 294, 295, 296, 297, 308, 309, 310, 311, 325, 326], "bbox": [0.1818735362997658, 0.41587500000000005, 0.8083606557377048, 0.9372343750000001], "iscrowd": 0, "area": 23928.40955, "rle": {"size": [640, 427], "counts": "iQa15ic06J6J5K6J6J5K6J5K4L4L4L3M4W_O_Om=e0oABj=b0QBEj=>RBIh=:TBLg=8SB0g=3UB3f=1UB6e=V3L2N2N2N2N2N3M2N2N2N2N2O100O1O100O1O1O100O1O100O2N1O01N1O100O1O2N1O100O1O2N1O100O1N3M2O1N2N2O1N3M2N2O1O1O2O0O1O101N2N101N2N1fMoAeNR>Z1oAdNS>Z1oAdNR>[1PBdNQ>Z1QBdNQ>Z1PBfNP>Y1RBeNP>Y1RBeNP>Y1PBhNQ>V1PBiNQ>V1oAjNS>T1nAlNS>R1mAnNT>Q1mAnNU>Q1jAPOW>n0jAQOW>e0ZAfM?e1^>>UAkM>g1`>:UAlM;j1e>4SAPN9k1i>0aA0c>L^A3g>VOkAj0Z>mNkAR1Y>eNlA[1Y>[NmAd1W`001O001O001O001O001O0O2O0O1O2O0O2N1O2O0O2N101N1O2O0O2N101N1010O01O010O010O010O010O0010O11O1N101O001N2O001N101O1O0O2O1O001N101O1NdAGS97nFJS94mFMS91nF0S9MnF4^>1O1O1N2O1O10000000000000O100000000000000000000000000004L5K2N1O2M3N2N2N2N2N2N2N1O2N2N2N1O0000000O10000O10O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O0O2O0O2O1O0O2O1N101N2_Oj\\O8Uc0Fo\\O9Qc0ES]O9lb0FW]O4WZc1"}, "label": "a young boy is touching his first birthday cake while his mother watches"}]}
{"id": 274642, "image": "COCO_train2014_000000274642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a naked child eats a birthday cake with his hands\"."}], "task": "refering", "answer_template": "The \"a naked child eats a birthday cake with his hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 153, 154, 155, 156, 167, 168, 169, 170, 171, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 243, 244, 245, 246, 258, 259, 260, 274, 294, 295, 296, 297, 308, 309, 310, 311, 325, 326], "bbox": [0.1818735362997658, 0.41587500000000005, 0.8083606557377048, 0.9372343750000001], "iscrowd": 0, "area": 23928.40955, "rle": {"size": [640, 427], "counts": "iQa15ic06J6J5K6J6J5K6J5K4L4L4L3M4W_O_Om=e0oABj=b0QBEj=>RBIh=:TBLg=8SB0g=3UB3f=1UB6e=V3L2N2N2N2N2N3M2N2N2N2N2O100O1O100O1O1O100O1O100O2N1O01N1O100O1O2N1O100O1O2N1O100O1N3M2O1N2N2O1N3M2N2O1O1O2O0O1O101N2N101N2N1fMoAeNR>Z1oAdNS>Z1oAdNR>[1PBdNQ>Z1QBdNQ>Z1PBfNP>Y1RBeNP>Y1RBeNP>Y1PBhNQ>V1PBiNQ>V1oAjNS>T1nAlNS>R1mAnNT>Q1mAnNU>Q1jAPOW>n0jAQOW>e0ZAfM?e1^>>UAkM>g1`>:UAlM;j1e>4SAPN9k1i>0aA0c>L^A3g>VOkAj0Z>mNkAR1Y>eNlA[1Y>[NmAd1W`001O001O001O001O001O0O2O0O1O2O0O2N1O2O0O2N101N1O2O0O2N101N1010O01O010O010O010O010O0010O11O1N101O001N2O001N101O1O0O2O1O001N101O1NdAGS97nFJS94mFMS91nF0S9MnF4^>1O1O1N2O1O10000000000000O100000000000000000000000000004L5K2N1O2M3N2N2N2N2N2N2N1O2N2N2N1O0000000O10000O10O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O0O2O0O2O1O0O2O1N101N2_Oj\\O8Uc0Fo\\O9Qc0ES]O9lb0FW]O4WZc1"}, "label": "a naked child eats a birthday cake with his hands"}]}
{"id": 282835, "image": "COCO_train2014_000000282835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small girl in a green shirt being held by her father\"."}], "task": "refering", "answer_template": "The \"a small girl in a green shirt being held by her father\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 123, 124, 125, 141, 142, 143, 158, 159, 160, 176, 177, 178, 193, 194, 195, 210, 211, 212, 227, 228, 229, 246], "bbox": [0.2876458333333333, 0.29625, 0.5093541666666667, 0.62884375], "iscrowd": 0, "area": 9482.04715, "rle": {"size": [640, 480], "counts": "c^f25ic09F9G7J3L3N3L4L4O1O1O2N1O1O1O1O1O001O1O6Ja0@2M100O1O100O100O100O2O1N2O2N0j_OSMY?l2d@YM[?e2d@_M\\?^2b@gMo0Ah<g2VBmMm0^Oo<c2SBSNh0\\OY=^2mAZNf0ZO_=[2hA`Nc0WOh=W2dAfN>VOQ>f3gA^L[>d3\\AbLe>[42O1N2O1N2O0O2O1N2\\Nf@`N[?^1i@^NY?_1k@_NV?_1m@^NU?_1o@^NR?a1RA\\Nn>b1VA[Nl>c1WAZNj>e1ZAWNg>g1]AVNe>h1^ATNd>l1_AmMe>S2]AgMg>Y2a1O00001O0002N4M2M3M4M21O2N2O0O1O0001O000f_O`MQ?a2Q@YNR`0i23M3M3M3L4M2N9F<D<E[1dN6J6J5K6JlSc4"}, "label": "a small girl in a green shirt being held by her father"}]}
{"id": 282835, "image": "COCO_train2014_000000282835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby in the green shorts\"."}], "task": "refering", "answer_template": "The \"the baby in the green shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 123, 124, 125, 141, 142, 143, 158, 159, 160, 176, 177, 178, 193, 194, 195, 210, 211, 212, 227, 228, 229, 246], "bbox": [0.2876458333333333, 0.29625, 0.5093541666666667, 0.62884375], "iscrowd": 0, "area": 9482.04715, "rle": {"size": [640, 480], "counts": "c^f25ic09F9G7J3L3N3L4L4O1O1O2N1O1O1O1O1O001O1O6Ja0@2M100O1O100O100O100O2O1N2O2N0j_OSMY?l2d@YM[?e2d@_M\\?^2b@gMo0Ah<g2VBmMm0^Oo<c2SBSNh0\\OY=^2mAZNf0ZO_=[2hA`Nc0WOh=W2dAfN>VOQ>f3gA^L[>d3\\AbLe>[42O1N2O1N2O0O2O1N2\\Nf@`N[?^1i@^NY?_1k@_NV?_1m@^NU?_1o@^NR?a1RA\\Nn>b1VA[Nl>c1WAZNj>e1ZAWNg>g1]AVNe>h1^ATNd>l1_AmMe>S2]AgMg>Y2a1O00001O0002N4M2M3M4M21O2N2O0O1O0001O000f_O`MQ?a2Q@YNR`0i23M3M3M3L4M2N9F<D<E[1dN6J6J5K6JlSc4"}, "label": "the baby in the green shorts"}]}
{"id": 282835, "image": "COCO_train2014_000000282835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a baby while on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man holding a baby while on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 74, 75, 76, 77, 92, 93, 94, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 158, 159, 161, 162, 163, 178, 179, 180, 194, 195, 196, 197, 211, 212, 213, 214, 228, 229, 230, 231, 245, 246, 247, 248, 262, 263, 264, 265, 279, 280, 281, 296, 297, 298, 313, 314, 315, 329, 330, 331, 332, 347, 348, 349], "bbox": [0.3223125, 0.141984375, 0.6190833333333333, 0.906015625], "iscrowd": 0, "area": 35042.51935000001, "rle": {"size": [640, 480], "counts": "STQ31oc01N2O1N2O1N3N1N2O1N2O1N2O1N2O1N2O0O100O100O100O100POTOk^Om0`7UOR10^Gk0Z7[OV1M_Gh0V7@Y1JbGe0U7AV1NeG`0U7BT10hG=U7BQ14jG9U7C`NDjKb0bN6U7D]NFiKb0fN3T7E\\NHgKc0iNOU7EZNLcKc0oNKT7FYNNaKd0ROGU7FVN2`Kc0P8\\O_L3^Kc0R8[O_L4\\Kd0e1mNP3<N5ZKd0h1mNWO:f00P37XKe0j1mNeNj0U1^OS38WKd0l1nNRN\\1d1kNV39VKc0n1oNlMb1e1eNX3:XKa0n1QOgMe1g1`N[3;WK`0P2RObMj1g1ZN_3<VK`0Q2RO^Mn1h1UNb3>TK>T2TOYMQ2h1PNf3`0SK;V2VOTMU2i1kMi3c0PK9Y2VOPMZ2i1eMm3f0mJ6\\2YOkL\\2l1_Mn3k0kJ3_2f2cN]LR4n0iJOb2i2_N[LU4Q1fJMe2j2\\NYLX4T1cJJk2k2UNYL[4V1aJGQ3k2oMYL^4Y1_JDU3l2iMXLa4Z3hM?cMXLd4[3jM?]MWLi4Z3kMa0YMULl4Z3mMc0RMTL9nNm3\\4iNc0mLTL9QOQ4Z4iNc0hLSL;SOT4Z4hNa0eLSL;UOY4Z4fNg1lLSJ^4Z4dNa1jLYJb4Y4cN_LmLZ3JRLf4X4bNPLXM_4V4DaN`KdMl4k3G`NQKoMX5a3JI68NC3=0_O1b02YOOg06SOKm0:mNGS1=hNDX1a0cN_O]1f0]NZOd1j0WNWOi1n0QNSOo1W1fMjNZ2b1YM_Ng2j1oLWNQ3R2eLoM[3[2ZLfMf3c2PL]MS4j2cKWM^4Q3XKPMi4S92N1O1O2M2O1O2N1O1O2N1O1N3N9G=C<D3M1O2M2O1\\LVJPIl5n6]KiGd4V8d3O2N1O2N1O1O2N1O1O2N1O1O2N1M4K4K5L5J5L4L5J5L4K6[Od0[Of0YOf0[O]2bMhTb3"}, "label": "a man holding a baby while on a skateboard"}]}
{"id": 282835, "image": "COCO_train2014_000000282835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white baseball hat\"."}], "task": "refering", "answer_template": "The \"a man wearing a white baseball hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 74, 75, 76, 77, 92, 93, 94, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 158, 159, 161, 162, 163, 178, 179, 180, 194, 195, 196, 197, 211, 212, 213, 214, 228, 229, 230, 231, 245, 246, 247, 248, 262, 263, 264, 265, 279, 280, 281, 296, 297, 298, 313, 314, 315, 329, 330, 331, 332, 347, 348, 349], "bbox": [0.3223125, 0.141984375, 0.6190833333333333, 0.906015625], "iscrowd": 0, "area": 35042.51935000001, "rle": {"size": [640, 480], "counts": "STQ31oc01N2O1N2O1N3N1N2O1N2O1N2O1N2O1N2O0O100O100O100O100POTOk^Om0`7UOR10^Gk0Z7[OV1M_Gh0V7@Y1JbGe0U7AV1NeG`0U7BT10hG=U7BQ14jG9U7C`NDjKb0bN6U7D]NFiKb0fN3T7E\\NHgKc0iNOU7EZNLcKc0oNKT7FYNNaKd0ROGU7FVN2`Kc0P8\\O_L3^Kc0R8[O_L4\\Kd0e1mNP3<N5ZKd0h1mNWO:f00P37XKe0j1mNeNj0U1^OS38WKd0l1nNRN\\1d1kNV39VKc0n1oNlMb1e1eNX3:XKa0n1QOgMe1g1`N[3;WK`0P2RObMj1g1ZN_3<VK`0Q2RO^Mn1h1UNb3>TK>T2TOYMQ2h1PNf3`0SK;V2VOTMU2i1kMi3c0PK9Y2VOPMZ2i1eMm3f0mJ6\\2YOkL\\2l1_Mn3k0kJ3_2f2cN]LR4n0iJOb2i2_N[LU4Q1fJMe2j2\\NYLX4T1cJJk2k2UNYL[4V1aJGQ3k2oMYL^4Y1_JDU3l2iMXLa4Z3hM?cMXLd4[3jM?]MWLi4Z3kMa0YMULl4Z3mMc0RMTL9nNm3\\4iNc0mLTL9QOQ4Z4iNc0hLSL;SOT4Z4hNa0eLSL;UOY4Z4fNg1lLSJ^4Z4dNa1jLYJb4Y4cN_LmLZ3JRLf4X4bNPLXM_4V4DaN`KdMl4k3G`NQKoMX5a3JI68NC3=0_O1b02YOOg06SOKm0:mNGS1=hNDX1a0cN_O]1f0]NZOd1j0WNWOi1n0QNSOo1W1fMjNZ2b1YM_Ng2j1oLWNQ3R2eLoM[3[2ZLfMf3c2PL]MS4j2cKWM^4Q3XKPMi4S92N1O1O2M2O1O2N1O1O2N1O1N3N9G=C<D3M1O2M2O1\\LVJPIl5n6]KiGd4V8d3O2N1O2N1O1O2N1O1O2N1O1O2N1M4K4K5L5J5L4L5J5L4K6[Od0[Of0YOf0[O]2bMhTb3"}, "label": "a man wearing a white baseball hat"}]}
{"id": 356569, "image": "COCO_train2014_000000356569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt\"."}], "task": "refering", "answer_template": "The \"man in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 31, 32, 55, 100, 101, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 287, 288], "bbox": [0.306046875, 0.0022500000000000003, 0.664234375, 0.7107708333333332], "iscrowd": 0, "area": 24664.1542, "rle": {"size": [480, 640], "counts": "SRl22m>3L3N3M3M3L3N3M3M2N3L4M3M2N3L4M3M2N3L4M3M2N3M3L3kJhM_M[2_2fM_M]2_2dM_M_2_2aM_Mb2`2_M^Md2`2]M^Mf2`2[M^Mh2`2YM^Mi2b2WM[Mm2c2TM[Mo2c2RM[MQ3c2PM[MR3d2nLZMV3d2kLZMX3d2iLZMY3e2hLYM[3e2fLXM^3f2cLXM`3f2aLXMa3g2`LWMc3g2^LWMe3g2\\LVMh3h2[LTMg3k2[LRMh3l2[LPMh3n2[LmLh3R3[LjLh3T3ZLiLi3U3ZLgLi3W3ZLeLh3Z3ZLbLj3\\3YL`Lj3^3YL^Lj3a3WL\\Lk3c3XLXLl3g3VLVLl3i3VLVLj3h3XLXLg3h3[LWLe3g3^LWLc3h3^LVLc3i3`LULa3j3aLTL`3j3bLUL^3k3dLRL\\3n3dLSL\\3m3cLTL\\3l3cLVL]3k3aLUL_3n3]LTLc3n3YLTLf3P4VLQLj3Q4RLPLo3R4mKPLS4R4jKoKV4m4lJUKT5Y5\\JhJe5Z700001N10WLhFi1X9TNkFl1U9RNmF_N4l2n8cNdG]1\\8aNeG`1[8]NgGd1Y8ZNhGh1V8WNkGj1U8UNmGj1S8TNoGk1R8SNQHk1o7UNRHj1o7VNQHj1o7UNRHj1o7VNQHi1o7XNPHh1Q8WNPHh1Q8XNnGh1S8WNnGh1R8YNmGg1T8XNmGh1S8XNlGh1U8XNkGg1U8YNkGg1V8YNkGe1V8ZNkGe1V8[NkGc1U8^NlGa1T8^NmGa1T8_NlG`1U8_NkGa1U8`NjG`1W8`NiG_1X8`NhG`1Y8`NfG`1Z8`NfG`1[8`NdG`1]8`NbG`1_8_NbG`1^8aNaG_1`8`N`G`1a8`N^G`1c8_N]Ga1c8_N^G`1c8`N\\G`1e8_N[Ga1e8`NZG`1g8_NYGa1g8_NZG`1f8aNZG^1g8aNZG^1f8bN[G]1f8cNZG\\1f8dNZG\\1f8eNZGZ1g8eNZGZ1f8gNZGX1g8gNYGY1g8hNYGW1g8iNZGV1g8jNXGV1h8jNYGU1h8kNXGT1h8lNXGT1h8mNXGR1i8mNWGS1i8nNWGQ1j8nNWGQ1i8POWGn0k8QOUGo0k8ROUGl0l8TOUGk0l8UOTGj0l8VOUGh0m8XOTGf0l8ZOVGd0j8]OVGa0l8^OVG`0j8AVG>k8AWG<j8DWG;i8EWG;j8DWG:j8GUG9l8FUG8l8HUG7k8IUG6m8ITG5m8LRG4o8KRG3o8MRG1P9NQG1o8OQG0P91PGNQ91PGMQ93PGKQ95oFKR94oFJR96oFHS97PGBT9>PGXOV9h0b2O000O2O001O000O10000000000O100000O10O1000O1000000O10O1000O1003M3M2N3M2M2N010O1M2JcbT3"}, "label": "man in white shirt"}]}
{"id": 356569, "image": "COCO_train2014_000000356569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt at a desk\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt at a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 31, 32, 55, 100, 101, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 287, 288], "bbox": [0.306046875, 0.0022500000000000003, 0.664234375, 0.7107708333333332], "iscrowd": 0, "area": 24664.1542, "rle": {"size": [480, 640], "counts": "SRl22m>3L3N3M3M3L3N3M3M2N3L4M3M2N3L4M3M2N3L4M3M2N3M3L3kJhM_M[2_2fM_M]2_2dM_M_2_2aM_Mb2`2_M^Md2`2]M^Mf2`2[M^Mh2`2YM^Mi2b2WM[Mm2c2TM[Mo2c2RM[MQ3c2PM[MR3d2nLZMV3d2kLZMX3d2iLZMY3e2hLYM[3e2fLXM^3f2cLXM`3f2aLXMa3g2`LWMc3g2^LWMe3g2\\LVMh3h2[LTMg3k2[LRMh3l2[LPMh3n2[LmLh3R3[LjLh3T3ZLiLi3U3ZLgLi3W3ZLeLh3Z3ZLbLj3\\3YL`Lj3^3YL^Lj3a3WL\\Lk3c3XLXLl3g3VLVLl3i3VLVLj3h3XLXLg3h3[LWLe3g3^LWLc3h3^LVLc3i3`LULa3j3aLTL`3j3bLUL^3k3dLRL\\3n3dLSL\\3m3cLTL\\3l3cLVL]3k3aLUL_3n3]LTLc3n3YLTLf3P4VLQLj3Q4RLPLo3R4mKPLS4R4jKoKV4m4lJUKT5Y5\\JhJe5Z700001N10WLhFi1X9TNkFl1U9RNmF_N4l2n8cNdG]1\\8aNeG`1[8]NgGd1Y8ZNhGh1V8WNkGj1U8UNmGj1S8TNoGk1R8SNQHk1o7UNRHj1o7VNQHj1o7UNRHj1o7VNQHi1o7XNPHh1Q8WNPHh1Q8XNnGh1S8WNnGh1R8YNmGg1T8XNmGh1S8XNlGh1U8XNkGg1U8YNkGg1V8YNkGe1V8ZNkGe1V8[NkGc1U8^NlGa1T8^NmGa1T8_NlG`1U8_NkGa1U8`NjG`1W8`NiG_1X8`NhG`1Y8`NfG`1Z8`NfG`1[8`NdG`1]8`NbG`1_8_NbG`1^8aNaG_1`8`N`G`1a8`N^G`1c8_N]Ga1c8_N^G`1c8`N\\G`1e8_N[Ga1e8`NZG`1g8_NYGa1g8_NZG`1f8aNZG^1g8aNZG^1f8bN[G]1f8cNZG\\1f8dNZG\\1f8eNZGZ1g8eNZGZ1f8gNZGX1g8gNYGY1g8hNYGW1g8iNZGV1g8jNXGV1h8jNYGU1h8kNXGT1h8lNXGT1h8mNXGR1i8mNWGS1i8nNWGQ1j8nNWGQ1i8POWGn0k8QOUGo0k8ROUGl0l8TOUGk0l8UOTGj0l8VOUGh0m8XOTGf0l8ZOVGd0j8]OVGa0l8^OVG`0j8AVG>k8AWG<j8DWG;i8EWG;j8DWG:j8GUG9l8FUG8l8HUG7k8IUG6m8ITG5m8LRG4o8KRG3o8MRG1P9NQG1o8OQG0P91PGNQ91PGMQ93PGKQ95oFKR94oFJR96oFHS97PGBT9>PGXOV9h0b2O000O2O001O000O10000000000O100000O10O1000O1000000O10O1000O1003M3M2N3M2M2N010O1M2JcbT3"}, "label": "a man in a white shirt at a desk"}]}
{"id": 356569, "image": "COCO_train2014_000000356569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a green computer\"."}], "task": "refering", "answer_template": "The \"a man holding a green computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.0, 0.618546875, 0.9865208333333333], "iscrowd": 0, "area": 109890.04254999994, "rle": {"size": [480, 640], "counts": "c;f0f=l1mIgM[LT4Y2gL]Ln3`3jLVKj3h4X3N2N2O1O1O1O2N1O1O1N2O1O100O100O100O1O100O100O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O100O1O1O100O1O1O100O0010000O100000000O100O1L4L4M3O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1N2N2N2N2N2N2N2M3M3L4L4L4M3N2M3N1N3M3N2M3N2M3N2M3N2M3J6N2N200O1O100O100O1O100O100O1O100O1O1M300O1O1O100O1O100000000000000000000000000000O100000000000000000000000001O0000000000001O0000000000001XGFh2:XMGg29XMIg27XMKg25YMLf24YMOe21[M0d21ZM1e2O[M3c2M\\M5c2K\\M8b2H^M:`2F_M=_2CaM`0\\2AbMc0[2]OeMf0X2ZOgMk0U2UOjMP1R2POnMQ1Q2oNnMS1Q2mNoMU1o1lNoMV1P2jNoMY1o1gNoM]1n1dNPN`1n1`NPNc1o1]NPNf1n1[NoMi1o1WNoMm1o1TNnMo1Q2RNlMR2R2oMkMT2T2mMeM[2Y2eMgM\\2X2eMfM_2W2bMgMb2V2_MiMc2U2]MjMf2T2[MkMg2S2ZMkMi2S2XMkMk2S2VMlMk2S2UMlMn2R2SMmMn2R2SMlMo2S2RMkMP3T2QM`M\\3^2dLaM^3^2cL`M_3_2bL_M`3`2aL^Ma3a2_L^Md3`2_L\\Md3b2_LZMc3e2`LWMc3g2`LUMc3i2`LSMb3l2aLhL]NkMU5[5bLdLi3[3ZL_Lk3_3XL\\Ll3b3WLYLl3f3WLVLl3h3WLTLl3j3WLRLk3m3XLoKk3o3XLmKj3R4ZLiKi3U4ZLgKh3X4[LdKg3[4\\L]Kj3b4[3O1O1O1N2O1O1O1O1O1nMdE^O]:a0cE_O^:`0bE@_:?aEA`:>`EBa:=_ECb:<^EDc::^EFd:8\\EHe:7[EIf:6ZEJh:4YEKh:4XELi:3WEMi:3WEMj:2VENj:2VENj:1WEOj:0VE0j:0VE0j:0VE0j:0VE0k:OVE0j:0VE0j:0VE0j:0VE0k:OUE1k:NVE2j:NWE1i:OWE1i:OWE1j:NVE2j:NVE2j:NWE1i:OWE1i:OWE1j:MWE3i:MWE3i:MXE2h:NXE2h:NXE2i:MWE3i:MWE3i:MXE2h:NXE2h:NXE1i:NXE2h:NYEOi:1WENj:2VELl:4UEJl:6TEIm:7TEFn::SEDn:<SEAo:?QE@P;?RE_Oo:a0RE^On:b0SE^Om:a0TE^Ol:b0UE]Ok:c0WE[Oj:d0WE[Oi:e0YEYOh:f0YEYOg:f0\\EXOd:h0^EWOb:h0`EVO`:j0aEUO`:j0bETO^:l0dERO]:m0dERO\\:n0fEPOZ:P1hEnNY:Q1hEnNX:Q1kEmNV:R1lElNT:T1nEiNS:W1mEiNS:W1nEgNS:Y1nEfNR:Z1j1000000O1000000000001O0001O01O001O001O010O001O001O0oNdB>o=00O2N100O2N1O1O2N1O2O0O2N2NQ_b3"}, "label": "a man holding a green computer"}]}
{"id": 356569, "image": "COCO_train2014_000000356569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a older man who is on the computer\"."}], "task": "refering", "answer_template": "The \"a older man who is on the computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.0, 0.618546875, 0.9865208333333333], "iscrowd": 0, "area": 109890.04254999994, "rle": {"size": [480, 640], "counts": "c;f0f=l1mIgM[LT4Y2gL]Ln3`3jLVKj3h4X3N2N2O1O1O1O2N1O1O1N2O1O100O100O100O1O100O100O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O100O1O1O100O1O1O100O0010000O100000000O100O1L4L4M3O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1N2N2N2N2N2N2N2M3M3L4L4L4M3N2M3N1N3M3N2M3N2M3N2M3N2M3J6N2N200O1O100O100O1O100O100O1O100O1O1M300O1O1O100O1O100000000000000000000000000000O100000000000000000000000001O0000000000001O0000000000001XGFh2:XMGg29XMIg27XMKg25YMLf24YMOe21[M0d21ZM1e2O[M3c2M\\M5c2K\\M8b2H^M:`2F_M=_2CaM`0\\2AbMc0[2]OeMf0X2ZOgMk0U2UOjMP1R2POnMQ1Q2oNnMS1Q2mNoMU1o1lNoMV1P2jNoMY1o1gNoM]1n1dNPN`1n1`NPNc1o1]NPNf1n1[NoMi1o1WNoMm1o1TNnMo1Q2RNlMR2R2oMkMT2T2mMeM[2Y2eMgM\\2X2eMfM_2W2bMgMb2V2_MiMc2U2]MjMf2T2[MkMg2S2ZMkMi2S2XMkMk2S2VMlMk2S2UMlMn2R2SMmMn2R2SMlMo2S2RMkMP3T2QM`M\\3^2dLaM^3^2cL`M_3_2bL_M`3`2aL^Ma3a2_L^Md3`2_L\\Md3b2_LZMc3e2`LWMc3g2`LUMc3i2`LSMb3l2aLhL]NkMU5[5bLdLi3[3ZL_Lk3_3XL\\Ll3b3WLYLl3f3WLVLl3h3WLTLl3j3WLRLk3m3XLoKk3o3XLmKj3R4ZLiKi3U4ZLgKh3X4[LdKg3[4\\L]Kj3b4[3O1O1O1N2O1O1O1O1O1nMdE^O]:a0cE_O^:`0bE@_:?aEA`:>`EBa:=_ECb:<^EDc::^EFd:8\\EHe:7[EIf:6ZEJh:4YEKh:4XELi:3WEMi:3WEMj:2VENj:2VENj:1WEOj:0VE0j:0VE0j:0VE0j:0VE0k:OVE0j:0VE0j:0VE0j:0VE0k:OUE1k:NVE2j:NWE1i:OWE1i:OWE1j:NVE2j:NVE2j:NWE1i:OWE1i:OWE1j:MWE3i:MWE3i:MXE2h:NXE2h:NXE2i:MWE3i:MWE3i:MXE2h:NXE2h:NXE1i:NXE2h:NYEOi:1WENj:2VELl:4UEJl:6TEIm:7TEFn::SEDn:<SEAo:?QE@P;?RE_Oo:a0RE^On:b0SE^Om:a0TE^Ol:b0UE]Ok:c0WE[Oj:d0WE[Oi:e0YEYOh:f0YEYOg:f0\\EXOd:h0^EWOb:h0`EVO`:j0aEUO`:j0bETO^:l0dERO]:m0dERO\\:n0fEPOZ:P1hEnNY:Q1hEnNX:Q1kEmNV:R1lElNT:T1nEiNS:W1mEiNS:W1nEgNS:Y1nEfNR:Z1j1000000O1000000000001O0001O01O001O001O010O001O001O0oNdB>o=00O2N100O2N1O1O2N1O2O0O2N2NQ_b3"}, "label": "a older man who is on the computer"}]}
{"id": 340197, "image": "COCO_train2014_000000340197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing red dress and she taking something\"."}], "task": "refering", "answer_template": "The \"a woman wearing red dress and she taking something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 128, 129, 142, 143, 144, 145, 158, 159, 160, 173, 174, 175, 176], "bbox": [0.48098360655737704, 0.318609375, 0.7448711943793911, 0.52596875], "iscrowd": 0, "area": 6072.638849999998, "rle": {"size": [640, 427], "counts": "`\\P41ec0:H80O2O1O0000O10000000O100000000000O10O100000000000000000O100O1VO@S^Oa0ja0CT^O>ja0ET^O<ja0GT^O:ia0KT^O5ka0OR^O2ka02S^OOla04Q^OMoa04o]OLRb06k]OKTb0Q1g_O\\NV>d1dAbN[>^1`AhN`>X1XAPOh>P1n@ZOQ?g0m@[OS?d0m@]OR?d0l@^OT?b0j@@U?`0k@@V?`0h@AY??e@A\\?`0b@A_??`@@d?>Z@Ch?=U@Cn?<P@ER`0;l_OEV`0;h_OEZ`0:e_OE^`0;`_OEc`09\\_OGf`09X_OGj`0g11O2N2O1N2N3M2N2O1N2O1N101N2O0O2O1N1O2O0O2O1N101N2O1N210O01OO2Mh0XO1N2O1O2M2O1O1N2O2M2O1O4K6K6JiiS2"}, "label": "a woman wearing red dress and she taking something"}]}
{"id": 340197, "image": "COCO_train2014_000000340197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with red dress playing in the field\"."}], "task": "refering", "answer_template": "The \"a woman with red dress playing in the field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 128, 129, 142, 143, 144, 145, 158, 159, 160, 173, 174, 175, 176], "bbox": [0.48098360655737704, 0.318609375, 0.7448711943793911, 0.52596875], "iscrowd": 0, "area": 6072.638849999998, "rle": {"size": [640, 427], "counts": "`\\P41ec0:H80O2O1O0000O10000000O100000000000O10O100000000000000000O100O1VO@S^Oa0ja0CT^O>ja0ET^O<ja0GT^O:ia0KT^O5ka0OR^O2ka02S^OOla04Q^OMoa04o]OLRb06k]OKTb0Q1g_O\\NV>d1dAbN[>^1`AhN`>X1XAPOh>P1n@ZOQ?g0m@[OS?d0m@]OR?d0l@^OT?b0j@@U?`0k@@V?`0h@AY??e@A\\?`0b@A_??`@@d?>Z@Ch?=U@Cn?<P@ER`0;l_OEV`0;h_OEZ`0:e_OE^`0;`_OEc`09\\_OGf`09X_OGj`0g11O2N2O1N2N3M2N2O1N2O1N101N2O0O2O1N1O2O0O2O1N101N2O1N210O01OO2Mh0XO1N2O1O2M2O1O1N2O2M2O1O4K6K6JiiS2"}, "label": "a woman with red dress playing in the field"}]}
{"id": 340197, "image": "COCO_train2014_000000340197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green shirt and grey pants\"."}], "task": "refering", "answer_template": "The \"a man in a green shirt and grey pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 117, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 164, 177, 178, 179, 192, 193, 194, 207, 208, 209, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 327, 328, 329, 343, 344], "bbox": [0.6969789227166276, 0.286125, 1.0, 0.9881562500000001], "iscrowd": 0, "area": 28692.460150000006, "rle": {"size": [640, 427], "counts": "jej5b0\\c03L4M4K4M3N2N2N2N2N2N1O1OM4K4M4L3K6J5O2O000O0100O101O0O101N10001O00001N10001O00001O0000001O000UASOd:n0TE\\Oi:d0PEFm::lD0Q;1gD9W;FcDc0Z;]O_Dm0^;TOZDV1c;jNVD`1g;`NSDi1[;gN]Dc1\\;aN[Dj1_;ZNYDQ2_;UNXDV2^;QNZDZ2k7mL]IQ1_N^2o7dL[IT1`Na2Q8]LYIX1eN^2n7\\LVI\\1kN[2l7[LSI_1POY2i7[LQIb1TOV2g7[LRIa1WOV2e7ZLRIa1XOX2d7XLRIb1XOX2f7VLPIc1YOZ2e7TLoHd1\\OZ2c7TLnHd1]O[2d7RLlHg7R7[HjHi7T7XHiHl7U7VHgHn7W7THeHP8Z7^2L6J7I6J7I6J3M1O1O1O1O100O100001O0000000001O[NSJWEm5f:WJYEi5c:\\J\\Ed5a:_J_Ea5]:dJbE]5Z:gJeEZ5V:jJjEX5Q:lJnEU5m9PKRFR5h9RKXFo4d9UK[Fm4_9XK`Fj4[9YKeFi4U9\\KjFg4P9]KoFf4j8^KVGf4c8^K\\Go5S7VJlHk5P7XJPIi5l6[JSIf5j6]JUId5g6_JYIb5d6aJ[I`5a6dJ^I]5_6eJaI\\5[6hJdIY5Y6jJ\\Ia5`6bJRIeM"}, "label": "a man in a green shirt and grey pants"}]}
{"id": 340197, "image": "COCO_train2014_000000340197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is standing on grass and watching a woman catch a frisbee\"."}], "task": "refering", "answer_template": "The \"a man is standing on grass and watching a woman catch a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 117, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 164, 177, 178, 179, 192, 193, 194, 207, 208, 209, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 327, 328, 329, 343, 344], "bbox": [0.6969789227166276, 0.286125, 1.0, 0.9881562500000001], "iscrowd": 0, "area": 28692.460150000006, "rle": {"size": [640, 427], "counts": "jej5b0\\c03L4M4K4M3N2N2N2N2N2N1O1OM4K4M4L3K6J5O2O000O0100O101O0O101N10001O00001N10001O00001O0000001O000UASOd:n0TE\\Oi:d0PEFm::lD0Q;1gD9W;FcDc0Z;]O_Dm0^;TOZDV1c;jNVD`1g;`NSDi1[;gN]Dc1\\;aN[Dj1_;ZNYDQ2_;UNXDV2^;QNZDZ2k7mL]IQ1_N^2o7dL[IT1`Na2Q8]LYIX1eN^2n7\\LVI\\1kN[2l7[LSI_1POY2i7[LQIb1TOV2g7[LRIa1WOV2e7ZLRIa1XOX2d7XLRIb1XOX2f7VLPIc1YOZ2e7TLoHd1\\OZ2c7TLnHd1]O[2d7RLlHg7R7[HjHi7T7XHiHl7U7VHgHn7W7THeHP8Z7^2L6J7I6J7I6J3M1O1O1O1O100O100001O0000000001O[NSJWEm5f:WJYEi5c:\\J\\Ed5a:_J_Ea5]:dJbE]5Z:gJeEZ5V:jJjEX5Q:lJnEU5m9PKRFR5h9RKXFo4d9UK[Fm4_9XK`Fj4[9YKeFi4U9\\KjFg4P9]KoFf4j8^KVGf4c8^K\\Go5S7VJlHk5P7XJPIi5l6[JSIf5j6]JUId5g6_JYIb5d6aJ[I`5a6dJ^I]5_6eJaI\\5[6hJdIY5Y6jJ\\Ia5`6bJRIeM"}, "label": "a man is standing on grass and watching a woman catch a frisbee"}]}
{"id": 561384, "image": "COCO_train2014_000000561384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a calf standing against a wall\"."}], "task": "refering", "answer_template": "The \"a calf standing against a wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 241, 242, 243, 244, 264], "bbox": [0.458609375, 0.38686182669789226, 0.6371875, 0.7894379391100703], "iscrowd": 0, "area": 11372.039899999994, "rle": {"size": [427, 640], "counts": "[ij32W=3N2N3M2N3M2iDA^9c0ZFCe9k0hEZOX:X1QEnNn:g101O3L4M3M3M3M4[GTMe6n2UIVMk6l2oHXMP7k2iHZMW7h2cH\\M]7g2\\H]Mc7o2nGUMR8R400O010O10000O10O1I6J7K50O01000O01jN]GdMc8Y2iG^MV8`2VHWMj7f2_1M2N3M3N2M3M3M3000O2O001O001QGbMo6^2oHdMQ7]2lHeMS7\\2kHfMU7[2hHgMX7Y2fHiMZ7X2cHiM^7W2`HkM`7V2]HlMc7T2[HnMe7S2XHoMg7R2WHPNi7Q2lGYNT8g1[GiNf8m21O1O1N2O001N2\\OPGlLQ9o2WGmLj8o2]GlLf8o2bGlL_8o2R1K3J7J6O1O001O1O001O1O0013N3M2O1oF`MR7f2aH`M`7g2RH_Mo7d2gGbMZ8a2[GfMf8]2PGiMQ9X31N2O1dLkF`2V9WMZGa2P:H8GZ1gNR1nNX^P3"}, "label": "a calf standing against a wall"}]}
{"id": 561384, "image": "COCO_train2014_000000561384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow that is behind other\"."}], "task": "refering", "answer_template": "The \"cow that is behind other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 241, 242, 243, 244, 264], "bbox": [0.458609375, 0.38686182669789226, 0.6371875, 0.7894379391100703], "iscrowd": 0, "area": 11372.039899999994, "rle": {"size": [427, 640], "counts": "[ij32W=3N2N3M2N3M2iDA^9c0ZFCe9k0hEZOX:X1QEnNn:g101O3L4M3M3M3M4[GTMe6n2UIVMk6l2oHXMP7k2iHZMW7h2cH\\M]7g2\\H]Mc7o2nGUMR8R400O010O10000O10O1I6J7K50O01000O01jN]GdMc8Y2iG^MV8`2VHWMj7f2_1M2N3M3N2M3M3M3000O2O001O001QGbMo6^2oHdMQ7]2lHeMS7\\2kHfMU7[2hHgMX7Y2fHiMZ7X2cHiM^7W2`HkM`7V2]HlMc7T2[HnMe7S2XHoMg7R2WHPNi7Q2lGYNT8g1[GiNf8m21O1O1N2O001N2\\OPGlLQ9o2WGmLj8o2]GlLf8o2bGlL_8o2R1K3J7J6O1O001O1O001O1O0013N3M2O1oF`MR7f2aH`M`7g2RH_Mo7d2gGbMZ8a2[GfMf8]2PGiMQ9X31N2O1dLkF`2V9WMZGa2P:H8GZ1gNR1nNX^P3"}, "label": "cow that is behind other"}]}
{"id": 561384, "image": "COCO_train2014_000000561384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the goat in front of the other goat\"."}], "task": "refering", "answer_template": "The \"the goat in front of the other goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 262, 263, 279, 280, 281, 282, 283, 285, 286, 302, 305, 306, 309], "bbox": [0.1363125, 0.4666744730679157, 0.477796875, 0.9299063231850119], "iscrowd": 0, "area": 24065.98585000001, "rle": {"size": [427, 640], "counts": "jbT12W=7I6J6J7I6L5K4L5K3M3aD[NS;g1jDZNT;k1gDWNV;Q2N2M3M2O0O2O0O2N2O0O2N101N2N101N1O2O1N100O1O010O1O10O01O100O1O010O100O0010O01O01O100O1O100O100O1O11O010O001O001O00001O001O0100O1O_FWMe0Ga6R3cHaMg0_Oe6P3^HlMg0VOk6n2WHVNi0nNP7Z4kHhKU7\\4eHfK[7U5O2O0000000O10001O000O01L4K5K5K4N3M3M3M3M2N3_NWG]Nl8c1[GSNh8l1`GPN]8P2jGTNj7l1ZHUNb7k1aHSN^7m1dHRN[7m1hHRNW7n1kHQNS7P2PInMo6Q2TInMk6R2WImMh6R2\\IlMc6T2_IiMa6X2aIdMa6[2bIaM`6_2dI[M^6d2fIWM\\6i2R210O10000O010O10000O01000O1000O100000000O10000000O0100000O10O10O100O10O10O100O1000O0100001O1N2O1O001O1dG[Mo5g2lI\\MU6e2gI]MZ6d2eHUMN:]7b2`H`MH0i7a2ZH`Nf7b1UHaNl7_1QHcNP8[31O2O0O1O1M3M3M3M3M3M4MoGiKR7T4oHSLl6j3UIWLl6f3UI\\Lk6a3VIaLj6b3QI_LP7d3kH^LU7f3dH\\L]7g3^H[Lc7g3XHZLi7d40O100O2O2K4L5K5K4cL^Gk1h8PNPHm0`8nNR3K8G9H9GdQ[4"}, "label": "the goat in front of the other goat"}]}
{"id": 561384, "image": "COCO_train2014_000000561384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black an white goat in front of another goat\"."}], "task": "refering", "answer_template": "The \"a black an white goat in front of another goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 262, 263, 279, 280, 281, 282, 283, 285, 286, 302, 305, 306, 309], "bbox": [0.1363125, 0.4666744730679157, 0.477796875, 0.9299063231850119], "iscrowd": 0, "area": 24065.98585000001, "rle": {"size": [427, 640], "counts": "jbT12W=7I6J6J7I6L5K4L5K3M3aD[NS;g1jDZNT;k1gDWNV;Q2N2M3M2O0O2O0O2N2O0O2N101N2N101N1O2O1N100O1O010O1O10O01O100O1O010O100O0010O01O01O100O1O100O100O1O11O010O001O001O00001O001O0100O1O_FWMe0Ga6R3cHaMg0_Oe6P3^HlMg0VOk6n2WHVNi0nNP7Z4kHhKU7\\4eHfK[7U5O2O0000000O10001O000O01L4K5K5K4N3M3M3M3M2N3_NWG]Nl8c1[GSNh8l1`GPN]8P2jGTNj7l1ZHUNb7k1aHSN^7m1dHRN[7m1hHRNW7n1kHQNS7P2PInMo6Q2TInMk6R2WImMh6R2\\IlMc6T2_IiMa6X2aIdMa6[2bIaM`6_2dI[M^6d2fIWM\\6i2R210O10000O010O10000O01000O1000O100000000O10000000O0100000O10O10O100O10O10O100O1000O0100001O1N2O1O001O1dG[Mo5g2lI\\MU6e2gI]MZ6d2eHUMN:]7b2`H`MH0i7a2ZH`Nf7b1UHaNl7_1QHcNP8[31O2O0O1O1M3M3M3M3M3M4MoGiKR7T4oHSLl6j3UIWLl6f3UI\\Lk6a3VIaLj6b3QI_LP7d3kH^LU7f3dH\\L]7g3^H[Lc7g3XHZLi7d40O100O2O2K4L5K5K4cL^Gk1h8PNPHm0`8nNR3K8G9H9GdQ[4"}, "label": "a black an white goat in front of another goat"}]}
{"id": 422127, "image": "COCO_train2014_000000422127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing brown color top and only left hand is visible\"."}], "task": "refering", "answer_template": "The \"a women wearing brown color top and only left hand is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 116, 117, 139, 140, 162, 163, 185, 186, 208, 209, 231, 232, 254, 255, 256, 277, 278, 279, 300, 301, 323, 324], "bbox": [0.040625, 0.24097916666666666, 0.18284375, 0.8978541666666667], "iscrowd": 0, "area": 8830.463249999999, "rle": {"size": [480, 640], "counts": "dY<5k>:G:E:G:E:G9F;F9F;F9F;F9F;F9F;F9F:F;F9F;F9F;F9H9F9G:G8G7I10O01O10gHdJl4[5nJlJQ5T5lJQKS5o4hJVKX5i4eJ\\KZ5c4cJbK[5_4aJeKb0TN5k0T1[5QNlKa0QN<_9POdHa0PN?\\9lNiH`0oMd0W9hNPI?mMi0R9eNUI?lMl0k;oNXDR1i;iN[DX1d;eN_D[1b;`NbD`1`<0O1000O1O1O010O1O1O1O010O1O1O1O010O1O1N2N1O2M3N2N2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2Nfid7"}, "label": "a women wearing brown color top and only left hand is visible"}]}
{"id": 422127, "image": "COCO_train2014_000000422127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman entering a kitchen to the left of another woman\"."}], "task": "refering", "answer_template": "The \"a woman entering a kitchen to the left of another woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 116, 117, 139, 140, 162, 163, 185, 186, 208, 209, 231, 232, 254, 255, 256, 277, 278, 279, 300, 301, 323, 324], "bbox": [0.040625, 0.24097916666666666, 0.18284375, 0.8978541666666667], "iscrowd": 0, "area": 8830.463249999999, "rle": {"size": [480, 640], "counts": "dY<5k>:G:E:G:E:G9F;F9F;F9F;F9F;F9F;F9F:F;F9F;F9F;F9H9F9G:G8G7I10O01O10gHdJl4[5nJlJQ5T5lJQKS5o4hJVKX5i4eJ\\KZ5c4cJbK[5_4aJeKb0TN5k0T1[5QNlKa0QN<_9POdHa0PN?\\9lNiH`0oMd0W9hNPI?mMi0R9eNUI?lMl0k;oNXDR1i;iN[DX1d;eN_D[1b;`NbD`1`<0O1000O1O1O010O1O1O1O010O1O1O1O010O1O1N2N1O2M3N2N2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2Nfid7"}, "label": "a woman entering a kitchen to the left of another woman"}]}
{"id": 454898, "image": "COCO_train2014_000000454898.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella on far right\"."}], "task": "refering", "answer_template": "The \"umbrella on far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178], "bbox": [0.594546875, 0.18652173913043477, 0.98121875, 0.48314009661835744], "iscrowd": 0, "area": 21181.822350000002, "rle": {"size": [414, 640], "counts": "^Vj48_<9F:I7L4M2N3M2M3N1O1O2N1N3N1O2N1O101N1O2O0O101N1O2O0O1O100O2N100O100O2N100O1O101N100O1O101N100O1O100O100O100O1O010O100O1O100O100O100O1O100O100O1O010O100O10000O100O100O100O100O10O010000O100O100O100O10000O10000000O100000000O10000O0100000O1000000O100000O01000000O10000O0100000O100000O01000000O100000O010000O100000O01000000O100O10O0100O100O2O0O100O100O100O100O101N100O100O100O100O100O2O0O1O100O100O101N100O100O2O0O100O100O2O0O100O1O2N100O1O1O2N1O1O1O1O2O0O1O1O1O2N2N2N2O1N2N2N2N2N2N2L4K5K5K5L5J5KTh4"}, "label": "umbrella on far right"}]}
{"id": 454898, "image": "COCO_train2014_000000454898.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first umbrella from the front\"."}], "task": "refering", "answer_template": "The \"the first umbrella from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178], "bbox": [0.594546875, 0.18652173913043477, 0.98121875, 0.48314009661835744], "iscrowd": 0, "area": 21181.822350000002, "rle": {"size": [414, 640], "counts": "^Vj48_<9F:I7L4M2N3M2M3N1O1O2N1N3N1O2N1O101N1O2O0O101N1O2O0O1O100O2N100O100O2N100O1O101N100O1O101N100O1O100O100O100O1O010O100O1O100O100O100O1O100O100O1O010O100O10000O100O100O100O100O10O010000O100O100O100O10000O10000000O100000000O10000O0100000O1000000O100000O01000000O10000O0100000O100000O01000000O100000O010000O100000O01000000O100O10O0100O100O2O0O100O100O100O100O101N100O100O100O100O100O2O0O1O100O100O101N100O100O2O0O100O100O2O0O100O1O2N100O1O1O2N1O1O1O1O2O0O1O1O1O2N2N2N2O1N2N2N2N2N2N2L4K5K5K5L5J5KTh4"}, "label": "the first umbrella from the front"}]}
{"id": 454898, "image": "COCO_train2014_000000454898.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella in the middle of the others\"."}], "task": "refering", "answer_template": "The \"umbrella in the middle of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172], "bbox": [0.41870312500000006, 0.28741545893719805, 0.65084375, 0.5041787439613526], "iscrowd": 0, "area": 6046.998349999999, "rle": {"size": [414, 640], "counts": "aa\\35f<3O2N2M3N1O1O1O1N2O1O1O1N101O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10O01O1O001O1O010O1O001O1O1O010O1O001G9N1O2O0011O0O100O10000O101O0O101N4M001N100O10000O10O010000O100O01000O100O10O010000O10O01O1BWEgNi:S1d0K5N1O2N2N2O1N2N101O1N2O1O1O1O1O1O1O1O1O1O100O100O10O0100O100O100O10Wkj2"}, "label": "umbrella in the middle of the others"}]}
{"id": 454898, "image": "COCO_train2014_000000454898.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second umbrella from the right\"."}], "task": "refering", "answer_template": "The \"second umbrella from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172], "bbox": [0.41870312500000006, 0.28741545893719805, 0.65084375, 0.5041787439613526], "iscrowd": 0, "area": 6046.998349999999, "rle": {"size": [414, 640], "counts": "aa\\35f<3O2N2M3N1O1O1O1N2O1O1O1N101O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10O01O1O001O1O010O1O001O1O1O010O1O001G9N1O2O0011O0O100O10000O101O0O101N4M001N100O10000O10O010000O100O01000O100O10O010000O10O01O1BWEgNi:S1d0K5N1O2N2N2O1N2N101O1N2O1O1O1O1O1O1O1O1O1O100O100O10O0100O100O100O10Wkj2"}, "label": "second umbrella from the right"}]}
{"id": 168179, "image": "COCO_train2014_000000168179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rear end of an elephant with it ' s tail swishing to the side\"."}], "task": "refering", "answer_template": "The \"the rear end of an elephant with it ' s tail swishing to the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.42134375, 0.004369158878504673, 1.0, 0.9886682242990653], "iscrowd": 0, "area": 144565.95454999997, "rle": {"size": [428, 640], "counts": "nPa3\\1^;`1`Na1mN;G7I8G8I7I7I7I7I7H8I7I8H5K5K5J6K5K4L5K5K5J6K4L5K5K5K5J5L5L4M3L4L3M4L4L4L4L2N1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O00100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000002N1Oh4XK^5bJdE"}, "label": "the rear end of an elephant with it ' s tail swishing to the side"}]}
{"id": 536820, "image": "COCO_train2014_000000536820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the middle\"."}], "task": "refering", "answer_template": "The \"the elephant in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220], "bbox": [0.22446875, 0.3029791666666667, 0.6355000000000001, 0.5762291666666667], "iscrowd": 0, "area": 22857.510649999997, "rle": {"size": [480, 640], "counts": "oeS2>[>>A`0B=G2N2N2N1O2N2N2N2N1O200000O10001O000O1000000O100000001O000000000000000000001O0000000000000001O100O001O1O1O1O1O1O1O1O001O1O0O2O001O001N101O001O0O2O001O002N2N1O2N2N2N2N1O2N2N1O0000O100000000O10000000000O100000000O10000000000O1O2N1O1O1O2N1O1001O10O01O1O0010O01O1O00100O001O10O01O001O1O0O2O1N101N101O1N101N100O100O100O2O0O100O100O101N4N1N01N101O001O001O0O2O001O1O001O0O2O001O001O001O001O001O001O001O1O001O010O001O001O010O1O1O1O1O100O1O1O1O1O1O1J6K5J6ZNdCb0a<^OPDKV<4[110O10000O\\CYOP;g0lD[OV;e0gD]OZ;b0cD@_;`0^DBc;=[DEf;;VDGl;9QDIP<6mCLU<4iCMX<3eCL_<3_CMd<3ZCLi<3UCMo<2nBNU=2iBMZ=2dBN_=2_BMd=3ZBMi=1VBNm=2QBMR>2mAMV><3M4L3M3M3M4M2MTP]3"}, "label": "the elephant in the middle"}]}
{"id": 536820, "image": "COCO_train2014_000000536820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant pictured between two other elephants\"."}], "task": "refering", "answer_template": "The \"elephant pictured between two other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220], "bbox": [0.22446875, 0.3029791666666667, 0.6355000000000001, 0.5762291666666667], "iscrowd": 0, "area": 22857.510649999997, "rle": {"size": [480, 640], "counts": "oeS2>[>>A`0B=G2N2N2N1O2N2N2N2N1O200000O10001O000O1000000O100000001O000000000000000000001O0000000000000001O100O001O1O1O1O1O1O1O1O001O1O0O2O001O001N101O001O0O2O001O002N2N1O2N2N2N2N1O2N2N1O0000O100000000O10000000000O100000000O10000000000O1O2N1O1O1O2N1O1001O10O01O1O0010O01O1O00100O001O10O01O001O1O0O2O1N101N101O1N101N100O100O100O2O0O100O100O101N4N1N01N101O001O001O0O2O001O1O001O0O2O001O001O001O001O001O001O001O1O001O010O001O001O010O1O1O1O1O100O1O1O1O1O1O1J6K5J6ZNdCb0a<^OPDKV<4[110O10000O\\CYOP;g0lD[OV;e0gD]OZ;b0cD@_;`0^DBc;=[DEf;;VDGl;9QDIP<6mCLU<4iCMX<3eCL_<3_CMd<3ZCLi<3UCMo<2nBNU=2iBMZ=2dBN_=2_BMd=3ZBMi=1VBNm=2QBMR>2mAMV><3M4L3M3M3M4M2MTP]3"}, "label": "elephant pictured between two other elephants"}]}
{"id": 536820, "image": "COCO_train2014_000000536820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant stepping on a large log\"."}], "task": "refering", "answer_template": "The \"an elephant stepping on a large log\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 251, 252, 274], "bbox": [0.7668125, 0.2556041666666667, 0.995515625, 0.6636875], "iscrowd": 0, "area": 18261.435999999998, "rle": {"size": [480, 640], "counts": "^[V7<_>a0@`0_Oa0@;E6K5J6K5K5J6K5[Oe0L2O1O100O010O1O10O0eNWFaMi9]2[FaMd9]2aF`M_9_2eF_M[9^2jF`MU9_2oF_MQ9^2i1O1N3N101OgEfMm7[2nGjMR8V2iGnMX8R2cGSN\\8m1_GYNa8g1ZG^Nf8b1UGbNl8^1oFgNP9Y1oFiNQ9W1nFiNS9W1mFiNS9W1lFjNS9V1mFkNS9U1lFkNU9U1jFlNV9T1iFmNV9S1jFmNW9S1hFnNX9R1gFoNY9Q1fFPOY9P1gFYNVOMT:j1eFZN\\OGo9o1dF[NABk9S2cF\\NE^Oh9U2cF^NGZOf9Y2bF]NJYOd9Z2aF^NMVOb9\\2`F_NW:`1iEaNV:a1hE`NX:c1dE^N\\:d1aE]N_:f1]E[Nc:h1YEYNg:i1VEXNk:n20O101O00001O00001O00001N1000000000000001N5L4L3M4L3M4L4L3M4jFSKQ8P5iGSKW8Q5cGQK]8S5]GoJc8T5XGnJh8V5RGlJm8c510000000000O1000001O000O100fMQGcNo8]1]GWNd8i1fGlMZ8T2PHbMP8^2TH]Mm7d2VHXMj7h2ZHTMf7l2^HPMc7o2aHmL_7T3dHhL\\7X3hHcLY7]3kH_LU7b3U2O0001O0000010O00001O01O0001O000lX1"}, "label": "an elephant stepping on a large log"}]}
{"id": 536820, "image": "COCO_train2014_000000536820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on far right with left leg on trunk\"."}], "task": "refering", "answer_template": "The \"elephant on far right with left leg on trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 251, 252, 274], "bbox": [0.7668125, 0.2556041666666667, 0.995515625, 0.6636875], "iscrowd": 0, "area": 18261.435999999998, "rle": {"size": [480, 640], "counts": "^[V7<_>a0@`0_Oa0@;E6K5J6K5K5J6K5[Oe0L2O1O100O010O1O10O0eNWFaMi9]2[FaMd9]2aF`M_9_2eF_M[9^2jF`MU9_2oF_MQ9^2i1O1N3N101OgEfMm7[2nGjMR8V2iGnMX8R2cGSN\\8m1_GYNa8g1ZG^Nf8b1UGbNl8^1oFgNP9Y1oFiNQ9W1nFiNS9W1mFiNS9W1lFjNS9V1mFkNS9U1lFkNU9U1jFlNV9T1iFmNV9S1jFmNW9S1hFnNX9R1gFoNY9Q1fFPOY9P1gFYNVOMT:j1eFZN\\OGo9o1dF[NABk9S2cF\\NE^Oh9U2cF^NGZOf9Y2bF]NJYOd9Z2aF^NMVOb9\\2`F_NW:`1iEaNV:a1hE`NX:c1dE^N\\:d1aE]N_:f1]E[Nc:h1YEYNg:i1VEXNk:n20O101O00001O00001O00001N1000000000000001N5L4L3M4L3M4L4L3M4jFSKQ8P5iGSKW8Q5cGQK]8S5]GoJc8T5XGnJh8V5RGlJm8c510000000000O1000001O000O100fMQGcNo8]1]GWNd8i1fGlMZ8T2PHbMP8^2TH]Mm7d2VHXMj7h2ZHTMf7l2^HPMc7o2aHmL_7T3dHhL\\7X3hHcLY7]3kH_LU7b3U2O0001O0000010O00001O01O0001O000lX1"}, "label": "elephant on far right with left leg on trunk"}]}
{"id": 110841, "image": "COCO_train2014_000000110841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player\"."}], "task": "refering", "answer_template": "The \"tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 383, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 432, 433, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 454, 455, 456, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480], "bbox": [0.131015625, 0.15243567753001716, 0.8740937499999999, 0.9883876500857633], "iscrowd": 0, "area": 131300.09575, "rle": {"size": [583, 640], "counts": "Y\\`12Sb02M3M3N2M3N2M3N2TLl3]Oc0]Oc0]Oc0oNQ1O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1XN[FnJe9i4hFRKY9e4TGVKm8`4bGZK_8]4nG_KR8Y4ZHbKg7[4`H`Ka7\\4gH_KZ7]4nH_KR7]4WI]Kj6`4`3L4L4L4L4M3L4L4M3O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O101O000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UEnK_7S4UGRLWOKd9T4cFcLHZOd9T4RFTM8iNf9T4aEdMh0XNg9Y6XFhIg9Z6WFfIi9[6UFgIj9Z6UFfIk9[6TFfIk9[6TFeIl9]6QFeIn9\\6QFdIo9]6PFcIP:^6oEbIQ:_6mEbIS:`6jEaIV:`6gEbIY:_6dEcI\\:^6aEdI_:^6]EdIc:P70000000000000000000001O00000000000000000000000000001O0000000000001O000000000SOiHjFW7b8]I^Gc6n7QJRHP6Y7eJfH\\5d6YK\\Ii4n5kKRJV4Y5^LgJd3b4QM^KQ3_4QM`KP3`4oL`KS3`4kL`KW3_4hLaKY3_4gL`K[3_4dLaK^3^4aLbK`3^4_LbKc3]4\\LcKf3\\4ZLcKg3]4XLcKj3]4TLcKm3]4RLcKP4\\4PLcKQ4]4nKcKS4]4lKcKV4\\4iKdKX4\\4gKdK[4[4dKeK]4[4bKeK_4\\4_KdKc4Z4]KfKd4Z4[KfKg4Y4XKgKi4Y4VKgKk4Y4TKgKn4X4QKhKP5W4PKiKR5V4mJjKT5V4kJjKV5V4iJjKY5U4fJkK[5T4eJlK]5S4bJmK_5S4`JmKa5S4^JmKd5R4[JnKf5R4YJnKi5P4WJPLj5P4UJPLl5P4SJPLo5o3PJQLQ6n3oIRLS6l3mITLT6l3kITLV6k3jIULX6i3hIWLY6h3gIXL[6f3eIZL\\6f3bI[L`6c3`I]La6b3_I^Lb6a3^I_Ld6_3\\IaLe6^3[IbLg6]3XIcLi6\\3WIdLj6[3VIeLl6Y3TIgLm6X3SIhLo6W3PIiLQ7V3oHjLS7T3mHlLT7R3mHnLU7P3kHPMW7n2iHRMY7l2gHTM[7j2eHVM]7h2cHXM_7e2bH[M`7c2`H]Mb7a2^H_Md7_2\\HaMf7]2ZHcMh7[2XHeMj7X2WHhMk7V2UHjMm7T2SHlMo7R2QHnMP8Q2PHoMR8n1oGRNS8l1mGTNU8j1kGVNV8i1jGWNV8i1jGWNW8h1iGXNW8g1jGYNW8f1iGZNX8e1hG[NX8e1hG[NY8d1gG\\NY8d1gG\\NZ8c1fG]N[8a1fG_NZ8a1fG_N[8`1dGaN\\8_1cGbN]8^1bGcN^8]1`GeN`8Z1`GgN`8Y1_GhNa8X1^GiNb8W1\\GkNd8U1[GlNe8T1ZGmNf8S1YGnNg8R1WGPOi8P1VGQOi8P1VGQOj8o0TGSOk8k0WGVOi8g0XG[Og8d0ZG]Of8`0[GBd8=]GDc89^GIa86`GK_84aGN^81bG1]8OcG2\\8NcG4\\8LdG5[8KfG5Y8KgG6X8JiG6V8JjG7U8KjG5V8LiG4W8LhG5X8LgG4Y8MfG3Z8MeG4Z8NeG2[8OdG1\\8OcG2]8ObG1^8OaG2_8O`G1`80_G0a80^G1a81^GOb82\\GOd81\\GOd82ZGOf84WGLi89QGHo8=lFCX9>bFCb9>YFBj9<TFEn96UFJh97[FHa9:aFF\\9:gFFW9:kFFT98PGGP97RGIn85TGKk83XGMh81[GNd82]GNd8O^G1b8N`G1`8NaG2_8MbG3^8KeG4\\8JgG4Y8KiG4W8JmG4S8KoG4R8JRH3n7LWH0i7N\\HOR=L2N2N2N1O2Eg^ODZa08kb]1"}, "label": "tennis player"}]}
{"id": 110841, "image": "COCO_train2014_000000110841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis player in white\"."}], "task": "refering", "answer_template": "The \"the tennis player in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 383, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 432, 433, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 454, 455, 456, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480], "bbox": [0.131015625, 0.15243567753001716, 0.8740937499999999, 0.9883876500857633], "iscrowd": 0, "area": 131300.09575, "rle": {"size": [583, 640], "counts": "Y\\`12Sb02M3M3N2M3N2M3N2TLl3]Oc0]Oc0]Oc0oNQ1O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1XN[FnJe9i4hFRKY9e4TGVKm8`4bGZK_8]4nG_KR8Y4ZHbKg7[4`H`Ka7\\4gH_KZ7]4nH_KR7]4WI]Kj6`4`3L4L4L4L4M3L4L4M3O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O101O000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UEnK_7S4UGRLWOKd9T4cFcLHZOd9T4RFTM8iNf9T4aEdMh0XNg9Y6XFhIg9Z6WFfIi9[6UFgIj9Z6UFfIk9[6TFfIk9[6TFeIl9]6QFeIn9\\6QFdIo9]6PFcIP:^6oEbIQ:_6mEbIS:`6jEaIV:`6gEbIY:_6dEcI\\:^6aEdI_:^6]EdIc:P70000000000000000000001O00000000000000000000000000001O0000000000001O000000000SOiHjFW7b8]I^Gc6n7QJRHP6Y7eJfH\\5d6YK\\Ii4n5kKRJV4Y5^LgJd3b4QM^KQ3_4QM`KP3`4oL`KS3`4kL`KW3_4hLaKY3_4gL`K[3_4dLaK^3^4aLbK`3^4_LbKc3]4\\LcKf3\\4ZLcKg3]4XLcKj3]4TLcKm3]4RLcKP4\\4PLcKQ4]4nKcKS4]4lKcKV4\\4iKdKX4\\4gKdK[4[4dKeK]4[4bKeK_4\\4_KdKc4Z4]KfKd4Z4[KfKg4Y4XKgKi4Y4VKgKk4Y4TKgKn4X4QKhKP5W4PKiKR5V4mJjKT5V4kJjKV5V4iJjKY5U4fJkK[5T4eJlK]5S4bJmK_5S4`JmKa5S4^JmKd5R4[JnKf5R4YJnKi5P4WJPLj5P4UJPLl5P4SJPLo5o3PJQLQ6n3oIRLS6l3mITLT6l3kITLV6k3jIULX6i3hIWLY6h3gIXL[6f3eIZL\\6f3bI[L`6c3`I]La6b3_I^Lb6a3^I_Ld6_3\\IaLe6^3[IbLg6]3XIcLi6\\3WIdLj6[3VIeLl6Y3TIgLm6X3SIhLo6W3PIiLQ7V3oHjLS7T3mHlLT7R3mHnLU7P3kHPMW7n2iHRMY7l2gHTM[7j2eHVM]7h2cHXM_7e2bH[M`7c2`H]Mb7a2^H_Md7_2\\HaMf7]2ZHcMh7[2XHeMj7X2WHhMk7V2UHjMm7T2SHlMo7R2QHnMP8Q2PHoMR8n1oGRNS8l1mGTNU8j1kGVNV8i1jGWNV8i1jGWNW8h1iGXNW8g1jGYNW8f1iGZNX8e1hG[NX8e1hG[NY8d1gG\\NY8d1gG\\NZ8c1fG]N[8a1fG_NZ8a1fG_N[8`1dGaN\\8_1cGbN]8^1bGcN^8]1`GeN`8Z1`GgN`8Y1_GhNa8X1^GiNb8W1\\GkNd8U1[GlNe8T1ZGmNf8S1YGnNg8R1WGPOi8P1VGQOi8P1VGQOj8o0TGSOk8k0WGVOi8g0XG[Og8d0ZG]Of8`0[GBd8=]GDc89^GIa86`GK_84aGN^81bG1]8OcG2\\8NcG4\\8LdG5[8KfG5Y8KgG6X8JiG6V8JjG7U8KjG5V8LiG4W8LhG5X8LgG4Y8MfG3Z8MeG4Z8NeG2[8OdG1\\8OcG2]8ObG1^8OaG2_8O`G1`80_G0a80^G1a81^GOb82\\GOd81\\GOd82ZGOf84WGLi89QGHo8=lFCX9>bFCb9>YFBj9<TFEn96UFJh97[FHa9:aFF\\9:gFFW9:kFFT98PGGP97RGIn85TGKk83XGMh81[GNd82]GNd8O^G1b8N`G1`8NaG2_8MbG3^8KeG4\\8JgG4Y8KiG4W8JmG4S8KoG4R8JRH3n7LWH0i7N\\HOR=L2N2N2N1O2Eg^ODZa08kb]1"}, "label": "the tennis player in white"}]}
{"id": 520445, "image": "COCO_train2014_000000520445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cockerspaniel\"."}], "task": "refering", "answer_template": "The \"a black and white cockerspaniel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 315, 316, 317], "bbox": [0.501484375, 0.2597658079625293, 0.980046875, 0.9306323185011709], "iscrowd": 0, "area": 54479.94575, "rle": {"size": [427, 640], "counts": "jUV42W=4L4L4L4L4L;E3L3N3N1N3N1N3M2O2M2O2M201ZETNj9m1SFWNk9l1oEXNo7M]Io1WN^NV8N\\I\\2]6lM`IU2Z6RNdIP2U6XNiIi1S6[NkIf1R6^NlIc1R6aNkIa1R6cNlI]1Q6hNlIY1R6kNkIV1R6POkIP1S6XOdIl0X6\\OaIf0]6B[I`0c6U3M3N2M4M2N2M3O1N2N2O1N2N2N2O1N2N2N2N2N2N2N101N2O1O1N2O1N2NcNSK`Jk4b5XK[Jg4f5]KVJc4i5aKTJ^4m5eKPJ[4P6iKlIV4U6mKhIR4X6SLcIn3]6UL`Ij3a6ZL[Ie3f6_LVIa3i6dLSI[3n6a10O100O100O100O2O0O100O100O1O100O101O000000000000000000000000O10O100000O100000000O100000000O1000001O0O10kJPIo3P7oKSIQ4m6mKUIR4k6lKXIS4h6lKZIS4f6kK\\IU4d6jK]IW4b6hK`IW4`6hKaIX4`6fKaIZ4_6dKcI]4\\6bKeI^4[6`KhI_4X6`KiI`4W6^KkIb4V6[KlIf4V6VKkIj4[701O01O0000001O001O1O001O001O0O2O001O001O001O001O001O1O0O2O001O2N1N2O1N2O2M2O1N3YO\\JcIf5[6g0M3M4L3M3M3N3L4L4L4L3N3L4L10O01O10O01O100O2N100O2N100O1O2O0O1O2O1N1O2O0O2O0O2O1N2O1N101O1N2O1N2O1O1O1O1O1O001O1N2O1O1O1O1O1O1O001O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N3N1O1N3N1N2N3M2N3M2N2N1O2N2N2N2N2N1O1O1O1O1O1O1N1N3N2M6J7I7J6I7IhS5"}, "label": "a black and white cockerspaniel"}]}
{"id": 520445, "image": "COCO_train2014_000000520445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the springer spaniel is mostly white\"."}], "task": "refering", "answer_template": "The \"the springer spaniel is mostly white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 315, 316, 317], "bbox": [0.501484375, 0.2597658079625293, 0.980046875, 0.9306323185011709], "iscrowd": 0, "area": 54479.94575, "rle": {"size": [427, 640], "counts": "jUV42W=4L4L4L4L4L;E3L3N3N1N3N1N3M2O2M2O2M201ZETNj9m1SFWNk9l1oEXNo7M]Io1WN^NV8N\\I\\2]6lM`IU2Z6RNdIP2U6XNiIi1S6[NkIf1R6^NlIc1R6aNkIa1R6cNlI]1Q6hNlIY1R6kNkIV1R6POkIP1S6XOdIl0X6\\OaIf0]6B[I`0c6U3M3N2M4M2N2M3O1N2N2O1N2N2N2O1N2N2N2N2N2N2N101N2O1O1N2O1N2NcNSK`Jk4b5XK[Jg4f5]KVJc4i5aKTJ^4m5eKPJ[4P6iKlIV4U6mKhIR4X6SLcIn3]6UL`Ij3a6ZL[Ie3f6_LVIa3i6dLSI[3n6a10O100O100O100O2O0O100O100O1O100O101O000000000000000000000000O10O100000O100000000O100000000O1000001O0O10kJPIo3P7oKSIQ4m6mKUIR4k6lKXIS4h6lKZIS4f6kK\\IU4d6jK]IW4b6hK`IW4`6hKaIX4`6fKaIZ4_6dKcI]4\\6bKeI^4[6`KhI_4X6`KiI`4W6^KkIb4V6[KlIf4V6VKkIj4[701O01O0000001O001O1O001O001O0O2O001O001O001O001O001O1O0O2O001O2N1N2O1N2O2M2O1N3YO\\JcIf5[6g0M3M4L3M3M3N3L4L4L4L3N3L4L10O01O10O01O100O2N100O2N100O1O2O0O1O2O1N1O2O0O2O0O2O1N2O1N101O1N2O1N2O1O1O1O1O1O001O1N2O1O1O1O1O1O1O001O1O1O1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N3N1O1N3N1N2N3M2N3M2N2N1O2N2N2N2N2N1O1O1O1O1O1O1N1N3N2M6J7I7J6I7IhS5"}, "label": "the springer spaniel is mostly white"}]}
{"id": 545022, "image": "COCO_train2014_000000545022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that can barely be seen\"."}], "task": "refering", "answer_template": "The \"the horse that can barely be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 146, 147, 148, 171, 173, 194, 195, 196, 217, 218, 219, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310], "bbox": [0.358515625, 0.36791569086651055, 0.606796875, 0.8997892271662764], "iscrowd": 0, "area": 11474.92869999999, "rle": {"size": [427, 640], "counts": "edo23X=6K001O1N010[CFX<:eCG\\<:bCG_<9^CIc<?00000O10O100O100000000000001N100O1O100O2O0O2O001O0O2O001O1hG<g2FUM>i2DRMb0n2]OmLh0Q3ZOkLk0R3VOjLP1S3QOiLT1U3mNiLV1U3lNgLV1Z3jNdLW1\\3kNaLV1_3nN[LT1f3nNVLS1j3QOPLQ1P4ROiKQ1X4RObK^NfNa1h54[K^NPO\\1f5:QK]N]OV1b5b0fJ]NKo0_5n0QJ]Nc0b0]5`2iJ\\MW5d2oJVMR5i2TKPMm4P3[KfLh4Y3aK[Lc4d3c2O1O1N101O1O1O1O10hH]Lh4b3SKcLo4Z3PKhLV5f0SIk0d1aN]5=SIQ1_1dNd5KWIa1R1fNP7Y1oHhNR7X1iHmNV7S1`HWOa7h0bG5^8L\\G:c8FUGb0l8n1000000000000000O1000000O10O1O010M3N1O2N2N2NmMfGZN4CW8P2YIhMg6W2W2N2N12O1N2O1N20]N]E?b:^OfE=Z:@oE:P:AYF=f9[OfFa0Z9VORGg0m:M3L3N3L4L4M2M^X65WgIm0ZO8H8\\D]No:R2NO1N2O1N7J<Ca0@7I4LdgX3"}, "label": "the horse that can barely be seen"}]}
{"id": 545022, "image": "COCO_train2014_000000545022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back leftmost horse who ' s ears you can see above the other horse ' s butt\"."}], "task": "refering", "answer_template": "The \"the back leftmost horse who ' s ears you can see above the other horse ' s butt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 146, 147, 148, 171, 173, 194, 195, 196, 217, 218, 219, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310], "bbox": [0.358515625, 0.36791569086651055, 0.606796875, 0.8997892271662764], "iscrowd": 0, "area": 11474.92869999999, "rle": {"size": [427, 640], "counts": "edo23X=6K001O1N010[CFX<:eCG\\<:bCG_<9^CIc<?00000O10O100O100000000000001N100O1O100O2O0O2O001O0O2O001O1hG<g2FUM>i2DRMb0n2]OmLh0Q3ZOkLk0R3VOjLP1S3QOiLT1U3mNiLV1U3lNgLV1Z3jNdLW1\\3kNaLV1_3nN[LT1f3nNVLS1j3QOPLQ1P4ROiKQ1X4RObK^NfNa1h54[K^NPO\\1f5:QK]N]OV1b5b0fJ]NKo0_5n0QJ]Nc0b0]5`2iJ\\MW5d2oJVMR5i2TKPMm4P3[KfLh4Y3aK[Lc4d3c2O1O1N101O1O1O1O10hH]Lh4b3SKcLo4Z3PKhLV5f0SIk0d1aN]5=SIQ1_1dNd5KWIa1R1fNP7Y1oHhNR7X1iHmNV7S1`HWOa7h0bG5^8L\\G:c8FUGb0l8n1000000000000000O1000000O10O1O010M3N1O2N2N2NmMfGZN4CW8P2YIhMg6W2W2N2N12O1N2O1N20]N]E?b:^OfE=Z:@oE:P:AYF=f9[OfFa0Z9VORGg0m:M3L3N3L4L4M2M^X65WgIm0ZO8H8\\D]No:R2NO1N2O1N7J<Ca0@7I4LdgX3"}, "label": "the back leftmost horse who ' s ears you can see above the other horse ' s butt"}]}
{"id": 545022, "image": "COCO_train2014_000000545022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse with 1 all black foot\"."}], "task": "refering", "answer_template": "The \"horse with 1 all black foot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315], "bbox": [0.47390625000000003, 0.3390632318501171, 0.7717031249999999, 0.907728337236534], "iscrowd": 0, "area": 24407.41925, "rle": {"size": [427, 640], "counts": "Rbn3>\\;I\\Ej0[:W1H9K41O010O01O00001O0001O000O0100O100O100O100O010O10O1N1N3N1O2N2N1O2N1O2N3M4L3M2M4M2NlI\\Nf0a1PKkNf2GZ2_1jJTOe2@a2c1]JUOl2\\Of2U3TMnLl2U3mLnLS3V3gLmLY3X3^LlLa3Z3ULkLk3X3mKlLS4X3eKmLZ4W3\\KPMd4S3SKRMm4T3iJPMW5V3_JnLb5]500001O001O00001O00001\\MYJbMg5X2mJXMV5g2XKhLl4W3e2O1O1O1O1O1O1O1O1O1O100O1O2N1O1O1O1O1O1O10O2cH\\LP5g3jJ[LW5g3cJ\\L]5g3]J\\Ld5f3RJaLo5a3XIUMh6o2jH[MW7\\41O001O001O000`McHRO]7m0hHoNY7P1jHmNV7S1TIcNl6]1^IYNb6f1eITN[6l1jIoMV6Q2PJiMP6W2VJcMj5\\2\\J_Md5a2bJYM^5g2hJSMX5m2mJnLS5Q3d201O000O13M5K5K4VH]Ld5h3TJ[Ll5j3lIYLT6k3eIXL[6m3^IULb6P4XIQLh6S4SInKm6W4lHkKT7Y4gHhKY7]4aHdK_7T5000000OVM]HNc75aHD_7?eHZO[7j0hHoNX7U1mHkLMT1V7U2iI^MW6f2nIQMR6S3eI[LUO8V7`3`I[LV7i3dHZL[7i3`HZL_7i3\\HYLd7j3WHYLh7f4O1N2O10000O10000O01N2M2O2N2jNkGSMV8f2SHVMP8`2[H]Mg7`2]H]Me7`2_H]Mc7^2cH^M`7V2mHgMU7n1WIoMj6m1\\IPNf6l1_IPNd6m1_2L4M6J6K6J5H7J8H9Fhdl1"}, "label": "horse with 1 all black foot"}]}
{"id": 545022, "image": "COCO_train2014_000000545022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse between the red wagon wheels\"."}], "task": "refering", "answer_template": "The \"a horse between the red wagon wheels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315], "bbox": [0.47390625000000003, 0.3390632318501171, 0.7717031249999999, 0.907728337236534], "iscrowd": 0, "area": 24407.41925, "rle": {"size": [427, 640], "counts": "Rbn3>\\;I\\Ej0[:W1H9K41O010O01O00001O0001O000O0100O100O100O100O010O10O1N1N3N1O2N2N1O2N1O2N3M4L3M2M4M2NlI\\Nf0a1PKkNf2GZ2_1jJTOe2@a2c1]JUOl2\\Of2U3TMnLl2U3mLnLS3V3gLmLY3X3^LlLa3Z3ULkLk3X3mKlLS4X3eKmLZ4W3\\KPMd4S3SKRMm4T3iJPMW5V3_JnLb5]500001O001O00001O00001\\MYJbMg5X2mJXMV5g2XKhLl4W3e2O1O1O1O1O1O1O1O1O1O100O1O2N1O1O1O1O1O1O10O2cH\\LP5g3jJ[LW5g3cJ\\L]5g3]J\\Ld5f3RJaLo5a3XIUMh6o2jH[MW7\\41O001O001O000`McHRO]7m0hHoNY7P1jHmNV7S1TIcNl6]1^IYNb6f1eITN[6l1jIoMV6Q2PJiMP6W2VJcMj5\\2\\J_Md5a2bJYM^5g2hJSMX5m2mJnLS5Q3d201O000O13M5K5K4VH]Ld5h3TJ[Ll5j3lIYLT6k3eIXL[6m3^IULb6P4XIQLh6S4SInKm6W4lHkKT7Y4gHhKY7]4aHdK_7T5000000OVM]HNc75aHD_7?eHZO[7j0hHoNX7U1mHkLMT1V7U2iI^MW6f2nIQMR6S3eI[LUO8V7`3`I[LV7i3dHZL[7i3`HZL_7i3\\HYLd7j3WHYLh7f4O1N2O10000O10000O01N2M2O2N2jNkGSMV8f2SHVMP8`2[H]Mg7`2]H]Me7`2_H]Mc7^2cH^M`7V2mHgMU7n1WIoMj6m1\\IPNf6l1_IPNd6m1_2L4M6J6K6J5H7J8H9Fhdl1"}, "label": "a horse between the red wagon wheels"}]}
{"id": 545022, "image": "COCO_train2014_000000545022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown horse in the front left of a four - horse team pulling a wagon\"."}], "task": "refering", "answer_template": "The \"a light brown horse in the front left of a four - horse team pulling a wagon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331], "bbox": [0.134875, 0.3313348946135831, 0.44765625, 0.9764871194379392], "iscrowd": 0, "area": 33785.84465, "rle": {"size": [427, 640], "counts": "g^T13W=2M3iDMS96bF6Z9L]F=a9EUFe0h9_OmEk0n8kNXG<AR1X9cNSG=[OY1c9\\NnF^2R9cMjF`2V9bMfF`2Z9bMbF_2`9bM\\F`2d9d000000O1O2N1O1O0000O10001O0O100000O1N1O2N2N2N101O1O100O00101N1O100O1O100O13L5VJZMf0R3lNQMS1V4cMnK\\2U4_MnK`2U4ZMoKe2T4RMSLm2P4kLVLT3m3dLYLZ3k3_LZLc3f3XL]Lk3c3oK`LS4b3fKaL]4_3]KdLd4_3VKcLj4a3oJbLQ5l51O001O001O001O001O000hLjJ[NW5d1mJXNS5g1RKUNo4j1UKRNk4m1ZKcLO_Oh4m3]K\\L3C`4P4bKTL7HX4R4RLbKO7o3W4hMeKX2Z4kMdKV2[4kMdKU2[4_3N10VHfKe6W4YInKf6Q4XITLf6j3YIZL=Eg4P4jJ`L4IR5g3gJdLH3a5X3dJkLXO=U6h2`JZN`5e1^J^Nc5d4100O10000O01O10O00001N10001N100O101O0O101O0O1O1N3L3N2O1000000000000000O10O1O1O1N2WOXInJi6Q5]IiJd6W5aIcJ`6\\5h000OZIaJb5_5]JbJc5]5]JeJa5\\5^JeJb5Z5^JhJ`5Y5_JiJ`5V5`JkJ_5V5aJkJ^5T5bJmJ]5T5bJnJ\\5R5dJPK[5P5dJQK[5n4fJTKX5h4kJZKU5b4kJcKS5Z4mJjKS5Z4eJjK\\5^4WJfKi5b4kIbKU6j51O00001O0000001O00001O01O01`McJoL^5n2iJkL[5Q3kJhL[5R3lJgLY5U3nJfLU5U3PKiLR5Q3TKnLm4l2YKRMh4i2^KVMc4c2dK[M^4_2hK_MZ4[2lKdMU4W2PLgMV4P2i3H8I8G8H8H9He0ZOXnb4"}, "label": "a light brown horse in the front left of a four - horse team pulling a wagon"}]}
{"id": 545022, "image": "COCO_train2014_000000545022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown horse in front\"."}], "task": "refering", "answer_template": "The \"brown horse in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331], "bbox": [0.134875, 0.3313348946135831, 0.44765625, 0.9764871194379392], "iscrowd": 0, "area": 33785.84465, "rle": {"size": [427, 640], "counts": "g^T13W=2M3iDMS96bF6Z9L]F=a9EUFe0h9_OmEk0n8kNXG<AR1X9cNSG=[OY1c9\\NnF^2R9cMjF`2V9bMfF`2Z9bMbF_2`9bM\\F`2d9d000000O1O2N1O1O0000O10001O0O100000O1N1O2N2N2N101O1O100O00101N1O100O1O100O13L5VJZMf0R3lNQMS1V4cMnK\\2U4_MnK`2U4ZMoKe2T4RMSLm2P4kLVLT3m3dLYLZ3k3_LZLc3f3XL]Lk3c3oK`LS4b3fKaL]4_3]KdLd4_3VKcLj4a3oJbLQ5l51O001O001O001O001O000hLjJ[NW5d1mJXNS5g1RKUNo4j1UKRNk4m1ZKcLO_Oh4m3]K\\L3C`4P4bKTL7HX4R4RLbKO7o3W4hMeKX2Z4kMdKV2[4kMdKU2[4_3N10VHfKe6W4YInKf6Q4XITLf6j3YIZL=Eg4P4jJ`L4IR5g3gJdLH3a5X3dJkLXO=U6h2`JZN`5e1^J^Nc5d4100O10000O01O10O00001N10001N100O101O0O101O0O1O1N3L3N2O1000000000000000O10O1O1O1N2WOXInJi6Q5]IiJd6W5aIcJ`6\\5h000OZIaJb5_5]JbJc5]5]JeJa5\\5^JeJb5Z5^JhJ`5Y5_JiJ`5V5`JkJ_5V5aJkJ^5T5bJmJ]5T5bJnJ\\5R5dJPK[5P5dJQK[5n4fJTKX5h4kJZKU5b4kJcKS5Z4mJjKS5Z4eJjK\\5^4WJfKi5b4kIbKU6j51O00001O0000001O00001O01O01`McJoL^5n2iJkL[5Q3kJhL[5R3lJgLY5U3nJfLU5U3PKiLR5Q3TKnLm4l2YKRMh4i2^KVMc4c2dK[M^4_2hK_MZ4[2lKdMU4W2PLgMV4P2i3H8I8G8H8H9He0ZOXnb4"}, "label": "brown horse in front"}]}
{"id": 291072, "image": "COCO_train2014_000000291072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in all white in the reflection\"."}], "task": "refering", "answer_template": "The \"the man in all white in the reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 142, 143, 144, 145, 157, 158, 159, 160, 173, 174, 189, 203, 204, 218, 219, 234, 235, 249, 250], "bbox": [0.5007259953161592, 0.363234375, 0.7057377049180329, 0.717484375], "iscrowd": 0, "area": 8774.372049999998, "rle": {"size": [640, 427], "counts": "aoU42nc03L3V]OKea08W^OJha09T^OIja0;R^OGma0;Q^OFna0=n]OFPb0=l]OERb0?j]OCUb0`0g]OBXb0T1O2M2O1O1O1N3N1O1O1N2O2Nf0ZO0O10001N1000001N11O1O010Of@SMb=l2]BYM`=f2_B^M^=c2\\BfMa=Y2YBQNd=n1WBZNf=f1TBdNi=[1QBoNl=Q1nAWOP>h0jABR>>hAKV>S3M3M2M4M3M2N3L3N3M00O001N2O001O1N2O001O1TKoCg1R<WNZD]1g;aNgDS1Z;kNSEi0n:TO_Ea0b:]OkE7U:GYFLi92dFB]9;PGZOQ9d0V5N3N2M2O2M3N2M2O2M3Na_^2"}, "label": "the man in all white in the reflection"}]}
{"id": 291072, "image": "COCO_train2014_000000291072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bearded man wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"the bearded man wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 142, 143, 144, 145, 157, 158, 159, 160, 173, 174, 189, 203, 204, 218, 219, 234, 235, 249, 250], "bbox": [0.5007259953161592, 0.363234375, 0.7057377049180329, 0.717484375], "iscrowd": 0, "area": 8774.372049999998, "rle": {"size": [640, 427], "counts": "aoU42nc03L3V]OKea08W^OJha09T^OIja0;R^OGma0;Q^OFna0=n]OFPb0=l]OERb0?j]OCUb0`0g]OBXb0T1O2M2O1O1O1N3N1O1O1N2O2Nf0ZO0O10001N1000001N11O1O010Of@SMb=l2]BYM`=f2_B^M^=c2\\BfMa=Y2YBQNd=n1WBZNf=f1TBdNi=[1QBoNl=Q1nAWOP>h0jABR>>hAKV>S3M3M2M4M3M2N3L3N3M00O001N2O001O1N2O001O1TKoCg1R<WNZD]1g;aNgDS1Z;kNSEi0n:TO_Ea0b:]OkE7U:GYFLi92dFB]9;PGZOQ9d0V5N3N2M2O2M3N2M2O2M3Na_^2"}, "label": "the bearded man wearing a white shirt"}]}
{"id": 307462, "image": "COCO_train2014_000000307462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tugboat to the right of the ship\"."}], "task": "refering", "answer_template": "The \"the tugboat to the right of the ship\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 200, 201, 223, 224, 225, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296], "bbox": [0.681015625, 0.4415866388308977, 0.9071093749999999, 0.742400835073069], "iscrowd": 0, "area": 9209.6702, "rle": {"size": [479, 640], "counts": "[W\\65g>5M4M2M3\\BF_<=YC0^<4ZC1f<T1O100O100L4M3L4K5J6B?M2N2O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N3N1O12N9G9gMREOV;kNhEU1Q<0000000000000001N1000gNY1O10000O10000001O<Dh0XO7I0000001O0000000O1000000O2O00L3O2M3O001O100O1C<101O3N1N2N2N2N001O0017H0100O000dNnBl0Q=POSCP1m<iNZCW1f<hN[CX1f<fN[C[1d<eN\\C[1T=000010O1O1O101N1O100OO1B>O10M3N2M8I8G5L3L4M3L3N3L4N3N1N2OQhk0"}, "label": "the tugboat to the right of the ship"}]}
{"id": 307462, "image": "COCO_train2014_000000307462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a steamboat sitting near the dock\"."}], "task": "refering", "answer_template": "The \"a steamboat sitting near the dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 200, 201, 223, 224, 225, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296], "bbox": [0.681015625, 0.4415866388308977, 0.9071093749999999, 0.742400835073069], "iscrowd": 0, "area": 9209.6702, "rle": {"size": [479, 640], "counts": "[W\\65g>5M4M2M3\\BF_<=YC0^<4ZC1f<T1O100O100L4M3L4K5J6B?M2N2O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N3N1O12N9G9gMREOV;kNhEU1Q<0000000000000001N1000gNY1O10000O10000001O<Dh0XO7I0000001O0000000O1000000O2O00L3O2M3O001O100O1C<101O3N1N2N2N2N001O0017H0100O000dNnBl0Q=POSCP1m<iNZCW1f<hN[CX1f<fN[C[1d<eN\\C[1T=000010O1O1O101N1O100OO1B>O10M3N2M8I8G5L3L4M3L3N3L4N3N1N2OQhk0"}, "label": "a steamboat sitting near the dock"}]}
{"id": 307462, "image": "COCO_train2014_000000307462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a row of tall flags\"."}], "task": "refering", "answer_template": "The \"a row of tall flags\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 200, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 288, 289, 290, 291], "bbox": [0.45746875, 0.06517745302713987, 0.743390625, 0.7393319415448851], "iscrowd": 0, "area": 25733.9015, "rle": {"size": [479, 640], "counts": "P[Y4Q1m=1O2N1O1O1O1O1O100O2O000O2fMcNPG]1o8dNQG\\1o8dNPG]1P9cNPG^1n8dNQG\\1o8eNoF\\1Q9gNlFZ1S9jNiFV1V9nNfFS1Z9QObFo0^9ARF`0l90cE2i8]NTHQ2aND]9^NmGi3T8[LfGe3]8]LYGh3i8\\LlFi3V9ZL_Fm3c9f02N2O1N3M2N2O2MH8H8G:G8QOo0K6J5K5K5N3L3N2N3L3N2M3N3L3N2M4N1O2O000O2O0O101N1O1N3N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2M2O1O2N1O1O2N1O1N3N1O1O1O1O1O1i2WM1O1O2M2O1O1O1O1O2N1O1O1O1O0O2UK_HV2b7nLbIi2^6RMPJe2Q6UM]Jb2c5YMkJ^2V5\\MXK[2h4`M^K_2b8O2N1O1O1N3N1O00O1O1N10N2N2N2M3cNkCOV<KaDCa;7f1I7N200O0100000O1000O10O100000000O0100000O1000O10O10000000O0100000km00URO0kj^2"}, "label": "a row of tall flags"}]}
{"id": 307462, "image": "COCO_train2014_000000307462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tall ship in the harbor with flags on its poles\"."}], "task": "refering", "answer_template": "The \"the tall ship in the harbor with flags on its poles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 200, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 288, 289, 290, 291], "bbox": [0.45746875, 0.06517745302713987, 0.743390625, 0.7393319415448851], "iscrowd": 0, "area": 25733.9015, "rle": {"size": [479, 640], "counts": "P[Y4Q1m=1O2N1O1O1O1O1O100O2O000O2fMcNPG]1o8dNQG\\1o8dNPG]1P9cNPG^1n8dNQG\\1o8eNoF\\1Q9gNlFZ1S9jNiFV1V9nNfFS1Z9QObFo0^9ARF`0l90cE2i8]NTHQ2aND]9^NmGi3T8[LfGe3]8]LYGh3i8\\LlFi3V9ZL_Fm3c9f02N2O1N3M2N2O2MH8H8G:G8QOo0K6J5K5K5N3L3N2N3L3N2M3N3L3N2M4N1O2O000O2O0O101N1O1N3N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2M2O1O2N1O1O2N1O1N3N1O1O1O1O1O1i2WM1O1O2M2O1O1O1O1O2N1O1O1O1O0O2UK_HV2b7nLbIi2^6RMPJe2Q6UM]Jb2c5YMkJ^2V5\\MXK[2h4`M^K_2b8O2N1O1O1N3N1O00O1O1N10N2N2N2M3cNkCOV<KaDCa;7f1I7N200O0100000O1000O10O100000000O0100000O1000O10O10000000O0100000km00URO0kj^2"}, "label": "the tall ship in the harbor with flags on its poles"}]}
{"id": 184583, "image": "COCO_train2014_000000184583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"big dark brown buffalo walking between 2 smaller lighter brown buffalo\"."}], "task": "refering", "answer_template": "The \"big dark brown buffalo walking between 2 smaller lighter brown buffalo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 219, 220, 221], "bbox": [0.29865624999999996, 0.34312499999999996, 0.6275000000000001, 0.5825208333333334], "iscrowd": 0, "area": 13680.048599999996, "rle": {"size": [480, 640], "counts": "kgi21l>3M4M2M3M3M3M4L3M3N2O1O2O0O1O1O100O2N1O1O100O2O000000001O00000O101O000001O100O1O1OO100O2N100K5O1O2N100O1O1O2O0O2N2N2N3N1N2N2O2N1O1O1O2N1O1O2N1O1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O1O0000000000000000000000QObD\\N^;d1iDUNW;k1oDoMQ;Q2SEkMm:U2WEgMi:Y2o00000000000000000001O000001O001O00001O00001O001O000010O0001O001O00001O00001O001N1O1O2OSDQNi:n1VESNl:k1SEWNm:h1REYNo:f1PE\\NQ;c1mD^NT;a1kDaNU;b1fD_N\\;e1]D\\Nd;i1UDXNl;m1mCTNU<Z21O1O2M2O1O2N1O1O1M4K4M3M4L3L4M4O01O100O1O1O100O1O100O1O1O100OXNPCa1o<`NWCZ1h<fN`CS1_<oNfCk0Y<WOkCd0T<^OlCa0S<AmC>S<CmC<R<FbCIEc0d<F[C13<]<:dCIW<7kCLP<4RDKm;6UDFn;8VDDQ<5RDHU<2nCJY<OkCM\\<LhC0_<JcC3b<HbC4i=L\\Y_3"}, "label": "big dark brown buffalo walking between 2 smaller lighter brown buffalo"}]}
{"id": 184583, "image": "COCO_train2014_000000184583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult bison\"."}], "task": "refering", "answer_template": "The \"the adult bison\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 219, 220, 221], "bbox": [0.29865624999999996, 0.34312499999999996, 0.6275000000000001, 0.5825208333333334], "iscrowd": 0, "area": 13680.048599999996, "rle": {"size": [480, 640], "counts": "kgi21l>3M4M2M3M3M3M4L3M3N2O1O2O0O1O1O100O2N1O1O100O2O000000001O00000O101O000001O100O1O1OO100O2N100K5O1O2N100O1O1O2O0O2N2N2N3N1N2N2O2N1O1O1O2N1O1O2N1O1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O1O0000000000000000000000QObD\\N^;d1iDUNW;k1oDoMQ;Q2SEkMm:U2WEgMi:Y2o00000000000000000001O000001O001O00001O00001O001O000010O0001O001O00001O00001O001N1O1O2OSDQNi:n1VESNl:k1SEWNm:h1REYNo:f1PE\\NQ;c1mD^NT;a1kDaNU;b1fD_N\\;e1]D\\Nd;i1UDXNl;m1mCTNU<Z21O1O2M2O1O2N1O1O1M4K4M3M4L3L4M4O01O100O1O1O100O1O100O1O1O100OXNPCa1o<`NWCZ1h<fN`CS1_<oNfCk0Y<WOkCd0T<^OlCa0S<AmC>S<CmC<R<FbCIEc0d<F[C13<]<:dCIW<7kCLP<4RDKm;6UDFn;8VDDQ<5RDHU<2nCJY<OkCM\\<LhC0_<JcC3b<HbC4i=L\\Y_3"}, "label": "the adult bison"}]}
{"id": 200971, "image": "COCO_train2014_000000200971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the refrigerator behind the man in black\"."}], "task": "refering", "answer_template": "The \"the refrigerator behind the man in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 25, 39, 40, 41, 42, 43, 57, 58, 59, 60, 75, 76, 77, 93, 94, 111, 112, 129], "bbox": [0.18034, 0.07415999999999999, 0.43146, 0.5820266666666667], "iscrowd": 0, "area": 11856.334599999996, "rle": {"size": [375, 500], "counts": "boP1=Z;[1eN[1eN]2fGdJ_5j5000O1000000O1000000O100KXJ[Jh5b58M4L3M3M3M3M3M3M3M3M3M3M3N2N2O1N2N2N2N2O1N2N2N2N2N2O1N3M2N2N2O1N2N2N2N2N2O1N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O2N1O100O1O1O100O100O100O100O10000O100O100O100O100O101N100O100O100O100O100O100Ob0_O4K6K4K5L2N1N2O1N2O1N2O0O2O1N2OgmW3"}, "label": "the refrigerator behind the man in black"}]}
{"id": 200971, "image": "COCO_train2014_000000200971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an off - white fridge with a blue magnet near it ' s handle\"."}], "task": "refering", "answer_template": "The \"an off - white fridge with a blue magnet near it ' s handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 25, 39, 40, 41, 42, 43, 57, 58, 59, 60, 75, 76, 77, 93, 94, 111, 112, 129], "bbox": [0.18034, 0.07415999999999999, 0.43146, 0.5820266666666667], "iscrowd": 0, "area": 11856.334599999996, "rle": {"size": [375, 500], "counts": "boP1=Z;[1eN[1eN]2fGdJ_5j5000O1000000O1000000O100KXJ[Jh5b58M4L3M3M3M3M3M3M3M3M3M3M3N2N2O1N2N2N2N2O1N2N2N2N2N2O1N3M2N2N2O1N2N2N2N2N2O1N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O2N1O100O1O1O100O100O100O100O10000O100O100O100O100O101N100O100O100O100O100O100Ob0_O4K6K4K5L2N1N2O1N2O1N2O0O2O1N2OgmW3"}, "label": "an off - white fridge with a blue magnet near it ' s handle"}]}
{"id": 200971, "image": "COCO_train2014_000000200971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"refridgator with white magnet\"."}], "task": "refering", "answer_template": "The \"refridgator with white magnet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 20, 21, 36, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 126, 127, 128, 129], "bbox": [0.0, 0.04944, 0.1854, 0.57304], "iscrowd": 0, "area": 17615.9335, "rle": {"size": [375, 500], "counts": "c0^5Y61O1O1O1O001O1O1O1O3M2N2N3M00000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O0001O00001O00001O0000001O0O101O0000001O00001O00001O0000001OR4nKcld4"}, "label": "refridgator with white magnet"}]}
{"id": 200971, "image": "COCO_train2014_000000200971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the refrigerator on the left\"."}], "task": "refering", "answer_template": "The \"the refrigerator on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 20, 21, 36, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 126, 127, 128, 129], "bbox": [0.0, 0.04944, 0.1854, 0.57304], "iscrowd": 0, "area": 17615.9335, "rle": {"size": [375, 500], "counts": "c0^5Y61O1O1O1O001O1O1O1O3M2N2N3M00000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O0001O00001O00001O0000001O0O101O0000001O00001O00001O0000001OR4nKcld4"}, "label": "the refrigerator on the left"}]}
{"id": 569613, "image": "COCO_train2014_000000569613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant leading another elephant across a dirt road\"."}], "task": "refering", "answer_template": "The \"elephant leading another elephant across a dirt road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 248, 249, 250, 251, 264, 265, 266, 267, 268, 271, 272, 290, 291, 294, 295], "bbox": [0.5092187499999999, 0.26232941176470587, 0.972265625, 0.8183764705882353], "iscrowd": 0, "area": 42660.36580000001, "rle": {"size": [425, 640], "counts": "RdW42o73QM0m24nLOR32iL2X3NdL5\\3L`L7`3J\\L:d3FXL=h3ERL?n3BnKa0R4@jKc0W4]OeKf0[4[OaKi0^4YO]Kj0d4VOWKn0i4SOSKP1m4QOoJR1Q5oNkJT1V5lNfJX1Y5jNaJZ1_5gN]J\\1d5dNXJ_1h5bNTJa1l5`NPJc1P6_NjIf1V6ZNfIi1Z6XNbIk1^6VN^Im1c6SNYIP2g6QNUIR2k6PNoHU2P7lMlHW2U7iMgHZ2Y7gMcH\\2]7eM_H^2b7cMXHa2h7`MTHc2l7^MPHf2o7=UH\\Ll7a3aHTL_7j3mHlKS7R4R1O1O1O1O1O001O1O1O1O001O1O1O1O001001O00001O01O01O00010O3M3M3N2M3M3M3N2M4M2N6I;F:F:F:E<E3M00O01000000O10O1000O1000000O0101O000000000O2iMjIYMV6f2RJSMn5l2VJQMj5m2ZJQMf5n2]JPMc5o2aJmL`5R3cJlL^5S3eJjL[5U3kJfLU5Z3VK[Lj4e3XKYLh4g3[KVLe4j3^KRLc4n3`KoKa4P4\\2O100000001O0O10000000001O00000O1000001O0O100000001N1000000O10O1000O10O1000O100O01000O100O010OWI]Ln3i3jKYLV4n3`KUL`4R4WKPLh4Y4nJiKQ5g4WJbKg5Q6O2M2O1N2N200O100O10000O10000O101O0O100O10000O100O10000O10nNTJhJj5W5]JdJc5Z5dJaJ[5e4hJjJ4<T5i4gKTKZ4k4iKRKY4l4jKQKW4o4kKoJW4n4kKPKW4Q5P201O01O00010O00010O0000010O00010O00010O000010O00010O00010[NmGRNT8k1THPNl7m1YHPNg7n1^HPNb7m1bHQN_7l1eHSN[7j1iHTNW7k1lHTNT7i1oHVNR7j1nHVNQ7j1PIUNCROb6i2kIUNASOe6g2kIUN^OUOi6f2iIUN[OWOl6d2jITNYOXOo6d2hITNVOYOT7c2eIUNTOZOX7`2fIVNoNZO\\7a2eIUNlN[Oa7_2dIVNiN\\Od7_2bIUNgN]Oi7^2aI`N_6a1aI]Na6c1_I[Nc6e1^IYNc6g1]IWNd6U2QIiMQ7b2eH\\M\\7o2YHoLi7V3RHhLo7Y3QHfLP8Y3RHdLP8\\3PHbLQ8_3oG_LS8`3oG^LR8b3nG\\LT8d3lGZLU8f3lGXLV8h3kGULX8j3hGRL]8l3b0O1C=^Ob0I7L4M3L4L[1fN5K5J6K5JRX7"}, "label": "elephant leading another elephant across a dirt road"}]}
{"id": 569613, "image": "COCO_train2014_000000569613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant crossing the path in front of a vehicle\"."}], "task": "refering", "answer_template": "The \"an elephant crossing the path in front of a vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 248, 249, 250, 251, 264, 265, 266, 267, 268, 271, 272, 290, 291, 294, 295], "bbox": [0.5092187499999999, 0.26232941176470587, 0.972265625, 0.8183764705882353], "iscrowd": 0, "area": 42660.36580000001, "rle": {"size": [425, 640], "counts": "RdW42o73QM0m24nLOR32iL2X3NdL5\\3L`L7`3J\\L:d3FXL=h3ERL?n3BnKa0R4@jKc0W4]OeKf0[4[OaKi0^4YO]Kj0d4VOWKn0i4SOSKP1m4QOoJR1Q5oNkJT1V5lNfJX1Y5jNaJZ1_5gN]J\\1d5dNXJ_1h5bNTJa1l5`NPJc1P6_NjIf1V6ZNfIi1Z6XNbIk1^6VN^Im1c6SNYIP2g6QNUIR2k6PNoHU2P7lMlHW2U7iMgHZ2Y7gMcH\\2]7eM_H^2b7cMXHa2h7`MTHc2l7^MPHf2o7=UH\\Ll7a3aHTL_7j3mHlKS7R4R1O1O1O1O1O001O1O1O1O001O1O1O1O001001O00001O01O01O00010O3M3M3N2M3M3M3N2M4M2N6I;F:F:F:E<E3M00O01000000O10O1000O1000000O0101O000000000O2iMjIYMV6f2RJSMn5l2VJQMj5m2ZJQMf5n2]JPMc5o2aJmL`5R3cJlL^5S3eJjL[5U3kJfLU5Z3VK[Lj4e3XKYLh4g3[KVLe4j3^KRLc4n3`KoKa4P4\\2O100000001O0O10000000001O00000O1000001O0O100000001N1000000O10O1000O10O1000O100O01000O100O010OWI]Ln3i3jKYLV4n3`KUL`4R4WKPLh4Y4nJiKQ5g4WJbKg5Q6O2M2O1N2N200O100O10000O10000O101O0O100O10000O100O10000O10nNTJhJj5W5]JdJc5Z5dJaJ[5e4hJjJ4<T5i4gKTKZ4k4iKRKY4l4jKQKW4o4kKoJW4n4kKPKW4Q5P201O01O00010O00010O0000010O00010O00010O000010O00010O00010[NmGRNT8k1THPNl7m1YHPNg7n1^HPNb7m1bHQN_7l1eHSN[7j1iHTNW7k1lHTNT7i1oHVNR7j1nHVNQ7j1PIUNCROb6i2kIUNASOe6g2kIUN^OUOi6f2iIUN[OWOl6d2jITNYOXOo6d2hITNVOYOT7c2eIUNTOZOX7`2fIVNoNZO\\7a2eIUNlN[Oa7_2dIVNiN\\Od7_2bIUNgN]Oi7^2aI`N_6a1aI]Na6c1_I[Nc6e1^IYNc6g1]IWNd6U2QIiMQ7b2eH\\M\\7o2YHoLi7V3RHhLo7Y3QHfLP8Y3RHdLP8\\3PHbLQ8_3oG_LS8`3oG^LR8b3nG\\LT8d3lGZLU8f3lGXLV8h3kGULX8j3hGRL]8l3b0O1C=^Ob0I7L4M3L4L[1fN5K5J6K5JRX7"}, "label": "an elephant crossing the path in front of a vehicle"}]}
{"id": 569613, "image": "COCO_train2014_000000569613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big south african wild elephant follow another elephant\"."}], "task": "refering", "answer_template": "The \"a big south african wild elephant follow another elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 195, 210, 211, 214, 215, 216, 218, 219, 232, 233, 234, 235, 237, 238, 239], "bbox": [0.115265625, 0.23470588235294118, 0.54990625, 0.733129411764706], "iscrowd": 0, "area": 35290.57744999999, "rle": {"size": [425, 640], "counts": "YPo03S=3M3M3M3M3VMM[H6R7f0bH]OQ7X1dHkNY7_1]HdNa7e1UH^Ni7e1RH_Nl7b1RHaNk7b1RHaNl7a1QHbNm7`1PHcNn7^1oGfNn7]1oGfNo7\\1nGgNP8[1mGhNQ8Y3O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O100O100O1O100O2O0O100O1O10O0100O1O1000_NcIVL]6h3kIRLU6l3oIRLQ6m3SJPLm5n3WJPLi5n3[JPLe5n3_JPLa5n3bJQL^5o3cJPL]5o3eJoK\\5P4fJoKZ5Q4gJnKY5Q4P200O100O1O010O10000000001N10000000000O2O00000000000O101O00000000000O2O0000000000000O101O0O100O100O100O100O100O100O01aHeLP5_3fJhLY5\\3^JiLb5[3TJlLk5X3lImLS6^3^IhL`6d3oHbLP7m4N2N1O2N1O2N1O1O1O1O010000O10000O101O0EYITJg6e5aIYJ`6a5gI^JY6[5nIeJR6U5UJiJl5Q5[JnJe5l4aJTK_5f4e1011N1O2N1O2N1O2N1O2O1N1O2N1O2N1Og0YO1O00100O00YMWIoNh6j0dIQO\\6h0PJTOo5f0[JVOf5e0aJYO^5g0eJVO[5j0gJRO[5n0hJiN^5V1eJcNa5\\1bJ]Nb5c1`JXNc5g1aJSNc5l1_JnMe5R2]JiMf5W2]JbMh5\\2c21O1O010O1O10O01O100O1O010O100O10O10O0100O001O001O1O001O001O10O8I7I6I4M2M4M3M3N1101O1L4K5L4L4L5J8I8H8G8I?Ag0XO2O1O1N2O1O1O1N100000O01000O1000O0100O10N2K4LhXg3"}, "label": "a big south african wild elephant follow another elephant"}]}
{"id": 569613, "image": "COCO_train2014_000000569613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a elephant following the herd\"."}], "task": "refering", "answer_template": "The \"a elephant following the herd\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 195, 210, 211, 214, 215, 216, 218, 219, 232, 233, 234, 235, 237, 238, 239], "bbox": [0.115265625, 0.23470588235294118, 0.54990625, 0.733129411764706], "iscrowd": 0, "area": 35290.57744999999, "rle": {"size": [425, 640], "counts": "YPo03S=3M3M3M3M3VMM[H6R7f0bH]OQ7X1dHkNY7_1]HdNa7e1UH^Ni7e1RH_Nl7b1RHaNk7b1RHaNl7a1QHbNm7`1PHcNn7^1oGfNn7]1oGfNo7\\1nGgNP8[1mGhNQ8Y3O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O100O100O1O100O2O0O100O1O10O0100O1O1000_NcIVL]6h3kIRLU6l3oIRLQ6m3SJPLm5n3WJPLi5n3[JPLe5n3_JPLa5n3bJQL^5o3cJPL]5o3eJoK\\5P4fJoKZ5Q4gJnKY5Q4P200O100O1O010O10000000001N10000000000O2O00000000000O101O00000000000O2O0000000000000O101O0O100O100O100O100O100O100O01aHeLP5_3fJhLY5\\3^JiLb5[3TJlLk5X3lImLS6^3^IhL`6d3oHbLP7m4N2N1O2N1O2N1O1O1O1O010000O10000O101O0EYITJg6e5aIYJ`6a5gI^JY6[5nIeJR6U5UJiJl5Q5[JnJe5l4aJTK_5f4e1011N1O2N1O2N1O2N1O2O1N1O2N1O2N1Og0YO1O00100O00YMWIoNh6j0dIQO\\6h0PJTOo5f0[JVOf5e0aJYO^5g0eJVO[5j0gJRO[5n0hJiN^5V1eJcNa5\\1bJ]Nb5c1`JXNc5g1aJSNc5l1_JnMe5R2]JiMf5W2]JbMh5\\2c21O1O010O1O10O01O100O1O010O100O10O10O0100O001O001O1O001O001O10O8I7I6I4M2M4M3M3N1101O1L4K5L4L4L5J8I8H8G8I?Ag0XO2O1O1N2O1O1O1N100000O01000O1000O0100O10N2K4LhXg3"}, "label": "a elephant following the herd"}]}
{"id": 454928, "image": "COCO_train2014_000000454928.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard tilted on one side being dragged by a child wearing a red helmet\"."}], "task": "refering", "answer_template": "The \"a skateboard tilted on one side being dragged by a child wearing a red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 285, 286, 287, 308, 309, 310, 311, 334, 335, 336, 357, 358, 359], "bbox": [0.374046875, 0.65875, 0.643015625, 0.8981041666666666], "iscrowd": 0, "area": 4001.341500000001, "rle": {"size": [480, 640], "counts": "Rj`34i>5L4M201N2N100O101N1O101N1O1[B\\Om<d0PC^OP=d0lB_OT=a0jB@W=`0hBAX=?hB@Y=`0fBAZ=?eBA\\=?cBA^=>bBC^==bBB_=>`BAb=?]B@e=`0ZB_Oh=a0WB^Ok=k01N2N110000000O1O1O1O10O01O1O001O1O0O2N2O01000O01000O01000O01000O01000O01000O01000O010000O01000O01000O01000O01000O01000O01000O01000O10O01000O01000O01TB^OY=b0bBD]=;bBG]=:bBF_=:`BFa=9_BH`=9_BGb=9\\BHe=7YBKg=f02O1N1O2O1N1N3O1O001O1O001N2O001O1O001O1O010O1O00001O0010O01O00001O0010O01O00001O0010O01O00001O0010O01O0O101O001N103M2Nf^Z3"}, "label": "a skateboard tilted on one side being dragged by a child wearing a red helmet"}]}
{"id": 454928, "image": "COCO_train2014_000000454928.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy ' s skateboard\"."}], "task": "refering", "answer_template": "The \"young boy ' s skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 285, 286, 287, 308, 309, 310, 311, 334, 335, 336, 357, 358, 359], "bbox": [0.374046875, 0.65875, 0.643015625, 0.8981041666666666], "iscrowd": 0, "area": 4001.341500000001, "rle": {"size": [480, 640], "counts": "Rj`34i>5L4M201N2N100O101N1O101N1O1[B\\Om<d0PC^OP=d0lB_OT=a0jB@W=`0hBAX=?hB@Y=`0fBAZ=?eBA\\=?cBA^=>bBC^==bBB_=>`BAb=?]B@e=`0ZB_Oh=a0WB^Ok=k01N2N110000000O1O1O1O10O01O1O001O1O0O2N2O01000O01000O01000O01000O01000O01000O01000O010000O01000O01000O01000O01000O01000O01000O01000O10O01000O01000O01TB^OY=b0bBD]=;bBG]=:bBF_=:`BFa=9_BH`=9_BGb=9\\BHe=7YBKg=f02O1N1O2O1N1N3O1O001O1O001N2O001O1O001O1O010O1O00001O0010O01O00001O0010O01O00001O0010O01O00001O0010O01O0O101O001N103M2Nf^Z3"}, "label": "young boy ' s skateboard"}]}
{"id": 307475, "image": "COCO_train2014_000000307475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player hitting a ball\"."}], "task": "refering", "answer_template": "The \"a baseball player hitting a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 59, 60, 61, 62, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 203, 204, 205, 206, 207, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 261, 262, 263, 272, 273, 274, 275, 276, 279, 280, 281, 290, 291, 292, 298, 299, 300, 318], "bbox": [0.12172, 0.1367, 0.7472000000000001, 0.9494400000000001], "iscrowd": 0, "area": 52132.4273, "rle": {"size": [500, 500], "counts": "dVn02`?4M3M3L3N3M3L4M3N2N1N2O1O1O1N2O1O2M2O1O1O1N2O1O1N2O001O0O10001O0O10001N1000001N10000O101O000OTO]BHd=7_BGa=9aBE^=<dBB\\=>gB_OX=b0jB\\OV=d0lBZOS=f0PCXOP=h0SCTOm<m0UCQOj<P1XCnNh<R1ZClNe<T1m0100O1O1hKoNmIQ1m5VOQJj0k5\\OTJd0f5CXJ>Z3jNoJo0e18W3UOgJj0Q21T3_O_Jg0[2JR3JZJ`0b2GP34UJ8j2Dn2=QJ2n2Bo2e0jILU3@o2l0dIG\\3]On2Q3PMoLo2T3oLmLo2V3PMjLo2X3oLiLo2Z3PMfLn2^3oLbLP3a3nL`LP3c3oL]Lo2f3oL[LP3g3nLZLP3i3oLWLo2l3oLTLP3o3nLRLP3R4nLnKP3U4nLlKP3W4oLiKP3Y4nLgKQ3\\4mLeKQ3^4nLbKP3a4nL`KP3c4nL^KP3f4nLZKP3i4nLWKR3k4lLVKS3l4lLTKT3m4jLTKU3n4jLRKV3Q5fLoJZ3U5bLkJ^3Y5_LgJa3\\5[LdJe3a5VL_Jk3d5RL\\Jm3h5oKXJQ4l5kKTJV4P6fKPJY4T6cKlI^4W6_KhIa4\\6[KeIc4`6YK`If4f6VKYIh4m6TKTIj4Q7RKoHl4m8O2M2N2100O1O010O1O100O1O010O1O100O1O010O1O100O00101N1O100O1O2O0O1O100O1O2O0O1O100O1O2N100VKaJ^O`5b0eJXO]5f0hJVOZ5h0kJROW5m0mJnNV5o0mJfN]5Y1eJ]Ne5`1_JXNg5g1[JXNe5f1^JZNb5d1`J[N`5d1bJ[N_5d1bJ\\N]5c1eJ\\N[5c1gJ]NX5c1iJ\\NW5c1kJ]NT5b1nJ]NQ5d1PK\\No4c1SK\\Nm4cMaI]3d1POj4`MeI`3c1nNg4_MiIc3b1nNd4[MnIf3`1nNb4YMPJj3_1mNR5R1PKmNo4S1SKmNl4S1UKlNj4T1XKlNg4S1[KlNe4T1\\KlNb4T1`KkN`4T1bKlN\\4U1dKkN\\4T1fKlNX4T1jKkNV4U1kKkNT4T1nKkNQ4U1QLkNn3U1SLjNl3V1ULkNj3T1XLkNg3V1ZLjNe3U1]LjNc3U1_LkN_3V1bLiN^3V1dLjN[3U1fLkNZ3T1hLlNW3S1kLmNU3Q1mLoNR3P1PMQOn2n0TMROk2m0WMSOh2l0YMUOf2j0\\MVOc2i0_MWO`2h0bMXO]2g0eMYOZ2e0iM[OdLbMP5Q3]N_O[LeMY5k2]NOd10^NNc11^NNc11_NMb12_NMb12_NNa11aNM`12aNMa11`NNa11aNM`12aNM`12aNM`12bNL_13bNL_13bNL_13cNK^14cNK^13dNL^12dNL]13dNL]13dNM\\12fNL[13fNL[13gNKZ14gNKZ14gNKZ14hNJY15hNJZ14gNKZ14hNJY15hNJY15hNJY15iNIX16iNIX16iNJW15kNIV15lNJV14kNKV14lNJT16mNIT16nNHS17nNHS17nNaN]Jl0f6b0ZO\\Og0c0[O[Of0d0]OYOd0f0^OXOb0h0@VOa0i0@WO`0h0AWO`0h0@XOa0g0@XOa0g0@XOa0g0_OYOa0f0AYO`0f0AYO`0f0@ZOa0e0@ZOa0e0@ZOa0e0_O[Ob0d0_O[Ob0d0_O[Oa0e0_O[Ob0d0_O\\Oa0c0@\\Oa0c0_O]Ob0b0_O]Ob0b0_O]Ob0b0^O]Od0b0]O]Oc0c0^O\\Oc0b0^O]Od0a0^O^Oc0a0^O^Oc0`0^O_Od0?^O@c0>_OAb0>^OAc0>[7N2N2O1N2N2O0O2NXim1"}, "label": "a baseball player hitting a ball"}]}
{"id": 307475, "image": "COCO_train2014_000000307475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging a bat at the ball\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging a bat at the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 59, 60, 61, 62, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 203, 204, 205, 206, 207, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 261, 262, 263, 272, 273, 274, 275, 276, 279, 280, 281, 290, 291, 292, 298, 299, 300, 318], "bbox": [0.12172, 0.1367, 0.7472000000000001, 0.9494400000000001], "iscrowd": 0, "area": 52132.4273, "rle": {"size": [500, 500], "counts": "dVn02`?4M3M3L3N3M3L4M3N2N1N2O1O1O1N2O1O2M2O1O1O1N2O1O1N2O001O0O10001O0O10001N1000001N10000O101O000OTO]BHd=7_BGa=9aBE^=<dBB\\=>gB_OX=b0jB\\OV=d0lBZOS=f0PCXOP=h0SCTOm<m0UCQOj<P1XCnNh<R1ZClNe<T1m0100O1O1hKoNmIQ1m5VOQJj0k5\\OTJd0f5CXJ>Z3jNoJo0e18W3UOgJj0Q21T3_O_Jg0[2JR3JZJ`0b2GP34UJ8j2Dn2=QJ2n2Bo2e0jILU3@o2l0dIG\\3]On2Q3PMoLo2T3oLmLo2V3PMjLo2X3oLiLo2Z3PMfLn2^3oLbLP3a3nL`LP3c3oL]Lo2f3oL[LP3g3nLZLP3i3oLWLo2l3oLTLP3o3nLRLP3R4nLnKP3U4nLlKP3W4oLiKP3Y4nLgKQ3\\4mLeKQ3^4nLbKP3a4nL`KP3c4nL^KP3f4nLZKP3i4nLWKR3k4lLVKS3l4lLTKT3m4jLTKU3n4jLRKV3Q5fLoJZ3U5bLkJ^3Y5_LgJa3\\5[LdJe3a5VL_Jk3d5RL\\Jm3h5oKXJQ4l5kKTJV4P6fKPJY4T6cKlI^4W6_KhIa4\\6[KeIc4`6YK`If4f6VKYIh4m6TKTIj4Q7RKoHl4m8O2M2N2100O1O010O1O100O1O010O1O100O1O010O1O100O00101N1O100O1O2O0O1O100O1O2O0O1O100O1O2N100VKaJ^O`5b0eJXO]5f0hJVOZ5h0kJROW5m0mJnNV5o0mJfN]5Y1eJ]Ne5`1_JXNg5g1[JXNe5f1^JZNb5d1`J[N`5d1bJ[N_5d1bJ\\N]5c1eJ\\N[5c1gJ]NX5c1iJ\\NW5c1kJ]NT5b1nJ]NQ5d1PK\\No4c1SK\\Nm4cMaI]3d1POj4`MeI`3c1nNg4_MiIc3b1nNd4[MnIf3`1nNb4YMPJj3_1mNR5R1PKmNo4S1SKmNl4S1UKlNj4T1XKlNg4S1[KlNe4T1\\KlNb4T1`KkN`4T1bKlN\\4U1dKkN\\4T1fKlNX4T1jKkNV4U1kKkNT4T1nKkNQ4U1QLkNn3U1SLjNl3V1ULkNj3T1XLkNg3V1ZLjNe3U1]LjNc3U1_LkN_3V1bLiN^3V1dLjN[3U1fLkNZ3T1hLlNW3S1kLmNU3Q1mLoNR3P1PMQOn2n0TMROk2m0WMSOh2l0YMUOf2j0\\MVOc2i0_MWO`2h0bMXO]2g0eMYOZ2e0iM[OdLbMP5Q3]N_O[LeMY5k2]NOd10^NNc11^NNc11_NMb12_NMb12_NNa11aNM`12aNMa11`NNa11aNM`12aNM`12aNM`12bNL_13bNL_13bNL_13cNK^14cNK^13dNL^12dNL]13dNL]13dNM\\12fNL[13fNL[13gNKZ14gNKZ14gNKZ14hNJY15hNJZ14gNKZ14hNJY15hNJY15hNJY15iNIX16iNIX16iNJW15kNIV15lNJV14kNKV14lNJT16mNIT16nNHS17nNHS17nNaN]Jl0f6b0ZO\\Og0c0[O[Of0d0]OYOd0f0^OXOb0h0@VOa0i0@WO`0h0AWO`0h0@XOa0g0@XOa0g0@XOa0g0_OYOa0f0AYO`0f0AYO`0f0@ZOa0e0@ZOa0e0@ZOa0e0_O[Ob0d0_O[Ob0d0_O[Oa0e0_O[Ob0d0_O\\Oa0c0@\\Oa0c0_O]Ob0b0_O]Ob0b0_O]Ob0b0^O]Od0b0]O]Oc0c0^O\\Oc0b0^O]Od0a0^O^Oc0a0^O^Oc0`0^O_Od0?^O@c0>_OAb0>^OAc0>[7N2N2O1N2N2O0O2NXim1"}, "label": "a baseball player swinging a bat at the ball"}]}
{"id": 307475, "image": "COCO_train2014_000000307475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man playing in ground\"."}], "task": "refering", "answer_template": "The \"man playing in ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 144, 145, 162, 163, 166, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 270, 271, 272, 288], "bbox": [0.0022400000000000002, 0.4045, 0.31236, 0.90786], "iscrowd": 0, "area": 16636.947650000002, "rle": {"size": [500, 500], "counts": "ne0l7i7N101N100O2N101N100O2O0O2N101N100O2O0O2N100O2O0O3mLfHhM\\7j1oHVNT7b1RI^NP7`1PI_NS7_1lHbNV7\\1jHcNY7[1gHeN\\7X1cHhN`7V1`HjNP8f0oGZOX8`0hG@^8:bGFd84[GLl8NTG2l8MTG3m8MSG3m8MSG2o8MPG4P9KQG5o8KPG5Q9KoF5R9IoF6R9JmF7S9ImF6T9ImF7S9IlF8U9GkF8V9HiF9W9FjF9W9GiF9X9FgF:Z9FfF:Z9FeF:\\9EeF;[9EeF;\\9DcF<^9DbF<^9DaF<`9D`F<a9B`F=a9C^F>b9B]F?c9A]F>d9B[F?f9_OZFa0g9_OYFa0g9_OXFa0i9_OWFa0j9^OUFc0k9]OTFc0m9\\OTFd0l9\\OSFd0n9\\OQFe0P:ZOPFe0Q:[OoEe0Q:[OoEe0Q:ZOPFf0Q:YOoEf0R:ZOnEf0R:ZOnEf0R:ZOoEe0Q:[OoEe0R:YOoEf0R:ZOnEf0R:ZOnEf0R:ZOoEe0Q:[OoEe0R:ZOnEf0R:YOoEf0R:ZOnEf0R:ZOoEe0R:ZOnEf0R:ZOnEf0R:YOoEf0R:[OmEe0S:[OnEd0S:[OmEe0S:\\OlEd0T:\\OlEd0T:\\OlEc0T:_OlE`0S:AmE?R:BnE>Q:DnE<Q:EoE;P:FQF8o9JPF6n9LRF3n9NRF1n91RFMn94TFIl98UFGj9;VFCj9>XF_Oh9c0XF[Og9g0[FVOe9k0\\FROe9P1[FmNh9R1ZFiNk9U1VFhNm9W1UFeNn9Z1SFcNQ:[1PFbNS:\\1T2O2N1N2O2N1N2O2N1O2M2O2N1N3N2N1O2M2O2N1N3N1O2M2O2N1OYeW5"}, "label": "man playing in ground"}]}
{"id": 307475, "image": "COCO_train2014_000000307475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher behind the batter\"."}], "task": "refering", "answer_template": "The \"the catcher behind the batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 144, 145, 162, 163, 166, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 216, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 270, 271, 272, 288], "bbox": [0.0022400000000000002, 0.4045, 0.31236, 0.90786], "iscrowd": 0, "area": 16636.947650000002, "rle": {"size": [500, 500], "counts": "ne0l7i7N101N100O2N101N100O2O0O2N101N100O2O0O2N100O2O0O3mLfHhM\\7j1oHVNT7b1RI^NP7`1PI_NS7_1lHbNV7\\1jHcNY7[1gHeN\\7X1cHhN`7V1`HjNP8f0oGZOX8`0hG@^8:bGFd84[GLl8NTG2l8MTG3m8MSG3m8MSG2o8MPG4P9KQG5o8KPG5Q9KoF5R9IoF6R9JmF7S9ImF6T9ImF7S9IlF8U9GkF8V9HiF9W9FjF9W9GiF9X9FgF:Z9FfF:Z9FeF:\\9EeF;[9EeF;\\9DcF<^9DbF<^9DaF<`9D`F<a9B`F=a9C^F>b9B]F?c9A]F>d9B[F?f9_OZFa0g9_OYFa0g9_OXFa0i9_OWFa0j9^OUFc0k9]OTFc0m9\\OTFd0l9\\OSFd0n9\\OQFe0P:ZOPFe0Q:[OoEe0Q:[OoEe0Q:ZOPFf0Q:YOoEf0R:ZOnEf0R:ZOnEf0R:ZOoEe0Q:[OoEe0R:YOoEf0R:ZOnEf0R:ZOnEf0R:ZOoEe0Q:[OoEe0R:ZOnEf0R:YOoEf0R:ZOnEf0R:ZOoEe0R:ZOnEf0R:ZOnEf0R:YOoEf0R:[OmEe0S:[OnEd0S:[OmEe0S:\\OlEd0T:\\OlEd0T:\\OlEc0T:_OlE`0S:AmE?R:BnE>Q:DnE<Q:EoE;P:FQF8o9JPF6n9LRF3n9NRF1n91RFMn94TFIl98UFGj9;VFCj9>XF_Oh9c0XF[Og9g0[FVOe9k0\\FROe9P1[FmNh9R1ZFiNk9U1VFhNm9W1UFeNn9Z1SFcNQ:[1PFbNS:\\1T2O2N1N2O2N1N2O2N1O2M2O2N1N3N2N1O2M2O2N1N3N1O2M2O2N1OYeW5"}, "label": "the catcher behind the batter"}]}
{"id": 307475, "image": "COCO_train2014_000000307475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player waiting for his turn at bat\"."}], "task": "refering", "answer_template": "The \"a baseball player waiting for his turn at bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 86, 87, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 211, 212, 213, 214, 229, 230, 231, 232, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 301], "bbox": [0.71846, 0.19932, 0.97748, 0.89978], "iscrowd": 0, "area": 25725.29675, "rle": {"size": [500, 500], "counts": "[g_52^?4L4L4L4N2N2O1O1SIZOfNl0S1[O[NW1^1POPNb1h1eNgMm1W2UNdMP2Z2QNcMS2\\2nMVM`2h2bMfLQ3W3PM[L_3c3cLVLd3i3\\LPLl3n3VLkKQ4S4QLfKV4X4kKaK]4^4dK_JnNDe5l5]K[JSOBg5S6WKUJXOAh5Z6PKnI@Be5_6kJjIGG\\5_6nJdIMLU5_6nJeIOLR5^6PKeIONo4]6SKdI0Ol4]6SKcI30j4\\6TKcI41c4`6YK^I53\\4d6^KYI82V4i6UMXIh2i6ZMWIb2l6^MTI_2o6bMQI[2P7fMQIV2R7kMmHS2U7mMlHo1W7RNiHj1Y7WNgHg1[7ZNeHb1^7^NcH^1_7dN`HZ1b7fN_HV1d7kN\\H_O<YNS7`2aHUOP1gMP6f3PI`NU8o1lGmMS8X2lGfMR8^2oG_MP8e2PHWMS8i2mGUMU8l2kGPMX8P3iGmLY8T3fGjL\\8V3eGiL[8X3eGiLY8W3gGlLV8U3jGlLT8T3mGnLP8S3PHnLn7R3RHQMk7P3UHQMi7o2XHSMe7n2ZHTMd7l2]HVM`7k2`HVM_7i2aHYM]7h2cHZMZ7f2gH[MW7f2hH]MU7c2lH^MU7`2kHbMW7Y2iHiMX7U2hHmMX7P2iHQNW7n1hHEh58YJIg56YJKh53WJOi5OXJ2i5LWJ5V6\\OjIf0T6[OlIf0S6ZOmIf0R6ZOnIh0Q6XOoIi0o5WORJh0o5XOPJh0P6XOQJh0o5XOQJg0o5ZOQJe0P6ZOPJf0P6[OPJd0Q6[OPJd0P6]OoIc0R6]OnIc0Q6]OPJb0P6_OoIa0R6^OoIa0Q6@oISLJm3X60bJJ^56hJDY5<mJ^OR5ZO`I\\Nd1T2m4XOfIeNb1m1h4VOmIoN`1e1d4TORJeNS2Q2k3ROZJmNQ2k1f3oN`JVOQ2e1_3oNfJ[Of2l0e2BjJBT4ZOR1o0oJGn3ZOT1j0RKLc:McE2^:IfE8[:BjE=W:]OPFa0d<N2M3M3N2M4L3NkS5"}, "label": "a baseball player waiting for his turn at bat"}]}
{"id": 307475, "image": "COCO_train2014_000000307475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player standing in the on deck circle while another player is at the plate\"."}], "task": "refering", "answer_template": "The \"a baseball player standing in the on deck circle while another player is at the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 86, 87, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 211, 212, 213, 214, 229, 230, 231, 232, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 301], "bbox": [0.71846, 0.19932, 0.97748, 0.89978], "iscrowd": 0, "area": 25725.29675, "rle": {"size": [500, 500], "counts": "[g_52^?4L4L4L4N2N2O1O1SIZOfNl0S1[O[NW1^1POPNb1h1eNgMm1W2UNdMP2Z2QNcMS2\\2nMVM`2h2bMfLQ3W3PM[L_3c3cLVLd3i3\\LPLl3n3VLkKQ4S4QLfKV4X4kKaK]4^4dK_JnNDe5l5]K[JSOBg5S6WKUJXOAh5Z6PKnI@Be5_6kJjIGG\\5_6nJdIMLU5_6nJeIOLR5^6PKeIONo4]6SKdI0Ol4]6SKcI30j4\\6TKcI41c4`6YK^I53\\4d6^KYI82V4i6UMXIh2i6ZMWIb2l6^MTI_2o6bMQI[2P7fMQIV2R7kMmHS2U7mMlHo1W7RNiHj1Y7WNgHg1[7ZNeHb1^7^NcH^1_7dN`HZ1b7fN_HV1d7kN\\H_O<YNS7`2aHUOP1gMP6f3PI`NU8o1lGmMS8X2lGfMR8^2oG_MP8e2PHWMS8i2mGUMU8l2kGPMX8P3iGmLY8T3fGjL\\8V3eGiL[8X3eGiLY8W3gGlLV8U3jGlLT8T3mGnLP8S3PHnLn7R3RHQMk7P3UHQMi7o2XHSMe7n2ZHTMd7l2]HVM`7k2`HVM_7i2aHYM]7h2cHZMZ7f2gH[MW7f2hH]MU7c2lH^MU7`2kHbMW7Y2iHiMX7U2hHmMX7P2iHQNW7n1hHEh58YJIg56YJKh53WJOi5OXJ2i5LWJ5V6\\OjIf0T6[OlIf0S6ZOmIf0R6ZOnIh0Q6XOoIi0o5WORJh0o5XOPJh0P6XOQJh0o5XOQJg0o5ZOQJe0P6ZOPJf0P6[OPJd0Q6[OPJd0P6]OoIc0R6]OnIc0Q6]OPJb0P6_OoIa0R6^OoIa0Q6@oISLJm3X60bJJ^56hJDY5<mJ^OR5ZO`I\\Nd1T2m4XOfIeNb1m1h4VOmIoN`1e1d4TORJeNS2Q2k3ROZJmNQ2k1f3oN`JVOQ2e1_3oNfJ[Of2l0e2BjJBT4ZOR1o0oJGn3ZOT1j0RKLc:McE2^:IfE8[:BjE=W:]OPFa0d<N2M3M3N2M4L3NkS5"}, "label": "a baseball player standing in the on deck circle while another player is at the plate"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the human arm that is behind the girl in the front\"."}], "task": "refering", "answer_template": "The \"the human arm that is behind the girl in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 257, 258, 259, 281, 282], "bbox": [0.133328125, 0.4123541666666667, 0.301859375, 0.7827083333333335], "iscrowd": 0, "area": 9868.819549999998, "rle": {"size": [480, 640], "counts": "STX11i>7I7H8I7L4N2O1N2N101N2O1N2N2O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1N2N2O1N2N2N2N2N2N2C=K5K5L4L4L4L4L4M3K5O1N2O1N2N2O1N2O1O1002N00001O001O00001O001O00001O001O00001O00001J`EiK`:R4hEkKY:n3`0K6I8I7H8I7H7J4K6K5J6K5J6K5J5K6K5J6K5J6K5JnXa6"}, "label": "the human arm that is behind the girl in the front"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man ' s arm\"."}], "task": "refering", "answer_template": "The \"man ' s arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 257, 258, 259, 281, 282], "bbox": [0.133328125, 0.4123541666666667, 0.301859375, 0.7827083333333335], "iscrowd": 0, "area": 9868.819549999998, "rle": {"size": [480, 640], "counts": "STX11i>7I7H8I7L4N2O1N2N101N2O1N2N2O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1N2N2O1N2N2N2N2N2N2C=K5K5L4L4L4L4L4M3K5O1N2O1N2N2O1N2O1O1002N00001O001O00001O001O00001O001O00001O00001J`EiK`:R4hEkKY:n3`0K6I8I7H8I7H7J4K6K5J6K5J6K5J5K6K5J6K5J6K5JnXa6"}, "label": "man ' s arm"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe without its tongue sticking out\"."}], "task": "refering", "answer_template": "The \"a giraffe without its tongue sticking out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 90, 91, 114, 137, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 270, 271, 272, 273], "bbox": [0.753375, 0.16179166666666667, 1.0, 0.6853958333333333], "iscrowd": 0, "area": 14918.607450000001, "rle": {"size": [480, 640], "counts": "ZWR72m>5J6J7I6K5J6J3N1N2O2O0O1O100O2O0O1O100O2O0O100O00100O1O1O0O2O1O1N101N2O1N2O10O0100O1O100O010O100O1O1K4I8H8L40000O10O1000O100000O010000O100000O01000000O10O1000O01O001O0000000000001O0O100O1O10TLTMeKk2]4ZM^Kd2d4\\M[Kd2f4]MXKb2j4^MUKa2m4`MQK`2P5aMnJ^2S5cMlJ\\2V5eMhJ[2Y5eMgJY2[5hMcJX2^5iM`JV2b5jM]JU2e5lMYJT2h5lMWJU2i5lMUJT2l5mMSJS2m5mMRJT2n5mMPJT2P6lMoIT2R6mMlIT2S6mMlIT2T6mMjIS2W6nMhIR2X6nMgIS2Y6nMeIR2\\6nMcIS2]6nMaIS2_6nM_IR2b6nM]IS2c6nM\\IR2d6nM[IR2f6oMXIR2h6oMVIR2i6WNnHi1S7YNjHh1V7YNhHh1X7YNgHf1Z7[NdHf1\\7[NbHe1_7[N`Hf1`7[N^Hf1b7[NWGfNg0Q3S8gNiG[1W8fNeG\\1\\8eNaG]1_8eN\\G^1d8cNYG^1i8bNSGa1m8aNoFa1R9_NjFc1X9R20O2M2M4L4K6Ka0^Oc0^Oa0_Ob0iMXM"}, "label": "a giraffe without its tongue sticking out"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second giraffe is looking at the lady ' s hands\"."}], "task": "refering", "answer_template": "The \"second giraffe is looking at the lady ' s hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 90, 91, 114, 137, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 270, 271, 272, 273], "bbox": [0.753375, 0.16179166666666667, 1.0, 0.6853958333333333], "iscrowd": 0, "area": 14918.607450000001, "rle": {"size": [480, 640], "counts": "ZWR72m>5J6J7I6K5J6J3N1N2O2O0O1O100O2O0O1O100O2O0O100O00100O1O1O0O2O1O1N101N2O1N2O10O0100O1O100O010O100O1O1K4I8H8L40000O10O1000O100000O010000O100000O01000000O10O1000O01O001O0000000000001O0O100O1O10TLTMeKk2]4ZM^Kd2d4\\M[Kd2f4]MXKb2j4^MUKa2m4`MQK`2P5aMnJ^2S5cMlJ\\2V5eMhJ[2Y5eMgJY2[5hMcJX2^5iM`JV2b5jM]JU2e5lMYJT2h5lMWJU2i5lMUJT2l5mMSJS2m5mMRJT2n5mMPJT2P6lMoIT2R6mMlIT2S6mMlIT2T6mMjIS2W6nMhIR2X6nMgIS2Y6nMeIR2\\6nMcIS2]6nMaIS2_6nM_IR2b6nM]IS2c6nM\\IR2d6nM[IR2f6oMXIR2h6oMVIR2i6WNnHi1S7YNjHh1V7YNhHh1X7YNgHf1Z7[NdHf1\\7[NbHe1_7[N`Hf1`7[N^Hf1b7[NWGfNg0Q3S8gNiG[1W8fNeG\\1\\8eNaG]1_8eN\\G^1d8cNYG^1i8bNSGa1m8aNoFa1R9_NjFc1X9R20O2M2M4L4K6Ka0^Oc0^Oa0_Ob0iMXM"}, "label": "second giraffe is looking at the lady ' s hands"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe sticking it ' s tongue out at a spectator\"."}], "task": "refering", "answer_template": "The \"a giraffe sticking it ' s tongue out at a spectator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 41, 42, 62, 64, 65, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 265, 266, 267, 268, 287, 288, 289, 290, 310, 311, 312, 333, 334], "bbox": [0.47634375, 0.10472916666666668, 0.9864843750000001, 0.8862708333333332], "iscrowd": 0, "area": 40272.802749999995, "rle": {"size": [480, 640], "counts": "l[_46f>8H6J5L3O1N2O1O1N101O001N2O000O1N2N1O2O1N1O2N2N101N2N1O2N2]O[NjCf1k;jNoCX1P<l0N3N1N2O1N20001O1O001O1O1O1O1O001O1O1O1O001N2O1O1N101O1N2O001N2O0\\KWLRNj3n1XLnMj3R2XLkMi3U2YLhMg3Y2[LcMg3]2[L`Mf3`2\\L]Me3c2\\LZMe3f2^LWMc3i2^LUMc3k2_LQMb3P3_LnLa3S3aLiL`3Y3`LeL_3]3cL`L]3a3eL[L\\3f3gLVLX3m3SMfKm2[4dMRK\\2P5gMlJY2U5jMfJW2\\5kM_JU2c5nMSJX2n5kMhI[2Y6hM]I]2e6dMSIb2o6_MhHe2[7]23N2M4L3N2O1O2N1O1O1O1O2N1O1O1O1O2N1O1O2N4L3M2N2M2O2N2N1O1O1O001OE;K6N1O1O1O1O1O2O0O1O1O1O100O2J5K5M3N2M3N3L3O1O1O1O1O1100O1O1O1O1O1O100O1O1O2N2N6J<E8G9G5K5L3L5L2N001O1O1O1O1O1N2OO10000000000000O010O000000000001O0N2M3M3N2000O1O1N2N101N2N10OA?2O0O1O100O1O\\MWF6h9FfF1[9CYG3i8J[G4f8I^G4e8I^G5c8HaG6`8GdG6^8HeG6\\8GhG6Z8GjG7W8GlG6V8GmG8T8EoG:R8DPH:R8CQH=P8_OSHa0n7\\OTHd0m7XOVHg0l7VOVHj0j7SOYHm0m:00O10O1000O1000O100O01000O100O1O1O100O1O1O010O1O100O100O10000O10000O0100O1O100O100O1O100O100O1O100O100O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O101N1O2N1O2O1N3M4L[R4"}, "label": "a giraffe sticking it ' s tongue out at a spectator"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the head of the giraffe that the woman is feeding\"."}], "task": "refering", "answer_template": "The \"the head of the giraffe that the woman is feeding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 41, 42, 62, 64, 65, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 265, 266, 267, 268, 287, 288, 289, 290, 310, 311, 312, 333, 334], "bbox": [0.47634375, 0.10472916666666668, 0.9864843750000001, 0.8862708333333332], "iscrowd": 0, "area": 40272.802749999995, "rle": {"size": [480, 640], "counts": "l[_46f>8H6J5L3O1N2O1O1N101O001N2O000O1N2N1O2O1N1O2N2N101N2N1O2N2]O[NjCf1k;jNoCX1P<l0N3N1N2O1N20001O1O001O1O1O1O1O001O1O1O1O001N2O1O1N101O1N2O001N2O0\\KWLRNj3n1XLnMj3R2XLkMi3U2YLhMg3Y2[LcMg3]2[L`Mf3`2\\L]Me3c2\\LZMe3f2^LWMc3i2^LUMc3k2_LQMb3P3_LnLa3S3aLiL`3Y3`LeL_3]3cL`L]3a3eL[L\\3f3gLVLX3m3SMfKm2[4dMRK\\2P5gMlJY2U5jMfJW2\\5kM_JU2c5nMSJX2n5kMhI[2Y6hM]I]2e6dMSIb2o6_MhHe2[7]23N2M4L3N2O1O2N1O1O1O1O2N1O1O1O1O2N1O1O2N4L3M2N2M2O2N2N1O1O1O001OE;K6N1O1O1O1O1O2O0O1O1O1O100O2J5K5M3N2M3N3L3O1O1O1O1O1100O1O1O1O1O1O100O1O1O2N2N6J<E8G9G5K5L3L5L2N001O1O1O1O1O1N2OO10000000000000O010O000000000001O0N2M3M3N2000O1O1N2N101N2N10OA?2O0O1O100O1O\\MWF6h9FfF1[9CYG3i8J[G4f8I^G4e8I^G5c8HaG6`8GdG6^8HeG6\\8GhG6Z8GjG7W8GlG6V8GmG8T8EoG:R8DPH:R8CQH=P8_OSHa0n7\\OTHd0m7XOVHg0l7VOVHj0j7SOYHm0m:00O10O1000O1000O100O01000O100O1O1O100O1O1O010O1O100O100O10000O10000O0100O1O100O100O1O100O100O1O100O100O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O101N1O2N1O2O1N3M4L[R4"}, "label": "the head of the giraffe that the woman is feeding"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde haired girl smiling at a giraffe\"."}], "task": "refering", "answer_template": "The \"a blonde haired girl smiling at a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 24, 25, 26, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 306, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0033750000000000004, 0.0449375, 0.43989062500000004, 0.9887708333333334], "iscrowd": 0, "area": 67340.2424, "rle": {"size": [480, 640], "counts": "^[1Z1Y9]4cK]4RLn300000000001O000001O0001O000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000000000001O00000000000000001O00000000000000001O1O001O1bIhNROX1n0ROhNo0V1YOcNh0[1[OcNf0[1\\OdNd0\\1]OcNd0[1_OcNb0[1@dN`0\\1AcN`0[1CcN>[1DdN=Z1EeN;[1GcN:[1HdN9Z1IeN7[1KcN6[1LdN5Z1MeN4Z1MeN3Z1OeN2Z1OeN2Y10fN0Z11eN0Y12fNOY12fNOY12fNNY15eNLZ15eNLY16fNJZ17eNJZ17eNJY18fNIY18fNHY1:fNGY1:fNGX1;gNEY1<fNEY1=eNDY1>fNCW1`0hN@T1e0kN\\OP1i0oNXOl0m0TOROh0S1WOnNd0W1[OjN>]1AdN8c1G]N4i1KXNJT24mMKT24lMKV24kMKV24kMJW25jMJW25iMKX24iMJY25hMJY25hMJY25gMJ\\24eMiN`M\\Nl4j2eMgNbM]Nk4k2fMcNcM`Nh4l2iM_NbMcNf4m2lMZNaMgNe4n2nMUNaMkNb4o2QNPNaMoN_4P3TNlM`MQO^4R3VNeMbMVOZ4T3WN]MhMZOT4X3XNTMlM@P4Z3XNlLe3S3U4O1O1O1O1O1O1O1O1O2N1O1O1O2N2N2N3M2N2N2N2N2N2N1O0000O1000000O10000O1000000O1000000O100O1O1O100O1O1O1O100O1O1O1O100O1OfMmCi1R<WNPDh1o;WNSDi1m;UNVDj1i;VNYDi1f;VN\\Dj1c;VN_Di1`;VNcDi1];UNeDk1Z;UNhDj1Y;SNhDn1U<0000000O1000000O1000000O1000001N1000000O100O1O100O1O01O000O2O00000O2O01O01O0001O01O00101N1O1O2O0O1O4L6I7G9G9G8H9G8Hk\\W5"}, "label": "a blonde haired girl smiling at a giraffe"}]}
{"id": 266515, "image": "COCO_train2014_000000266515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the gray shirt holding the green phone\"."}], "task": "refering", "answer_template": "The \"the woman in the gray shirt holding the green phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 24, 25, 26, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 306, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0033750000000000004, 0.0449375, 0.43989062500000004, 0.9887708333333334], "iscrowd": 0, "area": 67340.2424, "rle": {"size": [480, 640], "counts": "^[1Z1Y9]4cK]4RLn300000000001O000001O0001O000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000000000001O00000000000000001O00000000000000001O1O001O1bIhNROX1n0ROhNo0V1YOcNh0[1[OcNf0[1\\OdNd0\\1]OcNd0[1_OcNb0[1@dN`0\\1AcN`0[1CcN>[1DdN=Z1EeN;[1GcN:[1HdN9Z1IeN7[1KcN6[1LdN5Z1MeN4Z1MeN3Z1OeN2Z1OeN2Y10fN0Z11eN0Y12fNOY12fNOY12fNNY15eNLZ15eNLY16fNJZ17eNJZ17eNJY18fNIY18fNHY1:fNGY1:fNGX1;gNEY1<fNEY1=eNDY1>fNCW1`0hN@T1e0kN\\OP1i0oNXOl0m0TOROh0S1WOnNd0W1[OjN>]1AdN8c1G]N4i1KXNJT24mMKT24lMKV24kMKV24kMJW25jMJW25iMKX24iMJY25hMJY25hMJY25gMJ\\24eMiN`M\\Nl4j2eMgNbM]Nk4k2fMcNcM`Nh4l2iM_NbMcNf4m2lMZNaMgNe4n2nMUNaMkNb4o2QNPNaMoN_4P3TNlM`MQO^4R3VNeMbMVOZ4T3WN]MhMZOT4X3XNTMlM@P4Z3XNlLe3S3U4O1O1O1O1O1O1O1O1O2N1O1O1O2N2N2N3M2N2N2N2N2N2N1O0000O1000000O10000O1000000O1000000O100O1O1O100O1O1O1O100O1O1O1O100O1OfMmCi1R<WNPDh1o;WNSDi1m;UNVDj1i;VNYDi1f;VN\\Dj1c;VN_Di1`;VNcDi1];UNeDk1Z;UNhDj1Y;SNhDn1U<0000000O1000000O1000000O1000001N1000000O100O1O100O1O01O000O2O00000O2O01O01O0001O01O00101N1O1O2O0O1O4L6I7G9G9G8H9G8Hk\\W5"}, "label": "the woman in the gray shirt holding the green phone"}]}
{"id": 209178, "image": "COCO_train2014_000000209178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy in a yellow shirt\"."}], "task": "refering", "answer_template": "The \"young boy in a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 86, 87, 88, 89, 108, 109, 110, 111, 112, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.6713281249999999, 0.15955082742316784, 0.983234375, 0.7842789598108747], "iscrowd": 0, "area": 31613.66935, "rle": {"size": [423, 640], "counts": "ija5`0e<4L4L5J5L4L5J5L4L4L5J5M3N2M4M2N2N2M4M2N2M:G9G9G9F10000O10000O1000000O101O0O1ZNVLoIj3n5bLiI^3U6gLhIY3U6lLhIU3V6nLiIR3T6SMjIm2T6WMiIj2T6[MjIe2S6_MlIa2R6cMlI]2Q6hMlIY2R6jMmIV2P6oMmIS2P6QNlIQ2Q6TNlIm1R6VNkIl1R6f2N2M3O101O000O10000O10001N10000O10000O2O000O10000O101O0O1000000O101O0O10000O10001N10000O11O1O1O100O001O1O1O1\\KbJ]1_5\\NgJe1X5TNoJl1R5RNoJn1R5QNoJn1R5QNnJo1S5QNlJo1U5PNkJP2U5PNkJQ2U5nMlJQ2U5oMjJQ2W5nMiJR2X5mMhJS2Y5lMgJT2Y5lMhJS2Y5kMhJV2X5hMiJX2X5fMiJZ2X5cMjJ]2W5aMkJ^2U5aMlJ_2U5_MlJa2U5]MlJd2T5ZMnJe2S5XMoJh2W6QLjIo3g70001O0001O0001O000000010O000000002N2O1N2N2N2N2N2N2O2M2N2N2N2N2N2N2O1N2N2N2N3M2N2O1N2N2N2N2N2N2O1N2M4L3M3M:G=B>B4L2N3M2N2N2N3J5KmW4"}, "label": "young boy in a yellow shirt"}]}
{"id": 209178, "image": "COCO_train2014_000000209178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy smiling happily in a yellow shirt\"."}], "task": "refering", "answer_template": "The \"a young boy smiling happily in a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 86, 87, 88, 89, 108, 109, 110, 111, 112, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.6713281249999999, 0.15955082742316784, 0.983234375, 0.7842789598108747], "iscrowd": 0, "area": 31613.66935, "rle": {"size": [423, 640], "counts": "ija5`0e<4L4L5J5L4L5J5L4L4L5J5M3N2M4M2N2N2M4M2N2M:G9G9G9F10000O10000O1000000O101O0O1ZNVLoIj3n5bLiI^3U6gLhIY3U6lLhIU3V6nLiIR3T6SMjIm2T6WMiIj2T6[MjIe2S6_MlIa2R6cMlI]2Q6hMlIY2R6jMmIV2P6oMmIS2P6QNlIQ2Q6TNlIm1R6VNkIl1R6f2N2M3O101O000O10000O10001N10000O10000O2O000O10000O101O0O1000000O101O0O10000O10001N10000O11O1O1O100O001O1O1O1\\KbJ]1_5\\NgJe1X5TNoJl1R5RNoJn1R5QNoJn1R5QNnJo1S5QNlJo1U5PNkJP2U5PNkJQ2U5nMlJQ2U5oMjJQ2W5nMiJR2X5mMhJS2Y5lMgJT2Y5lMhJS2Y5kMhJV2X5hMiJX2X5fMiJZ2X5cMjJ]2W5aMkJ^2U5aMlJ_2U5_MlJa2U5]MlJd2T5ZMnJe2S5XMoJh2W6QLjIo3g70001O0001O0001O000000010O000000002N2O1N2N2N2N2N2N2O2M2N2N2N2N2N2N2O1N2N2N2N3M2N2O1N2N2N2N2N2N2O1N2M4L3M3M:G=B>B4L2N3M2N2N2N3J5KmW4"}, "label": "a young boy smiling happily in a yellow shirt"}]}
{"id": 209178, "image": "COCO_train2014_000000209178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing glasses and a gold crown\"."}], "task": "refering", "answer_template": "The \"the woman wearing glasses and a gold crown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 267, 268, 269, 281, 282, 283, 304, 305, 306], "bbox": [0.190546875, 0.06193853427895981, 0.7085781250000001, 0.9087943262411348], "iscrowd": 0, "area": 61324.147099999995, "rle": {"size": [423, 640], "counts": "`fb11S=5L3M3M3L4M4L3M3M3L5L3kNmNRFV1l9nNoEU1o9oNkEV1R:nNiEU1U:oNfET1W:QOdER1Z:P1M2N2N2M3N3M2N2N4K6K4L4L5K4K5L1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O3L3N2NiN`IjK^6W4fIfKX6Z4mIcKP6_4RJ`Km5`4TJ`Kk5_4WJaKg5`4ZJ[Kj5d4XJWKl5i4UJRKn5o4TJkJP6T5RJfJS6Z5l001000O0100O01000O010O10O10O10O01000O01000O10O10O10O0100O10O0100O10O01O100O100O101N100O2O0O100O2O0O101N100O1O2L3L4L5K4M3O1O2N1O1O1O1O1RO[IYKe6d4aIYK`6c4gIYKZ6c4lIYKV6c4PJZKQ6b4VJZKk5b4[J[Ke5a4c1L4L4L4L3M4O15K6J4L2M2O2N2N1OO1O0O2N2N2N2N2N1O2N2O1N2N2N1O2O11O000O2O004L5K6J2M3M1O2M2O1N2O1N2O2M2O1N2N2M4M2N200O1O101N10bJgHn4i70000001N10N2K5J6K5N23M2M4M3M2N3M3M001O0O101M2M3N3M2N3O0001OkK[Jk0b5TO`Jn0]5POfJR1W5lNkJV1S5iNoJX1o4fNTKZ1k4dNXK]1g4]N_Kc1`4XNfKi1X4RNnKn1Q4lMVLU2g3fM`LZ2_3fMbL[2[3eMgLZ2W3gMjLX2R3mMoLR2n2PNTMo1m2PNTMo1m2oMUMP2l2nMVMQ2k2nMVMQ2k2mMWMR2j2mMWMR2j2iM\\MU2e2fMaMX2a2bMeM\\2\\2^MkM_2W2\\MoMb2R2YMTNe2m1UMZNi2g1RM^Nm2c1mLdNQ3]1jLiNT3X1gLnNW3T1bLSO\\3^5N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2M6K7I7I6J7I7I7IXn\\2"}, "label": "the woman wearing glasses and a gold crown"}]}
{"id": 209178, "image": "COCO_train2014_000000209178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady with the crown\"."}], "task": "refering", "answer_template": "The \"the lady with the crown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 267, 268, 269, 281, 282, 283, 304, 305, 306], "bbox": [0.190546875, 0.06193853427895981, 0.7085781250000001, 0.9087943262411348], "iscrowd": 0, "area": 61324.147099999995, "rle": {"size": [423, 640], "counts": "`fb11S=5L3M3M3L4M4L3M3M3L5L3kNmNRFV1l9nNoEU1o9oNkEV1R:nNiEU1U:oNfET1W:QOdER1Z:P1M2N2N2M3N3M2N2N4K6K4L4L5K4K5L1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O2N1N2O1O1O1O3L3N2NiN`IjK^6W4fIfKX6Z4mIcKP6_4RJ`Km5`4TJ`Kk5_4WJaKg5`4ZJ[Kj5d4XJWKl5i4UJRKn5o4TJkJP6T5RJfJS6Z5l001000O0100O01000O010O10O10O10O01000O01000O10O10O10O0100O10O0100O10O01O100O100O101N100O2O0O100O2O0O101N100O1O2L3L4L5K4M3O1O2N1O1O1O1O1RO[IYKe6d4aIYK`6c4gIYKZ6c4lIYKV6c4PJZKQ6b4VJZKk5b4[J[Ke5a4c1L4L4L4L3M4O15K6J4L2M2O2N2N1OO1O0O2N2N2N2N2N1O2N2O1N2N2N1O2O11O000O2O004L5K6J2M3M1O2M2O1N2O1N2O2M2O1N2N2M4M2N200O1O101N10bJgHn4i70000001N10N2K5J6K5N23M2M4M3M2N3M3M001O0O101M2M3N3M2N3O0001OkK[Jk0b5TO`Jn0]5POfJR1W5lNkJV1S5iNoJX1o4fNTKZ1k4dNXK]1g4]N_Kc1`4XNfKi1X4RNnKn1Q4lMVLU2g3fM`LZ2_3fMbL[2[3eMgLZ2W3gMjLX2R3mMoLR2n2PNTMo1m2PNTMo1m2oMUMP2l2nMVMQ2k2nMVMQ2k2mMWMR2j2mMWMR2j2iM\\MU2e2fMaMX2a2bMeM\\2\\2^MkM_2W2\\MoMb2R2YMTNe2m1UMZNi2g1RM^Nm2c1mLdNQ3]1jLiNT3X1gLnNW3T1bLSO\\3^5N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2M6K7I7I6J7I7I7IXn\\2"}, "label": "the lady with the crown"}]}
{"id": 520479, "image": "COCO_train2014_000000520479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with formal place settings at a party\"."}], "task": "refering", "answer_template": "The \"a table with formal place settings at a party\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406], "bbox": [0.046, 0.3706060606060606, 0.744484375, 1.0], "iscrowd": 0, "area": 95004.5173, "rle": {"size": [495, 640], "counts": "Q`>1Y?5J6K5J6J6K5J6K5J6K5J6J6L4O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O11OO1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000[MVJeKj5Z4ZJcKf5\\4_J`Ka5`4bJ]K^5b4fJ[KZ5d4jJYKW5j3UJlJh0W1S5l3VJlJl0T1n4o3XJkJn0S1j4Q4YJkJQ1Q1f4S4[JkJR1o0c4T4]JkJU1n0^4V4_JiJX1n0Y4X4`JiJ\\1k0T4[4bJgJ_1k0P4\\4bJgJc1j0k3^4dJfJd1j0h3^4fJfJe1k0e3^4hJdJe1n0c3]4WMbKi2]4YMbKg2]4ZMcKf2\\4\\McKd2\\4^McKc2Z4_MfKa2Y4aMfK_2Y4cMfK]2Y4dMgK\\2X4fMgKZ2X4hMgKX2W4jMiKV2V4lMiKT2V4nMiKS2U4nMkKR2T4PNkKP2T4RNkKn1T4SNjKo1T4g4O100O1O1O2N1O1O1O1O1O1O102N3M2N2N1O001O001N1iL^Da2c;]MbD_2^;aMfD[2[;cMkDX2U;hMoDT2Q;kMQES2P;mMPEn1V;QNkDi1Z;VNgDe1^;[NcDa1`;^NaD_1c;`N^DX1i;hNWDj0W<VOiCc0^<]OcC=c<B]C:g<GYC3l<MTCOQ=0PCJU=6kBFZ=:m001O0010O01O0100O05LmQ_2"}, "label": "a table with formal place settings at a party"}]}
{"id": 520479, "image": "COCO_train2014_000000520479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table the children are at\"."}], "task": "refering", "answer_template": "The \"the table the children are at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406], "bbox": [0.046, 0.3706060606060606, 0.744484375, 1.0], "iscrowd": 0, "area": 95004.5173, "rle": {"size": [495, 640], "counts": "Q`>1Y?5J6K5J6J6K5J6K5J6K5J6J6L4O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O11OO1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000[MVJeKj5Z4ZJcKf5\\4_J`Ka5`4bJ]K^5b4fJ[KZ5d4jJYKW5j3UJlJh0W1S5l3VJlJl0T1n4o3XJkJn0S1j4Q4YJkJQ1Q1f4S4[JkJR1o0c4T4]JkJU1n0^4V4_JiJX1n0Y4X4`JiJ\\1k0T4[4bJgJ_1k0P4\\4bJgJc1j0k3^4dJfJd1j0h3^4fJfJe1k0e3^4hJdJe1n0c3]4WMbKi2]4YMbKg2]4ZMcKf2\\4\\McKd2\\4^McKc2Z4_MfKa2Y4aMfK_2Y4cMfK]2Y4dMgK\\2X4fMgKZ2X4hMgKX2W4jMiKV2V4lMiKT2V4nMiKS2U4nMkKR2T4PNkKP2T4RNkKn1T4SNjKo1T4g4O100O1O1O2N1O1O1O1O1O1O102N3M2N2N1O001O001N1iL^Da2c;]MbD_2^;aMfD[2[;cMkDX2U;hMoDT2Q;kMQES2P;mMPEn1V;QNkDi1Z;VNgDe1^;[NcDa1`;^NaD_1c;`N^DX1i;hNWDj0W<VOiCc0^<]OcC=c<B]C:g<GYC3l<MTCOQ=0PCJU=6kBFZ=:m001O0010O01O0100O05LmQ_2"}, "label": "the table the children are at"}]}
{"id": 209185, "image": "COCO_train2014_000000209185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couple on a bench\"."}], "task": "refering", "answer_template": "The \"a couple on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.44737499999999997, 0.2217948717948718, 0.69290625, 0.9835897435897437], "iscrowd": 0, "area": 26968.7686, "rle": {"size": [429, 640], "counts": "i`h33Y=3M3M3lMHnF<l82hF1R99fFKU98jFb0Y8BfGj0l7YOSHo0c7TO\\Ho0^7VO`Hn0Z7UOeHo0T7UOfHS1V7oNbH\\1Z7eNXHk1e7R2L4L3M6J6I6K6J6J6J6J5K5K4L1O100O1O1O1O1O1O100O1O1O100O1O100O2N1fNQHVMo7f2XHVMi7f2]HWMd7f2bHWM^7e2jHVMW7g2oHUMZ4_OWM[3dNRMQ4NTMP3mNoLm3<nLe2VOmLi3j0iLX2@lLf3U1cLo1HjLe3]1^Lh1OiLd3d1XLc15gLc3l1SL]1b5hNYJW1i5nNQJR1Q6QOjIo0X6TOWIW1l6d22M4M2N2M2O2[LgHR1[7gNjHZ1W7_NnHa1T7XNQIh1Q7PNUIP2m6eM\\I[2k80001O01O0000000000001O0000001O1XIlLU3U3hLVMo2k2nL]Mm2c2PMbMm2_2oLhMm2Y2PMmMl2T2QMQNl2P2nLYNn2h1kLbNQ3_1iLiNT3X1hLmNU3U1fLQOP3X1lLmNi2]1SMhNc2a1YMdN\\2j1TMcNb2Z6G9L4K3N2N2M3N2RN\\EQ1f:oN`Eh0c:XOdE?^:AhE7Z:HnENU:2QFFQ::VF]Ol9b0`1000000O10000O1000000O1000000O10001N101O001N2O1O0O2O1O1O1N2O1O0OWXb2"}, "label": "a couple on a bench"}]}
{"id": 209185, "image": "COCO_train2014_000000209185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in the bench who is wearing a jeans\"."}], "task": "refering", "answer_template": "The \"a man sitting in the bench who is wearing a jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.44737499999999997, 0.2217948717948718, 0.69290625, 0.9835897435897437], "iscrowd": 0, "area": 26968.7686, "rle": {"size": [429, 640], "counts": "i`h33Y=3M3M3lMHnF<l82hF1R99fFKU98jFb0Y8BfGj0l7YOSHo0c7TO\\Ho0^7VO`Hn0Z7UOeHo0T7UOfHS1V7oNbH\\1Z7eNXHk1e7R2L4L3M6J6I6K6J6J6J6J5K5K4L1O100O1O1O1O1O1O100O1O1O100O1O100O2N1fNQHVMo7f2XHVMi7f2]HWMd7f2bHWM^7e2jHVMW7g2oHUMZ4_OWM[3dNRMQ4NTMP3mNoLm3<nLe2VOmLi3j0iLX2@lLf3U1cLo1HjLe3]1^Lh1OiLd3d1XLc15gLc3l1SL]1b5hNYJW1i5nNQJR1Q6QOjIo0X6TOWIW1l6d22M4M2N2M2O2[LgHR1[7gNjHZ1W7_NnHa1T7XNQIh1Q7PNUIP2m6eM\\I[2k80001O01O0000000000001O0000001O1XIlLU3U3hLVMo2k2nL]Mm2c2PMbMm2_2oLhMm2Y2PMmMl2T2QMQNl2P2nLYNn2h1kLbNQ3_1iLiNT3X1hLmNU3U1fLQOP3X1lLmNi2]1SMhNc2a1YMdN\\2j1TMcNb2Z6G9L4K3N2N2M3N2RN\\EQ1f:oN`Eh0c:XOdE?^:AhE7Z:HnENU:2QFFQ::VF]Ol9b0`1000000O10000O1000000O1000000O10001N101O001N2O1O0O2O1O1O1N2O1O0OWXb2"}, "label": "a man sitting in the bench who is wearing a jeans"}]}
{"id": 274724, "image": "COCO_train2014_000000274724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing chritmas dress\"."}], "task": "refering", "answer_template": "The \"a man wearing chritmas dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.12883333333333333, 0.09228125000000001, 0.8089791666666668, 0.9754375], "iscrowd": 0, "area": 135667.28315, "rle": {"size": [640, 480], "counts": "SYW1`0Rc0`0M3oJP1mDSO`:c1ZEaN_:i1ZEZN`:P2YESNb:U2XEoMa:[2XEhMc:a2VEbMd:h2UE\\Md:m2VEVMe:S3TEPMf:Z3SEiLh:_3REeLi:c3PE`Lk:i3nDZLn:`6K4K5L5K4K5L5J5L4L4K5L2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M2O2N2M3N2N2N2O1O100O1O100O1O100O1O100O1O100O1O10O01O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1N101O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1N2O1O1O001O1O1N2O1N2SO`AXLb>d3fAWL[>d3oAULS>g3UBSLm=h3X1L4K5L4K5N2N2N2N2N2N2N2N2N1O2O100O100O1O100O100O10000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O2O0000000000001O0000000000001O000000001O001O001N101O1O001UBTM`:l2[EZMY8InFn2c0^MY84jF^2i0bMW8`0^FW2U1^MX8j0mET2g1WMU8U1^EQ2Z2lLR8\\5mGeJn7a5PH`Jj7f5UH\\Jd7k5ZHVJc7n5[HSJe7P6WHQJh7S6THoIi7V6RHlIl7Z6jGlIT8Y6dGnIX8X6_GoI\\8Y6YGoIc8Y6RGnIj8g8L4K5L4L5L5J5L4K5L4L4K6K4K6K6J6J6J6J6I7J6J7I6J6J6J6J9G=B>B=D;D6K5J6J6K5J6K5J<D;FmSi1"}, "label": "a man wearing chritmas dress"}]}
{"id": 274724, "image": "COCO_train2014_000000274724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man dressed as santa\"."}], "task": "refering", "answer_template": "The \"a man dressed as santa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.12883333333333333, 0.09228125000000001, 0.8089791666666668, 0.9754375], "iscrowd": 0, "area": 135667.28315, "rle": {"size": [640, 480], "counts": "SYW1`0Rc0`0M3oJP1mDSO`:c1ZEaN_:i1ZEZN`:P2YESNb:U2XEoMa:[2XEhMc:a2VEbMd:h2UE\\Md:m2VEVMe:S3TEPMf:Z3SEiLh:_3REeLi:c3PE`Lk:i3nDZLn:`6K4K5L5K4K5L5J5L4L4K5L2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M2O2N2M3N2N2N2O1O100O1O100O1O100O1O100O1O100O1O10O01O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1N101O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1N2O1O1O001O1O1N2O1N2SO`AXLb>d3fAWL[>d3oAULS>g3UBSLm=h3X1L4K5L4K5N2N2N2N2N2N2N2N2N1O2O100O100O1O100O100O10000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O2O0000000000001O0000000000001O000000001O001O001N101O1O001UBTM`:l2[EZMY8InFn2c0^MY84jF^2i0bMW8`0^FW2U1^MX8j0mET2g1WMU8U1^EQ2Z2lLR8\\5mGeJn7a5PH`Jj7f5UH\\Jd7k5ZHVJc7n5[HSJe7P6WHQJh7S6THoIi7V6RHlIl7Z6jGlIT8Y6dGnIX8X6_GoI\\8Y6YGoIc8Y6RGnIj8g8L4K5L4L5L5J5L4K5L4L4K6K4K6K6J6J6J6J6I7J6J7I6J6J6J6J9G=B>B=D;D6K5J6J6K5J6K5J<D;FmSi1"}, "label": "a man dressed as santa"}]}
{"id": 520486, "image": "COCO_train2014_000000520486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the opened umbrella on the right\"."}], "task": "refering", "answer_template": "The \"the opened umbrella on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 63, 64, 65, 66, 67, 68, 84], "bbox": [0.54404, 0.19809815950920245, 0.7989599999999999, 0.4097239263803681], "iscrowd": 0, "area": 3241.7738500000014, "rle": {"size": [326, 500], "counts": "kef26o93N1N1O100O10O0100O1O100O100O1O10O010O02O0O2N2O0O100O100O1O100O100O10000O1000000O100000000O100000O01000000O10O1000000000OOaG@\\7?PH6P8IoG9Q8EPHIc89d03M3M2N00000001O0O10001O00001O00001O00001O00001O001O001O1O001O1O001O1O001O1O0010O01O1O001O1O010O1O010O010N2O001N101N101N2OPRP1"}, "label": "the opened umbrella on the right"}]}
{"id": 520486, "image": "COCO_train2014_000000520486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella in front of the man with assless chaps\"."}], "task": "refering", "answer_template": "The \"the umbrella in front of the man with assless chaps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 63, 64, 65, 66, 67, 68, 84], "bbox": [0.54404, 0.19809815950920245, 0.7989599999999999, 0.4097239263803681], "iscrowd": 0, "area": 3241.7738500000014, "rle": {"size": [326, 500], "counts": "kef26o93N1N1O100O10O0100O1O100O100O1O10O010O02O0O2N2O0O100O100O1O100O100O10000O1000000O100000000O100000O01000000O10O1000000000OOaG@\\7?PH6P8IoG9Q8EPHIc89d03M3M2N00000001O0O10001O00001O00001O00001O00001O001O001O1O001O1O001O1O001O1O0010O01O1O001O1O010O1O010O010N2O001N101N101N2OPRP1"}, "label": "the umbrella in front of the man with assless chaps"}]}
{"id": 520486, "image": "COCO_train2014_000000520486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella on rearmost empty table\"."}], "task": "refering", "answer_template": "The \"umbrella on rearmost empty table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79], "bbox": [0.0674, 0.1776687116564417, 0.43662, 0.3866564417177914], "iscrowd": 0, "area": 8187.89425, "rle": {"size": [326, 500], "counts": "`W;2T:0O10001N100O8I7H2O0O10001N1O1O101N1O1O1O1O2N100O1O2N1O1O101N1O1O101N10000O101O0O10001N10000O2O000O101O0O10000O2O000O101O0O10001N10000O2O000O01000O0100O100O10000O100O2O000O100O100O10000001O000001O0000000001O01O0000000000010O00000000001O001O001N10001O001O0O2O00001O001N101O00001O0O2O001O00001N101O001O001N10001O001O0O2O00001O001N101O00001O0O2O001O1O1O2M2M3N2M3M4M2Mlei2"}, "label": "umbrella on rearmost empty table"}]}
{"id": 151848, "image": "COCO_train2014_000000151848.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white checkered table with two wooden chairs\"."}], "task": "refering", "answer_template": "The \"a red and white checkered table with two wooden chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212], "bbox": [0.31974, 0.805855855855856, 0.8079400000000001, 0.9885285285285286], "iscrowd": 0, "area": 8711.122099999999, "rle": {"size": [333, 500], "counts": "W[d11Z:2N2N2O1N2N2N2N2N2O1N2O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O100O1O100O100O100O1O100O100O1O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000OO2N2N2N20O10000]G^N[8j100O1001O000^Nb10000000000000000000001O000001O0000000001O00000000001O0006Je0[Oe0[O2OO00000000001O000000001O000000001O0000001O000000001O0000001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1OVWo0"}, "label": "a red and white checkered table with two wooden chairs"}]}
{"id": 151848, "image": "COCO_train2014_000000151848.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"attractive , country style , red - check round tablecloth\"."}], "task": "refering", "answer_template": "The \"attractive , country style , red - check round tablecloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212], "bbox": [0.31974, 0.805855855855856, 0.8079400000000001, 0.9885285285285286], "iscrowd": 0, "area": 8711.122099999999, "rle": {"size": [333, 500], "counts": "W[d11Z:2N2N2O1N2N2N2N2N2O1N2O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O100O1O100O100O100O1O100O100O1O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000OO2N2N2N20O10000]G^N[8j100O1001O000^Nb10000000000000000000001O000001O0000000001O00000000001O0006Je0[Oe0[O2OO00000000001O000000001O000000001O0000001O000000001O0000001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1OVWo0"}, "label": "attractive , country style , red - check round tablecloth"}]}
{"id": 28974, "image": "COCO_train2014_000000028974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing behind one giraffe , and in front of another\"."}], "task": "refering", "answer_template": "The \"a giraffe standing behind one giraffe , and in front of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 142, 143, 144, 145, 146, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346, 347], "bbox": [0.0016093750000000001, 0.289890590809628, 0.362640625, 0.9887746170678338], "iscrowd": 0, "area": 23294.456199999997, "rle": {"size": [457, 640], "counts": "mf0m1[<n1RN101N1O1O100O1O1O100O1O100O1L4L4M3Lc0]Oe0\\Od0^O3N0O100O100O100O100O10000O101N100O100O10000O100O100O100O10000O100O101N100O100O10L4L3M4L3M4L3M4L3M4L4L3M4L3L5L3M4L3XMSG2Q9M]GBg8=gGRO\\8n0RHaNR8^1\\HQNg7o1Z2O001O001O001O0O10001O00001O0000001O0000001N1000001O0000001O00001O00000O2O0000001O0000001O000000000O1000000000000000000000001O0O100000000000000000000000000O1000001O000000O100000000000O1000001O1O00100001O001O00001O00001O001O00001OcNbCo0]<kNkCT1g<O0O101N1O101N1O101N1O101N100O2N100O2N100O2O0O1N3M2M3N3M2M3Nmle5"}, "label": "a giraffe standing behind one giraffe , and in front of another"}]}
{"id": 28974, "image": "COCO_train2014_000000028974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large spotted giraffe walking out front of two other giraffes\"."}], "task": "refering", "answer_template": "The \"large spotted giraffe walking out front of two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 312, 313, 314, 328, 329, 330, 331, 332, 334, 335, 336, 337, 351, 352, 353, 354, 357, 358, 360], "bbox": [0.2974375, 0.11638949671772428, 0.99765625, 1.0], "iscrowd": 0, "area": 54447.44105, "rle": {"size": [457, 640], "counts": "]Ve21V>;E;E<C<E6J2N2N2N2M3N2N2N2N2N2M3N2L4K5L4L3M2N2N2N2N2M3N2N2N2N3M2N3M2N3O1O0100O10000O10O0100O10000O10VF_Ll8a3RGQM_8o2_GXM\\8h2aG\\M^8c2aG`M^8`2_GdM`8[2_GhM`8X2]GlMb8S2]GPNa8m3O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O10000O10N2M3M3L4M3M3M3M2N3M3M3M3M3M3M3L3N3M3M3M3M3M3M3M2N3M3L4M3M3M31O0O2O1O001O1O001O1O1N101O1O001OO1O10O01O1O1O100O1000O10000O1O1O1O1O1N2O1O001O1O1O1O1N2O1O1O1O1O1O0O2gHeKc4\\4XKiKg4X4UKlKj4U4QKQLm4P4nJULQ5l3jJYLU5h3fJ]LY5d3cJaLZ5a3aJdL^5]3]JhLb5Y3YJlLf5T3VJQMi5P3RJVMl5k2PJYMo5h2lI]MS6d2hIaMW6`2dIfMZ6[2aIjM^6Q5O101N1O1ON2N2N3N1N2N2N2O2M2N3M3N2M2N3M3N1N3M3M3N1N3M3M2O2M3M2N3N24L3L4M4L3L5L3M3L5L3M4L3L4M4L3L4M4L3M4K4M3M4K3N000O1000000O10000O10_L`JYN`5c1lJUNT5M^JeNl0U1e46eJ\\NS1U1X4>jJTN[1V1k3e0PKlMb1V1]3n0VKcMj1W1P3U1\\K[MQ2W1c2]1aKSMZ2W1T2f1YORNg0m1CjM=U2LcM4\\26[MIe2m500000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0O10000O100O10000O101O0O100O10000O100O101O0O100O10000O100O10001N10000O100O10000O100O10000O100O10000O10000O100O101O0O100O10000O100O10000001O1O1O1O1O1G90000O100O10001N100O12N2N3M2N2N3M2N2N3M2N3M2N1O1O1O1O010O1O1O001O1O1O010O1O1O100O1O1O101N1O1O100O1O1O100O1O2O0O1O1O100O1O1O100O2N1O100O1N2M3N2M3NW:"}, "label": "large spotted giraffe walking out front of two other giraffes"}]}
{"id": 28974, "image": "COCO_train2014_000000028974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe walking through the grass in front of two other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe walking through the grass in front of two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 312, 313, 314, 328, 329, 330, 331, 332, 334, 335, 336, 337, 351, 352, 353, 354, 357, 358, 360], "bbox": [0.2974375, 0.11638949671772428, 0.99765625, 1.0], "iscrowd": 0, "area": 54447.44105, "rle": {"size": [457, 640], "counts": "]Ve21V>;E;E<C<E6J2N2N2N2M3N2N2N2N2N2M3N2L4K5L4L3M2N2N2N2N2M3N2N2N2N3M2N3M2N3O1O0100O10000O10O0100O10000O10VF_Ll8a3RGQM_8o2_GXM\\8h2aG\\M^8c2aG`M^8`2_GdM`8[2_GhM`8X2]GlMb8S2]GPNa8m3O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O10000O10N2M3M3L4M3M3M3M2N3M3M3M3M3M3M3L3N3M3M3M3M3M3M3M2N3M3L4M3M3M31O0O2O1O001O1O001O1O1N101O1O001OO1O10O01O1O1O100O1000O10000O1O1O1O1O1N2O1O001O1O1O1O1N2O1O1O1O1O1O0O2gHeKc4\\4XKiKg4X4UKlKj4U4QKQLm4P4nJULQ5l3jJYLU5h3fJ]LY5d3cJaLZ5a3aJdL^5]3]JhLb5Y3YJlLf5T3VJQMi5P3RJVMl5k2PJYMo5h2lI]MS6d2hIaMW6`2dIfMZ6[2aIjM^6Q5O101N1O1ON2N2N3N1N2N2N2O2M2N3M3N2M2N3M3N1N3M3M3N1N3M3M2O2M3M2N3N24L3L4M4L3L5L3M3L5L3M4L3L4M4L3L4M4L3M4K4M3M4K3N000O1000000O10000O10_L`JYN`5c1lJUNT5M^JeNl0U1e46eJ\\NS1U1X4>jJTN[1V1k3e0PKlMb1V1]3n0VKcMj1W1P3U1\\K[MQ2W1c2]1aKSMZ2W1T2f1YORNg0m1CjM=U2LcM4\\26[MIe2m500000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0O10000O100O10000O101O0O100O10000O100O101O0O100O10000O100O10001N10000O100O10000O100O10000O100O10000O10000O100O101O0O100O10000O100O10000001O1O1O1O1O1G90000O100O10001N100O12N2N3M2N2N3M2N2N3M2N3M2N1O1O1O1O010O1O1O001O1O1O010O1O1O100O1O1O101N1O1O100O1O1O100O1O2O0O1O1O100O1O1O100O2N1O100O1N2M3N2M3NW:"}, "label": "a giraffe walking through the grass in front of two other giraffes"}]}
{"id": 28974, "image": "COCO_train2014_000000028974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the youngest giraffe out of three in the middle\"."}], "task": "refering", "answer_template": "The \"the youngest giraffe out of three in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258, 279, 280, 281, 303, 304, 326], "bbox": [0.13479687499999998, 0.40954048140043764, 0.4380625, 0.9421444201312911], "iscrowd": 0, "area": 12397.452649999996, "rle": {"size": [457, 640], "counts": "leV13U>;D=D;E;D<E:F1N101O1O10O01O1O00100O001O1O10O01O4L6J6I7I7J6I7I7J6I7K5K4L1O2N1O1OO010O1O1O1O100O1O1O001H7H8H8H8H9M2001O10O01O1O0010M2K5K5K6J5L4K5K5K6J5K5M30001O0O10001O000000001O000000001O000O101O000000001O000000001O000O101O000000001O0000001O00000O101O000000001O0000001O00000O010000000000O100000002O000O2O00000O2O02N101N1O100O1O100O001O3N2M4L3N0O1O100O1O100O001O100O1O1O1N101N2N2N2N2O1N2NQ^P5"}, "label": "the youngest giraffe out of three in the middle"}]}
{"id": 28974, "image": "COCO_train2014_000000028974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle giraffe in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the middle giraffe in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258, 279, 280, 281, 303, 304, 326], "bbox": [0.13479687499999998, 0.40954048140043764, 0.4380625, 0.9421444201312911], "iscrowd": 0, "area": 12397.452649999996, "rle": {"size": [457, 640], "counts": "leV13U>;D=D;E;D<E:F1N101O1O10O01O1O00100O001O1O10O01O4L6J6I7I7J6I7I7J6I7K5K4L1O2N1O1OO010O1O1O1O100O1O1O001H7H8H8H8H9M2001O10O01O1O0010M2K5K5K6J5L4K5K5K6J5K5M30001O0O10001O000000001O000000001O000O101O000000001O000000001O000O101O000000001O0000001O00000O101O000000001O0000001O00000O010000000000O100000002O000O2O00000O2O02N101N1O100O1O100O001O3N2M4L3N0O1O100O1O100O001O100O1O1O1N101N2N2N2N2O1N2NQ^P5"}, "label": "the middle giraffe in the right hand picture"}]}
{"id": 127284, "image": "COCO_train2014_000000127284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of head of someone on the edge of the frame\"."}], "task": "refering", "answer_template": "The \"back of head of someone on the edge of the frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 207, 230, 253, 276, 277, 299, 300, 301, 322, 323, 324], "bbox": [0.0015, 0.26259953161592503, 0.11668749999999999, 0.991288056206089], "iscrowd": 0, "area": 10057.112149999999, "rle": {"size": [427, 640], "counts": "ca0]5g0]MT3W3kLhLP3\\3QMdLo2\\3RMcLo2[3SMdLn2W3WMhLj2R3[MnLf2k2aMTM`2i2cMVM]2i2eMWM[2f2hMYMY2d2iM\\MX2a2kM^MV2`2lM_MT2_2oM`MR2]2QNbMP2\\2QNdMP2Y2SNfMn1W2UNhMk1V2XNiMi1U2YNjMh1S2ZNmMg1P2\\NoMe1n1^NQNb1n1`NQNa1l1bNSN_1j1cNVN^1g1eNXN\\1f1fNYNZ1e1iNZNX1c1jN]NW1a1kN^NV1_1mN`NW1Z1lNeNY1T1kNjNY1Q1iNnN\\1k0gNTO^1e0fNYO^1a0eN^O_1=cNB`18eNF^13gNL\\1MjN1Y1HlN7W1BnN=X1YOnNe0Z1oNkNP1]1dNhN[1f6N1O1O2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N2N2N2N2N2N2N]c[7"}, "label": "back of head of someone on the edge of the frame"}]}
{"id": 127284, "image": "COCO_train2014_000000127284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a back shot of a males red hair at the beach\"."}], "task": "refering", "answer_template": "The \"a back shot of a males red hair at the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 207, 230, 253, 276, 277, 299, 300, 301, 322, 323, 324], "bbox": [0.0015, 0.26259953161592503, 0.11668749999999999, 0.991288056206089], "iscrowd": 0, "area": 10057.112149999999, "rle": {"size": [427, 640], "counts": "ca0]5g0]MT3W3kLhLP3\\3QMdLo2\\3RMcLo2[3SMdLn2W3WMhLj2R3[MnLf2k2aMTM`2i2cMVM]2i2eMWM[2f2hMYMY2d2iM\\MX2a2kM^MV2`2lM_MT2_2oM`MR2]2QNbMP2\\2QNdMP2Y2SNfMn1W2UNhMk1V2XNiMi1U2YNjMh1S2ZNmMg1P2\\NoMe1n1^NQNb1n1`NQNa1l1bNSN_1j1cNVN^1g1eNXN\\1f1fNYNZ1e1iNZNX1c1jN]NW1a1kN^NV1_1mN`NW1Z1lNeNY1T1kNjNY1Q1iNnN\\1k0gNTO^1e0fNYO^1a0eN^O_1=cNB`18eNF^13gNL\\1MjN1Y1HlN7W1BnN=X1YOnNe0Z1oNkNP1]1dNhN[1f6N1O1O2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N2N2N2N2N2N2N]c[7"}, "label": "a back shot of a males red hair at the beach"}]}
{"id": 471352, "image": "COCO_train2014_000000471352.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"animal in middle\"."}], "task": "refering", "answer_template": "The \"animal in middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 261, 262, 263, 284, 285, 286, 309], "bbox": [0.3556875, 0.39135514018691586, 0.5205624999999999, 0.8817757009345794], "iscrowd": 0, "area": 12632.09415, "rle": {"size": [428, 640], "counts": "g`o27R=7J6I7RFVOo6Q1lHQOP7V1kHkNR7\\1aHmN\\7Z1fG@W8g0jF4S9T2M4M20000000001gGfK^7Z4WHRLh7n3mG]LS8^40000001O01O00000001O000_OlGkKS8U4ZH^Kf7c4a000010O001O001OWOjG\\LW8`3l0001O4L:F:G9bNmKoH]4i6cKUI`4j6`KTI`4n6aKoH`4R7l02N2N2N2N3M2N2N2N2N2N2N2N2NOF;ZOe0ROo0K4O2N0000001O0000000000001O0bMlEf1T:TNWFg1i9TNbFf1^9TNnFf1R9UNXGf1Y:M3M3M2N3M3M3M2N3M3M2N3M4L4L4K6JlRP4"}, "label": "animal in middle"}]}
{"id": 471352, "image": "COCO_train2014_000000471352.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"and ok being ridden in the field with a bowl on top of his back\"."}], "task": "refering", "answer_template": "The \"and ok being ridden in the field with a bowl on top of his back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 234, 235, 257, 258, 280, 281], "bbox": [0.162890625, 0.36352803738317757, 0.297921875, 0.8338317757009346], "iscrowd": 0, "area": 6755.590749999999, "rle": {"size": [428, 640], "counts": "Tf[12k<O`C6^<K^C;^<G^C>_<:M4K6K6I4L3O2N2M2O4L5K2N002N3M3M3M2O2M3M3N2M3M8G8Ho0QO5K6J5SObKaHb4V7n0L4L4005K6J8H8H4nNTH\\Lm7a3ZHYLg7d3aHSLc7j3R1L2N1M4L4M3L3L101O2O3bMPF`1P:^NUF_1k9`NXF^1h9aN\\F\\1f9ZNeFd1c:M3M3M3M3M3K5L3L5K4N3J5K6K4O2N1O1O20O0100O010O10OOokk5"}, "label": "and ok being ridden in the field with a bowl on top of his back"}]}
{"id": 471352, "image": "COCO_train2014_000000471352.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small white animal on the left side\"."}], "task": "refering", "answer_template": "The \"the small white animal on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 234, 235, 257, 258, 280, 281], "bbox": [0.162890625, 0.36352803738317757, 0.297921875, 0.8338317757009346], "iscrowd": 0, "area": 6755.590749999999, "rle": {"size": [428, 640], "counts": "Tf[12k<O`C6^<K^C;^<G^C>_<:M4K6K6I4L3O2N2M2O4L5K2N002N3M3M3M2O2M3M3N2M3M8G8Ho0QO5K6J5SObKaHb4V7n0L4L4005K6J8H8H4nNTH\\Lm7a3ZHYLg7d3aHSLc7j3R1L2N1M4L4M3L3L101O2O3bMPF`1P:^NUF_1k9`NXF^1h9aN\\F\\1f9ZNeFd1c:M3M3M3M3M3K5L3L5K4N3J5K6K4O2N1O1O20O0100O010O10OOokk5"}, "label": "the small white animal on the left side"}]}
{"id": 356665, "image": "COCO_train2014_000000356665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cot with one white pillow\"."}], "task": "refering", "answer_template": "The \"the cot with one white pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.533625, 0.413671875, 1.0], "iscrowd": 0, "area": 31408.523300000004, "rle": {"size": [480, 640], "counts": "^8d2\\<9F:G8H9F9H8H3L3N3M3L4M3M3L4M2N3L2O1O1N3N1N2O1O2M2O1O1N2O2N1N2O1O1N3N1O1N2O1O2M2001O1O1O1N3M2O1N2O1N2N3N1N2O1N2N2O1N101N2N101N100O1O100O1O001O1O1O1O1O1O1O001O1O100O1O1O1O001O1O1O1O1O001O001O1O0010O01O1O001O001O1O001O001O1O001O010O1O001O001O1O001O001O1O001O00100O001O001O1O001O001O1O001O1O0010O01O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001N101O1N101N101N100O2O0O2O0O2O0O2O0O2O1N2O001N2O0O2O1N2O0O2O1N101N2O0O3N1N2O1N2O2M2O100N3N1O1O2N1O1O2N1N2O1O1O1O1O1O2N0O2O1O1O1O1O1O1O1N2O1O2N1O2N1O2N1O0O2Nli_5"}, "label": "the cot with one white pillow"}]}
{"id": 356665, "image": "COCO_train2014_000000356665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long , twin size bed\"."}], "task": "refering", "answer_template": "The \"a long , twin size bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.533625, 0.413671875, 1.0], "iscrowd": 0, "area": 31408.523300000004, "rle": {"size": [480, 640], "counts": "^8d2\\<9F:G8H9F9H8H3L3N3M3L4M3M3L4M2N3L2O1O1N3N1N2O1O2M2O1O1N2O2N1N2O1O1N3N1O1N2O1O2M2001O1O1O1N3M2O1N2O1N2N3N1N2O1N2N2O1N101N2N101N100O1O100O1O001O1O1O1O1O1O1O001O1O100O1O1O1O001O1O1O1O1O001O001O1O0010O01O1O001O001O1O001O001O1O001O010O1O001O001O1O001O001O1O001O00100O001O001O1O001O001O1O001O1O0010O01O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001N101O1N101N101N100O2O0O2O0O2O0O2O0O2O1N2O001N2O0O2O1N2O0O2O1N101N2O0O3N1N2O1N2O2M2O100N3N1O1O2N1O1O2N1N2O1O1O1O1O1O2N0O2O1O1O1O1O1O1O1N2O1O2N1O2N1O2N1O0O2Nli_5"}, "label": "a long , twin size bed"}]}
{"id": 356665, "image": "COCO_train2014_000000356665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"queen bed with white linens\"."}], "task": "refering", "answer_template": "The \"queen bed with white linens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 307, 308, 309, 331, 332], "bbox": [0.1526875, 0.4856666666666667, 0.5503281250000001, 0.8480833333333334], "iscrowd": 0, "area": 23400.222449999997, "rle": {"size": [480, 640], "counts": "[W^12n>1N3M2O1N3M2O2gNCjC?U<ChC`0V<BgC?Y<CeC>Z<DcC>[<DcC=]<EaC<^<F_C<`<F]C<a<G]C;a<G\\C<b<E\\C=c<E[C=b<F[C=c<R1N1O2M101O000O2O000O2O0O101O0O101N10001N101N10001N100O2O000O2O000100O00100O00O2O000O2O0O101O0O101N10001N10001N100O2O001N100O2O000O2O000O2O0O1101N2N2O1N2N2O1N2N2O1N2N10O01O010O0001O000010O0001O001O00001O00010O00001O000010O0001O000010O0001O00010O00001O0000100O1O1O2N1O2N1O2N1O2O0O1O2N1O2N1O2N1O1O2O0O2NO100O100O1O101N1O100O100O1O100O101N1O100O100XOoD`MR;_2SE\\Mn:b2WE[Mi:d2\\EWMe:h2`ESMb:k2j0N2O1O1O1N2O1O1O1N3N1O1O1N2O1O1O1N2O1O2N1N2O1O1O1N2O1O1O1N3N1O1O1N3N1O2N1O2N101O0O2O002M2O2N1O2N2N1O2N1L5H7IofV4"}, "label": "queen bed with white linens"}]}
{"id": 356665, "image": "COCO_train2014_000000356665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the farther bed\"."}], "task": "refering", "answer_template": "The \"the farther bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 307, 308, 309, 331, 332], "bbox": [0.1526875, 0.4856666666666667, 0.5503281250000001, 0.8480833333333334], "iscrowd": 0, "area": 23400.222449999997, "rle": {"size": [480, 640], "counts": "[W^12n>1N3M2O1N3M2O2gNCjC?U<ChC`0V<BgC?Y<CeC>Z<DcC>[<DcC=]<EaC<^<F_C<`<F]C<a<G]C;a<G\\C<b<E\\C=c<E[C=b<F[C=c<R1N1O2M101O000O2O000O2O0O101O0O101N10001N101N10001N100O2O000O2O000100O00100O00O2O000O2O0O101O0O101N10001N10001N100O2O001N100O2O000O2O000O2O0O1101N2N2O1N2N2O1N2N2O1N2N10O01O010O0001O000010O0001O001O00001O00010O00001O000010O0001O000010O0001O00010O00001O0000100O1O1O2N1O2N1O2N1O2O0O1O2N1O2N1O2N1O1O2O0O2NO100O100O1O101N1O100O100O1O100O101N1O100O100XOoD`MR;_2SE\\Mn:b2WE[Mi:d2\\EWMe:h2`ESMb:k2j0N2O1O1O1N2O1O1O1N3N1O1O1N2O1O1O1N2O1O2N1N2O1O1O1N2O1O1O1N3N1O1O1N3N1O2N1O2N101O0O2O002M2O2N1O2N2N1O2N1L5H7IofV4"}, "label": "the farther bed"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older girl\"."}], "task": "refering", "answer_template": "The \"older girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 255, 256, 257, 278, 279, 280], "bbox": [0.033187499999999995, 0.21762499999999999, 0.21959374999999998, 0.7625416666666667], "iscrowd": 0, "area": 17317.4004, "rle": {"size": [480, 640], "counts": "YQ:6i=MRC;g<IUC;h<GTC=i<GRC>l<DPC?n<GjB=T=i0N1O1O1000O2N2O1N10WC\\NNOi;c1XDdNHNn;]1XDZOc;g0ZD@a;`0^DF^;:`DL[;5bD0\\;0bD5Z;k1M3N3M2M3N1O2M3N1N3N2N2jNbKfGa4T8eKhG^4T8hKhG[4T8jKhGY4S8QLeGR4W8QLfGS4V8oKhGU4U8mKiGV4S8lKkGY4Q8_1L3YHSIT7]700O2O000O2O100O1O2N2O1N2O001N101OO001N2O1O1O2PIkHP6W7nIjHR6X7iIkHW6W7dImH[6T7bInH]6V7_IlH`6X7[IiHe6j7[IjG^6\\8O2O001O001N101O1O0O2O1O1O1SL[G\\1f8`NdGZ1]8cNiGY1X8fNTHn0m7QO[Hf0g7YOcH=_7@kHM`72oH^OT7?WIdN]M0_9[1kI^NW6`1g3M2N2N3L3N2M3N3L3N2N3M2N2N2N3M3L4KmUZ7"}, "label": "older girl"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman in a t - shirt and jeans looking confused and holding a tin\"."}], "task": "refering", "answer_template": "The \"a young woman in a t - shirt and jeans looking confused and holding a tin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 255, 256, 257, 278, 279, 280], "bbox": [0.033187499999999995, 0.21762499999999999, 0.21959374999999998, 0.7625416666666667], "iscrowd": 0, "area": 17317.4004, "rle": {"size": [480, 640], "counts": "YQ:6i=MRC;g<IUC;h<GTC=i<GRC>l<DPC?n<GjB=T=i0N1O1O1000O2N2O1N10WC\\NNOi;c1XDdNHNn;]1XDZOc;g0ZD@a;`0^DF^;:`DL[;5bD0\\;0bD5Z;k1M3N3M2M3N1O2M3N1N3N2N2jNbKfGa4T8eKhG^4T8hKhG[4T8jKhGY4S8QLeGR4W8QLfGS4V8oKhGU4U8mKiGV4S8lKkGY4Q8_1L3YHSIT7]700O2O000O2O100O1O2N2O1N2O001N101OO001N2O1O1O2PIkHP6W7nIjHR6X7iIkHW6W7dImH[6T7bInH]6V7_IlH`6X7[IiHe6j7[IjG^6\\8O2O001O001N101O1O0O2O1O1O1SL[G\\1f8`NdGZ1]8cNiGY1X8fNTHn0m7QO[Hf0g7YOcH=_7@kHM`72oH^OT7?WIdN]M0_9[1kI^NW6`1g3M2N2N3L3N2M3N3L3N2N3M2N2N2N3M3L4KmUZ7"}, "label": "a young woman in a t - shirt and jeans looking confused and holding a tin"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large table with picnic food sitting on top of it\"."}], "task": "refering", "answer_template": "The \"large table with picnic food sitting on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.0033750000000000004, 0.6674166666666667, 0.8545, 0.9865208333333333], "iscrowd": 0, "area": 51538.67319999998, "rle": {"size": [480, 640], "counts": "Q\\1i0\\<k10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O0100000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O11O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00YR\\1"}, "label": "large table with picnic food sitting on top of it"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that is holding the baskets . the little girl has her back towards it and the woman is facing it\"."}], "task": "refering", "answer_template": "The \"the table that is holding the baskets . the little girl has her back towards it and the woman is facing it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.0033750000000000004, 0.6674166666666667, 0.8545, 0.9865208333333333], "iscrowd": 0, "area": 51538.67319999998, "rle": {"size": [480, 640], "counts": "Q\\1i0\\<k10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O0100000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O11O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00YR\\1"}, "label": "the table that is holding the baskets . the little girl has her back towards it and the woman is facing it"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table near to the child\"."}], "task": "refering", "answer_template": "The \"a wooden table near to the child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.507296875, 0.4108541666666667, 0.9050625, 0.5771458333333332], "iscrowd": 0, "area": 13345.957250000001, "rle": {"size": [480, 640], "counts": "eah44l>4L5K4K5L5K4L4K5L001O000O1000001O0O1000000O2O0000000O2O0000000O101O000O1000001O0O100000001N1000000O2O0000000O101O00000O10001O0O1000001O0O100000001N100001O0001O00000001O01O000000010O000000000010O00000000010O00000000010O00000001O0001O000001O0001O000001O000001O01O00000001O01O00O100O101N100O100O101N1O100O100O2O0O100O100O101N1O100O100O2O0O100O100O101N1O100O100O2O0O11O2N3M2N2O1N3M2N2N3M2O1N1OO1000001N10000O1000001N10000O10001N1000000O10001N10000O10001O0O1O100O1O1O2O0O1O1O100O2N100O1O1O101N1O1O100O4L5K5L5J5K][l0"}, "label": "a wooden table near to the child"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kitchen work station made out of wood\"."}], "task": "refering", "answer_template": "The \"a kitchen work station made out of wood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.507296875, 0.4108541666666667, 0.9050625, 0.5771458333333332], "iscrowd": 0, "area": 13345.957250000001, "rle": {"size": [480, 640], "counts": "eah44l>4L5K4K5L5K4L4K5L001O000O1000001O0O1000000O2O0000000O2O0000000O101O000O1000001O0O100000001N1000000O2O0000000O101O00000O10001O0O1000001O0O100000001N100001O0001O00000001O01O000000010O000000000010O00000000010O00000000010O00000001O0001O000001O0001O000001O000001O01O00000001O01O00O100O101N100O100O101N1O100O100O2O0O100O100O101N1O100O100O2O0O100O100O101N1O100O100O2O0O11O2N3M2N2O1N3M2N2N3M2O1N1OO1000001N10000O1000001N10000O10001N1000000O10001N10000O10001O0O1O100O1O1O2O0O1O1O100O2N100O1O1O101N1O1O100O4L5K5L5J5K][l0"}, "label": "a kitchen work station made out of wood"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girl\"."}], "task": "refering", "answer_template": "The \"the little girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317, 338, 339, 361, 362], "bbox": [0.69359375, 0.5141041666666667, 0.8265156250000001, 0.9378333333333333], "iscrowd": 0, "area": 10995.355200000004, "rle": {"size": [480, 640], "counts": "Z``65c><E;D<^Ob0\\Od0[Oe0\\Ob0I5L5L4K4L5L3L5K4L5L3L5K5M2O2O0O2N1O2O0O2N2N101N2N2N2O0O2O1O1O1O1O001O1O1O1ON2N2M3M3N2M3M4M3L3M4M2L5C<E:E<D<D<E;O02O001O1N\\MiDb1V;^NiDc1W;\\NiD6J>\\;\\OiD6O<W;]OiD847S;BhD793P;EgD7>0l:IfD6b0Mi:LfD6k0E^:6hD1o<NSCNP=2PCJT=5nBGU=9h00100OZic1"}, "label": "the little girl"}]}
{"id": 577850, "image": "COCO_train2014_000000577850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl with white t - shirt\"."}], "task": "refering", "answer_template": "The \"little girl with white t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317, 338, 339, 361, 362], "bbox": [0.69359375, 0.5141041666666667, 0.8265156250000001, 0.9378333333333333], "iscrowd": 0, "area": 10995.355200000004, "rle": {"size": [480, 640], "counts": "Z``65c><E;D<^Ob0\\Od0[Oe0\\Ob0I5L5L4K4L5L3L5K4L5L3L5K5M2O2O0O2N1O2O0O2N2N101N2N2N2O0O2O1O1O1O1O001O1O1O1ON2N2M3M3N2M3M4M3L3M4M2L5C<E:E<D<D<E;O02O001O1N\\MiDb1V;^NiDc1W;\\NiD6J>\\;\\OiD6O<W;]OiD847S;BhD793P;EgD7>0l:IfD6b0Mi:LfD6k0E^:6hD1o<NSCNP=2PCJT=5nBGU=9h00100OZic1"}, "label": "little girl with white t - shirt"}]}
{"id": 430394, "image": "COCO_train2014_000000430394.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blonde lady yellow dress\"."}], "task": "refering", "answer_template": "The \"blonde lady yellow dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 19, 20, 21, 22, 23, 24, 25, 34, 35, 36, 37, 38, 39, 40, 41, 50, 51, 52, 53, 54, 55, 56, 57, 66, 67, 68, 69, 70, 71, 72, 73, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 160, 161, 162, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 220, 221, 222, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287], "bbox": [0.0, 0.01348, 0.9943362831858407, 1.0], "iscrowd": 0, "area": 121482.73424999998, "rle": {"size": [500, 452], "counts": "k8g6l8100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100bM^2C=K5L4K5K5L4K5L4K5K5L4K5K5L4K5L4K5K5L4M3O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N200O100O1O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O2O00000000000000000O10000000000000000000000000000000000O100000000000000000000000000001O001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001UG^Ob2b0\\ME_2<_MJ\\26bM1[1YOdIg0o46W1UOiIe0n4=S1POnIc0n4d0m0lNSJa0m4l0g0fN[J>l4U1a0_NbJ=k4\\1;ZNhJ:k4a19WNkJ8k4b19XNkJ7i4c1UKeMa4e0_K4i4c1nJnMh4=`K2h4\\2^OkMiKJg4]2XOQNPLCe4`2VORNSL_Oe4a2SOUNWLZOd4e2POUN[LWOc4f2nNXN]LSOb4i2lNXNaLPO`4k2kNYNdLlN`4n2iNXNfLkN^4P3jNUNiLlN[4Q3jNTNkLkNZ4S3hNSNoLkNW4T3hNQNRMlNT4U3hNoMUMmNQ4V3hNmMXMmNo3X3gNkM[MnNl3X3hNjM]MoNi3Y3gNjM`MnNg3Z3gNhMcMnNe3\\3fNfMfMoNb3]3fNdMiMPO_3^3fNbMlMPO]3`3eN`MoMQOZ3a3dN_MSNQOW3a3eN^MUNROT3b3eN\\MXNROR3d3dNZM[NSOo2e3dNXM^NTOl2e3cNYMbNSOi2f3aNYMgNQOg2g3_NZMkNPOd2g3^NZMPOPO`2j3ZNXMWOnN^2Q4QNTMBlN[2V4kMPMKlNW2[4eMjL6nNQ2^4aMfL?oNl1b4\\M`Lj0QOf1f7\\N^H_1b7bNaHZ1^7hNeHT1Z7nNjHm0V7TOmHh0R7ZOQI`0P7BSI3S7OQIEV7<mHYOY7i0jHmN[7U1hH`N`7`1Y3100O1O100O100O100O100O100O100O100O100O100O1O100O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O101O000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000001O0000000000001O0000000000001O000000000000001O0000000000001O00002N2N1O2N1O2N1O2N1O2N2N1O2N1O2N2N2N3M2N2N3M2N3M2NO1N2N2N2N2N2nQ1"}, "label": "blonde lady yellow dress"}]}
{"id": 430394, "image": "COCO_train2014_000000430394.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a yellow shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 19, 20, 21, 22, 23, 24, 25, 34, 35, 36, 37, 38, 39, 40, 41, 50, 51, 52, 53, 54, 55, 56, 57, 66, 67, 68, 69, 70, 71, 72, 73, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 160, 161, 162, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 220, 221, 222, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287], "bbox": [0.0, 0.01348, 0.9943362831858407, 1.0], "iscrowd": 0, "area": 121482.73424999998, "rle": {"size": [500, 452], "counts": "k8g6l8100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100bM^2C=K5L4K5K5L4K5L4K5K5L4K5K5L4K5L4K5K5L4M3O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N200O100O1O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O2O00000000000000000O10000000000000000000000000000000000O100000000000000000000000000001O001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001UG^Ob2b0\\ME_2<_MJ\\26bM1[1YOdIg0o46W1UOiIe0n4=S1POnIc0n4d0m0lNSJa0m4l0g0fN[J>l4U1a0_NbJ=k4\\1;ZNhJ:k4a19WNkJ8k4b19XNkJ7i4c1UKeMa4e0_K4i4c1nJnMh4=`K2h4\\2^OkMiKJg4]2XOQNPLCe4`2VORNSL_Oe4a2SOUNWLZOd4e2POUN[LWOc4f2nNXN]LSOb4i2lNXNaLPO`4k2kNYNdLlN`4n2iNXNfLkN^4P3jNUNiLlN[4Q3jNTNkLkNZ4S3hNSNoLkNW4T3hNQNRMlNT4U3hNoMUMmNQ4V3hNmMXMmNo3X3gNkM[MnNl3X3hNjM]MoNi3Y3gNjM`MnNg3Z3gNhMcMnNe3\\3fNfMfMoNb3]3fNdMiMPO_3^3fNbMlMPO]3`3eN`MoMQOZ3a3dN_MSNQOW3a3eN^MUNROT3b3eN\\MXNROR3d3dNZM[NSOo2e3dNXM^NTOl2e3cNYMbNSOi2f3aNYMgNQOg2g3_NZMkNPOd2g3^NZMPOPO`2j3ZNXMWOnN^2Q4QNTMBlN[2V4kMPMKlNW2[4eMjL6nNQ2^4aMfL?oNl1b4\\M`Lj0QOf1f7\\N^H_1b7bNaHZ1^7hNeHT1Z7nNjHm0V7TOmHh0R7ZOQI`0P7BSI3S7OQIEV7<mHYOY7i0jHmN[7U1hH`N`7`1Y3100O1O100O100O100O100O100O100O100O100O100O1O100O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O101O000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000001O0000000000001O0000000000001O000000000000001O0000000000001O00002N2N1O2N1O2N1O2N1O2N2N1O2N1O2N2N2N3M2N2N3M2N3M2NO1N2N2N2N2N2nQ1"}, "label": "a woman in a yellow shirt"}]}
{"id": 127298, "image": "COCO_train2014_000000127298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kite laying on the grass\"."}], "task": "refering", "answer_template": "The \"a kite laying on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216], "bbox": [0.29304687500000004, 0.2966355140186916, 0.8622343749999999, 0.8381931464174455], "iscrowd": 0, "area": 27045.395000000008, "rle": {"size": [321, 640], "counts": "lTk18c98H8H8H8PNP2\\Od0D<D<I7O1N2O1O1N2O1N2O1N2O0O01O001O001N2O001O001O001O001O001O001N101O1O001O001O001O001O001N101O001O001O001O00000000O010000000000000O1000000000O10O100000000000000O1000O10000000O1000000000000O10O10O1000000O10000O1000000O01000O10000O1000000OmLiJ^1W5^NnJa1R5\\NRKc1m4ZNXKd1i4YN[Kf1e4VN`Ki1`4TNdKk1[4RNjKm1V4PNnKo1R4mMTLQ2l3lMXLS2h3iM]LV2c3gMaLX2_3dMfL[2Z3bMjL]2V3_MoL`2Z5N2O1O1O1O2N1O1O1O1O1O1O1O1N2O1O1O2N1O0000001O0000000O2O0000001O000000001O000O10001O0000001O000000001O0O10001O000000001O000000001N1000001O000000001O0000000O2O0000001O000O10001O00000O101O00000O2O0000000O1000001O0O100000000O100000000O10001O00000O100000000O100000000O101O00000O100000000O10000000001N100000000O100000000O1000001O0O100000000O100000000O1000001O000O100000000O100000000O101O00000O100000000O10000000001N100000000O100000000O10000000kcl0"}, "label": "a kite laying on the grass"}]}
{"id": 127298, "image": "COCO_train2014_000000127298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the colorful kite\"."}], "task": "refering", "answer_template": "The \"the colorful kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216], "bbox": [0.29304687500000004, 0.2966355140186916, 0.8622343749999999, 0.8381931464174455], "iscrowd": 0, "area": 27045.395000000008, "rle": {"size": [321, 640], "counts": "lTk18c98H8H8H8PNP2\\Od0D<D<I7O1N2O1O1N2O1N2O1N2O0O01O001O001N2O001O001O001O001O001O001N101O1O001O001O001O001O001N101O001O001O001O00000000O010000000000000O1000000000O10O100000000000000O1000O10000000O1000000000000O10O10O1000000O10000O1000000O01000O10000O1000000OmLiJ^1W5^NnJa1R5\\NRKc1m4ZNXKd1i4YN[Kf1e4VN`Ki1`4TNdKk1[4RNjKm1V4PNnKo1R4mMTLQ2l3lMXLS2h3iM]LV2c3gMaLX2_3dMfL[2Z3bMjL]2V3_MoL`2Z5N2O1O1O1O2N1O1O1O1O1O1O1O1N2O1O1O2N1O0000001O0000000O2O0000001O000000001O000O10001O0000001O000000001O0O10001O000000001O000000001N1000001O000000001O0000000O2O0000001O000O10001O00000O101O00000O2O0000000O1000001O0O100000000O100000000O10001O00000O100000000O100000000O101O00000O100000000O10000000001N100000000O100000000O1000001O0O100000000O100000000O1000001O000O100000000O100000000O101O00000O100000000O10000000001N100000000O100000000O10000000kcl0"}, "label": "the colorful kite"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"strawberries and cream in an fruit / snacks tray\"."}], "task": "refering", "answer_template": "The \"strawberries and cream in an fruit / snacks tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174], "bbox": [0.050859375, 0.056144329896907215, 0.5916718750000001, 0.4498762886597938], "iscrowd": 0, "area": 61612.66895, "rle": {"size": [485, 640], "counts": "hh?[1a=`0@;E;E;E;E;E;D<^Ob0_O?B>000000000000000001O000000000O1000000000000000001O000000000O100000000000000000000000001O000000000000000000000001O00000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000O100000000000001O00000000000O10000000000000000000000000000O100000000000000000000000000O2O0000000000000000000000000O1000000000000000000000000000000000000000001O0000000000000000000O1000000000000000000000000000000000000000000000001O0000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000001O000O100001O1O2N1O1O2N1O2Lf0ZOn0ROd0\\O;E<D;E;F:E;E;E;E;EP^k3"}, "label": "strawberries and cream in an fruit / snacks tray"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one bowl of strawberries and one of baby carrots\"."}], "task": "refering", "answer_template": "The \"one bowl of strawberries and one of baby carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174], "bbox": [0.050859375, 0.056144329896907215, 0.5916718750000001, 0.4498762886597938], "iscrowd": 0, "area": 61612.66895, "rle": {"size": [485, 640], "counts": "hh?[1a=`0@;E;E;E;E;E;D<^Ob0_O?B>000000000000000001O000000000O1000000000000000001O000000000O100000000000000000000000001O000000000000000000000001O00000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000O100000000000001O00000000000O10000000000000000000000000000O100000000000000000000000000O2O0000000000000000000000000O1000000000000000000000000000000000000000001O0000000000000000000O1000000000000000000000000000000000000000000000001O0000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000001O000O100001O1O2N1O1O2N1O2Lf0ZOn0ROd0\\O;E<D;E;F:E;E;E;E;EP^k3"}, "label": "one bowl of strawberries and one of baby carrots"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the container with carrots\"."}], "task": "refering", "answer_template": "The \"the container with carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181], "bbox": [0.5944375, 0.0571340206185567, 0.8759062499999999, 0.4675257731958763], "iscrowd": 0, "area": 31816.03435, "rle": {"size": [485, 640], "counts": "aUd57h<Y2kNU1ZOf0[Oe0\\Od0L4M3L3N0O10O0100O100O100O100O01000000000000000O01000000000000000O10O10000000000000O10O1000000000000000000O10000000000000000O10000000001O00000O1000000000000000000O10000000000000000O100000001O000000000O10000000000000000O100000000010O0001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O7I?A?A?A8H4L5K5K4L5K5L3L5K5K4L5@`0WOh0XOlXU1"}, "label": "the container with carrots"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the section with the carrots\"."}], "task": "refering", "answer_template": "The \"the section with the carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181], "bbox": [0.5944375, 0.0571340206185567, 0.8759062499999999, 0.4675257731958763], "iscrowd": 0, "area": 31816.03435, "rle": {"size": [485, 640], "counts": "aUd57h<Y2kNU1ZOf0[Oe0\\Od0L4M3L3N0O10O0100O100O100O100O01000000000000000O01000000000000000O10O10000000000000O10O1000000000000000000O10000000000000000O10000000001O00000O1000000000000000000O10000000000000000O100000001O000000000O10000000000000000O100000000010O0001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O7I?A?A?A8H4L5K5K4L5K5L3L5K5K4L5@`0WOh0XOlXU1"}, "label": "the section with the carrots"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"few strawberries and cumber and cherries with some designed items are in a box is displayed in the image\"."}], "task": "refering", "answer_template": "The \"few strawberries and cumber and cherries with some designed items are in a box is displayed in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358], "bbox": [0.015328125000000001, 0.4359587628865979, 0.601140625, 0.9168453608247422], "iscrowd": 0, "area": 81254.44895, "rle": {"size": [485, 640], "counts": "]T5:\\>`0A?@`0A?@`0@`0A?@`0A?@`0A?@`0K5N1O2N2N2N2O0O2N1O1O1O101N1O1O1O1O2O000000001O000000000000000000001O000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000001O00000000000000O10000000000000000000000000000000000O100000000000000000000000O10000000O10000000000000000000000000000000000O100O1O1O1O1O1O100000000000000000000O100000000000000000000000000000000000O10000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O100000000001O1O1O1O1O1O1O1O1O1N2N2N2N<Cb0_Oa0_O`0_Ob0_Oa0_O`0@a0^Oa0@a0_Oa0_O`0_OV\\h3"}, "label": "few strawberries and cumber and cherries with some designed items are in a box is displayed in the image"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tray of food at the bottom left\"."}], "task": "refering", "answer_template": "The \"tray of food at the bottom left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358], "bbox": [0.015328125000000001, 0.4359587628865979, 0.601140625, 0.9168453608247422], "iscrowd": 0, "area": 81254.44895, "rle": {"size": [485, 640], "counts": "]T5:\\>`0A?@`0A?@`0@`0A?@`0A?@`0A?@`0K5N1O2N2N2N2O0O2N1O1O1O101N1O1O1O1O2O000000001O000000000000000000001O000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000001O00000000000000O10000000000000000000000000000000000O100000000000000000000000O10000000O10000000000000000000000000000000000O100O1O1O1O1O1O100000000000000000000O100000000000000000000000000000000000O10000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O100000000001O1O1O1O1O1O1O1O1O1N2N2N2N<Cb0_Oa0_O`0_Ob0_Oa0_O`0@a0^Oa0@a0_Oa0_O`0_OV\\h3"}, "label": "tray of food at the bottom left"}]}
{"id": 225604, "image": "COCO_train2014_000000225604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the animal crackers and cookie\"."}], "task": "refering", "answer_template": "The \"the animal crackers and cookie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.590921875, 0.44719587628865975, 0.9110781250000001, 0.9145979381443299], "iscrowd": 0, "area": 41050.23990000001, "rle": {"size": [485, 640], "counts": "aXc53Q?8H8H8G9H8H9G8H8G9SDdNb9i3G8G:F9H8G9G2O1N2N2O1N2N2O1N2N2O1O1O1O1O1O001O000000001O000000001O0000001O000000001O0000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O100000000000000000000000000000000000001O001O0O2O00001O001O001O001O1O001O001O1O001O001O001O1O0O2O2N7H8H7I8I7H8H8H7I8H8H8I7H7I8H8H8H7I8I7H8H8H7I8\\OSdj0"}, "label": "the animal crackers and cookie"}]}
{"id": 274763, "image": "COCO_train2014_000000274763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl sitting in stripe shirt\"."}], "task": "refering", "answer_template": "The \"girl sitting in stripe shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 312, 313, 335], "bbox": [0.580859375, 0.6265320665083135, 0.775953125, 0.9454156769596199], "iscrowd": 0, "area": 4910.224, "rle": {"size": [421, 640], "counts": "aXi43o<5K5K5M3M3N1mD[Od9g0XF^Od9e0ZF[Of9g0WF[Og9h0VFYOj9i0TFWOl9k0QFWOm9l0PFUOP:l0nEUOQ:n0mEROT:o0iEROW:P1fEQOZ:Q1dEoN]:R1`EnNa:T1\\EmNd:U1YElNh:h10G9F;XOg0M3M4M2M3M4M200O2O0O1O2O0O1O[CI_<6aCN[<2eC1X<OgC5W<IjC;]<101O0000001O0A^O[Db0c;A\\D`0R;DiDN4>R;GfDN7;R;6mDJT;6kDKT;5lDKT;6kDJV;6iDKV;6iDJW;7hDIX;8gDHY;9fDHZ;7fDIZ;7fDIZ;7fDJY;6gDJY;6gDJY;7fDI[;6eDKZ;5fDK[;4eDL[;4eDL\\;3dDM\\;2eDN\\;0eD0\\;NeD2W:EZF8_O4W:EZF6_O4W:HZF2_O7W:HZFO_O9W:JYFL@8Z:MUFJA7\\:1RFGC5^:5mEGFO`:<hEEJKa:b0bECNHb:g0^EA2Dc:l0YE@V;b0hD_OX;c0eD]O[;V11O001O010O001O010O2O1N3M2O1N2O1N2N2O0O2WORD8o;JRD3n;OSDMn;4UDHk;:WDAj;`0=1O0001O1N5L4Fd`j1"}, "label": "girl sitting in stripe shirt"}]}
{"id": 274763, "image": "COCO_train2014_000000274763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing shorts sitting in a blue chair looking at an item in her hands\"."}], "task": "refering", "answer_template": "The \"a woman wearing shorts sitting in a blue chair looking at an item in her hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 312, 313, 335], "bbox": [0.580859375, 0.6265320665083135, 0.775953125, 0.9454156769596199], "iscrowd": 0, "area": 4910.224, "rle": {"size": [421, 640], "counts": "aXi43o<5K5K5M3M3N1mD[Od9g0XF^Od9e0ZF[Of9g0WF[Og9h0VFYOj9i0TFWOl9k0QFWOm9l0PFUOP:l0nEUOQ:n0mEROT:o0iEROW:P1fEQOZ:Q1dEoN]:R1`EnNa:T1\\EmNd:U1YElNh:h10G9F;XOg0M3M4M2M3M4M200O2O0O1O2O0O1O[CI_<6aCN[<2eC1X<OgC5W<IjC;]<101O0000001O0A^O[Db0c;A\\D`0R;DiDN4>R;GfDN7;R;6mDJT;6kDKT;5lDKT;6kDJV;6iDKV;6iDJW;7hDIX;8gDHY;9fDHZ;7fDIZ;7fDIZ;7fDJY;6gDJY;6gDJY;7fDI[;6eDKZ;5fDK[;4eDL[;4eDL\\;3dDM\\;2eDN\\;0eD0\\;NeD2W:EZF8_O4W:EZF6_O4W:HZF2_O7W:HZFO_O9W:JYFL@8Z:MUFJA7\\:1RFGC5^:5mEGFO`:<hEEJKa:b0bECNHb:g0^EA2Dc:l0YE@V;b0hD_OX;c0eD]O[;V11O001O010O001O010O2O1N3M2O1N2O1N2N2O0O2WORD8o;JRD3n;OSDMn;4UDHk;:WDAj;`0=1O0001O1N5L4Fd`j1"}, "label": "a woman wearing shorts sitting in a blue chair looking at an item in her hands"}]}
{"id": 414032, "image": "COCO_train2014_000000414032.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goat standing whose right hand side is fully visible\"."}], "task": "refering", "answer_template": "The \"a goat standing whose right hand side is fully visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 269, 270], "bbox": [0.5416875, 0.39419203747072595, 0.870921875, 0.7868618266978923], "iscrowd": 0, "area": 20030.768600000003, "rle": {"size": [427, 640], "counts": "WPa4a0f<5H7H8O1N10O000011M3N1_NjNmFX1m8A]Fa0c9ETF>l9HkE<S:_1O001O2M2O1O1O1O1O1O2N1O01O01O2O1O2M3N2N1N2O1N1000O0100O10O10O10O10O1000O10O1000G9J6O0O20000LYFeLf9[35LSFjLm9U3TFkLl9U3TFkLl9T3UFlLj9U3WFiLj9V36O1O1O1M3M3M3J6N2N2O1O10000001O001O001O0000001O0000001O00001O2NOdE_MP:a2lEeMR:l2TFlL]9Z3\\FgLd9Y3\\FgL5MW9]3bFgL6MJO[9d3fFcLZ9^3eFbL[9g31O0O[OWLjGe3Z8]LdG`3^8bLaG]3`8cLaG[3_8gL`GY3`8hL_GW3a8kL^GU3b8lL]GT3c8mL\\GS3d8nL[GR3e8oLZGQ3f8PMYGo2h8QMXGo2h8RMWGn2i8RMWGn2i8SMVGm2j8SMVGm2j8TMUGm2j8SMVGm2j8SMVGn2i8SMVGm2j8SMVGm2k8RMUGV2OXMO=m86SGV2=]M`8=SGW2>YM`8`3e0102M2N3N1N000000000O2M200O2O00O0O2O11O00001O00000000O100BiFhLW9P3YGgLh8S3k0J6N2N3N2N1O1O1O1TO^EnNa:R1eEhN\\:X1hEbN[:]1hE\\N]:c1f0O1N2N1L5A>N3F910O010O10O10O100O100O10O01O1O1O1O1O2N1N2NZ^R1"}, "label": "a goat standing whose right hand side is fully visible"}]}
{"id": 414032, "image": "COCO_train2014_000000414032.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small animal in snow to the left of picture\"."}], "task": "refering", "answer_template": "The \"small animal in snow to the left of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 256, 257, 258, 259, 279, 280, 281, 282, 303, 304, 305], "bbox": [0.160015625, 0.47782201405152225, 0.3143125, 0.9125058548009367], "iscrowd": 0, "area": 12541.689599999996, "rle": {"size": [427, 640], "counts": "UU[13X=2N1N3N2M2VF`0g5AkIU1l5kNkI_1S6cNfId1Y6\\N`Im1^6TN[IS2e6lMVIZ2i6fMVI]2h6dMUI`2i6`MVIc2h6^MUIf2i6ZMVIj2g6WMVIm2h6SMWIP3h6PMUIT3i6lLVIU3j6lLSIU3n6lLPIT3R7nLiHT3W7nLfHR3\\7PM_HQ3c7QMYHP3g7RMUHo2m7SMoGn2R8U10O10O1O010O010O10O01O10O0100O001O001OlNcH\\L]7d3dHZL]7f3eHXL[7h3gHVLY7j3iHSLX7n3R10dGPL]7P4TH_Ll7`4001O001O001N2M2O1000O101O00000O100iGRKP8U501O`K]HX3b7gLaHY3^7eLfHY3Y7hLhHX3W7gLkHX3T7hLnHX3Q7gLRIW3m6jLTIV3k6iLWIU3j6jLXIT3i6kLZIS3f6mL\\IP3d6PM`IZ2S7eMRIo1V7QNmHd1[7[NjH7aN0m8HfHNPOFb8;cHDf8;U2O1O2N`kf5"}, "label": "small animal in snow to the left of picture"}]}
{"id": 414032, "image": "COCO_train2014_000000414032.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lamb on the left\"."}], "task": "refering", "answer_template": "The \"lamb on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 256, 257, 258, 259, 279, 280, 281, 282, 303, 304, 305], "bbox": [0.160015625, 0.47782201405152225, 0.3143125, 0.9125058548009367], "iscrowd": 0, "area": 12541.689599999996, "rle": {"size": [427, 640], "counts": "UU[13X=2N1N3N2M2VF`0g5AkIU1l5kNkI_1S6cNfId1Y6\\N`Im1^6TN[IS2e6lMVIZ2i6fMVI]2h6dMUI`2i6`MVIc2h6^MUIf2i6ZMVIj2g6WMVIm2h6SMWIP3h6PMUIT3i6lLVIU3j6lLSIU3n6lLPIT3R7nLiHT3W7nLfHR3\\7PM_HQ3c7QMYHP3g7RMUHo2m7SMoGn2R8U10O10O1O010O010O10O01O10O0100O001O001OlNcH\\L]7d3dHZL]7f3eHXL[7h3gHVLY7j3iHSLX7n3R10dGPL]7P4TH_Ll7`4001O001O001N2M2O1000O101O00000O100iGRKP8U501O`K]HX3b7gLaHY3^7eLfHY3Y7hLhHX3W7gLkHX3T7hLnHX3Q7gLRIW3m6jLTIV3k6iLWIU3j6jLXIT3i6kLZIS3f6mL\\IP3d6PM`IZ2S7eMRIo1V7QNmHd1[7[NjH7aN0m8HfHNPOFb8;cHDf8;U2O1O2N`kf5"}, "label": "lamb on the left"}]}
{"id": 250192, "image": "COCO_train2014_000000250192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elaphant to the left in the distance\"."}], "task": "refering", "answer_template": "The \"elaphant to the left in the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 117, 118, 119, 120, 142, 143], "bbox": [0.00296875, 0.22980555555555557, 0.259359375, 0.4919444444444445], "iscrowd": 0, "area": 6047.73265, "rle": {"size": [360, 640], "counts": "Tj01U;4M210O001O1O1OO2O0O1O101N0001O00010O00001O01OO2O0O2O0O10OJH]E8c:LZE4e:701N101O0O1O2N3L4M2O001O001O010O10O10O10O10O10O01000O01000Oa`3VOcUL1T:l0F9G2M1UGdNj7]1_GWOb8j0[GWOg8c120O10000O10000O10000O01000O100000M3L4M4O0O100O2FkF_NU9`1mF^NT9a1nF^NR9a1QG\\NP9b1SG]Nn8`1>O1N3O00010O00001O000101\\FfNV9[1fFiNY9W1fFkNZ9T1fFmNY9S1fFoNY9Q1fFROY9`1N3MN3J5O14UGUNR8n1fGVNZ8n1^GVNc8W2001O1O1O1O1O1O1O1O1iNW1J6N3M4M3L4L4O01O010O010O01O010L3N3M2Nk_V5"}, "label": "elaphant to the left in the distance"}]}
{"id": 250192, "image": "COCO_train2014_000000250192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a elephant that is dark gray standing between two trees that is facing the left\"."}], "task": "refering", "answer_template": "The \"a elephant that is dark gray standing between two trees that is facing the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 117, 118, 119, 120, 142, 143], "bbox": [0.00296875, 0.22980555555555557, 0.259359375, 0.4919444444444445], "iscrowd": 0, "area": 6047.73265, "rle": {"size": [360, 640], "counts": "Tj01U;4M210O001O1O1OO2O0O1O101N0001O00010O00001O01OO2O0O2O0O10OJH]E8c:LZE4e:701N101O0O1O2N3L4M2O001O001O010O10O10O10O10O10O01000O01000Oa`3VOcUL1T:l0F9G2M1UGdNj7]1_GWOb8j0[GWOg8c120O10000O10000O10000O01000O100000M3L4M4O0O100O2FkF_NU9`1mF^NT9a1nF^NR9a1QG\\NP9b1SG]Nn8`1>O1N3O00010O00001O000101\\FfNV9[1fFiNY9W1fFkNZ9T1fFmNY9S1fFoNY9Q1fFROY9`1N3MN3J5O14UGUNR8n1fGVNZ8n1^GVNc8W2001O1O1O1O1O1O1O1O1iNW1J6N3M4M3L4L4O01O010O010O01O010L3N3M2Nk_V5"}, "label": "a elephant that is dark gray standing between two trees that is facing the left"}]}
{"id": 69971, "image": "COCO_train2014_000000069971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player in white uniform\"."}], "task": "refering", "answer_template": "The \"player in white uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333, 348, 349, 350, 351, 352, 353, 354], "bbox": [0.09275, 0.08059907834101382, 0.526078125, 0.9864285714285714], "iscrowd": 0, "area": 61793.233550000004, "rle": {"size": [434, 640], "counts": "RWi01_=<D<E:E<E;D<E;D<E;D<E6I3N2M2N3N2M3N1N3N2M2O2M3N2O0O2O1N101N2N2O0O2O1UJnJS3W6fKlIZ4T6dKnI[4T6bKnI]4S6aKoI_4R6]KQJb4P6\\KRJd4o5YKSJf4n5XKTJh4m5TKVJk4l5RKVJn4j5PKXJo4k5mJWJR5m5hJVJX5m5cJUJ\\5Q700O100O1O100O100\\MSHBn78_HAa78nH_OS7;VIBj6<ZIAg6=]I@d6>`I@`6?dI]O]6a0gI]OY6b0jI[OW6c0mI[OS6d0PJYOQ6e0TJWOm5g0ZJPOj5o0\\JgNi5W1Z3O100O100000000O100000000O100000000O100000000O100000000O100000000001O000000001O00000000001O00000000001O002N3M3M3M3M3M3M3M4YGWN]5l1]JYNa5h1[J[Ne5e1XJ_Ng5b1UJaNk5`1RJcNm5^1bHbMj0S1d6[1bHhMa0Q1l6X1bHmM9n0U7V1aHRN1l0]7R1aHYNIh0f7P1`H3_7N`H2`7N`H3_7N_H3a7N^H3a7M_H3a7N^H3a7N^H2b7N]H4b7M]H3c7N\\H3c7M]H3c7N[H4d7M[H3e7NZH3e7M[H3e7n2O001O001O001O001O001O1O001O001O001O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N2N1O2TObI`J`6^5l0M3M2O2M3M3POiGaLX8^3jG_LY8^3kG^LX8`3kG]LW8`3lG]LV8a3nG[LU8c3nGYLU8d3oGYLS8e3o0N2N3M2N2N2N2O1N2N2N2N2N2N2N3M2N2OZE]M]:a2`EdM`:Z2]EkMc:c2010O0001O01O0O1N2N3M2O1N2N3M2N2M2O1N2O1N3M2O1N2O1N2O1N3M2O1N3N1N2N3M2M3M3M4L3M3M4L5K7I6JSRP4"}, "label": "player in white uniform"}]}
{"id": 69971, "image": "COCO_train2014_000000069971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange and white uniform with a black cap on receiving an award\"."}], "task": "refering", "answer_template": "The \"a man in an orange and white uniform with a black cap on receiving an award\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333, 348, 349, 350, 351, 352, 353, 354], "bbox": [0.09275, 0.08059907834101382, 0.526078125, 0.9864285714285714], "iscrowd": 0, "area": 61793.233550000004, "rle": {"size": [434, 640], "counts": "RWi01_=<D<E:E<E;D<E;D<E;D<E6I3N2M2N3N2M3N1N3N2M2O2M3N2O0O2O1N101N2N2O0O2O1UJnJS3W6fKlIZ4T6dKnI[4T6bKnI]4S6aKoI_4R6]KQJb4P6\\KRJd4o5YKSJf4n5XKTJh4m5TKVJk4l5RKVJn4j5PKXJo4k5mJWJR5m5hJVJX5m5cJUJ\\5Q700O100O1O100O100\\MSHBn78_HAa78nH_OS7;VIBj6<ZIAg6=]I@d6>`I@`6?dI]O]6a0gI]OY6b0jI[OW6c0mI[OS6d0PJYOQ6e0TJWOm5g0ZJPOj5o0\\JgNi5W1Z3O100O100000000O100000000O100000000O100000000O100000000O100000000001O000000001O00000000001O00000000001O002N3M3M3M3M3M3M3M4YGWN]5l1]JYNa5h1[J[Ne5e1XJ_Ng5b1UJaNk5`1RJcNm5^1bHbMj0S1d6[1bHhMa0Q1l6X1bHmM9n0U7V1aHRN1l0]7R1aHYNIh0f7P1`H3_7N`H2`7N`H3_7N_H3a7N^H3a7M_H3a7N^H3a7N^H2b7N]H4b7M]H3c7N\\H3c7M]H3c7N[H4d7M[H3e7NZH3e7M[H3e7n2O001O001O001O001O001O1O001O001O001O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N2N1O2TObI`J`6^5l0M3M2O2M3M3POiGaLX8^3jG_LY8^3kG^LX8`3kG]LW8`3lG]LV8a3nG[LU8c3nGYLU8d3oGYLS8e3o0N2N3M2N2N2N2O1N2N2N2N2N2N2N3M2N2OZE]M]:a2`EdM`:Z2]EkMc:c2010O0001O01O0O1N2N3M2O1N2N3M2N2M2O1N2O1N3M2O1N2O1N2O1N3M2O1N3N1N2N3M2M3M3M4L3M3M4L5K7I6JSRP4"}, "label": "a man in an orange and white uniform with a black cap on receiving an award"}]}
{"id": 69971, "image": "COCO_train2014_000000069971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit offers a trophy to a stay player\"."}], "task": "refering", "answer_template": "The \"a man in a suit offers a trophy to a stay player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364], "bbox": [0.5628437500000001, 0.1234331797235023, 0.9063281250000002, 0.9980875576036867], "iscrowd": 0, "area": 52492.90490000002, "rle": {"size": [434, 640], "counts": "ihh42^=4L5K5VDBW:c0fE]OY:h0cEYO[:k0aEWO]:n0_EROb:P1[EQOd:S1XEmNh:W1UEiNj:[1REgNl:\\1REdNm:_1PEbNo:`1PEaNn:R2O1mIQNc0o1SO[Nl0g1mN`NR1`1iNeNV1\\1eNjNZ1W1_NoN`1R1[NSOd1n0lMfMfLb1^5i0dMH[29^MNb22XM4g2MSM9m2HPM:o2GRM8m2ISM7m2JRM6m2KSM5m2KSM5l2MTM2l2NTM2k2OUM1k20TM0k21VMNj22VMNi23WMMi24VMLi25XMJh26XMJg28XMHh28XMHg29YMGg2:YMEf2<ZMDf2<ZMDe2=[MCd2?[MAd2`0\\M@c2a0]M_Oa2d0_M[O`2f0`MZO_2g0aMYO]2j0bMVO]2k0cMUO\\2l0dMTOZ2o0eMQO[2o0fMPOZ2P1^L[L1e2a3P1\\L]L3c2a3Q1ZL^L4a2a3R1[L]L4a2a3S1ZL\\L5a2a3S1YL^L6^2a3U1XL]L7^2a3U1XL]L7^2a3V1VL]L7_2c3V1TL[L7a2e3U1PL]L:_2f3V1hKaL`0[2h3U1bKfLd0V2j3T2RLPNm3Q2PLRNP4o1`K`N`4Y5000000000000000000000000000000000000000000\\OUGaLk8[3dGZL\\8b3l0L4M3M3L4M3M3M3M3M3M3M3M3M3M3N2N2N2O1N2N200O1O1O100O1O1O100O1O10000000O10000000000000000000001O0000001O0000001O0000001O00000QGWNU6j1gI]NU6c1hIcNU6^1hIfNV6[1gIjNV6W1hIlNf3UOaLP2GnNe3UObLn1GPOe3TObLl1HROe3VO]JJ:2j0m1h0SOf3_O]JKT1d1g0UOe34aKh0h0VOf35_Kf0h0XOh35]Ke0g0YOk36ZKd0e0ZOP45XKc0c0]OS43WKb0a0_OW43TKa0?@\\42RK`0=B`41PK?;Ec4OoJ?9Ef40nJ=5Il4NkJX3T5e2N1O1O2N1O1O2N1N2N3L3N2N2N2N2N3M2K5K5L4L5K5K4K6I6K6I6I8I7I8B[DdNm;e0VDDT<3XDEn;1ZWi0"}, "label": "a man in a suit offers a trophy to a stay player"}]}
{"id": 69971, "image": "COCO_train2014_000000069971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit holding a trophy\"."}], "task": "refering", "answer_template": "The \"a man in a suit holding a trophy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364], "bbox": [0.5628437500000001, 0.1234331797235023, 0.9063281250000002, 0.9980875576036867], "iscrowd": 0, "area": 52492.90490000002, "rle": {"size": [434, 640], "counts": "ihh42^=4L5K5VDBW:c0fE]OY:h0cEYO[:k0aEWO]:n0_EROb:P1[EQOd:S1XEmNh:W1UEiNj:[1REgNl:\\1REdNm:_1PEbNo:`1PEaNn:R2O1mIQNc0o1SO[Nl0g1mN`NR1`1iNeNV1\\1eNjNZ1W1_NoN`1R1[NSOd1n0lMfMfLb1^5i0dMH[29^MNb22XM4g2MSM9m2HPM:o2GRM8m2ISM7m2JRM6m2KSM5m2KSM5l2MTM2l2NTM2k2OUM1k20TM0k21VMNj22VMNi23WMMi24VMLi25XMJh26XMJg28XMHh28XMHg29YMGg2:YMEf2<ZMDf2<ZMDe2=[MCd2?[MAd2`0\\M@c2a0]M_Oa2d0_M[O`2f0`MZO_2g0aMYO]2j0bMVO]2k0cMUO\\2l0dMTOZ2o0eMQO[2o0fMPOZ2P1^L[L1e2a3P1\\L]L3c2a3Q1ZL^L4a2a3R1[L]L4a2a3S1ZL\\L5a2a3S1YL^L6^2a3U1XL]L7^2a3U1XL]L7^2a3V1VL]L7_2c3V1TL[L7a2e3U1PL]L:_2f3V1hKaL`0[2h3U1bKfLd0V2j3T2RLPNm3Q2PLRNP4o1`K`N`4Y5000000000000000000000000000000000000000000\\OUGaLk8[3dGZL\\8b3l0L4M3M3L4M3M3M3M3M3M3M3M3M3M3N2N2N2O1N2N200O1O1O100O1O1O100O1O10000000O10000000000000000000001O0000001O0000001O0000001O00000QGWNU6j1gI]NU6c1hIcNU6^1hIfNV6[1gIjNV6W1hIlNf3UOaLP2GnNe3UObLn1GPOe3TObLl1HROe3VO]JJ:2j0m1h0SOf3_O]JKT1d1g0UOe34aKh0h0VOf35_Kf0h0XOh35]Ke0g0YOk36ZKd0e0ZOP45XKc0c0]OS43WKb0a0_OW43TKa0?@\\42RK`0=B`41PK?;Ec4OoJ?9Ef40nJ=5Il4NkJX3T5e2N1O1O2N1O1O2N1N2N3L3N2N2N2N2N3M2K5K5L4L5K5K4K6I6K6I6I8I7I8B[DdNm;e0VDDT<3XDEn;1ZWi0"}, "label": "a man in a suit holding a trophy"}]}
{"id": 69978, "image": "COCO_train2014_000000069978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very large tan crock holding decorative grasses that can almost touch the brick sidewalk\"."}], "task": "refering", "answer_template": "The \"a very large tan crock holding decorative grasses that can almost touch the brick sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294], "bbox": [0.710234375, 0.5169555035128806, 0.912234375, 0.869672131147541], "iscrowd": 0, "area": 12686.147150000004, "rle": {"size": [427, 640], "counts": "]Pn58k<:G:E:G9G9D<LP2oM3M3O100O100O100O100O00100O10000000000O1000O1000O10000O10000O1000O01000000001O00000O100O10000O100O100O2O0O1N2O1N2N2N2N3QOgFbMa9\\2dFYMd9e2b0004L4M3L4L4L4L4LO2K4_N`FSOd9m0bFiNb9Y1bF]Nb98lEd0g0iNa9b0jEf0R;WOPEk0o:SOSEn0l:POUER1j:lNWEV1h:gN[EZ1d:dN]E^1b:aN^Ea1`:_NaEa1_:^NaEc1_:[NbEf1^:YNcEh1\\:WNdEj1\\:UNdEl1n:0O00N3L3100O2N11011N5LJ5K6SOZDKl;5WDCo;;VD\\On;d0=000O1000000001N10001O001N1A^Rg0"}, "label": "a very large tan crock holding decorative grasses that can almost touch the brick sidewalk"}]}
{"id": 69978, "image": "COCO_train2014_000000069978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green plants are house in the concrete plant container on the far right\"."}], "task": "refering", "answer_template": "The \"green plants are house in the concrete plant container on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 200, 201, 202, 203, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294], "bbox": [0.710234375, 0.5169555035128806, 0.912234375, 0.869672131147541], "iscrowd": 0, "area": 12686.147150000004, "rle": {"size": [427, 640], "counts": "]Pn58k<:G:E:G9G9D<LP2oM3M3O100O100O100O100O00100O10000000000O1000O1000O10000O10000O1000O01000000001O00000O100O10000O100O100O2O0O1N2O1N2N2N2N3QOgFbMa9\\2dFYMd9e2b0004L4M3L4L4L4L4LO2K4_N`FSOd9m0bFiNb9Y1bF]Nb98lEd0g0iNa9b0jEf0R;WOPEk0o:SOSEn0l:POUER1j:lNWEV1h:gN[EZ1d:dN]E^1b:aN^Ea1`:_NaEa1_:^NaEc1_:[NbEf1^:YNcEh1\\:WNdEj1\\:UNdEl1n:0O00N3L3100O2N11011N5LJ5K6SOZDKl;5WDCo;;VD\\On;d0=000O1000000001N10001O001N1A^Rg0"}, "label": "green plants are house in the concrete plant container on the far right"}]}
{"id": 69978, "image": "COCO_train2014_000000069978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a concrete planter in front of two other planters\"."}], "task": "refering", "answer_template": "The \"a concrete planter in front of two other planters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338], "bbox": [0.482359375, 0.5529274004683841, 0.724296875, 0.9785948477751757], "iscrowd": 0, "area": 23938.29955, "rle": {"size": [427, 640], "counts": "hTQ4[1h;_1eN5J7J5F;G9L3M4L3M4L4L3O2M2N2O1N2N2O1N1O101N1O101N100O2N100O2N10001N10001O00000O10000000000O10000000000O100000000O10000000000O10000000000O10000001O00000000000000001O000000000000001O0O10000O1000000O2O000O101O1N2O001O1N2O1O1N2O1O1N101O1O1N2O2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2N2M3N000O2O00000O4M3M3C=_Ob0_O`0_Oa0_Oa0@lRY2"}, "label": "a concrete planter in front of two other planters"}]}
{"id": 69978, "image": "COCO_train2014_000000069978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a concrete planter in front of two other concrete planters\"."}], "task": "refering", "answer_template": "The \"a concrete planter in front of two other concrete planters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338], "bbox": [0.482359375, 0.5529274004683841, 0.724296875, 0.9785948477751757], "iscrowd": 0, "area": 23938.29955, "rle": {"size": [427, 640], "counts": "hTQ4[1h;_1eN5J7J5F;G9L3M4L3M4L4L3O2M2N2O1N2N2O1N1O101N1O101N100O2N100O2N10001N10001O00000O10000000000O10000000000O100000000O10000000000O10000000000O10000001O00000000000000001O000000000000001O0O10000O1000000O2O000O101O1N2O001O1N2O1O1N2O1O1N101O1O1N2O2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N2N2M3N000O2O00000O4M3M3C=_Ob0_O`0_Oa0_Oa0@lRY2"}, "label": "a concrete planter in front of two other concrete planters"}]}
{"id": 553308, "image": "COCO_train2014_000000553308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foot in the upper left corner\"."}], "task": "refering", "answer_template": "The \"the foot in the upper left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 115, 116, 117, 118, 139, 140, 141], "bbox": [0.0, 0.0, 0.16396875, 0.37435416666666665], "iscrowd": 0, "area": 10502.011349999999, "rle": {"size": [480, 640], "counts": "4X5d94000O10000000000O100000000O100000000O101O00000001O0000001O0000001O00000010O000001OiK_Fo2a9iLiFU3W9fLQGX3n8dLXGZ3h8bL_G[3a8`LgG]3Y8_LmG`3R8[LVHb3f9N3M2N3N2M2N3M2O2M2N3nMbD<S=N101N1O2N100O001O0001OO1000000000000000000000001O0O10000000001O00000O1O1O1O2N1O1N4M2N3LTdj7"}, "label": "the foot in the upper left corner"}]}
{"id": 553308, "image": "COCO_train2014_000000553308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing white shirt kneeling\"."}], "task": "refering", "answer_template": "The \"man wearing white shirt kneeling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 195, 196, 197, 213, 214, 215, 216, 218, 219, 220, 236, 237, 238, 239, 240, 242, 261, 262, 263, 264], "bbox": [0.204859375, 0.0400625, 0.650109375, 0.7105], "iscrowd": 0, "area": 57060.07659999999, "rle": {"size": [480, 640], "counts": "fbm1n0j=:G9F:G9F:G9G9F:K5L4L5L3L3M2N2O1N2N2N2O2M2N2O1N2N2N2O1N2N2N2O1N2N2O1N2O1O100O1O100O2O:E;E;F9F;E4M1N101N1O2O0O2N101N101N1O2O0O2N101O001O0O2O001O00001O0O2O001O001O001O0O2O001O001O0O2O001N101O0O2O001N10001N101O0O2O001N101O0O^MdJcK[5\\4gJcKY5[4kJdKS5\\4oJcKQ5\\4QKdKn4Z4UKeKj4[4XKeKg4Z4[KeKe4Y4]KhKa4X4`KhK`4V4cKjK\\4U4eKkKZ4U4hKkKW4S4kKmKU4R4mKmKR4S4oKnKP4P4SLoKm3P4TLQLj3o3XLPLh3n3ZLSLe3l3]LSLb3m3_LTL`3j3bLVL^3i3dLWLZ3i3gLWLY3g3jLYLU3f3lLZLS3f3oLZLP3e3QM[Lo2d3SM\\Ll2d3TM\\Ll2c3VM]Lh2d3XM\\Lh2c3YM^Lf2b3[M]Le2b3\\M_Lc2a3^M^La2b3`M_L_2a3bM^L^2a3cM^L^2b3cM]L]2b3dM^L[2c3fM[L[2d3fM\\LZ2c3gM\\LZ2d3gM[LY2d3hM[LX2f3iMYLW2g3iMXLX2g3jMWLW2i3iMVLX2i3jMULV2o3gMPLZ2S4dMlK\\2W4aMhK`2Z4_MdKb2_4[M`Ke2d4YMZKh2i4VMUKk2n4SMQKm2R5QMlJP3V5oLhJQ3\\5mLbJT3a5jL]JW3f5gLXJZ3k5dLTJ\\3o5aLPJ_3S6`LkIa3X6]LfId3]6ZLaIg3b6WL\\Ij3g6TLXIk3l6SLRIn3P7P22N1O1N3N1O2N1O2M2O1O2N1O2M2O1O2N1O2M2O1O001O1N2O1O1O1O0O2O1O1O1O1N2O00O2O000O101N100O101N1ZN^InJb6Q5cIjJ^6T5hIhJY6U5mIfJU6W5RJdJn5[5VJbJk5[5YJbJi5[5[JcJf5Z5]JdJe5Y5_JeJa5Z5bJcJ`5Z5dJdJ]5Y5gJdJ[5Y5hJfJY5W5kJfJW5X5lJfJT5X5PKeJR5X5]2M3N2M3M2N3YNnE\\NU:a1nE]NT:`1PF\\NS:a1QF[NR:c1PFZNR:d1RFYNP:d1SFYNP:d1g1N3N2M2N3M3M2N3N2M2N4L3M4M3L4L3M4L4K5K5L3L5L4K5LfmX3"}, "label": "man wearing white shirt kneeling"}]}
{"id": 553308, "image": "COCO_train2014_000000553308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man searching through an open trunk\"."}], "task": "refering", "answer_template": "The \"a man searching through an open trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 195, 196, 197, 213, 214, 215, 216, 218, 219, 220, 236, 237, 238, 239, 240, 242, 261, 262, 263, 264], "bbox": [0.204859375, 0.0400625, 0.650109375, 0.7105], "iscrowd": 0, "area": 57060.07659999999, "rle": {"size": [480, 640], "counts": "fbm1n0j=:G9F:G9F:G9G9F:K5L4L5L3L3M2N2O1N2N2N2O2M2N2O1N2N2N2O1N2N2N2O1N2N2O1N2O1O100O1O100O2O:E;E;F9F;E4M1N101N1O2O0O2N101N101N1O2O0O2N101O001O0O2O001O00001O0O2O001O001O001O0O2O001O001O0O2O001N101O0O2O001N10001N101O0O2O001N101O0O^MdJcK[5\\4gJcKY5[4kJdKS5\\4oJcKQ5\\4QKdKn4Z4UKeKj4[4XKeKg4Z4[KeKe4Y4]KhKa4X4`KhK`4V4cKjK\\4U4eKkKZ4U4hKkKW4S4kKmKU4R4mKmKR4S4oKnKP4P4SLoKm3P4TLQLj3o3XLPLh3n3ZLSLe3l3]LSLb3m3_LTL`3j3bLVL^3i3dLWLZ3i3gLWLY3g3jLYLU3f3lLZLS3f3oLZLP3e3QM[Lo2d3SM\\Ll2d3TM\\Ll2c3VM]Lh2d3XM\\Lh2c3YM^Lf2b3[M]Le2b3\\M_Lc2a3^M^La2b3`M_L_2a3bM^L^2a3cM^L^2b3cM]L]2b3dM^L[2c3fM[L[2d3fM\\LZ2c3gM\\LZ2d3gM[LY2d3hM[LX2f3iMYLW2g3iMXLX2g3jMWLW2i3iMVLX2i3jMULV2o3gMPLZ2S4dMlK\\2W4aMhK`2Z4_MdKb2_4[M`Ke2d4YMZKh2i4VMUKk2n4SMQKm2R5QMlJP3V5oLhJQ3\\5mLbJT3a5jL]JW3f5gLXJZ3k5dLTJ\\3o5aLPJ_3S6`LkIa3X6]LfId3]6ZLaIg3b6WL\\Ij3g6TLXIk3l6SLRIn3P7P22N1O1N3N1O2N1O2M2O1O2N1O2M2O1O2N1O2M2O1O001O1N2O1O1O1O0O2O1O1O1O1N2O00O2O000O101N100O101N1ZN^InJb6Q5cIjJ^6T5hIhJY6U5mIfJU6W5RJdJn5[5VJbJk5[5YJbJi5[5[JcJf5Z5]JdJe5Y5_JeJa5Z5bJcJ`5Z5dJdJ]5Y5gJdJ[5Y5hJfJY5W5kJfJW5X5lJfJT5X5PKeJR5X5]2M3N2M3M2N3YNnE\\NU:a1nE]NT:`1PF\\NS:a1QF[NR:c1PFZNR:d1RFYNP:d1SFYNP:d1g1N3N2M2N3M3M2N3N2M2N4L3M4M3L4L3M4L4K5K5L3L5L4K5LfmX3"}, "label": "a man searching through an open trunk"}]}
{"id": 94564, "image": "COCO_train2014_000000094564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bunch of carrots in the wood basket\"."}], "task": "refering", "answer_template": "The \"the bunch of carrots in the wood basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 198, 199, 200, 201, 213, 214, 215, 216, 229, 230, 231, 245], "bbox": [0.19796252927400468, 0.0625, 0.8236533957845434, 0.71815625], "iscrowd": 0, "area": 64038.33915000001, "rle": {"size": [640, 427], "counts": "S^e1T1Xb0f0K6TJUNfD1a3o1e7VNfD0]3o1i7WNeD0Z3o1l7VNfDL[3S2k7WNeDG]3W2j7XNdDC^3[2i7WNfD^O^3`2iL\\MT:`1PIS2m6oMPIT2m6oMQIT2k6oMTIR2i6RNUIP2h6RNWIo1f6TNXIo1d6TN[Im1b6VN]Ij1a6YN^Ih1_6[N_If1a6\\N]Id1c6]N\\Ic1d6^N[Ib1d6`N[I`1e6aNYI`1g6aNXI_1h6cNVI]1j6dNUI\\1j6fNTI[1l6fNSI[1l6fNSIZ1m6gNRIY1n6hNRIW1m6lNQIT1o6mNPIS1P7nNoHU1n6mNPIW1k6lNTIW1h6kNVIY1f6iNXI[1d6gNZI]1b6eN]I^1`6cN^Ia1_6`N_Ie1\\6]NbIg1[6ZNcIj1Z6WNcIn1Y6TNeIP2X6PNgIT2U6nMiIV2T6kMiIZ2T6gMjI]2R6eMlI]2S6dMkI]2U6dMhI_2V6fMeI[2[6hMaIZ2^6jM]IW2b6mMYIV2f6nMUIS2k6QNPIQ2n6SNnHn1R7a5O010O0010O01O0010O01O010O0010O01ON2O2N1O1N3N1E;C>C<A?A?@a0M2O101N1O1O1O2O0O1O2N101N1O2N2N2O1N1O2N2N2O0O2N2N2N2O0O2N2O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O001O1O100O1O1O1O001O1O1O00100O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O1O2N3M3aNSBfLo=W3RBhLQ>T3RBjLQ>S3QBkLR>R3oAmLT>P3nAnLU>n2mAQMV>l2lARMW>k2kASMX>j2iAUMY>h2jAVMX>h2iAWMY>g2iATM\\>j2fAQM_>l2cAQM`>n2bAmLc>Q3^AkLg>S3U1M4M3M3M3L4M4L3M3M3L4M4L3M3L4M2N2N2M3N2N2N2N2M3N2N1O2M3N2N2N6J6I7Jkb^1"}, "label": "the bunch of carrots in the wood basket"}]}
{"id": 94564, "image": "COCO_train2014_000000094564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"serveral bunchs of carrots\"."}], "task": "refering", "answer_template": "The \"serveral bunchs of carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 33, 34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 198, 199, 200, 201, 213, 214, 215, 216, 229, 230, 231, 245], "bbox": [0.19796252927400468, 0.0625, 0.8236533957845434, 0.71815625], "iscrowd": 0, "area": 64038.33915000001, "rle": {"size": [640, 427], "counts": "S^e1T1Xb0f0K6TJUNfD1a3o1e7VNfD0]3o1i7WNeD0Z3o1l7VNfDL[3S2k7WNeDG]3W2j7XNdDC^3[2i7WNfD^O^3`2iL\\MT:`1PIS2m6oMPIT2m6oMQIT2k6oMTIR2i6RNUIP2h6RNWIo1f6TNXIo1d6TN[Im1b6VN]Ij1a6YN^Ih1_6[N_If1a6\\N]Id1c6]N\\Ic1d6^N[Ib1d6`N[I`1e6aNYI`1g6aNXI_1h6cNVI]1j6dNUI\\1j6fNTI[1l6fNSI[1l6fNSIZ1m6gNRIY1n6hNRIW1m6lNQIT1o6mNPIS1P7nNoHU1n6mNPIW1k6lNTIW1h6kNVIY1f6iNXI[1d6gNZI]1b6eN]I^1`6cN^Ia1_6`N_Ie1\\6]NbIg1[6ZNcIj1Z6WNcIn1Y6TNeIP2X6PNgIT2U6nMiIV2T6kMiIZ2T6gMjI]2R6eMlI]2S6dMkI]2U6dMhI_2V6fMeI[2[6hMaIZ2^6jM]IW2b6mMYIV2f6nMUIS2k6QNPIQ2n6SNnHn1R7a5O010O0010O01O0010O01O010O0010O01ON2O2N1O1N3N1E;C>C<A?A?@a0M2O101N1O1O1O2O0O1O2N101N1O2N2N2O1N1O2N2N2O0O2N2N2N2O0O2N2O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O001O1O100O1O1O1O001O1O1O00100O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O1O2N3M3aNSBfLo=W3RBhLQ>T3RBjLQ>S3QBkLR>R3oAmLT>P3nAnLU>n2mAQMV>l2lARMW>k2kASMX>j2iAUMY>h2jAVMX>h2iAWMY>g2iATM\\>j2fAQM_>l2cAQM`>n2bAmLc>Q3^AkLg>S3U1M4M3M3M3L4M4L3M3M3L4M4L3M3L4M2N2N2M3N2N2N2N2M3N2N1O2M3N2N2N6J6I7Jkb^1"}, "label": "serveral bunchs of carrots"}]}
{"id": 160101, "image": "COCO_train2014_000000160101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing red sox outfit to the right of the man in the wheelchair\"."}], "task": "refering", "answer_template": "The \"man wearing red sox outfit to the right of the man in the wheelchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 83, 84, 85, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 291, 292, 293, 315, 316], "bbox": [0.5760625, 0.1213566739606127, 0.79590625, 0.8606783369803063], "iscrowd": 0, "area": 27703.053849999997, "rle": {"size": [457, 640], "counts": "`jT59f=;F9M4N2N2N1O2N2O1N101N2O1O2M3N2N2M3N2kNeNWE]1^:oN`EV1W:POfE^1i9hNTFd1^9`N`Fc1Z9cNdF`1Y9k1L3N3L4M2M4L4M4L4L3M4L4La0_O7I4L3M3M4L3L5L3M3Ni0VO4L4L4L4L2N1O000010O1O2O0O2N101N2N101N1O2O0O2N=C?A3M0O2O001O0O20O100O2O0O100O1O101N100O100O2O0O101N3N1O2N2N1O1O1OO1O1O1O1O1O1O1O1O1O001O1O1O1O2N2N2N2mJbIW2a6eMhIR2[6kMnIc0_ORNf6W1UJ5AaN]6W1[J^OMWOk5X1aJ^Nb06P5X1ZMdNj2X1ZMdNi2Y1[McNh2Y1^MbNe2[1_MaNd2\\1W5L4M4K4M5KTYj1"}, "label": "man wearing red sox outfit to the right of the man in the wheelchair"}]}
{"id": 160101, "image": "COCO_train2014_000000160101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing wrist watch in left hand\"."}], "task": "refering", "answer_template": "The \"a man wearing wrist watch in left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 83, 84, 85, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 291, 292, 293, 315, 316], "bbox": [0.5760625, 0.1213566739606127, 0.79590625, 0.8606783369803063], "iscrowd": 0, "area": 27703.053849999997, "rle": {"size": [457, 640], "counts": "`jT59f=;F9M4N2N2N1O2N2O1N101N2O1O2M3N2N2M3N2kNeNWE]1^:oN`EV1W:POfE^1i9hNTFd1^9`N`Fc1Z9cNdF`1Y9k1L3N3L4M2M4L4M4L4L3M4L4La0_O7I4L3M3M4L3L5L3M3Ni0VO4L4L4L4L2N1O000010O1O2O0O2N101N2N101N1O2O0O2N=C?A3M0O2O001O0O20O100O2O0O100O1O101N100O100O2O0O101N3N1O2N2N1O1O1OO1O1O1O1O1O1O1O1O1O001O1O1O1O2N2N2N2mJbIW2a6eMhIR2[6kMnIc0_ORNf6W1UJ5AaN]6W1[J^OMWOk5X1aJ^Nb06P5X1ZMdNj2X1ZMdNi2Y1[McNh2Y1^MbNe2[1_MaNd2\\1W5L4M4K4M5KTYj1"}, "label": "a man wearing wrist watch in left hand"}]}
{"id": 160101, "image": "COCO_train2014_000000160101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a wheelchair wearing a blue shirt who is missing one of his legs\"."}], "task": "refering", "answer_template": "The \"a man in a wheelchair wearing a blue shirt who is missing one of his legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 262, 263, 264, 265, 285, 286, 287, 288, 308, 310, 311], "bbox": [0.361046875, 0.13048140043763676, 0.5969218749999999, 0.8675711159737418], "iscrowd": 0, "area": 25465.905950000004, "rle": {"size": [457, 640], "counts": "`XW3P1S=8G9I5M3M3M3O1N2O0O2O1N2N2O1N2O1O1O1O100O1O1O1O1`I_MX1a2fNcMW1_2eNeMY1\\2eNhMX1Y2fNjMY1V2dNoMY1R2eNQNZ1o1dNUNZ1l1bNXN]1i1`N[N^1f1_N^N^1d1^N^Nc1f1WNZNk1k1nMTNU2n1fMSN\\2P2_MoMd2S2XMmMi2W2oIbLl2V1X3i2[LXMg3o2PLPMS4P3kKPMW4o2gKRM[4n2cKQM_4P3_KPMc4P3ZKPMi4P3UKPMm4P3QKoLR5Q3kJPMV5P3iJoLZ5Q3dJoL^5Q3`JnLb5S3[JnLg5R3WJnLm4o3RKPLl4T4RKmKl4U4TKjKk4Y4TKgKi4]4UKcKj4`4UK`Kk4a4TK^Km4c4QK^KP5a4PK^KR5b4mJ^KU5a4iJ`KX5`4gJ`K[5_4dJbK]5]4bJcK`5]4]JdKe5[4ZJfKg5Z4WJfKk5Z4RJgKP6X4oIiKR6W4lIiKV6V4hIkKZ6U4dIlK]6T4aIlKb6R4\\IoKf6Q4XIPLi6o3VIQLQ2aNb2^5ZKRLR2hNa2U5[KTLR2nN_2`6`MfI]2X6dMmIX2R6hMPJX2m5jMSJV2i5mMYJR2b5RN`Jm1\\5VNfJj1T5ZNmJf1o4]NSKb1i4aNYK^1c4eN_K[1\\4aNlKm1f3oM_LT2]3jMfLY2V3dMmL`2o2VM\\Mm2`2oLdMT3Y2jLjM_3m1\\LWNk3e1nK_NR4d1fKaN[4`1`KcNa4]1\\KeNe4\\1WKgNi4Z1SKhNn4X1oJjNQ5X1eJPO[5o4000O1O1000000kNXJiIh5Q6bJkI^5o5mJlIS5o5WKjIl4Q6\\KiIg4R6aKhIb4T6l1L4K5L3PMfGN_8OdGLa82aGHe85]GGh86[GEj89WGCm8;VG@P9<RG@U9<mF_OY9<jF@\\9<fF@a9;bF@d9<aF]Of9?_FYOh9b0S2M1N3M3N2L<D;Dcnb3"}, "label": "a man in a wheelchair wearing a blue shirt who is missing one of his legs"}]}
{"id": 160101, "image": "COCO_train2014_000000160101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a handicapped man wearing a blue tshirt sitting on a wheel chair with a golf bat in his left hand\"."}], "task": "refering", "answer_template": "The \"a handicapped man wearing a blue tshirt sitting on a wheel chair with a golf bat in his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 262, 263, 264, 265, 285, 286, 287, 288, 308, 310, 311], "bbox": [0.361046875, 0.13048140043763676, 0.5969218749999999, 0.8675711159737418], "iscrowd": 0, "area": 25465.905950000004, "rle": {"size": [457, 640], "counts": "`XW3P1S=8G9I5M3M3M3O1N2O0O2O1N2N2O1N2O1O1O1O100O1O1O1O1`I_MX1a2fNcMW1_2eNeMY1\\2eNhMX1Y2fNjMY1V2dNoMY1R2eNQNZ1o1dNUNZ1l1bNXN]1i1`N[N^1f1_N^N^1d1^N^Nc1f1WNZNk1k1nMTNU2n1fMSN\\2P2_MoMd2S2XMmMi2W2oIbLl2V1X3i2[LXMg3o2PLPMS4P3kKPMW4o2gKRM[4n2cKQM_4P3_KPMc4P3ZKPMi4P3UKPMm4P3QKoLR5Q3kJPMV5P3iJoLZ5Q3dJoL^5Q3`JnLb5S3[JnLg5R3WJnLm4o3RKPLl4T4RKmKl4U4TKjKk4Y4TKgKi4]4UKcKj4`4UK`Kk4a4TK^Km4c4QK^KP5a4PK^KR5b4mJ^KU5a4iJ`KX5`4gJ`K[5_4dJbK]5]4bJcK`5]4]JdKe5[4ZJfKg5Z4WJfKk5Z4RJgKP6X4oIiKR6W4lIiKV6V4hIkKZ6U4dIlK]6T4aIlKb6R4\\IoKf6Q4XIPLi6o3VIQLQ2aNb2^5ZKRLR2hNa2U5[KTLR2nN_2`6`MfI]2X6dMmIX2R6hMPJX2m5jMSJV2i5mMYJR2b5RN`Jm1\\5VNfJj1T5ZNmJf1o4]NSKb1i4aNYK^1c4eN_K[1\\4aNlKm1f3oM_LT2]3jMfLY2V3dMmL`2o2VM\\Mm2`2oLdMT3Y2jLjM_3m1\\LWNk3e1nK_NR4d1fKaN[4`1`KcNa4]1\\KeNe4\\1WKgNi4Z1SKhNn4X1oJjNQ5X1eJPO[5o4000O1O1000000kNXJiIh5Q6bJkI^5o5mJlIS5o5WKjIl4Q6\\KiIg4R6aKhIb4T6l1L4K5L3PMfGN_8OdGLa82aGHe85]GGh86[GEj89WGCm8;VG@P9<RG@U9<mF_OY9<jF@\\9<fF@a9;bF@d9<aF]Of9?_FYOh9b0S2M1N3M3N2L<D;Dcnb3"}, "label": "a handicapped man wearing a blue tshirt sitting on a wheel chair with a golf bat in his left hand"}]}
{"id": 160101, "image": "COCO_train2014_000000160101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player to the left of the man in the wheel chair\"."}], "task": "refering", "answer_template": "The \"a baseball player to the left of the man in the wheel chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.036828125, 0.14798687089715534, 0.36984374999999997, 0.8049234135667397], "iscrowd": 0, "area": 29504.392750000003, "rle": {"size": [457, 640], "counts": "XQ;3U>1N2N3M2O2M2N3M2N3N1N3M2N3N1N2N3M2O2O0O2O0O2O0O2O0O101N10001N12N2N3N1N2N2N01O01O00010O00010O000010O0001O01O01O010O00iMkNZGU1d8QOYGo0d8WOXGi0f8]OVGc0h8CUG<i8JSG7j8ORG1l85QGJm8<oFDP9a0lF@Q9e0mFZOQ9l0kFTOS9R1iFoNT9W1iFhNU9^1gFbNW9d1eF\\NU9n1hFRNQ9X2kFhMo8b2mF^Mn8Q4J7I6K5J7I6J6K5J7N1N2O1O2N001O1O001O1N101O001O1O001O1O001O010O10O010O0100O010O01G8I8G8I8H7H9N1O2N1O1O2N1O2N1O2N1O2NUIhJS4l5mKWJo3j5RLWJl3k5RLWJl3j5SLXJk3j5SLXJj3k5ULVJi3l5ULVJi3l5ULVJi3l5ULVJj3j5ULYJi3h5VLYJj3g5TL[Jl3e5RL]Jn3c5PL_JP4a5oK`JP4`5oKbJQ4^5mKdJS4\\5jKgJV4Y5gKkJX4U5fKmJY4h3dJYLP12\\4b3gJ[Lk06^4]3iJ]Lg08`4X3lJ`La0;c4R3oJcL;>e4m2RKfL6`0h4g2UKiL1b0j4c2WKkLLf0l4\\2[KmLGj0n4V2^KPMBl0P5Q2aKSM\\Oo0T5j1cKWMVOR1W5e1eKYMQOU1[5^1gKNY41hKOY40gK0Y40gK1Y4kMfJ8Q1m1Y4kMiJ4o0Q2Y4iMmJ2j0U2Y4hMRKOe0Y2Z4gMUKLa0]2Z4fMZKH=b2Z4cM_KG7f2Z4`MeKG1i2\\4ZMkKIIT2kNeMc58PLKBU2oNgMa54SLM^OT2TOiM^50RL4\\Oo1YOlM[5LSL9YOl1]OnMZ5FTL`0UOj1@oMY5BULd0SOh1CQN^78oHd1GSNZ79oHb1JTNW7:oH_1NWNQ7:RI\\11XNn6<QIZ14WNm6?oHW18XNi6b0oHS1<XNg6e0mHQ1?XNe6f0mHo0Q9ROoFk0T9UOlFi0V9WOjFf0X9ZOiFc0Z9]OfFa0\\9_OdF>^9CbF:a9E`F9b9DaF9a9DcF9`9BeF<]9@iF;Y9BmF9V9CQG8c;K6Jkcc5"}, "label": "a baseball player to the left of the man in the wheel chair"}]}
{"id": 332135, "image": "COCO_train2014_000000332135.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl at her birthday party\"."}], "task": "refering", "answer_template": "The \"a girl at her birthday party\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 214, 215, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281], "bbox": [0.17021875, 0.09004166666666666, 0.613484375, 0.7372291666666667], "iscrowd": 0, "area": 36797.46144999999, "rle": {"size": [480, 640], "counts": "\\]c11n>3N2M3M2O1N2N2O0O2N2N2O1N2N10O01O010O00010O003N2M3M3N2M1O0gMQOaFo0_9O`E5_:U201O001O1O10O01O1O1O001O5K;E1O2N100O2N1O2N1OkNVFWMj9T2PGhMn8j1eGRNZ8d1VHXNh7a1eHZNZ7g1o20O1O1O1O1O100O001O1O100O1O1PFRNn6o1cHaNZ7a1fH`NX7`1iHaNU7`1kH`NS7b1mH_NQ7b1oH_No6b1QI_Nl6c1TI^Nj6c1VI^Nh6c1YI]Nd6e1\\I\\Nb6e1^I\\N`6d1aI]N]6d1cI]NZ6f1eI[NY6m1`ITN^6S2\\InMa6[2XIfMf6b2SI_Mk6i2nHXMg5X4SJiKl5^4mIcKR6d4hI\\KW6k4bIVK]6P5]IQKb6V5WIkJg6h6O1O1O1O1O1O1N2O1O1O1O1O001N2O1O100000000000000000gNgIkIY6S6iImIW6Q6lInIT6P6nIPJR6n5QJQJo5l5TJTJl5j5VJVJj5h5YJWJg5g5[JYJe5e5^JZJb5e5_J[Ja5d5aJ[J_5c5cJ]J]5b5dJ^J\\5a5fJ^JZ5a5gJ_JY5_5jJ`JV5_5kJaJT5`5mJ_JS5`5nJ`JR5`5oJ_JQ5a5oJ_JQ5`5PK`JP5`5QK^JP5a5QK_Jo4a5RK^Jn4b5RK^Jo4`5RK`Jn4`5RK`Jo4^5SKaJn4^5RKbJo4\\5RKdJn4\\5RKdJo4[5QKeJP5Y5QKgJP5X5PKhJP5W5QKiJP5V5PKjJQ5T5PKlJQ5R5PKnJP5R5PKnJQ5P5PKPKQ5o4oJQKR5m4oJSKQ5l4PKTKQ5k4oJUKR5i4oJWKR5i4mJWKS5i4nJVKS5i4mJWKT5i4jJXKW5g4hJZKX5f4hJZKY5f4eJ[K\\5d4dJ\\K]5d4aJ]K_5c4aJ]K`5b4_J_Kb5a4]J^Ki5]4VJdKn5X4QJiKS6T4lIlKX6P4gIQL]6k3cIULb6g3\\IZLh6b3XI^Ll6^3SIcLQ7Y3oHgLV7T3iHmL[7o2dHRM`7j2`HVMd7f2[H[Mi7a2WH_Ml7^2SHcMo7[2QHeMQ8Y2nGhMU8U2kGkMW8S2hGnMZ8P2eGQN^8l1bGTN`8j1_GWNc8g1]GYNf8d1YG]Ni8a1WG_Nk8_1UGaNn8\\1UGaNm8]1VG`Nl8^1VG`Nm8]1VG`Nl8^1WG_Nk8^1YG_Nj8^1YG_Ni8^1[G_Ng8_1]FnM0a0f9]1ZFVN09i9_1WFZN14k9_1TF`N1NP:\\1PFhN1IS:[1kEoN3CW:X1gEXO\\;d0dD^O_;<bDG_;5`DNc;L^D6d;F\\D=f;]OZDf0h;VOXDm0j;mNWDU1i<2M`0A1O1O1N2O1O1O1N21O1N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2M2O2N1O1O2N1N3N1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1Nlmc3"}, "label": "a girl at her birthday party"}]}
{"id": 332135, "image": "COCO_train2014_000000332135.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a pink tanktop\"."}], "task": "refering", "answer_template": "The \"a girl in a pink tanktop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 214, 215, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281], "bbox": [0.17021875, 0.09004166666666666, 0.613484375, 0.7372291666666667], "iscrowd": 0, "area": 36797.46144999999, "rle": {"size": [480, 640], "counts": "\\]c11n>3N2M3M2O1N2N2O0O2N2N2O1N2N10O01O010O00010O003N2M3M3N2M1O0gMQOaFo0_9O`E5_:U201O001O1O10O01O1O1O001O5K;E1O2N100O2N1O2N1OkNVFWMj9T2PGhMn8j1eGRNZ8d1VHXNh7a1eHZNZ7g1o20O1O1O1O1O100O001O1O100O1O1PFRNn6o1cHaNZ7a1fH`NX7`1iHaNU7`1kH`NS7b1mH_NQ7b1oH_No6b1QI_Nl6c1TI^Nj6c1VI^Nh6c1YI]Nd6e1\\I\\Nb6e1^I\\N`6d1aI]N]6d1cI]NZ6f1eI[NY6m1`ITN^6S2\\InMa6[2XIfMf6b2SI_Mk6i2nHXMg5X4SJiKl5^4mIcKR6d4hI\\KW6k4bIVK]6P5]IQKb6V5WIkJg6h6O1O1O1O1O1O1N2O1O1O1O1O001N2O1O100000000000000000gNgIkIY6S6iImIW6Q6lInIT6P6nIPJR6n5QJQJo5l5TJTJl5j5VJVJj5h5YJWJg5g5[JYJe5e5^JZJb5e5_J[Ja5d5aJ[J_5c5cJ]J]5b5dJ^J\\5a5fJ^JZ5a5gJ_JY5_5jJ`JV5_5kJaJT5`5mJ_JS5`5nJ`JR5`5oJ_JQ5a5oJ_JQ5`5PK`JP5`5QK^JP5a5QK_Jo4a5RK^Jn4b5RK^Jo4`5RK`Jn4`5RK`Jo4^5SKaJn4^5RKbJo4\\5RKdJn4\\5RKdJo4[5QKeJP5Y5QKgJP5X5PKhJP5W5QKiJP5V5PKjJQ5T5PKlJQ5R5PKnJP5R5PKnJQ5P5PKPKQ5o4oJQKR5m4oJSKQ5l4PKTKQ5k4oJUKR5i4oJWKR5i4mJWKS5i4nJVKS5i4mJWKT5i4jJXKW5g4hJZKX5f4hJZKY5f4eJ[K\\5d4dJ\\K]5d4aJ]K_5c4aJ]K`5b4_J_Kb5a4]J^Ki5]4VJdKn5X4QJiKS6T4lIlKX6P4gIQL]6k3cIULb6g3\\IZLh6b3XI^Ll6^3SIcLQ7Y3oHgLV7T3iHmL[7o2dHRM`7j2`HVMd7f2[H[Mi7a2WH_Ml7^2SHcMo7[2QHeMQ8Y2nGhMU8U2kGkMW8S2hGnMZ8P2eGQN^8l1bGTN`8j1_GWNc8g1]GYNf8d1YG]Ni8a1WG_Nk8_1UGaNn8\\1UGaNm8]1VG`Nl8^1VG`Nm8]1VG`Nl8^1WG_Nk8^1YG_Nj8^1YG_Ni8^1[G_Ng8_1]FnM0a0f9]1ZFVN09i9_1WFZN14k9_1TF`N1NP:\\1PFhN1IS:[1kEoN3CW:X1gEXO\\;d0dD^O_;<bDG_;5`DNc;L^D6d;F\\D=f;]OZDf0h;VOXDm0j;mNWDU1i<2M`0A1O1O1N2O1O1O1N21O1N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2M2O2N1O1O2N1N3N1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1Nlmc3"}, "label": "a girl in a pink tanktop"}]}
{"id": 332135, "image": "COCO_train2014_000000332135.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with long blond hair and a blue shirt\"."}], "task": "refering", "answer_template": "The \"woman with long blond hair and a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 224, 225, 226, 227, 228, 229, 238, 239, 240, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.372484375, 0.002458333333333333, 0.9983281250000001, 0.9890208333333333], "iscrowd": 0, "area": 74731.38239999997, "rle": {"size": [480, 640], "counts": "kj_31l>3N2M4L3N3N1N3M2O2M2O1N3M2O2M2N3N1N200O2O0O10000O100O2O0O100O10000O2O0O100O10000O10O0010O02O0O100O1O2O0O100O101N100O10001N1000000O2O00000O101O0O1000001N1000O10O10000O10O1000O100000O010000O10O0100O100O10O010O010O010O010O010O010O010O0100O010O01O10O01O100O00100O1O010O1O10O01O100O00100O1O00100O1O010O1O10O01O100O00100O1O010O1O10O0100O10O0100O1O010O100O010O100O010O010O0010O0100O010O010O010O010O1O0O2N1O2N1O2N1O2N2N1N3M2M4M2M4M2M4M2N3H8_O`0B?M2O1N3M2N2O1N3M2N2O2M2]NVKoHl4P7YKjHj4S7]KgHd4X7aKbHa4e3_KcN6aM]4m3^K^N;_MZ4T4ZK]N=\\MZ4X4ZK[N<ZM\\4\\4YKYN;XM_4`4UKXN<UMa4d4TKVN;TMa4h4TKTNk5P2UJoMh5T2XJlMe5EgIOe0;b5BjI8d06_5@nI?c00\\5^OSJf0a0LY5\\OWJl0a0GV5YO[JU1>BT5WO^J\\1?\\OP5UOcJc1=XOm4oNkJn18ROk4iNRKY23nNh4bNZKe2OhNd4]NbKo2JdNa4VNkKY3EaN^4nMSLb3B`NX4hMZLj3A^NR4aMbLT4_OZNj3^MjL]4]OUNa3\\MWMd4XOPNX3\\McMj4UOjMo2[MQNo4QOfMl4FfKhN^Ob1g4UO^L\\OkN_1a4dNYMNWN^1[4YNkM<jM[1V4PN[Ng0`MY1o3hMmNQ1TMW1j3^M^O^1hLT1Y8nNhGR1V8POjGP1S8ROnGn0o7UOQHk0m7WOSHi0j7YOWHg0f7[O[He0c7\\O^Hd0_7^ObHb0[7@fH`0X7AiH?T7CmH=P7EQI;m6FTI\\NaNU1W8a0`IRN[N\\1R8d0nIeMRNf1n7f0]JYO`5i0bJVOZ5m0iJQOT5Q1nJnNo4T1SKkNj4W1XKhNe4W1`KhN]4W1hKhNU4W1QLgNl3W1ZLhNc3W1bLhN[3W1jLhNS3W1RMhNk2W1ZMhNc2V1cMiNZ2V1lMhNQ2V1UNiNh1a0aIkNm4c0_1a0gIiNo4e0W1a0nIeNQ5i0o0`0SJdNS5k0h0?XJbNW5m0`0>\\JbNY5o09=>B@<f0BYO9n0FPO6W1IgN3a1K]N1j1NUNMR22lMJ\\24bMHe27ZMDn2:PMBW3=d52N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1OIcBmN[=S1gBmNX=S1iBmNU=S1mBmNR=S1oBmNQ=Q1PCPOP=o0QCQOo<l0TCTOk<j0XCVOh<h0ZCXOf<e0]C[Oc<b0`C^O_<a0cC_O]<>fCBZ<;iCEW<9kCGU<6nCJQ<4RDLn;2TDNl;OWD1i;LZD4e;K]D5W=000O10000000000000000000000000000000000O1000000000000000000000^2bMg2YMg2YMg2YMg2YMd1"}, "label": "woman with long blond hair and a blue shirt"}]}
{"id": 332135, "image": "COCO_train2014_000000332135.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman decorate the cake\"."}], "task": "refering", "answer_template": "The \"woman decorate the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 224, 225, 226, 227, 228, 229, 238, 239, 240, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.372484375, 0.002458333333333333, 0.9983281250000001, 0.9890208333333333], "iscrowd": 0, "area": 74731.38239999997, "rle": {"size": [480, 640], "counts": "kj_31l>3N2M4L3N3N1N3M2O2M2O1N3M2O2M2N3N1N200O2O0O10000O100O2O0O100O10000O2O0O100O10000O10O0010O02O0O100O1O2O0O100O101N100O10001N1000000O2O00000O101O0O1000001N1000O10O10000O10O1000O100000O010000O10O0100O100O10O010O010O010O010O010O010O010O0100O010O01O10O01O100O00100O1O010O1O10O01O100O00100O1O00100O1O010O1O10O01O100O00100O1O010O1O10O0100O10O0100O1O010O100O010O100O010O010O0010O0100O010O010O010O010O1O0O2N1O2N1O2N1O2N2N1N3M2M4M2M4M2M4M2N3H8_O`0B?M2O1N3M2N2O1N3M2N2O2M2]NVKoHl4P7YKjHj4S7]KgHd4X7aKbHa4e3_KcN6aM]4m3^K^N;_MZ4T4ZK]N=\\MZ4X4ZK[N<ZM\\4\\4YKYN;XM_4`4UKXN<UMa4d4TKVN;TMa4h4TKTNk5P2UJoMh5T2XJlMe5EgIOe0;b5BjI8d06_5@nI?c00\\5^OSJf0a0LY5\\OWJl0a0GV5YO[JU1>BT5WO^J\\1?\\OP5UOcJc1=XOm4oNkJn18ROk4iNRKY23nNh4bNZKe2OhNd4]NbKo2JdNa4VNkKY3EaN^4nMSLb3B`NX4hMZLj3A^NR4aMbLT4_OZNj3^MjL]4]OUNa3\\MWMd4XOPNX3\\McMj4UOjMo2[MQNo4QOfMl4FfKhN^Ob1g4UO^L\\OkN_1a4dNYMNWN^1[4YNkM<jM[1V4PN[Ng0`MY1o3hMmNQ1TMW1j3^M^O^1hLT1Y8nNhGR1V8POjGP1S8ROnGn0o7UOQHk0m7WOSHi0j7YOWHg0f7[O[He0c7\\O^Hd0_7^ObHb0[7@fH`0X7AiH?T7CmH=P7EQI;m6FTI\\NaNU1W8a0`IRN[N\\1R8d0nIeMRNf1n7f0]JYO`5i0bJVOZ5m0iJQOT5Q1nJnNo4T1SKkNj4W1XKhNe4W1`KhN]4W1hKhNU4W1QLgNl3W1ZLhNc3W1bLhN[3W1jLhNS3W1RMhNk2W1ZMhNc2V1cMiNZ2V1lMhNQ2V1UNiNh1a0aIkNm4c0_1a0gIiNo4e0W1a0nIeNQ5i0o0`0SJdNS5k0h0?XJbNW5m0`0>\\JbNY5o09=>B@<f0BYO9n0FPO6W1IgN3a1K]N1j1NUNMR22lMJ\\24bMHe27ZMDn2:PMBW3=d52N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1OIcBmN[=S1gBmNX=S1iBmNU=S1mBmNR=S1oBmNQ=Q1PCPOP=o0QCQOo<l0TCTOk<j0XCVOh<h0ZCXOf<e0]C[Oc<b0`C^O_<a0cC_O]<>fCBZ<;iCEW<9kCGU<6nCJQ<4RDLn;2TDNl;OWD1i;LZD4e;K]D5W=000O10000000000000000000000000000000000O1000000000000000000000^2bMg2YMg2YMg2YMg2YMd1"}, "label": "woman decorate the cake"}]}
{"id": 225641, "image": "COCO_train2014_000000225641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white motorcycle\"."}], "task": "refering", "answer_template": "The \"a green and white motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 157, 158, 159, 160, 182, 183, 205, 206], "bbox": [0.434828125, 0.0022500000000000003, 0.9994375000000002, 0.5123541666666667], "iscrowd": 0, "area": 34624.97889999997, "rle": {"size": [480, 640], "counts": "j\\R41n>100O1O1O1O1O100O2N1O1O1O100O1O1O1O1O101N1O1O1O1kNZOoCg0P<ZOPDf0n;]OPDd0o;]OPDd0o;^OoCc0P<^OoCd0n;_OQDa0n;@QDa0n;APD`0o;APDa0o;_OPDb0P<_OnCb0R<^OnCc0P<_OnCb0R<^OmCd0R<]OlCd0T<\\OlCd0T<\\OkCf0S<\\OlCd0T<\\OkCe0U<\\OjCe0U<[OjCf0U<\\OiCf0V<ZOjCf0V<ZOiCg0W<ZOhCg0V<ZOiCg0W<ZOgCh0X<XOgCi0Y<XOfCh0Y<YOfCi0Y<o001O0O2O001O0O2O001N101O001N101N1N3N1O2N1O20O01O000000O100O2O0O100O2N100O2N1O1O2N1O1O1O2N1N2O2M2O1N3N1N2O2M2O101O0O101O00001N10001O0O101O000O101O000O2O0000000O1000000O0100000O1000O10000000O10O10000000000O10000000000O10000000000O10000000000O10002N2N2N000O2O000000001N100000001O000O10001O00000O101O001O1O1O1N101O1O1O001N2O1O1O001O1N2O0O2O1N2O0O3N2N2N2N2N2N2NR1nN1O001N101O001O001O001O001O001O0O2O001O001O001O001O001O001N101O001O001O001O001O001O0O2O001O00jJbGk3]8oKkGo3d901N2O0O2O1O0O2O0O2O1N103M3L3ON3L4M3M2N2N2N3M2N2N2N2N2N1O2N2N2N2N2O1N101N2O1O1N2O0O101N100O10N101N2O0O2O001N101N101N101N2O0O2O0O2O2M4M6I2O1O2M2O1N2O1N2O1N2O2iL_O"}, "label": "a green and white motorcycle"}]}
{"id": 225641, "image": "COCO_train2014_000000225641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green color bike in the right top most side of the image\"."}], "task": "refering", "answer_template": "The \"green color bike in the right top most side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 157, 158, 159, 160, 182, 183, 205, 206], "bbox": [0.434828125, 0.0022500000000000003, 0.9994375000000002, 0.5123541666666667], "iscrowd": 0, "area": 34624.97889999997, "rle": {"size": [480, 640], "counts": "j\\R41n>100O1O1O1O1O100O2N1O1O1O100O1O1O1O1O101N1O1O1O1kNZOoCg0P<ZOPDf0n;]OPDd0o;]OPDd0o;^OoCc0P<^OoCd0n;_OQDa0n;@QDa0n;APD`0o;APDa0o;_OPDb0P<_OnCb0R<^OnCc0P<_OnCb0R<^OmCd0R<]OlCd0T<\\OlCd0T<\\OkCf0S<\\OlCd0T<\\OkCe0U<\\OjCe0U<[OjCf0U<\\OiCf0V<ZOjCf0V<ZOiCg0W<ZOhCg0V<ZOiCg0W<ZOgCh0X<XOgCi0Y<XOfCh0Y<YOfCi0Y<o001O0O2O001O0O2O001N101O001N101N1N3N1O2N1O20O01O000000O100O2O0O100O2N100O2N1O1O2N1O1O1O2N1N2O2M2O1N3N1N2O2M2O101O0O101O00001N10001O0O101O000O101O000O2O0000000O1000000O0100000O1000O10000000O10O10000000000O10000000000O10000000000O10000000000O10002N2N2N000O2O000000001N100000001O000O10001O00000O101O001O1O1O1N101O1O1O001N2O1O1O001O1N2O0O2O1N2O0O3N2N2N2N2N2N2NR1nN1O001N101O001O001O001O001O001O0O2O001O001O001O001O001O001N101O001O001O001O001O001O0O2O001O00jJbGk3]8oKkGo3d901N2O0O2O1O0O2O0O2O1N103M3L3ON3L4M3M2N2N2N3M2N2N2N2N2N1O2N2N2N2N2O1N101N2O1O1N2O0O101N100O10N101N2O0O2O001N101N101N101N2O0O2O0O2O2M4M6I2O1O2M2O1N2O1N2O1N2O2iL_O"}, "label": "green color bike in the right top most side of the image"}]}
{"id": 225641, "image": "COCO_train2014_000000225641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle and a black motorcycle helmet sitting on the ground next to it\"."}], "task": "refering", "answer_template": "The \"a black motorcycle and a black motorcycle helmet sitting on the ground next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.15506250000000002, 0.232578125, 0.9887708333333334], "iscrowd": 0, "area": 34800.12745, "rle": {"size": [480, 640], "counts": "Z3a;_30O100O100O100O100O100O100O10O0100O100O100O100O100O100O1O100O100O100O100O100OjDXMV:h2jE]MQ:b2QFbMj9^2VFfMf9Y2[FlM`9T2`FQN[9n1fFQO[8o0eG[OQ8d0PHFf7:ZH1[7NfH<P7DPIf0f6YO[Il0`6TO`Io0]6POdIS1Y6mNgIU1W6jNjIY1S6gNmI[1Q6dNPJ_1m5`NTJc1i5]NWJe1g5ZN[Jh1b5XN^Jk1_5TNbJn1\\5RNdJQ2Y5oMgJT2V5lMjJV2T5jMlJY2Q5gMoJ[2o4eMQK^2l4bMTKa2i4_MWKc2g4]MYKf2d4ZM\\Ki2a4WM_Kk2_4UMaKn2\\4RMdKP3Z4PMfKS3W4mLiKV3T4jLlKX3R4hLnK[3o3eLRL]3k3cLUL_3i3aLWLb3f3^LZLd3d3\\L[Lg3c3YL]Li3a3WL_Lk3_3ULaLm3]3SLcLP4Z3PLfLR4X3nKhLT4V3lKjLV4T3jKlLX4R3hKnL[4o2eKQM]4m2cKSM_4k2aKUMa4i2_KWMc4g2]KYMe4e2[K[Mh4b2XK^Mj4`2WK_Mk4_2UKaMm4]2SKcMo4[2RKcMQ5[2oJeMR5Z2nJfMT5X2mJgMU5W2kJiMW5U2iJkMY5S2hJlMZ5R2fJnM\\5P2dJPN^5n1bJRN_5m1bJRN`5l1aJSNa5k1`JTNb5j1_JUNc5i1]JWNe5g1\\JXNf5f1[JYNf5f1[JYNg5e1ZJZNh5d1YJZNj5d1WJ[Nk5c1VJ\\Nl5b1UJ]Nm5a1TJ^Nn5`1SJ^No5a1QJ_NQ6_1PJ`NR6^1oI`NT6^1mIaNU6]1lIbNV6\\1kIbNX6\\1iIcNX6\\1iIcNY6[1hIcN[6[1fIdN\\6Z1eIeN]6Y1cIfN`6X1aIgNa6W1`IhNa6W1`IgNc6W1]IiNe6U1\\IjNf6T1[IjNh6T1d3N2NWgU7"}, "label": "a black motorcycle and a black motorcycle helmet sitting on the ground next to it"}]}
{"id": 225641, "image": "COCO_train2014_000000225641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small motorcycle that is barely visible off to the left with a shiny headlight\"."}], "task": "refering", "answer_template": "The \"small motorcycle that is barely visible off to the left with a shiny headlight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.15506250000000002, 0.232578125, 0.9887708333333334], "iscrowd": 0, "area": 34800.12745, "rle": {"size": [480, 640], "counts": "Z3a;_30O100O100O100O100O100O100O10O0100O100O100O100O100O100O1O100O100O100O100O100OjDXMV:h2jE]MQ:b2QFbMj9^2VFfMf9Y2[FlM`9T2`FQN[9n1fFQO[8o0eG[OQ8d0PHFf7:ZH1[7NfH<P7DPIf0f6YO[Il0`6TO`Io0]6POdIS1Y6mNgIU1W6jNjIY1S6gNmI[1Q6dNPJ_1m5`NTJc1i5]NWJe1g5ZN[Jh1b5XN^Jk1_5TNbJn1\\5RNdJQ2Y5oMgJT2V5lMjJV2T5jMlJY2Q5gMoJ[2o4eMQK^2l4bMTKa2i4_MWKc2g4]MYKf2d4ZM\\Ki2a4WM_Kk2_4UMaKn2\\4RMdKP3Z4PMfKS3W4mLiKV3T4jLlKX3R4hLnK[3o3eLRL]3k3cLUL_3i3aLWLb3f3^LZLd3d3\\L[Lg3c3YL]Li3a3WL_Lk3_3ULaLm3]3SLcLP4Z3PLfLR4X3nKhLT4V3lKjLV4T3jKlLX4R3hKnL[4o2eKQM]4m2cKSM_4k2aKUMa4i2_KWMc4g2]KYMe4e2[K[Mh4b2XK^Mj4`2WK_Mk4_2UKaMm4]2SKcMo4[2RKcMQ5[2oJeMR5Z2nJfMT5X2mJgMU5W2kJiMW5U2iJkMY5S2hJlMZ5R2fJnM\\5P2dJPN^5n1bJRN_5m1bJRN`5l1aJSNa5k1`JTNb5j1_JUNc5i1]JWNe5g1\\JXNf5f1[JYNf5f1[JYNg5e1ZJZNh5d1YJZNj5d1WJ[Nk5c1VJ\\Nl5b1UJ]Nm5a1TJ^Nn5`1SJ^No5a1QJ_NQ6_1PJ`NR6^1oI`NT6^1mIaNU6]1lIbNV6\\1kIbNX6\\1iIcNX6\\1iIcNY6[1hIcN[6[1fIdN\\6Z1eIeN]6Y1cIfN`6X1aIgNa6W1`IhNa6W1`IgNc6W1]IiNe6U1\\IjNf6T1[IjNh6T1d3N2NWgU7"}, "label": "small motorcycle that is barely visible off to the left with a shiny headlight"}]}
{"id": 225641, "image": "COCO_train2014_000000225641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown motorbike with the black seat\"."}], "task": "refering", "answer_template": "The \"the brown motorbike with the black seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319], "bbox": [0.20993750000000003, 0.1529375, 0.9531562499999999, 0.8075625], "iscrowd": 0, "area": 84017.71315000001, "rle": {"size": [480, 640], "counts": "Vmn11n>6J5J7J6I5L4K5L4L4K5L4K5L4K5L4M4L4K5L4L3M4L4L4K5L4L3M4L4K2O1O1O1O001N2O1O1O001O1N2O1O001O1N2O0O2O1O1O1O010O100O10O0100O1000O0100O100O010O10000O010O100O010O100O10O01O001O001O001O010O001O001O001O001O010O001O001O001O1O001O001N101O001O001O001O001O001O001O001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O0010O010O00010O010O01O01O010O01O01O010O01O010O01O01O010O010O01O010O0O2O001O0O2O001O0O2O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O001O0O2O00N2N2O0O2N2O1N1O2OO0M3M4M2M3M4M2M301O00101N2N1O2N2N1O2N2N101N2N2N1O2N2N2N2N101NiE`MR8^2oGfMn7X2RHmMk7Q2UHTNg7l1YHXNd7f1\\H`N`7^1`HgN]7W1cHnNY7R1gHROV7l0jHYOS7e0mH@P7>QIFl68TILi64VIOi6OWI4h6JYI8f6FZI=e6A[Ib0d6]O\\Ie0c6\\OZIg0d6\\OXIg0g6ZOWIh0h6ZOTIh0l6ZOPIi0o6YOmHj0R7WOlHk0S7b3O1O1N2O1O1O1O001O1O1N2O1O1O1O1O1O001000000O1000000O10O1000O1000000O1000O10O1000O01O00001O01O0eIbIU4]6iKeIW4[6gKhIX4X6fKjI[4U6cKnI\\4R6bKPJ^4P6`KSJ`4l5^KVJb4j5\\KYJc4g5\\KZJe4e5YK^Jf4c5WK_Ji4a5VKaJj4^5TKdJl4\\5RKgJm4Y5RKhJo4W5oJlJP5T5nJnJR5R5mJPKS5P5jJRKV5R5dJQK[5T5_JmJb5V5XJmJg5X5RJjJn5[5lIgJT6R701O010O0010O01O010O0010O01O010O0010O01O010O0010O01O010O0010O01O010O001O001O00001O001O001O001O001O001O0O2O0O2O0O2N101N101N101N101N101N101N1O2O1N4M2M4L4M2M4M3L3M4M3L3N3L4L3N3L4M2M4M3L3L5K5K4L5K5K5K7I7I7I7I7I7I7I7I7Ihg="}, "label": "the brown motorbike with the black seat"}]}
{"id": 225641, "image": "COCO_train2014_000000225641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ariel brown color bike standing position\"."}], "task": "refering", "answer_template": "The \"ariel brown color bike standing position\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319], "bbox": [0.20993750000000003, 0.1529375, 0.9531562499999999, 0.8075625], "iscrowd": 0, "area": 84017.71315000001, "rle": {"size": [480, 640], "counts": "Vmn11n>6J5J7J6I5L4K5L4L4K5L4K5L4K5L4M4L4K5L4L3M4L4L4K5L4L3M4L4K2O1O1O1O001N2O1O1O001O1N2O1O001O1N2O0O2O1O1O1O010O100O10O0100O1000O0100O100O010O10000O010O100O010O100O10O01O001O001O001O010O001O001O001O001O010O001O001O001O1O001O001N101O001O001O001O001O001O001O001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O00001O0010O010O00010O010O01O01O010O01O01O010O01O010O01O01O010O010O01O010O0O2O001O0O2O001O0O2O001O0O2O001O0O2O00001O0O2O001O0O2O001O0O2O001O0O2O00N2N2O0O2N2O1N1O2OO0M3M4M2M3M4M2M301O00101N2N1O2N2N1O2N2N101N2N2N1O2N2N2N2N101NiE`MR8^2oGfMn7X2RHmMk7Q2UHTNg7l1YHXNd7f1\\H`N`7^1`HgN]7W1cHnNY7R1gHROV7l0jHYOS7e0mH@P7>QIFl68TILi64VIOi6OWI4h6JYI8f6FZI=e6A[Ib0d6]O\\Ie0c6\\OZIg0d6\\OXIg0g6ZOWIh0h6ZOTIh0l6ZOPIi0o6YOmHj0R7WOlHk0S7b3O1O1N2O1O1O1O001O1O1N2O1O1O1O1O1O001000000O1000000O10O1000O1000000O1000O10O1000O01O00001O01O0eIbIU4]6iKeIW4[6gKhIX4X6fKjI[4U6cKnI\\4R6bKPJ^4P6`KSJ`4l5^KVJb4j5\\KYJc4g5\\KZJe4e5YK^Jf4c5WK_Ji4a5VKaJj4^5TKdJl4\\5RKgJm4Y5RKhJo4W5oJlJP5T5nJnJR5R5mJPKS5P5jJRKV5R5dJQK[5T5_JmJb5V5XJmJg5X5RJjJn5[5lIgJT6R701O010O0010O01O010O0010O01O010O0010O01O010O0010O01O010O0010O01O010O001O001O00001O001O001O001O001O001O0O2O0O2O0O2N101N101N101N101N101N101N1O2O1N4M2M4L4M2M4M3L3M4M3L3N3L4L3N3L4M2M4M3L3L5K5K4L5K5K5K7I7I7I7I7I7I7I7I7Ihg="}, "label": "ariel brown color bike standing position"}]}
{"id": 61803, "image": "COCO_train2014_000000061803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy and a girl riding on a horse\"."}], "task": "refering", "answer_template": "The \"a boy and a girl riding on a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 66, 67, 68, 80, 81, 82, 83, 84, 95, 96, 97, 98, 110, 111, 112, 113, 125, 126, 127, 139, 140, 141, 154, 155, 168, 169, 170, 183, 184, 198, 199], "bbox": [0.23342657342657344, 0.12698437499999998, 0.608904428904429, 0.6054375], "iscrowd": 0, "area": 14398.307499999999, "rle": {"size": [640, 429], "counts": "okn13ec08VOj0N2M3N2N3L3N2N2M3N2N2M3N2N2M3NO1O010O10O0100O010O10O01O010O010O010O010O0010O0100O2O1N3N1N3N1N2N3N1N2O_NfN`@W1`?mN^@R1a?RO]@l0c?YO[@e0c?@[@>e?FZ@8e?L\\@Od?6\\@Ec?`0^@\\Oa?h0_@TOc?o0]@mNd?U1\\@hNe?Z1Y@fNg?]1W@cNh?_1V@aNj?a1U@^Nk?e1R@[Nn?g1Q@XNo?k1n_OVNQ`0Q2i_OnMW`0X2c_OhM]`0^2^_OaMb`0n201O000000001O0O1000001O1O1O001O1O1N1XOi0N2O1N101O1N2O001N2O1O1N101N2O1O0010000O010O01N1O2N1O2N2N1O3M2M4M2N3M2N3M2N3M2N3M2N3L3N3M2N3M2N3M2N3M2N3M2N3N1O2N1O2N<C>C>B=C7I4L4L4LRkX3"}, "label": "a boy and a girl riding on a horse"}]}
{"id": 61803, "image": "COCO_train2014_000000061803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl sitting on the horse with red top\"."}], "task": "refering", "answer_template": "The \"girl sitting on the horse with red top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 66, 67, 68, 80, 81, 82, 83, 84, 95, 96, 97, 98, 110, 111, 112, 113, 125, 126, 127, 139, 140, 141, 154, 155, 168, 169, 170, 183, 184, 198, 199], "bbox": [0.23342657342657344, 0.12698437499999998, 0.608904428904429, 0.6054375], "iscrowd": 0, "area": 14398.307499999999, "rle": {"size": [640, 429], "counts": "okn13ec08VOj0N2M3N2N3L3N2N2M3N2N2M3N2N2M3NO1O010O10O0100O010O10O01O010O010O010O010O0010O0100O2O1N3N1N3N1N2N3N1N2O_NfN`@W1`?mN^@R1a?RO]@l0c?YO[@e0c?@[@>e?FZ@8e?L\\@Od?6\\@Ec?`0^@\\Oa?h0_@TOc?o0]@mNd?U1\\@hNe?Z1Y@fNg?]1W@cNh?_1V@aNj?a1U@^Nk?e1R@[Nn?g1Q@XNo?k1n_OVNQ`0Q2i_OnMW`0X2c_OhM]`0^2^_OaMb`0n201O000000001O0O1000001O1O1O001O1O1N1XOi0N2O1N101O1N2O001N2O1O1N101N2O1O0010000O010O01N1O2N1O2N2N1O3M2M4M2N3M2N3M2N3M2N3M2N3L3N3M2N3M2N3M2N3M2N3M2N3N1O2N1O2N<C>C>B=C7I4L4L4LRkX3"}, "label": "girl sitting on the horse with red top"}]}
{"id": 135538, "image": "COCO_train2014_000000135538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in ski gear , with a blue vest on\"."}], "task": "refering", "answer_template": "The \"a little girl in ski gear , with a blue vest on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 22, 23, 24, 25, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 73, 74, 75, 76, 90, 91, 92, 93, 94, 107, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314, 330], "bbox": [0.26845833333333335, 0.015875, 0.6502708333333334, 0.83690625], "iscrowd": 0, "area": 60896.3584, "rle": {"size": [640, 480], "counts": "Pl`29`c0<E<C<E;D=C<E<C<E<C<D=D;ZOg0J5J7J5J4L101N100O101N100O1kLjI_HW6\\7TJ^Hm5[7_J^Hb5]7iJ]HW5^7UKZHl4a7_KYHa4e7gKTH[4k7kKoGU4P8SLhGn3X8XLaGi3_8]L[Gc3d8eLTG]3k8iLoFY3o8mLjFV3S9n3N3M2M3`JPEl1S;nMREP2P;[KdEP1@c3^<XLhCf3X<WLlCh3U<TLPDi3R<SLRDl3n;QLWDm3j;oKZDo3h;mK]DQ4j=O2M2O1O2M2O2N1N3N1O2M2O2N1N3N1N3N1O2M2O2O0O1O10O0O101N1O2O0O101N101N1O101N100O2N101gMPL_DQ4`;XLXDh3g;aLPDa3n;gLkCY3S<QMdCQ3Z<XM]Ci2b<_MWCa2h<hMoBZ2n<]2L4L5M210O01O11O0OaIcC]5_=2O0O2O1N1[EjI_7W6ZHoIe7S6UHRJk7n5oGWJQ8k5hG[JX8e5bGaJ]8a5\\GdJe8]5TGiJk8X5oFnJQ9S5hFRKX9P5aFVK_9j4[F\\Kd9f4UF_Kl9b4mEdKR:]4hEiKX:o60100O00100O010O10O0100O010O10O0100O010bK]FoL0_Nd9`4bFhL0gN]9a4iF^LOPOY9`4nFWLOXOS9`4TGmK0AQ9\\4VGiKNKP9W4WGeKO2o8S4YGaKN;m8o3ZG]KOb0l8k3\\GYKMl0k8f3VIXLP7a3SI^LQ7Z3SIeLR7R3TImLP7k2UISMP7d2VI[Mn6]2VIaMP7V2VIgMP7Q2UIlMQ7k1TIRNS7f1RIWNT7`1RI]NT7[1PIcNV7U1oHhNW7R1Q6J6I7J6I7JnfX3"}, "label": "a little girl in ski gear , with a blue vest on"}]}
{"id": 135538, "image": "COCO_train2014_000000135538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child in a blue northstar jacket wearing a pair of cute little skis\"."}], "task": "refering", "answer_template": "The \"a child in a blue northstar jacket wearing a pair of cute little skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 22, 23, 24, 25, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 73, 74, 75, 76, 90, 91, 92, 93, 94, 107, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314, 330], "bbox": [0.26845833333333335, 0.015875, 0.6502708333333334, 0.83690625], "iscrowd": 0, "area": 60896.3584, "rle": {"size": [640, 480], "counts": "Pl`29`c0<E<C<E;D=C<E<C<E<C<D=D;ZOg0J5J7J5J4L101N100O101N100O1kLjI_HW6\\7TJ^Hm5[7_J^Hb5]7iJ]HW5^7UKZHl4a7_KYHa4e7gKTH[4k7kKoGU4P8SLhGn3X8XLaGi3_8]L[Gc3d8eLTG]3k8iLoFY3o8mLjFV3S9n3N3M2M3`JPEl1S;nMREP2P;[KdEP1@c3^<XLhCf3X<WLlCh3U<TLPDi3R<SLRDl3n;QLWDm3j;oKZDo3h;mK]DQ4j=O2M2O1O2M2O2N1N3N1O2M2O2N1N3N1N3N1O2M2O2O0O1O10O0O101N1O2O0O101N101N1O101N100O2N101gMPL_DQ4`;XLXDh3g;aLPDa3n;gLkCY3S<QMdCQ3Z<XM]Ci2b<_MWCa2h<hMoBZ2n<]2L4L5M210O01O11O0OaIcC]5_=2O0O2O1N1[EjI_7W6ZHoIe7S6UHRJk7n5oGWJQ8k5hG[JX8e5bGaJ]8a5\\GdJe8]5TGiJk8X5oFnJQ9S5hFRKX9P5aFVK_9j4[F\\Kd9f4UF_Kl9b4mEdKR:]4hEiKX:o60100O00100O010O10O0100O010O10O0100O010bK]FoL0_Nd9`4bFhL0gN]9a4iF^LOPOY9`4nFWLOXOS9`4TGmK0AQ9\\4VGiKNKP9W4WGeKO2o8S4YGaKN;m8o3ZG]KOb0l8k3\\GYKMl0k8f3VIXLP7a3SI^LQ7Z3SIeLR7R3TImLP7k2UISMP7d2VI[Mn6]2VIaMP7V2VIgMP7Q2UIlMQ7k1TIRNS7f1RIWNT7`1RI]NT7[1PIcNV7U1oHhNW7R1Q6J6I7J6I7JnfX3"}, "label": "a child in a blue northstar jacket wearing a pair of cute little skis"}]}
{"id": 217461, "image": "COCO_train2014_000000217461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in floral dress and black sweater sitting next to another girl on a bench\"."}], "task": "refering", "answer_template": "The \"a girl in floral dress and black sweater sitting next to another girl on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 332, 349, 350, 351, 354, 355, 372, 373, 374, 377, 378, 395, 396, 397, 400, 401, 418, 419, 420, 441, 442, 464, 465], "bbox": [0.1801875, 0.043921875, 0.49098437500000003, 0.90428125], "iscrowd": 0, "area": 51019.236650000006, "rle": {"size": [640, 640], "counts": "aQY27cc07I6J7I7J5J7N2N101N2O0O2N2O0O2O1_O?@?A`0@`0@`0@`0@?@a0@`0@`0K401O1O1O1O001O1O1O1O001O1O1O1O001O00N2N2N2N3M2O1N2N2NO1L4LG9F:E;E;H8H8H8H8L49G8H9G8H9J5O2N1O1O2N3M3M3M3M3M3M3M3M3M3M3[OlHYDW7b;mH\\DV7_;nH_DU7[;oHdDT7W;PIgDS7T;QIjDR7Q;QInDR7l:SIREo6j:UITEm6h:UIXEl6d:[1L3M3M4K4M4L3L4M4L3L5L3M4L3L4M4L3L5L300O2O001O001N10001O0O2O00001O0O2O00]NjG\\GV8d8kG\\GT8e8lGZGS8g8nGYGQ8g8PHXGP8i8PHWGo7i8RHVGm7l8SHSGm7m8THSGk7m8VHRGj7n8WHRGi1dNm3Z:[JQGf1hNm3X:^JoFc1kNo3V:_JoF_1nNQ4R:bJoF[1QOS4P:cJoFW1TOU4m9eJnFU1WOT4k9iJnFP1YOW4i9jJQGi0YO\\4f9lJTGb0XOb4d9mJWG;XOf4a9QKZG3XOk4^9SKoHn4P7SKPIo4l6SKSIo4k6RKTIQ5i6PKUIS5i6nJUIV5g6lJXIV5f6kJXIX5f6iJYIZ5c6hJ[I[5c6fJ[I^5K_Io3T1UL_5H`IR4R1TLb5F^IT4R1TLc5D^IW4P1TLe5@^I[4n0SLg5_O]I\\4n0TLg5\\O^I_4l0SLj5ZO\\Ib4k0RLl5XO\\Ie4i0RLn5TO[Ii4i0QLo5SOZIk4h0QLQ6oNZIo4f0PLS6mNZIQ5e0PLU6kNXIT5d0PLV6hNYIW5b0oKX6fNYIZ5`0nK[6dNWI\\5`0oK[6aNXI_5>nK^6^NVIa5`0PL\\6\\NVIa5a0QL]6YNUIc5c0PL[6YNUId5g0nKm7n3\\HcKn7Z4YHUKT8h4THfJY8P1TEb2W>TM_B]2d=ZMPCX2R=_MdCR2c?@?B?A?@a^[6"}, "label": "a girl in floral dress and black sweater sitting next to another girl on a bench"}]}
{"id": 217461, "image": "COCO_train2014_000000217461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in the flowered dress\"."}], "task": "refering", "answer_template": "The \"girl in the flowered dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 332, 349, 350, 351, 354, 355, 372, 373, 374, 377, 378, 395, 396, 397, 400, 401, 418, 419, 420, 441, 442, 464, 465], "bbox": [0.1801875, 0.043921875, 0.49098437500000003, 0.90428125], "iscrowd": 0, "area": 51019.236650000006, "rle": {"size": [640, 640], "counts": "aQY27cc07I6J7I7J5J7N2N101N2O0O2N2O0O2O1_O?@?A`0@`0@`0@`0@?@a0@`0@`0K401O1O1O1O001O1O1O1O001O1O1O1O001O00N2N2N2N3M2O1N2N2NO1L4LG9F:E;E;H8H8H8H8L49G8H9G8H9J5O2N1O1O2N3M3M3M3M3M3M3M3M3M3M3[OlHYDW7b;mH\\DV7_;nH_DU7[;oHdDT7W;PIgDS7T;QIjDR7Q;QInDR7l:SIREo6j:UITEm6h:UIXEl6d:[1L3M3M4K4M4L3L4M4L3L5L3M4L3L4M4L3L5L300O2O001O001N10001O0O2O00001O0O2O00]NjG\\GV8d8kG\\GT8e8lGZGS8g8nGYGQ8g8PHXGP8i8PHWGo7i8RHVGm7l8SHSGm7m8THSGk7m8VHRGj7n8WHRGi1dNm3Z:[JQGf1hNm3X:^JoFc1kNo3V:_JoF_1nNQ4R:bJoF[1QOS4P:cJoFW1TOU4m9eJnFU1WOT4k9iJnFP1YOW4i9jJQGi0YO\\4f9lJTGb0XOb4d9mJWG;XOf4a9QKZG3XOk4^9SKoHn4P7SKPIo4l6SKSIo4k6RKTIQ5i6PKUIS5i6nJUIV5g6lJXIV5f6kJXIX5f6iJYIZ5c6hJ[I[5c6fJ[I^5K_Io3T1UL_5H`IR4R1TLb5F^IT4R1TLc5D^IW4P1TLe5@^I[4n0SLg5_O]I\\4n0TLg5\\O^I_4l0SLj5ZO\\Ib4k0RLl5XO\\Ie4i0RLn5TO[Ii4i0QLo5SOZIk4h0QLQ6oNZIo4f0PLS6mNZIQ5e0PLU6kNXIT5d0PLV6hNYIW5b0oKX6fNYIZ5`0nK[6dNWI\\5`0oK[6aNXI_5>nK^6^NVIa5`0PL\\6\\NVIa5a0QL]6YNUIc5c0PL[6YNUId5g0nKm7n3\\HcKn7Z4YHUKT8h4THfJY8P1TEb2W>TM_B]2d=ZMPCX2R=_MdCR2c?@?B?A?@a^[6"}, "label": "girl in the flowered dress"}]}
{"id": 217461, "image": "COCO_train2014_000000217461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde young girl dressed in light blue is sitting on a bench\"."}], "task": "refering", "answer_template": "The \"a blonde young girl dressed in light blue is sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 57, 58, 59, 80, 81, 82, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 287, 288, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 338, 339, 357, 358, 359, 361, 362, 380, 381, 382, 384, 385, 404, 405, 407, 408, 409, 430, 431, 432, 433], "bbox": [0.47415624999999995, 0.06928125, 0.858421875, 0.82659375], "iscrowd": 0, "area": 47631.11945, "rle": {"size": [640, 640], "counts": "`am51ic0?@`0A?A>B?XMeMSCj2c<ZMTCU3c<mLYC]3^<gL_Cb3X<aLdCi3R<[LjCo3m;TLoCU4h;oKUDZ4b;hK[Db4[;bKaDh4V;[KfDo4P;UKmDU5i:nJSE]5b:gJZEa5_:aJ^Ea5`:cJ]E]5c:fJYE\\5e:hJWEY5i:jJSEX5k:P2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N100O_NnESIU:n6TFaHR:`7XFoGo9Q8o020O100O100O1O100OeGYI^3g6_L\\Ia3c6]L_Id3a6YLbIg3]6WLeIj3Z6TLiIl3W6QLlIo3S6oKoIR4Q6kKRJU4m5iKVJW4i5gKYJZ4g5cK\\J]4c5aK`J_4`5^KbJc4]5[KfJe4Z5XKiJh4V5VKlJk4S5RKQKn4o4oJSKR5l4lJWKT5i4iJZKW5e4gJ]KZ5c4cJ`Ka5[4]JhKh5R4VJPLo5l3nIWLW6c3gI`L]6\\3`IfLe6U3YInLl6m2QIUMQ7i2mHZMR7f2lH\\MU7d2hH]MZ7b2dHPL[OcNQ8^5aHkKFdNk7`5]HgK1fNc7d5YHbK=fN\\7g5UHaKe0fNW7h5RHbKl0bNS7m5nGbKR1^NR7o5jGcKY1ZNn6T6fGbK`1WNl6V6bGdKf1SNi6X6_GeKl1PNf6\\6[GdKS2mMd6^6WGfKT;a62N3N1N3M2O2M2N3N1N3M2O2M2N3N1N2N3N1N3M5L4K6J6K5J6J6K5J6J6K4K6J6K5001O001O1N2O1O1O001O1O1O1N101O1O1QD[J]9f5ZF_Ji9b5oEaJU:_5cEeJa:\\5VEhJm:Z5kDiJY;X5^DlJf;U5QDPKR<P5gCSK]<n4ZCVKj<n54L4L4M2M4L4L3M2N2O2M2N2N3M2N2O2M2N2N2N3Mi0XO:E;E<D;E;E;F:E<D;E;E;E2OO01O001O0010O01O001O0010O01O001O00010O001O001O010O001O001O010O001O000010O01O001O0010O01O001O0010O01Ol[h1"}, "label": "a blonde young girl dressed in light blue is sitting on a bench"}]}
{"id": 217461, "image": "COCO_train2014_000000217461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in white on the right\"."}], "task": "refering", "answer_template": "The \"the girl in white on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 57, 58, 59, 80, 81, 82, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 287, 288, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 338, 339, 357, 358, 359, 361, 362, 380, 381, 382, 384, 385, 404, 405, 407, 408, 409, 430, 431, 432, 433], "bbox": [0.47415624999999995, 0.06928125, 0.858421875, 0.82659375], "iscrowd": 0, "area": 47631.11945, "rle": {"size": [640, 640], "counts": "`am51ic0?@`0A?A>B?XMeMSCj2c<ZMTCU3c<mLYC]3^<gL_Cb3X<aLdCi3R<[LjCo3m;TLoCU4h;oKUDZ4b;hK[Db4[;bKaDh4V;[KfDo4P;UKmDU5i:nJSE]5b:gJZEa5_:aJ^Ea5`:cJ]E]5c:fJYE\\5e:hJWEY5i:jJSEX5k:P2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N100O_NnESIU:n6TFaHR:`7XFoGo9Q8o020O100O100O1O100OeGYI^3g6_L\\Ia3c6]L_Id3a6YLbIg3]6WLeIj3Z6TLiIl3W6QLlIo3S6oKoIR4Q6kKRJU4m5iKVJW4i5gKYJZ4g5cK\\J]4c5aK`J_4`5^KbJc4]5[KfJe4Z5XKiJh4V5VKlJk4S5RKQKn4o4oJSKR5l4lJWKT5i4iJZKW5e4gJ]KZ5c4cJ`Ka5[4]JhKh5R4VJPLo5l3nIWLW6c3gI`L]6\\3`IfLe6U3YInLl6m2QIUMQ7i2mHZMR7f2lH\\MU7d2hH]MZ7b2dHPL[OcNQ8^5aHkKFdNk7`5]HgK1fNc7d5YHbK=fN\\7g5UHaKe0fNW7h5RHbKl0bNS7m5nGbKR1^NR7o5jGcKY1ZNn6T6fGbK`1WNl6V6bGdKf1SNi6X6_GeKl1PNf6\\6[GdKS2mMd6^6WGfKT;a62N3N1N3M2O2M2N3N1N3M2O2M2N3N1N2N3N1N3M5L4K6J6K5J6J6K5J6J6K4K6J6K5001O001O1N2O1O1O001O1O1O1N101O1O1QD[J]9f5ZF_Ji9b5oEaJU:_5cEeJa:\\5VEhJm:Z5kDiJY;X5^DlJf;U5QDPKR<P5gCSK]<n4ZCVKj<n54L4L4M2M4L4L3M2N2O2M2N2N3M2N2O2M2N2N2N3Mi0XO:E;E<D;E;E;F:E<D;E;E;E2OO01O001O0010O01O001O0010O01O001O00010O001O001O010O001O001O010O001O000010O01O001O0010O01O001O0010O01Ol[h1"}, "label": "the girl in white on the right"}]}
{"id": 242039, "image": "COCO_train2014_000000242039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tail wing of a plane in front of the united plane ' s tail\"."}], "task": "refering", "answer_template": "The \"the tail wing of a plane in front of the united plane ' s tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161], "bbox": [0.0, 0.374519906323185, 0.376125, 0.514543325526932], "iscrowd": 0, "area": 3612.7789500000003, "rle": {"size": [427, 640], "counts": "P5k1a;N1N2M4L3N2M4M20001N101O0O101N101O0O2O0O2O001N10001N101N101O0O2O001O00001O001O001O001O00000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O00000000000001O00000000000000000000000000000000000010O0000000000000000000000000000000001O000000000000000001O0000000000000001O000000000000000O10000000000O10000000000O10000000000O101O0000000O1000001O000O1000000000dcV5"}, "label": "the tail wing of a plane in front of the united plane ' s tail"}]}
{"id": 242039, "image": "COCO_train2014_000000242039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the intersection of the two planes\"."}], "task": "refering", "answer_template": "The \"the intersection of the two planes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161], "bbox": [0.0, 0.374519906323185, 0.376125, 0.514543325526932], "iscrowd": 0, "area": 3612.7789500000003, "rle": {"size": [427, 640], "counts": "P5k1a;N1N2M4L3N2M4M20001N101O0O101N101O0O2O0O2O001N10001N101N101O0O2O001O00001O001O001O001O00000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O00000000000001O00000000000000000000000000000000000010O0000000000000000000000000000000001O000000000000000001O0000000000000001O000000000000000O10000000000O10000000000O10000000000O101O0000000O1000001O000O1000000000dcV5"}, "label": "the intersection of the two planes"}]}
{"id": 242039, "image": "COCO_train2014_000000242039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the united aircraft with the open cargo area and baggage carts around it\"."}], "task": "refering", "answer_template": "The \"the united aircraft with the open cargo area and baggage carts around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 117, 118, 140, 141, 142, 164, 165, 166, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.01196875, 0.3286885245901639, 0.966375, 0.7390163934426229], "iscrowd": 0, "area": 34763.60820000003, "rle": {"size": [427, 640], "counts": "`b31Y=1O10001O000000001O000000001O0000001O000001O01O000000001O000000001O000000001O0000001O01O000001O000000001O0dLEhI;X6HeI8[6KbI5^6N_I3`60]I0c63ZIMf66WIJi69TIGl6<QIEn6>oHBQ7a0lH_OT7d0jH[OV7h0gHYOX7j0dHWO\\7j0cHVO]7j0bHWO^7i0bHXO]7h0nFjNU1?l7g0oFmNQ1=P8f0oFPOn0;R8e0PGSOj0:U8c0QGVOg07X8c0QGYOd05Z8b0RG\\O`03^8a0RG_O=1`8`0SGB90d8=SGF6Mh8<RGJ3Jm8:PG0NGS98oF4KEV96oF8GC\\93mF=D@`92lF]1U9bNkF^1W9`NiFa1W9^NiFb1X9]NhFc1Z9[NfFe1[9ZNeFg1[9XNeFh1]9VNcFj1^9UNbFk1_9TNaFl1a9RN^FP2b9oM^FQ2c9nM]FR2e9lM[FT2f9kMZFV2f9iMZFW2h9gMXFY2i9fMWFZ2j9eMVF[2l9cMTF^2l9bMSF^2n9aMRF_2P:_MPFa2Q:^MoEc2\\:000000001O0000001O0000000010O2N1O100O2N1O1O101N1O1O100O2N1O1O101N1O1O101N1O1O1O101N1O0001O0000010O000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000001O00000000000000001O00001O001O001O001O001O001O001O001O0000001O0000000000000000000010O000000000001O3M2N2N2N3M000000000000O1LXOeMhFX2[9jMcFR2a9PN]Fm1f9UNXFk1h9WNVFi1j9ZNSFf1n9[NPFe1P:]NnEc1R:_NlEa1T:aNjE_1V:cNhE]1X:eNfE[1[:fNcEZ1]:j000000000001O00000000001O2N1N3M3L4M2N3M3N2N2N1O2N00000000000000000000001O000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000001O00000O100000000000000000000000000O100000001O0000000000000000000O1000000000000000000000001O000O100000000000000000000000000O10000000001O00000000000000000O1000000000001O1O001O1O1O001N2O1O001O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1O001O1O1N101O1O1O001O1O1M2N3M3MUm8"}, "label": "the united aircraft with the open cargo area and baggage carts around it"}]}
{"id": 242039, "image": "COCO_train2014_000000242039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flight with ' united ' written on it\"."}], "task": "refering", "answer_template": "The \"flight with ' united ' written on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 117, 118, 140, 141, 142, 164, 165, 166, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.01196875, 0.3286885245901639, 0.966375, 0.7390163934426229], "iscrowd": 0, "area": 34763.60820000003, "rle": {"size": [427, 640], "counts": "`b31Y=1O10001O000000001O000000001O0000001O000001O01O000000001O000000001O000000001O0000001O01O000001O000000001O0dLEhI;X6HeI8[6KbI5^6N_I3`60]I0c63ZIMf66WIJi69TIGl6<QIEn6>oHBQ7a0lH_OT7d0jH[OV7h0gHYOX7j0dHWO\\7j0cHVO]7j0bHWO^7i0bHXO]7h0nFjNU1?l7g0oFmNQ1=P8f0oFPOn0;R8e0PGSOj0:U8c0QGVOg07X8c0QGYOd05Z8b0RG\\O`03^8a0RG_O=1`8`0SGB90d8=SGF6Mh8<RGJ3Jm8:PG0NGS98oF4KEV96oF8GC\\93mF=D@`92lF]1U9bNkF^1W9`NiFa1W9^NiFb1X9]NhFc1Z9[NfFe1[9ZNeFg1[9XNeFh1]9VNcFj1^9UNbFk1_9TNaFl1a9RN^FP2b9oM^FQ2c9nM]FR2e9lM[FT2f9kMZFV2f9iMZFW2h9gMXFY2i9fMWFZ2j9eMVF[2l9cMTF^2l9bMSF^2n9aMRF_2P:_MPFa2Q:^MoEc2\\:000000001O0000001O0000000010O2N1O100O2N1O1O101N1O1O100O2N1O1O101N1O1O101N1O1O1O101N1O0001O0000010O000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000001O00000000000000001O00001O001O001O001O001O001O001O001O0000001O0000000000000000000010O000000000001O3M2N2N2N3M000000000000O1LXOeMhFX2[9jMcFR2a9PN]Fm1f9UNXFk1h9WNVFi1j9ZNSFf1n9[NPFe1P:]NnEc1R:_NlEa1T:aNjE_1V:cNhE]1X:eNfE[1[:fNcEZ1]:j000000000001O00000000001O2N1N3M3L4M2N3M3N2N2N1O2N00000000000000000000001O000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000001O00000O100000000000000000000000000O100000001O0000000000000000000O1000000000000000000000001O000O100000000000000000000000000O10000000001O00000000000000000O1000000000001O1O001O1O1O001N2O1O001O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1O001O1O1N101O1O1O001O1O1M2N3M3MUm8"}, "label": "flight with ' united ' written on it"}]}
{"id": 397687, "image": "COCO_train2014_000000397687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big book - like chair\"."}], "task": "refering", "answer_template": "The \"a big book - like chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267], "bbox": [0.338890625, 0.22381250000000003, 0.711, 0.6895208333333334], "iscrowd": 0, "area": 23844.837749999995, "rle": {"size": [480, 640], "counts": "XoU33h>5L4J5N3M3M3I7L41O001O00001N10001O00001O001O00001N101O000005K100O2N1O1O1O001O10O01O001ON2N2O1N2O1N2N101O10000O10O1000O10000O01000O10000O10000O10O2O000O100O110O00000SOBXC>d<3PCMk<V1L4L4M3L5M2M3N2M3N3L3N2M3N2H9F9O100O1O2O0O1O100O2N1_FnKT8S4kGmKT8U4jGlKU8V4jGjKV8X4gGiKW8Z4gGgKY8Z4fGfKY8]4dGdK\\8]4cGcK]8^4aGcK^8`4`G`K`8a4^G`Ka8c4\\G^Kd8c4[G]Ke8d4YG]Kf8f4XG[Kg8f4WG[Kh8g4WGYKi8i4TGXKl8i4SGWKl8k4RGVKn8l4oFUKP9m4oFSKQ9]510O01O010O1O0101N2O1O2M2N2O1O1N3N1O1O1N3N1O1O1N10001O0O10000O10001O0O1000000O2O000O10N2O001N101O0O2ZKfGe2[8RMQHj2o7mL]Ho2f7nL_Hn2b7PMbHm2_7QMfHk2[7SMiHi2Y7UMlHg2V7VMnHg2S7WMQIe2Q7YMRIe2o6YMTIe2m6YMVId2l6[MVIc2l6[MVIc2k6]MVI`2l6_MVI_2k6`MWI^2k6`MWI^2j6bMWI[2l6cMVI[2k6dMWIZ2k6dMWIY2k6gMVIW2k6hMWIV2k6hMXIT2j6kMXIS2j6lMWIQ2l6mMVIQ2l6mMVIQ2n9M3N2N2M3N2N2M3N2M3N2N2M3N3M2M3N2N2M3N2M3N2N2L4L4Lb_f2"}, "label": "a big book - like chair"}]}
{"id": 397687, "image": "COCO_train2014_000000397687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair that is shaped like an open book\"."}], "task": "refering", "answer_template": "The \"a chair that is shaped like an open book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267], "bbox": [0.338890625, 0.22381250000000003, 0.711, 0.6895208333333334], "iscrowd": 0, "area": 23844.837749999995, "rle": {"size": [480, 640], "counts": "XoU33h>5L4J5N3M3M3I7L41O001O00001N10001O00001O001O00001N101O000005K100O2N1O1O1O001O10O01O001ON2N2O1N2O1N2N101O10000O10O1000O10000O01000O10000O10000O10O2O000O100O110O00000SOBXC>d<3PCMk<V1L4L4M3L5M2M3N2M3N3L3N2M3N2H9F9O100O1O2O0O1O100O2N1_FnKT8S4kGmKT8U4jGlKU8V4jGjKV8X4gGiKW8Z4gGgKY8Z4fGfKY8]4dGdK\\8]4cGcK]8^4aGcK^8`4`G`K`8a4^G`Ka8c4\\G^Kd8c4[G]Ke8d4YG]Kf8f4XG[Kg8f4WG[Kh8g4WGYKi8i4TGXKl8i4SGWKl8k4RGVKn8l4oFUKP9m4oFSKQ9]510O01O010O1O0101N2O1O2M2N2O1O1N3N1O1O1N3N1O1O1N10001O0O10000O10001O0O1000000O2O000O10N2O001N101O0O2ZKfGe2[8RMQHj2o7mL]Ho2f7nL_Hn2b7PMbHm2_7QMfHk2[7SMiHi2Y7UMlHg2V7VMnHg2S7WMQIe2Q7YMRIe2o6YMTIe2m6YMVId2l6[MVIc2l6[MVIc2k6]MVI`2l6_MVI_2k6`MWI^2k6`MWI^2j6bMWI[2l6cMVI[2k6dMWIZ2k6dMWIY2k6gMVIW2k6hMWIV2k6hMXIT2j6kMXIS2j6lMWIQ2l6mMVIQ2l6mMVIQ2n9M3N2N2M3N2N2M3N2M3N2N2M3N3M2M3N2N2M3N2M3N2N2L4L4Lb_f2"}, "label": "a chair that is shaped like an open book"}]}
{"id": 397687, "image": "COCO_train2014_000000397687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with the halter tie multi - colored dress\"."}], "task": "refering", "answer_template": "The \"the woman with the halter tie multi - colored dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 77, 78, 100, 101, 102, 123, 124, 125, 126, 146, 147, 148, 149, 168, 169, 170, 171, 191, 192, 193, 194, 213, 214], "bbox": [0.27364062499999997, 0.15964583333333332, 0.487125, 0.565875], "iscrowd": 0, "area": 11276.6383, "rle": {"size": [480, 640], "counts": "ZYb23l>4L010O1O1O0000000001O0001O0000000001O01O1O2N1O2N1O101L3M4L3M4L3M4K4M4L3M4M01O1O1O1O1O1O1O1N2O1O1O1O3OVOlNhCT1V<nNjCQ1U<ROjCn0S<VOlCj0R<XOoCg0n;]OQDc0m;@RD`0l;BTD>i;FWD9g;JXD5k9\\OnFg0ROMo9FdFc0YOGQ:N^Fb0]O_OT:0\\Fh0\\OXOW:1[Fm1d9TNZFn1e9TNYFm1f9TNYFm1f9TNZFl1e9UNZFm1d9TN[Fm1d9TN[Fm1d9UNZFl1f9TNZFl1f9TNYFn1f9RNYFo1h9PNWFQ2i9PNUFQ2k9oMTFR2m9mMSFS2m9mMRFU2m9kMRFV2o9]10000010O00001O01O01O0000001iKQF[3o9cLTF\\3m9bLVF\\3l9`LWF^3l9_LVF`3l9]LUFc3m9YLVFf3l9WLUFi3a:01N101O00001N10001O000O2O001O002M2O2N2B\\D\\Me;^2bD`M`;Y2hDcMZ;W2mDgMU;R2Q1H7A`0M2M4L?Aa0@ebi4"}, "label": "the woman with the halter tie multi - colored dress"}]}
{"id": 397687, "image": "COCO_train2014_000000397687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a maxi dress and a tan shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a maxi dress and a tan shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 77, 78, 100, 101, 102, 123, 124, 125, 126, 146, 147, 148, 149, 168, 169, 170, 171, 191, 192, 193, 194, 213, 214], "bbox": [0.27364062499999997, 0.15964583333333332, 0.487125, 0.565875], "iscrowd": 0, "area": 11276.6383, "rle": {"size": [480, 640], "counts": "ZYb23l>4L010O1O1O0000000001O0001O0000000001O01O1O2N1O2N1O101L3M4L3M4L3M4K4M4L3M4M01O1O1O1O1O1O1O1N2O1O1O1O3OVOlNhCT1V<nNjCQ1U<ROjCn0S<VOlCj0R<XOoCg0n;]OQDc0m;@RD`0l;BTD>i;FWD9g;JXD5k9\\OnFg0ROMo9FdFc0YOGQ:N^Fb0]O_OT:0\\Fh0\\OXOW:1[Fm1d9TNZFn1e9TNYFm1f9TNYFm1f9TNZFl1e9UNZFm1d9TN[Fm1d9TN[Fm1d9UNZFl1f9TNZFl1f9TNYFn1f9RNYFo1h9PNWFQ2i9PNUFQ2k9oMTFR2m9mMSFS2m9mMRFU2m9kMRFV2o9]10000010O00001O01O01O0000001iKQF[3o9cLTF\\3m9bLVF\\3l9`LWF^3l9_LVF`3l9]LUFc3m9YLVFf3l9WLUFi3a:01N101O00001N10001O000O2O001O002M2O2N2B\\D\\Me;^2bD`M`;Y2hDcMZ;W2mDgMU;R2Q1H7A`0M2M4L?Aa0@ebi4"}, "label": "a woman wearing a maxi dress and a tan shirt"}]}
{"id": 12664, "image": "COCO_train2014_000000012664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is train on the tracks\"."}], "task": "refering", "answer_template": "The \"there is train on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226], "bbox": [0.28565625, 0.43274038461538467, 0.8569531250000001, 0.6524759615384615], "iscrowd": 0, "area": 13556.616000000004, "rle": {"size": [416, 640], "counts": "d`Z2e0[<9G01O0001O00000001O01O00000000010O00000000001O01O00000001O0001O000001O000001O0001O00000001O01O00000000010O000000000010O0000000001O01O00000001O0001O000001O000001O0001O000000000010O0000000001O00000000001O000001O0001O00000000001O00000000001O01O00000001O00000000001O00000001O01O00000000001O00000000001O01O00000001O000000000000001O0001O000000000001O01O000000001O0001O0001O0000000010O00000001O000000010O000000001O000001O01O000001O01O0001O0001O01O00000010O000001O0001O01O00000010O000001O0001O0001O0000010O00001O0001O01O00001O01O0001O000010O000001O00010O0000001O01O0001O000010O000001O00010O0000001O01O01O0000010O000010O000001O01O01O0001O01O0001O0000001O0000001O0000001N10000O1O10O01000O10000O10001N10000O10000O2O000O100O1002O5J7I7I6J7IPTU1"}, "label": "there is train on the tracks"}]}
{"id": 12664, "image": "COCO_train2014_000000012664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first three train cars\"."}], "task": "refering", "answer_template": "The \"first three train cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226], "bbox": [0.28565625, 0.43274038461538467, 0.8569531250000001, 0.6524759615384615], "iscrowd": 0, "area": 13556.616000000004, "rle": {"size": [416, 640], "counts": "d`Z2e0[<9G01O0001O00000001O01O00000000010O00000000001O01O00000001O0001O000001O000001O0001O00000001O01O00000000010O000000000010O0000000001O01O00000001O0001O000001O000001O0001O000000000010O0000000001O00000000001O000001O0001O00000000001O00000000001O01O00000001O00000000001O00000001O01O00000000001O00000000001O01O00000001O000000000000001O0001O000000000001O01O000000001O0001O0001O0000000010O00000001O000000010O000000001O000001O01O000001O01O0001O0001O01O00000010O000001O0001O01O00000010O000001O0001O0001O0000010O00001O0001O01O00001O01O0001O000010O000001O00010O0000001O01O0001O000010O000001O00010O0000001O01O01O0000010O000010O000001O01O01O0001O01O0001O0000001O0000001O0000001N10000O1O10O01000O10000O10001N10000O10000O2O000O100O1002O5J7I7I6J7IPTU1"}, "label": "first three train cars"}]}
{"id": 61819, "image": "COCO_train2014_000000061819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car with fuel on it\"."}], "task": "refering", "answer_template": "The \"a car with fuel on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229], "bbox": [0.760765625, 0.5413521126760563, 0.995703125, 0.7565633802816902], "iscrowd": 0, "area": 8101.233999999999, "rle": {"size": [355, 640], "counts": "_QY54n:2N2N2N1O2N2_EBZ:`0eEBW:`0hEBV:`0iEA63_9=[FA35U9o0hFUOU9^1N2N3N00000000000001O01O0000O100O100F:0000000000000000000000001O0001O000000000000000000000000lFZNk8f1nFbNQ9h10001O00001O1O1O1O10O01O00000000000000000000M3M3M3N2M31O1O1O1O1O00000000000001O00000001O000000000001O000000000000000010O0000bNPGj0o8QOXGn0g8PO[GQ1d8nN]GS1b8kN`GU1a8iN`GW1`8gNbGY1U9O01O0000010O0001O001O0GVFYOk9MXF=0Fh9I]F<NLo93RFMP:0QF1o9MRF3o9KQF7o9GRF9o9ERF<Z:0Zi0"}, "label": "a car with fuel on it"}]}
{"id": 61819, "image": "COCO_train2014_000000061819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"van to the right\"."}], "task": "refering", "answer_template": "The \"van to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229], "bbox": [0.760765625, 0.5413521126760563, 0.995703125, 0.7565633802816902], "iscrowd": 0, "area": 8101.233999999999, "rle": {"size": [355, 640], "counts": "_QY54n:2N2N2N1O2N2_EBZ:`0eEBW:`0hEBV:`0iEA63_9=[FA35U9o0hFUOU9^1N2N3N00000000000001O01O0000O100O100F:0000000000000000000000001O0001O000000000000000000000000lFZNk8f1nFbNQ9h10001O00001O1O1O1O10O01O00000000000000000000M3M3M3N2M31O1O1O1O1O00000000000001O00000001O000000000001O000000000000000010O0000bNPGj0o8QOXGn0g8PO[GQ1d8nN]GS1b8kN`GU1a8iN`GW1`8gNbGY1U9O01O0000010O0001O001O0GVFYOk9MXF=0Fh9I]F<NLo93RFMP:0QF1o9MRF3o9KQF7o9GRF9o9ERF<Z:0Zi0"}, "label": "van to the right"}]}
{"id": 512380, "image": "COCO_train2014_000000512380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white helmet playing baseball\"."}], "task": "refering", "answer_template": "The \"a man in white helmet playing baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 143, 144, 166, 167, 189, 190, 191, 211, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283], "bbox": [0.21134375, 0.3348333333333333, 0.338703125, 0.7403541666666666], "iscrowd": 0, "area": 8008.586349999998, "rle": {"size": [480, 640], "counts": "_ao11j>5L5J5N2N2O2K4E;L2O1O1O1mCmNZ:T1aE[OR:h0iEGg9c0QFJ`9`0WFO`94WF3i9b2O1O101N100O1O100O2N1WOUK]Gd33gL^8KZG^3:fL[80WGZ3`0eLY84TGR3i0iLS8k3THTLk7\\3XGWLV1:c7`3gH_LY7a3hH]LY7c3hH\\LX7e3hHYLY7g3gHYLY7h3bGTL>3o7k3`GVL`0MQ8m3]G[L?HT8n3YG`La0@W8Q4UGcLg9^3VFgLg9Z3UFlLh9X4M4MjK]Fe02?^9jNbFf02a0Z9gNgF;LkN7W2P9aNoF>LfN8]2j8\\NUG`0:V1_8VNZGd09V1\\8SN^Gf07W1[8gMjGS1JU1e9jN\\FT1W:ZNiEf1X:YNhEf1Z:YNgEe1\\:YNdEf1^:YNdEd1_:[NbEb1`:]NaEa1a:^N_F`0c9@eF7\\9IlFNU92kFMV93kFJW97hFHY98fFH[98eFG\\99dFE_9:aFEa98`FHb94aFKa92aFLb90aFOa9NaF1W<0S]V6"}, "label": "a man in white helmet playing baseball"}]}
{"id": 512380, "image": "COCO_train2014_000000512380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red baseball unpire uniform\"."}], "task": "refering", "answer_template": "The \"a man in a red baseball unpire uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 143, 144, 166, 167, 189, 190, 191, 211, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283], "bbox": [0.21134375, 0.3348333333333333, 0.338703125, 0.7403541666666666], "iscrowd": 0, "area": 8008.586349999998, "rle": {"size": [480, 640], "counts": "_ao11j>5L5J5N2N2O2K4E;L2O1O1O1mCmNZ:T1aE[OR:h0iEGg9c0QFJ`9`0WFO`94WF3i9b2O1O101N100O1O100O2N1WOUK]Gd33gL^8KZG^3:fL[80WGZ3`0eLY84TGR3i0iLS8k3THTLk7\\3XGWLV1:c7`3gH_LY7a3hH]LY7c3hH\\LX7e3hHYLY7g3gHYLY7h3bGTL>3o7k3`GVL`0MQ8m3]G[L?HT8n3YG`La0@W8Q4UGcLg9^3VFgLg9Z3UFlLh9X4M4MjK]Fe02?^9jNbFf02a0Z9gNgF;LkN7W2P9aNoF>LfN8]2j8\\NUG`0:V1_8VNZGd09V1\\8SN^Gf07W1[8gMjGS1JU1e9jN\\FT1W:ZNiEf1X:YNhEf1Z:YNgEe1\\:YNdEf1^:YNdEd1_:[NbEb1`:]NaEa1a:^N_F`0c9@eF7\\9IlFNU92kFMV93kFJW97hFHY98fFH[98eFG\\99dFE_9:aFEa98`FHb94aFKa92aFLb90aFOa9NaF1W<0S]V6"}, "label": "a man in a red baseball unpire uniform"}]}
{"id": 487806, "image": "COCO_train2014_000000487806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy wearing a wide brimmed blue hat\"."}], "task": "refering", "answer_template": "The \"a small boy wearing a wide brimmed blue hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 212, 213, 214, 215, 216, 230, 231, 232, 247, 248, 249, 250, 264, 265, 266, 267, 281, 282, 283, 298, 299, 300, 315, 316, 317], "bbox": [0.4983333333333333, 0.47371875, 0.730625, 0.8185625], "iscrowd": 0, "area": 12817.91575, "rle": {"size": [640, 480], "counts": "Qge42jc05J5@HT]O=kb0=N1O2N2O0O2O1O001O10O0\\OY]O0^4^Og9a0nA0Y4Bg9=SBOU4Gg97WB1Q4Kh91YB4h2CnM;P=L[B6d2FPN:Q=D`B<]2HQN9h=OTDLSN5i=0PDNVN3j=OiC5\\NMk=NcC;aNHl=M\\Cb0dNEP>JUCg0hNBS>GnBn0kN_OW>CgBV1nNZO[>@`Bc1kNPOf>]OXBQ3i=PMoAV3R>jLhAZ3X=WLSC?@^3[=YLRC9^Oa3_=[LQC5YOd3h=ZLlB[4U=fKjBX4X=hKhBV4Z=kKeBT4\\=lKdBR4]=PLbBm3a=SL_Bj3d=WL[Be3i=[LWBb3l=_LSB]3Q>cLoAY3T>iLkAT3X>lLhAP3\\>QMcAl2a>SM_Ai2e>XMYAf2k>YMl@S3Q?Q1N2M2N3N2L4hAnJ[=U5`BSKY=R5aBVKY=j5N2M2O1O2N1010O0001O00001OO2O0O1ROTCgJm<T5bCaJ_<n2iBTO^>m0gAmM2_O\\>d2gAdMY?\\2i@^MZ?c2i@TM]?k2g@kL`?[1P@9m0^Nc>V1d@;g0aNf>S1d@<b0eNi>o0f@?;eNo>k0g@c02fNW?g0h@f0JgN^?b0i@b1W?\\Nk@c1W?[Nk@b1X?\\Nk@^1Y?aNj@Z1Z?dNi@V1e?_N^@[1f?cN]@Y1e?gN^@S1f?lN]@n0f?RO]@h0g?WO]@c0f?\\O]@>f?A^@9f?FZ201O004L4KYY`2"}, "label": "a small boy wearing a wide brimmed blue hat"}]}
{"id": 487806, "image": "COCO_train2014_000000487806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child standing with father\"."}], "task": "refering", "answer_template": "The \"a child standing with father\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 212, 213, 214, 215, 216, 230, 231, 232, 247, 248, 249, 250, 264, 265, 266, 267, 281, 282, 283, 298, 299, 300, 315, 316, 317], "bbox": [0.4983333333333333, 0.47371875, 0.730625, 0.8185625], "iscrowd": 0, "area": 12817.91575, "rle": {"size": [640, 480], "counts": "Qge42jc05J5@HT]O=kb0=N1O2N2O0O2O1O001O10O0\\OY]O0^4^Og9a0nA0Y4Bg9=SBOU4Gg97WB1Q4Kh91YB4h2CnM;P=L[B6d2FPN:Q=D`B<]2HQN9h=OTDLSN5i=0PDNVN3j=OiC5\\NMk=NcC;aNHl=M\\Cb0dNEP>JUCg0hNBS>GnBn0kN_OW>CgBV1nNZO[>@`Bc1kNPOf>]OXBQ3i=PMoAV3R>jLhAZ3X=WLSC?@^3[=YLRC9^Oa3_=[LQC5YOd3h=ZLlB[4U=fKjBX4X=hKhBV4Z=kKeBT4\\=lKdBR4]=PLbBm3a=SL_Bj3d=WL[Be3i=[LWBb3l=_LSB]3Q>cLoAY3T>iLkAT3X>lLhAP3\\>QMcAl2a>SM_Ai2e>XMYAf2k>YMl@S3Q?Q1N2M2N3N2L4hAnJ[=U5`BSKY=R5aBVKY=j5N2M2O1O2N1010O0001O00001OO2O0O1ROTCgJm<T5bCaJ_<n2iBTO^>m0gAmM2_O\\>d2gAdMY?\\2i@^MZ?c2i@TM]?k2g@kL`?[1P@9m0^Nc>V1d@;g0aNf>S1d@<b0eNi>o0f@?;eNo>k0g@c02fNW?g0h@f0JgN^?b0i@b1W?\\Nk@c1W?[Nk@b1X?\\Nk@^1Y?aNj@Z1Z?dNi@V1e?_N^@[1f?cN]@Y1e?gN^@S1f?lN]@n0f?RO]@h0g?WO]@c0f?\\O]@>f?A^@9f?FZ201O004L4KYY`2"}, "label": "a child standing with father"}]}
{"id": 487806, "image": "COCO_train2014_000000487806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man standing who is wearing a hat and sunglasses\"."}], "task": "refering", "answer_template": "The \"the man standing who is wearing a hat and sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 39, 40, 41, 42, 56, 57, 58, 59, 73, 74, 75, 76, 89, 90, 91, 92, 93, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 313, 327, 328, 344, 345], "bbox": [0.2426875, 0.065171875, 0.5423125, 0.9146093750000001], "iscrowd": 0, "area": 48772.79655000001, "rle": {"size": [640, 480], "counts": "QgX22cc0k0UOj0VOj0WOj0ZOe0^Ob0_Ob0]Ob0^Oc0]Ob0^O>B=kEiHj6d7THYIh0eNf4`8_IhI_1QNn4Q;fJXEU5Q;^JWE^5Q;VJXEg5n:nIZEo5]<N101N1O100O2O0O1O101N100WNnAcMS>[2PBcMR>Z2RBdMn=[2UBbMl=\\2WBbMj=\\2ZBbMc=_2aB^M\\=e2fBZMW=g2mBVMP=l2TCQMj<Q3XCnLe<S3_CjL^<X3fCfLV<]3lCaLQ<c3QDZLk;k3WDSLe;P4]DnK`;V4cDgKX;]4TEWKi:m4Q24ZMe2O2M2O^ITDT5j;bJcD\\5\\;YJREf5l:TJ\\El5b:RJcEl5[:SJjEl5T:RJQFl5n9RJWFm5g9QJ]Fo5a9oIdFo5Z9PJkFo5S9oIRGo5m8oIXGP6];O0O2O1O1O1O001O1N2O1O1O001O1N2O1O001O00kMiJaEW5o9bKhE_4U:lKbEU4\\:SL]Em3c:UL[El3c:XLZEh3f:ZLWEh3g:\\LVEd3j:^LTEc3j:`LTE`3l:cLQE^3m:eLQE[3o:hLnDY3P;jLmDX3R;kLkDU3T;nLjDS3S;RMjDn2S;R3L3M4L3M4ZGiF^6Z9aIjFY6Z9eIjFU6Z9iIkFP6X9oIlFd4^OfJj9d0mFa4AaJe9m0nF]4E]Ja9T1PGY4GYJ\\9]1UGR4o9lKXFm3k9PL]Fh3h9SL_Ff3e9VLcFa3b9[LeF^3`9]LhF[3\\9aLkFX3Y9dLoFT3V9gLRGP3S9lLTGm2P9oLXGi2m8RMZGh2V=I6J7I7I6J7I6J7I6J7I6J7I7IRVY4"}, "label": "the man standing who is wearing a hat and sunglasses"}]}
{"id": 487806, "image": "COCO_train2014_000000487806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown wide brimmed hat\"."}], "task": "refering", "answer_template": "The \"a man in a brown wide brimmed hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 39, 40, 41, 42, 56, 57, 58, 59, 73, 74, 75, 76, 89, 90, 91, 92, 93, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 313, 327, 328, 344, 345], "bbox": [0.2426875, 0.065171875, 0.5423125, 0.9146093750000001], "iscrowd": 0, "area": 48772.79655000001, "rle": {"size": [640, 480], "counts": "QgX22cc0k0UOj0VOj0WOj0ZOe0^Ob0_Ob0]Ob0^Oc0]Ob0^O>B=kEiHj6d7THYIh0eNf4`8_IhI_1QNn4Q;fJXEU5Q;^JWE^5Q;VJXEg5n:nIZEo5]<N101N1O100O2O0O1O101N100WNnAcMS>[2PBcMR>Z2RBdMn=[2UBbMl=\\2WBbMj=\\2ZBbMc=_2aB^M\\=e2fBZMW=g2mBVMP=l2TCQMj<Q3XCnLe<S3_CjL^<X3fCfLV<]3lCaLQ<c3QDZLk;k3WDSLe;P4]DnK`;V4cDgKX;]4TEWKi:m4Q24ZMe2O2M2O^ITDT5j;bJcD\\5\\;YJREf5l:TJ\\El5b:RJcEl5[:SJjEl5T:RJQFl5n9RJWFm5g9QJ]Fo5a9oIdFo5Z9PJkFo5S9oIRGo5m8oIXGP6];O0O2O1O1O1O001O1N2O1O1O001O1N2O1O001O00kMiJaEW5o9bKhE_4U:lKbEU4\\:SL]Em3c:UL[El3c:XLZEh3f:ZLWEh3g:\\LVEd3j:^LTEc3j:`LTE`3l:cLQE^3m:eLQE[3o:hLnDY3P;jLmDX3R;kLkDU3T;nLjDS3S;RMjDn2S;R3L3M4L3M4ZGiF^6Z9aIjFY6Z9eIjFU6Z9iIkFP6X9oIlFd4^OfJj9d0mFa4AaJe9m0nF]4E]Ja9T1PGY4GYJ\\9]1UGR4o9lKXFm3k9PL]Fh3h9SL_Ff3e9VLcFa3b9[LeF^3`9]LhF[3\\9aLkFX3Y9dLoFT3V9gLRGP3S9lLTGm2P9oLXGi2m8RMZGh2V=I6J7I7I6J7I6J7I6J7I6J7I7IRVY4"}, "label": "a man in a brown wide brimmed hat"}]}
{"id": 127360, "image": "COCO_train2014_000000127360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and blue plane in back of a truck on a runway\"."}], "task": "refering", "answer_template": "The \"a white and blue plane in back of a truck on a runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 61, 62, 84, 106, 107, 129, 130, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 181, 182, 184, 185, 186, 187, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 259, 260, 261, 267, 268, 269, 270, 290, 291], "bbox": [7.8125e-05, 0.07544600938967136, 0.9976875, 0.8202347417840374], "iscrowd": 0, "area": 50083.88450000002, "rle": {"size": [426, 640], "counts": "i71U=400O10001O0O1000000O100000010O0000000000000001O01O00000000000001O01O00000000000001O0001O0000000001O00000001O00000001O0000000001O000001O000000000001O0001O00000000VDIa:7WE:`:G\\E>b:B]Eb0`:^O^Ef0a:YO]Ej0b:VO]El0b:TO]En0b:RO]Eo0c:QO\\EQ1c:oN\\ES1c:mN\\ET1d:mNZEU1e:kNZEW1f:hNZEX1f:hNYEZ1f:fNZEZ1f:fNYE\\1f:dNZE\\1f:dNYE^1f:bNZE^1f:bNYE`1g:`NXE`1h:`NWEa1i:_NWEb1h:^NXEb1h:^NXEb1h:^NXEc1g:]NYEc1g:]NXEd1h:\\NXEd1i:[NWEe1i:\\NVEd1j:\\NVEd1j:\\NVEd1j:\\NUEe1k:[NTEf1l:ZNSEg1m:YNREh1n:XNQEi1P;80O10000000000O101N10000O1O1N2H80001O000000000O100O1N]Od0L5KJ^EmM^:Q2=21O1N21O2N2N1O2N2N2N1O2N2N1O2N2N2O0O2N2O0O2N2TFoL]9_3O1O100O1O1O100O1O100O106J5K1N1000001N10N2M201O1005J2O001O1O0L400O101O01O2N2N1O1O01O001O1O2N1O2N001O1O1O01YO_GcLa8\\3dG`L\\8_3gG_LY8`3hG`LX8_3iGaLW8^3jGbLV8]3kGcLU8\\3lGdLT8[3mGdLT8\\3lGdLT8[3mGdLT8\\3lGdLT8\\3lGfLR8Y3oGhLP8X3PHkLm7U3SHoLj7o2WHSMg7m2bHgLa7Y3X1000000000000000000000001O0000000000000000000001O000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O01O001O1O1O001O1O1O001O1O1OG_FiLb9V3900001O000000000000000000000001O00000000000000000000000000001O0O1O100O1O1O1000000000hFUMQ8l2kGYMS8g2jG^MP8g2oG[Mk7i2THZMf7k2YHUMc7o2]HQM_7T3`HlL\\7Y3cHgLZ7\\3fHcLW7b3hH]LU7g3kHYLQ7l3nHSLo6Q4QInKl6W4SIhKj6]5M4L3M3N3L3M3M4L4L5K5K5K5K5K4K6K5K4K5L4L4K4M4K5L3M4K5LC]L[G`3h8bLTG`3l8bLPGa3k4hLcNH_Lb3a4UMoNU3c0\\MZOd28lMFU2LZN2f1@iN?W1SOXOm0g0fNGY1:XN5g1KkMd0T2\\O_MR1KVKf1h3QN`19hJg1f9XNYFi1m9QNSFo1Q:mMoES2P:nMPFS2o9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFS2o9lMRFT2m9mMSFS2m9mMSFS2m9mMSFS2m9mMRFT2n9lMRFS2o9mMQFR2P:nMPFR2P:mMQFR2P:nMPFQ2Q:oMoEQ2Q:oMoEP2Q:QNoEn1R:RNnEm1S:SNmEm1S:SNmE8DQ1b:fNkE9Ej0e:mNfE9Gg0d:POeE9Id0c:SOdE8Kc0b:UOcE8M`0a:XOaE966Z:A`E973Z:E_E890Y:G_E99NY:I^E9g;GYD8h;HXD8h;HXD8h;HXD8h;HXD8h;HXD8h;GYD9g;GYD9f;HZD7g;IYD7g;IYD7g;IYD7g;IXD8h;HXD8h;GYD9g;GYD9g;GYD8h;HXD8g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;HZD8f;HZD7g;IYD7g;IYD7g;IYD7g;IYD7f;JYD7g;IYD7g;HZD7g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;HZD7f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JYD7g;HZD7g;IYD7g;IYD7f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;I[D6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6e;K[D5e;KZD5g;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD5g;JZD6e;K[D5e;K[D5e;K[D5e;K[D5e;K[D5e;K[D5e;K[D4f;KZD6f;JZD6f;JZD6e;K[D5e;K[D5e;K[D5e;K[D5e;K[D4f;LZD4f;K[D5e;K[D5e;K[D5e;K[D5e;K[D5d;L\\D3e;M[D3e;M[D3e;L[D5e;K[D5e;K[D5e;K[D5e;K[D5e;K[D4f;LZD4f;LZD4e;L\\D4d;L\\D4d;L\\D4d;L\\D4d;L\\D4d;L\\D3e;M[D3e;M[D3e;L\\D4d;L[D5e;K[D5d;L\\D4d;L\\D4d;L\\D3e;M[D3e;M[D3e;L\\D4d;L\\D4d;L\\D4d;L\\D4d;L\\D3d;N\\D2d;N\\D2d;N\\D2d;M]D3d;L\\D4e;KZD6f;JZD6]<O100000000000000O1000000000O1000O100000f5"}, "label": "a white and blue plane in back of a truck on a runway"}]}
{"id": 127360, "image": "COCO_train2014_000000127360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a airplane being pulled by a automobile\"."}], "task": "refering", "answer_template": "The \"a airplane being pulled by a automobile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 61, 62, 84, 106, 107, 129, 130, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 181, 182, 184, 185, 186, 187, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 259, 260, 261, 267, 268, 269, 270, 290, 291], "bbox": [7.8125e-05, 0.07544600938967136, 0.9976875, 0.8202347417840374], "iscrowd": 0, "area": 50083.88450000002, "rle": {"size": [426, 640], "counts": "i71U=400O10001O0O1000000O100000010O0000000000000001O01O00000000000001O01O00000000000001O0001O0000000001O00000001O00000001O0000000001O000001O000000000001O0001O00000000VDIa:7WE:`:G\\E>b:B]Eb0`:^O^Ef0a:YO]Ej0b:VO]El0b:TO]En0b:RO]Eo0c:QO\\EQ1c:oN\\ES1c:mN\\ET1d:mNZEU1e:kNZEW1f:hNZEX1f:hNYEZ1f:fNZEZ1f:fNYE\\1f:dNZE\\1f:dNYE^1f:bNZE^1f:bNYE`1g:`NXE`1h:`NWEa1i:_NWEb1h:^NXEb1h:^NXEb1h:^NXEc1g:]NYEc1g:]NXEd1h:\\NXEd1i:[NWEe1i:\\NVEd1j:\\NVEd1j:\\NVEd1j:\\NUEe1k:[NTEf1l:ZNSEg1m:YNREh1n:XNQEi1P;80O10000000000O101N10000O1O1N2H80001O000000000O100O1N]Od0L5KJ^EmM^:Q2=21O1N21O2N2N1O2N2N2N1O2N2N1O2N2N2O0O2N2O0O2N2TFoL]9_3O1O100O1O1O100O1O100O106J5K1N1000001N10N2M201O1005J2O001O1O0L400O101O01O2N2N1O1O01O001O1O2N1O2N001O1O1O01YO_GcLa8\\3dG`L\\8_3gG_LY8`3hG`LX8_3iGaLW8^3jGbLV8]3kGcLU8\\3lGdLT8[3mGdLT8\\3lGdLT8[3mGdLT8\\3lGdLT8\\3lGfLR8Y3oGhLP8X3PHkLm7U3SHoLj7o2WHSMg7m2bHgLa7Y3X1000000000000000000000001O0000000000000000000001O000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O01O001O1O1O001O1O1O001O1O1OG_FiLb9V3900001O000000000000000000000001O00000000000000000000000000001O0O1O100O1O1O1000000000hFUMQ8l2kGYMS8g2jG^MP8g2oG[Mk7i2THZMf7k2YHUMc7o2]HQM_7T3`HlL\\7Y3cHgLZ7\\3fHcLW7b3hH]LU7g3kHYLQ7l3nHSLo6Q4QInKl6W4SIhKj6]5M4L3M3N3L3M3M4L4L5K5K5K5K5K4K6K5K4K5L4L4K4M4K5L3M4K5LC]L[G`3h8bLTG`3l8bLPGa3k4hLcNH_Lb3a4UMoNU3c0\\MZOd28lMFU2LZN2f1@iN?W1SOXOm0g0fNGY1:XN5g1KkMd0T2\\O_MR1KVKf1h3QN`19hJg1f9XNYFi1m9QNSFo1Q:mMoES2P:nMPFS2o9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFT2n9lMRFS2o9lMRFT2m9mMSFS2m9mMSFS2m9mMSFS2m9mMRFT2n9lMRFS2o9mMQFR2P:nMPFR2P:mMQFR2P:nMPFQ2Q:oMoEQ2Q:oMoEP2Q:QNoEn1R:RNnEm1S:SNmEm1S:SNmE8DQ1b:fNkE9Ej0e:mNfE9Gg0d:POeE9Id0c:SOdE8Kc0b:UOcE8M`0a:XOaE966Z:A`E973Z:E_E890Y:G_E99NY:I^E9g;GYD8h;HXD8h;HXD8h;HXD8h;HXD8h;HXD8h;GYD9g;GYD9f;HZD7g;IYD7g;IYD7g;IYD7g;IXD8h;HXD8h;GYD9g;GYD9g;GYD8h;HXD8g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;HZD8f;HZD7g;IYD7g;IYD7g;IYD7g;IYD7f;JYD7g;IYD7g;HZD7g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;IYD7g;HZD7f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JYD7g;HZD7g;IYD7g;IYD7f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;I[D6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6e;K[D5e;KZD5g;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD6f;JZD5g;JZD6e;K[D5e;K[D5e;K[D5e;K[D5e;K[D5e;K[D5e;K[D4f;KZD6f;JZD6f;JZD6e;K[D5e;K[D5e;K[D5e;K[D5e;K[D4f;LZD4f;K[D5e;K[D5e;K[D5e;K[D5e;K[D5d;L\\D3e;M[D3e;M[D3e;L[D5e;K[D5e;K[D5e;K[D5e;K[D5e;K[D4f;LZD4f;LZD4e;L\\D4d;L\\D4d;L\\D4d;L\\D4d;L\\D4d;L\\D3e;M[D3e;M[D3e;L\\D4d;L[D5e;K[D5d;L\\D4d;L\\D4d;L\\D3e;M[D3e;M[D3e;L\\D4d;L\\D4d;L\\D4d;L\\D4d;L\\D3d;N\\D2d;N\\D2d;N\\D2d;M]D3d;L\\D4e;KZD6f;JZD6]<O100000000000000O1000000000O1000O100000f5"}, "label": "a airplane being pulled by a automobile"}]}
{"id": 53632, "image": "COCO_train2014_000000053632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl with eyes closed smiling at camera\"."}], "task": "refering", "answer_template": "The \"little girl with eyes closed smiling at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.38087499999999996, 0.25077083333333333, 1.0, 0.9845625000000001], "iscrowd": 0, "area": 97856.5102, "rle": {"size": [480, 640], "counts": "iib3P1l=4L4L4L4M3L4L4L4L4L4L4L4N2O0O2N2N2N2N2N2N2N2O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100K5K5K5L4L4O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O1N20000O1000000O1000000O10000O1000000O1000000O100000000O1000000000000000000000000000O1000N2N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O10000O1000000000000000000000000000000O0100000000000000000000000000000O1000000000000000000000000000000O1001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1^NjJXIW5e6mJYIT5c6QK[IP5a6UK]Il4`6YK]Ig4`6^K^Ic4^6bK`I`4\\6dKbI^4Y6gKeI[4V6kKgIW4U6mKhIV4S6oKkIS4P6RLnIP4n5TLPJn3k5XLRJj3i5[LUJg3c5aL[Ja3]5gLaJ[3W5mLgJU3Q5TMlJn2l4ZMRKh2g4m3H8H7I8H8I7H8H8H8H8I7H8H8H8H8H8H8H8H8HPI"}, "label": "little girl with eyes closed smiling at camera"}]}
{"id": 53632, "image": "COCO_train2014_000000053632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with her eyes closed\"."}], "task": "refering", "answer_template": "The \"the girl with her eyes closed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.38087499999999996, 0.25077083333333333, 1.0, 0.9845625000000001], "iscrowd": 0, "area": 97856.5102, "rle": {"size": [480, 640], "counts": "iib3P1l=4L4L4L4M3L4L4L4L4L4L4L4N2O0O2N2N2N2N2N2N2N2O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100K5K5K5L4L4O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O1N20000O1000000O1000000O10000O1000000O1000000O100000000O1000000000000000000000000000O1000N2N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O10000O1000000000000000000000000000000O0100000000000000000000000000000O1000000000000000000000000000000O1001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1^NjJXIW5e6mJYIT5c6QK[IP5a6UK]Il4`6YK]Ig4`6^K^Ic4^6bK`I`4\\6dKbI^4Y6gKeI[4V6kKgIW4U6mKhIV4S6oKkIS4P6RLnIP4n5TLPJn3k5XLRJj3i5[LUJg3c5aL[Ja3]5gLaJ[3W5mLgJU3Q5TMlJn2l4ZMRKh2g4m3H8H7I8H8I7H8H8H8H8I7H8H8H8H8H8H8H8H8HPI"}, "label": "the girl with her eyes closed"}]}
{"id": 561543, "image": "COCO_train2014_000000561543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a knife to which a fork is also kept together\"."}], "task": "refering", "answer_template": "The \"a knife to which a fork is also kept together\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [321, 343, 366, 388, 410, 411, 433, 434, 456, 478, 479], "bbox": [0.6019281045751634, 0.6688725490196079, 0.779607843137255, 0.9938562091503268], "iscrowd": 0, "area": 2026.3111500000005, "rle": {"size": [612, 612], "counts": "nml62Rc02N2M3O2M2N2O1N2O1N2N2O1N2N2O2M2O1O1O1NXOZ^O4[b0100O1O100O2O000O10000O1O100O1O101O000001O00001O2N3M3M3N2M3KV\\2KQdM001O001O000O10001O001O00001O001N101O00001O001O0n]OGWa0:e^OJYa07d^OM[a04]^O4ba0i0O1O0000O100N2O0O2O1N2N2N2N2N2N2N2M2O2M3M3M3M3L4MWS`2"}, "label": "a knife to which a fork is also kept together"}]}
{"id": 561543, "image": "COCO_train2014_000000561543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a knife under a fork\"."}], "task": "refering", "answer_template": "The \"this is a knife under a fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [321, 343, 366, 388, 410, 411, 433, 434, 456, 478, 479], "bbox": [0.6019281045751634, 0.6688725490196079, 0.779607843137255, 0.9938562091503268], "iscrowd": 0, "area": 2026.3111500000005, "rle": {"size": [612, 612], "counts": "nml62Rc02N2M3O2M2N2O1N2O1N2N2O1N2N2O2M2O1O1O1NXOZ^O4[b0100O1O100O2O000O10000O1O100O1O101O000001O00001O2N3M3M3N2M3KV\\2KQdM001O001O000O10001O001O00001O001N101O00001O001O0n]OGWa0:e^OJYa07d^OM[a04]^O4ba0i0O1O0000O100N2O0O2O1N2N2N2N2N2N2N2M2O2M3M3M3M3L4MWS`2"}, "label": "this is a knife under a fork"}]}
{"id": 4490, "image": "COCO_train2014_000000004490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female skier\"."}], "task": "refering", "answer_template": "The \"female skier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 83, 84, 85, 97, 98, 99, 100, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 306, 307, 309, 310, 325], "bbox": [0.4243793911007026, 0.19775, 0.8117330210772834, 0.9348281249999999], "iscrowd": 0, "area": 45725.63579999999, "rle": {"size": [640, 427], "counts": "o[a3`1^b04L4L4L4L5_EUNe2o1VMXNd2l1WMWNg2j1WMXNg2k1UMYNi2j1PM\\Nn2h1cLdN\\3_1oKUOo3n0]KEa4?SKJl49hJ2V51nIj0P6YObGS3^8PMoFa3P9h3O0O2O1O001O1O001O1O001O1N100N2O1O1N2O1O10000O1000000000000O01L3M4M2M4L2O0O1O2A>D=9F:H8H8I7I7Il0TO4K4M4L3M4L3L3N2O1O1O1O1O1O1O1O1O1O1O100O1O2N4L3M3M4L3M2N2N100O2N1O1O2N1O1O2N1O1O10YJSKYLn4^3TLfKo3R4cLaK`3V4QM`KQ3X4^M^Kf2[4eM]K^2^4lMZKW2a4RNXKQ2b4ZNVKi1mNnI]4c4_Lc1lNQJe3j5mL>ROnIP4n6bK]O1lI[4h>O1N2N3N1N2N2O1N3N1N2N2O1N3N1N2N3M2O1N2N3M2N2N2N3M2N2N3N2M4L3M4L3M4L3M4L4L3N3L3M4L3M4L:Fj0VOkfa1"}, "label": "female skier"}]}
{"id": 4490, "image": "COCO_train2014_000000004490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person on skis wearing a beanie , ski goggles , a grey jacket and green pants\"."}], "task": "refering", "answer_template": "The \"a person on skis wearing a beanie , ski goggles , a grey jacket and green pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 83, 84, 85, 97, 98, 99, 100, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 306, 307, 309, 310, 325], "bbox": [0.4243793911007026, 0.19775, 0.8117330210772834, 0.9348281249999999], "iscrowd": 0, "area": 45725.63579999999, "rle": {"size": [640, 427], "counts": "o[a3`1^b04L4L4L4L5_EUNe2o1VMXNd2l1WMWNg2j1WMXNg2k1UMYNi2j1PM\\Nn2h1cLdN\\3_1oKUOo3n0]KEa4?SKJl49hJ2V51nIj0P6YObGS3^8PMoFa3P9h3O0O2O1O001O1O001O1O001O1N100N2O1O1N2O1O10000O1000000000000O01L3M4M2M4L2O0O1O2A>D=9F:H8H8I7I7Il0TO4K4M4L3M4L3L3N2O1O1O1O1O1O1O1O1O1O1O100O1O2N4L3M3M4L3M2N2N100O2N1O1O2N1O1O2N1O1O10YJSKYLn4^3TLfKo3R4cLaK`3V4QM`KQ3X4^M^Kf2[4eM]K^2^4lMZKW2a4RNXKQ2b4ZNVKi1mNnI]4c4_Lc1lNQJe3j5mL>ROnIP4n6bK]O1lI[4h>O1N2N3N1N2N2O1N3N1N2N2O1N3N1N2N3M2O1N2N3M2N2N2N3M2N2N3N2M4L3M4L3M4L3M4L4L3N3L3M4L3M4L:Fj0VOkfa1"}, "label": "a person on skis wearing a beanie , ski goggles , a grey jacket and green pants"}]}
{"id": 53643, "image": "COCO_train2014_000000053643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a carafe of water sitting on a table\"."}], "task": "refering", "answer_template": "The \"a carafe of water sitting on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 128, 129, 150, 151, 152, 173, 174, 175, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267], "bbox": [0.525921875, 0.16789583333333333, 0.6796249999999999, 0.6913541666666667], "iscrowd": 0, "area": 15624.326200000005, "rle": {"size": [480, 640], "counts": "WXn4`0P>b0]Ob0^Ob0^Ob0^Ob0M3M3N2XL`LmKb3R4`LkKb3S4aLjKa3T4bLiK`3V4bLfKa3X4aLfKa3X4bLeK`3Z4bLcK`3\\4bLaK`3^4bL_K_3a4cL[K`3d4eLRK`3n4eLgJa3X5fL[Ja3e5dLPJb3P6cLdId3\\6bLWIe3i6`LlHf3T7S200000000000001O00000000000000000000000000000O10000000001O0000000000XMQI_Mo6[2dIWM]6c2lIZMT6`2TJ]Mm5]2\\J`Md5Z2dJdM\\5W2lJeMU5X2PKfMP5X2UKdMl4Y2YKeMh4W2^KfMb4X2bKeM_4X2gKeMY4Y2nKaMS4i1oLkMQ3m1cMcMb2U2X5I7I7J7M2N3M3L3N3M3M2M4M3M2M4M3M2N3L4M2NTPP3"}, "label": "a carafe of water sitting on a table"}]}
{"id": 53643, "image": "COCO_train2014_000000053643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass vase with clear water in it , next to a wine glass with brown beer in it on a brown wood table\"."}], "task": "refering", "answer_template": "The \"a glass vase with clear water in it , next to a wine glass with brown beer in it on a brown wood table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 128, 129, 150, 151, 152, 173, 174, 175, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267], "bbox": [0.525921875, 0.16789583333333333, 0.6796249999999999, 0.6913541666666667], "iscrowd": 0, "area": 15624.326200000005, "rle": {"size": [480, 640], "counts": "WXn4`0P>b0]Ob0^Ob0^Ob0^Ob0M3M3N2XL`LmKb3R4`LkKb3S4aLjKa3T4bLiK`3V4bLfKa3X4aLfKa3X4bLeK`3Z4bLcK`3\\4bLaK`3^4bL_K_3a4cL[K`3d4eLRK`3n4eLgJa3X5fL[Ja3e5dLPJb3P6cLdId3\\6bLWIe3i6`LlHf3T7S200000000000001O00000000000000000000000000000O10000000001O0000000000XMQI_Mo6[2dIWM]6c2lIZMT6`2TJ]Mm5]2\\J`Md5Z2dJdM\\5W2lJeMU5X2PKfMP5X2UKdMl4Y2YKeMh4W2^KfMb4X2bKeM_4X2gKeMY4Y2nKaMS4i1oLkMQ3m1cMcMb2U2X5I7I7J7M2N3M3L3N3M3M2M4M3M2M4M3M2N3L4M2NTPP3"}, "label": "a glass vase with clear water in it , next to a wine glass with brown beer in it on a brown wood table"}]}
{"id": 53643, "image": "COCO_train2014_000000053643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table filled with many beer bottles and glasses\"."}], "task": "refering", "answer_template": "The \"a table filled with many beer bottles and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.0016875000000000002, 0.0, 0.9943750000000001, 0.9865208333333333], "iscrowd": 0, "area": 221120.58305, "rle": {"size": [480, 640], "counts": "gj0Q3g7X400O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O101N100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O1\\OSFULm9j3\\FmKe9S4d0O1O100O1O1O100O1O100O1O1O100O1O1007I7I6J1O00O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O1N2O1O1O1N2O1O1000000001O000000001O00000000001O0000O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100M3N2M3M3N2M3M3YORDZNP<d1g0N2O1N200001O0000000000JQC\\Nn<d1TCZNl<f1UCYNk<g1VCXNj<h17O1000000O100000000O100000000O1000000O11O1O1O1O2N1OImBbNR=^1800000000000000000000O100000000001O00000000000000QOo000000000000000O100000000000000004L3M1O0000000000000000001O0000000000000000001OR1nN00O10000POVCIk<6]CCc<=dC[O]<e0jCTOV<l0o0O100000000O100000000O100000000O1000000O100000000O100C=L4000000001O00000000000000001O00000000000000001O000000000kAN\\=2YB9g=a0O0000001O0000001O0000001O0000001O0000001O0000001ZBiNa=[10001O00000000001O000000001O0001O00cCkNo:V1gDAk:?lD5i:KoD`0n:l1M3M4L3M3M3M00O1O1O1O1O1N2O1O1O1O1O1O10000001O000000001O000000001O000000001O0000001O0O1N2M3N2M3M4M2M3N2M3N2M3M3N2M3N2M3iHcJi4`5SKcJj4`5SKdJi4_5UKcJi4_5TKdJi4_5TKeJi4]5TKfJi4]5TKfJj4\\5SKhJi4[5TKhJi4[5TKhJj4[5SKhJi4e5jJ^JT5n5_JUJ^5X6UJlIh5g7M3M4M2M3N3L3N2M4L3N2M4M2M3N3L3M3N3L3N2M4M2M3M[e1"}, "label": "a table filled with many beer bottles and glasses"}]}
{"id": 53643, "image": "COCO_train2014_000000053643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a hawaiian shirt and blue jeans sitting in a white chair\"."}], "task": "refering", "answer_template": "The \"a person in a hawaiian shirt and blue jeans sitting in a white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 54, 55, 56, 76, 77, 78, 79, 100, 101], "bbox": [0.33253125, 0.0, 0.5560937499999999, 0.2758541666666667], "iscrowd": 0, "area": 8062.6298499999975, "rle": {"size": [480, 640], "counts": "UnS32m>2N3M2N2N3L3N3M2N3M2N3M2N2N3M2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O1O000000001N1O2N1O2N2N1O2N1O2N1O2N2O0O2POVNhDk1X;o00O1000O0100O100O01000O10O10O1000O01000000O01000O1O1O1O1N2O1O1O1N2F:\\Od0A?N2M3N3L3N200O101O0O100O101O0O100O10001N100O10001N10000O101O001O1O1O1O2N1O1O1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O1OlRU4"}, "label": "a person in a hawaiian shirt and blue jeans sitting in a white chair"}]}
{"id": 53643, "image": "COCO_train2014_000000053643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person wearing floral shirt\"."}], "task": "refering", "answer_template": "The \"person wearing floral shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 54, 55, 56, 76, 77, 78, 79, 100, 101], "bbox": [0.33253125, 0.0, 0.5560937499999999, 0.2758541666666667], "iscrowd": 0, "area": 8062.6298499999975, "rle": {"size": [480, 640], "counts": "UnS32m>2N3M2N2N3L3N3M2N3M2N3M2N2N3M2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O1O000000001N1O2N1O2N2N1O2N1O2N1O2N2O0O2POVNhDk1X;o00O1000O0100O100O01000O10O10O1000O01000000O01000O1O1O1O1N2O1O1O1N2F:\\Od0A?N2M3N3L3N200O101O0O100O101O0O100O10001N100O10001N10000O101O001O1O1O1O2N1O1O1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O1OlRU4"}, "label": "person wearing floral shirt"}]}
{"id": 283018, "image": "COCO_train2014_000000283018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman , on skis , is wearing a red jacket\"."}], "task": "refering", "answer_template": "The \"a woman , on skis , is wearing a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386], "bbox": [0.5952968750000001, 0.19308333333333336, 0.8424843750000001, 0.9868125], "iscrowd": 0, "area": 34585.9072, "rle": {"size": [480, 640], "counts": "SQc54g>500O2N101N1O101N1O2O0O101N1O101N1O100O1O10O`NA`D`0Y;M`D4Y;9aDGW;g0bDZOW;R1cDoNV;^1dDbNT;i2I7I7H8I7I7I7H8J6J6J6J5K6J6J6J6J6I7I7H8I7H8I7H8I7H8H8I7H8I7H8I7H8J6N1O2N2N2O1O1eLkDj2V;TMlDl2T;RMnDn2R;PMPEo2Q;nLRER3n:kLUET3[;1000O1000000O10000O100000000001O00001O001O00001O00001O00001O002N1O1O1O2N1O1O2N1QF`L\\8a3XGjLh8h4O001O001O001iKmL_LS3[3VMbLk2W3_MeLa2U3hMhLY2T3mMiLT2T3RNhLn1W3VNfLk1W3[NeLe1Z3_NcLb1Z3cNaL_1^3fN[L_1a3fNXLa1d3dNTLb1h3cNRLb1j3m4M3L4M3L4L5L3L4M3L4L4M3L4M4K4G9G9G9G9F:G9GXT_1"}, "label": "a woman , on skis , is wearing a red jacket"}]}
{"id": 283018, "image": "COCO_train2014_000000283018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in the snow field with his colleague with smiling face\"."}], "task": "refering", "answer_template": "The \"a lady in the snow field with his colleague with smiling face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386], "bbox": [0.5952968750000001, 0.19308333333333336, 0.8424843750000001, 0.9868125], "iscrowd": 0, "area": 34585.9072, "rle": {"size": [480, 640], "counts": "SQc54g>500O2N101N1O101N1O2O0O101N1O101N1O100O1O10O`NA`D`0Y;M`D4Y;9aDGW;g0bDZOW;R1cDoNV;^1dDbNT;i2I7I7H8I7I7I7H8J6J6J6J5K6J6J6J6J6I7I7H8I7H8I7H8I7H8H8I7H8I7H8I7H8J6N1O2N2N2O1O1eLkDj2V;TMlDl2T;RMnDn2R;PMPEo2Q;nLRER3n:kLUET3[;1000O1000000O10000O100000000001O00001O001O00001O00001O00001O002N1O1O1O2N1O1O2N1QF`L\\8a3XGjLh8h4O001O001O001iKmL_LS3[3VMbLk2W3_MeLa2U3hMhLY2T3mMiLT2T3RNhLn1W3VNfLk1W3[NeLe1Z3_NcLb1Z3cNaL_1^3fN[L_1a3fNXLa1d3dNTLb1h3cNRLb1j3m4M3L4M3L4L5L3L4M3L4L4M3L4M4K4G9G9G9G9F:G9GXT_1"}, "label": "a lady in the snow field with his colleague with smiling face"}]}
{"id": 283018, "image": "COCO_train2014_000000283018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ice skating man\"."}], "task": "refering", "answer_template": "The \"ice skating man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 220, 221, 222, 243, 244, 266, 267, 289, 290, 311, 312, 313, 335, 336, 358, 359], "bbox": [0.5117499999999999, 0.11654166666666667, 0.687921875, 0.8973125], "iscrowd": 0, "area": 17785.183800000003, "rle": {"size": [480, 640], "counts": "ini4<_>7J6L4K6K2N200O1O100O00100O1O100O00O1M4M2M3O20O010O0ZGPO^3o0\\LXOc3i0UL_Ok3`0SLCl3>SLCm3<SLEl3<SLEm3:RLHl3:SLGk3:ULGi3;VLFc3`0]LAS3o0mLQOl2U1UMkN^LVOR5V2`NdN]LXOm4X2fNaN]LWOl4Y2gN_N]LZOk4V2iN`N[L\\Oj4U2kN_NZL^Oi4S2VKSN32d0:YO_Oi4S2PKcNA4Y1F^OAf4\\5hKSKACe4_5fKnJDDe4d5bKgJIGc4g5_KcJNFc4k5[K_J1Fd4Q6VKYJ5GT3_O`Mf6SOSJ9IR3D]Me6TOnI<JR3HYMe6UOiI`0JP3NVMe6UObIe0Ln23SMd6VO]Ih0Mn27oLZ72`Hm2;mLU77_Hk2b0iLP7;_Hk2f0eLl6?_Hk2i0bLi6b0_Hk2V9UMkFj2V9VMjFj2V9UMkFj2U9VMkFj2f:O100O10O10000gM\\MVHd2d7YNeGg1V8fNaG[1Y8lNfGS1U8TOjGl0P8\\OnGd0l7CSH=g7JWH7c70\\H0^77aHIY7>fHBU7e0iH\\OR7i0lHYOo6_4K5L4K4L5K5L4H8E;E;F:H9Ge0]LoDT2f;UMcDc2R<H8H8H7I8G9H7I8H8G9Hadm2"}, "label": "ice skating man"}]}
{"id": 283018, "image": "COCO_train2014_000000283018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing next to a woman with goggles on his head getting ready to go skiing\"."}], "task": "refering", "answer_template": "The \"a man standing next to a woman with goggles on his head getting ready to go skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 220, 221, 222, 243, 244, 266, 267, 289, 290, 311, 312, 313, 335, 336, 358, 359], "bbox": [0.5117499999999999, 0.11654166666666667, 0.687921875, 0.8973125], "iscrowd": 0, "area": 17785.183800000003, "rle": {"size": [480, 640], "counts": "ini4<_>7J6L4K6K2N200O1O100O00100O1O100O00O1M4M2M3O20O010O0ZGPO^3o0\\LXOc3i0UL_Ok3`0SLCl3>SLCm3<SLEl3<SLEm3:RLHl3:SLGk3:ULGi3;VLFc3`0]LAS3o0mLQOl2U1UMkN^LVOR5V2`NdN]LXOm4X2fNaN]LWOl4Y2gN_N]LZOk4V2iN`N[L\\Oj4U2kN_NZL^Oi4S2VKSN32d0:YO_Oi4S2PKcNA4Y1F^OAf4\\5hKSKACe4_5fKnJDDe4d5bKgJIGc4g5_KcJNFc4k5[K_J1Fd4Q6VKYJ5GT3_O`Mf6SOSJ9IR3D]Me6TOnI<JR3HYMe6UOiI`0JP3NVMe6UObIe0Ln23SMd6VO]Ih0Mn27oLZ72`Hm2;mLU77_Hk2b0iLP7;_Hk2f0eLl6?_Hk2i0bLi6b0_Hk2V9UMkFj2V9VMjFj2V9UMkFj2U9VMkFj2f:O100O10O10000gM\\MVHd2d7YNeGg1V8fNaG[1Y8lNfGS1U8TOjGl0P8\\OnGd0l7CSH=g7JWH7c70\\H0^77aHIY7>fHBU7e0iH\\OR7i0lHYOo6_4K5L4K4L5K5L4H8E;E;F:H9Ge0]LoDT2f;UMcDc2R<H8H8H7I8G9H7I8H8G9Hadm2"}, "label": "a man standing next to a woman with goggles on his head getting ready to go skiing"}]}
{"id": 283018, "image": "COCO_train2014_000000283018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man ' s left ski\"."}], "task": "refering", "answer_template": "The \"man ' s left ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [307, 308, 309, 310, 311, 317, 335, 336, 340, 357, 358], "bbox": [0.360640625, 0.7689166666666666, 0.817375, 0.9239583333333333], "iscrowd": 0, "area": 1867.7821500000023, "rle": {"size": [480, 640], "counts": "bd\\31o>2N2N2N2N000000000O2O000000000000001O0001O000O100000001N1000000O10001O0O10000000001O000001O000001O0001O00000001O01O0000000010O0000000001O01O000000000000N20KXA2h>3000O10000000000000O1oe5JWZJ3SN1bD3[;MeD2[;OeD1[;OeD1[;OeD0\\;0cD1];OcD1\\;0dDO];1cDO];1bD0^;0bD0^;0bDO^;2bDN^;2aDO_;1aDN`;2`DN`;2`DN_;2bDN^;2aDN`;2`DNb;0^D0_;KSC4^10f;NYD3i;KWD5Z=O1000O10O1000O0100000O010000O10O1O001N2O1O0O2O1O010O00100O00100O00100O010O1O010O00100O010O1O0101O2M2O3M3L4MhYX1NRfgN;E;M3M00O1O1O100O2O000O102N2M3N1N2O2N1N2O2N1NY^f1"}, "label": "man ' s left ski"}]}
{"id": 217487, "image": "COCO_train2014_000000217487.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table at which people are eating\"."}], "task": "refering", "answer_template": "The \"the table at which people are eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 143, 144, 145, 146, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.246078125, 0.34347319347319344, 0.809203125, 0.9876223776223775], "iscrowd": 0, "area": 31357.3538, "rle": {"size": [429, 640], "counts": "]nQ21Z=6J4L2O0O2\\ODSD`0m;`0000O1000000O100000RDXOW;g0jD[OT;e0kD^OR;c0nD_OP;`0PEBo:>QEDm:;SEHk:8SELk:4TEOj:0VE2i:NVE5h:JWE;f:EYE>e:AYEc0f:]OVEh0i:n001O1O4L00000O100000000000O10O10O011O1N10SF\\MR9b2oFaMo8^2QGfMl8Y2TGjMi8U2WGlMj8T2UGmMk8S2TGmMm8S2RGmMn8T2PGnMP9R2oFnMR9R2mFnMT9R2kFoMU9Q2jFoMV9R2hFoMY9Q2eFQN[9o1cFRN:\\O]8b2WGSN;_O[8_2XGRN=CY8^2UGPNb0EW8T3hGPMV8o2iGYMP8g2oG`Ml7`2RHaMU8Z2hGgMZ8Y2bGiM`8W2\\GkMf8V2VGkMl8Z32M2O1O1O1O1O1O1O1O1O1O1N2O001O1O1O1O001O001O0O101O000001O0000000100O001O1O1O010O1O00100O10O01O10O0100O10O01O10O0100O10O1000O0100000O1000O10O100000O01000000O10O1000kLZFb2f9[M^Fd2b9YM`Fi2_9TMcF?Am1\\:SNeEm1Z:RNhEo1V:QNkEo1T:PNnEQ2P:oMQFQ2n9nMTFS2j9mMWFS2h9mMXFW2d9iM]FZ2_9fMbF]2Z9cMfFa2k0SMQ7<UHd2c0SMX79VHg2;SM_76WHj23SMf74VHo2JQMP80WHf3j7ZLVHe3k7ZLUHf3l7YLUHf3l7ZLTHe3m7ZLTHe3m7[LSHd3n7[LSHd3n7[LSHd3n7\\LRHc3o7\\LRHc3o7]LQHb3P8]LQHb3P8]LQHb3P8]LQHb3Q8\\LoGd3R8[LoGd3m800O2O00001O00001O00001O00001O00001O0000001O000000001O000000001O000000001O000000001O0000001O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O00\\fb1"}, "label": "the table at which people are eating"}]}
{"id": 217487, "image": "COCO_train2014_000000217487.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown table\"."}], "task": "refering", "answer_template": "The \"brown table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 143, 144, 145, 146, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.246078125, 0.34347319347319344, 0.809203125, 0.9876223776223775], "iscrowd": 0, "area": 31357.3538, "rle": {"size": [429, 640], "counts": "]nQ21Z=6J4L2O0O2\\ODSD`0m;`0000O1000000O100000RDXOW;g0jD[OT;e0kD^OR;c0nD_OP;`0PEBo:>QEDm:;SEHk:8SELk:4TEOj:0VE2i:NVE5h:JWE;f:EYE>e:AYEc0f:]OVEh0i:n001O1O4L00000O100000000000O10O10O011O1N10SF\\MR9b2oFaMo8^2QGfMl8Y2TGjMi8U2WGlMj8T2UGmMk8S2TGmMm8S2RGmMn8T2PGnMP9R2oFnMR9R2mFnMT9R2kFoMU9Q2jFoMV9R2hFoMY9Q2eFQN[9o1cFRN:\\O]8b2WGSN;_O[8_2XGRN=CY8^2UGPNb0EW8T3hGPMV8o2iGYMP8g2oG`Ml7`2RHaMU8Z2hGgMZ8Y2bGiM`8W2\\GkMf8V2VGkMl8Z32M2O1O1O1O1O1O1O1O1O1O1N2O001O1O1O1O001O001O0O101O000001O0000000100O001O1O1O010O1O00100O10O01O10O0100O10O01O10O0100O10O1000O0100000O1000O10O100000O01000000O10O1000kLZFb2f9[M^Fd2b9YM`Fi2_9TMcF?Am1\\:SNeEm1Z:RNhEo1V:QNkEo1T:PNnEQ2P:oMQFQ2n9nMTFS2j9mMWFS2h9mMXFW2d9iM]FZ2_9fMbF]2Z9cMfFa2k0SMQ7<UHd2c0SMX79VHg2;SM_76WHj23SMf74VHo2JQMP80WHf3j7ZLVHe3k7ZLUHf3l7YLUHf3l7ZLTHe3m7ZLTHe3m7[LSHd3n7[LSHd3n7[LSHd3n7\\LRHc3o7\\LRHc3o7]LQHb3P8]LQHb3P8]LQHb3P8]LQHb3Q8\\LoGd3R8[LoGd3m800O2O00001O00001O00001O00001O00001O0000001O000000001O000000001O000000001O000000001O0000001O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O00\\fb1"}, "label": "brown table"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clear glass container of dip\"."}], "task": "refering", "answer_template": "The \"the clear glass container of dip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295], "bbox": [0.62825, 0.4840325865580448, 0.88225, 0.7245417515274949], "iscrowd": 0, "area": 14190.213300000001, "rle": {"size": [491, 640], "counts": "\\PQ63V?6K5K5J6K5J8I6J7H8I5L2N1N3N2N1N3N1O2N2M2O2N1O2M2O2N1N3N1O2N1N3N00000O10000O1000000O1000000000000000000O10000000000000001O0000000000000000000ZDeL];d30O100000001N100000000O2O0000000O2O0000000O101O000O1000001O0O100000001O0O1000001O001N2O001O001N2O001O1O0O2N1O2N2N1O2N2O0O2N1O2N2N1O2N2O0O2N1O2N2N1O2N2N101N1O2N2N2N3M3M4M2M3M4L3M3H8HVUT1"}, "label": "the clear glass container of dip"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small glass dish of some type of dip\"."}], "task": "refering", "answer_template": "The \"small glass dish of some type of dip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295], "bbox": [0.62825, 0.4840325865580448, 0.88225, 0.7245417515274949], "iscrowd": 0, "area": 14190.213300000001, "rle": {"size": [491, 640], "counts": "\\PQ63V?6K5K5J6K5J8I6J7H8I5L2N1N3N2N1N3N1O2N2M2O2N1O2M2O2N1N3N1O2N1N3N00000O10000O1000000O1000000000000000000O10000000000000001O0000000000000000000ZDeL];d30O100000001N100000000O2O0000000O2O0000000O101O000O1000001O0O100000001O0O1000001O001N2O001O001N2O001O1O0O2N1O2N2N1O2N2O0O2N1O2N2N1O2N2O0O2N1O2N2N1O2N2N101N1O2N2N2N3M3M4M2M3M4L3M3H8HVUT1"}, "label": "small glass dish of some type of dip"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl with red berries\"."}], "task": "refering", "answer_template": "The \"a bowl with red berries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330], "bbox": [0.051609375, 0.3766802443991853, 0.4644375, 0.8228716904276985], "iscrowd": 0, "area": 44016.65535, "rle": {"size": [491, 640], "counts": "jQ`0S1W>5K4L4L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L4L4L2N3L3N2N2N2N2N3M2N2N2N2O1O2M2O1O1O1N2O2N1N2O1O1O1N3N1O0O2O00000O2O000O101O00000O2O0000001O00000O2O0000001O000O101O0000001O000O101O0000001O000O101O0000001O0O10001O0000001O0O10001O0000001N1000001O0000001N1000XI`Ge6c81O0000010O0000O101O0O10000O2O000O10001N10000O101O0O10000O2O000O10001N10000O101O0O10000O2O000O101O000O10001N10000O101O0O10000O2O001N101O0O2O001N101O1N101O0O2O0O2N1N3N1O2M2O2M2O2M2O2M2O2N1N3N1N3N1N3N1O2M3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M2O2N2M4M4K4M3L4M3L4K5L4K5K5K5K5K6J5KjVT5"}, "label": "a bowl with red berries"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a striped bowl with red strawberries in it\"."}], "task": "refering", "answer_template": "The \"a striped bowl with red strawberries in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330], "bbox": [0.051609375, 0.3766802443991853, 0.4644375, 0.8228716904276985], "iscrowd": 0, "area": 44016.65535, "rle": {"size": [491, 640], "counts": "jQ`0S1W>5K4L4L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L4L4L2N3L3N2N2N2N2N3M2N2N2N2O1O2M2O1O1O1N2O2N1N2O1O1O1N3N1O0O2O00000O2O000O101O00000O2O0000001O00000O2O0000001O000O101O0000001O000O101O0000001O000O101O0000001O0O10001O0000001O0O10001O0000001N1000001O0000001N1000XI`Ge6c81O0000010O0000O101O0O10000O2O000O10001N10000O101O0O10000O2O000O10001N10000O101O0O10000O2O000O101O000O10001N10000O101O0O10000O2O001N101O0O2O001N101O1N101O0O2O0O2N1N3N1O2M2O2M2O2M2O2M2O2N1N3N1N3N1N3N1O2M3N1N3N1N3N1N3N1O2M2O2M2O2M2O2M2O2N2M4M4K4M3L4M3L4K5L4K5K5K5K5K6J5KjVT5"}, "label": "a striped bowl with red strawberries in it"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large knife sitting near small white items\"."}], "task": "refering", "answer_template": "The \"a large knife sitting near small white items\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 53, 54, 55, 77, 78, 79, 80, 101, 102, 103, 125, 126], "bbox": [0.27914062500000003, 0.0660692464358452, 0.49178125, 0.3009775967413442], "iscrowd": 0, "area": 4801.410250000002, "rle": {"size": [491, 640], "counts": "cke2:Q?2N1O1O1N101O001N2O00100O001O001O1O010O010O1O010O10O010O0100O010O0010O0100O010O010O10O01O010O010O10O010O01000O100O100O1000lA\\O`=c0VBHj=l010000O100O10000O100O10O1O010O001O10O01O010O1O0010O01O1O010O010O10O010O0100O0100O10O10O10O01000O0100O10O10O10O010000O100O10000O100O10000O100O1000Zgk4"}, "label": "a large knife sitting near small white items"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger knife next to the garlic\"."}], "task": "refering", "answer_template": "The \"the larger knife next to the garlic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 53, 54, 55, 77, 78, 79, 80, 101, 102, 103, 125, 126], "bbox": [0.27914062500000003, 0.0660692464358452, 0.49178125, 0.3009775967413442], "iscrowd": 0, "area": 4801.410250000002, "rle": {"size": [491, 640], "counts": "cke2:Q?2N1O1O1N101O001N2O00100O001O001O1O010O010O1O010O10O010O0100O010O0010O0100O010O010O10O01O010O010O10O010O01000O100O100O1000lA\\O`=c0VBHj=l010000O100O10000O100O10O1O010O001O10O01O010O1O0010O01O1O010O010O10O010O0100O0100O10O10O10O01000O0100O10O10O10O010000O100O10000O100O10000O100O1000Zgk4"}, "label": "the larger knife next to the garlic"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue dish containing baby carrots , green pepper strips , and a bowl of dip\"."}], "task": "refering", "answer_template": "The \"a blue dish containing baby carrots , green pepper strips , and a bowl of dip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 220, 221, 226, 227, 228, 241, 242, 243, 244, 249, 250, 251, 252, 264, 265, 266, 267, 272, 273, 274, 275, 288, 289, 290, 291, 292, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.5028125, 0.4436863543788187, 0.9745312500000001, 0.8378411405295315], "iscrowd": 0, "area": 24139.134950000018, "rle": {"size": [491, 640], "counts": "Qfj42Y?3L4L4M3L3N3L3M3N2M4H7K5M3M4M1N2O001O001O001O001K401O0O2O001N101O001O0O2O001O001N101O0O2M2O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1N3N101N1O2N1O2N101N1O2N1O2O0O2N1O2N101NfLPE[2o:bMYE[2e:cMaEZ2^:dMhEZ2W:cMoEZ2P:fMTFX2j9hM[FT2d9lMbFP2\\9PNjFk1U9TNRGh1m8WNYGd1f8\\N`G`1^8`NhG[1W8eNoGW1o7iNUHT1k7lNUHU1i7lNXHS1g7nNZHR1d7oN]HP1b7QO_Ho0`7PObHo0]7ROdHn0Z7SOfHm0Y7TOhHl0W7TOjHk0V7UOkHk0T7UOmHj0T7UOmHk0R7UOoHj0Q7VOPIj0o6VOQIj0P7UOQIk0n6TOSIl0m6TOSIm0l6SOUIl0l6SOTIn0k6ROUIn0k6ROUIo0j6QOWIn0i6ROWIo0i6POWIP1i6POXIP1g6POYIP1g6POYIQ1g6nNYIR1g6nNZIR1e6nN[IR1e6nN[IS1e6lN[IT1e6lN\\IT1c6lN]IT1c6lN]IU1c6jN]IV1c6kN]IU1b6kN^IU1b6kN^IU1c6jN^IV1a6jN_IV1a6jN`IU1`6kN`IU1a6jN_IV1a6jN`IU1`6kN`IU1a6jN`IV1_6jNaIV1_6jNaIV1`6iNaIV1_6jNaIV1`6iNaIV1_6jNaIV1_6jNaIW1_6iNaIV1_6jNaIV1`6iN`IW1`6iNaIV1_6jNaIV1`6iNaIW1^6iNbIW1_6hNaIX1_6hNbIW1^6iNbIW1_6hNbIW1^6iNbIW1_6hNaIY1^6gNcIX1]6hNcIX1^6gNbIY1^6gNbIY1_6gN_IZ1a6fN_IZ1a6fN^I\\1b6cN^I]1b6cN]I^1d6aN\\I_1d6aN\\I_1d6aN[I`1X:0O2O0O100O10001N100O100O100O2O000O100O100O2O0oLRNdHo1\\7QNdHo1\\7RNcHn1]7SNaHn1_7SN`Hm1`7SN_Ho1a7QN]HP2c7QN\\Ho1d7QN[HP2e7QNYHP2g7QNWHP2i7QNVHo1j7QNUHQ2j7PNTHQ2l7PNSHP2m7PNRHQ2o7oMoGR2Q8oMmGR2S8oMlGR2S8oMkGR2U8PNhGQ2X8PNgGP2Y8RNdGo1\\8RNbGo1^8RNaGn1_8TN^Gm1c8SN[Gn1e8TNXGl1i8UNVGk1j8WNSGi1n8XNPGi1P9YNlFi1T9XNhFj1Y9WNcFl1]9VN^Fl1d9TNXFo1h9SNTFo1l9RNPFP2Q:RNjEQ2V:PNfER2[:oMaET2_:\\10O100O2O0O1O1O1O100O1O2N101N1O1O2N101N1O1O2N101N1O1O2N2O2M3M2N3M3N1N3M2L4L5J5L4L5K4L4L4L3M4K4M3M4L3M4L3M4L3L4M4L3M4L<DX\\7"}, "label": "a blue dish containing baby carrots , green pepper strips , and a bowl of dip"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a checkered bowl\"."}], "task": "refering", "answer_template": "The \"a checkered bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.5094531250000001, 0.1574949083503055, 0.876703125, 0.3240325865580448], "iscrowd": 0, "area": 13880.333799999999, "rle": {"size": [491, 640], "counts": "o\\l43W?4L5K5K5K5K5K2N2N1O2N1O2N101O001O001O001O001O001O001O0O2O001O001O001O00001O0000001O00001O0000001O0000001O000O2O0000001O0000001O00001O0000001O0000001O00001O0000001O0O100010O000001O00001O01O0001O00000010O0001O00000010O00000001O000001O0000000000001O0001O0000000001O0000000000001O01O000000000001O0000000000010O0000000000001O0000001O00010O00001O00001O00001O00001N10001O001N10001O000O2O00001O0O101O00001N10001O000O2O00001O0O101O00001N10001O000O2O2N1O2N1N3N1O2N1O2M2O2N1O2G9GZhU1"}, "label": "a checkered bowl"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blurry dish behind the vegetable plate\"."}], "task": "refering", "answer_template": "The \"the blurry dish behind the vegetable plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.5094531250000001, 0.1574949083503055, 0.876703125, 0.3240325865580448], "iscrowd": 0, "area": 13880.333799999999, "rle": {"size": [491, 640], "counts": "o\\l43W?4L5K5K5K5K5K2N2N1O2N1O2N101O001O001O001O001O001O001O0O2O001O001O001O00001O0000001O00001O0000001O0000001O000O2O0000001O0000001O00001O0000001O0000001O00001O0000001O0O100010O000001O00001O01O0001O00000010O0001O00000010O00000001O000001O0000000000001O0001O0000000001O0000000000001O01O000000000001O0000000000010O0000000000001O0000001O00010O00001O00001O00001O00001N10001O001N10001O000O2O00001O0O101O00001N10001O000O2O00001O0O101O00001N10001O000O2O2N1O2N1N3N1O2N1O2M2O2N1O2G9GZhU1"}, "label": "the blurry dish behind the vegetable plate"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a granite counter top sitting underneath a bowl of strawberries\"."}], "task": "refering", "answer_template": "The \"a granite counter top sitting underneath a bowl of strawberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 240, 241, 253, 254, 263, 264, 276, 277, 278, 285, 286, 287, 299, 300, 301, 302, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356], "bbox": [0.031968750000000004, 0.5598981670061101, 0.49546874999999996, 0.8671894093686355], "iscrowd": 0, "area": 17687.585450000002, "rle": {"size": [491, 640], "counts": "lo9:l;U3ROn0000000000000O10000000gKUER4k:lKYER4R;N3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N3M2M3N3M2N2N2N3M2N2N2N1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001N10O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2N2O1N2M3M3M3L4M3M3M3M3M3000000000000000000000000000000Z2fMTaj4"}, "label": "a granite counter top sitting underneath a bowl of strawberries"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table the strawberries are on\"."}], "task": "refering", "answer_template": "The \"the table the strawberries are on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 240, 241, 253, 254, 263, 264, 276, 277, 278, 285, 286, 287, 299, 300, 301, 302, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356], "bbox": [0.031968750000000004, 0.5598981670061101, 0.49546874999999996, 0.8671894093686355], "iscrowd": 0, "area": 17687.585450000002, "rle": {"size": [491, 640], "counts": "lo9:l;U3ROn0000000000000O10000000gKUER4k:lKYER4R;N3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N3M2M3N3M2N2N2N3M2N2N2N1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001N10O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2O0O2N2O1N2N2N2O1N2M3M3M3L4M3M3M3M3M3000000000000000000000000000000Z2fMTaj4"}, "label": "the table the strawberries are on"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the darker table on the right cell\"."}], "task": "refering", "answer_template": "The \"the darker table on the right cell\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 126, 127, 134, 135, 149, 150, 151, 152, 153, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 226, 227, 228, 229, 241, 242, 251, 252, 275, 287, 288, 297, 298, 310, 311, 312, 313, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.5014062499999999, 0.254765784114053, 0.978078125, 0.8659674134419553], "iscrowd": 0, "area": 37317.65495, "rle": {"size": [491, 640], "counts": "WVj4;Z>e0[Oe0\\Od0PDiNX9m1hFFo6o0SI_Oh5V1ZJiNP5m1PKSNk4Q2WKnMk4o1WKPNk4n1UKRNm4k1UKTNm4i1UKVNm4g1TKYNm4f1TKYNn4d1TK[Nn4b1SK^No4`1RK_NP5^1RKaNP5\\1QKdNQ5Y1QKfNQ5X1oJhNR5V1PKiNR5T1PKkNR5S1nJmNS5Q1oJnNR5P1PKoNP5Q1PKoNP5P1RKoNo4o0SKPOm4o0TKQOm4n0TKQOl4n0VKQOk4m0VKSOj4m0WKROj4l0XKSOh4l0YKTOg4k0[KTOf4k0[KTOe4k0\\KUOe4i0]KVOc4j0^KUOc4i0^KWOb4h0`KWOa4h0`KWO`4h0aKXO_4g0cKXO^4f0dKYO\\4g0dKYO]4e0eKZO[4e0fK[O[4d0fK[OZ4d0gK\\OZ4b0hK]OX4b0iK^OW4b0jK]OW4a0jK_OV4`0lK_OU4`0kK@U4?mK@T4>mKBS4>nKAS4=oKBQ4=PLCP4<RLCo3<QLDo3;SLDn3:SLFm3:TLEm39TLGl38VLGk37VLIj37WLHi37XLIi35YLJg36ZLIg35ZLKf34\\LKd35\\LKd34^LKc33^LMb32`LM`33`LM`32bLM^32cLN]32dLM]31dLO\\30fLOZ30gL0Y30hLOX30jLOV30kL0V3OkL0U3OlL1T3NnL1R3NoL2Q3NPM1Q3MPM3P3LRM3n2MRM3n2LTM3l2LUM4k2LVM3k2KVM5j2JXM5h2JZM5f2KZM5f2J[M6e2J\\M5e2J[M6e2J[M6e2KZM5f2K[M4e2L[M4f2LYM4g2LYM4g2LZM3f2NYM2g2NYM2g2NYM2h2NXM1h2OXM1h2OXM1h20WM0i20XMOh21XMOi21VMOj21VMOj21WMNi23VMMj23VMMk22UMNk22VMMj23VMMj22WMNi22WMNi22XMMi21XMOh21XMOh20YM0g20ZMOf21ZMOf21ZMOf21ZMOf22ZMMf23ZMMf24YMLf25ZMKf26ZMIf28YMHg28YMHf2:YMFg2:ZMEf2;ZMEf2<YMDf2=ZMCf2=[MBe2?ZMAf2?ZMAe2a0ZM_Of2a0[M^Oe2b0[M^Oe2c0ZM]Oe2d0[M\\Oe2e0[MZOe2f0[MZOe2f0[MZOd2h0[MXOe2h0\\MWOd2i0\\MWOd2j0[MVOd2k0\\MUOd2l0\\MSOd2m0\\MSOd2m0\\MSOc2o0\\MQOd2o0]MPOc2Q1\\MoNc2R1]MnNc2R1]MnNc2S1]MlNc2T1]MlNb2U1^MkNb2V1]MjNc2V1^MiNb2X1]MhNb2Y1^MgNb2Y1^MgNb2Y1_MfNa2[1^MeNa2\\1_MdNa2\\1_MdNa2\\1`McN`2^1_MbN`2_1`MaN`2_1`MaN`2_1_MbNa2_1^MaNa2`1_M`Na2`1_M`Na2`1_M`Na2a1^M_Na2b1_M^Na2b1_M^Na2b1_M^Na2c1^M]Na2d1_M\\Na2d1_M\\Na2e1^M[Nb2e1^M[Na2f1_MZNa2f1_MZNa2g1^MYN`2i1`MWN_2j1aMVN^2k1bMUN\\2n1cMRN\\2o1dMQNZ2Q2fMoMY2R2gMnMW2U2hMkMW2V2iMjMV2W2jMiMU2X2kMhMU2Y2jMgMU2Z2kMfMU2Z2kMfMT2\\2kMdMU2\\2kMdMT2]2lMcMT2^2kMbMU2^2kMbMU2_2iMbMW2^2iMbMW2^2iMbMW2_2hMaMY2^2gMbMY2_2fMaM[2^2eMbM[2_2dMaM\\2_2dMaM]2^2cMbM]2_2bMaM_2^2aMbM_2_2`MaM`2_2`MaMa2^2_MbMa2_2^MaMc2^2]MbMc2\\2_MdMa2[2_MfMb2X2]MjMc2T2]MnMc2Q2\\MQNe2o1YMRNg2n1WMTNj2k1TMWNl2j1QMXNo2h1oLZNR3f1kL\\NU3d1iL^NX3a1gL`NY3a1dLaN\\3_1bLcN_3]1^LeNb3[1\\LgNe3X1YLjNg3X1ULjNk3W1SLjNn3V1oKlNQ4U1lKmNU4S1hKoNX4S1dKoN\\4R1aKPO`4P1]KROc4o0[KROf4n0WKTOi4l0UKVOk4j0SKXOn4h0oJZOQ5f0mJ\\OT5c0jJ_OV5a0eHQMm1_2^5`0cHTMl1^2b5=_HXMm1]2d5d0YJ^Og5b0WJ@j5`0SJBm5>QJDP6<nIER6;lIGT6:iIHX68eIJ[66cIL^64_INa62]I0c61[I0f6OXI3h6NUI4l6KRI7n6JoH8Q7[OZIg0g6^NQJd1o5bMSKV2n4oLTLP1gf6"}, "label": "the darker table on the right cell"}]}
{"id": 446864, "image": "COCO_train2014_000000446864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table on which two plates are plced\"."}], "task": "refering", "answer_template": "The \"the table on which two plates are plced\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 126, 127, 134, 135, 149, 150, 151, 152, 153, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 226, 227, 228, 229, 241, 242, 251, 252, 275, 287, 288, 297, 298, 310, 311, 312, 313, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.5014062499999999, 0.254765784114053, 0.978078125, 0.8659674134419553], "iscrowd": 0, "area": 37317.65495, "rle": {"size": [491, 640], "counts": "WVj4;Z>e0[Oe0\\Od0PDiNX9m1hFFo6o0SI_Oh5V1ZJiNP5m1PKSNk4Q2WKnMk4o1WKPNk4n1UKRNm4k1UKTNm4i1UKVNm4g1TKYNm4f1TKYNn4d1TK[Nn4b1SK^No4`1RK_NP5^1RKaNP5\\1QKdNQ5Y1QKfNQ5X1oJhNR5V1PKiNR5T1PKkNR5S1nJmNS5Q1oJnNR5P1PKoNP5Q1PKoNP5P1RKoNo4o0SKPOm4o0TKQOm4n0TKQOl4n0VKQOk4m0VKSOj4m0WKROj4l0XKSOh4l0YKTOg4k0[KTOf4k0[KTOe4k0\\KUOe4i0]KVOc4j0^KUOc4i0^KWOb4h0`KWOa4h0`KWO`4h0aKXO_4g0cKXO^4f0dKYO\\4g0dKYO]4e0eKZO[4e0fK[O[4d0fK[OZ4d0gK\\OZ4b0hK]OX4b0iK^OW4b0jK]OW4a0jK_OV4`0lK_OU4`0kK@U4?mK@T4>mKBS4>nKAS4=oKBQ4=PLCP4<RLCo3<QLDo3;SLDn3:SLFm3:TLEm39TLGl38VLGk37VLIj37WLHi37XLIi35YLJg36ZLIg35ZLKf34\\LKd35\\LKd34^LKc33^LMb32`LM`33`LM`32bLM^32cLN]32dLM]31dLO\\30fLOZ30gL0Y30hLOX30jLOV30kL0V3OkL0U3OlL1T3NnL1R3NoL2Q3NPM1Q3MPM3P3LRM3n2MRM3n2LTM3l2LUM4k2LVM3k2KVM5j2JXM5h2JZM5f2KZM5f2J[M6e2J\\M5e2J[M6e2J[M6e2KZM5f2K[M4e2L[M4f2LYM4g2LYM4g2LZM3f2NYM2g2NYM2g2NYM2h2NXM1h2OXM1h2OXM1h20WM0i20XMOh21XMOi21VMOj21VMOj21WMNi23VMMj23VMMk22UMNk22VMMj23VMMj22WMNi22WMNi22XMMi21XMOh21XMOh20YM0g20ZMOf21ZMOf21ZMOf21ZMOf22ZMMf23ZMMf24YMLf25ZMKf26ZMIf28YMHg28YMHf2:YMFg2:ZMEf2;ZMEf2<YMDf2=ZMCf2=[MBe2?ZMAf2?ZMAe2a0ZM_Of2a0[M^Oe2b0[M^Oe2c0ZM]Oe2d0[M\\Oe2e0[MZOe2f0[MZOe2f0[MZOd2h0[MXOe2h0\\MWOd2i0\\MWOd2j0[MVOd2k0\\MUOd2l0\\MSOd2m0\\MSOd2m0\\MSOc2o0\\MQOd2o0]MPOc2Q1\\MoNc2R1]MnNc2R1]MnNc2S1]MlNc2T1]MlNb2U1^MkNb2V1]MjNc2V1^MiNb2X1]MhNb2Y1^MgNb2Y1^MgNb2Y1_MfNa2[1^MeNa2\\1_MdNa2\\1_MdNa2\\1`McN`2^1_MbN`2_1`MaN`2_1`MaN`2_1_MbNa2_1^MaNa2`1_M`Na2`1_M`Na2`1_M`Na2a1^M_Na2b1_M^Na2b1_M^Na2b1_M^Na2c1^M]Na2d1_M\\Na2d1_M\\Na2e1^M[Nb2e1^M[Na2f1_MZNa2f1_MZNa2g1^MYN`2i1`MWN_2j1aMVN^2k1bMUN\\2n1cMRN\\2o1dMQNZ2Q2fMoMY2R2gMnMW2U2hMkMW2V2iMjMV2W2jMiMU2X2kMhMU2Y2jMgMU2Z2kMfMU2Z2kMfMT2\\2kMdMU2\\2kMdMT2]2lMcMT2^2kMbMU2^2kMbMU2_2iMbMW2^2iMbMW2^2iMbMW2_2hMaMY2^2gMbMY2_2fMaM[2^2eMbM[2_2dMaM\\2_2dMaM]2^2cMbM]2_2bMaM_2^2aMbM_2_2`MaM`2_2`MaMa2^2_MbMa2_2^MaMc2^2]MbMc2\\2_MdMa2[2_MfMb2X2]MjMc2T2]MnMc2Q2\\MQNe2o1YMRNg2n1WMTNj2k1TMWNl2j1QMXNo2h1oLZNR3f1kL\\NU3d1iL^NX3a1gL`NY3a1dLaN\\3_1bLcN_3]1^LeNb3[1\\LgNe3X1YLjNg3X1ULjNk3W1SLjNn3V1oKlNQ4U1lKmNU4S1hKoNX4S1dKoN\\4R1aKPO`4P1]KROc4o0[KROf4n0WKTOi4l0UKVOk4j0SKXOn4h0oJZOQ5f0mJ\\OT5c0jJ_OV5a0eHQMm1_2^5`0cHTMl1^2b5=_HXMm1]2d5d0YJ^Og5b0WJ@j5`0SJBm5>QJDP6<nIER6;lIGT6:iIHX68eIJ[66cIL^64_INa62]I0c61[I0f6OXI3h6NUI4l6KRI7n6JoH8Q7[OZIg0g6^NQJd1o5bMSKV2n4oLTLP1gf6"}, "label": "the table on which two plates are plced"}]}
{"id": 512400, "image": "COCO_train2014_000000512400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid wearing a blue hat holding a baseball bat\"."}], "task": "refering", "answer_template": "The \"a kid wearing a blue hat holding a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 104, 105, 106, 107, 108, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 179, 180, 181, 182, 183, 184, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 293, 294, 295, 296, 297, 298, 299, 300, 301, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 371, 372, 373, 374, 375, 376, 377, 378, 390, 391, 392, 393, 394, 395, 396, 397, 409, 410, 411, 412, 413, 414, 415, 416, 428, 429, 430, 431, 432, 433, 434, 435], "bbox": [0.21293577981651376, 0.192, 0.9488440366972477, 0.984], "iscrowd": 0, "area": 97400.30295000003, "rle": {"size": [640, 545], "counts": "djX24kc05J7J5J7I6K6I6K6H4L3M3M4L3N2N2N3M2M3N2N3L3N200O2O0O100O2O0O100O2O0O10001O0000010O0000001O0001O010O000010O0000010O0010O01O1O010O1O010O1O0010O01O1O010O0010O0001O00010O0000001O00001O01O0001O00001O00001O001N1O2N1O2N1O2N2N1O2N1O2O0O2N1O2N2N1O2N1O2N1O2O0O2N2N101N1O2N101N1O2N2N101N1O2N101N1fMVK]El4U:dKgE]4k9TLRFm3`9cL\\F_3^9iL^FX3b9kLZFW3d9mLYFS3g9PMUFR3i9QMTFP3k9SMRFo2m9TMPFl2o9WMnEk2Q:WMlEk2R:YMjEh2U:[MiEf2V:]MfEd2Y:c3O0O2O0O2N100O100O1O2O2M4WJ]Fo0e9lNnFc0U9XOQGe0Q9WOTGe0P9UOVGg0m8TOWGk0k8POXGo0k8lNXGT1k8fNXGY1k8cNWG\\1l8]NYGc1i8VN\\Gi1h8oM]GQ2e8hM`GW2c8aMcG_2_8ZMfGe2^8SMgGm2[8nLhGR3Z8jLhGV3Z8gLgGY3\\8bLfG^3\\8^LfGb3\\8ZLfGf3\\8VLfGj3\\8QLgGo3U<000000000000000000000000000O10000000000O10000000000O2O000000000O101O00001N1000001O000O2O00001O0O101O00001N10001O000O101O000kC[Lo7f3]GnLc8Q3\\GQMc8Q3YGRMf8S3TGPMj8T3PGoLo8V3kFlLT9Y3eFjLZ9Z3aFhL^9Z3_FiL_9Y3_FhL`9Z3]FhLc9Y3[FhLd9Z3YFhLf9Z3XFhLf9Z3XFgLg9[3VFgLi9[3UFfLj9\\3TFfLj9\\3TFeLk9]3SFdLl9^3QFdLo9]3oEeLo9]3oEdLP:g6O1O2N1O1O1O2N100O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2Nc0]OU1kN5K1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O2O1N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N3M2N2N2N3M2N2N2N2N3M2N2N3M8H7I8I7H8H7I8H8H5K3M3M4L6J5K5K6J7I6Jal`0"}, "label": "a kid wearing a blue hat holding a baseball bat"}]}
{"id": 512400, "image": "COCO_train2014_000000512400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy holding a baseball bat\"."}], "task": "refering", "answer_template": "The \"boy holding a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 104, 105, 106, 107, 108, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 179, 180, 181, 182, 183, 184, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 293, 294, 295, 296, 297, 298, 299, 300, 301, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 371, 372, 373, 374, 375, 376, 377, 378, 390, 391, 392, 393, 394, 395, 396, 397, 409, 410, 411, 412, 413, 414, 415, 416, 428, 429, 430, 431, 432, 433, 434, 435], "bbox": [0.21293577981651376, 0.192, 0.9488440366972477, 0.984], "iscrowd": 0, "area": 97400.30295000003, "rle": {"size": [640, 545], "counts": "djX24kc05J7J5J7I6K6I6K6H4L3M3M4L3N2N2N3M2M3N2N3L3N200O2O0O100O2O0O100O2O0O10001O0000010O0000001O0001O010O000010O0000010O0010O01O1O010O1O010O1O0010O01O1O010O0010O0001O00010O0000001O00001O01O0001O00001O00001O001N1O2N1O2N1O2N2N1O2N1O2O0O2N1O2N2N1O2N1O2N1O2O0O2N2N101N1O2N101N1O2N2N101N1O2N101N1fMVK]El4U:dKgE]4k9TLRFm3`9cL\\F_3^9iL^FX3b9kLZFW3d9mLYFS3g9PMUFR3i9QMTFP3k9SMRFo2m9TMPFl2o9WMnEk2Q:WMlEk2R:YMjEh2U:[MiEf2V:]MfEd2Y:c3O0O2O0O2N100O100O1O2O2M4WJ]Fo0e9lNnFc0U9XOQGe0Q9WOTGe0P9UOVGg0m8TOWGk0k8POXGo0k8lNXGT1k8fNXGY1k8cNWG\\1l8]NYGc1i8VN\\Gi1h8oM]GQ2e8hM`GW2c8aMcG_2_8ZMfGe2^8SMgGm2[8nLhGR3Z8jLhGV3Z8gLgGY3\\8bLfG^3\\8^LfGb3\\8ZLfGf3\\8VLfGj3\\8QLgGo3U<000000000000000000000000000O10000000000O10000000000O2O000000000O101O00001N1000001O000O2O00001O0O101O00001N10001O000O101O000kC[Lo7f3]GnLc8Q3\\GQMc8Q3YGRMf8S3TGPMj8T3PGoLo8V3kFlLT9Y3eFjLZ9Z3aFhL^9Z3_FiL_9Y3_FhL`9Z3]FhLc9Y3[FhLd9Z3YFhLf9Z3XFhLf9Z3XFgLg9[3VFgLi9[3UFfLj9\\3TFfLj9\\3TFeLk9]3SFdLl9^3QFdLo9]3oEeLo9]3oEdLP:g6O1O2N1O1O1O2N100O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2Nc0]OU1kN5K1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N1O2O1N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N3M2N2N2N3M2N2N2N2N3M2N2N3M8H7I8I7H8H7I8H8H5K3M3M4L6J5K5K6J7I6Jal`0"}, "label": "boy holding a baseball bat"}]}
{"id": 512400, "image": "COCO_train2014_000000512400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the young girl in the pink shirt and khaki pants\"."}], "task": "refering", "answer_template": "The \"the young girl in the pink shirt and khaki pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 148, 149, 150, 151, 167, 168, 169, 170, 185, 186, 187, 188, 189, 204, 205, 206, 207, 208, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 262, 263, 264, 265, 281, 282, 283, 284, 300, 301, 302, 303, 320, 321, 322, 339, 340, 341, 359, 360, 378, 379, 397, 398, 416, 417, 435, 436], "bbox": [0.7501834862385321, 0.262984375, 1.0, 0.98984375], "iscrowd": 0, "area": 29618.549399999996, "rle": {"size": [640, 545], "counts": "Tno74hc04L5K5K4K6K5K5K4L5K5K5J7J5K6J5M4LL5ZOf0CV@jNe=X1VBPOd=R1XBWOb=j0[BZOc=h0XB^Od=d0XBBe=?WBFe==XBHc=<XBId=:XBLc=g0iA^OR>\\3K4L5K5J5L5K4L5J5L5K4L5J6K4L5K4K6K4O2M2O2N1O2M3N1O2N1N3N1O2N1N3N2N1N3N1O2N1N3N2O1N3M2O2M3M2O2M2N3GUFXFl9g9UFWFn9g9TFVFo9h96O2N1O2N1O2N2N2N5K6J6J6J5K6J6J6J5J7J6J6J5K6J6J5K00O100Om1cElGU5V8gJUHo4n7nJ]Hh4f7SKeHd4]7XKoH]4T7`KVIk2TNmKi8S1_IU1ZOfMZ7R1fI^Oa0_Ol5o0cNoN`1n0bNPOb1k0`NUOb1h0^NXOe1d0]N[Of1a0\\N^Oh1>XNBk1:WNEl18UNGo14SNKP22PNNS2NoM1T2LmM4V2GkM8Y2DiM;Z2BgM<^2_OU9M\\E"}, "label": "the young girl in the pink shirt and khaki pants"}]}
{"id": 512400, "image": "COCO_train2014_000000512400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing a pink shirt watching a pitch being thrown\"."}], "task": "refering", "answer_template": "The \"a little girl wearing a pink shirt watching a pitch being thrown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 148, 149, 150, 151, 167, 168, 169, 170, 185, 186, 187, 188, 189, 204, 205, 206, 207, 208, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 262, 263, 264, 265, 281, 282, 283, 284, 300, 301, 302, 303, 320, 321, 322, 339, 340, 341, 359, 360, 378, 379, 397, 398, 416, 417, 435, 436], "bbox": [0.7501834862385321, 0.262984375, 1.0, 0.98984375], "iscrowd": 0, "area": 29618.549399999996, "rle": {"size": [640, 545], "counts": "Tno74hc04L5K5K4K6K5K5K4L5K5K5J7J5K6J5M4LL5ZOf0CV@jNe=X1VBPOd=R1XBWOb=j0[BZOc=h0XB^Od=d0XBBe=?WBFe==XBHc=<XBId=:XBLc=g0iA^OR>\\3K4L5K5J5L5K4L5J5L5K4L5J6K4L5K4K6K4O2M2O2N1O2M3N1O2N1N3N1O2N1N3N2N1N3N1O2N1N3N2O1N3M2O2M3M2O2M2N3GUFXFl9g9UFWFn9g9TFVFo9h96O2N1O2N1O2N2N2N5K6J6J6J5K6J6J6J5J7J6J6J5K6J6J5K00O100Om1cElGU5V8gJUHo4n7nJ]Hh4f7SKeHd4]7XKoH]4T7`KVIk2TNmKi8S1_IU1ZOfMZ7R1fI^Oa0_Ol5o0cNoN`1n0bNPOb1k0`NUOb1h0^NXOe1d0]N[Of1a0\\N^Oh1>XNBk1:WNEl18UNGo14SNKP22PNNS2NoM1T2LmM4V2GkM8Y2DiM;Z2BgM<^2_OU9M\\E"}, "label": "a little girl wearing a pink shirt watching a pitch being thrown"}]}
{"id": 512400, "image": "COCO_train2014_000000512400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white sweater watching children play a ballgame\"."}], "task": "refering", "answer_template": "The \"a woman in a white sweater watching children play a ballgame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 134, 135, 136, 153, 154, 155, 171, 172, 173, 174, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 228, 229, 230, 231, 232, 247, 248, 249, 250, 251, 267, 269, 270, 271, 288, 289, 290, 291, 306, 307, 309, 310, 325, 326, 327, 345, 346], "bbox": [0.007944954128440367, 0.28385937499999997, 0.36051376146788994, 0.7940156249999999], "iscrowd": 0, "area": 24112.311649999996, "rle": {"size": [640, 545], "counts": "bi2c0lb0g0A?E;E;E<D;D<J6J3M0O10001O00000O1O2N2N2N1O2N2N2N2N2N1O2N2N2L4K5K4L5K5K5H7I6K6I7N110O01000O01000O010O10O10000O01000O100000000001O001O00001O001O00001O01OjEdIh6\\6TIhIm6W6oHmIQ7T6jHPJX7n5dHVJ_7k3iE\\M`11i8]2kEbMS1:U9m1lEiMg0c0_9\\1oEQN;j0i9n0PFXN3n0P:c0QF_NJS1X:cMaEj1a0@BW1W;TO[EEYO]1];]N`DgNi0`1UO`1c;TNcDnNc0l4l:QLdDTO?k4n:lKgDZO:k4o:gKkD^O5k4Q;bKnDEOi4T;^KPEJKh4V;ZKSENFi4W;UKVE3Bh4Y;SKVE7_Of4\\;QKVE:\\Of4_;nJWE<YOg4`;kJXE?WOf4b;iJXEc0TOd4R=]KmBc4T=]KkBa3IdL\\=LjBR39RMl<MjBc2i0_M]<NjBU2Y1kMn;0hBg1i1XN_;1jBW1V2hNo:1mBl0]2SOf:2oBi0[2TOf:3QCi0W2TOi:3QCh0V2TOi:4TCg0R2TOj:5VCf0o1UOk:6WCe0m1TOl:7XCe0l1SOl:5]Ch0e1SOn:3`Cj0b1ROn:2dCl0\\1ROP;0gCn0Y1QOQ;NjCP1U1QOR;LlCT1P1POY;DlC[1l0oNR=R1mBmNU=R1kBmNW=S1iBkNX=U1hBjNZ=V1eBiN\\=W1dBhN]=Y1cBeN^=[1bBdN_=]1`BbNa=^1`B`Na=`1_B_Nb=b1]B\\Ne=d1\\21O0100O100O010O1N2O1O0O2O1N2O1O0O2O1O1N2O001N2O1OO010000O10000O10000O010O101O0O101N1O1O2N1O1O2N1O1O2M3M3M3Liei6"}, "label": "a woman in a white sweater watching children play a ballgame"}]}
{"id": 512400, "image": "COCO_train2014_000000512400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a white shirt and black leggings sitting\"."}], "task": "refering", "answer_template": "The \"a woman wearing a white shirt and black leggings sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 134, 135, 136, 153, 154, 155, 171, 172, 173, 174, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 228, 229, 230, 231, 232, 247, 248, 249, 250, 251, 267, 269, 270, 271, 288, 289, 290, 291, 306, 307, 309, 310, 325, 326, 327, 345, 346], "bbox": [0.007944954128440367, 0.28385937499999997, 0.36051376146788994, 0.7940156249999999], "iscrowd": 0, "area": 24112.311649999996, "rle": {"size": [640, 545], "counts": "bi2c0lb0g0A?E;E;E<D;D<J6J3M0O10001O00000O1O2N2N2N1O2N2N2N2N2N1O2N2N2L4K5K4L5K5K5H7I6K6I7N110O01000O01000O010O10O10000O01000O100000000001O001O00001O001O00001O01OjEdIh6\\6TIhIm6W6oHmIQ7T6jHPJX7n5dHVJ_7k3iE\\M`11i8]2kEbMS1:U9m1lEiMg0c0_9\\1oEQN;j0i9n0PFXN3n0P:c0QF_NJS1X:cMaEj1a0@BW1W;TO[EEYO]1];]N`DgNi0`1UO`1c;TNcDnNc0l4l:QLdDTO?k4n:lKgDZO:k4o:gKkD^O5k4Q;bKnDEOi4T;^KPEJKh4V;ZKSENFi4W;UKVE3Bh4Y;SKVE7_Of4\\;QKVE:\\Of4_;nJWE<YOg4`;kJXE?WOf4b;iJXEc0TOd4R=]KmBc4T=]KkBa3IdL\\=LjBR39RMl<MjBc2i0_M]<NjBU2Y1kMn;0hBg1i1XN_;1jBW1V2hNo:1mBl0]2SOf:2oBi0[2TOf:3QCi0W2TOi:3QCh0V2TOi:4TCg0R2TOj:5VCf0o1UOk:6WCe0m1TOl:7XCe0l1SOl:5]Ch0e1SOn:3`Cj0b1ROn:2dCl0\\1ROP;0gCn0Y1QOQ;NjCP1U1QOR;LlCT1P1POY;DlC[1l0oNR=R1mBmNU=R1kBmNW=S1iBkNX=U1hBjNZ=V1eBiN\\=W1dBhN]=Y1cBeN^=[1bBdN_=]1`BbNa=^1`B`Na=`1_B_Nb=b1]B\\Ne=d1\\21O0100O100O010O1N2O1O0O2O1N2O1O0O2O1O1N2O001N2O1OO010000O10000O10000O010O101O0O101N1O1O2N1O1O2N1O1O2M3M3M3Liei6"}, "label": "a woman wearing a white shirt and black leggings sitting"}]}
{"id": 496018, "image": "COCO_train2014_000000496018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald man wearing jeans and a dark shirt with an emblem on the back\"."}], "task": "refering", "answer_template": "The \"a bald man wearing jeans and a dark shirt with an emblem on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 179, 193, 194, 208, 209, 223, 224, 238, 239, 253, 254, 268, 269, 283, 284, 298, 299, 313, 314, 328, 329, 343, 344], "bbox": [0.8860187353629976, 0.45320312500000004, 0.9995316159250585, 0.974046875], "iscrowd": 0, "area": 10843.76065, "rle": {"size": [640, 427], "counts": "dW]73jc04k^OJV25U74YF5Z2L[71PFa0]2Ca7O\\EV1j2QOh7L\\EV1e2SOm7J\\EW1`2SOS8H[EY1[2ROY8G[EZ1V2SO^8EZE\\1Q2SOd8BZE_1k1SOj8_OYEb1f1TOo8\\OZEd1`1TOU9YOZEg1Z1TO[9WOYEh1V1UO`9TOYEk1o0UOg9ROYEm1g0WOo9mNYER2=WOY:hNYEV24XOb:cNXEd5g:k1N2O010N2N2N2M3N2N2M2O2N2M3N2N2N2N1O2O1N2N2N2O1N2N1O`L`3nJRL"}, "label": "a bald man wearing jeans and a dark shirt with an emblem on the back"}]}
{"id": 496018, "image": "COCO_train2014_000000496018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the far right who is cut off\"."}], "task": "refering", "answer_template": "The \"the man on the far right who is cut off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 179, 193, 194, 208, 209, 223, 224, 238, 239, 253, 254, 268, 269, 283, 284, 298, 299, 313, 314, 328, 329, 343, 344], "bbox": [0.8860187353629976, 0.45320312500000004, 0.9995316159250585, 0.974046875], "iscrowd": 0, "area": 10843.76065, "rle": {"size": [640, 427], "counts": "dW]73jc04k^OJV25U74YF5Z2L[71PFa0]2Ca7O\\EV1j2QOh7L\\EV1e2SOm7J\\EW1`2SOS8H[EY1[2ROY8G[EZ1V2SO^8EZE\\1Q2SOd8BZE_1k1SOj8_OYEb1f1TOo8\\OZEd1`1TOU9YOZEg1Z1TO[9WOYEh1V1UO`9TOYEk1o0UOg9ROYEm1g0WOo9mNYER2=WOY:hNYEV24XOb:cNXEd5g:k1N2O010N2N2N2M3N2N2M2O2N2M3N2N2N2N1O2O1N2N2N2O1N2N1O`L`3nJRL"}, "label": "the man on the far right who is cut off"}]}
{"id": 504211, "image": "COCO_train2014_000000504211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man playing in ground\"."}], "task": "refering", "answer_template": "The \"man playing in ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 338, 356, 357, 358], "bbox": [0.4916875, 0.574, 0.715359375, 0.9151666666666666], "iscrowd": 0, "area": 11500.994700000001, "rle": {"size": [480, 640], "counts": "TPd45f>6J6J5J7J6J5K6SDgNP:[1iEkNW:U1dEQO[:P1_EUO`:m0_ESOa:n0^ESOa:m0_ESO`:o0_EQOa:P1^EQOa:R1\\EnNc:V1ZEjNf:Y1VEiNi:Y1UEgNk:\\1REdNm:`1PEaNo:b1nD^NR;e1kD[NT;i1iDXNV;k1gDUNY;i2O2O000O2O001ZOXLQFi3n9[LoEe3Q:^LlEc3R:aLjE`3V:dLfE\\3Y:hLdEY3[:f0O101N1O1O2N100OoNoEVMo9l2TFPMk9R3WFiLk9W3YFaLj9a3YFXLi9i3[FPLg9R4b02O001000O100O01000O100O10O11N100O2O001M2O1aLlDP3V;jLoDU3R;gLQEX3];O1O001O1N101O001N101O001O1N101O001O0On0SO1O2N1N2O2N1O2I6K5J7I6K5J7I6KaV5M\\iJ<H7O10O0O1O2N100O2N1O1O2O0O2N1O101N1L_md2"}, "label": "man playing in ground"}]}
{"id": 504211, "image": "COCO_train2014_000000504211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the crouching guy with 55 in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the crouching guy with 55 in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 338, 356, 357, 358], "bbox": [0.4916875, 0.574, 0.715359375, 0.9151666666666666], "iscrowd": 0, "area": 11500.994700000001, "rle": {"size": [480, 640], "counts": "TPd45f>6J6J5J7J6J5K6SDgNP:[1iEkNW:U1dEQO[:P1_EUO`:m0_ESOa:n0^ESOa:m0_ESO`:o0_EQOa:P1^EQOa:R1\\EnNc:V1ZEjNf:Y1VEiNi:Y1UEgNk:\\1REdNm:`1PEaNo:b1nD^NR;e1kD[NT;i1iDXNV;k1gDUNY;i2O2O000O2O001ZOXLQFi3n9[LoEe3Q:^LlEc3R:aLjE`3V:dLfE\\3Y:hLdEY3[:f0O101N1O1O2N100OoNoEVMo9l2TFPMk9R3WFiLk9W3YFaLj9a3YFXLi9i3[FPLg9R4b02O001000O100O01000O100O10O11N100O2O001M2O1aLlDP3V;jLoDU3R;gLQEX3];O1O001O1N101O001N101O001O1N101O001O0On0SO1O2N1N2O2N1O2I6K5J7I6K5J7I6KaV5M\\iJ<H7O10O0O1O2N100O2N1O1O2O0O2N1O101N1L_md2"}, "label": "the crouching guy with 55 in the right hand picture"}]}
{"id": 504211, "image": "COCO_train2014_000000504211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coach of baseball players looking a players\"."}], "task": "refering", "answer_template": "The \"a coach of baseball players looking a players\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 314, 315, 316, 317, 337, 338, 339, 340, 360, 362, 363, 385, 386], "bbox": [0.636, 0.5356041666666667, 0.8039531249999999, 0.9969791666666665], "iscrowd": 0, "area": 13351.9819, "rle": {"size": [480, 640], "counts": "gSo51m>4A?C>F9I7K4K5O1N2_ESNm7o1_GQN_N6o9k1[GYN_N1T:g1WGQOf8Q1TGVOj8k0PG[On8g0hFCW9?`FH_9:XFNg94TF1i9d2APKlFR5Q9XKfFk4W9`0O1O1O1O1O1O1ON2N2M4M3L3N3\\Oc0M4L3N3M3N1O2O101N1000010O00101N1O1O101N1O5K5L9F10O10O1000O2O0O2O1N3N3L5K4M3L100O1O2O1O1O2N1O2N;E\\LkEm2d9RM^Fl2c9SM_Fk2b9TM`Fk2b9RM`Fl2e9oL]Fo2g9mL[FQ3i9dL`FZ3d:O1O1O1O1O1`MdD[1\\;cNiDY1X;eNmDW1V;fNnDV1U;bNTEZ1o:bNVEZ1m:cN[EU1i:fN`ER1_<H9H7ZOPUj1"}, "label": "a coach of baseball players looking a players"}]}
{"id": 504211, "image": "COCO_train2014_000000504211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire behind the catcher\"."}], "task": "refering", "answer_template": "The \"an umpire behind the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 314, 315, 316, 317, 337, 338, 339, 340, 360, 362, 363, 385, 386], "bbox": [0.636, 0.5356041666666667, 0.8039531249999999, 0.9969791666666665], "iscrowd": 0, "area": 13351.9819, "rle": {"size": [480, 640], "counts": "gSo51m>4A?C>F9I7K4K5O1N2_ESNm7o1_GQN_N6o9k1[GYN_N1T:g1WGQOf8Q1TGVOj8k0PG[On8g0hFCW9?`FH_9:XFNg94TF1i9d2APKlFR5Q9XKfFk4W9`0O1O1O1O1O1O1ON2N2M4M3L3N3\\Oc0M4L3N3M3N1O2O101N1000010O00101N1O1O101N1O5K5L9F10O10O1000O2O0O2O1N3N3L5K4M3L100O1O2O1O1O2N1O2N;E\\LkEm2d9RM^Fl2c9SM_Fk2b9TM`Fk2b9RM`Fl2e9oL]Fo2g9mL[FQ3i9dL`FZ3d:O1O1O1O1O1`MdD[1\\;cNiDY1X;eNmDW1V;fNnDV1U;bNTEZ1o:bNVEZ1m:cN[EU1i:fN`ER1_<H9H7ZOPUj1"}, "label": "an umpire behind the catcher"}]}
{"id": 504211, "image": "COCO_train2014_000000504211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the batter leaning forward at home plate\"."}], "task": "refering", "answer_template": "The \"the batter leaning forward at home plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 190, 191, 192, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 284, 285, 307, 308, 309], "bbox": [0.28209375, 0.2950416666666667, 0.47973437499999994, 0.7882916666666666], "iscrowd": 0, "area": 11202.74895, "rle": {"size": [480, 640], "counts": "QUe28_>9K5M3N3L3M3WNPO_ER1_:TO[Eo0b:WOXEl0e:[OREi0l:]OlDh0Q;^OgDg0V;E\\D`0c;^1O2O1O001O1O1O00[OnD^MQ;^2TEbMk:Z2ZEeMf:W2_EiM_:W2dEhM[:W2gEiMX:U2kEjMT:V2nEhMR:X2QFdMQ:[2Y11O4M3L4M2M4L4M2M4M3M3M2N3N1N2N2N2N2N2O1N2N2N2O2M;F<C<D=DbLjF[1l8cNZGb1^8\\NhGe1R8YNTHd1l7[NYHb1e7^N_H_1`7`NeH^1Z7aNkH[1T7dNQIY1o6fNVIW1h6iN^IR1a6mNgIm0X6SOoIh0o5XOXJb0f5^ObJ<]5DmJ3R5NVKJh47aKA^4`0eK]OZ4d0iKYOU4h0oKUOP4l0RLROm3n0WLoNg3S1\\LjNd3V1_LgN`3Y1cLeN]3[1dLdN[38\\H2[4DY38_H2Z4EV38cH2X4EU38fH1W4ER3:iH0V4EQ3:kHOW4En2;nHNU4Fm2;oHOT4Eo2:nH1T4Dn2;mH1V4Dn29mH3U4Dn28nH4T4Dn27nH6T4Co26mH6U4Co27lH6V4Bo27jH8W4Ao26kH9V4AP35jH9W4AP36hH:X4_OQ37gH:X4WOZ3=_H<^<DbC;`<@dC`0\\<YOlCf0W=0000000001O001O2N1O1O2N1O2J6Jolk4"}, "label": "the batter leaning forward at home plate"}]}
{"id": 504211, "image": "COCO_train2014_000000504211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the batter\"."}], "task": "refering", "answer_template": "The \"the batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 190, 191, 192, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 284, 285, 307, 308, 309], "bbox": [0.28209375, 0.2950416666666667, 0.47973437499999994, 0.7882916666666666], "iscrowd": 0, "area": 11202.74895, "rle": {"size": [480, 640], "counts": "QUe28_>9K5M3N3L3M3WNPO_ER1_:TO[Eo0b:WOXEl0e:[OREi0l:]OlDh0Q;^OgDg0V;E\\D`0c;^1O2O1O001O1O1O00[OnD^MQ;^2TEbMk:Z2ZEeMf:W2_EiM_:W2dEhM[:W2gEiMX:U2kEjMT:V2nEhMR:X2QFdMQ:[2Y11O4M3L4M2M4L4M2M4M3M3M2N3N1N2N2N2N2N2O1N2N2N2O2M;F<C<D=DbLjF[1l8cNZGb1^8\\NhGe1R8YNTHd1l7[NYHb1e7^N_H_1`7`NeH^1Z7aNkH[1T7dNQIY1o6fNVIW1h6iN^IR1a6mNgIm0X6SOoIh0o5XOXJb0f5^ObJ<]5DmJ3R5NVKJh47aKA^4`0eK]OZ4d0iKYOU4h0oKUOP4l0RLROm3n0WLoNg3S1\\LjNd3V1_LgN`3Y1cLeN]3[1dLdN[38\\H2[4DY38_H2Z4EV38cH2X4EU38fH1W4ER3:iH0V4EQ3:kHOW4En2;nHNU4Fm2;oHOT4Eo2:nH1T4Dn2;mH1V4Dn29mH3U4Dn28nH4T4Dn27nH6T4Co26mH6U4Co27lH6V4Bo27jH8W4Ao26kH9V4AP35jH9W4AP36hH:X4_OQ37gH:X4WOZ3=_H<^<DbC;`<@dC`0\\<YOlCf0W=0000000001O001O2N1O1O2N1O2J6Jolk4"}, "label": "the batter"}]}
{"id": 135577, "image": "COCO_train2014_000000135577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white cat laying down in the couch\"."}], "task": "refering", "answer_template": "The \"the white cat laying down in the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 95, 96, 97, 98, 99, 117, 118, 135, 136], "bbox": [0.2636, 0.352175, 0.6021, 0.51545], "iscrowd": 0, "area": 2376.4207000000006, "rle": {"size": [400, 500], "counts": "[gc13\\<2N100O01O010O1O010O1O100O10O0100O10O010O010O1O010O00100O010O010O10O0100O100N200O1000O0100O100O1O100O1O100O1O100O10O010001O000000001O1O1O100O00000000001O0000000000000000KXDBh;=ZDBf;>7O000001O0001O000001O01O00001O001O0O101N2O001O1O001O010O010O02O0O10001O0000001N2O1O3L4M?@3N0O1000O1O001O1O1O1O1O1O1O1O00000O101N10O0100O0_O^D6b;HbD6^;IeD5[;IiD5W;JkD5j;N001N1010OO2O0OZa]2"}, "label": "the white cat laying down in the couch"}]}
{"id": 135577, "image": "COCO_train2014_000000135577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white cat with back against sofa laying behind other white cat\"."}], "task": "refering", "answer_template": "The \"white cat with back against sofa laying behind other white cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 95, 96, 97, 98, 99, 117, 118, 135, 136], "bbox": [0.2636, 0.352175, 0.6021, 0.51545], "iscrowd": 0, "area": 2376.4207000000006, "rle": {"size": [400, 500], "counts": "[gc13\\<2N100O01O010O1O010O1O100O10O0100O10O010O010O1O010O00100O010O010O10O0100O100N200O1000O0100O100O1O100O1O100O1O100O10O010001O000000001O1O1O100O00000000001O0000000000000000KXDBh;=ZDBf;>7O000001O0001O000001O01O00001O001O0O101N2O001O1O001O010O010O02O0O10001O0000001N2O1O3L4M?@3N0O1000O1O001O1O1O1O1O1O1O1O00000O101N10O0100O0_O^D6b;HbD6^;IeD5[;IiD5W;JkD5j;N001N1010OO2O0OZa]2"}, "label": "white cat with back against sofa laying behind other white cat"}]}
{"id": 135577, "image": "COCO_train2014_000000135577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a striped head kitten\"."}], "task": "refering", "answer_template": "The \"a striped head kitten\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 157, 158, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213], "bbox": [0.5620599999999999, 0.6056, 0.9138, 0.821125], "iscrowd": 0, "area": 9096.273199999996, "rle": {"size": [400, 500], "counts": "cQ^31^<2N1O101N100O2N100O2N1O1I8L3M3M4O0O1O1O2O0O1CmN\\EU1b:mN]ES1c:nN\\ER1c:PO\\EP1d:ROZEo0e:ROZEn0f:SOYEm0g:UOWEk0j:>O0001O010O00001O010O00001O001O00001O0010O0001O001O001O01O001O0010O0001O001O010O00001N101N1O2O0O1O2N1O2N1O1O2O0010O010O0010HoMoEQ2Q:RNmEm1R:WNkEj1S::00O100O2N11O1O100O1O1O1O100O001O010O001O010O001O0010O01O0001O01O0000001O0O10001N10001N10000O2O00000O2N100O1O100O2N1O100O1O2N100O1N3N1N3N2N2M3N1N3N2M3N2N1N3N2Mm``0"}, "label": "a striped head kitten"}]}
{"id": 135577, "image": "COCO_train2014_000000135577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat sleeping near to the blue pillow\"."}], "task": "refering", "answer_template": "The \"cat sleeping near to the blue pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 157, 158, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213], "bbox": [0.5620599999999999, 0.6056, 0.9138, 0.821125], "iscrowd": 0, "area": 9096.273199999996, "rle": {"size": [400, 500], "counts": "cQ^31^<2N1O101N100O2N100O2N1O1I8L3M3M4O0O1O1O2O0O1CmN\\EU1b:mN]ES1c:nN\\ER1c:PO\\EP1d:ROZEo0e:ROZEn0f:SOYEm0g:UOWEk0j:>O0001O010O00001O010O00001O001O00001O0010O0001O001O001O01O001O0010O0001O001O010O00001N101N1O2O0O1O2N1O2N1O1O2O0010O010O0010HoMoEQ2Q:RNmEm1R:WNkEj1S::00O100O2N11O1O100O1O1O1O100O001O010O001O010O001O0010O01O0001O01O0000001O0O10001N10001N10000O2O00000O2N100O1O100O2N1O100O1O2N100O1N3N1N3N2N2M3N1N3N2M3N2N1N3N2Mm``0"}, "label": "cat sleeping near to the blue pillow"}]}
{"id": 135577, "image": "COCO_train2014_000000135577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cat lays on a couch and looks off into the distance\"."}], "task": "refering", "answer_template": "The \"a white cat lays on a couch and looks off into the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 111, 112, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.2013, 0.3957, 0.55226, 0.58065], "iscrowd": 0, "area": 6584.388399999998, "rle": {"size": [400, 500], "counts": "edW13[<5L3M4K4M4L3M4L4M2M3N2M3N2M10N1O2M2O2ZOe0010O0100O010O010O010O1O010O010O10O010O010O101N101N2O3L3N1N2O1N2O1N2O1O1N2O1N2O1O1N2O0O10O10000O0G:001O001O1O001O001O0O2O001O000000000000000000000O1000000001O1O1O1O1O1O10O00000000001O00000000001O0000001O00001O001N101O00001O001N101N1O101N101N1O2O0O21N1O10O01O01O01O00011N2N3N1N2N2O1N`0@2N101N1O7I8H1O1N2O1O2N1O1OPYg2"}, "label": "a white cat lays on a couch and looks off into the distance"}]}
{"id": 135577, "image": "COCO_train2014_000000135577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white cat that is closest to the edge of the couch\"."}], "task": "refering", "answer_template": "The \"the white cat that is closest to the edge of the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 111, 112, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 135], "bbox": [0.2013, 0.3957, 0.55226, 0.58065], "iscrowd": 0, "area": 6584.388399999998, "rle": {"size": [400, 500], "counts": "edW13[<5L3M4K4M4L3M4L4M2M3N2M3N2M10N1O2M2O2ZOe0010O0100O010O010O010O1O010O010O10O010O010O101N101N2O3L3N1N2O1N2O1N2O1O1N2O1N2O1O1N2O0O10O10000O0G:001O001O1O001O001O0O2O001O000000000000000000000O1000000001O1O1O1O1O1O10O00000000001O00000000001O0000001O00001O001N101O00001O001N101N1O101N101N1O2O0O21N1O10O01O01O01O00011N2N3N1N2N2O1N`0@2N101N1O7I8H1O1N2O1O2N1O1OPYg2"}, "label": "the white cat that is closest to the edge of the couch"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with its head over a fence in front of two other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe with its head over a fence in front of two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 151, 152, 153, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.0015, 0.37752336448598134, 0.793484375, 0.9887616822429908], "iscrowd": 0, "area": 54709.0279, "rle": {"size": [428, 640], "counts": "Zi0X1k:Y10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10001O000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O100000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10000O1O100O1O1O1O100O1O1O1O100O1O1O100O100O100O010O01O0010O01O001O010O001O010O1O0010O000000010O0000001O0000001O01O000001O0000001O0001O0001O000000010O01O001O010O001O0010O01O001O010O001O0010O01O001O010O1O001O010O001O1O010O001O00100O001O0010O01O1O0010O01O1O00O2O0O1O100O2N100N2O1N3mNQLfHQ4Y7XL^Hi3b7dLoG^3Q8n0010O1O100O1O100O1O100O101O100O1aKfGj3Q9M4L1O1000O0O2\\OQLhGQ4T8TLjGi3W8ZLeGf3[8_LaGa3_8bL]G`3c8e00O10000O10000001O1O0O20014M7HM3N2N3L3N4L5J6K5^OSG[LP9b3`0M2M4M2N2M2O1O1O1O1O1O001O1O1O1O1>BO1O001O1O1O1O1O1O001O1bMeEi1[:WNfE^10WN[::gEW1j:hNWEW1i:iNXEV1i:iNYEU1g:kNZET1g:kN[ES1j:hNWEW1`;O2N1O1N3O0O2N100O25J5K10N1O2M2O1O2N1O1O2N1O2_O`0O1O1O2M2O1O1N3N1N2O1Ohkf1"}, "label": "a giraffe with its head over a fence in front of two other giraffes"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe closest to camera\"."}], "task": "refering", "answer_template": "The \"the giraffe closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 151, 152, 153, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.0015, 0.37752336448598134, 0.793484375, 0.9887616822429908], "iscrowd": 0, "area": 54709.0279, "rle": {"size": [428, 640], "counts": "Zi0X1k:Y10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10001O000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O100000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10000O1O100O1O1O1O100O1O1O1O100O1O1O100O100O100O010O01O0010O01O001O010O001O010O1O0010O000000010O0000001O0000001O01O000001O0000001O0001O0001O000000010O01O001O010O001O0010O01O001O010O001O0010O01O001O010O1O001O010O001O1O010O001O00100O001O0010O01O1O0010O01O1O00O2O0O1O100O2N100N2O1N3mNQLfHQ4Y7XL^Hi3b7dLoG^3Q8n0010O1O100O1O100O1O100O101O100O1aKfGj3Q9M4L1O1000O0O2\\OQLhGQ4T8TLjGi3W8ZLeGf3[8_LaGa3_8bL]G`3c8e00O10000O10000001O1O0O20014M7HM3N2N3L3N4L5J6K5^OSG[LP9b3`0M2M4M2N2M2O1O1O1O1O1O001O1O1O1O1>BO1O001O1O1O1O1O1O001O1bMeEi1[:WNfE^10WN[::gEW1j:hNWEW1i:iNXEV1i:iNYEU1g:kNZET1g:kN[ES1j:hNWEW1`;O2N1O1N3O0O2N100O25J5K10N1O2M2O1O2N1O1O2N1O2_O`0O1O1O2M2O1O1N3N1N2O1Ohkf1"}, "label": "the giraffe closest to camera"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three giraffes getting snacks\"."}], "task": "refering", "answer_template": "The \"three giraffes getting snacks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 53, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 195, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256], "bbox": [0.0, 0.1163785046728972, 0.56409375, 0.8007710280373832], "iscrowd": 0, "area": 33688.587100000004, "rle": {"size": [428, 640], "counts": "V9a1j;01000O10000O01000O1000O0100O1000O0100O1O100O00100O1O1O010O1O100O00100O1O100O00100O1O1O010O1O100O00100O1O100O001O100O1O010O1O100O00100O1O100O001O100O1O010O1O100O00100O1O100O001O010O00010O00010O0010O0001O01O01O00010O000010O01O00010O000010O0001O01O01O0010O0001O01O01O00010O001O01O01O00010O000010O0001O010O1O1O100O001O100O1O1O100O1O1O100O1O1O10SOSF[Nl9f1XFUNh9m1[FoMd9R2`FjM`9V2dFfM[9[2iFaMV9`2j0100O2N1O100O1O1O100O2NYNdE`0[:@gE?Y:AiE=V:DkE;T:EnE:Q:GQF7o9IRF7l9JVF4i9LYF3g9M[F1d90]FOb92_FM`93cFK]95dFJ[97gFHW98kFEV9<lFAU9?lF_OT9b0nFZOS9g0nFWOd7]OeH]1HTO`7DgHX1KPO\\7MgHT1NmNW75iHo02iNR7=jHk05eNo6e0kHf08bNj6n0kHa0<_Ni6R1iH`0?[Nk6U1dHa0i8@VG`0k8@SGa0m8APG`0Q9@mFa0S9@lF`0U9@iFa0W9@gFa0Z9_OdFb0\\9@bF`0_9@_Fa0a9@]Fa0d9_OZFb0f9_OYFb0g9b1O2O0O101O0O100O2O0O101N100O01N2N3N1N3M2N3N1N3M3M2O2ML4I7J6I8H7F:G9N30O2N11O10O01O100O00100O00100O1O001001O2M3N2N3L3N2N3M2M3N3M1NJ7I7H7J7I7M201O1O001O1O0O2O1O001O1O001O11lKUHc2n7VMVHk2k7oLXHS3j7eL[H[3h7^L[Hd3g7UL]Hl3d7mKaHS4[82O0O01N1O2jMlF2V9KQGOP9OWGLj80^GJc84eGE]88jGBX8;oG@Q8>VH\\Ol7`0\\HZOe7f0`HTOb7j0dHQO]7m0a2O2O0O2N101N2N2N2O1N2N2N2O1N2N2O1N2N2N2O1Nf`d3"}, "label": "three giraffes getting snacks"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe who has its head up\"."}], "task": "refering", "answer_template": "The \"the giraffe who has its head up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 53, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 195, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256], "bbox": [0.0, 0.1163785046728972, 0.56409375, 0.8007710280373832], "iscrowd": 0, "area": 33688.587100000004, "rle": {"size": [428, 640], "counts": "V9a1j;01000O10000O01000O1000O0100O1000O0100O1O100O00100O1O1O010O1O100O00100O1O100O00100O1O1O010O1O100O00100O1O100O001O100O1O010O1O100O00100O1O100O001O100O1O010O1O100O00100O1O100O001O010O00010O00010O0010O0001O01O01O00010O000010O01O00010O000010O0001O01O01O0010O0001O01O01O00010O001O01O01O00010O000010O0001O010O1O1O100O001O100O1O1O100O1O1O100O1O1O10SOSF[Nl9f1XFUNh9m1[FoMd9R2`FjM`9V2dFfM[9[2iFaMV9`2j0100O2N1O100O1O1O100O2NYNdE`0[:@gE?Y:AiE=V:DkE;T:EnE:Q:GQF7o9IRF7l9JVF4i9LYF3g9M[F1d90]FOb92_FM`93cFK]95dFJ[97gFHW98kFEV9<lFAU9?lF_OT9b0nFZOS9g0nFWOd7]OeH]1HTO`7DgHX1KPO\\7MgHT1NmNW75iHo02iNR7=jHk05eNo6e0kHf08bNj6n0kHa0<_Ni6R1iH`0?[Nk6U1dHa0i8@VG`0k8@SGa0m8APG`0Q9@mFa0S9@lF`0U9@iFa0W9@gFa0Z9_OdFb0\\9@bF`0_9@_Fa0a9@]Fa0d9_OZFb0f9_OYFb0g9b1O2O0O101O0O100O2O0O101N100O01N2N3N1N3M2N3N1N3M3M2O2ML4I7J6I8H7F:G9N30O2N11O10O01O100O00100O00100O1O001001O2M3N2N3L3N2N3M2M3N3M1NJ7I7H7J7I7M201O1O001O1O0O2O1O001O1O001O11lKUHc2n7VMVHk2k7oLXHS3j7eL[H[3h7^L[Hd3g7UL]Hl3d7mKaHS4[82O0O01N1O2jMlF2V9KQGOP9OWGLj80^GJc84eGE]88jGBX8;oG@Q8>VH\\Ol7`0\\HZOe7f0`HTOb7j0dHQO]7m0a2O2O0O2N101N2N2N2O1N2N2N2O1N2N2O1N2N2N2O1Nf`d3"}, "label": "the giraffe who has its head up"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zookeeper wearing all khaki colors and a hat\"."}], "task": "refering", "answer_template": "The \"a zookeeper wearing all khaki colors and a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 85, 86, 87, 107, 108, 109, 110, 111, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 271, 272, 273, 296], "bbox": [0.6777656249999999, 0.16628504672897196, 0.9527812499999999, 0.8449532710280375], "iscrowd": 0, "area": 26078.02029999999, "rle": {"size": [428, 640], "counts": "Pce51Z=1O2N3M2N3aMGfG;X8HeG;Y8HcG:[8IbG:[8JaG8^8FcG=Z8BgG`0X8_OhGc0V8\\OkGg0R8ZOmGh0R8[OjGh0S8[OjGg0T8]OhGf0V8]OfGe0W8_OfGd0U8BgG`0U8FgG=T8HiG:S8JjG:R8JkG8Q8MlG6o7OnG3n7ORH4i7MXH5d7M\\H6b7I^H9`7FaH=]7BcH`0\\7_OcHd0[7[OfHh0X7WOhHk0W7WOfHl0W7WOeHl0Z7VOcHm0[7UOaHn0^7SO_HQ1^7RO_HP1`7`2N2M5L4L4L4K4M2N2M3N1O2M3N2OO1O1O001O1O1O1N2O1O001;EO1O1O1O1O010O12N5L<C10N1N3N2N1\\OXJbIi5]6d0N1`J\\IX4e6fKgIn3[6PLfIP4[6nKgIP4Z6oKgIQ4Z6mKgIR4[6lKfIT4[6jKgIU4Y6jKhIW4X6fKjIZ4W6cKlI]4T6`KnI`4S6\\KPJe4o5XKTJh4m5TKWJl4T700a0@1N1O2N101N1O10M2O1N2M4L3M3M4L3M3M3M4N11O0010000O0100000O010[H[Kb6e4UIdKj6]4mHlKS7S4eHVL[7m410O10000O01O1N2TNcH^LIm0e7d2PIZMQ7e2QIXMQ7f2RIWMP7h2RIVMo6h2TIUMo6i2SIPMT7m2oHkLY7R3jHeL_7Y3\\1M3N2M3N2N2N2N2N8H8H7I8G7J3M4L3M3M5K6J6J6IbX<"}, "label": "a zookeeper wearing all khaki colors and a hat"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is feeding giraffe\"."}], "task": "refering", "answer_template": "The \"a man is feeding giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 85, 86, 87, 107, 108, 109, 110, 111, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 271, 272, 273, 296], "bbox": [0.6777656249999999, 0.16628504672897196, 0.9527812499999999, 0.8449532710280375], "iscrowd": 0, "area": 26078.02029999999, "rle": {"size": [428, 640], "counts": "Pce51Z=1O2N3M2N3aMGfG;X8HeG;Y8HcG:[8IbG:[8JaG8^8FcG=Z8BgG`0X8_OhGc0V8\\OkGg0R8ZOmGh0R8[OjGh0S8[OjGg0T8]OhGf0V8]OfGe0W8_OfGd0U8BgG`0U8FgG=T8HiG:S8JjG:R8JkG8Q8MlG6o7OnG3n7ORH4i7MXH5d7M\\H6b7I^H9`7FaH=]7BcH`0\\7_OcHd0[7[OfHh0X7WOhHk0W7WOfHl0W7WOeHl0Z7VOcHm0[7UOaHn0^7SO_HQ1^7RO_HP1`7`2N2M5L4L4L4K4M2N2M3N1O2M3N2OO1O1O001O1O1O1N2O1O001;EO1O1O1O1O010O12N5L<C10N1N3N2N1\\OXJbIi5]6d0N1`J\\IX4e6fKgIn3[6PLfIP4[6nKgIP4Z6oKgIQ4Z6mKgIR4[6lKfIT4[6jKgIU4Y6jKhIW4X6fKjIZ4W6cKlI]4T6`KnI`4S6\\KPJe4o5XKTJh4m5TKWJl4T700a0@1N1O2N101N1O10M2O1N2M4L3M3M4L3M3M3M4N11O0010000O0100000O010[H[Kb6e4UIdKj6]4mHlKS7S4eHVL[7m410O10000O01O1N2TNcH^LIm0e7d2PIZMQ7e2QIXMQ7f2RIWMP7h2RIVMo6h2TIUMo6i2SIPMT7m2oHkLY7R3jHeL_7Y3\\1M3N2M3N2N2N2N2N8H8H7I8G7J3M4L3M3M5K6J6J6IbX<"}, "label": "a man is feeding giraffe"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl who is standing next to forest guard is giving food to giraffe\"."}], "task": "refering", "answer_template": "The \"a little girl who is standing next to forest guard is giving food to giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 205, 206, 229, 252, 274, 275, 297, 298, 319, 321, 342, 344], "bbox": [0.894671875, 0.5056308411214954, 1.0, 0.9876168224299066], "iscrowd": 0, "area": 3418.5263499999974, "rle": {"size": [428, 640], "counts": "Ql_7`0l<00O1000O010000O100O3N7IUi5g0PVJ4K5L400O2O0O1dMjNaHW1^7oN]HQ1a7UOZHl0e7ZOUHg0j7_OQHa0m7EnG<Q8GYHMf77nHROP7S1b23M4L3M3M3M3M2N3M2N2XEgMZ:Z2eEiMY:X2eElMX:U2gEnMV:R2iE^Nh9c1VFmN[9S1dF\\On8e0QG[On0YNn6X4RIgLo5Y3QJgLo5Y3PJhLP6X3PJhLo5X3RJhLc7c1fI"}, "label": "a little girl who is standing next to forest guard is giving food to giraffe"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl with black straight hair\"."}], "task": "refering", "answer_template": "The \"a young girl with black straight hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 205, 206, 229, 252, 274, 275, 297, 298, 319, 321, 342, 344], "bbox": [0.894671875, 0.5056308411214954, 1.0, 0.9876168224299066], "iscrowd": 0, "area": 3418.5263499999974, "rle": {"size": [428, 640], "counts": "Ql_7`0l<00O1000O010000O100O3N7IUi5g0PVJ4K5L400O2O0O1dMjNaHW1^7oN]HQ1a7UOZHl0e7ZOUHg0j7_OQHa0m7EnG<Q8GYHMf77nHROP7S1b23M4L3M3M3M3M2N3M2N2XEgMZ:Z2eEiMY:X2eElMX:U2gEnMV:R2iE^Nh9c1VFmN[9S1dF\\On8e0QG[On0YNn6X4RIgLo5Y3QJgLo5Y3PJhLP6X3PJhLo5X3RJhLc7c1fI"}, "label": "a young girl with black straight hair"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle zebra\"."}], "task": "refering", "answer_template": "The \"middle zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259], "bbox": [0.13315625, 0.4655373831775701, 0.46828125, 0.7720327102803739], "iscrowd": 0, "area": 10403.594199999998, "rle": {"size": [428, 640], "counts": "aXT11[=0O1O100O1O1O010O1O100O1O1O010O1O100O1O00100O1O100O1O1O010O1O100O1O1O010O1O100O1O1O10O01O100O1O1O10O01O100O1O1O100O00100O1O1O100O0100000O10000000O100000000000000O100000ROnNbER1^:PO`Eo0a:TO\\El0e:UOYEk0g:XOVEh0j:ZOTEf0l:]OQEb0P;AmD?S;CkD=V;EgD;Y;GeD8\\;l000000O2O00000O[NVNjGj1V8XNhGh1X8ZNgGe1Y8\\NfGc1[8_NcGa1]8`NbG`1^8bN`G^1`8cN_G]1a8eN^GY1c8hN\\GX1d8jNYGW1g8jNWGW1i8kNTGU1m8lNQGU1n8nNoFS1Q9nNmFS1S9oNjFQ1W9POgFQ1Y9QOdFP1\\9QObFP1^9RO_Fn0b9SO\\Fn0d9TOXFn0h9SOVFn0j9TOSFl0n9Y10000000O1000000O10O1000O1000000O2O2N1O2N1N3N1O2N1N3N2N1O2M2O2N1N3N1N3M2N3N1N3M3N1N3M2N101N1O1O100O1O100O1O1O101N1O1O100O1O100O1O1O10O01O0010O01O101N2N2N2O1N3M3M4M4K4L4M[k]4"}, "label": "middle zebra"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black tipped furry horns on giraffe head\"."}], "task": "refering", "answer_template": "The \"black tipped furry horns on giraffe head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259], "bbox": [0.13315625, 0.4655373831775701, 0.46828125, 0.7720327102803739], "iscrowd": 0, "area": 10403.594199999998, "rle": {"size": [428, 640], "counts": "aXT11[=0O1O100O1O1O010O1O100O1O1O010O1O100O1O00100O1O100O1O1O010O1O100O1O1O010O1O100O1O1O10O01O100O1O1O10O01O100O1O1O100O00100O1O1O100O0100000O10000000O100000000000000O100000ROnNbER1^:PO`Eo0a:TO\\El0e:UOYEk0g:XOVEh0j:ZOTEf0l:]OQEb0P;AmD?S;CkD=V;EgD;Y;GeD8\\;l000000O2O00000O[NVNjGj1V8XNhGh1X8ZNgGe1Y8\\NfGc1[8_NcGa1]8`NbG`1^8bN`G^1`8cN_G]1a8eN^GY1c8hN\\GX1d8jNYGW1g8jNWGW1i8kNTGU1m8lNQGU1n8nNoFS1Q9nNmFS1S9oNjFQ1W9POgFQ1Y9QOdFP1\\9QObFP1^9RO_Fn0b9SO\\Fn0d9TOXFn0h9SOVFn0j9TOSFl0n9Y10000000O1000000O10O1000O1000000O2O2N1O2N1N3N1O2N1N3N2N1O2M2O2N1N3N1N3M2N3N1N3M3N1N3M2N101N1O1O100O1O100O1O1O101N1O1O100O1O100O1O1O10O01O0010O01O101N2N2N2O1N3M3M4M4K4L4M[k]4"}, "label": "black tipped furry horns on giraffe head"}]}
{"id": 94618, "image": "COCO_train2014_000000094618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a head of a giraffe behind a zookeeper\"."}], "task": "refering", "answer_template": "The \"a head of a giraffe behind a zookeeper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 173, 174], "bbox": [0.5173125, 0.17936915887850466, 0.74971875, 0.5336214953271028], "iscrowd": 0, "area": 6128.223099999999, "rle": {"size": [428, 640], "counts": "SbZ41Y=2O1O1N3N1O1N3N1O1N2O2O0O1O1O2O0O1O1O100O001LVOeCk0[<UOeCj0[<WOdCj0[<4O001O10O01O001O1O010O1O001O00O2N1O1O00O1O1N2O1NWOQD;n;GYD1h;O^DJd;4cDEb:1aE9X;KfD3Z;1cDM_;7]DHd;;ZDDf;?WDAi;a0UD_Ok;a0VD^Oj;c0UD^Oj;c0UD]Ol;c0TD\\Ol;e0SD[Om;f0RDZOn;g0QDYOo;h0oCCg;>XDJ`;7_D2X;OgD5U;Q100010N1O1N3N1]OPEhNS;T1d0N2N3L301O0O1011N2O0O2O1O1O001O100M2N3M3M3L3N2N000O2O000O2hNVNkFl1T9eNZF^1d9U1N2012M2N3N1N3N1N3N1N3N1N3N1N3N1nMWE^1l:WN_Eg1U;M3N2M3M3M3N2M3M3M3N2M3M3M3M4M2M4L3M4MQgR2"}, "label": "a head of a giraffe behind a zookeeper"}]}
{"id": 242076, "image": "COCO_train2014_000000242076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man brushes his teeth with a ring on his ring finger\"."}], "task": "refering", "answer_template": "The \"a man brushes his teeth with a ring on his ring finger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 24, 25, 37, 38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 152], "bbox": [0.0068, 0.015714285714285715, 0.49494, 0.9887698412698414], "iscrowd": 0, "area": 48483.788349999995, "rle": {"size": [252, 500], "counts": "[o02U7e0M3O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1gJnNg3S1VLVOc3k0YL^Ob3b0[LG^3:_LOZ32cL7V3JfL`0S3AjLg0P3ZOmLo0l2ROQMV1i2kNTM[1h2fNVM^1g2cNYM`1c2aN]Mb1`2^N`Md1]2]NcMf1Y2[NgMg1V2ZNjMi1R2XNnMk1n1VNRNl1k1UNUNn1g1SNYNP2c1QN]NQ2`1PN`NS2\\1nMdNU2Y1kMgNW2V1jMjNY2R1hMnN[2n0fMRO\\2k0eMUO^2g0cMYO`2c0aM]Oa2`0`M@c2<^MDe28\\MHf25[MKh22XMNk2NVM2l2LTM4Z50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000005K?A`0@8HO1N2O1O1O1O1O1O1O]KmN_2R1`MQO_2n0aMSO_2m0`MTO`2m0^MUOa2l0]MUOc2l0XMYOg2j0]LOc3U201O001O00001O001O00001O001O00001O001O00001O001O00001O001ObTn1"}, "label": "a man brushes his teeth with a ring on his ring finger"}]}
{"id": 242076, "image": "COCO_train2014_000000242076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the picture of the man with the toothbrush sideways\"."}], "task": "refering", "answer_template": "The \"the picture of the man with the toothbrush sideways\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 24, 25, 37, 38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 152], "bbox": [0.0068, 0.015714285714285715, 0.49494, 0.9887698412698414], "iscrowd": 0, "area": 48483.788349999995, "rle": {"size": [252, 500], "counts": "[o02U7e0M3O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1gJnNg3S1VLVOc3k0YL^Ob3b0[LG^3:_LOZ32cL7V3JfL`0S3AjLg0P3ZOmLo0l2ROQMV1i2kNTM[1h2fNVM^1g2cNYM`1c2aN]Mb1`2^N`Md1]2]NcMf1Y2[NgMg1V2ZNjMi1R2XNnMk1n1VNRNl1k1UNUNn1g1SNYNP2c1QN]NQ2`1PN`NS2\\1nMdNU2Y1kMgNW2V1jMjNY2R1hMnN[2n0fMRO\\2k0eMUO^2g0cMYO`2c0aM]Oa2`0`M@c2<^MDe28\\MHf25[MKh22XMNk2NVM2l2LTM4Z50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000005K?A`0@8HO1N2O1O1O1O1O1O1O]KmN_2R1`MQO_2n0aMSO_2m0`MTO`2m0^MUOa2l0]MUOc2l0XMYOg2j0]LOc3U201O001O00001O001O00001O001O00001O001O00001O001O00001O001ObTn1"}, "label": "the picture of the man with the toothbrush sideways"}]}
{"id": 242076, "image": "COCO_train2014_000000242076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man brushing his teeth with the toothbrush straight instead of horizontal\"."}], "task": "refering", "answer_template": "The \"the man brushing his teeth with the toothbrush straight instead of horizontal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 51, 52, 64, 65, 66, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 89, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161], "bbox": [0.50544, 0.01146825396825397, 0.9904400000000001, 0.9870238095238095], "iscrowd": 0, "area": 46670.664250000016, "rle": {"size": [252, 500], "counts": "o_n15e72O1O1N2O1N2O1O1N2O1N2O1N2O1O1N200O1hJTO_3m0WLAd3`0RLNi33mK:o3GoK?m3AQLe0j3\\OSLi0j3XOTLm0h3TOVLP1g3QOWLT1e3mNYLW1d3jNZL[1c3eN\\L_1`3bN`L`1]3aNcLb1Y3_NgLc1V3^NjLe1R3\\NnLf1o2[NQMh1l2XNTMj1i2WNWMl1e2UN[Mm1b2TN^Mo1^2RNbMP2[2QNeMR2W2oMiMS2U2mMkMV2Q2kMoMW2n1jMRNY2j1hMVNZ2h1fMXN\\2e1eM[N]2b1dM^N^2`1bM`N`2]1aMcN`2[1aMeNa2Y1_MgNc2V1^MjNd2S1]MmNe2Q1[MoNf2o0[MQOg2m0YMSOi2j0XMVOj2g0WMYOj2f0VMZOl2c0UM]Om2`0TM@n2>RMBP3;QMEP39QMGQ37oLIS34nLLT31mLOT3OmL1S3NnL2R3MoL3Q3LPM4P3KQM5[5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O4M3M2N3M3M2N3M3M2N3MM3M3N2O1O100001O001O003M3M3M3M3M4gJdNj3b1oKbNn3c1lKaNQ4d1hK`NU4h1`K_N]4c2M3M2N1O001O1O1O001O00001O001O00001O00001O001O00001O001N101RLXMf2`V1"}, "label": "the man brushing his teeth with the toothbrush straight instead of horizontal"}]}
{"id": 373153, "image": "COCO_train2014_000000373153.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow about to finish crossing the road\"."}], "task": "refering", "answer_template": "The \"a brown cow about to finish crossing the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 241, 242, 243, 259, 260, 261, 265, 266, 267, 290], "bbox": [0.286359375, 0.33831249999999996, 0.7309687500000001, 0.7146666666666666], "iscrowd": 0, "area": 28594.954549999995, "rle": {"size": [480, 640], "counts": "cRf2<_>7aN^OXDi0e;XOSDP1k;ROmCV1Q<lNeC^1Y<f0N1O1O2N1hMeMjG\\2U8oM`GR2_8QN]GQ2b8QN\\G`1fNTNm9>WG^1]9cN[Fa1h9dNeEf1_:Y14L4M3N2O1O10001O100O0012N3M2N3M2N3M3M3L4M3N2M2N3M3M2N3M2N2N3N0O1O1O1O1O001OnN^FgLb9m2kFRMT9j2RGUMm8h2XGVMh8g2]GWMc8g2aGWM_8f2fGXMZ8e2lGXMT8f2PHXMQ8d2THZMl7d2WH[Mi7e2XHZMh7e2ZHZMf7f2ZHZMf7f2[HXMf7h2[HWMe7i2\\HUMe7k2Y20O0000001O001O1O10O01O4L1O001O0000O100O100O1000bDoLm:`32N2N2N2N2N2O1N20000O2O000O2O0000001O0000001OO100000001O001O001O00000001O01O100O001O100O1O1O100O1UEgLm9V410O1O1O1O1O1O000000000000WOaEQM_:l2lElLT:Q3TFjLl9T3XFjLh9U3[FiLe9V3^FhLb9V3U11O1O1O101N2N2N2N3M2N2N2N2O0O1O1O101N1O2N1O101N1O2N101N1O1O100O2N1O1O10oMXFiNg9V1jF[NU9e1mF[NQ9d1RG[Nn8d1TG\\Nj8d1YGZNf8e1]GZNb8f1`GXNa8f1bGYN]8f1gGWNY8h1jGWNU8g1PHVNQ8g1THWNk7h1WHWNi7g1ZHYNe7f1]HYNd7e1^HZNb7d1aHZN`7e1bHZN^7g1bHXN^7h1cHUN_7k1k21O0O1O2N1O101O00001O0O10001O000000000010O0000001O0000001O0000001O0000001O000000000O10O1O10RNVCf1i<XN[Cf1f<YN\\Cf1c<[N^Cd1b<[N`Cc1a<]NaCa1_<^NcC`1^<_NeC^1Q=N2M3N2M2O2B>E>GX]`2"}, "label": "a brown cow about to finish crossing the road"}]}
{"id": 373153, "image": "COCO_train2014_000000373153.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shaggy brown cow\"."}], "task": "refering", "answer_template": "The \"a shaggy brown cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 241, 242, 243, 259, 260, 261, 265, 266, 267, 290], "bbox": [0.286359375, 0.33831249999999996, 0.7309687500000001, 0.7146666666666666], "iscrowd": 0, "area": 28594.954549999995, "rle": {"size": [480, 640], "counts": "cRf2<_>7aN^OXDi0e;XOSDP1k;ROmCV1Q<lNeC^1Y<f0N1O1O2N1hMeMjG\\2U8oM`GR2_8QN]GQ2b8QN\\G`1fNTNm9>WG^1]9cN[Fa1h9dNeEf1_:Y14L4M3N2O1O10001O100O0012N3M2N3M2N3M3M3L4M3N2M2N3M3M2N3M2N2N3N0O1O1O1O1O001OnN^FgLb9m2kFRMT9j2RGUMm8h2XGVMh8g2]GWMc8g2aGWM_8f2fGXMZ8e2lGXMT8f2PHXMQ8d2THZMl7d2WH[Mi7e2XHZMh7e2ZHZMf7f2ZHZMf7f2[HXMf7h2[HWMe7i2\\HUMe7k2Y20O0000001O001O1O10O01O4L1O001O0000O100O100O1000bDoLm:`32N2N2N2N2N2O1N20000O2O000O2O0000001O0000001OO100000001O001O001O00000001O01O100O001O100O1O1O100O1UEgLm9V410O1O1O1O1O1O000000000000WOaEQM_:l2lElLT:Q3TFjLl9T3XFjLh9U3[FiLe9V3^FhLb9V3U11O1O1O101N2N2N2N3M2N2N2N2O0O1O1O101N1O2N1O101N1O2N101N1O1O100O2N1O1O10oMXFiNg9V1jF[NU9e1mF[NQ9d1RG[Nn8d1TG\\Nj8d1YGZNf8e1]GZNb8f1`GXNa8f1bGYN]8f1gGWNY8h1jGWNU8g1PHVNQ8g1THWNk7h1WHWNi7g1ZHYNe7f1]HYNd7e1^HZNb7d1aHZN`7e1bHZN^7g1bHXN^7h1cHUN_7k1k21O0O1O2N1O101O00001O0O10001O000000000010O0000001O0000001O0000001O0000001O000000000O10O1O10RNVCf1i<XN[Cf1f<YN\\Cf1c<[N^Cd1b<[N`Cc1a<]NaCa1_<^NcC`1^<_NeC^1Q=N2M3N2M2O2B>E>GX]`2"}, "label": "a shaggy brown cow"}]}
{"id": 373153, "image": "COCO_train2014_000000373153.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car closest to the camera\"."}], "task": "refering", "answer_template": "The \"the car closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 368, 369, 370], "bbox": [0.0016875000000000002, 0.6292083333333333, 0.286515625, 0.9865208333333333], "iscrowd": 0, "area": 20018.69495, "rle": {"size": [480, 640], "counts": "Qm0i0e<b1_Na1cN]10O10MiJ]FW5c9iJ]FW5c9300001O0000000000000000001O000000000000000000001O000000000000000000001O0000000000000O10001O000O1M3L4M3O1N2O1N2O2M2O1N2O1N2O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1O10000O1000000000000000000O1000000000000000001O0O1000001O1O001O1O1O0O2O1O001O1O001O1O001N2O002N2N2N2N2N2N2N2M3N2N2N4L3M4L3M3M4L3M3L5L2N1O1O2N1O1O1O1O2M2O1O1O1O2N1O1O1O1O1N2O1O1O2N1O1O1O1O1O1N2O1OTje6"}, "label": "the car closest to the camera"}]}
{"id": 373153, "image": "COCO_train2014_000000373153.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hood of a car with a large animal to the right of it\"."}], "task": "refering", "answer_template": "The \"the hood of a car with a large animal to the right of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 368, 369, 370], "bbox": [0.0016875000000000002, 0.6292083333333333, 0.286515625, 0.9865208333333333], "iscrowd": 0, "area": 20018.69495, "rle": {"size": [480, 640], "counts": "Qm0i0e<b1_Na1cN]10O10MiJ]FW5c9iJ]FW5c9300001O0000000000000000001O000000000000000000001O000000000000000000001O0000000000000O10001O000O1M3L4M3O1N2O1N2O2M2O1N2O1N2O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1O10000O1000000000000000000O1000000000000000001O0O1000001O1O001O1O1O0O2O1O001O1O001O1O001N2O002N2N2N2N2N2N2N2M3N2N2N4L3M4L3M3M4L3M3L5L2N1O1O2N1O1O1O1O2M2O1O1O1O2N1O1O1O1O1N2O1O1O2N1O1O1O1O1O1N2O1OTje6"}, "label": "the hood of a car with a large animal to the right of it"}]}
{"id": 45475, "image": "COCO_train2014_000000045475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stationary black motorcycle ridden by someone wearing dark boots and clothing\"."}], "task": "refering", "answer_template": "The \"stationary black motorcycle ridden by someone wearing dark boots and clothing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 166, 167, 168, 169, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 328, 329, 332, 333, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362, 383], "bbox": [0.210671875, 0.39789583333333334, 0.805625, 0.9462083333333333], "iscrowd": 0, "area": 61075.81174999999, "rle": {"size": [480, 640], "counts": "nco1:_>8N2M2O2M3[NWOPEk0[:KaE8[:N`E5\\:0aE2[:3aE0[:5bEMZ:9aEI]:;]EJ_:;ZEKc:9UENh:W2M3M3L4M3M2N2N2N2N3N1O1O1O2N1O1O1O1O2N1N2O1O1O2N1O2N2N2N2M4M2O1O1O1O2N1O1O1O1O2N001O001O1O001O001O001O001O001O001O0O2OfNfGoKY8o3kGoKU8P4mGPLR8o3PHPLP8o3a1O1O1O001O1O1O1O1O1O1ZOf0N101bMmEMU:ORFMn90XFMi9O\\FNf9N`FO`9NeFOY<M2M4M2M3N3L3NS\\3B[dL1O1O1M4K4M3M3L4M4L3L4M3L4M4L3L4M3M3L4M3M4K4M3M3L4M3M3L4J6J7I6J6J6J6J6J6K5J6J6L5J5L4L4O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O2N10000O100O10000O10000O10O010000O10O010000O10O100000000O100000O10O10000000000O100000000O0100000000000O1000000000000O101O0000000001O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O2N2N2N2N2N2N3M2N2N2N2N1O1O1O1O1O2N1O100O1O100O2O0O1O10O00001O00001dIRId4n6[KXI`4h6_K^I\\4c6aKdIZ4\\6eKeI[4\\6cKeI]4[6bKgI]4Y6bKhI^4Y6`KhI`4Z6]KhIb4Z6[KgIe4\\6VKfIj4\\6SKfIl4]6PKdIP5_6lJbIT5`6iJbIV5a6fJ`IZ5c6bJ^I^5d6_J^I_5R80001O00001O001O00001O00001O00001N10001O00001O000O2O00001O1O2N1O2M2O2N1O2N2N1N3M2N3M2N3M3L3N3M2N3M2N3M3L3N3M4L3M4L4L4K4M4K5J6J5J7J6J6J5K6J6J6Jahi1"}, "label": "stationary black motorcycle ridden by someone wearing dark boots and clothing"}]}
{"id": 45475, "image": "COCO_train2014_000000045475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle with a man in black sitting on it\"."}], "task": "refering", "answer_template": "The \"a black motorcycle with a man in black sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 166, 167, 168, 169, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 328, 329, 332, 333, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362, 383], "bbox": [0.210671875, 0.39789583333333334, 0.805625, 0.9462083333333333], "iscrowd": 0, "area": 61075.81174999999, "rle": {"size": [480, 640], "counts": "nco1:_>8N2M2O2M3[NWOPEk0[:KaE8[:N`E5\\:0aE2[:3aE0[:5bEMZ:9aEI]:;]EJ_:;ZEKc:9UENh:W2M3M3L4M3M2N2N2N2N3N1O1O1O2N1O1O1O1O2N1N2O1O1O2N1O2N2N2N2M4M2O1O1O1O2N1O1O1O1O2N001O001O1O001O001O001O001O001O001O0O2OfNfGoKY8o3kGoKU8P4mGPLR8o3PHPLP8o3a1O1O1O001O1O1O1O1O1O1ZOf0N101bMmEMU:ORFMn90XFMi9O\\FNf9N`FO`9NeFOY<M2M4M2M3N3L3NS\\3B[dL1O1O1M4K4M3M3L4M4L3L4M3L4M4L3L4M3M3L4M3M4K4M3M3L4M3M3L4J6J7I6J6J6J6J6J6K5J6J6L5J5L4L4O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1O2N10000O100O10000O10000O10O010000O10O010000O10O100000000O100000O10O10000000000O100000000O0100000000000O1000000000000O101O0000000001O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O2N2N2N2N2N2N3M2N2N2N2N1O1O1O1O1O2N1O100O1O100O2O0O1O10O00001O00001dIRId4n6[KXI`4h6_K^I\\4c6aKdIZ4\\6eKeI[4\\6cKeI]4[6bKgI]4Y6bKhI^4Y6`KhI`4Z6]KhIb4Z6[KgIe4\\6VKfIj4\\6SKfIl4]6PKdIP5_6lJbIT5`6iJbIV5a6fJ`IZ5c6bJ^I^5d6_J^I_5R80001O00001O001O00001O00001O00001N10001O00001O000O2O00001O1O2N1O2M2O2N1O2N2N1N3M2N3M2N3M3L3N3M2N3M2N3M3L3N3M4L3M4L4L4K4M4K5J6J5J7J6J6J5K6J6J6Jahi1"}, "label": "a black motorcycle with a man in black sitting on it"}]}
{"id": 45475, "image": "COCO_train2014_000000045475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 151, 152, 153, 169, 170, 171, 175, 176, 192, 193, 194, 215, 216, 217, 238, 239, 261, 262, 284, 285, 307, 308, 329, 330, 331, 352, 353, 354], "bbox": [0.336734375, 0.06197916666666667, 0.680265625, 0.9236666666666667], "iscrowd": 0, "area": 26685.37035, "rle": {"size": [480, 640], "counts": "VeU39]>;F:H8\\KoNRKW1i4QOnJT1n4TOiJR1S5TOeJR1V5VOaJP1Z5XO]Jn0^5ZOYJl0b5[OVJk0e5\\OTJh0i5]OQJh0j5_OoIf0l5@oIc0m5DlI`0Q6EiI?T6GeI=W6JbI9\\6L_I7]60\\I4a61YI3c6V4N3L4M3M2M4M3LTOPJPIm5n6YJPId5n6bJRI[5l6kJSIQ5k6VKTIg4j6^KVI_4g6hKWIV4g6RLVIk3g6^LVI_3h6jLTIR3k6VMQIh2o6_MmH^2T7gMiHW2W7i2O100O1O1O100O1O1O100O1O1O1O100N2D<D<D<D<H9I6J6J7H7L4N3M2O1N2O2M2N2O1N3N1N2O1O1001O1O1O001O1\\Od0O001N2O1O001O1O1O0O2O001O001O0O2OO1O100O2N2O1N2O2M2N2O1N2O1N2N3N1N2N2O1N2O1N3O01O2N100010O01O001O00010O001O00QNhD5Y;_OREb0m:TO^Ek0b:kNhEV1W:_NTFa1l9UN_Fj1\\;1O010O010O00010O0010N100O2O0O1O2O0O101N100O2O0O101N1O2O0O101N100O2O0O101N100O2N1O1O2N1O1O1O1O1N101O1O1O001O1O1O001O1O1O001O1O1O001O2N6Jklo2"}, "label": "a man on a motorcycle"}]}
{"id": 45475, "image": "COCO_train2014_000000045475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing all black riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man wearing all black riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 56, 57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 151, 152, 153, 169, 170, 171, 175, 176, 192, 193, 194, 215, 216, 217, 238, 239, 261, 262, 284, 285, 307, 308, 329, 330, 331, 352, 353, 354], "bbox": [0.336734375, 0.06197916666666667, 0.680265625, 0.9236666666666667], "iscrowd": 0, "area": 26685.37035, "rle": {"size": [480, 640], "counts": "VeU39]>;F:H8\\KoNRKW1i4QOnJT1n4TOiJR1S5TOeJR1V5VOaJP1Z5XO]Jn0^5ZOYJl0b5[OVJk0e5\\OTJh0i5]OQJh0j5_OoIf0l5@oIc0m5DlI`0Q6EiI?T6GeI=W6JbI9\\6L_I7]60\\I4a61YI3c6V4N3L4M3M2M4M3LTOPJPIm5n6YJPId5n6bJRI[5l6kJSIQ5k6VKTIg4j6^KVI_4g6hKWIV4g6RLVIk3g6^LVI_3h6jLTIR3k6VMQIh2o6_MmH^2T7gMiHW2W7i2O100O1O1O100O1O1O100O1O1O1O100N2D<D<D<D<H9I6J6J7H7L4N3M2O1N2O2M2N2O1N3N1N2O1O1001O1O1O001O1\\Od0O001N2O1O001O1O1O0O2O001O001O0O2OO1O100O2N2O1N2O2M2N2O1N2O1N2N3N1N2N2O1N2O1N3O01O2N100010O01O001O00010O001O00QNhD5Y;_OREb0m:TO^Ek0b:kNhEV1W:_NTFa1l9UN_Fj1\\;1O010O010O00010O0010N100O2O0O1O2O0O101N100O2O0O101N1O2O0O101N100O2O0O101N100O2N1O1O2N1O1O1O1O1N101O1O1O001O1O1O001O1O1O001O1O1O001O2N6Jklo2"}, "label": "a man wearing all black riding a motorcycle"}]}
{"id": 569769, "image": "COCO_train2014_000000569769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white cup in the background in front of the darker cake with the white frosting\"."}], "task": "refering", "answer_template": "The \"the white cup in the background in front of the darker cake with the white frosting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 47, 66, 67, 68, 69, 70, 88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 177, 178, 179], "bbox": [0.0, 0.1132843137254902, 0.21202614379084966, 0.3779575163398693], "iscrowd": 0, "area": 17080.175049999998, "rle": {"size": [612, 612], "counts": "j2>h11g>OYA2g>NYA2f>0XA1g>1XA0f>2XAOg>4WALh>6VAKj>6UAJa>DR@e0Z1Hb>e0\\A[Oa>Q1UAPOh>]1n@dNo>g1g@ZNX?Q2^@oMa?k2O1O1Oc0\\O9H00001O0O10001O00001N10001O000O2O00001O000O2O0000001O00001O000O2O00000000000000000O10000000000000000O1001O00000000000000001O000000000000001O0O10000O100O101O0O100O10001N100O10000O2O1O1N2O1SL^AU3d>jLbAP3_>nLhABJm2`>`MWB[2j=dM[BV2g=iM^BR2j=fM[BT2`?K5J6K4K6K5H8G9G9G9GZlo8"}, "label": "the white cup in the background in front of the darker cake with the white frosting"}]}
{"id": 569769, "image": "COCO_train2014_000000569769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white container near the cake\"."}], "task": "refering", "answer_template": "The \"a white container near the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 47, 66, 67, 68, 69, 70, 88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 177, 178, 179], "bbox": [0.0, 0.1132843137254902, 0.21202614379084966, 0.3779575163398693], "iscrowd": 0, "area": 17080.175049999998, "rle": {"size": [612, 612], "counts": "j2>h11g>OYA2g>NYA2f>0XA1g>1XA0f>2XAOg>4WALh>6VAKj>6UAJa>DR@e0Z1Hb>e0\\A[Oa>Q1UAPOh>]1n@dNo>g1g@ZNX?Q2^@oMa?k2O1O1Oc0\\O9H00001O0O10001O00001N10001O000O2O00001O000O2O0000001O00001O000O2O00000000000000000O10000000000000000O1001O00000000000000001O000000000000001O0O10000O100O101O0O100O10001N100O10000O2O1O1N2O1SL^AU3d>jLbAP3_>nLhABJm2`>`MWB[2j=dM[BV2g=iM^BR2j=fM[BT2`?K5J6K4K6K5H8G9G9G9GZlo8"}, "label": "a white container near the cake"}]}
{"id": 569769, "image": "COCO_train2014_000000569769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cream and fruit jam filled pastry\"."}], "task": "refering", "answer_template": "The \"a cream and fruit jam filled pastry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 33, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 215, 216], "bbox": [0.47864379084967323, 0.008986928104575163, 0.869656862745098, 0.4404575163398693], "iscrowd": 0, "area": 49841.803150000014, "rle": {"size": [612, 612], "counts": "RW_54nb06J7I7RA[OQ;k0hDYOU;n0bDVO[;Q1^DRO`;U1WDPOf;W1RDlNl;Z1lCjNR<]1fCfNX<a1`CcN]<c1[CaNb<g1VC\\Nh<k1oBZNn<l1kBWNS=P2dBTNZ=S2^BPN`=l3M2M3N3M2N2N2N3M2M3N2N3N1O101O000O101O0O101O0O10001O0O101O0O10001O0O101O0O10001N10001O0O10001N1000001N10001N1000001N0100O010O10O10O100O010O10O0100O10O01000O0100O010O100O010O10O010000O010O10O01O1O1O001O1O001O1O001N2O1O001O1O01000000O10000O1000000O100000000O100000O1000000000000000O1000001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O001O0000001N1000001O000000001O000O101O0000002N2N3M3M2N3M3L10000000O1000000O100000O10O1000000O1^MoDQMQ;d2ZE\\Mf:Y2ZFRMf9b2\\GgLe8n2\\H\\Ld7Y3X4E<E;E;D<E;E;D<E;E;DVj_1"}, "label": "a cream and fruit jam filled pastry"}]}
{"id": 569769, "image": "COCO_train2014_000000569769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dessert with blueberries and strawberry\"."}], "task": "refering", "answer_template": "The \"dessert with blueberries and strawberry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 33, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 215, 216], "bbox": [0.47864379084967323, 0.008986928104575163, 0.869656862745098, 0.4404575163398693], "iscrowd": 0, "area": 49841.803150000014, "rle": {"size": [612, 612], "counts": "RW_54nb06J7I7RA[OQ;k0hDYOU;n0bDVO[;Q1^DRO`;U1WDPOf;W1RDlNl;Z1lCjNR<]1fCfNX<a1`CcN]<c1[CaNb<g1VC\\Nh<k1oBZNn<l1kBWNS=P2dBTNZ=S2^BPN`=l3M2M3N3M2N2N2N3M2M3N2N3N1O101O000O101O0O101O0O10001O0O101O0O10001O0O101O0O10001N10001O0O10001N1000001N10001N1000001N0100O010O10O10O100O010O10O0100O10O01000O0100O010O100O010O10O010000O010O10O01O1O1O001O1O001O1O001N2O1O001O1O01000000O10000O1000000O100000000O100000O1000000000000000O1000001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O001O0000001N1000001O000000001O000O101O0000002N2N3M3M2N3M3L10000000O1000000O100000O10O1000000O1^MoDQMQ;d2ZE\\Mf:Y2ZFRMf9b2\\GgLe8n2\\H\\Ld7Y3X4E<E;E;D<E;E;D<E;E;DVj_1"}, "label": "dessert with blueberries and strawberry"}]}
{"id": 569769, "image": "COCO_train2014_000000569769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white china cup filled with a coffee drink and topped with cinnamon\"."}], "task": "refering", "answer_template": "The \"a blue and white china cup filled with a coffee drink and topped with cinnamon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378, 379, 380, 381, 396, 397, 398, 399, 400, 401, 402, 418, 419, 420, 421, 422, 423, 441, 442, 443], "bbox": [0.0, 0.3707843137254902, 0.615735294117647, 0.9213562091503268], "iscrowd": 0, "area": 88998.53209999998, "rle": {"size": [612, 612], "counts": "\\=l1f?c3^La1^Nb1nNS10O01O0O101O0O1000000O101O0O10000O1000001N10000O1000001N10000O10000O101O000O10000O101O0O1000000O101O000000000000001O0000000000001O0O10O1000000O1000000O1000000O1000000O1000O10O1000000O1000000O1000000O1000000O1000O10O1000000O1000000O1000000O1000000O1000O10O1000000O1000000O10000O100O1O1O100O1O00100O1O1O100O1O100O1O2N100O2N1O101N1O2N100O2N100O2N1O101N1O1O2O0O2N100O2N1O101N1O1O2O0O2N1O101N1O101N1O1O2O0N2O2M2O2M2O1N3N1N2O2M2O1N3N1N3N1N2O2M2O1N3N1N2O2M2O1N3N1N3N1O101O00001O00001O00001O001O00001O00001N100O2O0O2O2N4K4M3L5L3L4M4K4M3M4K4M3L4M4K010O01O00100O001O010O010O0100O10O10O10O0100O010O1000O0100O010O10O0100O10O10O10O0100O010O100O01000O010O1O1O1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O1O2N1N2O3M3M2N3M2N3M3M2N3M2N3L4M2N3K4H9H8H7H9H7IYT\\4"}, "label": "a blue and white china cup filled with a coffee drink and topped with cinnamon"}]}
{"id": 569769, "image": "COCO_train2014_000000569769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cappuccino to the lower left hand side of the fruit and the cake\"."}], "task": "refering", "answer_template": "The \"the cappuccino to the lower left hand side of the fruit and the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378, 379, 380, 381, 396, 397, 398, 399, 400, 401, 402, 418, 419, 420, 421, 422, 423, 441, 442, 443], "bbox": [0.0, 0.3707843137254902, 0.615735294117647, 0.9213562091503268], "iscrowd": 0, "area": 88998.53209999998, "rle": {"size": [612, 612], "counts": "\\=l1f?c3^La1^Nb1nNS10O01O0O101O0O1000000O101O0O10000O1000001N10000O1000001N10000O10000O101O000O10000O101O0O1000000O101O000000000000001O0000000000001O0O10O1000000O1000000O1000000O1000000O1000O10O1000000O1000000O1000000O1000000O1000O10O1000000O1000000O1000000O1000000O1000O10O1000000O1000000O10000O100O1O1O100O1O00100O1O1O100O1O100O1O2N100O2N1O101N1O2N100O2N100O2N1O101N1O1O2O0O2N100O2N1O101N1O1O2O0O2N1O101N1O101N1O1O2O0N2O2M2O2M2O1N3N1N2O2M2O1N3N1N3N1N2O2M2O1N3N1N2O2M2O1N3N1N3N1O101O00001O00001O00001O001O00001O00001N100O2O0O2O2N4K4M3L5L3L4M4K4M3M4K4M3L4M4K010O01O00100O001O010O010O0100O10O10O10O0100O010O1000O0100O010O10O0100O10O10O10O0100O010O100O01000O010O1O1O1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O2N1O1N2O1O1O2N1O1O1O1O1O2N1N2O3M3M2N3M2N3M3M2N3M2N3L4M2N3K4H9H8H7H9H7IYT\\4"}, "label": "the cappuccino to the lower left hand side of the fruit and the cake"}]}
{"id": 70059, "image": "COCO_train2014_000000070059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green coat\"."}], "task": "refering", "answer_template": "The \"a man in a green coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 49, 50, 51, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 259, 279, 280, 281, 282], "bbox": [0.03440625, 0.09887052341597796, 0.276578125, 0.9865289256198347], "iscrowd": 0, "area": 28244.194000000007, "rle": {"size": [363, 640], "counts": "km72X;4L5J5L3M1N2O1O2N1N2O2M2O2N2M3N2N2M3N2N2M3N2N2M3N3N2M3M2O2M3M4M2M4M2M4L4M4K5K6K4K5UIcM_4b2VKlMc4X2QKWNg4\\4K6I3N2M2O1N2O1N2O1N2O1N2O2M2O1N2N2O1N2O4K4J7C=F9K6gMnFo1X9O1N2N2N2N2N2N2M3N2000000O10O1000O1000000O1000000O10000OO2N2N2M5L3L4M4L3oHlMf4Y2WKiMf4Z2XKhMg4Y2WKiMj4W2SKkMn4T2QKmMP5S2mJoMT5P2kJQN\\OHb4W2PLRN\\OIe4T2nKTN\\OIg4S2fKZNBDi4Q2aK`ND@T5h1TKlNE]OW5i1QKmNF[OX5i1oJPOFYOZ5i1mJQOGVO\\5j1jJTOGRO`5l1dJWOJmNb5m1bJZOHiNg5o1]J\\OJeNd0GT4Y2\\K_OHaNi0Nl3V2]K_OK^Nk04g3e4ZL`Ka3`4_LcK^3\\4bLhKEYOl2o4_MkK[O_OS3f4bMhL[2X3eMkLX2T3hMPMU2P3jMTMS2l2lMXMQ2h2nM\\Mo1c2QNaMl1_2SNeMj1[2UNiMh1W2XNkMf1U2YNoM5oL`0R5ZOSN1oLc0n4[OWNMoLf0j4\\O\\NHnLj0f4]OdN@jLQ1a4_O_O>a0AB=>BE<<CF;;CH;;AI;=_OF?e5LdiS5"}, "label": "a man in a green coat"}]}
{"id": 70059, "image": "COCO_train2014_000000070059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman using her cell phone\"."}], "task": "refering", "answer_template": "The \"woman using her cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 49, 50, 51, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 259, 279, 280, 281, 282], "bbox": [0.03440625, 0.09887052341597796, 0.276578125, 0.9865289256198347], "iscrowd": 0, "area": 28244.194000000007, "rle": {"size": [363, 640], "counts": "km72X;4L5J5L3M1N2O1O2N1N2O2M2O2N2M3N2N2M3N2N2M3N2N2M3N3N2M3M2O2M3M4M2M4M2M4L4M4K5K6K4K5UIcM_4b2VKlMc4X2QKWNg4\\4K6I3N2M2O1N2O1N2O1N2O1N2O2M2O1N2N2O1N2O4K4J7C=F9K6gMnFo1X9O1N2N2N2N2N2N2M3N2000000O10O1000O1000000O1000000O10000OO2N2N2M5L3L4M4L3oHlMf4Y2WKiMf4Z2XKhMg4Y2WKiMj4W2SKkMn4T2QKmMP5S2mJoMT5P2kJQN\\OHb4W2PLRN\\OIe4T2nKTN\\OIg4S2fKZNBDi4Q2aK`ND@T5h1TKlNE]OW5i1QKmNF[OX5i1oJPOFYOZ5i1mJQOGVO\\5j1jJTOGRO`5l1dJWOJmNb5m1bJZOHiNg5o1]J\\OJeNd0GT4Y2\\K_OHaNi0Nl3V2]K_OK^Nk04g3e4ZL`Ka3`4_LcK^3\\4bLhKEYOl2o4_MkK[O_OS3f4bMhL[2X3eMkLX2T3hMPMU2P3jMTMS2l2lMXMQ2h2nM\\Mo1c2QNaMl1_2SNeMj1[2UNiMh1W2XNkMf1U2YNoM5oL`0R5ZOSN1oLc0n4[OWNMoLf0j4\\O\\NHnLj0f4]OdN@jLQ1a4_O_O>a0AB=>BE<<CF;;CH;;AI;=_OF?e5LdiS5"}, "label": "woman using her cell phone"}]}
{"id": 561582, "image": "COCO_train2014_000000561582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat with its eyes closed on the right\"."}], "task": "refering", "answer_template": "The \"the cat with its eyes closed on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 45, 46, 50, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197], "bbox": [0.47192, 0.1236, 0.9994400000000001, 0.8382133333333335], "iscrowd": 0, "area": 40588.7035, "rle": {"size": [375, 500], "counts": "Saf21f;001O1O1]OM`E3_:O`E2_:NaE3^:NaE3^:NaE2_:O`E2_:O`E2^:0aE1a0Kl85bF0?4i8LhF1;;^7^OSI7SO18b0]7XOVI6TO15g0]7UOXI4UO03k0_7SOWI3VO00n0a7ROVI2XOOMQ1d7POUI0ZOb1`7aNSIN\\Ob1a7aNQIN\\Oc1c7`NoHN]Ob1e7aNlHT2T7mMiHU2W7lMgHT2Z7nMcHS2]7nMaHR2_7PN_HQ2a7PN]HP2d7QNYHP2h7P10O100O100O1O100O100O100O1O100O100O100O1O1000O100000001O000000001O00000O1O2O0O1O1O1O2O0O1O00100O1O001O10O01O1O1O010O1O001O100O001O1O010O1O1O010O1O001RJcK`4^4]KdKd4\\4XKgKh4Y4VKiKj4X4SKjKm4W4PKkKP5V4mJlKT5S4iJPLW5Q4fJQLZ5P4cJRL]5n3aJTL_5m3^JULb5l3ZJWLf5j3WJYLg5h3WJZLi5j4000000O100000000O1000001O0O101O00001N10001O000O2O0O1O2N1O1NZKfJT3X5lLiJT3V5kLlJU3S5jLnJX3P5hLQKY3l4gLVKZ3h4fLYK[3e4dL\\K^3b4aL`K`3^4_LdKc3Y4\\LiKe3T4[LmKg3Q4WLRLj3l3ULVLl3h3SLZLo3c3PL^LR4_3mKdLS4[3lKgLT4X3kKjLV4U3hKlLY4U3dKmL\\4T3`KoL`4R3]KPMc4Q3ZKPMg4Q3VKQMj4Q3QKRMo4o2nJSMR5n2kJSMW5m2fJUMZ5l2bJWM^5a40O10000O101O000O1000000O10001OO10000000O10000000000000O01000000000000O10000000O100000O1000000000000000O10O100000001O000000001O0O1000001O00000000001N100000001O00000U4kKoG"}, "label": "the cat with its eyes closed on the right"}]}
{"id": 561582, "image": "COCO_train2014_000000561582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat that is sleeping\"."}], "task": "refering", "answer_template": "The \"a black cat that is sleeping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 45, 46, 50, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197], "bbox": [0.47192, 0.1236, 0.9994400000000001, 0.8382133333333335], "iscrowd": 0, "area": 40588.7035, "rle": {"size": [375, 500], "counts": "Saf21f;001O1O1]OM`E3_:O`E2_:NaE3^:NaE3^:NaE2_:O`E2_:O`E2^:0aE1a0Kl85bF0?4i8LhF1;;^7^OSI7SO18b0]7XOVI6TO15g0]7UOXI4UO03k0_7SOWI3VO00n0a7ROVI2XOOMQ1d7POUI0ZOb1`7aNSIN\\Ob1a7aNQIN\\Oc1c7`NoHN]Ob1e7aNlHT2T7mMiHU2W7lMgHT2Z7nMcHS2]7nMaHR2_7PN_HQ2a7PN]HP2d7QNYHP2h7P10O100O100O1O100O100O100O1O100O100O100O1O1000O100000001O000000001O00000O1O2O0O1O1O1O2O0O1O00100O1O001O10O01O1O1O010O1O001O100O001O1O010O1O1O010O1O001RJcK`4^4]KdKd4\\4XKgKh4Y4VKiKj4X4SKjKm4W4PKkKP5V4mJlKT5S4iJPLW5Q4fJQLZ5P4cJRL]5n3aJTL_5m3^JULb5l3ZJWLf5j3WJYLg5h3WJZLi5j4000000O100000000O1000001O0O101O00001N10001O000O2O0O1O2N1O1NZKfJT3X5lLiJT3V5kLlJU3S5jLnJX3P5hLQKY3l4gLVKZ3h4fLYK[3e4dL\\K^3b4aL`K`3^4_LdKc3Y4\\LiKe3T4[LmKg3Q4WLRLj3l3ULVLl3h3SLZLo3c3PL^LR4_3mKdLS4[3lKgLT4X3kKjLV4U3hKlLY4U3dKmL\\4T3`KoL`4R3]KPMc4Q3ZKPMg4Q3VKQMj4Q3QKRMo4o2nJSMR5n2kJSMW5m2fJUMZ5l2bJWM^5a40O10000O101O000O1000000O10001OO10000000O10000000000000O01000000000000O10000000O100000O1000000000000000O10O100000001O000000001O0O1000001O00000000001N100000001O00000U4kKoG"}, "label": "a black cat that is sleeping"}]}
{"id": 561582, "image": "COCO_train2014_000000561582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bigger black and white cat\"."}], "task": "refering", "answer_template": "The \"the bigger black and white cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 216, 217, 218, 219, 220, 221], "bbox": [0.00484, 0.42349333333333333, 0.79526, 0.98168], "iscrowd": 0, "area": 57881.2416, "rle": {"size": [375, 500], "counts": "an01V;j0[EFS9V2VO?]OP1ZO5K5J5L4L4L6I2O1O1N2O001O0O101O0O101O001N10001N10001O0O101O0O101O000O100O100O1O101N1000000O1000000000000O101O00000000001O000O100000001O000O100O100O1O100O10000O1000000O10000O1000O10O1000000O1000000O1000000O1010O0000000001O00000000000001O00000O1000000000000000001N100000000O0100000O10000O1O1BbJbJ^5k51EaJ]J`5b5;O100O100O10O0100O1000000O1000000000O10O100000000000000O1000000000000000000000001O00001O0O10001O00001O00001O00001O0O101O00001O00001O0O101O00001O00001O0O101O00001O00001N10001O1O1O001O1O1N101O1O00001O00001O00001O0O101O00001O0O10001O00001N101O1O001O1O001O001O1N101N2O0O2O0O2O1N3N1N2O1N3N1O1N2O2N1O1O1N2O2N1O1N2O1O001O1O1O010O1O001O1O001O1O001O1O001O1O00100O001O001O1O0`NmF<T9BnF=S9AoF>Q9ARG=o8ASG>m8AVG=k8AVG?k8_OVGa0j8^OXGa0i8\\OYGd0h8ZOYGf0g8YO[Gf0f8XO[Gg0g8WOZGi0f8VO[Gj0g8SO[Gl0h9000001O0000001O0000001O0000001O00001O001O001N2O001O0O2O001O1N101O001N101O1O0O2O001O1N1Mb]U1"}, "label": "the bigger black and white cat"}]}
{"id": 561582, "image": "COCO_train2014_000000561582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat with a paw resting on its head\"."}], "task": "refering", "answer_template": "The \"a black and white cat with a paw resting on its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 216, 217, 218, 219, 220, 221], "bbox": [0.00484, 0.42349333333333333, 0.79526, 0.98168], "iscrowd": 0, "area": 57881.2416, "rle": {"size": [375, 500], "counts": "an01V;j0[EFS9V2VO?]OP1ZO5K5J5L4L4L6I2O1O1N2O001O0O101O0O101O001N10001N10001O0O101O0O101O000O100O100O1O101N1000000O1000000000000O101O00000000001O000O100000001O000O100O100O1O100O10000O1000000O10000O1000O10O1000000O1000000O1000000O1010O0000000001O00000000000001O00000O1000000000000000001N100000000O0100000O10000O1O1BbJbJ^5k51EaJ]J`5b5;O100O100O10O0100O1000000O1000000000O10O100000000000000O1000000000000000000000001O00001O0O10001O00001O00001O00001O0O101O00001O00001O0O101O00001O00001O0O101O00001O00001N10001O1O1O001O1O1N101O1O00001O00001O00001O0O101O00001O0O10001O00001N101O1O001O1O001O001O1N101N2O0O2O0O2O1N3N1N2O1N3N1O1N2O2N1O1O1N2O2N1O1N2O1O001O1O1O010O1O001O1O001O1O001O1O001O1O00100O001O001O1O0`NmF<T9BnF=S9AoF>Q9ARG=o8ASG>m8AVG=k8AVG?k8_OVGa0j8^OXGa0i8\\OYGd0h8ZOYGf0g8YO[Gf0f8XO[Gg0g8WOZGi0f8VO[Gj0g8SO[Gl0h9000001O0000001O0000001O0000001O00001O001O001N2O001O0O2O001O1N101O001N101O1O0O2O001O1N1Mb]U1"}, "label": "a black and white cat with a paw resting on its head"}]}
{"id": 168366, "image": "COCO_train2014_000000168366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table supporting the pizza\"."}], "task": "refering", "answer_template": "The \"the table supporting the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 202, 203, 204, 205, 226, 227, 228, 249, 250, 251, 272, 273, 274, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.5031718749999999, 0.0, 1.0, 0.9844055944055944], "iscrowd": 0, "area": 46763.261199999986, "rle": {"size": [429, 640], "counts": "\\gW41[=1O100O10000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O100cCZOS<f0nC]On;c0RDAj;?WDDe;<[DHa;8_DM[;4eD0W;0iD4S;LmD8o:HQE<k:DUE`0g:@YEe0a:\\O_Eh0]:XOcEl0Y:TOgEP1U:POkET1Q:lNoEX1m9hNSF\\1i9dNWFa1c9`N]Fd1_9\\NaFh1_8_MQHi0@l1[8`MPHd0EP2W8`MPH`0IT2S8aMoG;NY2m7bMPH53]2i7`NWHd1e7\\N[Hh1`7YN`Hk1\\7UNdHo1X7QNhHT2R7mMnHW2n6iMRI[2j6eMVI_2f6aMZIc2a6^M_If2]6ZMcIj2Y6VMgIo2S6RMmIR3o5nLQJV3k5jLUJZ3f5gLZJ]3b5cL^Ja3^5_LbJf3Y5ZLgJj3T5WLlJm3P5SLPKQ4l4oKTKU4g4lKYKX4c4hK]K]4^4cKbKa4Y4`KgKd4U4\\KkKh4Q4XKoKl4m3TKSLP5h3QKXLS5d3mJ\\LX5_3hJaL\\5Z3eJfL_5V3aJjLc5R3]JnLg5m2ZJSMj5i2VJWMo5d2QJ\\MS6`2mI`MW6[2jIeMZ6W2fIiM^6S2bImMb6n1_IRNf6i1ZIWNj6e1VI[Nn6`1SI`NQ7\\1oHdNU7X1kHhNY7T1gHlN]7o0dHQOa7j0_HVOe7f0[HZOi7a0XH_Ol7=THCP89PHGT84mGLX8OhG1\\8KdG5j:000000000000000000000000000000000000000000000000000000000000000J6I7I7I7I7I7I6J7I7I7I7I7I7I7I7I7I7I7I7I7H8I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I6"}, "label": "the table supporting the pizza"}]}
{"id": 168366, "image": "COCO_train2014_000000168366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the area of table top to the right of the cutting board the pizza is on\"."}], "task": "refering", "answer_template": "The \"the area of table top to the right of the cutting board the pizza is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 202, 203, 204, 205, 226, 227, 228, 249, 250, 251, 272, 273, 274, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.5031718749999999, 0.0, 1.0, 0.9844055944055944], "iscrowd": 0, "area": 46763.261199999986, "rle": {"size": [429, 640], "counts": "\\gW41[=1O100O10000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O1000000000000O100cCZOS<f0nC]On;c0RDAj;?WDDe;<[DHa;8_DM[;4eD0W;0iD4S;LmD8o:HQE<k:DUE`0g:@YEe0a:\\O_Eh0]:XOcEl0Y:TOgEP1U:POkET1Q:lNoEX1m9hNSF\\1i9dNWFa1c9`N]Fd1_9\\NaFh1_8_MQHi0@l1[8`MPHd0EP2W8`MPH`0IT2S8aMoG;NY2m7bMPH53]2i7`NWHd1e7\\N[Hh1`7YN`Hk1\\7UNdHo1X7QNhHT2R7mMnHW2n6iMRI[2j6eMVI_2f6aMZIc2a6^M_If2]6ZMcIj2Y6VMgIo2S6RMmIR3o5nLQJV3k5jLUJZ3f5gLZJ]3b5cL^Ja3^5_LbJf3Y5ZLgJj3T5WLlJm3P5SLPKQ4l4oKTKU4g4lKYKX4c4hK]K]4^4cKbKa4Y4`KgKd4U4\\KkKh4Q4XKoKl4m3TKSLP5h3QKXLS5d3mJ\\LX5_3hJaL\\5Z3eJfL_5V3aJjLc5R3]JnLg5m2ZJSMj5i2VJWMo5d2QJ\\MS6`2mI`MW6[2jIeMZ6W2fIiM^6S2bImMb6n1_IRNf6i1ZIWNj6e1VI[Nn6`1SI`NQ7\\1oHdNU7X1kHhNY7T1gHlN]7o0dHQOa7j0_HVOe7f0[HZOi7a0XH_Ol7=THCP89PHGT84mGLX8OhG1\\8KdG5j:000000000000000000000000000000000000000000000000000000000000000J6I7I7I7I7I7I6J7I7I7I7I7I7I7I7I7I7I7I7I7H8I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I7I6"}, "label": "the area of table top to the right of the cutting board the pizza is on"}]}
{"id": 168366, "image": "COCO_train2014_000000168366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table top area to the left of the pizza and below the newspaper\"."}], "task": "refering", "answer_template": "The \"the table top area to the left of the pizza and below the newspaper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 277, 300, 323], "bbox": [0.0, 0.08969696969696969, 0.11996875, 0.9867832167832169], "iscrowd": 0, "area": 16033.920949999996, "rle": {"size": [429, 640], "counts": "^4[1Q<d2[M\\1eN2N1N3N1O2M3N1O2M2O2N1N3N2M2O2N1N3N1O2M3N1O2M2O2M2O2N2dKWIP2k6jM^IR2d6fMfIV2[6dMnIW2U6bMUJY2l5aM]J[2e5]MeJ^2]5\\MlJ`2V5YMRKc2Q5VMUKi2h7O0OO2M3N1N3N2N1N3N1N3I7D;E<D<D;E<C=D;E<D<D;E<D;E<D<C<E<D<D;E<D<D;E<D;D=D<Dmj[7"}, "label": "the table top area to the left of the pizza and below the newspaper"}]}
{"id": 479666, "image": "COCO_train2014_000000479666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white wooden table with small wooden chairs\"."}], "task": "refering", "answer_template": "The \"a white wooden table with small wooden chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 223, 224, 225, 226, 227, 247, 248, 249, 250, 251, 270, 274, 293, 297, 316, 320], "bbox": [0.703921875, 0.5889929742388759, 0.9539375, 0.8912412177985949], "iscrowd": 0, "area": 6799.620300000004, "rle": {"size": [427, 640], "counts": "mYl57T=3M1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2[FlNo6U1gHTOY7m0\\H]Od7e0QHEn7<hGMX84^G5b8LSG>n8CgFg0X9f10001O000000K5D<D<D<D<D<J600O1000000000000O1000000000000O10000000001O00000O10000O10000O10000O1000000cNYDU1h;kNZDS1f;mN[DR1e;nN]DP1c;POaDl0_;TOfDg0X<G100O1O1O000000000000O10001O0000000O1N2O1000000O101O001N2E;N20000O10O1O0O2O1O001O001O1O001N`E^OZ8a0ZGKf85nF7S9HaFd0`9[OTFQ1l9oNhE^1X:j01O001O1O0F;E:F;E;E:F;E;F9O2O1O0O2O1O0O2O00fi;"}, "label": "a white wooden table with small wooden chairs"}]}
{"id": 414130, "image": "COCO_train2014_000000414130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with dark hair and a mustache and beard\"."}], "task": "refering", "answer_template": "The \"a man with dark hair and a mustache and beard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 90, 91, 108, 109, 126, 127, 144, 145, 162, 163], "bbox": [0.00302, 0.3424024024024024, 0.11478000000000001, 0.8049849849849849], "iscrowd": 0, "area": 6458.4947, "rle": {"size": [333, 500], "counts": "bh0Q2W8S2RNc0]O000O1000000000000010O00001O00001O00001O00001RLeJh2[5WMfJj2Y5VMgJj2Z5UMfJk2\\5RMeJn2`5mL`JS3e5gL\\JY3i5bLWJ^3\\6N2N2N2N2N2N2N2L4H8M3O1N3M4M2M4M3L4M2M4M3\\ObGSOb8:cm_4"}, "label": "a man with dark hair and a mustache and beard"}]}
{"id": 414130, "image": "COCO_train2014_000000414130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown haired man with a striped shirt and a green pillow on his lap\"."}], "task": "refering", "answer_template": "The \"the brown haired man with a striped shirt and a green pillow on his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 90, 91, 108, 109, 126, 127, 144, 145, 162, 163], "bbox": [0.00302, 0.3424024024024024, 0.11478000000000001, 0.8049849849849849], "iscrowd": 0, "area": 6458.4947, "rle": {"size": [333, 500], "counts": "bh0Q2W8S2RNc0]O000O1000000000000010O00001O00001O00001O00001RLeJh2[5WMfJj2Y5VMgJj2Z5UMfJk2\\5RMeJn2`5mL`JS3e5gL\\JY3i5bLWJ^3\\6N2N2N2N2N2N2N2L4H8M3O1N3M4M2M4M3L4M2M4M3\\ObGSOb8:cm_4"}, "label": "the brown haired man with a striped shirt and a green pillow on his lap"}]}
{"id": 414130, "image": "COCO_train2014_000000414130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black man in a grey shirt watching tv\"."}], "task": "refering", "answer_template": "The \"a black man in a grey shirt watching tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 75, 76, 92, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 128, 129, 130, 145, 146, 147, 148, 163, 164, 165, 166], "bbox": [0.07765999999999999, 0.3085285285285285, 0.39722, 0.8062762762762763], "iscrowd": 0, "area": 10587.5542, "rle": {"size": [333, 500], "counts": "Qn<6n99K5O1N2O0O2O1O1N2O001N2O1O1O1O0O2O1O1O1O0O2O1O00O01O1O0010O01O0eNlN_IU1a6VOSIk0n6_OgH`0Z7Y110O100O100O2O0O100O100O100O100O101N100O100TIoL[6R3bIQM^6o2_ITM`6m2^IUMb6k2[IXMd6i2ZIYMf6W3K5H8H8J6L4L4M300O100O100O1O100O100000O2L3M3N3L3M3N3L3N2M4L3N4K7RM^If1h6SN^Ii1h6oM]IW1AQOY7@ZI^1@nNR8P1QHkNR8T1QHgNR8Y1PHcNR8\\1a0N1O101N1O1O101N1O1O101N1O101N1O1O101O00100O010O100O001O1O1O00100O1O1O001O1O10O01O1O1O1O010O1O1O010O1O100O001O100O`fQ3"}, "label": "a black man in a grey shirt watching tv"}]}
{"id": 414130, "image": "COCO_train2014_000000414130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in grey shirt\"."}], "task": "refering", "answer_template": "The \"man in grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 75, 76, 92, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 128, 129, 130, 145, 146, 147, 148, 163, 164, 165, 166], "bbox": [0.07765999999999999, 0.3085285285285285, 0.39722, 0.8062762762762763], "iscrowd": 0, "area": 10587.5542, "rle": {"size": [333, 500], "counts": "Qn<6n99K5O1N2O0O2O1O1N2O001N2O1O1O1O0O2O1O1O1O0O2O1O00O01O1O0010O01O0eNlN_IU1a6VOSIk0n6_OgH`0Z7Y110O100O100O2O0O100O100O100O100O101N100O100TIoL[6R3bIQM^6o2_ITM`6m2^IUMb6k2[IXMd6i2ZIYMf6W3K5H8H8J6L4L4M300O100O100O1O100O100000O2L3M3N3L3M3N3L3N2M4L3N4K7RM^If1h6SN^Ii1h6oM]IW1AQOY7@ZI^1@nNR8P1QHkNR8T1QHgNR8Y1PHcNR8\\1a0N1O101N1O1O101N1O1O101N1O101N1O1O101O00100O010O100O001O1O1O00100O1O1O001O1O10O01O1O1O1O010O1O1O010O1O100O001O100O`fQ3"}, "label": "man in grey shirt"}]}
{"id": 250293, "image": "COCO_train2014_000000250293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with pepperoni and jalapeno slices\"."}], "task": "refering", "answer_template": "The \"a pizza with pepperoni and jalapeno slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 233, 234, 235, 236, 237, 238, 239, 240, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.0016875000000000002, 0.17683333333333331, 1.0, 0.9835833333333334], "iscrowd": 0, "area": 133892.01815, "rle": {"size": [480, 640], "counts": "Sf06i>9F;E9H2M3M2O2M3M2O2M3M2O2M3M2O2M2N3N2M2N3M3N1O2N2O0O2O1N101N2O0O2O0O2N2O0O2O0O101N100O2N101N100O2O0O101N101N1O101N100O2O0O101N101N100O2O0O101N101N101N2O1O1N2O1N2O1N2O1N2O1N2O2N1N2O1N2O1N2O1N2O1N2O3M2M3N2N3M2M3N2N3M2M3N2N3M1O1N2O001O1N101O1O0O2O1O001N2O001O1O0O2O1O1N101O1O0O2O1O0O2O1OO010O10O10O100O01000O01000O010O10O10O10O10O10O010O10O10O010O10O10O10O010O10O10O10O010O10O10O10O010O10O0100O1O100O00100O1O100O10O01O100O1O10O01O100O1O100O010O1O100O1O010O1O100O1O10O0100O1O100O1O10O01O100O1O100O1O1O00100O1O1O100O1O1O00100O1O1O100O1O1O1O010O1O1O100O1O1O1O010O1O1O100O1O1O1O010O1O100O1O100O1O010O1O1O100O1O100O00100O1O100O1O100O1O010O1O100O1O100O1O00100O1O100O1O100O1O010O1O100O1O100O1O010O100O1O10O01O100O1O010O100O00100O1O10O0100O1O010O1O10O010000O10O10O1000O010000O01000O11O:E<E:F3M1O2M3N1O2N2N1N3N2N1O2N2M2O2N2N1O2M2O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1O1O2N100O2N1O1O2N101N1O2N2N1O2N101N2N1O2N1O2O1N1O2N2N101N1O2O1N1O2O1N2O1N2N2O2M2O1N2N2O1N3M2O1N2O1N2N3N0O10000O1000000O10000O1000000O0100000O10000O1000000O1000000O1000O010001O0O10000O10000O10000O101O0O10000O10000O10000O2O00O001O1O00100O001O1O010O1O001O0010O01O001O0010O01O001O10O01O001O0010O01O001O0010O01O001O001N2O001N101O0O2O001N2O2M2O1O1N2O1O1N3N1O1N2O1O1O1N2O2N1O1N2O1O1O1N3N4L3L5L4K4L5K5K4L5L3L4L4L5K4L4L5K4E;F:EPH"}, "label": "a pizza with pepperoni and jalapeno slices"}]}
{"id": 250293, "image": "COCO_train2014_000000250293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with jalapenos\"."}], "task": "refering", "answer_template": "The \"a pizza with jalapenos\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 233, 234, 235, 236, 237, 238, 239, 240, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.0016875000000000002, 0.17683333333333331, 1.0, 0.9835833333333334], "iscrowd": 0, "area": 133892.01815, "rle": {"size": [480, 640], "counts": "Sf06i>9F;E9H2M3M2O2M3M2O2M3M2O2M3M2O2M2N3N2M2N3M3N1O2N2O0O2O1N101N2O0O2O0O2N2O0O2O0O101N100O2N101N100O2O0O101N101N1O101N100O2O0O101N101N100O2O0O101N101N101N2O1O1N2O1N2O1N2O1N2O1N2O2N1N2O1N2O1N2O1N2O1N2O3M2M3N2N3M2M3N2N3M2M3N2N3M1O1N2O001O1N101O1O0O2O1O001N2O001O1O0O2O1O1N101O1O0O2O1O0O2O1OO010O10O10O100O01000O01000O010O10O10O10O10O10O010O10O10O010O10O10O10O010O10O10O10O010O10O10O10O010O10O0100O1O100O00100O1O100O10O01O100O1O10O01O100O1O100O010O1O100O1O010O1O100O1O10O0100O1O100O1O10O01O100O1O100O1O1O00100O1O1O100O1O1O00100O1O1O100O1O1O1O010O1O1O100O1O1O1O010O1O1O100O1O1O1O010O1O100O1O100O1O010O1O1O100O1O100O00100O1O100O1O100O1O010O1O100O1O100O1O00100O1O100O1O100O1O010O1O100O1O100O1O010O100O1O10O01O100O1O010O100O00100O1O10O0100O1O010O1O10O010000O10O10O1000O010000O01000O11O:E<E:F3M1O2M3N1O2N2N1N3N2N1O2N2M2O2N2N1O2M2O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1O1O2N100O2N1O1O2N101N1O2N2N1O2N101N2N1O2N1O2O1N1O2N2N101N1O2O1N1O2O1N2O1N2N2O2M2O1N2N2O1N3M2O1N2O1N2N3N0O10000O1000000O10000O1000000O0100000O10000O1000000O1000000O1000O010001O0O10000O10000O10000O101O0O10000O10000O10000O2O00O001O1O00100O001O1O010O1O001O0010O01O001O0010O01O001O10O01O001O0010O01O001O0010O01O001O001N2O001N101O0O2O001N2O2M2O1O1N2O1O1N3N1O1N2O1O1O1N2O2N1O1N2O1O1O1N3N4L3L5L4K4L5K5K4L5L3L4L4L5K4L4L5K4E;F:EPH"}, "label": "a pizza with jalapenos"}]}
{"id": 250293, "image": "COCO_train2014_000000250293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of pizza closest to the camera\"."}], "task": "refering", "answer_template": "The \"the slice of pizza closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 195, 196, 197, 198, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.19718750000000002, 0.46291666666666664, 0.851125, 0.9954999999999999], "iscrowd": 0, "area": 59904.43285, "rle": {"size": [480, 640], "counts": "W]k15k>4K3N3M3L4M2M2O2N1N3N1N2O2N1N3N2N2M3N2M3N2N2M3N2M3N3M2M3N001N101N1LdMdC]2[<dMdC\\2[<eMdC]2[<4O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2N2O0O2O0O2O001N101O001N101O001O0O2O001O0O2O001O001N101N1O2O0O2N1O2O0O2N1O2O0O2N100O2N1O2O0O2N101N1O1O1O2O0O1O1O2N100O1O2N100O101N100O101N100O101N100O101N100O101N100O1O2O0O100O2N100O100O1O100O100O100O100O100O100O100O100O100O100O1O100O100O100O1O100O100O1O100O1O1O1O100O001O100O1O1O100O1O1O100O1O100O1O100O1O1O100O100O1O100O10O0100O100O100O100O100O100O10000O100O100O100O10000O11O000O2O00001O001O00001O2N2N101N2N2N4L6J6J5K4L3M2N2N2N3M2O1N3M2N1O2N1O1O1O1O1O1O1O1O101N1O2N1O2N1O2N1O2N2N1O2N1O2O2M2N2N2N2N2N2N1O2N2N2N1O2N2O0O2N2N2N1O2N2O1N2N3M2N2O1N2N3M2N1O2O0O1O2N1O1O101N1O1O2N1O1O2N1O1O1O2N1O1O2N00001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O1O1O1O001O1O1O1O001O1O1O1O1O1N2O1N2O1N2O2M2O2N1N3M2N3M3L7DoR\\1"}, "label": "the slice of pizza closest to the camera"}]}
{"id": 250293, "image": "COCO_train2014_000000250293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza slice at 6 oclock\"."}], "task": "refering", "answer_template": "The \"pizza slice at 6 oclock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 195, 196, 197, 198, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.19718750000000002, 0.46291666666666664, 0.851125, 0.9954999999999999], "iscrowd": 0, "area": 59904.43285, "rle": {"size": [480, 640], "counts": "W]k15k>4K3N3M3L4M2M2O2N1N3N1N2O2N1N3N2N2M3N2M3N2N2M3N2M3N3M2M3N001N101N1LdMdC]2[<dMdC\\2[<eMdC]2[<4O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2N2O0O2O0O2O001N101O001N101O001O0O2O001O0O2O001O001N101N1O2O0O2N1O2O0O2N1O2O0O2N100O2N1O2O0O2N101N1O1O1O2O0O1O1O2N100O1O2N100O101N100O101N100O101N100O101N100O101N100O1O2O0O100O2N100O100O1O100O100O100O100O100O100O100O100O100O100O1O100O100O100O1O100O100O1O100O1O1O1O100O001O100O1O1O100O1O1O100O1O100O1O100O1O1O100O100O1O100O10O0100O100O100O100O100O100O10000O100O100O100O10000O11O000O2O00001O001O00001O2N2N101N2N2N4L6J6J5K4L3M2N2N2N3M2O1N3M2N1O2N1O1O1O1O1O1O1O1O101N1O2N1O2N1O2N1O2N2N1O2N1O2O2M2N2N2N2N2N2N1O2N2N2N1O2N2O0O2N2N2N1O2N2O1N2N3M2N2O1N2N3M2N1O2O0O1O2N1O1O101N1O1O2N1O1O2N1O1O1O2N1O1O2N00001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O1O1O1O001O1O1O1O001O1O1O1O1O1N2O1N2O1N2O2M2O2N1N3M2N3M3L7DoR\\1"}, "label": "pizza slice at 6 oclock"}]}
{"id": 561590, "image": "COCO_train2014_000000561590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black shirt and blue jeans standing next to a woman\"."}], "task": "refering", "answer_template": "The \"man in black shirt and blue jeans standing next to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 39, 40, 60, 61, 62, 63, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 222, 223, 224, 246, 247, 268, 269, 270, 289, 290, 292, 293], "bbox": [0.577921875, 0.037798594847775174, 0.797203125, 0.86903981264637], "iscrowd": 0, "area": 23711.306449999993, "rle": {"size": [427, 640], "counts": "]dj42U=5N2N101N101N2O010O00000O100O001001OkHB_1;^NJmJMP69oN3jJIT63oNl0o0TOmNS1P1lNmNZ1Q1fNkNa1_LWNP35`0j1TLSNmN5]4K`0S2QL]N^3^O>\\2nK[Nb3XO>c2kKZN]8k1_GXN]8m1_GVN^8n1_GTN]8R2^GQN_8`3M3N2M3M3N2M2O0O3N1O2M3N3M3M2M3N1O0O016J1N10002M4M1NM4L3M3M4L3M3M3M3MfKbJn0^5oNfJQ1]5iNfJW1]5dNfJ[1]5`NeJ`1^5ZNfJd1^5WNdJhM]Oh3S6[NbJjM@h3R6XNbJmM_Oi3S6UNkJS1o5hNTJU1^6YNfIc1c6TN`Ii1f6PN`Ik1g6nM_Im1g6lM[IS2j6gMYIX2m8O101N100000000O1000000001O0_OiMnEW2Q:lMmET2R:PNkEP2U:SNhEm1W:WNfEj1X:ZNeEf1[:f0kFSMi7n2UHUMj7l2SHWMh6n3UIUL`6V4\\InK[6\\4bIeKW6c4fI`KQ6m4dIZKS6T6H8YMlHiNa7T1iHWNd7h1kH]Mc7`2bHXMc7f2cHRMb7l2`1N1O2M2O2N1O3M2M4M3M6I7H8I6I8\\OWDYOQ<a0b0J6K6IhRf1"}, "label": "man in black shirt and blue jeans standing next to a woman"}]}
{"id": 561590, "image": "COCO_train2014_000000561590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in gray shirt\"."}], "task": "refering", "answer_template": "The \"man in gray shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 39, 40, 60, 61, 62, 63, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 222, 223, 224, 246, 247, 268, 269, 270, 289, 290, 292, 293], "bbox": [0.577921875, 0.037798594847775174, 0.797203125, 0.86903981264637], "iscrowd": 0, "area": 23711.306449999993, "rle": {"size": [427, 640], "counts": "]dj42U=5N2N101N101N2O010O00000O100O001001OkHB_1;^NJmJMP69oN3jJIT63oNl0o0TOmNS1P1lNmNZ1Q1fNkNa1_LWNP35`0j1TLSNmN5]4K`0S2QL]N^3^O>\\2nK[Nb3XO>c2kKZN]8k1_GXN]8m1_GVN^8n1_GTN]8R2^GQN_8`3M3N2M3M3N2M2O0O3N1O2M3N3M3M2M3N1O0O016J1N10002M4M1NM4L3M3M4L3M3M3M3MfKbJn0^5oNfJQ1]5iNfJW1]5dNfJ[1]5`NeJ`1^5ZNfJd1^5WNdJhM]Oh3S6[NbJjM@h3R6XNbJmM_Oi3S6UNkJS1o5hNTJU1^6YNfIc1c6TN`Ii1f6PN`Ik1g6nM_Im1g6lM[IS2j6gMYIX2m8O101N100000000O1000000001O0_OiMnEW2Q:lMmET2R:PNkEP2U:SNhEm1W:WNfEj1X:ZNeEf1[:f0kFSMi7n2UHUMj7l2SHWMh6n3UIUL`6V4\\InK[6\\4bIeKW6c4fI`KQ6m4dIZKS6T6H8YMlHiNa7T1iHWNd7h1kH]Mc7`2bHXMc7f2cHRMb7l2`1N1O2M2O2N1O3M2M4M3M6I7H8I6I8\\OWDYOQ<a0b0J6K6IhRf1"}, "label": "man in gray shirt"}]}
{"id": 561590, "image": "COCO_train2014_000000561590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing jeans and a black shirt with her arms crossed\"."}], "task": "refering", "answer_template": "The \"a woman wearing jeans and a black shirt with her arms crossed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 180, 181, 203, 226, 249], "bbox": [0.8006249999999999, 0.0853864168618267, 0.9550468750000001, 0.7730444964871194], "iscrowd": 0, "area": 13267.265150000001, "rle": {"size": [427, 640], "counts": "dde61W=4K6K4K6K4lCZO];l0ZD[Oc;X1L5L3L4M3M7I:F;E=C<D9J5J6J6K4K8H8I7H7I8I7H8I7I6J7I7H8D<_O3MG?A`0_Oa0@`0@`0M51O100O1O1O101N1O110O2N1O1O2O0O2N2N2NO1O2M2O1N2O1N3N1N2O1N3QOVHZLn7`3bHQLc7j3S1J7K4M3M3M4L4L4L5K4L4L3M4K4M3L4M4K4MY1fN4M3L3N3L4M3Lgm;"}, "label": "a woman wearing jeans and a black shirt with her arms crossed"}]}
{"id": 561590, "image": "COCO_train2014_000000561590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black shirt watches the horse\"."}], "task": "refering", "answer_template": "The \"a woman in a black shirt watches the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 180, 181, 203, 226, 249], "bbox": [0.8006249999999999, 0.0853864168618267, 0.9550468750000001, 0.7730444964871194], "iscrowd": 0, "area": 13267.265150000001, "rle": {"size": [427, 640], "counts": "dde61W=4K6K4K6K4lCZO];l0ZD[Oc;X1L5L3L4M3M7I:F;E=C<D9J5J6J6K4K8H8I7H7I8I7H8I7I6J7I7H8D<_O3MG?A`0_Oa0@`0@`0M51O100O1O1O101N1O110O2N1O1O2O0O2N2N2NO1O2M2O1N2O1N3N1N2O1N3QOVHZLn7`3bHQLc7j3S1J7K4M3M3M4L4L4L5K4L4L3M4K4M3L4M4K4MY1fN4M3L3N3L4M3Lgm;"}, "label": "a woman in a black shirt watches the horse"}]}
{"id": 561590, "image": "COCO_train2014_000000561590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"saddle bag on girls back\"."}], "task": "refering", "answer_template": "The \"saddle bag on girls back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 238, 239, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 328, 329, 330], "bbox": [0.29046875, 0.5895784543325527, 0.42965625, 0.9863934426229508], "iscrowd": 0, "area": 7699.395100000001, "rle": {"size": [427, 640], "counts": "`m]2m0\\<5K6J5K5L4K5K5K2N2O1N1O2N2N2N2O0O2N2N2N1O2O1N2N2N1O2O1N2N2N1O2N2N2N2N1O2M3N2N1O2N2N2M3N1O2N2N2N2M2O2N2N2N2N1N3N2N2NXKdG^4a8aKbGS4Q9H9L3M4L3N2M4L3M4L3M4TOXFQNj9j1n0K6K4K6K5M3L4M3M2N3M3M3M3L4M3M2N3Mkmg4"}, "label": "saddle bag on girls back"}]}
{"id": 561590, "image": "COCO_train2014_000000561590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a pink and black bag beneath a horse ' s head\"."}], "task": "refering", "answer_template": "The \"the side of a pink and black bag beneath a horse ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 238, 239, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 328, 329, 330], "bbox": [0.29046875, 0.5895784543325527, 0.42965625, 0.9863934426229508], "iscrowd": 0, "area": 7699.395100000001, "rle": {"size": [427, 640], "counts": "`m]2m0\\<5K6J5K5L4K5K5K2N2O1N1O2N2N2N2O0O2N2N2N1O2O1N2N2N1O2O1N2N2N1O2N2N2N2N1O2M3N2N1O2N2N2M3N1O2N2N2N2M2O2N2N2N2N1N3N2N2NXKdG^4a8aKbGS4Q9H9L3M4L3N2M4L3M4L3M4TOXFQNj9j1n0K6K4K6K5M3L4M3M2N3M3M3M3L4M3M2N3Mkmg4"}, "label": "the side of a pink and black bag beneath a horse ' s head"}]}
{"id": 479670, "image": "COCO_train2014_000000479670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"long green vegetable\"."}], "task": "refering", "answer_template": "The \"long green vegetable\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [254, 255, 256, 277, 278, 279, 280, 300, 301, 302, 303, 304, 324, 325, 326, 327, 328, 329, 350, 351, 352, 353, 354, 376, 377, 378, 379], "bbox": [0.046656249999999996, 0.6493528183716075, 0.4879375, 0.9844258872651357], "iscrowd": 0, "area": 10223.06035, "rle": {"size": [479, 640], "counts": "U\\>7_>:L3M3L5L3M3M4N2N2N2N2N2N2N2O1N2N2N4L4L3M4L4L4L10010O0001O01O01O01O01O00010O000001O00000000O100000000000002N2N1O2N2N1O2N1O2O0O1O2O0O2O0O2N100O2O0O2N100O1O100O100O00100O100O1O100O1O100O100O1O100O1O100O100O100O10O01O100O100O10O0100O100O010O010O010O0010O010O010O010O010O0010O010O010O010O0010O010O010O01O10O010O01O10O01O010O10O01O010O1O010O010O1O010O0010O0100O0010O01O10O01O010O10O01O010O0010O010O01O010O0010O010O01O01O01O010O010O0010O01O010O010O0010O00010O010O01O01O01O01O010O00010O000010O01O01O00001O00001O01O0001O00001O00001O00001O00001O0000001O00010O0O1N3N1O1N3NZoh4"}, "label": "long green vegetable"}]}
{"id": 479670, "image": "COCO_train2014_000000479670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stalk of broccoli on a plate with other vegetables\"."}], "task": "refering", "answer_template": "The \"a stalk of broccoli on a plate with other vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [254, 255, 256, 277, 278, 279, 280, 300, 301, 302, 303, 304, 324, 325, 326, 327, 328, 329, 350, 351, 352, 353, 354, 376, 377, 378, 379], "bbox": [0.046656249999999996, 0.6493528183716075, 0.4879375, 0.9844258872651357], "iscrowd": 0, "area": 10223.06035, "rle": {"size": [479, 640], "counts": "U\\>7_>:L3M3L5L3M3M4N2N2N2N2N2N2N2O1N2N2N4L4L3M4L4L4L10010O0001O01O01O01O01O00010O000001O00000000O100000000000002N2N1O2N2N1O2N1O2O0O1O2O0O2O0O2N100O2O0O2N100O1O100O100O00100O100O1O100O1O100O100O1O100O1O100O100O100O10O01O100O100O10O0100O100O010O010O010O0010O010O010O010O010O0010O010O010O010O0010O010O010O01O10O010O01O10O01O010O10O01O010O1O010O010O1O010O0010O0100O0010O01O10O01O010O10O01O010O0010O010O01O010O0010O010O01O01O01O010O010O0010O01O010O010O0010O00010O010O01O01O01O01O010O00010O000010O01O01O00001O00001O01O0001O00001O00001O00001O00001O0000001O00010O0O1N3N1O1N3NZoh4"}, "label": "a stalk of broccoli on a plate with other vegetables"}]}
{"id": 479670, "image": "COCO_train2014_000000479670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"head of a large piece of broccoli closest to the carrots\"."}], "task": "refering", "answer_template": "The \"head of a large piece of broccoli closest to the carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 233, 234, 256, 257, 258, 279, 280, 281, 282, 283, 303, 304, 305, 306], "bbox": [0.1315, 0.539937369519833, 0.331046875, 0.7930480167014614], "iscrowd": 0, "area": 7182.207300000002, "rle": {"size": [479, 640], "counts": "TRX13l>2M3N2M3N2gNMaC6Y<8]CI]<f0XC]Oc<h0[CZOc<b1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O1N2O1O1O1O2N1O1O1O1O1O1O2N1O001O0000000101N2N2N101N2fM`DQ1b;fNeDZ1\\;^NlDb1U;UNREk1P;QNSEn1P<N101O0O2O1O0O2O001O0O2O1O1N2O1O1N2O1O001O1O100O1O1O1O100O1O1O1O010O1O1O100O1O1O1O100O001O1O1O10O01O1O1O100O001O1O1O10O01O1O1O101N2N3M2N3N1NijW6"}, "label": "head of a large piece of broccoli closest to the carrots"}]}
{"id": 29116, "image": "COCO_train2014_000000029116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt skateboarding\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 143, 144, 147, 148, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 187], "bbox": [0.32072769953051644, 0.141578125, 0.9351643192488263, 0.539328125], "iscrowd": 0, "area": 28562.517099999994, "rle": {"size": [640, 426], "counts": "Rle22lc04M3M3M2M4M3M3L3N3M3M3L4M0SNlNo@U1P?SOh@m0X?\\O^@e0b?CV@>i?Jo_O6Q`02g_ONY`0:^_OHa`0k101O000O2O001O00001O0f@dLQ>\\3oAfLo=[3PBgLn=Y3RBiLl=X3SBiLl=W3SBlLk=a2RAeMS1Li=`2VAbMP10h=_2[A]Mm06f=^2`AYMi0;d=^2fASMf0a0b=]2jAoLd0f0`=\\2nBgMP=Z2oBhMo<Y2PCiMn<X2QCjMm<W2RCkMl<V2SClMk<U2SCmMl<T2SCnMk<S2TCoMj<R2UCPNi<Q2VCQNh<P2WCRNg<o1XCSNf<n1XCUNe<m1ZCUNd<l1[CVNd<j1[CXNc<i1\\CYNc<g1\\C[Nb<f1]C[Nc<e1[C^Nc<c1\\C_Nb<b1]C`Nb<`1]CbNa<_1^CcNa<]1^CeN`<[1_CgNa<Y1^ChNa<Y1^ChNa<Y1^ChNb<X1]CiNb<X1]ChNd<X1[CiNd<X1ZCjNf<V1YCkNf<V1YCkNf<V1YCUN]=k1bB`MS>a2lA_MU>a2jA^MW>c2gA^MZ>b2eA^M[>c2dA]M\\>d2cA[M_>R4001N10[BeJg<Z5UCkJj<T5TCPKk<P5QCUKo<j4mB[KR=d4kBaKU=^4gBgKX=Y4dBlK\\=R4bBRL]=Y5O10000000O1000000000001O00000000000000000000000000O100O100OXLdBn0[=ROfBn0Z=POhBP1W=POkBo0U=oNmBQ1R=oNoBQ1Q=nNPCR1o<mNSCS1m<lNTCT1k<kNWCJ5[Nd<j1YCH5^Na<i1\\CG5`N^<i1^CD7cN[<h1_CC9dNW<j1`C@;fNU<j1`C]O>iNQ<k1bCYO`0kNn;l1bCVOc0nNj;m1cCSOe0POh;l1dCROg0QOd;n1eCnNj0SOb;o1dClNl0UO_;P2eChNo0XO[;Q2gCdNQ1ZOX;R2gCbNS1\\OU;S2hC^NV1^OS;S2hC]NX1]OQ;W2gCZNZ1]OQ;Y2eCWN]1^OP;[2dCSN_1@P;\\2aCQNc1@o:^2^CoMf1Bn:]2]CnMh1Cm:`2ZCjM_?Y2]@dMg?^2V@_Mn?c2P@YMT`0i2a03L3N2NbNU_OHi`06Z_OLc`02`_OO^`0Oe_O2X`0Mj_O5T`0Io_O7P`0GS@8m?EW@:i?DZ@:g?D[@<d?C_@;b?Ca@<_?Bd@=\\?Ag@=_a00001N100O2O000O2O0O10001N100O2O000O2O0O101O0O101O1O0100O010O010O01000O010O010O10O10O010O10O01N101OO001N2O001N2O001O0O2O1O0O2O1O0O2O1O001N101O1N101O1N101OeWa0"}, "label": "a man in an orange shirt skateboarding"}]}
{"id": 29116, "image": "COCO_train2014_000000029116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing orange and doing a skateboard trick\"."}], "task": "refering", "answer_template": "The \"a boy wearing orange and doing a skateboard trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 66, 67, 68, 69, 70, 71, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 143, 144, 147, 148, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 187], "bbox": [0.32072769953051644, 0.141578125, 0.9351643192488263, 0.539328125], "iscrowd": 0, "area": 28562.517099999994, "rle": {"size": [640, 426], "counts": "Rle22lc04M3M3M2M4M3M3L3N3M3M3L4M0SNlNo@U1P?SOh@m0X?\\O^@e0b?CV@>i?Jo_O6Q`02g_ONY`0:^_OHa`0k101O000O2O001O00001O0f@dLQ>\\3oAfLo=[3PBgLn=Y3RBiLl=X3SBiLl=W3SBlLk=a2RAeMS1Li=`2VAbMP10h=_2[A]Mm06f=^2`AYMi0;d=^2fASMf0a0b=]2jAoLd0f0`=\\2nBgMP=Z2oBhMo<Y2PCiMn<X2QCjMm<W2RCkMl<V2SClMk<U2SCmMl<T2SCnMk<S2TCoMj<R2UCPNi<Q2VCQNh<P2WCRNg<o1XCSNf<n1XCUNe<m1ZCUNd<l1[CVNd<j1[CXNc<i1\\CYNc<g1\\C[Nb<f1]C[Nc<e1[C^Nc<c1\\C_Nb<b1]C`Nb<`1]CbNa<_1^CcNa<]1^CeN`<[1_CgNa<Y1^ChNa<Y1^ChNa<Y1^ChNb<X1]CiNb<X1]ChNd<X1[CiNd<X1ZCjNf<V1YCkNf<V1YCkNf<V1YCUN]=k1bB`MS>a2lA_MU>a2jA^MW>c2gA^MZ>b2eA^M[>c2dA]M\\>d2cA[M_>R4001N10[BeJg<Z5UCkJj<T5TCPKk<P5QCUKo<j4mB[KR=d4kBaKU=^4gBgKX=Y4dBlK\\=R4bBRL]=Y5O10000000O1000000000001O00000000000000000000000000O100O100OXLdBn0[=ROfBn0Z=POhBP1W=POkBo0U=oNmBQ1R=oNoBQ1Q=nNPCR1o<mNSCS1m<lNTCT1k<kNWCJ5[Nd<j1YCH5^Na<i1\\CG5`N^<i1^CD7cN[<h1_CC9dNW<j1`C@;fNU<j1`C]O>iNQ<k1bCYO`0kNn;l1bCVOc0nNj;m1cCSOe0POh;l1dCROg0QOd;n1eCnNj0SOb;o1dClNl0UO_;P2eChNo0XO[;Q2gCdNQ1ZOX;R2gCbNS1\\OU;S2hC^NV1^OS;S2hC]NX1]OQ;W2gCZNZ1]OQ;Y2eCWN]1^OP;[2dCSN_1@P;\\2aCQNc1@o:^2^CoMf1Bn:]2]CnMh1Cm:`2ZCjM_?Y2]@dMg?^2V@_Mn?c2P@YMT`0i2a03L3N2NbNU_OHi`06Z_OLc`02`_OO^`0Oe_O2X`0Mj_O5T`0Io_O7P`0GS@8m?EW@:i?DZ@:g?D[@<d?C_@;b?Ca@<_?Bd@=\\?Ag@=_a00001N100O2O000O2O0O10001N100O2O000O2O0O101O0O101O1O0100O010O010O01000O010O010O10O10O010O10O01N101OO001N2O001N2O001O0O2O1O0O2O1O0O2O1O001N101O1N101O1N101OeWa0"}, "label": "a boy wearing orange and doing a skateboard trick"}]}
{"id": 545213, "image": "COCO_train2014_000000545213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in brown that is swinging the controller\"."}], "task": "refering", "answer_template": "The \"the woman in brown that is swinging the controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 360, 361, 362, 363, 364, 382, 383, 384, 385, 386], "bbox": [0.633703125, 0.3955, 0.9286562500000001, 0.9955], "iscrowd": 0, "area": 32179.222700000002, "rle": {"size": [480, 640], "counts": "ihn53j>3M3N2M3N2M3N2M3M3N2M3N2M3N2M3gMiNgFY1W9lNcFW1\\9nN]FU1b9POXFR1g9ROSFQ1l9TOmEo0R:UOeEQ1Z:TO\\ER1c:ROUES1j:b1O1O1O1O1N200O1O100O1O100O1O100O1O100O1jMaLjH`3k6PMoHR3g6\\MSIe2e6gMVIZ2c6mM[IU2c6nM[IS2d6nM[IS2c6PN[IQ2c6RN[Io1c6SN\\In1b6UN\\Il1c6VN[Ik1e6UNZIl1e6VNYIk1g6UNXIl1g6VNWIk1h6WNVIj1j6V3O100O100O1O10000000000000000000000000000000000000000O1000000001O000001O0000000000000000000000000000M3K5K6H8H7J7I7N2M3N2N1N3N2M3N2M3N2M3M3M3M4L3M7I7I7I6J6J6J6J6J5K4M3L4L4M3L3M4MnMfM`GX2^8\\NQGc1m8fNmFZ1P9oNjFP1U9YOdFf0]9]2100O1N3I6J6K5K5K6J5O2N2O001O1N2O2N1N3TM[EZ1f:eNoEf0S:WOTFb0o9\\OVF?l9^O[F<f9C_F6c9IcFO`90gFG]96jFBY9<f2N4M5JP\\e0"}, "label": "the woman in brown that is swinging the controller"}]}
{"id": 545213, "image": "COCO_train2014_000000545213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in brown top with her hand in the air\"."}], "task": "refering", "answer_template": "The \"girl in brown top with her hand in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 360, 361, 362, 363, 364, 382, 383, 384, 385, 386], "bbox": [0.633703125, 0.3955, 0.9286562500000001, 0.9955], "iscrowd": 0, "area": 32179.222700000002, "rle": {"size": [480, 640], "counts": "ihn53j>3M3N2M3N2M3N2M3M3N2M3N2M3N2M3gMiNgFY1W9lNcFW1\\9nN]FU1b9POXFR1g9ROSFQ1l9TOmEo0R:UOeEQ1Z:TO\\ER1c:ROUES1j:b1O1O1O1O1N200O1O100O1O100O1O100O1O100O1jMaLjH`3k6PMoHR3g6\\MSIe2e6gMVIZ2c6mM[IU2c6nM[IS2d6nM[IS2c6PN[IQ2c6RN[Io1c6SN\\In1b6UN\\Il1c6VN[Ik1e6UNZIl1e6VNYIk1g6UNXIl1g6VNWIk1h6WNVIj1j6V3O100O100O1O10000000000000000000000000000000000000000O1000000001O000001O0000000000000000000000000000M3K5K6H8H7J7I7N2M3N2N1N3N2M3N2M3N2M3M3M3M4L3M7I7I7I6J6J6J6J6J5K4M3L4L4M3L3M4MnMfM`GX2^8\\NQGc1m8fNmFZ1P9oNjFP1U9YOdFf0]9]2100O1N3I6J6K5K5K6J5O2N2O001O1N2O2N1N3TM[EZ1f:eNoEf0S:WOTFb0o9\\OVF?l9^O[F<f9C_F6c9IcFO`90gFG]96jFBY9<f2N4M5JP\\e0"}, "label": "girl in brown top with her hand in the air"}]}
{"id": 545213, "image": "COCO_train2014_000000545213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman looking at the camera\"."}], "task": "refering", "answer_template": "The \"the woman looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 216, 217, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 376, 377, 378], "bbox": [0.295921875, 0.4399166666666667, 0.51190625, 0.9886666666666666], "iscrowd": 0, "area": 21168.63315, "rle": {"size": [480, 640], "counts": "YPi21k>6J6J6L4M1N100O100O101N100O1O100O2O0O100O100O2O0O100O100O2O0O100O100O2UCkNj;U1iCDh;W2oNP1POf0M3O1N2O1N3N1O1N2O1O1N2O1N2O1O1YNdJhI^5l5QKoIQ5k5WKPJl4o5WKlIl4T6VKhIl4X6WKdIj4[6YKcIg4]6[KaIe4_6]K^Id4a6_K]Ia4c6bKZI^4f6dKWI]4h6fKVIZ4j6hKTIX4l6n1O10000O10000O10000O10000O1001O001O00001O001O001O001O00001O001O00101N3M3YIlH\\5V7[JRId5Q7SJVIl5m6jI[IR6k7I8H8H8G8I8M3N2N2M2O2N2N2M2O2N2M4M3E;D<D<D<G9I7I7J8G<D=C<D<Efma4"}, "label": "the woman looking at the camera"}]}
{"id": 545213, "image": "COCO_train2014_000000545213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a brown hooded sweatshirt with her mouth open and eyes wide\"."}], "task": "refering", "answer_template": "The \"a woman wearing a brown hooded sweatshirt with her mouth open and eyes wide\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 192, 193, 194, 216, 217, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 376, 377, 378], "bbox": [0.295921875, 0.4399166666666667, 0.51190625, 0.9886666666666666], "iscrowd": 0, "area": 21168.63315, "rle": {"size": [480, 640], "counts": "YPi21k>6J6J6L4M1N100O100O101N100O1O100O2O0O100O100O2O0O100O100O2O0O100O100O2UCkNj;U1iCDh;W2oNP1POf0M3O1N2O1N3N1O1N2O1O1N2O1N2O1O1YNdJhI^5l5QKoIQ5k5WKPJl4o5WKlIl4T6VKhIl4X6WKdIj4[6YKcIg4]6[KaIe4_6]K^Id4a6_K]Ia4c6bKZI^4f6dKWI]4h6fKVIZ4j6hKTIX4l6n1O10000O10000O10000O10000O1001O001O00001O001O001O001O00001O001O00101N3M3YIlH\\5V7[JRId5Q7SJVIl5m6jI[IR6k7I8H8H8G8I8M3N2N2M2O2N2N2M2O2N2M4M3E;D<D<D<G9I7I7J8G<D=C<D<Efma4"}, "label": "a woman wearing a brown hooded sweatshirt with her mouth open and eyes wide"}]}
{"id": 569795, "image": "COCO_train2014_000000569795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white electrical toothbrush being held by a man\"."}], "task": "refering", "answer_template": "The \"white electrical toothbrush being held by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193], "bbox": [0.67222, 0.5354933333333334, 0.9589200000000001, 0.8291999999999999], "iscrowd": 0, "area": 5078.8995, "rle": {"size": [375, 500], "counts": "mZk32a;6I8I6J5M1O2N1O2N1N3N1O1O2N1O2N1N3N1O2N1O2O0010O010O01O01O01O010O0010O01O010O01O01O010O0010O010O00010O01O010O010O01O100O100O1O010O100O1O100O100O010O1O100O100O1O100O2O0O100O1O1O1N2N2N2N2N2N2L4K5O10O010O010OM4O0010O010O010O01O010O010O0010O010O01O010O010O010O0010O010O01O00001O001O0O2O0Ol_7"}, "label": "white electrical toothbrush being held by a man"}]}
{"id": 569795, "image": "COCO_train2014_000000569795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toothbrush in man ' s left hand\"."}], "task": "refering", "answer_template": "The \"toothbrush in man ' s left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193], "bbox": [0.67222, 0.5354933333333334, 0.9589200000000001, 0.8291999999999999], "iscrowd": 0, "area": 5078.8995, "rle": {"size": [375, 500], "counts": "mZk32a;6I8I6J5M1O2N1O2N1N3N1O1O2N1O2N1N3N1O2N1O2O0010O010O01O01O01O010O0010O01O010O01O01O010O0010O010O00010O01O010O010O01O100O100O1O010O100O1O100O100O010O1O100O100O1O100O2O0O100O1O1O1N2N2N2N2N2N2L4K5O10O010O010OM4O0010O010O010O01O010O010O0010O010O01O010O010O010O0010O010O01O00001O001O0O2O0Ol_7"}, "label": "toothbrush in man ' s left hand"}]}
{"id": 569795, "image": "COCO_train2014_000000569795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of the man who is brushing his teeth\"."}], "task": "refering", "answer_template": "The \"the reflection of the man who is brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 111, 112, 113, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.03048, 0.37866666666666665, 0.3894, 0.9858933333333334], "iscrowd": 0, "area": 27133.89535, "rle": {"size": [375, 500], "counts": "Sk56m:d0]Oc0\\Od0]Oc0\\Od0B>O1O1O100O1O1O1O1O1O1O100O10000000000000000O100000000000000000000O10000000000000000000001O0000000000000000000O1O1O1O1O1O1O1M3K5L4G9YOg0L4K5M3O1N2N2N2N2N2O1O1N2O1N2O1O1N2O1O1O1O100O100O1O1000000O10000O1000000O100001O001O001O1O001O001cKZKg1f4WN]Kh1d4VN^Ki1c4VN_Kh1a4XN`Kg1a4XN`Kg1`4YNaKf1`4YNbKe1_4YNcKf1]4ZNdKe1]4ZNdKe1\\4[NeKd1\\4ZNgKd1Y4[NiKd1X4ZNjKe1W4YNkKf1U4YNmKf1T4XNnKg1S4WNPLg1R4WNoKh1R4UNRLi1o3TNTLk1m3QNWLn1j3oMXLQ2i3lMZLS2g3jM\\LU2h3eMZL[2k3_MWL`2l3\\MULd2V4PMlKP3^6O001O001O1O001O1O001O001O1O001O001O1O008Hj0VOj0VO;E4L5K4L4L4LgZ_3"}, "label": "the reflection of the man who is brushing his teeth"}]}
{"id": 569795, "image": "COCO_train2014_000000569795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mirror image of a man brushing his teeth\"."}], "task": "refering", "answer_template": "The \"mirror image of a man brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 111, 112, 113, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.03048, 0.37866666666666665, 0.3894, 0.9858933333333334], "iscrowd": 0, "area": 27133.89535, "rle": {"size": [375, 500], "counts": "Sk56m:d0]Oc0\\Od0]Oc0\\Od0B>O1O1O100O1O1O1O1O1O1O100O10000000000000000O100000000000000000000O10000000000000000000001O0000000000000000000O1O1O1O1O1O1O1M3K5L4G9YOg0L4K5M3O1N2N2N2N2N2O1O1N2O1N2O1O1N2O1O1O1O100O100O1O1000000O10000O1000000O100001O001O001O1O001O001cKZKg1f4WN]Kh1d4VN^Ki1c4VN_Kh1a4XN`Kg1a4XN`Kg1`4YNaKf1`4YNbKe1_4YNcKf1]4ZNdKe1]4ZNdKe1\\4[NeKd1\\4ZNgKd1Y4[NiKd1X4ZNjKe1W4YNkKf1U4YNmKf1T4XNnKg1S4WNPLg1R4WNoKh1R4UNRLi1o3TNTLk1m3QNWLn1j3oMXLQ2i3lMZLS2g3jM\\LU2h3eMZL[2k3_MWL`2l3\\MULd2V4PMlKP3^6O001O001O1O001O1O001O001O1O001O001O1O008Hj0VOj0VO;E4L5K4L4L4LgZ_3"}, "label": "mirror image of a man brushing his teeth"}]}
{"id": 569795, "image": "COCO_train2014_000000569795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man looking at reflection brushing teeth\"."}], "task": "refering", "answer_template": "The \"man looking at reflection brushing teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 53, 71, 88, 89, 106, 107, 124, 125, 138, 142, 143, 154, 155, 156, 157, 161, 171, 172, 173, 174, 176, 179, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.41952, 0.07461333333333334, 1.0, 1.0], "iscrowd": 0, "area": 23282.25385, "rle": {"size": [375, 500], "counts": "aX]21e;2N2N2N1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1N2O1N2O1N2O1N2O1N2O1N2O1M3M3M3M3M3M3M3M3M3M3M3L4M3N2M3N2M3N2N2M3N2M300O100O1O100O100O100O100O1O100O100O100O100O1O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1O001N2O001O1N101OgMdH9Z7GlH4T7JTI1k6NVI3h6MYI3g6L[I4d6K]I5b6JaI6^6IcI7]6HdI9\\6EfI:[6CgI=[6@gI`0Y6^OhIb0Y6\\OhId0Y6YOjIg0W6VOjIj0W6TOkIk0V6SOkIn0U6oNmIQ1U6lNmIS1T6lNlIU1T6jNmIU1T6jNlIV1V6iNiIW1X6hNiIX1W6gNiIY1X6fNiIY1X6gNgIY1[6eNeI[1\\6dNeI[1\\6eNcI[1^6dNcI[1_6cNaI]1`6bN`I^1a6bN_I^1a6`NaI_1a6^NaIa1`6]NaIc1`6[NbId1Y8O1O00100O1O001O1O1O00100O1O001O100O001O1O10O01O1O10O01O1O1O010O1O1O2O1N1O2M3M3M2N3N2M2N3M3M3MfY83VfG4L5L4K5K5K4L5K3M3M3M3N2N2O001O1N2O1VGWNe7j1ZHZNb7f1mG_NGNW8e1oG_NJMV8e1mG`NNLS8e1mG`N0LR8e1kGbN3IP8g1jGbN6In7e1kGcN8Ha6F]Jf2SOdM^6K[Ja2WOfM\\6NXJ]2\\OeM?BP5b0oJY2AdM>En4c0oJU2DdM=Hl4e0nJP2HdM<Kj4f0nJl1KdM;Ni4g0kJj1ObM;2f4h0kJe12cM;4d4j3PKSL:7b4h3SKRL99a4`4]KcKa4^4]KeK`4]4^KfK_4\\4`KeK^4]4`KfK]4l5N2N2N2N2M2O2N2N2N2N2N2N2M2TO"}, "label": "man looking at reflection brushing teeth"}]}
{"id": 569795, "image": "COCO_train2014_000000569795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man brushing with his left hand\"."}], "task": "refering", "answer_template": "The \"a man brushing with his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 53, 71, 88, 89, 106, 107, 124, 125, 138, 142, 143, 154, 155, 156, 157, 161, 171, 172, 173, 174, 176, 179, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.41952, 0.07461333333333334, 1.0, 1.0], "iscrowd": 0, "area": 23282.25385, "rle": {"size": [375, 500], "counts": "aX]21e;2N2N2N1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1N2O1N2O1N2O1N2O1N2O1N2O1M3M3M3M3M3M3M3M3M3M3M3L4M3N2M3N2M3N2N2M3N2M300O100O1O100O100O100O100O1O100O100O100O100O1O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1O001N2O001O1N101OgMdH9Z7GlH4T7JTI1k6NVI3h6MYI3g6L[I4d6K]I5b6JaI6^6IcI7]6HdI9\\6EfI:[6CgI=[6@gI`0Y6^OhIb0Y6\\OhId0Y6YOjIg0W6VOjIj0W6TOkIk0V6SOkIn0U6oNmIQ1U6lNmIS1T6lNlIU1T6jNmIU1T6jNlIV1V6iNiIW1X6hNiIX1W6gNiIY1X6fNiIY1X6gNgIY1[6eNeI[1\\6dNeI[1\\6eNcI[1^6dNcI[1_6cNaI]1`6bN`I^1a6bN_I^1a6`NaI_1a6^NaIa1`6]NaIc1`6[NbId1Y8O1O00100O1O001O1O1O00100O1O001O100O001O1O10O01O1O10O01O1O1O010O1O1O2O1N1O2M3M3M2N3N2M2N3M3M3MfY83VfG4L5L4K5K5K4L5K3M3M3M3N2N2O001O1N2O1VGWNe7j1ZHZNb7f1mG_NGNW8e1oG_NJMV8e1mG`NNLS8e1mG`N0LR8e1kGbN3IP8g1jGbN6In7e1kGcN8Ha6F]Jf2SOdM^6K[Ja2WOfM\\6NXJ]2\\OeM?BP5b0oJY2AdM>En4c0oJU2DdM=Hl4e0nJP2HdM<Kj4f0nJl1KdM;Ni4g0kJj1ObM;2f4h0kJe12cM;4d4j3PKSL:7b4h3SKRL99a4`4]KcKa4^4]KeK`4]4^KfK_4\\4`KeK^4]4`KfK]4l5N2N2N2N2M2O2N2N2N2N2N2N2M2TO"}, "label": "a man brushing with his left hand"}]}
{"id": 504259, "image": "COCO_train2014_000000504259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheep standing in the pasture next to a sitting sheep\"."}], "task": "refering", "answer_template": "The \"a white sheep standing in the pasture next to a sitting sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 104, 105, 106, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 293, 294, 295, 316, 317, 340], "bbox": [0.44821875, 0.2455373831775701, 0.998109375, 0.9543224299065421], "iscrowd": 0, "area": 50000.07244999999, "rle": {"size": [428, 640], "counts": "lTh32Y=2N3M3M3M3M3M3N1N1N2O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2N2O1N2M3N2M3N2N2N2N2O1N2N2N2N2N2000000000000000000000000O10000000000000000000000000000000001N2O1O10O01O1O1O1O100O1O001O1O1O100O1O001O1O1O100O1O001O1O100O1O1O001O1O100O1O1O1O001O100O1O1O1O00100O1O1O1O1O010O1O1O1O1O10O01O1O1O1O100O1O001O1O1O100O1O001O1O1O1O1O10O01O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001N2O1O1O1O1kIdJX4\\5]KmJe4T5PKTKR5m4bJ\\K`5e4TJeKm5k52N2N2N2N2N2N2NfJXIT4f6iK_IW4_6fKgIY4W6dKnI\\4P6`KVJ`4h5]K]Jb4b5[KdJd4Z5YKkJg4S5UKSKj4l4SKZKl4d4QKaKn4^4nJhKR5Y6O2O0O101N100O10L3K6K4L4L5J5L5K4L5J5L5N101O0O2O1O0O2O1O1N2O1O1N101O1N2O1O1N2O0O2O1O1N2O1O0O2O1O1N2O1O1N101O1N2O1O1N20O0100000O10000O01000O10000O1000O010000O100O10O10O10000O10000O01000O10000O10O10O10000O10000O01000O10000O10O10O10000O1000O013Ml0TOm0SOm0SOl0TOm0SOm0ROe4"}, "label": "a white sheep standing in the pasture next to a sitting sheep"}]}
{"id": 504259, "image": "COCO_train2014_000000504259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front most goat\"."}], "task": "refering", "answer_template": "The \"the front most goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 104, 105, 106, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 293, 294, 295, 316, 317, 340], "bbox": [0.44821875, 0.2455373831775701, 0.998109375, 0.9543224299065421], "iscrowd": 0, "area": 50000.07244999999, "rle": {"size": [428, 640], "counts": "lTh32Y=2N3M3M3M3M3M3N1N1N2O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2N2O1N2M3N2M3N2N2N2N2O1N2N2N2N2N2000000000000000000000000O10000000000000000000000000000000001N2O1O10O01O1O1O1O100O1O001O1O1O100O1O001O1O1O100O1O001O1O100O1O1O001O1O100O1O1O1O001O100O1O1O1O00100O1O1O1O1O010O1O1O1O1O10O01O1O1O1O100O1O001O1O1O100O1O001O1O1O1O1O10O01O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001N2O1O1O1O1kIdJX4\\5]KmJe4T5PKTKR5m4bJ\\K`5e4TJeKm5k52N2N2N2N2N2N2NfJXIT4f6iK_IW4_6fKgIY4W6dKnI\\4P6`KVJ`4h5]K]Jb4b5[KdJd4Z5YKkJg4S5UKSKj4l4SKZKl4d4QKaKn4^4nJhKR5Y6O2O0O101N100O10L3K6K4L4L5J5L5K4L5J5L5N101O0O2O1O0O2O1O1N2O1O1N101O1N2O1O1N2O0O2O1O1N2O1O0O2O1O1N2O1O1N101O1N2O1O1N20O0100000O10000O01000O10000O1000O010000O100O10O10O10000O10000O01000O10000O10O10O10000O10000O01000O10000O10O10O10000O1000O013Ml0TOm0SOm0SOl0TOm0SOm0ROe4"}, "label": "the front most goat"}]}
{"id": 504259, "image": "COCO_train2014_000000504259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep sitting down in the grass\"."}], "task": "refering", "answer_template": "The \"a sheep sitting down in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 103, 104, 105, 106], "bbox": [0.4099375, 0.09672897196261682, 0.68525, 0.29254672897196266], "iscrowd": 0, "area": 10516.7965, "rle": {"size": [428, 640], "counts": "eo]34W=4L4L4L2N3M2N1O001O1O001N1M4N11000O1000000O1000000O1000O001OZDTOn:l0kD\\OT;c0fDEY;:bDM];S1N1O1O2N001O000010O00010O01O00010O0010O00010O01O01O0000010O00000000000000000001O000000000001O0001O00001O0000001O00001O00001O0000001O0000000000000000000000000000000000000000000000000001O00001O001O000O2O00001N101O000O2O001O0O101O001O0O101O0O2O0O2O1O1N2O0O2O1O1N2O0O2O1N3N1O2M3M2L5K8I?@\\mc2"}, "label": "a sheep sitting down in the grass"}]}
{"id": 504259, "image": "COCO_train2014_000000504259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooly sheep sitting in some sparse grass to one side of a standing sheep\"."}], "task": "refering", "answer_template": "The \"a wooly sheep sitting in some sparse grass to one side of a standing sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 103, 104, 105, 106], "bbox": [0.4099375, 0.09672897196261682, 0.68525, 0.29254672897196266], "iscrowd": 0, "area": 10516.7965, "rle": {"size": [428, 640], "counts": "eo]34W=4L4L4L2N3M2N1O001O1O001N1M4N11000O1000000O1000000O1000O001OZDTOn:l0kD\\OT;c0fDEY;:bDM];S1N1O1O2N001O000010O00010O01O00010O0010O00010O01O01O0000010O00000000000000000001O000000000001O0001O00001O0000001O00001O00001O0000001O0000000000000000000000000000000000000000000000000001O00001O001O000O2O00001N101O000O2O001O0O101O001O0O101O0O2O0O2O1O1N2O0O2O1O1N2O0O2O1N3N1O2M3M2L5K8I?@\\mc2"}, "label": "a wooly sheep sitting in some sparse grass to one side of a standing sheep"}]}
{"id": 152003, "image": "COCO_train2014_000000152003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of carrot on a plate\"."}], "task": "refering", "answer_template": "The \"a slice of carrot on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 353, 354], "bbox": [0.7511041666666666, 0.626375, 1.0, 0.87803125], "iscrowd": 0, "area": 16777.036300000003, "rle": {"size": [640, 480], "counts": "VQR7a0^c0b0]O9H8G9H7H9H8G8I7H7J7H6K4K5M2N3M3M2N3M2N3M3M2N3M2N3M1O000000O10001O0000000000000O10000000000000000O10000000000000000O10001O0000000000000O10000000000000000O10000000000000000O10001O00000000000O10000000002N2N2N2N2M4M2N2N2N2N2M4K4L4L4L4L4LZB"}, "label": "a slice of carrot on a plate"}]}
{"id": 152003, "image": "COCO_train2014_000000152003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of carrot under some red slices of pepper\"."}], "task": "refering", "answer_template": "The \"a piece of carrot under some red slices of pepper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 353, 354], "bbox": [0.7511041666666666, 0.626375, 1.0, 0.87803125], "iscrowd": 0, "area": 16777.036300000003, "rle": {"size": [640, 480], "counts": "VQR7a0^c0b0]O9H8G9H7H9H8G8I7H7J7H6K4K5M2N3M3M2N3M2N3M3M2N3M2N3M1O000000O10001O0000000000000O10000000000000000O10000000000000000O10001O0000000000000O10000000000000000O10000000000000000O10001O00000000000O10000000002N2N2N2N2M4M2N2N2N2N2M4K4L4L4L4L4LZB"}, "label": "a piece of carrot under some red slices of pepper"}]}
{"id": 152003, "image": "COCO_train2014_000000152003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark yellow piece of fruit or vegetable\"."}], "task": "refering", "answer_template": "The \"a dark yellow piece of fruit or vegetable\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [275, 276, 277, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 341, 342, 343, 344, 345, 346, 358, 359, 360, 361, 362, 363], "bbox": [0.0830625, 0.727390625, 0.43675, 0.95040625], "iscrowd": 0, "area": 18345.628650000002, "rle": {"size": [640, 480], "counts": "nai04hc06K4K6J6J5L5J5K4L4L5L3L4L4N3M2O1N2O1N3M2O1N2N3N1N2O1N3M2O1N2O1N3M2O1N2O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O101N100O100O2O0O100O100O10001O000000000000000000001O0000000001O00000001O0000000000000001O01O00000001O00000000010O00000000001O00000001O01O00000000001O0001O000001O00000O1M3M4L3M3N2M3M3M3M4L3N2M3M3M3M3M4M2M3M3M3M3M4M2M3M3M3M3M3N3L3M3M3Mo0QOXXX5"}, "label": "a dark yellow piece of fruit or vegetable"}]}
{"id": 53702, "image": "COCO_train2014_000000053702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with headband\"."}], "task": "refering", "answer_template": "The \"man with headband\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 58, 59, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 110, 111, 112, 113, 115, 116, 117, 128, 129, 130, 131, 132, 144, 145, 146, 147, 149, 150, 162, 168, 169, 186, 187], "bbox": [0.02256, 0.20033033033033032, 0.5321999999999999, 0.882042042042042], "iscrowd": 0, "area": 14574.618149999998, "rle": {"size": [333, 500], "counts": "mi3Z1S9D;E<O0010O10000O100O100O10O10O100O100O010O1000O0100O10O010000O010O100O01000O10O0100O10O01O1O1N101O1N2O001N2O1O0O2O1N2O001N2O1O0O2OO01O1O1O1O0O2O1O1O1N2O001O1O1N2O001O1N2O101O01O010O01O010O01O01O010O0010OZOSNmHn1R7ZNgHe1Z7]NcHc1]7g00O01O010O010O2O0O101O0O2O00001O001O000O2O00001O001O00001O000O2O001OjNSIlNl6S1XIkNh6S1]IkNb6S1bIkN^6S1fIlNX6S1lIkNT6S1PJkNP6S1SJmNl5Q1WJnNh5Q1[JPOc5n0`JVO[5j0fJZOU5f0lJ]Oo4c0SKAh4?YKEb4;_KI\\46fKNQ4]NbJf1_10f37[LM\\36fLNP37QMMf27[MM[28fMLT24QNLk14ZNKa16cNF\\1:jNBU1>nN_OQ1b0SOZOm0e0WOYOg0h0]OTOb0m0APO?P1ElN<T1FjN;U1HgN9X1JeN7[1LbN5=`L@\\315=aLAZ318:aLEW3099cLFS31;7dLGR30=5dLKn20?3eLMl2Oa00fL0i2Oc0JlL6`20f0BQM>Y2Og0\\OWMe0Q2NX21iMNX22gMN[21eMN\\21dMN^22bMM_23`MMa22`MLc23\\MMe23[MLf24YMKi24XMKi25VMJl26SMHQ36oLIS37mLFV3:iLEY3:gLD\\3<cLC_3=`LAc3>]LAe3?U30O10O10000O01000O100000O010000O10O10O10000O010000O001O1O001O1O001O1O00O1O100O1O1N2N2M3N2M3Nbm[2"}, "label": "man with headband"}]}
{"id": 53702, "image": "COCO_train2014_000000053702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man playing with headband\"."}], "task": "refering", "answer_template": "The \"man playing with headband\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 58, 59, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 110, 111, 112, 113, 115, 116, 117, 128, 129, 130, 131, 132, 144, 145, 146, 147, 149, 150, 162, 168, 169, 186, 187], "bbox": [0.02256, 0.20033033033033032, 0.5321999999999999, 0.882042042042042], "iscrowd": 0, "area": 14574.618149999998, "rle": {"size": [333, 500], "counts": "mi3Z1S9D;E<O0010O10000O100O100O10O10O100O100O010O1000O0100O10O010000O010O100O01000O10O0100O10O01O1O1N101O1N2O001N2O1O0O2O1N2O001N2O1O0O2OO01O1O1O1O0O2O1O1O1N2O001O1O1N2O001O1N2O101O01O010O01O010O01O01O010O0010OZOSNmHn1R7ZNgHe1Z7]NcHc1]7g00O01O010O010O2O0O101O0O2O00001O001O000O2O00001O001O00001O000O2O001OjNSIlNl6S1XIkNh6S1]IkNb6S1bIkN^6S1fIlNX6S1lIkNT6S1PJkNP6S1SJmNl5Q1WJnNh5Q1[JPOc5n0`JVO[5j0fJZOU5f0lJ]Oo4c0SKAh4?YKEb4;_KI\\46fKNQ4]NbJf1_10f37[LM\\36fLNP37QMMf27[MM[28fMLT24QNLk14ZNKa16cNF\\1:jNBU1>nN_OQ1b0SOZOm0e0WOYOg0h0]OTOb0m0APO?P1ElN<T1FjN;U1HgN9X1JeN7[1LbN5=`L@\\315=aLAZ318:aLEW3099cLFS31;7dLGR30=5dLKn20?3eLMl2Oa00fL0i2Oc0JlL6`20f0BQM>Y2Og0\\OWMe0Q2NX21iMNX22gMN[21eMN\\21dMN^22bMM_23`MMa22`MLc23\\MMe23[MLf24YMKi24XMKi25VMJl26SMHQ36oLIS37mLFV3:iLEY3:gLD\\3<cLC_3=`LAc3>]LAe3?U30O10O10000O01000O100000O010000O10O10O10000O010000O001O1O001O1O001O1O00O1O100O1O1N2N2M3N2M3Nbm[2"}, "label": "man playing with headband"}]}
{"id": 70094, "image": "COCO_train2014_000000070094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"drink can\"."}], "task": "refering", "answer_template": "The \"drink can\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 15, 16, 17, 18, 19, 31, 32, 33, 34, 46, 47, 48, 49, 61, 62, 63, 64, 76, 77, 78], "bbox": [0.052149532710280375, 0.0, 0.3212616822429907, 0.24028125], "iscrowd": 0, "area": 14610.3507, "rle": {"size": [640, 428], "counts": "Rh=1oc0;E<D<D;E<D;E<D<D;E<D:F1O2N1N2O1O1O1O2N1O1O1O1O1O2N1O1O000000001O00000000001O00000000001O00000000001O000000001O00000000001O00O1O10O01O100O1O1O100O1O1O100O1O100O1O1O100O1O1N2O1O1N2O1O1O2M2O1O1N3N1O1N3N1O2M5L3M3L5L3K6I6J6J7I6J7I>AR1oNnUe5"}, "label": "drink can"}]}
{"id": 70094, "image": "COCO_train2014_000000070094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black glass bottle with white writing , next to the coke cupm\"."}], "task": "refering", "answer_template": "The \"the black glass bottle with white writing , next to the coke cupm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 15, 16, 17, 18, 19, 31, 32, 33, 34, 46, 47, 48, 49, 61, 62, 63, 64, 76, 77, 78], "bbox": [0.052149532710280375, 0.0, 0.3212616822429907, 0.24028125], "iscrowd": 0, "area": 14610.3507, "rle": {"size": [640, 428], "counts": "Rh=1oc0;E<D<D;E<D;E<D<D;E<D:F1O2N1N2O1O1O1O2N1O1O1O1O1O2N1O1O000000001O00000000001O00000000001O00000000001O000000001O00000000001O00O1O10O01O100O1O1O100O1O1O100O1O100O1O1O100O1O1N2O1O1N2O1O1O2M2O1O1N3N1O1N3N1O2M5L3M3L5L3K6I6J6J7I6J7I>AR1oNnUe5"}, "label": "the black glass bottle with white writing , next to the coke cupm"}]}
{"id": 70094, "image": "COCO_train2014_000000070094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a paper coca cola cup on a table beside a hot dog\"."}], "task": "refering", "answer_template": "The \"a paper coca cola cup on a table beside a hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 22, 23, 24, 25, 26, 27, 28, 37, 38, 39, 40, 41, 42, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 83, 84, 85, 86, 87, 99, 100, 101], "bbox": [0.5051869158878505, 0.001125, 0.8924766355140187, 0.293921875], "iscrowd": 0, "area": 24877.44275, "rle": {"size": [640, 428], "counts": "RPW44lc0;E<D;E<D;E<D;E<D;E<D;E<D8H1O1O001O001O001O001O001O00001O001O001O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O1O001O1O000000000000000000000000000O10000000001O000000000000000000000000000000O100O100O100O1O1O1O1O1O1O1O1N2O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1K5J6J6J6K5J6J6J6K5J6J6K5J6J6J6K5J6J6J6K5J6J6J6K5J6J6KThl0"}, "label": "a paper coca cola cup on a table beside a hot dog"}]}
{"id": 70094, "image": "COCO_train2014_000000070094.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coca cola beverage container\"."}], "task": "refering", "answer_template": "The \"a coca cola beverage container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 22, 23, 24, 25, 26, 27, 28, 37, 38, 39, 40, 41, 42, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 83, 84, 85, 86, 87, 99, 100, 101], "bbox": [0.5051869158878505, 0.001125, 0.8924766355140187, 0.293921875], "iscrowd": 0, "area": 24877.44275, "rle": {"size": [640, 428], "counts": "RPW44lc0;E<D;E<D;E<D;E<D;E<D;E<D8H1O1O001O001O001O001O001O00001O001O001O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O1O001O1O000000000000000000000000000O10000000001O000000000000000000000000000000O100O100O100O1O1O1O1O1O1O1O1N2O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1K5J6J6J6K5J6J6J6K5J6J6K5J6J6J6K5J6J6J6K5J6J6J6K5J6J6KThl0"}, "label": "a coca cola beverage container"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a black and white striped sweater\"."}], "task": "refering", "answer_template": "The \"a person wearing a black and white striped sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 63, 64, 65, 86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 294, 295], "bbox": [0.6970156249999999, 0.10152083333333332, 0.8932968750000001, 0.7282291666666667], "iscrowd": 0, "area": 25848.788500000006, "rle": {"size": [480, 640], "counts": "V[a65d>9RMDlFf0n8HaFa0Y99jE0P:8dE2U:7`E2Z:6[E4^:[2J6J6I8I7H7J7RGXJZ8o5]GVJ_8X6L3N2000000000001O1O1O1O1O1000000O10000O10000O10000O1000000O1eN^IXJb6g5dITJ]6k5gIPJZ6o5nIjIR6U6UJeIk5Z6^1O2N1O100O1O2O0O100O2N101N1O2O0O2O0O2N101N101N1O2O0010O010O001N1O2M2O2M2M4H8G9QIbIS5e6lJeId4c6YK`I_4h6^KZI[4n6bKUIV4S7gKPIP4Y7nKhHn3^7nKeHo3`7kKdHS4`7eKeHY4`7_KfH^4P9M2O2M3M3M2O2M3M2N3M3M2M4M3M2M4M3M2N3L4M:F;E;D]1dNdeo0"}, "label": "a person wearing a black and white striped sweater"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a striped sweater and a black beanie\"."}], "task": "refering", "answer_template": "The \"a person wearing a striped sweater and a black beanie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 63, 64, 65, 86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 294, 295], "bbox": [0.6970156249999999, 0.10152083333333332, 0.8932968750000001, 0.7282291666666667], "iscrowd": 0, "area": 25848.788500000006, "rle": {"size": [480, 640], "counts": "V[a65d>9RMDlFf0n8HaFa0Y99jE0P:8dE2U:7`E2Z:6[E4^:[2J6J6I8I7H7J7RGXJZ8o5]GVJ_8X6L3N2000000000001O1O1O1O1O1000000O10000O10000O10000O1000000O1eN^IXJb6g5dITJ]6k5gIPJZ6o5nIjIR6U6UJeIk5Z6^1O2N1O100O1O2O0O100O2N101N1O2O0O2O0O2N101N101N1O2O0010O010O001N1O2M2O2M2M4H8G9QIbIS5e6lJeId4c6YK`I_4h6^KZI[4n6bKUIV4S7gKPIP4Y7nKhHn3^7nKeHo3`7kKdHS4`7eKeHY4`7_KfH^4P9M2O2M3M3M2O2M3M2N3M3M2M4M3M2M4M3M2N3L4M:F;E;D]1dNdeo0"}, "label": "a person wearing a striped sweater and a black beanie"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a hat that is dancing\"."}], "task": "refering", "answer_template": "The \"a man in a hat that is dancing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 216, 239], "bbox": [0.197984375, 0.07872916666666667, 0.436234375, 0.6089166666666668], "iscrowd": 0, "area": 23756.526199999997, "rle": {"size": [480, 640], "counts": "Uek1P1c=?M3M3L5L3jCUNR;n1iDWNT;l1gDYNV;j1eD[NX;i1bD\\N[;g1`D^N];e2M3M3M3M3M2N2N2N2N1O2N2N1O2N2N1OO1M3N2N2N2N2M30O3M3L4M3L4M3L4H7H9H8H8M4N3M2O1N3M2N3N3Ld0\\O4L4M3L4L4L10N10000O2O000O101O000O101O0O103L7J7H7J4K2O1N2O1N2O0O2O1N2O1N101N101N1M3I8G8I8O0100O10O01O100O100M3J6J600XG\\JR8e5\\GmJc8h5010YK_Gm2a8RMeGj2Z8VMjGf2W8XMoGd2P8\\MTH`2l7`MYH\\2f7cM_HY2b7fMcHU2]7jMhHS2W7mMnHn1R7QNSIl1m6SNXIh1h6XN\\Ie1c6ZNcIa1]6_NgI^1Y6`NmI[1S6eNQJX1n5gNXJT1i5kN\\JP1d5oNbJm0]5SOhJh0X5XOmJd0S5ZOSKa0m4_OWK>h4A^K9c4FcKOd4NcKGc48cK]Oc4a0dKTOc4i0k4O1N4M5J6K4KoSY5"}, "label": "a man in a hat that is dancing"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the thick beard and glasses , that looks like he ' s doing a silly dance\"."}], "task": "refering", "answer_template": "The \"the man with the thick beard and glasses , that looks like he ' s doing a silly dance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 216, 239], "bbox": [0.197984375, 0.07872916666666667, 0.436234375, 0.6089166666666668], "iscrowd": 0, "area": 23756.526199999997, "rle": {"size": [480, 640], "counts": "Uek1P1c=?M3M3L5L3jCUNR;n1iDWNT;l1gDYNV;j1eD[NX;i1bD\\N[;g1`D^N];e2M3M3M3M3M2N2N2N2N1O2N2N1O2N2N1OO1M3N2N2N2N2M30O3M3L4M3L4M3L4H7H9H8H8M4N3M2O1N3M2N3N3Ld0\\O4L4M3L4L4L10N10000O2O000O101O000O101O0O103L7J7H7J4K2O1N2O1N2O0O2O1N2O1N101N101N1M3I8G8I8O0100O10O01O100O100M3J6J600XG\\JR8e5\\GmJc8h5010YK_Gm2a8RMeGj2Z8VMjGf2W8XMoGd2P8\\MTH`2l7`MYH\\2f7cM_HY2b7fMcHU2]7jMhHS2W7mMnHn1R7QNSIl1m6SNXIh1h6XN\\Ie1c6ZNcIa1]6_NgI^1Y6`NmI[1S6eNQJX1n5gNXJT1i5kN\\JP1d5oNbJm0]5SOhJh0X5XOmJd0S5ZOSKa0m4_OWK>h4A^K9c4FcKOd4NcKGc48cK]Oc4a0dKTOc4i0k4O1N4M5J6K4KoSY5"}, "label": "the man with the thick beard and glasses , that looks like he ' s doing a silly dance"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with glasses wearing plaid\"."}], "task": "refering", "answer_template": "The \"man with glasses wearing plaid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 66, 67, 68, 89, 90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 389, 390], "bbox": [0.8905156249999999, 0.018625, 1.0, 0.98025], "iscrowd": 0, "area": 24401.88970000002, "rle": {"size": [480, 640], "counts": "Y[[85Q48V60dIi0b5^OWJY1]2`NZL>1`2W3ZMbL`0\\OQ30VLi1<jN`0XOX3LnKo1>jNa0TOm4o1eJlNb0oNm4Q2dJPOa0jNn4S2eJRO`0dNo4X2cJTOa0^No4[2cJVOe7f0_HYOa7d0bH[O_7c0dH\\O\\7a0gH^OZ7?iHAW7=lHAU7<nHBT7>lH@U7`0mH^OS7c0mH[OT7e0nHXOS7i0mHUOU7j0lHUOT7l0mHQOT7o0mHoNT7R1mHkNT7U1mHjNS7W1mHgNS7Z1oHcNR7^1nH`NS7`1oH\\NS7e1mHWNW7h1kHTNX7l1n2O100O1XNnCb0Z=M4M2M2O00O00100O1O00100O001N1O2M2YMn0_EUOY:U1bEoN[:V1_EnN^:V1]EmNa:k2M3M3K4M4L4L4eIe0lL^Oa2T1ZMSOR2_1eMYN"}, "label": "man with glasses wearing plaid"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a plaid shirt , beanie , and jeans facing another man in a beanie\"."}], "task": "refering", "answer_template": "The \"a man in a plaid shirt , beanie , and jeans facing another man in a beanie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 66, 67, 68, 89, 90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 389, 390], "bbox": [0.8905156249999999, 0.018625, 1.0, 0.98025], "iscrowd": 0, "area": 24401.88970000002, "rle": {"size": [480, 640], "counts": "Y[[85Q48V60dIi0b5^OWJY1]2`NZL>1`2W3ZMbL`0\\OQ30VLi1<jN`0XOX3LnKo1>jNa0TOm4o1eJlNb0oNm4Q2dJPOa0jNn4S2eJRO`0dNo4X2cJTOa0^No4[2cJVOe7f0_HYOa7d0bH[O_7c0dH\\O\\7a0gH^OZ7?iHAW7=lHAU7<nHBT7>lH@U7`0mH^OS7c0mH[OT7e0nHXOS7i0mHUOU7j0lHUOT7l0mHQOT7o0mHoNT7R1mHkNT7U1mHjNS7W1mHgNS7Z1oHcNR7^1nH`NS7`1oH\\NS7e1mHWNW7h1kHTNX7l1n2O100O1XNnCb0Z=M4M2M2O00O00100O1O00100O001N1O2M2YMn0_EUOY:U1bEoN[:V1_EnN^:V1]EmNa:k2M3M3K4M4L4L4eIe0lL^Oa2T1ZMSOR2_1eMYN"}, "label": "a man in a plaid shirt , beanie , and jeans facing another man in a beanie"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in glasses standing besides a table\"."}], "task": "refering", "answer_template": "The \"a woman in glasses standing besides a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 266], "bbox": [0.541765625, 0.13035416666666666, 0.7009062500000001, 0.6901875000000001], "iscrowd": 0, "area": 17871.19274999999, "rle": {"size": [480, 640], "counts": "_jR5e0U>>QDSOU9T1cFTOW9Q1dFUOY9n0cFVOY9n0cFVOY9m0eFWOV9m0fFWOV9m0gFVOU9m0gFXOV9k0eFZOX9i0cF\\OZ9h0_F^O_9e0YFCc9a0UFFh9k2NT1kN2N101[OcIeH^6X7hIdHZ6X7lIeHU6Y7oIcHR6\\7RJaHo5\\7VJaHh5`7i02N2ONiH_HU7d7kH[HS7h702M3M3N2M3M3N23M5K4L5L3M2M010000O10000O100O01000O10000O10001N2O2N1N2O2N1N2O101N100O2N100O2O0O2O0O1O2L3J6K6]MSGjNS9R1`2K6J5K6J5L4L5K4L4L7I9G9H8G9G9G9G_\\i2"}, "label": "a woman in glasses standing besides a table"}]}
{"id": 528851, "image": "COCO_train2014_000000528851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young lady with glasses and long hair\"."}], "task": "refering", "answer_template": "The \"young lady with glasses and long hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 266], "bbox": [0.541765625, 0.13035416666666666, 0.7009062500000001, 0.6901875000000001], "iscrowd": 0, "area": 17871.19274999999, "rle": {"size": [480, 640], "counts": "_jR5e0U>>QDSOU9T1cFTOW9Q1dFUOY9n0cFVOY9n0cFVOY9m0eFWOV9m0fFWOV9m0gFVOU9m0gFXOV9k0eFZOX9i0cF\\OZ9h0_F^O_9e0YFCc9a0UFFh9k2NT1kN2N101[OcIeH^6X7hIdHZ6X7lIeHU6Y7oIcHR6\\7RJaHo5\\7VJaHh5`7i02N2ONiH_HU7d7kH[HS7h702M3M3N2M3M3N23M5K4L5L3M2M010000O10000O100O01000O10000O10001N2O2N1N2O2N1N2O101N100O2N100O2O0O2O0O1O2L3J6K6]MSGjNS9R1`2K6J5K6J5L4L5K4L4L7I9G9H8G9G9G9G_\\i2"}, "label": "young lady with glasses and long hair"}]}
{"id": 553428, "image": "COCO_train2014_000000553428.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a parked white ford suv\"."}], "task": "refering", "answer_template": "The \"a parked white ford suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 249, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 271, 284, 285], "bbox": [0.6782083333333334, 0.48631250000000004, 1.0, 0.7348750000000001], "iscrowd": 0, "area": 18082.91895, "rle": {"size": [640, 480], "counts": "XS\\65ic06K4M3M4Lj0VO5K6J5K4N2N1K4L4L1O1O2N1O2N1O1O2M2O1O2N1O1O200j_O]Mn>l2c@\\M\\?d300O2O0O1O2N1O1O2N1O1O10O100O1O100O100O1XOh00000000O1000O100000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O01000000000000000000000000000O100000000000006l1nMWD"}, "label": "a parked white ford suv"}]}
{"id": 553428, "image": "COCO_train2014_000000553428.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light colored ford suv parked along the street\"."}], "task": "refering", "answer_template": "The \"a light colored ford suv parked along the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 249, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 271, 284, 285], "bbox": [0.6782083333333334, 0.48631250000000004, 1.0, 0.7348750000000001], "iscrowd": 0, "area": 18082.91895, "rle": {"size": [640, 480], "counts": "XS\\65ic06K4M3M4Lj0VO5K6J5K4N2N1K4L4L1O1O2N1O2N1O1O2M2O1O2N1O1O200j_O]Mn>l2c@\\M\\?d300O2O0O1O2N1O1O2N1O1O10O100O1O100O100O1XOh00000000O1000O100000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O01000000000000000000000000000O100000000000006l1nMWD"}, "label": "a light colored ford suv parked along the street"}]}
{"id": 135638, "image": "COCO_train2014_000000135638.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair bottom left corner\"."}], "task": "refering", "answer_template": "The \"chair bottom left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 351, 369, 370, 374], "bbox": [0.0, 0.48200000000000004, 0.28746875, 0.9852083333333332], "iscrowd": 0, "area": 28609.979999999996, "rle": {"size": [480, 640], "counts": "T<1o>1O2M2O1N3N1O1N2O2M2O1N3N1O1N2O2M2O1O2M2O1N2O2N1N2O2M2O1O1N3N1N2O2N1N2O1N3RNYN_Fh1R9kNiFV1h8]ORGf0b8HZG9d8MVG5g82TG0h86TGKj8;PGGm8`0nFBo8d0kF^OR9[3M2M3N3L3N3L3N2M4N1O2M2O1O2N1O1O0O100000001N100000M3J6J5K6M300O1000O10O10000O0100000O1000O01000000O01000O100000O010000O10O1000O01000O010000O01000O10O01000O0100O10O10O10O010000O010O11N1000001N10000O2O00000O2O000O101O000O101O0O10001O15K5K5K5L4K5K4L5KO1O2lJQJR1Q6cNXK=j4XOgKa0Z4TOVLe0l3POeLh0^3lNRMn0g8H7I8I6IdYe6"}, "label": "chair bottom left corner"}]}
{"id": 135638, "image": "COCO_train2014_000000135638.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dinning chair pushed under the long side of a table and does not have an open back\"."}], "task": "refering", "answer_template": "The \"a dinning chair pushed under the long side of a table and does not have an open back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 351, 369, 370, 374], "bbox": [0.0, 0.48200000000000004, 0.28746875, 0.9852083333333332], "iscrowd": 0, "area": 28609.979999999996, "rle": {"size": [480, 640], "counts": "T<1o>1O2M2O1N3N1O1N2O2M2O1N3N1O1N2O2M2O1O2M2O1N2O2N1N2O2M2O1O1N3N1N2O2N1N2O1N3RNYN_Fh1R9kNiFV1h8]ORGf0b8HZG9d8MVG5g82TG0h86TGKj8;PGGm8`0nFBo8d0kF^OR9[3M2M3N3L3N3L3N2M4N1O2M2O1O2N1O1O0O100000001N100000M3J6J5K6M300O1000O10O10000O0100000O1000O01000000O01000O100000O010000O10O1000O01000O010000O01000O10O01000O0100O10O10O10O010000O010O11N1000001N10000O2O00000O2O000O101O000O101O0O10001O15K5K5K5L4K5K4L5KO1O2lJQJR1Q6cNXK=j4XOgKa0Z4TOVLe0l3POeLh0^3lNRMn0g8H7I8I6IdYe6"}, "label": "a dinning chair pushed under the long side of a table and does not have an open back"}]}
{"id": 365015, "image": "COCO_train2014_000000365015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot basket on the left\"."}], "task": "refering", "answer_template": "The \"the carrot basket on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 175, 176, 189], "bbox": [0.0, 0.57248, 0.6071466666666667, 0.82904], "iscrowd": 0, "area": 11399.226, "rle": {"size": [500, 375], "counts": "l9X1\\>0O1O001O1O10O01O1O1O010O1O1O1O100O10O10O10000O100O10000O10000O10000O100O01000O10000O10000O1000000O10000O10O10O100000000O100000000O100000000O10O101O0000001O0000001O0000000000O10O100000O1000000000000000000000000000O10000000000000000000000000001O000O101O00001O0000001O0000001O001O1O1O2M2O1O1O001O00001O00001O00001O00001O0O1000001O01O1O2N1O1O2O0O2N100O2O0O2O000O2O0O2O0003N:E2OO0O101O001O00001O001O00001OiNnAk0R>TOQBk0n=TOUBj0k=UOXBi0h=VOZBj0e=UO^Bi0b=VOaBi0^=VOeBN@c0k=^OmBb0S=]OnBd0P>100000000000O1O1O1O1N101O1N4M5KikW2"}, "label": "the carrot basket on the left"}]}
{"id": 365015, "image": "COCO_train2014_000000365015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrots in the nearest basket\"."}], "task": "refering", "answer_template": "The \"carrots in the nearest basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 189, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.0, 0.68764, 0.5633066666666667, 0.98652], "iscrowd": 0, "area": 28254.182649999995, "rle": {"size": [500, 375], "counts": "X;`0S?o0QOo0QOo0ROk0TO1O1O10000000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O1000000000000O1000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000001O1O001O1O001O1O2N1O1O2N1O2N1O1O2N1O1O2L3N3M2N2N3M2N3M2N2M4K4L4L5K4L5K4L4L4L3N4K8G5G9I7J6J7H7JPf_2"}, "label": "carrots in the nearest basket"}]}
{"id": 365015, "image": "COCO_train2014_000000365015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large display of carrots in front of another tray\"."}], "task": "refering", "answer_template": "The \"a large display of carrots in front of another tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 189, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.0, 0.68764, 0.5633066666666667, 0.98652], "iscrowd": 0, "area": 28254.182649999995, "rle": {"size": [500, 375], "counts": "X;`0S?o0QOo0QOo0ROk0TO1O1O10000000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O1000000000000O1000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000001O1O001O1O001O1O2N1O1O2N1O2N1O1O2N1O1O2L3N3M2N2N3M2N3M2N2M4K4L4L5K4L5K4L4L4L3N4K8G5G9I7J6J7H7JPf_2"}, "label": "a large display of carrots in front of another tray"}]}
{"id": 307671, "image": "COCO_train2014_000000307671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man appears to be nailed to a couple of skateboards that are lying on a street\"."}], "task": "refering", "answer_template": "The \"a man appears to be nailed to a couple of skateboards that are lying on a street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 262, 263, 264, 285, 286, 287, 308, 309, 310, 332, 333], "bbox": [0.18696875, 0.22485915492957748, 0.6895625000000001, 0.9754225352112676], "iscrowd": 0, "area": 28049.251449999978, "rle": {"size": [426, 640], "counts": "\\Rb11Y=0O10O0100O1jBOS=1mBOR=2nBNR=6N10KPC0o<6II[C7b<L]C5c<K]C5c<K\\C6c<K]C5c<K\\C5e<K[C5e<80000000000000000000000001O0000001O0000001O00001OM300O100000O10O10000000000O010000000O1000O100000O100000O0100O10000O100O01000O10000O10O010000O100O10000O010O10000O10000O1000O010000000000O1000O100000O10000000000O011O001YEQOQ9o0lFUOR9l0kFYOS9h0jF[OU9e0hF@U9b0gF5e8KXG>a8D\\Gd0^8\\O_Gl0\\8UO`GT1Y8nNdGZ1V8o1N1N3N2N2M3VIjL\\3c6O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2O0O2N1O00000J7N1O1N2O2N1N2O1000001O00000O100O101OO2O00001O001O0O2O1O1bNhLhHY3R7TMfHm2U7WMkHm2m6WMRIn2f6WMYIn2^6VMaIl2Z6XMeIi2W6[MiIe2R6aMlI`2n5fMRJ[2h5jMWJW2c5PN\\JP2^5WN_Jk1[5\\NcJe1W5bNfJ`1T5gNjJ[1n4mNPKT1f4XOVKj0i38ULIi3m4M3M4L3M3M3M3M3nMhGPO[8o0P20000000000O2O0000000000001O0000000O101O0000000000001O000O100000001O0000000000000O2O01O0000000001O0000000000001O01O0000000001O0000000000001O0001O0000000001O0000000000001O01O0000000001O0000odb2"}, "label": "a man appears to be nailed to a couple of skateboards that are lying on a street"}]}
{"id": 307671, "image": "COCO_train2014_000000307671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with blue jean and brown t - shirt lying on the skateboard\"."}], "task": "refering", "answer_template": "The \"a man with blue jean and brown t - shirt lying on the skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 262, 263, 264, 285, 286, 287, 308, 309, 310, 332, 333], "bbox": [0.18696875, 0.22485915492957748, 0.6895625000000001, 0.9754225352112676], "iscrowd": 0, "area": 28049.251449999978, "rle": {"size": [426, 640], "counts": "\\Rb11Y=0O10O0100O1jBOS=1mBOR=2nBNR=6N10KPC0o<6II[C7b<L]C5c<K]C5c<K\\C6c<K]C5c<K\\C5e<K[C5e<80000000000000000000000001O0000001O0000001O00001OM300O100000O10O10000000000O010000000O1000O100000O100000O0100O10000O100O01000O10000O10O010000O100O10000O010O10000O10000O1000O010000000000O1000O100000O10000000000O011O001YEQOQ9o0lFUOR9l0kFYOS9h0jF[OU9e0hF@U9b0gF5e8KXG>a8D\\Gd0^8\\O_Gl0\\8UO`GT1Y8nNdGZ1V8o1N1N3N2N2M3VIjL\\3c6O2N1O2N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2O0O2N1O00000J7N1O1N2O2N1N2O1000001O00000O100O101OO2O00001O001O0O2O1O1bNhLhHY3R7TMfHm2U7WMkHm2m6WMRIn2f6WMYIn2^6VMaIl2Z6XMeIi2W6[MiIe2R6aMlI`2n5fMRJ[2h5jMWJW2c5PN\\JP2^5WN_Jk1[5\\NcJe1W5bNfJ`1T5gNjJ[1n4mNPKT1f4XOVKj0i38ULIi3m4M3M4L3M3M3M3M3nMhGPO[8o0P20000000000O2O0000000000001O0000000O101O0000000000001O000O100000001O0000000000000O2O01O0000000001O0000000000001O01O0000000001O0000000000001O0001O0000000001O0000000000001O01O0000000001O0000odb2"}, "label": "a man with blue jean and brown t - shirt lying on the skateboard"}]}
{"id": 307671, "image": "COCO_train2014_000000307671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black shirt with hammering nail\"."}], "task": "refering", "answer_template": "The \"man in black shirt with hammering nail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 40, 41, 42, 43, 64, 65, 66, 67, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 272, 273, 293], "bbox": [0.67909375, 0.002816901408450704, 0.97525, 0.8028169014084507], "iscrowd": 0, "area": 37759.034499999994, "rle": {"size": [426, 640], "counts": "SVe53V=3N3L4M3L4M3L4M2YMB]Ha0`7@`Hb0]7@aHb0^7^ObHd0[7^OcHe0Z7]OeHd0Y7]OfHf0X7[OgHg0V7ZOiHj0S7WOmHk0Q7UOnHo0n6RORIP1k6QOTIS1h6nNXIT1e6mNZIW1b6iN_IY1^6hNaI\\1[6eNeI]1Y6cNgI`1U6`NkId1Q6]NoIe1n5\\NQJh1k5YNUJj1g5VNYJm1d5TN[JP2a5QN_JP2^5RNaJP2[5RNeJP2X5QNhJP2U5SNjJo1R5SNoJn1QOdLR5b1lKk1QOdLP5c1oKk1oNcLQ5b1PLm1nNaLQ5c1PLo1lN_LT5a1PLS2iN]LV5a1QLT2fN\\LY5_1QLX2dNYL[5^1QL\\2aNWL]5^1QL^2_NUL`5\\1QLk2o3UMQLk2n3UMRLl2n3SMRLn2n3RMQLo2n3QMRLP3n3oLRLR3n3nLRLR3m3nLSLS3m3lLSLU3m3kLRLV3T2`KlMZ10W3Q2jKgMn09X3l1VLcMb0`0Y3j1[LbM<d0Z3h1]LbM<c0X3i1_LcM=?U3m1`LbM?=R3o1bLbMa0;m2R2dLaMd08i2V2eL`Mf06f2Y2eL`Mj02b2^2eL^Mm00_2a2fL]MP1NZ2d2hL\\MR1LW2g2iL[MU1IS2l2iLYM[6f2gIXMZ6g2hIWMY6i2hIUMY6j2hIUMY6j2T20000O1000O10O1000000O1000000O1000000O1000000O12N5K5K3M2N2N2N2N3M2N2N2N1O2L4L3M4M3L3M4L3eN\\1^Ob0F9O2M3N1N3N2M2N100O100O100O100O100O10O01O10O0100O10O0100O010O1O10O0100O1N3M2M3N3M2M4M2M4K4M3M6K5K5L4K5L3L5K5L4K5mMUF`0Q:oNcFg0X;H9F9H9F9H[_6"}, "label": "man in black shirt with hammering nail"}]}
{"id": 307671, "image": "COCO_train2014_000000307671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black hoodie pounding something into a hand\"."}], "task": "refering", "answer_template": "The \"a man with a black hoodie pounding something into a hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 40, 41, 42, 43, 64, 65, 66, 67, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 272, 273, 293], "bbox": [0.67909375, 0.002816901408450704, 0.97525, 0.8028169014084507], "iscrowd": 0, "area": 37759.034499999994, "rle": {"size": [426, 640], "counts": "SVe53V=3N3L4M3L4M3L4M2YMB]Ha0`7@`Hb0]7@aHb0^7^ObHd0[7^OcHe0Z7]OeHd0Y7]OfHf0X7[OgHg0V7ZOiHj0S7WOmHk0Q7UOnHo0n6RORIP1k6QOTIS1h6nNXIT1e6mNZIW1b6iN_IY1^6hNaI\\1[6eNeI]1Y6cNgI`1U6`NkId1Q6]NoIe1n5\\NQJh1k5YNUJj1g5VNYJm1d5TN[JP2a5QN_JP2^5RNaJP2[5RNeJP2X5QNhJP2U5SNjJo1R5SNoJn1QOdLR5b1lKk1QOdLP5c1oKk1oNcLQ5b1PLm1nNaLQ5c1PLo1lN_LT5a1PLS2iN]LV5a1QLT2fN\\LY5_1QLX2dNYL[5^1QL\\2aNWL]5^1QL^2_NUL`5\\1QLk2o3UMQLk2n3UMRLl2n3SMRLn2n3RMQLo2n3QMRLP3n3oLRLR3n3nLRLR3m3nLSLS3m3lLSLU3m3kLRLV3T2`KlMZ10W3Q2jKgMn09X3l1VLcMb0`0Y3j1[LbM<d0Z3h1]LbM<c0X3i1_LcM=?U3m1`LbM?=R3o1bLbMa0;m2R2dLaMd08i2V2eL`Mf06f2Y2eL`Mj02b2^2eL^Mm00_2a2fL]MP1NZ2d2hL\\MR1LW2g2iL[MU1IS2l2iLYM[6f2gIXMZ6g2hIWMY6i2hIUMY6j2hIUMY6j2T20000O1000O10O1000000O1000000O1000000O1000000O12N5K5K3M2N2N2N2N3M2N2N2N1O2L4L3M4M3L3M4L3eN\\1^Ob0F9O2M3N1N3N2M2N100O100O100O100O100O10O01O10O0100O10O0100O010O1O10O0100O1N3M2M3N3M2M4M2M4K4M3M6K5K5L4K5L3L5K5L4K5mMUF`0Q:oNcFg0X;H9F9H9F9H[_6"}, "label": "a man with a black hoodie pounding something into a hand"}]}
{"id": 479707, "image": "COCO_train2014_000000479707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white tableclothes table that all the people are sitting at\"."}], "task": "refering", "answer_template": "The \"the white tableclothes table that all the people are sitting at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 194, 195, 196, 197, 198, 199, 213, 214, 215, 217, 218, 219, 220, 221, 222, 232, 233, 237, 238, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 269, 289, 290, 291, 313, 314], "bbox": [0.0975625, 0.5569392523364486, 0.71753125, 0.8910514018691589], "iscrowd": 0, "area": 15138.222000000003, "rle": {"size": [428, 640], "counts": "Pdj01[=1O0O2O001O001O0O2O001O001O0O2O001O001OM2O2M3N2MlbU14S]jN1O1O1N3N1O1O1N3N1O1O1N2O2N1N2O1O1O2M2O001O1N2O001N2O001O1N101O1O1N101O1O0O2O1OO010O100O010O100O010O10O0100O10O07J7H7J7H7J7H7JQ]b01jb]O4L5L3L5K4M3E<_O`0@a0G82O0O2N2O0O2O0O1O01O001O010O000010O01O000001O000000001O000000N2I8H71O001O00001O001O00001O0HYElMf:S2\\EmMd:R2^EmMa:S2aElM_:R2cEnM\\:R2fEmMY:S2hEmMX:R2jEmMU:R2nEmMR:R2oEnMP:R2RFmMm9S2TFmMl9R2g00001O001O01O01O001O00001O001=B3M001O001O1O001O1O001O001O1O001O1O001O`0@8H001O1O001O0000001O000O10001O0000001O0000001O0000001OgMdFb0\\9]OnF:R9EmG]OS8c0mG^OR8a0PH^OP8a0RH^On7a0TH^Ol7a0UH@j7?XH@h7?ZH@f7?\\H@d7`0\\HAc7`0]H_Oc7b0]H]Oc7d0]H[Oc7f0\\H[Oc7f0]HYOc75eFEh15c74hFGe14b73mFI`14c71oFL^11d71QGOZ1Nj7LPG6V1LQ8FlF?Q1Jo87PGGQ9;nFCT9=lFAV9`0hF_OZ9b0eF\\O]9e0bFYO_9h0aFVOa9k0`11O1OO0N3L4L4M3L4M2M4M^[[2"}, "label": "the white tableclothes table that all the people are sitting at"}]}
{"id": 479707, "image": "COCO_train2014_000000479707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"restaurant table with white table cloth and customers\"."}], "task": "refering", "answer_template": "The \"restaurant table with white table cloth and customers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 194, 195, 196, 197, 198, 199, 213, 214, 215, 217, 218, 219, 220, 221, 222, 232, 233, 237, 238, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 269, 289, 290, 291, 313, 314], "bbox": [0.0975625, 0.5569392523364486, 0.71753125, 0.8910514018691589], "iscrowd": 0, "area": 15138.222000000003, "rle": {"size": [428, 640], "counts": "Pdj01[=1O0O2O001O001O0O2O001O001O0O2O001O001OM2O2M3N2MlbU14S]jN1O1O1N3N1O1O1N3N1O1O1N2O2N1N2O1O1O2M2O001O1N2O001N2O001O1N101O1O1N101O1O0O2O1OO010O100O010O100O010O10O0100O10O07J7H7J7H7J7H7JQ]b01jb]O4L5L3L5K4M3E<_O`0@a0G82O0O2N2O0O2O0O1O01O001O010O000010O01O000001O000000001O000000N2I8H71O001O00001O001O00001O0HYElMf:S2\\EmMd:R2^EmMa:S2aElM_:R2cEnM\\:R2fEmMY:S2hEmMX:R2jEmMU:R2nEmMR:R2oEnMP:R2RFmMm9S2TFmMl9R2g00001O001O01O01O001O00001O001=B3M001O001O1O001O1O001O001O1O001O1O001O`0@8H001O1O001O0000001O000O10001O0000001O0000001O0000001OgMdFb0\\9]OnF:R9EmG]OS8c0mG^OR8a0PH^OP8a0RH^On7a0TH^Ol7a0UH@j7?XH@h7?ZH@f7?\\H@d7`0\\HAc7`0]H_Oc7b0]H]Oc7d0]H[Oc7f0\\H[Oc7f0]HYOc75eFEh15c74hFGe14b73mFI`14c71oFL^11d71QGOZ1Nj7LPG6V1LQ8FlF?Q1Jo87PGGQ9;nFCT9=lFAV9`0hF_OZ9b0eF\\O]9e0bFYO_9h0aFVOa9k0`11O1OO0N3L4L4M3L4M2M4M^[[2"}, "label": "restaurant table with white table cloth and customers"}]}
{"id": 479707, "image": "COCO_train2014_000000479707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back table parallel to the mirrored wall\"."}], "task": "refering", "answer_template": "The \"the back table parallel to the mirrored wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 150, 152, 170, 173, 174, 175, 176, 177, 180, 195, 196, 198, 199], "bbox": [0.3695, 0.44331775700934584, 0.863984375, 0.5835046728971963], "iscrowd": 0, "area": 3803.6488, "rle": {"size": [428, 640], "counts": "PXS32Y=1O1N2O1000000000001O4L4M2M4L4M2M0000000000000001O0000000O1E;00O101O000O10000O10008H1NY`d09W_[OG\\C:d<HZC8P<2UDHI6P<;PDLm;l0I6O1000001O00000000M3M3M3M3M3L4M30000O103M1O1N2O001N10O10BcCMQ=Lob9OT]F1O1O1TCOZ<1dC0]<`0N1O2O0O20O010O01O010O001O001O001O001O001iCVOj;j0oC^OP<P1O001O5KO1N2O100000001O0000001N13MN2N2O1N2J6@`0M3N2O100O100O2O0O100O100O1O010O10O1000001O0O12N2Mo]Q10PbnN9I1O0000O3N4L3MTUT1"}, "label": "the back table parallel to the mirrored wall"}]}
{"id": 479707, "image": "COCO_train2014_000000479707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table covered in a white table cloth with boxes on it\"."}], "task": "refering", "answer_template": "The \"a table covered in a white table cloth with boxes on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 150, 152, 170, 173, 174, 175, 176, 177, 180, 195, 196, 198, 199], "bbox": [0.3695, 0.44331775700934584, 0.863984375, 0.5835046728971963], "iscrowd": 0, "area": 3803.6488, "rle": {"size": [428, 640], "counts": "PXS32Y=1O1N2O1000000000001O4L4M2M4L4M2M0000000000000001O0000000O1E;00O101O000O10000O10008H1NY`d09W_[OG\\C:d<HZC8P<2UDHI6P<;PDLm;l0I6O1000001O00000000M3M3M3M3M3L4M30000O103M1O1N2O001N10O10BcCMQ=Lob9OT]F1O1O1TCOZ<1dC0]<`0N1O2O0O20O010O01O010O001O001O001O001O001iCVOj;j0oC^OP<P1O001O5KO1N2O100000001O0000001N13MN2N2O1N2J6@`0M3N2O100O100O2O0O100O100O1O010O10O1000001O0O12N2Mo]Q10PbnN9I1O0000O3N4L3MTUT1"}, "label": "a table covered in a white table cloth with boxes on it"}]}
{"id": 201184, "image": "COCO_train2014_000000201184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one women wearing black dress is holding a glass\"."}], "task": "refering", "answer_template": "The \"there is one women wearing black dress is holding a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 75, 76, 77, 98, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 169, 170, 192, 193], "bbox": [0.249734375, 0.12947916666666667, 0.43565624999999997, 0.529375], "iscrowd": 0, "area": 11737.253399999998, "rle": {"size": [480, 640], "counts": "eU[26e>8I7J5L5L3L2O1O0O2O1O1N2O1N2I7I6M4M3L4M2N3L4O001O1O001O1N2M2O2N2M2O2N2N2N1N3N2N1O2M3N2N1O2M3N2N2N2O1O1O010000000000010O0000000]F\\Lj7d3TH`Lk7_3SHeLk7\\3SHgLk7\\3bGQLFf0g8\\3]GSLIb0j8^3WGTLL`0m8`3QGTLO=Q9P50mMnFXNR9h1nFXNS9f1nFZNT9d1lF\\NU9b1lF^NV9`1jF`NX9]1hFdNZ9Z1fFfN]9V1dFjN^9T1bFlN`9Q1aFoNb9n0^FROf9i0[FWOh9e0YF[Ok9a0UF_OW:4jELY:1gEO\\:MeE3^:JbE6a:F`E:c:C]E=f:@ZE`0Z;lNfDT1g;_NYDa1j;\\NVDd1i;]NXDb1g;_NYDa1f;_N[Da1e;_N\\D`1d;`N\\D`1d;`N]D_1c;aN]D_1c;aN_D\\1b;dN`DZ1`;fNbDW1^;iNeDT1\\;lNiDo0W;QOmDj0T;VOgEOY:1iELX:4iEIZ:6hEGY:9hEEZ::W2O5J5JhQY5"}, "label": "there is one women wearing black dress is holding a glass"}]}
{"id": 201184, "image": "COCO_train2014_000000201184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with dark shirt and white pants sitting on a couch\"."}], "task": "refering", "answer_template": "The \"a woman with dark shirt and white pants sitting on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 75, 76, 77, 98, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 169, 170, 192, 193], "bbox": [0.249734375, 0.12947916666666667, 0.43565624999999997, 0.529375], "iscrowd": 0, "area": 11737.253399999998, "rle": {"size": [480, 640], "counts": "eU[26e>8I7J5L5L3L2O1O0O2O1O1N2O1N2I7I6M4M3L4M2N3L4O001O1O001O1N2M2O2N2M2O2N2N2N1N3N2N1O2M3N2N1O2M3N2N2N2O1O1O010000000000010O0000000]F\\Lj7d3TH`Lk7_3SHeLk7\\3SHgLk7\\3bGQLFf0g8\\3]GSLIb0j8^3WGTLL`0m8`3QGTLO=Q9P50mMnFXNR9h1nFXNS9f1nFZNT9d1lF\\NU9b1lF^NV9`1jF`NX9]1hFdNZ9Z1fFfN]9V1dFjN^9T1bFlN`9Q1aFoNb9n0^FROf9i0[FWOh9e0YF[Ok9a0UF_OW:4jELY:1gEO\\:MeE3^:JbE6a:F`E:c:C]E=f:@ZE`0Z;lNfDT1g;_NYDa1j;\\NVDd1i;]NXDb1g;_NYDa1f;_N[Da1e;_N\\D`1d;`N\\D`1d;`N]D_1c;aN]D_1c;aN_D\\1b;dN`DZ1`;fNbDW1^;iNeDT1\\;lNiDo0W;QOmDj0T;VOgEOY:1iELX:4iEIZ:6hEGY:9hEEZ::W2O5J5JhQY5"}, "label": "a woman with dark shirt and white pants sitting on a couch"}]}
{"id": 201184, "image": "COCO_train2014_000000201184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady sitting on the floor handling a glass with smile\"."}], "task": "refering", "answer_template": "The \"a lady sitting on the floor handling a glass with smile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364], "bbox": [0.62021875, 0.3640416666666667, 0.886515625, 0.9258333333333334], "iscrowd": 0, "area": 28700.747400000007, "rle": {"size": [480, 640], "counts": "h_j5>`>2N2O2M2N2N3N1N2L4K6K4K5J6J7K4L4M4L3O1O1O2N1O1kL_MjIb2S6aMlIa2P6bMoI_2o5cMnI`2o5cMlIc2Q6_MiIg2V6ZMeIk2Z6WM_Io2`6RMZIU3e6kLVIZ3k6eLPI`3P7`LPIa3P7^LPIb3Q7]LnHd3R7^LlHb3U7`LhHa3W7bLfH^3[7dLbH\\3_7fL^HZ3b7iLZHY3f7iLWHW3i7lLTHT3m7nLPHR3Q8PMlGP3T8SMiGm2X8UMdGl2\\8WMaGj2`8VM^Gj2d8UM[Gk2f8UMYGk2i8TMVGl2l8SMRGn2P9QMoFo2R9QMmFo2U9PMjFQ3W9oLgFQ3[9Y11O2N2N2N1O2N2O1N1O2N2N1OO1O1O2N1O1O1VNkJhIV5V6nJgIS5W6PKgIQ5W6SKgIm4W6WKfIk4W6XKgIi4X6ZKeIg4Y6]KdId4Z6`KcIa4[6bKcI_4[6eKbI\\4\\6hKaIY4]6jKaIX4\\6lKaIU4]6oK`IR4_6PL_IQ4_6SL^In3`6VL^Ij3`6ZL]Ig3a6\\L]Ie3b6^L[Id3d6\\20000001O000001O0001O0000000001O01O000000000010O0000000001O00000O1OT1lNa0@2M3M4M2M3M3N2M3M3M3N3L3M3N2M3M4M5J6J6J6K4K6J6K5J5K6K5J;E>B=D=B>B=D=B\\jQ1"}, "label": "a lady sitting on the floor handling a glass with smile"}]}
{"id": 201184, "image": "COCO_train2014_000000201184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364], "bbox": [0.62021875, 0.3640416666666667, 0.886515625, 0.9258333333333334], "iscrowd": 0, "area": 28700.747400000007, "rle": {"size": [480, 640], "counts": "h_j5>`>2N2O2M2N2N3N1N2L4K6K4K5J6J7K4L4M4L3O1O1O2N1O1kL_MjIb2S6aMlIa2P6bMoI_2o5cMnI`2o5cMlIc2Q6_MiIg2V6ZMeIk2Z6WM_Io2`6RMZIU3e6kLVIZ3k6eLPI`3P7`LPIa3P7^LPIb3Q7]LnHd3R7^LlHb3U7`LhHa3W7bLfH^3[7dLbH\\3_7fL^HZ3b7iLZHY3f7iLWHW3i7lLTHT3m7nLPHR3Q8PMlGP3T8SMiGm2X8UMdGl2\\8WMaGj2`8VM^Gj2d8UM[Gk2f8UMYGk2i8TMVGl2l8SMRGn2P9QMoFo2R9QMmFo2U9PMjFQ3W9oLgFQ3[9Y11O2N2N2N1O2N2O1N1O2N2N1OO1O1O2N1O1O1VNkJhIV5V6nJgIS5W6PKgIQ5W6SKgIm4W6WKfIk4W6XKgIi4X6ZKeIg4Y6]KdId4Z6`KcIa4[6bKcI_4[6eKbI\\4\\6hKaIY4]6jKaIX4\\6lKaIU4]6oK`IR4_6PL_IQ4_6SL^In3`6VL^Ij3`6ZL]Ig3a6\\L]Ie3b6^L[Id3d6\\20000001O000001O0001O0000000001O01O000000000010O0000000001O00000O1OT1lNa0@2M3M4M2M3M3N2M3M3M3N3L3M3N2M3M4M5J6J6J6K4K6J6K5J5K6K5J;E>B=D=B>B=D=B\\jQ1"}, "label": "a woman wearing a white shirt"}]}
{"id": 201184, "image": "COCO_train2014_000000201184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a dress\"."}], "task": "refering", "answer_template": "The \"a woman in a dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 72, 73, 95, 96, 117, 118, 119, 120, 140, 141, 142, 143, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 280, 281, 282, 304], "bbox": [0.093953125, 0.1255, 0.35906249999999995, 0.7697916666666667], "iscrowd": 0, "area": 25247.207699999995, "rle": {"size": [480, 640], "counts": "`Zl0Y1d=5K4K5L4L4K5L4J6J6J6I7J6J6J7J4N3N1O2N1N2O2N1YN_LnGc3o7fLhG\\3W8fLgG\\3W8fLgG[3X8gLeG[3Y8hLeGZ3Y8hLeGY3Z8hLeGZ3Y8hLdGZ3Z8iLdGX3[8jLcGX3[8jLcGW3[8lLbGW3\\8g1O1O2N101O00001O001O00001O001N10001O0010000O100O2iIPI[4P7aKTI^4m6]KXIc4h6YK\\If4e6VK_Ij4a6QKdIn4]6oJfIR5f0SJn3h0^KW5`0UJP4a0cK\\59UJT4<fKa52TJY49gKe5LTJ]44jKi5EUJb40kKm5_OUJf4KmKS6YOTJk4EoKY6ROSJP5BPL\\6mNTJS5]OSLa6fNUJV5XOVLe6_NXJQ7m6O1O1O1N3N1N2N2N2N3M2N2O1N3M2N3M3M2iNRGXLQ9d3QG[LQ9c3PG]LR9_3oFaLT9[3nFcLU9Z3mFeLV9W3kFiLX9S3jFlLX9Q3jFnLY9n2hFRMX9n2iFQMW9n2kFQMU9n2lFRMU9l2mFSMS9l2oFSMQ9m2oFSMQ9l2QGSMo8l2SGSMm8l2TGTMl8k2VGTMj8k2XGSMi8m2XGkLn8U3SGhLP9W3RGeLQ9Z3QGcLQ9\\3Z100O1O1O1O100N2O0O2N2O1N1O2O1N1O2O1N2N10O01O010eNkCKV<7PD@R<a0RDWOQ<i0PDSOT<m0mCnNV<R1kCjNX<V1jCeNY<[1f0000000000001O0O100XO_BOc=1`BJb=5aBGa=8bBDd=7_BFe=6^BFh=5f0Lboo5"}, "label": "a woman in a dress"}]}
{"id": 201184, "image": "COCO_train2014_000000201184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a dress\"."}], "task": "refering", "answer_template": "The \"a woman wearing a dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 72, 73, 95, 96, 117, 118, 119, 120, 140, 141, 142, 143, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 280, 281, 282, 304], "bbox": [0.093953125, 0.1255, 0.35906249999999995, 0.7697916666666667], "iscrowd": 0, "area": 25247.207699999995, "rle": {"size": [480, 640], "counts": "`Zl0Y1d=5K4K5L4L4K5L4J6J6J6I7J6J6J7J4N3N1O2N1N2O2N1YN_LnGc3o7fLhG\\3W8fLgG\\3W8fLgG[3X8gLeG[3Y8hLeGZ3Y8hLeGY3Z8hLeGZ3Y8hLdGZ3Z8iLdGX3[8jLcGX3[8jLcGW3[8lLbGW3\\8g1O1O2N101O00001O001O00001O001N10001O0010000O100O2iIPI[4P7aKTI^4m6]KXIc4h6YK\\If4e6VK_Ij4a6QKdIn4]6oJfIR5f0SJn3h0^KW5`0UJP4a0cK\\59UJT4<fKa52TJY49gKe5LTJ]44jKi5EUJb40kKm5_OUJf4KmKS6YOTJk4EoKY6ROSJP5BPL\\6mNTJS5]OSLa6fNUJV5XOVLe6_NXJQ7m6O1O1O1N3N1N2N2N2N3M2N2O1N3M2N3M3M2iNRGXLQ9d3QG[LQ9c3PG]LR9_3oFaLT9[3nFcLU9Z3mFeLV9W3kFiLX9S3jFlLX9Q3jFnLY9n2hFRMX9n2iFQMW9n2kFQMU9n2lFRMU9l2mFSMS9l2oFSMQ9m2oFSMQ9l2QGSMo8l2SGSMm8l2TGTMl8k2VGTMj8k2XGSMi8m2XGkLn8U3SGhLP9W3RGeLQ9Z3QGcLQ9\\3Z100O1O1O1O100N2O0O2N2O1N1O2O1N1O2O1N2N10O01O010eNkCKV<7PD@R<a0RDWOQ<i0PDSOT<m0mCnNV<R1kCjNX<V1jCeNY<[1f0000000000001O0O100XO_BOc=1`BJb=5aBGa=8bBDd=7_BFe=6^BFh=5f0Lboo5"}, "label": "a woman wearing a dress"}]}
{"id": 242145, "image": "COCO_train2014_000000242145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow suitcase\"."}], "task": "refering", "answer_template": "The \"a yellow suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 241, 242, 258, 259, 260, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 331, 332], "bbox": [0.21935937499999997, 0.5755208333333334, 0.544734375, 0.8628541666666666], "iscrowd": 0, "area": 13508.537950000002, "rle": {"size": [480, 640], "counts": "P]R27h>6J7J5J7I6J7I1O2O0O2N1O1O2N100O2N1O1O2N101N1O1O2N1O101N1O1O2N1O2O0O1O2N1O1O2O0O1O2N1O2N100O2N1O1O2N100O21N8I7H8H8I2M2N100O1O100O100O2N100O1O100O1O10O01O10O0100O1O010O1O010O1O10O0100O0001N10001O00001O001N10001O00001O000O2O001O00001O000O2O00001O001O000O2O00001O00001O0O2O00001O00001O000O2O001O00001O0O1O2N1N2O2N1O00O01000O100O10O10O100O10O10O01O0O2N2N1I_MQDb2Q<42O1O2M2O1N3N1O1N3N1N2O2N1N2O2M2O1N3N1O1N3N1N3N7I7H8I7H8I7IbaX4"}, "label": "a yellow suitcase"}]}
{"id": 242145, "image": "COCO_train2014_000000242145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow suitcase on the ground\"."}], "task": "refering", "answer_template": "The \"the yellow suitcase on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 241, 242, 258, 259, 260, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 331, 332], "bbox": [0.21935937499999997, 0.5755208333333334, 0.544734375, 0.8628541666666666], "iscrowd": 0, "area": 13508.537950000002, "rle": {"size": [480, 640], "counts": "P]R27h>6J7J5J7I6J7I1O2O0O2N1O1O2N100O2N1O1O2N101N1O1O2N1O101N1O1O2N1O2O0O1O2N1O1O2O0O1O2N1O2N100O2N1O1O2N100O21N8I7H8H8I2M2N100O1O100O100O2N100O1O100O1O10O01O10O0100O1O010O1O010O1O10O0100O0001N10001O00001O001N10001O00001O000O2O001O00001O000O2O00001O001O000O2O00001O00001O0O2O00001O00001O000O2O001O00001O0O1O2N1N2O2N1O00O01000O100O10O10O100O10O10O01O0O2N2N1I_MQDb2Q<42O1O2M2O1N3N1O1N3N1N2O2N1N2O2M2O1N3N1O1N3N1N3N7I7H8I7H8I7IbaX4"}, "label": "the yellow suitcase on the ground"}]}
{"id": 242145, "image": "COCO_train2014_000000242145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown suitcase in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the brown suitcase in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 265, 266, 267, 288, 289, 290], "bbox": [0.510671875, 0.3910208333333333, 0.664046875, 0.7640416666666667], "iscrowd": 0, "area": 12994.15405, "rle": {"size": [480, 640], "counts": "Z_i4:_>8O1dCHn9:YE>f:CbDT1];Y1O001O001O001N101O003M5K5J6K5K5K4L5K5J6K5K8H4L1O001N2O1O1O1O000O0100000000000O10O100000001O2N1O1N3N1O2N1O2N1O1O2M2O2N1O2N1O2N1O1N3N1O2N1O2N1O2N1N2O2N1O2N1O2N1O1N3N1L5E:E<E:E<E:E<E:E<E:EYaT3"}, "label": "the brown suitcase in the right hand picture"}]}
{"id": 242145, "image": "COCO_train2014_000000242145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown suitcase in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the brown suitcase in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 265, 266, 267, 288, 289, 290], "bbox": [0.510671875, 0.3910208333333333, 0.664046875, 0.7640416666666667], "iscrowd": 0, "area": 12994.15405, "rle": {"size": [480, 640], "counts": "Z_i4:_>8O1dCHn9:YE>f:CbDT1];Y1O001O001O001N101O003M5K5J6K5K5K4L5K5J6K5K8H4L1O001N2O1O1O1O000O0100000000000O10O100000001O2N1O1N3N1O2N1O2N1O1O2M2O2N1O2N1O2N1O1N3N1O2N1O2N1O2N1N2O2N1O2N1O2N1O1N3N1L5E:E<E:E<E:E<E:E<E:EYaT3"}, "label": "the brown suitcase in the right hand picture"}]}
{"id": 496097, "image": "COCO_train2014_000000496097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair on the right of the two\"."}], "task": "refering", "answer_template": "The \"chair on the right of the two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 236, 237, 238, 259, 260, 261, 262, 263, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333], "bbox": [0.26990625, 0.6387058823529411, 0.485546875, 0.9892470588235293], "iscrowd": 0, "area": 12428.076649999999, "rle": {"size": [425, 640], "counts": "URX21X=3M2N3M2N3N1N3M3M2N3N1N4L4L3M4M3L3M4L4L3N3L4L4`FZNV7h1VHiNj7[1aGVO`8i21O1O001O1O1O001O1O1O001O1O1O001O1O1O010O1O1O1O3M3M3M3M2N3M3M3M3M3M3M3M3MN2M3L4M3M3L4M3N2001O001O001O010O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O0010O01O001O001O001O001O001O001O001O001O001O1O1O1O1O2N1O1O1OeVX4"}, "label": "chair on the right of the two"}]}
{"id": 496097, "image": "COCO_train2014_000000496097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan - and - brown patterned chair sits empty at a kitchen island\"."}], "task": "refering", "answer_template": "The \"a tan - and - brown patterned chair sits empty at a kitchen island\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 236, 237, 238, 259, 260, 261, 262, 263, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333], "bbox": [0.26990625, 0.6387058823529411, 0.485546875, 0.9892470588235293], "iscrowd": 0, "area": 12428.076649999999, "rle": {"size": [425, 640], "counts": "URX21X=3M2N3M2N3N1N3M3M2N3N1N4L4L3M4M3L3M4L4L3N3L4L4`FZNV7h1VHiNj7[1aGVO`8i21O1O001O1O1O001O1O1O001O1O1O001O1O1O010O1O1O1O3M3M3M3M2N3M3M3M3M3M3M3M3MN2M3L4M3M3L4M3N2001O001O001O010O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O0010O01O001O001O001O001O001O001O001O001O001O1O1O1O1O2N1O1O1OeVX4"}, "label": "a tan - and - brown patterned chair sits empty at a kitchen island"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange cat getting slapped in the face\"."}], "task": "refering", "answer_template": "The \"orange cat getting slapped in the face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 35, 36, 37, 38, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 203, 204, 205, 206, 221, 222, 243, 244, 266, 267, 288, 289], "bbox": [0.47726562499999997, 0.03017632241813602, 1.0, 0.9094710327455919], "iscrowd": 0, "area": 53156.82185000001, "rle": {"size": [397, 640], "counts": "aif31\\<0O100O101N100O100O100O2O0O1O100mM>ZGCc8`0]G@_8d0`G]O]8f0cG[OY8h0fGYOW8j0iGVOT8m0kGTOV8l0iGTOX8k0gGVOW8l0iGTOR8Q1mGQOl7U1THkNg7Z1XHgNg7[1XHeNg7\\1XHeNg7\\1YHdNe7^1ZHcNe7^1[HcNc7^1\\HcNc7_1\\HaNc7`1\\HaNb7a1^H_Nb7a1]H`Nb7a1^H_Na7b1^H_Na7c1^H^N`7c1_H^N`7c1`H]N_7d1aH\\N^7e1aH\\N_7d1aH\\N^7f1XHcNg7S3N2VMeK_M\\4`2fK^M[4a2fK^M\\4_2gK_MZ4`2hK^MZ4`2hK_MX4_2jK`MX4^2jK`MW4_2kK_MW4_2jK`MW4^2lK`MU4_2mK`MT4^2nK`MS4_2nK`MT4]2oKbMQ4]2QLaMQ4]2PLbMQ4]2QLbMo3\\2TLbMn3\\2SLcMn3\\2TLbMn3[2ULdMk3[2WLcMk3[2ULeMl3Z2RLiMo3T2QLmMP4R2oKoMR4P2mKRNU4j1jKYNX4d1gK^N[4^1eKdN]4X1bKkN_4S1`KnNc4n0]KTOe4i0ZKYOh4c0WK@j4>UKDm48SKJo43oJOT5JoJ8S5ZOWKh0j4kN`KW1b4[NgKh1[4jMoKX2S4YMXLi2c62O2N2N2M4M3M3L3N3M3M3L3N3M3L4M2N3M3L4M2N3L4M2N3M3L4M2N3L4M3M2N3L5L5K5J6K5K5K4K4M2N3L4M3M2N3L4M3M3N101O1O1O001O1O1O010O1O0000000000O1000001O0O100000000O1000001O000O100000000O1000001O0O100000000O1000001O000O100000000O1000001O0O10000000000O1000001O0O100000000O100000001VIZKk5f4TJ[Kl5e4SJ]Kl5c4RJ`Km5`4RJbKm5^4RJdKm5\\4QJgKn5Y4QJhKP6W4oIkKP6U4oImKP6S4nIPLQ6Q4mIQLR6o3mISLR6m3lIVLS6j3mIVLS6j3mIWLR6i3mIYLR6g3nIYLS6f3mIZLS6f3mIZLS6f3lI[LT6f3kIZLU6f3kIZLU6f3jIZLW6f3iIZLW6f3iIZLW6f3iIZLW6f3hI[LX6e3hI[LY6d3gI[LZ6e3eI\\L[6d3eI\\L[6e3dI[L\\6e3dI[L\\6e3cI\\L]6d3cI[L^6e3bI[L^6e3bI[L^6e3aI\\L`6c3`I]L`6c3`I]L`6g410000000000O10001O00000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000O10000000000O10000N2^Oe0ZOe0\\Oe0ZOe0\\Oe0[Od0[Oe0\\OUN"}, "label": "orange cat getting slapped in the face"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange cat on top of the other orange cat\"."}], "task": "refering", "answer_template": "The \"the orange cat on top of the other orange cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 35, 36, 37, 38, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 203, 204, 205, 206, 221, 222, 243, 244, 266, 267, 288, 289], "bbox": [0.47726562499999997, 0.03017632241813602, 1.0, 0.9094710327455919], "iscrowd": 0, "area": 53156.82185000001, "rle": {"size": [397, 640], "counts": "aif31\\<0O100O101N100O100O100O2O0O1O100mM>ZGCc8`0]G@_8d0`G]O]8f0cG[OY8h0fGYOW8j0iGVOT8m0kGTOV8l0iGTOX8k0gGVOW8l0iGTOR8Q1mGQOl7U1THkNg7Z1XHgNg7[1XHeNg7\\1XHeNg7\\1YHdNe7^1ZHcNe7^1[HcNc7^1\\HcNc7_1\\HaNc7`1\\HaNb7a1^H_Nb7a1]H`Nb7a1^H_Na7b1^H_Na7c1^H^N`7c1_H^N`7c1`H]N_7d1aH\\N^7e1aH\\N_7d1aH\\N^7f1XHcNg7S3N2VMeK_M\\4`2fK^M[4a2fK^M\\4_2gK_MZ4`2hK^MZ4`2hK_MX4_2jK`MX4^2jK`MW4_2kK_MW4_2jK`MW4^2lK`MU4_2mK`MT4^2nK`MS4_2nK`MT4]2oKbMQ4]2QLaMQ4]2PLbMQ4]2QLbMo3\\2TLbMn3\\2SLcMn3\\2TLbMn3[2ULdMk3[2WLcMk3[2ULeMl3Z2RLiMo3T2QLmMP4R2oKoMR4P2mKRNU4j1jKYNX4d1gK^N[4^1eKdN]4X1bKkN_4S1`KnNc4n0]KTOe4i0ZKYOh4c0WK@j4>UKDm48SKJo43oJOT5JoJ8S5ZOWKh0j4kN`KW1b4[NgKh1[4jMoKX2S4YMXLi2c62O2N2N2M4M3M3L3N3M3M3L3N3M3L4M2N3M3L4M2N3L4M2N3M3L4M2N3L4M3M2N3L5L5K5J6K5K5K4K4M2N3L4M3M2N3L4M3M3N101O1O1O001O1O1O010O1O0000000000O1000001O0O100000000O1000001O000O100000000O1000001O0O100000000O1000001O000O100000000O1000001O0O10000000000O1000001O0O100000000O100000001VIZKk5f4TJ[Kl5e4SJ]Kl5c4RJ`Km5`4RJbKm5^4RJdKm5\\4QJgKn5Y4QJhKP6W4oIkKP6U4oImKP6S4nIPLQ6Q4mIQLR6o3mISLR6m3lIVLS6j3mIVLS6j3mIWLR6i3mIYLR6g3nIYLS6f3mIZLS6f3mIZLS6f3lI[LT6f3kIZLU6f3kIZLU6f3jIZLW6f3iIZLW6f3iIZLW6f3iIZLW6f3hI[LX6e3hI[LY6d3gI[LZ6e3eI\\L[6d3eI\\L[6e3dI[L\\6e3dI[L\\6e3cI\\L]6d3cI[L^6e3bI[L^6e3bI[L^6e3aI\\L`6c3`I]L`6c3`I]L`6g410000000000O10001O00000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000O10000000000O10000N2^Oe0ZOe0\\Oe0ZOe0\\Oe0[Od0[Oe0\\OUN"}, "label": "the orange cat on top of the other orange cat"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown chair in the front on which cats is playing\"."}], "task": "refering", "answer_template": "The \"the brown chair in the front on which cats is playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 197, 198, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.00134375, 0.0, 0.84490625, 0.9892191435768263], "iscrowd": 0, "area": 85738.23634999999, "rle": {"size": [397, 640], "counts": "^<`6m5O10000O10000O10000O10000O10000O1000000O10000O10000O1000000O10001N1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O100000001O000O10000000000O1000000000O10O1000000000000O10000000000O10000000ZO_J`Ja5]5cJbJ]5\\5eJdJZ5[5iJdJW5Y5mJfJS5Y5oJfJQ5Y5QKfJo4X5TKgJl4X5VKgJj4X5XKgJh4W5[KhJe4W5]KhJc4V5`KiJ`4V5bKiJ^4V5cKjJ]4T5fKkJ[4S5fKmJZ4Q5iKnJW4Q5jKoJV4P5kKPKU4n4mKRKS4m4nKSKR4m4nKSKR4m4nKSKS4k4nKUKR4k4nKUKR4k4nKUKR4k4nKUKR4k4nKUKR4j4oKVKQ4j4oKVKR4i4nKWKR4i4nKWKR4i4nKWKR4i4nKWKR4h4oKXKQ4h4oKXKQ4h4oKXKQ4h4oKXKR4g4nKYKR4f4oKZKQ4f4oKZKQ4f4oKZKQ4f4oKZKQ4f4oKZKQ4f4oKZKR4d4oK\\KQ4d4oK\\KQ4d4oK\\KQ4d4oK\\KQ4d4oK\\KQ4c4oK^KQ4b4oK^KQ4b4oK^KR4a4nK_KR4`4oK`KQ4`4oK`KQ4`4oK`KQ4`4oK`KQ4`4oK`KQ4_4PLaKQ4^4oKbKQ4^4oKbKQ4^4oKbKQ4]4PLcKP4]4PLcKP4]4PLcKP4]4PLcKX3U5hLkJW3U5jLkJU3V5kLjJT3W5lLiJR3Y5nLgJQ3Y5PMgJo2Z5QMfJn2[5RMeJm2\\5SMdJk2^5UMbJj2^5WMbJh2_5XMaJg2`5YM`Je2b5[M^Jd2b5]M^Jb2c5^M]Ja2d5_M\\J`2d5aM\\J]2f5cMZJ\\2g5dMYJ[2g5fMYJY2h5gMXJX2h5iMXJU2j5kMVJT2k5lMUJT2j5mMVJR2k5nMUJR2j5oMVJP2k5PNUJP2j5QNVJn1k5RNUJn1k5RNUJm1k5TNUJl1k5TNUJk1k5VNUJj1k5VNUJi1k5XNUJh1k5XNUJg1l5YNTJg1k5ZNUJe1l5[NTJe1k5\\NUJc1l5]NTJc1l5]NTJb1l5_NTJa1l5_NTJ`1l5aNTJ_1l5aNTJ^1l5cNTJ]1l5cNTJ\\1m5dNSJ\\1l5eNTJZ1m5fNSJZ1l5gNTJX1l5iNTJW1l5iNTJV1l5kNTJT1l5mNTJS1l5mNTJR1l5oNTJQ1k5POUJo0l5QOTJo0k5ROUJm0k5TOTJm0l5SOTJl0l5UOTJk0k5VOUJi0k5XOUJh0k5XOUJg0k5ZOUJf0j5[OVJd0k5\\OUJd0j5]OVJb0j5_OVJa0j5_OVJ`0j5AVJ?i5BWJ=j5CVJ=i5DWJ;i5FWJ:h5GXJ9h5GXJ8h5IXJ7g5JYJ5h5KXJ5g5LYJ3g5NYJ2f5OZJ0g50YJ0f51ZJOe52[JMf53ZJMe54[JKe55\\JKc56]JId57\\JIc58]JGc5:]JFb5;^JEb5;^JDb5=^JCa5>_JAa5`0_J@a5`0_J_Oa5b0_J^O`5c0`J\\Oa5d0_J\\O`5e0`J[O_5f0aJYO_5h0aJXO_5h0aJWO_5j0aJVO^5k0bJTO^5m0bJSO^5m0bJRO^5o0bJQO]5P1bJQO]5P1cJoN^5Q1bJoN]5R1cJmN]5T1cJlN^5S1bJlN`5S1`JmNb5Q1^JnNd5Q1\\JoNf5o0ZJQOg5n0YJQOj5m0VJSOk5l0UJSOn5k0RJUOo5j0QJUOR6i0nIWOR6i0nIVOR6k0nIUOP6m0PJSOo5n0QJQOn5Q1RJoNm5R1SJmNo5R1QJnNo5R1QJmNQ6R1oInNQ6R1oImNS6R1mInNS6R1mInNT6Q1lInNU6R1kInNV6Q1jInNW6R1iIgN^6Y1bI_Nf6a1ZIWNm6j1SImMU7T2kHdM\\7]2dH[Md7e2\\HSMk7n2UHjLR8W3nGaLY8`3:0O1O100O1O1O100O1O1O1O1O1O1O1O1O1O100O10000O10YNiGXOW8g0mGVOR8k0PHSOP8l0THQOl7n0WHPOh7Q1ZHmNf7R1^HkNa7U1bHiN^7W1cHhN]7W1dHiN[7W1fHiNZ7W1gHhNi6nNjHY2?iNg6oNhHV2d0kNd6oNgHU2g0lNb6POeHR2l0nN_6POdHP2Q1oNZ6SOcHl1V1QOW6SObHk1Y1ROU6SOaHi1]1TOR6TO_Hf1b1VOn5UO`Hb1e1YOk5n0XJQOh5l0[JTOe5i0^JWOa5h0aJXO_5f0cJZO]5d0eJ\\O[5b0hJ]OX5a0jJ_OU5`0mJ@S5>oJBQ5<QKDo4;RKEm4:UKFk48WKHi46YKJg44[KLd43]KNc40_K0a4NaK2_4LcK4\\4JgK6Y4EjK=V4^OlKe0T4VOnKm0R4nNPLU1o3gNTL[1\\73M3M3N2N1O2M3N2N2N2M3N2N2N2N2N2O1N2O1N2O1N2O1N2O1N2O1M3N2M3M3N2M3M3N2M3O11O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O100O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N:Fe0[OVPV1"}, "label": "the brown chair in the front on which cats is playing"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown chair in the front on which cats is playing\"."}], "task": "refering", "answer_template": "The \"the brown chair in the front on which cats is playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 197, 198, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.00134375, 0.0, 0.84490625, 0.9892191435768263], "iscrowd": 0, "area": 85738.23634999999, "rle": {"size": [397, 640], "counts": "^<`6m5O10000O10000O10000O10000O10000O1000000O10000O10000O1000000O10001N1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O100000001O000O10000000000O1000000000O10O1000000000000O10000000000O10000000ZO_J`Ja5]5cJbJ]5\\5eJdJZ5[5iJdJW5Y5mJfJS5Y5oJfJQ5Y5QKfJo4X5TKgJl4X5VKgJj4X5XKgJh4W5[KhJe4W5]KhJc4V5`KiJ`4V5bKiJ^4V5cKjJ]4T5fKkJ[4S5fKmJZ4Q5iKnJW4Q5jKoJV4P5kKPKU4n4mKRKS4m4nKSKR4m4nKSKR4m4nKSKS4k4nKUKR4k4nKUKR4k4nKUKR4k4nKUKR4k4nKUKR4j4oKVKQ4j4oKVKR4i4nKWKR4i4nKWKR4i4nKWKR4i4nKWKR4h4oKXKQ4h4oKXKQ4h4oKXKQ4h4oKXKR4g4nKYKR4f4oKZKQ4f4oKZKQ4f4oKZKQ4f4oKZKQ4f4oKZKQ4f4oKZKR4d4oK\\KQ4d4oK\\KQ4d4oK\\KQ4d4oK\\KQ4d4oK\\KQ4c4oK^KQ4b4oK^KQ4b4oK^KR4a4nK_KR4`4oK`KQ4`4oK`KQ4`4oK`KQ4`4oK`KQ4`4oK`KQ4_4PLaKQ4^4oKbKQ4^4oKbKQ4^4oKbKQ4]4PLcKP4]4PLcKP4]4PLcKP4]4PLcKX3U5hLkJW3U5jLkJU3V5kLjJT3W5lLiJR3Y5nLgJQ3Y5PMgJo2Z5QMfJn2[5RMeJm2\\5SMdJk2^5UMbJj2^5WMbJh2_5XMaJg2`5YM`Je2b5[M^Jd2b5]M^Jb2c5^M]Ja2d5_M\\J`2d5aM\\J]2f5cMZJ\\2g5dMYJ[2g5fMYJY2h5gMXJX2h5iMXJU2j5kMVJT2k5lMUJT2j5mMVJR2k5nMUJR2j5oMVJP2k5PNUJP2j5QNVJn1k5RNUJn1k5RNUJm1k5TNUJl1k5TNUJk1k5VNUJj1k5VNUJi1k5XNUJh1k5XNUJg1l5YNTJg1k5ZNUJe1l5[NTJe1k5\\NUJc1l5]NTJc1l5]NTJb1l5_NTJa1l5_NTJ`1l5aNTJ_1l5aNTJ^1l5cNTJ]1l5cNTJ\\1m5dNSJ\\1l5eNTJZ1m5fNSJZ1l5gNTJX1l5iNTJW1l5iNTJV1l5kNTJT1l5mNTJS1l5mNTJR1l5oNTJQ1k5POUJo0l5QOTJo0k5ROUJm0k5TOTJm0l5SOTJl0l5UOTJk0k5VOUJi0k5XOUJh0k5XOUJg0k5ZOUJf0j5[OVJd0k5\\OUJd0j5]OVJb0j5_OVJa0j5_OVJ`0j5AVJ?i5BWJ=j5CVJ=i5DWJ;i5FWJ:h5GXJ9h5GXJ8h5IXJ7g5JYJ5h5KXJ5g5LYJ3g5NYJ2f5OZJ0g50YJ0f51ZJOe52[JMf53ZJMe54[JKe55\\JKc56]JId57\\JIc58]JGc5:]JFb5;^JEb5;^JDb5=^JCa5>_JAa5`0_J@a5`0_J_Oa5b0_J^O`5c0`J\\Oa5d0_J\\O`5e0`J[O_5f0aJYO_5h0aJXO_5h0aJWO_5j0aJVO^5k0bJTO^5m0bJSO^5m0bJRO^5o0bJQO]5P1bJQO]5P1cJoN^5Q1bJoN]5R1cJmN]5T1cJlN^5S1bJlN`5S1`JmNb5Q1^JnNd5Q1\\JoNf5o0ZJQOg5n0YJQOj5m0VJSOk5l0UJSOn5k0RJUOo5j0QJUOR6i0nIWOR6i0nIVOR6k0nIUOP6m0PJSOo5n0QJQOn5Q1RJoNm5R1SJmNo5R1QJnNo5R1QJmNQ6R1oInNQ6R1oImNS6R1mInNS6R1mInNT6Q1lInNU6R1kInNV6Q1jInNW6R1iIgN^6Y1bI_Nf6a1ZIWNm6j1SImMU7T2kHdM\\7]2dH[Md7e2\\HSMk7n2UHjLR8W3nGaLY8`3:0O1O100O1O1O100O1O1O1O1O1O1O1O1O1O100O10000O10YNiGXOW8g0mGVOR8k0PHSOP8l0THQOl7n0WHPOh7Q1ZHmNf7R1^HkNa7U1bHiN^7W1cHhN]7W1dHiN[7W1fHiNZ7W1gHhNi6nNjHY2?iNg6oNhHV2d0kNd6oNgHU2g0lNb6POeHR2l0nN_6POdHP2Q1oNZ6SOcHl1V1QOW6SObHk1Y1ROU6SOaHi1]1TOR6TO_Hf1b1VOn5UO`Hb1e1YOk5n0XJQOh5l0[JTOe5i0^JWOa5h0aJXO_5f0cJZO]5d0eJ\\O[5b0hJ]OX5a0jJ_OU5`0mJ@S5>oJBQ5<QKDo4;RKEm4:UKFk48WKHi46YKJg44[KLd43]KNc40_K0a4NaK2_4LcK4\\4JgK6Y4EjK=V4^OlKe0T4VOnKm0R4nNPLU1o3gNTL[1\\73M3M3N2N1O2M3N2N2N2M3N2N2N2N2N2O1N2O1N2O1N2O1N2O1N2O1M3N2M3M3N2M3M3N2M3O11O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O100O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N:Fe0[OVPV1"}, "label": "the brown chair in the front on which cats is playing"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat on its back\"."}], "task": "refering", "answer_template": "The \"the cat on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 78, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303], "bbox": [0.001390625, 0.19261964735516374, 0.5911718749999999, 0.9930982367758187], "iscrowd": 0, "area": 62258.99389999999, "rle": {"size": [397, 640], "counts": "Vc0_5n60O100O100O2O0O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1000000000000000000O100000000000000000000000000O1000001O000000000000000000000O100000000000000000000000000O100000000000000000000000WOPJUKP6k4RJSKn5m4SJRKm5n4UJPKk5o4WJPKi5P5YJnJg5R5ZJmJf5S5[JkJe5U5^JgJd5Y5]JdJe5\\5]J`Je5`5d0O1000000O1000000O1000000O1000000O1000000O0100000O100@RIaKn6_4UI^Kk6b4YIYKh6f4\\IWKd6i4`000O10000O1000000O10000O10000O10O1000O10O1O010O1O010O1O010BZKTIg4k6^KPIc4o6bKmH]4S7?00O001O100O001O1O010O1O1O010O1O001O1O1O001O1O1N101O1O001O1O1O001O1O001O1O1O0O2O1O001O1O1O001O1000O010O0010O01O00100O001O010O001O010O1O010O001O010O001O010O1O0010OPLlIh1U6WNkIi1U6WNlIh1U6WNlIi1S6WNnIh1S6XNmIg1T6XNmIg1S6YNmIh1S6WNnIh1R6XNoIg1R6XNoIg1Q6YNPJg1P6YNPJf1P6ZNPJf1Q6YNPJg1P6XNQJg1o5YNRJf1o5YNRJf1n5ZNRJg1n5YNRJf1n5ZNSJe1n5ZNSJe1m5[NTJe1l5ZNUJe1l5ZNTJf1l5ZNUJe1l5ZNUJf1j5[NVJd1k5[NVJd1j5\\NWJc1j5\\NVJe1i5[NXJd1i5[NXJd1i5[NXJd1h5]NXJc1h5\\NYJe1e5[N\\Jf1c5YN^Jg1a5YN`Jh1_5WNbJj1\\5VNeJk1Z5UNfJk1Z5TNgJm1W5SNjJn1U5QNlJP2R5PNoJP2Q5oMPKQ2o4oMRKQ2n4oMRKQ2m4oMTKR2k4mMVKS2i4mMXKS2h4lMYKT2g4kMZKU2e4kM\\KU2d4[MRJ@[1T3b4YMZJ_OU1X3a4TMWLl2h3PM]LQ3a3lLcLT3\\3hLiLX3W3dLmL\\3R60000000001O0000000000000000010O000L4L4M3O1O2M3N2N2M3N2N2O1O1N2O3M2N3M3L3N3Md0\\Ok0UOO010O1O10O0N3M2O2M2NP[U3"}, "label": "the cat on its back"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cat that has its mouth open and is upside down , swatting another orange cat\"."}], "task": "refering", "answer_template": "The \"an orange cat that has its mouth open and is upside down , swatting another orange cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 78, 79, 80, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303], "bbox": [0.001390625, 0.19261964735516374, 0.5911718749999999, 0.9930982367758187], "iscrowd": 0, "area": 62258.99389999999, "rle": {"size": [397, 640], "counts": "Vc0_5n60O100O100O2O0O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1000000000000000000O100000000000000000000000000O1000001O000000000000000000000O100000000000000000000000000O100000000000000000000000WOPJUKP6k4RJSKn5m4SJRKm5n4UJPKk5o4WJPKi5P5YJnJg5R5ZJmJf5S5[JkJe5U5^JgJd5Y5]JdJe5\\5]J`Je5`5d0O1000000O1000000O1000000O1000000O1000000O0100000O100@RIaKn6_4UI^Kk6b4YIYKh6f4\\IWKd6i4`000O10000O1000000O10000O10000O10O1000O10O1O010O1O010O1O010BZKTIg4k6^KPIc4o6bKmH]4S7?00O001O100O001O1O010O1O1O010O1O001O1O1O001O1O1N101O1O001O1O1O001O1O001O1O1O0O2O1O001O1O1O001O1000O010O0010O01O00100O001O010O001O010O1O010O001O010O001O010O1O0010OPLlIh1U6WNkIi1U6WNlIh1U6WNlIi1S6WNnIh1S6XNmIg1T6XNmIg1S6YNmIh1S6WNnIh1R6XNoIg1R6XNoIg1Q6YNPJg1P6YNPJf1P6ZNPJf1Q6YNPJg1P6XNQJg1o5YNRJf1o5YNRJf1n5ZNRJg1n5YNRJf1n5ZNSJe1n5ZNSJe1m5[NTJe1l5ZNUJe1l5ZNTJf1l5ZNUJe1l5ZNUJf1j5[NVJd1k5[NVJd1j5\\NWJc1j5\\NVJe1i5[NXJd1i5[NXJd1i5[NXJd1h5]NXJc1h5\\NYJe1e5[N\\Jf1c5YN^Jg1a5YN`Jh1_5WNbJj1\\5VNeJk1Z5UNfJk1Z5TNgJm1W5SNjJn1U5QNlJP2R5PNoJP2Q5oMPKQ2o4oMRKQ2n4oMRKQ2m4oMTKR2k4mMVKS2i4mMXKS2h4lMYKT2g4kMZKU2e4kM\\KU2d4[MRJ@[1T3b4YMZJ_OU1X3a4TMWLl2h3PM]LQ3a3lLcLT3\\3hLiLX3W3dLmL\\3R60000000001O0000000000000000010O000L4L4M3O1O2M3N2N2M3N2N2O1O1N2O3M2N3M3L3N3Md0\\Ok0UOO010O1O10O0N3M2O2M2NP[U3"}, "label": "an orange cat that has its mouth open and is upside down , swatting another orange cat"}]}
{"id": 430563, "image": "COCO_train2014_000000430563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark red fluffy pillow\"."}], "task": "refering", "answer_template": "The \"dark red fluffy pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 319, 320, 321], "bbox": [0.61571875, 0.009874055415617129, 1.0, 0.9777833753148615], "iscrowd": 0, "area": 29288.539699999994, "rle": {"size": [397, 640], "counts": "Vhh44Y<5K00O10000000000O1000000000000O10000000000O1000000000000O10000000000001O00001[DKP;5kD1T;OgD6Y;KaD:_;:01O000000001O0000001O0jKTOkLl0Q3YOnLg0n2^OQMc0k2ATM?h2FWM:e2KZM5c2O\\M1`23`MN[27dMIX2<gMDV2`0iM@S2e0lM\\Oo1i0PNWOm1m0RNSOj1R1UNnNg1W1XNiNe1[1ZNfNa1_1^NaN^1d1aN\\N_1e1`N[N`1e1`N\\N_1e1`N[N`1f1_NZNa1g1^NYNb1h1]NYNb1h1]NXNc1i1\\NWNd1j1[NVNe1k1ZNUNf1l1YNTNg1m1XNSNh1n1WNRNi1n1WNRNi1o1VNQNj1P2UNPNk1Q2TNoMl1R2SNmMn1T2QNlMo1U2PNkMP2V2oMjMQ2W2nMiMR2X2mMhMS2Y2mMfMS2[2lMeMT2\\2kMdMU2\\2kMdMU2]2jMcMV2^2iMbMW2_2hM`MY2a2fM_MZ2b2eM^M[2c2dM]M\\2d2cM\\M]2e2bM[M^2f2aMZM_2g2`MYM`2h2_MXMa2i2^MWMb2i2^MWMb2j2]MVMc2k2\\MTMe2m2ZMSMf2n2YMRMg2o2XMQMh2P3WMPMi2Q3VMoLk2Q3TMoLl2R3SMnLm2S3RMmLn2T3QMlLo2U3PMkLP3U3PMkLP3V3oLiLR3X3mLhLS3Y3lLgLT3Z3kLfLU3[3jLeLV3\\3iLdLW3]3hLcLX3^3gLbLY3_3fLaLZ3`3eL`L[3a3dL_L\\3C]Kk2V1bM]3C^Kj2U1bM^3C^Kl2S1aM_3C^Km2R1`M`3B_Ko2P1_Ma3B`Ko2n0_Mb3AaKQ3l0^Mc3AaKR3k0]Md3@cKS3h0]Me3@cKT3g0\\Mf3_OdKV3e0[Mg3_OeKV3c0[Mh3^OfKX3a0ZMi3^OfKY3`0YMj3]OgKZ3?YMj3]OhKZ3=XMl3^OgK[3<WMm3^OgK\\3;VMn3_OgK[3:VMo3_OgK\\39UMP4_OgK\\39UMP4_OhK\\37UMQ4_OhK\\37UMQ4@gK\\37TMR4@gK\\37TMR4@hK\\35TMS4@hK\\35TMS4@hK]34SMT4AhK\\33SMU4AhK\\33RMW4AfK^32QMX4AfK^32QMY4AdK_32PM[4@cK`32PM\\4_ObKb31oL^4^OaKc31oL_4]O_Kf31mLa4]O]Kf32mLa4]O]Kg31lLc4\\O\\Ki30kLe4[O[Kj30kLf4ZOZKl3OjLh4ZOXKl30jLi4YOWKn3OiLk4XOVKo3OhLm4XOTKQ4NgLn4YOSKP4OgLo4XOQKS4OeLQ5WOPKU4NdLS5VOoJV4NdLm5]3RJcLn5]3RJcLn5^3QJbLo5^3QJbLo5_3PJaLP6_3PJaLP6`3oI`LQ6`3oI`LQ6a3nI_LQ6c3nI\\LS6d3mI\\LS6e3lI[LT6e3lI[LT6f3kIZLU6f3kIZLU6g3jIYLV6g3jIYLV6g3jIYLV6g3kIXLU6h3kIXLT6i3lIWLT6i3lIWLT6i3lIWLT6i3lIWLT6i3lIVLU6j3kIVLU6j3kIVLU6j3kIVLU6j3kIVLU6j3kIULV6j3kIVLU6j3kIVLT6k3lIULT6k3lITLU6l3kITLU6l3kITLU6l3kITLU6l3kISLV6m3jISLV6m3jISLV6m3jISLV6m3jIRLW6n3iIRLW6n3iIRLV6o3jIQLV6o3jIPLW6P4iIPLW6P4iIPLW6P4iIPLW6P4iIoKX6Q4iInKW6Q4PJiKP6W4VJcKj5]4S10O1000XOh0^MSK"}, "label": "dark red fluffy pillow"}]}
{"id": 176617, "image": "COCO_train2014_000000176617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow frisbee on man ' s hand\"."}], "task": "refering", "answer_template": "The \"yellow frisbee on man ' s hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67], "bbox": [0.06413875598086125, 0.10426, 0.501866028708134, 0.27586], "iscrowd": 0, "area": 8212.7919, "rle": {"size": [500, 418], "counts": "YX=1b?2N101N1O2N101N1O2N2O0O2N2N101N1O2N2N101N2N1O2N2O0O2N2N1O2O1N1O2O10O01O001O10O01O1O010O1O001O10O01O1O0010O01O1O00100O001O10O01O1O0010O0001O0001O000000001O000000001O0001O01O000000001O000000001O0O10000O2O00000O101O0O10001N100O10001N10000O16K8G8H1O0001O000O10001O00001O1O001O1O1N2O1O1O001O1O1O1O1O1N101O1O1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O1N1N3N2N2N2M3N2N1O2M3N2N2NRnU3"}, "label": "yellow frisbee on man ' s hand"}]}
{"id": 176617, "image": "COCO_train2014_000000176617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow frisbee\"."}], "task": "refering", "answer_template": "The \"a yellow frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 63, 64, 65, 66, 67], "bbox": [0.06413875598086125, 0.10426, 0.501866028708134, 0.27586], "iscrowd": 0, "area": 8212.7919, "rle": {"size": [500, 418], "counts": "YX=1b?2N101N1O2N101N1O2N2O0O2N2N101N1O2N2N101N2N1O2N2O0O2N2N1O2O1N1O2O10O01O001O10O01O1O010O1O001O10O01O1O0010O01O1O00100O001O10O01O1O0010O0001O0001O000000001O000000001O0001O01O000000001O000000001O0O10000O2O00000O101O0O10001N100O10001N10000O16K8G8H1O0001O000O10001O00001O1O001O1O1N2O1O1O001O1O1O1O1O1N101O1O1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O1N1N3N2N2N2M3N2N1O2M3N2N2NRnU3"}, "label": "a yellow frisbee"}]}
{"id": 4587, "image": "COCO_train2014_000000004587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bowl of noodles with a pair of tongs is sitting in front of the buffet platter\"."}], "task": "refering", "answer_template": "The \"a white bowl of noodles with a pair of tongs is sitting in front of the buffet platter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 378, 379, 380, 381, 382, 383], "bbox": [0.43014062500000005, 0.7934166666666667, 0.690203125, 1.0], "iscrowd": 0, "area": 12529.43895, "rle": {"size": [480, 640], "counts": "bYQ43m>:F9G2N2N3L3N2N2N3L3N2N2N2N2M3N2N2N2N2M3N2N2N2N1N3N2N2N1O2M3N0000000O1000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O010000000O100000000O1000000O100000000O10000000000O100000000000000O100000000001O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2N2N2O1L4L4K5K5L4K5L4K5K5L^\\l2"}, "label": "a white bowl of noodles with a pair of tongs is sitting in front of the buffet platter"}]}
{"id": 4587, "image": "COCO_train2014_000000004587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white square bowl with onions and silver serving tongs\"."}], "task": "refering", "answer_template": "The \"white square bowl with onions and silver serving tongs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 378, 379, 380, 381, 382, 383], "bbox": [0.43014062500000005, 0.7934166666666667, 0.690203125, 1.0], "iscrowd": 0, "area": 12529.43895, "rle": {"size": [480, 640], "counts": "bYQ43m>:F9G2N2N3L3N2N2N3L3N2N2N2N2M3N2N2N2N2M3N2N2N2N1N3N2N2N1O2M3N0000000O1000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O010000000O100000000O1000000O100000000O10000000000O100000000000000O100000000001O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2N2N2O1L4L4K5K5L4K5L4K5K5L^\\l2"}, "label": "white square bowl with onions and silver serving tongs"}]}
{"id": 438769, "image": "COCO_train2014_000000438769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the underside of the black , white and red snowboard\"."}], "task": "refering", "answer_template": "The \"the underside of the black , white and red snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 23, 24, 25, 26, 39, 40, 41, 42, 56, 57, 58, 59, 72, 73, 74, 75, 89, 90, 91, 92, 106, 107, 108, 109, 122, 123, 124, 125, 139, 140, 141, 142, 155, 156, 157, 158, 172, 173, 174, 189], "bbox": [0.45094091903719913, 0.018140625, 0.9304595185995624, 0.5079375], "iscrowd": 0, "area": 22114.474050000008, "rle": {"size": [640, 457], "counts": "[iP41nc03M3M2M4M3M2N3M3M2N3L4M2N3M3M2N3M3L3N3M3N101N2N1O2N2N1O2N2N1O2N2JdMd^O]2[a0cMe^O_2Ya0aMg^Oa2Wa0_Mi^Ob2Ua0_Mk^Oc2Sa0]Mm^Oe2Qa0[Mo^Of2Pa0[Mo^Og2o`0YMQ_Oi2l`0XMU_Oi2i`0WMW_Oj2h`0VMX_Ol2f`0TMZ_On2d`0RM\\_Oo2b`0SM]_Oo2a`0QM__OQ3_`0oLa_OR3^`0nLb_OT3\\`0lLd_OV3Y`0kLg_OV3X`0jLh_OX3V`0hLj_OZ3T`0gLk_OZ3T`0fLl_O\\3R`0dLn_O^3o?cLQ@^3n?bLR@`3l?`LT@b3j?^LW@b3h?_LW@c3h?\\LX@f3f?ZLZ@g3f?XLZ@j3e?UL[@m3d?RL\\@o3d?PL\\@R4c?601O001O001O001O001O0O2O001O0100O2O1N101N2O0O2O1N101N2O0O2O1N101N10O0100O010O10O0100O010O010O10O0100O010O10O0100O010O010O10O0100O010O10O010O0100O010O10O0100O010O10O010O0100O010O10O0100O010O10O010O0100O010O10O0100O010O10O01N1N3L4M2N3L4M2N3L4M2N3L3N3L4M2N3L4M2N3L4M2N3L4M2M4M2N3L4M2N3L4M2M4M3M2M4M3M2Mdgc0"}, "label": "the underside of the black , white and red snowboard"}]}
{"id": 438769, "image": "COCO_train2014_000000438769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the snowboard suspended in the air\"."}], "task": "refering", "answer_template": "The \"the snowboard suspended in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 23, 24, 25, 26, 39, 40, 41, 42, 56, 57, 58, 59, 72, 73, 74, 75, 89, 90, 91, 92, 106, 107, 108, 109, 122, 123, 124, 125, 139, 140, 141, 142, 155, 156, 157, 158, 172, 173, 174, 189], "bbox": [0.45094091903719913, 0.018140625, 0.9304595185995624, 0.5079375], "iscrowd": 0, "area": 22114.474050000008, "rle": {"size": [640, 457], "counts": "[iP41nc03M3M2M4M3M2N3M3M2N3L4M2N3M3M2N3M3L3N3M3N101N2N1O2N2N1O2N2N1O2N2JdMd^O]2[a0cMe^O_2Ya0aMg^Oa2Wa0_Mi^Ob2Ua0_Mk^Oc2Sa0]Mm^Oe2Qa0[Mo^Of2Pa0[Mo^Og2o`0YMQ_Oi2l`0XMU_Oi2i`0WMW_Oj2h`0VMX_Ol2f`0TMZ_On2d`0RM\\_Oo2b`0SM]_Oo2a`0QM__OQ3_`0oLa_OR3^`0nLb_OT3\\`0lLd_OV3Y`0kLg_OV3X`0jLh_OX3V`0hLj_OZ3T`0gLk_OZ3T`0fLl_O\\3R`0dLn_O^3o?cLQ@^3n?bLR@`3l?`LT@b3j?^LW@b3h?_LW@c3h?\\LX@f3f?ZLZ@g3f?XLZ@j3e?UL[@m3d?RL\\@o3d?PL\\@R4c?601O001O001O001O001O0O2O001O0100O2O1N101N2O0O2O1N101N2O0O2O1N101N10O0100O010O10O0100O010O010O10O0100O010O10O0100O010O010O10O0100O010O10O010O0100O010O10O0100O010O10O010O0100O010O10O0100O010O10O010O0100O010O10O0100O010O10O01N1N3L4M2N3L4M2N3L4M2N3L3N3L4M2N3L4M2N3L4M2N3L4M2M4M2N3L4M2N3L4M2M4M3M2M4M3M2Mdgc0"}, "label": "the snowboard suspended in the air"}]}
{"id": 569851, "image": "COCO_train2014_000000569851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple sits under a bunch of grapes and beside a pile of kiwis , with another apple beneath it\"."}], "task": "refering", "answer_template": "The \"an apple sits under a bunch of grapes and beside a pile of kiwis , with another apple beneath it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 332, 333, 334], "bbox": [0.42543749999999997, 0.6687587822014052, 0.6586874999999999, 1.0], "iscrowd": 0, "area": 13237.003700000005, "rle": {"size": [427, 640], "counts": "Uia36d<c0E<K5K5J5L5K4L4L3M4K5L3M4M3M1O1O1N2O1O1O1O1O1N2O1O2N1O1O1N2O2N2N2N2N2O0O2O1N;F0O100O100O1O100O100O100O1000000O100000000O1000000O10000000000000D<N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2O1O1O1O1O1O1O1O1O1N3N1O1O2N1O2N1O2N1O2N1O1O2N101N102M2N2O1N2N2O1N2O1N2N2O2M2O1N2O1N2O1O1O1N2O1O1N2O2N1N2O1O3M3L4M3M3L3N3MVbj2"}, "label": "an apple sits under a bunch of grapes and beside a pile of kiwis , with another apple beneath it"}]}
{"id": 569851, "image": "COCO_train2014_000000569851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple next to the kiwi and in front of the grapes\"."}], "task": "refering", "answer_template": "The \"apple next to the kiwi and in front of the grapes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 332, 333, 334], "bbox": [0.42543749999999997, 0.6687587822014052, 0.6586874999999999, 1.0], "iscrowd": 0, "area": 13237.003700000005, "rle": {"size": [427, 640], "counts": "Uia36d<c0E<K5K5J5L5K4L4L3M4K5L3M4M3M1O1O1N2O1O1O1O1O1N2O1O2N1O1O1N2O2N2N2N2N2O0O2O1N;F0O100O100O1O100O100O100O1000000O100000000O1000000O10000000000000D<N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2O1O1O1O1O1O1O1O1O1N3N1O1O2N1O2N1O2N1O2N1O1O2N101N102M2N2O1N2N2O1N2O1N2N2O2M2O1N2O1N2O1O1O1N2O1O1N2O2N1N2O1O3M3L4M3M3L3N3MVbj2"}, "label": "apple next to the kiwi and in front of the grapes"}]}
{"id": 569851, "image": "COCO_train2014_000000569851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple in the middle of two other apples\"."}], "task": "refering", "answer_template": "The \"apple in the middle of two other apples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.534375, 0.8035128805620609, 0.809375, 0.9885245901639345], "iscrowd": 0, "area": 10630.5, "rle": {"size": [427, 640], "counts": "eP_41X=3M2N2N2O2M2N2N2N2N2N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O100O10000O10000O100O10000O100O10000O100O10000O100O100000000O1000000000000O10000000000000000000000000000000000000000000000000000001O001O001O00001O001O001O00001O001O00001O001O001O001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001N2N3M6J6J5K6K5I\\_b1"}, "label": "apple in the middle of two other apples"}]}
{"id": 373259, "image": "COCO_train2014_000000373259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue color car parked near a tree\"."}], "task": "refering", "answer_template": "The \"a blue color car parked near a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 105, 106, 107, 122, 123, 124, 125, 132, 133, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 214, 215, 233], "bbox": [0.035540275049115916, 0.24958250497017895, 0.9926522593320236, 0.6922862823061631], "iscrowd": 0, "area": 24149.71785, "rle": {"size": [503, 509], "counts": "aR9;\\?e0[Oc0]O001N1000001O00001N1000001O00001N1000001O00000O2O00001O0000001N10001O000000O010O1O100O100O010O100O100O1O010O100O100O10O010000O1000000O01000O10000O10O1000O10000O10O10O10000O1000O10O10000O1000O01000000O1000O010000O1000000O01000O10000O10O10O1000000O10O10O10000O1000O10O10000O1000O010000O100000O010000O10000O10O1000O10000O10O10O1000000O10O10O10000O1000O01000000O1000O01N2N2N2N1O2N3M4L4L5K4L5KR^e20oaZM2O3L4M3L4L4M3L5K4M3L4M3L5K4M3L4L5L3L4L4M3L4M1N2N2O1N2N2O0O2O1N2N2O1N2N2O0O2N2O1N100O2N101N1O101N100O2N101N1O101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O1O2O0O2N3N4K5L4K5K5L4K5K5L4K5K3N1N100O2N101N1O101N101N1O101N1O2O0O1O2O0O101R1mNX1iNW1hNX1hNW1jNkd1"}, "label": "a blue color car parked near a tree"}]}
{"id": 373259, "image": "COCO_train2014_000000373259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car parked behind a woman in sunglasses\"."}], "task": "refering", "answer_template": "The \"a blue car parked behind a woman in sunglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 105, 106, 107, 122, 123, 124, 125, 132, 133, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 214, 215, 233], "bbox": [0.035540275049115916, 0.24958250497017895, 0.9926522593320236, 0.6922862823061631], "iscrowd": 0, "area": 24149.71785, "rle": {"size": [503, 509], "counts": "aR9;\\?e0[Oc0]O001N1000001O00001N1000001O00001N1000001O00000O2O00001O0000001N10001O000000O010O1O100O100O010O100O100O1O010O100O100O10O010000O1000000O01000O10000O10O1000O10000O10O10O10000O1000O10O10000O1000O01000000O1000O010000O1000000O01000O10000O10O10O1000000O10O10O10000O1000O10O10000O1000O010000O100000O010000O10000O10O1000O10000O10O10O1000000O10O10O10000O1000O01000000O1000O01N2N2N2N1O2N3M4L4L5K4L5KR^e20oaZM2O3L4M3L4L4M3L5K4M3L4M3L5K4M3L4L5L3L4L4M3L4M1N2N2O1N2N2O0O2O1N2N2O1N2N2O0O2N2O1N100O2N101N1O101N100O2N101N1O101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O1O2O0O2N3N4K5L4K5K5L4K5K5L4K5K3N1N100O2N101N1O101N101N1O101N1O2O0O1O2O0O101R1mNX1iNW1hNX1hNW1jNkd1"}, "label": "a blue car parked behind a woman in sunglasses"}]}
{"id": 438795, "image": "COCO_train2014_000000438795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant standing second to last in a line of posing elephants\"."}], "task": "refering", "answer_template": "The \"an elephant standing second to last in a line of posing elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 59, 82, 105, 106, 129, 198, 221, 244], "bbox": [0.541125, 0.08348946135831381, 0.6445000000000001, 0.7010538641686181], "iscrowd": 0, "area": 2304.0441499999997, "rle": {"size": [427, 640], "counts": "_g`42X=101O000O2O00001O00001O001O1O00100O1O101N4M3M3L6K3Ld0TDeNa:T2N10000O11O0010000ZNVEi0j:RO`Ei0a:ROdEm0_;N101gNoCT1U<O[HROm2n0oLVOR3h0kL[OW3d0eL^O^3`0_LCc3;YLHj37PLMS41cK8_4GiJP1Y5POdJQ1^5oN_J@SNb0`7O[JWOcNd0U74_KHc48`KBd4<]KBf4=[K_Oh4a0S4O1N2N3M3L5KTmn2"}, "label": "an elephant standing second to last in a line of posing elephants"}]}
{"id": 438795, "image": "COCO_train2014_000000438795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second elephant rearing up that is barely visible\"."}], "task": "refering", "answer_template": "The \"the second elephant rearing up that is barely visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 59, 82, 105, 106, 129, 198, 221, 244], "bbox": [0.541125, 0.08348946135831381, 0.6445000000000001, 0.7010538641686181], "iscrowd": 0, "area": 2304.0441499999997, "rle": {"size": [427, 640], "counts": "_g`42X=101O000O2O00001O00001O001O1O00100O1O101N4M3M3L6K3Ld0TDeNa:T2N10000O11O0010000ZNVEi0j:RO`Ei0a:ROdEm0_;N101gNoCT1U<O[HROm2n0oLVOR3h0kL[OW3d0eL^O^3`0_LCc3;YLHj37PLMS41cK8_4GiJP1Y5POdJQ1^5oN_J@SNb0`7O[JWOcNd0U74_KHc48`KBd4<]KBf4=[K_Oh4a0S4O1N2N3M3L5KTmn2"}, "label": "the second elephant rearing up that is barely visible"}]}
{"id": 438795, "image": "COCO_train2014_000000438795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant plucking leaves\"."}], "task": "refering", "answer_template": "The \"an elephant plucking leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 286, 287, 288, 305, 306, 309, 311, 312, 328, 329], "bbox": [0.151765625, 0.3322014051522248, 0.572515625, 0.940304449648712], "iscrowd": 0, "area": 38534.94424999999, "rle": {"size": [427, 640], "counts": "ldX11W=4K4L5L4K4M4K4M4K5L3L5L3L5L3L5K5L3M4O010O10O01O010O010O010O01O00jEhMY9X2cFlM^9T2]FoMd9Q2XFSNi9l1SFXNm9h1oE[NS:d1iE`NW:Y22O000O2O0O2O00O001O1O001O1O100O1O1O1O1O1000000O2OV2jM`0@9F9H9G9F9H9G7I5J1000000000000001O00000000001O00000000001O0001O00000001O00000000001O1O2N1O2N2N2N2N2N1O2hN]JdJe5W5gJaJZ5Y5SK^Jm4_5a1L4M3M4K4M3M3L4M3M4K4M3O1001O01O01O00000010O0001O0000010O00010O0001O010O00010O000010O00010O00010O0001O010O0O101O00001O00001O00001O00[HiKW6V4bIQL^6o3[IXLe6h3TI_Lm6`3kHhLU7X3dHnL]7`41O00001N10001O001N10001O0O101O001O1N2O1O1^NUHbMl7]2\\H[Me7e2cHQM_7n2iHjLX7U3PIcLQ7\\3`100O1O100O1O00100O1O1O11O2N2N2N2N1O2N2N2N2N2N2N2N2N2M210<D<D<D<D9G1N2L4K6K4L4L4L5J5L5K5K4K6K5K4L5K5J[la3"}, "label": "an elephant plucking leaves"}]}
{"id": 438795, "image": "COCO_train2014_000000438795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant standing on all four feet , with two elephants behind it\"."}], "task": "refering", "answer_template": "The \"an elephant standing on all four feet , with two elephants behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 286, 287, 288, 305, 306, 309, 311, 312, 328, 329], "bbox": [0.151765625, 0.3322014051522248, 0.572515625, 0.940304449648712], "iscrowd": 0, "area": 38534.94424999999, "rle": {"size": [427, 640], "counts": "ldX11W=4K4L5L4K4M4K4M4K5L3L5L3L5L3L5K5L3M4O010O10O01O010O010O010O01O00jEhMY9X2cFlM^9T2]FoMd9Q2XFSNi9l1SFXNm9h1oE[NS:d1iE`NW:Y22O000O2O0O2O00O001O1O001O1O100O1O1O1O1O1000000O2OV2jM`0@9F9H9G9F9H9G7I5J1000000000000001O00000000001O00000000001O0001O00000001O00000000001O1O2N1O2N2N2N2N2N1O2hN]JdJe5W5gJaJZ5Y5SK^Jm4_5a1L4M3M4K4M3M3L4M3M4K4M3O1001O01O01O00000010O0001O0000010O00010O0001O010O00010O000010O00010O00010O0001O010O0O101O00001O00001O00001O00[HiKW6V4bIQL^6o3[IXLe6h3TI_Lm6`3kHhLU7X3dHnL]7`41O00001N10001O001N10001O0O101O001O1N2O1O1^NUHbMl7]2\\H[Me7e2cHQM_7n2iHjLX7U3PIcLQ7\\3`100O1O100O1O00100O1O1O11O2N2N2N2N1O2N2N2N2N2N2N2N2N2M210<D<D<D<D9G1N2L4K6K4L4L4L5J5L5K5K4K6K5K4L5K5J[la3"}, "label": "an elephant standing on all four feet , with two elephants behind it"}]}
{"id": 438795, "image": "COCO_train2014_000000438795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the back near the center of the dirt\"."}], "task": "refering", "answer_template": "The \"the elephant in the back near the center of the dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 82, 83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225], "bbox": [0.597640625, 0.1306791569086651, 0.8026093750000001, 0.645807962529274], "iscrowd": 0, "area": 14993.064200000004, "rle": {"size": [427, 640], "counts": "l[o49l<8L3N3M2M3N2N101N2HTOoCl0o;:N2O0O2O001O010O0SE`No9a1oEbNP:]1oEfNP:4XEi0f0XOP:K[EP1b0ZOY:g0dE_OW:d0eEAW:b0dECZ:i1M3cEZMP:P310O0001O00001OO3N2N2O001N2O1O1O1O1O9G4L4L2N1O2N2N2cGdKg05`5Y4`IlKk0Me5Y4WIULo0Di5Q5RJQKo5Q5mIPKS6T5hImJX6S61O001O001O001O1O2N1O1O1N2O1O1O1O3jNlHoKV7n3WIdKm6Y4^I\\Kc6b4U1L4M3N200lKWGd3Y900O010N2O0O3M2N3L3N2N1O22N8H3M1O0O2O1O1O010O100O1O101N101kFTLf8Y41O0O2]NXG_Ni8[1`GbN`8X1gGfN[8W1hGgNY8Z1gGdN[8\\1fGaN\\8n0lFbNh0>_8k0oFfNa0=c8h0QGkN;;f8f0UGnN59i8MoGIXO7o:JQEOV;0U1HSZd1"}, "label": "the elephant in the back near the center of the dirt"}]}
{"id": 438795, "image": "COCO_train2014_000000438795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the end of the line of three\"."}], "task": "refering", "answer_template": "The \"the elephant in the end of the line of three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 82, 83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225], "bbox": [0.597640625, 0.1306791569086651, 0.8026093750000001, 0.645807962529274], "iscrowd": 0, "area": 14993.064200000004, "rle": {"size": [427, 640], "counts": "l[o49l<8L3N3M2M3N2N101N2HTOoCl0o;:N2O0O2O001O010O0SE`No9a1oEbNP:]1oEfNP:4XEi0f0XOP:K[EP1b0ZOY:g0dE_OW:d0eEAW:b0dECZ:i1M3cEZMP:P310O0001O00001OO3N2N2O001N2O1O1O1O1O9G4L4L2N1O2N2N2cGdKg05`5Y4`IlKk0Me5Y4WIULo0Di5Q5RJQKo5Q5mIPKS6T5hImJX6S61O001O001O001O1O2N1O1O1N2O1O1O1O3jNlHoKV7n3WIdKm6Y4^I\\Kc6b4U1L4M3N200lKWGd3Y900O010N2O0O3M2N3L3N2N1O22N8H3M1O0O2O1O1O010O100O1O101N101kFTLf8Y41O0O2]NXG_Ni8[1`GbN`8X1gGfN[8W1hGgNY8Z1gGdN[8\\1fGaN\\8n0lFbNh0>_8k0oFfNa0=c8h0QGkN;;f8f0UGnN59i8MoGIXO7o:JQEOV;0U1HSZd1"}, "label": "the elephant in the end of the line of three"}]}
{"id": 258571, "image": "COCO_train2014_000000258571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl holding baseball bat\"."}], "task": "refering", "answer_template": "The \"little girl holding baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 122, 123, 145, 146, 147, 168, 169, 170, 192, 193, 215, 216, 238, 239, 262, 263, 285, 308], "bbox": [0.30846874999999996, 0.3109133489461358, 0.4644375, 0.8849414519906323], "iscrowd": 0, "area": 8125.837749999999, "rle": {"size": [427, 640], "counts": "gYb21S=9G9M3M3N2M3N3L3^DPOe:l1N2O1O1O1O1O1O1O1O1O1O1O1O1O001WFdMa8]2^GhM]8Y2bGlMY8U2fGPNT8R2kGSNo7o1QHUNi7m1VHXNd7j1[H[N^7h1aHYN]7i1bHXN\\7j1cHWN\\7j1cHWN]7i1bHXN]7i1cHWN]7h1cHYN]7f1cH[N\\7e1dH[N]7e1cH[N]7d1cH]N\\7c1eH]N[7b1eH_N[7`1fH`NY7a1gH_NY7`1gHaNY7^1hHcNV7FQH>i0MV7]OYHe0b0N[7oN[HS1:Of0`NU50RJa1309kN_6T1YI2IXOn6e0YI4_O@Y7;YI6TOHc72YIf1h6YNXIh1h6WNYIi1g6VNYIj1T7iMmHW2Y7bMhH]2_7\\MaHe2e7TM\\Hk2d7UM\\Hl2a7VM`Hi2_7XMbH`1nNWO]8\\OdH]1TOROW8BfH\\1UOPOS8EiHZ1VOoNS8EgH\\1WOnNS8EgH]1VOlNU8FfH]1WOjNV8GcH_1XOhNW8HbH`1WOfNY8IlHU1lNPOZ8J[Ie0]N^O[8KaI?TND^8K_I`0TNCb8I[Ic0UNAU;?kD_OW;`0jD^OX;a0i001N2O1N2O001N101N2O0O2O1O0O2O0O2O`f^4"}, "label": "little girl holding baseball bat"}]}
{"id": 258571, "image": "COCO_train2014_000000258571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl holding a yellow bat\"."}], "task": "refering", "answer_template": "The \"a little girl holding a yellow bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 122, 123, 145, 146, 147, 168, 169, 170, 192, 193, 215, 216, 238, 239, 262, 263, 285, 308], "bbox": [0.30846874999999996, 0.3109133489461358, 0.4644375, 0.8849414519906323], "iscrowd": 0, "area": 8125.837749999999, "rle": {"size": [427, 640], "counts": "gYb21S=9G9M3M3N2M3N3L3^DPOe:l1N2O1O1O1O1O1O1O1O1O1O1O1O1O001WFdMa8]2^GhM]8Y2bGlMY8U2fGPNT8R2kGSNo7o1QHUNi7m1VHXNd7j1[H[N^7h1aHYN]7i1bHXN\\7j1cHWN\\7j1cHWN]7i1bHXN]7i1cHWN]7h1cHYN]7f1cH[N\\7e1dH[N]7e1cH[N]7d1cH]N\\7c1eH]N[7b1eH_N[7`1fH`NY7a1gH_NY7`1gHaNY7^1hHcNV7FQH>i0MV7]OYHe0b0N[7oN[HS1:Of0`NU50RJa1309kN_6T1YI2IXOn6e0YI4_O@Y7;YI6TOHc72YIf1h6YNXIh1h6WNYIi1g6VNYIj1T7iMmHW2Y7bMhH]2_7\\MaHe2e7TM\\Hk2d7UM\\Hl2a7VM`Hi2_7XMbH`1nNWO]8\\OdH]1TOROW8BfH\\1UOPOS8EiHZ1VOoNS8EgH\\1WOnNS8EgH]1VOlNU8FfH]1WOjNV8GcH_1XOhNW8HbH`1WOfNY8IlHU1lNPOZ8J[Ie0]N^O[8KaI?TND^8K_I`0TNCb8I[Ic0UNAU;?kD_OW;`0jD^OX;a0i001N2O1N2O001N101N2O0O2O1O0O2O0O2O`f^4"}, "label": "a little girl holding a yellow bat"}]}
{"id": 455181, "image": "COCO_train2014_000000455181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection in a mirror of a tabby cat sitting on a cardboard box\"."}], "task": "refering", "answer_template": "The \"a reflection in a mirror of a tabby cat sitting on a cardboard box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272], "bbox": [0.583515625, 0.24354166666666668, 0.91325, 0.7133541666666666], "iscrowd": 0, "area": 21679.125799999998, "rle": {"size": [480, 640], "counts": "^c_54h>401O1O00100O001O1O001O10O01O001O1O001O10O01O1O001O001O01O01O001O00001O010O00001O001O01O01O00001O001O01O0001kMWOQFi0m99SEGl:W2N2O1N3M2O1N2O1N3N1N2N2O1N2O2M2O1N2O1N3M2O1O1O1O1N3N100O100O101N100O1O100O100O100O1O10O01000000O100N2O1O001O1N2O1O1O1O1O0O2O1O1O1O1O1N2O10O01O1O1O1O2N1O1O1O1O1O1O2O0O1O1O1O1O1O2N1O1O1O1O1O3N2MM1N3M3M3M2N3M3M3M5K6J6J6J7J5J6J6J6J6J6L3L5K5L4K5L3L4M3L4L4N2N2N2N2N2N1O2M3N2N2N2N2N2N2N2N2N1N3N2N2N2N2M3J6J6K5J5KTTj0"}, "label": "a reflection in a mirror of a tabby cat sitting on a cardboard box"}]}
{"id": 70161, "image": "COCO_train2014_000000070161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher of the game\"."}], "task": "refering", "answer_template": "The \"the catcher of the game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 227, 228, 229, 248, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 320, 321, 342, 343, 344], "bbox": [0.7935156250000001, 0.5640566037735849, 1.0, 0.9865094339622641], "iscrowd": 0, "area": 10895.670600000005, "rle": {"size": [424, 640], "counts": "\\eb6?T<e0J7O00001O00001O0000001O00000000O100O2O2N3L3N2M3N2M100O100000O0100000000000O01000000000O0100000O1000O10O10000oEVOd7j0[HWOe7i0ZHXOe7i0ZHXOf7h0YHYOg7g0XHZOh7f0WH[Oh7f0WH[Oi7e0WH[Oh7f0WH[Oh7f0WH[Oi7e0VH\\Om5MeIh0=\\On55\\I?e0]On5=TI7m0]Oo5>RI5n0^Oo5`0PI3P1^OP6a0nH1Q1_OP6d0kHNT1_OQ6g0fHKX1_OQ6j0dHG[1_OQ6n0_HD_1_OQ6Q1]HA`1_OS6T1XH^Oe1^OR6^2mIcMS6]2mIcMR6^2mIcMS6]2mIcMR6^2mIcMS6]2mIcMR6^2mIcMS6]2mIcMR6_2lIbMT6^2lIbMT6^2kIcMU6]2kIcMU6]2jIdMV6\\2jIdMV6]2hIdMX6\\2hIdMX6]2fIdMZ6\\2fIdMZ6]2dIdM\\6\\2dIdM\\6]2bIdM^6\\2bIdM^6]2`IdM`6\\2`IdM`6]2^IdMb6\\2^IdMb6]2\\IdMd6]2[IcMe6]2ZIdMf6]2YIcMg6]2XIdMh6]2WIcMj6]2UIcMk6^2TIbMm6]2RIdMn6]2QIcMP7]2oHcMQ7^2mHcMT7]2kHcMU7]2kHcMV7]2iHcMX7]2fHdMZ7]2dHdM]7]2`HdMa7^2[HcMf7^2WHcMk7^2QHcMQ8i31O5K5K5K2VOZH"}, "label": "the catcher of the game"}]}
{"id": 70161, "image": "COCO_train2014_000000070161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher\"."}], "task": "refering", "answer_template": "The \"a baseball catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 227, 228, 229, 248, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 320, 321, 342, 343, 344], "bbox": [0.7935156250000001, 0.5640566037735849, 1.0, 0.9865094339622641], "iscrowd": 0, "area": 10895.670600000005, "rle": {"size": [424, 640], "counts": "\\eb6?T<e0J7O00001O00001O0000001O00000000O100O2O2N3L3N2M3N2M100O100000O0100000000000O01000000000O0100000O1000O10O10000oEVOd7j0[HWOe7i0ZHXOe7i0ZHXOf7h0YHYOg7g0XHZOh7f0WH[Oh7f0WH[Oi7e0WH[Oh7f0WH[Oh7f0WH[Oi7e0VH\\Om5MeIh0=\\On55\\I?e0]On5=TI7m0]Oo5>RI5n0^Oo5`0PI3P1^OP6a0nH1Q1_OP6d0kHNT1_OQ6g0fHKX1_OQ6j0dHG[1_OQ6n0_HD_1_OQ6Q1]HA`1_OS6T1XH^Oe1^OR6^2mIcMS6]2mIcMR6^2mIcMS6]2mIcMR6^2mIcMS6]2mIcMR6^2mIcMS6]2mIcMR6_2lIbMT6^2lIbMT6^2kIcMU6]2kIcMU6]2jIdMV6\\2jIdMV6]2hIdMX6\\2hIdMX6]2fIdMZ6\\2fIdMZ6]2dIdM\\6\\2dIdM\\6]2bIdM^6\\2bIdM^6]2`IdM`6\\2`IdM`6]2^IdMb6\\2^IdMb6]2\\IdMd6]2[IcMe6]2ZIdMf6]2YIcMg6]2XIdMh6]2WIcMj6]2UIcMk6^2TIbMm6]2RIdMn6]2QIcMP7]2oHcMQ7^2mHcMT7]2kHcMU7]2kHcMV7]2iHcMX7]2fHdMZ7]2dHdM]7]2`HdMa7^2[HcMf7^2WHcMk7^2QHcMQ8i31O5K5K5K2VOZH"}, "label": "a baseball catcher"}]}
{"id": 70161, "image": "COCO_train2014_000000070161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging at a ball\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging at a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 80, 81, 82, 83, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 243, 244, 245, 246, 261, 262, 263, 267, 268, 269, 292], "bbox": [0.355125, 0.1261320754716981, 0.72815625, 0.8333254716981133], "iscrowd": 0, "area": 29513.52300000002, "rle": {"size": [424, 640], "counts": "mYn21V=2O1N2N2N1O100O1O1O1O1O100O001O1O1N2N2O1N2N2N1O2aLSOaJo0[5VObJl0Z5XOeJi0W5[OgJg0U5^OhJd0T5@jJb0R5BlJ`0Q5CmJ?Q5DlJ>S5CkJ?S5CkJ?S5DkJ=S5EkJ=T5DjJ>U5DhJ>W5CgJ?X5BfJ`0Y5AeJa0Z5_OeJc0Z5^OeJc0Z5^OdJd0[5]OcJe0\\5[OcJg0\\5ZOaJi0^5XO_Jk0`5VO^Jl0a5UO\\Jn0c5RO[JQ1d5POYJS1f5nNWJU1h5lNUJW1j5jNSJX1m5hNQJZ1o5gNoIZ1Q6gNlId0ROUNR7X1iId0P7\\OmHf0S7ZOlHg0T7ZOiHh0W7YOgHh0Y7YOdHh0]7YOaHh0_7YO^Hi0c7WOZHk0f7VOXHj0i7WOTHj0m7WOPHk0P8U2100O101N100O1O1N2O1N2O1N2M3J5L5O001O001O000010O1O1O1O1O1N2N2J6000i0XO4K5K5L4K5K5L4K5K3N1N1O2O1O0O2O001O1N101O1O0O2O001O0O2O001O0O2O001O001O001O001O001O001O010O01YNkJZKT5`4RK`Ko4X4YKhKf4T4^KlKc4Q4_KPL`4n3cKQL^4l3dKUL[4i3hKVLY4g3iKZLV4d3lK\\LU4a3nK_LR4^3PLbLP4\\3RLeLn3X3ULgLl3W3ULjLk3S3XLmLh3P3ZLPMf3o2[LRMe3k2^LTMc3j2^LWMb3f2`LZMa3c2bL]M]3`2fL`M\\3[2hLeMZ3U2iLkMY3P2jLQNX3h1mLWNV3c1mL^NT3[1RMdNQ3U1SMlNn2m0WMSOX39nLGY80010O0100O0010O00100O100O00100O010O00010O010O000100O1O1O1O1O1O2N101N3M2N3M4L2N0010O0001O000010O0001O01O1N2N2VNmE<`RX2"}, "label": "a baseball player swinging at a ball"}]}
{"id": 70161, "image": "COCO_train2014_000000070161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the bat\"."}], "task": "refering", "answer_template": "The \"the man with the bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 80, 81, 82, 83, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 243, 244, 245, 246, 261, 262, 263, 267, 268, 269, 292], "bbox": [0.355125, 0.1261320754716981, 0.72815625, 0.8333254716981133], "iscrowd": 0, "area": 29513.52300000002, "rle": {"size": [424, 640], "counts": "mYn21V=2O1N2N2N1O100O1O1O1O1O100O001O1O1N2N2O1N2N2N1O2aLSOaJo0[5VObJl0Z5XOeJi0W5[OgJg0U5^OhJd0T5@jJb0R5BlJ`0Q5CmJ?Q5DlJ>S5CkJ?S5CkJ?S5DkJ=S5EkJ=T5DjJ>U5DhJ>W5CgJ?X5BfJ`0Y5AeJa0Z5_OeJc0Z5^OeJc0Z5^OdJd0[5]OcJe0\\5[OcJg0\\5ZOaJi0^5XO_Jk0`5VO^Jl0a5UO\\Jn0c5RO[JQ1d5POYJS1f5nNWJU1h5lNUJW1j5jNSJX1m5hNQJZ1o5gNoIZ1Q6gNlId0ROUNR7X1iId0P7\\OmHf0S7ZOlHg0T7ZOiHh0W7YOgHh0Y7YOdHh0]7YOaHh0_7YO^Hi0c7WOZHk0f7VOXHj0i7WOTHj0m7WOPHk0P8U2100O101N100O1O1N2O1N2O1N2M3J5L5O001O001O000010O1O1O1O1O1N2N2J6000i0XO4K5K5L4K5K5L4K5K3N1N1O2O1O0O2O001O1N101O1O0O2O001O0O2O001O0O2O001O001O001O001O001O001O010O01YNkJZKT5`4RK`Ko4X4YKhKf4T4^KlKc4Q4_KPL`4n3cKQL^4l3dKUL[4i3hKVLY4g3iKZLV4d3lK\\LU4a3nK_LR4^3PLbLP4\\3RLeLn3X3ULgLl3W3ULjLk3S3XLmLh3P3ZLPMf3o2[LRMe3k2^LTMc3j2^LWMb3f2`LZMa3c2bL]M]3`2fL`M\\3[2hLeMZ3U2iLkMY3P2jLQNX3h1mLWNV3c1mL^NT3[1RMdNQ3U1SMlNn2m0WMSOX39nLGY80010O0100O0010O00100O100O00100O010O00010O010O000100O1O1O1O1O1O2N101N3M2N3M4L2N0010O0001O000010O0001O01O1N2N2VNmE<`RX2"}, "label": "the man with the bat"}]}
{"id": 193042, "image": "COCO_train2014_000000193042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch sitting in directly in front of the window\"."}], "task": "refering", "answer_template": "The \"the couch sitting in directly in front of the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236], "bbox": [0.090109375, 0.46189695550351284, 0.440171875, 0.7268384074941452], "iscrowd": 0, "area": 17567.36025, "rle": {"size": [427, 640], "counts": "m\\h02Y=1N2O2N1N2O2M2O1N2O2M2O1O2jDEX9<ZF6a9KWF?f9BRFe0n9]OjEi0V:XObEo0]:o01000BWMUFn2k99O010O100O0100000O10O1000O10O1000O1000O101O000O100000001N1000000O10001O000O1000000O2O0000000O10001O0O100000000O2O00000O100000000O1000000O100000000O1000000O100000O10O10O1000O10O100O10O10O100000O1000000O100000000O1000000O100000000O1000000O1000001O0O1000000O100000000O1000000O100000000O1000000O100000000O1000O10O10O1002N2M3N2N2N2M3N2N00000O10001N10000O1000000O2O000O10001N2O1O001N2O1O1N101O1N2O1O001N2Oi0WOUQe4"}, "label": "the couch sitting in directly in front of the window"}]}
{"id": 193042, "image": "COCO_train2014_000000193042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown couch in the back near the glass windows\"."}], "task": "refering", "answer_template": "The \"brown couch in the back near the glass windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236], "bbox": [0.090109375, 0.46189695550351284, 0.440171875, 0.7268384074941452], "iscrowd": 0, "area": 17567.36025, "rle": {"size": [427, 640], "counts": "m\\h02Y=1N2O2N1N2O2M2O1N2O2M2O1O2jDEX9<ZF6a9KWF?f9BRFe0n9]OjEi0V:XObEo0]:o01000BWMUFn2k99O010O100O0100000O10O1000O10O1000O1000O101O000O100000001N1000000O10001O000O1000000O2O0000000O10001O0O100000000O2O00000O100000000O1000000O100000000O1000000O100000O10O10O1000O10O100O10O10O100000O1000000O100000000O1000000O100000000O1000000O1000001O0O1000000O100000000O1000000O100000000O1000000O100000000O1000O10O10O1002N2M3N2N2N2M3N2N00000O10001N10000O1000000O2O000O10001N2O1O001N2O1O1N101O1N2O1O001N2Oi0WOUQe4"}, "label": "brown couch in the back near the glass windows"}]}
{"id": 193042, "image": "COCO_train2014_000000193042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch on the floor in front of a wall with no windows\"."}], "task": "refering", "answer_template": "The \"a couch on the floor in front of a wall with no windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 269, 270, 271], "bbox": [0.5064375, 0.47601873536299766, 0.847234375, 0.7858313817330209], "iscrowd": 0, "area": 16622.069350000005, "rle": {"size": [427, 640], "counts": "Z[W42S=6I7J6J6M30000O100IROQDn0o;TOnCm0R<TOmCl0S<7O0O101O0000001O0O101O00001O00001O00001O0000001O00001O01O100O2N10L3I7K5N3N1O101O00001O01O01O00010O0001O01O00010O000010O0001O010O0010O01[EQNl9P2jEYNV:_20O0001O01O01O00001O001O01O01O00001O01O01O00001O00001O00001N100012M1O010O01O01O01O01O010O00010O0001O00001O00010O00001O001O00010O00001O01O01O00010O000010O0001O00001O01O01O001O00001O000O2OO100O1000O01O1O1O1O100O2N1O1O1O1O1O1O1M3H8H8I8POiDB`;<jDSOg0Nc9n0eF^OX9G`Eb0o0Om9MkE;V:EbEb0_:]OZEk0f:UO[Ei0f:VO[Ei0g:VOYEi0h:WOYEh0g:WOZEh0g:XOYEg0i:WOXEd0o:ZOQE=Y;AhD6c;H^DNm;0SDLR<2f0NhcX1"}, "label": "a couch on the floor in front of a wall with no windows"}]}
{"id": 193042, "image": "COCO_train2014_000000193042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch that is between the tree and the end table\"."}], "task": "refering", "answer_template": "The \"the couch that is between the tree and the end table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 269, 270, 271], "bbox": [0.5064375, 0.47601873536299766, 0.847234375, 0.7858313817330209], "iscrowd": 0, "area": 16622.069350000005, "rle": {"size": [427, 640], "counts": "Z[W42S=6I7J6J6M30000O100IROQDn0o;TOnCm0R<TOmCl0S<7O0O101O0000001O0O101O00001O00001O00001O0000001O00001O01O100O2N10L3I7K5N3N1O101O00001O01O01O00010O0001O01O00010O000010O0001O010O0010O01[EQNl9P2jEYNV:_20O0001O01O01O00001O001O01O01O00001O01O01O00001O00001O00001N100012M1O010O01O01O01O01O010O00010O0001O00001O00010O00001O001O00010O00001O01O01O00010O000010O0001O00001O01O01O001O00001O000O2OO100O1000O01O1O1O1O100O2N1O1O1O1O1O1O1M3H8H8I8POiDB`;<jDSOg0Nc9n0eF^OX9G`Eb0o0Om9MkE;V:EbEb0_:]OZEk0f:UO[Ei0f:VO[Ei0g:VOYEi0h:WOYEh0g:WOZEh0g:XOYEg0i:WOXEd0o:ZOQE=Y;AhD6c;H^DNm;0SDLR<2f0NhcX1"}, "label": "the couch that is between the tree and the end table"}]}
{"id": 406034, "image": "COCO_train2014_000000406034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table behind the little girl\"."}], "task": "refering", "answer_template": "The \"the table behind the little girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 105, 106, 107, 108, 121, 122, 123, 137, 138, 152, 153, 154, 167, 168, 169, 182, 183, 184, 197, 198, 199, 212, 213, 227, 228], "bbox": [0.0, 0.29146875, 0.2941217798594848, 0.69209375], "iscrowd": 0, "area": 13248.6092, "rle": {"size": [640, 427], "counts": "R6`0_c0100O1000000000000O10000000O1000000000000000000000000000O1000000000O10000000000000000000000000O100000004L9G:F9G9G9G9G:F9G9G9G:F9G9G9G:E:G9G9G9G:F9G9G9G:F8H1O00N2O1O1N2O1O1N101O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N:G:F:E<E:F:E<E:F:F:E<E:F:E<E:F:E;F;E:E;F;EThk5"}, "label": "the table behind the little girl"}]}
{"id": 406034, "image": "COCO_train2014_000000406034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tall table no one is seated at\"."}], "task": "refering", "answer_template": "The \"the tall table no one is seated at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 105, 106, 107, 108, 121, 122, 123, 137, 138, 152, 153, 154, 167, 168, 169, 182, 183, 184, 197, 198, 199, 212, 213, 227, 228], "bbox": [0.0, 0.29146875, 0.2941217798594848, 0.69209375], "iscrowd": 0, "area": 13248.6092, "rle": {"size": [640, 427], "counts": "R6`0_c0100O1000000000000O10000000O1000000000000000000000000000O1000000000O10000000000000000000000000O100000004L9G:F9G9G9G9G:F9G9G9G:F9G9G9G:E:G9G9G9G:F9G9G9G:F8H1O00N2O1O1N2O1O1N101O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N:G:F:E<E:F:E<E:F:F:E<E:F:E<E:F:E;F;E:E;F;EThk5"}, "label": "the tall table no one is seated at"}]}
{"id": 406034, "image": "COCO_train2014_000000406034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden table with a food tray on it that hasa donut on it\"."}], "task": "refering", "answer_template": "The \"wooden table with a food tray on it that hasa donut on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.010070257611241218, 0.6927187499999999, 0.9947306791569087, 0.9864374999999999], "iscrowd": 0, "area": 71861.00085000001, "rle": {"size": [640, 427], "counts": "in2<cc0a3_LS1lN2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O0000000O1000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000000000000O2O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000000O1000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O0O1000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O3MT1lNS1mNT1lNS1mNhd0"}, "label": "wooden table with a food tray on it that hasa donut on it"}]}
{"id": 406034, "image": "COCO_train2014_000000406034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a dunkin donuts tray on it and a woman sitting at it\"."}], "task": "refering", "answer_template": "The \"a table with a dunkin donuts tray on it and a woman sitting at it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.010070257611241218, 0.6927187499999999, 0.9947306791569087, 0.9864374999999999], "iscrowd": 0, "area": 71861.00085000001, "rle": {"size": [640, 427], "counts": "in2<cc0a3_LS1lN2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O0000000O1000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000000000000O2O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000000O1000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O0O1000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O3MT1lNS1mNT1lNS1mNhd0"}, "label": "a table with a dunkin donuts tray on it and a woman sitting at it"}]}
{"id": 86549, "image": "COCO_train2014_000000086549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player , wearing a white and blue uniform , that is getting ready to hit a ball\"."}], "task": "refering", "answer_template": "The \"a baseball player , wearing a white and blue uniform , that is getting ready to hit a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 122, 123, 124, 125, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 353, 354, 357, 358], "bbox": [0.24117187499999998, 0.2342247191011236, 0.617015625, 1.0], "iscrowd": 0, "area": 35688.148050000025, "rle": {"size": [445, 640], "counts": "jbS24h=3L4M3M3L2O1N2O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O2N2O010O1O1O010O100O2O0O1O1O1O2N1009G00O1N2O0O2O1N2N101N2O1N1O2O1N2O1N1O2SMbNYI_1b6kNWIW1d6nN[IR1e6oNYIS1e6oNYIS1e6oNYIS1f6nNXIT1f6nNXIS1g6nNYIS1c2dMGZ1dMT1b2eMHX1eMU1_2hMIT1gMU1^2jMJP1hMX1Z2mMKl0jMY1X2nMLi0lM[1T2QNNc0nM]1R2SNN`0PN_1o1VNN:SNb1k1YN04UNd1i1]NONXNf1g1^N1KcMgNHP3b2aN3GcMjNFo2b2bN5DcMmNDn2b2cN7AcMPOAn2c2cN9^ObMTO@k2b2gN<YObMWO^Oj2a2iN?UObMZO[Oj2a2jNb0QObM]OYOi2`2lNf0lNaMl2g1ZNi0iN`Mn2d1\\Nl0eN`Mn2b1aNn0`N`Mn2a1eNP1[N_Mo2`1iNQ1WN_Mo2_1mNR1SN_Mo2^1QOS1oM_Mo2]1UOT1kM_Mo2[1ZOW1eM^MQ3[1[OW1cM^MR3Z1]OX1`M]MS3[1_OY1\\M\\MU3[1@Z1YM[MW3Z1B\\1UMZMY3Y1D^1QMYMZ3Z1F^1nLXM^3W1Fa1kLXMa3U1Ed1hLWMe3R1Eh1dLVMj3n0Dm1`LUMn3l0BQ2^LSMR4i0Bn2`0oLBQ3?mLBR3`0kLBU3?iLBV3`0gLnNSORMV4R4cLiN@QMP4T4]LmNFlLQ4T4VLQONgLn3W4QLTOk4n0oJSOT5o0fJSO[5P1_JROa5d4100O2O000O100O2N1O1O1O1N2O2N1O1O2N2N2N1O2N2N2N1N3N2N2N1O2N2N2N2N3M2N3L4M2N3M3M3M2=DO1O001O1O1N110O1O001O100N101N2N101N2N2O0011O0O10001N1000001N10000O2O000O10O1N2N1eMSHdNo7W1VHgNl7U1XHiNi7T1[HjNg7Q1^HmNd7o0aHnN`7o0fHmN\\7n0kHnNW7n0nHoNT7m0RIoNo6n0WInNk6n0[InNg6n0_InNb6o0cInN_6n0gInN[6n0kInNV6o0oInNS6o0QJnNQ6o0TJmNn5P1VJmNk5Q1YJlNi5Q1[JlNg5Q1\\JmNe5Q1f3M3M2N3M3M3M4L6J6JTUZ3"}, "label": "a baseball player , wearing a white and blue uniform , that is getting ready to hit a ball"}]}
{"id": 86549, "image": "COCO_train2014_000000086549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player getting ready to swing the bat\"."}], "task": "refering", "answer_template": "The \"a baseball player getting ready to swing the bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 122, 123, 124, 125, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 353, 354, 357, 358], "bbox": [0.24117187499999998, 0.2342247191011236, 0.617015625, 1.0], "iscrowd": 0, "area": 35688.148050000025, "rle": {"size": [445, 640], "counts": "jbS24h=3L4M3M3L2O1N2O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O2N2O010O1O1O010O100O2O0O1O1O1O2N1009G00O1N2O0O2O1N2N101N2O1N1O2O1N2O1N1O2SMbNYI_1b6kNWIW1d6nN[IR1e6oNYIS1e6oNYIS1e6oNYIS1f6nNXIT1f6nNXIS1g6nNYIS1c2dMGZ1dMT1b2eMHX1eMU1_2hMIT1gMU1^2jMJP1hMX1Z2mMKl0jMY1X2nMLi0lM[1T2QNNc0nM]1R2SNN`0PN_1o1VNN:SNb1k1YN04UNd1i1]NONXNf1g1^N1KcMgNHP3b2aN3GcMjNFo2b2bN5DcMmNDn2b2cN7AcMPOAn2c2cN9^ObMTO@k2b2gN<YObMWO^Oj2a2iN?UObMZO[Oj2a2jNb0QObM]OYOi2`2lNf0lNaMl2g1ZNi0iN`Mn2d1\\Nl0eN`Mn2b1aNn0`N`Mn2a1eNP1[N_Mo2`1iNQ1WN_Mo2_1mNR1SN_Mo2^1QOS1oM_Mo2]1UOT1kM_Mo2[1ZOW1eM^MQ3[1[OW1cM^MR3Z1]OX1`M]MS3[1_OY1\\M\\MU3[1@Z1YM[MW3Z1B\\1UMZMY3Y1D^1QMYMZ3Z1F^1nLXM^3W1Fa1kLXMa3U1Ed1hLWMe3R1Eh1dLVMj3n0Dm1`LUMn3l0BQ2^LSMR4i0Bn2`0oLBQ3?mLBR3`0kLBU3?iLBV3`0gLnNSORMV4R4cLiN@QMP4T4]LmNFlLQ4T4VLQONgLn3W4QLTOk4n0oJSOT5o0fJSO[5P1_JROa5d4100O2O000O100O2N1O1O1O1N2O2N1O1O2N2N2N1O2N2N2N1N3N2N2N1O2N2N2N2N3M2N3L4M2N3M3M3M2=DO1O001O1O1N110O1O001O100N101N2N101N2N2O0011O0O10001N1000001N10000O2O000O10O1N2N1eMSHdNo7W1VHgNl7U1XHiNi7T1[HjNg7Q1^HmNd7o0aHnN`7o0fHmN\\7n0kHnNW7n0nHoNT7m0RIoNo6n0WInNk6n0[InNg6n0_InNb6o0cInN_6n0gInN[6n0kInNV6o0oInNS6o0QJnNQ6o0TJmNn5P1VJmNk5Q1YJlNi5Q1[JlNg5Q1\\JmNe5Q1f3M3M2N3M3M3M4L6J6JTUZ3"}, "label": "a baseball player getting ready to swing the bat"}]}
{"id": 86549, "image": "COCO_train2014_000000086549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shortstop\"."}], "task": "refering", "answer_template": "The \"shortstop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 163, 164, 165, 185, 186, 187, 188, 209, 210, 211, 231, 232, 233, 234, 254, 255, 257, 258, 277, 280, 281, 300], "bbox": [0.051453125, 0.3657303370786517, 0.24787499999999998, 0.8276179775280899], "iscrowd": 0, "area": 11173.3799, "rle": {"size": [445, 640], "counts": "ne>3g=6I6J6J5L4K5K5K5L3L3M4M2M3M2N101TNSNhGo1Y8UN`Gm1`8XNYGi1h8ZNTGg1l8\\NoFe1S9\\NiFe1X9^NcFc1Y9eNcF[1\\9iN`FX1a9kNZFV1h9lNSFV1m9lNPFU1Q:Z1001O1O1O1O1O1O1O2N1O100000000O1000001O0001O00001H7O1NImEjLQ:Y3oEeLP:]3RF`Lm9c3TF[Ll9f3510O0100O010O1O10O0100O10O01O2O1N2O13M6K5JO2O1N2N2N2N2N2M3N2N2N2NiMaF1]9KjF4T9HRG8k8E\\G:b8BdG=_8\\OfGd0[8UOkGj0V8POPHm0R8oNSHn0P8lNVHQ1m7iNYHT1i7hN\\HV1X:M3M3M3N2M3M3M3M3N2M1O1O2N100O1O1O1N2O5JgSa6"}, "label": "shortstop"}]}
{"id": 86549, "image": "COCO_train2014_000000086549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball fielder in a dark shirt\"."}], "task": "refering", "answer_template": "The \"a baseball fielder in a dark shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 163, 164, 165, 185, 186, 187, 188, 209, 210, 211, 231, 232, 233, 234, 254, 255, 257, 258, 277, 280, 281, 300], "bbox": [0.051453125, 0.3657303370786517, 0.24787499999999998, 0.8276179775280899], "iscrowd": 0, "area": 11173.3799, "rle": {"size": [445, 640], "counts": "ne>3g=6I6J6J5L4K5K5K5L3L3M4M2M3M2N101TNSNhGo1Y8UN`Gm1`8XNYGi1h8ZNTGg1l8\\NoFe1S9\\NiFe1X9^NcFc1Y9eNcF[1\\9iN`FX1a9kNZFV1h9lNSFV1m9lNPFU1Q:Z1001O1O1O1O1O1O1O2N1O100000000O1000001O0001O00001H7O1NImEjLQ:Y3oEeLP:]3RF`Lm9c3TF[Ll9f3510O0100O010O1O10O0100O10O01O2O1N2O13M6K5JO2O1N2N2N2N2N2M3N2N2N2NiMaF1]9KjF4T9HRG8k8E\\G:b8BdG=_8\\OfGd0[8UOkGj0V8POPHm0R8oNSHn0P8lNVHQ1m7iNYHT1i7hN\\HV1X:M3M3M3N2M3M3M3M3N2M1O1O2N100O1O1O1N2O5JgSa6"}, "label": "a baseball fielder in a dark shirt"}]}
{"id": 365082, "image": "COCO_train2014_000000365082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of pizza on the plate\"."}], "task": "refering", "answer_template": "The \"the slice of pizza on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 323, 324, 325, 326, 327, 328, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5449414519906323, 0.88890625, 1.0, 1.0], "iscrowd": 0, "area": 8106.420300000002, "rle": {"size": [640, 427], "counts": "RVb43kc05L3N3M3N1N3M2N3M3M101M3I6M4N2O1N1O2O1N1O2N2O0O2N2O0O2N2N2O0O0001O00001O0000001O00001O0000001O00001O00001O00001O00001O00001OO100O1O1001O1O001O1O1O001O1O1O001O1O1O000000O10000O11O2N1O1O1O1O1O1O2N000000000000000000000000000000001O000000001O001O001O001O001O00000000O100000000O100000000O10000000000001O001O00001O001O001O00001O0000001O00000000001O00000000001O00000000001O001O1O001O1O001O1O00"}, "label": "the slice of pizza on the plate"}]}
{"id": 365082, "image": "COCO_train2014_000000365082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of pizza in a plate on a dinning table\"."}], "task": "refering", "answer_template": "The \"a piece of pizza in a plate on a dinning table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 323, 324, 325, 326, 327, 328, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5449414519906323, 0.88890625, 1.0, 1.0], "iscrowd": 0, "area": 8106.420300000002, "rle": {"size": [640, 427], "counts": "RVb43kc05L3N3M3N1N3M2N3M3M101M3I6M4N2O1N1O2O1N1O2N2O0O2N2O0O2N2N2O0O0001O00001O0000001O00001O0000001O00001O00001O00001O00001O00001OO100O1O1001O1O001O1O1O001O1O1O001O1O1O000000O10000O11O2N1O1O1O1O1O1O2N000000000000000000000000000000001O000000001O001O001O001O001O00000000O100000000O100000000O10000000000001O001O00001O001O001O00001O0000001O00000000001O00000000001O00000000001O001O1O001O1O001O1O00"}, "label": "a piece of pizza in a plate on a dinning table"}]}
{"id": 127515, "image": "COCO_train2014_000000127515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball umpire is crouching behind the catcher\"."}], "task": "refering", "answer_template": "The \"the baseball umpire is crouching behind the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 91, 92, 93, 108, 109, 110, 111, 112, 126, 127, 128, 129, 145, 146, 147, 163, 164, 165], "bbox": [0.04118, 0.3778978978978979, 0.26783999999999997, 0.8066966966966966], "iscrowd": 0, "area": 7018.069750000001, "rle": {"size": [333, 500], "counts": "]P7?l94L3N2M3N2M3M3lGROd6R1VISOg6P1VISOg6Q1VIROg6P1XIQOf6R1WIPOh6Q1TISOk6o0RISOl6P1QIROn6o0QIROm6Q1PIQOn6Q1PIQOo6P1oHQOP7Q1nHQOQ7P1mHQOR7V2N101OFQIXMo6h2RIWMn6i2RIWMn6i2RIWMm6j2TIUMl6k2TIUMl6j2UIVMk6j2VIUMj6k2VIUMj6k2XIQMj6o2:000ZIRMm5n2QJUMl5g2eITM<9j5b2^JaM_5_1jIPOi0D[5Y1QJoNe0KY5U1UJlNb02X5R1WJmN=4[5n0[JmN78^5j0\\JnN2;b5f0^JoNM>d5d0_JnNJa0g5`0aJoNDd0j5>bJnN@h0n59cJ>]5CbJ=^5DbJ;^5EbJ;^5FaJ9`5H_J8a5mNPJL?V1a5mNRJM<V1b5kNUJN9d0GoNk5>XJN5e0HoNk5=ZJO3d0HPOk5<\\J00d0IPOl5:]J2Md0JoNm5:_J2Ie0KoNm59aJ3Gd0KPOm57dJ5Dc0KQOo54dJ6Ce0JPOQ62fJ6Ag0HQOS6MhJ:^Og0GROT6IkJ<\\Oh0GQO[76oHa01UOR79nH9;[Oh6;nH1[8NR101O010O001O0000000000000O1000000O01000O1000O0100000O0100O010O001O001N2O1O0O[kf3"}, "label": "the baseball umpire is crouching behind the catcher"}]}
{"id": 127515, "image": "COCO_train2014_000000127515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umpire behind the catcher\"."}], "task": "refering", "answer_template": "The \"umpire behind the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 91, 92, 93, 108, 109, 110, 111, 112, 126, 127, 128, 129, 145, 146, 147, 163, 164, 165], "bbox": [0.04118, 0.3778978978978979, 0.26783999999999997, 0.8066966966966966], "iscrowd": 0, "area": 7018.069750000001, "rle": {"size": [333, 500], "counts": "]P7?l94L3N2M3N2M3M3lGROd6R1VISOg6P1VISOg6Q1VIROg6P1XIQOf6R1WIPOh6Q1TISOk6o0RISOl6P1QIROn6o0QIROm6Q1PIQOn6Q1PIQOo6P1oHQOP7Q1nHQOQ7P1mHQOR7V2N101OFQIXMo6h2RIWMn6i2RIWMn6i2RIWMm6j2TIUMl6k2TIUMl6j2UIVMk6j2VIUMj6k2VIUMj6k2XIQMj6o2:000ZIRMm5n2QJUMl5g2eITM<9j5b2^JaM_5_1jIPOi0D[5Y1QJoNe0KY5U1UJlNb02X5R1WJmN=4[5n0[JmN78^5j0\\JnN2;b5f0^JoNM>d5d0_JnNJa0g5`0aJoNDd0j5>bJnN@h0n59cJ>]5CbJ=^5DbJ;^5EbJ;^5FaJ9`5H_J8a5mNPJL?V1a5mNRJM<V1b5kNUJN9d0GoNk5>XJN5e0HoNk5=ZJO3d0HPOk5<\\J00d0IPOl5:]J2Md0JoNm5:_J2Ie0KoNm59aJ3Gd0KPOm57dJ5Dc0KQOo54dJ6Ce0JPOQ62fJ6Ag0HQOS6MhJ:^Og0GROT6IkJ<\\Oh0GQO[76oHa01UOR79nH9;[Oh6;nH1[8NR101O010O001O0000000000000O1000000O01000O1000O0100000O0100O010O001O001N2O1O0O[kf3"}, "label": "umpire behind the catcher"}]}
{"id": 127515, "image": "COCO_train2014_000000127515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man between the referee and batter\"."}], "task": "refering", "answer_template": "The \"the man between the referee and batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 112, 113, 114, 115, 116, 130, 131, 132, 147, 148, 149, 166, 167], "bbox": [0.21104, 0.42336336336336333, 0.4636, 0.7766366366366366], "iscrowd": 0, "area": 6462.419400000001, "rle": {"size": [333, 500], "counts": "WfR13W:3N3M3L3N3M3L3M4L4K5La0_O4M2N2O1N2N2O1N2O2M2O1N2O1N2O1O2M2O1N2O1E;M3L4L5N100000000O101O00000O100000001NI8I7N2O2M3M3M3N2M3M2N3N2M3M3M3N2J5VOmGRO\\8m0c000O1O1O<D1O1O1O00000000000000000001O0000000000000000000O10000000000000000000000001O00000000001O00000O1O2O0O4M7HTof2"}, "label": "the man between the referee and batter"}]}
{"id": 127515, "image": "COCO_train2014_000000127515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher in front of a baseball umpire\"."}], "task": "refering", "answer_template": "The \"a baseball catcher in front of a baseball umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 112, 113, 114, 115, 116, 130, 131, 132, 147, 148, 149, 166, 167], "bbox": [0.21104, 0.42336336336336333, 0.4636, 0.7766366366366366], "iscrowd": 0, "area": 6462.419400000001, "rle": {"size": [333, 500], "counts": "WfR13W:3N3M3L3N3M3L3M4L4K5La0_O4M2N2O1N2N2O1N2O2M2O1N2O1N2O1O2M2O1N2O1E;M3L4L5N100000000O101O00000O100000001NI8I7N2O2M3M3M3N2M3M2N3N2M3M3M3N2J5VOmGRO\\8m0c000O1O1O<D1O1O1O00000000000000000001O0000000000000000000O10000000000000000000000001O00000000001O00000O1O2O0O4M7HTof2"}, "label": "a baseball catcher in front of a baseball umpire"}]}
{"id": 127515, "image": "COCO_train2014_000000127515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 81, 82, 83, 99, 100, 117, 118, 119, 134, 135, 136, 137, 138, 152], "bbox": [0.4415, 0.2471771771771772, 0.7019200000000001, 0.6988888888888889], "iscrowd": 0, "area": 6497.955499999998, "rle": {"size": [333, 500], "counts": "\\RX26V:2N2O001N2O001N2O1O1O2M2O2N1N3NCeFG\\99hFCW9==000O0100O01mLDUL<j3MnK2R44hKMW49dKF]4?]KBb4g0UKYOk4Q1lJoNS5Z1dJgN[5b1]J]Nc5f1ZJ[Nc5i1[JWNd5l1[JTNd5n1ZJSNf5n1YJRNf5P2TJTNm5m1nIWNQ6l1hIYNX6h1cI\\N\\6j20100O10O0100000O10000O100O010O2K4L5K4L5L3L6M2O2O1O1O5QN]Hi0b7VOkH>U7BVI4i6KYI5f6K[I5d6J_I5`6KaI5^6JeI5Z6KgI5X6JjI5V6JmI5R6KoI5P6JRJ6m5JUJ5j5JXJ6g5I\\J6c5I_J7`5IaJ6`5HbJ8]5GeJ9Z5FiJ9W5EkJ;T5DnJ;R5DPK<o4CSK<`700O101O0O100O11O4L1N10O1O001O1O1O001O1O001O1O001O1O1O1O1O1O1O1O1N2N1OlW`1"}, "label": "a baseball player swinging a bat"}]}
{"id": 127515, "image": "COCO_train2014_000000127515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player swinging the bat\"."}], "task": "refering", "answer_template": "The \"the player swinging the bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 81, 82, 83, 99, 100, 117, 118, 119, 134, 135, 136, 137, 138, 152], "bbox": [0.4415, 0.2471771771771772, 0.7019200000000001, 0.6988888888888889], "iscrowd": 0, "area": 6497.955499999998, "rle": {"size": [333, 500], "counts": "\\RX26V:2N2O001N2O001N2O1O1O2M2O2N1N3NCeFG\\99hFCW9==000O0100O01mLDUL<j3MnK2R44hKMW49dKF]4?]KBb4g0UKYOk4Q1lJoNS5Z1dJgN[5b1]J]Nc5f1ZJ[Nc5i1[JWNd5l1[JTNd5n1ZJSNf5n1YJRNf5P2TJTNm5m1nIWNQ6l1hIYNX6h1cI\\N\\6j20100O10O0100000O10000O100O010O2K4L5K4L5L3L6M2O2O1O1O5QN]Hi0b7VOkH>U7BVI4i6KYI5f6K[I5d6J_I5`6KaI5^6JeI5Z6KgI5X6JjI5V6JmI5R6KoI5P6JRJ6m5JUJ5j5JXJ6g5I\\J6c5I_J7`5IaJ6`5HbJ8]5GeJ9Z5FiJ9W5EkJ;T5DnJ;R5DPK<o4CSK<`700O101O0O100O11O4L1N10O1O001O1O1O001O1O001O1O001O1O1O1O1O1O1O1O1N2N1OlW`1"}, "label": "the player swinging the bat"}]}
{"id": 160291, "image": "COCO_train2014_000000160291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of brown skiis attached to someone ' s feet\"."}], "task": "refering", "answer_template": "The \"a pair of brown skiis attached to someone ' s feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [358, 359, 361, 382, 383, 384, 385, 405, 406, 407, 408, 409, 429, 430, 431, 432, 433, 452, 453, 455, 456, 476, 477], "bbox": [0.592609375, 0.7146061643835616, 0.871484375, 0.9865239726027397], "iscrowd": 0, "area": 6889.9359500000055, "rle": {"size": [584, 640], "counts": "mah61Wb02M2O2N2M3N2N2M3N2M201O1O1O1O1O1O1O0O2O1O1O1O1O001O1O1O1O0010000O010000000O0100000000O0100O10O0100O100O00101N100O2O0O1O101WN^NgBc1W=^NjBb1T=`NkBb1R=_NoBa1P=`NPC`1n<bNRC^1l<cNVC]1g<eNYC[1f<fN[CY1c<hN^CX1`<jNaCV1\\<kNeCT1Z<nNgCo0Z<ROfCm0Y<TOhCk0W<WOiCh0V<ZOjCe0U<\\OlCc0T<^OkCb0T<_OmC`0S<AmC>S<CmC<T<ClC=T<DlC;U<EkC:U<FkC9W<GiC8W<IiC5Y<JhC3Z<NeC1]<NdCO^<2bCL_<4bCI`<8_CGb<9_CDd<;]CCd<>\\C_Of<a0ZC^Og<b0ZC[Oh<e0YCYOi<g0VCWOl<i0UCUOl<k0UCROm<o0W2O0001O00010O0O10001O00001O00001N10000000000000000O1000000000000000O10O10000000000000000001O00000O1O2M2O1O1O1N3N1O1O1N3N1O1O1N2O2N1O1N2O1O2M2O1O1O`W^1"}, "label": "a pair of brown skiis attached to someone ' s feet"}]}
{"id": 160291, "image": "COCO_train2014_000000160291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of skis on a smiling lady\"."}], "task": "refering", "answer_template": "The \"a pair of skis on a smiling lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [358, 359, 361, 382, 383, 384, 385, 405, 406, 407, 408, 409, 429, 430, 431, 432, 433, 452, 453, 455, 456, 476, 477], "bbox": [0.592609375, 0.7146061643835616, 0.871484375, 0.9865239726027397], "iscrowd": 0, "area": 6889.9359500000055, "rle": {"size": [584, 640], "counts": "mah61Wb02M2O2N2M3N2N2M3N2M201O1O1O1O1O1O1O0O2O1O1O1O1O001O1O1O1O0010000O010000000O0100000000O0100O10O0100O100O00101N100O2O0O1O101WN^NgBc1W=^NjBb1T=`NkBb1R=_NoBa1P=`NPC`1n<bNRC^1l<cNVC]1g<eNYC[1f<fN[CY1c<hN^CX1`<jNaCV1\\<kNeCT1Z<nNgCo0Z<ROfCm0Y<TOhCk0W<WOiCh0V<ZOjCe0U<\\OlCc0T<^OkCb0T<_OmC`0S<AmC>S<CmC<T<ClC=T<DlC;U<EkC:U<FkC9W<GiC8W<IiC5Y<JhC3Z<NeC1]<NdCO^<2bCL_<4bCI`<8_CGb<9_CDd<;]CCd<>\\C_Of<a0ZC^Og<b0ZC[Oh<e0YCYOi<g0VCWOl<i0UCUOl<k0UCROm<o0W2O0001O00010O0O10001O00001O00001N10000000000000000O1000000000000000O10O10000000000000000001O00000O1O2M2O1O1O1N3N1O1O1N3N1O1O1N2O2N1O1N2O1O2M2O1O1O`W^1"}, "label": "a pair of skis on a smiling lady"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player wearing a red top and white shorts\"."}], "task": "refering", "answer_template": "The \"a tennis player wearing a red top and white shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 127, 142, 143, 144, 145, 146, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 308, 327, 328, 329, 330, 350, 351, 352], "bbox": [0.18884375, 0.05310869565217391, 0.548828125, 1.0], "iscrowd": 0, "area": 42875.38720000001, "rle": {"size": [460, 640], "counts": "Sdf1f0a=>B=B>C=B>A?B>A=D9ZOf0H9H9G8G:G8G:G8G5L4L4K5L4TKZHm2j7SMWHe2Q8ZMPH_2W8_MkG^2X8aMiG]2Y8aMiG]2Y8bMiGZ2Z8eMgGY2[8ZL\\GU1;^2\\8[L[GV1:]2c8aM_Ge0FFo8C\\Gc0KFk8E\\Gb0MEj8HZG?2Eg8JXG>6Ce8MWG<9Dc8NUG;=Ba81TG9a0B]84SG8d0A[85RG9e0@Z85SG;e0^OX86TG;g0\\OW86TG>g0ZOU87UG>h0YO`8KjFl0h0UOe9j0o100O100O100O100O1001O00000000000000001O000000000000000N2K6K4L4N2N2M3N2O101N1O2N2N2N2O0O2N2N2N2O1N1O2QFhMg7Z2WHhMg7Z2XHhMo1b0o0g1QMiMn1b0n0g1SMiMn1a0l0g1UMjMo1`0i0h1WMjMn1`0h0h1YMjMaNb0n1MU2l1XM\\O<iNZ2n1VMC8`N_2R2UMA9_N_2T2UM^O<_N\\2X2TM[O?^NZ2[2mL_Oi0WNX2]2kL^Ol0WNV2^2kL]Oo0VNS2`2kL\\OR1UNP2c2bLC]1kMn1d2dLB]1lMl1e2dLA`1kMj1e2eLAa1kMg1g2fL_Ob1oMb1d2kL_Ob1PN^1[6aNiIZ1Y6dNmIW1T6hNPJS1R6kNTJo0o5oNXJi0j5UO]Jd0c5]OdJ<\\5ClJ5U5KRKMo42YKFh4:_K^Ob4a0eKXO\\4h0kKQOT4o0SLjNn3V1d4O100O100O00100O100O100O00100O100O1O100O010O100O00010O01O010O000010O01O010O000010O01O010O001O01O011N100O101N100O2O0O101N1O101N100O100O010O100O[OA^C`0b<A]C>c<D\\C;d<G[C9e<HYC8g<JXC6h<f0010O00010O0010O01O01O01N3M6I7J6DYeQ4"}, "label": "a tennis player wearing a red top and white shorts"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis player in the red shirt\"."}], "task": "refering", "answer_template": "The \"the tennis player in the red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 127, 142, 143, 144, 145, 146, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 308, 327, 328, 329, 330, 350, 351, 352], "bbox": [0.18884375, 0.05310869565217391, 0.548828125, 1.0], "iscrowd": 0, "area": 42875.38720000001, "rle": {"size": [460, 640], "counts": "Sdf1f0a=>B=B>C=B>A?B>A=D9ZOf0H9H9G8G:G8G:G8G5L4L4K5L4TKZHm2j7SMWHe2Q8ZMPH_2W8_MkG^2X8aMiG]2Y8aMiG]2Y8bMiGZ2Z8eMgGY2[8ZL\\GU1;^2\\8[L[GV1:]2c8aM_Ge0FFo8C\\Gc0KFk8E\\Gb0MEj8HZG?2Eg8JXG>6Ce8MWG<9Dc8NUG;=Ba81TG9a0B]84SG8d0A[85RG9e0@Z85SG;e0^OX86TG;g0\\OW86TG>g0ZOU87UG>h0YO`8KjFl0h0UOe9j0o100O100O100O100O1001O00000000000000001O000000000000000N2K6K4L4N2N2M3N2O101N1O2N2N2N2O0O2N2N2N2O1N1O2QFhMg7Z2WHhMg7Z2XHhMo1b0o0g1QMiMn1b0n0g1SMiMn1a0l0g1UMjMo1`0i0h1WMjMn1`0h0h1YMjMaNb0n1MU2l1XM\\O<iNZ2n1VMC8`N_2R2UMA9_N_2T2UM^O<_N\\2X2TM[O?^NZ2[2mL_Oi0WNX2]2kL^Ol0WNV2^2kL]Oo0VNS2`2kL\\OR1UNP2c2bLC]1kMn1d2dLB]1lMl1e2dLA`1kMj1e2eLAa1kMg1g2fL_Ob1oMb1d2kL_Ob1PN^1[6aNiIZ1Y6dNmIW1T6hNPJS1R6kNTJo0o5oNXJi0j5UO]Jd0c5]OdJ<\\5ClJ5U5KRKMo42YKFh4:_K^Ob4a0eKXO\\4h0kKQOT4o0SLjNn3V1d4O100O100O00100O100O100O00100O100O1O100O010O100O00010O01O010O000010O01O010O000010O01O010O001O01O011N100O101N100O2O0O101N1O101N100O100O010O100O[OA^C`0b<A]C>c<D\\C;d<G[C9e<HYC8g<JXC6h<f0010O00010O0010O01O01O01N3M6I7J6DYeQ4"}, "label": "the tennis player in the red shirt"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man spectating a tennis game\"."}], "task": "refering", "answer_template": "The \"a man spectating a tennis game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 201, 202, 203, 220, 221, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.519640625, 0.4452826086956522, 0.9821718749999999, 0.9886956521739131], "iscrowd": 0, "area": 38407.0774, "rle": {"size": [460, 640], "counts": "Roe4c0g=3M2N2O1N2N2N2N2N1O2O1N2N2N2N2N1O2O1N2N2N2N2N2N1O2O1N2N2N2N2N2N101N2N2N2N2N2N1N3J6J6K5J6J6J60O100000001O000O1000000000001O0O10000O100O1O1O101N1O1jLSFl1W;0O01O00100O001O03N7I7H8I7H8I7I7H8I7I0O10O001000O100O10000O10000O100O10000O100O10000O100O10000O2O000O100O10000O100N2L4L4L4L4I7H8I7H8H8H8I7H8O1O1N2O1O1N2O1O1N2O1O1N2O100000000001O000000000000010O000000000000001O000000000000001O1O1O1O1O1O2N1O1iJRIn2o6nLVIP3l6kLZIR3g6iL_IU3b6gLcIW3^6eLhIX3Y6dLlIZ3V6`LPJ^3Q6^LUJ_3l5]LYJa3h5ZL\\Jf3h5SL\\Jl3g5nK\\JR4h5gK\\JX4l70001O00001O001O000010O0001O001O00001O00001O001O000XOi0VOi0N22O0O101N1O100O2N100010O00100N2O001O1N101O1N101O1O0O2O1O001N2O1O001N2O001N2O001O1N101O2N1N3N1O2M2O2K4L4L5K4K6K4L5K4L5K4Ln`4"}, "label": "a man spectating a tennis game"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in grey shirt\"."}], "task": "refering", "answer_template": "The \"man in grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 201, 202, 203, 220, 221, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.519640625, 0.4452826086956522, 0.9821718749999999, 0.9886956521739131], "iscrowd": 0, "area": 38407.0774, "rle": {"size": [460, 640], "counts": "Roe4c0g=3M2N2O1N2N2N2N2N1O2O1N2N2N2N2N1O2O1N2N2N2N2N2N1O2O1N2N2N2N2N2N101N2N2N2N2N2N1N3J6J6K5J6J6J60O100000001O000O1000000000001O0O10000O100O1O1O101N1O1jLSFl1W;0O01O00100O001O03N7I7H8I7H8I7I7H8I7I0O10O001000O100O10000O10000O100O10000O100O10000O100O10000O2O000O100O10000O100N2L4L4L4L4I7H8I7H8H8H8I7H8O1O1N2O1O1N2O1O1N2O1O1N2O100000000001O000000000000010O000000000000001O000000000000001O1O1O1O1O1O2N1O1iJRIn2o6nLVIP3l6kLZIR3g6iL_IU3b6gLcIW3^6eLhIX3Y6dLlIZ3V6`LPJ^3Q6^LUJ_3l5]LYJa3h5ZL\\Jf3h5SL\\Jl3g5nK\\JR4h5gK\\JX4l70001O00001O001O000010O0001O001O00001O00001O001O000XOi0VOi0N22O0O101N1O100O2N100010O00100N2O001O1N101O1N101O1O0O2O1O001N2O1O001N2O001N2O001O1N101O2N1N3N1O2M2O2K4L4L5K4K6K4L5K4L5K4Ln`4"}, "label": "man in grey shirt"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a white shirt with a blue diamond pattern\"."}], "task": "refering", "answer_template": "The \"man in a white shirt with a blue diamond pattern\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 36, 37, 38, 59, 60, 61, 82, 83, 84, 106, 107, 108, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 244, 245, 246, 267, 268], "bbox": [0.53703125, 0.05304347826086956, 0.772296875, 0.7121956521739129], "iscrowd": 0, "area": 24064.842499999995, "rle": {"size": [460, 640], "counts": "Pgj42X>2M4M3M2N3L3N3N101O0100O100O10O01000O0gL_OgHb0V7AiH?T7EkH<Q7HnH8o6LPI5k60TI0h65WILd69[IGa6m0oHTOj6U1SIkNe6_1YIbNe6`1ZInNW6S1iI@c5b0\\J2o4O[JQ1Y5QOaJU1^5lN^JX1a5iN^JX1a5iN]JY1b5hN\\JY1e5gNZJZ1f5fNXJ\\1h5dNVJ^1j5bNTJ_1m5aNQJ^1R6bNlI^1V6bNhI]1[6cNcI\\1`6dN^I[1e6hNVIZ1j6S310000O100000^IoHi5P7XJVIb5j6^J[I]5e6bJaINAb4m6`KhIEEe4c6fKjI@Fj4`6eKhIAKj4\\6eKhI_O0l4X6eKfI^O4n4V6cKeI^O8o4S6bKfJ^4Y5bKgJ^4Z5bKfJ^4Z5aKgJ_4X5aKiJ_4W5aKhJ`4X5_KiJa4V5_KkJa4U5^KkJc4U5]KkJc4T5]KmJb4T5]KmJd4R5\\KmJg4Q5XKPKi4n4WKSKj4l4VKSKl4l4SKUKn4k4PKVKQ5j4mJWKU5i4iJVKY5j4eJWK\\5j4aJWK`5i62O2N2N2M3YIVJh4l5QKZJn4h5jJ_JT5d5cJdJ\\5_5SJQKk5i6N2O2M2O1N2O2M2O1N2O1N3N1N2O1N3N1N2O1N3M2M3N2N3L3N2N2M4M2N2M4hNTFjMP:o1iFSMc9g2U1I8H7L4L4L4M3L4L5L4K4L5L4K5K5L3L5K5L4K_\\Q2"}, "label": "man in a white shirt with a blue diamond pattern"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white and blue shirt on a tennis court shaking hands with another man\"."}], "task": "refering", "answer_template": "The \"a man with a white and blue shirt on a tennis court shaking hands with another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 36, 37, 38, 59, 60, 61, 82, 83, 84, 106, 107, 108, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 244, 245, 246, 267, 268], "bbox": [0.53703125, 0.05304347826086956, 0.772296875, 0.7121956521739129], "iscrowd": 0, "area": 24064.842499999995, "rle": {"size": [460, 640], "counts": "Pgj42X>2M4M3M2N3L3N3N101O0100O100O10O01000O0gL_OgHb0V7AiH?T7EkH<Q7HnH8o6LPI5k60TI0h65WILd69[IGa6m0oHTOj6U1SIkNe6_1YIbNe6`1ZInNW6S1iI@c5b0\\J2o4O[JQ1Y5QOaJU1^5lN^JX1a5iN^JX1a5iN]JY1b5hN\\JY1e5gNZJZ1f5fNXJ\\1h5dNVJ^1j5bNTJ_1m5aNQJ^1R6bNlI^1V6bNhI]1[6cNcI\\1`6dN^I[1e6hNVIZ1j6S310000O100000^IoHi5P7XJVIb5j6^J[I]5e6bJaINAb4m6`KhIEEe4c6fKjI@Fj4`6eKhIAKj4\\6eKhI_O0l4X6eKfI^O4n4V6cKeI^O8o4S6bKfJ^4Y5bKgJ^4Z5bKfJ^4Z5aKgJ_4X5aKiJ_4W5aKhJ`4X5_KiJa4V5_KkJa4U5^KkJc4U5]KkJc4T5]KmJb4T5]KmJd4R5\\KmJg4Q5XKPKi4n4WKSKj4l4VKSKl4l4SKUKn4k4PKVKQ5j4mJWKU5i4iJVKY5j4eJWK\\5j4aJWK`5i62O2N2N2M3YIVJh4l5QKZJn4h5jJ_JT5d5cJdJ\\5_5SJQKk5i6N2O2M2O1N2O2M2O1N2O1N3N1N2O1N3N1N2O1N3M2M3N2N3L3N2N2M4M2N2M4hNTFjMP:o1iFSMc9g2U1I8H7L4L4L4M3L4L5L4K4L5L4K5K5L3L5K5L4K_\\Q2"}, "label": "a man with a white and blue shirt on a tennis court shaking hands with another man"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis racket with a yellow rim\"."}], "task": "refering", "answer_template": "The \"a tennis racket with a yellow rim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 188, 238, 261, 262, 284, 285], "bbox": [0.14909375, 0.44632608695652176, 0.426078125, 0.7975217391304348], "iscrowd": 0, "area": 1774.5230999999994, "rle": {"size": [460, 640], "counts": "VlZ19R>101N101N2N101N1101N11O2O0O1O01N100O010O01O010O010O01O010O010O001J8G\\i^1OkVaN2N3N1O1O2N3M3M3M3M4L5J5L5K5K4L5KO001O1O1N2O1O1N2O2N1N2O2M2N2O1N3M2N2O3J8F:Go`T5"}, "label": "a tennis racket with a yellow rim"}]}
{"id": 291366, "image": "COCO_train2014_000000291366.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis recket being help by guy in a red shirt\"."}], "task": "refering", "answer_template": "The \"tennis recket being help by guy in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 188, 238, 261, 262, 284, 285], "bbox": [0.14909375, 0.44632608695652176, 0.426078125, 0.7975217391304348], "iscrowd": 0, "area": 1774.5230999999994, "rle": {"size": [460, 640], "counts": "VlZ19R>101N101N2N101N1101N11O2O0O1O01N100O010O01O010O010O01O010O010O001J8G\\i^1OkVaN2N3N1O1O2N3M3M3M3M4L5J5L5K5K4L5KO001O1O1N2O1O1N2O2N1N2O2M2N2O1N3M2N2O3J8F:Go`T5"}, "label": "tennis recket being help by guy in a red shirt"}]}
{"id": 537127, "image": "COCO_train2014_000000537127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair facing a window\"."}], "task": "refering", "answer_template": "The \"a brown chair facing a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 120, 121, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.668, 0.47125748502994014, 1.0, 0.9755389221556887], "iscrowd": 0, "area": 21863.895549999997, "rle": {"size": [334, 500], "counts": "nT]32Y:4K5L4K5K5K5THTOQ6o0kI[Oo5f0kIFm5<mI0l51nI;j5GPJ`0n5AlIe0R6]OhIi0W6XOcIm0\\6UO^IQ1a6^1O2O0O2O1N101N8I0O1001O0000001O00010O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00O1000000O10000O10000000000000000L4H8N2N2M3N2N2N2N2N2N2O1N2N2O1O1O100O1O100O100O1000000O2O00000000000001O00001O0000001O00001O0000001O00001O0000001nKfLn0Z3]NA"}, "label": "a brown chair facing a window"}]}
{"id": 537127, "image": "COCO_train2014_000000537127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather chair placed the farest away from the window\"."}], "task": "refering", "answer_template": "The \"a leather chair placed the farest away from the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 120, 121, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.668, 0.47125748502994014, 1.0, 0.9755389221556887], "iscrowd": 0, "area": 21863.895549999997, "rle": {"size": [334, 500], "counts": "nT]32Y:4K5L4K5K5K5THTOQ6o0kI[Oo5f0kIFm5<mI0l51nI;j5GPJ`0n5AlIe0R6]OhIi0W6XOcIm0\\6UO^IQ1a6^1O2O0O2O1N101N8I0O1001O0000001O00010O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00O1000000O10000O10000000000000000L4H8N2N2M3N2N2N2N2N2N2O1N2N2O1O1O100O1O100O100O1000000O2O00000000000001O00001O0000001O00001O0000001O00001O0000001nKfLn0Z3]NA"}, "label": "a leather chair placed the farest away from the window"}]}
{"id": 537127, "image": "COCO_train2014_000000537127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a two seater sofa in a living room\"."}], "task": "refering", "answer_template": "The \"a two seater sofa in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 47, 48, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 98, 101, 102, 103, 119, 120], "bbox": [0.4345, 0.210748502994012, 0.7427, 0.5096706586826348], "iscrowd": 0, "area": 9925.069799999994, "rle": {"size": [334, 500], "counts": "_lV21[:4K4M3N3N1N2O0O2`F@P9a0lFNg83VG4f8m0L01O010O001O1O010O001O0010O01ON2K5H90O1O001O1O010O1O10O10000O1000000O1000000000000000000O10000000000000000001O0000000000001O00001O001O00001O000000000000O100000000O10001N10000O101O000O10001O0_HRNb6n1XIXNh6h1PI`NP7`1hHhNX7W2000000O11O00001O00001O00001O00001O0000001O0000001O0000001O3M2L?@3M2O2M5POfGE^89oGXOV8e0k0O2M3M2M4KUoY1"}, "label": "a two seater sofa in a living room"}]}
{"id": 537127, "image": "COCO_train2014_000000537127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch on the left\"."}], "task": "refering", "answer_template": "The \"the couch on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166], "bbox": [0.0043, 0.24032934131736525, 0.42574, 0.7854191616766466], "iscrowd": 0, "area": 28256.984549999994, "rle": {"size": [334, 500], "counts": "`k08]9l0^MUObJg1V5]NSJ^2j5V1M3M3M3L4M3M2N1O00000O10000000001O0O1000000000001N1000000000000O101O0000000O100000001O000O10000000001O0O1000000000000O2O0000000001O1O100O1O1O1O001O1O1O100O1O1O1O1O1O1O1O100O1O001O1O1O1O100OfNhK_LW4b3iK^LV4b3lK]LS4c3nK]LQ4c3QL\\Ln3d3SL\\Ll3e3ULZLj3f3WLZLi3e3XL[Lg3e3[LYLe3g3\\LYLc3h3^LWLa3i3`LWL_3i3cLVL\\3j3eLVLZ3k3gLTLY3k3hLULW3k3kLTLT3l3mLTLR3l3PMSLo2n3QMRLn2n3TMQLk2o3VMQLi2o3XMQLg2o3[MPLc2R4]MmKb2T4`MkK]2W4dMgK\\2Z4fMcKZ2^4gM`KY2b4hM[KX2f4iMWKX2j4jMSKV2n4kMPKU2Q5mMkJT2W5lMgJT2Z5_11N2O2N1O1O1O1O2N1O1O1O1O2M2O1O1O1O2N1O1O1O1O1O1N2O1O1O2N1O1O1O1O1O1O1N2O1O1O1O1O2N1O1O1O1O1N2O1O1O1O1O1O1O2N3M5K5J7J1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1OU`m2"}, "label": "the couch on the left"}]}
{"id": 537127, "image": "COCO_train2014_000000537127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather couch with four throw pillows resting in its corners\"."}], "task": "refering", "answer_template": "The \"a brown leather couch with four throw pillows resting in its corners\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166], "bbox": [0.0043, 0.24032934131736525, 0.42574, 0.7854191616766466], "iscrowd": 0, "area": 28256.984549999994, "rle": {"size": [334, 500], "counts": "`k08]9l0^MUObJg1V5]NSJ^2j5V1M3M3M3L4M3M2N1O00000O10000000001O0O1000000000001N1000000000000O101O0000000O100000001O000O10000000001O0O1000000000000O2O0000000001O1O100O1O1O1O001O1O1O100O1O1O1O1O1O1O1O100O1O001O1O1O1O100OfNhK_LW4b3iK^LV4b3lK]LS4c3nK]LQ4c3QL\\Ln3d3SL\\Ll3e3ULZLj3f3WLZLi3e3XL[Lg3e3[LYLe3g3\\LYLc3h3^LWLa3i3`LWL_3i3cLVL\\3j3eLVLZ3k3gLTLY3k3hLULW3k3kLTLT3l3mLTLR3l3PMSLo2n3QMRLn2n3TMQLk2o3VMQLi2o3XMQLg2o3[MPLc2R4]MmKb2T4`MkK]2W4dMgK\\2Z4fMcKZ2^4gM`KY2b4hM[KX2f4iMWKX2j4jMSKV2n4kMPKU2Q5mMkJT2W5lMgJT2Z5_11N2O2N1O1O1O1O2N1O1O1O1O2M2O1O1O1O2N1O1O1O1O1O1N2O1O1O2N1O1O1O1O1O1O1N2O1O1O1O1O2N1O1O1O1O1N2O1O1O1O1O1O1O2N3M5K5J7J1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1OU`m2"}, "label": "a brown leather couch with four throw pillows resting in its corners"}]}
{"id": 315944, "image": "COCO_train2014_000000315944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey and white puppy eating out of a bowl\"."}], "task": "refering", "answer_template": "The \"a grey and white puppy eating out of a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 28, 29, 30, 31, 32, 33, 34, 35, 36, 51, 52, 53, 54, 55, 56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 283, 284, 285, 286, 287, 288, 307, 308, 309, 310], "bbox": [0.1681875, 0.0040416666666666665, 0.61515625, 0.8060625], "iscrowd": 0, "area": 65723.37505000002, "rle": {"size": [480, 640], "counts": "dnb11n>2N2O0O2N2O1N1O2N2O1N2N1dMZ1^DgNT;g1jD[Nf:U2XEmMV:e2iE[Mj9S3TFoLc9Z3[FhL\\9a3bFaLU9h3jFYLm8P4QGQLg8Y5G8H7I8H8K5N2M3M3N1N3N2M3M3N2M1O100O101N1O100O2N1O001O1O001O1O001O10001O1O001N2O001O1O1O3N2N3M2N3M2O1N3M2N3N1N3M2N3M2O1N3M2N3M2N3M2O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O001O001O00001O001O00001O001O00001O001O000010O01O00001O000000O10000000000O1000000O100O10000RInLm1S3oM]Me1c2VNnM]1S2^N]NW1c1dNbN[1_1`NgN_1Y1\\NlNc1U1ZNoNc1S1[NPOc1Q1\\NPOc1Q1\\NQOb1P1]NQOa1Q1^NQO`1P1_NQO`1P1_NRO_1o0`NRO`1n0_NTO_1m0`NTO`1l0_NVO_1k0_NXO`1h0_NYOa1g0^N[O`1f0_N[Oa1e0^NEY1;fN3m0MROb0a0]O^OQ15oNJX11gNN_1MaN1g1JXN5n1GQN8U2CkM<X2CgM<\\2BdM=_2AaM>b2@^M?e2_O[M`0h2^OXMa0j2^OVMa0m2]OSMc0o2[OQMd0R3ZOnLe0V3XOjLh0Y3UOgLj0]3SOcLm0`3PO`Lo0d3nN\\LQ1g3mNYLS1i3kNWLT1l3jNTLU1n3jNRLV1o3iNQLV1Q4hNPLX1Q4gNoKX1S4gNmKX1U4gNkKY1V4fNjKY1X4fNhKZ1Y4eNgKZ1[4eNeKZ1]4eNcK[1^4cNcK]1^4aNcK_1^4^NdKb1]4\\NdKd1]4YNeKg1\\4VNfKj1[4SNgKm1Z4iMoKW2]80000000000000000000000000000O10001O0000000000000000000O100000000000000000000M3L4K5L4K5L4K5L4H8F:F:FVZc3"}, "label": "a grey and white puppy eating out of a bowl"}]}
{"id": 315944, "image": "COCO_train2014_000000315944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey colored puppy eating food\"."}], "task": "refering", "answer_template": "The \"a grey colored puppy eating food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 28, 29, 30, 31, 32, 33, 34, 35, 36, 51, 52, 53, 54, 55, 56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 283, 284, 285, 286, 287, 288, 307, 308, 309, 310], "bbox": [0.1681875, 0.0040416666666666665, 0.61515625, 0.8060625], "iscrowd": 0, "area": 65723.37505000002, "rle": {"size": [480, 640], "counts": "dnb11n>2N2O0O2N2O1N1O2N2O1N2N1dMZ1^DgNT;g1jD[Nf:U2XEmMV:e2iE[Mj9S3TFoLc9Z3[FhL\\9a3bFaLU9h3jFYLm8P4QGQLg8Y5G8H7I8H8K5N2M3M3N1N3N2M3M3N2M1O100O101N1O100O2N1O001O1O001O1O001O10001O1O001N2O001O1O1O3N2N3M2N3M2O1N3M2N3N1N3M2N3M2O1N3M2N3M2N3M2O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O001O001O00001O001O00001O001O00001O001O000010O01O00001O000000O10000000000O1000000O100O10000RInLm1S3oM]Me1c2VNnM]1S2^N]NW1c1dNbN[1_1`NgN_1Y1\\NlNc1U1ZNoNc1S1[NPOc1Q1\\NPOc1Q1\\NQOb1P1]NQOa1Q1^NQO`1P1_NQO`1P1_NRO_1o0`NRO`1n0_NTO_1m0`NTO`1l0_NVO_1k0_NXO`1h0_NYOa1g0^N[O`1f0_N[Oa1e0^NEY1;fN3m0MROb0a0]O^OQ15oNJX11gNN_1MaN1g1JXN5n1GQN8U2CkM<X2CgM<\\2BdM=_2AaM>b2@^M?e2_O[M`0h2^OXMa0j2^OVMa0m2]OSMc0o2[OQMd0R3ZOnLe0V3XOjLh0Y3UOgLj0]3SOcLm0`3PO`Lo0d3nN\\LQ1g3mNYLS1i3kNWLT1l3jNTLU1n3jNRLV1o3iNQLV1Q4hNPLX1Q4gNoKX1S4gNmKX1U4gNkKY1V4fNjKY1X4fNhKZ1Y4eNgKZ1[4eNeKZ1]4eNcK[1^4cNcK]1^4aNcK_1^4^NdKb1]4\\NdKd1]4YNeKg1\\4VNfKj1[4SNgKm1Z4iMoKW2]80000000000000000000000000000O10001O0000000000000000000O100000000000000000000M3L4K5L4K5L4K5L4H8F:F:FVZc3"}, "label": "a grey colored puppy eating food"}]}
{"id": 315944, "image": "COCO_train2014_000000315944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white puppy struggles to reach the food\"."}], "task": "refering", "answer_template": "The \"black and white puppy struggles to reach the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293], "bbox": [0.431546875, 0.16374999999999998, 0.788203125, 0.765875], "iscrowd": 0, "area": 49965.82455000001, "rle": {"size": [480, 640], "counts": "h`Q4Z1`=6K6J5K6K4O2N1O2N2N1O2O1N2N2N1O2N2N2N101N8H:F;E:F;F:E=C9G3M2O2M3M2N3M3N2M3M4L3N3L3M4M2M3M2N2O1N2N2N2O1N1O2N2O0O2N101N2N101O0O2O1O001O001O001N101O1O001O001O1N101O1O001O1O001N2O1O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O01O001O00001O00001O00001O001O00001O01O01O00001O00010O00001O001O0010O01O001O001O0N3O1O001O001O001N101O001O001O001O001O001O1O001O001O001O001N1O2N1lMmH[LT7c3PIYLQ7e3SIXLo6e3VIULm6j3WIPLm6m3XInKj6Q4ZIkKh6R4\\IkKf6R4^IkKc6T4`IiKb6T4bIjK_6T4dIjK\\6T4hIjKY6T4jIjKW6S4mIjKT6T4PJjKQ6T4RJjKo5S4VJjKj5U4ZJhKg5U4^JhKc5T4f2L5L4L4K6K4L4M3M4L3M3M3M3M4L3WOi0Fjco1"}, "label": "black and white puppy struggles to reach the food"}]}
{"id": 315944, "image": "COCO_train2014_000000315944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black / white puppy\"."}], "task": "refering", "answer_template": "The \"black / white puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293], "bbox": [0.431546875, 0.16374999999999998, 0.788203125, 0.765875], "iscrowd": 0, "area": 49965.82455000001, "rle": {"size": [480, 640], "counts": "h`Q4Z1`=6K6J5K6K4O2N1O2N2N1O2O1N2N2N1O2N2N2N101N8H:F;E:F;F:E=C9G3M2O2M3M2N3M3N2M3M4L3N3L3M4M2M3M2N2O1N2N2N2O1N1O2N2O0O2N101N2N101O0O2O1O001O001O001N101O1O001O001O1N101O1O001O1O001N2O1O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O01O001O00001O00001O00001O001O00001O01O01O00001O00010O00001O001O0010O01O001O001O0N3O1O001O001O001N101O001O001O001O001O001O1O001O001O001O001N1O2N1lMmH[LT7c3PIYLQ7e3SIXLo6e3VIULm6j3WIPLm6m3XInKj6Q4ZIkKh6R4\\IkKf6R4^IkKc6T4`IiKb6T4bIjK_6T4dIjK\\6T4hIjKY6T4jIjKW6S4mIjKT6T4PJjKQ6T4RJjKo5S4VJjKj5U4ZJhKg5U4^JhKc5T4f2L5L4L4K6K4L4M3M4L3M3M3M3M4L3WOi0Fjco1"}, "label": "black / white puppy"}]}
{"id": 209449, "image": "COCO_train2014_000000209449.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"largest piece of broccoli touching the onions\"."}], "task": "refering", "answer_template": "The \"largest piece of broccoli touching the onions\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 294], "bbox": [0.59071875, 0.5011241217798594, 0.8456093750000001, 0.8157377049180328], "iscrowd": 0, "area": 16118.849950000002, "rle": {"size": [427, 640], "counts": "alm46Q==B=D=E:G5K2N3N2M2N3M3N2M2N3M3N1N3M3N1N3M3M2O1N1L3I8L40O0100O010O10O0100O010O10O0100O010O10O0100O010O10O01000O10000O10O1000O10000O10000O10000O01000O10000000001O0000001O0000001O0000010O001O001O001O1O001O1O001O1O0010O0100O1O2O1N2O2M2N2O0O010O100O1O010O100O00100O100O010O1O1O001O1O001O1O001O1O001O1O1O1O1O2N1N2O2N1M3N3L3N2M6J;FloX1"}, "label": "largest piece of broccoli touching the onions"}]}
{"id": 209449, "image": "COCO_train2014_000000209449.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brocolli\"."}], "task": "refering", "answer_template": "The \"the brocolli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 249, 250, 251, 252, 268, 269, 270, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.572703125, 0.4402107728337237, 1.0, 0.9964637002341921], "iscrowd": 0, "area": 38712.4911, "rle": {"size": [427, 640], "counts": "n\\i45S=5L4K5K6K4K5K5L5L3N2N2N3L3N2N3L3N2N2N3L1000000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000000000000000000000O1000000000000000000000000O10000000000000000000000000000001O00001O0000001O0000001O00001O0000001OhL\\MiKd2V4_MhKa2V4cMhK]2W4fMhKY2W4jMgKV2X4mMfKS2X4QNfKo1Y4TNfKk1Y4XNeKh1Y4\\NeKd1Z4]NgKb1X4`NgK`1W4cNhK]1W4eNhK[1W4fNjKY1U4iNjKW1T4lNkKT1T4nNkKR1T4oNmKP1R4ROmKn0S4SOlKm0S4UOmKj0R4WOnKi0Q4YOnKg0R4ZOlKg0S4[OlKe0S4\\OlKe0S4]OlKc0T4^OkKb0T4@jKa0U4@kK`0U4AiK`0V4BiK>V4DhK=W4DiK<W4EhK;W4GgK:X4HgK8X4IgK8Y4IdK9[4IbK9]4I_K:a4G[K<d4HVK;i4IQK:o4Q40000O100000000O100000000O100000000O100000000O10000000000O10000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001ULnK[Nl9E;C\\C"}, "label": "the brocolli"}]}
{"id": 209449, "image": "COCO_train2014_000000209449.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"raw broccoli in the far corner of the photo\"."}], "task": "refering", "answer_template": "The \"raw broccoli in the far corner of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 249, 250, 251, 252, 268, 269, 270, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.572703125, 0.4402107728337237, 1.0, 0.9964637002341921], "iscrowd": 0, "area": 38712.4911, "rle": {"size": [427, 640], "counts": "n\\i45S=5L4K5K6K4K5K5L5L3N2N2N3L3N2N3L3N2N2N3L1000000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000000000000000000000O1000000000000000000000000O10000000000000000000000000000001O00001O0000001O0000001O00001O0000001OhL\\MiKd2V4_MhKa2V4cMhK]2W4fMhKY2W4jMgKV2X4mMfKS2X4QNfKo1Y4TNfKk1Y4XNeKh1Y4\\NeKd1Z4]NgKb1X4`NgK`1W4cNhK]1W4eNhK[1W4fNjKY1U4iNjKW1T4lNkKT1T4nNkKR1T4oNmKP1R4ROmKn0S4SOlKm0S4UOmKj0R4WOnKi0Q4YOnKg0R4ZOlKg0S4[OlKe0S4\\OlKe0S4]OlKc0T4^OkKb0T4@jKa0U4@kK`0U4AiK`0V4BiK>V4DhK=W4DiK<W4EhK;W4GgK:X4HgK8X4IgK8Y4IdK9[4IbK9]4I_K:a4G[K<d4HVK;i4IQK:o4Q40000O100000000O100000000O100000000O100000000O10000000000O10000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001ULnK[Nl9E;C\\C"}, "label": "raw broccoli in the far corner of the photo"}]}
{"id": 274986, "image": "COCO_train2014_000000274986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a part of motorcycle on the right\"."}], "task": "refering", "answer_template": "The \"a part of motorcycle on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 91, 114, 137, 160, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 274, 275, 298, 321], "bbox": [0.8660625, 0.08051643192488263, 0.9991875, 0.8984976525821596], "iscrowd": 0, "area": 10271.447600000003, "rle": {"size": [426, 640], "counts": "XTW72X=3M2N2N2N1O1O001O00001O001O00001@@TD`0l;CQD>m;GPD8P<KmC5R<LnCh0^;XOcDi0Z;XOfDj0X;VOiDj0U;WOkDk0S;UOmDm0P;UOoDm0o:SOQEn0m:SOSEo0k:QOUEQ1h:POXEQ1g:SOUEo0i:j0O2N1O2N1N3N1O2N1O2M2N3N2M2N3M2N3M2O1N2N2N2N2Nd0\\O9H0RK\\KU1e4hN_KV1b4hNbKU1_4hNfKU1[4jNgKT1Z4jNjKS1W4lNlKQ1U4mNnKP1T4oNoKn0]4fNeKkMTOc2Y5@fKaMFc2f4JQMLQ32SMIP36TMDo2:UMAm2>ZMZOf2h0cMmN]2U1lM`NT2b1VNRNi1Q2`NdM`1^2iNWMW1k2a44M2M3NX<"}, "label": "a part of motorcycle on the right"}]}
{"id": 274986, "image": "COCO_train2014_000000274986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle with only a sliver of it showing\"."}], "task": "refering", "answer_template": "The \"the motorcycle with only a sliver of it showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 91, 114, 137, 160, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 274, 275, 298, 321], "bbox": [0.8660625, 0.08051643192488263, 0.9991875, 0.8984976525821596], "iscrowd": 0, "area": 10271.447600000003, "rle": {"size": [426, 640], "counts": "XTW72X=3M2N2N2N1O1O001O00001O001O00001@@TD`0l;CQD>m;GPD8P<KmC5R<LnCh0^;XOcDi0Z;XOfDj0X;VOiDj0U;WOkDk0S;UOmDm0P;UOoDm0o:SOQEn0m:SOSEo0k:QOUEQ1h:POXEQ1g:SOUEo0i:j0O2N1O2N1N3N1O2N1O2M2N3N2M2N3M2N3M2O1N2N2N2N2Nd0\\O9H0RK\\KU1e4hN_KV1b4hNbKU1_4hNfKU1[4jNgKT1Z4jNjKS1W4lNlKQ1U4mNnKP1T4oNoKn0]4fNeKkMTOc2Y5@fKaMFc2f4JQMLQ32SMIP36TMDo2:UMAm2>ZMZOf2h0cMmN]2U1lM`NT2b1VNRNi1Q2`NdM`1^2iNWMW1k2a44M2M3NX<"}, "label": "the motorcycle with only a sliver of it showing"}]}
{"id": 274986, "image": "COCO_train2014_000000274986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and red motorcycle parked on a street\"."}], "task": "refering", "answer_template": "The \"a black and red motorcycle parked on a street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 143, 144, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 326, 327], "bbox": [0.07964062499999999, 0.06037558685446009, 0.7858281249999999, 0.9430046948356808], "iscrowd": 0, "area": 86191.47880000004, "rle": {"size": [426, 640], "counts": "P`e0:`<e0E:J7J5J7J5J7J5J7J5J7J5J7J5J3N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2O1N2N100O101N100O1O100O100O2N100O100O100O1O101N100O1O100O100O2O0O1O100O100O1O2O00000000000000000000O100000000000O10000000000000000000000000O10000O2O001N10hNjI]KU6Z4TJfKk5R4_JnK`5h3kJWLU5`3UK`Li4X3`KhL`4U3dKkL[4T3gKkLX4V3iKjLV4U3kKkLU4T3mKlLQ4T3QLkLo3U3QLlLn3S3TLlLk3T3WLlLh3S3ZLlLf3T3ZLmLd3S3^LlLb3S3`LmL_3R3bLnL]3S3dLmL[3R3fLnLY3R3hLoLU3R3kLoLT3R3kLPMS3P3mLQMR3o2oLRMo2n2QMSMm2o2RMSMl2m2TMUMj2k2VMVMi2j2XMWMf2j2YMWMf2i2ZMYMc2i2\\MXMc2h2^MYM`2h2_MYM`2g2`MYM`2g2`MYM`2h2aMUM_2l2bMSM^2n2bMQM^2o2cMPM]2Q3cMmL^2S3cMlL\\2V3eMhL[2X3fMfL[2Z3fMeLZ2\\3fMcLZ2]3gMaLY2a3gM^LY2b3iM\\LW2e3iMZLW2f3jMXLV2j3jMULV2k3kMTLU2l3lMRLU2o3lMoKT2Q4mMnKR2T4nMkKR2U4oMiKR2X4nMgKR2Y4oMfKQ2[4PNbKP2_4QN`Ko1`4RN_Kn1b4RN]Ko1b4]300O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O2N1O1O101N1O1O2N1O1O1O2O0000000O100000001O0O100000000O100000000O1000000000O001O1O1O2ZI`L_3b3_LaL^3a3`LbL]3a3`LcL\\3_3bLdL[3^3cLeLY3^3eLeLX3]3eLgLX3[3fLiLV3Z3gLhLV3[3hLgLT3]3jLeLQ3`3mLbLo2b3oL`Ln2d3nL_LP3c3nL_LP3c3nL_LP3d3mL^LQ3d3mL^LQ3d3mL^LQ3e3lL]LR3e3kL^LS3e3jL]LU3d3iL^LV3c3hL_LW3c3fL_LY3b3eL`LY3c3bLaL]3`3]LfLb3[3WLlLh3V3QLPMn3Q3lKUMS4R6O1O1O1O1O1O0000O100O100O100O100O100O1O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N3O000O1000000O1000000O101OO10O1000O10O11N101O001N1N3N1O2N1O2N1O2N1O2M2O2N1O1O2O001O00001O00001N101O00001YLkFW3U9gLnFX3R9eLQG[3P9bLSG]3m8aLUG_3[9O2O001O0O101O0O2O0O101O0O101O0O2O000O2O001N100O2O001N10001N10000O101O0O100O2O000O10001N100O10001N100O100O2N1M3M5K9G9C>^Oa0_O^ih1"}, "label": "a black and red motorcycle parked on a street"}]}
{"id": 274986, "image": "COCO_train2014_000000274986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red bike\"."}], "task": "refering", "answer_template": "The \"red bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 143, 144, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 326, 327], "bbox": [0.07964062499999999, 0.06037558685446009, 0.7858281249999999, 0.9430046948356808], "iscrowd": 0, "area": 86191.47880000004, "rle": {"size": [426, 640], "counts": "P`e0:`<e0E:J7J5J7J5J7J5J7J5J7J5J7J5J3N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2O1N2N100O101N100O1O100O100O2N100O100O100O1O101N100O1O100O100O2O0O1O100O100O1O2O00000000000000000000O100000000000O10000000000000000000000000O10000O2O001N10hNjI]KU6Z4TJfKk5R4_JnK`5h3kJWLU5`3UK`Li4X3`KhL`4U3dKkL[4T3gKkLX4V3iKjLV4U3kKkLU4T3mKlLQ4T3QLkLo3U3QLlLn3S3TLlLk3T3WLlLh3S3ZLlLf3T3ZLmLd3S3^LlLb3S3`LmL_3R3bLnL]3S3dLmL[3R3fLnLY3R3hLoLU3R3kLoLT3R3kLPMS3P3mLQMR3o2oLRMo2n2QMSMm2o2RMSMl2m2TMUMj2k2VMVMi2j2XMWMf2j2YMWMf2i2ZMYMc2i2\\MXMc2h2^MYM`2h2_MYM`2g2`MYM`2g2`MYM`2h2aMUM_2l2bMSM^2n2bMQM^2o2cMPM]2Q3cMmL^2S3cMlL\\2V3eMhL[2X3fMfL[2Z3fMeLZ2\\3fMcLZ2]3gMaLY2a3gM^LY2b3iM\\LW2e3iMZLW2f3jMXLV2j3jMULV2k3kMTLU2l3lMRLU2o3lMoKT2Q4mMnKR2T4nMkKR2U4oMiKR2X4nMgKR2Y4oMfKQ2[4PNbKP2_4QN`Ko1`4RN_Kn1b4RN]Ko1b4]300O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O2N1O1O101N1O1O2N1O1O1O2O0000000O100000001O0O100000000O100000000O1000000000O001O1O1O2ZI`L_3b3_LaL^3a3`LbL]3a3`LcL\\3_3bLdL[3^3cLeLY3^3eLeLX3]3eLgLX3[3fLiLV3Z3gLhLV3[3hLgLT3]3jLeLQ3`3mLbLo2b3oL`Ln2d3nL_LP3c3nL_LP3c3nL_LP3d3mL^LQ3d3mL^LQ3d3mL^LQ3e3lL]LR3e3kL^LS3e3jL]LU3d3iL^LV3c3hL_LW3c3fL_LY3b3eL`LY3c3bLaL]3`3]LfLb3[3WLlLh3V3QLPMn3Q3lKUMS4R6O1O1O1O1O1O0000O100O100O100O100O100O1O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N3O000O1000000O1000000O101OO10O1000O10O11N101O001N1N3N1O2N1O2N1O2N1O2M2O2N1O1O2O001O00001O00001N101O00001YLkFW3U9gLnFX3R9eLQG[3P9bLSG]3m8aLUG_3[9O2O001O0O101O0O2O0O101O0O101O0O2O000O2O001N100O2O001N10001N10000O101O0O100O2O000O10001N100O10001N100O100O2N1M3M5K9G9C>^Oa0_O^ih1"}, "label": "red bike"}]}
{"id": 274986, "image": "COCO_train2014_000000274986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color bike parked on the road\"."}], "task": "refering", "answer_template": "The \"a black color bike parked on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 46, 47, 48, 69, 70, 71, 72, 92, 93, 94, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 184, 185, 186, 207, 208, 209, 230], "bbox": [0.0015, 0.10314553990610328, 0.30743750000000003, 0.6883333333333334], "iscrowd": 0, "area": 21388.9517, "rle": {"size": [426, 640], "counts": "f>j3a9j3VL00000O10O0O2O0O2O1N101N101N101N101N101N2O0O2O0O2O001O001O1O001O001O001O000000O1O1O1O2O0OiIhI\\5Y6fJjIT5W6mJmIm4V6SKmIh4U6XKoIa4T6^KPJ]4R6cKQJW4R6hKnIW4S6jKmIS4V6lKjIS4X6mKhIP4[6PLdIo3^6PLcIm3`6SL_Il3c6SL^Ij3e6VLZIi3h6VLYIg3j6YLUIf3m6YLTId3n6]LQIb3Q7^LoH_3T7`LlH_3V7aLjH]3X7bLhH\\3[7dLeHJ[OQ3R8TMcHK\\Oo2S8VMaHI_On2R8YM^HIAl2S8ZM]HIBj2S8]MZHHEi2S8^MYHHDi2T8`MXHEFh2U8bMUHFFg2W8cMSHEFg2Y8cMRHDFh2Z8dMoGDGg2\\8eMmGCGg2^8eM\\HY2f7gMYHX2i7gMWHW2l7iMSHV2o7iM]Hj1d7WN`Hc1b7]N]Hb1e7]N[Hb1g7^NWHb1k7]NUHb1m7^NRH`1Q8_NnGa1T8_NkG`1W8_NiG`1Y8`NfG_1\\8aNbG_1_8aNaG^1a8aN_G^1c8]N`Gc1T:O100O010O1O010O1O010O010O1O010O00100O010O1O010O010O10O100O10O100000000O100000000000000000000000000000000WOVD2j;NVD2j;NWD1i;OWD0j;0VD0j;0VD0j;OWD1i;OWD1i;OWD1i;OWD1i;OXD0h;OYD1g;OYD1g;OYD1g;OYD1h;NXD2h;MYD3h;LYD3g;J\\D6e;F^D:b;BbD>_;]OeDc0Q<01O00000000001N100000001O000000000M4J5JSTh5"}, "label": "a black color bike parked on the road"}]}
{"id": 274986, "image": "COCO_train2014_000000274986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"solid black bike\"."}], "task": "refering", "answer_template": "The \"solid black bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 46, 47, 48, 69, 70, 71, 72, 92, 93, 94, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 184, 185, 186, 207, 208, 209, 230], "bbox": [0.0015, 0.10314553990610328, 0.30743750000000003, 0.6883333333333334], "iscrowd": 0, "area": 21388.9517, "rle": {"size": [426, 640], "counts": "f>j3a9j3VL00000O10O0O2O0O2O1N101N101N101N101N101N2O0O2O0O2O001O001O1O001O001O001O000000O1O1O1O2O0OiIhI\\5Y6fJjIT5W6mJmIm4V6SKmIh4U6XKoIa4T6^KPJ]4R6cKQJW4R6hKnIW4S6jKmIS4V6lKjIS4X6mKhIP4[6PLdIo3^6PLcIm3`6SL_Il3c6SL^Ij3e6VLZIi3h6VLYIg3j6YLUIf3m6YLTId3n6]LQIb3Q7^LoH_3T7`LlH_3V7aLjH]3X7bLhH\\3[7dLeHJ[OQ3R8TMcHK\\Oo2S8VMaHI_On2R8YM^HIAl2S8ZM]HIBj2S8]MZHHEi2S8^MYHHDi2T8`MXHEFh2U8bMUHFFg2W8cMSHEFg2Y8cMRHDFh2Z8dMoGDGg2\\8eMmGCGg2^8eM\\HY2f7gMYHX2i7gMWHW2l7iMSHV2o7iM]Hj1d7WN`Hc1b7]N]Hb1e7]N[Hb1g7^NWHb1k7]NUHb1m7^NRH`1Q8_NnGa1T8_NkG`1W8_NiG`1Y8`NfG_1\\8aNbG_1_8aNaG^1a8aN_G^1c8]N`Gc1T:O100O010O1O010O1O010O010O1O010O00100O010O1O010O010O10O100O10O100000000O100000000000000000000000000000000WOVD2j;NVD2j;NWD1i;OWD0j;0VD0j;0VD0j;OWD1i;OWD1i;OWD1i;OWD1i;OXD0h;OYD1g;OYD1g;OYD1g;OYD1h;NXD2h;MYD3h;LYD3g;J\\D6e;F^D:b;BbD>_;]OeDc0Q<01O00000000001N100000001O000000000M4J5JSTh5"}, "label": "solid black bike"}]}
{"id": 528941, "image": "COCO_train2014_000000528941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leafy greens in the middle below the melon in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the leafy greens in the middle below the melon in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 288, 289], "bbox": [0.37084375, 0.48989583333333336, 0.638296875, 0.7239583333333334], "iscrowd": 0, "area": 11545.0121, "rle": {"size": [480, 640], "counts": "^\\_31h>8H7O2N2O1N1O2O1N2O1N1O2O0O2N101O001O1O001O001O00001O001O00010O00001O0O101N101N100O2O0O101N101N100O2O0O101N101N100O2O0O101N100O100O2O0O100O100O2O0O100O1O2O0O100O100O2O0O100O`NfC<Z<_OjCc0T<XOlCn0T<mNnCV1Q<gNQD[1o;eNoC]1P<cNoC_1Q<`NnCc1P<]NoCe1Q<ZNnCh1Q<XNnCj1R<UNnCl1Q<TNoCn1P<QNPDP2o;PNQDQ2o;nMQDS2n;mMRDT2n;kMRDV2m;jMSDX2l;hMSDY2l;gMTDZ2l;eMTD\\2k;eMTD\\2l;dMSD^2k;bMVD^2j;bMVD^2i;cMWD]2i;cMWD]2h;dMXD\\2h;cMXD_2g;aMYD_2g;aMYD_2g;aMYD_2g;`MZD`2f;`MZDa2e;^M[Dc2e;]M[Dc2d;]M]Dc2c;\\M^Dd2b;\\M]De2c;ZM]Dh2b;WM^Dj2b;VM]Dk2c;TM]Dm2c;RM^Dn2b;QM^DQ3a;oL^DR3b;mL^DT3b;kL^DV3a;kL^DV3b;jL^DV3b;jL]DX3b;gL_DY3a;gL_DY3a;gL_DY3a;gL_DU12j0_;QN_Dk0>S1R;RN`Dk0?R1Q;SN`Dk0`0Q1P;TNaDj0a0P1n:VNaDj0b0o0l:XNbDi0c0n0k:YNbDi0e0l0i:[NbDi0f0j0i:]NaDi0g0i0h:^NaDi0h0g0h:`N`Di0i0e0h:cN^Dh0k0c0h:eN]Dh0l0a0h:gN\\Dh0m0`0g:hN]Dg0m0?g:jN\\Dg0o0<f:mN[Dg0P1:e:PO\\De0P19e:SO[Db0R19d:UOZDb0S16o:HSE6n:JSE3P;KSE1o:OSENo:0SEMo:3SEIP;5SEHo:7REFP;9SS\\3"}, "label": "the leafy greens in the middle below the melon in the right hand picture"}]}
{"id": 528941, "image": "COCO_train2014_000000528941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leafy vegetable closest to the gord , apple , and bag of dates\"."}], "task": "refering", "answer_template": "The \"a leafy vegetable closest to the gord , apple , and bag of dates\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 288, 289], "bbox": [0.37084375, 0.48989583333333336, 0.638296875, 0.7239583333333334], "iscrowd": 0, "area": 11545.0121, "rle": {"size": [480, 640], "counts": "^\\_31h>8H7O2N2O1N1O2O1N2O1N1O2O0O2N101O001O1O001O001O00001O001O00010O00001O0O101N101N100O2O0O101N101N100O2O0O101N101N100O2O0O101N100O100O2O0O100O100O2O0O100O1O2O0O100O100O2O0O100O`NfC<Z<_OjCc0T<XOlCn0T<mNnCV1Q<gNQD[1o;eNoC]1P<cNoC_1Q<`NnCc1P<]NoCe1Q<ZNnCh1Q<XNnCj1R<UNnCl1Q<TNoCn1P<QNPDP2o;PNQDQ2o;nMQDS2n;mMRDT2n;kMRDV2m;jMSDX2l;hMSDY2l;gMTDZ2l;eMTD\\2k;eMTD\\2l;dMSD^2k;bMVD^2j;bMVD^2i;cMWD]2i;cMWD]2h;dMXD\\2h;cMXD_2g;aMYD_2g;aMYD_2g;aMYD_2g;`MZD`2f;`MZDa2e;^M[Dc2e;]M[Dc2d;]M]Dc2c;\\M^Dd2b;\\M]De2c;ZM]Dh2b;WM^Dj2b;VM]Dk2c;TM]Dm2c;RM^Dn2b;QM^DQ3a;oL^DR3b;mL^DT3b;kL^DV3a;kL^DV3b;jL^DV3b;jL]DX3b;gL_DY3a;gL_DY3a;gL_DY3a;gL_DU12j0_;QN_Dk0>S1R;RN`Dk0?R1Q;SN`Dk0`0Q1P;TNaDj0a0P1n:VNaDj0b0o0l:XNbDi0c0n0k:YNbDi0e0l0i:[NbDi0f0j0i:]NaDi0g0i0h:^NaDi0h0g0h:`N`Di0i0e0h:cN^Dh0k0c0h:eN]Dh0l0a0h:gN\\Dh0m0`0g:hN]Dg0m0?g:jN\\Dg0o0<f:mN[Dg0P1:e:PO\\De0P19e:SO[Db0R19d:UOZDb0S16o:HSE6n:JSE3P;KSE1o:OSENo:0SEMo:3SEIP;5SEHo:7REFP;9SS\\3"}, "label": "a leafy vegetable closest to the gord , apple , and bag of dates"}]}
{"id": 528941, "image": "COCO_train2014_000000528941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brocollie that is uncovered\"."}], "task": "refering", "answer_template": "The \"a brocollie that is uncovered\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 378, 379], "bbox": [0.35362499999999997, 0.6402083333333334, 0.6839375000000001, 0.9562708333333333], "iscrowd": 0, "area": 14590.744950000008, "rle": {"size": [480, 640], "counts": "QiZ35k>4L5K4L4K5L1O00001O0O1000001O0O10001O000O10001O00000O101O002N1N3N1O2N1O2N1N3N1O1O2N1N3N1O2N1O1N3N1O2N1O2N1N2O2N1O000O0100O100O10O10O1000O2O001N2O001O1N101O1O0O10O10O1000O10O100000O010000O0100000O0100000O10O10O1000O10O1000O01000000O0100000O10000O1000000O10O1000O10000O10O1O1O0O2O1O1O0O2O1O001N2O1O001O1O000001N100000000O2O00ZOfCgNZ<Z1fCeNZ<[1gCdNX<]1iCcNV<^1kC`NU<`1lC_NT<b1lC]NT<c1mC\\NR<e1PDZNo;g1QDXNo;h1RDWNn;\\1iCfN:Ml;_1kCcN:Lk;b1lC`N:Ni;b1oC^Ne<c1\\C\\Nc<e1^CYNa<i1`CVN_<k1bCSN^<m1:1O1N2O1O0011O00000O2N1N2N2N3M2O1N2O2M2O2N4K5L4K5L4K5L4K;F[[n2"}, "label": "a brocollie that is uncovered"}]}
{"id": 528941, "image": "COCO_train2014_000000528941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli type vegetable not in bag\"."}], "task": "refering", "answer_template": "The \"broccoli type vegetable not in bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 378, 379], "bbox": [0.35362499999999997, 0.6402083333333334, 0.6839375000000001, 0.9562708333333333], "iscrowd": 0, "area": 14590.744950000008, "rle": {"size": [480, 640], "counts": "QiZ35k>4L5K4L4K5L1O00001O0O1000001O0O10001O000O10001O00000O101O002N1N3N1O2N1O2N1N3N1O1O2N1N3N1O2N1O1N3N1O2N1O2N1N2O2N1O000O0100O100O10O10O1000O2O001N2O001O1N101O1O0O10O10O1000O10O100000O010000O0100000O0100000O10O10O1000O10O1000O01000000O0100000O10000O1000000O10O1000O10000O10O1O1O0O2O1O1O0O2O1O001N2O1O001O1O000001N100000000O2O00ZOfCgNZ<Z1fCeNZ<[1gCdNX<]1iCcNV<^1kC`NU<`1lC_NT<b1lC]NT<c1mC\\NR<e1PDZNo;g1QDXNo;h1RDWNn;\\1iCfN:Ml;_1kCcN:Lk;b1lC`N:Ni;b1oC^Ne<c1\\C\\Nc<e1^CYNa<i1`CVN_<k1bCSN^<m1:1O1N2O1O0011O00000O2N1N2N2N3M2O1N2O2M2O2N4K5L4K5L4K5L4K;F[[n2"}, "label": "broccoli type vegetable not in bag"}]}
{"id": 545325, "image": "COCO_train2014_000000545325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white toothbrush with white and blue bristles\"."}], "task": "refering", "answer_template": "The \"the white toothbrush with white and blue bristles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 90, 91, 92, 93, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 184, 185, 186, 203, 204, 205], "bbox": [0.0, 0.3399099099099099, 0.41514, 1.0], "iscrowd": 0, "area": 17377.53555, "rle": {"size": [333, 500], "counts": "f3f2g72N2N1N3N2N1N3N2N1N3N2N1N3N2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000O10000000000O10O100O100O010O100O1000O0100O100O10O010000O100O001O1O1N2O1O001O1N2O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O1O100O100O100O1O100O100O100O1O101O001O001O00010O001O001O0010O01O001O00003M4M2M4L3M4L3M3M10O0001O00001O01O01O00001O0001O01O00001O00010N1O1O2N1O1O2N1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2Nenn2"}, "label": "the white toothbrush with white and blue bristles"}]}
{"id": 545325, "image": "COCO_train2014_000000545325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and white toothbrush\"."}], "task": "refering", "answer_template": "The \"blue and white toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 90, 91, 92, 93, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 184, 185, 186, 203, 204, 205], "bbox": [0.0, 0.3399099099099099, 0.41514, 1.0], "iscrowd": 0, "area": 17377.53555, "rle": {"size": [333, 500], "counts": "f3f2g72N2N1N3N2N1N3N2N1N3N2N1N3N2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000O10000000000O10O100O100O010O100O1000O0100O100O10O010000O100O001O1O1N2O1O001O1N2O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O1O100O100O100O1O100O100O100O1O101O001O001O00010O001O001O0010O01O001O00003M4M2M4L3M4L3M3M10O0001O00001O01O01O00001O0001O01O00001O00010N1O1O2N1O1O2N1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2Nenn2"}, "label": "blue and white toothbrush"}]}
{"id": 545325, "image": "COCO_train2014_000000545325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow and white toothbrush places in the glass\"."}], "task": "refering", "answer_template": "The \"the yellow and white toothbrush places in the glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 47, 48, 49, 65, 66, 67, 83, 84, 85, 86, 102, 103, 121, 122, 139, 140, 157, 158, 175, 176, 193, 194, 210, 211, 212], "bbox": [0.6330800000000001, 0.08540540540540541, 0.81118, 0.9865165165165165], "iscrowd": 0, "area": 12718.534299999994, "rle": {"size": [333, 500], "counts": "[TW33Y:4M4K4UGF]7=\\HIa7;WHLg77RH0j75nG5l7NmG;m7Y1J5J6K4K100O010O10O0100O010O10O0100O010O100O010O11N01O1O0lNiLSKX3i4lLVKU3f4oLZKP3d4SM\\Km2`4WMaKj2Y4ZMhKk2o3WMSLX3Y3lLgLf3c2^L^MR4m1RLTN_4W1dKjNl4a0XK_OY5LkJ5f5VO]Jk0b72N3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4UJeL\\4^3\\KQMZ4R3^K\\MY4h2^KgMX4\\2`KRNW4Q2aK^NU4V4G:F9G:F5K5K5K5K5K6J5`N_HJ\\9FoWn0"}, "label": "the yellow and white toothbrush places in the glass"}]}
{"id": 545325, "image": "COCO_train2014_000000545325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow toothbrush with white parts\"."}], "task": "refering", "answer_template": "The \"a yellow toothbrush with white parts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 47, 48, 49, 65, 66, 67, 83, 84, 85, 86, 102, 103, 121, 122, 139, 140, 157, 158, 175, 176, 193, 194, 210, 211, 212], "bbox": [0.6330800000000001, 0.08540540540540541, 0.81118, 0.9865165165165165], "iscrowd": 0, "area": 12718.534299999994, "rle": {"size": [333, 500], "counts": "[TW33Y:4M4K4UGF]7=\\HIa7;WHLg77RH0j75nG5l7NmG;m7Y1J5J6K4K100O010O10O0100O010O10O0100O010O100O010O11N01O1O0lNiLSKX3i4lLVKU3f4oLZKP3d4SM\\Km2`4WMaKj2Y4ZMhKk2o3WMSLX3Y3lLgLf3c2^L^MR4m1RLTN_4W1dKjNl4a0XK_OY5LkJ5f5VO]Jk0b72N3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4UJeL\\4^3\\KQMZ4R3^K\\MY4h2^KgMX4\\2`KRNW4Q2aK^NU4V4G:F9G:F5K5K5K5K5K6J5`N_HJ\\9FoWn0"}, "label": "a yellow toothbrush with white parts"}]}
{"id": 545325, "image": "COCO_train2014_000000545325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange toothbrush\"."}], "task": "refering", "answer_template": "The \"the orange toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 25, 26, 27, 28, 43, 44, 45, 46, 61, 62, 63, 64, 80, 81, 82, 99, 100, 117, 118, 135, 136, 153, 154, 171, 172, 189, 190, 191, 207, 208, 209], "bbox": [0.43251999999999996, 0.0, 0.6211, 0.9865165165165165], "iscrowd": 0, "area": 15894.729099999995, "rle": {"size": [333, 500], "counts": "TZV29`9f0D=D;^Ob0M3Lj0VO2O1N2N2O1O10O0100O100O100O010O100O100O10O02O1N3N1O2N1O2N1O1O2N2N3M1O0YOiKQKOHX4V5mKkJOOT4\\4oKgKKM?0g3Z4_LdKJ2g3X4bLeKG3l3Q4TMoKW3e3jL[Ld8000000001O001O000010O01O1O1O1O1O1O1O2N1O6J5K6J6J5K6RJQOR2T1cM[OT2k0`MEV2a0bMKT2:eM2Q23]LgMl0b2]28_M4W2OdM<S2FZMU1\\2nN_M^1W2f3F:F9G6J1O1[OnFMS9NSG0n8KWG5i8F]G8^9N7IbTm1"}, "label": "the orange toothbrush"}]}
{"id": 545325, "image": "COCO_train2014_000000545325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle tooth brush is white with red / orange accents and designs all over it\"."}], "task": "refering", "answer_template": "The \"the middle tooth brush is white with red / orange accents and designs all over it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 25, 26, 27, 28, 43, 44, 45, 46, 61, 62, 63, 64, 80, 81, 82, 99, 100, 117, 118, 135, 136, 153, 154, 171, 172, 189, 190, 191, 207, 208, 209], "bbox": [0.43251999999999996, 0.0, 0.6211, 0.9865165165165165], "iscrowd": 0, "area": 15894.729099999995, "rle": {"size": [333, 500], "counts": "TZV29`9f0D=D;^Ob0M3Lj0VO2O1N2N2O1O10O0100O100O100O010O100O100O10O02O1N3N1O2N1O2N1O1O2N2N3M1O0YOiKQKOHX4V5mKkJOOT4\\4oKgKKM?0g3Z4_LdKJ2g3X4bLeKG3l3Q4TMoKW3e3jL[Ld8000000001O001O000010O01O1O1O1O1O1O1O2N1O6J5K6J6J5K6RJQOR2T1cM[OT2k0`MEV2a0bMKT2:eM2Q23]LgMl0b2]28_M4W2OdM<S2FZMU1\\2nN_M^1W2f3F:F9G6J1O1[OnFMS9NSG0n8KWG5i8F]G8^9N7IbTm1"}, "label": "the middle tooth brush is white with red / orange accents and designs all over it"}]}
{"id": 389681, "image": "COCO_train2014_000000389681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white jet\"."}], "task": "refering", "answer_template": "The \"a green and white jet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 201, 202, 203, 215, 216, 217, 218, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252], "bbox": [0.02695833333333333, 0.6018749999999999, 0.9842708333333333, 0.8760208333333334], "iscrowd": 0, "area": 22712.371550000007, "rle": {"size": [480, 480], "counts": "S_6<d>1O1N2O001O1O1O001N2O1O001O1O1O0O2O1O1O001K5L4K6M2O000O10000000000O1000000000000O10000000000O10000000000000000000000001O000000000000001O000O10000000001O000000000000001O0000000000000000000000001O000000000000000000000000000000000000001O00000O1000000000000000000000000000001O00000000000000000O100000000000000000001O000000000000000000000O10000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000000000000O100000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000000000001O00000001O000000000O1O1N2N2O1N2N2O1N2O1N2N2O1N2N22N9G9G000000000000O100000000000000000000000000000000000000O10000O1000000O10000O1000000O10O10O1000O10O10000O01000O10O1000O10O01O1N2O1O0O2O1O1O0O2O1O1N101O1N2O1O0O2O1O1O0O2O001N101O1N101O0O2O001O0O2O001N101O0O2O1O0O2O001O0O2O001N101000O2O0O10001N10000O2O0]OUDPNl;k1YDUNg;f1^DZNc;`1bD_N_;\\1fDdNZ;W1kDhNV;T1nDlNR;o0SEQOn:i0WEVOj:e0[E[Oe:`0`E@`:;eED]:6hEJX:2lEMU:NPF2P:IUF7Z<O1O100O1O100O100O1O10Z\\3"}, "label": "a green and white jet"}]}
{"id": 389681, "image": "COCO_train2014_000000389681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green plane behind other two\"."}], "task": "refering", "answer_template": "The \"the green plane behind other two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 146], "bbox": [0.4042291666666667, 0.2840625, 0.9631666666666667, 0.47383333333333333], "iscrowd": 0, "area": 8303.54995, "rle": {"size": [480, 480], "counts": "cTk22n>3L101N101N101N1O2O0O2O0O101N100O1O100O100O100O100O100O10000000000O100000000O100000000O100000O10O100000000O100000000O100000O11N4M3M00000O1001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000ATBJk=6WBIi=7YBGg=9ZBFf=:\\BDd=<^BBb==`BB`=>bB@^=`0cB_O]=a0a000O10000000000O100000000O100000000O10000000O10O100000000O1000O1000O1001O001O1N101O1O001O001O1O0O2O1O001O1O001O1N101O1O001O1O001O0O01000000000000O10O10000000O10O100O10000O010O100O1000O0100O100O100O01000O100N2O0O2O2N1N2O1N2O1O1N3N1N2O1O1N2O2N1N2O1N2O1O1N3N1N2O10O1QOiCUOW<a0TD^Om;7_DFb;1hDNY;GRE7n<N1N2O2N1N2O2M2O]W8"}, "label": "the green plane behind other two"}]}
{"id": 389681, "image": "COCO_train2014_000000389681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue plane in upper picture an is blurry\"."}], "task": "refering", "answer_template": "The \"blue plane in upper picture an is blurry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 146], "bbox": [0.4042291666666667, 0.2840625, 0.9631666666666667, 0.47383333333333333], "iscrowd": 0, "area": 8303.54995, "rle": {"size": [480, 480], "counts": "cTk22n>3L101N101N101N1O2O0O2O0O101N100O1O100O100O100O100O100O10000000000O100000000O100000000O100000O10O100000000O100000000O100000O11N4M3M00000O1001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000ATBJk=6WBIi=7YBGg=9ZBFf=:\\BDd=<^BBb==`BB`=>bB@^=`0cB_O]=a0a000O10000000000O100000000O100000000O10000000O10O100000000O1000O1000O1001O001O1N101O1O001O001O1O0O2O1O001O1O001O1N101O1O001O1O001O0O01000000000000O10O10000000O10O100O10000O010O100O1000O0100O100O100O01000O100N2O0O2O2N1N2O1N2O1O1N3N1N2O1O1N2O2N1N2O1N2O1O1N3N1N2O10O1QOiCUOW<a0TD^Om;7_DFb;1hDNY;GRE7n<N1N2O2N1N2O2M2O]W8"}, "label": "blue plane in upper picture an is blurry"}]}
{"id": 422452, "image": "COCO_train2014_000000422452.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the legs of a stool near another room\"."}], "task": "refering", "answer_template": "The \"the legs of a stool near another room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 224, 225, 281, 288, 304, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.241359375, 0.5281455399061034, 0.805046875, 0.987605633802817], "iscrowd": 0, "area": 7706.949350000004, "rle": {"size": [426, 640], "counts": "n]P23e<d0^Oc0O1O0O2O1O001N2O001O1OlNXDd0h;YOZDh0e;UO]Dk0e;QO]Do0e;mN]DS1m;01O000000001O000000001O000000001O000000001O01O000001O000000001O00000000001O000000001O000000001O000000001O00000000000001O000000000000000000000000000000000001O000000000000000000000000000000000001O00000000000O10000O100O10000O10000O10000O10000O10000O10000O10000O100O10001N10000O10000O10000O10000O10000O100O10000O10000O10000O1000O010000O01EZDROf;m0_DoN`;R1cDkN];U1<O010O01000O010O10O10M3L3N3L4L4L5K4L4LPPa1NSP_N3M3N2M3M3M3NO0000010O1O100O001O101O21O01OWDFb::XEKi:6QEOo:1kD5U;KfD;Y;EaD=d;BWD`0l;<3M3M3L4L5E:FXic1"}, "label": "the legs of a stool near another room"}]}
{"id": 397877, "image": "COCO_train2014_000000397877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girl sitting on the bed with the striped shirt\"."}], "task": "refering", "answer_template": "The \"the little girl sitting on the bed with the striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 226, 227, 228, 243, 244, 245, 260, 261, 262, 277, 278, 279, 280, 295, 296, 297, 312, 313, 314, 329, 330], "bbox": [0.31443750000000004, 0.535140625, 0.4867708333333333, 0.848078125], "iscrowd": 0, "area": 9927.1568, "rle": {"size": [640, 480], "counts": "Rhn23mc05K6J5K6J5K5K6J5K6J5K6J5K4L001O1O1O1H7F;E;E;H8O001O1O1O1O001O1O1O1O001O1O1O1O001O1l@hL^=m401O00001O001O000010O3N2M2N3N2M2O2M3M100O10O01O010N2M2O\\1dNk0TO5L4L5K4K5L4L5K4K5L4L5JVYi4"}, "label": "the little girl sitting on the bed with the striped shirt"}]}
{"id": 397877, "image": "COCO_train2014_000000397877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl wearing pink shirt and jeans sits on a bed next to a desk\"."}], "task": "refering", "answer_template": "The \"girl wearing pink shirt and jeans sits on a bed next to a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 226, 227, 228, 243, 244, 245, 260, 261, 262, 277, 278, 279, 280, 295, 296, 297, 312, 313, 314, 329, 330], "bbox": [0.31443750000000004, 0.535140625, 0.4867708333333333, 0.848078125], "iscrowd": 0, "area": 9927.1568, "rle": {"size": [640, 480], "counts": "Rhn23mc05K6J5K6J5K5K6J5K6J5K6J5K4L001O1O1O1H7F;E;E;H8O001O1O1O1O001O1O1O1O001O1O1O1O001O1l@hL^=m401O00001O001O000010O3N2M2N3N2M2O2M3M100O10O01O010N2M2O\\1dNk0TO5L4L5K4K5L4L5K4K5L4L5JVYi4"}, "label": "girl wearing pink shirt and jeans sits on a bed next to a desk"}]}
{"id": 315961, "image": "COCO_train2014_000000315961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small giraffe behind a fence\"."}], "task": "refering", "answer_template": "The \"a small giraffe behind a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 341], "bbox": [0.692703125, 0.6316041666666667, 0.92696875, 0.8855416666666667], "iscrowd": 0, "area": 6071.142299999994, "rle": {"size": [480, 640], "counts": "n``62n>1N1O2N2N1L5J6J5K6G8G62H800O1O1O1O10L3K5K5N3M2O1O1O1O21N1O2N2O0O2gNnf4R3XWK100O100O100O10M2H9H7N3N1N3M2O2M2N3N2M3M2M4M3M3N101N5K6I6J6J7D;Dl^1Ol`Ne0\\Oe0[Od0\\Oe0J5M4L3MaMYDi1c;WNdDe1Y;ZNnDd1n:\\NYE`1d:`NcE\\1Z:dNlEY1Q:gNVFQ1k9oN[Fj0f9VOaFb0`9^OfF;[9DmF4T9LRGMn84YGEg8;`G]Oa8c0n2000001N4M3M3M3M3M3Mhi22RVM5K6K5N101N1O1O1O2O0O1O1O1O2N1OO1000000O10001O1O0O2O1O1O2NXge0"}, "label": "a small giraffe behind a fence"}]}
{"id": 315961, "image": "COCO_train2014_000000315961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe standing near a fence\"."}], "task": "refering", "answer_template": "The \"a baby giraffe standing near a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 341], "bbox": [0.692703125, 0.6316041666666667, 0.92696875, 0.8855416666666667], "iscrowd": 0, "area": 6071.142299999994, "rle": {"size": [480, 640], "counts": "n``62n>1N1O2N2N1L5J6J5K6G8G62H800O1O1O1O10L3K5K5N3M2O1O1O1O21N1O2N2O0O2gNnf4R3XWK100O100O100O10M2H9H7N3N1N3M2O2M2N3N2M3M2M4M3M3N101N5K6I6J6J7D;Dl^1Ol`Ne0\\Oe0[Od0\\Oe0J5M4L3MaMYDi1c;WNdDe1Y;ZNnDd1n:\\NYE`1d:`NcE\\1Z:dNlEY1Q:gNVFQ1k9oN[Fj0f9VOaFb0`9^OfF;[9DmF4T9LRGMn84YGEg8;`G]Oa8c0n2000001N4M3M3M3M3M3Mhi22RVM5K6K5N101N1O1O1O2O0O1O1O1O2N1OO1000000O10001O1O0O2O1O1O2NXge0"}, "label": "a baby giraffe standing near a fence"}]}
{"id": 315961, "image": "COCO_train2014_000000315961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of a giraffe in the background behind the golf cart\"."}], "task": "refering", "answer_template": "The \"the half of a giraffe in the background behind the golf cart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 299, 322], "bbox": [0.0, 0.44987499999999997, 0.130640625, 0.8656458333333333], "iscrowd": 0, "area": 4493.6697, "rle": {"size": [480, 640], "counts": "e:[2m:VOYEj0c8Z1]GgNe8U1]GkNe8R1[GoNg8m0ZGTOg8j0YGWOi8e0XG[Ok8b0UG_Om8=TGDm8:SGGo85RGVO4YNl8^2PGjNW:S1kE^Nf:_1ZERNX;j1iDgMi;V2b0M4M3L4GdQ7bNkaH1[=i01O1O2NZBZOT=e0jBAS=>nBDP=;QCFo<8RCIm<6TCLj<3WCNh<1QC]ONc0Q=NQC@Nd0o<KSCBNe0m<HUCDNf0k<EWCFNg0j<AXCINh0h<^OZCJOj0e<[O\\CLOi0e<ZO\\CNOh0f<XO[C1Og0f<WO[C3Of0f<VO[C5Oe0f<UO[C7Od0g<SOZC:Oc0g<ROZC<Ob0g<QOZC>Oa0P=@PC`0o<AQC?n<BSC=m<CSC=n<BRC>o<AQC?Q=_OoBa0R=^OnBb0S=]OmBa0V=^OjB?Z=@fB>^=_OcB?`=@`B=d=B\\B<g=CYB;j=DVB9o=EQB9\\>M3N2NeiT8"}, "label": "the half of a giraffe in the background behind the golf cart"}]}
{"id": 315961, "image": "COCO_train2014_000000315961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe ' s body is cut off in the picture\"."}], "task": "refering", "answer_template": "The \"a giraffe ' s body is cut off in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 299, 322], "bbox": [0.0, 0.44987499999999997, 0.130640625, 0.8656458333333333], "iscrowd": 0, "area": 4493.6697, "rle": {"size": [480, 640], "counts": "e:[2m:VOYEj0c8Z1]GgNe8U1]GkNe8R1[GoNg8m0ZGTOg8j0YGWOi8e0XG[Ok8b0UG_Om8=TGDm8:SGGo85RGVO4YNl8^2PGjNW:S1kE^Nf:_1ZERNX;j1iDgMi;V2b0M4M3L4GdQ7bNkaH1[=i01O1O2NZBZOT=e0jBAS=>nBDP=;QCFo<8RCIm<6TCLj<3WCNh<1QC]ONc0Q=NQC@Nd0o<KSCBNe0m<HUCDNf0k<EWCFNg0j<AXCINh0h<^OZCJOj0e<[O\\CLOi0e<ZO\\CNOh0f<XO[C1Og0f<WO[C3Of0f<VO[C5Oe0f<UO[C7Od0g<SOZC:Oc0g<ROZC<Ob0g<QOZC>Oa0P=@PC`0o<AQC?n<BSC=m<CSC=n<BRC>o<AQC?Q=_OoBa0R=^OnBb0S=]OmBa0V=^OjB?Z=@fB>^=_OcB?`=@`B=d=B\\B<g=CYB;j=DVB9o=EQB9\\>M3N2NeiT8"}, "label": "a giraffe ' s body is cut off in the picture"}]}
{"id": 315961, "image": "COCO_train2014_000000315961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in a zoo looking left\"."}], "task": "refering", "answer_template": "The \"a giraffe in a zoo looking left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 46, 47, 69, 70, 71, 93, 94, 95, 117, 118, 119, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 256, 257, 259, 260, 279, 280, 282, 283, 284, 302, 303, 306, 307, 325, 326, 329, 330, 348, 349, 352, 353], "bbox": [0.00375, 0.08258333333333333, 0.39865625000000005, 0.9087916666666666], "iscrowd": 0, "area": 26454.017299999978, "rle": {"size": [480, 640], "counts": "Xo04l>Y1gN000000O101O0O1000000O2O000001Ob0^O0010O0001O00001O000010O3M2N3N1N2N3N1N3M20001O010O0O1O0010O000000102M2N7I=D0O2N1O101N2N3M2N2N3M1O1O1O1O1O0000000000002N3M3M3M3M3M3M2N3N2M3M3M1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UGoKd6R4SIWLh0Db3V4TKhLT1VOi3Y6QLkIP4V6jKnIW4S6TK`Jm4i71O1O1O1O2O0O1O00100O1O0dKQKgNP5X1PLiMP4V2QMiLo2W3^4O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O10O01O1O001O1O001O1O00100O001O001O1O001O001O1O0010O3N2N2M2O2N2bEjLl8j4L5J5L5KO001O010O1O001O001O1O001O1O1O13M4L5K4K5L4L5K4L3M3M0OULiGn0W8lNQHQ1P8jNXHS1g7jNZJXOg5c0]4M4QOSCEQ=8n0K4M41N101N10001N1011N1O100O2O0O1O102M3MmAMZ=2XB6P>87I7I6JXVd5"}, "label": "a giraffe in a zoo looking left"}]}
{"id": 315961, "image": "COCO_train2014_000000315961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest giraffe\"."}], "task": "refering", "answer_template": "The \"the tallest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 46, 47, 69, 70, 71, 93, 94, 95, 117, 118, 119, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 256, 257, 259, 260, 279, 280, 282, 283, 284, 302, 303, 306, 307, 325, 326, 329, 330, 348, 349, 352, 353], "bbox": [0.00375, 0.08258333333333333, 0.39865625000000005, 0.9087916666666666], "iscrowd": 0, "area": 26454.017299999978, "rle": {"size": [480, 640], "counts": "Xo04l>Y1gN000000O101O0O1000000O2O000001Ob0^O0010O0001O00001O000010O3M2N3N1N2N3N1N3M20001O010O0O1O0010O000000102M2N7I=D0O2N1O101N2N3M2N2N3M1O1O1O1O1O0000000000002N3M3M3M3M3M3M2N3N2M3M3M1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UGoKd6R4SIWLh0Db3V4TKhLT1VOi3Y6QLkIP4V6jKnIW4S6TK`Jm4i71O1O1O1O2O0O1O00100O1O0dKQKgNP5X1PLiMP4V2QMiLo2W3^4O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O10O01O1O001O1O001O1O00100O001O001O1O001O001O1O0010O3N2N2M2O2N2bEjLl8j4L5J5L5KO001O010O1O001O001O1O001O1O1O13M4L5K4K5L4L5K4L3M3M0OULiGn0W8lNQHQ1P8jNXHS1g7jNZJXOg5c0]4M4QOSCEQ=8n0K4M41N101N10001N1011N1O100O2O0O1O102M3MmAMZ=2XB6P>87I7I6JXVd5"}, "label": "the tallest giraffe"}]}
{"id": 463417, "image": "COCO_train2014_000000463417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby giraffe standing next to the two larger giraffes\"."}], "task": "refering", "answer_template": "The \"the baby giraffe standing next to the two larger giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 198, 199, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 290, 310, 311, 313, 333, 334, 336], "bbox": [0.50940625, 0.45833723653395786, 0.697234375, 0.958337236533958], "iscrowd": 0, "area": 9727.000350000002, "rle": {"size": [427, 640], "counts": "nYX4d0c<6[MJnG:o71dG5Y8OaG6\\8O]G4a82WG3g83QG0n86iFNW97aFM]9Q2000O010M2jNW1L4M26K6J6Ii0XO1O2M2O2N0OlNWGhMh8T2`GiM`8R2hGkMW8Q2QHlMn7P2ZHlMf7P2bHlM_7o1P2O1N3O01O0001O000010O000001O01O1O010O001O001O001O3M4L8H8H;Dh1YN4L4K2O1O1N2O1N2N2N1O2N2M3N2M3MnMQJmLm5X1gHCj1RO]5[1lHVOU2\\On4_1nHPO\\2^Od4b1SIlN_2_O]4d1WIiNb2_OV4i1ZIbNg2Cn3j1^I]Nn9a1VFZNo9c1UFXNn9e1m0M3N2M3M4M2L5M3000O10000002N101N1O3M5K2N101N2N1O2N2N2N1N3N2Nae`2"}, "label": "the baby giraffe standing next to the two larger giraffes"}]}
{"id": 463417, "image": "COCO_train2014_000000463417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe is displayed on the image along with its parents\"."}], "task": "refering", "answer_template": "The \"a baby giraffe is displayed on the image along with its parents\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 198, 199, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 290, 310, 311, 313, 333, 334, 336], "bbox": [0.50940625, 0.45833723653395786, 0.697234375, 0.958337236533958], "iscrowd": 0, "area": 9727.000350000002, "rle": {"size": [427, 640], "counts": "nYX4d0c<6[MJnG:o71dG5Y8OaG6\\8O]G4a82WG3g83QG0n86iFNW97aFM]9Q2000O010M2jNW1L4M26K6J6Ii0XO1O2M2O2N0OlNWGhMh8T2`GiM`8R2hGkMW8Q2QHlMn7P2ZHlMf7P2bHlM_7o1P2O1N3O01O0001O000010O000001O01O1O010O001O001O001O3M4L8H8H;Dh1YN4L4K2O1O1N2O1N2N2N1O2N2M3N2M3MnMQJmLm5X1gHCj1RO]5[1lHVOU2\\On4_1nHPO\\2^Od4b1SIlN_2_O]4d1WIiNb2_OV4i1ZIbNg2Cn3j1^I]Nn9a1VFZNo9c1UFXNn9e1m0M3N2M3M4M2L5M3000O10000002N101N1O3M5K2N101N2N1O2N2N2N1N3N2Nae`2"}, "label": "a baby giraffe is displayed on the image along with its parents"}]}
{"id": 463417, "image": "COCO_train2014_000000463417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bigger giraffe stretching its neck out\"."}], "task": "refering", "answer_template": "The \"the bigger giraffe stretching its neck out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 108, 109, 112, 113, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 217, 218, 219, 237, 238, 241, 260, 261, 264, 284, 286, 287, 307, 308, 309, 310, 330, 331, 333], "bbox": [0.337359375, 0.1252927400468384, 0.93225, 0.9604918032786885], "iscrowd": 0, "area": 29007.565400000018, "rle": {"size": [427, 640], "counts": "T\\j23T=7J6I6J7J6I7J6I6K6I7J6I7I6K6I7J6I7J5JPOcF]NV9Y1[GgN_8n0QHSOh7c0iH]On6a0\\I_OZ6g0jI\\Ol5g0YJ]O^5c0fJ_OT5a0QK]Om4b0WK]Of4b0_K]O_4b0fK\\OW4d0mK[OQ4d0TLZOi3f0[LYOc3f0bLXO[3h0iLWOU3h0oLVOo2j0VMTOh2k0\\MTOa2l0dMQO[2n0iMPOV2o0nMoNP2Q1UNlNj1W1VNfNj1]1WN`Ng1c1o42N2M3N2N2M2O2N1N2O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1N3N1O0O2O001N2O001O1O0010O0100O0kIQMo1P3[MhMc2W2iL`NT3a1VLWOh3i0SL]Ok3c0SLBj3?SLFk39SLLj35SL0k30RL9g3GWL<h3ETL`0k3_OSLd0l3]OQLf0o3ZOnK\\N3Fo3n1kK\\NX1gNl2n2iK[Na1dNf2Q3fKZNk1bN^2U3cKZNU2^NX2X3aKYN\\2]NR2[5UNbJk1[5[NbJd1[5cNbJ]1W5mNeJR1T5YOjJe0o4n3N200O1O1O1O01101O001O1`0@4L5K3M00001OO1O1O1N2O1O1O1N2O1N2C=J6K4L5N2M3N2N2M3N2M3N2N2M3K5^NXF]Ol9c0YFSOl9l0[FjNi9V1\\F`Ni9`1]FVNg9i1n00000O10O010000O10000O100O10000O10000O100O10000O1000O0100O10000O10000O10000O100O1000O01000O0100O10O10O10O10O10O010000O01000O010O1000O0100O010O10O0100O10O01O10O0100O010O100O010O10O0100O00100O10O0100O010O10O0100O10O0100O00100O010O100O010O10O0100O01L3N3M2O101O0O2O001O001O0011N3N2M3N3L3N0O0010O010O01001N2O1N2O001N2O1N2O1N110O100000O01000O010O10O01O10O0100O00100O001O1O001O1O010O2N1O1O1O1O2N100O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1N3N1ObWb0"}, "label": "the bigger giraffe stretching its neck out"}]}
{"id": 463417, "image": "COCO_train2014_000000463417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest and darkest giraffe standing by a wall and rocks\"."}], "task": "refering", "answer_template": "The \"the tallest and darkest giraffe standing by a wall and rocks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 108, 109, 112, 113, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 217, 218, 219, 237, 238, 241, 260, 261, 264, 284, 286, 287, 307, 308, 309, 310, 330, 331, 333], "bbox": [0.337359375, 0.1252927400468384, 0.93225, 0.9604918032786885], "iscrowd": 0, "area": 29007.565400000018, "rle": {"size": [427, 640], "counts": "T\\j23T=7J6I6J7J6I7J6I6K6I7J6I7I6K6I7J6I7J5JPOcF]NV9Y1[GgN_8n0QHSOh7c0iH]On6a0\\I_OZ6g0jI\\Ol5g0YJ]O^5c0fJ_OT5a0QK]Om4b0WK]Of4b0_K]O_4b0fK\\OW4d0mK[OQ4d0TLZOi3f0[LYOc3f0bLXO[3h0iLWOU3h0oLVOo2j0VMTOh2k0\\MTOa2l0dMQO[2n0iMPOV2o0nMoNP2Q1UNlNj1W1VNfNj1]1WN`Ng1c1o42N2M3N2N2M2O2N1N2O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1N3N1O0O2O001N2O001O1O0010O0100O0kIQMo1P3[MhMc2W2iL`NT3a1VLWOh3i0SL]Ok3c0SLBj3?SLFk39SLLj35SL0k30RL9g3GWL<h3ETL`0k3_OSLd0l3]OQLf0o3ZOnK\\N3Fo3n1kK\\NX1gNl2n2iK[Na1dNf2Q3fKZNk1bN^2U3cKZNU2^NX2X3aKYN\\2]NR2[5UNbJk1[5[NbJd1[5cNbJ]1W5mNeJR1T5YOjJe0o4n3N200O1O1O1O01101O001O1`0@4L5K3M00001OO1O1O1N2O1O1O1N2O1N2C=J6K4L5N2M3N2N2M3N2M3N2N2M3K5^NXF]Ol9c0YFSOl9l0[FjNi9V1\\F`Ni9`1]FVNg9i1n00000O10O010000O10000O100O10000O10000O100O10000O1000O0100O10000O10000O10000O100O1000O01000O0100O10O10O10O10O10O010000O01000O010O1000O0100O010O10O0100O10O01O10O0100O010O100O010O10O0100O00100O10O0100O010O10O0100O10O0100O00100O010O100O010O10O0100O01L3N3M2O101O0O2O001O001O0011N3N2M3N3L3N0O0010O010O01001N2O1N2O001N2O1N2O1N110O100000O01000O010O10O01O10O0100O00100O001O1O001O1O010O2N1O1O1O1O2N100O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1N3N1ObWb0"}, "label": "the tallest and darkest giraffe standing by a wall and rocks"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sits at a table working a computer\"."}], "task": "refering", "answer_template": "The \"a man sits at a table working a computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 102, 103, 104, 119, 120, 121, 122, 136, 137, 138, 139, 153, 154, 155, 156, 170, 172, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 238, 239, 240, 255], "bbox": [0.008958333333333334, 0.21326562500000001, 0.25710416666666663, 0.679640625], "iscrowd": 0, "area": 19043.128, "rle": {"size": [640, 480], "counts": "Xm2;[`0Y3fL[3fMY2101N1O2O0O2gKYFWNg9g1ZFYNh9e1YF[Ng9c1ZF]Nh9a1YF^Ni9_1XFbNh9\\1ZFcNh9[1XFfNh9X1ZFgNh9W1XFjNh9T1ZFkNh9S1XFmNi9n0\\FROe9h0_FXOc9d0`F\\O`9c0`F]Ob9`0`F_Oa9`0_F@c9>^FAc9>^FAd9=\\FCe9<\\FCf9>XFAj9?TFAn9`0PF_OR:b0kE^OW:c0gE\\O[:e0cEZO_:g0^EYOd:h0ZEWOh:i0WEVOk:k0REUOQ;j0nDVOS;i0mDVOU;j0iDWOX;i0gDVO[;i0eDWO\\;i0bDWO`;h0`DXOa;h0^DWOd;h0\\DWOf;h0ZDXOg;g0XDYOj;f0VDZOk;e0UDZOm;e0SD[Oo;c0QD\\OQ<c0oC]OR<b0mC_OT<`0lC_OV<`0jC@W<?iCAX<>hCAZ<>fCB[<=dCD]<;cCD_<;aCE`<:`CFa<9_CFd<8[CIf<6ZCIi<5WCKj<4VCKm<3SCMo<2PCMS=1lB0U=OkB0X=NhB2Z=LfB3]=KcB4`=J`B5b=J^B6d=H\\B7g=GYB8i=FXB9j=FVB:k=EUB:m=DUB:m=ESB:o=EQB:Q>DPB<Q>CoA<T>BlA=V>AkA=X>BiA<Y>BhA=Z>BfA=\\>BdA<_>BbA=`>AaA>a>@`A>c>_O`A?c>^O_A`0c>^O_A?d>^O^Aa0d>]O^A`0e>^O]A`0e>^O]A?f>@ZA?h>^O[A?h>]O\\Aa0f>ZO_Ac0e>XO^Ag0o`0N2M3M2N3L5L4Lgin6"}, "label": "a man sits at a table working a computer"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black laptop which is open and a person is typing on it\"."}], "task": "refering", "answer_template": "The \"a black laptop which is open and a person is typing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 126, 127, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210], "bbox": [0.14841666666666667, 0.280390625, 0.4978541666666666, 0.55525], "iscrowd": 0, "area": 15845.301150000001, "rle": {"size": [640, 480], "counts": "ae\\14nc08C2N3M3M3M3L4M2N3M3M3M3N20O1000000001O1O1O1O100O1`^O^Na`0b1]_O_Nc`0a1]_O_Nd`0`1[_OaNe`0_1Z_ObNg`0]1Y_OcNg`0]1X_OdN`?Lo@_1@fNa?Lh@d1G`Na?Mf@d1H`Nb?Me@c1HaNc?Md@b1IaNc?Oa@a1@ZNO7P`0O`@`1A\\NM5R`00^@_1C^NK4T`0:S@T1N`NH3W`0=n_OR11`NI1X`0a0j_OP12kNT`0V2j_OjMV`0P30001O00000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O1O100O1O100O1O1O1O1O1O1O1O2N1O1O1O1O00O1000000O1000000O100000000O1000000O1000000O100000O01000000O100000000O100000UO__OXN``0^1P@]NP`0W1a@dN_?Q1PAkNP?j0`AQO_>e0QBVOo=?n2Di^f4"}, "label": "a black laptop which is open and a person is typing on it"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black laptop with blue screen\"."}], "task": "refering", "answer_template": "The \"black laptop with blue screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 126, 127, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210], "bbox": [0.14841666666666667, 0.280390625, 0.4978541666666666, 0.55525], "iscrowd": 0, "area": 15845.301150000001, "rle": {"size": [640, 480], "counts": "ae\\14nc08C2N3M3M3M3L4M2N3M3M3M3N20O1000000001O1O1O1O100O1`^O^Na`0b1]_O_Nc`0a1]_O_Nd`0`1[_OaNe`0_1Z_ObNg`0]1Y_OcNg`0]1X_OdN`?Lo@_1@fNa?Lh@d1G`Na?Mf@d1H`Nb?Me@c1HaNc?Md@b1IaNc?Oa@a1@ZNO7P`0O`@`1A\\NM5R`00^@_1C^NK4T`0:S@T1N`NH3W`0=n_OR11`NI1X`0a0j_OP12kNT`0V2j_OjMV`0P30001O00000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O1O100O1O100O1O1O1O1O1O1O1O2N1O1O1O1O00O1000000O1000000O100000000O1000000O1000000O100000O01000000O100000000O100000UO__OXN``0^1P@]NP`0W1a@dN_?Q1PAkNP?j0`AQO_>e0QBVOo=?n2Di^f4"}, "label": "black laptop with blue screen"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands of the person on lower right corner\"."}], "task": "refering", "answer_template": "The \"hands of the person on lower right corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 235, 236, 237, 253, 254, 269, 271, 286, 287, 288, 303, 304, 305, 320, 321, 322], "bbox": [0.816625, 0.552125, 1.0, 0.8255156250000001], "iscrowd": 0, "area": 8249.089049999999, "rle": {"size": [640, 480], "counts": "j]e74lc04K2N2O100[]OFba0;R^ODC5[b07o]OGC5^b05j]O8Tb0Ii]O:Wb0Ff]O=Zb0Cc]O`0\\b0c000e]ObNRb0g100N2O1O1O1N2N1O2N200O1jMRNXBo1d=VNZBk1a=YN_Bh1]=[NcBf1[=[NeBe1Z=\\NfBe1X=[NiBf1U=ZNlBg1Q=[NoBf1o<ZNSCe1l<[NUCe1k<XNlAF:R2j=VNkAK;o1i=TNkA1=j1h=TNhA6`0f1h=SNdA<d0a1h=RN^Ad0j0Z1g=RN^Af0l0W1f=RN^Ag0m0W1e=QN^Ag0o0X1c=QN]Ag0R1W1a=SN\\Ae0T1X1a=YO_Bf0c=XO_Bg0b=WO_Bi0b=VO^Bj0c=TO^Bl0b=TO_Bk0b=SO[AeNj0X2k=SOZAhNi0T2o=ROYAlNg0Q2P>ROYAPOe0n1S>oNZATOb0m1T>nNZAVOc0k1T>mNZAXOb0k1T>kN\\AZO`0j1T>lN\\A[O`0i1T>jN^A\\O`0i1R>jN^A^O`0h1Q>kN_A]Oa0g1P>kN_A_Oa0f1P>jN^AAd0c1o=kNZAFh0^1n=lNVAJl0Z1o=kNSALP1X1n=jNRAOQ1V1n=jNPA0T1V1l=iNn@3V1T1m=gNm@6W1S1k=DWB;i=EXB:h=FYB:f=F[B9e=G]B8b=H_B7a=I`B7_=IbB6_=IbB7^=HcB7^=HcB8]=GdB9\\=cNiCOiM=k2"}, "label": "hands of the person on lower right corner"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hands of a person working on part of a computer\"."}], "task": "refering", "answer_template": "The \"the hands of a person working on part of a computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 235, 236, 237, 253, 254, 269, 271, 286, 287, 288, 303, 304, 305, 320, 321, 322], "bbox": [0.816625, 0.552125, 1.0, 0.8255156250000001], "iscrowd": 0, "area": 8249.089049999999, "rle": {"size": [640, 480], "counts": "j]e74lc04K2N2O100[]OFba0;R^ODC5[b07o]OGC5^b05j]O8Tb0Ii]O:Wb0Ff]O=Zb0Cc]O`0\\b0c000e]ObNRb0g100N2O1O1O1N2N1O2N200O1jMRNXBo1d=VNZBk1a=YN_Bh1]=[NcBf1[=[NeBe1Z=\\NfBe1X=[NiBf1U=ZNlBg1Q=[NoBf1o<ZNSCe1l<[NUCe1k<XNlAF:R2j=VNkAK;o1i=TNkA1=j1h=TNhA6`0f1h=SNdA<d0a1h=RN^Ad0j0Z1g=RN^Af0l0W1f=RN^Ag0m0W1e=QN^Ag0o0X1c=QN]Ag0R1W1a=SN\\Ae0T1X1a=YO_Bf0c=XO_Bg0b=WO_Bi0b=VO^Bj0c=TO^Bl0b=TO_Bk0b=SO[AeNj0X2k=SOZAhNi0T2o=ROYAlNg0Q2P>ROYAPOe0n1S>oNZATOb0m1T>nNZAVOc0k1T>mNZAXOb0k1T>kN\\AZO`0j1T>lN\\A[O`0i1T>jN^A\\O`0i1R>jN^A^O`0h1Q>kN_A]Oa0g1P>kN_A_Oa0f1P>jN^AAd0c1o=kNZAFh0^1n=lNVAJl0Z1o=kNSALP1X1n=jNRAOQ1V1n=jNPA0T1V1l=iNn@3V1T1m=gNm@6W1S1k=DWB;i=EXB:h=FYB:f=F[B9e=G]B8b=H_B7a=I`B7_=IbB6_=IbB7^=HcB7^=HcB8]=GdB9\\=cNiCOiM=k2"}, "label": "the hands of a person working on part of a computer"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white laptop computer\"."}], "task": "refering", "answer_template": "The \"a white laptop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 93, 94, 110, 111, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 178, 179, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235], "bbox": [0.48093749999999996, 0.184953125, 0.8720833333333334, 0.607328125], "iscrowd": 0, "area": 27205.222049999997, "rle": {"size": [640, 480], "counts": "\\d`49^c0<D=C=C<D=C=C=C<D=C=C<D=C=C=D;D=G90O1000000O10000O2N1L4M3L4L4M3L4M3L5L3L4M3L4M1N01O00000010O00000000001O0001O000001O000000000010O0000000001O000001O01O00000000001O10O01O001O1O001O1O010O1O001O1O00000001O000O1000O1N1O2N2N2N2O0O2N2N2O10O1000O10000O10000O1000000000001O00000000000000000001O0000001O00001O0000001O00001O0000001O000011N7I7I7J6I7I7J6I7I7J6J6J5K6J6J6J6J6J6J6J6J6J6JR\\V1"}, "label": "a white laptop computer"}]}
{"id": 356922, "image": "COCO_train2014_000000356922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white laptop\"."}], "task": "refering", "answer_template": "The \"a white laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 93, 94, 110, 111, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 178, 179, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235], "bbox": [0.48093749999999996, 0.184953125, 0.8720833333333334, 0.607328125], "iscrowd": 0, "area": 27205.222049999997, "rle": {"size": [640, 480], "counts": "\\d`49^c0<D=C=C<D=C=C=C<D=C=C<D=C=C=D;D=G90O1000000O10000O2N1L4M3L4L4M3L4M3L5L3L4M3L4M1N01O00000010O00000000001O0001O000001O000000000010O0000000001O000001O01O00000000001O10O01O001O1O001O1O010O1O001O1O00000001O000O1000O1N1O2N2N2N2O0O2N2N2O10O1000O10000O10000O1000000000001O00000000000000000001O0000001O00001O0000001O00001O0000001O000011N7I7I7J6I7I7J6I7I7J6J6J5K6J6J6J6J6J6J6J6J6J6JR\\V1"}, "label": "a white laptop"}]}
{"id": 184889, "image": "COCO_train2014_000000184889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and blue bag\"."}], "task": "refering", "answer_template": "The \"a red and blue bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 135, 136, 137, 138, 153, 154, 155, 156, 170, 171, 172, 173], "bbox": [0.45042000000000004, 0.538855421686747, 0.70094, 0.7935843373493977], "iscrowd": 0, "area": 4766.015950000002, "rle": {"size": [332, 500], "counts": "\\VY21X:3O11O00001O00O100O10000O100O11O000001O01O0001O00010O000ZOKUG6i8LWG3h8OXG0g82XGNh84WGKh87WGIh89XGFg8=WGCh8?XG@h8b0VG^Oi8d0UG]Oj8f0TGZOk8i0SGWOl8o0nFROR9T1O1O1OO1N3N11O2N2N101N2N2N1O2O0O2N2N1N3N1N2N2O1N2O1O1O1O001O00001O000O2O00010O0100O10O10O102M3N1O0O100O100O2O1O1N2O1N2O2M2N2M3N2M2ESGVOo8h0<O1N101N2O1N101N2N3N1N2N2N2N4Ld]`1"}, "label": "a red and blue bag"}]}
{"id": 184889, "image": "COCO_train2014_000000184889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman stands beside a train\"."}], "task": "refering", "answer_template": "The \"a woman stands beside a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 36, 37, 54, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 162, 163, 164, 181, 182], "bbox": [0.0134, 0.08771084337349398, 0.14292000000000002, 0.8679819277108434], "iscrowd": 0, "area": 11388.472249999999, "rle": {"size": [332, 500], "counts": "[\\2;b9`0ZGYOV7X1eHQOT7R1fHUOU7n0gHWOU7n0dHXOY7P2L4L4M4K5K8I7gJSLe3U4lK]LP3AiLZ4IgLn2XOWMg5_2aJ_Mc5\\2`JbMc5[2`JbMd5Y2cJ`Ma5]2Y1M101O1O001O001O00000001O000010O0001O00003M4L3M4VIWMi5l2oIZMP6j2gI]Mi45fKZ3T4lLjKV3P4PMnKR3l3SMSLo2g3VMXLk2c3YM\\Lj2^3[MaLg2Y3^MfLd2T3`MlLb2m2dMRM]2h2hMWM[2a2lM^MV2[2PNdMR2U2SNkMQ24]KQ1h2jNP2KeKn0`2WOm2=WM@T33QMFYQ[4"}, "label": "a woman stands beside a train"}]}
{"id": 184889, "image": "COCO_train2014_000000184889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman was standing\"."}], "task": "refering", "answer_template": "The \"a woman was standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 36, 37, 54, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 162, 163, 164, 181, 182], "bbox": [0.0134, 0.08771084337349398, 0.14292000000000002, 0.8679819277108434], "iscrowd": 0, "area": 11388.472249999999, "rle": {"size": [332, 500], "counts": "[\\2;b9`0ZGYOV7X1eHQOT7R1fHUOU7n0gHWOU7n0dHXOY7P2L4L4M4K5K8I7gJSLe3U4lK]LP3AiLZ4IgLn2XOWMg5_2aJ_Mc5\\2`JbMc5[2`JbMd5Y2cJ`Ma5]2Y1M101O1O001O001O00000001O000010O0001O00003M4L3M4VIWMi5l2oIZMP6j2gI]Mi45fKZ3T4lLjKV3P4PMnKR3l3SMSLo2g3VMXLk2c3YM\\Lj2^3[MaLg2Y3^MfLd2T3`MlLb2m2dMRM]2h2hMWM[2a2lM^MV2[2PNdMR2U2SNkMQ24]KQ1h2jNP2KeKn0`2WOm2=WM@T33QMFYQ[4"}, "label": "a woman was standing"}]}
{"id": 520767, "image": "COCO_train2014_000000520767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child in blue shirt\"."}], "task": "refering", "answer_template": "The \"child in blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 194, 212, 213, 216, 217], "bbox": [0.167921875, 0.26480093676814986, 0.48278125000000005, 0.6670491803278689], "iscrowd": 0, "area": 13086.703750000002, "rle": {"size": [427, 640], "counts": "oV]13W=3M3M2N3M2N3M3L3N3M2N3M2N3M3M2M4M3M3M3M3M2N3L4M3M3M3M3M3M2N3L4M3M3M3M3M3M2N3L5L4L4L4L4L3M4K5L4L1O1O1O1O2N1N2O1O1O1O1O1O1O2MiMQHVOn7l0VHoNi7S1\\HeNe7\\1`H]Nb7d1bHUN_7l1eHnM]7S2fHgM]7Y2fHaM]7^2gH\\M[7e2iHTMZ7k2jHoLY7Q3a100O100O1000O0100O10000O100O010O10000O100O10O01O1O105Jb0^Ob0_O<C001O010O0010O01O0010O01O010O001O010O00010O001O010O0010O01O0010O01O010O001O010O0010O01O0010O01O010O001O10O01O06K7H8I00O001O001O001O001O1O001N101O001N2N101N101N101N2_ObC2_<NaC2`<MaC1`<NaC2`<MaC1`<O`C1a<N`C0b<N_C2a<N`C0b<O_COb<1^COc<O_COb<1^COc<0<0_VZ4"}, "label": "child in blue shirt"}]}
{"id": 520767, "image": "COCO_train2014_000000520767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with a blue shirt feeding a giraffe\"."}], "task": "refering", "answer_template": "The \"a child with a blue shirt feeding a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 194, 212, 213, 216, 217], "bbox": [0.167921875, 0.26480093676814986, 0.48278125000000005, 0.6670491803278689], "iscrowd": 0, "area": 13086.703750000002, "rle": {"size": [427, 640], "counts": "oV]13W=3M3M2N3M2N3M3L3N3M2N3M2N3M3M2M4M3M3M3M3M2N3L4M3M3M3M3M3M2N3L4M3M3M3M3M3M2N3L5L4L4L4L4L3M4K5L4L1O1O1O1O2N1N2O1O1O1O1O1O1O2MiMQHVOn7l0VHoNi7S1\\HeNe7\\1`H]Nb7d1bHUN_7l1eHnM]7S2fHgM]7Y2fHaM]7^2gH\\M[7e2iHTMZ7k2jHoLY7Q3a100O100O1000O0100O10000O100O010O10000O100O10O01O1O105Jb0^Ob0_O<C001O010O0010O01O0010O01O010O001O010O00010O001O010O0010O01O0010O01O010O001O010O0010O01O0010O01O010O001O10O01O06K7H8I00O001O001O001O001O1O001N101O001N2N101N101N101N2_ObC2_<NaC2`<MaC1`<NaC2`<MaC1`<O`C1a<N`C0b<N_C2a<N`C0b<O_COb<1^COc<O_COb<1^COc<0<0_VZ4"}, "label": "a child with a blue shirt feeding a giraffe"}]}
{"id": 520767, "image": "COCO_train2014_000000520767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on the left\"."}], "task": "refering", "answer_template": "The \"man on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.003, 0.009133489461358314, 0.32984375, 0.9776814988290398], "iscrowd": 0, "area": 67469.25685, "rle": {"size": [427, 640], "counts": "el0R1`;V2kMV2[Nd1cN]1cN]1cN]1cN]1cN4L00000000000O10000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000O1000000000000000001O000000000000000000000000000000000000000O1000000000000000000000000000000000000000000kH_O`1a0_NC^1=`NH]19aNJ]16aNO\\11bN4[1LdN8Y1HeN=X1CgNa0U1@kNc0R1]OmNg0P1YOoNk0n0UOQOo0l0QOTOR1i0nNVOV1g0jNXOZ1e0fNZO^1c0bN]Oa1`0_N_Oe1>[NAi1<WNDl19TNFP27oMIU24kMKY22gMN\\2OdM0`2M`M2d2K\\M5g2HYM7k2FUM9o2DQM;S3BmL=W3@iL?Z3_OfLa0]3\\OcLc0a3ZO_Le0e3XO[Lg0i3VOWLi0m3TOSLl0P4POQLo0S4nNmKQ1W4lNhKT1\\4iNdKV1`4gN`KX1d4eN\\K[1g4bNYK]1k4`NUK_1o4^NQKa1R5]NnJb1V5[NjJe1Y5XNgJg1]5VNcJi1a5TN_Jk1e5RN[Jm1i5PNWJo1m5nMSJR2P6kMPJT2T6iMlIV2W6hMiIW2Y6hMgIW2\\6gMcIZ2^6eMbIZ2a6dM_I[2d6bM]I]2e6bM[I]2h6aMXI^2j6aMVI_2l6_MTI`2o6^MQIa2Q7^MoHa2T7]MlHb2W7\\MiHd2X7[MhHd2[7ZMeHe2]7YMdHf2_7XMaHg2b7WM^Hh2d7WM\\Hi2h7SMXHl2n7oLQHQ3U8jLkGU3[8fLeGV3d8eL\\GV3_9L4K5K5L4K6J<@`0A?@`0@gkb5"}, "label": "man on the left"}]}
{"id": 520767, "image": "COCO_train2014_000000520767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses that is holding a child\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses that is holding a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.003, 0.009133489461358314, 0.32984375, 0.9776814988290398], "iscrowd": 0, "area": 67469.25685, "rle": {"size": [427, 640], "counts": "el0R1`;V2kMV2[Nd1cN]1cN]1cN]1cN]1cN4L00000000000O10000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000O1000000000000000001O000000000000000000000000000000000000000O1000000000000000000000000000000000000000000kH_O`1a0_NC^1=`NH]19aNJ]16aNO\\11bN4[1LdN8Y1HeN=X1CgNa0U1@kNc0R1]OmNg0P1YOoNk0n0UOQOo0l0QOTOR1i0nNVOV1g0jNXOZ1e0fNZO^1c0bN]Oa1`0_N_Oe1>[NAi1<WNDl19TNFP27oMIU24kMKY22gMN\\2OdM0`2M`M2d2K\\M5g2HYM7k2FUM9o2DQM;S3BmL=W3@iL?Z3_OfLa0]3\\OcLc0a3ZO_Le0e3XO[Lg0i3VOWLi0m3TOSLl0P4POQLo0S4nNmKQ1W4lNhKT1\\4iNdKV1`4gN`KX1d4eN\\K[1g4bNYK]1k4`NUK_1o4^NQKa1R5]NnJb1V5[NjJe1Y5XNgJg1]5VNcJi1a5TN_Jk1e5RN[Jm1i5PNWJo1m5nMSJR2P6kMPJT2T6iMlIV2W6hMiIW2Y6hMgIW2\\6gMcIZ2^6eMbIZ2a6dM_I[2d6bM]I]2e6bM[I]2h6aMXI^2j6aMVI_2l6_MTI`2o6^MQIa2Q7^MoHa2T7]MlHb2W7\\MiHd2X7[MhHd2[7ZMeHe2]7YMdHf2_7XMaHg2b7WM^Hh2d7WM\\Hi2h7SMXHl2n7oLQHQ3U8jLkGU3[8fLeGV3d8eL\\GV3_9L4K5K5L4K6J<@`0A?@`0@gkb5"}, "label": "a man wearing glasses that is holding a child"}]}
{"id": 381509, "image": "COCO_train2014_000000381509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a cap on a rail with a skating board\"."}], "task": "refering", "answer_template": "The \"a man wearing a cap on a rail with a skating board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 286, 287], "bbox": [0.2403125, 0.08747072599531616, 0.6089687500000001, 0.8324355971896956], "iscrowd": 0, "area": 21577.0518, "rle": {"size": [427, 640], "counts": "S\\P22Y=2N1N2O1O2M10000O101O0O10O10O1000O01O1O1O001O1O001O1O1O001N101O00001O00001O01O010O00O2M2O1N2L5N1O1O1O1O1O1O2N1O1O1O1O1O1000M2M4L4L4O12M2O1O2N1O2M2O1N3N2M2K6H8M2M4ZFeMW8_2fGdMV8`2gGcMV8`2hGaMW8a2gGaMW8a2gGaMW8b2gG_MW8c2gG^MW8e2gG]MW8U3SGQMk8h301N101eHPLY5P4aIXM\\6m4O2M200O2O0O2O0O101N100O2N100001O1OiJ[Ig3e6XL[Ij3d6UL\\Im3b6SL^IP4`6oK`IT4^6kKbIZ4Z6eKgIa4S6]KnIc4S6\\KmId4U6ZKkIf4V6XKkIh4V6WKjIi4W6VKiIj4X6TKiIl4X6SKhIm4Y6QKhIo4Z6nJgIR5\\6iJfIX5\\6dJeI\\5U7SK[HT4f7hKaHT4_7hKfHW4Z7fKjHY4V7fKlHY4U7eKlH[4T7dKmH]4R7bKoH^4R7`KoH`4n70ZH`Ka6a4^IcK^6]4aIhKZ6Y4fIlKU6T4jIQLR6P4mISLP6m3PJULn5k3QJXLm5h3SJZLk5f3UJ\\Li5e3UJ_Lg5b3YJ`Le5`3[JbLc5^3\\JeLb5\\3]JfLa5Z3_JgL`5Y3_JjL_5V3aJlL\\5U3dJlLW5X3hJmLh1ZODm3dNQMX1]ONe3jNWMg0^O9_3oN?j0EVO>c0F]O<=HB<7HI;0I09KJ4`0_ODa0c601N10000O1000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O100O100O100O1O100O100O100O100O1O100O100O100O10`UX3"}, "label": "a man wearing a cap on a rail with a skating board"}]}
{"id": 381509, "image": "COCO_train2014_000000381509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in a plaid shirt doing tricks on a skateboard\"."}], "task": "refering", "answer_template": "The \"a young man in a plaid shirt doing tricks on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 286, 287], "bbox": [0.2403125, 0.08747072599531616, 0.6089687500000001, 0.8324355971896956], "iscrowd": 0, "area": 21577.0518, "rle": {"size": [427, 640], "counts": "S\\P22Y=2N1N2O1O2M10000O101O0O10O10O1000O01O1O1O001O1O001O1O1O001N101O00001O00001O01O010O00O2M2O1N2L5N1O1O1O1O1O1O2N1O1O1O1O1O1000M2M4L4L4O12M2O1O2N1O2M2O1N3N2M2K6H8M2M4ZFeMW8_2fGdMV8`2gGcMV8`2hGaMW8a2gGaMW8a2gGaMW8b2gG_MW8c2gG^MW8e2gG]MW8U3SGQMk8h301N101eHPLY5P4aIXM\\6m4O2M200O2O0O2O0O101N100O2N100001O1OiJ[Ig3e6XL[Ij3d6UL\\Im3b6SL^IP4`6oK`IT4^6kKbIZ4Z6eKgIa4S6]KnIc4S6\\KmId4U6ZKkIf4V6XKkIh4V6WKjIi4W6VKiIj4X6TKiIl4X6SKhIm4Y6QKhIo4Z6nJgIR5\\6iJfIX5\\6dJeI\\5U7SK[HT4f7hKaHT4_7hKfHW4Z7fKjHY4V7fKlHY4U7eKlH[4T7dKmH]4R7bKoH^4R7`KoH`4n70ZH`Ka6a4^IcK^6]4aIhKZ6Y4fIlKU6T4jIQLR6P4mISLP6m3PJULn5k3QJXLm5h3SJZLk5f3UJ\\Li5e3UJ_Lg5b3YJ`Le5`3[JbLc5^3\\JeLb5\\3]JfLa5Z3_JgL`5Y3_JjL_5V3aJlL\\5U3dJlLW5X3hJmLh1ZODm3dNQMX1]ONe3jNWMg0^O9_3oN?j0EVO>c0F]O<=HB<7HI;0I09KJ4`0_ODa0c601N10000O1000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O100O100O100O1O100O100O100O100O1O100O100O100O10`UX3"}, "label": "a young man in a plaid shirt doing tricks on a skateboard"}]}
{"id": 545351, "image": "COCO_train2014_000000545351.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the one with black cap , brown pant\"."}], "task": "refering", "answer_template": "The \"the one with black cap , brown pant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 155, 156, 169, 170, 171, 184, 185, 186, 199, 200, 201, 214, 215, 216, 229, 230, 231, 244, 245, 246, 259, 260, 261, 275, 276], "bbox": [0.2714953271028037, 0.40909375, 0.46623831775700936, 0.8120156250000001], "iscrowd": 0, "area": 13770.375150000003, "rle": {"size": [640, 428], "counts": "PkX22ic08J6J4M2M4M2N2N2N3M2M3N2N2S^OfN]O4[a0X1T_OXOd`0j0Y_OUOW2F`;W1UBSOe1`0T<>UBQOi1d0n;>TBPO?Cb0R1j<<QBoNb0L=l0n<m0dB^N7j0U=d0cBjNMPOFf1h==dBUOFh0f=0bBe1\\=XNcBn1\\=nMcBW2]=iM^B]2a=cMZBb2e=n1O001O0N3M2M4M2N3M2O2O001N101M2HRI]CP7[<>O1N10000OO3O11O0O2O1O1O1N2O001O1O100OcKTDf0m;XOXDe0g;[O\\Dc0c;\\O_Dc0b;\\O^De0a;nLRDXOMj1a0o1b;kLVDVOIo1a0n1b;kLZDV14l0]OVNW<EZDX13k0B]M27Q<7ZDV13d0[OiM<KL9S<5YDY12c0?kM^;LYDd1KdNYOc1S?IZBJk=5^B_Of=`0bBUOb=j0fBjN^=V1i2M2[Of0VOeV^4"}, "label": "the one with black cap , brown pant"}]}
{"id": 545351, "image": "COCO_train2014_000000545351.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a flannel ski jacket\"."}], "task": "refering", "answer_template": "The \"a woman in a flannel ski jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 155, 156, 169, 170, 171, 184, 185, 186, 199, 200, 201, 214, 215, 216, 229, 230, 231, 244, 245, 246, 259, 260, 261, 275, 276], "bbox": [0.2714953271028037, 0.40909375, 0.46623831775700936, 0.8120156250000001], "iscrowd": 0, "area": 13770.375150000003, "rle": {"size": [640, 428], "counts": "PkX22ic08J6J4M2M4M2N2N2N3M2M3N2N2S^OfN]O4[a0X1T_OXOd`0j0Y_OUOW2F`;W1UBSOe1`0T<>UBQOi1d0n;>TBPO?Cb0R1j<<QBoNb0L=l0n<m0dB^N7j0U=d0cBjNMPOFf1h==dBUOFh0f=0bBe1\\=XNcBn1\\=nMcBW2]=iM^B]2a=cMZBb2e=n1O001O0N3M2M4M2N3M2O2O001N101M2HRI]CP7[<>O1N10000OO3O11O0O2O1O1O1N2O001O1O100OcKTDf0m;XOXDe0g;[O\\Dc0c;\\O_Dc0b;\\O^De0a;nLRDXOMj1a0o1b;kLVDVOIo1a0n1b;kLZDV14l0]OVNW<EZDX13k0B]M27Q<7ZDV13d0[OiM<KL9S<5YDY12c0?kM^;LYDd1KdNYOc1S?IZBJk=5^B_Of=`0bBUOb=j0fBjN^=V1i2M2[Of0VOeV^4"}, "label": "a woman in a flannel ski jacket"}]}
{"id": 545351, "image": "COCO_train2014_000000545351.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skier with a yellow and black jacket and black pants standing next to a woman\"."}], "task": "refering", "answer_template": "The \"a man skier with a yellow and black jacket and black pants standing next to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 157, 158, 159, 173, 174, 187, 188, 189, 201, 202, 203, 204, 216, 217, 218, 219, 232, 233, 234, 247, 248, 249, 263, 264, 278, 279, 280, 293, 294, 295, 309, 310], "bbox": [0.4190654205607477, 0.413109375, 0.6913317757009345, 0.890015625], "iscrowd": 0, "area": 16365.135700000003, "rle": {"size": [640, 428], "counts": "ek`37hc06J6I2O2M2001O01O000000001O0000001O000000001O0001O01O0000001O000000l0TO4L4L5K4L5K4L4UM_MPDe2i;bMRDb2h;dMeCP3T<WM_CU3\\<PM]CW3]<oL\\CW3_<nL\\CU3a<PMYCU3d<oLVCU3i<nLQCW3m<lLmBY3Q=l1O2N1N3N1O2N2N1O2M2O2N1O2N1O2N2M2cDWHb:j7YEZHf:g7UE]Hl:d7oD`HP;a7kDdHT;U800O10O0100O1O10O0100O100O1O0101O1eLkDhMU;T2REiMo:Q2YEkMi:P2]EnMc:n1dEoM]:l1jEQNW:i1QFTNo9e1ZFYNg9b0`GZOa8EcH8^7YORIe0o6lNaIR1_6`NQJ^1P6YNZJd1h5TNaJi1`5PNiJl1i;M3L4M4K5L5J5L5C<B\\oa2"}, "label": "a man skier with a yellow and black jacket and black pants standing next to a woman"}]}
{"id": 545351, "image": "COCO_train2014_000000545351.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"human with a yellow jacket standing on top of skiis\"."}], "task": "refering", "answer_template": "The \"human with a yellow jacket standing on top of skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 157, 158, 159, 173, 174, 187, 188, 189, 201, 202, 203, 204, 216, 217, 218, 219, 232, 233, 234, 247, 248, 249, 263, 264, 278, 279, 280, 293, 294, 295, 309, 310], "bbox": [0.4190654205607477, 0.413109375, 0.6913317757009345, 0.890015625], "iscrowd": 0, "area": 16365.135700000003, "rle": {"size": [640, 428], "counts": "ek`37hc06J6I2O2M2001O01O000000001O0000001O000000001O0001O01O0000001O000000l0TO4L4L5K4L5K4L4UM_MPDe2i;bMRDb2h;dMeCP3T<WM_CU3\\<PM]CW3]<oL\\CW3_<nL\\CU3a<PMYCU3d<oLVCU3i<nLQCW3m<lLmBY3Q=l1O2N1N3N1O2N2N1O2M2O2N1O2N1O2N2M2cDWHb:j7YEZHf:g7UE]Hl:d7oD`HP;a7kDdHT;U800O10O0100O1O10O0100O100O1O0101O1eLkDhMU;T2REiMo:Q2YEkMi:P2]EnMc:n1dEoM]:l1jEQNW:i1QFTNo9e1ZFYNg9b0`GZOa8EcH8^7YORIe0o6lNaIR1_6`NQJ^1P6YNZJd1h5TNaJi1`5PNiJl1i;M3L4M4K5L5J5L5C<B\\oa2"}, "label": "human with a yellow jacket standing on top of skiis"}]}
{"id": 127560, "image": "COCO_train2014_000000127560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder jumping off ledge\"."}], "task": "refering", "answer_template": "The \"a skateboarder jumping off ledge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 73, 74, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 167, 168, 169, 170, 184, 185, 186, 201, 202], "bbox": [0.39564935064935064, 0.17781249999999998, 0.8807575757575757, 0.554265625], "iscrowd": 0, "area": 18356.429700000004, "rle": {"size": [640, 462], "counts": "\\`b31mc04L3N2M3N2O100O1O100O1O100O00100O00100O0100000O0o^O_Od>a0XABi>>SAEn>:n@JS?6i@NW?1e@3\\?La@7`?I[@;e?EY@=g?CV@a0i?_OT@d0l?\\OR@f0e0kN]=?kAi0f0kN]==jAl0e0kN`=9iAn0d0lNc=6fAR1d0kNe=3dAU1d0lNf=0dAW1b0mNi=LbA[1b0kNl=J_A_1a0kNo=F^Aa1`0mNQ>GWA`1d0lNT>S2iAPNV>P2gATNX>m1dAWN[>j1bAXN^>h1_A\\N_>f1]A^Nb>`301O1O1N100000O1O0A`0O1O1O1O1O1O1O1O1O1O1OoA_Ke<`4[CaKe<^4[CdKd<[4\\CfKd<Y4\\CiKc<V4]CkKc<T4]CnKb<P4_CQLa<k3bCWL]<d3gC]LY<^3kCdLT<[3lCfLT<Z3jCiLU<W3jCjLV<W3hCkLW<U3hCmLW<S3hCnLX<R3ZCWLWOi0^=P3ZC[LTOf0b=o2YC`Mg<_2YCaMg<`2WCbMi<]2VCdMg<_2XCcMa<c2^C^M[<i2dCYMU<m2kCSMT<o2jCSMU<c1UCnMe0`0V<`1XCoMa0c0U<]1\\CPN>d0V<[1^CPN;g0V<W1aCRN9g0U<W1cCUN4f0X<S1gC]NIa0`<Q1hCg0V<XOlCh0T<VOoCM]OPNb<S2RDM]OoMa<R2UDN\\OoM^<S2VDM^OoM[<T2XDKAPNV<T2[DIBRNS<U2[DGFSNm;V2^DEHSNk;X2^DALUNf;Y2_D@NUNd;[2^D^O2UNa;[2_D]O3WN_;[2^D\\O7VN\\;]2_DZO8XNZ;j0]C?Q1=<WNX;j0^C`0o0=>XNU;h0bCc0k0:b0YNR;g0eCd0g0:_=oNmAg0c08b=oNnAh0a06d=oNoAj0<5h=nNoAl094j=mNQBo050n=nNQBP110Q>mNQBR1ONR>nNRBT1KLb?2^@Ld?5\\@Hg?7Y@Gj?9U@Em?:U@_OR`0>P@\\OW`0a0l_OWO[`0g0a1N10100O00100O0010O01O10O01O10O01O10O01O010O1O01O0000000000O10001O000000000O100000000000001N10000000000000000O100000001O00000O11O00001O001O001O01O01O001O001O00I7IWVR1"}, "label": "a skateboarder jumping off ledge"}]}
{"id": 127560, "image": "COCO_train2014_000000127560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in air\"."}], "task": "refering", "answer_template": "The \"boy in air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 73, 74, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 167, 168, 169, 170, 184, 185, 186, 201, 202], "bbox": [0.39564935064935064, 0.17781249999999998, 0.8807575757575757, 0.554265625], "iscrowd": 0, "area": 18356.429700000004, "rle": {"size": [640, 462], "counts": "\\`b31mc04L3N2M3N2O100O1O100O1O100O00100O00100O0100000O0o^O_Od>a0XABi>>SAEn>:n@JS?6i@NW?1e@3\\?La@7`?I[@;e?EY@=g?CV@a0i?_OT@d0l?\\OR@f0e0kN]=?kAi0f0kN]==jAl0e0kN`=9iAn0d0lNc=6fAR1d0kNe=3dAU1d0lNf=0dAW1b0mNi=LbA[1b0kNl=J_A_1a0kNo=F^Aa1`0mNQ>GWA`1d0lNT>S2iAPNV>P2gATNX>m1dAWN[>j1bAXN^>h1_A\\N_>f1]A^Nb>`301O1O1N100000O1O0A`0O1O1O1O1O1O1O1O1O1O1OoA_Ke<`4[CaKe<^4[CdKd<[4\\CfKd<Y4\\CiKc<V4]CkKc<T4]CnKb<P4_CQLa<k3bCWL]<d3gC]LY<^3kCdLT<[3lCfLT<Z3jCiLU<W3jCjLV<W3hCkLW<U3hCmLW<S3hCnLX<R3ZCWLWOi0^=P3ZC[LTOf0b=o2YC`Mg<_2YCaMg<`2WCbMi<]2VCdMg<_2XCcMa<c2^C^M[<i2dCYMU<m2kCSMT<o2jCSMU<c1UCnMe0`0V<`1XCoMa0c0U<]1\\CPN>d0V<[1^CPN;g0V<W1aCRN9g0U<W1cCUN4f0X<S1gC]NIa0`<Q1hCg0V<XOlCh0T<VOoCM]OPNb<S2RDM]OoMa<R2UDN\\OoM^<S2VDM^OoM[<T2XDKAPNV<T2[DIBRNS<U2[DGFSNm;V2^DEHSNk;X2^DALUNf;Y2_D@NUNd;[2^D^O2UNa;[2_D]O3WN_;[2^D\\O7VN\\;]2_DZO8XNZ;j0]C?Q1=<WNX;j0^C`0o0=>XNU;h0bCc0k0:b0YNR;g0eCd0g0:_=oNmAg0c08b=oNnAh0a06d=oNoAj0<5h=nNoAl094j=mNQBo050n=nNQBP110Q>mNQBR1ONR>nNRBT1KLb?2^@Ld?5\\@Hg?7Y@Gj?9U@Em?:U@_OR`0>P@\\OW`0a0l_OWO[`0g0a1N10100O00100O0010O01O10O01O10O01O10O01O010O1O01O0000000000O10001O000000000O100000000000001N10000000000000000O100000001O00000O11O00001O001O001O01O01O001O001O00I7IWVR1"}, "label": "boy in air"}]}
{"id": 127560, "image": "COCO_train2014_000000127560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid with a green shirt and jeans\"."}], "task": "refering", "answer_template": "The \"a kid with a green shirt and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 194, 195, 209, 210, 211, 225, 226, 227, 228, 241, 242, 243, 244, 257, 258, 259, 273, 274, 289, 290, 305, 306, 321, 322], "bbox": [0.07636363636363637, 0.5019375, 0.2970995670995671, 0.894140625], "iscrowd": 0, "area": 11764.180849999999, "rle": {"size": [640, 462], "counts": "R]f04ac0=C<D<D=B=D=oMYNZAS2R>X2K6I7J6J6I7M2N100O100O100O10O0100O1O1O10O01O001WOkIgCT6X<XJ]Ci5b<k0N101N1O2O1O0100]OdCbI\\<V6mChIS<R6UDmIk;l5\\DSJe;f5cDYJ\\;`5mD_JS;Z5TEeJm:U5ZEjJe:P5cEoJ]:k4iETKX:V1_CT2a2eLo9X1aCl1g2kLh9X1bCf1m2QMb9Y1aC`1e?`N]@X1j?gNW@>d`0B\\_O=e`0@__O=d`0@^_O?c`0^Oa_O`0``0]Oc_Ob0^`0\\Od_Ob0^`0[Of_Oc0[`0[Oh_Oc0ka0O2M3N3M3O00100000000O1000000O1000000O0100000N1O2M3N2N1O2M3N2N1O2M3N2N1O^fZ6"}, "label": "a kid with a green shirt and jeans"}]}
{"id": 127560, "image": "COCO_train2014_000000127560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy watching a man skeing\"."}], "task": "refering", "answer_template": "The \"a boy watching a man skeing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 194, 195, 209, 210, 211, 225, 226, 227, 228, 241, 242, 243, 244, 257, 258, 259, 273, 274, 289, 290, 305, 306, 321, 322], "bbox": [0.07636363636363637, 0.5019375, 0.2970995670995671, 0.894140625], "iscrowd": 0, "area": 11764.180849999999, "rle": {"size": [640, 462], "counts": "R]f04ac0=C<D<D=B=D=oMYNZAS2R>X2K6I7J6J6I7M2N100O100O100O10O0100O1O1O10O01O001WOkIgCT6X<XJ]Ci5b<k0N101N1O2O1O0100]OdCbI\\<V6mChIS<R6UDmIk;l5\\DSJe;f5cDYJ\\;`5mD_JS;Z5TEeJm:U5ZEjJe:P5cEoJ]:k4iETKX:V1_CT2a2eLo9X1aCl1g2kLh9X1bCf1m2QMb9Y1aC`1e?`N]@X1j?gNW@>d`0B\\_O=e`0@__O=d`0@^_O?c`0^Oa_O`0``0]Oc_Ob0^`0\\Od_Ob0^`0[Of_Oc0[`0[Oh_Oc0ka0O2M3N3M3O00100000000O1000000O1000000O0100000N1O2M3N2N1O2M3N2N1O2M3N2N1O^fZ6"}, "label": "a boy watching a man skeing"}]}
{"id": 389705, "image": "COCO_train2014_000000389705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby zerba walking to the left of the other zebras\"."}], "task": "refering", "answer_template": "The \"the baby zerba walking to the left of the other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 98, 99, 100, 101, 109, 110, 111, 112, 121, 122, 123, 124, 134, 135, 136, 146, 147, 148, 158, 159, 160, 170, 171, 172], "bbox": [0.14426426426426425, 0.39327999999999996, 0.4467567567567568, 0.7962], "iscrowd": 0, "area": 13034.3727, "rle": {"size": [500, 333], "counts": "Ufg08Z??@`0A8G6K4K6J6K5J6L4L4L4M3L:F;E:F8H5K4L5K5K4L5K5L3L3N1O001N100ZOfJRG\\4l9iKoEQ4X:nKgEe4g9\\KXFS5Y9mJgFT5Y9lJfFT5Z9mJdFT5\\9lJdFT5]9lJbFT5^9lJbFT5^9mJ`FU5`9kJ_FU5a9kJ_FU5a9lJ^FT5c9>0H8H9G8H8H9G8c0]Oi0XOO01O01O01O010O000N3M2N3N1N1OL4M5K5J6K5K6J5J6K5L4Kd0]O4K5L5J5WNZD_Ok;=oDiNU;R1j1K4K5L4L5J5M300O2N3M2N2N2N2NRdi2"}, "label": "the baby zerba walking to the left of the other zebras"}]}
{"id": 389705, "image": "COCO_train2014_000000389705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small zebra standing on the left side in the picture\"."}], "task": "refering", "answer_template": "The \"the small zebra standing on the left side in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 98, 99, 100, 101, 109, 110, 111, 112, 121, 122, 123, 124, 134, 135, 136, 146, 147, 148, 158, 159, 160, 170, 171, 172], "bbox": [0.14426426426426425, 0.39327999999999996, 0.4467567567567568, 0.7962], "iscrowd": 0, "area": 13034.3727, "rle": {"size": [500, 333], "counts": "Ufg08Z??@`0A8G6K4K6J6K5J6L4L4L4M3L:F;E:F8H5K4L5K5K4L5K5L3L3N1O001N100ZOfJRG\\4l9iKoEQ4X:nKgEe4g9\\KXFS5Y9mJgFT5Y9lJfFT5Z9mJdFT5\\9lJdFT5]9lJbFT5^9lJbFT5^9mJ`FU5`9kJ_FU5a9kJ_FU5a9lJ^FT5c9>0H8H9G8H8H9G8c0]Oi0XOO01O01O01O010O000N3M2N3N1N1OL4M5K5J6K5K6J5J6K5L4Kd0]O4K5L5J5WNZD_Ok;=oDiNU;R1j1K4K5L4L5J5M300O2N3M2N2N2N2NRdi2"}, "label": "the small zebra standing on the left side in the picture"}]}
{"id": 389705, "image": "COCO_train2014_000000389705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 45, 46, 47, 57, 58, 59, 68, 69, 70, 71, 80, 81, 82, 83, 93, 94, 95, 105, 106, 107, 117, 118, 119, 130, 131, 142, 143, 154, 155, 166, 167, 178, 179], "bbox": [0.7286186186186187, 0.13152, 1.0, 0.78912], "iscrowd": 0, "area": 18868.9813, "rle": {"size": [500, 333], "counts": "mif3:S?<C<H8J6I7J6J7I6J6J6J6J6J6J4K5L3M4L4L4L4L4L4M3M3M2N3M3M3M3M3M3M3M2N3QIZKk2i4kL_KT3d4bLeK\\3^4ZLjKe3Y4QLoKn3T4gKULX4n3XK`Lg4b3iJmLV5V3YJYMf5[6O1O1O1O100O100O10000O10000O100O2O000O010O10jMbJVJ^5]5SK_Jl4U5eKgJ[4V5lKfJT4Y5RLbJm3]5YL^Jh3a5]L[Jc3c5cLYJ]3f5hLVJX3h5oLSJP3m5UMoIk2o5[MmIe2Q6aMkI_2T6gMfIZ2X6lMdIT2[6c3N2O0O2N2O1N2O1N[2fMfK"}, "label": "the zebra on the right"}]}
{"id": 389705, "image": "COCO_train2014_000000389705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mother zebra is leading her babies\"."}], "task": "refering", "answer_template": "The \"the mother zebra is leading her babies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 45, 46, 47, 57, 58, 59, 68, 69, 70, 71, 80, 81, 82, 83, 93, 94, 95, 105, 106, 107, 117, 118, 119, 130, 131, 142, 143, 154, 155, 166, 167, 178, 179], "bbox": [0.7286186186186187, 0.13152, 1.0, 0.78912], "iscrowd": 0, "area": 18868.9813, "rle": {"size": [500, 333], "counts": "mif3:S?<C<H8J6I7J6J7I6J6J6J6J6J6J4K5L3M4L4L4L4L4L4M3M3M2N3M3M3M3M3M3M3M2N3QIZKk2i4kL_KT3d4bLeK\\3^4ZLjKe3Y4QLoKn3T4gKULX4n3XK`Lg4b3iJmLV5V3YJYMf5[6O1O1O1O100O100O10000O10000O100O2O000O010O10jMbJVJ^5]5SK_Jl4U5eKgJ[4V5lKfJT4Y5RLbJm3]5YL^Jh3a5]L[Jc3c5cLYJ]3f5hLVJX3h5oLSJP3m5UMoIk2o5[MmIe2Q6aMkI_2T6gMfIZ2X6lMdIT2[6c3N2O0O2N2O1N2O1N[2fMfK"}, "label": "the mother zebra is leading her babies"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower vase with number of flower place beside other flower vase on the table\"."}], "task": "refering", "answer_template": "The \"a flower vase with number of flower place beside other flower vase on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 312, 313, 314, 334, 335, 336, 337, 358, 359, 360], "bbox": [0.39934375, 0.06493750000000001, 0.9311718750000001, 0.9194791666666667], "iscrowd": 0, "area": 60255.92349999997, "rle": {"size": [480, 640], "counts": "hSh3g0Q2@o9a0PF@P:a0nEAP:a0nE@Q:a0nEAP:a0nE@Q:b0mE@R:`0mEAR:a0lEAR:`0mEAR:a0jECT:?gEFX::dEJ[:8`EM^:5]EOb:2ZE3d:OWE5i:LSE8k:IQE;n:GmD>Q;DjD`0U;g1O0O2N1O2N1O2O0O2N1O2N1O2O0O2N1O2N1O100O1O1O1O1O100O1O1`FZKe8f4[GdK[8]4eGcKZ8^4fGbKZ8^4eGbK[8_4eGaKZ8`4fG`KY8a4fG`KY8a4gG^KY8c4fG^KZ8b4fG^KY8c4gG\\KY8e4fG\\KY8d4hG\\KW8e4iGZKW8g4hGZKX8f4hGZKW8g4iGXKW8i4hGXKW8i4iGWKV8k4iGUKV8l4iGTKW8m4iGSKV8o4iGQKV8P5iGPKX8Q5gGoJX8R5gGoJX8R5hGmJX8U5gGkJX8V5gGkJX8W5gGhJY8Y5gGgJX8[5fGfJZ8Z5fGfJZ8Z5fGeJ]8Z5aGgJa8W5_GiJc8V5\\GiJf8V5YGkJi8S5WGmJk8R5TGmJn8R5QGoJQ9_521N2O2N1N3N1O2M2O2N1N3fJbIj1a6SN`Ik1e6PN\\In1i6nMYIP2l6kMUIS2P7iMQIV2T7eMnHX2W7dMjH[2[7aMeH^2_7^MbH^2f7]M\\H]2l7_MUH]2S8^MnG]2Z8_MhG\\2`8_MaG]2f8`MZG[2o8`MRG]2T9_MnF_2W9\\MjFc2Z9YMgFe2_9VMcFh2a9UM_Fj2f9QM[Fn2i9nLXFP3n:0O2N13N6J6I7J04I;F9F2O1N2N2O1N2O1N2O0O2N2[FXLP8i3lG[LR8g3jG\\LV8e3fG_LX8c3eG`LY8b3cGbL\\8_3`GdL_8]3^GgLa8Z3[GjLc8X3YGkLg8V3UGnLj8S3RGQMl8Q3QGQMo8a4O1O00N2N2N2N2M3N2N2M3N`NeJ_IY5a6kJ]IR5d6RKYIm4g6WKWIg4h6^KVI`4j6dKTIY4m6lKPIR4o6SLoHk3Q7YLmHd3T7`LjH^3V7fLhHX3W7mLgHQ3Y7SMeHj2\\7ZMbHd2]7bM`H\\2`7hM^HU2c7oM[HQ2c7SN[Hn1a7WN\\Hj1l0gK`5c2bIf1l0iK`5e2bIc1k0iKa5i2bI^1k0kKa5k2bI[1j0lKb5X7\\JjHb5X7\\JiHb5[7[JgHc5[7[JgHc5[7[JgHb5]7[JdHd5^7ZJdHd5^7ZJdHd5_7YJbHe5a7YJaHe5a7YJaHe5^8M2O2N2N2M2O2N2M3N1O2N2M3N1O2M3N2N1O2O11O0O101O000O2O001O0O101ON2ZOe0K6O1O100O010O1O1jLjDb2U;^MlDa2U;_MkDa2U;^MlDa2U;^MlDa2T;_MmDa2S;_MlDa2U;^MlDa2T;_MmDe2o:ZMREo2d:QM]EX3Y:iLgE_3P:aLQF_3o9`LRF_3n9aLSF^3m9cLRF]3n9cLSF\\3m9dLTF[3m9dLTFZ3m9fLTFY3l9hLTFW3l9iLUFV3l9iLUFV3k9jLVFU3j9lLVFR3k9nLVFQ3k9nLVFQ3j9oLWFP3i9QMWFn2j9QMWFl0\\Of0_:\\NWFm0ZOf0b:ZNUFo0[Od0c:[NSFQ1ZOc0e:[NQFQ1\\Ob0f:ZNoES1]Oa0f:ZNnEU1\\O`0e;^O]D`0d;@\\D`0d;_O^D`0b;@_D?a;@`D`0`;@aD?`;_OaDa0_;_ObD`0^;_OdDa0[;_OeDa0[;^OgDa0Y;^OiDa0W;_OiDa0W;^OkDa0T;@lD`0T;_OnD`0Q;APE>P;AQE`0m:ATE>l:AUE?j:BWE=i:BYE=f:DZE<f:C\\E<c:D^E<b:D_E<_:DbE<^:DaE=^:CcE=]:CbE>^:AcE?\\:BcE?]:@cEa0\\:@dE?]:@cE?^:BbE<d:_O\\E`0j:[OWEc0o:YOPEf0V;UOkDi0[;SOdDk0c;PO^Dn0j<O2N1N2O2M2O2M2O2M2O2M2O2M2O2M2Og^d0"}, "label": "a flower vase with number of flower place beside other flower vase on the table"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple and white flowers in vase\"."}], "task": "refering", "answer_template": "The \"purple and white flowers in vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 312, 313, 314, 334, 335, 336, 337, 358, 359, 360], "bbox": [0.39934375, 0.06493750000000001, 0.9311718750000001, 0.9194791666666667], "iscrowd": 0, "area": 60255.92349999997, "rle": {"size": [480, 640], "counts": "hSh3g0Q2@o9a0PF@P:a0nEAP:a0nE@Q:a0nEAP:a0nE@Q:b0mE@R:`0mEAR:a0lEAR:`0mEAR:a0jECT:?gEFX::dEJ[:8`EM^:5]EOb:2ZE3d:OWE5i:LSE8k:IQE;n:GmD>Q;DjD`0U;g1O0O2N1O2N1O2O0O2N1O2N1O2O0O2N1O2N1O100O1O1O1O1O100O1O1`FZKe8f4[GdK[8]4eGcKZ8^4fGbKZ8^4eGbK[8_4eGaKZ8`4fG`KY8a4fG`KY8a4gG^KY8c4fG^KZ8b4fG^KY8c4gG\\KY8e4fG\\KY8d4hG\\KW8e4iGZKW8g4hGZKX8f4hGZKW8g4iGXKW8i4hGXKW8i4iGWKV8k4iGUKV8l4iGTKW8m4iGSKV8o4iGQKV8P5iGPKX8Q5gGoJX8R5gGoJX8R5hGmJX8U5gGkJX8V5gGkJX8W5gGhJY8Y5gGgJX8[5fGfJZ8Z5fGfJZ8Z5fGeJ]8Z5aGgJa8W5_GiJc8V5\\GiJf8V5YGkJi8S5WGmJk8R5TGmJn8R5QGoJQ9_521N2O2N1N3N1O2M2O2N1N3fJbIj1a6SN`Ik1e6PN\\In1i6nMYIP2l6kMUIS2P7iMQIV2T7eMnHX2W7dMjH[2[7aMeH^2_7^MbH^2f7]M\\H]2l7_MUH]2S8^MnG]2Z8_MhG\\2`8_MaG]2f8`MZG[2o8`MRG]2T9_MnF_2W9\\MjFc2Z9YMgFe2_9VMcFh2a9UM_Fj2f9QM[Fn2i9nLXFP3n:0O2N13N6J6I7J04I;F9F2O1N2N2O1N2O1N2O0O2N2[FXLP8i3lG[LR8g3jG\\LV8e3fG_LX8c3eG`LY8b3cGbL\\8_3`GdL_8]3^GgLa8Z3[GjLc8X3YGkLg8V3UGnLj8S3RGQMl8Q3QGQMo8a4O1O00N2N2N2N2M3N2N2M3N`NeJ_IY5a6kJ]IR5d6RKYIm4g6WKWIg4h6^KVI`4j6dKTIY4m6lKPIR4o6SLoHk3Q7YLmHd3T7`LjH^3V7fLhHX3W7mLgHQ3Y7SMeHj2\\7ZMbHd2]7bM`H\\2`7hM^HU2c7oM[HQ2c7SN[Hn1a7WN\\Hj1l0gK`5c2bIf1l0iK`5e2bIc1k0iKa5i2bI^1k0kKa5k2bI[1j0lKb5X7\\JjHb5X7\\JiHb5[7[JgHc5[7[JgHc5[7[JgHb5]7[JdHd5^7ZJdHd5^7ZJdHd5_7YJbHe5a7YJaHe5a7YJaHe5^8M2O2N2N2M2O2N2M3N1O2N2M3N1O2M3N2N1O2O11O0O101O000O2O001O0O101ON2ZOe0K6O1O100O010O1O1jLjDb2U;^MlDa2U;_MkDa2U;^MlDa2U;^MlDa2T;_MmDa2S;_MlDa2U;^MlDa2T;_MmDe2o:ZMREo2d:QM]EX3Y:iLgE_3P:aLQF_3o9`LRF_3n9aLSF^3m9cLRF]3n9cLSF\\3m9dLTF[3m9dLTFZ3m9fLTFY3l9hLTFW3l9iLUFV3l9iLUFV3k9jLVFU3j9lLVFR3k9nLVFQ3k9nLVFQ3j9oLWFP3i9QMWFn2j9QMWFl0\\Of0_:\\NWFm0ZOf0b:ZNUFo0[Od0c:[NSFQ1ZOc0e:[NQFQ1\\Ob0f:ZNoES1]Oa0f:ZNnEU1\\O`0e;^O]D`0d;@\\D`0d;_O^D`0b;@_D?a;@`D`0`;@aD?`;_OaDa0_;_ObD`0^;_OdDa0[;_OeDa0[;^OgDa0Y;^OiDa0W;_OiDa0W;^OkDa0T;@lD`0T;_OnD`0Q;APE>P;AQE`0m:ATE>l:AUE?j:BWE=i:BYE=f:DZE<f:C\\E<c:D^E<b:D_E<_:DbE<^:DaE=^:CcE=]:CbE>^:AcE?\\:BcE?]:@cEa0\\:@dE?]:@cE?^:BbE<d:_O\\E`0j:[OWEc0o:YOPEf0V;UOkDi0[;SOdDk0c;PO^Dn0j<O2N1N2O2M2O2M2O2M2O2M2O2M2O2M2Og^d0"}, "label": "purple and white flowers in vase"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant that is 2nd from the right\"."}], "task": "refering", "answer_template": "The \"plant that is 2nd from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 189, 190, 191, 192, 193, 213, 214, 215, 236, 237, 238, 258, 259, 260, 261, 282, 283, 284], "bbox": [0.011796875, 0.058416666666666665, 0.456734375, 0.7370833333333333], "iscrowd": 0, "area": 52419.453749999986, "rle": {"size": [480, 640], "counts": "\\n33k>3N2M4L3M3N2M4L3M3N2O1O2N1N2O1O2N1O1PNhNVFY1g9lNVFT1i9QOSFo0k9VORFj0l9\\OPFd0o9@nE`0P:DnE<P:GoE9P:IoE7o9MoE3P:OoE1P:1oEOP:4nELQ:6nEJP:9oEGP:<nEDQ:>nEB_8eNVHk1ZO@_8iNSHj1\\O]O`8lNQHm1ZOWOd8POnGn1ZOROg8SOlGQ2XOlNk8WOiGS2WOfNo8ZOgGU2VOaNR9^OdG`3[8cLbG^3]8eL`G\\3^8iL]GY3b8jL[GV3e8nLWGS3h8PMUGQ3j8c1O001O1001N1000000O10001O0O100000000O2O00000O10001O0O1000000O101O000O1000000O1000000O10001O0O1000000O100000000000000010O0000000O10000O101O0O1000000O10001N10000O10O100O010O10O0100O010O10O0100004L3M4L3M4L3M4L3mIgIY3]6bLhIZ3[6bLfI^3^6]LcIc3`6YLaIg3c6TL^Il3e6PL\\IP4h6jKZIW4h6eKYISOZOl4a7lKVIXOZOk4c7iKTI\\OZOj4f7eKQI@[Oj4g7bKoHDZOj4k7]KlHH[Oj4l7ZKjHL[Oi4n7VKiH1ZOh4Q8RKfH5[Oh4R9XKoFg4Q9XKPGh4P9VKSGi4m8UKVGj4a9OO10000O10000O101N101O0O2O001N101O001O001O0O2O001O00001O001N101O001O010O001O001bN_J[Ia5b6cJ]I^5_6fJ`IZ5]6jJcIV5Y6nJfIR5W6RKhIo4T6UKkIk4R6YKmIh4o5\\KPJd4m5`KRJa4k5bKTJ^4k5dKUJ\\4i5fKVJZ4h5iKWJW4h5kKWJU4h5mKWJR4i5PLVJo3i5TLVJk3j5WLUJh3k5[LRJe3m5^LQJb3o5`LoI`3Q6bLnI]3Q6fLmIZ3S6hLkIX3U6jLiIV3W6lLhIS3W6PMgIQ3X6QMfIQ3X6QMgIo2S5`JfKd2UOn2S5aJeKc2VOm2U5aJcKd2VOm2b2jJ]OHhNc2XOl2a2lJ\\OHgN\\8k1oG\\OLbNW8P2oG\\O5YNn7X2oG^OT9a0nF]OT9a0nF]OT9`0nF_OS9`0oF^OS9`0oF\\OT9b0oF[OT9c0mFZOW9d0kFYOW9e0lFWOX9g0iFVO[9h0gFUO[9i0`2N2N101N101N2O0O2O1N101O1N102M4K4K6J6J5KgQS5"}, "label": "plant that is 2nd from the right"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the flower vase at the left side\"."}], "task": "refering", "answer_template": "The \"the flower vase at the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 189, 190, 191, 192, 193, 213, 214, 215, 236, 237, 238, 258, 259, 260, 261, 282, 283, 284], "bbox": [0.011796875, 0.058416666666666665, 0.456734375, 0.7370833333333333], "iscrowd": 0, "area": 52419.453749999986, "rle": {"size": [480, 640], "counts": "\\n33k>3N2M4L3M3N2M4L3M3N2O1O2N1N2O1O2N1O1PNhNVFY1g9lNVFT1i9QOSFo0k9VORFj0l9\\OPFd0o9@nE`0P:DnE<P:GoE9P:IoE7o9MoE3P:OoE1P:1oEOP:4nELQ:6nEJP:9oEGP:<nEDQ:>nEB_8eNVHk1ZO@_8iNSHj1\\O]O`8lNQHm1ZOWOd8POnGn1ZOROg8SOlGQ2XOlNk8WOiGS2WOfNo8ZOgGU2VOaNR9^OdG`3[8cLbG^3]8eL`G\\3^8iL]GY3b8jL[GV3e8nLWGS3h8PMUGQ3j8c1O001O1001N1000000O10001O0O100000000O2O00000O10001O0O1000000O101O000O1000000O1000000O10001O0O1000000O100000000000000010O0000000O10000O101O0O1000000O10001N10000O10O100O010O10O0100O010O10O0100004L3M4L3M4L3M4L3mIgIY3]6bLhIZ3[6bLfI^3^6]LcIc3`6YLaIg3c6TL^Il3e6PL\\IP4h6jKZIW4h6eKYISOZOl4a7lKVIXOZOk4c7iKTI\\OZOj4f7eKQI@[Oj4g7bKoHDZOj4k7]KlHH[Oj4l7ZKjHL[Oi4n7VKiH1ZOh4Q8RKfH5[Oh4R9XKoFg4Q9XKPGh4P9VKSGi4m8UKVGj4a9OO10000O10000O101N101O0O2O001N101O001O001O0O2O001O00001O001N101O001O010O001O001bN_J[Ia5b6cJ]I^5_6fJ`IZ5]6jJcIV5Y6nJfIR5W6RKhIo4T6UKkIk4R6YKmIh4o5\\KPJd4m5`KRJa4k5bKTJ^4k5dKUJ\\4i5fKVJZ4h5iKWJW4h5kKWJU4h5mKWJR4i5PLVJo3i5TLVJk3j5WLUJh3k5[LRJe3m5^LQJb3o5`LoI`3Q6bLnI]3Q6fLmIZ3S6hLkIX3U6jLiIV3W6lLhIS3W6PMgIQ3X6QMfIQ3X6QMgIo2S5`JfKd2UOn2S5aJeKc2VOm2U5aJcKd2VOm2b2jJ]OHhNc2XOl2a2lJ\\OHgN\\8k1oG\\OLbNW8P2oG\\O5YNn7X2oG^OT9a0nF]OT9a0nF]OT9`0nF_OS9`0oF^OS9`0oF\\OT9b0oF[OT9c0mFZOW9d0kFYOW9e0lFWOX9g0iFVO[9h0gFUO[9i0`2N2N101N101N2O0O2O1N101O1N102M4K4K6J6J5KgQS5"}, "label": "the flower vase at the left side"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green vase with purple flowers to the left of another green vase with purple flowers\"."}], "task": "refering", "answer_template": "The \"a green vase with purple flowers to the left of another green vase with purple flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 213, 214, 215, 236, 237, 259, 260, 261, 282, 283, 284], "bbox": [0.230625, 0.39960416666666665, 0.38295312499999995, 0.7427708333333333], "iscrowd": 0, "area": 9292.50685, "rle": {"size": [480, 640], "counts": "XbU23m>3M4L3M3M3L5L3M3M3M4L3M3M3M4L3L4gDYNS9j1iFWNW9m1dFUN[9n1aFSN_9P2]FQNc9R2ZFoMe9U2VFlMj9W2RFkMm9X2oEiMQ:Z2kEgMT:^2hEbMX:a2dE`M\\:c2`E]Ma:f2[E[Me:i2VEXMj:k2SEUMb0Da9j3]F[La9e3]F^Lb9b3[FcLb9]3\\FgLc9Y3ZFjLf9V3WFlLj9S3TFoLm9Q3PFVL0f0Q:T3PFoLQ:P3mERMT:m30O1O001O1O1O1O100O1O1O001O1O1M3N2O11O1O2N1N10001O0000QObE\\M^:`2gE_MY:]2lEbMT:Y2SFfMl9V2YFiMg9S2^FlMb9o1dFPN\\9k1kFSNU9i1PGVNP9e1VGZNj8a1\\G_Nc8]1cGaN]8Z1iGeNV8W1PHhNP8T1UHkNk7P1\\HnNd7m0bHRO^7j0gHVOX7e0nHZOR7a0UI]Ok6?ZI@f6;`ID`67fIHZ64kILT6OSJO`:O1O1O[jh5"}, "label": "a green vase with purple flowers to the left of another green vase with purple flowers"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green plastic vase on the left holding a bunch of purple and white flowers\"."}], "task": "refering", "answer_template": "The \"a green plastic vase on the left holding a bunch of purple and white flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 213, 214, 215, 236, 237, 259, 260, 261, 282, 283, 284], "bbox": [0.230625, 0.39960416666666665, 0.38295312499999995, 0.7427708333333333], "iscrowd": 0, "area": 9292.50685, "rle": {"size": [480, 640], "counts": "XbU23m>3M4L3M3M3L5L3M3M3M4L3M3M3M4L3L4gDYNS9j1iFWNW9m1dFUN[9n1aFSN_9P2]FQNc9R2ZFoMe9U2VFlMj9W2RFkMm9X2oEiMQ:Z2kEgMT:^2hEbMX:a2dE`M\\:c2`E]Ma:f2[E[Me:i2VEXMj:k2SEUMb0Da9j3]F[La9e3]F^Lb9b3[FcLb9]3\\FgLc9Y3ZFjLf9V3WFlLj9S3TFoLm9Q3PFVL0f0Q:T3PFoLQ:P3mERMT:m30O1O001O1O1O1O100O1O1O001O1O1M3N2O11O1O2N1N10001O0000QObE\\M^:`2gE_MY:]2lEbMT:Y2SFfMl9V2YFiMg9S2^FlMb9o1dFPN\\9k1kFSNU9i1PGVNP9e1VGZNj8a1\\G_Nc8]1cGaN]8Z1iGeNV8W1PHhNP8T1UHkNk7P1\\HnNd7m0bHRO^7j0gHVOX7e0nHZOR7a0UI]Ok6?ZI@f6;`ID`67fIHZ64kILT6OSJO`:O1O1O[jh5"}, "label": "a green plastic vase on the left holding a bunch of purple and white flowers"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green color flower vase on a table\"."}], "task": "refering", "answer_template": "The \"a green color flower vase on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 338, 359], "bbox": [0.538109375, 0.5268958333333333, 0.73990625, 0.92825], "iscrowd": 0, "area": 14168.585699999996, "rle": {"size": [480, 640], "counts": "T`Q51o>3M4L4L4ZEBb6a0ZI@f6d0TI^Ol6f0oH[OQ7h0kHYOU7k0eHWO[7m0`HUO_7n0\\HTOd7P1WHQOi7S1QHoNo7U1lGlNT8W1hGjNX8Z1bGiN]8[1^GfNb8]1ZGdNf8`1TGbNl8b1oF_NQ9e1jF\\NU9h1fFZNZ9j1aFXN^9l1\\FVNd9m1XFTNh9P2SFQNm9S2mEoMS:U2hElMX:_3O0000000000001O00000000001O000O10c0]O4L4L3M4L1OO1O1O1O1O1O1O1O1N2N2N2N2M3N2N2N2O10O10001O000000001O00000000_NUFnMj9o1[FoMe9o1_FoMa9n1dFPN\\9m1iFRNV9l1nFRNR9k1SGSNm8k1WGSNi8j1\\GTNd8j1`GUN_8h1fGVNZ8h1jGVNV8g1oGWNP8h1UHUNk7h1ZHWNe7g1_HWNa7f1dHXN\\7f1hHXNX7e1mHZNR7d1RIZNn6c1WI[Ni6c1[I[Ne6b1`I\\N`6b1eI\\NZ6a1kI]NU6a1PJ\\No5b1i3N2M3M3N2M3N2M3N1N3N2M3N2M3N2M3N1N3N2M3NUQ^2"}, "label": "a green color flower vase on a table"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower vase closest to the edge of the table lengthwise\"."}], "task": "refering", "answer_template": "The \"a flower vase closest to the edge of the table lengthwise\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 338, 359], "bbox": [0.538109375, 0.5268958333333333, 0.73990625, 0.92825], "iscrowd": 0, "area": 14168.585699999996, "rle": {"size": [480, 640], "counts": "T`Q51o>3M4L4L4ZEBb6a0ZI@f6d0TI^Ol6f0oH[OQ7h0kHYOU7k0eHWO[7m0`HUO_7n0\\HTOd7P1WHQOi7S1QHoNo7U1lGlNT8W1hGjNX8Z1bGiN]8[1^GfNb8]1ZGdNf8`1TGbNl8b1oF_NQ9e1jF\\NU9h1fFZNZ9j1aFXN^9l1\\FVNd9m1XFTNh9P2SFQNm9S2mEoMS:U2hElMX:_3O0000000000001O00000000001O000O10c0]O4L4L3M4L1OO1O1O1O1O1O1O1O1N2N2N2N2M3N2N2N2O10O10001O000000001O00000000_NUFnMj9o1[FoMe9o1_FoMa9n1dFPN\\9m1iFRNV9l1nFRNR9k1SGSNm8k1WGSNi8j1\\GTNd8j1`GUN_8h1fGVNZ8h1jGVNV8g1oGWNP8h1UHUNk7h1ZHWNe7g1_HWNa7f1dHXN\\7f1hHXNX7e1mHZNR7d1RIZNn6c1WI[Ni6c1[I[Ne6b1`I\\N`6b1eI\\NZ6a1kI]NU6a1PJ\\No5b1i3N2M3M3N2M3N2M3N1N3N2M3N2M3N2M3N1N3N2M3NUQ^2"}, "label": "a flower vase closest to the edge of the table lengthwise"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a willow green resin chair next to the table with flowers\"."}], "task": "refering", "answer_template": "The \"a willow green resin chair next to the table with flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228], "bbox": [0.8241875000000001, 0.19354166666666667, 1.0, 0.58925], "iscrowd": 0, "area": 9830.923450000006, "rle": {"size": [480, 640], "counts": "aVg71o>0O2O0O1O2O0O100O2N100O101N1O100O2O0O1O2O0O1fB]OW<c0jC]OT<e0kC\\OS<e0nC\\OP<c0QD^Om;b0UD^Oj;b0VD_Oh;a0ZD_Od;b0\\D_Oc;`0_DB\\;`0dDBX;?jDCQ;?oDCm:>UECh:>XEEc:<_EE]:=cEFX:;jEFR:;oEHl9:UFGh99YFIc99^FI]98dFJX98iFJR97oFKm87TGJi86WGMe85YGOc83[G1b80[G5c8K[G9b8G]G=`8D^G?`8B]G`0c8A[G`0e8AYG?h8AWG`0i8ATGa0l8@RG`0o8AnFa0R9@cFj0]9VOZFR1g9S2000O10000O2O1O001N101O0O2O001N2O001N101O0O2O001N2O001N101O0O2O1O1N2O001N2O1O1O1N2O1O0O2O1L4K_2aMd0[O`J"}, "label": "a willow green resin chair next to the table with flowers"}]}
{"id": 528970, "image": "COCO_train2014_000000528970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green plastic chair\"."}], "task": "refering", "answer_template": "The \"green plastic chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228], "bbox": [0.8241875000000001, 0.19354166666666667, 1.0, 0.58925], "iscrowd": 0, "area": 9830.923450000006, "rle": {"size": [480, 640], "counts": "aVg71o>0O2O0O1O2O0O100O2N100O101N1O100O2O0O1O2O0O1fB]OW<c0jC]OT<e0kC\\OS<e0nC\\OP<c0QD^Om;b0UD^Oj;b0VD_Oh;a0ZD_Od;b0\\D_Oc;`0_DB\\;`0dDBX;?jDCQ;?oDCm:>UECh:>XEEc:<_EE]:=cEFX:;jEFR:;oEHl9:UFGh99YFIc99^FI]98dFJX98iFJR97oFKm87TGJi86WGMe85YGOc83[G1b80[G5c8K[G9b8G]G=`8D^G?`8B]G`0c8A[G`0e8AYG?h8AWG`0i8ATGa0l8@RG`0o8AnFa0R9@cFj0]9VOZFR1g9S2000O10000O2O1O001N101O0O2O001N2O001N101O0O2O001N2O001N101O0O2O1O1N2O001N2O1O1O1N2O1O0O2O1L4K_2aMd0[O`J"}, "label": "green plastic chair"}]}
{"id": 217676, "image": "COCO_train2014_000000217676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red pick up truck in front of a police motorcycle\"."}], "task": "refering", "answer_template": "The \"a red pick up truck in front of a police motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 306, 307, 308], "bbox": [0.248875, 0.4714591439688716, 0.5135, 0.7645719844357977], "iscrowd": 0, "area": 14382.268399999997, "rle": {"size": [514, 640], "counts": "hR`23n?4PAV1a<kN\\CZ1a<hN\\C[1c<eN[C^1c<cNZC`1e<bNXCa1f<`NWCc1h<_NUCc1k<]NSCf1l<ZNRCh1n<YNoBi1Q=WNmBl1R=b00001O000001O1O2N1O1OO1N1O1O1O100O100PCSMk<T3NIWCYMe<T3N2M3N2M3M4M2M4M3L4N2M3N2N2N00O1000O010000O10000O010O1O1O100O1O010O1O1O1O1N2O0O2000000]NRDROn;n0SDQOm;n0UDQOk;o0VDPOi;P1YDoNg;Q1ZDnNf;Q1\\DnNd;R1]DmNc;R1_DlNb;T1^DlNb;S1`DlN_;U1bDjN_;T1bDlN^;T1cDkN];T1dDlN\\;T1eDkN[;U1eDkN[;T1gDkNY;U1hDjNY;T1hDlNX;S1kDkNU;S1oDkNQ;T1SEiNm:V1VEhNj:W1YEgNg:X1\\EfNe:X1_EeNa:Y1dEdN\\:[1iEaNW:^1mE_NS:`1RF\\Nn9c1VFZNk9d1Z2O1O100O1O101N1O100O1O101N100001O000001O01O00000000010O0000000010O0001O00000010O000001O0001O0H9G:G8N2O2M2N2N3N1O1O2N1O1OPkk4"}, "label": "a red pick up truck in front of a police motorcycle"}]}
{"id": 217676, "image": "COCO_train2014_000000217676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red truck\"."}], "task": "refering", "answer_template": "The \"a red truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 306, 307, 308], "bbox": [0.248875, 0.4714591439688716, 0.5135, 0.7645719844357977], "iscrowd": 0, "area": 14382.268399999997, "rle": {"size": [514, 640], "counts": "hR`23n?4PAV1a<kN\\CZ1a<hN\\C[1c<eN[C^1c<cNZC`1e<bNXCa1f<`NWCc1h<_NUCc1k<]NSCf1l<ZNRCh1n<YNoBi1Q=WNmBl1R=b00001O000001O1O2N1O1OO1N1O1O1O100O100PCSMk<T3NIWCYMe<T3N2M3N2M3M4M2M4M3L4N2M3N2N2N00O1000O010000O10000O010O1O1O100O1O010O1O1O1O1N2O0O2000000]NRDROn;n0SDQOm;n0UDQOk;o0VDPOi;P1YDoNg;Q1ZDnNf;Q1\\DnNd;R1]DmNc;R1_DlNb;T1^DlNb;S1`DlN_;U1bDjN_;T1bDlN^;T1cDkN];T1dDlN\\;T1eDkN[;U1eDkN[;T1gDkNY;U1hDjNY;T1hDlNX;S1kDkNU;S1oDkNQ;T1SEiNm:V1VEhNj:W1YEgNg:X1\\EfNe:X1_EeNa:Y1dEdN\\:[1iEaNW:^1mE_NS:`1RF\\Nn9c1VFZNk9d1Z2O1O100O1O101N1O100O1O101N100001O000001O01O00000000010O0000000010O0001O00000010O000001O0001O0H9G:G8N2O2M2N2N3N1O1O2N1O1OPkk4"}, "label": "a red truck"}]}
{"id": 21070, "image": "COCO_train2014_000000021070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella carried by a girl in pink boots\"."}], "task": "refering", "answer_template": "The \"an umbrella carried by a girl in pink boots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 145, 165, 166, 167, 168, 169, 190, 191, 192, 213, 214, 215], "bbox": [0.166734375, 0.3464, 0.356890625, 0.6221555555555556], "iscrowd": 0, "area": 7255.9030500000035, "rle": {"size": [450, 640], "counts": "UV_13n=5K4K6K>B=C01F9N3N2M2N3M20100O0010O01O10O010O0100O010O010O0100O101N101O001O001O001O010O001O1]C]Oe;c0XDAh;?TDDl;=RDDn;=oCFP<:oCGR<9lCIS<S10010O1O3N1N4PElMj9W2RFkMm9X2oEkMP:W2lEjMT:Z2gEhMX:[2dEgM\\:o201O010O10O010O10O010O10O10O100O0100N101N2O0O2N101N2O0O3N1N2N2O1N2O1N3N1N2N2O1N2O1O2M2O1O1N2O1O1O1N101O2M4M4L4oNnCJQWe5"}, "label": "an umbrella carried by a girl in pink boots"}]}
{"id": 21070, "image": "COCO_train2014_000000021070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple umbrella\"."}], "task": "refering", "answer_template": "The \"purple umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 145, 165, 166, 167, 168, 169, 190, 191, 192, 213, 214, 215], "bbox": [0.166734375, 0.3464, 0.356890625, 0.6221555555555556], "iscrowd": 0, "area": 7255.9030500000035, "rle": {"size": [450, 640], "counts": "UV_13n=5K4K6K>B=C01F9N3N2M2N3M20100O0010O01O10O010O0100O010O010O0100O101N101O001O001O001O010O001O1]C]Oe;c0XDAh;?TDDl;=RDDn;=oCFP<:oCGR<9lCIS<S10010O1O3N1N4PElMj9W2RFkMm9X2oEkMP:W2lEjMT:Z2gEhMX:[2dEgM\\:o201O010O10O010O10O010O10O10O100O0100N101N2O0O2N101N2O0O3N1N2N2O1N2O1N3N1N2N2O1N2O1O2M2O1O1N2O1O1O1N101O2M4M4L4oNnCJQWe5"}, "label": "purple umbrella"}]}
{"id": 21070, "image": "COCO_train2014_000000021070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing a black hat and blue shoes , holding a pink umbrella\"."}], "task": "refering", "answer_template": "The \"a little girl wearing a black hat and blue shoes , holding a pink umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 173, 174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 265, 266, 288, 289], "bbox": [0.5374531250000001, 0.3738444444444444, 0.6763125000000001, 0.774511111111111], "iscrowd": 0, "area": 7684.454199999999, "rle": {"size": [450, 640], "counts": "l_g42o=4L3M3M2mLKVH7f7NWH5d7OZH3c70lF\\O;g0f80jF^O<e0h80fFA>b0k80`FDb0?l8O]FGb0>o8OXFIe0;S9n0jFTOU9n0hFTOW9Q1cFQO]9o0bFRO]9P1aFPO`9P1`FoN`9S1^FnNa9j2SOkKaGOKW4d8lKZG30Q4f8\\LYGe3g8\\LWGe3h8\\LWGT3Z9i0010OO10O001O2N1O2N1N3N2N101N2N2N2N2N200O1O1O101N2N101N2N2O1N2J8VM`Em0MSOk:M\\EP1IiNT;5VEQ1X;nNjDR1T;mNnDS1R;kNPEV1o:gNTEZ1j;0010O2O0O5L6I3N0O2O1OO0001O0N2O1N3M2O1M4L5J5Lefj2"}, "label": "a little girl wearing a black hat and blue shoes , holding a pink umbrella"}]}
{"id": 21070, "image": "COCO_train2014_000000021070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this girl is wearing blue shoes\"."}], "task": "refering", "answer_template": "The \"this girl is wearing blue shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 173, 174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 265, 266, 288, 289], "bbox": [0.5374531250000001, 0.3738444444444444, 0.6763125000000001, 0.774511111111111], "iscrowd": 0, "area": 7684.454199999999, "rle": {"size": [450, 640], "counts": "l_g42o=4L3M3M2mLKVH7f7NWH5d7OZH3c70lF\\O;g0f80jF^O<e0h80fFA>b0k80`FDb0?l8O]FGb0>o8OXFIe0;S9n0jFTOU9n0hFTOW9Q1cFQO]9o0bFRO]9P1aFPO`9P1`FoN`9S1^FnNa9j2SOkKaGOKW4d8lKZG30Q4f8\\LYGe3g8\\LWGe3h8\\LWGT3Z9i0010OO10O001O2N1O2N1N3N2N101N2N2N2N2N200O1O1O101N2N101N2N2O1N2J8VM`Em0MSOk:M\\EP1IiNT;5VEQ1X;nNjDR1T;mNnDS1R;kNPEV1o:gNTEZ1j;0010O2O0O5L6I3N0O2O1OO0001O0N2O1N3M2O1M4L5J5Lefj2"}, "label": "this girl is wearing blue shoes"}]}
{"id": 315994, "image": "COCO_train2014_000000315994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a notebook with kittens on it\"."}], "task": "refering", "answer_template": "The \"a notebook with kittens on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 112, 113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153], "bbox": [0.25956, 0.24682666666666667, 0.50898, 0.6760266666666667], "iscrowd": 0, "area": 18644.478349999998, "rle": {"size": [375, 500], "counts": "oj_19R;>A?B<D<D<D<D<C=D<D<D<E;00000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000iNW1dMYRj2"}, "label": "a notebook with kittens on it"}]}
{"id": 315994, "image": "COCO_train2014_000000315994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a post card with the cartoon of two cats and few flowers laying near a hand bag\"."}], "task": "refering", "answer_template": "The \"a post card with the cartoon of two cats and few flowers laying near a hand bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 112, 113, 114, 115, 116, 117, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153], "bbox": [0.25956, 0.24682666666666667, 0.50898, 0.6760266666666667], "iscrowd": 0, "area": 18644.478349999998, "rle": {"size": [375, 500], "counts": "oj_19R;>A?B<D<D<D<D<C=D<D<D<E;00000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000iNW1dMYRj2"}, "label": "a post card with the cartoon of two cats and few flowers laying near a hand bag"}]}
{"id": 373338, "image": "COCO_train2014_000000373338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey long truck with amherst and the canadian flag on the side\"."}], "task": "refering", "answer_template": "The \"grey long truck with amherst and the canadian flag on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 245, 246, 247, 248, 249, 250, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 339], "bbox": [0.20921875, 0.6938173302107729, 0.8780468750000001, 0.9481498829039813], "iscrowd": 0, "area": 32123.275449999997, "rle": {"size": [427, 640], "counts": "ZVh17T=8H000000000000001O00000001O00000001O00000DCjC=V<HeC8Z<=00O101O0O100O2O0O101Oc0]O7I2O1N2N1O001O10O0nNQEIn:1cEE\\:;nE[OR:e0[1O000000cDXOW:h0eE]OZ:c0bEB^:=aED_:<aEE^:;aEG^:9bEG^:9bEH]:8cEI]:6bEL]:4cEL]:4cEM\\:3dEN[:2dEO\\:2cEO]:1bE0]:0bE1^:0^E4a:MZE7g:ITE;l:R100001O0000001O0000001O000O1O2N0O2O1O1O1O1O1E;K5N2N1O2O10000000000O1000O100000O17I6J7]D]NP;o101O1O1O1O1O1N2O0000O100O100O100O100O100O1O0101O001O00001O00001O001O000O100000O100000000O1000000O100N1O2N2O1O1O1O1O1O1O1O001O1O1O1O1N2O1O1b0]O3N1O2N1OM3M3M3M3L301O1O1O1O1O11O1O1N101O1O0UETNW:m1eEVN[:k1aEYN^:h1^E[Nb:e1[E^Ne:U2O0O1000O1O1O1N2O1O1O01000000000001O2N1N3N1O1O2N000000000O0100000000O1000000O010000000O1000O10O100000O10O11O0O10001O0O101O00000O2O00000O2OO100000O1000000O1000000O010000000O1000000O2O000000000000001O00000000000000000000O10000O10000O10000O10000001O1O1O1O1O1O1O1O1O1O00000000O1000000000000000000mNRFiNn9W1XFbNi9_1]FSNj9m1j01O001O00001O001O001O001O00001O001O001O6J5K6J00000000O100000000O1000000O100000000O100O1O1O1O1O100000000O1000000000000O10000000000L4KY1gNcVP1"}, "label": "grey long truck with amherst and the canadian flag on the side"}]}
{"id": 373338, "image": "COCO_train2014_000000373338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an amherst crane truck body\"."}], "task": "refering", "answer_template": "The \"an amherst crane truck body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 245, 246, 247, 248, 249, 250, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 339], "bbox": [0.20921875, 0.6938173302107729, 0.8780468750000001, 0.9481498829039813], "iscrowd": 0, "area": 32123.275449999997, "rle": {"size": [427, 640], "counts": "ZVh17T=8H000000000000001O00000001O00000001O00000DCjC=V<HeC8Z<=00O101O0O100O2O0O101Oc0]O7I2O1N2N1O001O10O0nNQEIn:1cEE\\:;nE[OR:e0[1O000000cDXOW:h0eE]OZ:c0bEB^:=aED_:<aEE^:;aEG^:9bEG^:9bEH]:8cEI]:6bEL]:4cEL]:4cEM\\:3dEN[:2dEO\\:2cEO]:1bE0]:0bE1^:0^E4a:MZE7g:ITE;l:R100001O0000001O0000001O000O1O2N0O2O1O1O1O1O1E;K5N2N1O2O10000000000O1000O100000O17I6J7]D]NP;o101O1O1O1O1O1N2O0000O100O100O100O100O100O1O0101O001O00001O00001O001O000O100000O100000000O1000000O100N1O2N2O1O1O1O1O1O1O1O001O1O1O1O1N2O1O1b0]O3N1O2N1OM3M3M3M3L301O1O1O1O1O11O1O1N101O1O0UETNW:m1eEVN[:k1aEYN^:h1^E[Nb:e1[E^Ne:U2O0O1000O1O1O1N2O1O1O01000000000001O2N1N3N1O1O2N000000000O0100000000O1000000O010000000O1000O10O100000O10O11O0O10001O0O101O00000O2O00000O2OO100000O1000000O1000000O010000000O1000000O2O000000000000001O00000000000000000000O10000O10000O10000O10000001O1O1O1O1O1O1O1O1O1O00000000O1000000000000000000mNRFiNn9W1XFbNi9_1]FSNj9m1j01O001O00001O001O001O001O00001O001O001O6J5K6J00000000O100000000O1000000O100000000O100O1O1O1O1O100000000O1000000000000O10000000000L4KY1gNcVP1"}, "label": "an amherst crane truck body"}]}
{"id": 4704, "image": "COCO_train2014_000000004704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person doing jump on snowboard\"."}], "task": "refering", "answer_template": "The \"person doing jump on snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 21, 22, 36, 37, 51, 52, 53, 54, 66, 67, 68, 69, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 109, 112, 113, 127, 128, 142], "bbox": [0.2615456674473068, 0.03153125, 0.6437939110070257, 0.40737499999999993], "iscrowd": 0, "area": 12444.736250000002, "rle": {"size": [640, 427], "counts": "gUV28dc06N2N2N1O0000000O10001O00000000000N2K6M2O2N2O0O2O0O010O01O000010O01O001O0010O0001O0010O01OO1O100O2O0O100O101N100O100O101N10000O101gLnNbCS1_<PO]CP1d<SOXCm0i<VORCk0o<YOlBh0T=\\OfBe0[=_O`Ba0b=BZB=h=EUB:l=JPB5R>NjA2W>1eAN]>4`AKb>8ZAGh><TAEl>>QABo>a0m@@R?c0k@^OW?c0e@^O\\?d0`@^O`?d0\\@]Oe?_201kAnKZ<X4SCSLo<T4\\BWLg=P52M[B`Jn<^5PCeJP=Y5oBkJQ=S5mBQKS=m4kBWKU=f4jB]KW=a4gBaK[=\\4dBeK^=Z4`BhKb=V4]BkKe=W4UBkKm=S52N2N1N3N2N2M3L4M2N3L4M3M3L3N1O2_Nn@QNS?n1SAlMo>Q2WAjMj>U2\\AeMf>X2`AcMa>\\2eA^M]>`2hA[MY>c2nAWMS>h2d1N1N3N1N3N1O2M2O1O2M2O2N1N3N1N3N1O1N3E;E;N2O0O2O1N2O1O1N101N2O1N2L4J5J`ln2"}, "label": "person doing jump on snowboard"}]}
{"id": 4704, "image": "COCO_train2014_000000004704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboarder in mid - flight\"."}], "task": "refering", "answer_template": "The \"a snowboarder in mid - flight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 21, 22, 36, 37, 51, 52, 53, 54, 66, 67, 68, 69, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 109, 112, 113, 127, 128, 142], "bbox": [0.2615456674473068, 0.03153125, 0.6437939110070257, 0.40737499999999993], "iscrowd": 0, "area": 12444.736250000002, "rle": {"size": [640, 427], "counts": "gUV28dc06N2N2N1O0000000O10001O00000000000N2K6M2O2N2O0O2O0O010O01O000010O01O001O0010O0001O0010O01OO1O100O2O0O100O101N100O100O101N10000O101gLnNbCS1_<PO]CP1d<SOXCm0i<VORCk0o<YOlBh0T=\\OfBe0[=_O`Ba0b=BZB=h=EUB:l=JPB5R>NjA2W>1eAN]>4`AKb>8ZAGh><TAEl>>QABo>a0m@@R?c0k@^OW?c0e@^O\\?d0`@^O`?d0\\@]Oe?_201kAnKZ<X4SCSLo<T4\\BWLg=P52M[B`Jn<^5PCeJP=Y5oBkJQ=S5mBQKS=m4kBWKU=f4jB]KW=a4gBaK[=\\4dBeK^=Z4`BhKb=V4]BkKe=W4UBkKm=S52N2N1N3N2N2M3L4M2N3L4M3M3L3N1O2_Nn@QNS?n1SAlMo>Q2WAjMj>U2\\AeMf>X2`AcMa>\\2eA^M]>`2hA[MY>c2nAWMS>h2d1N1N3N1N3N1O2M2O1O2M2O2N1N3N1N3N1O1N3E;E;N2O0O2O1N2O1O1N101N2O1N2L4J5J`ln2"}, "label": "a snowboarder in mid - flight"}]}
{"id": 528992, "image": "COCO_train2014_000000528992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"scallots\"."}], "task": "refering", "answer_template": "The \"scallots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 284], "bbox": [0.21793749999999998, 0.4471458333333333, 0.5245625, 0.7172499999999999], "iscrowd": 0, "area": 18395.3766, "rle": {"size": [480, 640], "counts": "VlQ26g>5J6K5L3N3N2M3M2N3M3L4K4M3J6I7M3N2O1N2N2N2N5L3L3M3M3M4M2M2O001O1O1O00001O001O001O00001O001O00001O001O001O000000000001O1O1O1O010O1O1O1O1O1OO1O010O1O1O1O1000O010000O100O10O10O1000001O3M3M4L3M3M1O000000O10000000000O1000001O000O2O00001O000O2O00001O000O2O00001O000O2O001O000O2O00001O0O101O00001N10001O000O2O0O1O1O2N1O1O2N1O1_Oa001N10001O000O2O0000001N10001N10004K7J7I2M2O0O2N1O2N1O2N1O2N1O2N1I8E:FSX^4"}, "label": "scallots"}]}
{"id": 528992, "image": "COCO_train2014_000000528992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana snacks\"."}], "task": "refering", "answer_template": "The \"banana snacks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 284], "bbox": [0.21793749999999998, 0.4471458333333333, 0.5245625, 0.7172499999999999], "iscrowd": 0, "area": 18395.3766, "rle": {"size": [480, 640], "counts": "VlQ26g>5J6K5L3N3N2M3M2N3M3L4K4M3J6I7M3N2O1N2N2N2N5L3L3M3M3M4M2M2O001O1O1O00001O001O001O00001O001O00001O001O001O000000000001O1O1O1O010O1O1O1O1O1OO1O010O1O1O1O1000O010000O100O10O10O1000001O3M3M4L3M3M1O000000O10000000000O1000001O000O2O00001O000O2O00001O000O2O00001O000O2O001O000O2O00001O0O101O00001N10001O000O2O0O1O1O2N1O1O2N1O1_Oa001N10001O000O2O0000001N10001N10004K7J7I2M2O0O2N1O2N1O2N1O2N1O2N1I8E:FSX^4"}, "label": "banana snacks"}]}
{"id": 528992, "image": "COCO_train2014_000000528992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana toast sits on a cutting board to the right of another banana toast\"."}], "task": "refering", "answer_template": "The \"a banana toast sits on a cutting board to the right of another banana toast\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200], "bbox": [0.517375, 0.3160208333333333, 0.761796875, 0.521], "iscrowd": 0, "area": 10665.483850000002, "rle": {"size": [480, 640], "counts": "iZk45j>1O2N100^AHh=8]B1IKf=8_BNIM3K[==gB<W=FgB<W=FhB:W=HgB;W=h0N;F2M2N3N0O2O1O001`CcM[<d2M001O001O1O001O00001O001O0010O00010O01O0010O00010O2N1O2N2N2N0000000000000000N2O1N20000000001O000000000O10000000000001O0000000000001O000O1000000O101N1O1O2O0O1O2O3L2N1N3J6O001O1OO100O10O10O10O10O01HPC`NP=Z1QCcNO3P=Y1RCcNO3P=Y1SCbNO3n<[1YCcNi<\\1>N2N2N1O2000000001O0000000000000000EYB\\Oh=d0ZBZOf=e0\\BZOd=f0]BUOg=k08N10001N1O1N3L5L3L9H`RW2"}, "label": "a banana toast sits on a cutting board to the right of another banana toast"}]}
{"id": 528992, "image": "COCO_train2014_000000528992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced bananas to the left of a glass of brown liquid\"."}], "task": "refering", "answer_template": "The \"sliced bananas to the left of a glass of brown liquid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200], "bbox": [0.517375, 0.3160208333333333, 0.761796875, 0.521], "iscrowd": 0, "area": 10665.483850000002, "rle": {"size": [480, 640], "counts": "iZk45j>1O2N100^AHh=8]B1IKf=8_BNIM3K[==gB<W=FgB<W=FhB:W=HgB;W=h0N;F2M2N3N0O2O1O001`CcM[<d2M001O001O1O001O00001O001O0010O00010O01O0010O00010O2N1O2N2N2N0000000000000000N2O1N20000000001O000000000O10000000000001O0000000000001O000O1000000O101N1O1O2O0O1O2O3L2N1N3J6O001O1OO100O10O10O10O10O01HPC`NP=Z1QCcNO3P=Y1RCcNO3P=Y1SCbNO3n<[1YCcNi<\\1>N2N2N1O2000000001O0000000000000000EYB\\Oh=d0ZBZOf=e0\\BZOd=f0]BUOg=k08N10001N1O1N3L5L3L9H`RW2"}, "label": "sliced bananas to the left of a glass of brown liquid"}]}
{"id": 488033, "image": "COCO_train2014_000000488033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"partially opened green umbrella\"."}], "task": "refering", "answer_template": "The \"partially opened green umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 30, 31, 32, 33, 34, 35, 36, 38, 39, 40, 41, 42, 43, 45, 46, 47, 48, 49, 51, 54, 55, 56, 57, 58, 60, 61, 62, 63, 64, 66, 70, 71, 73, 76, 78, 81, 86, 96, 111, 126, 141, 172, 186, 187, 201, 202], "bbox": [0.0, 0.0, 0.9288056206088994, 0.5841093749999999], "iscrowd": 0, "area": 37207.23925000002, "rle": {"size": [640, 427], "counts": "_21g04[b0M[]O?ab0CU]Oh0hb0;L4L3N3L3M2N2N2N2O0O2N2N2N2O1N2N2N2N2O2M3M2N3M2O2M2N3M4L5K6K5Je0[O001O10O01O1O00100O11N2O1O1N2O1O1O1N2O1O1N2O1O1N200O1O100O100O1O100O100001O2N2N2N2N2N1O2N1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O4L4L5K4L00O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1O100O1O1^MSA1m>OVAMk>3XAJh>6[AFf>:]ACc>=`A_Oa>a0bA[O_>e0dAXO\\>h0gATOZ>l0lAnNT>S1QBfNQ>Y1VB_Nk=a1[BYNe=h1Y20000001O00001O00001O00001O00001O00001O0000001O00000000000000000000000000002VCQNhK;[:Z2RHeN\\MSOb:m2TF^OSOfMj:Y7oDiHQ;\\7hDfHX;Q80OWGbGR6^8mIcGS6j6SHkIi1\\OT6S6kH_JP1@T6Y5fITK6DT6Q4nJZLoNDS6V3QOiLo0`2I`M6i1b0VN^Ob1k0]NUOc1k0^NTOb1m0]NSOb1o0]NQOc1P1]NoNc1R1\\NnNc1T1\\NlNd1T1\\NlNc1U1^NjNb1U1_NkNa1U1_NkN`1V1aNiN^1W1cNiN\\1X1dNhN\\12\\NfH9W7Z14aN`H6\\7X14gNYH2e7V13lNPH0o7S11ROiGMW8Q10VObGJ`8o0NK34OI361E1:4AM?6\\OLc08YOIf0<TOFk0>QOCo0a0kNAU1b0fN@Z1c900O100000000O100000000O100000000O100000000O10000003M3M4L3M4L3M3M4L3M4L4L3M4L4L4L4L3M4L3M001O0000001O00001O0000001O00001O00001O0000001O00001O0S@`LZ?`3`@fL`?Z3Y@nLf?f30001O0000001O00001O0000010O000000@Y@hLf?X3[@gLe?Y3[@gLe?X3\\@iLc?W3]@iLc?V3^@jLb?V3_@iLa?V3`@kL_?U3a@kL_?T3b@lL^?T3b@lL^?S3d@mL[?S3e@mL[?R3f@nL[?Q3e@PM[?n2f@RM\\?l2e@SM\\?l2d@TM]?j2d@WM]?g2c@YM^?e2c@[M^?d2c@[M^?c2e@\\M\\?b2g@[MZ?d2i@YMX?f2k@XMV?f2n@VMS?i2S1O1O2N1N2O1O2N1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1lNf]O?\\b0Ad]O>]b0Ad]O>^b0@c]O`0]b0@c]O?^b0@c]O?^b0Ab]O>_b0Ab]O?_b0_Ob]O`0_b0@a]O?`b0@a]O`0_b0_Ob]O`0_b0@a]O?ab0_O`]O`0ab0@_]O`0ab0_O`]O`0ab0_Oa]O?`b0A`]O?ab0_O`]O`0ab0_O`]O`0ab0@_]O`0ab0_O`]O`0ab0@_]O?Rc0O100O1O100O1O1O100Obdb0"}, "label": "partially opened green umbrella"}]}
{"id": 488033, "image": "COCO_train2014_000000488033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green umbrella over a man ' s head\"."}], "task": "refering", "answer_template": "The \"a green umbrella over a man ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 30, 31, 32, 33, 34, 35, 36, 38, 39, 40, 41, 42, 43, 45, 46, 47, 48, 49, 51, 54, 55, 56, 57, 58, 60, 61, 62, 63, 64, 66, 70, 71, 73, 76, 78, 81, 86, 96, 111, 126, 141, 172, 186, 187, 201, 202], "bbox": [0.0, 0.0, 0.9288056206088994, 0.5841093749999999], "iscrowd": 0, "area": 37207.23925000002, "rle": {"size": [640, 427], "counts": "_21g04[b0M[]O?ab0CU]Oh0hb0;L4L3N3L3M2N2N2N2O0O2N2N2N2O1N2N2N2N2O2M3M2N3M2O2M2N3M4L5K6K5Je0[O001O10O01O1O00100O11N2O1O1N2O1O1O1N2O1O1N2O1O1N200O1O100O100O1O100O100001O2N2N2N2N2N1O2N1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O4L4L5K4L00O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1O100O1O1^MSA1m>OVAMk>3XAJh>6[AFf>:]ACc>=`A_Oa>a0bA[O_>e0dAXO\\>h0gATOZ>l0lAnNT>S1QBfNQ>Y1VB_Nk=a1[BYNe=h1Y20000001O00001O00001O00001O00001O00001O0000001O00000000000000000000000000002VCQNhK;[:Z2RHeN\\MSOb:m2TF^OSOfMj:Y7oDiHQ;\\7hDfHX;Q80OWGbGR6^8mIcGS6j6SHkIi1\\OT6S6kH_JP1@T6Y5fITK6DT6Q4nJZLoNDS6V3QOiLo0`2I`M6i1b0VN^Ob1k0]NUOc1k0^NTOb1m0]NSOb1o0]NQOc1P1]NoNc1R1\\NnNc1T1\\NlNd1T1\\NlNc1U1^NjNb1U1_NkNa1U1_NkN`1V1aNiN^1W1cNiN\\1X1dNhN\\12\\NfH9W7Z14aN`H6\\7X14gNYH2e7V13lNPH0o7S11ROiGMW8Q10VObGJ`8o0NK34OI361E1:4AM?6\\OLc08YOIf0<TOFk0>QOCo0a0kNAU1b0fN@Z1c900O100000000O100000000O100000000O100000000O10000003M3M4L3M4L3M3M4L3M4L4L3M4L4L4L4L3M4L3M001O0000001O00001O0000001O00001O00001O0000001O00001O0S@`LZ?`3`@fL`?Z3Y@nLf?f30001O0000001O00001O0000010O000000@Y@hLf?X3[@gLe?Y3[@gLe?X3\\@iLc?W3]@iLc?V3^@jLb?V3_@iLa?V3`@kL_?U3a@kL_?T3b@lL^?T3b@lL^?S3d@mL[?S3e@mL[?R3f@nL[?Q3e@PM[?n2f@RM\\?l2e@SM\\?l2d@TM]?j2d@WM]?g2c@YM^?e2c@[M^?d2c@[M^?c2e@\\M\\?b2g@[MZ?d2i@YMX?f2k@XMV?f2n@VMS?i2S1O1O2N1N2O1O2N1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1lNf]O?\\b0Ad]O>]b0Ad]O>^b0@c]O`0]b0@c]O?^b0@c]O?^b0Ab]O>_b0Ab]O?_b0_Ob]O`0_b0@a]O?`b0@a]O`0_b0_Ob]O`0_b0@a]O?ab0_O`]O`0ab0@_]O`0ab0_O`]O`0ab0_Oa]O?`b0A`]O?ab0_O`]O`0ab0_O`]O`0ab0@_]O`0ab0_O`]O`0ab0@_]O?Rc0O100O1O100O1O1O100Obdb0"}, "label": "a green umbrella over a man ' s head"}]}
{"id": 488033, "image": "COCO_train2014_000000488033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bamboo umbrella handle and leather suitcase handle in someone ' s hand at the same time\"."}], "task": "refering", "answer_template": "The \"a bamboo umbrella handle and leather suitcase handle in someone ' s hand at the same time\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 217, 218, 232, 233, 248, 263, 264, 279, 294], "bbox": [0.4691569086651054, 0.57371875, 0.6618032786885246, 0.871796875], "iscrowd": 0, "area": 4084.65105, "rle": {"size": [640, 427], "counts": "k\\m3k0gb0?L3M4K4M4L3K6N2N1O1O100O1O00O1O101N2N2nNR1G9O1O1O1O13M6J6J5L4K4M3L4M2M4L4M3L4M3L310O10000O10O10O10000O10O100fNc^OM\\a0Fo^O=Qa0UOZ_On0ja04M2N1OOh]OdNVb0Q1l]ORORb0l0n]OWOQb0g0P^O\\Ona0a0S^OBma0;T^OHja06V^OMia00Y^O3ea0K[^O6fa0G\\^O8fa0FZ^O:ha0DY^O;ia0BY^O=ja0@X^O=ka0AW^O=hb0N1O2N2N2N2M3NfRj2"}, "label": "a bamboo umbrella handle and leather suitcase handle in someone ' s hand at the same time"}]}
{"id": 488033, "image": "COCO_train2014_000000488033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cane is being held by a man\"."}], "task": "refering", "answer_template": "The \"a cane is being held by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 217, 218, 232, 233, 248, 263, 264, 279, 294], "bbox": [0.4691569086651054, 0.57371875, 0.6618032786885246, 0.871796875], "iscrowd": 0, "area": 4084.65105, "rle": {"size": [640, 427], "counts": "k\\m3k0gb0?L3M4K4M4L3K6N2N1O1O100O1O00O1O101N2N2nNR1G9O1O1O1O13M6J6J5L4K4M3L4M2M4L4M3L4M3L310O10000O10O10O10000O10O100fNc^OM\\a0Fo^O=Qa0UOZ_On0ja04M2N1OOh]OdNVb0Q1l]ORORb0l0n]OWOQb0g0P^O\\Ona0a0S^OBma0;T^OHja06V^OMia00Y^O3ea0K[^O6fa0G\\^O8fa0FZ^O:ha0DY^O;ia0BY^O=ja0@X^O=ka0AW^O=hb0N1O2N2N2N2M3NfRj2"}, "label": "a cane is being held by a man"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown suitcase with the sripes in the back\"."}], "task": "refering", "answer_template": "The \"brown suitcase with the sripes in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 163, 164, 165, 188, 189], "bbox": [0.0016718750000000002, 0.002458333333333333, 0.68959375, 0.51475], "iscrowd": 0, "area": 49767.39135, "rle": {"size": [480, 640], "counts": "ja0d0\\>d0\\O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O0O0100O001O10O01O1O010O1O0010jFnJ]8R5dGPKZ8o4fGTKX8k4iGWKT8j4kGXKT8g4mGXKT8g4lGZKS8f4nGZKR8f4mGZKT8e4mG[KR8e4oG[KQ8d4oG\\KR8d4nG\\KQ8e4nG[KS8d4nG\\KR8d4mG]KR8c4oG\\KR8d4mG]KS8b4nG^KQ8c4oG\\KR8c4nG^KR8b4nG]KR8d4mG]KS8b4nG^KR8b4mG^KS8b4nG^KR8b4mG_KS8`4nG_KR8b4nG^KR8b4mG^KT8a4mGVK[8k4m0O10O010O01000O0100O010O10O10O1O001O1N101O1O001O1O001O1O001N2O001O1O001O1O001O001O10O01000O01000O0100O01000O01000O010O10O10O10O10O10O01000O01000O0100O01000O01000O10O01000O01000O01000O0100O01000O01000O010O10O10O10O10O10O010000O10000O100O10000O10000O10000O100O10000O10000O010O10000O10000O100O10000O10000O100O10000O10000O10000O100O01000O10000O100O10000O10000O100O10000O10000O100O10000O1000O010000O100O10000O2O001N101N2O001N101O1N101N101O1N101O0O2O1O0O2O001N2O001N101O1N101O001N2O001N101O1N101O0O2O1O0O2O001N2O001N101N2O001N101N2O001N101O1N101N101O1N101N101O1N101N101O1N101O0O2O1N101O0O3N3L4M3M2MnVm2"}, "label": "brown suitcase with the sripes in the back"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and black suitcase chained to 2 other suitcases\"."}], "task": "refering", "answer_template": "The \"white and black suitcase chained to 2 other suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.2674166666666667, 0.40954687500000003, 1.0], "iscrowd": 0, "area": 56532.212400000004, "rle": {"size": [480, 640], "counts": "Q4o:R4O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K4L5K5K5K5K5K5K5K5K5K5K5K5K4L5K5K5K5K5KYg`5"}, "label": "white and black suitcase chained to 2 other suitcases"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white suitcase with black dots chained to other suitcases\"."}], "task": "refering", "answer_template": "The \"a white suitcase with black dots chained to other suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.2674166666666667, 0.40954687500000003, 1.0], "iscrowd": 0, "area": 56532.212400000004, "rle": {"size": [480, 640], "counts": "Q4o:R4O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K4L5K5K5K5K5K5K5K5K5K5K5K5K4L5K5K5K5K5KYg`5"}, "label": "a white suitcase with black dots chained to other suitcases"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a darker brown suit case in front of two light brown suit cases with it ' s latches visible\"."}], "task": "refering", "answer_template": "The \"a darker brown suit case in front of two light brown suit cases with it ' s latches visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 227, 228, 229, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.47973437499999994, 0.498875, 1.0, 0.9876041666666667], "iscrowd": 0, "area": 47327.342300000004, "rle": {"size": [480, 640], "counts": "f[`43g>6J6J6J6J6J6J6J6J6N200O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O2O0O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100ObN_1[Ne1\\Nf1YNSJ"}, "label": "a darker brown suit case in front of two light brown suit cases with it ' s latches visible"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown luggage to the right\"."}], "task": "refering", "answer_template": "The \"the brown luggage to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 227, 228, 229, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.47973437499999994, 0.498875, 1.0, 0.9876041666666667], "iscrowd": 0, "area": 47327.342300000004, "rle": {"size": [480, 640], "counts": "f[`43g>6J6J6J6J6J6J6J6J6N200O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O2O0O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100ObN_1[Ne1\\Nf1YNSJ"}, "label": "the brown luggage to the right"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white briefcase with brown lining chained to two other briefcases\"."}], "task": "refering", "answer_template": "The \"a white briefcase with brown lining chained to two other briefcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 354, 355, 356, 378], "bbox": [0.22753125000000002, 0.035958333333333335, 1.0, 0.9797708333333333], "iscrowd": 0, "area": 119788.08484999998, "rle": {"size": [480, 640], "counts": "VeT2S1l=3L4M3L4L5L3L4L4M3L4M3L4L4M4K4L4M3L3M3N2M3N2M3M3N2M3M3N2N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O001N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N101O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O001N3N6I7J6J6I7J7H7J6J4K2O1O1N101N2O1O0O2O1N2O001N2O1O0O2O1N101O1N2O0O2O1O1N101O1N2O0O2O1O1N101N2ON2N1N3N2M2O2M2O2M3M2O2M3N1N3N1N3N2J5H90O0100O010O10O010O0100O010O01000O010O10O010O0100O010O010O10O01000O010O01O10O010O0100O010O10O01O010O10O010O0100O010O1O010O010O10O010O0100O010O1O010O010O10O010O0100O00100O010O010O10O0100O0010O0100O010O010O10O01O10O010O0100O010O010O1O010O10O010O0100O010O00100O010O10O010O0100O0010O0100O010O10O010O01O10O010O0100O010O10O01O010O10O010O0100O010O10O01O010O10O010O0100O010O1O010O010O10O0100O010O00100O010O010O10O0100O0010O0100O010O010O10O01O010O010O10O010O0100O0010O010O0100O010O010O1O010O010O010O10O010O01O10O010O010O0100O010O00100O010O010O010O10O010O01O10O010O010O0100O010O00100O100O100O100O100O1O101N101N2O1N2O1N2O1N2N2O1N2O0O2O1N2O1N2O1N2N2O1N2nIbL`0W2"}, "label": "a white briefcase with brown lining chained to two other briefcases"}]}
{"id": 537188, "image": "COCO_train2014_000000537188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cream colored luggage with brown trim , secured with a metal chain\"."}], "task": "refering", "answer_template": "The \"a cream colored luggage with brown trim , secured with a metal chain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 354, 355, 356, 378], "bbox": [0.22753125000000002, 0.035958333333333335, 1.0, 0.9797708333333333], "iscrowd": 0, "area": 119788.08484999998, "rle": {"size": [480, 640], "counts": "VeT2S1l=3L4M3L4L5L3L4L4M3L4M3L4L4M4K4L4M3L3M3N2M3N2M3M3N2M3M3N2N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O001N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N101O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O001N3N6I7J6J6I7J7H7J6J4K2O1O1N101N2O1O0O2O1N2O001N2O1O0O2O1N101O1N2O0O2O1O1N101O1N2O0O2O1O1N101N2ON2N1N3N2M2O2M2O2M3M2O2M3N1N3N1N3N2J5H90O0100O010O10O010O0100O010O01000O010O10O010O0100O010O010O10O01000O010O01O10O010O0100O010O10O01O010O10O010O0100O010O1O010O010O10O010O0100O010O1O010O010O10O010O0100O00100O010O010O10O0100O0010O0100O010O010O10O01O10O010O0100O010O010O1O010O10O010O0100O010O00100O010O10O010O0100O0010O0100O010O10O010O01O10O010O0100O010O10O01O010O10O010O0100O010O10O01O010O10O010O0100O010O1O010O010O10O0100O010O00100O010O010O10O0100O0010O0100O010O010O10O01O010O010O10O010O0100O0010O010O0100O010O010O1O010O010O010O10O010O01O10O010O010O0100O010O00100O010O010O010O10O010O01O10O010O010O0100O010O00100O100O100O100O100O1O101N101N2O1N2O1N2O1N2N2O1N2O0O2O1N2O1N2O1N2N2O1N2nIbL`0W2"}, "label": "a cream colored luggage with brown trim , secured with a metal chain"}]}
{"id": 324200, "image": "COCO_train2014_000000324200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very long submarine sandwich\"."}], "task": "refering", "answer_template": "The \"a very long submarine sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 300, 301, 302, 303], "bbox": [0.032015625, 0.12583333333333332, 0.8190937500000001, 0.8157291666666666], "iscrowd": 0, "area": 47022.46804999999, "rle": {"size": [480, 640], "counts": "cf91l>4K6J6J5L5J5L5L3N3L3M4M3L3M4M2M4L3N3L3N3N2M2O2M2O2M2O0O101N101N100O2O0O2O0O101N101N10001N101N100O2O0O2N100O2O0O2OO010O0100O010O010O10O01O10O010O0100O010O010O10O010O01O10O0100O010O010O10O010O0100O010O001O1O0010O01O1O001O1O010O001O1O0010O01O00000010O00000000001O00000000001O0000001O1O001O1O1O001O1O1O1O010O100O1O010O100O00100O1O10O0100O1O10O0100O1O010O100O1O010O0001O001O01O01O00000010O00000001O01O0001O000001O01O00000010O00000001O01O00010O00001O01O01O001O01O01O00001O00001O010O00001O00001O010O00001O001O001O01O01O001O001O0010O01O010O0010O01O010O0010O01O0010O00010O01O010O0010O01O010O0010O01O0010O01O010O0010O00010O010O01O010O0010O01O010O01O01O010O0010O01O010O010O00010O0010O01O010O010O0010O00010O01O010O010O0010O00010O01O010O010O0010O01O01O01O010O010O0010O01O0010O0001O010O001O010O001O010O000010O01O0010O0001O010O00010O001O010O000010O01O00010O001O01O01O0010O01O100O1O100O2O0O1O100O2O0O1O101N100O100O2N100O100O2N100O100O2N100O101N100O1O101N100O1O101N100O1O2O0O1O101N100O1O2O0O1O2O0O100O2N2O1N2N2O1N2O1N2N2M3K_Yf1"}, "label": "a very long submarine sandwich"}]}
{"id": 324200, "image": "COCO_train2014_000000324200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long sub sandwich\"."}], "task": "refering", "answer_template": "The \"a long sub sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 300, 301, 302, 303], "bbox": [0.032015625, 0.12583333333333332, 0.8190937500000001, 0.8157291666666666], "iscrowd": 0, "area": 47022.46804999999, "rle": {"size": [480, 640], "counts": "cf91l>4K6J6J5L5J5L5L3N3L3M4M3L3M4M2M4L3N3L3N3N2M2O2M2O2M2O0O101N101N100O2O0O2O0O101N101N10001N101N100O2O0O2N100O2O0O2OO010O0100O010O010O10O01O10O010O0100O010O010O10O010O01O10O0100O010O010O10O010O0100O010O001O1O0010O01O1O001O1O010O001O1O0010O01O00000010O00000000001O00000000001O0000001O1O001O1O1O001O1O1O1O010O100O1O010O100O00100O1O10O0100O1O10O0100O1O010O100O1O010O0001O001O01O01O00000010O00000001O01O0001O000001O01O00000010O00000001O01O00010O00001O01O01O001O01O01O00001O00001O010O00001O00001O010O00001O001O001O01O01O001O001O0010O01O010O0010O01O010O0010O01O0010O00010O01O010O0010O01O010O0010O01O0010O01O010O0010O00010O010O01O010O0010O01O010O01O01O010O0010O01O010O010O00010O0010O01O010O010O0010O00010O01O010O010O0010O00010O01O010O010O0010O01O01O01O010O010O0010O01O0010O0001O010O001O010O001O010O000010O01O0010O0001O010O00010O001O010O000010O01O00010O001O01O01O0010O01O100O1O100O2O0O1O100O2O0O1O101N100O100O2N100O100O2N100O100O2N100O101N100O1O101N100O1O101N100O1O2O0O1O101N100O1O2O0O1O2O0O100O2N2O1N2N2O1N2O1N2N2M3K_Yf1"}, "label": "a long sub sandwich"}]}
{"id": 397929, "image": "COCO_train2014_000000397929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in a purple shirt and diapers with a woman pulling up her pants\"."}], "task": "refering", "answer_template": "The \"a little girl in a purple shirt and diapers with a woman pulling up her pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 333, 334, 356], "bbox": [0.43121875000000004, 0.364125, 0.6139218750000001, 0.9027083333333332], "iscrowd": 0, "area": 15231.320949999996, "rle": {"size": [480, 640], "counts": "^cQ43i>6M4L3M3M3M101N101N101O0O101OO10O100dF_OY4a0UIXO5;e6=eHKc0Kf6;lGg0Y1ROh6i2SIZMl6j2aG^Lm0o0]7QO^Gf34\\Lf0W1f7f2`GUL`0^1n7g2PH[Mn7h2PHWMo7m2oGPMR8T3mGjLc7`NaHj4JgL_7iNaHb4OfLZ7RO`H[41hL]7j5N2O0O2O0O101N10001N100O2O0O101O0O100O10YOWIVIj6i6YIUIf6k6\\ISIe6i6aITI`6h6eIWI[6h6gIVIZ6T5PI[Lf0_N[6U5PI\\Le0_N\\6S5QI^Lc0]N]6T5QI_Lc0\\N\\6S5TIaLe7X3UHVK7b1d7U3YHWK3e1d7S3dHnL\\7Q3eHnL]7P3cHQM]7n2cHRM^7m2bHgL_OPOo7X4bH\\LMZOb7Y4bHoK;EU7\\4`HmK?BT7`4_HkKh8S4ZGjKi8T4YGjKi8T4R1O2N2N2N1O2bNTEeNm:Q1gEdN[:V1`FRNb9o1h1O010O01O001O02O0O2Nn0RO7JO00010O0000010O000010O002N101N2I7H8Hcac3"}, "label": "a little girl in a purple shirt and diapers with a woman pulling up her pants"}]}
{"id": 397929, "image": "COCO_train2014_000000397929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child in a purple shirt\"."}], "task": "refering", "answer_template": "The \"a child in a purple shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 333, 334, 356], "bbox": [0.43121875000000004, 0.364125, 0.6139218750000001, 0.9027083333333332], "iscrowd": 0, "area": 15231.320949999996, "rle": {"size": [480, 640], "counts": "^cQ43i>6M4L3M3M3M101N101N101O0O101OO10O100dF_OY4a0UIXO5;e6=eHKc0Kf6;lGg0Y1ROh6i2SIZMl6j2aG^Lm0o0]7QO^Gf34\\Lf0W1f7f2`GUL`0^1n7g2PH[Mn7h2PHWMo7m2oGPMR8T3mGjLc7`NaHj4JgL_7iNaHb4OfLZ7RO`H[41hL]7j5N2O0O2O0O101N10001N100O2O0O101O0O100O10YOWIVIj6i6YIUIf6k6\\ISIe6i6aITI`6h6eIWI[6h6gIVIZ6T5PI[Lf0_N[6U5PI\\Le0_N\\6S5QI^Lc0]N]6T5QI_Lc0\\N\\6S5TIaLe7X3UHVK7b1d7U3YHWK3e1d7S3dHnL\\7Q3eHnL]7P3cHQM]7n2cHRM^7m2bHgL_OPOo7X4bH\\LMZOb7Y4bHoK;EU7\\4`HmK?BT7`4_HkKh8S4ZGjKi8T4YGjKi8T4R1O2N2N2N1O2bNTEeNm:Q1gEdN[:V1`FRNb9o1h1O010O01O001O02O0O2Nn0RO7JO00010O0000010O000010O002N101N2I7H8Hcac3"}, "label": "a child in a purple shirt"}]}
{"id": 201322, "image": "COCO_train2014_000000201322.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on which the kid is seated\"."}], "task": "refering", "answer_template": "The \"the chair on which the kid is seated\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 176, 177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 339, 340, 341, 342, 343, 344], "bbox": [0.6022482435597191, 0.256765625, 1.0, 0.9836406249999999], "iscrowd": 0, "area": 49141.9372, "rle": {"size": [640, 427], "counts": "\\lP51mc02SGNkM4R2NlM4S2NkM4S2MkM5T2LjM7T2JkM7S2KkM8S2IlM8S2IkM:S2GlM:R2HlM;R2FlM<S2ElM=R2DlM>S2ClM?Q2CmM?R2BlMa0R2@mMa0R2@lMc0Q2_OnMb0Q2_OmMd0Q2]OnMd0Q2]OmMf0Q2[OnMf0P2\\OoMf0o1\\OoMe0P2\\OnMg0P2ZOoMg0o1[OPNg0n1ZOQNh0m1YOQNj0m1WORNk0k1WOTNk0j1VOTNm0j1TOUNn0i1SOVNo0h1ROWNP1f1ROXNP1g1QOXNQ1c1SO\\No0`1TO_Nn0]1UOaNn0[1UOdNm0W1WOhNk0T1XOjNk0R1XOmNj0o0YOPOi0l0ZOSOh0h0]OUOf0g0]OfN\\M^JY3h6^OjN\\MZJX3h6_OnN\\MVJV3i6APO]MRJU3l6_OROQ1l0POSOR1l0nNTOS1j0nNUOT1j0lNVOU1h0lNWOV1h0jNXOW1f0jNYOX1f0hNZOY1d0hN\\OY1c0gN\\O[1b0fN^O[1a0eN^O\\1a0eN_O\\1`0dN_O^1?cNA^1>bNA`1=aNC`1<aNBa1<`NDa1;_NEb19_NFc19]NGd17]NHe17[NIf15[NJg15YNKg14ZNKh14XNMh11YNOh1kMhJfN`3_3i1iMiJfN_3a3i1gMiJhN^3a3j1eMjJhN]3c3j1cMkJiN[3d3k1aMkJjN[3e3k1_MlJkNY3g3n1YMkJnNX3i3[2jL^J\\OX3j3g2[LTJJU3k3T3mKiI6T3m3`3_K]Id0S3m3e3YKZIh0S3n3f3VKYIk0Q3o3j3QKWIn0P3R4k3lJVIR1o2R4o3gJTIU1n2T4R4bJRIY1l2U4U4^JPI\\1l2V4[5eKeJ[4`5_KaJa4e5YK[Jh4j5QKWJo4o5kJQJU5U6dJlI\\5Z6^JfIb5\\:0000001O0000000001O0001O0000000000001O000000000000001O0000000001O0001O0000000000001O000000000000001O000000000001O01^Ka4bIn0"}, "label": "the chair on which the kid is seated"}]}
{"id": 201322, "image": "COCO_train2014_000000201322.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ash color chair cover\"."}], "task": "refering", "answer_template": "The \"an ash color chair cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 176, 177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 339, 340, 341, 342, 343, 344], "bbox": [0.6022482435597191, 0.256765625, 1.0, 0.9836406249999999], "iscrowd": 0, "area": 49141.9372, "rle": {"size": [640, 427], "counts": "\\lP51mc02SGNkM4R2NlM4S2NkM4S2MkM5T2LjM7T2JkM7S2KkM8S2IlM8S2IkM:S2GlM:R2HlM;R2FlM<S2ElM=R2DlM>S2ClM?Q2CmM?R2BlMa0R2@mMa0R2@lMc0Q2_OnMb0Q2_OmMd0Q2]OnMd0Q2]OmMf0Q2[OnMf0P2\\OoMf0o1\\OoMe0P2\\OnMg0P2ZOoMg0o1[OPNg0n1ZOQNh0m1YOQNj0m1WORNk0k1WOTNk0j1VOTNm0j1TOUNn0i1SOVNo0h1ROWNP1f1ROXNP1g1QOXNQ1c1SO\\No0`1TO_Nn0]1UOaNn0[1UOdNm0W1WOhNk0T1XOjNk0R1XOmNj0o0YOPOi0l0ZOSOh0h0]OUOf0g0]OfN\\M^JY3h6^OjN\\MZJX3h6_OnN\\MVJV3i6APO]MRJU3l6_OROQ1l0POSOR1l0nNTOS1j0nNUOT1j0lNVOU1h0lNWOV1h0jNXOW1f0jNYOX1f0hNZOY1d0hN\\OY1c0gN\\O[1b0fN^O[1a0eN^O\\1a0eN_O\\1`0dN_O^1?cNA^1>bNA`1=aNC`1<aNBa1<`NDa1;_NEb19_NFc19]NGd17]NHe17[NIf15[NJg15YNKg14ZNKh14XNMh11YNOh1kMhJfN`3_3i1iMiJfN_3a3i1gMiJhN^3a3j1eMjJhN]3c3j1cMkJiN[3d3k1aMkJjN[3e3k1_MlJkNY3g3n1YMkJnNX3i3[2jL^J\\OX3j3g2[LTJJU3k3T3mKiI6T3m3`3_K]Id0S3m3e3YKZIh0S3n3f3VKYIk0Q3o3j3QKWIn0P3R4k3lJVIR1o2R4o3gJTIU1n2T4R4bJRIY1l2U4U4^JPI\\1l2V4[5eKeJ[4`5_KaJa4e5YK[Jh4j5QKWJo4o5kJQJU5U6dJlI\\5Z6^JfIb5\\:0000001O0000000001O0001O0000000000001O000000000000001O0000000001O0001O0000000000001O000000000000001O000000000001O01^Ka4bIn0"}, "label": "an ash color chair cover"}]}
{"id": 463467, "image": "COCO_train2014_000000463467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the suit\"."}], "task": "refering", "answer_template": "The \"the man in the suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 57, 58, 59, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 209, 210, 211, 212, 231, 232, 233, 234, 253, 254, 255, 256, 257, 275, 276, 277, 278, 279, 297, 298, 299, 300, 301, 318, 319, 320, 321, 322, 323, 340, 341, 342, 344, 345, 362, 363, 364, 366, 367, 384, 385, 386, 388, 389, 406, 407, 408, 410, 411, 412, 429, 430, 432, 433, 434], "bbox": [0.4768102073365231, 0.050265625, 0.776475279106858, 0.85759375], "iscrowd": 0, "area": 49229.335549999996, "rle": {"size": [640, 627], "counts": "Z\\k51oc01O001O1O1O1O1_CZ2f2gMXM^2d2cMZMc2[2dMcMb2b1UN\\NQ2i0fNUO]1RKQMc0Z4Z4gNgJ]M;]4l4YN]JhM4_4]5kMTJTNKb4o5]MjI_NCe4a6oL`Ie5^6P4O1N1O2O1N100O1O100O010O100O1O100O100O100O100O1O100O10kMPKZEo4Q:hKmEY4l9PLRFo3g9ZLXFf3`9dL]F\\3\\9mLcFR3W9VMhFj2P9`MmF`2l8iMSGW2e8RNZGm1b8ZN[Gg1b8]N]Gb1a8bN^G]1`8hN]GY1`8kN_GT1_8PO`GP1]8UO`Gl0\\8YOcGg0Z8\\OeGe0X8^OhGb0U8AkG>S8ElG<Q8HnG8o7KQH5k7OTH2i71WHOg73YHMj70UHZOd8d0\\GZOi8c0WG\\Om8i54K5L3M4K5L4L4K5L3M4L4L4L4L3M4L4L4L4L3M4L4L4L7J;E:E;F:F7I2N3M3MhDfIm8W6RGkIP9Q6PGPJS9m5lFUJV9h5iFYJY9e5fF\\J]9a5bFaJP;l3oDULQ;k3nDVLV1MX6l3aHYLV13R6b3gH]LW18k5[3kHaLY19h5U3mHdL[1<c5Q3oHeL^1?_5l2PIhL`1a0[5i2QIhLd1d0W5d2RIkLf1f0T5_2SImLi1j0n4Z2VInLl1m0j4V2VIPMo1o0f4R2XIQMR2R1b4m1YISMU2X1Z4f1^IUMW2U2\\3f0ZJnNc0j0e4:dJoNf0n0`43hJQOf0S1\\4MlJSOf0V1X4HPKTOg0Z1R4ETKSOh0_1n3_OXKTOi0d1h3YO]KUOi0j1c3RObKVOj0Y2S3cNPLVOl0i2b2RN`LWOl0R3Z2hMhLXOm0W3T2c1kMdNn1]1PNkNi1W1UNPOc1l9K3L3M4M2M4M2M4M2M4M2M4M2M6J100O1O1O100O1O2N100O1O00100O1O0001O01O0001O00001O01O9G8Ilof2"}, "label": "the man in the suit"}]}
{"id": 463467, "image": "COCO_train2014_000000463467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit is looking at his phone while walking\"."}], "task": "refering", "answer_template": "The \"a man in a suit is looking at his phone while walking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 57, 58, 59, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 209, 210, 211, 212, 231, 232, 233, 234, 253, 254, 255, 256, 257, 275, 276, 277, 278, 279, 297, 298, 299, 300, 301, 318, 319, 320, 321, 322, 323, 340, 341, 342, 344, 345, 362, 363, 364, 366, 367, 384, 385, 386, 388, 389, 406, 407, 408, 410, 411, 412, 429, 430, 432, 433, 434], "bbox": [0.4768102073365231, 0.050265625, 0.776475279106858, 0.85759375], "iscrowd": 0, "area": 49229.335549999996, "rle": {"size": [640, 627], "counts": "Z\\k51oc01O001O1O1O1O1_CZ2f2gMXM^2d2cMZMc2[2dMcMb2b1UN\\NQ2i0fNUO]1RKQMc0Z4Z4gNgJ]M;]4l4YN]JhM4_4]5kMTJTNKb4o5]MjI_NCe4a6oL`Ie5^6P4O1N1O2O1N100O1O100O010O100O1O100O100O100O100O1O100O10kMPKZEo4Q:hKmEY4l9PLRFo3g9ZLXFf3`9dL]F\\3\\9mLcFR3W9VMhFj2P9`MmF`2l8iMSGW2e8RNZGm1b8ZN[Gg1b8]N]Gb1a8bN^G]1`8hN]GY1`8kN_GT1_8PO`GP1]8UO`Gl0\\8YOcGg0Z8\\OeGe0X8^OhGb0U8AkG>S8ElG<Q8HnG8o7KQH5k7OTH2i71WHOg73YHMj70UHZOd8d0\\GZOi8c0WG\\Om8i54K5L3M4K5L4L4K5L3M4L4L4L4L3M4L4L4L4L3M4L4L4L7J;E:E;F:F7I2N3M3MhDfIm8W6RGkIP9Q6PGPJS9m5lFUJV9h5iFYJY9e5fF\\J]9a5bFaJP;l3oDULQ;k3nDVLV1MX6l3aHYLV13R6b3gH]LW18k5[3kHaLY19h5U3mHdL[1<c5Q3oHeL^1?_5l2PIhL`1a0[5i2QIhLd1d0W5d2RIkLf1f0T5_2SImLi1j0n4Z2VInLl1m0j4V2VIPMo1o0f4R2XIQMR2R1b4m1YISMU2X1Z4f1^IUMW2U2\\3f0ZJnNc0j0e4:dJoNf0n0`43hJQOf0S1\\4MlJSOf0V1X4HPKTOg0Z1R4ETKSOh0_1n3_OXKTOi0d1h3YO]KUOi0j1c3RObKVOj0Y2S3cNPLVOl0i2b2RN`LWOl0R3Z2hMhLXOm0W3T2c1kMdNn1]1PNkNi1W1UNPOc1l9K3L3M4M2M4M2M4M2M4M2M4M2M6J100O1O1O100O1O2N100O1O00100O1O0001O01O0001O00001O01O9G8Ilof2"}, "label": "a man in a suit is looking at his phone while walking"}]}
{"id": 463467, "image": "COCO_train2014_000000463467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two bikes on the street , behind to walking men\"."}], "task": "refering", "answer_template": "The \"two bikes on the street , behind to walking men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 117, 118, 119, 137, 140, 141, 159, 162, 163, 164, 181, 182, 184, 185, 186, 187, 199, 203, 204, 205, 206, 207, 208, 209, 221, 225, 226, 227, 228, 229, 230, 243, 245, 247, 248, 249, 251, 252, 253, 267, 269, 273, 274, 275], "bbox": [0.06460925039872407, 0.19059375, 0.5146251993620415, 0.570921875], "iscrowd": 0, "area": 18530.106050000002, "rle": {"size": [640, 627], "counts": "bli08fc08G9G9G9H7L5K4KL9Gifd0cNfZ[O3L3M4L3M3M4L3N2M4L3M3M4L3M3M4M2M3O25J5K5L5J5K5L4K6J5L4K6KVna05dQ^O:U@b0`:HfDg0Z;DZD>f;LoC4Q<6dCK]<gNcAZ1e10h<fNeA^1[1MP=eNhAa1Q1KX=cNjAe1h0I^=bNlAh1?Ge=aNoAk16Dl=aNPBn1MCS>^NSBR2D@Z>^NTBT2\\O_Oa>\\NWBQ4i=oKYBn3g=RL\\Bk3e=TL^Bj3b=VL`Bg3a=XLbBe3_=[LcBc3o>L4M3L4L4L4M3L5K4L4M3N2N2N2N2N2N2O1M3N2N2M3N3M2M3NO10O0100O010O1SL`NZE_1f:bNYE_1g:`NZE_1g:aNXE`1h:`NXE`1g:aNXE_1i:aNWE_1i:aNVE`1j:`NVE_1j:aNVE`1j:`NVE`1j:`NUE`1l:`NTE`1k:aNTE`1l:`NTE_1m:`NSEa1m:_NSEa1m:_NRE`1o:aNQE]1Q;cNnD[1U;eNkDP1`;oN`Di0h;XOXDf0j;ZOUDe0m;[OSDc0o;]OPDb0Q<^OPDa0Q<_OnC`0T<@lC>V<BiC=X<DhC:Z<EfC;[<EeC9]<GbC8`<H`C6a<K^C5c<J\\C7e<IZC6h<JVC7j<JUC6g<FS@4T37j<DS@8o24n<ES@9l22R=ES@:g22W=CS@>b2O\\=CR@`0^2N`=BT@a0Y2Md=BS@d0T2Kj=AR@f0Q2In=@S@h0l1HQ>AS@j0g1FW>_OS@m0\\1Kb>XOS@o0c0=[?dNR@Q12i0m?UNQ@Z3c0hL\\>OPAY3a0TMW>BXAZ3>`MR>VO_AZ3<lMm=jNcA\\3?UNe=_NjA^3>_N`=c1]BfN^=Y1`BkN_=SNXBa28_O^=oM^B_23E^=iMcB`2NJ^=dMgB`2JO^=^MlBa2E4]=ZMQC`2A9]=TMWC`2[O`0W>^OhAe0X>YOgAj0Y>TOfAo0[>[20000000O100O100O1O3N4K5L4K5L5J5K5L4K5L4K2OO0010O0010O0010O010O00010O010O01O010O00010O]MZ@l0g?oNh@g0W?UOXA`0h>\\OfA;Y>@WB5j=FeBO[=MTCIk<3cCC^<8g3L4Lohm5"}, "label": "two bikes on the street , behind to walking men"}]}
{"id": 463467, "image": "COCO_train2014_000000463467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing headphones\"."}], "task": "refering", "answer_template": "The \"the man wearing headphones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 46, 47, 48, 68, 69, 70, 89, 90, 91, 92, 111, 112, 113, 114, 115, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 243, 244, 245, 246, 265, 266, 267, 268, 287, 288, 290, 308, 309, 310, 312, 330, 331, 334, 352, 353, 356, 357, 378], "bbox": [0.022791068580542265, 0.082171875, 0.26885167464114834, 0.742890625], "iscrowd": 0, "area": 32415.28059999999, "rle": {"size": [640, 627], "counts": "YU91mc06J5J6K6J5K6J5K3M2N3M3M3M3M3M2N3aIPNiJT2T5TNdJo1Y5YN_Jj1]5_N[Jd1b5dNVJ^1g5kNQJX1l5POlIS1P6VOhIm0U6ZOdIh0Y6A_I`0`6HXI:e6OSI2l68jHJS7b0bH_O]7l0XHVOe7U1PHmNo7_1eGQNk8_2eF`M[9P3VFPMj9_3gE`LY:c60O1O3N1N2O2M2O1N3N1N2N2O2M2O1N3N1N2O2M2N2OPNRHQHl7R8YHiGe7Y8aH`G]7c8hHWGW7k8oHoFn6T9WIfFh6\\9^I]F`6e9fIVFW6m9l12M3N2M3N2N2N2N2N2N2N2N2N2N2N2N3M2N3M2N2N3N1N3M2N2N3M2N27J8G8I7H9G8I7H[HZEY6Y3VHh3\\1TIe6j2QHR4W1XIo6[2mG]4P1]I[7j1hGk4h0_Ig7h0WHl5K`IZ:c6`EaI_:b6[EbId:a6UEdIj:_6PEdIP;g7O1O1O1O001O1O1O001O1O1O1O001O1dNPGfGQ9m4\\JRKd5m3_KQLb4n3_KQLb4m3aKQL`4n3aKQLj4d3XKYLm5b2UJ]Mn6`1TI^NP8>QHAX>O001O1N2O1O1O1O001N2O1O1O1O001O1N[]n8"}, "label": "the man wearing headphones"}]}
{"id": 463467, "image": "COCO_train2014_000000463467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing blue sweater listening to music\"."}], "task": "refering", "answer_template": "The \"man wearing blue sweater listening to music\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 46, 47, 48, 68, 69, 70, 89, 90, 91, 92, 111, 112, 113, 114, 115, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 243, 244, 245, 246, 265, 266, 267, 268, 287, 288, 290, 308, 309, 310, 312, 330, 331, 334, 352, 353, 356, 357, 378], "bbox": [0.022791068580542265, 0.082171875, 0.26885167464114834, 0.742890625], "iscrowd": 0, "area": 32415.28059999999, "rle": {"size": [640, 627], "counts": "YU91mc06J5J6K6J5K6J5K3M2N3M3M3M3M3M2N3aIPNiJT2T5TNdJo1Y5YN_Jj1]5_N[Jd1b5dNVJ^1g5kNQJX1l5POlIS1P6VOhIm0U6ZOdIh0Y6A_I`0`6HXI:e6OSI2l68jHJS7b0bH_O]7l0XHVOe7U1PHmNo7_1eGQNk8_2eF`M[9P3VFPMj9_3gE`LY:c60O1O3N1N2O2M2O1N3N1N2N2O2M2O1N3N1N2O2M2N2OPNRHQHl7R8YHiGe7Y8aH`G]7c8hHWGW7k8oHoFn6T9WIfFh6\\9^I]F`6e9fIVFW6m9l12M3N2M3N2N2N2N2N2N2N2N2N2N2N2N3M2N3M2N2N3N1N3M2N2N3M2N27J8G8I7H9G8I7H[HZEY6Y3VHh3\\1TIe6j2QHR4W1XIo6[2mG]4P1]I[7j1hGk4h0_Ig7h0WHl5K`IZ:c6`EaI_:b6[EbId:a6UEdIj:_6PEdIP;g7O1O1O1O001O1O1O001O1O1O1O001O1dNPGfGQ9m4\\JRKd5m3_KQLb4n3_KQLb4m3aKQL`4n3aKQLj4d3XKYLm5b2UJ]Mn6`1TI^NP8>QHAX>O001O1N2O1O1O1O001N2O1O1O1O001O1N[]n8"}, "label": "man wearing blue sweater listening to music"}]}
{"id": 406121, "image": "COCO_train2014_000000406121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman is seated in front of a colorful selection of cupcakes\"."}], "task": "refering", "answer_template": "The \"a woman is seated in front of a colorful selection of cupcakes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 79, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 107, 108, 109, 110, 111, 112, 131, 132, 133, 134, 154, 155, 156, 177, 178, 200], "bbox": [0.22825, 0.0024528301886792454, 0.90259375, 0.570683962264151], "iscrowd": 0, "area": 48614.117750000005, "rle": {"size": [424, 640], "counts": "Sol12^<2VD0f;8TDJh;?RDAk;g0oC[Om;R1M2N3M2N3M2M3B?C<N2N3M2N3N1N2N3M2N2N3M2N3N1N2N3M3M2N3N2O1O001OoMlFOS9OPG1o8MTG2l8NVG1i8NYG1g8OZG0f8O\\G0d80]GOc80`GN`82aGM_83bGL^83dGL\\84fGJZ86fGJZ85hGJX86iGJV86kGIU86mGDROiNQ9c1mGAWOjNl8e1nG]OZOmNh8f1oGYO_8f0cGVO`8j0aGROb8n0_GmNe8R1\\GkNg8U1ZGgNi8Y1XGcNk8\\1WG`Nl8`1^10000O10000003M3M3M3M3M4L3M3M3M3M3M4L3M3MO1O1O1O1O1O100O1O1O10000000000O1000000000000O1000000000000O1000000000000O10000000000001O001O1O1O001O1O1O1O001O1O000000000000O1000000O1O1O1O1O1O1O1O1O1N200O1O1O100O1O1O100O1O100O1O11O001O001O001O001O001O001O001O001O1O001O001O1O001O1O001O1O1O001O1O001O0000O10000O10000O100000000000000000000000000000000000000000000000000000000001O001O1O1O001O1O001O1O1O001O1O001O1O1O009G:F000000001O000000001O000000001O000000001O0000001O1O1O1O1O1O2N1O2N1O2N1O6J7I6UIcK_4c4YK`Kf4c4TKbKj4`4QKeKm4^4mJfKR5\\4jJhKT5[4fJiKY5Y4bJlK\\5W4^JlKb5V4ZJkKg5j51N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N200O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K>B>B=C>B>B>B>B>BUbi0"}, "label": "a woman is seated in front of a colorful selection of cupcakes"}]}
{"id": 406121, "image": "COCO_train2014_000000406121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue and black shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue and black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 79, 81, 82, 83, 84, 85, 86, 87, 88, 89, 97, 98, 99, 107, 108, 109, 110, 111, 112, 131, 132, 133, 134, 154, 155, 156, 177, 178, 200], "bbox": [0.22825, 0.0024528301886792454, 0.90259375, 0.570683962264151], "iscrowd": 0, "area": 48614.117750000005, "rle": {"size": [424, 640], "counts": "Sol12^<2VD0f;8TDJh;?RDAk;g0oC[Om;R1M2N3M2N3M2M3B?C<N2N3M2N3N1N2N3M2N2N3M2N3N1N2N3M3M2N3N2O1O001OoMlFOS9OPG1o8MTG2l8NVG1i8NYG1g8OZG0f8O\\G0d80]GOc80`GN`82aGM_83bGL^83dGL\\84fGJZ86fGJZ85hGJX86iGJV86kGIU86mGDROiNQ9c1mGAWOjNl8e1nG]OZOmNh8f1oGYO_8f0cGVO`8j0aGROb8n0_GmNe8R1\\GkNg8U1ZGgNi8Y1XGcNk8\\1WG`Nl8`1^10000O10000003M3M3M3M3M4L3M3M3M3M3M4L3M3MO1O1O1O1O1O100O1O1O10000000000O1000000000000O1000000000000O1000000000000O10000000000001O001O1O1O001O1O1O1O001O1O000000000000O1000000O1O1O1O1O1O1O1O1O1N200O1O1O100O1O1O100O1O100O1O11O001O001O001O001O001O001O001O001O1O001O001O1O001O1O001O1O1O001O1O001O0000O10000O10000O100000000000000000000000000000000000000000000000000000000001O001O1O1O001O1O001O1O1O001O1O001O1O1O009G:F000000001O000000001O000000001O000000001O0000001O1O1O1O1O1O2N1O2N1O2N1O6J7I6UIcK_4c4YK`Kf4c4TKbKj4`4QKeKm4^4mJfKR5\\4jJhKT5[4fJiKY5Y4bJlK\\5W4^JlKb5V4ZJkKg5j51N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N200O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K>B>B=C>B>B>B>B>BUbi0"}, "label": "a woman wearing a blue and black shirt"}]}
{"id": 406121, "image": "COCO_train2014_000000406121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with teddy bear\"."}], "task": "refering", "answer_template": "The \"a child with teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 67, 68, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 319, 320, 321], "bbox": [0.7150000000000001, 0.07356132075471698, 1.0, 0.928183962264151], "iscrowd": 0, "area": 39898.3073, "rle": {"size": [424, 640], "counts": "hlm5=b<?@<L3M3M4L3N2M3M3M3M4M2M3M3N2N3L4M2N3M3M2N3M3L3N2N1O1O1N3N1O1O1N2O2N2M3N2N3L3N2N2M3N2M3M3M3N2M3M2O2M3M3N2N2N1O2N2N2N2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N101N2N1O2N2N2N101N2N2N2N2N2O1N2N1O2N2N2N2O1N2N1O2N2N1O2O0O2N101N101N101N2O0O2O0O2O0O2O0O101N100O100O2M2O1O1\\OVGhLl8V3XGfLk8W3ZGbLj8^3`0O100O2O0O100O101O0O100O2O0O100O101N1N2L4L4M4M1N3N2N2O1O1O0O2O001O001O001O1N101O001O010U4jKmJ"}, "label": "a child with teddy bear"}]}
{"id": 406121, "image": "COCO_train2014_000000406121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in red shirt holding a white bear\"."}], "task": "refering", "answer_template": "The \"a boy in red shirt holding a white bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 67, 68, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 319, 320, 321], "bbox": [0.7150000000000001, 0.07356132075471698, 1.0, 0.928183962264151], "iscrowd": 0, "area": 39898.3073, "rle": {"size": [424, 640], "counts": "hlm5=b<?@<L3M3M4L3N2M3M3M3M4M2M3M3N2N3L4M2N3M3M2N3M3L3N2N1O1O1N3N1O1O1N2O2N2M3N2N3L3N2N2M3N2M3M3M3N2M3M2O2M3M3N2N2N1O2N2N2N2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N101N2N1O2N2N2N101N2N2N2N2N2O1N2N1O2N2N2N2O1N2N1O2N2N1O2O0O2N101N101N101N2O0O2O0O2O0O2O0O101N100O100O2M2O1O1\\OVGhLl8V3XGfLk8W3ZGbLj8^3`0O100O2O0O100O101O0O100O2O0O100O101N1N2L4L4M4M1N3N2N2O1O1O0O2O001O001O001O1N101O001O010U4jKmJ"}, "label": "a boy in red shirt holding a white bear"}]}
{"id": 340598, "image": "COCO_train2014_000000340598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blonde fluffy puppy with belly up\"."}], "task": "refering", "answer_template": "The \"blonde fluffy puppy with belly up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 180, 181, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 271, 272, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 301, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 339, 340], "bbox": [0.028770685579196216, 0.42659375, 0.9776595744680852, 0.976171875], "iscrowd": 0, "area": 80501.79605000006, "rle": {"size": [640, 423], "counts": "fj72mc07I8H4L000000001O00000jB[OR7f0jH_OT7a0jHCT7=jHHS78jHMT74iH0U70gE]O<g0k9LiEE1d0T:GkEMG`0\\:DmE3\\O?d:^OPF;RO:m:[OQFc0hN5V;XOSFj0\\N3_;TOTFQ1SNNi;POUFY1hMJR<mNVF`3j9_LVFb3i9_LWFb3h9]LXFd3g9\\LYFe3g9ZLZFf3f9YLZFh3e9YLZFh3f9WL[Fg3f9YLZFf3h9YLXFf3i9[LWFc3k9\\LUFc3l9^LTF`3n9_LRF`3P:`LoE_3R:bLnE\\3T:cLlE\\3U:eLjEZ3X:eLiEY3X:hLgEW3[:hLeEl0`Mi0l<\\NdEl0`Me0n<_NaEm0bMa0n<bN`En0dM<n<fN^En0eM:m<hN^Eo0fM6n<kN\\EP1fM2P=nNYEQ1iMMo<TOVEP1lMIP=WOSEQ1nMEP=\\OoDR1TN]On<AmDS1YNSOn<LgDR1`NdNR=:\\DT1fNTNW=j0QDT1l<lNSCU1l<mNQCU1o<kNPCV1o<lNoBV1P=jNnBX1Q=jNmBX1R=hNlBZ1S=hNkBZ1T=fNkB[1U=fNhB]1V=dNiB]1W=dNgB^1W=cNgB_1Y=bNdBa1Z=`NdBc1[=^NbBd1]=]NaBf1]=\\N`Bf1_=[N_Bh1_=ZN^Bh1a=YN]Bj1a=XN\\Bj1c=XNZBk1d=cNmA_1R>Y2N2N1O2N1O2N1O2O0O2N1O2N1O2000O2O0O100O101N100O10001N100O100O2N1O1O1O1O2N1O1O1O1O1O0000000000000000000000000000000000000000SC[J_;e5^D^Jb;c5ZD`Jf;`5WDcJi;]5TDfJl;Z5PDkJP<T5mCnJV<Q5fCRK\\<l4aCVKb<h4[CZKi<c4TC_Ko<_4nBcKS=^4iBdKm0WOQ;Z6nDiIQ;V6nDmIQ;R6nDQJQ;n5nDTJQ;m5mDVJR;i5mDZJR;e5mD]JS;a5nD`JR;\\5PEfJP;W5QEkJn:S5SEoJm:m4VETKj:i4XEXKh:e4ZE\\Kf:b4[E_Kd:b4[E`Kd:`4ZEbKf:^4YEcKg:]4XEdKh:[4XEfKg:[4XEfKh:Z4VEhKj:X4UEiKk:W4TEjKl:V4SEkKl:V4SEkKm:U4RElKn:T4RElKn:S4REnKn:R4QEoKn:R4QEoKo:Q4QEoKo:Q4PEPLP;P4oDQLQ;o3oDQLP;P4oDQLQ;o3nDRLR;m3nDTLR;l3nDTLR;l3mDULR;l3mDULS;m3kDSLU;n3iDSLW;n3gDSLY;n3eDSL[;n3dDSLZ;o3dDRL\\;P4aDQL_;X60000O2O001O000O2O00001O0O2O00001O0O2O00001O0O2O000000000O100000000O1000000001O0000001O00000011Nh0XO7I010O000001O0000001O0000001O001O1O1O001O1O1O001O100O001O1O1O1O0O2N2O1N1O2O1N2N101N2N2O0O2N101N2N101N1O2SNdDTK\\;l4eDQK^;n4dDoJ];P5eDnJ];Q5dDlJ^;T5cDjJ_;T5dDiJ^;V5cDhJ^;X5cDeJ`;Y5cDdJ^;\\5cDbJ_;\\5cDbJ^;^5dD^J_;a5bD]J_;b5cD\\J_;c5cDYJ`;f5aDXJ`;g5bDWJ`;h5bDTJ`;l5aDQJb;m5`DPJb;P6`DmIb;Q6W101O0O2O1O0O2O0O2O1O1N2O1O2M2O1O1N3N1N2O1O2M2O1O1N2O2N1N2O1N3N1O1N3N2N1N3N2N2M2O2N2M2O2N2M3N1O2M3N2N1N3N2N2M2O2N2M3N1O2I7I7H7J7I7H8I6J7I7H8IVV5"}, "label": "blonde fluffy puppy with belly up"}]}
{"id": 29304, "image": "COCO_train2014_000000029304.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with mushrooms and no greens\"."}], "task": "refering", "answer_template": "The \"a pizza with mushrooms and no greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351], "bbox": [0.016859375, 0.33782881002087684, 0.46693750000000006, 0.9076617954070981], "iscrowd": 0, "area": 60901.42824999998, "rle": {"size": [479, 640], "counts": "W_55f>6I7I6J7gN[OoCl0l;V1M4K5K5L3L5L4K4M4K5K5L3N3L4M2N3L4M2M3N2M3N2N3L3N2M3N2M3N2M3N2N3M2N2O1N2N2N2N2N3M2N2N2N2M2O2N1O2N1O2M2O2N1O1O2N1N3N1O2N1O2M2O1O2N101O0O2O001N10001N101O0O2O001N100O2O001N101O0O2O000O2O001N101O0O10000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000000000O10001O0000000000000O10000000000000O10000O100O10000O10001N100O10000O10000O100O10000O100O2O000O10000O100O2O000O2O0O2O001N10001N101N101O0O101N101O0O2O001N100O2O001N101M3N1O2N2N2N1O2N2N2M2O2N2N2N1O2N2N1N3N2N2N1O2N2N2N1O2M3N2N2N3M2N3M2N3L4M2N3L3L5K4L5K4L5L4K4L6J5K6J6J6J5K6J6J5K6J6J5K6J6J`0@a0_O\\Wo4"}, "label": "a pizza with mushrooms and no greens"}]}
{"id": 29304, "image": "COCO_train2014_000000029304.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked circular pizza with mushroom , peppers , and white cheese\"."}], "task": "refering", "answer_template": "The \"an uncooked circular pizza with mushroom , peppers , and white cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351], "bbox": [0.016859375, 0.33782881002087684, 0.46693750000000006, 0.9076617954070981], "iscrowd": 0, "area": 60901.42824999998, "rle": {"size": [479, 640], "counts": "W_55f>6I7I6J7gN[OoCl0l;V1M4K5K5L3L5L4K4M4K5K5L3N3L4M2N3L4M2M3N2M3N2N3L3N2M3N2M3N2M3N2N3M2N2O1N2N2N2N2N3M2N2N2N2M2O2N1O2N1O2M2O2N1O1O2N1N3N1O2N1O2M2O1O2N101O0O2O001N10001N101O0O2O001N100O2O001N101O0O2O000O2O001N101O0O10000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000000000O10001O0000000000000O10000000000000O10000O100O10000O10001N100O10000O10000O100O10000O100O2O000O10000O100O2O000O2O0O2O001N10001N101N101O0O101N101O0O2O001N100O2O001N101M3N1O2N2N2N1O2N2N2M2O2N2N2N1O2N2N1N3N2N2N1O2N2N2N1O2M3N2N2N3M2N3M2N3L4M2N3L3L5K4L5K4L5L4K4L6J5K6J6J6J5K6J6J5K6J6J5K6J6J`0@a0_O\\Wo4"}, "label": "an uncooked circular pizza with mushroom , peppers , and white cheese"}]}
{"id": 29304, "image": "COCO_train2014_000000029304.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with spinach on top\"."}], "task": "refering", "answer_template": "The \"pizza with spinach on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 362, 363, 364], "bbox": [0.53315625, 0.33258872651356997, 0.9990312500000001, 0.8943841336116911], "iscrowd": 0, "area": 62110.666, "rle": {"size": [479, 640], "counts": "fio43c>;F:F9F;F:E;I7I6J7I7H8I7I7K4N3M3M3N2M3M2O2M3M3M3N2M2N3N2M3M3M3N1N3M3N2M3M3M2O2M1O2O1N1O2O1N1O2N2O0O2N2O0O2N2N101N2N101N2N1O2O1N1O2O1O001N2O001O1N101O1O0O2O1O001N2O001O1O0O2O001O0O2O001O0O2O00001N101O001O0O2O001O001O001O001O0O2O001O001O001O000O2O001O001O001O001O0O2O001O00000000000O1010O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000000O2O000O100O2O0O10001N101N2O1O1N101N2O1O0O2O1N2O1N101O1N2O0O2O1O1N2O0O2O1O1N101N2O1N2O0O2N2O1N1O2N2N2O1N1O2N2O1N2N2N2N2O1N3M2N2O1N2N2N2N2O1N2N2N2M3M4M2M3M3M3N2M3M3M3M4M4K5K5K4L5K5J6K5J5L5K5J6K5K4K6K5J6KS6"}, "label": "pizza with spinach on top"}]}
{"id": 29304, "image": "COCO_train2014_000000029304.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked spinach and feta pizza\"."}], "task": "refering", "answer_template": "The \"an uncooked spinach and feta pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 362, 363, 364], "bbox": [0.53315625, 0.33258872651356997, 0.9990312500000001, 0.8943841336116911], "iscrowd": 0, "area": 62110.666, "rle": {"size": [479, 640], "counts": "fio43c>;F:F9F;F:E;I7I6J7I7H8I7I7K4N3M3M3N2M3M2O2M3M3M3N2M2N3N2M3M3M3N1N3M3N2M3M3M2O2M1O2O1N1O2O1N1O2N2O0O2N2O0O2N2N101N2N101N2N1O2O1N1O2O1O001N2O001O1N101O1O0O2O1O001N2O001O1O0O2O001O0O2O001O0O2O00001N101O001O0O2O001O001O001O001O0O2O001O001O001O000O2O001O001O001O001O0O2O001O00000000000O1010O00000000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000000O2O000O100O2O0O10001N101N2O1O1N101N2O1O0O2O1N2O1N101O1N2O0O2O1O1N2O0O2O1O1N101N2O1N2O0O2N2O1N1O2N2N2O1N1O2N2O1N2N2N2N2O1N3M2N2O1N2N2N2N2O1N2N2N2M3M4M2M3M3M3N2M3M3M3M4M4K5K5K4L5K5J6K5J5L5K5J6K5K4K6K5J6KS6"}, "label": "an uncooked spinach and feta pizza"}]}
{"id": 168569, "image": "COCO_train2014_000000168569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small kid with tennis racket in the tennis court\"."}], "task": "refering", "answer_template": "The \"a small kid with tennis racket in the tennis court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 43, 44, 45, 53, 56, 57, 58, 59, 66, 69, 70, 71, 72, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 147, 148, 149, 150, 160, 161, 162, 163, 173, 174, 175, 176, 186, 187, 188, 189, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 224, 225, 226], "bbox": [0.10509333333333333, 0.16216, 0.7117066666666666, 1.0], "iscrowd": 0, "area": 38857.47675, "rle": {"size": [500, 375], "counts": "UUc0d0o>a0^O8I6I4M4K4N2N2O1O1OO001N2O1mNlBDU=:U1L3M4Lhj8N]UG3L3N2M4M3L4M3L0010O01O001O001O00001O001O001\\JfNlLZ1T3fNlL[1R3fNnLZ1R3fNnL[1P3fNPMZ1l2jNTMW1g2mNZMR1b2RO^Mn0]2WObMk0U2]OkMd0i1GWN9f1KXN7e1K[N6b1L]N6a1K_N5`1L_N6_1KaNP2CQN<R2@PN`0Q2^OPNa0S2\\OnMd0T2YOmMg0U2VOlMnNPObMV3]3lMoNSOaMS3]3lMoNVOaMP3\\3lMQOYO`Mm2[3lMSO\\O_Mi2[3mMTOITM\\2d3mMVOJUM[2a3mMXOKVMY2_3nMYOLWMX2\\3nM\\OLWMX2Y3nM_OLVMX2X3nMALVMX2U3nMDLVMW2S3oMFb45bKJ`4NfK1\\4HjK7Y4BlK=[4WOkKi0]4iNiKV1_4hMZLY2]83N2N3M2N3M3N2eN[1N2N2O1N2N2N2N2]Oc0N2M3N2M3N2N3L3N2N2O1N3M2O1O12O0O2N2N101N102N3L3N3L4M2N3Lg0ZO6I4M1N2O1O1N2O1N1010N1O2O0O2N100O3gF_KY7d4aH`K_7b4\\HbKb7`4YHdKf7_4THeKl7^4iGkKU8e4TGcKk8c510O1O100O001O1cIQJR3Q6mL`Ja0kNlNe6b0QKLbN@^6d0_KVO[N4W6g0cKhN^N?P6j0cK^NeNe0j5n0bKTNlNl0c5Q1bKiMUOS1[5T1bMjNa2U1_MiNc2W1]MfNg2Z1XMdNk2\\1TMbNo2^1PM`NS3^1SMUNW3i1\\5N2N3M2N1O1O1O1O001O1O1O100O2N1O2N1O1O2O0M4M2N2N3M3M2N3M3M3M3M4K5F]^d1"}, "label": "a small kid with tennis racket in the tennis court"}]}
{"id": 168569, "image": "COCO_train2014_000000168569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child holds a tennis racket\"."}], "task": "refering", "answer_template": "The \"a child holds a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 43, 44, 45, 53, 56, 57, 58, 59, 66, 69, 70, 71, 72, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 147, 148, 149, 150, 160, 161, 162, 163, 173, 174, 175, 176, 186, 187, 188, 189, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 224, 225, 226], "bbox": [0.10509333333333333, 0.16216, 0.7117066666666666, 1.0], "iscrowd": 0, "area": 38857.47675, "rle": {"size": [500, 375], "counts": "UUc0d0o>a0^O8I6I4M4K4N2N2O1O1OO001N2O1mNlBDU=:U1L3M4Lhj8N]UG3L3N2M4M3L4M3L0010O01O001O001O00001O001O001\\JfNlLZ1T3fNlL[1R3fNnLZ1R3fNnL[1P3fNPMZ1l2jNTMW1g2mNZMR1b2RO^Mn0]2WObMk0U2]OkMd0i1GWN9f1KXN7e1K[N6b1L]N6a1K_N5`1L_N6_1KaNP2CQN<R2@PN`0Q2^OPNa0S2\\OnMd0T2YOmMg0U2VOlMnNPObMV3]3lMoNSOaMS3]3lMoNVOaMP3\\3lMQOYO`Mm2[3lMSO\\O_Mi2[3mMTOITM\\2d3mMVOJUM[2a3mMXOKVMY2_3nMYOLWMX2\\3nM\\OLWMX2Y3nM_OLVMX2X3nMALVMX2U3nMDLVMW2S3oMFb45bKJ`4NfK1\\4HjK7Y4BlK=[4WOkKi0]4iNiKV1_4hMZLY2]83N2N3M2N3M3N2eN[1N2N2O1N2N2N2N2]Oc0N2M3N2M3N2N3L3N2N2O1N3M2O1O12O0O2N2N101N102N3L3N3L4M2N3Lg0ZO6I4M1N2O1O1N2O1N1010N1O2O0O2N100O3gF_KY7d4aH`K_7b4\\HbKb7`4YHdKf7_4THeKl7^4iGkKU8e4TGcKk8c510O1O100O001O1cIQJR3Q6mL`Ja0kNlNe6b0QKLbN@^6d0_KVO[N4W6g0cKhN^N?P6j0cK^NeNe0j5n0bKTNlNl0c5Q1bKiMUOS1[5T1bMjNa2U1_MiNc2W1]MfNg2Z1XMdNk2\\1TMbNo2^1PM`NS3^1SMUNW3i1\\5N2N3M2N1O1O1O1O001O1O1O100O2N1O2N1O1O2O0M4M2N2N3M3M2N3M3M3M3M4K5F]^d1"}, "label": "a child holds a tennis racket"}]}
{"id": 348794, "image": "COCO_train2014_000000348794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant carrying two people on its back and a stick in its mouth\"."}], "task": "refering", "answer_template": "The \"an elephant carrying two people on its back and a stick in its mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 328, 329, 330, 331], "bbox": [0.106453125, 0.3707962529274005, 0.45428125, 0.9865105386416863], "iscrowd": 0, "area": 42546.0385, "rle": {"size": [427, 640], "counts": "Tcl02U=6J6J6J7I6J6J6J6J7J5J6J6K5J6K6L3M3M3M3M4L3M3M3M3M3M4L3M3M3M3M3M4L8kHZKW5P5YJWKf5T5hITKW6V5YIQKf6j51N10000O1000001N1000000O101OOQNXI\\Mi6d2bIQM^6P3kIfLU6[3UJZLk5f3`JnK`5T4jJaKV5`4j1001N101O1O0O2O001N2hHfK]5T600O10000O2O000O100O10000O101O0O10000O1000001N1000000O1000000O2TJ`Ik4`7000000O1000000O100000000O11O001O00001O001O001O00001O001O001O00001O001O00001O00000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000_NaKVJ_4h5eKVJ[4i5iKTJW4k5lKSJT4k5QLRJo3m5TLQJl3m5YLPJg3o5k1O1N2O2M4H8F:F9PNWHTNS8h1eHZMe7c2d1L4L4L4M3L3M7J<C;E4L5L3L4L5L3L5K4LeYa4"}, "label": "an elephant carrying two people on its back and a stick in its mouth"}]}
{"id": 348794, "image": "COCO_train2014_000000348794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with the yellow strap\"."}], "task": "refering", "answer_template": "The \"elephant with the yellow strap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 328, 329, 330, 331], "bbox": [0.106453125, 0.3707962529274005, 0.45428125, 0.9865105386416863], "iscrowd": 0, "area": 42546.0385, "rle": {"size": [427, 640], "counts": "Tcl02U=6J6J6J7I6J6J6J6J7J5J6J6K5J6K6L3M3M3M3M4L3M3M3M3M3M4L3M3M3M3M3M4L8kHZKW5P5YJWKf5T5hITKW6V5YIQKf6j51N10000O1000001N1000000O101OOQNXI\\Mi6d2bIQM^6P3kIfLU6[3UJZLk5f3`JnK`5T4jJaKV5`4j1001N101O1O0O2O001N2hHfK]5T600O10000O2O000O100O10000O101O0O10000O1000001N1000000O1000000O2TJ`Ik4`7000000O1000000O100000000O11O001O00001O001O001O00001O001O001O00001O001O00001O00000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000_NaKVJ_4h5eKVJ[4i5iKTJW4k5lKSJT4k5QLRJo3m5TLQJl3m5YLPJg3o5k1O1N2O2M4H8F:F9PNWHTNS8h1eHZMe7c2d1L4L4L4M3L3M7J<C;E4L5L3L4L5L3L5K4LeYa4"}, "label": "elephant with the yellow strap"}]}
{"id": 348794, "image": "COCO_train2014_000000348794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with long trimmed tusks carrying a man and girl in light neutral colored clothes\"."}], "task": "refering", "answer_template": "The \"an elephant with long trimmed tusks carrying a man and girl in light neutral colored clothes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 290, 291, 292, 293, 294, 314, 315, 316, 317, 338, 339], "bbox": [0.604796875, 0.301288056206089, 0.8248906250000001, 0.9870023419203748], "iscrowd": 0, "area": 27616.734699999997, "rle": {"size": [427, 640], "counts": "jaQ5=\\<i0\\Od0H9G8H8H3N1N2N2N2N2N2N2O1N3M2N2N2O1N2N2N2O1N2N6K5K5K4L5iHiLP4\\3eJlMY5Y2fJdMY5a2gJ\\MW5h2iJVMU5o2mJkLR5Z3oJbLo4c3SKXLk4l3XKmKh4X4YKaKh4d4ZKTKg4P5\\KhJe4]5\\K]Jd4g5_KQJb4S6aKeI`4_6T14L4L5J5L4L4L4L4L1O2N1O2N1O2M2O2N1O2M2O2N1O2M2O=C5J10000O10O10001O00001N101O001O001O00001O001N1N3L3N3D;O2N1O101N1O2N1O2N1O2N1O2G8G9G:F:F:F:GYJ]JW4e5aK`J`4b5WKdJj4^5nJgJS5\\5cJjJ^5[4VJbK39h5k3aJlK_O>Q6\\3mJXMT5]2YK`Mh4V2eKfM\\4P2PLmMR4h1ZLUNg3a1fL[N\\3Y1QMdNR3V1UMfNn2T1XMiNk2Q1[MlNh2n0_MnNc2m0cMPO`2j0gMQO]2j0hMRO\\2h0jMTOZ2f0Z5J6J5K6J[`^1"}, "label": "an elephant with long trimmed tusks carrying a man and girl in light neutral colored clothes"}]}
{"id": 348794, "image": "COCO_train2014_000000348794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant that is to the far right\"."}], "task": "refering", "answer_template": "The \"an elephant that is to the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 290, 291, 292, 293, 294, 314, 315, 316, 317, 338, 339], "bbox": [0.604796875, 0.301288056206089, 0.8248906250000001, 0.9870023419203748], "iscrowd": 0, "area": 27616.734699999997, "rle": {"size": [427, 640], "counts": "jaQ5=\\<i0\\Od0H9G8H8H3N1N2N2N2N2N2N2O1N3M2N2N2O1N2N2N2O1N2N6K5K5K4L5iHiLP4\\3eJlMY5Y2fJdMY5a2gJ\\MW5h2iJVMU5o2mJkLR5Z3oJbLo4c3SKXLk4l3XKmKh4X4YKaKh4d4ZKTKg4P5\\KhJe4]5\\K]Jd4g5_KQJb4S6aKeI`4_6T14L4L5J5L4L4L4L4L1O2N1O2N1O2M2O2N1O2M2O2N1O2M2O=C5J10000O10O10001O00001N101O001O001O00001O001N1N3L3N3D;O2N1O101N1O2N1O2N1O2N1O2G8G9G:F:F:F:GYJ]JW4e5aK`J`4b5WKdJj4^5nJgJS5\\5cJjJ^5[4VJbK39h5k3aJlK_O>Q6\\3mJXMT5]2YK`Mh4V2eKfM\\4P2PLmMR4h1ZLUNg3a1fL[N\\3Y1QMdNR3V1UMfNn2T1XMiNk2Q1[MlNh2n0_MnNc2m0cMPO`2j0gMQO]2j0hMRO\\2h0jMTOZ2f0Z5J6J5K6J[`^1"}, "label": "an elephant that is to the far right"}]}
{"id": 479867, "image": "COCO_train2014_000000479867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse figurine on top of a cake\"."}], "task": "refering", "answer_template": "The \"a brown horse figurine on top of a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 237, 238], "bbox": [0.20484375, 0.2516041666666667, 0.42742187499999995, 0.6387083333333333], "iscrowd": 0, "area": 14155.74995, "rle": {"size": [480, 640], "counts": "bdm1l0f=>A`0E:01N101N1O2O0O2O0O2O0O2O0O2N010O01O0M4K4N3M2O1O2N1N3N1O2M2011N1O2O1N2O1N2O1N101N2N2O1N2N1O2O1N2N2N3N2M3M4L3N2M3M4L3N2M4L2N2O1N1O2N2N1O2M3N2N1O2N2N1O2N2N1O2N2N1O2N2N2M2O2N2N1NkMgFhNW9X1UG^Ni8a1_GYN`8g1cGXN[8g1hGYNV8g1mGWNS8h1PHVNP8i1RHVNm7j1WHSNi7l1]HmMd7S2aHeMc7Z2cH]M`7c2_2O1O00XD]MU;b2iDaMV;_2hDeMW;Y2hDkMV;U2hDnMX;P2gDTNW;k1gDYNY;e1fD_NY;_1fDdNY;[1fDiNY;V1eDmNZ;^2O0O2O1N101O0100O100O010O1O1M3M3L4M3M2N3L6Kg0YO7[Oe0@?A`0@k\\[5"}, "label": "a brown horse figurine on top of a cake"}]}
{"id": 479867, "image": "COCO_train2014_000000479867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown horse on left\"."}], "task": "refering", "answer_template": "The \"brown horse on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 237, 238], "bbox": [0.20484375, 0.2516041666666667, 0.42742187499999995, 0.6387083333333333], "iscrowd": 0, "area": 14155.74995, "rle": {"size": [480, 640], "counts": "bdm1l0f=>A`0E:01N101N1O2O0O2O0O2O0O2O0O2N010O01O0M4K4N3M2O1O2N1N3N1O2M2011N1O2O1N2O1N2O1N101N2N2O1N2N1O2O1N2N2N3N2M3M4L3N2M3M4L3N2M4L2N2O1N1O2N2N1O2M3N2N1O2N2N1O2N2N1O2N2N1O2N2N2M2O2N2N1NkMgFhNW9X1UG^Ni8a1_GYN`8g1cGXN[8g1hGYNV8g1mGWNS8h1PHVNP8i1RHVNm7j1WHSNi7l1]HmMd7S2aHeMc7Z2cH]M`7c2_2O1O00XD]MU;b2iDaMV;_2hDeMW;Y2hDkMV;U2hDnMX;P2gDTNW;k1gDYNY;e1fD_NY;_1fDdNY;[1fDiNY;V1eDmNZ;^2O0O2O1N101O0100O100O010O1O1M3M3L4M3M2N3L6Kg0YO7[Oe0@?A`0@k\\[5"}, "label": "brown horse on left"}]}
{"id": 479867, "image": "COCO_train2014_000000479867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and gray horse figurine\"."}], "task": "refering", "answer_template": "The \"a white and gray horse figurine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 127, 128, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244], "bbox": [0.5177343750000001, 0.28602083333333334, 0.6822656250000001, 0.6172083333333334], "iscrowd": 0, "area": 9199.294049999999, "rle": {"size": [480, 640], "counts": "RZk45i>7J7H8IP1oN7I7J2N100O0GiMkCV2U<lMiCT2X<lMhCS2X<mMhCP2\\<PNdCn1^<;00O003M6J9H5J2N2]EfL]9[3[FjLf9X3UFiLn9]3iEdLX:R412M4M2M4M2MN3N1N3N1N3N1N3M3N1N3M3L3N3K5M201O1O001O1O8H;ESDhMT;Q33N1N2N2N2O1N1O2M2M4M3L3M4M2M4L3N2M3M3M4L3M3M3N2O1O1O1O1O0102L3N2M001O100000000O1000YOg0N2N2N2TOaB5a=IgBOdjo2"}, "label": "a white and gray horse figurine"}]}
{"id": 479867, "image": "COCO_train2014_000000479867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and gray pony in middle of cake\"."}], "task": "refering", "answer_template": "The \"white and gray pony in middle of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 127, 128, 150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244], "bbox": [0.5177343750000001, 0.28602083333333334, 0.6822656250000001, 0.6172083333333334], "iscrowd": 0, "area": 9199.294049999999, "rle": {"size": [480, 640], "counts": "RZk45i>7J7H8IP1oN7I7J2N100O0GiMkCV2U<lMiCT2X<lMhCS2X<mMhCP2\\<PNdCn1^<;00O003M6J9H5J2N2]EfL]9[3[FjLf9X3UFiLn9]3iEdLX:R412M4M2M4M2MN3N1N3N1N3N1N3M3N1N3M3L3N3K5M201O1O001O1O8H;ESDhMT;Q33N1N2N2N2O1N1O2M2M4M3L3M4M2M4L3N2M3M3M4L3M3M3N2O1O1O1O1O0102L3N2M001O100000000O1000YOg0N2N2N2TOaB5a=IgBOdjo2"}, "label": "white and gray pony in middle of cake"}]}
{"id": 479867, "image": "COCO_train2014_000000479867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black horse decoration on a cake\"."}], "task": "refering", "answer_template": "The \"a black horse decoration on a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 38, 39, 40, 41, 58, 59, 60, 61, 62, 81, 82, 83, 84, 104, 107], "bbox": [0.5280781250000001, 0.014583333333333334, 0.8083125000000001, 0.26295833333333335], "iscrowd": 0, "area": 5958.030849999996, "rle": {"size": [480, 640], "counts": "X`n42m>4L3M2N1O2N2N3M6J3M2N3N2dBnNe<T1UCTOg<o0SCWOj<k0PC[Oo<\\1N1O2O1O1NH9M3M3M3M3O1O1O1O1O2O1N2N2N2N1O100O2N1000001O000001O00000000001O001O1O001O001O0000000000000000000000000001O000000aBnNP=Q1oBROP=m0PCUOo<k0oBXOP=g0oB[OQ=d0mB_OS=a0jBBV==iBEV=;hBHX=o01O0000000000000000000O100000O2O00000000000O10I7O1O1O0O3N1N3M3N1N3N2M2N3N2M20100O001O1O1O001O1O001O1O001O01O100O1O1O1O1O100O1O1O1O1O2M2O1O1O1O1?A001OO100O100O100O1000EfA0Z>NjA0W>MlA2T>MmA3a>O0O101O0OnRj1"}, "label": "a black horse decoration on a cake"}]}
{"id": 217725, "image": "COCO_train2014_000000217725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back chair , to the right of a window\"."}], "task": "refering", "answer_template": "The \"the back chair , to the right of a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 224, 225, 226, 247, 248], "bbox": [0.6465625, 0.3690208333333333, 0.87325, 0.62725], "iscrowd": 0, "area": 9319.7999, "rle": {"size": [480, 640], "counts": "]YR65e>7I6I7M4N10001N100O100O2O0O10000O2O0O100O100O2O0O1000001O0001O0001O02N5K4SOXB:X>00O001O001O01O00000mNAeC?V<FjC;Q<IoC7l;NTD2h;1YDOc;5]DKc;5]DLb;4^DLa;5`DJ`;6`DJ`;6`DJ`;5aDK_;5aDL^;4bDL];5cDK];5cDK];5cDK];5cDL];2dDN\\;2dDN\\;2dDN\\;2dDN\\;2cD0\\;0dD0FBd:>fE0_OIk:6gE4SOOV;LgEU1Y:AQE?o:JhD6X;ObD3];h10001O01O0001O00001O0001O01O1O1O2N00000001O0000000O100K5O2N100O1O1O100O1O2N1O100O1O1O10001O0O1000000000001iMTDT1P<hNUDT1n;hNVDT1n;hNVDT1m<L4L5K4L4K5L5K4L4L8HUhU1"}, "label": "the back chair , to the right of a window"}]}
{"id": 217725, "image": "COCO_train2014_000000217725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather chair turned halfway towards a coffee table with a couch to its left\"."}], "task": "refering", "answer_template": "The \"a black leather chair turned halfway towards a coffee table with a couch to its left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351], "bbox": [0.110234375, 0.49322916666666666, 0.45339062500000005, 0.9199375], "iscrowd": 0, "area": 33487.932349999995, "rle": {"size": [480, 640], "counts": "oaQ15i>6J5K6J6J5K6J6J5K8H`0@8H8I8H7I6J7J5J6J6J6K5J6J3M3N1N3M3M2N3M3M2N3M2O2M3M2N2N2N2N2N1O2O1O100O00100O1O01O000O2O000O2O0O1O2F:N101O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O001O0000000000001O0000000000000O10000000000000O1O100O1O1O100O10O1000O100O100O1000000O1001O1O2N1O2M100000000000000000000000000000000O100000000000000000001O000000000O10001O00000000001O00000000001O000O1000001O010O1000000O2O00000O2N2M;F7H100O1O100O1O1N2O2M2O2N1O2N1O2N1O2N1O3VOeDjM];c1[EVNh:V1l1XORiS5"}, "label": "a black leather chair turned halfway towards a coffee table with a couch to its left"}]}
{"id": 217725, "image": "COCO_train2014_000000217725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair to the left of the table\"."}], "task": "refering", "answer_template": "The \"the chair to the left of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351], "bbox": [0.110234375, 0.49322916666666666, 0.45339062500000005, 0.9199375], "iscrowd": 0, "area": 33487.932349999995, "rle": {"size": [480, 640], "counts": "oaQ15i>6J5K6J6J5K6J6J5K8H`0@8H8I8H7I6J7J5J6J6J6K5J6J3M3N1N3M3M2N3M3M2N3M2O2M3M2N2N2N2N2N1O2O1O100O00100O1O01O000O2O000O2O0O1O2F:N101O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O001O0000000000001O0000000000000O10000000000000O1O100O1O1O100O10O1000O100O100O1000000O1001O1O2N1O2M100000000000000000000000000000000O100000000000000000001O000000000O10001O00000000001O00000000001O000O1000001O010O1000000O2O00000O2N2M;F7H100O1O100O1O1N2O2M2O2N1O2N1O2N1O2N1O3VOeDjM];c1[EVNh:V1l1XORiS5"}, "label": "the chair to the left of the table"}]}
{"id": 520831, "image": "COCO_train2014_000000520831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child ' s skis\"."}], "task": "refering", "answer_template": "The \"child ' s skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 276, 277, 290, 291, 292, 295, 304, 305, 306, 310, 318, 319, 320, 325, 333, 334, 340], "bbox": [0.22515222482435598, 0.73990625, 0.7292505854800937, 0.9865468749999999], "iscrowd": 0, "area": 6131.880549999998, "rle": {"size": [640, 427], "counts": "Zcl11nc01O2N1O2M3N1O2N1O2N2N1O2N1O2M2O1O1O001O1O1O1O1O0O2O1O1O1O1O1O001O1N10O1000000O1000000O010000000O1000000O10O100000O1000000O1000O1000O100000F92O001O1O1N101O1O1N101O1O1N101O1O1N101O1O1N101O000O1000000O10001O0O1000000O10000O1000000O100000N1M4M3K5N110001O0O2O001N101O000O2O1O2N2M3N3M2N2M3N2N2NlUl0d2agSO01O001O00010O001O001O0010O01O00001O001C<C>B=B?BSoW2"}, "label": "child ' s skis"}]}
{"id": 520831, "image": "COCO_train2014_000000520831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis of a woman by a child\"."}], "task": "refering", "answer_template": "The \"the skis of a woman by a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 229, 230, 243, 244, 256, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 300, 301, 302, 316], "bbox": [0.0, 0.63078125, 0.4243793911007026, 0.9184218749999999], "iscrowd": 0, "area": 6743.328599999996, "rle": {"size": [640, 427], "counts": "ga0=jb0h00000000000000000001O000O10000000000000001O0000P^OoNRa0Q1n^OPOQa0P1o^OQOPa0o0Q_OQOn`0o0R_OSOl`0m0T_OTOk`0l0V_OTOi`0l0W_OUOh`0k0X_OVOh`0i0X_OXOg`0h0Z_OXOe`0h0[_OYOd`0g0\\_OZOc`0f0^_O[O``0e0`_O\\O_`0c0b_O^O]`0b0c_O_O\\`0a0e_O_OZ`0a0f_O@Y`0`0g_OAY`0>h_OBW`0>i_OCV`0=j_OET`0;l_OFS`0:n_OFQ`0:o_OFQ`0:o_OGP`09P@GP`09P@GP`09Q@Fo?:Q@FP`09P@GP`09P@Ho?8Q@Ho?8Q@Ho?8R@Gn?9R@Gn?9R@Hm?8S@Hm?8S@Hm?8S@Hn?6T@Il?7T@Jk?6U@Jk?6U@Jl?5T@Kl?5U@Jk?6U@Kj?5V@Kj?5V@Kk?4U@Lk?4U@Lk?4V@Li?4W@Lj?3V@Mj?3V@Mj?3V@Mj?3W@Mh?3X@Mi?2W@Ni?2W@Ni?2W@Ni?2W@Oi?0X@Oh?1X@Oh?1X@Oh?1X@Oi?OX@2g?NZ@1f?OZ@1f?OZ@1f?OZ@1g?NY@2g?NY@3f?M[@2e?N[@2f?MZ@3f?MZ@3f?MZ@4e?L[@4e?L\\@3d?M\\@3d?M\\@3d?M\\@4c?L]@4c?L^@3b?M^@3b?M^@3b?M^@4a?L_@4a?L_@4a?L`@3`?M`@3`?La@5^?Kb@5^?Kb@5^?Kc@4]?Lc@4]?Lc@5\\?Kd@5\\?Kd@5\\?Kd@5\\?Ke@4[?Le@5Z?Kf@5Z?Kf@5Z?Kf@5Z?Kg@4Y?Lg@5X?Kh@5X?Kh@5X?Kh@5X?Kh@5X?Ki@4W?Lj@4U?Lk@4U?Ll@3T?Mm@2S?Nm@2S?Mo@3P?MPA3P?MQA2o>NRA1n>ORA1n>OSA1l>OUA0k>0UA0k>MYA2g>H_A8a>BfA>n`0O001O1O010O001O1O010O1O001O10O01O001O10O01O1O1O2N100OPYi4"}, "label": "the skis of a woman by a child"}]}
{"id": 520831, "image": "COCO_train2014_000000520831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis with a woman standing on them next to a little boy also on skis\"."}], "task": "refering", "answer_template": "The \"skis with a woman standing on them next to a little boy also on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 229, 230, 243, 244, 256, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 300, 301, 302, 316], "bbox": [0.0, 0.63078125, 0.4243793911007026, 0.9184218749999999], "iscrowd": 0, "area": 6743.328599999996, "rle": {"size": [640, 427], "counts": "ga0=jb0h00000000000000000001O000O10000000000000001O0000P^OoNRa0Q1n^OPOQa0P1o^OQOPa0o0Q_OQOn`0o0R_OSOl`0m0T_OTOk`0l0V_OTOi`0l0W_OUOh`0k0X_OVOh`0i0X_OXOg`0h0Z_OXOe`0h0[_OYOd`0g0\\_OZOc`0f0^_O[O``0e0`_O\\O_`0c0b_O^O]`0b0c_O_O\\`0a0e_O_OZ`0a0f_O@Y`0`0g_OAY`0>h_OBW`0>i_OCV`0=j_OET`0;l_OFS`0:n_OFQ`0:o_OFQ`0:o_OGP`09P@GP`09P@GP`09Q@Fo?:Q@FP`09P@GP`09P@Ho?8Q@Ho?8Q@Ho?8R@Gn?9R@Gn?9R@Hm?8S@Hm?8S@Hm?8S@Hn?6T@Il?7T@Jk?6U@Jk?6U@Jl?5T@Kl?5U@Jk?6U@Kj?5V@Kj?5V@Kk?4U@Lk?4U@Lk?4V@Li?4W@Lj?3V@Mj?3V@Mj?3V@Mj?3W@Mh?3X@Mi?2W@Ni?2W@Ni?2W@Ni?2W@Oi?0X@Oh?1X@Oh?1X@Oh?1X@Oi?OX@2g?NZ@1f?OZ@1f?OZ@1f?OZ@1g?NY@2g?NY@3f?M[@2e?N[@2f?MZ@3f?MZ@3f?MZ@4e?L[@4e?L\\@3d?M\\@3d?M\\@3d?M\\@4c?L]@4c?L^@3b?M^@3b?M^@3b?M^@4a?L_@4a?L_@4a?L`@3`?M`@3`?La@5^?Kb@5^?Kb@5^?Kc@4]?Lc@4]?Lc@5\\?Kd@5\\?Kd@5\\?Kd@5\\?Ke@4[?Le@5Z?Kf@5Z?Kf@5Z?Kf@5Z?Kg@4Y?Lg@5X?Kh@5X?Kh@5X?Kh@5X?Kh@5X?Ki@4W?Lj@4U?Lk@4U?Ll@3T?Mm@2S?Nm@2S?Mo@3P?MPA3P?MQA2o>NRA1n>ORA1n>OSA1l>OUA0k>0UA0k>MYA2g>H_A8a>BfA>n`0O001O1O010O001O1O010O1O001O10O01O001O10O01O1O1O2N100OPYi4"}, "label": "skis with a woman standing on them next to a little boy also on skis"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult on skis\"."}], "task": "refering", "answer_template": "The \"an adult on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 58, 59, 60, 61, 62, 81, 82, 83, 84, 104, 105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 177, 196, 197, 198, 199, 219, 220, 221], "bbox": [0.50940625, 0.0022500000000000003, 0.766015625, 0.7325833333333334], "iscrowd": 0, "area": 22964.9491, "rle": {"size": [360, 640], "counts": "bcb35S;;E4L4L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M4K4M3M3M3M3M3M3M3M4ZLmLoNV3g0RMXOQ3a0TM^Oo2>RMBQ3:QMER37PMHS34oLKT30oLOR3OoL1Q3NQM1o2NSM1m2NUM1k2NWM1i2MYM3g2LZM4f2K[M5e2J\\M6d2I]M7c2H^M8b2G_M9a2EaM;_2DbM<^2CcM=]2BdM>\\2AeM?[2@eMa0[2`NdN`1b500000000000000000000000000000000000000000O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1dLiLfNX3W1VM]Nk2`1ZM^Nf2_1`M]Na2`1dM]N]2_1jM]NW2`1nM]NS2`1SN\\Nn1a1WN\\Nj1a1[N\\Nf1`1`N[Nc1b1cNYN_1d1fNXN\\1e1iNWNY1f1lNVNV1h1nNSNU1l1nNPNT1o1POlMR1R2ROjMP1U2SOgMo0X2TOcMo0[2P4O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2M3M3L4M3L4Mbgd1"}, "label": "an adult on skis"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing yellow shoes and red skis\"."}], "task": "refering", "answer_template": "The \"a person wearing yellow shoes and red skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 58, 59, 60, 61, 62, 81, 82, 83, 84, 104, 105, 106, 107, 128, 129, 130, 151, 152, 153, 174, 175, 176, 177, 196, 197, 198, 199, 219, 220, 221], "bbox": [0.50940625, 0.0022500000000000003, 0.766015625, 0.7325833333333334], "iscrowd": 0, "area": 22964.9491, "rle": {"size": [360, 640], "counts": "bcb35S;;E4L4L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M4K4M3M3M3M3M3M3M3M4ZLmLoNV3g0RMXOQ3a0TM^Oo2>RMBQ3:QMER37PMHS34oLKT30oLOR3OoL1Q3NQM1o2NSM1m2NUM1k2NWM1i2MYM3g2LZM4f2K[M5e2J\\M6d2I]M7c2H^M8b2G_M9a2EaM;_2DbM<^2CcM=]2BdM>\\2AeM?[2@eMa0[2`NdN`1b500000000000000000000000000000000000000000O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1dLiLfNX3W1VM]Nk2`1ZM^Nf2_1`M]Na2`1dM]N]2_1jM]NW2`1nM]NS2`1SN\\Nn1a1WN\\Nj1a1[N\\Nf1`1`N[Nc1b1cNYN_1d1fNXN\\1e1iNWNY1f1lNVNV1h1nNSNU1l1nNPNT1o1POlMR1R2ROjMP1U2SOgMo0X2TOcMo0[2P4O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2M3M3L4M3L4Mbgd1"}, "label": "a person wearing yellow shoes and red skis"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top blue railing behind the skiers\"."}], "task": "refering", "answer_template": "The \"the top blue railing behind the skiers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 32, 33, 34, 40, 41, 42, 43, 44, 45, 56, 79], "bbox": [0.41334375, 0.006750000000000001, 0.9973281249999999, 0.2719166666666667], "iscrowd": 0, "area": 4090.663999999998, "rle": {"size": [360, 640], "counts": "_Vm27k:6I7I7I8I6L4000001O0000002N4L4L4L4L40O4M3M3M2NO1L4KaNoFi0Q9WOXG`0h8_ObG8^8HkG0T80THHl78b10000000000000000000000000000000000000O100000000000O100000000000002N2N2N1O2N2NaS_1O^l`N2O1O1O1O1O1O1O1O1O1O1O1O100000O1000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O1000O100000000000000000000000000ce0"}, "label": "the top blue railing behind the skiers"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue ski rack with a child in front\"."}], "task": "refering", "answer_template": "The \"a blue ski rack with a child in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 32, 33, 34, 40, 41, 42, 43, 44, 45, 56, 79], "bbox": [0.41334375, 0.006750000000000001, 0.9973281249999999, 0.2719166666666667], "iscrowd": 0, "area": 4090.663999999998, "rle": {"size": [360, 640], "counts": "_Vm27k:6I7I7I8I6L4000001O0000002N4L4L4L4L40O4M3M3M2NO1L4KaNoFi0Q9WOXG`0h8_ObG8^8HkG0T80THHl78b10000000000000000000000000000000000000O100000000000O100000000000002N2N2N1O2N2NaS_1O^l`N2O1O1O1O1O1O1O1O1O1O1O1O100000O1000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O1000O100000000000000000000000000ce0"}, "label": "a blue ski rack with a child in front"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red skiis\"."}], "task": "refering", "answer_template": "The \"red skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 280, 281], "bbox": [0.15421875000000002, 0.6118055555555556, 0.795078125, 0.951111111111111], "iscrowd": 0, "area": 7197.335249999991, "rle": {"size": [360, 640], "counts": "W_S11V;101O001N1ROMgF4X90eFO\\91cF0\\91cFO]91cF0]90bF1]9OdF0]90bF1]9OcF1]90bF1^9NbF2^9ObF1^9NbF2^9OaF2^9NbF2^9OaF2^9NbF2^9ObF1]9OcF1]90bF1]9OcF1]9OcF2\\9OdF0\\90dF0\\91cFO^90bF0^90bF0^90bFO_91aFO_90bF0^90bF0^90bF0^90aF0`90`F0`9OaF1_9OaF1_9OaF0`90`F0_91aFO_90bF0^90bF0^90aF0`90`F0`90`F0`9OaF1^90bF0^90bFO_91aFO_91aFO_91`F0`9OaF0_91aFO_91aFO_91aFO_91aFO_90bFO_91aFO_91aFO^92aFO_91aFO_90bFO_91aFO_91aFO_91aFO^92bFN^91cFN^92bFN^92aFO_91aFO_91aFN`92`FN_92aFO_91aFO_91aFO_91`FOa91_FOa90`F0_91`F0`90`F0`90`FOa91^F0b9O_F1a9O_F1a9O^F2a9O_F0b90^F0b90]F1c9N^F1c9O]F1c9O]F1b90]F0d9O]F1c9O]F1c9O\\F2d9N\\F1e9O[F1e9N[F3d9N\\F1e9O[F1e9OZF2f9NZF1g9NZF2f9NYF3f9NZF2f9NZF1g9NYF3g9MYF3f9NZF1g9OXF2h9NXF2h9MYF2g9OXF2h9NXF2h9NXF2h9MYF2g9OXF2h9NXF2h9NXF1i9OVF2i9NXF2h9NXF1i9OVF2j9NVF2i9NXF2h9NWF2j9NVF2j9NVF2i9OVF1k9NVF2j9NVF2j9NUF2k9OUF1k9OUF1k9NUF3k9MUF2k9OUF1k9OUF1k9NUF2l9NTF2k9OUF1k9OTF1m9OSF1m9NTF2k9OTF2l9NTF1m9OSF1m9NSF3l9NTF1m9OSF1m9ORF2n9NRF1n9OSF1m9ORF2n9NRF2n9NRF1n90QF1o9NRF2n9NRF1o9OPF2o9OQF1o9NRF1o9OQF1o9OPF2o9OQF1o9OQF0P:OPF2P:NPF2o9OQF0P:0oE1Q:NPF2P:NPF1P:0oE1Q:OoE1Q:OoE0R:OnE2Q:OoE1Q:OoE1Q:OnE1S:OmE1S:NnE2Q:OnE1S:OmE1S:OmE1S:NnE1R:0mE1S:OmE1S:OmE1S:OlE1T:OmE1S:OmE1S:OlE1U:OkE1T:OmE1S:OlE1U:OkE1U:OkE1T:0kE1U:NlE1U:OkE1U:OjE2U:OkE0V:0jE0V:OjE2V:NjE1V:0jE0V:0iE1W:NjE2V:NjE1V:0jE0V:0iE1W:OiE0X:OiE1V:0iE1W:OiE1W:OiE0X:OhE2W:OiE1W:OiE1W:OhE1Y:OgE1X:OiE1W:OhE1Y:OgE1Y:OgE1X:0gE1Y:NhE1Y:OgE1Y:OfE2Y:OgE1Y:NhE1Y:OgE1Y:OfE2Y:OgE0Z:0fE0Z:OfE2Z:NfE2Y:OgE0Z:0eE1[:NfE2Z:NfE2Y:OfE1[:OeE1[:OeE1[:NeE2[:OeE1[:OeE1[:OdE2\\:NdE1\\:OeE1[:OdE2\\:NdE2\\:NdE1\\:OdE2]:McE3]:McE2^:NbE2^:<0AbE4^:LbE3_:;000AaE5_:92O000O10002N3L2O00000O01000O1000O10O1000O10O1000O10O1000O0100000O0100000O010000O0100000O0100000O01000O10O1000O10O1O1MEZE9d:J\\E6c:J]E7b:J^E5b:L^E4b:K_E5`:L_E5a:J`E5`:L`E4`:L`E4`:K`E6_:KaE4`:L`E4`:KaE5^:LaE4`:KaE5_:KaE5_:KaE5^:KcE4^:LaE5_:KaE5^:KcE5]:KcE4^:KcE5\\:LcE5]:KcE4^:KcE5\\:LdE4\\:LdE4\\:KdE5\\:LdE4\\:KeE5[:KeE5Z:LfE3[:LeE5[:KeE5[:KeE4[:LfE4Z:LfE4Z:KfE6Y:KgE4Z:LfE4Z:KgE5X:LhE4X:LgE4Z:KgE5X:LhE4X:KiE4X:LhE4W:MhE4X:KiE5W:KiE4X:KiE5V:LjE4V:LiE5W:JjE5W:KiE5V:LjE3W:LjE4V:LiE4X:LhE4W:LjE3W:MiE2X:NhE2X:MiE2W:OhE2X:MiE2Y:MgE3Z:LfE3h:0O01O100O10O0100O`W_1"}, "label": "red skiis"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red skis that the person with the yellow boots is on\"."}], "task": "refering", "answer_template": "The \"red skis that the person with the yellow boots is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 280, 281], "bbox": [0.15421875000000002, 0.6118055555555556, 0.795078125, 0.951111111111111], "iscrowd": 0, "area": 7197.335249999991, "rle": {"size": [360, 640], "counts": "W_S11V;101O001N1ROMgF4X90eFO\\91cF0\\91cFO]91cF0]90bF1]9OdF0]90bF1]9OcF1]90bF1^9NbF2^9ObF1^9NbF2^9OaF2^9NbF2^9OaF2^9NbF2^9ObF1]9OcF1]90bF1]9OcF1]9OcF2\\9OdF0\\90dF0\\91cFO^90bF0^90bF0^90bFO_91aFO_90bF0^90bF0^90bF0^90aF0`90`F0`9OaF1_9OaF1_9OaF0`90`F0_91aFO_90bF0^90bF0^90aF0`90`F0`90`F0`9OaF1^90bF0^90bFO_91aFO_91aFO_91`F0`9OaF0_91aFO_91aFO_91aFO_91aFO_90bFO_91aFO_91aFO^92aFO_91aFO_90bFO_91aFO_91aFO_91aFO^92bFN^91cFN^92bFN^92aFO_91aFO_91aFN`92`FN_92aFO_91aFO_91aFO_91`FOa91_FOa90`F0_91`F0`90`F0`90`FOa91^F0b9O_F1a9O_F1a9O^F2a9O_F0b90^F0b90]F1c9N^F1c9O]F1c9O]F1b90]F0d9O]F1c9O]F1c9O\\F2d9N\\F1e9O[F1e9N[F3d9N\\F1e9O[F1e9OZF2f9NZF1g9NZF2f9NYF3f9NZF2f9NZF1g9NYF3g9MYF3f9NZF1g9OXF2h9NXF2h9MYF2g9OXF2h9NXF2h9NXF2h9MYF2g9OXF2h9NXF2h9NXF1i9OVF2i9NXF2h9NXF1i9OVF2j9NVF2i9NXF2h9NWF2j9NVF2j9NVF2i9OVF1k9NVF2j9NVF2j9NUF2k9OUF1k9OUF1k9NUF3k9MUF2k9OUF1k9OUF1k9NUF2l9NTF2k9OUF1k9OTF1m9OSF1m9NTF2k9OTF2l9NTF1m9OSF1m9NSF3l9NTF1m9OSF1m9ORF2n9NRF1n9OSF1m9ORF2n9NRF2n9NRF1n90QF1o9NRF2n9NRF1o9OPF2o9OQF1o9NRF1o9OQF1o9OPF2o9OQF1o9OQF0P:OPF2P:NPF2o9OQF0P:0oE1Q:NPF2P:NPF1P:0oE1Q:OoE1Q:OoE0R:OnE2Q:OoE1Q:OoE1Q:OnE1S:OmE1S:NnE2Q:OnE1S:OmE1S:OmE1S:NnE1R:0mE1S:OmE1S:OmE1S:OlE1T:OmE1S:OmE1S:OlE1U:OkE1T:OmE1S:OlE1U:OkE1U:OkE1T:0kE1U:NlE1U:OkE1U:OjE2U:OkE0V:0jE0V:OjE2V:NjE1V:0jE0V:0iE1W:NjE2V:NjE1V:0jE0V:0iE1W:OiE0X:OiE1V:0iE1W:OiE1W:OiE0X:OhE2W:OiE1W:OiE1W:OhE1Y:OgE1X:OiE1W:OhE1Y:OgE1Y:OgE1X:0gE1Y:NhE1Y:OgE1Y:OfE2Y:OgE1Y:NhE1Y:OgE1Y:OfE2Y:OgE0Z:0fE0Z:OfE2Z:NfE2Y:OgE0Z:0eE1[:NfE2Z:NfE2Y:OfE1[:OeE1[:OeE1[:NeE2[:OeE1[:OeE1[:OdE2\\:NdE1\\:OeE1[:OdE2\\:NdE2\\:NdE1\\:OdE2]:McE3]:McE2^:NbE2^:<0AbE4^:LbE3_:;000AaE5_:92O000O10002N3L2O00000O01000O1000O10O1000O10O1000O10O1000O0100000O0100000O010000O0100000O0100000O01000O10O1000O10O1O1MEZE9d:J\\E6c:J]E7b:J^E5b:L^E4b:K_E5`:L_E5a:J`E5`:L`E4`:L`E4`:K`E6_:KaE4`:L`E4`:KaE5^:LaE4`:KaE5_:KaE5_:KaE5^:KcE4^:LaE5_:KaE5^:KcE5]:KcE4^:KcE5\\:LcE5]:KcE4^:KcE5\\:LdE4\\:LdE4\\:KdE5\\:LdE4\\:KeE5[:KeE5Z:LfE3[:LeE5[:KeE5[:KeE4[:LfE4Z:LfE4Z:KfE6Y:KgE4Z:LfE4Z:KgE5X:LhE4X:LgE4Z:KgE5X:LhE4X:KiE4X:LhE4W:MhE4X:KiE5W:KiE4X:KiE5V:LjE4V:LiE5W:JjE5W:KiE5V:LjE3W:LjE4V:LiE4X:LhE4W:LjE3W:MiE2X:NhE2X:MiE2W:OhE2X:MiE2Y:MgE3Z:LfE3h:0O01O100O10O0100O`W_1"}, "label": "red skis that the person with the yellow boots is on"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child wearing a purple snowsuit\"."}], "task": "refering", "answer_template": "The \"a child wearing a purple snowsuit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 56, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 169, 170, 171, 192, 193, 194, 216], "bbox": [0.332953125, 0.12583333333333332, 0.48809375, 0.7087777777777778], "iscrowd": 0, "area": 12830.994449999995, "rle": {"size": [360, 640], "counts": "iP[2e0]:8J7J5J7J5J6K6K1O2VHQNY6Q2dISNY6o1fITNW6l1jITNT6n1kISNS6o1mIQNQ6Q2oIoMo5Z1TIhNn0Nm5Y1\\IdNg02l5[1nJfNP5\\1PKdNP5[1oJgNP5Z1RJ`N@9^6Y1kIGT6<hIFX6d0ZI@f6a2O1IiK]IW4a6mK\\IT4b6:M3N2[Oe0N2N2N2N1lJcJ_4]5`KdJ_4]5_KfJ`4Z5_KgJa4Y5_KhJ`4Y5_KgJa4W5aKjJ^4S5eKmJZ4R5hKnJX4P5jKPKV4n4lKRKT4k4oKUKQ4i4QLWKo3j4jK\\KU4f4dK`K\\4f51O00000000001O000O1000001N2On0RO1O1N2O1O2N1N2M4L3K6K5UNTJRNS6\\1hJUN^5i1PIlMn8Y20O1OaGiMk7X2kGoMX8P2aGVN`8X23M2O2M2N3M5L5I6I8I4J7G8]Oe0ImTc3"}, "label": "a child wearing a purple snowsuit"}]}
{"id": 283263, "image": "COCO_train2014_000000283263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child is dressed warmly on skis\"."}], "task": "refering", "answer_template": "The \"a child is dressed warmly on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 56, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 169, 170, 171, 192, 193, 194, 216], "bbox": [0.332953125, 0.12583333333333332, 0.48809375, 0.7087777777777778], "iscrowd": 0, "area": 12830.994449999995, "rle": {"size": [360, 640], "counts": "iP[2e0]:8J7J5J7J5J6K6K1O2VHQNY6Q2dISNY6o1fITNW6l1jITNT6n1kISNS6o1mIQNQ6Q2oIoMo5Z1TIhNn0Nm5Y1\\IdNg02l5[1nJfNP5\\1PKdNP5[1oJgNP5Z1RJ`N@9^6Y1kIGT6<hIFX6d0ZI@f6a2O1IiK]IW4a6mK\\IT4b6:M3N2[Oe0N2N2N2N1lJcJ_4]5`KdJ_4]5_KfJ`4Z5_KgJa4Y5_KhJ`4Y5_KgJa4W5aKjJ^4S5eKmJZ4R5hKnJX4P5jKPKV4n4lKRKT4k4oKUKQ4i4QLWKo3j4jK\\KU4f4dK`K\\4f51O00000000001O000O1000001N2On0RO1O1N2O1O2N1N2M4L3K6K5UNTJRNS6\\1hJUN^5i1PIlMn8Y20O1OaGiMk7X2kGoMX8P2aGVN`8X23M2O2M2N3M5L5I6I8I4J7G8]Oe0ImTc3"}, "label": "a child is dressed warmly on skis"}]}
{"id": 512644, "image": "COCO_train2014_000000512644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white trike - cycle\"."}], "task": "refering", "answer_template": "The \"black and white trike - cycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 328, 329, 330, 331, 332, 333, 334, 346, 347, 348, 351, 352, 353, 354, 355, 356], "bbox": [0.009625, 0.024934497816593886, 0.7108125000000001, 0.9912882096069868], "iscrowd": 0, "area": 120573.66505000001, "rle": {"size": [458, 640], "counts": "To23V>;D<@`0E;F:E;E;E;E;E:G2M3M3N2M4L3M3N2M3M4M2M3M3N2N3M?A3M2N1O000000000000000000000000000000000001O0000000000000000lKoJoNQ5P1VKjNj4U1]KfNa4Z1fK`NZ4_1jK_NU4`1nK^NR4a1PL_No3`1TL^Nl3a1WL^Nh3a1ZL^Nf3a1]L^Nb3a1aL]N_3c1bL]N]3b1fL_NW3`1lL`NR3_1PMcNm2\\1WMcNg2\\1\\MdNb2[1bMcN]2\\1gMbNX2Z1oMdNP2U1YNjNf1V1]NhNb1X1aNfN^1Z1eNdNZ1\\1iNbNV1^1lNaNS1_1POdN@kMmMa3f2gNUOPNoMY3n2jNlNUNPNQ3V3nNbNYNRNi2]3nN^NdNlM]2h3oNYNnNgMS2R4nNTNh2m1XMPNj2Q2VMmMk2T2SMmMl2V2QMkMo2V2PMkMo2V2oLkMQ3`2cLaM]3g2[LYMd3o2SLSMm3U3jKlLV4[3aKgL_4a3WKaLi4f3oJ[LP5`60000O1O1O1O1O100O10000000000000000000000O1D<H8H8G9H8G9M3N2M3N2M3N2N2O1O1N2O1O1O1O1N2O1O10000O10000O10000O10001N3N2N2M3N2N3M2N2N2N2N3M2N2N2N2N2N3M2N1O1O001O1O001O001O1N101O1O001O1N101O001O1O001O1O001O1O001O000000000001O00000000001O000000001O2N1N3N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N001O1O001O1O001O1O001O1O001O001O00001O00001O00001O00001O0PHjLl4V3SKlLl4U3QKnLn4R3PKQMn4P3QKRMn4o2oJTMP5l2oJVMP5j2nJYMQ5g2mJ\\MR5d2mJ^MR5c2kJ`MT5`2kJbMT5^2jJeMU5[2iJhMV5X2iJjMV5W2gJlMX5T2gJnMX5R2fJQNY5o1kIjL\\OZ1h6l1jIPMYOV1l6k1hIVMUOR1R7h1hIUOW6k0gIXOV6k0gIXOW6j0gIXOV6j0iIYOS6j0kIXOS6j0kIXOR6j0mIXOP6k0nIWOP6j0oIXOn5k0PJWOm5l0QJVOm5k0RJWOk5l0SJSOm5P1QJoNP6U1lIjNT6d41001N2O2N1N3N1WOmGZKV8]4[HYKg7_4U1J7K4010O01O10O01O0010O01O11N2O1N101O1N2O1O1O1O1O001O1O1O00010O00001O00001O00010O00001O000000000001O001O001O001O1O001O001O1O001O001O1O0O2O001N2O0O2O001N2O0O2O001N2O0O4M2N2M3N3L3N2M4M2N2M3N3L3K5J7H7J6I7J>\\Og0WOP_b2"}, "label": "black and white trike - cycle"}]}
{"id": 512644, "image": "COCO_train2014_000000512644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black and white vehicle to the left of the motocycle\"."}], "task": "refering", "answer_template": "The \"the black and white vehicle to the left of the motocycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 328, 329, 330, 331, 332, 333, 334, 346, 347, 348, 351, 352, 353, 354, 355, 356], "bbox": [0.009625, 0.024934497816593886, 0.7108125000000001, 0.9912882096069868], "iscrowd": 0, "area": 120573.66505000001, "rle": {"size": [458, 640], "counts": "To23V>;D<@`0E;F:E;E;E;E;E:G2M3M3N2M4L3M3N2M3M4M2M3M3N2N3M?A3M2N1O000000000000000000000000000000000001O0000000000000000lKoJoNQ5P1VKjNj4U1]KfNa4Z1fK`NZ4_1jK_NU4`1nK^NR4a1PL_No3`1TL^Nl3a1WL^Nh3a1ZL^Nf3a1]L^Nb3a1aL]N_3c1bL]N]3b1fL_NW3`1lL`NR3_1PMcNm2\\1WMcNg2\\1\\MdNb2[1bMcN]2\\1gMbNX2Z1oMdNP2U1YNjNf1V1]NhNb1X1aNfN^1Z1eNdNZ1\\1iNbNV1^1lNaNS1_1POdN@kMmMa3f2gNUOPNoMY3n2jNlNUNPNQ3V3nNbNYNRNi2]3nN^NdNlM]2h3oNYNnNgMS2R4nNTNh2m1XMPNj2Q2VMmMk2T2SMmMl2V2QMkMo2V2PMkMo2V2oLkMQ3`2cLaM]3g2[LYMd3o2SLSMm3U3jKlLV4[3aKgL_4a3WKaLi4f3oJ[LP5`60000O1O1O1O1O100O10000000000000000000000O1D<H8H8G9H8G9M3N2M3N2M3N2N2O1O1N2O1O1O1O1N2O1O10000O10000O10000O10001N3N2N2M3N2N3M2N2N2N2N3M2N2N2N2N2N3M2N1O1O001O1O001O001O1N101O1O001O1N101O001O1O001O1O001O1O001O000000000001O00000000001O000000001O2N1N3N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N001O1O001O1O001O1O001O1O001O001O00001O00001O00001O00001O0PHjLl4V3SKlLl4U3QKnLn4R3PKQMn4P3QKRMn4o2oJTMP5l2oJVMP5j2nJYMQ5g2mJ\\MR5d2mJ^MR5c2kJ`MT5`2kJbMT5^2jJeMU5[2iJhMV5X2iJjMV5W2gJlMX5T2gJnMX5R2fJQNY5o1kIjL\\OZ1h6l1jIPMYOV1l6k1hIVMUOR1R7h1hIUOW6k0gIXOV6k0gIXOW6j0gIXOV6j0iIYOS6j0kIXOS6j0kIXOR6j0mIXOP6k0nIWOP6j0oIXOn5k0PJWOm5l0QJVOm5k0RJWOk5l0SJSOm5P1QJoNP6U1lIjNT6d41001N2O2N1N3N1WOmGZKV8]4[HYKg7_4U1J7K4010O01O10O01O0010O01O11N2O1N101O1N2O1O1O1O1O001O1O1O00010O00001O00001O00010O00001O000000000001O001O001O001O1O001O001O1O001O001O1O0O2O001N2O0O2O001N2O0O2O001N2O0O4M2N2M3N3L3N2M4M2N2M3N3L3K5J7H7J6I7J>\\Og0WOP_b2"}, "label": "the black and white vehicle to the left of the motocycle"}]}
{"id": 512644, "image": "COCO_train2014_000000512644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle with flames painted on it\"."}], "task": "refering", "answer_template": "The \"a motorcycle with flames painted on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 37, 38, 39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297], "bbox": [0.51621875, 0.011375545851528385, 0.9890156250000001, 0.8024017467248908], "iscrowd": 0, "area": 62183.25785, "rle": {"size": [458, 640], "counts": "cgc41W>;E;F9gNQOeDV1V;kNgD]1S;dNkD`1R;aNkDc1S;^NjDg1S;YNkDl1R;UNlDo1Q;RNjDU2HhMU;4oD^2P;cMkDa2U;;O001O001O001O001O001OO100O2N1O1O1O2O000010O000010O000010O0000001O00001O00O1O100O1O1O1O1O1O100001O0jMTMmHm2S7TMlHm2R7TMmHm2S7TMlHm2S7TMkHm2T7TMlHl2T7UMjHm2U7TMjHl2V7TMiHn2U7TMjHl2V7UMhHl2X7TMhHm2V7UMhHl2X7UMgHk2Y7UMfHm2Y7TMfHl2Y7VMeHl2Z7TMfHl2Z7UMeHk2Z7WMeHj2Z7VMeHl2Z7UMeHk2[7VMdHk2Z7VMeHl2Z7UMeHl2Z7UMeHl2Z7TMeHn2Z7TMbHP3\\7SM_HP3`7SM[HP3d7RMYHP3f7SMVHo2i7TMSHo2k7TMPHo2o7TMmGn2R8TMkGn2T8e1O2N1O1O1O1O1O2N1O3M4L3M1O2N1O2N101N1O2N1O2N1O2N2N1O2N1O2N1O2N101N1O2N2N2WJVHn4l7eJ`HZ5Y8O0O2N2O1N1O2N2O1N2N1N3N2N2N1O2M3N2N1O2N2O1O0O2O1O1N101O1O1N101O1N2O1O0O2O1O1N101O1O1N101O1N2O1O1O1N2O1O1O1N2O1O1O2M2O100O010O0001O00001O001O000010O01O00001O001O00002N3N1M3N3L3N2M4M2M3N3L3N2N3L3N3PJ]KR2f4kM\\KT2f4hM]KV2f4gM\\KX2g4dM[KZ2h4bMZK^2h4_MZK_2j4\\MYKc2i4ZMYKe2i4WMZKg2i4VMXKj2k4QMXKm2k4PMWKo2k4mLYKP3j4mLXKR3k4iLXKU3k4hLWKW3k4eLXKY3o4_LTK`3S5XLoJg3W5QLlJm3j7O1N2O1N101N2O1N2O1O1N101N2O1N2O1N2O1N2O1O1N1O2N2N2N2N2N3L3N2N2N3M2N2N2M5I7H8H8H8H7I8G;D<D=CUl2"}, "label": "a motorcycle with flames painted on it"}]}
{"id": 512644, "image": "COCO_train2014_000000512644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcycle with flame decals\"."}], "task": "refering", "answer_template": "The \"motorcycle with flame decals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 37, 38, 39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297], "bbox": [0.51621875, 0.011375545851528385, 0.9890156250000001, 0.8024017467248908], "iscrowd": 0, "area": 62183.25785, "rle": {"size": [458, 640], "counts": "cgc41W>;E;F9gNQOeDV1V;kNgD]1S;dNkD`1R;aNkDc1S;^NjDg1S;YNkDl1R;UNlDo1Q;RNjDU2HhMU;4oD^2P;cMkDa2U;;O001O001O001O001O001OO100O2N1O1O1O2O000010O000010O000010O0000001O00001O00O1O100O1O1O1O1O1O100001O0jMTMmHm2S7TMlHm2R7TMmHm2S7TMlHm2S7TMkHm2T7TMlHl2T7UMjHm2U7TMjHl2V7TMiHn2U7TMjHl2V7UMhHl2X7TMhHm2V7UMhHl2X7UMgHk2Y7UMfHm2Y7TMfHl2Y7VMeHl2Z7TMfHl2Z7UMeHk2Z7WMeHj2Z7VMeHl2Z7UMeHk2[7VMdHk2Z7VMeHl2Z7UMeHl2Z7UMeHl2Z7TMeHn2Z7TMbHP3\\7SM_HP3`7SM[HP3d7RMYHP3f7SMVHo2i7TMSHo2k7TMPHo2o7TMmGn2R8TMkGn2T8e1O2N1O1O1O1O1O2N1O3M4L3M1O2N1O2N101N1O2N1O2N1O2N2N1O2N1O2N1O2N101N1O2N2N2WJVHn4l7eJ`HZ5Y8O0O2N2O1N1O2N2O1N2N1N3N2N2N1O2M3N2N1O2N2O1O0O2O1O1N101O1O1N101O1N2O1O0O2O1O1N101O1O1N101O1N2O1O1O1N2O1O1O1N2O1O1O2M2O100O010O0001O00001O001O000010O01O00001O001O00002N3N1M3N3L3N2M4M2M3N3L3N2N3L3N3PJ]KR2f4kM\\KT2f4hM]KV2f4gM\\KX2g4dM[KZ2h4bMZK^2h4_MZK_2j4\\MYKc2i4ZMYKe2i4WMZKg2i4VMXKj2k4QMXKm2k4PMWKo2k4mLYKP3j4mLXKR3k4iLXKU3k4hLWKW3k4eLXKY3o4_LTK`3S5XLoJg3W5QLlJm3j7O1N2O1N101N2O1N2O1O1N101N2O1N2O1N2O1N2O1O1N1O2N2N2N2N2N3L3N2N2N3M2N2N2M5I7H8H8H8H7I8G;D<D=CUl2"}, "label": "motorcycle with flame decals"}]}
{"id": 496261, "image": "COCO_train2014_000000496261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foremost green circle\"."}], "task": "refering", "answer_template": "The \"the foremost green circle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390], "bbox": [0.6393096234309623, 0.790546875, 0.9740376569037656, 0.9887343749999999], "iscrowd": 0, "area": 16944.51069999999, "rle": {"size": [640, 478], "counts": "Rjo57gc07H8H8H8I7H8H8I7H7I4M3L3N3L3N3N100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O100000000001O000000000000000000001O000000000000000000001O00000000000000001O00000000000000001O000000000000001O0000001O001O001O001O1O001O001O001O001O1O001O001O001O001O2N2N2N2N2N3M2G9G9G9G:E:G9G9G9G9Fhn6"}, "label": "the foremost green circle"}]}
{"id": 496261, "image": "COCO_train2014_000000496261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a square doughnut\"."}], "task": "refering", "answer_template": "The \"a square doughnut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310], "bbox": [0.01202928870292887, 0.440453125, 0.5265481171548118, 0.808984375], "iscrowd": 0, "area": 44625.418800000014, "rle": {"size": [640, 478], "counts": "fR4`0Rc0c0J6J5L5J5K6K5J5K6J=D?@`0@9H4K4L5L5J6J6K5J6J3M3N2N1O2M3N2N2N1O2M3N2N2N1O1O0O2O00001O0O10001O0000001O0000001O00001O0000000000O10000000000000000000000000000000O10O1000000000O01000000O1000O10O100000000O10O1000O1000000O010000000O1000O10O1000000O10O100000O100000O01000000O1000O10O100000000O0100000O10O100O00100O00100O0010O01O10O01O10O01O010O1O010000O100O10000O10000O10000000000000000000000O010000000O100000000O10000O1O100O1O1O4M9F;E7I4M2M:F>B3N2M4L3M3N3L3M3M4M3L5K4L4L4M3L4K5K5J6K5J6K4K6K5J<E<C>Caj\\4"}, "label": "a square doughnut"}]}
{"id": 496261, "image": "COCO_train2014_000000496261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glazed donut\"."}], "task": "refering", "answer_template": "The \"glazed donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310], "bbox": [0.01202928870292887, 0.440453125, 0.5265481171548118, 0.808984375], "iscrowd": 0, "area": 44625.418800000014, "rle": {"size": [640, 478], "counts": "fR4`0Rc0c0J6J5L5J5K6K5J5K6J=D?@`0@9H4K4L5L5J6J6K5J6J3M3N2N1O2M3N2N2N1O2M3N2N2N1O1O0O2O00001O0O10001O0000001O0000001O00001O0000000000O10000000000000000000000000000000O10O1000000000O01000000O1000O10O100000000O10O1000O1000000O010000000O1000O10O1000000O10O100000O100000O01000000O1000O10O100000000O0100000O10O100O00100O00100O0010O01O10O01O10O01O010O1O010000O100O10000O10000O10000000000000000000000O010000000O100000000O10000O1O100O1O1O4M9F;E7I4M2M:F>B3N2M4L3M3N3L3M3M4M3L5K4L4L4M3L4K5K5J6K5J6K4K6K5J<E<C>Caj\\4"}, "label": "glazed donut"}]}
{"id": 135815, "image": "COCO_train2014_000000135815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow toy\"."}], "task": "refering", "answer_template": "The \"yellow toy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.274671875, 0.48052083333333334, 0.8181875, 0.9870208333333333], "iscrowd": 0, "area": 49705.05955, "rle": {"size": [480, 640], "counts": "nkb2<a>b0^Ob0^Ob0^O6J5K5K5K5K5K5K5L5J5K5K5K5H8D<C=C<I6N2M3N2N2N2N2N2N2N2N2M3N2O100O1O1O001O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O10000000000O10000000000O10000000000O^IgH_5Y7^JkHa5U7\\JoHc5Q7YJTIf5l6WJXIh5h6UJ]Ii5h6nI]IQ6j7O1O2N1O1O1O1O2N1O1N2O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1OO1@`0@`0G900000000O100000000O100000000O1000000O1000000003M4L5K4L4L4L4L4L5K4L4L4L4L4L4L3M3M4K4M3M4L1O1O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001OO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1M3N2N2M3N2M3N2N101001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O00001O00001O0000001O00001O00000000000000000oLmEU1S:kN^Fd0b9[O`Fd0`9[ObFd0^9[OdFd0\\9[OfFd0Z9\\OgFc0Y9\\OiFc0W9\\OkFc0U9\\OlFd0T9\\OmFc0S9\\OoFc0Q9\\OQGc0o8\\OSGc0m8\\OUGc0k8]OVGb0j8]OXGb0h8]OYGc0g8\\O[Gc0e8\\O]Gc0c8]O^Gb0b8]O`Gb0`8]ObGb0a8ZOaGe0a8YO`Gf0b8WO_Gi0d8SO^Gl0d8QO^Gn0[;O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O[\\f1"}, "label": "yellow toy"}]}
{"id": 135815, "image": "COCO_train2014_000000135815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow doll in the image\"."}], "task": "refering", "answer_template": "The \"yellow doll in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.274671875, 0.48052083333333334, 0.8181875, 0.9870208333333333], "iscrowd": 0, "area": 49705.05955, "rle": {"size": [480, 640], "counts": "nkb2<a>b0^Ob0^Ob0^O6J5K5K5K5K5K5K5L5J5K5K5K5H8D<C=C<I6N2M3N2N2N2N2N2N2N2N2M3N2O100O1O1O001O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O10000000000O10000000000O10000000000O^IgH_5Y7^JkHa5U7\\JoHc5Q7YJTIf5l6WJXIh5h6UJ]Ii5h6nI]IQ6j7O1O2N1O1O1O1O2N1O1N2O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1OO1@`0@`0G900000000O100000000O100000000O1000000O1000000003M4L5K4L4L4L4L4L5K4L4L4L4L4L4L3M3M4K4M3M4L1O1O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001OO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1M3N2N2M3N2M3N2N101001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O00001O00001O0000001O00001O00000000000000000oLmEU1S:kN^Fd0b9[O`Fd0`9[ObFd0^9[OdFd0\\9[OfFd0Z9\\OgFc0Y9\\OiFc0W9\\OkFc0U9\\OlFd0T9\\OmFc0S9\\OoFc0Q9\\OQGc0o8\\OSGc0m8\\OUGc0k8]OVGb0j8]OXGb0h8]OYGc0g8\\O[Gc0e8\\O]Gc0c8]O^Gb0b8]O`Gb0`8]ObGb0a8ZOaGe0a8YO`Gf0b8WO_Gi0d8SO^Gl0d8QO^Gn0[;O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O[\\f1"}, "label": "yellow doll in the image"}]}
{"id": 135815, "image": "COCO_train2014_000000135815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stuffed monkey\"."}], "task": "refering", "answer_template": "The \"stuffed monkey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 212, 213, 214, 216, 217, 218, 231, 232, 233, 235, 236, 237, 239, 240, 256, 257, 258, 259, 260, 279, 280, 281, 282, 303], "bbox": [0.061546875, 0.12552083333333333, 0.7740937499999999, 0.782375], "iscrowd": 0, "area": 51216.044050000026, "rle": {"size": [480, 640], "counts": "`ab01m>>C<C:G1N3M2O2M2N2O2M2N3N1N3M2O1N3M101N2N101O1N2O0O2O1O0O2O1N101O1N101N2O1O0O2O000O10O010000O010O10000O010O1000NkEVMY8i2hGYMU8f2lG]MR8a2PHaMn7]2THeMj7Y2XHiMf7V2ZHmMd7Q2^HRN_7l1cHVN[7h1gHZNV7f1jH]NT7`1oHbNo6[1TIgNj6V1YIlNe6R1\\IPOc6o0^ISO`6m0`IUO^6j0dIWOZ6i0fIXOX6h0jIYOT6g0lI[OR6d0PJ\\Oo5d0QJ^Om5a0TJAj5?WJBg5=ZJDe5<\\JEa5;`JG^59cJH[57fJJY56gJLW53kJNS52mJOR50PK1n4ORK3k4MWK4h4JYK7f4IZK9e4E]K<a4D_K=a4B`K>_4BaK>`4AaK>_4BaK?_4@aK`0_4@bK?_4@aKa0^4_OcK`0^4_ObKa0^4_OcKa0]4^OcKb0]4^OcKb0^4]OcKc0\\4]OdKc0]4\\OdKc0]4\\OcKe0c3YNZIQ1T3e0^3aN\\Ii0V3f0_3cNXIg0h1gN]Oo1c5fNVIe0W1\\OLY1g5hNTIc0n0J3k0k5jNSIa0c09<;o5mNoH`0:g0d0MR6<WIMd0FV6=TI3a0@[6>RI6`0]O^6>oH9a0XO`6[3^IeLc6^3ZIbLe6i3PIWLQ7k5001O001O1O1O1O100O1O1O1O1O1O1O1OO1N2O1N2O1N2N20000001O00000VOkHkIT7R6]I`Ic6]6gI]IX6`6nI]IR6a6W1M3M3O10001O00001O1O001O1O010O001N2O001O1O001O1O0O2O2N1O1O1O1N3N1O100O1O2O0O1O1O2O0O1O2O1N101O1N2O0O2O1O1N101O1O001O1OSGYKb7f4\\H\\Kd7e4XH_Kh7`4VHcKj7\\4SHhKm7W4PHmKo7S4oGPLQ8o3lGULT8l3hGWLX8j3bGZL_8g3[G^Le8o41O00001O00000O2O000000O1N2O1N101N2O1N2O1N2O1NL5K4L5K5K5K6J5N3N101N1dK\\F[3f9cLZF^3h9_LXFa3k9[LWFe3j9YLVFg3m9VLSFk3n9RLTFm3_:1O10TMkEl0U:TO]F;b9EcF7]9HhF4W9LnFOS9OSGMl83YGIg86^GDd8;aG^Ob8a0fGVO\\8j0mGkNU8T1l200O100O1O100O100O100O100O10000001O01O01O00001O001O00001O00001O00001O001O00001O00001O01O01O001O00001O000010O01O0010O01O0010O01O00100O001O010O001O010O0010O010O010O010O010O010O10O10O010O010O010O10O001O10O01O0010O01O1O0010O01O1O010O001O00100O001O00100O001O010O001O1O01O0001O01O1O001O10O01O1O00100O001O1O10O01O1N101N2O0O2N2O0O2O1N3M3L4M3M2M4M3M3L_iS2"}, "label": "stuffed monkey"}]}
{"id": 135815, "image": "COCO_train2014_000000135815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed monkey\"."}], "task": "refering", "answer_template": "The \"a stuffed monkey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 152, 153, 154, 155, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 212, 213, 214, 216, 217, 218, 231, 232, 233, 235, 236, 237, 239, 240, 256, 257, 258, 259, 260, 279, 280, 281, 282, 303], "bbox": [0.061546875, 0.12552083333333333, 0.7740937499999999, 0.782375], "iscrowd": 0, "area": 51216.044050000026, "rle": {"size": [480, 640], "counts": "`ab01m>>C<C:G1N3M2O2M2N2O2M2N3N1N3M2O1N3M101N2N101O1N2O0O2O1O0O2O1N101O1N101N2O1O0O2O000O10O010000O010O10000O010O1000NkEVMY8i2hGYMU8f2lG]MR8a2PHaMn7]2THeMj7Y2XHiMf7V2ZHmMd7Q2^HRN_7l1cHVN[7h1gHZNV7f1jH]NT7`1oHbNo6[1TIgNj6V1YIlNe6R1\\IPOc6o0^ISO`6m0`IUO^6j0dIWOZ6i0fIXOX6h0jIYOT6g0lI[OR6d0PJ\\Oo5d0QJ^Om5a0TJAj5?WJBg5=ZJDe5<\\JEa5;`JG^59cJH[57fJJY56gJLW53kJNS52mJOR50PK1n4ORK3k4MWK4h4JYK7f4IZK9e4E]K<a4D_K=a4B`K>_4BaK>`4AaK>_4BaK?_4@aK`0_4@bK?_4@aKa0^4_OcK`0^4_ObKa0^4_OcKa0]4^OcKb0]4^OcKb0^4]OcKc0\\4]OdKc0]4\\OdKc0]4\\OcKe0c3YNZIQ1T3e0^3aN\\Ii0V3f0_3cNXIg0h1gN]Oo1c5fNVIe0W1\\OLY1g5hNTIc0n0J3k0k5jNSIa0c09<;o5mNoH`0:g0d0MR6<WIMd0FV6=TI3a0@[6>RI6`0]O^6>oH9a0XO`6[3^IeLc6^3ZIbLe6i3PIWLQ7k5001O001O1O1O1O100O1O1O1O1O1O1O1OO1N2O1N2O1N2N20000001O00000VOkHkIT7R6]I`Ic6]6gI]IX6`6nI]IR6a6W1M3M3O10001O00001O1O001O1O010O001N2O001O1O001O1O0O2O2N1O1O1O1N3N1O100O1O2O0O1O1O2O0O1O2O1N101O1N2O0O2O1O1N101O1O001O1OSGYKb7f4\\H\\Kd7e4XH_Kh7`4VHcKj7\\4SHhKm7W4PHmKo7S4oGPLQ8o3lGULT8l3hGWLX8j3bGZL_8g3[G^Le8o41O00001O00000O2O000000O1N2O1N101N2O1N2O1N2O1NL5K4L5K5K5K6J5N3N101N1dK\\F[3f9cLZF^3h9_LXFa3k9[LWFe3j9YLVFg3m9VLSFk3n9RLTFm3_:1O10TMkEl0U:TO]F;b9EcF7]9HhF4W9LnFOS9OSGMl83YGIg86^GDd8;aG^Ob8a0fGVO\\8j0mGkNU8T1l200O100O1O100O100O100O100O10000001O01O01O00001O001O00001O00001O00001O001O00001O00001O01O01O001O00001O000010O01O0010O01O0010O01O00100O001O010O001O010O0010O010O010O010O010O010O10O10O010O010O010O10O001O10O01O0010O01O1O0010O01O1O010O001O00100O001O00100O001O010O001O1O01O0001O01O1O001O10O01O1O00100O001O1O10O01O1N101N2O0O2N2O0O2O1N3M3L4M3M2M4M3M3L_iS2"}, "label": "a stuffed monkey"}]}
{"id": 389772, "image": "COCO_train2014_000000389772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white shorts and shirt holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man in white shorts and shirt holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 56, 57, 79, 80, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 242, 261, 262, 264, 284, 285, 287, 288, 307, 308, 310, 311, 330, 331, 333, 352, 353, 354, 356, 357, 375, 376, 379, 380], "bbox": [0.31060937499999997, 0.11227083333333333, 0.544359375, 0.9737291666666666], "iscrowd": 0, "area": 27833.6805, "rle": {"size": [480, 640], "counts": "Vgm26h>9H1O1O1O100O1O100O1O100O1O100O1O100O1O100O1^HXOk0i0QO]Om0c0POBm0?oNFo0;mNKQ15kN1R10kN5R1LjN9T1HhN;X1EeN?Z1AdNb0[1^ObNe0hN\\N_No0f2i0eNgN[N`0n2l0aNSOXNOU3Q1^N_OSN@\\3U1[NKjMTOl3Q1TN:aMjN[4m0nMh0YM_Nj4i0hMb2Y2_MbMe2^2\\M]Mi2c2WMXMn2g2SMTMR3k2oLPMU3Q3kLjLZ3U3gLfL^3Z3bLbLb3]3_L^Le3b3\\L\\Lf3d3ZLYLi3f3XLWLk3i3ULcKbMM[6`4TLYKnM2P6e4mLTKV3l4_3000UMRKcJ0VOQ5S6XK\\JO]Ok4R6aKTJMFd4R6XMjIj2R6\\MiIg2S6bMfI`2V6X3K6K4L4L4L4L4KoNdJfHX5[7mJbHP5_7TK^Hj4d7ZKYHc4h7bKTH\\4m7iKoGU4n6VK`Ii0^Oo3Q7_K\\If0Ah3R7iKXId0Ba3V7QLTIb0B\\3X7YLQI`0CU3[7bLmHl4S7[KiHb4X7a1002O101O1O1O100O001002N3M2N3M3N1N3M2O2M3M3N2M3M3N:EV1jN3eLVF[1m9[NZFd1j9RN\\Fn1g9jM_Fa23PMm7O\\HS3CQMb9Q3YFQMh9Q3TFQMk9Q3QFRMo9o3101N3N2M3N00O010O1O1YNfKfH[41YLj4ASKW4IRMEjN]1KB7BR4CRO9ZM:GHm3CDZO^MS1VONi3Cf16fJ5e3Cl1OdJ=`3BZ2A^Jh0Y3ET46mKGZ44fKK`40aKNb4O`KOc4N`KOc4NaKMc41`KKc42X5Mk]X4"}, "label": "a man in white shorts and shirt holding a tennis racket"}]}
{"id": 389772, "image": "COCO_train2014_000000389772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with white color dress\"."}], "task": "refering", "answer_template": "The \"a man with white color dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 56, 57, 79, 80, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 242, 261, 262, 264, 284, 285, 287, 288, 307, 308, 310, 311, 330, 331, 333, 352, 353, 354, 356, 357, 375, 376, 379, 380], "bbox": [0.31060937499999997, 0.11227083333333333, 0.544359375, 0.9737291666666666], "iscrowd": 0, "area": 27833.6805, "rle": {"size": [480, 640], "counts": "Vgm26h>9H1O1O1O100O1O100O1O100O1O100O1O100O1O100O1^HXOk0i0QO]Om0c0POBm0?oNFo0;mNKQ15kN1R10kN5R1LjN9T1HhN;X1EeN?Z1AdNb0[1^ObNe0hN\\N_No0f2i0eNgN[N`0n2l0aNSOXNOU3Q1^N_OSN@\\3U1[NKjMTOl3Q1TN:aMjN[4m0nMh0YM_Nj4i0hMb2Y2_MbMe2^2\\M]Mi2c2WMXMn2g2SMTMR3k2oLPMU3Q3kLjLZ3U3gLfL^3Z3bLbLb3]3_L^Le3b3\\L\\Lf3d3ZLYLi3f3XLWLk3i3ULcKbMM[6`4TLYKnM2P6e4mLTKV3l4_3000UMRKcJ0VOQ5S6XK\\JO]Ok4R6aKTJMFd4R6XMjIj2R6\\MiIg2S6bMfI`2V6X3K6K4L4L4L4L4KoNdJfHX5[7mJbHP5_7TK^Hj4d7ZKYHc4h7bKTH\\4m7iKoGU4n6VK`Ii0^Oo3Q7_K\\If0Ah3R7iKXId0Ba3V7QLTIb0B\\3X7YLQI`0CU3[7bLmHl4S7[KiHb4X7a1002O101O1O1O100O001002N3M2N3M3N1N3M2O2M3M3N2M3M3N:EV1jN3eLVF[1m9[NZFd1j9RN\\Fn1g9jM_Fa23PMm7O\\HS3CQMb9Q3YFQMh9Q3TFQMk9Q3QFRMo9o3101N3N2M3N00O010O1O1YNfKfH[41YLj4ASKW4IRMEjN]1KB7BR4CRO9ZM:GHm3CDZO^MS1VONi3Cf16fJ5e3Cl1OdJ=`3BZ2A^Jh0Y3ET46mKGZ44fKK`40aKNb4O`KOc4N`KOc4NaKMc41`KKc42X5Mk]X4"}, "label": "a man with white color dress"}]}
{"id": 479886, "image": "COCO_train2014_000000479886.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the darkest giraffe shown\"."}], "task": "refering", "answer_template": "The \"the darkest giraffe shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 60, 82, 83, 105, 106, 107, 129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 266, 267, 269, 270, 271, 272, 289, 290, 292, 293, 294, 295, 312, 315, 316, 317, 318, 334, 335, 338, 339, 341, 362, 364], "bbox": [0.493984375, 0.0424375, 0.8690312499999999, 0.9456249999999999], "iscrowd": 0, "area": 31532.56080000002, "rle": {"size": [480, 640], "counts": "UVd45j>2N1O2N1L4O001O0O2O1O001O1N2O1O001O1N2N2M3O1O10000O100O10000O10O0100O001O001O1O0iLhNgHY1X7iNfHX1Y7jNeHV1[7mNaHV1]7lNaHV1]7lNaHW1\\7lNeHR1Y7oNiHP1U7POoHn0o6SOTIl0h6UO[Ij0b6WO_Ik0\\6XOeIi0W6XOjIj0P6ZOQJh0g5\\OZJf0`5^OaJc0Z5_OgJc0U5^OmJd0n4TOiGTOZ3j1h4TObKn0Z4SOgKP1S4ROoKo0m3POVLR1e3lN`LZOoK]1]7VOiLZOoKa1R7TOSMXOPLe1Z6A]Na0]1^OhNc0S1\\OROe0i0[O[Oe0a0ZODg07YOMg0OXO6i0EWO?i0\\OWOi0j0ROUOS1k0iNUO[1l0`NSOe1m0WNROn1o0mMQOW2o0eMPO_2R1\\MmNh2T1RMmNR3U1fLnN]3S1\\LoNh3S1PLQOR4o0iKSOZ4n0_KUOd4k0nInMTOZ1Q7i0fIYNkNQ1b7f0`IdNaNh0S8c0YIb0i6_OSIa0P7_OmHa0V7@eHa0^73kGMX8[33N2M3M3N2N3L3N3L3N3L4M2N3L3N3L3N3M2O1N3M2N2O1N3M2N2O101O0001O0O10001O3M2N3M3M3M3M3L;F;E:F8G:G:lHZIY5n6XJXIi5h6QJ]Io5d6kIaIV6_6\\InIe6V7001O]IkG]6S8bInG`6P8_IQHd6l7[IVHg6g7XIZHj6d7UI^Hm6_7RIbHP7f7101N1O101N1O101N1K6K3O2O1N2N2N2O1L3M4L4L4L4L4L3N3L3M3M4L3M3M3M4L3M3M3M4L3M3M4K4E;D=H8L4L3M4L4L4L4001O2M2O1O1O1O1O1^EWMj8k2bFgM^9Z2PFVNQ:]31O0O101N101O[GcKf6\\4]HaLd7^51O0O4M2N3YKkG[2Y8WMXHb2l7PMeHCROR2^8nMmHK[Ob1l7UNVI5@U1k9eN]FU1k;K6K5KcnV1"}, "label": "the darkest giraffe shown"}]}
{"id": 479886, "image": "COCO_train2014_000000479886.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mother zebra watching the calves\"."}], "task": "refering", "answer_template": "The \"mother zebra watching the calves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 60, 82, 83, 105, 106, 107, 129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 266, 267, 269, 270, 271, 272, 289, 290, 292, 293, 294, 295, 312, 315, 316, 317, 318, 334, 335, 338, 339, 341, 362, 364], "bbox": [0.493984375, 0.0424375, 0.8690312499999999, 0.9456249999999999], "iscrowd": 0, "area": 31532.56080000002, "rle": {"size": [480, 640], "counts": "UVd45j>2N1O2N1L4O001O0O2O1O001O1N2O1O001O1N2N2M3O1O10000O100O10000O10O0100O001O001O1O0iLhNgHY1X7iNfHX1Y7jNeHV1[7mNaHV1]7lNaHV1]7lNaHW1\\7lNeHR1Y7oNiHP1U7POoHn0o6SOTIl0h6UO[Ij0b6WO_Ik0\\6XOeIi0W6XOjIj0P6ZOQJh0g5\\OZJf0`5^OaJc0Z5_OgJc0U5^OmJd0n4TOiGTOZ3j1h4TObKn0Z4SOgKP1S4ROoKo0m3POVLR1e3lN`LZOoK]1]7VOiLZOoKa1R7TOSMXOPLe1Z6A]Na0]1^OhNc0S1\\OROe0i0[O[Oe0a0ZODg07YOMg0OXO6i0EWO?i0\\OWOi0j0ROUOS1k0iNUO[1l0`NSOe1m0WNROn1o0mMQOW2o0eMPO_2R1\\MmNh2T1RMmNR3U1fLnN]3S1\\LoNh3S1PLQOR4o0iKSOZ4n0_KUOd4k0nInMTOZ1Q7i0fIYNkNQ1b7f0`IdNaNh0S8c0YIb0i6_OSIa0P7_OmHa0V7@eHa0^73kGMX8[33N2M3M3N2N3L3N3L3N3L4M2N3L3N3L3N3M2O1N3M2N2O1N3M2N2O101O0001O0O10001O3M2N3M3M3M3M3L;F;E:F8G:G:lHZIY5n6XJXIi5h6QJ]Io5d6kIaIV6_6\\InIe6V7001O]IkG]6S8bInG`6P8_IQHd6l7[IVHg6g7XIZHj6d7UI^Hm6_7RIbHP7f7101N1O101N1O101N1K6K3O2O1N2N2N2O1L3M4L4L4L4L4L3N3L3M3M4L3M3M3M4L3M3M3M4L3M3M4K4E;D=H8L4L3M4L4L4L4001O2M2O1O1O1O1O1^EWMj8k2bFgM^9Z2PFVNQ:]31O0O101N101O[GcKf6\\4]HaLd7^51O0O4M2N3YKkG[2Y8WMXHb2l7PMeHCROR2^8nMmHK[Ob1l7UNVI5@U1k9eN]FU1k;K6K5KcnV1"}, "label": "mother zebra watching the calves"}]}
{"id": 479886, "image": "COCO_train2014_000000479886.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bigger giraffe facing the camera\"."}], "task": "refering", "answer_template": "The \"the bigger giraffe facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 102, 125, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 215, 217, 218, 238, 240, 241, 261, 263], "bbox": [0.35271875, 0.14506249999999998, 0.5011875, 0.7069166666666666], "iscrowd": 0, "area": 10666.796100000001, "rle": {"size": [480, 640], "counts": "lWZ33i>5H7J7I6mNT1M3L3eNRNbEQ2[:YNZEW1\\OhNW;<SE>;RO_:j0kDF[1[Oi9l1^FoM`9S2e10O101O006J_OiCfNm;Y1XDiNb;V1bDgN^;X1fDbN\\;_1R1O100O010O100001O01O1O1O1O1000000000000001O001O0O6I:F9Gm0TO7H8H9cNkK_G_4EiK\\70lH[5Q7hJlHZ5R7hJkH[5S7hJjHZ5T7hJjH[5Z6]J_I;4Z5\\6^J\\I;5Y5^6_JYI:7Y5`6_JUI;8Y5b6\\K\\If4d6k11O010O00O1O2N1N2O1N21PJ\\Ic3e6ULiIc3Y6ULUJc3m5ULaJj2YOgKW6U1nJi2Q6mL]Ji2i5nLeJg2`5RMkJg2Y5VMmJd2V5[MnJ5^MR1f7hN[LP1h3nN_Lj0c3VObLd0`3ZOgL?[3@iL;Z3BlL7W3HnL2T3Mcde4"}, "label": "the bigger giraffe facing the camera"}]}
{"id": 479886, "image": "COCO_train2014_000000479886.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shortest giraffe\"."}], "task": "refering", "answer_template": "The \"the shortest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 140, 141, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 214, 233, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 305], "bbox": [0.0915625, 0.32466666666666666, 0.309734375, 0.8181875000000001], "iscrowd": 0, "area": 11186.55815, "rle": {"size": [480, 640], "counts": "hjk06h>6J6J6K0O2N1O2N2N101N1O2N1O2N1101N2O1N2O1N2002N1O2N1O2N1O2O0O2N1O2N1O2N1O2N1O2N2N1O2N1O2N8H8H8H8H8H8oF`LQ6h3dIZL]6n3XITLi6R4nHoKT7U4eHmK\\7T4`HmKb7T4ZHnKg7S4SHQLn7P4jGULX8l3`GZLa8g3XGZLm8g3kF[LZ9a45L4K5K5K5L4^OgEWL_:c3gEXL]:c3hE^LW:]3oEbLQ:Z3SFgLk9U3[FkLd9P3bFPM\\9l2iFUMV9g2oFYMo8c2VG]Mj8^2\\GaMc8[2bGcM`8Y2eGdM\\8X2iGfMY8X2X201O0O101O00cEkMP8T2mGoMS8Q2jGRNW8m1fGVNZ8j1dGXN]8g1`G\\N`8d1]G_Nd8`1YGcNg8]1VGcNm8]1PGcNT9[1jFeNY9[1dFeN`9\\1\\FcNg9_1TFbNo9_1lEaNW:a1dE_N`:a1\\E_Ng:c1TE]NP;d1kD\\NX;f1cDZN`;g1\\DYNh;\\23M4L3M4L3M4L4L4LUEPNm8S2iFTNW9U2YFQNi9W2gEoMZ:Y2WEnMi:T3100L4H8G9H8G9H8G9H8G9G9H8G9H9F9HTn^6"}, "label": "the shortest giraffe"}]}
{"id": 479886, "image": "COCO_train2014_000000479886.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra walking away from his herd of three other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra walking away from his herd of three other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 140, 141, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 214, 233, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 305], "bbox": [0.0915625, 0.32466666666666666, 0.309734375, 0.8181875000000001], "iscrowd": 0, "area": 11186.55815, "rle": {"size": [480, 640], "counts": "hjk06h>6J6J6K0O2N1O2N2N101N1O2N1O2N1101N2O1N2O1N2002N1O2N1O2N1O2O0O2N1O2N1O2N1O2N1O2N2N1O2N1O2N8H8H8H8H8H8oF`LQ6h3dIZL]6n3XITLi6R4nHoKT7U4eHmK\\7T4`HmKb7T4ZHnKg7S4SHQLn7P4jGULX8l3`GZLa8g3XGZLm8g3kF[LZ9a45L4K5K5K5L4^OgEWL_:c3gEXL]:c3hE^LW:]3oEbLQ:Z3SFgLk9U3[FkLd9P3bFPM\\9l2iFUMV9g2oFYMo8c2VG]Mj8^2\\GaMc8[2bGcM`8Y2eGdM\\8X2iGfMY8X2X201O0O101O00cEkMP8T2mGoMS8Q2jGRNW8m1fGVNZ8j1dGXN]8g1`G\\N`8d1]G_Nd8`1YGcNg8]1VGcNm8]1PGcNT9[1jFeNY9[1dFeN`9\\1\\FcNg9_1TFbNo9_1lEaNW:a1dE_N`:a1\\E_Ng:c1TE]NP;d1kD\\NX;f1cDZN`;g1\\DYNh;\\23M4L3M4L3M4L4L4LUEPNm8S2iFTNW9U2YFQNi9W2gEoMZ:Y2WEnMi:T3100L4H8G9H8G9H8G9H8G9G9H8G9H9F9HTn^6"}, "label": "a zebra walking away from his herd of three other zebras"}]}
{"id": 479886, "image": "COCO_train2014_000000479886.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking at the foresty area with rocks to the right of them\"."}], "task": "refering", "answer_template": "The \"a giraffe looking at the foresty area with rocks to the right of them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 100, 101, 122, 123, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 213, 214, 237, 260], "bbox": [0.248328125, 0.19589583333333332, 0.44000000000000006, 0.7008333333333333], "iscrowd": 0, "area": 7461.635650000005, "rle": {"size": [480, 640], "counts": "ngZ22i>6K5J5L5N2N101N1O2O1N1O2O0O2N101N1O2O1N1O2O0O2N101N2N101N102M3M4M2Lb1^NN3M2O1N3M2PObD^N`;d1mDlMU;T2j01O2O0O1O2N10001O001hDkNd8Y1fFHk8X3N1O1O1O1N2ORO_G_K`8j3bHlK]7k3P2L3M3M3M4L3_NaD@b;6mDBV;=PEZOT;f0QEQOS;o0REhNS;W1RE`NR;`1V10001O00\\OaNiC_1Y<cNdC\\1^<fN^CZ1e<hNVCX1l<jNQCV1P=;2N2N3M2N3N1N3M2N3M2O2M2N2N2O100O10001OO100O100O1O1005K1N103M2N2N3M2N2N3M2M3NTfW5"}, "label": "a giraffe looking at the foresty area with rocks to the right of them"}]}
{"id": 78482, "image": "COCO_train2014_000000078482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"leftmost man\"."}], "task": "refering", "answer_template": "The \"leftmost man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 30, 31, 32, 33, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 238, 239, 259], "bbox": [0.205390625, 0.004485776805251641, 0.47496875000000005, 0.7123632385120351], "iscrowd": 0, "area": 37787.111500000006, "rle": {"size": [457, 640], "counts": "Udj1o0n<=B>K5L4L4L4L4L4L4L4L4L4L4K5M2N3N2M3N2M3N3L3N3M3M3M3M2O2M2N2O2M2O2N1Ng1ZN3M4K4M3L4M3M4K4M3MO1O001O1O1O1O0010000O10O10O10000O010000O1O100O1O100O1O1O100O1O100O1O100O1M3O1N2N2N2O1J5K6I7I7L4N2N1O2N2N2M3F9100000000O2O000000000O100000000O2O000000000O100000000O2O000O1O1O1O1M3M3N2M3N2M4L3N2M3M3]JUJf2m5XMVJe2l5XMWJf2k5XMXJe2j5YMYJd2i5YM[Jd2g5ZM\\Jc2g5ZM[Je2f5XM^Jf2c5XM`Je2b5YMaJc2b5ZMbJb2a5\\MaJa2b5]MaJ_2b5_MaJ]2g5[M]Ja2l5WMWJe2o5VMTJf2R6TMSJg2S6TMQJg2U6QMRJj2W6jLQJP3h8K7I8G:Fa0_O9G:G8G_ie4"}, "label": "leftmost man"}]}
{"id": 78482, "image": "COCO_train2014_000000078482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man standing beside a woman\"."}], "task": "refering", "answer_template": "The \"an older man standing beside a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 30, 31, 32, 33, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 238, 239, 259], "bbox": [0.205390625, 0.004485776805251641, 0.47496875000000005, 0.7123632385120351], "iscrowd": 0, "area": 37787.111500000006, "rle": {"size": [457, 640], "counts": "Udj1o0n<=B>K5L4L4L4L4L4L4L4L4L4L4K5M2N3N2M3N2M3N3L3N3M3M3M3M2O2M2N2O2M2O2N1Ng1ZN3M4K4M3L4M3M4K4M3MO1O001O1O1O1O0010000O10O10O10000O010000O1O100O1O100O1O1O100O1O100O1O100O1M3O1N2N2N2O1J5K6I7I7L4N2N1O2N2N2M3F9100000000O2O000000000O100000000O2O000000000O100000000O2O000O1O1O1O1M3M3N2M3N2M4L3N2M3M3]JUJf2m5XMVJe2l5XMWJf2k5XMXJe2j5YMYJd2i5YM[Jd2g5ZM\\Jc2g5ZM[Je2f5XM^Jf2c5XM`Je2b5YMaJc2b5ZMbJb2a5\\MaJa2b5]MaJ_2b5_MaJ]2g5[M]Ja2l5WMWJe2o5VMTJf2R6TMSJg2S6TMQJg2U6QMRJj2W6jLQJP3h8K7I8G:Fa0_O9G:G8G_ie4"}, "label": "an older man standing beside a woman"}]}
{"id": 78482, "image": "COCO_train2014_000000078482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman\"."}], "task": "refering", "answer_template": "The \"the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 103, 104, 105, 125, 126, 127, 148, 149, 171], "bbox": [0.46854687500000003, 0.05617067833698031, 0.6161875, 0.4606783369803064], "iscrowd": 0, "area": 7686.997950000004, "rle": {"size": [457, 640], "counts": "iRV45m=6I7J6I7iNX1O0001O00000000010O00000000001O01O0001O1O1O001N2O1O1O0OhMZNQHe1o7fNgGY1Z8PO]Go0c8\\OSGb0n8EkF;V9JeF5[90aFO`96[FIe9=UFBl9c0PF\\OQ:i0jEVOV:^21N101O00001N10001O0O1000000O101O1O1O1N2O2N1O1N2O2N1O1N2O1O2N1O1N2O1O2N1N2SNkDe0V;YOXE:i:D]E7e:FbE3`:KfE0[:NkEMV:1PFJR:3UFGl97T2N2N2N3MWh]3"}, "label": "the woman"}]}
{"id": 78482, "image": "COCO_train2014_000000078482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling woman in a black jacket\"."}], "task": "refering", "answer_template": "The \"a smiling woman in a black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 103, 104, 105, 125, 126, 127, 148, 149, 171], "bbox": [0.46854687500000003, 0.05617067833698031, 0.6161875, 0.4606783369803064], "iscrowd": 0, "area": 7686.997950000004, "rle": {"size": [457, 640], "counts": "iRV45m=6I7J6I7iNX1O0001O00000000010O00000000001O01O0001O1O1O001N2O1O1O0OhMZNQHe1o7fNgGY1Z8PO]Go0c8\\OSGb0n8EkF;V9JeF5[90aFO`96[FIe9=UFBl9c0PF\\OQ:i0jEVOV:^21N101O00001N10001O0O1000000O101O1O1O1N2O2N1O1N2O2N1O1N2O1O2N1O1N2O1O2N1N2SNkDe0V;YOXE:i:D]E7e:FbE3`:KfE0[:NkEMV:1PFJR:3UFGl97T2N2N2N3MWh]3"}, "label": "a smiling woman in a black jacket"}]}
{"id": 78482, "image": "COCO_train2014_000000078482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in navy attire pulling something off a large cake\"."}], "task": "refering", "answer_template": "The \"a man in navy attire pulling something off a large cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 89, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 366, 367], "bbox": [0.410546875, 0.027067833698030633, 0.9744531249999999, 1.0], "iscrowd": 0, "area": 86326.22025000003, "rle": {"size": [457, 640], "counts": "^ee36Q>7H8I5K2M4M3M2M4M3L3N3M3L3N3M3L3N3M2M3N2M3N2N3L3N2N2M3N3L3N2N2M3N2M2O1O1N2O1O1N2O3L3N3N1O2N2N1O2N2N2N2N2N2N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O001O001O1O001O001O001O001O001O001O002N2N2N2N2M3N1O1O001O1O001O1N101O1O002N1O1O1N2O2N1O1O1O1O2N1jN`IbJa6i4UJUKl5h4YJUKh5h4\\JVKe5g4`JVKb5e4cJYK^5d4gJYKZ5d4jJZKW5c4mJ\\KS5`4SK]Kn4`4VK^Kk4_4`201O0010O02N1O2N1O101N1O2N1O2N101N1O1O2N101N1O2O00001O0000O1O010O1O1O1O1O1O1O11N1000001O0000001O000O110O01O00010O001O0010O0001O001N1O2O0O1O2N101O101O0O2O000O2O001N10001O0O2O0O2H7O7J0O01O00001O001O001N1000000000000000000000000O1000000001O1O1O1O001O1O100O0100YKUGo3k8mKXGT4g8jK[GW4e8eK^G[4c8aK`G`4`8]KbGd4Q91O1O001O1O11N2O1O1O1O1N2O1O1O1O2M2O1O1O1O1N2O00O1O1N101O1O1O1O0O2O1O1O1O0O2O1O1O10O2O01O10O01O1O010O1O010O1O00100O00100O1O010O1O00100O005L7H8H8I7H8H8H8I7H8H8G9G:G8G9G9G9G9H8G9G9G9H8G9G9G:F9HU1jN>B>]Ocf6"}, "label": "a man in navy attire pulling something off a large cake"}]}
{"id": 78482, "image": "COCO_train2014_000000078482.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in sailor uniform cutting large cake and woman in uniform looking over his shoulder\"."}], "task": "refering", "answer_template": "The \"man in sailor uniform cutting large cake and woman in uniform looking over his shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 89, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 366, 367], "bbox": [0.410546875, 0.027067833698030633, 0.9744531249999999, 1.0], "iscrowd": 0, "area": 86326.22025000003, "rle": {"size": [457, 640], "counts": "^ee36Q>7H8I5K2M4M3M2M4M3L3N3M3L3N3M3L3N3M2M3N2M3N2N3L3N2N2M3N3L3N2N2M3N2M2O1O1N2O1O1N2O3L3N3N1O2N2N1O2N2N2N2N2N2N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O001O001O1O001O001O001O001O001O001O002N2N2N2N2M3N1O1O001O1O001O1N101O1O002N1O1O1N2O2N1O1O1O1O2N1jN`IbJa6i4UJUKl5h4YJUKh5h4\\JVKe5g4`JVKb5e4cJYK^5d4gJYKZ5d4jJZKW5c4mJ\\KS5`4SK]Kn4`4VK^Kk4_4`201O0010O02N1O2N1O101N1O2N1O2N101N1O1O2N101N1O2O00001O0000O1O010O1O1O1O1O1O1O11N1000001O0000001O000O110O01O00010O001O0010O0001O001N1O2O0O1O2N101O101O0O2O000O2O001N10001O0O2O0O2H7O7J0O01O00001O001O001N1000000000000000000000000O1000000001O1O1O1O001O1O100O0100YKUGo3k8mKXGT4g8jK[GW4e8eK^G[4c8aK`G`4`8]KbGd4Q91O1O001O1O11N2O1O1O1O1N2O1O1O1O2M2O1O1O1O1N2O00O1O1N101O1O1O1O0O2O1O1O1O0O2O1O1O10O2O01O10O01O1O010O1O010O1O00100O00100O1O010O1O00100O005L7H8H8I7H8H8H8I7H8H8G9G:G8G9G9G9G9H8G9G9G9H8G9G9G:F9HU1jN>B>]Ocf6"}, "label": "man in sailor uniform cutting large cake and woman in uniform looking over his shoulder"}]}
{"id": 471698, "image": "COCO_train2014_000000471698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table behind the people talking\"."}], "task": "refering", "answer_template": "The \"the table behind the people talking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 239, 240, 241, 263, 264, 286, 287, 309], "bbox": [0.43128124999999995, 0.3500249376558604, 0.587234375, 0.984563591022444], "iscrowd": 0, "area": 11816.858050000003, "rle": {"size": [401, 640], "counts": "j[\\3a0i7DWLP1X3XO\\LS1]3VOVLU1b3TORLW1g3ROlKY1m3oNgK\\1R4mNaK]1X4lN\\K_1\\4kNWK_1b4jNRK\\1k4nNiJS1X5R31O1N101O1O1O1N2O001O1O1O1N2O1O001O1L3MJ63M3WNRJmLQ6Q3UJhLn5U3YJdLj5Z3]J^Lg5_3_JZLd5e3aJSLc5k3dJmK_5R4`1O2M3N3M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2M3N2M3N2O1O1O1N3N001O0O1000001N100bNlE?T:\\OTFa0l9[O[Fb0f9ZO`Fc0`9\\ObFc0^9\\OeFb0[9]OhF`0Y9_OiF`0W9_OlF?U9_OnF?R9@PG?P9@SG>m8BUG<k8CWG;k8C[G8e8HaG2_8MhGMX83S2O2O0OooW3"}, "label": "the table behind the people talking"}]}
{"id": 471698, "image": "COCO_train2014_000000471698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table and the cutleries to which a boy is sitting\"."}], "task": "refering", "answer_template": "The \"the table and the cutleries to which a boy is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 239, 240, 241, 263, 264, 286, 287, 309], "bbox": [0.43128124999999995, 0.3500249376558604, 0.587234375, 0.984563591022444], "iscrowd": 0, "area": 11816.858050000003, "rle": {"size": [401, 640], "counts": "j[\\3a0i7DWLP1X3XO\\LS1]3VOVLU1b3TORLW1g3ROlKY1m3oNgK\\1R4mNaK]1X4lN\\K_1\\4kNWK_1b4jNRK\\1k4nNiJS1X5R31O1N101O1O1O1N2O001O1O1O1N2O1O001O1L3MJ63M3WNRJmLQ6Q3UJhLn5U3YJdLj5Z3]J^Lg5_3_JZLd5e3aJSLc5k3dJmK_5R4`1O2M3N3M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2M3N2M3N2O1O1O1N3N001O0O1000001N100bNlE?T:\\OTFa0l9[O[Fb0f9ZO`Fc0`9\\ObFc0^9\\OeFb0[9]OhF`0Y9_OiF`0W9_OlF?U9_OnF?R9@PG?P9@SG>m8BUG<k8CWG;k8C[G8e8HaG2_8MhGMX83S2O2O0OooW3"}, "label": "the table and the cutleries to which a boy is sitting"}]}
{"id": 471698, "image": "COCO_train2014_000000471698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table and wine glasses at the right\"."}], "task": "refering", "answer_template": "The \"the table and wine glasses at the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.49984375, 0.5370822942643392, 1.0, 0.9707980049875311], "iscrowd": 0, "area": 37448.6318, "rle": {"size": [401, 640], "counts": "cRn31_<1O100O1O1O1O1O1O100O1O1O1O1O100O2N1O1O1O100@^OPEc0o:^OPEc0o:_OoDb0Q;^OnDc0Q;_OmDb0R;_OmDb0R;@mD`0R;AmD`0R;BlD?T;AkD`0T;?O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1lNjMlGW2Q8RNiGn1U8[NcGf1Z8dN^G]1_8lNZGU1d8]1M3N2M3N20000O1000000O10000O1000000O1000000O1000000O10000001O001O1O001O001O001O001O001O1O001O001SMbGe1^8TNjGk1`8dMgG\\2W901O0000001OO1N2M3N2N2M3N2M3N2J6K5J6C=F:K5K5L4N2O1N2N2O1N2N2001O0000000000000000000000001O0000000000000000001O0000001O00000PLnHa2S7oL_In2a6lLgIR3Y6hLPJU3P6eLXJY3c7M2N5Le0ZO1O1O2N1O1O1O1O`NhMhHW2o6WNmHh1k6eNPI[1l6mNPIS1l6ROSIn0h6YOVIg0f6^OYIb0c6C\\I=d6D[I<e6EZI;e6GZI9f6HTI=l6EmH`0S7AgHd0Y7]OaHh0_7U200O10000000000001O00000000001O000000001O00000000001O2N2N2N1O2N2N2iKSHf3\\801O00000000000000000000000000O1N2O1N2O1N2O1N2M3N2M3O100O10000000000O10000000000O10000000000O10000001O00001O00001O00001O00001O001O000iMeJ"}, "label": "the table and wine glasses at the right"}]}
{"id": 471698, "image": "COCO_train2014_000000471698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tabletop with wine glasses on the far right\"."}], "task": "refering", "answer_template": "The \"tabletop with wine glasses on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.49984375, 0.5370822942643392, 1.0, 0.9707980049875311], "iscrowd": 0, "area": 37448.6318, "rle": {"size": [401, 640], "counts": "cRn31_<1O100O1O1O1O1O1O100O1O1O1O1O100O2N1O1O1O100@^OPEc0o:^OPEc0o:_OoDb0Q;^OnDc0Q;_OmDb0R;_OmDb0R;@mD`0R;AmD`0R;BlD?T;AkD`0T;?O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1lNjMlGW2Q8RNiGn1U8[NcGf1Z8dN^G]1_8lNZGU1d8]1M3N2M3N20000O1000000O10000O1000000O1000000O1000000O10000001O001O1O001O001O001O001O001O1O001O001SMbGe1^8TNjGk1`8dMgG\\2W901O0000001OO1N2M3N2N2M3N2M3N2J6K5J6C=F:K5K5L4N2O1N2N2O1N2N2001O0000000000000000000000001O0000000000000000001O0000001O00000PLnHa2S7oL_In2a6lLgIR3Y6hLPJU3P6eLXJY3c7M2N5Le0ZO1O1O2N1O1O1O1O`NhMhHW2o6WNmHh1k6eNPI[1l6mNPIS1l6ROSIn0h6YOVIg0f6^OYIb0c6C\\I=d6D[I<e6EZI;e6GZI9f6HTI=l6EmH`0S7AgHd0Y7]OaHh0_7U200O10000000000001O00000000001O000000001O00000000001O2N2N2N1O2N2N2iKSHf3\\801O00000000000000000000000000O1N2O1N2O1N2O1N2M3N2M3O100O10000000000O10000000000O10000000000O10000001O00001O00001O00001O00001O001O000iMeJ"}, "label": "tabletop with wine glasses on the far right"}]}
{"id": 463507, "image": "COCO_train2014_000000463507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with short hair playing wii\"."}], "task": "refering", "answer_template": "The \"the man with short hair playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 347, 349, 350, 351], "bbox": [0.09553125, 0.06417977528089887, 0.466671875, 0.9898651685393258], "iscrowd": 0, "area": 49471.38655000001, "rle": {"size": [445, 640], "counts": "`mj03d=:G7F]OTCc0m<_OPCa0P=51O1000O2O000O3N1O1O1N3N1O1N10L4N11O1O1O1O1O100O2M2N2O1N2N2O1N6K:E1O100OO1K6M2O2M2O110001O0O1mL[NPJf1]2_NcM:KW1j1[OnMM1i0HVNCb19K7>H[NH]14266JbNJY1067OLgNLW1K;6INkN1R1G`05C0PO3P1Fa01@1UO7m0cNTN`0^2?^O3XO:[1RO3;[O5]O>V1PO76XO7Ab0R1PO81VO:Ed0o0oN;LSO<Ii0j0mN>HPO?Mk0g0lNa0CmNa01o0d0kNj23gLR1`0iNi23jLS1?hNg22oLT1<hNd22TMU1;fNa22XMX1S6gNPJX1P6eNTJZ1k5dNYJ\\1T9O1O100O1O1O100O2N100O1O100O1O100000000000O1O2N1O1O1O1O1O1O2N1O1O1O1O1O100O2N1O2N1O2N2O0O2N2N1O2N101N2N1O2N2N101N1O2N2N1O2N2O0O2N1O2N2N1RGPMo6Q3mHRMR7P3jHSMU7V3`HnL^7[3WHhLh7a3mGbLR8g3cG]L\\8b4O1O2N1O101N1O1O2N100O2N2N2N2O2M2N2N3M2O1N2N3M2N2G:E9G:J6K4L5K4L5K5K4L5J6K4L5K4L5K5K4L5H8I6J7I6J7H8I6J7I7M2O2N101N2N1O2N2N1O2O0O2N2N1O2N2N2O1N1O2N2N2N2N101N2N2N1N3M3M3M5J6KZ^d4"}, "label": "the man with short hair playing wii"}]}
{"id": 463507, "image": "COCO_train2014_000000463507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black and white striped polo holding a wii remote\"."}], "task": "refering", "answer_template": "The \"a man in a black and white striped polo holding a wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 347, 349, 350, 351], "bbox": [0.09553125, 0.06417977528089887, 0.466671875, 0.9898651685393258], "iscrowd": 0, "area": 49471.38655000001, "rle": {"size": [445, 640], "counts": "`mj03d=:G7F]OTCc0m<_OPCa0P=51O1000O2O000O3N1O1O1N3N1O1N10L4N11O1O1O1O1O100O2M2N2O1N2N2O1N6K:E1O100OO1K6M2O2M2O110001O0O1mL[NPJf1]2_NcM:KW1j1[OnMM1i0HVNCb19K7>H[NH]14266JbNJY1067OLgNLW1K;6INkN1R1G`05C0PO3P1Fa01@1UO7m0cNTN`0^2?^O3XO:[1RO3;[O5]O>V1PO76XO7Ab0R1PO81VO:Ed0o0oN;LSO<Ii0j0mN>HPO?Mk0g0lNa0CmNa01o0d0kNj23gLR1`0iNi23jLS1?hNg22oLT1<hNd22TMU1;fNa22XMX1S6gNPJX1P6eNTJZ1k5dNYJ\\1T9O1O100O1O1O100O2N100O1O100O1O100000000000O1O2N1O1O1O1O1O1O2N1O1O1O1O1O100O2N1O2N1O2N2O0O2N2N1O2N101N2N1O2N2N101N1O2N2N1O2N2O0O2N1O2N2N1RGPMo6Q3mHRMR7P3jHSMU7V3`HnL^7[3WHhLh7a3mGbLR8g3cG]L\\8b4O1O2N1O101N1O1O2N100O2N2N2N2O2M2N2N3M2O1N2N3M2N2G:E9G:J6K4L5K4L5K5K4L5J6K4L5K4L5K5K4L5H8I6J7I6J7H8I6J7I7M2O2N101N2N1O2N2N1O2O0O2N2N1O2N2N2O1N1O2N2N2N2N101N2N2N1N3M3M3M5J6KZ^d4"}, "label": "a man in a black and white striped polo holding a wii remote"}]}
{"id": 463507, "image": "COCO_train2014_000000463507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a shirt in that says bo ( red )\"."}], "task": "refering", "answer_template": "The \"a guy in a shirt in that says bo ( red )\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.2546875, 0.0696629213483146, 0.765625, 0.9842696629213483], "iscrowd": 0, "area": 76563.0, "rle": {"size": [445, 640], "counts": "ZXW21k=1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1N2N2K5K5K5K5J6I7I7G9E;I7L4M3L4L4M3L4L4M3J6H8H8H8I7H8L4O1O100O1O100O1O100O1O100O1O100O1O1nKnJYOR5>[K]Of4=dK^O]4;nKAR49WLCj39[LEf3;\\LCd3<_LAb3>aL@_3?cL_O^3`0eL]O\\3b0gL\\OY3c0iL[OX3d0kLYOV3g0lLWOT3h0nLVOS3i0oLUOR3j0PMUOP3j0SMSOn2l0TMROm2m0VMPOk2P1VMoNj2Q1XMlNi2T1XMjNi2V1o400000000000000001O00000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O0000000000000000000001O02N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O5K7I7I8H7I;E?A:F4L5K4L5K4L4L5K4L4L5K4L4L5K2N1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O1O1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2M3N2N1O2N2N1O2N2N2N1O2N2N1O2N3M4L5K4L5K4L4L5Kd0[OS1mNS1nNS1lNohP2"}, "label": "a guy in a shirt in that says bo ( red )"}]}
{"id": 463507, "image": "COCO_train2014_000000463507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with beard\"."}], "task": "refering", "answer_template": "The \"a man with beard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362], "bbox": [0.2546875, 0.0696629213483146, 0.765625, 0.9842696629213483], "iscrowd": 0, "area": 76563.0, "rle": {"size": [445, 640], "counts": "ZXW21k=1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1N2N2K5K5K5K5J6I7I7G9E;I7L4M3L4L4M3L4L4M3J6H8H8H8I7H8L4O1O100O1O100O1O100O1O100O1O100O1O1nKnJYOR5>[K]Of4=dK^O]4;nKAR49WLCj39[LEf3;\\LCd3<_LAb3>aL@_3?cL_O^3`0eL]O\\3b0gL\\OY3c0iL[OX3d0kLYOV3g0lLWOT3h0nLVOS3i0oLUOR3j0PMUOP3j0SMSOn2l0TMROm2m0VMPOk2P1VMoNj2Q1XMlNi2T1XMjNi2V1o400000000000000001O00000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O0000000000000000000001O02N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O5K7I7I8H7I;E?A:F4L5K4L5K4L4L5K4L4L5K4L4L5K2N1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O1O1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2M3N2N1O2N2N1O2N2N2N1O2N2N1O2N3M4L5K4L5K4L4L5Kd0[OS1mNS1nNS1lNohP2"}, "label": "a man with beard"}]}
{"id": 291478, "image": "COCO_train2014_000000291478.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bigger animal\"."}], "task": "refering", "answer_template": "The \"the bigger animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291], "bbox": [0.41237500000000005, 0.5395, 0.688078125, 0.7305833333333333], "iscrowd": 0, "area": 7962.094399999999, "rle": {"size": [480, 640], "counts": "eQl32l>5L2M3N2M3N1N100O2O1N2O1N2O1N2O001O1O2O0O1O1O100O1O1O2O0N3I6I8N12O1N2000OaNjBT1V=gNnB[1Z=2O7H1O100O00010O0000010OYNkBc1Y=O2O00001O000000000000O10O0100O100O100000001O0001O000000001O000001O0001O0001O0001O01O01O010O010O0000O101O00000O10001O00000001O00001O00001O001O00001O01O01N10001O00001O000O101O00001O00001O001O10O01O001O001O1O0O2O1O001O1N101O1O0O2O1N1O2N2O1N2N101N2N3Mk]m2"}, "label": "the bigger animal"}]}
{"id": 176790, "image": "COCO_train2014_000000176790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couple in formal attire\"."}], "task": "refering", "answer_template": "The \"couple in formal attire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 59, 60, 61, 76, 77, 78, 93, 94, 95, 108, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 131, 132, 137, 138, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 159, 160, 161, 162, 163, 164, 165, 166, 171, 172, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 296, 297, 298, 299, 300, 301, 302, 313, 314, 315, 316, 317, 318, 319, 330, 331, 332, 333, 334, 335, 336, 347, 348, 349, 350, 351, 352, 364, 365, 366, 367, 368, 369, 382, 383, 384, 385, 386], "bbox": [0.07789583333333333, 0.084703125, 0.8179791666666667, 0.985828125], "iscrowd": 0, "area": 88727.04204999997, "rle": {"size": [640, 480], "counts": "U\\g0;dc05K3M3M1O1O2N1J6M3N2N2O1N2N2M3N2M300O100O100O100O100000ZOo]O]OPb0f0P^OXOPb0h0Q^OVOPb0j0U^OQOka0o0c00O1000O0100OlNZ]OP1Qc0J4L`[a1YOTe^N7H8K4L2N2O0O2O1N2O1N101O0O2N1O2N1O2N1O2N1O2N2N1O2N2N1N3N2F9ImM]^OX2ba04O1O1O2N2N2N1O2N3M2N2N2N2N3MS2mM6J6J5VCTKU:S5TEdKf:R7I6J8H>B8H5K5K9Gk1UN4L3M4L4L4L4L5K`0[MhBbM_=V2TCXMo<c2TC[Mo<b2SC]MP=_2RC`MP=\\2SCcMm<S2^ClMb<Q2iCgMW<W2lChMT<U2PDjMP<Q2VDnMj;m1\\DRNd;n1]DQNd;m1^DRNb;m1aDQN`;n1aDQN_;n1dDPN];n1fDPNZ;P2gDoMZ;o1hDPNX;o1jDPNU;Q2lDnMT;Q2nDnMQ;R2QEmMo:S2^30100O100O100O10000O100000000000000000O01000O100O2O000O100L5gNY1O1N2O1N2O1N2O1N2002NZ1fN3M00O2O00001N2N2N2N3i@\\Ml<h2iB`MV=c2fB_MY=g2`B\\M^=k2ZBWMe=j2YBXMf=j2WBXMh=k2TBWMk=o2nASMQ>R3hAQMW>]4O1O2N1O1O1O1O2N1O1O2N1O1O1O001O1O1O001O1N101O1O001O001O001N101O001O001O001O001O001N10001O001O001N1O1\\Oe0\\Oc0H9K4M3]Oc0@a0M2O1O1N2O1O1N3N1N4M2N2G:H:I7Jh0jNZDdHV<h6P1B>B^1bN<D;E;PMP3TOYQf1"}, "label": "couple in formal attire"}]}
{"id": 176790, "image": "COCO_train2014_000000176790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man wearing a suit and has a nice smile\"."}], "task": "refering", "answer_template": "The \"a young man wearing a suit and has a nice smile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 59, 60, 61, 76, 77, 78, 93, 94, 95, 108, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 131, 132, 137, 138, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 159, 160, 161, 162, 163, 164, 165, 166, 171, 172, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 296, 297, 298, 299, 300, 301, 302, 313, 314, 315, 316, 317, 318, 319, 330, 331, 332, 333, 334, 335, 336, 347, 348, 349, 350, 351, 352, 364, 365, 366, 367, 368, 369, 382, 383, 384, 385, 386], "bbox": [0.07789583333333333, 0.084703125, 0.8179791666666667, 0.985828125], "iscrowd": 0, "area": 88727.04204999997, "rle": {"size": [640, 480], "counts": "U\\g0;dc05K3M3M1O1O2N1J6M3N2N2O1N2N2M3N2M300O100O100O100O100000ZOo]O]OPb0f0P^OXOPb0h0Q^OVOPb0j0U^OQOka0o0c00O1000O0100OlNZ]OP1Qc0J4L`[a1YOTe^N7H8K4L2N2O0O2O1N2O1N101O0O2N1O2N1O2N1O2N1O2N2N1O2N2N1N3N2F9ImM]^OX2ba04O1O1O2N2N2N1O2N3M2N2N2N2N3MS2mM6J6J5VCTKU:S5TEdKf:R7I6J8H>B8H5K5K9Gk1UN4L3M4L4L4L4L5K`0[MhBbM_=V2TCXMo<c2TC[Mo<b2SC]MP=_2RC`MP=\\2SCcMm<S2^ClMb<Q2iCgMW<W2lChMT<U2PDjMP<Q2VDnMj;m1\\DRNd;n1]DQNd;m1^DRNb;m1aDQN`;n1aDQN_;n1dDPN];n1fDPNZ;P2gDoMZ;o1hDPNX;o1jDPNU;Q2lDnMT;Q2nDnMQ;R2QEmMo:S2^30100O100O100O10000O100000000000000000O01000O100O2O000O100L5gNY1O1N2O1N2O1N2O1N2002NZ1fN3M00O2O00001N2N2N2N3i@\\Ml<h2iB`MV=c2fB_MY=g2`B\\M^=k2ZBWMe=j2YBXMf=j2WBXMh=k2TBWMk=o2nASMQ>R3hAQMW>]4O1O2N1O1O1O1O2N1O1O2N1O1O1O001O1O1O001O1N101O1O001O001O001N101O001O001O001O001O001N10001O001O001N1O1\\Oe0\\Oc0H9K4M3]Oc0@a0M2O1O1N2O1O1N3N1N4M2N2G:H:I7Jh0jNZDdHV<h6P1B>B^1bN<D;E;PMP3TOYQf1"}, "label": "a young man wearing a suit and has a nice smile"}]}
{"id": 176790, "image": "COCO_train2014_000000176790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue evening dress\"."}], "task": "refering", "answer_template": "The \"a woman in a blue evening dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 88, 89, 90, 91, 104, 105, 106, 107, 108, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 279, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 341, 342, 343, 344, 345, 346, 347, 358, 359, 360, 361, 362, 363, 364, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.05991666666666667, 0.17078125, 0.4524375, 0.9887656249999999], "iscrowd": 0, "area": 76541.82104999998, "rle": {"size": [640, 480], "counts": "ncb03kc05K4K5L4L5K4fHTOYKR1a4SOZGMo0U1b7SOiFo0o02S8TOXFQ2o0QOd8ROgEU3o0nMT9ROXEW4n0kLf9k4eEZKU:]7K5L3ZGQGV6S9mHgGR7d:O1O1O1O1O100O1O1O1O1O2N100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1K5J6J6J6K5J6M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M300O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O11O001O00001O00001O00001O0000001O0000001O00001O0000001O00001O0000003iBeLc9]3YFfLf9]3UFeLk9]3QFfLn9]3lEfLT:]3gEfLX:\\3dEgL[:\\3`EfL`:\\3\\EgLc:\\3XEgLg:\\3TEfLl:^3nDeLQ;^3jDdLV;`3dDcL[;`3`DcL_;a3[DaLe;b3VDaLi;c3QD_LP<i5O1O001O1O001O001O1O001O2N3M3M4L3M3M3M3M3M3M3M3M3M3kLcFaKa9Y4_FiKd9Q4]FQLg>iNYAY1m>^NTAd1Q?TNRAl1S?mMeA]1a>ZNXBP1m=iNjBb0[=VO^C4h<CTPT5"}, "label": "a woman in a blue evening dress"}]}
{"id": 176790, "image": "COCO_train2014_000000176790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman in a blue prom dress with matching eyeshadow\"."}], "task": "refering", "answer_template": "The \"a young woman in a blue prom dress with matching eyeshadow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 88, 89, 90, 91, 104, 105, 106, 107, 108, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 279, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 324, 325, 326, 327, 328, 329, 330, 341, 342, 343, 344, 345, 346, 347, 358, 359, 360, 361, 362, 363, 364, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.05991666666666667, 0.17078125, 0.4524375, 0.9887656249999999], "iscrowd": 0, "area": 76541.82104999998, "rle": {"size": [640, 480], "counts": "ncb03kc05K4K5L4L5K4fHTOYKR1a4SOZGMo0U1b7SOiFo0o02S8TOXFQ2o0QOd8ROgEU3o0nMT9ROXEW4n0kLf9k4eEZKU:]7K5L3ZGQGV6S9mHgGR7d:O1O1O1O1O100O1O1O1O1O2N100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1K5J6J6J6K5J6M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M300O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O11O001O00001O00001O00001O0000001O0000001O00001O0000001O00001O0000003iBeLc9]3YFfLf9]3UFeLk9]3QFfLn9]3lEfLT:]3gEfLX:\\3dEgL[:\\3`EfL`:\\3\\EgLc:\\3XEgLg:\\3TEfLl:^3nDeLQ;^3jDdLV;`3dDcL[;`3`DcL_;a3[DaLe;b3VDaLi;c3QD_LP<i5O1O001O1O001O001O1O001O2N3M3M4L3M3M3M3M3M3M3M3M3M3kLcFaKa9Y4_FiKd9Q4]FQLg>iNYAY1m>^NTAd1Q?TNRAl1S?mMeA]1a>ZNXBP1m=iNjBb0[=VO^C4h<CTPT5"}, "label": "a young woman in a blue prom dress with matching eyeshadow"}]}
{"id": 561818, "image": "COCO_train2014_000000561818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe laying down to the right of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe laying down to the right of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 109, 110, 131, 132, 133, 134, 153, 154, 155, 157, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 269, 289, 290, 291, 292, 293], "bbox": [0.518140625, 0.2556, 0.86209375, 0.8542588235294117], "iscrowd": 0, "area": 16223.775899999993, "rle": {"size": [425, 640], "counts": "lQZ41W=1O1O2N1O2O1N1O2N2N2N1O2N2O0O2N2N1O2N2N101N2N2N2N2N101N2N2N2N101N2N2N2N1O2O1N3M4L3M3N3L3M4ZF^M^8e2[G`Me8b2TGcMk8a2lFeMS9^2fFgMY9V3O0O1O1O100O2N1O100O1O2N100O1O1O101N1O1O100O2N1O100O1O1O2O0O1O1O101N1O1O1OkMfG]OY8c0iG\\OU8d0oG[Oo7d0TH[Ok7e0XHYOg7f0\\HYOc7g0_HXO`7g0dHXOZ7h0hHWOW7h0mHVOR7j0PIUOo6j0UITOj6l0XISOg6l0\\ITOb6l0aIRO^6m0eIROZ6m0jIoM`Nb0e7^1nIhMdNj0]7]1`JcN_5\\1dJcN[5\\1gJdNX5[1jJdNV5[1lJeNS5Z1PKeNo4Z1SKfNl4Y1VKgNi4X1YKhNf4W1]KhNb4W1`KiN_4V1cKjN\\4U1gKjNX4U1jKjNV4U1lKkNS4T1oKlNP4S1SLlNl3S1VLmNi3R1YLnNf3Q1\\LoNc3P1`LoN_3P1cLPO[3P1gLPOX3o0kLPOS3P1oLoNQ3P1QMPOn2o0TMQOj2o0YMPOf2o0\\MQOb2o0`MQO_2n0cMRO\\2m0gMROW2o0jMQOU2n0mMROQ2o0PNPOP2o0SNPOk1Q1VNoNi1P1YNPOf1P1\\NoNb1Q1`NoN_1Q1bNoN\\1R1eNnNZ1Q1iNnNU1S1lNmNS1R1oNnNQ1Q1POnNP1Q1SOnNl0R1UOnNj0Q1XOoNh0P1ZOoNe0P1]OPO4^1MbN2^1ObN1\\12cNN]12cNN]12bNO^11bNO^12\\N3d1i50001O2O0O2N2N1O2N1O2N2N1O2N1O1O2O0O2N1O1O2N1O1O1O001O1O10O0100O10O100000O10O01O1O1N101O1O1O1N101O1O1N101O1O1O0Od[T1"}, "label": "a giraffe laying down to the right of another giraffe"}]}
{"id": 561818, "image": "COCO_train2014_000000561818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe sitting behind another giraffe facing to the right\"."}], "task": "refering", "answer_template": "The \"giraffe sitting behind another giraffe facing to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 109, 110, 131, 132, 133, 134, 153, 154, 155, 157, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 269, 289, 290, 291, 292, 293], "bbox": [0.518140625, 0.2556, 0.86209375, 0.8542588235294117], "iscrowd": 0, "area": 16223.775899999993, "rle": {"size": [425, 640], "counts": "lQZ41W=1O1O2N1O2O1N1O2N2N2N1O2N2O0O2N2N1O2N2N101N2N2N2N2N101N2N2N2N101N2N2N2N1O2O1N3M4L3M3N3L3M4ZF^M^8e2[G`Me8b2TGcMk8a2lFeMS9^2fFgMY9V3O0O1O1O100O2N1O100O1O2N100O1O1O101N1O1O100O2N1O100O1O1O2O0O1O1O101N1O1O1OkMfG]OY8c0iG\\OU8d0oG[Oo7d0TH[Ok7e0XHYOg7f0\\HYOc7g0_HXO`7g0dHXOZ7h0hHWOW7h0mHVOR7j0PIUOo6j0UITOj6l0XISOg6l0\\ITOb6l0aIRO^6m0eIROZ6m0jIoM`Nb0e7^1nIhMdNj0]7]1`JcN_5\\1dJcN[5\\1gJdNX5[1jJdNV5[1lJeNS5Z1PKeNo4Z1SKfNl4Y1VKgNi4X1YKhNf4W1]KhNb4W1`KiN_4V1cKjN\\4U1gKjNX4U1jKjNV4U1lKkNS4T1oKlNP4S1SLlNl3S1VLmNi3R1YLnNf3Q1\\LoNc3P1`LoN_3P1cLPO[3P1gLPOX3o0kLPOS3P1oLoNQ3P1QMPOn2o0TMQOj2o0YMPOf2o0\\MQOb2o0`MQO_2n0cMRO\\2m0gMROW2o0jMQOU2n0mMROQ2o0PNPOP2o0SNPOk1Q1VNoNi1P1YNPOf1P1\\NoNb1Q1`NoN_1Q1bNoN\\1R1eNnNZ1Q1iNnNU1S1lNmNS1R1oNnNQ1Q1POnNP1Q1SOnNl0R1UOnNj0Q1XOoNh0P1ZOoNe0P1]OPO4^1MbN2^1ObN1\\12cNN]12cNN]12bNO^11bNO^12\\N3d1i50001O2O0O2N2N1O2N1O2N2N1O2N1O1O2O0O2N1O1O2N1O1O1O001O1O10O0100O10O100000O10O01O1O1N101O1O1O1N101O1O1N101O1O1O0Od[T1"}, "label": "giraffe sitting behind another giraffe facing to the right"}]}
{"id": 561818, "image": "COCO_train2014_000000561818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of giraffes\"."}], "task": "refering", "answer_template": "The \"a pair of giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 94, 95, 96, 97, 118, 119, 120, 121, 143, 144, 166, 167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289], "bbox": [0.11937500000000001, 0.1820235294117647, 0.5998906249999999, 0.8337176470588237], "iscrowd": 0, "area": 26145.19974999999, "rle": {"size": [425, 640], "counts": "nRP14S=4L3M4L4L3M2O1O1O100O1O10O01O001O0O1O2N1O2N2N2O0O2N2O1O1N2O001L4L4N200O001O100O1O10O4O1O1O1O1O1O1OVJ\\NOb1LbN4_1HdN8]1EdN<]1BbN?`1^OaNc0`1[O_Nf0c1WO^Nj0c1TO\\Nm0f1QOZNP1g1nNXNS1j1kNVNU1l1iNTNX1m1eNSN\\1o1bNQN^1P2aNPN`1Q2^NoMb1S2\\NmMe1T2YNlMg1V2VNjMl1W2RNiMn1Y2PNgMQ2Z2mMfMT2[2jMeMV2\\2iMeMW2\\2fMeMZ25iIf1k3UN]24jIg1g3VN^22mIi1c3UNa21mIk1`3TNd2OnIn1\\3TNf2MoIP2Y3SNi2KPJS2U3RNl2JQJS2R3SNn2IQJU2o2SNP3FSJX2k2RNS3ESJZ2h2QNV3CTJ]2d2PNo3P2oKPNR4Q2lKoMU4R2iKnMX4R2gKnMZ4S2dKmM^4S2`KmMa4T2]KlMd4U2ZKkMh4U2VKkMk4V2SKiMP5V2oJjMR5W2\\J`MZN9\\7V2UJgM]N3_7W2mImMbNLc7V2iIQNbNIf7W2eISNcNFi7W2bIVNcNCm7W2^IeNc6[1[IfNg6Z1VIgNk6Z1RIgNP7X1nHhNT7Y1iHhNX7X1gHhN[7X1bHiN_7W1^HkNd7U1XHmNi7S1THoNn7Q1nGQOS8o0jGSOW8n0eGTO]8k0^GYOc8g20O1O1O1O1O001O1O1O10O01O1O1O1O1O001O1O100O001O1O1O1O1O010O1O1O10O01O100O1O10O01O100O1O001O1O1O1O1N101O1O1O1N101O1O1O0O2O001O1O0O2O001O1O0O2O1O001O001O1O10O01O1O10O01O1O001O1O1N101O1O1O010O1O10O01O1O010O1O10O01O1O01O00001O00001N10001O00001O000000000O10000001O000000001O000O10001O00001HUEPNk:Q2UEnMl:Q2UEnMk:S2TEmMm:R2TEmMm:S2SElMm:T25O010O10O010O010O01O1O001O001O001O1O001O001O2N2N2N2N2N3M2N2N3M2N2N3M2M3I8I6I7JjmY3"}, "label": "a pair of giraffes"}]}
{"id": 561818, "image": "COCO_train2014_000000561818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe laying to west\"."}], "task": "refering", "answer_template": "The \"giraffe laying to west\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 94, 95, 96, 97, 118, 119, 120, 121, 143, 144, 166, 167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289], "bbox": [0.11937500000000001, 0.1820235294117647, 0.5998906249999999, 0.8337176470588237], "iscrowd": 0, "area": 26145.19974999999, "rle": {"size": [425, 640], "counts": "nRP14S=4L3M4L4L3M2O1O1O100O1O10O01O001O0O1O2N1O2N2N2O0O2N2O1O1N2O001L4L4N200O001O100O1O10O4O1O1O1O1O1O1OVJ\\NOb1LbN4_1HdN8]1EdN<]1BbN?`1^OaNc0`1[O_Nf0c1WO^Nj0c1TO\\Nm0f1QOZNP1g1nNXNS1j1kNVNU1l1iNTNX1m1eNSN\\1o1bNQN^1P2aNPN`1Q2^NoMb1S2\\NmMe1T2YNlMg1V2VNjMl1W2RNiMn1Y2PNgMQ2Z2mMfMT2[2jMeMV2\\2iMeMW2\\2fMeMZ25iIf1k3UN]24jIg1g3VN^22mIi1c3UNa21mIk1`3TNd2OnIn1\\3TNf2MoIP2Y3SNi2KPJS2U3RNl2JQJS2R3SNn2IQJU2o2SNP3FSJX2k2RNS3ESJZ2h2QNV3CTJ]2d2PNo3P2oKPNR4Q2lKoMU4R2iKnMX4R2gKnMZ4S2dKmM^4S2`KmMa4T2]KlMd4U2ZKkMh4U2VKkMk4V2SKiMP5V2oJjMR5W2\\J`MZN9\\7V2UJgM]N3_7W2mImMbNLc7V2iIQNbNIf7W2eISNcNFi7W2bIVNcNCm7W2^IeNc6[1[IfNg6Z1VIgNk6Z1RIgNP7X1nHhNT7Y1iHhNX7X1gHhN[7X1bHiN_7W1^HkNd7U1XHmNi7S1THoNn7Q1nGQOS8o0jGSOW8n0eGTO]8k0^GYOc8g20O1O1O1O1O001O1O1O10O01O1O1O1O1O001O1O100O001O1O1O1O1O010O1O1O10O01O100O1O10O01O100O1O001O1O1O1O1N101O1O1O1N101O1O1O0O2O001O1O0O2O001O1O0O2O1O001O001O1O10O01O1O10O01O1O001O1O1N101O1O1O010O1O10O01O1O010O1O10O01O1O01O00001O00001N10001O00001O000000000O10000001O000000001O000O10001O00001HUEPNk:Q2UEnMl:Q2UEnMk:S2TEmMm:R2TEmMm:S2SElMm:T25O010O10O010O010O01O1O001O001O001O1O001O001O2N2N2N2N2N3M2N2N3M2N2N3M2M3I8I6I7JjmY3"}, "label": "giraffe laying to west"}]}
{"id": 209563, "image": "COCO_train2014_000000209563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"angry - looking woman\"."}], "task": "refering", "answer_template": "The \"angry - looking woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 297, 298, 321], "bbox": [0.725703125, 0.32268749999999996, 1.0, 0.7889791666666667], "iscrowd": 0, "area": 18660.525749999993, "rle": {"size": [480, 640], "counts": "jXj62m>2N3N1N2O2N1O1N2O000O2O0000001O000000001O0000000O10000001O0010O01O1O1O2O0N2O2N1O1O1N2N3N]O3;6A?J7J>B4L4L30100O1O1O1O1O1O1O1O1O1N2O1kNkM^EV2k9UN\\EGh0T2h9ZN]EEi0o1i9aN[ECj0m1g9ROWFP1f9ZORFf0m9_OnEc0o9BmE`0R:BkE`0S:BkE?T:CiE?W:U2O100O2O0O10000O2O000O2O0O10001N100O100O101N1O100O2O0O1O2O0O2N101N1O101O0O2O001O1O1UGYJ[8i5bGXJ_8h5^GZJb8g5XG_Jg8j1YGm1g8QN[Go1f8oM[GR2f8kM[GU2f8jM[GU2g8hMZGY2g8dMZG\\2h8YMVGgN2P4j8UMaGl2`8PMbGP3^8fLlGZ3V8bLlG_3Z8WLjGh3X8TLjGl3U8SLmGn3S8oKPHP4P8mKSHS4]91O00001O1O101N2N2N2N1O1O100O000001O01O01O001O1O00010O001M2N3M2O2M2N3L4LbH"}, "label": "angry - looking woman"}]}
{"id": 209563, "image": "COCO_train2014_000000209563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a napkin looking away from the camera\"."}], "task": "refering", "answer_template": "The \"a man holding a napkin looking away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 117, 118, 119, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 300, 301], "bbox": [0.006734374999999999, 0.28089583333333334, 0.296625, 0.8247291666666666], "iscrowd": 0, "area": 27011.991949999992, "rle": {"size": [480, 640], "counts": "lS2Z2d<2N3M3M2N3M3M2O2M3M3M3M3M3M3M3O1N2O1O1N2O1O2M5L8H7H9H8H8G8IO010O10O0100O010O10O0100O010O10O10OdNjJoHW5P7kJoHT5R7mJmHS5R7SKhHm4Y7YKaHg4^7\\K`Hc4_7aK^H`4^7fK`HY4\\7oKaHQ4Z7VLcHj3Y7]LeHc3V7dLhH[3Y7gLdHZ3\\7T2010O100O10O0100O10O010000O100O100O100O100O100O100O100O100O101N2O0O2O1N2O0O2O1O1N101N2O1N101N2TKcGS3^8kLeGR3\\8lLhGR3Y8kLjGS3X8kLjGT3W8iLkGV3V8hLmGW3T8gLmGS3Z8jLiGP3a8iLaGQ3k8hLVGR3V9fLmFV3]9aLeF_3_:O0O2kNbDfN^;Y1cDgN^;X1bDiN];V1dDTOS;j0nD@h:?YEBf:=[ECf:;[EDf:<ZEDg::ZEEg::ZEFg:8ZEGg:8ZEHg:6ZEIg:7YEIh:5YEJh:5YEKh:3YEMg:2ZEMg:L`E4a:DfE;[:CgE=Z:BfE=[:DdE<]:CcE<^:DbE<_:D`E;a:E_E;b:D^E;e:C[E=g:BXE>j:@VE?m:_OSEa0o:]ORE>T;AkD;[;CeD:`;D`D;c;D\\D<f;BZD>h;@XDa0i;]OWDc0k;\\OTDd0n;ZORDf0P<XOPDh0R<VOnCk0S<TOlCl0V<ROjCn0X<POhCn0V=N2N2N2N2N3N1N2N2N3M2N2N2N2NjSc6"}, "label": "a man holding a napkin looking away from the camera"}]}
{"id": 209563, "image": "COCO_train2014_000000209563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 117, 118, 119, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 300, 301], "bbox": [0.006734374999999999, 0.28089583333333334, 0.296625, 0.8247291666666666], "iscrowd": 0, "area": 27011.991949999992, "rle": {"size": [480, 640], "counts": "lS2Z2d<2N3M3M2N3M3M2O2M3M3M3M3M3M3M3O1N2O1O1N2O1O2M5L8H7H9H8H8G8IO010O10O0100O010O10O0100O010O10O10OdNjJoHW5P7kJoHT5R7mJmHS5R7SKhHm4Y7YKaHg4^7\\K`Hc4_7aK^H`4^7fK`HY4\\7oKaHQ4Z7VLcHj3Y7]LeHc3V7dLhH[3Y7gLdHZ3\\7T2010O100O10O0100O10O010000O100O100O100O100O100O100O100O100O101N2O0O2O1N2O0O2O1O1N101N2O1N101N2TKcGS3^8kLeGR3\\8lLhGR3Y8kLjGS3X8kLjGT3W8iLkGV3V8hLmGW3T8gLmGS3Z8jLiGP3a8iLaGQ3k8hLVGR3V9fLmFV3]9aLeF_3_:O0O2kNbDfN^;Y1cDgN^;X1bDiN];V1dDTOS;j0nD@h:?YEBf:=[ECf:;[EDf:<ZEDg::ZEEg::ZEFg:8ZEGg:8ZEHg:6ZEIg:7YEIh:5YEJh:5YEKh:3YEMg:2ZEMg:L`E4a:DfE;[:CgE=Z:BfE=[:DdE<]:CcE<^:DbE<_:D`E;a:E_E;b:D^E;e:C[E=g:BXE>j:@VE?m:_OSEa0o:]ORE>T;AkD;[;CeD:`;D`D;c;D\\D<f;BZD>h;@XDa0i;]OWDc0k;\\OTDd0n;ZORDf0P<XOPDh0R<VOnCk0S<TOlCl0V<ROjCn0X<POhCn0V=N2N2N2N2N3N1N2N2N3M2N2N2N2NjSc6"}, "label": "a man wearing glasses"}]}
{"id": 209563, "image": "COCO_train2014_000000209563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a fork near her mouth smiling at the camera\"."}], "task": "refering", "answer_template": "The \"a woman with a fork near her mouth smiling at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 152, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 267, 268, 269], "bbox": [0.438984375, 0.33337500000000003, 0.725390625, 0.6664583333333333], "iscrowd": 0, "area": 18530.692050000005, "rle": {"size": [480, 640], "counts": "hPT45\\>?I8H7M3N2N2M3N2N2I7K5L4K4M400O1O1O100O1O1O100O1O1O1O10O01O001O100O2O0O100O1O101NOmCWMR<j211O1O100N2N1L5G9H8J6J5M4M3M3M3M3O010O1000O0100000O010000O1000O100000O10000000001O0000001OnLmEW1S:hNPFV1P:jNRFT1n9kNTFT1k9mNVFR1j9nNXFP1h9POYFo0g9ROYFm0g9TOYFk0g9VOYFi0g9YOWFh0h9[OVFd0j9^OTFb0l9_OTF`0l9BSF=m9DQF=o9EoEoMJY2W:InEoMKW2W:LlEoMNR2W:OjESN5e1P::jEUN;W1l9S30^NiE[NX:f1gEZNX:i0hEdN0b0]:e0dEhN0b0b:`0^EnN0c0c:>]EoN0b0e:=[EQO0b0h::XETO0b0k:7VEVO0c0k:5UEXO0c0P;0PE^O0a0V;LiDC1a0\\;FcDJ1?_;D`DM3=^;F_DM4;];H_D036_;I_D224_;J^D44O_;L]D83K_;M_D92H`;O^D93G_;0_D92E`;2^D:2Ba;4]D;3^O`;8]D;3[Oa;:\\D>1VOc;=\\D>2ROc;`0ZD?4oNb;c0ZD?4lNc;e0ZD?3jNc;h0ZD`01gNf;i0YDb0OdNh;k0YDb0OaNi;m0XDb0OaNh;n0XDb01^Ng;Q1XDa0P<_OPDa0o;@RD?n;AQD`0n;@SD`0m;@SD?m;BSD=n;CRD<n;DRD<o;DQD;P<EPD:P<GPD8Q<GoC9R<HmC7S<JmC5T<KkC5V<LiC3W<NiC1X<[10dNhC0Y<0gCO]<McC3f<EZC:g<FYC9i<EXC:j<EVC9m<ETC9R=CnB:X=DgB:^=DaB:d=D\\B9g=>3M3M2\\OQB3S>LmA1X>MhA0^>LcA2b>KYWb2"}, "label": "a woman with a fork near her mouth smiling at the camera"}]}
{"id": 209563, "image": "COCO_train2014_000000209563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue shirt eating in between man and woman\"."}], "task": "refering", "answer_template": "The \"woman in blue shirt eating in between man and woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 152, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 267, 268, 269], "bbox": [0.438984375, 0.33337500000000003, 0.725390625, 0.6664583333333333], "iscrowd": 0, "area": 18530.692050000005, "rle": {"size": [480, 640], "counts": "hPT45\\>?I8H7M3N2N2M3N2N2I7K5L4K4M400O1O1O100O1O1O100O1O1O1O10O01O001O100O2O0O100O1O101NOmCWMR<j211O1O100N2N1L5G9H8J6J5M4M3M3M3M3O010O1000O0100000O010000O1000O100000O10000000001O0000001OnLmEW1S:hNPFV1P:jNRFT1n9kNTFT1k9mNVFR1j9nNXFP1h9POYFo0g9ROYFm0g9TOYFk0g9VOYFi0g9YOWFh0h9[OVFd0j9^OTFb0l9_OTF`0l9BSF=m9DQF=o9EoEoMJY2W:InEoMKW2W:LlEoMNR2W:OjESN5e1P::jEUN;W1l9S30^NiE[NX:f1gEZNX:i0hEdN0b0]:e0dEhN0b0b:`0^EnN0c0c:>]EoN0b0e:=[EQO0b0h::XETO0b0k:7VEVO0c0k:5UEXO0c0P;0PE^O0a0V;LiDC1a0\\;FcDJ1?_;D`DM3=^;F_DM4;];H_D036_;I_D224_;J^D44O_;L]D83K_;M_D92H`;O^D93G_;0_D92E`;2^D:2Ba;4]D;3^O`;8]D;3[Oa;:\\D>1VOc;=\\D>2ROc;`0ZD?4oNb;c0ZD?4lNc;e0ZD?3jNc;h0ZD`01gNf;i0YDb0OdNh;k0YDb0OaNi;m0XDb0OaNh;n0XDb01^Ng;Q1XDa0P<_OPDa0o;@RD?n;AQD`0n;@SD`0m;@SD?m;BSD=n;CRD<n;DRD<o;DQD;P<EPD:P<GPD8Q<GoC9R<HmC7S<JmC5T<KkC5V<LiC3W<NiC1X<[10dNhC0Y<0gCO]<McC3f<EZC:g<FYC9i<EXC:j<EVC9m<ETC9R=CnB:X=DgB:^=DaB:d=D\\B9g=>3M3M2\\OQB3S>LmA1X>MhA0^>LcA2b>KYWb2"}, "label": "woman in blue shirt eating in between man and woman"}]}
{"id": 94877, "image": "COCO_train2014_000000094877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse with a rider in a brown jacket\"."}], "task": "refering", "answer_template": "The \"the horse with a rider in a brown jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 152, 153, 154, 167, 169, 170, 171, 184, 185, 187, 188, 189], "bbox": [0.26548, 0.5966366366366367, 0.5711400000000001, 0.920990990990991], "iscrowd": 0, "area": 3786.703999999999, "rle": {"size": [333, 500], "counts": "o`[13V:5M3M2N2\\FCW9>hFCW9>hFDU9>jFCP9c0nF@n8c0PG@o8o00O100O100O0O2O1O1O011L3M3O2M2O1O2M2O1O10O01H9L3Lf]9N`bF1O000O2O001O0O2O0nFJm76RH0i71UH0j71UHOl71THOl71THNP8OoG1S8NlG2U8NjG1W80hG1X8OgG1Z8OeG1T87jGJW86fGLZ84eGN[82dGN\\83bGN_82_GOa82]GOd81YG1g8i02O\\OhNUHW1i7PORHP1l7XOoGf0m7EmG9V8l03O0O2N2N1N3001N2O1O2M2O@dGPO[8P1fGPOX8P1kGoNT8P1nGPOQ8n0SHQOk7o0WHROg7i0`HVO_7f0fH[OW7d0mH\\OQ7c0QI]On6a0UI@h6`0[I_Od6?_IB_6<eICY6=iICW6:mIES69oIGQ66RJKk54YJKd56^JJa54cJK[57eJIZ58gJGW5;iJDW5<jJBV5`0kJ^OU5c0kJ[OU5g0lJWOT5i0Z2O3M4K101O000O2L3N2N3L3N^fU2"}, "label": "the horse with a rider in a brown jacket"}]}
{"id": 94877, "image": "COCO_train2014_000000094877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3 brown horses next to a lake caring 3 people\"."}], "task": "refering", "answer_template": "The \"3 brown horses next to a lake caring 3 people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 152, 153, 154, 167, 169, 170, 171, 184, 185, 187, 188, 189], "bbox": [0.26548, 0.5966366366366367, 0.5711400000000001, 0.920990990990991], "iscrowd": 0, "area": 3786.703999999999, "rle": {"size": [333, 500], "counts": "o`[13V:5M3M2N2\\FCW9>hFCW9>hFDU9>jFCP9c0nF@n8c0PG@o8o00O100O100O0O2O1O1O011L3M3O2M2O1O2M2O1O10O01H9L3Lf]9N`bF1O000O2O001O0O2O0nFJm76RH0i71UH0j71UHOl71THOl71THNP8OoG1S8NlG2U8NjG1W80hG1X8OgG1Z8OeG1T87jGJW86fGLZ84eGN[82dGN\\83bGN_82_GOa82]GOd81YG1g8i02O\\OhNUHW1i7PORHP1l7XOoGf0m7EmG9V8l03O0O2N2N1N3001N2O1O2M2O@dGPO[8P1fGPOX8P1kGoNT8P1nGPOQ8n0SHQOk7o0WHROg7i0`HVO_7f0fH[OW7d0mH\\OQ7c0QI]On6a0UI@h6`0[I_Od6?_IB_6<eICY6=iICW6:mIES69oIGQ66RJKk54YJKd56^JJa54cJK[57eJIZ58gJGW5;iJDW5<jJBV5`0kJ^OU5c0kJ[OU5g0lJWOT5i0Z2O3M4K101O000O2L3N2N3L3N^fU2"}, "label": "3 brown horses next to a lake caring 3 people"}]}
{"id": 94877, "image": "COCO_train2014_000000094877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown horse with grey saddle blanket\"."}], "task": "refering", "answer_template": "The \"brown horse with grey saddle blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 182, 183, 184, 186, 187], "bbox": [0.13262000000000002, 0.6212912912912912, 0.415, 0.918948948948949], "iscrowd": 0, "area": 6124.665650000001, "rle": {"size": [333, 500], "counts": "Zge08Q:5L4L1O2O000O011UG\\Om7d0PHAm7a0RH@m7a0RHAl7`0THAi7a0VH@i7`0WHAj7>UHCl7<THDl7<SHEn79RHHo75RHKP82QHOo7ORH1P8LQH4Q8JoG7Q8JmG6U8JiG7X8N[G8g8f00O2N2O1O1O0010O000100001O1N3N1O2N2N3M2N4L3M2N1O1O1O1O10POTH^Ok7b0WH]Oh7b0[H]Od7c0_H[O`7e0bHZO]7f0eHYOZ7f0jHXOV7g0kHYOT7g0nHXOQ7h0PIWOP7h0RIVOo6j0RIUOn6k0Z1001O0000000000001OO010000000000O001O101N1O1O100O2N1O100O101N1O2O0O2N1O2N2N2N2O1N3N1N3M4M2M3N1O1O1O2N`N]H7b7DfH:Y7CmH;R7DRI:m6FVI8i6HYI7f6IgIKX65mIFS6:oIEP6<QJCn5=TJBl5>TJBk5?VJ_Ok5`0VJ@k5?UJ@m5>UJ@m5?SJ@o5?RJ_OQ6>PJAR6>oI_OT6?o1N3M2M^gn2"}, "label": "brown horse with grey saddle blanket"}]}
{"id": 94877, "image": "COCO_train2014_000000094877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the back\"."}], "task": "refering", "answer_template": "The \"the horse in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 182, 183, 184, 186, 187], "bbox": [0.13262000000000002, 0.6212912912912912, 0.415, 0.918948948948949], "iscrowd": 0, "area": 6124.665650000001, "rle": {"size": [333, 500], "counts": "Zge08Q:5L4L1O2O000O011UG\\Om7d0PHAm7a0RH@m7a0RHAl7`0THAi7a0VH@i7`0WHAj7>UHCl7<THDl7<SHEn79RHHo75RHKP82QHOo7ORH1P8LQH4Q8JoG7Q8JmG6U8JiG7X8N[G8g8f00O2N2O1O1O0010O000100001O1N3N1O2N2N3M2N4L3M2N1O1O1O1O10POTH^Ok7b0WH]Oh7b0[H]Od7c0_H[O`7e0bHZO]7f0eHYOZ7f0jHXOV7g0kHYOT7g0nHXOQ7h0PIWOP7h0RIVOo6j0RIUOn6k0Z1001O0000000000001OO010000000000O001O101N1O1O100O2N1O100O101N1O2O0O2N1O2N2N2N2O1N3N1N3M4M2M3N1O1O1O2N`N]H7b7DfH:Y7CmH;R7DRI:m6FVI8i6HYI7f6IgIKX65mIFS6:oIEP6<QJCn5=TJBl5>TJBk5?VJ_Ok5`0VJ@k5?UJ@m5>UJ@m5?SJ@o5?RJ_OQ6>PJAR6>oI_OT6?o1N3M2M^gn2"}, "label": "the horse in the back"}]}
{"id": 94877, "image": "COCO_train2014_000000094877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that the person in the green jacket is sitting on\"."}], "task": "refering", "answer_template": "The \"the horse that the person in the green jacket is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 153, 154, 155, 168, 169, 171, 172, 186, 187, 188, 189, 190], "bbox": [0.34692, 0.5818018018018019, 0.6540199999999999, 0.9060960960960962], "iscrowd": 0, "area": 2867.55485, "rle": {"size": [333, 500], "counts": "_ah11T:9G8011N1012MO2@XF6Q:O1O1O1O1N2Okc00R\\O4K6J500O2O0O2O0OEYF2f9J_F5`9GeF;g93M0O01O1O5K^[<KbdC6J5L401[OEUG;S8EhG952Q8FjG96@J5U84jG9a0Ad77kG9b0^Oa7;lG9d0ZO_7>mG5n8MQG1Q9c02N101N310O10O010O010O010VOROTHm0k7VOSHk0l7XORHh0m7ZOQHf0n7\\ORHd0m7]OSHb0m7@QH0EMX85QHMJMU87oGKONQ88mGK4Ln7<kGG:Mj7=jG]Og05^7e0eH[OZ7d0hH\\OB0j6b0gI_OPO<X75kI7T6HnI7R6GQJ9m5G]2N2O1O1001O0000001O0000001O00001O1O001O1O1N2N101N27I2N2M3N2M3L2N3M2M3N3MZQh1"}, "label": "the horse that the person in the green jacket is sitting on"}]}
{"id": 53928, "image": "COCO_train2014_000000053928.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suitcase with handle extension\"."}], "task": "refering", "answer_template": "The \"a black suitcase with handle extension\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362, 363, 383, 384, 385, 386], "bbox": [0.59909375, 0.5004935622317597, 0.823984375, 0.9840343347639485], "iscrowd": 0, "area": 22614.748250000008, "rle": {"size": [466, 640], "counts": "ho^5:X>9G9F4MO1N101N2N101N2O0O2N2O0O2O1N1O2O1N1O2OfB7m;GaCo0\\<f0N2N2M3N2N2M3N2N3L`0A?A`0_O`0A?A?@`0A?A5K1N101O0O2O001O010O001O001O001O000010O01O001O001O001O0001O00000001O0000000001O00000000001O00010O0000001O0000001O00001O01O0001O0000001O00001O0000010N1O1N3N2N2N3M2M3N2N2N2M3N2N2N3M<C>C<D<D<C=D<D<D<D<C=@`0\\OaQc1"}, "label": "a black suitcase with handle extension"}]}
{"id": 53928, "image": "COCO_train2014_000000053928.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a traveller bag near the red color traveller color bag\"."}], "task": "refering", "answer_template": "The \"a traveller bag near the red color traveller color bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362, 363, 383, 384, 385, 386], "bbox": [0.59909375, 0.5004935622317597, 0.823984375, 0.9840343347639485], "iscrowd": 0, "area": 22614.748250000008, "rle": {"size": [466, 640], "counts": "ho^5:X>9G9F4MO1N101N2N101N2O0O2N2O0O2O1N1O2O1N1O2OfB7m;GaCo0\\<f0N2N2M3N2N2M3N2N3L`0A?A`0_O`0A?A?@`0A?A5K1N101O0O2O001O010O001O001O001O000010O01O001O001O001O0001O00000001O0000000001O00000000001O00010O0000001O0000001O00001O01O0001O0000001O00001O0000010N1O1N3N2N2N3M2M3N2N2N2M3N2N2N3M<C>C<D<D<C=D<D<D<D<C=@`0\\OaQc1"}, "label": "a traveller bag near the red color traveller color bag"}]}
{"id": 53928, "image": "COCO_train2014_000000053928.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red luggage on a train\"."}], "task": "refering", "answer_template": "The \"red luggage on a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 174, 175, 197, 198, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337, 356, 357, 358, 359, 360, 379, 380, 381, 382, 383], "bbox": [0.47426562499999997, 0.368175965665236, 0.667703125, 0.9998927038626609], "iscrowd": 0, "area": 24466.96579999999, "rle": {"size": [466, 640], "counts": "feZ4l1_<;E`0C;E9G9G8I7I6J5J5L4M3M3N2N2M3N2N2N2N2O0O2N2O1N2O1N2O1N2O1N2N1000000000000O100000000001O000000O010O10000O1000000000000000000O100oLXJXMh5e2RKdLn4\\3^LXKb3h4T300000000000000O100000000000O100000000000000000000000001O2N2N`3`L00001O1O1O001O2N1O2N1O5K4L5K5K5K4EjERL[:f3jEXL[:`3b0H=\\Of0kNWcP3"}, "label": "red luggage on a train"}]}
{"id": 53928, "image": "COCO_train2014_000000053928.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red / pink suitcase with a handle\"."}], "task": "refering", "answer_template": "The \"a red / pink suitcase with a handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 174, 175, 197, 198, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337, 356, 357, 358, 359, 360, 379, 380, 381, 382, 383], "bbox": [0.47426562499999997, 0.368175965665236, 0.667703125, 0.9998927038626609], "iscrowd": 0, "area": 24466.96579999999, "rle": {"size": [466, 640], "counts": "feZ4l1_<;E`0C;E9G9G8I7I6J5J5L4M3M3N2N2M3N2N2N2N2O0O2N2O1N2O1N2O1N2O1N2N1000000000000O100000000001O000000O010O10000O1000000000000000000O100oLXJXMh5e2RKdLn4\\3^LXKb3h4T300000000000000O100000000000O100000000000000000000000001O2N2N`3`L00001O1O1O001O2N1O2N1O5K4L5K5K5K4EjERL[:f3jEXL[:`3b0H=\\Of0kNWcP3"}, "label": "a red / pink suitcase with a handle"}]}
{"id": 127657, "image": "COCO_train2014_000000127657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty blue cot under a white tenet\"."}], "task": "refering", "answer_template": "The \"empty blue cot under a white tenet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 57, 58, 59, 60, 61, 62, 63, 64, 65, 70, 71, 75, 89], "bbox": [0.16013999999999998, 0.09144144144144144, 1.0, 0.34762762762762767], "iscrowd": 0, "area": 14366.682, "rle": {"size": [333, 500], "counts": "^Qj07W:>A>B>B>B?A7IO2O000000000O10001O0000000O1000001O00000O1000001O0000000O1000001O0000000000001O000000000000001N100000000000001O00001O001O00001O00001O001O00001O0O101O001O00001O001O00001O00001O001O00001O0kNgG6Y8\\OUHd0j8000000O100000O1000000000000000000O1000000000000000000O10O1000000000000000O1000000000000000000O100000000000O100000O1000000000000000000O1000000000000000000O10O1000000000000000O1000000000000000000O100000000000O100000O100000000000000000000O100000O1O1O1O1O1O1O001O1O1O1O1O1O1O100O010O100O100O100O100O00100O100O100O100O11N10000000002N0000O10O10O10000O10000O1000O0100O10000O10000O10O10O100O100O10000O10O0100O100O10000O10\\:0ZW40Z^K001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O1O001O001O1O001O1O001O1O001N2O2N1O2N1O1O2N5K3M4L1O000000000000000000000000000000000000001O000000000000000000000RN"}, "label": "empty blue cot under a white tenet"}]}
{"id": 127657, "image": "COCO_train2014_000000127657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with a blue blanket and nobody on it\"."}], "task": "refering", "answer_template": "The \"a bed with a blue blanket and nobody on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 57, 58, 59, 60, 61, 62, 63, 64, 65, 70, 71, 75, 89], "bbox": [0.16013999999999998, 0.09144144144144144, 1.0, 0.34762762762762767], "iscrowd": 0, "area": 14366.682, "rle": {"size": [333, 500], "counts": "^Qj07W:>A>B>B>B?A7IO2O000000000O10001O0000000O1000001O00000O1000001O0000000O1000001O0000000000001O000000000000001N100000000000001O00001O001O00001O00001O001O00001O0O101O001O00001O001O00001O00001O001O00001O0kNgG6Y8\\OUHd0j8000000O100000O1000000000000000000O1000000000000000000O10O1000000000000000O1000000000000000000O100000000000O100000O1000000000000000000O1000000000000000000O10O1000000000000000O1000000000000000000O100000000000O100000O100000000000000000000O100000O1O1O1O1O1O1O001O1O1O1O1O1O1O100O010O100O100O100O100O00100O100O100O100O11N10000000002N0000O10O10O10000O10000O1000O0100O10000O10000O10O10O100O100O10000O10O0100O100O10000O10\\:0ZW40Z^K001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O1O001O001O1O001O1O001O1O001N2O2N1O2N1O1O2N5K3M4L1O000000000000000000000000000000000000001O000000000000000000000RN"}, "label": "a bed with a blue blanket and nobody on it"}]}
{"id": 127657, "image": "COCO_train2014_000000127657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue cot the boy is laying on\"."}], "task": "refering", "answer_template": "The \"blue cot the boy is laying on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 54, 55, 56, 72, 73, 74, 90, 91, 92, 106, 107, 108, 109, 110, 111, 123, 124, 125, 126, 127, 128, 140, 141, 142, 143, 144, 145, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.0, 0.24045045045045044, 1.0, 0.9887687687687687], "iscrowd": 0, "area": 51177.66584999999, "rle": {"size": [333, 500], "counts": "a2n0_1]2V3VNWLj3i1VLWNj3i1VLWNj3i1WLWNh3i1XLWNh3i1XLWNh3h1ZLWNf3i1ZLWNf3i1[LWNd3i1\\LWNd3i1\\LWNd3i1]LVNc3j1]LVNV1:]N`1=WNT1<]N]1`0VNQ1?_N[1`0VNP1`0`NZ1`0VNo0a0aNY1a0UNm0c0bNX1a0VNj0d0eNW1`0UNj0e0fNW1?TNj0f0gNV1?TNi0g0hNV1>SNh0i0jNU1=RNh0i0lNU1=QNf0k0mNU1<QNe0k0oNU1;PNe0l0POT1;PNc0n0ROS1:oMc0n0TOT18nMc0o0UOS18nMb0P1VOS17nMa0P1XOR17nM?R1ZOQ16mM?R1\\OR15lM=S1^OQ15lM<T1_OQ15kM:U1AQ15jM8V1CP15jM8U1DR14iM6V1FR13hM6W1GQ13iM4W1IQ13gM4X1IR12gM3W1LR11gM2X1MR10gM2W1NS1OgM1W10R10fMOY11R1OfMNY13R1NeMOY13R1NfMMY15R1MeMMZ16R1MdMK[18Q1MdMK[18R1LdMJ[1:Q1LdMI\\1;Q1KcMJ\\1;R1JbMJ]1<Q1KbMG^1>Q1JaMG_1?Q1I`MH^1`0R1H`MG_1a0R1FaMG^1c0R1E`MG_1d0Q1DaMH^1d0R1BaMI^1e0R1@bMJ\\1f0R1@bMI]1g0R1^ObMK\\1g0S1\\ObMM[1g0S1ZOeMMY1i0S1XOeMOX1i0T1VOeM1W1i0T1UOgM0V1k0T1SOgM2U1k0T1QOiM4S1k0U1oNjM4R1m0U1mNjM6Q1m0U1lNlM6o0n0V1jNlM7o0o0V1hNmM8m0P1V1gNnM9l0P1W1eNnM:k0R1V3mNjLS1V3mNjLS1W3lNiLT1W3lNiLT1X3jNiLV1W3jNiLV1W3jNiLV1X3iNhLW1X3iNhLW1X3iNhLW1Y3hNgLX1Y3hNgLX1Y3gNhLY1Y3fNgLZ1Y3fNgLZ1Y3fNgLZ1Z3eNfL[1Z3eNeL\\1[3dNeL\\1\\3cNdL]1\\3bNeL^1[3bNeL^1U60000000000000000000000000O100000000O1000000O1000000O1000000O1000000O100000000O100000000O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O10000000000O1O100O1O100O1O100O1O1001O00000000001OO1O100O100O1O100O1O100O1O100O1O100O1000000000000000000O1000000O1000000O1000000O1O010O100O1O100O1000000O100000000O100000000O10000O100O10000O10000O100O10000O10000O10000000000O10000000000O1000000000000001O0000001O0000001O0000001O000000001O000000000000001O0000000000O10000000000O1000000000000O100000000000000O100000000000000O10000O10000O10000O10000O10000O100O100O100O100O1L4K5O10000000000O10000000000O10000O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1N2M3M3N2M3M3N2M3N2000000001O000000001O0000001O0000001O000000001O0000001O0lMT2fMfM"}, "label": "blue cot the boy is laying on"}]}
{"id": 53929, "image": "COCO_train2014_000000053929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown color couch in the image\"."}], "task": "refering", "answer_template": "The \"brown color couch in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 81, 99, 100, 101, 104, 105, 106, 107, 121, 122, 123, 124, 127, 128, 129, 130, 144, 145, 146, 151, 152, 153, 167, 168, 174, 175, 176, 187, 188, 189, 190, 191, 199, 200, 210, 211, 212, 213, 214, 221, 222, 223, 224, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 256, 257, 258, 259, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.133515625, 0.2268565400843882, 0.7468437499999999, 0.996097046413502], "iscrowd": 0, "area": 80530.70420000001, "rle": {"size": [474, 640], "counts": "_eW16^>g1YNh0WOi0XO`0@8H8H8L3N3SOUKgGm4U8WKhGl4S8YKkGi4Q8[KmGg4n7^KoGd4n7`KPHb4k7cKSH_4j7dKSH_4j7dKTH^4h7fKVH\\4g7gKVH[4i7]1NO10O010000O100O100O10000O100O10000O100000000000000001O000O100000000000001O0000000000000000000000001O0000000O10000000000O10001O00000ON3YKXHY2k7bM\\HY2g7cM`HW2c7dMeHX2\\7gMgHW2X7iMkHU2U7kMlHT2k6TNWIk1`6^NaIa1U6hNmIV1k5SOWJl0`4QM]J]2U1b0[4=fKCX4=jKCT4>mKBP4?RLBk3?VLAh3?[L@c3a0^L_OgN[LZ4W4QM^O_N`L`4R4QMKo24RMLm25SMKm25SMKm24SMMm23SMMm23SMMl23TMNl22TMNl2^NiKcN[1o2j2_NQL\\NV1U3f2aNVLYNS1W3e2aNYLXNR1W3b2cN^LTNQ1Y3_2cNcLSNm0[3]2dNhLPNk0\\3Z2fNmLmMi0]3Y2fNoLlMi0^3W2fNRMkMf0`3X2dNTMkMd0a3X2bNWMjMc0d3V2^N\\MhMc0j3Q2[N_MeMf0P4k1XNcMbMg0V4f1TN4l1KRN8n1HoM;Q2DmM?S2AlM?U2@jMb0V2^OiMc0W2\\OhMf0X2ZOgMg0Y2XOfMj0Z2VOeMk0[2SOeMo0[2oNfMR1Z2lNfMU1[2jNdMX1\\2hNcMY1]2fNbM\\1^2dNaM]1_2cN_M^1b2bN]M_1c2aN[M`1f2_NZMb1f2^NXMd1h2\\NWMd1j2\\NVMd1j2[NVMe1k2[NUMe1k2[NTMf1l2ZNTMe1m2[NRMf1n2YNSMf1n2ZNQMg1o2YNQMg1o2YNQMf1P3ZNoLg1Q3YNoLg1Q3YNnLg1S3YNlLh1T3XNlLg1U3YNjLh1V3XNjLh1V3XNiLh1X3XNhLh1X3XNgLi1Y3WNgLi1Y3WNfLj1Z3VNeLj1\\3VNdLj1\\3VNcLk1]3UNbLl1^3TNbLl1^3TNaLl1`3TN_Lm1a3SN^Ln1b3WNXLj1h3[NnKj1R4]NbKh1^4Y4100000000O10000000000O100000000O11O001O001O001O001O0000O100000001O000O10000O100O10000O100O10000O100O10000O100O100O10000O100O10000O100mJnJ=S5_OQKa0o4[OUKe0k4XOWKh0j4TOZKl0f4PO^Ko0c4mNaKS1_4iNeKW1[4fNgKZ1Z4cNiK]1W4cNjK[1W4dNmKY1S4gNPLV1P4jNRLS1o3mNRLR1n3nNRLQ1o3nNSLQ1m3oNTLP1l3POTLo0m3QOTLn0l3ROULl0l3TOTLl0l3UOTLj0l3VOULh0l3YOSLg0m3YOTLe0m3[OTLd0l3]OTLb0l3^OTLa0m3@SL?m3ATL=m3DUL9k3GXL6h3J\\L1e30]LMc33`LIa38bLD^3<eLA[3?hL]OY3d0jLXOV3h0mLTOT3m0lLROT3n0nLPOR3Q1nLmNS3S1nLlNR3T1PMiNQ3X1oLgNQ3Y1nLhNR3Y1mLfNT3Z1lLfNT3[1kLdNV3\\1jLdNV3\\1iLeNW3\\1hLcNY3]1gLbNZ3_1eL_N[3c1eL\\N[3e1eLZN[3h1dLWN]3i1cLUN_3l1`LSNa3m1_LRNb3o1]LoMe3Q2[LnMf3R2ZLmMg3T2YLjMh3V2XLhMj3Y2ULfMl3Z2TLeMn3Z2RLdMP4]2oKbMR4^2nKaMS4`2lK_MU4a2lK\\MV4e2iKZMX4f2hKYMY4g2gKWM[4j2dKUM]4k2cKTM^4m2aKRM`4n2`KPMb4P3_KnLb4S3]KlLd4T3_KgLd4Y3^KcLc4]3_K`Lb4a3`K[La4e3bKVL`4j3fJPLdN3g6n3nIoLTOlNR7U4hITMUO`NX7]4`IWMg6i2WI\\Mf6e2VI`Mh6b2TIcMi6`2RIcMm6_2oHcMQ7`2jHbMV7a2dHbM]7n4N1O1O1O1O2O0O1O1O1O2H7F:G9F:G:E:G9F:jL[GNn82n2N100O1O101N2O2M2O1N3N1N3J5J6J7J5JlgZ2"}, "label": "brown color couch in the image"}]}
{"id": 53929, "image": "COCO_train2014_000000053929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch on which the lady is sitting\"."}], "task": "refering", "answer_template": "The \"couch on which the lady is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 81, 99, 100, 101, 104, 105, 106, 107, 121, 122, 123, 124, 127, 128, 129, 130, 144, 145, 146, 151, 152, 153, 167, 168, 174, 175, 176, 187, 188, 189, 190, 191, 199, 200, 210, 211, 212, 213, 214, 221, 222, 223, 224, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 256, 257, 258, 259, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.133515625, 0.2268565400843882, 0.7468437499999999, 0.996097046413502], "iscrowd": 0, "area": 80530.70420000001, "rle": {"size": [474, 640], "counts": "_eW16^>g1YNh0WOi0XO`0@8H8H8L3N3SOUKgGm4U8WKhGl4S8YKkGi4Q8[KmGg4n7^KoGd4n7`KPHb4k7cKSH_4j7dKSH_4j7dKTH^4h7fKVH\\4g7gKVH[4i7]1NO10O010000O100O100O10000O100O10000O100000000000000001O000O100000000000001O0000000000000000000000001O0000000O10000000000O10001O00000ON3YKXHY2k7bM\\HY2g7cM`HW2c7dMeHX2\\7gMgHW2X7iMkHU2U7kMlHT2k6TNWIk1`6^NaIa1U6hNmIV1k5SOWJl0`4QM]J]2U1b0[4=fKCX4=jKCT4>mKBP4?RLBk3?VLAh3?[L@c3a0^L_OgN[LZ4W4QM^O_N`L`4R4QMKo24RMLm25SMKm25SMKm24SMMm23SMMm23SMMl23TMNl22TMNl2^NiKcN[1o2j2_NQL\\NV1U3f2aNVLYNS1W3e2aNYLXNR1W3b2cN^LTNQ1Y3_2cNcLSNm0[3]2dNhLPNk0\\3Z2fNmLmMi0]3Y2fNoLlMi0^3W2fNRMkMf0`3X2dNTMkMd0a3X2bNWMjMc0d3V2^N\\MhMc0j3Q2[N_MeMf0P4k1XNcMbMg0V4f1TN4l1KRN8n1HoM;Q2DmM?S2AlM?U2@jMb0V2^OiMc0W2\\OhMf0X2ZOgMg0Y2XOfMj0Z2VOeMk0[2SOeMo0[2oNfMR1Z2lNfMU1[2jNdMX1\\2hNcMY1]2fNbM\\1^2dNaM]1_2cN_M^1b2bN]M_1c2aN[M`1f2_NZMb1f2^NXMd1h2\\NWMd1j2\\NVMd1j2[NVMe1k2[NUMe1k2[NTMf1l2ZNTMe1m2[NRMf1n2YNSMf1n2ZNQMg1o2YNQMg1o2YNQMf1P3ZNoLg1Q3YNoLg1Q3YNnLg1S3YNlLh1T3XNlLg1U3YNjLh1V3XNjLh1V3XNiLh1X3XNhLh1X3XNgLi1Y3WNgLi1Y3WNfLj1Z3VNeLj1\\3VNdLj1\\3VNcLk1]3UNbLl1^3TNbLl1^3TNaLl1`3TN_Lm1a3SN^Ln1b3WNXLj1h3[NnKj1R4]NbKh1^4Y4100000000O10000000000O100000000O11O001O001O001O001O0000O100000001O000O10000O100O10000O100O10000O100O10000O100O100O10000O100O10000O100mJnJ=S5_OQKa0o4[OUKe0k4XOWKh0j4TOZKl0f4PO^Ko0c4mNaKS1_4iNeKW1[4fNgKZ1Z4cNiK]1W4cNjK[1W4dNmKY1S4gNPLV1P4jNRLS1o3mNRLR1n3nNRLQ1o3nNSLQ1m3oNTLP1l3POTLo0m3QOTLn0l3ROULl0l3TOTLl0l3UOTLj0l3VOULh0l3YOSLg0m3YOTLe0m3[OTLd0l3]OTLb0l3^OTLa0m3@SL?m3ATL=m3DUL9k3GXL6h3J\\L1e30]LMc33`LIa38bLD^3<eLA[3?hL]OY3d0jLXOV3h0mLTOT3m0lLROT3n0nLPOR3Q1nLmNS3S1nLlNR3T1PMiNQ3X1oLgNQ3Y1nLhNR3Y1mLfNT3Z1lLfNT3[1kLdNV3\\1jLdNV3\\1iLeNW3\\1hLcNY3]1gLbNZ3_1eL_N[3c1eL\\N[3e1eLZN[3h1dLWN]3i1cLUN_3l1`LSNa3m1_LRNb3o1]LoMe3Q2[LnMf3R2ZLmMg3T2YLjMh3V2XLhMj3Y2ULfMl3Z2TLeMn3Z2RLdMP4]2oKbMR4^2nKaMS4`2lK_MU4a2lK\\MV4e2iKZMX4f2hKYMY4g2gKWM[4j2dKUM]4k2cKTM^4m2aKRM`4n2`KPMb4P3_KnLb4S3]KlLd4T3_KgLd4Y3^KcLc4]3_K`Lb4a3`K[La4e3bKVL`4j3fJPLdN3g6n3nIoLTOlNR7U4hITMUO`NX7]4`IWMg6i2WI\\Mf6e2VI`Mh6b2TIcMi6`2RIcMm6_2oHcMQ7`2jHbMV7a2dHbM]7n4N1O1O1O1O2O0O1O1O1O2H7F:G9F:G:E:G9F:jL[GNn82n2N100O1O101N2O2M2O1N3N1N3J5J6J7J5JlgZ2"}, "label": "couch on which the lady is sitting"}]}
{"id": 152238, "image": "COCO_train2014_000000152238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother elephant\"."}], "task": "refering", "answer_template": "The \"a mother elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 470, 471, 472, 473, 475, 476, 477, 479, 480], "bbox": [0.19506535947712417, 0.09181372549019608, 1.0, 0.9886764705882352], "iscrowd": 0, "area": 149257.00755, "rle": {"size": [612, 612], "counts": "c\\W2Z1ga03L4L4L4M3L3M4M3L4M3O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2N101N2O1N2O1N200O1O1O10O01O1O1O100O1O1O100O001O1O100O1O1O100O1O1O1O100O1O2N100O1O1O2N100O1O1O2O0O1O1O1O101N1O1O100O2N1O1O100O2N1O100O1L4M4K4M3L4L4M4K4N2O1O1O2N1O1O102M3M4L3M4L3M3M4L3M4L3M4M2M2N2N2O1N3M2N2ZKRITMo6P2lIoMW6k1mITNU6f1oIZNS6`1RJ`No5[1UJdNm5V1XJiNk5Q1YJnNi5l0\\JTOe5g0_JXOc5b0aJ^Oa5=cJB`57eJI\\52hJMZ5MkJ2W5ImJ6U5DoJ=S5]OQKb0Q5XOTKg0n4TOVKk0l4oNYKQ1h4jN\\KU1g4aNaK^1a4WNhKi1Z4lMPLT2Q4aMYL^2i3WMaLh2n9O10O01O1O100O1O1O010O1O1O100O010000000O1000O10000000000O1000000000000`LYNbEg1[:AaD?[;J`D6\\;2`DN\\;;_DE];d0^D\\O^;l0]DUO_;T1\\DlN`;]1[DcNa;b4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L44L4L3M4L4L3M4K5L3M4L3M4L4L3M4L4L7I;E:F;E;D;F;D<Eh1WN2O1N2O1N2O1N2O1N2N2O0O2O0O100O2O0O100O2N100O101N100O1010O000001O0001O01O0000010O0000001O01O0001O0001O00000000001O000000001O0000000iLZNPEf1V:YOeEg0a98ZFHk8X1PGhNV8W2eGiM`7W3[HiLb7_3YHaLg7d3TH\\Lk7j3PHVLP8o3kGQLT8U4gGkKY8Z4bGfK^8_4]GaKb8e4YG[Kg8j4TGVKk8P5PGPKP9U5kFkJT9[5gFeJ\\9`74L5K4L4L5K4L7I9G9G8H9G9G8H9G8I8H8H7I8H8H7J7H8H7I8H8H7I8H7I8I6IO2N1O2N1O2N1O2M2O2N1O2N1O1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N1O2N1O2N1N3N4ZOh0oNP1QOP1oNP1QOP1POo0POQ1H7K6J5L5J6J5K6UN]CjLg<l2`CRMf<a2bC^Mc<_1\\D`Nn;T1UDlNU<j0kCUO_<b0aC]Oj<9VCFm<8SCHm<8SCGn<9RCFP=9PCFQ=:oBFQ=:oBER=;nBDS=<mBCT==lBCT==lBBU=>kBAV=?jBAV=?kB_OV=a0jB^OW=b0iB]OX=c0hB]OX=c0hB\\OY=d0gB[OZ=e0eB[O\\=e0dB[O\\=e0cB[O^=e0bBZO_=f0`BZOa=f0^B[Ob=e0^BZOc=f0\\BZOe=e0\\BZOe=f0ZB[Of=e0ZBZOg=f0XBZOi=f0VBZOl=e0TBZOm=f0RBZOo=f0QBYOP>g0oAYOR>g0nAXOS>h0lAXOU>h0jAXOW>h0iAWOX>i0gAWOZ>i0fAVO[>j0dAVO]>j0bAVO_>j0aAUO`>k0_AUOb>k0^ATOc>l0\\ATOe>l0[ASOf>m0YASOh>m0WASOj>l0WASOj>l0VATOk>k0VATOk>k0UAUOl>k0TATOm>k0SAUOn>j0RAVOo>i0RAVOP?h0PAXOQ?h0o@WOR?h0n@XOS?g0n@XOS?g0m@YOT?g0m@WOT?h0QASOP?l0UAoNl>P1e1O100O1O1O1O100O1O1O1@WE"}, "label": "a mother elephant"}]}
{"id": 152238, "image": "COCO_train2014_000000152238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large adult elephant\"."}], "task": "refering", "answer_template": "The \"the large adult elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 470, 471, 472, 473, 475, 476, 477, 479, 480], "bbox": [0.19506535947712417, 0.09181372549019608, 1.0, 0.9886764705882352], "iscrowd": 0, "area": 149257.00755, "rle": {"size": [612, 612], "counts": "c\\W2Z1ga03L4L4L4M3L3M4M3L4M3O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2N101N2O1N2O1N200O1O1O10O01O1O1O100O1O1O100O001O1O100O1O1O100O1O1O1O100O1O2N100O1O1O2N100O1O1O2O0O1O1O1O101N1O1O100O2N1O1O100O2N1O100O1L4M4K4M3L4L4M4K4N2O1O1O2N1O1O102M3M4L3M4L3M3M4L3M4L3M4M2M2N2N2O1N3M2N2ZKRITMo6P2lIoMW6k1mITNU6f1oIZNS6`1RJ`No5[1UJdNm5V1XJiNk5Q1YJnNi5l0\\JTOe5g0_JXOc5b0aJ^Oa5=cJB`57eJI\\52hJMZ5MkJ2W5ImJ6U5DoJ=S5]OQKb0Q5XOTKg0n4TOVKk0l4oNYKQ1h4jN\\KU1g4aNaK^1a4WNhKi1Z4lMPLT2Q4aMYL^2i3WMaLh2n9O10O01O1O100O1O1O010O1O1O100O010000000O1000O10000000000O1000000000000`LYNbEg1[:AaD?[;J`D6\\;2`DN\\;;_DE];d0^D\\O^;l0]DUO_;T1\\DlN`;]1[DcNa;b4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L44L4L3M4L4L3M4K5L3M4L3M4L4L3M4L4L7I;E:F;E;D;F;D<Eh1WN2O1N2O1N2O1N2O1N2N2O0O2O0O100O2O0O100O2N100O101N100O1010O000001O0001O01O0000010O0000001O01O0001O0001O00000000001O000000001O0000000iLZNPEf1V:YOeEg0a98ZFHk8X1PGhNV8W2eGiM`7W3[HiLb7_3YHaLg7d3TH\\Lk7j3PHVLP8o3kGQLT8U4gGkKY8Z4bGfK^8_4]GaKb8e4YG[Kg8j4TGVKk8P5PGPKP9U5kFkJT9[5gFeJ\\9`74L5K4L4L5K4L7I9G9G8H9G9G8H9G8I8H8H7I8H8H7J7H8H7I8H8H7I8H7I8I6IO2N1O2N1O2N1O2M2O2N1O2N1O1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N1O2N1O2N1N3N4ZOh0oNP1QOP1oNP1QOP1POo0POQ1H7K6J5L5J6J5K6UN]CjLg<l2`CRMf<a2bC^Mc<_1\\D`Nn;T1UDlNU<j0kCUO_<b0aC]Oj<9VCFm<8SCHm<8SCGn<9RCFP=9PCFQ=:oBFQ=:oBER=;nBDS=<mBCT==lBCT==lBBU=>kBAV=?jBAV=?kB_OV=a0jB^OW=b0iB]OX=c0hB]OX=c0hB\\OY=d0gB[OZ=e0eB[O\\=e0dB[O\\=e0cB[O^=e0bBZO_=f0`BZOa=f0^B[Ob=e0^BZOc=f0\\BZOe=e0\\BZOe=f0ZB[Of=e0ZBZOg=f0XBZOi=f0VBZOl=e0TBZOm=f0RBZOo=f0QBYOP>g0oAYOR>g0nAXOS>h0lAXOU>h0jAXOW>h0iAWOX>i0gAWOZ>i0fAVO[>j0dAVO]>j0bAVO_>j0aAUO`>k0_AUOb>k0^ATOc>l0\\ATOe>l0[ASOf>m0YASOh>m0WASOj>l0WASOj>l0VATOk>k0VATOk>k0UAUOl>k0TATOm>k0SAUOn>j0RAVOo>i0RAVOP?h0PAXOQ?h0o@WOR?h0n@XOS?g0n@XOS?g0m@YOT?g0m@WOT?h0QASOP?l0UAoNl>P1e1O100O1O1O1O100O1O1O1@WE"}, "label": "the large adult elephant"}]}
{"id": 152238, "image": "COCO_train2014_000000152238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant\"."}], "task": "refering", "answer_template": "The \"a baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 396, 399, 400, 401, 402, 403, 404, 405, 418, 421, 422, 423, 424, 425, 426, 440, 443, 444, 445, 446, 447, 448, 468, 469, 470], "bbox": [0.004493464052287581, 0.5209803921568628, 0.4561764705882353, 0.9883823529411765], "iscrowd": 0, "area": 53694.21265, "rle": {"size": [612, 612], "counts": "[W2c0]b0m0SOl0UOl0SOl0TOd0\\O01O000O2O00001N1000001O3M3N2M4M2M3N2N3L3N2N2M4M2M3N30O2N1O100O1O100O1O1O101N1O100O000010O01O00010O00010O000010O0001O01O01O00010O001O01O01O00010O1O4L4M3L4L4L4M7H=C=C<E<C<E<C=D8H000O1000000O1000000O1000001O0O1000000O1000000O1000000O10000000001O0000000000000000000000000000000000001O00N1N3M3N2M3M3M3M3L4M3L4M3L4M3L4M3L4M3L4L4M3L4M3N21O1O1O1N2O1O1[EUJo7l5aGcJ_8^5RGQKm8P5cF`K\\9a4TFnKl9S4dE]L[:V6000000000001O0000000000001O0000000O10001O000000000000001O0000000000001O01O2N2M2O2N2N2N2N1O2N2N2M3N1O3M2N3M2N2N3M2M4M2N3WOh0B>B>B=D<K5L4K5K5K5L3L5K5L4K5K5L4K5K5L4K5J6I6J7I7J6I7I7I7J6I7I7IYdV6"}, "label": "a baby elephant"}]}
{"id": 152238, "image": "COCO_train2014_000000152238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant next to adult elephant\"."}], "task": "refering", "answer_template": "The \"baby elephant next to adult elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 396, 399, 400, 401, 402, 403, 404, 405, 418, 421, 422, 423, 424, 425, 426, 440, 443, 444, 445, 446, 447, 448, 468, 469, 470], "bbox": [0.004493464052287581, 0.5209803921568628, 0.4561764705882353, 0.9883823529411765], "iscrowd": 0, "area": 53694.21265, "rle": {"size": [612, 612], "counts": "[W2c0]b0m0SOl0UOl0SOl0TOd0\\O01O000O2O00001N1000001O3M3N2M4M2M3N2N3L3N2N2M4M2M3N30O2N1O100O1O100O1O1O101N1O100O000010O01O00010O00010O000010O0001O01O01O00010O001O01O01O00010O1O4L4M3L4L4L4M7H=C=C<E<C<E<C=D8H000O1000000O1000000O1000001O0O1000000O1000000O1000000O10000000001O0000000000000000000000000000000000001O00N1N3M3N2M3M3M3M3L4M3L4M3L4M3L4M3L4M3L4L4M3L4M3N21O1O1O1N2O1O1[EUJo7l5aGcJ_8^5RGQKm8P5cF`K\\9a4TFnKl9S4dE]L[:V6000000000001O0000000000001O0000000O10001O000000000000001O0000000000001O01O2N2M2O2N2N2N2N1O2N2N2M3N1O3M2N3M2N2N3M2M4M2N3WOh0B>B>B=D<K5L4K5K5K5L3L5K5L4K5K5L4K5K5L4K5J6I6J7I7J6I7I7I7J6I7I7IYdV6"}, "label": "baby elephant next to adult elephant"}]}
{"id": 168622, "image": "COCO_train2014_000000168622.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a van parked in front of two trees\"."}], "task": "refering", "answer_template": "The \"a van parked in front of two trees\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305], "bbox": [0.04553125, 0.5264583333333334, 0.39681249999999996, 0.8215], "iscrowd": 0, "area": 22973.297800000008, "rle": {"size": [480, 640], "counts": "Wm=2i01V=7dB1Q=8iBMP=;iBIR=W1J6J5M2M2N3N1N3N2M2N2O1N2N2O1N2N2HPM_DQ3`;QM]DQ3b;700O103M2M3N1O1N2O0O2N2N100O10000O1000O10O100000O010O100O00100O010O1N2N1O20000000O10000000000000000O010000O1000000O01000O1000000000O1000O1000000000O1000O10000000000O100000O100000O1001O0000000O101O00O100000000O0100000O10000O1000O103M3M3M2N1O001O001O1O00000000000001O0O101N2O0O100O2O0O1N2M3]OhDWM_;h2>O0O1O101N1O2O1N101N101N2O001N101N2O1O1DcCRN_<i1hCTNY<j1`0O1O1O1O100O101O00M4M2O1O00001O01O1N3N4L3M1O1O2N1O1O2N101M2O2N1N3@f0DZdd5"}, "label": "a van parked in front of two trees"}]}
{"id": 168622, "image": "COCO_train2014_000000168622.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"th silver vehicle on the right\"."}], "task": "refering", "answer_template": "The \"th silver vehicle on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305], "bbox": [0.04553125, 0.5264583333333334, 0.39681249999999996, 0.8215], "iscrowd": 0, "area": 22973.297800000008, "rle": {"size": [480, 640], "counts": "Wm=2i01V=7dB1Q=8iBMP=;iBIR=W1J6J5M2M2N3N1N3N2M2N2O1N2N2O1N2N2HPM_DQ3`;QM]DQ3b;700O103M2M3N1O1N2O0O2N2N100O10000O1000O10O100000O010O100O00100O010O1N2N1O20000000O10000000000000000O010000O1000000O01000O1000000000O1000O1000000000O1000O10000000000O100000O100000O1001O0000000O101O00O100000000O0100000O10000O1000O103M3M3M2N1O001O001O1O00000000000001O0O101N2O0O100O2O0O1N2M3]OhDWM_;h2>O0O1O101N1O2O1N101N101N2O001N101N2O1O1DcCRN_<i1hCTNY<j1`0O1O1O1O100O101O00M4M2O1O00001O01O1N3N4L3M1O1O2N1O1O2N101M2O2N1N3@f0DZdd5"}, "label": "th silver vehicle on the right"}]}
{"id": 348850, "image": "COCO_train2014_000000348850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big giraffe in the distance\"."}], "task": "refering", "answer_template": "The \"the big giraffe in the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 61, 62, 84, 85, 86, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 180, 201, 203, 224, 225, 226], "bbox": [0.625453125, 0.03336673346693386, 0.8714999999999999, 0.5472344689378758], "iscrowd": 0, "area": 11144.635650000006, "rle": {"size": [499, 640], "counts": "_nR61b?2N001O001O00000DLXA5g>MVA5i>NSA=d>;0O1000N1O2N2M2O2M3L4M3NgNnAS1Y>01oNkAb0T>\\OnAe0Q>ZOPBf0o=ZOSBf0j=[OWBf0g=YO[Bg0d=YO]Bh0a=XO`Bi0_=TOdBl0[=SOgBn0o=1O1O004L001O00001O001N10001O001O000000000000001iBdNT<\\1cCnN\\<R1^CUOa<n0XCVOh<m0RCWOm<h103TC]Ma<l2O1O1O3PFnLX7Y3^HiLc7\\3VHfLj7m3]GXLc8l3kFcLU9_3^FkLc9a41O2K4N3N1WO\\FbKf9o3V1MnERLf8o3VGSLk8n3QGTLo8^4\\FfKd9S51O2N1O2N1VMUF^N;Ec9l1`FnM8L[9T2h2O1O1O1O1O1O2O0O1O1O1O1O1O1O010O1O1O0O2O1O1O00001O0O100100hD`N[8a1dG_N\\8b1bG_N^8b1^GbNa8_1[GdNf8\\1UGhNk8Z1oFjNQ9Y1hFkNX9X1`FhN_N[OQ;P2XFiNjNUOn:\\2iEcN_:l1lDXNV;m26I6K6I7J5KlDoLQ:l2QFVMP:f2QF\\Mo9b2PFaMQ:j2^E[Mc:j3002N1O2IcDTL^;W1jDm0_;mNeDPOBe0i0GT;9iD^O[O=R1Cl:T1RDQO03\\=Z1eB^Na=h0eBDc=0kBOT=DYC9R>DnhW1"}, "label": "the big giraffe in the distance"}]}
{"id": 348850, "image": "COCO_train2014_000000348850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mature giraffe\"."}], "task": "refering", "answer_template": "The \"a mature giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 61, 62, 84, 85, 86, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 180, 201, 203, 224, 225, 226], "bbox": [0.625453125, 0.03336673346693386, 0.8714999999999999, 0.5472344689378758], "iscrowd": 0, "area": 11144.635650000006, "rle": {"size": [499, 640], "counts": "_nR61b?2N001O001O00000DLXA5g>MVA5i>NSA=d>;0O1000N1O2N2M2O2M3L4M3NgNnAS1Y>01oNkAb0T>\\OnAe0Q>ZOPBf0o=ZOSBf0j=[OWBf0g=YO[Bg0d=YO]Bh0a=XO`Bi0_=TOdBl0[=SOgBn0o=1O1O004L001O00001O001N10001O001O000000000000001iBdNT<\\1cCnN\\<R1^CUOa<n0XCVOh<m0RCWOm<h103TC]Ma<l2O1O1O3PFnLX7Y3^HiLc7\\3VHfLj7m3]GXLc8l3kFcLU9_3^FkLc9a41O2K4N3N1WO\\FbKf9o3V1MnERLf8o3VGSLk8n3QGTLo8^4\\FfKd9S51O2N1O2N1VMUF^N;Ec9l1`FnM8L[9T2h2O1O1O1O1O1O2O0O1O1O1O1O1O1O010O1O1O0O2O1O1O00001O0O100100hD`N[8a1dG_N\\8b1bG_N^8b1^GbNa8_1[GdNf8\\1UGhNk8Z1oFjNQ9Y1hFkNX9X1`FhN_N[OQ;P2XFiNjNUOn:\\2iEcN_:l1lDXNV;m26I6K6I7J5KlDoLQ:l2QFVMP:f2QF\\Mo9b2PFaMQ:j2^E[Mc:j3002N1O2IcDTL^;W1jDm0_;mNeDPOBe0i0GT;9iD^O[O=R1Cl:T1RDQO03\\=Z1eB^Na=h0eBDc=0kBOT=DYC9R>DnhW1"}, "label": "a mature giraffe"}]}
{"id": 62131, "image": "COCO_train2014_000000062131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair in the front\"."}], "task": "refering", "answer_template": "The \"the chair in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357], "bbox": [0.003234375, 0.6877874186550976, 0.564921875, 0.9821691973969631], "iscrowd": 0, "area": 43637.20275, "rle": {"size": [461, 640], "counts": "RW19]>e0ROn0ROn0ROe0[O000O10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O100000001O0000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000001N10000000000000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000000000000000O10000000000000000000001O00000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O1000000000001O000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O10000000000000000000000000000000000O101O0000000000000000000000000000000O100000N2ZOf0ZOf0ZOf0ZOf0ZOf0ZOgkl3"}, "label": "the chair in the front"}]}
{"id": 62131, "image": "COCO_train2014_000000062131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wicker chair with a bird perched on it\"."}], "task": "refering", "answer_template": "The \"a wicker chair with a bird perched on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 266, 267, 289, 290, 312, 313, 335, 336, 358, 359], "bbox": [0.005234375, 0.5707158351409979, 0.6466406250000001, 0.9986334056399132], "iscrowd": 0, "area": 23213.284050000002, "rle": {"size": [461, 640], "counts": "Ue11R>:E;E;M300000000000000000000O1000000000000000000000000000000O10000000000000000000000000O10O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O100000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000O10O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000O100000O100000000000000O1000000000000O1000000000000O1000000000000O10000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000iF^Oc4b0oIl0Q6TO_H\\2a7X2000O1000000000000O10000000000O10000oIfGo5\\801O001N101O001O001N101O001O001N1I7YOh0WOh0YOh0XOg0XOi0XOg0YOm_U3"}, "label": "a wicker chair with a bird perched on it"}]}
{"id": 62131, "image": "COCO_train2014_000000062131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair in the back\"."}], "task": "refering", "answer_template": "The \"the chair in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 36, 45, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 167, 168, 169, 170, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 190, 191, 192, 193, 194, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 226, 227, 228, 229, 275, 297, 298], "bbox": [0.28253125, 0.09913232104121476, 1.0, 0.7887852494577007], "iscrowd": 0, "area": 86220.06044999998, "rle": {"size": [461, 640], "counts": "_aa2:c0T1[;]1B?C<C>BU2lMO010O010O00010O010O0010O0010O1O2O100O000000000000000000O2O000000000000000O10001O000000000O10000000001OcNoG^LQ8Z3VHgLj7W3XHiLh7U3YHlLg7S3ZHmLf7Q3[HPMe7o2\\HQMd7m2]HTMc7k2^HUMb7i2_HXMa7f2aHZM_7d2bH]M^7a2dH_M\\7_2eHcMZ7\\2gHdMY7\\2fHeMZ7Z2gHfMY7Z2gHfMZ7X2hHgMX7Y2iHfMW7Y2jHgMV7Y2kHfMU7Y2lHgMT7Y2mHfMS7Y2oHfMQ7Z2oHfMQ7Y2QIgMn6Y2RIgMn6X2TIgMl6Y2UIfMk6Y2VIgMj6Y2YIdMg6]2_I]M`6c2gIVMY6j2lIQMT6P3lIoLT6Q3mInLS6S3mIlLS6T3nIkLR6U3oIkLP6V3QJhLo5X3RJgLn5Z3RJeLn5[3SJdLm5]3SJbLm5^3TJbLl5]3UJbLk5_3VJ_Lj5a3WJ^Li5c3ZJYLf5g3^JULb5k3`21O000010O0001O001O00001O01O01O00001O000010O01O00001O001O00100O001O1O001O1O010O1O001O1O001O1O010O1O001O1O1O1O100O001O1O1O1O1O1O100O001O1O1O1O1O1O1O002N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O0O1O00001O00001O00001O0000001O00001O000000000000001O000001O000001O0000000000001O0000000000001O000000000000001O000000000000001O00000000000000000000000000001O00000000000000000000000000010O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000001O0000000001O00000001O00000O1000000000001O0000000000000000001O000000000O100000001O00000000000000001O0000000000000mJaIX2`6fMaIY2`6fMaIZ2_6eMbI[2^6eMcIZ2^6eMbI[2^6dMcI\\2]6dMcI\\2]6cMeI\\2\\6cMdI]2\\6cMdI]2\\6bMfI]2Z6cMfI]2[6bMeI^2[6aMfI_2Z6aMgI^2Y6aMhI_2Y6`MgI`2Y6`MgI`2Y6_MiI_2X6aMhI_2Y6`MgI`2Y6_MhIa2X6_MiI`2W6`MiI`2X6^MiIb2W6^MiIb2W6]MkIb2V6]MiId2W6\\MiId2W6[MjIe2V6[MjIe2W6ZMiIf2W6YMjIg2V6YMjIg2V6XMkIg2W6XMiIh2W6WMjIi2V6WMjIi2V6VMkIi2W6VMiIj2W6UMjIk2V6UMiIl2W6SMjIl2X6SMhIm2X6SMhIm2X6RMiIm2X6SMhIm2Y6QMhIo2X6QMhIo2X6PMiIo2X6QMhIo2Y6oLhIQ3X6oLhIQ3X6oLdIT3^6lL[IZ3g0WL_3`0]Ke38cMi0RO@d3"}, "label": "the chair in the back"}]}
{"id": 398017, "image": "COCO_train2014_000000398017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man attcks the camera with his skateboard\"."}], "task": "refering", "answer_template": "The \"a man attcks the camera with his skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 105, 106, 107, 124, 125, 143, 154, 176, 177, 178, 193, 194, 195, 196, 212, 213, 215, 227, 228, 229], "bbox": [0.55574, 0.00112, 1.0, 0.9898666666666667], "iscrowd": 0, "area": 27485.308299999993, "rle": {"size": [375, 500], "counts": "hWV32e;0O1O1^I3Z1NaN9[1HeN8[1HeN8Z1IgN6Y1KfN5Y1LgN4X1MhN3X1NhN0X11hNOX11hNNX13iNKX16gNH_OK_M>R3EA0YM<V3BB5VM9h75iGMV86gGJY89cGH\\8<aGE^8>^GCb8?[GBe8a0XG@f8d0VG]Ok8d0SG]Ol8d0RG]On8e0PG\\Oo8e0oF\\OQ9f0lF[OU9d1O001O001O00001O001O001O00001O0O2O001O000oNgMdHZ2[7gMdHY2\\7hMcHY2]7gMbHY2^7gMbHZ2^7fMaHZ2_7gM`HY2a7gM^HZ2a7fM_HZ2a7gM^HZ2b7fM]HZ2c7fM]H[2c7eM\\H[2d7fM[H[2e7eMZH[2f7eMZH\\2f7cMZH]2f7cMZH^2e7bM[H^2f7aMZH`2e7`M[H`2f7_MZHb2e7^M[Hb2f7]MZHd2e7\\M[Hd2e7\\M[Hd2f7[MZHf2e7YM\\Hg2e7XM[Hi2d7WM\\Hi2e7VM[Hk2d7UM\\Hk2d7UM\\Hl2d7SM\\Hm2d7SM\\Hn2d7QM\\Ho2d7QM\\HP3d7oL\\HQ3d7oL\\HR3d7mL\\HS3d7mL\\HT3d5jLmK2_NU3c5kLmKO`NV3c5lLlKNaNW3b5lLmKKbNZ3`5lLmKJbN\\3a5jLmKIbN]3a5lLkKGdN^3`5lLkKFeN_3_5lLlKDeNa3^5kLmKDeNa3^5kLnKCdNc3]5kLnKBeNd3\\5jLPLAdNf3[5jLPL@eNf3Z5kLRL^OdNh3Y5jLSL^OdNi3W5kLUL[OdNk3U5kLWLZOdNk3U5kL^LV3`3lL_LU3`3kL`LV3_3kL`LU3`3kL`LV3^3kLbLV3]3kLbLV3]3kLbLU3^3mL`LT3_3mL`LS3`3nL_LS3h2QL[Ll0l0T3h2QL[Ll0l0S3i2RL[Lk0k0T3i2QL\\Lk0k0T3i2RL\\Li0k0V3h2QL]Li0k0W3g2PL_Li0i0W3h2PL_Li0i0X3g2oKaLh0h0Y3g2oKaLi0g0Y3g2nKcLh0f0[3e2nKeLg0f0[3e2nKeLg0f0\\3d2mKfLh0e0\\3d2lKgLh0f0[3c2mKhLg0f0\\3`2nKjLg0e0\\OiN]3h3`LjLg0e0ROSOh3\\3`LlLf0W1k2k1`LnLe0X1j2f1eLRMb0W1j2g1cLSMb0V1l2f1bLTMb0V1l2f1bLTMb0V1m2e1aLUMc0V1l2d1aLVMb0W1m2c1aLWM?X1Q3a1`LVM>Z1S3_1_LWM=\\1T3]1^LWM=]1U3\\1^LWM;_1X3Z1\\LWM;`1Z3X1[LYM9a1\\3V1ZLYM9b1]3V1YLXM8d1`3R1YLZM7d1a3Q1WL\\M7d1c3U2\\LlMc3T2\\LmMe3R2ZLoMg3Q2XLoMi3P2VLRNi3o1VLQNk3n1TLSNm3m1QLTNP4k1PLVNo3j1PLWNQ4i1nKWNS4h1lKYNU4g1iKZNX4e1gK]NY4b1eK`N\\4]400100O1O1O1PLeKb0\\4]OdKd0\\4ZOfKe0Z4ZOhKe0Y4ZOhKe0Y4ZOgKg0Y4XOhKg0Y4YOfKg0[4XOcKk0]4TO`Ko0a4PO]KR1d4nNZKT1e4lNZKU1g4jNYKW1g4l21O100O101N1O2WMSKeNm4W1VKiNl4T1UKmNk4R1UKnNm4P1SKPOn4o0RKROo4l0QKTOP5k0PKVOQ5h0oJXOR5g0nJYOT5e0lJ\\OT5d0kJ\\OW5b0hJ@X5_320^Oa0mNS1nNR1mNT1mN@"}, "label": "a man attcks the camera with his skateboard"}]}
{"id": 398017, "image": "COCO_train2014_000000398017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the face and torso of the boy with skateboard\"."}], "task": "refering", "answer_template": "The \"the face and torso of the boy with skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 105, 106, 107, 124, 125, 143, 154, 176, 177, 178, 193, 194, 195, 196, 212, 213, 215, 227, 228, 229], "bbox": [0.55574, 0.00112, 1.0, 0.9898666666666667], "iscrowd": 0, "area": 27485.308299999993, "rle": {"size": [375, 500], "counts": "hWV32e;0O1O1^I3Z1NaN9[1HeN8[1HeN8Z1IgN6Y1KfN5Y1LgN4X1MhN3X1NhN0X11hNOX11hNNX13iNKX16gNH_OK_M>R3EA0YM<V3BB5VM9h75iGMV86gGJY89cGH\\8<aGE^8>^GCb8?[GBe8a0XG@f8d0VG]Ok8d0SG]Ol8d0RG]On8e0PG\\Oo8e0oF\\OQ9f0lF[OU9d1O001O001O00001O001O001O00001O0O2O001O000oNgMdHZ2[7gMdHY2\\7hMcHY2]7gMbHY2^7gMbHZ2^7fMaHZ2_7gM`HY2a7gM^HZ2a7fM_HZ2a7gM^HZ2b7fM]HZ2c7fM]H[2c7eM\\H[2d7fM[H[2e7eMZH[2f7eMZH\\2f7cMZH]2f7cMZH^2e7bM[H^2f7aMZH`2e7`M[H`2f7_MZHb2e7^M[Hb2f7]MZHd2e7\\M[Hd2e7\\M[Hd2f7[MZHf2e7YM\\Hg2e7XM[Hi2d7WM\\Hi2e7VM[Hk2d7UM\\Hk2d7UM\\Hl2d7SM\\Hm2d7SM\\Hn2d7QM\\Ho2d7QM\\HP3d7oL\\HQ3d7oL\\HR3d7mL\\HS3d7mL\\HT3d5jLmK2_NU3c5kLmKO`NV3c5lLlKNaNW3b5lLmKKbNZ3`5lLmKJbN\\3a5jLmKIbN]3a5lLkKGdN^3`5lLkKFeN_3_5lLlKDeNa3^5kLmKDeNa3^5kLnKCdNc3]5kLnKBeNd3\\5jLPLAdNf3[5jLPL@eNf3Z5kLRL^OdNh3Y5jLSL^OdNi3W5kLUL[OdNk3U5kLWLZOdNk3U5kL^LV3`3lL_LU3`3kL`LV3_3kL`LU3`3kL`LV3^3kLbLV3]3kLbLV3]3kLbLU3^3mL`LT3_3mL`LS3`3nL_LS3h2QL[Ll0l0T3h2QL[Ll0l0S3i2RL[Lk0k0T3i2QL\\Lk0k0T3i2RL\\Li0k0V3h2QL]Li0k0W3g2PL_Li0i0W3h2PL_Li0i0X3g2oKaLh0h0Y3g2oKaLi0g0Y3g2nKcLh0f0[3e2nKeLg0f0[3e2nKeLg0f0\\3d2mKfLh0e0\\3d2lKgLh0f0[3c2mKhLg0f0\\3`2nKjLg0e0\\OiN]3h3`LjLg0e0ROSOh3\\3`LlLf0W1k2k1`LnLe0X1j2f1eLRMb0W1j2g1cLSMb0V1l2f1bLTMb0V1l2f1bLTMb0V1m2e1aLUMc0V1l2d1aLVMb0W1m2c1aLWM?X1Q3a1`LVM>Z1S3_1_LWM=\\1T3]1^LWM=]1U3\\1^LWM;_1X3Z1\\LWM;`1Z3X1[LYM9a1\\3V1ZLYM9b1]3V1YLXM8d1`3R1YLZM7d1a3Q1WL\\M7d1c3U2\\LlMc3T2\\LmMe3R2ZLoMg3Q2XLoMi3P2VLRNi3o1VLQNk3n1TLSNm3m1QLTNP4k1PLVNo3j1PLWNQ4i1nKWNS4h1lKYNU4g1iKZNX4e1gK]NY4b1eK`N\\4]400100O1O1O1PLeKb0\\4]OdKd0\\4ZOfKe0Z4ZOhKe0Y4ZOhKe0Y4ZOgKg0Y4XOhKg0Y4YOfKg0[4XOcKk0]4TO`Ko0a4PO]KR1d4nNZKT1e4lNZKU1g4jNYKW1g4l21O100O101N1O2WMSKeNm4W1VKiNl4T1UKmNk4R1UKnNm4P1SKPOn4o0RKROo4l0QKTOP5k0PKVOQ5h0oJXOR5g0nJYOT5e0lJ\\OT5d0kJ\\OW5b0hJ@X5_320^Oa0mNS1nNR1mNT1mN@"}, "label": "the face and torso of the boy with skateboard"}]}
{"id": 168643, "image": "COCO_train2014_000000168643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"floral three seater couch to the left of the picture\"."}], "task": "refering", "answer_template": "The \"floral three seater couch to the left of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 280, 281, 282, 283, 284, 285, 304, 305, 306, 328], "bbox": [0.047031250000000004, 0.5462441314553991, 0.447546875, 0.9530985915492958], "iscrowd": 0, "area": 25059.776799999996, "rle": {"size": [426, 640], "counts": "nh<3V=9G9G9G9G9G2N001O001O00001O001O001O001O00001O0010O010O0010O010O0010O0101N100O101N100O2N100O2O0O101N100O2O0O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O10000O10000000000O100000000O100000000O010000000O>Cc0]O3M1O1N2O2N1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O2N001N10001O000O2O001O00001N101O00001N101O00O10O01O1O100O001O100O1O00100O1O1O10O01O1O101N2N3N1N2N3M2O1N3M2N2O1N3M2N2O2M2N2N3N1N2N2N3N1N1O100O2N1O100O1O1O2O0O1O1O100O2N1O100O1O1O2N1O1N2O1O1N3N1O1N2O1O1O2M2O1O1N4M2N3L4M2N3L4M2N3L4M2N3M3L3N3M3L3N3M3L3N3MTob4"}, "label": "floral three seater couch to the left of the picture"}]}
{"id": 168643, "image": "COCO_train2014_000000168643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa with floral print\"."}], "task": "refering", "answer_template": "The \"a sofa with floral print\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 280, 281, 282, 283, 284, 285, 304, 305, 306, 328], "bbox": [0.047031250000000004, 0.5462441314553991, 0.447546875, 0.9530985915492958], "iscrowd": 0, "area": 25059.776799999996, "rle": {"size": [426, 640], "counts": "nh<3V=9G9G9G9G9G2N001O001O00001O001O001O001O00001O0010O010O0010O010O0010O0101N100O101N100O2N100O2O0O101N100O2O0O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O10000O10000000000O100000000O100000000O010000000O>Cc0]O3M1O1N2O2N1O1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O2N001N10001O000O2O001O00001N101O00001N101O00O10O01O1O100O001O100O1O00100O1O1O10O01O1O101N2N3N1N2N3M2O1N3M2N2O1N3M2N2O2M2N2N3N1N2N2N3N1N1O100O2N1O100O1O1O2O0O1O1O100O2N1O100O1O1O2N1O1N2O1O1N3N1O1N2O1O1O2M2O1O1N4M2N3L4M2N3L4M2N3L4M2N3M3L3N3M3L3N3M3L3N3MTob4"}, "label": "a sofa with floral print"}]}
{"id": 168643, "image": "COCO_train2014_000000168643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is a brown wicker arm chair with floral cushions sitting underneath a hanging floral arrangement\"."}], "task": "refering", "answer_template": "The \"it is a brown wicker arm chair with floral cushions sitting underneath a hanging floral arrangement\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343], "bbox": [0.705171875, 0.6658450704225352, 0.9902031250000001, 0.9826760563380281], "iscrowd": 0, "area": 16527.82145, "rle": {"size": [426, 640], "counts": "dPl51U=5I7I8H7I7I7I8H7N2N2O000000000000000O1000000000O100000O10000000000000000O10000CRN^En1^:XN`Eh1\\:^NbEb1^:_NaEa1_:aN_E_1a:cN]E]1c:e000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000O1N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O100000000000000000000000000000000000000000000000000000000000000000000000000000O1L4L4K6K4L5K4L5K4K6K4L4L5K4K6K4L5K4L4K6K4L5K4L5J5L>B`V2"}, "label": "it is a brown wicker arm chair with floral cushions sitting underneath a hanging floral arrangement"}]}
{"id": 168643, "image": "COCO_train2014_000000168643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a single sofa in the right corner of the room\"."}], "task": "refering", "answer_template": "The \"a single sofa in the right corner of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 251, 252, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343], "bbox": [0.705171875, 0.6658450704225352, 0.9902031250000001, 0.9826760563380281], "iscrowd": 0, "area": 16527.82145, "rle": {"size": [426, 640], "counts": "dPl51U=5I7I8H7I7I7I8H7N2N2O000000000000000O1000000000O100000O10000000000000000O10000CRN^En1^:XN`Eh1\\:^NbEb1^:_NaEa1_:aN_E_1a:cN]E]1c:e000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000O1N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O100000000000000000000000000000000000000000000000000000000000000000000000000000O1L4L4K6K4L5K4L5K4K6K4L4L5K4K6K4L5K4L4K6K4L5K4L5J5L>B`V2"}, "label": "a single sofa in the right corner of the room"}]}
{"id": 291526, "image": "COCO_train2014_000000291526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing glasses keeping right hand on the lips and watching other person\"."}], "task": "refering", "answer_template": "The \"a person wearing glasses keeping right hand on the lips and watching other person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.45740625, 0.21235416666666668, 1.0, 1.0], "iscrowd": 0, "area": 85428.82065, "rle": {"size": [480, 640], "counts": "ehY46e>6J6I6K5J6K5J7J5J6K5J7M2N2O0O2N2O1N2N2O0O2N2O1N2N1O2O1N2N2O1N101N2O1O1N101O1N2O001N101O0O2O0O2N1O2O1N1O2N1O2O0O2N1O2N1O2N101N2O001N101O0O2O0O2N101N1O2N101N1O2N1O2O0O2N101O001O00001O001O001WHjJh5U5YJkJf5V5YJkJg5T5ZJlJf5T5YJmJf5S5[JmJe5S5ZJnJe5R5\\JnJd5Q5\\JPKc5Q5]JoJc5P5^JPKb5P5^JPKa5P5_JQKa5o4_JQK`5o4aJQK_5o4aJQK^5o4cJQK]5o4bJRK^5m4cJSK]5m4cJSK\\5m4eJSK[5l4eJUK[5k4eJUK[5j4eJWK[5i4eJWK[5h4eJYK[5g4eJXK\\5g4dJZK\\5f4cJ[K]5d4cJ]K]5c4aJ_K_5b4ZIeKCLR7g4PIlKA_O_7m4cH]L]7c500000000O10000O100O1oNdH^J\\7a5gH]JY7b5iH\\JX7c5jH\\JV7c5lH[JU7d5mH[JS7e5nHYJS7f5oHYJQ7f5QIXJP7g5SIWJm6h5UIVJl6i5VIVJj6i5XIUJi6k5YIRJh6m5ZIQJg6n5V1WO_GRKb8^4X1M3M3N2M3N2N2M3N2N2N2N2M3N2N20000O0100000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000O10O10000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00003M2N3M2N3M2M4M2dGhKl5[4nIiKQ6Y4jIkKU6X4eIkK[6W4`ImK_6V4[InKd6T4WIPLh6S4RIPLn6R4mHRLR7o5O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N00000000000000001O000000001O000000001O000000001O0O1000001O000000001O000000001O00VH"}, "label": "a person wearing glasses keeping right hand on the lips and watching other person"}]}
{"id": 291526, "image": "COCO_train2014_000000291526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the glasses on the right\"."}], "task": "refering", "answer_template": "The \"the man with the glasses on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.45740625, 0.21235416666666668, 1.0, 1.0], "iscrowd": 0, "area": 85428.82065, "rle": {"size": [480, 640], "counts": "ehY46e>6J6I6K5J6K5J7J5J6K5J7M2N2O0O2N2O1N2N2O0O2N2O1N2N1O2O1N2N2O1N101N2O1O1N101O1N2O001N101O0O2O0O2N1O2O1N1O2N1O2O0O2N1O2N1O2N101N2O001N101O0O2O0O2N101N1O2N101N1O2N1O2O0O2N101O001O00001O001O001WHjJh5U5YJkJf5V5YJkJg5T5ZJlJf5T5YJmJf5S5[JmJe5S5ZJnJe5R5\\JnJd5Q5\\JPKc5Q5]JoJc5P5^JPKb5P5^JPKa5P5_JQKa5o4_JQK`5o4aJQK_5o4aJQK^5o4cJQK]5o4bJRK^5m4cJSK]5m4cJSK\\5m4eJSK[5l4eJUK[5k4eJUK[5j4eJWK[5i4eJWK[5h4eJYK[5g4eJXK\\5g4dJZK\\5f4cJ[K]5d4cJ]K]5c4aJ_K_5b4ZIeKCLR7g4PIlKA_O_7m4cH]L]7c500000000O10000O100O1oNdH^J\\7a5gH]JY7b5iH\\JX7c5jH\\JV7c5lH[JU7d5mH[JS7e5nHYJS7f5oHYJQ7f5QIXJP7g5SIWJm6h5UIVJl6i5VIVJj6i5XIUJi6k5YIRJh6m5ZIQJg6n5V1WO_GRKb8^4X1M3M3N2M3N2N2M3N2N2N2N2M3N2N20000O0100000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000O10O10000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00003M2N3M2N3M2M4M2dGhKl5[4nIiKQ6Y4jIkKU6X4eIkK[6W4`ImK_6V4[InKd6T4WIPLh6S4RIPLn6R4mHRLR7o5O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N00000000000000001O000000001O000000001O000000001O0O1000001O000000001O000000001O00VH"}, "label": "the man with the glasses on the right"}]}
{"id": 291526, "image": "COCO_train2014_000000291526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a pullover talks to a man wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man in a pullover talks to a man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0033750000000000004, 0.2674166666666667, 0.535953125, 0.9887708333333334], "iscrowd": 0, "area": 74946.3166, "rle": {"size": [480, 640], "counts": "]Y1Q1l=`2aM2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O10000O10000O10000O100iMcIjK^6S4hIjKX6T4lIjKT6T4PJiKQ6T4UJiKk5U4YJiKg5T4_JhKb5V4bJhK^5V4gJgKY5V4lJgKU5W4PKfKP5X4TKfKl4W4ZKeKg4Y4]KeKc4Y4bKcK_4Z4fKdKZ4Z4kKcKU4[4oKbKR4[4SLcKm3[4VLdKj3Z4YLdKh3[4ZLdKf3[4]LcKc3]4^LaKc3_4^L`Kb3_4`L`K`3`4bL]K_3c4bL\\K^3c4dL[K]3e4eLYK[3g4_300O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O11O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001eHlKh3T4mKYLQ4h3cKcL]4^3WKmLf4V3PKTMm4P3gJ[MV5h2eJ]MX5g2fJ[MV5h2jJXMR5m2lJTMR5o2mJQMS5Q3jJPMV5S3gJmLX5V3eJlLZ5W3cJiL]5Z3_JgLa5[3\\JfLc5^3ZJbLf5\\601O000O10001O0O1000000O2O000O1O100O2N1O1O1O1O101N1O1O1O1O2O0O1O1O1O1O2N100O1O1O1O2N1N2M3M3M3M4L3M3M3M4L5J5M4L3L4M4L3L4M4L3L5L3M3L5L3M4K4M3M4L3M4L3M3M4L3M3M4L3M4L3M3M4LnlZ4"}, "label": "a man in a pullover talks to a man wearing glasses"}]}
{"id": 291526, "image": "COCO_train2014_000000291526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a gray zip - up jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing a gray zip - up jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0033750000000000004, 0.2674166666666667, 0.535953125, 0.9887708333333334], "iscrowd": 0, "area": 74946.3166, "rle": {"size": [480, 640], "counts": "]Y1Q1l=`2aM2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O10000O10000O10000O100iMcIjK^6S4hIjKX6T4lIjKT6T4PJiKQ6T4UJiKk5U4YJiKg5T4_JhKb5V4bJhK^5V4gJgKY5V4lJgKU5W4PKfKP5X4TKfKl4W4ZKeKg4Y4]KeKc4Y4bKcK_4Z4fKdKZ4Z4kKcKU4[4oKbKR4[4SLcKm3[4VLdKj3Z4YLdKh3[4ZLdKf3[4]LcKc3]4^LaKc3_4^L`Kb3_4`L`K`3`4bL]K_3c4bL\\K^3c4dL[K]3e4eLYK[3g4_300O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O11O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001eHlKh3T4mKYLQ4h3cKcL]4^3WKmLf4V3PKTMm4P3gJ[MV5h2eJ]MX5g2fJ[MV5h2jJXMR5m2lJTMR5o2mJQMS5Q3jJPMV5S3gJmLX5V3eJlLZ5W3cJiL]5Z3_JgLa5[3\\JfLc5^3ZJbLf5\\601O000O10001O0O1000000O2O000O1O100O2N1O1O1O1O101N1O1O1O1O2O0O1O1O1O1O2N100O1O1O1O2N1N2M3M3M3M4L3M3M3M4L5J5M4L3L4M4L3L4M4L3L5L3M3L5L3M4K4M3M4L3M4L3M3M4L3M3M4L3M4L3M3M4LnlZ4"}, "label": "a man wearing a gray zip - up jacket"}]}
{"id": 250569, "image": "COCO_train2014_000000250569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green suit jacket\"."}], "task": "refering", "answer_template": "The \"a woman in a green suit jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262], "bbox": [0.17706249999999998, 0.27949999999999997, 0.416640625, 0.70675], "iscrowd": 0, "area": 21294.39475, "rle": {"size": [480, 640], "counts": "iXe11j>6J6J6J6J6J6J6J6J6J6J6J6L4L4L4L4L4L4L4L3M4L4L4M200O010O2O000O100O101N10000O100O2O000O100O101N10000O100O1O1O1O0M4I7I7I7I7H8M2O2N2N2O1N2NM3002O1O1O1O100O1O1O1O1O1O10000000000O1000000001O00001O001O001O00001O001O001O001O00010O001O00103L3M4M3L4L3M4L4Lm0SON2N2N2N23M2M3N2N3M1O1O001O000O2O0O2N1O2M2O1O2N3UOl0QOP1C<D<D<Eoc^5"}, "label": "a woman in a green suit jacket"}]}
{"id": 250569, "image": "COCO_train2014_000000250569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"news reporter with blonde hair and green jacket\"."}], "task": "refering", "answer_template": "The \"news reporter with blonde hair and green jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262], "bbox": [0.17706249999999998, 0.27949999999999997, 0.416640625, 0.70675], "iscrowd": 0, "area": 21294.39475, "rle": {"size": [480, 640], "counts": "iXe11j>6J6J6J6J6J6J6J6J6J6J6J6L4L4L4L4L4L4L4L3M4L4L4M200O010O2O000O100O101N10000O100O2O000O100O101N10000O100O1O1O1O0M4I7I7I7I7H8M2O2N2N2O1N2NM3002O1O1O1O100O1O1O1O1O1O10000000000O1000000001O00001O001O001O00001O001O001O001O00010O001O00103L3M4M3L4L3M4L4Lm0SON2N2N2N23M2M3N2N3M1O1O001O000O2O0O2N1O2M2O1O2N3UOl0QOP1C<D<D<Eoc^5"}, "label": "news reporter with blonde hair and green jacket"}]}
{"id": 250569, "image": "COCO_train2014_000000250569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing black coat sitting along with a woman\"."}], "task": "refering", "answer_template": "The \"a man wearing black coat sitting along with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271], "bbox": [0.5297031249999999, 0.25695833333333334, 0.79371875, 0.7031458333333334], "iscrowd": 0, "area": 23876.9123, "rle": {"size": [480, 640], "counts": "PWo44f>8J6K6J5L4K3M3N2M3M3N2M3L3L5L4K5K5L4K5K5K4M4K5K5L4N10100O010O100O010ON3M2NSNWMiGf2Y8bM_G\\2b8iM[GS2h8PNVGm1k8WNRGg1HUM`8W1fGa1J[M^8X1eG[1N_M\\8Y1dGU11dMY8[1dGR2[8QNbGQ2[8SNcGm1\\8WN`Gj1_8YN_Gg1_8]N_Gc1`8`N]Ga1c8aN[G_1d8dNZG\\1e8`200O2O0O1O100O1O10000000001O00000000000000000O2O01O1O001O1O1O00100O3M2N3nJZGd3h8UL]Gk3e8oK_GQ4d8gKbGY4_8`KfG`4\\8ZKhGf4Z9O00001O01O01O00001O00001O00001O01O01O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O1O1O1O1O1O1O1O2N1O3M4L3M3N2M3M4L3M3oNQ1bNiem1"}, "label": "a man wearing black coat sitting along with a woman"}]}
{"id": 250569, "image": "COCO_train2014_000000250569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man in a suit and tie\"."}], "task": "refering", "answer_template": "The \"an older man in a suit and tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271], "bbox": [0.5297031249999999, 0.25695833333333334, 0.79371875, 0.7031458333333334], "iscrowd": 0, "area": 23876.9123, "rle": {"size": [480, 640], "counts": "PWo44f>8J6K6J5L4K3M3N2M3M3N2M3L3L5L4K5K5L4K5K5K4M4K5K5L4N10100O010O100O010ON3M2NSNWMiGf2Y8bM_G\\2b8iM[GS2h8PNVGm1k8WNRGg1HUM`8W1fGa1J[M^8X1eG[1N_M\\8Y1dGU11dMY8[1dGR2[8QNbGQ2[8SNcGm1\\8WN`Gj1_8YN_Gg1_8]N_Gc1`8`N]Ga1c8aN[G_1d8dNZG\\1e8`200O2O0O1O100O1O10000000001O00000000000000000O2O01O1O001O1O1O00100O3M2N3nJZGd3h8UL]Gk3e8oK_GQ4d8gKbGY4_8`KfG`4\\8ZKhGf4Z9O00001O01O01O00001O00001O00001O01O01O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O1O1O1O1O1O1O1O2N1O3M4L3M3N2M3M4L3M3oNQ1bNiem1"}, "label": "an older man in a suit and tie"}]}
{"id": 447179, "image": "COCO_train2014_000000447179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cake table with the cake decorated like a present\"."}], "task": "refering", "answer_template": "The \"the cake table with the cake decorated like a present\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.117203125, 0.529906103286385, 0.7753125, 0.9881455399061034], "iscrowd": 0, "area": 42125.96149999999, "rle": {"size": [426, 640], "counts": "dPP11Y=0O1O100O100O1O100O100O1O2O0O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O101N1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O2O0O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O101N100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O2O0O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O101N1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O10bKTH^3k7^LZHb3f7YL_Hg3`7XLbHh3]7XLdHh3\\7WLeHi3Z7WLgHi3Y7VLhHj3W7VLiHk3V7ULkHk3U7ULkHk3T7ULmHl3R7SLnHn3Q7RLPIn3o6RLRIn3n6QLSIo3l6QLUIo3k6PLUIQ4j6PLVIP4i6PLXIP4h6oKYIQ4f6oKZIR4f6mK[IS4d6mK]IS4b6nK^IR4b6mK^IT4a6mK_IS4a6lK_IU4`6lK`IT4_6mKaIS4_6lKaIU4^6lKbIT4^6kKcIU4\\6lKcIU4\\6kKeIU4[6kKeIU4Z6kKfIV4Z6hKhIX4X6gKhIZ4X6eKiI[4W6cKkI]4U6bKkI_4U6`KlI`4T6^KnIb4S6\\KmIe4S6ZKnIf4Z710000O100O100O100O100O2N100O100O100O100O1O100O100O100O100O100O1000000O1000000O01000O1000L4M2M4TMjGa0Y8YOnGc0U8YOPHb0S8_OnG=U8ClG9W8GjG4Z8KhG1[8OfGM^82cGIa87`GEc8:_GAe8?\\G]Og8c0ZGYOi8g0XGTOm8j0UGROn8n0SGnNP9R1QGiNS9W1nFeNU9[1lF`NX9`1iF\\N[9c1V100000001O00000000001O00000O10001aN\\DV1d;fNaDY1i;N2O2N1O1N5L4L4L5K4K5L5Kgck1"}, "label": "the cake table with the cake decorated like a present"}]}
{"id": 447179, "image": "COCO_train2014_000000447179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white table with wine glasses on it\"."}], "task": "refering", "answer_template": "The \"a red and white table with wine glasses on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.117203125, 0.529906103286385, 0.7753125, 0.9881455399061034], "iscrowd": 0, "area": 42125.96149999999, "rle": {"size": [426, 640], "counts": "dPP11Y=0O1O100O100O1O100O100O1O2O0O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O101N1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O2O0O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O101N100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O2O0O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100O101N1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O10bKTH^3k7^LZHb3f7YL_Hg3`7XLbHh3]7XLdHh3\\7WLeHi3Z7WLgHi3Y7VLhHj3W7VLiHk3V7ULkHk3U7ULkHk3T7ULmHl3R7SLnHn3Q7RLPIn3o6RLRIn3n6QLSIo3l6QLUIo3k6PLUIQ4j6PLVIP4i6PLXIP4h6oKYIQ4f6oKZIR4f6mK[IS4d6mK]IS4b6nK^IR4b6mK^IT4a6mK_IS4a6lK_IU4`6lK`IT4_6mKaIS4_6lKaIU4^6lKbIT4^6kKcIU4\\6lKcIU4\\6kKeIU4[6kKeIU4Z6kKfIV4Z6hKhIX4X6gKhIZ4X6eKiI[4W6cKkI]4U6bKkI_4U6`KlI`4T6^KnIb4S6\\KmIe4S6ZKnIf4Z710000O100O100O100O100O2N100O100O100O100O1O100O100O100O100O100O1000000O1000000O01000O1000L4M2M4TMjGa0Y8YOnGc0U8YOPHb0S8_OnG=U8ClG9W8GjG4Z8KhG1[8OfGM^82cGIa87`GEc8:_GAe8?\\G]Og8c0ZGYOi8g0XGTOm8j0UGROn8n0SGnNP9R1QGiNS9W1nFeNU9[1lF`NX9`1iF\\N[9c1V100000001O00000000001O00000O10001aN\\DV1d;fNaDY1i;N2O2N1O1N5L4L4L5K4K5L5Kgck1"}, "label": "a red and white table with wine glasses on it"}]}
{"id": 160459, "image": "COCO_train2014_000000160459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"suv seen through window\"."}], "task": "refering", "answer_template": "The \"suv seen through window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 132, 133, 134, 150, 151, 152, 153, 155, 156, 157, 176, 178, 179], "bbox": [0.549703125, 0.19618750000000001, 0.83978125, 0.4252291666666667], "iscrowd": 0, "area": 6718.637449999998, "rle": {"size": [480, 640], "counts": "cUU5;e>001O0001O000001O00000000000_NE^D;`;F`D:_;GaD9^;HbD8\\;JdD6[;KeD5Z;LfD5X;LhD4W;MiD3V;NjD2U;OkD1T;0mDOR;2nDOQ;1oDOQ;1oDOP;2PENP;2PENP;2PENo:3QENP=0;CaP63]oIb0^O1O1\\CEa:=_EC`:>`EB`:>`EB`:>`EB`:>`EB`:>`EB_:?aEA_:?aEA_:?aEA_:?aEA_:?aEA_:?aEA_:?aEA^:`0jEXOV:h0Q20000000O1000000^CD`:<`ED`:<`EE_:;aEE^:<bED^:<bED^:<cEC]:=cEC]:=cEC]:=cEC\\:>dEC[:=eEC[:=eEC[:=eEC[:=eEC[:=eECZ:>fEBZ:>T2000000000O100000B>^Ogb6h0k\\I800QCoNo;Q1QDUOZOd0[;7[EUOZOd0[;7[EUOZOd0Z;8\\ETOZOd0Z;8\\ETOZOd0Z;8\\ETOZOd0Z;8\\EUOYOc0[;8\\EUOYOc0[;8\\EUOYOb0\\;9[EUOYOb0\\;9[EUOYOb0\\;9[EUOYOb0\\;9[EUOYOb0\\;9[EUOZOa0Z;;\\ETOZOa0Z;;\\ETOZOa0Z;;\\ETOZOa0Z;;\\ETOZO`0\\;;ZEUOZO`0^;9XEWOZO`0`;7VEYOZO`0a;6UEZOZO`0a;6UEZOZO`0a;6UEZOZO`0a;6UEZOZO;I@h;k0UEZOZO;K^Of;l0VE[OYO;e;;REZOYO;e;;RE[OYO8f;=QE[OYO8f;=QE[OYO8f;=QE[OYO8f;=QE[OYO9e;<RE[OYO:d;;SE[OYO;c;:TE[OYO<b;9UE[OYO<b;9UE[OYO=a;8VE[OYO=a;8VE[OYO=a;8VE[OYO>`;7WE[OYO>`;7WE[OYO>`;7WE[OZO=_;8WE[OZO=^;9XEZOZO=^;9XEZOZO=_;8WE[OZO>`;5VE]OZO=a;6UE^OYO<b;6UE^OYO<b;6UE^OYO;b;7VE^OXO;b;7VE^OXO;b;7VE^OXO:c;8UE^OXO9d;9TE^OXO8f;9RE_OXO7h;9PE@YO6g;:PE@YO5i;:nDAYO3l;;kDBYO1o;<hDCYO0P<=gDCYONS<>iD@SO2ja`1"}, "label": "suv seen through window"}]}
{"id": 160459, "image": "COCO_train2014_000000160459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a station wagon with writing on the back window\"."}], "task": "refering", "answer_template": "The \"a station wagon with writing on the back window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 132, 133, 134, 150, 151, 152, 153, 155, 156, 157, 176, 178, 179], "bbox": [0.549703125, 0.19618750000000001, 0.83978125, 0.4252291666666667], "iscrowd": 0, "area": 6718.637449999998, "rle": {"size": [480, 640], "counts": "cUU5;e>001O0001O000001O00000000000_NE^D;`;F`D:_;GaD9^;HbD8\\;JdD6[;KeD5Z;LfD5X;LhD4W;MiD3V;NjD2U;OkD1T;0mDOR;2nDOQ;1oDOQ;1oDOP;2PENP;2PENP;2PENo:3QENP=0;CaP63]oIb0^O1O1\\CEa:=_EC`:>`EB`:>`EB`:>`EB`:>`EB`:>`EB_:?aEA_:?aEA_:?aEA_:?aEA_:?aEA_:?aEA_:?aEA^:`0jEXOV:h0Q20000000O1000000^CD`:<`ED`:<`EE_:;aEE^:<bED^:<bED^:<cEC]:=cEC]:=cEC]:=cEC\\:>dEC[:=eEC[:=eEC[:=eEC[:=eEC[:=eECZ:>fEBZ:>T2000000000O100000B>^Ogb6h0k\\I800QCoNo;Q1QDUOZOd0[;7[EUOZOd0[;7[EUOZOd0Z;8\\ETOZOd0Z;8\\ETOZOd0Z;8\\ETOZOd0Z;8\\EUOYOc0[;8\\EUOYOc0[;8\\EUOYOb0\\;9[EUOYOb0\\;9[EUOYOb0\\;9[EUOYOb0\\;9[EUOYOb0\\;9[EUOZOa0Z;;\\ETOZOa0Z;;\\ETOZOa0Z;;\\ETOZOa0Z;;\\ETOZO`0\\;;ZEUOZO`0^;9XEWOZO`0`;7VEYOZO`0a;6UEZOZO`0a;6UEZOZO`0a;6UEZOZO`0a;6UEZOZO;I@h;k0UEZOZO;K^Of;l0VE[OYO;e;;REZOYO;e;;RE[OYO8f;=QE[OYO8f;=QE[OYO8f;=QE[OYO8f;=QE[OYO9e;<RE[OYO:d;;SE[OYO;c;:TE[OYO<b;9UE[OYO<b;9UE[OYO=a;8VE[OYO=a;8VE[OYO=a;8VE[OYO>`;7WE[OYO>`;7WE[OYO>`;7WE[OZO=_;8WE[OZO=^;9XEZOZO=^;9XEZOZO=_;8WE[OZO>`;5VE]OZO=a;6UE^OYO<b;6UE^OYO<b;6UE^OYO;b;7VE^OXO;b;7VE^OXO;b;7VE^OXO:c;8UE^OXO9d;9TE^OXO8f;9RE_OXO7h;9PE@YO6g;:PE@YO5i;:nDAYO3l;;kDBYO1o;<hDCYO0P<=gDCYONS<>iD@SO2ja`1"}, "label": "a station wagon with writing on the back window"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl was teasing\"."}], "task": "refering", "answer_template": "The \"a girl was teasing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 157, 158, 159, 160, 174, 175, 176, 177, 191, 192, 193, 194, 208, 209, 210, 211, 225, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 311, 312, 313, 328, 329, 330, 331, 345, 346, 362], "bbox": [0.2516875, 0.366296875, 0.47939583333333335, 0.92584375], "iscrowd": 0, "area": 24912.161249999997, "rle": {"size": [640, 480], "counts": "Wn[25hc07I8H7I7I7I7I8k_OfNW=a1[B]On<j0eB2d<6oB:e<OmB;n<OdB;W=Y3K6J5M4K4M4K5K4M4K4M4L301N:G>dEZGa8T9_FcG]9d9K6K3L3N1N2O1N3N0O02M2O1N3N1SMbGWJ_8i5fGRJZ8n5jGnIW8Q6nGjIR8V6SHfIl7Z6YHaIh7^6]H]Ic7c6aHYI_7g6fHTI[7l6iHoHW7Q7nHcHY7]7f2O00VEWH]9i7ZF`Hg9_7QFiHP:V7jEPIV:P7eEVI[:i6`E\\Ia:c6ZEbIg:k7001O1O1O002O0O2N2N1O2N2N1O2N1O2N2N3M3UO^DRIe;c6cD]Ia;P6oDoIT;l5nDTJU;i5lDVJW;g5iDYJY;d5iD[JY;c5gD]J[;a5fD^J\\;`5dD`J^;]5dDbJ];Y3eCVNn0aN_;h2VDbN=fN^;U2kDoNIkN^;g1ZEYOYOPO_;c1_EWOTOUO`;_1bEWOoNZOa;[1gEUOjN_Ob;V1kEVOdNDd;Q1oEUO^NJf;l0RFUOZNMh;i0UFTOTN3i;e0YFSOoM7Sb000O1O2O]gk4"}, "label": "a girl was teasing"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"barefoot girl in glasses sticking tongue out at camera\"."}], "task": "refering", "answer_template": "The \"barefoot girl in glasses sticking tongue out at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 157, 158, 159, 160, 174, 175, 176, 177, 191, 192, 193, 194, 208, 209, 210, 211, 225, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 311, 312, 313, 328, 329, 330, 331, 345, 346, 362], "bbox": [0.2516875, 0.366296875, 0.47939583333333335, 0.92584375], "iscrowd": 0, "area": 24912.161249999997, "rle": {"size": [640, 480], "counts": "Wn[25hc07I8H7I7I7I7I8k_OfNW=a1[B]On<j0eB2d<6oB:e<OmB;n<OdB;W=Y3K6J5M4K4M4K5K4M4K4M4L301N:G>dEZGa8T9_FcG]9d9K6K3L3N1N2O1N3N0O02M2O1N3N1SMbGWJ_8i5fGRJZ8n5jGnIW8Q6nGjIR8V6SHfIl7Z6YHaIh7^6]H]Ic7c6aHYI_7g6fHTI[7l6iHoHW7Q7nHcHY7]7f2O00VEWH]9i7ZF`Hg9_7QFiHP:V7jEPIV:P7eEVI[:i6`E\\Ia:c6ZEbIg:k7001O1O1O002O0O2N2N1O2N2N1O2N1O2N2N3M3UO^DRIe;c6cD]Ia;P6oDoIT;l5nDTJU;i5lDVJW;g5iDYJY;d5iD[JY;c5gD]J[;a5fD^J\\;`5dD`J^;]5dDbJ];Y3eCVNn0aN_;h2VDbN=fN^;U2kDoNIkN^;g1ZEYOYOPO_;c1_EWOTOUO`;_1bEWOoNZOa;[1gEUOjN_Ob;V1kEVOdNDd;Q1oEUO^NJf;l0RFUOZNMh;i0UFTOTN3i;e0YFSOoM7Sb000O1O2O]gk4"}, "label": "barefoot girl in glasses sticking tongue out at camera"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black color sofa in the living room\"."}], "task": "refering", "answer_template": "The \"black color sofa in the living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 170, 171, 172, 173, 174, 177, 178, 179, 187, 188, 189, 190, 191, 194, 195, 196, 204, 205, 206, 207, 208, 212, 213, 221, 222, 223, 224, 225, 228, 229, 230, 238, 241, 242, 245, 246, 247, 262, 263, 264, 279, 280], "bbox": [0.0, 0.42204687500000004, 0.5710208333333333, 0.7025625], "iscrowd": 0, "area": 21447.48905, "rle": {"size": [640, 480], "counts": "W9P3Pa0000000000000001O000000000000000O1000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000O10000000001O1O001O1O001O1N2O001O1O001O1O001N2O001O1O001O1O0O2O1O001O1O001O1O1N101O1O001O1OmMS@3m?IY@5f?I`@3a?Je@1]?Ki@2X?Kn@0T?MRAOo>MWA0j>M\\ANf>O_ANb>NeAM\\>1jAKW>2nAJT>2SBJn=3XBHj=4\\BIe=4U3M3LmS[11UldN4M2N3M3M2M4VA]OQ:f0dEB\\:`0ZEHf:;oDMQ;6dD2[;1[D7e;LPD<P<GfC`0Z<B`Cb0`<AXCd0g<@RCe0m<]OmBg0S=\\OeBi0[=ZO^Bj0a=YOXBl0h=WOQBm0o=VOjAn0V>TOcAQ1]>Y2O100000000O100000000O100000000O100000000O11O3N2M3M3K5J6I7J6I7J6I6J7J6I7J6I7J6I7C=]Oc0\\Od0]OS^P4"}, "label": "black color sofa in the living room"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is black sofa back of ladies\"."}], "task": "refering", "answer_template": "The \"there is black sofa back of ladies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 170, 171, 172, 173, 174, 177, 178, 179, 187, 188, 189, 190, 191, 194, 195, 196, 204, 205, 206, 207, 208, 212, 213, 221, 222, 223, 224, 225, 228, 229, 230, 238, 241, 242, 245, 246, 247, 262, 263, 264, 279, 280], "bbox": [0.0, 0.42204687500000004, 0.5710208333333333, 0.7025625], "iscrowd": 0, "area": 21447.48905, "rle": {"size": [640, 480], "counts": "W9P3Pa0000000000000001O000000000000000O1000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000O10000000001O1O001O1O001O1N2O001O1O001O1O001N2O001O1O001O1O0O2O1O001O1O001O1O1N101O1O001O1OmMS@3m?IY@5f?I`@3a?Je@1]?Ki@2X?Kn@0T?MRAOo>MWA0j>M\\ANf>O_ANb>NeAM\\>1jAKW>2nAJT>2SBJn=3XBHj=4\\BIe=4U3M3LmS[11UldN4M2N3M3M2M4VA]OQ:f0dEB\\:`0ZEHf:;oDMQ;6dD2[;1[D7e;LPD<P<GfC`0Z<B`Cb0`<AXCd0g<@RCe0m<]OmBg0S=\\OeBi0[=ZO^Bj0a=YOXBl0h=WOQBm0o=VOjAn0V>TOcAQ1]>Y2O100000000O100000000O100000000O100000000O11O3N2M3M3K5J6I7J6I7J6I6J7J6I7J6I7J6I7C=]Oc0\\Od0]OS^P4"}, "label": "there is black sofa back of ladies"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red blanket on the couch\"."}], "task": "refering", "answer_template": "The \"red blanket on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 255, 256, 257, 258, 259, 272, 273, 274, 275, 276, 290, 291, 292, 293, 308, 309, 310, 326, 327], "bbox": [0.0, 0.606453125, 0.2982083333333333, 0.8602187499999999], "iscrowd": 0, "area": 12586.43505, "rle": {"size": [640, 480], "counts": "[<1oc03L4M3M3M3M3L4M3M3M3L4M3M3M3L4M3M3M3M3L4M3M3M3L2O1O001O1N101O1O10O01O1O001O1O1O010O1O001O1O1O1O2O0O2N2N2N1O2N2N2O0O1O1O1O1O100O010O100O10O01O100O10O0100O100O010O1O100O010O100O10O01O100O10O0100O10O0100O010O1000O01000O010000O01000O10O10O10O1000O01F:K4PNi_O6Y`0nNf@m0\\?eNVAU1n`0L3M2N3M2N3M3M2N3M2N3M3M2N3M3MkSb6"}, "label": "red blanket on the couch"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red blanket thrown on a couch\"."}], "task": "refering", "answer_template": "The \"a red blanket thrown on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 255, 256, 257, 258, 259, 272, 273, 274, 275, 276, 290, 291, 292, 293, 308, 309, 310, 326, 327], "bbox": [0.0, 0.606453125, 0.2982083333333333, 0.8602187499999999], "iscrowd": 0, "area": 12586.43505, "rle": {"size": [640, 480], "counts": "[<1oc03L4M3M3M3M3L4M3M3M3L4M3M3M3L4M3M3M3M3L4M3M3M3L2O1O001O1N101O1O10O01O1O001O1O1O010O1O001O1O1O1O2O0O2N2N2N1O2N2N2O0O1O1O1O1O100O010O100O10O01O100O10O0100O100O010O1O100O010O100O10O01O100O10O0100O10O0100O010O1000O01000O010000O01000O10O10O10O1000O01F:K4PNi_O6Y`0nNf@m0\\?eNVAU1n`0L3M2N3M2N3M3M2N3M2N3M3M2N3M3MkSb6"}, "label": "a red blanket thrown on a couch"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in an orange shirt and pink shorts\"."}], "task": "refering", "answer_template": "The \"a woman in an orange shirt and pink shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 95, 96, 112, 113, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 179, 180, 181, 182, 196, 197, 198, 199, 213, 214, 215, 216, 230, 231, 232, 233, 247, 248, 249, 250, 264], "bbox": [0.5052708333333333, 0.19359375, 0.7754583333333334, 0.6840468749999999], "iscrowd": 0, "area": 22154.293650000003, "rle": {"size": [640, 480], "counts": "jSh44ic03L5L3L5L3L5K6Ke0]O4L3L4M3M3M3M3M2mBbMa8`2[GbMe8a2VGaMi8b2RGbMm8a2fDXMc09f:a2cDZMd06i:b2`DZMe06j:c2cCTM48\\13l:c2^C^MO1d10n:d2]CXN_1VOT;e2ZCZN\\1SOY;`4`DbK_;c4[D^Ke;g4UD[Kj;j4PDXKo;\\6O200O10O0]O`ChI`<X6aCgI^<Z6cCdI^<\\6bCXIi<i6XCVI^;<YE]6ZOWIU;f0^ET6h:QJUEo5i:VJTEj5j:bJkD^5T;eJjD\\5V;fJgD[5X;gJeD[5Z;e110O100O01N101M2M4L3L5L3N3M3L3N3L4M3L5L3004M3L4L5L3L4L4M3L5L]IZDn4c;oJ`DP5`;mJcDS5a;fJcDX5b;aJaD_5c;ZJaDe5b;UJaDk5a;lIgDR6\\;kIeDU6];iIcDW6`;fIaDY6a;dI`D[6d<O001O0O2O0`LZBf0g=XOZBi0e=WO\\Bi0d=VO\\BgM1b2c=G]BfM0d2c=D_BgMNf2b=D`BeMNg2c=C`BfMLh2c=BbBeMKj2c=AbBdMKg2g=E_BcMM_2k=MXBdMOQ1LJT>`1QBdM2`0GH;5Q>o1lAcM4<HK`0Ho=]2eAdM68JMV?V2j@eM83M0Q?X2k@dM;NM4n>Y2j@dMR12U>]1_@YOl1XOh=\\1^@ZOk1YOj=Y1^@YOha0d0Z^OUOoa0f0T^OWOPb0f0R^OWORb0e0Q^OXOSb0e0g0L5K4M2M3N3L3N[XS2"}, "label": "a woman in an orange shirt and pink shorts"}]}
{"id": 398036, "image": "COCO_train2014_000000398036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing an orange shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 95, 96, 112, 113, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 179, 180, 181, 182, 196, 197, 198, 199, 213, 214, 215, 216, 230, 231, 232, 233, 247, 248, 249, 250, 264], "bbox": [0.5052708333333333, 0.19359375, 0.7754583333333334, 0.6840468749999999], "iscrowd": 0, "area": 22154.293650000003, "rle": {"size": [640, 480], "counts": "jSh44ic03L5L3L5L3L5K6Ke0]O4L3L4M3M3M3M3M2mBbMa8`2[GbMe8a2VGaMi8b2RGbMm8a2fDXMc09f:a2cDZMd06i:b2`DZMe06j:c2cCTM48\\13l:c2^C^MO1d10n:d2]CXN_1VOT;e2ZCZN\\1SOY;`4`DbK_;c4[D^Ke;g4UD[Kj;j4PDXKo;\\6O200O10O0]O`ChI`<X6aCgI^<Z6cCdI^<\\6bCXIi<i6XCVI^;<YE]6ZOWIU;f0^ET6h:QJUEo5i:VJTEj5j:bJkD^5T;eJjD\\5V;fJgD[5X;gJeD[5Z;e110O100O01N101M2M4L3L5L3N3M3L3N3L4M3L5L3004M3L4L5L3L4L4M3L5L]IZDn4c;oJ`DP5`;mJcDS5a;fJcDX5b;aJaD_5c;ZJaDe5b;UJaDk5a;lIgDR6\\;kIeDU6];iIcDW6`;fIaDY6a;dI`D[6d<O001O0O2O0`LZBf0g=XOZBi0e=WO\\Bi0d=VO\\BgM1b2c=G]BfM0d2c=D_BgMNf2b=D`BeMNg2c=C`BfMLh2c=BbBeMKj2c=AbBdMKg2g=E_BcMM_2k=MXBdMOQ1LJT>`1QBdM2`0GH;5Q>o1lAcM4<HK`0Ho=]2eAdM68JMV?V2j@eM83M0Q?X2k@dM;NM4n>Y2j@dMR12U>]1_@YOl1XOh=\\1^@ZOk1YOj=Y1^@YOha0d0Z^OUOoa0f0T^OWOPb0f0R^OWORb0e0Q^OXOSb0e0g0L5K4M2M3N3L3N[XS2"}, "label": "a woman wearing an orange shirt"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mahou branded lunch cup\"."}], "task": "refering", "answer_template": "The \"a mahou branded lunch cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147], "bbox": [0.24478125, 0.019538106235565823, 0.45610937499999993, 0.45773672055427256], "iscrowd": 0, "area": 20620.138600000006, "rle": {"size": [433, 640], "counts": "e]R24^=3WO3gC0T<<aCIZ<k0K6K4K6J5L5J5L5J5L5J5L5J5L5K4L5K4L4L5K4K6K4L5K4L5K4L5K4L5K4K6K0000000000001O0O10000000000000000O1000000000000000000000001O00000000000000000000000000000000O10000O100O10000O10000O100O10000O10000O100O2O000O2O0O101O0O2O000O2O0O101O0O101M2E<D;E;E<D;E;E<C<E<D;E;E<D;EcTc4"}, "label": "a mahou branded lunch cup"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cup has mahou written on it\"."}], "task": "refering", "answer_template": "The \"the cup has mahou written on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147], "bbox": [0.24478125, 0.019538106235565823, 0.45610937499999993, 0.45773672055427256], "iscrowd": 0, "area": 20620.138600000006, "rle": {"size": [433, 640], "counts": "e]R24^=3WO3gC0T<<aCIZ<k0K6K4K6J5L5J5L5J5L5J5L5J5L5K4L5K4L4L5K4K6K4L5K4L5K4L5K4L5K4K6K0000000000001O0O10000000000000000O1000000000000000000000001O00000000000000000000000000000000O10000O100O10000O10000O100O10000O10000O100O2O000O2O0O101O0O2O000O2O0O101O0O101M2E<D;E;E<D;E;E<C<E<D;E;E<D;EcTc4"}, "label": "the cup has mahou written on it"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a hot dog next to the drink in the clear cup\"."}], "task": "refering", "answer_template": "The \"there is a hot dog next to the drink in the clear cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 78, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130], "bbox": [0.42662500000000003, 0.15057736720554274, 0.687703125, 0.3976443418013857], "iscrowd": 0, "area": 9677.56855, "rle": {"size": [433, 640], "counts": "aac34W=8H8H7I8H8]Oc0O0O2N2O0O1O010O01O01O010O00010O00010O010O00010O00010O010O00010O010O01O010O00010O010O0010O01O010O010O0010O01O01O010O10O010O0100O010O010O010O1O010O010O010O10O010O0100O010O0010O01O01O00001O00001O00001O00001O00001O00001O00001O00001O00001O000O2O00001O00001O00001O00001O001N2O1O1N2O001N2O1O1N2O1N2O2N1N2O1O1N2O1O2M2O1O2M4L4K5Jb]d2"}, "label": "there is a hot dog next to the drink in the clear cup"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog right next to a beer\"."}], "task": "refering", "answer_template": "The \"a hot dog right next to a beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 78, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130], "bbox": [0.42662500000000003, 0.15057736720554274, 0.687703125, 0.3976443418013857], "iscrowd": 0, "area": 9677.56855, "rle": {"size": [433, 640], "counts": "aac34W=8H8H7I8H8]Oc0O0O2N2O0O1O010O01O01O010O00010O00010O010O00010O00010O010O00010O010O01O010O00010O010O0010O01O010O010O0010O01O01O010O10O010O0100O010O010O010O1O010O010O010O10O010O0100O010O0010O01O01O00001O00001O00001O00001O00001O00001O00001O00001O00001O000O2O00001O00001O00001O00001O001N2O1O1N2O001N2O1O1N2O1N2O2N1N2O1O1N2O1O2M2O1O2M4L4K5Jb]d2"}, "label": "a hot dog right next to a beer"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a full hot - dog in front of two others\"."}], "task": "refering", "answer_template": "The \"a full hot - dog in front of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.30440625, 0.5987990762124711, 0.8501249999999999, 1.0], "iscrowd": 0, "area": 34187.7956, "rle": {"size": [433, 640], "counts": "ghb2b0m<9F5K5K5K5L4M3M3N2M3M3M3N2M3M1O2O0O2O001O001O0CcMfE^2X:dMgE\\2X:gMeE[2Y:gMfEY2Y:iMfEX2X:jMgEV2W:nMgES2W:d0O2N1O2N1N3N1O2N1O2N1N3N1O2N1O2M201O001O001O00001O0010O01O0010O01O010O0010O01O0010O2O0O100O100O100O100O100O100O100O100O100O01O010O001O010O001O0010O01O0010O01O001O010O001O010O001O010O001O10O010O01O010O01O00001O0O2O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O2N2N2N2N00O1000001O000001O1O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1M2N3M2N3M2L5K5K4L_VX1"}, "label": "a full hot - dog in front of two others"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cup that shows a coke bottle on it\"."}], "task": "refering", "answer_template": "The \"the cup that shows a coke bottle on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 272, 273], "bbox": [0.703921875, 0.004503464203233256, 1.0, 0.7707852193995381], "iscrowd": 0, "area": 47446.8842, "rle": {"size": [433, 640], "counts": "lfn5>U=OM3M2O2M3M3O001O1cE;l6P4UOk0hJQJX1j6cM`I]2[7_LnHa3b8O00001O0000001O00001O0000001O00001O00001O0000010O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001OO100O100O100O10000O100O100O1M3K5J6J6K5J6K5J6K5J6K5J6K5J7J5J6J6K5J6K5J6K5J6K5J6K5J6K5K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4K5L4L4K5L4L4K5L0"}, "label": "the cup that shows a coke bottle on it"}]}
{"id": 406230, "image": "COCO_train2014_000000406230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red coca cola cup\"."}], "task": "refering", "answer_template": "The \"the red coca cola cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 272, 273], "bbox": [0.703921875, 0.004503464203233256, 1.0, 0.7707852193995381], "iscrowd": 0, "area": 47446.8842, "rle": {"size": [433, 640], "counts": "lfn5>U=OM3M2O2M3M3O001O1cE;l6P4UOk0hJQJX1j6cM`I]2[7_LnHa3b8O00001O0000001O00001O0000001O00001O00001O0000010O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001OO100O100O100O10000O100O100O1M3K5J6J6K5J6K5J6K5J6K5J6K5J7J5J6J6K5J6K5J6K5J6K5J6K5J6K5K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4K5L4L4K5L4L4K5L0"}, "label": "the red coca cola cup"}]}
{"id": 258774, "image": "COCO_train2014_000000258774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a baseball cap sitting in a chair in front of a desk\"."}], "task": "refering", "answer_template": "The \"a man in a baseball cap sitting in a chair in front of a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 177, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 246], "bbox": [0.6644062500000001, 0.3983177570093458, 0.9045937500000001, 0.6760514018691589], "iscrowd": 0, "area": 7719.0645, "rle": {"size": [428, 640], "counts": "nla51Z=3N2M4M0O2O0O2O0O2K4M4K4L5L3L5K4M3O1O2N1N2O1O1O1O1O1O1O1N02L4M3L4M3L43N7H7IO10000O100O100O2O0O10000O100O100O101N10000O2O000O10001N10000O1000O010O1O1O00O2N1O1O1N3O01O1O00100O001O010O1O00100WDfN_;Z1aDgN^;Z1aDgN^;Y1bDhN^;X1aDiN^;X1bDiN\\;W1dDjN\\;V1dDjN[;W1dDjN[;V1eDkNZ;V1fDjNZ;U1fDlNY;U1gDlNW;U1iDkNW;T1iDmNV;T1jDlNU;U1kDkN[:EXFa1]OjN[:GVF_1_OjN[:JRF]1CiN[:LoE\\1FgN\\:1iEY1KeN]:5dEW1ObN_:S2aElM`:T2`EjMb:V2^EiMc:V2^EiMc:W2]EgMd:Z2[EfMf:Z260O100O10000O100O11O001M2N3N2M2N3M2N3N1N3M2N3N1N3Mc0]O6K5J6J6JZYi0"}, "label": "a man in a baseball cap sitting in a chair in front of a desk"}]}
{"id": 258774, "image": "COCO_train2014_000000258774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sits at home watching tv\"."}], "task": "refering", "answer_template": "The \"a man sits at home watching tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 177, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 246], "bbox": [0.6644062500000001, 0.3983177570093458, 0.9045937500000001, 0.6760514018691589], "iscrowd": 0, "area": 7719.0645, "rle": {"size": [428, 640], "counts": "nla51Z=3N2M4M0O2O0O2O0O2K4M4K4L5L3L5K4M3O1O2N1N2O1O1O1O1O1O1O1N02L4M3L4M3L43N7H7IO10000O100O100O2O0O10000O100O100O101N10000O2O000O10001N10000O1000O010O1O1O00O2N1O1O1N3O01O1O00100O001O010O1O00100WDfN_;Z1aDgN^;Z1aDgN^;Y1bDhN^;X1aDiN^;X1bDiN\\;W1dDjN\\;V1dDjN[;W1dDjN[;V1eDkNZ;V1fDjNZ;U1fDlNY;U1gDlNW;U1iDkNW;T1iDmNV;T1jDlNU;U1kDkN[:EXFa1]OjN[:GVF_1_OjN[:JRF]1CiN[:LoE\\1FgN\\:1iEY1KeN]:5dEW1ObN_:S2aElM`:T2`EjMb:V2^EiMc:V2^EiMc:W2]EgMd:Z2[EfMf:Z260O100O10000O100O11O001M2N3N2M2N3M2N3N1N3M2N3N1N3Mc0]O6K5J6J6JZYi0"}, "label": "a man sits at home watching tv"}]}
{"id": 258774, "image": "COCO_train2014_000000258774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather chair with a red pillow\"."}], "task": "refering", "answer_template": "The \"a brown leather chair with a red pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.5956542056074766, 0.30771875, 1.0], "iscrowd": 0, "area": 28487.385499999997, "rle": {"size": [428, 640], "counts": "[8Q5Z81000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10000001O0000001O00001O00001O0000001O00001O00001O1O1O2N1O1O1O2N1O7I:F1O00000000000000000000000000O100000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O1000000000000000000000O1O1O1O1O1O1N2O1O1N2O1N2O1O1N2O1O1N101O1N3N3M3[MQFg1R:UNXFc1k9WN`F`1d9ZNhF]1[9]NPG[1h:G9H8H8G9HXih5"}, "label": "a brown leather chair with a red pillow"}]}
{"id": 258774, "image": "COCO_train2014_000000258774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather love seat\"."}], "task": "refering", "answer_template": "The \"a brown leather love seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.5956542056074766, 0.30771875, 1.0], "iscrowd": 0, "area": 28487.385499999997, "rle": {"size": [428, 640], "counts": "[8Q5Z81000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10000001O0000001O00001O00001O0000001O00001O00001O1O1O2N1O1O1O2N1O7I:F1O00000000000000000000000000O100000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O1000000000000000000000O1O1O1O1O1O1N2O1O1N2O1N2O1O1N2O1O1N101O1N3N3M3[MQFg1R:UNXFc1k9WN`F`1d9ZNhF]1[9]NPG[1h:G9H8H8G9HXih5"}, "label": "a brown leather love seat"}]}
{"id": 258774, "image": "COCO_train2014_000000258774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown sofa\"."}], "task": "refering", "answer_template": "The \"a brown sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 210, 230, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.016515625, 0.6149766355140186, 0.30878125, 0.9680607476635513], "iscrowd": 0, "area": 20310.889450000006, "rle": {"size": [428, 640], "counts": "dk46V=7[EGh7=THIf7;WHIe79XHLc78YHMc76ZHNa76[H0_73^H2^71_H3\\71`H3]7O`H2_70`H0_72_HOa72]HOb74\\HMc74[HMe74YHMg74XHLh75VHLj75THLl75SHKm76QHKo77nGJR87lGKS86kGKU86iGKW86gGKY86eGK[86cGK]87`GJa86]GKc86[GKe86YGLf85XGLh8\\200000000000001O00000000000000001O00000000000000001O0000000001O1O1O2O0O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O0O1000001O000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000O100O1000O0100O100O10O1O001N101O001N2O001O4L4K5L4L4L4K5L4L4L4K5L4L4L4K5L4L4L4K5L4L4L4K5L4L[\\h5"}, "label": "a brown sofa"}]}
{"id": 258774, "image": "COCO_train2014_000000258774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather chair with a red blanket sits next to wrapped gifts\"."}], "task": "refering", "answer_template": "The \"a leather chair with a red blanket sits next to wrapped gifts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 210, 230, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.016515625, 0.6149766355140186, 0.30878125, 0.9680607476635513], "iscrowd": 0, "area": 20310.889450000006, "rle": {"size": [428, 640], "counts": "dk46V=7[EGh7=THIf7;WHIe79XHLc78YHMc76ZHNa76[H0_73^H2^71_H3\\71`H3]7O`H2_70`H0_72_HOa72]HOb74\\HMc74[HMe74YHMg74XHLh75VHLj75THLl75SHKm76QHKo77nGJR87lGKS86kGKU86iGKW86gGKY86eGK[86cGK]87`GJa86]GKc86[GKe86YGLf85XGLh8\\200000000000001O00000000000000001O00000000000000001O0000000001O1O1O2O0O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O0O1000001O000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000O100O1000O0100O100O10O1O001N101O001N2O001O4L4K5L4L4L4K5L4L4L4K5L4L4L4K5L4L4L4K5L4L4L4K5L4L[\\h5"}, "label": "a leather chair with a red blanket sits next to wrapped gifts"}]}
{"id": 537304, "image": "COCO_train2014_000000537304.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an uncooked pizza with four hotdogs\"."}], "task": "refering", "answer_template": "The \"an uncooked pizza with four hotdogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.1275625, 0.7302352941176471, 0.502734375, 0.9533882352941178], "iscrowd": 0, "area": 17151.93975, "rle": {"size": [425, 640], "counts": "h\\R15T=2N1O2M3N2N1O2N1O1O001O1N101O001ROXOYEh0f:YOZEh0e:YOZEh0d:YO\\Eg0c:ZO]Eg0a:ZO^Eg0b:YO^Eh0`:YO`Eg0_:[O`Ef0^:[ObEf0]:ZOcEf0\\:[OdEf0Z:[OfEe0Y:\\OfEf0Y:[OfEe0Y:\\OgEd0X:]OhEc0W:^OiEb0W:^OiEb0V:_OjEa0U:AjE?U:BkE>U:BjE?U:CjE=V:CjE=U:DkE<U:EjE;V:EjE;V:FiE:V:GjE9V:GjE9V:HhE9X:GhE9W:IhE7X:JgE6Y:KfE5Z:LdE5[:McE4]:MaE4_:M`E3`:N^E3a:O]E2c:Z10000O100000000O10000000000O10000000000O10000000000O100O1O1O1O1O10000000000O11O001O1O1O001O1O1O00000000O1000000O10000O1000000O10000O1000000000000000000000000000000001O00000000000000000000000000001O0000000000000000001O000000001O000000001O000000001O0O10001O0O10001N10000O2O00000O101O0O100O2O0O1O1O2N101N2N1O2O1N2N1O2O1N1O2N2O0O2N2N2O0O2N2N101N2N1O3K5L3L5K4L5L3L5K`dS4"}, "label": "an uncooked pizza with four hotdogs"}]}
{"id": 537304, "image": "COCO_train2014_000000537304.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza furthest left from the child\"."}], "task": "refering", "answer_template": "The \"the pizza furthest left from the child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.1275625, 0.7302352941176471, 0.502734375, 0.9533882352941178], "iscrowd": 0, "area": 17151.93975, "rle": {"size": [425, 640], "counts": "h\\R15T=2N1O2M3N2N1O2N1O1O001O1N101O001ROXOYEh0f:YOZEh0e:YOZEh0d:YO\\Eg0c:ZO]Eg0a:ZO^Eg0b:YO^Eh0`:YO`Eg0_:[O`Ef0^:[ObEf0]:ZOcEf0\\:[OdEf0Z:[OfEe0Y:\\OfEf0Y:[OfEe0Y:\\OgEd0X:]OhEc0W:^OiEb0W:^OiEb0V:_OjEa0U:AjE?U:BkE>U:BjE?U:CjE=V:CjE=U:DkE<U:EjE;V:EjE;V:FiE:V:GjE9V:GjE9V:HhE9X:GhE9W:IhE7X:JgE6Y:KfE5Z:LdE5[:McE4]:MaE4_:M`E3`:N^E3a:O]E2c:Z10000O100000000O10000000000O10000000000O10000000000O100O1O1O1O1O10000000000O11O001O1O1O001O1O1O00000000O1000000O10000O1000000O10000O1000000000000000000000000000000001O00000000000000000000000000001O0000000000000000001O000000001O000000001O000000001O0O10001O0O10001N10000O2O00000O101O0O100O2O0O1O1O2N101N2N1O2O1N2N1O2O1N1O2N2O0O2N2N2O0O2N2N101N2N1O3K5L3L5K4L5L3L5K`dS4"}, "label": "the pizza furthest left from the child"}]}
{"id": 250588, "image": "COCO_train2014_000000250588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in red tie\"."}], "task": "refering", "answer_template": "The \"man in red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 94, 95, 96, 97, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 366, 367, 368, 369, 370, 371, 372, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5249791666666667, 0.20379687500000002, 0.99925, 1.0], "iscrowd": 0, "area": 82911.50695, "rle": {"size": [640, 480], "counts": "mkm4b0bb0R1WOj0VOj0RNm1I8HV1jN8H8H7I8H8ZMXHeHm7U7VHiHn7S7THjHR8Q7PHmHU8n6mGPIW8l6kGRIZ8i6hGUI]8f6eGXI_8d6dGYId8_6^G`Il8U6VGiIU9l5mFSJ]9b5eF\\Jf9X5]FgJm9n4UFPKV:e4lEZK^:[OcDQ4Q1bLg:dNgDh4e0cLY;Z3kDdLV;Z3lDfLT;W3PEgLQ;W3REhLn:V3UEhLl:W3UEiLk:W3VEgLk:X3WEgLi:Y3WEfLj:Y3XEfLh:Z3YEdLh:[3ZEdLf:\\3ZEcLg:\\3[EcLe:]3\\EaLe:_3\\E`Ld:_3_E^Lb:b3`E\\L`:c3bE[L_:e3cEYL]:f3eEXL\\:h3fEVLZ:i3hEULY:k3iESLW:l3lEQLU:o3lEPLT:o3oEnKR:R4n2O11O001O001O00001O001O001O001O001O00001O001O001O001N1M3M4L3M5L3L4L5K4M3L4L5K4mBUJm;n5PDYJh;l5UDZJc;j5ZD\\J_;h5^D^J\\;e5bD]JZ;h5bDZJ`;e5]D^Jf;_5XDcJj;[5SDgJP<X5lCkJV<_62M3N3M2M3N2M3M4L3L2O1O0O10000O10000O1000000O2O000O10000O1001O000001O01O0000001O0000001O0001O0001O0000001O00000010O000001O000000001O0hL\\ITId6i6WJ_Hi5^7QKlGo4P8kKYGU4d8eLeF[3X9kLeFV3W9nLhFS3T9P4K5K4K6K4L5K5K4L5K4L5K4L5K5H7I8H7I8I7H7I8H7I8I6I8K5L3N3L3M4L3N3L4L3N3L3M4M3L3M4M2M4L5Ke0\\Od0[Oe0[Of0[OaD"}, "label": "man in red tie"}]}
{"id": 250588, "image": "COCO_train2014_000000250588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with his tongue sticking out , wearing a nice jacket , tie , and hat\"."}], "task": "refering", "answer_template": "The \"a man with his tongue sticking out , wearing a nice jacket , tie , and hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 94, 95, 96, 97, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 179, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 366, 367, 368, 369, 370, 371, 372, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5249791666666667, 0.20379687500000002, 0.99925, 1.0], "iscrowd": 0, "area": 82911.50695, "rle": {"size": [640, 480], "counts": "mkm4b0bb0R1WOj0VOj0RNm1I8HV1jN8H8H7I8H8ZMXHeHm7U7VHiHn7S7THjHR8Q7PHmHU8n6mGPIW8l6kGRIZ8i6hGUI]8f6eGXI_8d6dGYId8_6^G`Il8U6VGiIU9l5mFSJ]9b5eF\\Jf9X5]FgJm9n4UFPKV:e4lEZK^:[OcDQ4Q1bLg:dNgDh4e0cLY;Z3kDdLV;Z3lDfLT;W3PEgLQ;W3REhLn:V3UEhLl:W3UEiLk:W3VEgLk:X3WEgLi:Y3WEfLj:Y3XEfLh:Z3YEdLh:[3ZEdLf:\\3ZEcLg:\\3[EcLe:]3\\EaLe:_3\\E`Ld:_3_E^Lb:b3`E\\L`:c3bE[L_:e3cEYL]:f3eEXL\\:h3fEVLZ:i3hEULY:k3iESLW:l3lEQLU:o3lEPLT:o3oEnKR:R4n2O11O001O001O00001O001O001O001O001O00001O001O001O001N1M3M4L3M5L3L4L5K4M3L4L5K4mBUJm;n5PDYJh;l5UDZJc;j5ZD\\J_;h5^D^J\\;e5bD]JZ;h5bDZJ`;e5]D^Jf;_5XDcJj;[5SDgJP<X5lCkJV<_62M3N3M2M3N2M3M4L3L2O1O0O10000O10000O1000000O2O000O10000O1001O000001O01O0000001O0000001O0001O0001O0000001O00000010O000001O000000001O0hL\\ITId6i6WJ_Hi5^7QKlGo4P8kKYGU4d8eLeF[3X9kLeFV3W9nLhFS3T9P4K5K4K6K4L5K5K4L5K4L5K4L5K5H7I8H7I8I7H7I8H7I8I6I8K5L3N3L3M4L3N3L4L3N3L3M4M3L3M4M2M4L5Ke0\\Od0[Oe0[Of0[OaD"}, "label": "a man with his tongue sticking out , wearing a nice jacket , tie , and hat"}]}
{"id": 250588, "image": "COCO_train2014_000000250588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing checked blazer standing with her colleague\"."}], "task": "refering", "answer_template": "The \"a man wearing checked blazer standing with her colleague\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 73, 74, 75, 76, 90, 91, 92, 93, 105, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 213, 223, 224, 225, 226, 227, 228, 229, 230, 240, 241, 242, 243, 244, 245, 246, 247, 257, 258, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 330, 331, 341, 342, 343, 344, 345, 346, 347, 348, 358, 359, 360, 361, 362, 363, 364, 365, 376, 377, 378, 380, 381, 382], "bbox": [0.11385416666666666, 0.152953125, 0.5363333333333333, 0.982171875], "iscrowd": 0, "area": 79209.28940000001, "rle": {"size": [640, 480], "counts": "_lR1j2Ra04M4K4L4L4M3]ImLULW3]3]MYLh2X3mM]LW2]3UNYLo1c3\\NSLg1j3bNlKb1Q4gNeK]1X4mNUK`1g4iN_Jk1^5^NSJk1j5_NgIi1V6`N[Ii1b6aNnHi1n6`NcHi1Z7`5L4M3M3M4L3N2O1N2O2N1N2O1O1O1N3N1O1N2O1O101O000001O01O00000000001O01O000001O01O000000001M2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N1O2N2NlMgDdKW;W4UEeKg:X4dEcKY:a4kE[KQ:h4TFTKh9P5\\FkJb9X5bFdJZ9`5jF[JS9h5QGUJk8o5YGmIc8W6aGdI\\8`6hG\\IU8f6PHVIl7n6XHmHe7W7^2eJbD\\2a>O101O000O101O000O101O000O101O0O10001O0O10000O2O000O101O0O10001O0O10001O00000010O1XB\\LR;e3mD_Lo:b3PEcLl:\\3SEiLi:X3VEkLg:V3XEkLg:V3WElLi:T3VEnLh:S3WEnLh:S3VEoLi:R3VEoLi:R3VEPMi:P3VEQMi:o2VESMi:n2VETMh:m2WETMi:l2UEVMj:k2UEWMi:j2VEVMj:k2UEVMk:j2SEXMl:i2SEXMl:i2SEXMl:h2SEZMf3VOT1`3UK[Mg3O9g2oK[Mh3h0^OP2gLYMk3f1^NT1cMXMo3k3WLoNGVMS4`4`K^N9SMW4T5jJlMi0TM]4P:]KTFc4n9WKVFi4l9QKXFP5i9iJ\\FW5W<00000000000O1000000000000000mKTCQ1a`0]ObP[4"}, "label": "a man wearing checked blazer standing with her colleague"}]}
{"id": 250588, "image": "COCO_train2014_000000250588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in plaid is wearing a fedora\"."}], "task": "refering", "answer_template": "The \"a man in plaid is wearing a fedora\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 73, 74, 75, 76, 90, 91, 92, 93, 105, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 213, 223, 224, 225, 226, 227, 228, 229, 230, 240, 241, 242, 243, 244, 245, 246, 247, 257, 258, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 330, 331, 341, 342, 343, 344, 345, 346, 347, 348, 358, 359, 360, 361, 362, 363, 364, 365, 376, 377, 378, 380, 381, 382], "bbox": [0.11385416666666666, 0.152953125, 0.5363333333333333, 0.982171875], "iscrowd": 0, "area": 79209.28940000001, "rle": {"size": [640, 480], "counts": "_lR1j2Ra04M4K4L4L4M3]ImLULW3]3]MYLh2X3mM]LW2]3UNYLo1c3\\NSLg1j3bNlKb1Q4gNeK]1X4mNUK`1g4iN_Jk1^5^NSJk1j5_NgIi1V6`N[Ii1b6aNnHi1n6`NcHi1Z7`5L4M3M3M4L3N2O1N2O2N1N2O1O1O1N3N1O1N2O1O101O000001O01O00000000001O01O000001O01O000000001M2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N1O2N2NlMgDdKW;W4UEeKg:X4dEcKY:a4kE[KQ:h4TFTKh9P5\\FkJb9X5bFdJZ9`5jF[JS9h5QGUJk8o5YGmIc8W6aGdI\\8`6hG\\IU8f6PHVIl7n6XHmHe7W7^2eJbD\\2a>O101O000O101O000O101O000O101O0O10001O0O10000O2O000O101O0O10001O0O10001O00000010O1XB\\LR;e3mD_Lo:b3PEcLl:\\3SEiLi:X3VEkLg:V3XEkLg:V3WElLi:T3VEnLh:S3WEnLh:S3VEoLi:R3VEoLi:R3VEPMi:P3VEQMi:o2VESMi:n2VETMh:m2WETMi:l2UEVMj:k2UEWMi:j2VEVMj:k2UEVMk:j2SEXMl:i2SEXMl:i2SEXMl:h2SEZMf3VOT1`3UK[Mg3O9g2oK[Mh3h0^OP2gLYMk3f1^NT1cMXMo3k3WLoNGVMS4`4`K^N9SMW4T5jJlMi0TM]4P:]KTFc4n9WKVFi4l9QKXFP5i9iJ\\FW5W<00000000000O1000000000000000mKTCQ1a`0]ObP[4"}, "label": "a man in plaid is wearing a fedora"}]}
{"id": 447197, "image": "COCO_train2014_000000447197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl wearing the paper crown\"."}], "task": "refering", "answer_template": "The \"the girl wearing the paper crown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.298390625, 0.1419375, 0.8129062499999999, 0.7720416666666666], "iscrowd": 0, "area": 51399.59750000001, "rle": {"size": [480, 640], "counts": "Wki27f>=B>B9G5L4K4L5K4M4L3O2N1O2O0O2N2N1O2N1O2O0O2N1O1N2O1O0O2O001N2O0O2O1O0O2O001N2O0O2L4M2M4M3L3N3L3N3L4M2M400O0100O100O100O010O100O100O01N101N2N101N101N101N101N1dLXK`Lk4`3_KTLj4e3`KnKc4R4hKaKY4`4QLSKQ4_5hKUJY4f5WLmIk3S6aL`I`3a6jLRIX3n6]21O1O1O1O001O1O00000000O1000000O100000O01000000O1000000O1000000O100000000O10000YOTHZJl7c5XH[Ji7b5[H\\Jf7a5_HYJe7e5^HWJe7i5\\HSJg7m5f00000001O00000000000O1000000000000000f0ZO9G1O000O101N10001N100O101N100O101O0O101N100000N2L5L3M3K5K5L5J5N21O0010O01O001O001O00:FP1POQ1oNg0YO1O2N1O1O2N1O2N1O2N1O2N1N2O2N1O2N1O2N1O1O2N1O2N1O2N2N2N2N2N2N2N2N2N2O1N2N2N2N2O4K6J5K10O0O1000001N1000@UDkMk;U2YDhMe;Y2_DcMa;]2dD_M[;a2iD^MT;c2c0100001O1O010O1O001O1O010O1O001O100O001O1O0010O0001O000010O000001O00010O0O101O0O101O0O101O0O101N10000O2O000O2O000O2M2O1N3N2N5J5L4K5L4L4K6K4K5L4LYng1"}, "label": "the girl wearing the paper crown"}]}
{"id": 447197, "image": "COCO_train2014_000000447197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl in a pink jacket and ' party time ' paper crown\"."}], "task": "refering", "answer_template": "The \"little girl in a pink jacket and ' party time ' paper crown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.298390625, 0.1419375, 0.8129062499999999, 0.7720416666666666], "iscrowd": 0, "area": 51399.59750000001, "rle": {"size": [480, 640], "counts": "Wki27f>=B>B9G5L4K4L5K4M4L3O2N1O2O0O2N2N1O2N1O2O0O2N1O1N2O1O0O2O001N2O0O2O1O0O2O001N2O0O2L4M2M4M3L3N3L3N3L4M2M400O0100O100O100O010O100O100O01N101N2N101N101N101N101N1dLXK`Lk4`3_KTLj4e3`KnKc4R4hKaKY4`4QLSKQ4_5hKUJY4f5WLmIk3S6aL`I`3a6jLRIX3n6]21O1O1O1O001O1O00000000O1000000O100000O01000000O1000000O1000000O100000000O10000YOTHZJl7c5XH[Ji7b5[H\\Jf7a5_HYJe7e5^HWJe7i5\\HSJg7m5f00000001O00000000000O1000000000000000f0ZO9G1O000O101N10001N100O101N100O101O0O101N100000N2L5L3M3K5K5L5J5N21O0010O01O001O001O00:FP1POQ1oNg0YO1O2N1O1O2N1O2N1O2N1O2N1N2O2N1O2N1O2N1O1O2N1O2N1O2N2N2N2N2N2N2N2N2N2O1N2N2N2N2O4K6J5K10O0O1000001N1000@UDkMk;U2YDhMe;Y2_DcMa;]2dD_M[;a2iD^MT;c2c0100001O1O010O1O001O1O010O1O001O100O001O1O0010O0001O000010O000001O00010O0O101O0O101O0O101O0O101N10000O2O000O2O000O2M2O1N3N2N5J5L4K5L4L4K6K4K5L4LYng1"}, "label": "little girl in a pink jacket and ' party time ' paper crown"}]}
{"id": 512734, "image": "COCO_train2014_000000512734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with girl scatting in a snow\"."}], "task": "refering", "answer_template": "The \"a man with girl scatting in a snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 172, 173, 174, 175, 176, 197, 198, 199, 200, 221, 222, 223, 224, 245], "bbox": [0.49967187500000004, 0.22065420560747662, 0.759328125, 0.6879205607476636], "iscrowd": 0, "area": 13590.608950000002, "rle": {"size": [428, 640], "counts": "omU4a0a<?E:F6K5J6K5N2O1O1O1O1O1O1ON1O2N2N101O001N]Od0I71O1O2N1O2M26J5L5J5L2M3M3N2M3O1N9G8I0O1O100O1O100O1O100O100000000000000001O00000000000000001O0000O1O1O10_LQGd2n8]MSGb2l8^MUGb2k8]MWGb2h8_MYG`2f8`M\\G_2c8aM^G_2b8`M`G_2_8bMbG]2]8cMeG\\2Z8dMhG[2X8eMhG[2W8eMkGZ2T8fMnGY2R8fMPHY2o7hMQHW1[OlNc8MTHR1\\OQO`8MWHm0ZOXO_8KYHh0YO_O]8I]He0VOB]8I_Hd0ROE_8FbHc0nNH`8EbHc0nNGa8EcHc0kNIb8CdHd0iNId8CcHc0kNIb8CeHc0jNI`8DgHc0jNG`8FgHb0jNG_8FhHc0jNE_8GkHa0gNG^8EPIb0bNI^8CTIb0_NI^8BXIc0[NJ]8A\\Ic0XNJ]8@^If0VNI\\8\\OdIj0QNIZ8ZOiIm0nMGl:9UEFk:9VEFk::VEEi:;XEEh:;YECh:<YEBh:?YE^Oi:a0XE]Oj:a0YE\\Oi:c0YEZOh:f0YEXOi:g0o0O001O1O001O1O1O001000O101O0O10000O10000O2O000O10000O10O1N101N101O0O2O1N101O0O5L7H\\SP2"}, "label": "a man with girl scatting in a snow"}]}
{"id": 512734, "image": "COCO_train2014_000000512734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue jacket and purple pants skiis\"."}], "task": "refering", "answer_template": "The \"a woman in a blue jacket and purple pants skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 172, 173, 174, 175, 176, 197, 198, 199, 200, 221, 222, 223, 224, 245], "bbox": [0.49967187500000004, 0.22065420560747662, 0.759328125, 0.6879205607476636], "iscrowd": 0, "area": 13590.608950000002, "rle": {"size": [428, 640], "counts": "omU4a0a<?E:F6K5J6K5N2O1O1O1O1O1O1ON1O2N2N101O001N]Od0I71O1O2N1O2M26J5L5J5L2M3M3N2M3O1N9G8I0O1O100O1O100O1O100O100000000000000001O00000000000000001O0000O1O1O10_LQGd2n8]MSGb2l8^MUGb2k8]MWGb2h8_MYG`2f8`M\\G_2c8aM^G_2b8`M`G_2_8bMbG]2]8cMeG\\2Z8dMhG[2X8eMhG[2W8eMkGZ2T8fMnGY2R8fMPHY2o7hMQHW1[OlNc8MTHR1\\OQO`8MWHm0ZOXO_8KYHh0YO_O]8I]He0VOB]8I_Hd0ROE_8FbHc0nNH`8EbHc0nNGa8EcHc0kNIb8CdHd0iNId8CcHc0kNIb8CeHc0jNI`8DgHc0jNG`8FgHb0jNG_8FhHc0jNE_8GkHa0gNG^8EPIb0bNI^8CTIb0_NI^8BXIc0[NJ]8A\\Ic0XNJ]8@^If0VNI\\8\\OdIj0QNIZ8ZOiIm0nMGl:9UEFk:9VEFk::VEEi:;XEEh:;YECh:<YEBh:?YE^Oi:a0XE]Oj:a0YE\\Oi:c0YEZOh:f0YEXOi:g0o0O001O1O001O1O1O001000O101O0O10000O10000O2O000O10000O10O1N101N101O0O2O1N101O0O5L7H\\SP2"}, "label": "a woman in a blue jacket and purple pants skiis"}]}
{"id": 512734, "image": "COCO_train2014_000000512734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with orange goggles who is skiing behind another girl\"."}], "task": "refering", "answer_template": "The \"a girl with orange goggles who is skiing behind another girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 213, 214], "bbox": [0.197671875, 0.33294392523364486, 0.365109375, 0.6429672897196261], "iscrowd": 0, "area": 5754.610549999999, "rle": {"size": [428, 640], "counts": "fYe15U=6J2O2M2N3NKZCFd<;41M3M5K4K5L4M30O2O00JnNSDQ1l;8O0101N@TOcDl0];a00O100O21N6L4K3M2N2N3M7J2M2N2N2O1O1O1O1N11O9GO1O1O1O1O1O2N2N2N2N001O01O0ROPFaNo9`1PF`NQ:KlEn04VOQ:LQFh0N\\OP:LSFh0M[O1ZOh9c0[Fg0L\\OOZOk9c0[Ff0L\\OL]On9a0ZFf0L[OK^OQ:`0YFg0L^Ol9KXFg0L\\Om9MYFg0I[Oo9L[Fh0FZOR:K[Fj0CZOS:J^Fj0@YOT:K_Fk0]OYOV:I`Fm0ZOXOX:EfFQ1QOZOc;d0]DZOi;b0VD^OQ<<oCCS<<lCCV<<kCCW<;iCDY<<gC_O^<`08O1O010O1O1O001O10O01O0O101N100O2N100O2O1N2O^^Y5"}, "label": "a girl with orange goggles who is skiing behind another girl"}]}
{"id": 512734, "image": "COCO_train2014_000000512734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in purple shirt and black pant is skiing behind the other\"."}], "task": "refering", "answer_template": "The \"a lady in purple shirt and black pant is skiing behind the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 213, 214], "bbox": [0.197671875, 0.33294392523364486, 0.365109375, 0.6429672897196261], "iscrowd": 0, "area": 5754.610549999999, "rle": {"size": [428, 640], "counts": "fYe15U=6J2O2M2N3NKZCFd<;41M3M5K4K5L4M30O2O00JnNSDQ1l;8O0101N@TOcDl0];a00O100O21N6L4K3M2N2N3M7J2M2N2N2O1O1O1O1N11O9GO1O1O1O1O1O2N2N2N2N001O01O0ROPFaNo9`1PF`NQ:KlEn04VOQ:LQFh0N\\OP:LSFh0M[O1ZOh9c0[Fg0L\\OOZOk9c0[Ff0L\\OL]On9a0ZFf0L[OK^OQ:`0YFg0L^Ol9KXFg0L\\Om9MYFg0I[Oo9L[Fh0FZOR:K[Fj0CZOS:J^Fj0@YOT:K_Fk0]OYOV:I`Fm0ZOXOX:EfFQ1QOZOc;d0]DZOi;b0VD^OQ<<oCCS<<lCCV<<kCCW<;iCDY<<gC_O^<`08O1O010O1O1O001O10O01O0O101N100O2N100O2O1N2O^^Y5"}, "label": "a lady in purple shirt and black pant is skiing behind the other"}]}
{"id": 365284, "image": "COCO_train2014_000000365284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan umbrella next to a table with no umbrella\"."}], "task": "refering", "answer_template": "The \"a tan umbrella next to a table with no umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 80], "bbox": [0.323984375, 0.0, 0.6816093750000001, 0.2649648711943794], "iscrowd": 0, "area": 11820.4779, "rle": {"size": [427, 640], "counts": "mgf24V=2N1O1O1O1O1O1O1O1O1N3O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O2N1O1O1O1O00O100000000O100000000O100000000O1000000O100000000O1000QFQNb8o1PG^NQ9b1`FmN`9^21H8E;D<E;O10000O1000000000000O1000000000000O100000000000O01000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000BPEaNP;X1e0L4M3M3N2M3M3N2O1N2N2N2N2O1N2O1N2O1N2N2OURe2"}, "label": "a tan umbrella next to a table with no umbrella"}]}
{"id": 365284, "image": "COCO_train2014_000000365284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella on the right\"."}], "task": "refering", "answer_template": "The \"the umbrella on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 31, 32, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 80], "bbox": [0.323984375, 0.0, 0.6816093750000001, 0.2649648711943794], "iscrowd": 0, "area": 11820.4779, "rle": {"size": [427, 640], "counts": "mgf24V=2N1O1O1O1O1O1O1O1O1N3O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O2N1O1O1O1O00O100000000O100000000O100000000O1000000O100000000O1000QFQNb8o1PG^NQ9b1`FmN`9^21H8E;D<E;O10000O1000000000000O1000000000000O100000000000O01000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000BPEaNP;X1e0L4M3M3N2M3M3N2O1N2N2N2N2O1N2O1N2O1N2N2OURe2"}, "label": "the umbrella on the right"}]}
{"id": 176871, "image": "COCO_train2014_000000176871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black refrigerator beside a stainless steel one\"."}], "task": "refering", "answer_template": "The \"a black refrigerator beside a stainless steel one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 64, 65, 66, 67, 81, 82, 83, 84, 98, 99, 100, 101, 115, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 302, 303], "bbox": [0.7550625, 0.089890625, 1.0, 0.7438125], "iscrowd": 0, "area": 46167.768800000005, "rle": {"size": [640, 480], "counts": "hfR74bb0Z1fNZ1fNZ1fNZ1gNY1fNZ1fNZ1fNZ1fNZ1kNU10000000O10000000000000000000000000000000000000000000000000000O1000001O00000000000000000000000000000000000000000000000O10000000000000000000000000001O00000000000000000000000O1000000000000000000000000000000000000@`0RKn4QKS3"}, "label": "a black refrigerator beside a stainless steel one"}]}
{"id": 176871, "image": "COCO_train2014_000000176871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black fridge to the right\"."}], "task": "refering", "answer_template": "The \"the black fridge to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 64, 65, 66, 67, 81, 82, 83, 84, 98, 99, 100, 101, 115, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 302, 303], "bbox": [0.7550625, 0.089890625, 1.0, 0.7438125], "iscrowd": 0, "area": 46167.768800000005, "rle": {"size": [640, 480], "counts": "hfR74bb0Z1fNZ1fNZ1fNZ1gNY1fNZ1fNZ1fNZ1fNZ1kNU10000000O10000000000000000000000000000000000000000000000000000O1000001O00000000000000000000000000000000000000000000000O10000000000000000000000000001O00000000000000000000000O1000000000000000000000000000000000000@`0RKn4QKS3"}, "label": "the black fridge to the right"}]}
{"id": 176871, "image": "COCO_train2014_000000176871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a two - door stainless steel refrigerator with ice and water dispenser in the left door\"."}], "task": "refering", "answer_template": "The \"a two - door stainless steel refrigerator with ice and water dispenser in the left door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 343, 344, 345, 346, 347], "bbox": [0.1498125, 0.085546875, 0.7490624999999999, 0.8967812500000001], "iscrowd": 0, "area": 134027.47079999998, "rle": {"size": [640, 480], "counts": "UR]17hc0?A>B?A?A>B?A?A?A>B?B>A>B?A?B=C>B>B>B=C>B>B=C>B>B=C>B>B>B=C>B>B=C>B>B=C6J3M2N3M2N3M000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O010000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000UOdKRB\\4j<lLQCU3k<R2O100O10000O100O10l3SL\\4eKZ4eKiaZ2"}, "label": "a two - door stainless steel refrigerator with ice and water dispenser in the left door"}]}
{"id": 176871, "image": "COCO_train2014_000000176871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stainless steel refrigetator\"."}], "task": "refering", "answer_template": "The \"a stainless steel refrigetator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 343, 344, 345, 346, 347], "bbox": [0.1498125, 0.085546875, 0.7490624999999999, 0.8967812500000001], "iscrowd": 0, "area": 134027.47079999998, "rle": {"size": [640, 480], "counts": "UR]17hc0?A>B?A?A>B?A?A?A>B?B>A>B?A?B=C>B>B>B=C>B>B=C>B>B=C>B>B>B=C>B>B=C>B>B=C6J3M2N3M2N3M000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O010000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000UOdKRB\\4j<lLQCU3k<R2O100O10000O100O10l3SL\\4eKZ4eKiaZ2"}, "label": "a stainless steel refrigetator"}]}
{"id": 176871, "image": "COCO_train2014_000000176871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of fridge on the left , rest of fridge not pictured\"."}], "task": "refering", "answer_template": "The \"part of fridge on the left , rest of fridge not pictured\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 51, 52, 53, 68, 69, 70, 85, 86, 87, 102, 103, 104, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325, 326, 340, 341, 342, 343], "bbox": [0.006, 0.092140625, 0.21272916666666666, 0.896625], "iscrowd": 0, "area": 42482.235049999996, "rle": {"size": [640, 480], "counts": "lm1Q7o<Q9oF0000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000000000000002N`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0_Oa0@`0@a0_O`0@`0@`0@_f[7"}, "label": "part of fridge on the left , rest of fridge not pictured"}]}
{"id": 176871, "image": "COCO_train2014_000000176871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light gray refrigerator mostly out of frame\"."}], "task": "refering", "answer_template": "The \"light gray refrigerator mostly out of frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 51, 52, 53, 68, 69, 70, 85, 86, 87, 102, 103, 104, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325, 326, 340, 341, 342, 343], "bbox": [0.006, 0.092140625, 0.21272916666666666, 0.896625], "iscrowd": 0, "area": 42482.235049999996, "rle": {"size": [640, 480], "counts": "lm1Q7o<Q9oF0000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000000000000002N`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0_Oa0@`0@a0_O`0@`0@`0@_f[7"}, "label": "light gray refrigerator mostly out of frame"}]}
{"id": 504554, "image": "COCO_train2014_000000504554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of yellow banana\"."}], "task": "refering", "answer_template": "The \"a piece of yellow banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 20, 21, 22, 23, 24, 25, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 129, 130, 131, 132, 133], "bbox": [0.13236, 0.004853333333333334, 0.45530000000000004, 0.6213333333333333], "iscrowd": 0, "area": 31869.3615, "rle": {"size": [375, 500], "counts": "TYh04[;>A>B>C=B>B?B=B>F:J6I7I4M2M2O2M3M3N2M3N2M3N2M2N3N2M3N2M3N2M3M3N1N3N2M3N2O1O1O1O001O1O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O00001O0000001O00001O0000001O00001O00001O000000O100000000O100000000O1000000O100000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O1O1O8H9G9G9H8G9G9G9G9G9G9G9G9cN]1TNcaS3"}, "label": "a piece of yellow banana"}]}
{"id": 504554, "image": "COCO_train2014_000000504554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana which is in the center\"."}], "task": "refering", "answer_template": "The \"a banana which is in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 25, 26, 27, 28, 29, 43, 44, 45, 46, 47, 48, 62, 63, 64, 65, 66, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 172, 173, 174], "bbox": [0.43236, 0.0023466666666666666, 0.7553, 0.7246933333333333], "iscrowd": 0, "area": 32072.50975, "rle": {"size": [375, 500], "counts": "UT_2a0]:P1I7I7I7I7I7I6J7oISMj2S3dL\\MY3k2SLdMk3b2cKlM[4[2QKTNm4]4N2N2N3M2N2N2N2N2N3M2M3N2N1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O0O2O00001O001O00001O001O001O00001O001O001O001O001O001N101O1O001O001O001O1O001O001O0000O102N3M4L3M4L3M4L3M3L5L3M2N3M2N3M2N3M3M2M4M2N3M2N3M2N3M2N3M2M4M2N3M2N3M2N3M3M2N3L3N3L4K4L5L4K5L3L5L4K5K4M4K5L4K7J6I7I8I6I7J7H7J6I7J7H7I7Jej\\1"}, "label": "a banana which is in the center"}]}
{"id": 504554, "image": "COCO_train2014_000000504554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana in the middle . it has a sticker on it\"."}], "task": "refering", "answer_template": "The \"the banana in the middle . it has a sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 25, 26, 27, 28, 29, 43, 44, 45, 46, 47, 48, 62, 63, 64, 65, 66, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 172, 173, 174], "bbox": [0.43236, 0.0023466666666666666, 0.7553, 0.7246933333333333], "iscrowd": 0, "area": 32072.50975, "rle": {"size": [375, 500], "counts": "UT_2a0]:P1I7I7I7I7I7I6J7oISMj2S3dL\\MY3k2SLdMk3b2cKlM[4[2QKTNm4]4N2N2N3M2N2N2N2N2N3M2M3N2N1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O0O2O00001O001O00001O001O001O00001O001O001O001O001O001N101O1O001O001O001O1O001O001O0000O102N3M4L3M4L3M4L3M3L5L3M2N3M2N3M2N3M3M2M4M2N3M2N3M2N3M2N3M2M4M2N3M2N3M2N3M3M2N3L3N3L4K4L5L4K5L3L5L4K5K4M4K5L4K7J6I7I8I6I7J7H7J6I7J7H7I7Jej\\1"}, "label": "the banana in the middle . it has a sticker on it"}]}
{"id": 504554, "image": "COCO_train2014_000000504554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bananna on the right\"."}], "task": "refering", "answer_template": "The \"the bananna on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 29, 30, 31, 32, 47, 48, 49, 50, 66, 67, 68, 69, 84, 85, 86, 87, 102, 103, 104, 105, 121, 122, 123, 139, 140, 141], "bbox": [0.61588, 0.01176, 0.86294, 0.62824], "iscrowd": 0, "area": 16639.84465, "rle": {"size": [375, 500], "counts": "`i`32e;3M3M3M3M3M3M2N3M3M3N2M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M4L3M4L3M4L3M4L4L3M4L3M4L4L3M4L3M4L4L4L5K4L5K4L5K4L5K4L5K2N1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0000O100O10001N5L4K5L4K6K4K5L4K5L4K5L4J6K5J6J6K5J6J6K5J6J6K5J7I7J5J7I7J6G9E:E<D<E;DRTi0"}, "label": "the bananna on the right"}]}
{"id": 504554, "image": "COCO_train2014_000000504554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rearmost banana out of focus\"."}], "task": "refering", "answer_template": "The \"the rearmost banana out of focus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 29, 30, 31, 32, 47, 48, 49, 50, 66, 67, 68, 69, 84, 85, 86, 87, 102, 103, 104, 105, 121, 122, 123, 139, 140, 141], "bbox": [0.61588, 0.01176, 0.86294, 0.62824], "iscrowd": 0, "area": 16639.84465, "rle": {"size": [375, 500], "counts": "`i`32e;3M3M3M3M3M3M2N3M3M3N2M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M4L3M4L3M4L3M4L4L3M4L3M4L4L3M4L3M4L4L4L5K4L5K4L5K4L5K4L5K2N1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0000O100O10001N5L4K5L4K6K4K5L4K5L4K5L4J6K5J6J6K5J6J6K5J6J6K5J7I7J5J7I7J6G9E:E<D<E;DRTi0"}, "label": "the rearmost banana out of focus"}]}
{"id": 471788, "image": "COCO_train2014_000000471788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long pair of scissors\"."}], "task": "refering", "answer_template": "The \"a long pair of scissors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 134, 135, 152, 153, 169, 170, 171, 187, 188, 189, 202, 203, 204, 206, 207, 220, 221, 222, 224, 225, 226, 243, 244, 261, 262], "bbox": [0.24944, 0.34382, 0.59776, 0.8], "iscrowd": 0, "area": 7518.47975, "rle": {"size": [500, 500], "counts": "f[m12`?4L3M4L3M4L3M3M4M2M4L3N3N1O2M0100000000000000000000O010000000000000000000O0O3L4K6K4K6J6K4K6Kh@9n>2O1O2N11O100O1O2N100O1000O10O01O100O10O01O100O010O1O01O010O01O01O010O00010O01O01O010O01O01O010O00010O010O00010O0010O0010O00010OmCBR9?oFBn8?SGBj8>WGDg8<ZGEc8<^GE_8;bGG[8:fGGX89iGHT88nGIo78`G;]8EYGh0e8XOQGT1l8mNiF`1T9`NbFm1[9TNZFY2`9kMVF`2e9cMQFj2h9[MmEn3P9_1M2O2O00ZNRGgLl8e2jGZMV8R2_HnM`7]1VIbNi6j0mIVOR6CVK<j4AYK`0e4_O^K`0b4]OaKd0^4YOeKg0[4VOhKk0V4SOnKl0R4QOQLP1n3mNULS1j3lNXLU1g3hN\\LX1d3fN^L[1`3cNcL]1]3aNfL]1[3aNgL^1Y3`NjL_1W3_NkL`1e800O1O1O2N100O1O1O1O101N1O1O1O100O1Og0YORiQ3"}, "label": "a long pair of scissors"}]}
{"id": 471788, "image": "COCO_train2014_000000471788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gold - handled longest scissors\"."}], "task": "refering", "answer_template": "The \"gold - handled longest scissors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 134, 135, 152, 153, 169, 170, 171, 187, 188, 189, 202, 203, 204, 206, 207, 220, 221, 222, 224, 225, 226, 243, 244, 261, 262], "bbox": [0.24944, 0.34382, 0.59776, 0.8], "iscrowd": 0, "area": 7518.47975, "rle": {"size": [500, 500], "counts": "f[m12`?4L3M4L3M4L3M3M4M2M4L3N3N1O2M0100000000000000000000O010000000000000000000O0O3L4K6K4K6J6K4K6Kh@9n>2O1O2N11O100O1O2N100O1000O10O01O100O10O01O100O010O1O01O010O01O01O010O00010O01O01O010O01O01O010O00010O010O00010O0010O0010O00010OmCBR9?oFBn8?SGBj8>WGDg8<ZGEc8<^GE_8;bGG[8:fGGX89iGHT88nGIo78`G;]8EYGh0e8XOQGT1l8mNiF`1T9`NbFm1[9TNZFY2`9kMVF`2e9cMQFj2h9[MmEn3P9_1M2O2O00ZNRGgLl8e2jGZMV8R2_HnM`7]1VIbNi6j0mIVOR6CVK<j4AYK`0e4_O^K`0b4]OaKd0^4YOeKg0[4VOhKk0V4SOnKl0R4QOQLP1n3mNULS1j3lNXLU1g3hN\\LX1d3fN^L[1`3cNcL]1]3aNfL]1[3aNgL^1Y3`NjL_1W3_NkL`1e800O1O1O2N100O1O1O1O101N1O1O1O100O1Og0YORiQ3"}, "label": "gold - handled longest scissors"}]}
{"id": 155645, "image": "COCO_train2014_000000155645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the body of the person holding the colorful umbrella\"."}], "task": "refering", "answer_template": "The \"the body of the person holding the colorful umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [300, 301, 322, 323, 324, 345, 346, 347, 348, 349, 350, 367, 368, 369, 370, 371, 372, 389, 390, 391, 392, 393, 394, 412, 413, 414, 415, 416, 435, 436, 437, 438, 457, 458, 459, 460, 480, 481, 482], "bbox": [0.6696568627450981, 0.6179738562091504, 0.9348366013071896, 0.9932516339869281], "iscrowd": 0, "area": 19846.864150000005, "rle": {"size": [612, 612], "counts": "R]e75Pc06I8H7J3L1O2O0O2N101N1O101N1O2N101N1O101N2N101N2N2^_O^N_?b1]@dN`?^1[@hNb?Z1X@lNg?U1T@QOi?P1S@UOm?S2N1O2O0O0010O00001O01O01O00010O00001O01O0O1O1O2N100O1O1O2N100O1O1O2N100O1O1O2N100O1O1O2N100O1O1O2N2O1N2N2N2O1N2N3M2N2O1Nh0XO00000000000010O00000000000001O01O000000000001O0001O0000000001O00000001O000001O0000000001O0001O000000000000001O00000000dKbBR3^=lLoBi2Q=UM\\C`2d<^MiCX2V<gMgF]OY9a0TGTOl8j0h4N2N2N2En_g0"}, "label": "the body of the person holding the colorful umbrella"}]}
{"id": 155645, "image": "COCO_train2014_000000155645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holds a red bule yellow and green umbrella wearing a coat\"."}], "task": "refering", "answer_template": "The \"woman holds a red bule yellow and green umbrella wearing a coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [300, 301, 322, 323, 324, 345, 346, 347, 348, 349, 350, 367, 368, 369, 370, 371, 372, 389, 390, 391, 392, 393, 394, 412, 413, 414, 415, 416, 435, 436, 437, 438, 457, 458, 459, 460, 480, 481, 482], "bbox": [0.6696568627450981, 0.6179738562091504, 0.9348366013071896, 0.9932516339869281], "iscrowd": 0, "area": 19846.864150000005, "rle": {"size": [612, 612], "counts": "R]e75Pc06I8H7J3L1O2O0O2N101N1O101N1O2N101N1O101N2N101N2N2^_O^N_?b1]@dN`?^1[@hNb?Z1X@lNg?U1T@QOi?P1S@UOm?S2N1O2O0O0010O00001O01O01O00010O00001O01O0O1O1O2N100O1O1O2N100O1O1O2N100O1O1O2N100O1O1O2N100O1O1O2N2O1N2N2N2O1N2N3M2N2O1Nh0XO00000000000010O00000000000001O01O000000000001O0001O0000000001O00000001O000001O0000000001O0001O000000000000001O00000000dKbBR3^=lLoBi2Q=UM\\C`2d<^MiCX2V<gMgF]OY9a0TGTOl8j0h4N2N2N2En_g0"}, "label": "woman holds a red bule yellow and green umbrella wearing a coat"}]}
{"id": 78578, "image": "COCO_train2014_000000078578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in wedding gown cutting a cake\"."}], "task": "refering", "answer_template": "The \"woman in wedding gown cutting a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 84, 85, 86, 87, 99, 100, 101, 102, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 233, 234, 237, 238, 239, 252, 253], "bbox": [0.5468075117370892, 0.174953125, 0.9681690140845071, 0.7196093749999999], "iscrowd": 0, "area": 42164.28384999998, "rle": {"size": [640, 426], "counts": "hPb4e08BP`0h0S@P1LXNo?U3P@lLm?W3Q@jLm?Y3R@hLi?]3U@dLc?Q4G8H8H9G8H9H7H8H8L5N1O1N2O2N1N101O0O2O0[OTIXDm6_;aI\\D^6\\;PJ`Do5_;VJ^Dh5b;\\J\\Dc5c;bJZD\\5f;hJXDX5f;b1N1O2N2N2N2N2N2N2N2N2N2N2NO1M31O101000O100O10000O010O10000O1000O0100O10000O10O010000O100O10O10O10001N2O1N101O1N2O1O100000000000O1000001O0000000000000O100000000001O10O01O1O001O1O1O001O1O001O13L5L3MbHcDZ6Y;bIjDd6R;XIQEl6l:PIWEP7k:jHYEU7i:gHZEW7j:dHYEY7k:bHYEZ7i;N2N2O1O0O2O1O1N2O1O1O1O2N1O100O1O2N1PKiBU3X=eLoBY3R=bLSC]3o<]LVCa3o<WLWCe3n<TLWCi3o<oKVCm3P=lKUCQ4Q=gKUCU4]>L4M4L3J6E;G:G8G9H<D=B>C>Bc]8"}, "label": "woman in wedding gown cutting a cake"}]}
{"id": 78578, "image": "COCO_train2014_000000078578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman that is cutting a cake while wearing a dress\"."}], "task": "refering", "answer_template": "The \"a woman that is cutting a cake while wearing a dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 84, 85, 86, 87, 99, 100, 101, 102, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 233, 234, 237, 238, 239, 252, 253], "bbox": [0.5468075117370892, 0.174953125, 0.9681690140845071, 0.7196093749999999], "iscrowd": 0, "area": 42164.28384999998, "rle": {"size": [640, 426], "counts": "hPb4e08BP`0h0S@P1LXNo?U3P@lLm?W3Q@jLm?Y3R@hLi?]3U@dLc?Q4G8H8H9G8H9H7H8H8L5N1O1N2O2N1N101O0O2O0[OTIXDm6_;aI\\D^6\\;PJ`Do5_;VJ^Dh5b;\\J\\Dc5c;bJZD\\5f;hJXDX5f;b1N1O2N2N2N2N2N2N2N2N2N2N2NO1M31O101000O100O10000O010O10000O1000O0100O10000O10O010000O100O10O10O10001N2O1N101O1N2O1O100000000000O1000001O0000000000000O100000000001O10O01O1O001O1O1O001O1O001O13L5L3MbHcDZ6Y;bIjDd6R;XIQEl6l:PIWEP7k:jHYEU7i:gHZEW7j:dHYEY7k:bHYEZ7i;N2N2O1O0O2O1O1N2O1O1O1O2N1O100O1O2N1PKiBU3X=eLoBY3R=bLSC]3o<]LVCa3o<WLWCe3n<TLWCi3o<oKVCm3P=lKUCQ4Q=gKUCU4]>L4M4L3J6E;G:G8G9H<D=B>C>Bc]8"}, "label": "a woman that is cutting a cake while wearing a dress"}]}
{"id": 78578, "image": "COCO_train2014_000000078578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man watching a couple cut a cake\"."}], "task": "refering", "answer_template": "The \"a man watching a couple cut a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 105, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 168, 180, 181, 182, 183, 195, 196, 197, 211], "bbox": [0.0020657276995305163, 0.276890625, 0.2102582159624413, 0.6612500000000001], "iscrowd": 0, "area": 12822.567199999996, "rle": {"size": [640, 426], "counts": "_m0U2j>Q3TOm0000000O10O10000000000O10000000000O10O12N5K4L5J6K4L3M00O10O01O100O100O00100O100O1O10O0101N1O101N100O2N101NkNgBdKY=X4Y1N2M4M2M4M2M3N3O1N4L4L4L4M3L5K4L4L4L4M3L4L4L5K4M3L4L4L4L4L4M>Ac0]OZea6"}, "label": "a man watching a couple cut a cake"}]}
{"id": 78578, "image": "COCO_train2014_000000078578.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man to the left of the couple\"."}], "task": "refering", "answer_template": "The \"a man to the left of the couple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 105, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 168, 180, 181, 182, 183, 195, 196, 197, 211], "bbox": [0.0020657276995305163, 0.276890625, 0.2102582159624413, 0.6612500000000001], "iscrowd": 0, "area": 12822.567199999996, "rle": {"size": [640, 426], "counts": "_m0U2j>Q3TOm0000000O10O10000000000O10000000000O10O12N5K4L5J6K4L3M00O10O01O100O100O00100O100O1O10O0101N1O101N100O2N101NkNgBdKY=X4Y1N2M4M2M4M2M3N3O1N4L4L4L4M3L5K4L4L4L4M3L4L4L5K4M3L4L4L4L4L4M>Ac0]OZea6"}, "label": "a man to the left of the couple"}]}
{"id": 119540, "image": "COCO_train2014_000000119540.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a speed boat with writing on it\"."}], "task": "refering", "answer_template": "The \"a speed boat with writing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.15509375, 0.5069064748201438, 0.956671875, 0.7795923261390887], "iscrowd": 0, "area": 39535.082899999994, "rle": {"size": [417, 640], "counts": "maX11o<6K2N1O1O1O1N3N1O1O1O1O2M2O1O1O1O1O2M2O1O1O1O1O2M2O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O001O1O001O1O001O001O1O001O000O10001O000000001O00000O101O000000001O00000O101O000000000000001O0O100000000000000000000O1000000000000000000O100000000000000000000O100000001O00000000000O1000000000000000000O100000000000000000000O100000000000000000001N1000000000000000000O100000000000000O10O0100O100O010O100O10O0100O100O010O1000O10O10000O0100000O10O10O100O10O0100O10O0100O100O10000000000000000O10000000000000000O10000001O0000000000000000000000000000000000000000001O000000000O1000001O00001O00001O00000000000000001N10000000000000000000000000001O000000000O1KoEZMQ:f2oEZMQ:f2PFYMP:g2PFYMP:g2QFXMo9h2QFXMo9h2RFWMn9i2RFWMo9h250001O0000001N1000001O0000001O0000001O0000001O000O101O0000001O0000001O0000001O0000001N1000001O00000000000000O010000000O1000O01000000O1000O100000000000000O1000000000000O10O1000000000O1000000000000O10000000001O00001O0000001O00001O00001O00001O00001O00000O2O00001O00001O00001O0000001O001O0O2O001N101O0O2O001N101O0O2O001N101O0O2O001N101O0O2O001N101O0O2O001N101O1O0O2K4L5K4LlT;"}, "label": "a speed boat with writing on it"}]}
{"id": 119540, "image": "COCO_train2014_000000119540.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the speedboat in the water facing the right\"."}], "task": "refering", "answer_template": "The \"the speedboat in the water facing the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.15509375, 0.5069064748201438, 0.956671875, 0.7795923261390887], "iscrowd": 0, "area": 39535.082899999994, "rle": {"size": [417, 640], "counts": "maX11o<6K2N1O1O1O1N3N1O1O1O1O2M2O1O1O1O1O2M2O1O1O1O1O2M2O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O001O1O001O1O001O001O1O001O000O10001O000000001O00000O101O000000001O00000O101O000000000000001O0O100000000000000000000O1000000000000000000O100000000000000000000O100000001O00000000000O1000000000000000000O100000000000000000000O100000000000000000001N1000000000000000000O100000000000000O10O0100O100O010O100O10O0100O100O010O1000O10O10000O0100000O10O10O100O10O0100O10O0100O100O10000000000000000O10000000000000000O10000001O0000000000000000000000000000000000000000001O000000000O1000001O00001O00001O00000000000000001N10000000000000000000000000001O000000000O1KoEZMQ:f2oEZMQ:f2PFYMP:g2PFYMP:g2QFXMo9h2QFXMo9h2RFWMn9i2RFWMo9h250001O0000001N1000001O0000001O0000001O0000001O000O101O0000001O0000001O0000001O0000001N1000001O00000000000000O010000000O1000O01000000O1000O100000000000000O1000000000000O10O1000000000O1000000000000O10000000001O00001O0000001O00001O00001O00001O00001O00000O2O00001O00001O00001O0000001O001O0O2O001N101O0O2O001N101O0O2O001N101O0O2O001N101O0O2O001N101O0O2O001N101O1O0O2K4L5K4LlT;"}, "label": "the speedboat in the water facing the right"}]}
{"id": 13045, "image": "COCO_train2014_000000013045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bunch of cilantro on top of mash potato\"."}], "task": "refering", "answer_template": "The \"a bunch of cilantro on top of mash potato\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 53, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 169, 170, 171, 193], "bbox": [0.310875, 0.0803125, 0.5751250000000001, 0.47927083333333337], "iscrowd": 0, "area": 18738.35775, "rle": {"size": [480, 640], "counts": "P]m2a0g=j0M3L4M3M3L4M3M3L4M3M3L3N2N3L3N2N2M3N3M2N2N2O1N3M2N2N2N3N1N2N2N2N3N1N2N2N3M2N2ON1MnEoK\\9n3_F\\L_9a3\\FhLc9U3YFSMg9i2TFaMj9k3O1O2N100O2N1O1O2N100O2N1O2O000O2O0O101O0O101O0O01N2O001N2O001O1N101O1N101O1N2O2N2M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N2M3N2N2N2M3N2N1O2N2N2N2O1N1O2N2N2O1N2N1O2N2O1N2N1O2N2O1N2N2N1O2O1N2N2N2N101N2M3N2M3N1O2M3N2M3N2M2O2N2M3N2M3N1N3N2N2M3N2M2O2M3N2M3N2Na\\o3"}, "label": "a bunch of cilantro on top of mash potato"}]}
{"id": 13045, "image": "COCO_train2014_000000013045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"italian parsly on top of mashed potatoes\"."}], "task": "refering", "answer_template": "The \"italian parsly on top of mashed potatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 53, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 169, 170, 171, 193], "bbox": [0.310875, 0.0803125, 0.5751250000000001, 0.47927083333333337], "iscrowd": 0, "area": 18738.35775, "rle": {"size": [480, 640], "counts": "P]m2a0g=j0M3L4M3M3L4M3M3L4M3M3L3N2N3L3N2N2M3N3M2N2N2O1N3M2N2N2N3N1N2N2N2N3N1N2N2N3M2N2ON1MnEoK\\9n3_F\\L_9a3\\FhLc9U3YFSMg9i2TFaMj9k3O1O2N100O2N1O1O2N100O2N1O2O000O2O0O101O0O101O0O01N2O001N2O001O1N101O1N101O1N2O2N2M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N2M3N2N2N2M3N2N1O2N2N2N2O1N1O2N2N2O1N2N1O2N2O1N2N1O2N2O1N2N2N1O2O1N2N2N2N101N2M3N2M3N1O2M3N2M3N2M2O2N2M3N2M3N1N3N2N2M3N2M2O2M3N2M3N2Na\\o3"}, "label": "italian parsly on top of mashed potatoes"}]}
{"id": 13045, "image": "COCO_train2014_000000013045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 front broccoli floweretts\"."}], "task": "refering", "answer_template": "The \"2 front broccoli floweretts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 224, 225, 226, 247, 248, 249, 250, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 319, 320], "bbox": [0.75884375, 0.50825, 0.9498750000000001, 0.7723958333333333], "iscrowd": 0, "area": 8430.06515, "rle": {"size": [480, 640], "counts": "jST76h>8I7H6XOf0E:\\Od0N3M2N3M2N2N3M2N3M1O1O2N1O2N1O2N1O1010O010O01O01O01O010O0100O1O100O1O00100O1O1kL`Dh2P<L5K4M3L1O1O00100O1O001O1O10O01O1O1O10O01O1O1O00100O001O1O10O01O1O1O00100O1O001O100O001O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N2N2N1O2N2N2N2I7Jnd>"}, "label": "2 front broccoli floweretts"}]}
{"id": 13045, "image": "COCO_train2014_000000013045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli florets rest behind the steak knife on the plate\"."}], "task": "refering", "answer_template": "The \"broccoli florets rest behind the steak knife on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 224, 225, 226, 247, 248, 249, 250, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297, 319, 320], "bbox": [0.75884375, 0.50825, 0.9498750000000001, 0.7723958333333333], "iscrowd": 0, "area": 8430.06515, "rle": {"size": [480, 640], "counts": "jST76h>8I7H6XOf0E:\\Od0N3M2N3M2N2N3M2N3M1O1O2N1O2N1O2N1O1010O010O01O01O01O010O0100O1O100O1O00100O1O1kL`Dh2P<L5K4M3L1O1O00100O1O001O1O10O01O1O1O10O01O1O1O00100O001O1O10O01O1O1O00100O1O001O100O001O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N2N2N1O2N2N2N2I7Jnd>"}, "label": "broccoli florets rest behind the steak knife on the plate"}]}
{"id": 209654, "image": "COCO_train2014_000000209654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black colour car in the road\"."}], "task": "refering", "answer_template": "The \"a black colour car in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96], "bbox": [0.005375, 0.045149999999999996, 0.244625, 0.333325], "iscrowd": 0, "area": 14068.0055, "rle": {"size": [400, 640], "counts": "XV16Z<l1TNZ1eN2O00000O100000000O1000000O100000000O100000O10O10O100O100O100O101N100000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000001UOj0L5L3oNnE@W:>n001O0O2O001N101O0O2O001N101O0O2O001N101M2N3L3N3MZdl5"}, "label": "a black colour car in the road"}]}
{"id": 209654, "image": "COCO_train2014_000000209654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red car in motion\"."}], "task": "refering", "answer_template": "The \"a red car in motion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96], "bbox": [0.005375, 0.045149999999999996, 0.244625, 0.333325], "iscrowd": 0, "area": 14068.0055, "rle": {"size": [400, 640], "counts": "XV16Z<l1TNZ1eN2O00000O100000000O1000000O100000000O100000O10O10O100O100O100O101N100000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000001UOj0L5L3oNnE@W:>n001O0O2O001N101O0O2O001N101O0O2O001N101M2N3L3N3MZdl5"}, "label": "a red car in motion"}]}
{"id": 209654, "image": "COCO_train2014_000000209654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white vehicle with a red stripe diagonally on its passenger door\"."}], "task": "refering", "answer_template": "The \"a white vehicle with a red stripe diagonally on its passenger door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 30, 31, 32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 106, 107], "bbox": [0.340671875, 0.031325, 0.730953125, 0.3103], "iscrowd": 0, "area": 15016.551699999996, "rle": {"size": [400, 640], "counts": "QVe28W<6J1O1O1N2O1O1O1O1N2O1O1O100O1000000000001O00000O1000000000000000001O1N3N1O2N1O2N1O2N1O2N1O1N3N1O2N1O2N1O2N1O2N1N2O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N01O000000000000000000001O00000000000000000000001O000000000N2N2N2N2N2O2O00000O2O001O1N2O001O1N2O001O1N1010000O01000O10O10O1000O01000O0100O1O001O1O1O001N2O1O001O1O001O100O010O01O1O001O010O001O1O1O1O1O1O00100O1O1O00001O001O000010O01O001O0000000000O101O0000000O100000000O1O2N1O1O1O1N2O1O1O2N1010O01O001O001O001O001O00001N100O101N100O2O3M3L4M2M4M3K5G9H[SS2"}, "label": "a white vehicle with a red stripe diagonally on its passenger door"}]}
{"id": 209654, "image": "COCO_train2014_000000209654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white car\"."}], "task": "refering", "answer_template": "The \"white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 30, 31, 32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 106, 107], "bbox": [0.340671875, 0.031325, 0.730953125, 0.3103], "iscrowd": 0, "area": 15016.551699999996, "rle": {"size": [400, 640], "counts": "QVe28W<6J1O1O1N2O1O1O1O1N2O1O1O100O1000000000001O00000O1000000000000000001O1N3N1O2N1O2N1O2N1O2N1O1N3N1O2N1O2N1O2N1O2N1N2O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N01O000000000000000000001O00000000000000000000001O000000000N2N2N2N2N2O2O00000O2O001O1N2O001O1N2O001O1N1010000O01000O10O10O1000O01000O0100O1O001O1O1O001N2O1O001O1O001O100O010O01O1O001O010O001O1O1O1O1O1O00100O1O1O00001O001O000010O01O001O0000000000O101O0000000O100000000O1O2N1O1O1O1N2O1O1O2N1010O01O001O001O001O001O00001N100O101N100O2O3M3L4M2M4M3K5G9H[SS2"}, "label": "white car"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right - side duplicate of a skateboarder in a yellow shirt\"."}], "task": "refering", "answer_template": "The \"the right - side duplicate of a skateboarder in a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 197, 198, 199, 200, 220, 221, 222, 244, 245], "bbox": [0.595796875, 0.04503464203233257, 0.94475, 0.6914318706697459], "iscrowd": 0, "area": 32220.073449999985, "rle": {"size": [433, 640], "counts": "oZQ5=T=4L3M4L4L3N3L4L5K6J5K1dLQOTJl0`4d0_KVOo3b1RLYNQ4i1nKWNR4i1nKVNR4l1lKUNS4l1mKTNS4m1kKTNT4n1kKRNZ3ROgJm2n1QNX3[OdJf2S2nMY3C^J_2X2nMY3S3fLmLY3T3fLlLY3W3fLPMS3P3kLTMS3m2jLWMT3i2jLZMT3h2iL\\MU3e2hL^MW3b2gLbMW3_2fLcMY3f2]L\\Mc3e2ZL]Mf3d2WL]Mj3d2SL^Ml3d2PL_MP4b2mK`MS4a2iKbMW4_2fKbMZ4`2bKcM^4^2_KdMa4\\2ZKiMf4`5100000000O101O000O1O1N2O1O2J5I7J6K5J6K6I7N2O2M2O1N2N2O1N2O1N2O1N101N2O1N1O2O0O2O0O100O101N100O101N1O100O2O0O100O2O0O100O2O0O100O2N100O101N100O101N100O101N100O1O2O0O2O0O2O0O2O0O2O1N1O2N1O2O1N2N2iNPGjMR9U2oFiMS9U2oFiMS9V2nFhMT9V2oFhMR9W2oFgMS9W2oFgMS9X2QGcMQ9\\2SG[MS9d2o0M2O2N1N3N1010O010O01O010O00100O1N2N100O0001O000O2O0000001N10001O00000000O10000O1000000O101O1N1M4M2N3M3L3N3L3N3L4N2N2O1N3M2O1N2O1N2N3N1N3N2M3F:Ele>"}, "label": "the right - side duplicate of a skateboarder in a yellow shirt"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a skateboard in the air\"."}], "task": "refering", "answer_template": "The \"a man on a skateboard in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 197, 198, 199, 200, 220, 221, 222, 244, 245], "bbox": [0.595796875, 0.04503464203233257, 0.94475, 0.6914318706697459], "iscrowd": 0, "area": 32220.073449999985, "rle": {"size": [433, 640], "counts": "oZQ5=T=4L3M4L4L3N3L4L5K6J5K1dLQOTJl0`4d0_KVOo3b1RLYNQ4i1nKWNR4i1nKVNR4l1lKUNS4l1mKTNS4m1kKTNT4n1kKRNZ3ROgJm2n1QNX3[OdJf2S2nMY3C^J_2X2nMY3S3fLmLY3T3fLlLY3W3fLPMS3P3kLTMS3m2jLWMT3i2jLZMT3h2iL\\MU3e2hL^MW3b2gLbMW3_2fLcMY3f2]L\\Mc3e2ZL]Mf3d2WL]Mj3d2SL^Ml3d2PL_MP4b2mK`MS4a2iKbMW4_2fKbMZ4`2bKcM^4^2_KdMa4\\2ZKiMf4`5100000000O101O000O1O1N2O1O2J5I7J6K5J6K6I7N2O2M2O1N2N2O1N2O1N2O1N101N2O1N1O2O0O2O0O100O101N100O101N1O100O2O0O100O2O0O100O2O0O100O2N100O101N100O101N100O101N100O1O2O0O2O0O2O0O2O0O2O1N1O2N1O2O1N2N2iNPGjMR9U2oFiMS9U2oFiMS9V2nFhMT9V2oFhMR9W2oFgMS9W2oFgMS9X2QGcMQ9\\2SG[MS9d2o0M2O2N1N3N1010O010O01O010O00100O1N2N100O0001O000O2O0000001N10001O00000000O10000O1000000O101O1N1M4M2N3M3L3N3L3N3L4N2N2O1N3M2O1N2O1N2N3N1N3N2M3F:Ele>"}, "label": "a man on a skateboard in the air"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy in the green shorts to the right of an identical version of himself\"."}], "task": "refering", "answer_template": "The \"the guy in the green shorts to the right of an identical version of himself\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 334, 335], "bbox": [0.5154062500000001, 0.02023094688221709, 0.644640625, 0.9595612009237876], "iscrowd": 0, "area": 22641.307650000002, "rle": {"size": [433, 640], "counts": "^e[4g8S1g300O10000000001O00000000000000001O00000000000O101O00000000000000001O00000000000000J6F:G9G8G:G9G9G9kLnLgKV3S4WMaKn2Z4^M[Kf2_4hMUK\\2f4PNoJT2l4XNeJP2U5Y3K5J6\\J_HV32ZNc7`NeHa2:fNS7hNmHn1a0ROd6POTI[1i0[OV6YO\\Ig0o0Hg5AcI3X13W5JnLMT32U510O100O100O100O10000O100O10\\jP3"}, "label": "the guy in the green shorts to the right of an identical version of himself"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with green color dress\"."}], "task": "refering", "answer_template": "The \"the man with green color dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 334, 335], "bbox": [0.5154062500000001, 0.02023094688221709, 0.644640625, 0.9595612009237876], "iscrowd": 0, "area": 22641.307650000002, "rle": {"size": [433, 640], "counts": "^e[4g8S1g300O10000000001O00000000000000001O00000000000O101O00000000000000001O00000000000000J6F:G9G8G:G9G9G9kLnLgKV3S4WMaKn2Z4^M[Kf2_4hMUK\\2f4PNoJT2l4XNeJP2U5Y3K5J6\\J_HV32ZNc7`NeHa2:fNS7hNmHn1a0ROd6POTI[1i0[OV6YO\\Ig0o0Hg5AcI3X13W5JnLMT32U510O100O100O100O10000O100O10\\jP3"}, "label": "the man with green color dress"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man jumping on his skateboard wearing a yellow tshirt and dark colored shorts\"."}], "task": "refering", "answer_template": "The \"a man jumping on his skateboard wearing a yellow tshirt and dark colored shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 186, 187, 188, 189, 190, 210, 211, 212], "bbox": [0.119234375, 0.042725173210161664, 0.47850000000000004, 0.6681755196304849], "iscrowd": 0, "area": 33221.455500000004, "rle": {"size": [433, 640], "counts": "V\\P11`=3iJ9mGNo3Ml3l1SKWNe4R2VKRNe4Q2YKSNc4o1[KUNb4l1\\KXN`4j1^KWNb4j1]KVNb4l1\\KVNb4k1]KVNc4j1]KWNb4j1\\KWNc4j1]KXNa4i1^KXNa4h1_KZN_4g1`K[N]4f1cKYN^4g1cKXN]4i1cKVN\\4k1dKUN\\4l1cKTN]4l1cKTN\\4n1cKRN]4n1bKSNZ3nNYKP3\\1RNZ3ROVKm2`1PN[3VORKk2a1PN]3WO^JM1n2T2nM]3U3aLkL_3V3`LkL`3V3^LkLb3V3\\LjLd3W3\\LiLSNjNY5^4bLRM]3n2bLTM\\3n2bLSM^3m2aLUM^3m2^LUMa3m2\\LVMc3l2ZLUMf3m2VLUMj3m2RLUMm3n2oKTMQ4n2lKSMT4o2hKSMW4o2fKRM[4o2cKQM^4Q3^KQMa4P3]KRMc4o2YKUMf4k2WKXMh4j2UKXMk4h2SKZMm4g2QK[Mm4f2RK[Mn4f2oJZMS5f2lJYMV5f2jJZMV5`51O1O1O1N2L4M3N2@aJkHc5P7a0L4M301O1O1N2O1O1N3N2N2M2O2N1N3N2N1N3N000O10000O101O0O10000O100O101N100O100O100O100O2O0O100O100O100O2O0O101N100O2O0O101N100O2O0O101N101N2O0O2N1O2N1O2O0O2N2N2gNRGlMo8S2UGiMm8V2UGhMl8W2VGeMm8Y2WGbMk8^2WG_Mk8`2XGPMA6X9h2aGUMa8j2bGRM`8m2S1N2N2N2N2N2O1O1N2O1011N2O1N01O01O01O010N101N101N01N101O001N11O00001O010N100O10000O101N1O1O1O1N2O1O1N2N2N2M3M4M2N3N2O0O2O1O1N3N2M3N2L4J6I6L5L4K5LcS]4"}, "label": "a man jumping on his skateboard wearing a yellow tshirt and dark colored shorts"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left image of a boy doing a skateboard stunts\"."}], "task": "refering", "answer_template": "The \"the left image of a boy doing a skateboard stunts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 186, 187, 188, 189, 190, 210, 211, 212], "bbox": [0.119234375, 0.042725173210161664, 0.47850000000000004, 0.6681755196304849], "iscrowd": 0, "area": 33221.455500000004, "rle": {"size": [433, 640], "counts": "V\\P11`=3iJ9mGNo3Ml3l1SKWNe4R2VKRNe4Q2YKSNc4o1[KUNb4l1\\KXN`4j1^KWNb4j1]KVNb4l1\\KVNb4k1]KVNc4j1]KWNb4j1\\KWNc4j1]KXNa4i1^KXNa4h1_KZN_4g1`K[N]4f1cKYN^4g1cKXN]4i1cKVN\\4k1dKUN\\4l1cKTN]4l1cKTN\\4n1cKRN]4n1bKSNZ3nNYKP3\\1RNZ3ROVKm2`1PN[3VORKk2a1PN]3WO^JM1n2T2nM]3U3aLkL_3V3`LkL`3V3^LkLb3V3\\LjLd3W3\\LiLSNjNY5^4bLRM]3n2bLTM\\3n2bLSM^3m2aLUM^3m2^LUMa3m2\\LVMc3l2ZLUMf3m2VLUMj3m2RLUMm3n2oKTMQ4n2lKSMT4o2hKSMW4o2fKRM[4o2cKQM^4Q3^KQMa4P3]KRMc4o2YKUMf4k2WKXMh4j2UKXMk4h2SKZMm4g2QK[Mm4f2RK[Mn4f2oJZMS5f2lJYMV5f2jJZMV5`51O1O1O1N2L4M3N2@aJkHc5P7a0L4M301O1O1N2O1O1N3N2N2M2O2N1N3N2N1N3N000O10000O101O0O10000O100O101N100O100O100O100O2O0O100O100O100O2O0O101N100O2O0O101N100O2O0O101N101N2O0O2N1O2N1O2O0O2N2N2gNRGlMo8S2UGiMm8V2UGhMl8W2VGeMm8Y2WGbMk8^2WG_Mk8`2XGPMA6X9h2aGUMa8j2bGRM`8m2S1N2N2N2N2N2O1O1N2O1011N2O1N01O01O01O010N101N101N01N101O001N11O00001O010N100O10000O101N1O1O1O1N2O1O1N2N2N2M3M4M2N3N2O0O2O1O1N3N2M3N2L4J6I6L5L4K5LcS]4"}, "label": "the left image of a boy doing a skateboard stunts"}]}
{"id": 78583, "image": "COCO_train2014_000000078583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person standing up in a green shirt and green shorts\"."}], "task": "refering", "answer_template": "The \"a person standing up in a green shirt and green shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323], "bbox": [0.009109375, 0.013718244803695151, 0.147140625, 0.9576674364896074], "iscrowd": 0, "area": 27116.603499999997, "rle": {"size": [433, 640], "counts": "\\a2T1]<d1\\Nd1\\Nc1]Nd1\\Nd1]Nc1\\Nd1\\N5K00000000001O00000001O00000001O00000000000000010O00000000000000001O000001O0000000001OL4D<E;D<E;D<E<H700O1O1O1O1O1O1O1N2N2M3SLXMnLl2o2XMmLj2Q3[MiLh2Q3aMjLa2W1@dNc0WOe1d0]N[Nf2a1\\M^Ni2[1ZMdNk2V1WMiNm2Q1VMnNo2l0SMTOQ3f0QMYO\\38gLGc8N10gfW7"}, "label": "a person standing up in a green shirt and green shorts"}]}
{"id": 496374, "image": "COCO_train2014_000000496374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of a pizza that is shown on a box\"."}], "task": "refering", "answer_template": "The \"a picture of a pizza that is shown on a box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216], "bbox": [0.053125, 0.22471544715447156, 0.515671875, 0.7528184281842819], "iscrowd": 0, "area": 47534.6731, "rle": {"size": [369, 640], "counts": "\\]<?n:7H8I7I7I6K6I7L4L4L2N2N1O2N2N2N2M3N2N2N1O2M3N2N2O1N2O1N101N2N2O1N2N2O1N2N101N2N2N2O1N2N2N2N1O2N2N2N2N2O1N2O1N101N2O0O2O0O2O001N100O2O001O0O101O001O000O2O001O00001O0O1000001O0000000O10001O00000000000O2O00000000001O000O1000001O0000000O101O00000000001O0O100000001O0000000O101O00000000001N10000000001O00000O101O000000001O0O1000000000001O000O1000000000000000000001O01O000000000000O100000001O0000000000000O10000000001N100O100O10000O100O100O101O0O100O100O100O100O100O101N100O100O100O100O100O100O1O101N100O100O100O100O100O100O100O100O100O100O1000000000000000000O101O0O2O0O2O1O0O2O0O2O001N2O001Ne1\\Nj1UNW`_3"}, "label": "a picture of a pizza that is shown on a box"}]}
{"id": 496374, "image": "COCO_train2014_000000496374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of a pizza on a pizza box\"."}], "task": "refering", "answer_template": "The \"a picture of a pizza on a pizza box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216], "bbox": [0.053125, 0.22471544715447156, 0.515671875, 0.7528184281842819], "iscrowd": 0, "area": 47534.6731, "rle": {"size": [369, 640], "counts": "\\]<?n:7H8I7I7I6K6I7L4L4L2N2N1O2N2N2N2M3N2N2N1O2M3N2N2O1N2O1N101N2N2O1N2N2O1N2N101N2N2N2O1N2N2N2N1O2N2N2N2N2O1N2O1N101N2O0O2O0O2O001N100O2O001O0O101O001O000O2O001O00001O0O1000001O0000000O10001O00000000000O2O00000000001O000O1000001O0000000O101O00000000001O0O100000001O0000000O101O00000000001N10000000001O00000O101O000000001O0O1000000000001O000O1000000000000000000001O01O000000000000O100000001O0000000000000O10000000001N100O100O10000O100O100O101O0O100O100O100O100O100O101N100O100O100O100O100O100O1O101N100O100O100O100O100O100O100O100O100O100O1000000000000000000O101O0O2O0O2O1O0O2O0O2O001N2O001Ne1\\Nj1UNW`_3"}, "label": "a picture of a pizza on a pizza box"}]}
{"id": 496374, "image": "COCO_train2014_000000496374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a burn pepperoni pizza on a rack with orange paper\"."}], "task": "refering", "answer_template": "The \"a burn pepperoni pizza on a rack with orange paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226], "bbox": [0.522140625, 0.1865040650406504, 0.9315625000000001, 0.7595392953929538], "iscrowd": 0, "area": 42772.797900000005, "rle": {"size": [369, 640], "counts": "U`h37W;h0YO9G4K5L4K5L4K5L4K5L4K5UG]Mc8k2M2N3M3M2N3M3M2N3M3M2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2O1N2O1O1N2O1N2O0O2O1N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O1000000000001N100000000000000000001O00000000000000000000001O000000000000000000001O000001O000000001O0001O01O000000001O000000010O0000001O000000001O01O000001O0000001O000O101O1N101O0O2O001N101O1O0O2O001N101O0O2O1O0O2O001N101O0O2O001N2O1N1O2N2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N101N2N2N101N2N1O3N1N3M3M2O2L4M2N3M3L3N3M3L4M2N3M3L3N3M3M2M4M3M2K6I:F:F:F:F:Fle?"}, "label": "a burn pepperoni pizza on a rack with orange paper"}]}
{"id": 496374, "image": "COCO_train2014_000000496374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"frozen spicy italian pizza right out of the oven\"."}], "task": "refering", "answer_template": "The \"frozen spicy italian pizza right out of the oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226], "bbox": [0.522140625, 0.1865040650406504, 0.9315625000000001, 0.7595392953929538], "iscrowd": 0, "area": 42772.797900000005, "rle": {"size": [369, 640], "counts": "U`h37W;h0YO9G4K5L4K5L4K5L4K5L4K5UG]Mc8k2M2N3M3M2N3M3M2N3M3M2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2O1N2O1O1N2O1N2O0O2O1N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O2O001N101N101N101O0O2O0O2O0O1000000000001N100000000000000000001O00000000000000000000001O000000000000000000001O000001O000000001O0001O01O000000001O000000010O0000001O000000001O01O000001O0000001O000O101O1N101O0O2O001N101O1O0O2O001N101O0O2O1O0O2O001N101O0O2O001N2O1N1O2N2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N101N2N2N101N2N1O3N1N3M3M2O2L4M2N3M3L3N3M3L4M2N3M3L3N3M3M2M4M3M2K6I:F:F:F:F:Fle?"}, "label": "frozen spicy italian pizza right out of the oven"}]}
{"id": 561913, "image": "COCO_train2014_000000561913.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire in the black shirt and gray pants\"."}], "task": "refering", "answer_template": "The \"the umpire in the black shirt and gray pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 57, 58, 59, 74, 75, 76, 92, 93, 94, 110, 111, 112, 129, 130], "bbox": [0.1207, 0.29348360655737704, 0.28854, 0.8521311475409836], "iscrowd": 0, "area": 6024.510149999998, "rle": {"size": [244, 500], "counts": "b]>=U73M2M3N2^J[OS4h0eK_OZ4c0_KC`4?\\KDc4>ZKDe4>WKEh4=UKEj4=RKFm4<PKGn4;nJHQ5^1O1O1O1O1OFUKmMj4T2XKjMg4V2ZKjMe4V2]KjMa4W2_KiM`4W2bKhM]4Y2cKgM\\4Y2fKeMZ4\\2fKcM[4\\2>@lJbNS5_1?2O1O1O2N100O100oKnM`2S2[MQNe2o1VMVNj2j1QM[No2e1mL_NS3a1iLcNW3^1dLfN[3[1`LjN`3V1\\LnNd3R1XLROh3]2O2O00000O1O10ZMYLR10TN`3P1cLj0NUN^3Q1hLg0JXN]3Q1lLe0HYN[3R1PMd0EYN[3R1TMb0B[NY3S1XM`0@\\NX3S1[M;BaNS3R1_M7CfNm2S1dM0DlNh2S1fMLGPOc2S1hMHJSO_2j0hLZOQ1d0Z32QMIS38nLCU3=lL^OX3b0n1O09G5K10O0DlIYOS6f0?N2M2L5L5JUhd2"}, "label": "the umpire in the black shirt and gray pants"}]}
{"id": 561913, "image": "COCO_train2014_000000561913.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire with a black shirt and dark grey pants\"."}], "task": "refering", "answer_template": "The \"an umpire with a black shirt and dark grey pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 57, 58, 59, 74, 75, 76, 92, 93, 94, 110, 111, 112, 129, 130], "bbox": [0.1207, 0.29348360655737704, 0.28854, 0.8521311475409836], "iscrowd": 0, "area": 6024.510149999998, "rle": {"size": [244, 500], "counts": "b]>=U73M2M3N2^J[OS4h0eK_OZ4c0_KC`4?\\KDc4>ZKDe4>WKEh4=UKEj4=RKFm4<PKGn4;nJHQ5^1O1O1O1O1OFUKmMj4T2XKjMg4V2ZKjMe4V2]KjMa4W2_KiM`4W2bKhM]4Y2cKgM\\4Y2fKeMZ4\\2fKcM[4\\2>@lJbNS5_1?2O1O1O2N100O100oKnM`2S2[MQNe2o1VMVNj2j1QM[No2e1mL_NS3a1iLcNW3^1dLfN[3[1`LjN`3V1\\LnNd3R1XLROh3]2O2O00000O1O10ZMYLR10TN`3P1cLj0NUN^3Q1hLg0JXN]3Q1lLe0HYN[3R1PMd0EYN[3R1TMb0B[NY3S1XM`0@\\NX3S1[M;BaNS3R1_M7CfNm2S1dM0DlNh2S1fMLGPOc2S1hMHJSO_2j0hLZOQ1d0Z32QMIS38nLCU3=lL^OX3b0n1O09G5K10O0DlIYOS6f0?N2M2L5L5JUhd2"}, "label": "an umpire with a black shirt and dark grey pants"}]}
{"id": 561913, "image": "COCO_train2014_000000561913.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher wearing green\"."}], "task": "refering", "answer_template": "The \"a baseball catcher wearing green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 63, 78, 79, 80, 81, 95, 96, 97, 113, 114, 115, 116, 131, 132, 133], "bbox": [0.30328, 0.4087704918032787, 0.52792, 0.846639344262295], "iscrowd": 0, "area": 5175.516849999999, "rle": {"size": [244, 500], "counts": "S\\T15\\77J4J5K5N2N2N2eIUOj5l0TJWOi5l0VJUOh5l0WJVOg5l0WJVOg5k0XJWOf5\\1O1N2O1O1O2N001N2O100O5K2N100O1O2N100O1O2N100O1O100O1O1M3I7ODbKdM^4Z2eKeMZ4Z2iKeMW4Y2kKgMT4X2oKgMQ4X2QLhMm3W2ULiMk3V2WLiMh3W2ZLhMf3V2l001N2N2N2N2N2F:1OXObJ[O^5c0dJ\\O\\5d0eJ\\O[5b0fJ_OY5a0hJ^OY5`0hJAW5?jJAV5>kJAW5KYJ6c0Nb5N`J1_5OcJO]51eJ0X5OjJ2T5NmJ4P5LQK5m4KTK3m4MUK0l4OWKOh42]100000O01000N2N101N2O0O2O1O1O001O1O100000000001N2O2N2N2M3M3N2M2OjSh1"}, "label": "a baseball catcher wearing green"}]}
{"id": 561913, "image": "COCO_train2014_000000561913.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher in green behind a batter swinging at a ball\"."}], "task": "refering", "answer_template": "The \"a catcher in green behind a batter swinging at a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 63, 78, 79, 80, 81, 95, 96, 97, 113, 114, 115, 116, 131, 132, 133], "bbox": [0.30328, 0.4087704918032787, 0.52792, 0.846639344262295], "iscrowd": 0, "area": 5175.516849999999, "rle": {"size": [244, 500], "counts": "S\\T15\\77J4J5K5N2N2N2eIUOj5l0TJWOi5l0VJUOh5l0WJVOg5l0WJVOg5k0XJWOf5\\1O1N2O1O1O2N001N2O100O5K2N100O1O2N100O1O2N100O1O100O1O1M3I7ODbKdM^4Z2eKeMZ4Z2iKeMW4Y2kKgMT4X2oKgMQ4X2QLhMm3W2ULiMk3V2WLiMh3W2ZLhMf3V2l001N2N2N2N2N2F:1OXObJ[O^5c0dJ\\O\\5d0eJ\\O[5b0fJ_OY5a0hJ^OY5`0hJAW5?jJAV5>kJAW5KYJ6c0Nb5N`J1_5OcJO]51eJ0X5OjJ2T5NmJ4P5LQK5m4KTK3m4MUK0l4OWKOh42]100000O01000N2N101N2O0O2O1O1O001O1O100000000001N2O2N2N2M3M3N2M2OjSh1"}, "label": "a catcher in green behind a batter swinging at a ball"}]}
{"id": 119543, "image": "COCO_train2014_000000119543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left sausage on the yellow plate on the left\"."}], "task": "refering", "answer_template": "The \"the left sausage on the yellow plate on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 231, 232, 233, 255, 256, 257, 278, 279, 280, 302, 303, 304, 325, 326, 327, 349, 350], "bbox": [0.06540625, 0.5716458333333333, 0.23987499999999998, 0.9077708333333334], "iscrowd": 0, "area": 7712.6512999999995, "rle": {"size": [480, 640], "counts": "onc03l>3M3N2M4L3M3M3M3N2M3M3M3M3M4M2M3N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2O0010000O01000O01000O10O10000O0100000O10000O1000000O10O1000O10000O100O1O1N2O1N2N101N2O1N1O2O1N2M3N1N3N2N2M2O2N2M3N1N3N2N2M2O2N2M3N2N2M3N2M3N2N<]OW]S7"}, "label": "the left sausage on the yellow plate on the left"}]}
{"id": 119543, "image": "COCO_train2014_000000119543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hotdog leaning up onto the side of a plate\"."}], "task": "refering", "answer_template": "The \"a hotdog leaning up onto the side of a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 231, 232, 233, 255, 256, 257, 278, 279, 280, 302, 303, 304, 325, 326, 327, 349, 350], "bbox": [0.06540625, 0.5716458333333333, 0.23987499999999998, 0.9077708333333334], "iscrowd": 0, "area": 7712.6512999999995, "rle": {"size": [480, 640], "counts": "onc03l>3M3N2M4L3M3M3M3N2M3M3M3M3M4M2M3N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2N2N2N2N3M2N2N2N2O0010000O01000O01000O10O10000O0100000O10000O1000000O10O1000O10000O100O1O1N2O1N2N101N2O1N1O2O1N2M3N1N3N2N2M2O2N2M3N1N3N2N2M2O2N2M3N2N2M3N2M3N2N<]OW]S7"}, "label": "a hotdog leaning up onto the side of a plate"}]}
{"id": 463613, "image": "COCO_train2014_000000463613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair to the right of another chair\"."}], "task": "refering", "answer_template": "The \"a white chair to the right of another chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 225, 226, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 321], "bbox": [0.780640625, 0.49677083333333333, 0.983875, 0.8279583333333332], "iscrowd": 0, "area": 8744.782499999998, "rle": {"size": [480, 640], "counts": "YgZ73k>3M3L4M3M3LO2N2O10TMCTE0Y1>c9DRE4U19g9FRE7Q14l9GRE9l02Q:GQE=h0MY:EnDc0c0I`:DlDg0>Gg:BkDk08Dn:AiDP13@U;@gDT1O]O[;_OfDX1H[Oc;]OdDY2];gMbDY2_;gM`DX2b;hM^DX2c;gM\\DZ2d;fM[D[2e;eM[D\\2d;dM[D]2e;cMZD^2f;bMZD_2e;aMZD`2f;`MYDb2f;^MZDb2f;^MYDc2g;<fEnLe8Q3SGWMP9f2hFbM\\9[2[FlMj9P2mEXNY:d1^EbNh:g27I6J7I8H5L1N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N28I0O01O001O001O1O001EYC[Ng<e1;O1O0010O01O1O001O10O01O001O1O01O01O00000001O0000000000001O000001O0001O01O0V1jN0001O000O101N3I7I7I7I7E;D<D<F:EV\\4"}, "label": "a white chair to the right of another chair"}]}
{"id": 78590, "image": "COCO_train2014_000000078590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the class of water in front of the bowl of bread\"."}], "task": "refering", "answer_template": "The \"the class of water in front of the bowl of bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 36, 37, 38, 49, 50, 51, 62, 63, 64, 75, 76, 88, 89, 90, 101, 102], "bbox": [0.7610399999999999, 0.06966, 0.9887733333333333, 0.4045], "iscrowd": 0, "area": 9158.262000000002, "rle": {"size": [500, 375], "counts": "ej[41b?2N1YM=\\EEZ:e0dE\\OZ:g0dEZOY:j0eEXOX:j0gEWOV:m0hEUOU:n0iESOT:Q1jEPOT:R1kEoNU9R2iFPNV9Q2hFPNX9Q2fFPNY9Q2fFPNZ9Q2dFQN[9P2cFQN\\9Q2bFPN^9P2aFRN]9P2aFQN_9P2]FSNc9n1XFVNg9l1fEgNY:Y1dEjN[:X1aEkN_:V3O0O100O2O0000001O00001O0000000000[NeESMO8\\:^2nEWMG:\\:Z2SFYMD<Y:[2UFUMD`0W:Z2XFSMCb0U:[2YFPMDe0S:[2[FmLCh0R:Z2kFeMU9[2kFdMV9\\2kFcMU9\\2lFdMT9\\2lFdMT9\\2mFcMS9]2mFcMS9\\2nFdMR9\\2oFbMR9^2nFbMR9]2PGbMP9^2PGbMP9^2PGaMQ9^2QG^MR9b2nFZMV9f2jFVMZ9i2hFRM\\9n2e10000O10001N2O0K6K4K6J6J5kNdBMb=LV1K6I7I[m1"}, "label": "the class of water in front of the bowl of bread"}]}
{"id": 78590, "image": "COCO_train2014_000000078590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goblet with water\"."}], "task": "refering", "answer_template": "The \"a goblet with water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 36, 37, 38, 49, 50, 51, 62, 63, 64, 75, 76, 88, 89, 90, 101, 102], "bbox": [0.7610399999999999, 0.06966, 0.9887733333333333, 0.4045], "iscrowd": 0, "area": 9158.262000000002, "rle": {"size": [500, 375], "counts": "ej[41b?2N1YM=\\EEZ:e0dE\\OZ:g0dEZOY:j0eEXOX:j0gEWOV:m0hEUOU:n0iESOT:Q1jEPOT:R1kEoNU9R2iFPNV9Q2hFPNX9Q2fFPNY9Q2fFPNZ9Q2dFQN[9P2cFQN\\9Q2bFPN^9P2aFRN]9P2aFQN_9P2]FSNc9n1XFVNg9l1fEgNY:Y1dEjN[:X1aEkN_:V3O0O100O2O0000001O00001O0000000000[NeESMO8\\:^2nEWMG:\\:Z2SFYMD<Y:[2UFUMD`0W:Z2XFSMCb0U:[2YFPMDe0S:[2[FmLCh0R:Z2kFeMU9[2kFdMV9\\2kFcMU9\\2lFdMT9\\2lFdMT9\\2mFcMS9]2mFcMS9\\2nFdMR9\\2oFbMR9^2nFbMR9]2PGbMP9^2PGbMP9^2PGaMQ9^2QG^MR9b2nFZMV9f2jFVMZ9i2hFRM\\9n2e10000O10001N2O0K6K4K6J6J5kNdBMb=LV1K6I7I[m1"}, "label": "a goblet with water"}]}
{"id": 78590, "image": "COCO_train2014_000000078590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of red wine on a table\"."}], "task": "refering", "answer_template": "The \"a glass of red wine on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 13, 14, 15, 16, 17, 27, 28, 29, 30, 40, 41, 42, 43, 54, 55, 68, 80, 81, 82, 93], "bbox": [0.047946666666666665, 0.00674, 0.3535466666666666, 0.41348], "iscrowd": 0, "area": 12046.31855, "rle": {"size": [500, 375], "counts": "[i83a?4L4L5K4L5K4L5K4L5K4L5K4L5K4L2N2N2N2N2N2N1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O001O1O1O0nEPM`7Q3_HRM^7o2`HUM]7k2cHXMZ7i2dH[MY7f2eH^MX7b2hHaMU7`2iHdMT7]2jHdMV7]2iHcMW7]2hHcMY7^2fHbMZ7_2dHaM]7_2bHbM^7_2aH`M`7a2^H`Mb7`2^H_Mc7b2[H_Me7b2YH^Mh7c2WH\\Mj7d2UH]Mk7d2SH\\Mn7e2QH[Mo7e2PH[MQ8f2nGZMR8g2lGYMU8g2jGZMV8g2iGXMX8l2cGUM]8Q3_GlLb8Z3ZGdLf8b3VG[Lk8k3RGRLn8T4nFiKS9]4jF`KV9f4fFWK[9o4bFmJ_9Y5]FeJc9a5YF\\Jh9h50O100iMnFkMS9S2XGdMh8[2aG\\M`8c2jGTMV8j2THnLl7Q3WHlLj7S3WHmLi7R3XHnLh7P3YHQMg7n2ZHRMf7m2ZHTMf7j2\\HVMd7i2]HWMc7h2]HYMc7e2_HZMb7e2^H\\Mb7c2_H]Ma7a2aH_M_7`2aHaM_7^2bHbM^7\\2cHeM]7Z2dHfM\\7V2jHhM[7n1mHnM[7f1lHXN[7\\1mHaNZ7S1m3D<D<DkSf3"}, "label": "a glass of red wine on a table"}]}
{"id": 78590, "image": "COCO_train2014_000000078590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of red wine by the pizza\"."}], "task": "refering", "answer_template": "The \"a glass of red wine by the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 13, 14, 15, 16, 17, 27, 28, 29, 30, 40, 41, 42, 43, 54, 55, 68, 80, 81, 82, 93], "bbox": [0.047946666666666665, 0.00674, 0.3535466666666666, 0.41348], "iscrowd": 0, "area": 12046.31855, "rle": {"size": [500, 375], "counts": "[i83a?4L4L5K4L5K4L5K4L5K4L5K4L5K4L2N2N2N2N2N2N1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O001O1O1O0nEPM`7Q3_HRM^7o2`HUM]7k2cHXMZ7i2dH[MY7f2eH^MX7b2hHaMU7`2iHdMT7]2jHdMV7]2iHcMW7]2hHcMY7^2fHbMZ7_2dHaM]7_2bHbM^7_2aH`M`7a2^H`Mb7`2^H_Mc7b2[H_Me7b2YH^Mh7c2WH\\Mj7d2UH]Mk7d2SH\\Mn7e2QH[Mo7e2PH[MQ8f2nGZMR8g2lGYMU8g2jGZMV8g2iGXMX8l2cGUM]8Q3_GlLb8Z3ZGdLf8b3VG[Lk8k3RGRLn8T4nFiKS9]4jF`KV9f4fFWK[9o4bFmJ_9Y5]FeJc9a5YF\\Jh9h50O100iMnFkMS9S2XGdMh8[2aG\\M`8c2jGTMV8j2THnLl7Q3WHlLj7S3WHmLi7R3XHnLh7P3YHQMg7n2ZHRMf7m2ZHTMf7j2\\HVMd7i2]HWMc7h2]HYMc7e2_HZMb7e2^H\\Mb7c2_H]Ma7a2aH_M_7`2aHaM_7^2bHbM^7\\2cHeM]7Z2dHfM\\7V2jHhM[7n1mHnM[7f1lHXN[7\\1mHaNZ7S1m3D<D<DkSf3"}, "label": "a glass of red wine by the pizza"}]}
{"id": 365314, "image": "COCO_train2014_000000365314.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the further of two yellow motorcycles\"."}], "task": "refering", "answer_template": "The \"the further of two yellow motorcycles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 263, 264, 265, 266], "bbox": [0.256234375, 0.3531666666666667, 0.593140625, 0.6956458333333333], "iscrowd": 0, "area": 16497.40300000001, "rle": {"size": [480, 640], "counts": "Ra]21m>101O010O010O0cB2j;NQD7o;HPD9P<HnC:Q<FnC<Q<ElC>S<CkC?T<BkC?T<BlC=T<DlC<S<EmC;S<EmC;S<DmC=R<DnC<R<DnC=Q<CoC>P<BPD?n;CfCGEg0d<DeCEFi0d<CdCEHi0c<CdCEHi0b<DdCDJi0a<CeCEIh0b<CeCEIi0a<BfCFGj0b<@gCHEi0b<@iCHDi0b<_OdCS1[<i00N110O100O010O100O010O1O10O0100O10O010000000O101bCgMS<Y2kCiMU<c0fCl03cNW<a0gCk00fNY<>hCl0MhN[<<iCU1W<jNkCU1U<jNmCU1S<kNnCT1R<kNoCU1Q<kNPDT1P<kNRDT1n;lNSDS1m;lNTDT1l;lNUDS1k;lNWDS1i;mNXDR1h;mNYDS1g;mNZDS1e;mN\\DR1d;nN]DQ1c;oN^DP1b;PO^DP1d;nN]DQ1e;mN\\DR1g;kNZDT1f;lNZDT1f;lN[DS1e;mN\\DR1ZOQOh;MoDQ1XOTOh;KPEQ1WOVOi;HQEQ1UOYOi;ETEQ1RO\\Oi;CVEP1QO^Oh;BWEP1PO@h;@YEP1mNBi;^O[Eh1e:WN\\Ei1c:WN\\Ek1c:UN\\Em1c:SN]En1b:QN^ER2f;2O1N1O2O1N2O1O1O0O2O1O1O1O1N2O2N2N2M3N2N2N2N2M3N1O1O2N1N201N100O01BfDTM];i2eDUM_;g2bDXMa;e2aDYMb;d2_D\\M_;e2cDYM];h2bDXM];j2bDVM^;k2bDTM];m2cDSM];n2bDRM];P3cDPM\\;Q3cDoL\\;S3cDmL];T3cDkL\\;V3dDjL\\;W3cDiL\\;Y342L3M4M3M3O1N101N2O100O01O1O001O1O3M1O1O00000dLhDP3X;oLjDP3V;PMkDo2U;PMmDo2S;QMmDn2T;QMnDn2R;RMoDm2Q;RMPEn2P;RMQEm2P;QMREm2o:SMREl2n:SMSEm2m:SMTEl2a;O0O2O1O1N1O2N2N101N2N2N1O2N2O2M3M3M3L4K5L4K5L3L5K9F<D<DTci3"}, "label": "the further of two yellow motorcycles"}]}
{"id": 365314, "image": "COCO_train2014_000000365314.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bike with black wheels sitting to the right of an identical bike\"."}], "task": "refering", "answer_template": "The \"a yellow bike with black wheels sitting to the right of an identical bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 263, 264, 265, 266], "bbox": [0.256234375, 0.3531666666666667, 0.593140625, 0.6956458333333333], "iscrowd": 0, "area": 16497.40300000001, "rle": {"size": [480, 640], "counts": "Ra]21m>101O010O010O0cB2j;NQD7o;HPD9P<HnC:Q<FnC<Q<ElC>S<CkC?T<BkC?T<BlC=T<DlC<S<EmC;S<EmC;S<DmC=R<DnC<R<DnC=Q<CoC>P<BPD?n;CfCGEg0d<DeCEFi0d<CdCEHi0c<CdCEHi0b<DdCDJi0a<CeCEIh0b<CeCEIi0a<BfCFGj0b<@gCHEi0b<@iCHDi0b<_OdCS1[<i00N110O100O010O100O010O1O10O0100O10O010000000O101bCgMS<Y2kCiMU<c0fCl03cNW<a0gCk00fNY<>hCl0MhN[<<iCU1W<jNkCU1U<jNmCU1S<kNnCT1R<kNoCU1Q<kNPDT1P<kNRDT1n;lNSDS1m;lNTDT1l;lNUDS1k;lNWDS1i;mNXDR1h;mNYDS1g;mNZDS1e;mN\\DR1d;nN]DQ1c;oN^DP1b;PO^DP1d;nN]DQ1e;mN\\DR1g;kNZDT1f;lNZDT1f;lN[DS1e;mN\\DR1ZOQOh;MoDQ1XOTOh;KPEQ1WOVOi;HQEQ1UOYOi;ETEQ1RO\\Oi;CVEP1QO^Oh;BWEP1PO@h;@YEP1mNBi;^O[Eh1e:WN\\Ei1c:WN\\Ek1c:UN\\Em1c:SN]En1b:QN^ER2f;2O1N1O2O1N2O1O1O0O2O1O1O1O1N2O2N2N2M3N2N2N2N2M3N1O1O2N1N201N100O01BfDTM];i2eDUM_;g2bDXMa;e2aDYMb;d2_D\\M_;e2cDYM];h2bDXM];j2bDVM^;k2bDTM];m2cDSM];n2bDRM];P3cDPM\\;Q3cDoL\\;S3cDmL];T3cDkL\\;V3dDjL\\;W3cDiL\\;Y342L3M4M3M3O1N101N2O100O01O1O001O1O3M1O1O00000dLhDP3X;oLjDP3V;PMkDo2U;PMmDo2S;QMmDn2T;QMnDn2R;RMoDm2Q;RMPEn2P;RMQEm2P;QMREm2o:SMREl2n:SMSEm2m:SMTEl2a;O0O2O1O1N1O2N2N101N2N2N1O2N2O2M3M3M3L4K5L4K5L3L5K9F<D<DTci3"}, "label": "a yellow bike with black wheels sitting to the right of an identical bike"}]}
{"id": 365314, "image": "COCO_train2014_000000365314.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bike to the left of another yellow bike\"."}], "task": "refering", "answer_template": "The \"a yellow bike to the left of another yellow bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 161, 162, 163, 164, 165, 166, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 328, 330, 331], "bbox": [0.026968750000000003, 0.40675, 0.461796875, 0.8426875], "iscrowd": 0, "area": 30781.214650000005, "rle": {"size": [480, 640], "counts": "gU81nm00PA001I7N110O10O010O10000O010O10000O01000O0101O0001O010O0001O00001O0iCg0g8ZOnFS1o8mNhF_1U9bNeFf1X9ZNfFl1V9UNhFo1U9RNhFS2U9mMjFW2S9jMkFZ2R9fMlF^2R9bMmFa2Q9_MnFd2o8]MoFg2o8YMPGj2n8VMPGo2m8QMRGQ3m8oL^FD4^3^9nL]FF2^3a9lL\\FH1^3b9jL[FK1\\3c9jL[FL0[3e9iLZFm3e9g00001OO001O1O1O001O1O1M3K5K5K4K6K5K5O1000000O10001O0O11OjLoD]2Q;cMQE\\2n:dMTEZ2l:fMVEY2i:gMXEX2h:hMZEV2f:jM[EV2d:jM]EU2c:kM^EW2_:iMbE[2Y:eMgE_2U:`MmEc2o9]MRFe2k9[MVFf2h9YMYFi2e9WM\\Fj2b9VM_Fj2`9UMbFk2]9UMcFl2\\9SMfFm2Y9RMiFn2W9QMiFo2W9PMkFP3T9oLnFQ3Q9nLPGR3P9mLRGS3m8lLTGT3l8kLVGU3j8iLXGV3h8iLYGX3g8gLZGX3f8hLZGX3f8hL[GX3e8gL[GY3e8hL[GW3e8iL[GW3f8hL[GW3f8hLZGX3g8gLYGY3g8hLYGX3g8hLXGX3i8gLWGY3i8gLWGY3j8gLUGZ3j8fLVGZ3k8eLUG[3k8eLUG\\3j8dLVG]3i8cLWG]3j8cLUG^3j8bLUG`3j8`LUGa3k8_LUGb3j8_LTGc3k8]LTGd3EYL]83nGd3EYL]83mGf3EWL_82lGg3EWL_82lGh3DWL_81mGh3DWL_81lGi3EVL_81lGj3DVLd9j3\\FVLd9j3\\FVLd9k3ZFWLe9i3[FWLd9k3ZFVLf9j3YFXLf9h3YFYLg9g3WF[Li9e3VF]Li9c3VF^Lj9b3UF_Lk9a3TFaLk9_3UFaLk9_3TF^LJDQ:n3TF^LMCo9o3RF_LU:a3hEaL5@k9b4:NoE`Kf9_4>O1N2O1N3M201N101O3M4K3N1O001OO100O101N2N4L2N2N2N2N2N2O0O2N2O001N2O001O0O2O00dKhEQ4W:nKjER4V:lKlEU4S:iKoEW4R:fKPF[4o9dKQF]4Y:1O001O0000001O0O1000000000000000000000000000eKlEj3T:ULnEj3S:TLoEk3Q:TLQFk3o9TLSFk3m9TLUFk3a:O1O1O1O1O1O1O1O1O1O1N2O1O0O2O1N2O1O1N2O1O1N2O2N2N2N2M3N1O2DgCmM[<k1kCUNW<d1mC\\NV<\\1PDcNS<V1QDhNT<Q1Q1I8I7H8I7IjmP5"}, "label": "a yellow bike to the left of another yellow bike"}]}
{"id": 365314, "image": "COCO_train2014_000000365314.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow motorbiked parked to the left of other yellow bike\"."}], "task": "refering", "answer_template": "The \"yellow motorbiked parked to the left of other yellow bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 161, 162, 163, 164, 165, 166, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 328, 330, 331], "bbox": [0.026968750000000003, 0.40675, 0.461796875, 0.8426875], "iscrowd": 0, "area": 30781.214650000005, "rle": {"size": [480, 640], "counts": "gU81nm00PA001I7N110O10O010O10000O010O10000O01000O0101O0001O010O0001O00001O0iCg0g8ZOnFS1o8mNhF_1U9bNeFf1X9ZNfFl1V9UNhFo1U9RNhFS2U9mMjFW2S9jMkFZ2R9fMlF^2R9bMmFa2Q9_MnFd2o8]MoFg2o8YMPGj2n8VMPGo2m8QMRGQ3m8oL^FD4^3^9nL]FF2^3a9lL\\FH1^3b9jL[FK1\\3c9jL[FL0[3e9iLZFm3e9g00001OO001O1O1O001O1O1M3K5K5K4K6K5K5O1000000O10001O0O11OjLoD]2Q;cMQE\\2n:dMTEZ2l:fMVEY2i:gMXEX2h:hMZEV2f:jM[EV2d:jM]EU2c:kM^EW2_:iMbE[2Y:eMgE_2U:`MmEc2o9]MRFe2k9[MVFf2h9YMYFi2e9WM\\Fj2b9VM_Fj2`9UMbFk2]9UMcFl2\\9SMfFm2Y9RMiFn2W9QMiFo2W9PMkFP3T9oLnFQ3Q9nLPGR3P9mLRGS3m8lLTGT3l8kLVGU3j8iLXGV3h8iLYGX3g8gLZGX3f8hLZGX3f8hL[GX3e8gL[GY3e8hL[GW3e8iL[GW3f8hL[GW3f8hLZGX3g8gLYGY3g8hLYGX3g8hLXGX3i8gLWGY3i8gLWGY3j8gLUGZ3j8fLVGZ3k8eLUG[3k8eLUG\\3j8dLVG]3i8cLWG]3j8cLUG^3j8bLUG`3j8`LUGa3k8_LUGb3j8_LTGc3k8]LTGd3EYL]83nGd3EYL]83mGf3EWL_82lGg3EWL_82lGh3DWL_81mGh3DWL_81lGi3EVL_81lGj3DVLd9j3\\FVLd9j3\\FVLd9k3ZFWLe9i3[FWLd9k3ZFVLf9j3YFXLf9h3YFYLg9g3WF[Li9e3VF]Li9c3VF^Lj9b3UF_Lk9a3TFaLk9_3UFaLk9_3TF^LJDQ:n3TF^LMCo9o3RF_LU:a3hEaL5@k9b4:NoE`Kf9_4>O1N2O1N3M201N101O3M4K3N1O001OO100O101N2N4L2N2N2N2N2N2O0O2N2O001N2O001O0O2O00dKhEQ4W:nKjER4V:lKlEU4S:iKoEW4R:fKPF[4o9dKQF]4Y:1O001O0000001O0O1000000000000000000000000000eKlEj3T:ULnEj3S:TLoEk3Q:TLQFk3o9TLSFk3m9TLUFk3a:O1O1O1O1O1O1O1O1O1O1N2O1O0O2O1N2O1O1N2O1O1N2O2N2N2N2M3N1O2DgCmM[<k1kCUNW<d1mC\\NV<\\1PDcNS<V1QDhNT<Q1Q1I8I7H8I7IjmP5"}, "label": "yellow motorbiked parked to the left of other yellow bike"}]}
{"id": 234244, "image": "COCO_train2014_000000234244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a canned drink observing the game but not playing\"."}], "task": "refering", "answer_template": "The \"a man holding a canned drink observing the game but not playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 81, 82, 83, 104, 105, 106, 107, 128, 129, 130, 151, 152, 174, 175, 197, 198, 220, 221, 222, 243, 244, 245, 266, 267, 289, 290, 313], "bbox": [0.5443749999999999, 0.167875, 0.691015625, 0.7858333333333333], "iscrowd": 0, "area": 12525.362949999999, "rle": {"size": [480, 640], "counts": "ceS54k>101O00001N10001WB<T<DlC6Z<KeCOa<1_C0`<0aC?_1g0c6jNmGa0U1T1i6ZNRHc0k0b1m6kMYHc0?R2a0SMb47^Je05`2d0dLk45]Jh0Jn2g0VLS53]Ji0_O\\3k0iKZ50]Jl0TOj3n0[Kb5N\\JR6KQJm6S6oHmIQ7V6kHkIV7V6gHkIY7X6cHiI]7Z6_HgIa7[6\\HfId7]6XHdIi7]6UHcIk7i600ROUIgIk6P6`InI`6P6eInIZ6R6hIlIW6T6lIjIT6V6oIgIR6X6PJfIQ6Y6QJeIP6Z6SJdIm5[6UJcIl5\\6]1O10^LgI]NX6^1lIcNS6Z1QJfNo5U1VJkNi5R1[JnNe5n0_JRO`5j0eJVO[5f0hJ[OX5`0mJ@R5=RKCn48WKIg44]KLc4ObK1]4LfK5Z4FkK:U4PMPJj1o1c0c4^McIl1n12S5lMTIn1n1LW5QNQIo1l1Ko52TJJP65RJFS67PJCU6<mI_OX6?jI\\OZ6b0iIYO[6f0fIVO_6h0cISOa6k0bIoNd6o0^IlNf6R1]IiNg6V1ZIfNk6X1_3N3N1O1N3N1N3N1O3L3N3M2N3M3M2N3M2O2Mbel2"}, "label": "a man holding a canned drink observing the game but not playing"}]}
{"id": 234244, "image": "COCO_train2014_000000234244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in all brown behind the others\"."}], "task": "refering", "answer_template": "The \"the man in all brown behind the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 81, 82, 83, 104, 105, 106, 107, 128, 129, 130, 151, 152, 174, 175, 197, 198, 220, 221, 222, 243, 244, 245, 266, 267, 289, 290, 313], "bbox": [0.5443749999999999, 0.167875, 0.691015625, 0.7858333333333333], "iscrowd": 0, "area": 12525.362949999999, "rle": {"size": [480, 640], "counts": "ceS54k>101O00001N10001WB<T<DlC6Z<KeCOa<1_C0`<0aC?_1g0c6jNmGa0U1T1i6ZNRHc0k0b1m6kMYHc0?R2a0SMb47^Je05`2d0dLk45]Jh0Jn2g0VLS53]Ji0_O\\3k0iKZ50]Jl0TOj3n0[Kb5N\\JR6KQJm6S6oHmIQ7V6kHkIV7V6gHkIY7X6cHiI]7Z6_HgIa7[6\\HfId7]6XHdIi7]6UHcIk7i600ROUIgIk6P6`InI`6P6eInIZ6R6hIlIW6T6lIjIT6V6oIgIR6X6PJfIQ6Y6QJeIP6Z6SJdIm5[6UJcIl5\\6]1O10^LgI]NX6^1lIcNS6Z1QJfNo5U1VJkNi5R1[JnNe5n0_JRO`5j0eJVO[5f0hJ[OX5`0mJ@R5=RKCn48WKIg44]KLc4ObK1]4LfK5Z4FkK:U4PMPJj1o1c0c4^McIl1n12S5lMTIn1n1LW5QNQIo1l1Ko52TJJP65RJFS67PJCU6<mI_OX6?jI\\OZ6b0iIYO[6f0fIVO_6h0cISOa6k0bIoNd6o0^IlNf6R1]IiNg6V1ZIfNk6X1_3N3N1O1N3N1N3N1O3L3N3M2N3M3M2N3M2O2Mbel2"}, "label": "the man in all brown behind the others"}]}
{"id": 234244, "image": "COCO_train2014_000000234244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt with a skull on it\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt with a skull on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 75, 76, 77, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 281, 282, 283, 284, 305, 306, 307], "bbox": [0.2545, 0.15885416666666666, 0.42471875, 0.8127916666666667], "iscrowd": 0, "area": 23124.586200000005, "rle": {"size": [480, 640], "counts": "ec\\22n>2hDc0i6_OPIZ1[6jN\\I\\1d6fNUI^1l6dNlHb1T7`NdHe1\\4mMSL`0[Oh1b4jMQL>WOn1g4fMQL=QOR2n4cMPL;oNU2Q5aMPL:kNX2T5`MPL8jNY2W5aMnK6gN\\2Z5`MnK4eN^2]5`MnK2aNa2a5^MaKAoM`0m0c2b5^M]KESN:j0e2f5^MXKIVN4i0h2h5\\MUKNYNNf0j2m5[MPK2[NIf0l2n5ZMmJ7^NCc0o2Q6XMjJo0ROl1S6VMgJS1ROk1U6SMeJV1TOi1V6RMdJX1ROj1X6oLcJZ1ROj1[6mL_J]1ROj1]6jL^Jf5a5[J\\Ji5b5XJ[Jk5d5VJZJm5d5TJYJo5\\5n11N2N2N2N2M3N2N2N3L3N3M4L3M3M3O1;E;E;F:E3M000000O10001O1O1O1O1N2O1O2N1QHoIn6Q6PIQJP7o5nHSJS7l5kHVJU7j5jHWJV7i5hHYJY7g5dH[J\\7h5_HZJa7i5ZHYJg7i5UHXJk7k5PHWJP8\\62NM4J5K5L4K5K5L4K6J5L4K5K5J6K5J7J5K5J6K5K5L4M4K<lM]EXOn:OiEGc:@UF6V:QObFe0U<E:GRX\\5"}, "label": "a man in a black shirt with a skull on it"}]}
{"id": 234244, "image": "COCO_train2014_000000234244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy with the fiery skull t - shirt and baseball cap who is playing wii\"."}], "task": "refering", "answer_template": "The \"the guy with the fiery skull t - shirt and baseball cap who is playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 75, 76, 77, 98, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 281, 282, 283, 284, 305, 306, 307], "bbox": [0.2545, 0.15885416666666666, 0.42471875, 0.8127916666666667], "iscrowd": 0, "area": 23124.586200000005, "rle": {"size": [480, 640], "counts": "ec\\22n>2hDc0i6_OPIZ1[6jN\\I\\1d6fNUI^1l6dNlHb1T7`NdHe1\\4mMSL`0[Oh1b4jMQL>WOn1g4fMQL=QOR2n4cMPL;oNU2Q5aMPL:kNX2T5`MPL8jNY2W5aMnK6gN\\2Z5`MnK4eN^2]5`MnK2aNa2a5^MaKAoM`0m0c2b5^M]KESN:j0e2f5^MXKIVN4i0h2h5\\MUKNYNNf0j2m5[MPK2[NIf0l2n5ZMmJ7^NCc0o2Q6XMjJo0ROl1S6VMgJS1ROk1U6SMeJV1TOi1V6RMdJX1ROj1X6oLcJZ1ROj1[6mL_J]1ROj1]6jL^Jf5a5[J\\Ji5b5XJ[Jk5d5VJZJm5d5TJYJo5\\5n11N2N2N2N2M3N2N2N3L3N3M4L3M3M3O1;E;E;F:E3M000000O10001O1O1O1O1N2O1O2N1QHoIn6Q6PIQJP7o5nHSJS7l5kHVJU7j5jHWJV7i5hHYJY7g5dH[J\\7h5_HZJa7i5ZHYJg7i5UHXJk7k5PHWJP8\\62NM4J5K5L4K5K5L4K6J5L4K5K5J6K5J7J5K5J6K5K5L4M4K<lM]EXOn:OiEGc:@UF6V:QObFe0U<E:GRX\\5"}, "label": "the guy with the fiery skull t - shirt and baseball cap who is playing wii"}]}
{"id": 234244, "image": "COCO_train2014_000000234244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man blue shirt , blue jeans , and glasses playing game with two remotes in his hand\"."}], "task": "refering", "answer_template": "The \"man blue shirt , blue jeans , and glasses playing game with two remotes in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 63, 64, 65, 86, 87, 88, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386], "bbox": [0.5934062499999999, 0.0828125, 0.930875, 1.0], "iscrowd": 0, "area": 54609.42100000001, "rle": {"size": [480, 640], "counts": "ebb53k>4L4K4M4L3L3N2N2M3N2N2M3]Oc0E;N2fJQNaMQ2l1cNQN_1m1eNoM]1n1gNoM[1o1iNnMX1P2kNmMW1P2nNlMT1R2oNnLP2P3TNiLQ2U3RNeLS2X3PNcLU2[3nM^LX2`3kMZLZ2d3iMVL\\2g3gMUL]2i3fMRL^2l3eMoK_2n3eMmK_2Q4dMjK`2T4bMiKa2U4bMfKb2W4bMdKb2Z4aMaKc2]4`M^Kd2_4_M]Ke2a4^M[Ke2c4^MYKe2e4^MWKe2h4]MTKf2k4d32O2N1O2N2M2O2N1O2N2M2O2O2O1N2O0O2O1O1N101N2O1N101O1N101N2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2V1jNZ1eN>C00000O1gMPFBP:;XF@h9`0]F[Oc9d0cFVO^9i0hFROX9n0lFnNT9Q1oFmNQ9R1QGlNP9T1QGkNo8T1TGjNl8U1VGiNk8V1XGcNm8]1[2O1O100O1O1O100O1000001O0O10000000000000000O100000000000000O1000I7I7I7J6I8H7J7H8H8I7H7J7H8H8I7TH`KQ5e4iJaKV5a4dJeK[5\\4`JiK^5Y4^JjKb5W4ZJmKd5T4ZJoKe5R4XJPLg5R4WJPLQ5AaI`4\\1QLe4T5XKnJ[4b5aK`JQ4T6gKnIk3f6lK]Ie3W7oKoHh3Z9O10O01O1O001O10O01O1O00100O1O001O1O10O01O1O1O010O1O1O001M6K4L4L4L4K5L4L4L5K4L4K6K4H8B=D=D<F^]d0"}, "label": "man blue shirt , blue jeans , and glasses playing game with two remotes in his hand"}]}
{"id": 234244, "image": "COCO_train2014_000000234244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian on the right wearing a blue shirt\"."}], "task": "refering", "answer_template": "The \"an asian on the right wearing a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 63, 64, 65, 86, 87, 88, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386], "bbox": [0.5934062499999999, 0.0828125, 0.930875, 1.0], "iscrowd": 0, "area": 54609.42100000001, "rle": {"size": [480, 640], "counts": "ebb53k>4L4K4M4L3L3N2N2M3N2N2M3]Oc0E;N2fJQNaMQ2l1cNQN_1m1eNoM]1n1gNoM[1o1iNnMX1P2kNmMW1P2nNlMT1R2oNnLP2P3TNiLQ2U3RNeLS2X3PNcLU2[3nM^LX2`3kMZLZ2d3iMVL\\2g3gMUL]2i3fMRL^2l3eMoK_2n3eMmK_2Q4dMjK`2T4bMiKa2U4bMfKb2W4bMdKb2Z4aMaKc2]4`M^Kd2_4_M]Ke2a4^M[Ke2c4^MYKe2e4^MWKe2h4]MTKf2k4d32O2N1O2N2M2O2N1O2N2M2O2O2O1N2O0O2O1O1N101N2O1N101O1N101N2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2V1jNZ1eN>C00000O1gMPFBP:;XF@h9`0]F[Oc9d0cFVO^9i0hFROX9n0lFnNT9Q1oFmNQ9R1QGlNP9T1QGkNo8T1TGjNl8U1VGiNk8V1XGcNm8]1[2O1O100O1O1O100O1000001O0O10000000000000000O100000000000000O1000I7I7I7J6I8H7J7H8H8I7H7J7H8H8I7TH`KQ5e4iJaKV5a4dJeK[5\\4`JiK^5Y4^JjKb5W4ZJmKd5T4ZJoKe5R4XJPLg5R4WJPLQ5AaI`4\\1QLe4T5XKnJ[4b5aK`JQ4T6gKnIk3f6lK]Ie3W7oKoHh3Z9O10O01O1O001O10O01O1O00100O1O001O1O10O01O1O1O010O1O1O001M6K4L4L4L4K5L4L4L5K4L4K6K4H8B=D=D<F^]d0"}, "label": "an asian on the right wearing a blue shirt"}]}
{"id": 316170, "image": "COCO_train2014_000000316170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"some foods on a container\"."}], "task": "refering", "answer_template": "The \"some foods on a container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.295390625, 0.5293691588785047, 0.703234375, 0.8567289719626169], "iscrowd": 0, "area": 32401.468650000003, "rle": {"size": [428, 640], "counts": "WZ_26n<9K5L4K5K6J5K5K5K5K6J5K5L4K6J5K5K5K5K6J5K5M2O1O100O100O1O100O100O1O100O1O100O100O1O100O2O0O1O1000000000000000000000000000000000000000000000000001O000001O0000000000000000000000000000000000000000000001O000000000000000001O0000000000000000000000000000000001O0000000000000000000000000001O0000000000000000000001O000000000000000000000000000000001O000001O00000000O10000000000000001O00000O10000000000000000000000O101O0000000000000000000O1000000000001O000000000O100000000000001O1O1O001N2O1O001O1O1N1O2N2N2N1O2O1N2N1O2N2N7Id0\\Od0\\Oe0[Od0\\OjR_2"}, "label": "some foods on a container"}]}
{"id": 316170, "image": "COCO_train2014_000000316170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plate in the center with eggs , sausage links and patties , bacon , and a cooked tomato slice\"."}], "task": "refering", "answer_template": "The \"plate in the center with eggs , sausage links and patties , bacon , and a cooked tomato slice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.295390625, 0.5293691588785047, 0.703234375, 0.8567289719626169], "iscrowd": 0, "area": 32401.468650000003, "rle": {"size": [428, 640], "counts": "WZ_26n<9K5L4K5K6J5K5K5K5K6J5K5L4K6J5K5K5K5K6J5K5M2O1O100O100O1O100O100O1O100O1O100O100O1O100O2O0O1O1000000000000000000000000000000000000000000000000001O000001O0000000000000000000000000000000000000000000001O000000000000000001O0000000000000000000000000000000001O0000000000000000000000000001O0000000000000000000001O000000000000000000000000000000001O000001O00000000O10000000000000001O00000O10000000000000000000000O101O0000000000000000000O1000000000001O000000000O100000000000001O1O1O001N2O1O001O1O1N1O2N2N2N1O2O1N2N1O2N2N7Id0\\Od0\\Oe0[Od0\\OjR_2"}, "label": "plate in the center with eggs , sausage links and patties , bacon , and a cooked tomato slice"}]}
{"id": 316170, "image": "COCO_train2014_000000316170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a round glass on the right side of a tray\"."}], "task": "refering", "answer_template": "The \"a round glass on the right side of a tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 272, 293, 294], "bbox": [0.728421875, 0.4664953271028037, 0.9114374999999999, 0.8321028037383177], "iscrowd": 0, "area": 12274.984700000003, "rle": {"size": [428, 640], "counts": "eQS62W=;F9G8G9H9G:F:E:G6I4M3L3M4M2M4L4M2M4M2M4L3N2M3N2M3N2M3JhKVGZ4i8500O2N1O100O1O1O100O1O100O1O0100O100O100O100O100O100O100O1N2O1N2N2N2GTGPLn8o38O2O0O1O1O100O2N1O100O2N1O101N1O1O1N3M2N2N3M2M4M2N3M3M2N3M3M2N3M2N3L4M2N3M3M2N3M3N3L3N3M2M3N3L3N3L3M6K8Glbg0"}, "label": "a round glass on the right side of a tray"}]}
{"id": 316170, "image": "COCO_train2014_000000316170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty clear cup on a tray of food\"."}], "task": "refering", "answer_template": "The \"an empty clear cup on a tray of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 181, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 270, 271, 272, 293, 294], "bbox": [0.728421875, 0.4664953271028037, 0.9114374999999999, 0.8321028037383177], "iscrowd": 0, "area": 12274.984700000003, "rle": {"size": [428, 640], "counts": "eQS62W=;F9G8G9H9G:F:E:G6I4M3L3M4M2M4L4M2M4M2M4L3N2M3N2M3N2M3JhKVGZ4i8500O2N1O100O1O1O100O1O100O1O0100O100O100O100O100O100O100O1N2O1N2N2N2GTGPLn8o38O2O0O1O1O100O2N1O100O2N1O101N1O1O1N3M2N2N3M2M4M2N3M3M2N3M3M2N3M2N3L4M2N3M3M2N3M3N3L3N3M2M3N3L3N3L3M6K8Glbg0"}, "label": "an empty clear cup on a tray of food"}]}
{"id": 316170, "image": "COCO_train2014_000000316170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dish of coloured food in the bowl\"."}], "task": "refering", "answer_template": "The \"a dish of coloured food in the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 189, 190], "bbox": [0.18551562500000002, 0.28210280373831775, 0.462296875, 0.5751635514018691], "iscrowd": 0, "area": 15608.706600000003, "rle": {"size": [428, 640], "counts": "Xma1:m<?A8K4M3M2N3M3M3M3L4M3M3M2N2N2M2O2N2N1O2N101N2N101N2O0O2O1N1O2O0O2O1N100O2N100O2O0O2N100O2O0O101N10001N101O0O10001N10000O10001N10000O100O01O10O01O1O001N2G8K6O1O001O1O10000000000000000000000000000000000000000000000000000000000000001O000000000000001O000000001O0000001O0000001O000000001O000O101O0000001O1O2N1O2N1O1N3N1O2L3N2N3M2N3M2N2N3M2M4I6J6K6I6K5J7J5JPd_4"}, "label": "a dish of coloured food in the bowl"}]}
{"id": 316170, "image": "COCO_train2014_000000316170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate of fruit salad with other dishes on the tray\"."}], "task": "refering", "answer_template": "The \"a plate of fruit salad with other dishes on the tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 189, 190], "bbox": [0.18551562500000002, 0.28210280373831775, 0.462296875, 0.5751635514018691], "iscrowd": 0, "area": 15608.706600000003, "rle": {"size": [428, 640], "counts": "Xma1:m<?A8K4M3M2N3M3M3M3L4M3M3M2N2N2M2O2N2N1O2N101N2N101N2O0O2O1N1O2O0O2O1N100O2N100O2O0O2N100O2O0O101N10001N101O0O10001N10000O10001N10000O100O01O10O01O1O001N2G8K6O1O001O1O10000000000000000000000000000000000000000000000000000000000000001O000000000000001O000000001O0000001O0000001O000000001O000O101O0000001O1O2N1O2N1O1N3N1O2L3N2N3M2N3M2N2N3M2M4I6J6K6I6K5J7J5JPd_4"}, "label": "a plate of fruit salad with other dishes on the tray"}]}
{"id": 70415, "image": "COCO_train2014_000000070415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza half served on the table\"."}], "task": "refering", "answer_template": "The \"pizza half served on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 181, 182, 183, 205, 206], "bbox": [0.70646875, 0.3393240093240093, 1.0, 0.5887645687645687], "iscrowd": 0, "area": 11354.49665, "rle": {"size": [429, 640], "counts": "Uam5:R=2N2O1N2N2N2N2N2O1N2N2O1N101O1N10000O2O000O1O2N100O1O1O2O0O1O1O2O0O100O101N1ONZD_Nf;a1ZD_Ne;b1[D^Ne;b1[D_Nc;b1^D]Na;d1_D\\Na;d1_D\\N`;e1`D[N`;e1aDZN_;f1aDZN_;f1aDZN_;f1aDZN_;f140O10000000000000000O10O1000000O100O1O100O1O100O100O1O100O1O2N1O1O1O1O2N1O11O001O001O1O001O1O001O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O00100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O1O100O1OO101O0000001O00000O2O0000001O0000001E:oNQ1oNSL"}, "label": "pizza half served on the table"}]}
{"id": 70415, "image": "COCO_train2014_000000070415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three fourth of a pizza on a silver pan\"."}], "task": "refering", "answer_template": "The \"three fourth of a pizza on a silver pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 181, 182, 183, 205, 206], "bbox": [0.70646875, 0.3393240093240093, 1.0, 0.5887645687645687], "iscrowd": 0, "area": 11354.49665, "rle": {"size": [429, 640], "counts": "Uam5:R=2N2O1N2N2N2N2N2O1N2N2O1N101O1N10000O2O000O1O2N100O1O1O2O0O1O1O2O0O100O101N1ONZD_Nf;a1ZD_Ne;b1[D^Ne;b1[D_Nc;b1^D]Na;d1_D\\Na;d1_D\\N`;e1`D[N`;e1aDZN_;f1aDZN_;f1aDZN_;f1aDZN_;f140O10000000000000000O10O1000000O100O1O100O1O100O100O1O100O1O2N1O1O1O1O2N1O11O001O001O1O001O1O001O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O00100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O1O100O1OO101O0000001O00000O2O0000001O0000001E:oNQ1oNSL"}, "label": "three fourth of a pizza on a silver pan"}]}
{"id": 70415, "image": "COCO_train2014_000000070415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fresh pizza cut and sitting on a wooden tray directly next to the wine bottle\"."}], "task": "refering", "answer_template": "The \"a fresh pizza cut and sitting on a wooden tray directly next to the wine bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214], "bbox": [0.09037500000000001, 0.3685314685314685, 0.429296875, 0.6247086247086248], "iscrowd": 0, "area": 18141.9527, "rle": {"size": [429, 640], "counts": "g`h03Y=3L4M4L3L4M3L5L3M3L4M4K4M3M3L2O1N3N1N2O1O1N2O2M2O100O1O100O2N100O1O100O1O2N100O1O100O1O2O0O1O100O1O1O100O1O100O1O100O1O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000O10O1000000O1000000000O10000000001O000000000O1000000000000000000O100000000O10000O100O100O2O000O100O100O10000O100O100O10000O2O0O101N10000O2O0O100O2O0O101O0O100O2O0O101N100O101N100O101N2O0O2O1N2O1N2O1N2O1N2O1N101N2O1N2O1L4J6J6J6J6J6J5K_gh4"}, "label": "a fresh pizza cut and sitting on a wooden tray directly next to the wine bottle"}]}
{"id": 70415, "image": "COCO_train2014_000000070415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white pizza with brown sauce to the right of the other similar pizza\"."}], "task": "refering", "answer_template": "The \"the white pizza with brown sauce to the right of the other similar pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.22393749999999998, 0.6614452214452214, 0.777015625, 0.9865501165501166], "iscrowd": 0, "area": 38064.67045, "rle": {"size": [429, 640], "counts": "dgl13W=3N2M3M3M3M3N2M3M3M3N2M3M3M3N2M3M3M3N2O1O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O10000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000001O000000001O000000001O000000001O0000001O000000001O00000O101O000000001O000000001O000000001O0000001O000000001O000000001O0000001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O002N4L4L4L4L4L4L3M4L4L4L4L4L4L3H9G9Gg`k1"}, "label": "the white pizza with brown sauce to the right of the other similar pizza"}]}
{"id": 70415, "image": "COCO_train2014_000000070415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza in front in the right hand picture\"."}], "task": "refering", "answer_template": "The \"pizza in front in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.22393749999999998, 0.6614452214452214, 0.777015625, 0.9865501165501166], "iscrowd": 0, "area": 38064.67045, "rle": {"size": [429, 640], "counts": "dgl13W=3N2M3M3M3M3N2M3M3M3N2M3M3M3N2M3M3M3N2O1O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O10000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000001O000000001O000000001O000000001O0000001O000000001O00000O101O000000001O000000001O000000001O0000001O000000001O000000001O0000001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O002N4L4L4L4L4L4L3M4L4L4L4L4L4L3H9G9Gg`k1"}, "label": "pizza in front in the right hand picture"}]}
{"id": 21268, "image": "COCO_train2014_000000021268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple tablecloth with flowers on it\"."}], "task": "refering", "answer_template": "The \"purple tablecloth with flowers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 232, 233, 251, 252, 253, 254, 255, 256, 274, 275, 276, 277, 278, 279, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.0, 0.5723541666666667, 1.0, 0.9813333333333334], "iscrowd": 0, "area": 63443.743699999985, "rle": {"size": [480, 640], "counts": "o:l1S=l1TN1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1001O1O1O2N1O1O1O1O1O1OiNWKYH\\1kNk1l8PMSHS1SOl1j8WMnGk0[Ol1g8`MhGc0Ck1e8hMcG;Jl1c8jMcG8Ml1`8mMcG5Om1^8oMcG31l1\\8QNeG02m1Y8TNeGM4n1W8VNeGK6m1U8YNeGH8n1S8[NeGE:o1Q8]NeGB;Q2P8^NeG@<Q2o7`NfG\\O=S2m7bNfGYO?T2k7dNfGVOa0U2i7fNfGTOa0V2i7gNfGPOd0X2f7hNgGfNm0a2\\7jNYIU1g6lNXIT1h6mNXIR1h6oNXIP1h6QOXIn0h6SOXIl0h6UOWIk0j6ROYIm0g6oN^IP1b6lNcIS1]6iNhIV1X6fNlIZ1T6bNQJ]1o5^NWJa1i5[N\\Jd1d5XN_Ji1a5SNcJm1]5oMgJQ2Y5kMjJV2P90O100O10000O100O100O10000O100O100O1000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O10001O0000000O100000000000000O1000000000000O1000000000000O1O100O1O01O010O00010O010O00010O01O01O010O01O01O00001O001O0000VOj0D<000000O100000000O101O1O0O2O001O1N101O001N2O04M4K4M3M4K4M4K4M3M4K4M4K4M4L3L4M4K4M4L3L4M4L3L5L3L5L3M3L5LSD"}, "label": "purple tablecloth with flowers on it"}]}
{"id": 21268, "image": "COCO_train2014_000000021268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark red table cloth\"."}], "task": "refering", "answer_template": "The \"the dark red table cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 232, 233, 251, 252, 253, 254, 255, 256, 274, 275, 276, 277, 278, 279, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.0, 0.5723541666666667, 1.0, 0.9813333333333334], "iscrowd": 0, "area": 63443.743699999985, "rle": {"size": [480, 640], "counts": "o:l1S=l1TN1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1001O1O1O2N1O1O1O1O1O1OiNWKYH\\1kNk1l8PMSHS1SOl1j8WMnGk0[Ol1g8`MhGc0Ck1e8hMcG;Jl1c8jMcG8Ml1`8mMcG5Om1^8oMcG31l1\\8QNeG02m1Y8TNeGM4n1W8VNeGK6m1U8YNeGH8n1S8[NeGE:o1Q8]NeGB;Q2P8^NeG@<Q2o7`NfG\\O=S2m7bNfGYO?T2k7dNfGVOa0U2i7fNfGTOa0V2i7gNfGPOd0X2f7hNgGfNm0a2\\7jNYIU1g6lNXIT1h6mNXIR1h6oNXIP1h6QOXIn0h6SOXIl0h6UOWIk0j6ROYIm0g6oN^IP1b6lNcIS1]6iNhIV1X6fNlIZ1T6bNQJ]1o5^NWJa1i5[N\\Jd1d5XN_Ji1a5SNcJm1]5oMgJQ2Y5kMjJV2P90O100O10000O100O100O10000O100O100O1000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O10001O0000000O100000000000000O1000000000000O1000000000000O1O100O1O01O010O00010O010O00010O01O01O010O01O01O00001O001O0000VOj0D<000000O100000000O101O1O0O2O001O1N101O001N2O04M4K4M3M4K4M4K4M3M4K4M4K4M4L3L4M4K4M4L3L4M4L3L5L3L5L3M3L5LSD"}, "label": "the dark red table cloth"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt holding a game control\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt holding a game control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 89, 90, 91, 112, 113, 114, 135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367], "bbox": [0.8006718749999999, 0.17752232142857144, 1.0, 0.9887723214285715], "iscrowd": 0, "area": 34702.89805, "rle": {"size": [448, 640], "counts": "PWP71m=<eCHX:R2M3M3M2M4M3M3M3M3M2N3L4M3M3M3M2N3L4M3O5J6K6I6TI[LS3k3dKXM[4o2]JSNc5U5O100O100O100O10000O100O100O100O100O010O100O100O100O1gNeGmL[8n2kGPMV8m2nGRMR8l2QHRMP8k2UHSMk7k2XHSMi7j2[HUMe7i2_HTMb7i2bHVM^7h2eHWM[7f2jHWMW7g2oHUMQ7h2XIQMi6n2T20O1O2N1O1O1O2N1O1O1O1O2O0O2N2N2N2N2N2N2N2N3M2N21O2N2N3M2N2N3M2O1N3M2N2N3M2N3M2N2O2M2N2N2N0000000001O0PGXMh6j2SIZMl6h2oH\\MP7g2jH]MU7e2fH_MY7c2bHaM]7a2\\HeMc7]2SHlMl7V2iGTNh6]3lHmLX2f4"}, "label": "a man in a blue shirt holding a game control"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blond boy with a blue shirt playing wii\"."}], "task": "refering", "answer_template": "The \"a blond boy with a blue shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 89, 90, 91, 112, 113, 114, 135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367], "bbox": [0.8006718749999999, 0.17752232142857144, 1.0, 0.9887723214285715], "iscrowd": 0, "area": 34702.89805, "rle": {"size": [448, 640], "counts": "PWP71m=<eCHX:R2M3M3M2M4M3M3M3M3M2N3L4M3M3M3M2N3L4M3O5J6K6I6TI[LS3k3dKXM[4o2]JSNc5U5O100O100O100O10000O100O100O100O100O010O100O100O100O1gNeGmL[8n2kGPMV8m2nGRMR8l2QHRMP8k2UHSMk7k2XHSMi7j2[HUMe7i2_HTMb7i2bHVM^7h2eHWM[7f2jHWMW7g2oHUMQ7h2XIQMi6n2T20O1O2N1O1O1O2N1O1O1O1O2O0O2N2N2N2N2N2N2N2N3M2N21O2N2N3M2N2N3M2O1N3M2N2N3M2N3M2N2O2M2N2N2N0000000001O0PGXMh6j2SIZMl6h2oH\\MP7g2jH]MU7e2fH_MY7c2bHaM]7a2\\HeMc7]2SHlMl7V2iGTNh6]3lHmLX2f4"}, "label": "a blond boy with a blue shirt playing wii"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing green playing wii\"."}], "task": "refering", "answer_template": "The \"a lady wearing green playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267, 268, 269, 270, 271, 290, 291, 292, 293, 313, 314, 315, 316, 336, 337, 339, 340, 359, 360, 362, 363], "bbox": [0.5744374999999999, 0.26895089285714285, 0.8051562499999999, 0.9886830357142857], "iscrowd": 0, "area": 25836.10465000001, "rle": {"size": [448, 640], "counts": "RYQ53k=3M2M4N2N2N1N3N2N00000O101O000O100N2O1O1O1O1O1O1O1N2N2N2N3WHkNd2W1UMQOh2Q1PMXOm2j0lL^OU2`1hMdNU2^1iMeNU2\\1jMfNT2[1jMhNU2X1]KfMc0U1o3U1\\KSN9j0[4R1[KaNL`0h4o0[KmN@7T5l0[KVOXO0\\5j0ZK_OQOJd5U1lJXOXOFk5e1YJg0g5i3O1O1TOl0A?M3M3M3O10000O10000O100O1000M2L5J6K4K6J6K4L5J6K4K6K5J6L4M3M4L3M4L3M4L3N3L3M3M4L3013N2N2M[JlH`4R7[KUId4j6VKZIn4d6mJ_IW5^6dJgI_5W6\\JlIh5R6RJRJR6l64M2M4L4L4oHaIZ6X7J5K5K6K4K4L01O0000001O00000RJWJe3j5WL[Jg3e5ULaJi3_5ULfJh3Z5XLhJf3Y5ZLhJd3Y5[LjJb3Y5[LiJc3[5YLgJd3]5ZLeJc3_5YLcJe3`5XLbJf3b5VLbJZ1PODa6POeJf0_O4P6ROhJ4Le0_5TOlJA:U1m4WObMd0`2[OeM?]2_OhM<Z2BlM8V2FoM5S2JRN0P2NUNMk13[NGe1:_NAb1>P6M3HP_f1"}, "label": "a lady wearing green playing wii"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a wiimote\"."}], "task": "refering", "answer_template": "The \"a woman holding a wiimote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267, 268, 269, 270, 271, 290, 291, 292, 293, 313, 314, 315, 316, 336, 337, 339, 340, 359, 360, 362, 363], "bbox": [0.5744374999999999, 0.26895089285714285, 0.8051562499999999, 0.9886830357142857], "iscrowd": 0, "area": 25836.10465000001, "rle": {"size": [448, 640], "counts": "RYQ53k=3M2M4N2N2N1N3N2N00000O101O000O100N2O1O1O1O1O1O1O1N2N2N2N3WHkNd2W1UMQOh2Q1PMXOm2j0lL^OU2`1hMdNU2^1iMeNU2\\1jMfNT2[1jMhNU2X1]KfMc0U1o3U1\\KSN9j0[4R1[KaNL`0h4o0[KmN@7T5l0[KVOXO0\\5j0ZK_OQOJd5U1lJXOXOFk5e1YJg0g5i3O1O1TOl0A?M3M3M3O10000O10000O100O1000M2L5J6K4K6J6K4L5J6K4K6K5J6L4M3M4L3M4L3M4L3N3L3M3M4L3013N2N2M[JlH`4R7[KUId4j6VKZIn4d6mJ_IW5^6dJgI_5W6\\JlIh5R6RJRJR6l64M2M4L4L4oHaIZ6X7J5K5K6K4K4L01O0000001O00000RJWJe3j5WL[Jg3e5ULaJi3_5ULfJh3Z5XLhJf3Y5ZLhJd3Y5[LjJb3Y5[LiJc3[5YLgJd3]5ZLeJc3_5YLcJe3`5XLbJf3b5VLbJZ1PODa6POeJf0_O4P6ROhJ4Le0_5TOlJA:U1m4WObMd0`2[OeM?]2_OhM<Z2BlM8V2FoM5S2JRN0P2NUNMk13[NGe1:_NAb1>P6M3HP_f1"}, "label": "a woman holding a wiimote"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a sweatsuit playing the wii\"."}], "task": "refering", "answer_template": "The \"a man wearing a sweatsuit playing the wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 100, 101, 102, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 307, 308, 309, 330, 331, 332, 353, 354, 355], "bbox": [0.3271875, 0.22696428571428573, 0.537984375, 0.9887723214285715], "iscrowd": 0, "area": 26652.2904, "rle": {"size": [448, 640], "counts": "Ygk23Q=m0SOn0YOf0ZOf0\\Od0N3M2O2M=C>B=C=C=D<C<D<D<D=C<VJbG\\5a8O1O1O1O1O100O1O1O1O1O1O1O1N2O1oN]GdLd8Y3mGXLT8e3U1M3M3M3O1000000O10000O10K5H8I7H8I7I8G8I76J;F:E<D;E4MO000000001O5K8nF\\LX7l3]HXLb7P4RHULm7S4hGQLW8m4O1O1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N2N2N5nMT2lMT2J6J6I7J6Jl0SO2O001N2O1O1N2N2N2N1O2N2N2N2N2M3N2N1O0000001O00000000000O101O000N2J_VQ4"}, "label": "a man wearing a sweatsuit playing the wii"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male wearing glasses and a black and grey jacket playing wii\"."}], "task": "refering", "answer_template": "The \"male wearing glasses and a black and grey jacket playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 100, 101, 102, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 307, 308, 309, 330, 331, 332, 353, 354, 355], "bbox": [0.3271875, 0.22696428571428573, 0.537984375, 0.9887723214285715], "iscrowd": 0, "area": 26652.2904, "rle": {"size": [448, 640], "counts": "Ygk23Q=m0SOn0YOf0ZOf0\\Od0N3M2O2M=C>B=C=C=D<C<D<D<D=C<VJbG\\5a8O1O1O1O1O100O1O1O1O1O1O1O1N2O1oN]GdLd8Y3mGXLT8e3U1M3M3M3O1000000O10000O10K5H8I7H8I7I8G8I76J;F:E<D;E4MO000000001O5K8nF\\LX7l3]HXLb7P4RHULm7S4hGQLW8m4O1O1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N2N2N5nMT2lMT2J6J6I7J6Jl0SO2O001N2O1O1N2N2N2N1O2N2N2N2N2M3N2N1O0000001O00000000000O101O000N2J_VQ4"}, "label": "male wearing glasses and a black and grey jacket playing wii"}]}
{"id": 62233, "image": "COCO_train2014_000000062233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with beanie\"."}], "task": "refering", "answer_template": "The \"man with beanie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 47, 48, 49, 70, 71, 72, 93, 94, 95, 116, 117, 118, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351], "bbox": [0.0, 0.11908482142857144, 0.2970625, 0.9881473214285714], "iscrowd": 0, "area": 52689.66315, "rle": {"size": [448, 640], "counts": "l5^4b91N2N2N2O1N2N2N2O1N2N2O1N2N2N2O3L8H8H9H7HR2nM1O100kNbGgL_8Y3fGaL[8_3iG]LW8c3mGXLT8h3PHSLQ8m3oGRLR8Y3]GmLa0JR8P3hGTM6KS8g2RH]MKKT8^2]HeM_OLU8n1nHUNmNMU8n1PISNkNNV8n1RIRNhNOW8n1SIQNgN0W8n1UIPNdN2W8m1WIPNbN2X8n1XInM`N3Y8n1ZImM]N5Y8m1oIRNR6m1oIRNR6n1nIQNS6n1nIRNR6m1oIRNR6m1oIRNR6m1Q3000000000000O1000000000000O1000000000000O11O001O001O001O001O001O001O001O1O001O001O2N1O2N2N1O2oFeMY6]2bIjMZ6W2cIkM]6W2^IlMb6V2YImMg6T2UIoMk6S2QIoMo6S2lHQNS7Q2iHQNW7R2dHPN\\7S2`HnM`7U2[HmMe7V2WHkMi7X2SHjMl7Y2oGiMQ8S40000000001O00000000001O0000000000001O1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N2N2N3M2N3N1N3M2N3M3M2N3M4L3M3M4L3^MQImMS7m1YIiMj6R2cIcMa6W2lI^MW6^2j2J5K6J5L8G;E<D<E;DeaT6"}, "label": "man with beanie"}]}
{"id": 62245, "image": "COCO_train2014_000000062245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman hugging man\"."}], "task": "refering", "answer_template": "The \"woman hugging man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 194, 195, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 312, 332, 333, 334], "bbox": [0.43537499999999996, 0.3697916666666667, 0.5730625, 0.8731458333333334], "iscrowd": 0, "area": 8425.018600000003, "rle": {"size": [480, 640], "counts": "ToR45h>5K6J5K6J5oCUOg9_1eEgNV:m1UEXNh:[2cDjMZ;R3Le0[Om0SO`0A9F:G:F3NN2O1O1O1O1O3M3M2N3MO1OjNYGTLf8`3kG\\LT8X3]HdLb7P3oHlLP7h2aISM_6j1nGfMS2=n5P2jJmMU5U2mJhMR5[2oJbMP5`2RK]Mn4d2TKYMl4i2e31O1O2N1N2O2?@2O1N2N2O1NO101O000O2O0014K5K5L5J5K6K3L5K4M4K21K6K6J5K6JO1N101Na0@:E4L3N3L4M2M4M3L3Mibo3"}, "label": "woman hugging man"}]}
{"id": 62245, "image": "COCO_train2014_000000062245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue jacket and skis\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue jacket and skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 194, 195, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 312, 332, 333, 334], "bbox": [0.43537499999999996, 0.3697916666666667, 0.5730625, 0.8731458333333334], "iscrowd": 0, "area": 8425.018600000003, "rle": {"size": [480, 640], "counts": "ToR45h>5K6J5K6J5oCUOg9_1eEgNV:m1UEXNh:[2cDjMZ;R3Le0[Om0SO`0A9F:G:F3NN2O1O1O1O1O3M3M2N3MO1OjNYGTLf8`3kG\\LT8X3]HdLb7P3oHlLP7h2aISM_6j1nGfMS2=n5P2jJmMU5U2mJhMR5[2oJbMP5`2RK]Mn4d2TKYMl4i2e31O1O2N1N2O2?@2O1N2N2O1NO101O000O2O0014K5K5L5J5K6K3L5K4M4K21K6K6J5K6JO1N101Na0@:E4L3N3L4M2M4M3L3Mibo3"}, "label": "a woman wearing a blue jacket and skis"}]}
{"id": 504616, "image": "COCO_train2014_000000504616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man was cooking in kitchen\"."}], "task": "refering", "answer_template": "The \"man was cooking in kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 91, 92, 93, 108, 109, 110, 125, 126, 127, 128, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 210, 211, 212, 213, 226, 227, 228, 229, 230, 244, 245, 246, 247, 261, 262, 263, 264, 278, 279, 280, 281, 295, 296, 297, 298, 312, 313, 314, 315, 329, 330, 331, 332, 346, 347, 348, 349, 363, 364, 365], "bbox": [0.2230625, 0.20218750000000002, 0.5799166666666667, 0.9539843749999999], "iscrowd": 0, "area": 41345.640750000006, "rle": {"size": [640, 480], "counts": "iTS23hc05K5J6N200O101O000000000O10001O0000000O1000001O0000000O101O0000000010O001O000010O01O00001O01O01O001O00010O0O2O00001O0000[_O@i=a0SB1]=1^Bc0h2hNR5f0RHT1\\2ZN_5c0PHf1P2lMm5?oGW2d1^MY6=nGe2[1SMd69mGS3T1eLm69jGb3n0WLT79jGn3h0kK[77iG^4a0]Kc76hGl4:PKj4CVLb0aNZ56aJj4?iKHnNl54oIk4f0QL@[Nk6>QIR5g0SLm7`N]G\\5g0QLW9o3jFoKW9P4kFnKV9R4jFkKZ9T4gFhK\\9W4fFeK]9[4dFaK_9^4cF^Ka9a4`F[Kc9e4^FWKe9h4]FTKf9l4[FPKi9n4YFmJk9S5VFiJm9W5^2O2O000O101N10000O2O00000010O001TOoJgBQ5X=VKcBj4\\=\\K^Bd4a=bKZB_4e=gKUBY4j=P1O0O2O0O2O0O2O0O2O0O20O100O2O0O100O1005L6I8H7I2M2N1N2O1O1N2O1O1N2O1N2O1O1N2QN\\H`Gf7T8gHhG[7m7RIPHo6e7_IWHb6^7kI^HW6W7VJfHk5n6dJmH^5h6oJUIR5`6[K]If4X6hKcIZ4R6SLkIn3j5_LRJc3b5lLZJU3[5XMbJi2S5dMiJ_2^2<^MEV2Y9D<D<D<D<DcQn3"}, "label": "man was cooking in kitchen"}]}
{"id": 504616, "image": "COCO_train2014_000000504616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baker in a long white apron holding a metal paddle\"."}], "task": "refering", "answer_template": "The \"a baker in a long white apron holding a metal paddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 91, 92, 93, 108, 109, 110, 125, 126, 127, 128, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 210, 211, 212, 213, 226, 227, 228, 229, 230, 244, 245, 246, 247, 261, 262, 263, 264, 278, 279, 280, 281, 295, 296, 297, 298, 312, 313, 314, 315, 329, 330, 331, 332, 346, 347, 348, 349, 363, 364, 365], "bbox": [0.2230625, 0.20218750000000002, 0.5799166666666667, 0.9539843749999999], "iscrowd": 0, "area": 41345.640750000006, "rle": {"size": [640, 480], "counts": "iTS23hc05K5J6N200O101O000000000O10001O0000000O1000001O0000000O101O0000000010O001O000010O01O00001O01O01O001O00010O0O2O00001O0000[_O@i=a0SB1]=1^Bc0h2hNR5f0RHT1\\2ZN_5c0PHf1P2lMm5?oGW2d1^MY6=nGe2[1SMd69mGS3T1eLm69jGb3n0WLT79jGn3h0kK[77iG^4a0]Kc76hGl4:PKj4CVLb0aNZ56aJj4?iKHnNl54oIk4f0QL@[Nk6>QIR5g0SLm7`N]G\\5g0QLW9o3jFoKW9P4kFnKV9R4jFkKZ9T4gFhK\\9W4fFeK]9[4dFaK_9^4cF^Ka9a4`F[Kc9e4^FWKe9h4]FTKf9l4[FPKi9n4YFmJk9S5VFiJm9W5^2O2O000O101N10000O2O00000010O001TOoJgBQ5X=VKcBj4\\=\\K^Bd4a=bKZB_4e=gKUBY4j=P1O0O2O0O2O0O2O0O2O0O20O100O2O0O100O1005L6I8H7I2M2N1N2O1O1N2O1O1N2O1N2O1O1N2QN\\H`Gf7T8gHhG[7m7RIPHo6e7_IWHb6^7kI^HW6W7VJfHk5n6dJmH^5h6oJUIR5`6[K]If4X6hKcIZ4R6SLkIn3j5_LRJc3b5lLZJU3[5XMbJi2S5dMiJ_2^2<^MEV2Y9D<D<D<D<DcQn3"}, "label": "a baker in a long white apron holding a metal paddle"}]}
{"id": 430889, "image": "COCO_train2014_000000430889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a keyboard and monitor with a video playing on the screen\"."}], "task": "refering", "answer_template": "The \"a keyboard and monitor with a video playing on the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.037078125, 0.305625, 0.49381250000000004, 0.8134791666666668], "iscrowd": 0, "area": 63441.7142, "rle": {"size": [480, 640], "counts": "Qd;1m>2O2M2N2O1N2N2O1N2N2O1N2N2N2O1aI@\\Nb0c14fMNY2i0PMXOo2_1YLcNf3S2cKPN\\4f2mJ[MT5Y3UJiLm5j3\\IXLd6V4mHkKS7c4]H_Kc7o5000000000000000000000000000000000000000000O10000000000000000000000000001O00000000000000000000000000001O000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000O100000000O100O100O10000O100O100O1000000000000000000000000000000000000000000000000000O100000000000000000000000000000001O000000000000000000000000000000000004L_1aN^1bN^1bN^1bN<D8H7I8H8HU`g4"}, "label": "a keyboard and monitor with a video playing on the screen"}]}
{"id": 430889, "image": "COCO_train2014_000000430889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dekstop computer\"."}], "task": "refering", "answer_template": "The \"a dekstop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.037078125, 0.305625, 0.49381250000000004, 0.8134791666666668], "iscrowd": 0, "area": 63441.7142, "rle": {"size": [480, 640], "counts": "Qd;1m>2O2M2N2O1N2N2O1N2N2O1N2N2N2O1aI@\\Nb0c14fMNY2i0PMXOo2_1YLcNf3S2cKPN\\4f2mJ[MT5Y3UJiLm5j3\\IXLd6V4mHkKS7c4]H_Kc7o5000000000000000000000000000000000000000000O10000000000000000000000000001O00000000000000000000000000001O000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000O100000000O100O100O10000O100O100O1000000000000000000000000000000000000000000000000000O100000000000000000000000000000001O000000000000000000000000000000000004L_1aN^1bN^1bN^1bN<D8H7I8H8HU`g4"}, "label": "a dekstop computer"}]}
{"id": 430889, "image": "COCO_train2014_000000430889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open laptop on a desk\"."}], "task": "refering", "answer_template": "The \"an open laptop on a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294], "bbox": [0.6016874999999999, 0.41122916666666665, 0.857859375, 0.7460625000000001], "iscrowd": 0, "area": 23056.0599, "rle": {"size": [480, 640], "counts": "Yed5_1a=T2lM2N2N2M3N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N00000000000000000000000000000000O100000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000O10000000000000O100000000000000000000000000000000O1L4L4K5L4L4L4K5L4L4N2O1000000O10000O1000000O10000O0100000SOgDRNZ;DSG9l;N2M4M2N3M2Ma[Z1"}, "label": "an open laptop on a desk"}]}
{"id": 430889, "image": "COCO_train2014_000000430889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lap top with a weird lady head on it\"."}], "task": "refering", "answer_template": "The \"a lap top with a weird lady head on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294], "bbox": [0.6016874999999999, 0.41122916666666665, 0.857859375, 0.7460625000000001], "iscrowd": 0, "area": 23056.0599, "rle": {"size": [480, 640], "counts": "Yed5_1a=T2lM2N2N2M3N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N00000000000000000000000000000000O100000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000O10000000000000O100000000000000000000000000000000O1L4L4K5L4L4L4K5L4L4N2O1000000O10000O1000000O10000O0100000SOgDRNZ;DSG9l;N2M4M2N3M2Ma[Z1"}, "label": "a lap top with a weird lady head on it"}]}
{"id": 430889, "image": "COCO_train2014_000000430889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black keyboard sitting beside a computer monitor on a desk\"."}], "task": "refering", "answer_template": "The \"a black keyboard sitting beside a computer monitor on a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.035390624999999995, 0.7101041666666668, 0.49887499999999996, 0.8247291666666666], "iscrowd": 0, "area": 12774.42675, "rle": {"size": [480, 640], "counts": "TU;1m>2O2M2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O2M2O1O1000000000000001O000000000000000000000001O0000000000000000000000000001O00000000000000000000000000000000000000000001O00000001O0000000001O0000000000001O000000000000001O0000000000O100000000000000000000000000O1000000000000000O1000000000O100000000000000000000000000O100000000000000000O1000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000O1000000000O10000000000000O1000000000000000000000O01000000000000000000000000O1000000000O100000000000O10000000000000000000000O0100000000000000000000000O104L7I7I6J7I6IWSf4"}, "label": "a black keyboard sitting beside a computer monitor on a desk"}]}
{"id": 430889, "image": "COCO_train2014_000000430889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black computer keyboard of a desktop computer\"."}], "task": "refering", "answer_template": "The \"a black computer keyboard of a desktop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.035390624999999995, 0.7101041666666668, 0.49887499999999996, 0.8247291666666666], "iscrowd": 0, "area": 12774.42675, "rle": {"size": [480, 640], "counts": "TU;1m>2O2M2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O2M2O1O1000000000000001O000000000000000000000001O0000000000000000000000000001O00000000000000000000000000000000000000000001O00000001O0000000001O0000000000001O000000000000001O0000000000O100000000000000000000000000O1000000000000000O1000000000O100000000000000000000000000O100000000000000000O1000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000O1000000000O10000000000000O1000000000000000000000O01000000000000000000000000O1000000000O100000000000O10000000000000000000000O0100000000000000000000000O104L7I7I6J7I6IWSf4"}, "label": "a black computer keyboard of a desktop computer"}]}
{"id": 267049, "image": "COCO_train2014_000000267049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a painting of a man in glasses , standing beside a woman\"."}], "task": "refering", "answer_template": "The \"a painting of a man in glasses , standing beside a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 240, 259, 260, 261, 262, 283, 284, 285], "bbox": [0.21140625000000002, 0.255375, 0.45259375, 0.7427083333333333], "iscrowd": 0, "area": 21560.916600000004, "rle": {"size": [480, 640], "counts": "o`o11n>5L4K4L3M3N2M3M3N2M3N3M2M3N2N3L3N2N1O2M3N2N2M2O2N2N2M101N2O1N2O1O0O2O1N2O1O1N2O0O2O1O1N2O1N101O1N2O0O2N2N1O2N2VNfLkG[3U8hLgGZ3X8jLcGX3\\8kL`GV3`8lL]GV3b8lL[GV3d8lLYGU3g8nLUGT3j8nLSGS3m8oLPGS3o8PMmFR3R9RMiFo2W9UMdFm2g8UL_GR1Ek2i8Q2N3M3M2O2M3M2N3M2O2M2N3O0O2O0O2O0O100O101N100O100O1000001O000000000001O01O0000000000000001O00000O1O1O1ZLVH5k7IXH5i7H\\H6e7H^H5c7J_H4b7KaH2`7McH0^70cHM`71cHJ`75cHG`77bHFa78bHDa79cHC_7<cH@a7>bH^Oa7?cH]O_7`0eH]O^7`0fH[O^7a0gH[O\\7b0gH[O[7c0iHZOY7a0mH[OV7b0mH[OU7c0oHYOT7c0QIYOR7d0QIYOR7d0RIXOQ7e0SIXOo6d0UIYOm6a0[IUOn6b0o3I:F:FVmS5"}, "label": "a painting of a man in glasses , standing beside a woman"}]}
{"id": 267049, "image": "COCO_train2014_000000267049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man from an image on a clock , who is next to a woman\"."}], "task": "refering", "answer_template": "The \"a man from an image on a clock , who is next to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 240, 259, 260, 261, 262, 283, 284, 285], "bbox": [0.21140625000000002, 0.255375, 0.45259375, 0.7427083333333333], "iscrowd": 0, "area": 21560.916600000004, "rle": {"size": [480, 640], "counts": "o`o11n>5L4K4L3M3N2M3M3N2M3N3M2M3N2N3L3N2N1O2M3N2N2M2O2N2N2M101N2O1N2O1O0O2O1N2O1O1N2O0O2O1O1N2O1N101O1N2O0O2N2N1O2N2VNfLkG[3U8hLgGZ3X8jLcGX3\\8kL`GV3`8lL]GV3b8lL[GV3d8lLYGU3g8nLUGT3j8nLSGS3m8oLPGS3o8PMmFR3R9RMiFo2W9UMdFm2g8UL_GR1Ek2i8Q2N3M3M2O2M3M2N3M2O2M2N3O0O2O0O2O0O100O101N100O100O1000001O000000000001O01O0000000000000001O00000O1O1O1ZLVH5k7IXH5i7H\\H6e7H^H5c7J_H4b7KaH2`7McH0^70cHM`71cHJ`75cHG`77bHFa78bHDa79cHC_7<cH@a7>bH^Oa7?cH]O_7`0eH]O^7`0fH[O^7a0gH[O\\7b0gH[O[7c0iHZOY7a0mH[OV7b0mH[OU7c0oHYOT7c0QIYOR7d0QIYOR7d0RIXOQ7e0SIXOo6d0UIYOm6a0[IUOn6b0o3I:F:FVmS5"}, "label": "a man from an image on a clock , who is next to a woman"}]}
{"id": 267049, "image": "COCO_train2014_000000267049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of a woman beside a man in a clock\"."}], "task": "refering", "answer_template": "The \"a picture of a woman beside a man in a clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 125, 126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 195, 196, 197, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 309], "bbox": [0.40325, 0.28752083333333334, 0.66234375, 0.770625], "iscrowd": 0, "area": 21205.1061, "rle": {"size": [480, 640], "counts": "SYi32l>2M4M3L4M2N3L4M3M3L3N3L4M3M2M4M3M3L3N2N2N2O0O2N2O1N2N2O1N2N101N2N2O1N2O1O1O1oL`MaI`2X6kMdIV2T6UNhIl1P6_NlIb1S6bNhI`1W6dNeI]1Z6hNaIY1^6kN^IV1a6nN[IR1e6QOXIP1g6SOVIn0i6UOTIl0k6XOQIi0n6ZOoHf0Q7]OlHd0S7@iHa0V7BfH`0Y7CdH=\\7G`H:_7I^H8a7K\\H6c7NYH2g71VH0i73THNk75RHKo7c300000O10000O101O000O1000000O10000O1000000O1000000O2O000O10000O10000O10000O100O10000O2O0O101O2iLbGG`81hGNY8OjG0W8MkG2W8KlG4U8InG5T8InG6S8GoG9R8DQH:Q8CRH<o7ATH=n7ASH?n7^OUHa0l7\\OWHb0k7[OXHd0i7ZOXHe0j7XOYHg0h7VO[Hi0f7TO]Hj0e7SO]Hm0d7nNaHQ1`7fNiHX1Y7`NoH_1S7WNVIg1l6QN[Io1l9O1N2O0O2O1O1N101M3N2M3N1O2M3N2M3N1N3N2N2M3N1N3N2M3N4L3L4M3LenT3"}, "label": "a picture of a woman beside a man in a clock"}]}
{"id": 267049, "image": "COCO_train2014_000000267049.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a womans face that is on a clock\"."}], "task": "refering", "answer_template": "The \"a womans face that is on a clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 125, 126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 195, 196, 197, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 309], "bbox": [0.40325, 0.28752083333333334, 0.66234375, 0.770625], "iscrowd": 0, "area": 21205.1061, "rle": {"size": [480, 640], "counts": "SYi32l>2M4M3L4M2N3L4M3M3L3N3L4M3M2M4M3M3L3N2N2N2O0O2N2O1N2N2O1N2N101N2N2O1N2O1O1O1oL`MaI`2X6kMdIV2T6UNhIl1P6_NlIb1S6bNhI`1W6dNeI]1Z6hNaIY1^6kN^IV1a6nN[IR1e6QOXIP1g6SOVIn0i6UOTIl0k6XOQIi0n6ZOoHf0Q7]OlHd0S7@iHa0V7BfH`0Y7CdH=\\7G`H:_7I^H8a7K\\H6c7NYH2g71VH0i73THNk75RHKo7c300000O10000O101O000O1000000O10000O1000000O1000000O2O000O10000O10000O10000O100O10000O2O0O101O2iLbGG`81hGNY8OjG0W8MkG2W8KlG4U8InG5T8InG6S8GoG9R8DQH:Q8CRH<o7ATH=n7ASH?n7^OUHa0l7\\OWHb0k7[OXHd0i7ZOXHe0j7XOYHg0h7VO[Hi0f7TO]Hj0e7SO]Hm0d7nNaHQ1`7fNiHX1Y7`NoH_1S7WNVIg1l6QN[Io1l9O1N2O0O2O1O1N101M3N2M3N1O2M3N2M3N1N3N2N2M3N1N3N2M3N4L3L4M3LenT3"}, "label": "a womans face that is on a clock"}]}
{"id": 21292, "image": "COCO_train2014_000000021292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby holding a banana waste in his hand\"."}], "task": "refering", "answer_template": "The \"a baby holding a banana waste in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.16507812500000002, 0.04, 0.78809375, 1.0], "iscrowd": 0, "area": 87746.36444999998, "rle": {"size": [426, 640], "counts": "T]\\1l0Z<6I7J7I6J4O2N2N1O2O0O2O0O2O000O10000000000000000000000000000000000000000000O1000000000O10000000000000000O10000000000000000O1000O100000000000O10000000000000000O10000O100O100O100O100O010O100O100O100O100O1O1O1O1N10000O1000000O1cGZMW6f2gI^MW6b2fIcMW6^2gIfMW6[2eIjMY6V2eImMZ6S2cIQN[6Q2aISN^6m1^IXNa6i1ZI\\Ne6d1WIaNh6`1VIbNj6]1VIdNi6]1UIeNj6[1UIgNk6Y1SIiNl6W1SIkNm6U1RIlNm6T1RInNm6S1QIoNo6P1PIROo6o0oHSOQ7l0oHUOP7l0nHVOQ7j0nHXOR7h0lHZOS7f0lH\\OT7d0iH_OV7b0fHBY7>gHCY7=eHEZ7;fHFZ7:dHH[78eHIZ78dHJ\\75dHL[7R300O100O100]JQKj2P5UM]M?c2A`M<`2CdM9]2FgM7Y2HkM4V2KmM3S2MPN0P2OTNMm12WNKi14[NHf18]NEc1:_NEa1:`NEa1:aNE_1:cNE]1:eNE[1:fNFZ19hNFX19jNEW1:jNFV19lNFT18oNGQ18QOGo08ROGo08SOGm08UOGk08WOGi08XOGi07ZOHf07[OIe07[OIe07[OHf07\\OHd08\\OHd08\\OGe08\\OHd08\\OHd08]OGc09^OEc0:_OEa0;AC?=BA??C_O=`0F^O:b0G]O9c0W6O1000000O2O00000O100000000000000000000O10000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O1O001O001N2O001O001O1O001N2O001O001O001N101O001O001O0O2O001O001O0ZIkNW1V1dNnN[1S1bNQO]1P1^NUOa1k0[NYOe1h0VN]Oi1c0TN@l1a0oMDP2=kMGT2;gMJX27dML\\25_M0`21[M3e2OUM6j2KQM9o2HmL<R3EiL?V3CeLb0Z3_OaLe0_3\\O\\Lj0b3WOYLo0e3XOPLn0n3^ObKh0\\4AWKe0f4^OTKh0j4[OoJk0o4XOkJm0S5VOfJP1X5RObJT1\\5_3N001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O3M3M2N3M2M4M3M2N3L:G:F:E8I4L4K4M4L3L5L4L3L5L4L3L5L3M4L4K4M4L4KkUh1"}, "label": "a baby holding a banana waste in his hand"}]}
{"id": 21292, "image": "COCO_train2014_000000021292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby holding a bannana peal\"."}], "task": "refering", "answer_template": "The \"a baby holding a bannana peal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.16507812500000002, 0.04, 0.78809375, 1.0], "iscrowd": 0, "area": 87746.36444999998, "rle": {"size": [426, 640], "counts": "T]\\1l0Z<6I7J7I6J4O2N2N1O2O0O2O0O2O000O10000000000000000000000000000000000000000000O1000000000O10000000000000000O10000000000000000O1000O100000000000O10000000000000000O10000O100O100O100O100O010O100O100O100O100O1O1O1O1N10000O1000000O1cGZMW6f2gI^MW6b2fIcMW6^2gIfMW6[2eIjMY6V2eImMZ6S2cIQN[6Q2aISN^6m1^IXNa6i1ZI\\Ne6d1WIaNh6`1VIbNj6]1VIdNi6]1UIeNj6[1UIgNk6Y1SIiNl6W1SIkNm6U1RIlNm6T1RInNm6S1QIoNo6P1PIROo6o0oHSOQ7l0oHUOP7l0nHVOQ7j0nHXOR7h0lHZOS7f0lH\\OT7d0iH_OV7b0fHBY7>gHCY7=eHEZ7;fHFZ7:dHH[78eHIZ78dHJ\\75dHL[7R300O100O100]JQKj2P5UM]M?c2A`M<`2CdM9]2FgM7Y2HkM4V2KmM3S2MPN0P2OTNMm12WNKi14[NHf18]NEc1:_NEa1:`NEa1:aNE_1:cNE]1:eNE[1:fNFZ19hNFX19jNEW1:jNFV19lNFT18oNGQ18QOGo08ROGo08SOGm08UOGk08WOGi08XOGi07ZOHf07[OIe07[OIe07[OHf07\\OHd08\\OHd08\\OGe08\\OHd08\\OHd08]OGc09^OEc0:_OEa0;AC?=BA??C_O=`0F^O:b0G]O9c0W6O1000000O2O00000O100000000000000000000O10000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O1O001O001N2O001O001O1O001N2O001O001O001N101O001O001O0O2O001O001O0ZIkNW1V1dNnN[1S1bNQO]1P1^NUOa1k0[NYOe1h0VN]Oi1c0TN@l1a0oMDP2=kMGT2;gMJX27dML\\25_M0`21[M3e2OUM6j2KQM9o2HmL<R3EiL?V3CeLb0Z3_OaLe0_3\\O\\Lj0b3WOYLo0e3XOPLn0n3^ObKh0\\4AWKe0f4^OTKh0j4[OoJk0o4XOkJm0S5VOfJP1X5RObJT1\\5_3N001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O3M3M2N3M2M4M3M2N3L:G:F:E8I4L4K4M4L3L5L4L3L5L4L3L5L3M4L4K4M4L4KkUh1"}, "label": "a baby holding a bannana peal"}]}
{"id": 21292, "image": "COCO_train2014_000000021292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"adult\"."}], "task": "refering", "answer_template": "The \"adult\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215], "bbox": [0.0850625, 0.002511737089201878, 0.422359375, 0.6616901408450704], "iscrowd": 0, "area": 30807.424549999996, "rle": {"size": [426, 640], "counts": "Rof01Y=0O2O1O001O1O1O001O1N101O1O001O1O1O001N10001O0000000O100O2O000O10000O101O0O10000O100O2O0O1O100O1O1O2N1O100O1O1O1O2N1O1O1O1O1O101N1O1OO10O000O2N101N1O2N101N1O2O0O2N101N1O2O0O2N1O2O0011N2O1N2O1N2OcJZNXOd1`0jN\\OT1;\\O@c06LF21=KFHe07^O@h0`0\\OVOj0j0YOnNl0R1WOfNm0[1WO[No0e1TOTLbMV1_3f2NnL8R3MdL8\\3NYL7g33iK3W4`4000000000000000O2O00000000000000000000000000000O100000000000000L4M3L4L4L4L4L4N21O1O1O1O1O1O1O001O1O1O1O1O1O1O00O1O1O1O1O1O1O1O1O1O1O1O1O2N1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1N2_JTKb2n4VMSMo0o2hNeMk0]2mNQNk0Q2mN^Ni0e1oNiNi0S7H8H9G\\gi4"}, "label": "adult"}]}
{"id": 21292, "image": "COCO_train2014_000000021292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing in the back with hands on counter\"."}], "task": "refering", "answer_template": "The \"woman standing in the back with hands on counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215], "bbox": [0.0850625, 0.002511737089201878, 0.422359375, 0.6616901408450704], "iscrowd": 0, "area": 30807.424549999996, "rle": {"size": [426, 640], "counts": "Rof01Y=0O2O1O001O1O1O001O1N101O1O001O1O1O001N10001O0000000O100O2O000O10000O101O0O10000O100O2O0O1O100O1O1O2N1O100O1O1O1O2N1O1O1O1O1O101N1O1OO10O000O2N101N1O2N101N1O2O0O2N101N1O2O0O2N1O2O0011N2O1N2O1N2OcJZNXOd1`0jN\\OT1;\\O@c06LF21=KFHe07^O@h0`0\\OVOj0j0YOnNl0R1WOfNm0[1WO[No0e1TOTLbMV1_3f2NnL8R3MdL8\\3NYL7g33iK3W4`4000000000000000O2O00000000000000000000000000000O100000000000000L4M3L4L4L4L4L4N21O1O1O1O1O1O1O001O1O1O1O1O1O1O00O1O1O1O1O1O1O1O1O1O1O1O1O2N1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1N2_JTKb2n4VMSMo0o2hNeMk0]2mNQNk0Q2mN^Ni0e1oNiNi0S7H8H9G\\gi4"}, "label": "woman standing in the back with hands on counter"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwich with no toothpick\"."}], "task": "refering", "answer_template": "The \"the half of the sandwich with no toothpick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 258, 259, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349, 350, 351, 352, 372, 373, 374], "bbox": [0.150671875, 0.6202083333333334, 0.33182812500000003, 0.9678333333333333], "iscrowd": 0, "area": 12435.7805, "rle": {"size": [480, 640], "counts": "ik]12l>7I5K6J5K4L5K5K4L5K5K4L6J5K5K3M3M2O1N3N1O2M2O1O2M2O2N1N2O2N1O2M2O2N1N2O2N1N3N1N2O1N2O1N101N2O1N2O0O2O1N2O1N101N2O1N101N2O1N2O0O2O1N2O10O01O1O10O01O1O1002M2O2N2M3M3M2N3M3M3M2N3M3M3M3M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M3L5L4L3M4L4K4M5K_XX6"}, "label": "the half of the sandwich with no toothpick"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwich next to the tomato slices\"."}], "task": "refering", "answer_template": "The \"the half of the sandwich next to the tomato slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 258, 259, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349, 350, 351, 352, 372, 373, 374], "bbox": [0.150671875, 0.6202083333333334, 0.33182812500000003, 0.9678333333333333], "iscrowd": 0, "area": 12435.7805, "rle": {"size": [480, 640], "counts": "ik]12l>7I5K6J5K4L5K5K4L5K5K4L6J5K5K3M3M2O1N3N1O2M2O1O2M2O2N1N2O2N1O2M2O2N1N2O2N1N3N1N2O1N2O1N101N2O1N2O0O2O1N2O1N101N2O1N101N2O1N2O0O2O1N2O10O01O1O10O01O1O1002M2O2N2M3M3M2N3M3M3M2N3M3M3M3M2N3M2N3M2N3M3M2N3M2N3M2N3M2N3M3L5L4L3M4L4K4M5K_XX6"}, "label": "the half of the sandwich next to the tomato slices"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food on tabule\"."}], "task": "refering", "answer_template": "The \"a food on tabule\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 301, 302, 324], "bbox": [0.06425, 0.5550625, 0.257953125, 0.8525208333333334], "iscrowd": 0, "area": 10236.4574, "rle": {"size": [480, 640], "counts": "l`c02i>8H8H8H8H8H8M3L4M3L4L4M3L4L4M2M4L4M3L4L3N3L3M4M2M4M2M4L3N3L9G2OO001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O2M2O1O1N3N1O1N2O1N3N1O1N2O2N1N3N2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O3M3M2N3M2N3M2N3M2N3M3M2N3M2N3M2N`[n6"}, "label": "a food on tabule"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left half of sandwich\"."}], "task": "refering", "answer_template": "The \"left half of sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 301, 302, 324], "bbox": [0.06425, 0.5550625, 0.257953125, 0.8525208333333334], "iscrowd": 0, "area": 10236.4574, "rle": {"size": [480, 640], "counts": "l`c02i>8H8H8H8H8H8M3L4M3L4L4M3L4L4M2M4L4M3L4L3N3L3M4M2M4M2M4L3N3L9G2OO001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O2M2O1O1N3N1O1N2O1N3N1O1N2O2N1N3N2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O3M3M2N3M2N3M2N3M2N3M3M2N3M2N3M2N`[n6"}, "label": "left half of sandwich"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl of soup on the right plate\"."}], "task": "refering", "answer_template": "The \"the bowl of soup on the right plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227], "bbox": [0.757046875, 0.29810416666666667, 0.9583124999999999, 0.5540208333333333], "iscrowd": 0, "area": 12395.70085, "rle": {"size": [480, 640], "counts": "QaS7i0n=<J6K4K6K5K5K3M3M4M2M3M3O2M2N2N3N1N2N2N3N1N2N2N2O0O2N1O2O000O2O001N10001O0O2O001O0O101O000O10000O100000000001O00000000000010O000000000001O0000000000001N10000O2O001N101O0O2O001N101O0O2O001N101O1N101O1N2N2N2N2N2M2O2N2N2N2N2N2N2N3L3N3M2N3L4M2M4M2N4K6K4K6G8En^<"}, "label": "the bowl of soup on the right plate"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl of tomato soup\"."}], "task": "refering", "answer_template": "The \"bowl of tomato soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227], "bbox": [0.757046875, 0.29810416666666667, 0.9583124999999999, 0.5540208333333333], "iscrowd": 0, "area": 12395.70085, "rle": {"size": [480, 640], "counts": "QaS7i0n=<J6K4K6K5K5K3M3M4M2M3M3O2M2N2N3N1N2N2N3N1N2N2N2O0O2N1O2O000O2O001N10001O0O2O001O0O101O000O10000O100000000001O00000000000010O000000000001O0000000000001N10000O2O001N101O0O2O001N101O0O2O001N101O1N101O1N2N2N2N2N2M2O2N2N2N2N2N2N2N3L3N3M2N3L4M2M4M2N4K6K4K6G8En^<"}, "label": "bowl of tomato soup"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the spoon sitting on top of the plate\"."}], "task": "refering", "answer_template": "The \"the spoon sitting on top of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 198, 199, 200, 223, 246, 247, 270, 293, 294], "bbox": [0.613484375, 0.3550625, 0.8022500000000001, 0.7595416666666668], "iscrowd": 0, "area": 4377.835900000004, "rle": {"size": [480, 640], "counts": "R]h5`0P>a0J7N1N3N1N3N1N3O1N101N1O2O0O2N101N2N10O000010O0001O010O00101N1O2N101N2N101N1O2N2O001000O100O01ROhB2S>10000O100000O01000000O10000O0100000O10000O1000O10O10000O100000O010000O1000000O01000O1000000O1000O01000000O1000000ON3N2M2O2N2M3N1O2M3N1O2M3N2MkUk1"}, "label": "the spoon sitting on top of the plate"}]}
{"id": 176945, "image": "COCO_train2014_000000176945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver spoon sitting on a white plate\"."}], "task": "refering", "answer_template": "The \"a silver spoon sitting on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 198, 199, 200, 223, 246, 247, 270, 293, 294], "bbox": [0.613484375, 0.3550625, 0.8022500000000001, 0.7595416666666668], "iscrowd": 0, "area": 4377.835900000004, "rle": {"size": [480, 640], "counts": "R]h5`0P>a0J7N1N3N1N3N1N3O1N101N1O2O0O2N101N2N10O000010O0001O010O00101N1O2N101N2N101N1O2N2O001000O100O01ROhB2S>10000O100000O01000000O10000O0100000O10000O1000O10O10000O100000O010000O1000000O01000O1000000O1000O01000000O1000000ON3N2M2O2N2M3N1O2M3N1O2M3N2MkUk1"}, "label": "a silver spoon sitting on a white plate"}]}
{"id": 37682, "image": "COCO_train2014_000000037682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfboard near a girl resting on some rocks\"."}], "task": "refering", "answer_template": "The \"a surfboard near a girl resting on some rocks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 225, 226, 227, 228, 246, 247, 248], "bbox": [0.7099375, 0.5349882903981265, 0.9991875, 0.7314519906323185], "iscrowd": 0, "area": 3304.8002999999994, "rle": {"size": [427, 640], "counts": "fPn51Z=1O1N101O1SCKa<6]CLc<5[CLe<5XCLh<5WCLi<;0O01001O0O101O001N101O1N10GK[C6e<JZC7e<JYC7h<600000O1000O01000O0100O01O10O0100O010O010O10O010O01000O010O0100O010O010O10O010O10O10O010O10O10O010O10O10O0100O010O010O10O010O0100O010O00100O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O010O10O10O10O01000O0100O01000O01W6"}, "label": "a surfboard near a girl resting on some rocks"}]}
{"id": 37682, "image": "COCO_train2014_000000037682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfboard leaning up against the hill\"."}], "task": "refering", "answer_template": "The \"the surfboard leaning up against the hill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 225, 226, 227, 228, 246, 247, 248], "bbox": [0.7099375, 0.5349882903981265, 0.9991875, 0.7314519906323185], "iscrowd": 0, "area": 3304.8002999999994, "rle": {"size": [427, 640], "counts": "fPn51Z=1O1N101O1SCKa<6]CLc<5[CLe<5XCLh<5WCLi<;0O01001O0O101O001N101O1N10GK[C6e<JZC7e<JYC7h<600000O1000O01000O0100O01O10O0100O010O010O10O010O01000O010O0100O010O010O10O010O10O10O010O10O10O010O10O10O0100O010O010O10O010O0100O010O00100O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O010O10O10O010O10O10O10O01000O0100O01000O01W6"}, "label": "the surfboard leaning up against the hill"}]}
{"id": 37682, "image": "COCO_train2014_000000037682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a bikini top and surf suit standing on a beach looking at the ocean\"."}], "task": "refering", "answer_template": "The \"a woman in a bikini top and surf suit standing on a beach looking at the ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 60, 61, 83, 84, 85, 106, 107, 108, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 243, 244, 245, 246, 266, 267, 268, 269, 291, 312, 313, 314], "bbox": [0.58809375, 0.0809367681498829, 0.714078125, 0.9339812646370024], "iscrowd": 0, "area": 18263.10220000001, "rle": {"size": [427, 640], "counts": "aPm41S=8J6K5K5K4L5mEVO`7m0^HWO]7l0nGJg0SOl4U1UJ0n0mNj4U1PJ6V1fNf4W1lI<\\1^Nc4[1lI<a1YNl0Ia0f1mL=f1TNk0JjN5[ONmNg19<j1PNj0KiN7ZOk1SO9P2jMh0MiN8UO`4[1[Ke0OiN9oNc4c1VK5HRN8U1:lN`4l1VKMc0XOJkN\\4k1RKPN5P2f0WOJjNX4R2TKmM4P2h0TOIjNV4Y2UKiM4o1k0QOg4[1ZJeM4n1n0nNb4c1\\JaM3n1R1kN\\4j1_J]M3m1T1eN]4T2\\JZM3m1V1[N`4a2WJWM3m1W1WN_4h2WJSM4n1X1SN_4l2UJRM5n1Y1RN`4l2RJSM6o1d7h0cHXO`7e0`H[Od7`0]H@g7<YHDj79VHGm74UHLm71THOo7MRH3Z8@gG`0b8VO_Gj0g8oNZGQ1l8hNUGX1R9UNZGk1S:000000001O00001O001O2gEPNS9S2hFRNU9T2eFoMX9[2^FfMb9[2\\FfMc9[2\\FfMb9]2[FeMd9\\2[FeMd9R3N1D_LnFb3Q9bLiFa3V9;N2cLnKaL5[Nn3h4^LbL6WN_3X2fK5W1QOY4f0eJIf1VOj3n0eJGk1oNf3U1eJHS2`Na3a1cJK_8MhGO]8LgG0_8LcG0b8B^F[OR1P1g8WO`FHj0n0d9PO_F3[OMh;OgT\\2"}, "label": "a woman in a bikini top and surf suit standing on a beach looking at the ocean"}]}
{"id": 37682, "image": "COCO_train2014_000000037682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a hat , standing on the beach\"."}], "task": "refering", "answer_template": "The \"a woman in a hat , standing on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 60, 61, 83, 84, 85, 106, 107, 108, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 243, 244, 245, 246, 266, 267, 268, 269, 291, 312, 313, 314], "bbox": [0.58809375, 0.0809367681498829, 0.714078125, 0.9339812646370024], "iscrowd": 0, "area": 18263.10220000001, "rle": {"size": [427, 640], "counts": "aPm41S=8J6K5K5K4L5mEVO`7m0^HWO]7l0nGJg0SOl4U1UJ0n0mNj4U1PJ6V1fNf4W1lI<\\1^Nc4[1lI<a1YNl0Ia0f1mL=f1TNk0JjN5[ONmNg19<j1PNj0KiN7ZOk1SO9P2jMh0MiN8UO`4[1[Ke0OiN9oNc4c1VK5HRN8U1:lN`4l1VKMc0XOJkN\\4k1RKPN5P2f0WOJjNX4R2TKmM4P2h0TOIjNV4Y2UKiM4o1k0QOg4[1ZJeM4n1n0nNb4c1\\JaM3n1R1kN\\4j1_J]M3m1T1eN]4T2\\JZM3m1V1[N`4a2WJWM3m1W1WN_4h2WJSM4n1X1SN_4l2UJRM5n1Y1RN`4l2RJSM6o1d7h0cHXO`7e0`H[Od7`0]H@g7<YHDj79VHGm74UHLm71THOo7MRH3Z8@gG`0b8VO_Gj0g8oNZGQ1l8hNUGX1R9UNZGk1S:000000001O00001O001O2gEPNS9S2hFRNU9T2eFoMX9[2^FfMb9[2\\FfMc9[2\\FfMb9]2[FeMd9\\2[FeMd9R3N1D_LnFb3Q9bLiFa3V9;N2cLnKaL5[Nn3h4^LbL6WN_3X2fK5W1QOY4f0eJIf1VOj3n0eJGk1oNf3U1eJHS2`Na3a1cJK_8MhGO]8LgG0_8LcG0b8B^F[OR1P1g8WO`FHj0n0d9PO_F3[OMh;OgT\\2"}, "label": "a woman in a hat , standing on the beach"}]}
{"id": 144179, "image": "COCO_train2014_000000144179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large multi striped chair\"."}], "task": "refering", "answer_template": "The \"large multi striped chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340], "bbox": [0.539546875, 0.5898529411764706, 0.840046875, 0.9830637254901962], "iscrowd": 0, "area": 20514.415299999997, "rle": {"size": [408, 640], "counts": "TUZ45b<3L2O001N101O1N101O0O2O1N1010O100O010O1O010O1O10O01O10O01OiFDT6<gIIY66cIO\\62_I3a6L[I9e6FSIc0l6^OkHk0U7TOdHT1\\7kN\\H^1c7cNTHf1l7YNlGQ2R8oMgGY2Y8gMaG_2^8aM^Gd2b8[M[Gi2e8WMVGn2i8RMTGR3l8=O100O100O10000O100O100M3M3M3M3M3M3M3M3L4M3N20000000000001O00000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O0000000000001O00000000000I7G:F9G9G9G9I7H8I7I8H7L4M3M3M3M3M4L3M5L3M3N2N3L3N2N2N2M4M2NkhX1"}, "label": "large multi striped chair"}]}
{"id": 144179, "image": "COCO_train2014_000000144179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"single sofa with vertical geometrical lines on which a red cloth is placed\"."}], "task": "refering", "answer_template": "The \"single sofa with vertical geometrical lines on which a red cloth is placed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340], "bbox": [0.539546875, 0.5898529411764706, 0.840046875, 0.9830637254901962], "iscrowd": 0, "area": 20514.415299999997, "rle": {"size": [408, 640], "counts": "TUZ45b<3L2O001N101O1N101O0O2O1N1010O100O010O1O010O1O10O01O10O01OiFDT6<gIIY66cIO\\62_I3a6L[I9e6FSIc0l6^OkHk0U7TOdHT1\\7kN\\H^1c7cNTHf1l7YNlGQ2R8oMgGY2Y8gMaG_2^8aM^Gd2b8[M[Gi2e8WMVGn2i8RMTGR3l8=O100O100O10000O100O100M3M3M3M3M3M3M3M3L4M3N20000000000001O00000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O0000000000001O00000000000I7G:F9G9G9G9I7H8I7I8H7L4M3M3M3M3M4L3M5L3M3N2N3L3N2N2N2M4M2NkhX1"}, "label": "single sofa with vertical geometrical lines on which a red cloth is placed"}]}
{"id": 193333, "image": "COCO_train2014_000000193333.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man without a shirt , red shorts , dark hair , and is standing up with his hands spread out\"."}], "task": "refering", "answer_template": "The \"the man without a shirt , red shorts , dark hair , and is standing up with his hands spread out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 58, 59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 266, 267, 268, 269, 270, 289, 290, 291, 292, 312, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360], "bbox": [0.48420312499999996, 0.10552083333333333, 0.775390625, 0.9384791666666666], "iscrowd": 0, "area": 37913.674049999994, "rle": {"size": [480, 640], "counts": "dca42m>2M2N3M3N1N3M3M2O2M1O1O100O1O1O100O1O00100O1O1O1O1O1O1hFlNd4U1[KmNdNOh2U1cNnNdN8]2k0nNoNdN`0S2b0XOPOdNg0j1:AQOeNg0f18ESOdNh0c16HTOdNh0a15KTOdNi0]14NUOdNj0Y132TOeNl0S137SOfNk0o04:SOfNl0j06=POhNm0f08>mNkNm0b0==iNQOl0<a0?eNUOk08f0>aNZOj03k0?]N]Ok0On0?YNBj0JS1`0UNFj0EW1?RNLh0@a1;hM6i0ZOi17`M>i0lM]3\\1lKh0h0jMd3U1fKQ1h0fMl3o0^K[1g0dMT4f0XKe1g0aMZ4a0QKn1f0^Mc4:iJX2e0\\Mj43cJa2e0XMP7h2d2O010000000001O001O1N1O2O1N1O2SNWMaGj2U8cMfG`2o7lMmGV2i7VNTHk1b7bNYH`1`7iN\\HZ1`7kN]HV1_7QO\\HQ1`7UO\\Hn0`7WO]Hj0_7]O\\He0`7A\\Hb0`7C]H>_7d3L5N3O0O2N2N5K00O10000O103M4K4M4L3L6K5K4K5L4L4K5L4L4K4M4L4K5L6I8H9H8G7I01N10001N100O2N3J6K4L5K5K4L5K5K4YN^JkIg5Q6dJcI`5Z6jJ[I[5a6PKSIU5i6c1M4L4L3M4L4L3M4L3M4K4M3M4L3N2M4L3M4L3N2M4L3M3N3L3M3M4L3N3L4L6J6J7J5J6J6J6J7D;B>A?B?ASYS2"}, "label": "the man without a shirt , red shorts , dark hair , and is standing up with his hands spread out"}]}
{"id": 193333, "image": "COCO_train2014_000000193333.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfer in orange trucks and no shirt riding a surfboard\"."}], "task": "refering", "answer_template": "The \"a surfer in orange trucks and no shirt riding a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 58, 59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 266, 267, 268, 269, 270, 289, 290, 291, 292, 312, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360], "bbox": [0.48420312499999996, 0.10552083333333333, 0.775390625, 0.9384791666666666], "iscrowd": 0, "area": 37913.674049999994, "rle": {"size": [480, 640], "counts": "dca42m>2M2N3M3N1N3M3M2O2M1O1O100O1O1O100O1O00100O1O1O1O1O1O1hFlNd4U1[KmNdNOh2U1cNnNdN8]2k0nNoNdN`0S2b0XOPOdNg0j1:AQOeNg0f18ESOdNh0c16HTOdNh0a15KTOdNi0]14NUOdNj0Y132TOeNl0S137SOfNk0o04:SOfNl0j06=POhNm0f08>mNkNm0b0==iNQOl0<a0?eNUOk08f0>aNZOj03k0?]N]Ok0On0?YNBj0JS1`0UNFj0EW1?RNLh0@a1;hM6i0ZOi17`M>i0lM]3\\1lKh0h0jMd3U1fKQ1h0fMl3o0^K[1g0dMT4f0XKe1g0aMZ4a0QKn1f0^Mc4:iJX2e0\\Mj43cJa2e0XMP7h2d2O010000000001O001O1N1O2O1N1O2SNWMaGj2U8cMfG`2o7lMmGV2i7VNTHk1b7bNYH`1`7iN\\HZ1`7kN]HV1_7QO\\HQ1`7UO\\Hn0`7WO]Hj0_7]O\\He0`7A\\Hb0`7C]H>_7d3L5N3O0O2N2N5K00O10000O103M4K4M4L3L6K5K4K5L4L4K5L4L4K4M4L4K5L6I8H9H8G7I01N10001N100O2N3J6K4L5K5K4L5K5K4YN^JkIg5Q6dJcI`5Z6jJ[I[5a6PKSIU5i6c1M4L4L3M4L4L3M4L3M4K4M3M4L3N2M4L3M4L3N2M4L3M3N3L3M3M4L3N3L4L6J6J7J5J6J6J6J7D;B>A?B?ASYS2"}, "label": "a surfer in orange trucks and no shirt riding a surfboard"}]}
{"id": 226102, "image": "COCO_train2014_000000226102.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a english muffin breakfast sandwhich\"."}], "task": "refering", "answer_template": "The \"a english muffin breakfast sandwhich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 293, 294, 295, 296, 297, 298, 311, 312, 313], "bbox": [0.23370833333333335, 0.617984375, 0.6591666666666666, 0.8292187500000001], "iscrowd": 0, "area": 16698.67025, "rle": {"size": [640, 480], "counts": "m^V22hc07I7K5M3M3M3N2M3M3M3M3N2M3M3M3N2O001N101N100O2O001N101N100O2O001N101N101O0O2O0O2O1N101O0O2O0O2O1N2O1O0O2O1N2O1O1N101N2O1N2O001N2O1N2O1N101O1N100O2O000O100O100O2O000O100OI80O01O10O010O01O10O010O01O010O10O01O010O010O1O010O010O1O010O010O1O100O100O1O100O101N1O100O100O1O100O2O0O1O100O100O1O2O0O100O1O100O101N1O100O100O1O101N100O1O100O1O1O2O0O1O1O100O1O1O2O0O1O1O2O0O2N1O2O0O2N1O2O0O2N2N3N1N2N3M2O1N3M2N2O2M2N9G`RV3"}, "label": "a english muffin breakfast sandwhich"}]}
{"id": 226102, "image": "COCO_train2014_000000226102.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left piece of sandwich in the right hand picture\"."}], "task": "refering", "answer_template": "The \"left piece of sandwich in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 293, 294, 295, 296, 297, 298, 311, 312, 313], "bbox": [0.23370833333333335, 0.617984375, 0.6591666666666666, 0.8292187500000001], "iscrowd": 0, "area": 16698.67025, "rle": {"size": [640, 480], "counts": "m^V22hc07I7K5M3M3M3N2M3M3M3M3N2M3M3M3N2O001N101N100O2O001N101N100O2O001N101N101O0O2O0O2O1N101O0O2O0O2O1N2O1O0O2O1N2O1O1N101N2O1N2O001N2O1N2O1N101O1N100O2O000O100O100O2O000O100OI80O01O10O010O01O10O010O01O010O10O01O010O010O1O010O010O1O010O010O1O100O100O1O100O101N1O100O100O1O100O2O0O1O100O100O1O2O0O100O1O100O101N1O100O100O1O101N100O1O100O1O1O2O0O1O1O100O1O1O2O0O1O1O2O0O2N1O2O0O2N1O2O0O2N2N3N1N2N3M2O1N3M2N2O2M2N9G`RV3"}, "label": "left piece of sandwich in the right hand picture"}]}
{"id": 226102, "image": "COCO_train2014_000000226102.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"breakfast sandwich with egg , meat and toasted bread\"."}], "task": "refering", "answer_template": "The \"breakfast sandwich with egg , meat and toasted bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 316, 317, 318, 319, 320, 321, 322, 336, 337, 338, 339, 356], "bbox": [0.5988333333333333, 0.6286562499999999, 1.0, 0.8790312499999999], "iscrowd": 0, "area": 20915.840000000007, "rle": {"size": [640, 480], "counts": "Zkc51mc04K5L4K5L4K5L3L4M2M3O2N1N2O2N1O1N3N1O1O2M2O1O1O2ROSNe_Oo1[`0SNb_On1]`0VN__Ol1``0VN]_Ok1c`0WNY_Ok1f`0XNW_Oj1h`0YNT_Oh1l`0ZNQ_Og1n`0g0O00001N10000O2O000O2O000O2O000O2O000O2O000O101O000O2O000O2O000O2O000O2O000O2O000O101O0O110O000010O0001O01O01O01O0001O01O01O00010O00010O000010O0000010O00010O0001O01O01O00010O0001O01O00010O00010O00010O00010O0001O01O01O01O01O01O01O01O01O01O01O01O01O00010O00010O00010O00010O00010O000010O00010O00010O00010O02N2O0O2N1O2O1N1OPB"}, "label": "breakfast sandwich with egg , meat and toasted bread"}]}
{"id": 226102, "image": "COCO_train2014_000000226102.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the portion of the sandwhich nearest the utensil\"."}], "task": "refering", "answer_template": "The \"the portion of the sandwhich nearest the utensil\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 316, 317, 318, 319, 320, 321, 322, 336, 337, 338, 339, 356], "bbox": [0.5988333333333333, 0.6286562499999999, 1.0, 0.8790312499999999], "iscrowd": 0, "area": 20915.840000000007, "rle": {"size": [640, 480], "counts": "Zkc51mc04K5L4K5L4K5L3L4M2M3O2N1N2O2N1O1N3N1O1O2M2O1O1O2ROSNe_Oo1[`0SNb_On1]`0VN__Ol1``0VN]_Ok1c`0WNY_Ok1f`0XNW_Oj1h`0YNT_Oh1l`0ZNQ_Og1n`0g0O00001N10000O2O000O2O000O2O000O2O000O2O000O101O000O2O000O2O000O2O000O2O000O2O000O101O0O110O000010O0001O01O01O01O0001O01O01O00010O00010O000010O0000010O00010O0001O01O01O00010O0001O01O00010O00010O00010O00010O0001O01O01O01O01O01O01O01O01O01O01O01O01O00010O00010O00010O00010O00010O000010O00010O00010O00010O02N2O0O2N1O2O1N1OPB"}, "label": "the portion of the sandwhich nearest the utensil"}]}
{"id": 406328, "image": "COCO_train2014_000000406328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on right\"."}], "task": "refering", "answer_template": "The \"elephant on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 272, 273, 274, 291, 297], "bbox": [0.641734375, 0.34517647058823525, 0.9901406249999999, 0.8182823529411765], "iscrowd": 0, "area": 35312.10725000001, "rle": {"size": [425, 640], "counts": "\\iZ5b3]9b0^O;F9M4L4M3M2O2O1O1O1N101OO10O010000O10O10O100O10O10O10000O0ROQHdLo7\\3THaLl7_3XH\\Li7d3[HXLe7h3_HSLa7n3bHoK^7Q4fHkKZ7U4jHfKW7Z4l00O10O100000O1000000O100000000O100000000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000000a0_O001O00001O00001O001O00001O00001O000000000000O10001N1000000O100O100O1O1O1O1O1O1O1O12N2N2N2N2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O000000000000O1M3M3M3K5J6J6J6J6J6K6K4N3N2M2N3L4K5L3M4K5L3L5L4L4K4M4L4K4M4K5L4L3_Ob0YOg0YOV[2"}, "label": "elephant on right"}]}
{"id": 406328, "image": "COCO_train2014_000000406328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant facing directly to the right\"."}], "task": "refering", "answer_template": "The \"the elephant facing directly to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 272, 273, 274, 291, 297], "bbox": [0.641734375, 0.34517647058823525, 0.9901406249999999, 0.8182823529411765], "iscrowd": 0, "area": 35312.10725000001, "rle": {"size": [425, 640], "counts": "\\iZ5b3]9b0^O;F9M4L4M3M2O2O1O1O1N101OO10O010000O10O10O100O10O10O10000O0ROQHdLo7\\3THaLl7_3XH\\Li7d3[HXLe7h3_HSLa7n3bHoK^7Q4fHkKZ7U4jHfKW7Z4l00O10O100000O1000000O100000000O100000000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000000a0_O001O00001O00001O001O00001O00001O000000000000O10001N1000000O100O100O1O1O1O1O1O1O1O12N2N2N2N2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O000000000000O1M3M3M3K5J6J6J6J6J6K6K4N3N2M2N3L4K5L3M4K5L3L5L4L4K4M4L4K4M4K5L4L3_Ob0YOg0YOV[2"}, "label": "the elephant facing directly to the right"}]}
{"id": 406328, "image": "COCO_train2014_000000406328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on left\"."}], "task": "refering", "answer_template": "The \"elephant on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 244, 261, 262, 263, 264, 265, 267, 284, 285, 286, 287, 288], "bbox": [0.341, 0.32995294117647056, 0.65209375, 0.8299529411764707], "iscrowd": 0, "area": 29066.039250000005, "rle": {"size": [425, 640], "counts": "bej22W=5J6J6K5J7I5L4K5L3L5K5L7H9hFnMh6[2gHPNV7Y2ZHPNc7Y2mGQNQ8e3N1O0O2O2N1O2M2O2O0O2O0O10O0100O10O0100O00100O10O010001N100O100O01N101N1O2]Ob0C>J6L3N3M2N3L32O1N3N1N2O1N2O0O2O1YHRLP6Q4bI\\L]6h3XI_Lg6e3fHkLZ7c4N101N100O2O0O2O0O10O10001O0O10001N1[OlHUKT7g4UITKk6b3nHRM?XOd6e3gJULZ5j3S20001N10001N10000O20N1_GZL5OT6g3_IjL0E`6a3XI[MJZOl6\\3PI\\No6P4O000O2O001O0O2O001NN3M3L4M2M4L4M2M4L4M2N3L4M2N3M3M2N3L4M2N3M3M2G:C=K4O2N23L3N3M2N100O1O1O2N1O1O1O1O01N1N2O1N2M3L4M43M2[HfLU5^3^JkLc5X3\\IeMe6_41O1O1O00100O1O1O001O1O1O10O1000bJRIe3AhL`8k2iGSM^8`2kG]M[8V2oGbMZ8S2oGfMY8`1W2D<C=DgZl2"}, "label": "elephant on left"}]}
{"id": 406328, "image": "COCO_train2014_000000406328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant whose trunk is fully exposed\"."}], "task": "refering", "answer_template": "The \"elephant whose trunk is fully exposed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 244, 261, 262, 263, 264, 265, 267, 284, 285, 286, 287, 288], "bbox": [0.341, 0.32995294117647056, 0.65209375, 0.8299529411764707], "iscrowd": 0, "area": 29066.039250000005, "rle": {"size": [425, 640], "counts": "bej22W=5J6J6K5J7I5L4K5L3L5K5L7H9hFnMh6[2gHPNV7Y2ZHPNc7Y2mGQNQ8e3N1O0O2O2N1O2M2O2O0O2O0O10O0100O10O0100O00100O10O010001N100O100O01N101N1O2]Ob0C>J6L3N3M2N3L32O1N3N1N2O1N2O0O2O1YHRLP6Q4bI\\L]6h3XI_Lg6e3fHkLZ7c4N101N100O2O0O2O0O10O10001O0O10001N1[OlHUKT7g4UITKk6b3nHRM?XOd6e3gJULZ5j3S20001N10001N10000O20N1_GZL5OT6g3_IjL0E`6a3XI[MJZOl6\\3PI\\No6P4O000O2O001O0O2O001NN3M3L4M2M4L4M2M4L4M2N3L4M2N3M3M2N3L4M2N3M3M2G:C=K4O2N23L3N3M2N100O1O1O2N1O1O1O1O01N1N2O1N2M3L4M43M2[HfLU5^3^JkLc5X3\\IeMe6_41O1O1O00100O1O1O001O1O1O10O1000bJRIe3AhL`8k2iGSM^8`2kG]M[8V2oGbMZ8S2oGfMY8`1W2D<C=DgZl2"}, "label": "elephant whose trunk is fully exposed"}]}
{"id": 258878, "image": "COCO_train2014_000000258878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red train\"."}], "task": "refering", "answer_template": "The \"red train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285], "bbox": [0.1214375, 0.2516861826697892, 0.73015625, 0.8179859484777517], "iscrowd": 0, "area": 74165.26185000002, "rle": {"size": [427, 640], "counts": "VjP18o<;E;E;E6I5L5K5K4L5K5K4K6K5K4L5K5K4K6K5K4L5K5K4L5J6K4L4L4L4L4K6K4L4L4L4L4L4K5L4L4L4L4M300O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O10000000001O000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000001O0000001O0000001O00001O00000010O000001O0000001O00001O0000001O0000001O00001O0000001O0O1N2E<\\Oc0G901O0000010O0000001O000000000001O000001O00000000000000000000001O00000O1000000000000000001O0000000000000000000000001O0000000O10000000000000001O0000000000000000000000001O00000000000O10000000001O0000000000000000000000001O000000000000000O100000001O0000000000000000000000001O0000000O10000000000000001O000000000000000000000O101O00000000000000000000001O00000000000O1000001O00001O001O001O00001O001O008H8H8G:XOmFfL[9k2P1@k0TOl0UOZmW2"}, "label": "red train"}]}
{"id": 258878, "image": "COCO_train2014_000000258878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red train car on tracks\"."}], "task": "refering", "answer_template": "The \"a red train car on tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285], "bbox": [0.1214375, 0.2516861826697892, 0.73015625, 0.8179859484777517], "iscrowd": 0, "area": 74165.26185000002, "rle": {"size": [427, 640], "counts": "VjP18o<;E;E;E6I5L5K5K4L5K5K4K6K5K4L5K5K4K6K5K4L5K5K4L5J6K4L4L4L4L4K6K4L4L4L4L4L4K5L4L4L4L4M300O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O10000000001O000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000001O0000001O0000001O00001O00000010O000001O0000001O00001O0000001O0000001O00001O0000001O0O1N2E<\\Oc0G901O0000010O0000001O000000000001O000001O00000000000000000000001O00000O1000000000000000001O0000000000000000000000001O0000000O10000000000000001O0000000000000000000000001O00000000000O10000000001O0000000000000000000000001O000000000000000O100000001O0000000000000000000000001O0000000O10000000000000001O000000000000000000000O101O00000000000000000000001O00000000000O1000001O00001O001O001O00001O001O008H8H8G:XOmFfL[9k2P1@k0TOl0UOZmW2"}, "label": "a red train car on tracks"}]}
{"id": 447297, "image": "COCO_train2014_000000447297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vespa that is parked up on the sidewalk on this side of the rainy street\"."}], "task": "refering", "answer_template": "The \"the vespa that is parked up on the sidewalk on this side of the rainy street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139], "bbox": [0.005046875, 0.025905292479108638, 0.18571875, 0.5146518105849582], "iscrowd": 0, "area": 9824.22945, "rle": {"size": [359, 640], "counts": "TV1Z1h7T22O1O1N2O1N2O1O1N2O1O1N2O1N2N2N2O2M2N2O1N2N2N2O1N2O100OgNlM`Ig1iNcNe7GcIe1iNcNc7IeIc1hNdNb7JgIa1hNeN_7LiI^1iNfN\\7MlI[1iNhN[7MmIZ1iNiNX7NPJX1iNjNV7NRJW1hNlNT7NTJV1iNkNR70VJT1iNlNo62XJP1jNoNm61ZJo0jNPOj62]Jm0jNQOg63`Jk0iNROf64bJi0iNSOc65eJf0jNUO`65gJe0iNUO`68gJb0jNUO_69gJb0jNTO`6:gJa0jNSO_6=hJ?jNRO_6a0fJ;lNTO]6e0eJ6oNSO]6i0cJ3QORO]6o0_JNV68aIL_68nH6R7O[H`0e7a100O10O010000O1000O010000O10O10O10000O0100O010O10O01O0aL]H\\3c7cL^Hd02n1e7lMbHS2o6hMjH3;R2d6RNRIK<R2a6RNUIK=P2_6TNUIK?o1[6VNWIJ`0o1X6WNYIIb0P2S6VN\\IIc0Q2o5VN_IHe0Q2j5VNcIHf0Q2h5TNdIIf0S2k5mM`IOh0S2j7O001O001O1O001M4L6J6J5K6J6J6J5AaENo:L]`f5"}, "label": "the vespa that is parked up on the sidewalk on this side of the rainy street"}]}
{"id": 447297, "image": "COCO_train2014_000000447297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"scooter sitting unattended on the street\"."}], "task": "refering", "answer_template": "The \"scooter sitting unattended on the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139], "bbox": [0.005046875, 0.025905292479108638, 0.18571875, 0.5146518105849582], "iscrowd": 0, "area": 9824.22945, "rle": {"size": [359, 640], "counts": "TV1Z1h7T22O1O1N2O1N2O1O1N2O1O1N2O1N2N2N2O2M2N2O1N2N2N2O1N2O100OgNlM`Ig1iNcNe7GcIe1iNcNc7IeIc1hNdNb7JgIa1hNeN_7LiI^1iNfN\\7MlI[1iNhN[7MmIZ1iNiNX7NPJX1iNjNV7NRJW1hNlNT7NTJV1iNkNR70VJT1iNlNo62XJP1jNoNm61ZJo0jNPOj62]Jm0jNQOg63`Jk0iNROf64bJi0iNSOc65eJf0jNUO`65gJe0iNUO`68gJb0jNUO_69gJb0jNTO`6:gJa0jNSO_6=hJ?jNRO_6a0fJ;lNTO]6e0eJ6oNSO]6i0cJ3QORO]6o0_JNV68aIL_68nH6R7O[H`0e7a100O10O010000O1000O010000O10O10O10000O0100O010O10O01O0aL]H\\3c7cL^Hd02n1e7lMbHS2o6hMjH3;R2d6RNRIK<R2a6RNUIK=P2_6TNUIK?o1[6VNWIJ`0o1X6WNYIIb0P2S6VN\\IIc0Q2o5VN_IHe0Q2j5VNcIHf0Q2h5TNdIIf0S2k5mM`IOh0S2j7O001O001O1O001M4L6J6J5K6J6J6J5AaENo:L]`f5"}, "label": "scooter sitting unattended on the street"}]}
{"id": 37698, "image": "COCO_train2014_000000037698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a grey shirt\"."}], "task": "refering", "answer_template": "The \"a man in a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.48209375000000004, 0.21519906323185012, 0.8192812500000001, 0.9872365339578455], "iscrowd": 0, "area": 44547.2007, "rle": {"size": [427, 640], "counts": "[TQ4\\1j;d0]O4L5J6K5K4K6I7I6I7I7J6N1O2M3N2N1O2N2O1O1O0O2O1O1O001O1O1N110O100[H`L^5`3UJnLk5R3jIYMU6o400O100O100O100O1O100O100O100O100O10000O1eNjHaLV7Z3_IVLa6c3gI\\LY6[3QJcLP6S3[JlLe5o2`JQM`5n2bJQM^5o2cJoL^5P3cJPM]5P3dJoL\\5P3fJnL[5Q3gJiL^5W3bJeLb5Z3X2O100O1O100O100O1O100O1O100O1000000O100000000O1000000O10000001O001O001O001O001O001O1O001O001O001O001O001O001O2N2N1O2_H\\LW5f3aJ`L_5b3XJeLh5h3aIcL^6o3gHZLY7l401O0000001O00001O0000001O0000001O00001O000\\KPJm1P6UNVJc1k5]N\\J[1d5gNcJP1_5POgJh0[5WOmJ`0U5@TK5n4K\\KIf46dK_O^4a0gKXO[4h0fKUO\\4j0fKSO\\4m0eKPO]4P1dKmN_4Q1bKmN`4S1aKjNa4V1`KgNb4X1`KeNb4[1^KcNd4]1[KbNf4_1ZK_Nh4a1WK^Nk4b1TK]Nm4d1SKZNo4f1PKYNR5g1nJWNS5j1lJUNV5k1iJTNX5m1hJRNY5m1gJRN[5m1eJRN\\5m1fJQNZ5o1gJPNZ5o1gJPNZ5P2gJnMZ5Q2gJnMZ5Q2gJnMZ5Q2hJmMY5S2gJlMZ5S2gJlMZ5S2hJkMY5T2hJkMY5U2gJjMZ5U2hJiMY5V2hJiMY5W2o2N1O2N101N1O2N1O2L3N5K5J7E:E<DmR`1"}, "label": "a man in a grey shirt"}]}
{"id": 37698, "image": "COCO_train2014_000000037698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in grey shirt playing video game\"."}], "task": "refering", "answer_template": "The \"man in grey shirt playing video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.48209375000000004, 0.21519906323185012, 0.8192812500000001, 0.9872365339578455], "iscrowd": 0, "area": 44547.2007, "rle": {"size": [427, 640], "counts": "[TQ4\\1j;d0]O4L5J6K5K4K6I7I6I7I7J6N1O2M3N2N1O2N2O1O1O0O2O1O1O001O1O1N110O100[H`L^5`3UJnLk5R3jIYMU6o400O100O100O100O1O100O100O100O100O10000O1eNjHaLV7Z3_IVLa6c3gI\\LY6[3QJcLP6S3[JlLe5o2`JQM`5n2bJQM^5o2cJoL^5P3cJPM]5P3dJoL\\5P3fJnL[5Q3gJiL^5W3bJeLb5Z3X2O100O1O100O100O1O100O1O100O1000000O100000000O1000000O10000001O001O001O001O001O001O1O001O001O001O001O001O001O2N2N1O2_H\\LW5f3aJ`L_5b3XJeLh5h3aIcL^6o3gHZLY7l401O0000001O00001O0000001O0000001O00001O000\\KPJm1P6UNVJc1k5]N\\J[1d5gNcJP1_5POgJh0[5WOmJ`0U5@TK5n4K\\KIf46dK_O^4a0gKXO[4h0fKUO\\4j0fKSO\\4m0eKPO]4P1dKmN_4Q1bKmN`4S1aKjNa4V1`KgNb4X1`KeNb4[1^KcNd4]1[KbNf4_1ZK_Nh4a1WK^Nk4b1TK]Nm4d1SKZNo4f1PKYNR5g1nJWNS5j1lJUNV5k1iJTNX5m1hJRNY5m1gJRN[5m1eJRN\\5m1fJQNZ5o1gJPNZ5o1gJPNZ5P2gJnMZ5Q2gJnMZ5Q2gJnMZ5Q2hJmMY5S2gJlMZ5S2gJlMZ5S2hJkMY5T2hJkMY5U2gJjMZ5U2hJiMY5V2hJiMY5W2o2N1O2N101N1O2N1O2L3N5K5J7E:E<DmR`1"}, "label": "man in grey shirt playing video game"}]}
{"id": 37698, "image": "COCO_train2014_000000037698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right arm of the man in brown\"."}], "task": "refering", "answer_template": "The \"the right arm of the man in brown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.081140625, 0.042786885245901636, 0.47184375, 0.7702810304449648], "iscrowd": 0, "area": 50031.4089, "rle": {"size": [427, 640], "counts": "QPf02W=2M3M3M3M3M3M3M3N2M3PNSO]GR1a8]ObFS1\\9QOaFP1]9SO`Fo0_9TO^Fm0`9VO^Fk0`9YO\\Fi0b9ZO\\Fg0c9[O[Ff0c9^OZFc0f9^OWFd0h9_OUFb0j9`100O100O2O0O1O100O100O1O100O100O1O100O1O1O100O1O1O1N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2O1N2N2N2O1L4L4M3L4L4L4L4M3O1N2O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1O001O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O001O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1000000O100000000000000O1000O10000000O100000000000000O101OQL\\LWNd3h1`LUN`3k1cLRN]3m1gLPNY3P2kLlMU3S2oLjMP3W2SMfMm2Y2WMdMi2\\2YMcMf2\\2]MbMc2^2^MaMb2^2aM`M_2`2cM^M\\2b2gM\\MY2d2iMZMW2e2lMYMU2f2lMYMT2f2oMYMQ2f2QNXMP2f2SNXMm1h2UNVMl1h2WNVMi1j2YNTMh1j2[NTMf1k2\\NSMd1l2`NQMa1n2aNQM_1m2dNQM\\1o2fNoL[1o2lNkLU1T3YO^Lg0a3HQL9n30iK0V44gKMX45fKLX48fKGZ4;dKFZ4>cKC\\4`0aK@^4c0`K^O_4g4O2O1J5K6POhFfM]9X2mF[MY9b2i0N3M3M2N3M2iN\\EEg:3^1G:G:Ejl\\4"}, "label": "the right arm of the man in brown"}]}
{"id": 37698, "image": "COCO_train2014_000000037698.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in the left side with white color speaker\"."}], "task": "refering", "answer_template": "The \"man in the left side with white color speaker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.081140625, 0.042786885245901636, 0.47184375, 0.7702810304449648], "iscrowd": 0, "area": 50031.4089, "rle": {"size": [427, 640], "counts": "QPf02W=2M3M3M3M3M3M3M3N2M3PNSO]GR1a8]ObFS1\\9QOaFP1]9SO`Fo0_9TO^Fm0`9VO^Fk0`9YO\\Fi0b9ZO\\Fg0c9[O[Ff0c9^OZFc0f9^OWFd0h9_OUFb0j9`100O100O2O0O1O100O100O1O100O100O1O100O1O1O100O1O1O1N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2O1N2N2N2O1L4L4M3L4L4L4L4M3O1N2O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1O001O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O001O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1000000O100000000000000O1000O10000000O100000000000000O101OQL\\LWNd3h1`LUN`3k1cLRN]3m1gLPNY3P2kLlMU3S2oLjMP3W2SMfMm2Y2WMdMi2\\2YMcMf2\\2]MbMc2^2^MaMb2^2aM`M_2`2cM^M\\2b2gM\\MY2d2iMZMW2e2lMYMU2f2lMYMT2f2oMYMQ2f2QNXMP2f2SNXMm1h2UNVMl1h2WNVMi1j2YNTMh1j2[NTMf1k2\\NSMd1l2`NQMa1n2aNQM_1m2dNQM\\1o2fNoL[1o2lNkLU1T3YO^Lg0a3HQL9n30iK0V44gKMX45fKLX48fKGZ4;dKFZ4>cKC\\4`0aK@^4c0`K^O_4g4O2O1J5K6POhFfM]9X2mF[MY9b2i0N3M3M2N3M2iN\\EEg:3^1G:G:Ejl\\4"}, "label": "man in the left side with white color speaker"}]}
{"id": 54083, "image": "COCO_train2014_000000054083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant on a filing cabinet\"."}], "task": "refering", "answer_template": "The \"a plant on a filing cabinet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 22, 23, 24, 37, 38, 51, 52, 65, 66], "bbox": [0.5784754521963824, 0.002248062015503876, 0.7892248062015504, 0.31839793281653744], "iscrowd": 0, "area": 5654.2611499999975, "rle": {"size": [387, 387], "counts": "ced28h;6K6K0000010O1O2O0O2O1O1N2O1O001N101O0O2O001O1N101O1N101N2O1O001N2O001O002N2N1TG[NX7g1cH]N\\7e1`H^N_7d1\\H`Nc7a1ZHbNe7`1WHcNh7_1SHeNl7]1PHfNo7\\1lGgNT8Z1gGkNX8V1eGmNZ8_2O1O1O1O100O1O1O0000000000000000000001O00O10000000000000QNo1eN[1Hlon0"}, "label": "a plant on a filing cabinet"}]}
{"id": 54083, "image": "COCO_train2014_000000054083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"potted plant on shelf in distance\"."}], "task": "refering", "answer_template": "The \"potted plant on shelf in distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 22, 23, 24, 37, 38, 51, 52, 65, 66], "bbox": [0.5784754521963824, 0.002248062015503876, 0.7892248062015504, 0.31839793281653744], "iscrowd": 0, "area": 5654.2611499999975, "rle": {"size": [387, 387], "counts": "ced28h;6K6K0000010O1O2O0O2O1O1N2O1O001N101O0O2O001O1N101O1N101N2O1O001N2O001O002N2N1TG[NX7g1cH]N\\7e1`H^N_7d1\\H`Nc7a1ZHbNe7`1WHcNh7_1SHeNl7]1PHfNo7\\1lGgNT8Z1gGkNX8V1eGmNZ8_2O1O1O1O100O1O1O0000000000000000000001O00O10000000000000QNo1eN[1Hlon0"}, "label": "potted plant on shelf in distance"}]}
{"id": 496457, "image": "COCO_train2014_000000496457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch to the right of the snack table\"."}], "task": "refering", "answer_template": "The \"the couch to the right of the snack table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 135, 136, 137, 138, 139, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193, 194, 195, 210, 211, 212, 213, 214, 228, 229, 230, 231, 232], "bbox": [0.54298, 0.46984, 0.9526800000000001, 0.99936], "iscrowd": 0, "area": 18123.03485, "rle": {"size": [375, 500], "counts": "n]S32b;8J7aE@[9U1lETOo9Z1O10000O10000O10O0VOTFJm91\\FJd95cFE^99iF^OZ9b0n00O0100O01000000O2O001O1O1N101kE\\OU9c1000000O2O00001O1O0O2O1O1O1O1O1N101O001O00001O001O001O001O1O001O1O1O001O1O001O1O1O001O1\\HTO[4U4O000000000000cK[Jh2e5UM_Jj2a5SMbJm2^5QMeJn2[5oLhJQ3X5nLjJQ3V5mLlJS3T5lLnJS3R5kLPKU3P5jLRKU3m4jLUKV3k4iLVKX3i4gLYKX3g4gLZKY3e4eL^K[3b4cLaK\\3_4aLdK_3\\4^LgKb3Y4\\LjKc3U4[LnKe3R4XLQLh3o3VLTLi3l3TLWLl3h3SLZLm3h501O000001O00000000000001O00000000000000000001O01O00000000000000000000001O00001O001O0O2O001O001O001O005K5K4L5K5K5K>B4L2N2N3M2N3M2N2N3M2N3M2N3M3L3N3M3M3M2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1Oe]8"}, "label": "the couch to the right of the snack table"}]}
{"id": 349007, "image": "COCO_train2014_000000349007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the triangle of sandwich in the front\"."}], "task": "refering", "answer_template": "The \"the triangle of sandwich in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 253, 268, 269, 270, 271, 272, 273, 274, 275, 276, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 402, 403, 404, 405, 406, 425], "bbox": [0.2, 0.4563235294117647, 0.7101143790849673, 0.8743137254901959], "iscrowd": 0, "area": 50298.829150000005, "rle": {"size": [612, 612], "counts": "WkY28hb06I8I6J7H7J7H7J7I6I7J7H7J7I6I8K4N3M2N3M2O2M2N2N3M2N3M2N3M2N3M2N3M2N3N1N2N3M2N3M2N3M2N3M2N3M2N3M2N3N1N2N3M2N3M2N3M2N3N1N3M2N3N1N2N3M2O2M2N3N1N3M2N3NO0O10001N10001O0O10001N10000O2O00000O2O000O2O000O101O000O101O0O101O0O10001N1000001N10001N10000O2O00000O2O000O101O0O101O0O10001O0O10001N10001N1000001N10000O2O000O101O000O2O000O101O0O10001N10001O0O10001N10000O2O001O0O2O1O1N101O1N101O1N2O001O1N2O001N2O001N2O1O001N2O1O0O2O1O0O2O1O1O0O2O1O1N101O1N2O001N2O001O1N2O001N2O1O0O2O1O001N2O1O0O2O1O1N101O1N2O001O1N101O1N2O001N2O1O001N2O001N2O1O0O2O1O1O0O2O1O1N101O1N101O1N2O001O1N2O001N2O001N2O1O001N2O1O0O2O1O0O2O1O1N101O1O1N101O1A?]Ob0]Ol\\Y3"}, "label": "the triangle of sandwich in the front"}]}
{"id": 349007, "image": "COCO_train2014_000000349007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"closest sandwich wedge on right\"."}], "task": "refering", "answer_template": "The \"closest sandwich wedge on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 368, 369], "bbox": [0.6485294117647058, 0.3242647058823529, 0.9932026143790849, 0.7573692810457516], "iscrowd": 0, "area": 35948.80819999999, "rle": {"size": [612, 612], "counts": "a`]74nb09H7H8H9G8I8G8H9G8I8G8H8H9H7H9G8H9H7H9G8H8I7H4L4L4M3L4L4L4M3L4N2N2N2O1N2N2O1N2O1N3M2O1N2nDaH[:a7d01lD^Hb:d7\\E\\He:e7XE[Hi:g7UEYHk:i7REXHo:i7oDVHR;n7001N10001N101O0O101O0O101O0O2O000O2O000O2O000O2O001N10001N10001N101O0O101O0O101O0O101O0O2O000O2O000O2O001N10001N10001N2O001N2O001N2O1O0O2O1O0O2O1O0O2O1O1N101O1N101O1N2O001N2O001N2O001N2O1O0O2O1O0O2O1O001N2O1O0O2O1O0O2O1O1N101O1RNjAkNW>T1kAkNV>R1mAmNT>Q1nAmNS>Q1QBmNP>Q1RBnNo=o0TBoNn=o0TBPOl=n0XBPOi=n0YBPOi=m0ZBROf=m0\\BROe=k0_BROc=l0_BSOa=k0bBTO_=j0cBTO_=i0eBUO[=j0gBUOZ=h0iBVOY=h0iBWOW=g0mBWOT=g0nBWOT=f0oBVOT=i0nBSOV=j0mBSOV=k0mBPOV=n0mBoNV=o0lBmNX=P1`2N1N3N2M2O2M3N1N3N2M2O2M3N1N3N]f2"}, "label": "closest sandwich wedge on right"}]}
{"id": 349007, "image": "COCO_train2014_000000349007.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich on the top left\"."}], "task": "refering", "answer_template": "The \"the sandwich on the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 133, 134, 135, 136, 137, 138, 139, 155, 156, 157, 158, 159, 160, 161, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 266, 267, 268, 288], "bbox": [0.044950980392156864, 0.25857843137254904, 0.3685457516339869, 0.6046405228758169], "iscrowd": 0, "area": 26421.763900000005, "rle": {"size": [612, 612], "counts": "bm`05ob05J6K5K5K5J6K6J5K5K5J6K5K5K6J5J6K5K5K5J6K6J5K5K5J6K5K5K6I6K5K5K5K5J3NO10000O01000O10O010000O10O10O1000O0100O1000O010000O10O10O1000O0100O1000O010000O01000O100O01000O10O10O10000O010O10000O01000O10O10N2O1O1O0O2O1O1O0O2O1O1O1O0O2O1O1O1N101O1O1N101O1O1N2O001O1N2O001O1O1N2O001O1N2O1O001N2O1O001N2O1O1O0O2O1O1O001N2O1O1O1N2O001O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1M3J6J6I6K6J6J6I7J6J6I7H8IaaV7"}, "label": "the sandwich on the top left"}]}
{"id": 299859, "image": "COCO_train2014_000000299859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pc monitor on a table\"."}], "task": "refering", "answer_template": "The \"a pc monitor on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 187, 199, 200, 201, 202, 203], "bbox": [0.0, 0.42345821325648414, 0.45568000000000003, 1.0], "iscrowd": 0, "area": 32353.80280000001, "rle": {"size": [347, 500], "counts": "X7a0Z:3L5L3L4M4K4M4K4M3L5L3L4M4K4M4K4M3L5L1N1O2OO0100O010O101N101N2O0O2O1N101N2O0O2O1N101N2O0O2O0O2O1N101N101N1O2O1N101N101N2O0O2O0O2O0O2O0O2O0O2O0O2H7K6I6L5M2N3L3N4M2M6K1N2O1O1O1O1O100O100O100O100O10O10O01O100O1O010O1O100O1O101N1O100O1O100O100010O1O001O001O001O010O1O001O00O1O101N1O1O1O100O1O2N100O1O1O1O101N1O1O100O1O2N1O100O1O1O101N1O1O1O100O2N1O100O2N1O1O101N1O1O100O2N1O1O101N1000001O0000000010O0000O2O0000001O00000O2O000O10001N1O100O2N100O=Ca0]Oc0_Ob0^Oa0_Ob0^Oa0^Omlk2"}, "label": "a pc monitor on a table"}]}
{"id": 299859, "image": "COCO_train2014_000000299859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large computer moniter sitting on a table by the bed\"."}], "task": "refering", "answer_template": "The \"a large computer moniter sitting on a table by the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 187, 199, 200, 201, 202, 203], "bbox": [0.0, 0.42345821325648414, 0.45568000000000003, 1.0], "iscrowd": 0, "area": 32353.80280000001, "rle": {"size": [347, 500], "counts": "X7a0Z:3L5L3L4M4K4M4K4M3L5L3L4M4K4M4K4M3L5L1N1O2OO0100O010O101N101N2O0O2O1N101N2O0O2O1N101N2O0O2O0O2O1N101N101N1O2O1N101N101N2O0O2O0O2O0O2O0O2O0O2O0O2H7K6I6L5M2N3L3N4M2M6K1N2O1O1O1O1O100O100O100O100O10O10O01O100O1O010O1O100O1O101N1O100O1O100O100010O1O001O001O001O010O1O001O00O1O101N1O1O1O100O1O2N100O1O1O1O101N1O1O100O1O2N1O100O1O1O101N1O1O1O100O2N1O100O2N1O1O101N1O1O100O2N1O1O101N1000001O0000000010O0000O2O0000001O00000O2O000O10001N1O100O2N100O=Ca0]Oc0_Ob0^Oa0_Ob0^Oa0^Omlk2"}, "label": "a large computer moniter sitting on a table by the bed"}]}
{"id": 299859, "image": "COCO_train2014_000000299859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor sitting at the end of the bed\"."}], "task": "refering", "answer_template": "The \"a computer monitor sitting at the end of the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179], "bbox": [0.73264, 0.47265129682997115, 1.0, 0.784956772334294], "iscrowd": 0, "area": 10441.978299999999, "rle": {"size": [347, 500], "counts": "SXl31h:4L3M4K5I6J7I7J5J7I7I6K6I7J5K6J6I7N1O2N1O10O000001O0000010O00000001O0O2O0O1O2O001N10001O000O2O001O00001O001O00010O000010O01O00010O000010O01O00010O001O00010O00001O001O01O01O00001O001O00001O001O00001O00001O00001O00001O00001O00001O00001L3M4M3L3M4L4L4M3L4L4L4L4M3L4L4L4L_I"}, "label": "a computer monitor sitting at the end of the bed"}]}
{"id": 299859, "image": "COCO_train2014_000000299859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a old computer monitor sitting on a blue bedspread\"."}], "task": "refering", "answer_template": "The \"a old computer monitor sitting on a blue bedspread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179], "bbox": [0.73264, 0.47265129682997115, 1.0, 0.784956772334294], "iscrowd": 0, "area": 10441.978299999999, "rle": {"size": [347, 500], "counts": "SXl31h:4L3M4K5I6J7I7J5J7I7I6K6I7J5K6J6I7N1O2N1O10O000001O0000010O00000001O0O2O0O1O2O001N10001O000O2O001O00001O001O00010O000010O01O00010O000010O01O00010O001O00010O00001O001O01O01O00001O001O00001O001O00001O00001O00001O00001O00001O00001O00001L3M4M3L3M4L4L4M3L4L4L4L4M3L4L4L4L_I"}, "label": "a old computer monitor sitting on a blue bedspread"}]}
{"id": 398164, "image": "COCO_train2014_000000398164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a blue headband holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a woman with a blue headband holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.359875, 0.07160377358490566, 0.7279375, 0.9876650943396227], "iscrowd": 0, "area": 42894.00200000001, "rle": {"size": [424, 640], "counts": "bPP35n<5L5K4L5K4L5K5K4K6K4L5K4L5K4L5K5J5M4L3Mk0UO3M3M3M3M3N2M3M3M3M3M3M3M3N2M3N2N2M3N2M3N2N2M3N2N3L3N2M3N2N2M3N2M3N2I7G9G9G9F:G9K5M3N2N2N2N2N2M3N2N2N2N3M2N2UNQG[OQ9c0ZGSOg8k0dGjN^8S1nGbNT8\\1XHXNj7f1j10N3M3L4O14M3M3M3M3L4M3M3M3M3M3L21O000000000000001O000000000000001O0000000000001O000000000000001O0000000bMkNiHU1S7TOhHl0U7[OgHe0U7DfH=U7LfH5U74fHMU1gNW3d1_KFX1hNT3k1_K_OY1iNS3P2`KXOZ1kNQ3V2`KPO]1lNo2\\2_KiN_1mNn2d2]K`Nb1oNl2k2\\KXNe1oNk2R3ZKRNf1oNk2Y3YKlMf1nNl2\\5oLfJl2b5nL`Jo2i5jLYJQ3T6dLoIW3^6^LeI]3g6ZL[Ic3V801N2O1O1O001O1O1O1N109G=C<D=^NoD3];ZOjDf0k;10O0100O100O10000O010O010O01000O010O010O10O010O10O10O100O10001N100O10000O2O0O100O10001N100O10000O2O0O10000O101N10000O100O2O0O109GWjW2"}, "label": "a woman with a blue headband holding a tennis racket"}]}
{"id": 398164, "image": "COCO_train2014_000000398164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player who is wearing a blue band on her head\"."}], "task": "refering", "answer_template": "The \"a player who is wearing a blue band on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.359875, 0.07160377358490566, 0.7279375, 0.9876650943396227], "iscrowd": 0, "area": 42894.00200000001, "rle": {"size": [424, 640], "counts": "bPP35n<5L5K4L5K4L5K5K4K6K4L5K4L5K4L5K5J5M4L3Mk0UO3M3M3M3M3N2M3M3M3M3M3M3M3N2M3N2N2M3N2M3N2N2M3N2N3L3N2M3N2N2M3N2M3N2I7G9G9G9F:G9K5M3N2N2N2N2N2M3N2N2N2N3M2N2UNQG[OQ9c0ZGSOg8k0dGjN^8S1nGbNT8\\1XHXNj7f1j10N3M3L4O14M3M3M3M3L4M3M3M3M3M3L21O000000000000001O000000000000001O0000000000001O000000000000001O0000000bMkNiHU1S7TOhHl0U7[OgHe0U7DfH=U7LfH5U74fHMU1gNW3d1_KFX1hNT3k1_K_OY1iNS3P2`KXOZ1kNQ3V2`KPO]1lNo2\\2_KiN_1mNn2d2]K`Nb1oNl2k2\\KXNe1oNk2R3ZKRNf1oNk2Y3YKlMf1nNl2\\5oLfJl2b5nL`Jo2i5jLYJQ3T6dLoIW3^6^LeI]3g6ZL[Ic3V801N2O1O1O001O1O1O1N109G=C<D=^NoD3];ZOjDf0k;10O0100O100O10000O010O010O01000O010O010O10O010O10O10O100O10001N100O10000O2O0O100O10001N100O10000O2O0O10000O101N10000O100O2O0O109GWjW2"}, "label": "a player who is wearing a blue band on her head"}]}
{"id": 398164, "image": "COCO_train2014_000000398164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis racket held by a woman playing tennis\"."}], "task": "refering", "answer_template": "The \"tennis racket held by a woman playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205], "bbox": [0.701546875, 0.3447877358490566, 0.953921875, 0.5687971698113208], "iscrowd": 0, "area": 9780.3089, "rle": {"size": [424, 640], "counts": "YTj52U=4M3L101N100O100O100O100O100O100O101N100O1O1O1O100O1O1O1O100O1O1O2N100O1O1O100O1O1O1O100O1O1O2N1N2N2O1N2O1N2N2O1N2N2O1O1O1O2N1O100O1O1O1O1O1O1O100O1O1O2N1O1O1O100O100O100000000O10001O0O100000000O1000000O1000001O0O1000000O100000000O10O2O0O10000O2O0O100O2O0O10001N100O101N10000O101N100O3N1O1N3N1N3N1N3L3M3M4L3M4L3M3M4L3M4Ld0\\OVj;"}, "label": "tennis racket held by a woman playing tennis"}]}
{"id": 398164, "image": "COCO_train2014_000000398164.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis racket being held by the woman in the front\"."}], "task": "refering", "answer_template": "The \"the tennis racket being held by the woman in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205], "bbox": [0.701546875, 0.3447877358490566, 0.953921875, 0.5687971698113208], "iscrowd": 0, "area": 9780.3089, "rle": {"size": [424, 640], "counts": "YTj52U=4M3L101N100O100O100O100O100O100O101N100O1O1O1O100O1O1O1O100O1O1O2N100O1O1O100O1O1O1O100O1O1O2N1N2N2O1N2O1N2N2O1N2N2O1O1O1O2N1O100O1O1O1O1O1O1O100O1O1O2N1O1O1O100O100O100000000O10001O0O100000000O1000000O1000001O0O1000000O100000000O10O2O0O10000O2O0O100O2O0O10001N100O101N10000O101N100O3N1O1N3N1N3N1N3L3M3M4L3M4L3M3M4L3M4Ld0\\OVj;"}, "label": "the tennis racket being held by the woman in the front"}]}
{"id": 103251, "image": "COCO_train2014_000000103251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl in pink and white pajamas\"."}], "task": "refering", "answer_template": "The \"a young girl in pink and white pajamas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 329, 330], "bbox": [0.0960625, 0.07022916666666666, 0.45168749999999996, 0.8365208333333333], "iscrowd": 0, "area": 45355.384099999996, "rle": {"size": [480, 640], "counts": "\\fl02X>h0XOh0K5L4M3M3L4M3M3M2M4M3M3M3L4M3M3L4N2N2N2N1O2M3N2N2N2N1O2N2M2O2N2N1XFmKa8U4[GQLa8Q4ZGVLb8k3[G[L`8h3[G_La8c3[GcLa8^3[GiLa8o4L4L4L4K5Lc1]Nf0ZO2N1O2N1N3N1O2N1O2N1O2M3N1O2N2N2N2N2N2M3N2N2N2N2N2N2N2O1N2N1O001O01O0001O0001O0O10000O2O0O1000001O000000100O1O1O100O1O1O1O100O1O2N100O1O1O100O1O1O3M9H7H9H2M2O2M2N3N1N3N1N3M2gHeIj5\\6nIkIQ6X6fIoIZ6R6]IVJb6m5UIZJk6g5mH`JR7h6O010O010O0`NSHmKn7Q4UHnKj7R4XHlKi7R4ZHmKe7R4^HlKc7S4_HlK`7S4bHlK_7R4cHmK]7S4cHnK]7P4dHPL\\7o3fHPL[7n3fHSLY7m3gHSLZ7k3hHTLX7k3iHULW7k3iHULX7i3jHVLV7i3kHWLV7h3jHXLV7g3lHXLU7f3lHZLT7e3mH\\LS7c3nH\\LR7c3oH]LR7m0oG`0P1cNQ7n0QH<o0eNQ7n0TH8l0jNP7o0WH2j0oNP7n0YH0h0QOo6o0]HKe0VOo6o0^HGd0ZOn6o0bHBa0@m6m0fH_O=Cn6o0hHYO;Hn6n0kHUO8Lo6o0mHPO41P7m0QImN05Q7m0SIiNM9R7m0UIfNI=T7k0XIbNEb0U7k0QKTOQ5j0PKTOS5k0mJROW5m0iJoN\\5P1eJlN`5R1eJeN`5Y1U4O1O1O2N1N101O0000000O2O00000O101O000O10001O0O2O1O0O2O001O0O2O1O001N7J;EfYT5"}, "label": "a young girl in pink and white pajamas"}]}
{"id": 103251, "image": "COCO_train2014_000000103251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child where a white and pink shirt\"."}], "task": "refering", "answer_template": "The \"a small child where a white and pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 329, 330], "bbox": [0.0960625, 0.07022916666666666, 0.45168749999999996, 0.8365208333333333], "iscrowd": 0, "area": 45355.384099999996, "rle": {"size": [480, 640], "counts": "\\fl02X>h0XOh0K5L4M3M3L4M3M3M2M4M3M3M3L4M3M3L4N2N2N2N1O2M3N2N2N2N1O2N2M2O2N2N1XFmKa8U4[GQLa8Q4ZGVLb8k3[G[L`8h3[G_La8c3[GcLa8^3[GiLa8o4L4L4L4K5Lc1]Nf0ZO2N1O2N1N3N1O2N1O2N1O2M3N1O2N2N2N2N2N2M3N2N2N2N2N2N2N2O1N2N1O001O01O0001O0001O0O10000O2O0O1000001O000000100O1O1O100O1O1O1O100O1O2N100O1O1O100O1O1O3M9H7H9H2M2O2M2N3N1N3N1N3M2gHeIj5\\6nIkIQ6X6fIoIZ6R6]IVJb6m5UIZJk6g5mH`JR7h6O010O010O0`NSHmKn7Q4UHnKj7R4XHlKi7R4ZHmKe7R4^HlKc7S4_HlK`7S4bHlK_7R4cHmK]7S4cHnK]7P4dHPL\\7o3fHPL[7n3fHSLY7m3gHSLZ7k3hHTLX7k3iHULW7k3iHULX7i3jHVLV7i3kHWLV7h3jHXLV7g3lHXLU7f3lHZLT7e3mH\\LS7c3nH\\LR7c3oH]LR7m0oG`0P1cNQ7n0QH<o0eNQ7n0TH8l0jNP7o0WH2j0oNP7n0YH0h0QOo6o0]HKe0VOo6o0^HGd0ZOn6o0bHBa0@m6m0fH_O=Cn6o0hHYO;Hn6n0kHUO8Lo6o0mHPO41P7m0QImN05Q7m0SIiNM9R7m0UIfNI=T7k0XIbNEb0U7k0QKTOQ5j0PKTOS5k0mJROW5m0iJoN\\5P1eJlN`5R1eJeN`5Y1U4O1O1O2N1N101O0000000O2O00000O101O000O10001O0O2O1O0O2O001O0O2O1O001N7J;EfYT5"}, "label": "a small child where a white and pink shirt"}]}
{"id": 103251, "image": "COCO_train2014_000000103251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a blue shirt\"."}], "task": "refering", "answer_template": "The \"a girl wearing a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 388, 389, 390], "bbox": [0.5309062499999999, 0.07639583333333334, 0.9994375, 0.9842708333333333], "iscrowd": 0, "area": 73411.46170000003, "rle": {"size": [480, 640], "counts": "neo49a><D9F;J4M2M3N2M3M3N2M3N2M3M3N1N2O1N3M2N2N2N3M2XOh0O1O2N1O1O1O1O1cLYL`Kh3]4_L^Kb3_4eL[K]3b4jLYKW3e4QMTKP3i4ZMoJg2n4cMjJ^2S5kMfJV2W5TNaJm1\\5\\N]Je1`5aN\\J`1b5cN\\J^1b5eN\\J\\1a5hN^JX1`5kN^JV1_5oN^JR1`5QO_Jo0_5UO^Jl0_5XO`Jh0`5ZO]Jg0b5[O]Je0b5]O\\Jd0c5_OZJb0e5@ZJ`0f5BWJ?h5CWJ=h5FUJ;j5HSJ9m5JPJ6o5^4O1O2N2O1N2N2O1N2N2O1N2N2O1N2M3N1N3N2O1N2O1O1N2O1O0O2O1O001N101O1N101O1N101O1O0O2O1O0O2O001N2O001O0YLUEX3m:fLUEY3k:fLWEY3i:fLXEY3j:eLXEZ3h:eLZEZ3g:dL[EZ3f:eL[E[3e:dL]E[3U;O1N2O001N2O2N1N2O001O1N3N1O1O1O100O100O1O100O010O100O00100O01O0O2O00001N101O001N10001O001N101O000cHSMf2o2VMUMi2k2UMWMk2j2TMWMj2j2VMWMi2j2UMXMi2i2WMXMg2i2YMWMf2k2XMWMg2i2YMXMe2j2YMXMf2h2YMZMf2g2YMYMf2h2YMZMg2f2XM[Mg2h2UMZMj2h2SMYMn2i2oLXMP3j2mLXMS3i2jLYMV3i2gLWMZ3j2cLXM]3j2_LXM`3j2^LWMb3Q3TLPMm3[3fKgLY4i4VJYKj5g4SJ[Kn5f4lI_KS6c4gIbKY6^4bIfK_6[4[IjKd6Z600O01000O01000O01000O10000O01000O10000O10O0100O1O2N100O1O100O1O010O1O1O10O01O100O001O100O1O010O1O1O010O1O100O2N3M2O2kLkF9W9ClF<W9_OkFb0W9ZOjFf0Y9UOjFj0Z9QOgFP1[9kNhFT1\\9gNeFY1^9bNdF^1_9^NcFb1^9ZNdFf1_9VNbFj1`9RNcFn1_9nMbFR2`9jMcFU2`9gMaFY2e9_M]Fb2j9TMYFk2n:NPDTMQ<h2SDXMm;e2UD[Mk;b2<M3M2M4M3M3M2N3N2N1O2N2M2O2N2N2M3E<E:E<EmC"}, "label": "a girl wearing a blue shirt"}]}
{"id": 103251, "image": "COCO_train2014_000000103251.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with a kitty at in a pink crown on her shirt\"."}], "task": "refering", "answer_template": "The \"a girl with a kitty at in a pink crown on her shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 388, 389, 390], "bbox": [0.5309062499999999, 0.07639583333333334, 0.9994375, 0.9842708333333333], "iscrowd": 0, "area": 73411.46170000003, "rle": {"size": [480, 640], "counts": "neo49a><D9F;J4M2M3N2M3M3N2M3N2M3M3N1N2O1N3M2N2N2N3M2XOh0O1O2N1O1O1O1O1cLYL`Kh3]4_L^Kb3_4eL[K]3b4jLYKW3e4QMTKP3i4ZMoJg2n4cMjJ^2S5kMfJV2W5TNaJm1\\5\\N]Je1`5aN\\J`1b5cN\\J^1b5eN\\J\\1a5hN^JX1`5kN^JV1_5oN^JR1`5QO_Jo0_5UO^Jl0_5XO`Jh0`5ZO]Jg0b5[O]Je0b5]O\\Jd0c5_OZJb0e5@ZJ`0f5BWJ?h5CWJ=h5FUJ;j5HSJ9m5JPJ6o5^4O1O2N2O1N2N2O1N2N2O1N2N2O1N2M3N1N3N2O1N2O1O1N2O1O0O2O1O001N101O1N101O1N101O1O0O2O1O0O2O001N2O001O0YLUEX3m:fLUEY3k:fLWEY3i:fLXEY3j:eLXEZ3h:eLZEZ3g:dL[EZ3f:eL[E[3e:dL]E[3U;O1N2O001N2O2N1N2O001O1N3N1O1O1O100O100O1O100O010O100O00100O01O0O2O00001N101O001N10001O001N101O000cHSMf2o2VMUMi2k2UMWMk2j2TMWMj2j2VMWMi2j2UMXMi2i2WMXMg2i2YMWMf2k2XMWMg2i2YMXMe2j2YMXMf2h2YMZMf2g2YMYMf2h2YMZMg2f2XM[Mg2h2UMZMj2h2SMYMn2i2oLXMP3j2mLXMS3i2jLYMV3i2gLWMZ3j2cLXM]3j2_LXM`3j2^LWMb3Q3TLPMm3[3fKgLY4i4VJYKj5g4SJ[Kn5f4lI_KS6c4gIbKY6^4bIfK_6[4[IjKd6Z600O01000O01000O01000O10000O01000O10000O10O0100O1O2N100O1O100O1O010O1O1O10O01O100O001O100O1O010O1O1O010O1O100O2N3M2O2kLkF9W9ClF<W9_OkFb0W9ZOjFf0Y9UOjFj0Z9QOgFP1[9kNhFT1\\9gNeFY1^9bNdF^1_9^NcFb1^9ZNdFf1_9VNbFj1`9RNcFn1_9nMbFR2`9jMcFU2`9gMaFY2e9_M]Fb2j9TMYFk2n:NPDTMQ<h2SDXMm;e2UD[Mk;b2<M3M2M4M3M3M2N3N2N1O2N2M2O2N2N2M3E<E:E<EmC"}, "label": "a girl with a kitty at in a pink crown on her shirt"}]}
{"id": 463702, "image": "COCO_train2014_000000463702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase on the near table\"."}], "task": "refering", "answer_template": "The \"the vase on the near table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 134, 135, 136, 147, 148, 160, 161, 162, 173, 174, 175, 186, 187, 188, 199, 200, 201, 212, 213, 214, 225, 226, 227], "bbox": [0.30712, 0.52386, 0.5146666666666667, 0.96714], "iscrowd": 0, "area": 12146.08265, "rle": {"size": [500, 375], "counts": "a]h11c?3M2N3hD3c60RIY1m5iNiIc1S6`NcIg1]6\\NXIm1g6UNPIR2P7QNiHT2V7oMcHV2\\7lM^HX2b7kMWHZ2h7iMQH\\2n7fMlG^2T8eMeG`2Z8cM_Gb2`8cMWGa2i8jMhF[2W9R2O001O1O001O1O001O1OO0100000000000O1000000000000O100000000000O01000000000O1OSKoFX3P9eL^GR3a8jLmGm2R8oL\\Hg2c7VMkHa2T7[MTIa2l6[M\\Ib2b6\\MdIa2\\6^MiI^2Y6`MjI^2W6_MnI]2U6`MnI^2S6`MQJ\\2Q6aMSJ\\2o5bMUJZ2n5cMZJT2i5jMcJi1`5TNmJ_1P:D=B?Ae]h2"}, "label": "the vase on the near table"}]}
{"id": 463702, "image": "COCO_train2014_000000463702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase half filled with water\"."}], "task": "refering", "answer_template": "The \"a vase half filled with water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 134, 135, 136, 147, 148, 160, 161, 162, 173, 174, 175, 186, 187, 188, 199, 200, 201, 212, 213, 214, 225, 226, 227], "bbox": [0.30712, 0.52386, 0.5146666666666667, 0.96714], "iscrowd": 0, "area": 12146.08265, "rle": {"size": [500, 375], "counts": "a]h11c?3M2N3hD3c60RIY1m5iNiIc1S6`NcIg1]6\\NXIm1g6UNPIR2P7QNiHT2V7oMcHV2\\7lM^HX2b7kMWHZ2h7iMQH\\2n7fMlG^2T8eMeG`2Z8cM_Gb2`8cMWGa2i8jMhF[2W9R2O001O1O001O1O001O1OO0100000000000O1000000000000O100000000000O01000000000O1OSKoFX3P9eL^GR3a8jLmGm2R8oL\\Hg2c7VMkHa2T7[MTIa2l6[M\\Ib2b6\\MdIa2\\6^MiI^2Y6`MjI^2W6_MnI]2U6`MnI^2S6`MQJ\\2Q6aMSJ\\2o5bMUJZ2n5cMZJT2i5jMcJi1`5TNmJ_1P:D=B?Ae]h2"}, "label": "a vase half filled with water"}]}
{"id": 463702, "image": "COCO_train2014_000000463702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown table with a vase sitting on it\"."}], "task": "refering", "answer_template": "The \"a brown table with a vase sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.003466666666666667, 0.61038, 0.9974133333333334, 1.0], "iscrowd": 0, "area": 48399.67294999999, "rle": {"size": [500, 375], "counts": "gn0>d:b4A?00O1000000O1000000O1000001N1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O010O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O0010O01O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001OZ1"}, "label": "a brown table with a vase sitting on it"}]}
{"id": 463702, "image": "COCO_train2014_000000463702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tabletop the closest flowers are sitting on\"."}], "task": "refering", "answer_template": "The \"the tabletop the closest flowers are sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.003466666666666667, 0.61038, 0.9974133333333334, 1.0], "iscrowd": 0, "area": 48399.67294999999, "rle": {"size": [500, 375], "counts": "gn0>d:b4A?00O1000000O1000000O1000001N1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O010O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O0010O01O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001OZ1"}, "label": "the tabletop the closest flowers are sitting on"}]}
{"id": 463702, "image": "COCO_train2014_000000463702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table with the candle\"."}], "task": "refering", "answer_template": "The \"the table with the candle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129], "bbox": [0.22813333333333333, 0.46656, 1.0, 0.56116], "iscrowd": 0, "area": 4731.38375, "rle": {"size": [500, 375], "counts": "ZWZ11c?001O00001O001O00001O001O00001N101O00001O001O001O00001O03N1N2O2M2N2OV[k00edTO4L4L5L3L4N2O2M2O2aASOT>9eA;4_OV>5hA<OAY>1jAg0U>XOmAg0S>XOoAh0P>XOPBh0P>XOPBi0o=WORBh0m=YOSBh0l=XOTBg0m=YOSBg0m=YOSBg0m=YOSBg0m=YOTBf0l=ZOTBf0l=ZOTBf0l=ZOTBf0l=[OSBd0n=\\ORBd0n=\\ORBd0n=\\ORBd0n=\\ORBd0m=]OSBc0_>0000O010000000000000O3N1O1O00000000000O10O1000000O100O010O1O1O1000000O010AXORBh0n=[OmAg0T><0000000O101O000000001O0O100000001O0000000O101O000000001O0O100000001O00000O101O00000000001N100000001O00000O10001O00000J6K6O00000000001O000000001O0004L1O00010O00000001O000001O0001O00000001OO2O000000001O000O1000001O0000000O2O00000000001N100000001O00000OQH"}, "label": "the table with the candle"}]}
{"id": 463702, "image": "COCO_train2014_000000463702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden table in the back\"."}], "task": "refering", "answer_template": "The \"the wooden table in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129], "bbox": [0.22813333333333333, 0.46656, 1.0, 0.56116], "iscrowd": 0, "area": 4731.38375, "rle": {"size": [500, 375], "counts": "ZWZ11c?001O00001O001O00001O001O00001N101O00001O001O001O00001O03N1N2O2M2N2OV[k00edTO4L4L5L3L4N2O2M2O2aASOT>9eA;4_OV>5hA<OAY>1jAg0U>XOmAg0S>XOoAh0P>XOPBh0P>XOPBi0o=WORBh0m=YOSBh0l=XOTBg0m=YOSBg0m=YOSBg0m=YOSBg0m=YOTBf0l=ZOTBf0l=ZOTBf0l=ZOTBf0l=[OSBd0n=\\ORBd0n=\\ORBd0n=\\ORBd0n=\\ORBd0m=]OSBc0_>0000O010000000000000O3N1O1O00000000000O10O1000000O100O010O1O1O1000000O010AXORBh0n=[OmAg0T><0000000O101O000000001O0O100000001O0000000O101O000000001O0O100000001O00000O101O00000000001N100000001O00000O10001O00000J6K6O00000000001O000000001O0004L1O00010O00000001O000001O0001O00000001OO2O000000001O000O1000001O0000000O2O00000000001N100000001O00000OQH"}, "label": "the wooden table in the back"}]}
{"id": 152406, "image": "COCO_train2014_000000152406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting with her knees up , talking on a phone\"."}], "task": "refering", "answer_template": "The \"a woman sitting with her knees up , talking on a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287], "bbox": [0.19871875, 0.27451097804391217, 0.564015625, 0.7068463073852296], "iscrowd": 0, "area": 25571.066500000015, "rle": {"size": [501, 640], "counts": "S_n12a?3M3O000O101N100O10000O1000000O1000000O1000000O100O1O100O100O100O1O100N2O1O1O0O2N2M3N2M3N2M3N2K4L5K5K5L4L4K5FgMTC]2i<90O0100MNYC\\Md<c2`C]M\\<b2iC^MT<\\2UDdMh;]2ZDaMf;_2g00001O000000001O1N3N1O1O2N12O6WDcMn9a2jEdMT:`2gEbMY:`2dEaM\\:b2`E_M`:d2]E]Mb:f2ZE[Mf:h2VEYMj:i2TEWMm:i2QEXMo:j2oDVMQ;k2mDVMS;l2kDTMU;n2hDSMY;m2eDTM[;n2cDRM];o2aDRM_;b3O1N2N200O100000000000000000O100000000000001O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1M3M01O2M3QOYLnEi3o9[LmEh3Q:eLaE^3^:gL\\E[3c:l0O1O1POZKoFh4P9ZKjFk4V9VKhFk4W9WKfFk4Z9g001O100O1O1O100O1O1O100O1O1O100O1O1O010000O2O000O1000000O10000O101O0O101O0O2O1O0O2O1O0O2O1N2N2N2O1N2N2N4L4M3[KdEj3`:hKVFa3V;L4L4M3L4L4M3L3M4N2M3N1N3N2M2N3N2M2N3J6K4K6K5J8I8H9H:EnXX4"}, "label": "a woman sitting with her knees up , talking on a phone"}]}
{"id": 152406, "image": "COCO_train2014_000000152406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady on the steps\"."}], "task": "refering", "answer_template": "The \"the lady on the steps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287], "bbox": [0.19871875, 0.27451097804391217, 0.564015625, 0.7068463073852296], "iscrowd": 0, "area": 25571.066500000015, "rle": {"size": [501, 640], "counts": "S_n12a?3M3O000O101N100O10000O1000000O1000000O1000000O100O1O100O100O100O1O100N2O1O1O0O2N2M3N2M3N2M3N2K4L5K5K5L4L4K5FgMTC]2i<90O0100MNYC\\Md<c2`C]M\\<b2iC^MT<\\2UDdMh;]2ZDaMf;_2g00001O000000001O1N3N1O1O2N12O6WDcMn9a2jEdMT:`2gEbMY:`2dEaM\\:b2`E_M`:d2]E]Mb:f2ZE[Mf:h2VEYMj:i2TEWMm:i2QEXMo:j2oDVMQ;k2mDVMS;l2kDTMU;n2hDSMY;m2eDTM[;n2cDRM];o2aDRM_;b3O1N2N200O100000000000000000O100000000000001O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1M3M01O2M3QOYLnEi3o9[LmEh3Q:eLaE^3^:gL\\E[3c:l0O1O1POZKoFh4P9ZKjFk4V9VKhFk4W9WKfFk4Z9g001O100O1O1O100O1O1O100O1O1O100O1O1O010000O2O000O1000000O10000O101O0O101O0O2O1O0O2O1O0O2O1N2N2N2O1N2N2N4L4M3[KdEj3`:hKVFa3V;L4L4M3L4L4M3L3M4N2M3N1N3N2M2N3N2M2N3J6K4K6K5J8I8H9H:EnXX4"}, "label": "the lady on the steps"}]}
{"id": 471900, "image": "COCO_train2014_000000471900.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a red coat with brown fur on the collar\"."}], "task": "refering", "answer_template": "The \"a woman wearing a red coat with brown fur on the collar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 326, 327, 328, 329, 330], "bbox": [0.095125, 0.12616707616707618, 0.38809375, 1.0], "iscrowd": 0, "area": 43819.9034, "rle": {"size": [407, 640], "counts": "^^h0d0m;;G<D<E8nERNn8U2fFYNS9m2H7Ig0YO2N2N1O2N2M3N2N2N1_NXKRKj4g4`KSKc4h4eKRK^4k4fKRK[4m4gKQK[4l4hKRKY4l4jKRKX4k4kKSKW4k4kKTKU4k4mKSKU4k4mKSK_4a4cK]Kd4[4_KcKf4X4\\KeKg4Y4[KfKf4X4\\KgKe4W4]KgKe4W4]KhKe4U4\\KjKf4S4]KlKd4R4^KlKd4Q4S2N2O0O2O1N101N2O00O100O02O1O1O002M2O1O1O2N2M2Kb0]O:J1O1O1O1O1O1O1N2O1O1O1N200O1O100O1O2O0O1O1O100O1O100O1000000001O000000000000000000O1000000O1000O1O1N2O2M3N2000O0100000O01000O01O1O001O1O001O2N2N2N2fFVN\\4AeM[2hMZNa4]OfM[2fMYNb4@jJ0^2Y2RNYNd4BgJ38Dk1`2[N]Nh4=PKlNOMo0_2UO^Nl4=kJZOS1X3P4\\NoKe1n3^NQLb1m3aNRL`1l3bNSL_1i3eNWL[1d3jN[LW1b3lN]LU1a3mN^LU1_3mN`LT1o0nK=n2dNU1j0XL:d2lNU1d0eL6X2UOT1?mL6Q2[OS1:RM8m1]OT13VM;i1BS1CdMb0a1Jg3G`L9P8O001O001O001O1O001O001O00001O001O1O1O1O1O2N1N2O1N2O2NU]k4"}, "label": "a woman wearing a red coat with brown fur on the collar"}]}
{"id": 471900, "image": "COCO_train2014_000000471900.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing red jacket\"."}], "task": "refering", "answer_template": "The \"a woman wearing red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 326, 327, 328, 329, 330], "bbox": [0.095125, 0.12616707616707618, 0.38809375, 1.0], "iscrowd": 0, "area": 43819.9034, "rle": {"size": [407, 640], "counts": "^^h0d0m;;G<D<E8nERNn8U2fFYNS9m2H7Ig0YO2N2N1O2N2M3N2N2N1_NXKRKj4g4`KSKc4h4eKRK^4k4fKRK[4m4gKQK[4l4hKRKY4l4jKRKX4k4kKSKW4k4kKTKU4k4mKSKU4k4mKSK_4a4cK]Kd4[4_KcKf4X4\\KeKg4Y4[KfKf4X4\\KgKe4W4]KgKe4W4]KhKe4U4\\KjKf4S4]KlKd4R4^KlKd4Q4S2N2O0O2O1N101N2O00O100O02O1O1O002M2O1O1O2N2M2Kb0]O:J1O1O1O1O1O1O1N2O1O1O1N200O1O100O1O2O0O1O1O100O1O100O1000000001O000000000000000000O1000000O1000O1O1N2O2M3N2000O0100000O01000O01O1O001O1O001O2N2N2N2fFVN\\4AeM[2hMZNa4]OfM[2fMYNb4@jJ0^2Y2RNYNd4BgJ38Dk1`2[N]Nh4=PKlNOMo0_2UO^Nl4=kJZOS1X3P4\\NoKe1n3^NQLb1m3aNRL`1l3bNSL_1i3eNWL[1d3jN[LW1b3lN]LU1a3mN^LU1_3mN`LT1o0nK=n2dNU1j0XL:d2lNU1d0eL6X2UOT1?mL6Q2[OS1:RM8m1]OT13VM;i1BS1CdMb0a1Jg3G`L9P8O001O001O001O1O001O001O00001O001O1O1O1O1O2N1N2O1N2O2NU]k4"}, "label": "a woman wearing red jacket"}]}
{"id": 471900, "image": "COCO_train2014_000000471900.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding skis\"."}], "task": "refering", "answer_template": "The \"man holding skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 35, 36, 37, 38, 57, 58, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 134, 135, 149, 150, 151, 152, 153, 154, 155, 157, 158, 172, 173, 174, 175, 176, 177, 178, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339], "bbox": [0.335046875, 0.03447174447174447, 0.9287656250000002, 0.9987223587223586], "iscrowd": 0, "area": 70033.87084999998, "rle": {"size": [407, 640], "counts": "Vge25_<4L4L5K4L4L5K4L5K4N2N3M2N2N3M2O1O1O01O00001O001O001O100O1O2N1O1O1O1O1O1O1O1O2N1O1O0000O101O001N101O001N101O000O1000000O1000000O1000000O10000O10000O1O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2cLVLjMm3f1dLWN]3h1fLTN]3k1eLRN\\3n1eLoM^3P2dLlM^3U2cLfMa3Y2aLcMa3]2`L_Md3`2^L[Me3QOUKc2X1WNg3SO\\K]2o0YNj3XOcKV2d0[No3\\OhKP2:^NS4@nKi11_NX4FQL6VNd0c1YO[4KRL5PNf0e1SO_4OPL7lMf0f1mNd44nK7iMh0V8nNUJ9eMh0V8mNZJ9`Mi0W8lN\\J9^Mj0P;UOPEj0R;UOnDj0T;UOlDk0U;SOlDl0`;O1O1O1O100O1O100O100O100O1O100O100O1O100O100O100O1010O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O000010O04L3M4L4L4L4L4L4L3M4L4L1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O00100O2N2N2N1O2N2N2N2N1O2N2N1O1O001O00001O001O001O001O00001O001ZOeLnG[3d7VMYHk2X7eMfH[2l6P2A>B?A>B>Ik1]NQ2oMN2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N101O1N2O1O0O10O10O0100O010O1OWOj0\\Nc1B?0O2O0O1O2O0O101N100O1O2O0O101N100O101N1O101N2O2M3N2M3M3N2M4M2M3N2M3N2M3M3N2M4M2M3M3K5L4K5L4K5K6K4K5K5L4K5L4K5K6K4K6K5JVRb0"}, "label": "man holding skis"}]}
{"id": 471900, "image": "COCO_train2014_000000471900.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elderly man in a white ski jacket holding red skis and black ski poles\"."}], "task": "refering", "answer_template": "The \"an elderly man in a white ski jacket holding red skis and black ski poles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 35, 36, 37, 38, 57, 58, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 134, 135, 149, 150, 151, 152, 153, 154, 155, 157, 158, 172, 173, 174, 175, 176, 177, 178, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339], "bbox": [0.335046875, 0.03447174447174447, 0.9287656250000002, 0.9987223587223586], "iscrowd": 0, "area": 70033.87084999998, "rle": {"size": [407, 640], "counts": "Vge25_<4L4L5K4L4L5K4L5K4N2N3M2N2N3M2O1O1O01O00001O001O001O100O1O2N1O1O1O1O1O1O1O1O2N1O1O0000O101O001N101O001N101O000O1000000O1000000O1000000O10000O10000O1O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2cLVLjMm3f1dLWN]3h1fLTN]3k1eLRN\\3n1eLoM^3P2dLlM^3U2cLfMa3Y2aLcMa3]2`L_Md3`2^L[Me3QOUKc2X1WNg3SO\\K]2o0YNj3XOcKV2d0[No3\\OhKP2:^NS4@nKi11_NX4FQL6VNd0c1YO[4KRL5PNf0e1SO_4OPL7lMf0f1mNd44nK7iMh0V8nNUJ9eMh0V8mNZJ9`Mi0W8lN\\J9^Mj0P;UOPEj0R;UOnDj0T;UOlDk0U;SOlDl0`;O1O1O1O100O1O100O100O100O1O100O100O1O100O100O100O1010O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O000010O04L3M4L4L4L4L4L4L3M4L4L1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O00100O2N2N2N1O2N2N2N2N1O2N2N1O1O001O00001O001O001O001O00001O001ZOeLnG[3d7VMYHk2X7eMfH[2l6P2A>B?A>B>Ik1]NQ2oMN2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N101O1N2O1O0O10O10O0100O010O1OWOj0\\Nc1B?0O2O0O1O2O0O101N100O1O2O0O101N100O101N1O101N2O2M3N2M3M3N2M4M2M3N2M3N2M3M3N2M4M2M3M3K5L4K5L4K5K6K4K5K5L4K5L4K5K6K4K6K5JVRb0"}, "label": "an elderly man in a white ski jacket holding red skis and black ski poles"}]}
{"id": 160614, "image": "COCO_train2014_000000160614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the book cover looks like a real face\"."}], "task": "refering", "answer_template": "The \"the book cover looks like a real face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194], "bbox": [0.40740625, 0.1851875, 0.5851875, 0.5061666666666667], "iscrowd": 0, "area": 12320.7445, "rle": {"size": [480, 640], "counts": "Saj32m>4L4M3L=C<D<D3M01O00001O001O1O0010kMFdE9Z:O`E2\\:5aEJ\\:;cED\\:?bEB^:?aE@`:a0^E@b:a0]E_Ob:c0]E\\Od:e0ZE\\Of:e0YE[Oa:cNXES26YOa:gNXEQ25YOb:hNXEP25XOb:S1\\EnNd:S1[ElNf:U1XElNh:f20O1000000O10000O100000O10000000000000000000000000000000000000001O0O10000000000000000000000000000001O1O100O1O1O1O1O1O001O1O2O1N3M2N3M2N4L5K5KT1lN2N3M2N3M3L6KT1iNYPl3"}, "label": "the book cover looks like a real face"}]}
{"id": 160614, "image": "COCO_train2014_000000160614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the face of a man wearing glasses\"."}], "task": "refering", "answer_template": "The \"the face of a man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194], "bbox": [0.40740625, 0.1851875, 0.5851875, 0.5061666666666667], "iscrowd": 0, "area": 12320.7445, "rle": {"size": [480, 640], "counts": "Saj32m>4L4M3L=C<D<D3M01O00001O001O1O0010kMFdE9Z:O`E2\\:5aEJ\\:;cED\\:?bEB^:?aE@`:a0^E@b:a0]E_Ob:c0]E\\Od:e0ZE\\Of:e0YE[Oa:cNXES26YOa:gNXEQ25YOb:hNXEP25XOb:S1\\EnNd:S1[ElNf:U1XElNh:f20O1000000O10000O100000O10000000000000000000000000000000000000001O0O10000000000000000000000000000001O1O100O1O1O1O1O1O001O1O2O1N3M2N3M2N4L5K5KT1lN2N3M2N3M3L6KT1iNYPl3"}, "label": "the face of a man wearing glasses"}]}
{"id": 160614, "image": "COCO_train2014_000000160614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in bed reading a book\"."}], "task": "refering", "answer_template": "The \"a guy in bed reading a book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 79, 80, 81, 130, 131, 132, 154, 155, 169, 170, 177, 178, 179, 190, 191, 192, 193, 199, 200, 201, 202, 213, 223, 224, 225, 226, 248, 249, 250, 273, 274], "bbox": [0.259546875, 0.1528125, 0.9286562500000001, 0.6741666666666667], "iscrowd": 0, "area": 12493.005150000005, "rle": {"size": [480, 640], "counts": "Ya^21n>10O01O1O001O10O01O1O001O1O010O1O001O1O010O1O001O1O010O1O1O001O10O01O1O001O10O0100O010O10O010000O010O10O0100O10O0100O01000O10O0100O01000000O10001N10000O1000000O10000O101O000O10000O10001O3L4M2N3Lcf9O_YF0O010O100O1O100O100O010O100O1O100O100O010O1O100O100O100O010O1O100O11O00000O101O000000001O000O10001O0000001O000O10001O000000001O0O3N2Nikj00WTUO001O001O001O0O2O001O001O001O001N101O001O0]MF`F;_9G_F9`9J^F7`9L^F4a9N]F4b9N\\F2c91[F0c93[FMd96ZFKe97YFIf9:XFGf9<XFDg9?WFBg9a0VF@j9a0UF@i9c0UF]Oj9f0TF[Oj9h0TFXOl9j0RFWOm9k0QFVOo9n0lEROT:W1cEjN\\:]1]EcNc:]1]EdNb:]1]EcNd:\\1]EdNb:]1]EcNc:]1]EdNb:]1]EcNc:]1]EdNb:]1]EcNd:\\1\\EeNc:\\1\\EdNd:\\1\\EeNc:\\1\\EdNd:\\1\\EeNc:\\1\\EdNe:[1[EfNd:[1[EeNe:[1\\EeNc:]1[EcNe:^1ZEcNe:_1YEaNh:_1WEbNh:`1VE`Nj:a1UE`Nj:b1TE^Nl:c1SE^Nl:d1RE\\No:k21N101N2O1N101N2O1N101N2O0O2O1N2O0WM]DS2e;dMdD[2];[MmDc2i;01O1N2O0O2O1N2O001N2O0O2O1N2O0O2O1O1N101N2O0O2O1O1N101N2O1N101N2O001N2O0O100O100O10000O100O100O10000O100O100O100O1O100O1O00100O1O1O100O1O1O1O2O0O1O2N100O1O2N100O2N1O1O101N1O1O2O0O1O1OnWe0"}, "label": "a guy in bed reading a book"}]}
{"id": 160614, "image": "COCO_train2014_000000160614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person reading a book in bed\"."}], "task": "refering", "answer_template": "The \"a person reading a book in bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 79, 80, 81, 130, 131, 132, 154, 155, 169, 170, 177, 178, 179, 190, 191, 192, 193, 199, 200, 201, 202, 213, 223, 224, 225, 226, 248, 249, 250, 273, 274], "bbox": [0.259546875, 0.1528125, 0.9286562500000001, 0.6741666666666667], "iscrowd": 0, "area": 12493.005150000005, "rle": {"size": [480, 640], "counts": "Ya^21n>10O01O1O001O10O01O1O001O1O010O1O001O1O010O1O001O1O010O1O1O001O10O01O1O001O10O0100O010O10O010000O010O10O0100O10O0100O01000O10O0100O01000000O10001N10000O1000000O10000O101O000O10000O10001O3L4M2N3Lcf9O_YF0O010O100O1O100O100O010O100O1O100O100O010O1O100O100O100O010O1O100O11O00000O101O000000001O000O10001O0000001O000O10001O000000001O0O3N2Nikj00WTUO001O001O001O0O2O001O001O001O001N101O001O0]MF`F;_9G_F9`9J^F7`9L^F4a9N]F4b9N\\F2c91[F0c93[FMd96ZFKe97YFIf9:XFGf9<XFDg9?WFBg9a0VF@j9a0UF@i9c0UF]Oj9f0TF[Oj9h0TFXOl9j0RFWOm9k0QFVOo9n0lEROT:W1cEjN\\:]1]EcNc:]1]EdNb:]1]EcNd:\\1]EdNb:]1]EcNc:]1]EdNb:]1]EcNc:]1]EdNb:]1]EcNd:\\1\\EeNc:\\1\\EdNd:\\1\\EeNc:\\1\\EdNd:\\1\\EeNc:\\1\\EdNe:[1[EfNd:[1[EeNe:[1\\EeNc:]1[EcNe:^1ZEcNe:_1YEaNh:_1WEbNh:`1VE`Nj:a1UE`Nj:b1TE^Nl:c1SE^Nl:d1RE\\No:k21N101N2O1N101N2O1N101N2O0O2O1N2O0WM]DS2e;dMdD[2];[MmDc2i;01O1N2O0O2O1N2O001N2O0O2O1N2O0O2O1O1N101N2O0O2O1O1N101N2O1N101N2O001N2O0O100O100O10000O100O100O10000O100O100O100O1O100O1O00100O1O1O100O1O1O1O2O0O1O2N100O1O2N100O2N1O1O101N1O1O2O0O1O1OnWe0"}, "label": "a person reading a book in bed"}]}
{"id": 430950, "image": "COCO_train2014_000000430950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the man in the red shirt is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the man in the red shirt is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 165, 166, 167, 168, 182, 190, 202, 203, 204, 221, 222, 223, 224, 225, 243, 244, 245, 246, 265, 266, 267, 287, 288, 310], "bbox": [0.06032679738562092, 0.3064542483660131, 0.6787418300653596, 0.6663235294117648], "iscrowd": 0, "area": 8237.187750000001, "rle": {"size": [612, 612], "counts": "W]f0c0`b04K5K5K5K5K5M3M3M3L4M4L3M3M3M3M3M3M3M3M3M3N2M3N2N2N3M0OL5J6I7I7L3M3N1O2M2O2N1N3N1N2N3M2L5O001N1O2N1O2N101N1O2O001O000010O01O010O001O010O00001O001O0010O01O0001O01O00010O0000010O01O01O00010O00001O000010O0001O001O000010O00010O00010O01O000O101O00001O0000001O00001O001O0O2O0010O10O010O010O1O001O1O001O1O2N3L6JmSQ2MXlnM00001O000000000000000O100000000000000000000O100000000000000000000O10000000001O00000000000O100000000000000000000000000000000N3L3M3N2O1O1O1O1N2O1O1O1O1O1O1O2N100O1O100O100000000001O001O001O001O001O010O001O2N1O1O1O1O1O1O1O2N3M2M3N3Lm`e3"}, "label": "the chair the man in the red shirt is sitting in"}]}
{"id": 430950, "image": "COCO_train2014_000000430950.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown chair the man is sitting on\"."}], "task": "refering", "answer_template": "The \"brown chair the man is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 165, 166, 167, 168, 182, 190, 202, 203, 204, 221, 222, 223, 224, 225, 243, 244, 245, 246, 265, 266, 267, 287, 288, 310], "bbox": [0.06032679738562092, 0.3064542483660131, 0.6787418300653596, 0.6663235294117648], "iscrowd": 0, "area": 8237.187750000001, "rle": {"size": [612, 612], "counts": "W]f0c0`b04K5K5K5K5K5M3M3M3L4M4L3M3M3M3M3M3M3M3M3M3N2M3N2N2N3M0OL5J6I7I7L3M3N1O2M2O2N1N3N1N2N3M2L5O001N1O2N1O2N101N1O2O001O000010O01O010O001O010O00001O001O0010O01O0001O01O00010O0000010O01O01O00010O00001O000010O0001O001O000010O00010O00010O01O000O101O00001O0000001O00001O001O0O2O0010O10O010O010O1O001O1O001O1O2N3L6JmSQ2MXlnM00001O000000000000000O100000000000000000000O100000000000000000000O10000000001O00000000000O100000000000000000000000000000000N3L3M3N2O1O1O1O1N2O1O1O1O1O1O1O2N100O1O100O100000000001O001O001O001O001O010O001O2N1O1O1O1O1O1O1O2N3M2M3N3Lm`e3"}, "label": "brown chair the man is sitting on"}]}
{"id": 258927, "image": "COCO_train2014_000000258927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman petting an elephant\"."}], "task": "refering", "answer_template": "The \"a woman petting an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 123, 124, 125, 140, 141, 142, 144, 145, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 207, 208, 209, 210, 224, 225, 226, 227, 241, 242, 243, 244, 258, 259, 260, 261, 275, 276, 277, 278, 293, 294, 295, 310, 311, 312, 327, 328, 329, 344, 345, 346, 361, 362, 378, 379, 380], "bbox": [0.18535416666666665, 0.299546875, 0.5441041666666666, 0.9901406250000001], "iscrowd": 0, "area": 33940.48015, "rle": {"size": [640, 480], "counts": "`og14hc06I7I7I8H7I7I7I7I7I7I8H<[Od0YOh0G9G9H7H9H8G8H9H8G9H7H9H8G8H9H8]FPIf5Y7dIVI[6Q7oH_In6k6\\HdIa7_9M3M2M4M3K5L4K5L3N3O1N2N2N2N1O2O1N200O00E<0O101N101N10001O001O00VO_IVD`6h;cIWD]6h;dIXD\\6f;gIZDX6e;jIZDV6d;lI]DS6a;PJ^DP6a;RJ^Dn5`;TJ`Dl5_;VJ`Dj5^;YJaDg5];[JcDe5\\;]JcDc5[;`JdD`5[;bJeD]5Y;eJgD[5X;gJgDY5S;nJlDR5m:UKSEk4e:^KXEd4a:dK]E]4[:kKdEV4U:RLiEP4n9YLPFh3i9_LVFb3c9fLZF]3]9g3I7H9H9G;D;F:E;F9G9F:kJfBH7o1\\=nMfB3a0T1S=^NeB>l09h<QObBe0Z1_O^<E]Bl0``0nNd_OR1fa0N1000000O2O000O10000O2O000O10001N1000000O2O000O0100O001O1O00100O001O1O1O00100O0000O10O0O2N1O2N1O1O2N1O2O0O1O2N1O20O2O7H9G8I7H8I[dX4"}, "label": "a woman petting an elephant"}]}
{"id": 258927, "image": "COCO_train2014_000000258927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman petting an elephant\"."}], "task": "refering", "answer_template": "The \"a woman petting an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 123, 124, 125, 140, 141, 142, 144, 145, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 207, 208, 209, 210, 224, 225, 226, 227, 241, 242, 243, 244, 258, 259, 260, 261, 275, 276, 277, 278, 293, 294, 295, 310, 311, 312, 327, 328, 329, 344, 345, 346, 361, 362, 378, 379, 380], "bbox": [0.18535416666666665, 0.299546875, 0.5441041666666666, 0.9901406250000001], "iscrowd": 0, "area": 33940.48015, "rle": {"size": [640, 480], "counts": "`og14hc06I7I7I8H7I7I7I7I7I7I8H<[Od0YOh0G9G9H7H9H8G8H9H8G9H7H9H8G8H9H8]FPIf5Y7dIVI[6Q7oH_In6k6\\HdIa7_9M3M2M4M3K5L4K5L3N3O1N2N2N2N1O2O1N200O00E<0O101N101N10001O001O00VO_IVD`6h;cIWD]6h;dIXD\\6f;gIZDX6e;jIZDV6d;lI]DS6a;PJ^DP6a;RJ^Dn5`;TJ`Dl5_;VJ`Dj5^;YJaDg5];[JcDe5\\;]JcDc5[;`JdD`5[;bJeD]5Y;eJgD[5X;gJgDY5S;nJlDR5m:UKSEk4e:^KXEd4a:dK]E]4[:kKdEV4U:RLiEP4n9YLPFh3i9_LVFb3c9fLZF]3]9g3I7H9H9G;D;F:E;F9G9F:kJfBH7o1\\=nMfB3a0T1S=^NeB>l09h<QObBe0Z1_O^<E]Bl0``0nNd_OR1fa0N1000000O2O000O10000O2O000O10001N1000000O2O000O0100O001O1O00100O001O1O1O00100O0000O10O0O2N1O2N1O1O2N1O2O0O1O2N1O20O2O7H9G8I7H8I[dX4"}, "label": "a woman petting an elephant"}]}
{"id": 545651, "image": "COCO_train2014_000000545651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza which has all the toppings collected in the center\"."}], "task": "refering", "answer_template": "The \"the pizza which has all the toppings collected in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 242], "bbox": [0.463109375, 0.5214681440443213, 0.661234375, 0.7787534626038781], "iscrowd": 0, "area": 9285.243900000001, "rle": {"size": [361, 640], "counts": "bbX3?d::F:J3N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N1N3N1O101N1O101N1O101N1O101N100O2O0O101O0000001O00000O1O100000000000000000000000O10000000000000000000000O10000000001O00001N101O00001N10001O0O2O0O101O0O2O0O2O1N101O1N2O0O2N2O1N1O2N2O1N1O2N2O2M2N2O1N2N3M2M3M3M3M6J7I7HiX\\2"}, "label": "the pizza which has all the toppings collected in the center"}]}
{"id": 545651, "image": "COCO_train2014_000000545651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza closest to the little girl in the yellow sweater\"."}], "task": "refering", "answer_template": "The \"the pizza closest to the little girl in the yellow sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 242], "bbox": [0.463109375, 0.5214681440443213, 0.661234375, 0.7787534626038781], "iscrowd": 0, "area": 9285.243900000001, "rle": {"size": [361, 640], "counts": "bbX3?d::F:J3N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N1N3N1O101N1O101N1O101N1O101N100O2O0O101O0000001O00000O1O100000000000000000000000O10000000000000000000000O10000000001O00001N101O00001N10001O0O2O0O101O0O2O0O2O1N101O1N2O0O2N2O1N1O2N2O1N1O2N2O2M2N2O1N2N3M2M3M3M3M6J7I7HiX\\2"}, "label": "the pizza closest to the little girl in the yellow sweater"}]}
{"id": 545651, "image": "COCO_train2014_000000545651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right hand pizza with pepperoni topping\"."}], "task": "refering", "answer_template": "The \"right hand pizza with pepperoni topping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 246, 247, 248, 249], "bbox": [0.675859375, 0.5461772853185595, 0.8778593750000001, 0.8277008310249306], "iscrowd": 0, "area": 10211.054050000006, "rle": {"size": [361, 640], "counts": "alh48i::F9L5M2N3L3N3M2N3M2N3M3M2N3L3N2N2N2N1O2N2N2N2N2O001N2O001N10000O2O0O10001N10001N10000O2O0O10001N1000010O00001O0001O00000000010O00000000001O00000000001O00000000010O000000000O2O000O2O000O2O0O2O000O2O001N10001N101N10001N101N1N3N1N2O2M2O2M3N3L5L3L5L3L4L5K4C>BjXk0"}, "label": "right hand pizza with pepperoni topping"}]}
{"id": 545651, "image": "COCO_train2014_000000545651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with pepperonis scattered on it\"."}], "task": "refering", "answer_template": "The \"a pizza with pepperonis scattered on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 246, 247, 248, 249], "bbox": [0.675859375, 0.5461772853185595, 0.8778593750000001, 0.8277008310249306], "iscrowd": 0, "area": 10211.054050000006, "rle": {"size": [361, 640], "counts": "alh48i::F9L5M2N3L3N3M2N3M2N3M3M2N3L3N2N2N2N1O2N2N2N2N2O001N2O001N10000O2O0O10001N10001N10000O2O0O10001N1000010O00001O0001O00000000010O00000000001O00000000001O00000000010O000000000O2O000O2O000O2O0O2O000O2O001N10001N101N10001N101N1N3N1N2O2M2O2M3N3L5L3L5L3L4L5K4C>BjXk0"}, "label": "a pizza with pepperonis scattered on it"}]}
{"id": 480115, "image": "COCO_train2014_000000480115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"some sort of fighter jet in the sky flying along side two other fighter jets ; it is the biggest of the three\"."}], "task": "refering", "answer_template": "The \"some sort of fighter jet in the sky flying along side two other fighter jets ; it is the biggest of the three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143], "bbox": [0.096703125, 0.1353827160493827, 0.36546874999999995, 0.46123456790123457], "iscrowd": 0, "area": 9841.304349999997, "rle": {"size": [405, 640], "counts": "Zdh02c<3M2M3N2M4M1O0O010O10O11N1010O01O001O0010O0001O001O0kN[O[Fe0a9B[F=c9JYF6d92WFMi98SFHm9<oECR:a0kE^OV:f0eEZO[:j0aEVOV1^Oo7_2bGdM]8Z2bGjM^8S2bGPN]8n1bGVN]8h1cG[N\\8f1`G^N`8c1\\GaNb8a1YGcNf8`1UGcNk8d2OO10001N1000001N100O1O2N1O100O2J5J6N2N3M2O1O1010O1O1O010O1O1O010O1O1O010O1O1O010O1O00100O1O00100O1O000O1O010O01O001O010O001O001N1K5K6J5L5J5K6J5L6J5O1O100O2N1O1O1O1O1O2N1O100O00001O001M2O2N1N3N1O1O2M2O2O001O001O000010O01O001O001O00001O001O1O1O1O1O1O2N\\mP5"}, "label": "some sort of fighter jet in the sky flying along side two other fighter jets ; it is the biggest of the three"}]}
{"id": 480115, "image": "COCO_train2014_000000480115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plane that is bigger than the other and chunkier\"."}], "task": "refering", "answer_template": "The \"a plane that is bigger than the other and chunkier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143], "bbox": [0.096703125, 0.1353827160493827, 0.36546874999999995, 0.46123456790123457], "iscrowd": 0, "area": 9841.304349999997, "rle": {"size": [405, 640], "counts": "Zdh02c<3M2M3N2M4M1O0O010O10O11N1010O01O001O0010O0001O001O0kN[O[Fe0a9B[F=c9JYF6d92WFMi98SFHm9<oECR:a0kE^OV:f0eEZO[:j0aEVOV1^Oo7_2bGdM]8Z2bGjM^8S2bGPN]8n1bGVN]8h1cG[N\\8f1`G^N`8c1\\GaNb8a1YGcNf8`1UGcNk8d2OO10001N1000001N100O1O2N1O100O2J5J6N2N3M2O1O1010O1O1O010O1O1O010O1O1O010O1O1O010O1O00100O1O00100O1O000O1O010O01O001O010O001O001N1K5K6J5L5J5K6J5L6J5O1O100O2N1O1O1O1O1O2N1O100O00001O001M2O2N1N3N1O1O2M2O2O001O001O000010O01O001O001O00001O001O1O1O1O1O1O2N\\mP5"}, "label": "a plane that is bigger than the other and chunkier"}]}
{"id": 537461, "image": "COCO_train2014_000000537461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white toothbrush with green , blue and white bristles\"."}], "task": "refering", "answer_template": "The \"a white toothbrush with green , blue and white bristles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 89, 90, 91, 92, 106, 107, 108, 109, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330, 344, 345, 346, 347, 361, 362, 363, 364, 378, 379, 380, 381], "bbox": [0.21116666666666667, 0.19729687499999998, 0.5064583333333333, 1.0], "iscrowd": 0, "area": 52373.18129999999, "rle": {"size": [640, 480], "counts": "jRP23mc03hNK_^O7^a0L_^O8\\a0Lb^O6[a0Mb^O7Ya0Me^O5Xa0Nf^O5Q`0_OU@`0G>d?]O]@9Lo0P>[NWBi0E;FS1m=nMbBk0CP3a=YLgBY5R=kJiB]5a<VKZCS5R<i1J4L4K5L4L4K4M4L4K5gNkF_GY9[8lFbGW9Z8mFbGX9X8mFeGX9T8mFiGY9o7lFmG[9l7iFQH]9j2kE^1m0eK^9_2mEj1i0dKa9U2mEV2f0bKc9k1nEa2d0aKd9`1PFm2k;fL\\DY3k;ZL\\Dd3l;nK\\DP4k=O1N2N2N20000O1000000O1000000O1000000O1000000O100000000O1001O001O001O0010O0001O001O001O001O00001O001O001O001O00001O001O001WNQL_Co3X<`LbCa3T<oLeCQ3R<_MgCa2P<nMjCS2l;]NmCc1k;lNnCU1h;f3I8M:F7I3M3M3L4M3M2N3TF^FU9f9hF`FR9c9kFcFo8`9nFfFl8]9PGjFi8Z9TGlFf8\\:J7I;D<E;E;E9G8G:G8H8H8G:G8H8H9F]2dMPhc4"}, "label": "a white toothbrush with green , blue and white bristles"}]}
{"id": 537461, "image": "COCO_train2014_000000537461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green toothbrush with blue brushes on it\"."}], "task": "refering", "answer_template": "The \"a green toothbrush with blue brushes on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 316, 317, 318, 319, 333, 334, 335, 336, 350, 351, 352, 353, 367, 368, 369, 370, 384, 385, 386], "bbox": [0.5632916666666666, 0.341578125, 0.8509375, 0.993265625], "iscrowd": 0, "area": 45996.1924, "rle": {"size": [640, 480], "counts": "jPY5k4S?6K4K5L4K6K4K5L4K6K4L4K5L5J5L4K5L5J5L5J8I7I7H8I6I8PHoFd4X9m28RH`Fo4f9lJZFR5R:cJoE\\5d:RJ\\El5W;bIjD]6i;QIWDl6`<N2M3N2N2M3O10000000O100000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000001O001N101J5L5K4K5L5K4K6K4K6K4L5J5L5J5L4L5J5L5J5L5K4K6K4L5J5L4K6F9G:E:G:E:G:E:F;F9F:Gb0]OX1iNV1iNW1jNje\\1"}, "label": "a green toothbrush with blue brushes on it"}]}
{"id": 537461, "image": "COCO_train2014_000000537461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue , orange , white , and green toothbrush\"."}], "task": "refering", "answer_template": "The \"a blue , orange , white , and green toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 316, 317, 318, 319, 333, 334, 335, 336, 350, 351, 352, 353, 367, 368, 369, 370, 384, 385, 386], "bbox": [0.5632916666666666, 0.341578125, 0.8509375, 0.993265625], "iscrowd": 0, "area": 45996.1924, "rle": {"size": [640, 480], "counts": "jPY5k4S?6K4K5L4K6K4K5L4K6K4L4K5L5J5L4K5L5J5L5J8I7I7H8I6I8PHoFd4X9m28RH`Fo4f9lJZFR5R:cJoE\\5d:RJ\\El5W;bIjD]6i;QIWDl6`<N2M3N2N2M3O10000000O100000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000001O001N101J5L5K4K5L5K4K6K4K6K4L5J5L5J5L4L5J5L5J5L5K4K6K4L5J5L4K6F9G:E:G:E:G:E:F;F9F:Gb0]OX1iNV1iNW1jNje\\1"}, "label": "a blue , orange , white , and green toothbrush"}]}
{"id": 414588, "image": "COCO_train2014_000000414588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red sweater holding onto a wii controller\"."}], "task": "refering", "answer_template": "The \"a man in a red sweater holding onto a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 89, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 289, 290, 291, 312, 313, 314], "bbox": [0.55971875, 0.05969267139479906, 0.9348437500000001, 0.8907565011820332], "iscrowd": 0, "area": 44972.94784999999, "rle": {"size": [423, 640], "counts": "`Rd41U=2M2O001O0O101N3RHGk2=QMHk2:SMHk2;RMFm2<QMDo2=PMDo2>oLCo2`0TI[Oo27l3>SIGg2KV4?PI5]2\\Oc4?nH?W2SOj4>nHd0T2nNn4>mHi0Q2jNQ5>mHX1b1ZNa5>lHb1Z1QNi5=mHl1P1gMS6>kHT2j0_MY6>lHc2<oLg6?mHf3R7[LmHg3Q7ZLnHg3Q7ZLnHg3R7YLmHi3Q7XLoHh3P7YLoHh3Q7X1N100O2O00000O1000000000000000000000O2O00001O001O001O001N101O00001O001O00000O100000O^NoHPMQ7P3VIiLj6[3YI^Lg6e3\\IULd6k3bInK_6S4bIkK^6U4dIiK\\6W4gIfKY6[4Y100001O00O100O1O1mNS1000gLcLgL^3X3dLfL]3Z3cLeL^3Z3dLcL^3R3nLkLT3R3PMjLT3S3oLkLR3T3]M\\Le2b3]M]Ld2a3_M]Lb2b3`M[Lb2b3bM[La2b3aM\\La2b3bM[L`2c3cMZL_2d3cMZL`2_3gM^L[2a3fM\\L]2c3eMYL_2f3bMWL`2i3`MRLe2n3[MnKj2Q4WMlKk2T4UMdKS3[4nLdKS3\\4mLcKU3\\4lLbKU3^4kLaKV3_4jL`KX3^4iLaKX3_4iL_KX3a4hL_KX3`4mL[KU3d4g200O1000000O1000000O10001N10000O10000O1001O010O00001O001O001O001O001O001O1O001O001O1O001O001O006jHSKf5S5QJQKn5U5iIoJV6R6O1O1O1O1O001O00001O001O001N2O1O1O1O1N2O1O2N1N2O1O1O1N2O2M2N2O1N2O1N2N3N1N2O1N2N2O3L4M3L4L4M1N3N1N2N3K4L4G=F<F9G:F9H8mNPE@Z;2W1CSUa0"}, "label": "a man in a red sweater holding onto a wii controller"}]}
{"id": 414588, "image": "COCO_train2014_000000414588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in red shirt , sitting on black stool\"."}], "task": "refering", "answer_template": "The \"the man in red shirt , sitting on black stool\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 89, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 289, 290, 291, 312, 313, 314], "bbox": [0.55971875, 0.05969267139479906, 0.9348437500000001, 0.8907565011820332], "iscrowd": 0, "area": 44972.94784999999, "rle": {"size": [423, 640], "counts": "`Rd41U=2M2O001O0O101N3RHGk2=QMHk2:SMHk2;RMFm2<QMDo2=PMDo2>oLCo2`0TI[Oo27l3>SIGg2KV4?PI5]2\\Oc4?nH?W2SOj4>nHd0T2nNn4>mHi0Q2jNQ5>mHX1b1ZNa5>lHb1Z1QNi5=mHl1P1gMS6>kHT2j0_MY6>lHc2<oLg6?mHf3R7[LmHg3Q7ZLnHg3Q7ZLnHg3R7YLmHi3Q7XLoHh3P7YLoHh3Q7X1N100O2O00000O1000000000000000000000O2O00001O001O001O001N101O00001O001O00000O100000O^NoHPMQ7P3VIiLj6[3YI^Lg6e3\\IULd6k3bInK_6S4bIkK^6U4dIiK\\6W4gIfKY6[4Y100001O00O100O1O1mNS1000gLcLgL^3X3dLfL]3Z3cLeL^3Z3dLcL^3R3nLkLT3R3PMjLT3S3oLkLR3T3]M\\Le2b3]M]Ld2a3_M]Lb2b3`M[Lb2b3bM[La2b3aM\\La2b3bM[L`2c3cMZL_2d3cMZL`2_3gM^L[2a3fM\\L]2c3eMYL_2f3bMWL`2i3`MRLe2n3[MnKj2Q4WMlKk2T4UMdKS3[4nLdKS3\\4mLcKU3\\4lLbKU3^4kLaKV3_4jL`KX3^4iLaKX3_4iL_KX3a4hL_KX3`4mL[KU3d4g200O1000000O1000000O10001N10000O10000O1001O010O00001O001O001O001O001O001O1O001O001O1O001O001O006jHSKf5S5QJQKn5U5iIoJV6R6O1O1O1O1O001O00001O001O001N2O1O1O1O1N2O1O2N1N2O1O1O1N2O2M2N2O1N2O1N2N3N1N2O1N2N2O3L4M3L4L4M1N3N1N2N3K4L4G=F<F9G:F9H8mNPE@Z;2W1CSUa0"}, "label": "the man in red shirt , sitting on black stool"}]}
{"id": 414588, "image": "COCO_train2014_000000414588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bean bag the man wearing khakis is sitting on\"."}], "task": "refering", "answer_template": "The \"the bean bag the man wearing khakis is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 339, 340, 341], "bbox": [0.698015625, 0.621678486997636, 0.9532812500000001, 0.9479669030732861], "iscrowd": 0, "area": 16120.754149999999, "rle": {"size": [423, 640], "counts": "VPi5>U<e0[Od0\\Oe0A>0001O0010O01O0010O0001O0010O01O001O01O01O0010O01O001O01O01O001O010O001O01O01O001O010O001O0001O0000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000O10O01O100O100O1O100O100O10O01O100O100O1O100O100O00100O100O100O1O1O1O1N100O10O0100O100O010001O2N1N2O2N1O1O1N3N1O1O1O2N2M4M3H8^N`U<"}, "label": "the bean bag the man wearing khakis is sitting on"}]}
{"id": 414588, "image": "COCO_train2014_000000414588.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cushion being sat on by a man in a dark red shirt\"."}], "task": "refering", "answer_template": "The \"a black cushion being sat on by a man in a dark red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 339, 340, 341], "bbox": [0.698015625, 0.621678486997636, 0.9532812500000001, 0.9479669030732861], "iscrowd": 0, "area": 16120.754149999999, "rle": {"size": [423, 640], "counts": "VPi5>U<e0[Od0\\Oe0A>0001O0010O01O0010O0001O0010O01O001O01O01O0010O01O001O01O01O001O010O001O01O01O001O010O001O0001O0000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000O10O01O100O100O1O100O100O10O01O100O100O1O100O100O00100O100O100O1O1O1O1N100O10O0100O100O010001O2N1N2O2N1O1O1N3N1O1O1O2N2M4M3H8^N`U<"}, "label": "a black cushion being sat on by a man in a dark red shirt"}]}
{"id": 209791, "image": "COCO_train2014_000000209791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the entire backside of the vehicle\"."}], "task": "refering", "answer_template": "The \"the entire backside of the vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.00671875, 0.018104166666666664, 0.75671875, 0.9867916666666667], "iscrowd": 0, "area": 203595.79015, "rle": {"size": [480, 640], "counts": "^Z28c0Ii:o2WERMP8f5PH[JX5\\8hJfG^2R;bMoDFh=:?00000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000001O00000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O0000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000001O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2O0O2N1O2N1O2N1O2N1O2N1O2M2O2N1N3N1N3N1O2M2O1N3N1O2M2O2N1N3N1N3N1O2M2O2M2O2N1N3N1O2M2O2M2O1O2M2O2M2O2N1N3N1N3N1O2M4M5K6I6K5J6K6J5J6K5J6K6J5J6K5J7J5K5J6K6J5J6K5J7J?Ac0\\Od0]Oc0\\Od0]Od0\\Oc0\\Od0]OffX2"}, "label": "the entire backside of the vehicle"}]}
{"id": 209791, "image": "COCO_train2014_000000209791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surf board is shown in the back of the truck\"."}], "task": "refering", "answer_template": "The \"a surf board is shown in the back of the truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.00671875, 0.018104166666666664, 0.75671875, 0.9867916666666667], "iscrowd": 0, "area": 203595.79015, "rle": {"size": [480, 640], "counts": "^Z28c0Ii:o2WERMP8f5PH[JX5\\8hJfG^2R;bMoDFh=:?00000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000001O00000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O0000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000001O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2O0O2N1O2N1O2N1O2N1O2N1O2M2O2N1N3N1N3N1O2M2O1N3N1O2M2O2N1N3N1N3N1O2M2O2M2O2N1N3N1O2M2O2M2O1O2M2O2M2O2N1N3N1N3N1O2M4M5K6I6K5J6K6J5J6K5J6K6J5J6K5J7J5K5J6K6J5J6K5J7J?Ac0\\Od0]Oc0\\Od0]Od0\\Oc0\\Od0]OffX2"}, "label": "a surf board is shown in the back of the truck"}]}
{"id": 4993, "image": "COCO_train2014_000000004993.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black hat and dark jacket holding two cellphones\"."}], "task": "refering", "answer_template": "The \"a man with a black hat and dark jacket holding two cellphones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406], "bbox": [0.149171875, 0.1325910931174089, 0.76840625, 0.9887651821862348], "iscrowd": 0, "area": 106947.8128, "rle": {"size": [494, 640], "counts": "]U^12Z?d0\\O6J6I6K5K6J5K5J7J5K5K6J4K4M4L3L5L4L3L5L3M4K4M4L3M3M4M2M3M4M1N3M3N1N3M3N1N3N1N3M3N1N3M3N1N2N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1O1O10000O100O100O100O100O1000O0100O100O100O100O10000O100O1UImH[5S7cJoH\\5R7aJQI_5o6^JTIa5l6^JVIb5i6\\JZIc5g6ZJ\\If5c6YJ_If5a6XJbIh5]6VJfIj5Y6UJiIj5W6TJlIl5U6PJnIe5^6XJdIf5_6VJdIh5`6SJcIk5`6RJbIl5Q8N2N2oMTG^Mn8n1jGmMW8l1THoMm7P2XHlMh7S2]HiMc7V2aHfM`7Y2eHcM[7]2iH_MW7`2mH\\MT7c2PIZMP7e2TIXMl6g2n200O1O1O1O100O1O1O1O100O1O1O1O100O1O10000O10000000000O1000000000000O100000000000000001O0000001O00001O0000001O00001O0000001O00001O00001O001O1O001O1O001O1O001O1O1O001O1O2N3M2N2N3M2N3M2N2N3M2N4L=C5K4L5K8H2N1O1O1O1O1O1O1O1O1O00001O001O001O0000001O00001O00001O00001O00001O001O1O1O1O001O1O1O1O1O1O1O2N001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O001O00001O1O001O1O1O1O1O1O1O1O1O1O1hM]IcKc6]4mIQKV6m4ZJ\\Jo5c5o1N3N1O2M2O2M2O2N2M2O2M2O2M2O2N1N3N1N3N2M2O2N1N3N1N3N1N3M3N1N3M2O2M2N3M2O2M2N3N2M2N3M2O2M2N3N1N3M4M2M4L3M4L3L5K4L6J7I8H7H8G9F:G9FPRW2"}, "label": "a man with a black hat and dark jacket holding two cellphones"}]}
{"id": 4993, "image": "COCO_train2014_000000004993.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man holding two cell phones\"."}], "task": "refering", "answer_template": "The \"the man holding two cell phones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406], "bbox": [0.149171875, 0.1325910931174089, 0.76840625, 0.9887651821862348], "iscrowd": 0, "area": 106947.8128, "rle": {"size": [494, 640], "counts": "]U^12Z?d0\\O6J6I6K5K6J5K5J7J5K5K6J4K4M4L3L5L4L3L5L3M4K4M4L3M3M4M2M3M4M1N3M3N1N3M3N1N3N1N3M3N1N3M3N1N2N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1O1O10000O100O100O100O100O1000O0100O100O100O100O10000O100O1UImH[5S7cJoH\\5R7aJQI_5o6^JTIa5l6^JVIb5i6\\JZIc5g6ZJ\\If5c6YJ_If5a6XJbIh5]6VJfIj5Y6UJiIj5W6TJlIl5U6PJnIe5^6XJdIf5_6VJdIh5`6SJcIk5`6RJbIl5Q8N2N2oMTG^Mn8n1jGmMW8l1THoMm7P2XHlMh7S2]HiMc7V2aHfM`7Y2eHcM[7]2iH_MW7`2mH\\MT7c2PIZMP7e2TIXMl6g2n200O1O1O1O100O1O1O1O100O1O1O1O100O1O10000O10000000000O1000000000000O100000000000000001O0000001O00001O0000001O00001O0000001O00001O00001O001O1O001O1O001O1O001O1O1O001O1O2N3M2N2N3M2N3M2N2N3M2N4L=C5K4L5K8H2N1O1O1O1O1O1O1O1O1O00001O001O001O0000001O00001O00001O00001O00001O001O1O1O1O001O1O1O1O1O1O1O2N001O001O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O001O00001O1O001O1O1O1O1O1O1O1O1O1O1hM]IcKc6]4mIQKV6m4ZJ\\Jo5c5o1N3N1O2M2O2M2O2N2M2O2M2O2M2O2N1N3N1N3N2M2O2N1N3N1N3N1N3M3N1N3M2O2M2N3M2O2M2N3N2M2N3M2O2M2N3N1N3M4M2M4L3M4L3L5K4L6J7I8H7H8G9F:G9FPRW2"}, "label": "the man holding two cell phones"}]}
{"id": 455554, "image": "COCO_train2014_000000455554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white polka dot womens tie untied\"."}], "task": "refering", "answer_template": "The \"black and white polka dot womens tie untied\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 152, 153, 163, 173, 174, 184, 185], "bbox": [0.7673076923076924, 0.7444791666666667, 0.9834935897435898, 0.979125], "iscrowd": 0, "area": 3240.1116499999994, "rle": {"size": [480, 312], "counts": "hm`34j>4L4L4L4L3M4N1N4L3M2N2O2M2N2N3M2O0]O_NfCb1W<eNdC\\1[<jN_CW1`<e00O00010O00O1O1L4I8I6O1O1O2N100001O10O`NaD@`;>fD]OZ;c0kDWOX;f0mDTOV;j0WEhNl:W1_1N2N2O2M2O1N2O1O1O2N1O1O1O1O1N3N1O1O1O3M5Kbo1"}, "label": "black and white polka dot womens tie untied"}]}
{"id": 455554, "image": "COCO_train2014_000000455554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black bow tie with white polka dots\"."}], "task": "refering", "answer_template": "The \"black bow tie with white polka dots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 152, 153, 163, 173, 174, 184, 185], "bbox": [0.7673076923076924, 0.7444791666666667, 0.9834935897435898, 0.979125], "iscrowd": 0, "area": 3240.1116499999994, "rle": {"size": [480, 312], "counts": "hm`34j>4L4L4L4L3M4N1N4L3M2N2O2M2N2N3M2O0]O_NfCb1W<eNdC\\1[<jN_CW1`<e00O00010O00O1O1L4I8I6O1O1O2N100001O10O`NaD@`;>fD]OZ;c0kDWOX;f0mDTOV;j0WEhNl:W1_1N2N2O2M2O1N2O1O1O2N1O1O1O1O1N3N1O1O1O3M5Kbo1"}, "label": "black bow tie with white polka dots"}]}
{"id": 111490, "image": "COCO_train2014_000000111490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy on a base\"."}], "task": "refering", "answer_template": "The \"a boy on a base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 245, 265, 266, 267, 268], "bbox": [0.494921875, 0.06810304449648712, 0.69340625, 0.7649180327868853], "iscrowd": 0, "area": 19705.279250000003, "rle": {"size": [427, 640], "counts": "g_T41Y=2O0O100O10000oGLj26UMJl25TMLk24TMMh27XMIg28YMHg28YMIf27YMJg27XMIg28YMH_M1b27NJPM?Q3GM1]Lh0f3WOJ;SLe0n0hNR19k1=QLe0P1iNR15k1a0nKd0l0`NbN:g23i1c0nKc0o0POX1Jj1f0jKc0S1oNW1Il1f0gKd0T1oNY1Gk1i0dKc0W1lN[1Ij1i0bKb0Y1lNjNc0NeNm0a0_3j2dLfLQOh0Om0i0kNb3k2cLBIeMc3i2bLEJbMc3j2cL3ZOTMS4i2bLS1]3nNiKl1W4d3O3M4M1O1O000000000000000000001O00`NgLnHY3o6kLPIU3S6YLgIe05R3i5PMdI3`0n2g5\\NWJe1f5`NWJa1g5cNVJ^1h5eNVJ[1i5iNUJW1i5mNSJU1j5X3N1YK`Hb3;^Kh6n0nHc33iKn6a0RIY4m6eKUI\\4k6`KXIa4g6YK^Ii4b6oJeIR5Z6iJkIX5T6eJoI\\5o62O1O1O1N2O1O1O1O1fI[Jk4f5RK\\Jn4e5oJ\\JR5e5kJ]JU5d5hJ]JZ5d5aJ_J_5`5`JaJa5_5]JbJd5_5YJcJg5b3iIgM:jNn5\\3lIlMNlNW6V3nIiNS6U1QJhNP6U1TJiNm5U1VJiNk5T1ZJiNg5U1\\JjNe5S1^JkNc5R1aJlNa5Q1cJlN^5Q1fJmN[5P1jJmNX5o0PKkNQ5R1XKgNj4U1bKbN^4\\1gK`NZ4]1lK_NV4^1mK`NT4]1PLaNQ4]1RL`NP4]1TLaNm3\\1ULdNl3Z1TLgNm3V1ULiNm3U1[LdNf3Y1aLbN`3\\1eL_N]3^1[4M4K5K5L4K5L4K8I<CYba2"}, "label": "a boy on a base"}]}
{"id": 111490, "image": "COCO_train2014_000000111490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a blue shirt a grey pants\"."}], "task": "refering", "answer_template": "The \"a person with a blue shirt a grey pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 245, 265, 266, 267, 268], "bbox": [0.494921875, 0.06810304449648712, 0.69340625, 0.7649180327868853], "iscrowd": 0, "area": 19705.279250000003, "rle": {"size": [427, 640], "counts": "g_T41Y=2O0O100O10000oGLj26UMJl25TMLk24TMMh27XMIg28YMHg28YMIf27YMJg27XMIg28YMH_M1b27NJPM?Q3GM1]Lh0f3WOJ;SLe0n0hNR19k1=QLe0P1iNR15k1a0nKd0l0`NbN:g23i1c0nKc0o0POX1Jj1f0jKc0S1oNW1Il1f0gKd0T1oNY1Gk1i0dKc0W1lN[1Ij1i0bKb0Y1lNjNc0NeNm0a0_3j2dLfLQOh0Om0i0kNb3k2cLBIeMc3i2bLEJbMc3j2cL3ZOTMS4i2bLS1]3nNiKl1W4d3O3M4M1O1O000000000000000000001O00`NgLnHY3o6kLPIU3S6YLgIe05R3i5PMdI3`0n2g5\\NWJe1f5`NWJa1g5cNVJ^1h5eNVJ[1i5iNUJW1i5mNSJU1j5X3N1YK`Hb3;^Kh6n0nHc33iKn6a0RIY4m6eKUI\\4k6`KXIa4g6YK^Ii4b6oJeIR5Z6iJkIX5T6eJoI\\5o62O1O1O1N2O1O1O1O1fI[Jk4f5RK\\Jn4e5oJ\\JR5e5kJ]JU5d5hJ]JZ5d5aJ_J_5`5`JaJa5_5]JbJd5_5YJcJg5b3iIgM:jNn5\\3lIlMNlNW6V3nIiNS6U1QJhNP6U1TJiNm5U1VJiNk5T1ZJiNg5U1\\JjNe5S1^JkNc5R1aJlNa5Q1cJlN^5Q1fJmN[5P1jJmNX5o0PKkNQ5R1XKgNj4U1bKbN^4\\1gK`NZ4]1lK_NV4^1mK`NT4]1PLaNQ4]1RL`NP4]1TLaNm3\\1ULdNl3Z1TLgNm3V1ULiNm3U1[LdNf3Y1aLbN`3\\1eL_N]3^1[4M4K5K5L4K5L4K8I<CYba2"}, "label": "a person with a blue shirt a grey pants"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a drinking glass with a knife resting on it\"."}], "task": "refering", "answer_template": "The \"a drinking glass with a knife resting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 351, 352, 353, 354], "bbox": [0.214515625, 0.5621875000000001, 0.481796875, 0.9091041666666667], "iscrowd": 0, "area": 18174.4177, "rle": {"size": [480, 640], "counts": "abP25[>b0^O`0B>N2N1O2N2N1O2N2N1O2N2N1O2N2N110O1]DeM`:[2`EhM^:X2bEiM\\:X2cEkMZ:U2gEmMW:S2iEoMT:R2kEQNR:o1oERNo9o1QFSNm9m1RFUNl9k1UFVNi9k1WFUNi9k1VFWNh9i1YFWNf9j1YFXNf9h1ZFXNe9h1\\FYNb9h1]FYNc9g1]FZNa9f1`FZN_9g1`F[N_9e1aF[N^9e1cF\\N[9e1dF\\N\\9d1dF]NZ9c1fF^NY9c1gF]NY9c1gF^NW9b1iF_NV9b1jF_NU9a1kF_NT9a1lFaNS9_1mFaNS9_1mFbNQ9^1oFcNQ9]1oFdNP9\\1oFeNQ9Z1PGgNo8Y1QGgNo8Y1PGiNo8V1RGjNn8V1RGjNm8W1RGkNm8T1TGlNl8T1TGlNl8T1SGmNm8R1TGnNl8R1SGoNm8Q1SGoNm8P1TGPOk8Q1TGPOl8P1TGPOl8o0UGQOk8o0TGROl8n0TGROl8m0UGSOk8m0TGTOl8l0TGTOl8k0TGVOl8j0TGVOl8j0TGVOl8i0TGXOl8h0TGXOl8h0TGXOl8g0TGZOm8e0SG[Om8e0SG[Om8d0SG]Om8c0SG]Om8c0RG^On8a0SG_Om8a0SG_Om8a0RG@n8?SGAm8?SGAm8?RGBn8=SGCm8=SGCn8<QGEo8:RGFn8:QGGo89QGGo88RGHn88QGHP98PGHP97QGIo87PGJP96PGJQ94PGLP94oFMR92nFNR91nF0R90nF0S9OmF1S9NmF3S9MmF3T9LlF4T9KlF6T9JlF5V9JjF6V9IjF7W9IiF6Y9IfF8Z9GgF8Z9HfF7\\9HcF8^9GcF9]9GcF8_9G`F9a9F`F:`9F`F9b9F]F:d9E]F;c9E\\F;f9DZF;g9DZF<f9DZF;h9DWF<j9CWF=i9CWF<k9BVF=l9BTF>l9BTF=n9ASF>o9APF?R:@nE`0R:_OoE`0S:_OmE`0U:_OkEa0V:]OkEb0W:]OiEb0X:^OgEc0Z:[OgEd0[:[OeEc0^:\\ObEb0a:\\O`Ec0a:]O_Ea0d:^O\\Ea0g:\\OZEb0j:\\OVEb0o:[OPEe0S;XOnDf0W;WOiDh0Z;VOfDh0^;UOcDi0b;TO]Dk0g;SOYDj0P=M3N1N3M3MPYk4"}, "label": "a drinking glass with a knife resting on it"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear glass with knife on it\"."}], "task": "refering", "answer_template": "The \"clear glass with knife on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 351, 352, 353, 354], "bbox": [0.214515625, 0.5621875000000001, 0.481796875, 0.9091041666666667], "iscrowd": 0, "area": 18174.4177, "rle": {"size": [480, 640], "counts": "abP25[>b0^O`0B>N2N1O2N2N1O2N2N1O2N2N1O2N2N110O1]DeM`:[2`EhM^:X2bEiM\\:X2cEkMZ:U2gEmMW:S2iEoMT:R2kEQNR:o1oERNo9o1QFSNm9m1RFUNl9k1UFVNi9k1WFUNi9k1VFWNh9i1YFWNf9j1YFXNf9h1ZFXNe9h1\\FYNb9h1]FYNc9g1]FZNa9f1`FZN_9g1`F[N_9e1aF[N^9e1cF\\N[9e1dF\\N\\9d1dF]NZ9c1fF^NY9c1gF]NY9c1gF^NW9b1iF_NV9b1jF_NU9a1kF_NT9a1lFaNS9_1mFaNS9_1mFbNQ9^1oFcNQ9]1oFdNP9\\1oFeNQ9Z1PGgNo8Y1QGgNo8Y1PGiNo8V1RGjNn8V1RGjNm8W1RGkNm8T1TGlNl8T1TGlNl8T1SGmNm8R1TGnNl8R1SGoNm8Q1SGoNm8P1TGPOk8Q1TGPOl8P1TGPOl8o0UGQOk8o0TGROl8n0TGROl8m0UGSOk8m0TGTOl8l0TGTOl8k0TGVOl8j0TGVOl8j0TGVOl8i0TGXOl8h0TGXOl8h0TGXOl8g0TGZOm8e0SG[Om8e0SG[Om8d0SG]Om8c0SG]Om8c0RG^On8a0SG_Om8a0SG_Om8a0RG@n8?SGAm8?SGAm8?RGBn8=SGCm8=SGCn8<QGEo8:RGFn8:QGGo89QGGo88RGHn88QGHP98PGHP97QGIo87PGJP96PGJQ94PGLP94oFMR92nFNR91nF0R90nF0S9OmF1S9NmF3S9MmF3T9LlF4T9KlF6T9JlF5V9JjF6V9IjF7W9IiF6Y9IfF8Z9GgF8Z9HfF7\\9HcF8^9GcF9]9GcF8_9G`F9a9F`F:`9F`F9b9F]F:d9E]F;c9E\\F;f9DZF;g9DZF<f9DZF;h9DWF<j9CWF=i9CWF<k9BVF=l9BTF>l9BTF=n9ASF>o9APF?R:@nE`0R:_OoE`0S:_OmE`0U:_OkEa0V:]OkEb0W:]OiEb0X:^OgEc0Z:[OgEd0[:[OeEc0^:\\ObEb0a:\\O`Ec0a:]O_Ea0d:^O\\Ea0g:\\OZEb0j:\\OVEb0o:[OPEe0S;XOnDf0W;WOiDh0Z;VOfDh0^;UOcDi0b;TO]Dk0g;SOYDj0P=M3N1N3M3MPYk4"}, "label": "clear glass with knife on it"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup with a brown rim inside of it\"."}], "task": "refering", "answer_template": "The \"a white cup with a brown rim inside of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 225, 226, 248, 249], "bbox": [0.680296875, 0.2506041666666667, 0.9243593749999999, 0.602625], "iscrowd": 0, "area": 18250.096899999997, "rle": {"size": [480, 640], "counts": "_S\\67i=Q1N3L3N2M4M2M4M2M3N3M2M4M2M3N3L3N2M3N2M3N2N2M3O1N2O0O2O000O10001N100O10001N10000O101N10000O2O0O10000O2O000O100O2O000O10001N100O10001N100O013M7H8I6J4L4K2O001N101O1O0O10001N10000000000000000001O000000000O100000000N3N1N2N2O2M2N2WOlE`LZ:`3d0N10000O2O0O2O0O2O001N101N101N2O1O0O2O1N2N2M3N2M3N2N2M2O2M3N2M3N2M4M3M3L3L5L6J:C=]Od0]ObZf0"}, "label": "a white cup with a brown rim inside of it"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty white coffee cup , with coffee rings inside\"."}], "task": "refering", "answer_template": "The \"an empty white coffee cup , with coffee rings inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 225, 226, 248, 249], "bbox": [0.680296875, 0.2506041666666667, 0.9243593749999999, 0.602625], "iscrowd": 0, "area": 18250.096899999997, "rle": {"size": [480, 640], "counts": "_S\\67i=Q1N3L3N2M4M2M4M2M3N3M2M4M2M3N3L3N2M3N2M3N2N2M3O1N2O0O2O000O10001N100O10001N10000O101N10000O2O0O10000O2O000O100O2O000O10001N100O10001N100O013M7H8I6J4L4K2O001N101O1O0O10001N10000000000000000001O000000000O100000000N3N1N2N2O2M2N2WOlE`LZ:`3d0N10000O2O0O2O0O2O001N101N101N2O1O0O2O1N2N2M3N2M3N2N2M2O2M3N2M3N2M4M3M3L3L5L6J:C=]Od0]ObZf0"}, "label": "an empty white coffee cup , with coffee rings inside"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a knife under the lid\"."}], "task": "refering", "answer_template": "The \"a knife under the lid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 276, 277, 278, 279, 301, 302, 303, 327, 353, 354, 355], "bbox": [0.0050625, 0.6853958333333333, 0.45337500000000003, 0.9348333333333334], "iscrowd": 0, "area": 2857.4415500000027, "rle": {"size": [480, 640], "counts": "hg11m>3L3N3L3M3N3O010O0010O0010O0010O0010O0010O010O00010O010O01O01O010O01O010O01O01O010O010O00010O010O00010O010O0010O0010O0010O010O00010O010O01O01O010O01O01O010O010O000010O010O010O010O0010O010O0100O010O0010O010O001O0O2O0O101N101N101N101O0O2O0O101N101N101O0OlS60UlI1O001O001O001O01O100O100O010O100O10iUS10VjlN0O1000001O0O100000001O0O100000001N1000000O2N1O1O1O1O1O2N1O11O00010O00001O0010O00O2O0O101N101N100O2O000M_dS5"}, "label": "a knife under the lid"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver butter knife\"."}], "task": "refering", "answer_template": "The \"silver butter knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 325, 326, 327], "bbox": [0.13314062499999998, 0.5595416666666666, 0.746625, 0.8763958333333333], "iscrowd": 0, "area": 11878.615299999998, "rle": {"size": [480, 640], "counts": "[WX1`0_>NbAF\\>;fACY>>hABU>a0kA^OT>c041N2O1O2M2O1O0O1000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O1000O01000O01000O01000O01000O01000O01000O01000O01000O0100O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10000O10000O10000O10000O10000O10000O1000O01000O01000O01000O01000O01000O010O10O10O10O10O1000O01000O01000O01000O0100O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O01000O01000O01000O01000O01000O01000O0100O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O10O10O01000O01000O01000O01000O01000O01000O0100O01000O010O1O001O1O1O001O1O1O1O1O1O010O1O1O1O1O2L4M4K4M`e[2"}, "label": "silver butter knife"}]}
{"id": 373639, "image": "COCO_train2014_000000373639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a butter knife\"."}], "task": "refering", "answer_template": "The \"a butter knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 325, 326, 327], "bbox": [0.13314062499999998, 0.5595416666666666, 0.746625, 0.8763958333333333], "iscrowd": 0, "area": 11878.615299999998, "rle": {"size": [480, 640], "counts": "[WX1`0_>NbAF\\>;fACY>>hABU>a0kA^OT>c041N2O1O2M2O1O0O1000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O1000O01000O01000O01000O01000O01000O01000O01000O01000O0100O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10O10000O10000O10000O10000O10000O10000O1000O01000O01000O01000O01000O01000O010O10O10O10O10O1000O01000O01000O01000O0100O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O01000O01000O01000O01000O01000O01000O0100O01000O01000O01000O01000O01000O01000O010O10O10O10O10O10O10O10O10O10O10O10O10O10O01000O01000O01000O01000O01000O01000O0100O01000O010O1O001O1O1O001O1O1O1O1O1O010O1O1O1O1O2L4M4K4M`e[2"}, "label": "a butter knife"}]}
{"id": 144272, "image": "COCO_train2014_000000144272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a turtleneck and jeans\"."}], "task": "refering", "answer_template": "The \"a woman wearing a turtleneck and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 125, 146, 147, 148, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.09818750000000001, 0.2635416666666667, 0.592546875, 0.983625], "iscrowd": 0, "area": 59243.61979999999, "rle": {"size": [480, 640], "counts": "Pnm0:\\>?F:I5L2M3N2M3N3L3N2M3N2O1O1N2O1N2O1O1N2O1N10001N100O10001N10000O101N10000O2O0O101O0O101N5L4L0O10000O100O100000000O100000000O100000000O100000000O1^OnLVER3g:YMQEg2m:g0M3N2N2O1O1O1N2O1O1O1M3M3M3L4M3K5I7I7I7J6H8\\Od0O1O1N2O1O1O1O1O1N2O001O1M3L4M3M3M3L4O100O10000O100O10000O100O10000O10000O10000O1000000O10000O1000000O1000000000000000000000O10ZNeHgK[7U4kHiKU7S4QIlKn6S4UIkKk6T4XIiKi6V4[IgKe6X4^IfKb6Y4aIeK_6Z4gIaKY6^4U2O1O1O1O100O100O100O1O1O100O10000000000000000000000000000001O0000001O00001O001O1N101O1O001O1O1O2N1O1O1O2N1O1O1O2N1O1O2hGXKT6k4TIkKk6W4mHPLR7S4eHTLZ7n3]HZLb7c5O2N1O1O1O1fLeHmN\\7l0jHTOV7f0PIZOQ7?UIAl68ZIHf64^ILc61_IOb6M`I4a6IaI7_6GcI9^6DdI<]6@fI`0Z6^OhIb0Y6[OiIe0X6WOkIi0V6TOlIl0T6ROnIn0S6nNPJR1Q6kNQJU1o5iNSJW1n5dNVJ\\1k5_NYJa1h5YN]Jg1d5TN_Jl1c5nMbJR2_5iMeJW2\\5cMiJ]2X5^MlJb2U5YMoJg2R5SMSKm2n4nLVKR3\\800[OUDUNk;j1VDVNj;j1WDUNi;k1WDUNi;j1YDUNg;k1YDUNg;k1YDUNg;j1[DUNe;k1\\DTNd;l1\\DTNd;k1^DTNb;l1_DSNa;m1_DSNa;l1aDSN_;m1bDRN_;m1aDSN`;k1bDTN_;k1aDUN`;j1aDUN`;i1bDVN_;i1bDVN_;h1bDXN_;g1bDXN_;f1cDYN^;f1bDZN_;d1cD[N^;d1eDYN\\;e1Q1O1N3N1N2M4L4M2M4M2M4L4M2M4J6K5K]ni3"}, "label": "a woman wearing a turtleneck and jeans"}]}
{"id": 144272, "image": "COCO_train2014_000000144272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a light green turtleneck sweater sitting on the beach at night\"."}], "task": "refering", "answer_template": "The \"a woman in a light green turtleneck sweater sitting on the beach at night\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 125, 146, 147, 148, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.09818750000000001, 0.2635416666666667, 0.592546875, 0.983625], "iscrowd": 0, "area": 59243.61979999999, "rle": {"size": [480, 640], "counts": "Pnm0:\\>?F:I5L2M3N2M3N3L3N2M3N2O1O1N2O1N2O1O1N2O1N10001N100O10001N10000O101N10000O2O0O101O0O101N5L4L0O10000O100O100000000O100000000O100000000O100000000O1^OnLVER3g:YMQEg2m:g0M3N2N2O1O1O1N2O1O1O1M3M3M3L4M3K5I7I7I7J6H8\\Od0O1O1N2O1O1O1O1O1N2O001O1M3L4M3M3M3L4O100O10000O100O10000O100O10000O10000O10000O1000000O10000O1000000O1000000000000000000000O10ZNeHgK[7U4kHiKU7S4QIlKn6S4UIkKk6T4XIiKi6V4[IgKe6X4^IfKb6Y4aIeK_6Z4gIaKY6^4U2O1O1O1O100O100O100O1O1O100O10000000000000000000000000000001O0000001O00001O001O1N101O1O001O1O1O2N1O1O1O2N1O1O1O2N1O1O2hGXKT6k4TIkKk6W4mHPLR7S4eHTLZ7n3]HZLb7c5O2N1O1O1O1fLeHmN\\7l0jHTOV7f0PIZOQ7?UIAl68ZIHf64^ILc61_IOb6M`I4a6IaI7_6GcI9^6DdI<]6@fI`0Z6^OhIb0Y6[OiIe0X6WOkIi0V6TOlIl0T6ROnIn0S6nNPJR1Q6kNQJU1o5iNSJW1n5dNVJ\\1k5_NYJa1h5YN]Jg1d5TN_Jl1c5nMbJR2_5iMeJW2\\5cMiJ]2X5^MlJb2U5YMoJg2R5SMSKm2n4nLVKR3\\800[OUDUNk;j1VDVNj;j1WDUNi;k1WDUNi;j1YDUNg;k1YDUNg;k1YDUNg;j1[DUNe;k1\\DTNd;l1\\DTNd;k1^DTNb;l1_DSNa;m1_DSNa;l1aDSN_;m1bDRN_;m1aDSN`;k1bDTN_;k1aDUN`;j1aDUN`;i1bDVN_;i1bDVN_;h1bDXN_;g1bDXN_;f1cDYN^;f1bDZN_;d1cD[N^;d1eDYN\\;e1Q1O1N3N1N2M4L4M2M4M2M4L4M2M4J6K5K]ni3"}, "label": "a woman in a light green turtleneck sweater sitting on the beach at night"}]}
{"id": 144272, "image": "COCO_train2014_000000144272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a blue shirt and brown pants with a band on his head\"."}], "task": "refering", "answer_template": "The \"man wearing a blue shirt and brown pants with a band on his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 335, 336, 337, 338], "bbox": [0.456734375, 0.16854166666666667, 0.82753125, 0.8853958333333334], "iscrowd": 0, "area": 47393.684750000015, "rle": {"size": [480, 640], "counts": "n`Y45j>3M4K5L3L5L3M4K5L3M4K4M4K2O2N1N3N2N1N3N1N3N2jNTNUEm1_:kNSEX1`2fMX5P4dJULY5n3aJXL]5j3^J\\L^5f3]J`La5b3ZJcLc5_3YJgLc5\\3WJjLg5X3TJmLi5U3RJRMk5P3PJUMm5n2mIXMQ6i2kI]MS6d2gIbMX6^2dIgM[6Z2`IkM^6W2\\IPNb6P2ZIUNe6T5O2N1O1N3N1O1O1O2N1N2O1O1O2N1N2O1O1O2N1O1N2O1O2N1O1O100O2N1O1O1O101N1O1O1O1O2O0O1O1O1O20O1O1O2N1O100O1O2N1N2O1ZH_Ji5c5mIfJR6\\5cImJ^6U5VITKj6n4kH[KU7g4`HbK`7o510O0001O000010O000001O01O01O00001O01O01O00000010O0001O000010O000001O00010O00O1O1O2N1O100O1O1O2N1O1O1O1O1O2N1O100O1O1O2N1O1O1O1O2N1O1O100O1O2N1O1O1O1O1O2N1O1O100`N`G_La8_3gGZL[8d3kGVLV8i3PHQLQ8n3UHlKl7R4\\HgKe7X4aHbK`7]4c1N1N2O1O1O2N1N2O1O1O1N3N1O1O1N2O1O2N1N2O1O1O1N3N1O1O1O1N3N1O1O1N2O1O2N1N2O6J:F:E<E:F;E:E<E:FaYc1"}, "label": "man wearing a blue shirt and brown pants with a band on his head"}]}
{"id": 144272, "image": "COCO_train2014_000000144272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on his cellphone\"."}], "task": "refering", "answer_template": "The \"a man on his cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 335, 336, 337, 338], "bbox": [0.456734375, 0.16854166666666667, 0.82753125, 0.8853958333333334], "iscrowd": 0, "area": 47393.684750000015, "rle": {"size": [480, 640], "counts": "n`Y45j>3M4K5L3L5L3M4K5L3M4K4M4K2O2N1N3N2N1N3N1N3N2jNTNUEm1_:kNSEX1`2fMX5P4dJULY5n3aJXL]5j3^J\\L^5f3]J`La5b3ZJcLc5_3YJgLc5\\3WJjLg5X3TJmLi5U3RJRMk5P3PJUMm5n2mIXMQ6i2kI]MS6d2gIbMX6^2dIgM[6Z2`IkM^6W2\\IPNb6P2ZIUNe6T5O2N1O1N3N1O1O1O2N1N2O1O1O2N1N2O1O1O2N1O1N2O1O2N1O1O100O2N1O1O1O101N1O1O1O1O2O0O1O1O1O20O1O1O2N1O100O1O2N1N2O1ZH_Ji5c5mIfJR6\\5cImJ^6U5VITKj6n4kH[KU7g4`HbK`7o510O0001O000010O000001O01O01O00001O01O01O00000010O0001O000010O000001O00010O00O1O1O2N1O100O1O1O2N1O1O1O1O1O2N1O100O1O1O2N1O1O1O1O2N1O1O100O1O2N1O1O1O1O1O2N1O1O100`N`G_La8_3gGZL[8d3kGVLV8i3PHQLQ8n3UHlKl7R4\\HgKe7X4aHbK`7]4c1N1N2O1O1O2N1N2O1O1O1N3N1O1O1N2O1O2N1N2O1O1O1N3N1O1O1O1N3N1O1O1N2O1O2N1N2O6J:F:E<E:F;E:E<E:FaYc1"}, "label": "a man on his cellphone"}]}
{"id": 37779, "image": "COCO_train2014_000000037779.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a close - up of the front of a motorcycle , including the mirrors\"."}], "task": "refering", "answer_template": "The \"a close - up of the front of a motorcycle , including the mirrors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 336, 341, 342, 343, 346, 347, 353, 354, 355, 356, 357, 358, 359, 363, 364, 365, 366, 369, 370, 371, 376, 377, 378, 379, 380, 381, 382, 383, 385, 386, 387, 388], "bbox": [0.047296875, 0.6790625, 0.93075, 0.989875], "iscrowd": 0, "area": 28346.320099999994, "rle": {"size": [480, 640], "counts": "``>2j>5K5K6L3M3N3L3N1N2O1N2O1000000000000000000000000000000000000001O001O001O00001O001O001O001O001O00001O001O001O002N1O1O2N1O2N1O1O2N1O2N001O001O001O001O001O001O001O001O001O00\\da11b[^N2M3N2N3M2M3N2N2TB\\OX=g0gBYOX=i0fBYOX=i0gBWOX=k0gBVOV=m0iBTOU=o0hBROk;LXET1lNQOi;0VER1POnNi;3TEQ1ROmNh;5SEP1SOlNi;8PEn0VOkNh;:nDX1Q;lNkDV1S;mNjDT1U;POfDS1X;POeDQ1Z;SObDo0\\;TOaDn0=jMa:a3^E`L`:c3^E^La:c3^E^La:d3]E]Lb:d3]E]Lb:e3[E]Ld:d3[E]Ld:e3ZE\\Le:P4O1O1O1O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1^OTEnLm:Q3UEmLl:R3VElLk:S3UEmLl:Q3UEoLl:P3TEPMm:o2TEPMm:=VE`1MSNn:3_Ei1CTNo:HjER2WOVNS<i1mCWNT<h1lCXNU<g1kCYNV<[1UDeNl;Q1dCjNa05l;o0dClN`05m;l0eCnN?6n;i0dCPO?7n;f0dCTO?5n;e0dCUO?6n;b0eCXO=6o;`0eCYO=7o;=fC\\O;7P<;eC^O<7Q<7eCA;8Q<5eCC:8R<2fCE99R<0fCG89S<NbD2_;KfCL79T<IdD6S=0000000000000000000001O00000000000000000000001O0000001O0000001O00001O0000001O00001O00001O000000oj91oTF100O100O100O100O100O1O100O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O100O1000G810000O010O1000O0100O1000O1000000O2O00000O10001O0O10000N2O1O2M2O1N2O2M3N2N2M4M2M3H9G8HTWd0"}, "label": "a close - up of the front of a motorcycle , including the mirrors"}]}
{"id": 37779, "image": "COCO_train2014_000000037779.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dashboard of motorcycle clostest to camera\"."}], "task": "refering", "answer_template": "The \"dashboard of motorcycle clostest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 336, 341, 342, 343, 346, 347, 353, 354, 355, 356, 357, 358, 359, 363, 364, 365, 366, 369, 370, 371, 376, 377, 378, 379, 380, 381, 382, 383, 385, 386, 387, 388], "bbox": [0.047296875, 0.6790625, 0.93075, 0.989875], "iscrowd": 0, "area": 28346.320099999994, "rle": {"size": [480, 640], "counts": "``>2j>5K5K6L3M3N3L3N1N2O1N2O1000000000000000000000000000000000000001O001O001O00001O001O001O001O001O00001O001O001O002N1O1O2N1O2N1O1O2N1O2N001O001O001O001O001O001O001O001O001O00\\da11b[^N2M3N2N3M2M3N2N2TB\\OX=g0gBYOX=i0fBYOX=i0gBWOX=k0gBVOV=m0iBTOU=o0hBROk;LXET1lNQOi;0VER1POnNi;3TEQ1ROmNh;5SEP1SOlNi;8PEn0VOkNh;:nDX1Q;lNkDV1S;mNjDT1U;POfDS1X;POeDQ1Z;SObDo0\\;TOaDn0=jMa:a3^E`L`:c3^E^La:c3^E^La:d3]E]Lb:d3]E]Lb:e3[E]Ld:d3[E]Ld:e3ZE\\Le:P4O1O1O1O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1^OTEnLm:Q3UEmLl:R3VElLk:S3UEmLl:Q3UEoLl:P3TEPMm:o2TEPMm:=VE`1MSNn:3_Ei1CTNo:HjER2WOVNS<i1mCWNT<h1lCXNU<g1kCYNV<[1UDeNl;Q1dCjNa05l;o0dClN`05m;l0eCnN?6n;i0dCPO?7n;f0dCTO?5n;e0dCUO?6n;b0eCXO=6o;`0eCYO=7o;=fC\\O;7P<;eC^O<7Q<7eCA;8Q<5eCC:8R<2fCE99R<0fCG89S<NbD2_;KfCL79T<IdD6S=0000000000000000000001O00000000000000000000001O0000001O0000001O00001O0000001O00001O00001O000000oj91oTF100O100O100O100O100O1O100O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O100O1000G810000O010O1000O0100O1000O1000000O2O00000O10001O0O10000N2O1O2M2O1N2O2M3N2N2M4M2M3H9G8HTWd0"}, "label": "dashboard of motorcycle clostest to camera"}]}
{"id": 406420, "image": "COCO_train2014_000000406420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench holding a man in a red jacket\"."}], "task": "refering", "answer_template": "The \"a wooden bench holding a man in a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 234, 235, 236, 237, 238, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 286, 287, 288, 303, 310], "bbox": [0.172265625, 0.5659251559251559, 0.5759218749999999, 0.7906444906444906], "iscrowd": 0, "area": 11911.75575, "rle": {"size": [481, 640], "counts": "WPd11l>4M4L3O1O2N1O1O2H:Ab0^Oa0B?I7I6J1O1O1O1O1N2O1O1O1O1O1O010000QOo0000000000000000000O100000000000000000O10000000000000000000000000000000000000000000O100000O100000000000000000000000000000000000000000000000000O1000O10000000002N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2M2N2N3M2N2O1N3M2N2N3M2N2N3M2N2N3M2NQo?c0\\P@1N101N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2SCcNU<^1dCiN\\<W1]CPOc<h1N101O00001O001O00001O001O[Oe0G900QORCKn<H_C8a<H`C7`<HbC7^<IcC6]<JdC4]<KeC4[<LeC4[<LfC3Y<MiC2W<NjC1V<OkCOV<0lCLW<4[10000O10000O10000O10000O10000O1000P?0PA0`_o3"}, "label": "a wooden bench holding a man in a red jacket"}]}
{"id": 406420, "image": "COCO_train2014_000000406420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table that a person is sitting on\"."}], "task": "refering", "answer_template": "The \"a wooden table that a person is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 234, 235, 236, 237, 238, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 286, 287, 288, 303, 310], "bbox": [0.172265625, 0.5659251559251559, 0.5759218749999999, 0.7906444906444906], "iscrowd": 0, "area": 11911.75575, "rle": {"size": [481, 640], "counts": "WPd11l>4M4L3O1O2N1O1O2H:Ab0^Oa0B?I7I6J1O1O1O1O1N2O1O1O1O1O1O010000QOo0000000000000000000O100000000000000000O10000000000000000000000000000000000000000000O100000O100000000000000000000000000000000000000000000000000O1000O10000000002N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2M2N2N3M2N2O1N3M2N2N3M2N2N3M2N2N3M2NQo?c0\\P@1N101N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2SCcNU<^1dCiN\\<W1]CPOc<h1N101O00001O001O00001O001O[Oe0G900QORCKn<H_C8a<H`C7`<HbC7^<IcC6]<JdC4]<KeC4[<LeC4[<LfC3Y<MiC2W<NjC1V<OkCOV<0lCLW<4[10000O10000O10000O10000O10000O1000P?0PA0`_o3"}, "label": "a wooden table that a person is sitting on"}]}
{"id": 373653, "image": "COCO_train2014_000000373653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus on the road\"."}], "task": "refering", "answer_template": "The \"a red bus on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354], "bbox": [0.241015625, 0.24045833333333333, 0.72471875, 0.894375], "iscrowd": 0, "area": 88754.42285, "rle": {"size": [480, 640], "counts": "TZX2h0W>X6iI2N2N2N2M3N1O2N2N2M3N2N2N2N3L3N2N2N2N2M4M2N2N2N2M3N3M2N2N2M3N3M2N2N2M3N1O00000O100000000O1000000000000O10O1000000000000000O1000000000000000000O0100000000000000000O1000000000000000000O0100000000000000000O100000000000000000O0100000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000001O00000000001O0O1000001O00000000001O00000000001O00000O101O00000000001O00000000001O000000000O2O00000000001O00000000001O00000000001N100000001O00000000001O00000000001O00000O10001O00000000001O0000000000000000001N102N2N1O2N2N2N1Od0\\Oe0[Oe0[Oe0[Oe0[Of0ZOe0[Oe0[Ob0^O<D;D=D;E<D;E<DTSb2"}, "label": "a red bus on the road"}]}
{"id": 373653, "image": "COCO_train2014_000000373653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red metrobus\"."}], "task": "refering", "answer_template": "The \"a red metrobus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354], "bbox": [0.241015625, 0.24045833333333333, 0.72471875, 0.894375], "iscrowd": 0, "area": 88754.42285, "rle": {"size": [480, 640], "counts": "TZX2h0W>X6iI2N2N2N2M3N1O2N2N2M3N2N2N2N3L3N2N2N2N2M4M2N2N2N2M3N3M2N2N2M3N3M2N2N2M3N1O00000O100000000O1000000000000O10O1000000000000000O1000000000000000000O0100000000000000000O1000000000000000000O0100000000000000000O100000000000000000O0100000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000001O00000000001O0O1000001O00000000001O00000000001O00000O101O00000000001O00000000001O000000000O2O00000000001O00000000001O00000000001N100000001O00000000001O00000000001O00000O10001O00000000001O0000000000000000001N102N2N1O2N2N2N1Od0\\Oe0[Oe0[Oe0[Oe0[Of0ZOe0[Oe0[Ob0^O<D;D=D;E<D;E<DTSb2"}, "label": "a red metrobus"}]}
{"id": 373653, "image": "COCO_train2014_000000373653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bus\"."}], "task": "refering", "answer_template": "The \"the white bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 86, 87, 89, 90, 91, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.52803125, 0.19954166666666667, 1.0, 0.5515625], "iscrowd": 0, "area": 30408.088999999996, "rle": {"size": [480, 640], "counts": "aPo41o>0O101N1O100O100O1O2N1O1O1O1O100O101O000O10000O101O0O1000000O101O0O10000000001O0000001O00001O0000001O0000001O00010O0000001O00001O0000001O00001O0000001O00001O0000001O0000010O00001O0000001O00001O0000001O00001O001O001O1O=Db0]Ob0^Oc0]Oc0]Ob0^O=C1O0000000000000000001O00000001O0000000001O0000000000000000001O00000001O0000000001O00000000000000000000O1000001O000000000000000000000000001O4L3M4L4L3M4I6K6J5XMak4LoTKPO]EU2X:lMmDY3o:>00O1O100O1O100O1N2G9G9F:O1000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000j0VOf2ZMXI"}, "label": "the white bus"}]}
{"id": 373653, "image": "COCO_train2014_000000373653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large white bus driving on the road\"."}], "task": "refering", "answer_template": "The \"a large white bus driving on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 86, 87, 89, 90, 91, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.52803125, 0.19954166666666667, 1.0, 0.5515625], "iscrowd": 0, "area": 30408.088999999996, "rle": {"size": [480, 640], "counts": "aPo41o>0O101N1O100O100O1O2N1O1O1O1O100O101O000O10000O101O0O1000000O101O0O10000000001O0000001O00001O0000001O0000001O00010O0000001O00001O0000001O00001O0000001O00001O0000001O0000010O00001O0000001O00001O0000001O00001O001O001O1O=Db0]Ob0^Oc0]Oc0]Ob0^O=C1O0000000000000000001O00000001O0000000001O0000000000000000001O00000001O0000000001O00000000000000000000O1000001O000000000000000000000000001O4L3M4L4L3M4I6K6J5XMak4LoTKPO]EU2X:lMmDY3o:>00O1O100O1O100O1N2G9G9F:O1000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000j0VOf2ZMXI"}, "label": "a large white bus driving on the road"}]}
{"id": 95124, "image": "COCO_train2014_000000095124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman walking up steps\"."}], "task": "refering", "answer_template": "The \"woman walking up steps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 235, 236, 237, 259, 260, 282, 283], "bbox": [0.229640625, 0.25309133489461355, 0.389828125, 0.8122950819672131], "iscrowd": 0, "area": 13630.096850000004, "rle": {"size": [427, 640], "counts": "^`m13U=7J6J5K3M2M4M3M2N3M3L3XOkNXEX1`:RO[EQ1i0lNV89mFm0e0TOZ84lFk0e0VO\\8j1_G\\N]8f1_G_N^8Z3K4M4L1O1O1O1O1O2N1O1O1ON2M3N2M3M32Nc0]O8I8H7fIPJm4V6jJmIU5[6^JkIa5P701N1O1001O00001O001O001O000I8B=M4M2N4L3N3M2M4QNoHVMGoN\\7g3TISMZ7i2lHeLc7W3\\1M3O0O2N1O1O1O1O1O1O1O1O1O1O1O100O1O9G2O1N2O0O1O100N2N2N2O1_N`E4c:EiE3Y:FRFE`:4\\mR5"}, "label": "woman walking up steps"}]}
{"id": 95124, "image": "COCO_train2014_000000095124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady walking up the steps\"."}], "task": "refering", "answer_template": "The \"a lady walking up the steps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 235, 236, 237, 259, 260, 282, 283], "bbox": [0.229640625, 0.25309133489461355, 0.389828125, 0.8122950819672131], "iscrowd": 0, "area": 13630.096850000004, "rle": {"size": [427, 640], "counts": "^`m13U=7J6J5K3M2M4M3M2N3M3L3XOkNXEX1`:RO[EQ1i0lNV89mFm0e0TOZ84lFk0e0VO\\8j1_G\\N]8f1_G_N^8Z3K4M4L1O1O1O1O1O2N1O1O1ON2M3N2M3M32Nc0]O8I8H7fIPJm4V6jJmIU5[6^JkIa5P701N1O1001O00001O001O001O000I8B=M4M2N4L3N3M2M4QNoHVMGoN\\7g3TISMZ7i2lHeLc7W3\\1M3O0O2N1O1O1O1O1O1O1O1O1O1O1O100O1O9G2O1N2O0O1O100N2N2N2O1_N`E4c:EiE3Y:FRFE`:4\\mR5"}, "label": "a lady walking up the steps"}]}
{"id": 242583, "image": "COCO_train2014_000000242583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man without glasses\"."}], "task": "refering", "answer_template": "The \"a man without glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 34, 35, 36, 37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 312, 313], "bbox": [0.376265625, 0.033278688524590164, 0.8065468750000001, 0.9001405152224823], "iscrowd": 0, "area": 61916.519300000014, "rle": {"size": [427, 640], "counts": "khT31W=3N1N3N2M2N3N1N3N1N3N2M2hLVOoIl0Q6WOiIk0X6XObIk0]6YO^Ih0c6[OWIh0i6[ORIg0m6\\OnHf0S7]OhHe0X7^ObHd0^7@]Hb0c7AWHa0j7BQH`0n7DlG?T8DfG>[8DaG>^8F\\G<e8GVG;j8HPG:P9JkF8U9n1M4D<C<E<D;L5O001N2O001O0O2O001O1N101O001N101O1O0O2O001O001N2O001O0O2O001O1N202O1O1N101O1O1O1O001N2M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3N1N3N2M3N2M2O3L:G9F:G8G5L0O2O1N101N2O0O2O1N101O1N101N2O0O100O10O0100O10O0100O10O2O000O2O00001N10001O0O10001O0O101O000O2O00001N1000001N10001O0O101O001N101O000O2O001N101O000O2O001O0O2O00001N101O001N10001O0O2O001O0O101O001PGXMZ7i2YHnM]7S2UHdNa7m3F:E:G:F:E;F4SM_HHb77^HIc75_HIc75^HKc74^HKc73^HMd71]HMe71\\HOe7O\\H1e7UOUHnN7l1e7TOYHkN3P2f7SO\\HiNOS2g7QO_HgNKX2g7nNdHeNF]2g7lNVIR1l6lNUIT1l6jNVIU1k6iNVIW1l6fNUIY1m6eNUIZ1m6cNTI]1n6`NTI_1n6]NTIb1o6[NSId1o6YNRIg1P7VNQIj1Q7SNQIk1R7RNoHn1T7nMnHQ2T7lMmHT2W9M4M2N3M3L4M2N3L4M3M2N3L4M3M2N3L4M2N3M3L4L3Nl[c1"}, "label": "a man without glasses"}]}
{"id": 242583, "image": "COCO_train2014_000000242583.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing santa hat\"."}], "task": "refering", "answer_template": "The \"man wearing santa hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 34, 35, 36, 37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 312, 313], "bbox": [0.376265625, 0.033278688524590164, 0.8065468750000001, 0.9001405152224823], "iscrowd": 0, "area": 61916.519300000014, "rle": {"size": [427, 640], "counts": "khT31W=3N1N3N2M2N3N1N3N1N3N2M2hLVOoIl0Q6WOiIk0X6XObIk0]6YO^Ih0c6[OWIh0i6[ORIg0m6\\OnHf0S7]OhHe0X7^ObHd0^7@]Hb0c7AWHa0j7BQH`0n7DlG?T8DfG>[8DaG>^8F\\G<e8GVG;j8HPG:P9JkF8U9n1M4D<C<E<D;L5O001N2O001O0O2O001O1N101O001N101O1O0O2O001O001N2O001O0O2O001O1N202O1O1N101O1O1O1O001N2M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3N1N3N2M3N2M2O3L:G9F:G8G5L0O2O1N101N2O0O2O1N101O1N101N2O0O100O10O0100O10O0100O10O2O000O2O00001N10001O0O10001O0O101O000O2O00001N1000001N10001O0O101O001N101O000O2O001N101O000O2O001O0O2O00001N101O001N10001O0O2O001O0O101O001PGXMZ7i2YHnM]7S2UHdNa7m3F:E:G:F:E;F4SM_HHb77^HIc75_HIc75^HKc74^HKc73^HMd71]HMe71\\HOe7O\\H1e7UOUHnN7l1e7TOYHkN3P2f7SO\\HiNOS2g7QO_HgNKX2g7nNdHeNF]2g7lNVIR1l6lNUIT1l6jNVIU1k6iNVIW1l6fNUIY1m6eNUIZ1m6cNTI]1n6`NTI_1n6]NTIb1o6[NSId1o6YNRIg1P7VNQIj1Q7SNQIk1R7RNoHn1T7nMnHQ2T7lMmHT2W9M4M2N3M3L4M2N3L4M3M2N3L4M3M2N3L4M2N3M3L4L3Nl[c1"}, "label": "man wearing santa hat"}]}
{"id": 357272, "image": "COCO_train2014_000000357272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"full size bed with two pillows\"."}], "task": "refering", "answer_template": "The \"full size bed with two pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.597609375, 0.3793958333333334, 0.8713593750000002, 0.5992916666666667], "iscrowd": 0, "area": 14217.423549999994, "rle": {"size": [480, 640], "counts": "oXc52l>c0\\OU1`N>D<00000000000000O10000000000000000000O10000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O1000000000000000002N:F0000001NO2F:O10000000000O10000002N9G0O100000E;000000000000000000000O100000000000000000000000000O10000000000000000000000000000O100000000V1jN001O001O001O001O001O001O001O001O1O001O001O0H9D;F;DRWV1"}, "label": "full size bed with two pillows"}]}
{"id": 357272, "image": "COCO_train2014_000000357272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 beds pushed together in front of a display\"."}], "task": "refering", "answer_template": "The \"2 beds pushed together in front of a display\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.597609375, 0.3793958333333334, 0.8713593750000002, 0.5992916666666667], "iscrowd": 0, "area": 14217.423549999994, "rle": {"size": [480, 640], "counts": "oXc52l>c0\\OU1`N>D<00000000000000O10000000000000000000O10000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O1000000000000000002N:F0000001NO2F:O10000000000O10000002N9G0O100000E;000000000000000000000O100000000000000000000000000O10000000000000000000000000000O100000000V1jN001O001O001O001O001O001O001O001O1O001O001O0H9D;F;DRWV1"}, "label": "2 beds pushed together in front of a display"}]}
{"id": 357272, "image": "COCO_train2014_000000357272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"round bed was in the room\"."}], "task": "refering", "answer_template": "The \"round bed was in the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.126296875, 0.5613125, 0.76359375, 0.8152083333333333], "iscrowd": 0, "area": 35055.93970000001, "rle": {"size": [480, 640], "counts": "]YV13l>3L5L4L3M4K5L4L3M4K5L2N1N3N1O1O1N3N1O1N2O2N1000001N1000001N1000001N1000001O0O10001O0O10001O0O10001O0O10001O0O10001O000O101O000O101O000O101O000O101O000O101O00000O2O00000O2O00000O2O00000O1000000O100000001N1000000O1000000O100000000O10001O000O100000000O100000000O10001O000O100000000O100000000O10001O0000000000000O100000000000001OSMcDW2];fMgDY2Y;eMiD[2W;bMmD]2S;aMPE^2P;_MSEa2m:\\MWEc2i:[MYEe2b;O001O1O001O010O001O001O1O001O001O001O001O1O01O000O1000000O10000O10000O10000O10000O10000O01000O10000O10000O1000O010000O10000O100oN_McEa2^:fMYEZ2h:nMPER2P;oMoDQ2Q;PNnDP2R;RNkDo1U;RNjDn1V;TNhDl1X;n000000O101O00000000000000000O10001O0000000000000O100000001O000000000O1000000000001O0000000O1000000000001O00000O2O000000001O0000001N100000001O0000001O000O101O00000O101N10000O2O0O10001N100O10001N100O101O0O100O2O000O100O2O000O101O0O100O2O000O100O2O000O101N10001N2O0O2O1O0O2O1N101O1N101N2O001N2O0O2O1O0O2O1N101O1N1N3K5L3M4K5L3M4K5L3MW_V2"}, "label": "round bed was in the room"}]}
{"id": 357272, "image": "COCO_train2014_000000357272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a circular checker pattern mattress\"."}], "task": "refering", "answer_template": "The \"a circular checker pattern mattress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.126296875, 0.5613125, 0.76359375, 0.8152083333333333], "iscrowd": 0, "area": 35055.93970000001, "rle": {"size": [480, 640], "counts": "]YV13l>3L5L4L3M4K5L4L3M4K5L2N1N3N1O1O1N3N1O1N2O2N1000001N1000001N1000001N1000001O0O10001O0O10001O0O10001O0O10001O0O10001O000O101O000O101O000O101O000O101O000O101O00000O2O00000O2O00000O2O00000O1000000O100000001N1000000O1000000O100000000O10001O000O100000000O100000000O10001O000O100000000O100000000O10001O0000000000000O100000000000001OSMcDW2];fMgDY2Y;eMiD[2W;bMmD]2S;aMPE^2P;_MSEa2m:\\MWEc2i:[MYEe2b;O001O1O001O010O001O001O1O001O001O001O001O1O01O000O1000000O10000O10000O10000O10000O10000O01000O10000O10000O1000O010000O10000O100oN_McEa2^:fMYEZ2h:nMPER2P;oMoDQ2Q;PNnDP2R;RNkDo1U;RNjDn1V;TNhDl1X;n000000O101O00000000000000000O10001O0000000000000O100000001O000000000O1000000000001O0000000O1000000000001O00000O2O000000001O0000001N100000001O0000001O000O101O00000O101N10000O2O0O10001N100O10001N100O101O0O100O2O000O100O2O000O101O0O100O2O000O100O2O000O101N10001N2O0O2O1O0O2O1N101O1N101N2O001N2O0O2O1O0O2O1N101O1N1N3K5L3M4K5L3M4K5L3MW_V2"}, "label": "a circular checker pattern mattress"}]}
{"id": 512920, "image": "COCO_train2014_000000512920.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the main zebra\"."}], "task": "refering", "answer_template": "The \"the main zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 285, 286, 287, 288, 309, 310, 311], "bbox": [0.35071875, 0.24875586854460094, 0.649203125, 0.894507042253521], "iscrowd": 0, "area": 26097.569950000005, "rle": {"size": [426, 640], "counts": "d\\m21S=:E;E;F:G9K5L4K5K5L4K4L5L4K4L5L4K5K5L4K4L5L8GT1mN8G9H1N100O100O100O10O010O010G8D=4K4N2M3M3M3M3M01O00000eNiHcLW7X3SIcLn6W3\\IdLd6V3hIcLY6X3QJcLo5X3[JcLe5Z3bJbL^5\\3gJaLZ5\\3[2N2N2N3M2O20O0001O001O001O00_FFS6h3010O001O0000000001O000000N2@?B?C<D:hNX13M3N2N1O2M3N2N2M2O2mH\\L^4f3XKgLc4[3RKTMh4o2lJ`Mo4b2fJkMU5W2`JUN^5l4N3N1O1N3N1O2N1O1OjNTKQJj4m5^KnIa4P6fKmIX4Q6nKlIQ4R6TLlIk3S6ZLkId3R6cLjI]3S6iLjIW3S6Q2M3L5K4L5M2O2M3N2M2M4L4L4L4K4I8I7WOi0E;F:000000001O1O1O001O1O1O1O001N2O1O1O0000000000SNRF>m9_OYF?g9^O]Fa0c9ZOdFd0\\9XOjFf0X9SOnFl0j:N1O2N2N1N3M2M3M3M3N2M3M3M3N2M3M3M3Nc]m2"}, "label": "the main zebra"}]}
{"id": 512920, "image": "COCO_train2014_000000512920.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing in front of water\"."}], "task": "refering", "answer_template": "The \"zebra standing in front of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 285, 286, 287, 288, 309, 310, 311], "bbox": [0.35071875, 0.24875586854460094, 0.649203125, 0.894507042253521], "iscrowd": 0, "area": 26097.569950000005, "rle": {"size": [426, 640], "counts": "d\\m21S=:E;E;F:G9K5L4K5K5L4K4L5L4K4L5L4K5K5L4K4L5L8GT1mN8G9H1N100O100O100O10O010O010G8D=4K4N2M3M3M3M3M01O00000eNiHcLW7X3SIcLn6W3\\IdLd6V3hIcLY6X3QJcLo5X3[JcLe5Z3bJbL^5\\3gJaLZ5\\3[2N2N2N3M2O20O0001O001O001O00_FFS6h3010O001O0000000001O000000N2@?B?C<D:hNX13M3N2N1O2M3N2N2M2O2mH\\L^4f3XKgLc4[3RKTMh4o2lJ`Mo4b2fJkMU5W2`JUN^5l4N3N1O1N3N1O2N1O1OjNTKQJj4m5^KnIa4P6fKmIX4Q6nKlIQ4R6TLlIk3S6ZLkId3R6cLjI]3S6iLjIW3S6Q2M3L5K4L5M2O2M3N2M2M4L4L4L4K4I8I7WOi0E;F:000000001O1O1O001O1O1O1O001N2O1O1O0000000000SNRF>m9_OYF?g9^O]Fa0c9ZOdFd0\\9XOjFf0X9SOnFl0j:N1O2N2N1N3M2M3M3M3N2M3M3M3N2M3M3M3Nc]m2"}, "label": "zebra standing in front of water"}]}
{"id": 365464, "image": "COCO_train2014_000000365464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zebras looking to their left\"."}], "task": "refering", "answer_template": "The \"two zebras looking to their left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 127, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 247, 248, 265, 270], "bbox": [0.43429687499999997, 0.30492957746478877, 0.83128125, 0.7645774647887325], "iscrowd": 0, "area": 24908.518449999996, "rle": {"size": [426, 640], "counts": "`ic32T=5K5L4K5N2O1N2O1O1O1O1O003M>A`0A<D4L4L4L4L3M000001O01O2N2N2N2N2N2N2N2N2N2O1N2N2O11O1O10O01O1O2N2NO1O1N2O1N2O1N010O01O010O0010O01O010O10O01O010O0015J7J6I6L51L2I1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1N2O1\\Od0000001O00000000000000O1000000000000000000O1000000000000000000O1000O1000000O10000O0100000O1000000O01000O1000000O010O10000O10O10O10000O100O01000O10000O10O10O100O10000O01000O100O1O02O4L4L4M3L5K4L4L4L4L4L5K4L000000O10000000000O100000000O2O001O001O0O2O00001O0O2O001O001O0O2O1O2N2N2M3N2N2N3M2N2iNTFkNo9h0^FVOl;N3N1N3N2M2O2M3N1N3NYd\\1"}, "label": "two zebras looking to their left"}]}
{"id": 365464, "image": "COCO_train2014_000000365464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in on the right , in front of the other zebra and looking at the camera\"."}], "task": "refering", "answer_template": "The \"the zebra in on the right , in front of the other zebra and looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 127, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 247, 248, 265, 270], "bbox": [0.43429687499999997, 0.30492957746478877, 0.83128125, 0.7645774647887325], "iscrowd": 0, "area": 24908.518449999996, "rle": {"size": [426, 640], "counts": "`ic32T=5K5L4K5N2O1N2O1O1O1O1O003M>A`0A<D4L4L4L4L3M000001O01O2N2N2N2N2N2N2N2N2N2O1N2N2O11O1O10O01O1O2N2NO1O1N2O1N2O1N010O01O010O0010O01O010O10O01O010O0015J7J6I6L51L2I1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1N2O1\\Od0000001O00000000000000O1000000000000000000O1000000000000000000O1000O1000000O10000O0100000O1000000O01000O1000000O010O10000O10O10O10000O100O01000O10000O10O10O100O10000O01000O100O1O02O4L4L4M3L5K4L4L4L4L4L5K4L000000O10000000000O100000000O2O001O001O0O2O00001O0O2O001O001O0O2O1O2N2N2M3N2N2N3M2N2iNTFkNo9h0^FVOl;N3N1N3N2M2O2M3N1N3NYd\\1"}, "label": "the zebra in on the right , in front of the other zebra and looking at the camera"}]}
{"id": 471966, "image": "COCO_train2014_000000471966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batter swings at a ball during a baseball game\"."}], "task": "refering", "answer_template": "The \"a batter swings at a ball during a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 333, 334, 335, 336, 337, 355, 356, 359, 360], "bbox": [0.443265625, 0.496625, 0.694375, 0.9550625], "iscrowd": 0, "area": 14338.658049999998, "rle": {"size": [480, 640], "counts": "aaU4a0Y>7O1O1O1O1O1O1O1OM3N2M3N2M3M3N2M3O00100O010OXOJhB6W=KhB5Y=LfB4Y=NfB1Z=1dB0[=2dBMX;OUF7aNJT;8XF0aNHR;a0YFJcNEm:k0\\FDcNAm:S1\\F_OeN^Om:X1[F\\OfN[On:\\1[FAc9b0\\F]Ob9g0]FYOa9j0^FUOa9n0^FRO_9R1`FmN_9W1_FiN_9Z1`FeN^9_1aFaN]9b1bF]N\\9g1cFYN[9j1dFUNZ9o1eFQNX9S2hFkMW9Y2gFgMV9]2iFbMV9a2jF^MS9l2fFTMY9T3aFjL`9\\3ZFdLe9d3UFZLk9^40O010O01O010O0100O10O0O2O0O2N101N1011N100O100O100O100O100fMWF\\Ok9d0[FUOe9k0`FoNa9Q1aFnN_9R1bFmN]9S1eFmNY9S1iFlNW9T1jFkNU9U1mFkNQ9U1QGjNo8V1RGjNl8V1VGiNi8X1XGhNg8W1ZGiNe8XO]FS1P1Ea8WOcFR1m0F`8WOfFP1l0I\\8VOlFn0j0KZ8VOlFP1k0JW8UOPGP1k0JT8VOQGP1l0KR8SOTGR1k0Ko7SOWGQ1l0NP8iNUGZ1l0OT8^NQGb1m03d8L^G3a8N_G1b8N_G1b8N_G1a80_GOb80_GOb80_GOa82_GLc83^GLc84]GKc85^GJc85^GJb87^GHc87^GHc87^GHd87\\GHf86[GHg87ZGHh87XGHj86WGIj87VGHl86UGIl86VGHl87TGHn86TGGn88TGFn88SGGo87SGGn87TGHm87TGHm87UGGl87VGHl86UGHm87TGHm86UGIl86UGIl86UGIm84UGKl84UGKk85VGJk84WGKi85XGIj86WGIi86YGIh86YGIg86S3O1O1N3N^ak2"}, "label": "a batter swings at a ball during a baseball game"}]}
{"id": 471966, "image": "COCO_train2014_000000471966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a bat\"."}], "task": "refering", "answer_template": "The \"a man with a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 333, 334, 335, 336, 337, 355, 356, 359, 360], "bbox": [0.443265625, 0.496625, 0.694375, 0.9550625], "iscrowd": 0, "area": 14338.658049999998, "rle": {"size": [480, 640], "counts": "aaU4a0Y>7O1O1O1O1O1O1O1OM3N2M3N2M3M3N2M3O00100O010OXOJhB6W=KhB5Y=LfB4Y=NfB1Z=1dB0[=2dBMX;OUF7aNJT;8XF0aNHR;a0YFJcNEm:k0\\FDcNAm:S1\\F_OeN^Om:X1[F\\OfN[On:\\1[FAc9b0\\F]Ob9g0]FYOa9j0^FUOa9n0^FRO_9R1`FmN_9W1_FiN_9Z1`FeN^9_1aFaN]9b1bF]N\\9g1cFYN[9j1dFUNZ9o1eFQNX9S2hFkMW9Y2gFgMV9]2iFbMV9a2jF^MS9l2fFTMY9T3aFjL`9\\3ZFdLe9d3UFZLk9^40O010O01O010O0100O10O0O2O0O2N101N1011N100O100O100O100O100fMWF\\Ok9d0[FUOe9k0`FoNa9Q1aFnN_9R1bFmN]9S1eFmNY9S1iFlNW9T1jFkNU9U1mFkNQ9U1QGjNo8V1RGjNl8V1VGiNi8X1XGhNg8W1ZGiNe8XO]FS1P1Ea8WOcFR1m0F`8WOfFP1l0I\\8VOlFn0j0KZ8VOlFP1k0JW8UOPGP1k0JT8VOQGP1l0KR8SOTGR1k0Ko7SOWGQ1l0NP8iNUGZ1l0OT8^NQGb1m03d8L^G3a8N_G1b8N_G1b8N_G1a80_GOb80_GOb80_GOa82_GLc83^GLc84]GKc85^GJc85^GJb87^GHc87^GHc87^GHd87\\GHf86[GHg87ZGHh87XGHj86WGIj87VGHl86UGIl86VGHl87TGHn86TGGn88TGFn88SGGo87SGGn87TGHm87TGHm87UGGl87VGHl86UGHm87TGHm86UGIl86UGIl86UGIm84UGKl84UGKk85VGJk84WGKi85XGIj86WGIi86YGIh86YGIg86S3O1O1N3N^ak2"}, "label": "a man with a bat"}]}
{"id": 471966, "image": "COCO_train2014_000000471966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire\"."}], "task": "refering", "answer_template": "The \"the umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 257, 258, 259, 260, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 348, 349, 350, 351, 371, 372, 373, 374], "bbox": [0.111234375, 0.6091875000000001, 0.324515625, 0.9887708333333334], "iscrowd": 0, "area": 14211.4798, "rle": {"size": [480, 640], "counts": "ceQ12i>7I6J7I7K5L3N1O1O2M2O2N1O1N3N1O2N1N3N1O1VDYNZ:i1cE[NY:g1dE^NY:d1dE`NY:a1eEbNY:_1dEdNZ:_1cEcN\\:^1bEdN]:^1`EdN_:]1_EeN_:^1_EcNa:^1\\EdNc:h1PEZNP;k2O1O100O100O1O100O100O100O1O100O100O1O100O2O0O1O01O001O001O00001N101ROm01O2O1O001]EdL`9]3[FgLe9Y3VFlLj9U3QFPMm9R400000000000000O100000000000000F:I7O1OoMZFjNe9P1cFoN\\9k0lFTOS9g0TGXOj8i0XGVOh8h0[GWOd8h0_GWO`8i0bGVO^8h0eGWOZ8h0iGWOW8h0kGWOT8h0oGWOQ8g0RHXOn7g0THXOk7g0XHXOh7g0ZHXOf7f0]HWOe7g0^HROg74\\F^Oo17k7<WFYOQ23o7e0QFTOS20R8l0mEQOS2LU8U1iEkNU2HZ8]1bEgNm<Y1UCdNl<]1<0O00100O00100O0001O1O3K5K5J6J6KYVZ6"}, "label": "the umpire"}]}
{"id": 471966, "image": "COCO_train2014_000000471966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire is wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"the umpire is wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 257, 258, 259, 260, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 324, 325, 326, 327, 328, 348, 349, 350, 351, 371, 372, 373, 374], "bbox": [0.111234375, 0.6091875000000001, 0.324515625, 0.9887708333333334], "iscrowd": 0, "area": 14211.4798, "rle": {"size": [480, 640], "counts": "ceQ12i>7I6J7I7K5L3N1O1O2M2O2N1O1N3N1O2N1N3N1O1VDYNZ:i1cE[NY:g1dE^NY:d1dE`NY:a1eEbNY:_1dEdNZ:_1cEcN\\:^1bEdN]:^1`EdN_:]1_EeN_:^1_EcNa:^1\\EdNc:h1PEZNP;k2O1O100O100O1O100O100O100O1O100O100O1O100O2O0O1O01O001O001O00001N101ROm01O2O1O001]EdL`9]3[FgLe9Y3VFlLj9U3QFPMm9R400000000000000O100000000000000F:I7O1OoMZFjNe9P1cFoN\\9k0lFTOS9g0TGXOj8i0XGVOh8h0[GWOd8h0_GWO`8i0bGVO^8h0eGWOZ8h0iGWOW8h0kGWOT8h0oGWOQ8g0RHXOn7g0THXOk7g0XHXOh7g0ZHXOf7f0]HWOe7g0^HROg74\\F^Oo17k7<WFYOQ23o7e0QFTOS20R8l0mEQOS2LU8U1iEkNU2HZ8]1bEgNm<Y1UCdNl<]1<0O00100O00100O0001O1O3K5K5J6J6KYVZ6"}, "label": "the umpire is wearing a black shirt"}]}
{"id": 471966, "image": "COCO_train2014_000000471966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher with a umpire behind him and a player in front of him\"."}], "task": "refering", "answer_template": "The \"a catcher with a umpire behind him and a player in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [284, 285, 305, 306, 307, 308, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 374, 376, 377], "bbox": [0.26110937500000003, 0.7308541666666667, 0.52778125, 0.9851875], "iscrowd": 0, "area": 9599.720750000002, "rle": {"size": [480, 640], "counts": "[T_24l>1N3N1N3N1N3N1N2O2M2O2M2O2M2O2M2O2N1N2O2UChNi;Z1kCPOV<R1]CYOb<b110O10O10O10O010O10O10O10O10O010O10O10O10O10O01000O010O01000O0100O0100O0100O011O2M2O2M3L4L3M4K5L4L3L5N2N1O1O001O1O010O1O001O1O0000000001O00000O100UOhDnMX;Q2jDnMW;P2kDoMV;o1mDoMT;o1nDPNS;a0eDBOJ<1Q;`0hDCLJ=2P;?jDCIK?2o:>lDCGKa02m:=oDDDKb03l:<PEEBKe02j:;QEHAIf03i:9REK@Fh05g:4VE1[OEk04e:0ZE7WOCl05d:K^E=9Fa;9`DF`;:aDE_;;cDC];=c101O000000O1000000000000O100O100O100O1O100O100O2N100O100O1O100O100O1000001O01O001O00001O001O00010O002N7I7I7IjT]4"}, "label": "a catcher with a umpire behind him and a player in front of him"}]}
{"id": 471966, "image": "COCO_train2014_000000471966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting to keep the ball\"."}], "task": "refering", "answer_template": "The \"a man sitting to keep the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [284, 285, 305, 306, 307, 308, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 374, 376, 377], "bbox": [0.26110937500000003, 0.7308541666666667, 0.52778125, 0.9851875], "iscrowd": 0, "area": 9599.720750000002, "rle": {"size": [480, 640], "counts": "[T_24l>1N3N1N3N1N3N1N2O2M2O2M2O2M2O2M2O2N1N2O2UChNi;Z1kCPOV<R1]CYOb<b110O10O10O10O010O10O10O10O10O010O10O10O10O10O01000O010O01000O0100O0100O0100O011O2M2O2M3L4L3M4K5L4L3L5N2N1O1O001O1O010O1O001O1O0000000001O00000O100UOhDnMX;Q2jDnMW;P2kDoMV;o1mDoMT;o1nDPNS;a0eDBOJ<1Q;`0hDCLJ=2P;?jDCIK?2o:>lDCGKa02m:=oDDDKb03l:<PEEBKe02j:;QEHAIf03i:9REK@Fh05g:4VE1[OEk04e:0ZE7WOCl05d:K^E=9Fa;9`DF`;:aDE_;;cDC];=c101O000000O1000000000000O100O100O100O1O100O100O2N100O100O1O100O100O1000001O01O001O00001O001O00010O002N7I7I7IjT]4"}, "label": "a man sitting to keep the ball"}]}
{"id": 168865, "image": "COCO_train2014_000000168865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a husband celebrating his wife ' s birthday\"."}], "task": "refering", "answer_template": "The \"a husband celebrating his wife ' s birthday\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 97, 98, 99, 100, 120, 121, 122, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 262, 263, 264, 280, 286], "bbox": [0.060390624999999996, 0.17402083333333335, 0.4850625, 0.7428541666666668], "iscrowd": 0, "area": 34538.578150000016, "rle": {"size": [480, 640], "counts": "kbb04h>5K4K5L5K4L4L5M2O1N3M2N2N3N1N2N3M2N2O2M2N2N3M2O1N3M2N2O2M2N2N3M2O1N3M2N2N3N1N2N3M2O1O2M2O1O2N1O1N3N1O1O2N1N20001N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101N10000O01N101O0O2O1O0O2O001N101N1XO_ERMb:m2fEkL[:T3jEfLW:Z3YOeLYGZ3e8kLXGU3e8QMYGo2d8VMYGk2d8ZMZGf2c8_MZGc2b8bM\\G^2a8gM]GY2`8lM]GV2_8oM_GZ2U8jMiGa2k7cMSH_2j7dMSH]2k7gMSHY2l7jMQHV2o7mMoGS2P8PNnGP2P8TNmGm1R8`201O1N101O1O1O0OROWHiJi7W5\\HdJc7]5aH^J`7b5eHYJ[7g5jHTJV7l5oHoIP7Q6P100000iGRJ[7o5^HXJb7g5ZH]Jg7c5XH]Ji7c5VH^Ji7c5VH]Jk7c5TH^Jl7b5RH_Jo7`5QHaJn7`5QH`JP8`5oGaJQ8V610O0100_OXHiIi7U6b0M3O2O10O010O10O1O0OfGSJd7k5WH[Ji7e5QHaJP8]5mGgJT8X5hGlJX8o50O001O001O10O01O001WLiGg0X8SOmGl0V8mNPHo0U8iNQHU1R8eNSHX1R8`NUH\\1P8]NUHa1n7XNYHd1l7UNYHh1l7QNYHm1k7kM\\HQ2h7iM]HV2f7bMaH]2Q:10O000010O01O000010O01O00010O001O010O00WDbMP;^2kDgMV;Y2dDmM[;S2aDQN`;n1[DWNe;c2O01O010O0010O00010O01O01O03O10000O1000001O0O10000O1M1O100O100O100O100O010O1O1N2M3M3N2M3M5K6J6J6J6J5K6K5J6J6J6J6JT\\j4"}, "label": "a husband celebrating his wife ' s birthday"}]}
{"id": 168865, "image": "COCO_train2014_000000168865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with gray beard in patterned sweater sitting at table with chocolate birthday cake\"."}], "task": "refering", "answer_template": "The \"man with gray beard in patterned sweater sitting at table with chocolate birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 97, 98, 99, 100, 120, 121, 122, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 262, 263, 264, 280, 286], "bbox": [0.060390624999999996, 0.17402083333333335, 0.4850625, 0.7428541666666668], "iscrowd": 0, "area": 34538.578150000016, "rle": {"size": [480, 640], "counts": "kbb04h>5K4K5L5K4L4L5M2O1N3M2N2N3N1N2N3M2N2O2M2N2N3M2O1N3M2N2O2M2N2N3M2O1N3M2N2N3N1N2N3M2O1O2M2O1O2N1O1N3N1O1O2N1N20001N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101N10000O01N101O0O2O1O0O2O001N101N1XO_ERMb:m2fEkL[:T3jEfLW:Z3YOeLYGZ3e8kLXGU3e8QMYGo2d8VMYGk2d8ZMZGf2c8_MZGc2b8bM\\G^2a8gM]GY2`8lM]GV2_8oM_GZ2U8jMiGa2k7cMSH_2j7dMSH]2k7gMSHY2l7jMQHV2o7mMoGS2P8PNnGP2P8TNmGm1R8`201O1N101O1O1O0OROWHiJi7W5\\HdJc7]5aH^J`7b5eHYJ[7g5jHTJV7l5oHoIP7Q6P100000iGRJ[7o5^HXJb7g5ZH]Jg7c5XH]Ji7c5VH^Ji7c5VH]Jk7c5TH^Jl7b5RH_Jo7`5QHaJn7`5QH`JP8`5oGaJQ8V610O0100_OXHiIi7U6b0M3O2O10O010O10O1O0OfGSJd7k5WH[Ji7e5QHaJP8]5mGgJT8X5hGlJX8o50O001O001O10O01O001WLiGg0X8SOmGl0V8mNPHo0U8iNQHU1R8eNSHX1R8`NUH\\1P8]NUHa1n7XNYHd1l7UNYHh1l7QNYHm1k7kM\\HQ2h7iM]HV2f7bMaH]2Q:10O000010O01O000010O01O00010O001O010O00WDbMP;^2kDgMV;Y2dDmM[;S2aDQN`;n1[DWNe;c2O01O010O0010O00010O01O01O03O10000O1000001O0O10000O1M1O100O100O100O100O010O1O1N2M3M3N2M3M5K6J6J6J6J5K6K5J6J6J6J6JT\\j4"}, "label": "man with gray beard in patterned sweater sitting at table with chocolate birthday cake"}]}
{"id": 168865, "image": "COCO_train2014_000000168865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman sitting in front of cake\"."}], "task": "refering", "answer_template": "The \"woman sitting in front of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 361, 363, 364, 365, 366, 367], "bbox": [0.456734375, 0.16591666666666666, 0.991015625, 0.9322083333333333], "iscrowd": 0, "area": 77383.59525, "rle": {"size": [480, 640], "counts": "gWY42^>>N5L3M4L3M4K4M3M4L3M4K4M4L3M4L3L5M2M3M4M2M3M3N3L3O1O2N1O1O1O2N1N2O1O2N1O1O1O2N1N2O1O1O2N1O1O1O1O2M2O1O1O1O2N1O100O1O2N1O1O1O101N1O1O1O1O2N100O1O1O2N1O100O1O2N1O1O1O100O2N1O1O1O101N1O1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O101N1O1O1O1O2N100O1O101N1O100O101N100fG_IT8b6kG`IT8`6kGbIS8g6O1N2N2O1N2O1N2O1N100O2O000O101O000O2O000O2O000O2O0001O000O10001O000000000O2O000000001O000O1000TOl0L4L4L4L4K5L4L4L4M3M3N2N2N2N2M3N2N2N2N2M3N2N2N2N2N200000000O100000001O00000O2O0000001O000O2O0000001O000O101O0000001O000O10001O0000000O2O000000001O000O101O000000001N100000001O000000010O0001O0010O0001O00010O0kHWLS3j3iL[LU3e3hLaLcLDU6k3UMfL`LD[6f3QMUMo2k2nLYMQ3h2kL\\MT3d2hL`MX3`2eLdM[3\\2aLhM^3X2_LkMa3U2\\LoMc3R2ZLPNf3P2XLSNh3m1VLUNi3m1SLUNm3l1PLWNo3k1mKXNS4h1kKYNU4i1gKZNX4g1eK\\NZ4f1bK\\N^4e1`K]N`4d1\\K^Nd4c1YK`Nf4a1WKbNh4`1UKaNk4`1RKcNn4^1nJeNQ5\\1[JWOe5S5O1O010O1O001O001O1O001O2M2M3J6K6I6H8A?I8H7I7I7J9G9G:F9G:F9G:Fe`2"}, "label": "woman sitting in front of cake"}]}
{"id": 168865, "image": "COCO_train2014_000000168865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady holding a camera\"."}], "task": "refering", "answer_template": "The \"a lady holding a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 361, 363, 364, 365, 366, 367], "bbox": [0.456734375, 0.16591666666666666, 0.991015625, 0.9322083333333333], "iscrowd": 0, "area": 77383.59525, "rle": {"size": [480, 640], "counts": "gWY42^>>N5L3M4L3M4K4M3M4L3M4K4M4L3M4L3L5M2M3M4M2M3M3N3L3O1O2N1O1O1O2N1N2O1O2N1O1O1O2N1N2O1O1O2N1O1O1O1O2M2O1O1O1O2N1O100O1O2N1O1O1O101N1O1O1O1O2N100O1O1O2N1O100O1O2N1O1O1O100O2N1O1O1O101N1O1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O101N1O1O1O1O2N100O1O101N1O100O101N100fG_IT8b6kG`IT8`6kGbIS8g6O1N2N2O1N2O1N2O1N100O2O000O101O000O2O000O2O000O2O0001O000O10001O000000000O2O000000001O000O1000TOl0L4L4L4L4K5L4L4L4M3M3N2N2N2N2M3N2N2N2N2M3N2N2N2N2N200000000O100000001O00000O2O0000001O000O2O0000001O000O101O0000001O000O10001O0000000O2O000000001O000O101O000000001N100000001O000000010O0001O0010O0001O00010O0kHWLS3j3iL[LU3e3hLaLcLDU6k3UMfL`LD[6f3QMUMo2k2nLYMQ3h2kL\\MT3d2hL`MX3`2eLdM[3\\2aLhM^3X2_LkMa3U2\\LoMc3R2ZLPNf3P2XLSNh3m1VLUNi3m1SLUNm3l1PLWNo3k1mKXNS4h1kKYNU4i1gKZNX4g1eK\\NZ4f1bK\\N^4e1`K]N`4d1\\K^Nd4c1YK`Nf4a1WKbNh4`1UKaNk4`1RKcNn4^1nJeNQ5\\1[JWOe5S5O1O010O1O001O001O1O001O2M2M3J6K6I6H8A?I8H7I7I7J9G9G:F9G:F9G:Fe`2"}, "label": "a lady holding a camera"}]}
{"id": 381858, "image": "COCO_train2014_000000381858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe bending its head under a beam\"."}], "task": "refering", "answer_template": "The \"a giraffe bending its head under a beam\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 283, 284, 285, 286, 287, 288, 289, 294, 295, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.20804687500000002, 0.29774999999999996, 0.892625, 0.8794166666666666], "iscrowd": 0, "area": 40669.34915, "rle": {"size": [480, 640], "counts": "UWo11n>100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100N2O1N2O1N2O1N2O1O10000000000000000000000000000000000000000000000000O10O1N2N2N2N2N2O1O100O1O1O1O100O1O1O100O1O1O100O1O1O100O100O100O1O100O100O100O100O100N2N2O1N2N2O1N2N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O0000O0100000000000O101O00000000000O10000000001O000O1000000000000O10001O000000000O10000000001O000O2O001O00001O000O2O001O00001O0O101O00001O001O0O101O00001O00000O100000O1O100O001O1O2N2O1N2M3M3M3M2N3M3M3N2M3M3N2O1N2O0O2O1N2N2L4L4L5K4L401O000O2O00001N10001O001N100010O1O010O1O10O01O103L9G3N2M3M2OF9O1O2O0O101N100O2O0O101N10001N10001O01O00001O0O101O000TK`FX4a9_KiF<F[3b9QLRG:Ed3S:QLVFn3_:O1O1O0O2O00001O1N2O1O1O1O1O3L4M3M3M3M2M3N3M2N3M2N2O2N1N2O0O100O100O100O010O100O10000O010O100O100O010O100O10O0100N2O0O2O1O1N101N2O1N101N2O1N101N2N2N1O2M3N2N3L3O2M2O2N1O2M3N1O2N1N3N1O2N2M2O2N1OlPP1"}, "label": "a giraffe bending its head under a beam"}]}
{"id": 381858, "image": "COCO_train2014_000000381858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small giraffe standing under a big giraffe\"."}], "task": "refering", "answer_template": "The \"a small giraffe standing under a big giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 283, 284, 285, 286, 287, 288, 289, 294, 295, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.20804687500000002, 0.29774999999999996, 0.892625, 0.8794166666666666], "iscrowd": 0, "area": 40669.34915, "rle": {"size": [480, 640], "counts": "UWo11n>100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100N2O1N2O1N2O1N2O1O10000000000000000000000000000000000000000000000000O10O1N2N2N2N2N2O1O100O1O1O1O100O1O1O100O1O1O100O1O1O100O100O100O1O100O100O100O100O100N2N2O1N2N2O1N2N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O0000O0100000000000O101O00000000000O10000000001O000O1000000000000O10001O000000000O10000000001O000O2O001O00001O000O2O001O00001O0O101O00001O001O0O101O00001O00000O100000O1O100O001O1O2N2O1N2M3M3M3M2N3M3M3N2M3M3N2O1N2O0O2O1N2N2L4L4L5K4L401O000O2O00001N10001O001N100010O1O010O1O10O01O103L9G3N2M3M2OF9O1O2O0O101N100O2O0O101N10001N10001O01O00001O0O101O000TK`FX4a9_KiF<F[3b9QLRG:Ed3S:QLVFn3_:O1O1O0O2O00001O1N2O1O1O1O1O3L4M3M3M3M2M3N3M2N3M2N2O2N1N2O0O100O100O100O010O100O10000O010O100O100O010O100O10O0100N2O0O2O1O1N101N2O1N101N2O1N101N2N2N1O2M3N2N3L3O2M2O2N1O2M3N1O2N1N3N1O2N2M2O2N1OlPP1"}, "label": "a small giraffe standing under a big giraffe"}]}
{"id": 381858, "image": "COCO_train2014_000000381858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older giraffe towering over a younger giraffe\"."}], "task": "refering", "answer_template": "The \"an older giraffe towering over a younger giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 50, 51, 52, 53, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342, 343, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.013484375000000002, 0.08314583333333332, 1.0, 0.9842708333333334], "iscrowd": 0, "area": 51476.68440000001, "rle": {"size": [480, 640], "counts": "^\\43k>3M3N2M3M3M3M3M3N3L3M4N1N3N1N3N0O101O0O100O100O100O10000O100O100O100O101O0O100O100O10000O100O100O100O10000O2O0O100O100O10000O100O100O100O10001N100O100O100O10000O100O100O10000O1O2K4L4M3L4M3O1O1O100O1O1O1N2I7L40000O2O000000000O100000000O100000000000010O000001O0000001O00001O0000001O0000010O0000O1N2O1O0O2O1O1N2O1O1O0O2O1O1N2O1O1N101O1N2O1O1N2O001N2O1O1O1N2O001N2001O001N2O010O010O0010O01jKmE]3T:`LPF_3o9^LUFa3l9[LYFd3f9ZL^Fd3c9XLaFg3_9WLeFh3[:N1O2O0O2O1N1O2O1N1O2N1O2N2N101N001O1O1O001O1O001O1O001O1O001O10O01O1O1O001O1O001O1O001O1O001O1O010O1O1O001O1O001O1O001O1O001O1O00100O1O001O1O001O1O001O1O1O001O1O1O010O1O1O001O1O001O1O1O001O1O1O001O100O001O1O1O001O1O1O001O1O001O1O1O00100O1O001O1O1O001O1O1O001O1O1O001O100O001O1O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O1O1O1O00fDOf70YH1Z;2O1N2N2_LHWH:h7FWH<V;0O2N2N2N2N2O1N2N101N2O1O1O1N2O001N2O1O1N2O1O1O0O2O1O1N2O1O1O00100O1O100O1O010O100O1O100O1O010O1O100O1O10O01O100O1O100O1O010O1O100O1O10O01O100O100O1O010O1O1]EdMa8\\2[GgMe8Z2WGiMi8X2RGlMo8T2mFoMS9S2hFPNX9Q2cFSN^9m1^FVNb9k1ZFXNf9j1TFZNm9f1oE]NQ:i1fEZNZ:V31O0000000000O1000000000000000000O10000000000001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001OO1000000000000000000O10000000000000000O1000000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O001O1O001O001O001O1O001O001O1O001O001O001OSB"}, "label": "an older giraffe towering over a younger giraffe"}]}
{"id": 381858, "image": "COCO_train2014_000000381858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe whose head is above the wooden frame\"."}], "task": "refering", "answer_template": "The \"a giraffe whose head is above the wooden frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 50, 51, 52, 53, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342, 343, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.013484375000000002, 0.08314583333333332, 1.0, 0.9842708333333334], "iscrowd": 0, "area": 51476.68440000001, "rle": {"size": [480, 640], "counts": "^\\43k>3M3N2M3M3M3M3M3N3L3M4N1N3N1N3N0O101O0O100O100O100O10000O100O100O100O101O0O100O100O10000O100O100O100O10000O2O0O100O100O10000O100O100O100O10001N100O100O100O10000O100O100O10000O1O2K4L4M3L4M3O1O1O100O1O1O1N2I7L40000O2O000000000O100000000O100000000000010O000001O0000001O00001O0000001O0000010O0000O1N2O1O0O2O1O1N2O1O1O0O2O1O1N2O1O1N101O1N2O1O1N2O001N2O1O1O1N2O001N2001O001N2O010O010O0010O01jKmE]3T:`LPF_3o9^LUFa3l9[LYFd3f9ZL^Fd3c9XLaFg3_9WLeFh3[:N1O2O0O2O1N1O2O1N1O2N1O2N2N101N001O1O1O001O1O001O1O001O1O001O10O01O1O1O001O1O001O1O001O1O001O1O010O1O1O001O1O001O1O001O1O001O1O00100O1O001O1O001O1O001O1O1O001O1O1O010O1O1O001O1O001O1O1O001O1O1O001O100O001O1O1O001O1O1O001O1O001O1O1O00100O1O001O1O1O001O1O1O001O1O1O001O100O001O1O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O1O1O1O00fDOf70YH1Z;2O1N2N2_LHWH:h7FWH<V;0O2N2N2N2N2O1N2N101N2O1O1O1N2O001N2O1O1N2O1O1O0O2O1O1N2O1O1O00100O1O100O1O010O100O1O100O1O010O1O100O1O10O01O100O1O100O1O010O1O100O1O10O01O100O100O1O010O1O1]EdMa8\\2[GgMe8Z2WGiMi8X2RGlMo8T2mFoMS9S2hFPNX9Q2cFSN^9m1^FVNb9k1ZFXNf9j1TFZNm9f1oE]NQ:i1fEZNZ:V31O0000000000O1000000000000000000O10000000000001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001OO1000000000000000000O10000000000000000O1000000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O001O1O001O001O001O1O001O001O1O001O001O001OSB"}, "label": "a giraffe whose head is above the wooden frame"}]}
{"id": 357289, "image": "COCO_train2014_000000357289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red truck turning to the left\"."}], "task": "refering", "answer_template": "The \"a red truck turning to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293], "bbox": [0.52584375, 0.5254583333333334, 0.792140625, 0.7232291666666666], "iscrowd": 0, "area": 14497.396149999997, "rle": {"size": [480, 640], "counts": "YXn4<b>?B>B5K1O1N2O1O2M2O1O1O1N2O2N1N2O1O1N2O2N1O1N2O1O1N2O1O2N10000O100O10000O2O0000000O1000001O000000000000001O00000000000000001O000000000000000O10000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O10000O1000000O10000O01000O10000O10000Om0TOX1hNPQn1"}, "label": "a red truck turning to the left"}]}
{"id": 357289, "image": "COCO_train2014_000000357289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red fire services truck\"."}], "task": "refering", "answer_template": "The \"a red fire services truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293], "bbox": [0.52584375, 0.5254583333333334, 0.792140625, 0.7232291666666666], "iscrowd": 0, "area": 14497.396149999997, "rle": {"size": [480, 640], "counts": "YXn4<b>?B>B5K1O1N2O1O2M2O1O1O1N2O2N1N2O1O1N2O2N1O1N2O1O1N2O1O2N10000O100O10000O2O0000000O1000001O000000000000001O00000000000000001O000000000000000O10000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O10000O1000000O10000O01000O10000O10000Om0TOX1hNPQn1"}, "label": "a red fire services truck"}]}
{"id": 185258, "image": "COCO_train2014_000000185258.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a steel and wood bench sits empty next to a couple\"."}], "task": "refering", "answer_template": "The \"a steel and wood bench sits empty next to a couple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 257, 258, 259, 280, 282], "bbox": [0.11709375, 0.46514583333333337, 0.3065, 0.7476458333333333], "iscrowd": 0, "area": 8738.060649999998, "rle": {"size": [480, 640], "counts": "U^S16]>>B>I601O001O1O001O001OL4C<2O1O1O001O1O1O1O1O1O001O1O1O1O1N2O001O1O1N2M3M3N1N3M3M5L7H8H8H7J7H8H;E2O1N2N2N2O1N21O0M4VOi001O0O101O001N101O001N101O001N101O000O2O0010O100O10O0100O100O010O100O1O10O0100O100O01ROVDeNi;X1\\DgNd;U1aDjN^;S1gDlNY;P1lDnNT;o0QEPOo:l0VESOi:j0\\ESOe:j0_EUOb:g0cEWO]:e0iEYOX:c0mE[OS:b0RF\\Oo9`0VF^Oj9?[F_Of9=_FAa9<d2L3M4LUl_6"}, "label": "a steel and wood bench sits empty next to a couple"}]}
{"id": 185258, "image": "COCO_train2014_000000185258.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long sitting chair near by two mens standing\"."}], "task": "refering", "answer_template": "The \"a long sitting chair near by two mens standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 257, 258, 259, 280, 282], "bbox": [0.11709375, 0.46514583333333337, 0.3065, 0.7476458333333333], "iscrowd": 0, "area": 8738.060649999998, "rle": {"size": [480, 640], "counts": "U^S16]>>B>I601O001O1O001O001OL4C<2O1O1O001O1O1O1O1O1O001O1O1O1O1N2O001O1O1N2M3M3N1N3M3M5L7H8H8H7J7H8H;E2O1N2N2N2O1N21O0M4VOi001O0O101O001N101O001N101O001N101O000O2O0010O100O10O0100O100O010O100O1O10O0100O100O01ROVDeNi;X1\\DgNd;U1aDjN^;S1gDlNY;P1lDnNT;o0QEPOo:l0VESOi:j0\\ESOe:j0_EUOb:g0cEWO]:e0iEYOX:c0mE[OS:b0RF\\Oo9`0VF^Oj9?[F_Of9=_FAa9<d2L3M4LUl_6"}, "label": "a long sitting chair near by two mens standing"}]}
{"id": 504748, "image": "COCO_train2014_000000504748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear with a thing on its chest that says , i love you this much\"."}], "task": "refering", "answer_template": "The \"a teddy bear with a thing on its chest that says , i love you this much\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 293, 294, 295, 296, 297], "bbox": [0.48168749999999994, 0.12098326359832635, 0.9986249999999999, 0.7636610878661088], "iscrowd": 0, "area": 73497.93190000003, "rle": {"size": [478, 640], "counts": "\\P`42k>5K5J6K5K5J5L5K5J6K5K5J6K6J6I6K6J6I6K6J2M2O2N1O2N2N1O2nNXLkFi3S9ZLkFh3Q9[LmFg3P9]LnFe3o8^LoFc3n8`LQGb3l8aLRG`3k8cLTG_3h8eLVG\\3i8fLUG\\3j8dLUG]3k8dLSG^3k8cLTG_3k8bLSG_3m8aLRGa3m8`LRG`3m8bLQG`3n8`LQGa3o8`LoFb3o8_LQGa3o8W100O10000O1000000O10000O10000O10000000O0100000000000000O1000000000000O10000fNnIhIR6S6UJkIk5Q6\\JkId5Q6cJiIa5S6fJfI^5Y6eJaI_5^6eJ[I_5d6dJVI`5i6[100O1O1O1O10000O1000000O10000O10O1000O1000000O10000O100000O01O100O100O1O100O1O10O01O100O10000000000O010000000000001O001O0O2O001O00001O001O001O001O001O001O00010O0000001O0000001O01O000001O0000010O0000000010O0000000001O01O000001O0001O0001O000000010O0000000010O0000000010O000001O01O000001O01O000001hIWJY3i5cL\\J]3c5_LbJ`3^5\\LgJc3Z5WLlJh3T5ULPKj3P5ULSKj3l4TLUKm3k4RLVKn3k4oKVKR4j4mKWKS4i4kKYKU4g4jKYKX4f4fK\\KZ4e4dK\\K\\4d4bK]K_4d4_K]Ka4d4]K\\Ke4d4XK^Kh4c4VK^Kj4c4SK^Kn4c4oJ_KQ5b4jJbKV5_4dJeK]5\\4]JiKc5X4WJlKj5U4PJPLP6Q4jITLV6Y610000O2O0O100000000001O00000000001O000001O01O000000001O000N2L4M3M4L3L4M3M4L3M3N3N1O1O2N1N2O2N1O2N1O1N3N1O1O2N1N2O2N1O2N5QOU1RNf:"}, "label": "a teddy bear with a thing on its chest that says , i love you this much"}]}
{"id": 504748, "image": "COCO_train2014_000000504748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bear on the circular stand\"."}], "task": "refering", "answer_template": "The \"the bear on the circular stand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 293, 294, 295, 296, 297], "bbox": [0.48168749999999994, 0.12098326359832635, 0.9986249999999999, 0.7636610878661088], "iscrowd": 0, "area": 73497.93190000003, "rle": {"size": [478, 640], "counts": "\\P`42k>5K5J6K5K5J5L5K5J6K5K5J6K6J6I6K6J6I6K6J2M2O2N1O2N2N1O2nNXLkFi3S9ZLkFh3Q9[LmFg3P9]LnFe3o8^LoFc3n8`LQGb3l8aLRG`3k8cLTG_3h8eLVG\\3i8fLUG\\3j8dLUG]3k8dLSG^3k8cLTG_3k8bLSG_3m8aLRGa3m8`LRG`3m8bLQG`3n8`LQGa3o8`LoFb3o8_LQGa3o8W100O10000O1000000O10000O10000O10000000O0100000000000000O1000000000000O10000fNnIhIR6S6UJkIk5Q6\\JkId5Q6cJiIa5S6fJfI^5Y6eJaI_5^6eJ[I_5d6dJVI`5i6[100O1O1O1O10000O1000000O10000O10O1000O1000000O10000O100000O01O100O100O1O100O1O10O01O100O10000000000O010000000000001O001O0O2O001O00001O001O001O001O001O001O00010O0000001O0000001O01O000001O0000010O0000000010O0000000001O01O000001O0001O0001O000000010O0000000010O0000000010O000001O01O000001O01O000001hIWJY3i5cL\\J]3c5_LbJ`3^5\\LgJc3Z5WLlJh3T5ULPKj3P5ULSKj3l4TLUKm3k4RLVKn3k4oKVKR4j4mKWKS4i4kKYKU4g4jKYKX4f4fK\\KZ4e4dK\\K\\4d4bK]K_4d4_K]Ka4d4]K\\Ke4d4XK^Kh4c4VK^Kj4c4SK^Kn4c4oJ_KQ5b4jJbKV5_4dJeK]5\\4]JiKc5X4WJlKj5U4PJPLP6Q4jITLV6Y610000O2O0O100000000001O00000000001O000001O01O000000001O000N2L4M3M4L3L4M3M4L3M3N3N1O1O2N1N2O2N1O2N1O1N3N1O1O2N1N2O2N1O2N5QOU1RNf:"}, "label": "the bear on the circular stand"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis that are being worn by the woman who is squatting . she is wearing a blue and white cap\"."}], "task": "refering", "answer_template": "The \"the skis that are being worn by the woman who is squatting . she is wearing a blue and white cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 343, 344], "bbox": [0.003565737051792829, 0.774703125, 1.0, 0.829078125], "iscrowd": 0, "area": 5440.220449999993, "rle": {"size": [640, 502], "counts": "Wh11oc00O2N1O20O01O010O0010O01O010ZOJ`]O6`b0L^]O5bb0L]]O3db0NZ]O3fb0MY]O3gb0NX]O3hb0LX]O4ib0LV]O5jb0KU]O5kb0KU]O5lb0KS]O4nb0MQ]O3Pc0LP]O4Pc0Mo\\O3Rc0Mm\\O3Sc0Mm\\O3Sc0Nl\\O2Uc0Mk\\O3Uc0Nj\\O2Vc0Nj\\O2Wc0Mi\\O3Wc0Mi\\O3Wc0Mi\\O3Wc0Mi\\O3Wc0Nh\\O2Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Le\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Le\\O5[c0Jf\\O6Zc0Jf\\O6Zc0Jf\\O6Zc0Jf\\O6Zc0Jf\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O3]c0Mc\\O3]c0Mc\\O3]c0Mc\\O3]c0Mc\\O3]c0Mc\\O3\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Me\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2[c0Oe\\O1[c0Oe\\O1[c0Oe\\O1[c0Oe\\O1[c0Oe\\O1[c0Od\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Ne\\O3[c0Me\\O3[c0Mg\\O1]c0Ke\\O3fc0N2N\\_k04jlSO2Zc0Nc\\O5]c0Kc\\O5]c0Kc\\O5\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O3\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3Zc0Nf\\O2Zc0Ne\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O2\\c0Nd\\O2]c0Mc\\O3_c0Ka\\O5`c0J`\\O6bc0H^\\O8ec00001O1O1OBKQ]O5ob0LQ]O3ob0NQ]O1ob01P]ONPc03P]OKQc06<1O1O1[\\OGac0930]\\OG_c09a\\OH^c08b\\OI]c07c\\OJ\\c06d\\OK[c05e\\OLZc04f\\OLZc04f\\OLZc04f\\OLZc04e\\OM[c03e\\OM[c03e\\OM\\c02d\\ON]c01c\\OO^c0Oc\\O1^c0Nb\\O2^c0Nb\\O2^c0Nb\\O2^c0Nb\\O2^c0Nb\\O2_c0Ma\\O3_c0M`\\O4`c0L`\\O4`c0L`\\O4`c0L`\\O4`c0L`\\O4`c0L`\\O3ac0M_\\O3ac0M_\\O3bc0L^\\O4bc0L^\\O4bc0L^\\O4bc0L^\\O4bc0L]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc04O100000O100000000000000000000000000000000G]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0400000000000000000000000000000000000O1000O1000000O1000000O1000000O10000O10000O10000O10Fc\\O1]c0Oc\\O1\\c00d\\O0\\c00d\\O0\\c00d\\O0\\c0:1O001O1O001N2E^\\O6bc0J^\\O6cc0I]\\O7cc0I]\\O7dc0H\\\\O8\\c0Hl\\O8Tc0Hl\\O8Tc0Ik\\O7Uc0Ik\\O7Uc0Jk\\O5Uc0Lk\\O3Uc0Mk\\O3Uc0Nk\\O1Uc0Ol\\O0Tc01l\\ONTc02l\\ONTc02m\\OMSc03n\\OLRc04o\\OKQc04P]OLPc04>0Y\\O1]c0Ob\\ONbc02^\\OOac01_\\O0`c00`\\O2^c0Nb\\O3]c0Mc\\O4\\c0Ke\\O5[c0Ke\\O5[c0Kf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Zc0Ig\\O7Yc0Ig\\O7Yc0Ig\\O7Yc0Ig\\O7Yc0Ig\\O7ac000000000000000000001O00000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000MY@"}, "label": "the skis that are being worn by the woman who is squatting . she is wearing a blue and white cap"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis the woman in blue is wearing\"."}], "task": "refering", "answer_template": "The \"the skis the woman in blue is wearing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 343, 344], "bbox": [0.003565737051792829, 0.774703125, 1.0, 0.829078125], "iscrowd": 0, "area": 5440.220449999993, "rle": {"size": [640, 502], "counts": "Wh11oc00O2N1O20O01O010O0010O01O010ZOJ`]O6`b0L^]O5bb0L]]O3db0NZ]O3fb0MY]O3gb0NX]O3hb0LX]O4ib0LV]O5jb0KU]O5kb0KU]O5lb0KS]O4nb0MQ]O3Pc0LP]O4Pc0Mo\\O3Rc0Mm\\O3Sc0Mm\\O3Sc0Nl\\O2Uc0Mk\\O3Uc0Nj\\O2Vc0Nj\\O2Wc0Mi\\O3Wc0Mi\\O3Wc0Mi\\O3Wc0Mi\\O3Wc0Nh\\O2Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Yc0Mg\\O3Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Le\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Le\\O5[c0Jf\\O6Zc0Jf\\O6Zc0Jf\\O6Zc0Jf\\O6Zc0Jf\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O5[c0Ke\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O3]c0Mc\\O3]c0Mc\\O3]c0Mc\\O3]c0Mc\\O3]c0Mc\\O3\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Me\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2[c0Oe\\O1[c0Oe\\O1[c0Oe\\O1[c0Oe\\O1[c0Oe\\O1[c0Od\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Nf\\O2Zc0Ne\\O3[c0Me\\O3[c0Mg\\O1]c0Ke\\O3fc0N2N\\_k04jlSO2Zc0Nc\\O5]c0Kc\\O5]c0Kc\\O5\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O4\\c0Ld\\O3\\c0Nd\\O2\\c0Nd\\O2\\c0Nd\\O2\\c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O3Zc0Nf\\O2Zc0Ne\\O3[c0Me\\O3[c0Me\\O3[c0Me\\O2\\c0Nd\\O2]c0Mc\\O3_c0Ka\\O5`c0J`\\O6bc0H^\\O8ec00001O1O1OBKQ]O5ob0LQ]O3ob0NQ]O1ob01P]ONPc03P]OKQc06<1O1O1[\\OGac0930]\\OG_c09a\\OH^c08b\\OI]c07c\\OJ\\c06d\\OK[c05e\\OLZc04f\\OLZc04f\\OLZc04f\\OLZc04e\\OM[c03e\\OM[c03e\\OM\\c02d\\ON]c01c\\OO^c0Oc\\O1^c0Nb\\O2^c0Nb\\O2^c0Nb\\O2^c0Nb\\O2^c0Nb\\O2_c0Ma\\O3_c0M`\\O4`c0L`\\O4`c0L`\\O4`c0L`\\O4`c0L`\\O4`c0L`\\O3ac0M_\\O3ac0M_\\O3bc0L^\\O4bc0L^\\O4bc0L^\\O4bc0L^\\O4bc0L]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc04O100000O100000000000000000000000000000000G]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0K]\\O5cc0400000000000000000000000000000000000O1000O1000000O1000000O1000000O10000O10000O10000O10Fc\\O1]c0Oc\\O1\\c00d\\O0\\c00d\\O0\\c00d\\O0\\c0:1O001O1O001N2E^\\O6bc0J^\\O6cc0I]\\O7cc0I]\\O7dc0H\\\\O8\\c0Hl\\O8Tc0Hl\\O8Tc0Ik\\O7Uc0Ik\\O7Uc0Jk\\O5Uc0Lk\\O3Uc0Mk\\O3Uc0Nk\\O1Uc0Ol\\O0Tc01l\\ONTc02l\\ONTc02m\\OMSc03n\\OLRc04o\\OKQc04P]OLPc04>0Y\\O1]c0Ob\\ONbc02^\\OOac01_\\O0`c00`\\O2^c0Nb\\O3]c0Mc\\O4\\c0Ke\\O5[c0Ke\\O5[c0Kf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Lf\\O4Zc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Yc0Kg\\O5Zc0Ig\\O7Yc0Ig\\O7Yc0Ig\\O7Yc0Ig\\O7Yc0Ig\\O7ac000000000000000000001O00000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000MY@"}, "label": "the skis the woman in blue is wearing"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady with the blue hat\"."}], "task": "refering", "answer_template": "The \"the lady with the blue hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 209, 210, 211, 212, 226, 227, 228, 229, 244, 245, 246, 262, 263, 264, 281, 282, 299, 300, 301, 316, 317, 318, 319, 334, 335, 336], "bbox": [0.4239840637450199, 0.192765625, 0.8592629482071713, 0.789828125], "iscrowd": 0, "area": 31853.122700000007, "rle": {"size": [640, 502], "counts": "dXU47hc08G:G8H4K6K4K6K4L5L2M100O100O2O0O100O100O2O0O100O100O2O0O1b_OeMV?\\2h@gMW?Z2g@iMV?Y2g@kMW?U2g@nMW?T2g@oMX?Q2e@SNY?m1f@UNY?l1f@VNX?k1g@WNW?i1h@YNX?g1g@YNY?h1f@YNY?g1f@ZNZ?g1e@YN[?j1a@WN`?l1\\@UNc?o1Y@QNg?T301O01O01O00001O01O1O2O1N2N2O1N2N2O1N2N2O1N2ObB^Lj:T3UEoLi:R3TERMk:m2TEVMj:k2TEXMk:g2SE]Mk:d2SE`Mj:a2SEcMl:\\2SEgMk:Z2SEiMY1QOj5U3kHnMT1ROQ6Q3iHPNP1SOW6n2fHTNk0SO`6h2dHXNe0UOg6d2bHZNd0SOk6b2_H_Nc0POn6b2]HaNc0mNP7c2[HdNa0jNU7a2XHiNa0fNW7b2VHkNa0cNZ7a2SHPO`0`N]7a2QHRO`0]N_7a2PHUO>[Nc7`2lGYO?WNe7a2jG[O>TNi7a2gG_O>PNl7a2dGB=nMo7`2cGE<kMR8`2_GI<hMU8`2]GJ=fMV8`2[GN<cMZ8_2XG1<`M\\8_2WG4;]M_8_2TG7:[Mb8_2QG:;WMe8^2oF>9UMh8^2mF`09RMk8]2kFd07PMn8]2iFf07mLP9^2gFh06kLT9]2dFk06hLV9^2aFn05fL[9\\2^FQ14cL_9]2[FS12bLd9[2XFV10aLh9Z2VFX1O_Ll9Y2SF[1M^LP:W2RFh2o9XMoEi2Q:c31O001O001RLjEPNV:o1mEoMS:Q2PFlMQ:R2RFlMn9S2UFkMl9S2WFjMj9U2YFiMh9U2[FiMe9V2^FhMc9V2`FhM`9W2cFgM^9W2eFgM[9X2hFfMY9X2jFfMV9Y2nFdMS9Z2PGdMP9\\2RGbMo8\\2UGaMl8]2WGaMj8]2ZG`Mf8_2]G_Md8_2_G_Mb8_2bG]M_8b2dG\\M]8b2gG[MZ8c2iG[MW8d2mGYMT8e2oGYMR8e2RHXMn7g2UHWMl7h2WHUMj7i2YHRMk7l2YHnLj7Q3YHjLk7T3]HbLg7\\3bHYLa7g3gHnK]7P4[4N1O2N2N101N2N2N2N2N2O1N2N2N2N3N1N2N2N2N2N2O1N2N2N2N2O1N2N2N3M2N2O1N3M2N2N2N2M4M2M3M3N3L3M4L3N3L3M4M2M3M4L3N3L3M4M[S\\1"}, "label": "the lady with the blue hat"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman leaning forward on skis\"."}], "task": "refering", "answer_template": "The \"a woman leaning forward on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 209, 210, 211, 212, 226, 227, 228, 229, 244, 245, 246, 262, 263, 264, 281, 282, 299, 300, 301, 316, 317, 318, 319, 334, 335, 336], "bbox": [0.4239840637450199, 0.192765625, 0.8592629482071713, 0.789828125], "iscrowd": 0, "area": 31853.122700000007, "rle": {"size": [640, 502], "counts": "dXU47hc08G:G8H4K6K4K6K4L5L2M100O100O2O0O100O100O2O0O100O100O2O0O1b_OeMV?\\2h@gMW?Z2g@iMV?Y2g@kMW?U2g@nMW?T2g@oMX?Q2e@SNY?m1f@UNY?l1f@VNX?k1g@WNW?i1h@YNX?g1g@YNY?h1f@YNY?g1f@ZNZ?g1e@YN[?j1a@WN`?l1\\@UNc?o1Y@QNg?T301O01O01O00001O01O1O2O1N2N2O1N2N2O1N2N2O1N2ObB^Lj:T3UEoLi:R3TERMk:m2TEVMj:k2TEXMk:g2SE]Mk:d2SE`Mj:a2SEcMl:\\2SEgMk:Z2SEiMY1QOj5U3kHnMT1ROQ6Q3iHPNP1SOW6n2fHTNk0SO`6h2dHXNe0UOg6d2bHZNd0SOk6b2_H_Nc0POn6b2]HaNc0mNP7c2[HdNa0jNU7a2XHiNa0fNW7b2VHkNa0cNZ7a2SHPO`0`N]7a2QHRO`0]N_7a2PHUO>[Nc7`2lGYO?WNe7a2jG[O>TNi7a2gG_O>PNl7a2dGB=nMo7`2cGE<kMR8`2_GI<hMU8`2]GJ=fMV8`2[GN<cMZ8_2XG1<`M\\8_2WG4;]M_8_2TG7:[Mb8_2QG:;WMe8^2oF>9UMh8^2mF`09RMk8]2kFd07PMn8]2iFf07mLP9^2gFh06kLT9]2dFk06hLV9^2aFn05fL[9\\2^FQ14cL_9]2[FS12bLd9[2XFV10aLh9Z2VFX1O_Ll9Y2SF[1M^LP:W2RFh2o9XMoEi2Q:c31O001O001RLjEPNV:o1mEoMS:Q2PFlMQ:R2RFlMn9S2UFkMl9S2WFjMj9U2YFiMh9U2[FiMe9V2^FhMc9V2`FhM`9W2cFgM^9W2eFgM[9X2hFfMY9X2jFfMV9Y2nFdMS9Z2PGdMP9\\2RGbMo8\\2UGaMl8]2WGaMj8]2ZG`Mf8_2]G_Md8_2_G_Mb8_2bG]M_8b2dG\\M]8b2gG[MZ8c2iG[MW8d2mGYMT8e2oGYMR8e2RHXMn7g2UHWMl7h2WHUMj7i2YHRMk7l2YHnLj7Q3YHjLk7T3]HbLg7\\3bHYLa7g3gHnK]7P4[4N1O2N2N101N2N2N2N2N2O1N2N2N2N3N1N2N2N2N2N2O1N2N2N2N2O1N2N2N3M2N2O1N3M2N2N2N2M4M2M3M3N3L3M4L3N3L3M4M2M3M4L3N3L3M4M[S\\1"}, "label": "a woman leaning forward on skis"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child in the purple pants and greenish top\"."}], "task": "refering", "answer_template": "The \"the child in the purple pants and greenish top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 131, 132, 133, 149, 150, 151, 167, 168, 169, 170, 185, 186, 187, 188, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 275, 276, 277, 278, 293, 294, 295, 312, 313, 330, 331, 348, 349, 365, 366, 367, 382, 383, 384, 385], "bbox": [0.18926294820717132, 0.26975, 0.4986852589641435, 0.9349218749999999], "iscrowd": 0, "area": 30962.244300000006, "rle": {"size": [640, 502], "counts": "_gk12mc02M3N2M3N2N2M3N1O2N1O2M201N101N101N1N3L3M4M2M4L300O100O100ON3N101O001N1000\\DbNg4]1ZKdNe4\\1[KeNd4[1\\KgNc4X1]KiNd4U1[KnNd4Q1\\KPOd4o0\\KSOc4m0XKXOh4g0WK\\O]3]OWGV1]5]OX3KnFj0j5]OR39fF<Y6[Om2V2TMjMg2Z2lFUMm5a0R3`2oFTMk5<P3h2RGQMj57k2S3XGkLh53j2Y3\\GeLj52e2V4ZMjK`2\\4_MdK\\2b4WMlJkJb0i7h4[MiJiJ?e7o4bMcJgJ>[7]5mMVJgJ=V7d5QNQJhJ;U7f5SNfJl1]5SNcJj1a5WJaI`2n0X3b5^IeIM1[3h0Y3d5VIRJKHd3a0[3Y7cLgH\\3U8hKlGX4V8fKjGY4X8fKhGY4Z8WIfFW2Q1a4d8\\K\\Gd4g8XKZGg4k8TKVGl4l8RKTGn4n8oJSGQ5o8lJRGT5P9iJQGW5Q9fJPGZ5]9WJeFi5i;000000000000000000000000000000000000001O0010O01O001O006J5K5K5Kg0YO5K6J2N2N1N3N1SLUGcLl8Y3ZGhJLFS9^5UG^J9Me8a5WGTJd0Jd8m5mFkIP13T8o5PG`I]1<e7o5_IlIc6P6eIkI]6P6l3K6J6]Ob0C>E:J7N3WNY@ROk?j0[@QOi?j0^@POe?m0`@mNd?o0c@kNa?P1f@jN^?Q1h@iN[?T1j1O2M3M2N3M3N2L3L5L6I_Vm4"}, "label": "the child in the purple pants and greenish top"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small kid learning how to skii\"."}], "task": "refering", "answer_template": "The \"a small kid learning how to skii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 131, 132, 133, 149, 150, 151, 167, 168, 169, 170, 185, 186, 187, 188, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 275, 276, 277, 278, 293, 294, 295, 312, 313, 330, 331, 348, 349, 365, 366, 367, 382, 383, 384, 385], "bbox": [0.18926294820717132, 0.26975, 0.4986852589641435, 0.9349218749999999], "iscrowd": 0, "area": 30962.244300000006, "rle": {"size": [640, 502], "counts": "_gk12mc02M3N2M3N2N2M3N1O2N1O2M201N101N101N1N3L3M4M2M4L300O100O100ON3N101O001N1000\\DbNg4]1ZKdNe4\\1[KeNd4[1\\KgNc4X1]KiNd4U1[KnNd4Q1\\KPOd4o0\\KSOc4m0XKXOh4g0WK\\O]3]OWGV1]5]OX3KnFj0j5]OR39fF<Y6[Om2V2TMjMg2Z2lFUMm5a0R3`2oFTMk5<P3h2RGQMj57k2S3XGkLh53j2Y3\\GeLj52e2V4ZMjK`2\\4_MdK\\2b4WMlJkJb0i7h4[MiJiJ?e7o4bMcJgJ>[7]5mMVJgJ=V7d5QNQJhJ;U7f5SNfJl1]5SNcJj1a5WJaI`2n0X3b5^IeIM1[3h0Y3d5VIRJKHd3a0[3Y7cLgH\\3U8hKlGX4V8fKjGY4X8fKhGY4Z8WIfFW2Q1a4d8\\K\\Gd4g8XKZGg4k8TKVGl4l8RKTGn4n8oJSGQ5o8lJRGT5P9iJQGW5Q9fJPGZ5]9WJeFi5i;000000000000000000000000000000000000001O0010O01O001O006J5K5K5Kg0YO5K6J2N2N1N3N1SLUGcLl8Y3ZGhJLFS9^5UG^J9Me8a5WGTJd0Jd8m5mFkIP13T8o5PG`I]1<e7o5_IlIc6P6eIkI]6P6l3K6J6]Ob0C>E:J7N3WNY@ROk?j0[@QOi?j0^@POe?m0`@mNd?o0c@kNa?P1f@jN^?Q1h@iN[?T1j1O2M3M2N3M3N2L3L5L6I_Vm4"}, "label": "a small kid learning how to skii"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis that are being worn by the child who is facing the woman\"."}], "task": "refering", "answer_template": "The \"the skis that are being worn by the child who is facing the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [363, 364, 365, 368, 369, 370, 371, 372, 373, 378, 379, 380, 381, 382, 383, 386, 387, 388, 389, 390], "bbox": [0.0034262948207171313, 0.8733437500000001, 0.8003386454183268, 0.9416406250000001], "iscrowd": 0, "area": 5213.5953999999865, "rle": {"size": [640, 502], "counts": "Vj12oc01N2N2N2N2N2N2N00000001O01O00000000000000000000000000000000000010O0000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000G@R]O`0nb0AQ]O`0mb0AS]O?mb0AS]O?lb0CS]O<nb0DR]O<mb0FR]O:nb0FR]O:mb0GS]O9mb0HR]O8nb0HR]O8mb0IS]O7mb0JR]O6mb0KS]O5mb0LR]O4mb0MS]O3mb0MS]O3lb0NT]O1mb0OS]O1lb00T]O0lb00T]O0lb00T]O0kb00V]O0jb00V]O0ib01W]OOib01W]OOhb01Y]OOgb01Y]OOfb02Z]ONfb02Z]ONeb03[]OLfb03[]OMeb03[]OMdb04\\]OLdb04\\]OLcb04^]OLbb04^]OLab05_]OKab05_]OKbb03`]OL`b04`]OLab03`]OLab03_]OMab02a]OLab03`]OL`b04`]OLab03`]OLab02a]OM_b03a]OM`b02a]OM_b02b]ON_b01b]ON^b02c]OM^b01c]OO^b00c]OO]b01d]ON]b01c]ON^b01d]ON]b01c]OO^b00l`Z10l_eN1O1O1O1O2^\\O4nb0LR]O4mb0MS]O3lb0NT]O2kb0OU]O1jb00V]O0ib01W]OOhb02X]ONgb04X]OLgb05Y]OKhb04X]OLhb04W]OMib03W]OMib03W]OMib03W]OMib02X]ONhb02X]ONhb02X]ONib01W]OOib01W]OOib01W]OOib01W]OOib01W]OOib00X]O0hb00X]O0ib0OW]O1ib0OW]O1ib0OW]O1ib0OW]O1ib0OW]O1ib0NX]O2hb0NX]O2hb0NX]O2ib0MW]O3ib0MW]O3ib0MW]O3ib0MW]O3ib0LX]O4hb0LX]O4hb0LX]O4ib0KW]O5ib0KW]O5ib0KW]O5ib0KW]O5ib0KW]O5ib0JX]O6hb0JW]O7ib0IW]O7jb0HV]O8jb0GW]O9ib0GW]O9ib0GW]O9ib0GW]O9ib0GW]O9ib0FX]O:ib0EW]O;ib0EW]O;ib0EW]O;ib0DX]O<hb0DX]O<hb0DX]O<hb0DX]O<hb0DX]O<gb0DZ]O<fb0DZ]O<fb0DZ]O<fb0DZ]O<fb0DZ]O<Uc0000@D\\]O<db0D\\]O<Tc00000000000001O000O1000000000000000000000000O100000000000000000000O1000O10000O100O10000O01000O100O10000O101O0O10000O100O10000O10000O100O10000O101O0O1000000000000000000000000000000000kaX2"}, "label": "the skis that are being worn by the child who is facing the woman"}]}
{"id": 570285, "image": "COCO_train2014_000000570285.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ski on which a child is standing\"."}], "task": "refering", "answer_template": "The \"the ski on which a child is standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [363, 364, 365, 368, 369, 370, 371, 372, 373, 378, 379, 380, 381, 382, 383, 386, 387, 388, 389, 390], "bbox": [0.0034262948207171313, 0.8733437500000001, 0.8003386454183268, 0.9416406250000001], "iscrowd": 0, "area": 5213.5953999999865, "rle": {"size": [640, 502], "counts": "Vj12oc01N2N2N2N2N2N2N00000001O01O00000000000000000000000000000000000010O0000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000G@R]O`0nb0AQ]O`0mb0AS]O?mb0AS]O?lb0CS]O<nb0DR]O<mb0FR]O:nb0FR]O:mb0GS]O9mb0HR]O8nb0HR]O8mb0IS]O7mb0JR]O6mb0KS]O5mb0LR]O4mb0MS]O3mb0MS]O3lb0NT]O1mb0OS]O1lb00T]O0lb00T]O0lb00T]O0kb00V]O0jb00V]O0ib01W]OOib01W]OOhb01Y]OOgb01Y]OOfb02Z]ONfb02Z]ONeb03[]OLfb03[]OMeb03[]OMdb04\\]OLdb04\\]OLcb04^]OLbb04^]OLab05_]OKab05_]OKbb03`]OL`b04`]OLab03`]OLab03_]OMab02a]OLab03`]OL`b04`]OLab03`]OLab02a]OM_b03a]OM`b02a]OM_b02b]ON_b01b]ON^b02c]OM^b01c]OO^b00c]OO]b01d]ON]b01c]ON^b01d]ON]b01c]OO^b00l`Z10l_eN1O1O1O1O2^\\O4nb0LR]O4mb0MS]O3lb0NT]O2kb0OU]O1jb00V]O0ib01W]OOhb02X]ONgb04X]OLgb05Y]OKhb04X]OLhb04W]OMib03W]OMib03W]OMib03W]OMib02X]ONhb02X]ONhb02X]ONib01W]OOib01W]OOib01W]OOib01W]OOib01W]OOib00X]O0hb00X]O0ib0OW]O1ib0OW]O1ib0OW]O1ib0OW]O1ib0OW]O1ib0NX]O2hb0NX]O2hb0NX]O2ib0MW]O3ib0MW]O3ib0MW]O3ib0MW]O3ib0LX]O4hb0LX]O4hb0LX]O4ib0KW]O5ib0KW]O5ib0KW]O5ib0KW]O5ib0KW]O5ib0JX]O6hb0JW]O7ib0IW]O7jb0HV]O8jb0GW]O9ib0GW]O9ib0GW]O9ib0GW]O9ib0GW]O9ib0FX]O:ib0EW]O;ib0EW]O;ib0EW]O;ib0DX]O<hb0DX]O<hb0DX]O<hb0DX]O<hb0DX]O<gb0DZ]O<fb0DZ]O<fb0DZ]O<fb0DZ]O<fb0DZ]O<Uc0000@D\\]O<db0D\\]O<Tc00000000000001O000O1000000000000000000000000O100000000000000000000O1000O10000O100O10000O01000O100O10000O101O0O10000O100O10000O10000O100O10000O101O0O1000000000000000000000000000000000kaX2"}, "label": "the ski on which a child is standing"}]}
{"id": 562092, "image": "COCO_train2014_000000562092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow eating grasses\"."}], "task": "refering", "answer_template": "The \"a cow eating grasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275, 297, 298], "bbox": [0.8440625, 0.3501890756302521, 1.0, 0.7356092436974789], "iscrowd": 0, "area": 12879.965699999992, "rle": {"size": [476, 640], "counts": "UWk75]>a0E;G9H7I5[C]Ng;j1RDZNk;k1oCYNo;l1kCWNR<[2N101N1\\DUMW;k2hDYMT;h2kD[MR;f2nDmM_:S2`EnM_:S2aEmM^:T2aEnM^:R2bEnM]:T2aElM_:U2fEfMZ:Z2TFWMl9f2W1K4H8102M2N2O1O1O1N3N2N000cDoLo:P3PETMm:l2REVMn:i2REYMm:f2SE[Mm:f2QE\\Mn:f2oD[MQ;Y3N10001N2O1O2M2O2N2M4M3M3M3L3N3M3L7J=C4K3N001N10000000000000O10O10000000O1000000000001N2O001]OTGTKm8j4d0N1N3N2N2M2O2G9FV2_MVI"}, "label": "a cow eating grasses"}]}
{"id": 562092, "image": "COCO_train2014_000000562092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow furthest to the right of the picture\"."}], "task": "refering", "answer_template": "The \"the cow furthest to the right of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275, 297, 298], "bbox": [0.8440625, 0.3501890756302521, 1.0, 0.7356092436974789], "iscrowd": 0, "area": 12879.965699999992, "rle": {"size": [476, 640], "counts": "UWk75]>a0E;G9H7I5[C]Ng;j1RDZNk;k1oCYNo;l1kCWNR<[2N101N1\\DUMW;k2hDYMT;h2kD[MR;f2nDmM_:S2`EnM_:S2aEmM^:T2aEnM^:R2bEnM]:T2aElM_:U2fEfMZ:Z2TFWMl9f2W1K4H8102M2N2O1O1O1N3N2N000cDoLo:P3PETMm:l2REVMn:i2REYMm:f2SE[Mm:f2QE\\Mn:f2oD[MQ;Y3N10001N2O1O2M2O2N2M4M3M3M3L3N3M3L7J=C4K3N001N10000000000000O10O10000000O1000000000001N2O001]OTGTKm8j4d0N1N3N2N2M2O2G9FV2_MVI"}, "label": "the cow furthest to the right of the picture"}]}
{"id": 562092, "image": "COCO_train2014_000000562092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow with 949 tag\"."}], "task": "refering", "answer_template": "The \"cow with 949 tag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 35, 37, 38, 39, 40, 41, 44, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 305, 309, 310, 311, 312, 313, 314, 326, 327, 328, 332, 333, 334, 336, 337, 350, 351, 355, 356, 357, 359, 360, 373, 374, 378, 379, 382, 383], "bbox": [0.1340625, 0.03004201680672269, 0.98634375, 0.9892647058823529], "iscrowd": 0, "area": 126821.46390000003, "rle": {"size": [476, 640], "counts": "aTX14e>6J5K6K5J7I6K6I6J7I6K6I6K4N3L4M2M4L4M2M4M3L3M4M2M3M4M2M4L3VGPLd6S4aHhL[7[3bHjLY7Z3dHiLX7Z3eHjLX7Y3dHkLX7X3fHlLU7W3hHmLT7Z3eHjLX7\\3bHhLY7^3aHfL\\7`5L5K4M3L5K5L5J6K6I6J6K5JoMZKeIn4U6P25K5K4M4K5K4L5K5K4L5K5K4M1N3M2N3M2N3M2N3M2N_1aN1O001O1O1O1O1O1O1O1O1O0000UMjLPJV3g5XMTJh2f5cMUJ\\2i5lMSJS2j5UNQJk1l5\\NPJd1m5cNnI]1o5jNnIV1o5QOlIo0R6WOjIj0T6\\OhIc0Y6@cIa0\\6D`I;_6K\\I6b61XIOh67SIIk6>oHBQ7U4O0100O010000O10000O100000000001O00000O2O000000001O000000001O00000000001O000000001O000000000O2O000000001O0000000O2O00000000001O000O1000000000000000001N10000000000000000O100000000000O10000O100O10000O10000O27H`0@a0_O`0@a0_O`0@_1aN00000000000000000O10000000000O1RO]DcNc;\\1cD_NY;e1kDWNP;m1SEQNi:R2[EkM`:Z2cEbM[:`2iE]MV:d2mEYMQ:h2Q11O0O101N101O0O101hNVM\\Fl2]9^M_Fc2Y9iMaFY2W9RNeFo1T9\\NgFe1Q9fNjF\\1n8oNnFR1m8a21O001O1O001O1O1O001O1O1O001O1O1O1O001O1O1N101O1O1O10O0101O0O101O0O10001O0O101O0O10001N10000O0100O001O1O1O010O0000O2N6K5J6J6J7I6J6I7J6J6J7H7G9E8I8K4M3L4M3L3N1O1N2O100O00100O100O1O100O1O100OBUOmBk0n<]OoBb0l<GQC8k<OSC1e<:XCEX<h1VOi0F:G:E;F:F:L3M4L4K5L4M2N3M3L4L4M2N3M3M3M3N1N3N2M3N2M3O1O1N2O1O1O1O1O1O1O1N2O2N1O1O1O101O0000001O000000002N2N2N2N2N1O1O0000000000001O000O100O100O1O2O0O3N1N2O2M2oMTHPMm7k2XHSMk7g2[HWMf7d2_HZMc7a2bH^Ml7Q2YHlMi7o1\\HoMg7k1^HTNc7T1WIhNk6U1]IcNf6Z1bIZNe6c1\\3M3L5L3M3M3M3M30N100O101N2O1O1N2O1N2O1N2O1N2O9F5LO01O100O00100O100O00100O1O010O100O10O1000000O1000000O0100000O1O1N2N2O1N2N2N2O1N2N2N1O2O1O1O1O1O1O100O1O1O1O2N1O1O1O1O1O1O2N2N3M2N3M2N3M2M3L5L3MjS4"}, "label": "cow with 949 tag"}]}
{"id": 562092, "image": "COCO_train2014_000000562092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow with an ear tag with the number 949 on it\"."}], "task": "refering", "answer_template": "The \"a cow with an ear tag with the number 949 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 35, 37, 38, 39, 40, 41, 44, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 305, 309, 310, 311, 312, 313, 314, 326, 327, 328, 332, 333, 334, 336, 337, 350, 351, 355, 356, 357, 359, 360, 373, 374, 378, 379, 382, 383], "bbox": [0.1340625, 0.03004201680672269, 0.98634375, 0.9892647058823529], "iscrowd": 0, "area": 126821.46390000003, "rle": {"size": [476, 640], "counts": "aTX14e>6J5K6K5J7I6K6I6J7I6K6I6K4N3L4M2M4L4M2M4M3L3M4M2M3M4M2M4L3VGPLd6S4aHhL[7[3bHjLY7Z3dHiLX7Z3eHjLX7Y3dHkLX7X3fHlLU7W3hHmLT7Z3eHjLX7\\3bHhLY7^3aHfL\\7`5L5K4M3L5K5L5J6K6I6J6K5JoMZKeIn4U6P25K5K4M4K5K4L5K5K4L5K5K4M1N3M2N3M2N3M2N3M2N_1aN1O001O1O1O1O1O1O1O1O1O0000UMjLPJV3g5XMTJh2f5cMUJ\\2i5lMSJS2j5UNQJk1l5\\NPJd1m5cNnI]1o5jNnIV1o5QOlIo0R6WOjIj0T6\\OhIc0Y6@cIa0\\6D`I;_6K\\I6b61XIOh67SIIk6>oHBQ7U4O0100O010000O10000O100000000001O00000O2O000000001O000000001O00000000001O000000001O000000000O2O000000001O0000000O2O00000000001O000O1000000000000000001N10000000000000000O100000000000O10000O100O10000O10000O27H`0@a0_O`0@a0_O`0@_1aN00000000000000000O10000000000O1RO]DcNc;\\1cD_NY;e1kDWNP;m1SEQNi:R2[EkM`:Z2cEbM[:`2iE]MV:d2mEYMQ:h2Q11O0O101N101O0O101hNVM\\Fl2]9^M_Fc2Y9iMaFY2W9RNeFo1T9\\NgFe1Q9fNjF\\1n8oNnFR1m8a21O001O1O001O1O1O001O1O1O001O1O1O1O001O1O1N101O1O1O10O0101O0O101O0O10001O0O101O0O10001N10000O0100O001O1O1O010O0000O2N6K5J6J6J7I6J6I7J6J6J7H7G9E8I8K4M3L4M3L3N1O1N2O100O00100O100O1O100O1O100OBUOmBk0n<]OoBb0l<GQC8k<OSC1e<:XCEX<h1VOi0F:G:E;F:F:L3M4L4K5L4M2N3M3L4L4M2N3M3M3M3N1N3N2M3N2M3O1O1N2O1O1O1O1O1O1O1N2O2N1O1O1O101O0000001O000000002N2N2N2N2N1O1O0000000000001O000O100O100O1O2O0O3N1N2O2M2oMTHPMm7k2XHSMk7g2[HWMf7d2_HZMc7a2bH^Ml7Q2YHlMi7o1\\HoMg7k1^HTNc7T1WIhNk6U1]IcNf6Z1bIZNe6c1\\3M3L5L3M3M3M3M30N100O101N2O1O1N2O1N2O1N2O1N2O9F5LO01O100O00100O100O00100O1O010O100O10O1000000O1000000O0100000O1O1N2N2O1N2N2N2O1N2N2N1O2O1O1O1O1O1O100O1O1O1O2N1O1O1O1O1O1O2N2N3M2N3M2N3M2M3L5L3MjS4"}, "label": "a cow with an ear tag with the number 949 on it"}]}
{"id": 324528, "image": "COCO_train2014_000000324528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left\"."}], "task": "refering", "answer_template": "The \"the zebra on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 186, 187, 188, 199, 200, 204, 205, 206], "bbox": [0.02104, 0.24096096096096095, 0.6554800000000001, 0.9813813813813814], "iscrowd": 0, "area": 32305.35745000001, "rle": {"size": [333, 500], "counts": "eh3<Q:0O1O1O10O01O1O010O00100O001O2O1N2N2O1N2N3L4M3L4M3L4M3L4N2M3`I[NT4g1cKbNZ4a1]KhNa4c1lJgNQ5f1YJeNf5S3M3N1N3N2M2O2M3NN1O2O0O2O1N101N101N101M2O2N1O2M2O2N1O2M2O2N1O2M3N2N3N1O1N2O2N1N2O1O1N3N1O1N2O2L3O1000001O000001O01O0000001O01O0000000000010O000000000000001O0000000000001O0000000000O10000000000O1000O1000O100000000O10O100000000000000O10000000000000000O10000000000000000O100000000000000O100VJYMn3g2jKaMV4_2bKiM^4W2VKVNi4j1jJcNV5Z30001N100N2O2N1O1O1O2N1N2O1O2N1O1O1O2N1N2O1O00001N10001N10000010G8G9G9[Oe0N3M2M3N2N2N3O0O1000001N10O10O01O10N2O1N2O2M2N2O1N2O1N3N1N2N2O1N201O0O100O10001N100O10002M4M2N2M3N3L3N2N201N1O1O101N1O100O1O1O100O00100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O0O2O1O1O4K;F8H9G9Fbhg1"}, "label": "the zebra on the left"}]}
{"id": 324528, "image": "COCO_train2014_000000324528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a complete profile view of a zebra\"."}], "task": "refering", "answer_template": "The \"a complete profile view of a zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 186, 187, 188, 199, 200, 204, 205, 206], "bbox": [0.02104, 0.24096096096096095, 0.6554800000000001, 0.9813813813813814], "iscrowd": 0, "area": 32305.35745000001, "rle": {"size": [333, 500], "counts": "eh3<Q:0O1O1O10O01O1O010O00100O001O2O1N2N2O1N2N3L4M3L4M3L4M3L4N2M3`I[NT4g1cKbNZ4a1]KhNa4c1lJgNQ5f1YJeNf5S3M3N1N3N2M2O2M3NN1O2O0O2O1N101N101N101M2O2N1O2M2O2N1O2M2O2N1O2M3N2N3N1O1N2O2N1N2O1O1N3N1O1N2O2L3O1000001O000001O01O0000001O01O0000000000010O000000000000001O0000000000001O0000000000O10000000000O1000O1000O100000000O10O100000000000000O10000000000000000O10000000000000000O100000000000000O100VJYMn3g2jKaMV4_2bKiM^4W2VKVNi4j1jJcNV5Z30001N100N2O2N1O1O1O2N1N2O1O2N1O1O1O2N1N2O1O00001N10001N10000010G8G9G9[Oe0N3M2M3N2N2N3O0O1000001N10O10O01O10N2O1N2O2M2N2O1N2O1N3N1N2N2O1N201O0O100O10001N100O10002M4M2N2M3N3L3N2N201N1O1O101N1O100O1O1O100O00100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O0O2O1O1O4K;F8H9G9Fbhg1"}, "label": "a complete profile view of a zebra"}]}
{"id": 324528, "image": "COCO_train2014_000000324528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra who is looking other way with rear facing forward\"."}], "task": "refering", "answer_template": "The \"zebra who is looking other way with rear facing forward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 83, 84, 85, 96, 97, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 171, 172, 174, 175, 189, 190, 192, 193, 207, 208, 210, 211], "bbox": [0.375, 0.3277177177177177, 0.7965, 0.9786186186186187], "iscrowd": 0, "area": 15403.6575, "rle": {"size": [333, 500], "counts": "lXm17T:6J2N2_OBRG`0i8GRG<j8KPG6m8g00O1O2N1O2O0O2N2N1O1O001O001O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O001O1O1O001O1Omg13oWN2N2N3L3N2N3M2N2M4M2Nb0^O6I7J5K<Db0^Ob0^O2N3M3M2N3M3M2N3M`NjJmMT5c1mKoMP4n1X24L4L4M4L4M3L4M3L>C2N2N3M2mI\\LZ5e3VJiLj5m300001O000hMaJB_5>_KdNb4[1eK^N[4a1kKZNU4f1oKVNQ4i1TLSNm3l1XLoMh3Q2Z2O101O0O10001N10001N10000O2O000O101OO01000O010O10O10O0\\ObN[H^1]7m0G8K6000O20O3M3M2N3M3M3M3M2N3M2N2N1O1bKSLZ2o3^MWLb2j3VM\\Lk2e3mLbLS3`3eLeL\\3\\3_LhLa3Y3[LiLf3Y3TLkLl3V3PLlLR4T3iKoLX4S3cKPM]4_42O0O101N1]O`KYKa4`4hK\\KY4_4oK]KR4^4VL_Kk3[4V1K5K6J5K5L5M2M3M4L3M3M4M2M3M4L3M4L4L5K4K5L4L4L4L4L4L4L5J5L4L6J9G\\PQ1"}, "label": "zebra who is looking other way with rear facing forward"}]}
{"id": 324528, "image": "COCO_train2014_000000324528.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"head of zebra , standing near behind of another zebra\"."}], "task": "refering", "answer_template": "The \"head of zebra , standing near behind of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 83, 84, 85, 96, 97, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 171, 172, 174, 175, 189, 190, 192, 193, 207, 208, 210, 211], "bbox": [0.375, 0.3277177177177177, 0.7965, 0.9786186186186187], "iscrowd": 0, "area": 15403.6575, "rle": {"size": [333, 500], "counts": "lXm17T:6J2N2_OBRG`0i8GRG<j8KPG6m8g00O1O2N1O2O0O2N2N1O1O001O001O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O001O1O1O001O1Omg13oWN2N2N3L3N2N3M2N2M4M2Nb0^O6I7J5K<Db0^Ob0^O2N3M3M2N3M3M2N3M`NjJmMT5c1mKoMP4n1X24L4L4M4L4M3L4M3L>C2N2N3M2mI\\LZ5e3VJiLj5m300001O000hMaJB_5>_KdNb4[1eK^N[4a1kKZNU4f1oKVNQ4i1TLSNm3l1XLoMh3Q2Z2O101O0O10001N10001N10000O2O000O101OO01000O010O10O10O0\\ObN[H^1]7m0G8K6000O20O3M3M2N3M3M3M3M2N3M2N2N1O1bKSLZ2o3^MWLb2j3VM\\Lk2e3mLbLS3`3eLeL\\3\\3_LhLa3Y3[LiLf3Y3TLkLl3V3PLlLR4T3iKoLX4S3cKPM]4_42O0O101N1]O`KYKa4`4hK\\KY4_4oK]KR4^4VL_Kk3[4V1K5K6J5K5L5M2M3M4L3M3M4M2M3M4L3M4L4L5K4K5L4L4L4L4L4L4L5J5L4L6J9G\\PQ1"}, "label": "head of zebra , standing near behind of another zebra"}]}
{"id": 160688, "image": "COCO_train2014_000000160688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blond lady in orange jacket\"."}], "task": "refering", "answer_template": "The \"blond lady in orange jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 31, 32, 33, 34, 44, 45, 46, 47, 48, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 112, 113, 114, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 224, 238, 239], "bbox": [0.0, 0.038359375, 0.6937939698492462, 0.79115625], "iscrowd": 0, "area": 72530.33574999998, "rle": {"size": [640, 398], "counts": "\\=5cc0:G9G9F:G9G:E:G9F:G9G9F;F9F:G9C=A?A?A`0_O`0A?A?A?A?A`0HXOjERHS:e7\\FXH`9a7mF]HP9a7XG\\Hd8f7aGWH\\8j7iGTHS8n7PHPHl7R8YHkGd7V8Q22O1N2O2N1N2O1N2O2N1N2O1N2O2N1O100O1O2O0O1O100O2O0O1O100O2N100O100M4L3N2M3N2M4L3N2M3M3N3L3M3N2M3M3N2M3M3N2M3N3L3M3N2M3M3N2MYOQHcEn7^:UH_Ei7c:[HYEd7g:`HVE^7l:fHPEY7P;lHlDR7V;QIgDn6Y;k02O1N101N101N2O0O2O0O2O1N101N101N2O0010O10001N10000010O000010O0001O0010O0001O010O000010O01O00010O000001O0000000O2O0000001N2O1O2N1O1N2O1O1O1N3N1O1O1O1N2O1O2N1O1N2O1O1O1O2M2O1O1O1OZ1eN\\1eN2N2N2M3N2N2N2N2M3N2N2N2N2M3N2N1O2N1O2N1O101N2N4L3M4L4L4L4L4L3M4L4L4L4L4L3M4L7I:G9F:F:F:F5K3M4L3M4L3M3M4L3M3M4L3M4L3M3M4M2M4L3M3M2N2N1O1O2N1N3N1N2O2M2N3N1N3N1N2O2M_k[2"}, "label": "blond lady in orange jacket"}]}
{"id": 160688, "image": "COCO_train2014_000000160688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue eyed blonde woman showing her teet in a smile\"."}], "task": "refering", "answer_template": "The \"a blue eyed blonde woman showing her teet in a smile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 31, 32, 33, 34, 44, 45, 46, 47, 48, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 84, 85, 86, 87, 88, 89, 90, 91, 98, 99, 100, 101, 102, 103, 104, 105, 112, 113, 114, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 154, 155, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 224, 238, 239], "bbox": [0.0, 0.038359375, 0.6937939698492462, 0.79115625], "iscrowd": 0, "area": 72530.33574999998, "rle": {"size": [640, 398], "counts": "\\=5cc0:G9G9F:G9G:E:G9F:G9G9F;F9F:G9C=A?A?A`0_O`0A?A?A?A?A`0HXOjERHS:e7\\FXH`9a7mF]HP9a7XG\\Hd8f7aGWH\\8j7iGTHS8n7PHPHl7R8YHkGd7V8Q22O1N2O2N1N2O1N2O2N1N2O1N2O2N1O100O1O2O0O1O100O2O0O1O100O2N100O100M4L3N2M3N2M4L3N2M3M3N3L3M3N2M3M3N2M3M3N2M3N3L3M3N2M3M3N2MYOQHcEn7^:UH_Ei7c:[HYEd7g:`HVE^7l:fHPEY7P;lHlDR7V;QIgDn6Y;k02O1N101N101N2O0O2O0O2O1N101N101N2O0010O10001N10000010O000010O0001O0010O0001O010O000010O01O00010O000001O0000000O2O0000001N2O1O2N1O1N2O1O1O1N3N1O1O1O1N2O1O2N1O1N2O1O1O1O2M2O1O1O1OZ1eN\\1eN2N2N2M3N2N2N2N2M3N2N2N2N2M3N2N1O2N1O2N1O101N2N4L3M4L4L4L4L4L3M4L4L4L4L4L3M4L7I:G9F:F:F:F5K3M4L3M4L3M3M4L3M3M4L3M4L3M3M4M2M4L3M3M2N2N1O1O2N1N3N1N2O2M2N3N1N3N1N2O2M_k[2"}, "label": "a blue eyed blonde woman showing her teet in a smile"}]}
{"id": 275380, "image": "COCO_train2014_000000275380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large airplane showing both the nose and tail\"."}], "task": "refering", "answer_template": "The \"large airplane showing both the nose and tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 209, 210, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 313, 314], "bbox": [0.028421875000000003, 0.5372131147540983, 0.72553125, 0.8825058548009367], "iscrowd": 0, "area": 22293.45245, "rle": {"size": [427, 640], "counts": "_d81Z=00001O0000001O00001O01O01O0000001O00001O00001O0000001O00001O00001O0jMLUG4k82oFNP98kFIT9=fFCY9c0cF\\O]9j0]FWOa9P1YFQOf9T1UFmNi9Z1QFgNn9_1lEaNS:e1hE\\NW:j1cEWN[:^2O1N101N2O1O11N2O1O001O1N1O2O1N2N1O2O1N1O2O1N2N1O2O1N1O2N2O1N1O2O1N1O2N2O0O2N2O1N1O2N2O0O1O1O01O000010O000001O00010O0000001O01O0001O00010O0000001O01O0001O00010O0000001O01O01O00000010O000001O0001O000000000010O0000000001O00000001O0000000000000000001O0000000000000000001O0000001O00001O00001O00001O00001O0000O100O100O1O100O1001O1O2N2N2N1O2N2N000001O01O0000000000000000001O0000N2O1N2N2000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000O1O1N2O1N2O1N2000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O01O000001O0000`DbNS;^1hDhNW;X1dDmN\\;a110O00001O01O01OO10001N10000O1M4E:K5O100O2O001N101N2O001N2O1N101O1N101O1N101N2O001N2O0O2O1O1N101N2O001N2O0O2O1OlQY2"}, "label": "large airplane showing both the nose and tail"}]}
{"id": 275380, "image": "COCO_train2014_000000275380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of a blue and orange airplane being taxied to the runway\"."}], "task": "refering", "answer_template": "The \"a picture of a blue and orange airplane being taxied to the runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 209, 210, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 313, 314], "bbox": [0.028421875000000003, 0.5372131147540983, 0.72553125, 0.8825058548009367], "iscrowd": 0, "area": 22293.45245, "rle": {"size": [427, 640], "counts": "_d81Z=00001O0000001O00001O01O01O0000001O00001O00001O0000001O00001O00001O0jMLUG4k82oFNP98kFIT9=fFCY9c0cF\\O]9j0]FWOa9P1YFQOf9T1UFmNi9Z1QFgNn9_1lEaNS:e1hE\\NW:j1cEWN[:^2O1N101N2O1O11N2O1O001O1N1O2O1N2N1O2O1N1O2O1N2N1O2O1N1O2N2O1N1O2O1N1O2N2O0O2N2O1N1O2N2O0O1O1O01O000010O000001O00010O0000001O01O0001O00010O0000001O01O0001O00010O0000001O01O01O00000010O000001O0001O000000000010O0000000001O00000001O0000000000000000001O0000000000000000001O0000001O00001O00001O00001O00001O0000O100O100O1O100O1001O1O2N2N2N1O2N2N000001O01O0000000000000000001O0000N2O1N2N2000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000O1O1N2O1N2O1N2000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O01O000001O0000`DbNS;^1hDhNW;X1dDmN\\;a110O00001O01O01OO10001N10000O1M4E:K5O100O2O001N101N2O001N2O1N101O1N101O1N101N2O001N2O0O2O1O1N101N2O001N2O0O2O1OlQY2"}, "label": "a picture of a blue and orange airplane being taxied to the runway"}]}
{"id": 152501, "image": "COCO_train2014_000000152501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the pink shirt\"."}], "task": "refering", "answer_template": "The \"the woman in the pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.5570468749999999, 0.4207916666666666, 0.9966406249999998, 0.9890208333333333], "iscrowd": 0, "area": 38492.698150000004, "rle": {"size": [480, 640], "counts": "ncW52k>4M3M2N3M3L3N3M3M2N3L4M2N3M3L4M3M3L4M2N3L4M3M3L4M3M2M4M3M3L4M3M3L3O4M3M4L3L4M3M4L3L4M4L1O2M2O1O2N1N2O2N1O2M2O1O2N1N3N1O1O2M2O2N1O101N1O1O2O0O2N100nF^Jd8P6O1O100O1O1O100O1O010O1O1O1000000000000000eJSHa3n7mKcHT4]7cKkH^4l80O1O1O10O01O100O1O1O010O1O1O100O001O100O1O010O1O1O100O1O0011O00000000001O0O1000001O0000000O101O00N2O1N2O1N2N101N2O1N2N2O0O010O100000O10001O000O2O0000001N1000003M3dKZFW3k9bLXF^3k9\\LYFc3j9WLYFi3k9PLYF3@d3l:ZLXEd3i:YLZEe3R;O1O101N1O1O1O2N1O1O1O2NoMjLYHU3a7TM]Hk2]7^MaHa2Z7hMdHV2\\7mM`HT2`7kMaHU2_7jMaHW2_7iMaHW2_7hMbHX2^7gMbHZ2^7fMbHZ2^7bMeH_2a7SMgHn2e90O10000O100O10000O10000O100O10000O100O10000O2O000O100O10000O10000O100O10000O10000O100O10001N100O10000O10000O100O10000O10000O100O10000O101N1]c0"}, "label": "the woman in the pink shirt"}]}
{"id": 152501, "image": "COCO_train2014_000000152501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"women playing game\"."}], "task": "refering", "answer_template": "The \"women playing game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.5570468749999999, 0.4207916666666666, 0.9966406249999998, 0.9890208333333333], "iscrowd": 0, "area": 38492.698150000004, "rle": {"size": [480, 640], "counts": "ncW52k>4M3M2N3M3L3N3M3M2N3L4M2N3M3L4M3M3L4M2N3L4M3M3L4M3M2M4M3M3L4M3M3L3O4M3M4L3L4M3M4L3L4M4L1O2M2O1O2N1N2O2N1O2M2O1O2N1N3N1O1O2M2O2N1O101N1O1O2O0O2N100nF^Jd8P6O1O100O1O1O100O1O010O1O1O1000000000000000eJSHa3n7mKcHT4]7cKkH^4l80O1O1O10O01O100O1O1O010O1O1O100O001O100O1O010O1O1O100O1O0011O00000000001O0O1000001O0000000O101O00N2O1N2O1N2N101N2O1N2N2O0O010O100000O10001O000O2O0000001N1000003M3dKZFW3k9bLXF^3k9\\LYFc3j9WLYFi3k9PLYF3@d3l:ZLXEd3i:YLZEe3R;O1O101N1O1O1O2N1O1O1O2NoMjLYHU3a7TM]Hk2]7^MaHa2Z7hMdHV2\\7mM`HT2`7kMaHU2_7jMaHW2_7iMaHW2_7hMbHX2^7gMbHZ2^7fMbHZ2^7bMeH_2a7SMgHn2e90O10000O100O10000O10000O100O10000O100O10000O2O000O100O10000O10000O100O10000O10000O100O10001N100O10000O10000O100O10000O10000O100O10000O101N1]c0"}, "label": "women playing game"}]}
{"id": 152501, "image": "COCO_train2014_000000152501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chubby man in a green mario shirt sits beside a chubby woman in a pink shirt as she plays wii\"."}], "task": "refering", "answer_template": "The \"a chubby man in a green mario shirt sits beside a chubby woman in a pink shirt as she plays wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.20056250000000003, 0.28952083333333334, 0.798875, 0.9861458333333334], "iscrowd": 0, "area": 67661.04110000002, "rle": {"size": [480, 640], "counts": "`\\l11k>:G9F:G9F;F9F:G9F;F9G9H6K001O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100lLlJ[LU5`3TK[Lm4_3]K\\Ld4_3fK[L[4_3]LoKc3m3RMaKo2^4TM_Km2_4XM]Ki2b4[MZKf2d4_MYKa2e4dMWK]2h4fMUK[2i4jMSKW2k4nMQKS2n4QNoJo1o4VNmJk1Q5S4O1N2N2O1O1O100000000O10000000000O2O000000000O100000000O100000000O100000000O100000000O10000001O00001O0000001O00001O00001O0000001O00001O001O1O001O001O001O1O001O001O001O1O001O001O1O2N3M2N3M3M2N3M2N3M2N5K7Ic0]O6J6J6J5K5K4L4L2N2N3M4L7I6J6J1O001O1O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O1O001O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O3M2N2N3M2N3M2N3M2N2N3M2N2N2N1O2N1O2N2N1O2N2N1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1gMgCi1Z<jMSDU2]<O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O001O001O001O001O00001O001O001O001O002N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N001O1O001OX_l1"}, "label": "a chubby man in a green mario shirt sits beside a chubby woman in a pink shirt as she plays wii"}]}
{"id": 152501, "image": "COCO_train2014_000000152501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses , wearing a green mushroom shirt\"."}], "task": "refering", "answer_template": "The \"a man with glasses , wearing a green mushroom shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.20056250000000003, 0.28952083333333334, 0.798875, 0.9861458333333334], "iscrowd": 0, "area": 67661.04110000002, "rle": {"size": [480, 640], "counts": "`\\l11k>:G9F:G9F;F9F:G9F;F9G9H6K001O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100lLlJ[LU5`3TK[Lm4_3]K\\Ld4_3fK[L[4_3]LoKc3m3RMaKo2^4TM_Km2_4XM]Ki2b4[MZKf2d4_MYKa2e4dMWK]2h4fMUK[2i4jMSKW2k4nMQKS2n4QNoJo1o4VNmJk1Q5S4O1N2N2O1O1O100000000O10000000000O2O000000000O100000000O100000000O100000000O100000000O10000001O00001O0000001O00001O00001O0000001O00001O001O1O001O001O001O1O001O001O001O1O001O001O1O2N3M2N3M3M2N3M2N3M2N5K7Ic0]O6J6J6J5K5K4L4L2N2N3M4L7I6J6J1O001O1O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O1O001O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O3M2N2N3M2N3M2N3M2N2N3M2N2N2N1O2N1O2N2N1O2N2N1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1gMgCi1Z<jMSDU2]<O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O001O001O001O001O00001O001O001O001O002N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N001O1O001OX_l1"}, "label": "a man with glasses , wearing a green mushroom shirt"}]}
{"id": 152501, "image": "COCO_train2014_000000152501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in brown holding something up to his ear\"."}], "task": "refering", "answer_template": "The \"a man in brown holding something up to his ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.8401249999999999, 0.27020833333333333, 1.0, 0.5881041666666665], "iscrowd": 0, "area": 10035.609050000003, "rle": {"size": [480, 640], "counts": "^]l7`0]>:F4M2M4L4N1O2N1O2N2N2N3N1N3M2N3M2N2N2N1O2N2N2O0O001O1O001O10O010O1000O001N2N1O2N2O0O2N2N1O2M2O2^Ob0J501N2O001N2O0J700O10O0100O10O0100O10O0100O01000000O1000000O10000O1O100O1O2O0O1O100O100O1O2O0O2`NnDoNS;P1PEnNQ;Q1QElNP;T1REiNS<2PDRO`7"}, "label": "a man in brown holding something up to his ear"}]}
{"id": 152501, "image": "COCO_train2014_000000152501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the far right\"."}], "task": "refering", "answer_template": "The \"the man on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.8401249999999999, 0.27020833333333333, 1.0, 0.5881041666666665], "iscrowd": 0, "area": 10035.609050000003, "rle": {"size": [480, 640], "counts": "^]l7`0]>:F4M2M4L4N1O2N1O2N2N2N3N1N3M2N3M2N2N2N1O2N2N2O0O001O1O001O10O010O1000O001N2N1O2N2O0O2N2N1O2M2O2^Ob0J501N2O001N2O0J700O10O0100O10O0100O10O0100O01000000O1000000O10000O1O100O1O2O0O1O100O100O1O2O0O2`NnDoNS;P1PEnNQ;Q1QElNP;T1REiNS<2PDRO`7"}, "label": "the man on the far right"}]}
{"id": 299959, "image": "COCO_train2014_000000299959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red boat that the two guys and dog are sitting on\"."}], "task": "refering", "answer_template": "The \"the red boat that the two guys and dog are sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.090796875, 0.5035266821345707, 0.9049843750000001, 0.932737819025522], "iscrowd": 0, "area": 56307.26150000001, "rle": {"size": [431, 640], "counts": "ldh01]=2N1N3N2M2O2N2O4L5K6J6J5K3M0O2O1O1O1O1O1O1O1O1O001O001O001O11N2OO010O001O10O01O001O001O1O001O010O1O001O001O1O001O001O1O001O00100O001O001O10O01O00100O001O010O1O0010O01O1O010O1O001O010O1O001O001O1O001O010O1O001O001O001O0O101N101N1L5L3L4M4L3O2N1O2M2O101O001O001O00001O001N101O00001O001O001O00001O001O001O001O00001O001O001N100O2O0O2O0O2O00001O001O001N10001O001O00001O001O001O00001N101O001O00001O001O00001N101O0006J5Kl0UOO000O1000001O000000001O00000000001O01O00000001O00000000001O00000010O000001O000O10001O0000001O0O10001O0000001N100000001O000O101O00000O2O0L4H801O0O100O100O02O000O101O000O10001O0O10O1O1O010O1O001O1O10O01O1O10O1000001N10001N1000001N1000000O2O000000001O0001O0001O0001O0001O0001O0001O0001O01O000001O01O0000010O00000kL`G[1a8bNaG_1^8_NdGa1\\8]NfGc1Z8ZNjGe1W8UNnGk1S8jMWHW2k7bMYH^2l7YMXHg2W91O101N1O100N2M3M4M2M3N2M3M301N1O100OdM_Fl0a9QOcFn0\\9QOgFo0X9oNkFP1X9jNlFU1W9bNPG]1S9YNTGg1o8oMXGQ2S:001O01O00000001O0N2O1O1O1_Ob0H7O1O1O1O1O2N1O100001O001O1O100O1O1O1O4L4L3M3M2N3M3N2M3M2N000O1O1O2NTN^NmGa1n7QOfGn0U8XOkGg0V8ZOiGe0X8\\OhGb0Z8]OfGb0[8_OdG`0^8_ObG`0_8A`G>a8B_G=c8C\\G<e8EYG;i8DWG;j8FUG9l8GSG9o8GPG8Q9HoF7R9JlF6V9IjF6W9JiF5Y9KeF5\\9KdF4]9MbF2`9M_F3b9N[F3g9LVF5l9KPF8Q:IkE9W:^11O2N1O1O2N1O2N1O1O2N1N2N3N1N3M3N2M3N2N2M3N2N2M3N2N1O2N2N1O2N2O0O2N1O2N2N1O2O0O2N2N2N3M2K6J5L5JZ\\i0"}, "label": "the red boat that the two guys and dog are sitting on"}]}
{"id": 299959, "image": "COCO_train2014_000000299959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red boat with a dog on top in the water\"."}], "task": "refering", "answer_template": "The \"a red boat with a dog on top in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.090796875, 0.5035266821345707, 0.9049843750000001, 0.932737819025522], "iscrowd": 0, "area": 56307.26150000001, "rle": {"size": [431, 640], "counts": "ldh01]=2N1N3N2M2O2N2O4L5K6J6J5K3M0O2O1O1O1O1O1O1O1O1O001O001O001O11N2OO010O001O10O01O001O001O1O001O010O1O001O001O1O001O001O1O001O00100O001O001O10O01O00100O001O010O1O0010O01O1O010O1O001O010O1O001O001O1O001O010O1O001O001O001O0O101N101N1L5L3L4M4L3O2N1O2M2O101O001O001O00001O001N101O00001O001O001O00001O001O001O001O00001O001O001N100O2O0O2O0O2O00001O001O001N10001O001O00001O001O001O00001N101O001O00001O001O00001N101O0006J5Kl0UOO000O1000001O000000001O00000000001O01O00000001O00000000001O00000010O000001O000O10001O0000001O0O10001O0000001N100000001O000O101O00000O2O0L4H801O0O100O100O02O000O101O000O10001O0O10O1O1O010O1O001O1O10O01O1O10O1000001N10001N1000001N1000000O2O000000001O0001O0001O0001O0001O0001O0001O0001O01O000001O01O0000010O00000kL`G[1a8bNaG_1^8_NdGa1\\8]NfGc1Z8ZNjGe1W8UNnGk1S8jMWHW2k7bMYH^2l7YMXHg2W91O101N1O100N2M3M4M2M3N2M3M301N1O100OdM_Fl0a9QOcFn0\\9QOgFo0X9oNkFP1X9jNlFU1W9bNPG]1S9YNTGg1o8oMXGQ2S:001O01O00000001O0N2O1O1O1_Ob0H7O1O1O1O1O2N1O100001O001O1O100O1O1O1O4L4L3M3M2N3M3N2M3M2N000O1O1O2NTN^NmGa1n7QOfGn0U8XOkGg0V8ZOiGe0X8\\OhGb0Z8]OfGb0[8_OdG`0^8_ObG`0_8A`G>a8B_G=c8C\\G<e8EYG;i8DWG;j8FUG9l8GSG9o8GPG8Q9HoF7R9JlF6V9IjF6W9JiF5Y9KeF5\\9KdF4]9MbF2`9M_F3b9N[F3g9LVF5l9KPF8Q:IkE9W:^11O2N1O1O2N1O2N1O1O2N1N2N3N1N3M3N2M3N2N2M3N2N2M3N2N1O2N2N1O2N2O0O2N1O2N2N1O2O0O2N2N2N3M2K6J5L5JZ\\i0"}, "label": "a red boat with a dog on top in the water"}]}
{"id": 299959, "image": "COCO_train2014_000000299959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat tied up at dock\"."}], "task": "refering", "answer_template": "The \"boat tied up at dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 128, 129, 130], "bbox": [0.5039374999999999, 0.13498839907192575, 0.8081250000000001, 0.3754292343387471], "iscrowd": 0, "area": 12976.800299999999, "rle": {"size": [431, 640], "counts": "\\QX45Z=1N2O001O0O101O00001O00001O001O00001O0000;XC@k;R10001O001O1N2O1O1O1O1N10001O000O101N1O1O2N1O2N2N1O2O0O2N1O2N1O2N2N1O2O0O2N1O2O0O2O000O100O2O0O10000O100O100O101O000O10000O10000O101O0000O10000O100O10000O100O01000O100O10000O10000O100O10O10O100O10000O10000O1O100O100O00100O100O2N101O00001O1O001O1O0010O01O1O1O001O1O1O1O100O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O2N1O2N1N3M2N3L4M2N3M3L4M3M3L4M2N3M2N2Mndc1"}, "label": "boat tied up at dock"}]}
{"id": 299959, "image": "COCO_train2014_000000299959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white boat next to a dock\"."}], "task": "refering", "answer_template": "The \"a white boat next to a dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 128, 129, 130], "bbox": [0.5039374999999999, 0.13498839907192575, 0.8081250000000001, 0.3754292343387471], "iscrowd": 0, "area": 12976.800299999999, "rle": {"size": [431, 640], "counts": "\\QX45Z=1N2O001O0O101O00001O00001O001O00001O0000;XC@k;R10001O001O1N2O1O1O1O1N10001O000O101N1O1O2N1O2N2N1O2O0O2N1O2N1O2N2N1O2O0O2N1O2O0O2O000O100O2O0O10000O100O100O101O000O10000O10000O101O0000O10000O100O10000O100O01000O100O10000O10000O100O10O10O100O10000O10000O1O100O100O00100O100O2N101O00001O1O001O1O0010O01O1O1O001O1O1O1O100O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O2N1O2N1N3M2N3L4M2N3M3L4M3M3L4M2N3M2N2Mndc1"}, "label": "a white boat next to a dock"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza pie with one slice missing is in front of a person with a purdue t shirt\"."}], "task": "refering", "answer_template": "The \"a pizza pie with one slice missing is in front of a person with a purdue t shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126], "bbox": [0.279796875, 0.14044496487119437, 0.646390625, 0.348711943793911], "iscrowd": 0, "area": 14771.153149999998, "rle": {"size": [427, 640], "counts": "WgZ21Y=2O0O2N2O1N2XCKW<h0M3L4L2O0O2O1N2O0O2O1N101N2O0O3N1N2O1N2O1N100O2O0O2O000O2O0O2O000O2O0O2O000O2O0O101O0O2O0O101O001O00001O00001O001O0000001O01O001O1O001O1O010O1O001O1O001O1O010O1O1O001O1O00100O1O1O001O1O1O001O1O1O1N102N1O1O1O1O1O1O1O1O2NE`NlD_1S;iNgDV1W;`0O10O10O10000O10000O100001O000000000O2O000000000000K5N2O100O010O11O1O001O1NO2M3N2N2M3N2N101001O0000001N10001O00000O2O000O101O0O10001O0O101O0O10001O0O101O0O10001O0O101O0O10001O0O2O1O1N2O001N2O1O1N1O2N2N2M3N1O2N2N2N2N2N4L4L4L4L4L4Lgan2"}, "label": "a pizza pie with one slice missing is in front of a person with a purdue t shirt"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the partial pizza next to the man with the gray shirt , that says purdue\"."}], "task": "refering", "answer_template": "The \"the partial pizza next to the man with the gray shirt , that says purdue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126], "bbox": [0.279796875, 0.14044496487119437, 0.646390625, 0.348711943793911], "iscrowd": 0, "area": 14771.153149999998, "rle": {"size": [427, 640], "counts": "WgZ21Y=2O0O2N2O1N2XCKW<h0M3L4L2O0O2O1N2O0O2O1N101N2O0O3N1N2O1N2O1N100O2O0O2O000O2O0O2O000O2O0O2O000O2O0O101O0O2O0O101O001O00001O00001O001O0000001O01O001O1O001O1O010O1O001O1O001O1O010O1O1O001O1O00100O1O1O001O1O1O001O1O1O1N102N1O1O1O1O1O1O1O1O2NE`NlD_1S;iNgDV1W;`0O10O10O10000O10000O100001O000000000O2O000000000000K5N2O100O010O11O1O001O1NO2M3N2N2M3N2N101001O0000001N10001O00000O2O000O101O0O10001O0O101O0O10001O0O101O0O10001O0O101O0O10001O0O2O1O1N2O001N2O1O1N1O2N2N2M3N1O2N2N2N2N2N4L4L4L4L4L4Lgan2"}, "label": "the partial pizza next to the man with the gray shirt , that says purdue"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with slices taken\"."}], "task": "refering", "answer_template": "The \"a pizza with slices taken\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 207, 208, 230], "bbox": [0.0, 0.29213114754098357, 0.373328125, 0.7101170960187353], "iscrowd": 0, "area": 22182.931450000004, "rle": {"size": [427, 640], "counts": "h4V1U<[2dMW1iN01O001O1O00100O001O1O001O100O001O1O001O10O01O1O001O1O00100O001O1O1O00100O001O1O001O1cNhFfNX9Z1PG_NP9a1WGWNj8h1^GPNb8Q2fGfM[8Z2lG^MU8a2Z110O10000O10000O010O10000O100O10O10O100O12N2N3L4M3M3M3MO1O1O1O001O1O1O1O1N2O1O001O1O1O1O1O1O10O101O000000000O10001O000000000O10001O000000000O10001O0000000000000010O00000000000010O000000000000010O000000000001O01O000000000001O01O000000000000001O01O000000000001O0000000001O0001O0000000000000010O1O001O1O1O1O010O1O1O1O1O001O100O1O002N2N2N2N2O1N2N3M2N2N2N2N2N2N2M4M2N3L7J7I6I8I6JQPW5"}, "label": "a pizza with slices taken"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with a knife under it\"."}], "task": "refering", "answer_template": "The \"pizza with a knife under it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 207, 208, 230], "bbox": [0.0, 0.29213114754098357, 0.373328125, 0.7101170960187353], "iscrowd": 0, "area": 22182.931450000004, "rle": {"size": [427, 640], "counts": "h4V1U<[2dMW1iN01O001O1O00100O001O1O001O100O001O1O001O10O01O1O001O1O00100O001O1O1O00100O001O1O001O1cNhFfNX9Z1PG_NP9a1WGWNj8h1^GPNb8Q2fGfM[8Z2lG^MU8a2Z110O10000O10000O010O10000O100O10O10O100O12N2N3L4M3M3M3MO1O1O1O001O1O1O1O1N2O1O001O1O1O1O1O1O10O101O000000000O10001O000000000O10001O000000000O10001O0000000000000010O00000000000010O000000000000010O000000000001O01O000000000001O01O000000000000001O01O000000000001O0000000001O0001O0000000000000010O1O001O1O1O1O010O1O1O1O1O001O100O1O002N2N2N2N2O1N2N3M2N2N2N2N2N2N2M4M2N3L7J7I6I8I6JQPW5"}, "label": "pizza with a knife under it"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with eggs on it\"."}], "task": "refering", "answer_template": "The \"pizza with eggs on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343], "bbox": [0.46153125, 0.3788290398126464, 1.0, 0.9623419203747071], "iscrowd": 0, "area": 68011.13955, "rle": {"size": [427, 640], "counts": "oWk31X==D<C=C=D3L4M3L4L4M3L4L5L3L4M3L4L4M3L4L5L3M3M3M3M3N2M2N2N2N2N2N101N2N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2O001N2O001N101O1N101N101O1N101O0O2O001N2O001N101N2O001N101O0O2O1O0O2O001N2O00001O000O2O00001O0O101O001O000O2O00001O0O101O00001O0O101O00001N10001O00001O001O0O101O00001O00001O00001O00001O00001O0O101O001O00001O00001O00001O00001O000O2O00001O00001O00001O001O00001O00010O000010O0001O00010O00000001O00000O2O000000001O0000001O00000O101O000000001O0000001O00000O101O000000001O0000001O00000O101O000000001O0000001O000O2O1O001O1O001O1O001O1O001O0O2O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O001O1O0O2O001O1O001O1O001O1O1O1O0O2O1O1O1O1O1O1O1O1O1N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N4L6J6J7I6J7I6J6J7IQG"}, "label": "pizza with eggs on it"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the whole pizza with fresh greens\"."}], "task": "refering", "answer_template": "The \"the whole pizza with fresh greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343], "bbox": [0.46153125, 0.3788290398126464, 1.0, 0.9623419203747071], "iscrowd": 0, "area": 68011.13955, "rle": {"size": [427, 640], "counts": "oWk31X==D<C=C=D3L4M3L4L4M3L4L5L3L4M3L4L4M3L4L5L3M3M3M3M3N2M2N2N2N2N2N101N2N2N2N2N1O2O1N2N2N2N1O2O1N2N2N2O001N2O001N101O1N101N101O1N101O0O2O001N2O001N101N2O001N101O0O2O1O0O2O001N2O00001O000O2O00001O0O101O001O000O2O00001O0O101O00001O0O101O00001N10001O00001O001O0O101O00001O00001O00001O00001O00001O0O101O001O00001O00001O00001O00001O000O2O00001O00001O00001O001O00001O00010O000010O0001O00010O00000001O00000O2O000000001O0000001O00000O101O000000001O0000001O00000O101O000000001O0000001O00000O101O000000001O0000001O000O2O1O001O1O001O1O001O1O001O0O2O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O001O1O0O2O001O1O001O1O001O1O1O1O0O2O1O1O1O1O1O1O1O1O1N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N4L6J6J7I6J7I6J6J7IQG"}, "label": "the whole pizza with fresh greens"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray purdue shirt\"."}], "task": "refering", "answer_template": "The \"a gray purdue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 109], "bbox": [0.437796875, 0.0022482435597189696, 0.8965781249999999, 0.2876346604215457], "iscrowd": 0, "area": 23115.072500000002, "rle": {"size": [427, 640], "counts": "Yhd36U=f0ZOf0ZO:F0000000000000000O1001O0000000000000000000000000000000000000000000000000000000000000000000000O10000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O0000001O001O00001O001O001O001O00001O001O001O00001O001O001O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00000000L4L4K5L4L4K5L4K5N2000000000000000000000000000000000000000000000000000000000000000000000000000000_OZE\\Ng:d1]EXNc:g1bEUN^:k1fEQNZ:n1b00000O100O100O10000O100O10000O100O104K4M4L4K4M4K4M4K5L3M4K4M4K5LR_k0"}, "label": "a gray purdue shirt"}]}
{"id": 545721, "image": "COCO_train2014_000000545721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray t - shirt with purdue in bold letters\"."}], "task": "refering", "answer_template": "The \"a gray t - shirt with purdue in bold letters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 109], "bbox": [0.437796875, 0.0022482435597189696, 0.8965781249999999, 0.2876346604215457], "iscrowd": 0, "area": 23115.072500000002, "rle": {"size": [427, 640], "counts": "Yhd36U=f0ZOf0ZO:F0000000000000000O1001O0000000000000000000000000000000000000000000000000000000000000000000000O10000001O000000001O00000000001O00000000001O000000001O00000000001O00000000001O0000001O001O00001O001O001O001O00001O001O001O00001O001O001O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00000000L4L4K5L4L4K5L4K5N2000000000000000000000000000000000000000000000000000000000000000000000000000000_OZE\\Ng:d1]EXNc:g1bEUN^:k1fEQNZ:n1b00000O100O100O10000O100O10000O100O104K4M4L4K4M4K4M4K5L3M4K4M4K5LR_k0"}, "label": "a gray t - shirt with purdue in bold letters"}]}
{"id": 127945, "image": "COCO_train2014_000000127945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"space between two train cars\"."}], "task": "refering", "answer_template": "The \"space between two train cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 218, 219], "bbox": [0.40254687499999997, 0.25787735849056603, 0.66846875, 0.6542688679245282], "iscrowd": 0, "area": 15212.346450000003, "rle": {"size": [424, 640], "counts": "mmZ32V=2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2O1N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N5Ko0QOo0QO9G001O000000000000000000000000000000000000001O00000000000000001O0000000000001O000000000000001O0000000000ZOTHPLl7d3aH[L_7Y3nHeLS7o2[InLf6l2aISM_6k2eIRM\\6m2fIQM[6m2iIPMX6o2jIPMV6n2S2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O4L6I\\eg2"}, "label": "space between two train cars"}]}
{"id": 127945, "image": "COCO_train2014_000000127945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red train in the back\"."}], "task": "refering", "answer_template": "The \"the red train in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 218, 219], "bbox": [0.40254687499999997, 0.25787735849056603, 0.66846875, 0.6542688679245282], "iscrowd": 0, "area": 15212.346450000003, "rle": {"size": [424, 640], "counts": "mmZ32V=2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2O1N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N5Ko0QOo0QO9G001O000000000000000000000000000000000000001O00000000000000001O0000000000001O000000000000001O0000000000ZOTHPLl7d3aH[L_7Y3nHeLS7o2[InLf6l2aISM_6k2eIRM\\6m2fIQM[6m2iIPMX6o2jIPMV6n2S2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O4L6I\\eg2"}, "label": "the red train in the back"}]}
{"id": 127945, "image": "COCO_train2014_000000127945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two trains that are facing one another\"."}], "task": "refering", "answer_template": "The \"two trains that are facing one another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.0, 0.12134433962264152, 1.0, 0.950566037735849], "iscrowd": 0, "area": 165438.43010000006, "rle": {"size": [424, 640], "counts": "X8V2d6^4iMW21O01O000000001O000000010O000000001O0000001O000000001O0000002N2N1O2N2N1O2N2N2N1O2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000000000O1000000000000000001O00000000000000000000000000000000000000000000000O100000000000000000000000000000000000000O10000000000O10000000000O10000000000O10000000000O10000000001N10000000000O10000000000O10000000000O10000000000O1N2M3M3L4M3M3M3L4M3M3M3L4M3O100001O00001O0000001O00001O01O101N1O100O1O1O101N1O100O1O100O1O2N1N2O1N2O1N2O2M2O1N2O1O1N2O1N3N1N2O1N2O1N3N2M4M2M3N2N2M3N2M3N2M3N2M4M2M3N2M3N2N2M3N2M3N2N3M2N2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N`0QN\\FZOb:\\OjEd0];0O10000O1000000O0100000O1000000O1000000O100000000O11O1O1O2N1N2O1O1O1O1O1O1O1O00O100000000000O10O10000000000O1N2N2N2UNHdF:Z9M_F5_93YFOf98RFJl9j1N2N2N3M2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2O1O3L3N2N3M2N2M4M2N2N3M2M3N3M2N2N4K4M3M3M4L3L4M3M4L3M3L4M4K4L4L4L5K4L4L4I8I6I7J400O1O1O1O100O1O1O1O100O1O2N1O100O1O1O1O1000000O10000O2O000O1000000O10000O10000O10001O0O100000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000001O01O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0001O0000000K5^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^OaN"}, "label": "two trains that are facing one another"}]}
{"id": 127945, "image": "COCO_train2014_000000127945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red and silver trains in the foreground\"."}], "task": "refering", "answer_template": "The \"the red and silver trains in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 301, 302, 303, 304, 305, 314, 315, 316, 317, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.0, 0.12134433962264152, 1.0, 0.950566037735849], "iscrowd": 0, "area": 165438.43010000006, "rle": {"size": [424, 640], "counts": "X8V2d6^4iMW21O01O000000001O000000010O000000001O0000001O000000001O0000002N2N1O2N2N1O2N2N2N1O2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000000000O1000000000000000001O00000000000000000000000000000000000000000000000O100000000000000000000000000000000000000O10000000000O10000000000O10000000000O10000000000O10000000001N10000000000O10000000000O10000000000O10000000000O1N2M3M3L4M3M3M3L4M3M3M3L4M3O100001O00001O0000001O00001O01O101N1O100O1O1O101N1O100O1O100O1O2N1N2O1N2O1N2O2M2O1N2O1O1N2O1N3N1N2O1N2O1N3N2M4M2M3N2N2M3N2M3N2M3N2M4M2M3N2M3N2N2M3N2M3N2N3M2N2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N`0QN\\FZOb:\\OjEd0];0O10000O1000000O0100000O1000000O1000000O100000000O11O1O1O2N1N2O1O1O1O1O1O1O1O00O100000000000O10O10000000000O1N2N2N2UNHdF:Z9M_F5_93YFOf98RFJl9j1N2N2N3M2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2O1O3L3N2N3M2N2M4M2N2N3M2M3N3M2N2N4K4M3M3M4L3L4M3M4L3M3L4M4K4L4L4L5K4L4L4I8I6I7J400O1O1O1O100O1O1O1O100O1O2N1O100O1O1O1O1000000O10000O2O000O1000000O10000O10000O10001O0O100000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000001O01O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0001O0000000K5^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^Ob0^OaN"}, "label": "the red and silver trains in the foreground"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a sofa with his arms crossed\"."}], "task": "refering", "answer_template": "The \"a man sitting on a sofa with his arms crossed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 300, 301], "bbox": [0.04596875, 0.4295208333333333, 0.258703125, 0.774625], "iscrowd": 0, "area": 14561.175850000001, "rle": {"size": [480, 640], "counts": "_l=1g><G9I6I8I5K5L4M3SCdNP<`1hClNP<X1jCoN3VOh;l2VDUMh;S3N101N1O100O100O1O100O10000O10000O1XOdLlE]3l9kLSFU3g9RMWFo2e9VMZFj2e9XMZFh2e9ZMYFg2f9[MYFe2f9]MXFd2h9^MVFb2i9aMTF_2l9eMQF[2n9\\100O010O100O01000O10O0100000O1000000O100O2O0O101O0O2O0O101N2O1N2O1O0dLcET2_:jMcEU2^:iMeET2]:kMdET2]:jMeET2]:kMdES2_:jMbEV2`:hM`EW2c:fM_EY2d:`M`E_2d:ZM`Eg2b:RMcEm2X;0001O0000001O01O01O001O00001O01O01O00001O001O101N1O1O1O1O2N1N2O1O1N2O1O2M2O1O1N2XOTCZOm<>[CAf<:_CDc<9`CFc<2cCM_<FmC9_=M3M4LVZn6"}, "label": "a man sitting on a sofa with his arms crossed"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a sofa\"."}], "task": "refering", "answer_template": "The \"a man sitting on a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 300, 301], "bbox": [0.04596875, 0.4295208333333333, 0.258703125, 0.774625], "iscrowd": 0, "area": 14561.175850000001, "rle": {"size": [480, 640], "counts": "_l=1g><G9I6I8I5K5L4M3SCdNP<`1hClNP<X1jCoN3VOh;l2VDUMh;S3N101N1O100O100O1O100O10000O10000O1XOdLlE]3l9kLSFU3g9RMWFo2e9VMZFj2e9XMZFh2e9ZMYFg2f9[MYFe2f9]MXFd2h9^MVFb2i9aMTF_2l9eMQF[2n9\\100O010O100O01000O10O0100000O1000000O100O2O0O101O0O2O0O101N2O1N2O1O0dLcET2_:jMcEU2^:iMeET2]:kMdET2]:jMeET2]:kMdES2_:jMbEV2`:hM`EW2c:fM_EY2d:`M`E_2d:ZM`Eg2b:RMcEm2X;0001O0000001O01O01O001O00001O01O01O00001O001O101N1O1O1O1O2N1N2O1O1N2O1O2M2O1O1N2XOTCZOm<>[CAf<:_CDc<9`CFc<2cCM_<FmC9_=M3M4LVZn6"}, "label": "a man sitting on a sofa"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in white\"."}], "task": "refering", "answer_template": "The \"girl in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 64, 84, 85, 86, 87, 103, 104, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 358, 359, 360, 361, 362, 384], "bbox": [0.4955, 0.092125, 0.8325781250000001, 0.9820208333333333], "iscrowd": 0, "area": 51670.1473, "rle": {"size": [480, 640], "counts": "Vhd4;`>S1PNhNWDR2T;n0O1O1O1O1O100O1O1O000000O10000O10000O100PN`D`0`;XOkDd0V;UOSEi0m:PO\\En0d:kNfEQ1[:gNnEX1l;10O01O1O00001O00010O00001O00001O00001O01O01O0hEaNP7_1mHfNQ7[1kHjNS7V1jHoNT7Q1iHTO\\OGW5V1YKl1c4TNZKT2a4lM\\K\\2_4eM]Kc2[4`MbKh2W4ZMfKn2R4ZMdKo2U4YMaKn2]4XMXKn2h4VMmJQ3S5SMcJS3]5QMYJU3g5S3O000000000000001O0000001O1O1O1O1O1O001O1O1O1N2O1O0O2L4M2N2O10000O1`NWFiMj9U2[FfMf9X2`FdM`9Z2fFbMZ9\\2lF_MU9_2QG]Mo8a2VG\\Mi8c2]GTMh8j2g101N1O2N2N1O2N1O20001N2O1O001N2O1O1O0O2O1O1O1N2O001O1N2O1O1N101O1O1N2O001O1N2ON2L4K4M4K5L4K4M4K6K5I6K6I7J5K6I7J6J5L5K5L3L5K5K5L3L5M3N2N101XOYJTHi5`7gJZHZ5[7VK_Hl4U7f1E;N101N2RKmGm2T8QMoGl2T8QMnGn2V8mLmGP3X8kLkGS3X8hLlGU3Y8fLjGX3Z8cLiGZ3[8bLhG\\3\\8^LhG_3]8\\LeGc3_8XLcGf3a8VLaGi3c8QLaGl3k9O1N2L4H9G8I7I7J6K6I6K=B`0@?B?@gna1"}, "label": "girl in white"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with a white top playing a game\"."}], "task": "refering", "answer_template": "The \"a girl with a white top playing a game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 64, 84, 85, 86, 87, 103, 104, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 358, 359, 360, 361, 362, 384], "bbox": [0.4955, 0.092125, 0.8325781250000001, 0.9820208333333333], "iscrowd": 0, "area": 51670.1473, "rle": {"size": [480, 640], "counts": "Vhd4;`>S1PNhNWDR2T;n0O1O1O1O1O100O1O1O000000O10000O10000O100PN`D`0`;XOkDd0V;UOSEi0m:PO\\En0d:kNfEQ1[:gNnEX1l;10O01O1O00001O00010O00001O00001O00001O01O01O0hEaNP7_1mHfNQ7[1kHjNS7V1jHoNT7Q1iHTO\\OGW5V1YKl1c4TNZKT2a4lM\\K\\2_4eM]Kc2[4`MbKh2W4ZMfKn2R4ZMdKo2U4YMaKn2]4XMXKn2h4VMmJQ3S5SMcJS3]5QMYJU3g5S3O000000000000001O0000001O1O1O1O1O1O001O1O1O1N2O1O0O2L4M2N2O10000O1`NWFiMj9U2[FfMf9X2`FdM`9Z2fFbMZ9\\2lF_MU9_2QG]Mo8a2VG\\Mi8c2]GTMh8j2g101N1O2N2N1O2N1O20001N2O1O001N2O1O1O0O2O1O1O1N2O001O1N2O1O1N101O1O1N2O001O1N2ON2L4K4M4K5L4K4M4K6K5I6K6I7J5K6I7J6J5L5K5L3L5K5K5L3L5M3N2N101XOYJTHi5`7gJZHZ5[7VK_Hl4U7f1E;N101N2RKmGm2T8QMoGl2T8QMnGn2V8mLmGP3X8kLkGS3X8hLlGU3Y8fLjGX3Z8cLiGZ3[8bLhG\\3\\8^LhG_3]8\\LeGc3_8XLcGf3a8VLaGi3c8QLaGl3k9O1N2L4H9G8I7I7J6K6I6K=B`0@?B?@gna1"}, "label": "a girl with a white top playing a game"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman is sitting by wearing black color top and phant\"."}], "task": "refering", "answer_template": "The \"a woman is sitting by wearing black color top and phant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 205, 226, 227, 228, 248, 249, 250, 251, 271, 272, 273, 274, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 339, 340, 341, 342, 343, 362, 363, 364], "bbox": [0.7365781250000001, 0.44316666666666665, 0.9580468750000002, 0.9219166666666666], "iscrowd": 0, "area": 18979.531949999997, "rle": {"size": [480, 640], "counts": "ZVm61l>9H8G3N3L3N2M3N2M3M3N2M3N2M3N2M3M3N2M4M2M3N2M3M3N2M3N2M3N2M3N2M3M3M4M2M3M3M3M3N2M3M3M3M3N2M3M3M3M4M2M2N3_NXKiHh4W7]KeHc4[7aK`H_4`7fK\\HZ4c7kKYHT4h7mKWHS4h7oKWHP4i7SLUHm3j7ULTHl3l7ULSHj3m7XLRHh3m7[LQHd3P8]LoGc3P8_LnGa3R8aLmG_3R8dLlG[3U8fLjGZ3U8hLjGW3V8lLhGT3X8mLfGS3Z8oLeGQ3[8PMdGo2\\8TMbGl2^8UMaGj2_8XM`Gh2`8l1O010O010O010O010O010O010O010O010O10O010O2O1N3N1N2O1gKWGX2k8fMVGZ2l8dMUGY2n8eMSGZ2o8eMRGX2Q9fMPGY2R9fMnFX2V9eMlFY2V9fMjFX2Y9fMiFW2Z9hMfFW2\\9gMeFW2_9fMcFX2_9fMbFX2a9\\MkFb2W9RMTGl2o8hL]GU3X:O1N3N101N1O1O2O0O2N1O2N100O2N1O2O0O1O2N1POf\\<"}, "label": "a woman is sitting by wearing black color top and phant"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dress lady sitting on a sofa\"."}], "task": "refering", "answer_template": "The \"a black dress lady sitting on a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 205, 226, 227, 228, 248, 249, 250, 251, 271, 272, 273, 274, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 339, 340, 341, 342, 343, 362, 363, 364], "bbox": [0.7365781250000001, 0.44316666666666665, 0.9580468750000002, 0.9219166666666666], "iscrowd": 0, "area": 18979.531949999997, "rle": {"size": [480, 640], "counts": "ZVm61l>9H8G3N3L3N2M3N2M3M3N2M3N2M3N2M3M3N2M4M2M3N2M3M3N2M3N2M3N2M3N2M3M3M4M2M3M3M3M3N2M3M3M3M3N2M3M3M3M4M2M2N3_NXKiHh4W7]KeHc4[7aK`H_4`7fK\\HZ4c7kKYHT4h7mKWHS4h7oKWHP4i7SLUHm3j7ULTHl3l7ULSHj3m7XLRHh3m7[LQHd3P8]LoGc3P8_LnGa3R8aLmG_3R8dLlG[3U8fLjGZ3U8hLjGW3V8lLhGT3X8mLfGS3Z8oLeGQ3[8PMdGo2\\8TMbGl2^8UMaGj2_8XM`Gh2`8l1O010O010O010O010O010O010O010O010O10O010O2O1N3N1N2O1gKWGX2k8fMVGZ2l8dMUGY2n8eMSGZ2o8eMRGX2Q9fMPGY2R9fMnFX2V9eMlFY2V9fMjFX2Y9fMiFW2Z9hMfFW2\\9gMeFW2_9fMcFX2_9fMbFX2a9\\MkFb2W9RMTGl2o8hL]GU3X:O1N3N101N1O1O2O0O2N1O2N100O2N1O2O0O1O2N1POf\\<"}, "label": "a black dress lady sitting on a sofa"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark couch holding a woman wearing all black\"."}], "task": "refering", "answer_template": "The \"a dark couch holding a woman wearing all black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 247, 248, 251, 252, 264, 265, 266, 270, 271, 275, 287, 288, 289, 293, 294, 298, 310, 311, 312, 320, 321, 333, 334, 335, 342, 343, 344, 357, 358, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.47903124999999996, 0.5806458333333333, 0.9967812500000001, 0.9827916666666667], "iscrowd": 0, "area": 19490.1694, "rle": {"size": [480, 640], "counts": "cX`48c>9H7H9G8H9G8K6L4L3M4L3M2M2O0000001N1000001O0O100VObMRE_2m:gMmDY2R;nMhDR2X;k0O2N100O101N100O2O0O100O2O0N2N2N3L3N2N2M4M2N2M3O2O0001O0PNhEYOW:c0RFXOn9d0\\FVOe9XOhE>n05Y9[OmE:S16P9^OQF6Y16f8BUF2^17^8DXFOd18S8G]FKj18i7JbFHn1:_7LgFDT2:V7OPJMo51YJIg55c4Ngh[14QWdN6K5J6K5J6N2O10000000000000000000000000000O1YMnN`GR1h6g0WIYOi6g0WIYOi6h0VIXOj6i0UIWOl6g0UIYOk6f0VIZOj6d0XI\\Oh6a0[I_Oe6?]IAc6<`ID`6:bIF^67eII\\64fILZ61hI0X6NjI2V6KmI5S6IoI7Q6FRJ:n5BVJ>j5]O[Jc0f5WO_Ji0a5ROdJn0\\5mNiJS1W5iNmJW1S5iNmJW1S5iNmJW1S5iNmJW1T5hNlJX1T5hNkJY1U5gNkJY1^90000000000000000000000000O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1N2@`000O10000O100O10000O100UMYM\\Ih2e6YMYIg2g6\\MVId2j6^MSIc2m6`MPI`2Q7aMlH`2T7cMiH]2W7eMeH]2[7eM^H`2c7bMVHb2j7`MPHd2P8_MhGf2X8\\MbGh2_8ZMZGj2f8XMTGl2l8WMlFn2T9TMfFP3[9Z10000001O0000001O0000001O0000001O0_Oa0XOh0iMjf0"}, "label": "a dark couch holding a woman wearing all black"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sofa that lady in black is sitting on\"."}], "task": "refering", "answer_template": "The \"the sofa that lady in black is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 247, 248, 251, 252, 264, 265, 266, 270, 271, 275, 287, 288, 289, 293, 294, 298, 310, 311, 312, 320, 321, 333, 334, 335, 342, 343, 344, 357, 358, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.47903124999999996, 0.5806458333333333, 0.9967812500000001, 0.9827916666666667], "iscrowd": 0, "area": 19490.1694, "rle": {"size": [480, 640], "counts": "cX`48c>9H7H9G8H9G8K6L4L3M4L3M2M2O0000001N1000001O0O100VObMRE_2m:gMmDY2R;nMhDR2X;k0O2N100O101N100O2O0O100O2O0N2N2N3L3N2N2M4M2N2M3O2O0001O0PNhEYOW:c0RFXOn9d0\\FVOe9XOhE>n05Y9[OmE:S16P9^OQF6Y16f8BUF2^17^8DXFOd18S8G]FKj18i7JbFHn1:_7LgFDT2:V7OPJMo51YJIg55c4Ngh[14QWdN6K5J6K5J6N2O10000000000000000000000000000O1YMnN`GR1h6g0WIYOi6g0WIYOi6h0VIXOj6i0UIWOl6g0UIYOk6f0VIZOj6d0XI\\Oh6a0[I_Oe6?]IAc6<`ID`6:bIF^67eII\\64fILZ61hI0X6NjI2V6KmI5S6IoI7Q6FRJ:n5BVJ>j5]O[Jc0f5WO_Ji0a5ROdJn0\\5mNiJS1W5iNmJW1S5iNmJW1S5iNmJW1S5iNmJW1T5hNlJX1T5hNkJY1U5gNkJY1^90000000000000000000000000O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1N2@`000O10000O100O10000O100UMYM\\Ih2e6YMYIg2g6\\MVId2j6^MSIc2m6`MPI`2Q7aMlH`2T7cMiH]2W7eMeH]2[7eM^H`2c7bMVHb2j7`MPHd2P8_MhGf2X8\\MbGh2_8ZMZGj2f8XMTGl2l8WMlFn2T9TMfFP3[9Z10000001O0000001O0000001O0000001O0_Oa0XOh0iMjf0"}, "label": "the sofa that lady in black is sitting on"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue sofa with a man seated on it\"."}], "task": "refering", "answer_template": "The \"a blue sofa with a man seated on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 276, 277, 282, 283, 299, 300, 305, 306, 322], "bbox": [0.0016093750000000001, 0.5828333333333333, 0.34603125, 0.8296041666666666], "iscrowd": 0, "area": 4283.38375, "rle": {"size": [480, 640], "counts": "Wj0W1_;3bEL^:4bEL^:5bEJ]:7cEI]:7dEH\\:8eEF\\::dEF\\::eEE[:;eEE[:<eECZ:>fEBZ:>gEAY:?gEAY:?hE_OY:a0hE^OX:b0hE^OX:c0hE\\OW:e0iE[OW:e0jEZOV:f0jEZOV:f0kEYOU:h0jEWOW:j0iEUOV:n0iEQOW:P1hEPOX:Q1hEnNX:S1gEmNY:U1fEjNZ:W1eEiN[:X1eEgNZ:[1eEdN\\:^1cEaN]:`1cE_N]:b1bE^N^:c1bE\\Nb:b1\\E^Nl:[1TEdNT;S1mDmN[;f0lDWO];;lDB\\;1lDLW=Md_k10[`TN5L4L3M4L4K5L4L4L4K2OK5J6K5K500O101O0O10001N10001N17I9G9H8G1OM3M4M2M3M4M2O1O101N1O1O1O2N1O101N1O5KlQT6"}, "label": "a blue sofa with a man seated on it"}]}
{"id": 218057, "image": "COCO_train2014_000000218057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue couch is a good place for the man to rest\"."}], "task": "refering", "answer_template": "The \"the blue couch is a good place for the man to rest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 276, 277, 282, 283, 299, 300, 305, 306, 322], "bbox": [0.0016093750000000001, 0.5828333333333333, 0.34603125, 0.8296041666666666], "iscrowd": 0, "area": 4283.38375, "rle": {"size": [480, 640], "counts": "Wj0W1_;3bEL^:4bEL^:5bEJ]:7cEI]:7dEH\\:8eEF\\::dEF\\::eEE[:;eEE[:<eECZ:>fEBZ:>gEAY:?gEAY:?hE_OY:a0hE^OX:b0hE^OX:c0hE\\OW:e0iE[OW:e0jEZOV:f0jEZOV:f0kEYOU:h0jEWOW:j0iEUOV:n0iEQOW:P1hEPOX:Q1hEnNX:S1gEmNY:U1fEjNZ:W1eEiN[:X1eEgNZ:[1eEdN\\:^1cEaN]:`1cE_N]:b1bE^N^:c1bE\\Nb:b1\\E^Nl:[1TEdNT;S1mDmN[;f0lDWO];;lDB\\;1lDLW=Md_k10[`TN5L4L3M4L4K5L4L4L4K2OK5J6K5K500O101O0O10001N10001N17I9G9H8G1OM3M4M2M3M4M2O1O101N1O1O1O2N1O101N1O5KlQT6"}, "label": "the blue couch is a good place for the man to rest"}]}
{"id": 308180, "image": "COCO_train2014_000000308180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa with a teddy bear\"."}], "task": "refering", "answer_template": "The \"a sofa with a teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332], "bbox": [0.26246875, 0.5465105386416862, 0.47046875000000005, 1.0], "iscrowd": 0, "area": 14547.97955, "rle": {"size": [427, 640], "counts": "Q\\V23V=6K6IKWC0W2IQ88kE1m1JV87oEOe1N[84RFM_12]82VFKZ16^81XFJU19a8OZFHR1<b8N\\FGn0>e8K_FFh0c0g8IaFDe0g0g8GdFC:P1P9^OgFBL]1\\9ROhFc1V9_NkF`1S9aNnF_1P9cNPG]1o8dNQG\\1h7jMgHl0BY1f7VN\\Hb0OX1e7`NPH:;V1e7L[H4d7L]H4c7L^H3a7N_H2a7MaH2_7NaH2^7OcH0]70cH0]7OdH2Z7OgH0Y7OhH1X7NjH1U7NmH2S7MnH3Q7MQI2o6MRI3n6LSI4l6LTI5l6JUI6k6IVI7i6IWI8i6FYI:f6F[I:e6E[I<e6C\\I=c6C]I>c6B]I?b6A^I?a6A_I`0a6@_I`0`6A`I?`6@`Ia0`6_O`Ia0_6_ObIa0^6_OaIb0_6^OaIb0^6^ObIc0^6]ObIc0^6\\OcId0\\6]OcId0]6\\OcId0]6[OdIe0\\6ZOdIg0\\6YOdIg0[6YOfIg0Z6YOeIi0Z6VOgIj0Y6VOgIj0Y6UOgIl0Y6SOhIm0X6SOgIn0Y6QOhIo0W6ROiIn0W6QOiIP1W6POiIP1W6oNkIP1U6POlIo0T6POnIo0R6POoIP1P6QOQJn0o5QOSJn0m5ROSJn0m5QOUJn0k5ROVJm0j5SOVJn0i5QOYJn0g5ROZJm0e5SO\\Jm0d5SO]Jl0c5TO]Jl0c5SO]Jn0c5RO]Jn0c5RO]Jn0c5QO^Jo0b5QO^Jo0a5QO`Jo0`5QO`Jo0`5QO`Jo0`5PO`JQ1`5oN`JQ1`5nNaJR1_5nNaJR1_5nNaJS1]5mNdJS1\\5mNdJS1\\5lNeJT1[5lNdJU1\\5kNdJU1\\5jNfJU1Z5kNQKj0o4VO\\K?c4AiK4W4[OnH]OV3m0l3BSM4m2H_M0a2KcM6]2FgM:Y2BjM?U2^OoMb0\\7100O2O0O1004L5K4L4L5KYn\\4"}, "label": "a sofa with a teddy bear"}]}
{"id": 308180, "image": "COCO_train2014_000000308180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teddy bear that is sitting down\"."}], "task": "refering", "answer_template": "The \"the teddy bear that is sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332], "bbox": [0.26246875, 0.5465105386416862, 0.47046875000000005, 1.0], "iscrowd": 0, "area": 14547.97955, "rle": {"size": [427, 640], "counts": "Q\\V23V=6K6IKWC0W2IQ88kE1m1JV87oEOe1N[84RFM_12]82VFKZ16^81XFJU19a8OZFHR1<b8N\\FGn0>e8K_FFh0c0g8IaFDe0g0g8GdFC:P1P9^OgFBL]1\\9ROhFc1V9_NkF`1S9aNnF_1P9cNPG]1o8dNQG\\1h7jMgHl0BY1f7VN\\Hb0OX1e7`NPH:;V1e7L[H4d7L]H4c7L^H3a7N_H2a7MaH2_7NaH2^7OcH0]70cH0]7OdH2Z7OgH0Y7OhH1X7NjH1U7NmH2S7MnH3Q7MQI2o6MRI3n6LSI4l6LTI5l6JUI6k6IVI7i6IWI8i6FYI:f6F[I:e6E[I<e6C\\I=c6C]I>c6B]I?b6A^I?a6A_I`0a6@_I`0`6A`I?`6@`Ia0`6_O`Ia0_6_ObIa0^6_OaIb0_6^OaIb0^6^ObIc0^6]ObIc0^6\\OcId0\\6]OcId0]6\\OcId0]6[OdIe0\\6ZOdIg0\\6YOdIg0[6YOfIg0Z6YOeIi0Z6VOgIj0Y6VOgIj0Y6UOgIl0Y6SOhIm0X6SOgIn0Y6QOhIo0W6ROiIn0W6QOiIP1W6POiIP1W6oNkIP1U6POlIo0T6POnIo0R6POoIP1P6QOQJn0o5QOSJn0m5ROSJn0m5QOUJn0k5ROVJm0j5SOVJn0i5QOYJn0g5ROZJm0e5SO\\Jm0d5SO]Jl0c5TO]Jl0c5SO]Jn0c5RO]Jn0c5RO]Jn0c5QO^Jo0b5QO^Jo0a5QO`Jo0`5QO`Jo0`5QO`Jo0`5PO`JQ1`5oN`JQ1`5nNaJR1_5nNaJR1_5nNaJS1]5mNdJS1\\5mNdJS1\\5lNeJT1[5lNdJU1\\5kNdJU1\\5jNfJU1Z5kNQKj0o4VO\\K?c4AiK4W4[OnH]OV3m0l3BSM4m2H_M0a2KcM6]2FgM:Y2BjM?U2^OoMb0\\7100O2O0O1004L5K4L4L5KYn\\4"}, "label": "the teddy bear that is sitting down"}]}
{"id": 308180, "image": "COCO_train2014_000000308180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a moustache\"."}], "task": "refering", "answer_template": "The \"a man with a moustache\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 39, 40, 41, 42, 62, 63, 64, 65, 86, 87, 88, 89, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320], "bbox": [0.6729375, 0.02477751756440281, 0.9662656249999999, 0.9342622950819672], "iscrowd": 0, "area": 45017.04799999999, "rle": {"size": [427, 640], "counts": "Yoc5a0j<2M2O1N2O1N2O1N2O1N1`DUO`:k0ZE[Oe:f0ZE\\Oe:c0[E_Oc:b0\\E@c:?\\EDb:=]EEb:;]EGa:9_EI`:7_EJ`:6`EL_:4`EN^:3aEO^:0bE2\\:OcE3\\:McE5[:KdE7\\:IcE9[:GeE;Z:EeE=Y:DgE=X:BhE`0V:AjE>V:BjE=X:ChE:Z:GeE8\\4SOo0d0fJ7o3DY16gJ5P4HX12iJ4o3MV10jJ4o3NV1NjJ5n31U1JnJ4l35U1GnJ5j0mNk1_1V2o0gKlMo1c1n1`0lKWNR2_1k1=kK_NV2W1j1=hKfNZ2Q1i1o1RNTNi1R2UNPNf1V2WNlMd1Y2ZNiM`1^2]NcM_1b2_N`M\\1e2bN]MX1j2dNXMX1m2fNTMV1R3gNoLT1W3jNiLS1]3jNdLQ1_3POaLm0b3SO^Lj0d3WO]Ld0g3\\OYL6U4JlKXOQ5g0PKWOR5i0oJUOR5k0nJSOT5m0mJPOU5P1kJnNW5R1jJkNX5U1hJjNY5V1hJgNZ5Y1\\300O1000000000000000000O10001O00000000000001O001O1O1O1O1O1O1O1O1O1O001O1O1UE_Nj9b1PFlNg9U1TFQOj9P1QFVOm9k0nE\\Oo9R2N2N3M2N2N1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1I7H8H9M2N8H=C>B>B=C>Ba0_Oa0_Oa0_O=C6J5K5K5K5K5D<A`0@?AYl8"}, "label": "a man with a moustache"}]}
{"id": 308180, "image": "COCO_train2014_000000308180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"a smiling man wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 39, 40, 41, 42, 62, 63, 64, 65, 86, 87, 88, 89, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320], "bbox": [0.6729375, 0.02477751756440281, 0.9662656249999999, 0.9342622950819672], "iscrowd": 0, "area": 45017.04799999999, "rle": {"size": [427, 640], "counts": "Yoc5a0j<2M2O1N2O1N2O1N2O1N1`DUO`:k0ZE[Oe:f0ZE\\Oe:c0[E_Oc:b0\\E@c:?\\EDb:=]EEb:;]EGa:9_EI`:7_EJ`:6`EL_:4`EN^:3aEO^:0bE2\\:OcE3\\:McE5[:KdE7\\:IcE9[:GeE;Z:EeE=Y:DgE=X:BhE`0V:AjE>V:BjE=X:ChE:Z:GeE8\\4SOo0d0fJ7o3DY16gJ5P4HX12iJ4o3MV10jJ4o3NV1NjJ5n31U1JnJ4l35U1GnJ5j0mNk1_1V2o0gKlMo1c1n1`0lKWNR2_1k1=kK_NV2W1j1=hKfNZ2Q1i1o1RNTNi1R2UNPNf1V2WNlMd1Y2ZNiM`1^2]NcM_1b2_N`M\\1e2bN]MX1j2dNXMX1m2fNTMV1R3gNoLT1W3jNiLS1]3jNdLQ1_3POaLm0b3SO^Lj0d3WO]Ld0g3\\OYL6U4JlKXOQ5g0PKWOR5i0oJUOR5k0nJSOT5m0mJPOU5P1kJnNW5R1jJkNX5U1hJjNY5V1hJgNZ5Y1\\300O1000000000000000000O10001O00000000000001O001O1O1O1O1O1O1O1O1O1O001O1O1UE_Nj9b1PFlNg9U1TFQOj9P1QFVOm9k0nE\\Oo9R2N2N3M2N2N1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1I7H8H9M2N8H=C>B>B=C>Ba0_Oa0_Oa0_O=C6J5K5K5K5K5D<A`0@?AYl8"}, "label": "a smiling man wearing a white shirt"}]}
{"id": 308180, "image": "COCO_train2014_000000308180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old woman playing wii\"."}], "task": "refering", "answer_template": "The \"an old woman playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 129, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337], "bbox": [0.304234375, 0.10660421545667448, 0.6698125, 0.9892505854800936], "iscrowd": 0, "area": 48730.615500000014, "rle": {"size": [427, 640], "counts": "[Rb21Y=10O01\\O0lC1Q<2oCNo;4PDMn;5QDKn;7QDJm;8SDHk;:TDGk;:TDFl;;TDEk;<TDEl;;SDFl;;TDEk;<TDDl;=SDDm;;TDEk;<TDEk;;UDFj;;VDDk;<TDEk;;VDEi;<VDEj;:WDFh;;WDEi;;WDFh;;XDEh;;WDFh;:YDEg;<XDEg;;YDFg;:YDFf;;YDFf;;ZDDf;=YDDg;<YDDg;<XDEh;;WDFi;:WDEj;;UDCn;=RDAP<k02N2O11O001O1O001O1O001O001O1O0O2O1O0O2O001N2O0O2O000O2O0O10000hIjN:W1WOYOh0g0WO[Of0g0XO\\Of0e0XO^Of0c0YO_Oe0b0YOAe0`0YOCe0>ZODd0=ZOFd0<ZOFd0<YOGe0<WOGg0<VOFg0=VOFh0=TOF7XNQLj2P3POj0d2WN_Md1c2]N_M_1c2aN`MY1b2iN_MR1c2PO_Mj0c2XO_Mb0e2^O]M<j2AWM:o2DTM5R3IPM1W3LlLOX3OkLMX31kLJZ33gLK\\33dLL^33bLL_34`LKb35^LJc36]LId37[LHg38YLGh39YLEh3;ZLAh3?YL_Oh3a0ZL\\Og3d0[LYOf3g0\\LUOf3k0\\LROe3m0_LoNb3Q1`LkNb3U1_LiNb3W1_LgNb3Y1^LfNc3Z1^LcNd3]1]LaNd3_1\\L`Ne3`1\\L]Nf3c1Q40001O0000000000000000000000000000000000000000000001O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2SFXNS8i1VGoNi8k2N2N2N2N2N2N2N001O001O001O001O001O002N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N5K5K4L5nKQJ>S6ROSK4R5[OTLKP4EV5_O`0@a0_Omhg2"}, "label": "an old woman playing wii"}]}
{"id": 308180, "image": "COCO_train2014_000000308180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with flowers on her sweating holding a remote\"."}], "task": "refering", "answer_template": "The \"a woman with flowers on her sweating holding a remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 129, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337], "bbox": [0.304234375, 0.10660421545667448, 0.6698125, 0.9892505854800936], "iscrowd": 0, "area": 48730.615500000014, "rle": {"size": [427, 640], "counts": "[Rb21Y=10O01\\O0lC1Q<2oCNo;4PDMn;5QDKn;7QDJm;8SDHk;:TDGk;:TDFl;;TDEk;<TDEl;;SDFl;;TDEk;<TDDl;=SDDm;;TDEk;<TDEk;;UDFj;;VDDk;<TDEk;;VDEi;<VDEj;:WDFh;;WDEi;;WDFh;;XDEh;;WDFh;:YDEg;<XDEg;;YDFg;:YDFf;;YDFf;;ZDDf;=YDDg;<YDDg;<XDEh;;WDFi;:WDEj;;UDCn;=RDAP<k02N2O11O001O1O001O1O001O001O1O0O2O1O0O2O001N2O0O2O000O2O0O10000hIjN:W1WOYOh0g0WO[Of0g0XO\\Of0e0XO^Of0c0YO_Oe0b0YOAe0`0YOCe0>ZODd0=ZOFd0<ZOFd0<YOGe0<WOGg0<VOFg0=VOFh0=TOF7XNQLj2P3POj0d2WN_Md1c2]N_M_1c2aN`MY1b2iN_MR1c2PO_Mj0c2XO_Mb0e2^O]M<j2AWM:o2DTM5R3IPM1W3LlLOX3OkLMX31kLJZ33gLK\\33dLL^33bLL_34`LKb35^LJc36]LId37[LHg38YLGh39YLEh3;ZLAh3?YL_Oh3a0ZL\\Og3d0[LYOf3g0\\LUOf3k0\\LROe3m0_LoNb3Q1`LkNb3U1_LiNb3W1_LgNb3Y1^LfNc3Z1^LcNd3]1]LaNd3_1\\L`Ne3`1\\L]Nf3c1Q40001O0000000000000000000000000000000000000000000001O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2SFXNS8i1VGoNi8k2N2N2N2N2N2N2N001O001O001O001O001O002N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N5K5K4L5nKQJ>S6ROSK4R5[OTLKP4EV5_O`0@a0_Omhg2"}, "label": "a woman with flowers on her sweating holding a remote"}]}
{"id": 488404, "image": "COCO_train2014_000000488404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dessert that is is the person ' s left hand\"."}], "task": "refering", "answer_template": "The \"the dessert that is is the person ' s left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179], "bbox": [0.586171875, 0.3465, 0.8346875, 0.6014999999999999], "iscrowd": 0, "area": 11082.90145, "rle": {"size": [360, 640], "counts": "boS43U;3BM`E6[:2_E2^:>M4M7H5L1O1N2O1O1O001O1O1N101N1O2N2O001N2O001O001O1O0O2O1O001O1O001N2O001O1N101O1O0O10001O0000001O000O101O0000001O0000001O000000001O000000001O0O1000001N100000001O000000001OO100000000O100000000O2O00001O0O101O001O0O2O1O001N2O001O1O0O2O001O001N101O001N101O000O2O001O0O101O1O0O2O1O0O2O0O2N1O2N1O2N2O1M3N2M3M4M2TOlE`0i:EoRU1"}, "label": "the dessert that is is the person ' s left hand"}]}
{"id": 488404, "image": "COCO_train2014_000000488404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small donut\"."}], "task": "refering", "answer_template": "The \"a small donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179], "bbox": [0.586171875, 0.3465, 0.8346875, 0.6014999999999999], "iscrowd": 0, "area": 11082.90145, "rle": {"size": [360, 640], "counts": "boS43U;3BM`E6[:2_E2^:>M4M7H5L1O1N2O1O1O001O1O1N101N1O2N2O001N2O001O001O1O0O2O1O001O1O001N2O001O1N101O1O0O10001O0000001O000O101O0000001O0000001O000000001O000000001O0O1000001N100000001O000000001OO100000000O100000000O2O00001O0O101O001O0O2O1O001N2O001O1O0O2O001O001N101O001N101O000O2O001O0O101O1O0O2O1O0O2O0O2N1O2N1O2N2O1M3N2M3M4M2TOlE`0i:EoRU1"}, "label": "a small donut"}]}
{"id": 488404, "image": "COCO_train2014_000000488404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the park bench on the left\"."}], "task": "refering", "answer_template": "The \"the back of the park bench on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 142, 198, 199, 200, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.1315, 0.07013888888888889, 0.9978906249999999, 0.9844166666666666], "iscrowd": 0, "area": 35060.10419999999, "rle": {"size": [360, 640], "counts": "Sdm01S;6J6K5J6J6K5J6J6K5J6J6K5J6J6K5J6J6K5J6M3N2N2M3N2N2N1O2N2M3N2N2N2N2N2M3N2O1N2O1O1N2O1O1N2O1N01M3M2N3N2M2N3M3N1N3M3M2N3N1N3M3M2N3O1O001O1O001O1O001O0O2O1O001O1O001O1O001O001O1O001O1O001O1N101O1O001O001O1O001O1O0O2O1O001O1O001O001N2O001O1O001O1O0O2O001O1O001O1O001N2O001O1O001O001N2O001O1O001O1O0H9_Oa0K401O001O1O010O1O001O10O01O001O10O01O1O00100O001O1O010O001O1N1O2N2N1O2Nib=OY]B0O100O100O100O10000O100O100O100O100O100O10000O101N100O100O100O100O100O1000000000000001O0000000000001OO1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2^NfMTJ\\2l5mMhIV2X6SN]Io1c6ZNRIh1n6Y11O00000000001O00000000001O000000001O00000000001O000000001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1N101O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O`0"}, "label": "the back of the park bench on the left"}]}
{"id": 488404, "image": "COCO_train2014_000000488404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench\"."}], "task": "refering", "answer_template": "The \"a wooden bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 142, 198, 199, 200, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.1315, 0.07013888888888889, 0.9978906249999999, 0.9844166666666666], "iscrowd": 0, "area": 35060.10419999999, "rle": {"size": [360, 640], "counts": "Sdm01S;6J6K5J6J6K5J6J6K5J6J6K5J6J6K5J6J6K5J6M3N2N2M3N2N2N1O2N2M3N2N2N2N2N2M3N2O1N2O1O1N2O1O1N2O1N01M3M2N3N2M2N3M3N1N3M3M2N3N1N3M3M2N3O1O001O1O001O1O001O0O2O1O001O1O001O1O001O001O1O001O1O001O1N101O1O001O001O1O001O1O0O2O1O001O1O001O001N2O001O1O001O1O0O2O001O1O001O1O001N2O001O1O001O001N2O001O1O001O1O0H9_Oa0K401O001O1O010O1O001O10O01O001O10O01O1O00100O001O1O010O001O1N1O2N2N1O2Nib=OY]B0O100O100O100O10000O100O100O100O100O100O10000O101N100O100O100O100O100O1000000000000001O0000000000001OO1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2^NfMTJ\\2l5mMhIV2X6SN]Io1c6ZNRIh1n6Y11O00000000001O00000000001O000000001O00000000001O000000001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1N101O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O`0"}, "label": "a wooden bench"}]}
{"id": 488404, "image": "COCO_train2014_000000488404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest pastry\"."}], "task": "refering", "answer_template": "The \"the biggest pastry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 288], "bbox": [0.20564062500000002, 0.3983611111111111, 0.61621875, 0.9453333333333334], "iscrowd": 0, "area": 37874.236300000004, "rle": {"size": [360, 640], "counts": "Wc^1j0]:5KO010O03H7I8N1N3N1O2M2O2N1N2O1O2N100O2N1O3M3N2M6J>Bb0^O3N2M2O2N2N1O1O100O1O1O1O1O001O001O001O001O001O00001O0O101O1O1O1O1O1O001O1O1O1O1O1O1O1O2N1N3N2N2M10000O10000000000000000O1000000000000000001O00001O0000001O00001O0O1O2N1O1O1O2O0O101O00000001O000000001O01O000001O0000001O00000O101O000O10001O000O10001O000O0100000O10000000000000000000000001O00000000000001O01O01O1O00100O1O00100O001O1O10O0100O010O10O0100O0102N1O2M10N2N1N3N2N1O2N2N1O2N2O001N2O001O1N101N101O1N101O001O001O0010O0:F1N2O1N101N2O1N2N2N2N2O1M3N2M3M3M4L3L8QO\\GmNn8f0\\GPOP9f0WGUOV9;SGAg`f2"}, "label": "the biggest pastry"}]}
{"id": 488404, "image": "COCO_train2014_000000488404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pastry in the right hand\"."}], "task": "refering", "answer_template": "The \"the pastry in the right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 288], "bbox": [0.20564062500000002, 0.3983611111111111, 0.61621875, 0.9453333333333334], "iscrowd": 0, "area": 37874.236300000004, "rle": {"size": [360, 640], "counts": "Wc^1j0]:5KO010O03H7I8N1N3N1O2M2O2N1N2O1O2N100O2N1O3M3N2M6J>Bb0^O3N2M2O2N2N1O1O100O1O1O1O1O001O001O001O001O001O00001O0O101O1O1O1O1O1O001O1O1O1O1O1O1O1O2N1N3N2N2M10000O10000000000000000O1000000000000000001O00001O0000001O00001O0O1O2N1O1O1O2O0O101O00000001O000000001O01O000001O0000001O00000O101O000O10001O000O10001O000O0100000O10000000000000000000000001O00000000000001O01O01O1O00100O1O00100O001O1O10O0100O010O10O0100O0102N1O2M10N2N1N3N2N1O2N2N1O2N2O001N2O001O1N101N101O1N101O001O001O0010O0:F1N2O1N101N2O1N2N2N2N2O1M3N2M3M3M4L3L8QO\\GmNn8f0\\GPOP9f0WGUOV9;SGAg`f2"}, "label": "the pastry in the right hand"}]}
{"id": 119765, "image": "COCO_train2014_000000119765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white shirt , green swim trunks is surfing on a green surfboard\"."}], "task": "refering", "answer_template": "The \"a man with a white shirt , green swim trunks is surfing on a green surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 121, 122, 123, 141, 142, 143, 144, 145, 164, 165, 167, 168, 190, 191, 213, 214, 237], "bbox": [0.158203125, 0.30345070422535214, 0.35953124999999997, 0.696338028169014], "iscrowd": 0, "area": 6344.810150000001, "rle": {"size": [426, 640], "counts": "bVZ1:Q=O000000000000000000000000000000000000000000000000000000O100000000000O1000O100O010O1O10O010O010O010O01O001O1N101O001O001N101O1O1POAUE?i:GRE:k:MQE3l:4oDMQ;4nDLR;5lDLQ2ZOg6k0VGMo1]Oi6g0WGNS1ZOXO7]8`0WG1Q1@RO4d8<XG1o0<h7DWG2m0?j7d0RH_Om7c0oG@P8c0hGBX8a0aGD_8c201O001O1O0010O015J<EO00XOPGWMo8g2TGYMk8e2XG[Mg8c2\\G]Mb8b2aG^M^8a2dG_M[8`2gG`MX8^2kGbMT8]2nGcMQ8[2UHaMk7^2ZH_Me7`2`H\\M`7c2fHYMY7e2mHWMS7h2h1O1O1O1O1N2gMaEf1`:XNjE^1W:aNSFP1T:POQFe0U:ZOmEc0U:]OlE`0V:_OlE?U:AmE;T:EnE6V:JkEN\\:1_100O100O1O1Nc]Z5"}, "label": "a man with a white shirt , green swim trunks is surfing on a green surfboard"}]}
{"id": 119765, "image": "COCO_train2014_000000119765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man rides a waterboard while being pulled by something attached to him by ropes\"."}], "task": "refering", "answer_template": "The \"a man rides a waterboard while being pulled by something attached to him by ropes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 121, 122, 123, 141, 142, 143, 144, 145, 164, 165, 167, 168, 190, 191, 213, 214, 237], "bbox": [0.158203125, 0.30345070422535214, 0.35953124999999997, 0.696338028169014], "iscrowd": 0, "area": 6344.810150000001, "rle": {"size": [426, 640], "counts": "bVZ1:Q=O000000000000000000000000000000000000000000000000000000O100000000000O1000O100O010O1O10O010O010O010O01O001O1N101O001O001N101O1O1POAUE?i:GRE:k:MQE3l:4oDMQ;4nDLR;5lDLQ2ZOg6k0VGMo1]Oi6g0WGNS1ZOXO7]8`0WG1Q1@RO4d8<XG1o0<h7DWG2m0?j7d0RH_Om7c0oG@P8c0hGBX8a0aGD_8c201O001O1O0010O015J<EO00XOPGWMo8g2TGYMk8e2XG[Mg8c2\\G]Mb8b2aG^M^8a2dG_M[8`2gG`MX8^2kGbMT8]2nGcMQ8[2UHaMk7^2ZH_Me7`2`H\\M`7c2fHYMY7e2mHWMS7h2h1O1O1O1O1N2gMaEf1`:XNjE^1W:aNSFP1T:POQFe0U:ZOmEc0U:]OlE`0V:_OlE?U:AmE;T:EnE6V:JkEN\\:1_100O100O1O1Nc]Z5"}, "label": "a man rides a waterboard while being pulled by something attached to him by ropes"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep being fed by a little girl\"."}], "task": "refering", "answer_template": "The \"a sheep being fed by a little girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 358, 359], "bbox": [0.41900000000000004, 0.32025, 0.8016249999999999, 0.9521041666666666], "iscrowd": 0, "area": 40528.15400000002, "rle": {"size": [480, 640], "counts": "_km34j>5K5L4L4L4L3L4M3M3M2N2O1N1O2N2N2N2N1O101N1O2N1O2N1O2O0O1O2N100O00100O1O010O001O001O0010O01O1O4L3M2N3N1N2N3M2\\DVMo:m2iD[MT;[3M3N2M3M3M3M3M3N2N1O2N2N2O1N2N2N2N2O1N2N2N2N2O1N1O2N2N1O2N2O0O2N2N1O2N2N1O2bHeJS5]5bJmJ]5T5YJUKh5l4nI]KQ6e4iI_KW6b4eIbKZ6`4aIcK`6_6O1O0010O01O010OK5K6J5K5L5K4L4L4L5K4L4M3N2O1N2O1O1N2O100O01O1N2N2O1N3M2O2M2O2M2cHXIa6j6[I\\I`6f6]IaI_6`6[IhIa6^7M2N1OO1O1O2N1N3M2O2L3J6J7K4N3M2O2N101O0010O01O001O1O010O10000O010O1000O0100O100O1O1N2N2N2N2N2O1M3M3N2M3M3M3N3M200O101N1O101N101N101N1O2O0O2O0O101N1O2O0O2O0O2M3N1N3N1O2M3N3L3N2N3L3N2M3N3N1N2N2M3L4K6UNTE[OP;O]2A_[k1"}, "label": "a sheep being fed by a little girl"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the animal that ' s being pet by the people\"."}], "task": "refering", "answer_template": "The \"the animal that ' s being pet by the people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 358, 359], "bbox": [0.41900000000000004, 0.32025, 0.8016249999999999, 0.9521041666666666], "iscrowd": 0, "area": 40528.15400000002, "rle": {"size": [480, 640], "counts": "_km34j>5K5L4L4L4L3L4M3M3M2N2O1N1O2N2N2N2N1O101N1O2N1O2N1O2O0O1O2N100O00100O1O010O001O001O0010O01O1O4L3M2N3N1N2N3M2\\DVMo:m2iD[MT;[3M3N2M3M3M3M3M3N2N1O2N2N2O1N2N2N2N2O1N2N2N2N2O1N1O2N2N1O2N2O0O2N2N1O2N2N1O2bHeJS5]5bJmJ]5T5YJUKh5l4nI]KQ6e4iI_KW6b4eIbKZ6`4aIcK`6_6O1O0010O01O010OK5K6J5K5L5K4L4L4L5K4L4M3N2O1N2O1O1N2O100O01O1N2N2O1N3M2O2M2O2M2cHXIa6j6[I\\I`6f6]IaI_6`6[IhIa6^7M2N1OO1O1O2N1N3M2O2L3J6J7K4N3M2O2N101O0010O01O001O1O010O10000O010O1000O0100O100O1O1N2N2N2N2N2O1M3M3N2M3M3M3N3M200O101N1O101N101N101N1O2O0O2O0O101N1O2O0O2O0O2M3N1N3N1O2M3N3L3N2N3L3N2M3N3N1N2N2M3L4K6UNTE[OP;O]2A_[k1"}, "label": "the animal that ' s being pet by the people"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep standing next another sheep in a field\"."}], "task": "refering", "answer_template": "The \"a sheep standing next another sheep in a field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 189, 190, 193, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304], "bbox": [0.18284375, 0.32741666666666663, 0.5393125, 0.8112499999999999], "iscrowd": 0, "area": 12509.570300000001, "rle": {"size": [480, 640], "counts": "aUg12l>5K4L5K5K5K5L3L5K5K5K5K4L5L4K5K5K4L2N100O2N2N3M2M3N2M4L3M3N2M3M4L3M3N2M4L3M3N2M3M4L3N2O1N01L4M2M4L3011O0O10001N10lMjE_OU:9XFCh94fFHY9OUGLj8LdGO\\8[O_H`0`7^OhH=X7CmH9R7HRI3m6NXIMh64\\IGd6:_IC_6>Q41N2O001N101000O0100O0100O01000O0100O0100O01000O0100O0100O01000O0100O01000O101O0O2O0O2O000O2O001N101N101O0O2O001N101N101O0O2O001O0O3N2N2N2M3N3M2M3N2N>B9FN3M3M3N1N3M3N2M2N3N2O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1N101O1O1O001O1O1O1O1O1O1O1O1N2O1O1O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2O1OTTZ4"}, "label": "a sheep standing next another sheep in a field"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the animal on the left with its head down\"."}], "task": "refering", "answer_template": "The \"the animal on the left with its head down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 189, 190, 193, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304], "bbox": [0.18284375, 0.32741666666666663, 0.5393125, 0.8112499999999999], "iscrowd": 0, "area": 12509.570300000001, "rle": {"size": [480, 640], "counts": "aUg12l>5K4L5K5K5K5L3L5K5K5K5K4L5L4K5K5K4L2N100O2N2N3M2M3N2M4L3M3N2M3M4L3M3N2M4L3M3N2M3M4L3N2O1N01L4M2M4L3011O0O10001N10lMjE_OU:9XFCh94fFHY9OUGLj8LdGO\\8[O_H`0`7^OhH=X7CmH9R7HRI3m6NXIMh64\\IGd6:_IC_6>Q41N2O001N101000O0100O0100O01000O0100O0100O01000O0100O0100O01000O0100O01000O101O0O2O0O2O000O2O001N101N101O0O2O001N101N101O0O2O001O0O3N2N2N2M3N3M2M3N2N>B9FN3M3M3N1N3M3N2M2N3N2O1O001O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1N101O1O1O001O1O1O1O1O1O1O1O1N2O1O1O0O2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2O1OTTZ4"}, "label": "the animal on the left with its head down"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women feeding animal\"."}], "task": "refering", "answer_template": "The \"a women feeding animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 219, 220, 221, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 335, 336, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.54484375, 0.3004583333333333, 1.0, 0.9887916666666666], "iscrowd": 0, "area": 57847.39555, "rle": {"size": [480, 640], "counts": "PlS53j>5K5K5L5J5K5I7iFiN^4Z1`KiN_4X1]KlNb4U1[KmNf4S1WKPOh4P1UKSOk4n0RKUOn4k0nJYOQ5g0mJ[OS5f0iJ^OW5b0fJ@Z5a0bJC^5<aJF^5;aJE_5<_JFa59_JGa5:^JGa5:^JFc5:[JHd58\\JIc58\\JHe58YJJf57YJIh56XJKg56WJLh54XJLi54VJMi54UJMk53UJNk52TJOk52TJNl52SJ0m50RJ0n51QJ0o5OPJ2P6OoI1R6NnI2S6NkI3U6NjI2W6MiI3W6NhI2Y6NeI3[6MeI3\\6McI3]6MbI4_6L`I4`6M_I3b6L]I5c6L\\I4e6LZI4f6LZI4g6LWI5i6LWI3j6LWI3i6NWI1j6NWI1i60WIOj61VINj62WIMj63VILj65VIJk65WIIi68WIGj68WIGi6:WIEj6;VIDk6;VIDj6=VIBk6>UIAk6?VI@k6`0UI_Ok6b0UI]Ol6b0UI]Ok6d0VIZOk6e0VIZOj6g0VIXOk6h0UIWOk6i0VIVOk6j0UIUOk6l0UISOl6l0UISOk6n0UIQOl6o0TIPOl6P1UIoNl6Q1TInNl6R1UImNk6T1VIjNk6V1UIiNk6W1VIhNj6Y1VIfNk6Z1UIeNk6[1VIdNk6\\1UIcNk6]1VIbNj6_1VI`Nk6`1UI_Nk6a1VI^Nj6c1VI\\Nk6d1UI[Nk6e1WIYNi6h1WIWNj6i1VIVNj6j1WIUNj6k1VITNj6l1WISNi6n1WIQNj6o1VIPNj6P2WIoMi6R2oHUNR7k1cH_N]7a1YHiNg7X1PHPOQ8P1gGWOY8i0aG]O`8c0XGDh8<QGKo86jF0W90aF7_9I[F=e9DSFc0n9W2001O00001O00001O00001O001OO100O100O100O100O100VLQKbMP5^2QKaMo4^2SK`Mn4_2UK_Mk4`2WKiJe0k0U4[4YKaJT1i0c3e4[K[JZ1m0\\3g4]KSJa1S1R3i4eMTK\\2l4gMQKY2n4jMoJW2P5mMmJS2R5QNjJP2U5SNiJm1V5WNfJj1Y5YNeJg1[5\\NaJe1^5_N_Ja1`5bN]J_1b5eN[J[1e5P400O1000000O10000O10000O10000O1000000O10000O10000O1000000O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1000000000000O1000000000000O1000000000000O10000000000O2O0001O1O1O001O1O1O0000001O0000001O000000001O00000mNT1]Lb3]LiN"}, "label": "a women feeding animal"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a plaid shirt reaching out\"."}], "task": "refering", "answer_template": "The \"a woman in a plaid shirt reaching out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 219, 220, 221, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 335, 336, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.54484375, 0.3004583333333333, 1.0, 0.9887916666666666], "iscrowd": 0, "area": 57847.39555, "rle": {"size": [480, 640], "counts": "PlS53j>5K5K5L5J5K5I7iFiN^4Z1`KiN_4X1]KlNb4U1[KmNf4S1WKPOh4P1UKSOk4n0RKUOn4k0nJYOQ5g0mJ[OS5f0iJ^OW5b0fJ@Z5a0bJC^5<aJF^5;aJE_5<_JFa59_JGa5:^JGa5:^JFc5:[JHd58\\JIc58\\JHe58YJJf57YJIh56XJKg56WJLh54XJLi54VJMi54UJMk53UJNk52TJOk52TJNl52SJ0m50RJ0n51QJ0o5OPJ2P6OoI1R6NnI2S6NkI3U6NjI2W6MiI3W6NhI2Y6NeI3[6MeI3\\6McI3]6MbI4_6L`I4`6M_I3b6L]I5c6L\\I4e6LZI4f6LZI4g6LWI5i6LWI3j6LWI3i6NWI1j6NWI1i60WIOj61VINj62WIMj63VILj65VIJk65WIIi68WIGj68WIGi6:WIEj6;VIDk6;VIDj6=VIBk6>UIAk6?VI@k6`0UI_Ok6b0UI]Ol6b0UI]Ok6d0VIZOk6e0VIZOj6g0VIXOk6h0UIWOk6i0VIVOk6j0UIUOk6l0UISOl6l0UISOk6n0UIQOl6o0TIPOl6P1UIoNl6Q1TInNl6R1UImNk6T1VIjNk6V1UIiNk6W1VIhNj6Y1VIfNk6Z1UIeNk6[1VIdNk6\\1UIcNk6]1VIbNj6_1VI`Nk6`1UI_Nk6a1VI^Nj6c1VI\\Nk6d1UI[Nk6e1WIYNi6h1WIWNj6i1VIVNj6j1WIUNj6k1VITNj6l1WISNi6n1WIQNj6o1VIPNj6P2WIoMi6R2oHUNR7k1cH_N]7a1YHiNg7X1PHPOQ8P1gGWOY8i0aG]O`8c0XGDh8<QGKo86jF0W90aF7_9I[F=e9DSFc0n9W2001O00001O00001O00001O001OO100O100O100O100O100VLQKbMP5^2QKaMo4^2SK`Mn4_2UK_Mk4`2WKiJe0k0U4[4YKaJT1i0c3e4[K[JZ1m0\\3g4]KSJa1S1R3i4eMTK\\2l4gMQKY2n4jMoJW2P5mMmJS2R5QNjJP2U5SNiJm1V5WNfJj1Y5YNeJg1[5\\NaJe1^5_N_Ja1`5bN]J_1b5eN[J[1e5P400O1000000O10000O10000O10000O1000000O10000O10000O1000000O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1000000000000O1000000000000O1000000000000O10000000000O2O0001O1O1O001O1O1O0000001O0000001O000000001O00000mNT1]Lb3]LiN"}, "label": "a woman in a plaid shirt reaching out"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl feeding a sheep\"."}], "task": "refering", "answer_template": "The \"a little girl feeding a sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 329, 330, 331, 332, 352, 353, 354, 355, 375, 376, 377], "bbox": [0.27421875, 0.43695833333333334, 0.51190625, 0.9955208333333334], "iscrowd": 0, "area": 25015.156900000013, "rle": {"size": [480, 640], "counts": "bhb2<V=GPDe0h;GgC\\1[;POUDY1f;T1L5K4L5K4M4K4L5L3L5K5L3L7J6I7J;E2N2N1O2N2N1N3N1O2M3N1N3N2M2O2N2M2O2N2N2N1O2N3PIVId5n6RJXIn5k6iI[IW6i6_I]Ia6a7000000003M0O2O1O001O1O001O001O001O001O001O001O001O001O002N2N3M3M2N3Me0[O6J5J2O1O001O1O001O2N1O1N3SJQGc5Z9K5L3L2N2N2O1N2N2N2N2O1I7E;L4LVOULgFg3\\9\\LbF]3c9fL\\FX1^OfNj;5fDT1h;mNWDS1i;POSDP1m;Q110O0100O010O2N100O2O0O1O101N2O1N2N2O2M2O1N3N1N2N3N1N2O2L3L5L3L5K4M4Kf0ZO^na4"}, "label": "a little girl feeding a sheep"}]}
{"id": 578519, "image": "COCO_train2014_000000578519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl feeding sheep\"."}], "task": "refering", "answer_template": "The \"a girl feeding sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 329, 330, 331, 332, 352, 353, 354, 355, 375, 376, 377], "bbox": [0.27421875, 0.43695833333333334, 0.51190625, 0.9955208333333334], "iscrowd": 0, "area": 25015.156900000013, "rle": {"size": [480, 640], "counts": "bhb2<V=GPDe0h;GgC\\1[;POUDY1f;T1L5K4L5K4M4K4L5L3L5K5L3L7J6I7J;E2N2N1O2N2N1N3N1O2M3N1N3N2M2O2N2M2O2N2N2N1O2N3PIVId5n6RJXIn5k6iI[IW6i6_I]Ia6a7000000003M0O2O1O001O1O001O001O001O001O001O001O001O001O002N2N3M3M2N3Me0[O6J5J2O1O001O1O001O2N1O1N3SJQGc5Z9K5L3L2N2N2O1N2N2N2N2O1I7E;L4LVOULgFg3\\9\\LbF]3c9fL\\FX1^OfNj;5fDT1h;mNWDS1i;POSDP1m;Q110O0100O010O2N100O2O0O1O101N2O1N2N2O2M2O1N3N1N2N3N1N2O2L3L5L3L5K4M4Kf0ZO^na4"}, "label": "a girl feeding sheep"}]}
{"id": 365527, "image": "COCO_train2014_000000365527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle being ridden by a man\"."}], "task": "refering", "answer_template": "The \"a bicycle being ridden by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 117, 119, 128, 129, 130, 131, 132, 133, 142, 143, 144, 145, 146, 149, 156, 157, 158, 159, 160, 162, 163, 170, 171, 172, 173, 174, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 225, 226, 227, 228, 229, 230, 231, 232, 233, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 267, 268, 269, 270, 282, 283], "bbox": [0.11970443349753696, 0.34709375, 0.7555665024630543, 0.8930624999999999], "iscrowd": 0, "area": 42113.623599999984, "rle": {"size": [640, 406], "counts": "nRo0V1`b0>B>I7I7J6I7I4L5L4K4L5L4K4L5L4K4L4N2M3fM`KPEb4n:`KQEb4l:`KREc4e:eKYE^4^:kK`EW4W:QLgER4Q:ULnEm3b9bL\\F`3[8iMcGZ2Z8iMdGX2[8iMcGY2\\8hMbGZ2]8gMbG[2\\8gMaG[2]8gMaG[2^8fMaG\\2\\8fMcG[2\\8fMcG[2\\8fMdG[2Z8fMeG[2Z8fMeG[2Z8fMfGZ2Z8fMeG\\2Y8eMfG\\2Z8dMfG\\2Y8eMfG\\2Z8dMeG^2Y8cMfG^2Z8bMfG^2Y8cMfG^2Y8cMfG^2Z8bMfG^2Y8cMfG]2[8cMeG\\2[8eMdG\\2\\8dMcG\\2]8eMcG[2]8eMbG[2^8fMbGZ2]8gMbGY2_8gMYGXLCQ6T9hMUG[LGm5T9hMSG]LHk5V9hMRG^LGj5W9hMQGd2P9\\MPGc2Q9]MnFd2R9]MmFb2T9^MlFb2S9_MlFa2U9_MkFa2U9_MjFa2W9_MiF`2Y9_MfFb2Z9^MfFa2[9_MeFa2[9_MeF`2\\9`McF`2^9`MbF_2_9aMaF_2`9`M`F_2a9aM_F^2b9bM]F^2d9cM[F\\2f9dMZF[2g9fMXFY2i9gMWFX2j9iMTFW2n9l31O1O100N2N2N2N2M4M2N2N2N3M2M4M2M4M2M4L4L3M4L3M4M2QL_D\\Oe;a0`DYOc;e0cDTOa;i0dDPO`;R1aDhNc;Y1^D`Ng;a1[DXNj;g1YDSNj;n1XDmMj;T2XDfMl;Z2VDaMl;`2cDjLc;W3oDPLW;Q4]2ROU@eMj?V2]@iMb?S2c@mM]?n1i@QNV?k1o@UNP?g1UAYNj>c1\\A\\Nc>`1bA`N]>\\1hAdNW>W1oAhNP>U1UBkNi=Q1]BnNmNKW=T1QDPO`N4^=h0XDSOVN:a=>_DWOkMa0d=4gDZOaMg0f=KoD]OWMl0i=CWEP1g:nN^ER1`:jNfEU1Y:hNlEV1T:gNQFW1o9fNTF[1k9bNXF\\1i9cNUF`1l9_NSFb1n9^NQFc1n9^NQFb1n9`NQF`1o9aNQF_1o9bNoE^1Q:cN`E@SMn1]=[NYBLU30PMj1c=YNYBMR34mLi1g=VN[BMo2Z2g:iMZBMl2]2i:fM\\BMj2]2k:eM\\BNh2P1[MOa=SO]BMf2`0QN<l<WO^BMR3MmMl0c<ZO^BM^4g0R9\\OaBMa4c0n8@aBMe4?j8CbBNh4:f8IbBMm42d80`BNmb03S]OMmb02T]ONlb02T]ONkb03U]OMkb02V]ONjb02V]ONjb02V]ONib02X]ONhb02X]OMjb01X]OMib03b0O10Vf7LaYH`0I6L5o\\OCVb0V1XM]NoBe1o<`NkBc1R=cNiB`1T=eNgB]1W=hN`ADNg1`>iNZAI1a1a>2YA1d>f2N2O1O2O0O1O1O1O2O0O100O1O010O01O2N2N2N2N2N3M5K5J7QLe@o2^?kLi@P3[?nLh@n2[?QMf@k2^?TMc@h2b?VM_@e2g?YM[@b2f`0J6H7K8dNV^O8Rb0_Oi^OG^a01c^ONgb010000O010O1000O014L\\_m1"}, "label": "a bicycle being ridden by a man"}]}
{"id": 365527, "image": "COCO_train2014_000000365527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man with purple gloves and long sleeved shirt riding a bicycle\"."}], "task": "refering", "answer_template": "The \"an older man with purple gloves and long sleeved shirt riding a bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 34, 35, 36, 47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 78, 79, 80, 88, 89, 90, 91, 92, 93, 94, 101, 102, 103, 104, 105, 106, 107, 108, 115, 117, 118, 119, 120, 121, 132, 133, 134, 135, 136, 146, 147, 148, 149, 160, 161, 162, 163, 173, 174, 175, 176, 187, 188, 189, 190, 202, 203, 204, 216, 217, 218, 230, 231, 232, 245, 246], "bbox": [0.24152709359605912, 0.055640625, 0.7467980295566502, 0.7754531250000001], "iscrowd": 0, "area": 42299.82960000003, "rle": {"size": [640, 406], "counts": "R_m11lc02O2N2N2N2N2M2O2N2N2N2N1N3N2N2N2N1O2M3N2O1N1O2N2N2N2O1N2N2N2N3M2O1N2N2N3M2O1N2N3M2N2N2O1N3M2O1N2O2M2XCoL`8R3_GPM^8R3bGnL]8S3bGoL[8T3dGmLZ8T3eGmLY8U3fGmLX8T3hGmLU8V3iGkLV8V3jGkLS8W3lGkLR8V3mGkLQ8W3oGjLo7X3oGiLo7Y3QHhLm7Y3RHiLk7Y3UHgLj7[3THgLj7[3THgLk7[3RHfLm7\\3PHgLo7Z3nGiLQ8Y3kGiLU8X3iGjLU8X3hGjLX8X3dGkL[8V3bGmL\\8U3aGmL_8U3^GmLa8T3\\GoLb8S3[GoLe8S3WGPMh8Q3UGRMi8P3TGRMl8P3SGPMl8Q3TGnLk8T3VGkLi8V3WGjLh8V3YGiLf8Y3ZGgL]NfNT9c4`HgLVN[Oo7mN\\HQ5`1fLTNK^7eNjHk4f1dLPNM]7jNjHe4j1dLmMOZ7POkH]4o1cLkM1W7XOiHU4V2bLhM2X7]OeHo3Y1mJEe1SO4X7e4mIdIIb1QO6X7e4lIdIMa1mN7Z7d4iIfI3]1iN;Y7c4iIgI6Y3P6P3iIhI9W3m5R3hIiI<U3j5R3iIjI?S3g5S3hIlIc0o2f5U3eInIg0j2^OhLT5`6fJoIj0g2ZOnLT5\\6gJPJm0d2VOTMU5X6fJRJP1`2RO[MV5S6fJTJT1]2mN`MW5o5gJVJV1Y2jNgMW5i5hJYJX1U2gNlMX5g5jJXJX1R2cNSNZ5c5mJXJU1Q2bNWNY5b5RKVJS1o1`N\\NZ5`5TKTJR1o1]NaN]5[5VKUJP1m1[NfN_5W5gJ_I3g0[1l1ZNjNb5R5eJdI4d0g0]2mN\\Nc5n4dJhI5`0a0Y4f4^1cJlI6=:_4n4V1aJPJ8;3e4S5P1`JSJ:8Lk4Z5i0_JWJ;4FR5a5a0]J\\J<o5W5D\\J_J>l5V5D\\JbJ>i5W5CZJgJ?e5W5CYJkJ?b5X5d0iJ[OX5e0gJ[OY5f0eJZO\\5f0dJYO^5g0`JZO`5f0`JYOa5h0]JYOd5f0\\JYOf5g0XJYOj5f0VJZOl5e0RJ[OQ6c0oI]OR6c0lI]OW6b0gI^O[6a0eI_O]6P72O2M2oB^Ia<e6VCbIj<j6O0bCmH_:0eFd7T9eHfF^7U9kHaF\\7X9mH`FX7Z9PI_FS7^9SI[FQ7a9VIYFn6b9j1mMlE]IMGZ:a6XF^IGKT:`6`F[ID0P:b6bFTIG3k9g6\\2N3M2N2N2N3N1N2N2N3M2N2N3M2N2O1N3N1O1N3N3L4M2M4L3M4aNUAcMn>Z2WAaMm>[2YA`Mj>\\2\\A^Mg>_2^A\\Mf>`2`AZMf>`2`AZMj>\\2[A_Mn>X2f1F9G9G8H8H8H8H8HQVP2"}, "label": "an older man with purple gloves and long sleeved shirt riding a bicycle"}]}
{"id": 365527, "image": "COCO_train2014_000000365527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old man riding a bike that has a purple glove on his left hand\"."}], "task": "refering", "answer_template": "The \"an old man riding a bike that has a purple glove on his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 34, 35, 36, 47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 78, 79, 80, 88, 89, 90, 91, 92, 93, 94, 101, 102, 103, 104, 105, 106, 107, 108, 115, 117, 118, 119, 120, 121, 132, 133, 134, 135, 136, 146, 147, 148, 149, 160, 161, 162, 163, 173, 174, 175, 176, 187, 188, 189, 190, 202, 203, 204, 216, 217, 218, 230, 231, 232, 245, 246], "bbox": [0.24152709359605912, 0.055640625, 0.7467980295566502, 0.7754531250000001], "iscrowd": 0, "area": 42299.82960000003, "rle": {"size": [640, 406], "counts": "R_m11lc02O2N2N2N2N2M2O2N2N2N2N1N3N2N2N2N1O2M3N2O1N1O2N2N2N2O1N2N2N2N3M2O1N2N2N3M2O1N2N3M2N2N2O1N3M2O1N2O2M2XCoL`8R3_GPM^8R3bGnL]8S3bGoL[8T3dGmLZ8T3eGmLY8U3fGmLX8T3hGmLU8V3iGkLV8V3jGkLS8W3lGkLR8V3mGkLQ8W3oGjLo7X3oGiLo7Y3QHhLm7Y3RHiLk7Y3UHgLj7[3THgLj7[3THgLk7[3RHfLm7\\3PHgLo7Z3nGiLQ8Y3kGiLU8X3iGjLU8X3hGjLX8X3dGkL[8V3bGmL\\8U3aGmL_8U3^GmLa8T3\\GoLb8S3[GoLe8S3WGPMh8Q3UGRMi8P3TGRMl8P3SGPMl8Q3TGnLk8T3VGkLi8V3WGjLh8V3YGiLf8Y3ZGgL]NfNT9c4`HgLVN[Oo7mN\\HQ5`1fLTNK^7eNjHk4f1dLPNM]7jNjHe4j1dLmMOZ7POkH]4o1cLkM1W7XOiHU4V2bLhM2X7]OeHo3Y1mJEe1SO4X7e4mIdIIb1QO6X7e4lIdIMa1mN7Z7d4iIfI3]1iN;Y7c4iIgI6Y3P6P3iIhI9W3m5R3hIiI<U3j5R3iIjI?S3g5S3hIlIc0o2f5U3eInIg0j2^OhLT5`6fJoIj0g2ZOnLT5\\6gJPJm0d2VOTMU5X6fJRJP1`2RO[MV5S6fJTJT1]2mN`MW5o5gJVJV1Y2jNgMW5i5hJYJX1U2gNlMX5g5jJXJX1R2cNSNZ5c5mJXJU1Q2bNWNY5b5RKVJS1o1`N\\NZ5`5TKTJR1o1]NaN]5[5VKUJP1m1[NfN_5W5gJ_I3g0[1l1ZNjNb5R5eJdI4d0g0]2mN\\Nc5n4dJhI5`0a0Y4f4^1cJlI6=:_4n4V1aJPJ8;3e4S5P1`JSJ:8Lk4Z5i0_JWJ;4FR5a5a0]J\\J<o5W5D\\J_J>l5V5D\\JbJ>i5W5CZJgJ?e5W5CYJkJ?b5X5d0iJ[OX5e0gJ[OY5f0eJZO\\5f0dJYO^5g0`JZO`5f0`JYOa5h0]JYOd5f0\\JYOf5g0XJYOj5f0VJZOl5e0RJ[OQ6c0oI]OR6c0lI]OW6b0gI^O[6a0eI_O]6P72O2M2oB^Ia<e6VCbIj<j6O0bCmH_:0eFd7T9eHfF^7U9kHaF\\7X9mH`FX7Z9PI_FS7^9SI[FQ7a9VIYFn6b9j1mMlE]IMGZ:a6XF^IGKT:`6`F[ID0P:b6bFTIG3k9g6\\2N3M2N2N2N3N1N2N2N3M2N2N3M2N2O1N3N1O1N3N3L4M2M4L3M4aNUAcMn>Z2WAaMm>[2YA`Mj>\\2\\A^Mg>_2^A\\Mf>`2`AZMf>`2`AZMj>\\2[A_Mn>X2f1F9G9G8H8H8H8H8HQVP2"}, "label": "an old man riding a bike that has a purple glove on his left hand"}]}
{"id": 234457, "image": "COCO_train2014_000000234457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cat that is sitting behind two other cats\"."}], "task": "refering", "answer_template": "The \"an orange cat that is sitting behind two other cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 99, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 144, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 217], "bbox": [0.1328125, 0.1419375, 0.54296875, 0.5976666666666667], "iscrowd": 0, "area": 21759.375, "rle": {"size": [480, 640], "counts": "omW14e>9G9N2N1O2N2N2N2N3M4MU1jN2N2N2O1O1N3N1O00O001O01O000O101O00000O2O0000O1O1O001N2O1O1O001N2O1O1O10O10O10000O10O1000O1000000O0100000O10O1O0O2O1O101N101N2O001N101N2O0O2O1N101N10100O00100O0010O0100005K5L4L4L4L4K5L1O00O0100001N2O1O2M2O1O1O2M2O1O0O01N01O1O1O1O1O001O1N2O1O1O001O1O1O1O1O1O001O1O2N3M3M2N3M3M2N3M3M2N3M3M2N2NO0010O01O1O0010O01O001O10O011N2O001N101N101N101O0O2O0O2O1O0O2O0ORLjEU3U:jLPFR3P:mLUFP3i9PM[Fm2e9RM`Fk2_9TMeFi2[9VMjFf2V9YMoFb2R9]MRG^2P9cMSGX2n8iMWGP2j8QN]Ge1e8[NbG]1_8dNhGS1Y8nNmGi0U8XORH?o7BWH6j7K]HKe76aHB`7>hHYOY7h0_31O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O00O1O100O1O100O1O2O2M3M3N2M3M3M3L4M3L3N3L4M3LUVY4"}, "label": "an orange cat that is sitting behind two other cats"}]}
{"id": 234457, "image": "COCO_train2014_000000234457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a orange cat is laying on a bed with his black and white cat friend\"."}], "task": "refering", "answer_template": "The \"a orange cat is laying on a bed with his black and white cat friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 352, 353, 354, 355, 357, 358, 374, 375, 376, 377], "bbox": [0.288703125, 0.2903125, 1.0, 0.9849374999999999], "iscrowd": 0, "area": 85440.55045, "rle": {"size": [480, 640], "counts": "VUg2=b>6J1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O2N100O100O100O100O1O100O100O100O100O10O00001O001[O^ClNa<T1dCgN]<X1gCdNY<]1jC_NW<`1mC\\NS<d1e00O02O0O1O2O0O1O2O0OdHeNU1Z1kNjNGI`K\\1h4nNDHdKY1g4SO@HhKU1g4WO[OHnKo0g4\\OVOHSLk0g4AQOGWLh0h4ClNH]Lc0i4FeNKaL?k4H^NMfL;l4JZNMkL8l4LTN0oL4n4NmM1VM0n40hM3YMMo47[M2fMGP5?kL0VN@P5g0\\L0dNXOQ5P1kKOUOoNQ5[1[KLEhNQ5Y5QKeJP5Z5QKeJP5Z5QKeJo4\\5QKcJP5\\5QKcJP5\\5RKbJo4]5RKbJn4^5SKaJn4^5SK`Jo4_5RK`Jo4`5RK^Jn4b5SK]Jn4b5SK]Jn4b5QK_JP5`5PK`JP5`5PK`JQ5`5mJaJT5^5lJbJU5]5jJcJW5]5iJbJY5]5fJbJ]5_5aJ_Jb5b5[J]Jg5d5WJ[Jl5f5QJYJR6h5kIYJV6h5gIZJX6h5eIYJ\\6g5bIZJ_6g5^I[Jb6f5[I[Je6g5XI[Jh6g5TIZJm6i5nHXJR7l5iHVJT7P6gHQJX7m6M2N2O2M2N3M2O1N3M2O101O001O00001N100000000000000000O10000000O1000000000O1000000000000000O10O1000000000000000000O1000O10000000000000O100000O1O1O1O1O1O001O1O1O1O1O1O1O1O10O0100000000001O00000O2O0000001O000000001N1000001O1O2N2N2N2M3N2N2N2N2N2N2N2N2N2M2O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2M2O1O2N1O1O2N1O1O2N1O000000000000000000000000000O1000000000001O000000000000000O10000O1000000O1000000O10000O101O000O0100000O1000000O01000O1000000O0100000O1000000O01000O1000000O0100000O10000O10O1000O1000000O0100000O10000O10O1000O1000000O0100000O10000O10O1000O1000000O10O1000O1000000O01000O1000000O10O1000O1000000O0100000O1000O10O10000O0100000O0100kI"}, "label": "a orange cat is laying on a bed with his black and white cat friend"}]}
{"id": 234457, "image": "COCO_train2014_000000234457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cat to the right of a black and white cat\"."}], "task": "refering", "answer_template": "The \"an orange cat to the right of a black and white cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 352, 353, 354, 355, 357, 358, 374, 375, 376, 377], "bbox": [0.288703125, 0.2903125, 1.0, 0.9849374999999999], "iscrowd": 0, "area": 85440.55045, "rle": {"size": [480, 640], "counts": "VUg2=b>6J1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O2N100O100O100O100O1O100O100O100O100O10O00001O001[O^ClNa<T1dCgN]<X1gCdNY<]1jC_NW<`1mC\\NS<d1e00O02O0O1O2O0O1O2O0OdHeNU1Z1kNjNGI`K\\1h4nNDHdKY1g4SO@HhKU1g4WO[OHnKo0g4\\OVOHSLk0g4AQOGWLh0h4ClNH]Lc0i4FeNKaL?k4H^NMfL;l4JZNMkL8l4LTN0oL4n4NmM1VM0n40hM3YMMo47[M2fMGP5?kL0VN@P5g0\\L0dNXOQ5P1kKOUOoNQ5[1[KLEhNQ5Y5QKeJP5Z5QKeJP5Z5QKeJo4\\5QKcJP5\\5QKcJP5\\5RKbJo4]5RKbJn4^5SKaJn4^5SK`Jo4_5RK`Jo4`5RK^Jn4b5SK]Jn4b5SK]Jn4b5QK_JP5`5PK`JP5`5PK`JQ5`5mJaJT5^5lJbJU5]5jJcJW5]5iJbJY5]5fJbJ]5_5aJ_Jb5b5[J]Jg5d5WJ[Jl5f5QJYJR6h5kIYJV6h5gIZJX6h5eIYJ\\6g5bIZJ_6g5^I[Jb6f5[I[Je6g5XI[Jh6g5TIZJm6i5nHXJR7l5iHVJT7P6gHQJX7m6M2N2O2M2N3M2O1N3M2O101O001O00001N100000000000000000O10000000O1000000000O1000000000000000O10O1000000000000000000O1000O10000000000000O100000O1O1O1O1O1O001O1O1O1O1O1O1O1O10O0100000000001O00000O2O0000001O000000001N1000001O1O2N2N2N2M3N2N2N2N2N2N2N2N2N2M2O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2M2O1O2N1O1O2N1O1O2N1O000000000000000000000000000O1000000000001O000000000000000O10000O1000000O1000000O10000O101O000O0100000O1000000O01000O1000000O0100000O1000000O01000O1000000O0100000O10000O10O1000O1000000O0100000O10000O10O1000O1000000O0100000O10000O10O1000O1000000O10O1000O1000000O01000O1000000O10O1000O1000000O0100000O1000O10O10000O0100000O0100kI"}, "label": "an orange cat to the right of a black and white cat"}]}
{"id": 37847, "image": "COCO_train2014_000000037847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black hair and glasses drinking wine\"."}], "task": "refering", "answer_template": "The \"a man with black hair and glasses drinking wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 52, 53, 54, 55, 56, 71, 72, 73, 75, 76, 77, 78, 79, 94, 95, 96, 98, 99, 100, 101, 102, 116, 117, 118, 119, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374], "bbox": [0.0, 0.058374999999999996, 0.5134218749999999, 0.9845624999999999], "iscrowd": 0, "area": 78020.86435000002, "rle": {"size": [480, 640], "counts": "b73i>7J7H7J6I7J3L5L3L4M4K4M3L5M2O1O2M2O1O2N1O1O2N1O1O2N1N2O2mHWMl1j2RNZMk1g2TN\\Mi1e2UN_Mh1c2VN`Mg1a2WNcMf1^2YNeMd1]2YNgMd1Z2ZNjMd1V2[NmMb1T2]NoMa1Q2_NQN^1Q2aNPN^1P2bNQN\\1P2cNSN[1m1eNTNY1m1gNTNX1l1gNVNW1l1hNUNW1k1iNWNR1l1mNVN>^2BdM:^2FcM8^2GdM7^2HcM5_2KaM4`2L`M3a2L`M2b2N^M1c2O]M0d2O]MOf20ZMOg21YMNh21YMMi23WMLj24VMKk24VMJm25SMJn26RMIo27QMGQ38PMGQ39oLFR3:nLDU3:lLEi0ZLAQ4FDg0]LDo3EBg0aLDl3FBe0dLEj3GAa0gLIh3FA?iLLd3FC;lLOa3ED:lL2`3DC9oL3]3ED7PM4\\3ED6PM^OT4<lN5QM]OU4=kN5PM^OU4>kN3QM^OU4?iN3SM]OU4`0hN2SM^OV4`0gN1TM^OU4b0gN0TM]OV4b0gN0SM^OW4b0fNOTM^OW4c0dNOUM]OY4d0bNNVM]OX4f0bNLWM]OX4f0bNLVM^OY4f0aNKXM]OX4h0`NJYM]OX4i0^NKZM[OX4k0^NI[M[OX4l0]NH\\M[OX4l0]NH\\MZOY4n0[NG]MZOY4o0ZNF_MYOX4Q1XNFaMXOW4S1XNDbMXOW4S1XNDbMXOW4T1WNCcMXOW4U1VNBdMXOW4V1UNBeMVOV4Y1TNAgMUOV4Z1SN@hMTOW4[1RN@hMVOU4Z1SN_OiMWOT4Z1TN]OiMZOQ4Z1WNZOjM[Oo3[1XNXOiM^Oo3Y1ZNWOfMCo3V1[NVOeMFP4T1\\NTOcMKP4Q1^NRObMNm3S1bNnN_M2g3W1jNfN^M5b3Z1RO_N[M9\\3_1ZOVNZM=T3d1CRMTN[1R2j1P2WNnMj1R2VNmMk1R2VNnMi1S2XNkMi1T2XNkMi1U2XNiMi1W2WNiMi1V2YNhMg1Y2YNfMh1Z2XNeMi1Z2YNeMg1[2YNdMh1\\2YNbMh1]2YNcMf1^2]N^Md1a2_N\\Mb1d2aNXM`1h2cNUM]1j2fNSMZ1n2iNnLX1R3kNjLV1U3mNiLS1W3POeLQ1[3ROaLo0^3UO^Lk0c3WOWLm0i3VOPLn0o3VOjKn0V4TOdKP1\\4SO]KP1d4SOUKQ1j4SOoJQ1Q5QOiJS1W5PObJT1^5oN[JU1e5mNUJV1l5o3000000O010000000000O1000O1[OUGUKk8a4bG\\K]8Z4PHdKP8R4]HkKc7j3kHSLU7a3ZI\\Lf6X3hIfLX6T3l2VO]DUNg;j1aDjMd;V2c0O10O0100O100O100O100O100O100O100O100O100O010O100O100O100O1O100O100O100O100001O00001O00001N101O00001O00001O00001O00001O00001O00001O001O00001N10001O00001O00001O00001O001O2N1O1O2N1O2N1TG[M[5f2bJ]M]5e2_J]Ma5d2]J^MeMKn6i2ZK_MgMJm6i2YK^MjMKk6h2XK`MlMJj6g2WKbMnMJh6f2WKbMPNJg6e2TKeMUNHe6e2PKhMZNFc6c2nJlM^NCb6b2kJoMcNA`6b2gJRNhN_O^6`2eJVNlN\\Ol0XOn3W3ULYNQOZOh0]Ok1MdNU3B\\NVOXOd0Be16cNe2I`NZOUOa0H]1`0cNT20dN^ORO?LV1j0bNf15fNDQO:1P1S1aNV1<jNHnN77i0o4ROQLMkN4<b0o4XOnK3iNNb0<m4^OmK7fNKh05l4CkK<cNIl0Mm7:YGDZ;=e1000O10000O1000000O10000O1000000O10000O10000O01G9Fnba4"}, "label": "a man with black hair and glasses drinking wine"}]}
{"id": 37847, "image": "COCO_train2014_000000037847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white man wearing glasses and a checkered shirt is drinking wine out of a wine glass\"."}], "task": "refering", "answer_template": "The \"a white man wearing glasses and a checkered shirt is drinking wine out of a wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 52, 53, 54, 55, 56, 71, 72, 73, 75, 76, 77, 78, 79, 94, 95, 96, 98, 99, 100, 101, 102, 116, 117, 118, 119, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374], "bbox": [0.0, 0.058374999999999996, 0.5134218749999999, 0.9845624999999999], "iscrowd": 0, "area": 78020.86435000002, "rle": {"size": [480, 640], "counts": "b73i>7J7H7J6I7J3L5L3L4M4K4M3L5M2O1O2M2O1O2N1O1O2N1O1O2N1N2O2mHWMl1j2RNZMk1g2TN\\Mi1e2UN_Mh1c2VN`Mg1a2WNcMf1^2YNeMd1]2YNgMd1Z2ZNjMd1V2[NmMb1T2]NoMa1Q2_NQN^1Q2aNPN^1P2bNQN\\1P2cNSN[1m1eNTNY1m1gNTNX1l1gNVNW1l1hNUNW1k1iNWNR1l1mNVN>^2BdM:^2FcM8^2GdM7^2HcM5_2KaM4`2L`M3a2L`M2b2N^M1c2O]M0d2O]MOf20ZMOg21YMNh21YMMi23WMLj24VMKk24VMJm25SMJn26RMIo27QMGQ38PMGQ39oLFR3:nLDU3:lLEi0ZLAQ4FDg0]LDo3EBg0aLDl3FBe0dLEj3GAa0gLIh3FA?iLLd3FC;lLOa3ED:lL2`3DC9oL3]3ED7PM4\\3ED6PM^OT4<lN5QM]OU4=kN5PM^OU4>kN3QM^OU4?iN3SM]OU4`0hN2SM^OV4`0gN1TM^OU4b0gN0TM]OV4b0gN0SM^OW4b0fNOTM^OW4c0dNOUM]OY4d0bNNVM]OX4f0bNLWM]OX4f0bNLVM^OY4f0aNKXM]OX4h0`NJYM]OX4i0^NKZM[OX4k0^NI[M[OX4l0]NH\\M[OX4l0]NH\\MZOY4n0[NG]MZOY4o0ZNF_MYOX4Q1XNFaMXOW4S1XNDbMXOW4S1XNDbMXOW4T1WNCcMXOW4U1VNBdMXOW4V1UNBeMVOV4Y1TNAgMUOV4Z1SN@hMTOW4[1RN@hMVOU4Z1SN_OiMWOT4Z1TN]OiMZOQ4Z1WNZOjM[Oo3[1XNXOiM^Oo3Y1ZNWOfMCo3V1[NVOeMFP4T1\\NTOcMKP4Q1^NRObMNm3S1bNnN_M2g3W1jNfN^M5b3Z1RO_N[M9\\3_1ZOVNZM=T3d1CRMTN[1R2j1P2WNnMj1R2VNmMk1R2VNnMi1S2XNkMi1T2XNkMi1U2XNiMi1W2WNiMi1V2YNhMg1Y2YNfMh1Z2XNeMi1Z2YNeMg1[2YNdMh1\\2YNbMh1]2YNcMf1^2]N^Md1a2_N\\Mb1d2aNXM`1h2cNUM]1j2fNSMZ1n2iNnLX1R3kNjLV1U3mNiLS1W3POeLQ1[3ROaLo0^3UO^Lk0c3WOWLm0i3VOPLn0o3VOjKn0V4TOdKP1\\4SO]KP1d4SOUKQ1j4SOoJQ1Q5QOiJS1W5PObJT1^5oN[JU1e5mNUJV1l5o3000000O010000000000O1000O1[OUGUKk8a4bG\\K]8Z4PHdKP8R4]HkKc7j3kHSLU7a3ZI\\Lf6X3hIfLX6T3l2VO]DUNg;j1aDjMd;V2c0O10O0100O100O100O100O100O100O100O100O100O010O100O100O100O1O100O100O100O100001O00001O00001N101O00001O00001O00001O00001O00001O00001O001O00001N10001O00001O00001O00001O001O2N1O1O2N1O2N1TG[M[5f2bJ]M]5e2_J]Ma5d2]J^MeMKn6i2ZK_MgMJm6i2YK^MjMKk6h2XK`MlMJj6g2WKbMnMJh6f2WKbMPNJg6e2TKeMUNHe6e2PKhMZNFc6c2nJlM^NCb6b2kJoMcNA`6b2gJRNhN_O^6`2eJVNlN\\Ol0XOn3W3ULYNQOZOh0]Ok1MdNU3B\\NVOXOd0Be16cNe2I`NZOUOa0H]1`0cNT20dN^ORO?LV1j0bNf15fNDQO:1P1S1aNV1<jNHnN77i0o4ROQLMkN4<b0o4XOnK3iNNb0<m4^OmK7fNKh05l4CkK<cNIl0Mm7:YGDZ;=e1000O10000O1000000O10000O1000000O10000O10000O01G9Fnba4"}, "label": "a white man wearing glasses and a checkered shirt is drinking wine out of a wine glass"}]}
{"id": 37847, "image": "COCO_train2014_000000037847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older woman drinking wine\"."}], "task": "refering", "answer_template": "The \"an older woman drinking wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 317, 318, 319, 320, 321, 332, 333, 334, 335, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.46453125, 0.009000000000000001, 0.9983125000000002, 0.9864791666666667], "iscrowd": 0, "area": 85629.58695000001, "rle": {"size": [480, 640], "counts": "nb[42a>b0]Od0VDlNn8g1iF]Nj8U2mFoMQ9X2fFlMZ9Y2]FkMc9[2SFiMm9\\2jEhMV:]2aEgM_:]3O100000000000000O1000000000000O101O000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000VNkEiNU:Q1RFnNm9DmEL<?g9D\\FYO4S1_9CiG;W8DkG;U8DmG;S8DPH:P8ERH:m7EVH:j7EYH9g7F[H9e7F]H9c7F`H8`7GbH8^7GdH8^7EeH9b7^OaHa0g7VOYHk0n:0000000O100000000O10000000000O100000000O10000000000O100000000O10000000001N100000000O100000000O1000000O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100`HWNh1j1WNZNf1f1XN`Nc1a1\\NdN`1\\1_NiN\\1X1bNmN[1S1dNROW1o0gNWOT1j0kNZOR1f0mN_On0b0PODl0=ROGj0:eMbNVMZ1i1\\N:h1c0c1mNjL<d1d0h1lNiL;`1e0o1jNfL>[1d0W2iNcL>W1e0]2hNaL`0R1d0e2gN^L`0n0f0k2dN\\Lc0i0e0R3dNZLb0e0f0Y3bNWLd0a0f0`3`NULf0;f0g3`NRLf08g0m3]NQLh02g0U4\\NlKj00g0[4YNgKP1Ne0`4WNcKS1Nd0U7ZOlHd0V7\\OjHc0W7\\OjHb0Y7]OgHa0[7^OfH`0\\7@dH?^7_OcH?_7@bH>`7B`H<c7B^H=c7C]H;e7D\\H:g7EYH9i7FXH9i7FXH8k7GUH7m7HTH6n7JRH5P8IQH5Q8KoG4R8KoG4S8JnG5S8KmG5S8JnG5T8JlG5U8JlG6T8JlG5V8IkG6V8IkG6V8JjG6V8IkG6W8IiG6X8IiG7W8IiG6Y8HiG6X8IiG7W8IiG6Y8HhG7Y8IgG7Y8HhG8Y8GgG9Y8FhG:X8EiG;^;000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000b8^GZ6"}, "label": "an older woman drinking wine"}]}
{"id": 37847, "image": "COCO_train2014_000000037847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman drinking wine from a tall stem glass beside a man drinking while at the table\"."}], "task": "refering", "answer_template": "The \"woman drinking wine from a tall stem glass beside a man drinking while at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 317, 318, 319, 320, 321, 332, 333, 334, 335, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.46453125, 0.009000000000000001, 0.9983125000000002, 0.9864791666666667], "iscrowd": 0, "area": 85629.58695000001, "rle": {"size": [480, 640], "counts": "nb[42a>b0]Od0VDlNn8g1iF]Nj8U2mFoMQ9X2fFlMZ9Y2]FkMc9[2SFiMm9\\2jEhMV:]2aEgM_:]3O100000000000000O1000000000000O101O000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000VNkEiNU:Q1RFnNm9DmEL<?g9D\\FYO4S1_9CiG;W8DkG;U8DmG;S8DPH:P8ERH:m7EVH:j7EYH9g7F[H9e7F]H9c7F`H8`7GbH8^7GdH8^7EeH9b7^OaHa0g7VOYHk0n:0000000O100000000O10000000000O100000000O10000000000O100000000O10000000001N100000000O100000000O1000000O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100`HWNh1j1WNZNf1f1XN`Nc1a1\\NdN`1\\1_NiN\\1X1bNmN[1S1dNROW1o0gNWOT1j0kNZOR1f0mN_On0b0PODl0=ROGj0:eMbNVMZ1i1\\N:h1c0c1mNjL<d1d0h1lNiL;`1e0o1jNfL>[1d0W2iNcL>W1e0]2hNaL`0R1d0e2gN^L`0n0f0k2dN\\Lc0i0e0R3dNZLb0e0f0Y3bNWLd0a0f0`3`NULf0;f0g3`NRLf08g0m3]NQLh02g0U4\\NlKj00g0[4YNgKP1Ne0`4WNcKS1Nd0U7ZOlHd0V7\\OjHc0W7\\OjHb0Y7]OgHa0[7^OfH`0\\7@dH?^7_OcH?_7@bH>`7B`H<c7B^H=c7C]H;e7D\\H:g7EYH9i7FXH9i7FXH8k7GUH7m7HTH6n7JRH5P8IQH5Q8KoG4R8KoG4S8JnG5S8KmG5S8JnG5T8JlG5U8JlG6T8JlG5V8IkG6V8IkG6V8JjG6V8IkG6W8IiG6X8IiG7W8IiG6Y8HiG6X8IiG7W8IiG6Y8HhG7Y8IgG7Y8HhG8Y8GgG9Y8FhG:X8EiG;^;000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000b8^GZ6"}, "label": "woman drinking wine from a tall stem glass beside a man drinking while at the table"}]}
{"id": 37847, "image": "COCO_train2014_000000037847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine glass in woman ' s hands\"."}], "task": "refering", "answer_template": "The \"wine glass in woman ' s hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 233, 234, 235, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 263, 264, 265, 266, 267, 279, 280, 281], "bbox": [0.127515625, 0.33579166666666665, 0.76175, 0.7407083333333334], "iscrowd": 0, "area": 39906.469799999984, "rle": {"size": [480, 640], "counts": "afV18c>9H8G8I8G9G9H8G9I7L4K5L4K5L4K2O0O2O1N101N2O0O2O0O2O1N101N2O00100O010000O010O1000O010000O1N101N2O1N101N2O1N101N2O1N3N2M4L3L4M4L3M4K4M3M4L2N3L3N3aNoBj0T=SOTCe0m<[OZC<h<D`C3a<MfCKZ<6X100O10O1000000O0100000O10O1000O100000O01000000O10O1000O1000O10O1000000O010000000O10O1000O100000O01000000O1000000O1000000O2O00000O1000000O1000000O1000000O1000000N2N2N2O1N2N2O2M2N3N2M3M3N2M3M3N2M3M3M2O2M3N2M3M3N1N3N2M3M3N2M2O2M3M3N2M3N1N3N2M3M3N2N101N2N2O1N2N101N1O2O0O2N100O2N101N1O101N1O2O0O101N1O101O001N10001O001O00001O00O10O1000000000000000000O10000000O10000000000O010000000O100000000O10O100000O100000000O10O100000O10000000000O10O100000O100000000O100000000O100000000O100000000O100000000O100000000O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O1000O010000O100O1000O0102M6K5K5J5L5K5J5L5J6K5K4K6K5J8I6J6I7J7H7H8H8H_oV2"}, "label": "wine glass in woman ' s hands"}]}
{"id": 37847, "image": "COCO_train2014_000000037847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass in the right hand of the lady\"."}], "task": "refering", "answer_template": "The \"a wine glass in the right hand of the lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 233, 234, 235, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 263, 264, 265, 266, 267, 279, 280, 281], "bbox": [0.127515625, 0.33579166666666665, 0.76175, 0.7407083333333334], "iscrowd": 0, "area": 39906.469799999984, "rle": {"size": [480, 640], "counts": "afV18c>9H8G8I8G9G9H8G9I7L4K5L4K5L4K2O0O2O1N101N2O0O2O0O2O1N101N2O00100O010000O010O1000O010000O1N101N2O1N101N2O1N101N2O1N3N2M4L3L4M4L3M4K4M3M4L2N3L3N3aNoBj0T=SOTCe0m<[OZC<h<D`C3a<MfCKZ<6X100O10O1000000O0100000O10O1000O100000O01000000O10O1000O1000O10O1000000O010000000O10O1000O100000O01000000O1000000O1000000O2O00000O1000000O1000000O1000000O1000000N2N2N2O1N2N2O2M2N3N2M3M3N2M3M3N2M3M3M2O2M3N2M3M3N1N3N2M3M3N2M2O2M3M3N2M3N1N3N2M3M3N2N101N2N2O1N2N101N1O2O0O2N100O2N101N1O101N1O2O0O101N1O101O001N10001O001O00001O00O10O1000000000000000000O10000000O10000000000O010000000O100000000O10O100000O100000000O10O100000O10000000000O10O100000O100000000O100000000O100000000O100000000O100000000O100000000O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O1000O010000O100O1000O0102M6K5K5J5L5K5J5L5J6K5K4K6K5J8I6J6I7J7H7H8H8H_oV2"}, "label": "a wine glass in the right hand of the lady"}]}
{"id": 259028, "image": "COCO_train2014_000000259028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person hiking through the snow wearing a green parka\"."}], "task": "refering", "answer_template": "The \"a person hiking through the snow wearing a green parka\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 162, 163, 164, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 229, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353, 366, 367, 368, 369, 370, 383, 384, 385, 386, 387], "bbox": [0.4185416666666667, 0.37693750000000004, 0.8729375, 0.999625], "iscrowd": 0, "area": 49862.059299999986, "rle": {"size": [640, 480], "counts": "YQn33gc06J7I6L5O0O2O001N101N101O0O101N101O0O100O10000O10000O100O1000O010000O100O1000d0[O>C2M3N3M2M3N3M2M4K4L4L5J5L4L5J5Lg0YOo0QOn0ULeIVHj6\\7gISH_6h7gIQH_6i7gIQH`6i7eIPHb6j7dIoGc6l7bI^GT7\\8^2I6K5J4M1N3N1N2O2M2N2O2O001O0000000000001O000000000000001O000000000000001O01O0I7mM]HaGl7a8_HjFh7X9dHRFd7P:V12M2N2O2M2N2O2M2N[1fNO00001O00009G=C=C1O00001O00000000000000000000001O00000000000000000000001O00000000000000001O1O1O001mKUHoKl7`3dI`K]6]4jI^KV6`4RJZKo5c4YJoJP6o4VJkJl5R5\\JhJe5U5d4N3M3M3N2M2N3M3M3N1N3M3M3M3M2N3M3M3M3M2N3L4M3M2N3M3M3L4M6J7I7I7H8I7I7I7I7H8I7I7I7I7HogU1"}, "label": "a person hiking through the snow wearing a green parka"}]}
{"id": 259028, "image": "COCO_train2014_000000259028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in winter gear walking through a snowy hill\"."}], "task": "refering", "answer_template": "The \"a man in winter gear walking through a snowy hill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 162, 163, 164, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 229, 230, 231, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 297, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 349, 350, 351, 352, 353, 366, 367, 368, 369, 370, 383, 384, 385, 386, 387], "bbox": [0.4185416666666667, 0.37693750000000004, 0.8729375, 0.999625], "iscrowd": 0, "area": 49862.059299999986, "rle": {"size": [640, 480], "counts": "YQn33gc06J7I6L5O0O2O001N101N101O0O101N101O0O100O10000O10000O100O1000O010000O100O1000d0[O>C2M3N3M2M3N3M2M4K4L4L5J5L4L5J5Lg0YOo0QOn0ULeIVHj6\\7gISH_6h7gIQH_6i7gIQH`6i7eIPHb6j7dIoGc6l7bI^GT7\\8^2I6K5J4M1N3N1N2O2M2N2O2O001O0000000000001O000000000000001O000000000000001O01O0I7mM]HaGl7a8_HjFh7X9dHRFd7P:V12M2N2O2M2N2O2M2N[1fNO00001O00009G=C=C1O00001O00000000000000000000001O00000000000000000000001O00000000000000001O1O1O001mKUHoKl7`3dI`K]6]4jI^KV6`4RJZKo5c4YJoJP6o4VJkJl5R5\\JhJe5U5d4N3M3M3N2M2N3M3M3N1N3M3M3M3M2N3M3M3M3M2N3L4M3M2N3M3M3L4M6J7I7I7H8I7I7I7I7H8I7I7I7I7HogU1"}, "label": "a man in winter gear walking through a snowy hill"}]}
{"id": 340958, "image": "COCO_train2014_000000340958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a lime that is in focus\"."}], "task": "refering", "answer_template": "The \"this is a lime that is in focus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.28901869158878507, 0.4225, 1.0], "iscrowd": 0, "area": 68706.7228, "rle": {"size": [428, 640], "counts": "j4b8i41O100O100O1O100O100O1O100O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000000000000000001O0000000000000000000000001O001O00001O001O00001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1bLWJWOj5f0\\JVOe5h0_JUOb5i0cJSO^5k0gJQO[5l0iJoNZ5o0kJkNX5S1mJPMm6n2j1N3M2O1O2N1O2N1N2O2N1O2N1O1O2M2O2N1O1O2N1O1N3N2N3M2N3M2N3L4M2N3M2N3M2cN]Dl0X<J6I7Je[j4"}, "label": "this is a lime that is in focus"}]}
{"id": 340958, "image": "COCO_train2014_000000340958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ripe lime sits slightly behind a banana and slightly in front of an apple\"."}], "task": "refering", "answer_template": "The \"a ripe lime sits slightly behind a banana and slightly in front of an apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.28901869158878507, 0.4225, 1.0], "iscrowd": 0, "area": 68706.7228, "rle": {"size": [428, 640], "counts": "j4b8i41O100O100O1O100O100O1O100O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000000000000000001O0000000000000000000000001O001O00001O001O00001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1bLWJWOj5f0\\JVOe5h0_JUOb5i0cJSO^5k0gJQO[5l0iJoNZ5o0kJkNX5S1mJPMm6n2j1N3M2O1O2N1O2N1N2O2N1O2N1O1O2M2O2N1O1O2N1O1N3N2N3M2N3M2N3L4M2N3M2N3M2cN]Dl0X<J6I7Je[j4"}, "label": "a ripe lime sits slightly behind a banana and slightly in front of an apple"}]}
{"id": 340958, "image": "COCO_train2014_000000340958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red apple closest to the camera\"."}], "task": "refering", "answer_template": "The \"red apple closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246], "bbox": [0.46262499999999995, 0.16432242990654206, 1.0, 0.6937149532710281], "iscrowd": 0, "area": 58583.358049999995, "rle": {"size": [428, 640], "counts": "\\mk3k0]<n0RO8I6I8H7I8H7I8I6I8K4L4K2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O001N2O1N2O1N2O1N2O1N2O1N2O1N110O1O100O1O100O1O100O1O100O1O00100O1O100O1O100O1O100O1O100O001O100O1O100O1O100O1O100O1O100O100O101O0O10000O10000O10000O2N1O1O1O100000000000001O000O10000000000000001O0O1000000000000000001O0O100000000000000000001N100000000000000000001N10000000O1M3L4M3M3M3L3N3M4N1O1O1O1O1O1O2N1N2O1O100O1O2N1O1O1O1O1O1O2N1O1O1O1000001O0000000O10001O0000000000001O0000000000001O0000000000001O00000000001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O0O2O00001O001N10001O001O001N10001O001O000O2O001O001O1O001O001O0O2O001O001O1O001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O001O001O001O001O0^O`L"}, "label": "red apple closest to the camera"}]}
{"id": 340958, "image": "COCO_train2014_000000340958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red apple\"."}], "task": "refering", "answer_template": "The \"a red apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246], "bbox": [0.46262499999999995, 0.16432242990654206, 1.0, 0.6937149532710281], "iscrowd": 0, "area": 58583.358049999995, "rle": {"size": [428, 640], "counts": "\\mk3k0]<n0RO8I6I8H7I8H7I8I6I8K4L4K2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O001N2O1N2O1N2O1N2O1N2O1N2O1N110O1O100O1O100O1O100O1O100O1O00100O1O100O1O100O1O100O1O100O001O100O1O100O1O100O1O100O1O100O100O101O0O10000O10000O10000O2N1O1O1O100000000000001O000O10000000000000001O0O1000000000000000001O0O100000000000000000001N100000000000000000001N10000000O1M3L4M3M3M3L3N3M4N1O1O1O1O1O1O2N1N2O1O100O1O2N1O1O1O1O1O1O2N1O1O1O1000001O0000000O10001O0000000000001O0000000000001O0000000000001O00000000001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O0O2O00001O001N10001O001O001N10001O001O000O2O001O001O1O001O001O0O2O001O001O1O001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O001O001O001O001O0^O`L"}, "label": "a red apple"}]}
{"id": 340958, "image": "COCO_train2014_000000340958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red apple behind another red apple\"."}], "task": "refering", "answer_template": "The \"a red apple behind another red apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 89, 90, 91, 114], "bbox": [0.56934375, 0.08558411214953271, 1.0, 0.3108177570093458], "iscrowd": 0, "area": 7621.416700000003, "rle": {"size": [428, 640], "counts": "kdh41Z=001O1O0010O01O001O001O1O001O001O001O1O001O001O001O00100O001O001O001O001O1O0010O010O10O10O100O01000000O10000O0100000O1000000O10000O0100000O1000000O1000O01000000O10000O100000O01000000O10000O1000O10O1000000O10000O1000O10O2O00000O101O000000001O0001O01O0000000010O00000001O0001O0001O000001O01O0000000010O00000001O0001O0001O00000010O00000001O01O000001O000001O01O0000000010O000000010O00100O0100O0100O010O010O10O010O0100O010O010O10O10O10O010O0100O010O010O10O010O0100O0100O0100O010O010O100O101N100O100O100O101O0O100O100O100O2O0O100O10000O2O000O101O3M3L5L3MlK"}, "label": "a red apple behind another red apple"}]}
{"id": 340958, "image": "COCO_train2014_000000340958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two limes in a basket\"."}], "task": "refering", "answer_template": "The \"two limes in a basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 145, 146, 147, 148, 169, 170, 171, 193], "bbox": [0.02953125, 0.01997663551401869, 0.48809375, 0.5680373831775701], "iscrowd": 0, "area": 35760.77515000001, "rle": {"size": [428, 640], "counts": "aR81Y=2N1O2N2N2N1N3N2N2N1O2N2N2N1O2N2N2N1O2N2N2N101O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O2N100O101O0O100O101N10000O2O0O100O2O000O101N100O10001N100O101N10000O2O0O100O2O000O100O2O0O101O0O2O0O2O0O2O001N101N101N101O0O2O001O001O1N101O001O001O001O001O1O1O1O1O1O1O001N2O1O1O1O1O1O001O1O1O1O1O1O1O001O1N2O1O1O1O001O1O1O1O1O1O1O001O1O1O1O100O1O010O1O100O1O10O01O100O100O1O2O1N1O2O0O2N2O0O2N1O2N2N1O2N101N2N1O2N2N1O2N101N1ON2O1N2O1N2O1N2O1N2O1N2O1N2O2M2O2M2O2M3N1N3N1N3N2M2O2M2O4K6J7J5J7J2M3N2M3N2M3N2M3L5mL_FS2c:I7I7I7H8I7I7H8I7IZoX4"}, "label": "two limes in a basket"}]}
{"id": 283615, "image": "COCO_train2014_000000283615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown coffee table near a couch and a chair\"."}], "task": "refering", "answer_template": "The \"a brown coffee table near a couch and a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 172, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 243, 244, 245], "bbox": [0.490296875, 0.528, 0.70940625, 0.8485277777777779], "iscrowd": 0, "area": 7831.770199999999, "rle": {"size": [360, 640], "counts": "gb^31W;1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1O0O2O1O1O1O001N2O1O001O1O1N101O1O1O001N2O1O001O1THaNn5_1[IWOf6h0eHM[7P21O001N101O1N101O001N101O1N101O001N101O1N101O0O2O001O0O2O1O0OC>001O0O101O0O101O0O101O000O2O00001O001O01b0]O00N2K5L5KF>\\Oe0[Od0N3O0O101O001N101O0O101O001N10001K4L5K4KnUQ2"}, "label": "a brown coffee table near a couch and a chair"}]}
{"id": 283615, "image": "COCO_train2014_000000283615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown wooden living room table with a newspaper on top\"."}], "task": "refering", "answer_template": "The \"a light brown wooden living room table with a newspaper on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 172, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 243, 244, 245], "bbox": [0.490296875, 0.528, 0.70940625, 0.8485277777777779], "iscrowd": 0, "area": 7831.770199999999, "rle": {"size": [360, 640], "counts": "gb^31W;1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1O0O2O1O1O001O1N2O001O1O1O0O2O1O1O1O001N2O1O001O1O1N101O1O1O001N2O1O001O1THaNn5_1[IWOf6h0eHM[7P21O001N101O1N101O001N101O1N101O001N101O1N101O0O2O001O0O2O1O0OC>001O0O101O0O101O0O101O000O2O00001O001O01b0]O00N2K5L5KF>\\Oe0[Od0N3O0O101O001N101O0O101O001N10001K4L5K4KnUQ2"}, "label": "a light brown wooden living room table with a newspaper on top"}]}
{"id": 398305, "image": "COCO_train2014_000000398305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bed on the top bunk\"."}], "task": "refering", "answer_template": "The \"bed on the top bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84], "bbox": [0.12481641468682506, 0.10778125000000001, 0.9860043196544277, 0.204859375], "iscrowd": 0, "area": 18528.553099999997, "rle": {"size": [640, 463], "counts": "`[T11nc03M3M2N3M3M2N3L3N3M3M2NM3WOi0O1Ojd:6dZE>B>H8N2O1N2O1N20000000000000O100000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O1000000000000000000000000000000000000000000000000`0@o0QO\\d3"}, "label": "bed on the top bunk"}]}
{"id": 398305, "image": "COCO_train2014_000000398305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top bunk mattress above a girl sitting on the lower bunk\"."}], "task": "refering", "answer_template": "The \"the top bunk mattress above a girl sitting on the lower bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84], "bbox": [0.12481641468682506, 0.10778125000000001, 0.9860043196544277, 0.204859375], "iscrowd": 0, "area": 18528.553099999997, "rle": {"size": [640, 463], "counts": "`[T11nc03M3M2N3M3M2N3L3N3M3M2NM3WOi0O1Ojd:6dZE>B>H8N2O1N2O1N20000000000000O100000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O1000000000000000000000000000000000000000000000000`0@o0QO\\d3"}, "label": "the top bunk mattress above a girl sitting on the lower bunk"}]}
{"id": 398305, "image": "COCO_train2014_000000398305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheet on the bottom bunk\"."}], "task": "refering", "answer_template": "The \"a sheet on the bottom bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 242, 243, 244, 245, 246, 252, 253, 254, 258, 259, 260, 261, 262, 270, 271, 274, 275, 276, 277, 278, 279, 280, 281, 282, 284, 285, 288, 292, 293, 294, 295, 296, 297, 298, 299, 301, 302, 303, 305, 309, 310, 311, 312, 313, 318, 319], "bbox": [0.1391576673866091, 0.572921875, 1.0, 0.79215625], "iscrowd": 0, "area": 21939.034949999994, "rle": {"size": [640, 463], "counts": "XRY12mc02N2N1O2N2N2O1N2N2N2N2N2N25J8H8EeT87QkG8aNBR_Of0k`01Y^O7ea0T1N2W^OkMba0Y2100000O10000O1000000O100O1N101O1O1N2O1O1O100O1O1O1O1O1O001O1O100O1O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O10000000000001O000000001O000000001O00000O101O000000001O000000001aMi@3W?Kk@5U?In@6R?Io@7Q?GQA9o>ESA;n>BTA>l>@VA`0j>^OXAb0h>^OYAa0g>^OZAb0g>\\OZAd0f>[O[Ae0e>kNY@GR1^1e>jNlAV1T>hNoAW1Q>hNPBX1P>gNQBY1P>eNQB[1o=cNSB\\1n=cNSB]1m=bNTB^1l=aNVB^1j=`NXB`1T`000000001O0000000000001N1000000000001O000O1000001O00000000000O2O001O2N1O2N2N2N2N2M3N2N2NkhQ1_OcWnN3M3N2M3H7^Oc0N2O100O10O0100O100O10O0100O100O010O100O1000O0100001O1O2M2O1O1O1O2M2O1O1O1O1O2M2O2N2N2N2N2M3N2N2N2N2N2M3NP]21mbM3N1N3O100O010O100O10O10O100O10O0100O100O010N2O1N2N103L3N2M4M2M4M2N3L3N2M4M2N3L3N2M4M2N6K6I6J7J4K10O0010O010O00010O01O010gM`_OP1``0`N^@S1o2"}, "label": "a sheet on the bottom bunk"}]}
{"id": 398305, "image": "COCO_train2014_000000398305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed the lady is sitting on\"."}], "task": "refering", "answer_template": "The \"the bed the lady is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 242, 243, 244, 245, 246, 252, 253, 254, 258, 259, 260, 261, 262, 270, 271, 274, 275, 276, 277, 278, 279, 280, 281, 282, 284, 285, 288, 292, 293, 294, 295, 296, 297, 298, 299, 301, 302, 303, 305, 309, 310, 311, 312, 313, 318, 319], "bbox": [0.1391576673866091, 0.572921875, 1.0, 0.79215625], "iscrowd": 0, "area": 21939.034949999994, "rle": {"size": [640, 463], "counts": "XRY12mc02N2N1O2N2N2O1N2N2N2N2N2N25J8H8EeT87QkG8aNBR_Of0k`01Y^O7ea0T1N2W^OkMba0Y2100000O10000O1000000O100O1N101O1O1N2O1O1O100O1O1O1O1O1O001O1O100O1O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O10000000000001O000000001O000000001O00000O101O000000001O000000001aMi@3W?Kk@5U?In@6R?Io@7Q?GQA9o>ESA;n>BTA>l>@VA`0j>^OXAb0h>^OYAa0g>^OZAb0g>\\OZAd0f>[O[Ae0e>kNY@GR1^1e>jNlAV1T>hNoAW1Q>hNPBX1P>gNQBY1P>eNQB[1o=cNSB\\1n=cNSB]1m=bNTB^1l=aNVB^1j=`NXB`1T`000000001O0000000000001N1000000000001O000O1000001O00000000000O2O001O2N1O2N2N2N2N2M3N2N2NkhQ1_OcWnN3M3N2M3H7^Oc0N2O100O10O0100O100O10O0100O100O010O100O1000O0100001O1O2M2O1O1O1O2M2O1O1O1O1O2M2O2N2N2N2N2M3N2N2N2N2N2M3NP]21mbM3N1N3O100O010O100O10O10O100O10O0100O100O010N2O1N2N103L3N2M4M2M4M2N3L3N2M4M2N3L3N2M4M2N6K6I6J7J4K10O0010O010O00010O01O010gM`_OP1``0`N^@S1o2"}, "label": "the bed the lady is sitting on"}]}
{"id": 373730, "image": "COCO_train2014_000000373730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young man in glasses tapping glasses with a woman\"."}], "task": "refering", "answer_template": "The \"young man in glasses tapping glasses with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 155, 156, 157, 178, 179, 180, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 366], "bbox": [0.498046875, 0.34020833333333333, 1.0, 0.903375], "iscrowd": 0, "area": 30053.79505, "rle": {"size": [480, 640], "counts": "\\ie42m>3N1HMaA3_>MaA5]>KcA7Z>KeA6Z>JfA6Z>JfA6Y>KgA5Y>KgA5Y>KgA5Y>KgA5Y>KgA50HS>3mA50IR>2nA50IR>2nA5OJS>1nA5OJS>1nA5OJS>1nA6NIT>2mA5OJS>1nA5OJS>1nA5OJS>1mA60IS>?mAAS>f0001O00000000001O0001O000001O001O00010O001O000010O0001O001O0001O0000000000001O000000000000001O0000000001O000001O00000000000000001O00000000000000001O0O10000000000O10001O000O100000000O101O0000000000000000000000000000O100000001O000000000000000O010000O1000000O10000]NQOUEo0a:TORDO[1n0a:WOQDM]1k0^:EaE;[:IdE8n9SOkDi0W14m9:PFFo9`0lE@NYOg8[1XG]O0B^8[1YGSO8C^8[1ZGRO7DSOBo8k1eGPO7DVOKe8`1nGQO6EWO3\\8W1WHQO5EXO>S8l0^HQO7FXOg0i7b0hHQO5G[Og0g7a0hHRO5G\\Of0g7b0gHQO5G^Of0f7b0fHRO5G_Oe0e7c0gHQO4GAe0d7d0eHQO5GBd0c7e0fHPO4GCe0c7d0eHSO3DEe0c7e0dHSO2CHe0f0_Nl5W2cITO1BJe0d0_Nn5V2bIVO0@Me00^N83X6T2cIWOO^OOf0O`N52Z6V2aIVOO^O2d01`N03]6W2^IWON\\O6d01aNM3^6W2_I95lN1cNJ2a6X2\\I79lNOfNH0c6R5fIXLOSOZ6d4hIYLNSO[6b4hI[LMSO[6b4gI\\LNRO[6a4hI]LMRO[6a4hI]LMQO[6b4iI]LLQO[6b4iI^LKoN\\6c4iI_LKnN]6a4iIaLJmN^6b4hIaLJlN`6a4gIcLIlN`6c4eIaLJlNb6h4^I]L0kNc6l4XIYL5kNc6g6]IYId6f6\\IZId6f6\\IZId6g6[IYIf6f6ZIZIf6f6ZIZIg6e6YIZIh6f6YIXIh6h6XIWIj6h6VIXIj6h6WIVIk6i6UIVIl6j6UITIl6l6TISIn6m6QIRIR7l6oHRIT7l6lHTIW7Q1bH^48_JY7o0aHc45]J^7l0_Hf43]Ja7?ZHD4`52[Jd7;\\HF0c5Q910O001O010O1O00100O1O0ZOiFcKX9[4kFdKU9[4lFdKT9\\4mFcKT9\\4mFcKT9[4mFdKT9\\4lFdKU9[4jFeKX9Y4iFgKX9X4gFhKZ9X4fFgK\\9X4dFhK]9W4bFiK_9W4bFgK`9X4`FhKa9W4WFdK14OLd9[4\\FfK30L0e9Z4\\FgK80]9X4[FiK61_9U4[FjK52a9S4ZFlK32d9P4ZFPLO1h9n3YFRLN0j9m3WFVLLNn9k3VFcLj9\\3VFdLk9[3UFeLl9Z3TFfLm9Y3SFgLm9Y3SFhLm9W3SFiLn9V3RFjL5ZO]9k3]FlL6ZO]9i3]FmL6ZO^9i3[FmL6\\O_9f3[FoL5[Oa9e3ZFPM5\\Oa9c3[FPM4]Ob9b3ZFQM3^Od9`3YFQM4_Od9_3XFRM4_Oe9^3XFQM3Af9^3WFQM3Ah9\\3UFSM3Aj9Z3SFTM4Bk9X3QFVM3Cn9U3PFWM2CQ:T3mEXM3DR:R3kEZM2EU:o2iE\\M2EW:m2gE]M3FX:l2eE]M3F[:k2bE_M2G^:h2`E`M3H_:f2^EbM3Ha:d2\\EdM2He:b2ZEdM2Jf:a2WEeM3Jh:_2UEfM4Ki:]2SEhM3Ll:Z2QEjM3Ko:Y2oDjM3MP;X2lDkM4MR;V2jDmM3NU;S2hDnM4NW;R2eDPN4M[;P2bDRN2N`;n1]DSN4Md;m1XDVN4Lh;k1TDYN4Kl;i1PD[N4KQ<h1kC\\N4KU<f1gC_N3KZ<c1cCaN3L^<`1_CdNo<Y1QCfNS=X1mBgNV=V1=M3M3M4M2M3J6H8HaD"}, "label": "young man in glasses tapping glasses with a woman"}]}
{"id": 373730, "image": "COCO_train2014_000000373730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the white t - shirt holding glass\"."}], "task": "refering", "answer_template": "The \"a man in the white t - shirt holding glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 155, 156, 157, 178, 179, 180, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 366], "bbox": [0.498046875, 0.34020833333333333, 1.0, 0.903375], "iscrowd": 0, "area": 30053.79505, "rle": {"size": [480, 640], "counts": "\\ie42m>3N1HMaA3_>MaA5]>KcA7Z>KeA6Z>JfA6Z>JfA6Y>KgA5Y>KgA5Y>KgA5Y>KgA5Y>KgA50HS>3mA50IR>2nA50IR>2nA5OJS>1nA5OJS>1nA5OJS>1nA6NIT>2mA5OJS>1nA5OJS>1nA5OJS>1mA60IS>?mAAS>f0001O00000000001O0001O000001O001O00010O001O000010O0001O001O0001O0000000000001O000000000000001O0000000001O000001O00000000000000001O00000000000000001O0O10000000000O10001O000O100000000O101O0000000000000000000000000000O100000001O000000000000000O010000O1000000O10000]NQOUEo0a:TORDO[1n0a:WOQDM]1k0^:EaE;[:IdE8n9SOkDi0W14m9:PFFo9`0lE@NYOg8[1XG]O0B^8[1YGSO8C^8[1ZGRO7DSOBo8k1eGPO7DVOKe8`1nGQO6EWO3\\8W1WHQO5EXO>S8l0^HQO7FXOg0i7b0hHQO5G[Og0g7a0hHRO5G\\Of0g7b0gHQO5G^Of0f7b0fHRO5G_Oe0e7c0gHQO4GAe0d7d0eHQO5GBd0c7e0fHPO4GCe0c7d0eHSO3DEe0c7e0dHSO2CHe0f0_Nl5W2cITO1BJe0d0_Nn5V2bIVO0@Me00^N83X6T2cIWOO^OOf0O`N52Z6V2aIVOO^O2d01`N03]6W2^IWON\\O6d01aNM3^6W2_I95lN1cNJ2a6X2\\I79lNOfNH0c6R5fIXLOSOZ6d4hIYLNSO[6b4hI[LMSO[6b4gI\\LNRO[6a4hI]LMRO[6a4hI]LMQO[6b4iI]LLQO[6b4iI^LKoN\\6c4iI_LKnN]6a4iIaLJmN^6b4hIaLJlN`6a4gIcLIlN`6c4eIaLJlNb6h4^I]L0kNc6l4XIYL5kNc6g6]IYId6f6\\IZId6f6\\IZId6g6[IYIf6f6ZIZIf6f6ZIZIg6e6YIZIh6f6YIXIh6h6XIWIj6h6VIXIj6h6WIVIk6i6UIVIl6j6UITIl6l6TISIn6m6QIRIR7l6oHRIT7l6lHTIW7Q1bH^48_JY7o0aHc45]J^7l0_Hf43]Ja7?ZHD4`52[Jd7;\\HF0c5Q910O001O010O1O00100O1O0ZOiFcKX9[4kFdKU9[4lFdKT9\\4mFcKT9\\4mFcKT9[4mFdKT9\\4lFdKU9[4jFeKX9Y4iFgKX9X4gFhKZ9X4fFgK\\9X4dFhK]9W4bFiK_9W4bFgK`9X4`FhKa9W4WFdK14OLd9[4\\FfK30L0e9Z4\\FgK80]9X4[FiK61_9U4[FjK52a9S4ZFlK32d9P4ZFPLO1h9n3YFRLN0j9m3WFVLLNn9k3VFcLj9\\3VFdLk9[3UFeLl9Z3TFfLm9Y3SFgLm9Y3SFhLm9W3SFiLn9V3RFjL5ZO]9k3]FlL6ZO]9i3]FmL6ZO^9i3[FmL6\\O_9f3[FoL5[Oa9e3ZFPM5\\Oa9c3[FPM4]Ob9b3ZFQM3^Od9`3YFQM4_Od9_3XFRM4_Oe9^3XFQM3Af9^3WFQM3Ah9\\3UFSM3Aj9Z3SFTM4Bk9X3QFVM3Cn9U3PFWM2CQ:T3mEXM3DR:R3kEZM2EU:o2iE\\M2EW:m2gE]M3FX:l2eE]M3F[:k2bE_M2G^:h2`E`M3H_:f2^EbM3Ha:d2\\EdM2He:b2ZEdM2Jf:a2WEeM3Jh:_2UEfM4Ki:]2SEhM3Ll:Z2QEjM3Ko:Y2oDjM3MP;X2lDkM4MR;V2jDmM3NU;S2hDnM4NW;R2eDPN4M[;P2bDRN2N`;n1]DSN4Md;m1XDVN4Lh;k1TDYN4Kl;i1PD[N4KQ<h1kC\\N4KU<f1gC_N3KZ<c1cCaN3L^<`1_CdNo<Y1QCfNS=X1mBgNV=V1=M3M3M4M2M3J6H8HaD"}, "label": "a man in the white t - shirt holding glass"}]}
{"id": 373730, "image": "COCO_train2014_000000373730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table top where the people are seated\"."}], "task": "refering", "answer_template": "The \"the table top where the people are seated\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.0016875000000000002, 0.5704166666666667, 0.785390625, 0.9883958333333334], "iscrowd": 0, "area": 75361.5928, "rle": {"size": [480, 640], "counts": "al0Y1f=100O100O100O1O100O100O100O1O100O100O100O1O100O100O100L4F:G9G9O1O100O1O1O100O1O100O1O1O100O1O100000000000000000000000000000000000000000000000000001O001O001O001O001O001O000000O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O100O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000000000000000O100000000000000000000000000O100000000000000000000O100O1O100O100O100O100O100O100O1O100O100O100O100OkIZGR6j801O0000000000001O0000000000001O0000000000001O001O00001O00001O001O00001O001O000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001N1000000000000000000000000000000000000000000000000000d0\\O6J7I7I6J7I7I6J7I7I6J7I7I6J7I7I6J7I7I6J7I6J7I7I6IZho1"}, "label": "the table top where the people are seated"}]}
{"id": 373730, "image": "COCO_train2014_000000373730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table\"."}], "task": "refering", "answer_template": "The \"table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.0016875000000000002, 0.5704166666666667, 0.785390625, 0.9883958333333334], "iscrowd": 0, "area": 75361.5928, "rle": {"size": [480, 640], "counts": "al0Y1f=100O100O100O1O100O100O100O1O100O100O100O1O100O100O100L4F:G9G9O1O100O1O1O100O1O100O1O1O100O1O100000000000000000000000000000000000000000000000000001O001O001O001O001O001O000000O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O100O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000000000000000O100000000000000000000000000O100000000000000000000O100O1O100O100O100O100O100O100O1O100O100O100O100OkIZGR6j801O0000000000001O0000000000001O0000000000001O001O00001O00001O001O00001O001O000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001N1000000000000000000000000000000000000000000000000000d0\\O6J7I7I6J7I7I6J7I7I6J7I7I6J7I7I6J7I7I6J7I6J7I7I6IZho1"}, "label": "table"}]}
{"id": 373731, "image": "COCO_train2014_000000373731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ups truck that is in the middle lane\"."}], "task": "refering", "answer_template": "The \"the ups truck that is in the middle lane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245], "bbox": [0.496765625, 0.34158333333333335, 0.946765625, 0.8381944444444445], "iscrowd": 0, "area": 32216.07565, "rle": {"size": [360, 640], "counts": "PP`3P2g8b0^Ob0^Ob0F:L4M3O1O1N2O000O2O000O101O000O10001N10000O2O00000O2O000O10001N1000001N10000O2O0000000O2O0000001O0000001O0O1000001O0000001O00000O101O0000001O0000001O000000001O0000001O0000001O000000001O0000001OO1M3N2M3M3N2M3N2M3M3N2M3N2M3O1N2O1N2N2O1N2O1N2O1N2O1N2O1O1O1O100O100O1O100O1O100O100O1O010O1O100O10O01O100O10O100000O10O10000000000O1000O1000O10000000000000O1000000000000000000O1000000000000O100000000O100000000O1000000O100000001N100000000O101O0000000000001O000000004L8H8H5K2N1O1O1O2N1O1O1O2N1O1O1O2N1N2O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O3Mjg;"}, "label": "the ups truck that is in the middle lane"}]}
{"id": 373731, "image": "COCO_train2014_000000373731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black delivery truck is parked directly behind a red sign\"."}], "task": "refering", "answer_template": "The \"a black delivery truck is parked directly behind a red sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239], "bbox": [0.13271875, 0.39033333333333337, 0.43609375000000006, 0.83975], "iscrowd": 0, "area": 25411.5512, "rle": {"size": [360, 640], "counts": "_Qn03U;5K6J5J7J5Km0SOi0WO5K0O100000001O0O10000000000O1000001O3M2M4M2N3M2N1O0O2O1O001O1O0O2O00001O000O101O00001O0O10001O00000O2O0000001O0O101O00000O2O0000001N2O2N2N2N2M2O00001O0O101O00001N101O00000O100000000O100000000O10O100000O100000000O10000000O0100000000O10000000O010000000O0100000O10O100000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000001N4M2N2N2M3N3M2K5K5K5L4K6J5K?Aa0POegn3"}, "label": "a black delivery truck is parked directly behind a red sign"}]}
{"id": 373731, "image": "COCO_train2014_000000373731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black truck coming behind the white car in same lane\"."}], "task": "refering", "answer_template": "The \"black truck coming behind the white car in same lane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239], "bbox": [0.13271875, 0.39033333333333337, 0.43609375000000006, 0.83975], "iscrowd": 0, "area": 25411.5512, "rle": {"size": [360, 640], "counts": "_Qn03U;5K6J5J7J5Km0SOi0WO5K0O100000001O0O10000000000O1000001O3M2M4M2N3M2N1O0O2O1O001O1O0O2O00001O000O101O00001O0O10001O00000O2O0000001O0O101O00000O2O0000001N2O2N2N2N2M2O00001O0O101O00001N101O00000O100000000O100000000O10O100000O100000000O10000000O0100000000O10000000O010000000O0100000O10O100000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000001N4M2N2N2M3N3M2K5K5K5L4K6J5K?Aa0POegn3"}, "label": "black truck coming behind the white car in same lane"}]}
{"id": 373731, "image": "COCO_train2014_000000373731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver suv in front of ups trucks\"."}], "task": "refering", "answer_template": "The \"silver suv in front of ups trucks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.6483125000000001, 0.6113888888888889, 1.0, 1.0], "iscrowd": 0, "area": 28400.90235, "rle": {"size": [360, 640], "counts": "]Wb4c0d:10000VOKaF5o8Z1N2O1O100O1N2M3M3M3N2N2N2N2M3N2N2N2N2N2N2N2N2O1O1N2O1N2O1N2O1N2O1O1N2O1O1O100O1O100000000O1000000O100O100000000O100000000O100000000O100000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000001O000000000000000000000000001N100000000000000000000000001O00000000000000000000001O000000000000000000001[LVKj0hM"}, "label": "silver suv in front of ups trucks"}]}
{"id": 373731, "image": "COCO_train2014_000000373731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver suv with black roof rails\"."}], "task": "refering", "answer_template": "The \"silver suv with black roof rails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.6483125000000001, 0.6113888888888889, 1.0, 1.0], "iscrowd": 0, "area": 28400.90235, "rle": {"size": [360, 640], "counts": "]Wb4c0d:10000VOKaF5o8Z1N2O1O100O1N2M3M3M3N2N2N2N2M3N2N2N2N2N2N2N2N2O1O1N2O1N2O1N2O1N2O1O1N2O1O1O100O1O100000000O1000000O100O100000000O100000000O100000000O100000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000001O000000000000000000000000001N100000000000000000000000001O00000000000000000000001O000000000000000000001[LVKj0hM"}, "label": "silver suv with black roof rails"}]}
{"id": 455649, "image": "COCO_train2014_000000455649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left half of the sandwich on the plate\"."}], "task": "refering", "answer_template": "The \"the left half of the sandwich on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 201, 202, 203, 204, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 334, 335, 336, 337], "bbox": [0.08086601307189543, 0.39973856209150327, 0.36236928104575167, 0.7240196078431373], "iscrowd": 0, "area": 21956.18425, "rle": {"size": [612, 612], "counts": "UXn04lb06K4G9K5L4L4L3O2N2N2M3N5K2N1N3N1OO10O01O0001O7I3M3M3N3M2M3N2N2N2N1O2_O`0N3L3N3K4O2N1O20ON2N2N2N2N2O1O1O1O1O2N1O1O1O2N1O1O2N100O1O2N100O1O1QBXKa=W5O2M3M1O10O01O010O1O011N5L0O1O101O00000000N2O1N2N14M1O2N1O0O2O003M2N000O1000O10000O100O10000000000000001O00002M2O1O1O1O2N1O1O1eJdBl4`=mJeBQ5i=M2O2N2N3L5L001aKdAT4\\>kKgAR4Z>lKjAR4V>kKoAR4d>hNo@YNS?e1n@ZNT?c1n@[NW?a1j@^NY?^1i@`N_?Y1b@fNe?R1]@lNj?m0V@TOQ`0d0P@[OZ`0<e_OETa0Am^O?ka0100O10O0101N1010O1An]OKQb04S^OERb09Q^ODQb0:=O4L5JPfX7"}, "label": "the left half of the sandwich on the plate"}]}
{"id": 455649, "image": "COCO_train2014_000000455649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"big bread piece\"."}], "task": "refering", "answer_template": "The \"big bread piece\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 201, 202, 203, 204, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 334, 335, 336, 337], "bbox": [0.08086601307189543, 0.39973856209150327, 0.36236928104575167, 0.7240196078431373], "iscrowd": 0, "area": 21956.18425, "rle": {"size": [612, 612], "counts": "UXn04lb06K4G9K5L4L4L3O2N2N2M3N5K2N1N3N1OO10O01O0001O7I3M3M3N3M2M3N2N2N2N1O2_O`0N3L3N3K4O2N1O20ON2N2N2N2N2O1O1O1O1O2N1O1O1O2N1O1O2N100O1O2N100O1O1QBXKa=W5O2M3M1O10O01O010O1O011N5L0O1O101O00000000N2O1N2N14M1O2N1O0O2O003M2N000O1000O10000O100O10000000000000001O00002M2O1O1O1O2N1O1O1eJdBl4`=mJeBQ5i=M2O2N2N3L5L001aKdAT4\\>kKgAR4Z>lKjAR4V>kKoAR4d>hNo@YNS?e1n@ZNT?c1n@[NW?a1j@^NY?^1i@`N_?Y1b@fNe?R1]@lNj?m0V@TOQ`0d0P@[OZ`0<e_OETa0Am^O?ka0100O10O0101N1010O1An]OKQb04S^OERb09Q^ODQb0:=O4L5JPfX7"}, "label": "big bread piece"}]}
{"id": 37863, "image": "COCO_train2014_000000037863.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right - most umbrella at the restaurant\"."}], "task": "refering", "answer_template": "The \"right - most umbrella at the restaurant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 156, 179, 202], "bbox": [0.6268281250000001, 0.3219722222222222, 0.9774375, 0.67175], "iscrowd": 0, "area": 5089.289399999999, "rle": {"size": [360, 640], "counts": "X_]41W;0O10001N100O100O10000O101N10000O10000O10001N10000O10000O10000O2O000O10000O10000O2O000O1000O010000O10000O01000O10000O10O10O10000O10000O01000O10000O10O1000O10000O10O10O10000O10000O01000O10001N10000O10;Ee0ZOf0[Of0ZO>BN2M2O2M3N2fMZ20000001O000000010O0001L3K5L4L5K4L4L5K4LoD:k:6J6KO000000000000000000000010O00000000000000000001O0000000000000000000O2O000000000O10000000001O0O10000000000O1000001O000O1000000000000O101O0000000O2O00000O2O000O101O0O10001O0O1000Rk5"}, "label": "right - most umbrella at the restaurant"}]}
{"id": 37863, "image": "COCO_train2014_000000037863.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lighted umbrella sitting above a table in an outside diner along with other tables to the left\"."}], "task": "refering", "answer_template": "The \"a lighted umbrella sitting above a table in an outside diner along with other tables to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 156, 179, 202], "bbox": [0.6268281250000001, 0.3219722222222222, 0.9774375, 0.67175], "iscrowd": 0, "area": 5089.289399999999, "rle": {"size": [360, 640], "counts": "X_]41W;0O10001N100O100O10000O101N10000O10000O10001N10000O10000O10000O2O000O10000O10000O2O000O1000O010000O10000O01000O10000O10O10O10000O10000O01000O10000O10O1000O10000O10O10O10000O10000O01000O10001N10000O10;Ee0ZOf0[Of0ZO>BN2M2O2M3N2fMZ20000001O000000010O0001L3K5L4L5K4L4L5K4LoD:k:6J6KO000000000000000000000010O00000000000000000001O0000000000000000000O2O000000000O10000000001O0O10000000000O1000001O000O1000000000000O101O0000000O2O00000O2O000O101O0O10001O0O1000Rk5"}, "label": "a lighted umbrella sitting above a table in an outside diner along with other tables to the left"}]}
{"id": 439273, "image": "COCO_train2014_000000439273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller boy with the blue tie\"."}], "task": "refering", "answer_template": "The \"the taller boy with the blue tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364], "bbox": [0.47214062500000004, 0.04352297592997812, 0.8572656249999999, 0.9851859956236323], "iscrowd": 0, "area": 76813.93629999999, "rle": {"size": [457, 640], "counts": "QPW4T1h<h3XLh2\\M9O10000O2O000O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O100O100nLeG6[8GkG5V8HPH5P8GVH7j7F\\H6e7FbH7^7FhH6Y7FmH8S7ESI8m6DZI8g6E_I8a6FcI8]6GgI5Z6KiI2X6LkI1V6NmI0S6OoINS62oIKR64QJIP66SJGn58UJEl5;VJBk5=XJ@i5?YJ_Oh5a0ZJ\\Og5c0\\JZOe5e0^JWOd5h0_JUOb5k0`JROa5m0bJPO_5o0cJoN^5P1eJmN\\5S1m3O1O100O1O1O100O1O100000000000000O10000000000000000000000000001O00000001O0000000000001O00000000001O00000000001O00001O1O2N1O1O1O2N1O1O1O1O2`FcNP6^1jIjNS6W1hIPOU6Q1fIVOW6k0cI]OZ6e0`IB]6?_IG^6:_IK^66_IM`65\\INc63ZIOf62WI0j60SI2m6OPI3P7OlH3T7NhH5X70aH2_74YHNg78QHJo7;jGHU8>cGD]8T300000000000001O0000000000000000001lK]J@c5j0dJeN\\5d1mJjMS5`2TKoLl4[3\\KSLd4V4dKYK\\4Q5kK^JU4l5SLbIm3g6g19F9H9F:G8G3N0O01O001O001O00001O001O0010O01O001O001O001O3M9G9G9G9H8G9G:F9G9G9G9G9G9G9G9G:F9G9H8G9G9G9G9G9G9G:FVVX1"}, "label": "the taller boy with the blue tie"}]}
{"id": 439273, "image": "COCO_train2014_000000439273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with blue color tie\"."}], "task": "refering", "answer_template": "The \"a boy with blue color tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364], "bbox": [0.47214062500000004, 0.04352297592997812, 0.8572656249999999, 0.9851859956236323], "iscrowd": 0, "area": 76813.93629999999, "rle": {"size": [457, 640], "counts": "QPW4T1h<h3XLh2\\M9O10000O2O000O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O100O100nLeG6[8GkG5V8HPH5P8GVH7j7F\\H6e7FbH7^7FhH6Y7FmH8S7ESI8m6DZI8g6E_I8a6FcI8]6GgI5Z6KiI2X6LkI1V6NmI0S6OoINS62oIKR64QJIP66SJGn58UJEl5;VJBk5=XJ@i5?YJ_Oh5a0ZJ\\Og5c0\\JZOe5e0^JWOd5h0_JUOb5k0`JROa5m0bJPO_5o0cJoN^5P1eJmN\\5S1m3O1O100O1O1O100O1O100000000000000O10000000000000000000000000001O00000001O0000000000001O00000000001O00000000001O00001O1O2N1O1O1O2N1O1O1O1O2`FcNP6^1jIjNS6W1hIPOU6Q1fIVOW6k0cI]OZ6e0`IB]6?_IG^6:_IK^66_IM`65\\INc63ZIOf62WI0j60SI2m6OPI3P7OlH3T7NhH5X70aH2_74YHNg78QHJo7;jGHU8>cGD]8T300000000000001O0000000000000000001lK]J@c5j0dJeN\\5d1mJjMS5`2TKoLl4[3\\KSLd4V4dKYK\\4Q5kK^JU4l5SLbIm3g6g19F9H9F:G8G3N0O01O001O001O00001O001O0010O01O001O001O001O3M9G9G9G9H8G9G:F9G9G9G9G9G9G9G9G:F9G9H8G9G9G9G9G9G9G:FVVX1"}, "label": "a boy with blue color tie"}]}
{"id": 390125, "image": "COCO_train2014_000000390125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two woman in a room , one wearing white pants , the other wearing blue jeans\"."}], "task": "refering", "answer_template": "The \"two woman in a room , one wearing white pants , the other wearing blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 179, 180, 201, 202, 203, 204, 224, 225, 226, 227, 248, 249, 271, 272, 294, 295], "bbox": [0.7398906249999999, 0.3842708333333333, 0.8966249999999999, 0.7662916666666667], "iscrowd": 0, "area": 7068.065799999999, "rle": {"size": [480, 640], "counts": "^^n67g>3M1O2O0O2N1O2N2N1O2O0O2N101O0O2O0O01000O0100O10000O100OhDFm7:QHKl75aEMc11l8U1kFoNT9W1dFlN[9[1\\FhNc9^1TFgNj9_1nEdNQ:b1fEbNY:S3O101^OfK[F[4b9mKXFT4e9e0N2M3M4M2O11O001O100N2M2M4L4L4L4L3M4L4J>@c0]O8H9F9H8H8F;F9G9M3O2N1O1O1O2N1O1O2N1O1O101N1O1O1O2N1O1O1O2N2N3M2N2O1N3M2Nben0"}, "label": "two woman in a room , one wearing white pants , the other wearing blue jeans"}]}
{"id": 390125, "image": "COCO_train2014_000000390125.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in tan long sleeve shirt and white pants standing with woman in pink shirt and blue geans\"."}], "task": "refering", "answer_template": "The \"woman in tan long sleeve shirt and white pants standing with woman in pink shirt and blue geans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 179, 180, 201, 202, 203, 204, 224, 225, 226, 227, 248, 249, 271, 272, 294, 295], "bbox": [0.7398906249999999, 0.3842708333333333, 0.8966249999999999, 0.7662916666666667], "iscrowd": 0, "area": 7068.065799999999, "rle": {"size": [480, 640], "counts": "^^n67g>3M1O2O0O2N1O2N2N1O2O0O2N101O0O2O0O01000O0100O10000O100OhDFm7:QHKl75aEMc11l8U1kFoNT9W1dFlN[9[1\\FhNc9^1TFgNj9_1nEdNQ:b1fEbNY:S3O101^OfK[F[4b9mKXFT4e9e0N2M3M4M2O11O001O100N2M2M4L4L4L4L3M4L4J>@c0]O8H9F9H8H8F;F9G9M3O2N1O1O1O2N1O1O2N1O1O101N1O1O1O2N1O1O1O2N2N3M2N2O1N3M2Nben0"}, "label": "woman in tan long sleeve shirt and white pants standing with woman in pink shirt and blue geans"}]}
{"id": 218096, "image": "COCO_train2014_000000218096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting at the right side of the image\"."}], "task": "refering", "answer_template": "The \"man sitting at the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 48, 49, 50, 51, 66, 67, 68, 69, 85, 86, 87, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215], "bbox": [0.6685800000000001, 0.1056923076923077, 1.0, 0.9784307692307692], "iscrowd": 0, "area": 31270.72075, "rle": {"size": [325, 500], "counts": "ZUZ31S:4M3L3N3L4L3N3L4L4M3L5K5L4K6J5L4K6J5L4K5YM^MbMh2n1mMhMY2Q2RNdMZ20eL3^1BW20\\L=k1UOU22QLg0X2jNR25fKo0\\2jNS21bKT1^2hNW2M\\KY1a2gNY2JWK]1d2gNb2Y1`MdNa2\\1aMaN`2^1dM_N\\2a1T30000O10000O10000O10000O1000000O10000O10000O10000O10000O11O001O00001O001O001O001O00010O001O1O1O001O1O001O1O1O001O1mI]Nb3c1\\L_Nd3b1YLaNf3`1XLaNh3_1VLdNi3]1TLeNl3\\1QLfNo3[1nKhNQ4Z1jKiNV4Y1fKiNZ4Y1aKjN_4X1]KkNb4[1UKhNk4]1mJfNS5_1_JkN`5m2000001O00001O001O00001O001O00001O001O00001O1O1O1O1O1O1O1O1O1O1O2N9G>B:F3M2N3M2N3M2N3M2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3MoG"}, "label": "man sitting at the right side of the image"}]}
{"id": 218096, "image": "COCO_train2014_000000218096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a blue color sweater\"."}], "task": "refering", "answer_template": "The \"a person in a blue color sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 48, 49, 50, 51, 66, 67, 68, 69, 85, 86, 87, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215], "bbox": [0.6685800000000001, 0.1056923076923077, 1.0, 0.9784307692307692], "iscrowd": 0, "area": 31270.72075, "rle": {"size": [325, 500], "counts": "ZUZ31S:4M3L3N3L4L3N3L4L4M3L5K5L4K6J5L4K6J5L4K5YM^MbMh2n1mMhMY2Q2RNdMZ20eL3^1BW20\\L=k1UOU22QLg0X2jNR25fKo0\\2jNS21bKT1^2hNW2M\\KY1a2gNY2JWK]1d2gNb2Y1`MdNa2\\1aMaN`2^1dM_N\\2a1T30000O10000O10000O10000O1000000O10000O10000O10000O10000O11O001O00001O001O001O001O00010O001O1O1O001O1O001O1O1O001O1mI]Nb3c1\\L_Nd3b1YLaNf3`1XLaNh3_1VLdNi3]1TLeNl3\\1QLfNo3[1nKhNQ4Z1jKiNV4Y1fKiNZ4Y1aKjN_4X1]KkNb4[1UKhNk4]1mJfNS5_1_JkN`5m2000001O00001O001O00001O001O00001O001O00001O1O1O1O1O1O1O1O1O1O1O2N9G>B:F3M2N3M2N3M2N3M2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3MoG"}, "label": "a person in a blue color sweater"}]}
{"id": 218096, "image": "COCO_train2014_000000218096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"head cushion behind the man on the right\"."}], "task": "refering", "answer_template": "The \"head cushion behind the man on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 69, 83, 84, 87, 101, 102], "bbox": [0.6485599999999999, 0.27058461538461537, 0.88024, 0.4984923076923077], "iscrowd": 0, "area": 2330.8028, "rle": {"size": [325, 500], "counts": "[oV36^9a0K6I7J5K6M3M2N3M1O2N1O1O2O0010O0010O01O010O0010O01O0010O02N9G:G7H4L4L4M4K5K9GQRb0d0]m]OO1O1O01N100O1O1O2N1M3M3O1101J6K4L9FPnb0"}, "label": "head cushion behind the man on the right"}]}
{"id": 218096, "image": "COCO_train2014_000000218096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey chair behind man with beer bottle\"."}], "task": "refering", "answer_template": "The \"grey chair behind man with beer bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 69, 83, 84, 87, 101, 102], "bbox": [0.6485599999999999, 0.27058461538461537, 0.88024, 0.4984923076923077], "iscrowd": 0, "area": 2330.8028, "rle": {"size": [325, 500], "counts": "[oV36^9a0K6I7J5K6M3M2N3M1O2N1O1O2O0010O0010O01O010O0010O01O0010O02N9G:G7H4L4L4M4K5K9GQRb0d0]m]OO1O1O01N100O1O1O2N1M3M3O1101J6K4L9FPnb0"}, "label": "grey chair behind man with beer bottle"}]}
{"id": 218096, "image": "COCO_train2014_000000218096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a striped shirt kissing a flask\"."}], "task": "refering", "answer_template": "The \"a man in a striped shirt kissing a flask\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 43, 44, 45, 46, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210], "bbox": [0.20304, 0.09212307692307693, 0.71866, 0.986523076923077], "iscrowd": 0, "area": 50442.85064999999, "rle": {"size": [325, 500], "counts": "ReP1m0g8a0M3L4M3L4M3M3L4M3L4L4L4L4L4L4K5L4L4L4L4L4L4L4L4L4L4M3L4M3M3L4M2M4M3L4N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N200O1O100O1O100O1O1O1O1O1O1O1O1O1J6N2O1000000O10000O1000000O1000000O1[OPIUNP7h1YIRNg6k1bIoM^6n1n0M3M3M3N2O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O1000000001O00000000001O0000000000001O00000000001O001O001O001O001O001O1O001O001O001O001O0_JaN\\2`1\\MgNd2Z1TMmNl2S1PMROo2o0oLROQ3o0mLROS3P1kLQOT3Q1iLPOW3R1fLoNZ3S1dLnN[3T1bLmN^3U1_LlNa3V1]LjNc3X1ZLjNe3]1SLdNm3c1lK]NT4i1eKXN[4o1]KRNc4U2VKkMj4[2oJgMP5`2hJaMX5f30000000001O00000000001O1O001O1O001O1O001O001O1O3M4L5K4L4L5K4L5K4L4L5K4L4L5K4L4L5K4L5K4L4L5K4L4L5K4L4L5K4L5K4L4L5K4L4L5KU^\\1"}, "label": "a man in a striped shirt kissing a flask"}]}
{"id": 218096, "image": "COCO_train2014_000000218096.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with a mustache and green and lighter green striped shirt drinking from a flask on a bus\"."}], "task": "refering", "answer_template": "The \"man with a mustache and green and lighter green striped shirt drinking from a flask on a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 43, 44, 45, 46, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210], "bbox": [0.20304, 0.09212307692307693, 0.71866, 0.986523076923077], "iscrowd": 0, "area": 50442.85064999999, "rle": {"size": [325, 500], "counts": "ReP1m0g8a0M3L4M3L4M3M3L4M3L4L4L4L4L4L4K5L4L4L4L4L4L4L4L4L4L4M3L4M3M3L4M2M4M3L4N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N200O1O100O1O100O1O1O1O1O1O1O1O1O1J6N2O1000000O10000O1000000O1000000O1[OPIUNP7h1YIRNg6k1bIoM^6n1n0M3M3M3N2O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O1000000001O00000000001O0000000000001O00000000001O001O001O001O001O001O1O001O001O001O001O0_JaN\\2`1\\MgNd2Z1TMmNl2S1PMROo2o0oLROQ3o0mLROS3P1kLQOT3Q1iLPOW3R1fLoNZ3S1dLnN[3T1bLmN^3U1_LlNa3V1]LjNc3X1ZLjNe3]1SLdNm3c1lK]NT4i1eKXN[4o1]KRNc4U2VKkMj4[2oJgMP5`2hJaMX5f30000000001O00000000001O1O001O1O001O1O001O001O1O3M4L5K4L4L5K4L5K4L4L5K4L4L5K4L4L5K4L5K4L4L5K4L4L5K4L4L5K4L5K4L4L5K4L4L5KU^\\1"}, "label": "man with a mustache and green and lighter green striped shirt drinking from a flask on a bus"}]}
{"id": 308208, "image": "COCO_train2014_000000308208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup of brocolli cheese soup\"."}], "task": "refering", "answer_template": "The \"a white cup of brocolli cheese soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 359, 360, 361], "bbox": [0.510671875, 0.19256250000000003, 0.9893281249999999, 0.9341458333333333], "iscrowd": 0, "area": 75820.44254999999, "rle": {"size": [480, 640], "counts": "c^i46f>8H8H8H8H8H9G8H8H8H8H8J6K6K4K5K5L4K5K5L5J5L4K5K5L4K5K6K4K5K5L4K5L4K5K6K4K5K5L4K5L4K6J5L4K5K4M1N1O2O0O2O0O2N2O0O2O0O2O001N2O0O2O001N101N101O1N101N101N101O1N101N101O0O2O0O2^KfF\\3[9WLRGh3n8mK]GR4i901O0000001O00010O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O0000O1O1N2O1O1N2O1N2N2L4L4M3L4L4L4N2000000O1000000O1000000O1000000O1000000O1000000O2O00000O1000000O1000001N100O2N1O2O0O1O2N1O1O2O0O1O2N1O1O2N100O2N1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O2N1O1N3N1N2N3M2O1N3M2N2O2M2N3M2O1N3M2N2O2M2N2N3N1N2N3M2O1N3M2N3N1N2N3M2O1N3M2N2O2M3M3M4L3N2M3M3M4M2M3M3M3N3L3M3M3N2M4L3M3N2M3M4L3N2M3M3M4M2M3M3M3N3L3M3M3N2M3M4L3N2M3M3MeS3"}, "label": "a white cup of brocolli cheese soup"}]}
{"id": 308208, "image": "COCO_train2014_000000308208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup full of beverage\"."}], "task": "refering", "answer_template": "The \"a cup full of beverage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 359, 360, 361], "bbox": [0.510671875, 0.19256250000000003, 0.9893281249999999, 0.9341458333333333], "iscrowd": 0, "area": 75820.44254999999, "rle": {"size": [480, 640], "counts": "c^i46f>8H8H8H8H8H9G8H8H8H8H8J6K6K4K5K5L4K5K5L5J5L4K5K5L4K5K6K4K5K5L4K5L4K5K6K4K5K5L4K5L4K6J5L4K5K4M1N1O2O0O2O0O2N2O0O2O0O2O001N2O0O2O001N101N101O1N101N101N101O1N101N101O0O2O0O2^KfF\\3[9WLRGh3n8mK]GR4i901O0000001O00010O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O0000O1O1N2O1O1N2O1N2N2L4L4M3L4L4L4N2000000O1000000O1000000O1000000O1000000O1000000O2O00000O1000000O1000001N100O2N1O2O0O1O2N1O1O2O0O1O2N1O1O2N100O2N1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O2N1O1N3N1N2N3M2O1N3M2N2O2M2N3M2O1N3M2N2O2M2N2N3N1N2N3M2O1N3M2N3N1N2N3M2O1N3M2N2O2M3M3M4L3N2M3M3M4M2M3M3M3N3L3M3M3N2M4L3M3N2M3M4L3N2M3M3M4M2M3M3M3N3L3M3M3N2M3M4L3N2M3M3MeS3"}, "label": "a cup full of beverage"}]}
{"id": 373747, "image": "COCO_train2014_000000373747.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of pizza with ham which is behind the piece of pizza with olives and pepperoni\"."}], "task": "refering", "answer_template": "The \"the piece of pizza with ham which is behind the piece of pizza with olives and pepperoni\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 44, 45, 46], "bbox": [0.32626, 0.0045045045045045045, 0.6989200000000001, 0.18426426426426426], "iscrowd": 0, "area": 8564.656949999999, "rle": {"size": [333, 500], "counts": "jPe13U:6L5L3L4N2M4M2O1O1O1O001O1O001O1O001O1O001O1O001O1O00001O001O001O00001O001O00001O001O00001O001O00001O001O00000000001O00000000001O000000001O000000000000001O0000000000001O000000000000001O00000000000000000000000000O1000000000000O100000000000000O100O100O100O100O1O100O100O100O100O1O100O100O100O10000O100O10000O10000O100O10000O100O100O1O100O100O100O1O1O100O1O1O100O1O1O100O1OZSa1"}, "label": "the piece of pizza with ham which is behind the piece of pizza with olives and pepperoni"}]}
{"id": 373747, "image": "COCO_train2014_000000373747.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half visible pizza at the top edge\"."}], "task": "refering", "answer_template": "The \"the half visible pizza at the top edge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 44, 45, 46], "bbox": [0.32626, 0.0045045045045045045, 0.6989200000000001, 0.18426426426426426], "iscrowd": 0, "area": 8564.656949999999, "rle": {"size": [333, 500], "counts": "jPe13U:6L5L3L4N2M4M2O1O1O1O001O1O001O1O001O1O001O1O001O1O00001O001O001O00001O001O00001O001O00001O001O00001O001O00000000001O00000000001O000000001O000000000000001O0000000000001O000000000000001O00000000000000000000000000O1000000000000O100000000000000O100O100O100O100O1O100O100O100O100O1O100O100O100O10000O100O10000O10000O100O10000O100O100O1O100O100O100O1O1O100O1O1O100O1O1O100O1OZSa1"}, "label": "the half visible pizza at the top edge"}]}
{"id": 373747, "image": "COCO_train2014_000000373747.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza set in front of the camera , with 6 black things on top\"."}], "task": "refering", "answer_template": "The \"the pizza set in front of the camera , with 6 black things on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192], "bbox": [0.22898, 0.3011111111111111, 0.79172, 0.9258558558558558], "iscrowd": 0, "area": 43752.40095000001, "rle": {"size": [333, 500], "counts": "ZXU11Z:4L5K4L5K4K5L5K4L5K4L5K4L5K4L5K4M4L3M4L3M3M2N2M3N2N2N3M2N2N2N2M3N2N2N3M2N2N2N2N2N2N2N2O2M2N2N1O2N1O2N1O2O0O2N1O2N101N101O0O2O001N101O0O2O0O2O001N101O0O2O001N101O0O2O0O2O001N101O0O101O0O10000O2O0O10000O2O000O101O0O1000001O000000001O000000001O000000001O000000001O000000001O000000001O0001O0001O000000001O0001O01O000000001O000001O0000000O1000000O2O0000000O10001O0O1000000O101O000O1000001N2O1O1O1N2O1O1O1O1N1000000O10000O2O0O1O100O1O100O1O100O101N2N2O1N2N2O1N1O2O1N2N101N2O1N1O2O1N2N1O2N2N2N1O2N3M2N2N3M2N3M2N2O2M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N8H9G9E7H7I8G8I8HdcQ1"}, "label": "the pizza set in front of the camera , with 6 black things on top"}]}
{"id": 373747, "image": "COCO_train2014_000000373747.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza that is on the plate has black olives , cheese , and tomato sauce\"."}], "task": "refering", "answer_template": "The \"the pizza that is on the plate has black olives , cheese , and tomato sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192], "bbox": [0.22898, 0.3011111111111111, 0.79172, 0.9258558558558558], "iscrowd": 0, "area": 43752.40095000001, "rle": {"size": [333, 500], "counts": "ZXU11Z:4L5K4L5K4K5L5K4L5K4L5K4L5K4L5K4M4L3M4L3M3M2N2M3N2N2N3M2N2N2N2M3N2N2N3M2N2N2N2N2N2N2N2O2M2N2N1O2N1O2N1O2O0O2N1O2N101N101O0O2O001N101O0O2O0O2O001N101O0O2O001N101O0O2O0O2O001N101O0O101O0O10000O2O0O10000O2O000O101O0O1000001O000000001O000000001O000000001O000000001O000000001O000000001O0001O0001O000000001O0001O01O000000001O000001O0000000O1000000O2O0000000O10001O0O1000000O101O000O1000001N2O1O1O1N2O1O1O1O1N1000000O10000O2O0O1O100O1O100O1O100O101N2N2O1N2N2O1N1O2O1N2N101N2O1N1O2O1N2N1O2N2N2N1O2N3M2N2N3M2N3M2N2O2M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N8H9G9E7H7I8G8I8HdcQ1"}, "label": "the pizza that is on the plate has black olives , cheese , and tomato sauce"}]}
{"id": 300021, "image": "COCO_train2014_000000300021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dish with layers of yellow squash and herbs\"."}], "task": "refering", "answer_template": "The \"dish with layers of yellow squash and herbs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.533484375, 0.3922175732217573, 1.0, 0.8285774058577406], "iscrowd": 0, "area": 54404.11335, "rle": {"size": [478, 640], "counts": "kjo46g>8I6J6J7I6J6J7I6J6J7I6J6J7I6J6J7I6J6J7I6I7J7I6J6J7I6J6J7I6J6J00000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000001O000000000000000O100000000000000000000000000000000000000O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1Of0ZO\\1dN\\1dNZD"}, "label": "dish with layers of yellow squash and herbs"}]}
{"id": 300021, "image": "COCO_train2014_000000300021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow baked squash dish\"."}], "task": "refering", "answer_template": "The \"yellow baked squash dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.533484375, 0.3922175732217573, 1.0, 0.8285774058577406], "iscrowd": 0, "area": 54404.11335, "rle": {"size": [478, 640], "counts": "kjo46g>8I6J6J7I6J6J7I6J6J7I6J6J7I6J6J7I6J6J7I6I7J7I6J6J7I6J6J7I6J6J00000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000001O000000000000000O100000000000000000000000000000000000000O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1Of0ZO\\1dN\\1dNZD"}, "label": "yellow baked squash dish"}]}
{"id": 300021, "image": "COCO_train2014_000000300021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the food with the pepperoni on it\"."}], "task": "refering", "answer_template": "The \"the food with the pepperoni on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.070578125, 0.36403765690376566, 0.533015625, 0.6907531380753137], "iscrowd": 0, "area": 35506.422299999984, "rle": {"size": [478, 640], "counts": "hYe02f>6N2N2O1N2O1N2O1N2O4L6K5K3M1O0O1N3M2N2N2N3M2N2O2M2N2O2M2N2N3N1N2N2O0O2N2N101M3N1N3J6L4M3N2O100O1O1O100O1N2N2O1N2O1N2O1N2M3N2O1N20000000000001N101OO10000O100O1O100O100O100O00100O100O100O100O100O1O1O1N10100O1O1O100O1O1001O1O1O1N100000001OO10000O10O0100O100001N2O001O1OO10O1000000000000O10O100O1O10O02O00000O10001O000000000000000001O1O1O1O1O1O1O1O00000000O10000000O1000O10O10001O1O1O1O00100O000001O00000O2N1O1O1O2N1O1O1O1000000000001O0000000000O1000000000000O100000000000000000000001O0000001O01O001O1O001O001O1N101O001O1O1O1O1O1O1N3N1O1O1O5K7H3N7I1O1O1O0O2O1O001O1O2N3L3N1O1O1O1O1gMQD`1Q<^NUD]1l;`N`DV1a;hNdDT1g<L:F`0@fX[4"}, "label": "the food with the pepperoni on it"}]}
{"id": 300021, "image": "COCO_train2014_000000300021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a square pepperoni pizza\"."}], "task": "refering", "answer_template": "The \"a square pepperoni pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.070578125, 0.36403765690376566, 0.533015625, 0.6907531380753137], "iscrowd": 0, "area": 35506.422299999984, "rle": {"size": [478, 640], "counts": "hYe02f>6N2N2O1N2O1N2O1N2O4L6K5K3M1O0O1N3M2N2N2N3M2N2O2M2N2O2M2N2N3N1N2N2O0O2N2N101M3N1N3J6L4M3N2O100O1O1O100O1N2N2O1N2O1N2O1N2M3N2O1N20000000000001N101OO10000O100O1O100O100O100O00100O100O100O100O100O1O1O1N10100O1O1O100O1O1001O1O1O1N100000001OO10000O10O0100O100001N2O001O1OO10O1000000000000O10O100O1O10O02O00000O10001O000000000000000001O1O1O1O1O1O1O1O00000000O10000000O1000O10O10001O1O1O1O00100O000001O00000O2N1O1O1O2N1O1O1O1000000000001O0000000000O1000000000000O100000000000000000000001O0000001O01O001O1O001O001O1N101O001O1O1O1O1O1O1N3N1O1O1O5K7H3N7I1O1O1O0O2O1O001O1O2N3L3N1O1O1O1O1gMQD`1Q<^NUD]1l;`N`DV1a;hNdDT1g<L:F`0@fX[4"}, "label": "a square pepperoni pizza"}]}
{"id": 300021, "image": "COCO_train2014_000000300021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a knife\"."}], "task": "refering", "answer_template": "The \"a person with a knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 67, 68, 85, 90, 113], "bbox": [0.612609375, 0.0, 0.9969531249999999, 0.2959414225941423], "iscrowd": 0, "area": 10465.451699999998, "rle": {"size": [478, 640], "counts": "bof5;c>`0@1O1O2N1O1O2N1O1O1O2N1O1O1O001N101O1O001O001O001O1O001O001O001O00O1N2O1O1N2O1N24L6J5K6J6J4LO10000O10000O100O10000O100O10000007I7I7I7I1ON2N2N2N2N2N1O2N2N2N2N2N2iM^CP2d<QN]Ck1e<VN]Ce1e<\\N\\CV1P=lNPCc0_=B_BKQ>985K4L5K4L5K4L1O000000000000000001O4L4L4L3M4L4L4L4L4L4L4LYl?OgS@001O1O1N2O1O1O1O001N2O1O1O1O1O1N10001O001N10010O010O01O010O010O010O0010O010O01O010O010O1O05L4K5K5K6J5eDgNh8^1eFQOY9U1TFZOk9k0aEC`:Y24L4L3M4L4O1N3N2N2N2N2N2N2N2N2M3N1O2N2N2N2N2N2N2M3N2N2N2N2M3N2M3N2M3N2N2M3N2M3F:_O[n0"}, "label": "a person with a knife"}]}
{"id": 300021, "image": "COCO_train2014_000000300021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands holding kitchen tongs\"."}], "task": "refering", "answer_template": "The \"hands holding kitchen tongs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 67, 68, 85, 90, 113], "bbox": [0.612609375, 0.0, 0.9969531249999999, 0.2959414225941423], "iscrowd": 0, "area": 10465.451699999998, "rle": {"size": [478, 640], "counts": "bof5;c>`0@1O1O2N1O1O2N1O1O1O2N1O1O1O001N101O1O001O001O001O1O001O001O001O00O1N2O1O1N2O1N24L6J5K6J6J4LO10000O10000O100O10000O100O10000007I7I7I7I1ON2N2N2N2N2N1O2N2N2N2N2N2iM^CP2d<QN]Ck1e<VN]Ce1e<\\N\\CV1P=lNPCc0_=B_BKQ>985K4L5K4L5K4L1O000000000000000001O4L4L4L3M4L4L4L4L4L4L4LYl?OgS@001O1O1N2O1O1O1O001N2O1O1O1O1O1N10001O001N10010O010O01O010O010O010O0010O010O01O010O010O1O05L4K5K5K6J5eDgNh8^1eFQOY9U1TFZOk9k0aEC`:Y24L4L3M4L4O1N3N2N2N2N2N2N2N2N2M3N1O2N2N2N2N2N2N2M3N2N2N2N2M3N2M3N2M3N2N2M3N2M3F:_O[n0"}, "label": "hands holding kitchen tongs"}]}
{"id": 365573, "image": "COCO_train2014_000000365573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike closest to the man\"."}], "task": "refering", "answer_template": "The \"the bike closest to the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 153, 155, 156], "bbox": [0.44448000000000004, 0.4287687687687688, 0.71174, 0.6927027027027027], "iscrowd": 0, "area": 5800.5826, "rle": {"size": [333, 500], "counts": "hfX22Z:2N2O2M1O010O1O101N2N1O2O1N2N2O1N2N2O1N1O2O1N2O1O1O1N101O1O001N2O0L5K5K4K6K5L310000O101O0O1N2N2N2N3M3M2M4M3M3M2N3M3N1O1OO00100O1O100O1O100O00100O1O1O100O1O010O1O2O1N2N3K4I7I7J6K5O01O1O010O001O10O01O001O010O1O`0@:G0O001O1O010O001O10O01O001O1N1M4L3N3L4L3M4L4L3M4M2M4Lad^1"}, "label": "the bike closest to the man"}]}
{"id": 365573, "image": "COCO_train2014_000000365573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike with a blue water bottle\"."}], "task": "refering", "answer_template": "The \"the bike with a blue water bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 153, 155, 156], "bbox": [0.44448000000000004, 0.4287687687687688, 0.71174, 0.6927027027027027], "iscrowd": 0, "area": 5800.5826, "rle": {"size": [333, 500], "counts": "hfX22Z:2N2O2M1O010O1O101N2N1O2O1N2N2O1N2N2O1N1O2O1N2O1O1O1N101O1O001N2O0L5K5K4K6K5L310000O101O0O1N2N2N2N3M3M2M4M3M3M2N3M3N1O1OO00100O1O100O1O100O00100O1O1O100O1O010O1O2O1N2N3K4I7I7J6K5O01O1O010O001O10O01O001O010O1O`0@:G0O001O1O010O001O10O01O001O1N1M4L3N3L4L3M4L4L3M4M2M4Lad^1"}, "label": "the bike with a blue water bottle"}]}
{"id": 365573, "image": "COCO_train2014_000000365573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bicycle in center front\"."}], "task": "refering", "answer_template": "The \"bicycle in center front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 111, 112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 187, 188], "bbox": [0.19436, 0.4527927927927928, 0.5145, 0.8497897897897898], "iscrowd": 0, "area": 13180.43245, "rle": {"size": [333, 500], "counts": "Vho01Z:`0oNh0J4L3M2E<K4M3M4M2O2O0O2N1O3N1M2N2N2N2N2N1O2N2N1O1O1O1O10O01O010O001O010O0010O010O1O1O1O1O1O1O1O1O1O1O1cMfHf1[7UNlHh1V7QNRIl1o6oMVIR2b72N1O0000000000010O1O1O1O001NYOPHSOP8m0SHPOm7o0WHoNg7R1]HjNc7V1bHfN]7Z1hHaNW7`1k0O10O100Ob0_O6I2M3L4M2N3M4L3M4L3O2M3N2M3N1O1N2O0O2O1O1N2O0O2O001N101O0OoLUJb1k5ZNZJe1f5WN`Jg1_5WNdJi1\\5TNgJm1X5PNkJP2Z5hMiJX2[5aMhJ_2c600000O1000O101O000O2O001O0O2O00001N101O001N2N2O2M2N2N2O2M2N2O2M2N2N2N3L3N2N3F9Hfi^2"}, "label": "bicycle in center front"}]}
{"id": 365573, "image": "COCO_train2014_000000365573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle with a jacket hanging on the handle bars with a red bag in the back\"."}], "task": "refering", "answer_template": "The \"a bicycle with a jacket hanging on the handle bars with a red bag in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 111, 112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 187, 188], "bbox": [0.19436, 0.4527927927927928, 0.5145, 0.8497897897897898], "iscrowd": 0, "area": 13180.43245, "rle": {"size": [333, 500], "counts": "Vho01Z:`0oNh0J4L3M2E<K4M3M4M2O2O0O2N1O3N1M2N2N2N2N2N1O2N2N1O1O1O1O10O01O010O001O010O0010O010O1O1O1O1O1O1O1O1O1O1O1cMfHf1[7UNlHh1V7QNRIl1o6oMVIR2b72N1O0000000000010O1O1O1O001NYOPHSOP8m0SHPOm7o0WHoNg7R1]HjNc7V1bHfN]7Z1hHaNW7`1k0O10O100Ob0_O6I2M3L4M2N3M4L3M4L3O2M3N2M3N1O1N2O0O2O1O1N2O0O2O001N101O0OoLUJb1k5ZNZJe1f5WN`Jg1_5WNdJi1\\5TNgJm1X5PNkJP2Z5hMiJX2[5aMhJ_2c600000O1000O101O000O2O001O0O2O00001N101O001N2N2O2M2N2N2O2M2N2O2M2N2N2N3L3N2N3F9Hfi^2"}, "label": "a bicycle with a jacket hanging on the handle bars with a red bag in the back"}]}
{"id": 13318, "image": "COCO_train2014_000000013318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a side bag of a lady\"."}], "task": "refering", "answer_template": "The \"a side bag of a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [327, 328, 350, 351, 372, 373, 374, 375, 394, 395, 396, 397, 398, 399, 418, 419, 420, 421, 422, 423, 442, 443, 444, 445, 446], "bbox": [0.1703627760252366, 0.60290625, 0.4, 0.8686562500000001], "iscrowd": 0, "area": 11721.278300000002, "rle": {"size": [640, 634], "counts": "jnS29fc0c0]O:F3M2N2N3M2N2N3M1O1O2N1O2N10001N100O2O000O2O0O2O000O2O0O101O0O101N101O0O101N100O2N1N2N3M2N3M2[ORMV@P3i?SMT@o2j?UMQ@m2m?XMn_Oj2Q`0d0N1O1N3N1O2M2O1O2M2O1O2M2O1O2M2O2N1N2O2l1SN1O100O1O101N100O1O100O1O100O1O100O2N100O1O1O10G8O1O100?BO01O0010O01O001O010O001O0010O01O0010O01O0O2O001N101O0O101N101O0O2O0O2N2M5K4M3L4L4M3G:D;EXP]7"}, "label": "a side bag of a lady"}]}
{"id": 13318, "image": "COCO_train2014_000000013318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black purse the woman int he coat is carrying in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the black purse the woman int he coat is carrying in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [327, 328, 350, 351, 372, 373, 374, 375, 394, 395, 396, 397, 398, 399, 418, 419, 420, 421, 422, 423, 442, 443, 444, 445, 446], "bbox": [0.1703627760252366, 0.60290625, 0.4, 0.8686562500000001], "iscrowd": 0, "area": 11721.278300000002, "rle": {"size": [640, 634], "counts": "jnS29fc0c0]O:F3M2N2N3M2N2N3M1O1O2N1O2N10001N100O2O000O2O0O2O000O2O0O101O0O101N101O0O101N100O2N1N2N3M2N3M2[ORMV@P3i?SMT@o2j?UMQ@m2m?XMn_Oj2Q`0d0N1O1N3N1O2M2O1O2M2O1O2M2O1O2M2O2N1N2O2l1SN1O100O1O101N100O1O100O1O100O1O100O2N100O1O1O10G8O1O100?BO01O0010O01O001O010O001O0010O01O0010O01O0O2O001N101O0O101N101O0O2O0O2N2M5K4M3L4L4M3G:D;EXP]7"}, "label": "the black purse the woman int he coat is carrying in the right hand picture"}]}
{"id": 578567, "image": "COCO_train2014_000000578567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy playing baseball in a white uniform with orange lettering\"."}], "task": "refering", "answer_template": "The \"a small boy playing baseball in a white uniform with orange lettering\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 226, 227, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319, 320, 340, 341, 342, 343, 362, 363, 365, 366, 385, 386, 388, 389, 408, 409, 411, 412, 431], "bbox": [0.75675, 0.37275, 0.9436875, 0.7984218750000001], "iscrowd": 0, "area": 17166.3507, "rle": {"size": [640, 640], "counts": "[o^93ac0`0G`0C3M2O1N2N2O1N3A>I7L4N2M1M4K5J5L5L3N3nMSMPCk2P=^MiB`2W=kM`BS2d=XNnAg1U>n14L3M4L3N3M2O2N1O1O1N2O11O1O1O1N2C=L3M4M3L4000O010O1O1O00100L3H:O1O1O10O0100O100O001=C8H1O1N101O1O1O1N10`0@3M3L4M7I2N2N2N2O0O2N11O05K1O1O112N4MNo0QOO100O1O2N10O1N101cLgBOZ=InB6T=BTC<o<[OXCd0k<TO[Ck0h<mN_CQ1b<fNgCh0ZNXNP>`0\\Dm0kM`Nj=`0fDc0gMjNd=?PE;n;BZ4L4M3MPVf0"}, "label": "a small boy playing baseball in a white uniform with orange lettering"}]}
{"id": 578567, "image": "COCO_train2014_000000578567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy wearing white baseball helmet , white baseball uniform with orange writing , standing slightly bent at the waist\"."}], "task": "refering", "answer_template": "The \"boy wearing white baseball helmet , white baseball uniform with orange writing , standing slightly bent at the waist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 226, 227, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319, 320, 340, 341, 342, 343, 362, 363, 365, 366, 385, 386, 388, 389, 408, 409, 411, 412, 431], "bbox": [0.75675, 0.37275, 0.9436875, 0.7984218750000001], "iscrowd": 0, "area": 17166.3507, "rle": {"size": [640, 640], "counts": "[o^93ac0`0G`0C3M2O1N2N2O1N3A>I7L4N2M1M4K5J5L5L3N3nMSMPCk2P=^MiB`2W=kM`BS2d=XNnAg1U>n14L3M4L3N3M2O2N1O1O1N2O11O1O1O1N2C=L3M4M3L4000O010O1O1O00100L3H:O1O1O10O0100O100O001=C8H1O1N101O1O1O1N10`0@3M3L4M7I2N2N2N2O0O2N11O05K1O1O112N4MNo0QOO100O1O2N10O1N101cLgBOZ=InB6T=BTC<o<[OXCd0k<TO[Ck0h<mN_CQ1b<fNgCh0ZNXNP>`0\\Dm0kM`Nj=`0fDc0gMjNd=?PE;n;BZ4L4M3MPVf0"}, "label": "boy wearing white baseball helmet , white baseball uniform with orange writing , standing slightly bent at the waist"}]}
{"id": 578567, "image": "COCO_train2014_000000578567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a number four baseball jersey pitching the ball to a player\"."}], "task": "refering", "answer_template": "The \"a boy with a number four baseball jersey pitching the ball to a player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 349, 350, 351, 352, 353, 354, 372, 373, 375, 376, 377, 395, 396, 397, 398, 399, 400, 417, 418, 419, 421, 422, 423, 445, 446, 468, 469, 491, 492, 514, 515], "bbox": [0.15021875, 0.268953125, 0.443953125, 0.9774687499999999], "iscrowd": 0, "area": 41084.139650000005, "rle": {"size": [640, 640], "counts": "n_l11nc01O001O1O1O1O1O1O001N2O1mKH]D9b;MXD4g;2SDOk;9nCGR<?hCBW<d0cC]O\\<j0]CWOa<P1YCQOf<U1TClNk<[1nBfNP=a1iB`NW=f1cB[N\\=l1]BUNb=Q2XBPNh=U2RBlMn=Z2kAgMU>^2eAbM\\>c2^A^Mb>h2WAYMi>l2QAUMo>n300000O10000000001OSMXKbFg4^9_K]Fa4b9eKYF[4f9lKTFT4k9RLPFn3o9XLlEh3S:_LgE1WNU2Q<PNcEL_No1n;[N^EFhNk1j;bN[EDnNf1f;iNZEBSOa1b;POYE_OYO]1];WOXE]O_OW1X;_OWE[ODQ1U;GUEYOIl0Q;NTEVOOh0l:5SETO4c0h:<RERO9>d:c0QEoN`09^:k0PEmNe04Z:R1oDkNj0NW:Z1mDiNo0IT:`1kDgNU1Eo9g1jDeNZ1@k9n1iDcN`1ZOf9V2hD`Nf1VOc8XOSFT3C_Ni1RO[8BWFP3C]Nm1mNZ8GTFR3C[NQ2iNX8JSFU3BXNV2fNU8NQFW3CUNY2cNS81PFZ3BSN]2_NQ85nE\\3BPNa2]No78lEY5W2\\Jm7;kE[5m<fJRC[5m<fJPC\\5P=dJoB^5P=cJnB_5Q=aJnB`5R=aJlBa5S=g0N2N1N3N2N1N3N2N2M2O2N2M2O2N2M3N1O2N2O1N4M4L3M4K4M3M4L3L5L3M4M3L6J5K6J5L5J5K6J5L4K4L5K4L5L3L5K4L5K4M4K4iIgC_4_<XKdCh4h<jJZCW5Q=[JQCe5e=2N2N1O2O1WDhIf9Z6TFkIk9W6oEmIQ:U6jEnIV:S6fEPJZ:R6aEQJ_:Q6\\ERJd:P6VETJj:n5QEUJo:m5lDWJS;j5iDYJi7QOSKh6oLZJj7VOTKa6mL\\Jk7[OTKf7g4cHUK]7g4kHUKU7g4RIVKn6f4ZIVKf6e4cIWK^6d4jIXKV6c4SJYKm5c4[JYKe5b4dJZK\\5b4mJYKS5b4VK[Ki4a4`KXKc4c4fKTKa4f4gKQK`4j4gKmJ`4n4gKjJ_4P5hKhJ_4S5[5K5K5J6K5K6I6K5iNW1F:F=Ce0ZOe0\\OhUn6"}, "label": "a boy with a number four baseball jersey pitching the ball to a player"}]}
{"id": 578567, "image": "COCO_train2014_000000578567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy pitching a ball with the number 4 on his jersey\"."}], "task": "refering", "answer_template": "The \"a boy pitching a ball with the number 4 on his jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 349, 350, 351, 352, 353, 354, 372, 373, 375, 376, 377, 395, 396, 397, 398, 399, 400, 417, 418, 419, 421, 422, 423, 445, 446, 468, 469, 491, 492, 514, 515], "bbox": [0.15021875, 0.268953125, 0.443953125, 0.9774687499999999], "iscrowd": 0, "area": 41084.139650000005, "rle": {"size": [640, 640], "counts": "n_l11nc01O001O1O1O1O1O1O001N2O1mKH]D9b;MXD4g;2SDOk;9nCGR<?hCBW<d0cC]O\\<j0]CWOa<P1YCQOf<U1TClNk<[1nBfNP=a1iB`NW=f1cB[N\\=l1]BUNb=Q2XBPNh=U2RBlMn=Z2kAgMU>^2eAbM\\>c2^A^Mb>h2WAYMi>l2QAUMo>n300000O10000000001OSMXKbFg4^9_K]Fa4b9eKYF[4f9lKTFT4k9RLPFn3o9XLlEh3S:_LgE1WNU2Q<PNcEL_No1n;[N^EFhNk1j;bN[EDnNf1f;iNZEBSOa1b;POYE_OYO]1];WOXE]O_OW1X;_OWE[ODQ1U;GUEYOIl0Q;NTEVOOh0l:5SETO4c0h:<RERO9>d:c0QEoN`09^:k0PEmNe04Z:R1oDkNj0NW:Z1mDiNo0IT:`1kDgNU1Eo9g1jDeNZ1@k9n1iDcN`1ZOf9V2hD`Nf1VOc8XOSFT3C_Ni1RO[8BWFP3C]Nm1mNZ8GTFR3C[NQ2iNX8JSFU3BXNV2fNU8NQFW3CUNY2cNS81PFZ3BSN]2_NQ85nE\\3BPNa2]No78lEY5W2\\Jm7;kE[5m<fJRC[5m<fJPC\\5P=dJoB^5P=cJnB_5Q=aJnB`5R=aJlBa5S=g0N2N1N3N2N1N3N2N2M2O2N2M2O2N2M3N1O2N2O1N4M4L3M4K4M3M4L3L5L3M4M3L6J5K6J5L5J5K6J5L4K4L5K4L5L3L5K4L5K4M4K4iIgC_4_<XKdCh4h<jJZCW5Q=[JQCe5e=2N2N1O2O1WDhIf9Z6TFkIk9W6oEmIQ:U6jEnIV:S6fEPJZ:R6aEQJ_:Q6\\ERJd:P6VETJj:n5QEUJo:m5lDWJS;j5iDYJi7QOSKh6oLZJj7VOTKa6mL\\Jk7[OTKf7g4cHUK]7g4kHUKU7g4RIVKn6f4ZIVKf6e4cIWK^6d4jIXKV6c4SJYKm5c4[JYKe5b4dJZK\\5b4mJYKS5b4VK[Ki4a4`KXKc4c4fKTKa4f4gKQK`4j4gKmJ`4n4gKjJ_4P5hKhJ_4S5[5K5K5J6K5K6I6K5iNW1F:F=Ce0ZOe0\\OhUn6"}, "label": "a boy pitching a ball with the number 4 on his jersey"}]}
{"id": 283666, "image": "COCO_train2014_000000283666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in front with the red shirted jockey\"."}], "task": "refering", "answer_template": "The \"the horse in front with the red shirted jockey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 135, 136, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 271], "bbox": [0.65425, 0.3125058548009368, 0.9360781250000001, 0.7801639344262294], "iscrowd": 0, "area": 15683.31029999999, "rle": {"size": [427, 640], "counts": "To^51X=2N2N2N2N2N2N2N2N2N2N2O1bD]OS:d0iECS:<kEKQ:5lE2m9WOYEf0i08k9RO\\Ef0f0;m9QO\\Ec0e0`0n9mN]Ec0b0d0o9@PFb0P:^OmEf0Q:ZOmEj0R:VOkEn0S:ROlEP1T:POiET1U:lNkEV1T:P1O2O1N101O1N10J6H701000O10000O1O1O1O2N1O1O100001O1O1O1O1O001O1O100O1O00100O1O100O001O100O1O[F^Mc8b2\\G_Mc8a2]GaMb8^2_GbMa8^2^GdMa8[2_GfMa8Y2_GiM`8W2^GkMa8U2^GmMb8R2]GQNb8o1\\GSNd8l1[GWNd8h1[GZNe8e1ZG^Ne8b1YG`Ng8`1WGbNi8_1TGdNk8^1QGdNo8]1oFeNP9\\1mFfNR9]1jFfNU9[1hFgNX9g2O000eGnK]7S4XHWLh7i3nG_LT8Z42N2N2N3M2N2N2N3M2N2N2N2N2N2N2N2O2N1O1N2O1O1O1O1O1O1N2O1O2N1O1O1N2O1O1O1O1O1O1N2O1N3M1O00O100O100O100O10O010O010O010O10O04L6D<A?A?3MN2M2O2M3N2M4M2N2M3L5L3M3L5L4LPm`0"}, "label": "the horse in front with the red shirted jockey"}]}
{"id": 283666, "image": "COCO_train2014_000000283666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown horse carrying the jockey in the red shirt\"."}], "task": "refering", "answer_template": "The \"the brown horse carrying the jockey in the red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 135, 136, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 271], "bbox": [0.65425, 0.3125058548009368, 0.9360781250000001, 0.7801639344262294], "iscrowd": 0, "area": 15683.31029999999, "rle": {"size": [427, 640], "counts": "To^51X=2N2N2N2N2N2N2N2N2N2N2O1bD]OS:d0iECS:<kEKQ:5lE2m9WOYEf0i08k9RO\\Ef0f0;m9QO\\Ec0e0`0n9mN]Ec0b0d0o9@PFb0P:^OmEf0Q:ZOmEj0R:VOkEn0S:ROlEP1T:POiET1U:lNkEV1T:P1O2O1N101O1N10J6H701000O10000O1O1O1O2N1O1O100001O1O1O1O1O001O1O100O1O00100O1O100O001O100O1O[F^Mc8b2\\G_Mc8a2]GaMb8^2_GbMa8^2^GdMa8[2_GfMa8Y2_GiM`8W2^GkMa8U2^GmMb8R2]GQNb8o1\\GSNd8l1[GWNd8h1[GZNe8e1ZG^Ne8b1YG`Ng8`1WGbNi8_1TGdNk8^1QGdNo8]1oFeNP9\\1mFfNR9]1jFfNU9[1hFgNX9g2O000eGnK]7S4XHWLh7i3nG_LT8Z42N2N2N3M2N2N2N3M2N2N2N2N2N2N2N2O2N1O1N2O1O1O1O1O1O1N2O1O2N1O1O1N2O1O1O1O1O1O1N2O1N3M1O00O100O100O100O10O010O010O010O10O04L6D<A?A?3MN2M2O2M3N2M4M2N2M3L5L3M3L5L4LPm`0"}, "label": "the brown horse carrying the jockey in the red shirt"}]}
{"id": 537621, "image": "COCO_train2014_000000537621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with green and white striped pants\"."}], "task": "refering", "answer_template": "The \"woman with green and white striped pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 91, 113, 114, 135, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 249, 250], "bbox": [0.843, 0.17112412177985947, 1.0, 0.6880562060889929], "iscrowd": 0, "area": 5378.879700000002, "rle": {"size": [427, 640], "counts": "]^Q79Z:8lGJQ89nGHo7<oGEo7=oGDo7?PHAo7a0oG@o7c0PH]Oo7e0oG\\OP8f0oGZOP8k0kGWOT8X1]GhNb8Y1]GhNc8Y1\\GgNd8Y1\\GgNc8Z1\\GgNd8Z1[GfNd8[1[GfNe8Z1[GfNe8[1YGfNf8[1ZGeNf8[1ZGeNe8]1ZGcNe8^1[GbNd8`1ZGaNe8`1[G`Nd8b1[G^Nd8c1\\G]Nc8e1[G\\Nd8l1UGTNj8S2oFnMQ9T2mFkMT9W2iFjMV9Y2hFgMX9Z2fFgMZ9U300000000000K4K6M3bLXFV3k9hLXFT3Q:XNi1D<H7J7I]d6EP\\I4L5K4WOj0K4L4L5L3N3M2N3OWOlDUOY;h0g000O100000O10O1000000O1000000O10001N6HdM"}, "label": "woman with green and white striped pants"}]}
{"id": 537621, "image": "COCO_train2014_000000537621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing glasses and green pajamas\"."}], "task": "refering", "answer_template": "The \"a woman wearing glasses and green pajamas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 91, 113, 114, 135, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 249, 250], "bbox": [0.843, 0.17112412177985947, 1.0, 0.6880562060889929], "iscrowd": 0, "area": 5378.879700000002, "rle": {"size": [427, 640], "counts": "]^Q79Z:8lGJQ89nGHo7<oGEo7=oGDo7?PHAo7a0oG@o7c0PH]Oo7e0oG\\OP8f0oGZOP8k0kGWOT8X1]GhNb8Y1]GhNc8Y1\\GgNd8Y1\\GgNc8Z1\\GgNd8Z1[GfNd8[1[GfNe8Z1[GfNe8[1YGfNf8[1ZGeNf8[1ZGeNe8]1ZGcNe8^1[GbNd8`1ZGaNe8`1[G`Nd8b1[G^Nd8c1\\G]Nc8e1[G\\Nd8l1UGTNj8S2oFnMQ9T2mFkMT9W2iFjMV9Y2hFgMX9Z2fFgMZ9U300000000000K4K6M3bLXFV3k9hLXFT3Q:XNi1D<H7J7I]d6EP\\I4L5K4WOj0K4L4L5L3N3M2N3OWOlDUOY;h0g000O100000O10O1000000O1000000O10001N6HdM"}, "label": "a woman wearing glasses and green pajamas"}]}
{"id": 537621, "image": "COCO_train2014_000000537621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little boy in red footies sitting on the floor\"."}], "task": "refering", "answer_template": "The \"little boy in red footies sitting on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319], "bbox": [0.3973125, 0.06756440281030446, 0.916078125, 0.90576112412178], "iscrowd": 0, "area": 53842.24044999997, "rle": {"size": [427, 640], "counts": "PWZ32T=6J7I6_L_OQJg0l5FdI`0Z6LVI:g64hH2V7:ZHLc7a0mGFP8g0_G_O^8i2N2N2M3N2M3N3L3N2M3N2M3O1N3N1O1N2O1N2O1O2M2O0O101N10000O100O10000O100O100O10000O100O100dM[JRMe5b2[KiLf4k2gKSMZ4i2kKVMV4i2jKVMW4i2kKUMV4k2kKTMU4l2lKRMU4m2mKQMT4o2mKPMS4o2oKoLR4Q3oKmLR4S3oKlLQ4S3QLkLP4U3PLjLQ4V3PLiLP4V3RLhLo3X3RLfLo3Y3SLfLm3Z3TLdLm3\\3Q3O100000001N100000000O1000000O10000000000001O0000001O01O0001O000000001O0000001O0000001O000000001O0001O01O0000001O1O1O1O1O1O1O1O1bHZLU5g3iJ^LS5c3jJdLQ5]3mJiLn4W3PKoLl4R3RKTMi4m2UKYMf4h2WK_Md4b2ZKcMc4]2ZKjMa4W2[KQN`4R2YKVNc4n1TK[Nh4j1nJ\\NQ5h1gJ^NW5f1`J`N_5h4N2N1O2N2N1O2N2N101N101O1N101O1N101O1N101O0O2O1N101O1N101O1N101N2N1O2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N2O1N2N101N2N2N2O1N2N1O2O1N2N2N2O0O2N2N2N101N1O1O1O2O0O1O1O2N1O100O2N1O1O101N1O1O1O2N100O00001O0001O01O000010O0001O000010O000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O0000001O0000001O00001N1O1O1N3N1O1N2O2M2O1O2M2O1O1N3N1O1N2O2M2KhUf0"}, "label": "little boy in red footies sitting on the floor"}]}
{"id": 537621, "image": "COCO_train2014_000000537621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in moose pajamas\"."}], "task": "refering", "answer_template": "The \"the boy in moose pajamas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 57, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319], "bbox": [0.3973125, 0.06756440281030446, 0.916078125, 0.90576112412178], "iscrowd": 0, "area": 53842.24044999997, "rle": {"size": [427, 640], "counts": "PWZ32T=6J7I6_L_OQJg0l5FdI`0Z6LVI:g64hH2V7:ZHLc7a0mGFP8g0_G_O^8i2N2N2M3N2M3N3L3N2M3N2M3O1N3N1O1N2O1N2O1O2M2O0O101N10000O100O10000O100O100O10000O100O100dM[JRMe5b2[KiLf4k2gKSMZ4i2kKVMV4i2jKVMW4i2kKUMV4k2kKTMU4l2lKRMU4m2mKQMT4o2mKPMS4o2oKoLR4Q3oKmLR4S3oKlLQ4S3QLkLP4U3PLjLQ4V3PLiLP4V3RLhLo3X3RLfLo3Y3SLfLm3Z3TLdLm3\\3Q3O100000001N100000000O1000000O10000000000001O0000001O01O0001O000000001O0000001O0000001O000000001O0001O01O0000001O1O1O1O1O1O1O1O1bHZLU5g3iJ^LS5c3jJdLQ5]3mJiLn4W3PKoLl4R3RKTMi4m2UKYMf4h2WK_Md4b2ZKcMc4]2ZKjMa4W2[KQN`4R2YKVNc4n1TK[Nh4j1nJ\\NQ5h1gJ^NW5f1`J`N_5h4N2N1O2N2N1O2N2N101N101O1N101O1N101O1N101O0O2O1N101O1N101O1N101N2N1O2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N2O1N2N101N2N2N2O1N2N1O2O1N2N2N2O0O2N2N2N101N1O1O1O2O0O1O1O2N1O100O2N1O1O101N1O1O1O2N100O00001O0001O01O000010O0001O000010O000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O0000001O0000001O00001N1O1O1N3N1O1N2O2M2O1O2M2O1O1N3N1O1N2O2M2KhUf0"}, "label": "the boy in moose pajamas"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli is in a container next to cauliflower and cucumbers\"."}], "task": "refering", "answer_template": "The \"broccoli is in a container next to cauliflower and cucumbers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 69, 70, 71, 72, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 210, 211, 212, 234, 235], "bbox": [0.003015625, 0.14035416666666667, 0.370125, 0.6134583333333334], "iscrowd": 0, "area": 26604.91004999999, "rle": {"size": [480, 640], "counts": "eR1k0^<g1@`00O1O1O1O1O1O1O1O1O100O100O1O10O01000000000001O1O2N100O1O2N1001M2O2M2O2\\DZMi:h2VE\\Mf:f2XE^Mc:c2\\EaMa:`2^EdMU:@`En2:dMS:AaEl2;fMP:AbEj2=hMn9c2QF_Ml9c2SF`Mi9a2VFbMh9_2WFdMe9^2ZFdMc9^2\\FdMa9]2^FfM`9[2_FfMa9Z2^FgMa9Y2_FiM_9X2_FjM`9W2_FjM`9V2`FlM^9U2aFlM^9m3NO1O1N2O1O100O1000001O1N3N1O1O1O2N1N2O2N2N2N2N1O2N2O1O100O0000000000hEeKi9\\4PFlKo9d4O2N1O2N1O2N1O1O0000O100O00100O100O100O010O100O1O100O010O100O10O0000000000O1O1O1O100O1O1O1O100O1O1O1O1O01O01O003M2O1M4M2N2N2N2N1N3N2N1O2N2M2O2N2N1O2N2M3N3M4L3M3M2M3N2N1O2N2M2O2N2N2N1O2M3N1O2N2N2M2O2N2N1N2O1O1O1N2O1O1O1O1N2O001O1N2O1O1O1N2O1O1O1O1N3N1O1O2M2O1O2N1N3N1O1O2N1N2O2N1O`gl5"}, "label": "broccoli is in a container next to cauliflower and cucumbers"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli floret with other vegetables in a plastic container\"."}], "task": "refering", "answer_template": "The \"broccoli floret with other vegetables in a plastic container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 69, 70, 71, 72, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 210, 211, 212, 234, 235], "bbox": [0.003015625, 0.14035416666666667, 0.370125, 0.6134583333333334], "iscrowd": 0, "area": 26604.91004999999, "rle": {"size": [480, 640], "counts": "eR1k0^<g1@`00O1O1O1O1O1O1O1O1O100O100O1O10O01000000000001O1O2N100O1O2N1001M2O2M2O2\\DZMi:h2VE\\Mf:f2XE^Mc:c2\\EaMa:`2^EdMU:@`En2:dMS:AaEl2;fMP:AbEj2=hMn9c2QF_Ml9c2SF`Mi9a2VFbMh9_2WFdMe9^2ZFdMc9^2\\FdMa9]2^FfM`9[2_FfMa9Z2^FgMa9Y2_FiM_9X2_FjM`9W2_FjM`9V2`FlM^9U2aFlM^9m3NO1O1N2O1O100O1000001O1N3N1O1O1O2N1N2O2N2N2N2N1O2N2O1O100O0000000000hEeKi9\\4PFlKo9d4O2N1O2N1O2N1O1O0000O100O00100O100O100O010O100O1O100O010O100O10O0000000000O1O1O1O100O1O1O1O100O1O1O1O1O01O01O003M2O1M4M2N2N2N2N1N3N2N1O2N2M2O2N2N1O2N2M3N3M4L3M3M2M3N2N1O2N2M2O2N2N2N1O2M3N1O2N2N2M2O2N2N1N2O1O1O1N2O1O1O1O1N2O001O1N2O1O1O1N2O1O1O1O1N3N1O1O2M2O1O2N1N3N1O1O2N1N2O2N1O`gl5"}, "label": "broccoli floret with other vegetables in a plastic container"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a container holding broccoli , cauliflower , cucumber , and carrots\"."}], "task": "refering", "answer_template": "The \"a container holding broccoli , cauliflower , cucumber , and carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257], "bbox": [0.0, 0.00022916666666666666, 0.4843125, 0.6930416666666667], "iscrowd": 0, "area": 78901.50859999999, "rle": {"size": [480, 640], "counts": "R8i1g9`3bM_2M2O2N1O2N1O2N1O2N1O2N1N2O1O2N1O1O1O2N1O1O2M2O1O1O2N1O1O1O2N1O1N2O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N001O1O1O1O1000000000000000O100000000000000000000000000O1000000000000000000000000O10000O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O10O0100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N101O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N3N3M3L3N3L3N3M2M4L3L5J6J6K6I6K6I7I6K6I7Ildj4"}, "label": "a container holding broccoli , cauliflower , cucumber , and carrots"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the container holding fresh vegetables\"."}], "task": "refering", "answer_template": "The \"the container holding fresh vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257], "bbox": [0.0, 0.00022916666666666666, 0.4843125, 0.6930416666666667], "iscrowd": 0, "area": 78901.50859999999, "rle": {"size": [480, 640], "counts": "R8i1g9`3bM_2M2O2N1O2N1O2N1O2N1O2N1N2O1O2N1O1O1O2N1O1O2M2O1O1O2N1O1O1O2N1O1N2O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N001O1O1O1O1000000000000000O100000000000000000000000000O1000000000000000000000000O10000O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O10O0100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N101O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N3N3M3L3N3L3N3M2M4L3L5J6J6K6I6K6I7I6K6I7Ildj4"}, "label": "the container holding fresh vegetables"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bowl filled with cooked rice\"."}], "task": "refering", "answer_template": "The \"a white bowl filled with cooked rice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 249, 250, 251, 252], "bbox": [0.49604687500000005, 0.003958333333333333, 1.0, 0.6495416666666666], "iscrowd": 0, "area": 61615.975399999996, "rle": {"size": [480, 640], "counts": "Qdd45j>o0QO5K2N2M4M2N2N2N[1eN1N2O1O1O1O1O1O1N2O001O1O1O1N2O1O1O100000001O000000001O000000001O00000000001O000000001O000000001O000000001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O1O1O001O1O5K7I6J6J6J5K2N3M2N2N1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O100O10O10O100O10000O01000O100O10O10O100O1000O0100O10000O10O010000O100O01000O010O10O10O010O10O100O10O010000O100O10O10O100O10000O010O10000O10O107I7I00O1O1N101O1O1O1O1O1O001O1N2O1O1O001O1O1O1O1O1N101O1O1O1O1O1O001O1N2O1O1O1O001O1O1O1O1N101O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O001N2O1O1O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O0O2OmK"}, "label": "a white bowl filled with cooked rice"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white fluffy rice is a good compliment to the stir fry\"."}], "task": "refering", "answer_template": "The \"white fluffy rice is a good compliment to the stir fry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 249, 250, 251, 252], "bbox": [0.49604687500000005, 0.003958333333333333, 1.0, 0.6495416666666666], "iscrowd": 0, "area": 61615.975399999996, "rle": {"size": [480, 640], "counts": "Qdd45j>o0QO5K2N2M4M2N2N2N[1eN1N2O1O1O1O1O1O1N2O001O1O1O1N2O1O1O100000001O000000001O000000001O00000000001O000000001O000000001O000000001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O1O1O001O1O5K7I6J6J6J5K2N3M2N2N1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O100O10O10O100O10000O01000O100O10O10O100O1000O0100O10000O10O010000O100O01000O010O10O10O010O10O100O10O010000O100O10O10O100O10000O010O10000O10O107I7I00O1O1N101O1O1O1O1O1O001O1N2O1O1O001O1O1O1O1O1N101O1O1O1O1O1O001O1N2O1O1O1O001O1O1O1O1N101O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O001N2O1O1O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O0O2OmK"}, "label": "white fluffy rice is a good compliment to the stir fry"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the food with the brown sauce\"."}], "task": "refering", "answer_template": "The \"the food with the brown sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.189453125, 0.5638124999999999, 0.8828125, 0.9778749999999999], "iscrowd": 0, "area": 75251.5625, "rle": {"size": [480, 640], "counts": "nQi14i>`0@?A`0@?A`0@?A6J5K5K5K4N3M3M3N2M3M2N3N1N2N2N1O2O1N1O2N2N101N2N1O2O1N1O2N1O2O1N1O2O1O001O00000000000000000O10000000000000000000O100000O100000000000000000000000000O10000000O100000000000000000O100000000000000000000000O10O100000000000000000000000000O100000000000O10000000000000O10000000000000000000000000O0100000000000000000000000000O10000000000000O100000000000O100000000000000000000000000O10O100000000000000000000000O100000000000000000O10000000O100000000000000000000000000O1000O1000000000000000000000O10000000000000000000O100000O100000000000000000000000000O10000000O100000000000000000O100000000000000000000000O10O100000000000000000000000000O1000000000O1000000000000000O100000000000000000000000O10O100000000000000000000000O10O100000000000000000000000O100000000000001O00000000000O101O00000000001O1O0O2N2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N101N2N1O2O1N1O2O1N2N4L4M2M4L4L3N3L4L3M4M3K4I<D<D<D<D=C<D<D]iR1"}, "label": "the food with the brown sauce"}]}
{"id": 242709, "image": "COCO_train2014_000000242709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cooked vegetables with a brown sauce in a white container\"."}], "task": "refering", "answer_template": "The \"cooked vegetables with a brown sauce in a white container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.189453125, 0.5638124999999999, 0.8828125, 0.9778749999999999], "iscrowd": 0, "area": 75251.5625, "rle": {"size": [480, 640], "counts": "nQi14i>`0@?A`0@?A`0@?A6J5K5K5K4N3M3M3N2M3M2N3N1N2N2N1O2O1N1O2N2N101N2N1O2O1N1O2N1O2O1N1O2O1O001O00000000000000000O10000000000000000000O100000O100000000000000000000000000O10000000O100000000000000000O100000000000000000000000O10O100000000000000000000000000O100000000000O10000000000000O10000000000000000000000000O0100000000000000000000000000O10000000000000O100000000000O100000000000000000000000000O10O100000000000000000000000O100000000000000000O10000000O100000000000000000000000000O1000O1000000000000000000000O10000000000000000000O100000O100000000000000000000000000O10000000O100000000000000000O100000000000000000000000O10O100000000000000000000000000O1000000000O1000000000000000O100000000000000000000000O10O100000000000000000000000O10O100000000000000000000000O100000000000001O00000000000O101O00000000001O1O0O2N2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N101N2N1O2O1N1O2O1N2N4L4M2M4L4L3N3L4L3M4M3K4I<D<D<D<D=C<D<D]iR1"}, "label": "cooked vegetables with a brown sauce in a white container"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the black pants\"."}], "task": "refering", "answer_template": "The \"the person in the black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 64, 65], "bbox": [0.5484600000000001, 0.0, 0.8558, 0.2439], "iscrowd": 0, "area": 11520.540299999999, "rle": {"size": [400, 500], "counts": "PQ[34k09Y:T1K6J5K6J5K6J2N000000000000000000002N2N1O2N2N1O2N2N2N1O1O000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000M3K5K5K5K5K6J5K5K5K5K5K5L4K5O13M3M3M3M3M3M3M4L3M5K5K4L5K5K4L5K4L0000000000000000000000001O000000000000000000000000000000000000O10000000000000000000003M3M4L3M3M4L3M3M4K4L7I8H8H7I8H8HgQl0"}, "label": "the person in the black pants"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone in dark pants standing behind a table full of food\"."}], "task": "refering", "answer_template": "The \"someone in dark pants standing behind a table full of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 64, 65], "bbox": [0.5484600000000001, 0.0, 0.8558, 0.2439], "iscrowd": 0, "area": 11520.540299999999, "rle": {"size": [400, 500], "counts": "PQ[34k09Y:T1K6J5K6J5K6J2N000000000000000000002N2N1O2N2N1O2N2N2N1O1O000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000M3K5K5K5K5K6J5K5K5K5K5K5L4K5O13M3M3M3M3M3M3M4L3M5K5K4L5K5K4L5K4L0000000000000000000000001O000000000000000000000000000000000000O10000000000000000000003M3M4L3M3M4L3M3M4K4L7I8H8H7I8H8HgQl0"}, "label": "someone in dark pants standing behind a table full of food"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a black shirt sitting with their back to the table\"."}], "task": "refering", "answer_template": "The \"a person in a black shirt sitting with their back to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 51, 52, 53, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 230, 231, 232, 233, 249, 250, 251], "bbox": [0.75866, 0.08539999999999999, 0.9941599999999999, 0.986525], "iscrowd": 0, "area": 32804.19195000001, "rle": {"size": [400, 500], "counts": "lZd42l;e0mNR1gNZ1C<I8J5L5J5L5J5L5J5L5J5L5J5L5J5L5J5L5J5L5J4M2M4L4M2M4M3L3N3L3N3L4M2M4L4RO^HjLd7V3l0N1O2O1N1O2O1N1O2O0O2N2O0O2O1N1O2O0O2N2O0O2N2O0O2N101N2N101N101N1O101N1O2O0O1O2O0O2N100O2N010O100O1O100O1O10O01O100O1O100O1O10O0100O1O100O1O100O00100O13M3jK[X1"}, "label": "a person in a black shirt sitting with their back to the table"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the black sweatshirt\"."}], "task": "refering", "answer_template": "The \"the person in the black sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 51, 52, 53, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 230, 231, 232, 233, 249, 250, 251], "bbox": [0.75866, 0.08539999999999999, 0.9941599999999999, 0.986525], "iscrowd": 0, "area": 32804.19195000001, "rle": {"size": [400, 500], "counts": "lZd42l;e0mNR1gNZ1C<I8J5L5J5L5J5L5J5L5J5L5J5L5J5L5J5L5J5L5J4M2M4L4M2M4M3L3N3L3N3L4M2M4L4RO^HjLd7V3l0N1O2O1N1O2O1N1O2O0O2N2O0O2O1N1O2O0O2N2O0O2N2O0O2N101N2N101N101N1O101N1O2O0O1O2O0O2N100O2N010O100O1O100O1O10O01O100O1O100O1O10O0100O1O100O1O100O00100O13M3jK[X1"}, "label": "the person in the black sweatshirt"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plastic cup with liquid in it\"."}], "task": "refering", "answer_template": "The \"plastic cup with liquid in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 168, 169, 184, 185, 186, 187, 202, 203, 204, 205, 220, 221, 222, 239, 240], "bbox": [0.23156000000000002, 0.64035, 0.4397, 0.9976499999999998], "iscrowd": 0, "area": 9087.78265, "rle": {"size": [400, 500], "counts": "jb]1;R<4L4N3M2N2N2N4L5K5K5K6K4L4K5L4K5L6J5J6K5J7J7I7H8I5J2O1N2O1O000O10000000001O0O100000000000000O1000000000000O10aNTHZNl7`1\\H^Nd7^1bH`N_7^1dHaN[7^1iH_NW7`1lH^NT7a1oH]NQ7b1RI\\Nn6c1VIZNj6f1ZIVNg6h1^ITNb6k1dIPN\\6o1iImMW6R2nIjMR6U2U2O2O0O10000O100O2O000O10000O2O0O2O001N101O0O2O001N101O0G;D<I6M4L4L4GaS]3"}, "label": "plastic cup with liquid in it"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty plastic cup\"."}], "task": "refering", "answer_template": "The \"empty plastic cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 153, 154, 155, 156, 171, 172, 173, 174, 189, 190, 191, 192, 207, 208, 209, 210, 226, 227], "bbox": [0.5102599999999999, 0.55925, 0.71392, 0.9040250000000001], "iscrowd": 0, "area": 9478.321700000002, "rle": {"size": [400, 500], "counts": "ZkS31]<4L4L4L4L4L5K7J6I7K5K5K5J5L5K5J6K5K4L3L4M3M00N1O2N2N2N1O2NnF`MX8^2iGjMP8T2PHRNk7m1VHTNj7k1VHWNi7h1VHZNi7f1WH[Ni7d1VH^Nj7a1VHaNi7^1WHcNh7]1WHeNi7Z1WHhNh7W1WHkNi7T1WHmNh7S1XHoNg7P1XHROh7m0XHTOh7k0WHXOh7g0XHZOh7e0XH\\Oh7c0WH_Oj7?VHBj7=UHEk7:UHGk78UHIk75UHMk72UHOl7OTH2l7NRH4n7MPH4P8NmG3S8OjG2V8T21O000001O0001N1O1O1O1O1O2N1O1O1O1O1O2L3K5K5L4K5K6J5L4K6J6J6K5J6J6J6K5J5L5M3N2N2Nncg1"}, "label": "empty plastic cup"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty plastic cup immediately to the right of another cup half full of orange liquid\"."}], "task": "refering", "answer_template": "The \"an empty plastic cup immediately to the right of another cup half full of orange liquid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 153, 154, 155, 156, 171, 172, 173, 174, 189, 190, 191, 192, 207, 208, 209, 210, 226, 227], "bbox": [0.5102599999999999, 0.55925, 0.71392, 0.9040250000000001], "iscrowd": 0, "area": 9478.321700000002, "rle": {"size": [400, 500], "counts": "ZkS31]<4L4L4L4L4L5K7J6I7K5K5K5J5L5K5J6K5K4L3L4M3M00N1O2N2N2N1O2NnF`MX8^2iGjMP8T2PHRNk7m1VHTNj7k1VHWNi7h1VHZNi7f1WH[Ni7d1VH^Nj7a1VHaNi7^1WHcNh7]1WHeNi7Z1WHhNh7W1WHkNi7T1WHmNh7S1XHoNg7P1XHROh7m0XHTOh7k0WHXOh7g0XHZOh7e0XH\\Oh7c0WH_Oj7?VHBj7=UHEk7:UHGk78UHIk75UHMk72UHOl7OTH2l7NRH4n7MPH4P8NmG3S8OjG2V8T21O000001O0001N1O1O1O1O1O2N1O1O1O1O1O2L3K5K5L4K5K6J5L4K6J6J6K5J6J6J6K5J5L5M3N2N2Nncg1"}, "label": "an empty plastic cup immediately to the right of another cup half full of orange liquid"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza not cut , at top of the picture\"."}], "task": "refering", "answer_template": "The \"pizza not cut , at top of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98], "bbox": [0.1618, 0.24045000000000002, 0.50158, 0.413475], "iscrowd": 0, "area": 9611.907299999997, "rle": {"size": [400, 500], "counts": "kho01^<2O1M3M3M3M3M3K5K5M3L3O2O1N1O2N101N1O101N1O2O0O101N101N100O101N100O100O101O0O100O100O100O2O00000000001O0000000000001O000000000000001O00000000001O00000000000000000001O0000000000000000000000000000000000000000000000000O1000000O1000000O1000000O1000000000000O1000000000000000000O1000000000000O100O101N101O0O2O0O1O2O0O2N2N2O1M3N3M2M4M3J6K5K6H7JjTQ3"}, "label": "pizza not cut , at top of the picture"}]}
{"id": 283673, "image": "COCO_train2014_000000283673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza behind another pizza\"."}], "task": "refering", "answer_template": "The \"a pizza behind another pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98], "bbox": [0.1618, 0.24045000000000002, 0.50158, 0.413475], "iscrowd": 0, "area": 9611.907299999997, "rle": {"size": [400, 500], "counts": "kho01^<2O1M3M3M3M3M3K5K5M3L3O2O1N1O2N101N1O101N1O2O0O101N101N100O101N100O100O101O0O100O100O100O2O00000000001O0000000000001O000000000000001O00000000001O00000000000000000001O0000000000000000000000000000000000000000000000000O1000000O1000000O1000000O1000000000000O1000000000000000000O1000000000000O100O101N101O0O2O0O1O2O0O2N2N2O1M3N3M2M4M3J6K5K6H7JjTQ3"}, "label": "a pizza behind another pizza"}]}
{"id": 226329, "image": "COCO_train2014_000000226329.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beige love seat facing fire place\"."}], "task": "refering", "answer_template": "The \"beige love seat facing fire place\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 329, 330, 331], "bbox": [0.101453125, 0.4290248962655602, 0.55996875, 0.8823858921161826], "iscrowd": 0, "area": 42570.0764, "rle": {"size": [482, 640], "counts": "djn0Q1P>1O1O1O1O2N2N2N1WChNg;Z1XDjNc;X1[DlN`;W1_DlN[;X1cDlNX;W1gDlNT;W1jDmNQ;V1nDmNl:W1REnNh:U1WEnNd:T1[EPO`:S1^EQO^:Q1aERO[:P1cETOZ:m0eEVOX:k0fEWOY:j0fEWOY:j0eEWO[:j0dEWO[:j0cEWO\\:j0dEWO[:i2O001O1O001O0O2O1O001O2N1O1N2O1O1O001O00000010O0001O0001O0O1N2N2O20O001O01O01O001O01O01O001O010O00001O010O00001O010O000010O01O000010O01O000010O01O000010O01O000010O01O00010O001O00010O001O00010O001O00010O001O00010O001O01O01O001O01O01O001O01O01O001O01O01O0010O0010O0010O0010O0010O0010O3M5L2M01O00001O001O001O2N2N2N2N2N2N2N2N1O00001O000000001N1F:L4O101N100O1O2O0O100O1O2O0O100O1O2O0O100O101N1O100O2O0O1O100O2O0O1O100O2O0O010O1O10O0100O1O010O100O00100O10O01O100O010O100O002O0O2O0O1O2O0O101N1O101N100O2O0O1O2O1N2O2M3M2M4K5K4L5K4J=C`0\\O`^T4"}, "label": "beige love seat facing fire place"}]}
{"id": 226329, "image": "COCO_train2014_000000226329.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandal color three seater sofa placed in a living room\"."}], "task": "refering", "answer_template": "The \"a sandal color three seater sofa placed in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 208, 209], "bbox": [0.0016250000000000001, 0.2790871369294606, 0.47232812500000004, 0.5678838174273859], "iscrowd": 0, "area": 20375.489000000005, "rle": {"size": [482, 640], "counts": "nd05_>MjA;P>HmA=c0B^<4kB`0=De<l0SCXOl<_1O2N100O1O100O2O1O003M1O001O001N2O0O2O0O2O1O0O2O001O1O001O1O0O2O001O1N101O0O2O1O001N2O001O001N2O001O001O1N101O[O^DkMa;T2cDiM];U2hDiMW;U2mDiMS;V2QEhMn:V2VEiMi:W2ZEfMf:Y2^EeMa:[2aEcM_:]2dEaM[:^2hEaMW:_2U10100O100O1K5K5N2N2N10100O1O100O1O100O00100O100O2O0O100O2O0O2O1ON201O00000O100000001O0O10001O000O101O0000000O2O0000001O0O10001O000O101O0000000O2O0000001O0O10001O00000O101O00000O2O000O1N3O0O1000001O000O101O0001O10O01O1OO101O0000000O10001O000O10000000001N100000000O1000001O0O10000000000O2O0000000O10O10000O10O01000000O2O2N1N5L4L5K5J2O00000O1000000O1000000O1000000O10000O1000<C101O0O10001N10000O2O000O10000O10000O10000O10000O2O1O0O2O1O3L3N7I1N101O1N102N4K3N00n[o4"}, "label": "a sandal color three seater sofa placed in a living room"}]}
{"id": 226329, "image": "COCO_train2014_000000226329.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large couch sitting along side of blue wall in front of fireplace\"."}], "task": "refering", "answer_template": "The \"large couch sitting along side of blue wall in front of fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 208, 209], "bbox": [0.0016250000000000001, 0.2790871369294606, 0.47232812500000004, 0.5678838174273859], "iscrowd": 0, "area": 20375.489000000005, "rle": {"size": [482, 640], "counts": "nd05_>MjA;P>HmA=c0B^<4kB`0=De<l0SCXOl<_1O2N100O1O100O2O1O003M1O001O001N2O0O2O0O2O1O0O2O001O1O001O1O0O2O001O1N101O0O2O1O001N2O001O001N2O001O001O1N101O[O^DkMa;T2cDiM];U2hDiMW;U2mDiMS;V2QEhMn:V2VEiMi:W2ZEfMf:Y2^EeMa:[2aEcM_:]2dEaM[:^2hEaMW:_2U10100O100O1K5K5N2N2N10100O1O100O1O100O00100O100O2O0O100O2O0O2O1ON201O00000O100000001O0O10001O000O101O0000000O2O0000001O0O10001O000O101O0000000O2O0000001O0O10001O00000O101O00000O2O000O1N3O0O1000001O000O101O0001O10O01O1OO101O0000000O10001O000O10000000001N100000000O1000001O0O10000000000O2O0000000O10O10000O10O01000000O2O2N1N5L4L5K5J2O00000O1000000O1000000O1000000O10000O1000<C101O0O10001N10000O2O000O10000O10000O10000O10000O2O1O0O2O1O3L3N7I1N101O1N102N4K3N00n[o4"}, "label": "large couch sitting along side of blue wall in front of fireplace"}]}
{"id": 226329, "image": "COCO_train2014_000000226329.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown couch that is facing the fireplace\"."}], "task": "refering", "answer_template": "The \"the brown couch that is facing the fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332], "bbox": [0.103671875, 0.4292116182572614, 0.5555468749999999, 0.8863692946058092], "iscrowd": 0, "area": 42488.3478, "rle": {"size": [482, 640], "counts": "fYo03n>m0TO4K2N2O1N2N2O1N2UCbNo;`1PDeNk;\\1SDkNf;W1YDnNa;T1]DRO^;o0aDWOV;m0hDYOa:Z1^EkN[:X1dEnNV:S1hEPOV:Q1iEoNW:R1gEPOX:Q1gEoNX:S1gEnNX:S1fEnNZ:S1eEnNZ:S1dEnN[:n2O001O000O2O001O001O0010O01O00001O01O0001O000010O000001O00010O00001O00010O001O001O010O001O0010O01O00001O010O001O001O010O001O001O010O001O001O010O00001O0010O01O001O010O001O001O010O001O001O010O00001O0010O01O001O0010O01O001O0010O01O001O0010O0001O0010O01O001O0010O01O001O0010O01O001O0010O01O00001O0100O2O0O2O0O1O2N1O2N1O2M2O1N3N1O2M2O2M2O1O2M10000O2O0O1O100O2N100O1O101N1O100O1O2O0O1O100O2N100O1O101N1O100O1O2O0O1O100O2N100O1O101N1O100O1O2O0O001O100O00100O1O010O1O10O01O100O00100O1N101O1N2O0O5L3M4K4M4K4M4L3L5L3L5L3M4K5L5K4K5H8F:F:FP\\U4"}, "label": "the brown couch that is facing the fireplace"}]}
{"id": 226329, "image": "COCO_train2014_000000226329.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller couch\"."}], "task": "refering", "answer_template": "The \"smaller couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332], "bbox": [0.103671875, 0.4292116182572614, 0.5555468749999999, 0.8863692946058092], "iscrowd": 0, "area": 42488.3478, "rle": {"size": [482, 640], "counts": "fYo03n>m0TO4K2N2O1N2N2O1N2UCbNo;`1PDeNk;\\1SDkNf;W1YDnNa;T1]DRO^;o0aDWOV;m0hDYOa:Z1^EkN[:X1dEnNV:S1hEPOV:Q1iEoNW:R1gEPOX:Q1gEoNX:S1gEnNX:S1fEnNZ:S1eEnNZ:S1dEnN[:n2O001O000O2O001O001O0010O01O00001O01O0001O000010O000001O00010O00001O00010O001O001O010O001O0010O01O00001O010O001O001O010O001O001O010O001O001O010O00001O0010O01O001O010O001O001O010O001O001O010O00001O0010O01O001O0010O01O001O0010O01O001O0010O0001O0010O01O001O0010O01O001O0010O01O001O0010O01O00001O0100O2O0O2O0O1O2N1O2N1O2M2O1N3N1O2M2O2M2O1O2M10000O2O0O1O100O2N100O1O101N1O100O1O2O0O1O100O2N100O1O101N1O100O1O2O0O1O100O2N100O1O101N1O100O1O2O0O001O100O00100O1O010O1O10O01O100O00100O1N101O1N2O0O5L3M4K4M4K4M4L3L5L3L5L3M4K5L5K4K5H8F:F:FP\\U4"}, "label": "smaller couch"}]}
{"id": 111646, "image": "COCO_train2014_000000111646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a metal bowl full of fried food next to a plastic bowl with green food\"."}], "task": "refering", "answer_template": "The \"a metal bowl full of fried food next to a plastic bowl with green food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.47814988290398125, 0.871890625, 0.9489695550351288, 1.0], "iscrowd": 0, "area": 13311.769500000002, "rle": {"size": [640, 427], "counts": "dRP41nc05J6K4K6K5K3L3N2N2M3N2N2M2O2N2M3N2M3N2N2M3N101N2O1O1N2O1N2O1O0O10000O100O1000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O001O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O2N1O2N2N2N2N1O2N2L4M2N3L4M3L4M2N3LaT="}, "label": "a metal bowl full of fried food next to a plastic bowl with green food"}]}
{"id": 111646, "image": "COCO_train2014_000000111646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"spicy curry in a plate\"."}], "task": "refering", "answer_template": "The \"spicy curry in a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.47814988290398125, 0.871890625, 0.9489695550351288, 1.0], "iscrowd": 0, "area": 13311.769500000002, "rle": {"size": [640, 427], "counts": "dRP41nc05J6K4K6K5K3L3N2N2M3N2N2M2O2N2M3N2M3N2N2M3N101N2O1O1N2O1N2O1O0O10000O100O1000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O001O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O2N1O2N2N2N2N1O2N2L4M2N3L4M3L4M2N3LaT="}, "label": "spicy curry in a plate"}]}
{"id": 103455, "image": "COCO_train2014_000000103455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a lady wearing a baseball cap , gray sweatshirt , pants and a white apron operating a doughnut making machine\"."}], "task": "refering", "answer_template": "The \"this is a lady wearing a baseball cap , gray sweatshirt , pants and a white apron operating a doughnut making machine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 94, 95, 96, 97, 117, 118, 119, 123, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 278, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 347, 348, 349, 350, 370, 371, 372], "bbox": [0.055625, 0.20675, 0.391015625, 1.0], "iscrowd": 0, "area": 32466.702800000003, "rle": {"size": [480, 640], "counts": "bQa0:e>4M2M3N2N2M3N2M2O1N101O1N2O1N101O3nEkNR6Y1gInNV6U1UHaNAb0U8Q1THaNBe0W8n0PHbNCg0Z8j0mGcNDk0Z8g0kGcNFl0\\8P1\\GXN4l0_8T2]GQN`8R2[GRNd8o1YGUNe8l1XGWNf8l1WGZNd8W4YNdI^Jb6W5eIfJb6Q5bInJk6d4XIZK]7P4fHnKn7^3TH`LZ8Q3hGnLX8Q3kGmLU8S3mGkLS8T3\\H^Le7`3_H\\Lb7d3_H[La7d3aH[L_7e3bHZL^7f3dHXL\\7h3eHWL[7h3gHWLY7i3iHULm7U3o100O1000000O1000000001O000001O0000000001O000000000`FlLT7T3lHmLS7S3mHnLR7R3oHoLo6R3PIoLo6Q3QIPMn6Q3QIPMn6P3RIQMm6P3SIoLm6Q3SIoLn6Q3QIoLo6Q3PIQMo6P3PIPMP7P3oHQMQ7P3nHQMQ7o2nHRMR7o2lHRMT7n2jHTMV7m2hHUMW7k2gHWMX7k2eHWMR7S3mHmLi6`3SIbLc6h3\\IXLb6k3]IULb6m3fH^KJf0^7o3gH^KHd0_7P4gHaKE`0c7Q4gHbKC=e7S4gHdLW7^3hHbLW7`3gHaLo0^Ng4T5YJ^LR1`N`4T5]J]LR1dN[4Q5bJ[LR1iNW4m4gJZLP1oNT4i4jJYLP1UO@DS3R5]LVLm0ZOYOJS3k4fLQLm0^OTOV1h1_3WNmKk0DnN`2?U2GgKj0V:UOlEh0W:WOiEh0X:XOhEf0Z:YOgEg0Z:WOhEh0X:WOiEi0X:UOiEk0W:TOkEk0U<000001N1000001O0000001O00001N0100O10001N10000O10000O2O0O10000O10000O01O00001O001O000010O0001O00001O001O00001O00001O010O00001O0000O1O1O100O1O1O100100O1O1O001O1O1^OoA4Q>HUB7[>O0O2O001N101O0OaSg5"}, "label": "this is a lady wearing a baseball cap , gray sweatshirt , pants and a white apron operating a doughnut making machine"}]}
{"id": 103455, "image": "COCO_train2014_000000103455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue cap , grey sweatshirt , and white apron\"."}], "task": "refering", "answer_template": "The \"a woman in a blue cap , grey sweatshirt , and white apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 94, 95, 96, 97, 117, 118, 119, 123, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 278, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 347, 348, 349, 350, 370, 371, 372], "bbox": [0.055625, 0.20675, 0.391015625, 1.0], "iscrowd": 0, "area": 32466.702800000003, "rle": {"size": [480, 640], "counts": "bQa0:e>4M2M3N2N2M3N2M2O1N101O1N2O1N101O3nEkNR6Y1gInNV6U1UHaNAb0U8Q1THaNBe0W8n0PHbNCg0Z8j0mGcNDk0Z8g0kGcNFl0\\8P1\\GXN4l0_8T2]GQN`8R2[GRNd8o1YGUNe8l1XGWNf8l1WGZNd8W4YNdI^Jb6W5eIfJb6Q5bInJk6d4XIZK]7P4fHnKn7^3TH`LZ8Q3hGnLX8Q3kGmLU8S3mGkLS8T3\\H^Le7`3_H\\Lb7d3_H[La7d3aH[L_7e3bHZL^7f3dHXL\\7h3eHWL[7h3gHWLY7i3iHULm7U3o100O1000000O1000000001O000001O0000000001O000000000`FlLT7T3lHmLS7S3mHnLR7R3oHoLo6R3PIoLo6Q3QIPMn6Q3QIPMn6P3RIQMm6P3SIoLm6Q3SIoLn6Q3QIoLo6Q3PIQMo6P3PIPMP7P3oHQMQ7P3nHQMQ7o2nHRMR7o2lHRMT7n2jHTMV7m2hHUMW7k2gHWMX7k2eHWMR7S3mHmLi6`3SIbLc6h3\\IXLb6k3]IULb6m3fH^KJf0^7o3gH^KHd0_7P4gHaKE`0c7Q4gHbKC=e7S4gHdLW7^3hHbLW7`3gHaLo0^Ng4T5YJ^LR1`N`4T5]J]LR1dN[4Q5bJ[LR1iNW4m4gJZLP1oNT4i4jJYLP1UO@DS3R5]LVLm0ZOYOJS3k4fLQLm0^OTOV1h1_3WNmKk0DnN`2?U2GgKj0V:UOlEh0W:WOiEh0X:XOhEf0Z:YOgEg0Z:WOhEh0X:WOiEi0X:UOiEk0W:TOkEk0U<000001N1000001O0000001O00001N0100O10001N10000O10000O2O0O10000O10000O01O00001O001O000010O0001O00001O001O00001O00001O010O00001O0000O1O1O100O1O1O100100O1O1O001O1O1^OoA4Q>HUB7[>O0O2O001N101O0OaSg5"}, "label": "a woman in a blue cap , grey sweatshirt , and white apron"}]}
{"id": 111649, "image": "COCO_train2014_000000111649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat closest to the boy\"."}], "task": "refering", "answer_template": "The \"the boat closest to the boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 135, 136, 137, 138, 140, 141, 142, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 171, 172, 173, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241], "bbox": [0.0033723653395784543, 0.36825, 1.0, 0.7015781249999999], "iscrowd": 0, "area": 32611.134700000002, "rle": {"size": [640, 427], "counts": "`k0^6b=0000000O101O000000000000000O1000000000000000000O2O0000iJeBi3[=WLfBh3Z=WLiBg3W=XLkBg3U=YLmBe3S=ZLoBe3Q=[LPCd3o<\\LRCd3n<[LSCe3m<[LSCe3m<ZLTCf3l<ZLTCf3l<ZLUCe3k<[LUCe3k<\\LTCd3l<\\LUCc3k<^LTCc3k<^LUCa3k<_LVC`3j<aLVC^3k<bLUC]3k<dLTCa2GgLU=i0TC7T2kNh:o0TCYOT3Di9R1TCoNb3JZ9X1SCeNQ4Nl8^1T510O01O001O010O1O0010O0000000010O000001O01O0001O000000001O0000010O000010O01O002O3L4L4L4M3L2N1001N3N1OO1O10O01O1O100O001O100N2O0O2O6I<E\\gb03\\X]O5M4M2M4L3H9K410O01000O6J5L4K1O2O0O2N101N1O2O0O2N101N1O2OO01O010O00010O0j^OoM]`0R2c_OPN[`0o1e_OUNW`0l1i_OVNT`0j1l_OYNR`0g1n_O\\Nn?d1R@_Nk?a1V@aNh?_1W@eNe?[1\\@gNa?Z1^@iN`?V1a@mN[?T1d@oNY?Q1h@YOn>g0QAGa>9`A4R>MmA4S>KnA4R>LnA4R>MnA2R>NnA2S>NmA1S>OmA1S>0mAOT>0lA0T>1lANT>2lANU>1lANT>3kAMU>3lALU>4jALV>4kAKU>6jAJW>5jAJV>7iAIW>7jAHW>dNTAl0d0`0X>aNWAo0b0?W>^N\\AR1=`0X>YN`AW18?X>VNeAZ13`0X>SNiA]1O?T?@m@?T?Ak@?V?@k@`0U?@j@`0W?_Oj@`0V?Ai@?X?@i@?X?@h@`0Y?@g@?Z?@f@`0[?@e@?\\?@d@`0]?@c@?^?@b@`0_?@a@?`?@`@`0`?@a@?`?A_@?b?@_@?b?A]@?d?@]@?d?A[@?f?@Z@`0g?@X@`0i?_OW@a0j?_OU@a0l?^OU@a0k?@T@`0m?_OS@a0n?_OQ@a0P`0_Oo_Oa0R`0^Oo_Oa0R`0_Om_Oa0T`0^Ol_Ob0U`0^Oj_Ob0W`0^Oh_Ob0Y`0]Oe_Oe0\\`0[O`_Oh0a`0WO\\_Ol0e`0TOW_Oo0i`0QOT_OR1m`0P11O1O1O0000001O00000000000O1O1O1O1O1O1O1O2NShg0jMUZXO:E<D<D;E<D;E4L000000000000O1000001O000000000000000000000O100000001O0000000000000000000O1000000000001O000000000000000O100000000000001O0000000000000O1000000000000000001O000000000O1000000D<hNRF"}, "label": "the boat closest to the boy"}]}
{"id": 111649, "image": "COCO_train2014_000000111649.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white yatch right behind the boy\"."}], "task": "refering", "answer_template": "The \"a white yatch right behind the boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 135, 136, 137, 138, 140, 141, 142, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 171, 172, 173, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241], "bbox": [0.0033723653395784543, 0.36825, 1.0, 0.7015781249999999], "iscrowd": 0, "area": 32611.134700000002, "rle": {"size": [640, 427], "counts": "`k0^6b=0000000O101O000000000000000O1000000000000000000O2O0000iJeBi3[=WLfBh3Z=WLiBg3W=XLkBg3U=YLmBe3S=ZLoBe3Q=[LPCd3o<\\LRCd3n<[LSCe3m<[LSCe3m<ZLTCf3l<ZLTCf3l<ZLUCe3k<[LUCe3k<\\LTCd3l<\\LUCc3k<^LTCc3k<^LUCa3k<_LVC`3j<aLVC^3k<bLUC]3k<dLTCa2GgLU=i0TC7T2kNh:o0TCYOT3Di9R1TCoNb3JZ9X1SCeNQ4Nl8^1T510O01O001O010O1O0010O0000000010O000001O01O0001O000000001O0000010O000010O01O002O3L4L4L4M3L2N1001N3N1OO1O10O01O1O100O001O100N2O0O2O6I<E\\gb03\\X]O5M4M2M4L3H9K410O01000O6J5L4K1O2O0O2N101N1O2O0O2N101N1O2OO01O010O00010O0j^OoM]`0R2c_OPN[`0o1e_OUNW`0l1i_OVNT`0j1l_OYNR`0g1n_O\\Nn?d1R@_Nk?a1V@aNh?_1W@eNe?[1\\@gNa?Z1^@iN`?V1a@mN[?T1d@oNY?Q1h@YOn>g0QAGa>9`A4R>MmA4S>KnA4R>LnA4R>MnA2R>NnA2S>NmA1S>OmA1S>0mAOT>0lA0T>1lANT>2lANU>1lANT>3kAMU>3lALU>4jALV>4kAKU>6jAJW>5jAJV>7iAIW>7jAHW>dNTAl0d0`0X>aNWAo0b0?W>^N\\AR1=`0X>YN`AW18?X>VNeAZ13`0X>SNiA]1O?T?@m@?T?Ak@?V?@k@`0U?@j@`0W?_Oj@`0V?Ai@?X?@i@?X?@h@`0Y?@g@?Z?@f@`0[?@e@?\\?@d@`0]?@c@?^?@b@`0_?@a@?`?@`@`0`?@a@?`?A_@?b?@_@?b?A]@?d?@]@?d?A[@?f?@Z@`0g?@X@`0i?_OW@a0j?_OU@a0l?^OU@a0k?@T@`0m?_OS@a0n?_OQ@a0P`0_Oo_Oa0R`0^Oo_Oa0R`0_Om_Oa0T`0^Ol_Ob0U`0^Oj_Ob0W`0^Oh_Ob0Y`0]Oe_Oe0\\`0[O`_Oh0a`0WO\\_Ol0e`0TOW_Oo0i`0QOT_OR1m`0P11O1O1O0000001O00000000000O1O1O1O1O1O1O1O2NShg0jMUZXO:E<D<D;E<D;E4L000000000000O1000001O000000000000000000000O100000001O0000000000000000000O1000000000001O000000000000000O100000000000001O0000000000000O1000000000000000001O000000000O1000000D<hNRF"}, "label": "a white yatch right behind the boy"}]}
{"id": 46118, "image": "COCO_train2014_000000046118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a four door sedan parked in a snowy driveway\"."}], "task": "refering", "answer_template": "The \"a four door sedan parked in a snowy driveway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.082453125, 0.40324009324009324, 0.462234375, 0.5653146853146853], "iscrowd": 0, "area": 10236.857249999997, "rle": {"size": [429, 640], "counts": "Y]f02Z=3N2N2N2N2M2O2N2N2YC\\Oa<h0O100000000O100000000O1000000O100000000O010000000O100000O10000000O100000000000001O00000O100000000000001O00000O2O0000001O0O1O2N1O1O10O01O1O1O1O001O100O001O1O1O1O00100O1O001O100O1O010O1O1O10O01O1O101N1000001O0000001O0000001O000000001O01O0001O00000000000000000000000000000000000000000000O100000000000000O101O000000000O2O00001O001O00001O00001O001O002N2N1O2N2N2N2N2N1O2N2N2M3N00001N10000O10001O0O1000001N10000O10001N100O100O101N100001O00010O00001O00010O00001O0001AXC8S=K4M4K^l_4"}, "label": "a four door sedan parked in a snowy driveway"}]}
{"id": 46118, "image": "COCO_train2014_000000046118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white car in the driveway\"."}], "task": "refering", "answer_template": "The \"the white car in the driveway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.082453125, 0.40324009324009324, 0.462234375, 0.5653146853146853], "iscrowd": 0, "area": 10236.857249999997, "rle": {"size": [429, 640], "counts": "Y]f02Z=3N2N2N2N2M2O2N2N2YC\\Oa<h0O100000000O100000000O1000000O100000000O010000000O100000O10000000O100000000000001O00000O100000000000001O00000O2O0000001O0O1O2N1O1O10O01O1O1O1O001O100O001O1O1O1O00100O1O001O100O1O010O1O1O10O01O1O101N1000001O0000001O0000001O000000001O01O0001O00000000000000000000000000000000000000000000O100000000000000O101O000000000O2O00001O001O00001O00001O001O002N2N1O2N2N2N2N2N1O2N2N2M3N00001N10000O10001O0O1000001N10000O10001N100O100O101N100001O00010O00001O00010O00001O0001AXC8S=K4M4K^l_4"}, "label": "the white car in the driveway"}]}
{"id": 324650, "image": "COCO_train2014_000000324650.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with red and black striped shorts\"."}], "task": "refering", "answer_template": "The \"the man with red and black striped shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 103, 104, 105, 126, 127, 149, 150, 151, 171, 172, 173, 194, 195, 196, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 311, 312], "bbox": [0.46032812500000003, 0.17661870503597124, 0.6075625, 0.9064988009592326], "iscrowd": 0, "area": 15708.23285, "rle": {"size": [417, 640], "counts": "g^h3;a<<C=D<D;D:RMRN\\JS2a5eNdIa1X6YOmHl0P7d2M3[NmJYKV5d4QKUKR5h4UKPKo4m4XKlJk4R5\\KfJg4Y5^K`Je4^5`1O1N2O^OfJ_IX5U6WKjIh4j5eKVJZ4^5TLaJj3`5XL_Jg3`5[L`Jc3`5`L_J_3`5dL_J[3a5gL^JW3b5kL^JT3a5oL^Jo2b5TM]JV2J\\Li5_1]JT2OXLc5g1]Jo16TL]5P2\\Jk1;QLX5V2]Jh1X6ZNgId1Y6_NfI`1Y6cNfI\\1Y6WNkISNLe3\\6QNmI[NFb3a6lMoI]2S6]MSJa2P6XMUJf2o5UMUJi2n5RMVJl2l5QMVJn2k5PMXJm2i5QM[Jm2f5QM\\Jn2e5PM^Jo2a5oLcJP3Y2XMYNF@R34bL1g0:BDU30jLJa0`0_OGW3NPME<d0[OKY3LWM]O7k0VOO]3G]MYO1o0TO3^3DeMSOKT1PO7`3BiMnNIW1lN<b3^OjMoNJU1hN`0d3[OjMPOL_3[2`MjMQOL^3Z2aMiMROOc2eN[Md3n0iMSO0a2fN]Ma3n0iMUO1]2hN_M^3o0hMUO5X2hNdM[3n0iMVO5U2jNfMX3n0iMWO8P2kNiMU3P1hMWO9m1mNkMR3P1hMYO:i1oNmMP3o0hMZO<e1POQNl2o0hM\\O=a1ROSNi2P1hM[O`0^1QOWNg2o0hM]Oa0Y1TOZNc2o0hM_Ob0U1VO\\N`2P1hM^Oe0Q1WO_N]2Q1gM@f0m0YOaNZ2Q1hM@g0k0ZOcNX2Q1gMAi0f0[OhNU2P1hMBi0c0]OjNR2;^N76`0]OmNh4c0lK<@POe4b0mK;X6EjI6Y6JhI3Z6LhI1Z6OhIL[63gIJ\\65fIF]6:dIC^6<dIAd69Y3HVQV3"}, "label": "the man with red and black striped shorts"}]}
{"id": 324650, "image": "COCO_train2014_000000324650.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man carrying a surfboard\"."}], "task": "refering", "answer_template": "The \"a man carrying a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 103, 104, 105, 126, 127, 149, 150, 151, 171, 172, 173, 194, 195, 196, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 311, 312], "bbox": [0.46032812500000003, 0.17661870503597124, 0.6075625, 0.9064988009592326], "iscrowd": 0, "area": 15708.23285, "rle": {"size": [417, 640], "counts": "g^h3;a<<C=D<D;D:RMRN\\JS2a5eNdIa1X6YOmHl0P7d2M3[NmJYKV5d4QKUKR5h4UKPKo4m4XKlJk4R5\\KfJg4Y5^K`Je4^5`1O1N2O^OfJ_IX5U6WKjIh4j5eKVJZ4^5TLaJj3`5XL_Jg3`5[L`Jc3`5`L_J_3`5dL_J[3a5gL^JW3b5kL^JT3a5oL^Jo2b5TM]JV2J\\Li5_1]JT2OXLc5g1]Jo16TL]5P2\\Jk1;QLX5V2]Jh1X6ZNgId1Y6_NfI`1Y6cNfI\\1Y6WNkISNLe3\\6QNmI[NFb3a6lMoI]2S6]MSJa2P6XMUJf2o5UMUJi2n5RMVJl2l5QMVJn2k5PMXJm2i5QM[Jm2f5QM\\Jn2e5PM^Jo2a5oLcJP3Y2XMYNF@R34bL1g0:BDU30jLJa0`0_OGW3NPME<d0[OKY3LWM]O7k0VOO]3G]MYO1o0TO3^3DeMSOKT1PO7`3BiMnNIW1lN<b3^OjMoNJU1hN`0d3[OjMPOL_3[2`MjMQOL^3Z2aMiMROOc2eN[Md3n0iMSO0a2fN]Ma3n0iMUO1]2hN_M^3o0hMUO5X2hNdM[3n0iMVO5U2jNfMX3n0iMWO8P2kNiMU3P1hMWO9m1mNkMR3P1hMYO:i1oNmMP3o0hMZO<e1POQNl2o0hM\\O=a1ROSNi2P1hM[O`0^1QOWNg2o0hM]Oa0Y1TOZNc2o0hM_Ob0U1VO\\N`2P1hM^Oe0Q1WO_N]2Q1gM@f0m0YOaNZ2Q1hM@g0k0ZOcNX2Q1gMAi0f0[OhNU2P1hMBi0c0]OjNR2;^N76`0]OmNh4c0lK<@POe4b0mK;X6EjI6Y6JhI3Z6LhI1Z6OhIL[63gIJ\\65fIF]6:dIC^6<dIAd69Y3HVQV3"}, "label": "a man carrying a surfboard"}]}
{"id": 13355, "image": "COCO_train2014_000000013355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"art project\"."}], "task": "refering", "answer_template": "The \"art project\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.01884375, 0.43295081967213117, 0.9921406249999999, 0.983536299765808], "iscrowd": 0, "area": 54443.632100000024, "rle": {"size": [427, 640], "counts": "iV5>l<2O001N10001N101N101O0O101O0O2O001N101N10001N101O0O2O0O2O000O2O001N101N101O0O101O0O2O001N101N10001N101O0O2O0O101O0O2O001N101O0O101N101O0O2O001N100O2O001O0010O0010O0010O00010O00010O01O01O01O010O00010O01O01O010O00010O0010O00010O00010O01O01O010O01O01O01O010O00010O00010O0010O00010O01O01O010O00010O01O01O01O010O00010O00010O0010O0010O0010O00010O00010O01O01O01O010O00010O010O00010O00010O0010O00010O01O01O01O01O01O010O01O01O010O00010O00010O0010O00010O01O01O010O01O01O01O00010O001O00010O001O00010O00001O010O00001O010O00001O01O01O001O00010O001O00010O00001O010O00001O010O00001O010O00001O01O01O001O01O01O000010O0001O00010O00001O0001O01O000010O0001O00010O00001O01O01O000010O0001O00010O00001O01O01O000010O0001O000010O0001O00010O00000010O0001O00010O00001O01O01O000010O0001O00010O00001O01O01O000010O00010O0001O01O01O01O01O00010O00010O0000010O00010O000010O00010O0001O01O01O01O01O00010O00010O000010O00010O0001O01O01O01O01O00010O00010O000010O000010O000010O00010O0001O01O01O01O01O00010O00010O000001O00010O0000001O00001O01O0001O00001O0000010O00001O0000001O01O01O0000001O00010O0000001O00001O01O0001O00001O0000010O00001O0000001O01O01O0000001O000010O000001O00001O0001O01O00001O00000010O0001O0000001O01O01O0000001O000O2O0000001O000O2O0000001O000O2O0000001O000O2O00000hN_i1"}, "label": "art project"}]}
{"id": 13355, "image": "COCO_train2014_000000013355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfboard with a blue stripe along it located behind the man\"."}], "task": "refering", "answer_template": "The \"the surfboard with a blue stripe along it located behind the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 30, 31, 32, 33, 34, 55, 136, 158, 159, 160, 182, 183, 205, 206], "bbox": [0.27471875, 0.00234192037470726, 0.996859375, 0.5670491803278689], "iscrowd": 0, "area": 7959.6156999999985, "rle": {"size": [427, 640], "counts": "niY21Z=1O1O001O1O001O1O001O1O001O1O1O001O1N101O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O0O2O1O001O1O001O1O1O001O1O001O1O001O1O1O00100O00100O010O100O00100O010O10O01O10O0N3M2M4L3M3M4M2M4L3M4O0O101N101N101N101N100O2O0O2O001N100O2O001O001O000010O01O001O00001O001O001O00001O001O001O000010O01O001O001O01O01O00SoR3<RPmLd0G8L5K5K4L2M2O00001O00001O00001O0010O0010O01O010O0010O0100O0010O01O010O0010O010O01O10O01O010O0010O010O01O010O0TOhe0"}, "label": "the surfboard with a blue stripe along it located behind the man"}]}
{"id": 13355, "image": "COCO_train2014_000000013355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white surfboard behind a man\"."}], "task": "refering", "answer_template": "The \"a white surfboard behind a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 30, 31, 32, 33, 34, 55, 136, 158, 159, 160, 182, 183, 205, 206], "bbox": [0.27471875, 0.00234192037470726, 0.996859375, 0.5670491803278689], "iscrowd": 0, "area": 7959.6156999999985, "rle": {"size": [427, 640], "counts": "niY21Z=1O1O001O1O001O1O001O1O001O1O1O001O1N101O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O0O2O1O001O1O001O1O1O001O1O001O1O001O1O1O00100O00100O010O100O00100O010O10O01O10O0N3M2M4L3M3M4M2M4L3M4O0O101N101N101N101N100O2O0O2O001N100O2O001O001O000010O01O001O00001O001O001O00001O001O001O000010O01O001O001O01O01O00SoR3<RPmLd0G8L5K5K4L2M2O00001O00001O00001O0010O0010O01O010O0010O0100O0010O01O010O0010O010O01O10O01O010O0010O010O01O010O0TOhe0"}, "label": "a white surfboard behind a man"}]}
{"id": 13355, "image": "COCO_train2014_000000013355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two biciycles in the background\"."}], "task": "refering", "answer_template": "The \"two biciycles in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 122, 123, 124, 146, 147, 148, 149, 170, 171, 172, 193, 194, 195, 196, 197, 198, 199, 220, 221, 222], "bbox": [0.31278125, 0.2835128805620609, 0.6643125, 0.6211475409836066], "iscrowd": 0, "area": 6899.375549999998, "rle": {"size": [427, 640], "counts": "lmc21[=0O1O2N100O1O100O1O1O2O1eCG`;;YDJg;8SDMl;5mC1S<c001O10O01O1O10O01O1O010OZOaDA^;?bDA_;>bDB];>dDB[;>fDAZ;?fDCY;<hDDW;<jDDU;<lDDS;<nDEQ;:oDGP;9QEGn:9SEGm:8TEIj:7VEJi:6XEJg:6[EIe:6]EJa:6aEI^:7cEI\\:7fEG[:8gEGX:9jEFU::lEEU::lEFS::mEGR:9nEIP:7PF0j9NWF9b9F_F`0[9^OgFi0S9TOPGn0m8QOTGo0l8oNVGQ1j8nNWGR1a:1O0000010O00001O0AlNgDU1V;nNjDQ1U;QOjDo0U;SOjDn0T;TOkDl0S;WOlDj0R;YOlDg0T;ZOlDe0T;\\OkDe0T;\\OkDd0U;]OjDd0U;]OhDe0X;g001O001O00001O000010O000001O0001O01O00001O01O0001O0000010O00000011N2`NmDd0U;YOoDe0R;XOQEf0Q;VOTEg0n:VOUEh0m:UOTEl0m:POVEo0d;O001O1O010O1O002N4L4L3N3L4L2N001O0010O01O001O00O101O00000O1000001O0O100000001N1000000O10001O0O10000O10001N10000O10000O2O000O10000O2O000O1000010O01O001O0010O01O1O001O010O001O001O10O01O001O1O010O001OQ]i2"}, "label": "two biciycles in the background"}]}
{"id": 226350, "image": "COCO_train2014_000000226350.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting on the motor bike\"."}], "task": "refering", "answer_template": "The \"the man sitting on the motor bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 264, 286, 287, 288], "bbox": [0.39159375, 0.33951388888888884, 0.595984375, 0.8528935185185185], "iscrowd": 0, "area": 9337.101999999999, "rle": {"size": [432, 640], "counts": "mSZ34[=1N2O1O00O01001O1O1000jBMo<;N3M2O00O1I71O001O1O00IPCOo<0UCMk<390000001O00001O00000O100O1PDLd:4[E1a:O^E4_:MaE5]:KbE8\\:HcE<Y:EgE=W:ChE`0V:AhEa0X1SOf6>PHb0T1TOl6<nGb0S1SOo6<mGc0R1QOP7>lGc0R1POR7=lGe0P1mNU7>jGi0m0jNX7=lGl0h0hN\\7<lGo0e0eN_7<kGR1c0cNa7<lGT1?aNe7:lGY1;]Nj7:kG[18\\Nl7:kG^15YNP88lGa10YNT86lGd1KXNX85lGf1HVN\\6FPK=mNj1AVNa6KjJ5TO5[Ol02QOe6MeJ2XO5]Od1f6WNcJO[O6^Ob1d6[NaJM]O6_Oa1b6_N_JK_O6A^1b6cN[JIC6C[1_6hNSJMK0GW1[6nNkI12KLR1V6VOcI2<FOn0R6\\O\\I5c0A3i0o5P1nIWN7c0m5U1lIYN;=i5[1kIYN`07f5`1iIZNe01c5d1hI]Ng0M`5f1hI^Nj0J^5h1hI^Nl0I[5h1jI_Nn0FY5j1hIaNQ1CW5l1hIaNS1BU5k1hIdNV1^OR5n1hIdNX1\\OQ5o1gIeNZ1YOP5Q2fIgN]1TOn4U2eIgN_1POo4X2aIiNb1lNn4Z2aIjNg8U1YGkNg8T1YGlNk8Q1TGPOn8m0RGTOQ9QO^F\\1a0CS9nN_F^1=EV9iN_Fa1;G`98_FHb9=XFDi9?SFAm9d0mE\\OU:o1001O0O2O0O2O001N2O001N3N2M3N2N2M3M3N2M3N5J6K6I7ATDmNS<k0QDSOT<e0c0HTQ]3"}, "label": "the man sitting on the motor bike"}]}
{"id": 226350, "image": "COCO_train2014_000000226350.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with shorts sitting on a motorcycle\"."}], "task": "refering", "answer_template": "The \"man with shorts sitting on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 264, 286, 287, 288], "bbox": [0.39159375, 0.33951388888888884, 0.595984375, 0.8528935185185185], "iscrowd": 0, "area": 9337.101999999999, "rle": {"size": [432, 640], "counts": "mSZ34[=1N2O1O00O01001O1O1000jBMo<;N3M2O00O1I71O001O1O00IPCOo<0UCMk<390000001O00001O00000O100O1PDLd:4[E1a:O^E4_:MaE5]:KbE8\\:HcE<Y:EgE=W:ChE`0V:AhEa0X1SOf6>PHb0T1TOl6<nGb0S1SOo6<mGc0R1QOP7>lGc0R1POR7=lGe0P1mNU7>jGi0m0jNX7=lGl0h0hN\\7<lGo0e0eN_7<kGR1c0cNa7<lGT1?aNe7:lGY1;]Nj7:kG[18\\Nl7:kG^15YNP88lGa10YNT86lGd1KXNX85lGf1HVN\\6FPK=mNj1AVNa6KjJ5TO5[Ol02QOe6MeJ2XO5]Od1f6WNcJO[O6^Ob1d6[NaJM]O6_Oa1b6_N_JK_O6A^1b6cN[JIC6C[1_6hNSJMK0GW1[6nNkI12KLR1V6VOcI2<FOn0R6\\O\\I5c0A3i0o5P1nIWN7c0m5U1lIYN;=i5[1kIYN`07f5`1iIZNe01c5d1hI]Ng0M`5f1hI^Nj0J^5h1hI^Nl0I[5h1jI_Nn0FY5j1hIaNQ1CW5l1hIaNS1BU5k1hIdNV1^OR5n1hIdNX1\\OQ5o1gIeNZ1YOP5Q2fIgN]1TOn4U2eIgN_1POo4X2aIiNb1lNn4Z2aIjNg8U1YGkNg8T1YGlNk8Q1TGPOn8m0RGTOQ9QO^F\\1a0CS9nN_F^1=EV9iN_Fa1;G`98_FHb9=XFDi9?SFAm9d0mE\\OU:o1001O0O2O0O2O001N2O001N3N2M3N2N2M3M3N2M3N5J6K6I7ATDmNS<k0QDSOT<e0c0HTQ]3"}, "label": "man with shorts sitting on a motorcycle"}]}
{"id": 341039, "image": "COCO_train2014_000000341039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stop sign on the right furthest from the man\"."}], "task": "refering", "answer_template": "The \"the stop sign on the right furthest from the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 107, 108, 109, 110, 111, 124, 125, 126, 127, 140, 141, 142, 143, 157, 158], "bbox": [0.7415812917594655, 0.14751562499999998, 0.9695322939866371, 0.40203124999999995], "iscrowd": 0, "area": 13233.7506, "rle": {"size": [640, 449], "counts": "SX`6>ac0j0VOk0UO:F3M2M4M3M2N3M3M2N3L4M2N3M3M2N3M3M2M4M2N3M3M2N3M3M2M4M3M2N2N1001O000001O00000000001O000001O000001O0000000000001O01O0000000001O0000000000O2N3L3N3L4M2N3L3N3L4M2N3L3N3L4M2M4M3M2M4M2M4M3M2M4M2M4M3M2M4M`c8"}, "label": "the stop sign on the right furthest from the man"}]}
{"id": 341039, "image": "COCO_train2014_000000341039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stop sign in the right side near a white truck\"."}], "task": "refering", "answer_template": "The \"stop sign in the right side near a white truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 107, 108, 109, 110, 111, 124, 125, 126, 127, 140, 141, 142, 143, 157, 158], "bbox": [0.7415812917594655, 0.14751562499999998, 0.9695322939866371, 0.40203124999999995], "iscrowd": 0, "area": 13233.7506, "rle": {"size": [640, 449], "counts": "SX`6>ac0j0VOk0UO:F3M2M4M3M2N3M3M2N3L4M2N3M3M2N3M3M2M4M2N3M3M2N3M3M2M4M3M2N2N1001O000001O00000000001O000001O000001O0000000000001O01O0000000001O0000000000O2N3L3N3L4M2N3L3N3L4M2N3L3N3L4M2M4M3M2M4M2M4M3M2M4M2M4M3M2M4M`c8"}, "label": "stop sign in the right side near a white truck"}]}
{"id": 341039, "image": "COCO_train2014_000000341039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red octagonal shaped sign with the word stop written on it in white lettering\"."}], "task": "refering", "answer_template": "The \"a red octagonal shaped sign with the word stop written on it in white lettering\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 33, 34, 35, 36, 37, 38, 49, 50, 51, 52, 53, 54, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 114, 115, 116, 117, 118, 131, 132, 133], "bbox": [0.08648106904231626, 0.0516875, 0.43563474387527845, 0.366296875], "iscrowd": 0, "area": 25905.395150000004, "rle": {"size": [640, 449], "counts": "j^h0d0Zc0n0SOm0RO;F2M3N2N3L3N2M3N2M4M2N2M3N2M4M2M3N2N2M4M2M3N2M3N3M2M3N2M3N3L3N2N2M3N3L3N2N2O1O2N1O1O1O1O2N1O1O1O1O2N00000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000O1O1N2O1O3L4M2N3L4M2N3L4M2N3L3N3M3L3N3M3L3N3M3L3N3M2M4M3M2M4M3L3N3M3L3N3M2M4M3Mc0\\Om1TNinm4"}, "label": "a red octagonal shaped sign with the word stop written on it in white lettering"}]}
{"id": 341039, "image": "COCO_train2014_000000341039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the octagon of the leftmost stop sign\"."}], "task": "refering", "answer_template": "The \"the octagon of the leftmost stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 33, 34, 35, 36, 37, 38, 49, 50, 51, 52, 53, 54, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 114, 115, 116, 117, 118, 131, 132, 133], "bbox": [0.08648106904231626, 0.0516875, 0.43563474387527845, 0.366296875], "iscrowd": 0, "area": 25905.395150000004, "rle": {"size": [640, 449], "counts": "j^h0d0Zc0n0SOm0RO;F2M3N2N3L3N2M3N2M4M2N2M3N2M4M2M3N2N2M4M2M3N2M3N3M2M3N2M3N3L3N2N2M3N3L3N2N2O1O2N1O1O1O1O2N1O1O1O1O2N00000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000O1O1N2O1O3L4M2N3L4M2N3L4M2N3L3N3M3L3N3M3L3N3M3L3N3M2M4M3M2M4M3L3N3M3L3N3M2M4M3Mc0\\Om1TNinm4"}, "label": "the octagon of the leftmost stop sign"}]}
{"id": 136240, "image": "COCO_train2014_000000136240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging the bat at the baseball\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging the bat at the baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 216, 217, 239, 262], "bbox": [0.399828125, 0.3484777517564403, 0.5551406249999999, 0.7605854800936768], "iscrowd": 0, "area": 6380.080749999997, "rle": {"size": [427, 640], "counts": "dnZ38W2LZ8>bGOn78oGLi7:THKd7:[HJ]7;bHIV7=gHDV7`0hH_OX7d0gHYOZ7j0dHTO]7o0aHPO_7S1_HkNb7X1[HhNe7[1XHdNj7^1SHbNm7a1PH_NP8d1mG[NT8h1jGWNX8l1cGUN]8o1^GRNb8Q2ZGPNg8U2RGkMo8X300001O0O101O1O1N2O1O1N2O1N2O1O2M2PNaF3`9JoFIS97VG]Ok8d0YGWOf8l0]GoNc8S1^GjNb8W1`GgN_8[1aGcN_8_1aG`N^8b1bG]N]8d1eGYN\\8h1dGWN\\8j1dGUN\\8k1b11O1O010O00iNeD?Z;@gD`0Z;_OgDa0Y;]OiDb0W;^OjDb0V;\\OlDc0T;]OmDc0T;ZOmDe0X;WOiDh0[;TOfDk0];QOeDn0];POdDo0m;O001O1O1O001O1O1O0O2O1O1O001O1O1O1O1O1O1N2O002M3N000O1000adf3"}, "label": "a baseball player swinging the bat at the baseball"}]}
{"id": 136240, "image": "COCO_train2014_000000136240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man hitting the baseball with number 8 on t - shirt\"."}], "task": "refering", "answer_template": "The \"the man hitting the baseball with number 8 on t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 216, 217, 239, 262], "bbox": [0.399828125, 0.3484777517564403, 0.5551406249999999, 0.7605854800936768], "iscrowd": 0, "area": 6380.080749999997, "rle": {"size": [427, 640], "counts": "dnZ38W2LZ8>bGOn78oGLi7:THKd7:[HJ]7;bHIV7=gHDV7`0hH_OX7d0gHYOZ7j0dHTO]7o0aHPO_7S1_HkNb7X1[HhNe7[1XHdNj7^1SHbNm7a1PH_NP8d1mG[NT8h1jGWNX8l1cGUN]8o1^GRNb8Q2ZGPNg8U2RGkMo8X300001O0O101O1O1N2O1O1N2O1N2O1O2M2PNaF3`9JoFIS97VG]Ok8d0YGWOf8l0]GoNc8S1^GjNb8W1`GgN_8[1aGcN_8_1aG`N^8b1bG]N]8d1eGYN\\8h1dGWN\\8j1dGUN\\8k1b11O1O010O00iNeD?Z;@gD`0Z;_OgDa0Y;]OiDb0W;^OjDb0V;\\OlDc0T;]OmDc0T;ZOmDe0X;WOiDh0[;TOfDk0];QOeDn0];POdDo0m;O001O1O1O001O1O1O0O2O1O1O001O1O1O1O1O1O1N2O002M3N000O1000adf3"}, "label": "the man hitting the baseball with number 8 on t - shirt"}]}
{"id": 136240, "image": "COCO_train2014_000000136240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"catcher\"."}], "task": "refering", "answer_template": "The \"catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 266, 285, 286, 287, 288, 308, 310, 311], "bbox": [0.39771875, 0.6024355971896955, 0.577234375, 0.9011241217798596], "iscrowd": 0, "area": 7257.238799999997, "rle": {"size": [427, 640], "counts": "`eZ3388[<LcC7Y<LdC6Z<>N3H8K4M3M2O2M3N2M2OH72O2N2N2N2N3M2M3N2N2N2O1N3N1O1N2O1O1N2O2M101O0O2O001O001O010O001O0010O01O001O01O2N1O3M2N2N3MXF_Mi8_2SGgMn8V2oFoMQ9o1mFUNR9j1kF[NT9d1jF_NV9k20000000O1000OWMdFe1\\9[NnF[1R9eNnF[1R9eNnF[1R9eNnF[1R9eNoFi0\\OlNe9;oFg0^OnNc9;oFe0@POc9:nFc0@SOd97nFc0@VOd94nFc0@YOc92PGb0_OZOd91oFb0_O\\Oe9OnF1^OE1:e9NnF3MNd:1\\ENg:1XENj:1WEMk:3W1O10000O01O0010N10000000O01000O100O1001O1O1O1Ofe`3"}, "label": "catcher"}]}
{"id": 136240, "image": "COCO_train2014_000000136240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher wearing a red helmet\"."}], "task": "refering", "answer_template": "The \"a baseball catcher wearing a red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 266, 285, 286, 287, 288, 308, 310, 311], "bbox": [0.39771875, 0.6024355971896955, 0.577234375, 0.9011241217798596], "iscrowd": 0, "area": 7257.238799999997, "rle": {"size": [427, 640], "counts": "`eZ3388[<LcC7Y<LdC6Z<>N3H8K4M3M2O2M3N2M2OH72O2N2N2N2N3M2M3N2N2N2O1N3N1O1N2O1O1N2O2M101O0O2O001O001O010O001O0010O01O001O01O2N1O3M2N2N3MXF_Mi8_2SGgMn8V2oFoMQ9o1mFUNR9j1kF[NT9d1jF_NV9k20000000O1000OWMdFe1\\9[NnF[1R9eNnF[1R9eNnF[1R9eNnF[1R9eNoFi0\\OlNe9;oFg0^OnNc9;oFe0@POc9:nFc0@SOd97nFc0@VOd94nFc0@YOc92PGb0_OZOd91oFb0_O\\Oe9OnF1^OE1:e9NnF3MNd:1\\ENg:1XENj:1WEMk:3W1O10000O01O0010N10000000O01000O100O1001O1O1O1Ofe`3"}, "label": "a baseball catcher wearing a red helmet"}]}
{"id": 537648, "image": "COCO_train2014_000000537648.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black office chair sits empty in a living room\"."}], "task": "refering", "answer_template": "The \"a black office chair sits empty in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 190, 191, 192, 213, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 283, 284, 285, 306, 308], "bbox": [0.25770312500000003, 0.44966666666666666, 0.45021875, 0.7932708333333333], "iscrowd": 0, "area": 9128.623399999993, "rle": {"size": [480, 640], "counts": "Pb]23m>4L4L5K2N1O001O1O1O1O0O2O1O1O1O1N101O1O1O1N3N4L4L3M4L4K7J7I7`CkMg;i2N1N2O1O1O1O1N2O1O1O1O1]EcL`9_3^FbLb9_3\\FbLYO0S:_3bFbLZO3Q:[3dFcL[O5n9Y3eFcL\\O5n9Y3eFcL\\O4P:Y3cFdL\\O4Q:Y3bFoL_9Q3aFoL_9R3`FmLb9R3^FmLd9S3[FmLf9R3ZFmLh9R3XFnLi9R3VFmLm9Q3SFnLP:Q3oEoLT:n2lEQMW:n2`EYMc:a33L3N3L3N2N3N1N3M2O1N3M2O2_NkDQOV;n0REjNn:V1TEiNk:W1UEiNk:W1UEiNj:X1VEiNi:W1WEiNi:W1WEiNi:V1WElNg:U1YE`NSO1d;_1YE`NTO0d;_1XEaNTO0d;_1XEaNUO0b;_1YEaNVOOa;`1YEaNVOOa;`1YEaNWO1];^1\\EaNXO3Y;\\1_EaNXO4X;Z1aEaNYO4V;[1aEaNZO4T;[1bEaN[O4S;Z1bEbN[O5R;Y1cEbN\\O3R;[1bEbN]O1R;]1aEbN^OOR;_1`EbN_OMR;a1_EbN@KR;b1_EcN@IR;d1^EdNR;\\1nDdNS;[1lDfNT;Y1mDgNS;Y1mDkNo:T1SEkNm:U1TEkNk:T1WEkNj:T1WEkNj:S1XEkNj:T1WEjNk:T1XEiNj::TD3U1@i:1]D?k=N101N2O0L5LhfT5"}, "label": "a black office chair sits empty in a living room"}]}
{"id": 537648, "image": "COCO_train2014_000000537648.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black office chair on wheels with armrests\"."}], "task": "refering", "answer_template": "The \"a black office chair on wheels with armrests\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 190, 191, 192, 213, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 283, 284, 285, 306, 308], "bbox": [0.25770312500000003, 0.44966666666666666, 0.45021875, 0.7932708333333333], "iscrowd": 0, "area": 9128.623399999993, "rle": {"size": [480, 640], "counts": "Pb]23m>4L4L5K2N1O001O1O1O1O0O2O1O1O1O1N101O1O1O1N3N4L4L3M4L4K7J7I7`CkMg;i2N1N2O1O1O1O1N2O1O1O1O1]EcL`9_3^FbLb9_3\\FbLYO0S:_3bFbLZO3Q:[3dFcL[O5n9Y3eFcL\\O5n9Y3eFcL\\O4P:Y3cFdL\\O4Q:Y3bFoL_9Q3aFoL_9R3`FmLb9R3^FmLd9S3[FmLf9R3ZFmLh9R3XFnLi9R3VFmLm9Q3SFnLP:Q3oEoLT:n2lEQMW:n2`EYMc:a33L3N3L3N2N3N1N3M2O1N3M2O2_NkDQOV;n0REjNn:V1TEiNk:W1UEiNk:W1UEiNj:X1VEiNi:W1WEiNi:W1WEiNi:V1WElNg:U1YE`NSO1d;_1YE`NTO0d;_1XEaNTO0d;_1XEaNUO0b;_1YEaNVOOa;`1YEaNVOOa;`1YEaNWO1];^1\\EaNXO3Y;\\1_EaNXO4X;Z1aEaNYO4V;[1aEaNZO4T;[1bEaN[O4S;Z1bEbN[O5R;Y1cEbN\\O3R;[1bEbN]O1R;]1aEbN^OOR;_1`EbN_OMR;a1_EbN@KR;b1_EcN@IR;d1^EdNR;\\1nDdNS;[1lDfNT;Y1mDgNS;Y1mDkNo:T1SEkNm:U1TEkNk:T1WEkNj:T1WEkNj:S1XEkNj:T1WEjNk:T1XEiNj::TD3U1@i:1]D?k=N101N2O0L5LhfT5"}, "label": "a black office chair on wheels with armrests"}]}
{"id": 365614, "image": "COCO_train2014_000000365614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man is jumping cycle\"."}], "task": "refering", "answer_template": "The \"the man is jumping cycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 31, 32, 33, 54, 55, 56, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 197, 214, 215, 216, 237, 238, 239, 261, 262, 284, 285, 307, 308, 330, 331, 353, 354], "bbox": [0.28090625, 0.033703125, 0.617984375, 0.6674218749999999], "iscrowd": 0, "area": 30148.549199999998, "rle": {"size": [640, 640], "counts": "Pg`33jc04L5bMEPA`0m>Bo@b0o>Al@d0Q?^Ol@f0Q?\\Ok@h0S?ZOi@k0T?WOh@m0W?TOe@Q1X?ROc@S1Z?oNb@U1\\?mNa@W1\\?kN`@Z1]?hN_@\\1_?fN]@_1a?bN[@b1c?aNX@d1e?^NW@g1f?[NV@i1h?Z1M2N3N2M2N4L4L3M4L4M3L3M4L4L3M4L4M3L3M4L4L4L;Ec0^Oa0^Oc0]O?A5K6J5L4K6J5K5K6J2hNfEgH[:X7gEgHY:W7jEgHW:W7lEgHU:X7mEfHT:X7PFeHQ:Z7QFdHP:Z7SFdHn9Z7UFeHk9Z7WFdHj9Z7YFdHh9Z7\\FcHe9\\7]FbHd9\\7_FbHb9]7`FaHa9]7bF^Hb9`7`1O1N2O10YKaHmL_7P3kHiLU7U3TIdLl6Y3^I`La6_3hIZLX6c3RJVLn5h3[JQLe5l3eJmK[5P4oJiKQ5U4XKdKg4Z4dK^K\\4`4mKYKS4d4WLUKi3i4`LPK`3m4mLiJS3T5ZMbJe2[5jMZJV2_5[NWJe1b5lNTJT1d5h6I7I6I8J61O[MSA8l>GVA9i>EYA<e>C^A=a>BaA>^>@dAa0[>]OhAc0W>[OlAe0S>ZOnAg0W>QOlAo0Z>hNiAX1]>`NdAa1c>UN`Ak1f>lM]AT2i>dMXA]2n>ZMUAf2S`00O10000O10000O10O0L5L4K4L5K5K4L5O1N101O1N101O1N101O1N101O1N101N2O001N2O010001N10001N10000O2O000100O1O2O0O1O1O100O1O100O0001O00000000000000000000O10000000000000000O10001O000000000M3L4L4L4L4L4L4Lemh4"}, "label": "the man is jumping cycle"}]}
{"id": 365614, "image": "COCO_train2014_000000365614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man doing a wheelie on his red bike\"."}], "task": "refering", "answer_template": "The \"a man doing a wheelie on his red bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 31, 32, 33, 54, 55, 56, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 197, 214, 215, 216, 237, 238, 239, 261, 262, 284, 285, 307, 308, 330, 331, 353, 354], "bbox": [0.28090625, 0.033703125, 0.617984375, 0.6674218749999999], "iscrowd": 0, "area": 30148.549199999998, "rle": {"size": [640, 640], "counts": "Pg`33jc04L5bMEPA`0m>Bo@b0o>Al@d0Q?^Ol@f0Q?\\Ok@h0S?ZOi@k0T?WOh@m0W?TOe@Q1X?ROc@S1Z?oNb@U1\\?mNa@W1\\?kN`@Z1]?hN_@\\1_?fN]@_1a?bN[@b1c?aNX@d1e?^NW@g1f?[NV@i1h?Z1M2N3N2M2N4L4L3M4L4M3L3M4L4L3M4L4M3L3M4L4L4L;Ec0^Oa0^Oc0]O?A5K6J5L4K6J5K5K6J2hNfEgH[:X7gEgHY:W7jEgHW:W7lEgHU:X7mEfHT:X7PFeHQ:Z7QFdHP:Z7SFdHn9Z7UFeHk9Z7WFdHj9Z7YFdHh9Z7\\FcHe9\\7]FbHd9\\7_FbHb9]7`FaHa9]7bF^Hb9`7`1O1N2O10YKaHmL_7P3kHiLU7U3TIdLl6Y3^I`La6_3hIZLX6c3RJVLn5h3[JQLe5l3eJmK[5P4oJiKQ5U4XKdKg4Z4dK^K\\4`4mKYKS4d4WLUKi3i4`LPK`3m4mLiJS3T5ZMbJe2[5jMZJV2_5[NWJe1b5lNTJT1d5h6I7I6I8J61O[MSA8l>GVA9i>EYA<e>C^A=a>BaA>^>@dAa0[>]OhAc0W>[OlAe0S>ZOnAg0W>QOlAo0Z>hNiAX1]>`NdAa1c>UN`Ak1f>lM]AT2i>dMXA]2n>ZMUAf2S`00O10000O10000O10O0L5L4K4L5K5K4L5O1N101O1N101O1N101O1N101O1N101N2O001N2O010001N10001N10000O2O000100O1O2O0O1O1O100O1O100O0001O00000000000000000000O10000000000000000O10001O000000000M3L4L4L4L4L4L4Lemh4"}, "label": "a man doing a wheelie on his red bike"}]}
{"id": 414772, "image": "COCO_train2014_000000414772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a set table with a white tablecloth with a turkey and salads\"."}], "task": "refering", "answer_template": "The \"a set table with a white tablecloth with a turkey and salads\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 238, 239, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 294, 295, 296, 297, 298], "bbox": [0.351875, 0.5955277777777778, 1.0, 1.0], "iscrowd": 0, "area": 31324.83080000001, "rle": {"size": [360, 640], "counts": "g[_23l::F9O1O2O0O101N100O2O001N100O2O0O2O0O2O0O2O0O2O0O101N101N101N101O001N10001O0O2O001O0O2O001NkNfF>Z9BmF7S9HTG3k8M[GMd84bGG]88kGAU8?QH\\On7c0[11000O10O1000000O2O000O101O000O2O00000O2O00000O2O00001N1000001N10000O2O00001N10001O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O`NhFQ1W9mNlFT1R9jNPGW1n8gNUGZ1j8dNXG`1e8]N^Gf1P9f0ZOQOZGIZ86jGFV89nGDR8;SHAm7>WH@i7?[H]Oe7b0_H[O`7e0eHWO[7i0hHUOV7k0oHQOP7o0UImNk6R1YIkNf6V1^IgNa6X1g1002O0O2N100O2O0O2N101N100O2O001N101N10001N101N100O2O001N101N100O2O001N100O2O001O001O00001O0010O0001O0010O01O000010O01O00100O1O1O1O100O2N1O1O100O1O1O2N100O1O1PIPMf5U3QJoLo5W3hIlLY6X3^IlLb6Y3UIkLk6i300001O0000001O000000001O000000001oNRIfMn6Z2WIaMi6_2\\I\\Md6d2aIWM`6g2fITMZ6l2kIoLU6Q3Q101O00000O101O0000001O0O10001O0001O01O0001O0001O01O0001O01O0001O01O00010O00000010O000001O01O0001O0000010O00001O01O0001O0001O01O0000010O00001O01O0001O0001O01O001O010O001O0010O01O001O0010O01O001O010O001O0010O01O001O10O01O1O0010O000000001O000O10001O000000001O0UMTHZ2l7[M_He2S8O00001N10001O00001O00001O00001O000O2O000000000000000000000000000O100000000000000000O10000[G"}, "label": "a set table with a white tablecloth with a turkey and salads"}]}
{"id": 414772, "image": "COCO_train2014_000000414772.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table with seven chairs shown around it\"."}], "task": "refering", "answer_template": "The \"the table with seven chairs shown around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 238, 239, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 294, 295, 296, 297, 298], "bbox": [0.351875, 0.5955277777777778, 1.0, 1.0], "iscrowd": 0, "area": 31324.83080000001, "rle": {"size": [360, 640], "counts": "g[_23l::F9O1O2O0O101N100O2O001N100O2O0O2O0O2O0O2O0O2O0O101N101N101N101O001N10001O0O2O001O0O2O001NkNfF>Z9BmF7S9HTG3k8M[GMd84bGG]88kGAU8?QH\\On7c0[11000O10O1000000O2O000O101O000O2O00000O2O00000O2O00001N1000001N10000O2O00001N10001O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O`NhFQ1W9mNlFT1R9jNPGW1n8gNUGZ1j8dNXG`1e8]N^Gf1P9f0ZOQOZGIZ86jGFV89nGDR8;SHAm7>WH@i7?[H]Oe7b0_H[O`7e0eHWO[7i0hHUOV7k0oHQOP7o0UImNk6R1YIkNf6V1^IgNa6X1g1002O0O2N100O2O0O2N101N100O2O001N101N10001N101N100O2O001N101N100O2O001N100O2O001O001O00001O0010O0001O0010O01O000010O01O00100O1O1O1O100O2N1O1O100O1O1O2N100O1O1PIPMf5U3QJoLo5W3hIlLY6X3^IlLb6Y3UIkLk6i300001O0000001O000000001O000000001oNRIfMn6Z2WIaMi6_2\\I\\Md6d2aIWM`6g2fITMZ6l2kIoLU6Q3Q101O00000O101O0000001O0O10001O0001O01O0001O0001O01O0001O01O0001O01O00010O00000010O000001O01O0001O0000010O00001O01O0001O0001O01O0000010O00001O01O0001O0001O01O001O010O001O0010O01O001O0010O01O001O010O001O0010O01O001O10O01O1O0010O000000001O000O10001O000000001O0UMTHZ2l7[M_He2S8O00001N10001O00001O00001O00001O000O2O000000000000000000000000000O100000000000000000O10000[G"}, "label": "the table with seven chairs shown around it"}]}
{"id": 382005, "image": "COCO_train2014_000000382005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car on the left\"."}], "task": "refering", "answer_template": "The \"a white car on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238], "bbox": [0.0, 0.5595800000000001, 0.3318, 0.7738], "iscrowd": 0, "area": 11681.338549999999, "rle": {"size": [500, 500], "counts": "h8j2j<0000000000000000000000000000000001O000000000000000000000001O1O2N1O1O1O1O2N1O1O1O2N0010O0000000000000000000001O0001O1O1O001O1N2O001N2O1O1N101O1N2O001O1N2O1L3N3M3L4M2N3L4O1O001N2O1O1O00000000001O000000000000000000001O000000000000000O10001O0000000000000000001O000000000000000000001O000000000O10000O100O2N1O1O4L5K6J5K5K1O1O101N1O1N2J^hR5"}, "label": "a white car on the left"}]}
{"id": 382005, "image": "COCO_train2014_000000382005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car with a square trunk\"."}], "task": "refering", "answer_template": "The \"a white car with a square trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238], "bbox": [0.0, 0.5595800000000001, 0.3318, 0.7738], "iscrowd": 0, "area": 11681.338549999999, "rle": {"size": [500, 500], "counts": "h8j2j<0000000000000000000000000000000001O000000000000000000000001O1O2N1O1O1O1O2N1O1O1O2N0010O0000000000000000000001O0001O1O1O001O1N2O001N2O1O1N101O1N2O001O1N2O1L3N3M3L4M2N3L4O1O001N2O1O1O00000000001O000000000000000000001O000000000000000O10001O0000000000000000001O000000000000000000001O000000000O10000O100O2N1O1O4L5K6J5K5K1O1O101N1O1N2J^hR5"}, "label": "a white car with a square trunk"}]}
{"id": 382005, "image": "COCO_train2014_000000382005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back half of a white compact car with the letter n on the side\"."}], "task": "refering", "answer_template": "The \"back half of a white compact car with the letter n on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233, 247, 248, 249, 250, 251], "bbox": [0.71404, 0.55876, 1.0, 0.7721], "iscrowd": 0, "area": 11701.106450000001, "rle": {"size": [500, 500], "counts": "Pe^55V?;L3N3L4M3UO\\O`Bf0`=CUB?k=e000000001O000O1000000000000O10001O000000001N100000001O000O10001N100O100O2O0O100O101N2O1N2O1O1N101N2O1N101N101N101O0O2O0O2O0O2O0O101N101N101O0O10O0100O100O01000000O100000000O100000000O0100000000000000000000000000000000000000O10001O00000000000000000000000000000000000O100YG"}, "label": "back half of a white compact car with the letter n on the side"}]}
{"id": 382005, "image": "COCO_train2014_000000382005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white compact car facing right\"."}], "task": "refering", "answer_template": "The \"a white compact car facing right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233, 247, 248, 249, 250, 251], "bbox": [0.71404, 0.55876, 1.0, 0.7721], "iscrowd": 0, "area": 11701.106450000001, "rle": {"size": [500, 500], "counts": "Pe^55V?;L3N3L4M3UO\\O`Bf0`=CUB?k=e000000001O000O1000000000000O10001O000000001N100000001O000O10001N100O100O2O0O100O101N2O1N2O1O1N101N2O1N101N101N101O0O2O0O2O0O2O0O101N101N101O0O10O0100O100O01000000O100000000O100000000O0100000000000000000000000000000000000000O10001O00000000000000000000000000000000000O100YG"}, "label": "a white compact car facing right"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray purse being held by a caucasian woman\"."}], "task": "refering", "answer_template": "The \"a gray purse being held by a caucasian woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 161, 162, 163, 174, 175, 176, 177, 187, 188, 189, 190, 200, 201, 202], "bbox": [0.38171581769436996, 0.46517187499999996, 0.6670509383378016, 0.66965625], "iscrowd": 0, "area": 7396.023899999997, "rle": {"size": [640, 373], "counts": "lTi21cc0=L5M3N1O2M2O2M2O2N2M10O10O100O10000O10000O100O0bNCW_Oa0^?f22N1O1O2N1O2N1O2N1O1N3N1O2N001O001O001O1O001O0O2O001O1O001O00000000000O100000O10O1O100O3M2O2M3M2N3N2M2N3M3N1N3M2O2M3M2N3N2M2N3M3N1N3M3N1N3M3M2O2M4L3N3L3M4L3N3L3MhS]2"}, "label": "a gray purse being held by a caucasian woman"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey leather bag\"."}], "task": "refering", "answer_template": "The \"grey leather bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 161, 162, 163, 174, 175, 176, 177, 187, 188, 189, 190, 200, 201, 202], "bbox": [0.38171581769436996, 0.46517187499999996, 0.6670509383378016, 0.66965625], "iscrowd": 0, "area": 7396.023899999997, "rle": {"size": [640, 373], "counts": "lTi21cc0=L5M3N1O2M2O2M2O2N2M10O10O100O10000O10000O100O0bNCW_Oa0^?f22N1O1O2N1O2N1O2N1O1N3N1O2N001O001O001O1O001O0O2O001O1O001O00000000000O100000O10O1O100O3M2O2M3M2N3N2M2N3M3N1N3M2O2M3M2N3N2M2N3M3N1N3M3N1N3M3M2O2M4L3N3L3M4L3N3L3MhS]2"}, "label": "grey leather bag"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a black shirt talking on a cellphone who looks like she ' s crying\"."}], "task": "refering", "answer_template": "The \"a girl in a black shirt talking on a cellphone who looks like she ' s crying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 27, 28, 29, 30, 40, 41, 42, 43, 53, 54, 55, 56, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 118, 119, 120, 121, 122, 123, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 157, 158, 159, 160, 161, 162, 163, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 189, 190, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 235, 236, 237, 238, 239, 248, 249, 250, 251, 252, 261, 262, 264, 274, 275, 277, 278, 287, 288, 289, 290, 291], "bbox": [0.030911528150134048, 0.079953125, 0.6569705093833781, 0.989859375], "iscrowd": 0, "area": 67108.97094999999, "rle": {"size": [640, 373], "counts": "je7:]c0<E;K4K6K5K2M4M3M2M4M2M4M3M2N3M2O2M3M2N3N2M2N3M3[Oe0K5L4K6K5J6lCcL^7c3[HjLc0eNi2d4]LVM<kNe2R4iL^M7XOZ2^3XMgM2^OV2n2bMoMLZO_2j2^MYNGTOf2g2]M`N@ROo2a2ZMkNWOnN[3Z2XM\\OfNdNm3T2VMX3JYGb1c5\\Ne4b1^KmMR5P2b6M2N2M4M2N2N2O0O1O100O100O1O100O100O1O10000000TMDTB<`=2^BOT=`0jB@i<o0UCQO`<\\1^CdN]<b1bC^NX<j1fCVNU<Q2iCoMQ<Y2mCgMm;a2QD_Mj;i2SDWMg;Q3WDoLe;X3XDhLd;^3ZDbLb;e3[D[La;k3]DUL_;R4^DnK^;X4`DhK\\;^4bDbKY;f4dD[KW;k4gDUKU;n6K5K5M3`0@`0@>B8H9GZ1fN9G:F9G9G9H4M\\BdKa;\\4RDoKP<Q4iCSLZ<m3aCULb<k3ZCWLh<Y4dBhK_=EaBl3HaLc>]42O0O2N2OO01O01O01O010O000010O01O01O01O01O0nMiK[DW4e8bK[H=oNR4b7XMfHlNGm3^7hMYHaN8h3^7SNjGZNg0d3_71_HOa72^HOa72^HOa72^HOa72^HOa72]HOc72\\HOc72\\HOd71[H0d71[H0d71[H1c71[H1b71]H2`7O_H3_7O_H3_7N`H5]7LbH6[7McH>R7ClHk0g6VOXIS1_6nN`IT1^6mNaIT1^6mNaIT1^6mNaIT1^6mN`IU1_6lN`IU1_6lN`IU1_6kN`IW1_6jN`IW1_6jN`IW1_6V6O1O1O1O1O1O2N1O2N2N2N2M3M3N2M3M4M4K5K5L4K4L5L4K5L5J6J6K5J3M4TMb_OP2``0lMc_OR2``0jMc_OU2_`0fMf_OW2]`0eMf_OY2^`0bMe_O]2Pa0N2O2M2O1N2N2O2M2O1N2N2O2M2O1N3N1N3M2O2M3N1N3M2O2M2O2J5J7J6I[d_2"}, "label": "a girl in a black shirt talking on a cellphone who looks like she ' s crying"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman crying qith a brown napkin in her hand\"."}], "task": "refering", "answer_template": "The \"a young woman crying qith a brown napkin in her hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 27, 28, 29, 30, 40, 41, 42, 43, 53, 54, 55, 56, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 118, 119, 120, 121, 122, 123, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 157, 158, 159, 160, 161, 162, 163, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 189, 190, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 235, 236, 237, 238, 239, 248, 249, 250, 251, 252, 261, 262, 264, 274, 275, 277, 278, 287, 288, 289, 290, 291], "bbox": [0.030911528150134048, 0.079953125, 0.6569705093833781, 0.989859375], "iscrowd": 0, "area": 67108.97094999999, "rle": {"size": [640, 373], "counts": "je7:]c0<E;K4K6K5K2M4M3M2M4M2M4M3M2N3M2O2M3M2N3N2M2N3M3[Oe0K5L4K6K5J6lCcL^7c3[HjLc0eNi2d4]LVM<kNe2R4iL^M7XOZ2^3XMgM2^OV2n2bMoMLZO_2j2^MYNGTOf2g2]M`N@ROo2a2ZMkNWOnN[3Z2XM\\OfNdNm3T2VMX3JYGb1c5\\Ne4b1^KmMR5P2b6M2N2M4M2N2N2O0O1O100O100O1O100O100O1O10000000TMDTB<`=2^BOT=`0jB@i<o0UCQO`<\\1^CdN]<b1bC^NX<j1fCVNU<Q2iCoMQ<Y2mCgMm;a2QD_Mj;i2SDWMg;Q3WDoLe;X3XDhLd;^3ZDbLb;e3[D[La;k3]DUL_;R4^DnK^;X4`DhK\\;^4bDbKY;f4dD[KW;k4gDUKU;n6K5K5M3`0@`0@>B8H9GZ1fN9G:F9G9G9H4M\\BdKa;\\4RDoKP<Q4iCSLZ<m3aCULb<k3ZCWLh<Y4dBhK_=EaBl3HaLc>]42O0O2N2OO01O01O01O010O000010O01O01O01O01O0nMiK[DW4e8bK[H=oNR4b7XMfHlNGm3^7hMYHaN8h3^7SNjGZNg0d3_71_HOa72^HOa72^HOa72^HOa72^HOa72]HOc72\\HOc72\\HOd71[H0d71[H0d71[H1c71[H1b71]H2`7O_H3_7O_H3_7N`H5]7LbH6[7McH>R7ClHk0g6VOXIS1_6nN`IT1^6mNaIT1^6mNaIT1^6mNaIT1^6mN`IU1_6lN`IU1_6lN`IU1_6kN`IW1_6jN`IW1_6jN`IW1_6V6O1O1O1O1O1O2N1O2N2N2N2M3M3N2M3M4M4K5K5L4K4L5L4K5L5J6J6K5J3M4TMb_OP2``0lMc_OR2``0jMc_OU2_`0fMf_OW2]`0eMf_OY2^`0bMe_O]2Pa0N2O2M2O1N2N2O2M2O1N2N2O2M2O1N3N1N3M2O2M3N1N3M2O2M2O2J5J7J6I[d_2"}, "label": "a young woman crying qith a brown napkin in her hand"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing headphones and has her arms crossed\"."}], "task": "refering", "answer_template": "The \"a woman wearing headphones and has her arms crossed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 21, 22, 23, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 58, 59, 60, 61, 62, 63, 71, 72, 73, 74, 75, 76, 77, 85, 86, 87, 88, 89, 90, 98, 99, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 203, 204, 206, 207, 216, 217, 219, 220, 229, 230, 232, 233, 242, 243, 245, 246, 256, 257, 269, 270], "bbox": [0.5211260053619303, 0.0, 1.0, 0.8916875], "iscrowd": 0, "area": 57300.704950000014, "rle": {"size": [640, 373], "counts": "n[i31mc0:F9G:F:F9G:E:G:F:F9G:F9G:F1O4M2M4M2N3M2N2N3M2N3M2N3L3N2N3M2N3M2N3L3N2K6F9WFmI[5]6`JkIS5^6hJlIj4^6QKjIb4`6XKjI\\4_6^KiI[4[6_KoIZ4U6`KSJZ4Q6_KYJZ4k5`K]JY4h5`KbJY4b5aKfJY4^5`KkJY4Y5bKPKW4T5bKUKX4o4bKZKV4k4dK]KV4g4cKcKV4a4dKgKU4^4dKlKU4W4fKQLT4S4eKWLU4o9J7J5K6J5K6J5K5K1O001O010O001O001O00WKcMPH\\2m7oMkGR2P8ZNiGe1T8HbF8Z90cFOZ97cFIY9>eFAX9e0eF[OX9k0fFTOW9R1fFoNW9X1eFgNY9a1aF_N]9i1^FVN_9S2[FmMc9[2XFdMf9b2VF_Mf9h2WFXMf9n2VFSMg9S3VFmLg9Y3UFhLg9_3VFaLi9c3SF^Lo9a3PF_LR:`3lEaLW:^3gEcLZ:\\3dEhL[:W3dEkL]:T3`EPM_:o2`ETM_:k2_EXMb:g2\\E\\Mc:c2[E_Mf:`2YEbMh:l50O101N10000O2O000O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O101N100O2N100O2O0O100O2N100O2O0OO2M2M3N3M2N3VHUHX3V9XKgGQ4c8_KgG]4c8SKgGh4d8gJgGU5c8[JgGa5`;K5L4L4L4K5L4L:E?B=C=C=BR1oNZ1fN^@"}, "label": "a woman wearing headphones and has her arms crossed"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black woman with headphones\"."}], "task": "refering", "answer_template": "The \"a black woman with headphones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 21, 22, 23, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 58, 59, 60, 61, 62, 63, 71, 72, 73, 74, 75, 76, 77, 85, 86, 87, 88, 89, 90, 98, 99, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 203, 204, 206, 207, 216, 217, 219, 220, 229, 230, 232, 233, 242, 243, 245, 246, 256, 257, 269, 270], "bbox": [0.5211260053619303, 0.0, 1.0, 0.8916875], "iscrowd": 0, "area": 57300.704950000014, "rle": {"size": [640, 373], "counts": "n[i31mc0:F9G:F:F9G:E:G:F:F9G:F9G:F1O4M2M4M2N3M2N2N3M2N3M2N3L3N2N3M2N3M2N3L3N2K6F9WFmI[5]6`JkIS5^6hJlIj4^6QKjIb4`6XKjI\\4_6^KiI[4[6_KoIZ4U6`KSJZ4Q6_KYJZ4k5`K]JY4h5`KbJY4b5aKfJY4^5`KkJY4Y5bKPKW4T5bKUKX4o4bKZKV4k4dK]KV4g4cKcKV4a4dKgKU4^4dKlKU4W4fKQLT4S4eKWLU4o9J7J5K6J5K6J5K5K1O001O010O001O001O00WKcMPH\\2m7oMkGR2P8ZNiGe1T8HbF8Z90cFOZ97cFIY9>eFAX9e0eF[OX9k0fFTOW9R1fFoNW9X1eFgNY9a1aF_N]9i1^FVN_9S2[FmMc9[2XFdMf9b2VF_Mf9h2WFXMf9n2VFSMg9S3VFmLg9Y3UFhLg9_3VFaLi9c3SF^Lo9a3PF_LR:`3lEaLW:^3gEcLZ:\\3dEhL[:W3dEkL]:T3`EPM_:o2`ETM_:k2_EXMb:g2\\E\\Mc:c2[E_Mf:`2YEbMh:l50O101N10000O2O000O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O101N100O2N100O2O0O100O2N100O2O0OO2M2M3N3M2N3VHUHX3V9XKgGQ4c8_KgG]4c8SKgGh4d8gJgGU5c8[JgGa5`;K5L4L4L4K5L4L:E?B=C=C=BR1oNZ1fN^@"}, "label": "a black woman with headphones"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy in a grey sweater behind the crying girl\"."}], "task": "refering", "answer_template": "The \"the guy in a grey sweater behind the crying girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 18, 19, 20, 21, 30, 31, 32, 33, 43, 44, 45, 46, 56, 57, 58, 69, 70, 71, 83, 84, 96, 97], "bbox": [0.3411260053619303, 0.0019375, 0.6256300268096515, 0.33890625], "iscrowd": 0, "area": 13233.635949999998, "rle": {"size": [640, 373], "counts": "n__22lc03M3L5L3M3L5L3M4K4M3M4mNR1K5K6K4K5L5`N`MXAe2d>`MVAf2h>ZMVAi2i>WMVAk2i>UMUAn2j>RMUAP3j>PMUAR3j>oLSAT3l>lLSAV3l>jLSAX3l>hLRA[3m>m0O1O1O1O1O1O1O5K4L5K4L5K4L5K5K4L01O01O00001O00001O0000001O00001O00001O00M3L4ZMdDdL`;Z3d2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1O1O1N2O1O1O1O3L4M2N3M3L4M2N3M3M3L3K6D<D<D__g2"}, "label": "the guy in a grey sweater behind the crying girl"}]}
{"id": 242745, "image": "COCO_train2014_000000242745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in a grey sweatshirt standing behind two girls\"."}], "task": "refering", "answer_template": "The \"person in a grey sweatshirt standing behind two girls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 18, 19, 20, 21, 30, 31, 32, 33, 43, 44, 45, 46, 56, 57, 58, 69, 70, 71, 83, 84, 96, 97], "bbox": [0.3411260053619303, 0.0019375, 0.6256300268096515, 0.33890625], "iscrowd": 0, "area": 13233.635949999998, "rle": {"size": [640, 373], "counts": "n__22lc03M3L5L3M3L5L3M4K4M3M4mNR1K5K6K4K5L5`N`MXAe2d>`MVAf2h>ZMVAi2i>WMVAk2i>UMUAn2j>RMUAP3j>PMUAR3j>oLSAT3l>lLSAV3l>jLSAX3l>hLRA[3m>m0O1O1O1O1O1O1O5K4L5K4L5K4L5K5K4L01O01O00001O00001O0000001O00001O00001O00M3L4ZMdDdL`;Z3d2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1O1O1N2O1O1O1O3L4M2N3M3L4M2N3M3M3L3K6D<D<D__g2"}, "label": "person in a grey sweatshirt standing behind two girls"}]}
{"id": 291898, "image": "COCO_train2014_000000291898.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man\"."}], "task": "refering", "answer_template": "The \"a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 385, 386, 387, 388, 389, 390], "bbox": [0.480859375, 0.14843096234309625, 1.0, 0.9984309623430963], "iscrowd": 0, "area": 89020.91254999998, "rle": {"size": [478, 640], "counts": "lV`47f>2L3N2XEMc66WI3d6NYI9b6I[I=`6D]I`0b6AZIc0d6_OYIe0d6\\OZIg0d6[OZIh0d6YOZIi0d6XOZIl0c6UO\\Im0c6SO\\IP1a6PO_IS1_6mNaIT1^6lNaIW1\\6jNdIW1[6iNdIY1Z6hNfIY1Y6fNhI[1W6eNhI^1U6cNkI^1T6bNlI_1R6bNnI_1Q6aNoI_1Q6`NoIb1o5_NQJb1n5^NRJc1l5^NTJb1l5^NTJc1k5aNQJ`1n5aNQJ`1m5aNSJ`1l5`NTJ`1l5`NTJa1j5_NWJb1h5^NXJc1g5]NYJd1f5[N[Je1d5\\N\\Je1c5[N]Je1c5ZN^Jf1b5ZN_Jf1_5[NaJe1_5ZNbJf1^5ZNbJg1\\5YNeJg1[5YNeJg1[5YNeJh1Z5XNgJg1Y5YNgJg1X5ZNiJf1V5ZNjJf1V5ZNkJe1U5[NkJf1c1YMmNQ1Ae1_1bMjNj0Gd1\\1bNPNJe0c1X1eNTNHd0c1U1hNWNEe0c1o0kN]NBd0c1l0nN_N_Og0b1f0ROdN\\Of0b1nN[MAk1k0XOg0a1hNaMAj1Q1TOf0b1bNfM@j1W1oNg0a1[NnMAf1^1kNg0`1SNVNAf1e1dNg0`1lM^N@d1n1^Nf0d1bMaNE^1U2^Nd0e3TMmMY2`Nc0d3QMkM]2cNb0^6_OcIb0[6]OgIc0W6^OjIb0U6]OmId0P6\\ORJd0l5]OTJe0j5ZOXJh0d5YO\\Jj0`5SOeJo0V5nNPKS1m4hNYK[1c4`NdKb1W4ZNPLh1l3SNZLP2b3kMeLV2W3gMnL[2o2^MZMc2b2hLUNZ3h1`L_Nb3a1[LaNg3_1VLaNk3a1QL`NP4W60O100O1O1O1N2O1N2O1O1N2O0O2O0O1O1N2N2O2M2N3N1N3M3N2M3M3N2M3M3N3L3M3N2003M3M3N2M3M3M3M3M3M2N100O1O1O1O2N1O100O1O1O2N100O1O4M3L3M3N0O100O1O100O1O100O100O100O1O01O001O01O01O0010O7J5J2N01O00001O000010O000001O0000001O00001O001O001O001O001O1O001O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O001O1O2N3M4L3M3M2N2N1O2N2O0O2N2N2N1O2N2N1O2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1OPG"}, "label": "a man"}]}
{"id": 291898, "image": "COCO_train2014_000000291898.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing over a pizza box wearing a grey t - shirt and a dark colored watch\"."}], "task": "refering", "answer_template": "The \"a man standing over a pizza box wearing a grey t - shirt and a dark colored watch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 385, 386, 387, 388, 389, 390], "bbox": [0.480859375, 0.14843096234309625, 1.0, 0.9984309623430963], "iscrowd": 0, "area": 89020.91254999998, "rle": {"size": [478, 640], "counts": "lV`47f>2L3N2XEMc66WI3d6NYI9b6I[I=`6D]I`0b6AZIc0d6_OYIe0d6\\OZIg0d6[OZIh0d6YOZIi0d6XOZIl0c6UO\\Im0c6SO\\IP1a6PO_IS1_6mNaIT1^6lNaIW1\\6jNdIW1[6iNdIY1Z6hNfIY1Y6fNhI[1W6eNhI^1U6cNkI^1T6bNlI_1R6bNnI_1Q6aNoI_1Q6`NoIb1o5_NQJb1n5^NRJc1l5^NTJb1l5^NTJc1k5aNQJ`1n5aNQJ`1m5aNSJ`1l5`NTJ`1l5`NTJa1j5_NWJb1h5^NXJc1g5]NYJd1f5[N[Je1d5\\N\\Je1c5[N]Je1c5ZN^Jf1b5ZN_Jf1_5[NaJe1_5ZNbJf1^5ZNbJg1\\5YNeJg1[5YNeJg1[5YNeJh1Z5XNgJg1Y5YNgJg1X5ZNiJf1V5ZNjJf1V5ZNkJe1U5[NkJf1c1YMmNQ1Ae1_1bMjNj0Gd1\\1bNPNJe0c1X1eNTNHd0c1U1hNWNEe0c1o0kN]NBd0c1l0nN_N_Og0b1f0ROdN\\Of0b1nN[MAk1k0XOg0a1hNaMAj1Q1TOf0b1bNfM@j1W1oNg0a1[NnMAf1^1kNg0`1SNVNAf1e1dNg0`1lM^N@d1n1^Nf0d1bMaNE^1U2^Nd0e3TMmMY2`Nc0d3QMkM]2cNb0^6_OcIb0[6]OgIc0W6^OjIb0U6]OmId0P6\\ORJd0l5]OTJe0j5ZOXJh0d5YO\\Jj0`5SOeJo0V5nNPKS1m4hNYK[1c4`NdKb1W4ZNPLh1l3SNZLP2b3kMeLV2W3gMnL[2o2^MZMc2b2hLUNZ3h1`L_Nb3a1[LaNg3_1VLaNk3a1QL`NP4W60O100O1O1O1N2O1N2O1O1N2O0O2O0O1O1N2N2O2M2N3N1N3M3N2M3M3N2M3M3N3L3M3N2003M3M3N2M3M3M3M3M3M2N100O1O1O1O2N1O100O1O1O2N100O1O4M3L3M3N0O100O1O100O1O100O100O100O1O01O001O01O01O0010O7J5J2N01O00001O000010O000001O0000001O00001O001O001O001O001O1O001O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O001O1O2N3M4L3M3M2N2N1O2N2O0O2N2N2N1O2N2N1O2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1OPG"}, "label": "a man standing over a pizza box wearing a grey t - shirt and a dark colored watch"}]}
{"id": 578619, "image": "COCO_train2014_000000578619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white stuffed animal eating a piece of cake\"."}], "task": "refering", "answer_template": "The \"a white stuffed animal eating a piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 166, 167, 168, 169, 170, 173, 174, 188, 189, 190, 191, 192, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258, 279], "bbox": [0.05065625, 0.15325, 0.5961093749999999, 0.7246666666666666], "iscrowd": 0, "area": 36577.960649999986, "rle": {"size": [480, 640], "counts": "kd?4k>3M3M3M3M4M0O2N1O1O2N1O1O2N1O1O2N101N1O1O2N1O00O1O1O1O1N2O20O000100O00100O010O1O10O01000O10O1000O10000O1PGSOm3m0nKXOR4h0iK]OV4d0dKB\\4=`KH`48ZKNf42SK5l4LmJ;S5EfJb0Z5^OdJd0[5]OeJc0[5]OdJd0\\5\\OcJe0]5[ObJf0]5ZOcJg0]5YObJg0_5YO`Jh0_5YO`Jg0a5YO^Jh0b5XO]Jh0d5XO\\Jh0c5YO[Jh0f5XOYJi0g5WOXJi0h5XOVJj0j5UOVJk0k5UOSJm0m5SOQJn0o5SOnIP1R6POjIS1W6lNeIX1\\6gN`I^1`6aN[Id1e6\\NXIh1h6WNYIh1h6WNXIj1h6UNYIj1h6UNYIk1g6TNZIk1f6UN[Ij1f6UN[Ik1e6TN\\Ik1e6TN[Il1f6RN\\Im1d6SN]Il1d6SN]Il1d6SN]Im1c6RN^Im1c6RN^In1a6RN_In1a6RN`In1_6RNbIn1]6RNdIm1\\6SNeIm1[6RNfIn1Y6RNhIm1X6WNeIi1Z6\\NbId1]6bN]I^1c6gNYIY1f6kNWIU1i6mNUIR1k6QORIP1m6SOPIm0P7VOmHk0R7YOjHh0U7[OiHd0X7^OeHc0Z7@cHa0]7A`H?`7C^H>a7E\\H<d7FZH9f7IXH8g7KVH6h7MVH3j7OTH2k70TH0k72SHNm74QHLn78oGIP89nGGR8;mGDS8>kGCS8`0kG@U8b0iG^OW8e0fG\\OY8f0dG[O\\8c3100O1O100O100O1O100O101N2N2O1N101N2N2O1N2O0O2N2O1N2O0O2N2O0O2O0O1O1O1O10O10001O0O100010O1O1O1O1O1O100O1O1O1O1O1O1O100O10000O100O10000O100O10001N101O0O2O0O2O001M2N3M2N3M2N3M3M3M2N3M3M2N200O001O1O1O1O100O1O1M2N3M3M3M3N2M3M3N11000001O0O1000001O000O101O2N1O1N3N1O1O2N1O1O1N2O1O1O1O001O1OA_D`M`;^2aDdM^;Y2cDiM];j1dDiM1?Z;e1QE^Nn:^1UEdNk:Y1UEiNk:V1UElNj:R1WEPOi:m0WEVOh:h0YEZOg:d0XE^Oi:?XECh::YEHg:5YENg:0YE1h:MWE6i:HYE8h:FXE;i:BXE?h:_OZE`0h:]OYEd0h:ZOYEf0h:WOYEi0h:PO^EQ1Y<O0100O00100O010O00100O01O010O00001O001O001O1O1O001O1O001O1O10O01M3L5L4L3M4L3M5K7I6JPhh3"}, "label": "a white stuffed animal eating a piece of cake"}]}
{"id": 578619, "image": "COCO_train2014_000000578619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white stuffed dog sitting next to a cupcake\"."}], "task": "refering", "answer_template": "The \"a white stuffed dog sitting next to a cupcake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 166, 167, 168, 169, 170, 173, 174, 188, 189, 190, 191, 192, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258, 279], "bbox": [0.05065625, 0.15325, 0.5961093749999999, 0.7246666666666666], "iscrowd": 0, "area": 36577.960649999986, "rle": {"size": [480, 640], "counts": "kd?4k>3M3M3M3M4M0O2N1O1O2N1O1O2N1O1O2N101N1O1O2N1O00O1O1O1O1N2O20O000100O00100O010O1O10O01000O10O1000O10000O1PGSOm3m0nKXOR4h0iK]OV4d0dKB\\4=`KH`48ZKNf42SK5l4LmJ;S5EfJb0Z5^OdJd0[5]OeJc0[5]OdJd0\\5\\OcJe0]5[ObJf0]5ZOcJg0]5YObJg0_5YO`Jh0_5YO`Jg0a5YO^Jh0b5XO]Jh0d5XO\\Jh0c5YO[Jh0f5XOYJi0g5WOXJi0h5XOVJj0j5UOVJk0k5UOSJm0m5SOQJn0o5SOnIP1R6POjIS1W6lNeIX1\\6gN`I^1`6aN[Id1e6\\NXIh1h6WNYIh1h6WNXIj1h6UNYIj1h6UNYIk1g6TNZIk1f6UN[Ij1f6UN[Ik1e6TN\\Ik1e6TN[Il1f6RN\\Im1d6SN]Il1d6SN]Il1d6SN]Im1c6RN^Im1c6RN^In1a6RN_In1a6RN`In1_6RNbIn1]6RNdIm1\\6SNeIm1[6RNfIn1Y6RNhIm1X6WNeIi1Z6\\NbId1]6bN]I^1c6gNYIY1f6kNWIU1i6mNUIR1k6QORIP1m6SOPIm0P7VOmHk0R7YOjHh0U7[OiHd0X7^OeHc0Z7@cHa0]7A`H?`7C^H>a7E\\H<d7FZH9f7IXH8g7KVH6h7MVH3j7OTH2k70TH0k72SHNm74QHLn78oGIP89nGGR8;mGDS8>kGCS8`0kG@U8b0iG^OW8e0fG\\OY8f0dG[O\\8c3100O1O100O100O1O100O101N2N2O1N101N2N2O1N2O0O2N2O1N2O0O2N2O0O2O0O1O1O1O10O10001O0O100010O1O1O1O1O1O100O1O1O1O1O1O1O100O10000O100O10000O100O10001N101O0O2O0O2O001M2N3M2N3M2N3M3M3M2N3M3M2N200O001O1O1O1O100O1O1M2N3M3M3M3N2M3M3N11000001O0O1000001O000O101O2N1O1N3N1O1O2N1O1O1N2O1O1O1O001O1OA_D`M`;^2aDdM^;Y2cDiM];j1dDiM1?Z;e1QE^Nn:^1UEdNk:Y1UEiNk:V1UElNj:R1WEPOi:m0WEVOh:h0YEZOg:d0XE^Oi:?XECh::YEHg:5YENg:0YE1h:MWE6i:HYE8h:FXE;i:BXE?h:_OZE`0h:]OYEd0h:ZOYEf0h:WOYEi0h:PO^EQ1Y<O0100O00100O010O00100O01O010O00001O001O001O1O1O001O1O001O1O10O01M3L5L4L3M4L3M5K7I6JPhh3"}, "label": "a white stuffed dog sitting next to a cupcake"}]}
{"id": 578619, "image": "COCO_train2014_000000578619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown stuffed dog toy\"."}], "task": "refering", "answer_template": "The \"brown stuffed dog toy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.469640625, 0.42756249999999996, 1.0, 0.9987916666666666], "iscrowd": 0, "area": 56141.82430000001, "rle": {"size": [480, 640], "counts": "d^]4e0Q>;L4L4K5L5J5L4N2O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1O1O1O1O1O1O001O1O1O001O1O001O1O1O010O1O001O1O1O001000000O1000000iDnL^:S3]ESMb:m2XEZMg:f2TE`Mk:[301N2OO1O1O001O1O1O1O001O1O1O1O2N1O1O2N1O1O1O2N1O1O2NfMbMSH]2k7hMSHW2k7nMTHP2k7TNSHk1k7ZNSHe1k7`NTH^1k7fNSHY1k7lNSHS1k7ROTHl0j7YOTHf0k7_ORH`0l7FRH8l7NQH1n70SHOk73WHKf78[HGb7<_HC^7?dH@Y7c0iH[OU7g0lHXOT7h0kHYOU7g0kHYOT7h0lHXOT7h0lHXOT7h0kHYOU7g0kHYOU7g0kHYOT7h0lHXOT7h0lHXOT7h0kHYOU7g0kHYOU7h0jHXOU7i0kHWOU7j0iHWOW7j0hHVOX7j0hHVOX7k0gHUOX7l0gHUOY7l0fHTOZ7l0fHTOZ7m0eHSOZ7n0eHSO[7n0dHRO\\7n0dHRO\\7o0cHQO]7P1aHQO^7P1bHPO]7R1bHnN]7S1cHmN\\7U1bHlN]7U1cHkN\\7W1cHiN\\7Y1cHgN\\7\\1aHeN^7^1`HbN_7a1_H_N_7e1_H[NQ7R5L4L4L400O100O1O100O100O100O1O100O100O1001O0000000000001O00000000001O0000000000001O001O1O001O1O001O1O002N3M2N2N?A001O001O001O1O001O001O001O001O1O001O001O00O1002N2N1O2N2N1O2N2N2N1O2N2N1O2N4L6J6J001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O2N4L4L4L3M4L2N001O001O001O001O00001O1ZMmD]1Y;XNQEc1V;QNTE[1\\OcNf;HWE]1UOlNm<m0TCTOQ=e0PC\\OU==lBDY=5hBL]=MdB4a=E`B<e=]O\\Bd0o=1O100O1O1O1O1O1O1O1O1O1O1O100O1OC=@hB"}, "label": "brown stuffed dog toy"}]}
{"id": 578619, "image": "COCO_train2014_000000578619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stuffed brown puppy on handmade blue , white and yellow washcloth , with a chocolate cupcake\"."}], "task": "refering", "answer_template": "The \"stuffed brown puppy on handmade blue , white and yellow washcloth , with a chocolate cupcake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.469640625, 0.42756249999999996, 1.0, 0.9987916666666666], "iscrowd": 0, "area": 56141.82430000001, "rle": {"size": [480, 640], "counts": "d^]4e0Q>;L4L4K5L5J5L4N2O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1O1O1O1O1O1O001O1O1O001O1O001O1O1O010O1O001O1O1O001000000O1000000iDnL^:S3]ESMb:m2XEZMg:f2TE`Mk:[301N2OO1O1O001O1O1O1O001O1O1O1O2N1O1O2N1O1O1O2N1O1O2NfMbMSH]2k7hMSHW2k7nMTHP2k7TNSHk1k7ZNSHe1k7`NTH^1k7fNSHY1k7lNSHS1k7ROTHl0j7YOTHf0k7_ORH`0l7FRH8l7NQH1n70SHOk73WHKf78[HGb7<_HC^7?dH@Y7c0iH[OU7g0lHXOT7h0kHYOU7g0kHYOT7h0lHXOT7h0lHXOT7h0kHYOU7g0kHYOU7g0kHYOT7h0lHXOT7h0lHXOT7h0kHYOU7g0kHYOU7h0jHXOU7i0kHWOU7j0iHWOW7j0hHVOX7j0hHVOX7k0gHUOX7l0gHUOY7l0fHTOZ7l0fHTOZ7m0eHSOZ7n0eHSO[7n0dHRO\\7n0dHRO\\7o0cHQO]7P1aHQO^7P1bHPO]7R1bHnN]7S1cHmN\\7U1bHlN]7U1cHkN\\7W1cHiN\\7Y1cHgN\\7\\1aHeN^7^1`HbN_7a1_H_N_7e1_H[NQ7R5L4L4L400O100O1O100O100O100O1O100O100O1001O0000000000001O00000000001O0000000000001O001O1O001O1O001O1O002N3M2N2N?A001O001O001O1O001O001O001O001O1O001O001O00O1002N2N1O2N2N1O2N2N2N1O2N2N1O2N4L6J6J001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O2N4L4L4L3M4L2N001O001O001O001O00001O1ZMmD]1Y;XNQEc1V;QNTE[1\\OcNf;HWE]1UOlNm<m0TCTOQ=e0PC\\OU==lBDY=5hBL]=MdB4a=E`B<e=]O\\Bd0o=1O100O1O1O1O1O1O1O1O1O1O1O100O1OC=@hB"}, "label": "stuffed brown puppy on handmade blue , white and yellow washcloth , with a chocolate cupcake"}]}
{"id": 70718, "image": "COCO_train2014_000000070718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the brown pants and shirt\"."}], "task": "refering", "answer_template": "The \"the man in the brown pants and shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 40, 41, 42, 43, 62, 63, 64, 65, 66, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 338], "bbox": [0.66271875, 0.0073364485981308415, 0.952546875, 0.9824999999999999], "iscrowd": 0, "area": 50539.60735, "rle": {"size": [428, 640], "counts": "Y]a5a0b<<D;]Oc0J6J5K5K6L4L3M5K6J6J5Ke0[O5K3F:J6J6J5M4M2M4L4M2M4N1N2O1N6K4K5Am0^Oc0^Oi0VO>gKSHY2c9L5L00O0O2N1O2O1N1O2N1O2N2M2O2M2O2M2O2M3N1N3M2N3M3K4L5K4L5O01O1O1O1O1O1O100O001O1O1O1O100O1O1O100O1O1O1N2O1N2O1D=M2M3N2N2M3N200O1000000O2O000O101O0000001O0000001O0000001O000000001O0000001O0000000000000000000001O000000000fFfMb7Z2ZHlMd7T2XHSNe7n1WHWNg7i1SH_Nj7b1QHdNn7\\1lGkNS8X3O1O1O1O1O1O1M4M3L3N2M3N2kNiJQJ]5`5UKXJo4d5XKTJm4i5_1K5L4K7I7J7H7J6I7J7H8H8H8H8B>bN^1C\\[<"}, "label": "the man in the brown pants and shirt"}]}
{"id": 70718, "image": "COCO_train2014_000000070718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a plaid shirt\"."}], "task": "refering", "answer_template": "The \"a man in a plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 40, 41, 42, 43, 62, 63, 64, 65, 66, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 338], "bbox": [0.66271875, 0.0073364485981308415, 0.952546875, 0.9824999999999999], "iscrowd": 0, "area": 50539.60735, "rle": {"size": [428, 640], "counts": "Y]a5a0b<<D;]Oc0J6J5K5K6L4L3M5K6J6J5Ke0[O5K3F:J6J6J5M4M2M4L4M2M4N1N2O1N6K4K5Am0^Oc0^Oi0VO>gKSHY2c9L5L00O0O2N1O2O1N1O2N1O2N2M2O2M2O2M2O2M3N1N3M2N3M3K4L5K4L5O01O1O1O1O1O1O100O001O1O1O1O100O1O1O100O1O1O1N2O1N2O1D=M2M3N2N2M3N200O1000000O2O000O101O0000001O0000001O0000001O000000001O0000001O0000000000000000000001O000000000fFfMb7Z2ZHlMd7T2XHSNe7n1WHWNg7i1SH_Nj7b1QHdNn7\\1lGkNS8X3O1O1O1O1O1O1M4M3L3N2M3N2kNiJQJ]5`5UKXJo4d5XKTJm4i5_1K5L4K7I7J7H7J6I7J7H8H8H8H8B>bN^1C\\[<"}, "label": "a man in a plaid shirt"}]}
{"id": 70718, "image": "COCO_train2014_000000070718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a pink shirt holding a blue cup\"."}], "task": "refering", "answer_template": "The \"a man in a pink shirt holding a blue cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 128, 129, 130, 152, 174, 175, 197, 198, 220], "bbox": [0.42904687499999994, 0.005186915887850468, 0.6930624999999999, 0.6597429906542056], "iscrowd": 0, "area": 18631.082200000004, "rle": {"size": [428, 640], "counts": "\\Rc34S=5K4L5L4K5L3M4K5L4L3M4L4K5L3M4L4L4K4M4L4N20O0100O10O0100O10O0100O1O010O100O010O100O010O100O010O1O10O0100O10O0100O10O0100O10O01O100O011N101N1L4J7I6M4O001O00001O001O00001O0010O0001O001O00001QKnKb0R4ZORLf0o3UOVLj0j3ROZLn0f3nN_LQ1b3jNbLV1^3fNgLY1Z3bNjL^1V3^NoLa1Q3[NSMe1n2VNVMj1j2RN[Mm1f2nM^MR2b2jMcMU2_2eMeM[2]2_MhM`2Z2ZMjMf2X2TMjMn2X2lLhMX3[2aLeMc3]2WLcMm3_2mKaMW4\\NjJ]3i0WNa4ZNiJ`3`0VNk4WNhJe37TNU5TNhJi3MSN_5RNgJl3DRNi5oMgJW6]5fIgJU6]5iIfJR6^5kIfJP6^6K5K5L4K5K5K5K5K5K5K5K4L5nNlFfMY9Y2PGZMU9e2UGmLP9R3f0O1O1N2O1O1O1O1O1O1O1O7I:F9G8H2N3K4L5K4M3L5K4L5L3L4L5K4Mdia2"}, "label": "a man in a pink shirt holding a blue cup"}]}
{"id": 70718, "image": "COCO_train2014_000000070718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing pink shirt\"."}], "task": "refering", "answer_template": "The \"man wearing pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 128, 129, 130, 152, 174, 175, 197, 198, 220], "bbox": [0.42904687499999994, 0.005186915887850468, 0.6930624999999999, 0.6597429906542056], "iscrowd": 0, "area": 18631.082200000004, "rle": {"size": [428, 640], "counts": "\\Rc34S=5K4L5L4K5L3M4K5L4L3M4L4K5L3M4L4L4K4M4L4N20O0100O10O0100O10O0100O1O010O100O010O100O010O100O010O1O10O0100O10O0100O10O0100O10O01O100O011N101N1L4J7I6M4O001O00001O001O00001O0010O0001O001O00001QKnKb0R4ZORLf0o3UOVLj0j3ROZLn0f3nN_LQ1b3jNbLV1^3fNgLY1Z3bNjL^1V3^NoLa1Q3[NSMe1n2VNVMj1j2RN[Mm1f2nM^MR2b2jMcMU2_2eMeM[2]2_MhM`2Z2ZMjMf2X2TMjMn2X2lLhMX3[2aLeMc3]2WLcMm3_2mKaMW4\\NjJ]3i0WNa4ZNiJ`3`0VNk4WNhJe37TNU5TNhJi3MSN_5RNgJl3DRNi5oMgJW6]5fIgJU6]5iIfJR6^5kIfJP6^6K5K5L4K5K5K5K5K5K5K5K4L5nNlFfMY9Y2PGZMU9e2UGmLP9R3f0O1O1N2O1O1O1O1O1O1O1O7I:F9G8H2N3K4L5K4M3L5K4L5L3L4L5K4Mdia2"}, "label": "man wearing pink shirt"}]}
{"id": 70718, "image": "COCO_train2014_000000070718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy seeing and standing\"."}], "task": "refering", "answer_template": "The \"a boy seeing and standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 301, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.09896875000000001, 0.2939719626168224, 0.715234375, 0.9845560747663551], "iscrowd": 0, "area": 52339.41359999999, "rle": {"size": [428, 640], "counts": "iTk0<\\<d0001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00000000000000000000000000000000O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2O1N2N2O1N2N2O1N2O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O100O10000O100O10000O100O10000O100O10000O100O10000O1fLbLfL^3V3iLfLX3U3PMhLP3T3WMhLj2S3^MjLb2T3dMgL]2W3iMeLW2Y3TN^Ll1`3]NXLd1f3aNWL_1g3eNVL\\1h3hNVLX1h3lNULU1i3POTLP1j3TOSLm0k3WOSLi0l3ZOQLg0m3^OPLb0o3AnK`0Q4DkK=S4HjK8U4KhK6W4NeK3Y42cKO\\44aKM^47^KJb49[KGd4a4O100O1O100O1O100O100O100O10000O100O10000O10000O1000000O10000O10000O100000000O10000000000O10000000000000000000O101O0000000000000000001O0000000000000000001O000000000000000000001O001O1O001O1O1ZKaK=`4@cK?^4^OeKa0\\4[OiKc0X4ZOkKe0V4WOnKh0R4VORLh0o3UOTLj0m3ROWLm0j3POYLo0h3mN]LQ1d3lN_LS1b3iNbLV1_3fNeLY1\\3cNhL\\1Y3`NkL_1W3\\NmLc1U3XNoLg1S3TNQMk1Q3PNSMo1o2lMUMS2m2dM[M[2g2[MbMd2a2RMgMm2\\2hLmMW3V2^LSNa3m5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O2N3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3Mee[2"}, "label": "a boy seeing and standing"}]}
{"id": 70718, "image": "COCO_train2014_000000070718.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in orange shirt\"."}], "task": "refering", "answer_template": "The \"boy in orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 301, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.09896875000000001, 0.2939719626168224, 0.715234375, 0.9845560747663551], "iscrowd": 0, "area": 52339.41359999999, "rle": {"size": [428, 640], "counts": "iTk0<\\<d0001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00000000000000000000000000000000O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2O1N2N2O1N2N2O1N1O2O1N2N2O1N2O1N2N2O1N2N2O1N2O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O100O10000O100O10000O100O10000O100O10000O100O10000O1fLbLfL^3V3iLfLX3U3PMhLP3T3WMhLj2S3^MjLb2T3dMgL]2W3iMeLW2Y3TN^Ll1`3]NXLd1f3aNWL_1g3eNVL\\1h3hNVLX1h3lNULU1i3POTLP1j3TOSLm0k3WOSLi0l3ZOQLg0m3^OPLb0o3AnK`0Q4DkK=S4HjK8U4KhK6W4NeK3Y42cKO\\44aKM^47^KJb49[KGd4a4O100O1O100O1O100O100O100O10000O100O10000O10000O1000000O10000O10000O100000000O10000000000O10000000000000000000O101O0000000000000000001O0000000000000000001O000000000000000000001O001O1O001O1O1ZKaK=`4@cK?^4^OeKa0\\4[OiKc0X4ZOkKe0V4WOnKh0R4VORLh0o3UOTLj0m3ROWLm0j3POYLo0h3mN]LQ1d3lN_LS1b3iNbLV1_3fNeLY1\\3cNhL\\1Y3`NkL_1W3\\NmLc1U3XNoLg1S3TNQMk1Q3PNSMo1o2lMUMS2m2dM[M[2g2[MbMd2a2RMgMm2\\2hLmMW3V2^LSNa3m5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O2N3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3Mee[2"}, "label": "boy in orange shirt"}]}
{"id": 570440, "image": "COCO_train2014_000000570440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a second base man getting ready to play baseball\"."}], "task": "refering", "answer_template": "The \"a second base man getting ready to play baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 85, 86, 87, 99, 100, 101, 102, 114, 115, 116, 117, 118, 130, 131, 132, 145, 146, 147, 160, 161, 162, 175, 176, 177, 190, 191, 192], "bbox": [0.6548591549295776, 0.20725, 0.8766197183098593, 0.5411093749999999], "iscrowd": 0, "area": 11491.413449999993, "rle": {"size": [640, 426], "counts": "Pb^5`0Zc08H8H7M4L4L4Lh0XO8H8H3M3O1O1N2O1O7H8I6J7H8aAiKe<`4QCcKl<g4jB\\KT=m4bBWKZ=S5\\BPKb=h5N3L3N2N2O2O00001O000000000001N101O1[OlB\\JU=]5j0M8M5KN2O0O2O1N2O000O100O010O100OTOeATL[>i3iAVLV>h3oAULR>7mAY25_Mm=:`Ba1GSNj=<fBW1D[Nh=>jBl0CdNd=`0oBc0@lNc=a0SC9^OUO`=a0XC1\\O[O^=e0ZCH[OA^=f0[CB[OE\\=j0\\CZO\\OI[=l0]CTO[ON[=n0R40100O001006I01O001N2M2O2M3N3L4M3K6IamP1"}, "label": "a second base man getting ready to play baseball"}]}
{"id": 570440, "image": "COCO_train2014_000000570440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball pitcher with jersey number 21 holding the ball in his right hand\"."}], "task": "refering", "answer_template": "The \"a baseball pitcher with jersey number 21 holding the ball in his right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 127, 128, 141, 142, 143, 144, 156, 157, 158, 159, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 246, 247, 248, 261, 262, 263, 264, 276, 277, 278, 279, 291, 292, 293, 294, 305, 306, 307, 308, 309, 322, 323], "bbox": [0.3575821596244132, 0.304171875, 0.680868544600939, 0.9198281249999999], "iscrowd": 0, "area": 28325.642900000003, "rle": {"size": [640, 426], "counts": "gao21ic06N3N2O00000O1000000UKJkE7n91QFOk97SFIk9:TFFk9=RFDl9=UFCi9?WFAh9`0XF@f94UBCT4:d95ZB@R4;_99`B\\OQ4;\\9=cBWOP4<\\9d0_BPOT4=[9f0nFZOP9i0nFXOP9j0PGVOn8m0PGTOn8m0SGSOl8m0TGTOh8e0RChN=4k2a0e8]OcDU15AV2No8@aDR154m1\\OZ9B_DP14l0`1cNj9E]Dm04W1`1XNl9H[Dk02c1a1kMV9GXE3Jj05i1]1eMZ9S2VEn2h:V3N2N2N2N2N2O1O2N100O1O1O1O1hN_FQHa9g7oFPHQ9_O]Fn7l0`Hh8B^Fk7k0bHh8B_Fk7h0cHi8C_Fi7j0cHg8C`Fj7n0]HV7JgH0VOg7S1YHn6;aHD_Oh7U1UHi6k0YHYOJe7_9ZI`FSO1c7\\9QJeFn5Y9UJhFj5g7YHUIo1UOg5b7gHnHe1@d5]7RIjH]1I`5c7TIQHh1=T5h7fKXHY4o7`KSH_4P8^KPHb4R8[KoGe4S8XKnGi4S8TKnGl4T8RKlGn4U8PKlGQ5X8hJjGX5Z8cJgG^5[8]JgGc5\\8XJfGh5]8SJeGn5]8kIgGU6\\8cIiG^6^8VIgGi6a8fHhGZ7^:2O0PEVHj9P8mETHQ:T8eEPHZ:S8`EQH_:Q8[ETHd:e8N2O001O0O2O1O001N2O1lKfF`MZ9Y2PGcLESN\\9R5aGZL\\O`NU9n4iGYL[OfNm8h4SHYLXOmNf8b4\\HXLVOSO`8\\4eHXLSOZOY8\\4hHQLWO@S8^4iHbKg9]4o2N2O1O1O1O1N2O1O2N2M3N1O6I6K5J6K5K5J4M4K5L3L4eNQ_O_ORa06\\_O@k`05c_OAc`05^gd2"}, "label": "a baseball pitcher with jersey number 21 holding the ball in his right hand"}]}
{"id": 570440, "image": "COCO_train2014_000000570440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball pitcher holding the baseball\"."}], "task": "refering", "answer_template": "The \"the baseball pitcher holding the baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 127, 128, 141, 142, 143, 144, 156, 157, 158, 159, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 246, 247, 248, 261, 262, 263, 264, 276, 277, 278, 279, 291, 292, 293, 294, 305, 306, 307, 308, 309, 322, 323], "bbox": [0.3575821596244132, 0.304171875, 0.680868544600939, 0.9198281249999999], "iscrowd": 0, "area": 28325.642900000003, "rle": {"size": [640, 426], "counts": "gao21ic06N3N2O00000O1000000UKJkE7n91QFOk97SFIk9:TFFk9=RFDl9=UFCi9?WFAh9`0XF@f94UBCT4:d95ZB@R4;_99`B\\OQ4;\\9=cBWOP4<\\9d0_BPOT4=[9f0nFZOP9i0nFXOP9j0PGVOn8m0PGTOn8m0SGSOl8m0TGTOh8e0RChN=4k2a0e8]OcDU15AV2No8@aDR154m1\\OZ9B_DP14l0`1cNj9E]Dm04W1`1XNl9H[Dk02c1a1kMV9GXE3Jj05i1]1eMZ9S2VEn2h:V3N2N2N2N2N2O1O2N100O1O1O1O1hN_FQHa9g7oFPHQ9_O]Fn7l0`Hh8B^Fk7k0bHh8B_Fk7h0cHi8C_Fi7j0cHg8C`Fj7n0]HV7JgH0VOg7S1YHn6;aHD_Oh7U1UHi6k0YHYOJe7_9ZI`FSO1c7\\9QJeFn5Y9UJhFj5g7YHUIo1UOg5b7gHnHe1@d5]7RIjH]1I`5c7TIQHh1=T5h7fKXHY4o7`KSH_4P8^KPHb4R8[KoGe4S8XKnGi4S8TKnGl4T8RKlGn4U8PKlGQ5X8hJjGX5Z8cJgG^5[8]JgGc5\\8XJfGh5]8SJeGn5]8kIgGU6\\8cIiG^6^8VIgGi6a8fHhGZ7^:2O0PEVHj9P8mETHQ:T8eEPHZ:S8`EQH_:Q8[ETHd:e8N2O001O0O2O1O001N2O1lKfF`MZ9Y2PGcLESN\\9R5aGZL\\O`NU9n4iGYL[OfNm8h4SHYLXOmNf8b4\\HXLVOSO`8\\4eHXLSOZOY8\\4hHQLWO@S8^4iHbKg9]4o2N2O1O1O1O1N2O1O2N2M3N1O6I6K5J6K5K5J4M4K5L3L4eNQ_O_ORa06\\_O@k`05c_OAc`05^gd2"}, "label": "the baseball pitcher holding the baseball"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in blue jacket on skis\"."}], "task": "refering", "answer_template": "The \"person in blue jacket on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 29, 30, 31, 32, 52, 53, 54, 55, 74, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123], "bbox": [0.258296875, 0.002701949860724234, 0.410640625, 0.4449303621169916], "iscrowd": 0, "area": 12962.719799999997, "rle": {"size": [359, 640], "counts": "Sni1a0_:g0YO;E:F:H8J6I7]Oc0O0M1M4M3M2O2N2N1M3N2O2N1O1N2O1O1O1N3N1N2O1O1N2O2N1N2O0O2O1000000000O100000000000000000000000001O0000000000000001O0000000000000000000000O100O1O1M3N2N2N2jL^Ie1d6ZN`Ia1b6^NbI]1`6aNdIY1`6fNbIU1b6jNaIo0d6PO_Ij0e6UO^Id0g6[O\\I>j6@\\I4k6K\\IGl68S2N6JWUT4"}, "label": "person in blue jacket on skis"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the backpack of the last skier\"."}], "task": "refering", "answer_template": "The \"the backpack of the last skier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 29, 30, 31, 32, 52, 53, 54, 55, 74, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123], "bbox": [0.258296875, 0.002701949860724234, 0.410640625, 0.4449303621169916], "iscrowd": 0, "area": 12962.719799999997, "rle": {"size": [359, 640], "counts": "Sni1a0_:g0YO;E:F:H8J6I7]Oc0O0M1M4M3M2O2N2N1M3N2O2N1O1N2O1O1O1N3N1N2O1O1N2O2N1N2O0O2O1000000000O100000000000000000000000001O0000000000000001O0000000000000000000000O100O1O1M3N2N2N2jL^Ie1d6ZN`Ia1b6^NbI]1`6aNdIY1`6fNbIU1b6jNaIo0d6PO_Ij0e6UO^Id0g6[O\\I>j6@\\I4k6K\\IGl68S2N6JWUT4"}, "label": "the backpack of the last skier"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is in the middle of two men riding on skis\"."}], "task": "refering", "answer_template": "The \"a man is in the middle of two men riding on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 151, 152, 174, 175, 197, 198], "bbox": [0.575046875, 0.14239554317548747, 0.687140625, 0.6730362116991644], "iscrowd": 0, "area": 7318.504949999998, "rle": {"size": [359, 640], "counts": "mTQ45W:2UF7f9g0L5K4L5K5`H\\NY5h1bJ\\N\\5g1lHRN`1:b5k1YJYNf5i1dImNY6W1YIEU6f2O2N2N1O2N1010O010O010lNWJiLj5T3U1L5I7L44nHdLP6a3eInLS6V3cI\\MQ6R4I5_O>O10O10O10O010FdJiJ[5V5iJgJX5W5kJgJU5Y5=\\O\\JcKe5a3VJdLb0GZ5_3]JaLa0KV5]3bJcL^6f0oHi1m7TNWHn0MSOQ8LUHk05oNl71SHi0>kNf77oGg0U9TOnFg0l90O4M3M2N4K4N2NO1O1O1O1N2N4L9H;D2N3HjoU2"}, "label": "a man is in the middle of two men riding on skis"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person skiing wearing a white hat\"."}], "task": "refering", "answer_template": "The \"a person skiing wearing a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 151, 152, 174, 175, 197, 198], "bbox": [0.575046875, 0.14239554317548747, 0.687140625, 0.6730362116991644], "iscrowd": 0, "area": 7318.504949999998, "rle": {"size": [359, 640], "counts": "mTQ45W:2UF7f9g0L5K4L5K5`H\\NY5h1bJ\\N\\5g1lHRN`1:b5k1YJYNf5i1dImNY6W1YIEU6f2O2N2N1O2N1010O010O010lNWJiLj5T3U1L5I7L44nHdLP6a3eInLS6V3cI\\MQ6R4I5_O>O10O10O10O010FdJiJ[5V5iJgJX5W5kJgJU5Y5=\\O\\JcKe5a3VJdLb0GZ5_3]JaLa0KV5]3bJcL^6f0oHi1m7TNWHn0MSOQ8LUHk05oNl71SHi0>kNf77oGg0U9TOnFg0l90O4M3M2N4K4N2NO1O1O1O1N2N4L9H;D2N3HjoU2"}, "label": "a person skiing wearing a white hat"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis being worn by man with yellow stripes on his shoes\"."}], "task": "refering", "answer_template": "The \"skis being worn by man with yellow stripes on his shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 261, 262, 263, 281, 282, 283, 284], "bbox": [0.23537499999999997, 0.7942339832869081, 0.480625, 1.0], "iscrowd": 0, "area": 1014.9710500000015, "rle": {"size": [359, 640], "counts": "mXe11U;1O10O01O010O00100O0010O01O10O010O0100O010O010O10O010O2O0O101O`EOj90VF1i90WF0i9OWF2i9NTF4l9MTF3l9LTF4m9LRF5m9KTF4m9LRF5n9KQF5o9KRF5n9KQF5P:JQF6n9KQF5P:JPF7P:IPF6Q:JnE7Q:IoE7R:InE7`:0BImE6U:JkE5c:0CKjE5c:100BJkE6b:100O100O1000O1O1O10001O0O1000O100O10O0O2O001O0100O010O10O^OOPF1P:0nE0R:1nEOR:0nE0S:0mE0R:0oE0Q:0nE1R:NoE2P:OoE1R:NkE4V:MiE4W:KjE4W:LiE4V:LjE5V:KjE4V:LjE5V:KjE4W:KjE5U:LjE4W:KjE5U:LkE4U:KkE6U:JkE6T:JlE7T:IlE6U:IlE6T:KkE4W:KjE4V:MjE3T:NlE2S:0mE0Q:1oEOP:3PFMn94SFKl94VFMh9O^F1\\:O010O010O10O0100O010O010O10O010O0100O010O10O010O0O2N101N1O1ON2N3MhSd3"}, "label": "skis being worn by man with yellow stripes on his shoes"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier skiing behind two other skiers\"."}], "task": "refering", "answer_template": "The \"a skier skiing behind two other skiers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 282, 283, 284], "bbox": [0.166390625, 0.002256267409470752, 0.455046875, 0.9460724233983286], "iscrowd": 0, "area": 33602.4437, "rle": {"size": [359, 640], "counts": "aYU11U;:E8I1N2O1N20000O2O0O100O100O100O10001N010O1N1O2N2M2O2N1O2N2N1O2N2N1O2N2O0O2N00000010O000000[O]NhGd1X8d01O0010O6K6I7J5K:Ge0ZOe0\\Od0[Oe0\\Od0[Oe0[Od0]O0O100O1fMUHX1k7bN_HY1a7cNhHX1W7gNoHV1o6hNUIW1k6gNXIY1f6gN\\IX1d6hN\\IY1b6hN_IU1b6kN`IR1b6nN_IP1a6PO`In0b6RO_Ik0b6UO`Ii0`6XO`If0b6ZO_Ic0b6]O`Ia0a6_O`I?`6AaI?^6BcI>\\6AfI?X6BiI>U6CkI=T6CnI=P6DQJ<m5DTJ<k5KPJ5n5b0\\I_Ob6j0VIWOh6l0VITOi6o0UIROi6Q1UIPOi6T1TImNj6V1UIjNi6Y1UIhMGg0R7d1_I\\N_6i1^IWN`6d1gI\\NX6T1YJlNe5c0nJ]OP53bKM\\4BWL>g3A\\L?c3@_L`0_3@cL`0[3AfL?X3AjL?U3@mL`0Q3APM?n2ATM?j2AXM?g2AZM?d2A^M?`2AbM?\\2AfM?Y2AhM?V2AlM?R2APN?n1BSN>l1AVN?h1AZN?d1B]N>a1BaN>^1AdN?Z1BgN>W1BkN>S1BoN>P1ARO?l0BUO>j0AXO?i0^ORN0_Mb0a4[OiM>cM7e4XObMk0gMMe7=XHC`7g0^HYOZ7R1cHnNU7\\1iHdNl6j1QIVNR5YOmKQ3oNfMT5\\O]K^3\\OVMW5_OlJk3KfLX5R4eJnK[5U4bJkK^5o400O2O00000O10000O1000000J6hKRJQ3T6lLQJl2U6RMnIh2W6VMmIe2W6XMmIb2X6[MlI_2X6_MlI[2Y6bMjIY2[6dMhIW2\\6gMfIV2^6gMeIT2i7L4L5K4M3L5K6J6J6J8IaYj3"}, "label": "a skier skiing behind two other skiers"}]}
{"id": 136267, "image": "COCO_train2014_000000136267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skater follwing two other skaters on snow\"."}], "task": "refering", "answer_template": "The \"skater follwing two other skaters on snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 28, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 282, 283, 284], "bbox": [0.166390625, 0.002256267409470752, 0.455046875, 0.9460724233983286], "iscrowd": 0, "area": 33602.4437, "rle": {"size": [359, 640], "counts": "aYU11U;:E8I1N2O1N20000O2O0O100O100O100O10001N010O1N1O2N2M2O2N1O2N2N1O2N2N1O2N2O0O2N00000010O000000[O]NhGd1X8d01O0010O6K6I7J5K:Ge0ZOe0\\Od0[Oe0\\Od0[Oe0[Od0]O0O100O1fMUHX1k7bN_HY1a7cNhHX1W7gNoHV1o6hNUIW1k6gNXIY1f6gN\\IX1d6hN\\IY1b6hN_IU1b6kN`IR1b6nN_IP1a6PO`In0b6RO_Ik0b6UO`Ii0`6XO`If0b6ZO_Ic0b6]O`Ia0a6_O`I?`6AaI?^6BcI>\\6AfI?X6BiI>U6CkI=T6CnI=P6DQJ<m5DTJ<k5KPJ5n5b0\\I_Ob6j0VIWOh6l0VITOi6o0UIROi6Q1UIPOi6T1TImNj6V1UIjNi6Y1UIhMGg0R7d1_I\\N_6i1^IWN`6d1gI\\NX6T1YJlNe5c0nJ]OP53bKM\\4BWL>g3A\\L?c3@_L`0_3@cL`0[3AfL?X3AjL?U3@mL`0Q3APM?n2ATM?j2AXM?g2AZM?d2A^M?`2AbM?\\2AfM?Y2AhM?V2AlM?R2APN?n1BSN>l1AVN?h1AZN?d1B]N>a1BaN>^1AdN?Z1BgN>W1BkN>S1BoN>P1ARO?l0BUO>j0AXO?i0^ORN0_Mb0a4[OiM>cM7e4XObMk0gMMe7=XHC`7g0^HYOZ7R1cHnNU7\\1iHdNl6j1QIVNR5YOmKQ3oNfMT5\\O]K^3\\OVMW5_OlJk3KfLX5R4eJnK[5U4bJkK^5o400O2O00000O10000O1000000J6hKRJQ3T6lLQJl2U6RMnIh2W6VMmIe2W6XMmIb2X6[MlI_2X6_MlI[2Y6bMjIY2[6dMhIW2\\6gMfIV2^6gMeIT2i7L4L5K4M3L5K6J6J6J8IaYj3"}, "label": "skater follwing two other skaters on snow"}]}
{"id": 463953, "image": "COCO_train2014_000000463953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing behind the man wearing white shirt\"."}], "task": "refering", "answer_template": "The \"a man standing behind the man wearing white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 137, 229, 252, 275, 297, 298, 320, 321, 343, 344, 366, 367, 389, 390], "bbox": [0.9174687499999999, 0.19891666666666669, 1.0, 0.9912916666666667], "iscrowd": 0, "area": 8091.5823499999915, "rle": {"size": [480, 640], "counts": "ZXc81n>4XL0YH3_73`H0Y75fHNS77lHMk69SIKf6:YII_6=`IEZ6`0eIBU6c0kI]Oo5i0QJWOj5n0VJROd5T1\\JmN^5X1bJhNY5]1gJcNS5d1lJ]Nn4h1RKXNi4m1WKSNc4S2]KnM]4W2cKiMW4]2iKcMS4a2mK`Mo3c2QL]Ml3f2TL[Mg3i2YLXMc3k2]LVM_3m2aLTM[3o2eLRMW3Q3iLPMR3T3nLmLn2V3RMkLj2X3VMiLf2[3XMgLd2\\3\\MeL_2_3aMbL[2a3fM_LV2d3kM\\LQ2g3PNYLl1j3TNVLh1n3YNRLc1Q4^NoK^1T4cNlKY1W4gNjKU1Y4lNgKo0]4ROcKj0`4WO`Ke0c4[O^Ka0e4@[K<i4DWK7m4JRK3Q5MfJ9]5NoIm2c3aI"}, "label": "a man standing behind the man wearing white shirt"}]}
{"id": 463953, "image": "COCO_train2014_000000463953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man watching two adults playing a video game\"."}], "task": "refering", "answer_template": "The \"a man watching two adults playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 137, 229, 252, 275, 297, 298, 320, 321, 343, 344, 366, 367, 389, 390], "bbox": [0.9174687499999999, 0.19891666666666669, 1.0, 0.9912916666666667], "iscrowd": 0, "area": 8091.5823499999915, "rle": {"size": [480, 640], "counts": "ZXc81n>4XL0YH3_73`H0Y75fHNS77lHMk69SIKf6:YII_6=`IEZ6`0eIBU6c0kI]Oo5i0QJWOj5n0VJROd5T1\\JmN^5X1bJhNY5]1gJcNS5d1lJ]Nn4h1RKXNi4m1WKSNc4S2]KnM]4W2cKiMW4]2iKcMS4a2mK`Mo3c2QL]Ml3f2TL[Mg3i2YLXMc3k2]LVM_3m2aLTM[3o2eLRMW3Q3iLPMR3T3nLmLn2V3RMkLj2X3VMiLf2[3XMgLd2\\3\\MeL_2_3aMbL[2a3fM_LV2d3kM\\LQ2g3PNYLl1j3TNVLh1n3YNRLc1Q4^NoK^1T4cNlKY1W4gNjKU1Y4lNgKo0]4ROcKj0`4WO`Ke0c4[O^Ka0e4@[K<i4DWK7m4JRK3Q5MfJ9]5NoIm2c3aI"}, "label": "a man watching two adults playing a video game"}]}
{"id": 463953, "image": "COCO_train2014_000000463953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing glasses and a blue shirt holding a wii controller\"."}], "task": "refering", "answer_template": "The \"a woman wearing glasses and a blue shirt holding a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 101, 102, 103, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.22921874999999997, 0.21797916666666667, 0.5561875, 0.9887708333333334], "iscrowd": 0, "area": 48350.77419999999, "rle": {"size": [480, 640], "counts": "jRU24l>5K6J6J6J6J5K5K1O1O1O100O1O1O1O1O2N1O1O1O1O1O1O010O0010O10O0100O010O10O0100O010O10O0100OVIZN:f1B^N?b1[OcNf0\\1VOhNj0a1iNcNX1W2iMmMX2[2[MiMd2`2PMdMn2`2lLdMS3`2fLdMY3`2`LeM^3^2kKWNT4m1mJoNR5`5O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O12N2N2N1O2N2O0O2N2NKaGjI^8W6cGiI[8W6hGhIV8Y6kGfIT8[6nGdIP8]6QHcIm7]6VHbIh7_6<1O00100O001O1O00SNeHVLZ7g3kHXLT7d3RIZLn6b3XI]Lg6_3_I`L`6\\3fIaL[6[3kIcLU6Y3QJdLQ6[3QJbLP6]3RJbLn5^3TJ_Lm5a3UJ]Lk5b3XJ[Li5e3YJYLg5f3j200O10000O100O2O000O1000000000000001O000001O000001O0000000000001O0001O00000001XFbLl7^3jGPMR8P3gG[MU8e2kG_MQ8a2nGdMn7]2PHgMm7Y2RHlMj7T2VHlMj7U2THlMl7T2SHnMm7S2QHmMo7U2nGlMR8V2kGkMU8V2iGlMV8V2hGjMX8X2eGiM[8Y2bGhM^8Z2`GgM_8[2^GfMb8\\2[GeMe8[401O0001O0001O0000001O0000001O0000010E:B>C=B>B`0A`0_Oa0_Oa0@`0_Oa0@a0Db0^Ob0_Ob0]Oc0]Oc0]OgkT4"}, "label": "a woman wearing glasses and a blue shirt holding a wii controller"}]}
{"id": 463953, "image": "COCO_train2014_000000463953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing tan pants\"."}], "task": "refering", "answer_template": "The \"a woman wearing tan pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 101, 102, 103, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.22921874999999997, 0.21797916666666667, 0.5561875, 0.9887708333333334], "iscrowd": 0, "area": 48350.77419999999, "rle": {"size": [480, 640], "counts": "jRU24l>5K6J6J6J6J5K5K1O1O1O100O1O1O1O1O2N1O1O1O1O1O1O010O0010O10O0100O010O10O0100O010O10O0100OVIZN:f1B^N?b1[OcNf0\\1VOhNj0a1iNcNX1W2iMmMX2[2[MiMd2`2PMdMn2`2lLdMS3`2fLdMY3`2`LeM^3^2kKWNT4m1mJoNR5`5O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O12N2N2N1O2N2O0O2N2NKaGjI^8W6cGiI[8W6hGhIV8Y6kGfIT8[6nGdIP8]6QHcIm7]6VHbIh7_6<1O00100O001O1O00SNeHVLZ7g3kHXLT7d3RIZLn6b3XI]Lg6_3_I`L`6\\3fIaL[6[3kIcLU6Y3QJdLQ6[3QJbLP6]3RJbLn5^3TJ_Lm5a3UJ]Lk5b3XJ[Li5e3YJYLg5f3j200O10000O100O2O000O1000000000000001O000001O000001O0000000000001O0001O00000001XFbLl7^3jGPMR8P3gG[MU8e2kG_MQ8a2nGdMn7]2PHgMm7Y2RHlMj7T2VHlMj7U2THlMl7T2SHnMm7S2QHmMo7U2nGlMR8V2kGkMU8V2iGlMV8V2hGjMX8X2eGiM[8Y2bGhM^8Z2`GgM_8[2^GfMb8\\2[GeMe8[401O0001O0001O0000001O0000001O0000010E:B>C=B>B`0A`0_Oa0_Oa0@`0_Oa0@a0Db0^Ob0_Ob0]Oc0]Oc0]OgkT4"}, "label": "a woman wearing tan pants"}]}
{"id": 283729, "image": "COCO_train2014_000000283729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl on a skateboard\"."}], "task": "refering", "answer_template": "The \"a girl on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 20, 21, 22, 23, 34, 35, 36, 37, 38, 39, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 80, 81, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 101, 102, 103, 110, 111, 112, 113, 125, 126, 127, 128, 140, 141, 142, 143, 155, 156, 157, 158, 170, 171, 172, 173, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277], "bbox": [0.2802112676056338, 0.02303125, 0.8879107981220657, 0.8162968749999999], "iscrowd": 0, "area": 50519.994000000006, "rle": {"size": [640, 426], "counts": "SS[25hc04L5K4L5K5K4L5K4L3M4M2N3N2M2O2M2O2M3M2O2N1O2O1N101N101N2N10Z1eN;F_1`N:G9F9H9F8I6hEoHT7W7WHbI_7e6XHhI]7_6[HlI[7Z6]HRJX7T6bHWLT5m3gJXLU5l3fJXLW5l3cJZLY5k3aJYM\\4k2_KWM`4m2[KUMd4T9O2N1O1O1O2N1O1O1O1O100O1O1O10000000000O010000O10O010000O3K4L5K5K4L5K3M3M3N21O1O2N1O1O00000000000001O0000001O0hNlJdCT5X<PKhCP5T<UKkCl4P<XKPDh4l;]KSDd4h;`KXDa4c;dK\\D]4_;gK`D[4\\;iKcDX4Y;kKfDW4U;nKjDS4R;QLlDQ4P;SLoDn3l:WLREm3h:WLVEn3c:WL[El3^:YL`Ek3e9g3^L`3E;E;I7L5K4M3L4L4L4L5J5J6K5J6K6I6K5J6K5L410O01O001OO101N1O1O107H9G1O100O1O1O100O1O1O2O0O1O1O100O2N2N3M2O2M2N2N100O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O00001O000000001O01O000001O000000001O000000001O0001O00O2O00O0100000O010000O01000O10O1000O10O12M3N1O2N2M3M3N2M3N2M3M[jm0"}, "label": "a girl on a skateboard"}]}
{"id": 283729, "image": "COCO_train2014_000000283729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl who is riding on a skateboard wearing a white shirt and a necklace\"."}], "task": "refering", "answer_template": "The \"a girl who is riding on a skateboard wearing a white shirt and a necklace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 20, 21, 22, 23, 34, 35, 36, 37, 38, 39, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 80, 81, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 101, 102, 103, 110, 111, 112, 113, 125, 126, 127, 128, 140, 141, 142, 143, 155, 156, 157, 158, 170, 171, 172, 173, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277], "bbox": [0.2802112676056338, 0.02303125, 0.8879107981220657, 0.8162968749999999], "iscrowd": 0, "area": 50519.994000000006, "rle": {"size": [640, 426], "counts": "SS[25hc04L5K4L5K5K4L5K4L3M4M2N3N2M2O2M2O2M3M2O2N1O2O1N101N101N2N10Z1eN;F_1`N:G9F9H9F8I6hEoHT7W7WHbI_7e6XHhI]7_6[HlI[7Z6]HRJX7T6bHWLT5m3gJXLU5l3fJXLW5l3cJZLY5k3aJYM\\4k2_KWM`4m2[KUMd4T9O2N1O1O1O2N1O1O1O1O100O1O1O10000000000O010000O10O010000O3K4L5K5K4L5K3M3M3N21O1O2N1O1O00000000000001O0000001O0hNlJdCT5X<PKhCP5T<UKkCl4P<XKPDh4l;]KSDd4h;`KXDa4c;dK\\D]4_;gK`D[4\\;iKcDX4Y;kKfDW4U;nKjDS4R;QLlDQ4P;SLoDn3l:WLREm3h:WLVEn3c:WL[El3^:YL`Ek3e9g3^L`3E;E;I7L5K4M3L4L4L4L5J5J6K5J6K6I6K5J6K5L410O01O001OO101N1O1O107H9G1O100O1O1O100O1O1O2O0O1O1O100O2N2N3M2O2M2N2N100O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O00001O000000001O01O000001O000000001O000000001O0001O00O2O00O0100000O010000O01000O10O1000O10O12M3N1O2N2M3M3N2M3N2M3M[jm0"}, "label": "a girl who is riding on a skateboard wearing a white shirt and a necklace"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle rier in black\"."}], "task": "refering", "answer_template": "The \"the motorcycle rier in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 106, 107, 129, 130, 151, 152, 153, 154, 175, 176, 198, 199, 220, 221, 222, 243, 244], "bbox": [0.58940625, 0.2866666666666667, 0.7375156250000001, 0.8090277777777778], "iscrowd": 0, "area": 6436.18215, "rle": {"size": [360, 640], "counts": "`iT42T;:G8H7J0O2O1gMTOSJm0i5YOUJh0e5^OYJc0c5C[J>a5F]J;`5J^J6`5M^J3`52^JLa59\\JCf5a0YJXOk5m0RJROn5Q1QJmNn5X1nIgNR6]1lIcNT6`1hIbNW6a1fI_NZ6c1dI]N\\6f1aI[NP6@\\IW2b0YNk5_2RJbMk5b2SJ^Mm5e2PJ\\Mo5g2nIYMR6i2lIWMT6l2iIUMV6m2hISMX6P3eIQMZ6Q3dIoL\\6T3aIlL_6o3O001O0100O1O010O100SNbIoN_6YOiI9L;\\6\\OoI0J`0X6@SJIId0U6DRJDOe0P6FQJE2a0n5JPJE4>n5LoIE6;l50nIE88l53kIE<4j57kID>1i5:iIF?Mi5=hIFb0Ih5a0fIEd0Gg5d0eIEg0Dd5g0fIDh0Bd5i0dIES8;mGET8;lGDT85SHKn7MYH4g7E`H:`7@fH`0Q901O001O1O1O1O100O1O1O001O00O10001O000000000O12M4L4LZlj1"}, "label": "the motorcycle rier in black"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a guy on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 106, 107, 129, 130, 151, 152, 153, 154, 175, 176, 198, 199, 220, 221, 222, 243, 244], "bbox": [0.58940625, 0.2866666666666667, 0.7375156250000001, 0.8090277777777778], "iscrowd": 0, "area": 6436.18215, "rle": {"size": [360, 640], "counts": "`iT42T;:G8H7J0O2O1gMTOSJm0i5YOUJh0e5^OYJc0c5C[J>a5F]J;`5J^J6`5M^J3`52^JLa59\\JCf5a0YJXOk5m0RJROn5Q1QJmNn5X1nIgNR6]1lIcNT6`1hIbNW6a1fI_NZ6c1dI]N\\6f1aI[NP6@\\IW2b0YNk5_2RJbMk5b2SJ^Mm5e2PJ\\Mo5g2nIYMR6i2lIWMT6l2iIUMV6m2hISMX6P3eIQMZ6Q3dIoL\\6T3aIlL_6o3O001O0100O1O010O100SNbIoN_6YOiI9L;\\6\\OoI0J`0X6@SJIId0U6DRJDOe0P6FQJE2a0n5JPJE4>n5LoIE6;l50nIE88l53kIE<4j57kID>1i5:iIF?Mi5=hIFb0Ih5a0fIEd0Gg5d0eIEg0Dd5g0fIDh0Bd5i0dIES8;mGET8;lGDT85SHKn7MYH4g7E`H:`7@fH`0Q901O001O1O1O1O100O1O1O001O00O10001O000000000O12M4L4LZlj1"}, "label": "a guy on a motorcycle"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white van in the background\"."}], "task": "refering", "answer_template": "The \"the white van in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 103, 104, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 155, 156, 169, 170, 171, 172, 174], "bbox": [0.368953125, 0.29891666666666666, 0.815328125, 0.5956944444444444], "iscrowd": 0, "area": 14621.538150000004, "rle": {"size": [360, 640], "counts": "`Tc2<Y:d0O1O1O1N2O1O1O2N1O1N2O1O000000000O10000000000O2O3M4L3M2N0O10000000000O1000000000000O10000000000O100O1O100O1O1O100O1O010O00010O0000010O00010O0000010O]ORGmNn8U1XGcNi8]1]G]Nb8e1^GZNa8g1_GYNa8g1`GXN_8j1`GVN_8k1aGUN^8m1aGSN_8m1aGSN^8n1bGRN]8P2=001N1000001N1000000O101O000O1000001N1000000O101O000O1000001N1000000O101O000O1000001N1000000O101O000O100000O001O1O10O01O2N1O1O1O1O1I7F:G9J6O1O100O1O1O2N1O1O1O1O1]OkE2V:DXF8`:K5KXd5Oi[J5L3M4K5L2N0O100nN[O`Gf0`8ZO_Gg0a8ZO]Gh0b8YO]Gg0c8YO\\Gh0d8YOZGh0f8XORGNDk0Z9XOmF4Fd0]9YOhFV1X9jNdFZ1\\99000000000000000000001O0001O00N2O1N2[O]FDd9=\\FAe9?\\F^Of9b0[F]Oe9c0\\F\\Od9d0]F\\Ob9e0]F\\Ob9d0_F[Oa9e0`F[O_9e0bF[O]9e0dFZO\\9f0eFZOZ9g0fFYOY9g0gF[OW9e0jF\\OT9d0mF\\OR9d0oF]Oo8c0RG^Om8b0SG^Ol8b0UG_Oi8a0WG_Oi8a0XG]Oi8c0]GWOc8j0bGoN_8Q1l0001O1O001000O10001O0O10001N10000O2O000O101O0O101O1O0O2O001N110O001O0010O0fZY1"}, "label": "the white van in the background"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver van\"."}], "task": "refering", "answer_template": "The \"a silver van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 103, 104, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 155, 156, 169, 170, 171, 172, 174], "bbox": [0.368953125, 0.29891666666666666, 0.815328125, 0.5956944444444444], "iscrowd": 0, "area": 14621.538150000004, "rle": {"size": [360, 640], "counts": "`Tc2<Y:d0O1O1O1N2O1O1O2N1O1N2O1O000000000O10000000000O2O3M4L3M2N0O10000000000O1000000000000O10000000000O100O1O100O1O1O100O1O010O00010O0000010O00010O0000010O]ORGmNn8U1XGcNi8]1]G]Nb8e1^GZNa8g1_GYNa8g1`GXN_8j1`GVN_8k1aGUN^8m1aGSN_8m1aGSN^8n1bGRN]8P2=001N1000001N1000000O101O000O1000001N1000000O101O000O1000001N1000000O101O000O1000001N1000000O101O000O100000O001O1O10O01O2N1O1O1O1O1I7F:G9J6O1O100O1O1O2N1O1O1O1O1]OkE2V:DXF8`:K5KXd5Oi[J5L3M4K5L2N0O100nN[O`Gf0`8ZO_Gg0a8ZO]Gh0b8YO]Gg0c8YO\\Gh0d8YOZGh0f8XORGNDk0Z9XOmF4Fd0]9YOhFV1X9jNdFZ1\\99000000000000000000001O0001O00N2O1N2[O]FDd9=\\FAe9?\\F^Of9b0[F]Oe9c0\\F\\Od9d0]F\\Ob9e0]F\\Ob9d0_F[Oa9e0`F[O_9e0bF[O]9e0dFZO\\9f0eFZOZ9g0fFYOY9g0gF[OW9e0jF\\OT9d0mF\\OR9d0oF]Oo8c0RG^Om8b0SG^Ol8b0UG_Oi8a0WG_Oi8a0XG]Oi8c0]GWOc8j0bGoN_8Q1l0001O1O001000O10001O0O10001N10000O2O000O101O0O101O1O0O2O001N110O001O0010O0fZY1"}, "label": "a silver van"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color car\"."}], "task": "refering", "answer_template": "The \"a black color car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165], "bbox": [0.0, 0.35155555555555557, 0.21140625000000002, 0.5931666666666667], "iscrowd": 0, "area": 10003.980650000001, "rle": {"size": [360, 640], "counts": "^4;l:g0ZOf0YOa0@000O2O0O10000O10000O100O10000O100O10000O100O10000O10000O100000000000000000000000O010000000O10000000000O10000000000O01000O1000000O10000O1000000O1000000O1001O0000000000000000000000000000001O00000000001O00000000O100O2O0O2N1O1O2N1O101M2L5K4L4L5K4L4L5M2M3N3M2N5K6J7Ig\\a5"}, "label": "a black color car"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"compact car\"."}], "task": "refering", "answer_template": "The \"compact car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165], "bbox": [0.0, 0.35155555555555557, 0.21140625000000002, 0.5931666666666667], "iscrowd": 0, "area": 10003.980650000001, "rle": {"size": [360, 640], "counts": "^4;l:g0ZOf0YOa0@000O2O0O10000O10000O100O10000O100O10000O100O10000O10000O100000000000000000000000O010000000O10000000000O10000000000O01000O1000000O10000O1000000O1000000O1001O0000000000000000000000000000001O00000000001O00000000O100O2O0O2N1O1O2N1O101M2L5K4L4L5K4L4L5M2M3N3M2N5K6J7Ig\\a5"}, "label": "compact car"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white dressed a man riding a white sports bike\"."}], "task": "refering", "answer_template": "The \"a white dressed a man riding a white sports bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235], "bbox": [0.183875, 0.35052777777777777, 0.330234375, 0.8236666666666667], "iscrowd": 0, "area": 7032.481649999999, "rle": {"size": [360, 640], "counts": "leY19j:9F:G9VGoNP7W1oHnNjNHi7]1\\IoNfNEl7_1]IPO`6T1]ImN`6W1_IjN^6Y1`IhN^6[1aIfN\\6]1bIdN\\6^1cIdNZ6_1dIbNY6b1eI`NX6b1gI_NV6e1hI\\NU6g1jI[NS6h1kIYNR6j1mIXNl5o1RJRNh5T2WJnMd5W2YJkMd5Y2ZJiMb5Z2]JgMc5Z2[JhMd5X2ZJ^M0ROf5a3XJ]MU6d2iI\\MX6i31O1O1O1O1O1O1O1O1O1O1O2N1RN\\IWOf6i0bImN_6S1cIjN_6NVIJ<6_60VII=4_61VIK<1_6CnHK8`0<0`6DnHJ7a0=N`6FmHJ6b0?K_6InHH5d0>Ia6IdI=MG`6LdI<MEb6NbI<NCa61bI<N@b63aI<N_Ob64aI=O\\Ob66`I=0ZOa69`I=OXOc69`I>g7AYH?g7@ZH`0W91O0000001OO100000000O100000000O1000000EiEEW:;<O000O2O00001O3M3MTdf4"}, "label": "a white dressed a man riding a white sports bike"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white riding suit in the back on cycle\"."}], "task": "refering", "answer_template": "The \"man in white riding suit in the back on cycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235], "bbox": [0.183875, 0.35052777777777777, 0.330234375, 0.8236666666666667], "iscrowd": 0, "area": 7032.481649999999, "rle": {"size": [360, 640], "counts": "leY19j:9F:G9VGoNP7W1oHnNjNHi7]1\\IoNfNEl7_1]IPO`6T1]ImN`6W1_IjN^6Y1`IhN^6[1aIfN\\6]1bIdN\\6^1cIdNZ6_1dIbNY6b1eI`NX6b1gI_NV6e1hI\\NU6g1jI[NS6h1kIYNR6j1mIXNl5o1RJRNh5T2WJnMd5W2YJkMd5Y2ZJiMb5Z2]JgMc5Z2[JhMd5X2ZJ^M0ROf5a3XJ]MU6d2iI\\MX6i31O1O1O1O1O1O1O1O1O1O1O2N1RN\\IWOf6i0bImN_6S1cIjN_6NVIJ<6_60VII=4_61VIK<1_6CnHK8`0<0`6DnHJ7a0=N`6FmHJ6b0?K_6InHH5d0>Ia6IdI=MG`6LdI<MEb6NbI<NCa61bI<N@b63aI<N_Ob64aI=O\\Ob66`I=0ZOa69`I=OXOc69`I>g7AYH?g7@ZH`0W91O0000001OO100000000O100000000O1000000EiEEW:;<O000O2O00001O3M3MTdf4"}, "label": "man in white riding suit in the back on cycle"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white motorcycle type vehicle , not including the driver\"."}], "task": "refering", "answer_template": "The \"the white motorcycle type vehicle , not including the driver\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 162, 163, 164, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 260, 261, 262], "bbox": [0.008875, 0.507888888888889, 0.49409375000000005, 0.9065277777777778], "iscrowd": 0, "area": 26689.628200000006, "rle": {"size": [360, 640], "counts": "`[2k0Z:6J7J5J7J3O1N2O1O1O0O2O1O1O0O2O00001N101O001N101O000O2O001O010O0000001O00000@kMnGU2P8RNjGn1U8c0N2O1O2N100O101O0O10000O10O10O01000O0100O010O1O100O100O1O100O1O100O1O1O1O1O1O1O101N100O100O101O000000001O1O1O1O1O10O01O001O001O1O010O001O001O1O0000001O000001O0001O000000001N1000000000O1000000O10000O1000O01000000O10000O1O100O1O100O1O100O00101N100O2O0O101N100O2O000O2O00001O001O000O2O00000O1O1O1O1O1O1O2O0N2H8M32N2N2O1N1O01O00001O001O0O1O2O1N2N2O1N3J4O2N1O2O0O1O2O001O010O0010O001O01O01O00010O001O012M2N2N2O0O2WLgH_3c7N3M3M2M4M2N001N2O001N2O001N2O001O0O1M4M2M4M2M4M2M4O0O101N101N101N101N100O2O0O2O001O0O2O001O001O1O001O1O1O001O1O3M3M2M4L7H<E4L4KVda3"}, "label": "the white motorcycle type vehicle , not including the driver"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white motorcycle\"."}], "task": "refering", "answer_template": "The \"white motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 162, 163, 164, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 260, 261, 262], "bbox": [0.008875, 0.507888888888889, 0.49409375000000005, 0.9065277777777778], "iscrowd": 0, "area": 26689.628200000006, "rle": {"size": [360, 640], "counts": "`[2k0Z:6J7J5J7J3O1N2O1O1O0O2O1O1O0O2O00001N101O001N101O000O2O001O010O0000001O00000@kMnGU2P8RNjGn1U8c0N2O1O2N100O101O0O10000O10O10O01000O0100O010O1O100O100O1O100O1O100O1O1O1O1O1O1O101N100O100O101O000000001O1O1O1O1O10O01O001O001O1O010O001O001O1O0000001O000001O0001O000000001N1000000000O1000000O10000O1000O01000000O10000O1O100O1O100O1O100O00101N100O2O0O101N100O2O000O2O00001O001O000O2O00000O1O1O1O1O1O1O2O0N2H8M32N2N2O1N1O01O00001O001O0O1O2O1N2N2O1N3J4O2N1O2O0O1O2O001O010O0010O001O01O01O00010O001O012M2N2N2O0O2WLgH_3c7N3M3M2M4M2N001N2O001N2O001N2O001O0O1M4M2M4M2M4M2M4O0O101N101N101N101N100O2O0O2O001O0O2O001O001O1O001O1O1O001O1O3M3M2M4L7H<E4L4KVda3"}, "label": "white motorcycle"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle with a person on riding it\"."}], "task": "refering", "answer_template": "The \"a black motorcycle with a person on riding it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 149, 150, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 248, 249], "bbox": [0.49803125000000004, 0.4041111111111111, 0.862078125, 0.8041111111111111], "iscrowd": 0, "area": 16603.591999999997, "rle": {"size": [360, 640], "counts": "dZ`32V;2M4M2M3WFK[87\\G4_8N]G9_8I[G?a8X1L5K3M2N100O2N1O2N101N1O0010O0001O0010O0001O00010O0010O0010O010O01O010O000N3N1N3N1N3N1N3M2O2M2O2N1O2N1N2O2N1O2M2O2N1O2O0O2N102M2N2O1N3M2O1N3M2N2N2TOeEh0_:N2MWF]OR9a0l0NSFAT9=lFDU9:kFGV97jFJW94jFKY92gFOZ9OfF2W:1O00100O1O1O1O100O1O1O100O1XNCeH=Z7FcH;\\7HbH8]7K`H6^7N`H2_71^H0a73]HMb76[HKd78ZHHe7<WHEh7?UHAk7a0RH@n7c0oG]OQ8e0lG\\OT8g0iGYOW8j0eGWO[8k0cGUO]8n0_GSOa8o0]GQOc8k1O00000000000000001O000000000000000VO^MnHc2Q7^MnHb2R7`MlH`2U7`MjH`2V7bMhH^2X7cMgH]2Y7cMgH]2Y7dMeH]2[7cMcH_2^7aM`H`2b7^M]Hc2f7[MUHi2n773M2NYORMUIm2d6[M]Ic2d6_MZI`2f6eMUI[2k6iMQIW2P7mMjHT2V7PNfHP2Z7o0001O00001O0N3J6I610000000000O1000000001O01O0000001O0001O0O1O101N100O100lMQHm0Q8oNQHR1o7lNSHS1n7jNTHW1l7gNVHX1k7fNVHZ1l7cNVH\\1k7`NXH`1i7\\N[Hc1f7ZN\\Hf1f7UN^Hj1`801O1O001O1O001O1O0N3M2N3M3M2N3M3M3M3H;E:Fdfn0"}, "label": "a black motorcycle with a person on riding it"}]}
{"id": 275544, "image": "COCO_train2014_000000275544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey sporty motorcycle with a rider\"."}], "task": "refering", "answer_template": "The \"a grey sporty motorcycle with a rider\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 149, 150, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 248, 249], "bbox": [0.49803125000000004, 0.4041111111111111, 0.862078125, 0.8041111111111111], "iscrowd": 0, "area": 16603.591999999997, "rle": {"size": [360, 640], "counts": "dZ`32V;2M4M2M3WFK[87\\G4_8N]G9_8I[G?a8X1L5K3M2N100O2N1O2N101N1O0010O0001O0010O0001O00010O0010O0010O010O01O010O000N3N1N3N1N3N1N3M2O2M2O2N1O2N1N2O2N1O2M2O2N1O2O0O2N102M2N2O1N3M2O1N3M2N2N2TOeEh0_:N2MWF]OR9a0l0NSFAT9=lFDU9:kFGV97jFJW94jFKY92gFOZ9OfF2W:1O00100O1O1O1O100O1O1O100O1XNCeH=Z7FcH;\\7HbH8]7K`H6^7N`H2_71^H0a73]HMb76[HKd78ZHHe7<WHEh7?UHAk7a0RH@n7c0oG]OQ8e0lG\\OT8g0iGYOW8j0eGWO[8k0cGUO]8n0_GSOa8o0]GQOc8k1O00000000000000001O000000000000000VO^MnHc2Q7^MnHb2R7`MlH`2U7`MjH`2V7bMhH^2X7cMgH]2Y7cMgH]2Y7dMeH]2[7cMcH_2^7aM`H`2b7^M]Hc2f7[MUHi2n773M2NYORMUIm2d6[M]Ic2d6_MZI`2f6eMUI[2k6iMQIW2P7mMjHT2V7PNfHP2Z7o0001O00001O0N3J6I610000000000O1000000001O01O0000001O0001O0O1O101N100O100lMQHm0Q8oNQHR1o7lNSHS1n7jNTHW1l7gNVHX1k7fNVHZ1l7cNVH\\1k7`NXH`1i7\\N[Hc1f7ZN\\Hf1f7UN^Hj1`801O1O001O1O001O1O0N3M2N3M3M2N3M3M3M3H;E:Fdfn0"}, "label": "a grey sporty motorcycle with a rider"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver serving tray with lemonade\"."}], "task": "refering", "answer_template": "The \"the silver serving tray with lemonade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 241, 242, 243, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 365, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.48034375, 0.5393333333333333, 0.9994375000000002, 0.9865208333333333], "iscrowd": 0, "area": 24655.995099999993, "rle": {"size": [480, 640], "counts": "PW`42_>a0L5K5L4K5K5L4K5K5K7J5J7I6K1N3M3M2O2M3M2O2M3M2N3N2M3M2O2O10O0100O010O10O01O10O0100O10O0100O010O10O0100O010O1O010O100O01YMVEW1i:[NeEf1[:jMTFV2\\;0O10O0100O00100O010O1O10O0100O00100O010O1O010O10O0100O1O010O10O01O10O0100O00100O010O1O1O0O2M3M2O2M3M2O2M3N1N3M3N1N3M3N20O10000O100O10000O10O10O100O10000O100O10000O100O10000O10000O010O10000O100O10000O100O10000O100O10000O01000O100O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000O01000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O11O1N101O1O001O001O1O001O1O001OjA"}, "label": "the silver serving tray with lemonade"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a platter where to put food on\"."}], "task": "refering", "answer_template": "The \"a platter where to put food on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 241, 242, 243, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 365, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.48034375, 0.5393333333333333, 0.9994375000000002, 0.9865208333333333], "iscrowd": 0, "area": 24655.995099999993, "rle": {"size": [480, 640], "counts": "PW`42_>a0L5K5L4K5K5L4K5K5K7J5J7I6K1N3M3M2O2M3M2O2M3M2N3N2M3M2O2O10O0100O010O10O01O10O0100O10O0100O010O10O0100O010O1O010O100O01YMVEW1i:[NeEf1[:jMTFV2\\;0O10O0100O00100O010O1O10O0100O00100O010O1O010O10O0100O1O010O10O01O10O0100O00100O010O1O1O0O2M3M2O2M3M2O2M3N1N3M3N1N3M3N20O10000O100O10000O10O10O100O10000O100O10000O100O10000O10000O010O10000O100O10000O100O10000O100O10000O01000O100O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000O01000000O10000O1000000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O11O1N101O1O001O001O1O001O1O001OjA"}, "label": "a platter where to put food on"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pieces of bread in a gold bowl\"."}], "task": "refering", "answer_template": "The \"pieces of bread in a gold bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 262, 263], "bbox": [0.272359375, 0.38685416666666667, 0.5508124999999999, 0.6605624999999999], "iscrowd": 0, "area": 16514.718749999996, "rle": {"size": [480, 640], "counts": "iia22Y>d0N21O01O002O2M3M2O2M3M3N1N3M3M2O2M3M3N1N3M3M2O2M3M3N1N3M3N1N3M3M3N1N3M3N1N3M3M3N1N2N100O1O2O0O1O1O100O1O100O1O1O2O000000000000000001O0001O000000000001O0000000001O0001O00000000000000010O00000000000000001O0001O000000000001O0000000001O0001O00000000000000010O001O001O1N2M2N3M3M2N3M3N2M2N3M3M3M2N3M3M3M2O2M3M3M2N3M3M2N3M3M3M2O2M3M3M2N3M3M3M2N3M3M2O2M3M3M2N3MmhV4"}, "label": "pieces of bread in a gold bowl"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small bread basket\"."}], "task": "refering", "answer_template": "The \"a small bread basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 262, 263], "bbox": [0.272359375, 0.38685416666666667, 0.5508124999999999, 0.6605624999999999], "iscrowd": 0, "area": 16514.718749999996, "rle": {"size": [480, 640], "counts": "iia22Y>d0N21O01O002O2M3M2O2M3M3N1N3M3M2O2M3M3N1N3M3M2O2M3M3N1N3M3N1N3M3M3N1N3M3N1N3M3M3N1N2N100O1O2O0O1O1O100O1O100O1O1O2O000000000000000001O0001O000000000001O0000000001O0001O00000000000000010O00000000000000001O0001O000000000001O0000000001O0001O00000000000000010O001O001O1N2M2N3M3M2N3M3N2M2N3M3M3M2N3M3M3M2O2M3M3M2N3M3M2N3M3M3M2O2M3M3M2N3M3M3M2N3M3M2O2M3M3M2N3MmhV4"}, "label": "a small bread basket"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lemonade\"."}], "task": "refering", "answer_template": "The \"lemonade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 177, 178, 179, 180, 201, 202, 203, 224, 225, 226], "bbox": [0.6859531249999999, 0.17304166666666668, 0.92359375, 0.5820208333333333], "iscrowd": 0, "area": 18710.873099999993, "rle": {"size": [480, 640], "counts": "To]64f>6I7J7H7J6I7J7H7J6J7I6J6J6I8I6L4O101N100O2O0O100O2O0O2O0O2O1N2O1N2O1N2O0O2O1jFoKZ7S4bHPL^7Q4^HSL`7o3\\HTLd7m3XHVLg7l3UHWLk7i3RHZLn7g3nG]LQ8d3kG_LV8a3fGbLZ8_3bGdL^8]3^GfLb8[3ZGiLf8W3VGlLj8T3SGoLm8b40010O00000001O0000001O000O10001O0000001N100000001O000O101O000000001N1000001M2M3M3M3M4L3N2N2M4M2M3N2M3N3M2M3N2M4M2M3N2N3L3N3L4M2M4M2N3L3N3L4M2M4M2N3L3N3L3N3L4M2N3L3N3L3N3L4M2M4M2N3L3N3L4M2M4L3M4L3NZkf0"}, "label": "lemonade"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass with the drink that has lemons in it\"."}], "task": "refering", "answer_template": "The \"the glass with the drink that has lemons in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 177, 178, 179, 180, 201, 202, 203, 224, 225, 226], "bbox": [0.6859531249999999, 0.17304166666666668, 0.92359375, 0.5820208333333333], "iscrowd": 0, "area": 18710.873099999993, "rle": {"size": [480, 640], "counts": "To]64f>6I7J7H7J6I7J7H7J6J7I6J6J6I8I6L4O101N100O2O0O100O2O0O2O0O2O1N2O1N2O1N2O0O2O1jFoKZ7S4bHPL^7Q4^HSL`7o3\\HTLd7m3XHVLg7l3UHWLk7i3RHZLn7g3nG]LQ8d3kG_LV8a3fGbLZ8_3bGdL^8]3^GfLb8[3ZGiLf8W3VGlLj8T3SGoLm8b40010O00000001O0000001O000O10001O0000001N100000001O000O101O000000001N1000001M2M3M3M3M4L3N2N2M4M2M3N2M3N3M2M3N2M4M2M3N2N3L3N3L4M2M4M2N3L3N3L4M2M4M2N3L3N3L3N3L4M2N3L3N3L3N3L4M2M4M2N3L3N3L4M2M4L3M4L3NZkf0"}, "label": "the glass with the drink that has lemons in it"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass half full of beer with no logo visible , in front of a man in a striped shirt\"."}], "task": "refering", "answer_template": "The \"a glass half full of beer with no logo visible , in front of a man in a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 37, 38, 57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 126, 127, 128, 129, 150, 151, 173, 174, 196, 197, 219], "bbox": [0.502890625, 0.048749999999999995, 0.668640625, 0.5402083333333334], "iscrowd": 0, "area": 15130.924350000003, "rle": {"size": [480, 640], "counts": "kPg4n0l==B>F8I7H7J5J5L4K5L4^EmLP9X3kFlLQ9Y3jFkLS9Y3hFkLT9Z3gFjLV9Z3eFjLW9[3dFiLY9[3bFiLZ9\\3aFhL\\9_4O2M3M3N3oGUJh6n5SIVJk6l5QIXJn6j5mHZJQ7i5kHYJU7j5fHXJY7h600000000O1000eN\\HZKd7d4_H[Ka7d4bHZK^7c4fH\\KY7c4kH[KU7b4PI\\KP7a4TI^Kl6_4XI`Kh6^4ZIbKf6[4^IaKe6\\4^IQKTONa7n4^ISKS7k4oHTKR7i4QIVKP7i4RIUKo6j4RIUKo6j4RITKP7l4c1N1O1O100N2O1O1O1O1O1O1O2N1O1N2O2N1O1O1O2N1O1O1O2M2O1O1O2N1N2O1N3N1N2N2O2M2O1N2N3L3L5K5L3L5K5K5L4K7I8In0QOlYS3"}, "label": "a glass half full of beer with no logo visible , in front of a man in a striped shirt"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass of beer to the right of the basket of bread\"."}], "task": "refering", "answer_template": "The \"the glass of beer to the right of the basket of bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 37, 38, 57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 126, 127, 128, 129, 150, 151, 173, 174, 196, 197, 219], "bbox": [0.502890625, 0.048749999999999995, 0.668640625, 0.5402083333333334], "iscrowd": 0, "area": 15130.924350000003, "rle": {"size": [480, 640], "counts": "kPg4n0l==B>F8I7H7J5J5L4K5L4^EmLP9X3kFlLQ9Y3jFkLS9Y3hFkLT9Z3gFjLV9Z3eFjLW9[3dFiLY9[3bFiLZ9\\3aFhL\\9_4O2M3M3N3oGUJh6n5SIVJk6l5QIXJn6j5mHZJQ7i5kHYJU7j5fHXJY7h600000000O1000eN\\HZKd7d4_H[Ka7d4bHZK^7c4fH\\KY7c4kH[KU7b4PI\\KP7a4TI^Kl6_4XI`Kh6^4ZIbKf6[4^IaKe6\\4^IQKTONa7n4^ISKS7k4oHTKR7i4QIVKP7i4RIUKo6j4RIUKo6j4RITKP7l4c1N1O1O100N2O1O1O1O1O1O1O2N1O1N2O2N1O1O1O2N1O1O1O2M2O1O1O2N1N2O1N3N1N2N2O2M2O1N2N3L3L5K5L3L5K5K5L4K7I8In0QOlYS3"}, "label": "the glass of beer to the right of the basket of bread"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with ring on her finger eating a hamburger\"."}], "task": "refering", "answer_template": "The \"girl with ring on her finger eating a hamburger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 47, 48, 49, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 142, 143, 144, 145], "bbox": [0.0050625, 0.0045000000000000005, 0.37584375000000003, 0.37527083333333333], "iscrowd": 0, "area": 26356.72744999999, "rle": {"size": [480, 640], "counts": "S^18l=IfBP1j<QOnBV1R=;10O0100O1000O101O00000O100000000O100000000O10000000001O00000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000005K5K4L5K3M3M2N3hE^Ml7d2fGiMY8Y2ZGSNe8P2lF\\NT9f1^FgNa9]3O1O1O001O1O1O1O1O001O1O000000000000001O0000000000000001O01O0000000000000000O100000000000000O10000000000001O1O001O1O1O00000000000000000000000000000000000000000000O1000000000000000000000000000000N2N2N2N2N2N2N2O1N2N2L4G9G9G9H8F:G9G9G;E>D;I8G8H8H9H7H8Hkoj5"}, "label": "girl with ring on her finger eating a hamburger"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding a piece of bread\"."}], "task": "refering", "answer_template": "The \"the woman holding a piece of bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 47, 48, 49, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 142, 143, 144, 145], "bbox": [0.0050625, 0.0045000000000000005, 0.37584375000000003, 0.37527083333333333], "iscrowd": 0, "area": 26356.72744999999, "rle": {"size": [480, 640], "counts": "S^18l=IfBP1j<QOnBV1R=;10O0100O1000O101O00000O100000000O100000000O10000000001O00000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000005K5K4L5K3M3M2N3hE^Ml7d2fGiMY8Y2ZGSNe8P2lF\\NT9f1^FgNa9]3O1O1O001O1O1O1O1O001O1O000000000000001O0000000000000001O01O0000000000000000O100000000000000O10000000000001O1O001O1O1O00000000000000000000000000000000000000000000O1000000000000000000000000000000N2N2N2N2N2N2N2O1N2N2L4G9G9G9H8F:G9G9G;E>D;I8G8H8H9H7H8Hkoj5"}, "label": "the woman holding a piece of bread"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in striped shirt\"."}], "task": "refering", "answer_template": "The \"person in striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 31, 32, 33, 34, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 84, 85, 88, 89, 90, 91, 100, 101, 102, 103, 106, 107, 108, 112, 113, 114, 123, 124, 125, 126, 127, 130, 131, 147, 148, 149, 150], "bbox": [0.354828125, 0.002125, 1.0, 0.4012291666666667], "iscrowd": 0, "area": 38144.32655, "rle": {"size": [480, 640], "counts": "n`Z36a>:F:F;D;F:F;E:E;F;E:G9I7H9H7J6O2N1O1O2N1O1O1O2N1O1O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001OO1000000000000000000O1000000000000000000O100TMoFIQ9UOUHg0k7TO^Hh0b7WOcHe0]7YOiHc0W7[OnHb0R7\\ORIb0n6\\OUIc0k6[OYIb0h6\\O\\Ib0d6]O_Ia0a6]OcIa0]6]OgIa0Y6]OkIa0U6]OoIa0Q6]OUJ?k5@YJ=g5A_J;a5CeJ8\\5FRKNn40dK@\\4>T5N2O1N20000000000000000O10000000000000000O1000000000000000000O1000000000000000000O10000000000000000O1000000000000000000O10000000000001O2N2N2N2N2N1WE[OU7g0eH_OY7c0aHA_7a0[HCe7?UHEk7=oGGQ8;iGJV88dGL\\86^GNb84WG1i81QG3o8OjF6V9LdF8\\9J]F;c9GWF=i9EPF`0P:BjEc0U:_OdEf0\\:o100000000000001N1000000000000000001O00000000000000N2I7I7I7I7I7I7I7K5M3L4O100000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O001O001O1O001>A8I7I7H8I7H5LO1O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001N2O001O00O100O100O10O0100O100O100O100O1000O100000O105K<B=A?A?AhL"}, "label": "person in striped shirt"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a striped polo shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a striped polo shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 31, 32, 33, 34, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 84, 85, 88, 89, 90, 91, 100, 101, 102, 103, 106, 107, 108, 112, 113, 114, 123, 124, 125, 126, 127, 130, 131, 147, 148, 149, 150], "bbox": [0.354828125, 0.002125, 1.0, 0.4012291666666667], "iscrowd": 0, "area": 38144.32655, "rle": {"size": [480, 640], "counts": "n`Z36a>:F:F;D;F:F;E:E;F;E:G9I7H9H7J6O2N1O1O2N1O1O1O2N1O1O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001OO1000000000000000000O1000000000000000000O100TMoFIQ9UOUHg0k7TO^Hh0b7WOcHe0]7YOiHc0W7[OnHb0R7\\ORIb0n6\\OUIc0k6[OYIb0h6\\O\\Ib0d6]O_Ia0a6]OcIa0]6]OgIa0Y6]OkIa0U6]OoIa0Q6]OUJ?k5@YJ=g5A_J;a5CeJ8\\5FRKNn40dK@\\4>T5N2O1N20000000000000000O10000000000000000O1000000000000000000O1000000000000000000O10000000000000000O1000000000000000000O10000000000001O2N2N2N2N2N1WE[OU7g0eH_OY7c0aHA_7a0[HCe7?UHEk7=oGGQ8;iGJV88dGL\\86^GNb84WG1i81QG3o8OjF6V9LdF8\\9J]F;c9GWF=i9EPF`0P:BjEc0U:_OdEf0\\:o100000000000001N1000000000000000001O00000000000000N2I7I7I7I7I7I7I7K5M3L4O100000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O001O001O1O001>A8I7I7H8I7H5LO1O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001N2O001O00O100O100O10O0100O100O100O100O1000O100000O105K<B=A?A?AhL"}, "label": "a man wearing a striped polo shirt"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beer glass with complete stella artois label showing\"."}], "task": "refering", "answer_template": "The \"beer glass with complete stella artois label showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 185, 186, 187, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256], "bbox": [0.003359375, 0.11208333333333333, 0.19798437500000002, 0.6870208333333334], "iscrowd": 0, "area": 24152.825850000005, "rle": {"size": [480, 640], "counts": "QS17T>g0XOh0YOg0ZOf0M2M4M3M3L4M3[H`Ld3d3WLcLd3_3WLjLc3Y3XLoLa3U3ZLRMa3Q3ZLRMe3Q3ULRMj3R3RLPMm3S3PLnLP4S3oKmLQ4U3mKkLS4W3jKjLV4X3hKgLY4[3eKeL[4]3bKdL]4_3aKaL_4a3_K_La4b3^K^Lb4d3[K\\Lf4f3XKZLh4h3VKXLj4j3SKWLm4k3QKULo4m3oJSLQ5n3nJQLS5Q4jJPLV5R4hJnKX5T4fJlKY5W4dJjK\\5\\4^JdKb5e4UJZKl5n4kISKU6U5cIkJ]6l600000O10000000O10O10000000000O1000000000O10O10000000000O100^N[IiJe6T5^IlJc6P5aIoJ_6n4dIRK]6j4gITKZ6i4iIWKX6e4kI[KU6b4oI]KR6_4QJaKo5]4TJbKl5\\4VJcKl5Z4WJeKi5Z4XJfKi5W4YJiKg5U4\\JjKe5T4\\JlKd5R4_JlKc5Q4_JoKa5P4aJoK`5n3bJRL^5l3dJTL\\5k3fJTL[5i3gJVLZ5h3iJWLX5g3iJYLW5e3lJZLU5c3mJ]LS5b3nJ^LS5_3PK_LQ5_3QKaLP5]3VK^Lj4`3fKRL[4k3^3O1N3M2O2M2N3N3L7I7J6I7H8C>C<D<D<D<C=DU[`7"}, "label": "beer glass with complete stella artois label showing"}]}
{"id": 365659, "image": "COCO_train2014_000000365659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass that says stella artois\"."}], "task": "refering", "answer_template": "The \"a glass that says stella artois\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 185, 186, 187, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256], "bbox": [0.003359375, 0.11208333333333333, 0.19798437500000002, 0.6870208333333334], "iscrowd": 0, "area": 24152.825850000005, "rle": {"size": [480, 640], "counts": "QS17T>g0XOh0YOg0ZOf0M2M4M3M3L4M3[H`Ld3d3WLcLd3_3WLjLc3Y3XLoLa3U3ZLRMa3Q3ZLRMe3Q3ULRMj3R3RLPMm3S3PLnLP4S3oKmLQ4U3mKkLS4W3jKjLV4X3hKgLY4[3eKeL[4]3bKdL]4_3aKaL_4a3_K_La4b3^K^Lb4d3[K\\Lf4f3XKZLh4h3VKXLj4j3SKWLm4k3QKULo4m3oJSLQ5n3nJQLS5Q4jJPLV5R4hJnKX5T4fJlKY5W4dJjK\\5\\4^JdKb5e4UJZKl5n4kISKU6U5cIkJ]6l600000O10000000O10O10000000000O1000000000O10O10000000000O100^N[IiJe6T5^IlJc6P5aIoJ_6n4dIRK]6j4gITKZ6i4iIWKX6e4kI[KU6b4oI]KR6_4QJaKo5]4TJbKl5\\4VJcKl5Z4WJeKi5Z4XJfKi5W4YJiKg5U4\\JjKe5T4\\JlKd5R4_JlKc5Q4_JoKa5P4aJoK`5n3bJRL^5l3dJTL\\5k3fJTL[5i3gJVLZ5h3iJWLX5g3iJYLW5e3lJZLU5c3mJ]LS5b3nJ^LS5_3PK_LQ5_3QKaLP5]3VK^Lj4`3fKRL[4k3^3O1N3M2O2M2N3N3L7I7J6I7H8C>C<D<D<D<C=DU[`7"}, "label": "a glass that says stella artois"}]}
{"id": 5215, "image": "COCO_train2014_000000005215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leg . of a woman wearing jeans and a shoe\"."}], "task": "refering", "answer_template": "The \"the leg . of a woman wearing jeans and a shoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.0016718750000000002, 0.23145569620253162, 0.577515625, 0.9977426160337552], "iscrowd": 0, "area": 85296.42974999998, "rle": {"size": [474, 640], "counts": "Um06c>9G1O1N2O0iJ_OVLb0`3I]L9X33fLNo2>oLCf2i0XMXO]2T1`MnNU2_1hMaNS2f1kM[NP2k1mMWNo1n1oMSNm1R2QNoMj1W2TNjMh1\\2TNfMg1`2WNaM>nN`Lh3P3ZM>o3_OSL=R4AoK;V4CkK<X4AiK?X4^OjKa0X4]OiKb0X4]OhKc0Z4ZOhKd0[4ZOfKd0]4ZOdKd0_4ZObKd0a4ZO`Ke0b4YO_Ke0d4YO]Ke0f4YOZKf0i4XOXKf0k4XOVKg0l4VOVKj0k4TOVKl0k4ROVKn0k4POVKP1k4nNVKR1k4lNVKS1l4kNUKU1k4jNVKV1k4hNVKX1k4fNVKZ1k4dNVK\\1k4bNVK^1k4`NVK`1k4^NVKb1k4\\NVKd1k4ZNVKf1k4XNVKh1k4UNWKk1j4SNWKm1j4QNWKo1j4oMWKR2i4lMXKU2h4iMYKW2h4gMYKZ2g4dMZK]2f4aM[K_2e4`M\\Ka2d4]M]Kd2c4ZM^Kf2c4XM^Ki2b4UM_Kl2a4RM`Kn2a4PM`KS3^4kLcKW3\\4fLfK]3X4aLiKb3U4\\LlKf3S4XLnKk3P4SLQLo3n3oKSLT4k3jKVLX4i3fKXL\\4g3bKZL`4e3^K\\Ld4b3[K_Lh4_3VKbLn4c6<D;E`0@0000001O000000001O000000001O0000001O000000001O000000001O0000000000000000000000O100000001O000000000000000O100000000000000000000000O01000000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000000000000000000000000000001O000000000000001O00000000000000001O000000001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O001O1O2N1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O001O2N5K4L5K5K5K3M3M2N3M2N3M2N2N3M2N001O00000000O1N2O1N2N2N2O1O100000000000000000000000fK]G]2c8aMbG\\2^8cMfGZ2[8dMiGY2W8fMmGW2T8gMPHV2Q8hMSHU2n7iMVHT2j7kMZHR2g7lM]HQ2d7lMaHQ2`7lMeHQ2[7lMkHQ2V7lMoHQ2R7kMTIR2l6kM[IQ2f6lM_IQ2b6iMfIT2Z6gMnIV2S6cMVJZ2X9M2N3M3M2N3M3M2N3M2N3M3M2N3M3M2N3M3M2N3M3M2M4G8G^al3"}, "label": "the leg . of a woman wearing jeans and a shoe"}]}
{"id": 5215, "image": "COCO_train2014_000000005215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing blue jeans and black shoes\"."}], "task": "refering", "answer_template": "The \"a person wearing blue jeans and black shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.0016718750000000002, 0.23145569620253162, 0.577515625, 0.9977426160337552], "iscrowd": 0, "area": 85296.42974999998, "rle": {"size": [474, 640], "counts": "Um06c>9G1O1N2O0iJ_OVLb0`3I]L9X33fLNo2>oLCf2i0XMXO]2T1`MnNU2_1hMaNS2f1kM[NP2k1mMWNo1n1oMSNm1R2QNoMj1W2TNjMh1\\2TNfMg1`2WNaM>nN`Lh3P3ZM>o3_OSL=R4AoK;V4CkK<X4AiK?X4^OjKa0X4]OiKb0X4]OhKc0Z4ZOhKd0[4ZOfKd0]4ZOdKd0_4ZObKd0a4ZO`Ke0b4YO_Ke0d4YO]Ke0f4YOZKf0i4XOXKf0k4XOVKg0l4VOVKj0k4TOVKl0k4ROVKn0k4POVKP1k4nNVKR1k4lNVKS1l4kNUKU1k4jNVKV1k4hNVKX1k4fNVKZ1k4dNVK\\1k4bNVK^1k4`NVK`1k4^NVKb1k4\\NVKd1k4ZNVKf1k4XNVKh1k4UNWKk1j4SNWKm1j4QNWKo1j4oMWKR2i4lMXKU2h4iMYKW2h4gMYKZ2g4dMZK]2f4aM[K_2e4`M\\Ka2d4]M]Kd2c4ZM^Kf2c4XM^Ki2b4UM_Kl2a4RM`Kn2a4PM`KS3^4kLcKW3\\4fLfK]3X4aLiKb3U4\\LlKf3S4XLnKk3P4SLQLo3n3oKSLT4k3jKVLX4i3fKXL\\4g3bKZL`4e3^K\\Ld4b3[K_Lh4_3VKbLn4c6<D;E`0@0000001O000000001O000000001O0000001O000000001O000000001O0000000000000000000000O100000001O000000000000000O100000000000000000000000O01000000000000000000000000O1000000000000000000000000O1000000000000000000000000O10000000000000000000000000000001O000000000000001O00000000000000001O000000001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O001O1O2N1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O001O2N5K4L5K5K5K3M3M2N3M2N3M2N2N3M2N001O00000000O1N2O1N2N2N2O1O100000000000000000000000fK]G]2c8aMbG\\2^8cMfGZ2[8dMiGY2W8fMmGW2T8gMPHV2Q8hMSHU2n7iMVHT2j7kMZHR2g7lM]HQ2d7lMaHQ2`7lMeHQ2[7lMkHQ2V7lMoHQ2R7kMTIR2l6kM[IQ2f6lM_IQ2b6iMfIT2Z6gMnIV2S6cMVJZ2X9M2N3M3M2N3M3M2N3M2N3M3M2N3M3M2N3M3M2N3M3M2M4G8G^al3"}, "label": "a person wearing blue jeans and black shoes"}]}
{"id": 5215, "image": "COCO_train2014_000000005215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a womans jeans with red socks and brown boots\"."}], "task": "refering", "answer_template": "The \"a womans jeans with red socks and brown boots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 34, 35, 43, 44, 45, 56, 57, 58, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.459359375, 0.024345991561181432, 1.0, 0.9883966244725738], "iscrowd": 0, "area": 85391.26159999998, "rle": {"size": [474, 640], "counts": "kTX44e>7I6J7I7J6I7I6J3M3M3M3M3N2M3N2M3N2M3N1N3N2M3N2N2M3N2M3N2N2N2N1]EQMY9Q3UF`Mi9k3N2N2N1O2N2M201N2O1N101N2O0O2O1N2O0O01O001O2N2N2N3M2O1N2N2N2N2N2N2N1O2N1O101N1O1O2N1O1O1O2N1O01N100O01O01O01O001O1O1O1O1O100O1O1O6J7I1O1O100O2N1O1O;E1O001001N4M3L4M3M3L4L4L4L4L4L4K5L4L4L4L4L3M4L4M3M3L4M1N2O1O1N2O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O2M2O1O1N1000001N10001N10000O2O00000O2O00001O0O10001O0OO2N2N2N2N1O2N2N2N1N3N2N2N2N1O2N200000O1000O10O100000000O1000000O10O100000O100000008Gk0VOk0TOl0TOk0VOk0TOl0UO2M1O101N100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O2O0O1O100O100O1O100O100O100O10000O10000O100O10000O100O10001N10000O100O10000O100O10000O100O10000O10000O100O10000O10001N100D"}, "label": "a womans jeans with red socks and brown boots"}]}
{"id": 5215, "image": "COCO_train2014_000000005215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man crossing his legs with red socks and brown shoes\"."}], "task": "refering", "answer_template": "The \"a man crossing his legs with red socks and brown shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 34, 35, 43, 44, 45, 56, 57, 58, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.459359375, 0.024345991561181432, 1.0, 0.9883966244725738], "iscrowd": 0, "area": 85391.26159999998, "rle": {"size": [474, 640], "counts": "kTX44e>7I6J7I7J6I7I6J3M3M3M3M3N2M3N2M3N2M3N1N3N2M3N2N2M3N2M3N2N2N2N1]EQMY9Q3UF`Mi9k3N2N2N1O2N2M201N2O1N101N2O0O2O1N2O0O01O001O2N2N2N3M2O1N2N2N2N2N2N2N1O2N1O101N1O1O2N1O1O1O2N1O01N100O01O01O01O001O1O1O1O1O100O1O1O6J7I1O1O100O2N1O1O;E1O001001N4M3L4M3M3L4L4L4L4L4L4K5L4L4L4L4L3M4L4M3M3L4M1N2O1O1N2O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O2M2O1O1N1000001N10001N10000O2O00000O2O00001O0O10001O0OO2N2N2N2N1O2N2N2N1N3N2N2N2N1O2N200000O1000O10O100000000O1000000O10O100000O100000008Gk0VOk0TOl0TOk0VOk0TOl0UO2M1O101N100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O2O0O1O100O100O1O100O100O100O10000O10000O100O10000O100O10001N10000O100O10000O100O10000O100O10000O10000O100O10000O10001N100D"}, "label": "a man crossing his legs with red socks and brown shoes"}]}
{"id": 275556, "image": "COCO_train2014_000000275556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cake sitting on a table next to some people\"."}], "task": "refering", "answer_template": "The \"a cake sitting on a table next to some people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275], "bbox": [0.71678125, 0.5506791569086651, 1.0, 0.7826463700234192], "iscrowd": 0, "area": 14487.819700000002, "rle": {"size": [427, 640], "counts": "[do52Y=4L4_DNa96QF4o90dE9\\:KWE>i:R11O00000001O000000000001O00000001O0000001O0001O01O0000001O0000001O01O01O0000001O0000001O0000001O0000001O0001O01O000000000000001O00000000000000001O000000000001O0000O2O00000000000O10001O000000000O1000001O00000O1000000000000O2O0000000000000O100000001O000O1000000000000O10001O000000000O1000000000000O101O000000000O100000000000000O101O0000000000000O1000PNeJ"}, "label": "a cake sitting on a table next to some people"}]}
{"id": 275556, "image": "COCO_train2014_000000275556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheet cake\"."}], "task": "refering", "answer_template": "The \"a sheet cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275], "bbox": [0.71678125, 0.5506791569086651, 1.0, 0.7826463700234192], "iscrowd": 0, "area": 14487.819700000002, "rle": {"size": [427, 640], "counts": "[do52Y=4L4_DNa96QF4o90dE9\\:KWE>i:R11O00000001O000000000001O00000001O0000001O0001O01O0000001O0000001O01O01O0000001O0000001O0000001O0000001O0001O01O000000000000001O00000000000000001O000000000001O0000O2O00000000000O10001O000000000O1000001O00000O1000000000000O2O0000000000000O100000001O000O1000000000000O10001O000000000O1000000000000O101O000000000O100000000000000O101O0000000000000O1000PNeJ"}, "label": "a sheet cake"}]}
{"id": 185447, "image": "COCO_train2014_000000185447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a riderless black motorcycle parked sideways in front of a brick building\"."}], "task": "refering", "answer_template": "The \"a riderless black motorcycle parked sideways in front of a brick building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 138, 139, 140, 141], "bbox": [0.64442, 0.4202710843373494, 0.86124, 0.6674096385542169], "iscrowd": 0, "area": 5063.396649999999, "rle": {"size": [332, 500], "counts": "ZbX33W:7H6J5L4K4L3N3L3M4N1O101N1O100O100O100O00100O100O1O10O0100O101N100:F3M000001O01O0000000010O000000010O10O3N2N1OO1O1O1N2O1O001O1O1N101O1O001N2O1O001O1O0O2O00001O0O1000001O000O100000000O01000000O1O0O2N2WO^GFd88cGA_8>hGZOZ8e0f0O2O2M;F;D^Vf0"}, "label": "a riderless black motorcycle parked sideways in front of a brick building"}]}
{"id": 185447, "image": "COCO_train2014_000000185447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle parked in front of a brick house\"."}], "task": "refering", "answer_template": "The \"a black motorcycle parked in front of a brick house\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 138, 139, 140, 141], "bbox": [0.64442, 0.4202710843373494, 0.86124, 0.6674096385542169], "iscrowd": 0, "area": 5063.396649999999, "rle": {"size": [332, 500], "counts": "ZbX33W:7H6J5L4K4L3N3L3M4N1O101N1O100O100O100O00100O100O1O10O0100O101N100:F3M000001O01O0000000010O000000010O10O3N2N1OO1O1O1N2O1O001O1O1N101O1O001N2O1O001O1O0O2O00001O0O1000001O000O100000000O01000000O1O0O2N2WO^GFd88cGA_8>hGZOZ8e0f0O2O2M;F;D^Vf0"}, "label": "a black motorcycle parked in front of a brick house"}]}
{"id": 185447, "image": "COCO_train2014_000000185447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle with a man standing next to it\"."}], "task": "refering", "answer_template": "The \"a motorcycle with a man standing next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 94, 95, 96, 97, 112, 113, 114, 115, 130, 131, 132, 133], "bbox": [0.21597999999999998, 0.39713855421686745, 0.44838, 0.6696686746987951], "iscrowd": 0, "area": 5277.3021, "rle": {"size": [332, 500], "counts": "jUS13U:4O2O001O0XFH\\99aFI_97_FJb97ZFLf9=0O1O00iFAc8>[GEe8;ZGFf8;WGLd84[G0b81[G2d8l0O1O1O1O1O1O001O001O0O101O0000000O01000000O10O10O0100O10O00LVNhGi1X8XNgGg1Z8ZNfGe1Z8\\NeGc1\\8_NcG`1^852L401N100O2O0O10YG`N[8OhGa1M`NX82jG^1O`NW81kG_1M`NY81jGa1K]N]82gGc1[8\\NdGc1]8^NbG_1N_Nb82^G_1f822O001O1N1000O1O010BdNkG]1T8cNmG]1S8bNmG^1T8aNmG_1S8bNkG^1W8cNgG]1[8bNdG]1^8dN`G\\1a8dN]G\\1f841N200100O0010aNXGW1n8001N101OkN[Gc0e8]O[Gc0g8[OYGd0j8ZOVGf0k8XOVGh0l8VOUGi0m8UOSGk0W900O102M3L3N3L4JnYi2"}, "label": "a motorcycle with a man standing next to it"}]}
{"id": 185447, "image": "COCO_train2014_000000185447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle with a man standing next to it\"."}], "task": "refering", "answer_template": "The \"a motorcycle with a man standing next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 94, 95, 96, 97, 112, 113, 114, 115, 130, 131, 132, 133], "bbox": [0.21597999999999998, 0.39713855421686745, 0.44838, 0.6696686746987951], "iscrowd": 0, "area": 5277.3021, "rle": {"size": [332, 500], "counts": "jUS13U:4O2O001O0XFH\\99aFI_97_FJb97ZFLf9=0O1O00iFAc8>[GEe8;ZGFf8;WGLd84[G0b81[G2d8l0O1O1O1O1O1O001O001O0O101O0000000O01000000O10O10O0100O10O00LVNhGi1X8XNgGg1Z8ZNfGe1Z8\\NeGc1\\8_NcG`1^852L401N100O2O0O10YG`N[8OhGa1M`NX82jG^1O`NW81kG_1M`NY81jGa1K]N]82gGc1[8\\NdGc1]8^NbG_1N_Nb82^G_1f822O001O1N1000O1O010BdNkG]1T8cNmG]1S8bNmG^1T8aNmG_1S8bNkG^1W8cNgG]1[8bNdG]1^8dN`G\\1a8dN]G\\1f841N200100O0010aNXGW1n8001N101OkN[Gc0e8]O[Gc0g8[OYGd0j8ZOVGf0k8XOVGh0l8VOUGi0m8UOSGk0W900O102M3L3N3L4JnYi2"}, "label": "a motorcycle with a man standing next to it"}]}
{"id": 185447, "image": "COCO_train2014_000000185447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car behind a motercycle\"."}], "task": "refering", "answer_template": "The \"a black car behind a motercycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 95, 96, 109, 110, 111, 127, 128, 129], "bbox": [0.05232, 0.4246084337349398, 0.3494, 0.6142168674698796], "iscrowd": 0, "area": 4659.46165, "rle": {"size": [332, 500], "counts": "kb83d00m8i0O1O0O2O1O1N101O1N2O1N10000O10000O10O10O10000O100O10000O10000O01000O10000O100O10000O10001N2O1O001O0O2O001O001O001N10O1000000O10O10O1000000O10OO2J6L4000000000000O1\\OVGFj8:XGDh8;ZGDf8<[GCe8=]G@d8`0d0000O100000O100000000000000O4M3M3M3M3M3L4Mai65ZVI9F2O00O01001O001N1O2N1N3N1O2N1O2N^aY3"}, "label": "a black car behind a motercycle"}]}
{"id": 185447, "image": "COCO_train2014_000000185447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car parked next to a man with a motorbike\"."}], "task": "refering", "answer_template": "The \"a black car parked next to a man with a motorbike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 95, 96, 109, 110, 111, 127, 128, 129], "bbox": [0.05232, 0.4246084337349398, 0.3494, 0.6142168674698796], "iscrowd": 0, "area": 4659.46165, "rle": {"size": [332, 500], "counts": "kb83d00m8i0O1O0O2O1O1N101O1N2O1N10000O10000O10O10O10000O100O10000O10000O01000O10000O100O10000O10001N2O1O001O0O2O001O001O001N10O1000000O10O10O1000000O10OO2J6L4000000000000O1\\OVGFj8:XGDh8;ZGDf8<[GCe8=]G@d8`0d0000O100000O100000000000000O4M3M3M3M3M3L4Mai65ZVI9F2O00O01001O001N1O2N1N3N1O2N1O2N^aY3"}, "label": "a black car parked next to a man with a motorbike"}]}
{"id": 455791, "image": "COCO_train2014_000000455791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in black throwing a ball\"."}], "task": "refering", "answer_template": "The \"the man in black throwing a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 138, 139, 151, 152, 162, 163, 164, 165, 175, 176, 177, 178, 179, 190, 191, 192, 203, 205, 206, 216, 217, 219, 220], "bbox": [0.5193888888888889, 0.39375, 0.9420000000000001, 0.7402656249999999], "iscrowd": 0, "area": 9824.813650000002, "rle": {"size": [640, 360], "counts": "WWe33lc04M4K5K5K3N001N101O0O2O000O2O000O10O100M2O2N2M2O2N2M210O0100O1d_O_OX=b0n@B=Nc>?PAI8In<1UC>DN4En<8SC5K21Bl<`0SCLN70]Ok<f0TCG0:NYOj<j0WCC1>KUOj<n0YC_O0b0KROh<P1\\C\\O1f0HnNg<T1_CYO1i0GiNf<X1bCVO0l0EgNh<X1bCUO0Q1AdNm<V1aCVO0T1]ObNQ=V1aCSO1Y1XO`NV=T1`CTO1[1VO^NX=T1aCSOOo1`<nN`CSO0P2_<nN`CSO0Q2_<PO]CnN3T2_<UOWChN9T2`<ZOPCcN?:^Oc0R=iNkB_1OjNf0<A?o<lNkBW2c0_NH:i<POmBV2b0`NL7e<SOmBV2a0aN12`<WOoBU2?cN6N\\<ZOPCm3h0fLW<\\ORCm3k0dLT<^OQCm3P1aLo;ARCn3S1]Lk;DSCn3V1ZLi;FSCn3Y1VLf;KQCP4\\1PLe;OPCP4`1kKb;5nBP4o>nKRAQ4o>nKRAQ4o>nKRAT4l>kKUAW4X?2O1O0015K6J<D001N101O001O002N9G:F:F9G:E4M4L3M4L3M4L4WOV]O3mb0CZ]O=Uc00000O010000000000O100000000O01000000000O>C0000000O1000000000O0O2O1N2N2N2N101N2N2N2O0O2N2N2N101N2N1O2N101Nje<"}, "label": "the man in black throwing a ball"}]}
{"id": 455791, "image": "COCO_train2014_000000455791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black outfit pitching\"."}], "task": "refering", "answer_template": "The \"a man in a black outfit pitching\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 138, 139, 151, 152, 162, 163, 164, 165, 175, 176, 177, 178, 179, 190, 191, 192, 203, 205, 206, 216, 217, 219, 220], "bbox": [0.5193888888888889, 0.39375, 0.9420000000000001, 0.7402656249999999], "iscrowd": 0, "area": 9824.813650000002, "rle": {"size": [640, 360], "counts": "WWe33lc04M4K5K5K3N001N101O0O2O000O2O000O10O100M2O2N2M2O2N2M210O0100O1d_O_OX=b0n@B=Nc>?PAI8In<1UC>DN4En<8SC5K21Bl<`0SCLN70]Ok<f0TCG0:NYOj<j0WCC1>KUOj<n0YC_O0b0KROh<P1\\C\\O1f0HnNg<T1_CYO1i0GiNf<X1bCVO0l0EgNh<X1bCUO0Q1AdNm<V1aCVO0T1]ObNQ=V1aCSO1Y1XO`NV=T1`CTO1[1VO^NX=T1aCSOOo1`<nN`CSO0P2_<nN`CSO0Q2_<PO]CnN3T2_<UOWChN9T2`<ZOPCcN?:^Oc0R=iNkB_1OjNf0<A?o<lNkBW2c0_NH:i<POmBV2b0`NL7e<SOmBV2a0aN12`<WOoBU2?cN6N\\<ZOPCm3h0fLW<\\ORCm3k0dLT<^OQCm3P1aLo;ARCn3S1]Lk;DSCn3V1ZLi;FSCn3Y1VLf;KQCP4\\1PLe;OPCP4`1kKb;5nBP4o>nKRAQ4o>nKRAQ4o>nKRAT4l>kKUAW4X?2O1O0015K6J<D001N101O001O002N9G:F:F9G:E4M4L3M4L3M4L4WOV]O3mb0CZ]O=Uc00000O010000000000O100000000O01000000000O>C0000000O1000000000O0O2O1N2N2N2N101N2N2N2O0O2N2N2N101N2N1O2N101Nje<"}, "label": "a man in a black outfit pitching"}]}
{"id": 136310, "image": "COCO_train2014_000000136310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy holding up a half full wine glass\"."}], "task": "refering", "answer_template": "The \"guy holding up a half full wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 130, 150, 151, 152, 153, 154, 155, 157, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 245, 246, 247, 248, 249, 251, 266, 271], "bbox": [0.4971875, 0.16141666666666668, 0.967421875, 0.6962500000000001], "iscrowd": 0, "area": 32260.77345, "rle": {"size": [480, 640], "counts": "dUe41l>7J5K6J5K5J7J5K5K4L5K4M3M4L3M3nEjMU7Z2dHRNS7Q2gH[NP7i1iHbNo6`1lHlNk6W1nHVOi6l0RI@e6b0VIEf6>VIEh6=VIDj6>SIDk6>TIBl6`0QIBn6?PICo6>oHCQ7?lHCS7>kHCU7?iHBV7?hHCW7?fHBZ7?dHC[7>cHC]7?`HC_7>`HC_7?^HBa7`0]HBb7`0[HAf7m3O1O010O001O10O01O1O100O0001O00001O001O001L3K6K4K6K5J7K6J5J7J5J7J5K6I6K6I5L4K4M3L5PMSEg1Q;VNSEf1P;XNSEc1R;YNREc1W;TNmDj1[;fMoDX2n;N1O2N2N2N2O1N2N2O1O00001O000<D>B8H4L5K5L3L5K5K5K4L5Ka0_O00N2N2M3O2N1O1O100O1O1O1O1O100O1O2O000000000O10000000000O10000000000O2O0000000000000O10000001O001O1O001O001O001O001O001O0V1kN8G100O10000O010O100O01O001O001O001O001O@lCTNT<l1nCSNQ<m1QDQNo;o1RDQNl;P2VDoMh;R2ZDnMb;a1QDjN?E];b1UDhN?GY;b1YDeNa0IR;c1_DcNa0Jm:d1cDaNb0Kg:f1hD^Nc0Lb:f1mD]Ng0IY:k1QE[NR<f1oCYNo;i1RDUNn;k1TDTNk;m1UDSNi;o1XDPNg;P2ZDPNe;Q2[DoMc;R2_DmM`;T2`DlM_;T2bDlM];T2eDkMZ;V2g0001O001O0000000001O1O1O1O100O101O2N3SN[CX1X=O1O1O2N1O1O100O100000000000O3M3N1N1O1O2N100O1O1O1O1O1O1O00000000000001O0O2O2N2M2O2N2N2M2O2N4L3L4M3M4K2O2M2O1O2M2N2N3M`b9"}, "label": "guy holding up a half full wine glass"}]}
{"id": 136310, "image": "COCO_train2014_000000136310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man at a dinner table cheersing a glass of red wine\"."}], "task": "refering", "answer_template": "The \"a man at a dinner table cheersing a glass of red wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 130, 150, 151, 152, 153, 154, 155, 157, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 245, 246, 247, 248, 249, 251, 266, 271], "bbox": [0.4971875, 0.16141666666666668, 0.967421875, 0.6962500000000001], "iscrowd": 0, "area": 32260.77345, "rle": {"size": [480, 640], "counts": "dUe41l>7J5K6J5K5J7J5K5K4L5K4M3M4L3M3nEjMU7Z2dHRNS7Q2gH[NP7i1iHbNo6`1lHlNk6W1nHVOi6l0RI@e6b0VIEf6>VIEh6=VIDj6>SIDk6>TIBl6`0QIBn6?PICo6>oHCQ7?lHCS7>kHCU7?iHBV7?hHCW7?fHBZ7?dHC[7>cHC]7?`HC_7>`HC_7?^HBa7`0]HBb7`0[HAf7m3O1O010O001O10O01O1O100O0001O00001O001O001L3K6K4K6K5J7K6J5J7J5J7J5K6I6K6I5L4K4M3L5PMSEg1Q;VNSEf1P;XNSEc1R;YNREc1W;TNmDj1[;fMoDX2n;N1O2N2N2N2O1N2N2O1O00001O000<D>B8H4L5K5L3L5K5K5K4L5Ka0_O00N2N2M3O2N1O1O100O1O1O1O1O100O1O2O000000000O10000000000O10000000000O2O0000000000000O10000001O001O1O001O001O001O001O001O0V1kN8G100O10000O010O100O01O001O001O001O001O@lCTNT<l1nCSNQ<m1QDQNo;o1RDQNl;P2VDoMh;R2ZDnMb;a1QDjN?E];b1UDhN?GY;b1YDeNa0IR;c1_DcNa0Jm:d1cDaNb0Kg:f1hD^Nc0Lb:f1mD]Ng0IY:k1QE[NR<f1oCYNo;i1RDUNn;k1TDTNk;m1UDSNi;o1XDPNg;P2ZDPNe;Q2[DoMc;R2_DmM`;T2`DlM_;T2bDlM];T2eDkMZ;V2g0001O001O0000000001O1O1O1O100O101O2N3SN[CX1X=O1O1O2N1O1O100O100000000000O3M3N1N1O1O2N100O1O1O1O1O1O1O00000000000001O0O2O2N2M2O2N2N2M2O2N4L3L4M3M4K2O2M2O1O2M2N2N3M`b9"}, "label": "a man at a dinner table cheersing a glass of red wine"}]}
{"id": 136310, "image": "COCO_train2014_000000136310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"display of coat on the chair\"."}], "task": "refering", "answer_template": "The \"display of coat on the chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 345, 346, 368], "bbox": [0.0016875000000000002, 0.5394791666666666, 0.12134375, 0.9844166666666666], "iscrowd": 0, "area": 10490.48315, "rle": {"size": [480, 640], "counts": "gm02i8?hLDX3V2kJmMU5m3mHWLR7l500000000O100000000O10O1O1N2O0O2O1N2O1N101N2O1N2O0O2O1N2N2N1N3N2N2M3N1N3N2N2N2N1O2M3N2N2N1O2N2M3N2N1N3M3L4L4M2M4L4K5L4K4nMS2M3M3M3M2O2M3M3M3N1N3MPeW8"}, "label": "display of coat on the chair"}]}
{"id": 136310, "image": "COCO_train2014_000000136310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"coat draped over the back of the chair behind the woman in a flower dress\"."}], "task": "refering", "answer_template": "The \"coat draped over the back of the chair behind the woman in a flower dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 345, 346, 368], "bbox": [0.0016875000000000002, 0.5394791666666666, 0.12134375, 0.9844166666666666], "iscrowd": 0, "area": 10490.48315, "rle": {"size": [480, 640], "counts": "gm02i8?hLDX3V2kJmMU5m3mHWLR7l500000000O100000000O10O1O1N2O0O2O1N2O1N101N2O1N2O0O2O1N2N2N1N3N2N2M3N1N3N2N2N2N1O2M3N2N2N1O2N2M3N2N1N3M3L4L4M2M4L4K5L4K4nMS2M3M3M3M2O2M3M3M3N1N3MPeW8"}, "label": "coat draped over the back of the chair behind the woman in a flower dress"}]}
{"id": 136310, "image": "COCO_train2014_000000136310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with the floral blouse to the left of the man\"."}], "task": "refering", "answer_template": "The \"the woman with the floral blouse to the left of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.031875, 0.224125, 0.4798593749999999, 0.9981875], "iscrowd": 0, "area": 54052.90610000001, "rle": {"size": [480, 640], "counts": "nj91a>>M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2N2N2N2N2N2O1N2N2N2O1M3H8H8H8G9J6J6K5J6J6J6K5J6J6J6L4L4K5L4K5L4L4K5L4L4K5L4J6K5K5K5K5O1O010O1O100O100O1O100O1O100O10000000000000000000000000000000000000000000000000000000000nM]HgLc7W3hH`LW7`3SIWLk6i3[IRLe6n3^InKc6Q4aIlK^6U4eIgK\\6X4hIeKX6[4jIcKV6\\4nI`KS6`4PJ]Ko5c4UJYKl5g4WJVKi5i4ZJUKf5k4W2O10O10000O2O000O101N10000O2O000O10001N100O101O0O10001N100O101N1N3M3N1N3N2N2O100O100O010O10000O100O10O0100O10000O100O010O100O1O100O101N2N2O1N2N2O1N2O1N2UK[IX1h6eNZIZ1g6dN\\IY1f6eN\\IZ1e6dN^IZ1c6dN_IZ1c6dN_I[1c6bN`I\\1a6bNaI\\1a6bNbI\\1`6`NcI^1_6^NfI`1[6]NhIb1Z6ZNhIe1Z6WNjIh1X6TNjIl1W6PNlIo1V6nMmIQ2U6kMmIU2T6gMoIX2T6dMnI\\2W6]MjId2Z6UMgIl2_6mLbIT3b6fL^I\\3g6\\L[Id3j6VLWIj3o6oKQIS4m801O100O1O100O1O010O1O10O01O1QOUEiMj:T2_EfMb:S1oD]Oh0YO[:Y1QEWOl0ZOT:`1SEoNW<P1mCiNW<W1lCbNW<_1e0O10O010000O100O10000O1XOaBO_=1aBN`=2aBM_=3aBM_=4aBJ`=6`BJ`=6aBH`=8`BH`=8aBG_=9aBG_=:aBE_=;aBD`=<aBC_==aBC_==bBB^=>bBB^=>cB@^=5\\BL6N^=3`BN2O^=0cB10N]=NfB4MM^=OfB2NN\\=OgB3NM[=OhB3OMY=OjB3MNY=NkB3NMX=OkB3OMi=3c0N100OTkk4"}, "label": "the woman with the floral blouse to the left of the man"}]}
{"id": 136310, "image": "COCO_train2014_000000136310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with grey hair holding a wine glass\"."}], "task": "refering", "answer_template": "The \"a woman with grey hair holding a wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.031875, 0.224125, 0.4798593749999999, 0.9981875], "iscrowd": 0, "area": 54052.90610000001, "rle": {"size": [480, 640], "counts": "nj91a>>M3N2N2M3N2N2N2M3N2N2M3N2N2M3N2N2N2N2N2N2N2N2O1N2N2N2O1M3H8H8H8G9J6J6K5J6J6J6K5J6J6J6L4L4K5L4K5L4L4K5L4L4K5L4J6K5K5K5K5O1O010O1O100O100O1O100O1O100O10000000000000000000000000000000000000000000000000000000000nM]HgLc7W3hH`LW7`3SIWLk6i3[IRLe6n3^InKc6Q4aIlK^6U4eIgK\\6X4hIeKX6[4jIcKV6\\4nI`KS6`4PJ]Ko5c4UJYKl5g4WJVKi5i4ZJUKf5k4W2O10O10000O2O000O101N10000O2O000O10001N100O101O0O10001N100O101N1N3M3N1N3N2N2O100O100O010O10000O100O10O0100O10000O100O010O100O1O100O101N2N2O1N2N2O1N2O1N2UK[IX1h6eNZIZ1g6dN\\IY1f6eN\\IZ1e6dN^IZ1c6dN_IZ1c6dN_I[1c6bN`I\\1a6bNaI\\1a6bNbI\\1`6`NcI^1_6^NfI`1[6]NhIb1Z6ZNhIe1Z6WNjIh1X6TNjIl1W6PNlIo1V6nMmIQ2U6kMmIU2T6gMoIX2T6dMnI\\2W6]MjId2Z6UMgIl2_6mLbIT3b6fL^I\\3g6\\L[Id3j6VLWIj3o6oKQIS4m801O100O1O100O1O010O1O10O01O1QOUEiMj:T2_EfMb:S1oD]Oh0YO[:Y1QEWOl0ZOT:`1SEoNW<P1mCiNW<W1lCbNW<_1e0O10O010000O100O10000O1XOaBO_=1aBN`=2aBM_=3aBM_=4aBJ`=6`BJ`=6aBH`=8`BH`=8aBG_=9aBG_=:aBE_=;aBD`=<aBC_==aBC_==bBB^=>bBB^=>cB@^=5\\BL6N^=3`BN2O^=0cB10N]=NfB4MM^=OfB2NN\\=OgB3NM[=OhB3OMY=OjB3MNY=NkB3NMX=OkB3OMi=3c0N100OTkk4"}, "label": "a woman with grey hair holding a wine glass"}]}
{"id": 234616, "image": "COCO_train2014_000000234616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball hitter in the ground\"."}], "task": "refering", "answer_template": "The \"a baseball hitter in the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 127, 128, 129, 151, 152, 153, 175, 176, 198, 199, 221, 222, 243, 244, 245], "bbox": [0.544078125, 0.3032552693208431, 0.689359375, 0.7278454332552694], "iscrowd": 0, "area": 7421.236450000002, "rle": {"size": [427, 640], "counts": "gea42W=3N1O1N2O1O1O1M3M3N200O2O2SD_Ok:d0REAj:a0SEAk:`0TEBk:>TEDk:<SEGk:;REHU3YOW4o0bHKV3VOW4U1]HGZ3UOY4U2fKkMZ4U2eKlM[4T2aKPN_4P2_KRNb4m1]KSNe4l1ZKUNg4i1ZKWNf4i1YKXNh4f1XK[Ni4d1UK^Nl4a1SK`Nn4^1QKdNP5[1oJfNS5X1lJiNU5X1hJhNZ5X1dJiN]5X1_JjNa5W1ZJmNf5U1TJnNm5T1mIPOS6Q1hHZNFi0b7P1^HbNJ`0h7\\2SHeMn7l3001O000M4M2M3M4L4L4L4N3M2N2N2N2O1O1O1O2M3XO_FcMc9W2aFiM`9S2cFmM<]O\\8`2\\GRN0Fg8T2[GVNKHk8o1]GYNFHP9l1\\G\\NBIT9g1]G`N]OJX9c1]GcNYOK\\9^1^GgNTOKa9[1]GXOe8e0]GZOe8c0]G\\Oe8a0]G]Og8>\\GAf8<[GDg89[GEh86[GJg81]GNe8M_G0e8IaG5b8DdG:_8_OgG>e:K5L5Lafb2"}, "label": "a baseball hitter in the ground"}]}
{"id": 234616, "image": "COCO_train2014_000000234616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white\"."}], "task": "refering", "answer_template": "The \"man in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 127, 128, 129, 151, 152, 153, 175, 176, 198, 199, 221, 222, 243, 244, 245], "bbox": [0.544078125, 0.3032552693208431, 0.689359375, 0.7278454332552694], "iscrowd": 0, "area": 7421.236450000002, "rle": {"size": [427, 640], "counts": "gea42W=3N1O1N2O1O1O1M3M3N200O2O2SD_Ok:d0REAj:a0SEAk:`0TEBk:>TEDk:<SEGk:;REHU3YOW4o0bHKV3VOW4U1]HGZ3UOY4U2fKkMZ4U2eKlM[4T2aKPN_4P2_KRNb4m1]KSNe4l1ZKUNg4i1ZKWNf4i1YKXNh4f1XK[Ni4d1UK^Nl4a1SK`Nn4^1QKdNP5[1oJfNS5X1lJiNU5X1hJhNZ5X1dJiN]5X1_JjNa5W1ZJmNf5U1TJnNm5T1mIPOS6Q1hHZNFi0b7P1^HbNJ`0h7\\2SHeMn7l3001O000M4M2M3M4L4L4L4N3M2N2N2N2O1O1O1O2M3XO_FcMc9W2aFiM`9S2cFmM<]O\\8`2\\GRN0Fg8T2[GVNKHk8o1]GYNFHP9l1\\G\\NBIT9g1]G`N]OJX9c1]GcNYOK\\9^1^GgNTOKa9[1]GXOe8e0]GZOe8c0]G\\Oe8a0]G]Og8>\\GAf8<[GDg89[GEh86[GJg81]GNe8M_G0e8IaG5b8DdG:_8_OgG>e:K5L5Lafb2"}, "label": "man in white"}]}
{"id": 234616, "image": "COCO_train2014_000000234616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man behind the striker\"."}], "task": "refering", "answer_template": "The \"a man behind the striker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 129, 130, 153, 154, 176, 177, 199, 200, 222, 223, 246], "bbox": [0.6017187500000001, 0.305152224824356, 0.7375, 0.7013348946135832], "iscrowd": 0, "area": 3736.733300000001, "rle": {"size": [427, 640], "counts": "ceP51Z=4K3N0O100O1O100O100O11O01O1O10O01O100O001O11O10001N107IO1N2N2O1N1O2O1N2N101N2N101N3N1N3N1N3jDlNP:U1kESOP:P1jEWOS:j0gE^OU:e0eEBX:?cEGZ:i1M2M3N1N3N2N1O2N2N2NO1O1N2O1\\MRFh1=bMR9c0eFi17gMT9>gFd0HN<ROV99jF?J64TOY95kF9Oe0DoNc90mF70V1T9aNoF5N\\1S9\\NRHc1o7ZNVHc1l9M3M4L2N3M2N2N2N2M4M2N>BhhU2"}, "label": "a man behind the striker"}]}
{"id": 234616, "image": "COCO_train2014_000000234616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire\"."}], "task": "refering", "answer_template": "The \"the umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 129, 130, 153, 154, 176, 177, 199, 200, 222, 223, 246], "bbox": [0.6017187500000001, 0.305152224824356, 0.7375, 0.7013348946135832], "iscrowd": 0, "area": 3736.733300000001, "rle": {"size": [427, 640], "counts": "ceP51Z=4K3N0O100O1O100O100O11O01O1O10O01O100O001O11O10001N107IO1N2N2O1N1O2O1N2N101N2N101N3N1N3N1N3jDlNP:U1kESOP:P1jEWOS:j0gE^OU:e0eEBX:?cEGZ:i1M2M3N1N3N2N1O2N2N2NO1O1N2O1\\MRFh1=bMR9c0eFi17gMT9>gFd0HN<ROV99jF?J64TOY95kF9Oe0DoNc90mF70V1T9aNoF5N\\1S9\\NRHc1o7ZNVHc1l9M3M4L2N3M2N2N2N2M4M2N>BhhU2"}, "label": "the umpire"}]}
{"id": 160893, "image": "COCO_train2014_000000160893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smiling man\"."}], "task": "refering", "answer_template": "The \"the smiling man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.2650625, 0.15405152224824356, 0.68225, 0.9892740046838406], "iscrowd": 0, "area": 65200.80225000001, "rle": {"size": [427, 640], "counts": "TVW2b0f<c0\\O6K3M3L4M3L4M3L4L5L3N2N2M3N2N2M4M2N2M3N2N2N2N2M4M2N2N2N2M3N2N3M2aLVL[Ml3c2YLXMi3g2[LTMg3j2]LRMe3l2`LoLb3o2cLlL_3S3eLhL]3V3hLfLY3X3kLdLX3Y3mLbLU3]3oL^LS3`3RM\\Lo2b3VMYLl2f3YMULh2i3_MPLd2n3bMkK`2T4gMbK]2]4jMZKZ2i1lJe0R3YMU2o1lJg0[8UOhGj0Z8QOjGm0Y8oNjGP1X8gNPHX1R8_NVH_1l9O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10O1000001O0000001O0000001O0000001O0000001O0000001O001O1O1O001O1O1O1O001O1O1O001O1O1O2M3N1O2N2N2N3M2O2M3M2N2N1O2N2N1cG\\LQ7e3gHbLY7_3_HiL`7W3YHPMg7Q3PHWMP8R401O000000001O00000001O0001O0000000000001O0000000000001O000000001O000000001N100000001O0000000O2O00000000001O001N2O001O1O1O001O1N2O001O1O1O001O1O1N101O1O1O001O1O1O1O100O100O1O101N1O1iNdJ]J]5^5[1J6J6M3N2M4M2M3M3N2M3N2M3N3J7G8H9H7I7@a0J5J7J5K6IY]d2"}, "label": "the smiling man"}]}
{"id": 160893, "image": "COCO_train2014_000000160893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy with a black t - shirt\"."}], "task": "refering", "answer_template": "The \"the guy with a black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.2650625, 0.15405152224824356, 0.68225, 0.9892740046838406], "iscrowd": 0, "area": 65200.80225000001, "rle": {"size": [427, 640], "counts": "TVW2b0f<c0\\O6K3M3L4M3L4M3L4L5L3N2N2M3N2N2M4M2N2M3N2N2N2N2M4M2N2N2N2M3N2N3M2aLVL[Ml3c2YLXMi3g2[LTMg3j2]LRMe3l2`LoLb3o2cLlL_3S3eLhL]3V3hLfLY3X3kLdLX3Y3mLbLU3]3oL^LS3`3RM\\Lo2b3VMYLl2f3YMULh2i3_MPLd2n3bMkK`2T4gMbK]2]4jMZKZ2i1lJe0R3YMU2o1lJg0[8UOhGj0Z8QOjGm0Y8oNjGP1X8gNPHX1R8_NVH_1l9O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10O1000001O0000001O0000001O0000001O0000001O0000001O001O1O1O001O1O1O1O001O1O1O001O1O1O2M3N1O2N2N2N3M2O2M3M2N2N1O2N2N1cG\\LQ7e3gHbLY7_3_HiL`7W3YHPMg7Q3PHWMP8R401O000000001O00000001O0001O0000000000001O0000000000001O000000001O000000001N100000001O0000000O2O00000000001O001N2O001O1O1O001O1N2O001O1O1O001O1O1N101O1O1O001O1O1O1O100O100O1O101N1O1iNdJ]J]5^5[1J6J6M3N2M4M2M3M3N2M3N2M3N3J7G8H9H7I7@a0J5J7J5K6IY]d2"}, "label": "the guy with a black t - shirt"}]}
{"id": 193663, "image": "COCO_train2014_000000193663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soft ball player standing idle\"."}], "task": "refering", "answer_template": "The \"a soft ball player standing idle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 105, 106, 107, 123, 124, 125, 140, 141, 142, 143, 159, 160, 161, 178, 179, 197, 215, 233, 251, 269, 287], "bbox": [0.8063671875, 0.19629687499999998, 1.0, 0.6686875], "iscrowd": 0, "area": 9429.034700000002, "rle": {"size": [640, 512], "counts": "_ZR81nc01N2O1O1O1O1N2O1O1O1O1O1N2O1000000O100000000O10000000000O100S]O\\O^b0d0b]O]Oha0O^^Od0J^Ofa0O`^Oc0J_Oda0Ob^Ob0J@ba0Nd^Oc0J@`a0Nf^Ob0JA^a0Nh^Oa0JB\\a0Nj^O`0JCZa0Nl^O?IEZa0Lm^O?IFXa0Ln^O?JFVa0LP_O>JHSa0KS_O<KJPa0JV_O<JKn`0JX_O;JLm`0IX_O<KLk`0IZ_O;KLj`0J[_O:KMi`0I\\_O:KNg`0I^_O9KNf`0J__O8KOe`0Ha_O9J0c`0Hc_O8J0b`0Id_O7J1a`0He_O7J2_`0Hg_O6J2^`0Ih_O5J3]`0Hh_O6K3[`0Hj_O5K3Z`0Ik_O4K4Y`0Gm_O5J7U`0EQ@4J8S`0ES@3J8S`0ES@3J9Q`0EU@2J:o?EW@1J;n?DX@1J<l?EY@OK<k?GY@ML=j?GY@LM>h?HZ@JN?g?IW@J2>0nNQ?l0k@I4>LRO\\3Mi7k0hDJ6=IWOi0I\\10Q9k0iDI8_11ZNe1EV9m1TE50iNV1WOd9k1VE5OVOj0mNn9i1YE4O@1TOe:X1[E4OQ1c:lN^E3OS1a:jN`E3OU1^:iNcE2OV1]:hNdE2OV1\\:iNeE1OV1\\:iNeE1OV1[:jNfE0OV1[:jNfE0OW1Y:jNhEOOW1Y:jNhEOOW1Y:jNhE0NV1Y:kNiEONV1Y:kNiEONV1Y:kNiEONV1Y:kNiEONV1X:lNjENNW1W:kNkENNW1W:kNkENNW1W:kNkENN=S7"}, "label": "a soft ball player standing idle"}]}
{"id": 193663, "image": "COCO_train2014_000000193663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player in blue on the outside of the fence\"."}], "task": "refering", "answer_template": "The \"player in blue on the outside of the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 105, 106, 107, 123, 124, 125, 140, 141, 142, 143, 159, 160, 161, 178, 179, 197, 215, 233, 251, 269, 287], "bbox": [0.8063671875, 0.19629687499999998, 1.0, 0.6686875], "iscrowd": 0, "area": 9429.034700000002, "rle": {"size": [640, 512], "counts": "_ZR81nc01N2O1O1O1O1N2O1O1O1O1O1N2O1000000O100000000O10000000000O100S]O\\O^b0d0b]O]Oha0O^^Od0J^Ofa0O`^Oc0J_Oda0Ob^Ob0J@ba0Nd^Oc0J@`a0Nf^Ob0JA^a0Nh^Oa0JB\\a0Nj^O`0JCZa0Nl^O?IEZa0Lm^O?IFXa0Ln^O?JFVa0LP_O>JHSa0KS_O<KJPa0JV_O<JKn`0JX_O;JLm`0IX_O<KLk`0IZ_O;KLj`0J[_O:KMi`0I\\_O:KNg`0I^_O9KNf`0J__O8KOe`0Ha_O9J0c`0Hc_O8J0b`0Id_O7J1a`0He_O7J2_`0Hg_O6J2^`0Ih_O5J3]`0Hh_O6K3[`0Hj_O5K3Z`0Ik_O4K4Y`0Gm_O5J7U`0EQ@4J8S`0ES@3J8S`0ES@3J9Q`0EU@2J:o?EW@1J;n?DX@1J<l?EY@OK<k?GY@ML=j?GY@LM>h?HZ@JN?g?IW@J2>0nNQ?l0k@I4>LRO\\3Mi7k0hDJ6=IWOi0I\\10Q9k0iDI8_11ZNe1EV9m1TE50iNV1WOd9k1VE5OVOj0mNn9i1YE4O@1TOe:X1[E4OQ1c:lN^E3OS1a:jN`E3OU1^:iNcE2OV1]:hNdE2OV1\\:iNeE1OV1\\:iNeE1OV1[:jNfE0OV1[:jNfE0OW1Y:jNhEOOW1Y:jNhEOOW1Y:jNhE0NV1Y:kNiEONV1Y:kNiEONV1Y:kNiEONV1Y:kNiEONV1X:lNjENNW1W:kNkENNW1W:kNkENNW1W:kNkENN=S7"}, "label": "player in blue on the outside of the fence"}]}
{"id": 193663, "image": "COCO_train2014_000000193663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man at bat\"."}], "task": "refering", "answer_template": "The \"the man at bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 39, 40, 41, 57, 58, 59, 60, 75, 76, 77, 92, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 295, 296, 297, 298, 307, 308, 309, 310, 314, 315, 316, 317, 325, 326, 327, 328, 332, 333, 334, 335, 342, 343, 344, 345, 351, 352, 353, 360, 361, 362, 370, 371], "bbox": [0.0259765625, 0.07748437500000001, 0.668828125, 0.9112500000000001], "iscrowd": 0, "area": 70767.02719999998, "rle": {"size": [640, 512], "counts": "fe81lc05J7J5J7J5K6I5L3N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N200O1O1O1O1O1O100OO1O1O1O1O1O1O2N100000000000001O000000000001O0000000000000SGiM2W2@[N<e1WOlNe0T1oN\\Om0d0SOAh0`0^MfMRLm1\\6=[MSNTLc1]6;WM^NWL[1^67UMiNYLS1^65YMlNTLS1_61^MPOnKS1`6NaMTOjKR1a6JfMXOeKR1`6GkM\\O_KQ1b6CPN@YKQ1c6_OTNETKP1d6\\OXNHPKo0d6YO]NV3`MiJk0R2e1X3mL\\KW1\\1l1\\3gL`KW1U1R2_3`LeKV1m0[2a3XLlKU1e0c2b3RLRLS1=k2e3lKWLQ16S3f3eK^LP1M\\3h3^KcLn0Hd3h3XKiLl0@l3k3QKoLk0XOT4m7eKTH\\4R8\\KPHc4W8UKjGl4[8mJgGS5a;2N1O2N2M3N1O2N2N2N1O2N2N2N1N3N2N2N1O2O1N2O1N3N1N2O2M2O2M2O2M2O2M2O2M2O2M2O2M2O1N3N1N3N1N3N1N3N1N3N6I8I7H5L2M2O2M2O2PFcGY8_8dGdG[8^8bGdG\\8_8aGcG^8^8`GeG]8^8`GdG_8]8^GfG`8]8]GeGb8\\8\\GgGa8\\8\\GfGc8m9M2O2M3O0100O2O000O101N11O2N100O1O2O0O1O1O10N1O1O1O2N100O2N1O1O1O2NdIkDn3T;iKYET4f:iKaET4^:kKgES4W:mKmEP4R:oKSFn3l9QLYFm3e9SL^Fk3a9TLdFi3[9VLjFg3U9YLnFf3Q9XLRGg3m8YLVGe3i8[LYGd3f8[L^Gd3`8\\LbGc3]8]LfGa3Y8_LiGa3U8_LnG_3Q8aLQH^3n7bLUH\\3j7dLYH[3e7dL^H[3a7eLbHY3]7gLeHX3Z7hLiHW3U7iLmHW3Q7iLRIU3l6lLVIT3h6lL[IS3c6mL_IR3`6mLdIR3Z6nLhIR3V6nLlIR3R6nLPJQ3o5oLSJQ3k5oLXJP3f5PM\\Jo2c5QM_Jo2_5PMdJP3Z5PMhJP3V5PMlJo2S5QMoJo2o4QMTKn2i4SMYKl2f4TM\\Kl2b4TM`Kl2^4SMeKm2Y4SMiKl2V4TMlKl2R4TMPLl2n3TMULj2j3VMXLj2f3VM\\Lj2b3VM`Lj2^3VMdLi2[3VMhLj2V3VMlLj2R3VMQMh2n2XMTMh2i2YMYMf2f2ZM\\Me2c2[M_Md2`2\\MbMc2]2\\MfMc2Y2]MjMa2U2_MmM`2R2`MPN_2o1aMSN]2m1cMUN\\2j1dMXN[2g1eM[NZ2d1fM^NY2a1gMbNW2]1iMeNV2Z1jMhNU2X1jMjNV2T1jMnNU2R1jMPOV2n0jMTOU2l0jMWOU2h0jMZOV2d0jM^OU2b0jM@V2?iMCW2;iMGV29iMIW26hMLX22hM0W20hM2V2OiM3U2MkM6S2JlM8R2ImM9Q2GoM;P2EoM=o1DoM?o1APNb0o1^OPNd0n1\\OQNg0m1ZOQNi0n1WOQNk0m1UORNo0k1ROSNQ1l1oNSNS1k1mNTNV1j1kNTNX1k1hNTNZ1j1fNUN]1i1eNTN^1j1cNUN_1j1aNTNb1j1_NTNd1j1]NUNe1j1[NTNi1i1YNTNj1j1WNUNk1h1WNVNl1g1VNWNm1g1TNXNP2c1[8K4L5L4K4L5L4K4L5L4K4LYfY3"}, "label": "the man at bat"}]}
{"id": 193663, "image": "COCO_train2014_000000193663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging a baseball bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 39, 40, 41, 57, 58, 59, 60, 75, 76, 77, 92, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 295, 296, 297, 298, 307, 308, 309, 310, 314, 315, 316, 317, 325, 326, 327, 328, 332, 333, 334, 335, 342, 343, 344, 345, 351, 352, 353, 360, 361, 362, 370, 371], "bbox": [0.0259765625, 0.07748437500000001, 0.668828125, 0.9112500000000001], "iscrowd": 0, "area": 70767.02719999998, "rle": {"size": [640, 512], "counts": "fe81lc05J7J5J7J5K6I5L3N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N200O1O1O1O1O1O100OO1O1O1O1O1O1O2N100000000000001O000000000001O0000000000000SGiM2W2@[N<e1WOlNe0T1oN\\Om0d0SOAh0`0^MfMRLm1\\6=[MSNTLc1]6;WM^NWL[1^67UMiNYLS1^65YMlNTLS1_61^MPOnKS1`6NaMTOjKR1a6JfMXOeKR1`6GkM\\O_KQ1b6CPN@YKQ1c6_OTNETKP1d6\\OXNHPKo0d6YO]NV3`MiJk0R2e1X3mL\\KW1\\1l1\\3gL`KW1U1R2_3`LeKV1m0[2a3XLlKU1e0c2b3RLRLS1=k2e3lKWLQ16S3f3eK^LP1M\\3h3^KcLn0Hd3h3XKiLl0@l3k3QKoLk0XOT4m7eKTH\\4R8\\KPHc4W8UKjGl4[8mJgGS5a;2N1O2N2M3N1O2N2N2N1O2N2N2N1N3N2N2N1O2O1N2O1N3N1N2O2M2O2M2O2M2O2M2O2M2O2M2O2M2O1N3N1N3N1N3N1N3N1N3N6I8I7H5L2M2O2M2O2PFcGY8_8dGdG[8^8bGdG\\8_8aGcG^8^8`GeG]8^8`GdG_8]8^GfG`8]8]GeGb8\\8\\GgGa8\\8\\GfGc8m9M2O2M3O0100O2O000O101N11O2N100O1O2O0O1O1O10N1O1O1O2N100O2N1O1O1O2NdIkDn3T;iKYET4f:iKaET4^:kKgES4W:mKmEP4R:oKSFn3l9QLYFm3e9SL^Fk3a9TLdFi3[9VLjFg3U9YLnFf3Q9XLRGg3m8YLVGe3i8[LYGd3f8[L^Gd3`8\\LbGc3]8]LfGa3Y8_LiGa3U8_LnG_3Q8aLQH^3n7bLUH\\3j7dLYH[3e7dL^H[3a7eLbHY3]7gLeHX3Z7hLiHW3U7iLmHW3Q7iLRIU3l6lLVIT3h6lL[IS3c6mL_IR3`6mLdIR3Z6nLhIR3V6nLlIR3R6nLPJQ3o5oLSJQ3k5oLXJP3f5PM\\Jo2c5QM_Jo2_5PMdJP3Z5PMhJP3V5PMlJo2S5QMoJo2o4QMTKn2i4SMYKl2f4TM\\Kl2b4TM`Kl2^4SMeKm2Y4SMiKl2V4TMlKl2R4TMPLl2n3TMULj2j3VMXLj2f3VM\\Lj2b3VM`Lj2^3VMdLi2[3VMhLj2V3VMlLj2R3VMQMh2n2XMTMh2i2YMYMf2f2ZM\\Me2c2[M_Md2`2\\MbMc2]2\\MfMc2Y2]MjMa2U2_MmM`2R2`MPN_2o1aMSN]2m1cMUN\\2j1dMXN[2g1eM[NZ2d1fM^NY2a1gMbNW2]1iMeNV2Z1jMhNU2X1jMjNV2T1jMnNU2R1jMPOV2n0jMTOU2l0jMWOU2h0jMZOV2d0jM^OU2b0jM@V2?iMCW2;iMGV29iMIW26hMLX22hM0W20hM2V2OiM3U2MkM6S2JlM8R2ImM9Q2GoM;P2EoM=o1DoM?o1APNb0o1^OPNd0n1\\OQNg0m1ZOQNi0n1WOQNk0m1UORNo0k1ROSNQ1l1oNSNS1k1mNTNV1j1kNTNX1k1hNTNZ1j1fNUN]1i1eNTN^1j1cNUN_1j1aNTNb1j1_NTNd1j1]NUNe1j1[NTNi1i1YNTNj1j1WNUNk1h1WNVNl1g1VNWNm1g1TNXNP2c1[8K4L5L4K4L5L4K4L5L4K4LYfY3"}, "label": "a baseball player swinging a baseball bat"}]}
{"id": 185474, "image": "COCO_train2014_000000185474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a circular bowl just behind the flowers\"."}], "task": "refering", "answer_template": "The \"a circular bowl just behind the flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 146, 148, 149], "bbox": [0.09818, 0.5454666666666667, 0.35064, 0.8207666666666666], "iscrowd": 0, "area": 5437.553850000001, "rle": {"size": [300, 500], "counts": "i`>6U9=C=D9F2N2O1N2O1N2N2O1N2N101O1O1O1O1O1O1O100O1O1O1OO1O1O1O2M2O1O1O1O1O1N2O1O2N1O1000000000001O0000000001OK5H]HPO[7f0iH_OW79lHMS7JRI9o6_OTIc0g73M3M2O2M3mHPOk5P1SJUOn5h0QJZOP6d0eIUOC8h6`0fIXOB:h6<fI[OB:h69fI^OA:j65fIAA;i62eIEB:i6OeIHA:k6KeIKA;k6GdIO@<l6CdI2@;m6@dI5_O=m6\\OdI8_O=g7DXH=g7DYH=e7C[H?b7C^H=a7D_H=_7CaH?]7BcH?[7BdH?[7AfH`0X7AgHa0W7@iH`0Y7]OgHe0Y7ZOgHg0Y7XOfHi0[7UOfHl0[7ROeHo0[7POdHQ1]7mNdHT1\\7kNcHW1]7hNcHX1_7eNaH]1f71N2O0011N1O1N3N1O2M2O2M2O2N1N2O2N1N4M2N2M3N2M3N2N3L3NdPo2"}, "label": "a circular bowl just behind the flowers"}]}
{"id": 185474, "image": "COCO_train2014_000000185474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sink behind some green leafy plant\"."}], "task": "refering", "answer_template": "The \"a white sink behind some green leafy plant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 146, 148, 149], "bbox": [0.09818, 0.5454666666666667, 0.35064, 0.8207666666666666], "iscrowd": 0, "area": 5437.553850000001, "rle": {"size": [300, 500], "counts": "i`>6U9=C=D9F2N2O1N2O1N2N2O1N2N101O1O1O1O1O1O1O100O1O1O1OO1O1O1O2M2O1O1O1O1O1N2O1O2N1O1000000000001O0000000001OK5H]HPO[7f0iH_OW79lHMS7JRI9o6_OTIc0g73M3M2O2M3mHPOk5P1SJUOn5h0QJZOP6d0eIUOC8h6`0fIXOB:h6<fI[OB:h69fI^OA:j65fIAA;i62eIEB:i6OeIHA:k6KeIKA;k6GdIO@<l6CdI2@;m6@dI5_O=m6\\OdI8_O=g7DXH=g7DYH=e7C[H?b7C^H=a7D_H=_7CaH?]7BcH?[7BdH?[7AfH`0X7AgHa0W7@iH`0Y7]OgHe0Y7ZOgHg0Y7XOfHi0[7UOfHl0[7ROeHo0[7POdHQ1]7mNdHT1\\7kNcHW1]7hNcHX1_7eNaH]1f71N2O0011N1O1N3N1O2M2O2M2O2N1N2O2N1N4M2N2M3N2M3N2N3L3NdPo2"}, "label": "a white sink behind some green leafy plant"}]}
{"id": 185474, "image": "COCO_train2014_000000185474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bath tub in the middle of the room that is shaped like a bowl\"."}], "task": "refering", "answer_template": "The \"a white bath tub in the middle of the room that is shaped like a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 133, 134, 135, 136, 137], "bbox": [0.37376, 0.4298, 0.69778, 0.7001666666666667], "iscrowd": 0, "area": 9319.028250000001, "rle": {"size": [300, 500], "counts": "kmf12k81`G3[81cG1Y81hG0U81kG2R8MoG5o7KQH7m7HTH:j7FWH<e7E[H=c7B^H`0`7@aHa0]7_OcHc0Z7]OhHc0W7]OiHd0V7\\OkHd0T7[OmHe0S7[OnHe0Q7\\OnHe0Q7\\OoHd0P7]OoHc0Q7^OoHb0P7_OoHb0P7_OPIa0o6@QI`0n6AQI?o6BQI>o6DmH>R7EjH=U7IdH8\\7m0O1O1O001O1O1O001O00001O00001O01O0001O000010O000001O00010O0000001O01O00000001O0001O00000001O01O00000000010O00000000001O000000001O000000001O00000O2O000000001O000000001O00000000001O000000001O0O10000O2O0O2O0O2O0O1O2O0O2N100O2N100O2N100O2N101N1N3M2N3M2O2M3M2N3M3M2L6IgR\\1"}, "label": "a white bath tub in the middle of the room that is shaped like a bowl"}]}
{"id": 185474, "image": "COCO_train2014_000000185474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fancy white bath tub\"."}], "task": "refering", "answer_template": "The \"a fancy white bath tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 133, 134, 135, 136, 137], "bbox": [0.37376, 0.4298, 0.69778, 0.7001666666666667], "iscrowd": 0, "area": 9319.028250000001, "rle": {"size": [300, 500], "counts": "kmf12k81`G3[81cG1Y81hG0U81kG2R8MoG5o7KQH7m7HTH:j7FWH<e7E[H=c7B^H`0`7@aHa0]7_OcHc0Z7]OhHc0W7]OiHd0V7\\OkHd0T7[OmHe0S7[OnHe0Q7\\OnHe0Q7\\OoHd0P7]OoHc0Q7^OoHb0P7_OoHb0P7_OPIa0o6@QI`0n6AQI?o6BQI>o6DmH>R7EjH=U7IdH8\\7m0O1O1O001O1O1O001O00001O00001O01O0001O000010O000001O00010O0000001O01O00000001O0001O00000001O01O00000000010O00000000001O000000001O000000001O00000O2O000000001O000000001O00000000001O000000001O0O10000O2O0O2O0O2O0O1O2O0O2N100O2N100O2N100O2N101N1N3M2N3M2O2M3M2N3M3M2L6IgR\\1"}, "label": "a fancy white bath tub"}]}
{"id": 185474, "image": "COCO_train2014_000000185474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bathroom sink next to flowers in vase\"."}], "task": "refering", "answer_template": "The \"bathroom sink next to flowers in vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 146, 148, 149], "bbox": [0.09662000000000001, 0.5550666666666667, 0.34753999999999996, 0.82], "iscrowd": 0, "area": 6088.026849999999, "rle": {"size": [300, 500], "counts": "dW>1Y9?B=C4L3M2M4M3M2O2N2N1N3N2N1N2O1O001O1O001O1N101O1OO1O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2000001O0002N5K4L4L4LN2N2N2N3N1N2O10bHfNk6Y1oHlNR7T1hHQOY7o0aHVO?A`6i1_IXNb6g1^IZNc6e1[I\\Nf6c1ZI]Ng6b1XI`Nh6_1XIaNj6]1UIeNk6Z1TIgNm6X1SIiNm6V1RIkNP7T1oHmNQ7R1nHoNS7P1mHQOS7n0lHSOU7l0kHUOU7j0jHWOW7i0hHXOX7g0gHZOZ7f0eH[O[7d0dH]O^7X10O1O2O0O2O0O101N101N101N100O2O0O2O0O1O2O0O2O0O2O0O102M3N1N3N1N3N2M2L5L3M4K4MoYo2"}, "label": "bathroom sink next to flowers in vase"}]}
{"id": 185474, "image": "COCO_train2014_000000185474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sink on left side\"."}], "task": "refering", "answer_template": "The \"sink on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 146, 148, 149], "bbox": [0.09662000000000001, 0.5550666666666667, 0.34753999999999996, 0.82], "iscrowd": 0, "area": 6088.026849999999, "rle": {"size": [300, 500], "counts": "dW>1Y9?B=C4L3M2M4M3M2O2N2N1N3N2N1N2O1O001O1O001O1N101O1OO1O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2000001O0002N5K4L4L4LN2N2N2N3N1N2O10bHfNk6Y1oHlNR7T1hHQOY7o0aHVO?A`6i1_IXNb6g1^IZNc6e1[I\\Nf6c1ZI]Ng6b1XI`Nh6_1XIaNj6]1UIeNk6Z1TIgNm6X1SIiNm6V1RIkNP7T1oHmNQ7R1nHoNS7P1mHQOS7n0lHSOU7l0kHUOU7j0jHWOW7i0hHXOX7g0gHZOZ7f0eH[O[7d0dH]O^7X10O1O2O0O2O0O101N101N101N100O2O0O2O0O1O2O0O2O0O2O0O102M3N1N3N1N3N2M2L5L3M4K4MoYo2"}, "label": "sink on left side"}]}
{"id": 144519, "image": "COCO_train2014_000000144519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shaggy sheep sticking its head out of the fence to eat some grass\"."}], "task": "refering", "answer_template": "The \"a shaggy sheep sticking its head out of the fence to eat some grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 264, 265, 266, 267, 268, 285, 287, 289, 291], "bbox": [0.305484375, 0.3664851485148515, 0.74884375, 0.9082178217821782], "iscrowd": 0, "area": 36311.863600000026, "rle": {"size": [404, 640], "counts": "k`]27[<9H6I5K6K4K6K5J6K6I7I6K6I7J5J4M3M3M3M4M2M3M3M3M3M3M3M3N3L4L4L3N3L3M2N2O1N2N2N2O1N2O0O2O1N2O0O2O1N2O3L4M4K5L3L3N3LoNgIkKW6U4kIjKS6V4PJhKP6X4SJfKk5[4WJdKh5\\4[JbKd5^4^J`Kb5`4`J_K_5a4cJ^K\\5a4gJ^KX5b4jJ\\KV5d4lJ[KS5e4oJZKP5f4RKXKo4g4SKUKo4k4WKmJk4S5[1O1000000000000O100000000000001N1000000O1000001O00001O00001O00001O010O01O0`ITKY5k4bJZK^5g4\\J^Kd5b4WJcKj5`4lIfKT6]50N3M2N2N3M2N3M2N2O2M2J7H7O2O000O2O001O0O2O00001O0O2O001O001O0010O0001O0010O01O001O01O100O100O2O0O100O10:E:F5LO001O001N101O1O0O2O001O0O2O1O001O0O2O1O0O2O0O2O1O0O2O0O2O1O0O2O0O2O1N101N3N3L3M3N2M4M2N2O1N2N3M2O1N201N2O0O2O1N101N3N2M4N1NO00O1O1O100O1O1O010O1O1O010O1L3E<^Ob0J6K5J6J6L4N2N2N2N2O1O001O1O10O01O1O1O001O1O1O001O100O1O1O001O1O1Oaco1"}, "label": "a shaggy sheep sticking its head out of the fence to eat some grass"}]}
{"id": 144519, "image": "COCO_train2014_000000144519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep closest to the fence\"."}], "task": "refering", "answer_template": "The \"sheep closest to the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 264, 265, 266, 267, 268, 285, 287, 289, 291], "bbox": [0.305484375, 0.3664851485148515, 0.74884375, 0.9082178217821782], "iscrowd": 0, "area": 36311.863600000026, "rle": {"size": [404, 640], "counts": "k`]27[<9H6I5K6K4K6K5J6K6I7I6K6I7J5J4M3M3M3M4M2M3M3M3M3M3M3M3N3L4L4L3N3L3M2N2O1N2N2N2O1N2O0O2O1N2O0O2O1N2O3L4M4K5L3L3N3LoNgIkKW6U4kIjKS6V4PJhKP6X4SJfKk5[4WJdKh5\\4[JbKd5^4^J`Kb5`4`J_K_5a4cJ^K\\5a4gJ^KX5b4jJ\\KV5d4lJ[KS5e4oJZKP5f4RKXKo4g4SKUKo4k4WKmJk4S5[1O1000000000000O100000000000001N1000000O1000001O00001O00001O00001O010O01O0`ITKY5k4bJZK^5g4\\J^Kd5b4WJcKj5`4lIfKT6]50N3M2N2N3M2N3M2N2O2M2J7H7O2O000O2O001O0O2O00001O0O2O001O001O0010O0001O0010O01O001O01O100O100O2O0O100O10:E:F5LO001O001N101O1O0O2O001O0O2O1O001O0O2O1O0O2O0O2O1O0O2O0O2O1O0O2O0O2O1N101N3N3L3M3N2M4M2N2O1N2N3M2O1N201N2O0O2O1N101N3N2M4N1NO00O1O1O100O1O1O010O1O1O010O1L3E<^Ob0J6K5J6J6L4N2N2N2N2O1O001O1O10O01O1O1O001O1O1O001O100O1O1O001O1O1Oaco1"}, "label": "sheep closest to the fence"}]}
{"id": 144519, "image": "COCO_train2014_000000144519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second sheep from the right\"."}], "task": "refering", "answer_template": "The \"second sheep from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 202, 203, 204, 205, 206, 222, 223, 225, 226, 228], "bbox": [0.565015625, 0.24227722772277227, 0.96646875, 0.6876732673267326], "iscrowd": 0, "area": 26471.015750000006, "rle": {"size": [404, 640], "counts": "To^4;V<:F4L4L3M4L4L4L4K5L3M3M4L3M3M4M200O2O000O101O000O2O000O2O1O1N101O1N2O001O1N2O1O2M2O2N2M2O2N1N3N2N1O2N1O2N1O2N2N1O2O3L5K4L5K5K4L3M4L3M2N10POdHgL[7W3iHhLV7U3nHkLQ7S3RImLm6Q3WInLh6P3[IPMd6o2_IPM`6P3aIPM^6o2eIPMZ6o2hIQMW6o2jIQMV6m2mIRMR6m2PJSMo5m2SJQMm5n2UJQMk5o2WJnLj5R3XJlLh5T3]JgLc5Y3bJbL^5^3fJ^LZ5b3m1O00000O2O0000001O0000001O00000O2O0O100O2O0O100O2O0O100O2O0O10001O01O00010O000010O0000010O0001O:GT1kN7I7JN1O1O2M2M3M3^Oc0K4L4M4L3O1O2N1O1O1O20O2N2O0O2N2O1N1O2N101N1O2O0O2O0O100O100O010O10000O100O100O100O01YO_GZMa8d2aG\\M^8c2dG]M\\8a2fG]M[8d2eGZM]8e2dGYM]8h2cGVM_8i2bGUM`8k2`GRMb8n2_GmLf8R3>01O1O001O101N2N2N2N2N2N2N2O1N2N1O2N1O1O2N100101N100O1O100O1O1O1O00O1O1O1O001O1O1O1O1O2N2N2N2N2N2N2L5K9G8H9G8H?A[Q8"}, "label": "second sheep from the right"}]}
{"id": 144519, "image": "COCO_train2014_000000144519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep going last and the head is covered behind another sheep\"."}], "task": "refering", "answer_template": "The \"sheep going last and the head is covered behind another sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 231, 233, 234, 235], "bbox": [0.038015625, 0.37205445544554455, 0.33667187499999995, 0.7827970297029703], "iscrowd": 0, "area": 22233.208700000003, "rle": {"size": [404, 640], "counts": "oe91]<9F:F:F:F:G9RFYNb8Q2RGlNU8\\1_GoN^8f2M3L3N3EhKUH[4h7gKVH\\4g7fKWH]4e7<M3L21O0000001O00000AcHgK]7U4iHiKX7R4mHmKS7o3RIPLn6l3WISLi6i3]IULd6f3aIYL_6c3fI\\LZ6`3_1N21O1O1O1O0O2O1O1O1O1O002N1O2N1O2N1O2N1O2N1O2N1O001O1O1O001O1O1O0O2O1O1O001O1OSOUHRMj7l2YHSMg7k2\\HUMc7j2_HVM`7h2cHXM\\7g2fHXMZ7h2gHXMX7h2iHXMV7h2lHWMS7i2nHVMR7j2oHVMP7j2QIVMn6j2SIUMm6k2TIUMl6j2UIVMj6j2XITMh6l2YITMf6l2[ISMe6m2\\IQMe6o2\\IPMd6P3^ImLc6S3^IlLb6T3_IjLb6V3_IiLb6V3`IhL`6X3aIfL`6Z3aIeL_6[3bIcL_6]3cIaL]6_3\\10000000001O000000000000000000001O0000jGgLQ7Y3mHlLP7T3nHnLR7R3lHPMT7Q3iHRMV7n2iHSMW7m2gHUMY7k2eHXMZ7k2aHWM_7n2ZHTMf7n3O00001O0000CXHoKg7n3a0N2M3M3N2M3N2N2O1O1O1O1O1O100O1O1O1O1O1O100000001O001O001O0]NRG@o83`GJ`83gGIZ85jGHV86PHFQ88SHEm7:WHCj7:ZHDf7;^HCb7:cHC]7<gHAZ7=jHAU7=oHAR7=RI@n6>f2N1O2M2O2M3M3NZ`W5"}, "label": "sheep going last and the head is covered behind another sheep"}]}
{"id": 144519, "image": "COCO_train2014_000000144519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep who ' s head is not visible , standing next to three other sheep\"."}], "task": "refering", "answer_template": "The \"a sheep who ' s head is not visible , standing next to three other sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 231, 233, 234, 235], "bbox": [0.038015625, 0.37205445544554455, 0.33667187499999995, 0.7827970297029703], "iscrowd": 0, "area": 22233.208700000003, "rle": {"size": [404, 640], "counts": "oe91]<9F:F:F:F:G9RFYNb8Q2RGlNU8\\1_GoN^8f2M3L3N3EhKUH[4h7gKVH\\4g7fKWH]4e7<M3L21O0000001O00000AcHgK]7U4iHiKX7R4mHmKS7o3RIPLn6l3WISLi6i3]IULd6f3aIYL_6c3fI\\LZ6`3_1N21O1O1O1O0O2O1O1O1O1O002N1O2N1O2N1O2N1O2N1O2N1O001O1O1O001O1O1O0O2O1O1O001O1OSOUHRMj7l2YHSMg7k2\\HUMc7j2_HVM`7h2cHXM\\7g2fHXMZ7h2gHXMX7h2iHXMV7h2lHWMS7i2nHVMR7j2oHVMP7j2QIVMn6j2SIUMm6k2TIUMl6j2UIVMj6j2XITMh6l2YITMf6l2[ISMe6m2\\IQMe6o2\\IPMd6P3^ImLc6S3^IlLb6T3_IjLb6V3_IiLb6V3`IhL`6X3aIfL`6Z3aIeL_6[3bIcL_6]3cIaL]6_3\\10000000001O000000000000000000001O0000jGgLQ7Y3mHlLP7T3nHnLR7R3lHPMT7Q3iHRMV7n2iHSMW7m2gHUMY7k2eHXMZ7k2aHWM_7n2ZHTMf7n3O00001O0000CXHoKg7n3a0N2M3M3N2M3N2N2O1O1O1O1O1O100O1O1O1O1O1O100000001O001O001O0]NRG@o83`GJ`83gGIZ85jGHV86PHFQ88SHEm7:WHCj7:ZHDf7;^HCb7:cHC]7<gHAZ7=jHAU7=oHAR7=RI@n6>f2N1O2M2O2M3M3NZ`W5"}, "label": "a sheep who ' s head is not visible , standing next to three other sheep"}]}
{"id": 406666, "image": "COCO_train2014_000000406666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffes graing over the grass in a cage\"."}], "task": "refering", "answer_template": "The \"two giraffes graing over the grass in a cage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 87, 88, 89, 101, 102, 103, 104, 117, 118, 119, 120, 133, 134, 135, 136, 149, 150, 151, 152, 165, 166, 168], "bbox": [0.3197333333333333, 0.37612426035502955, 0.5870222222222221, 0.9099112426035502], "iscrowd": 0, "area": 9497.483899999997, "rle": {"size": [338, 450], "counts": "]i_1a0h9n0QO:G9F:G9ROo0OO1O1O1D<L3N3M3M3M3N2N2O0O2O0O2O000O2O001N10001N101O1NbHcNj5]1nIlNR6S1fIVOY6i0_IAa6X200O100O10001NO2L3M4L4L3M4L3M4L4N1O2O0O1O2O0O1O2O0O1O1O2O0O11O100O1O2O0O1O1O2O0O1O1O101N1SJRMU4o2bKZM^4f2YKcMf4_2PKjMo4W2hJRNW5o1`JZN`5c1h1[Oe0\\Oe0\\Oc0O100O2O0001O010O0N3O010O0010O010O01O010O10O0100O1O010O10OeXm1"}, "label": "two giraffes graing over the grass in a cage"}]}
{"id": 406666, "image": "COCO_train2014_000000406666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller of the two giraffes\"."}], "task": "refering", "answer_template": "The \"the smaller of the two giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 87, 88, 89, 101, 102, 103, 104, 117, 118, 119, 120, 133, 134, 135, 136, 149, 150, 151, 152, 165, 166, 168], "bbox": [0.3197333333333333, 0.37612426035502955, 0.5870222222222221, 0.9099112426035502], "iscrowd": 0, "area": 9497.483899999997, "rle": {"size": [338, 450], "counts": "]i_1a0h9n0QO:G9F:G9ROo0OO1O1O1D<L3N3M3M3M3N2N2O0O2O0O2O000O2O001N10001N101O1NbHcNj5]1nIlNR6S1fIVOY6i0_IAa6X200O100O10001NO2L3M4L4L3M4L3M4L4N1O2O0O1O2O0O1O2O0O1O1O2O0O11O100O1O2O0O1O1O2O0O1O1O101N1SJRMU4o2bKZM^4f2YKcMf4_2PKjMo4W2hJRNW5o1`JZN`5c1h1[Oe0\\Oe0\\Oc0O100O2O0001O010O0N3O010O0010O010O01O010O10O0100O1O010O10OeXm1"}, "label": "the smaller of the two giraffes"}]}
{"id": 406666, "image": "COCO_train2014_000000406666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffe is eating leaves\"."}], "task": "refering", "answer_template": "The \"two giraffe is eating leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 69, 70, 71, 72, 73, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 160, 161, 162, 163, 164, 176, 177, 179, 180], "bbox": [0.03713333333333334, 0.24680473372781064, 0.6633333333333333, 0.9884023668639053], "iscrowd": 0, "area": 23653.19205, "rle": {"size": [338, 450], "counts": "hj57Y::F9G:E;_HeNY5d1TJhNj5b1aIlN\\6f2N2N2M4M2N2N2N2NG8]Od0E;7J7J6I7I8I6I7I3N1N100O2N100O2O0O1O2O0OL5H7H8H9G8H9G8M301N1O2O000010O01O01O010O0010O0100O1O100O100O1O100O1jJnLl2R3bL`M]3a2QLQNn3P2kKWNU4j1cK]N\\4d1^KbNb4^1XKhNg4b3O100O100ON2M4L3M4L3M3M4L3M4L3M3M4L3M4L4L3MWK^Lj2a3lLkLS3T3cLWM]3h2YLcMf3^2RLjMn3W2lKnMT4S2gKQNY4P2aKUN^4P4000O10O10O1000OD=B>B>K5L3M4L4L4L3M4L4M3M2N3M3M3[NnHGT7:QI]OS7b0TIUOo6k0VImNm6S1WIfNk6Z1VIcNm6]1RIaNQ7^1PI_NS7a1mH\\NU7e1l0O100O01000O10O010000O010O1000O010000O01000O10O10O1000O0100000O01000O010O10O0100O0010O001O001O010O001O001O000010O01O001O001O0010O01O000010O01O010O010O0010O010O00010O0O1O2N1N3N11O100O100O00100O100O1O10O01O100O100O1O010O100O1O100O10O01O100O1O100O100O1O100O100O1O102M3N1N3M3N2M3M3N2M3NZRb1"}, "label": "two giraffe is eating leaves"}]}
{"id": 406666, "image": "COCO_train2014_000000406666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest giraffe eating grasses\"."}], "task": "refering", "answer_template": "The \"the biggest giraffe eating grasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 69, 70, 71, 72, 73, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 160, 161, 162, 163, 164, 176, 177, 179, 180], "bbox": [0.03713333333333334, 0.24680473372781064, 0.6633333333333333, 0.9884023668639053], "iscrowd": 0, "area": 23653.19205, "rle": {"size": [338, 450], "counts": "hj57Y::F9G:E;_HeNY5d1TJhNj5b1aIlN\\6f2N2N2M4M2N2N2N2NG8]Od0E;7J7J6I7I8I6I7I3N1N100O2N100O2O0O1O2O0OL5H7H8H9G8H9G8M301N1O2O000010O01O01O010O0010O0100O1O100O100O1O100O1jJnLl2R3bL`M]3a2QLQNn3P2kKWNU4j1cK]N\\4d1^KbNb4^1XKhNg4b3O100O100ON2M4L3M4L3M3M4L3M4L3M3M4L3M4L4L3MWK^Lj2a3lLkLS3T3cLWM]3h2YLcMf3^2RLjMn3W2lKnMT4S2gKQNY4P2aKUN^4P4000O10O10O1000OD=B>B>K5L3M4L4L4L3M4L4M3M2N3M3M3[NnHGT7:QI]OS7b0TIUOo6k0VImNm6S1WIfNk6Z1VIcNm6]1RIaNQ7^1PI_NS7a1mH\\NU7e1l0O100O01000O10O010000O010O1000O010000O01000O10O10O1000O0100000O01000O010O10O0100O0010O001O001O010O001O001O000010O01O001O001O0010O01O000010O01O010O010O0010O010O00010O0O1O2N1N3N11O100O100O00100O100O1O10O01O100O100O1O010O100O1O100O10O01O100O1O100O100O1O100O100O1O102M3N1N3M3N2M3M3N2M3NZRb1"}, "label": "the biggest giraffe eating grasses"}]}
{"id": 242827, "image": "COCO_train2014_000000242827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"several pieces of brown cake\"."}], "task": "refering", "answer_template": "The \"several pieces of brown cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 12, 13, 14, 15, 16, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 330, 331, 332], "bbox": [0.230546875, 0.011770833333333335, 0.7215156250000001, 0.8454999999999999], "iscrowd": 0, "area": 92146.46735, "rle": {"size": [480, 640], "counts": "WcU2^1^=<E8G9fNnMXEZ2`:nMXE[2_:nMYEZ2Y9mLcG]4Q8kKkG^4i7iKSH`4h7`KVHf4e7[KYHl4b7TK\\Ho4b7RK[HS5c7lJYHZ5e7gJVH_5i7aJRHf5k7i0O1N3N1N3N1N2O2M2O1O2M3N1O2N2M3N2N1O2M3N2N2M2O2N1O0O010000O01000O10O100[N^HlKb7R4QI]Ko6a4SI_Km6_4UIaKk6]4WIdKh6Z4ZIgKe6W4^IiKa6U4aIlK^6R4dInK\\6P4fIQLY6m3iITLV6j3mIVLS6g3oIZLP6d3RJ\\Ln5c3SJ]Lm5a3UJ^Ll5`3VJ`Lj5^3YJaLg5]3[JbLf5^3ZJbLf5^3ZJaLg5`3XJ`Lh5`3YJ_Lg5a3YJ^Lh5c3WJ]Li5c3WJ\\Lj5d3VJ\\Lj5d3VJ\\Lj5e3VJYLk5g3UJYLk5g3UJYLk5h3TJXLl5h3TJWLm5i3TJVLl5k3SJULn5j3RJULo5k3QJVLn5k3QJULo5k3QJULo5k3RJULm5l3RJTLn5l3RJULm5k3SJULm5l3RJTLn5i0cGV2`2QMl5g0fGX2^2QMl5e0hG[2[2PMm5b0kG^2X2QMl5?mGa2W2nLn5>nGd2`;0000000000000O0100O10O10O10O0100O010O10VGPMd5o2SJPM^M1^8P3UJoL[M2`8o2UJoL[M2`8n2VJPMZM2_8o2WJoLYM3`8m2XJPMXM4_8l2YJSMUM1b8k2[J[Me5d2\\J\\Md5c2]J]Md5a2]J_Mc5_2_J`Mb5_2`J`M`5_2aJaM_5^2bJaM_5^2bJbM^5^2bJbM^5^2bJbM^5^2cJbM]5^2bJbM^5^2bJbM^5^2bJbM]5_2cJaM]5`2cJ_M\\5b2dJ^M\\5b2dJ^M\\5b2dJ_MZ5d2dJ\\M\\5g2aJYM^5k2`JTM`5n2^JRMa5R3\\JnLd5U3YJkLf5X3XJhLh5[3VJdLi5]3WJcLi5]3WJbLi5_3WJaLi5_3WJaLh5`3YJ_Lg5a3YJ_Lf5b3ZJ^Lf5b3ZJ]Lf5d3ZJ\\Lf5d3ZJ[Lf5f3j200O010000O100O10000O10000O1000O010000O100O1000O0100O100O1O100O00100O1O1O100O0\\NPLYHQ4`7YL]Hf3a7`L[H`3f7cLVH]3k7eLSH[3l7gLRHY3o7iLnGW3S8jLlGU3U8lLiGT3X8mLfGS3[8nLdGQ3\\8QMbGP3^8RM_Gn2b8SM]Gl2d8UMZGl2f8UMXGl2h8UMWGk2h8WMVGj2j8WMTGi2m8WMSGi2m8XMQGi2o8XMPGh2o8ZMoFg2Q9ZMnFe2S9[MmFe2S9d100O10O100000O10000000000O10000gJTJ^1l5bNVJ\\1j5eNWJX1j5hNXJV1h5jN[JS1e5mN]JQ1c5oN_Jo0a5QObJk0_5UOcJi0]5WOeJg0[5YOhJd0X5\\OjJ8\\ORLj5f3lJ3_OTLf5i3nJN`52bJJ`56bJF`5:cJ@`5`0`J^Ob5c0^J\\Ob5d0_JZOb5f0_JYOa5g0`JWOa5i0`JVO`5j0`JVO`5j0aJTO`5l0aJSO_5m0bJQO_5o0bJPO^5P1cJnN^5R1cJmN]5S1cJlN^5T1cJkN]5U1dJiN]5W1dJhN\\5X1eJfN\\5Z1dJfN\\5Z1eJdN\\5\\1dJdN\\5\\1dJdN\\5\\1eJbN\\5^1dJbN\\5^1dJaN]5_1dJ`N\\5`1dJ_N]5a1cJ_N]5a1dJfL^NP1o6T2iJeLbNT1e6j1VKkL^NX1]6b1_L[Na3]1iL`NX3X1SMeNm2R1^MkNd2l0gMQOY2g0QNVOP2e0XNXOh1e0]NYOc1e0bNWO`1e0gNWOY1g0nNSOT1i0TOROm0k0f6L3N3L8I8GX]c2"}, "label": "several pieces of brown cake"}]}
{"id": 242827, "image": "COCO_train2014_000000242827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of cake\"."}], "task": "refering", "answer_template": "The \"a piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 12, 13, 14, 15, 16, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 330, 331, 332], "bbox": [0.230546875, 0.011770833333333335, 0.7215156250000001, 0.8454999999999999], "iscrowd": 0, "area": 92146.46735, "rle": {"size": [480, 640], "counts": "WcU2^1^=<E8G9fNnMXEZ2`:nMXE[2_:nMYEZ2Y9mLcG]4Q8kKkG^4i7iKSH`4h7`KVHf4e7[KYHl4b7TK\\Ho4b7RK[HS5c7lJYHZ5e7gJVH_5i7aJRHf5k7i0O1N3N1N3N1N2O2M2O1O2M3N1O2N2M3N2N1O2M3N2N2M2O2N1O0O010000O01000O10O100[N^HlKb7R4QI]Ko6a4SI_Km6_4UIaKk6]4WIdKh6Z4ZIgKe6W4^IiKa6U4aIlK^6R4dInK\\6P4fIQLY6m3iITLV6j3mIVLS6g3oIZLP6d3RJ\\Ln5c3SJ]Lm5a3UJ^Ll5`3VJ`Lj5^3YJaLg5]3[JbLf5^3ZJbLf5^3ZJaLg5`3XJ`Lh5`3YJ_Lg5a3YJ^Lh5c3WJ]Li5c3WJ\\Lj5d3VJ\\Lj5d3VJ\\Lj5e3VJYLk5g3UJYLk5g3UJYLk5h3TJXLl5h3TJWLm5i3TJVLl5k3SJULn5j3RJULo5k3QJVLn5k3QJULo5k3QJULo5k3RJULm5l3RJTLn5l3RJULm5k3SJULm5l3RJTLn5i0cGV2`2QMl5g0fGX2^2QMl5e0hG[2[2PMm5b0kG^2X2QMl5?mGa2W2nLn5>nGd2`;0000000000000O0100O10O10O10O0100O010O10VGPMd5o2SJPM^M1^8P3UJoL[M2`8o2UJoL[M2`8n2VJPMZM2_8o2WJoLYM3`8m2XJPMXM4_8l2YJSMUM1b8k2[J[Me5d2\\J\\Md5c2]J]Md5a2]J_Mc5_2_J`Mb5_2`J`M`5_2aJaM_5^2bJaM_5^2bJbM^5^2bJbM^5^2bJbM^5^2cJbM]5^2bJbM^5^2bJbM^5^2bJbM]5_2cJaM]5`2cJ_M\\5b2dJ^M\\5b2dJ^M\\5b2dJ_MZ5d2dJ\\M\\5g2aJYM^5k2`JTM`5n2^JRMa5R3\\JnLd5U3YJkLf5X3XJhLh5[3VJdLi5]3WJcLi5]3WJbLi5_3WJaLi5_3WJaLh5`3YJ_Lg5a3YJ_Lf5b3ZJ^Lf5b3ZJ]Lf5d3ZJ\\Lf5d3ZJ[Lf5f3j200O010000O100O10000O10000O1000O010000O100O1000O0100O100O1O100O00100O1O1O100O0\\NPLYHQ4`7YL]Hf3a7`L[H`3f7cLVH]3k7eLSH[3l7gLRHY3o7iLnGW3S8jLlGU3U8lLiGT3X8mLfGS3[8nLdGQ3\\8QMbGP3^8RM_Gn2b8SM]Gl2d8UMZGl2f8UMXGl2h8UMWGk2h8WMVGj2j8WMTGi2m8WMSGi2m8XMQGi2o8XMPGh2o8ZMoFg2Q9ZMnFe2S9[MmFe2S9d100O10O100000O10000000000O10000gJTJ^1l5bNVJ\\1j5eNWJX1j5hNXJV1h5jN[JS1e5mN]JQ1c5oN_Jo0a5QObJk0_5UOcJi0]5WOeJg0[5YOhJd0X5\\OjJ8\\ORLj5f3lJ3_OTLf5i3nJN`52bJJ`56bJF`5:cJ@`5`0`J^Ob5c0^J\\Ob5d0_JZOb5f0_JYOa5g0`JWOa5i0`JVO`5j0`JVO`5j0aJTO`5l0aJSO_5m0bJQO_5o0bJPO^5P1cJnN^5R1cJmN]5S1cJlN^5T1cJkN]5U1dJiN]5W1dJhN\\5X1eJfN\\5Z1dJfN\\5Z1eJdN\\5\\1dJdN\\5\\1dJdN\\5\\1eJbN\\5^1dJbN\\5^1dJaN]5_1dJ`N\\5`1dJ_N]5a1cJ_N]5a1dJfL^NP1o6T2iJeLbNT1e6j1VKkL^NX1]6b1_L[Na3]1iL`NX3X1SMeNm2R1^MkNd2l0gMQOY2g0QNVOP2e0XNXOh1e0]NYOc1e0bNWO`1e0gNWOY1g0nNSOT1i0TOROm0k0f6L3N3L8I8GX]c2"}, "label": "a piece of cake"}]}
{"id": 578702, "image": "COCO_train2014_000000578702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccoli at the bottom of the dish\"."}], "task": "refering", "answer_template": "The \"the broccoli at the bottom of the dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381], "bbox": [0.13583682008368203, 0.7189687499999999, 0.5125732217573222, 0.975171875], "iscrowd": 0, "area": 19476.45845, "rle": {"size": [640, 478], "counts": "jSY12mc05K4L5AFo\\O>kb0`0K6J5K5J5L3N2N3M2N3M2N2N2O1O001O1O0U_ObMT`0^2j_OeMU`0[2i_OhMV`0X2h_OkMX`0T2e_OPN[`0n1__OZN``0d200O0O2O1N2O1N2O1N100O101O000000001O01O0001O000000001O1O1O1O1O1^@TLn>m3n@VLR?k3j@XLV?i3f@ZLZ?f3c@]L]?S40000O100O2O0O100O100O1000000O1000001O0O2O1O001N101O00001O000O2O001O00000000000000000000001N1000000002N2N2N3M2N2N2N2M4M3M3L3N2N2M4M2[Lm_OY3^`0O2M2O1N3N1O2M2O1N3N1N2O2M3N2M2O2M3M2O2M1O100O1O2N2O1N2N1O2N2N2N1O2N2N2N1O2N2N2N2M2L5L4K5M4K6K5J6K]Ra4"}, "label": "the broccoli at the bottom of the dish"}]}
{"id": 578702, "image": "COCO_train2014_000000578702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoli closest to the bottom of the plate\"."}], "task": "refering", "answer_template": "The \"a piece of broccoli closest to the bottom of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381], "bbox": [0.13583682008368203, 0.7189687499999999, 0.5125732217573222, 0.975171875], "iscrowd": 0, "area": 19476.45845, "rle": {"size": [640, 478], "counts": "jSY12mc05K4L5AFo\\O>kb0`0K6J5K5J5L3N2N3M2N3M2N2N2O1O001O1O0U_ObMT`0^2j_OeMU`0[2i_OhMV`0X2h_OkMX`0T2e_OPN[`0n1__OZN``0d200O0O2O1N2O1N2O1N100O101O000000001O01O0001O000000001O1O1O1O1O1^@TLn>m3n@VLR?k3j@XLV?i3f@ZLZ?f3c@]L]?S40000O100O2O0O100O100O1000000O1000001O0O2O1O001N101O00001O000O2O001O00000000000000000000001N1000000002N2N2N3M2N2N2N2M4M3M3L3N2N2M4M2[Lm_OY3^`0O2M2O1N3N1O2M2O1N3N1N2O2M3N2M2O2M3M2O2M1O100O1O2N2O1N2N1O2N2N2N1O2N2N2N1O2N2N2N2M2L5L4K5M4K6K5J6K]Ra4"}, "label": "a piece of broccoli closest to the bottom of the plate"}]}
{"id": 578702, "image": "COCO_train2014_000000578702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate of stir fried vegetables with tofu on a plate\"."}], "task": "refering", "answer_template": "The \"a plate of stir fried vegetables with tofu on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.000606694560669456, 0.056499999999999995, 1.0, 1.0], "iscrowd": 0, "area": 252992.5543, "rle": {"size": [640, 478], "counts": "_?\\4_5U:O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O101N100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O2N100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O101N100O100O1O100O100O100O100O100O1O10000O10000000000O1000000000000O10000000000O100001O001O00001O001O001O001O00001O001O001O001O1O001O001O1O001O001O1O001O001O001O10O01O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O100O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1N2NO`@dK_?`4M0d@aKZ?a4f@_KX?c4h@\\KV?f4k@ZKS?h4l@YKQ?j4o@VKo>k4RAUKl>m4TARKj>Q5VAoJh>R5XAoJf>S5ZAmJc>V5]AiJb>X5_AhJ^>[5bAeJ\\>]5dAcJZ>^5fAbJX>a5hA_JV>c5jA]JT>d5mA\\JP>g5PBYJn=i5QBWJn=j5SBVJj=m5VBSJh=o5XBQJe=Q642WBmIb=U6^BkI_=W6bBiI\\=Y6dBgIZ=Z643bBcIV=_6jBaIR=b6oB^In<e643M2QCVIe<n6[CRIb<Q742N3L4`CfHV<\\7kCdHR<_753L3N3oCXHf;k752N3L4M2N3M3L3O2M2N3M3N_J"}, "label": "a plate of stir fried vegetables with tofu on a plate"}]}
{"id": 578702, "image": "COCO_train2014_000000578702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plate filled with vegetables\"."}], "task": "refering", "answer_template": "The \"a white plate filled with vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.000606694560669456, 0.056499999999999995, 1.0, 1.0], "iscrowd": 0, "area": 252992.5543, "rle": {"size": [640, 478], "counts": "_?\\4_5U:O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O101N100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O2N100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O101N100O100O1O100O100O100O100O100O1O10000O10000000000O1000000000000O10000000000O100001O001O00001O001O001O001O00001O001O001O001O1O001O001O1O001O001O1O001O001O001O10O01O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O100O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1N2NO`@dK_?`4M0d@aKZ?a4f@_KX?c4h@\\KV?f4k@ZKS?h4l@YKQ?j4o@VKo>k4RAUKl>m4TARKj>Q5VAoJh>R5XAoJf>S5ZAmJc>V5]AiJb>X5_AhJ^>[5bAeJ\\>]5dAcJZ>^5fAbJX>a5hA_JV>c5jA]JT>d5mA\\JP>g5PBYJn=i5QBWJn=j5SBVJj=m5VBSJh=o5XBQJe=Q642WBmIb=U6^BkI_=W6bBiI\\=Y6dBgIZ=Z643bBcIV=_6jBaIR=b6oB^In<e643M2QCVIe<n6[CRIb<Q742N3L4`CfHV<\\7kCdHR<_753L3N3oCXHf;k752N3L4M2N3M3L3O2M2N3M3N_J"}, "label": "a white plate filled with vegetables"}]}
{"id": 578702, "image": "COCO_train2014_000000578702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a single broccoli floret sitting on the top\"."}], "task": "refering", "answer_template": "The \"a single broccoli floret sitting on the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 260, 261, 262, 263, 264, 265, 280, 281], "bbox": [0.25518828451882847, 0.46353125000000006, 0.6489748953974895, 0.71059375], "iscrowd": 0, "area": 20779.098349999997, "rle": {"size": [640, 478], "counts": "]c\\2:fc0a0^O1O2N2N2N2N1O2N2N2N2N2N1O2N2O1N2N1O2N2N2M3N2N1N3N2M3N2N1N3N2N2M3N2N1N3N2N2N2N1N3N2N2N2N1O2N2N2N2N2N1O1O1N2O1O100O10000000000000000O1000000000000000000O10000000001O001O1N101O001O1O001O001N101O1O001O001O001O1N101O001O001O00000010O001O000010O01O001O0010O01O000010O01O001O0010O0001O001O010_Nb@^N^?a1d@]N^?a1d@^N\\?b1e@]N\\?a1f@]N[?b1g@]NY?b1i@]NX?b1i@\\NX?c1j@\\NW?b1k@]NU?b1m@\\NU?c1l@\\NT?c1m@\\NS?d1o@[NQ?d1QA[Nn>e1TAYNl>h1UAWNj>i1XATNi>l1YAlMm>T2UAcMS?]2W100000010O0000010O3M2N3N2M3M2N3N2M2N3M3N1N3M3WO`fX3"}, "label": "a single broccoli floret sitting on the top"}]}
{"id": 578702, "image": "COCO_train2014_000000578702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoli on top and in the middle of the plate\"."}], "task": "refering", "answer_template": "The \"a piece of broccoli on top and in the middle of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 260, 261, 262, 263, 264, 265, 280, 281], "bbox": [0.25518828451882847, 0.46353125000000006, 0.6489748953974895, 0.71059375], "iscrowd": 0, "area": 20779.098349999997, "rle": {"size": [640, 478], "counts": "]c\\2:fc0a0^O1O2N2N2N2N1O2N2N2N2N2N1O2N2O1N2N1O2N2N2M3N2N1N3N2M3N2N1N3N2N2M3N2N1N3N2N2N2N1N3N2N2N2N1O2N2N2N2N2N1O1O1N2O1O100O10000000000000000O1000000000000000000O10000000001O001O1N101O001O1O001O001N101O1O001O001O001O1N101O001O001O00000010O001O000010O01O001O0010O01O000010O01O001O0010O0001O001O010_Nb@^N^?a1d@]N^?a1d@^N\\?b1e@]N\\?a1f@]N[?b1g@]NY?b1i@]NX?b1i@\\NX?c1j@\\NW?b1k@]NU?b1m@\\NU?c1l@\\NT?c1m@\\NS?d1o@[NQ?d1QA[Nn>e1TAYNl>h1UAWNj>i1XATNi>l1YAlMm>T2UAcMS?]2W100000010O0000010O3M2N3N2M3M2N3N2M2N3M3N1N3M3WO`fX3"}, "label": "a piece of broccoli on top and in the middle of the plate"}]}
{"id": 562330, "image": "COCO_train2014_000000562330.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rear end of a military vehicle with a camp cover\"."}], "task": "refering", "answer_template": "The \"the rear end of a military vehicle with a camp cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 243, 244, 245], "bbox": [0.39914062499999997, 0.2060747663551402, 0.725984375, 0.7263551401869158], "iscrowd": 0, "area": 38881.96639999999, "rle": {"size": [428, 640], "counts": "ejZ32m<c0\\Od0^Na1iNX1ROn0I5K3M3L4M3M2N3M3L4M2N1O1O1N2O1O2N1O1O1N2O1O1O1O100O100000000O100000000O100000000O10N2M3N2M2N3N2J6J61N101O001O0O2O001O001N101OO1O1N101O1N2O1O0O2001O001N10001O000O2O00001O00001O01O01O00001O00001OO1N2O1O1O1N2O1O11O1O001O010O2N1O1O1O1O2N1O1O1O2N1O1O00O100O100O100O2O0002N2N1O2N2N2N1O001O01O0000010O00001O0001O01O00000010O0001O01O00010O0000010O000010I6H8H9M2O10001N10000O2O0O1J7I6I7J6eM`GDg88^G@h8=]G[Oi8a0]GVOj8g0[GQOl8k0YGmNm8o0YGiNm8T1WGeNo8X1VG`NT9X1g1H8H9F9H8H_PY2"}, "label": "the rear end of a military vehicle with a camp cover"}]}
{"id": 562330, "image": "COCO_train2014_000000562330.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an army truck on the road\"."}], "task": "refering", "answer_template": "The \"an army truck on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 243, 244, 245], "bbox": [0.39914062499999997, 0.2060747663551402, 0.725984375, 0.7263551401869158], "iscrowd": 0, "area": 38881.96639999999, "rle": {"size": [428, 640], "counts": "ejZ32m<c0\\Od0^Na1iNX1ROn0I5K3M3L4M3M2N3M3L4M2N1O1O1N2O1O2N1O1O1N2O1O1O1O100O100000000O100000000O100000000O10N2M3N2M2N3N2J6J61N101O001O0O2O001O001N101OO1O1N101O1N2O1O0O2001O001N10001O000O2O00001O00001O01O01O00001O00001OO1N2O1O1O1N2O1O11O1O001O010O2N1O1O1O1O2N1O1O1O2N1O1O00O100O100O100O2O0002N2N1O2N2N2N1O001O01O0000010O00001O0001O01O00000010O0001O01O00010O0000010O000010I6H8H9M2O10001N10000O2O0O1J7I6I7J6eM`GDg88^G@h8=]G[Oi8a0]GVOj8g0[GQOl8k0YGmNm8o0YGiNm8T1WGeNo8X1VG`NT9X1g1H8H9F9H8H_PY2"}, "label": "an army truck on the road"}]}
{"id": 226460, "image": "COCO_train2014_000000226460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unoccupied boat , docked right next to an occupied boat\"."}], "task": "refering", "answer_template": "The \"an unoccupied boat , docked right next to an occupied boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 269], "bbox": [0.588484375, 0.48765000000000003, 1.0, 0.7955], "iscrowd": 0, "area": 22092.983500000002, "rle": {"size": [400, 640], "counts": "Wac45V<5K5L5J5K6J5L5J5K6K401O0O2O001O001O0O2O001O001O0O2O1O001O001N101O001O0O2O001O1O001N101O001O001N101O001O001N1000001O0O10001O00000O2O0000001O0O10001O00000O2O0000001O0O10O1O1O10O01O1O1O100O001O1O1O100O01000O10000O0100000O10O10O10000O10O10O100000O010000O10O10O1000000O01000O1000O01000000000O010000000O100000O1000O100000000O01000000000O100000O10O10000000000O010000000O100000O1000000000000000O10001O00000000001O0000000O10001O00000000001O00000O1000001O00000000001O00000O1000001O00000000001O000O0100O100O10000O100O100O10O010000O1@`0]N`K"}, "label": "an unoccupied boat , docked right next to an occupied boat"}]}
{"id": 226460, "image": "COCO_train2014_000000226460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat in the water and snow\"."}], "task": "refering", "answer_template": "The \"a boat in the water and snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 269], "bbox": [0.588484375, 0.48765000000000003, 1.0, 0.7955], "iscrowd": 0, "area": 22092.983500000002, "rle": {"size": [400, 640], "counts": "Wac45V<5K5L5J5K6J5L5J5K6K401O0O2O001O001O0O2O001O001O0O2O1O001O001N101O001O0O2O001O1O001N101O001O001N101O001O001N1000001O0O10001O00000O2O0000001O0O10001O00000O2O0000001O0O10O1O1O10O01O1O1O100O001O1O1O100O01000O10000O0100000O10O10O10000O10O10O100000O010000O10O10O1000000O01000O1000O01000000000O010000000O100000O1000O100000000O01000000000O100000O10O10000000000O010000000O100000O1000000000000000O10001O00000000001O0000000O10001O00000000001O00000O1000001O00000000001O00000O1000001O00000000001O000O0100O100O10000O100O100O10O010000O1@`0]N`K"}, "label": "a boat in the water and snow"}]}
{"id": 226460, "image": "COCO_train2014_000000226460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of skis in the boat\"."}], "task": "refering", "answer_template": "The \"a pair of skis in the boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 155, 156, 157, 178, 179, 180, 200, 201, 222, 223, 244, 245], "bbox": [0.630015625, 0.40909999999999996, 0.85728125, 0.794125], "iscrowd": 0, "area": 3783.1089500000003, "rle": {"size": [400, 640], "counts": "hSn42\\<2O2M2O2N1N3JCVD>g;DYD=e;EZD<d;E\\D;d;E]D;b;F]D;b;E^D;b;F]D;b;E^D<a;D`D;`;F_D;`;E`D;`;;O[OaD:^;<0O100O100O10000O1E^DIb;7_DHa;8`DF`;;`DE`;:aDF_;:aDF_;:aDF_;:`DG`;9`DG`;9`DG`;8aDH_;8aDH_;8aDH_;8aDH_;8aDH_;8aDG_;:aDF_;:aDF_;:aDF_;:aDF_;;`DE`;;`DE`;;`DE`;;`DE`;;`DE`;;`DE`;;`DE`;;`DE_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD^;=bDD];<cDD];<cDD\\;=dDC\\;<eDD[;<eDDZ;=fDCZ;=fDDY;;hDEW;<iDDW;<iDDW;<iDDV;<kDDU;<kDDU;<kDES;;nDER;;nDER;;nDEQ;<oDDQ;;PEEP;;oDFQ;:oDFP;;PEFo:9REGn:9REGm::SEFm::TEEl::UEFj:;VEEj:;VEFi:9YEFf:;ZEEf:;ZEEf:;ZEEe:;]EDc:<]EDc:<]EEb:;^EEb::`EE`:;`EEa::_EFa::_EFa:9aEF_::aEF_::aEF_:9bEG^:9cEF]::cEF]::cEF]:9dEG\\:9dEG\\:9eEF[::eEF[:9eEH[:8eEH\\:8bEI_:6aEJ`:5_ELb:4]ELe:2ZEOg:1XEOi:0WE0j:0TE1m:ORE1P;NnD3S;LmD4T;LjD5W;KhD5Y;KeD6\\;IdD7^;HaD8`;H\\D;f;5L42M3N2N2N2_Om\\S1"}, "label": "a pair of skis in the boat"}]}
{"id": 226460, "image": "COCO_train2014_000000226460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of skis sitting by themselves at the front of a boat\"."}], "task": "refering", "answer_template": "The \"a pair of skis sitting by themselves at the front of a boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 155, 156, 157, 178, 179, 180, 200, 201, 222, 223, 244, 245], "bbox": [0.630015625, 0.40909999999999996, 0.85728125, 0.794125], "iscrowd": 0, "area": 3783.1089500000003, "rle": {"size": [400, 640], "counts": "hSn42\\<2O2M2O2N1N3JCVD>g;DYD=e;EZD<d;E\\D;d;E]D;b;F]D;b;E^D;b;F]D;b;E^D<a;D`D;`;F_D;`;E`D;`;;O[OaD:^;<0O100O100O10000O1E^DIb;7_DHa;8`DF`;;`DE`;:aDF_;:aDF_;:aDF_;:`DG`;9`DG`;9`DG`;8aDH_;8aDH_;8aDH_;8aDH_;8aDH_;8aDG_;:aDF_;:aDF_;:aDF_;:aDF_;;`DE`;;`DE`;;`DE`;;`DE`;;`DE`;;`DE`;;`DE`;;`DE_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD_;<aDD^;=bDD];<cDD];<cDD\\;=dDC\\;<eDD[;<eDDZ;=fDCZ;=fDDY;;hDEW;<iDDW;<iDDW;<iDDV;<kDDU;<kDDU;<kDES;;nDER;;nDER;;nDEQ;<oDDQ;;PEEP;;oDFQ;:oDFP;;PEFo:9REGn:9REGm::SEFm::TEEl::UEFj:;VEEj:;VEFi:9YEFf:;ZEEf:;ZEEf:;ZEEe:;]EDc:<]EDc:<]EEb:;^EEb::`EE`:;`EEa::_EFa::_EFa:9aEF_::aEF_::aEF_:9bEG^:9cEF]::cEF]::cEF]:9dEG\\:9dEG\\:9eEF[::eEF[:9eEH[:8eEH\\:8bEI_:6aEJ`:5_ELb:4]ELe:2ZEOg:1XEOi:0WE0j:0TE1m:ORE1P;NnD3S;LmD4T;LjD5W;KhD5Y;KeD6\\;IdD7^;HaD8`;H\\D;f;5L42M3N2N2N2_Om\\S1"}, "label": "a pair of skis sitting by themselves at the front of a boat"}]}
{"id": 226460, "image": "COCO_train2014_000000226460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden john boat carries 5 boys with skis\"."}], "task": "refering", "answer_template": "The \"a wooden john boat carries 5 boys with skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 191, 192, 210, 211, 214, 215, 233, 234, 235, 236, 237, 238, 239, 240, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 312], "bbox": [0.125, 0.60675, 0.683984375, 0.95955], "iscrowd": 0, "area": 16622.6574, "rle": {"size": [400, 640], "counts": "R`o01^<3N2M2N3N2M3N2M3N2M2O2M3N2M3M3N2M2O2M3O0000O1O100O100O100O100O100O100O100O1O100O100O2O0O100O100O101N1O100O2O0O100O2O0O100O10O001N10001O001O0O2O001O001O001N101O00001O0O2O001O001O0O2O001O00001N101O001O001N101O001O001O0O101O001O0O2N101N101N1O2O0N2K6I6K5K6J5J6L5O0001O01O01O01O0001O01O01O01O00010O0001O01O00VNVGNk8POVHQ1d90000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001OJ6F:E;E;Fn`5MU_J:F9F;F9J6100O100O100O2O0O100O100O1O100O101N100O1001O00001O00010O001O00001O000102M3N2M3M2O2M3N2M3M3N1N3M3N2M3N2M3MWN0g15L4L4K5L4K5L4L4K5L4L4K5oEdNg8`1cFRO\\9R20O100O100O100O100O2O0O2O001N101N101N101N101N101N101N101O0O101N101N101N101N101N101O0O2O0O2O0O2O0O2O0N3K4L5K4L5L3L4L5K4L5K4MYc^2"}, "label": "a wooden john boat carries 5 boys with skis"}]}
{"id": 226460, "image": "COCO_train2014_000000226460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a canoe with people inside\"."}], "task": "refering", "answer_template": "The \"a canoe with people inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 191, 192, 210, 211, 214, 215, 233, 234, 235, 236, 237, 238, 239, 240, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 312], "bbox": [0.125, 0.60675, 0.683984375, 0.95955], "iscrowd": 0, "area": 16622.6574, "rle": {"size": [400, 640], "counts": "R`o01^<3N2M2N3N2M3N2M3N2M2O2M3N2M3M3N2M2O2M3O0000O1O100O100O100O100O100O100O100O1O100O100O2O0O100O100O101N1O100O2O0O100O2O0O100O10O001N10001O001O0O2O001O001O001N101O00001O0O2O001O001O0O2O001O00001N101O001O001N101O001O001O0O101O001O0O2N101N101N1O2O0N2K6I6K5K6J5J6L5O0001O01O01O01O0001O01O01O01O00010O0001O01O00VNVGNk8POVHQ1d90000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001OJ6F:E;E;Fn`5MU_J:F9F;F9J6100O100O100O2O0O100O100O1O100O101N100O1001O00001O00010O001O00001O000102M3N2M3M2O2M3N2M3M3N1N3M3N2M3N2M3MWN0g15L4L4K5L4K5L4L4K5L4L4K5oEdNg8`1cFRO\\9R20O100O100O100O100O2O0O2O001N101N101N101N101N101N101N101O0O101N101N101N101N101N101O0O2O0O2O0O2O0O2O0N3K4L5K4L5L3L4L5K4L5K4MYc^2"}, "label": "a canoe with people inside"}]}
{"id": 234653, "image": "COCO_train2014_000000234653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing up against a large wooden fence made of posts\"."}], "task": "refering", "answer_template": "The \"zebra standing up against a large wooden fence made of posts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 78, 79, 82, 83, 84, 85, 86, 87, 88, 95, 96, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 138, 139, 142, 172, 173, 174, 190, 191, 192, 208, 210, 226, 228], "bbox": [0.3112627291242363, 0.17808000000000002, 0.9729124236252547, 0.7274400000000001], "iscrowd": 0, "area": 17140.130400000002, "rle": {"size": [500, 491], "counts": "dkZ21`?2N3M2N2O1N2N2N21O10O01O1O0010O10O10O10O10O01000O010O10O10O01000O010000O100O10000O10000O10000O100O10000O10000O100O10000O10000O100O10000O1000aQl05TnSO5K5@a0J5L5J5dJaN`Lc1`3_NYLg1g3ZNSLj1n3WNkKo1U4RNeKR2\\4oM^KU2c4lMiJg2V5\\MhJc2Y5^MgJa2Y5`MfJ_2[5bMdJh1R6YNmIa1X6aNhIZ1\\6gNcIU1a6lN^IQ1e6PO[Ik0i6VOVIf0m6]OQI?S7BlH;W7FiH5[7LdH0_72`HJd77[HEi7<WH_Om7b0RHZOQ8h0nGTOV8m0iGoN[8S1Q32O1O1O1N2O1O1O1N2O2N1O1iDROe7P1[H9[6HeI9Y6HgI:U6HkI9S6HmI9Q6HoI:n5GRJ:k5IUJ7i5JWJ8f5IZJ8d5I\\J9`5I`J9]5HcJ:Z5GfJlNYM9m7m0jJ]NhMe0\\7o0mJnMWNn0l6V1ZLeNg3]1ZL_Ng3b1\\LYNd3i1_LRNb3o1`LmMa3T2bLgM^3[2eL`M\\3a2fL[M[3f2h42O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1NO2L4M2N3M3M2N3L3O2O10O01O1O001O001O1O001O1O001O1O001O001O1O001O1O0011N101O1N101O1N101O001O10O01O001O1O010O001O1O001O10O01O001O1O010O2N1O2N1O2N101N1O2M0100000000000OaMUDW1k;aN`D\\1`;\\NlD`1T;XNWEe1`<K5K4M4K5K4L5L2M3M2N3N2M3M3M3N2M3M3NVU6"}, "label": "zebra standing up against a large wooden fence made of posts"}]}
{"id": 234653, "image": "COCO_train2014_000000234653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in the back side of the image\"."}], "task": "refering", "answer_template": "The \"zebra in the back side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 78, 79, 82, 83, 84, 85, 86, 87, 88, 95, 96, 101, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 125, 138, 139, 142, 172, 173, 174, 190, 191, 192, 208, 210, 226, 228], "bbox": [0.3112627291242363, 0.17808000000000002, 0.9729124236252547, 0.7274400000000001], "iscrowd": 0, "area": 17140.130400000002, "rle": {"size": [500, 491], "counts": "dkZ21`?2N3M2N2O1N2N2N21O10O01O1O0010O10O10O10O10O01000O010O10O10O01000O010000O100O10000O10000O10000O100O10000O10000O100O10000O10000O100O10000O1000aQl05TnSO5K5@a0J5L5J5dJaN`Lc1`3_NYLg1g3ZNSLj1n3WNkKo1U4RNeKR2\\4oM^KU2c4lMiJg2V5\\MhJc2Y5^MgJa2Y5`MfJ_2[5bMdJh1R6YNmIa1X6aNhIZ1\\6gNcIU1a6lN^IQ1e6PO[Ik0i6VOVIf0m6]OQI?S7BlH;W7FiH5[7LdH0_72`HJd77[HEi7<WH_Om7b0RHZOQ8h0nGTOV8m0iGoN[8S1Q32O1O1O1N2O1O1O1N2O2N1O1iDROe7P1[H9[6HeI9Y6HgI:U6HkI9S6HmI9Q6HoI:n5GRJ:k5IUJ7i5JWJ8f5IZJ8d5I\\J9`5I`J9]5HcJ:Z5GfJlNYM9m7m0jJ]NhMe0\\7o0mJnMWNn0l6V1ZLeNg3]1ZL_Ng3b1\\LYNd3i1_LRNb3o1`LmMa3T2bLgM^3[2eL`M\\3a2fL[M[3f2h42O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1NO2L4M2N3M3M2N3L3O2O10O01O1O001O001O1O001O1O001O1O001O001O1O001O1O0011N101O1N101O1N101O001O10O01O001O1O010O001O1O001O10O01O001O1O010O2N1O2N1O2N101N1O2M0100000000000OaMUDW1k;aN`D\\1`;\\NlD`1T;XNWEe1`<K5K4M4K5K4L5L2M3M2N3N2M3M3M3N2M3M3NVU6"}, "label": "zebra in the back side of the image"}]}
{"id": 234653, "image": "COCO_train2014_000000234653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra whose back legs are not shown\"."}], "task": "refering", "answer_template": "The \"the zebra whose back legs are not shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 173, 174, 175, 180, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 272, 273, 274, 290, 291, 292, 309, 310], "bbox": [0.0, 0.2184, 0.7723625254582485, 0.98364], "iscrowd": 0, "area": 61940.49455000001, "rle": {"size": [500, 491], "counts": "R9n1P;g2O1O001O1O1N101O1O1O001O1O1O001O1O1N101O1O1O001O1O1O000000000O100000000000000000O100000O10000000000000000000000hGfJV6Z5eIPKV6P5eIZKV6f4dIdKX6\\4cInKX6R4dIWLW6h3gI^LV6b3fIcLY6]3dIfL\\6Z3aIjL^6V3^IoLa6Q3\\IRMd6n2XIWMg6i2TI]Mj6d2RI`Mn6`2mHfMR7_5O001O1O1ON2N2M3N2N2M3N2@`0M3N2O1N2O1004L5K4K?B?A8H1O1N2O1N2N2O1N2O1N010O010O010ON3M2N3M2N3cN\\IeIh6Y6hIeHl6Y7h0N1O2N2N2N2N2N1O2N2N2N2N3M3M3M3M201O1O1N2O1O0O2O1O1O1N101O1O1N10000O1000000O10O1000O10000O10001O0O2O1O001N23L4M4K4M4L3L4M4K4M3M4K3NO010000O100O1000O0100O10000O100O01eMaGaM_8^2gG]MX8c2nGXMR8h2SHSMm7l2ZHnLe7R3aHiL_7V3gHeLY7Z3mH`LS7`3RI\\Ln6c3VIZLj6e3YIYLf6h3]IULc6j3aISL_6l3eIQLZ6o3jInKV6Q4mIlKT6S4PJjKP6U4TJhKk5X4YJeKg5Z4j20O01O1O00O1000000O2O000000000001O00010O00001O0000001O011N1O2O000O2O001O001O001O001N10001O001O001O001O001O100O1O1O100O1O1O2O0O1O1O100O1O1O100O2N1O100O1O1O100O1O1O2O0O100O1O100O1O100O2O0O1O100O1O100O100O2N100O10000O100000001O0000000000000000001O0O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2M2N2N1O2N1O1O2N1O5K4L5K4L5K6J:FR^f1"}, "label": "the zebra whose back legs are not shown"}]}
{"id": 234653, "image": "COCO_train2014_000000234653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra that is not close the wood fence\"."}], "task": "refering", "answer_template": "The \"the zebra that is not close the wood fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 162, 163, 164, 165, 166, 167, 168, 169, 170, 173, 174, 175, 180, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 272, 273, 274, 290, 291, 292, 309, 310], "bbox": [0.0, 0.2184, 0.7723625254582485, 0.98364], "iscrowd": 0, "area": 61940.49455000001, "rle": {"size": [500, 491], "counts": "R9n1P;g2O1O001O1O1N101O1O1O001O1O1O001O1O1N101O1O1O001O1O1O000000000O100000000000000000O100000O10000000000000000000000hGfJV6Z5eIPKV6P5eIZKV6f4dIdKX6\\4cInKX6R4dIWLW6h3gI^LV6b3fIcLY6]3dIfL\\6Z3aIjL^6V3^IoLa6Q3\\IRMd6n2XIWMg6i2TI]Mj6d2RI`Mn6`2mHfMR7_5O001O1O1ON2N2M3N2N2M3N2@`0M3N2O1N2O1004L5K4K?B?A8H1O1N2O1N2N2O1N2O1N010O010O010ON3M2N3M2N3cN\\IeIh6Y6hIeHl6Y7h0N1O2N2N2N2N2N1O2N2N2N2N3M3M3M3M201O1O1N2O1O0O2O1O1O1N101O1O1N10000O1000000O10O1000O10000O10001O0O2O1O001N23L4M4K4M4L3L4M4K4M3M4K3NO010000O100O1000O0100O10000O100O01eMaGaM_8^2gG]MX8c2nGXMR8h2SHSMm7l2ZHnLe7R3aHiL_7V3gHeLY7Z3mH`LS7`3RI\\Ln6c3VIZLj6e3YIYLf6h3]IULc6j3aISL_6l3eIQLZ6o3jInKV6Q4mIlKT6S4PJjKP6U4TJhKk5X4YJeKg5Z4j20O01O1O00O1000000O2O000000000001O00010O00001O0000001O011N1O2O000O2O001O001O001O001N10001O001O001O001O001O100O1O1O100O1O1O2O0O1O1O100O1O1O100O2N1O100O1O1O100O1O1O2O0O100O1O100O1O100O2O0O1O100O1O100O100O2N100O10000O100000001O0000000000000000001O0O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2M2N2N1O2N1O1O2N1O5K4L5K4L5K6J:FR^f1"}, "label": "the zebra that is not close the wood fence"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with the carrots\"."}], "task": "refering", "answer_template": "The \"the bowl with the carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 40, 41, 42, 43, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 128, 129, 130, 147], "bbox": [0.09437999999999999, 0.13933333333333334, 0.4264, 0.6269600000000001], "iscrowd": 0, "area": 16727.74989999999, "rle": {"size": [375, 500], "counts": "X\\a0>W;4L5K4L5K3M3M3M3N3L3N2M3N2M3M3N2M3N1N3N2M3M2O2M3N2M2O2M3M2O2N2N2N1O2M3N2N1O2N2N1N3N2N1O2N2N1N3N2N1O1O1O1N2O1O001O001O001O001O10O01O001O0001O0HkHnKV7R4lHlKT7T4mHkKS7V460001O000001O0001O000001O0001O00010O001O000010O0001O0000100O001O1O010N2O0O2N2O0O2N2O0O2O1N1O2O1N3M2O1N2O1N3M2O1N2N2O1N3M2O1N2O1N3M2N2O1N2N2O2M2N3N1N3M2N3M2N3M2N3L3N3M2N3M2N3L3N3M8H9GlPY3"}, "label": "the bowl with the carrots"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced carrots in a green basket\"."}], "task": "refering", "answer_template": "The \"sliced carrots in a green basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 40, 41, 42, 43, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 128, 129, 130, 147], "bbox": [0.09437999999999999, 0.13933333333333334, 0.4264, 0.6269600000000001], "iscrowd": 0, "area": 16727.74989999999, "rle": {"size": [375, 500], "counts": "X\\a0>W;4L5K4L5K3M3M3M3N3L3N2M3N2M3M3N2M3N1N3N2M3M2O2M3N2M2O2M3M2O2N2N2N1O2M3N2N1O2N2N1N3N2N1O2N2N1N3N2N1O1O1O1N2O1O001O001O001O001O10O01O001O0001O0HkHnKV7R4lHlKT7T4mHkKS7V460001O000001O0001O000001O0001O00010O001O000010O0001O0000100O001O1O010N2O0O2N2O0O2N2O0O2O1N1O2O1N3M2O1N2O1N3M2O1N2N2O1N3M2O1N2O1N3M2N2O1N2N2O2M2N3N1N3M2N3M2N3M2N3L3N3M2N3M2N3L3N3M8H9GlPY3"}, "label": "sliced carrots in a green basket"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green container of food between two other containers\"."}], "task": "refering", "answer_template": "The \"a green container of food between two other containers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 47, 61, 62, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 102, 113, 114, 115, 116, 117, 118, 119, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171], "bbox": [0.28924, 0.18162666666666666, 0.6861, 0.76008], "iscrowd": 0, "area": 28537.698350000006, "rle": {"size": [375, 500], "counts": "\\Ye1g0n:6K5J6J6K4K6K3L3N3M3M2N3M3M2M4M3M2N3M2N2N2M3N2N2N2N2N2N2N2N1N3N1O2N1O2N1O2M2O2N1O2M2O2N1O1N3N1O1O2M2O1O1N3M2O1N3M2O1000000O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O001N2O1O1O1O1O001O1O1O1O1O1O1O010O001O001O001O10O010O10O010O10O10O01O1O0O2O1O1N2O001N2O1O1N101O1N2O1O1N101O1N2O1O001N2O1O1N2O1O1N2O1O1O1O1O1O1N2O1O2N1O1O2N1O1O1O2M2O1N3M2N2N3M2N2O2M2N2O2M2N2O1N3M2N2O2M4L4M2M4L4L3M4L4L3M4hNbF1b9IgFO]9LmFLW9MTGKP9O_1Ki\\i1"}, "label": "a green container of food between two other containers"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bread and potatoes in a green bowl\"."}], "task": "refering", "answer_template": "The \"bread and potatoes in a green bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 47, 61, 62, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 102, 113, 114, 115, 116, 117, 118, 119, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171], "bbox": [0.28924, 0.18162666666666666, 0.6861, 0.76008], "iscrowd": 0, "area": 28537.698350000006, "rle": {"size": [375, 500], "counts": "\\Ye1g0n:6K5J6J6K4K6K3L3N3M3M2N3M3M2M4M3M2N3M2N2N2M3N2N2N2N2N2N2N2N1N3N1O2N1O2N1O2M2O2N1O2M2O2N1O1N3N1O1O2M2O1O1N3M2O1N3M2O1000000O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O001N2O1O1O1O1O001O1O1O1O1O1O1O010O001O001O001O10O010O10O010O10O10O01O1O0O2O1O1N2O001N2O1O1N101O1N2O1O1N101O1N2O1O001N2O1O1N2O1O1N2O1O1O1O1O1O1N2O1O2N1O1O2N1O1O1O2M2O1N3M2N2N3M2N2O2M2N2O2M2N2O1N3M2N2O2M4L4M2M4L4L3M4L4L3M4hNbF1b9IgFO]9LmFLW9MTGKP9O_1Ki\\i1"}, "label": "bread and potatoes in a green bowl"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green bowl with carrots\"."}], "task": "refering", "answer_template": "The \"green bowl with carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 75, 76, 77, 78, 92, 93, 94, 95, 96, 110, 111, 112, 113, 128, 129, 130], "bbox": [0.12504, 0.29704, 0.38056, 0.57632], "iscrowd": 0, "area": 8038.972599999999, "rle": {"size": [375, 500], "counts": "XXg03c;2N3M3M3ZOClE`0R:DjE?S:EiE>T:EiE>U:EgE>V:c0N2N100O2N1O101N1O1O2N100O2N1L4E;L4N2N2O10000O100000O010000O1000000O11O00N2M3N2O1O1O1O0O2O10O100O1O100O1O2O0O2N101N101N1O100O1O100O100O1O1O1N2O1N2N2O1N2N2O1N2N2O1N2N3N1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O2M2O1N2N3N1N2O1N3M2O1N3N1NT]a3"}, "label": "green bowl with carrots"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange carrots in a container\"."}], "task": "refering", "answer_template": "The \"orange carrots in a container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 75, 76, 77, 78, 92, 93, 94, 95, 96, 110, 111, 112, 113, 128, 129, 130], "bbox": [0.12504, 0.29704, 0.38056, 0.57632], "iscrowd": 0, "area": 8038.972599999999, "rle": {"size": [375, 500], "counts": "XXg03c;2N3M3M3ZOClE`0R:DjE?S:EiE>T:EiE>U:EgE>V:c0N2N100O2N1O101N1O1O2N100O2N1L4E;L4N2N2O10000O100000O010000O1000000O11O00N2M3N2O1O1O1O0O2O10O100O1O100O1O2O0O2N101N101N1O100O1O100O100O1O1O1N2O1N2N2O1N2N2O1N2N2O1N2N3N1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O2M2O1N2N3N1N2O1N3M2O1N3N1NT]a3"}, "label": "orange carrots in a container"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dish of yogurt and orange slices with a candy\"."}], "task": "refering", "answer_template": "The \"a dish of yogurt and orange slices with a candy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 226, 227, 228, 229], "bbox": [0.52028, 0.3845866666666667, 0.94188, 0.9755466666666668], "iscrowd": 0, "area": 30258.6937, "rle": {"size": [375, 500], "counts": "__o2?X;`0^O9H3M3L4M3M3L5L3M3L3N2N2M3N2N2M3N2N2N2M3N2N2N2M3N2N2N2M3N2N2N1N3N1O2N1N3N1O2N2M2O2N1O2M2O2N1O2M2O2N1N2O2M2O1N2N3N1N2N3N1N2O1N2O1O1O1O1N2O1O2N1N2O2N1O1N101O1O1N2O1O1O1O1O001O1O100O1O1O1O00100O010O0100O0100O010O010O010O01000O0100O0010O01O100O100N101O1N2O1O1N101O1N2O1O1N2O2N1N2O1O1N2O1O1O1N2O1O1N3N1O1N2O1O1N2O2N1N2O1O1N2N3N1N2N2O2M3M2O2M2O2M3M2O2M2O2M2N3N2M2N3N1N3N2M3M3N2M3M3N2M3N2M3M3N1N3M3N3L3N2M4L3N3M2N2M4M5J5J7I?A^]:"}, "label": "a dish of yogurt and orange slices with a candy"}]}
{"id": 38046, "image": "COCO_train2014_000000038046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plastic container filled with oranges\"."}], "task": "refering", "answer_template": "The \"a white plastic container filled with oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 226, 227, 228, 229], "bbox": [0.52028, 0.3845866666666667, 0.94188, 0.9755466666666668], "iscrowd": 0, "area": 30258.6937, "rle": {"size": [375, 500], "counts": "__o2?X;`0^O9H3M3L4M3M3L5L3M3L3N2N2M3N2N2M3N2N2N2M3N2N2N2M3N2N2N2M3N2N2N1N3N1O2N1N3N1O2N2M2O2N1O2M2O2N1O2M2O2N1N2O2M2O1N2N3N1N2N3N1N2O1N2O1O1O1O1N2O1O2N1N2O2N1O1N101O1O1N2O1O1O1O1O001O1O100O1O1O1O00100O010O0100O0100O010O010O010O01000O0100O0010O01O100O100N101O1N2O1O1N101O1N2O1O1N2O2N1N2O1O1N2O1O1O1N2O1O1N3N1O1N2O1O1N2O2N1N2O1O1N2N3N1N2N2O2M3M2O2M2O2M3M2O2M2O2M2N3N2M2N3N1N3N2M3M3N2M3M3N2M3N2M3M3N1N3M3N3L3N2M4L3N3M2N2M4M5J5J7I?A^]:"}, "label": "a white plastic container filled with oranges"}]}
{"id": 414881, "image": "COCO_train2014_000000414881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man drinking coffee and eating a pastry\"."}], "task": "refering", "answer_template": "The \"a man drinking coffee and eating a pastry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 148, 149, 150, 170, 171, 172, 173, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 257, 258, 259, 260, 279, 280, 281, 300, 301, 302, 322, 323, 324, 344, 345, 346], "bbox": [0.6532165605095542, 0.332579185520362, 0.9031050955414014, 0.9865158371040724], "iscrowd": 0, "area": 22674.702650000003, "rle": {"size": [442, 628], "counts": "f\\a52b=6K5J6K5L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4M3L4L4L4eMoLiIU3T6oLfIV3W6nLbIX3\\6kL^IZ3_6jLZI\\3c6hLWI]3g6gLSI]3j6gLSI[3j6iLTIX3j6kLSIW3j6mLSIU3j6oLTIR3j6QMSIQ3k6SMQIo2n6SMPIn2o6TMoHl2Q7VMmHk2R7WMlHi2T7YMjHg2W7n1O001O001O001O010O00001O001O0000O2N1O100O1O1O1O1O2XNVI_Lj6b3f11L3L5K5K4L5K5J5L5K5J510000O10000O1000000O10O10O10000O10000O1O1O100O2N1O2O1N1O2O1N1O2N2O0O2N2N101N2N101N2N1O2O1N1SLWGh2k8QM[Gn2g8jLaGU3_8eLhGX3[8`LmG]3Y9N3M3M3M2N3M3M3M2N3M3L4G8I8G9G9I6M4K5L4L3L5L4L4K4M4LiSj0"}, "label": "a man drinking coffee and eating a pastry"}]}
{"id": 414881, "image": "COCO_train2014_000000414881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in a wooden chair next to some camping supplies\"."}], "task": "refering", "answer_template": "The \"a man sitting in a wooden chair next to some camping supplies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 148, 149, 150, 170, 171, 172, 173, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 257, 258, 259, 260, 279, 280, 281, 300, 301, 302, 322, 323, 324, 344, 345, 346], "bbox": [0.6532165605095542, 0.332579185520362, 0.9031050955414014, 0.9865158371040724], "iscrowd": 0, "area": 22674.702650000003, "rle": {"size": [442, 628], "counts": "f\\a52b=6K5J6K5L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4M3L4L4L4eMoLiIU3T6oLfIV3W6nLbIX3\\6kL^IZ3_6jLZI\\3c6hLWI]3g6gLSI]3j6gLSI[3j6iLTIX3j6kLSIW3j6mLSIU3j6oLTIR3j6QMSIQ3k6SMQIo2n6SMPIn2o6TMoHl2Q7VMmHk2R7WMlHi2T7YMjHg2W7n1O001O001O001O010O00001O001O0000O2N1O100O1O1O1O1O2XNVI_Lj6b3f11L3L5K5K4L5K5J5L5K5J510000O10000O1000000O10O10O10000O10000O1O1O100O2N1O2O1N1O2O1N1O2N2O0O2N2N101N2N101N2N1O2O1N1SLWGh2k8QM[Gn2g8jLaGU3_8eLhGX3[8`LmG]3Y9N3M3M3M2N3M3M3M2N3M3L4G8I8G9G9I6M4K5L4L3L5L4L4K4M4LiSj0"}, "label": "a man sitting in a wooden chair next to some camping supplies"}]}
{"id": 414881, "image": "COCO_train2014_000000414881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady in a white dress and dark coat\"."}], "task": "refering", "answer_template": "The \"lady in a white dress and dark coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 96, 97, 118, 119, 120, 140, 141, 142, 162, 163, 164, 184, 185, 186, 206, 207, 208, 228, 229, 230, 250, 251, 252], "bbox": [0.3574203821656051, 0.22617647058823528, 0.5054299363057325, 0.7205656108597285], "iscrowd": 0, "area": 12909.9061, "rle": {"size": [442, 628], "counts": "dYQ33d=4M3M3M3M3M3L5L3M3M3M3N6J7I6J6Je0[OY1gN6iFSLV8i4O1O1N3N1O1O1O1O2N1VOgJRIZ5m6oJjHR5U7WKbHk4\\7j0O1O1O2N1O1O1O1O1O1O1O1O1N2O11O2N1N3N2N2N1O2N3M3M4L3M3M4K3N1O2N1O2N1O2N1O2N1O2M2O2N1O3M2N3M2iLTG\\1o8`NYGY1j8cN^GV1d8gNdGQ1`8kNiGm0Y8POoGi0T8SOTHf0o7VOYHc0i7ZO_H?d7]OeH;^7AjH8X7EPI4S7HUI1m6MPUV4"}, "label": "lady in a white dress and dark coat"}]}
{"id": 414881, "image": "COCO_train2014_000000414881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white photo of a lady standing looking at the photographer\"."}], "task": "refering", "answer_template": "The \"a black and white photo of a lady standing looking at the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 96, 97, 118, 119, 120, 140, 141, 142, 162, 163, 164, 184, 185, 186, 206, 207, 208, 228, 229, 230, 250, 251, 252], "bbox": [0.3574203821656051, 0.22617647058823528, 0.5054299363057325, 0.7205656108597285], "iscrowd": 0, "area": 12909.9061, "rle": {"size": [442, 628], "counts": "dYQ33d=4M3M3M3M3M3L5L3M3M3M3N6J7I6J6Je0[OY1gN6iFSLV8i4O1O1N3N1O1O1O1O2N1VOgJRIZ5m6oJjHR5U7WKbHk4\\7j0O1O1O2N1O1O1O1O1O1O1O1O1N2O11O2N1N3N2N2N1O2N3M3M4L3M3M4K3N1O2N1O2N1O2N1O2N1O2M2O2N1O3M2N3M2iLTG\\1o8`NYGY1j8cN^GV1d8gNdGQ1`8kNiGm0Y8POoGi0T8SOTHf0o7VOYHc0i7ZO_H?d7]OeH;^7AjH8X7EPI4S7HUI1m6MPUV4"}, "label": "a black and white photo of a lady standing looking at the photographer"}]}
{"id": 390310, "image": "COCO_train2014_000000390310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle halfway out of the frame\"."}], "task": "refering", "answer_template": "The \"the motorcycle halfway out of the frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 270, 271, 272, 273, 293, 294, 295, 296, 317, 318, 319, 320, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390, 410, 411, 412, 413, 433, 434, 435, 436, 457, 458, 459, 480, 481, 503], "bbox": [0.761796875, 0.44232812499999996, 0.9955, 0.9209687499999999], "iscrowd": 0, "area": 23626.04559999999, "rle": {"size": [640, 640], "counts": "dZa95gc05L4K6K4DYOY]Ol0fb0UOV]OP1hb05O1O2N1O2N1O1O2N1O1O1O001O1O1O001O1O1O1O0n_O]Ng=d1mAgNS>Z1bAQO]>P1^AUOa>l0\\AXOb>h0]AZOb>g0[A]Od>c0YA@f>a0XABf>>XAEc>`0[AC[>f0bA^O[>d0cA^OZ>d0dA@Y>b0eABV>a0gACV>>hAGS><hAJU>T3K3N1N3N1N3N2M2010O10O10O010O02O2N3M200O10000O2O5J6d0]O4KN3H7I8H7H6KO001O000000000000001O00000000000000001O000O1O100O1O101N1O100O1O100O1O101N7I8I6I8H5L2M2N3N1M4K4L5K4L5K4N2O2N1O2N1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O_^1"}, "label": "the motorcycle halfway out of the frame"}]}
{"id": 390310, "image": "COCO_train2014_000000390310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the partially visible back end of a black motorcycle to the right of the blue one\"."}], "task": "refering", "answer_template": "The \"the partially visible back end of a black motorcycle to the right of the blue one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 270, 271, 272, 273, 293, 294, 295, 296, 317, 318, 319, 320, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390, 410, 411, 412, 413, 433, 434, 435, 436, 457, 458, 459, 480, 481, 503], "bbox": [0.761796875, 0.44232812499999996, 0.9955, 0.9209687499999999], "iscrowd": 0, "area": 23626.04559999999, "rle": {"size": [640, 640], "counts": "dZa95gc05L4K6K4DYOY]Ol0fb0UOV]OP1hb05O1O2N1O2N1O1O2N1O1O1O001O1O1O001O1O1O1O0n_O]Ng=d1mAgNS>Z1bAQO]>P1^AUOa>l0\\AXOb>h0]AZOb>g0[A]Od>c0YA@f>a0XABf>>XAEc>`0[AC[>f0bA^O[>d0cA^OZ>d0dA@Y>b0eABV>a0gACV>>hAGS><hAJU>T3K3N1N3N1N3N2M2010O10O10O010O02O2N3M200O10000O2O5J6d0]O4KN3H7I8H7H6KO001O000000000000001O00000000000000001O000O1O100O1O101N1O100O1O100O1O101N7I8I6I8H5L2M2N3N1M4K4L5K4L5K4N2O2N1O2N1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O_^1"}, "label": "the partially visible back end of a black motorcycle to the right of the blue one"}]}
{"id": 390310, "image": "COCO_train2014_000000390310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and black motorcycle with a man riding it\"."}], "task": "refering", "answer_template": "The \"a blue and black motorcycle with a man riding it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 244, 245, 246, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 317, 331, 333, 334, 335, 336, 337, 338, 339, 340, 349, 350, 353, 354, 357, 358, 359, 360, 361, 362, 363, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 512, 513, 514, 515, 517, 518, 519, 522, 523, 524, 525], "bbox": [0.194765625, 0.40745312499999997, 0.8797812500000001, 1.0], "iscrowd": 0, "area": 88790.18714999998, "rle": {"size": [640, 640], "counts": "Sa^2n0Qc05K2N2O1O1O1N2O1O1O1O1O1O1O1O1h^O_Nl`0f0n^O^ORa0c0h^OBXa0?c^OE]a0<^^OHba09Y^OKga06T^OOka0S100f_ORNb>n1o@gNk>Y1g@[OS?e0^@1[?`2J7I5L0O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O0000000001O000000000000000000000000001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1OO1O1O1O1O1O1kNSKWCm4g<WKVCj4h<YKVCh4h<\\KUCe4i<^KUCc4h<bKUC_4i<eKTCZ4l<iKSCV4l<nKQCQ4n<SLPCm3o<WLnBo0H`0Z=dNlBi0O`0U=kNiBb07a0o<QOgB;?a0i<S1\\CkNc<S1bCjN^<T1gCiN[<S1kCiNX<R1mClNU<n0QDoNR<l0SDQOQ<k0RDROQ<k0RDSOP<k0RDROR<j0QDSOR<j0QDSOR<k0PDROS<k0QDROR<j0QDSOR<k0PDROS<k0PDROS<k0PDSOS<j0oCSOT<j0oCTOS<i0PDWOQ<f0RDYOn;d0UD[Ol;b0WD^Oi;`0YD@h;<[DDe;9^DGb;7`DIa;3bDM^;0fDOZ;OhD1X;LkD4V;HkD9V;EkD<U;A[CISOh0c=\\OXC1SOd0e=XOWC9QOa0P`0Em_O<o?In_O9o?Kn_O7n?NP@3l?3P@Nm?7Q@Jk?;R@Gj??S@Bi?c0T@^Oj?f0S@ZOm?j0P@VOo?m0n_OTOQ`0o0m_OQOQ`0S1l_OnNS`0`2O1O1O1N2O1O1O1N2O1O1O1O1N2O1kLgKbFZ4]9iK`FX4^9kK`FV4_9lK^FV4`9nK]FS4b9oK\\FR4c9PL[FT4a9oK[FU4b9mK\\FR4e9PLYFn3i9ULTFk3l9WLRFj3m9XLPFj3o9YLnEh3R:YLlEh3S:ZLkEg3T:\\LiEe3V:]LgEe3X:]LfEd3Y:_LdEb3[:`LcEa3\\:aLaEa3^:bL_E_3`:cL^E^3a:dL[E_3d:dLVE`3i:bLREb3m:`LnDd3Q;_LiDe3n:eLnD_3P;cLkDa3T;f2O1DYG^Eh8_:[G`Ef8]:]GaEe8^:\\GaEe8_:[G`Ef8`:ZG_Eg8a:YG^Eh8b:XG]Ei8c::000000000000000001O03M7I3M1O000000001O00000000O1O1O1O10O]OPGWFQ9i9UGPFk8Q:[GhEe8Y:a01N1O100O2N100O1ML52M3M4L3N2MoNfGbFZ8]9iGbFU8_9lG`FS8`9PH^Fo7b9SH^Fk7b9VH^Fi7c9WH]Fh7d9XH\\Fg7e9YH\\Fe7e9\\HZFc7g9]HYFb7i9]HXFb7h9^HXFa7i9_HWFZ7P:fHQFR7V:nHjEj6^:VIbEc6e:]I\\EX6o:gIQEe5d;ZJ\\Df5f;XJ[Dg5f;XJZDh5g;XJYDg5i;XJVDi5j;WJUDi5m;WJRDh5n;ZJPDf5P<\\JnCe5Q<]JnCb5R<`JlC`5T<bJjC^5n0jJk8KVFZ5e0[Ko8n5fF^JX9b5dFdJZ9\\5dFhJZ9X5cFmJ[9n7N2N2M3N2N2N2N2N2N2M3N2N2O100OO2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N4L3M4QJPEf2T;SMoDm2U;lLoDS3T;hLnDX3V;aLmD_3Z;WLjDh3_;lKdDT4e;`K_D_4k;UKWDk4R<iJQDW5X<]JlCb5Y=0001O001O00001O001O00001O001O00001O0cJeAU5d>O2N1O2N1O2N1O2N1O2M2M3M4L3M4L3M4L3M4L3M4L3M4L6J7I7D<C=D<C=D<D<CQc_1"}, "label": "a blue and black motorcycle with a man riding it"}]}
{"id": 390310, "image": "COCO_train2014_000000390310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike on which a lady sites semi nude\"."}], "task": "refering", "answer_template": "The \"bike on which a lady sites semi nude\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 244, 245, 246, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 317, 331, 333, 334, 335, 336, 337, 338, 339, 340, 349, 350, 353, 354, 357, 358, 359, 360, 361, 362, 363, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 512, 513, 514, 515, 517, 518, 519, 522, 523, 524, 525], "bbox": [0.194765625, 0.40745312499999997, 0.8797812500000001, 1.0], "iscrowd": 0, "area": 88790.18714999998, "rle": {"size": [640, 640], "counts": "Sa^2n0Qc05K2N2O1O1O1N2O1O1O1O1O1O1O1O1h^O_Nl`0f0n^O^ORa0c0h^OBXa0?c^OE]a0<^^OHba09Y^OKga06T^OOka0S100f_ORNb>n1o@gNk>Y1g@[OS?e0^@1[?`2J7I5L0O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O0000000001O000000000000000000000000001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1OO1O1O1O1O1O1kNSKWCm4g<WKVCj4h<YKVCh4h<\\KUCe4i<^KUCc4h<bKUC_4i<eKTCZ4l<iKSCV4l<nKQCQ4n<SLPCm3o<WLnBo0H`0Z=dNlBi0O`0U=kNiBb07a0o<QOgB;?a0i<S1\\CkNc<S1bCjN^<T1gCiN[<S1kCiNX<R1mClNU<n0QDoNR<l0SDQOQ<k0RDROQ<k0RDSOP<k0RDROR<j0QDSOR<j0QDSOR<k0PDROS<k0QDROR<j0QDSOR<k0PDROS<k0PDROS<k0PDSOS<j0oCSOT<j0oCTOS<i0PDWOQ<f0RDYOn;d0UD[Ol;b0WD^Oi;`0YD@h;<[DDe;9^DGb;7`DIa;3bDM^;0fDOZ;OhD1X;LkD4V;HkD9V;EkD<U;A[CISOh0c=\\OXC1SOd0e=XOWC9QOa0P`0Em_O<o?In_O9o?Kn_O7n?NP@3l?3P@Nm?7Q@Jk?;R@Gj??S@Bi?c0T@^Oj?f0S@ZOm?j0P@VOo?m0n_OTOQ`0o0m_OQOQ`0S1l_OnNS`0`2O1O1O1N2O1O1O1N2O1O1O1O1N2O1kLgKbFZ4]9iK`FX4^9kK`FV4_9lK^FV4`9nK]FS4b9oK\\FR4c9PL[FT4a9oK[FU4b9mK\\FR4e9PLYFn3i9ULTFk3l9WLRFj3m9XLPFj3o9YLnEh3R:YLlEh3S:ZLkEg3T:\\LiEe3V:]LgEe3X:]LfEd3Y:_LdEb3[:`LcEa3\\:aLaEa3^:bL_E_3`:cL^E^3a:dL[E_3d:dLVE`3i:bLREb3m:`LnDd3Q;_LiDe3n:eLnD_3P;cLkDa3T;f2O1DYG^Eh8_:[G`Ef8]:]GaEe8^:\\GaEe8_:[G`Ef8`:ZG_Eg8a:YG^Eh8b:XG]Ei8c::000000000000000001O03M7I3M1O000000001O00000000O1O1O1O10O]OPGWFQ9i9UGPFk8Q:[GhEe8Y:a01N1O100O2N100O1ML52M3M4L3N2MoNfGbFZ8]9iGbFU8_9lG`FS8`9PH^Fo7b9SH^Fk7b9VH^Fi7c9WH]Fh7d9XH\\Fg7e9YH\\Fe7e9\\HZFc7g9]HYFb7i9]HXFb7h9^HXFa7i9_HWFZ7P:fHQFR7V:nHjEj6^:VIbEc6e:]I\\EX6o:gIQEe5d;ZJ\\Df5f;XJ[Dg5f;XJZDh5g;XJYDg5i;XJVDi5j;WJUDi5m;WJRDh5n;ZJPDf5P<\\JnCe5Q<]JnCb5R<`JlC`5T<bJjC^5n0jJk8KVFZ5e0[Ko8n5fF^JX9b5dFdJZ9\\5dFhJZ9X5cFmJ[9n7N2N2M3N2N2N2N2N2N2M3N2N2O100OO2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N4L3M4QJPEf2T;SMoDm2U;lLoDS3T;hLnDX3V;aLmD_3Z;WLjDh3_;lKdDT4e;`K_D_4k;UKWDk4R<iJQDW5X<]JlCb5Y=0001O001O00001O001O00001O001O00001O0cJeAU5d>O2N1O2N1O2N1O2N1O2M2M3M4L3M4L3M4L3M4L3M4L3M4L6J7I7D<C=D<C=D<D<CQc_1"}, "label": "bike on which a lady sites semi nude"}]}
{"id": 537770, "image": "COCO_train2014_000000537770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in glasses and suit standing at podium\"."}], "task": "refering", "answer_template": "The \"man in glasses and suit standing at podium\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 67, 68, 69, 85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214], "bbox": [0.64042, 0.22169642857142854, 0.96516, 0.9867857142857143], "iscrowd": 0, "area": 24348.96195, "rle": {"size": [336, 500], "counts": "mWY32X:7I8H7I7I8H7I7I8H7I7I8H7I7I8I6O1N3N1bJ]Ll3e3QL]Lo3e3nK\\LQ4f3mK[LS4f3kK[LT4h3hKZLX4g3fKZLY4h3eKYL[4i3bKXL]4j3`KXL`4i3^KXLa4k3\\KVLd4k3YKWLf4k3XKWLg4k3VKVLi4l3UKULk4l3RKVLm4m3PKTLP5f4O100O100O100O100O100O100O100O100O100O100O101N_O\\KXKf3@VMW5VOXKb3KoLm4AWK]36kLa4KXKW3a0fLW45WKR3o5PMPJn2Q6SMoIj2S6XMlIe2U6^MjI_2X6cMgI[2Z6gMeIV2]6lMbIQ2`6PN`IP2^6RNbIn1]6RNdIm1\\6SNeIm1Z6TNfIl1Y6TNhIl1V6UNkIj1U6VNlIj1S6WNmIi1R6WNoIi1P6WNQJh1n5YNSJg1l5YNUJg1j5ZNVJf1l5WNVJg1l5WNUJi1l5UNUJk1l5TNTJl1n5QNSJn1o5PNRJP2o5nMRJR2P6lMPJT2Q6jMPJU2Q6jMPJV2n5kMSJU2k5mMUJS2e4_McK`0IQ2d4bM^K>Oo1c4fMZK<4n1a4iMVK<9k1`4mMRK9?j1^4PNnJ8e0g1]4TNjJ6j0f1[4WNgJ5n0d1Z4UOgKk0X4UOiKl0V4SOkKo0R4QOoKQ1n3PORLQ1l3oNULS1h3mNYLU1d3kN^LU1`3lN`LV1]3jNdLX1Y3hNhLY1W3fNjL\\1S3eNmL]1P3cNQM^1m2bNTM`1i2`NXMb1e2_N[Mb1c2^N^Md1ITMo0X1YOf1]OdMP1f0Dg1ROmMY1<Ei1gNTNc13Ge38[LIf35ZLLf33[LMe32[LOf3OZL2f3NYL3g3MXL4i3KWL5i3JWL7i3IVL8k3GTL:l3FTL:l3FSL<m3CfLJZ36lMdNT2[1b3O00001O00001N10001O00001O0O101O00<D<D=Ci[5"}, "label": "man in glasses and suit standing at podium"}]}
{"id": 537770, "image": "COCO_train2014_000000537770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black colored suit giving speech to the audience\"."}], "task": "refering", "answer_template": "The \"a man with black colored suit giving speech to the audience\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 67, 68, 69, 85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214], "bbox": [0.64042, 0.22169642857142854, 0.96516, 0.9867857142857143], "iscrowd": 0, "area": 24348.96195, "rle": {"size": [336, 500], "counts": "mWY32X:7I8H7I7I8H7I7I8H7I7I8H7I7I8I6O1N3N1bJ]Ll3e3QL]Lo3e3nK\\LQ4f3mK[LS4f3kK[LT4h3hKZLX4g3fKZLY4h3eKYL[4i3bKXL]4j3`KXL`4i3^KXLa4k3\\KVLd4k3YKWLf4k3XKWLg4k3VKVLi4l3UKULk4l3RKVLm4m3PKTLP5f4O100O100O100O100O100O100O100O100O100O100O101N_O\\KXKf3@VMW5VOXKb3KoLm4AWK]36kLa4KXKW3a0fLW45WKR3o5PMPJn2Q6SMoIj2S6XMlIe2U6^MjI_2X6cMgI[2Z6gMeIV2]6lMbIQ2`6PN`IP2^6RNbIn1]6RNdIm1\\6SNeIm1Z6TNfIl1Y6TNhIl1V6UNkIj1U6VNlIj1S6WNmIi1R6WNoIi1P6WNQJh1n5YNSJg1l5YNUJg1j5ZNVJf1l5WNVJg1l5WNUJi1l5UNUJk1l5TNTJl1n5QNSJn1o5PNRJP2o5nMRJR2P6lMPJT2Q6jMPJU2Q6jMPJV2n5kMSJU2k5mMUJS2e4_McK`0IQ2d4bM^K>Oo1c4fMZK<4n1a4iMVK<9k1`4mMRK9?j1^4PNnJ8e0g1]4TNjJ6j0f1[4WNgJ5n0d1Z4UOgKk0X4UOiKl0V4SOkKo0R4QOoKQ1n3PORLQ1l3oNULS1h3mNYLU1d3kN^LU1`3lN`LV1]3jNdLX1Y3hNhLY1W3fNjL\\1S3eNmL]1P3cNQM^1m2bNTM`1i2`NXMb1e2_N[Mb1c2^N^Md1ITMo0X1YOf1]OdMP1f0Dg1ROmMY1<Ei1gNTNc13Ge38[LIf35ZLLf33[LMe32[LOf3OZL2f3NYL3g3MXL4i3KWL5i3JWL7i3IVL8k3GTL:l3FTL:l3FSL<m3CfLJZ36lMdNT2[1b3O00001O00001N10001O00001O0O101O00<D<D=Ci[5"}, "label": "a man with black colored suit giving speech to the audience"}]}
{"id": 537770, "image": "COCO_train2014_000000537770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in neutral colored suitcoat\"."}], "task": "refering", "answer_template": "The \"man in neutral colored suitcoat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 56, 57, 58, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 204], "bbox": [0.0, 0.1676190476190476, 0.35648, 0.9847619047619047], "iscrowd": 0, "area": 36125.4756, "rle": {"size": [336, 500], "counts": "W6:n9f0oNe1hN9H7K6K4L5L4M2M4L4L3M3M2N2N2O1N2N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O001N2O001O0O2O001N101O0O2O001O1N1N3N2M2O2SOTJRMo5f2^JQMe5l2dJjLa5S3o0N3N2M2N3M3N1N3M3N1N3O1O0CRHXNo7g1SHWNn7h1THVNl7j1VHTNk7k1WHSNj7l1XHRNh7n1=00O100000000000O01001O0000001O000000001O0000001O000000001O000000001O0000001N2O001O1N2N101O1O001O1O001O1O001O010O1O2O2M3N5J9G8I2M3M4L3M4L3M3M4L3M3M4K4M4L3M3M4L3M3M7I6J5K5K4L5K4L4L3M3M3M3M4L3M3L4J7I6J6cNUHJ^eY3"}, "label": "man in neutral colored suitcoat"}]}
{"id": 537770, "image": "COCO_train2014_000000537770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the lighter colored suit\"."}], "task": "refering", "answer_template": "The \"the person in the lighter colored suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 56, 57, 58, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 204], "bbox": [0.0, 0.1676190476190476, 0.35648, 0.9847619047619047], "iscrowd": 0, "area": 36125.4756, "rle": {"size": [336, 500], "counts": "W6:n9f0oNe1hN9H7K6K4L5L4M2M4L4L3M3M2N2N2O1N2N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O001N2O001O0O2O001N101O0O2O001O1N1N3N2M2O2SOTJRMo5f2^JQMe5l2dJjLa5S3o0N3N2M2N3M3N1N3M3N1N3O1O0CRHXNo7g1SHWNn7h1THVNl7j1VHTNk7k1WHSNj7l1XHRNh7n1=00O100000000000O01001O0000001O000000001O0000001O000000001O000000001O0000001N2O001O1N2N101O1O001O1O001O1O001O010O1O2O2M3N5J9G8I2M3M4L3M4L3M3M4L3M3M4K4M4L3M3M4L3M3M7I6J5K5K4L5K4L4L3M3M3M3M4L3M3L4J7I6J6cNUHJ^eY3"}, "label": "the person in the lighter colored suit"}]}
{"id": 365739, "image": "COCO_train2014_000000365739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with the potbelly\"."}], "task": "refering", "answer_template": "The \"the person with the potbelly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 163, 164, 165, 186, 187, 188, 189, 209, 210, 211, 232, 233, 234, 235, 255, 256, 257], "bbox": [0.0044843750000000005, 0.002453271028037383, 0.24685937500000002, 0.7787383177570094], "iscrowd": 0, "area": 30929.861599999997, "rle": {"size": [428, 640], "counts": "nY1f0e<P1oN;F2M3N2M3M3N2M2O2M3N2M3N2M3N1N3O1N2N2N2N2O1N1O2N2O1N2N2N2N101N2N2N2O1N2N2N1O2O1N2N2N2O1N:F>B:G6I8H7I8I7H7I8H8H7J7I7I6J7I5K2N3M2N2N2N3M2O1N2N2N3M2N2N`0@00001O001O001O0000WMoLXKP3e4XMVKh2f4aMVK^2f4lMTKT2i4XNnJh1n4hNgJX1T5POhJP1U5UOjJj0R5[OmJe0o4AoJ?S5@mJ?V5@hJ`0Z5@dJ`0_5@^J`0d5@ZJ`0i5[OYIeMj0Q3P6aNmI_N1P3\\6bMZJ_OWOQ3e8oLYGS3f8mLXGV3h8jLVGX3i8iLUGY3k8fLSG]3m8cLQG]3P9dLnF]3S9bLlF^3U9cLhF_3X9bLfF^3Z9cLeF^3Z9dLdF\\3\\9;O0O2O001O001O001O1O0O2O1O4L5mNcKmHa4R6bK`I9<Y4n5^LPJf3X4^KYLo0^Og3b3RLiL;Ch3l2eLYMGJh3b2mL]M]O0k3^2oLZMXO8m3X2RMYMTO>n3T2jNlM[1m1gNSN]1h1dNXNo1S1SNmNW8O00000O10001N1N2N2M3N[UY6"}, "label": "the person with the potbelly"}]}
{"id": 365739, "image": "COCO_train2014_000000365739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the corner wearing a green shirt\"."}], "task": "refering", "answer_template": "The \"the man in the corner wearing a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 163, 164, 165, 186, 187, 188, 189, 209, 210, 211, 232, 233, 234, 235, 255, 256, 257], "bbox": [0.0044843750000000005, 0.002453271028037383, 0.24685937500000002, 0.7787383177570094], "iscrowd": 0, "area": 30929.861599999997, "rle": {"size": [428, 640], "counts": "nY1f0e<P1oN;F2M3N2M3M3N2M2O2M3N2M3N2M3N1N3O1N2N2N2N2O1N1O2N2O1N2N2N2N101N2N2N2O1N2N2N1O2O1N2N2N2O1N:F>B:G6I8H7I8I7H7I8H8H7J7I7I6J7I5K2N3M2N2N2N3M2O1N2N2N3M2N2N`0@00001O001O001O0000WMoLXKP3e4XMVKh2f4aMVK^2f4lMTKT2i4XNnJh1n4hNgJX1T5POhJP1U5UOjJj0R5[OmJe0o4AoJ?S5@mJ?V5@hJ`0Z5@dJ`0_5@^J`0d5@ZJ`0i5[OYIeMj0Q3P6aNmI_N1P3\\6bMZJ_OWOQ3e8oLYGS3f8mLXGV3h8jLVGX3i8iLUGY3k8fLSG]3m8cLQG]3P9dLnF]3S9bLlF^3U9cLhF_3X9bLfF^3Z9cLeF^3Z9dLdF\\3\\9;O0O2O001O001O001O1O0O2O1O4L5mNcKmHa4R6bK`I9<Y4n5^LPJf3X4^KYLo0^Og3b3RLiL;Ch3l2eLYMGJh3b2mL]M]O0k3^2oLZMXO8m3X2RMYMTO>n3T2jNlM[1m1gNSN]1h1dNXNo1S1SNmNW8O00000O10001N1N2N2M3N[UY6"}, "label": "the man in the corner wearing a green shirt"}]}
{"id": 365739, "image": "COCO_train2014_000000365739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a plaid shirt holding hands with a man wearing blue jeans\"."}], "task": "refering", "answer_template": "The \"a little boy in a plaid shirt holding hands with a man wearing blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 285, 286, 287, 308, 309, 310, 311, 332], "bbox": [0.356671875, 0.3333411214953271, 0.5479375, 0.9397897196261683], "iscrowd": 0, "area": 16773.568100000004, "rle": {"size": [428, 640], "counts": "j^o25n<`0A?D<I5J3N3L3N3L3N3L3N2N3L3N35J5J2N2N2M5L<C3N2N2M2OAiNhDW1U;lNkDS1S;POmDP1o:TOQEk0m:YOREg0k:\\OUE`2m8bMRGa2j8aMVGa2g8aMXG`2e8bM[Ga2`8aM`Gc2Z8_MfGf2f0_LY6m0PIj2>aL]6g0SIj2;fL]6b0WIh29nLZ6<]If24WMY66aIc23^MW61eIb20eMV6KhIW4V6b1M2N3N1O2N10O0001O0000001O00001O0000001O00000000QO\\JUJe5i5_JUJa5i5dJTJ\\5j5hJTJX5j5mJRJU5k5PKRJQ5m5SKoIm4Q6VKlIk4S6X1O100O12N8G6K1OgJ[Ij3e6SL^In3b6nKaIR4`6iKdIX4[6eKiI[4d7000000O3N4L5J5L4RMZGm0k8lNZGP1k8jNYGT1k8fNZGV1l8cNYGY1l8bN[GW1j8cNgGm0_8lNnGh0W8ROnGj0W8POoGk0]:L4L2N1O1O001O1O1O1O1O10O01Fa]h3"}, "label": "a little boy in a plaid shirt holding hands with a man wearing blue jeans"}]}
{"id": 365739, "image": "COCO_train2014_000000365739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child in a plaid shirt and jeans holding hands with an adult\"."}], "task": "refering", "answer_template": "The \"a small child in a plaid shirt and jeans holding hands with an adult\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 285, 286, 287, 308, 309, 310, 311, 332], "bbox": [0.356671875, 0.3333411214953271, 0.5479375, 0.9397897196261683], "iscrowd": 0, "area": 16773.568100000004, "rle": {"size": [428, 640], "counts": "j^o25n<`0A?D<I5J3N3L3N3L3N3L3N2N3L3N35J5J2N2N2M5L<C3N2N2M2OAiNhDW1U;lNkDS1S;POmDP1o:TOQEk0m:YOREg0k:\\OUE`2m8bMRGa2j8aMVGa2g8aMXG`2e8bM[Ga2`8aM`Gc2Z8_MfGf2f0_LY6m0PIj2>aL]6g0SIj2;fL]6b0WIh29nLZ6<]If24WMY66aIc23^MW61eIb20eMV6KhIW4V6b1M2N3N1O2N10O0001O0000001O00001O0000001O00000000QO\\JUJe5i5_JUJa5i5dJTJ\\5j5hJTJX5j5mJRJU5k5PKRJQ5m5SKoIm4Q6VKlIk4S6X1O100O12N8G6K1OgJ[Ij3e6SL^In3b6nKaIR4`6iKdIX4[6eKiI[4d7000000O3N4L5J5L4RMZGm0k8lNZGP1k8jNYGT1k8fNZGV1l8cNYGY1l8bN[GW1j8cNgGm0_8lNnGh0W8ROnGj0W8POoGk0]:L4L2N1O1O001O1O1O1O1O10O01Fa]h3"}, "label": "a small child in a plaid shirt and jeans holding hands with an adult"}]}
{"id": 365739, "image": "COCO_train2014_000000365739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt holding a child ' s hand\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt holding a child ' s hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 236, 237, 238, 259, 260, 261, 282, 283, 284, 305, 306, 307, 328, 329, 330], "bbox": [0.198375, 0.006752336448598131, 0.38171875, 0.9865186915887849], "iscrowd": 0, "area": 33857.010299999994, "rle": {"size": [428, 640], "counts": "PSe12X=3L4L5N3M3M2N3M3M2Na0_O:F6J6J7I6J6J4L5K4L5K4mFmLe7`4M3M3M4L3M3M3M4KP2QN<D3MO1O1O1O1OkJeH[4\\8J6Jm1SN>B>B>B>B:F0000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000002N3M3M2N3M2N3M2N3M2N3M2N3M3M3jGLhNXOQ1o0M9aNXOZ1b01;\\NCR1;=8YNLZOe1W2dNSNd4h1aKmMj4o1ZKfMQ5:]HR1f2]NW6W1U3I7Io\\U5"}, "label": "a man in a red shirt holding a child ' s hand"}]}
{"id": 365739, "image": "COCO_train2014_000000365739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with red t shirt holding a child in his hand\"."}], "task": "refering", "answer_template": "The \"a man with red t shirt holding a child in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 236, 237, 238, 259, 260, 261, 282, 283, 284, 305, 306, 307, 328, 329, 330], "bbox": [0.198375, 0.006752336448598131, 0.38171875, 0.9865186915887849], "iscrowd": 0, "area": 33857.010299999994, "rle": {"size": [428, 640], "counts": "PSe12X=3L4L5N3M3M2N3M3M2Na0_O:F6J6J7I6J6J4L5K4L5K4mFmLe7`4M3M3M4L3M3M3M4KP2QN<D3MO1O1O1O1OkJeH[4\\8J6Jm1SN>B>B>B>B:F0000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000002N3M3M2N3M2N3M2N3M2N3M2N3M3M3jGLhNXOQ1o0M9aNXOZ1b01;\\NCR1;=8YNLZOe1W2dNSNd4h1aKmMj4o1ZKfMQ5:]HR1f2]NW6W1U3I7Io\\U5"}, "label": "a man with red t shirt holding a child in his hand"}]}
{"id": 562348, "image": "COCO_train2014_000000562348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with dark hair holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a woman with dark hair holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 87, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 215, 218, 219, 220, 221, 222, 223, 224, 225, 237, 238, 239, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.31565625, 0.10077647058823529, 0.946953125, 0.9841882352941176], "iscrowd": 0, "area": 90653.96184999998, "rle": {"size": [425, 640], "counts": "^Ud2b0U<b0N3N2M2N3M2O2M3M2N3N1N3M3M2O2M301N100O2N100O1O2O0O1O2ZF^Mb8b2VGeMj8\\2nFkMS9V2cFRN]9m2001OO1C=G90000000000O100002N4L3M4L3M2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N5K5K5K5K5K5K5K4L1O1O1O1O1O1O1O2NhJjN^NU1`1PO^No0`1VO^Ni0`1\\O^Nc0UOkNVNg0c2=SOQOWNg0d27QOXOWNf0f21oN^OXNe0h2LkNEZNd0i2FiNLZNc0k2@gN2[Nc0l2ZOeN9[Nb0n2TObN`0]Na0o2nN`Nf0^Na0Q3gN]Nn0^N`0S3aN\\NS1^Nb0S3ZN^NV1\\Nh0Q3QNbNZ1VNP1S3eMfN]1QNY1T3YMjNa1kM]1Z3QMjNd1eMa1b3jLgNh1aMc1]2PN1^4JgK5[4IfK6[4IfK6[4JeK5]4IdK5^4JcK5^4JcK5_4JaK5`4JaK5`4JaK5a4I`K5b4K^K5b4J_K5b4J_K5c4I^K7b4I^K6c4I^K6d4H]K8c4G^K8c4H]K7e4G\\K9d4F]K9d4F]K:d4E\\K:e4E\\K:e4E\\K<c4C^K=c4B]K>c4A^K?b4@_K`0b4^O_Kb0a4^O_Kc0`4\\OaKd0`4ZOaKf0_4YObKg0^4YObKg0^4XOcKh0^4VOcKj0]4UOdKl0[4TOeKl0\\4ROeKn0[4QOfKo0Z4oNhKQ1Y4lNiKT1W4jNkKW1U4eNnK[1S4bNoK^1R4_NPLa1Q4[NRLe1o3XNSLh1n3TNULl1d7O1000000O1000000O100000O01000000O1000000O1000000O1000000O1000000O10000000000000000000000O100000000000000000000O10000000000000000001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1nJSMFn28VMEk28ZMEg28^MEc29aMD`29eMD\\29iMDX2:lMCT2;QNBP2<TNAm1<XNAi1<]N@d1>`N_Oa1>dN_O]1>gN@Z1>hNAY1<jNCW1;kNDV19mNFT17oNHQ17ROGo06TOIm04VOKk03WOLm0NVO1n0IUO6P1DRO;S1^OPOa0U1YOmNf0W1TOlNk0Y1nNjNQ1[1`NPO_1T1TNVOk1o0gM[OX2j0YMAf2b5O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O2N2N2N1O2N2N2N1O2N2N2N1N3N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N2N2N2Ncf="}, "label": "a woman with dark hair holding an umbrella"}]}
{"id": 562348, "image": "COCO_train2014_000000562348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older lady wearing a multicolored neck kerchief and holding an umbrella while on a sidewalk\"."}], "task": "refering", "answer_template": "The \"an older lady wearing a multicolored neck kerchief and holding an umbrella while on a sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 87, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 215, 218, 219, 220, 221, 222, 223, 224, 225, 237, 238, 239, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.31565625, 0.10077647058823529, 0.946953125, 0.9841882352941176], "iscrowd": 0, "area": 90653.96184999998, "rle": {"size": [425, 640], "counts": "^Ud2b0U<b0N3N2M2N3M2O2M3M2N3N1N3M3M2O2M301N100O2N100O1O2O0O1O2ZF^Mb8b2VGeMj8\\2nFkMS9V2cFRN]9m2001OO1C=G90000000000O100002N4L3M4L3M2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N5K5K5K5K5K5K5K4L1O1O1O1O1O1O1O2NhJjN^NU1`1PO^No0`1VO^Ni0`1\\O^Nc0UOkNVNg0c2=SOQOWNg0d27QOXOWNf0f21oN^OXNe0h2LkNEZNd0i2FiNLZNc0k2@gN2[Nc0l2ZOeN9[Nb0n2TObN`0]Na0o2nN`Nf0^Na0Q3gN]Nn0^N`0S3aN\\NS1^Nb0S3ZN^NV1\\Nh0Q3QNbNZ1VNP1S3eMfN]1QNY1T3YMjNa1kM]1Z3QMjNd1eMa1b3jLgNh1aMc1]2PN1^4JgK5[4IfK6[4IfK6[4JeK5]4IdK5^4JcK5^4JcK5_4JaK5`4JaK5`4JaK5a4I`K5b4K^K5b4J_K5b4J_K5c4I^K7b4I^K6c4I^K6d4H]K8c4G^K8c4H]K7e4G\\K9d4F]K9d4F]K:d4E\\K:e4E\\K:e4E\\K<c4C^K=c4B]K>c4A^K?b4@_K`0b4^O_Kb0a4^O_Kc0`4\\OaKd0`4ZOaKf0_4YObKg0^4YObKg0^4XOcKh0^4VOcKj0]4UOdKl0[4TOeKl0\\4ROeKn0[4QOfKo0Z4oNhKQ1Y4lNiKT1W4jNkKW1U4eNnK[1S4bNoK^1R4_NPLa1Q4[NRLe1o3XNSLh1n3TNULl1d7O1000000O1000000O100000O01000000O1000000O1000000O1000000O1000000O10000000000000000000000O100000000000000000000O10000000000000000001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1nJSMFn28VMEk28ZMEg28^MEc29aMD`29eMD\\29iMDX2:lMCT2;QNBP2<TNAm1<XNAi1<]N@d1>`N_Oa1>dN_O]1>gN@Z1>hNAY1<jNCW1;kNDV19mNFT17oNHQ17ROGo06TOIm04VOKk03WOLm0NVO1n0IUO6P1DRO;S1^OPOa0U1YOmNf0W1TOlNk0Y1nNjNQ1[1`NPO_1T1TNVOk1o0gM[OX2j0YMAf2b5O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O2N2N2N1O2N2N2N1O2N2N2N1N3N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N2N2N2Ncf="}, "label": "an older lady wearing a multicolored neck kerchief and holding an umbrella while on a sidewalk"}]}
{"id": 21678, "image": "COCO_train2014_000000021678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plane that ' s near a beige building and some smaller vehicles\"."}], "task": "refering", "answer_template": "The \"a plane that ' s near a beige building and some smaller vehicles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 105, 106, 128, 129, 142, 143, 151, 152, 166, 173, 174, 175, 177, 178, 189, 195, 196, 197, 198, 200, 201, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292], "bbox": [0.2071875, 0.2055, 0.789625, 0.7549166666666667], "iscrowd": 0, "area": 29614.767849999993, "rle": {"size": [480, 640], "counts": "f`n14k>6K4K5L5J5L5J53NJ5O2N1O2N1O1O2N101O010O0100M2L5K4L5K4L5K5K4L5K4L2N00O10000000000000000O10000000000000000O100000001O0000000O100000000000000O100000000jAJb=6[BNd=2ZB1e=OXB5g=c0N2O1O1O1O1O1O1O1O1N101O00001O0N2N3L3N2M4O001YCYNS<g1lCZNT<f1lCZNT<g1kCYNU<g1kCYNU<g1jCZNV<g1iCYNW<g1iCYNW<g1hCZNX<g1gCZNX<f1hCZNX<g1gCYNY<g1fCZNZ<g1eCYN[<j1bCVN^<V2001O0000000000001O0001O00000001O0000000000000010O00001O000010O0001O00001O01OO2O00001O00001O0O101O00001O000O1000001O00000O10001O00000O1O1O2N1O100O1O1O2N1O100O1O1O2N1O100O1O1O1O1O100O1O1O1O1M3N2M3N2N2M3N2M3N2N2M3N2M3N2M2O2N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2M3N3L3N2M3N2N2M3N3L3N2M3N2N2M3NaM[HhMa7V2dHjMZ7R2lHnMQ7o1UIQNi6l1\\ITNb6h1cIYNZ6d1lIWLoNh1R7n1TJZLjNh1o6k1]J]LdNh1m6g1eJ`L_Ni1l6b1jJeL[Nh1k6]1PKkLUNh1j6X1WKPMoMh1j6S1[KUMlMi1h6l0bK[MfMi1h6f0hKaMaMh1g6a0nKgM[Mh1g6<SLkMWMi1f66YLQNQMi1e61`LVNkLi1e6LeL[NgLi1c6FlL`NbLj1b6@QMgN]Li1b6[OVMlNXLi1b6UO\\MRORLi1b6oNbMXOmKh1`6lNhM[OiKi1_6fNnMAcKh1`6aNSNG]Kg1a6]NWNLYKe1a6YN[N2UKd1a6TN`N8oJc1Q<]NoCb1Q<_NoC`1Q<`NQD^1o;cNQD\\1o;eNQDZ1o;gNQDX1o;hNSDV1l;lNVDQ1j;POXDm0h;TOYDj0g;WO[Df0e;[O]Db0c;_O_D`0_;AcD=[;EgD9X;HjD6U;KmD3R;NoD1P;OSE0k:1WEMh:4ZEJd:8\\EHc:9]EGb::_EE`:<`EE^:<bED]:=cEC\\:>eEAY:a0gE_OX:b0hE^OW:c0jE]OT:d0lE\\OS:e0mE[OR:f0nEZOQ:g0PFXOn9j0RFVOm9k0SFVOk9k0VFTOi9l0XFTOg9m0YFSOf9m0[FSOd9l0_FSO_9n0bFSO\\9l0fFTOY9k0jFTOU9l0lFTOS9k0oFUOP9k0QGUOn8j0UGVOi8i0YGUOg8l0ZGQOh8n0[GoNf8Q1[GlNg8S1[GkNf8T1\\GiNf8W1a2O010O1O100O100O001O2N2N2N3N1N2N2N2N2N3M2O1N2N2N2N3M2N2OfRo1"}, "label": "a plane that ' s near a beige building and some smaller vehicles"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"neck of giraffe on the right\"."}], "task": "refering", "answer_template": "The \"neck of giraffe on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 180, 202, 203], "bbox": [0.76790625, 0.26067901234567903, 0.99828125, 0.5101028806584362], "iscrowd": 0, "area": 6825.867149999998, "rle": {"size": [486, 640], "counts": "geY73T?2N3L4M2N1O2J6B>N1O2O1N1O2N2N001O000N2O2M2O1O1N3N1N2O1O2M2O1N2O2MkMSCT2l<lMTCU2k<jMVCW2h<iMXCZ2i<1OF9L5N2N1O2N2N2N1O2N2N1O2N2N1O2O10O010000O01000O01000O010O10O10O1000O01000O0010O01O10O01O10O01O10O01O010O1O010O1O010O1O010O00100O00100O01000O0100O010000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000W;"}, "label": "neck of giraffe on the right"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with its head bowed down\"."}], "task": "refering", "answer_template": "The \"a giraffe with its head bowed down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 180, 202, 203], "bbox": [0.76790625, 0.26067901234567903, 0.99828125, 0.5101028806584362], "iscrowd": 0, "area": 6825.867149999998, "rle": {"size": [486, 640], "counts": "geY73T?2N3L4M2N1O2J6B>N1O2O1N1O2N2N001O000N2O2M2O1O1N3N1N2O1O2M2O1N2O2MkMSCT2l<lMTCU2k<jMVCW2h<iMXCZ2i<1OF9L5N2N1O2N2N2N1O2N2N1O2N2N1O2O10O010000O01000O01000O010O10O10O1000O01000O0010O01O10O01O10O01O10O01O010O1O010O1O010O1O010O00100O00100O01000O0100O010000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000W;"}, "label": "a giraffe with its head bowed down"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the lead\"."}], "task": "refering", "answer_template": "The \"the zebra in the lead\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 252, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 389, 390], "bbox": [0.816203125, 0.5759670781893005, 0.9987187500000001, 0.9886625514403292], "iscrowd": 0, "area": 13933.966799999993, "rle": {"size": [486, 640], "counts": "lTh72i>>B=B?@`0L4L4L4M3L3M4M3L4L4M3L4L3N3L4L4M3L4M2N0O100000N2O1O001O1O1O1O1XOg0O2001O00001N101O001O001O001O001O001O001O001O001O00001O001O3M2N3M2N2M4L3N3L2N100O1O100O1OL5J5J6M31M2N3M2N3`NVLbGk3_8YLZGj3e8ZLUGh3l8\\LmFg3S9\\LgFg3Y9]L`Fe3`9_LZFd3f9n0002N1O2N1O2O02O0O1O2M2M3N2N3M2M3N3M0O100T6"}, "label": "the zebra in the lead"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra to the right of the other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra to the right of the other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 252, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 389, 390], "bbox": [0.816203125, 0.5759670781893005, 0.9987187500000001, 0.9886625514403292], "iscrowd": 0, "area": 13933.966799999993, "rle": {"size": [486, 640], "counts": "lTh72i>>B=B?@`0L4L4L4M3L3M4M3L4L4M3L4L3N3L4L4M3L4M2N0O100000N2O1O001O1O1O1O1XOg0O2001O00001N101O001O001O001O001O001O001O001O001O00001O001O3M2N3M2N2M4L3N3L2N100O1O100O1OL5J5J6M31M2N3M2N3`NVLbGk3_8YLZGj3e8ZLUGh3l8\\LmFg3S9\\LgFg3Y9]L`Fe3`9_LZFd3f9n0002N1O2N1O2O02O0O1O2M2M3N2N3M2M3N3M0O100T6"}, "label": "a zebra to the right of the other zebras"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating leaves\"."}], "task": "refering", "answer_template": "The \"a giraffe eating leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 85, 107, 108, 109, 130, 131, 132, 133, 134, 154, 155, 156, 158, 159, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 251, 271, 272, 294, 295], "bbox": [0.5800468750000001, 0.021790123456790125, 0.9958124999999999, 0.7736008230452674], "iscrowd": 0, "area": 27023.306399999998, "rle": {"size": [486, 640], "counts": "iS`51U?3L4L4M3L4M3L4L8Ib0]O=D1N2O1N2FjMaCX2^<jM`CW2_<jM_CX2`<jM^CW2a<8O00100N2N1OO2N1O1O1O100O1O2N1O1O1Ol0UOO01O12N5K5J6K6J\\NUCS1f<mN\\CV1_<iNeCY1U<gNnC\\1m;cNVD`1e;_N^De1];ZNfDh1W;UNlDn1P;PNSES2j:jMZEX2c:eM`E\\2^:cMdE^2[:_MiE`2_;1O01O01O0010O00010O01O01O01O010O000010O01O01O01O010O000010O01O01O01O010O00010O001O01O01O010bDbMQ:^2fEkMZ:U2]ETNd:l1RE]Nn:n2001O00O010O100O100O1O10O0100O100O100O010O1O100O100O10O0kLTEQ2l:nMTES2k:lMWET2i:jMXEW2g:iMZEY2d:eM^Ea2[:^MfEj2R:VMoEP3k9nLWFY3a9gL_F`3[9^LgFh3R9WLoFQ4j8nKWGX4b8gK`G`4Y8_KgGh4R8WKPHP5i7nJYHV5b7jJ^HU5c7jJ_HU5b7iJ_HV5b7jJ_HU5b7iJ`HU5a7kJ_HT5c7jJ_HU5a7jJaHT5a7kJ_HU5a7jJaHT5a7kJ`HS5a7lJ`HT5`7kJbHS5`7lJ`HT5`7kJbHS5`7kJbHS5_7mJaHS5j8N2O0O2O0O1bMkEJU:7QFCo9=XF[Oi9e0]FUOb9l0eFlN\\9T1jFeNW9[1PG^No8c1WGVNj8j1]GoMc8Q2cGhM]8Y2jG_MW8a2W20O100O101O0O10000O10000O10000O10000O10000O2O000O2O1O1N200O1O010O100O100O100O1O10O0100O100O100O1O100O10O0100OUOk0M4N101O001O001O00001O001O001N101O00001O001O2Ng0YOg0YO]U1"}, "label": "a giraffe eating leaves"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe facing the camera\"."}], "task": "refering", "answer_template": "The \"a giraffe facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 85, 107, 108, 109, 130, 131, 132, 133, 134, 154, 155, 156, 158, 159, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 251, 271, 272, 294, 295], "bbox": [0.5800468750000001, 0.021790123456790125, 0.9958124999999999, 0.7736008230452674], "iscrowd": 0, "area": 27023.306399999998, "rle": {"size": [486, 640], "counts": "iS`51U?3L4L4M3L4M3L4L8Ib0]O=D1N2O1N2FjMaCX2^<jM`CW2_<jM_CX2`<jM^CW2a<8O00100N2N1OO2N1O1O1O100O1O2N1O1O1Ol0UOO01O12N5K5J6K6J\\NUCS1f<mN\\CV1_<iNeCY1U<gNnC\\1m;cNVD`1e;_N^De1];ZNfDh1W;UNlDn1P;PNSES2j:jMZEX2c:eM`E\\2^:cMdE^2[:_MiE`2_;1O01O01O0010O00010O01O01O01O010O000010O01O01O01O010O000010O01O01O01O010O00010O001O01O01O010bDbMQ:^2fEkMZ:U2]ETNd:l1RE]Nn:n2001O00O010O100O100O1O10O0100O100O100O010O1O100O100O10O0kLTEQ2l:nMTES2k:lMWET2i:jMXEW2g:iMZEY2d:eM^Ea2[:^MfEj2R:VMoEP3k9nLWFY3a9gL_F`3[9^LgFh3R9WLoFQ4j8nKWGX4b8gK`G`4Y8_KgGh4R8WKPHP5i7nJYHV5b7jJ^HU5c7jJ_HU5b7iJ_HV5b7jJ_HU5b7iJ`HU5a7kJ_HT5c7jJ_HU5a7jJaHT5a7kJ_HU5a7jJaHT5a7kJ`HS5a7lJ`HT5`7kJbHS5`7lJ`HT5`7kJbHS5`7kJbHS5_7mJaHS5j8N2O0O2O0O1bMkEJU:7QFCo9=XF[Oi9e0]FUOb9l0eFlN\\9T1jFeNW9[1PG^No8c1WGVNj8j1]GoMc8Q2cGhM]8Y2jG_MW8a2W20O100O101O0O10000O10000O10000O10000O10000O2O000O2O1O1N200O1O010O100O100O100O1O10O0100O100O100O1O100O10O0100OUOk0M4N101O001O001O00001O001O001N101O00001O001O2Ng0YOg0YO]U1"}, "label": "a giraffe facing the camera"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe in the front\"."}], "task": "refering", "answer_template": "The \"the giraffe in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 52, 53, 54, 55, 75, 76, 77, 78, 79, 101, 102, 124, 125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 269, 287, 288, 290, 292, 310, 311, 315, 333, 356], "bbox": [0.29626562500000003, 0.058559670781893, 0.72703125, 0.9173045267489712], "iscrowd": 0, "area": 28469.64164999999, "rle": {"size": [486, 640], "counts": "[Xj26i>6K6J6L301O10O01O1O10O01O1O010O1O1O010O1O001O1N2O1O1N2O1O1O2M2O1G9000000000000000000000000000000000002N3M3N3O0O100O4LfMTC\\2i<5L4K3M2N2O1N2N2O0O2N2O1N2N2O1N2[FdL\\7]3_HeLd7Z3XHiLj7W3QHlLQ8T3kGmLX8S3cGPM_8P3\\GSMf8o2SGSMP9o2iFTMY9n2`FTMc9m2XFUMj9m2oEVMR:m2gEUM\\:j32N2M3N2N2M210O1O100O01N100O2N100O2O0O101N3M4M3L3N3L4L4M2M3MO1O1O1N1O2N2NkGRKU6l4kIbKi5\\4XJQL\\5m3dJlLd4S3]KmLd4R3\\KmLf4P3[KQMf4g2aKYM`4_2fKbM[4W2kKiMW4S2kKmMV4V2eKkM\\4[2]KeMd4`2ZK[Mh4j2ZKnLh4j3fJnK[5W4hJ`KY5e4iJSKX5Q5kJgJW5]5W24L5L4KO1TMYGTOl8j0]GkNf8T1dG`N_8_1gGYN]8e1gGTN]8k1gGoM\\8P2_2O0O1O1O2N1O1O2N1O1O101N1O000010O2N1O101N1O2N100O2N1O2O0O2N1O101N1O2N100O2N1O2O0O1O2N100O1O1O10hNUEUNi:k1XEUNg:k1YEVNf:j1[EVNe:i1[EXNd:h1]EXNb:h1_EYN_:g1aEZN_:e1bEYN_:g1aEUNc:j1_EQNe:o1\\EmMh:R2XEjMl:V2UEfMn:Z2o001O0000001N1000001O0000001O0000001O0000010O0000001O0000001O1O001O1O1O001O1O1O1OaCROf:n0XEVOg:i0WE[Oh:d0VE@i:?TEFj::TEJk:5SEOl:0RE4m:KPE:n:R2N1O1OZNYEmNf:n0eEmNY:n0SFmNl9n0_FmN`9n0lFlNS9n0YGlNg<Gkeb2"}, "label": "the giraffe in the front"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe in the lead\"."}], "task": "refering", "answer_template": "The \"the giraffe in the lead\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 52, 53, 54, 55, 75, 76, 77, 78, 79, 101, 102, 124, 125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 269, 287, 288, 290, 292, 310, 311, 315, 333, 356], "bbox": [0.29626562500000003, 0.058559670781893, 0.72703125, 0.9173045267489712], "iscrowd": 0, "area": 28469.64164999999, "rle": {"size": [486, 640], "counts": "[Xj26i>6K6J6L301O10O01O1O10O01O1O010O1O1O010O1O001O1N2O1O1N2O1O1O2M2O1G9000000000000000000000000000000000002N3M3N3O0O100O4LfMTC\\2i<5L4K3M2N2O1N2N2O0O2N2O1N2N2O1N2[FdL\\7]3_HeLd7Z3XHiLj7W3QHlLQ8T3kGmLX8S3cGPM_8P3\\GSMf8o2SGSMP9o2iFTMY9n2`FTMc9m2XFUMj9m2oEVMR:m2gEUM\\:j32N2M3N2N2M210O1O100O01N100O2N100O2O0O101N3M4M3L3N3L4L4M2M3MO1O1O1N1O2N2NkGRKU6l4kIbKi5\\4XJQL\\5m3dJlLd4S3]KmLd4R3\\KmLf4P3[KQMf4g2aKYM`4_2fKbM[4W2kKiMW4S2kKmMV4V2eKkM\\4[2]KeMd4`2ZK[Mh4j2ZKnLh4j3fJnK[5W4hJ`KY5e4iJSKX5Q5kJgJW5]5W24L5L4KO1TMYGTOl8j0]GkNf8T1dG`N_8_1gGYN]8e1gGTN]8k1gGoM\\8P2_2O0O1O1O2N1O1O2N1O1O101N1O000010O2N1O101N1O2N100O2N1O2O0O2N1O101N1O2N100O2N1O2O0O1O2N100O1O1O10hNUEUNi:k1XEUNg:k1YEVNf:j1[EVNe:i1[EXNd:h1]EXNb:h1_EYN_:g1aEZN_:e1bEYN_:g1aEUNc:j1_EQNe:o1\\EmMh:R2XEjMl:V2UEfMn:Z2o001O0000001N1000001O0000001O0000001O0000010O0000001O0000001O1O001O1O1O001O1O1O1OaCROf:n0XEVOg:i0WE[Oh:d0VE@i:?TEFj::TEJk:5SEOl:0RE4m:KPE:n:R2N1O1OZNYEmNf:n0eEmNY:n0SFmNl9n0_FmN`9n0lFlNS9n0YGlNg<Gkeb2"}, "label": "the giraffe in the lead"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left\"."}], "task": "refering", "answer_template": "The \"the zebra on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 264, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373], "bbox": [0.116203125, 0.6592798353909466, 0.510515625, 0.9952674897119342], "iscrowd": 0, "area": 21649.941450000006, "rle": {"size": [486, 640], "counts": "jaS11U?6E8I4K5K6J5L5K4L4L5K4L4L5K4L4L5kCeM[;g2TD`Ml;m20000O10000O1000000O1000000O10000O1000000O10O10O10000000000O100000000000O10000000O01O1O1O1O1O1O1O1O1O0O2M3L4M3M3L400001O00000001O00010O00001O00010O00001O000010O000001O000010O0001O00001O00001O00O1000000000000000000O100000000000000000000O100000000000000000000O100N2O1N2O1N2O1N2O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O10O00001O00001O0000000000000000001O00000000J52M3N2N3M2N2N2N2N3M2N2M3N3M2N2N2N2O1N1O2O0O2N102M2N2O2M20001O01O0010O00010O00O100O001O1O01N1O1POP1I8HmWd4"}, "label": "the zebra on the left"}]}
{"id": 554168, "image": "COCO_train2014_000000554168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the entire side view of a zebra\"."}], "task": "refering", "answer_template": "The \"the entire side view of a zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 264, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373], "bbox": [0.116203125, 0.6592798353909466, 0.510515625, 0.9952674897119342], "iscrowd": 0, "area": 21649.941450000006, "rle": {"size": [486, 640], "counts": "jaS11U?6E8I4K5K6J5L5K4L4L5K4L4L5K4L4L5kCeM[;g2TD`Ml;m20000O10000O1000000O1000000O10000O1000000O10O10O10000000000O100000000000O10000000O01O1O1O1O1O1O1O1O1O0O2M3L4M3M3L400001O00000001O00010O00001O00010O00001O000010O000001O000010O0001O00001O00001O00O1000000000000000000O100000000000000000000O100000000000000000000O100N2O1N2O1N2O1N2O1O100O1O100O1O100O1O100O1O100O00100O1O100O1O100O1O10O00001O00001O0000000000000000001O00000000J52M3N2N3M2N2N2N2N3M2N2M3N3M2N2N2N2O1N1O2O0O2N102M2N2O2M20001O01O0010O00010O00O100O001O1O01N1O1POP1I8HmWd4"}, "label": "the entire side view of a zebra"}]}
{"id": 144574, "image": "COCO_train2014_000000144574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue dress\"."}], "task": "refering", "answer_template": "The \"a woman in a blue dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 129, 130, 131, 132, 146, 147, 148, 149, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 248, 249, 250, 251, 252, 253, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 351, 352, 353, 354, 368, 369, 370, 371, 386, 387, 388], "bbox": [0.5926041666666666, 0.283765625, 0.9271874999999999, 1.0], "iscrowd": 0, "area": 49469.64365000003, "rle": {"size": [640, 480], "counts": "Uka5;^c0;G6H7l]OoNb`0h1]_OZNY`0P2f_OPNV`0T2i_OmMS`0X2l_OhMP`0]2n_OdMo?_2Q@aMk?d2S@]Mk?e2U@[Mj?f2U@ZMj?h2U@YM68k<CPDm2oNYMMc0Q=XORDl2oNYMIj0T=TOQDi2QO[MDP1W=oNQDf2TOiNi<dNPDd2UOkNc<iNVD\\2VOnN[<PO]DQ2YOPOV<TO^Dn1ZO_1f9gLmFm1ZOg1]9^LYGl1VOT2U9RLcG^7V8cHiG^7V8bHiG`7V8aHiG_7V8cHhG^7X8cHgG]7Y8cHfG^7Z8cHeG]7Z8eHeG[7[8eHdG\\7\\8eHcG[7]8fHbGZ7^8fHaG[7^8gHaGY7^8hHbGX7\\8kHbGV7\\8lHdGT7Z8nHeGS7Y8oHgGQ7W8QIhGP7V8RIjGn6U8SIkGm6MSGl7Q2UH]7f7dHZHc7_7f2I5J7J5J7J4L5J6K4L5K5K3M0000000001O001O001O001O00001O001O001O1O1@ZJdBf5U=cJiB^5P=hJPCZ5h<mJWCV5_<RK`CS5c1gIP9d7oF]HV9^7jFbHY9[7fFfH\\9Y7cFhH^9V7bFjH`9T7_FmHc9Q7\\FPIf9n6ZFSIg9l6WFUIk9i6TFXIn9f6RFZIQ:c6nE_IT:_6jEbIY:[6gEeI[:Y6cEjI_:S6`EnIc:o5\\ERJf:l5XEWJg:i5XEXJh:h5VEZJj:f5UE[Jk:e5SE^Jl:g5lD\\JT;h5cD]Jn8nNSIj6fM]JU9mNRIR8n6QHoHo7P7UHmHl7R7WHkHi7T7ZHjHf7V7\\HhHe7b6kGQHa0\\1d7_6WHmG7b1b7]6\\IbIe6Y6aIeI_6W6fIhIZ6S6oIiIR6S6TJjIl5S6[JiIe5U6`JhIa5S6iJgIW5T6^4I8I6M3N2N2O1M3N200O23M3M2NN2N2N2N2N1O2N2N1N3N1O2N1`JnAP5T>mJSBm4o=PK\\Bf4f=WKfB^4j>iMh@nNk?8b@BTb0I6J5KV]e0"}, "label": "a woman in a blue dress"}]}
{"id": 144574, "image": "COCO_train2014_000000144574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the dress\"."}], "task": "refering", "answer_template": "The \"the woman in the dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 129, 130, 131, 132, 146, 147, 148, 149, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 248, 249, 250, 251, 252, 253, 265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 351, 352, 353, 354, 368, 369, 370, 371, 386, 387, 388], "bbox": [0.5926041666666666, 0.283765625, 0.9271874999999999, 1.0], "iscrowd": 0, "area": 49469.64365000003, "rle": {"size": [640, 480], "counts": "Uka5;^c0;G6H7l]OoNb`0h1]_OZNY`0P2f_OPNV`0T2i_OmMS`0X2l_OhMP`0]2n_OdMo?_2Q@aMk?d2S@]Mk?e2U@[Mj?f2U@ZMj?h2U@YM68k<CPDm2oNYMMc0Q=XORDl2oNYMIj0T=TOQDi2QO[MDP1W=oNQDf2TOiNi<dNPDd2UOkNc<iNVD\\2VOnN[<PO]DQ2YOPOV<TO^Dn1ZO_1f9gLmFm1ZOg1]9^LYGl1VOT2U9RLcG^7V8cHiG^7V8bHiG`7V8aHiG_7V8cHhG^7X8cHgG]7Y8cHfG^7Z8cHeG]7Z8eHeG[7[8eHdG\\7\\8eHcG[7]8fHbGZ7^8fHaG[7^8gHaGY7^8hHbGX7\\8kHbGV7\\8lHdGT7Z8nHeGS7Y8oHgGQ7W8QIhGP7V8RIjGn6U8SIkGm6MSGl7Q2UH]7f7dHZHc7_7f2I5J7J5J7J4L5J6K4L5K5K3M0000000001O001O001O001O00001O001O001O1O1@ZJdBf5U=cJiB^5P=hJPCZ5h<mJWCV5_<RK`CS5c1gIP9d7oF]HV9^7jFbHY9[7fFfH\\9Y7cFhH^9V7bFjH`9T7_FmHc9Q7\\FPIf9n6ZFSIg9l6WFUIk9i6TFXIn9f6RFZIQ:c6nE_IT:_6jEbIY:[6gEeI[:Y6cEjI_:S6`EnIc:o5\\ERJf:l5XEWJg:i5XEXJh:h5VEZJj:f5UE[Jk:e5SE^Jl:g5lD\\JT;h5cD]Jn8nNSIj6fM]JU9mNRIR8n6QHoHo7P7UHmHl7R7WHkHi7T7ZHjHf7V7\\HhHe7b6kGQHa0\\1d7_6WHmG7b1b7]6\\IbIe6Y6aIeI_6W6fIhIZ6S6oIiIR6S6TJjIl5S6[JiIe5U6`JhIa5S6iJgIW5T6^4I8I6M3N2N2O1M3N200O23M3M2NN2N2N2N2N1O2N2N1N3N1O2N1`JnAP5T>mJSBm4o=PK\\Bf4f=WKfB^4j>iMh@nNk?8b@BTb0I6J5KV]e0"}, "label": "the woman in the dress"}]}
{"id": 144574, "image": "COCO_train2014_000000144574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt\"."}], "task": "refering", "answer_template": "The \"man in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 141, 142, 143, 157, 158, 159, 160, 174, 175, 176, 177, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 274, 275, 276, 277, 278, 279, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 326, 327, 328, 329, 330, 344, 345, 346, 347, 348, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382], "bbox": [0.15879166666666666, 0.32515625, 0.5243541666666667, 0.983578125], "iscrowd": 0, "area": 48185.428100000005, "rle": {"size": [640, 480], "counts": "Vl_1n2i`0>C=C=C=E8H8I7M3N2M3M3M3M3N2M3M2N3N1N2N3M2N3M2O1N3M2N2N3M2N3M2O1N3M2M4M2N2N3M2jEZHo7i7mGYHR8i7kGYHT8i7jGXHU8k7gGWHX8k7eGWHZ8l7bGVH]8l7aGUH]8n7aGSH^8P8_GQH`8Q8^GPHa8S8\\GnGc8T8[GmGd8U8ZGlGW8e8fG\\Gh7W9VHjFd7l:N2N2O1N2N3N1N2N2O1N2O1N2N2O1N2N2O1N2O1N200O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O11O001O1O001O001O1O001O001O1O001O1O001mFkHY5U7fJPIV5Q7hJUIS5l6lJXIP5h6oJ]Im4d6oJdIl4\\6RKjIj4W6RKQJi4P6TKVJh4j5UK[Ji4f5TK_Ji4a5VKbJh4_5VKdJh4]5UKhJh4X5[JbHoNY2d6V5[JeHmNX2f6S5[JkHjNU2i6Q5[JnHgNU2l6n4[JRIdNS2o6k4[JWIbNQ2Q7i4[JWLc5i3[J\\Lb5f3[J]Lo2WLQN]7mN`Lh2]LYNT7mNhL[2]LfNl6iNPO5_Jo0c6fN`1X1bNbNd1\\1gNUN`1h1^8N2N3M2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N2N3M2M3N_`^4"}, "label": "man in white shirt"}]}
{"id": 144574, "image": "COCO_train2014_000000144574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elderly man wearing white shirt and sunglasses carrying blue tote bag\"."}], "task": "refering", "answer_template": "The \"elderly man wearing white shirt and sunglasses carrying blue tote bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 141, 142, 143, 157, 158, 159, 160, 174, 175, 176, 177, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 274, 275, 276, 277, 278, 279, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 326, 327, 328, 329, 330, 344, 345, 346, 347, 348, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382], "bbox": [0.15879166666666666, 0.32515625, 0.5243541666666667, 0.983578125], "iscrowd": 0, "area": 48185.428100000005, "rle": {"size": [640, 480], "counts": "Vl_1n2i`0>C=C=C=E8H8I7M3N2M3M3M3M3N2M3M2N3N1N2N3M2N3M2O1N3M2N2N3M2N3M2O1N3M2M4M2N2N3M2jEZHo7i7mGYHR8i7kGYHT8i7jGXHU8k7gGWHX8k7eGWHZ8l7bGVH]8l7aGUH]8n7aGSH^8P8_GQH`8Q8^GPHa8S8\\GnGc8T8[GmGd8U8ZGlGW8e8fG\\Gh7W9VHjFd7l:N2N2O1N2N3N1N2N2O1N2O1N2N2O1N2N2O1N2O1N200O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O11O001O1O001O001O1O001O001O1O001O1O001mFkHY5U7fJPIV5Q7hJUIS5l6lJXIP5h6oJ]Im4d6oJdIl4\\6RKjIj4W6RKQJi4P6TKVJh4j5UK[Ji4f5TK_Ji4a5VKbJh4_5VKdJh4]5UKhJh4X5[JbHoNY2d6V5[JeHmNX2f6S5[JkHjNU2i6Q5[JnHgNU2l6n4[JRIdNS2o6k4[JWIbNQ2Q7i4[JWLc5i3[J\\Lb5f3[J]Lo2WLQN]7mN`Lh2]LYNT7mNhL[2]LfNl6iNPO5_Jo0c6fN`1X1bNbNd1\\1gNUN`1h1^8N2N3M2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N2N3M2M3N_`^4"}, "label": "elderly man wearing white shirt and sunglasses carrying blue tote bag"}]}
{"id": 144574, "image": "COCO_train2014_000000144574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue backpack on a guys back\"."}], "task": "refering", "answer_template": "The \"a blue backpack on a guys back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 206, 207, 222, 223, 238, 239, 240, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 308, 309, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378], "bbox": [0.0019375000000000002, 0.4764375, 0.33187500000000003, 1.0], "iscrowd": 0, "area": 23963.653599999998, "rle": {"size": [640, 480], "counts": "\\V1d1V?V3SMm200O10000O1000000O10000O10000O1000000O10O10O10000001O001O001O1O001O001O00O1N2O1O1N2O1O1N2O1O1O1N2O1O1NjH`DP6_;nIiDm5V;RJPEj5n:UJZEf5e:YJaEc5^:[JjE`5T:_JSF]5l9bJ[FY5d9eJcFW5[9iJlFR5S9lJTGP5k8nJ]Gm4b8RKdGj4Z8UKnGf4Q8YKUHb4k7^K[H]4c7cKcHY4\\7gKhHV4W7jKlHT4R7mKQIQ4n6nKVIP4i6PL[Im3c6TL`Ij3_6ULeIi3Z6WLiIg3U6ZLoIc3P6]LSJa3l5^LXJ`3g5`L\\J^3b5cLbJZ3]5fLfJX3Y5gLjJX3T5iLoJU3P5kLSKS3l4mLWKQ3g4oL]Ko2b4QMaKm2^4SMdKl2\\4RMfKn2Y4RMhKn2W4RMkKm2U4RMlKn2S4QMPLn2o3RMRLn2m3RMULm2k3RMVLn2i3QMYLo2f3QM\\Ln2d3QM]Lo2b3PMaLo2^3QMcLo2]3PMeLo2Z3QMgLo2X3PMkLo2U3PMlLP3S3PMoLn2Q3RMPMm2P3TMQMj2P3UMQMi2P3XMQMf2o2[MQMd2P3[MSMb2m2_MTM_2l2aMWM[2k2eMVMY2j2hMXMU2i2kMXMS2i2mMYMP2h2PNYMm1j2SNWMj1j2VNWMh1j2XNXMc1k2^NUM\\1P3dNRMT1T3lNmLn0X3SOiLf0]3YOdLa0a3_OaL:e3E]L4h3MXLLo33o7010O01O10O01O02O0O10000O101O0001O1O1O010O1M2N3MdYX6"}, "label": "a blue backpack on a guys back"}]}
{"id": 144574, "image": "COCO_train2014_000000144574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man ' s duffel bag\"."}], "task": "refering", "answer_template": "The \"man ' s duffel bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 206, 207, 222, 223, 238, 239, 240, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 308, 309, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375, 376, 377, 378], "bbox": [0.0019375000000000002, 0.4764375, 0.33187500000000003, 1.0], "iscrowd": 0, "area": 23963.653599999998, "rle": {"size": [640, 480], "counts": "\\V1d1V?V3SMm200O10000O1000000O10000O10000O1000000O10O10O10000001O001O001O1O001O001O00O1N2O1O1N2O1O1N2O1O1O1N2O1O1NjH`DP6_;nIiDm5V;RJPEj5n:UJZEf5e:YJaEc5^:[JjE`5T:_JSF]5l9bJ[FY5d9eJcFW5[9iJlFR5S9lJTGP5k8nJ]Gm4b8RKdGj4Z8UKnGf4Q8YKUHb4k7^K[H]4c7cKcHY4\\7gKhHV4W7jKlHT4R7mKQIQ4n6nKVIP4i6PL[Im3c6TL`Ij3_6ULeIi3Z6WLiIg3U6ZLoIc3P6]LSJa3l5^LXJ`3g5`L\\J^3b5cLbJZ3]5fLfJX3Y5gLjJX3T5iLoJU3P5kLSKS3l4mLWKQ3g4oL]Ko2b4QMaKm2^4SMdKl2\\4RMfKn2Y4RMhKn2W4RMkKm2U4RMlKn2S4QMPLn2o3RMRLn2m3RMULm2k3RMVLn2i3QMYLo2f3QM\\Ln2d3QM]Lo2b3PMaLo2^3QMcLo2]3PMeLo2Z3QMgLo2X3PMkLo2U3PMlLP3S3PMoLn2Q3RMPMm2P3TMQMj2P3UMQMi2P3XMQMf2o2[MQMd2P3[MSMb2m2_MTM_2l2aMWM[2k2eMVMY2j2hMXMU2i2kMXMS2i2mMYMP2h2PNYMm1j2SNWMj1j2VNWMh1j2XNXMc1k2^NUM\\1P3dNRMT1T3lNmLn0X3SOiLf0]3YOdLa0a3_OaL:e3E]L4h3MXLLo33o7010O01O10O01O02O0O10000O101O0001O1O1O010O1M2N3MdYX6"}, "label": "man ' s duffel bag"}]}
{"id": 488641, "image": "COCO_train2014_000000488641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a mt . zion shirt sits on a couch\"."}], "task": "refering", "answer_template": "The \"a boy in a mt . zion shirt sits on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 190, 191, 192, 212, 213, 214, 215, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 300, 301], "bbox": [0.058203125, 0.4972877358490566, 0.3881875, 0.8955660377358491], "iscrowd": 0, "area": 11814.148549999998, "rle": {"size": [424, 640], "counts": "We?6P=3O1N3N1N2N3N:E3N2M2OO101O1O001O0O2O0O2N1O0O2N1O110O0100O010O1O01O010O001O000001O00000001O01O010O010O00100O010O00100O010O010O010O01O0100O010O10O0100O0fCWOW<i0hCXOW<i0hCXOX<g0hCZOW<l000O100O100O10000O10O001O001O001O101N2O0O0001O001N101O001O0LQOoCQ1l;7O1O2`DdNR;]1nDfNn:\\1QEgNk:Z1UEiNh:X1VEjNi:V1WEkNg:W1XEjNg:V1XElNf:V1WEmNh:S1WEoNg:j1O1N2O1O1N2O1O1O1O1O1O1O1O1O100O001O1O10000O010000000000O1000WO^MlFb2S9bMjF^2T9fMjFZ2R9lMlFS2U9oMiFQ2V9Q10000O01000O100O10O1000O0100O10000O101O2M4M3M3jLdFT2e9iM\\Fn1o9oMQFo1U:hMPFW2b:O0001O100O1O001O01O000O2O1O1FQEYNQ;c1TEZNm:d1=M3N3\\Oc0N2N:F7H4M3MVQR5"}, "label": "a boy in a mt . zion shirt sits on a couch"}]}
{"id": 488641, "image": "COCO_train2014_000000488641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in white shirt sitting on left side of couch\"."}], "task": "refering", "answer_template": "The \"boy in white shirt sitting on left side of couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 190, 191, 192, 212, 213, 214, 215, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 300, 301], "bbox": [0.058203125, 0.4972877358490566, 0.3881875, 0.8955660377358491], "iscrowd": 0, "area": 11814.148549999998, "rle": {"size": [424, 640], "counts": "We?6P=3O1N3N1N2N3N:E3N2M2OO101O1O001O0O2O0O2N1O0O2N1O110O0100O010O1O01O010O001O000001O00000001O01O010O010O00100O010O00100O010O010O010O01O0100O010O10O0100O0fCWOW<i0hCXOW<i0hCXOX<g0hCZOW<l000O100O100O10000O10O001O001O001O101N2O0O0001O001N101O001O0LQOoCQ1l;7O1O2`DdNR;]1nDfNn:\\1QEgNk:Z1UEiNh:X1VEjNi:V1WEkNg:W1XEjNg:V1XElNf:V1WEmNh:S1WEoNg:j1O1N2O1O1N2O1O1O1O1O1O1O1O1O100O001O1O10000O010000000000O1000WO^MlFb2S9bMjF^2T9fMjFZ2R9lMlFS2U9oMiFQ2V9Q10000O01000O100O10O1000O0100O10000O101O2M4M3M3jLdFT2e9iM\\Fn1o9oMQFo1U:hMPFW2b:O0001O100O1O001O01O000O2O1O1FQEYNQ;c1TEZNm:d1=M3N3\\Oc0N2N:F7H4M3MVQR5"}, "label": "boy in white shirt sitting on left side of couch"}]}
{"id": 488641, "image": "COCO_train2014_000000488641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty sofa on the right\"."}], "task": "refering", "answer_template": "The \"the empty sofa on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.8747812500000001, 0.5548349056603774, 1.0, 1.0], "iscrowd": 0, "area": 11730.805500000004, "rle": {"size": [424, 640], "counts": "jUX75Q=9H7I2M2O1O1N2aEMh74PH5n7LjG<T8FdGc0Z8^ObGg0\\8[O`Gi0]8YO`Gk0^8VOUGW1k8c1O10000O100O100O100O01000O100O1M3L4K5L4K5L4L4K5L4K5L4K5O100000000001O0000000000000001O000000000000000000000001O00000000000FeHgJ[7Q5b0H8@SI"}, "label": "the empty sofa on the right"}]}
{"id": 488641, "image": "COCO_train2014_000000488641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an armchair to the right of a brown couch\"."}], "task": "refering", "answer_template": "The \"an armchair to the right of a brown couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.8747812500000001, 0.5548349056603774, 1.0, 1.0], "iscrowd": 0, "area": 11730.805500000004, "rle": {"size": [424, 640], "counts": "jUX75Q=9H7I2M2O1O1N2aEMh74PH5n7LjG<T8FdGc0Z8^ObGg0\\8[O`Gi0]8YO`Gk0^8VOUGW1k8c1O10000O100O100O100O01000O100O1M3L4K5L4K5L4L4K5L4K5L4K5O100000000001O0000000000000001O000000000000000000000001O00000000000FeHgJ[7Q5b0H8@SI"}, "label": "an armchair to the right of a brown couch"}]}
{"id": 488641, "image": "COCO_train2014_000000488641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man lounging on a couch leaning against his left hand with feet on the coffee table\"."}], "task": "refering", "answer_template": "The \"a man lounging on a couch leaning against his left hand with feet on the coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 307, 308, 309], "bbox": [0.34396875, 0.5186320754716981, 0.6565156249999999, 0.9271933962264151], "iscrowd": 0, "area": 13368.357549999997, "rle": {"size": [424, 640], "counts": "V]k29n<6J2O1N3N>B7I4L2001OO1O1O1N2O1OO10000000000000O1O0000O01000O10O1000O0O2M3M3N1010000O10000O01000O1000O001O010O1O001O10O01O1fNnN\\FS1a9SORFX1n9iNoEY1P:hNoEX1R:iNkEY1T:hNkEY1S:jNkEV1U:kNiEW1V:kNhEV1V:n001O0O2O1O001O1O010O100O1O010O10000O10O100000O10O1000O1000O100000000O10O100000O001O001O1O001O1O001O1O001O1O001O001O0000N1N3M3O1N2O1O10000001N1000001O1O1O1O1N3K4K5L4L4K5L4M3O1O2N2O1N2N2O0O2N7J7H2N2N100O2N101N1O100O1O2O0O1O1O101N1O101N2N1O2N1O2N2N1O2N2N2N101OQkj2"}, "label": "a man lounging on a couch leaning against his left hand with feet on the coffee table"}]}
{"id": 488641, "image": "COCO_train2014_000000488641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt holding a remote control\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt holding a remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 307, 308, 309], "bbox": [0.34396875, 0.5186320754716981, 0.6565156249999999, 0.9271933962264151], "iscrowd": 0, "area": 13368.357549999997, "rle": {"size": [424, 640], "counts": "V]k29n<6J2O1N3N>B7I4L2001OO1O1O1N2O1OO10000000000000O1O0000O01000O10O1000O0O2M3M3N1010000O10000O01000O1000O001O010O1O001O10O01O1fNnN\\FS1a9SORFX1n9iNoEY1P:hNoEX1R:iNkEY1T:hNkEY1S:jNkEV1U:kNiEW1V:kNhEV1V:n001O0O2O1O001O1O010O100O1O010O10000O10O100000O10O1000O1000O100000000O10O100000O001O001O1O001O1O001O1O001O1O001O001O0000N1N3M3O1N2O1O10000001N1000001O1O1O1O1N3K4K5L4L4K5L4M3O1O2N2O1N2N2O0O2N7J7H2N2N100O2N101N1O100O1O2O0O1O1O101N1O101N2N1O2N1O2N2N1O2N2N2N101OQkj2"}, "label": "a man in a black shirt holding a remote control"}]}
{"id": 423107, "image": "COCO_train2014_000000423107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blond woman wearing sunglasses and a black shirt\"."}], "task": "refering", "answer_template": "The \"a blond woman wearing sunglasses and a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 275, 292, 293, 294, 295, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.7125, 0.7486416861826698, 1.0, 0.9851756440281031], "iscrowd": 0, "area": 10695.5, "rle": {"size": [427, 640], "counts": "g_n54V=9H4K00001O3hD_Ob9b0YFBf9a0TFCk9c0lEAS:h0\\EAd:c1O100O100N2N2N2M3N2N2M3N2N2O10000001O000000001O000000001O0000001O000000001O000000001O000000002N2N3M2N3N1N3M2N3M2N3[NiDP1Y;`03ZNdDW1^;fNcDZ1`;cN`D]1b;`N_D`1f;000O100000000000000O1000000000000000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000010O0000001O0L4L401O000O101O0000001O000000001O0000001O0000001O0000001O00000000O1M3L4M3M3L4M3M3L4M3M3L4M3M4M2M3N2MH9\\Oe0\\Oe0ZO_F"}, "label": "a blond woman wearing sunglasses and a black shirt"}]}
{"id": 423107, "image": "COCO_train2014_000000423107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman watches kites\"."}], "task": "refering", "answer_template": "The \"a blonde woman watches kites\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 275, 292, 293, 294, 295, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.7125, 0.7486416861826698, 1.0, 0.9851756440281031], "iscrowd": 0, "area": 10695.5, "rle": {"size": [427, 640], "counts": "g_n54V=9H4K00001O3hD_Ob9b0YFBf9a0TFCk9c0lEAS:h0\\EAd:c1O100O100N2N2N2M3N2N2M3N2N2O10000001O000000001O000000001O0000001O000000001O000000001O000000002N2N3M2N3N1N3M2N3M2N3[NiDP1Y;`03ZNdDW1^;fNcDZ1`;cN`D]1b;`N_D`1f;000O100000000000000O1000000000000000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000010O0000001O0L4L401O000O101O0000001O000000001O0000001O0000001O0000001O00000000O1M3L4M3M3L4M3M3L4M3M3L4M3M4M2M3N2MH9\\Oe0\\Oe0ZO_F"}, "label": "a blonde woman watches kites"}]}
{"id": 496839, "image": "COCO_train2014_000000496839.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a green frisbee\"."}], "task": "refering", "answer_template": "The \"a woman with a green frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 87, 88, 89, 100, 101, 102, 103, 113, 114, 115, 116, 126, 127, 128, 129, 140, 141, 142, 154, 155, 167, 168, 180, 181, 193, 194, 207], "bbox": [0.7101066666666668, 0.32133999999999996, 1.0, 0.8404400000000001], "iscrowd": 0, "area": 12786.2697, "rle": {"size": [500, 375], "counts": "YTR43_?8G8H5K4B?E9O1O001O001O0cNlN[Dm0YOeNV<f0^Dd0AbNR<m0XDa0R<CjC<X<FeC:\\<I`C6b<L[C3f<]11N2N2N101N2002M4M3M2N3L4M1O0001O100O1O0010jCTMc;k2\\D]M^;b2aDfMY;Y2gDmMT;S2jDoMV;P2iDSNV;l1iDVNW;i1iDYNV;g1hD[NX;d1gD_NX;`1gDbNY;]1fDfNY;Z1fDhNX;X1gDjNY;nNgD[3X;cLiD^3W;`LjDb3U;\\LlDe3T;YLmDi3R;ULoDl3Q;QLQEQ4n:mKSEU4U;1O2N1O2N101N10102M6WFZKW8l4aGXK]8n4[GVKc8o4UGWKh8i4VG[Ki8c4VGbKg8]4XGgKf8X4YGkKf8T4XGnKi8P4WGQLj8n3UGTLl8@mFb36oLn8_OnF_32TMQ9\\OQG\\3MYMS9ZOSGZ3I]MU9YOTGW3FaMW9WOVGT3BgMY9TOXGR3^OjM]9SOWGP3[OlMa9SOVGZ4m8fKUGU4n8jKTGR4o8mKSGo3P9QLQGk3S9SLoFh3U9RLSGi3_:L4K5L5K4K5L4K5L5M2000bF"}, "label": "a woman with a green frisbee"}]}
{"id": 496839, "image": "COCO_train2014_000000496839.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman throwing a green frisbee\"."}], "task": "refering", "answer_template": "The \"woman throwing a green frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 87, 88, 89, 100, 101, 102, 103, 113, 114, 115, 116, 126, 127, 128, 129, 140, 141, 142, 154, 155, 167, 168, 180, 181, 193, 194, 207], "bbox": [0.7101066666666668, 0.32133999999999996, 1.0, 0.8404400000000001], "iscrowd": 0, "area": 12786.2697, "rle": {"size": [500, 375], "counts": "YTR43_?8G8H5K4B?E9O1O001O001O0cNlN[Dm0YOeNV<f0^Dd0AbNR<m0XDa0R<CjC<X<FeC:\\<I`C6b<L[C3f<]11N2N2N101N2002M4M3M2N3L4M1O0001O100O1O0010jCTMc;k2\\D]M^;b2aDfMY;Y2gDmMT;S2jDoMV;P2iDSNV;l1iDVNW;i1iDYNV;g1hD[NX;d1gD_NX;`1gDbNY;]1fDfNY;Z1fDhNX;X1gDjNY;nNgD[3X;cLiD^3W;`LjDb3U;\\LlDe3T;YLmDi3R;ULoDl3Q;QLQEQ4n:mKSEU4U;1O2N1O2N101N10102M6WFZKW8l4aGXK]8n4[GVKc8o4UGWKh8i4VG[Ki8c4VGbKg8]4XGgKf8X4YGkKf8T4XGnKi8P4WGQLj8n3UGTLl8@mFb36oLn8_OnF_32TMQ9\\OQG\\3MYMS9ZOSGZ3I]MU9YOTGW3FaMW9WOVGT3BgMY9TOXGR3^OjM]9SOWGP3[OlMa9SOVGZ4m8fKUGU4n8jKTGR4o8mKSGo3P9QLQGk3S9SLoFh3U9RLSGi3_:L4K5L5K4K5L4K5L5M2000bF"}, "label": "woman throwing a green frisbee"}]}
{"id": 496839, "image": "COCO_train2014_000000496839.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy dancing with mother at the park\"."}], "task": "refering", "answer_template": "The \"young boy dancing with mother at the park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 145, 146, 147, 157, 158, 159, 160, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 223, 224, 225, 226], "bbox": [0.03885333333333334, 0.55404, 0.43648000000000003, 1.0], "iscrowd": 0, "area": 18372.748499999998, "rle": {"size": [500, 375], "counts": "Pf7:Y?;E;E<D;E<D:G0O1O2N1O1O100O2N1O1O1O101N1O1O1O1O2N100O1O1O2N1O100O1O2N2N3M2N3N1N3M3M2N3N1N3M2O2M3N1N3M2O2M2N3N2M2gNgKUG[4h8iKTGY4HSKf8h0^GX4FTKk8h0\\G^4c8eKZG\\4e8hKWGY4h8Z1O1000001O0O10000000000O100000000O2O000000000000N2L4K5M3001O01O100O100O100O100O1mJkF>1\\1U9RNnF>4[1o8UNQG;6\\1k8VNRG9:^1e8UNVG7=_1^8XNXG4a0a1X8XN[G1e0b1R8YN^GOg0e1k7ZNaGLk0f1P8QNYG3o0h1T9SNSGi1_;N0O100O2N100O100O1O2N1O1O1O1O1O2N1O1O1O2N2N3M3M2N3M3M2N3M2N3M3M2N3MoPW3"}, "label": "young boy dancing with mother at the park"}]}
{"id": 496839, "image": "COCO_train2014_000000496839.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy ready to jump\"."}], "task": "refering", "answer_template": "The \"a boy ready to jump\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 145, 146, 147, 157, 158, 159, 160, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 223, 224, 225, 226], "bbox": [0.03885333333333334, 0.55404, 0.43648000000000003, 1.0], "iscrowd": 0, "area": 18372.748499999998, "rle": {"size": [500, 375], "counts": "Pf7:Y?;E;E<D;E<D:G0O1O2N1O1O100O2N1O1O1O101N1O1O1O1O2N100O1O1O2N1O100O1O2N2N3M2N3N1N3M3M2N3N1N3M2O2M3N1N3M2O2M2N3N2M2gNgKUG[4h8iKTGY4HSKf8h0^GX4FTKk8h0\\G^4c8eKZG\\4e8hKWGY4h8Z1O1000001O0O10000000000O100000000O2O000000000000N2L4K5M3001O01O100O100O100O100O1mJkF>1\\1U9RNnF>4[1o8UNQG;6\\1k8VNRG9:^1e8UNVG7=_1^8XNXG4a0a1X8XN[G1e0b1R8YN^GOg0e1k7ZNaGLk0f1P8QNYG3o0h1T9SNSGi1_;N0O100O2N100O100O1O2N1O1O1O1O1O2N1O1O1O2N2N3M3M2N3M3M2N3M2N3M3M2N3MoPW3"}, "label": "a boy ready to jump"}]}
{"id": 267463, "image": "COCO_train2014_000000267463.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toyota sienna at a stop sign\"."}], "task": "refering", "answer_template": "The \"a toyota sienna at a stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.343265625, 0.6222169811320755, 0.790171875, 1.0], "iscrowd": 0, "area": 38007.722350000004, "rle": {"size": [424, 640], "counts": "U^k2`0n;GbDj0n:]OjDi0Q;i0M3M3M4L3M3N2N2M3N2N3M2M3N1O2N1N3N1O2N1N2O2N1O2M2O2N1O2M2O2N1O2N1O2N1O2N1O2N2N1O2N1O2N2N1O1O1O1O1O100000000O10000000000O101O0000000O100000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000001O000000001O000000001O000000001O0M3N2M3M4L3O1O100O2N100O100O2N100O100O2N100O2O001N2O1N101N2O0O2O1N2O0O2O1N101N2O1N101N2O001N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O01O1O00100O010001N1N;F2M2O1N3M2O1N3L7I;^Ob0^OfTg1"}, "label": "a toyota sienna at a stop sign"}]}
{"id": 349386, "image": "COCO_train2014_000000349386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady learning to surf\"."}], "task": "refering", "answer_template": "The \"lady learning to surf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 236, 237, 238, 258, 259, 260, 281, 282, 283, 305], "bbox": [0.1701875, 0.25, 0.531640625, 0.8730210772833723], "iscrowd": 0, "area": 20890.580199999997, "rle": {"size": [427, 640], "counts": "me]1`0i<4M4K4L4M3L4M1N1O100O1O100O2N1O2M2N3M2N3M2N3M2N3N1N3M4L4N1O2N2O0O2O0O2N2O0O2O0O2N1`H]MX4d2fK]MY4d2eK^MZ4d2dK^M[4b2cK`M\\4b2bK_M^4a2`KaM_4a2_K`M`4`2_KcM`4\\2`KeM_4\\2_KfMa4Y2_KhM`4Y2^KiMa4W2_KkM`4U2^KmMa4S2_KnMa4R2]KPNb4P2^KQNa4o1^KTNa4k1_KVN`4j1_KXN`4h1`KYN`4f1_K\\N`4a1cKaN\\4`1aKbN^4`1`KaN_4a1^K`Nc4a1[K`Nd4b1ZK_Nf4b1XK_Ng4c1VK_Ni4c1UK^Nk4c1SKiMcNXOZ6Q3QKgMgNWOW6T3PKeMjNWOV6U3nJcMoNWOR6X3mJaMSOVOP6Z3kJ`MWOUOm5]3iJ^M]OSOj5a3gJ\\M@SOi5a3fJ\\MCROf5d3eJZMFROe5e3cJXMKROa5h3bJVMNRO`5i3`JUM1QO_5l3^JRM5RO]5m3[JRM9QO[5o3ZJPM<QOZ5R4VJmLb0POW5W4SJhLh0POV5Z4oIfLm0oNT5f5nJYJQ5g5QKWJP5g5Z101O0O2O1O001N101O0O2O1O001N2O1O001N2O1gMPHVOP8i0SHTOo7k0RHTOo7j0^HjNb7V1eH`N_7^1hHYN\\7f1jHQNY7n1oHhMS7Y2SI^MP7a2m1000O0100O10O0100O010O01001O100O1O1O100O1O100O2N100O1O100O1O100001O00001O01O011N1O100O1lNQEJP;2TENm:MWE2j:J[E4f:H^E8b:DaE<`:@dE`0]:\\OeEd0\\:XOhEg0^;10O0010O01O010O0010O01O0010O01O1O001O1O001O100O2N1O1O1O1O00O101O0O100O100O1N2O10000000O1000000O1000002N2M3K4Kaml3"}, "label": "lady learning to surf"}]}
{"id": 349386, "image": "COCO_train2014_000000349386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older lady learning to surf\"."}], "task": "refering", "answer_template": "The \"an older lady learning to surf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 236, 237, 238, 258, 259, 260, 281, 282, 283, 305], "bbox": [0.1701875, 0.25, 0.531640625, 0.8730210772833723], "iscrowd": 0, "area": 20890.580199999997, "rle": {"size": [427, 640], "counts": "me]1`0i<4M4K4L4M3L4M1N1O100O1O100O2N1O2M2N3M2N3M2N3M2N3N1N3M4L4N1O2N2O0O2O0O2N2O0O2O0O2N1`H]MX4d2fK]MY4d2eK^MZ4d2dK^M[4b2cK`M\\4b2bK_M^4a2`KaM_4a2_K`M`4`2_KcM`4\\2`KeM_4\\2_KfMa4Y2_KhM`4Y2^KiMa4W2_KkM`4U2^KmMa4S2_KnMa4R2]KPNb4P2^KQNa4o1^KTNa4k1_KVN`4j1_KXN`4h1`KYN`4f1_K\\N`4a1cKaN\\4`1aKbN^4`1`KaN_4a1^K`Nc4a1[K`Nd4b1ZK_Nf4b1XK_Ng4c1VK_Ni4c1UK^Nk4c1SKiMcNXOZ6Q3QKgMgNWOW6T3PKeMjNWOV6U3nJcMoNWOR6X3mJaMSOVOP6Z3kJ`MWOUOm5]3iJ^M]OSOj5a3gJ\\M@SOi5a3fJ\\MCROf5d3eJZMFROe5e3cJXMKROa5h3bJVMNRO`5i3`JUM1QO_5l3^JRM5RO]5m3[JRM9QO[5o3ZJPM<QOZ5R4VJmLb0POW5W4SJhLh0POV5Z4oIfLm0oNT5f5nJYJQ5g5QKWJP5g5Z101O0O2O1O001N101O0O2O1O001N2O1O001N2O1gMPHVOP8i0SHTOo7k0RHTOo7j0^HjNb7V1eH`N_7^1hHYN\\7f1jHQNY7n1oHhMS7Y2SI^MP7a2m1000O0100O10O0100O010O01001O100O1O1O100O1O100O2N100O1O100O1O100001O00001O01O011N1O100O1lNQEJP;2TENm:MWE2j:J[E4f:H^E8b:DaE<`:@dE`0]:\\OeEd0\\:XOhEg0^;10O0010O01O010O0010O01O0010O01O1O001O1O001O100O2N1O1O1O1O00O101O0O100O100O1N2O10000000O1000000O1000002N2M3K4Kaml3"}, "label": "an older lady learning to surf"}]}
{"id": 349386, "image": "COCO_train2014_000000349386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black swimming in the sea\"."}], "task": "refering", "answer_template": "The \"a man in black swimming in the sea\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 187, 188, 209, 210, 211, 212, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281], "bbox": [0.06314062499999999, 0.49784543325526937, 0.281515625, 0.8644964871194379], "iscrowd": 0, "area": 10483.1494, "rle": {"size": [427, 640], "counts": "hPa01W=9H8G4M3L3N2M3M3N3L3N2M3M3N3L3N2M2OL3M3M4L31001N1001O100O1O1VEcNg9]1WFfNh9Z1UFjNj9V1TFmNl9R1QFROn9n0PFUOo9l0mEXOS:o0cETO\\:n1O1O10O01O100O1O100O1O100O1O1O1O1mNRMhGo2U8\\MbGe2\\8aM^Ga2`8cM]G^2a8fM\\GZ2d8iMYGX2f8W1N2O10000000000OcLaGP2_8nMeGP2[8nMgGR2Y8lMjGS2V8kMlGU2T8iMnGW2R8gMPHY2P8dMTHZ2m7dMUH\\2k7aMXH_2h7_MZHa2f7\\M^Hc2b7[M`He2a7WMbHi2_7TMaHn2`7nLaHT3i82N2N2O10jLTFm2l9QMVFo2k9nLWFR3Q:00O10001O000000001ROSF\\Nm9_1R1K6J5L4M3N3M2O1100O10000001O00000O10001O`DmNg:S1jDYOY;[12M3M4L3M3M4L4L4M3L4L4L4L5K4L4Lb`o5"}, "label": "a man in black swimming in the sea"}]}
{"id": 349386, "image": "COCO_train2014_000000349386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding surfboard\"."}], "task": "refering", "answer_template": "The \"man holding surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 187, 188, 209, 210, 211, 212, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281], "bbox": [0.06314062499999999, 0.49784543325526937, 0.281515625, 0.8644964871194379], "iscrowd": 0, "area": 10483.1494, "rle": {"size": [427, 640], "counts": "hPa01W=9H8G4M3L3N2M3M3N3L3N2M3M3N3L3N2M2OL3M3M4L31001N1001O100O1O1VEcNg9]1WFfNh9Z1UFjNj9V1TFmNl9R1QFROn9n0PFUOo9l0mEXOS:o0cETO\\:n1O1O10O01O100O1O100O1O100O1O1O1O1mNRMhGo2U8\\MbGe2\\8aM^Ga2`8cM]G^2a8fM\\GZ2d8iMYGX2f8W1N2O10000000000OcLaGP2_8nMeGP2[8nMgGR2Y8lMjGS2V8kMlGU2T8iMnGW2R8gMPHY2P8dMTHZ2m7dMUH\\2k7aMXH_2h7_MZHa2f7\\M^Hc2b7[M`He2a7WMbHi2_7TMaHn2`7nLaHT3i82N2N2O10jLTFm2l9QMVFo2k9nLWFR3Q:00O10001O000000001ROSF\\Nm9_1R1K6J5L4M3N3M2O1100O10000001O00000O10001O`DmNg:S1jDYOY;[12M3M4L3M3M4L4L4M3L4L4L4L5K4L4Lb`o5"}, "label": "man holding surfboard"}]}
{"id": 423114, "image": "COCO_train2014_000000423114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"short giraffe standing next to a tall giraffe\"."}], "task": "refering", "answer_template": "The \"short giraffe standing next to a tall giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 209, 210, 211, 233, 234, 235], "bbox": [0.1273125, 0.3078551532033426, 0.2773125, 0.8449303621169916], "iscrowd": 0, "area": 5362.902850000001, "rle": {"size": [359, 640], "counts": "idl0;h:3L4N11000hNCgG<Y8IcG6]8O_G0a85UG0k86jFOf0ZOa7f1[HgN\\7Y1_HoN`7\\2O010O00G:[OhGXNZ8d1oGUNT8f1UHSNn7i1i0OoGXNg6f1WI^Ni6_1XIbNi6\\1XIeNg6Z1YIhNg6V1ZIjNg6T1ZIlNg6R1ZImNg6S1XInNi6R1VImNl6S1SImNn6S1QIlNP7U1oHkNR7U1lHkNV7U1iHkNX7W1eHiN[7]1ZHhNg7\\1oGiNR8[1cGkN^8n1N1O2O1N2NM5J6O16J6J<D=cHaLY6W4O1M3M3MbLbIY2[6iMhI:c0Ab57mI4Z1oNf4o0SJMk75WHGi7<YH\\Ok7c0]HQOg7o0aHfNc7Y1U100O1O1O1K6H9I7L3L5O1O1O01O1N2O22M4L1O10O01O1O100O1O101N1O1OVVR5"}, "label": "short giraffe standing next to a tall giraffe"}]}
{"id": 423114, "image": "COCO_train2014_000000423114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the short giraffe standing beside a taller fellow\"."}], "task": "refering", "answer_template": "The \"the short giraffe standing beside a taller fellow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 209, 210, 211, 233, 234, 235], "bbox": [0.1273125, 0.3078551532033426, 0.2773125, 0.8449303621169916], "iscrowd": 0, "area": 5362.902850000001, "rle": {"size": [359, 640], "counts": "idl0;h:3L4N11000hNCgG<Y8IcG6]8O_G0a85UG0k86jFOf0ZOa7f1[HgN\\7Y1_HoN`7\\2O010O00G:[OhGXNZ8d1oGUNT8f1UHSNn7i1i0OoGXNg6f1WI^Ni6_1XIbNi6\\1XIeNg6Z1YIhNg6V1ZIjNg6T1ZIlNg6R1ZImNg6S1XInNi6R1VImNl6S1SImNn6S1QIlNP7U1oHkNR7U1lHkNV7U1iHkNX7W1eHiN[7]1ZHhNg7\\1oGiNR8[1cGkN^8n1N1O2O1N2NM5J6O16J6J<D=cHaLY6W4O1M3M3MbLbIY2[6iMhI:c0Ab57mI4Z1oNf4o0SJMk75WHGi7<YH\\Ok7c0]HQOg7o0aHfNc7Y1U100O1O1O1K6H9I7L3L5O1O1O01O1N2O22M4L1O10O01O1O100O1O101N1O1OVVR5"}, "label": "the short giraffe standing beside a taller fellow"}]}
{"id": 423114, "image": "COCO_train2014_000000423114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the furthest left that is tall\"."}], "task": "refering", "answer_template": "The \"the giraffe on the furthest left that is tall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 72, 94, 95, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253], "bbox": [0.02384375, 0.15467966573816155, 0.171921875, 0.8616155988857939], "iscrowd": 0, "area": 7861.128549999999, "rle": {"size": [359, 640], "counts": "_a53P;7I6\\LBVLe0f3HiK`0R4NoJh0n4EfI\\1X6j1O100O1O100O1O10GfIeKY6Z4:M4J5[NgIYN^6e1c1N1O2N2N1O2O001NfH_NR5`1jJfNT5Z1iJkNV5T1hJoNX5o0gJUOW5k0fJZOX5f0fJ]OZ5a0dJ[Oc5e0ZJYOk5g0RJYOQ6g0kI[OW6f0eI[O]6e0]I_Oe6a0TIWOY7l0aHWO`7j0\\HXOc7U21O2O13M6J8I:E;F9F7J1OjMQJTOm5k0\\JPOa5n0jJlMYO8j5j1YK_MZOb0Y5o1\\LkMb3T2eLfM[3Z2lL_MT3`2i2N2O0VOiGfNY8X1PH_NR8_1k0N2OQOcNWH]1m7bNQH]1R8cNlG[1X8fNeGY1^8gN`GX1c8hN[GW1i8iNSGV1Q9jNmFT1W9lNgFR1]9oN^FQ1f94KkN_FR1e9QOXFj0m944K;F71K2N2N2N1K55K8H3AaE0TRj5"}, "label": "the giraffe on the furthest left that is tall"}]}
{"id": 423114, "image": "COCO_train2014_000000423114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe that is farthest to the left\"."}], "task": "refering", "answer_template": "The \"the giraffe that is farthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 72, 94, 95, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253], "bbox": [0.02384375, 0.15467966573816155, 0.171921875, 0.8616155988857939], "iscrowd": 0, "area": 7861.128549999999, "rle": {"size": [359, 640], "counts": "_a53P;7I6\\LBVLe0f3HiK`0R4NoJh0n4EfI\\1X6j1O100O1O100O1O10GfIeKY6Z4:M4J5[NgIYN^6e1c1N1O2N2N1O2O001NfH_NR5`1jJfNT5Z1iJkNV5T1hJoNX5o0gJUOW5k0fJZOX5f0fJ]OZ5a0dJ[Oc5e0ZJYOk5g0RJYOQ6g0kI[OW6f0eI[O]6e0]I_Oe6a0TIWOY7l0aHWO`7j0\\HXOc7U21O2O13M6J8I:E;F9F7J1OjMQJTOm5k0\\JPOa5n0jJlMYO8j5j1YK_MZOb0Y5o1\\LkMb3T2eLfM[3Z2lL_MT3`2i2N2O0VOiGfNY8X1PH_NR8_1k0N2OQOcNWH]1m7bNQH]1R8cNlG[1X8fNeGY1^8gN`GX1c8hN[GW1i8iNSGV1Q9jNmFT1W9lNgFR1]9oN^FQ1f94KkN_FR1e9QOXFj0m944K;F71K2N2N2N1K55K8H3AaE0TRj5"}, "label": "the giraffe that is farthest to the left"}]}
{"id": 300239, "image": "COCO_train2014_000000300239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in plaid drinking a beer\"."}], "task": "refering", "answer_template": "The \"a man in plaid drinking a beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 29, 30, 31, 32, 48, 49, 50, 66, 67, 68, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213], "bbox": [0.5881799999999999, 0.015735735735735737, 0.9398799999999998, 1.0], "iscrowd": 0, "area": 35024.15830000002, "rle": {"size": [333, 500], "counts": "jmo21Z:2O1N2O1N2N2O1N2N2O1N1O2O1N2O1N2N2O1N2N2O1cNROYIP1b6WOZIk0a6[O]If0_6_O^Ic0]6D`I=[6IbI9Y6NcI4P5Z1nJgNT4Y2hKiMm3c2PL_Mm3e2QL\\Ml3i2PLYMm3k2PLVMQ1]Nf1Y6UNWJ\\1m5`N\\JX1g5cN`JX1c5cNeJW1^5hNfJR1]5mNgJo0Z5QOhJm0X5SOkJi0V5VOnJg0R5YOQKc0P5]ORKa0n4_OUK=l4BWK<i4DZK8g4H[K6e4I_K3b4M`K1`4ObKN_42n200O10O1000O100000000O1000000O01O1N3N1N2N2O1N3N1N2O1N2O2:E001O10O10O10O10O10O10O01O1O00101N2N3gG[Od6h0WI@c6b0ZI@f6b0UI@k6b0QIAn6b0mH@S7c0hH@W7b0fH@Y7c0bH_O^7c0^H@a7f101O2O3L4L4L2N1O1O001O1O001O10OjNjLUKW3i4QMPKP3n4XMkJh2T5`MeJa2Y5gM`JZ2^5[1O2N2N2O1N1O2N2N2N2N1O2N2N2N2N1O2O1N2N2N2N1O2N2N2N2N1O2N3M7J5J7I6J7I6J7I6J7I6J7I6J7I6K6^Oga9"}, "label": "a man in plaid drinking a beer"}]}
{"id": 300239, "image": "COCO_train2014_000000300239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person that is not a woman\"."}], "task": "refering", "answer_template": "The \"the person that is not a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 29, 30, 31, 32, 48, 49, 50, 66, 67, 68, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213], "bbox": [0.5881799999999999, 0.015735735735735737, 0.9398799999999998, 1.0], "iscrowd": 0, "area": 35024.15830000002, "rle": {"size": [333, 500], "counts": "jmo21Z:2O1N2O1N2N2O1N2N2O1N1O2O1N2O1N2N2O1N2N2O1cNROYIP1b6WOZIk0a6[O]If0_6_O^Ic0]6D`I=[6IbI9Y6NcI4P5Z1nJgNT4Y2hKiMm3c2PL_Mm3e2QL\\Ml3i2PLYMm3k2PLVMQ1]Nf1Y6UNWJ\\1m5`N\\JX1g5cN`JX1c5cNeJW1^5hNfJR1]5mNgJo0Z5QOhJm0X5SOkJi0V5VOnJg0R5YOQKc0P5]ORKa0n4_OUK=l4BWK<i4DZK8g4H[K6e4I_K3b4M`K1`4ObKN_42n200O10O1000O100000000O1000000O01O1N3N1N2N2O1N3N1N2O1N2O2:E001O10O10O10O10O10O10O01O1O00101N2N3gG[Od6h0WI@c6b0ZI@f6b0UI@k6b0QIAn6b0mH@S7c0hH@W7b0fH@Y7c0bH_O^7c0^H@a7f101O2O3L4L4L2N1O1O001O1O001O10OjNjLUKW3i4QMPKP3n4XMkJh2T5`MeJa2Y5gM`JZ2^5[1O2N2N2O1N1O2N2N2N2N1O2N2N2N2N1O2O1N2N2N2N1O2N2N2N2N1O2N3M7J5J7I6J7I6J7I6J7I6J7I6J7I6K6^Oga9"}, "label": "the person that is not a woman"}]}
{"id": 300239, "image": "COCO_train2014_000000300239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl wearing white frock looking at the pizza is on the image\"."}], "task": "refering", "answer_template": "The \"a young girl wearing white frock looking at the pizza is on the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 26, 27, 28, 44, 45, 46, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 134, 135, 136, 137, 152, 153, 154, 155, 173], "bbox": [0.4142, 0.06732732732732734, 0.6555799999999999, 0.7787387387387387], "iscrowd": 0, "area": 19246.888850000007, "rle": {"size": [333, 500], "counts": "\\^S22V:6cFc0d7T1O2O0O1M3J6N2O2M200001O02O4K4TIdMa5c2UJ`Mj5b30000K6F9G9K5J7I6J6M4M2M4L3N2M4N1O2N1O1O2N1O1O1O1O2N1O100O101O000O1000000O1001O0000000000001O0000000000000O2O00100O001O101N1O2N101N2O002N2N1N4M3M4L3N5J9G3M4L3M3L4M2M100O1O100O10O]MhKQOY4m0RLgNP4Y1ZLTNo3k1[2O001N101N2J7H8G9G9^O_jg1"}, "label": "a young girl wearing white frock looking at the pizza is on the image"}]}
{"id": 300239, "image": "COCO_train2014_000000300239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women in center with white dress\"."}], "task": "refering", "answer_template": "The \"a women in center with white dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 26, 27, 28, 44, 45, 46, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 134, 135, 136, 137, 152, 153, 154, 155, 173], "bbox": [0.4142, 0.06732732732732734, 0.6555799999999999, 0.7787387387387387], "iscrowd": 0, "area": 19246.888850000007, "rle": {"size": [333, 500], "counts": "\\^S22V:6cFc0d7T1O2O0O1M3J6N2O2M200001O02O4K4TIdMa5c2UJ`Mj5b30000K6F9G9K5J7I6J6M4M2M4L3N2M4N1O2N1O1O2N1O1O1O1O2N1O100O101O000O1000000O1001O0000000000001O0000000000000O2O00100O001O101N1O2N101N2O002N2N1N4M3M4L3N5J9G3M4L3M3L4M2M100O1O100O10O]MhKQOY4m0RLgNP4Y1ZLTNo3k1[2O001N101N2J7H8G9G9^O_jg1"}, "label": "a women in center with white dress"}]}
{"id": 300239, "image": "COCO_train2014_000000300239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in black holding the pizza\"."}], "task": "refering", "answer_template": "The \"the woman in black holding the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 22, 23, 24, 40, 41, 42, 43, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 170, 171, 172, 181, 182, 183, 184, 199, 200, 201, 202, 203], "bbox": [0.08979999999999999, 0.07192192192192191, 0.60614, 1.0], "iscrowd": 0, "area": 34334.0179, "rle": {"size": [333, 500], "counts": "]n>8f9?D<L4L4L4L4L4M3L4L4L4M3L4M3L4M3L4M3L4M3L4_Oa0^Ob0G9J6J6^OSK^KS5_4?N2M3M3N2M3M3N2N2O1O1O1O100O1O1O1O2N1N2O1O1N200O100O100O100O10000O10000O10000O100N2N2N2J6I7K5LkNaMdJZ2`5gM`JU2c5mM\\Jn1h5SNXJi1l5XNSJf1n5[NRJd1n5^NQJ`1P6bNoI\\1R6eNnIY1S6iNlIU1V6kNjIT1V6nNiIP1X6QOhIm0Y6UOfIi0[6XOeIg0[6[OdIc0^6]ObIa0_6A`I=a6D_I:b6H]I7c6J]I5c6M\\I2d6O\\I0e61ZINf63ZILf66YIIg68YIGg6;XIDh6?VI@k6b0SI]Om6f0gHfNLd0\\7i0bHhN1?\\7W1dHiN[7Y1dHgN[7Z1eHfN[7Z1eHfNZ7\\1eHdNZ7]1fHcNY7_1fHaNY7`1gH`NX7a1hH`NW7a1hH_NW7b1iH^NV7]2O1O1O101N1O1O1O1O2O1N1VO_LgJc3W5bLdJ_3\\5eL_J]3`5gL[J[3c5jLXJW3h5h0O02O3M3L4M3M10N1O2O1N1O2N2N101N2N2N101N2[L^K[1c4`NcK^1_4\\NgKc1Z4WNlKg1V4SNPLk1R4PNSLn1n3nMWLP2k3kMZLT2g3gM^LW2d3eMiLP2Y3kMjMU1X2fNkN;U1ARO;P1@SO`0]5001O0000001O00000010O001O010O001O010O001O010O0010O01O0010O0000010O0000001O0000001O01O00000000000O10000O10000O10000000000000000001O1O010O100O010O10O00001N101N101O1N2O0O2O1N2O1OUjo1"}, "label": "the woman in black holding the pizza"}]}
{"id": 300239, "image": "COCO_train2014_000000300239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two women holding and looking at pizza\"."}], "task": "refering", "answer_template": "The \"two women holding and looking at pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 22, 23, 24, 40, 41, 42, 43, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 170, 171, 172, 181, 182, 183, 184, 199, 200, 201, 202, 203], "bbox": [0.08979999999999999, 0.07192192192192191, 0.60614, 1.0], "iscrowd": 0, "area": 34334.0179, "rle": {"size": [333, 500], "counts": "]n>8f9?D<L4L4L4L4L4M3L4L4L4M3L4M3L4M3L4M3L4M3L4_Oa0^Ob0G9J6J6^OSK^KS5_4?N2M3M3N2M3M3N2N2O1O1O1O100O1O1O1O2N1N2O1O1N200O100O100O100O10000O10000O10000O100N2N2N2J6I7K5LkNaMdJZ2`5gM`JU2c5mM\\Jn1h5SNXJi1l5XNSJf1n5[NRJd1n5^NQJ`1P6bNoI\\1R6eNnIY1S6iNlIU1V6kNjIT1V6nNiIP1X6QOhIm0Y6UOfIi0[6XOeIg0[6[OdIc0^6]ObIa0_6A`I=a6D_I:b6H]I7c6J]I5c6M\\I2d6O\\I0e61ZINf63ZILf66YIIg68YIGg6;XIDh6?VI@k6b0SI]Om6f0gHfNLd0\\7i0bHhN1?\\7W1dHiN[7Y1dHgN[7Z1eHfN[7Z1eHfNZ7\\1eHdNZ7]1fHcNY7_1fHaNY7`1gH`NX7a1hH`NW7a1hH_NW7b1iH^NV7]2O1O1O101N1O1O1O1O2O1N1VO_LgJc3W5bLdJ_3\\5eL_J]3`5gL[J[3c5jLXJW3h5h0O02O3M3L4M3M10N1O2O1N1O2N2N101N2N2N101N2[L^K[1c4`NcK^1_4\\NgKc1Z4WNlKg1V4SNPLk1R4PNSLn1n3nMWLP2k3kMZLT2g3gM^LW2d3eMiLP2Y3kMjMU1X2fNkN;U1ARO;P1@SO`0]5001O0000001O00000010O001O010O001O010O001O010O0010O01O0010O0000010O0000001O0000001O01O00000000000O10000O10000O10000000000000000001O1O010O100O010O10O00001N101N101O1N2O0O2O1N2O1OUjo1"}, "label": "two women holding and looking at pizza"}]}
{"id": 308433, "image": "COCO_train2014_000000308433.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle being ridden by the man\"."}], "task": "refering", "answer_template": "The \"the bicycle being ridden by the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [287, 288, 289, 290, 312, 335, 336, 357, 358, 359, 380, 381, 382, 402, 403, 404, 425, 426, 427, 449, 450, 472, 473, 496], "bbox": [0.483046875, 0.5552564102564103, 0.6549531249999999, 0.9632211538461538], "iscrowd": 0, "area": 9044.472149999998, "rle": {"size": [624, 640], "counts": "odl51]c02N3L3N3M2N3N1N2O2M201YA[Oa:a0bEFX:NRFb0`9]O_Fe0a9[O]Fg0c9YO\\Fi0c9WO]Fi0c9WO\\Fj0d9VO\\Fj0d9VO\\Fj0d9VOZFl0f9TOWFo0i9QOkE[1U:eN`Eg1_:YN`Eh1`:XN_Ei1a:WN_Ei1TM`N]<G^Fl2`9TM`Fn2^9RMaFR3\\9nLdFT3Z9lLeFW3Y9iLgFX3Y9gLfFZ3Z9fLfFZ3Z9fLeF[3[9eLdF\\3\\9cLeF]3[9cLdF^3\\9bLdFY3a9gL^FS3i9mLWFl2P:TMoEf2X:ZMhE_2_:aM`EX2h:hMXEQ2o:oMoDl1X;TNeDh1b;XNZDm1e;RNYDR2f;TNPDQ2PNiM]=j0lC`1[NYNb=g0_CU2Y<\\NSCn1f<e2I7I7I5K5J5LdITDX5h;^JbDg5Y;oIRET6j:dI_E[6a:eI`EZ6`:eIbEY6^:hIbE[1POU3^;_KdEW1YOS3S;fKeES1AQ3j:kKfEP1IP3a:PLgEk02o2W:VLlEd04m2S:`LTF8Oj2U:nLSF1Md2X:[MkE04W2Y:jMbEO:j1\\:WNZEO`0]1]:dNRE0f0o0`:QOjDOm0b0a:_ObD0Q14e:LZD0R?0n@1Q?On@3Q?Mo@c0a>^O^Al0X>TOhAV1n=jNWBQ1i=oNgBa0Y=_OlB<T=DPC8P=HTC4l<LVC2j<NVC2j<OUC1k<OUC1k<OUC0l<0TC0l<0TC0l<0TCOn<0RC0n<0RC0n<1U@Ma22[=;gBCY==jB@W=?kB^OV=b0l20000GmZV4"}, "label": "the bicycle being ridden by the man"}]}
{"id": 308433, "image": "COCO_train2014_000000308433.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle being used by a man\"."}], "task": "refering", "answer_template": "The \"a bicycle being used by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [287, 288, 289, 290, 312, 335, 336, 357, 358, 359, 380, 381, 382, 402, 403, 404, 425, 426, 427, 449, 450, 472, 473, 496], "bbox": [0.483046875, 0.5552564102564103, 0.6549531249999999, 0.9632211538461538], "iscrowd": 0, "area": 9044.472149999998, "rle": {"size": [624, 640], "counts": "odl51]c02N3L3N3M2N3N1N2O2M201YA[Oa:a0bEFX:NRFb0`9]O_Fe0a9[O]Fg0c9YO\\Fi0c9WO]Fi0c9WO\\Fj0d9VO\\Fj0d9VO\\Fj0d9VOZFl0f9TOWFo0i9QOkE[1U:eN`Eg1_:YN`Eh1`:XN_Ei1a:WN_Ei1TM`N]<G^Fl2`9TM`Fn2^9RMaFR3\\9nLdFT3Z9lLeFW3Y9iLgFX3Y9gLfFZ3Z9fLfFZ3Z9fLeF[3[9eLdF\\3\\9cLeF]3[9cLdF^3\\9bLdFY3a9gL^FS3i9mLWFl2P:TMoEf2X:ZMhE_2_:aM`EX2h:hMXEQ2o:oMoDl1X;TNeDh1b;XNZDm1e;RNYDR2f;TNPDQ2PNiM]=j0lC`1[NYNb=g0_CU2Y<\\NSCn1f<e2I7I7I5K5J5LdITDX5h;^JbDg5Y;oIRET6j:dI_E[6a:eI`EZ6`:eIbEY6^:hIbE[1POU3^;_KdEW1YOS3S;fKeES1AQ3j:kKfEP1IP3a:PLgEk02o2W:VLlEd04m2S:`LTF8Oj2U:nLSF1Md2X:[MkE04W2Y:jMbEO:j1\\:WNZEO`0]1]:dNRE0f0o0`:QOjDOm0b0a:_ObD0Q14e:LZD0R?0n@1Q?On@3Q?Mo@c0a>^O^Al0X>TOhAV1n=jNWBQ1i=oNgBa0Y=_OlB<T=DPC8P=HTC4l<LVC2j<NVC2j<OUC1k<OUC1k<OUC0l<0TC0l<0TC0l<0TCOn<0RC0n<0RC0n<1U@Ma22[=;gBCY==jB@W=?kB^OV=b0l20000GmZV4"}, "label": "a bicycle being used by a man"}]}
{"id": 185558, "image": "COCO_train2014_000000185558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in blue t - shirt and white shorts sitting on a park bench\"."}], "task": "refering", "answer_template": "The \"a lady in blue t - shirt and white shorts sitting on a park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 167, 168, 190, 191, 192, 213, 214, 215, 216, 237, 238, 239, 261, 262, 285], "bbox": [0.285703125, 0.44948477751756444, 0.442, 0.8140983606557377], "iscrowd": 0, "area": 5803.90765, "rle": {"size": [427, 640], "counts": "\\a\\29l<8M4L3M4L5K4L4M5UOgNYE`1^:fN_E`1Y:dNeE_1W:dNfE^1X:fNeE[1Z:gNcEZ1\\:hNcEX1\\:l0O100O10001O0O10000O100001O0nMhEQ1Y:nNgER1Z:lNgET1Y:lNhES1Y:kNhEU1Y:jNgEV1[:fNhEY1[:bNgE_1[:[NhEf1Z:TNjEk1k:10O10O03N5J35K00O10000O0O000O10OQOaD3_;LdD2\\;MfD2Z;MiD1V;OlD1S;NoD1P;ORE1l:OWE0g:0[E0d:O^E0a:0aE0]:0eE0Y:OjE0U:0mE0Q:0QF0n9OSF1l9OUEN:3`:i0]EXOa:l0\\ETOc:Q1YEmNh:X1TEgNl:Z1TEeNm:[1SEdNm:^1SE_Nn:b1RE]Nn:d1REZNo:g1REWNn:j1REUNo:k1RERNQ;b0hEJa:KiE4h;N2N2N0010O01O1O001N1O2M20Vad4"}, "label": "a lady in blue t - shirt and white shorts sitting on a park bench"}]}
{"id": 185558, "image": "COCO_train2014_000000185558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couple of friends are sitting on a bench and hanging out\"."}], "task": "refering", "answer_template": "The \"a couple of friends are sitting on a bench and hanging out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 167, 168, 190, 191, 192, 213, 214, 215, 216, 237, 238, 239, 261, 262, 285], "bbox": [0.285703125, 0.44948477751756444, 0.442, 0.8140983606557377], "iscrowd": 0, "area": 5803.90765, "rle": {"size": [427, 640], "counts": "\\a\\29l<8M4L3M4L5K4L4M5UOgNYE`1^:fN_E`1Y:dNeE_1W:dNfE^1X:fNeE[1Z:gNcEZ1\\:hNcEX1\\:l0O100O10001O0O10000O100001O0nMhEQ1Y:nNgER1Z:lNgET1Y:lNhES1Y:kNhEU1Y:jNgEV1[:fNhEY1[:bNgE_1[:[NhEf1Z:TNjEk1k:10O10O03N5J35K00O10000O0O000O10OQOaD3_;LdD2\\;MfD2Z;MiD1V;OlD1S;NoD1P;ORE1l:OWE0g:0[E0d:O^E0a:0aE0]:0eE0Y:OjE0U:0mE0Q:0QF0n9OSF1l9OUEN:3`:i0]EXOa:l0\\ETOc:Q1YEmNh:X1TEgNl:Z1TEeNm:[1SEdNm:^1SE_Nn:b1RE]Nn:d1REZNo:g1REWNn:j1REUNo:k1RERNQ;b0hEJa:KiE4h;N2N2N0010O01O1O001N1O2M20Vad4"}, "label": "a couple of friends are sitting on a bench and hanging out"}]}
{"id": 185558, "image": "COCO_train2014_000000185558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on a bench in a black shirt\"."}], "task": "refering", "answer_template": "The \"a woman on a bench in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 169, 170, 192, 193, 194, 215, 216, 217, 218, 240, 241, 242, 263], "bbox": [0.357171875, 0.44735362997658085, 0.5435156250000001, 0.7951990632318502], "iscrowd": 0, "area": 4796.387300000001, "rle": {"size": [427, 640], "counts": "jfo248`0R<=N3N1O2O1N101AiNhDW1V;lNiDU1T;mNkDU1Q;oNnDQ1Q;QOnDP1Q;QOnDo0Q;ROnDP1P;QOPEo0n:i0N1O2O0O100O10000000mNTEHk:NVEVO6f0d:4WETO8f0a:5ZERO8g0^:6lEIU:5mEJT:3mENS:0oE0R:MoE4R:IPF7U:BlE?X:\\OiEd0^;103M6I7J2M100N2NVOYD4f;K\\D5d;H_D7a;GaD:_;DcD<\\;BgD>Y;@iD`0P<100000O10cC1^;c1[O0O101O0000000000DWEVNi:f1\\EYNd:b1aE_N^:\\1hEcNY:T1PFkNP:n0WFROk9g0[FXOf9h0ZFWOh9g0XFYOj9f0WFXOk9g0WFVOk9h0]101N2O1N2O1O03N3M00O0O2O0O2O0O2O0O1O1O001O001O00001O0O100000001O000000000N3M2NXli3"}, "label": "a woman on a bench in a black shirt"}]}
{"id": 185558, "image": "COCO_train2014_000000185558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl sitting on the bench reading the magazine\"."}], "task": "refering", "answer_template": "The \"girl sitting on the bench reading the magazine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 169, 170, 192, 193, 194, 215, 216, 217, 218, 240, 241, 242, 263], "bbox": [0.357171875, 0.44735362997658085, 0.5435156250000001, 0.7951990632318502], "iscrowd": 0, "area": 4796.387300000001, "rle": {"size": [427, 640], "counts": "jfo248`0R<=N3N1O2O1N101AiNhDW1V;lNiDU1T;mNkDU1Q;oNnDQ1Q;QOnDP1Q;QOnDo0Q;ROnDP1P;QOPEo0n:i0N1O2O0O100O10000000mNTEHk:NVEVO6f0d:4WETO8f0a:5ZERO8g0^:6lEIU:5mEJT:3mENS:0oE0R:MoE4R:IPF7U:BlE?X:\\OiEd0^;103M6I7J2M100N2NVOYD4f;K\\D5d;H_D7a;GaD:_;DcD<\\;BgD>Y;@iD`0P<100000O10cC1^;c1[O0O101O0000000000DWEVNi:f1\\EYNd:b1aE_N^:\\1hEcNY:T1PFkNP:n0WFROk9g0[FXOf9h0ZFWOh9g0XFYOj9f0WFXOk9g0WFVOk9h0]101N2O1N2O1O03N3M00O0O2O0O2O0O2O0O1O1O001O001O00001O0O100000001O000000000N3M2NXli3"}, "label": "girl sitting on the bench reading the magazine"}]}
{"id": 169179, "image": "COCO_train2014_000000169179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a grey pullover\"."}], "task": "refering", "answer_template": "The \"a man in a grey pullover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 87, 88, 89, 90, 100, 101, 102, 103, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 137, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 191, 192, 193, 194, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233], "bbox": [0.5530666666666667, 0.30642, 1.0, 0.9858], "iscrowd": 0, "area": 39123.100249999996, "rle": {"size": [500, 375], "counts": "fjU38W?7I7H8L4M3N1N3M3M3N2M3M3M2O2M3M3M3N2M3M2N3N2YFaMT6b2iIaMT6b2jI`MS6c2kI_MS6c2jI_MT6c2kI_MR6d2lI^MQ6e2lI^MQ6e2mI]MQ6d2mI^MQ6e2mI]MP6e2oI`Mk5c2RJeMg5\\2XJkM`5W2_JQNX5R2eJVNS5l1lJUNS5l1lJTNS5o1jJRNU5P2jJPNU5S2iJmMW5T2hJlMW5V2hJjMW5Y2gJgMX5[2gJeMY5\\2fJdMY5_2eJaMZ5a2eJ_MZ5c2eJ]M[5e2cJ[M\\5g2cJYM\\5j2bJVM]5l2bJTM^5m2aJSM^5P3`JPM_5R3`JnL_5T3`JlL`5V3^JjL_5Z3`JfL\\5`3bJ`LY5f3fJZLV5l3hJTLT5Q4kJoKP5W4oJiKm4]4oJeKl4a4QKaKk4e4RK\\Kk4h4RKZKn4g4oJ[KP5h4lJZKS5h4jJZKU5i4gJYKX5i4eJYKZ5i4cJYK\\5j4`JXK_5j4^JXKa5^700O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O11O001O001O000010O01O001O001O00001O001O001O001O00001O001O1O1O1O1O001O1oG`Jl5a5PJcJo5^5mIeJS6\\5iIhJV6Y5fIkJY6V5bInJ^6V5[InJd6U5VIoJi6T5QIoJo6U5jHoJU7T5eHPKZ7T5_HoJa7c6O1O001O001O1O001O1O0O2L4M2M4L4L3M4M3L3M4J6YO`H"}, "label": "a man in a grey pullover"}]}
{"id": 169179, "image": "COCO_train2014_000000169179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a short , balding man with glasses in a grey sweater\"."}], "task": "refering", "answer_template": "The \"a short , balding man with glasses in a grey sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 87, 88, 89, 90, 100, 101, 102, 103, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 137, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 191, 192, 193, 194, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233], "bbox": [0.5530666666666667, 0.30642, 1.0, 0.9858], "iscrowd": 0, "area": 39123.100249999996, "rle": {"size": [500, 375], "counts": "fjU38W?7I7H8L4M3N1N3M3M3N2M3M3M2O2M3M3M3N2M3M2N3N2YFaMT6b2iIaMT6b2jI`MS6c2kI_MS6c2jI_MT6c2kI_MR6d2lI^MQ6e2lI^MQ6e2mI]MQ6d2mI^MQ6e2mI]MP6e2oI`Mk5c2RJeMg5\\2XJkM`5W2_JQNX5R2eJVNS5l1lJUNS5l1lJTNS5o1jJRNU5P2jJPNU5S2iJmMW5T2hJlMW5V2hJjMW5Y2gJgMX5[2gJeMY5\\2fJdMY5_2eJaMZ5a2eJ_MZ5c2eJ]M[5e2cJ[M\\5g2cJYM\\5j2bJVM]5l2bJTM^5m2aJSM^5P3`JPM_5R3`JnL_5T3`JlL`5V3^JjL_5Z3`JfL\\5`3bJ`LY5f3fJZLV5l3hJTLT5Q4kJoKP5W4oJiKm4]4oJeKl4a4QKaKk4e4RK\\Kk4h4RKZKn4g4oJ[KP5h4lJZKS5h4jJZKU5i4gJYKX5i4eJYKZ5i4cJYK\\5j4`JXK_5j4^JXKa5^700O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O11O001O001O000010O01O001O001O00001O001O001O001O00001O001O1O1O1O1O001O1oG`Jl5a5PJcJo5^5mIeJS6\\5iIhJV6Y5fIkJY6V5bInJ^6V5[InJd6U5VIoJi6T5QIoJo6U5jHoJU7T5eHPKZ7T5_HoJa7c6O1O001O001O1O001O1O0O2L4M2M4L4L3M4M3L3M4J6YO`H"}, "label": "a short , balding man with glasses in a grey sweater"}]}
{"id": 169179, "image": "COCO_train2014_000000169179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man hold a remote with smile\"."}], "task": "refering", "answer_template": "The \"a man hold a remote with smile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 27, 28, 29, 40, 41, 42, 43, 53, 54, 55, 56, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 156, 157, 158, 159, 160, 161, 162, 163, 164, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 189, 190, 191, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226], "bbox": [0.017973333333333334, 0.09662000000000001, 0.7490666666666668, 0.98426], "iscrowd": 0, "area": 66212.54460000002, "rle": {"size": [500, 375], "counts": "[j3:V82[N>Z13QN<d16eM8R26YM7a23lL<n2MTLn0f3[OZKb1`4n4J3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M2N2O10000fNaE`M^:Q2SFmMm9Q2WFnMh9Q2[FmMe9R2^FlMb9S2aFlM]9U2eFiM[9V2R200O100O10000O100O100O10O010000O100O100O10000O100O10O0100O10000O100O11O0E<M2N3M2N3M2N2N3N1100O101N1O10000001O1O001O00001O010O001O00001O002N2N3L4L5K4L5Kk0UO5K4L3M4L4L1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1jMaJUJ`5m4dKlJ]4Y4eL^K^3a4RMdJY3[5b3O1N2O1N2O1N1O2O3L4M2M4M3L4M3L3M4M3L4M2M4M2M2O1N2O1N1O2O1N2O1N2O1N2O1N1O2O1N2O1N2O1N2O0O2N2O1N2O1N2O1N101N100O2N100O2O1N2O1N3N1N2N2O1N2O1N101N2O0O2N2O0O2O1N2O0O2O1N101N2N2O0O2O1N101N2O0O2N100O100O100O101N100O1O102M2O2M2O1N3N1N3M2O1N2O1N2O0Oc_]1"}, "label": "a man hold a remote with smile"}]}
{"id": 169179, "image": "COCO_train2014_000000169179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one person wearing red t - shirt holding a device in his hand\"."}], "task": "refering", "answer_template": "The \"there is one person wearing red t - shirt holding a device in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 27, 28, 29, 40, 41, 42, 43, 53, 54, 55, 56, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 156, 157, 158, 159, 160, 161, 162, 163, 164, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 189, 190, 191, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226], "bbox": [0.017973333333333334, 0.09662000000000001, 0.7490666666666668, 0.98426], "iscrowd": 0, "area": 66212.54460000002, "rle": {"size": [500, 375], "counts": "[j3:V82[N>Z13QN<d16eM8R26YM7a23lL<n2MTLn0f3[OZKb1`4n4J3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M2N2O10000fNaE`M^:Q2SFmMm9Q2WFnMh9Q2[FmMe9R2^FlMb9S2aFlM]9U2eFiM[9V2R200O100O10000O100O100O10O010000O100O100O10000O100O10O0100O10000O100O11O0E<M2N3M2N3M2N2N3N1100O101N1O10000001O1O001O00001O010O001O00001O002N2N3L4L5K4L5Kk0UO5K4L3M4L4L1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1jMaJUJ`5m4dKlJ]4Y4eL^K^3a4RMdJY3[5b3O1N2O1N2O1N1O2O3L4M2M4M3L4M3L3M4M3L4M2M4M2M2O1N2O1N1O2O1N2O1N2O1N2O1N1O2O1N2O1N2O1N2O0O2N2O1N2O1N2O1N101N100O2N100O2O1N2O1N3N1N2N2O1N2O1N101N2O0O2N2O0O2O1N2O0O2O1N101N2N2O0O2O1N101N2O0O2N100O100O100O101N100O1O102M2O2M2O1N3N1N3M2O1N2O1N2O0Oc_]1"}, "label": "there is one person wearing red t - shirt holding a device in his hand"}]}
{"id": 349403, "image": "COCO_train2014_000000349403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elderly woman in black dress holding an umbrella\"."}], "task": "refering", "answer_template": "The \"elderly woman in black dress holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 260, 261, 262, 283, 284, 285, 307], "bbox": [0.22178124999999999, 0.10327083333333334, 0.46557812500000006, 0.7714375], "iscrowd": 0, "area": 25569.346899999986, "rle": {"size": [480, 640], "counts": "UiR25b>9I8J5L5N100O1O1O1O1O1O1O1N2N2N2N1O1O1O1N2N1O1O1O00O2O1O1H8M3M3N2O1N2O0001[CL`:4[E6_:KaE;Y:FfE`0S:BkEg0l9ZOTFb1P9_NoFf1k8[NUGi1g8XNXGl1c8VN[Go1a8RN]GS2`8mM^GY2^8gM`G_2\\8bMaGd2\\8[MaGl2[8TMcGQ3Z8oLcGV3[8jLbG\\3\\8cLbGa3\\8_LaGf3]8`1N2N3N1N2O1N10000O100iNWIWJi6h5`IoIa6o5iIiIY6T6jIjIX6S6jIkIY6Q6kImI`6g5bIXJX7m4kHPKW7m4mHQKS7n4oHQKR7n4`1O2O001O1O00000000UNVK`Ij4_6[K]Ie4a6mKoHS4o6RLnHn3Q7VLlHj3S7\\LhHd3X7bLaH_3_7iLYHW3f7R2000[IYHQ6g7nI\\HP6c7PJaHm5_7RJdHl5\\7SJhHj5X7UJjHj5U7VJmHi5S7VJoHi5T801O1O2O3MTH\\J^6`5aIaJ`6_5^IcJb6\\5]IeJd6[5ZIgJf6Y5YIgJh6Z5UIhJk6[5lHjJV7`6L3J6J7I6J6J6J7I6J6J7I6J6J7I6J7I6J7H7I7J8G9H7H9D<B=VO_CgNQ=R1d0I7K3M2N3M2N2N3M2N5K5KZUP5"}, "label": "elderly woman in black dress holding an umbrella"}]}
{"id": 349403, "image": "COCO_train2014_000000349403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holing a white ubmrella and wearing a dress\"."}], "task": "refering", "answer_template": "The \"a woman holing a white ubmrella and wearing a dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 260, 261, 262, 283, 284, 285, 307], "bbox": [0.22178124999999999, 0.10327083333333334, 0.46557812500000006, 0.7714375], "iscrowd": 0, "area": 25569.346899999986, "rle": {"size": [480, 640], "counts": "UiR25b>9I8J5L5N100O1O1O1O1O1O1O1N2N2N2N1O1O1O1N2N1O1O1O00O2O1O1H8M3M3N2O1N2O0001[CL`:4[E6_:KaE;Y:FfE`0S:BkEg0l9ZOTFb1P9_NoFf1k8[NUGi1g8XNXGl1c8VN[Go1a8RN]GS2`8mM^GY2^8gM`G_2\\8bMaGd2\\8[MaGl2[8TMcGQ3Z8oLcGV3[8jLbG\\3\\8cLbGa3\\8_LaGf3]8`1N2N3N1N2O1N10000O100iNWIWJi6h5`IoIa6o5iIiIY6T6jIjIX6S6jIkIY6Q6kImI`6g5bIXJX7m4kHPKW7m4mHQKS7n4oHQKR7n4`1O2O001O1O00000000UNVK`Ij4_6[K]Ie4a6mKoHS4o6RLnHn3Q7VLlHj3S7\\LhHd3X7bLaH_3_7iLYHW3f7R2000[IYHQ6g7nI\\HP6c7PJaHm5_7RJdHl5\\7SJhHj5X7UJjHj5U7VJmHi5S7VJoHi5T801O1O2O3MTH\\J^6`5aIaJ`6_5^IcJb6\\5]IeJd6[5ZIgJf6Y5YIgJh6Z5UIhJk6[5lHjJV7`6L3J6J7I6J6J6J7I6J6J7I6J6J7I6J7I6J7H7I7J8G9H7H9D<B=VO_CgNQ=R1d0I7K3M2N3M2N2N3M2N5K5KZUP5"}, "label": "a woman holing a white ubmrella and wearing a dress"}]}
{"id": 562401, "image": "COCO_train2014_000000562401.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli florets\"."}], "task": "refering", "answer_template": "The \"broccoli florets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 127, 128, 129, 144, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.162359375, 0.24706250000000002, 0.744703125, 0.8823541666666667], "iscrowd": 0, "area": 67993.67785, "rle": {"size": [480, 640], "counts": "aQa1g0c=f000O1O2O0O100O101N100O101N100O100O2O0O1eNVN_Ek1a:XN\\Eh1d:[NXEg1g:\\NVEd1i:aNRE`1n:cNoD]1Q;fNkD\\1T;gNiDY1V;kNfDV1Z;nNbDR1^;QO^DP1b;SO[Dn0d;W1O100mDSMS:n31N100O100O10001NO2L3M4M2M4L3N3L3M4L3N3L4L3N3L3M4L3N3L3fGUJ\\7i6N3L3M4M2O21N100O100O100O100O100O100O100O101N100O100O100O100N2N2O1N2O1N2O2M2O1N2N2O7H9HM2O2M2N3M2O1N3M2O20O2O0O1O10000100O1O1O1O100O1O1O1O1O100O1O1O1O00O2O000000001O00000000001O000000001O0000001O0000001O0000001N10001O0000001O0000001O0000001O0000O1O1O1O1O1O1POaI_I_6_6eI^I\\6_6iI^IX6`6lI]IU6`6PJ]IQ6`6TJ]Il5d6VJYIk5h6VJUIk5k6XJRIh5n6ZJoHg5R7S10000001O00001O0000001O0000_NXHkKg7U4ZHjKf7V4[HiKe7W4\\HhKd7X4]HhKb7X4_HgK`7Z4aHeK_7[4bHeK]7[4dHdK\\7[4fHdKZ7\\4gHcKY7]4hHcKW7]4jHbKV7^4lH`KT7`4mHYKY7g4hHRK^7n4cHkJc7U5^HdJh7\\5YH]Jm7c5h0000000O10000000O100000000000000000O1000eL]JYMc5g2^JXMb5g2_JYMa5g2`JXM`5h2aJWM_5h2cJWM]5i2dJVM\\5j2dJVM\\5i2fJVMZ5j2gJUMY5j2iJUMW5k2iJUMV5l2kJSMU5l2mJSMS5R2`JfK>W2R5R2eJbK;[2P5R2kJ^K5`2P5Q2PK[K1c2o4Q2UKWKNg2m4Q2RLnMn3Q2TLnMl3Q2ULoMk3P2VLPNj3o1XLPNm3j1TLVNX4]1iKcNc4P1_KoNl4e0UK[Ol4c0UK]Ok4b0UK_Ol4>VKBk4<VKDj4<UKEl49UKGl48TKHl48SKIn45SKKm45SKKn44RKLo42QKOo41QKNQ51oJOR5OnJ2R5NnJ2S5MmJ3S5LmJ5T5JlJ5V5JjJ6V5IjJ8W5GiJ9X5FhJ:\\5AeJ?i5SOVJm0Y6dNhI\\1f6UN[Ik1S7WM\\Ij2]9000O01000O100000O010000O1000O011O0000000O10001O00000O100O2N1N2O1N2O1O2M2O1N2O1O1N2O2M2O1N4M3M4K5L3L5L4L3L5L4K4M4L3L5L4K4MeQ\\2"}, "label": "broccoli florets"}]}
{"id": 562401, "image": "COCO_train2014_000000562401.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccolis in the plate\"."}], "task": "refering", "answer_template": "The \"the broccolis in the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 127, 128, 129, 144, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.162359375, 0.24706250000000002, 0.744703125, 0.8823541666666667], "iscrowd": 0, "area": 67993.67785, "rle": {"size": [480, 640], "counts": "aQa1g0c=f000O1O2O0O100O101N100O101N100O100O2O0O1eNVN_Ek1a:XN\\Eh1d:[NXEg1g:\\NVEd1i:aNRE`1n:cNoD]1Q;fNkD\\1T;gNiDY1V;kNfDV1Z;nNbDR1^;QO^DP1b;SO[Dn0d;W1O100mDSMS:n31N100O100O10001NO2L3M4M2M4L3N3L3M4L3N3L4L3N3L3M4L3N3L3fGUJ\\7i6N3L3M4M2O21N100O100O100O100O100O100O100O101N100O100O100O100N2N2O1N2O1N2O2M2O1N2N2O7H9HM2O2M2N3M2O1N3M2O20O2O0O1O10000100O1O1O1O100O1O1O1O1O100O1O1O1O00O2O000000001O00000000001O000000001O0000001O0000001O0000001N10001O0000001O0000001O0000001O0000O1O1O1O1O1O1POaI_I_6_6eI^I\\6_6iI^IX6`6lI]IU6`6PJ]IQ6`6TJ]Il5d6VJYIk5h6VJUIk5k6XJRIh5n6ZJoHg5R7S10000001O00001O0000001O0000_NXHkKg7U4ZHjKf7V4[HiKe7W4\\HhKd7X4]HhKb7X4_HgK`7Z4aHeK_7[4bHeK]7[4dHdK\\7[4fHdKZ7\\4gHcKY7]4hHcKW7]4jHbKV7^4lH`KT7`4mHYKY7g4hHRK^7n4cHkJc7U5^HdJh7\\5YH]Jm7c5h0000000O10000000O100000000000000000O1000eL]JYMc5g2^JXMb5g2_JYMa5g2`JXM`5h2aJWM_5h2cJWM]5i2dJVM\\5j2dJVM\\5i2fJVMZ5j2gJUMY5j2iJUMW5k2iJUMV5l2kJSMU5l2mJSMS5R2`JfK>W2R5R2eJbK;[2P5R2kJ^K5`2P5Q2PK[K1c2o4Q2UKWKNg2m4Q2RLnMn3Q2TLnMl3Q2ULoMk3P2VLPNj3o1XLPNm3j1TLVNX4]1iKcNc4P1_KoNl4e0UK[Ol4c0UK]Ok4b0UK_Ol4>VKBk4<VKDj4<UKEl49UKGl48TKHl48SKIn45SKKm45SKKn44RKLo42QKOo41QKNQ51oJOR5OnJ2R5NnJ2S5MmJ3S5LmJ5T5JlJ5V5JjJ6V5IjJ8W5GiJ9X5FhJ:\\5AeJ?i5SOVJm0Y6dNhI\\1f6UN[Ik1S7WM\\Ij2]9000O01000O100000O010000O1000O011O0000000O10001O00000O100O2N1N2O1N2O1O2M2O1N2O1O1N2O2M2O1N4M3M4K5L3L5L4L3L5L4K4M4L3L5L4K4MeQ\\2"}, "label": "the broccolis in the plate"}]}
{"id": 62690, "image": "COCO_train2014_000000062690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing looking the rails at the ocean with a man seated on a brown bench\"."}], "task": "refering", "answer_template": "The \"woman standing looking the rails at the ocean with a man seated on a brown bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 28, 29, 46, 47, 48, 64, 65, 66, 100, 101, 118, 119], "bbox": [0.551, 0.015818181818181818, 0.682, 0.8544545454545454], "iscrowd": 0, "area": 5320.5, "rle": {"size": [220, 500], "counts": "_Zk17b68H5L2N2N1O4L5K4L4K6K4M3N2_K^Nd3e1RLbNn1K_Oe1ZNgNg17Lh13[NKe15]NHd18^NFb1:`ND`1<cNA^1>dN@\\1a0dN^O\\1b0dN^O\\1b0dN]O]1c0cN]O]1c0cN]O]1c0cN]O]1d0bN\\O_1c0aN]O_1c0aN_O]1a0cNA[1?eNCY1=gNFV1;jNFT1:lNHR18nNIR16nNKQ15oNLP14POMo04POLP14POMo03QONn02ROOm01SOOn00RO1m00nN5Q1KjN;U1EfN`0Z1@aNf0^1ZO]Nl0b1TOZNQ1e1S213L5K4L4L4L:F7I6K5J6J6JYRR1"}, "label": "woman standing looking the rails at the ocean with a man seated on a brown bench"}]}
{"id": 62690, "image": "COCO_train2014_000000062690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person reaching back while getting up off of a bench\"."}], "task": "refering", "answer_template": "The \"a person reaching back while getting up off of a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 28, 29, 46, 47, 48, 64, 65, 66, 100, 101, 118, 119], "bbox": [0.551, 0.015818181818181818, 0.682, 0.8544545454545454], "iscrowd": 0, "area": 5320.5, "rle": {"size": [220, 500], "counts": "_Zk17b68H5L2N2N1O4L5K4L4K6K4M3N2_K^Nd3e1RLbNn1K_Oe1ZNgNg17Lh13[NKe15]NHd18^NFb1:`ND`1<cNA^1>dN@\\1a0dN^O\\1b0dN^O\\1b0dN]O]1c0cN]O]1c0cN]O]1c0cN]O]1d0bN\\O_1c0aN]O_1c0aN_O]1a0cNA[1?eNCY1=gNFV1;jNFT1:lNHR18nNIR16nNKQ15oNLP14POMo04POLP14POMo03QONn02ROOm01SOOn00RO1m00nN5Q1KjN;U1EfN`0Z1@aNf0^1ZO]Nl0b1TOZNQ1e1S213L5K4L4L4L:F7I6K5J6J6JYRR1"}, "label": "a person reaching back while getting up off of a bench"}]}
{"id": 62690, "image": "COCO_train2014_000000062690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green button up looking at the view with a woman\"."}], "task": "refering", "answer_template": "The \"a man wearing a green button up looking at the view with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 44, 45, 46, 62, 63, 64, 98, 99, 116, 117], "bbox": [0.44504000000000005, 0.15136363636363634, 0.5718799999999999, 0.8456818181818182], "iscrowd": 0, "area": 3526.27865, "rle": {"size": [220, 500], "counts": "]o_1e0U63M101N2O1O1N2O1O1N2O1N2O1O1N2N2K5O1`L^Ne1c1ZN`Nc1a1]NaN_1a1aNaN\\1_1eNcNX1^1hNgNR1Z1nNXO>j0BWO<j0CWO=i0CXO<h0DXO=g0CZO<f0DZO<f0DZO<f0DZO<f0C[O=e0C[O=e0C[O?c0A\\Oa0c0_O]Oc0a0]O_Od0`0\\O@f0>YOCh0;YOEh0:XOEj0:VOFk09UOFm09SOFo09QOVOaNF`2S1nNhNTO2P2U1JiN8V1IiN8V1h2O1O2N3M4L3M4L4L3M4L4K5L_l]1"}, "label": "a man wearing a green button up looking at the view with a woman"}]}
{"id": 62690, "image": "COCO_train2014_000000062690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green shirt looking out over a pier\"."}], "task": "refering", "answer_template": "The \"a man in a green shirt looking out over a pier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 44, 45, 46, 62, 63, 64, 98, 99, 116, 117], "bbox": [0.44504000000000005, 0.15136363636363634, 0.5718799999999999, 0.8456818181818182], "iscrowd": 0, "area": 3526.27865, "rle": {"size": [220, 500], "counts": "]o_1e0U63M101N2O1O1N2O1O1N2O1N2O1O1N2N2K5O1`L^Ne1c1ZN`Nc1a1]NaN_1a1aNaN\\1_1eNcNX1^1hNgNR1Z1nNXO>j0BWO<j0CWO=i0CXO<h0DXO=g0CZO<f0DZO<f0DZO<f0DZO<f0C[O=e0C[O=e0C[O?c0A\\Oa0c0_O]Oc0a0]O_Od0`0\\O@f0>YOCh0;YOEh0:XOEj0:VOFk09UOFm09SOFo09QOVOaNF`2S1nNhNTO2P2U1JiN8V1IiN8V1h2O1O2N3M4L3M4L4L3M4L4K5L_l]1"}, "label": "a man in a green shirt looking out over a pier"}]}
{"id": 79083, "image": "COCO_train2014_000000079083.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bright yellow car\"."}], "task": "refering", "answer_template": "The \"the bright yellow car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 89, 90, 91, 92, 94, 95, 106, 107, 118, 119, 130, 131], "bbox": [0.40885885885885886, 0.33588, 1.0, 0.58856], "iscrowd": 0, "area": 6140.6509, "rle": {"size": [500, 333], "counts": "YcR23[?8H6K5N2N2N2N2O10O01000O10O010000O00100O10O0100O1O010O100O10O2O00000O10001O00000000001O00000000001O0000000000O10000000001O0000000000001O000O100000001O0000000000001O0000000000001O000000000000001O0000001O001O00001O002N1O2N2N2N1O2N1O2N1O2N1OeRb00Zm]O3M3`B4S;O`D<^;HVDb0i;AaCS1^<PO^CR1a<QO[CQ1d<ROWCQ1g<o00001O00000001O00000001O0001O000000000O2N1O2N101N1YNoBj0S=UOmBj0T=UOnBi0T=VOnBe0X=XOkBa0^=ZOeB>g6"}, "label": "the bright yellow car"}]}
{"id": 431340, "image": "COCO_train2014_000000431340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck that has the word bread written on it in green\"."}], "task": "refering", "answer_template": "The \"a truck that has the word bread written on it in green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 202, 203, 204, 205, 206, 207, 208, 209, 210, 224, 225, 226], "bbox": [0.1595, 0.04050666666666666, 0.8981, 0.9924], "iscrowd": 0, "area": 98220.87670000001, "rle": {"size": [375, 500], "counts": "ham01f;1]E1[91aF4]9M_F8_9G_F>^9B`Fb0_9^O^Fe0b9ZOYFl0g9c0N100K5^Oc0^Oa0_Oa0^Ob0_Ob0^Oa0^Ob0_Ob0^Oa0^Ob0_Oa0@a0M2O1N3N1N2O101O00001N1000001O000O101O00001O000O101O00001O000O101O00001O0O10001O00001O0O10001O00000O2O00001O00000O2O00001O00000O2O00001O000O101O0000001O0O101O0000001N10001O0000001N10001O0000001N1000001O000O2O0000001O000O2O0000001O000O2O0000001O0O101O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O0O2O00000000000O2O000000000O10001O0000000O1000001O000O100000000000O10000O10000O101O0O2O000hMDTI>i6ESI>l6DoHa0n6CmH`0R7BkHa0R7DiH>V7IbH:[7K`H8^7K^H7_7N^H4U7:gHHX7;dHH[7:bHG^7:_HI`7X2O001O001N2O001O001O0101O2M3N1N3N1O0O2O0O101N100nLfHf1[7UNhHk1Y7RNjHn1W7mMmHR2T7gMRIZ2U81N101O0O2O001N2O001N101O1N101N101N101O1N101N101N2O001N101K4L5J6K4N3N1N3N2N1N3N1O2N101N2O001N101N2O0O2O0O2O001N2O0O2O0O2O1N101O0O2O0O2O1N101N101N2O001N101N101N2O0O2O001N2O0O20O100O10000OO2O001O1N101O001N101M3M2H9J5M4K5L3L5N;DQ1POQ1lNS1gN\\_b0"}, "label": "a truck that has the word bread written on it in green"}]}
{"id": 431340, "image": "COCO_train2014_000000431340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck in the road\"."}], "task": "refering", "answer_template": "The \"a truck in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 202, 203, 204, 205, 206, 207, 208, 209, 210, 224, 225, 226], "bbox": [0.1595, 0.04050666666666666, 0.8981, 0.9924], "iscrowd": 0, "area": 98220.87670000001, "rle": {"size": [375, 500], "counts": "ham01f;1]E1[91aF4]9M_F8_9G_F>^9B`Fb0_9^O^Fe0b9ZOYFl0g9c0N100K5^Oc0^Oa0_Oa0^Ob0_Ob0^Oa0^Ob0_Ob0^Oa0^Ob0_Oa0@a0M2O1N3N1N2O101O00001N1000001O000O101O00001O000O101O00001O000O101O00001O0O10001O00001O0O10001O00000O2O00001O00000O2O00001O00000O2O00001O000O101O0000001O0O101O0000001N10001O0000001N10001O0000001N1000001O000O2O0000001O000O2O0000001O000O2O0000001O0O101O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O0O2O00000000000O2O000000000O10001O0000000O1000001O000O100000000000O10000O10000O101O0O2O000hMDTI>i6ESI>l6DoHa0n6CmH`0R7BkHa0R7DiH>V7IbH:[7K`H8^7K^H7_7N^H4U7:gHHX7;dHH[7:bHG^7:_HI`7X2O001O001N2O001O001O0101O2M3N1N3N1O0O2O0O101N100nLfHf1[7UNhHk1Y7RNjHn1W7mMmHR2T7gMRIZ2U81N101O0O2O001N2O001N101O1N101N101N101O1N101N101N2O001N101K4L5J6K4N3N1N3N2N1N3N1O2N101N2O001N101N2O0O2O0O2O001N2O0O2O0O2O1N101O0O2O0O2O1N101N101N2O001N101N101N2O0O2O001N2O0O20O100O10000OO2O001O1N101O001N101M3M2H9J5M4K5L3L5N;DQ1POQ1lNS1gN\\_b0"}, "label": "a truck in the road"}]}
{"id": 169197, "image": "COCO_train2014_000000169197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair behind the cat\"."}], "task": "refering", "answer_template": "The \"the chair behind the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 56, 57, 58, 79, 80, 81, 82, 102, 103, 104, 105, 126, 127, 128, 129, 130, 131, 132, 152, 153, 154, 155, 178, 201, 224, 247], "bbox": [0.45225, 0.0020416666666666665, 0.774140625, 0.6179166666666667], "iscrowd": 0, "area": 14017.357549999997, "rle": {"size": [480, 640], "counts": "ToW46j>6J6I7J6J6J6I7J5J7I7I7J6I7I7I7J6I6J7J3L1O0010O01O1O100O00100O1O1O010O1O1O1N2M2N3M300000O100000000O100001O02N4L5L3L5K4L4M4K4L5K4M3L5K4M4K4L4L5L3L5K4L4M4K4M4L3L4M4K4M4Lc0\\Omj90RUF6K5K6I6K5K6J1O00000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O100007I=]CSOY:T3F2N1O2N2M3N1O2N2N2N001O1O001>BPjS2"}, "label": "the chair behind the cat"}]}
{"id": 169197, "image": "COCO_train2014_000000169197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair behind a kitten\"."}], "task": "refering", "answer_template": "The \"a chair behind a kitten\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 56, 57, 58, 79, 80, 81, 82, 102, 103, 104, 105, 126, 127, 128, 129, 130, 131, 132, 152, 153, 154, 155, 178, 201, 224, 247], "bbox": [0.45225, 0.0020416666666666665, 0.774140625, 0.6179166666666667], "iscrowd": 0, "area": 14017.357549999997, "rle": {"size": [480, 640], "counts": "ToW46j>6J6I7J6J6J6I7J5J7I7I7J6I7I7I7J6I6J7J3L1O0010O01O1O100O00100O1O1O010O1O1O1N2M2N3M300000O100000000O100001O02N4L5L3L5K4L4M4K4L5K4M3L5K4M4K4L4L5L3L5K4L4M4K4M4L3L4M4K4M4Lc0\\Omj90RUF6K5K6I6K5K6J1O00000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O100007I=]CSOY:T3F2N1O2N2M3N1O2N2N2N001O1O001>BPjS2"}, "label": "a chair behind a kitten"}]}
{"id": 414961, "image": "COCO_train2014_000000414961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing pink dress\"."}], "task": "refering", "answer_template": "The \"woman wearing pink dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 142, 143, 144, 145, 146, 166, 167, 168, 169, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 283, 303, 304, 306, 326, 327, 328, 329], "bbox": [0.16078125000000001, 0.1092271662763466, 0.36665625, 0.9898594847775175], "iscrowd": 0, "area": 25098.0791, "rle": {"size": [427, 640], "counts": "RS[1?j<;D3N3M2M4M2M4M2N2N2O1O1O1O1O1O1O1N2O2N1O2_KUN_Ml1^MhNZ2F5d1aMoNf1If0Y1dMYOT1GU1Q1gMBh0A^1o0jMC\\OQOWO`0`3n0mMDTOj0m2DnMDoNm0Q3@QNDgNQ1W3\\ORNF_NS1^3YORNFZNU1c3VOTNP3j1RMUNn2j1SMVNm2i1TMXNl2g1TMYNl2g1TMYNl2g1TMZNl2d1UM\\Nk2d1UM\\N\\1dNmLT3g1YNY1TO^Le2Z2WNU1V3kNjLR1Y3mNgLQ1\\3oNdLn0^3SObLj0a3VO_Lg0d3[OZLb0i3ASL=o3HmK5V4N^K7f4LkJ=W5GYJb0i5]33N3L4M4K4L4M4L30001O00001O00000000O1WOeIbJ[6Y5nIbJS6Z5TJcJl5Z5T100O1O1O1O1O1O1O102N5J6K5hMeGVO`8<RHZOh8F^G;o8TOUGl0W9dNlF\\1b:000O10000010O00kE_N`8a1_GaN`8`1^GaNb8_1]GcNb8]1]GeNc8[1[GfNe8[1YGgNf8Z1XGgN`07_4NdJU1;hN>8^44fJl0<iN>:Z4:hJd0>iN><W4>jJ?=jN><V4a0nJ:<jN>>R4d0TK49lN>?]2lNmMk1nN_1e0mM\\1[3nMiNb0oMZ1_3SNfN=nMX1d3[N`N6PNW1h3cNZNNUNS1m3nN4m03POMm0[6K4L5K4L5M2N3N2M2ObSY5"}, "label": "woman wearing pink dress"}]}
{"id": 414961, "image": "COCO_train2014_000000414961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white dress and dark jacket\"."}], "task": "refering", "answer_template": "The \"a woman in a white dress and dark jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 142, 143, 144, 145, 146, 166, 167, 168, 169, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 283, 303, 304, 306, 326, 327, 328, 329], "bbox": [0.16078125000000001, 0.1092271662763466, 0.36665625, 0.9898594847775175], "iscrowd": 0, "area": 25098.0791, "rle": {"size": [427, 640], "counts": "RS[1?j<;D3N3M2M4M2M4M2N2N2O1O1O1O1O1O1O1N2O2N1O2_KUN_Ml1^MhNZ2F5d1aMoNf1If0Y1dMYOT1GU1Q1gMBh0A^1o0jMC\\OQOWO`0`3n0mMDTOj0m2DnMDoNm0Q3@QNDgNQ1W3\\ORNF_NS1^3YORNFZNU1c3VOTNP3j1RMUNn2j1SMVNm2i1TMXNl2g1TMYNl2g1TMYNl2g1TMZNl2d1UM\\Nk2d1UM\\N\\1dNmLT3g1YNY1TO^Le2Z2WNU1V3kNjLR1Y3mNgLQ1\\3oNdLn0^3SObLj0a3VO_Lg0d3[OZLb0i3ASL=o3HmK5V4N^K7f4LkJ=W5GYJb0i5]33N3L4M4K4L4M4L30001O00001O00000000O1WOeIbJ[6Y5nIbJS6Z5TJcJl5Z5T100O1O1O1O1O1O1O102N5J6K5hMeGVO`8<RHZOh8F^G;o8TOUGl0W9dNlF\\1b:000O10000010O00kE_N`8a1_GaN`8`1^GaNb8_1]GcNb8]1]GeNc8[1[GfNe8[1YGgNf8Z1XGgN`07_4NdJU1;hN>8^44fJl0<iN>:Z4:hJd0>iN><W4>jJ?=jN><V4a0nJ:<jN>>R4d0TK49lN>?]2lNmMk1nN_1e0mM\\1[3nMiNb0oMZ1_3SNfN=nMX1d3[N`N6PNW1h3cNZNNUNS1m3nN4m03POMm0[6K4L5K4L5M2N3N2M2ObSY5"}, "label": "a woman in a white dress and dark jacket"}]}
{"id": 414961, "image": "COCO_train2014_000000414961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black jacket holding an umbrella walking beside a woman\"."}], "task": "refering", "answer_template": "The \"a man in a black jacket holding an umbrella walking beside a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 262, 263, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 333], "bbox": [0.324625, 0.13138173302107728, 0.532546875, 0.9901405152224825], "iscrowd": 0, "area": 24904.633949999996, "rle": {"size": [427, 640], "counts": "elf26o<8H9G9G8K6K3M2N2N3M2N2N3N1N2N3M2M3M4K4L4M4M2M3N3O0010O01aFnLd8Q3SG\\Ml8b2kFjMh9d2`JVM8o2BZM6k2DWM<m2]OWMa0n2YOTMf0P3UOSMi0R3QOPMo0T3kNoLS1V3gNnLV1W3dNmLX1W3dNmLX1U3fNoLObNUNa4j1PMKhNUN[4m1oLMiNSNY4n1PMNjNQNW4o1QM0jNoMV4k1UM5hNkMV4k1WM9S4@SL`0m3[OXLd0i3WO\\Li0e3QO`Lo0g3eN^LZ1j3YN\\Lg1i6kM`FU2^9QN^Fn1`9YN\\Fg1b9_NZFa1d9dNYF[1f9jNWFV1g9oNVFQ1h9TOUFl0h9]1N2N2N2N2N2M3N2M3N4K4M4L41O001NRHaKQ7^4jHfKX7[4aHjK`7[4THjKn7j41N2O1N2N2N^MTKeLj4W3]KiLa4Q3hKnLV4n2PLRMm3k2YLUMe3h2`LXM]3e2iL[MU3b2QM]Mm2_2YMaMd2]2aMcM]2Z2hMfMU2Z2nMiMm1V2VNoMb1P2cNlM^1S2eNiM]1V2fNgMZ1Y2iNdMX1[2kNbMV1]2mN`MU1^2oN^MR1]2UO`Mm0V2^OfMd0Q2HkM:Q2KlM7S2LiM5V2NgM4W2OfMN^23`M1\\22`M3]2NaM5\\2NaM6\\25XMOd23WM4f2MVM9f2Y5M4K5K5K5J5J7G9G9A?[OU`l3"}, "label": "a man in a black jacket holding an umbrella walking beside a woman"}]}
{"id": 414961, "image": "COCO_train2014_000000414961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue jeans walking beside a woman\"."}], "task": "refering", "answer_template": "The \"a man in blue jeans walking beside a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 262, 263, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 333], "bbox": [0.324625, 0.13138173302107728, 0.532546875, 0.9901405152224825], "iscrowd": 0, "area": 24904.633949999996, "rle": {"size": [427, 640], "counts": "elf26o<8H9G9G8K6K3M2N2N3M2N2N3N1N2N3M2M3M4K4L4M4M2M3N3O0010O01aFnLd8Q3SG\\Ml8b2kFjMh9d2`JVM8o2BZM6k2DWM<m2]OWMa0n2YOTMf0P3UOSMi0R3QOPMo0T3kNoLS1V3gNnLV1W3dNmLX1W3dNmLX1U3fNoLObNUNa4j1PMKhNUN[4m1oLMiNSNY4n1PMNjNQNW4o1QM0jNoMV4k1UM5hNkMV4k1WM9S4@SL`0m3[OXLd0i3WO\\Li0e3QO`Lo0g3eN^LZ1j3YN\\Lg1i6kM`FU2^9QN^Fn1`9YN\\Fg1b9_NZFa1d9dNYF[1f9jNWFV1g9oNVFQ1h9TOUFl0h9]1N2N2N2N2N2M3N2M3N4K4M4L41O001NRHaKQ7^4jHfKX7[4aHjK`7[4THjKn7j41N2O1N2N2N^MTKeLj4W3]KiLa4Q3hKnLV4n2PLRMm3k2YLUMe3h2`LXM]3e2iL[MU3b2QM]Mm2_2YMaMd2]2aMcM]2Z2hMfMU2Z2nMiMm1V2VNoMb1P2cNlM^1S2eNiM]1V2fNgMZ1Y2iNdMX1[2kNbMV1]2mN`MU1^2oN^MR1]2UO`Mm0V2^OfMd0Q2HkM:Q2KlM7S2LiM5V2NgM4W2OfMN^23`M1\\22`M3]2NaM5\\2NaM6\\25XMOd23WM4f2MVM9f2Y5M4K5K5K5J5J7G9G9A?[OU`l3"}, "label": "a man in blue jeans walking beside a woman"}]}
{"id": 21750, "image": "COCO_train2014_000000021750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a purse hanging from it\"."}], "task": "refering", "answer_template": "The \"a chair with a purse hanging from it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 100, 101, 102, 123, 124, 125, 146, 147, 148, 169, 170, 171, 192, 193, 194, 195, 196, 197, 215, 218, 219, 220, 238], "bbox": [0.34784375, 0.21953161592505854, 0.572734375, 0.7543793911007026], "iscrowd": 0, "area": 12847.250499999996, "rle": {"size": [427, 640], "counts": "cRm2?l<`0A`0_O`0@a0@?@a0@?\\HjLS4T63M4L3M3M4K4M3M4L3M3M4L3M3M4K4M4L3M3M4L3M3M4O00001O00001O00001O001O00001O00001O00001O000010O0001O001O00001O00001O8H;E;E<D;E<D;E;E<D:F001O0001O1O1O010O1O1O1O100O1O001O1O100O1O1O001O100O1O1O100O010O1O01N10001O0O2O00001N101O0O101O000O2O001O0O101O001N100010O0001O010O00001C<Bbca3"}, "label": "a chair with a purse hanging from it"}]}
{"id": 21750, "image": "COCO_train2014_000000021750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table on which the crockerys and cuttleries are present\"."}], "task": "refering", "answer_template": "The \"the table on which the crockerys and cuttleries are present\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0075, 0.6898829039812646, 0.9985312499999999, 0.9865105386416863], "iscrowd": 0, "area": 59656.703850000005, "rle": {"size": [427, 640], "counts": "R_2b0i<8G100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O1000000000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001N1000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O00000000001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O\\2"}, "label": "the table on which the crockerys and cuttleries are present"}]}
{"id": 21750, "image": "COCO_train2014_000000021750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dinning table with cake and drinks\"."}], "task": "refering", "answer_template": "The \"a dinning table with cake and drinks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0075, 0.6898829039812646, 0.9985312499999999, 0.9865105386416863], "iscrowd": 0, "area": 59656.703850000005, "rle": {"size": [427, 640], "counts": "R_2b0i<8G100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O1000000000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001N1000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O00000000001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O\\2"}, "label": "a dinning table with cake and drinks"}]}
{"id": 300279, "image": "COCO_train2014_000000300279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry car with a big red streak on the back from the lights\"."}], "task": "refering", "answer_template": "The \"a blurry car with a big red streak on the back from the lights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.52078125, 0.5910208333333333, 0.9994375, 1.0], "iscrowd": 0, "area": 51864.9163, "rle": {"size": [480, 640], "counts": "d^l41n>5J6K5K5J6K5K5J7J5K5K4K6K5D<K5K5L3L3M4L4L2N1O1O1N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2O101O0O10000O2O001O0O2O1O0O2O001O0O2O001O0O2O001N101O001N101O001N101O0O2O001O1N3N1O2N1O2N2M10000000000O10000000000O10000000000O10000000000O10000000000O10000000O10O10000000000O10000000000O10000000000O10000000000000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000O1000001O000000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O00000000000000000000000000001O00001O0O101O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0H8eM\\2cMeH"}, "label": "a blurry car with a big red streak on the back from the lights"}]}
{"id": 300279, "image": "COCO_train2014_000000300279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark car drives through a green light\"."}], "task": "refering", "answer_template": "The \"a dark car drives through a green light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.52078125, 0.5910208333333333, 0.9994375, 1.0], "iscrowd": 0, "area": 51864.9163, "rle": {"size": [480, 640], "counts": "d^l41n>5J6K5K5J6K5K5J7J5K5K4K6K5D<K5K5L3L3M4L4L2N1O1O1N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2O101O0O10000O2O001O0O2O1O0O2O001O0O2O001O0O2O001N101O001N101O001N101O0O2O001O1N3N1O2N1O2N2M10000000000O10000000000O10000000000O10000000000O10000000000O10000000O10O10000000000O10000000000O10000000000O10000000000000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000O1000001O000000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O00000000000000000000000000001O00001O0O101O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0H8eM\\2cMeH"}, "label": "a dark car drives through a green light"}]}
{"id": 275709, "image": "COCO_train2014_000000275709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant with its trunck curled under\"."}], "task": "refering", "answer_template": "The \"the elephant with its trunck curled under\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 237, 238, 239, 240, 241, 254, 255, 256, 257, 260, 261, 263, 264, 278, 279, 280, 281, 283, 284, 285, 286, 287, 306, 307, 308], "bbox": [0.04925, 0.11235294117647059, 0.6252656249999999, 0.8786588235294118], "iscrowd": 0, "area": 80119.03289999999, "rle": {"size": [425, 640], "counts": "Y]=:m<:F:kGSO\\3X1ZLoN^3[1XLnN^3]1YLjN_3a1WLdNc3f1SL]Nj3k2oJVMo4d5L3N2N3M2N1O0O2O00aN]JZKc5Y4PKbKo4Q4eKhKZ4m3XLmKh3W4ZLdKe3`4_24L4L4M3N2M3N2M3N3L3N2M3N2M3N4K7J6I7YK[I^2j6[MXId2o6TMSIk2T7nLmHR3Y7fLjHX3]7`LeH`3`7ZLaHe3f7SL]Hl3i7lKYHS4_800O2O0O101N100O101N1O100O2O0O101N100O101N100O2O0000001O000O101O000000O1000O10000000WMbLgK^3P4oLlKQ3o3WMnKi2Q4\\MkKc2V4aMfK_2Y4eMcK\\2\\4iM`KW2_4nM]KR2c4QNZKo1e4VNWKj1h4[NTKe1l4^NQKa1o4dNmJ\\1R5iNjJW1U5nNgJR1Y5QOdJo0[5VOaJj0^5[O]Jf0b5^O[Jb0e5BWJ=i5HSJ8l5LQJ4n51nIOR6d3O1001N100000000000000O100000000000000O101O00000000000001O000000001O000000010O00000000001O000000001O01bLlIEU68mIHT66mIJT63nINR60oI0R6NoI2R6KPJ5Q6IPJ7Q6FQJ;R6@oI`0]6SOdIm0g6fN[IZ1P7ZNQIf1[7lMgHU2c7_M^Ha2n7RMSHn2X8eLkGZ3R9001O000000001O000000001O0000000010O00000001O000000001O0000000hNdLcH]3X7iLfHW3V7oLhHQ3T7TMkHl2P7[MoHd2m6bMQI_2j6gMTIY2h6mMVIS2e6SNZIm1b6YN\\Ig1`6_N^Ib1]6dNaI\\1Z6kNdIU1X6POgIP1Y6ROeIn0\\6SObIn0^6SO`Im0b6SO\\Im0e6SOZIm0g6j22O0O1O2N1O101fKPIW2Q7fMQIZ2P7cMSI\\2n6`MUI`2m6\\MUIe2k6WMXIi2l6QMVIo2Q7gLRIY3U7^LnHa3X7VLkHk3[7lKgHT4W8001O10O01O1O001O10O01O1O001O1fKdKE\\4;jK@V4>QL\\Oo3d0VLWOj3i0\\LQOd3n0aLnN`3Q1dLkN\\3T1hLjNW3V1mLfNS3Z1PMcNQ3[1SMbNm2^1WM^Ni2b1ZM[Nf2f1]MWNc2i1_MTNa2m1aMPN_2Q2dMkM\\2V2fMlLlM_O_4e3hMdLPNDX4i3jM[LVNJo3l3nMRLYN0i3o3PNjK^N4c3R4QNcKcN8\\3V4\\OgKd0Z4^OcKb0^4@`K`0`4iKTLg2YO`1f5Q32O0O1O1O101N1O0001O0000001O00000O2O00001O0000001O0000001O00000O2N1N2O2M2N2O2M2O1N3M2O1I8G8eMiG\\O`8b0SHdNU8[1m1O2N1N2O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O1O2M8I8H7I7I[_S3"}, "label": "the elephant with its trunck curled under"}]}
{"id": 275709, "image": "COCO_train2014_000000275709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant\"."}], "task": "refering", "answer_template": "The \"an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 237, 238, 239, 240, 241, 254, 255, 256, 257, 260, 261, 263, 264, 278, 279, 280, 281, 283, 284, 285, 286, 287, 306, 307, 308], "bbox": [0.04925, 0.11235294117647059, 0.6252656249999999, 0.8786588235294118], "iscrowd": 0, "area": 80119.03289999999, "rle": {"size": [425, 640], "counts": "Y]=:m<:F:kGSO\\3X1ZLoN^3[1XLnN^3]1YLjN_3a1WLdNc3f1SL]Nj3k2oJVMo4d5L3N2N3M2N1O0O2O00aN]JZKc5Y4PKbKo4Q4eKhKZ4m3XLmKh3W4ZLdKe3`4_24L4L4M3N2M3N2M3N3L3N2M3N2M3N4K7J6I7YK[I^2j6[MXId2o6TMSIk2T7nLmHR3Y7fLjHX3]7`LeH`3`7ZLaHe3f7SL]Hl3i7lKYHS4_800O2O0O101N100O101N1O100O2O0O101N100O101N100O2O0000001O000O101O000000O1000O10000000WMbLgK^3P4oLlKQ3o3WMnKi2Q4\\MkKc2V4aMfK_2Y4eMcK\\2\\4iM`KW2_4nM]KR2c4QNZKo1e4VNWKj1h4[NTKe1l4^NQKa1o4dNmJ\\1R5iNjJW1U5nNgJR1Y5QOdJo0[5VOaJj0^5[O]Jf0b5^O[Jb0e5BWJ=i5HSJ8l5LQJ4n51nIOR6d3O1001N100000000000000O100000000000000O101O00000000000001O000000001O000000010O00000000001O000000001O01bLlIEU68mIHT66mIJT63nINR60oI0R6NoI2R6KPJ5Q6IPJ7Q6FQJ;R6@oI`0]6SOdIm0g6fN[IZ1P7ZNQIf1[7lMgHU2c7_M^Ha2n7RMSHn2X8eLkGZ3R9001O000000001O000000001O0000000010O00000001O000000001O0000000hNdLcH]3X7iLfHW3V7oLhHQ3T7TMkHl2P7[MoHd2m6bMQI_2j6gMTIY2h6mMVIS2e6SNZIm1b6YN\\Ig1`6_N^Ib1]6dNaI\\1Z6kNdIU1X6POgIP1Y6ROeIn0\\6SObIn0^6SO`Im0b6SO\\Im0e6SOZIm0g6j22O0O1O2N1O101fKPIW2Q7fMQIZ2P7cMSI\\2n6`MUI`2m6\\MUIe2k6WMXIi2l6QMVIo2Q7gLRIY3U7^LnHa3X7VLkHk3[7lKgHT4W8001O10O01O1O001O10O01O1O001O1fKdKE\\4;jK@V4>QL\\Oo3d0VLWOj3i0\\LQOd3n0aLnN`3Q1dLkN\\3T1hLjNW3V1mLfNS3Z1PMcNQ3[1SMbNm2^1WM^Ni2b1ZM[Nf2f1]MWNc2i1_MTNa2m1aMPN_2Q2dMkM\\2V2fMlLlM_O_4e3hMdLPNDX4i3jM[LVNJo3l3nMRLYN0i3o3PNjK^N4c3R4QNcKcN8\\3V4\\OgKd0Z4^OcKb0^4@`K`0`4iKTLg2YO`1f5Q32O0O1O1O101N1O0001O0000001O00000O2O00001O0000001O0000001O00000O2N1N2O2M2N2O2M2O1N3M2O1I8G8eMiG\\O`8b0SHdNU8[1m1O2N1N2O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O1O2M8I8H7I7I[_S3"}, "label": "an elephant"}]}
{"id": 275709, "image": "COCO_train2014_000000275709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby elephant is in between two other elephants\"."}], "task": "refering", "answer_template": "The \"the baby elephant is in between two other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 174, 175, 194, 195, 197, 198, 199, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 266, 267, 268, 287, 289, 290, 291], "bbox": [0.464546875, 0.49988235294117644, 0.6789531249999999, 0.8720941176470588], "iscrowd": 0, "area": 12024.533799999996, "rle": {"size": [425, 640], "counts": "`_k31X=:F:F7I3M4L3M3M3M3M3M3M3M3M3M3`FSN^7P2]HSNb7P2YHTNe7o1UHUNj7n1QHUNP8l1kGXNT88mFc0h0YO\\8NSGi0<]O`8EZGP10^Og8\\O`GW1BAn8SOfG]1WOBU9kNkGj2\\9O0O001O100O1TNiEd0X:[OWF6k9HeFI[97TGYOl8g0k1001O0000001O000O101O0000001N10001O000O101O0000001O0O1K6K4dEdNf8b1XG`Nc8e1[G^N^8h1`GZNZ8l1dGUNW8Q2gGPNT8V2jGkMR8Z2lGgMo7l2bGUMY8_3WGbLd8T4L4K51O000000001O000000010O0000010O000000010ZOe0[Oe03M3N3M2N2N2N2M4M2N2N00N2O1N102M2O2M2O2N2M2O2M2O2M2O2M2O2M2O2M2O2N1O2M[1fNg0YOh0XOigd2"}, "label": "the baby elephant is in between two other elephants"}]}
{"id": 275709, "image": "COCO_train2014_000000275709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby elephant\"."}], "task": "refering", "answer_template": "The \"the baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 174, 175, 194, 195, 197, 198, 199, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 266, 267, 268, 287, 289, 290, 291], "bbox": [0.464546875, 0.49988235294117644, 0.6789531249999999, 0.8720941176470588], "iscrowd": 0, "area": 12024.533799999996, "rle": {"size": [425, 640], "counts": "`_k31X=:F:F7I3M4L3M3M3M3M3M3M3M3M3M3`FSN^7P2]HSNb7P2YHTNe7o1UHUNj7n1QHUNP8l1kGXNT88mFc0h0YO\\8NSGi0<]O`8EZGP10^Og8\\O`GW1BAn8SOfG]1WOBU9kNkGj2\\9O0O001O100O1TNiEd0X:[OWF6k9HeFI[97TGYOl8g0k1001O0000001O000O101O0000001N10001O000O101O0000001O0O1K6K4dEdNf8b1XG`Nc8e1[G^N^8h1`GZNZ8l1dGUNW8Q2gGPNT8V2jGkMR8Z2lGgMo7l2bGUMY8_3WGbLd8T4L4K51O000000001O000000010O0000010O000000010ZOe0[Oe03M3N3M2N2N2N2M4M2N2N00N2O1N102M2O2M2O2N2M2O2M2O2M2O2M2O2M2O2M2O2N1O2M[1fNg0YOh0XOigd2"}, "label": "the baby elephant"}]}
{"id": 275709, "image": "COCO_train2014_000000275709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant who is standing behind and baby elephant whose front portion is seen\"."}], "task": "refering", "answer_template": "The \"the elephant who is standing behind and baby elephant whose front portion is seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291], "bbox": [0.46364062500000003, 0.16778823529411765, 0.8854218750000001, 0.8254588235294118], "iscrowd": 0, "area": 34915.47289999999, "rle": {"size": [425, 640], "counts": "bUl31W=1000001O000O2O0000001O0O10001O00000O2O0000001O000O101O00001O00000O2O0000001O00000O2O0000001O00000O2O01O0001O00010O0000001O01O0001O0001O01O00000bE[OX8e0gGLj73UH>[7CcHj0Q7VOnHk0R7UOmHm0R7SOnHm0R7SOmHo0R7ROlHo0T7QOkHP1U7POjHR1U7nNjHS1V7mNiHT1W7mNgHU1X7kNhHU1X7kNgHW1X7iNgHX1Y7hNfHY1[7fNdH\\1[7eNcH\\1]7dNbH]1^7cNaH_1^7eN^H[1b7iNYHY1f7kNUHV1k7nNPHS1P8QOkGQ1T8SOgGn0Y8UOcGl0]8XO^Gi0b8[OZGf0e8o1000001O0000001O1O2N1O2O1N2N1O2N2O1N1O2N2N1O2O1N2N1O2N2N101N>Bd0\\Oc0]Oc0]Ob0_ON1O100O100O2N100O1O100O1O100O100O1O100O1O10000O10001N100O101O0O10001N100O10001N10000O2O0VMZJPNg5m1fJiMZ5S2TKbMm4[2]K]Md4_2fKZM[4d2mKoLZ4Q3mKbLY4^3i20000001O00000000001O00000001O01O01O000010O3M3N2M3M3M2N3M3M3M3M3M7I:F:F9G:F<E:E00001O01O01O0000010O000010O001O010O0010OD=B=D=B=C>B>ULdGZ2l8]M[GU2\\9[MkFV2]:gNX1M4L3M3M5K4L5K5Knan0"}, "label": "the elephant who is standing behind and baby elephant whose front portion is seen"}]}
{"id": 275709, "image": "COCO_train2014_000000275709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large elephant with his trunk hanging down towards the ground\"."}], "task": "refering", "answer_template": "The \"a large elephant with his trunk hanging down towards the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291], "bbox": [0.46364062500000003, 0.16778823529411765, 0.8854218750000001, 0.8254588235294118], "iscrowd": 0, "area": 34915.47289999999, "rle": {"size": [425, 640], "counts": "bUl31W=1000001O000O2O0000001O0O10001O00000O2O0000001O000O101O00001O00000O2O0000001O00000O2O0000001O00000O2O01O0001O00010O0000001O01O0001O0001O01O00000bE[OX8e0gGLj73UH>[7CcHj0Q7VOnHk0R7UOmHm0R7SOnHm0R7SOmHo0R7ROlHo0T7QOkHP1U7POjHR1U7nNjHS1V7mNiHT1W7mNgHU1X7kNhHU1X7kNgHW1X7iNgHX1Y7hNfHY1[7fNdH\\1[7eNcH\\1]7dNbH]1^7cNaH_1^7eN^H[1b7iNYHY1f7kNUHV1k7nNPHS1P8QOkGQ1T8SOgGn0Y8UOcGl0]8XO^Gi0b8[OZGf0e8o1000001O0000001O1O2N1O2O1N2N1O2N2O1N1O2N2N1O2O1N2N1O2N2N101N>Bd0\\Oc0]Oc0]Ob0_ON1O100O100O2N100O1O100O1O100O100O1O100O1O10000O10001N100O101O0O10001N100O10001N10000O2O0VMZJPNg5m1fJiMZ5S2TKbMm4[2]K]Md4_2fKZM[4d2mKoLZ4Q3mKbLY4^3i20000001O00000000001O00000001O01O01O000010O3M3N2M3M3M2N3M3M3M3M3M7I:F:F9G:F<E:E00001O01O01O0000010O000010O001O010O0010OD=B=D=B=C>B>ULdGZ2l8]M[GU2\\9[MkFV2]:gNX1M4L3M3M5K4L5K5Knan0"}, "label": "a large elephant with his trunk hanging down towards the ground"}]}
{"id": 546046, "image": "COCO_train2014_000000546046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus on the right side of the blue bus\"."}], "task": "refering", "answer_template": "The \"a red bus on the right side of the blue bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296], "bbox": [0.6871875000000001, 0.28981250000000003, 0.9568906250000001, 0.7208541666666667], "iscrowd": 0, "area": 31802.38399999999, "rle": {"size": [480, 640], "counts": "Yb^6;U>`0@`0A?@`0@`0@`0@a0_O`0A?@`0@`0J6N2000000000000000000000000O1000001O0000000000000000000O1000000000000000000000001O000O100000000000000000000000000O1000000000001O0000000000000O10000000000000000000000000001N100000000000000000000000000O10000000000000001O00000000000O10000001O3L4M3L3N2M2O1N3N1O1N2O1N2O2M2O1N2O1N3N1O1N2O1N3N1N2O1N2O2M2O1N2N2M3N3M2N2M3N2N3Mnm<"}, "label": "a red bus on the right side of the blue bus"}]}
{"id": 546046, "image": "COCO_train2014_000000546046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red bus on the right\"."}], "task": "refering", "answer_template": "The \"the red bus on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296], "bbox": [0.6871875000000001, 0.28981250000000003, 0.9568906250000001, 0.7208541666666667], "iscrowd": 0, "area": 31802.38399999999, "rle": {"size": [480, 640], "counts": "Yb^6;U>`0@`0A?@`0@`0@`0@a0_O`0A?@`0@`0J6N2000000000000000000000000O1000001O0000000000000000000O1000000000000000000000001O000O100000000000000000000000000O1000000000001O0000000000000O10000000000000000000000000001N100000000000000000000000000O10000000000000001O00000000000O10000001O3L4M3L3N2M2O1N3N1O1N2O1N2O2M2O1N2O1N3N1O1N2O1N3N1N2O1N2O2M2O1N2N2M3N3M2N2M3N2N3Mnm<"}, "label": "the red bus on the right"}]}
{"id": 546046, "image": "COCO_train2014_000000546046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue bus in the middle\"."}], "task": "refering", "answer_template": "The \"the blue bus in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 288], "bbox": [0.374078125, 0.34075, 0.701859375, 0.7185208333333333], "iscrowd": 0, "area": 32014.07625, "rle": {"size": [480, 640], "counts": "gY`31h>?A>C=B>B?A>mC[MW;R3\\DTM];[3\\N_LbGh3^8XL[Go3e8QLTGV4l8jKnF\\4Q9fKgFa4Y9_K`Fh4`9XKYFo4g9500000000000000O1000000000000000000O100000000000000000000O10O10000000000000000O0100000000000000000O0100000000000000000O10O100000000000000000000O1000000000000000000O100000000000:FO10000O10001O0O101O0O101O000O2O000O2O00001N10001N1000001N10001N10001O0O101O000O2O000O2O00001N10001N10001O0O101O0O101O000O2O000O2O00000O2O000O2O00001N10001N10001O0O101O000O2O000O2O00001N10001N10001O0I7\\Oe0[Od0]Oc0\\Oe0[O_[i2"}, "label": "the blue bus in the middle"}]}
{"id": 546046, "image": "COCO_train2014_000000546046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bus between two other buses\"."}], "task": "refering", "answer_template": "The \"a blue bus between two other buses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 288], "bbox": [0.374078125, 0.34075, 0.701859375, 0.7185208333333333], "iscrowd": 0, "area": 32014.07625, "rle": {"size": [480, 640], "counts": "gY`31h>?A>C=B>B?A>mC[MW;R3\\DTM];[3\\N_LbGh3^8XL[Go3e8QLTGV4l8jKnF\\4Q9fKgFa4Y9_K`Fh4`9XKYFo4g9500000000000000O1000000000000000000O100000000000000000000O10O10000000000000000O0100000000000000000O0100000000000000000O10O100000000000000000000O1000000000000000000O100000000000:FO10000O10001O0O101O0O101O000O2O000O2O00001N10001N1000001N10001N10001O0O101O000O2O000O2O00001N10001N10001O0O101O0O101O000O2O000O2O00000O2O000O2O00001N10001N10001O0O101O000O2O000O2O00001N10001N10001O0I7\\Oe0[Od0]Oc0\\Oe0[O_[i2"}, "label": "a blue bus between two other buses"}]}
{"id": 472320, "image": "COCO_train2014_000000472320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small serving of beans in a small white dish\"."}], "task": "refering", "answer_template": "The \"a small serving of beans in a small white dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 156, 157, 158, 159, 169, 170, 171, 172, 182, 183, 184, 185, 195, 196], "bbox": [0.0, 0.49421875000000004, 0.3136288088642659, 0.683], "iscrowd": 0, "area": 10321.78135, "rle": {"size": [640, 361], "counts": "h:?^c0n0QOc0@4L3N3L4L3M3N1N3M3N2M5K6K3L2O001O0O10000O1000O10O10000O10000O10000O1000O010000O10O1O001O1O1O1O1O001O1O1O1O1O100O100O00101N100O100O100O100O2O0O100O100O100000001O000000001N1O2O0O2M2O2N2L3M4N1N3N2N2N1O2O1N2N2M3L6I7L3L5K5K7IRej4"}, "label": "a small serving of beans in a small white dish"}]}
{"id": 472320, "image": "COCO_train2014_000000472320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of beans\"."}], "task": "refering", "answer_template": "The \"a bowl of beans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 156, 157, 158, 159, 169, 170, 171, 172, 182, 183, 184, 185, 195, 196], "bbox": [0.0, 0.49421875000000004, 0.3136288088642659, 0.683], "iscrowd": 0, "area": 10321.78135, "rle": {"size": [640, 361], "counts": "h:?^c0n0QOc0@4L3N3L4L3M3N1N3M3N2M5K6K3L2O001O0O10000O1000O10O10000O10000O10000O1000O010000O10O1O001O1O1O1O1O001O1O1O1O1O100O100O00101N100O100O100O100O2O0O100O100O100000001O000000001N1O2O0O2M2O2N2L3M4N1N3N2N2N1O2O1N2N2M3L6I7L3L5K5K7IRej4"}, "label": "a bowl of beans"}]}
{"id": 472320, "image": "COCO_train2014_000000472320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"spoon in cup of beans\"."}], "task": "refering", "answer_template": "The \"spoon in cup of beans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 107, 108, 120, 132, 133, 144, 145, 156, 157, 158], "bbox": [0.05072022160664819, 0.333296875, 0.343601108033241, 0.5578749999999999], "iscrowd": 0, "area": 2424.7153500000004, "rle": {"size": [640, 361], "counts": "QW<2mc00O2O2M3N1N100O101O1N20O0100O01000O010O10O10O10O1000000001O00000O2O0O1O2N1O1N2N1O2N2N2NAU]OMib02\\]OLcb02b]OK]b05g]OHWb08m]OFRb0:R^OBna0>i01N1000000O1000001N100O1O2O0O2N1O101N1O2N101N100O2O0O2O0O101N1O2O0O1100O2O0O2O1N1O3N1N2O1N2N3N2N2M3N2N2M3N5KPmc4"}, "label": "spoon in cup of beans"}]}
{"id": 5377, "image": "COCO_train2014_000000005377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing in front of sign\"."}], "task": "refering", "answer_template": "The \"woman standing in front of sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 146, 147, 148, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209], "bbox": [0.0025825825825825824, 0.61652, 0.5212612612612613, 0.9990399999999999], "iscrowd": 0, "area": 18791.540000000005, "rle": {"size": [500, 333], "counts": "Wm03\\?6L4N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N1O2M3N2N2N2N1O2N2M3N2N2N1O2N2N2M3N2N1N3L4L4L4N2O0O2O1O1N2O1N101O0O10000O100O1\\DeLo:[3mDiLS;W3iDmLV;T3fDPMZ;P3bDSM_;a3O100O1M3L4M3L4M3M3L4M3L4M3M3N2M3M3N2M300O1O100O1O100O1O100O100O1O10O2O001O00001O00001O001O00001O001O00001O00001O1O;E;[K`E`3Z;00000001O2N1O1O1O2N1N2O2N1O1O1O2N1O1CnCVMT<h2QDSMP<k2>N3N1N2O1N3N1N2O2M2N2L4M4L3M3M4L4K4M4L3M4L3M4K4M4L3M4L3M4K4Mdf]2"}, "label": "woman standing in front of sign"}]}
{"id": 5377, "image": "COCO_train2014_000000005377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women looking at the water\"."}], "task": "refering", "answer_template": "The \"a women looking at the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 146, 147, 148, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209], "bbox": [0.0025825825825825824, 0.61652, 0.5212612612612613, 0.9990399999999999], "iscrowd": 0, "area": 18791.540000000005, "rle": {"size": [500, 333], "counts": "Wm03\\?6L4N2N2N2M2O2N2N2N2N2N1N3N2N2N2N2N1O2M3N2N2N2N1O2N2M3N2N2N1O2N2N2M3N2N1N3L4L4L4N2O0O2O1O1N2O1N101O0O10000O100O1\\DeLo:[3mDiLS;W3iDmLV;T3fDPMZ;P3bDSM_;a3O100O1M3L4M3L4M3M3L4M3L4M3M3N2M3M3N2M300O1O100O1O100O1O100O100O1O10O2O001O00001O00001O001O00001O001O00001O00001O1O;E;[K`E`3Z;00000001O2N1O1O1O2N1N2O2N1O1O1O2N1O1CnCVMT<h2QDSMP<k2>N3N1N2O1N3N1N2O2M2N2L4M4L3M3M4L4K4M4L3M4L3M4K4M4L3M4L3M4K4Mdf]2"}, "label": "a women looking at the water"}]}
{"id": 5377, "image": "COCO_train2014_000000005377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of someone holding a child on his shoulders\"."}], "task": "refering", "answer_template": "The \"the back of someone holding a child on his shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 212, 213, 214, 215], "bbox": [0.5792792792792792, 0.74374, 1.0, 0.99612], "iscrowd": 0, "area": 14967.755999999996, "rle": {"size": [500, 333], "counts": "aen25^?9G6XO]OSBf0i=_OTBb0j=CRB?j=GQB;m=h0L4M2M4M2M4L3N3L3N1N3N1N3N2O010O00100O010O0nC^MS;c2cDgM];X2[DQNd;P301N100O100000000000000000000000000000000000000000000000000001O000000000000000001O000000000000000000000000000000000000000000000000000001O0000000000000000000001O00000000000000000001O000fNXDhNh;Z1]D_Nd;a1aDYN_;h1gDQNY;P2mDiMS;X2SEaMm:a2WEYMj:g2R1000100N2N1O2N2QOn0oNiD"}, "label": "the back of someone holding a child on his shoulders"}]}
{"id": 5377, "image": "COCO_train2014_000000005377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of the dad , standing on right\"."}], "task": "refering", "answer_template": "The \"back of the dad , standing on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 199, 200, 201, 202, 203, 212, 213, 214, 215], "bbox": [0.5792792792792792, 0.74374, 1.0, 0.99612], "iscrowd": 0, "area": 14967.755999999996, "rle": {"size": [500, 333], "counts": "aen25^?9G6XO]OSBf0i=_OTBb0j=CRB?j=GQB;m=h0L4M2M4M2M4L3N3L3N1N3N1N3N2O010O00100O010O0nC^MS;c2cDgM];X2[DQNd;P301N100O100000000000000000000000000000000000000000000000000001O000000000000000001O000000000000000000000000000000000000000000000000000001O0000000000000000000001O00000000000000000001O000fNXDhNh;Z1]D_Nd;a1aDYN_;h1gDQNY;P2mDiMS;X2SEaMm:a2WEYMj:g2R1000100N2N1O2N2QOn0oNiD"}, "label": "back of the dad , standing on right"}]}
{"id": 5377, "image": "COCO_train2014_000000005377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby on the man ' s shoulders\"."}], "task": "refering", "answer_template": "The \"the baby on the man ' s shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 116, 117, 118, 128, 129, 130, 140, 141, 142, 152, 153, 154, 164, 165, 166, 167], "bbox": [0.6748348348348349, 0.49213999999999997, 0.9312612612612613, 0.7527999999999999], "iscrowd": 0, "area": 7456.230249999998, "rle": {"size": [500, 333], "counts": "[W^31`?3N2M3N2N2M3N2M3_N@lCb0Q<FhC<V<JdC7Z<1`C0^<6]CKa<<YCEd<c0VC^Oh<h0SCYOj<j1L4L4K5L4K5L4O1O1O1O1N2O1O1O1O1N2O1O1000O1000000000001O000000000000000000001O2N1O2N1O2N2N1O2N1O2N1O2M3N1O2N7iN]CoNk<g0YCWOo<`0SC_OU=7oBGZ=NjB0a=DaB;^>N1O2N2N1O2N1M4KRl:"}, "label": "the baby on the man ' s shoulders"}]}
{"id": 5377, "image": "COCO_train2014_000000005377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl sitting on man ' s shoulders\"."}], "task": "refering", "answer_template": "The \"a girl sitting on man ' s shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 116, 117, 118, 128, 129, 130, 140, 141, 142, 152, 153, 154, 164, 165, 166, 167], "bbox": [0.6748348348348349, 0.49213999999999997, 0.9312612612612613, 0.7527999999999999], "iscrowd": 0, "area": 7456.230249999998, "rle": {"size": [500, 333], "counts": "[W^31`?3N2M3N2N2M3N2M3_N@lCb0Q<FhC<V<JdC7Z<1`C0^<6]CKa<<YCEd<c0VC^Oh<h0SCYOj<j1L4L4K5L4K5L4O1O1O1O1N2O1O1O1O1N2O1O1000O1000000000001O000000000000000000001O2N1O2N1O2N2N1O2N1O2N1O2M3N1O2N7iN]CoNk<g0YCWOo<`0SC_OU=7oBGZ=NjB0a=DaB;^>N1O2N2N1O2N1M4KRl:"}, "label": "a girl sitting on man ' s shoulders"}]}
{"id": 349442, "image": "COCO_train2014_000000349442.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat plane parked by the shore\"."}], "task": "refering", "answer_template": "The \"a boat plane parked by the shore\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 152, 153, 155], "bbox": [0.016640625, 0.06942748091603054, 0.9240937499999999, 0.7849999999999999], "iscrowd": 0, "area": 49417.643950000005, "rle": {"size": [262, 640], "counts": "ck22Q83N2N2M3N2N2O100O1O100O1O100O2N100O1O100O1O100O1O100O100O1000000000000000001O000O1000000000000000000000000O1000000jKnNX1R1fNPOZ1P1eNROZ1o0cNTO]1k0aNWO_1i0`NYO_1g0_N[Oa1e0^N]Oa1c0]N@b1`0]NAc1?]NBb1>^NBb1>^NCb1<^NEa1;_NDb1<^NDb1<^NCc1=]NCc1=]NCc1=]NBd1>\\NBd1>\\NAe1?[NAf1>ZNAg1?YNAg1`0XN@h1`0XN_Oi1a0WN_Oi1a0WN_Oi1a0VNAi1?WNAi1?WNAj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBk1=UNCk1=UNCk1=UNCk1=UNCk1=UNCk1=UNCk1>TNBl1>TNBl1>TNBm1=RNDn1<RNDn1<RNDn1<RNEm1;SNEm1;SNEm1;SNEm1;SNEm1;SNEm1;SNEn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFo1:PNFP2:PNFP2:PNFP2:oMGQ29oMGQ29oMGQ29oMGQ29oMGQ29oMGQ29oMHQ27oMIQ27oMIQ27oMIQ27oMIQ27oMIQ27oMIQ2h0^MXOb2`1fL`NZ3X2nKhMS4U30000000000000000001O000000000001O0001O0000000000000000001O0000000000000000001O000001O000000000001O0000000000000000001O0000000000000000010O0000000000000000001O0000000000000000001O0000000001O00000001O0000000000000000001O0000000000000000001O0001O0000000001O000000000000001O000000000000001O0000000000000010O00000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0001O0000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000001O0000001O2N1O2N2N1O2N2M3N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2M3N1O2NN2M3N2M3M3M3M3M3M3M3M3M2N3M3M3M4M20000000000000000000000000000000001O0000000001O1O1O001O1O1O001O1O1O00100O001O1O1O001O1O1O001O1O001O1O1O0010SOiKjMV4U2nKiMQ4W2RLfMn3Z2ULdMj3[2ZLcMe3]2^LaMa3_2bL^M^3b2eL\\MZ3c2jL[MU3e2nLXMR3h2QMVMn2j2TMUMk2j2YMTMf2l2Z10000O10001dMgJi1Y5VNiJh1X5VNjJj1V5VNkJi1U5WNkJi1U5WNlJW1e5iN]JT1d5lN]JQ1e5oN]Jl0f5TO\\Jh0f5XO\\Jc0g5]OZJ?i5AWJ>j5BWJ<j5DWJ;i5EWJ:j5FUJ;k5EUJ:m5ERJ<n5DRJ;o5EPJ7U6IkI1[6OdIN`62`IMa63^INb62^IMc63\\IMe63[ILf64YILh64XILh64WILj64VIKk65TIKm65SIJn66QIKo65QIJP76oHJR76?0000000O1000000000000000000000000O1000O10000O100O10000O10000O10000O10000O102N1N2O[_<"}, "label": "a boat plane parked by the shore"}]}
{"id": 349442, "image": "COCO_train2014_000000349442.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white , black and red seaplane\"."}], "task": "refering", "answer_template": "The \"a white , black and red seaplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 152, 153, 155], "bbox": [0.016640625, 0.06942748091603054, 0.9240937499999999, 0.7849999999999999], "iscrowd": 0, "area": 49417.643950000005, "rle": {"size": [262, 640], "counts": "ck22Q83N2N2M3N2N2O100O1O100O1O100O2N100O1O100O1O100O1O100O100O1000000000000000001O000O1000000000000000000000000O1000000jKnNX1R1fNPOZ1P1eNROZ1o0cNTO]1k0aNWO_1i0`NYO_1g0_N[Oa1e0^N]Oa1c0]N@b1`0]NAc1?]NBb1>^NBb1>^NCb1<^NEa1;_NDb1<^NDb1<^NCc1=]NCc1=]NCc1=]NBd1>\\NBd1>\\NAe1?[NAf1>ZNAg1?YNAg1`0XN@h1`0XN_Oi1a0WN_Oi1a0WN_Oi1a0VNAi1?WNAi1?WNAj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBj1>VNBk1=UNCk1=UNCk1=UNCk1=UNCk1=UNCk1=UNCk1>TNBl1>TNBl1>TNBm1=RNDn1<RNDn1<RNDn1<RNEm1;SNEm1;SNEm1;SNEm1;SNEm1;SNEm1;SNEn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFn1:RNFo1:PNFP2:PNFP2:PNFP2:oMGQ29oMGQ29oMGQ29oMGQ29oMGQ29oMGQ29oMHQ27oMIQ27oMIQ27oMIQ27oMIQ27oMIQ27oMIQ2h0^MXOb2`1fL`NZ3X2nKhMS4U30000000000000000001O000000000001O0001O0000000000000000001O0000000000000000001O000001O000000000001O0000000000000000001O0000000000000000010O0000000000000000001O0000000000000000001O0000000001O00000001O0000000000000000001O0000000000000000001O0001O0000000001O000000000000001O000000000000001O0000000000000010O00000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0001O0000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000001O0000001O2N1O2N2N1O2N2M3N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2M3N1O2NN2M3N2M3M3M3M3M3M3M3M3M2N3M3M3M4M20000000000000000000000000000000001O0000000001O1O1O001O1O1O001O1O1O00100O001O1O1O001O1O1O001O1O001O1O1O0010SOiKjMV4U2nKiMQ4W2RLfMn3Z2ULdMj3[2ZLcMe3]2^LaMa3_2bL^M^3b2eL\\MZ3c2jL[MU3e2nLXMR3h2QMVMn2j2TMUMk2j2YMTMf2l2Z10000O10001dMgJi1Y5VNiJh1X5VNjJj1V5VNkJi1U5WNkJi1U5WNlJW1e5iN]JT1d5lN]JQ1e5oN]Jl0f5TO\\Jh0f5XO\\Jc0g5]OZJ?i5AWJ>j5BWJ<j5DWJ;i5EWJ:j5FUJ;k5EUJ:m5ERJ<n5DRJ;o5EPJ7U6IkI1[6OdIN`62`IMa63^INb62^IMc63\\IMe63[ILf64YILh64XILh64WILj64VIKk65TIKm65SIJn66QIKo65QIJP76oHJR76?0000000O1000000000000000000000000O1000O10000O100O10000O10000O10000O10000O102N1N2O[_<"}, "label": "a white , black and red seaplane"}]}
{"id": 324871, "image": "COCO_train2014_000000324871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white shirt riding an elephant\"."}], "task": "refering", "answer_template": "The \"a man in white shirt riding an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 36, 37, 48, 49, 50, 51, 52, 53, 64, 65, 66, 67, 68, 69, 80, 81, 82, 83, 84], "bbox": [0.25992974238875877, 0.077703125, 0.6392974238875878, 0.262390625], "iscrowd": 0, "area": 9915.11295, "rle": {"size": [640, 427], "counts": "P_U29gc01N2O1O1O1O001O1N2O1O1O1O1O1O1N2O1O1O001O1O001O00001O001O001O1O2N1O2N1O2N1O2N1O001O00001O001O001O0100O1O10O0100O2O0O2O10POmNo^OU1j`0ROU_OP1f`0UOW_Om0e`0XOY_Oi0d`0ZOZ_Oh0b`0[O^_Of0_`0\\O`_Of0\\`0]Oc_Od0Z`0@d_O`0[`0Bd_O>Z`0Ee_O;Z`0Hd_O8[`0Jd_O5\\`0Mc_O3\\`0i1O1O1000000000000000O100000001O000000001O0O101O001O001O001O00001N101O002N1O2N2SNT_Ob0Pa0UOV_Oj0m`0QOU_Oo0o`0jNU_OU1ka0O2O00001O001O001O00001N101O001O00001O001O1O001N2O1O1O1O002N1O2N1O2N1N3N1O2N1O2N5K4L4L4LSSP3"}, "label": "a man in white shirt riding an elephant"}]}
{"id": 324871, "image": "COCO_train2014_000000324871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on an elephant resting his arm on his friend ' s shoulder\"."}], "task": "refering", "answer_template": "The \"a man on an elephant resting his arm on his friend ' s shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 36, 37, 48, 49, 50, 51, 52, 53, 64, 65, 66, 67, 68, 69, 80, 81, 82, 83, 84], "bbox": [0.25992974238875877, 0.077703125, 0.6392974238875878, 0.262390625], "iscrowd": 0, "area": 9915.11295, "rle": {"size": [640, 427], "counts": "P_U29gc01N2O1O1O1O001O1N2O1O1O1O1O1O1N2O1O1O001O1O001O00001O001O001O1O2N1O2N1O2N1O2N1O001O00001O001O001O0100O1O10O0100O2O0O2O10POmNo^OU1j`0ROU_OP1f`0UOW_Om0e`0XOY_Oi0d`0ZOZ_Oh0b`0[O^_Of0_`0\\O`_Of0\\`0]Oc_Od0Z`0@d_O`0[`0Bd_O>Z`0Ee_O;Z`0Hd_O8[`0Jd_O5\\`0Mc_O3\\`0i1O1O1000000000000000O100000001O000000001O0O101O001O001O001O00001N101O002N1O2N2SNT_Ob0Pa0UOV_Oj0m`0QOU_Oo0o`0jNU_OU1ka0O2O00001O001O001O00001N101O001O00001O001O1O001N2O1O1O1O002N1O2N1O2N1N3N1O2N1O2N5K4L4L4LSSP3"}, "label": "a man on an elephant resting his arm on his friend ' s shoulder"}]}
{"id": 324871, "image": "COCO_train2014_000000324871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a purple shirt riding an elephant\"."}], "task": "refering", "answer_template": "The \"a man with a purple shirt riding an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 32, 33, 34, 48, 49, 62, 63, 64, 65, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 111, 122, 123, 124, 125, 126], "bbox": [0.10988290398126464, 0.0698125, 0.44578454332552697, 0.37740625000000005], "iscrowd": 0, "area": 10035.158600000002, "rle": {"size": [640, 427], "counts": "Vam0:cc05L3L4M2M3N2N2N2N2M3N1O2N2N2M3N2N2N2N2N2M3N20O01O100O100O1e_ORNc>m1^AVN`>i1aAXN]>h1dAZNZ>g1eAYNa=GbAR2l0WN`=JcAP2l0VNi<ORBL8Q2l0TNg<<iAAd0o1l0UNc<h0_BT1n0TNa<S1XBi0W1TNa<T1VBi0X1TNa<U1VBg0Y1UN`<U1UBg0[1TN_<W1UBe0\\1TN_<W1TBf0]1SN^<Y1TBd0^1SN^<Y1TBc0_1TN]<Z1SBb0`1TN]<Z1RBc0a1SN\\<f3dCZL\\<f3dC[L[<e3eC[L[<e3eC[L[<e3eC[L[<e3eC[L[<d3eC]LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC]LY<b3hC^LX<U1WBf0a1TNZ<S1ZBf0\\1WNZ<Q1`Bd0V1[NZ<o0cBe0R1]N[<n0cBe0R1]N\\<l0dBf0P1^N\\<l0eBe0o0_N\\<l0eBe0o0_N\\<k0gBd0n0aN\\<j0fBe0n0`N^<j0eBe0m0aN`<h0cBg0m0aNa<g0cBg0l0bNb<f0bBh0l0bNc<d0cBi0j0cNf<<eBQ1e0cNj<3gBY1?dNn<JhBb1:dNa>Z1_AfNc>Y1]AgNc>Y1]AgNd>X1\\AhNd>X1\\AhNe>W1[AiNe>W1[AiNf>V1ZAjNf>V1ZAjNg>T1ZAkNj>R1VAnNl>P1TAPOo>m0QASOP?l0PASOS?k0l@VOT?j0l@VOU?i0k@WOV?h0j@XOW?g0i@XOX?g0i@YOX?f0h@ZOY?e0g@[OY?e0g@ZO[?e0e@[O\\?d0d@\\O]?c0c@]O^?b0b@]O`?b0`@^Oa?<d@D\\?7h@JY?0l@OV?Ko@5^a0000000O100000000O1000O100000O10000000000O100000000M2NWmc4"}, "label": "a man with a purple shirt riding an elephant"}]}
{"id": 324871, "image": "COCO_train2014_000000324871.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the maroon\"."}], "task": "refering", "answer_template": "The \"the man in the maroon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 32, 33, 34, 48, 49, 62, 63, 64, 65, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 111, 122, 123, 124, 125, 126], "bbox": [0.10988290398126464, 0.0698125, 0.44578454332552697, 0.37740625000000005], "iscrowd": 0, "area": 10035.158600000002, "rle": {"size": [640, 427], "counts": "Vam0:cc05L3L4M2M3N2N2N2N2M3N1O2N2N2M3N2N2N2N2N2M3N20O01O100O100O1e_ORNc>m1^AVN`>i1aAXN]>h1dAZNZ>g1eAYNa=GbAR2l0WN`=JcAP2l0VNi<ORBL8Q2l0TNg<<iAAd0o1l0UNc<h0_BT1n0TNa<S1XBi0W1TNa<T1VBi0X1TNa<U1VBg0Y1UN`<U1UBg0[1TN_<W1UBe0\\1TN_<W1TBf0]1SN^<Y1TBd0^1SN^<Y1TBc0_1TN]<Z1SBb0`1TN]<Z1RBc0a1SN\\<f3dCZL\\<f3dC[L[<e3eC[L[<e3eC[L[<e3eC[L[<e3eC[L[<d3eC]LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC\\LZ<d3fC]LY<b3hC^LX<U1WBf0a1TNZ<S1ZBf0\\1WNZ<Q1`Bd0V1[NZ<o0cBe0R1]N[<n0cBe0R1]N\\<l0dBf0P1^N\\<l0eBe0o0_N\\<l0eBe0o0_N\\<k0gBd0n0aN\\<j0fBe0n0`N^<j0eBe0m0aN`<h0cBg0m0aNa<g0cBg0l0bNb<f0bBh0l0bNc<d0cBi0j0cNf<<eBQ1e0cNj<3gBY1?dNn<JhBb1:dNa>Z1_AfNc>Y1]AgNc>Y1]AgNd>X1\\AhNd>X1\\AhNe>W1[AiNe>W1[AiNf>V1ZAjNf>V1ZAjNg>T1ZAkNj>R1VAnNl>P1TAPOo>m0QASOP?l0PASOS?k0l@VOT?j0l@VOU?i0k@WOV?h0j@XOW?g0i@XOX?g0i@YOX?f0h@ZOY?e0g@[OY?e0g@ZO[?e0e@[O\\?d0d@\\O]?c0c@]O^?b0b@]O`?b0`@^Oa?<d@D\\?7h@JY?0l@OV?Ko@5^a0000000O100000000O1000O100000O10000000000O100000000M2NWmc4"}, "label": "the man in the maroon"}]}
{"id": 177418, "image": "COCO_train2014_000000177418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and yellow taxi in the traffic having number 6897\"."}], "task": "refering", "answer_template": "The \"a black and yellow taxi in the traffic having number 6897\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298], "bbox": [0.71790625, 0.5484235294117648, 0.9975937500000001, 0.8299529411764707], "iscrowd": 0, "area": 17085.27495, "rle": {"size": [425, 640], "counts": "ohn51V=8H8H8H8H8H8H8H6J2N2O1O1O1O1O1O1O1O1O100O1O0000001O000000001O2O001O1N101O1O0O2O1O001N101O1O1O1N2O2N1O1N2O1O1PFhLm9Z3O10001O0O1000000O1000001N1000000O1000000O2O00000001O0000000001O01O00000000000010O0000000001O000001O000001O000000000010O00000000O101O00000000001O0O10000000001O00000O1000001O1O1O1O1O2M2O1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O1O1N2O0000000000001O0O10000000Rb0"}, "label": "a black and yellow taxi in the traffic having number 6897"}]}
{"id": 177418, "image": "COCO_train2014_000000177418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and yellow car driving down the road\"."}], "task": "refering", "answer_template": "The \"black and yellow car driving down the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298], "bbox": [0.71790625, 0.5484235294117648, 0.9975937500000001, 0.8299529411764707], "iscrowd": 0, "area": 17085.27495, "rle": {"size": [425, 640], "counts": "ohn51V=8H8H8H8H8H8H8H6J2N2O1O1O1O1O1O1O1O1O100O1O0000001O000000001O2O001O1N101O1O0O2O1O001N101O1O1O1N2O2N1O1N2O1O1PFhLm9Z3O10001O0O1000000O1000001N1000000O1000000O2O00000001O0000000001O01O00000000000010O0000000001O000001O000001O000000000010O00000000O101O00000000001O0O10000000001O00000O1000001O1O1O1O1O2M2O1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O1O1O1N2O0000000000001O0O10000000Rb0"}, "label": "black and yellow car driving down the road"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottle in the middle\"."}], "task": "refering", "answer_template": "The \"the bottle in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 102, 103, 104, 125, 126, 127, 148, 149, 150, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 377, 378, 379, 380, 381], "bbox": [0.391890625, 0.1835625, 0.59121875, 0.9876041666666667], "iscrowd": 0, "area": 36683.348150000005, "rle": {"size": [480, 640], "counts": "hRf3R1^<`1`N`1`N`1mNS1N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1E;[Oe0[Oe0[Oe0[Oe0\\Od0N2M3N2O100000000000000000O1000001O000000000000000000000000000000000000000000000000000000000hDhNh8[1TGNd78UHKi7;oGHP8f3N2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2VOi0`Na1^Nb1_NXcj3"}, "label": "the bottle in the middle"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle wine bottle made by township seven sitting at a table with two other bottles\"."}], "task": "refering", "answer_template": "The \"the middle wine bottle made by township seven sitting at a table with two other bottles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 102, 103, 104, 125, 126, 127, 148, 149, 150, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 377, 378, 379, 380, 381], "bbox": [0.391890625, 0.1835625, 0.59121875, 0.9876041666666667], "iscrowd": 0, "area": 36683.348150000005, "rle": {"size": [480, 640], "counts": "hRf3R1^<`1`N`1`N`1mNS1N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1E;[Oe0[Oe0[Oe0[Oe0\\Od0N2M3N2O100000000000000000O1000001O000000000000000000000000000000000000000000000000000000000hDhNh8[1TGNd78UHKi7;oGHP8f3N2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2VOi0`Na1^Nb1_NXcj3"}, "label": "the middle wine bottle made by township seven sitting at a table with two other bottles"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty wine glass next to three bottles of red wine\"."}], "task": "refering", "answer_template": "The \"empty wine glass next to three bottles of red wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 203, 204, 205, 206, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 366, 367, 389], "bbox": [0.85434375, 0.46549999999999997, 1.0, 1.0], "iscrowd": 0, "area": 17999.495499999997, "rle": {"size": [480, 640], "counts": "adP83j>4M4K4WE5Y6NRIg0j6^O`HX1[7lNWHb1g7`NkGo1R8TN_GZ2`8gMRGh2m8XMfFU3Y9[1N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N001O0010O01O1O001O001O1O001O001oG`J`6P7000000000000001O0000000O1H8H8G9H8H8K500O10000O100O100O2O0O100O10000O100O100O100O100O10000O1O1N3M2NnH"}, "label": "empty wine glass next to three bottles of red wine"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass on the far right\"."}], "task": "refering", "answer_template": "The \"a wine glass on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 203, 204, 205, 206, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 366, 367, 389], "bbox": [0.85434375, 0.46549999999999997, 1.0, 1.0], "iscrowd": 0, "area": 17999.495499999997, "rle": {"size": [480, 640], "counts": "adP83j>4M4K4WE5Y6NRIg0j6^O`HX1[7lNWHb1g7`NkGo1R8TN_GZ2`8gMRGh2m8XMfFU3Y9[1N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N001O0010O01O1O001O001O1O001O001oG`J`6P7000000000000001O0000000O1H8H8G9H8H8K500O10000O100O100O2O0O100O10000O100O100O100O100O10000O1O1N3M2NnH"}, "label": "a wine glass on the far right"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a township 7 wine bottle standing on a table next to two similar wine bottles and two glasses\"."}], "task": "refering", "answer_template": "The \"a township 7 wine bottle standing on a table next to two similar wine bottles and two glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 73, 74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375], "bbox": [0.10278125, 0.11583333333333333, 0.345296875, 0.9895], "iscrowd": 0, "area": 46621.37685, "rle": {"size": [480, 640], "counts": "g[o0i0S=U1PCYNb;T3mNS1lNT1]Oc0M3M3M3L4M3M3M3N2N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O100O100O1N2N2N2N2N2K5F:E;F:F:E;F:E;F:hNQETNZ;g1lDnM^;n1gDfMd;V2e0K5L40000O1000000000000000O100000000000000000000000000000001O00001O0000001O0000001O00001O0000001O001O001nCHZ99aFi1a7WN[Hn1d7i2N1O2N2N1O2N1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O3M2N3M3M3M2N3M3eMjVT6"}, "label": "a township 7 wine bottle standing on a table next to two similar wine bottles and two glasses"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leftmost of three bottles of township wine\"."}], "task": "refering", "answer_template": "The \"the leftmost of three bottles of township wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 73, 74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375], "bbox": [0.10278125, 0.11583333333333333, 0.345296875, 0.9895], "iscrowd": 0, "area": 46621.37685, "rle": {"size": [480, 640], "counts": "g[o0i0S=U1PCYNb;T3mNS1lNT1]Oc0M3M3M3L4M3M3M3N2N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O100O100O1N2N2N2N2N2K5F:E;F:F:E;F:E;F:hNQETNZ;g1lDnM^;n1gDfMd;V2e0K5L40000O1000000000000000O100000000000000000000000000000001O00001O0000001O0000001O00001O0000001O001O001nCHZ99aFi1a7WN[Hn1d7i2N1O2N2N1O2N1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O3M2N3M3M3M2N3M3eMjVT6"}, "label": "the leftmost of three bottles of township wine"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine glass between the two bottles\"."}], "task": "refering", "answer_template": "The \"wine glass between the two bottles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 314, 336, 359, 382], "bbox": [0.5544687500000001, 0.45725, 0.688890625, 1.0], "iscrowd": 0, "area": 10493.627649999999, "rle": {"size": [480, 640], "counts": "YdV53l>4K5L4K5L4K5L4K5L3M4M3M3M3M3M?Ad0\\Od0\\Oe0[Od0\\O7I1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1^HjJX5W5iIgKW6a600000001O000000H8F:E;E;F:E;G9D<A?@`0A`0PMQF=_:B`2N3N1N3N1N3N1N3N1N2O2M2O2M2O2M2O2M2OmQm2"}, "label": "wine glass between the two bottles"}]}
{"id": 54541, "image": "COCO_train2014_000000054541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"winge glass behind the wine bottles\"."}], "task": "refering", "answer_template": "The \"winge glass behind the wine bottles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 314, 336, 359, 382], "bbox": [0.5544687500000001, 0.45725, 0.688890625, 1.0], "iscrowd": 0, "area": 10493.627649999999, "rle": {"size": [480, 640], "counts": "YdV53l>4K5L4K5L4K5L4K5L3M4M3M3M3M3M?Ad0\\Od0\\Oe0[Od0\\O7I1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1^HjJX5W5iIgKW6a600000001O000000H8F:E;E;F:E;G9D<A?@`0A`0PMQF=_:B`2N3N1N3N1N3N1N3N1N2O2M2O2M2O2M2O2M2OmQm2"}, "label": "winge glass behind the wine bottles"}]}
{"id": 111889, "image": "COCO_train2014_000000111889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy in grey shirt giving a thumbs up\"."}], "task": "refering", "answer_template": "The \"guy in grey shirt giving a thumbs up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 282, 283], "bbox": [0.222578125, 0.057416666666666664, 0.46935937499999997, 0.7197708333333334], "iscrowd": 0, "area": 28723.361649999984, "rle": {"size": [480, 640], "counts": "jiR21m>3L4L5K4M3L5K4I7I8H8K5M4K4L5aDmMj9V2RFnMj9W2QFnMk9U2QFoMk9T2RFPNj9T2RFoMl9S2RFoMk9U2QFmMn9U2PFmMf9^2VFdMV9Q3hFQMS9U3iFmLR9Y3lFiLo8]3mFeLn8j4K4L5N1O2^GeI\\8`6O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O0O2O1N101O1N101O0O2O1O0O2O1I6C>N2N1N3N2N1O2N1OQOo02N3M201N101O001O00001N101O001O00L3H9H8G9H8G9G8E<A?kMYHQMg7e2bH\\M^7[2iHgMW7P2PIRNo6f1XI\\Nh6c1XI^Nh6b1VI`Nj6_1SIeNm6Z1PIjNP7U1nHnNR7Q1kHSOV7k0gHYOZ7e0hHZOY7d0iH[OX7d0iH[OW7d0kH[OW7b0XIQOi6l0YISOi6j0YIUOi6g0ZIXOh6c0[I]Og6>\\IAf6<]ICe68^IGe64^IKe60_INd6M_I3c6I_I6d6E_I:d6A_I>d6]O_Ic0b6YOaIf0b6VOaIh0c:H7J7HaUo4"}, "label": "guy in grey shirt giving a thumbs up"}]}
{"id": 111889, "image": "COCO_train2014_000000111889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man thumbs up holding his blue color bullet\"."}], "task": "refering", "answer_template": "The \"a man thumbs up holding his blue color bullet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 282, 283], "bbox": [0.222578125, 0.057416666666666664, 0.46935937499999997, 0.7197708333333334], "iscrowd": 0, "area": 28723.361649999984, "rle": {"size": [480, 640], "counts": "jiR21m>3L4L5K4M3L5K4I7I8H8K5M4K4L5aDmMj9V2RFnMj9W2QFnMk9U2QFoMk9T2RFPNj9T2RFoMl9S2RFoMk9U2QFmMn9U2PFmMf9^2VFdMV9Q3hFQMS9U3iFmLR9Y3lFiLo8]3mFeLn8j4K4L5N1O2^GeI\\8`6O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O0O2O1N101O1N101O0O2O1O0O2O1I6C>N2N1N3N2N1O2N1OQOo02N3M201N101O001O00001N101O001O00L3H9H8G9H8G9G8E<A?kMYHQMg7e2bH\\M^7[2iHgMW7P2PIRNo6f1XI\\Nh6c1XI^Nh6b1VI`Nj6_1SIeNm6Z1PIjNP7U1nHnNR7Q1kHSOV7k0gHYOZ7e0hHZOY7d0iH[OX7d0iH[OW7d0kH[OW7b0XIQOi6l0YISOi6j0YIUOi6g0ZIXOh6c0[I]Og6>\\IAf6<]ICe68^IGe64^IKe60_INd6M_I3c6I_I6d6E_I:d6A_I>d6]O_Ic0b6YOaIf0b6VOaIh0c:H7J7HaUo4"}, "label": "a man thumbs up holding his blue color bullet"}]}
{"id": 111889, "image": "COCO_train2014_000000111889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in green shirt\"."}], "task": "refering", "answer_template": "The \"woman in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 138, 139, 161, 162, 185], "bbox": [0.0, 0.12218749999999999, 0.14165624999999998, 0.492125], "iscrowd": 0, "area": 9962.843799999999, "rle": {"size": [480, 640], "counts": "i3o1k;Q2iN<O1O1O1N2O2N1O1N2O1O101O00000000000010O00000100O1O100O001O100O1O1O010O1O10000O01000O100000jNTFVMm9f2[FVMe9f2dFUM\\9g2lFRMW9j2XGgLl8U3b1L4M2M4L4L4L4N2M3N2M3N2M3N2M3N2M3N2N2M2O2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N4L4K5L[WQ8"}, "label": "woman in green shirt"}]}
{"id": 111889, "image": "COCO_train2014_000000111889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in green shirt\"."}], "task": "refering", "answer_template": "The \"woman in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 138, 139, 161, 162, 185], "bbox": [0.0, 0.12218749999999999, 0.14165624999999998, 0.492125], "iscrowd": 0, "area": 9962.843799999999, "rle": {"size": [480, 640], "counts": "i3o1k;Q2iN<O1O1O1N2O2N1O1N2O1O101O00000000000010O00000100O1O100O001O100O1O1O010O1O10000O01000O100000jNTFVMm9f2[FVMe9f2dFUM\\9g2lFRMW9j2XGgLl8U3b1L4M2M4L4L4L4N2M3N2M3N2M3N2M3N2M3N2N2M2O2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N4L4K5L[WQ8"}, "label": "woman in green shirt"}]}
{"id": 128282, "image": "COCO_train2014_000000128282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player bending down in right picture\"."}], "task": "refering", "answer_template": "The \"a tennis player bending down in right picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 271, 272, 273, 274, 294, 295, 296, 297, 298, 317, 318, 319, 320], "bbox": [0.7021875, 0.38651972157772624, 0.977625, 0.8988863109048725], "iscrowd": 0, "area": 18692.69869999999, "rle": {"size": [431, 640], "counts": "dUm53g<e0O2O1N2N2N1O2O1N2N1O2O1N2O1O00100O1O100O2N1O101N1O101N1O2O0O1O2N100O2O0O1O100O100UEnMW:S2gEPNX:P2fERNY:n1fEUN=Io8S2bFVN=Jo8Q2cFWN:KQ9P2cFWN9MR9m1cFYN7NT9j1cFZN7NU9i1bF\\N5OV9f1dF]N30X9^2eFfMY9Z3M2O2N1O2M3N1O2iH^K\\5d4cJ^K[5d4cJ^K[5d4cJ]K\\5d4bJ_K[5d4cJ\\K^5d4aJ\\K_5f4^J[Kb5g4\\JYKd5h4[JXKe5j4XJWKi5i4VJXKi5j4UJVKk5l4RJUKn5l4QJTKo5n4nISKR6n4mIRKT6o4iIRKW6P5fIQKZ6P5dIRK[6P5bIQK^6Q5_IPKa6g3VImL7]Od6f3VImL3^Og6e3XIlLO@i6e3ZIcMf6]2\\IaMd6`2]I_Mc6a2^I]Mb6c2`I[Ma6e2`IYM`6g2bIWM_6i2bIUM_6j2cITM]6m2dIQM]6n2fIoLZ6R3gIkL[6U3fIiLZ6W3hIfLZ6Z3gIdLY6\\3iIbLX6^3hI`LY6`3gI`LZ6`3fI^L\\6b3dI]L\\6c3eI\\L\\6d3cI[L^6e3cIZL^6f3bIXL_6h3bIWL_6i3aIVL_6j3aIULa6k3_ITLb6l3^IRLc6j2`HhMR1YO_6m2aHjMU1SOZ6R3dHiMT1ROZ6S3eHjMg8U2\\GiMe8U2^GjMa8V2`GjM`8T2cGjM]8U2fGjMZ8T2iGjMX8T2kGkMT8T2oGjMR8U2oGkMP8T2SHkMm7S2VHkMk7S2XHlMh7R2YHnMi7o1YHQNg7n1ZHQNKCa6Z2eITNBIj6Q2fIUNZO1Q7g1gIBZ6<gIDZ6:hIEY69iIFY68gIHZ66hIIY65iIJX64iILY61iINX60jIOW6OjI1W6MkI2W6JlI4V6IlI6V6HlI6V6GmI7W6EjI:[6@hI>\\6^OeIa0_6ZOcIe0a6WO`Ih0d6SO^Il0f6PO[Io0c9O2N1O1O1O1N2O1O2N2M2N3N2M2O2M3M2O2M2Oga5"}, "label": "a tennis player bending down in right picture"}]}
{"id": 128282, "image": "COCO_train2014_000000128282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy wearing a white t - shirt and red shorts is playing tennis\"."}], "task": "refering", "answer_template": "The \"a guy wearing a white t - shirt and red shorts is playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 271, 272, 273, 274, 294, 295, 296, 297, 298, 317, 318, 319, 320], "bbox": [0.7021875, 0.38651972157772624, 0.977625, 0.8988863109048725], "iscrowd": 0, "area": 18692.69869999999, "rle": {"size": [431, 640], "counts": "dUm53g<e0O2O1N2N2N1O2O1N2N1O2O1N2O1O00100O1O100O2N1O101N1O101N1O2O0O1O2N100O2O0O1O100O100UEnMW:S2gEPNX:P2fERNY:n1fEUN=Io8S2bFVN=Jo8Q2cFWN:KQ9P2cFWN9MR9m1cFYN7NT9j1cFZN7NU9i1bF\\N5OV9f1dF]N30X9^2eFfMY9Z3M2O2N1O2M3N1O2iH^K\\5d4cJ^K[5d4cJ^K[5d4cJ]K\\5d4bJ_K[5d4cJ\\K^5d4aJ\\K_5f4^J[Kb5g4\\JYKd5h4[JXKe5j4XJWKi5i4VJXKi5j4UJVKk5l4RJUKn5l4QJTKo5n4nISKR6n4mIRKT6o4iIRKW6P5fIQKZ6P5dIRK[6P5bIQK^6Q5_IPKa6g3VImL7]Od6f3VImL3^Og6e3XIlLO@i6e3ZIcMf6]2\\IaMd6`2]I_Mc6a2^I]Mb6c2`I[Ma6e2`IYM`6g2bIWM_6i2bIUM_6j2cITM]6m2dIQM]6n2fIoLZ6R3gIkL[6U3fIiLZ6W3hIfLZ6Z3gIdLY6\\3iIbLX6^3hI`LY6`3gI`LZ6`3fI^L\\6b3dI]L\\6c3eI\\L\\6d3cI[L^6e3cIZL^6f3bIXL_6h3bIWL_6i3aIVL_6j3aIULa6k3_ITLb6l3^IRLc6j2`HhMR1YO_6m2aHjMU1SOZ6R3dHiMT1ROZ6S3eHjMg8U2\\GiMe8U2^GjMa8V2`GjM`8T2cGjM]8U2fGjMZ8T2iGjMX8T2kGkMT8T2oGjMR8U2oGkMP8T2SHkMm7S2VHkMk7S2XHlMh7R2YHnMi7o1YHQNg7n1ZHQNKCa6Z2eITNBIj6Q2fIUNZO1Q7g1gIBZ6<gIDZ6:hIEY69iIFY68gIHZ66hIIY65iIJX64iILY61iINX60jIOW6OjI1W6MkI2W6JlI4V6IlI6V6HlI6V6GmI7W6EjI:[6@hI>\\6^OeIa0_6ZOcIe0a6WO`Ih0d6SO^Il0f6PO[Io0c9O2N1O1O1O1N2O1O2N2M2N3N2M2O2M3M2O2M2Oga5"}, "label": "a guy wearing a white t - shirt and red shorts is playing tennis"}]}
{"id": 128282, "image": "COCO_train2014_000000128282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in air on the far right picture\"."}], "task": "refering", "answer_template": "The \"the man in air on the far right picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 173, 174, 175, 197, 198, 220, 221, 243, 244, 245, 266, 267, 289], "bbox": [0.522109375, 0.13392111368909512, 0.70753125, 0.81122969837587], "iscrowd": 0, "area": 14537.134250000005, "rle": {"size": [431, 640], "counts": "ef\\4b1i;d0\\O5K6J5K5K6K4N2N3M2N2N3M2N2N3N1N3^OUL[Gm3^8]L]Ge3\\8dL`G_3\\8l0N2N3M2N2ZJlJo2V5oLlJn2V5PMnJm2S5RMPKj2S5TMoJj2R5TMRKh2P5WMSKf2n4XMVKe2k4YMXKe2i4XM\\Ke2e4XM`Ke2a4XMcKf2^4WMfKh2Z4UMhKk2Z4QMiKn2Z4lLiKT3\\4eLfK\\3^4]LeKb3`4WLbKi3c4PL_KQ4e4hK^KW4g4bK[K^4m5]LQHb3m7R1N1N2N2O2M8H5L3gKkIU1V6hNlIX1U6fNlIZ1T6QMiI:5e2S6PMgI;7e2S6ZLhI1Oo07f2S6VLjI5Lo07f2U6mKnI>Gn08f2a6]L\\Ig04l2`6\\LbIb0OR3^6]LhI;KX3\\6^LnI4HW2ESNe6ATJOCBAa2>WNY6HZJH_OFB_2f0UNn5N`JCAT27ZNg50`Kb1mN^Nc5ObK]1oNdN_50cKW1ROhN[51dKR1UOmNX51dKl0YOROS52eKh0[OUOQ54eKa0_OZOm44eK3LI`43fK1LJ`45gKLJO`44iKII2^45lKEG6^45mK@H9]46aNJ`15`NKa15_NIc16]NJc17]NHd17\\NIe17[NGg18YNHh18XNGh19XNFj1:VNEk1:UNFl1;SNCo1?_53N2N1N10000O0100O0010000O010O1O00;E1N102JTb^2"}, "label": "the man in air on the far right picture"}]}
{"id": 128282, "image": "COCO_train2014_000000128282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left picture man on left jumping\"."}], "task": "refering", "answer_template": "The \"left picture man on left jumping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 162, 163, 164, 185, 186, 187, 209, 210, 231, 232, 233, 254, 255, 256, 278], "bbox": [0.0309375, 0.13436194895591647, 0.22214062500000004, 0.8259628770301625], "iscrowd": 0, "area": 16047.336299999995, "rle": {"size": [431, 640], "counts": "Pb8g0d0^Ol:S1mDTOm:S1kDTOo:i1L5J6J6K5K4M3M3M3L4N2N2N101N2N2N2VIRL`34TJm3Z2SL]37RJj3^2SLZ3a4cLaKZ3c4dL^KY3f4eL[KX3i4fLYKV3j4hLXKV3i4jLXKU3i4jLXKT3i4lLYKQ3i4mLYKR3g4nL[Ko2g4PM[Kn2f4PM]Km2e4QM^Km2b4SM`Kj2b4TMaKj2e4mL`KP3i4fL[KX3l4^LYKa3l4WLXKh3i4SL\\Kl3f4oK^KP4c4lKaKR4a4iKdKV4]4fKeK[4e63M3M3QO\\KoHf4Q7eKcH]4\\7nKWHT4i7j0O001O1N3N1O2N7I2N1O1O2O=_I[IP11d2d6]L^Ij02f2`6`LbIe02i2[6cLfI>3m2W6eLjI91R3T6fLnI22V3P6hLRJM0[3n5hLVJF0`3j5jLZJANd3i5kL\\J\\ONh3e5lLbKS3_4mLbKV1jNlNd5McKn0ROTO[5OdKd0YO\\OT5OdK<AEj4OgK8BHh40fK6DJe40iK3DLd41hK1FMb42jKMG0`43iKKI1^44kKHI2^46iKFK2]48jKBL4\\4:hK@N4\\4;hK^ON5\\4=fK\\O05[4?_N_Oc1a0]N]Oe1b0\\N]Oe1b0e500O0100000001O0O10001O00000O101O0000000010O001J5I7IU_a6"}, "label": "left picture man on left jumping"}]}
{"id": 128282, "image": "COCO_train2014_000000128282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the left side of the picture\"."}], "task": "refering", "answer_template": "The \"the man on the left side of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 162, 163, 164, 185, 186, 187, 209, 210, 231, 232, 233, 254, 255, 256, 278], "bbox": [0.0309375, 0.13436194895591647, 0.22214062500000004, 0.8259628770301625], "iscrowd": 0, "area": 16047.336299999995, "rle": {"size": [431, 640], "counts": "Pb8g0d0^Ol:S1mDTOm:S1kDTOo:i1L5J6J6K5K4M3M3M3L4N2N2N101N2N2N2VIRL`34TJm3Z2SL]37RJj3^2SLZ3a4cLaKZ3c4dL^KY3f4eL[KX3i4fLYKV3j4hLXKV3i4jLXKU3i4jLXKT3i4lLYKQ3i4mLYKR3g4nL[Ko2g4PM[Kn2f4PM]Km2e4QM^Km2b4SM`Kj2b4TMaKj2e4mL`KP3i4fL[KX3l4^LYKa3l4WLXKh3i4SL\\Kl3f4oK^KP4c4lKaKR4a4iKdKV4]4fKeK[4e63M3M3QO\\KoHf4Q7eKcH]4\\7nKWHT4i7j0O001O1N3N1O2N7I2N1O1O2O=_I[IP11d2d6]L^Ij02f2`6`LbIe02i2[6cLfI>3m2W6eLjI91R3T6fLnI22V3P6hLRJM0[3n5hLVJF0`3j5jLZJANd3i5kL\\J\\ONh3e5lLbKS3_4mLbKV1jNlNd5McKn0ROTO[5OdKd0YO\\OT5OdK<AEj4OgK8BHh40fK6DJe40iK3DLd41hK1FMb42jKMG0`43iKKI1^44kKHI2^46iKFK2]48jKBL4\\4:hK@N4\\4;hK^ON5\\4=fK\\O05[4?_N_Oc1a0]N]Oe1b0\\N]Oe1b0e500O0100000001O0O10001O00000O101O0000000010O001J5I7IU_a6"}, "label": "the man on the left side of the picture"}]}
{"id": 570656, "image": "COCO_train2014_000000570656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the beard and the suit who is looking away\"."}], "task": "refering", "answer_template": "The \"the man with the beard and the suit who is looking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 234, 235, 236, 237, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390], "bbox": [0.68225, 0.19740625, 1.0, 0.9818125], "iscrowd": 0, "area": 61064.13735, "rle": {"size": [640, 480], "counts": "co\\61ac0>C=B>B>C=B>]DbMbL0S6l2[MaMhK2^6j2iMcMnJ2P7b2QNXOk1k0TNXOf1l0ZNVOa1m0^NVO\\1n0cNUOX1n0hNTOS1o0lNTOn0P1QOROk0Q1TOROf0R1ZOPOa0S1^OPO=S1BPO8T1GoN4T1LnNNV11mNJV15lNFX1:jNAY1>jN]OY1b0jNXOZ1g0iNTOZ1l0hNnN\\1Q1dNmN_1R1bNkNa1T1`NhNd1X1\\NeNg1Z1ZNbNj1]1WN`Nl1`1SN]NQ2b1PN[NS2d1nMYNU2f1lMVNX2j1hMSN[2l1fMPN^2o1cMnM`2Q2`MmMc2S2]MiMg2V2ZMgMi2X2XMdMl2\\2TMaMo2^2RM_MQ3`2oL]MU3b2lL[MW3e2nLRMV3m2UMfLn2Y3]MYLg2f3dMmK_2S4kM`KX2_4e6O100O1O1O100O1O1O100O1O100O1O1O11O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001QEnJn6R5mHTKR7m4iHXKV7h4eH_KY7b4bHcK]7]4^HiKa7W4[HoKc7R4WHTLh7l3SHZLl7g3oG_Lo7d3iGbLV8i3[G\\Ld8o3lFXLR9S4_FRL`9Y4PFmKo9T7N1O1O2N1O\\1dNf2ZMf2ZMg_O"}, "label": "the man with the beard and the suit who is looking away"}]}
{"id": 570656, "image": "COCO_train2014_000000570656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit looking away from the camera being pulled by his tie\"."}], "task": "refering", "answer_template": "The \"a man in a suit looking away from the camera being pulled by his tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 234, 235, 236, 237, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390], "bbox": [0.68225, 0.19740625, 1.0, 0.9818125], "iscrowd": 0, "area": 61064.13735, "rle": {"size": [640, 480], "counts": "co\\61ac0>C=B>B>C=B>]DbMbL0S6l2[MaMhK2^6j2iMcMnJ2P7b2QNXOk1k0TNXOf1l0ZNVOa1m0^NVO\\1n0cNUOX1n0hNTOS1o0lNTOn0P1QOROk0Q1TOROf0R1ZOPOa0S1^OPO=S1BPO8T1GoN4T1LnNNV11mNJV15lNFX1:jNAY1>jN]OY1b0jNXOZ1g0iNTOZ1l0hNnN\\1Q1dNmN_1R1bNkNa1T1`NhNd1X1\\NeNg1Z1ZNbNj1]1WN`Nl1`1SN]NQ2b1PN[NS2d1nMYNU2f1lMVNX2j1hMSN[2l1fMPN^2o1cMnM`2Q2`MmMc2S2]MiMg2V2ZMgMi2X2XMdMl2\\2TMaMo2^2RM_MQ3`2oL]MU3b2lL[MW3e2nLRMV3m2UMfLn2Y3]MYLg2f3dMmK_2S4kM`KX2_4e6O100O1O1O100O1O1O100O1O100O1O1O11O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001QEnJn6R5mHTKR7m4iHXKV7h4eH_KY7b4bHcK]7]4^HiKa7W4[HoKc7R4WHTLh7l3SHZLl7g3oG_Lo7d3iGbLV8i3[G\\Ld8o3lFXLR9S4_FRL`9Y4PFmKo9T7N1O1O2N1O\\1dNf2ZMf2ZMg_O"}, "label": "a man in a suit looking away from the camera being pulled by his tie"}]}
{"id": 570656, "image": "COCO_train2014_000000570656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with brown hair and white top carrying an unidentified item\"."}], "task": "refering", "answer_template": "The \"a lady with brown hair and white top carrying an unidentified item\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 273, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 344, 345, 346, 347, 348, 349, 350, 359, 360, 361, 362, 363, 364, 365, 366, 367, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.019895833333333335, 0.24721875, 0.6504791666666667, 1.0], "iscrowd": 0, "area": 84809.6896, "rle": {"size": [640, 480], "counts": "_^6;bc0:F:l]OWO``0R1V_OTOe`0U1Q_OQOj`0W1m^OoNn`0V2J5M3N2M3M3M3M3M3N2eCZLW8i3fG[LV8g3hG]LT8f3iG^LS8d3kGaL\\M]Od9S4nHhLSMYOn9R4lHlLlLWOV:o3lH`MS7c2jH^MU7d2iH^MU7d2iH]MW7e2fH]MX7e2fH\\MY7g2dHZM[7h2dHYM[7i2bHXM]7j2aHWM_7j2iFgK1a1U9j2\\FUL:R1Z9j2UF_L;i0_9j2RFdL9c0d9k2QFgL6?i9k2oElL2;n9`701O0O2O00001O0O3N4L4L4L1N2O1O1O1O0O2O1aJYEc1h:WNgE_1Z:[NkEe1V:VNnEi1T:QNQFo1o9lMVFT2l9eMYF[2h9`M\\F`2e9[M_Fd2d9VM`Fj2a9QMcFo2^9lLfFT3\\9gLgFY3Z9bLjF]3X9^LlFb3V9XLnFh3S9SLQGm3P9oKSGQ4o8iKUGV4m8eKWG[4j8`KZG`4h8ZK\\Gg4d8UK_Gk4k;00001O0000001O00001O0000001O0000001O00002N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O2N2N3M2N2M3N2N2N2N3M2N2N2N2N2N2N2N3M2M3N2N3M3M4L4L3M4L3M4L3M4L4gE`Gj8c8PGaGo8c8jFaGT9c8fF`GZ9d8_F`G`9d8YF`Gg9b8SFbGl9Z9O1O001O1O1O1O1O1O001O1O1O1O1O1O00100O1O1O1O1O001O1O1O1O1O1O001O1O1O1O100O001O1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N2N1O2N3M4L3M3M3M3M4L3M3M3M4L3M3M3M5K6I6K5K6J5K4L3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M5L4L5K5K5K5K4L5K5K5K5K4L5K5KX\\X3"}, "label": "a lady with brown hair and white top carrying an unidentified item"}]}
{"id": 570656, "image": "COCO_train2014_000000570656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling woman in a white shirt holding onto a man ' s tie\"."}], "task": "refering", "answer_template": "The \"a smiling woman in a white shirt holding onto a man ' s tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 273, 274, 275, 276, 277, 278, 279, 280, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 344, 345, 346, 347, 348, 349, 350, 359, 360, 361, 362, 363, 364, 365, 366, 367, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.019895833333333335, 0.24721875, 0.6504791666666667, 1.0], "iscrowd": 0, "area": 84809.6896, "rle": {"size": [640, 480], "counts": "_^6;bc0:F:l]OWO``0R1V_OTOe`0U1Q_OQOj`0W1m^OoNn`0V2J5M3N2M3M3M3M3M3N2eCZLW8i3fG[LV8g3hG]LT8f3iG^LS8d3kGaL\\M]Od9S4nHhLSMYOn9R4lHlLlLWOV:o3lH`MS7c2jH^MU7d2iH^MU7d2iH]MW7e2fH]MX7e2fH\\MY7g2dHZM[7h2dHYM[7i2bHXM]7j2aHWM_7j2iFgK1a1U9j2\\FUL:R1Z9j2UF_L;i0_9j2RFdL9c0d9k2QFgL6?i9k2oElL2;n9`701O0O2O00001O0O3N4L4L4L1N2O1O1O1O0O2O1aJYEc1h:WNgE_1Z:[NkEe1V:VNnEi1T:QNQFo1o9lMVFT2l9eMYF[2h9`M\\F`2e9[M_Fd2d9VM`Fj2a9QMcFo2^9lLfFT3\\9gLgFY3Z9bLjF]3X9^LlFb3V9XLnFh3S9SLQGm3P9oKSGQ4o8iKUGV4m8eKWG[4j8`KZG`4h8ZK\\Gg4d8UK_Gk4k;00001O0000001O00001O0000001O0000001O00002N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O2N2N3M2N2M3N2N2N2N3M2N2N2N2N2N2N2N3M2M3N2N3M3M4L4L3M4L3M4L3M4L4gE`Gj8c8PGaGo8c8jFaGT9c8fF`GZ9d8_F`G`9d8YF`Gg9b8SFbGl9Z9O1O001O1O1O1O1O1O001O1O1O1O1O1O00100O1O1O1O1O001O1O1O1O1O1O001O1O1O1O100O001O1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N2N1O2N3M4L3M3M3M3M4L3M3M3M4L3M3M3M5K6I6K5K6J5K4L3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M5L4L5K5K5K5K4L5K5K5K5K4L5K5KX\\X3"}, "label": "a smiling woman in a white shirt holding onto a man ' s tie"}]}
{"id": 513324, "image": "COCO_train2014_000000513324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing sunglasses and a tan coat , talking on a cell phone\"."}], "task": "refering", "answer_template": "The \"a man wearing sunglasses and a tan coat , talking on a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.32359375, 0.1075, 0.97753125, 0.9816458333333333], "iscrowd": 0, "area": 103711.84514999998, "rle": {"size": [480, 640], "counts": "m^Q34h>;E<D;F3L4L4L4L4N2O1N2O1O1N2O1O1O1N2O1O1N2O1O1iNfMfE\\2X:iMdEX2Z:lMcEU2[:PN`ER2^:RN_Eo1_:VN]Ek1a:YN\\Eh1b:\\NZEf1d:_NXEb1f:bNWE_1g:^1N2N2N2N2M3N2M3N2M3N2iIcKh1`4VNiKa1Y4^NPLW1S4bN[LS1i3eNgL_MYN^2T5KRMUMXNi2j4ORNJR26PNDT2;PN^OT2a0oMYOU2g0nMROV2m0nMlNV2S1mMhNV2X1mMcNU2\\1nMaNS2^1PN^NR2b1PN[NQ2d1RNXNP2g1SNVNn1j1TNRNn1m1UNPNl1o1WNmMk1S2WNjMj1U2YNgMi1X2ZNeMg1[2[NbMf1]2_5O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O1000000000000000000000000000000000000000000001O00000000000000000O100000000000000000000000001O00000000000000001O0000001O00001O0000001O00001O0000001O00001O001O00001O00001O00001O00001O00001O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N3M3M2N3M3M2N3M3M2N3M3M3M2N4L6J5K6J6J5K6J5K5K5K4L2N1O2N2N2N2N2N1O2N2N2N2N3M5K4L5eI^Hn4f7fJcHY5b7YJhHf5Z8O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1Oo0QO^S6"}, "label": "a man wearing sunglasses and a tan coat , talking on a cell phone"}]}
{"id": 513324, "image": "COCO_train2014_000000513324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in a brown coat who is talking on his cell phone\"."}], "task": "refering", "answer_template": "The \"the man in a brown coat who is talking on his cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.32359375, 0.1075, 0.97753125, 0.9816458333333333], "iscrowd": 0, "area": 103711.84514999998, "rle": {"size": [480, 640], "counts": "m^Q34h>;E<D;F3L4L4L4L4N2O1N2O1O1N2O1O1O1N2O1O1N2O1O1iNfMfE\\2X:iMdEX2Z:lMcEU2[:PN`ER2^:RN_Eo1_:VN]Ek1a:YN\\Eh1b:\\NZEf1d:_NXEb1f:bNWE_1g:^1N2N2N2N2M3N2M3N2M3N2iIcKh1`4VNiKa1Y4^NPLW1S4bN[LS1i3eNgL_MYN^2T5KRMUMXNi2j4ORNJR26PNDT2;PN^OT2a0oMYOU2g0nMROV2m0nMlNV2S1mMhNV2X1mMcNU2\\1nMaNS2^1PN^NR2b1PN[NQ2d1RNXNP2g1SNVNn1j1TNRNn1m1UNPNl1o1WNmMk1S2WNjMj1U2YNgMi1X2ZNeMg1[2[NbMf1]2_5O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O1000000000000000000000000000000000000000000001O00000000000000000O100000000000000000000000001O00000000000000001O0000001O00001O0000001O00001O0000001O00001O001O00001O00001O00001O00001O00001O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N3M3M2N3M3M2N3M3M2N3M3M3M2N4L6J5K6J6J5K6J5K5K5K4L2N1O2N2N2N2N2N1O2N2N2N2N3M5K4L5eI^Hn4f7fJcHY5b7YJhHf5Z8O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1Oo0QO^S6"}, "label": "the man in a brown coat who is talking on his cell phone"}]}
{"id": 546093, "image": "COCO_train2014_000000546093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue suitcase on roller\"."}], "task": "refering", "answer_template": "The \"the blue suitcase on roller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 158, 159, 160, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 225], "bbox": [0.11957333333333334, 0.65696, 0.40658666666666665, 0.9529200000000001], "iscrowd": 0, "area": 11458.456799999998, "rle": {"size": [500, 375], "counts": "_\\f0<V?5J6K5bAUOk=P1RBTOh=Q1UBPOh=U1UBkNi=Z1TBfNj=c1[Oe0@`0N3M2N2N2N2N3MN2N4N2O1O1O100000000001O0000000O1O1O1O1O1O1O1O1O1G9N2M301N10000O10000O10000O10000O10000O1010O00000000000000000000001O00001O00010O09H2M100a0^O9H00O01O1O001O1O0O2N1O2N2G9L5L4L3M4K5L4La`\\3"}, "label": "the blue suitcase on roller"}]}
{"id": 546093, "image": "COCO_train2014_000000546093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a medium blue carryon bag\"."}], "task": "refering", "answer_template": "The \"a medium blue carryon bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 158, 159, 160, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 225], "bbox": [0.11957333333333334, 0.65696, 0.40658666666666665, 0.9529200000000001], "iscrowd": 0, "area": 11458.456799999998, "rle": {"size": [500, 375], "counts": "_\\f0<V?5J6K5bAUOk=P1RBTOh=Q1UBPOh=U1UBkNi=Z1TBfNj=c1[Oe0@`0N3M2N2N2N2N3MN2N4N2O1O1O100000000001O0000000O1O1O1O1O1O1O1O1O1G9N2M301N10000O10000O10000O10000O10000O1010O00000000000000000000001O00001O00010O09H2M100a0^O9H00O01O1O001O1O0O2N1O2N2G9L5L4L3M4K5L4La`\\3"}, "label": "a medium blue carryon bag"}]}
{"id": 546093, "image": "COCO_train2014_000000546093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black rolling suitcase pulled by a woman in a pink jacket\"."}], "task": "refering", "answer_template": "The \"a black rolling suitcase pulled by a woman in a pink jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 125, 126, 138, 139, 150, 151, 152, 153, 163, 164, 165, 166, 176, 177, 178, 179, 189, 190, 191, 192, 202, 203, 204, 205], "bbox": [0.5697333333333333, 0.47002, 0.8351999999999999, 0.9040199999999999], "iscrowd": 0, "area": 14144.940550000001, "rle": {"size": [500, 375], "counts": "UjX3;W?=C<D=C<D=C<D=C<D=C6K2O1cMlKYIU4e6mKYIT4f6oKWIR4h6PLVIQ4i6QLUIQ4h6RLTIn3n6UL^HDPOW4d8WLZHGjNR4V9RLnGY4S8iKkGW4U8lKhGT4X8nKfGR4Z8PLbGR4]8QL`GP4`8RL]Go3c8TLVGP4j8RLPGR4P9PLjFT4V9U100000001O00000000000O1N2000000000000009G=C>B=C>B1O1O00000000001O00000000001O000000001O0001O1O1O100O01H701M3N1N3M7I;F:E<D;F;D;E;E<E:E<DPlm0"}, "label": "a black rolling suitcase pulled by a woman in a pink jacket"}]}
{"id": 546093, "image": "COCO_train2014_000000546093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black color trolley bag standing on its wheel , holded by a lady\"."}], "task": "refering", "answer_template": "The \"black color trolley bag standing on its wheel , holded by a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 125, 126, 138, 139, 150, 151, 152, 153, 163, 164, 165, 166, 176, 177, 178, 179, 189, 190, 191, 192, 202, 203, 204, 205], "bbox": [0.5697333333333333, 0.47002, 0.8351999999999999, 0.9040199999999999], "iscrowd": 0, "area": 14144.940550000001, "rle": {"size": [500, 375], "counts": "UjX3;W?=C<D=C<D=C<D=C<D=C6K2O1cMlKYIU4e6mKYIT4f6oKWIR4h6PLVIQ4i6QLUIQ4h6RLTIn3n6UL^HDPOW4d8WLZHGjNR4V9RLnGY4S8iKkGW4U8lKhGT4X8nKfGR4Z8PLbGR4]8QL`GP4`8RL]Go3c8TLVGP4j8RLPGR4P9PLjFT4V9U100000001O00000000000O1N2000000000000009G=C>B=C>B1O1O00000000001O00000000001O000000001O0001O1O1O100O01H701M3N1N3M7I;F:E<D;F;D;E;E<E:E<DPlm0"}, "label": "black color trolley bag standing on its wheel , holded by a lady"}]}
{"id": 259375, "image": "COCO_train2014_000000259375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown surfboard laid up on the beach\"."}], "task": "refering", "answer_template": "The \"a light brown surfboard laid up on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.057499999999999996, 0.5296519721577726, 0.5387500000000001, 0.8015777262180975], "iscrowd": 0, "area": 23568.32195, "rle": {"size": [431, 640], "counts": "Ql?:g<>B>K6N1O2O0O1O2N1O2N100O2N1O1O2O0O2N1O1O2O0O2N1O1O2O0O2N1O1O2N100O2N1O2N100O2O001O0O101O001N10001O0O101O001N10001O0O2O00001N101O000O1000000O1000000O0100000O1000000O10O1000O1000000O100000O01000000O10000000O10000000000000000O1000000000O10000000O100000000000000O100000000000000O100000LhLPFY3P:gLPFY3P:gLPFY3P:hLoEX3Q:hLoEW3R:iLnEW3R:300000000000O10000000000000000O100000000000000O100000000000000O4M4L5K5K5K9G;E;D=D;E<D>BkY64gcIh0RFCl9P2O2N2O1N1O2N2O1N1O2N22N0O1000000O1000000O1000O10O1000001N4M4L3M4K4M3M4L3L4M3M3L4M3M3M2M4M3M3M2M3N2N2N2M3NXNC`F;_9I`F5^91`FL`98_FF_9?`F_O_9f0_FXO_9m0_15K4N3L3N2M4M2M4M2M4M2M3NO0O2M2O2N1O1O2N1O2N1O2M2O2N1O2N1O2N5K4L4K5L5K4L4L4L5Kelk3"}, "label": "a light brown surfboard laid up on the beach"}]}
{"id": 259375, "image": "COCO_train2014_000000259375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfboard on the left\"."}], "task": "refering", "answer_template": "The \"the surfboard on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.057499999999999996, 0.5296519721577726, 0.5387500000000001, 0.8015777262180975], "iscrowd": 0, "area": 23568.32195, "rle": {"size": [431, 640], "counts": "Ql?:g<>B>K6N1O2O0O1O2N1O2N100O2N1O1O2O0O2N1O1O2O0O2N1O1O2O0O2N1O1O2N100O2N1O2N100O2O001O0O101O001N10001O0O101O001N10001O0O2O00001N101O000O1000000O1000000O0100000O1000000O10O1000O1000000O100000O01000000O10000000O10000000000000000O1000000000O10000000O100000000000000O100000000000000O100000LhLPFY3P:gLPFY3P:gLPFY3P:hLoEX3Q:hLoEW3R:iLnEW3R:300000000000O10000000000000000O100000000000000O100000000000000O4M4L5K5K5K9G;E;D=D;E<D>BkY64gcIh0RFCl9P2O2N2O1N1O2N2O1N1O2N22N0O1000000O1000000O1000O10O1000001N4M4L3M4K4M3M4L3L4M3M3L4M3M3M2M4M3M3M2M3N2N2N2M3NXNC`F;_9I`F5^91`FL`98_FF_9?`F_O_9f0_FXO_9m0_15K4N3L3N2M4M2M4M2M4M2M3NO0O2M2O2N1O1O2N1O2N1O2M2O2N1O2N1O2N5K4L4K5L5K4L4L4L5Kelk3"}, "label": "the surfboard on the left"}]}
{"id": 259375, "image": "COCO_train2014_000000259375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black swim suit\"."}], "task": "refering", "answer_template": "The \"a man in a black swim suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 121, 122, 123, 124, 125, 143, 144, 146, 147, 148, 165, 166, 167, 169, 170, 171, 188, 192, 194, 215, 217, 218, 238, 240, 241, 260, 261, 263, 264], "bbox": [0.193671875, 0.2248723897911833, 0.522296875, 0.8060556844547564], "iscrowd": 0, "area": 17195.89270000001, "rle": {"size": [431, 640], "counts": "U]d11^=3M2M4M00000O10000000000O10000000000O10O1O1O10N1O1N2O1N2O2ZMAYHa0b7F[H;c7H[H9d7IYH9e7KXH7f7KWH7i7JUH7l7JPH8Q8HlG:T8GjG:W8GeG:]8G`G9a8I[G9f8HWG8k8HRG:n8HnF9T9HiF9X9HdF:]9G`F9a9I[F9f9HVF9l9HQF9o9ImE9T:IhE7Z:^1001N2O0O2O1N3N2M3N1O2M3N1N3N2M2O2M3N2M2O3A[DnNi;Q1YDkNj;S1YDhNk;V170O101N100O101N100O100]IcNS1_1lNaNT1_1lNbNS1^1mNbNR1_1nNbNQ1^1oNbNP1`1oNaNP1_1POaNo0`1QOaNn0_1XKhNl1Jj2`1VKMl0cNn3`1SKU1J\\MQ5a1PKf1_OjLa5a1lJn1\\OcLf5a1jJV2WO[Lo5_1gJk2W5WMeJl2[5UMaJo2]5SM^JQ3b5PMZJT3d5mLYJV3g5kLUJY3j5V2O2O1O0100O010O100O1C<WMoIfM`6W2fIXMh6e2_IeLS7W3TIPLa7n3R1L4M3M2M4M3M3L3N3M3L4M2N3L4N2N102N2N1O2M2O2N14L6K5J5K6J6K5J6J5L1N2N2N3O0O1O10000O10001OO1O1O1O1O3L9H9G8J6J6J6K3K3K5hMTEc1_;N2N2M3L4L4M2M3M3M3M3M2N00100O1N2N2N1O2N2N2O1O1N2Om^P4"}, "label": "a man in a black swim suit"}]}
{"id": 259375, "image": "COCO_train2014_000000259375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a wetsuit\"."}], "task": "refering", "answer_template": "The \"a man in a wetsuit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 121, 122, 123, 124, 125, 143, 144, 146, 147, 148, 165, 166, 167, 169, 170, 171, 188, 192, 194, 215, 217, 218, 238, 240, 241, 260, 261, 263, 264], "bbox": [0.193671875, 0.2248723897911833, 0.522296875, 0.8060556844547564], "iscrowd": 0, "area": 17195.89270000001, "rle": {"size": [431, 640], "counts": "U]d11^=3M2M4M00000O10000000000O10000000000O10O1O1O10N1O1N2O1N2O2ZMAYHa0b7F[H;c7H[H9d7IYH9e7KXH7f7KWH7i7JUH7l7JPH8Q8HlG:T8GjG:W8GeG:]8G`G9a8I[G9f8HWG8k8HRG:n8HnF9T9HiF9X9HdF:]9G`F9a9I[F9f9HVF9l9HQF9o9ImE9T:IhE7Z:^1001N2O0O2O1N3N2M3N1O2M3N1N3N2M2O2M3N2M2O3A[DnNi;Q1YDkNj;S1YDhNk;V170O101N100O101N100O100]IcNS1_1lNaNT1_1lNbNS1^1mNbNR1_1nNbNQ1^1oNbNP1`1oNaNP1_1POaNo0`1QOaNn0_1XKhNl1Jj2`1VKMl0cNn3`1SKU1J\\MQ5a1PKf1_OjLa5a1lJn1\\OcLf5a1jJV2WO[Lo5_1gJk2W5WMeJl2[5UMaJo2]5SM^JQ3b5PMZJT3d5mLYJV3g5kLUJY3j5V2O2O1O0100O010O100O1C<WMoIfM`6W2fIXMh6e2_IeLS7W3TIPLa7n3R1L4M3M2M4M3M3L3N3M3L4M2N3L4N2N102N2N1O2M2O2N14L6K5J5K6J6K5J6J5L1N2N2N3O0O1O10000O10001OO1O1O1O1O3L9H9G8J6J6J6K3K3K5hMTEc1_;N2N2M3L4L4M2M3M3M3M3M2N00100O1N2N2N1O2N2N2O1O1N2Om^P4"}, "label": "a man in a wetsuit"}]}
{"id": 259375, "image": "COCO_train2014_000000259375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in swim trunks by the ocean preparing surfboard to surf\"."}], "task": "refering", "answer_template": "The \"man in swim trunks by the ocean preparing surfboard to surf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 197, 198, 199, 200, 219, 220, 221, 242, 243, 244, 266, 267, 288, 289, 290, 291], "bbox": [0.5230625, 0.2293967517401392, 0.744, 0.8611832946635731], "iscrowd": 0, "area": 20179.713799999994, "rle": {"size": [431, 640], "counts": "PU]4<P=?A?B=D8G3N1N2O1N2O1O1N2O1O1N4M3M3L4L4L4L4L5L3^GnLg0Dc3b3aKQMa0Cm3_3^KUM:BV4^3[KWM4A`4[3XK[MO_Og4Z3UK`MLWOn4\\3QKeMLoNQ5c3lJfMLhNW5i3eJgMO`NZ5W700O101N10000O2O0O100O2O000O100O100O1000000O10000000000O10000000gNQKRJo4j5WKTJh4j5]KTJc4h5bKVJ_4l0QKg2i0\\LV4m0WK\\2i0fLP4n0]KS2g0nLl3o0^Ko1j0QMh3P1cKf1j0YMc3Q1iK[1j0cM]3R1oKQ1i0lMX3S1UL5Z1gNa2T1[LLX1oN]2U1dLBQ1TO\\M_Oo4j1RMUOe0A]2Z1oLROh0BY2\\1PMPOi0CW2]1QMnNj0DU2^1RMlNk0ES2_1RMkNm0ER2^1SMlNl0EQ2_1SMkNn0Eo1_1UMjNm0Gn1_1VMiNm0Gm1_1YMgNl0Ik1c1WMbNP1Ij1i1SM]NS1Jk1k1QMXNV1Kj1o1PMTNV1Lk1Q2e0mM]OT2n51O010O010O0010N101N2O10O1_EWN[9Q2WFVNh9i1SF^N<[Og8W2jFcN<YOi8S2gFjN=UOl8V3PGnLo8j30O0016JO1O1O101N1O1O1QMnF_1S9`NUG9C^OX97^G0AER99bGKAIP99bGCH3g88mHHT76mHIU75mHJT73nHLT72mHNU7MoH2T7GPI8U:O001L3M4K`iT2"}, "label": "man in swim trunks by the ocean preparing surfboard to surf"}]}
{"id": 259375, "image": "COCO_train2014_000000259375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing shoes and shorts holding the back of a surf board\"."}], "task": "refering", "answer_template": "The \"a man wearing shoes and shorts holding the back of a surf board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 197, 198, 199, 200, 219, 220, 221, 242, 243, 244, 266, 267, 288, 289, 290, 291], "bbox": [0.5230625, 0.2293967517401392, 0.744, 0.8611832946635731], "iscrowd": 0, "area": 20179.713799999994, "rle": {"size": [431, 640], "counts": "PU]4<P=?A?B=D8G3N1N2O1N2O1O1N2O1O1N4M3M3L4L4L4L4L5L3^GnLg0Dc3b3aKQMa0Cm3_3^KUM:BV4^3[KWM4A`4[3XK[MO_Og4Z3UK`MLWOn4\\3QKeMLoNQ5c3lJfMLhNW5i3eJgMO`NZ5W700O101N10000O2O0O100O2O000O100O100O1000000O10000000000O10000000gNQKRJo4j5WKTJh4j5]KTJc4h5bKVJ_4l0QKg2i0\\LV4m0WK\\2i0fLP4n0]KS2g0nLl3o0^Ko1j0QMh3P1cKf1j0YMc3Q1iK[1j0cM]3R1oKQ1i0lMX3S1UL5Z1gNa2T1[LLX1oN]2U1dLBQ1TO\\M_Oo4j1RMUOe0A]2Z1oLROh0BY2\\1PMPOi0CW2]1QMnNj0DU2^1RMlNk0ES2_1RMkNm0ER2^1SMlNl0EQ2_1SMkNn0Eo1_1UMjNm0Gn1_1VMiNm0Gm1_1YMgNl0Ik1c1WMbNP1Ij1i1SM]NS1Jk1k1QMXNV1Kj1o1PMTNV1Lk1Q2e0mM]OT2n51O010O010O0010N101N2O10O1_EWN[9Q2WFVNh9i1SF^N<[Og8W2jFcN<YOi8S2gFjN=UOl8V3PGnLo8j30O0016JO1O1O101N1O1O1QMnF_1S9`NUG9C^OX97^G0AER99bGKAIP99bGCH3g88mHHT76mHIU75mHJT73nHLT72mHNU7MoH2T7GPI8U:O001L3M4K`iT2"}, "label": "a man wearing shoes and shorts holding the back of a surf board"}]}
{"id": 46385, "image": "COCO_train2014_000000046385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a purple shirt lighting birthday candles\"."}], "task": "refering", "answer_template": "The \"a woman wearing a purple shirt lighting birthday candles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 56, 57, 58, 59, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 172, 173, 174, 175, 176, 193, 194, 196, 197, 198, 199, 216, 217, 219, 220, 221, 222, 239, 240, 241, 242, 244, 245, 262, 263], "bbox": [0.39270312500000004, 0.06066666666666667, 0.6859531249999999, 0.6988749999999999], "iscrowd": 0, "area": 28971.525800000007, "rle": {"size": [480, 640], "counts": "kke31n>7J6K6I6J7J5J6K6I6eLYNQIm1i6UNVIQ2b6QN]IU2]6mMbIY2V6iMiI]2Q6dMoIa2j5bMTJe2e5\\MZJj2_5YM`Jm2Y5TMfJl2Y5VMgJh2Z5ZMdJf2[5\\MdJc2]5_MbJ`2]5bMbJ]2_5dMaJN[M`1T8eN_JI`Ma1Q8gN^JEfMa1l7lN^J@iMc1i7oN\\J\\OnMc1f7SO\\JWORNd1b7WOZJSOWNd1_7[OZJnNZNe1\\7_OYJiN`Nf1W7CWJeNeNf1T7GWJ`NiNg1P7K\\K3d4O\\K0d41\\KMd46[KHf49ZKFe4=ZKAg4`0XK@g4b0YK\\Oh4f0WKYOh4i0XKUOi4m0UKROk4P1UKoNk4R1UKlNk4W1TKhNo3RO^IW2c2eNk3AYIm1j2bNg3OWI`1R3_Nc3n2]LPM_3W3`LhL[3^3dLaL\\3b3cL]L[3f3eLXL[3j3eLULY3o3eLPL[3R4eLmKY3V4gLhKY3[4fLcKY3`4fL`KY3c4fL[KY3h4gLWKX3k4hLSKY3o4fLPKZ3Q5eLnJ\\3T5cLjJ^3W5bLhJ^3[5`LcJa3_5]LaJc3b5[L\\Jf3f5YLYJg3j5WLTJj3n5SLRJn3Q6nKPJR4S6iKnIX4T6eKmI[4V6aKjI`4X6]KiIc4Z6XKgIi4[6TKeIm4^6oJcIQ5_6kJbIV5a6fJ`IZ5c6bJ]I_5e6]J\\Id5g6XJZIh5h6UJXIl5k6PJUIQ6m6kIUIU6n6gIRI[6o6bIRI^6Q7]IPIe6R7WInHj6d70O2O0O1O2O0O2O0O1O2O0O101N1O2O0O101N1O2O0O101N7Id0]O>A2O0O2N101N2N101N101N1O2O1N101N1O2O1N12O2M3N1O2M3N1N3N2N1N2O1N2O1O1N01N1O2O8Ga0_O`0@a0_Oa0_O`0A`0_O`0@;E:F:F5K3N2M3M3M2N3N2M3N1N3N2N2M3N1O2M\\nm2"}, "label": "a woman wearing a purple shirt lighting birthday candles"}]}
{"id": 46385, "image": "COCO_train2014_000000046385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman\"."}], "task": "refering", "answer_template": "The \"woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 56, 57, 58, 59, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 172, 173, 174, 175, 176, 193, 194, 196, 197, 198, 199, 216, 217, 219, 220, 221, 222, 239, 240, 241, 242, 244, 245, 262, 263], "bbox": [0.39270312500000004, 0.06066666666666667, 0.6859531249999999, 0.6988749999999999], "iscrowd": 0, "area": 28971.525800000007, "rle": {"size": [480, 640], "counts": "kke31n>7J6K6I6J7J5J6K6I6eLYNQIm1i6UNVIQ2b6QN]IU2]6mMbIY2V6iMiI]2Q6dMoIa2j5bMTJe2e5\\MZJj2_5YM`Jm2Y5TMfJl2Y5VMgJh2Z5ZMdJf2[5\\MdJc2]5_MbJ`2]5bMbJ]2_5dMaJN[M`1T8eN_JI`Ma1Q8gN^JEfMa1l7lN^J@iMc1i7oN\\J\\OnMc1f7SO\\JWORNd1b7WOZJSOWNd1_7[OZJnNZNe1\\7_OYJiN`Nf1W7CWJeNeNf1T7GWJ`NiNg1P7K\\K3d4O\\K0d41\\KMd46[KHf49ZKFe4=ZKAg4`0XK@g4b0YK\\Oh4f0WKYOh4i0XKUOi4m0UKROk4P1UKoNk4R1UKlNk4W1TKhNo3RO^IW2c2eNk3AYIm1j2bNg3OWI`1R3_Nc3n2]LPM_3W3`LhL[3^3dLaL\\3b3cL]L[3f3eLXL[3j3eLULY3o3eLPL[3R4eLmKY3V4gLhKY3[4fLcKY3`4fL`KY3c4fL[KY3h4gLWKX3k4hLSKY3o4fLPKZ3Q5eLnJ\\3T5cLjJ^3W5bLhJ^3[5`LcJa3_5]LaJc3b5[L\\Jf3f5YLYJg3j5WLTJj3n5SLRJn3Q6nKPJR4S6iKnIX4T6eKmI[4V6aKjI`4X6]KiIc4Z6XKgIi4[6TKeIm4^6oJcIQ5_6kJbIV5a6fJ`IZ5c6bJ]I_5e6]J\\Id5g6XJZIh5h6UJXIl5k6PJUIQ6m6kIUIU6n6gIRI[6o6bIRI^6Q7]IPIe6R7WInHj6d70O2O0O1O2O0O2O0O1O2O0O101N1O2O0O101N1O2O0O101N7Id0]O>A2O0O2N101N2N101N101N1O2O1N101N1O2O1N12O2M3N1O2M3N1N3N2N1N2O1N2O1O1N01N1O2O8Ga0_O`0@a0_Oa0_O`0A`0_O`0@;E:F:F5K3N2M3M3M2N3N2M3N1N3N2N2M3N1O2M\\nm2"}, "label": "woman"}]}
{"id": 46385, "image": "COCO_train2014_000000046385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a striped dress shirt lighting candles\"."}], "task": "refering", "answer_template": "The \"a man wearing a striped dress shirt lighting candles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 194, 195, 207, 208, 209, 210, 211, 212, 213, 217, 218, 219, 230, 231, 232, 233, 234, 235, 241, 242, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 368, 369], "bbox": [0.0016718750000000002, 0.002458333333333333, 0.543625, 0.9823125], "iscrowd": 0, "area": 97559.00150000001, "rle": {"size": [480, 640], "counts": "P`0j0V>c1\\Nd1\\Nd1]Nc1\\Nd1]Nc1\\Nd1]Nc1\\N6J10O010O010O010O01O010O01O010O010O01O010O010O010O010O0010O100O10O001N101O0O20O100O2O0O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O101N100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O101N100O10O0100O1H7I8G9G9H7H9G9G9H8G8H9H8G9G9G9H8G9G9H8M3O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O010O100O10000O100O100O11O00001O00001O00001O00001O00001O00001O00001O00001O00001OeLgGOY80jGNV82lGMS82PHLP83SHKm74WHJh75[HIe76^HHb77aHH^77eHG[78iHEW7:lHES7:oHEQ7:QIEo6:TIEk6;VIDj6;YICf6=\\ICc6<_ICa6<bIB^6=dIC[6;iICW6<kICU6<mIDR6;QJCo5<SJCm5<VJCi5<YJCeNlMS6`2[KB`NQNT6\\2^KCZNVNV6V2bKCVNZNW6R2fKBQN`NW6m1jKCkMdNZ6h1nKBfMiN]6b1oKDbMnN_6[1QLG]MQOc6U1SLHWMXOf6m0ULJSM\\Oi6g0WLLmL@n6`0WL[1i3bNYL^1h3_N[La1e3\\N]Ld1d3YN_Lf1b3WN`Lj1a3RNbLm1_3PNcLP2^3mMdLR2^3kMeLT2\\3iMfLV2]3fMfLY2[3dMgL[2[3bMgL^2Z3_MiL_2Y3^MiLb2Y3ZMjLd2X3YMjLg2i7N100O2O1N2O1N2O1N2O1N2O1N2O1N2O2M2O1N2O1N2N2L4M3L4M3M3L4M3M3L4M4K4M3M3L4M3M`cX4"}, "label": "a man wearing a striped dress shirt lighting candles"}]}
{"id": 46385, "image": "COCO_train2014_000000046385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in purple lighting a birthday cake with a woman\"."}], "task": "refering", "answer_template": "The \"a man in purple lighting a birthday cake with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 194, 195, 207, 208, 209, 210, 211, 212, 213, 217, 218, 219, 230, 231, 232, 233, 234, 235, 241, 242, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 368, 369], "bbox": [0.0016718750000000002, 0.002458333333333333, 0.543625, 0.9823125], "iscrowd": 0, "area": 97559.00150000001, "rle": {"size": [480, 640], "counts": "P`0j0V>c1\\Nd1\\Nd1]Nc1\\Nd1]Nc1\\Nd1]Nc1\\N6J10O010O010O010O01O010O01O010O010O01O010O010O010O010O0010O100O10O001N101O0O20O100O2O0O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O101N100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O101N100O10O0100O1H7I8G9G9H7H9G9G9H8G8H9H8G9G9G9H8G9G9H8M3O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O010O100O10000O100O100O11O00001O00001O00001O00001O00001O00001O00001O00001O00001OeLgGOY80jGNV82lGMS82PHLP83SHKm74WHJh75[HIe76^HHb77aHH^77eHG[78iHEW7:lHES7:oHEQ7:QIEo6:TIEk6;VIDj6;YICf6=\\ICc6<_ICa6<bIB^6=dIC[6;iICW6<kICU6<mIDR6;QJCo5<SJCm5<VJCi5<YJCeNlMS6`2[KB`NQNT6\\2^KCZNVNV6V2bKCVNZNW6R2fKBQN`NW6m1jKCkMdNZ6h1nKBfMiN]6b1oKDbMnN_6[1QLG]MQOc6U1SLHWMXOf6m0ULJSM\\Oi6g0WLLmL@n6`0WL[1i3bNYL^1h3_N[La1e3\\N]Ld1d3YN_Lf1b3WN`Lj1a3RNbLm1_3PNcLP2^3mMdLR2^3kMeLT2\\3iMfLV2]3fMfLY2[3dMgL[2[3bMgL^2Z3_MiL_2Y3^MiLb2Y3ZMjLd2X3YMjLg2i7N100O2O1N2O1N2O1N2O1N2O1N2O1N2O2M2O1N2O1N2N2L4M3L4M3M3L4M3M3L4M4K4M3M3L4M3M`cX4"}, "label": "a man in purple lighting a birthday cake with a woman"}]}
{"id": 554310, "image": "COCO_train2014_000000554310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in front of two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra in front of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 178], "bbox": [0.69316, 0.5662133333333333, 0.99828, 0.73336], "iscrowd": 0, "area": 5739.533600000002, "rle": {"size": [375, 500], "counts": "eYo38\\;5M3N2eE_O[9d0^FAb9a0XFCh9?RFEm9>lEGT:n00O0100O10000O10000O11N100000000000001O000O100000001O000000000O101O0000000000001O0O10000000001O00000N2O1N2O1N2O1N2O100001O00000000000000000000000O100000001O000000000O1O100O2N2N1O2N1O2O001O1O001N10N2K5K500O00100O1O1O010O1O1O100O010O100O1000O2O001O1O1O0O2O1O001O1O0012N2N2M3N2N2M4M4L4K5L3Kj3"}, "label": "a zebra in front of two other zebras"}]}
{"id": 554310, "image": "COCO_train2014_000000554310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rightmost zebra\"."}], "task": "refering", "answer_template": "The \"the rightmost zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 178], "bbox": [0.69316, 0.5662133333333333, 0.99828, 0.73336], "iscrowd": 0, "area": 5739.533600000002, "rle": {"size": [375, 500], "counts": "eYo38\\;5M3N2eE_O[9d0^FAb9a0XFCh9?RFEm9>lEGT:n00O0100O10000O10000O11N100000000000001O000O100000001O000000000O101O0000000000001O0O10000000001O00000N2O1N2O1N2O1N2O100001O00000000000000000000000O100000001O000000000O1O100O2N2N1O2N1O2O001O1O001N10N2K5K500O00100O1O1O010O1O1O100O010O100O1000O2O001O1O1O0O2O1O001O1O0012N2N2M3N2N2M4M4L4K5L3Kj3"}, "label": "the rightmost zebra"}]}
{"id": 333127, "image": "COCO_train2014_000000333127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teenage boy skateboarding while wearing a green beanie and a stripped shirt\"."}], "task": "refering", "answer_template": "The \"a teenage boy skateboarding while wearing a green beanie and a stripped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 69, 70, 71, 72, 84, 85, 86, 98, 99, 100, 101, 112, 113, 114, 115, 116, 128, 129, 130, 131, 142, 143, 144, 159], "bbox": [0.5063231850117096, 0.153609375, 0.857400468384075, 0.44346874999999997], "iscrowd": 0, "area": 9910.216149999995, "rle": {"size": [640, 427], "counts": "jkW42nc00_N0Q_O2m`0OR_O2l`00T_O1i`02U_OOj`02V_ONj`02V_OOi`01W_OOi`02V_OOi`01V_O0j`00V_O0j`01U_O0j`00V_O0j`0OW_O2h`0NX_O2h`0NX_O3g`0Ml^O`0Ta0_Ol^Ob0Ta0^Ok^Od0Sa0]Ok^Oe0Ua0ZOk^Oh0Sa0YOl^Oh0Ta0XOk^Oi0Ta0XOk^Oj0Sa0VOm^Ok0Sa0UOl^Om0Ra0TOm^Om0Sa0SOl^On0Sa0ROm^Oo0Sa0QOk^Oo0Va0ROh^Om0[a0ROd^On06iN``0:X_Om09jN_`09V_Ol0=lN\\`09U_Ok0`0nNY`07V_Oj0c0POW`06U_Oi0e0ROU`06T_Oh0h0SOT`05R_Og0l0VOP`03S_Og0m0XOP`0V1R@jNm?T1V@mNi?Q1X@POg?o0[@QOe?l0^@UO`?j0b@VO]?i0d@XO\\?e0^@kNQOa0_`0c0`@POoN<b`0b0_@VOmN7d`0d0\\@Ke?6X@Lg?6U@Ll?4R@Mn?5o_OLR`04k_ONU`04h_OMY?kNQAY1ELT?TOUAQ1EKS?[OTAl0GIU?^ORAi0HIV?Ao@h0IGW?Do@e0IGX?Gl@d0JEY?Im@b0HF[?Il@c0GDZ?Mo@?FWOf?=c@Z1^?gNa@X1a?gN^@X1d?hN[@X1f?hNX@Y1i?gNU@Z1l?fNR@[1o?Y12O1O1O1N2O1O1000107H1O100O1O1O10O00001O0000001O01O01O00001O0O1O000O10O1WMc_OP2]`0oMl_Oi1T`0WNS@b1m?^NZ@[1e?fNc@R1]?nNj@j0W?VOPAc0Q?\\OWA<j>C]A6c>JdAN^>1jAGW>8PBAP>?i2O1O0O2O1O001O1O00N2N2O1N2O2M23N2M2O2M2O2M2O2M2O2M_PV1"}, "label": "a teenage boy skateboarding while wearing a green beanie and a stripped shirt"}]}
{"id": 333127, "image": "COCO_train2014_000000333127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green beanie and striped hoodie skating\"."}], "task": "refering", "answer_template": "The \"a man wearing a green beanie and striped hoodie skating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 69, 70, 71, 72, 84, 85, 86, 98, 99, 100, 101, 112, 113, 114, 115, 116, 128, 129, 130, 131, 142, 143, 144, 159], "bbox": [0.5063231850117096, 0.153609375, 0.857400468384075, 0.44346874999999997], "iscrowd": 0, "area": 9910.216149999995, "rle": {"size": [640, 427], "counts": "jkW42nc00_N0Q_O2m`0OR_O2l`00T_O1i`02U_OOj`02V_ONj`02V_OOi`01W_OOi`02V_OOi`01V_O0j`00V_O0j`01U_O0j`00V_O0j`0OW_O2h`0NX_O2h`0NX_O3g`0Ml^O`0Ta0_Ol^Ob0Ta0^Ok^Od0Sa0]Ok^Oe0Ua0ZOk^Oh0Sa0YOl^Oh0Ta0XOk^Oi0Ta0XOk^Oj0Sa0VOm^Ok0Sa0UOl^Om0Ra0TOm^Om0Sa0SOl^On0Sa0ROm^Oo0Sa0QOk^Oo0Va0ROh^Om0[a0ROd^On06iN``0:X_Om09jN_`09V_Ol0=lN\\`09U_Ok0`0nNY`07V_Oj0c0POW`06U_Oi0e0ROU`06T_Oh0h0SOT`05R_Og0l0VOP`03S_Og0m0XOP`0V1R@jNm?T1V@mNi?Q1X@POg?o0[@QOe?l0^@UO`?j0b@VO]?i0d@XO\\?e0^@kNQOa0_`0c0`@POoN<b`0b0_@VOmN7d`0d0\\@Ke?6X@Lg?6U@Ll?4R@Mn?5o_OLR`04k_ONU`04h_OMY?kNQAY1ELT?TOUAQ1EKS?[OTAl0GIU?^ORAi0HIV?Ao@h0IGW?Do@e0IGX?Gl@d0JEY?Im@b0HF[?Il@c0GDZ?Mo@?FWOf?=c@Z1^?gNa@X1a?gN^@X1d?hN[@X1f?hNX@Y1i?gNU@Z1l?fNR@[1o?Y12O1O1O1N2O1O1000107H1O100O1O1O10O00001O0000001O01O01O00001O0O1O000O10O1WMc_OP2]`0oMl_Oi1T`0WNS@b1m?^NZ@[1e?fNc@R1]?nNj@j0W?VOPAc0Q?\\OWA<j>C]A6c>JdAN^>1jAGW>8PBAP>?i2O1O0O2O1O001O1O00N2N2O1N2O2M23N2M2O2M2O2M2O2M2O2M_PV1"}, "label": "a man wearing a green beanie and striped hoodie skating"}]}
{"id": 423250, "image": "COCO_train2014_000000423250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man smoking a cigarette\"."}], "task": "refering", "answer_template": "The \"a man smoking a cigarette\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 85, 86, 102, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 187, 188, 189, 204, 205, 206, 221, 222, 223, 238, 239, 240, 255, 256, 272, 273, 289, 290, 306, 307, 323, 324, 340, 341, 357, 358, 374, 375], "bbox": [0.0, 0.184375, 0.1764375, 0.9987812500000001], "iscrowd": 0, "area": 27151.703149999994, "rle": {"size": [640, 480], "counts": "g3X`0g310000000000000000001O000000001O001O001N2O0O2O0O2O0dA_LX<c3dCaL[<a3cC`L[<b3dC_L[;\\OWDV4<`LZ;]OYDU4:_L\\;_OWDW47]L`;^ObCGb0i40TLj;^OcCG`0m4MPLd;HnCB:f43QLc;9PDl36RLh;j6^OoGRER8l:RHQEP8l:THQEm7m:WHPEk7S9VHTH2fNj7U8dIcHQ7Y7aIWHb6e7o2K4M4K5L3L5L4L3G:N2N2N2L4J6K4M4M3N2_Oa0K5L4K5K5K5J6K5K6J6J6K5J7I6B>@b0_Oa0I8J5L4L<Dd0\\Oibf7"}, "label": "a man smoking a cigarette"}]}
{"id": 423250, "image": "COCO_train2014_000000423250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing green t - shirt holding cigarette\"."}], "task": "refering", "answer_template": "The \"man wearing green t - shirt holding cigarette\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 85, 86, 102, 119, 120, 121, 136, 137, 138, 153, 154, 155, 170, 171, 172, 187, 188, 189, 204, 205, 206, 221, 222, 223, 238, 239, 240, 255, 256, 272, 273, 289, 290, 306, 307, 323, 324, 340, 341, 357, 358, 374, 375], "bbox": [0.0, 0.184375, 0.1764375, 0.9987812500000001], "iscrowd": 0, "area": 27151.703149999994, "rle": {"size": [640, 480], "counts": "g3X`0g310000000000000000001O000000001O001O001N2O0O2O0O2O0dA_LX<c3dCaL[<a3cC`L[<b3dC_L[;\\OWDV4<`LZ;]OYDU4:_L\\;_OWDW47]L`;^ObCGb0i40TLj;^OcCG`0m4MPLd;HnCB:f43QLc;9PDl36RLh;j6^OoGRER8l:RHQEP8l:THQEm7m:WHPEk7S9VHTH2fNj7U8dIcHQ7Y7aIWHb6e7o2K4M4K5L3L5L4L3G:N2N2N2L4J6K4M4M3N2_Oa0K5L4K5K5K5J6K5K6J6J6K5J7I6B>@b0_Oa0I8J5L4L<Dd0\\Oibf7"}, "label": "man wearing green t - shirt holding cigarette"}]}
{"id": 423250, "image": "COCO_train2014_000000423250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white male with buzzed haircut wearing a white colorado long sleve shirt\"."}], "task": "refering", "answer_template": "The \"white male with buzzed haircut wearing a white colorado long sleve shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 90, 91, 106, 107, 108, 122, 123, 124, 125, 139, 140, 141, 142, 156, 157, 158, 159, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 326, 327, 328, 329, 343, 344, 345, 346, 359, 360, 361, 362, 363, 364, 380], "bbox": [0.14947916666666666, 0.21103125, 0.4843125, 0.9688749999999999], "iscrowd": 0, "area": 42564.43545, "rle": {"size": [640, 480], "counts": "gb]12mc02N3M2WHNbK3S4<gKGn3g0lK[Oi3T1QLnNe3_1VLbNf3g1TLYNi3^2cKcMY4a2fK`MV4d2iK]MS4h2kKXMR4l2mKUMo3o2PLRMl3R3SLoLi3V3VLiLg3[3XLfLd3^3ZLdLc3_3[LcLa3b3\\L`L`3d3^L]L_3g3_L[L]3i3SHaMW1hNb6l3kGmN;YMf7n3dGD4_LU8\\:L4L4L3N3O010O010O010O01O010O010O0010O010O01O010O010O0010OiN`HoEa7[9XIbFg6i8RJSGo5a8`J[G`5\\8mJaGS5T8[KhGe4o7hKnGX4g7U4F;F\\OPIkDl6V;[IgDd6W;aIfD^6X;iIeDU6Z;QJbDo5[;WJcDg5\\;_J`D`5_;fJ^DX5`;e1N11O101N2N1O2O15K7iI`G?g8XO\\Gh0k8nNXGR1n8dNVG[1R9[NQGe1U9RNoFm1X9ULXFAc0Z4[9kKXFF`0^4`9aKVFL=c4c9WKVF0;i4e9nJUF4:l4i9eJSF:7Q5l9[JSF?4V5T;eJoD\\5R;^JREc5o:XJUEg5b<1O1O001]CUJP;k5nDWJQ;i5mDZJU;d5iD^JY;`5dDcJ^;Z5aDgJb;W5[DlJg;R5WDQKj;l4TD\\Kj;a4TDfKi;X4UDnKh;P4XDVLa;l3]D[L];f3aDbLW;`3gDjLP;Z3kDPMl:U3nDUMh:P3SEZMd:k2WE^M_:U6ZNdFbH]9V7mFfHU9S7SGkHn8k6]GSId8Y3aFdNV1PN[8Z3PGmMQ1fNP8\\3_GUMl0\\Og7^3nG]Le03^7^3VK_Ll4`3WK]Lj4b3XK[Lj4c3YK[Lh4d3ZKZLg4e3\\KWLf4g3]KWLd4h3^KULd4j3^KTLc4j3k5O1O0O2N2K5L4K4M4K5L4K5L3I8F:F:F:F9G:G9F^fj4"}, "label": "white male with buzzed haircut wearing a white colorado long sleve shirt"}]}
{"id": 423250, "image": "COCO_train2014_000000423250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue jeans and a white shirt that has the number 8 on it\"."}], "task": "refering", "answer_template": "The \"a man wearing blue jeans and a white shirt that has the number 8 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 90, 91, 106, 107, 108, 122, 123, 124, 125, 139, 140, 141, 142, 156, 157, 158, 159, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 326, 327, 328, 329, 343, 344, 345, 346, 359, 360, 361, 362, 363, 364, 380], "bbox": [0.14947916666666666, 0.21103125, 0.4843125, 0.9688749999999999], "iscrowd": 0, "area": 42564.43545, "rle": {"size": [640, 480], "counts": "gb]12mc02N3M2WHNbK3S4<gKGn3g0lK[Oi3T1QLnNe3_1VLbNf3g1TLYNi3^2cKcMY4a2fK`MV4d2iK]MS4h2kKXMR4l2mKUMo3o2PLRMl3R3SLoLi3V3VLiLg3[3XLfLd3^3ZLdLc3_3[LcLa3b3\\L`L`3d3^L]L_3g3_L[L]3i3SHaMW1hNb6l3kGmN;YMf7n3dGD4_LU8\\:L4L4L3N3O010O010O010O01O010O010O0010O010O01O010O010O0010OiN`HoEa7[9XIbFg6i8RJSGo5a8`J[G`5\\8mJaGS5T8[KhGe4o7hKnGX4g7U4F;F\\OPIkDl6V;[IgDd6W;aIfD^6X;iIeDU6Z;QJbDo5[;WJcDg5\\;_J`D`5_;fJ^DX5`;e1N11O101N2N1O2O15K7iI`G?g8XO\\Gh0k8nNXGR1n8dNVG[1R9[NQGe1U9RNoFm1X9ULXFAc0Z4[9kKXFF`0^4`9aKVFL=c4c9WKVF0;i4e9nJUF4:l4i9eJSF:7Q5l9[JSF?4V5T;eJoD\\5R;^JREc5o:XJUEg5b<1O1O001]CUJP;k5nDWJQ;i5mDZJU;d5iD^JY;`5dDcJ^;Z5aDgJb;W5[DlJg;R5WDQKj;l4TD\\Kj;a4TDfKi;X4UDnKh;P4XDVLa;l3]D[L];f3aDbLW;`3gDjLP;Z3kDPMl:U3nDUMh:P3SEZMd:k2WE^M_:U6ZNdFbH]9V7mFfHU9S7SGkHn8k6]GSId8Y3aFdNV1PN[8Z3PGmMQ1fNP8\\3_GUMl0\\Og7^3nG]Le03^7^3VK_Ll4`3WK]Lj4b3XK[Lj4c3YK[Lh4d3ZKZLg4e3\\KWLf4g3]KWLd4h3^KULd4j3^KTLc4j3k5O1O0O2N2K5L4K4M4K5L4K5L3I8F:F:F:F9G:G9F^fj4"}, "label": "a man wearing blue jeans and a white shirt that has the number 8 on it"}]}
{"id": 423250, "image": "COCO_train2014_000000423250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a dark blue tshirt holding a white game controller\"."}], "task": "refering", "answer_template": "The \"a man wearing a dark blue tshirt holding a white game controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 366, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.38052083333333336, 0.21587499999999998, 0.9498125, 0.98890625], "iscrowd": 0, "area": 73103.95849999998, "rle": {"size": [640, 480], "counts": "[bb3<]c0<D=H7J7I6J3M3M3M3L4M3M3M3N2M4M2N2N2N2N2N2N2N2M3N001N1O2N1N3N1O2M2O2N1N3N1O2M2O2N1O2N1O2N1O2N1O2N1O2N101O001O0O2O001O001OR1mN3N2N3M2N3M2M4M2N2N3M2M4M2N3M2N2N3L3N3M2N3M2M3N3M2dGWJ]2]<YO5K6I6K6J2OO0010O01O00001O001O001N101O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00000000000000000001O0000000000000000000000000001N10000O10000O100O100001O001O1O1O1O1O001O0000O1PDkJU9U5k2000XDkJe8U5^GkJ_8U5cGkJZ8V5iGjJT8V5oGjJn7W5SHjJj7V5YHjJ]20_L18U5o0kJV21cL41P5X1mJo13gL6Ik4d1mJh14kL9Bf4m1oJa16nL;]Oa4V2nJ\\18QM?UO[4a2oJT1;UM`0oNY4g2nJQ1;YMc0hN`4e2cJU1>]Me0`Ng4V1nI]O9[3a0aMf0ZNi4T1kIF7W3a0dMg0WNi4S1lIL4Q3d0hMb0WNm4o0nI01n2d0lM?VNQ5l0nI5Ni2h0oM:VNU5h0PJ:Je2i0SN7TNZ5e0PJ?H`2k0WN2TN^5a0RJ`0Ga2i0YN0TNa5>RJ`0J`2g0]NKTNe5:TJ?Lb2b0aNHSNi56UJ?N^2e0iN^OSNm52WJ?OY2g0SOUOQNi7>bHU2h0ZOV8\\NYGP2h0EP8VN]Gl1i00V:FQF;Q:[OVFf0b>1N101O1N2O1O0O2O1O1N2O1O001N2OO1O001O1O00100O1O2N2N1O2N2N101N2N2N1O2N2N2O0O2N2N2N1O2N2N101N2N2M2N3L4M5J5K6K4K6JV]>"}, "label": "a man wearing a dark blue tshirt holding a white game controller"}]}
{"id": 423250, "image": "COCO_train2014_000000423250.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man who is ducking while playing a wi video game\"."}], "task": "refering", "answer_template": "The \"a man who is ducking while playing a wi video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 366, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.38052083333333336, 0.21587499999999998, 0.9498125, 0.98890625], "iscrowd": 0, "area": 73103.95849999998, "rle": {"size": [640, 480], "counts": "[bb3<]c0<D=H7J7I6J3M3M3M3L4M3M3M3N2M4M2N2N2N2N2N2N2N2M3N001N1O2N1N3N1O2M2O2N1N3N1O2M2O2N1O2N1O2N1O2N1O2N1O2N101O001O0O2O001O001OR1mN3N2N3M2N3M2M4M2N2N3M2M4M2N3M2N2N3L3N3M2N3M2M3N3M2dGWJ]2]<YO5K6I6K6J2OO0010O01O00001O001O001N101O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00000000000000000001O0000000000000000000000000001N10000O10000O100O100001O001O1O1O1O1O001O0000O1PDkJU9U5k2000XDkJe8U5^GkJ_8U5cGkJZ8V5iGjJT8V5oGjJn7W5SHjJj7V5YHjJ]20_L18U5o0kJV21cL41P5X1mJo13gL6Ik4d1mJh14kL9Bf4m1oJa16nL;]Oa4V2nJ\\18QM?UO[4a2oJT1;UM`0oNY4g2nJQ1;YMc0hN`4e2cJU1>]Me0`Ng4V1nI]O9[3a0aMf0ZNi4T1kIF7W3a0dMg0WNi4S1lIL4Q3d0hMb0WNm4o0nI01n2d0lM?VNQ5l0nI5Ni2h0oM:VNU5h0PJ:Je2i0SN7TNZ5e0PJ?H`2k0WN2TN^5a0RJ`0Ga2i0YN0TNa5>RJ`0J`2g0]NKTNe5:TJ?Lb2b0aNHSNi56UJ?N^2e0iN^OSNm52WJ?OY2g0SOUOQNi7>bHU2h0ZOV8\\NYGP2h0EP8VN]Gl1i00V:FQF;Q:[OVFf0b>1N101O1N2O1O0O2O1O1N2O1O001N2OO1O001O1O00100O1O2N2N1O2N2N101N2N2N1O2N2N2O0O2N2N2N1O2N2N101N2N2M2N3L4M5J5K6K4K6JV]>"}, "label": "a man who is ducking while playing a wi video game"}]}
{"id": 275811, "image": "COCO_train2014_000000275811.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red couch\"."}], "task": "refering", "answer_template": "The \"a red couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 298, 313, 314, 315, 316, 317, 318, 321, 339, 340, 341, 344], "bbox": [0.60834375, 0.6423887587822015, 1.0, 0.9592271662763466], "iscrowd": 0, "area": 21981.45870000001, "rle": {"size": [427, 640], "counts": "TaR51S=>C<C>C=B=C>C=B:G:L3M3M10O01O0000000000000000000000000000000000001O0000000000000000000000000000000001O01O0000000000000000000000000000000000001O0000000000000000000000000000000000010O0000000000000000000000000000000000001O0000000000000000000000000000000000010O00000000000000000000000000000000001O0000000000001O0000000000mMhF3X9KnF2Q9LTG1l8MYGMj81\\GDk8:ZG]Ol8a0YGVOn8g0XGnNo8Q1UGfNQ9Y1a1O100O1O1O1O1O100O1O1O1O1O100O1N2N2M3N2N2N2N2O1N2O1N2O1N2O1N2NVf9k0oXF101O1N2O1N2O1K5K5K5K5K4N300O10000O100O10000O01000O10nN_F"}, "label": "a red couch"}]}
{"id": 275811, "image": "COCO_train2014_000000275811.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue couch\"."}], "task": "refering", "answer_template": "The \"the blue couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.0, 0.6266978922716628, 0.230015625, 0.9885714285714285], "iscrowd": 0, "area": 15056.890799999997, "rle": {"size": [427, 640], "counts": "m8Y4Q910000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000002N5K6J6J2N2N2N2N2N1O2N2N1O2N1N3N1O2N1O1O1O1O2N1O1O1O1O1O1O2N00001O000000001O000O100O010O100O100O010O100O100O01000O100O1O0O2O1N2O1O0O2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2K5M3O1O1O2N1O2N2N2N2N3M3M4L4L4L4L4L4L3MbW]6"}, "label": "the blue couch"}]}
{"id": 13670, "image": "COCO_train2014_000000013670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty space above the catcher\"."}], "task": "refering", "answer_template": "The \"the empty space above the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 148, 154, 155], "bbox": [0.431796875, 0.2606791569086651, 0.8216093750000001, 0.467423887587822], "iscrowd": 0, "area": 12028.4623, "rle": {"size": [427, 640], "counts": "cVc31Z=9G9F;F9G9G7I000O100000O10O1000000000000000000000000000000000O10000000000000000000000000000000O1000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000O10000000000000000000000000O1000000000000000000000000000000000000O10000000000000000000000000O1000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000O1000U1kNO1000000000000000000nNR1000000000000001O00^O\\D]Od;7hDIX;LSE4m:KUE4k:LUE4k:KVE5i:JZE5f:E`E;h;0001O00001O01O01O0000001O00001O00001O00001O1O100O1O1O1O1O1O[i_1"}, "label": "the empty space above the catcher"}]}
{"id": 537960, "image": "COCO_train2014_000000537960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and tan cat lying on a hardwood floor with its belly facing upwards\"."}], "task": "refering", "answer_template": "The \"a brown and tan cat lying on a hardwood floor with its belly facing upwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 265, 266, 267, 268, 269, 270, 285, 291, 292, 312, 313, 314, 334, 335, 336], "bbox": [0.38890625, 0.5808272506082724, 0.7805937500000001, 0.9838442822384429], "iscrowd": 0, "area": 19243.86635, "rle": {"size": [411, 640], "counts": "RXT32g<3N2M3cNGTF<c9MkE0_O5e:MhE3_O3f:LgEh0Y:YOdEj0Z:XOdEi0\\:WOcEk0[:VOeEj0[:VOdEl0Z:UOfEk0Z:VOdEl0Z:UOfEk0Z:UOeEl0Z:UOeEl0[:TOeEl0Z:UOeEk0\\:UOdEk0[:VOdEk0\\:UOcEk0]:j000O01000O0100O10O0100O10O01O100O01000O10O0100O10O10O10O0100O01000O010O10O010O10O100O10O10O100000O010000O1000000O10000O100000000000000O10000000000000SHoMP5Q2lJTNT5k1kJWNT5i1lJWNT5h1lJZNS5f1mJ[NR5e1mJ\\NS5d1mJ]NR5c1nJ]NR5c1mJ^NT5a1lJ`NS5`1lJaNT5_1lJaNT5_1kJbNU5^1kJbNU5^1jJcNV5]1jJcNV5]1jJcNW5\\1hJeNX5[1hJeNX5[1gJfNY5Z1gJfNY5[1eJfN[5Z1eJfN[5Z1dJgN\\5Y1dJgN]5Y1bJgN^5Y1bJgN^5Y1bJgN^5Y1aJhN_5Y1`JgN`5Y1`JgN`5Y1`JgN`5Y1`JgN`5Z1_JfNa5Z1^JgNb5Y1^JgNb5Y1^JgNb5Y1^JgNc5Y1\\JgNd5Y1\\JgNd5Y1\\JgNd5Y1[JhNe5X1[JhNe5Y1ZJgNf5Y1ZJfNg5Z1YJfNg5Z1XJgNh5Y1XJgNh5Y1WJhNi5X1VJhNk5X1TJiNl5W1TJiNm5V1RJjNo5V1PJkNP6U1PJjNR6U1mIlNS6T1lIlNU6U1iIlNX6S1gImNZ6T1dIlN]6T1bImN_6S1_ImNb6U1[IkNf6V1XIjNi6W1UIiNl6X1RIhNo6Y1oHgNR7[1kHeNW7[1gHeNZ7\\1dHdN]7^1^HdNc7b1TH`Nm7e1kG]NV8T30O1O2N1O100O1O1N2O1O1O1N2N2N3M2O1N3KeFfL]9X35M_FkL`9S31mL_FS3b9mL^FQ3d912N101N2N2N1O2N2N2O1N3N2M3N2M3N3M2N2N2M3N2L4M3M3L3N3O001O001O00N2O1M3I7K5N2O100O1000O10O10O10001N100jNSE<o:@]E5f:GhEK\\:1Y1L4Leng1"}, "label": "a brown and tan cat lying on a hardwood floor with its belly facing upwards"}]}
{"id": 537960, "image": "COCO_train2014_000000537960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange cat on their back\"."}], "task": "refering", "answer_template": "The \"orange cat on their back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 265, 266, 267, 268, 269, 270, 285, 291, 292, 312, 313, 314, 334, 335, 336], "bbox": [0.38890625, 0.5808272506082724, 0.7805937500000001, 0.9838442822384429], "iscrowd": 0, "area": 19243.86635, "rle": {"size": [411, 640], "counts": "RXT32g<3N2M3cNGTF<c9MkE0_O5e:MhE3_O3f:LgEh0Y:YOdEj0Z:XOdEi0\\:WOcEk0[:VOeEj0[:VOdEl0Z:UOfEk0Z:VOdEl0Z:UOfEk0Z:UOeEl0Z:UOeEl0[:TOeEl0Z:UOeEk0\\:UOdEk0[:VOdEk0\\:UOcEk0]:j000O01000O0100O10O0100O10O01O100O01000O10O0100O10O10O10O0100O01000O010O10O010O10O100O10O10O100000O010000O1000000O10000O100000000000000O10000000000000SHoMP5Q2lJTNT5k1kJWNT5i1lJWNT5h1lJZNS5f1mJ[NR5e1mJ\\NS5d1mJ]NR5c1nJ]NR5c1mJ^NT5a1lJ`NS5`1lJaNT5_1lJaNT5_1kJbNU5^1kJbNU5^1jJcNV5]1jJcNV5]1jJcNW5\\1hJeNX5[1hJeNX5[1gJfNY5Z1gJfNY5[1eJfN[5Z1eJfN[5Z1dJgN\\5Y1dJgN]5Y1bJgN^5Y1bJgN^5Y1bJgN^5Y1aJhN_5Y1`JgN`5Y1`JgN`5Y1`JgN`5Y1`JgN`5Z1_JfNa5Z1^JgNb5Y1^JgNb5Y1^JgNb5Y1^JgNc5Y1\\JgNd5Y1\\JgNd5Y1\\JgNd5Y1[JhNe5X1[JhNe5Y1ZJgNf5Y1ZJfNg5Z1YJfNg5Z1XJgNh5Y1XJgNh5Y1WJhNi5X1VJhNk5X1TJiNl5W1TJiNm5V1RJjNo5V1PJkNP6U1PJjNR6U1mIlNS6T1lIlNU6U1iIlNX6S1gImNZ6T1dIlN]6T1bImN_6S1_ImNb6U1[IkNf6V1XIjNi6W1UIiNl6X1RIhNo6Y1oHgNR7[1kHeNW7[1gHeNZ7\\1dHdN]7^1^HdNc7b1TH`Nm7e1kG]NV8T30O1O2N1O100O1O1N2O1O1O1N2N2N3M2O1N3KeFfL]9X35M_FkL`9S31mL_FS3b9mL^FQ3d912N101N2N2N1O2N2N2O1N3N2M3N2M3N3M2N2N2M3N2L4M3M3L3N3O001O001O00N2O1M3I7K5N2O100O1000O10O10O10001N100jNSE<o:@]E5f:GhEK\\:1Y1L4Leng1"}, "label": "orange cat on their back"}]}
{"id": 537960, "image": "COCO_train2014_000000537960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a portion of a black piece of furniture with a patchwork quilt on it\"."}], "task": "refering", "answer_template": "The \"a portion of a black piece of furniture with a patchwork quilt on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110], "bbox": [0.57328125, 0.016739659367396593, 0.8227187499999999, 0.3386374695863747], "iscrowd": 0, "area": 14811.712699999998, "rle": {"size": [411, 640], "counts": "lZc4:R<c0K6J6K4L5J5L5K4K6K5K:F;D;F:F3L1000000O01000O1000000O01000O10000O10O1000O10000O10O1000O10000O10O10O1000000O10O101O1O1O2N1O1N2O1O2N1O1O1O1N2O2N1O1O1O1O2M2O1O1O1O2N1O1N2O1O2N1O1O1O1N3N1O1O1O1O1O2M2O1O1O1O2N1O1N2O1O2NO1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O010001O1O1O001N2O1O001O1dMb\\]1"}, "label": "a portion of a black piece of furniture with a patchwork quilt on it"}]}
{"id": 537960, "image": "COCO_train2014_000000537960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black leather chair\"."}], "task": "refering", "answer_template": "The \"black leather chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110], "bbox": [0.57328125, 0.016739659367396593, 0.8227187499999999, 0.3386374695863747], "iscrowd": 0, "area": 14811.712699999998, "rle": {"size": [411, 640], "counts": "lZc4:R<c0K6J6K4L5J5L5K4K6K5K:F;D;F:F3L1000000O01000O1000000O01000O10000O10O1000O10000O10O1000O10000O10O10O1000000O10O101O1O1O2N1O1N2O1O2N1O1O1O1N2O2N1O1O1O1O2M2O1O1O1O2N1O1N2O1O2N1O1O1O1N3N1O1O1O1O1O2M2O1O1O1O2N1O1N2O1O2NO1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O010001O1O1O001N2O1O001O1dMb\\]1"}, "label": "black leather chair"}]}
{"id": 537960, "image": "COCO_train2014_000000537960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown chair to the left of the cat\"."}], "task": "refering", "answer_template": "The \"the brown chair to the left of the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.0, 0.0, 0.40049999999999997, 0.9892457420924574], "iscrowd": 0, "area": 90907.70219999999, "rle": {"size": [411, 640], "counts": "S6d6X2o3O10000O10000O100O10000O10000O10000O100VNWEb1i:]NXEb1i:^NWEb1i:^NWEa1j:_NWE`1h:aNXE_1h:aNXE^1i:bNWE^1i:bNXE]1h:cNXE\\1i:dNWE\\1i:dNWE\\1i:dNXEZ1i:fNWEZ1i:fNWEZ1i:fNWEY1j:gNWEX1i:hNWEW1j:iNVEW1j:iNVEW1j:iNWEU1j:kNVEU1j:kNVEU1j:kNVET1k:lNVES1j:mNVES1j:mNVER1k:nNUER1k:nNVEQ1j:oNVEP1k:POUEP1k:POUEo0l:QOUEn0k:ROUEn0k:ROUEm0l:SOTEm0l:SOUEl0k:TOUEk0l:UOTEk0l:UOTEk0l:UOUEi0l:WOTEi0l:WOTEh0m:XOSEh0m:XOTEg0l:YOTEf0m:YOTEg0l:YOTEg0l:YOUEe0l:[OTEe0l:[OTEe0l:[OTEd0m:\\OTEc0l:]OTEc0l:]OUEa0l:_OTEa0l:_OUE?l:AUE>k:BUE>k:BVE<k:DUE<k:DVE;j:EWE9j:GVE9j:GWE8i:HXE6i:JWE6i:JXE5h:KYE3h:MXE3h:MYE1h:OYE0g:0YE0g:0ZENg:2YENg:2ZEMf:3[EKf:5ZEKf:5P1000000000000000000000000O1000000000000000000000000000000000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000000000001O0000000000000000001O0000000000000O1O1O1O2N1N2O1O1O1O1N2O1O1O1O2M2O1O1O1O2M:G:F9G9G:E:G:F9G9G:E:G:F9G9G:E:G:F9G:F9G9F;F9G:F9G9F;F9G:F9G9F;F9G:F9G:E:G9Gjii4"}, "label": "the brown chair to the left of the cat"}]}
{"id": 537960, "image": "COCO_train2014_000000537960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a couch is shown\"."}], "task": "refering", "answer_template": "The \"the side of a couch is shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.0, 0.0, 0.40049999999999997, 0.9892457420924574], "iscrowd": 0, "area": 90907.70219999999, "rle": {"size": [411, 640], "counts": "S6d6X2o3O10000O10000O100O10000O10000O10000O100VNWEb1i:]NXEb1i:^NWEb1i:^NWEa1j:_NWE`1h:aNXE_1h:aNXE^1i:bNWE^1i:bNXE]1h:cNXE\\1i:dNWE\\1i:dNWE\\1i:dNXEZ1i:fNWEZ1i:fNWEZ1i:fNWEY1j:gNWEX1i:hNWEW1j:iNVEW1j:iNVEW1j:iNWEU1j:kNVEU1j:kNVEU1j:kNVET1k:lNVES1j:mNVES1j:mNVER1k:nNUER1k:nNVEQ1j:oNVEP1k:POUEP1k:POUEo0l:QOUEn0k:ROUEn0k:ROUEm0l:SOTEm0l:SOUEl0k:TOUEk0l:UOTEk0l:UOTEk0l:UOUEi0l:WOTEi0l:WOTEh0m:XOSEh0m:XOTEg0l:YOTEf0m:YOTEg0l:YOTEg0l:YOUEe0l:[OTEe0l:[OTEe0l:[OTEd0m:\\OTEc0l:]OTEc0l:]OUEa0l:_OTEa0l:_OUE?l:AUE>k:BUE>k:BVE<k:DUE<k:DVE;j:EWE9j:GVE9j:GWE8i:HXE6i:JWE6i:JXE5h:KYE3h:MXE3h:MYE1h:OYE0g:0YE0g:0ZENg:2YENg:2ZEMf:3[EKf:5ZEKf:5P1000000000000000000000000O1000000000000000000000000000000000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000000000001O0000000000000000001O0000000000000O1O1O1O2N1N2O1O1O1O1N2O1O1O1O2M2O1O1O1O2M:G:F9G9G:E:G:F9G9G:E:G:F9G9G:E:G:F9G:F9G9F;F9G:F9G9F;F9G:F9G9F;F9G:F9G:E:G9Gjii4"}, "label": "the side of a couch is shown"}]}
{"id": 54632, "image": "COCO_train2014_000000054632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat sitting near the window looking at camera\"."}], "task": "refering", "answer_template": "The \"cat sitting near the window looking at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.41451562500000005, 0.3720416666666667, 0.9903281249999999, 0.9978541666666667], "iscrowd": 0, "area": 61247.12759999999, "rle": {"size": [480, 640], "counts": "U`l31o>1O1O2M2O2N1O1O2N1O2jN0WC2f<a0iB@T=Y1M2WNVNXFm1f9XNUFi1j9\\NPFf1n9`NmEa1R:eNgE]1X:hNcEY1[:mN_EV1_:f1O1N2O1O1N2001O100O1001O2N2N2N1O2N2N2N1dFcKR8_4cGkK\\8W4YGSLf8o3PGYLP9m4O1O010O1O001O1O1O001O1N1O2N2O0O2N2N101N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N3M6J1O1O2N1O1O1O000O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100000000001O000O10000000a0_Ob0^O6J5K5K4L5K5K2N1O2N1O1O2N1O2N1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O00001O0000001O000000001O000000001N100000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O1O001O1O1O001O1O1N101O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N1O2N2N2N2N2N2N2N2N1O2N2N2N2J6I7I7Jnj2"}, "label": "cat sitting near the window looking at camera"}]}
{"id": 54632, "image": "COCO_train2014_000000054632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat sitting in tall grass\"."}], "task": "refering", "answer_template": "The \"a cat sitting in tall grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.41451562500000005, 0.3720416666666667, 0.9903281249999999, 0.9978541666666667], "iscrowd": 0, "area": 61247.12759999999, "rle": {"size": [480, 640], "counts": "U`l31o>1O1O2M2O2N1O1O2N1O2jN0WC2f<a0iB@T=Y1M2WNVNXFm1f9XNUFi1j9\\NPFf1n9`NmEa1R:eNgE]1X:hNcEY1[:mN_EV1_:f1O1N2O1O1N2001O100O1001O2N2N2N1O2N2N2N1dFcKR8_4cGkK\\8W4YGSLf8o3PGYLP9m4O1O010O1O001O1O1O001O1N1O2N2O0O2N2N101N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N3M6J1O1O2N1O1O1O000O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100000000001O000O10000000a0_Ob0^O6J5K5K4L5K5K2N1O2N1O1O2N1O2N1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O00001O0000001O000000001O000000001N100000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O1O001O1O1O001O1O1N101O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N1O2N2N2N2N2N2N2N2N1O2N2N2N2J6I7I7Jnj2"}, "label": "a cat sitting in tall grass"}]}
{"id": 406895, "image": "COCO_train2014_000000406895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man without a hat sitting on the rail\"."}], "task": "refering", "answer_template": "The \"the man without a hat sitting on the rail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 22, 23, 39, 40, 41, 57, 58, 59, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 112, 113, 114, 130, 131, 132], "bbox": [0.19358, 0.06399399399399398, 0.39782000000000006, 0.6320120120120121], "iscrowd": 0, "area": 10268.417349999998, "rle": {"size": [333, 500], "counts": "jdo0h0]9;J6J4K4M3L5J5D<N3O0O2N100O2N100O2N100O2N101N101N1aI_MR5b2jJbMU5^2iJdMV5]2iJeM[4^OdKn2OfM[4AbKi22hMZ4D_Kf25gMZ4I]K`28iMY4LZK]2;hM[4V3dKlLZ4V3dKkL\\4V3bKlL]4T3bKmL^4T3`KnL_4S3_KnLa4S3]KoLb4Q3^KoLb4R3]KoLb4R3\\KoLd4R3[KoLd4R3[KnLe4R3[KoLd4]410O001O001O001O]K`KY3a4eLbKZ3]4eLgKX3Y4gLRLo2n3PMQMR2P3lMRMS2Q3iMPMW2T3dMnL[2U3aMmL^2b5O1O1O1O1O1O1O1O1O1O1O1O0^O[H\\Nf7_1`H_N`7^1cHbN]7]1eHbN\\7\\1fHcNZ7\\1gHdNY7[1mH`NT7^1n0O1O2N1O1O2N1O1M4L3M3N3M2M3N3L3N2M6GXhQ3"}, "label": "the man without a hat sitting on the rail"}]}
{"id": 406895, "image": "COCO_train2014_000000406895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young guy who is sitting near and old man\"."}], "task": "refering", "answer_template": "The \"a young guy who is sitting near and old man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 22, 23, 39, 40, 41, 57, 58, 59, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 112, 113, 114, 130, 131, 132], "bbox": [0.19358, 0.06399399399399398, 0.39782000000000006, 0.6320120120120121], "iscrowd": 0, "area": 10268.417349999998, "rle": {"size": [333, 500], "counts": "jdo0h0]9;J6J4K4M3L5J5D<N3O0O2N100O2N100O2N100O2N101N101N1aI_MR5b2jJbMU5^2iJdMV5]2iJeM[4^OdKn2OfM[4AbKi22hMZ4D_Kf25gMZ4I]K`28iMY4LZK]2;hM[4V3dKlLZ4V3dKkL\\4V3bKlL]4T3bKmL^4T3`KnL_4S3_KnLa4S3]KoLb4Q3^KoLb4R3]KoLb4R3\\KoLd4R3[KoLd4R3[KnLe4R3[KoLd4]410O001O001O001O]K`KY3a4eLbKZ3]4eLgKX3Y4gLRLo2n3PMQMR2P3lMRMS2Q3iMPMW2T3dMnL[2U3aMmL^2b5O1O1O1O1O1O1O1O1O1O1O1O0^O[H\\Nf7_1`H_N`7^1cHbN]7]1eHbN\\7\\1fHcNZ7\\1gHdNY7[1mH`NT7^1n0O1O2N1O1O2N1O1M4L3M3N3M2M3N3L3N2M6GXhQ3"}, "label": "a young guy who is sitting near and old man"}]}
{"id": 136559, "image": "COCO_train2014_000000136559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"book on the shelf near the cat ' s head\"."}], "task": "refering", "answer_template": "The \"book on the shelf near the cat ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113], "bbox": [0.08794, 0.26549333333333336, 0.33074000000000003, 0.5085066666666667], "iscrowd": 0, "area": 7413.4701, "rle": {"size": [375, 500], "counts": "bX`0`0P;:F7O2M3M3N100O100O1O100O2O0O100O1O100O100O101N1O100O100O100O1O101N100O100O1O100O100O2O0O1O100O100O100O1O2O0O100O100O1O100O2O0O100O1O100O100O101N1000000000000000001O0000000000000000010O000000PN^Gk0b8SObGk0^8SOeGl0[8SOhGk0X8SOkGl0U8SOnGk0R8SORHl0m7SOWHj0i7UO[Hh0e7VO`Hg0`7XOdHe0\\7ZOfHe0Z7YOiHf0W7VOoHh0Q9O1O100N2N2N2O1N1O2N2M]`j3"}, "label": "book on the shelf near the cat ' s head"}]}
{"id": 136559, "image": "COCO_train2014_000000136559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book disguised as a cat pillow\"."}], "task": "refering", "answer_template": "The \"a book disguised as a cat pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113], "bbox": [0.08794, 0.26549333333333336, 0.33074000000000003, 0.5085066666666667], "iscrowd": 0, "area": 7413.4701, "rle": {"size": [375, 500], "counts": "bX`0`0P;:F7O2M3M3N100O100O1O100O2O0O100O1O100O100O101N1O100O100O100O1O101N100O100O1O100O100O2O0O1O100O100O100O1O2O0O100O100O1O100O2O0O100O1O100O100O101N1000000000000000001O0000000000000000010O000000PN^Gk0b8SObGk0^8SOeGl0[8SOhGk0X8SOkGl0U8SOnGk0R8SORHl0m7SOWHj0i7UO[Hh0e7VO`Hg0`7XOdHe0\\7ZOfHe0Z7YOiHf0W7VOoHh0Q9O1O100N2N2N2O1N1O2N2M]`j3"}, "label": "a book disguised as a cat pillow"}]}
{"id": 136563, "image": "COCO_train2014_000000136563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow bus with a small black arrow on a sign pointing to it\"."}], "task": "refering", "answer_template": "The \"yellow bus with a small black arrow on a sign pointing to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 108, 109, 110, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181], "bbox": [0.0033799999999999998, 0.38584, 0.17696, 0.7993066666666667], "iscrowd": 0, "area": 10523.289900000002, "rle": {"size": [375, 500], "counts": "Ql0\\2[9]2bM10000O1000000001O001O001O001O001O001O00001N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2O001O0O2O001O001O001O001O001O1O1O1N2O001O1O1O1O1O1O1O001O1O1O1N2O1O1O001O1O1O2N3M4L3M3M4L3L4M_ef4"}, "label": "yellow bus with a small black arrow on a sign pointing to it"}]}
{"id": 136563, "image": "COCO_train2014_000000136563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus on the far left , half off screen\"."}], "task": "refering", "answer_template": "The \"the bus on the far left , half off screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 108, 109, 110, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181], "bbox": [0.0033799999999999998, 0.38584, 0.17696, 0.7993066666666667], "iscrowd": 0, "area": 10523.289900000002, "rle": {"size": [375, 500], "counts": "Ql0\\2[9]2bM10000O1000000001O001O001O001O001O001O00001N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2O001O0O2O001O001O001O001O001O1O1O1N2O001O1O1O1O1O1O1O001O1O1O1N2O1O1O001O1O1O2N3M4L3M3M4L3L4M_ef4"}, "label": "the bus on the far left , half off screen"}]}
{"id": 136563, "image": "COCO_train2014_000000136563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus to right of other buses\"."}], "task": "refering", "answer_template": "The \"bus to right of other buses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 190, 191, 192, 193], "bbox": [0.5097200000000001, 0.40026666666666666, 0.7931400000000001, 0.8109066666666668], "iscrowd": 0, "area": 16903.18375, "rle": {"size": [375, 500], "counts": "oam21d;4M1O2O0O100O2O0O2O0iDIm:h0I2M2O0O100O100O100O1`FVOS8j0^GG_8:\\GMb83]GOa82_G0_80`G3]8NbG3^8MaG5]8M`G5`8M\\G5d8_1O107H>C=B3N1N2O2M100O101N100O100O100O2O0000000O10001O000O10O1O100O1O010M3N2000O10000000000O1000000000000001O0001O1O1O001O1O001O0000001O00001O000001O00000O10000O100O10000O100K50000000000O2O0000001O1O001O001O1O002N:F;WOh0hNY1[NQbU1"}, "label": "bus to right of other buses"}]}
{"id": 136563, "image": "COCO_train2014_000000136563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bus waits for a large crowd to board\"."}], "task": "refering", "answer_template": "The \"a yellow bus waits for a large crowd to board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 190, 191, 192, 193], "bbox": [0.5097200000000001, 0.40026666666666666, 0.7931400000000001, 0.8109066666666668], "iscrowd": 0, "area": 16903.18375, "rle": {"size": [375, 500], "counts": "oam21d;4M1O2O0O100O2O0O2O0iDIm:h0I2M2O0O100O100O100O1`FVOS8j0^GG_8:\\GMb83]GOa82_G0_80`G3]8NbG3^8MaG5]8M`G5`8M\\G5d8_1O107H>C=B3N1N2O2M100O101N100O100O100O2O0000000O10001O000O10O1O100O1O010M3N2000O10000000000O1000000000000001O0001O1O1O001O1O001O0000001O00001O000001O00000O10000O100O10000O100K50000000000O2O0000001O1O001O001O1O002N:F;WOh0hNY1[NQbU1"}, "label": "a yellow bus waits for a large crowd to board"}]}
{"id": 103797, "image": "COCO_train2014_000000103797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting next to a woman on a bench\"."}], "task": "refering", "answer_template": "The \"a man sitting next to a woman on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 225, 226, 248, 249, 250, 251, 271, 272, 273, 274, 295, 296, 297, 298, 319, 320, 342, 343], "bbox": [0.78934375, 0.583887587822014, 0.98140625, 0.9549648711943793], "iscrowd": 0, "area": 6364.848599999994, "rle": {"size": [427, 640], "counts": "Ykb61Z=1O2N2N2N3ZOKPD8i;NUD5g;NWD5d;NZD6b;M\\D6`;L_D7^;o0L4L4L=C2N2M3O0O2O1N2O1N01O003M5bM]Ej1h:lM]ET2m:01O00001O0O101O001O1O1O001O1O1O010O0bNdDn0\\;QOfDm0Z;TOgDk0X;UOjDj0V;VOkDh0U;XOmDg0S;XOmDh0S;VOoDl0o:ROSEQ1k:mNVEV1g:jNYEY1d:gN\\E]1`:cNaE`1[:`NeEe1V:[NjEi1S:VNmEn1o9RNQFQ2l9oMTFS2j9mMWFU2f9kM[Fb2W9^MkF`2V9`MjF`2U9`MlF`2S9`MnF`2Q9`MPG_2P9aMRG]2n8cMSG\\2m8dMTG[2l8eMUGn0VO5e9mNWGl0CHV9\\OXGk0DGT9^OYGj0FEQ9AZGi0L_Oj8IZGf00^Of8L\\Gc00@d8M]Ga02@a8O^G?4@^81_G=5A\\82`G<6_O[85aG<3^O\\86bG>i9CWF>g9BZF?d9B]F>a9B`F`0^9_OcFb0[9_OeF`0[9@fF`0Y9AhF=X9CiF=V9_OZELa1e0U9]O]EM]1f0V9]O^EL]1e0V9_OPG`0Q9_OPGa0P9_OPG`0R9_OoF?S9_OnFa0n:O1O101N100O01O010O010O0101L3M4M2NSe4"}, "label": "a man sitting next to a woman on a bench"}]}
{"id": 103797, "image": "COCO_train2014_000000103797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 225, 226, 248, 249, 250, 251, 271, 272, 273, 274, 295, 296, 297, 298, 319, 320, 342, 343], "bbox": [0.78934375, 0.583887587822014, 0.98140625, 0.9549648711943793], "iscrowd": 0, "area": 6364.848599999994, "rle": {"size": [427, 640], "counts": "Ykb61Z=1O2N2N2N3ZOKPD8i;NUD5g;NWD5d;NZD6b;M\\D6`;L_D7^;o0L4L4L=C2N2M3O0O2O1N2O1N01O003M5bM]Ej1h:lM]ET2m:01O00001O0O101O001O1O1O001O1O1O010O0bNdDn0\\;QOfDm0Z;TOgDk0X;UOjDj0V;VOkDh0U;XOmDg0S;XOmDh0S;VOoDl0o:ROSEQ1k:mNVEV1g:jNYEY1d:gN\\E]1`:cNaE`1[:`NeEe1V:[NjEi1S:VNmEn1o9RNQFQ2l9oMTFS2j9mMWFU2f9kM[Fb2W9^MkF`2V9`MjF`2U9`MlF`2S9`MnF`2Q9`MPG_2P9aMRG]2n8cMSG\\2m8dMTG[2l8eMUGn0VO5e9mNWGl0CHV9\\OXGk0DGT9^OYGj0FEQ9AZGi0L_Oj8IZGf00^Of8L\\Gc00@d8M]Ga02@a8O^G?4@^81_G=5A\\82`G<6_O[85aG<3^O\\86bG>i9CWF>g9BZF?d9B]F>a9B`F`0^9_OcFb0[9_OeF`0[9@fF`0Y9AhF=X9CiF=V9_OZELa1e0U9]O]EM]1f0V9]O^EL]1e0V9_OPG`0Q9_OPGa0P9_OPG`0R9_OoF?S9_OnFa0n:O1O101N100O01O010O010O0101L3M4M2NSe4"}, "label": "a man wearing glasses"}]}
{"id": 103797, "image": "COCO_train2014_000000103797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting on a bench\"."}], "task": "refering", "answer_template": "The \"a woman sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 246, 247, 248, 269, 270, 271, 293, 294, 295, 317, 318, 340], "bbox": [0.7226874999999999, 0.6005620608899297, 0.866171875, 0.9553864168618267], "iscrowd": 0, "area": 5332.133600000001, "rle": {"size": [427, 640], "counts": "^[Q68Q=8H7I6J?B7I3N2M2O1N1O1CXNWEh1g:`NREa1m:?N2N2O1O1O1O1O2N1O1000000001O000O2N1O20O2O1O4L4L3M4LWMQFY2g0[MU84\\GU3GjLb9`3O010O0010O102M2N3M3[MoEk1S:QNPFm1X:kMiEW13SOf:i0cEoN_:n0gEnNZ:Q1iElNY:Q1iEmNY:R1iElNY:Q1R1O2M3N1N3M3N2M210O1000O01O00000O100N2O1O1O001O0O1O1O1O101N2NP`S1"}, "label": "a woman sitting on a bench"}]}
{"id": 103797, "image": "COCO_train2014_000000103797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman sitting on the bench\"."}], "task": "refering", "answer_template": "The \"the woman sitting on the bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 246, 247, 248, 269, 270, 271, 293, 294, 295, 317, 318, 340], "bbox": [0.7226874999999999, 0.6005620608899297, 0.866171875, 0.9553864168618267], "iscrowd": 0, "area": 5332.133600000001, "rle": {"size": [427, 640], "counts": "^[Q68Q=8H7I6J?B7I3N2M2O1N1O1CXNWEh1g:`NREa1m:?N2N2O1O1O1O1O2N1O1000000001O000O2N1O20O2O1O4L4L3M4LWMQFY2g0[MU84\\GU3GjLb9`3O010O0010O102M2N3M3[MoEk1S:QNPFm1X:kMiEW13SOf:i0cEoN_:n0gEnNZ:Q1iElNY:Q1iEmNY:R1iElNY:Q1R1O2M3N1N3M3N2M210O1000O01O00000O100N2O1O1O001O0O1O1O1O101N2NP`S1"}, "label": "the woman sitting on the bench"}]}
{"id": 103797, "image": "COCO_train2014_000000103797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bench on which a man and woman is sitting\"."}], "task": "refering", "answer_template": "The \"bench on which a man and woman is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 250, 251, 252, 268, 269, 273, 274, 275, 292, 293, 295, 296, 297, 298, 316, 321], "bbox": [0.688171875, 0.6940046838407493, 1.0, 0.9322014051522247], "iscrowd": 0, "area": 6341.53165, "rle": {"size": [427, 640], "counts": "ehg52Y=g0YOd0[O2O00001O0O101O001N2O2N1O1O2M2O1O000O1h0XO3M3M3M3M3M1O001O1O10O01o0QOO1O010O100O1O010oN[D:e;FaD5^;JjDOU;2QEHo:8SEGl:8XEEh:;[ECc:>U1O100O01000O101N3N3M2M4MZ]:0fbE2N2N2N2N1O2N2M3NO1000000O2O00000O1000001N10000O10nNHQE8m:JTE5k:LUE3j:OVE1i:0WE0g:2YENf:4YEKf:7ZEIf:7[EHe:8[EGf:9ZEEh:;XECj:>Q100000000001O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N100O2O000O2O0O2O0O101N1O1O000O100000000F:001N10000000000000000O100000ZEZNg9f1QFcNo9\\1iElNW:T1aETO_:l1000001O0000ZOcE^N]:_1nEYNS:c1WFVNo9b1a1ZOf0ZO\\E"}, "label": "bench on which a man and woman is sitting"}]}
{"id": 103797, "image": "COCO_train2014_000000103797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden bench in which two are sitting\"."}], "task": "refering", "answer_template": "The \"wooden bench in which two are sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 250, 251, 252, 268, 269, 273, 274, 275, 292, 293, 295, 296, 297, 298, 316, 321], "bbox": [0.688171875, 0.6940046838407493, 1.0, 0.9322014051522247], "iscrowd": 0, "area": 6341.53165, "rle": {"size": [427, 640], "counts": "ehg52Y=g0YOd0[O2O00001O0O101O001N2O2N1O1O2M2O1O000O1h0XO3M3M3M3M3M1O001O1O10O01o0QOO1O010O100O1O010oN[D:e;FaD5^;JjDOU;2QEHo:8SEGl:8XEEh:;[ECc:>U1O100O01000O101N3N3M2M4MZ]:0fbE2N2N2N2N1O2N2M3NO1000000O2O00000O1000001N10000O10nNHQE8m:JTE5k:LUE3j:OVE1i:0WE0g:2YENf:4YEKf:7ZEIf:7[EHe:8[EGf:9ZEEh:;XECj:>Q100000000001O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N100O2O000O2O0O2O0O101N1O1O000O100000000F:001N10000000000000000O100000ZEZNg9f1QFcNo9\\1iElNW:T1aETO_:l1000001O0000ZOcE^N]:_1nEYNS:c1WFVNo9b1a1ZOf0ZO\\E"}, "label": "wooden bench in which two are sitting"}]}
{"id": 46454, "image": "COCO_train2014_000000046454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batter\"."}], "task": "refering", "answer_template": "The \"a batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 80, 81, 82, 94, 95, 96, 97, 109, 110, 111, 112, 124, 125, 126, 127, 140, 141, 142, 155, 156, 157, 170, 171, 172, 185, 186, 187, 200, 201, 202, 215, 216, 217, 231, 232, 246, 247, 260, 261, 262, 276, 277, 290, 291, 292], "bbox": [0.3027166276346604, 0.205140625, 0.527400468384075, 0.84003125], "iscrowd": 0, "area": 22982.6875, "rle": {"size": [640, 427], "counts": "Qk`24dc0=B>XOg0ZO_Ne^Ob1Za0^Nf^Ob1Za0_Ne^Oa1Za0`Nf^Oa1Ya0_Ng^Oa1Ya0_Ng^Oa1Ya0`Nf^Oa1Xa0`Nh^O`1Xa0`Nh^O`1Xa0aNb^Oe1]a0[Na^Og1_a0ZN_^Og1ba0:10oEfM\\2Z2cMiM\\2V2cMnM]2n1bFhNa5\\Om3l1\\FPOe5UOo3j1WFWOh5POR4[3mKfLR4[3mKeLS4]3UGYL^3<\\5^3bFkLn3G`5a3\\FPMn3@f5c3UFUMQ4XOi5T5RJlJn5X5mIiJS6]5gIcJY6b5bI^Jo5_OQGY6j2XJhM@Z70TH]6e2SJlMAZ7OVHc6^2mIQNCY7NXHg6Y2hIUNDX7O[Hj6Q2dI[NEW7N]Ho6j1_IaNEW7N^HQ7f1]IdNFV7MaH2Ba6R2jIeNGT7:VHB2c6m1kIfNGT7f0ZHh5j1lIgNGS7g0\\Hg5g1mIiNFP7j0`Hd5e1mIjNFm6m0dHa5c1mIlNEj6P1gH_5a1mImNEj6P1hH`5]1lIQOEi6P1iH`5\\1lIPOEj6P1jH`5[1kIPOFj6Q1jHb5Y1fIROHj6Q1kHf5T1aIVOJj6o0lHl5n0[I[OKk6n0lHU6e0RICLl6m0lHi6McH4POMf0W7n0kHn7N]GW7e0kHn7M^GX7d0kHn7L_GY7c0kHn7K`GZ7b0kHn7K`GZ7b0kHR9V7nFjHR9V7nFjHR9V7nFjHR9V7nFjHR9U7oFkHQ9U7PGjHP9V7PGjHP9V7PGjHP9V7PGjHP9W7oFiHQ9W7oFiHQ9X7nFhHQ9Y7oFgHQ9Z7nFfHR9Z7nFfHR9[7mFeHR9\\7nFdHR9]7mFcHe34jJ1g4X7jLcH\\3NUKg0b4i6lLbHk11mMQ1X3\\6PMbHj14jMV1U3V6WM_Hj16hMY1T3S6YM^Hj19dM[1UOcNb3^7YM]H[OOP3:_Mn1k2_5TMcH]OGS3<[MQ2n2\\5QMoHf2EVMV2P3Z5mLPIk2FoLV2o1^MgNi7YORIQ3g2gNdLnNR9X2_J`NfLVOo8W2cJTNiL@g8Y2iJhMlLI_8e1RFbNn6CU3c1^FZNe6MR3`1Y8HZQn3"}, "label": "a batter"}]}
{"id": 46454, "image": "COCO_train2014_000000046454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player holding a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player holding a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 80, 81, 82, 94, 95, 96, 97, 109, 110, 111, 112, 124, 125, 126, 127, 140, 141, 142, 155, 156, 157, 170, 171, 172, 185, 186, 187, 200, 201, 202, 215, 216, 217, 231, 232, 246, 247, 260, 261, 262, 276, 277, 290, 291, 292], "bbox": [0.3027166276346604, 0.205140625, 0.527400468384075, 0.84003125], "iscrowd": 0, "area": 22982.6875, "rle": {"size": [640, 427], "counts": "Qk`24dc0=B>XOg0ZO_Ne^Ob1Za0^Nf^Ob1Za0_Ne^Oa1Za0`Nf^Oa1Ya0_Ng^Oa1Ya0_Ng^Oa1Ya0`Nf^Oa1Xa0`Nh^O`1Xa0`Nh^O`1Xa0aNb^Oe1]a0[Na^Og1_a0ZN_^Og1ba0:10oEfM\\2Z2cMiM\\2V2cMnM]2n1bFhNa5\\Om3l1\\FPOe5UOo3j1WFWOh5POR4[3mKfLR4[3mKeLS4]3UGYL^3<\\5^3bFkLn3G`5a3\\FPMn3@f5c3UFUMQ4XOi5T5RJlJn5X5mIiJS6]5gIcJY6b5bI^Jo5_OQGY6j2XJhM@Z70TH]6e2SJlMAZ7OVHc6^2mIQNCY7NXHg6Y2hIUNDX7O[Hj6Q2dI[NEW7N]Ho6j1_IaNEW7N^HQ7f1]IdNFV7MaH2Ba6R2jIeNGT7:VHB2c6m1kIfNGT7f0ZHh5j1lIgNGS7g0\\Hg5g1mIiNFP7j0`Hd5e1mIjNFm6m0dHa5c1mIlNEj6P1gH_5a1mImNEj6P1hH`5]1lIQOEi6P1iH`5\\1lIPOEj6P1jH`5[1kIPOFj6Q1jHb5Y1fIROHj6Q1kHf5T1aIVOJj6o0lHl5n0[I[OKk6n0lHU6e0RICLl6m0lHi6McH4POMf0W7n0kHn7N]GW7e0kHn7M^GX7d0kHn7L_GY7c0kHn7K`GZ7b0kHn7K`GZ7b0kHR9V7nFjHR9V7nFjHR9V7nFjHR9V7nFjHR9U7oFkHQ9U7PGjHP9V7PGjHP9V7PGjHP9V7PGjHP9W7oFiHQ9W7oFiHQ9X7nFhHQ9Y7oFgHQ9Z7nFfHR9Z7nFfHR9[7mFeHR9\\7nFdHR9]7mFcHe34jJ1g4X7jLcH\\3NUKg0b4i6lLbHk11mMQ1X3\\6PMbHj14jMV1U3V6WM_Hj16hMY1T3S6YM^Hj19dM[1UOcNb3^7YM]H[OOP3:_Mn1k2_5TMcH]OGS3<[MQ2n2\\5QMoHf2EVMV2P3Z5mLPIk2FoLV2o1^MgNi7YORIQ3g2gNdLnNR9X2_J`NfLVOo8W2cJTNiL@g8Y2iJhMlLI_8e1RFbNn6CU3c1^FZNe6MR3`1Y8HZQn3"}, "label": "a baseball player holding a bat"}]}
{"id": 398712, "image": "COCO_train2014_000000398712.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant who ' s right eye is visible\"."}], "task": "refering", "answer_template": "The \"the elephant who ' s right eye is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 220, 221, 222, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 244, 245, 253, 254, 255, 256, 258, 260, 261, 262, 263, 264, 267, 268, 276, 277, 278, 279, 283, 284, 285, 286, 287, 290, 291, 299, 300, 301, 306, 307, 309, 310, 313, 314, 322, 323, 324, 329, 330, 332, 333, 337, 338, 345, 346, 347, 348, 351, 352, 353, 355, 356, 360, 361], "bbox": [0.00021875000000000003, 0.09779166666666667, 0.714453125, 0.9342708333333333], "iscrowd": 0, "area": 123435.38875000004, "rle": {"size": [480, 640], "counts": "b4V3Z3gMW4\\2`JiNb5Y1PJQOR6T1\\IWOf6P1gHYO[7T42N2I7G8G:G9F:B=F;3M3M3L4M3M3M3L4M4eJ]JR1f5dNfJW1]5`NlJ\\1X5\\NnJc1U5TNRKk1Q5lMVKS2m4eMYKY2k4_M\\K_2g4XM`Kg2c4QMcKo2`4hLfKW3^4bLfK]3_4ZLfKe3^4TLgKj3]4oKgKP4e700O1O1O2N1O100O1O1O1O1O100O1O1O1O100O1O100O100O1O100O100O1O100O100O100O1O100O10bMgMUHX2c7WNWHh1b7eNYH[1g7jNSHV1n7POlGP1U8SOgGl0Z8UOdGk0]8VObGj0^8XO`Gg0b8YO\\Gh0d8YO[Gf0f8[OYGd0i8]OUGc0j8@SG`0m8j2O00001N101O0\\OoIZHQ6e7YJRHh5l7aJlG_5T8f0O2N10010O010O100O01000O100O10O100000000O10000000000O2O0000000O10000000000O100000000O1000000000000O100000000000000O100000000001O00000000001O000O1000001O000O1000001O00000O10001O0000nLVHnNj7i0_HWO`7g0cHYO]7d0fH]OY7`0jH@V7=mHCS7:PIFP77SIIm64VILi62ZIOe6L`I4`6GeI9[6BjI>V6]OoIc0Q6fNfJZ1Z5PN\\KQ2c4[MQLe2U8000000010O00000000001O000000001O00000000001O000000001O00000ROTMhEl2U:YMiEh2T:]MiEc2T:bMiE_2T:gMiEY2T:lMjET2Q:SNmEn1n9WNQFi1j9^NTFb1h9m1K6K4K5L4K5J7J5J6J6K5J7L3N2O1N3N1N2N2M3M4L3L47I8H8I8G8H8H8I7H8H8VLbFb1f9nMbFR2f9dM\\F]2k9YMWFg2R:nLPFR3X:dLjE\\3m:1O0000001O0O1000001O0000001O0000001O000000001O000000001O0000000000001WO_LQFa3g9iLWFW3^9UMaFk2T9aMkF_2j8mMUGS2`8ZN^Gf1W8fNhGZ1h4^LnMc2ZMn0i4mLTMa2RNb0j4[M[L_2jN6l4hMPKo23YOm4Z2RKfMo4Y2QKgMo4Z2PKeMQ5\\2nJdMS5T6001O001N101O00001O0O2O00001O001O00001O00001O0O2O00001O001O00001O1O100O001O1O1O1O1O1O1O1O1O1O2N3M3M4L3M3M3M3M2N2N3M2O1N2N2O2M2O`0_O6K:F:E;F2M2O0O2O1O1N200001O1O2O1N2N01O0OO2O1O2N3M4L3M4L4ZMdJ_K^5Z4lJaKU5Y4SKdKn4S4[KmKd4j3fKUL[4`3PL`LP4Y3VLgLl3M^IP2k2SNh3BmIg1i2gNc3oNQJo1d2SOj4:]KHg4KaK6b4^OfKb0]90O2ON2N1O2N1O2N1O1O2N1N2O4L3L5Kh[e2"}, "label": "the elephant who ' s right eye is visible"}]}
{"id": 398712, "image": "COCO_train2014_000000398712.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray elephant that we see the entire side of\"."}], "task": "refering", "answer_template": "The \"the gray elephant that we see the entire side of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 220, 221, 222, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 244, 245, 253, 254, 255, 256, 258, 260, 261, 262, 263, 264, 267, 268, 276, 277, 278, 279, 283, 284, 285, 286, 287, 290, 291, 299, 300, 301, 306, 307, 309, 310, 313, 314, 322, 323, 324, 329, 330, 332, 333, 337, 338, 345, 346, 347, 348, 351, 352, 353, 355, 356, 360, 361], "bbox": [0.00021875000000000003, 0.09779166666666667, 0.714453125, 0.9342708333333333], "iscrowd": 0, "area": 123435.38875000004, "rle": {"size": [480, 640], "counts": "b4V3Z3gMW4\\2`JiNb5Y1PJQOR6T1\\IWOf6P1gHYO[7T42N2I7G8G:G9F:B=F;3M3M3L4M3M3M3L4M4eJ]JR1f5dNfJW1]5`NlJ\\1X5\\NnJc1U5TNRKk1Q5lMVKS2m4eMYKY2k4_M\\K_2g4XM`Kg2c4QMcKo2`4hLfKW3^4bLfK]3_4ZLfKe3^4TLgKj3]4oKgKP4e700O1O1O2N1O100O1O1O1O1O100O1O1O1O100O1O100O100O1O100O100O1O100O100O100O1O100O10bMgMUHX2c7WNWHh1b7eNYH[1g7jNSHV1n7POlGP1U8SOgGl0Z8UOdGk0]8VObGj0^8XO`Gg0b8YO\\Gh0d8YO[Gf0f8[OYGd0i8]OUGc0j8@SG`0m8j2O00001N101O0\\OoIZHQ6e7YJRHh5l7aJlG_5T8f0O2N10010O010O100O01000O100O10O100000000O10000000000O2O0000000O10000000000O100000000O1000000000000O100000000000000O100000000001O00000000001O000O1000001O000O1000001O00000O10001O0000nLVHnNj7i0_HWO`7g0cHYO]7d0fH]OY7`0jH@V7=mHCS7:PIFP77SIIm64VILi62ZIOe6L`I4`6GeI9[6BjI>V6]OoIc0Q6fNfJZ1Z5PN\\KQ2c4[MQLe2U8000000010O00000000001O000000001O00000000001O000000001O00000ROTMhEl2U:YMiEh2T:]MiEc2T:bMiE_2T:gMiEY2T:lMjET2Q:SNmEn1n9WNQFi1j9^NTFb1h9m1K6K4K5L4K5J7J5J6J6K5J7L3N2O1N3N1N2N2M3M4L3L47I8H8I8G8H8H8I7H8H8VLbFb1f9nMbFR2f9dM\\F]2k9YMWFg2R:nLPFR3X:dLjE\\3m:1O0000001O0O1000001O0000001O0000001O000000001O000000001O0000000000001WO_LQFa3g9iLWFW3^9UMaFk2T9aMkF_2j8mMUGS2`8ZN^Gf1W8fNhGZ1h4^LnMc2ZMn0i4mLTMa2RNb0j4[M[L_2jN6l4hMPKo23YOm4Z2RKfMo4Y2QKgMo4Z2PKeMQ5\\2nJdMS5T6001O001N101O00001O0O2O00001O001O00001O00001O0O2O00001O001O00001O1O100O001O1O1O1O1O1O1O1O1O1O2N3M3M4L3M3M3M3M2N2N3M2O1N2N2O2M2O`0_O6K:F:E;F2M2O0O2O1O1N200001O1O2O1N2N01O0OO2O1O2N3M4L3M4L4ZMdJ_K^5Z4lJaKU5Y4SKdKn4S4[KmKd4j3fKUL[4`3PL`LP4Y3VLgLl3M^IP2k2SNh3BmIg1i2gNc3oNQJo1d2SOj4:]KHg4KaK6b4^OfKb0]90O2ON2N1O2N1O2N1O1O2N1N2O4L3L5Kh[e2"}, "label": "the gray elephant that we see the entire side of"}]}
{"id": 398712, "image": "COCO_train2014_000000398712.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant facing the camera\"."}], "task": "refering", "answer_template": "The \"elephant facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 60, 61, 62, 63, 64, 83, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 338, 339, 340, 361, 362, 363], "bbox": [0.63565625, 0.10527083333333334, 0.8895781249999999, 0.9146874999999999], "iscrowd": 0, "area": 37006.73855000001, "rle": {"size": [480, 640], "counts": "dkn53m>2M3M4L3M4M2M4L3M3M4L3N3L3M3M4L3N3L3M4TJSN^NP2X1ZNeNi1P1bNnN`1h0jNUOZ1?PO_OS17WOFl0O_ONd0HF6<_OO>5SO8k0KfNd0W1LmMc0P25lL:Q39PL5o3`5M3M3N1N3M3N2M2N3N2M4M2N2O1O2M2O1N3N1N2O00O001O001O001O1O0010O01000O10O001O00100O001O0010O01O1O001O010O001O001N101O001N101O001N101O1O0bMYLbIi3W6gL_IY3\\6VMZIk2_6eMWI[2d6TNQIn1i6aNmH_1n6POhHQ1S7`3M4L3M4K4M4L3M4L3L5L3M4L4L3N3M2N3M2O2M2N3M2O2M2N3M2N3N1L5L3M4K5L3M4K4M5J7J6J6I7J6J6I7J6J6I7J6J6I7J6J6I7J6J6ITTQ1"}, "label": "elephant facing the camera"}]}
{"id": 398712, "image": "COCO_train2014_000000398712.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant who is looking straight at you\"."}], "task": "refering", "answer_template": "The \"the elephant who is looking straight at you\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 60, 61, 62, 63, 64, 83, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 338, 339, 340, 361, 362, 363], "bbox": [0.63565625, 0.10527083333333334, 0.8895781249999999, 0.9146874999999999], "iscrowd": 0, "area": 37006.73855000001, "rle": {"size": [480, 640], "counts": "dkn53m>2M3M4L3M4M2M4L3M3M4L3N3L3M3M4L3N3L3M4TJSN^NP2X1ZNeNi1P1bNnN`1h0jNUOZ1?PO_OS17WOFl0O_ONd0HF6<_OO>5SO8k0KfNd0W1LmMc0P25lL:Q39PL5o3`5M3M3N1N3M3N2M2N3N2M4M2N2O1O2M2O1N3N1N2O00O001O001O001O1O0010O01000O10O001O00100O001O0010O01O1O001O010O001O001N101O001N101O001N101O1O0bMYLbIi3W6gL_IY3\\6VMZIk2_6eMWI[2d6TNQIn1i6aNmH_1n6POhHQ1S7`3M4L3M4K4M4L3M4L3L5L3M4L4L3N3M2N3M2O2M2N3M2O2M2N3M2N3N1L5L3M4K5L3M4K4M5J7J6J6I7J6J6I7J6J6I7J6J6I7J6J6I7J6J6ITTQ1"}, "label": "the elephant who is looking straight at you"}]}
{"id": 226681, "image": "COCO_train2014_000000226681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of bread topped with egg\"."}], "task": "refering", "answer_template": "The \"a slice of bread topped with egg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215], "bbox": [0.113390625, 0.0, 0.755953125, 0.439421875], "iscrowd": 0, "area": 72798.38999999998, "rle": {"size": [640, 640], "counts": "dg]15jc05K5K5K4L5K4L5K4L4L5K4L4L5J5L4L5K2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O2N7I6I8I7I6J1O1O1O2N1N2O1O2N1O1O1O000O01000000O1000O10O100000000O10O1000O10000O10O10O10000O10000O0100000O10000O01000O10000O1000O010000O100000O010000O10000O10O10O10000O1000O01000000O10000O01000O10000O10O10O10000O0100O0100O010O010O10O010O0100O010O10O010O0100O010O010O10O010O01000O0100O010O010O10O010O0100O010O010O10O0100O010O010O10O010O0100O010O010O10O0100O0100O0100O010O010O10O010O0100O010O10O010O0100O010O010O10O010O0100O101N100O100O2O0O10000O2O0O100O101N100O100O2O0O100O101N100O100O2O0O1000001N1000000O2O000O10001O0O1000001N1000000O2O000O101O000O10001O0O1000001N10000O101O000O10001O0O1000001N10000O101O000O4M3M3M2M4M3M3L4M2N3M3L4M3M2N3L4M3M3K4L5K5K5K5K4L6J7I6J7J6I6J7I7I6J7In[Q3"}, "label": "a slice of bread topped with egg"}]}
{"id": 226681, "image": "COCO_train2014_000000226681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of toast with the egg on top\"."}], "task": "refering", "answer_template": "The \"the slice of toast with the egg on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215], "bbox": [0.113390625, 0.0, 0.755953125, 0.439421875], "iscrowd": 0, "area": 72798.38999999998, "rle": {"size": [640, 640], "counts": "dg]15jc05K5K5K4L5K4L5K4L4L5K4L4L5J5L4L5K2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O2N7I6I8I7I6J1O1O1O2N1N2O1O2N1O1O1O000O01000000O1000O10O100000000O10O1000O10000O10O10O10000O10000O0100000O10000O01000O10000O1000O010000O100000O010000O10000O10O10O10000O1000O01000000O10000O01000O10000O10O10O10000O0100O0100O010O010O10O010O0100O010O10O010O0100O010O010O10O010O01000O0100O010O010O10O010O0100O010O010O10O0100O010O010O10O010O0100O010O010O10O0100O0100O0100O010O010O10O010O0100O010O10O010O0100O010O010O10O010O0100O101N100O100O2O0O10000O2O0O100O101N100O100O2O0O100O101N100O100O2O0O1000001N1000000O2O000O10001O0O1000001N1000000O2O000O101O000O10001O0O1000001N10000O101O000O10001O0O1000001N10000O101O000O4M3M3M2M4M3M3L4M2N3M3L4M3M2N3L4M3M3K4L5K5K5K5K4L6J7I6J7J6I6J7I7I6J7In[Q3"}, "label": "the slice of toast with the egg on top"}]}
{"id": 226681, "image": "COCO_train2014_000000226681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate of food has one sausage that has bite marks\"."}], "task": "refering", "answer_template": "The \"the plate of food has one sausage that has bite marks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [402, 403, 404, 405, 424, 425, 426, 427, 428, 429, 430, 431, 446, 447, 448, 449, 450, 451, 452, 453, 454, 469, 470, 471, 472, 473, 474, 475, 476, 477, 493, 494, 495, 496, 497, 498, 499, 500, 519, 520, 521, 522], "bbox": [0.4211875, 0.760265625, 0.78325, 0.974546875], "iscrowd": 0, "area": 23767.29455, "rle": {"size": [640, 640], "counts": "UYY56ec07I8I6I8I6L5L3M3M4L3N1N3M2N2N2N2O1N2N2N2N2N2O1N2N2N2N3M2O1N2N2N2N2O1N100O100O2O0O10000O100O2O0O100O10000O2O0O100O100O101N10000O100000001O000001O000001O0001O00000001O0001O00000001O01O00000000001O01O0000000001O000001O0001O0000000000010O00000000001O000001O000001O00000000010O00000000001O0001O00000001O00000000010O0000001O001O00010O001O00001O001O01O01O001O001O000010O01O00001O001O000010O01O0O1O2N101N1O2N1O101N2N2N2N3N1N2N2N2O1N2N2N3M2O1N2N2N2O1N2E;B?J5L4K5L4K5K5LmZf2"}, "label": "the plate of food has one sausage that has bite marks"}]}
{"id": 226681, "image": "COCO_train2014_000000226681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog next to bacon\"."}], "task": "refering", "answer_template": "The \"a hot dog next to bacon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [402, 403, 404, 405, 424, 425, 426, 427, 428, 429, 430, 431, 446, 447, 448, 449, 450, 451, 452, 453, 454, 469, 470, 471, 472, 473, 474, 475, 476, 477, 493, 494, 495, 496, 497, 498, 499, 500, 519, 520, 521, 522], "bbox": [0.4211875, 0.760265625, 0.78325, 0.974546875], "iscrowd": 0, "area": 23767.29455, "rle": {"size": [640, 640], "counts": "UYY56ec07I8I6I8I6L5L3M3M4L3N1N3M2N2N2N2O1N2N2N2N2N2O1N2N2N2N3M2O1N2N2N2N2O1N100O100O2O0O10000O100O2O0O100O10000O2O0O100O100O101N10000O100000001O000001O000001O0001O00000001O0001O00000001O01O00000000001O01O0000000001O000001O0001O0000000000010O00000000001O000001O000001O00000000010O00000000001O0001O00000001O00000000010O0000001O001O00010O001O00001O001O01O01O001O001O000010O01O00001O001O000010O01O0O1O2N101N1O2N1O101N2N2N2N3N1N2N2N2O1N2N2N3M2O1N2N2N2O1N2E;B?J5L4K5L4K5K5LmZf2"}, "label": "a hot dog next to bacon"}]}
{"id": 152954, "image": "COCO_train2014_000000152954.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a blue and white plaid shirt shearing a sheep\"."}], "task": "refering", "answer_template": "The \"a person in a blue and white plaid shirt shearing a sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 175, 176, 177, 194, 195, 198, 199, 221, 222, 244, 245, 266, 267, 268, 288, 289, 290, 291], "bbox": [0.389015625, 0.11229166666666666, 0.77053125, 0.7598958333333333], "iscrowd": 0, "area": 32501.021700000016, "rle": {"size": [480, 640], "counts": "Tjd36^>=C=H8N2N2O0O2N2N2N2N2O1N2N1O2N2N101N2N1O2O1O001O1O001O010O1O001O1O1O001O010O0kFdMf5]2VJkMe5U2XJQNe5P2WJRNj5n1SJUNm5k1PJWNQ6j1kIYNU6g1hI[NZ6d1dI^N\\6b1VIbMWNk0d8c1TIfMTNh0h8b1SIiMRNe0m8a1PInMPNa0Q9`1nHoNT7P1kHQOW7m0hHSOZ7l0eHUO\\7j0cHVO_7k0^HUOd7m0XHSOj7o0RHQOP8Q1kGPOW8R1eGnN]8T1fFE\\9m23N2N2N1O1O2M2O1O2N1N2O2N1O1O1N3N1O1O1O1O100O1O100O1O100O1O100O1O1N2N2O1N2N2N2O1N2N2N2O00100O10eIoLg0P3YOQMg0o2XORMh0n2WOSMi0l2WOUMi0k2VOVMj0i2VOXMj0i2TOXMl0i2ROXMn0h2QOXMP1i2nNXMR1h2lNZMT1g2jNZMV1g2hNZMX1f2gN[MY1f2eN[M[1f2cN[M]1e2bN\\M^1e2`N\\M`1e2^N\\Mb1d2]N]Mc1c2\\N^Md1b2[N_Me1a2ZN`Mf1`2YNaMg1_2XNbMg1_2XNbMh1]2WNeMi1[2VNfMj1Z2UNgMl1X2SNiMm1W2RNjMn1V2QNkMo1U2mMoMS2Q2jMRNV2n1VKaLX1d1b3k1TKkLo0]1m3g1SKUMf0W1X4c1QK^M=R1b4`1oJhM3l0n4\\1mJQNKf0X5Y1lJYNBa0b5V1jJcNYO:m5S1iJU1W5kNgJW1Z5hNdJZ1\\5fNcJ[1]5dNbJ^1^5bN`J`1`5`N_Ja1a5_N]Jc1c5]N[Je1e5[NZJf1g5YNWJi1i5WNUJj1l5VNSJi1o5WNoIj1R6UNmIk1U6UNjIk1W6UNgIh1^6Y35J5K5K5K5K5K5K6J5K5K5L4K5K6J5K5M3M3L4M4L3M3L4M3M3M4K4M3M3M3L4M4M2N2O1N2O1N3N1N2O1N2O1N3N1N2O1N3N3L3N2M4L3N2M4M2M3N3L3N2M4M2M3L5K4L5K4L4L5K5K8G9H8HlhT2"}, "label": "a person in a blue and white plaid shirt shearing a sheep"}]}
{"id": 152954, "image": "COCO_train2014_000000152954.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in blue jeans , bending over a sheep and shearing its wool off\"."}], "task": "refering", "answer_template": "The \"a person in blue jeans , bending over a sheep and shearing its wool off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 175, 176, 177, 194, 195, 198, 199, 221, 222, 244, 245, 266, 267, 268, 288, 289, 290, 291], "bbox": [0.389015625, 0.11229166666666666, 0.77053125, 0.7598958333333333], "iscrowd": 0, "area": 32501.021700000016, "rle": {"size": [480, 640], "counts": "Tjd36^>=C=H8N2N2O0O2N2N2N2N2O1N2N1O2N2N101N2N1O2O1O001O1O001O010O1O001O1O1O001O010O0kFdMf5]2VJkMe5U2XJQNe5P2WJRNj5n1SJUNm5k1PJWNQ6j1kIYNU6g1hI[NZ6d1dI^N\\6b1VIbMWNk0d8c1TIfMTNh0h8b1SIiMRNe0m8a1PInMPNa0Q9`1nHoNT7P1kHQOW7m0hHSOZ7l0eHUO\\7j0cHVO_7k0^HUOd7m0XHSOj7o0RHQOP8Q1kGPOW8R1eGnN]8T1fFE\\9m23N2N2N1O1O2M2O1O2N1N2O2N1O1O1N3N1O1O1O1O100O1O100O1O100O1O100O1O1N2N2O1N2N2N2O1N2N2N2O00100O10eIoLg0P3YOQMg0o2XORMh0n2WOSMi0l2WOUMi0k2VOVMj0i2VOXMj0i2TOXMl0i2ROXMn0h2QOXMP1i2nNXMR1h2lNZMT1g2jNZMV1g2hNZMX1f2gN[MY1f2eN[M[1f2cN[M]1e2bN\\M^1e2`N\\M`1e2^N\\Mb1d2]N]Mc1c2\\N^Md1b2[N_Me1a2ZN`Mf1`2YNaMg1_2XNbMg1_2XNbMh1]2WNeMi1[2VNfMj1Z2UNgMl1X2SNiMm1W2RNjMn1V2QNkMo1U2mMoMS2Q2jMRNV2n1VKaLX1d1b3k1TKkLo0]1m3g1SKUMf0W1X4c1QK^M=R1b4`1oJhM3l0n4\\1mJQNKf0X5Y1lJYNBa0b5V1jJcNYO:m5S1iJU1W5kNgJW1Z5hNdJZ1\\5fNcJ[1]5dNbJ^1^5bN`J`1`5`N_Ja1a5_N]Jc1c5]N[Je1e5[NZJf1g5YNWJi1i5WNUJj1l5VNSJi1o5WNoIj1R6UNmIk1U6UNjIk1W6UNgIh1^6Y35J5K5K5K5K5K5K6J5K5K5L4K5K6J5K5M3M3L4M4L3M3L4M3M3M4K4M3M3M3L4M4M2N2O1N2O1N3N1N2O1N2O1N3N1N2O1N3N3L3N2M4L3N2M4M2M3N3L3N2M4M2M3L5K4L5K4L4L5K5K8G9H8HlhT2"}, "label": "a person in blue jeans , bending over a sheep and shearing its wool off"}]}
{"id": 152954, "image": "COCO_train2014_000000152954.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown cow right behind woman\"."}], "task": "refering", "answer_template": "The \"brown cow right behind woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 167, 170], "bbox": [0.21276562499999999, 0.18570833333333334, 0.490609375, 0.4513125], "iscrowd": 0, "area": 12622.2792, "rle": {"size": [480, 640], "counts": "Qno11m>2N1O2M3N1O2N2O001N2O0O2jM_OlEb0S:GdE9\\:K_E7_:N]E3b:1ZE0e:4VEMj:8QEIn:<lDFR;?hDCX;b0`DAa;c0XD_Oi;e0PD]OQ<i10O1O1O1O1O12N2M4M<D1O100O001O10N100O2N1O1O1L5K4L4N3O01O010O0000O2O0O101N100O100O1O2O0O11O1O010O1O1O001O100O001OO101O0O10000O101O0O1000001O00000000010O000000001O0001OO10001O0O102N2N2M2O00000O100000M302UDiMi:Z2oDkMQ;W2hDnMX;U2`DPNb;f21O2M2O001O000N3L5L4K6J4L3M3N2M3N2O3M3M100O1O1O1O1O101N1O2N2N101N1O1O1O1O010O3M4L4L4L4M2M4L4L4L4Modh4"}, "label": "brown cow right behind woman"}]}
{"id": 152954, "image": "COCO_train2014_000000152954.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bull directly behind the woman\"."}], "task": "refering", "answer_template": "The \"a bull directly behind the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 167, 170], "bbox": [0.21276562499999999, 0.18570833333333334, 0.490609375, 0.4513125], "iscrowd": 0, "area": 12622.2792, "rle": {"size": [480, 640], "counts": "Qno11m>2N1O2M3N1O2N2O001N2O0O2jM_OlEb0S:GdE9\\:K_E7_:N]E3b:1ZE0e:4VEMj:8QEIn:<lDFR;?hDCX;b0`DAa;c0XD_Oi;e0PD]OQ<i10O1O1O1O1O12N2M4M<D1O100O001O10N100O2N1O1O1L5K4L4N3O01O010O0000O2O0O101N100O100O1O2O0O11O1O010O1O1O001O100O001OO101O0O10000O101O0O1000001O00000000010O000000001O0001OO10001O0O102N2N2M2O00000O100000M302UDiMi:Z2oDkMQ;W2hDnMX;U2`DPNb;f21O2M2O001O000N3L5L4K6J4L3M3N2M3N2O3M3M100O1O1O1O1O101N1O2N2N101N1O1O1O1O010O3M4L4L4L4M2M4L4L4L4Modh4"}, "label": "a bull directly behind the woman"}]}
{"id": 38266, "image": "COCO_train2014_000000038266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef stirring something with a large wooden spoon\"."}], "task": "refering", "answer_template": "The \"a chef stirring something with a large wooden spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 59, 79, 80, 81, 82, 102, 103, 104, 125, 126, 127, 148, 149, 150, 151, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.3558125, 0.11910377358490566, 0.6342187499999999, 0.9752830188679245], "iscrowd": 0, "area": 43507.43880000001, "rle": {"size": [424, 640], "counts": "Zen2X2d:>M2M4M2M4M3L3N3L3N3L3N3L3N3M3VHTLR6o3eIYLZ6j3\\I^Lb6e3UIbLk6a3kHgLS7\\3dHlL[7a4N2O1N2O1N2O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O001O100O1O1bNYJZKh5c4`JWKa5g4a1VMfHC]7=WIlNl6T1gIVN\\6j1a20000000000000000000000000000000000O01000000000000000000000000000000000000000000000001O0000001O0O1000001O0000001O0000001O000000001O0000001O0O10001O0TGeNk5[1lIIY57^J3a5MVJ=i5DmIf0R6ZOYI\\1f6d2O001O1O1O1O1oNQ1O:E:F3N2M3M3M3N2M2N3M3N2M3M3lNdGWM^8\\2nGdMT8W2oGhMT8S2nGmMU8n1nGRNT8j1nGUNW8d1lG[NY8^1iGbN[8X1hGhN\\8k0mGTOT;O1O010O1OX=0WTP3"}, "label": "a chef stirring something with a large wooden spoon"}]}
{"id": 38266, "image": "COCO_train2014_000000038266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef holding a long wooden handle\"."}], "task": "refering", "answer_template": "The \"a chef holding a long wooden handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 59, 79, 80, 81, 82, 102, 103, 104, 125, 126, 127, 148, 149, 150, 151, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335], "bbox": [0.3558125, 0.11910377358490566, 0.6342187499999999, 0.9752830188679245], "iscrowd": 0, "area": 43507.43880000001, "rle": {"size": [424, 640], "counts": "Zen2X2d:>M2M4M2M4M3L3N3L3N3L3N3L3N3M3VHTLR6o3eIYLZ6j3\\I^Lb6e3UIbLk6a3kHgLS7\\3dHlL[7a4N2O1N2O1N2O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O001O100O1O1bNYJZKh5c4`JWKa5g4a1VMfHC]7=WIlNl6T1gIVN\\6j1a20000000000000000000000000000000000O01000000000000000000000000000000000000000000000001O0000001O0O1000001O0000001O0000001O000000001O0000001O0O10001O0TGeNk5[1lIIY57^J3a5MVJ=i5DmIf0R6ZOYI\\1f6d2O001O1O1O1O1oNQ1O:E:F3N2M3M3M3N2M2N3M3N2M3M3lNdGWM^8\\2nGdMT8W2oGhMT8S2nGmMU8n1nGRNT8j1nGUNW8d1lG[NY8^1iGbN[8X1hGhN\\8k0mGTOT;O1O010O1OX=0WTP3"}, "label": "a chef holding a long wooden handle"}]}
{"id": 38266, "image": "COCO_train2014_000000038266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person working in the far left corner\"."}], "task": "refering", "answer_template": "The \"a person working in the far left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 93, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255], "bbox": [0.001484375, 0.2561792452830189, 0.20842187499999998, 0.7932547169811321], "iscrowd": 0, "area": 18424.232850000004, "rle": {"size": [424, 640], "counts": "]f07[7P4TJ^L]5f3`J^L\\5d3aJ]L_5f3^JZLb5h3\\JXLd5k3YJULg5n3UJSLk5o3SJQLm5R4PJnKP6T4nIkKS6X4iIiKW6Y4gIgKY6\\4dIdK\\6_4aIaK_6a4^I`Kb6c4[I^Kc6e4[I[Ke6d500000000000001O00000002N3M3M3N1N3M3M2O2M2N00100O001O10O01O1O1O010O1O00100O001O10O01O1O010O1O00100O16I10O1O001O1N1O2J6K4K6L4L3N3L4N2O2N2M4M3L4M2N3L4M3L3N3M3L4M2N4L5K5K5K5K6K4K5K5K5K5K5K5K5K1O1O1O100O1O1O1O1O1O1O1O1O1O1O3M4M3L3MZda6"}, "label": "a person working in the far left corner"}]}
{"id": 38266, "image": "COCO_train2014_000000038266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women cooking\"."}], "task": "refering", "answer_template": "The \"a women cooking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 93, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255], "bbox": [0.001484375, 0.2561792452830189, 0.20842187499999998, 0.7932547169811321], "iscrowd": 0, "area": 18424.232850000004, "rle": {"size": [424, 640], "counts": "]f07[7P4TJ^L]5f3`J^L\\5d3aJ]L_5f3^JZLb5h3\\JXLd5k3YJULg5n3UJSLk5o3SJQLm5R4PJnKP6T4nIkKS6X4iIiKW6Y4gIgKY6\\4dIdK\\6_4aIaK_6a4^I`Kb6c4[I^Kc6e4[I[Ke6d500000000000001O00000002N3M3M3N1N3M3M2O2M2N00100O001O10O01O1O1O010O1O00100O001O10O01O1O010O1O00100O16I10O1O001O1N1O2J6K4K6L4L3N3L4N2O2N2M4M3L4M2N3L4M3L3N3M3L4M2N4L5K5K5K5K6K4K5K5K5K5K5K5K5K1O1O1O100O1O1O1O1O1O1O1O1O1O1O3M4M3L3MZda6"}, "label": "a women cooking"}]}
{"id": 38266, "image": "COCO_train2014_000000038266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chef in the back with his back turned\"."}], "task": "refering", "answer_template": "The \"the chef in the back with his back turned\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 282, 283, 284, 305, 306, 307, 328, 329, 330], "bbox": [0.230765625, 0.059976415094339625, 0.43918749999999995, 0.9835849056603774], "iscrowd": 0, "area": 28929.598450000005, "rle": {"size": [424, 640], "counts": "X^m1c0^<e0[O;K1O2O0O2O0O2O0O2O0O2O0O2O2M5L3WFhN[7]1YHf2l3^MhKj2W4[MaKi2_4P3O100O010O100O100O1O1O1O1O1O1O1N2O001O1O1O1O1O1O1VOmF_MS9^2^GSMc8k2dGoL]8k2S1C=B=D=B>I71O000000001O000000000O2O00000000001O000000001O0O1UKd1jJ\\NT5i1iJWNU5n1hJRNW5R2fJoMW5U2gJkMW5Z2fJfMY5^2dJbMZ5c2cJ]M[5g2cJYM\\5k2aJVM\\5o2aJQM]5T3`JlL_5W3_JiL_5\\3]JeLa5`3ZJcLd5a3TJdLj5a3nIdLP6[5N2O1N3oLmHIU77VI]Ok6c0WIYOk6g0UIWOm6i0TIUOm6k0SITOo6k0RISOo6m0RIPOP7P1PIoNQ7Q1PImNQ7S1oHlNS7S1nHkNS7I]HC`0c0T7kN^I>@e0S9UOoFi0V9ROmFl0W9oNkFP1h:N2N1O2N1O2M4M3M3M4L3M3M3L5L3M3MT_d4"}, "label": "the chef in the back with his back turned"}]}
{"id": 38266, "image": "COCO_train2014_000000038266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef wearing a white shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"a chef wearing a white shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 282, 283, 284, 305, 306, 307, 328, 329, 330], "bbox": [0.230765625, 0.059976415094339625, 0.43918749999999995, 0.9835849056603774], "iscrowd": 0, "area": 28929.598450000005, "rle": {"size": [424, 640], "counts": "X^m1c0^<e0[O;K1O2O0O2O0O2O0O2O0O2O0O2O2M5L3WFhN[7]1YHf2l3^MhKj2W4[MaKi2_4P3O100O010O100O100O1O1O1O1O1O1O1N2O001O1O1O1O1O1O1VOmF_MS9^2^GSMc8k2dGoL]8k2S1C=B=D=B>I71O000000001O000000000O2O00000000001O000000001O0O1UKd1jJ\\NT5i1iJWNU5n1hJRNW5R2fJoMW5U2gJkMW5Z2fJfMY5^2dJbMZ5c2cJ]M[5g2cJYM\\5k2aJVM\\5o2aJQM]5T3`JlL_5W3_JiL_5\\3]JeLa5`3ZJcLd5a3TJdLj5a3nIdLP6[5N2O1N3oLmHIU77VI]Ok6c0WIYOk6g0UIWOm6i0TIUOm6k0SITOo6k0RISOo6m0RIPOP7P1PIoNQ7Q1PImNQ7S1oHlNS7S1nHkNS7I]HC`0c0T7kN^I>@e0S9UOoFi0V9ROmFl0W9oNkFP1h:N2N1O2N1O2M4M3M3M4L3M3M3L5L3M3MT_d4"}, "label": "a chef wearing a white shirt and blue jeans"}]}
{"id": 562559, "image": "COCO_train2014_000000562559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest broccoli to the camera\"."}], "task": "refering", "answer_template": "The \"the closest broccoli to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 259, 260, 261, 262, 263, 278, 279, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 321, 322, 323, 324, 325, 326, 327, 328, 329, 343, 344, 345, 346, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.41411764705882353, 0.4635294117647059, 0.9952941176470589, 1.0], "iscrowd": 0, "area": 79093.3248, "rle": {"size": [612, 612], "counts": "Thg42ob0c0]O;E7I7I7I7I7K5K3M1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1N2O1N2O1O1N2O1N2O1O0O2O1O1N2O100O100O100O100O100O100O001O1O1O1O100O1O1O1O1O1O1O1O100O001O1O1O1O1O100O100O100O10000O10O010000O10000O100O100\\MXKlFh4T9ZKiFg4V9\\KhFd4X9]KeFe4Z9^KbFd4^9^K_Fb4a9`K\\Fb4d9`KXFb4g9`KWFa4i9aKSFa4l9bKPF`4P:aKmEa4S:aKiEa4V:bKfE`4Z:aKcEa4\\:bK`E`4`:bK\\E`4d:aKZE`4e:cKWE_4i:cKSE_4m:l1O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O100000000001O000000000000001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O001O1O1O1O1O1O0000O10000O1000000O10000O10000O1000000O101O0gNoH^FR7`9QI_Fo6`9SI^Fn6`9UI^Fl6a9VI]Fk6b9WI\\Fj6b9YI\\Fh6c9ZIZFh6e9ZIYFg6e9\\IXFf6g9\\IVFf6i9\\ITFf6j9]ISFe6l9Z1N2O1N2O1[OhF\\GZ9d8lFUGU9k8PGnFR9R9`000000000000000000000O100000000000000000000000000000000000000000000001O000000001O000000001O000000001O000000001O00000000001O000000001O001O001O1O1O001O1O1O001O1O001O1O001O1O1O001O1M2J7J7I7I7I8Ic_1"}, "label": "the closest broccoli to the camera"}]}
{"id": 562559, "image": "COCO_train2014_000000562559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three single pieces of broccoli\"."}], "task": "refering", "answer_template": "The \"three single pieces of broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 132, 133, 134, 135, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 159, 160, 161, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 220, 221], "bbox": [0.0, 0.0, 0.4929575163398693, 0.49436274509803924], "iscrowd": 0, "area": 56660.128749999996, "rle": {"size": [612, 612], "counts": "d2j6X<2M3N1N3N2M2O2M2O2M3N101O10O0100O0010O0100O00100O010O1O010O010O1O010O1O0100mNgD[JZ;c5jDYJW;g5lDVJT;j5PERJP;n5SEnIo:Q6TElIl:T6WEhIj:X6YEeIg:Z6]EcId:\\6R1000001O0O1000000000O10O1000O10000O011O000O2O00001N101O00001O00001O0010O0001O00001O01N10000O10000O101O0O100mNdC^K\\<b4jCWKW<i4oCQKQ<o4TDlJm;S5YDfJh;Z5Q1000000000001O0000000000001O01O001O1O1O010O1O0N3N2N2M2O2N2M3N1O2M3N1O2M3N2N1N3N2N1N3N2N2M20000010O000010O0001O01O01O00010O0000O2M2N2O1N3M2O1N2N3N1N2N2O2O0000001O0000001O0000DlDhHT;S7SElHl:o6[EoHe:l6bERI^:i6Q1K5K5K5L4K5K5K5K500O100O10O010000O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O101N3N2M3N1N3M:G;D<E4K3M3N2M4M2M3N3L3M3N3L3N2M3N1N2N3N1N3M2M3L5K4L4M4K4L5L3L4L5K4M3Lmhi5"}, "label": "three single pieces of broccoli"}]}
{"id": 128385, "image": "COCO_train2014_000000128385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table in front of the two men in front\"."}], "task": "refering", "answer_template": "The \"the table in front of the two men in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.0, 0.8344533333333334, 1.0, 0.9898666666666668], "iscrowd": 0, "area": 20982.02664999999, "rle": {"size": [375, 640], "counts": "jd32e;0000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O0000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O0000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O00000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O00000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000001N1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000001O00000000000000001O00000000000000001O000000000000001O001O1O001O1O1O001O1O1O001O1O1O001O1O001A?^OUF"}, "label": "the table in front of the two men in front"}]}
{"id": 128385, "image": "COCO_train2014_000000128385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white table under various dining objects\"."}], "task": "refering", "answer_template": "The \"a white table under various dining objects\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.0, 0.8344533333333334, 1.0, 0.9898666666666668], "iscrowd": 0, "area": 20982.02664999999, "rle": {"size": [375, 640], "counts": "jd32e;0000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O0000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O0000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O00000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O2O00000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000001N1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000001O00000000000000001O00000000000000001O000000000000001O001O1O001O1O1O001O1O1O001O1O1O001O1O001A?^OUF"}, "label": "a white table under various dining objects"}]}
{"id": 5508, "image": "COCO_train2014_000000005508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of the chair that the man in rainbow shirt is resting his arm on\"."}], "task": "refering", "answer_template": "The \"the arm of the chair that the man in rainbow shirt is resting his arm on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 195, 196, 197, 198, 317, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5016093749999999, 0.4580625, 1.0, 0.9871041666666667], "iscrowd": 0, "area": 17688.614199999993, "rle": {"size": [480, 640], "counts": "VUg41n>1O1O1O1O1O1O1MI\\A8d>20000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000gHFL:1IO7NL24KO51I17OF4:LC7=IA:>EB<>DA=?BA??A@`0`0_O@b0`0^O_Oc0a0\\O_Oe0a0[O^Og0a0XO_Oi0a0WO^Oj0b0UO^Ol0b0TO]Om0c0RO]Oo0c0QO\\OP1d0oN\\OR1d0nN[OT1d0kN\\OV1d0jN[OW1e0hN[OY1e0gNZOZ1f0eNZO\\1f0dNYO]1U71N1O1O1O1O1O2N12N1O1O2M2O1O1O1O0000000000N2N2N1O2N2N2N2N2N2N2N2NO3K5K4M4N21O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N10000000000O10000000000O10000000000O10000000000O10O100000O10000000000O10000000000O10000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O0000000O2O0000000000000000001O0000000000000000001O000000000000000000001O0000O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O10000O10000O100\\D"}, "label": "the arm of the chair that the man in rainbow shirt is resting his arm on"}]}
{"id": 398729, "image": "COCO_train2014_000000398729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch with blue pillows\"."}], "task": "refering", "answer_template": "The \"a white couch with blue pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.06401562499999999, 0.478107476635514, 0.7138125000000001, 1.0], "iscrowd": 0, "area": 64986.84135000001, "rle": {"size": [428, 640], "counts": "Z[a02Y=5L3L5L3L5K4M4K4L5L3L5L3L5K5L3L5L3M4L3N3L3M4M2M4L3NO01O01O000101N7I7J7H6J100O00100O00100O0010O01O10O01O10O0100O0G:D<O001O00100O001O10O01O1O00100O001O1O101N1O1O1O2O0O1O1O2O0O1O1O2N10001N10000O2O000O01000O1000O0100O10000O01000O1000O010000O10O10O10000O1O0O2O1N2O1N101N2O1O0O100O100O100O100O100001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O001O1O001O1O001O1O001O1O001O1O0000000000000000000000O10000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000O100000000000oJ_HW4a7gKaHY4_7eKcH[4]7dKdH\\4\\7bKeH_4]7^KdHb4d7TK^Hl4P8100O10000O100O01O1O100O1O1O2N100O1O1O1O101N1O1000001O000000001O100O1O1O1O001O1O1O1O000001O01O00000000001O000000000O2O000O10001N100O101O0O10001N100O101N1M3M3M4K4@`0O2N101N101N2N101N101N101N1O2O0O2O0O2N2M2M9F;E;E;XOSV\\2"}, "label": "a white couch with blue pillows"}]}
{"id": 398729, "image": "COCO_train2014_000000398729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white leather couch with blue cushions\"."}], "task": "refering", "answer_template": "The \"a white leather couch with blue cushions\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.06401562499999999, 0.478107476635514, 0.7138125000000001, 1.0], "iscrowd": 0, "area": 64986.84135000001, "rle": {"size": [428, 640], "counts": "Z[a02Y=5L3L5L3L5K4M4K4L5L3L5L3L5K5L3L5L3M4L3N3L3M4M2M4L3NO01O01O000101N7I7J7H6J100O00100O00100O0010O01O10O01O10O0100O0G:D<O001O00100O001O10O01O1O00100O001O1O101N1O1O1O2O0O1O1O2O0O1O1O2N10001N10000O2O000O01000O1000O0100O10000O01000O1000O010000O10O10O10000O1O0O2O1N2O1N101N2O1O0O100O100O100O100O100001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O001O1O001O1O001O1O001O1O001O1O0000000000000000000000O10000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000O100000000000oJ_HW4a7gKaHY4_7eKcH[4]7dKdH\\4\\7bKeH_4]7^KdHb4d7TK^Hl4P8100O10000O100O01O1O100O1O1O2N100O1O1O1O101N1O1000001O000000001O100O1O1O1O001O1O1O1O000001O01O00000000001O000000000O2O000O10001N100O101O0O10001N100O101N1M3M3M4K4@`0O2N101N101N2N101N101N101N1O2O0O2O0O2N2M2M9F;E;E;XOSV\\2"}, "label": "a white leather couch with blue cushions"}]}
{"id": 398729, "image": "COCO_train2014_000000398729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey sofa\"."}], "task": "refering", "answer_template": "The \"a grey sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 189, 190], "bbox": [0.030203124999999997, 0.3210280373831776, 0.45878125000000003, 0.6006074766355141], "iscrowd": 0, "area": 16657.119450000006, "rle": {"size": [428, 640], "counts": "kS81Z=5L3L5K5K;F?@a0_O`0A2N0O100000000O100000000O1000O100000O100000000OG:E;F:F:0O10001O000O10001O000O10001O00000O101O00000O101O00001N10001O000O2O001O000O2O00001O0O10[Oe00000001N1000001O0O1000001O0O10001O000O10001O000O101O0000000O2O0000001N10001O1O1N101O1O001N2O001O1O0O2O1O1O001N2O001O1O001N2O001O1O0O2O1OO100O010O100000O010000O10O10O10000O10O10O100000O100000001N100000000O101O00000O10001O000O1000000000O01000O1000O01000000O01000O1000O02O2N1N3N2N1N3N1O2M2O2N1N3N2N1N3E:N3M2O001N10001N101O0O101N10001N10001N100O2O000O2O000O2O001N100O2O1O0O2O001N2O2M3N2N2M2O2Nh]`4"}, "label": "a grey sofa"}]}
{"id": 398729, "image": "COCO_train2014_000000398729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark gray couch in a living room\"."}], "task": "refering", "answer_template": "The \"a dark gray couch in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 189, 190], "bbox": [0.030203124999999997, 0.3210280373831776, 0.45878125000000003, 0.6006074766355141], "iscrowd": 0, "area": 16657.119450000006, "rle": {"size": [428, 640], "counts": "kS81Z=5L3L5K5K;F?@a0_O`0A2N0O100000000O100000000O1000O100000O100000000OG:E;F:F:0O10001O000O10001O000O10001O00000O101O00000O101O00001N10001O000O2O001O000O2O00001O0O10[Oe00000001N1000001O0O1000001O0O10001O000O10001O000O101O0000000O2O0000001N10001O1O1N101O1O001N2O001O1O0O2O1O1O001N2O001O1O001N2O001O1O0O2O1OO100O010O100000O010000O10O10O10000O10O10O100000O100000001N100000000O101O00000O10001O000O1000000000O01000O1000O01000000O01000O1000O02O2N1N3N2N1N3N1O2M2O2N1N3N2N1N3E:N3M2O001N10001N101O0O101N10001N10001N100O2O000O2O000O2O001N100O2O1O0O2O001N2O2M3N2N2M2O2Nh]`4"}, "label": "a dark gray couch in a living room"}]}
{"id": 103818, "image": "COCO_train2014_000000103818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a plain flannel shirt\"."}], "task": "refering", "answer_template": "The \"a person wearing a plain flannel shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 153, 154, 168, 169, 170, 183, 184, 185, 197, 198, 199, 200, 212, 213, 214, 215, 227, 228, 229, 230], "bbox": [0.1871896955503513, 0.42331250000000004, 0.40187353629976585, 0.69825], "iscrowd": 0, "area": 10279.9369, "rle": {"size": [640, 427], "counts": "g\\b1`0Zb0Af^Oh0Qa0\\Ol^Ok0k`0[O`^OK2S1Ua0VOh^O_1o`0gNP_O]1g`0hNX_O]1``0hN^_O\\1]`0fNb_O^1[`0cNd_O^1[`0S101O1O100O1O1O001O100O1O1O00O1000iNWLXBi3h=aLnA`3R>fLgA[3Y>iLbAX3]>nL[AU3e>S100O10000O101O1O001N101O001O00000O1000O102N1O2N1O2N2N1O2Me0[O2L4K5E;[Of0_O`0N2N2M3N1O2N1N3N1O2N1O2N2N2N2N2N2N3N5J6I6K6J5K7H8IiPo4"}, "label": "a person wearing a plain flannel shirt"}]}
{"id": 103818, "image": "COCO_train2014_000000103818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"reflection in glass of the back of a man in a plaid shirt\"."}], "task": "refering", "answer_template": "The \"reflection in glass of the back of a man in a plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 153, 154, 168, 169, 170, 183, 184, 185, 197, 198, 199, 200, 212, 213, 214, 215, 227, 228, 229, 230], "bbox": [0.1871896955503513, 0.42331250000000004, 0.40187353629976585, 0.69825], "iscrowd": 0, "area": 10279.9369, "rle": {"size": [640, 427], "counts": "g\\b1`0Zb0Af^Oh0Qa0\\Ol^Ok0k`0[O`^OK2S1Ua0VOh^O_1o`0gNP_O]1g`0hNX_O]1``0hN^_O\\1]`0fNb_O^1[`0cNd_O^1[`0S101O1O100O1O1O001O100O1O1O00O1000iNWLXBi3h=aLnA`3R>fLgA[3Y>iLbAX3]>nL[AU3e>S100O10000O101O1O001N101O001O00000O1000O102N1O2N1O2N2N1O2Me0[O2L4K5E;[Of0_O`0N2N2M3N1O2N1N3N1O2N1O2N2N2N2N2N2N3N5J6I6K6J5K7H8IiPo4"}, "label": "reflection in glass of the back of a man in a plaid shirt"}]}
{"id": 398733, "image": "COCO_train2014_000000398733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman eating a hot dog\"."}], "task": "refering", "answer_template": "The \"a woman eating a hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 295, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.15596875, 0.033770491803278686, 0.8872656249999998, 0.9792271662763465], "iscrowd": 0, "area": 94645.76640000002, "rle": {"size": [427, 640], "counts": "lQZ14U=5J6J6J6J6K5J5K6J6J6K5J6J6J6J6K4K3M3M3M3kJSMJo2HeMO^2DUN3n1_OfN8]1[OVO<m0VOGb0;PO9g0JlNi0k0ZOgNZ1P1iNlNa1k0bNUOd1c0^N]Oh1:[NFk11XNNn1JUN6Q2ARN?T2YOnMg0X2POkMP1[2gNhMX1V700000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O11O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O0]JaNZO`1>hNAX1;mNDT1:mNFT19mNFS19oNFR19nNGR18POGQ18POGP18QOHP16QOJo06POKQ13oNNQ12nNOS1OmN2S1MnN3S1LlN5T1JlN7U1HjN9W1EiN<W1DiNPMAo0g1o1hNRMKf0]1W2hNSM5=T1_2fNSM`06j0f2fNTMj0Ma0n2dNUMR1G:S3eNTMT1H8R3dNVMU1H7R3cNVMX1G6Q3bNWMZ1H4Q3aNWM]1G3P3`NYM^1G3n2`NYM`1H0o2_NYMb1H0m2^N[Md1GNn2]N[Mf1GNl2\\N\\Mi1GKm2\\N[Mj1HMi2YN_Ml1GNf2VNbMn1HNd2SNdMP2H0`2PNhMR2G1^2lMjMU2H2Z2jMmMV2H2X2hMPNW2H4U2dMRN[2H4R2aMVN\\2H5P2^MXN_2G6n1ZMZNb2H7k1WM\\Nd2H8j1RMaMC3T3b09h1oLcME1V3c09f1kLeMH1U3d0;c1gLhMJ0U3e0<a1dLjMLOV3e0=`1`LkMNNW3g0>]1^LlMNOX3g0>\\1]LlMN0X3h0`0Y1[LmMN0Z3i0`0X1XLlM02Y3j0b0U1VLmM03Z3j0b0T1ULmM03[3l0c0Q1SLnM04\\3l0d0P1PLnM15\\3m0e0n0oKnM15^3n0e0l0mKoM16^3o0g0i0^LWNm2o0h0h0ZLWNP3Q1h0f0XLXNQ3R1j0c0ULZNS3R1k0a0RL[NU3T1k0`0:_OI>7BL;5DM:3F080G36MJ63KL72IN:OG0:0E0;0FO;0E0<OD1<0DO=0C0=0DO=0C0>0BO>1BO?0BO>1BO?1TLWOT1g0i21SL]Oo0b0n21TL@k0>R31SLFf09W32SLHb06\\31RLN>0`32RL39Kf31RL75Gj32PL<1Bo32QL`0L]OT42PLe0HYOX42PLj0DSO]42PLn0dNK]5WOnK[3R4eLoKZ3R4eLnK[3R4eLnK[3S4eLmKZ3T4eLlK[3T4eLlK[3U4dLlK[3T4fLkKZ3V4eLkKZ3U4fLkKZ3V4eLjK[3W4eLiKZ3W4fLiKZ3X4eLiKZ3W4fLiKZ3X4fLgKZ3Z4eLgKZ3Y4fLgKZ3Z4eLgKZ3Y4gLfKY3[4fLeKZ3\\4eLeKZ3[4fLeKZ3\\4eLeKZ3[4gLdKY3]4fLcKZ3]4fLdKY3]4fLcKZ3^4fLbKY3^4gLbKY3_4fLaKZ3_4fLbKY3_4gLaKX3`4gLbKW3^4iLcKV3^4iLcKV3]4kLcKT3^4kLdKS3]4lLdKS3\\4mLeKR3\\4mLfKQ3Z4nLhKQ3Y4nLhKQ3Y4nLiKP3W4oLkKP3V4oLkKP3U4PMmKn2T4PMnKo2R4QMoKn2Y7N1O1O2N1O1O1O2N1O1O2N1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O1O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001OTNGfF9\\9K_F4c9OXF1j92RFMP:6kEJX:8dEG^:=\\ECf:`0VE_Ol:d0oD\\OS;g0iDXOY;l0aDTOa;o0[DPOg;Y12N2N3M2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2Ngcm0"}, "label": "a woman eating a hot dog"}]}
{"id": 398733, "image": "COCO_train2014_000000398733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl eating hotdog\"."}], "task": "refering", "answer_template": "The \"girl eating hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 27, 28, 29, 30, 31, 32, 33, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 266, 267, 268, 279, 280, 281, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 295, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.15596875, 0.033770491803278686, 0.8872656249999998, 0.9792271662763465], "iscrowd": 0, "area": 94645.76640000002, "rle": {"size": [427, 640], "counts": "lQZ14U=5J6J6J6J6K5J5K6J6J6K5J6J6J6J6K4K3M3M3M3kJSMJo2HeMO^2DUN3n1_OfN8]1[OVO<m0VOGb0;PO9g0JlNi0k0ZOgNZ1P1iNlNa1k0bNUOd1c0^N]Oh1:[NFk11XNNn1JUN6Q2ARN?T2YOnMg0X2POkMP1[2gNhMX1V700000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O11O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O0]JaNZO`1>hNAX1;mNDT1:mNFT19mNFS19oNFR19nNGR18POGQ18POGP18QOHP16QOJo06POKQ13oNNQ12nNOS1OmN2S1MnN3S1LlN5T1JlN7U1HjN9W1EiN<W1DiNPMAo0g1o1hNRMKf0]1W2hNSM5=T1_2fNSM`06j0f2fNTMj0Ma0n2dNUMR1G:S3eNTMT1H8R3dNVMU1H7R3cNVMX1G6Q3bNWMZ1H4Q3aNWM]1G3P3`NYM^1G3n2`NYM`1H0o2_NYMb1H0m2^N[Md1GNn2]N[Mf1GNl2\\N\\Mi1GKm2\\N[Mj1HMi2YN_Ml1GNf2VNbMn1HNd2SNdMP2H0`2PNhMR2G1^2lMjMU2H2Z2jMmMV2H2X2hMPNW2H4U2dMRN[2H4R2aMVN\\2H5P2^MXN_2G6n1ZMZNb2H7k1WM\\Nd2H8j1RMaMC3T3b09h1oLcME1V3c09f1kLeMH1U3d0;c1gLhMJ0U3e0<a1dLjMLOV3e0=`1`LkMNNW3g0>]1^LlMNOX3g0>\\1]LlMN0X3h0`0Y1[LmMN0Z3i0`0X1XLlM02Y3j0b0U1VLmM03Z3j0b0T1ULmM03[3l0c0Q1SLnM04\\3l0d0P1PLnM15\\3m0e0n0oKnM15^3n0e0l0mKoM16^3o0g0i0^LWNm2o0h0h0ZLWNP3Q1h0f0XLXNQ3R1j0c0ULZNS3R1k0a0RL[NU3T1k0`0:_OI>7BL;5DM:3F080G36MJ63KL72IN:OG0:0E0;0FO;0E0<OD1<0DO=0C0=0DO=0C0>0BO>1BO?0BO>1BO?1TLWOT1g0i21SL]Oo0b0n21TL@k0>R31SLFf09W32SLHb06\\31RLN>0`32RL39Kf31RL75Gj32PL<1Bo32QL`0L]OT42PLe0HYOX42PLj0DSO]42PLn0dNK]5WOnK[3R4eLoKZ3R4eLnK[3R4eLnK[3S4eLmKZ3T4eLlK[3T4eLlK[3U4dLlK[3T4fLkKZ3V4eLkKZ3U4fLkKZ3V4eLjK[3W4eLiKZ3W4fLiKZ3X4eLiKZ3W4fLiKZ3X4fLgKZ3Z4eLgKZ3Y4fLgKZ3Z4eLgKZ3Y4gLfKY3[4fLeKZ3\\4eLeKZ3[4fLeKZ3\\4eLeKZ3[4gLdKY3]4fLcKZ3]4fLdKY3]4fLcKZ3^4fLbKY3^4gLbKY3_4fLaKZ3_4fLbKY3_4gLaKX3`4gLbKW3^4iLcKV3^4iLcKV3]4kLcKT3^4kLdKS3]4lLdKS3\\4mLeKR3\\4mLfKQ3Z4nLhKQ3Y4nLhKQ3Y4nLiKP3W4oLkKP3V4oLkKP3U4PMmKn2T4PMnKo2R4QMoKn2Y7N1O1O2N1O1O1O2N1O1O2N1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O1O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001OTNGfF9\\9K_F4c9OXF1j92RFMP:6kEJX:8dEG^:=\\ECf:`0VE_Ol:d0oD\\OS;g0iDXOY;l0aDTOa;o0[DPOg;Y12N2N3M2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2Ngcm0"}, "label": "girl eating hotdog"}]}
{"id": 267661, "image": "COCO_train2014_000000267661.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black keyboard in front of a man in white shirt\"."}], "task": "refering", "answer_template": "The \"a black keyboard in front of a man in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 275, 276, 277, 278, 279, 296, 297, 298, 299, 317], "bbox": [0.7127433628318584, 0.46517187499999996, 0.99529203539823, 0.662921875], "iscrowd": 0, "area": 10533.062549999997, "rle": {"size": [640, 565], "counts": "Pjl72nc01O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1000000O10000000001O0O10000000O1N2O1O0O2O1N2O1N2O1O0O2O1N2O1N2O1O1N101N2O1O1N2O1N101N2O1O2M2O1N2O1O2M2O1N2O1N3N1O1N2O1N3N1N2O1O1N2O2M2O1O1N2O2M2O1N2O1O2M2O1NX`1"}, "label": "a black keyboard in front of a man in white shirt"}]}
{"id": 267661, "image": "COCO_train2014_000000267661.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer keyboard in front of the person is black in color\"."}], "task": "refering", "answer_template": "The \"the computer keyboard in front of the person is black in color\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 275, 276, 277, 278, 279, 296, 297, 298, 299, 317], "bbox": [0.7127433628318584, 0.46517187499999996, 0.99529203539823, 0.662921875], "iscrowd": 0, "area": 10533.062549999997, "rle": {"size": [640, 565], "counts": "Pjl72nc01O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N2O2N1000000O10000000001O0O10000000O1N2O1O0O2O1N2O1N2O1O0O2O1N2O1N2O1O1N101N2O1O1N2O1N101N2O1O2M2O1N2O1O2M2O1N2O1N3N1O1N2O1N3N1N2O1O1N2O2M2O1O1N2O2M2O1N2O1O2M2O1NX`1"}, "label": "the computer keyboard in front of the person is black in color"}]}
{"id": 406933, "image": "COCO_train2014_000000406933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red double decker bus on a road\"."}], "task": "refering", "answer_template": "The \"a red double decker bus on a road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 27, 28, 29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 400, 401], "bbox": [0.103859375, 0.026288032454361055, 0.9278437500000001, 0.9793711967545639], "iscrowd": 0, "area": 184831.9951, "rle": {"size": [493, 640], "counts": "\\UP14m=^1aN^1bLdN]Gi2l7h2K5J6J6K5J6J6K6I6J6J6K5J6J6K5J6J7J5J6J6K5J6L4N2O1N3M2O0O1O1O101N1O100O1O2N100O1O1O101N1O100O1O2N100O1O101N1O1O100O2N1O100O1O2O0O1O1O101N1O100O1O1O2O0O1O1O101N1O100O1O2N100O1O101N1O1O010O1O100000O01000000O01000O1000O10O1000000O0100000O10O10O1000000O0100000O10O1000O1000O01000000O1000000O1000000O10000O1000000O10000001O00001O001O00001O000100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100O00100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100OO1O2N100O2N101N1O1O2O0O1O2O0O1O2O0O1O2N100O2N100O2N1O101N1O101N1O101N1O1O2O0O1O2O0O1O2O0O1O2N100O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N101O0O101O0O101O0O101O0O101O0O101O0O101O000O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N10001N10001N10001N10001N10001N10001N10001O0O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N10001N10001N10001N10001N10001N10001N10001O0O101O0O101O0O101O0O101O0O101O0O101O0O101O001N10001N10001N10001N10001N10001N10001N10001O0O101O0O101O0O101O0O101O0O101O0O101O0O101O000O2O000O2O000O2O000O2O000O2O000O2O000O2O001O0O101O0O3N3M4K4M3lKUH`0o7[OmHHX74eI`Nn6]1mIXMg6c2U3L4M3L5K4M3L5K4M3L4L5K4M3L5K4M3L4L5L3L4L5L3Lcme0"}, "label": "a red double decker bus on a road"}]}
{"id": 406933, "image": "COCO_train2014_000000406933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red double - decker bus with a lit interior\"."}], "task": "refering", "answer_template": "The \"red double - decker bus with a lit interior\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 27, 28, 29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 400, 401], "bbox": [0.103859375, 0.026288032454361055, 0.9278437500000001, 0.9793711967545639], "iscrowd": 0, "area": 184831.9951, "rle": {"size": [493, 640], "counts": "\\UP14m=^1aN^1bLdN]Gi2l7h2K5J6J6K5J6J6K6I6J6J6K5J6J6K5J6J7J5J6J6K5J6L4N2O1N3M2O0O1O1O101N1O100O1O2N100O1O1O101N1O100O1O2N100O1O101N1O1O100O2N1O100O1O2O0O1O1O101N1O100O1O1O2O0O1O1O101N1O100O1O2N100O1O101N1O1O010O1O100000O01000000O01000O1000O10O1000000O0100000O10O10O1000000O0100000O10O1000O1000O01000000O1000000O1000000O10000O1000000O10000001O00001O001O00001O000100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100O00100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100OO1O2N100O2N101N1O1O2O0O1O2O0O1O2O0O1O2N100O2N100O2N1O101N1O101N1O101N1O1O2O0O1O2O0O1O2O0O1O2N100O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N101O0O101O0O101O0O101O0O101O0O101O0O101O000O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N10001N10001N10001N10001N10001N10001N10001O0O2O000O2O000O2O000O2O000O2O000O2O000O2O00001N10001N10001N10001N10001N10001N10001N10001O0O101O0O101O0O101O0O101O0O101O0O101O0O101O001N10001N10001N10001N10001N10001N10001N10001O0O101O0O101O0O101O0O101O0O101O0O101O0O101O000O2O000O2O000O2O000O2O000O2O000O2O000O2O001O0O101O0O3N3M4K4M3lKUH`0o7[OmHHX74eI`Nn6]1mIXMg6c2U3L4M3L5K4M3L5K4M3L4L5K4M3L5K4M3L4L5L3L4L5L3Lcme0"}, "label": "red double - decker bus with a lit interior"}]}
{"id": 333207, "image": "COCO_train2014_000000333207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light blue umbrella with bright blue spots\"."}], "task": "refering", "answer_template": "The \"a light blue umbrella with bright blue spots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.801921875, 0.38932203389830505, 1.0, 0.6679661016949152], "iscrowd": 0, "area": 10294.035700000002, "rle": {"size": [413, 640], "counts": "Y`_63h<3N2N2N2N3L3N2N2N2N3L3N2N2N2N3M2M3N2N2N3M2M3N2O1O1N3N1O1O1N2O2N1N2O1O1O2M10000O2O00000O2O000O101O000O101O0O101O0O10001O0O10001O0O10001O000O101O000O101O000O101O000O2O00000O2O00000O2O0000001O0000001O0000001O0000001O00001O0000001O01O0001O0000001O000Z1fN`1aNSH"}, "label": "a light blue umbrella with bright blue spots"}]}
{"id": 333207, "image": "COCO_train2014_000000333207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue spotted umbreall with a woman sitting under it\"."}], "task": "refering", "answer_template": "The \"a blue spotted umbreall with a woman sitting under it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.801921875, 0.38932203389830505, 1.0, 0.6679661016949152], "iscrowd": 0, "area": 10294.035700000002, "rle": {"size": [413, 640], "counts": "Y`_63h<3N2N2N2N3L3N2N2N2N3L3N2N2N2N3M2M3N2N2N3M2M3N2O1O1N3N1O1O1N2O2N1N2O1O1O2M10000O2O00000O2O000O101O000O101O0O101O0O10001O0O10001O0O10001O000O101O000O101O000O101O000O2O00000O2O00000O2O0000001O0000001O0000001O0000001O00001O0000001O01O0001O0000001O000Z1fN`1aNSH"}, "label": "a blue spotted umbreall with a woman sitting under it"}]}
{"id": 333207, "image": "COCO_train2014_000000333207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue umbrella with flower designs , being held by a woman wearing a blue hujab and a white shirt\"."}], "task": "refering", "answer_template": "The \"a blue umbrella with flower designs , being held by a woman wearing a blue hujab and a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 177, 178, 179, 198], "bbox": [0.481453125, 0.26966101694915257, 0.819328125, 0.6045036319612591], "iscrowd": 0, "area": 15916.548550000001, "rle": {"size": [413, 640], "counts": "T\\l32i<6J7H7J6J7I6J6J3O0O100O10000O10000O100O10000O100O10000O10000O101N10001N101O0O2O0O101O0O2O000O2O0O2O001N10001N101N100O2N1O2N1O1O2O0O2N1O2N1O101O00000000000000001O001O001O0O2]FWMP9l3_O00O100O1O100O1M3M3L4M3L4M3K5F:M3N2M3M201N2O1O2M2O1N2O1N2N200000000000000000001O00000001O1O001O001O1O001O1O2O1N2N2N2N2N2N3M2N2N000002O0O1O101N1O1O1O2N1O1O1O2N1O1O1O010O1O1O1O001O1O1O001O1O1O001O1O1N1O2N2N2O0O2N2N1O2N2M2N3L4M3M2N3M3N1O2N2N1O2N2N3M3M3M3M3M3Mfc^1"}, "label": "a blue umbrella with flower designs , being held by a woman wearing a blue hujab and a white shirt"}]}
{"id": 333207, "image": "COCO_train2014_000000333207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green flower umbrella\"."}], "task": "refering", "answer_template": "The \"green flower umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 177, 178, 179, 198], "bbox": [0.481453125, 0.26966101694915257, 0.819328125, 0.6045036319612591], "iscrowd": 0, "area": 15916.548550000001, "rle": {"size": [413, 640], "counts": "T\\l32i<6J7H7J6J7I6J6J3O0O100O10000O10000O100O10000O100O10000O10000O101N10001N101O0O2O0O101O0O2O000O2O0O2O001N10001N101N100O2N1O2N1O1O2O0O2N1O2N1O101O00000000000000001O001O001O0O2]FWMP9l3_O00O100O1O100O1M3M3L4M3L4M3K5F:M3N2M3M201N2O1O2M2O1N2O1N2N200000000000000000001O00000001O1O001O001O1O001O1O2O1N2N2N2N2N2N3M2N2N000002O0O1O101N1O1O1O2N1O1O1O2N1O1O1O010O1O1O1O001O1O1O001O1O1O001O1O1N1O2N2N2O0O2N2N1O2N2M2N3L4M3M2N3M3N1O2N2N1O2N2N3M3M3M3M3M3Mfc^1"}, "label": "green flower umbrella"}]}
{"id": 333207, "image": "COCO_train2014_000000333207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a umbrella\"."}], "task": "refering", "answer_template": "The \"a umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280], "bbox": [0.00290625, 0.3803389830508475, 0.308875, 0.8275302663438256], "iscrowd": 0, "area": 29170.006149999997, "rle": {"size": [413, 640], "counts": "TQ1m1V:k0N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O2M2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N1O10O01O1O1O100O1O1O10O01O1O1O100O1O1000O10000000000000000O10000000000000000O10000000000000000O1000000000000000O010000000000000000O10000000000000000O10000000000000000O10000000000000000O10001O1O001O1O1O0O2O1O001O1O0O2M3M3M2N3N2M2N3M3M2O2M3M3M2N3M3N1N3M3M3M2N3N2M2N3M3M2O2M3M3M2O2_NhE9[:BlE9V:CRF8P:BXF9k9B\\F:f9AbF9`9CgF8\\9BlF:P;Jd]b5"}, "label": "a umbrella"}]}
{"id": 333207, "image": "COCO_train2014_000000333207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple umbrella\"."}], "task": "refering", "answer_template": "The \"purple umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280], "bbox": [0.00290625, 0.3803389830508475, 0.308875, 0.8275302663438256], "iscrowd": 0, "area": 29170.006149999997, "rle": {"size": [413, 640], "counts": "TQ1m1V:k0N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O2M2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N1O10O01O1O1O100O1O1O10O01O1O1O100O1O1000O10000000000000000O10000000000000000O10000000000000000O1000000000000000O010000000000000000O10000000000000000O10000000000000000O10000000000000000O10001O1O001O1O1O0O2O1O001O1O0O2M3M3M2N3N2M2N3M3M2O2M3M3M2N3M3N1N3M3M3M2N3N2M2N3M3M2O2M3M3M2O2_NhE9[:BlE9V:CRF8P:BXF9k9B\\F:f9AbF9`9CgF8\\9BlF:P;Jd]b5"}, "label": "purple umbrella"}]}
{"id": 374171, "image": "COCO_train2014_000000374171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stop sign above a pedestrian crossing sign\"."}], "task": "refering", "answer_template": "The \"a stop sign above a pedestrian crossing sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 20, 21, 22, 23, 24, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132], "bbox": [0.07036363636363636, 0.03302, 0.41084848484848485, 0.43301999999999996], "iscrowd": 0, "area": 25902.0609, "rle": {"size": [500, 495], "counts": "SUa0;X??A`0@`0@?A?A2N2N2N2N3M2N2N2N2N2O2M2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2N2N2N2N2N3M2O1N2N2N3M2N2N2N2N4L6J5K1O1O1O1O1O101O0001O0000001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O01O01O0000001O1M3N3M2N3M2N3L3N3M2N3M2M4M2N3M2N2N3L3N3M2N3M2M4M2N3M2N3M2M4M2N3M2N2M4M2N3M2N3M2M4M2N__^4"}, "label": "a stop sign above a pedestrian crossing sign"}]}
{"id": 374171, "image": "COCO_train2014_000000374171.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stop sign with a tomato on it , directly above another sign that says pedestrian crossing\"."}], "task": "refering", "answer_template": "The \"a stop sign with a tomato on it , directly above another sign that says pedestrian crossing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 20, 21, 22, 23, 24, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 97, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132], "bbox": [0.07036363636363636, 0.03302, 0.41084848484848485, 0.43301999999999996], "iscrowd": 0, "area": 25902.0609, "rle": {"size": [500, 495], "counts": "SUa0;X??A`0@`0@?A?A2N2N2N2N3M2N2N2N2N2O2M2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2N2N2N2N2N3M2O1N2N2N3M2N2N2N2N4L6J5K1O1O1O1O1O101O0001O0000001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O01O01O0000001O1M3N3M2N3M2N3L3N3M2N3M2M4M2N3M2N2N3L3N3M2N3M2M4M2N3M2N3M2M4M2N3M2N2M4M2N3M2N3M2M4M2N__^4"}, "label": "a stop sign with a tomato on it , directly above another sign that says pedestrian crossing"}]}
{"id": 112029, "image": "COCO_train2014_000000112029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl laying in a bed next to a bear\"."}], "task": "refering", "answer_template": "The \"a girl laying in a bed next to a bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 41, 42, 43, 48, 49, 63, 64, 65, 66, 67, 71, 72, 73, 78, 85, 86, 87, 88, 89, 90, 93, 94, 95, 96, 100, 101, 108, 109, 110, 111, 112, 113, 116, 117, 118, 119, 120, 121, 122, 123, 124, 131, 132, 133, 134, 135, 136, 139, 140, 141, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 175, 176, 177, 178, 179, 180, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 220, 221, 222, 223, 224, 225, 226, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 295, 296, 297, 298], "bbox": [0.0050625, 0.06029166666666667, 1.0, 0.7524375], "iscrowd": 0, "area": 89865.37949999998, "rle": {"size": [480, 640], "counts": "hg1a0m<b1M3N2N2M3N1N3N2N2M3N2M3N2N1N3N2M3N2N2M3N2M2O2N2M3N2M3SMSLeJo3Z5SLcJo3\\5SL`Jo3a5RL\\JP4c5RLYJQ4f5QLWJQ4h5QLUJQ4k5PLQJS4n5oKoIS4P6nKmIU4R6mKkIT4V6mKgIU4X6mKdIV4[6lKbIV4]6lK_IW4a6jK\\IX4c6jKZIX4e6jKWIY4h6iKUIX4l6iKPIZ4o6hKnHZ4Q7hKlHZ4S7gKjH\\4V7eKgH]4X7eKdH^4[7dKbH^4\\7eKbH[4]7hKaHY4^7e1N2O1N2N2O1N1O2O1N2O1N2N2O1N1O2O1[NoGYLS8f3UHRLl7m3[HkKg7T4`HdKb7\\4eH\\K\\7c4kHUKV7k4`1O1001O1O1O1O1N2_LnJRMS5j2oJWMR5d2QK]Mo4`2TK`Mm4[2VKfMk4V2WKkMj4Q2YKoMi4k1ZKUNi4f1YK[Nj4^1YKcNi4X1ZKhNh4R1[KoNh4k0ZKVOh4e0[K[Og4?\\KBg48[KHh42\\KNf4M\\K4g4E\\K<f4_O]Ka0e4YO^Kh0e4RO]Ko0e4lN]KU1e4eN_KZ1e4`N]Ka1e4]NZKd1h4ZNXKf1k4WNTKj1n4SNRKn1P5PNPKP2S5mMlJT2V5jMiJa0QN[OX72gJb0RN\\OZ7NdJe0SN^O[7KaJg0UN^O\\7I^Ji0VN_O_7D\\Jk0WNA_7BYJm0YNA`7_OWJQ1XNAd7ZOUJT1XNBe7XORJV1YNCh7SOoI[1YNBj7POnI]1YNBm7nNiI`1ZNCo7jNgId1ZNBR8gNdIf1[NCS8dNbIi1[NDV8_N_In1[NCX8]N\\IP2\\NDV9YNcGR2XOES9fNYGe1EEP9SOoFY10Eo8d1RG\\Nn8c1SG]Nm8c1RG^Nn8a1SG_Nn8_1RGbNn8^1RGbNo8\\1RGdNn8[1RGfNo8Y1QGgNo8X1QGiNP9U1QGkNP9T1PGlNP9S1PGnNQ9P1PGPOP9P1PGPOP9o0PGQOP9o0QGQOo8o0PGROP9m0QGSOn8m0SGSOm8m0RGTOm8l0TGTOl8k0UGUOj8k0VGVOj8j0VGVOj8i0VGXOi8h0XGXOh8h0XGXOg8h0YGYOg8f0ZGZOf8f0YG[Of8e0[G[Oe8d0\\G\\Oc8e0\\G\\Od8c0]G]Oc8b0]G_Ob8b0]G_Oc8`0]GAb8?^GBb8>]GBd8=\\GDc8<]GEc8;[GGd89\\GHd87\\GJd86[GKd85\\GLd83\\GNc83\\GNd81]GOc80]G1b8O^G2b8N^G2a8N_G3a8L`G4`8L_G5`8K`G6`8J`G6_8K`G6`8J_G7a8I_G7`8J_G7a8I_G7`8J_G6b8J]G7c8I]G7b8J]G7c8I\\G8c8I]G7c8I\\G8c8I]G7c8I\\G8c8I\\G8c8I]G7b8J]G7b8J^G6a8K^G6b8J^G6a8K_G5`8L`G4_8MaG2_8OaG0_80bGO_81aGN_83aGL_85`GJa87_GHa89_GFa8;_GDa8=_GBb8>^G@c8a0]G^Oc86\\G[M1^2c84aG]ML^2c82gG^MF_2c80lG`M@_2f8MoGcM[O_2f8LTHcMVOa2e8IZHeMQOa2e8GcH8]7EgH:Y7CkH=U7_OoH`0Q7]OSIb0m6[OWId0h6ZO[Ig0c6WOaIh0_6UOeIj0Z6TOjIk0V6ROnIn0P6POTJo0l5oNWJP1h5nN\\JQ1c5mNaJS1_5iNeJV1[5gNhJY1Y5cNkJ\\1U5aNoJ_1Q5]NTKa1l4\\N[K`1f4\\N`Ka1`4\\NgKa1Y4[NmKb1S4[NTLa1h8M4M2M4L3M4M2M4L3M4M2M4L3M4M2Mg]]17mabN8M3M3N2M3N1N2N7J0O101N100O100O2O0O1O1O2N1O1O1O2N1O1O100O2N1O1O2N1O2N2N1O2N2N1O2O1N1O2N2N1O2N2N1O2O1N1O2N2N1O2N1O2N2N1O2N2M2O2N2N1O2N2M2O2N2N1M4L4L3L5L4K4M4L4K4M4K5J5K6J6J5L5L4L3M4L4L3N1N2N2N2N3M2N2N2N2N2N3M2N2N2N21O1O010O1O1O1O001O100O1O001O12N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O1N3N1O1OiLbJlL^5e2RKZMn4]2\\KbMd4Z2bKdM^4X2gKhMX4T2mKkMS4Q2SLmMm3o1XLPNg3P2\\LnMd3R2]LmMc3R2_LmMa3R2bLlM^3S2dLlM\\3T2eLkM[3T2hLkMX3S2jLlMV3S2mLkMS3U2nLjMR3U2PMjMP3U2SMiMm2V2UMiMk2W2WMgMi2Y2XMfMh2Z2YMeMg2[2[McMe2]2\\McMc2]2^MbMc2]2_MaMa2_2`M`M`2`2bM^M^2b2cM]M]2c2dM\\M\\2d2fMZMZ2f2gMYMY2g2hMXMX2h2jMVMV2j2kMUMV2j2lMTMT2l2nMRMR2n2nMRMS2m2mMTMR2l2mMUMT2j2kMWMU2i2jMXMV2i2iMWMX2h2gMYMY2g2fMZM[2e2dM\\M\\2e2cM[M]2e2bM\\M_2c2`M^M`2b2_M_Mb2a2]M_Mc2`2]MaMc2_2\\MbMe2\\2[MeMf2Z2ZMgMi2T2WMmMl2o1TMRNo2k1PMVNS3f1nLZNU3c1jL^NY3^1gLcN\\3Z1cLgN`3U1aLkNd3o0\\LROj3h0ULYOP4a0PL@U4ZOlLf0Y3hNSMY1Y81O2O0O1O100O1O1O100O1O1O100O1O1O101N1O1O100O1O100O1O1O100O1O100O2NM3kNV1jNWI"}, "label": "a girl laying in a bed next to a bear"}]}
{"id": 112029, "image": "COCO_train2014_000000112029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl wearing shoes and lying on bed\"."}], "task": "refering", "answer_template": "The \"girl wearing shoes and lying on bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 41, 42, 43, 48, 49, 63, 64, 65, 66, 67, 71, 72, 73, 78, 85, 86, 87, 88, 89, 90, 93, 94, 95, 96, 100, 101, 108, 109, 110, 111, 112, 113, 116, 117, 118, 119, 120, 121, 122, 123, 124, 131, 132, 133, 134, 135, 136, 139, 140, 141, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 175, 176, 177, 178, 179, 180, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 220, 221, 222, 223, 224, 225, 226, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 295, 296, 297, 298], "bbox": [0.0050625, 0.06029166666666667, 1.0, 0.7524375], "iscrowd": 0, "area": 89865.37949999998, "rle": {"size": [480, 640], "counts": "hg1a0m<b1M3N2N2M3N1N3N2N2M3N2M3N2N1N3N2M3N2N2M3N2M2O2N2M3N2M3SMSLeJo3Z5SLcJo3\\5SL`Jo3a5RL\\JP4c5RLYJQ4f5QLWJQ4h5QLUJQ4k5PLQJS4n5oKoIS4P6nKmIU4R6mKkIT4V6mKgIU4X6mKdIV4[6lKbIV4]6lK_IW4a6jK\\IX4c6jKZIX4e6jKWIY4h6iKUIX4l6iKPIZ4o6hKnHZ4Q7hKlHZ4S7gKjH\\4V7eKgH]4X7eKdH^4[7dKbH^4\\7eKbH[4]7hKaHY4^7e1N2O1N2N2O1N1O2O1N2O1N2N2O1N1O2O1[NoGYLS8f3UHRLl7m3[HkKg7T4`HdKb7\\4eH\\K\\7c4kHUKV7k4`1O1001O1O1O1O1N2_LnJRMS5j2oJWMR5d2QK]Mo4`2TK`Mm4[2VKfMk4V2WKkMj4Q2YKoMi4k1ZKUNi4f1YK[Nj4^1YKcNi4X1ZKhNh4R1[KoNh4k0ZKVOh4e0[K[Og4?\\KBg48[KHh42\\KNf4M\\K4g4E\\K<f4_O]Ka0e4YO^Kh0e4RO]Ko0e4lN]KU1e4eN_KZ1e4`N]Ka1e4]NZKd1h4ZNXKf1k4WNTKj1n4SNRKn1P5PNPKP2S5mMlJT2V5jMiJa0QN[OX72gJb0RN\\OZ7NdJe0SN^O[7KaJg0UN^O\\7I^Ji0VN_O_7D\\Jk0WNA_7BYJm0YNA`7_OWJQ1XNAd7ZOUJT1XNBe7XORJV1YNCh7SOoI[1YNBj7POnI]1YNBm7nNiI`1ZNCo7jNgId1ZNBR8gNdIf1[NCS8dNbIi1[NDV8_N_In1[NCX8]N\\IP2\\NDV9YNcGR2XOES9fNYGe1EEP9SOoFY10Eo8d1RG\\Nn8c1SG]Nm8c1RG^Nn8a1SG_Nn8_1RGbNn8^1RGbNo8\\1RGdNn8[1RGfNo8Y1QGgNo8X1QGiNP9U1QGkNP9T1PGlNP9S1PGnNQ9P1PGPOP9P1PGPOP9o0PGQOP9o0QGQOo8o0PGROP9m0QGSOn8m0SGSOm8m0RGTOm8l0TGTOl8k0UGUOj8k0VGVOj8j0VGVOj8i0VGXOi8h0XGXOh8h0XGXOg8h0YGYOg8f0ZGZOf8f0YG[Of8e0[G[Oe8d0\\G\\Oc8e0\\G\\Od8c0]G]Oc8b0]G_Ob8b0]G_Oc8`0]GAb8?^GBb8>]GBd8=\\GDc8<]GEc8;[GGd89\\GHd87\\GJd86[GKd85\\GLd83\\GNc83\\GNd81]GOc80]G1b8O^G2b8N^G2a8N_G3a8L`G4`8L_G5`8K`G6`8J`G6_8K`G6`8J_G7a8I_G7`8J_G7a8I_G7`8J_G6b8J]G7c8I]G7b8J]G7c8I\\G8c8I]G7c8I\\G8c8I]G7c8I\\G8c8I\\G8c8I]G7b8J]G7b8J^G6a8K^G6b8J^G6a8K_G5`8L`G4_8MaG2_8OaG0_80bGO_81aGN_83aGL_85`GJa87_GHa89_GFa8;_GDa8=_GBb8>^G@c8a0]G^Oc86\\G[M1^2c84aG]ML^2c82gG^MF_2c80lG`M@_2f8MoGcM[O_2f8LTHcMVOa2e8IZHeMQOa2e8GcH8]7EgH:Y7CkH=U7_OoH`0Q7]OSIb0m6[OWId0h6ZO[Ig0c6WOaIh0_6UOeIj0Z6TOjIk0V6ROnIn0P6POTJo0l5oNWJP1h5nN\\JQ1c5mNaJS1_5iNeJV1[5gNhJY1Y5cNkJ\\1U5aNoJ_1Q5]NTKa1l4\\N[K`1f4\\N`Ka1`4\\NgKa1Y4[NmKb1S4[NTLa1h8M4M2M4L3M4M2M4L3M4M2M4L3M4M2Mg]]17mabN8M3M3N2M3N1N2N7J0O101N100O100O2O0O1O1O2N1O1O1O2N1O1O100O2N1O1O2N1O2N2N1O2N2N1O2O1N1O2N2N1O2N2N1O2O1N1O2N2N1O2N1O2N2N1O2N2M2O2N2N1O2N2M2O2N2N1M4L4L3L5L4K4M4L4K4M4K5J5K6J6J5L5L4L3M4L4L3N1N2N2N2N3M2N2N2N2N2N3M2N2N2N21O1O010O1O1O1O001O100O1O001O12N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O1N3N1O1OiLbJlL^5e2RKZMn4]2\\KbMd4Z2bKdM^4X2gKhMX4T2mKkMS4Q2SLmMm3o1XLPNg3P2\\LnMd3R2]LmMc3R2_LmMa3R2bLlM^3S2dLlM\\3T2eLkM[3T2hLkMX3S2jLlMV3S2mLkMS3U2nLjMR3U2PMjMP3U2SMiMm2V2UMiMk2W2WMgMi2Y2XMfMh2Z2YMeMg2[2[McMe2]2\\McMc2]2^MbMc2]2_MaMa2_2`M`M`2`2bM^M^2b2cM]M]2c2dM\\M\\2d2fMZMZ2f2gMYMY2g2hMXMX2h2jMVMV2j2kMUMV2j2lMTMT2l2nMRMR2n2nMRMS2m2mMTMR2l2mMUMT2j2kMWMU2i2jMXMV2i2iMWMX2h2gMYMY2g2fMZM[2e2dM\\M\\2e2cM[M]2e2bM\\M_2c2`M^M`2b2_M_Mb2a2]M_Mc2`2]MaMc2_2\\MbMe2\\2[MeMf2Z2ZMgMi2T2WMmMl2o1TMRNo2k1PMVNS3f1nLZNU3c1jL^NY3^1gLcN\\3Z1cLgN`3U1aLkNd3o0\\LROj3h0ULYOP4a0PL@U4ZOlLf0Y3hNSMY1Y81O2O0O1O100O1O1O100O1O1O100O1O1O101N1O1O100O1O100O1O1O100O1O100O2NM3kNV1jNWI"}, "label": "girl wearing shoes and lying on bed"}]}
{"id": 193953, "image": "COCO_train2014_000000193953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and silver office chair with wheels\"."}], "task": "refering", "answer_template": "The \"a black and silver office chair with wheels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261], "bbox": [0.172875, 0.40491803278688526, 0.385734375, 0.7989227166276346], "iscrowd": 0, "area": 6392.10645, "rle": {"size": [427, 640], "counts": "lb^12X=2O1N2N2VCH_<8aCI]<8bCJ\\<7cCK^:;hFJjNL]::hFKjNM]:9hFJkNN\\:8iFJkNO[:7jFJjN1[:6jFHlN3Y:5kFHlN2Z:7iFGlN3[:6iFGlN2\\:8gFFmN2\\:8gFFlN3]:7gFFlN2^:9eFEmN2^:9eFElN2`::cFCnN3_::cFCnN3_:;bFBnN3a:<`FAoN3a:=_F@PO2b:?]F_OPO3c:?\\F^OQO3c:`0[F]ORO2d:b0YF\\ORO3e:c0WFYOUO3e:e0UFXOVO3e:f0TFWOVO4f:f0SFVOWO3g:h0QFUOXO3g:i0PF@Q:a0nE_OR:b0mE^OS:b0mE]OT:c0lE]OT:c0lE\\OU:d0kE\\OU:d0kE\\OU:d0kE[OV:e0jE[OV:e0jEZOW:f0iEZOV:g0jEYOV:g0jEXOW:h0iEXOW:h0iEWOX:i0hEWOX:i0hEWOX:i0hEVOY:j0gEVOY:k0fETO[:l0eETO[:T1]ElNc:n10001O1O001O1O001OkNXEHg:8_EC`:<bEC^:=bED]:;dEF[::eEF[:9fEHY:8hEHW:7jEIV:7jEJU:5lELS:4mELS:3nENQ:2PFNo91RFOn91RF0m9OTF2k9NUF3j9LXF3h9MXF_O[O3\\:=ZF@[O3Z:=[F@[O3Z:<]F@ZO4X:;_FBYO2X:;`FCXO3W:9cFCWO3V:9dFDWO2U:9eFEVO2U:8fFFVO1T:9gFEVO2R:8iFGYOLn9<jFH\\OHk7IdIf0VOI@Ce7OeId0XOIj7CnHc0YOJh7DoHa0ZOKg7EnH?\\OMd7EoH>_OLb7FoH=@M`7GPI;AN_7GPI:BO]7HQI9CN\\7IQI4H4V7HQK8o4HRK7n4IRK7n4IRK7n4IRK7n4ISK7l4ITK7l4ITK7l4ITK7m4HTK7l4IUK6j4LVK3j4MWK3h4MYK2g4NZK1f4O[K0d41]KNc42]KNc42^KMb43_KLa44`KK`42dKM[42hKMX41kKNU40ZkS5"}, "label": "a black and silver office chair with wheels"}]}
{"id": 193953, "image": "COCO_train2014_000000193953.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair pushed into a desk\"."}], "task": "refering", "answer_template": "The \"a chair pushed into a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261], "bbox": [0.172875, 0.40491803278688526, 0.385734375, 0.7989227166276346], "iscrowd": 0, "area": 6392.10645, "rle": {"size": [427, 640], "counts": "lb^12X=2O1N2N2VCH_<8aCI]<8bCJ\\<7cCK^:;hFJjNL]::hFKjNM]:9hFJkNN\\:8iFJkNO[:7jFJjN1[:6jFHlN3Y:5kFHlN2Z:7iFGlN3[:6iFGlN2\\:8gFFmN2\\:8gFFlN3]:7gFFlN2^:9eFEmN2^:9eFElN2`::cFCnN3_::cFCnN3_:;bFBnN3a:<`FAoN3a:=_F@PO2b:?]F_OPO3c:?\\F^OQO3c:`0[F]ORO2d:b0YF\\ORO3e:c0WFYOUO3e:e0UFXOVO3e:f0TFWOVO4f:f0SFVOWO3g:h0QFUOXO3g:i0PF@Q:a0nE_OR:b0mE^OS:b0mE]OT:c0lE]OT:c0lE\\OU:d0kE\\OU:d0kE\\OU:d0kE[OV:e0jE[OV:e0jEZOW:f0iEZOV:g0jEYOV:g0jEXOW:h0iEXOW:h0iEWOX:i0hEWOX:i0hEWOX:i0hEVOY:j0gEVOY:k0fETO[:l0eETO[:T1]ElNc:n10001O1O001O1O001OkNXEHg:8_EC`:<bEC^:=bED]:;dEF[::eEF[:9fEHY:8hEHW:7jEIV:7jEJU:5lELS:4mELS:3nENQ:2PFNo91RFOn91RF0m9OTF2k9NUF3j9LXF3h9MXF_O[O3\\:=ZF@[O3Z:=[F@[O3Z:<]F@ZO4X:;_FBYO2X:;`FCXO3W:9cFCWO3V:9dFDWO2U:9eFEVO2U:8fFFVO1T:9gFEVO2R:8iFGYOLn9<jFH\\OHk7IdIf0VOI@Ce7OeId0XOIj7CnHc0YOJh7DoHa0ZOKg7EnH?\\OMd7EoH>_OLb7FoH=@M`7GPI;AN_7GPI:BO]7HQI9CN\\7IQI4H4V7HQK8o4HRK7n4IRK7n4IRK7n4IRK7n4ISK7l4ITK7l4ITK7l4ITK7m4HTK7l4IUK6j4LVK3j4MWK3h4MYK2g4NZK1f4O[K0d41]KNc42]KNc42^KMb43_KLa44`KK`42dKM[42hKMX41kKNU40ZkS5"}, "label": "a chair pushed into a desk"}]}
{"id": 87458, "image": "COCO_train2014_000000087458.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boat next to a larger boat\"."}], "task": "refering", "answer_template": "The \"a small boat next to a larger boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.38059375, 0.6900263852242744, 0.9568125000000001, 0.8832981530343007], "iscrowd": 0, "area": 16156.642549999991, "rle": {"size": [379, 640], "counts": "icj25b;4M4N2N1O2N2O0O100000001O000000000000000000001O0000000000000O1O100O2O0O1000000O100000000O2O00000O10000001O00000001O000001O000000000000010O000000000000001O0001O000000000000000000000O1000000000000O1000001O000O100O10000O100O100O10000O100O100O10000O100000000000000000000000000000000000000000000000000000000O10000000001O000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000000000000000O1000000000000000000000000000000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O10000O1000O0100O10000O100O10000O10000O100O10O10O100000000000000000000000000000O10001O001O001O001O001O001O001O0O101O001O001N101O001N101O1O001N101O1O0O2O001O001N2O001O0O2O001O001N101O1O001N101O001N101O0010O00010O0000010O00001O01O01O06JUR:"}, "label": "a small boat next to a larger boat"}]}
{"id": 87458, "image": "COCO_train2014_000000087458.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller boat\"."}], "task": "refering", "answer_template": "The \"the smaller boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.38059375, 0.6900263852242744, 0.9568125000000001, 0.8832981530343007], "iscrowd": 0, "area": 16156.642549999991, "rle": {"size": [379, 640], "counts": "icj25b;4M4N2N1O2N2O0O100000001O000000000000000000001O0000000000000O1O100O2O0O1000000O100000000O2O00000O10000001O00000001O000001O000000000000010O000000000000001O0001O000000000000000000000O1000000000000O1000001O000O100O10000O100O100O10000O100O100O10000O100000000000000000000000000000000000000000000000000000000O10000000001O000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000000000000000O1000000000000000000000000000000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O10000O1000O0100O10000O100O10000O10000O100O10O10O100000000000000000000000000000O10001O001O001O001O001O001O001O0O101O001O001N101O001N101O1O001N101O1O0O2O001O001N2O001O0O2O001O001N101O1O001N101O001N101O0010O00010O0000010O00001O01O01O06JUR:"}, "label": "the smaller boat"}]}
{"id": 87458, "image": "COCO_train2014_000000087458.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey boat sits in the harbor\"."}], "task": "refering", "answer_template": "The \"a grey boat sits in the harbor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.048343750000000005, 0.6379419525065964, 0.656640625, 0.8420316622691293], "iscrowd": 0, "area": 13036.028549999994, "rle": {"size": [379, 640], "counts": "Zg;9^;a0_Oa0_Ob0^O=D31O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O000000001O000000000000000000000O10000000001O000000000000000000000000000000001O0000000000000000000000000000001O00000000000000000O100000000000001O0000000000000000000000000000001O000000000000000000000000000000001O00000000000O1000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000O10000000000000000000001O000000000000000000000000000000001O0000000C=0000000000000000000001O00000000000O100000000000000000001O0000000000000000000000000000001O0000000000000O1000000000000000001O00000001O000000000000000000000010O00000000000O1000000O100000000O2O00000O1000000O1000000O1000000O101O000O1000000O100000000000000O101O00000000000O100000000000000O10001O00000000000O100000000000000O101O00000000000O10000000000000a`a2"}, "label": "a grey boat sits in the harbor"}]}
{"id": 87458, "image": "COCO_train2014_000000087458.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"naval vessel k377\"."}], "task": "refering", "answer_template": "The \"naval vessel k377\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.048343750000000005, 0.6379419525065964, 0.656640625, 0.8420316622691293], "iscrowd": 0, "area": 13036.028549999994, "rle": {"size": [379, 640], "counts": "Zg;9^;a0_Oa0_Ob0^O=D31O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O000000001O000000000000000000000O10000000001O000000000000000000000000000000001O0000000000000000000000000000001O00000000000000000O100000000000001O0000000000000000000000000000001O000000000000000000000000000000001O00000000000O1000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000O10000000000000000000001O000000000000000000000000000000001O0000000C=0000000000000000000001O00000000000O100000000000000000001O0000000000000000000000000000001O0000000000000O1000000000000000001O00000001O000000000000000000000010O00000000000O1000000O100000000O2O00000O1000000O1000000O1000000O101O000O1000000O100000000000000O101O00000000000O100000000000000O10001O00000000000O100000000000000O101O00000000000O10000000000000a`a2"}, "label": "naval vessel k377"}]}
{"id": 374180, "image": "COCO_train2014_000000374180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a completely visible orange wedge in the 3 o ' clock position on a plate\"."}], "task": "refering", "answer_template": "The \"a completely visible orange wedge in the 3 o ' clock position on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 334, 335, 336], "bbox": [0.543109375, 0.3219284294234593, 0.811890625, 0.8172166998011928], "iscrowd": 0, "area": 27932.804599999996, "rle": {"size": [503, 640], "counts": "nY[54`?5L4L4K5K5K5K5K5K5K4L4M3L4L4L4L4L4L4L5K3M2N3M2J7J5K6I6K6J5J7J6I6K6J5J7J5K6I6K6I6K6J5J7J5J7J6M2100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O2O1N101N2O0O2O1N101O1N101N2O0O2O1N101N2O1N101N2O001N2O2M2O1M4L3M3M3M4L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M4L7I8H9G9E:ROofj1"}, "label": "a completely visible orange wedge in the 3 o ' clock position on a plate"}]}
{"id": 374180, "image": "COCO_train2014_000000374180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half orange slice at the right side of the plate\"."}], "task": "refering", "answer_template": "The \"a half orange slice at the right side of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 334, 335, 336], "bbox": [0.543109375, 0.3219284294234593, 0.811890625, 0.8172166998011928], "iscrowd": 0, "area": 27932.804599999996, "rle": {"size": [503, 640], "counts": "nY[54`?5L4L4K5K5K5K5K5K5K4L4M3L4L4L4L4L4L4L5K3M2N3M2J7J5K6I6K6J5J7J6I6K6J5J7J5K6I6K6I6K6J5J7J5J7J6M2100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O2O1N101N2O0O2O1N101O1N101N2O0O2O1N101N2O1N101N2O001N2O2M2O1M4L3M3M3M4L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M4L7I8H9G9E:ROofj1"}, "label": "a half orange slice at the right side of the plate"}]}
{"id": 513445, "image": "COCO_train2014_000000513445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baked stacked on another on a plate\"."}], "task": "refering", "answer_template": "The \"a baked stacked on another on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.249796875, 0.5775352112676057, 0.6027968749999999, 0.9842723004694836], "iscrowd": 0, "area": 32463.344750000004, "rle": {"size": [426, 640], "counts": "alR2>Y<e0ZOf0F:M3M4L3L4M3M3M4L3L4M3M3M3M4L3L4M3M3M3M1N3O1O0O2O1N2O001N2O001N2O0O2O1O0O2O1O0O2O1N10000O100O100O100O10000O100O100O2O0O10000O100O100O100O100000000000001O00000000000000000001O00000000000001O0000000000000000000000000000000000001O000000000001O0000000000000000000001O0000000001O00001O001O001O00001O001O01O01O001O001O00001O001O00001O001O0010O0001O001O00001O001O001O001O1O1O1O100O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1N2M4L3M3M3L4K5L4L4K5L5J5L4L7H;F:F:E<E:E;F^ZY3"}, "label": "a baked stacked on another on a plate"}]}
{"id": 513445, "image": "COCO_train2014_000000513445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut or piece of bread on a plate\"."}], "task": "refering", "answer_template": "The \"a donut or piece of bread on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.249796875, 0.5775352112676057, 0.6027968749999999, 0.9842723004694836], "iscrowd": 0, "area": 32463.344750000004, "rle": {"size": [426, 640], "counts": "alR2>Y<e0ZOf0F:M3M4L3L4M3M3M4L3L4M3M3M3M4L3L4M3M3M3M1N3O1O0O2O1N2O001N2O001N2O0O2O1O0O2O1O0O2O1N10000O100O100O100O10000O100O100O2O0O10000O100O100O100O100000000000001O00000000000000000001O00000000000001O0000000000000000000000000000000000001O000000000001O0000000000000000000001O0000000001O00001O001O001O00001O001O01O01O001O001O00001O001O00001O001O0010O0001O001O00001O001O001O001O1O1O1O100O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1N2M4L3M3M3L4K5L4L4K5L5J5L4L7H;F:F:E<E:E;F^ZY3"}, "label": "a donut or piece of bread on a plate"}]}
{"id": 431526, "image": "COCO_train2014_000000431526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white airplane\"."}], "task": "refering", "answer_template": "The \"the white airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 221, 222], "bbox": [0.316546875, 0.35872916666666665, 0.685546875, 0.5426249999999999], "iscrowd": 0, "area": 5117.371950000002, "rle": {"size": [480, 640], "counts": "i[o23n>4M3L01O0O1O2N1O001O0000000001O0001O0000000000000001O00000001O0000000001O00000000000001O000001O0000000000000001O0001O000000000000000000010O000000000000000000001O01O0000000000000000RN7eDJZ;c0fC[OM2]<o0ZCSO5N`<b1\\C^Nf<b1YC]Nj<a1UC_Nn<_1QCaNR=\\1nBdNU=Z1jBfNZ=W1eBiN^=\\13L4L3M2N1O1O2NSBYOb=f0]B\\Oc=c0\\B^Oe=a0ZB@g=l01O1O0001O000O10000O100O10hN\\BT1d=lN\\BT1d=lN\\BU1c=jN^BV1f=002N0000N2GXBXOj=h070001O0001O000000001O00000000010O00000000001O000000000010O0000000001OJRBZOn=g0500010O1O2N2N0000O10001O000000000000001O00000000000001O010O00001O01O00000010O1O2OO001O001O001O1O001O0O2O1N102N1N3N2M2O2M2O2N1N3LXom2"}, "label": "the white airplane"}]}
{"id": 431526, "image": "COCO_train2014_000000431526.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plane with a red tail\"."}], "task": "refering", "answer_template": "The \"a plane with a red tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 221, 222], "bbox": [0.316546875, 0.35872916666666665, 0.685546875, 0.5426249999999999], "iscrowd": 0, "area": 5117.371950000002, "rle": {"size": [480, 640], "counts": "i[o23n>4M3L01O0O1O2N1O001O0000000001O0001O0000000000000001O00000001O0000000001O00000000000001O000001O0000000000000001O0001O000000000000000000010O000000000000000000001O01O0000000000000000RN7eDJZ;c0fC[OM2]<o0ZCSO5N`<b1\\C^Nf<b1YC]Nj<a1UC_Nn<_1QCaNR=\\1nBdNU=Z1jBfNZ=W1eBiN^=\\13L4L3M2N1O1O2NSBYOb=f0]B\\Oc=c0\\B^Oe=a0ZB@g=l01O1O0001O000O10000O100O10hN\\BT1d=lN\\BT1d=lN\\BU1c=jN^BV1f=002N0000N2GXBXOj=h070001O0001O000000001O00000000010O00000000001O000000000010O0000000001OJRBZOn=g0500010O1O2N2N0000O10001O000000000000001O00000000000001O010O00001O01O00000010O1O2OO001O001O001O1O001O0O2O1N102N1N3N2M2O2M2O2N1N3LXom2"}, "label": "a plane with a red tail"}]}
{"id": 210344, "image": "COCO_train2014_000000210344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female eatting pizza\"."}], "task": "refering", "answer_template": "The \"female eatting pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.32021875, 0.07416666666666667, 0.82753125, 0.9887708333333334], "iscrowd": 0, "area": 102975.77414999997, "rle": {"size": [480, 640], "counts": "S_P36d>=D;E9G8H8G9H6J7I7H7J7I6J7I7jMoLSHX3f7SMPHS3j7XMkGn2R8YMdGm2Z8n1M3N2N2M3N2M3N2N2M2O1N2O001N200O1O100O1O1O100O001O1O1N2O1N2O1N2N2O1N2kN`HhJa7T5oH^JT7\\5[1J6K5J6J6N2N2M3N2M2K6G9L4N2N3N1N2O2M2O1N2O2M2O101N1O1O2N1O101N1O1O2N1O100O2N1O1O2N100O2O0O101N100O1O2O0O101N100O2O0O101N1O100O2O0O101N100O2O00001N10000O2O000O2O000O2O000O2O0000001O00001O00001O00001O000000000010O0000001O1O001O001O010O001O001O1O001O001O010O001O001O1O001O001O10O01O1O1O1O1O1O001O1O100O1O2N1O1O1O1O1O1O2O1N1O2N2N2N1O2N2N3M2N4M2M2N2M2O2N1O2N1O2N3M2N2N2N3M2M3N3M2N2YGeKe6]4SIiKm6\\4hHjKX7\\4\\HjKd7\\4PHjKP8d5O00000O10001O00000000001O00000O101O00000000001O00001O2M4M2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2M3M3N2M3M3M3M3M3M3M2N3L5L3M4K4M4L4K4M4I8PLlFi1[9kMQG7]NT1h<dNbCS1^=F:G8G8I6InTc1"}, "label": "female eatting pizza"}]}
{"id": 210344, "image": "COCO_train2014_000000210344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman eating like a food like pizza\"."}], "task": "refering", "answer_template": "The \"a woman eating like a food like pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.32021875, 0.07416666666666667, 0.82753125, 0.9887708333333334], "iscrowd": 0, "area": 102975.77414999997, "rle": {"size": [480, 640], "counts": "S_P36d>=D;E9G8H8G9H6J7I7H7J7I6J7I7jMoLSHX3f7SMPHS3j7XMkGn2R8YMdGm2Z8n1M3N2N2M3N2M3N2N2M2O1N2O001N200O1O100O1O1O100O001O1O1N2O1N2O1N2N2O1N2kN`HhJa7T5oH^JT7\\5[1J6K5J6J6N2N2M3N2M2K6G9L4N2N3N1N2O2M2O1N2O2M2O101N1O1O2N1O101N1O1O2N1O100O2N1O1O2N100O2O0O101N100O1O2O0O101N100O2O0O101N1O100O2O0O101N100O2O00001N10000O2O000O2O000O2O000O2O0000001O00001O00001O00001O000000000010O0000001O1O001O001O010O001O001O1O001O001O010O001O001O1O001O001O10O01O1O1O1O1O1O001O1O100O1O2N1O1O1O1O1O1O2O1N1O2N2N2N1O2N2N3M2N4M2M2N2M2O2N1O2N1O2N3M2N2N2N3M2M3N3M2N2YGeKe6]4SIiKm6\\4hHjKX7\\4\\HjKd7\\4PHjKP8d5O00000O10001O00000000001O00000O101O00000000001O00001O2M4M2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2M3M3N2M3M3M3M3M3M3M2N3L5L3M4K4M4L4K4M4I8PLlFi1[9kMQG7]NT1h<dNbCS1^=F:G8G8I6InTc1"}, "label": "a woman eating like a food like pizza"}]}
{"id": 333225, "image": "COCO_train2014_000000333225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of broccoli furthest left on the plate that is centered\"."}], "task": "refering", "answer_template": "The \"the piece of broccoli furthest left on the plate that is centered\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234], "bbox": [0.0, 0.3406042884990253, 0.26818749999999997, 0.6137816764132553], "iscrowd": 0, "area": 15443.528049999999, "rle": {"size": [513, 640], "counts": "R6[2f=2N101N2N101N1O2N2O0O2N102M4L4L4M1N00010O000010O00010O01O00010O00001O00000O2O000O10001O0O10001N10000O101O0O1000O010000O100O10000O10O010000O100O10000O10O01000000000000000000O10000000000000O1O10O01O100O00O2O00001O01N10000O103L5L4K4M4L4K5L4K4M4K5L3L3M4M3L3M4L4L3M3O00001O0000000O10001O00000000001O000000003M3M3M3N2M3M2N3M1O1O1O1O1O001M3N2M3N1NifZ7"}, "label": "the piece of broccoli furthest left on the plate that is centered"}]}
{"id": 333225, "image": "COCO_train2014_000000333225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of broccoi sitting on the edge of a plate\"."}], "task": "refering", "answer_template": "The \"a piece of broccoi sitting on the edge of a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 73, 74, 75], "bbox": [0.0, 0.0023586744639376217, 0.299875, 0.2023586744639376], "iscrowd": 0, "area": 12964.8342, "rle": {"size": [513, 640], "counts": "j21n?3M3M2N3M3M3M3M2N3O1O1N2O0O100O10000O100O100O10000O100O100O10000O100O100O1O100O1O100O1O100O100O1N2N2N2N2N2N2N2O1O1O100O1O0010O01O0010O01O0010O0001O010O001O010N101N101N101O0O2O0O2O0O2O0010O100O100O100O100000000001O000000001O000000001O000000001O000000002N2N2N2N2N1O2N2N1O001O0000000000010O0001O001O0010O01O00001O001O001O00001O001O001O00001O1O1O2N2N2O1N1O2N2M3N2M2O2M3M3N2M2O3L5L3L5L4B^\\P7"}, "label": "a piece of broccoi sitting on the edge of a plate"}]}
{"id": 333225, "image": "COCO_train2014_000000333225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a floret of brocollie lying on the plate at the left most corner\"."}], "task": "refering", "answer_template": "The \"a floret of brocollie lying on the plate at the left most corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 73, 74, 75], "bbox": [0.0, 0.0023586744639376217, 0.299875, 0.2023586744639376], "iscrowd": 0, "area": 12964.8342, "rle": {"size": [513, 640], "counts": "j21n?3M3M2N3M3M3M3M2N3O1O1N2O0O100O10000O100O100O10000O100O100O10000O100O100O1O100O1O100O1O100O100O1N2N2N2N2N2N2N2O1O1O100O1O0010O01O0010O01O0010O0001O010O001O010N101N101N101O0O2O0O2O0O2O0010O100O100O100O100000000001O000000001O000000001O000000001O000000002N2N2N2N2N1O2N2N1O001O0000000000010O0001O001O0010O01O00001O001O001O00001O001O001O00001O1O1O2N2N2O1N1O2N2M3N2M2O2M3M3N2M2O3L5L3L5L4B^\\P7"}, "label": "a floret of brocollie lying on the plate at the left most corner"}]}
{"id": 546218, "image": "COCO_train2014_000000546218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table cloth\"."}], "task": "refering", "answer_template": "The \"a table cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 139, 140, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.34121779859484774, 1.0, 0.9876112412177985], "iscrowd": 0, "area": 84978.62720000002, "rle": {"size": [427, 640], "counts": "\\:P1j;h0TOm0ZOf0_Ob0@a0B7H5K5N2N3M2N2N2O1N2N2N2O1O1O1O1N2O1O2N1O1O1O1O1O1O1N2O1O1O1O1O1O2N1`NYJ[Kh5c4]JXKe5e4`JXKa5f4cJWK^5g4jJRKW5l4e1L41O2N1O1O00100O1O1O1O1O1O2N2N1O2N2N2N2N3N2M2N3M3M1O2N1O2N1O2N1O2N1O101N1O1O2N1O2N1O1O2N3M2N2N2N2O2M2N2N4L4VKdH]3`7PLYIf3[8K00000000000001O000000000000000000000000010O000000000000000000000000001O000000000000000000000000010O000000000000000000000000001O000000000000000000000001O01O0000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O0O1000000000000000000000000000000000000O100000000000001O0000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000001O000000000000000000000O100000000000000000000000000000001O000000000000000000000000000000000000000O100000000000001O0000000000000000000000000000000000000000000000000000001O000O100000000000000000000000000000000000000000000000001O00000000000000000000000O1000000000000000000000000000001O00000000000000000000000000000000000000000O1000000000001O0000000000000000000000000000000000000000000000O10000000000000O0100000000000000O1000000000000O1000000cNdLkH\\3U7fLiHZ3V7hLiHX3W7jLgHV3Y7kLeHV3[7lLcHT3\\7nLcHR3]7PMaHP3_7QM`Ho2a7RM]Hn2c7SM\\Hm2d7UMZHk2f7VMXHk2h7WMVHi2k7WMTHi2l7YMRHg2n7ZMQHf2o7\\MoGc2R8_MlGa2U8_MjGa2V8aMgG`2Z8`MeG`2\\8aMbG_2^8cM`G]2a8cM^G]2b8eM\\G[2e8eMZG[2f8U11O001O001O001O001O00001O0000001O0000001O000000001O00000O2O0000001O0000001O0000001O0000000000000000000000000000000000O10000000000000O100000000000\\2dM:F;E:FQD"}, "label": "a table cloth"}]}
{"id": 546218, "image": "COCO_train2014_000000546218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom part of the tablecloth\"."}], "task": "refering", "answer_template": "The \"the bottom part of the tablecloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 139, 140, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.34121779859484774, 1.0, 0.9876112412177985], "iscrowd": 0, "area": 84978.62720000002, "rle": {"size": [427, 640], "counts": "\\:P1j;h0TOm0ZOf0_Ob0@a0B7H5K5N2N3M2N2N2O1N2N2N2O1O1O1O1N2O1O2N1O1O1O1O1O1O1N2O1O1O1O1O1O2N1`NYJ[Kh5c4]JXKe5e4`JXKa5f4cJWK^5g4jJRKW5l4e1L41O2N1O1O00100O1O1O1O1O1O2N2N1O2N2N2N2N3N2M2N3M3M1O2N1O2N1O2N1O2N1O101N1O1O2N1O2N1O1O2N3M2N2N2N2O2M2N2N4L4VKdH]3`7PLYIf3[8K00000000000001O000000000000000000000000010O000000000000000000000000001O000000000000000000000000010O000000000000000000000000001O000000000000000000000001O01O0000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000001O0O1000000000000000000000000000000000000O100000000000001O0000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000001O000000000000000000000O100000000000000000000000000000001O000000000000000000000000000000000000000O100000000000001O0000000000000000000000000000000000000000000000000000001O000O100000000000000000000000000000000000000000000000001O00000000000000000000000O1000000000000000000000000000001O00000000000000000000000000000000000000000O1000000000001O0000000000000000000000000000000000000000000000O10000000000000O0100000000000000O1000000000000O1000000cNdLkH\\3U7fLiHZ3V7hLiHX3W7jLgHV3Y7kLeHV3[7lLcHT3\\7nLcHR3]7PMaHP3_7QM`Ho2a7RM]Hn2c7SM\\Hm2d7UMZHk2f7VMXHk2h7WMVHi2k7WMTHi2l7YMRHg2n7ZMQHf2o7\\MoGc2R8_MlGa2U8_MjGa2V8aMgG`2Z8`MeG`2\\8aMbG_2^8cM`G]2a8cM^G]2b8eM\\G[2e8eMZG[2f8U11O001O001O001O001O00001O0000001O0000001O000000001O00000O2O0000001O0000001O0000001O0000000000000000000000000000000000O10000000000000O100000000000\\2dM:F;E:FQD"}, "label": "the bottom part of the tablecloth"}]}
{"id": 546218, "image": "COCO_train2014_000000546218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a creased white long sleeved shirt\"."}], "task": "refering", "answer_template": "The \"a creased white long sleeved shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 50, 51, 52, 53, 54, 56, 72, 73, 74, 76, 77, 79, 96, 100, 102], "bbox": [0.160421875, 0.0, 0.506765625, 0.3078688524590164], "iscrowd": 0, "area": 12355.31295, "rle": {"size": [427, 640], "counts": "eQ[15R=6K5J5L5N2N1O1O1O1O2N1O1dk2oNXUM5J5L4K5L4L4K500gDPOW:P1iEROU:o0iETOU:l0jEXOS:h0lE[OR:e0nE]OP:c0oE@o9`0PFCn9=QFEn9;RFGl99SFJk97SFLk94TFOj91VF1h9OWF3h9MXF4g9LYF5f9KZF6d9K\\F6c9J\\F8c9I\\F7d9I\\F8c9H]F9b9H]F9b9G^F:a9F^F<a9D_F=`9D_F<a9D_F=`9C`F>_9C`F>_9BaF?^9AbF`0]9AbF`0]9@dF?\\9BcF?\\9BcF@8_OT9Q1dF_O<^Oo8T1eF\\O`0^Oj8V1fFZOe0^Od8Y1fFXOk0\\O^8\\1gFWOb:i0_EUOb:k0^ETOc:l0]ESOd:m0\\EROe:n0\\EPOe:P1\\EnNe:R1[ElNg:T1ZEjNg:V1ZEhNg:X1YEgNh:Y1YEdNi:\\1XEbNi:^1b00000O1000000000000000000000000000000000000000O10000000000jDiNX:W1bESOZ:m0fEWOV:i0kEYOR:g0nE]On9b0SFAj9?VFEf9;ZFHc98]FL_94bFN[92eF2W9NiF5T9KlF9P9GPG<m8DSG?j8AVG?j8AWG=j8CVG=j8CVG<k8DUG;l8EUG:k8FWG7j8IXG5h8K[G1f8O\\GNe82]GLc84`GHa88oGWOR8i0WHnNi7R1ZHjNg7V1\\HgNd7Y1W20000000000O1000000000000000000000000000000000O1000000000000000000TEjNc9V1]F_On8a0RG@m8`0SGAl8?TGCj8<WGEh8;XGGf89ZGHe88[GId87\\GKb85^GLa84_GN_82aGO^81bG0]80cG2[8NeG3Z8MfG4Y8LgG6W8JiG7V8IjG0]80cGHe88[GAn8=RG[OX9c0iFTO`9k0`FmNj9Q1Z1N2N1O2N2N2N2N1O2N2N2N1N3N2N2N1O2N2N2N2NYgS4"}, "label": "a creased white long sleeved shirt"}]}
{"id": 546218, "image": "COCO_train2014_000000546218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the white shirt behind the wine glasses\"."}], "task": "refering", "answer_template": "The \"the person in the white shirt behind the wine glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 50, 51, 52, 53, 54, 56, 72, 73, 74, 76, 77, 79, 96, 100, 102], "bbox": [0.160421875, 0.0, 0.506765625, 0.3078688524590164], "iscrowd": 0, "area": 12355.31295, "rle": {"size": [427, 640], "counts": "eQ[15R=6K5J5L5N2N1O1O1O1O2N1O1dk2oNXUM5J5L4K5L4L4K500gDPOW:P1iEROU:o0iETOU:l0jEXOS:h0lE[OR:e0nE]OP:c0oE@o9`0PFCn9=QFEn9;RFGl99SFJk97SFLk94TFOj91VF1h9OWF3h9MXF4g9LYF5f9KZF6d9K\\F6c9J\\F8c9I\\F7d9I\\F8c9H]F9b9H]F9b9G^F:a9F^F<a9D_F=`9D_F<a9D_F=`9C`F>_9C`F>_9BaF?^9AbF`0]9AbF`0]9@dF?\\9BcF?\\9BcF@8_OT9Q1dF_O<^Oo8T1eF\\O`0^Oj8V1fFZOe0^Od8Y1fFXOk0\\O^8\\1gFWOb:i0_EUOb:k0^ETOc:l0]ESOd:m0\\EROe:n0\\EPOe:P1\\EnNe:R1[ElNg:T1ZEjNg:V1ZEhNg:X1YEgNh:Y1YEdNi:\\1XEbNi:^1b00000O1000000000000000000000000000000000000000O10000000000jDiNX:W1bESOZ:m0fEWOV:i0kEYOR:g0nE]On9b0SFAj9?VFEf9;ZFHc98]FL_94bFN[92eF2W9NiF5T9KlF9P9GPG<m8DSG?j8AVG?j8AWG=j8CVG=j8CVG<k8DUG;l8EUG:k8FWG7j8IXG5h8K[G1f8O\\GNe82]GLc84`GHa88oGWOR8i0WHnNi7R1ZHjNg7V1\\HgNd7Y1W20000000000O1000000000000000000000000000000000O1000000000000000000TEjNc9V1]F_On8a0RG@m8`0SGAl8?TGCj8<WGEh8;XGGf89ZGHe88[GId87\\GKb85^GLa84_GN_82aGO^81bG0]80cG2[8NeG3Z8MfG4Y8LgG6W8JiG7V8IjG0]80cGHe88[GAn8=RG[OX9c0iFTO`9k0`FmNj9Q1Z1N2N1O2N2N2N2N1O2N2N2N1N3N2N2N1O2N2N2N2NYgS4"}, "label": "the person in the white shirt behind the wine glasses"}]}
{"id": 488880, "image": "COCO_train2014_000000488880.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second elephant from the front\"."}], "task": "refering", "answer_template": "The \"the second elephant from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 267, 268, 282, 283, 284, 285, 286, 290, 291, 305, 306, 307, 308, 313, 314, 328, 329, 330, 331, 336, 337, 353, 354, 360], "bbox": [0.20515625, 0.3945, 0.702921875, 0.8989999999999999], "iscrowd": 0, "area": 31934.008349999993, "rle": {"size": [480, 640], "counts": "Yem11h><E;D<D<J6J6J6J7I6J4L3M2N3M3M3M2N3M3M2N3M3M3M2N3M2N1O1O1O1O1O1O2N1O1O1O1O1VFoKd8R4]GVLY8k3gG_Ln7b3RHeLf7\\3[HcLe7]3[HdLc7]3]HcLc7]3]HdLa7]3`HbL`7_3_HbL_7_3aHaL_7_3bHaL\\7`3dH`L\\7`3dHaL[7_3eHaLZ7`3gH^LZ7b3fH^LY7c3gH]LY7c3hH[LX7f3hHZLX7f3hHYLX7h3hHgKnN4Z8U4iHfKoN3X8X4hHeKQO1W8[4hHdKSOOU8]4iHcKTOMT8`4hHcKUOLR8b4iHbKWOJP8d4iHcKe7]4\\HcKc7]4]HfK`7Z4`HjK\\7V4eHlKX7T4hHoKU7Q4kHPLT7P4lHPLT7P4mHPLR7Q4mHPLR7P4nHQLQ7o3PIQLo6o3QIQLo6o3QIQLo6o3QIQLn6P4SIoKm6Q4SIoKm6Q4SIoKm6Q4TIoKk6Q4UIoKk6Q4UIoKk6Q4UIoKk6Q4VInKj6S4UImKk6S4UImKk6S4WIkKi6U4YIhKh6X4YIgKg6Y4[IdKf6\\4\\IbKd6^4]I`Kc6a4_I]Ka6c4aIZKa6e4`IZK`6f4bIWK_6i4l1001O000001O000000WGXKZ7h4fH^KT7b4lHeKm6[4SIkKh6T4XISLa6m3_I`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3mI_LS6a3mI_LS6a3mI_LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI_LV6`3jI`LV6`3kI_LU6a3kI_LU6a3kI_LU6a3kI_LU6a3kI_LU6a3kI_LU6a3kI_LV6`3jI`LV6`3jI`LV6`3jI`LV6`3kI_LU6a3kI_LU6a3kI^LV6b3jI^LW6a3oIXLR6h3TJRLl5n3WJnKj5Q4ZJkKg5U4[JhKf5X4]JdKd5\\4_J`Kb5`4aJ\\Ka5c4^20000O100O1O2O0O1O1K5L4L4M3M4K4N2N2O1N2O10001O000000000000001O000000000000001O0000000hNgDiNY;i0XETOh:k0[ESOf:k0^EROb:n0aEoN_:P1dEnN\\:R1gEjNZ:U1iEgNY:X1kEbNX:^1e1O2N100O1O1N2O1O2N4L4L4L4K5L4L5K4L\\kb0OdT]O2O2N1N3N1O2M2O1O2^C@`:b0^EA`:a0]EBb:>[EFd::ZEIe:7YELf:5VEOi:1WE1g:OYE2f:NZE3e:M[E4d:L\\E5c:K^E5a:K_E6`:J`E7_:J`E8^:HbE9]:GcE:\\:FdE:\\:FeE9[:GeE9[:GeE9[:GeE:Z:FfE:Z:GeE9[:GfE8Z:HfE8Z:HfE8Z:HfE8[:GeE9]:EcE;_:CbE=^:BbE>`:@aE?`:_ObE`0`:^ObE`0_:^OcEa0_:]OcEa0_:\\OdEa0^:^OcEa0_:]OcE`0`<N3N1N3M2O2M2N3Nndh2"}, "label": "the second elephant from the front"}]}
{"id": 488880, "image": "COCO_train2014_000000488880.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two elephants\"."}], "task": "refering", "answer_template": "The \"the two elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 267, 268, 282, 283, 284, 285, 286, 290, 291, 305, 306, 307, 308, 313, 314, 328, 329, 330, 331, 336, 337, 353, 354, 360], "bbox": [0.20515625, 0.3945, 0.702921875, 0.8989999999999999], "iscrowd": 0, "area": 31934.008349999993, "rle": {"size": [480, 640], "counts": "Yem11h><E;D<D<J6J6J6J7I6J4L3M2N3M3M3M2N3M3M2N3M3M3M2N3M2N1O1O1O1O1O1O2N1O1O1O1O1VFoKd8R4]GVLY8k3gG_Ln7b3RHeLf7\\3[HcLe7]3[HdLc7]3]HcLc7]3]HdLa7]3`HbL`7_3_HbL_7_3aHaL_7_3bHaL\\7`3dH`L\\7`3dHaL[7_3eHaLZ7`3gH^LZ7b3fH^LY7c3gH]LY7c3hH[LX7f3hHZLX7f3hHYLX7h3hHgKnN4Z8U4iHfKoN3X8X4hHeKQO1W8[4hHdKSOOU8]4iHcKTOMT8`4hHcKUOLR8b4iHbKWOJP8d4iHcKe7]4\\HcKc7]4]HfK`7Z4`HjK\\7V4eHlKX7T4hHoKU7Q4kHPLT7P4lHPLT7P4mHPLR7Q4mHPLR7P4nHQLQ7o3PIQLo6o3QIQLo6o3QIQLo6o3QIQLn6P4SIoKm6Q4SIoKm6Q4SIoKm6Q4TIoKk6Q4UIoKk6Q4UIoKk6Q4UIoKk6Q4VInKj6S4UImKk6S4UImKk6S4WIkKi6U4YIhKh6X4YIgKg6Y4[IdKf6\\4\\IbKd6^4]I`Kc6a4_I]Ka6c4aIZKa6e4`IZK`6f4bIWK_6i4l1001O000001O000000WGXKZ7h4fH^KT7b4lHeKm6[4SIkKh6T4XISLa6m3_I`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3mI_LS6a3mI_LS6a3mI_LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI`LT6`3lI_LV6`3jI`LV6`3kI_LU6a3kI_LU6a3kI_LU6a3kI_LU6a3kI_LU6a3kI_LU6a3kI_LV6`3jI`LV6`3jI`LV6`3jI`LV6`3kI_LU6a3kI_LU6a3kI^LV6b3jI^LW6a3oIXLR6h3TJRLl5n3WJnKj5Q4ZJkKg5U4[JhKf5X4]JdKd5\\4_J`Kb5`4aJ\\Ka5c4^20000O100O1O2O0O1O1K5L4L4M3M4K4N2N2O1N2O10001O000000000000001O000000000000001O0000000hNgDiNY;i0XETOh:k0[ESOf:k0^EROb:n0aEoN_:P1dEnN\\:R1gEjNZ:U1iEgNY:X1kEbNX:^1e1O2N100O1O1N2O1O2N4L4L4L4K5L4L5K4L\\kb0OdT]O2O2N1N3N1O2M2O1O2^C@`:b0^EA`:a0]EBb:>[EFd::ZEIe:7YELf:5VEOi:1WE1g:OYE2f:NZE3e:M[E4d:L\\E5c:K^E5a:K_E6`:J`E7_:J`E8^:HbE9]:GcE:\\:FdE:\\:FeE9[:GeE9[:GeE9[:GeE:Z:FfE:Z:GeE9[:GfE8Z:HfE8Z:HfE8Z:HfE8[:GeE9]:EcE;_:CbE=^:BbE>`:@aE?`:_ObE`0`:^ObE`0_:^OcEa0_:]OcEa0_:\\OdEa0^:^OcEa0_:]OcE`0`<N3N1N3M2O2M2N3Nndh2"}, "label": "the two elephants"}]}
{"id": 472502, "image": "COCO_train2014_000000472502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy serving a tennis ball\"."}], "task": "refering", "answer_template": "The \"a young boy serving a tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 219, 220, 233, 234, 235, 236, 237, 249, 251, 252, 253, 267, 268, 269, 283, 284, 285, 299, 300, 315, 316, 330, 331, 332, 347, 348], "bbox": [0.55945295404814, 0.5205625, 0.8437855579868709, 0.956390625], "iscrowd": 0, "area": 11797.570650000005, "rle": {"size": [640, 457], "counts": "Q]P51nc03M2N2O1N2N2N1O2O1O010O01N101N1O100O1O010O01000O010O10O010O10O10O010O10O0100O010O10O10SAGR::nEHP:7QFKl96TFLj94VFNg92ZFOe91[FOd92\\FNd91\\F1b90^F0b90^F0b9O_F1`90`F0`90`F1^9OcF1]9OcF1\\90cF1\\9OeF2Z9NfF2Y9OfF2Z9MgF3X9NgF4X9LhF4W9MiF3j7_OfD?^33jK^O\\;3YE=a32iK@[;3ZE<`33iK_O\\;5XE;^O\\Oa3e0[LB];3XE:ZOEb3<^LC];9RE5[OKd35bLC[;b0kDL^O3e3NfLBZ;g0gDg0Q3TOnL^OY;i0fDi0m2SOSM]OY;g0gDm0f2ROZMZOX;i0fDo0a2SO`MUOX;k0fDQ1[2ROgMROW;m0eDT1Z1VNgNk03oNU;n0fDV1n01WOkMU;o0dDZ1c06BbMV;P1dDb1155YMV;\\7jDdHV;\\7jDdHV;]7iDcHW;]7iDcHW;c6eDeI4HW;h6]EYIc:l6XETIh:R7REnHn:W7mDiHS;\\7hDdHX;a7cD_H\\;T8O0010O1O101N1YHcDa40kL];_NhDd43fLW;`NlDh45`LP;cNPEl46[Lk:dNUEo44YLf;f3^DVLd;h3_DULb;j3bDRL_;m3eDoK];n3hDnKY;Q4jDlKX;R4lDjKU;S4QEiKQ;m3[EoKf:h3fETL^:`3mE]LW:R1iBn0g`0oN[_Og0Qa0TOR_Oc0Ya0XOj^Oc0\\a0YOg^Oc0]a0ZOf^Oa0_a0\\Od^O`0_a0_Oc^O<ba0Ab^O:ba0Ca^O8da0E_^O7da0G_^O1aT]1"}, "label": "a young boy serving a tennis ball"}]}
{"id": 472502, "image": "COCO_train2014_000000472502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy about to hit a tennis ball\"."}], "task": "refering", "answer_template": "The \"a boy about to hit a tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 219, 220, 233, 234, 235, 236, 237, 249, 251, 252, 253, 267, 268, 269, 283, 284, 285, 299, 300, 315, 316, 330, 331, 332, 347, 348], "bbox": [0.55945295404814, 0.5205625, 0.8437855579868709, 0.956390625], "iscrowd": 0, "area": 11797.570650000005, "rle": {"size": [640, 457], "counts": "Q]P51nc03M2N2O1N2N2N1O2O1O010O01N101N1O100O1O010O01000O010O10O010O10O10O010O10O0100O010O10O10SAGR::nEHP:7QFKl96TFLj94VFNg92ZFOe91[FOd92\\FNd91\\F1b90^F0b90^F0b9O_F1`90`F0`90`F1^9OcF1]9OcF1\\90cF1\\9OeF2Z9NfF2Y9OfF2Z9MgF3X9NgF4X9LhF4W9MiF3j7_OfD?^33jK^O\\;3YE=a32iK@[;3ZE<`33iK_O\\;5XE;^O\\Oa3e0[LB];3XE:ZOEb3<^LC];9RE5[OKd35bLC[;b0kDL^O3e3NfLBZ;g0gDg0Q3TOnL^OY;i0fDi0m2SOSM]OY;g0gDm0f2ROZMZOX;i0fDo0a2SO`MUOX;k0fDQ1[2ROgMROW;m0eDT1Z1VNgNk03oNU;n0fDV1n01WOkMU;o0dDZ1c06BbMV;P1dDb1155YMV;\\7jDdHV;\\7jDdHV;]7iDcHW;]7iDcHW;c6eDeI4HW;h6]EYIc:l6XETIh:R7REnHn:W7mDiHS;\\7hDdHX;a7cD_H\\;T8O0010O1O101N1YHcDa40kL];_NhDd43fLW;`NlDh45`LP;cNPEl46[Lk:dNUEo44YLf;f3^DVLd;h3_DULb;j3bDRL_;m3eDoK];n3hDnKY;Q4jDlKX;R4lDjKU;S4QEiKQ;m3[EoKf:h3fETL^:`3mE]LW:R1iBn0g`0oN[_Og0Qa0TOR_Oc0Ya0XOj^Oc0\\a0YOg^Oc0]a0ZOf^Oa0_a0\\Od^O`0_a0_Oc^O<ba0Ab^O:ba0Ca^O8da0E_^O7da0G_^O1aT]1"}, "label": "a boy about to hit a tennis ball"}]}
{"id": 46519, "image": "COCO_train2014_000000046519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the pink shirt\"."}], "task": "refering", "answer_template": "The \"the girl in the pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 153, 154, 155, 156, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.5842968749999999, 0.2506041666666667, 0.8446406249999999, 1.0], "iscrowd": 0, "area": 45161.271649999995, "rle": {"size": [480, 640], "counts": "mW`51n>1TO5]B1a=8QBOm=c0aJTOiLm0o2@jLa0n2MjL6n26hL0Q3<dLJ[3<ZLJf3;oKKQ49eKMZ48\\KNd46QK1o42gJ5X50\\J7e5KRJ<n5FiIa0V6B`If0`6\\OXIk0g6WORIn0m6TOmHQ1S7QOfHT1Z7Y3O10000O10000O10000O1BZHbIf7Z6cHaI]7\\6jH_IW7^6g0L4M3M3L4M3M3L4M3K5J6K5K5K5J7J5L4000000000000000000O1O1O1O1O1N2O1O1O1N2O1O1O1O11O1O001O00100O001O001O1O001OO10000O10000O10000O1001O1O1O1O2N1O1O1O1O2O0O1O1O1O2N1O1O1O1O2N1O1O1O1O;E;D<E6I6K6VHdId6a6VI`Ii6g6oH\\IP7i6jHYIT7n6dHTI\\7Y7O0O2O1O001O001O1O001O001O1O001O2N1O1N2^NPHXK03Q8b3QHcL1E4Ol7d3VHdLMG7Ii7f3ZHfLKH7Eh7g3]HhLGId8Z3lGhLDLa8V3SHjL_ONZ9X2]GgM\\ONY:U1e3lNTR^1"}, "label": "the girl in the pink shirt"}]}
{"id": 46519, "image": "COCO_train2014_000000046519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the pink outfit\"."}], "task": "refering", "answer_template": "The \"the girl in the pink outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 153, 154, 155, 156, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.5842968749999999, 0.2506041666666667, 0.8446406249999999, 1.0], "iscrowd": 0, "area": 45161.271649999995, "rle": {"size": [480, 640], "counts": "mW`51n>1TO5]B1a=8QBOm=c0aJTOiLm0o2@jLa0n2MjL6n26hL0Q3<dLJ[3<ZLJf3;oKKQ49eKMZ48\\KNd46QK1o42gJ5X50\\J7e5KRJ<n5FiIa0V6B`If0`6\\OXIk0g6WORIn0m6TOmHQ1S7QOfHT1Z7Y3O10000O10000O10000O1BZHbIf7Z6cHaI]7\\6jH_IW7^6g0L4M3M3L4M3M3L4M3K5J6K5K5K5J7J5L4000000000000000000O1O1O1O1O1N2O1O1O1N2O1O1O1O11O1O001O00100O001O001O1O001OO10000O10000O10000O1001O1O1O1O2N1O1O1O1O2O0O1O1O1O2N1O1O1O1O2N1O1O1O1O;E;D<E6I6K6VHdId6a6VI`Ii6g6oH\\IP7i6jHYIT7n6dHTI\\7Y7O0O2O1O001O001O1O001O001O1O001O2N1O1N2^NPHXK03Q8b3QHcL1E4Ol7d3VHdLMG7Ii7f3ZHfLKH7Eh7g3]HhLGId8Z3lGhLDLa8V3SHjL_ONZ9X2]GgM\\ONY:U1e3lNTR^1"}, "label": "the girl in the pink outfit"}]}
{"id": 46519, "image": "COCO_train2014_000000046519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a cast and his arm in a sling playing a video game\"."}], "task": "refering", "answer_template": "The \"a person with a cast and his arm in a sling playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 122, 141, 142, 143, 144, 145, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329, 349, 350, 351, 352, 372, 373, 374, 375], "bbox": [0.131453125, 0.2561875, 0.35225, 0.9865208333333333], "iscrowd": 0, "area": 27925.0149, "rle": {"size": [480, 640], "counts": "PaW11k>6K5J6K5J5L5J6K5J6M3N2N1O2N2M2O1O1bDQNi9P2TFVNh9k1UFZNh9g1UF_Ng9b1UFdNh9]1UFiNf9Y1WFmNe9T1XFQOe9P1XFVOd9k0YFZOd9g0XF^Of9c0WFAg9`0VFDh9=UFGh9;UFIi98SFMk9g2M2N2N2N2N2M3N2N2N2N2N3\\JWJdMMQ3m5SOcKm0_4jNhKV1Z4aNmK_1V4WNQLj1Q4lMVLT2m3cMYL]2j3YM^Lf2d3SMaLn2a3kLcLU3`3eLcL[3`3_LcLa3_3ZLdLg3^3SLeLm3^3lKfLT4S7001O000000000000O11O4L4L3M4M3jGYKk5j4RJYKk5k4PJZKl5P5jITKR6W5bInJZ6^5ZIdJd6b5UI`Jj6_5WIcJg6\\5YIfJf6Y5[IhJd6X5\\IjJb6U5^ImJa6R5`IPK^6o4bISK]6m4cIUK[6j4eIXKZ6g4gIZKX6e4hI^KV6a4kI`KT6`4kIcKS6\\4nIeKQ6Z4oIhKQ6V4PJlKn5T4RJmKm5R4SJQLk5n3VJSLi5l3WJVLh5j3XJXLf5g3ZJ[Le5d3\\J^Lb5a3^JaLa5_3_JbL`5]3`JfL^5Y3cJhL\\5V3fJlLX5R3iJQMU5n2kJUMS5i2nJZMP5d2QK^MROXMj5Y5UK]MSO\\Mf5U5XK^MUO^Ma5R5[K_MWO`M\\5o4^K_MZOcMV5m4`K_M\\OdMU5j4`KaM^OcMS5j4aK`M@dMP5l4_K_M[5b2cJ^Mg1gMi0k4_M_Mg1mMd0d4dMaMf1SN?]4iMaMg1XN;W4nMaMf1_N6Q4RNaMf1fN2i3WNbMf1lNMb3\\NcMf1QOI]3_NcMg1WODV3dNgMd1ZOBP3iNjM`1]OAi2nNnM]1_O@c2ROSNX1B@\\2VOVNV1E^OU2SObNQ1JFh5@SK;o9J6J6K4K\\RR6"}, "label": "a person with a cast and his arm in a sling playing a video game"}]}
{"id": 46519, "image": "COCO_train2014_000000046519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid with a cast and no shirt playing wii tennis\"."}], "task": "refering", "answer_template": "The \"a kid with a cast and no shirt playing wii tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 122, 141, 142, 143, 144, 145, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329, 349, 350, 351, 352, 372, 373, 374, 375], "bbox": [0.131453125, 0.2561875, 0.35225, 0.9865208333333333], "iscrowd": 0, "area": 27925.0149, "rle": {"size": [480, 640], "counts": "PaW11k>6K5J6K5J5L5J6K5J6M3N2N1O2N2M2O1O1bDQNi9P2TFVNh9k1UFZNh9g1UF_Ng9b1UFdNh9]1UFiNf9Y1WFmNe9T1XFQOe9P1XFVOd9k0YFZOd9g0XF^Of9c0WFAg9`0VFDh9=UFGh9;UFIi98SFMk9g2M2N2N2N2N2M3N2N2N2N2N3\\JWJdMMQ3m5SOcKm0_4jNhKV1Z4aNmK_1V4WNQLj1Q4lMVLT2m3cMYL]2j3YM^Lf2d3SMaLn2a3kLcLU3`3eLcL[3`3_LcLa3_3ZLdLg3^3SLeLm3^3lKfLT4S7001O000000000000O11O4L4L3M4M3jGYKk5j4RJYKk5k4PJZKl5P5jITKR6W5bInJZ6^5ZIdJd6b5UI`Jj6_5WIcJg6\\5YIfJf6Y5[IhJd6X5\\IjJb6U5^ImJa6R5`IPK^6o4bISK]6m4cIUK[6j4eIXKZ6g4gIZKX6e4hI^KV6a4kI`KT6`4kIcKS6\\4nIeKQ6Z4oIhKQ6V4PJlKn5T4RJmKm5R4SJQLk5n3VJSLi5l3WJVLh5j3XJXLf5g3ZJ[Le5d3\\J^Lb5a3^JaLa5_3_JbL`5]3`JfL^5Y3cJhL\\5V3fJlLX5R3iJQMU5n2kJUMS5i2nJZMP5d2QK^MROXMj5Y5UK]MSO\\Mf5U5XK^MUO^Ma5R5[K_MWO`M\\5o4^K_MZOcMV5m4`K_M\\OdMU5j4`KaM^OcMS5j4aK`M@dMP5l4_K_M[5b2cJ^Mg1gMi0k4_M_Mg1mMd0d4dMaMf1SN?]4iMaMg1XN;W4nMaMf1_N6Q4RNaMf1fN2i3WNbMf1lNMb3\\NcMf1QOI]3_NcMg1WODV3dNgMd1ZOBP3iNjM`1]OAi2nNnM]1_O@c2ROSNX1B@\\2VOVNV1E^OU2SObNQ1JFh5@SK;o9J6J6K4K\\RR6"}, "label": "a kid with a cast and no shirt playing wii tennis"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in tan pants and a gray shirt sitting on a sofa\"."}], "task": "refering", "answer_template": "The \"a man in tan pants and a gray shirt sitting on a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 155, 156, 157, 158, 159, 175, 176, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 267, 268, 290, 291, 312, 313], "bbox": [0.591875, 0.31871635610766047, 0.9242812499999999, 0.801863354037267], "iscrowd": 0, "area": 17629.882700000006, "rle": {"size": [483, 640], "counts": "[Tc52P?3M3M3M2N1O100O001O1O1O001O1O1O1N101O1EXOZBi0`=DXB<d=d0J6eLlNbHW1]7nN[HV1f7mNSHV1o7kNeG^1]8dNmFn1U9TNhFk1Y9XNdFg1^9[N_Fd1c9]N[Fb1g9`NWF^1k9dNRF\\1o9fNnEY1U:iNgEW1Z:kNcET1_:nN^ER1c:QOZEm0h:UOUEk0m:c13M3N1N3N2M2O2M3M2N3M3M2M4M3M2N3M3M3M2N3M3M3N2N2N101N2O1N1O2O0O2O0O2N100O2N101N101N1O2O0O2O0000000O1000000000O001O1N2O1O0O2^M^NbGc1^8aN]G`1c8cNXG_1h8eNSG\\1l8YO_Fg0b9[OZFg0f9[OWFf0i9\\OTFe0k9^ORFc0n9_OnEb0S:@jEa0V:AgE`0X:CdE>]:JZE7f:R210O10O100O010O10000O100O10000O101N10000O100O10000O100O100O100O100O101N100O100O100O100O10000O100O1O1O2N1O1N2O1O1O1O1N2O1O2N1O1N2O1O1O1O2N1N3N1O2N1O2M3N1O2N1O2M2N3L3N3L3N3M2M4M3L3L5H7I8I6L5L3LW_f0"}, "label": "a man in tan pants and a gray shirt sitting on a sofa"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with gray clothes , black shoes , and glasses relaxing on a couch\"."}], "task": "refering", "answer_template": "The \"a man with gray clothes , black shoes , and glasses relaxing on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 155, 156, 157, 158, 159, 175, 176, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 267, 268, 290, 291, 312, 313], "bbox": [0.591875, 0.31871635610766047, 0.9242812499999999, 0.801863354037267], "iscrowd": 0, "area": 17629.882700000006, "rle": {"size": [483, 640], "counts": "[Tc52P?3M3M3M2N1O100O001O1O1O001O1O1O1N101O1EXOZBi0`=DXB<d=d0J6eLlNbHW1]7nN[HV1f7mNSHV1o7kNeG^1]8dNmFn1U9TNhFk1Y9XNdFg1^9[N_Fd1c9]N[Fb1g9`NWF^1k9dNRF\\1o9fNnEY1U:iNgEW1Z:kNcET1_:nN^ER1c:QOZEm0h:UOUEk0m:c13M3N1N3N2M2O2M3M2N3M3M2M4M3M2N3M3M3M2N3M3M3N2N2N101N2O1N1O2O0O2O0O2N100O2N101N101N1O2O0O2O0000000O1000000000O001O1N2O1O0O2^M^NbGc1^8aN]G`1c8cNXG_1h8eNSG\\1l8YO_Fg0b9[OZFg0f9[OWFf0i9\\OTFe0k9^ORFc0n9_OnEb0S:@jEa0V:AgE`0X:CdE>]:JZE7f:R210O10O100O010O10000O100O10000O101N10000O100O10000O100O100O100O100O101N100O100O100O100O10000O100O1O1O2N1O1N2O1O1O1O1N2O1O2N1O1N2O1O1O1O2N1N3N1O2N1O2M3N1O2N1O2M2N3L3N3L3N3M2M4M3L3L5H7I8I6L5L3LW_f0"}, "label": "a man with gray clothes , black shoes , and glasses relaxing on a couch"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a beard sitting with his legs crossed and holding up the peace sign\"."}], "task": "refering", "answer_template": "The \"a man with a beard sitting with his legs crossed and holding up the peace sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 122, 123, 124, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 235, 236, 259], "bbox": [0.1626875, 0.26053830227743274, 0.46396875, 0.7105383022774328], "iscrowd": 0, "area": 14020.509399999999, "rle": {"size": [483, 640], "counts": "eha12P?2N2O1N2N100O2N100O10001N10000O101A_OYBb0d=B[B>c=D]B<a=F^B;`=H_B9^=IbB7\\=KcB6\\=LcB4]=LbB5]=LcB4]=KdB5[=LdB4]=LcB4]=LcB3]=NbB2_=NaB2^=ObB0_=ObB1]=0cBO^=1aB0^=1bBN_=2aBN_=2aBM_=4aBL_=4aBL^=5bBJ_=6aBJ_=6aBJ_=5bBK^=5aBK`=5`BK`=5`BK`=4aBL_=4aBK`=4aBL_=4aBL_=3bBN]=2cBN]=1cB0]=0cBe1h;ZNYDk1b;UN^DP2];oMdDV2W;jMiD[2R;dMoD`2m:`MSEc2j:\\MWEe2h:[MXEf2g:ZMYEg2f:XM[Ei2d:WM\\Ej2c:VM]Ek2a:VM_Ek2^:VMcEk2[:VMeEk2Y:VMgEk2V:VMkEk2S:VMmES2KgMU:9oEj1n:WNREd1R;^NlD^1X;cNhDe0o;]OPD?T<AlC>T<DkC<U<DjC=U<EjC;V<EjC:V<HiC8V<IjC7V<JhC7W<KhC4Y<LgC4X<NgC2X<0gC0Y<0fC0Z<]100O1O100O100O100O0O2O1N2N2O1N2O1[Oe0C=0000O100O10000O10000O10O01000O01000O01001N2O1O1N2M3N3M2M3N2M9`LTEZ2k;N3M2N2O1N3M2O100O2N100O2N1O10O000H8H9G8N23N3L3N3L3N3M3L3N3M2N3L5L5K5J4M3M3L4L4M2M3M3N2M3M3M3N2M[bQ5"}, "label": "a man with a beard sitting with his legs crossed and holding up the peace sign"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on left with peace sign\"."}], "task": "refering", "answer_template": "The \"man on left with peace sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 122, 123, 124, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 235, 236, 259], "bbox": [0.1626875, 0.26053830227743274, 0.46396875, 0.7105383022774328], "iscrowd": 0, "area": 14020.509399999999, "rle": {"size": [483, 640], "counts": "eha12P?2N2O1N2N100O2N100O10001N10000O101A_OYBb0d=B[B>c=D]B<a=F^B;`=H_B9^=IbB7\\=KcB6\\=LcB4]=LbB5]=LcB4]=KdB5[=LdB4]=LcB4]=LcB3]=NbB2_=NaB2^=ObB0_=ObB1]=0cBO^=1aB0^=1bBN_=2aBN_=2aBM_=4aBL_=4aBL^=5bBJ_=6aBJ_=6aBJ_=5bBK^=5aBK`=5`BK`=5`BK`=4aBL_=4aBK`=4aBL_=4aBL_=3bBN]=2cBN]=1cB0]=0cBe1h;ZNYDk1b;UN^DP2];oMdDV2W;jMiD[2R;dMoD`2m:`MSEc2j:\\MWEe2h:[MXEf2g:ZMYEg2f:XM[Ei2d:WM\\Ej2c:VM]Ek2a:VM_Ek2^:VMcEk2[:VMeEk2Y:VMgEk2V:VMkEk2S:VMmES2KgMU:9oEj1n:WNREd1R;^NlD^1X;cNhDe0o;]OPD?T<AlC>T<DkC<U<DjC=U<EjC;V<EjC:V<HiC8V<IjC7V<JhC7W<KhC4Y<LgC4X<NgC2X<0gC0Y<0fC0Z<]100O1O100O100O100O0O2O1N2N2O1N2O1[Oe0C=0000O100O10000O10000O10O01000O01000O01001N2O1O1N2M3N3M2M3N2M9`LTEZ2k;N3M2N2O1N3M2O100O2N100O2N1O10O000H8H9G8N23N3L3N3L3N3M3L3N3M2N3L5L5K5J4M3M3L4L4M2M3M3N2M3M3M3N2M[bQ5"}, "label": "man on left with peace sign"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"section of couch touching wall\"."}], "task": "refering", "answer_template": "The \"section of couch touching wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 124, 125, 126, 127, 128, 129, 130, 144, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290], "bbox": [0.18756250000000002, 0.3100207039337474, 0.773515625, 0.7284265010351968], "iscrowd": 0, "area": 35842.93029999999, "rle": {"size": [483, 640], "counts": "jYi12P?1O0cAOi=2VB8`=I`B`0V=AiBi0l<XOUCQ1a<PO_CY1W<hNhCb1m;`NSDa1k;_NUDb1j;_NVDb1i;^NWDb1i;^NVDc1k;\\NUDd1k;[NUDf1k;ZNUDg1j;YNVDg1k;XNTDi1l;VNUDj1k;VNTDk1k;VNUDk1i;VNWDj1i;UNWDl1h;UNXDk1h;UNXDk1g;VNYDk1f;TN[Dl1d;UN\\Dk1d;UN[Dl1e;UNZDl1d;UN\\Dk1d;UN\\Dk1d;UN\\Dk1d;UN\\Dk1d;UN\\Dl1c;TN]Dl1c;TN]Dl1c;TN]Dl1c;TN]Dl1c;TN]Dk0DIm;]O_D>41];A_D3c08n:E^DHT1?^:IXF3h9M\\F0S9CRE<P2Mh8LYE7S2I^84aE3U2ET8=hEMl<7VCId<;]CF]<?dCAX<a0jC^OV<b0kC^OS<b0oC^OP<a0RD^On;b0SD^Ok;b0WD^Og10n6b0\\G^OP1f0c7K_G^O?W1P8[OeG\\O;Y1o7[OhG[O8[1n7[OmGWO5_1m7ZOPHVO3`1m7ZORHTO0c1n7XOdIj0[6UOfIk0Z6UOeIl0[6SOeIn0[6QOeIP1[6POeIP1[6oNeIR1[6mNeIT1[6kNfIV1Y6jNfIW1Z6hNfIY1Z6fNgIZ1Y6fNfI\\1Y6cNgI^1Y6aNhI_1X6aNhI_1X6`NiI`1W6_NjIb1U6]NmIb1S6^NmIb1S6]NnIc1R6\\NoId1Q6\\NoIe1P6[NPJe1P6[NPJe1P6\\NoId1Q6\\NoId1Q6\\NPJd1P6[NPJe1P6YNRJg1n5TNWJl1i5PN[JP2\\91O0000001N100O100O2O0O100O101N100O100O2O0O10000O2OXOPD]No;l0gC]OR=`0UC[Oh9FmHn0_MXOd9LlHm0cMQOc93iHm0l7TOSHm0l7TOSHm0l7TOSHm0l7TOSHm0k7UOUHj0k7XOSHh0m7YORHg0n7ZOQHf0o7[OPHe0o7\\OQHd0o7[ORHe0n7YOTHg0P8TORHk0S;0000000000A?10O01O0aLAZH`0g7ERH<o7GlG:T8IhG8W8KgG5X8NeG3Z80cG2[81bG0]83`GN_85^GLb85\\GLc87ZGJe89YGHe8:YGGf8:YGGf8;XGF_OmNk7_1dHE\\OPOP8]1bHDYOSOU8[1`HCWOUOY8Z1^HCSOWO_8X1]HAoN[Od8V1[H@mN]Oh8T1ZH@iN@m8P1ZHAeNBQ9l0[Hf0f7YOZHg0f7YOZHg0f7XO[Hh0e7XOZHi0f7WOYHj0g7UOYHl0g7TOWHo0h7QOWHP1i7POVHQ1j7nNVHS1k7lNTHU1l7kNSHV1m7iNSHX1m7iNPHY1P8gNoGZ1Q8fNnG[1R8eNmG\\1S8eNkG\\1U8dNjG]1W8bNhG_1X8aNgG`1Y8aNdGa1\\8_NcGb1]8^NbGd1]8]NaGd1_8]N_Gd1a8]N]Gd1d8\\NZGe1f8\\NWGf1i8[NUGf1k8[NSGf1m8[NQGf1o8[NoFf1Q9V2001O000000000001O01O000000000000000000000000000001O00000000000000000000000000000000001N100000000000000000000000000000001O00000000TLlFh1U9WNnFc1V9\\NmF]1X9aNlFY1X9fNlFR1Y9lNkFm0Z9ROiFh0[9WOhFb0]9\\OhF=\\9BgF7^9HeF2_9ObFXOU:h0nEUOR:k0PFROQ:n0QFoNP:Q1RFmNn9S1UFiNl9W1UFgNl9Z1UFdNk9\\1WFaNj9_1WF_Nj9a1XF\\Ni9d1XF[Nh9e1k100001O00002N1O1O2O01O01O00010O00010O00OO2O000O101O0O101O0O100O2O000O1010O2N2O0O2N1O01O00001O0000001O00001O0000001OO1N2M3N2N2N2N3M5K5K5J6K4L5K5K5K4L5KUVT2"}, "label": "section of couch touching wall"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch with two occupants\"."}], "task": "refering", "answer_template": "The \"a couch with two occupants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 124, 125, 126, 127, 128, 129, 130, 144, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290], "bbox": [0.18756250000000002, 0.3100207039337474, 0.773515625, 0.7284265010351968], "iscrowd": 0, "area": 35842.93029999999, "rle": {"size": [483, 640], "counts": "jYi12P?1O0cAOi=2VB8`=I`B`0V=AiBi0l<XOUCQ1a<PO_CY1W<hNhCb1m;`NSDa1k;_NUDb1j;_NVDb1i;^NWDb1i;^NVDc1k;\\NUDd1k;[NUDf1k;ZNUDg1j;YNVDg1k;XNTDi1l;VNUDj1k;VNTDk1k;VNUDk1i;VNWDj1i;UNWDl1h;UNXDk1h;UNXDk1g;VNYDk1f;TN[Dl1d;UN\\Dk1d;UN[Dl1e;UNZDl1d;UN\\Dk1d;UN\\Dk1d;UN\\Dk1d;UN\\Dk1d;UN\\Dl1c;TN]Dl1c;TN]Dl1c;TN]Dl1c;TN]Dl1c;TN]Dk0DIm;]O_D>41];A_D3c08n:E^DHT1?^:IXF3h9M\\F0S9CRE<P2Mh8LYE7S2I^84aE3U2ET8=hEMl<7VCId<;]CF]<?dCAX<a0jC^OV<b0kC^OS<b0oC^OP<a0RD^On;b0SD^Ok;b0WD^Og10n6b0\\G^OP1f0c7K_G^O?W1P8[OeG\\O;Y1o7[OhG[O8[1n7[OmGWO5_1m7ZOPHVO3`1m7ZORHTO0c1n7XOdIj0[6UOfIk0Z6UOeIl0[6SOeIn0[6QOeIP1[6POeIP1[6oNeIR1[6mNeIT1[6kNfIV1Y6jNfIW1Z6hNfIY1Z6fNgIZ1Y6fNfI\\1Y6cNgI^1Y6aNhI_1X6aNhI_1X6`NiI`1W6_NjIb1U6]NmIb1S6^NmIb1S6]NnIc1R6\\NoId1Q6\\NoIe1P6[NPJe1P6[NPJe1P6\\NoId1Q6\\NoId1Q6\\NPJd1P6[NPJe1P6YNRJg1n5TNWJl1i5PN[JP2\\91O0000001N100O100O2O0O100O101N100O100O2O0O10000O2OXOPD]No;l0gC]OR=`0UC[Oh9FmHn0_MXOd9LlHm0cMQOc93iHm0l7TOSHm0l7TOSHm0l7TOSHm0l7TOSHm0k7UOUHj0k7XOSHh0m7YORHg0n7ZOQHf0o7[OPHe0o7\\OQHd0o7[ORHe0n7YOTHg0P8TORHk0S;0000000000A?10O01O0aLAZH`0g7ERH<o7GlG:T8IhG8W8KgG5X8NeG3Z80cG2[81bG0]83`GN_85^GLb85\\GLc87ZGJe89YGHe8:YGGf8:YGGf8;XGF_OmNk7_1dHE\\OPOP8]1bHDYOSOU8[1`HCWOUOY8Z1^HCSOWO_8X1]HAoN[Od8V1[H@mN]Oh8T1ZH@iN@m8P1ZHAeNBQ9l0[Hf0f7YOZHg0f7YOZHg0f7XO[Hh0e7XOZHi0f7WOYHj0g7UOYHl0g7TOWHo0h7QOWHP1i7POVHQ1j7nNVHS1k7lNTHU1l7kNSHV1m7iNSHX1m7iNPHY1P8gNoGZ1Q8fNnG[1R8eNmG\\1S8eNkG\\1U8dNjG]1W8bNhG_1X8aNgG`1Y8aNdGa1\\8_NcGb1]8^NbGd1]8]NaGd1_8]N_Gd1a8]N]Gd1d8\\NZGe1f8\\NWGf1i8[NUGf1k8[NSGf1m8[NQGf1o8[NoFf1Q9V2001O000000000001O01O000000000000000000000000000001O00000000000000000000000000000000001N100000000000000000000000000000001O00000000TLlFh1U9WNnFc1V9\\NmF]1X9aNlFY1X9fNlFR1Y9lNkFm0Z9ROiFh0[9WOhFb0]9\\OhF=\\9BgF7^9HeF2_9ObFXOU:h0nEUOR:k0PFROQ:n0QFoNP:Q1RFmNn9S1UFiNl9W1UFgNl9Z1UFdNk9\\1WFaNj9_1WF_Nj9a1XF\\Ni9d1XF[Nh9e1k100001O00002N1O1O2O01O01O00010O00010O00OO2O000O101O0O101O0O100O2O000O1010O2N2O0O2N1O01O00001O0000001O00001O0000001OO1N2M3N2N2N2N3M5K5K5J6K4L5K5K5K4L5KUVT2"}, "label": "a couch with two occupants"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing black t - shirt\"."}], "task": "refering", "answer_template": "The \"a women wearing black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 340, 341, 342, 343, 344, 357, 358, 359], "bbox": [0.5393125000000001, 0.5213457556935818, 0.9989062499999999, 0.9415734989648032], "iscrowd": 0, "area": 21042.430449999993, "rle": {"size": [483, 640], "counts": "nTS52Q?5K1O1O001O1O1O1O001O1O000000000000O100000O100O01000O101N100O100O100O2N3M3M2N1N101O1O1O1OLUBROh=P123M3N2000O10000O100O1O1O1O100O0000000O100O1O2N1O1O100010O0010000O2O00010O01O10O0100O010O10O0100O1000000001N100O100O1O100O2N1O1O1O1O10001O01O0001O00001O01O01O0000001O00001O0000001O001O001O00001O0O2O001N100O2O001N101N10001N101O0O2O0O101O0O2O0O101O0O101O0O100O2O001N101N1O1O2N1O2N1O2N1O2O0O1O2N101N1O2O0O2N100O1O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1G\\LQEf3m:^LPEc3m:;M3M3N2O1N2N2O1N2N2N2O001N2O1N20O1000O100O010O10000O10O0100O1000O10001O000O10001O000O101O0O1O1O2N1N2O1O1O2N1N2N2N3L3N2N2N2M4M2O1N2N3M3M3M3M3SOcDPN`;d1V1A?YO^7"}, "label": "a women wearing black t - shirt"}]}
{"id": 366009, "image": "COCO_train2014_000000366009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady is sitting in the chair next to a box of tissue\"."}], "task": "refering", "answer_template": "The \"the lady is sitting in the chair next to a box of tissue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 340, 341, 342, 343, 344, 357, 358, 359], "bbox": [0.5393125000000001, 0.5213457556935818, 0.9989062499999999, 0.9415734989648032], "iscrowd": 0, "area": 21042.430449999993, "rle": {"size": [483, 640], "counts": "nTS52Q?5K1O1O001O1O1O1O001O1O000000000000O100000O100O01000O101N100O100O100O2N3M3M2N1N101O1O1O1OLUBROh=P123M3N2000O10000O100O1O1O1O100O0000000O100O1O2N1O1O100010O0010000O2O00010O01O10O0100O010O10O0100O1000000001N100O100O1O100O2N1O1O1O1O10001O01O0001O00001O01O01O0000001O00001O0000001O001O001O00001O0O2O001N100O2O001N101N10001N101O0O2O0O101O0O2O0O101O0O101O0O100O2O001N101N1O1O2N1O2N1O2N1O2O0O1O2N101N1O2O0O2N100O1O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1G\\LQEf3m:^LPEc3m:;M3M3N2O1N2N2O1N2N2N2O001N2O1N20O1000O100O010O10000O10O0100O1000O10001O000O10001O000O101O0O1O1O2N1N2O1O1O2N1N2N2N3L3N2N2N2M4M2O1N2N3M3M3M3M3SOcDPN`;d1V1A?YO^7"}, "label": "the lady is sitting in the chair next to a box of tissue"}]}
{"id": 472506, "image": "COCO_train2014_000000472506.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra was eting in gras\"."}], "task": "refering", "answer_template": "The \"a zebra was eting in gras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 47, 48, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 117, 118, 119, 120, 122, 123, 124, 133, 137, 138, 141, 142, 151, 155, 156, 160, 173], "bbox": [0.39174000000000003, 0.19183183183183183, 0.94402, 0.7887687687687689], "iscrowd": 0, "area": 26948.601300000006, "rle": {"size": [333, 500], "counts": "[jo11[:2N3N1N3bGIb68YIOc64WI3WOCf6=nI=PO_Oa0E`5a0kJQ2n4SNnJS2l4PNPKS2n4QNnJR2o4h1N2N2jJWKe4k4WKZKg4g4UK\\Kj4S510O10O0VNmK\\MS4a2XLVMi3h2]LSMc3k2cLQM^3m2gLPMY3n2PMjLQ3T3T201O1O1O1O1O2N1O1O1O2N1000010O01O1O010O001O010O1O0010O01O1O010O1O010O000010O01O010O000010O000001O01O000000000000001O0000000000001O00000O100000001O0000O010000000000O1000000000RKSM\\2l2aM^MY2h2VMcMj2i2fL\\MY3]3iKhLW4^3^KgLb4c4O00000000O100O0C`KnJa4^4TLaKl3Z4ZLdKg3W4_LgKb3R4`1E;F:N2N3M2N2O[InLQ6m3J5K5aJgK`4^4VKkKf4[4nJkKS5g4000010O0001O0010O00`NWL\\Lj3W3l1M3M4M2M4L3N2N3N1O2N1N3N1O1O2O001O00010O001O0010O0001O010O001O0010O01O0010O01O1O010O001O010O001O010O010O010O00010O010O0100O010O1O100jGUNh7l1TH[Ni7V2N2O0O100O100O1O100O2O1N2O001N2O1O00O010O100O10O01O010O10O010O01O1O0O2O1O0O2O0O1000OSN_Ho0a7QObHm0]7TOeHj0Z7WOjHd0W7\\OeIH[67QJ]OP6c0m102I?APo8"}, "label": "a zebra was eting in gras"}]}
{"id": 472506, "image": "COCO_train2014_000000472506.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left who is facing the camera\"."}], "task": "refering", "answer_template": "The \"the zebra on the left who is facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 111, 112, 113, 114, 129, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167, 168, 184], "bbox": [0.15016, 0.20210210210210208, 0.4194, 0.8544144144144143], "iscrowd": 0, "area": 18178.700299999997, "rle": {"size": [333, 500], "counts": "Zah04R::F9G9I8J5L4L4L5nG\\NS7h1kHZNP7k1oHVNm6n1RISNj6Q2VIoMf6U2YIlMd0Fk4b2`JiM?Km4`2cJfM:0o4b2cJ^M87Q5d2`JVM9<S5T3eJRMX5U4K4M2N3M2O2M2O1N3N1O2M2O2M10O10O10O0100O2O001N101N101N101OO0100O10000O10O0^O\\L]Jd3[5j0N3L4M2jNmJQMV5k2nJQMU5l2PKPMS5l2QKfL\\5k2UKTMk4l2UKSMl4m2TKSMk4n2UKRMk4m2VKRMj4o2VKQMj4o2VKPMk4P3UKPMj4Q3VKoLj4P3VKPMHKV4V3RLoLGOT4R3ULnLG3Q4P3XLmLF5Q4o2XLkLG8P4o2WLiLH9Q4P3ULgLH<Q4P3ULcLGa0T4n2SL]LJh0S4m2`LVM`3m2ZLVMg3l2SLVMo3l2kKWMV4k2eKVM\\4]41N2M2L6K4L5mMdJoN`5P1cJjNb5T1`JgNd5X1_JaNg5^1ZJ]Nj5b1YJXNl5f1WJUNl5j1WJPNn5n1VJkMn5V2SJdMR6\\2o010O0S1RI`Lj4d4M4M2M4M=B01O010fNTKQMl4h2aKRM`4b2PLZMP4W2dLcM\\3Q2[M\\Mn2]2d2J7I6J9K4L5K5J5L4L4L4L5K4CiVn2"}, "label": "the zebra on the left who is facing the camera"}]}
{"id": 472506, "image": "COCO_train2014_000000472506.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra facing camera\"."}], "task": "refering", "answer_template": "The \"zebra facing camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 111, 112, 113, 114, 129, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167, 168, 184], "bbox": [0.15016, 0.20210210210210208, 0.4194, 0.8544144144144143], "iscrowd": 0, "area": 18178.700299999997, "rle": {"size": [333, 500], "counts": "Zah04R::F9G9I8J5L4L4L5nG\\NS7h1kHZNP7k1oHVNm6n1RISNj6Q2VIoMf6U2YIlMd0Fk4b2`JiM?Km4`2cJfM:0o4b2cJ^M87Q5d2`JVM9<S5T3eJRMX5U4K4M2N3M2O2M2O1N3N1O2M2O2M10O10O10O0100O2O001N101N101N101OO0100O10000O10O0^O\\L]Jd3[5j0N3L4M2jNmJQMV5k2nJQMU5l2PKPMS5l2QKfL\\5k2UKTMk4l2UKSMl4m2TKSMk4n2UKRMk4m2VKRMj4o2VKQMj4o2VKPMk4P3UKPMj4Q3VKoLj4P3VKPMHKV4V3RLoLGOT4R3ULnLG3Q4P3XLmLF5Q4o2XLkLG8P4o2WLiLH9Q4P3ULgLH<Q4P3ULcLGa0T4n2SL]LJh0S4m2`LVM`3m2ZLVMg3l2SLVMo3l2kKWMV4k2eKVM\\4]41N2M2L6K4L5mMdJoN`5P1cJjNb5T1`JgNd5X1_JaNg5^1ZJ]Nj5b1YJXNl5f1WJUNl5j1WJPNn5n1VJkMn5V2SJdMR6\\2o010O0S1RI`Lj4d4M4M2M4M=B01O010fNTKQMl4h2aKRM`4b2PLZMP4W2dLcM\\3Q2[M\\Mn2]2d2J7I6J9K4L5K5J5L4L4L4L5K4CiVn2"}, "label": "zebra facing camera"}]}
{"id": 447934, "image": "COCO_train2014_000000447934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant leading the way\"."}], "task": "refering", "answer_template": "The \"the elephant leading the way\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 228, 229, 244, 245, 246, 248, 249, 251, 252, 271, 272, 275], "bbox": [0.611796875, 0.298875, 0.994375, 0.6719166666666666], "iscrowd": 0, "area": 27838.7284, "rle": {"size": [480, 640], "counts": "Yog58f>:G8VCWOm:R1kDoNS;Z1fDfNX;c1`D^N^;k1[DUNc;T2UDmMi;\\2PDeMm;h2N2N2N2O2M2N2N2N2N3M2N2N2N2N3N1N2N2N2N3M2NM3L4K5L4O2010O010O10O01000O0100O0100O0100O010O10O1=B`0A5J010O10000O100O10gNPFbMo9_2SF_Mm9`2WF\\Mj9d2YFYMf9h2]FUMc9j2`FTM_9m2dFPM\\9P3gFmLX9S3lFjLT9V3nFhLR9X3QGeLn8[3VGbLj8^3YG^Lg8c3[1O01000O0100O01000O010O10O2O0O10000O10000O10YF`Ln7`3QHaLo7^3QHcLo7]3PHdLP8\\3nGgLQ8X3oGiLQ8W3nGjLR8V3mGkLS8T3mGmLS8S3lGnLT8R3jGPMV8o2jGSMU8m2jGTMW8k2hGVMX8k2gGUMY8o2bGRM^8R3]GoLc8U3XGmLg8W3UGiLk8\\3oFeLQ9_3jFbLV9b3fF^LZ9f3aF[L_9d4O00000000001O00000001O0001O00000000001O00000000001O000000000010G8@`0@`0@`0@`0A?00000000001O0000000000000001O0000000000000001O2N1O2N2N2N1O2N2N1O2N2N101N2N1O2N2N1O200O010O10O10O10O10O10O10O10O010000O0h0YOQ1oNQ1nNa0@O10O10O10O0100O010OO2N2N1O2N1N3N1O2NQb1"}, "label": "the elephant leading the way"}]}
{"id": 447934, "image": "COCO_train2014_000000447934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant walking in the grass\"."}], "task": "refering", "answer_template": "The \"an elephant walking in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 228, 229, 244, 245, 246, 248, 249, 251, 252, 271, 272, 275], "bbox": [0.611796875, 0.298875, 0.994375, 0.6719166666666666], "iscrowd": 0, "area": 27838.7284, "rle": {"size": [480, 640], "counts": "Yog58f>:G8VCWOm:R1kDoNS;Z1fDfNX;c1`D^N^;k1[DUNc;T2UDmMi;\\2PDeMm;h2N2N2N2O2M2N2N2N2N3M2N2N2N2N3N1N2N2N2N3M2NM3L4K5L4O2010O010O10O01000O0100O0100O0100O010O10O1=B`0A5J010O10000O100O10gNPFbMo9_2SF_Mm9`2WF\\Mj9d2YFYMf9h2]FUMc9j2`FTM_9m2dFPM\\9P3gFmLX9S3lFjLT9V3nFhLR9X3QGeLn8[3VGbLj8^3YG^Lg8c3[1O01000O0100O01000O010O10O2O0O10000O10000O10YF`Ln7`3QHaLo7^3QHcLo7]3PHdLP8\\3nGgLQ8X3oGiLQ8W3nGjLR8V3mGkLS8T3mGmLS8S3lGnLT8R3jGPMV8o2jGSMU8m2jGTMW8k2hGVMX8k2gGUMY8o2bGRM^8R3]GoLc8U3XGmLg8W3UGiLk8\\3oFeLQ9_3jFbLV9b3fF^LZ9f3aF[L_9d4O00000000001O00000001O0001O00000000001O00000000001O000000000010G8@`0@`0@`0@`0A?00000000001O0000000000000001O0000000000000001O2N1O2N2N2N1O2N2N1O2N2N101N2N1O2N2N1O200O010O10O10O10O10O10O10O10O010000O0h0YOQ1oNQ1nNa0@O10O10O10O0100O010OO2N2N1O2N1N3N1O2NQb1"}, "label": "an elephant walking in the grass"}]}
{"id": 570822, "image": "COCO_train2014_000000570822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench\"."}], "task": "refering", "answer_template": "The \"a wooden bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 311, 325, 326, 327, 328, 341, 342, 343, 344, 345, 346, 358, 359, 360, 361, 362, 363, 364, 376, 377, 378, 379, 380, 381], "bbox": [0.09713389121338913, 0.785828125, 0.4635564853556486, 1.0], "iscrowd": 0, "area": 14639.872949999997, "rle": {"size": [640, 478], "counts": "eYm01oc07H7I8I6I4M0O2N2O0O2O1N1O2O1N101N2N101N2N1cN]N]@d1D[Nm>5[Ab1GYNm>8ZA`1HXNn>:WA`1JVNn>=UA_1KUNP??RA]1MTNP?b0PA\\1ORNR?c0m@[11RNR?f0i@Z14PNT?g0e@[15oMV?h0c@Z16nMX?i0_@[18lMY?l0\\@Y1:kM[?m0X@Z1;jM]?n0U@Z1=hM_?\\3`@dL`?\\3`@eL`?[3_@eLa?\\3]@eLd?[3[@eLe?\\3Z@dLg?\\3X@dLi?\\3V@dLj?]3T@dLm?[3S@eLm?\\3R@dLo?\\3P@dLP`0]3n_OdLS`0d3001O001O001O1O001O001O001O001O001O001O001O1O000001O000000000O100000000000000O1000000000000O100000000000000O1000O1000000000O1000SMf_OT2Z`0lMf_OT2Z`0kMh_OT2X`0kMj_OT2V`0lMk_OS2T`0mMm_OS2S`0lMo_OS2Q`0mMP@R2P`0mMQ@S2o?gMX@X2i`0O1O001O1O1O001O1N2O001O1O1O001O1O1O001O1O1O001`NV^Oh0ka0TOZ^Oj0ga0RO]^Om0ca0QOa^Om0`a0oNd^OP1]a0lNh^OR1Tb0O2N1O2N1N3N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1OUPP5"}, "label": "a wooden bench"}]}
{"id": 570822, "image": "COCO_train2014_000000570822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table leaning against a white wall\"."}], "task": "refering", "answer_template": "The \"a table leaning against a white wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 311, 325, 326, 327, 328, 341, 342, 343, 344, 345, 346, 358, 359, 360, 361, 362, 363, 364, 376, 377, 378, 379, 380, 381], "bbox": [0.09713389121338913, 0.785828125, 0.4635564853556486, 1.0], "iscrowd": 0, "area": 14639.872949999997, "rle": {"size": [640, 478], "counts": "eYm01oc07H7I8I6I4M0O2N2O0O2O1N1O2O1N101N2N101N2N1cN]N]@d1D[Nm>5[Ab1GYNm>8ZA`1HXNn>:WA`1JVNn>=UA_1KUNP??RA]1MTNP?b0PA\\1ORNR?c0m@[11RNR?f0i@Z14PNT?g0e@[15oMV?h0c@Z16nMX?i0_@[18lMY?l0\\@Y1:kM[?m0X@Z1;jM]?n0U@Z1=hM_?\\3`@dL`?\\3`@eL`?[3_@eLa?\\3]@eLd?[3[@eLe?\\3Z@dLg?\\3X@dLi?\\3V@dLj?]3T@dLm?[3S@eLm?\\3R@dLo?\\3P@dLP`0]3n_OdLS`0d3001O001O001O1O001O001O001O001O001O001O001O1O000001O000000000O100000000000000O1000000000000O100000000000000O1000O1000000000O1000SMf_OT2Z`0lMf_OT2Z`0kMh_OT2X`0kMj_OT2V`0lMk_OS2T`0mMm_OS2S`0lMo_OS2Q`0mMP@R2P`0mMQ@S2o?gMX@X2i`0O1O001O1O1O001O1N2O001O1O1O001O1O1O001O1O1O001`NV^Oh0ka0TOZ^Oj0ga0RO]^Om0ca0QOa^Om0`a0oNd^OP1]a0lNh^OR1Tb0O2N1O2N1N3N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1OUPP5"}, "label": "a table leaning against a white wall"}]}
{"id": 505288, "image": "COCO_train2014_000000505288.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 72, 73, 74, 75, 95, 96, 97, 98, 99, 121, 122, 123, 124, 125, 145, 146, 147, 148, 149, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 265, 266, 285, 286, 288, 289, 290, 308, 309, 311, 312, 313, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 359, 377, 378, 379, 381], "bbox": [0.12303125, 0.13708333333333333, 0.613484375, 0.9617916666666667], "iscrowd": 0, "area": 33964.68280000001, "rle": {"size": [480, 640], "counts": "lTU12m>2M2O2M3N2M3N2M2O2M3N2M3N1N101N101N101N2N1O2N1O2M3N1O2M2O2N1O2M3O0100O100O10000O100O100O2O2N1N2O1N2O2N1N2O1N2O1N10001O01O1O100O001O100O1O1O100O1O1O100O1O12N2N2N2M2O1OO10000O10000O10000O100O10000O01000O10000O10000O100O10000O10O10O10O10O10O010O10O10O10O10O10O10O010O10O10O10O10O10O00100O001O10O01O010O1O00100O00100eC]NW;c1gDbNV;_1fDhNV;X1hDmNV;S1gDSOU;n0gDXOV;h0hD^OT;c0iDCT;=iDHT;8iDOS;2jD3T;LjD:R;n1L3M4M2aJhK0\\4IgK7]4BgK>[4\\OhKd0\\4TOhKl0[4nNhKR1\\4gNgKZ1\\4_NgKa1\\4YNgKg1]4QNhKn1[4lMhKU2[4dMhK\\2\\4]MgKc2\\4WMgKj2\\4nLhKR3[4hLhKX3Q8O01O00001O0001O01O0000001O01O01O0000001O01O01O1O1O0fJfJn0[5QORMbNo2]1[5O1O1O1O1O001N2O1O1O1O1O1O1O001O1O1O1N2O1O\\IeLd1Z3YNjLf1V3WNnLi1Q3TNSMl1l2QNXMo1g2nM]MR2b2kMbMU2]2hMgMX2X2eMkM\\2T2aMPN_2o1^MUNa2j1]MXNe2g1XMZNk2e1RM[NR3d1kL]NX3b1eL^N_3a1^L`Ne3`1WLaNl3`1oK`NU4a1fK`N\\4b1_K_Nd4b1WK^Nm4c1nJ^NT5d1gJ]N\\5e1^J[Nf5f1UJ[Nm5g1nIZNU6g1fIYN^6h1]IYNe6i1VIXNm6j1mHVNV7n1cHSN`7o1ZHRNi7Q2PHoMS8U2fGlM]8W2\\GjMg8X2SGhMP9R43L4M2N3L3N3M3L3N3M3L3N3M3L3N3M2M42N8gG[Ke5n4QJTKo5S5gIPKY6X5\\IkJd6]5RIdJS7_5cHdJ`7a5UHbJo7b5gG`J\\8i1kGd1X8QNQHn1S8eMWHZ2l7[M]Hd2g7PMaHP3b7dLhHZ3\\7[LmHa3\\9M3M2N3L4\\N[DHh;FVEKm:CQFNQ:AmFOWgc3"}, "label": "the giraffe on the left"}]}
{"id": 554452, "image": "COCO_train2014_000000554452.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy wearing a london sweatshirt playing with a windsock walking down a cobblestone street\"."}], "task": "refering", "answer_template": "The \"a young boy wearing a london sweatshirt playing with a windsock walking down a cobblestone street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 67, 68, 69, 82, 83, 84, 95, 96, 97, 98, 99, 100, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 148, 149, 156, 157, 158, 159, 171, 172, 173, 174, 186, 187, 188, 189, 201, 202, 203, 204, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 279, 291, 292, 293, 294, 307, 308, 309], "bbox": [0.13384976525821596, 0.132859375, 0.965962441314554, 0.899578125], "iscrowd": 0, "area": 51190.845850000005, "rle": {"size": [640, 426], "counts": "VkS11nc0101N3M4L3M4M2M1O1O1O1O2O0O1O1O1O100O1O1O2N100O1O1O1O100O1O2O0000000O100000001O0O10000000000O10001O00000O10000000001N1000000000O0100000O1000O01000000O10O1000O100000O010000O1000O10O1000000O01000O0100O1O00101N1O1O2O0O1O2N101N1g^O[NV`0f1e_OhNQ`0Z1j_OSOm?m0n_O@i?b0R@Kd?6W@8`?\\2H9G9F9H9G9G9F:G9G9F:G9G9G9F:G9G?Ae0ZOf0[Od0\\O3M0O2O1O0O2N1O2M3M2O2M2O2M3N1N3M2mNT1G8J7N1O2N1O2O1N2N2N3M2N2N3M2N2N3M2N2N2O2M10000O1000001N1000000O1000000O10000O100010O1O1O1M3L3L5J6`CgJ`9Q6_FoIk8h6SGYIV8_7hGcHa7T8^HlGk6l8SIUGW6b9gI`Fn5k9QJUFf5U:XJlE_5]:^JgEZ5a:_JeE[5b:^JdE\\5j:VJ]Ec5Q;oIUEU4mNQMh=g2mBeLY=T3]CWLi<b3o1H7I7I8I6N2O2M2O1N3N1N3N1N2O2M2O1N3N1O1N3N1N201O001O1O1O1O2N100O1O2N1O1O1O1O2N1O1O101N1O1O00000000O10001O0000000000000O10001O0000000000000O11O0000001O000000001O0000001O0000001O0000001O0000001O0001O2O1N2N2N2N010O000100O100O00100O100O1O010O100O10O01O100O100N1K6J6I7J6JoT9"}, "label": "a young boy wearing a london sweatshirt playing with a windsock walking down a cobblestone street"}]}
{"id": 554452, "image": "COCO_train2014_000000554452.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy holding a kite\"."}], "task": "refering", "answer_template": "The \"a boy holding a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 67, 68, 69, 82, 83, 84, 95, 96, 97, 98, 99, 100, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 148, 149, 156, 157, 158, 159, 171, 172, 173, 174, 186, 187, 188, 189, 201, 202, 203, 204, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 279, 291, 292, 293, 294, 307, 308, 309], "bbox": [0.13384976525821596, 0.132859375, 0.965962441314554, 0.899578125], "iscrowd": 0, "area": 51190.845850000005, "rle": {"size": [640, 426], "counts": "VkS11nc0101N3M4L3M4M2M1O1O1O1O2O0O1O1O1O100O1O1O2N100O1O1O1O100O1O2O0000000O100000001O0O10000000000O10001O00000O10000000001N1000000000O0100000O1000O01000000O10O1000O100000O010000O1000O10O1000000O01000O0100O1O00101N1O1O2O0O1O2N101N1g^O[NV`0f1e_OhNQ`0Z1j_OSOm?m0n_O@i?b0R@Kd?6W@8`?\\2H9G9F9H9G9G9F:G9G9F:G9G9G9F:G9G?Ae0ZOf0[Od0\\O3M0O2O1O0O2N1O2M3M2O2M2O2M3N1N3M2mNT1G8J7N1O2N1O2O1N2N2N3M2N2N3M2N2N3M2N2N2O2M10000O1000001N1000000O1000000O10000O100010O1O1O1M3L3L5J6`CgJ`9Q6_FoIk8h6SGYIV8_7hGcHa7T8^HlGk6l8SIUGW6b9gI`Fn5k9QJUFf5U:XJlE_5]:^JgEZ5a:_JeE[5b:^JdE\\5j:VJ]Ec5Q;oIUEU4mNQMh=g2mBeLY=T3]CWLi<b3o1H7I7I8I6N2O2M2O1N3N1N3N1N2O2M2O1N3N1O1N3N1N201O001O1O1O1O2N100O1O2N1O1O1O1O2N1O1O101N1O1O00000000O10001O0000000000000O10001O0000000000000O11O0000001O000000001O0000001O0000001O0000001O0000001O0001O2O1N2N2N2N010O000100O100O00100O100O1O010O100O10O01O100O100N1K6J6I7J6JoT9"}, "label": "a boy holding a kite"}]}
{"id": 54743, "image": "COCO_train2014_000000054743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair with a guy sitting on it\"."}], "task": "refering", "answer_template": "The \"a brown chair with a guy sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 227, 250, 273, 296, 318, 319, 338, 339, 340, 341, 342, 361, 362, 363, 364, 365, 384, 387, 388, 407, 410, 411, 430, 433, 456], "bbox": [0.69990625, 0.36974999999999997, 0.9156093750000001, 0.8479375000000001], "iscrowd": 0, "area": 8856.860800000011, "rle": {"size": [640, 640], "counts": "Q_h85ac09H:E<D<E;D<E;N2M3N2M3M3N0O100O100O10aMZ_Od1e`0]NS@k0m?UOj@4U?Nm@OS?1o@MP?4SAIm>7VAFi>;YABh>>[A_Oe>a0^A\\Oa>f0aAWO_>i0[2100O100O101O0O1000000000001N1000000000001O00000O100000001O000000000O101O000000000000001O0O10000000001N1O1O100O1O1O101N1O1O1O100Q_O9d=HPBd0o=^OeAm0Z>TO[AW1d>S2O1O100O1N2MK6H8H7XOi02M3N2M3YL[LlFl3T9iLQF]3o9iLeE]3[:iLXE^3h:iLiDYO\\O?V1^Oe:P1YDVO86e1jNj9T3mDlMb=m1kBlMV=n1UClMl<m1`ClMb<m1jCmMW<l1UDnMl;k1X3I7I7I7I7I7J7O0000000000:FX`Q1"}, "label": "a brown chair with a guy sitting on it"}]}
{"id": 54743, "image": "COCO_train2014_000000054743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair with a man sitting on it\"."}], "task": "refering", "answer_template": "The \"a wooden chair with a man sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 227, 250, 273, 296, 318, 319, 338, 339, 340, 341, 342, 361, 362, 363, 364, 365, 384, 387, 388, 407, 410, 411, 430, 433, 456], "bbox": [0.69990625, 0.36974999999999997, 0.9156093750000001, 0.8479375000000001], "iscrowd": 0, "area": 8856.860800000011, "rle": {"size": [640, 640], "counts": "Q_h85ac09H:E<D<E;D<E;N2M3N2M3M3N0O100O100O10aMZ_Od1e`0]NS@k0m?UOj@4U?Nm@OS?1o@MP?4SAIm>7VAFi>;YABh>>[A_Oe>a0^A\\Oa>f0aAWO_>i0[2100O100O101O0O1000000000001N1000000000001O00000O100000001O000000000O101O000000000000001O0O10000000001N1O1O100O1O1O101N1O1O1O100Q_O9d=HPBd0o=^OeAm0Z>TO[AW1d>S2O1O100O1N2MK6H8H7XOi02M3N2M3YL[LlFl3T9iLQF]3o9iLeE]3[:iLXE^3h:iLiDYO\\O?V1^Oe:P1YDVO86e1jNj9T3mDlMb=m1kBlMV=n1UClMl<m1`ClMb<m1jCmMW<l1UDnMl;k1X3I7I7I7I7I7J7O0000000000:FX`Q1"}, "label": "a wooden chair with a man sitting on it"}]}
{"id": 54743, "image": "COCO_train2014_000000054743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair with no one in it\"."}], "task": "refering", "answer_template": "The \"the chair with no one in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376, 393, 394, 395, 396, 397, 398, 399, 417, 418, 419, 420, 440, 441], "bbox": [0.05915625, 0.355421875, 0.376421875, 0.86328125], "iscrowd": 0, "area": 40905.38905, "rle": {"size": [640, 640], "counts": "hog0?[c07N101O001O1O002N4L4L3L5L4L4L4L4L4L4L4L4L4L4K5L3M4L4L4L4L4L4L4ZChL\\8\\3oFVMo8o2\\FbMd9c2fEnMZ:[2lDUNU;S2RD^Nn;j40000000O10O1000000000O100000000O001O1O1O1O1O1N2O1O1OjEaHl7_7VG_Ik8`6VF`Jj9k7000000000000000O0100000000000000000O010000000M3H8H8G9I70O10O1000000O100000000O010000000O100000000O0100000O100000000O1000O1000O1000000O100000O01000000[HkD_6U;`IWET6j:lIaEi5_:WJmE]5S:bJYFS5g9mJdFg4\\9YKQG[4o8eKTGX4[<O1N101O1O001N2O1O001N2O001O1N101O1O0O2O1O001O1N101O1O0O2O1O1O0O2O1O001N2O001O1O0O2O1O001N2QMa@[1_?bN`Ab0a>ZO_BIb=3W3N3L4L4M2M]Pi7"}, "label": "the chair with no one in it"}]}
{"id": 54743, "image": "COCO_train2014_000000054743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty wooden chair\"."}], "task": "refering", "answer_template": "The \"an empty wooden chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376, 393, 394, 395, 396, 397, 398, 399, 417, 418, 419, 420, 440, 441], "bbox": [0.05915625, 0.355421875, 0.376421875, 0.86328125], "iscrowd": 0, "area": 40905.38905, "rle": {"size": [640, 640], "counts": "hog0?[c07N101O001O1O002N4L4L3L5L4L4L4L4L4L4L4L4L4L4K5L3M4L4L4L4L4L4L4ZChL\\8\\3oFVMo8o2\\FbMd9c2fEnMZ:[2lDUNU;S2RD^Nn;j40000000O10O1000000000O100000000O001O1O1O1O1O1N2O1O1OjEaHl7_7VG_Ik8`6VF`Jj9k7000000000000000O0100000000000000000O010000000M3H8H8G9I70O10O1000000O100000000O010000000O100000000O0100000O100000000O1000O1000O1000000O100000O01000000[HkD_6U;`IWET6j:lIaEi5_:WJmE]5S:bJYFS5g9mJdFg4\\9YKQG[4o8eKTGX4[<O1N101O1O001N2O1O001N2O001O1N101O1O0O2O1O001O1N101O1O0O2O1O1O0O2O1O001N2O001O1O0O2O1O001N2QMa@[1_?bN`Ab0a>ZO_BIb=3W3N3L4L4M2M]Pi7"}, "label": "an empty wooden chair"}]}
{"id": 161242, "image": "COCO_train2014_000000161242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a yellow shirt , sitting down and holding a birthday cake\"."}], "task": "refering", "answer_template": "The \"a woman in a yellow shirt , sitting down and holding a birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 38, 39, 40, 41, 42, 53, 54, 55, 56, 57, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 117, 118, 119, 126, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 171, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 215, 216, 217, 218, 219, 220, 221, 222, 223, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299], "bbox": [0.2780470588235294, 0.0489375, 0.9979294117647058, 0.8483437500000001], "iscrowd": 0, "area": 101806.62504999996, "rle": {"size": [640, 425], "counts": "RVZ23mc0;F:E:F;E8ION2O2M2O1N3M2O2M2O1N3N1N2O2M2O1N3N1N3M2O1N3N1N2O2M2O1N3N1N3N1N2O2M2N2O2M2O1N3N1N2O2M2O2M2O1iJcLjI_3h5SMRJn2m5YMlIh2R6`MfIc2X6cMbI^2]6hM]IY2b6mMXIU2f6QNTIP2k6VNnHm1P7YNjHh1U7_NdHb1[7dN_H^1_7hN[HY1d7oNSHS1l7XOiGj0U8A`G`0_8JWG7h8b5N1O2N1O1O2N1N2O2O0O100O2O000O100O101N100O100O101N100O100O101N100O100O100O1J6H8H8G9H8H8I7M3M3N2M3M3M3M3N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O10000000000000000000000000000000000000000000000000000000000000001O0000000000000000000001O001O1O001O001O001O1O001O001O001O1O001O001O1O003M2N3M3M3M3M2N3M3M3L3N3M3M3M3M2N3M4L8H9G9G8H9G9G8Hg0YO1O001N101O1O001PKRGlNo8o0WGoNi8POSGnM9Q3e8kNjG]MGf3`8fNeIY1[6bNlI\\1U6]NRJb1o5XNXJf1h5UN^Jj1c5oMeJo1[5lMkJS2V5fMRKY2n4aMXK^2h4]M_Ka2b4XMfKiN[KZ3o8gMlKkNXK^3l8aMfL_2Z3[MlLd2W3TMoLk2S3mLTMS3m2fLYMY3h9O100O1O010O1O101N1O101N1O101N100O1H9Lj1VN8H7I8Hi3"}, "label": "a woman in a yellow shirt , sitting down and holding a birthday cake"}]}
{"id": 161242, "image": "COCO_train2014_000000161242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl sitting holding cake\"."}], "task": "refering", "answer_template": "The \"girl sitting holding cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 38, 39, 40, 41, 42, 53, 54, 55, 56, 57, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 101, 102, 103, 111, 112, 113, 114, 115, 116, 117, 118, 119, 126, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 171, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 215, 216, 217, 218, 219, 220, 221, 222, 223, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299], "bbox": [0.2780470588235294, 0.0489375, 0.9979294117647058, 0.8483437500000001], "iscrowd": 0, "area": 101806.62504999996, "rle": {"size": [640, 425], "counts": "RVZ23mc0;F:E:F;E8ION2O2M2O1N3M2O2M2O1N3N1N2O2M2O1N3N1N3M2O1N3N1N2O2M2O1N3N1N3N1N2O2M2N2O2M2O1N3N1N2O2M2O2M2O1iJcLjI_3h5SMRJn2m5YMlIh2R6`MfIc2X6cMbI^2]6hM]IY2b6mMXIU2f6QNTIP2k6VNnHm1P7YNjHh1U7_NdHb1[7dN_H^1_7hN[HY1d7oNSHS1l7XOiGj0U8A`G`0_8JWG7h8b5N1O2N1O1O2N1N2O2O0O100O2O000O100O101N100O100O101N100O100O101N100O100O100O1J6H8H8G9H8H8I7M3M3N2M3M3M3M3N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2N2O1N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O10000000000000000000000000000000000000000000000000000000000000001O0000000000000000000001O001O1O001O001O001O1O001O001O001O1O001O001O1O003M2N3M3M3M3M2N3M3M3L3N3M3M3M3M2N3M4L8H9G9G8H9G9G8Hg0YO1O001N101O1O001PKRGlNo8o0WGoNi8POSGnM9Q3e8kNjG]MGf3`8fNeIY1[6bNlI\\1U6]NRJb1o5XNXJf1h5UN^Jj1c5oMeJo1[5lMkJS2V5fMRKY2n4aMXK^2h4]M_Ka2b4XMfKiN[KZ3o8gMlKkNXK^3l8aMfL_2Z3[MlLd2W3TMoLk2S3mLTMS3m2fLYMY3h9O100O1O010O1O101N1O101N1O101N100O1H9Lj1VN8H7I8Hi3"}, "label": "girl sitting holding cake"}]}
{"id": 38365, "image": "COCO_train2014_000000038365.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep that has just given birth in a grassy field\"."}], "task": "refering", "answer_template": "The \"a sheep that has just given birth in a grassy field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 82, 83, 84, 85, 86, 88, 89, 90, 91, 92, 93, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 229, 230, 232, 233, 234, 242, 243, 245, 246, 247, 248, 249, 250, 258, 259, 261, 262, 263, 264, 265, 266, 274, 275], "bbox": [0.06739696312364425, 0.14646874999999998, 0.975466377440347, 0.76353125], "iscrowd": 0, "area": 89511.49449999994, "rle": {"size": [640, 461], "counts": "nbc0e0Vc0;F7I9g]OeN`a0T2I5K4L4M3L4L4M3L5K6K5P@dLR?^4G9F:G6I5L4K5L4K4N2M2O2M2O1N2O1N2N101N101N3N3L3N2M3M6K6I6K6I6K5J5L4K4L6K=aEiFZ9`9YFiFc9m9L100O2O0O10001N100OgMPGeIo8n4mFnIh0i0[8[5[GYIa0Y1T8^5mH_JS7h5iHUJW7l5kHPJV7Q6kHmIT7T6nHjIR7U6QIiIo6W6RIhIn6X6TIeIm6Y6WIeIh6[6[IbIf6^6^I]Ic6c6bIWI_6i6eIRI\\6n6[31O000001O0000000O100000001O0001O001O001O001O001O001O00001O000000000001O00UCTIb<Y7L4M2M7I:F101N2N4M5J4L4L5L3L3M6J<E1N00O101O0000001O00000eM]FYJd9a5fF[JY9d5nFVJR9i5VGPJj8o5]GkId8S6aGiI_8V6eGgI[8X6hGgIX8X6kGeIU8[6lGdIT8]6mGbIR8^6UH[Il7d6XHXIh7h6[HVIe7j6\\HTId7l6^HRIb7n6i20OO101N101O00001O0O101O000O2O0OUEgIR8V6dGUJ]8i5cGYJ]8g5bGZJ^8f5aGZJ`8e5_G]Ja8c5]G^Jd8b5[G^Jf8c5XG^Jh8b5WG^Jj8c5TG^Jl8b5RG`Jn8a5oF`JQ9b5lF`JT9a5iFaJW9`5fFbJZ9^5dFdJ\\9]5aFdJ`9\\5]FgJc9Y5XFlJg9T5XFnJg9S5WFoJg9R5XFPKg9l4[FVKe9f4^FVKg9l4UFnJS:T5hElJ[:U5aEmJ`:U5WEPKl:o62O1N3M2O1N3M2N2O1N3N100O2O0O2O001N2N1O2O1N2M2YOkCTI40W<k6c0O1O0010O01O001O001O0O2N2N1O2O10O104mBeIY<a6]CeIa<P7gCiHj1Ka7_7^FZIY1_OX8a8YGeGg8b8gFiGY9_9O10000O10O10O01000O010O1OeNjFnGV9l7a1M3M4L3VOj0L5M200100O010SO_HZEb7b:eHZE[7g:hHUEY7k:iHREW7n:lHoDU7Q;mHlDS7U;oHhDQ7Y;QIdDP7\\;i01O1O100O100O1O101N1O1O1O2N1N2L5K4O1N2O1O100O10000O100000000O2O000O1000O0100O100O1O1O1O001O100O1O1O1O1O1N2O1N2O1N2N2N2M3L4L4L5K4M3K5K5L5E:F:F;K4O1O2M2N3M3L4K5E:B?L4N2N1^NV_OMk`0Ma_OIc`02h_OUO^OMm`0l0g1N2O1000O10O1000O100O001000O010O100O010O100O1O100O100O100O100O1O100O1O2N1O100O2N2N3M2Kih6"}, "label": "a sheep that has just given birth in a grassy field"}]}
{"id": 38365, "image": "COCO_train2014_000000038365.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white big sheep\"."}], "task": "refering", "answer_template": "The \"a white big sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 82, 83, 84, 85, 86, 88, 89, 90, 91, 92, 93, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 229, 230, 232, 233, 234, 242, 243, 245, 246, 247, 248, 249, 250, 258, 259, 261, 262, 263, 264, 265, 266, 274, 275], "bbox": [0.06739696312364425, 0.14646874999999998, 0.975466377440347, 0.76353125], "iscrowd": 0, "area": 89511.49449999994, "rle": {"size": [640, 461], "counts": "nbc0e0Vc0;F7I9g]OeN`a0T2I5K4L4M3L4L4M3L5K6K5P@dLR?^4G9F:G6I5L4K5L4K4N2M2O2M2O1N2O1N2N101N101N3N3L3N2M3M6K6I6K6I6K5J5L4K4L6K=aEiFZ9`9YFiFc9m9L100O2O0O10001N100OgMPGeIo8n4mFnIh0i0[8[5[GYIa0Y1T8^5mH_JS7h5iHUJW7l5kHPJV7Q6kHmIT7T6nHjIR7U6QIiIo6W6RIhIn6X6TIeIm6Y6WIeIh6[6[IbIf6^6^I]Ic6c6bIWI_6i6eIRI\\6n6[31O000001O0000000O100000001O0001O001O001O001O001O001O00001O000000000001O00UCTIb<Y7L4M2M7I:F101N2N4M5J4L4L5L3L3M6J<E1N00O101O0000001O00000eM]FYJd9a5fF[JY9d5nFVJR9i5VGPJj8o5]GkId8S6aGiI_8V6eGgI[8X6hGgIX8X6kGeIU8[6lGdIT8]6mGbIR8^6UH[Il7d6XHXIh7h6[HVIe7j6\\HTId7l6^HRIb7n6i20OO101N101O00001O0O101O000O2O0OUEgIR8V6dGUJ]8i5cGYJ]8g5bGZJ^8f5aGZJ`8e5_G]Ja8c5]G^Jd8b5[G^Jf8c5XG^Jh8b5WG^Jj8c5TG^Jl8b5RG`Jn8a5oF`JQ9b5lF`JT9a5iFaJW9`5fFbJZ9^5dFdJ\\9]5aFdJ`9\\5]FgJc9Y5XFlJg9T5XFnJg9S5WFoJg9R5XFPKg9l4[FVKe9f4^FVKg9l4UFnJS:T5hElJ[:U5aEmJ`:U5WEPKl:o62O1N3M2O1N3M2N2O1N3N100O2O0O2O001N2N1O2O1N2M2YOkCTI40W<k6c0O1O0010O01O001O001O0O2N2N1O2O10O104mBeIY<a6]CeIa<P7gCiHj1Ka7_7^FZIY1_OX8a8YGeGg8b8gFiGY9_9O10000O10O10O01000O010O1OeNjFnGV9l7a1M3M4L3VOj0L5M200100O010SO_HZEb7b:eHZE[7g:hHUEY7k:iHREW7n:lHoDU7Q;mHlDS7U;oHhDQ7Y;QIdDP7\\;i01O1O100O100O1O101N1O1O1O2N1N2L5K4O1N2O1O100O10000O100000000O2O000O1000O0100O100O1O1O1O001O100O1O1O1O1O1N2O1N2O1N2N2N2M3L4L4L5K4M3K5K5L5E:F:F;K4O1O2M2N3M3L4K5E:B?L4N2N1^NV_OMk`0Ma_OIc`02h_OUO^OMm`0l0g1N2O1000O10O1000O100O001000O010O100O010O100O1O100O100O100O100O1O100O1O2N1O100O2N2N3M2Kih6"}, "label": "a white big sheep"}]}
{"id": 87518, "image": "COCO_train2014_000000087518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat being held by a man with a helmet\"."}], "task": "refering", "answer_template": "The \"a black cat being held by a man with a helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 148, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 211, 212, 213, 215], "bbox": [0.41860416666666667, 0.372015625, 0.7254375000000001, 0.563890625], "iscrowd": 0, "area": 8066.707049999997, "rle": {"size": [640, 480], "counts": "Qom32jc05K5M3L4L4N2M1O101O0O2O0XNTO]@l0b?XO[@i0d?ZOY@e0g?_OV@`0j?DS@:m?KP@4P`00m_ONS`08i_OFW`0?f_O@\\`0b0a_O^O``0d0]_O\\Oe`0e0X_O[Oi`0h0R_OYOo`0j0l^OWOTa0m0f^OUO[a0h10O1O0010O001O0O2N1O2O04L5K2N3M2O2Na^O^Na`0a1]_ObNc`0]1\\_OfNc`0Y1[_OiNf`0V1X_OmNh`0Q1W_OQOj`0m0U_OUOl`0i0S_OYOn`0e0P_O^OQa0Y1S_OhMQa0T2`0N2O1N2N1O2N2N2N2N2N1O2N1O100O0Hg]OjNZb0U1h]OiNXb0V1j]OiNWb0V1j]OiNVb0X18000O21O1O2N2N2NN2N2N3N02O1N2O1O1O1O1O1O1O1O2N101N2N1000O10000O3N1O1N9H9F1O2O0O10O0N2M3O11O000O2L3[Of0G8M4L3N3N1O0OO2O011ROn^OmNSa0n0Y_OiNh`0V1^_OeNc`0Y1b_OcN^`0\\1g_O_N[`0_1j_O]NW`0a1T1N3M3M3L4AlXb2"}, "label": "a black cat being held by a man with a helmet"}]}
{"id": 87518, "image": "COCO_train2014_000000087518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a photo of a man holding a cat while wearing a football helmet\"."}], "task": "refering", "answer_template": "The \"a photo of a man holding a cat while wearing a football helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 92, 93, 94, 95, 109, 110, 111, 112, 126, 127, 128, 142, 143, 144, 145, 146, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266], "bbox": [0.38554166666666667, 0.19365625, 0.7058541666666667, 0.690078125], "iscrowd": 0, "area": 32910.041849999994, "rle": {"size": [640, 480], "counts": "klc3=Qc0j0\\Od0G9G9F:G9G6J5K4L5L6I:G8G:F9H8G:G8G:G8^MhIkFX6T9RJdFn5Z9\\J]Fe5a9dJXF\\5f9nJQFS5m9WKkEi4R:cKdE^4Z:lK^ES4a:PL]EQ4a:QL_Eo3_:TL_Em3^:VLbEj3\\:YLbEh3\\:ZLdEe3\\:]LbEd3]:]LcEc3\\:_LbEb3\\:`LdE`3[:S3O1O0O2O1O10000000O1000000000000O100000O1000O1000000000000O2O001O001O001N101O001O001O0O2O001O001O0O4M2N3M3M3M2N1N2[IkD]4W;YKSEe4n:QK]Em4d:jJfET5\\:bJmE]5T:ZJVFd5k9SJ_Fj5n;N2N2N2N2N2M2O2N2N2N2N2N2M2O1O1O1O2N1O1N2O2N1O1O1O2N1N2O1O2N1O1O1H9D;F:E;F:E<E:L4L6K6I7J5J7J5J7J5J7J5Jkig2"}, "label": "a photo of a man holding a cat while wearing a football helmet"}]}
{"id": 87518, "image": "COCO_train2014_000000087518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a fully brown dog\"."}], "task": "refering", "answer_template": "The \"a man holding a fully brown dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 92, 93, 94, 95, 109, 110, 111, 112, 126, 127, 128, 142, 143, 144, 145, 146, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266], "bbox": [0.38554166666666667, 0.19365625, 0.7058541666666667, 0.690078125], "iscrowd": 0, "area": 32910.041849999994, "rle": {"size": [640, 480], "counts": "klc3=Qc0j0\\Od0G9G9F:G9G6J5K4L5L6I:G8G:F9H8G:G8G:G8^MhIkFX6T9RJdFn5Z9\\J]Fe5a9dJXF\\5f9nJQFS5m9WKkEi4R:cKdE^4Z:lK^ES4a:PL]EQ4a:QL_Eo3_:TL_Em3^:VLbEj3\\:YLbEh3\\:ZLdEe3\\:]LbEd3]:]LcEc3\\:_LbEb3\\:`LdE`3[:S3O1O0O2O1O10000000O1000000000000O100000O1000O1000000000000O2O001O001O001N101O001O001O0O2O001O001O0O4M2N3M3M3M2N1N2[IkD]4W;YKSEe4n:QK]Em4d:jJfET5\\:bJmE]5T:ZJVFd5k9SJ_Fj5n;N2N2N2N2N2M2O2N2N2N2N2N2M2O1O1O1O2N1O1N2O2N1O1O1O2N1N2O1O2N1O1O1H9D;F:E;F:E<E:L4L6K6I7J5J7J5J7J5J7J5Jkig2"}, "label": "a man holding a fully brown dog"}]}
{"id": 87518, "image": "COCO_train2014_000000087518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange cat sitting in front of television on a stand\"."}], "task": "refering", "answer_template": "The \"orange cat sitting in front of television on a stand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 232, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 317, 318, 319, 320, 321], "bbox": [0.6479374999999999, 0.52896875, 0.8970416666666666, 0.8285625], "iscrowd": 0, "area": 16231.942499999997, "rle": {"size": [640, 480], "counts": "RlR62lc04L4L2SMJoA8e=;QBGm=c0jA^OZ<CYCV17XO_<EaB0BU1\\1WO_<H_B7\\Om0c1UOb<I\\Bk1Q1\\Nd<IXBo1Q1ZNf<IVBo1R1YNi<IQBR2S1VNl<JmAS2V1SNn<KiAU2V1QNR=JeAX2W1oMT=h2iBYMX=g2gBYMZ=h4N101O0O01O10O01O1O001O00010O000000001O0O100O2O000O10001N100010O01O0010O0001O001O00001O00000O2O00000000001O000`KTAl3m>QLUAo3m>lKVAT4l>gKWAX4l>cKWA]4V?00000001O0O100O2O0O101N100O102M2O1N2O2M2O1N3N1N2O2M4M3L4M3L4M3L6K7H8I6I9H8G;Fb0]O<E\\Tn0"}, "label": "orange cat sitting in front of television on a stand"}]}
{"id": 87518, "image": "COCO_train2014_000000087518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat sitting in front of the television\"."}], "task": "refering", "answer_template": "The \"a cat sitting in front of the television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 232, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 317, 318, 319, 320, 321], "bbox": [0.6479374999999999, 0.52896875, 0.8970416666666666, 0.8285625], "iscrowd": 0, "area": 16231.942499999997, "rle": {"size": [640, 480], "counts": "RlR62lc04L4L2SMJoA8e=;QBGm=c0jA^OZ<CYCV17XO_<EaB0BU1\\1WO_<H_B7\\Om0c1UOb<I\\Bk1Q1\\Nd<IXBo1Q1ZNf<IVBo1R1YNi<IQBR2S1VNl<JmAS2V1SNn<KiAU2V1QNR=JeAX2W1oMT=h2iBYMX=g2gBYMZ=h4N101O0O01O10O01O1O001O00010O000000001O0O100O2O000O10001N100010O01O0010O0001O001O00001O00000O2O00000000001O000`KTAl3m>QLUAo3m>lKVAT4l>gKWAX4l>cKWA]4V?00000001O0O100O2O0O101N100O102M2O1N2O2M2O1N3N1N2O2M4M3L4M3L4M3L6K7H8I6I9H8G;Fb0]O<E\\Tn0"}, "label": "a cat sitting in front of the television"}]}
{"id": 439784, "image": "COCO_train2014_000000439784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass table\"."}], "task": "refering", "answer_template": "The \"glass table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.687640625, 0.7505624999999999, 0.9960625000000001, 0.9887708333333334], "iscrowd": 0, "area": 17271.455700000006, "rle": {"size": [480, 640], "counts": "ee^68f>>B4L4L4L4M3L5K4L2N2N2N2O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000001N1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O10000001O000000000000001O000000000000001O00000000000000001O0VNXE\\Oh:`0^E^Ob:?cE_O]:=iEAW:;oECQ:9UFEk97[FGf94`FJa92dFL]90hFNY9NlF1U9JPG4Q9HTG6l;O2N1O2N1OWn0"}, "label": "glass table"}]}
{"id": 439784, "image": "COCO_train2014_000000439784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear glass table with a glass cup being held above it\"."}], "task": "refering", "answer_template": "The \"a clear glass table with a glass cup being held above it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.687640625, 0.7505624999999999, 0.9960625000000001, 0.9887708333333334], "iscrowd": 0, "area": 17271.455700000006, "rle": {"size": [480, 640], "counts": "ee^68f>>B4L4L4L4M3L5K4L2N2N2N2O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000001N1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O10000001O000000000000001O000000000000001O00000000000000001O0VNXE\\Oh:`0^E^Ob:?cE_O]:=iEAW:;oECQ:9UFEk97[FGf94`FJa92dFL]90hFNY9NlF1U9JPG4Q9HTG6l;O2N1O2N1OWn0"}, "label": "a clear glass table with a glass cup being held above it"}]}
{"id": 54764, "image": "COCO_train2014_000000054764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a basket full of flowering plants sitting on top of a stack of cardboard boxes\"."}], "task": "refering", "answer_template": "The \"a basket full of flowering plants sitting on top of a stack of cardboard boxes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 243, 244, 245, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.00840625, 0.5426041666666667, 0.761765625, 0.9865416666666665], "iscrowd": 0, "area": 70705.3644, "rle": {"size": [480, 640], "counts": "ii21j>5L4K5L4K5L4K5K5L4K5L4K5L4K5K5L4K5L4K5K5L4K5L4K5L4K5K5L4K5L4K4M4O1000000O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000dNVKaHk4_7YK]Hg4c7^KXHb4g7cKUH]4k7hKPHX4P8lKlGT4T8PLhGP4W8VLcGk3]8YL_Gg3a8\\100O100000000O1000000O1000000O100000000O1000000O100000000O1000000O1000000O14L4L5K4L5K3M00O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100OeJbGV4^8iKbGX4]8gKcG[4]8dKcG]4]8aKdG`4\\8^KdGd4]8ZKcGg4]8WKdGj4\\8UKdGl4]8QKcGQ5]8mJdGT5]8jJcGW5]8gJcG[5]8dJcG]5^8`JcGa5]8]JdGd5\\8[JcGg5^8VJcGk5]8TJcGm5]8QJcGQ6e81O1O1N2O1O1N2O1O1O10000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000004L<D;E1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O00000000000000001O00000000000000000000001O0000000000000O10000000001O0000000000001O1O1O1O1O1O1O1O1OVOjKkFU4U9RLdFn3\\9YL^Ff3b9aLXF^3h9m000001O000N2O1N2N2O1N2N2O1N3M2O1N2N2O1N2N2O1N2N3N1N2N2O1N2N101N2N101N2N101N2N101N1L5G9H7H9K50O10000O100O10000O11O3M2N2N2N3M2N2N3M2N2N1O000000000000001O000000000000000000000000000000001O000000000ROcB9]=DfB<Z=AjB>V=^OnBb0R=[ORCd0n<YOVCf0b=01O001O001O001O001O2N3M2N3M3M3M3MYlV2"}, "label": "a basket full of flowering plants sitting on top of a stack of cardboard boxes"}]}
{"id": 54764, "image": "COCO_train2014_000000054764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pale yellow color flowers arranged on the table\"."}], "task": "refering", "answer_template": "The \"pale yellow color flowers arranged on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 243, 244, 245, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.00840625, 0.5426041666666667, 0.761765625, 0.9865416666666665], "iscrowd": 0, "area": 70705.3644, "rle": {"size": [480, 640], "counts": "ii21j>5L4K5L4K5L4K5K5L4K5L4K5L4K5K5L4K5L4K5K5L4K5L4K5L4K5K5L4K5L4K4M4O1000000O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000dNVKaHk4_7YK]Hg4c7^KXHb4g7cKUH]4k7hKPHX4P8lKlGT4T8PLhGP4W8VLcGk3]8YL_Gg3a8\\100O100000000O1000000O1000000O100000000O1000000O100000000O1000000O1000000O14L4L5K4L5K3M00O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100OeJbGV4^8iKbGX4]8gKcG[4]8dKcG]4]8aKdG`4\\8^KdGd4]8ZKcGg4]8WKdGj4\\8UKdGl4]8QKcGQ5]8mJdGT5]8jJcGW5]8gJcG[5]8dJcG]5^8`JcGa5]8]JdGd5\\8[JcGg5^8VJcGk5]8TJcGm5]8QJcGQ6e81O1O1N2O1O1N2O1O1O10000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000004L<D;E1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O00000000000000001O00000000000000000000001O0000000000000O10000000001O0000000000001O1O1O1O1O1O1O1O1OVOjKkFU4U9RLdFn3\\9YL^Ff3b9aLXF^3h9m000001O000N2O1N2N2O1N2N2O1N3M2O1N2N2O1N2N2O1N2N3N1N2N2O1N2N101N2N101N2N101N2N101N1L5G9H7H9K50O10000O100O10000O11O3M2N2N2N3M2N2N3M2N2N1O000000000000001O000000000000000000000000000000001O000000000ROcB9]=DfB<Z=AjB>V=^OnBb0R=[ORCd0n<YOVCf0b=01O001O001O001O001O2N3M2N3M3M3M3MYlV2"}, "label": "pale yellow color flowers arranged on the table"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the legs of a woman sitting on the floor\"."}], "task": "refering", "answer_template": "The \"the legs of a woman sitting on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 253, 254, 276, 277, 278, 279, 299, 300, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348], "bbox": [0.0, 0.5819583333333332, 0.22985937500000003, 0.9340208333333333], "iscrowd": 0, "area": 9922.694799999996, "rle": {"size": [480, 640], "counts": "i8k4S:40N1O1O100O1O1O101N1cNoEiMR:V2nEkMR:S2oEmMR:R2nEnMS:Q2mEoMT:Q2kEoMV:P2kEoMV:P2jEPNX:o1gEQNZ:n1fESNZ:l1fETN[:l1dETN]:k1dETN]:k1cEUN^:i1cEWN^:g1cEYN^:e1dEZN^:d1bE\\N_:b1bE^N_:`1cE`N]:^1dEbN]:\\1dEdN]:[1cEeN^:Y1dEgN\\:X1dEhN^:U1cEkN^:T1cEkN^:S1cEmN^:R1bEnN_:P1cEoN`:n0`EROc:k0]EUOe:h0\\EXOg:e0ZEZOh:;^DmNk0h0j:7aDlNe0m0l:3`EL`:3aEM_:2bEN^:0dE0\\:OeE1[:NfE2Z:LhE5V:KkE5U:JmE5S:JnE7Q:HPF8P:FRF:n9FRF;m9DTF<k9DVF=i9BXF>h9BYF=g9CZF;g9DZF<f9DZF;g9EYF:h9FWF;i9EWF;i9EWF;i9EWF;i9EWF;i9DXF<h9DXF<i9BXF>h9BXF>h9AYF?g9AYF?g9BYF>g9DVF<j9GTF8l9LPF4P:0mEOS:5jEJW:9fEF[:=bEB`:?^E@c:c0YE]Oh:f0UEYOl:j0_DcN2c0`;l0YDcN7a0b;o0PDbN>?c;V1\\DiNf;V1[DiNf;V1\\DhNe;W1]DgNd;X1^DfNc;Y1_DeNb;Y1aDdNa;[1aDcN`;\\1bDbN_;]1S1O1N3N2N1O2N2M2O5K1O1O1N2O1O1O10O010O01O010O010O001O001O00000O10000O100O10000O100O1O2M4KifV7"}, "label": "the legs of a woman sitting on the floor"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a grey shirt sitting on the floor\"."}], "task": "refering", "answer_template": "The \"a woman in a grey shirt sitting on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 253, 254, 276, 277, 278, 279, 299, 300, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348], "bbox": [0.0, 0.5819583333333332, 0.22985937500000003, 0.9340208333333333], "iscrowd": 0, "area": 9922.694799999996, "rle": {"size": [480, 640], "counts": "i8k4S:40N1O1O100O1O1O101N1cNoEiMR:V2nEkMR:S2oEmMR:R2nEnMS:Q2mEoMT:Q2kEoMV:P2kEoMV:P2jEPNX:o1gEQNZ:n1fESNZ:l1fETN[:l1dETN]:k1dETN]:k1cEUN^:i1cEWN^:g1cEYN^:e1dEZN^:d1bE\\N_:b1bE^N_:`1cE`N]:^1dEbN]:\\1dEdN]:[1cEeN^:Y1dEgN\\:X1dEhN^:U1cEkN^:T1cEkN^:S1cEmN^:R1bEnN_:P1cEoN`:n0`EROc:k0]EUOe:h0\\EXOg:e0ZEZOh:;^DmNk0h0j:7aDlNe0m0l:3`EL`:3aEM_:2bEN^:0dE0\\:OeE1[:NfE2Z:LhE5V:KkE5U:JmE5S:JnE7Q:HPF8P:FRF:n9FRF;m9DTF<k9DVF=i9BXF>h9BYF=g9CZF;g9DZF<f9DZF;g9EYF:h9FWF;i9EWF;i9EWF;i9EWF;i9EWF;i9DXF<h9DXF<i9BXF>h9BXF>h9AYF?g9AYF?g9BYF>g9DVF<j9GTF8l9LPF4P:0mEOS:5jEJW:9fEF[:=bEB`:?^E@c:c0YE]Oh:f0UEYOl:j0_DcN2c0`;l0YDcN7a0b;o0PDbN>?c;V1\\DiNf;V1[DiNf;V1\\DhNe;W1]DgNd;X1^DfNc;Y1_DeNb;Y1aDdNa;[1aDcN`;\\1bDbN_;]1S1O1N3N2N1O2N2M2O5K1O1O1N2O1O1O10O010O01O010O010O001O001O00000O10000O100O10000O100O1O2M4KifV7"}, "label": "a woman in a grey shirt sitting on the floor"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a multi patterned blue shirt sits on a couch\"."}], "task": "refering", "answer_template": "The \"a woman in a multi patterned blue shirt sits on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 217, 218, 219, 220], "bbox": [0.153875, 0.2504166666666667, 0.6038125, 0.5476041666666667], "iscrowd": 0, "area": 30647.721600000004, "rle": {"size": [480, 640], "counts": "nS^12c>a0H8J5J5L4K5L4K5L3_OYNbCl1Z<WNbC[1LPO]<IdCT14ROW<X1nCfNQ<\\1oCcNQ<]1QD_NP<c1QDYNQ<g1c02O0O101N100O10N1O2N1O2N1O2O0010O010O010O010H9I7O001O1O1O1O1O4L4L1O1O1O001O1nCZMh;o20001O00000O10000000000000001O00000000000000000000001O0O100000000000000000001O00000000000000000000000O2O00000000000000000000001O0000000000000000001N10000000000000001O01O00000000000001O0001O0000O100O2N1O1O1O1O1O100O1O20O00001O00001O00010O00001O00001O00010O00001O00001O00UE^LX:b3hE^LX:b3gE_LY:a3fE`LZ:`3eEaL[:_3dEbL\\:^3cEcL]:]3bEdL^:\\3bEdL^:\\3aEeL_:[3`EfL`:Z3_EgLa:Y3^EhLb:Z3[EgLe:[3XEfLh:i300000000000000000000001O00000000000000000000000000000000000000000000000000000000001O0000000000000000000000000O10000O10000O1000000O2O000O10000O101O000O10000O2D;D<C`0ASnf3"}, "label": "a woman in a multi patterned blue shirt sits on a couch"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing glasses and a watch sitting on a sofa by herself\"."}], "task": "refering", "answer_template": "The \"a woman wearing glasses and a watch sitting on a sofa by herself\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 217, 218, 219, 220], "bbox": [0.153875, 0.2504166666666667, 0.6038125, 0.5476041666666667], "iscrowd": 0, "area": 30647.721600000004, "rle": {"size": [480, 640], "counts": "nS^12c>a0H8J5J5L4K5L4K5L3_OYNbCl1Z<WNbC[1LPO]<IdCT14ROW<X1nCfNQ<\\1oCcNQ<]1QD_NP<c1QDYNQ<g1c02O0O101N100O10N1O2N1O2N1O2O0010O010O010O010H9I7O001O1O1O1O1O4L4L1O1O1O001O1nCZMh;o20001O00000O10000000000000001O00000000000000000000001O0O100000000000000000001O00000000000000000000000O2O00000000000000000000001O0000000000000000001N10000000000000001O01O00000000000001O0001O0000O100O2N1O1O1O1O1O100O1O20O00001O00001O00010O00001O00001O00010O00001O00001O00UE^LX:b3hE^LX:b3gE_LY:a3fE`LZ:`3eEaL[:_3dEbL\\:^3cEcL]:]3bEdL^:\\3bEdL^:\\3aEeL_:[3`EfL`:Z3_EgLa:Y3^EhLb:Z3[EgLe:[3XEfLh:i300000000000000000000001O00000000000000000000000000000000000000000000000000000000001O0000000000000000000000000O10000O10000O1000000O2O000O10000O101O000O10000O2D;D<C`0ASnf3"}, "label": "a woman wearing glasses and a watch sitting on a sofa by herself"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on couch with dog toy\"."}], "task": "refering", "answer_template": "The \"man on couch with dog toy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 155, 156, 157, 177, 178, 179, 180, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 266, 267, 289, 290, 312], "bbox": [0.577046875, 0.32435416666666667, 0.86153125, 0.7941875], "iscrowd": 0, "area": 14647.631599999999, "rle": {"size": [480, 640], "counts": "PZ]5?^>8H7I3L5L3M4L3L4M4L3L4N3M2N3M2N1O1O1000000000I7D<N2N2NmN_NlD_1X;aNfD\\1^;eN`DX1d;hN[Dk0S<VOkCh0X<XOgCf0[<[OdCb0`<^O^Cb0d<_OZC`0h<@WC?k<IkB7W=2_BMc=g01N1O2O1O1001O00000O101O000000001O0O100000001O000O10001O000O1000001O0O100000000O100000000O101oN^NaDc1[;eNaD[1\\;lN_DU1^;RO^Dn0_;]OXDd0h;[1N1O100O100O2O0O100O100O2N100O100O1O2N1O1O1O1N2O1O1O1O2N1O1O1N2O1O1O1N2N2N3N010O1O1O100O1O100O001O1O1O10000000O1O2N1M3M3M3M3M4M2M3M3M3M3M4M2N2O1O1N3N1O3L7J5K6J5J4M4K5L3L5K5L5J6J6Ka`Y1"}, "label": "man on couch with dog toy"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"balding man sitting on couch wearing brown hoodie\"."}], "task": "refering", "answer_template": "The \"balding man sitting on couch wearing brown hoodie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 155, 156, 157, 177, 178, 179, 180, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 266, 267, 289, 290, 312], "bbox": [0.577046875, 0.32435416666666667, 0.86153125, 0.7941875], "iscrowd": 0, "area": 14647.631599999999, "rle": {"size": [480, 640], "counts": "PZ]5?^>8H7I3L5L3M4L3L4M4L3L4N3M2N3M2N1O1O1000000000I7D<N2N2NmN_NlD_1X;aNfD\\1^;eN`DX1d;hN[Dk0S<VOkCh0X<XOgCf0[<[OdCb0`<^O^Cb0d<_OZC`0h<@WC?k<IkB7W=2_BMc=g01N1O2O1O1001O00000O101O000000001O0O100000001O000O10001O000O1000001O0O100000000O100000000O101oN^NaDc1[;eNaD[1\\;lN_DU1^;RO^Dn0_;]OXDd0h;[1N1O100O100O2O0O100O100O2N100O100O1O2N1O1O1O1N2O1O1O1O2N1O1O1N2O1O1O1N2N2N3N010O1O1O100O1O100O001O1O1O10000000O1O2N1M3M3M3M3M4M2M3M3M3M3M4M2N2O1O1N3N1O3L7J5K6J5J4M4K5L3L5K5L5J6J6Ka`Y1"}, "label": "balding man sitting on couch wearing brown hoodie"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch the man in brown is sitting on\"."}], "task": "refering", "answer_template": "The \"the couch the man in brown is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 157, 158, 176, 177, 180, 181, 182, 199, 203, 204, 205, 206, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.587796875, 0.3578125, 1.0, 0.8974166666666666], "iscrowd": 0, "area": 38843.91755000001, "rle": {"size": [480, 640], "counts": "Zb`51o>2PD0l82gF9X9K[F?e9CnEh0R:[OaEo0_:SOUEW1j:lNQEW1o:lNlDV1T;lNgDW1Y;lNaDW1_;kN]DW1b;lNYDW1g;Q100O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O100001O000000YMjLWHoNT1W4d6oLaIQ3^6TM^Il2b6TM_Ik2`6VM`Ij2`6VM`Ij2_6WMaIi2^6XMbIh2^6WMcIi2\\6XMeIg2[6YMeIg2_5oKgI[1j0f2_5SLbIX1o0e2_5XL]IS1T1e2^5]LYIo0Y1d2^5bLTIi0`1d2[5WNeJi1[5WNeJi1[5WNeJi1[5WNeJi1\\5VNdJj1\\5UNfJj1[5UNeJk1[5UNeJk1\\5`MUIkN_1e3\\5_M_IcNU1n3\\5]M[Kc2f4ZM]Ke2c4YM_Kg2b4WM_Ki2a4UMaKk2`4RMbKn2^4QMcKo2]4oLfKP3[4mLgKS3Y4lLhKT3Y4jLhKV3X4iLiKW3X4hLhKX3X4gLiKY3W4gLiKY3X4eLiK[3W4dLjK\\3W4cLiK]3W4bLjK^3W4aLjK^3V4aLkK_3U4aLkK_3V4_LkKa3U4^LlKb3U4]LkKc3U4\\LlKd3U4[LkKe3U4ZLlKf3T4YLmKg3T4XLlKh3T4WLmKi3T4ULmKk3S4ULmKk3T4SLmKm3S4SLmKm3S4RLnKn3R4QLoKo3P4RLPLn3P4QLQLo3n3RLRLn3n3QLSLo3m3PLTLP4k3QLULo3k3PLVLP4j3PLVLP4i3PLYLo3\\70ZLQL[Lo3[7000000000000000000000000000000000000000O1M3M3M3M3M3M3M3M3M3M3N2O1N2O1O1N2O1O1N2O1O1N2O1jMaJjJ`5U5aJiJa5V5aJhJ`5W5bJgJ_5W5cJgJ_5X5cJfJ^5Y5cJfJ^5Y5dJdJ^5[5dJcJ]5]5eJ`J\\5`5fJ\\J\\5d5fJYJ[5g5gJVJZ5j5hJRJZ5o5gJnIZ5R6hJkIY5U6hJhIZ5X6hJeIY5[6iJbIX5^6jJ^IX5b6jJ[IW5f6jJWIW5i6kJSIW5m6kJPIV5o6^1000000O10N2O1O2N3L4M2N3M3L3N3M2N3L4M2N3M3L3N3M2N3M3L3N3M3M2M4M2N3M3M2O2N2O0O2N1iNXFTMj9f2_FVMb9d2gFWMZ9d2nFXMT9b2UGYMm8a2[G\\Me8_2`G`Mb8Z2bGfM_8U2fGjM\\8P2hGQNY8i1lGVNU8e1oG[NS8_1RH`NP8Z1THgNl7T1YHkNi7o0[HQOf7j0_HUOc7e0aH\\O`7>eHA]79gHG[73jHLX7NlH3U7GPI8l:010O01O0010O01O0010O01N1N3LfC"}, "label": "the couch the man in brown is sitting on"}]}
{"id": 521709, "image": "COCO_train2014_000000521709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beige sofa with a man sitting\"."}], "task": "refering", "answer_template": "The \"beige sofa with a man sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 157, 158, 176, 177, 180, 181, 182, 199, 203, 204, 205, 206, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.587796875, 0.3578125, 1.0, 0.8974166666666666], "iscrowd": 0, "area": 38843.91755000001, "rle": {"size": [480, 640], "counts": "Zb`51o>2PD0l82gF9X9K[F?e9CnEh0R:[OaEo0_:SOUEW1j:lNQEW1o:lNlDV1T;lNgDW1Y;lNaDW1_;kN]DW1b;lNYDW1g;Q100O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O100001O000000YMjLWHoNT1W4d6oLaIQ3^6TM^Il2b6TM_Ik2`6VM`Ij2`6VM`Ij2_6WMaIi2^6XMbIh2^6WMcIi2\\6XMeIg2[6YMeIg2_5oKgI[1j0f2_5SLbIX1o0e2_5XL]IS1T1e2^5]LYIo0Y1d2^5bLTIi0`1d2[5WNeJi1[5WNeJi1[5WNeJi1[5WNeJi1\\5VNdJj1\\5UNfJj1[5UNeJk1[5UNeJk1\\5`MUIkN_1e3\\5_M_IcNU1n3\\5]M[Kc2f4ZM]Ke2c4YM_Kg2b4WM_Ki2a4UMaKk2`4RMbKn2^4QMcKo2]4oLfKP3[4mLgKS3Y4lLhKT3Y4jLhKV3X4iLiKW3X4hLhKX3X4gLiKY3W4gLiKY3X4eLiK[3W4dLjK\\3W4cLiK]3W4bLjK^3W4aLjK^3V4aLkK_3U4aLkK_3V4_LkKa3U4^LlKb3U4]LkKc3U4\\LlKd3U4[LkKe3U4ZLlKf3T4YLmKg3T4XLlKh3T4WLmKi3T4ULmKk3S4ULmKk3T4SLmKm3S4SLmKm3S4RLnKn3R4QLoKo3P4RLPLn3P4QLQLo3n3RLRLn3n3QLSLo3m3PLTLP4k3QLULo3k3PLVLP4j3PLVLP4i3PLYLo3\\70ZLQL[Lo3[7000000000000000000000000000000000000000O1M3M3M3M3M3M3M3M3M3M3N2O1N2O1O1N2O1O1N2O1O1N2O1jMaJjJ`5U5aJiJa5V5aJhJ`5W5bJgJ_5W5cJgJ_5X5cJfJ^5Y5cJfJ^5Y5dJdJ^5[5dJcJ]5]5eJ`J\\5`5fJ\\J\\5d5fJYJ[5g5gJVJZ5j5hJRJZ5o5gJnIZ5R6hJkIY5U6hJhIZ5X6hJeIY5[6iJbIX5^6jJ^IX5b6jJ[IW5f6jJWIW5i6kJSIW5m6kJPIV5o6^1000000O10N2O1O2N3L4M2N3M3L3N3M2N3L4M2N3M3L3N3M2N3M3L3N3M3M2M4M2N3M3M2O2N2O0O2N1iNXFTMj9f2_FVMb9d2gFWMZ9d2nFXMT9b2UGYMm8a2[G\\Me8_2`G`Mb8Z2bGfM_8U2fGjM\\8P2hGQNY8i1lGVNU8e1oG[NS8_1RH`NP8Z1THgNl7T1YHkNi7o0[HQOf7j0_HUOc7e0aH\\O`7>eHA]79gHG[73jHLX7NlH3U7GPI8l:010O01O0010O01O0010O01N1N3LfC"}, "label": "beige sofa with a man sitting"}]}
{"id": 456176, "image": "COCO_train2014_000000456176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown cow behind another cow\"."}], "task": "refering", "answer_template": "The \"a light brown cow behind another cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 141, 142, 150, 151, 152, 153, 154, 164, 173, 174, 175, 176, 177, 199, 200, 222, 223, 224, 245, 247], "bbox": [0.13921875, 0.2415625, 0.7584375, 0.60259765625], "iscrowd": 0, "area": 13120.4103, "rle": {"size": [512, 640], "counts": "kf\\13e?7N2M3N3L3N2O101O0O10001O000O10001O00000010O1O1O101N1O1O1N1O2N2N2N2N2N2N1O2N2O1O1O1O010Og^P30ZaoL1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1kA@h<a0UCFf<:XCLd<5ZC1a<0\\C7_<J^C<^<DaC>^<C_C`0`<A]Ca0c<@ZCc0e<]OZCe0e<\\OQCn0n<R101O1O1O0000O10000O1000000O1000000O10000O100000O0100000O01O1O1O001O1O010O1000O100000O010000O10O1001N2O2N2N2N2N2N1O2N2N7I8H7J7H;E;RE`Kf9]5000IhEfJX:Z5hEeJY:Z5hEfJX:Y5jEfJV:Z5jEfJW:X5jEgJW:Y5iEgJW:X5jEhJV:W5lEhJT:X5lEgJU:X5lEhJT:X5lEhJT:a3kFlL`9T3f1100O001O001O1O001O00iNmCbNS<Y1RDfNn;V1XDiNi;R1[DnNf;l0`DTO_;g0fD[OY;?mDAS;9SEHl:3ZEM]=0100O10000O100O100O10000O1O0N3M3M3M3M4L4M3LjV]2"}, "label": "a light brown cow behind another cow"}]}
{"id": 456176, "image": "COCO_train2014_000000456176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light brown cattle in background grazing\"."}], "task": "refering", "answer_template": "The \"light brown cattle in background grazing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 141, 142, 150, 151, 152, 153, 154, 164, 173, 174, 175, 176, 177, 199, 200, 222, 223, 224, 245, 247], "bbox": [0.13921875, 0.2415625, 0.7584375, 0.60259765625], "iscrowd": 0, "area": 13120.4103, "rle": {"size": [512, 640], "counts": "kf\\13e?7N2M3N3L3N2O101O0O10001O000O10001O00000010O1O1O101N1O1O1N1O2N2N2N2N2N2N1O2N2O1O1O1O010Og^P30ZaoL1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1kA@h<a0UCFf<:XCLd<5ZC1a<0\\C7_<J^C<^<DaC>^<C_C`0`<A]Ca0c<@ZCc0e<]OZCe0e<\\OQCn0n<R101O1O1O0000O10000O1000000O1000000O10000O100000O0100000O01O1O1O001O1O010O1000O100000O010000O10O1001N2O2N2N2N2N2N1O2N2N7I8H7J7H;E;RE`Kf9]5000IhEfJX:Z5hEeJY:Z5hEfJX:Y5jEfJV:Z5jEfJW:X5jEgJW:Y5iEgJW:X5jEhJV:W5lEhJT:X5lEgJU:X5lEhJT:X5lEhJT:a3kFlL`9T3f1100O001O001O1O001O00iNmCbNS<Y1RDfNn;V1XDiNi;R1[DnNf;l0`DTO_;g0fD[OY;?mDAS;9SEHl:3ZEM]=0100O10000O100O100O10000O1O0N3M3M3M3M4L4M3LjV]2"}, "label": "light brown cattle in background grazing"}]}
{"id": 456176, "image": "COCO_train2014_000000456176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow with it ' s tail twisted in the air\"."}], "task": "refering", "answer_template": "The \"a brown cow with it ' s tail twisted in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 219, 220, 232, 235, 236, 237, 242, 243, 258, 260, 261, 265, 266], "bbox": [0.082828125, 0.2271484375, 0.5855625, 0.67134765625], "iscrowd": 0, "area": 31738.107299999992, "rle": {"size": [512, 640], "counts": "]ij05Z>2kB4R=3fB2b<DPCc06Lg<EoBg02Hj<FPCj0MCR=GkBg1V=a01O001N2O001O001O1N1000000000000N2F:O000O101O0000001N1000001O0O10001O000O101O0000001N10000001O1O1N2O1O1O001O1O1O1O1O001O1N2O1O1N101O1N101O1N101O0O2O001N101O010O1]EXMn7g2QH[Mn7f2PH]Mo7b2PH`Mo7a2oGaMQ8^2nGdMQ8]2nGeMQ8Z2oGgMQ8Y2mGiMR8W2nGjMR8V2mGlMQ8T2oGmMQ8S2mGnMS8S2fGTNZ8n1ZG\\Ne8j1mF_NS9f1_FbNa9d1PFfNP:e3O10FjEjJV:Q5RFkJn9R5YFkJg9Q5e0M4L3N3M2N3N1N3M3M2N3N2M2N3M3M3M3M22O3M2N2N3L3N3M2N2N2M2O1O001O1N2O001O1OfNlDhMS;Y2SEbMl:]2ZE_Me:`2bE\\M\\:c2iE[MU:d2nE[MQ:e2PF\\Mn9c2TF]Mk9b2XF^Mf9c2ZF]Md9d2]F]Ma9c2`F]M_9c2cF]M[9c2fF]MY9c2hF^MnN\\Oj9V3ZG]MiN@l9S3\\G^MeNAn9Q3^G^MbNDn9n2aGjM`8V2bGhM^8Y2bGeM_8[2bGcM_8]2cGaM]8_2dG_M]8a2dG]M]8c2dGoLXN2T:o2fGkLYN4R:Q3fGhL[N5P:S3PHlLP8T3QHjLP8V3QHiLo7W3QHhLP8W3]2000000O10000O10000O10000O10000O1000000O01000O10000O10000O10000O10000O1000000O1000000O1000O1000O10O10000000O0100000000O01000^OTMPDl2m;XMRDh2k;\\MSDd2o;[MQDe2P<[MoCe2Q<[MoCe2R<[MmCe2T<[MjCf2V<ZMkCe2V<[MiCd2X<\\MhCd2Y<\\MfCe2Z<[MeCe2[<[MeCf2[<ZMdCf2]<\\M`Ce2_<>G910O001O001O00010O06dLgCf2i<100O2N1O100O1O101N1O1O100O1O2O0O1O1O100O2O00[M[Do0d;POXE6g:I[E7c:I^E9`:FbE:]:DeE=Z:AiE`0U:]OnEd0R:YOQFh0YOeNZ9>`GP1QOcNa99aGV1jNaNi95`G\\1bNaNo9ObGb1[NaNe<`1WCbNi<`1QCeNo<\\1kBgNV=Z1eBiN\\=l110O10dMdBT2i=J6K5K5K5J6K5J6F:EcVT4"}, "label": "a brown cow with it ' s tail twisted in the air"}]}
{"id": 456176, "image": "COCO_train2014_000000456176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown cow with a red tag in it ' s ear\"."}], "task": "refering", "answer_template": "The \"the brown cow with a red tag in it ' s ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 219, 220, 232, 235, 236, 237, 242, 243, 258, 260, 261, 265, 266], "bbox": [0.082828125, 0.2271484375, 0.5855625, 0.67134765625], "iscrowd": 0, "area": 31738.107299999992, "rle": {"size": [512, 640], "counts": "]ij05Z>2kB4R=3fB2b<DPCc06Lg<EoBg02Hj<FPCj0MCR=GkBg1V=a01O001N2O001O001O1N1000000000000N2F:O000O101O0000001N1000001O0O10001O000O101O0000001N10000001O1O1N2O1O1O001O1O1O1O1O001O1N2O1O1N101O1N101O1N101O0O2O001N101O010O1]EXMn7g2QH[Mn7f2PH]Mo7b2PH`Mo7a2oGaMQ8^2nGdMQ8]2nGeMQ8Z2oGgMQ8Y2mGiMR8W2nGjMR8V2mGlMQ8T2oGmMQ8S2mGnMS8S2fGTNZ8n1ZG\\Ne8j1mF_NS9f1_FbNa9d1PFfNP:e3O10FjEjJV:Q5RFkJn9R5YFkJg9Q5e0M4L3N3M2N3N1N3M3M2N3N2M2N3M3M3M3M22O3M2N2N3L3N3M2N2N2M2O1O001O1N2O001O1OfNlDhMS;Y2SEbMl:]2ZE_Me:`2bE\\M\\:c2iE[MU:d2nE[MQ:e2PF\\Mn9c2TF]Mk9b2XF^Mf9c2ZF]Md9d2]F]Ma9c2`F]M_9c2cF]M[9c2fF]MY9c2hF^MnN\\Oj9V3ZG]MiN@l9S3\\G^MeNAn9Q3^G^MbNDn9n2aGjM`8V2bGhM^8Y2bGeM_8[2bGcM_8]2cGaM]8_2dG_M]8a2dG]M]8c2dGoLXN2T:o2fGkLYN4R:Q3fGhL[N5P:S3PHlLP8T3QHjLP8V3QHiLo7W3QHhLP8W3]2000000O10000O10000O10000O10000O1000000O01000O10000O10000O10000O10000O1000000O1000000O1000O1000O10O10000000O0100000000O01000^OTMPDl2m;XMRDh2k;\\MSDd2o;[MQDe2P<[MoCe2Q<[MoCe2R<[MmCe2T<[MjCf2V<ZMkCe2V<[MiCd2X<\\MhCd2Y<\\MfCe2Z<[MeCe2[<[MeCf2[<ZMdCf2]<\\M`Ce2_<>G910O001O001O00010O06dLgCf2i<100O2N1O100O1O101N1O1O100O1O2O0O1O1O100O2O00[M[Do0d;POXE6g:I[E7c:I^E9`:FbE:]:DeE=Z:AiE`0U:]OnEd0R:YOQFh0YOeNZ9>`GP1QOcNa99aGV1jNaNi95`G\\1bNaNo9ObGb1[NaNe<`1WCbNi<`1QCeNo<\\1kBgNV=Z1eBiN\\=l110O10dMdBT2i=J6K5K5K5J6K5J6F:EcVT4"}, "label": "the brown cow with a red tag in it ' s ear"}]}
{"id": 357877, "image": "COCO_train2014_000000357877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle blonde in a black hoodie that reads michigan in gold letters\"."}], "task": "refering", "answer_template": "The \"a bottle blonde in a black hoodie that reads michigan in gold letters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 153, 154, 155, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.52625, 0.26491803278688525, 0.9295625000000001, 0.9907494145199064], "iscrowd": 0, "area": 48869.809, "rle": {"size": [427, 640], "counts": "mm\\43T=4L4L4K5L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4M3M4M2[Oe0O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1000000O10000O1000000O10000O100000000000000K5K5J6J6ZNXIdLm6\\3a1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O100O1O1O10000O100000000O10000000000O10000000000001O1O1O2N1O1O2N1O1O2N1O1O2N4L4L5K4L4L5K5K4jHnJl5V5_I\\K_6j5N1O2N2N2N1O2N1O1O001O1O001O1O001O1O001O1O002N1O1O2N1O1O2N1O1O1O2N1O1O1O1O1O001O1O1O001O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1XOh0J6I7J7N100O2O1N101N101N1O2L4M2N3L3N3L4M2M4M2M4L3M4M3L4L4L3L5I7I7H8I6J7I`^b0"}, "label": "a bottle blonde in a black hoodie that reads michigan in gold letters"}]}
{"id": 357877, "image": "COCO_train2014_000000357877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a dark blue sweatshirt is talking on her phone\"."}], "task": "refering", "answer_template": "The \"a woman in a dark blue sweatshirt is talking on her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 153, 154, 155, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.52625, 0.26491803278688525, 0.9295625000000001, 0.9907494145199064], "iscrowd": 0, "area": 48869.809, "rle": {"size": [427, 640], "counts": "mm\\43T=4L4L4K5L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L4M3M4M2[Oe0O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1000000O10000O1000000O10000O100000000000000K5K5J6J6ZNXIdLm6\\3a1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O100O1O1O10000O100000000O10000000000O10000000000001O1O1O2N1O1O2N1O1O2N1O1O2N4L4L5K4L4L5K5K4jHnJl5V5_I\\K_6j5N1O2N2N2N1O2N1O1O001O1O001O1O001O1O001O1O002N1O1O2N1O1O2N1O1O1O2N1O1O1O1O1O001O1O1O001O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1XOh0J6I7J7N100O2O1N101N101N1O2L4M2N3L3N3L4M2M4M2M4L3M4M3L4L4L3L5I7I7H8I6J7I`^b0"}, "label": "a woman in a dark blue sweatshirt is talking on her phone"}]}
{"id": 431613, "image": "COCO_train2014_000000431613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red jacket sitting on the ground with his ski equipment having something to eat and drink with his buddy next to them\"."}], "task": "refering", "answer_template": "The \"a man in a red jacket sitting on the ground with his ski equipment having something to eat and drink with his buddy next to them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 225, 245, 246, 247, 248, 249, 270, 271], "bbox": [0.6353125000000001, 0.43587822014051525, 0.8482968750000002, 0.7681498829039812], "iscrowd": 0, "area": 8443.721650000001, "rle": {"size": [427, 640], "counts": "ToY5=k<3N3L3M4L3M4L3O2M2O2M2N3N1N3N1N3N1M4M2L6J7J5K3B=O2N2N2O01000O01000000O10O1001O001O1O001O1O2N1O2N1O2N1O2O2OlE[Mb9c2YFaMj9]2RFeMo9\\2mEfMT:[2hEgMY:e22O0O1O1O100O1O1O1O100O001O2O0O1O1O1O100O1N2003N2M2NO2M2O1O2N1O1N3N1O4L:E7J1O2N1O2M2O2N2N2M5L3M3M0O10O100000O1000000O100O100O1O100O101N1O2O1N2N1O2N2N1O2NWbX1"}, "label": "a man in a red jacket sitting on the ground with his ski equipment having something to eat and drink with his buddy next to them"}]}
{"id": 431613, "image": "COCO_train2014_000000431613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on the right of a snowy hill\"."}], "task": "refering", "answer_template": "The \"a man sitting on the right of a snowy hill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 225, 245, 246, 247, 248, 249, 270, 271], "bbox": [0.6353125000000001, 0.43587822014051525, 0.8482968750000002, 0.7681498829039812], "iscrowd": 0, "area": 8443.721650000001, "rle": {"size": [427, 640], "counts": "ToY5=k<3N3L3M4L3M4L3O2M2O2M2N3N1N3N1N3N1M4M2L6J7J5K3B=O2N2N2O01000O01000000O10O1001O001O1O001O1O2N1O2N1O2N1O2O2OlE[Mb9c2YFaMj9]2RFeMo9\\2mEfMT:[2hEgMY:e22O0O1O1O100O1O1O1O100O001O2O0O1O1O1O100O1N2003N2M2NO2M2O1O2N1O1N3N1O4L:E7J1O2N1O2M2O2N2N2M5L3M3M0O10O100000O1000000O100O100O1O100O101N1O2O1N2N1O2N2N1O2NWbX1"}, "label": "a man sitting on the right of a snowy hill"}]}
{"id": 570878, "image": "COCO_train2014_000000570878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the 5th ram from the left\"."}], "task": "refering", "answer_template": "The \"the 5th ram from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.520328125, 0.4068632075471698, 1.0, 0.9905660377358491], "iscrowd": 0, "area": 50636.847100000006, "rle": {"size": [424, 640], "counts": "eTZ44n<6I7J5K4K6K5L301O1O1O001O10O01O1O001O100O001O1O001001O0O10000O10000O1000000O2O001N101O001N12O1N3N1N2O1O1N3N1O1O1O1O2N1O1O001O001O1UH]No3c1kKeNS4\\1gKlNV4T1eKTOX4m0cKYO[4h0_K@^4`0]KH`49ZKNe41WK6f4KUK<i4CSKd0j4]OQKi0m4XOnJo0P5POmJV1P5kNkJ[1T5TOXJS1f5Z3N2N1O2N2N1OO1O1O1O001O1O1O1O1O1N2O1O1O1O2N3M1O0000O10000O1000O01O00O100001O001O00001O001O00001O0100O1000O2O001O001N101O001O0O2O001O2N1O2M2O2N1O1O0O100000000O10000000000000001O00000000000000000000000000000000000000O10001O2N3M2N3M2N2N3L3N2N3M2N2N3M2N2N3L3N3M2N2N3M2N2kJbH\\4a7^KbHb4`7YKbHh4`7SKcHl4m70O100O10000O01000O1000O010000O1000O010000O1000O010000O100O01000O1000O100000O1000O10000O1O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N201000O2O00001O00001O00001N10001O00001O001O0O101O0P1PO^1gMSF"}, "label": "the 5th ram from the left"}]}
{"id": 570878, "image": "COCO_train2014_000000570878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an animal with large curving horns on the furthest to the right of the group , closest to the photographer staring directly into the camera\"."}], "task": "refering", "answer_template": "The \"an animal with large curving horns on the furthest to the right of the group , closest to the photographer staring directly into the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.520328125, 0.4068632075471698, 1.0, 0.9905660377358491], "iscrowd": 0, "area": 50636.847100000006, "rle": {"size": [424, 640], "counts": "eTZ44n<6I7J5K4K6K5L301O1O1O001O10O01O1O001O100O001O1O001001O0O10000O10000O1000000O2O001N101O001N12O1N3N1N2O1O1N3N1O1O1O1O2N1O1O001O001O1UH]No3c1kKeNS4\\1gKlNV4T1eKTOX4m0cKYO[4h0_K@^4`0]KH`49ZKNe41WK6f4KUK<i4CSKd0j4]OQKi0m4XOnJo0P5POmJV1P5kNkJ[1T5TOXJS1f5Z3N2N1O2N2N1OO1O1O1O001O1O1O1O1O1N2O1O1O1O2N3M1O0000O10000O1000O01O00O100001O001O00001O001O00001O0100O1000O2O001O001N101O001O0O2O001O2N1O2M2O2N1O1O0O100000000O10000000000000001O00000000000000000000000000000000000000O10001O2N3M2N3M2N2N3L3N2N3M2N2N3M2N2N3L3N3M2N2N3M2N2kJbH\\4a7^KbHb4`7YKbHh4`7SKcHl4m70O100O10000O01000O1000O010000O1000O010000O1000O010000O100O01000O1000O100000O1000O10000O1O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N201000O2O00001O00001O00001N10001O00001O001O0O101O0P1PO^1gMSF"}, "label": "an animal with large curving horns on the furthest to the right of the group , closest to the photographer staring directly into the camera"}]}
{"id": 570878, "image": "COCO_train2014_000000570878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ram in front with his head down\"."}], "task": "refering", "answer_template": "The \"the ram in front with his head down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 121, 122, 123, 124, 125, 168, 169, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 308, 326, 328, 329, 330], "bbox": [0.071234375, 0.2795754716981132, 0.482984375, 0.9870990566037736], "iscrowd": 0, "area": 29226.65075, "rle": {"size": [424, 640], "counts": "_\\c02T=3N2M2O2M3N2M3N2N2M3N3L3N3L4M3L4M3M3L4M3L4M2M4M3M3L4M2M2O1N2O1N2O1O0O2O1N2O1M3M3M3M3L4M3M3M3M2M4O1O100O1O100O1O1O100O1O010O1O100O1O1O100O1O100O1O10O0o0QO00100O1O100O1O100O1O100O1OH8K6N1O1O2N1O1O2N1O1O200000000000O1000O10000000000000000000O10000009G;EoLjKYMm3b2XL^Mh3a2ZL^Mk3]2ULcMP4W2QLiMT4Q2mKoMX4l1hKTN\\4g1eKYN`4a1`K`Ne4[1[KeNf4Y1[KgNe4X1\\KhNd4X1[KiNe4V1\\KiNe4V1\\K`Nn4`1RKnM`5Q2`JPN`5o1aJPN`5P2`JPN`5o1`JQNa5n1`JRN`5n1`JRN`5m1aJRN`5n1_JSNa5l1`JSNa5l1`JTN`5l1`JTN`5k1`JUNa5k1_JUNa5k1_JUNa5j1_JVNT4SOYKg2c0VNQ4Z3PLeLn3^3RLbLk3`3ULaLi3a3WL^Li3c3WL]Ll3_3TLaLP4\\3PLdLR4Y3oKgL[4o2eKPMi4b2WK_MW5S2iJmMc5f1^JYNc5g1]JYNc5g1\\JYNe5f1\\JZNd5f1\\JZNd5f1[JZNf5f1ZJZNf5g1YJXNh5h1XJXNh5i1XJVNh5j1XJUNi5l1VJTNj5l1VJTNj5l1VJSNk5n1TJRNl5n1TJ`N^5a1aJROl4n0TKFX4;hK2j3NVL`0\\3AcLl0P3TOPM]1_2dN`MP2l1PNTNS2i1mMWNV2f1kMYNX2d1hM\\N\\2`1eM`N]2]1cMcN`2Z1aMeN`2Y1aMgN`2T1eMlN\\2n0hMTOW2g0mM[OR2a0RN@m1;YNFf14aNK^10iNOV1LQO3n0HYO7g0C@=g6O1M3L2OO1O001O001O001O0O101O001O001O00001O0101N100O100O10000O10000O101O000O10000O10000O101O0O101O0O101O0O`ZY4"}, "label": "the ram in front with his head down"}]}
{"id": 570878, "image": "COCO_train2014_000000570878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long horned sheep whose head is above and to the left of the others\"."}], "task": "refering", "answer_template": "The \"the long horned sheep whose head is above and to the left of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 121, 141, 142, 143, 144, 163, 164, 165, 166, 185, 186, 187, 188, 189, 208, 209, 210, 231, 232, 253, 254, 255, 276, 277], "bbox": [0.016609375000000003, 0.28485849056603774, 0.27446875, 0.834245283018868], "iscrowd": 0, "area": 13193.346849999998, "rle": {"size": [424, 640], "counts": "`Y51W=001O1O001O1K4M4K5L3L5L3L5L4K4M4K5L3M4L4M2N3M2M4M3M2N2M3N1O2N1N3N1O2N10000O1000000O100000000O1000000O0100O1O2O0O1O100O1O100O101N1O1N2O2M2O2N2M2O2M3N1N3N2M3N1O2N2OZO\\NiEc1]9ZObFd0P9MoF2P92mFMS96lFHT9;kFCU9a0hF_OW9d0hFZOX9j0eFWOY9k0gFTOX9l0jFTOT9m0lFSOS9n0nFROQ9l0RGSOo8k0RGVOm8i0UGVOk8i0WGWOi8f0YGZOg8e0[G[Oe8c0\\G]Od8b0^G^Ob8?aG@_8?bGB^8<dGC\\8<eGE[89gGFY8:hGFW8:iGFX89iGGW88iGHX87iGIW86iGJX86hGJX86gGJZ86fGI[88dGH\\88cGH^88bGG_89`GGa89_GFb8;]GEc8;]GDd8=ZGCg8=YGBh8?WG@j8a0UG_Ok8a0TG_Om8b0RG]OP9c0oF\\OR9d0nF[OT9f0iF[OW9g0gFXO[9i0cFVO^9l0`FSOb9n0[FROf9X21O001O000O2O001O1O001O1O001N2O0HeEcM\\:W2kEgMV:T2PFjMP:Q2WFmMj9n1k0K5K5L4K5KT1lNfno5"}, "label": "the long horned sheep whose head is above and to the left of the others"}]}
{"id": 570878, "image": "COCO_train2014_000000570878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ram in the middle of a herd\"."}], "task": "refering", "answer_template": "The \"a ram in the middle of a herd\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 331, 332, 333], "bbox": [0.26471875, 0.338561320754717, 0.64390625, 0.9886556603773585], "iscrowd": 0, "area": 39927.6749, "rle": {"size": [424, 640], "counts": "\\UV2W1k;d0\\O7I7I7K6K3L4M3O100O1O1O00_Ob0N1O1O1O101N1O101N2O1N2O0O2O1N2O1N2O0O2O1N2N2O1N101N2001O1O1O1O1O001O1O1O001O1O1O1O1O1O3M7I7XEaNY9T3C9G8H8Hl0TO3M3N2M3N3L3N2M3M2O1N3N1O101O000010O0001O0N2N2N3M2N2N3L3M4M3L4M3L4M2M100O100O1O2O0O100O100O100O1O100O1O1000POo010O1000O010000O03iJhHZ4d1ZKe35PK[4\\1bKd3NRK`4[1cKd3GTKe4Y1fKW4[4jKgKS4Y4nKhKP4X4PLjKn3U4TLlKi3U4XLlKf3U4ZLkKe3U4[LkKe3U4\\LkKb3V4_LjK`3V4`LkK_3U4bLkK]3V4cLiK\\3X4eLhKZ3X4fLiKY3W4hLiKW3W4jLiKT3Y4lLfKT3Z4mLfKR3Z4oLfKo2[4QMfKn2Z4SMfKl2[4UMdKh2^4^M\\K]2i4PNkJj1Z5WNfJe1^5[NbJc1_5]NbJa1_5`N`J_1a5bN_J\\1c5cN]J\\1d5dN]JZ1d5cN_J[1c5bN`J^1a5^NcJa1]5^NdJa1]5]NeJc1[5\\NgJc1Z5ZNhJf1X5YNjJf1V5YNkJg1U5WNmJh1U5VNmJi1S5UNoJk1Q5TNoJm1Q5RNoJo1R5nMoJS2R5kMnJU2T5hMlJZ2U5dMkJ]2W5_MjJb2W5\\MiJe2X5YMhJh2Y5UMhJk2[5RMeJo2]5mLdJT3_5hLaJY3b5cL^J^3e5]L\\Jd3h5WLXJj3n5nKSJS4`71N2O1O1O1O1O1O1O1O1O1O1000O100O1O1N2O1N2N2N2N2O1N2O2M3M3N2M3M3N2M3M3N2M3M4M3L3F;ROn0^Ob0J6K5K5J6K4L5J6KmTn2"}, "label": "a ram in the middle of a herd"}]}
{"id": 570878, "image": "COCO_train2014_000000570878.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a herd of sheep standing up in a field\"."}], "task": "refering", "answer_template": "The \"a herd of sheep standing up in a field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 331, 332, 333], "bbox": [0.26471875, 0.338561320754717, 0.64390625, 0.9886556603773585], "iscrowd": 0, "area": 39927.6749, "rle": {"size": [424, 640], "counts": "\\UV2W1k;d0\\O7I7I7K6K3L4M3O100O1O1O00_Ob0N1O1O1O101N1O101N2O1N2O0O2O1N2O1N2O0O2O1N2N2O1N101N2001O1O1O1O1O001O1O1O001O1O1O1O1O1O3M7I7XEaNY9T3C9G8H8Hl0TO3M3N2M3N3L3N2M3M2O1N3N1O101O000010O0001O0N2N2N3M2N2N3L3M4M3L4M3L4M2M100O100O1O2O0O100O100O100O1O100O1O1000POo010O1000O010000O03iJhHZ4d1ZKe35PK[4\\1bKd3NRK`4[1cKd3GTKe4Y1fKW4[4jKgKS4Y4nKhKP4X4PLjKn3U4TLlKi3U4XLlKf3U4ZLkKe3U4[LkKe3U4\\LkKb3V4_LjK`3V4`LkK_3U4bLkK]3V4cLiK\\3X4eLhKZ3X4fLiKY3W4hLiKW3W4jLiKT3Y4lLfKT3Z4mLfKR3Z4oLfKo2[4QMfKn2Z4SMfKl2[4UMdKh2^4^M\\K]2i4PNkJj1Z5WNfJe1^5[NbJc1_5]NbJa1_5`N`J_1a5bN_J\\1c5cN]J\\1d5dN]JZ1d5cN_J[1c5bN`J^1a5^NcJa1]5^NdJa1]5]NeJc1[5\\NgJc1Z5ZNhJf1X5YNjJf1V5YNkJg1U5WNmJh1U5VNmJi1S5UNoJk1Q5TNoJm1Q5RNoJo1R5nMoJS2R5kMnJU2T5hMlJZ2U5dMkJ]2W5_MjJb2W5\\MiJe2X5YMhJh2Y5UMhJk2[5RMeJo2]5mLdJT3_5hLaJY3b5cL^J^3e5]L\\Jd3h5WLXJj3n5nKSJS4`71N2O1O1O1O1O1O1O1O1O1O1000O100O1O1N2O1N2N2N2N2O1N2O2M3M3N2M3M3N2M3M3N2M3M4M3L3F;ROn0^Ob0J6K5K5J6K4L5J6KmTn2"}, "label": "a herd of sheep standing up in a field"}]}
{"id": 407038, "image": "COCO_train2014_000000407038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very wrinkly baby elephant walking around\"."}], "task": "refering", "answer_template": "The \"a very wrinkly baby elephant walking around\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 268, 278, 279, 280, 281, 282, 284, 285, 287, 288, 289, 291, 301, 302, 304, 305, 306, 307, 308, 311, 312, 324, 325, 327, 328, 329, 330, 331, 334, 335], "bbox": [0.09125, 0.30196936542669583, 0.6904843749999999, 0.9313566739606127], "iscrowd": 0, "area": 68183.6825, "rle": {"size": [457, 640], "counts": "dXj02Q>?E4L3M4L3M3M3M3N2M3M3M3N1N3N2N2N2N2N2N2N2M2O1O1O1O1O1N2M3M3L4M3M3M3M2NCREcMl:^2[E\\Mc:e2>1O2N2N3mMaMZHa2e7bNWGa1f8Q2O1O2O0O101O0O2O000O2O001O1O100O10000O100O101N1000O010O010001O1O010O001O001O001O01O00000001iGXJl7h5RH]Jj0HR6k5RIgJ?KW6V6bIUJX6m5cI\\JV6Q7J5K4M2N1O1N3N0000O01000000O1000000O100000O01000000O1000000O1000000O1000000O1000000O`M^K[Kb4d4fKUKZ4j4mKPKS4P5ZLcJe3]5_L`Ja3_5bL_J^3`5dL^J\\3b5gL[JZ3e5hLXJY3g5kLUJV3k5nLPJS3o5d20001N1O1O100O1O1O1O1O1O100O1O1000iI^JnM6U5\\5aL^K_3b4VLiKj3W4mKRLS4n3dKZL]4f3\\KaLd4a3TKeLl4\\3mJjLS5W3fJoLZ5o52O00000O100000000O2O00000O2O00001O0O100000001O000000000@YJ^Hh5^7\\JaHd5R7jJmHV5Q7mJnHS5o6QKoHP5n6SKRIm4k6XKSIi4i6_KRIa4k6`1N2N2N2N3L3N2N2M3M3L4M3M3N2O2M2N2O1N2O1N2O1001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O010O1O2N100O1O1O2O0O1O1O100O2N1O100O1O2N100O1O2N2O1N2N2N2N2O2McLiHKT76nHIa0jLc5]3nIb0S6^OPJ?Q6@QJ>o5BTJ;l5EVJ8k5HXJ5h5KZJ3f5N\\JNe52]JLc55_JHa58aJF_5;cJA^5?dJ_O\\5b0fJ[OZ5e0hJYOX5h0iJVOX5i0kJTOU5m0lJQOT5o0nJoNR5R1oJlNQ5T1WKdNj4\\1]K[Nd4e1aKVN`4i1bKUN_4j1dKSN\\4n1fKoM[4P2hKmMY4R2kKjMU4W2nKeMS4Z2QLbMP4^2TL]Ml3c2XLYMi3g2ZLTMh3l2[LPMe3P3]LnLd3R3[LnLf3R3ZLlLg3T3ZLjLh3V3XLhLj3X3WLeLk3Z3VLeLl3Z3TLdLn3[3TLbLn3]3SLaLP4]3RL`LP4_3_3N1EUEVMm:h2VEUMk:j2WESMk:m2:N3M2O2M2O2N1O4L3M3M4L3M3L3N1O2N2N2N2N2N1O2N1O1O1O1O12N2dDcNf9`1QFeNP:b1fEaNZ:j1YEWNh:T2jDmMW;b22N2N2N2K5L4K5K5L4K6K4K5L4J6Gc0]OQRh2"}, "label": "a very wrinkly baby elephant walking around"}]}
{"id": 407038, "image": "COCO_train2014_000000407038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant calf\"."}], "task": "refering", "answer_template": "The \"an elephant calf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 268, 278, 279, 280, 281, 282, 284, 285, 287, 288, 289, 291, 301, 302, 304, 305, 306, 307, 308, 311, 312, 324, 325, 327, 328, 329, 330, 331, 334, 335], "bbox": [0.09125, 0.30196936542669583, 0.6904843749999999, 0.9313566739606127], "iscrowd": 0, "area": 68183.6825, "rle": {"size": [457, 640], "counts": "dXj02Q>?E4L3M4L3M3M3M3N2M3M3M3N1N3N2N2N2N2N2N2N2M2O1O1O1O1O1N2M3M3L4M3M3M3M2NCREcMl:^2[E\\Mc:e2>1O2N2N3mMaMZHa2e7bNWGa1f8Q2O1O2O0O101O0O2O000O2O001O1O100O10000O100O101N1000O010O010001O1O010O001O001O001O01O00000001iGXJl7h5RH]Jj0HR6k5RIgJ?KW6V6bIUJX6m5cI\\JV6Q7J5K4M2N1O1N3N0000O01000000O1000000O100000O01000000O1000000O1000000O1000000O1000000O`M^K[Kb4d4fKUKZ4j4mKPKS4P5ZLcJe3]5_L`Ja3_5bL_J^3`5dL^J\\3b5gL[JZ3e5hLXJY3g5kLUJV3k5nLPJS3o5d20001N1O1O100O1O1O1O1O1O100O1O1000iI^JnM6U5\\5aL^K_3b4VLiKj3W4mKRLS4n3dKZL]4f3\\KaLd4a3TKeLl4\\3mJjLS5W3fJoLZ5o52O00000O100000000O2O00000O2O00001O0O100000001O000000000@YJ^Hh5^7\\JaHd5R7jJmHV5Q7mJnHS5o6QKoHP5n6SKRIm4k6XKSIi4i6_KRIa4k6`1N2N2N2N3L3N2N2M3M3L4M3M3N2O2M2N2O1N2O1N2O1001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O1O010O1O2N100O1O1O2O0O1O1O100O2N1O100O1O2N100O1O2N2O1N2N2N2N2O2McLiHKT76nHIa0jLc5]3nIb0S6^OPJ?Q6@QJ>o5BTJ;l5EVJ8k5HXJ5h5KZJ3f5N\\JNe52]JLc55_JHa58aJF_5;cJA^5?dJ_O\\5b0fJ[OZ5e0hJYOX5h0iJVOX5i0kJTOU5m0lJQOT5o0nJoNR5R1oJlNQ5T1WKdNj4\\1]K[Nd4e1aKVN`4i1bKUN_4j1dKSN\\4n1fKoM[4P2hKmMY4R2kKjMU4W2nKeMS4Z2QLbMP4^2TL]Ml3c2XLYMi3g2ZLTMh3l2[LPMe3P3]LnLd3R3[LnLf3R3ZLlLg3T3ZLjLh3V3XLhLj3X3WLeLk3Z3VLeLl3Z3TLdLn3[3TLbLn3]3SLaLP4]3RL`LP4_3_3N1EUEVMm:h2VEUMk:j2WESMk:m2:N3M2O2M2O2N1O4L3M3M4L3M3L3N1O2N2N2N2N2N1O2N1O1O1O1O12N2dDcNf9`1QFeNP:b1fEaNZ:j1YEWNh:T2jDmMW;b22N2N2N2K5L4K5K5L4K6K4K5L4J6Gc0]OQRh2"}, "label": "an elephant calf"}]}
{"id": 407038, "image": "COCO_train2014_000000407038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"adult elephant tail\"."}], "task": "refering", "answer_template": "The \"adult elephant tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 182, 183, 202, 203, 205, 206, 225, 226, 228, 229, 248, 251, 252, 271, 273, 274, 275, 295, 296, 297, 298, 320, 321, 343, 344], "bbox": [0.781453125, 0.0018818380743982493, 1.0, 0.8805251641137857], "iscrowd": 0, "area": 32342.137499999997, "rle": {"size": [457, 640], "counts": "P^o61W>2N2M3N2M3N3M2M3N2N2M3N2N2M3N2M3N2N2M4M2C=C=B>B>B9G7J6I7I7I7I7J6I6^N_1hNY19G9G9G9G9F:G9H8I7I7I7J6I7I7I7IkLkKdLo3a3RL_Lg3g3[LXL^3m3eLRLT3T4mLlKl2Z4VMeKd2`4]M`K\\2e4gMZK\\2b4fM]K^2^4cMbKa2X4bMgKb2T4`MkKd2P4]MPLg2j3\\MULh2f3ZMYLj2b3WM^Lm2\\3VMcLn2X3SMhLQ3S3QMlLT3m2oLRMU3i2lLWMX3d2jL[MZ3_2iL`M[3[2fLeM^3V2dLiM`3Q2cLnMa3m1`LSNd3h1^LWNf3c1\\L]Nh3^1ZLaNj3Z1XLeNl3U1VLkNn3P1TLoNP4l0RLSOS4g0nKYOV4a0mK^OW4=kKBY49hKG\\43gKL^4NiKLb4HeK2f4AcK8h4\\O_K>l4VOUKj0V5[NZKe1e800000000000000000000000000000000000000000000000000000000000000000000000000000000000000Z7fHeH"}, "label": "adult elephant tail"}]}
{"id": 407038, "image": "COCO_train2014_000000407038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big elephant leading another small elephant\"."}], "task": "refering", "answer_template": "The \"a big elephant leading another small elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 182, 183, 202, 203, 205, 206, 225, 226, 228, 229, 248, 251, 252, 271, 273, 274, 275, 295, 296, 297, 298, 320, 321, 343, 344], "bbox": [0.781453125, 0.0018818380743982493, 1.0, 0.8805251641137857], "iscrowd": 0, "area": 32342.137499999997, "rle": {"size": [457, 640], "counts": "P^o61W>2N2M3N2M3N3M2M3N2N2M3N2N2M3N2M3N2N2M4M2C=C=B>B>B9G7J6I7I7I7I7J6I6^N_1hNY19G9G9G9G9F:G9H8I7I7I7J6I7I7I7IkLkKdLo3a3RL_Lg3g3[LXL^3m3eLRLT3T4mLlKl2Z4VMeKd2`4]M`K\\2e4gMZK\\2b4fM]K^2^4cMbKa2X4bMgKb2T4`MkKd2P4]MPLg2j3\\MULh2f3ZMYLj2b3WM^Lm2\\3VMcLn2X3SMhLQ3S3QMlLT3m2oLRMU3i2lLWMX3d2jL[MZ3_2iL`M[3[2fLeM^3V2dLiM`3Q2cLnMa3m1`LSNd3h1^LWNf3c1\\L]Nh3^1ZLaNj3Z1XLeNl3U1VLkNn3P1TLoNP4l0RLSOS4g0nKYOV4a0mK^OW4=kKBY49hKG\\43gKL^4NiKLb4HeK2f4AcK8h4\\O_K>l4VOUKj0V5[NZKe1e800000000000000000000000000000000000000000000000000000000000000000000000000000000000000Z7fHeH"}, "label": "a big elephant leading another small elephant"}]}
{"id": 128512, "image": "COCO_train2014_000000128512.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"navy blue luggage with tall handle raised\"."}], "task": "refering", "answer_template": "The \"navy blue luggage with tall handle raised\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 38, 39, 61, 62, 84, 85, 107, 108, 130, 131, 153, 154, 155, 156, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 359, 360, 361, 362, 363, 364, 365, 366, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.608421875, 0.029208333333333333, 0.967421875, 0.9820208333333333], "iscrowd": 0, "area": 59536.71720000001, "rle": {"size": [480, 640], "counts": "fif51a>>B>A?B>B>B>B>A?B>B>B>B>B>A?C=N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2dIfIP4\\6PLgLk0[3UOW61000000000000000000000000000000000000000OhBo0c:QO\\E`2T9`MkFQ4e7oKZH`4X7`KfHb4Z7^KeHc4[7]KdHd4\\7\\KcHd4^7\\KaHe4_7[K`Hf4`7ZK_Hg4a7YK^Hh4b7XK]Hi4c7WK[Hj4f7VKYHk4g7UKXHl4h7TKnGV5R8l00O1000O1000O100000000O10000N2M3L4M3M3M3M30000O10000O10000000000001O001O001O00002N1O2N1N2O2N1O2N1O1O001O00001O00000000000000000000000000001O00000000000000001O00000000001N1000001O1O001O001O001O1O001O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2M2O1O1O2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N1O2XOh0C=D<C`0C=C=C?B`0_O`0TO`a9"}, "label": "navy blue luggage with tall handle raised"}]}
{"id": 5632, "image": "COCO_train2014_000000005632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting on the couch using a laptop\"."}], "task": "refering", "answer_template": "The \"man sitting on the couch using a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 339, 340, 341, 357, 362, 363, 364, 383, 384, 385, 386, 387, 388], "bbox": [0.030343750000000003, 0.11235416666666667, 0.87978125, 0.9887708333333334], "iscrowd": 0, "area": 76131.85070000001, "rle": {"size": [480, 640], "counts": "Wg93k>3L4M3M3M3M2M3N2M3I7I7J6K6J5L4K5L9F:E;F;E:F;D;F9G5J6K5N1N3M3M3M3M3N2N2O100O100O1O100O010O1O100O100O100O1O100O100O10O0100O1O100O100O100O100O100O010O100O11O00000000000000001OO00100O1O1O100nN`IcIa6Z6cIdI^6Y6gIdIZ6Y6jIfIV6W6oIfIR6V6SJhIn5V6VJhIj5^4`IaLi0oNg5`4iIWLd0VOd5b4QJoK=]Oc5c4WJhK9D`5c4_J`K5J]5f4dJXK20[5g4ZLWKg3h4[LWKe3h4\\LWKe3i4Y3O1O1O100O1O001O100O1O1O100O1O1O1O100O1O010000000O1000000O1000000O1000O10O100000000O1000000O100000O01000000O10000001O00001O00001N10001O00001O00001O00001O001N101O1O001O001O001O1O0RJZKc1f4ZN\\Kf1d4WN`Kh1a4UNaKk1`4SNaKm1_4QNcKP2]4nMdKR2\\4mMeKS2\\4jMgKU2Z4iMgKW2Y4gMiKZ2W4dMjK\\2V4[MSLe2n3mL`LR3`3hLfLX3[3gLfLY3Z3eLhLZ3X3fLiLY3X3fLiLY3X3fLhLZ3Y3eLhL[3X3dLiL[3Y3cLhL\\3Y3cLgL]3Z3aLhL^3Y3aLhL^3Y3aLhL_3Y3_LhL`3Y3_LjL^3]3[LeLc3i3oKZLn3P4hKRLW4n3gKVLV4k3iKWLU4k3iKXLT4i3kKYLS4h3lK[LR4e3mK]LQ4d3nK_Lo3b3PL`Ln3b3oKbLn3_3QLcLn3]3QLfLl3[3SLgLk3Z3TLiLi3Y3ULiLi3X3VLkLh3U3VLnLh3S3WLPMf3Q3YLQMe3P3ZLSMc3o2[LSMc3n2\\LUMb3k2]LXM`3i2_LYM_3h2_L\\M^3U7N3N1N3M2N3M5L4K2N3M2O2M3M2O1N0010O01O10O01O0010O01O10O01O010O00100O0010O01O1O010O010O00100O0010O010O01O10O000010O0001O001O00010O00001O0010O0001O000010O01OO100lNaCF`<:eCA[<?jC[OW<f0mCUOR<l0TDnNl;R1Q10O101O0O10000O101N10000O101O0O10000O101O0000000000^OcBA]=?cBA]=>eBA[=?eBA[=>gBAY=?gBAY=>iBAW=?jB@V=?kBAU=?lB@T=?mBAS=?nB@R=?oB@R=`0oB\\OT=c0mB[OU=e0c0001O1O1O2N001O1O1O1O1O001O1O1O1O1O001O11O1O1O1N3L3N2N2M4M2M3N2N2M4M2N2MeCOk9ORF4[<200000000000000000000000000000000000000000000000000O1N2N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2O10000001O000000001O000000001O000000001O000000001O000000001O00000000001O0000001O1O001O1O1O1O1O1O1O2N1O2N1O1O2N1O2N1O2N1O4L5K4L5K4L5K4LXdS1"}, "label": "man sitting on the couch using a laptop"}]}
{"id": 5632, "image": "COCO_train2014_000000005632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a hat using his laptop\"."}], "task": "refering", "answer_template": "The \"a man with a hat using his laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 339, 340, 341, 357, 362, 363, 364, 383, 384, 385, 386, 387, 388], "bbox": [0.030343750000000003, 0.11235416666666667, 0.87978125, 0.9887708333333334], "iscrowd": 0, "area": 76131.85070000001, "rle": {"size": [480, 640], "counts": "Wg93k>3L4M3M3M3M2M3N2M3I7I7J6K6J5L4K5L9F:E;F;E:F;D;F9G5J6K5N1N3M3M3M3M3N2N2O100O100O1O100O010O1O100O100O100O1O100O100O10O0100O1O100O100O100O100O100O010O100O11O00000000000000001OO00100O1O1O100nN`IcIa6Z6cIdI^6Y6gIdIZ6Y6jIfIV6W6oIfIR6V6SJhIn5V6VJhIj5^4`IaLi0oNg5`4iIWLd0VOd5b4QJoK=]Oc5c4WJhK9D`5c4_J`K5J]5f4dJXK20[5g4ZLWKg3h4[LWKe3h4\\LWKe3i4Y3O1O1O100O1O001O100O1O1O100O1O1O1O100O1O010000000O1000000O1000000O1000O10O100000000O1000000O100000O01000000O10000001O00001O00001N10001O00001O00001O00001O001N101O1O001O001O001O1O0RJZKc1f4ZN\\Kf1d4WN`Kh1a4UNaKk1`4SNaKm1_4QNcKP2]4nMdKR2\\4mMeKS2\\4jMgKU2Z4iMgKW2Y4gMiKZ2W4dMjK\\2V4[MSLe2n3mL`LR3`3hLfLX3[3gLfLY3Z3eLhLZ3X3fLiLY3X3fLiLY3X3fLhLZ3Y3eLhL[3X3dLiL[3Y3cLhL\\3Y3cLgL]3Z3aLhL^3Y3aLhL^3Y3aLhL_3Y3_LhL`3Y3_LjL^3]3[LeLc3i3oKZLn3P4hKRLW4n3gKVLV4k3iKWLU4k3iKXLT4i3kKYLS4h3lK[LR4e3mK]LQ4d3nK_Lo3b3PL`Ln3b3oKbLn3_3QLcLn3]3QLfLl3[3SLgLk3Z3TLiLi3Y3ULiLi3X3VLkLh3U3VLnLh3S3WLPMf3Q3YLQMe3P3ZLSMc3o2[LSMc3n2\\LUMb3k2]LXM`3i2_LYM_3h2_L\\M^3U7N3N1N3M2N3M5L4K2N3M2O2M3M2O1N0010O01O10O01O0010O01O10O01O010O00100O0010O01O1O010O010O00100O0010O010O01O10O000010O0001O001O00010O00001O0010O0001O000010O01OO100lNaCF`<:eCA[<?jC[OW<f0mCUOR<l0TDnNl;R1Q10O101O0O10000O101N10000O101O0O10000O101O0000000000^OcBA]=?cBA]=>eBA[=?eBA[=>gBAY=?gBAY=>iBAW=?jB@V=?kBAU=?lB@T=?mBAS=?nB@R=?oB@R=`0oB\\OT=c0mB[OU=e0c0001O1O1O2N001O1O1O1O1O001O1O1O1O1O001O11O1O1O1N3L3N2N2M4M2M3N2N2M4M2N2MeCOk9ORF4[<200000000000000000000000000000000000000000000000000O1N2N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2O10000001O000000001O000000001O000000001O000000001O000000001O00000000001O0000001O1O001O1O1O1O1O1O1O2N1O2N1O1O2N1O2N1O2N1O4L5K4L5K4L5K4LXdS1"}, "label": "a man with a hat using his laptop"}]}
{"id": 5632, "image": "COCO_train2014_000000005632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in the background with glasses watching his phone\"."}], "task": "refering", "answer_template": "The \"a person in the background with glasses watching his phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 365, 366, 367, 390], "bbox": [0.459453125, 0.24887499999999999, 0.9983125000000002, 0.9786041666666666], "iscrowd": 0, "area": 45172.10345000001, "rle": {"size": [480, 640], "counts": "iRZ4>j=k0[Oe0K5L3M4K5L4K5L4K5L3M4K5L4K5L4K5L3M3M2O001O00001N10001O00001N10001O0lNgKaGY4Z8oKcGR4W8WLdGj3X8]LeGd3U8dLhG\\3T8kLiGU3R8TMjGm2P8[MmGe2o7bMmG_2n7]2K4O100O1000000O0100000O10O10O1000000O0100000O10O1000O10000O1000000O1000000O100001O0000000000000000000000000000000000000000000000000RL[Ha0e7\\O^Hd0c7XOaHg0_7VOeHi0[7TOhHl0X7POmHo0T7mNPIR1R7iNRIV1o6fNTIZ1m6bNWI]1j6_NZI`1g6[N]Ie1d6WN`Ih1a6TNcIm1]6nMgIS2X6iMkIY2T6cMPJ]2P6^MUJc2j5TM^Jn2a5fLkJ[3T5YLXKh3P80O2O1M100O10000O100O10000O100O2O000O100O10000O100O10000O10001O000001O000000000000001O000001O0000000001O000000000000002O1RMZE^1h:YNbEd1_:UNjEh1X:oMRFn1P:iMZFR2j9eM_FW2Y;K5K5K5K5K4L3M2N3M2N3M2N3M2N3L3N3M`CBb:<]EGd:6\\EMc:1\\E2e:K[E8d:F[E=a<3N3L3M4M2M3N3L3M4M2M3N3L3M4M2M4M2M3M4M2O20O00100O010O1ONfM^C[2a<2010O01O00001O001O01O01O001O000010O01O001O000010O01O00001O0010O0001O001O001O01O01O001O000010O01O00001O0010O01O00001O2N100O2N1O2N1O2N1O101N1O2N1O1O2N101N1O2NO100O101N100O100O1O100O2O0Og0ZOh0WOj0WOi0VOQ1"}, "label": "a person in the background with glasses watching his phone"}]}
{"id": 5632, "image": "COCO_train2014_000000005632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy wearing glasses\"."}], "task": "refering", "answer_template": "The \"the boy wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 365, 366, 367, 390], "bbox": [0.459453125, 0.24887499999999999, 0.9983125000000002, 0.9786041666666666], "iscrowd": 0, "area": 45172.10345000001, "rle": {"size": [480, 640], "counts": "iRZ4>j=k0[Oe0K5L3M4K5L4K5L4K5L3M4K5L4K5L4K5L3M3M2O001O00001N10001O00001N10001O0lNgKaGY4Z8oKcGR4W8WLdGj3X8]LeGd3U8dLhG\\3T8kLiGU3R8TMjGm2P8[MmGe2o7bMmG_2n7]2K4O100O1000000O0100000O10O10O1000000O0100000O10O1000O10000O1000000O1000000O100001O0000000000000000000000000000000000000000000000000RL[Ha0e7\\O^Hd0c7XOaHg0_7VOeHi0[7TOhHl0X7POmHo0T7mNPIR1R7iNRIV1o6fNTIZ1m6bNWI]1j6_NZI`1g6[N]Ie1d6WN`Ih1a6TNcIm1]6nMgIS2X6iMkIY2T6cMPJ]2P6^MUJc2j5TM^Jn2a5fLkJ[3T5YLXKh3P80O2O1M100O10000O100O10000O100O2O000O100O10000O100O10000O10001O000001O000000000000001O000001O0000000001O000000000000002O1RMZE^1h:YNbEd1_:UNjEh1X:oMRFn1P:iMZFR2j9eM_FW2Y;K5K5K5K5K4L3M2N3M2N3M2N3M2N3L3N3M`CBb:<]EGd:6\\EMc:1\\E2e:K[E8d:F[E=a<3N3L3M4M2M3N3L3M4M2M3N3L3M4M2M4M2M3M4M2O20O00100O010O1ONfM^C[2a<2010O01O00001O001O01O01O001O000010O01O001O000010O01O00001O0010O0001O001O001O01O01O001O000010O01O00001O0010O01O00001O2N100O2N1O2N1O2N1O101N1O2N1O1O2N101N1O2NO100O101N100O100O1O100O2O0Og0ZOh0WOj0WOi0VOQ1"}, "label": "the boy wearing glasses"}]}
{"id": 194050, "image": "COCO_train2014_000000194050.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the snowboard attached to the person in the air\"."}], "task": "refering", "answer_template": "The \"the snowboard attached to the person in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 141, 142, 143, 166, 167, 168, 190, 191, 192, 193, 215, 216], "bbox": [0.104421875, 0.3388290398126464, 0.435484375, 0.6348243559718969], "iscrowd": 0, "area": 3899.1776500000005, "rle": {"size": [427, 640], "counts": "kRl02W=4L4L3N3M2N3N2N101N100O1O010O010O010O01O010O010O10O010O01O010O010O010O010O10O01O010O010O010O001O0O2N2N1O2N1O2O0O2N1O2N1O2O1N1ORQ50mnJ2N3M2N3M3M2N3M210O10O10O10O010O01000O010O0100O0100O0100O010O01000O010O010O10O10O010O10O01000O010O0100O0100O0100O010O01M3M2N3M2N3M2M4MRQ52knJ4K5K4M4M3O010O10O010O0100O010O10O0100O010O10O0100O01O00001O001O001O001O001O0010O01O001O001O001O001O001O1N3J6JRYf4"}, "label": "the snowboard attached to the person in the air"}]}
{"id": 194050, "image": "COCO_train2014_000000194050.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sketch bord , a sketcher on it and flying high\"."}], "task": "refering", "answer_template": "The \"the sketch bord , a sketcher on it and flying high\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 141, 142, 143, 166, 167, 168, 190, 191, 192, 193, 215, 216], "bbox": [0.104421875, 0.3388290398126464, 0.435484375, 0.6348243559718969], "iscrowd": 0, "area": 3899.1776500000005, "rle": {"size": [427, 640], "counts": "kRl02W=4L4L3N3M2N3N2N101N100O1O010O010O010O01O010O010O10O010O01O010O010O010O010O10O01O010O010O010O001O0O2N2N1O2N1O2O0O2N1O2N1O2O1N1ORQ50mnJ2N3M2N3M3M2N3M210O10O10O10O010O01000O010O0100O0100O0100O010O01000O010O010O10O10O010O10O01000O010O0100O0100O0100O010O01M3M2N3M2N3M2M4MRQ52knJ4K5K4M4M3O010O10O010O0100O010O10O0100O010O10O0100O01O00001O001O001O001O001O0010O01O001O001O001O001O001O1N3J6JRYf4"}, "label": "the sketch bord , a sketcher on it and flying high"}]}
{"id": 505347, "image": "COCO_train2014_000000505347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse walking in the water\"."}], "task": "refering", "answer_template": "The \"a brown horse walking in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 265, 267], "bbox": [0.382546875, 0.36343457943925234, 0.75215625, 0.7569859813084112], "iscrowd": 0, "area": 22881.5735, "rle": {"size": [428, 640], "counts": "\\dV3^1j;6K5J6J1O4M4L4L4L4L3M4L4L4K5O2Nb0]O2O1O1O1N2O000O2O00000O101OAUG]Lk8o2R11000000O2O0000000O1000000O101O000O101O001O1N101O001O1OTO^FnMc9o1aFPN^9o1eFoM[9o1iFPNV9o1mFPNR9o1PGPNP9o1SGPNl8o1VGPNj8P2XGoMg8R2YGnMf8R2[GmMe8T2\\GkMc8V2]GiMc8W2_GeMc8\\2W10O00001O001O001O00010O001O001O0O101O0O100O100O10000O100O10000O100O1mFUMe7k2XHXMg7i2VHZMj7f2SH]Mm7d2oG_MP8c2jGcMU8^2aGkM_8V2VGTNj8W300O01N2O1M3J5J7I7J6M23N2N2M3M3M3M3M3M3M3M3M3M3M3L4M3M3N2N2N0000O1000000O10000aNSH_Mm7_2\\HYMe7e2dHTM\\7i2PInLP7P3g1N2N20000O10000O2O000O100O10000O100O1O101N1O1O1O100O1O1O1O1O100M3I8K4M3N2O1000O1O1O1O1O11O1O2O0O1000000O10000O0100000O10001O2M2M3N2N3M2N2N3L3N3M2N3L5L5K6IjWR2"}, "label": "a brown horse walking in the water"}]}
{"id": 505347, "image": "COCO_train2014_000000505347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse , closer to the photographer\"."}], "task": "refering", "answer_template": "The \"a horse , closer to the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 265, 267], "bbox": [0.382546875, 0.36343457943925234, 0.75215625, 0.7569859813084112], "iscrowd": 0, "area": 22881.5735, "rle": {"size": [428, 640], "counts": "\\dV3^1j;6K5J6J1O4M4L4L4L4L3M4L4L4K5O2Nb0]O2O1O1O1N2O000O2O00000O101OAUG]Lk8o2R11000000O2O0000000O1000000O101O000O101O001O1N101O001O1OTO^FnMc9o1aFPN^9o1eFoM[9o1iFPNV9o1mFPNR9o1PGPNP9o1SGPNl8o1VGPNj8P2XGoMg8R2YGnMf8R2[GmMe8T2\\GkMc8V2]GiMc8W2_GeMc8\\2W10O00001O001O001O00010O001O001O0O101O0O100O100O10000O100O10000O100O1mFUMe7k2XHXMg7i2VHZMj7f2SH]Mm7d2oG_MP8c2jGcMU8^2aGkM_8V2VGTNj8W300O01N2O1M3J5J7I7J6M23N2N2M3M3M3M3M3M3M3M3M3M3M3L4M3M3N2N2N0000O1000000O10000aNSH_Mm7_2\\HYMe7e2dHTM\\7i2PInLP7P3g1N2N20000O10000O2O000O100O10000O100O1O101N1O1O1O100O1O1O1O1O100M3I8K4M3N2O1000O1O1O1O1O11O1O2O0O1000000O10000O0100000O10001O2M2M3N2N3M2N2N3L3N3M2N3L5L5K6IjWR2"}, "label": "a horse , closer to the photographer"}]}
{"id": 505347, "image": "COCO_train2014_000000505347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse in back\"."}], "task": "refering", "answer_template": "The \"horse in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 154, 155, 178, 221, 222, 245], "bbox": [0.6285624999999999, 0.33175233644859814, 0.767765625, 0.7030140186915887], "iscrowd": 0, "area": 1623.9057499999988, "rle": {"size": [428, 640], "counts": "VYX5a0h<N2N3O22M20001O001O00010O011O1O1O1O1O2N1O2N1N3[LQOhJn0X5ROiJm0W5TOiJk0W5UOjJj0V5VOmJg0S5YOQKc0n4^OSKa0m4_OTK`0l4AWK;i4E[K7e4I_K3a4MbK0]41eKM[43_4000O2OO100O10O01O1N2O0103M2N1O0O2O0O2O0O2N2O2M3N2M3N2M4L3O19H0O1O2N1O1O1O000010O1O001O3L2ON10000N3L4MgRn1"}, "label": "horse in back"}]}
{"id": 267774, "image": "COCO_train2014_000000267774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black pants and a blue jacket skating\"."}], "task": "refering", "answer_template": "The \"a man with black pants and a blue jacket skating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 80, 81, 82, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 219], "bbox": [0.39070312500000004, 0.19245901639344265, 0.60271875, 0.6100702576112412], "iscrowd": 0, "area": 12336.085100000006, "rle": {"size": [427, 640], "counts": "a]X32V=3N2O1O1O1QCHj<9UCHk<=N2N2N1O2N2N1O1O1O1N2O1O1OC=1O001O1QE[OW9e0fF@X9a0gFBV9>kFCR9?nFBP9?PGCm8>SGBl8>UGCi8>WGCf8?ZGBd8?\\GAb8`0_GA^8a0bG@[8b0eG_OY8a0hG_OW8a0fGDW8=gGGW8:fGJX86eGOY81eG2Y8OdG5[8KcG8\\8IaG:^8G^G=a8D\\G`0b8BZGa0e8BVGa0i8BRGa0m8BnFa0P9ClF?P9P2N1000010O01O0N2N2N2N2M010O2M3F:J62O001O1O1000O101O00aGlLg6T3WIQMf6n2[ISMd6m2[IUMe6j2[IWMe6h2[IYMe6f2[IZMf6d2[I]Md6c2\\I^Md6a2\\I_Me6`2[IaMe6^2XIeMj6Y2bHPMHk0f7T2^HbNa7]1\\HfNf7Y1VHkNj7U1nGSOS8l0lGUOT8k0kGVOV8i0jGVOW8i0iGXOW8h0iGXOX8l22O1N2N2O0N3L3N2M3M3M3D=C<G9O1N2N2N2O1N2N2O1N3N2M2O2M2O2M3ZN\\E`0j:WObEa0c:WOeEc0g;I>BWhY3"}, "label": "a man with black pants and a blue jacket skating"}]}
{"id": 267774, "image": "COCO_train2014_000000267774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy riding on the concrete wall\"."}], "task": "refering", "answer_template": "The \"the boy riding on the concrete wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 80, 81, 82, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 219], "bbox": [0.39070312500000004, 0.19245901639344265, 0.60271875, 0.6100702576112412], "iscrowd": 0, "area": 12336.085100000006, "rle": {"size": [427, 640], "counts": "a]X32V=3N2O1O1O1QCHj<9UCHk<=N2N2N1O2N2N1O1O1O1N2O1O1OC=1O001O1QE[OW9e0fF@X9a0gFBV9>kFCR9?nFBP9?PGCm8>SGBl8>UGCi8>WGCf8?ZGBd8?\\GAb8`0_GA^8a0bG@[8b0eG_OY8a0hG_OW8a0fGDW8=gGGW8:fGJX86eGOY81eG2Y8OdG5[8KcG8\\8IaG:^8G^G=a8D\\G`0b8BZGa0e8BVGa0i8BRGa0m8BnFa0P9ClF?P9P2N1000010O01O0N2N2N2N2M010O2M3F:J62O001O1O1000O101O00aGlLg6T3WIQMf6n2[ISMd6m2[IUMe6j2[IWMe6h2[IYMe6f2[IZMf6d2[I]Md6c2\\I^Md6a2\\I_Me6`2[IaMe6^2XIeMj6Y2bHPMHk0f7T2^HbNa7]1\\HfNf7Y1VHkNj7U1nGSOS8l0lGUOT8k0kGVOV8i0jGVOW8i0iGXOW8h0iGXOX8l22O1N2N2O0N3L3N2M3M3M3D=C<G9O1N2N2N2O1N2N2O1N3N2M2O2M2O2M3ZN\\E`0j:WObEa0c:WOeEc0g;I>BWhY3"}, "label": "the boy riding on the concrete wall"}]}
{"id": 513541, "image": "COCO_train2014_000000513541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the old plane is painted dark blue with white lettering and numbers\"."}], "task": "refering", "answer_template": "The \"the old plane is painted dark blue with white lettering and numbers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 288, 289, 290, 292, 293, 294, 311, 312, 313, 316, 317, 334, 335, 336], "bbox": [0.0, 0.37274999999999997, 0.922296875, 0.8592291666666666], "iscrowd": 0, "area": 61167.99409999997, "rle": {"size": [480, 640], "counts": "c84l>9fALV=o00001O00000000000000000000001O000000000000000000001O000000000000000000001O0O1000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000O100000000000001O000000000000000000001O00000000000000000J6D<C>C<H8N2O1O100O100O100O2O0001O1O1O1O1O1O1O2N101N1O1O2N1O2N1ORDcMZ;Q32O0O2N1O1O2N1O2N1^OXDkMi;S2YDkMh;S2ZDlMg;S2e0M2010O0010O010O0O101N1O2N1O1O1O00O100O100O01000O100O100O100O10000O010O100O10000O100O100O10O10O100O100O10000O100O10O010000O100O100O10000O1O001O1O1O100O1O1O1O1O1N2O0O2N2O1N2O1N2O1N2O1N2O0O2N200O100O10000O10O10O10000O10000O100O10O10O10000O1000O0100O10000O10O10O10000O100O01000O10000O10O10O100O10000O01000O10000O10O010000O10000O01000O100O1000O010000O6K8kF_K[7h4ZH_Kb7j4QH^Kl7j4gG^KU8j5M2N1N101O1N2O001N2O1N10001N10000O10000O2O\\M\\HlMd7T2aHgM^7Z2eHcM\\7[2fHdM[7[2gHcMZ7\\2gHcM[7[2fHdM[7[2gHcM[7Z2gHeMZ7Z2gHeMZ7Z2hHdMZ7Z2gHeMZ7Z2gHeMZ7Z2gHdM\\7Y2fHfM[7Y2eHfM^7X2cHgM]7Y2cHfM^7Z2cHeM]7Z2dHeM]7[2dHdM[7]2eHbM\\7^2eHaM[7_2eH`M\\7`2eH^M[7c2fH[M[7d2gH[MY7e2hHYMY7g2hHWMX7j2iHUMW7k2jHSMX7l2iHQMY7o2hHmL[7S3hHgL\\7X3iHaLY7_3lHYLX7f3Q2O100000O10000O100000O10O100000000O100000O1000O100000000O1000O1000O10000000000O100000000O10000000000O100000000O100O100O100O100]FPLU8S4[G_La8e3TGfLg8k4O1O001O1O001O1O001O001N1000001O000000PNYGfMh8Z2ZGdMf8[2\\GdMd8\\2]GbMd8]2_GaMa8^2aGaM_8]2dGaM]8^2fG`MZ8_2hG`MW8`2mG\\MT8c2oG[MQ8d2RHYMo7f2THXMl7b2]HZMd7e2_HYMa7g2aHVM`7j2bHSM_7l2cHRM^7n2cHQM]7o2dHoL]7Q3dHkL_7T3W20000O10000O10000O10001N101O0O101O001N3N1O1N3N1O2M2O1O2M2O2N1N3N1N2M4L3N3L3N2M4K4L4L5L3L4^Oc0N1O100O1O1O2N0001O01O0000010O00001O01O0001O00001O00001O002M4L4L^Vg0"}, "label": "the old plane is painted dark blue with white lettering and numbers"}]}
{"id": 513541, "image": "COCO_train2014_000000513541.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane with the number 215 on it\"."}], "task": "refering", "answer_template": "The \"an airplane with the number 215 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 156, 157, 165, 166, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 288, 289, 290, 292, 293, 294, 311, 312, 313, 316, 317, 334, 335, 336], "bbox": [0.0, 0.37274999999999997, 0.922296875, 0.8592291666666666], "iscrowd": 0, "area": 61167.99409999997, "rle": {"size": [480, 640], "counts": "c84l>9fALV=o00001O00000000000000000000001O000000000000000000001O000000000000000000001O0O1000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000O100000000000001O000000000000000000001O00000000000000000J6D<C>C<H8N2O1O100O100O100O2O0001O1O1O1O1O1O1O2N101N1O1O2N1O2N1ORDcMZ;Q32O0O2N1O1O2N1O2N1^OXDkMi;S2YDkMh;S2ZDlMg;S2e0M2010O0010O010O0O101N1O2N1O1O1O00O100O100O01000O100O100O100O10000O010O100O10000O100O100O10O10O100O100O10000O100O10O010000O100O100O10000O1O001O1O1O100O1O1O1O1O1N2O0O2N2O1N2O1N2O1N2O1N2O0O2N200O100O10000O10O10O10000O10000O100O10O10O10000O1000O0100O10000O10O10O10000O100O01000O10000O10O10O100O10000O01000O10000O10O010000O10000O01000O100O1000O010000O6K8kF_K[7h4ZH_Kb7j4QH^Kl7j4gG^KU8j5M2N1N101O1N2O001N2O1N10001N10000O10000O2O\\M\\HlMd7T2aHgM^7Z2eHcM\\7[2fHdM[7[2gHcMZ7\\2gHcM[7[2fHdM[7[2gHcM[7Z2gHeMZ7Z2gHeMZ7Z2hHdMZ7Z2gHeMZ7Z2gHeMZ7Z2gHdM\\7Y2fHfM[7Y2eHfM^7X2cHgM]7Y2cHfM^7Z2cHeM]7Z2dHeM]7[2dHdM[7]2eHbM\\7^2eHaM[7_2eH`M\\7`2eH^M[7c2fH[M[7d2gH[MY7e2hHYMY7g2hHWMX7j2iHUMW7k2jHSMX7l2iHQMY7o2hHmL[7S3hHgL\\7X3iHaLY7_3lHYLX7f3Q2O100000O10000O100000O10O100000000O100000O1000O100000000O1000O1000O10000000000O100000000O10000000000O100000000O100O100O100O100]FPLU8S4[G_La8e3TGfLg8k4O1O001O1O001O1O001O001N1000001O000000PNYGfMh8Z2ZGdMf8[2\\GdMd8\\2]GbMd8]2_GaMa8^2aGaM_8]2dGaM]8^2fG`MZ8_2hG`MW8`2mG\\MT8c2oG[MQ8d2RHYMo7f2THXMl7b2]HZMd7e2_HYMa7g2aHVM`7j2bHSM_7l2cHRM^7n2cHQM]7o2dHoL]7Q3dHkL_7T3W20000O10000O10000O10001N101O0O101O001N3N1O1N3N1O2M2O1O2M2O2N1N3N1N2M4L3N3L3N2M4K4L4L5L3L4^Oc0N1O100O1O1O2N0001O01O0000010O00001O01O0001O00001O00001O002M4L4L^Vg0"}, "label": "an airplane with the number 215 on it"}]}
{"id": 390663, "image": "COCO_train2014_000000390663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a well - dressed man in an old photo on the left side of another well - dressed man\"."}], "task": "refering", "answer_template": "The \"a well - dressed man in an old photo on the left side of another well - dressed man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 101, 120, 121, 122, 123, 124, 144, 145, 146, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376, 396, 397, 398, 399], "bbox": [0.170421875, 0.19904109589041094, 0.48743749999999997, 0.9636986301369863], "iscrowd": 0, "area": 51056.91230000001, "rle": {"size": [511, 640], "counts": "_ff1f0T?;I4L3L4M2N3L3C_NYBd1c=bNXBa1d=?M4M3L4@`M[Cd2`<`0K6J6J6I7K4M4K5K5L4L3M4_EWKf9m4jEeKP:Z5J6L4L1N2O1O1O1N2O001N2O1O1O1N3]HUIm5m6oIVIQ6k6kIXIT6j6hIYIW6i6dI[I\\6f6`I]I_6e6]I^Ie4]O^LW7hN`Ih4]O]LU7gNaIk4\\O]LS7eNdIn4ZO[LT7dNdIP5YO[LU7aNeIT5WOZLU7^NgIW5UOZL]8f3dGXL]8g3dGYL\\8g3eGWL\\8h3eGWL\\8i3eGVL[8i3fGWLZ8i3fGWLZ8h3hGVLY8i3hGWLX8i3PHoKP8P4RHoKn7a3WGjKl0c0n7a3eH^L[7`3hH_LX7_3kH`LU7^3nH`LS7_3PI_LP7_3RIaLn6]3UIbLk6]3WIaLj6^3XI`Li6`3WI^Lk6a3VI^Lk6b3UI\\Lm6c3SI\\Lo6c3VIXLk6h3ZIQLh6n3e200O100O100O100O100O100O100O1000000000000001O001O1O001O1O1O1O1O1O2N1O1O1O1O1O2N1O3M2eF`Kh50]Ic4e0aKj50`Ia4a0bKo5N_Ic4<cKT6L^Ic48fKY6H]If45eK]6F]Ie44hK^6HYI`47kK^6c5`I_J`6a5[IeJd6[5RIPKm6P5RIRKm6m4TITKk6l4TIUKl6k4TIUKl6k4SIVKl6k4TIUKl6k4SIVKm6j4SIVKm6j4RIWKn6j4PIWKP7i4PIWKP7i4oHXKP7i4nHYKR7h4lHYKT7g4jH[KV7e4iH\\KW7d4gH^KY7c4dH_K[7d6O0O2O0O2N101N101N101N101N101N100O2O001O0010O01O001O3M2N3N:E1XMWFXOi9f0[FXOf9f0]FYOc9d0`F[Oa9XNTF`1?6_9UNVFb1a05\\9TNVFd1c05Y9RNWFg1f04T9PNZFj1f03e9JaF2c9HfF1a9FlFN`9EmFMW9Lc3H7J7K4L^\\S5"}, "label": "a well - dressed man in an old photo on the left side of another well - dressed man"}]}
{"id": 390663, "image": "COCO_train2014_000000390663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the left with the paisley tie\"."}], "task": "refering", "answer_template": "The \"the man on the left with the paisley tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 101, 120, 121, 122, 123, 124, 144, 145, 146, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376, 396, 397, 398, 399], "bbox": [0.170421875, 0.19904109589041094, 0.48743749999999997, 0.9636986301369863], "iscrowd": 0, "area": 51056.91230000001, "rle": {"size": [511, 640], "counts": "_ff1f0T?;I4L3L4M2N3L3C_NYBd1c=bNXBa1d=?M4M3L4@`M[Cd2`<`0K6J6J6I7K4M4K5K5L4L3M4_EWKf9m4jEeKP:Z5J6L4L1N2O1O1O1N2O001N2O1O1O1N3]HUIm5m6oIVIQ6k6kIXIT6j6hIYIW6i6dI[I\\6f6`I]I_6e6]I^Ie4]O^LW7hN`Ih4]O]LU7gNaIk4\\O]LS7eNdIn4ZO[LT7dNdIP5YO[LU7aNeIT5WOZLU7^NgIW5UOZL]8f3dGXL]8g3dGYL\\8g3eGWL\\8h3eGWL\\8i3eGVL[8i3fGWLZ8i3fGWLZ8h3hGVLY8i3hGWLX8i3PHoKP8P4RHoKn7a3WGjKl0c0n7a3eH^L[7`3hH_LX7_3kH`LU7^3nH`LS7_3PI_LP7_3RIaLn6]3UIbLk6]3WIaLj6^3XI`Li6`3WI^Lk6a3VI^Lk6b3UI\\Lm6c3SI\\Lo6c3VIXLk6h3ZIQLh6n3e200O100O100O100O100O100O100O1000000000000001O001O1O001O1O1O1O1O1O2N1O1O1O1O1O2N1O3M2eF`Kh50]Ic4e0aKj50`Ia4a0bKo5N_Ic4<cKT6L^Ic48fKY6H]If45eK]6F]Ie44hK^6HYI`47kK^6c5`I_J`6a5[IeJd6[5RIPKm6P5RIRKm6m4TITKk6l4TIUKl6k4TIUKl6k4SIVKl6k4TIUKl6k4SIVKm6j4SIVKm6j4RIWKn6j4PIWKP7i4PIWKP7i4oHXKP7i4nHYKR7h4lHYKT7g4jH[KV7e4iH\\KW7d4gH^KY7c4dH_K[7d6O0O2O0O2N101N101N101N101N101N100O2O001O0010O01O001O3M2N3N:E1XMWFXOi9f0[FXOf9f0]FYOc9d0`F[Oa9XNTF`1?6_9UNVFb1a05\\9TNVFd1c05Y9RNWFg1f04T9PNZFj1f03e9JaF2c9HfF1a9FlFN`9EmFMW9Lc3H7J7K4L^\\S5"}, "label": "the man on the left with the paisley tie"}]}
{"id": 390663, "image": "COCO_train2014_000000390663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting in the right side of the image\"."}], "task": "refering", "answer_template": "The \"the man sitting in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 380, 381, 382, 383, 384, 403, 404, 405, 406, 407, 408], "bbox": [0.4894375, 0.13767123287671232, 0.819171875, 0.9636399217221134], "iscrowd": 0, "area": 60797.51065, "rle": {"size": [511, 640], "counts": "Y^l42j?8I7H8H8I7H8I7H8H8I7H8H8VDTMU:S3cETMV:T3bERMX:T3cERMW:R3dETMW:P3dEWMW:l2dEZMY:h2bE_Mh7bNeJf5U5]JjJf5R5\\JmJh5m4[JQKi5j4ZJUKj5f4XJXKl5c4WJ\\Km5^4VJaKn5Z4TJdKP6X4RJgKR6T4PJkKS6R4nIlKU6Q4mInKT6P4oImKS6Q4oInKS6P4nIoKS6o3oIoKS6o3oIPLR6n3PJQLQ6n3PJPLS6m3oIRLR6l3QJQLQ6m3QJRLP6m3RJQLo5m3SJQLo5m3TJQLY7a2iH]MX7c2jH[MV7d2lH[MT7d2oHYMR7e2RIYMn6f2TIYMl6e2XIXMi6g2YIXMg6f2]IWMd6g2_IXMa6g2bIWM^6g2eIWM\\6g2jIUMV6j2nISMR6k2SJQMn5n2WJnLi5P3d3N2O1N2O1N2N2O11O000O2O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O001O00<D9ZESLh8n3VGTLi8l3UGXLi8i3UGZLi8f3UG]Lj8d3TG_Lj8a3TGcLj8]3UGeLj8\\3SGhLk8X3TGkLj8V3TGlLk8T3SGPMk8Q3SGQMl8o2RGSMn8m2QGTMo8m2nFUMR9k2mFVMS9k2jFWMV9i2iFXMW9Q3^FQMb9f4O100000001O0000000000000000000000000000000TOhIQHX6o7kInGU6Q8nImGS6Q8PJmGP6R8SJlGm5S8VJkGj5T8ZJiGf5V8]JhGc5W8`JgG`5Y8bJdG_5[8eJbG[5]8hJaGX5^8U1O1O1O1O100O1O1O2N1O1O3M2O1N3M2N2N3M2N3WLbGL`80eGN\\8OgGO\\8LhG2Z8KiG3Y8JjG4Y8GlG6^8_OeG?f8UO]Gi0n8jNVGU1T9_NoF_1[9VNiFg1b9lMbFR2i9bMZF\\2Q:XMRFf2f;M2N2O1N2N3M2N2N2XOh0UO]bi1"}, "label": "the man sitting in the right side of the image"}]}
{"id": 390663, "image": "COCO_train2014_000000390663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a bench with his arms folded across his chest and stomach\"."}], "task": "refering", "answer_template": "The \"a man sitting on a bench with his arms folded across his chest and stomach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 380, 381, 382, 383, 384, 403, 404, 405, 406, 407, 408], "bbox": [0.4894375, 0.13767123287671232, 0.819171875, 0.9636399217221134], "iscrowd": 0, "area": 60797.51065, "rle": {"size": [511, 640], "counts": "Y^l42j?8I7H8H8I7H8I7H8H8I7H8H8VDTMU:S3cETMV:T3bERMX:T3cERMW:R3dETMW:P3dEWMW:l2dEZMY:h2bE_Mh7bNeJf5U5]JjJf5R5\\JmJh5m4[JQKi5j4ZJUKj5f4XJXKl5c4WJ\\Km5^4VJaKn5Z4TJdKP6X4RJgKR6T4PJkKS6R4nIlKU6Q4mInKT6P4oImKS6Q4oInKS6P4nIoKS6o3oIoKS6o3oIPLR6n3PJQLQ6n3PJPLS6m3oIRLR6l3QJQLQ6m3QJRLP6m3RJQLo5m3SJQLo5m3TJQLY7a2iH]MX7c2jH[MV7d2lH[MT7d2oHYMR7e2RIYMn6f2TIYMl6e2XIXMi6g2YIXMg6f2]IWMd6g2_IXMa6g2bIWM^6g2eIWM\\6g2jIUMV6j2nISMR6k2SJQMn5n2WJnLi5P3d3N2O1N2O1N2N2O11O000O2O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O001O00<D9ZESLh8n3VGTLi8l3UGXLi8i3UGZLi8f3UG]Lj8d3TG_Lj8a3TGcLj8]3UGeLj8\\3SGhLk8X3TGkLj8V3TGlLk8T3SGPMk8Q3SGQMl8o2RGSMn8m2QGTMo8m2nFUMR9k2mFVMS9k2jFWMV9i2iFXMW9Q3^FQMb9f4O100000001O0000000000000000000000000000000TOhIQHX6o7kInGU6Q8nImGS6Q8PJmGP6R8SJlGm5S8VJkGj5T8ZJiGf5V8]JhGc5W8`JgG`5Y8bJdG_5[8eJbG[5]8hJaGX5^8U1O1O1O1O100O1O1O2N1O1O3M2O1N3M2N2N3M2N3WLbGL`80eGN\\8OgGO\\8LhG2Z8KiG3Y8JjG4Y8GlG6^8_OeG?f8UO]Gi0n8jNVGU1T9_NoF_1[9VNiFg1b9lMbFR2i9bMZF\\2Q:XMRFf2f;M2N2O1N2N3M2N2N2XOh0UO]bi1"}, "label": "a man sitting on a bench with his arms folded across his chest and stomach"}]}
{"id": 259591, "image": "COCO_train2014_000000259591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a seat cushion that has a giraffe on it\"."}], "task": "refering", "answer_template": "The \"a chair with a seat cushion that has a giraffe on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 327, 328, 329, 340, 341, 342, 343, 344, 345, 346, 357, 358, 359, 360, 361, 362, 363, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.002875, 0.6109843749999999, 0.3965416666666667, 0.988140625], "iscrowd": 0, "area": 40098.51845, "rle": {"size": [640, 480], "counts": "]P1_4a?R1mNS1nNh0XO000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000O1000000000000O010000000000000O100000000000000O100000000000O10O100000000000000O10001O1O001O1N2O001O1O3M4L3M4K4M1O1O1O2N1O1O1N2O1O1O2N1O1O1O1N2O2N1O1O1O1O1O2M21O1O010000O1O1N2N2N2N2N2N2N2N2N2ZLQBW1Q>gNgBa0[=\\O]CMe<0TDXOo;e0h3M4M2L5K4L5K4L_hd5"}, "label": "a chair with a seat cushion that has a giraffe on it"}]}
{"id": 259591, "image": "COCO_train2014_000000259591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pillow with an image of a giraffe rests on a chair next o a desk\"."}], "task": "refering", "answer_template": "The \"a pillow with an image of a giraffe rests on a chair next o a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 327, 328, 329, 340, 341, 342, 343, 344, 345, 346, 357, 358, 359, 360, 361, 362, 363, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.002875, 0.6109843749999999, 0.3965416666666667, 0.988140625], "iscrowd": 0, "area": 40098.51845, "rle": {"size": [640, 480], "counts": "]P1_4a?R1mNS1nNh0XO000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000O1000000000000O010000000000000O100000000000000O100000000000O10O100000000000000O10001O1O001O1N2O001O1O3M4L3M4K4M1O1O1O2N1O1O1N2O1O1O2N1O1O1O1N2O2N1O1O1O1O1O2M21O1O010000O1O1N2N2N2N2N2N2N2N2N2ZLQBW1Q>gNgBa0[=\\O]CMe<0TDXOo;e0h3M4M2L5K4L5K4L_hd5"}, "label": "a pillow with an image of a giraffe rests on a chair next o a desk"}]}
{"id": 382472, "image": "COCO_train2014_000000382472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing the snow kit and dress with his parents\"."}], "task": "refering", "answer_template": "The \"a girl wearing the snow kit and dress with his parents\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 101, 122, 123, 124, 145, 146, 147, 148, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 214, 215, 216, 237, 238, 239, 259, 260, 261, 262, 283, 285, 308], "bbox": [0.28546875, 0.21395833333333333, 0.456078125, 0.7837916666666666], "iscrowd": 0, "area": 16739.044499999996, "rle": {"size": [480, 640], "counts": "XPf22l>3M2O2M3M2N3M2O1N2_D^Ob8d0hFc0f8^OSGT1`8nNZGY1V8WOcGP1S8YOgGm0R8ZOiGl0P8ZOjGl0R8ZOeGn0W8YO\\GP1b8h2N2N20000000000O100kNRJXIn5d6^JTIb5h6dJVI[5g6lJVIS5h6RKVIm4k6UKRIk4n6YKnHg4S7b10M4L3N3L3I8G9B=B?C=H82N3N1N2N3M2N2O1N3M2N2N2O2M2N2O101N5K6J5L2M1O100000000f0[ON1O5kG_I`7f6RHbIo7g61dL]HXOe7b0hHVOY7d0QIWOP7c0YIXOi6<fI@[6?iI]OX6b0lIYOV6f0nIVOS6i0QJROP6m0UJnNm5Q1WJjNl5T1XJgNk5W1YJeNi5Y1[JbNg5]1^J]Ne5a1aJXNd5d1cJUNb5e1dJUNc5e1o3M3M2N3K4M3M3M4L3M7I8BZmR5"}, "label": "a girl wearing the snow kit and dress with his parents"}]}
{"id": 382472, "image": "COCO_train2014_000000382472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in blue jumpsuit\"."}], "task": "refering", "answer_template": "The \"boy in blue jumpsuit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 101, 122, 123, 124, 145, 146, 147, 148, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 214, 215, 216, 237, 238, 239, 259, 260, 261, 262, 283, 285, 308], "bbox": [0.28546875, 0.21395833333333333, 0.456078125, 0.7837916666666666], "iscrowd": 0, "area": 16739.044499999996, "rle": {"size": [480, 640], "counts": "XPf22l>3M2O2M3M2N3M2O1N2_D^Ob8d0hFc0f8^OSGT1`8nNZGY1V8WOcGP1S8YOgGm0R8ZOiGl0P8ZOjGl0R8ZOeGn0W8YO\\GP1b8h2N2N20000000000O100kNRJXIn5d6^JTIb5h6dJVI[5g6lJVIS5h6RKVIm4k6UKRIk4n6YKnHg4S7b10M4L3N3L3I8G9B=B?C=H82N3N1N2N3M2N2O1N3M2N2N2O2M2N2O101N5K6J5L2M1O100000000f0[ON1O5kG_I`7f6RHbIo7g61dL]HXOe7b0hHVOY7d0QIWOP7c0YIXOi6<fI@[6?iI]OX6b0lIYOV6f0nIVOS6i0QJROP6m0UJnNm5Q1WJjNl5T1XJgNk5W1YJeNi5Y1[JbNg5]1^J]Ne5a1aJXNd5d1cJUNb5e1dJUNc5e1o3M3M2N3K4M3M3M4L3M7I8BZmR5"}, "label": "boy in blue jumpsuit"}]}
{"id": 382472, "image": "COCO_train2014_000000382472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on skiis\"."}], "task": "refering", "answer_template": "The \"a woman on skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336], "bbox": [0.546296875, 0.11358333333333334, 0.696296875, 0.8419791666666666], "iscrowd": 0, "area": 22758.583, "rle": {"size": [480, 640], "counts": "VXT5`0W>=G8H9J6I7I7Jd0[Oh2YM5^MhJSK_5`4PKWKW5c4QKRKV5j4RKkJU5Q5SKcJU5Y5SKZJW5a5W2L4K5L4L4M3M3N4L3M3M4L3M3M3M2N0001O00000YLmKaLS4_3RL\\Lo3b3WLYLi3g3]LSLc3m3bLnK^3R4gLhKZ3X4jLdKW3[4mLaKS3g4jLTKV3T5fL_Jd3h5XLjIS4S7]KUHm4e8;g0XO15K6J5K6K5J3M03M1O0000O10O10O1000000O0L5L4K6K4L6I6^MYIdLl6R3`IfLg6P3eIhL`6n2lIjLZ6m2RJkLT6k2XJmLm5l2\\JlLj5m2`JkLf5m2cJkLc5n2gJjL^5o2_3H7UOl0G9G9G9H8I7I7I7Iihj2"}, "label": "a woman on skiis"}]}
{"id": 382472, "image": "COCO_train2014_000000382472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman skiing with a man and a boy\"."}], "task": "refering", "answer_template": "The \"a woman skiing with a man and a boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336], "bbox": [0.546296875, 0.11358333333333334, 0.696296875, 0.8419791666666666], "iscrowd": 0, "area": 22758.583, "rle": {"size": [480, 640], "counts": "VXT5`0W>=G8H9J6I7I7Jd0[Oh2YM5^MhJSK_5`4PKWKW5c4QKRKV5j4RKkJU5Q5SKcJU5Y5SKZJW5a5W2L4K5L4L4M3M3N4L3M3M4L3M3M3M2N0001O00000YLmKaLS4_3RL\\Lo3b3WLYLi3g3]LSLc3m3bLnK^3R4gLhKZ3X4jLdKW3[4mLaKS3g4jLTKV3T5fL_Jd3h5XLjIS4S7]KUHm4e8;g0XO15K6J5K6K5J3M03M1O0000O10O10O1000000O0L5L4K6K4L6I6^MYIdLl6R3`IfLg6P3eIhL`6n2lIjLZ6m2RJkLT6k2XJmLm5l2\\JlLj5m2`JkLf5m2cJkLc5n2gJjL^5o2_3H7UOl0G9G9G9H8I7I7I7Iihj2"}, "label": "a woman skiing with a man and a boy"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing behind person hitting ball\"."}], "task": "refering", "answer_template": "The \"man standing behind person hitting ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 27, 28, 29, 30, 31, 45, 46, 47, 48, 49, 63, 64, 65, 66, 81, 82, 83, 84, 99, 100, 101, 102, 118, 119, 120, 121, 138, 139], "bbox": [0.4973, 0.0, 0.74098, 0.5957336956521739], "iscrowd": 0, "area": 14995.9795, "rle": {"size": [368, 500], "counts": "\\ai27U;8I7I6J7G6I6J7I6N3N1O2O0O1O2N2N3N2M3M3M3]H^MS6e2QIZNi6\\3J6J6J6I7J7I6I7J5KN3M3M3L4M3M3M3L4M3M3M3L4L4L4L4K5L4L4L5L32N1O1O2N1O1O1O2O0O1O2N1O1O1O2N100O2N3M3M3M3N3L3N2M3M2O1N3N1N2O2N1O1O2N1O1O7I8I3LO0QMiJEX5ROnJmN2m1Q5UORKdN5T2j4VOWK]N6Z2d4WO[KVN9`2]4YO_Lc0a3\\OcLa0^3^OeL`0[3^OjL>W3@mL=T3BoL;R3CSM9n2EVM8l2FWM7j2G[M5f2J]M3d2K_M3b2JcM3^2JfM4\\2HiM4Z2IiM5Y2GkM7X2DmM9U2CoM;S2ARN<Q2@RN=Q2_OTN>a6N3Mb`^1"}, "label": "man standing behind person hitting ball"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt behind the fence\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt behind the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 27, 28, 29, 30, 31, 45, 46, 47, 48, 49, 63, 64, 65, 66, 81, 82, 83, 84, 99, 100, 101, 102, 118, 119, 120, 121, 138, 139], "bbox": [0.4973, 0.0, 0.74098, 0.5957336956521739], "iscrowd": 0, "area": 14995.9795, "rle": {"size": [368, 500], "counts": "\\ai27U;8I7I6J7G6I6J7I6N3N1O2O0O1O2N2N3N2M3M3M3]H^MS6e2QIZNi6\\3J6J6J6I7J7I6I7J5KN3M3M3L4M3M3M3L4M3M3M3L4L4L4L4K5L4L4L5L32N1O1O2N1O1O1O2O0O1O2N1O1O1O2N100O2N3M3M3M3N3L3N2M3M2O1N3N1N2O2N1O1O2N1O1O7I8I3LO0QMiJEX5ROnJmN2m1Q5UORKdN5T2j4VOWK]N6Z2d4WO[KVN9`2]4YO_Lc0a3\\OcLa0^3^OeL`0[3^OjL>W3@mL=T3BoL;R3CSM9n2EVM8l2FWM7j2G[M5f2J]M3d2K_M3b2JcM3^2JfM4\\2HiM4Z2IiM5Y2GkM7X2DmM9U2CoM;S2ARN<Q2@RN=Q2_OTN>a6N3Mb`^1"}, "label": "a man in a black shirt behind the fence"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a red and white baseball uniform\"."}], "task": "refering", "answer_template": "The \"a boy in a red and white baseball uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 42, 43, 44, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 115, 116, 117, 133, 134, 135, 136, 151, 152, 153, 154, 155, 171, 172, 173, 174, 190, 191, 192], "bbox": [0.34048, 0.11027173913043478, 0.693, 0.8069021739130434], "iscrowd": 0, "area": 18227.4491, "rle": {"size": [368, 500], "counts": "ZUm12Z;6I8H7J6I7TF4o7OkG5T8OdG7Z8`1O10001O000O2O0000001JjLoGW3Q8jLnGV3R8kLmGV3R84000001O00001O0000001O000100O1O1OQJdLk3\\3nKkLS4S3hKSMX4l2bK[M`4b2[KdMf4Z2XKiMj4U2SKnMn4P2PKSNR5k1lJWNU5g1iJ\\NY5b1dJaN]5]1bJeN`5Y1]JjNd5T1ZJoNg5o0WJTOj5k0RJYOn5f0PJ]OQ6b0lIAT6>hIGW6:cIK^6d2O01O010O010O10O01O010O010O10O01O010O010O010O0010O010O0010O0001O010O0010O10OQL^IP3c6PM^In2b6RM_In2a6RM`Il2a6SM`Im2_6TMbIk2^6TMcIk2]6VMdIi2\\6WMeIg2[6YMhIe2X6[MiId2W6[MkI;5<P6ZOjKc0V4\\OkKd0T4]OkKd0U4[OkKf0U4ZOkKf0T4ZOlKg0T4YOlKg0S4YOmKh0S4XOmKg0S4YOmKi0R4WOnKj0P4VOPLl0o3TOQLm0m3SOSLn0m3RORLP1l3POULQ1j3oNULS1j3lNWLU1h3kNWLW1h3hNXLZ1g3eNZL[1f3dNZL\\1g3cNYL]1h3aNZL^1g3aNYL`1f3`NZL`1g3_NZL`1f3`NZL`1g3^N[Lb1d3^N\\Lb1e3]N[Lc1e3]N\\Lb1e3]N[Lc1e3]N[Ld1e3ZN]Le1c3[N]Le1i3UNXLi1h601N2O001N2O1O2M2O2M3N1O2M3N2M3N2N2M3N3L3N2N2M3N2N3L3N2M3N2N1NVRg1"}, "label": "a boy in a red and white baseball uniform"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a red shirt\"."}], "task": "refering", "answer_template": "The \"a boy wearing a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 42, 43, 44, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 96, 97, 98, 99, 100, 101, 115, 116, 117, 133, 134, 135, 136, 151, 152, 153, 154, 155, 171, 172, 173, 174, 190, 191, 192], "bbox": [0.34048, 0.11027173913043478, 0.693, 0.8069021739130434], "iscrowd": 0, "area": 18227.4491, "rle": {"size": [368, 500], "counts": "ZUm12Z;6I8H7J6I7TF4o7OkG5T8OdG7Z8`1O10001O000O2O0000001JjLoGW3Q8jLnGV3R8kLmGV3R84000001O00001O0000001O000100O1O1OQJdLk3\\3nKkLS4S3hKSMX4l2bK[M`4b2[KdMf4Z2XKiMj4U2SKnMn4P2PKSNR5k1lJWNU5g1iJ\\NY5b1dJaN]5]1bJeN`5Y1]JjNd5T1ZJoNg5o0WJTOj5k0RJYOn5f0PJ]OQ6b0lIAT6>hIGW6:cIK^6d2O01O010O010O10O01O010O010O10O01O010O010O010O0010O010O0010O0001O010O0010O10OQL^IP3c6PM^In2b6RM_In2a6RM`Il2a6SM`Im2_6TMbIk2^6TMcIk2]6VMdIi2\\6WMeIg2[6YMhIe2X6[MiId2W6[MkI;5<P6ZOjKc0V4\\OkKd0T4]OkKd0U4[OkKf0U4ZOkKf0T4ZOlKg0T4YOlKg0S4YOmKh0S4XOmKg0S4YOmKi0R4WOnKj0P4VOPLl0o3TOQLm0m3SOSLn0m3RORLP1l3POULQ1j3oNULS1j3lNWLU1h3kNWLW1h3hNXLZ1g3eNZL[1f3dNZL\\1g3cNYL]1h3aNZL^1g3aNYL`1f3`NZL`1g3_NZL`1f3`NZL`1g3^N[Lb1d3^N\\Lb1e3]N[Lc1e3]N\\Lb1e3]N[Lc1e3]N[Ld1e3ZN]Le1c3[N]Le1i3UNXLi1h601N2O001N2O1O2M2O2M3N1O2M3N2M3N2N2M3N3L3N2N2M3N2N3L3N2M3N2N1NVRg1"}, "label": "a boy wearing a red shirt"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stands\"."}], "task": "refering", "answer_template": "The \"the stands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 33, 34, 35, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89], "bbox": [0.69204, 0.13505434782608697, 1.0, 0.3272826086956522], "iscrowd": 0, "area": 8595.674899999998, "rle": {"size": [368, 500], "counts": "Pjl31_;1N100O2O0O101N100O2O0O101N1O100O1O100O1O1O1cNAiG?W8GbG:^8I^G8b8I]G7c8J[G7e84oFNP9?cFA]9Y1000000001O00000000000000001O01O000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O100000000000000O100000000O1000000000O0100000000O100000000000000000000000000000000000000000000000000000O1O100O2N4L8H8HmM"}, "label": "the stands"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"metal bleachers sitting behind a man\"."}], "task": "refering", "answer_template": "The \"metal bleachers sitting behind a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 33, 34, 35, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89], "bbox": [0.69204, 0.13505434782608697, 1.0, 0.3272826086956522], "iscrowd": 0, "area": 8595.674899999998, "rle": {"size": [368, 500], "counts": "Pjl31_;1N100O2O0O101N100O2O0O101N1O100O1O100O1O1O1cNAiG?W8GbG:^8I^G8b8I]G7c8J[G7e84oFNP9?cFA]9Y1000000001O00000000000000001O01O000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000O100000000000000O100000000O1000000000O0100000000O100000000000000000000000000000000000000000000000000000O1O100O2N4L8H8HmM"}, "label": "metal bleachers sitting behind a man"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white picnic bench with a cooler on top\"."}], "task": "refering", "answer_template": "The \"a white picnic bench with a cooler on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 139, 140, 141], "bbox": [0.6870599999999999, 0.3598369565217391, 0.8959599999999999, 0.6086684782608696], "iscrowd": 0, "area": 6699.583750000001, "rle": {"size": [368, 500], "counts": "[hk34Z;5L4M2N3M3M3M2N3M3M2N3M2N2M3N2N2N4L3M4L4L2N001O0000001O0000000000000O10000O10001O0000000000000000000000000O2O01O0000001O0000001O0000100O2NO100000000000000O101O:F6J00001O000000000000001O000000O1O1O2K5J6J7Jh0WOSab0"}, "label": "a white picnic bench with a cooler on top"}]}
{"id": 144906, "image": "COCO_train2014_000000144906.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and red cooler sitting on the bleachers at a baseball game\"."}], "task": "refering", "answer_template": "The \"a white and red cooler sitting on the bleachers at a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 139, 140, 141], "bbox": [0.6870599999999999, 0.3598369565217391, 0.8959599999999999, 0.6086684782608696], "iscrowd": 0, "area": 6699.583750000001, "rle": {"size": [368, 500], "counts": "[hk34Z;5L4M2N3M3M3M2N3M3M2N3M2N2M3N2N2N4L3M4L4L2N001O0000001O0000000000000O10000O10001O0000000000000000000000000O2O01O0000001O0000001O0000100O2NO100000000000000O101O:F6J00001O000000000000001O000000O1O1O2K5J6J7Jh0WOSab0"}, "label": "a white and red cooler sitting on the bleachers at a baseball game"}]}
{"id": 259595, "image": "COCO_train2014_000000259595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby getting his teeth brushed\"."}], "task": "refering", "answer_template": "The \"baby getting his teeth brushed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 260, 261, 262, 263, 264, 276, 277, 278, 299, 300, 301, 303, 304, 305, 322, 323, 325, 326, 327, 328, 329, 330, 345, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0016875000000000002, 0.01954166666666667, 0.618546875, 0.9633541666666666], "iscrowd": 0, "area": 77891.66234999997, "rle": {"size": [480, 640], "counts": "ml01[;c3K4M4O1000O010000O010O1000O010000O01000O10O10O1000O010000O01000O10O10O1000O010000O010O1000O010000O01000O10O10O1000O01000lLVES2i:nM[Em1f:RN_Ej1a:VNbEg1]:ZNgEb1Y:]NlE^1U:bNnE[1Q:fNSFV1m9iNXFS1i9lN\\Fn0e9RO`Fi0a9UOeFf0\\9[OfFa0Z9@jF:X9GkF4V9LnFOR93PGIQ97QGEP9=PGAQ9?PG_OQ9b0oF\\OQ9e0PGXOR9i0nFUOR9l0oFQOl4LPMT1TNnNg46SMl0T5[OjJg0R5^OlJb0R5BkJ?R5FlJ;Q5ImJ7Q5MlJ5P50nJ0P54nJMW2oNcLY1S1I`1H[Md0R1EU14fM=Q1@W15fMa0o0[OX16fMf0n0TO[17eMk0k0PO]16gMP1h0jN_18hMR1e0gNb18hMV1b0cNc19kMX1>_Ne1;lMZ1;]Ng1:mM]19YNh1;oM`16UNi1=QNa12TNk1<RNd10PNl1>SNf1MmMn1?UNg1JkMo1?VNi1IhMn1b0XNh1HhMl1c0\\Nf1GgMj1f0^Ne1FfMl1e0]Ng1FeMl1d0]Ni1FcMm1d0]Nj1DdMn1b0]Nk1EcMn1b0\\Nm1DbMP2a0\\Nn1CbMP2`0\\NP2C`MQ2`0[NR2B_MS2?[NS2A_MS2>[NU2A]MT2>ZNW2@]MU2<[NX2_O\\MV2<ZNZ2^O[MX2;YN\\2^OZMX2:ZN]2]OYMY2:YN_2\\OYMZ28ZN`2ZOYM\\27YNb2YOXM^26XNc2YOYM^24YNd2XOXM_24XNf2WOWMa23WNh2VOWMb21XNi2TOWMd20WNk2SOWMe2NXNl2QOWMf2NXNm2QOUMg2NWNo2POUMh2LXNj5h1VJWNl5h1TJWNm5h1TJXNl5h1TJWNn5h1RJXNn5h1RJXNn5h1RJXNo5g1QJYNo5f1RJZNo5e1QJ[No5e1QJ[No5e1QJZNQ6e1oI[NQ6d1PJ\\NQ6c1oI]NQ6c1oI]NQ6c1oI]NR6b1nI^NR6a1oI^NS6a1mI_NS6a1mI`NR6`1nI`NS6_1mIaNS6`1lI`NT6a1kI_NV6`1jI`NV6a1iI_NX6a1gI`NX6a1gI_NZ6`1fI`NZ6a1eI_N\\6a1cI_N]6b1bI^N_6b1`I^N`6b1`I^Na6b1]I`Nb6a1]I_Nd6a1[I_Ne6b1ZI^Ng6a1YI_Ng6b1XI^Ni6b1VI^Nj6c1UI^Nk6b1TI^Nl6b1TI^Nm6b1RI^Nn6c1QI]NP7c1oH]NQ7c1oH]NR7c1mH]NS7d1lH]NT7c1kH]NU7d1jH\\NW7c1iH]NW7d1hH\\NY7d1fH\\NZ7e1eH[N\\7e1cH\\N\\7d1dH\\N]7d1bH\\N^7e1`H\\Na7d1^H\\Nb7e1]H[Nd7d1\\H\\Nd7e1[H[Ne7f1ZH[Nf7e1YH[Ng7e1YH[Ng7f1XHZNi7f1VHZNj7g1UHYNk7h1THXNl7h1THXNm7h1RHYNm7h1RHXNn7i1QHWNP8i1oGWNQ8i1oGYNo7h1PHYNP8g1oGZNP8g1oGZNP8g1oG[NP8d1PH]No7d1PH]No7d1QH\\Nn7e1QH\\No7d1PH^Nn7b1RH_Nm7b1RH_Nm7b1RH_Nn7a1QH`Nn7`1RHaNm7`1RHaNm7`1RHbNm7^1RHcNm7^1RHcNm7]1SHdNl7]1SHdNl7]1SHdNm7\\1RHeNm7\\1RHfNl7Z1THgNk7Z1THgNl7Y1SHhNl7Y1SHhNl7X1THiNk7W1UHjNk7U1UHlNj7U1UHmNi7S1WHoNg7Q1YHQOe7P1ZHSOd7l0\\HVOb7j0^HXO`7i0_HZO^7f0bH\\Om1XMk2[3XK_Og1[MP3W3XK@f1ZMQ3V3YKCc1XMS3U3ZKE`1XMV3S3YKG_1WMW3R3ZKJ\\1UMY3Q3[KLY1TM\\3Q3YKNY1RM^3o2XK3V1oLb3o2VK5V1lLd3o2TK:U1gLg3R3PK<T1cLl3U3kJ=U1^LH_OX3P8lLaHHC\\3P8gL_HID`3Q8cL[HIHd3Q8^LXHJKi3P8YLVHINn3P8ULTHG0T4P8PLnHP4V7lKjHU4P90000001O0000000O2O0000001O0000001O000000001O00000010O2N1O0O2N1O1gN_ERNc:m1iEfMX:[2nE]MS:d2nEYMT:g2mEVMT:k2mEQMU:o2n011O00000O1O100O2N1O3N1N3M2O0O001O010O001N2O0O2O0O2O1N101N101N101O1N101N101N2O0O2O0O2N101N2N1O2N1O2O1N1O2N1O2N1O2N2N1O2N1O1O1O1O1O1O1O1O1O2N1N3N3M3L5IQTb3"}, "label": "baby getting his teeth brushed"}]}
{"id": 259595, "image": "COCO_train2014_000000259595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby is getting his teeth brushed by a woman while sitting on a sink\"."}], "task": "refering", "answer_template": "The \"a baby is getting his teeth brushed by a woman while sitting on a sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 260, 261, 262, 263, 264, 276, 277, 278, 299, 300, 301, 303, 304, 305, 322, 323, 325, 326, 327, 328, 329, 330, 345, 348, 349, 350, 351, 352, 353, 354, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0016875000000000002, 0.01954166666666667, 0.618546875, 0.9633541666666666], "iscrowd": 0, "area": 77891.66234999997, "rle": {"size": [480, 640], "counts": "ml01[;c3K4M4O1000O010000O010O1000O010000O01000O10O10O1000O010000O01000O10O10O1000O010000O010O1000O010000O01000O10O10O1000O01000lLVES2i:nM[Em1f:RN_Ej1a:VNbEg1]:ZNgEb1Y:]NlE^1U:bNnE[1Q:fNSFV1m9iNXFS1i9lN\\Fn0e9RO`Fi0a9UOeFf0\\9[OfFa0Z9@jF:X9GkF4V9LnFOR93PGIQ97QGEP9=PGAQ9?PG_OQ9b0oF\\OQ9e0PGXOR9i0nFUOR9l0oFQOl4LPMT1TNnNg46SMl0T5[OjJg0R5^OlJb0R5BkJ?R5FlJ;Q5ImJ7Q5MlJ5P50nJ0P54nJMW2oNcLY1S1I`1H[Md0R1EU14fM=Q1@W15fMa0o0[OX16fMf0n0TO[17eMk0k0PO]16gMP1h0jN_18hMR1e0gNb18hMV1b0cNc19kMX1>_Ne1;lMZ1;]Ng1:mM]19YNh1;oM`16UNi1=QNa12TNk1<RNd10PNl1>SNf1MmMn1?UNg1JkMo1?VNi1IhMn1b0XNh1HhMl1c0\\Nf1GgMj1f0^Ne1FfMl1e0]Ng1FeMl1d0]Ni1FcMm1d0]Nj1DdMn1b0]Nk1EcMn1b0\\Nm1DbMP2a0\\Nn1CbMP2`0\\NP2C`MQ2`0[NR2B_MS2?[NS2A_MS2>[NU2A]MT2>ZNW2@]MU2<[NX2_O\\MV2<ZNZ2^O[MX2;YN\\2^OZMX2:ZN]2]OYMY2:YN_2\\OYMZ28ZN`2ZOYM\\27YNb2YOXM^26XNc2YOYM^24YNd2XOXM_24XNf2WOWMa23WNh2VOWMb21XNi2TOWMd20WNk2SOWMe2NXNl2QOWMf2NXNm2QOUMg2NWNo2POUMh2LXNj5h1VJWNl5h1TJWNm5h1TJXNl5h1TJWNn5h1RJXNn5h1RJXNn5h1RJXNo5g1QJYNo5f1RJZNo5e1QJ[No5e1QJ[No5e1QJZNQ6e1oI[NQ6d1PJ\\NQ6c1oI]NQ6c1oI]NQ6c1oI]NR6b1nI^NR6a1oI^NS6a1mI_NS6a1mI`NR6`1nI`NS6_1mIaNS6`1lI`NT6a1kI_NV6`1jI`NV6a1iI_NX6a1gI`NX6a1gI_NZ6`1fI`NZ6a1eI_N\\6a1cI_N]6b1bI^N_6b1`I^N`6b1`I^Na6b1]I`Nb6a1]I_Nd6a1[I_Ne6b1ZI^Ng6a1YI_Ng6b1XI^Ni6b1VI^Nj6c1UI^Nk6b1TI^Nl6b1TI^Nm6b1RI^Nn6c1QI]NP7c1oH]NQ7c1oH]NR7c1mH]NS7d1lH]NT7c1kH]NU7d1jH\\NW7c1iH]NW7d1hH\\NY7d1fH\\NZ7e1eH[N\\7e1cH\\N\\7d1dH\\N]7d1bH\\N^7e1`H\\Na7d1^H\\Nb7e1]H[Nd7d1\\H\\Nd7e1[H[Ne7f1ZH[Nf7e1YH[Ng7e1YH[Ng7f1XHZNi7f1VHZNj7g1UHYNk7h1THXNl7h1THXNm7h1RHYNm7h1RHXNn7i1QHWNP8i1oGWNQ8i1oGYNo7h1PHYNP8g1oGZNP8g1oGZNP8g1oG[NP8d1PH]No7d1PH]No7d1QH\\Nn7e1QH\\No7d1PH^Nn7b1RH_Nm7b1RH_Nm7b1RH_Nn7a1QH`Nn7`1RHaNm7`1RHaNm7`1RHbNm7^1RHcNm7^1RHcNm7]1SHdNl7]1SHdNl7]1SHdNm7\\1RHeNm7\\1RHfNl7Z1THgNk7Z1THgNl7Y1SHhNl7Y1SHhNl7X1THiNk7W1UHjNk7U1UHlNj7U1UHmNi7S1WHoNg7Q1YHQOe7P1ZHSOd7l0\\HVOb7j0^HXO`7i0_HZO^7f0bH\\Om1XMk2[3XK_Og1[MP3W3XK@f1ZMQ3V3YKCc1XMS3U3ZKE`1XMV3S3YKG_1WMW3R3ZKJ\\1UMY3Q3[KLY1TM\\3Q3YKNY1RM^3o2XK3V1oLb3o2VK5V1lLd3o2TK:U1gLg3R3PK<T1cLl3U3kJ=U1^LH_OX3P8lLaHHC\\3P8gL_HID`3Q8cL[HIHd3Q8^LXHJKi3P8YLVHINn3P8ULTHG0T4P8PLnHP4V7lKjHU4P90000001O0000000O2O0000001O0000001O000000001O00000010O2N1O0O2N1O1gN_ERNc:m1iEfMX:[2nE]MS:d2nEYMT:g2mEVMT:k2mEQMU:o2n011O00000O1O100O2N1O3N1N3M2O0O001O010O001N2O0O2O0O2O1N101N101N101O1N101N101N2O0O2O0O2N101N2N1O2N1O2O1N1O2N1O2N1O2N2N1O2N1O1O1O1O1O1O1O1O1O2N1N3N3M3L5IQTb3"}, "label": "a baby is getting his teeth brushed by a woman while sitting on a sink"}]}
{"id": 259595, "image": "COCO_train2014_000000259595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in a green shirt holding a thermometer in the babies mouth\"."}], "task": "refering", "answer_template": "The \"the person in a green shirt holding a thermometer in the babies mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 80, 88, 89, 90, 91, 102, 103, 104, 105, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 233, 234, 235, 236, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 263, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.09767187499999999, 0.134875, 1.0, 1.0], "iscrowd": 0, "area": 128582.75775000005, "rle": {"size": [480, 640], "counts": "`im07b>8L4O1N2N2N3M2N2O1N2N3N1O1O2N2N1O2QDfNT:\\1gEhNX:Z1bEkN]:V1_EnN_:U1[EPOe:Q1VESOj:b21O0O1O1O2O0O1O2O0O1O2O3L1O01O010O001O1O010O001O000001O00000001O01O000000001O0O10001O000O101O00001O0O101O0O2N101N1O2O0O2N101N1O2N101N1O2O0O2O001O001O0010O01O001O0010O01O001O0010O`0Ac0\\O;F0O100O010O100O100O010O100O010O01O010O01O010O00010O010O00010O0010O00010O0001O000010O0001O0010O0001O00010O00001O01N100O100O1O2O0O101N100O2O001N101N10001N101N10010O010O01O001N101O001YKnL]MS3b2nL^MS3`2oL_MQ3`2PM`MQ3_2PM_MQ3`2PM`MQ3^2QMaMo2^2SM`Mo2^2RMbMP3[2RMdMo2Z2RMfMP3W2RMgMQ3V2PMjMQ3U2PMjMR3S2oLmMS3P2nLQNR3m1oLSNS3j1mLWNU3f1lL[NT3c1mL]NU3`1lL`NV3]1kLdNU3[1kLeNV3Y1kLhNT3W1lLjNU3T1lLlNU3R1lLoNT3P1lLPOU3n0lLROU3m0kLTOU30mKYLn0g3V3MPLYLk0k3T3KULWLh0m3S3JZLVLc0P4S3H^LUL`0T4Q3F`LVL`0S4Q3DaLXL?T4P3BcLZL=U4o2@dL\\L>T4m2^OgL]L=V4k2\\OiL^L<W4j2[OjL^L=W4h2ZOlL^L=Y4f2YOlL_L>Y4e2WOnL`L>Y4c2WOoL`L?Y4a2VOQMaL?Y4`2UOQMcL>Y4`2SORMeL?X4^2SOSMeL`0X4]2QOTMgL`0X4\\2POTMhLa0X4[2nNUMjLa0X4Z2mNTMlLb0X4Y2lNZMdLa0`4U2lNUOS1l0lNUOS1l0lNVOR1k0mNVOR1k0nNUOP1l0POVOn0k0QOVOn0k0QOVOn0j0SOWOk0i0UOXOj0h0VOYOh0h0XOYOg0g0ZOZOd0f0\\O[Oc0e0^O[Oa0e0_O\\O`0d0@]O>d0C\\O<d0D]O;c0E_O9a0H_O7a0I@6`0K@4`0LA3?MB2>OB0>0B0>0CO=2CM=3CM=4CK=5CK=5DJ<7DH<8DH<8EH:9EH:8GG9:FG99GH88IH69IG7:HG7:IF6:JG5:KE5<JE5<JE5<KD4<LB6?J_O7b0H\\O:e0EYO=h0CUO?l0@ROb0n0_OoNc0R1\\OlNf0U1YOjNh0W1XOfNj0[1UOcNm0^1TO^Nn0b1SO[No0f1ROVNP1k1QOQNQ1P2POmMQ1T2POhMR1X2oNeMS1^2kN`MV1b2iN[MY1g2fNWM[1k2dNRM^1P3aNnL`1U3^NhLd1Z3[NdLf1^3YN`Lh1b3VN\\Ll1g3QNXLP2j3oMSLS2o3kMPLV2R4hMlKZ2V4dMiK]2Z4`MeKa2]4^M`Kd2b4ZM]Kg2d4XMZKj2g4UMXKl2i4SMUKo2l4QMRKP3o4oLPKR3Q5mLnJT3S5kLlJV3T5jLlJV3U5jLiJW3X5hLgJY3Z5fLeJ[3\\5dLcJ]3^5bLaJ_3`5`L_Ja3b5]L^Jd3c5[L]Je3d5ZL[Jg3e5XL[Ji3e5WLZJj3f5VLYJk3g5ULXJl3h5SLXJn3h5RLWJo3i5QLWJo3i5QLVJP4i5PLWJQ4i5oKVJR4j5nKUJS4k5mKUJS4k5lKUJU4k5kKTJV4l5jKSJW4m5iKRJX4n5hKRJX4n5hKQJY4o5gKPJZ4P6fKoI[4Q6eKnI\\4R6dKnI\\4R6dKmI]4S6cKlI^4T6W200000000000000000000O10000000000000000000000N2O1N2N2VOPGcKQ9[4k0N2O1N2N2N2O1N2N2O1N2N2N2O1N2N200O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O1000000O100O100O100O100O100O100O100O1O100O100O100O100O1O1O1O1O100O1O1O1O1O100"}, "label": "the person in a green shirt holding a thermometer in the babies mouth"}]}
{"id": 259595, "image": "COCO_train2014_000000259595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"probably mom\"."}], "task": "refering", "answer_template": "The \"probably mom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 80, 88, 89, 90, 91, 102, 103, 104, 105, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 229, 232, 233, 234, 235, 236, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 263, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.09767187499999999, 0.134875, 1.0, 1.0], "iscrowd": 0, "area": 128582.75775000005, "rle": {"size": [480, 640], "counts": "`im07b>8L4O1N2N2N3M2N2O1N2N3N1O1O2N2N1O2QDfNT:\\1gEhNX:Z1bEkN]:V1_EnN_:U1[EPOe:Q1VESOj:b21O0O1O1O2O0O1O2O0O1O2O3L1O01O010O001O1O010O001O000001O00000001O01O000000001O0O10001O000O101O00001O0O101O0O2N101N1O2O0O2N101N1O2N101N1O2O0O2O001O001O0010O01O001O0010O01O001O0010O`0Ac0\\O;F0O100O010O100O100O010O100O010O01O010O01O010O00010O010O00010O0010O00010O0001O000010O0001O0010O0001O00010O00001O01N100O100O1O2O0O101N100O2O001N101N10001N101N10010O010O01O001N101O001YKnL]MS3b2nL^MS3`2oL_MQ3`2PM`MQ3_2PM_MQ3`2PM`MQ3^2QMaMo2^2SM`Mo2^2RMbMP3[2RMdMo2Z2RMfMP3W2RMgMQ3V2PMjMQ3U2PMjMR3S2oLmMS3P2nLQNR3m1oLSNS3j1mLWNU3f1lL[NT3c1mL]NU3`1lL`NV3]1kLdNU3[1kLeNV3Y1kLhNT3W1lLjNU3T1lLlNU3R1lLoNT3P1lLPOU3n0lLROU3m0kLTOU30mKYLn0g3V3MPLYLk0k3T3KULWLh0m3S3JZLVLc0P4S3H^LUL`0T4Q3F`LVL`0S4Q3DaLXL?T4P3BcLZL=U4o2@dL\\L>T4m2^OgL]L=V4k2\\OiL^L<W4j2[OjL^L=W4h2ZOlL^L=Y4f2YOlL_L>Y4e2WOnL`L>Y4c2WOoL`L?Y4a2VOQMaL?Y4`2UOQMcL>Y4`2SORMeL?X4^2SOSMeL`0X4]2QOTMgL`0X4\\2POTMhLa0X4[2nNUMjLa0X4Z2mNTMlLb0X4Y2lNZMdLa0`4U2lNUOS1l0lNUOS1l0lNVOR1k0mNVOR1k0nNUOP1l0POVOn0k0QOVOn0k0QOVOn0j0SOWOk0i0UOXOj0h0VOYOh0h0XOYOg0g0ZOZOd0f0\\O[Oc0e0^O[Oa0e0_O\\O`0d0@]O>d0C\\O<d0D]O;c0E_O9a0H_O7a0I@6`0K@4`0LA3?MB2>OB0>0B0>0CO=2CM=3CM=4CK=5CK=5DJ<7DH<8DH<8EH:9EH:8GG9:FG99GH88IH69IG7:HG7:IF6:JG5:KE5<JE5<JE5<KD4<LB6?J_O7b0H\\O:e0EYO=h0CUO?l0@ROb0n0_OoNc0R1\\OlNf0U1YOjNh0W1XOfNj0[1UOcNm0^1TO^Nn0b1SO[No0f1ROVNP1k1QOQNQ1P2POmMQ1T2POhMR1X2oNeMS1^2kN`MV1b2iN[MY1g2fNWM[1k2dNRM^1P3aNnL`1U3^NhLd1Z3[NdLf1^3YN`Lh1b3VN\\Ll1g3QNXLP2j3oMSLS2o3kMPLV2R4hMlKZ2V4dMiK]2Z4`MeKa2]4^M`Kd2b4ZM]Kg2d4XMZKj2g4UMXKl2i4SMUKo2l4QMRKP3o4oLPKR3Q5mLnJT3S5kLlJV3T5jLlJV3U5jLiJW3X5hLgJY3Z5fLeJ[3\\5dLcJ]3^5bLaJ_3`5`L_Ja3b5]L^Jd3c5[L]Je3d5ZL[Jg3e5XL[Ji3e5WLZJj3f5VLYJk3g5ULXJl3h5SLXJn3h5RLWJo3i5QLWJo3i5QLVJP4i5PLWJQ4i5oKVJR4j5nKUJS4k5mKUJS4k5lKUJU4k5kKTJV4l5jKSJW4m5iKRJX4n5hKRJX4n5hKQJY4o5gKPJZ4P6fKoI[4Q6eKnI\\4R6dKnI\\4R6dKmI]4S6cKlI^4T6W200000000000000000000O10000000000000000000000N2O1N2N2VOPGcKQ9[4k0N2O1N2N2N2O1N2N2O1N2N2N2O1N2N200O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O1000000O100O100O100O100O100O100O100O1O100O100O100O100O1O1O1O1O100O1O1O1O1O100"}, "label": "probably mom"}]}
{"id": 480779, "image": "COCO_train2014_000000480779.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom half of a sandwich with a piece of fish and vegetables on top\"."}], "task": "refering", "answer_template": "The \"the bottom half of a sandwich with a piece of fish and vegetables on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 199], "bbox": [0.45317187499999995, 0.10639583333333333, 0.8324843749999999, 0.4857083333333333], "iscrowd": 0, "area": 27510.829549999995, "rle": {"size": [480, 640], "counts": "^RX41n>1N3N1O1N3N1N2O2N1N2O2N1O2N2N2N2J7I6K5N2N2O1N2N3N1N2M3L4L4L4L5K4M3N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N1O2O1N2N2N101O1O1N2O001O1N2O00000O100000000O1000000O100000000O2O000001O01O0001O0001O00010O1O1O100O100O10O0100O100O1O10O010O0010O010O01O01O010O01O01O001O00001O010O0000O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O2O2M3N2M3M3N3L3N1N2N2O1N2O1N2N2O0O2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N3N1N3M2O2M3M2N3M3M2N3M2O2M3M2N3M3M2N3M3M2O2M2N3M3M2N3M3M2N3N1N3MPPb1"}, "label": "the bottom half of a sandwich with a piece of fish and vegetables on top"}]}
{"id": 480779, "image": "COCO_train2014_000000480779.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of bruschetta\"."}], "task": "refering", "answer_template": "The \"a piece of bruschetta\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 199], "bbox": [0.45317187499999995, 0.10639583333333333, 0.8324843749999999, 0.4857083333333333], "iscrowd": 0, "area": 27510.829549999995, "rle": {"size": [480, 640], "counts": "^RX41n>1N3N1O1N3N1N2O2N1N2O2N1O2N2N2N2J7I6K5N2N2O1N2N3N1N2M3L4L4L4L5K4M3N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N1O2O1N2N2N101O1O1N2O001O1N2O00000O100000000O1000000O100000000O2O000001O01O0001O0001O00010O1O1O100O100O10O0100O100O1O10O010O0010O010O01O01O010O01O01O001O00001O010O0000O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O2O2M3N2M3M3N3L3N1N2N2O1N2O1N2N2O0O2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N3N1N3M2O2M3M2N3M3M2N3M2O2M3M2N3M3M2N3M3M2O2M2N3M3M2N3M3M2N3N1N3MPPb1"}, "label": "a piece of bruschetta"}]}
{"id": 480779, "image": "COCO_train2014_000000480779.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of toasted bread with nothing on it\"."}], "task": "refering", "answer_template": "The \"a piece of toasted bread with nothing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 167, 168], "bbox": [0.150390625, 0.15914583333333335, 0.461734375, 0.4190625], "iscrowd": 0, "area": 19041.758349999996, "rle": {"size": [480, 640], "counts": "nT]15[>b0L4M2N3L3N2N1N3N1O1N3N1O2M2O2N1N3N1O1N3N1O2M2O2N1N2O2N1N3N1O2N1O2N101N1O2N1O2N2O0O2N1O2N101N1O2O0O2O1N100O2O000O101N100O101O0O10001O0O10001O0O10001O0O10001O0O1000000O1000000000000000000000000000000000000000O10000000000000000O1000000000000O100O100O100O10000O100O2O0O100O100O100O100O100O101N10000O100O100O100O1O1O2N1O1O100O1O1O1O1O1O1O1O1O2N1O1O1O1O1O101N1O2N2N2N1O2N2N1N3N2M2N3M3M3M2N3N2M4L4L5K4E<BRUQ5"}, "label": "a piece of toasted bread with nothing on it"}]}
{"id": 480779, "image": "COCO_train2014_000000480779.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of toast on a plate that is part of a larger meal\"."}], "task": "refering", "answer_template": "The \"a piece of toast on a plate that is part of a larger meal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 167, 168], "bbox": [0.150390625, 0.15914583333333335, 0.461734375, 0.4190625], "iscrowd": 0, "area": 19041.758349999996, "rle": {"size": [480, 640], "counts": "nT]15[>b0L4M2N3L3N2N1N3N1O1N3N1O2M2O2N1N3N1O1N3N1O2M2O2N1N2O2N1N3N1O2N1O2N101N1O2N1O2N2O0O2N1O2N101N1O2O0O2O1N100O2O000O101N100O101O0O10001O0O10001O0O10001O0O10001O0O1000000O1000000000000000000000000000000000000000O10000000000000000O1000000000000O100O100O100O10000O100O2O0O100O100O100O100O100O101N10000O100O100O100O1O1O2N1O1O100O1O1O1O1O1O1O1O1O2N1O1O1O1O1O101N1O2N2N2N1O2N2N1N3N2M2N3M3M3M2N3N2M4L4L5K4E<BRUQ5"}, "label": "a piece of toast on a plate that is part of a larger meal"}]}
{"id": 87569, "image": "COCO_train2014_000000087569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 323, 324, 325, 326, 327, 328, 331, 332, 333, 341, 342, 343, 344, 345, 346, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379], "bbox": [0.0, 0.21940624999999997, 0.6183125, 0.98590625], "iscrowd": 0, "area": 66405.87754999999, "rle": {"size": [640, 480], "counts": "P:9cc0d0]Ob0]Od0]O9F:G8G:G8H8J6J5K5L5J5K5K6J5K5TBZKU<l4[CkKY<Y4WC_L\\<h5J6J5K6lFiGT6]8TIVHk6P8]HbHd7b7eGQIZ8[9O001O10O01O1O10O01O100O001O100O001O1O1N1N3N2N2VOi0H9G9J6M2N3L4M3L3N3L4O10O01O1O1O1O001O1O1O001O1O1O1O1O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O10000000000000000O1000kEPKZ5P5fJUKU5k4kJYKQ5g4oJZKP5f4PK[Ko4e4QK[Ko4e4QK\\Kn4d4RK\\Kn4c4RK_Kn4a4QK_Ko4a4QK`Kn4a4QK_Ko4a4QK`Kn4a4PKaKo4_4QKaKo4`4oJbKP5^4PKbKQ5^4mJdKR5\\4mJeKkK]Og8o4]KeKcKFn8e4^KZL`4g3_K[L_4e3`K^L^4c3`K`L^4`3aKcL]4^3aKeL^4Z3aKiL]4X3aKkL]4V3aKmL]4S3bKPM\\4Q3bKRM\\4o2bKTMnNQNS2l4mNVMhNVNY2e4mNXMbNZN`2_4lNZM\\N_Nd2Y4oN\\MUNaNj2U4oN]MoMeNP3P4oN^MjMiNS3l3PO_MfMkNX3h3PObM_MmN^3b3ROfMWMnNe3^3QOjMQMoNj3Z3SOg0l0[OROf0m0\\OQOe0m0^OROb0cNYId0V6i0b0aN[Ic0U6j0a0cN\\I?V6m0>dN^I<V6n0=fN_I9V6P1;gNaI6U6S1:gNcI3U6T1:hNcI1U6c0k0[OSIOT6c0k0^OSIKU6c0k0BRIHU6c0k0ESIDS6e0l0GSIAS6a0Q1MnH_OS6=U14jH\\OS69Y1:gHZOR68Z1>fHVOR69[1a0eHSOS66]1f0bHQOT64]1k0aHnNU62]1P1`HkNV60]1U1_HhNW6M^1Z1^HfN\\?Z1f1O0000000O101O000000000O2O00000000000O10000000000O100000O100000000000000O100000000000000O11O0O10001O01O1O1O010O1O1O10O0001O01O01O001O01O01O01O010O0100O010O100O100O001O100O1O1O010O00O1N2N3M2O1N2N3O02N1O100O10001O0000N2N2O1N2N2O1L4J?@lka3"}, "label": "a woman wearing a black shirt"}]}
{"id": 87569, "image": "COCO_train2014_000000087569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman wearing a brown sweater\"."}], "task": "refering", "answer_template": "The \"a young woman wearing a brown sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 323, 324, 325, 326, 327, 328, 331, 332, 333, 341, 342, 343, 344, 345, 346, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379], "bbox": [0.0, 0.21940624999999997, 0.6183125, 0.98590625], "iscrowd": 0, "area": 66405.87754999999, "rle": {"size": [640, 480], "counts": "P:9cc0d0]Ob0]Od0]O9F:G8G:G8H8J6J5K5L5J5K5K6J5K5TBZKU<l4[CkKY<Y4WC_L\\<h5J6J5K6lFiGT6]8TIVHk6P8]HbHd7b7eGQIZ8[9O001O10O01O1O10O01O100O001O100O001O1O1N1N3N2N2VOi0H9G9J6M2N3L4M3L3N3L4O10O01O1O1O1O001O1O1O001O1O1O1O1O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O10000000000000000O1000kEPKZ5P5fJUKU5k4kJYKQ5g4oJZKP5f4PK[Ko4e4QK[Ko4e4QK\\Kn4d4RK\\Kn4c4RK_Kn4a4QK_Ko4a4QK`Kn4a4QK_Ko4a4QK`Kn4a4PKaKo4_4QKaKo4`4oJbKP5^4PKbKQ5^4mJdKR5\\4mJeKkK]Og8o4]KeKcKFn8e4^KZL`4g3_K[L_4e3`K^L^4c3`K`L^4`3aKcL]4^3aKeL^4Z3aKiL]4X3aKkL]4V3aKmL]4S3bKPM\\4Q3bKRM\\4o2bKTMnNQNS2l4mNVMhNVNY2e4mNXMbNZN`2_4lNZM\\N_Nd2Y4oN\\MUNaNj2U4oN]MoMeNP3P4oN^MjMiNS3l3PO_MfMkNX3h3PObM_MmN^3b3ROfMWMnNe3^3QOjMQMoNj3Z3SOg0l0[OROf0m0\\OQOe0m0^OROb0cNYId0V6i0b0aN[Ic0U6j0a0cN\\I?V6m0>dN^I<V6n0=fN_I9V6P1;gNaI6U6S1:gNcI3U6T1:hNcI1U6c0k0[OSIOT6c0k0^OSIKU6c0k0BRIHU6c0k0ESIDS6e0l0GSIAS6a0Q1MnH_OS6=U14jH\\OS69Y1:gHZOR68Z1>fHVOR69[1a0eHSOS66]1f0bHQOT64]1k0aHnNU62]1P1`HkNV60]1U1_HhNW6M^1Z1^HfN\\?Z1f1O0000000O101O000000000O2O00000000000O10000000000O100000O100000000000000O100000000000000O11O0O10001O01O1O1O010O1O1O10O0001O01O01O001O01O01O01O010O0100O010O100O100O001O100O1O1O010O00O1N2N3M2O1N2N3O02N1O100O10001O0000N2N2O1N2N2O1L4J?@lka3"}, "label": "a young woman wearing a brown sweater"}]}
{"id": 87569, "image": "COCO_train2014_000000087569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue sweater grabbing a piece of cake\"."}], "task": "refering", "answer_template": "The \"woman in blue sweater grabbing a piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 70, 71, 72, 73, 74, 87, 88, 89, 90, 91, 107, 108, 141, 142, 158, 159, 174, 175, 176, 177, 191, 192, 193, 194, 208, 209, 210, 211, 212, 226, 227, 228, 229, 230, 243, 244, 245, 246, 247, 263, 264, 265, 281, 282], "bbox": [0.1313125, 0.149546875, 0.6225625000000001, 0.715140625], "iscrowd": 0, "area": 21311.84065, "rle": {"size": [640, 480], "counts": "VaW12jc03M3M4L3L4M4O010O0010O010O01O10O01O010O01O2N1O1O2N101N1O2N1O1O2N101N1O1OO100O2N2O1N2N101N2N2O1N2O1O1N101O1N2O1N2O1O1YBPO\\8Q1cGSO[8n0cGTO\\8m0aGVO]8l0aGVO^8j0`GZO^8g0`G[O_8f0^G]O`8e0^G^O`8b0_G@`8a0^GC^8?`GD^8<`GI]88aGL[86cGM[84cG0Z80eG4W8NgG5W8LgG8V8HiG<T8EjG>S8DjGa0S8@kGd0R8\\OmGg0Q8ZOmGj0P8WOnGm0o7SOQHo0m7ROQHR1l7oNSHS1k7nNTHT1j7lNVHU1PMQMR:l1kHV1nLSMT:i1mHV1iLVMY:e1lHX1fLXM\\:a1mHn1R7SNlHP2R7QNmHP2R7QNlHQ2S7PNlHQ2S7PNkHS2S7oMjHS2U7nMQHeM_N`4_9lMoGhM_N]4b9lMmGiM_N\\4d9lMkGjM_N[4f9lMhGlM`NY4h9lMfGmM`NX4k9kMcGPN^NV4Q:kM]GRN`NT4S:kM^GPN]NV4V:jM]GPN[NW4X:jM^Gh2c8XM]Gn0hNhMl9Z1]Gf0VOgM]9d1]Gd0^O_MW9m1\\Gb0FXMo8V2[Ga0OPMg8`2ZGa0b:_O^Ed0_:\\ObEf0\\:YOdEi0Z:WOgEk0V:UOjEm0T:SOmEo0Q:POoER1o9mNSFU1j9gNZF[1d9aNaFa1]9ZNgFh1W9UNmFi1T9RNQGn1o8lMXGS2h8gM^GY2c8`MdG_2\\8[MjGd2W8VMPHi2P8QMVHo2k7jL\\HU3X<01O001O001O0O2O1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O0000001O000O10001O000000001O000000001O001O1O1O1O1O1O1O1N2O1O1O1N2N2O1N100O1O100O1O101N100O1O100O1O1O1O2N2N1N3L4L4L3L5K5O1N101007H01O1O0O2N2N1N3N1N3N2L3K_f`3"}, "label": "woman in blue sweater grabbing a piece of cake"}]}
{"id": 87569, "image": "COCO_train2014_000000087569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with blue colored sweater\"."}], "task": "refering", "answer_template": "The \"girl with blue colored sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 70, 71, 72, 73, 74, 87, 88, 89, 90, 91, 107, 108, 141, 142, 158, 159, 174, 175, 176, 177, 191, 192, 193, 194, 208, 209, 210, 211, 212, 226, 227, 228, 229, 230, 243, 244, 245, 246, 247, 263, 264, 265, 281, 282], "bbox": [0.1313125, 0.149546875, 0.6225625000000001, 0.715140625], "iscrowd": 0, "area": 21311.84065, "rle": {"size": [640, 480], "counts": "VaW12jc03M3M4L3L4M4O010O0010O010O01O10O01O010O01O2N1O1O2N101N1O2N1O1O2N101N1O1OO100O2N2O1N2N101N2N2O1N2O1O1N101O1N2O1N2O1O1YBPO\\8Q1cGSO[8n0cGTO\\8m0aGVO]8l0aGVO^8j0`GZO^8g0`G[O_8f0^G]O`8e0^G^O`8b0_G@`8a0^GC^8?`GD^8<`GI]88aGL[86cGM[84cG0Z80eG4W8NgG5W8LgG8V8HiG<T8EjG>S8DjGa0S8@kGd0R8\\OmGg0Q8ZOmGj0P8WOnGm0o7SOQHo0m7ROQHR1l7oNSHS1k7nNTHT1j7lNVHU1PMQMR:l1kHV1nLSMT:i1mHV1iLVMY:e1lHX1fLXM\\:a1mHn1R7SNlHP2R7QNmHP2R7QNlHQ2S7PNlHQ2S7PNkHS2S7oMjHS2U7nMQHeM_N`4_9lMoGhM_N]4b9lMmGiM_N\\4d9lMkGjM_N[4f9lMhGlM`NY4h9lMfGmM`NX4k9kMcGPN^NV4Q:kM]GRN`NT4S:kM^GPN]NV4V:jM]GPN[NW4X:jM^Gh2c8XM]Gn0hNhMl9Z1]Gf0VOgM]9d1]Gd0^O_MW9m1\\Gb0FXMo8V2[Ga0OPMg8`2ZGa0b:_O^Ed0_:\\ObEf0\\:YOdEi0Z:WOgEk0V:UOjEm0T:SOmEo0Q:POoER1o9mNSFU1j9gNZF[1d9aNaFa1]9ZNgFh1W9UNmFi1T9RNQGn1o8lMXGS2h8gM^GY2c8`MdG_2\\8[MjGd2W8VMPHi2P8QMVHo2k7jL\\HU3X<01O001O001O0O2O1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O0000001O000O10001O000000001O000000001O001O1O1O1O1O1O1O1N2O1O1O1N2N2O1N100O1O100O1O101N100O1O100O1O1O1O2N2N1N3L4L4L3L5K5O1N101007H01O1O0O2N2N1N3N1N3N2L3K_f`3"}, "label": "girl with blue colored sweater"}]}
{"id": 87569, "image": "COCO_train2014_000000087569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a pink sweater cutting a cake\"."}], "task": "refering", "answer_template": "The \"a woman in a pink sweater cutting a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 212, 213, 214, 215, 229, 230, 231, 232, 244, 245, 261, 262, 278, 279, 295, 296], "bbox": [0.3611666666666667, 0.18228125, 0.7343958333333334, 0.7579062499999999], "iscrowd": 0, "area": 30264.648499999996, "rle": {"size": [640, 480], "counts": "Pc\\31jc05K5K5K5K5K5K6J5K5J6K5K5K5K5K5_KcMZGb2c8cMVGc2g8aMRGd2l8_MPGd2m8aMRG^2l8fMSGY2j8lMUGS2i8PNWGo1g8UNXGj1e8[NZGd1d8`N[G_1b8eN^G[1_8iNaGU1]8oNbGP1[8UOdG=aLRNi;d1fG8cLSNd;j1hG1fLTN`;o1iGLhLTN\\;U2kGEkLUNX;Z2lG_OnLVNT;^2nG[OoLVNP;d2PHTORMXNk:h2SHmNTMZNf:n2UHfNWM[Nb:R3WHbNXM[N_:W3XHYN^M_NW:]3ZHnMb8V2]GcMf8a2ZGYMi8k2VGoLm8U3RGeLP9`3oFYLU9j3lFoKV9V4W34L4K4M4K5M3M2N3N2M3M3N1N3N2M3M2O2M3N2M2N3N2M3N1N3N2M3N1N3N2M3N2N1N3N2N2N1N3N2NO1O0000000001O00000000N2L4M3L4L4L4L4M4K4L4O15K5K5L4K5K5K5K6J5L4K5K3M2N2O5K00O100O010O1O100O10000000000000000[Oe0H8H9F9H8H8H8G9H8H9G9G9F9H9G9G9H7O2M3N2M3N2M3M3N2N3L3N2M4M5J6K5J6K4L5J6KlV_2"}, "label": "a woman in a pink sweater cutting a cake"}]}
{"id": 87569, "image": "COCO_train2014_000000087569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a long sleeve sweater cutting into a desert\"."}], "task": "refering", "answer_template": "The \"a woman wearing a long sleeve sweater cutting into a desert\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 212, 213, 214, 215, 229, 230, 231, 232, 244, 245, 261, 262, 278, 279, 295, 296], "bbox": [0.3611666666666667, 0.18228125, 0.7343958333333334, 0.7579062499999999], "iscrowd": 0, "area": 30264.648499999996, "rle": {"size": [640, 480], "counts": "Pc\\31jc05K5K5K5K5K5K6J5K5J6K5K5K5K5K5_KcMZGb2c8cMVGc2g8aMRGd2l8_MPGd2m8aMRG^2l8fMSGY2j8lMUGS2i8PNWGo1g8UNXGj1e8[NZGd1d8`N[G_1b8eN^G[1_8iNaGU1]8oNbGP1[8UOdG=aLRNi;d1fG8cLSNd;j1hG1fLTN`;o1iGLhLTN\\;U2kGEkLUNX;Z2lG_OnLVNT;^2nG[OoLVNP;d2PHTORMXNk:h2SHmNTMZNf:n2UHfNWM[Nb:R3WHbNXM[N_:W3XHYN^M_NW:]3ZHnMb8V2]GcMf8a2ZGYMi8k2VGoLm8U3RGeLP9`3oFYLU9j3lFoKV9V4W34L4K4M4K5M3M2N3N2M3M3N1N3N2M3M2O2M3N2M2N3N2M3N1N3N2M3N1N3N2M3N2N1N3N2N2N1N3N2NO1O0000000001O00000000N2L4M3L4L4L4L4M4K4L4O15K5K5L4K5K5K5K6J5L4K5K3M2N2O5K00O100O010O1O100O10000000000000000[Oe0H8H9F9H8H8H8G9H8H9G9G9F9H9G9G9H7O2M3N2M3N2M3M3N2N3L3N2M4M5J6K5J6K4L5J6KlV_2"}, "label": "a woman wearing a long sleeve sweater cutting into a desert"}]}
{"id": 136721, "image": "COCO_train2014_000000136721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman tennis player with the blue outfit and getting ready to hit the ball\"."}], "task": "refering", "answer_template": "The \"the woman tennis player with the blue outfit and getting ready to hit the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 128, 140, 141, 143, 156, 157, 158, 172, 173, 174, 187, 188, 189, 202, 203, 204, 217, 218, 219, 232, 233, 234, 247, 248, 262, 263, 277, 278], "bbox": [0.3596244131455399, 0.34696875, 0.6385915492957746, 0.796640625], "iscrowd": 0, "area": 10294.364650000005, "rle": {"size": [640, 426], "counts": "Yko21mc05K4L5K5K4M4M2N1O1O001O1O1O001O1O10O1000000O10000O10000O1000000O10011N2O1O1N;F000O10000O11[AG_99]FKc96YFLh92VF2j9MSF6n9HPF<P:CmE`0T:_OiEe0W:YOgEj0Z:UOcEn0]:TO^EP1mLmNa<4_FR1nLkNd<4ZFT1PMjNe<4WFU1RMhNg<4UFV1RMhNi<3QFW1TMhNk<2nEY1UMfNm<3jEZ1VMcNEB[=b0eE]1XMbNH^O[=d0bE_1XMbNW=1]E`1YMaN[=0XE]2h:cMZDYOROW3i<\\MmCJQOl2U=YMbC7POc2_=VMYCS4g<oKQCW4P=jKhB\\4`;QKfEc0cNb4d;TKcE<aNe4k;XK^E`5_:jJZEV5c:SKXEl4e:^KTEa4l:fKQEW4o:PLoDj3R;\\LlD_3T;[1^DZJb;d5bDYJ_;f5dDWJ];h5fDUJ[;@dD]55QKW;YOSE_5IUK\\<e4gCXKa<_4cC^Kf<Y4]CeKf<W4]CfKf<W4]CfKf<V4^CgKc<X4`CeKa<Z4bCdK]<[4gCbKZ<]4f1N1N3N2N2M3N1O4`LgAR1]>jNfAR1`>iNdAS1`>gNeAV1f`0K4M4L4K4M4L3L5L3M4LTko2"}, "label": "the woman tennis player with the blue outfit and getting ready to hit the ball"}]}
{"id": 136721, "image": "COCO_train2014_000000136721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis playing in blue dress serves the ball\"."}], "task": "refering", "answer_template": "The \"tennis playing in blue dress serves the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 128, 140, 141, 143, 156, 157, 158, 172, 173, 174, 187, 188, 189, 202, 203, 204, 217, 218, 219, 232, 233, 234, 247, 248, 262, 263, 277, 278], "bbox": [0.3596244131455399, 0.34696875, 0.6385915492957746, 0.796640625], "iscrowd": 0, "area": 10294.364650000005, "rle": {"size": [640, 426], "counts": "Yko21mc05K4L5K5K4M4M2N1O1O001O1O1O001O1O10O1000000O10000O10000O1000000O10011N2O1O1N;F000O10000O11[AG_99]FKc96YFLh92VF2j9MSF6n9HPF<P:CmE`0T:_OiEe0W:YOgEj0Z:UOcEn0]:TO^EP1mLmNa<4_FR1nLkNd<4ZFT1PMjNe<4WFU1RMhNg<4UFV1RMhNi<3QFW1TMhNk<2nEY1UMfNm<3jEZ1VMcNEB[=b0eE]1XMbNH^O[=d0bE_1XMbNW=1]E`1YMaN[=0XE]2h:cMZDYOROW3i<\\MmCJQOl2U=YMbC7POc2_=VMYCS4g<oKQCW4P=jKhB\\4`;QKfEc0cNb4d;TKcE<aNe4k;XK^E`5_:jJZEV5c:SKXEl4e:^KTEa4l:fKQEW4o:PLoDj3R;\\LlD_3T;[1^DZJb;d5bDYJ_;f5dDWJ];h5fDUJ[;@dD]55QKW;YOSE_5IUK\\<e4gCXKa<_4cC^Kf<Y4]CeKf<W4]CfKf<W4]CfKf<V4^CgKc<X4`CeKa<Z4bCdK]<[4gCbKZ<]4f1N1N3N2N2M3N1O4`LgAR1]>jNfAR1`>iNdAS1`>gNeAV1f`0K4M4L4K4M4L3L5L3M4LTko2"}, "label": "tennis playing in blue dress serves the ball"}]}
{"id": 46612, "image": "COCO_train2014_000000046612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"big bear\"."}], "task": "refering", "answer_template": "The \"big bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 223], "bbox": [0.58678125, 0.4014117647058823, 0.867171875, 0.7661176470588236], "iscrowd": 0, "area": 15240.457750000001, "rle": {"size": [340, 640], "counts": "aRm3?d9b0^Ob0]Ob0K5M3M4L3M3M3M3O2N1O1O1O1O1O100O1O1O1O1O@iHeMV7V2RIhMm6T2YIkMf6S2^IkMb6U2`IjM_6W2aIiM_6V2cIiM\\6W2eIiMZ6W2hIgMX6Y2iIfMX6Y2Q101O01O01O000100O101N2N3N1N3N3M4hHSMa6a3O1O1O1O00001O00001N100001O00000000000000000@XITMi6g2\\IWMe6g2]IYMc6d2aI[M_6c2dI\\M\\6b2fI]M[6a2hI^MX6_2lI`MT6]2oIbMR6\\2U1O01O1O001O1O2N2oH[Mo5f2nI\\MR6f2kI[MU6g2gI[MY6h2bIZM^6h2\\I\\Me6f2TI]Mm6U3000O10000O100000000000O10001O00000000000000000O100000kN^IXNb6a1lIXNT6f1TJSNm5l1[JlMf5T2^1O00100O1O10000O100O10001O01O0000001O001O00001O001O00O1O1N2N2M3N2M3N3K4K5L5K4J6J6K3L5O1O1N200O100O2N2M`Ql0"}, "label": "big bear"}]}
{"id": 46612, "image": "COCO_train2014_000000046612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bear looking at a rock\"."}], "task": "refering", "answer_template": "The \"a brown bear looking at a rock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 223], "bbox": [0.58678125, 0.4014117647058823, 0.867171875, 0.7661176470588236], "iscrowd": 0, "area": 15240.457750000001, "rle": {"size": [340, 640], "counts": "aRm3?d9b0^Ob0]Ob0K5M3M4L3M3M3M3O2N1O1O1O1O1O100O1O1O1O1O@iHeMV7V2RIhMm6T2YIkMf6S2^IkMb6U2`IjM_6W2aIiM_6V2cIiM\\6W2eIiMZ6W2hIgMX6Y2iIfMX6Y2Q101O01O01O000100O101N2N3N1N3N3M4hHSMa6a3O1O1O1O00001O00001N100001O00000000000000000@XITMi6g2\\IWMe6g2]IYMc6d2aI[M_6c2dI\\M\\6b2fI]M[6a2hI^MX6_2lI`MT6]2oIbMR6\\2U1O01O1O001O1O2N2oH[Mo5f2nI\\MR6f2kI[MU6g2gI[MY6h2bIZM^6h2\\I\\Me6f2TI]Mm6U3000O10000O100000000000O10001O00000000000000000O100000kN^IXNb6a1lIXNT6f1TJSNm5l1[JlMf5T2^1O00100O1O10000O100O10001O01O0000001O001O00001O001O00O1O1N2N2M3N2M3N3K4K5L5K4J6J6K3L5O1O1N200O100O2N2M`Ql0"}, "label": "a brown bear looking at a rock"}]}
{"id": 46612, "image": "COCO_train2014_000000046612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bear behind the tree on the left\"."}], "task": "refering", "answer_template": "The \"the bear behind the tree on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 139, 140, 141, 143, 162, 163, 164, 166, 185, 186, 189, 190], "bbox": [0.038140625, 0.48047058823529415, 0.31021875, 0.7413823529411765], "iscrowd": 0, "area": 3982.452949999999, "rle": {"size": [340, 640], "counts": "]V81b:1N2N2N2O1N2N2N2O1N2N2N3M2N2M3N2N2N2O1O1N2O1O1O2N1O1O1O1O1O1O1O10M2M4001N100O10000O100O2O00000000000O101OO100O1O100O00100D<I8I6I7I7Ild?:oZ@4L3N2M3M3M3M2O0RO_OiGb0l70iG5U8U11O1O1O2SOgG\\O\\8`0hG]OZ8a0iG\\OY8EbG`09Hf83]GKg8JbG3b9N1;D2O1O1N0100O1O001O1O001O1O1O001O0010O01O001O001O001O00001O001O001O001O001Nc`b4"}, "label": "the bear behind the tree on the left"}]}
{"id": 46612, "image": "COCO_train2014_000000046612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little brown bear is laying under the birch trees for shade\"."}], "task": "refering", "answer_template": "The \"the little brown bear is laying under the birch trees for shade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 139, 140, 141, 143, 162, 163, 164, 166, 185, 186, 189, 190], "bbox": [0.038140625, 0.48047058823529415, 0.31021875, 0.7413823529411765], "iscrowd": 0, "area": 3982.452949999999, "rle": {"size": [340, 640], "counts": "]V81b:1N2N2N2O1N2N2N2O1N2N2N3M2N2M3N2N2N2O1O1N2O1O1O2N1O1O1O1O1O1O1O10M2M4001N100O10000O100O2O00000000000O101OO100O1O100O00100D<I8I6I7I7Ild?:oZ@4L3N2M3M3M3M2O0RO_OiGb0l70iG5U8U11O1O1O2SOgG\\O\\8`0hG]OZ8a0iG\\OY8EbG`09Hf83]GKg8JbG3b9N1;D2O1O1N0100O1O001O1O001O1O1O001O0010O01O001O001O001O00001O001O001O001O001Nc`b4"}, "label": "the little brown bear is laying under the birch trees for shade"}]}
{"id": 398872, "image": "COCO_train2014_000000398872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person walking behind subject on bench\"."}], "task": "refering", "answer_template": "The \"person walking behind subject on bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 16, 17, 18, 30, 31, 32, 33, 45, 46, 47, 48, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 105, 106, 107, 120, 121, 122, 136, 137, 151, 152, 166, 167, 180, 181, 182], "bbox": [0.004637002341920374, 0.0, 0.25, 0.5507500000000001], "iscrowd": 0, "area": 21809.234599999992, "rle": {"size": [640, 427], "counts": "g\\13ic05K5K5K5\\B_Ob7l0XHTOd7Q1ZHPOb7T1^HmN]7X1bHhNZ7\\1fHeNT7a1jH`NR7d1YESNn19e8h1TEZNU2N^8GTDV2n0aN]2D]8GUDX2h0hNd2YO[8X2jDgNi2QOY8\\2jDhNk2mNW8`2jDhNm2hNU8d2lDhNl2eNT8h2mDgNm2bNR8j2oDhNl2_NQ8S5mGmJo7X5oGiJn7\\5oGeJo7_5oGaJn7d5PH]Jm7g5QHYJm7j5RHVJk7P6RHPJc6DQHa6X1lIg6GjGc6\\1fIj6JeGh6IVI2IO?a80XGi7NnG7;b8W9]GiFc8X9\\GhFd8Y9ZGhFf8Y9YGgFg8Z9WGgFi8[9TGfFl8W:000000000O100000000000000O100000000000000O10000000000lKZG[KOSOg8^5dGPK0^O^8^5mGgJLGZ8^5TH_JJOT8^5]HWJE8P8]5eHoIC?k7]5`J_Jc5]5bJ]Jc5`5bJZJb5d5^4N2O100O1O1O1O1O1O1N2K5VKXAW4m>iKTAQ4Q?nKRAl3R?TLPAe3U?[Lm@`3V?`LPAV3T?jLSAk2Q?UMWA_2m>`MZAT2l>kMXAl1n>SNWA1gNR1Y`0lNTAITOj0n?\\OU3N2O1O1O1OQPX6"}, "label": "person walking behind subject on bench"}]}
{"id": 398872, "image": "COCO_train2014_000000398872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person walking behind bus shelter\"."}], "task": "refering", "answer_template": "The \"person walking behind bus shelter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 16, 17, 18, 30, 31, 32, 33, 45, 46, 47, 48, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 105, 106, 107, 120, 121, 122, 136, 137, 151, 152, 166, 167, 180, 181, 182], "bbox": [0.004637002341920374, 0.0, 0.25, 0.5507500000000001], "iscrowd": 0, "area": 21809.234599999992, "rle": {"size": [640, 427], "counts": "g\\13ic05K5K5K5\\B_Ob7l0XHTOd7Q1ZHPOb7T1^HmN]7X1bHhNZ7\\1fHeNT7a1jH`NR7d1YESNn19e8h1TEZNU2N^8GTDV2n0aN]2D]8GUDX2h0hNd2YO[8X2jDgNi2QOY8\\2jDhNk2mNW8`2jDhNm2hNU8d2lDhNl2eNT8h2mDgNm2bNR8j2oDhNl2_NQ8S5mGmJo7X5oGiJn7\\5oGeJo7_5oGaJn7d5PH]Jm7g5QHYJm7j5RHVJk7P6RHPJc6DQHa6X1lIg6GjGc6\\1fIj6JeGh6IVI2IO?a80XGi7NnG7;b8W9]GiFc8X9\\GhFd8Y9ZGhFf8Y9YGgFg8Z9WGgFi8[9TGfFl8W:000000000O100000000000000O100000000000000O10000000000lKZG[KOSOg8^5dGPK0^O^8^5mGgJLGZ8^5TH_JJOT8^5]HWJE8P8]5eHoIC?k7]5`J_Jc5]5bJ]Jc5`5bJZJb5d5^4N2O100O1O1O1O1O1O1N2K5VKXAW4m>iKTAQ4Q?nKRAl3R?TLPAe3U?[Lm@`3V?`LPAV3T?jLSAk2Q?UMWA_2m>`MZAT2l>kMXAl1n>SNWA1gNR1Y`0lNTAITOj0n?\\OU3N2O1O1O1OQPX6"}, "label": "person walking behind bus shelter"}]}
{"id": 398872, "image": "COCO_train2014_000000398872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a hat which is covered in buttons\"."}], "task": "refering", "answer_template": "The \"a man wearing a hat which is covered in buttons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 49, 50, 51, 52, 53, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 217, 218, 219, 220, 221, 228, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 248, 249, 250, 251, 258, 259, 260, 261, 263, 264, 265, 266, 273, 274, 275, 278, 279, 280, 281, 288, 289, 290, 293, 294, 295, 303, 304, 305, 308, 309, 310, 318, 319, 320, 322, 323, 324, 325, 333, 334, 335, 337, 338, 339, 340], "bbox": [0.20791569086651054, 0.11353125, 0.7529508196721312, 0.99809375], "iscrowd": 0, "area": 83016.24610000002, "rle": {"size": [640, 427], "counts": "hWh14X:0RO6h01RO4g04RO2g06RO0h08QOMh0;QOKh0=QOIi0>QOGh0a0SOCf0e0YO[Oa0l0]OUOcNl2[1UMbNP3S1YMiNl2j0`MSOd2`0hM]O\\26oMHU2JXN4k1_OaN`0a1SOkNm0V1eNWOZ1l0XN@h1n0^M^Oa2Q900O100O100O100O100O100O100O100O10O0100O100O100O100lMZAmNg>S1RBTNn=k1lB[MU=e2S2O100O100O100O100OJ7[Od0\\Od0\\Oe0A>N3L3N3M2N3M2M4M2N3M2M3N3M2N3M2M4M2N3M2N3L3N3M3M3M4K4M4L3M3M4K4N3M2N2N3L3N3M2N2M4M2N3M2N3L3N2N3M2N3L3O12O0O2O0O2N101N101N1O2O0O101N101N1O10O01]LhE]MX:Q2YFoMg9_1kFaNV9l0\\GUOc89PHFQ8OYH2g7J\\H6d7G_H9b7CaH>_7_OcHb0\\7[OgHg0X7UOkHn0R7oNQIS1n6iNVIX1i6eNYI^1d6_N_Ic1`6YNcIi1aMYLZ7j1XKP2ZMYL\\7d1]KU2TMYL_7_1_KZ2nLZLb7Y1cK`2gLYLf7S1fKf2aLYLi7m0jKl2YLYLl7h0nKQ3SLZLn7b0QLV3nKZLP8=UL\\3gKYLT87XLb3aKYLW81[Lh3[KYLY8L_Ln3TKYL\\8FbLS4nJZL`8_OeLZ5Z3cJjL]5V3_JmLb5R3[JQMg5n2VJTMk5`81O1O1O2N100O1O1O4L;E<D;E;E;ZFQHf6Y8XH`Hf7k7YGlHf8X9O1O2N1O1O1O101N1O1O1O2N1jMlD^KU;_4PE^KQ;_4TE_Km:]4XE`Ki:]4\\E`Ke:]4`E_Kb:^4cE^K_:_4fE]K]:_4hE^KY:_4lE]KV:`4oE\\KS:a4RF[KQ:a4TF[Kn9b4XFYKj9d4[FXKg9e4^FWKd9f4aFVKb9f4cFWK^9f4P3L4M4L3M3M3M4L3M3M3M3A`0^Ol0TOf1[NbVQ2"}, "label": "a man wearing a hat which is covered in buttons"}]}
{"id": 398872, "image": "COCO_train2014_000000398872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a blue bench talking on a cell phone\"."}], "task": "refering", "answer_template": "The \"a man sitting on a blue bench talking on a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 49, 50, 51, 52, 53, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 217, 218, 219, 220, 221, 228, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 248, 249, 250, 251, 258, 259, 260, 261, 263, 264, 265, 266, 273, 274, 275, 278, 279, 280, 281, 288, 289, 290, 293, 294, 295, 303, 304, 305, 308, 309, 310, 318, 319, 320, 322, 323, 324, 325, 333, 334, 335, 337, 338, 339, 340], "bbox": [0.20791569086651054, 0.11353125, 0.7529508196721312, 0.99809375], "iscrowd": 0, "area": 83016.24610000002, "rle": {"size": [640, 427], "counts": "hWh14X:0RO6h01RO4g04RO2g06RO0h08QOMh0;QOKh0=QOIi0>QOGh0a0SOCf0e0YO[Oa0l0]OUOcNl2[1UMbNP3S1YMiNl2j0`MSOd2`0hM]O\\26oMHU2JXN4k1_OaN`0a1SOkNm0V1eNWOZ1l0XN@h1n0^M^Oa2Q900O100O100O100O100O100O100O100O10O0100O100O100O100lMZAmNg>S1RBTNn=k1lB[MU=e2S2O100O100O100O100OJ7[Od0\\Od0\\Oe0A>N3L3N3M2N3M2M4M2N3M2M3N3M2N3M2M4M2N3M2N3L3N3M3M3M4K4M4L3M3M4K4N3M2N2N3L3N3M2N2M4M2N3M2N3L3N2N3M2N3L3O12O0O2O0O2N101N101N1O2O0O101N101N1O10O01]LhE]MX:Q2YFoMg9_1kFaNV9l0\\GUOc89PHFQ8OYH2g7J\\H6d7G_H9b7CaH>_7_OcHb0\\7[OgHg0X7UOkHn0R7oNQIS1n6iNVIX1i6eNYI^1d6_N_Ic1`6YNcIi1aMYLZ7j1XKP2ZMYL\\7d1]KU2TMYL_7_1_KZ2nLZLb7Y1cK`2gLYLf7S1fKf2aLYLi7m0jKl2YLYLl7h0nKQ3SLZLn7b0QLV3nKZLP8=UL\\3gKYLT87XLb3aKYLW81[Lh3[KYLY8L_Ln3TKYL\\8FbLS4nJZL`8_OeLZ5Z3cJjL]5V3_JmLb5R3[JQMg5n2VJTMk5`81O1O1O2N100O1O1O4L;E<D;E;E;ZFQHf6Y8XH`Hf7k7YGlHf8X9O1O2N1O1O1O101N1O1O1O2N1jMlD^KU;_4PE^KQ;_4TE_Km:]4XE`Ki:]4\\E`Ke:]4`E_Kb:^4cE^K_:_4fE]K]:_4hE^KY:_4lE]KV:`4oE\\KS:a4RF[KQ:a4TF[Kn9b4XFYKj9d4[FXKg9e4^FWKd9f4aFVKb9f4cFWK^9f4P3L4M4L3M3M3M4L3M3M3M3A`0^Ol0TOf1[NbVQ2"}, "label": "a man sitting on a blue bench talking on a cell phone"}]}
{"id": 226840, "image": "COCO_train2014_000000226840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle of the three zebras\"."}], "task": "refering", "answer_template": "The \"the middle of the three zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 239, 240, 241, 242, 262, 264, 265, 285, 286, 287, 288], "bbox": [0.341984375, 0.238625, 0.6225625, 0.7658958333333333], "iscrowd": 0, "area": 23393.5375, "rle": {"size": [480, 640], "counts": "TkV39a>8K6J6I7J6J6K5L4K5K5K4L5K5K5K5K5L4L3M4L4M2M4L4L3N2M3M3M3N3L3M3M3N3L5K4L5L4L4L4L3N3L4L4L4M2M4L3M2O1N3M2N2O0ON3eNYHZKh7\\4dHaK^7T4oHhKR7T4UIhKm6U4ZIeKg6Y4R2N2M2N3N1N3M2O2L3L5K4L5K4O2O001O010O0010O6Jf0[Oe0YO5RG[KY7g4\\HcKb7g4RH]Km7R5bGRK[8k5O2M2O0O101N^NRHTLn7h3cHkK\\7S4RIbKn6c3UHWLT13f6a3YJ]Lf5^3cJ_L\\5\\3lJbLT5X3TKfLk4Z3]31O001O1N2N2M3N2N2N4M3L4L4M5J9I8G8I6I6K6I[1fN1N2O1N2O2dHdHo6_7lHdHT7e7N10^NUIPKj6g4dITK\\6c4RJXKn5P4nJkKS5o3VKnKj4m3^KoKc4k3fKRLZ4h3PLSLQ4h3WLTLj3f3_LWLa3d3P400O100O10000O100O2O1N1M4M2N3_Oa0N9G2M2O0O1O100PO]CCd<6hCCZ<7QDBR<>RD[OP<e0U10100O1O100O2N3N5FTZa3"}, "label": "the middle of the three zebras"}]}
{"id": 226840, "image": "COCO_train2014_000000226840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in the middle of two other zebras\"."}], "task": "refering", "answer_template": "The \"zebra in the middle of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 239, 240, 241, 242, 262, 264, 265, 285, 286, 287, 288], "bbox": [0.341984375, 0.238625, 0.6225625, 0.7658958333333333], "iscrowd": 0, "area": 23393.5375, "rle": {"size": [480, 640], "counts": "TkV39a>8K6J6I7J6J6K5L4K5K5K4L5K5K5K5K5L4L3M4L4M2M4L4L3N2M3M3M3N3L3M3M3N3L5K4L5L4L4L4L3N3L4L4L4M2M4L3M2O1N3M2N2O0ON3eNYHZKh7\\4dHaK^7T4oHhKR7T4UIhKm6U4ZIeKg6Y4R2N2M2N3N1N3M2O2L3L5K4L5K4O2O001O010O0010O6Jf0[Oe0YO5RG[KY7g4\\HcKb7g4RH]Km7R5bGRK[8k5O2M2O0O101N^NRHTLn7h3cHkK\\7S4RIbKn6c3UHWLT13f6a3YJ]Lf5^3cJ_L\\5\\3lJbLT5X3TKfLk4Z3]31O001O1N2N2M3N2N2N4M3L4L4M5J9I8G8I6I6K6I[1fN1N2O1N2O2dHdHo6_7lHdHT7e7N10^NUIPKj6g4dITK\\6c4RJXKn5P4nJkKS5o3VKnKj4m3^KoKc4k3fKRLZ4h3PLSLQ4h3WLTLj3f3_LWLa3d3P400O100O10000O100O2O1N1M4M2N3_Oa0N9G2M2O0O1O100PO]CCd<6hCCZ<7QDBR<>RD[OP<e0U10100O1O100O2N3N5FTZa3"}, "label": "zebra in the middle of two other zebras"}]}
{"id": 226840, "image": "COCO_train2014_000000226840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra to the left , facing away from the others\"."}], "task": "refering", "answer_template": "The \"zebra to the left , facing away from the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282], "bbox": [0.15671875, 0.2921666666666667, 0.29957812500000003, 0.7677083333333333], "iscrowd": 0, "area": 14900.217900000001, "rle": {"size": [480, 640], "counts": "kQ_11j>:F:F:G:I;Ea0_O:F7aF[MV6l2`IXM_6o2XIUMg6T3lHPMS7\\3^HhLa7d3QH_Ln7m3cGWL\\8T4WGoKh8V5N1O1O100O1O2N01H7J6J6L5J5K5dN]11N2N2N2O1N2N3N8H:F`0ZFeKg7j500000000000000O10001O00000001N1O1N3N1N2O2M2O2M2O1N3M2C>B=D<F;[1dN>C2O2N2N2N2OM3N2M4L3M4[NbG]Le8h2YHoLl7[2j2A?H<C=D;E^ia6"}, "label": "zebra to the left , facing away from the others"}]}
{"id": 226840, "image": "COCO_train2014_000000226840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back end of a zebra standing to the left of two other zebras\"."}], "task": "refering", "answer_template": "The \"back end of a zebra standing to the left of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282], "bbox": [0.15671875, 0.2921666666666667, 0.29957812500000003, 0.7677083333333333], "iscrowd": 0, "area": 14900.217900000001, "rle": {"size": [480, 640], "counts": "kQ_11j>:F:F:G:I;Ea0_O:F7aF[MV6l2`IXM_6o2XIUMg6T3lHPMS7\\3^HhLa7d3QH_Ln7m3cGWL\\8T4WGoKh8V5N1O1O100O1O2N01H7J6J6L5J5K5dN]11N2N2N2O1N2N3N8H:F`0ZFeKg7j500000000000000O10001O00000001N1O1N3N1N2O2M2O2M2O1N3M2C>B=D<F;[1dN>C2O2N2N2N2OM3N2M4L3M4[NbG]Le8h2YHoLl7[2j2A?H<C=D;E^ia6"}, "label": "back end of a zebra standing to the left of two other zebras"}]}
{"id": 226840, "image": "COCO_train2014_000000226840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right - hand zebra eating the ground where there does not appear to be any grass\"."}], "task": "refering", "answer_template": "The \"the right - hand zebra eating the ground where there does not appear to be any grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 314, 315, 317, 318], "bbox": [0.608109375, 0.30066666666666664, 0.90034375, 0.8051875000000001], "iscrowd": 0, "area": 24452.95195, "rle": {"size": [480, 640], "counts": "daf54e><C=C<E<I7J6J6J5K5L4K5K5K5K5K5K5iF]Lh6i3QI[Ll6S4dHRLY7\\4XHgKf7g4jG]KT8P5^GSK`8h5N2N2N2N1O2N1O@jGYJT8U5aHiJ^7e4XIYKg6P4RJmKn5n3[JnKf5o3`JnK_5o3j2N3M2MlE[LU9b3lFdLo8Y3SGlLj8Q3WGUMd8i2]G]M_8`2bGeMZ8\\2dGjMW8X2eGnMX8S2eGSNV8o1gGWNU8Y2YGlMc8]4L4L3M4L3N2M4L3N2M3M3NhMgHjLV7S3TIiLj6T3[IjLd6T3bIhL^6V3gIgLZ6U3lIhLT6V3QJfLP6X3UJeLl5W3ZJfLf5X3_JeLa5Y3eJcL\\5Z3iJbLX5[3nJbLR5\\3SKaLn4^3UK_Lk4a3XK\\Lh4d3[KXLg4h3\\KTLd4l3X301O01O01O01O01O01O01O01O102M2O2N1O2N1N3N5K5K5J6K5K5K4K4L4M3LPNjFZNR9e1YGVNb8i1jGPNR8o1YHlMb7R2eHoMU7P2QImMn6R2WIiMi6V2^IcMc6]2bI^M^6b2hIXMX6h2nIRMS6l2[IdLiN;m7Q3TIjLQO3l7S3kHRMZOIl7c3THlL1_Ol7S4^GeLg0UOm7W5UHgJl7Y5UHeJl7[5m0008H7H10001O00001O0000O012N3M3M3M3L3L5K5K5K5\\LiEW2Z:dMlEW2[:`MlE[2Z:^MkE^2\\:ZMjEa2]:VMjEe2Z;O0O1fNWD\\Ok;>^1L4I7E;00O1O1O100O2N100O1O100O1O1O100O00100O00100O00RVn0"}, "label": "the right - hand zebra eating the ground where there does not appear to be any grass"}]}
{"id": 226840, "image": "COCO_train2014_000000226840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra eating with two zebras to its left\"."}], "task": "refering", "answer_template": "The \"a zebra eating with two zebras to its left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 314, 315, 317, 318], "bbox": [0.608109375, 0.30066666666666664, 0.90034375, 0.8051875000000001], "iscrowd": 0, "area": 24452.95195, "rle": {"size": [480, 640], "counts": "daf54e><C=C<E<I7J6J6J5K5L4K5K5K5K5K5K5iF]Lh6i3QI[Ll6S4dHRLY7\\4XHgKf7g4jG]KT8P5^GSK`8h5N2N2N2N1O2N1O@jGYJT8U5aHiJ^7e4XIYKg6P4RJmKn5n3[JnKf5o3`JnK_5o3j2N3M2MlE[LU9b3lFdLo8Y3SGlLj8Q3WGUMd8i2]G]M_8`2bGeMZ8\\2dGjMW8X2eGnMX8S2eGSNV8o1gGWNU8Y2YGlMc8]4L4L3M4L3N2M4L3N2M3M3NhMgHjLV7S3TIiLj6T3[IjLd6T3bIhL^6V3gIgLZ6U3lIhLT6V3QJfLP6X3UJeLl5W3ZJfLf5X3_JeLa5Y3eJcL\\5Z3iJbLX5[3nJbLR5\\3SKaLn4^3UK_Lk4a3XK\\Lh4d3[KXLg4h3\\KTLd4l3X301O01O01O01O01O01O01O01O102M2O2N1O2N1N3N5K5K5J6K5K5K4K4L4M3LPNjFZNR9e1YGVNb8i1jGPNR8o1YHlMb7R2eHoMU7P2QImMn6R2WIiMi6V2^IcMc6]2bI^M^6b2hIXMX6h2nIRMS6l2[IdLiN;m7Q3TIjLQO3l7S3kHRMZOIl7c3THlL1_Ol7S4^GeLg0UOm7W5UHgJl7Y5UHeJl7[5m0008H7H10001O00001O0000O012N3M3M3M3L3L5K5K5K5\\LiEW2Z:dMlEW2[:`MlE[2Z:^MkE^2\\:ZMjEa2]:VMjEe2Z;O0O1fNWD\\Ok;>^1L4I7E;00O1O1O100O2N100O1O100O1O1O100O00100O00100O00RVn0"}, "label": "a zebra eating with two zebras to its left"}]}
{"id": 13856, "image": "COCO_train2014_000000013856.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man ' s hand with ring on it\"."}], "task": "refering", "answer_template": "The \"man ' s hand with ring on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 48, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236], "bbox": [0.005984375, 0.004483568075117371, 0.336546875, 0.6876291079812207], "iscrowd": 0, "area": 37105.9286, "rle": {"size": [426, 640], "counts": "]j1m0V;Y1fNY1gNZ1fNZ1J5O2N1O2N1O2N101N1O2N2N1O2N1O2N1O2N1O2N1O2O1N1O2N1101N101N2O0O2O1O0O2O0O2O1N101N2O0O2O1O0O2O0O2O1N101N2O0O2O1N1000N1O1O00000001O0000000000000000000000000000000000000000001O01O0000000000000000000000000000000000000001O00000000000000000000001O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O101N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1N3M2N2N3L3N3M2M3N3M2N2M4M2N2N3L3N2N3M2M4M2N2N3L3N2N3M2M3N3M2N2M4M2N3M2M3N[c`5"}, "label": "man ' s hand with ring on it"}]}
{"id": 13856, "image": "COCO_train2014_000000013856.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a persion sitting on left chair\"."}], "task": "refering", "answer_template": "The \"a persion sitting on left chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 48, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236], "bbox": [0.005984375, 0.004483568075117371, 0.336546875, 0.6876291079812207], "iscrowd": 0, "area": 37105.9286, "rle": {"size": [426, 640], "counts": "]j1m0V;Y1fNY1gNZ1fNZ1J5O2N1O2N1O2N101N1O2N2N1O2N1O2N1O2N1O2N1O2O1N1O2N1101N101N2O0O2O1O0O2O0O2O1N101N2O0O2O1O0O2O0O2O1N101N2O0O2O1N1000N1O1O00000001O0000000000000000000000000000000000000000001O01O0000000000000000000000000000000000000001O00000000000000000000001O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O101N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1N3M2N2N3L3N3M2M3N3M2N2M4M2N2N3L3N2N3M2M4M2N2N3L3N2N3M2M3N3M2N2M4M2N3M2M3N[c`5"}, "label": "a persion sitting on left chair"}]}
{"id": 13856, "image": "COCO_train2014_000000013856.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair behind the person on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the chair behind the person on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 97, 98, 99, 121, 122], "bbox": [0.070453125, 0.13852112676056338, 0.3448125, 0.39302816901408455], "iscrowd": 0, "area": 8858.54175, "rle": {"size": [426, 640], "counts": "nhb01Y=1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O00000000000000000000000000000O10000000000000000000000000000000000000000000001O001O1O001O001O1O001O001O1O001O1O001O001O1O0010O01O1O001O001O1O001O010O1O001O001O1O001O010O1O001O1O001O1O2N3M2O2M2N3M2N3M2N3M2N3M3M2O2M2N2N0000000000000001O00000000000000000O2O00000O6J5L5J5K6K4K6J5L5J5K6J5L5J5K5L5J5K6K4KXU^5"}, "label": "the chair behind the person on the left in the right hand picture"}]}
{"id": 13856, "image": "COCO_train2014_000000013856.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a whoman eating\"."}], "task": "refering", "answer_template": "The \"a whoman eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 34, 35, 36, 37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 191, 192], "bbox": [0.324578125, 0.003802816901408451, 0.792765625, 0.5925821596244132], "iscrowd": 0, "area": 34072.11335, "rle": {"size": [426, 640], "counts": "ihf23R=5K5L3L5K5L4N101N2O001N2O1N101N2O001N2O1N101O1N100O101O0O100O101O0O100O2O000O100O2O000O100O2O0O10000O2O0O10001N100O10001N100O10001N100O101O0O100O101O0O100O1O2M2N2N2N2N3M2N2N2N3N100O100O2O0O100O1O2O0O100O101N100O101N1O100O101N100O100O[OfF^MX9R36`0A>A3M3M3M3M3M2N3M3M3M3M3O01000O10000000O100000O1000O1000000000000O10000000000O10000000000O10000000000000000000000000000000000000000000000000000000000000010O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O2N2N101N2N2N2N1O2N2N2N1O2N2N1O1O001O01O01N101O000O2O001O0O101O001N10001O1N3N2N2N2M3N2N2N2M3N8H;E:E;F2N2M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N3N1N2N2N2N3M2NmUg1"}, "label": "a whoman eating"}]}
{"id": 13856, "image": "COCO_train2014_000000013856.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady trying to eat a food\"."}], "task": "refering", "answer_template": "The \"a lady trying to eat a food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 34, 35, 36, 37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 191, 192], "bbox": [0.324578125, 0.003802816901408451, 0.792765625, 0.5925821596244132], "iscrowd": 0, "area": 34072.11335, "rle": {"size": [426, 640], "counts": "ihf23R=5K5L3L5K5L4N101N2O001N2O1N101N2O001N2O1N101O1N100O101O0O100O101O0O100O2O000O100O2O000O100O2O0O10000O2O0O10001N100O10001N100O10001N100O101O0O100O101O0O100O1O2M2N2N2N2N3M2N2N2N3N100O100O2O0O100O1O2O0O100O101N100O101N1O100O101N100O100O[OfF^MX9R36`0A>A3M3M3M3M3M2N3M3M3M3M3O01000O10000000O100000O1000O1000000000000O10000000000O10000000000O10000000000000000000000000000000000000000000000000000000000000010O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O2N2N101N2N2N2N1O2N2N2N1O2N2N1O1O001O01O01N101O000O2O001O0O101O001N10001O1N3N2N2N2M3N2N2N2M3N8H;E:E;F2N2M2N2N2N2N3M2N2N2N3M2N2N2N2N3M2N2N2N3N1N2N2N2N3M2NmUg1"}, "label": "a lady trying to eat a food"}]}
{"id": 292386, "image": "COCO_train2014_000000292386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"measuring cup with yellow substance in it\"."}], "task": "refering", "answer_template": "The \"measuring cup with yellow substance in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 74, 75, 89, 90, 91, 92, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 138, 139, 140, 141, 155, 156, 157, 158, 159, 160, 161, 162, 163, 177, 178, 179, 180, 181, 182, 183, 184, 185, 199, 200, 202, 203, 204, 205, 206, 207, 224, 225, 226, 227], "bbox": [0.0698202614379085, 0.13625816993464052, 0.4504575163398693, 0.46509803921568627], "iscrowd": 0, "area": 36590.597850000006, "rle": {"size": [612, 612], "counts": "Vki0;gb0>A?B>A>B?B>B>C8G3N3M2M4M2N2O1N2O1N2O1N2O0O2O1N10O10000000O1000O1YMlA=U>^O]B3e=HmBJS=2_C_Ob<?oCQOR<P1Q3O01O001O001O0O101N1O2N4M3L4N2M3M3N2M3N6I8I7H8I7H7J7I7I7H8I6J7H8I7I7I2M10001O0O1000001N10000O2O00000O101O000O1000000O2O00000O10000O1000001N10000000000000001O0000000000000000000000000000000000000000000000000000000000000001O0O100000000000000000000000000O10000000000O1000000O10000O1000000O1000000O1000000O10001O00000O10000000000O10000O1N3M3M2N3M3M3N2J6H8H7J7H8H8H8H8H7I9G9G8I8G8G:G9G8H9G9G8HcfX6"}, "label": "measuring cup with yellow substance in it"}]}
{"id": 292386, "image": "COCO_train2014_000000292386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"measuring cup full of an orange substance\"."}], "task": "refering", "answer_template": "The \"measuring cup full of an orange substance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 74, 75, 89, 90, 91, 92, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 138, 139, 140, 141, 155, 156, 157, 158, 159, 160, 161, 162, 163, 177, 178, 179, 180, 181, 182, 183, 184, 185, 199, 200, 202, 203, 204, 205, 206, 207, 224, 225, 226, 227], "bbox": [0.0698202614379085, 0.13625816993464052, 0.4504575163398693, 0.46509803921568627], "iscrowd": 0, "area": 36590.597850000006, "rle": {"size": [612, 612], "counts": "Vki0;gb0>A?B>A>B?B>B>C8G3N3M2M4M2N2O1N2O1N2O1N2O0O2O1N10O10000000O1000O1YMlA=U>^O]B3e=HmBJS=2_C_Ob<?oCQOR<P1Q3O01O001O001O0O101N1O2N4M3L4N2M3M3N2M3N6I8I7H8I7H7J7I7I7H8I6J7H8I7I7I2M10001O0O1000001N10000O2O00000O101O000O1000000O2O00000O10000O1000001N10000000000000001O0000000000000000000000000000000000000000000000000000000000000001O0O100000000000000000000000000O10000000000O1000000O10000O1000000O1000000O1000000O10001O00000O10000000000O10000O1N3M3M2N3M3M3N2J6H8H7J7H8H8H8H8H7I9G9G8I8G8G:G9G8H9G9G8HcfX6"}, "label": "measuring cup full of an orange substance"}]}
{"id": 292386, "image": "COCO_train2014_000000292386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a container of maggio rocotta cheese on the countier\"."}], "task": "refering", "answer_template": "The \"a container of maggio rocotta cheese on the countier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 209, 210], "bbox": [0.4051960784313725, 0.057140522875816994, 0.6909150326797386, 0.41558823529411765], "iscrowd": 0, "area": 27121.076749999993, "rle": {"size": [612, 612], "counts": "Qld42Qc01O2N2N101N1O2N1O2O2M5K5K6J5L4K6J5K6J5L4K6J5K5K6J5L5J5K5K6J5L4K6K5K5K5K5K4L5K5K5K5L4K5K5K5K5K5K5K5K1O000000001O000000000001O000000000000000000000001O00000000000001O00000000000000000001O000000000001O000000000000001O0000000O100O101O0O100O10000O101N10000O10000O100O2O000O100O1aMWCcNi<\\1aC[N_<d1lCQNV<n1SDiMm;V2^D_Mc;a2fDVMZ;i2QEmLo:R3[EdLf:\\3a2O101N100O1O100O100O1O2O0O100O1O100O1O2D;E;F:E;E;F:E;E<E^m`3"}, "label": "a container of maggio rocotta cheese on the countier"}]}
{"id": 292386, "image": "COCO_train2014_000000292386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"maggio premium part skim ricotta cheese container with lid\"."}], "task": "refering", "answer_template": "The \"maggio premium part skim ricotta cheese container with lid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 37, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 209, 210], "bbox": [0.4051960784313725, 0.057140522875816994, 0.6909150326797386, 0.41558823529411765], "iscrowd": 0, "area": 27121.076749999993, "rle": {"size": [612, 612], "counts": "Qld42Qc01O2N2N101N1O2N1O2O2M5K5K6J5L4K6J5K6J5L4K6J5K5K6J5L5J5K5K6J5L4K6K5K5K5K5K4L5K5K5K5L4K5K5K5K5K5K5K5K1O000000001O000000000001O000000000000000000000001O00000000000001O00000000000000000001O000000000001O000000000000001O0000000O100O101O0O100O10000O101N10000O10000O100O2O000O100O1aMWCcNi<\\1aC[N_<d1lCQNV<n1SDiMm;V2^D_Mc;a2fDVMZ;i2QEmLo:R3[EdLf:\\3a2O101N100O1O100O100O1O2O0O100O1O100O1O2D;E;F:E;E;F:E;E<E^m`3"}, "label": "maggio premium part skim ricotta cheese container with lid"}]}
{"id": 374308, "image": "COCO_train2014_000000374308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog wearing a harness\"."}], "task": "refering", "answer_template": "The \"dog wearing a harness\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 166, 167, 168, 190, 191, 192, 214, 215, 216, 217, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 283, 285, 286], "bbox": [0.25382812499999996, 0.37133333333333335, 0.497265625, 0.7377708333333334], "iscrowd": 0, "area": 11012.962550000002, "rle": {"size": [480, 640], "counts": "QT\\21o>;E1O010O1O1O1dALj=5VBO7CP=>hB13EU=:hB2NIY=8fB<[=e00O1O100M3O100OO2O1O001O1O0O2O000000OL5L4L4L3M6K4N3N101N01000O109G5L1O00O10`ESMk8m2PGVMR9i2kFYMW9f2fF\\M\\97`Eh1S1QN_93`En1o0PNf9I^E\\2f0mMX:W2cEjM]:V2bEjM_:V2`EiMb:X2\\EhMe:X2ZEhMg:X2WEhMk:^2hDhMY;n22O001N2O1N2N101N2I600O2O1N2O1\\ORDSNQ<l1b001N1O1O1O1O11N1YCRN[<^2L4L3M2NO001O1O1O1O001O1O10O01O1O1O10O10000000O1O100O001O100O1O100O001O1O1N2O1N2N2N101N2N2O1N2N2N2O0O2N2O1N2N2N2O1N2O1N5L4L4L3L5K3L306Igcf4"}, "label": "dog wearing a harness"}]}
{"id": 374308, "image": "COCO_train2014_000000374308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white puppy\"."}], "task": "refering", "answer_template": "The \"the white puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 166, 167, 168, 190, 191, 192, 214, 215, 216, 217, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 283, 285, 286], "bbox": [0.25382812499999996, 0.37133333333333335, 0.497265625, 0.7377708333333334], "iscrowd": 0, "area": 11012.962550000002, "rle": {"size": [480, 640], "counts": "QT\\21o>;E1O010O1O1O1dALj=5VBO7CP=>hB13EU=:hB2NIY=8fB<[=e00O1O100M3O100OO2O1O001O1O0O2O000000OL5L4L4L3M6K4N3N101N01000O109G5L1O00O10`ESMk8m2PGVMR9i2kFYMW9f2fF\\M\\97`Eh1S1QN_93`En1o0PNf9I^E\\2f0mMX:W2cEjM]:V2bEjM_:V2`EiMb:X2\\EhMe:X2ZEhMg:X2WEhMk:^2hDhMY;n22O001N2O1N2N101N2I600O2O1N2O1\\ORDSNQ<l1b001N1O1O1O1O11N1YCRN[<^2L4L3M2NO001O1O1O1O001O1O10O01O1O1O10O10000000O1O100O001O100O1O100O001O1O1N2O1N2N2N101N2N2O1N2N2N2O0O2N2O1N2N2N2O1N2O1N5L4L4L3L5K3L306Igcf4"}, "label": "the white puppy"}]}
{"id": 448046, "image": "COCO_train2014_000000448046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the partial black tabletop on the right\"."}], "task": "refering", "answer_template": "The \"the partial black tabletop on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.68004, 0.81232, 1.0, 1.0], "iscrowd": 0, "area": 9057.547199999999, "rle": {"size": [375, 500], "counts": "Rll31e;1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000"}, "label": "the partial black tabletop on the right"}]}
{"id": 448046, "image": "COCO_train2014_000000448046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table cut off closest to camera\"."}], "task": "refering", "answer_template": "The \"table cut off closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.68004, 0.81232, 1.0, 1.0], "iscrowd": 0, "area": 9057.547199999999, "rle": {"size": [375, 500], "counts": "Rll31e;1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000"}, "label": "table cut off closest to camera"}]}
{"id": 398901, "image": "COCO_train2014_000000398901.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis racket of the boy in a striped shirt with glasses\"."}], "task": "refering", "answer_template": "The \"the tennis racket of the boy in a striped shirt with glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 122, 123, 124, 125, 127, 128, 129, 137, 140, 141, 142, 144, 145, 146, 148, 149, 150, 151, 159, 160, 162, 163, 164, 166, 168, 169, 172, 181, 182, 183, 210, 216, 223, 224, 225, 226, 232, 233, 235, 236, 237, 239, 240, 246, 247, 248, 249, 250, 255, 256, 258, 259, 260, 262, 263, 269, 270, 271, 272, 279, 282, 285, 292, 295, 302, 305, 308, 315, 318, 325, 328, 331, 338, 341], "bbox": [0.045234375, 0.2936192468619247, 0.9883906250000001, 0.8564644351464434], "iscrowd": 0, "area": 19309.108900000003, "rle": {"size": [478, 640], "counts": "Wh=1l>3N1N1O010O0002O0O2N10hQ6OXnI2O1N2N100O010O0010O01WDJj85TG4e8MXG;b8E\\Gd0^8\\O_Gm0Z8TOdGP1Z8oNdGU1Z8lNcGX1[8hNdG[1[8eNeG\\1Y8dNhG]1V8dNjG]1T8cNmG]1S8cNlG_1R8aNoG`1o7aNQH`1m7aNSH`1k7`NVHa1i7_NWHb1g7^NZHd1c7]N]Hk1X7WNiHR2h6TNXIU2Z6PNeIY2R6hMnIa2h5aMWJh2_5YMaJo2V5RMjJj0eM>Z7iNQKh0nM6Q7SOQKf0WNNh6ZOTKg0]NF`6@XKf0bN@W6H\\Ke0fNZOn50`Kc0kNTOe5:bK?SOmN[5d0eK<f6C]I:d6E_I8a6HcI4^6LdI1]6NVJ_Oj5a0Z4O1O010O1000O10000O100000000001O0000001O0001O010O2N11L4K6J5L4KQm10RSN1N10O0100O010O10O0100O010O10OTDOl82SGk0Q8UOoGn0m7ROSHP1l7POSHR1m7mNRHT1o7jNQHX1o7gNQHZ1h:1O2N101N2N1O2O10O01N101Ni0XOk0TO9H0O100O101N11UOk0SOl0GfLZNoId1R6^NmI_1T6cNlI[1T6hNjIV1W6lNiIR1X6POgIn0Y6TOPGEk1U1V7YOeFJV2k0U7^OaFH^2e0R7E\\FIe2>P7KWFIl28m6b0XIXOi6i0ZISOg6m0\\InNf6R1^IiNc6W1`IeNa6Z1a3100000000O010000O10O02N101N]BjN`=V1_BjNc=T1]BmNc=R1]BnNd=V11O0J[BROg=k0ZBVOf=i0[BVOf=h0[BYOf=e0[B[Oe=c0\\B]Oj=<WBEn=5SBJT>NmA3_>O0100O010O010O10O010O2O2Mo3OlK8G:G8L4N2N2N3M2N2N2N2N3M2N2N2N7Ia0_Oa0@`0oKaLbKP4m3QLSLQ4j3oKWLQ4h3QLWLn3i3TLVLl3i3VLVLa3S4`LlKd2o4^MPKh1i5ZNVJd1k5]NUJa1m5aNQJ]1Q6gNkIX1U6lNhIR1Z6\\OYIa0i6_OXI?i6AXI<j6DWI:i6GXI6j6IXI5i6KXI2j6MXI1h60YIMi62YILh64YIIi66YIHh67ZIFg6;_I^Ob6a0eIWO]6i0iIPOY6n0l301N1O2M2O2M2N3N1N5L8G`1MdN1O2N1O00001O000000010O00000000001O000000001O0001O000001O000000001O01O3M3MS\\1OmcN2N2O1N2O1O1O1O100O1O1O1O100O1O1O2O0O1O1O200O2N1O2N1O;FO01O00O1N2O2N1O1O1O1O2N1N4L5J6K5JVng0LoQXO>B>B4K4M3M3M3O1N2N2N2N2N2N2O1YETNY8n1oFiNQ9b3001O000cNnEnMR:j1cFcMc9V2f1K6M2N2N2M3N3M2N2M3N2N3L5L5K4L5K4L8I8GdV;N]iD6I7J6J5J7J6L4M2M3M101N2N1O2O0O2N2O0O2N1O2004K01O00:F:F;E:GcN[DJZ;5SEIc:5jEIm96_F_Oa9`0kFUOU9j0XGjNh8T1eGaN\\8]1c2O1O]NdNhEY1X:iNiES1X:POhEm0Y:SOiEi0Z:UOiEg0Z:XOhEe0Z:YOiEc0\\<Lob4Ab]K1N2O1N1O01O010O010O01O010O01O010O01O010O010O01O01O010O010O010O00010O010O001N1O2O0O1O2O1M4L3L4M3M3O1O1O1O2O0O1O1O1O100O1O1R1nNcR3"}, "label": "the tennis racket of the boy in a striped shirt with glasses"}]}
{"id": 120376, "image": "COCO_train2014_000000120376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a calf being bottle - fed\"."}], "task": "refering", "answer_template": "The \"a calf being bottle - fed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 145, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 293, 310, 311, 312, 313, 316, 317, 333, 334, 335, 357, 358], "bbox": [0.33040625, 0.33547884187082405, 0.7966093750000001, 0.9612917594654788], "iscrowd": 0, "area": 26665.358399999986, "rle": {"size": [449, 640], "counts": "PTm23n=1O2M3N1O1N3N1O1N3N1]O]OfCd0Y<@dC`0[<CbC?\\<CbC=^<DaC=^<E`C<_<EaC;^<G`C9`<H_C9`<H_C9`<H`C8`<e0O01000O10O10O10000O00001O001O001O001O00001O001O001O001O00001O0010O001O001O001@hCUOX<k0jCSOW<l0jCSOV<m0lCQOU<m0lCSOT<m0lCSOT<m0lCROV<m0jCSOW<k0jCUOV<k0jCSOY<l0gCSO[<l0eCSO]<l0cCSO^<l0cCSO_<l0<O1O1O0O2O1O1O0O2OlC]Ok:a0UEBj:=UEFj:8UELh:5WENg:1XE3f:MYE7d:H[E<c:D\\E?b:A]Ec0_:^O_Ee0`:ZO`Eh0_:XO_Ek0`:UO_El0a:TO^En0a:RO_Eo0_:QOaEP1_:POaEQ1^:oNbER1]:nNbET1]:lNcET1\\:lNeEU1Z:kNfEV1Y:iNgEX1Y:gNhEZ1k1TOk4A[I\\1`1_OR5UO^I]1[1DV5mN`I_1V1JY5eNbIb1:d0S6iMdId10m0[6]MfIg1GT1b6TMhIh1DX1c6nLjIk1@Z1i7gNUH\\1i7dNUH_1i7cNTHa1j7`NSHc1j7`NQHe1l77[HfMe7o1PIgMQ7S2^IdMa6\\2fI^MY6b2kI[MT6e2PJXMP6h2SJTMm5l2XJPMg5P3d20010O00000003lFgL]7[3_HmLg0F_4]3fJUMb0Ef4W3dJ[M>Ek4Q3bJcM:@T5m2\\JGd5;UJJk5V40001O000000000000N2N2mMeI^L]6a3gIZL[6d3jIVLY6h3lIRLW6l3nImKV6R4UJ^KQ6`4^JkJj5S5fInJk6Q5SIRKm6e4ZI]Kg6Y4_IjKa6l3fIWLZ6`3mIbLS6V3SJlLm5R3SJQMl5n2SJTMm5j2TJWMl5g2TJ[Mm5c2RJ_Mn5_2RJcMn5[2RJfMo5X2RJhMo5W2PJjMQ6T2oIlMS6R2mIoMU6o1jIRNW6n1hIRNY6o1cIQN`6Q2[InMi6T2RIjMS7W2hHkMZ7W2aHjMa7W2[HiMh7Y2RHiMP8Y2kGhMW8Y2cGjM_8X2[GjMg8W2VGiMl8d32N2N2N2N2O1N2N3M2N2O1N3N1[NWFSOh9m0YFSOg9l0[FjNOjNf9\\2\\FfN3mN`9]2^FbN8oNY9_2aF^NR:a1oE[NT:e1mEWNV:i1XFfMk9[2n00O1O1O1O1O100O1O1O1O1O1O1O1O1O10000000O1000O10000000001O000O27H8I9F:F7J6I6J10O001O00001O0010O000nN]Cf0b<YO_Cg0a<XO`Ci0`<UOaCk0_<SOcCm0^<QOcCc1i;\\NXDg1f;WN[Dk1m;2O1000000:F1N10000O100M2N3M2N1O2N2N1O2N2N1O2N3L6K4LYdh1"}, "label": "a calf being bottle - fed"}]}
{"id": 120376, "image": "COCO_train2014_000000120376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow that is being fed by a bottle\"."}], "task": "refering", "answer_template": "The \"a cow that is being fed by a bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 145, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 293, 310, 311, 312, 313, 316, 317, 333, 334, 335, 357, 358], "bbox": [0.33040625, 0.33547884187082405, 0.7966093750000001, 0.9612917594654788], "iscrowd": 0, "area": 26665.358399999986, "rle": {"size": [449, 640], "counts": "PTm23n=1O2M3N1O1N3N1O1N3N1]O]OfCd0Y<@dC`0[<CbC?\\<CbC=^<DaC=^<E`C<_<EaC;^<G`C9`<H_C9`<H_C9`<H`C8`<e0O01000O10O10O10000O00001O001O001O001O00001O001O001O001O00001O0010O001O001O001@hCUOX<k0jCSOW<l0jCSOV<m0lCQOU<m0lCSOT<m0lCSOT<m0lCROV<m0jCSOW<k0jCUOV<k0jCSOY<l0gCSO[<l0eCSO]<l0cCSO^<l0cCSO_<l0<O1O1O0O2O1O1O0O2OlC]Ok:a0UEBj:=UEFj:8UELh:5WENg:1XE3f:MYE7d:H[E<c:D\\E?b:A]Ec0_:^O_Ee0`:ZO`Eh0_:XO_Ek0`:UO_El0a:TO^En0a:RO_Eo0_:QOaEP1_:POaEQ1^:oNbER1]:nNbET1]:lNcET1\\:lNeEU1Z:kNfEV1Y:iNgEX1Y:gNhEZ1k1TOk4A[I\\1`1_OR5UO^I]1[1DV5mN`I_1V1JY5eNbIb1:d0S6iMdId10m0[6]MfIg1GT1b6TMhIh1DX1c6nLjIk1@Z1i7gNUH\\1i7dNUH_1i7cNTHa1j7`NSHc1j7`NQHe1l77[HfMe7o1PIgMQ7S2^IdMa6\\2fI^MY6b2kI[MT6e2PJXMP6h2SJTMm5l2XJPMg5P3d20010O00000003lFgL]7[3_HmLg0F_4]3fJUMb0Ef4W3dJ[M>Ek4Q3bJcM:@T5m2\\JGd5;UJJk5V40001O000000000000N2N2mMeI^L]6a3gIZL[6d3jIVLY6h3lIRLW6l3nImKV6R4UJ^KQ6`4^JkJj5S5fInJk6Q5SIRKm6e4ZI]Kg6Y4_IjKa6l3fIWLZ6`3mIbLS6V3SJlLm5R3SJQMl5n2SJTMm5j2TJWMl5g2TJ[Mm5c2RJ_Mn5_2RJcMn5[2RJfMo5X2RJhMo5W2PJjMQ6T2oIlMS6R2mIoMU6o1jIRNW6n1hIRNY6o1cIQN`6Q2[InMi6T2RIjMS7W2hHkMZ7W2aHjMa7W2[HiMh7Y2RHiMP8Y2kGhMW8Y2cGjM_8X2[GjMg8W2VGiMl8d32N2N2N2N2O1N2N3M2N2O1N3N1[NWFSOh9m0YFSOg9l0[FjNOjNf9\\2\\FfN3mN`9]2^FbN8oNY9_2aF^NR:a1oE[NT:e1mEWNV:i1XFfMk9[2n00O1O1O1O1O100O1O1O1O1O1O1O1O1O10000000O1000O10000000001O000O27H8I9F:F7J6I6J10O001O00001O0010O000nN]Cf0b<YO_Cg0a<XO`Ci0`<UOaCk0_<SOcCm0^<QOcCc1i;\\NXDg1f;WN[Dk1m;2O1000000:F1N10000O100M2N3M2N1O2N2N1O2N2N1O2N3L6K4LYdh1"}, "label": "a cow that is being fed by a bottle"}]}
{"id": 71232, "image": "COCO_train2014_000000071232.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white male wearing an orange and grey shirt with white shorts holding a tennis racket while on the court with 3 other people\"."}], "task": "refering", "answer_template": "The \"a white male wearing an orange and grey shirt with white shorts holding a tennis racket while on the court with 3 other people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 108, 109, 110, 132, 133, 134, 155, 156, 179, 201, 202, 224, 225, 247, 248, 270, 271, 293, 294], "bbox": [0.724796875, 0.22697399527186762, 0.8317343749999999, 0.8561702127659574], "iscrowd": 0, "area": 8080.322349999999, "rle": {"size": [423, 640], "counts": "`io58c<<L5N1O1O1O101N1bHSOa2n0QMBl2?fL1W30\\L?a3C^L?`3A`La0]3BaL`0\\3DaL>\\3EbL=[3FdL;Z3GdL;Y3HeL:X3IfL9W3JgL9V3IhL9V3IhL9V3IhL9W3HgL:W3HgL:W3HgL:W3HgL:W3HgL:X3GgL:W3HgL:W3HgL:W3HgL:W3HgL:X3GfL;X3GfL;X3GfL;X3GfL;X3GfL;Y3FeL<Y3FeL<Y3FeL:[3HcL8]3JbL4`3M^L2c30[LOf33YLKh37^L@c3b0cLUO^3m0hLjNZ3W1lL`NX3^2oKZM\\4T2TLbMX4l1WLkMP4h1_LoMd3h1kLoMW3h1YMoMj2g1fMPN]2g1f4I8J6I7I6J7J6I6JP_\\1"}, "label": "a white male wearing an orange and grey shirt with white shorts holding a tennis racket while on the court with 3 other people"}]}
{"id": 71232, "image": "COCO_train2014_000000071232.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player in orange and gray shirt\"."}], "task": "refering", "answer_template": "The \"tennis player in orange and gray shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 108, 109, 110, 132, 133, 134, 155, 156, 179, 201, 202, 224, 225, 247, 248, 270, 271, 293, 294], "bbox": [0.724796875, 0.22697399527186762, 0.8317343749999999, 0.8561702127659574], "iscrowd": 0, "area": 8080.322349999999, "rle": {"size": [423, 640], "counts": "`io58c<<L5N1O1O1O101N1bHSOa2n0QMBl2?fL1W30\\L?a3C^L?`3A`La0]3BaL`0\\3DaL>\\3EbL=[3FdL;Z3GdL;Y3HeL:X3IfL9W3JgL9V3IhL9V3IhL9V3IhL9W3HgL:W3HgL:W3HgL:W3HgL:W3HgL:X3GgL:W3HgL:W3HgL:W3HgL:W3HgL:X3GfL;X3GfL;X3GfL;X3GfL;X3GfL;Y3FeL<Y3FeL<Y3FeL:[3HcL8]3JbL4`3M^L2c30[LOf33YLKh37^L@c3b0cLUO^3m0hLjNZ3W1lL`NX3^2oKZM\\4T2TLbMX4l1WLkMP4h1_LoMd3h1kLoMW3h1YMoMj2g1fMPN]2g1f4I8J6I7I6J7J6I6JP_\\1"}, "label": "tennis player in orange and gray shirt"}]}
{"id": 71232, "image": "COCO_train2014_000000071232.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person wearing khaki - colored shorts\"."}], "task": "refering", "answer_template": "The \"the person wearing khaki - colored shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 97, 98, 99, 120, 121, 143, 144, 145, 166, 167, 168, 189, 190, 212, 213, 235, 236, 258, 259, 281, 282, 304, 305], "bbox": [0.214875, 0.2376595744680851, 0.3334375, 0.9147990543735225], "iscrowd": 0, "area": 13102.336250000002, "rle": {"size": [423, 640], "counts": "jTi1R1T<9iDhN^1c1_4k0RKAk4d0hJHT5=oI>n5a3L3M3M4L3M3M3M2N2N2N2AaHSKa7k4eHoJ]7n4?O1N2N2N3M2O1N3O00001N10\\NeK\\J[4^5kKaIGa0^4h5QLfICa0[4d5XLkI^O`0Z4a5\\LoI[O?Y4`5]LRJZO>Y4]5`LTJYOb0R4X5hLVJWOa0Q4V5kLYJUO`0P4T5nL\\JSO>P4T5oL^JQO>P4P5SMaJoN>o3l4VMfJlN=o3h4YMiJkN>m3d4\\MlJjN?k3R3fK`Mi1mNhNa0j3Q3hKZMj1ROgNa0i3R3hKTMn1VOdNc0g3R3jKmLn20X1S3mKdLR38R1S3oK\\LV3?m0T3HjL:?hKX1P4VN;`0iKW1m3WN<`0jKW1k3VN>`0kKX1h3VN>a0mKW1f3VN?a0mKX1j8fNYGX1h8eN\\GY1e8eN^GX1d8fN_GX1b8fN`GY1`8eNbG[1_8cNbG]1_8aNcG]1_8aNbG^1T:O1O1O2N1VO[D1f;KcDN_;NiDMX;OnW`5"}, "label": "the person wearing khaki - colored shorts"}]}
{"id": 71232, "image": "COCO_train2014_000000071232.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black jacket holds a racket on the left of the net\"."}], "task": "refering", "answer_template": "The \"a man in a black jacket holds a racket on the left of the net\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 97, 98, 99, 120, 121, 143, 144, 145, 166, 167, 168, 189, 190, 212, 213, 235, 236, 258, 259, 281, 282, 304, 305], "bbox": [0.214875, 0.2376595744680851, 0.3334375, 0.9147990543735225], "iscrowd": 0, "area": 13102.336250000002, "rle": {"size": [423, 640], "counts": "jTi1R1T<9iDhN^1c1_4k0RKAk4d0hJHT5=oI>n5a3L3M3M4L3M3M3M2N2N2N2AaHSKa7k4eHoJ]7n4?O1N2N2N3M2O1N3O00001N10\\NeK\\J[4^5kKaIGa0^4h5QLfICa0[4d5XLkI^O`0Z4a5\\LoI[O?Y4`5]LRJZO>Y4]5`LTJYOb0R4X5hLVJWOa0Q4V5kLYJUO`0P4T5nL\\JSO>P4T5oL^JQO>P4P5SMaJoN>o3l4VMfJlN=o3h4YMiJkN>m3d4\\MlJjN?k3R3fK`Mi1mNhNa0j3Q3hKZMj1ROgNa0i3R3hKTMn1VOdNc0g3R3jKmLn20X1S3mKdLR38R1S3oK\\LV3?m0T3HjL:?hKX1P4VN;`0iKW1m3WN<`0jKW1k3VN>`0kKX1h3VN>a0mKW1f3VN?a0mKX1j8fNYGX1h8eN\\GY1e8eN^GX1d8fN_GX1b8fN`GY1`8eNbG[1_8cNbG]1_8aNcG]1_8aNbG^1T:O1O1O2N1VO[D1f;KcDN_;NiDMX;OnW`5"}, "label": "a man in a black jacket holds a racket on the left of the net"}]}
{"id": 292416, "image": "COCO_train2014_000000292416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl filled with fruit\"."}], "task": "refering", "answer_template": "The \"the bowl filled with fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147], "bbox": [0.213375, 0.2041079812206573, 0.462203125, 0.4373943661971831], "iscrowd": 0, "area": 12042.887700000001, "rle": {"size": [426, 640], "counts": "PTi17Q=5K4L5K4L4L4L4L4L4N1N2O1N2O1N2N2O0O2O1N2O1N2N2O1N2O0O2N100O101N100O1O2O0O101N100O1O2O0O101N1O2O00001O0O101O001O00000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000O2O0O1O100O2O0O1O100O2O0O1O100O2O0O1O100O2O0O100O1O2O0O100O1O2O0O100O1O2O0O2O1N2N2O1N2O1N2L3M4L4L4L4L4M3L5K7I6J7Iio^4"}, "label": "the bowl filled with fruit"}]}
{"id": 292416, "image": "COCO_train2014_000000292416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fruits in white bowl . also seen parts of another bowl and plate\"."}], "task": "refering", "answer_template": "The \"fruits in white bowl . also seen parts of another bowl and plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147], "bbox": [0.213375, 0.2041079812206573, 0.462203125, 0.4373943661971831], "iscrowd": 0, "area": 12042.887700000001, "rle": {"size": [426, 640], "counts": "PTi17Q=5K4L5K4L4L4L4L4L4N1N2O1N2O1N2N2O0O2O1N2O1N2N2O1N2O0O2N100O101N100O1O2O0O101N100O1O2O0O101N1O2O00001O0O101O001O00000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000O2O0O1O100O2O0O1O100O2O0O1O100O2O0O1O100O2O0O100O1O2O0O100O1O2O0O100O1O2O0O2O1N2N2O1N2O1N2L3M4L4L4L4L4M3L5K7I6J7Iio^4"}, "label": "fruits in white bowl . also seen parts of another bowl and plate"}]}
{"id": 153154, "image": "COCO_train2014_000000153154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green houseplant next to a cat\"."}], "task": "refering", "answer_template": "The \"a green houseplant next to a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 17, 18, 19, 20, 30, 31, 32, 33, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 70, 71, 72, 73, 83, 96], "bbox": [0.3116111111111111, 0.006734374999999999, 0.7111111111111111, 0.357296875], "iscrowd": 0, "area": 15824.8323, "rle": {"size": [640, 360], "counts": "ZPV22nc07I8G8I7I8H7I7I8G8I7I7I8H6J4K5L4L4L3M4L4K5L4L3M1O2N1N3N1O1O2N1O1N3O0O2N1O101Nm0SO3M3M4M2M3M3McNaB\\L]=c3kBXLR=h3WCRLf<n3bCmK[<S4nCgKo;Y4m1O0O101O00001O000O2O00001O0O101O00001N10001O001O000010O0001O00001O001O001O001O001O001N101O001O0RMZ@b1g?ZN^@c1c?[Na@c1`?YNf@d1Z?YNk@e1V?WNo@g1Q?WNTAe1n>WNWAg1i>WN[Ag1f>VN_Ag1a>WNcAg1^>UNhAh1X>VNmAf1U>WNPBf1P>WNWBe1U`0M4L4K2O001O1O1N1O2O1N2N1O2O1N2N1O2O1N2M2O2M3M3N1N3N1N1O10l_Q2"}, "label": "a green houseplant next to a cat"}]}
{"id": 153154, "image": "COCO_train2014_000000153154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant is there above the cat\"."}], "task": "refering", "answer_template": "The \"a plant is there above the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 17, 18, 19, 20, 30, 31, 32, 33, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 70, 71, 72, 73, 83, 96], "bbox": [0.3116111111111111, 0.006734374999999999, 0.7111111111111111, 0.357296875], "iscrowd": 0, "area": 15824.8323, "rle": {"size": [640, 360], "counts": "ZPV22nc07I8G8I7I8H7I7I8G8I7I7I8H6J4K5L4L4L3M4L4K5L4L3M1O2N1N3N1O1O2N1O1N3O0O2N1O101Nm0SO3M3M4M2M3M3McNaB\\L]=c3kBXLR=h3WCRLf<n3bCmK[<S4nCgKo;Y4m1O0O101O00001O000O2O00001O0O101O00001N10001O001O000010O0001O00001O001O001O001O001O001N101O001O0RMZ@b1g?ZN^@c1c?[Na@c1`?YNf@d1Z?YNk@e1V?WNo@g1Q?WNTAe1n>WNWAg1i>WN[Ag1f>VN_Ag1a>WNcAg1^>UNhAh1X>VNmAf1U>WNPBf1P>WNWBe1U`0M4L4K2O001O1O1N1O2O1N2N1O2O1N2N1O2O1N2M2O2M3M3N1N3N1N1O10l_Q2"}, "label": "a plant is there above the cat"}]}
{"id": 374340, "image": "COCO_train2014_000000374340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the basset hound that is laying down\"."}], "task": "refering", "answer_template": "The \"the basset hound that is laying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 362, 363, 367, 385, 386], "bbox": [0.74353125, 0.5409583333333334, 1.0, 0.9849166666666667], "iscrowd": 0, "area": 17241.446099999997, "rle": {"size": [480, 640], "counts": "oao67d>6O1N2O1N2O2M2O1N2O1N2O1N2N2O2M2O1N01O001O001O001OROo0B=0000O10000000000O2O0000O1N2N2M2O2N2N2M3N3M3M4L3L4M3M30001O000000N2N2M3cMbNVG`1i8gNnF\\1R9kNdFX1[9oN]FS1b9TOUFo0k9VOnEl0Q:YOhEj0W:[ObEh0^:\\O\\Ef0c:^OWEe0h:j100O1O100O1O100O1O100O1O100O1O2O1N1O2O0O2O0010O01O010O010O1O010O010O0010O010O01O01O010O00001O010O000010O01O000010O01O000010O01O00010O001O00010O001O00010O001O01O01O001O0gLdEo1[:kMmES2T:gMTFV2l9dM]FY2d9`MeFT2d9fMdFY1[:aNnE>Q2"}, "label": "the basset hound that is laying down"}]}
{"id": 374340, "image": "COCO_train2014_000000374340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and white dog wearing a black collar in a crate between two other dogs\"."}], "task": "refering", "answer_template": "The \"a brown and white dog wearing a black collar in a crate between two other dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 268, 269, 270], "bbox": [0.46118750000000003, 0.1330625, 0.79990625, 0.6880833333333334], "iscrowd": 0, "area": 23306.882949999996, "rle": {"size": [480, 640], "counts": "^P[43l>2O2M2N2O1N00100O1O01000000O1000000O10O10000000O010000O1O1O1O001O1O00001O000O101O0N2N2N3M3M4M2O2M3N1N3O0O2O1N101N2O0O2O1O1N101N2O1O0O2N2N2O0nNdN[Db1Y;Y1K5K4L5J6N2M2O2mMoKTIS4j6RLPIQ4n6TLmHm3R7WLjHg3kNbKZ8h0hHi3nN_KX8j0hHh3QO^KU8l0hHh3TOZKS8P1gHh3VOXKQ8S1eHh3ZOUKo7U1eHg3_7ZL^Hh3b7YL[Hi3d7_LTHb3l7i1O101N100O10000O100O10000O100000000O100000000O10O1000000000O01000000O1000000O2O^HTIP7l6mHXIR7g6kH^IT7W701O100O01000O10O1O100O1O10O01O100O1O2QOgHTJ[7d5kH[JY7\\5mHcJV7U5oHmJQ7n4SITKl66[H^3n0]Lg60aH]3k0fLc6HhH[3j0lL`6DlHo2S1\\MZ7X2nHhMU7R2oHnM[7d1iH[Nc7W1_HiNn7g0THZOX87kGIf;00000O10O100000O1000O10O1O1O1O000000001O00001O001N100O2OO0O1ON4M3001O10O01O0O2M4M4K4L4M3N2M7IZgk1"}, "label": "a brown and white dog wearing a black collar in a crate between two other dogs"}]}
{"id": 374340, "image": "COCO_train2014_000000374340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white dog in cage\"."}], "task": "refering", "answer_template": "The \"white dog in cage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 268, 269, 270], "bbox": [0.46118750000000003, 0.1330625, 0.79990625, 0.6880833333333334], "iscrowd": 0, "area": 23306.882949999996, "rle": {"size": [480, 640], "counts": "^P[43l>2O2M2N2O1N00100O1O01000000O1000000O10O10000000O010000O1O1O1O001O1O00001O000O101O0N2N2N3M3M4M2O2M3N1N3O0O2O1N101N2O0O2O1O1N101N2O1O0O2N2N2O0nNdN[Db1Y;Y1K5K4L5J6N2M2O2mMoKTIS4j6RLPIQ4n6TLmHm3R7WLjHg3kNbKZ8h0hHi3nN_KX8j0hHh3QO^KU8l0hHh3TOZKS8P1gHh3VOXKQ8S1eHh3ZOUKo7U1eHg3_7ZL^Hh3b7YL[Hi3d7_LTHb3l7i1O101N100O10000O100O10000O100000000O100000000O10O1000000000O01000000O1000000O2O^HTIP7l6mHXIR7g6kH^IT7W701O100O01000O10O1O100O1O10O01O100O1O2QOgHTJ[7d5kH[JY7\\5mHcJV7U5oHmJQ7n4SITKl66[H^3n0]Lg60aH]3k0fLc6HhH[3j0lL`6DlHo2S1\\MZ7X2nHhMU7R2oHnM[7d1iH[Nc7W1_HiNn7g0THZOX87kGIf;00000O10O100000O1000O10O1O1O1O000000001O00001O001N100O2OO0O1ON4M3001O10O01O0O2M4M4K4L4M3N2M7IZgk1"}, "label": "white dog in cage"}]}
{"id": 243268, "image": "COCO_train2014_000000243268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a baseball uniform and cap throwing a ball\"."}], "task": "refering", "answer_template": "The \"a man in a baseball uniform and cap throwing a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 266, 267, 268, 269, 283, 284, 285, 286, 287, 291, 292, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 357, 378, 379, 380], "bbox": [0.321421875, 0.289625, 0.72365625, 0.9851666666666666], "iscrowd": 0, "area": 45906.9582, "rle": {"size": [480, 640], "counts": "XmP3g1n<P1QO:F:E;F:I7I7H9H7I7I7I7H8I7J6O1O101N1O1O100O1O100O1O1O101N1O100O1O100O1O1O101N1O100O1O100O1O1O101N1O100O1O1O100O1O101N1O100O1O1O100O1O100O2N3M4M4K4L4M3L4L4M2M3M2N100O1O100O1O10000O10000O1000000O10000O1000000O10000O2O0000UMZJ\\Le5`3aJ_L_5\\3gJcLY5W3nJhLQ5T3UKkLk4P3\\KnLd4l2cKgLjNSNc5Q5iKgLkNUN[5o4QLhLiNWNV5l4WLkL]4P3jKnLU4m2RLRMn3i2[LTMd3g2dLVM\\3j2gLSMX3m2lLPMT3o2PMoLo2Q3TMlLk2T3YMiLg2W3]MfLb2Y3bMdL^2\\3eMaLZ2_3jM_LU2`3oM]LQ2c3RNZLm1f3WNTLl1l3XNlKl1S4a40O0100O00100O010O1001N2O1O1N2O1O1O1N2O001M3N2N2N2M3N2N2N2M10000O2O0O101N100O101000100O001O1O00100O001O1OeLeFQ1\\9mNiFP1V9oNoFn0P9RORGl0n8SOSGn0l8QOUGP1k8nNUGS1k8lNVGU1i8jNXGW1g8hNYGY1g8fNZG[1e8dN\\G]1d8aN\\Gc0[OaNY9k0]Gd0ZObNX9i0_Ge0ZObNV9i0_Gf0[OaNV9h0`Gf0\\ObNT9f0aGh0\\ObNS9d0aGk0\\OaNU9a0`Gm0]ObNT9>_GQ1]ObNT9;`GS1]OaNU99_GU1]OcNU95`GW1\\OdNU92`GZ1\\OcNV90`G[1[OfNV9DiGe1ROgNV9XOSHP2gNhNR;W1PEiNP;V1PEkNP;T1QEkNP;T1QElNn:S1SEnNm:Q1TEnNm:Q1SEPOl:o0VEQOj:n0VEROk:m0VESOj:k0XETOh:l0XEUOh:j0YEVOg:h0ZEXOg:g0ZEYOe:g0[EZOT;5nDJc;_OdDa0m<10O0100O01000O010O10O001N2M4M4K4L5L3L5K^db2"}, "label": "a man in a baseball uniform and cap throwing a ball"}]}
{"id": 243268, "image": "COCO_train2014_000000243268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player throw over the ball\"."}], "task": "refering", "answer_template": "The \"the player throw over the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 266, 267, 268, 269, 283, 284, 285, 286, 287, 291, 292, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 357, 378, 379, 380], "bbox": [0.321421875, 0.289625, 0.72365625, 0.9851666666666666], "iscrowd": 0, "area": 45906.9582, "rle": {"size": [480, 640], "counts": "XmP3g1n<P1QO:F:E;F:I7I7H9H7I7I7I7H8I7J6O1O101N1O1O100O1O100O1O1O101N1O100O1O100O1O1O101N1O100O1O100O1O1O101N1O100O1O1O100O1O101N1O100O1O1O100O1O100O2N3M4M4K4L4M3L4L4M2M3M2N100O1O100O1O10000O10000O1000000O10000O1000000O10000O2O0000UMZJ\\Le5`3aJ_L_5\\3gJcLY5W3nJhLQ5T3UKkLk4P3\\KnLd4l2cKgLjNSNc5Q5iKgLkNUN[5o4QLhLiNWNV5l4WLkL]4P3jKnLU4m2RLRMn3i2[LTMd3g2dLVM\\3j2gLSMX3m2lLPMT3o2PMoLo2Q3TMlLk2T3YMiLg2W3]MfLb2Y3bMdL^2\\3eMaLZ2_3jM_LU2`3oM]LQ2c3RNZLm1f3WNTLl1l3XNlKl1S4a40O0100O00100O010O1001N2O1O1N2O1O1O1N2O001M3N2N2N2M3N2N2N2M10000O2O0O101N100O101000100O001O1O00100O001O1OeLeFQ1\\9mNiFP1V9oNoFn0P9RORGl0n8SOSGn0l8QOUGP1k8nNUGS1k8lNVGU1i8jNXGW1g8hNYGY1g8fNZG[1e8dN\\G]1d8aN\\Gc0[OaNY9k0]Gd0ZObNX9i0_Ge0ZObNV9i0_Gf0[OaNV9h0`Gf0\\ObNT9f0aGh0\\ObNS9d0aGk0\\OaNU9a0`Gm0]ObNT9>_GQ1]ObNT9;`GS1]OaNU99_GU1]OcNU95`GW1\\OdNU92`GZ1\\OcNV90`G[1[OfNV9DiGe1ROgNV9XOSHP2gNhNR;W1PEiNP;V1PEkNP;T1QEkNP;T1QElNn:S1SEnNm:Q1TEnNm:Q1SEPOl:o0VEQOj:n0VEROk:m0VESOj:k0XETOh:l0XEUOh:j0YEVOg:h0ZEXOg:g0ZEYOe:g0[EZOT;5nDJc;_OdDa0m<10O0100O01000O010O10O001N2M4M4K4L5L3L5K^db2"}, "label": "the player throw over the ball"}]}
{"id": 521797, "image": "COCO_train2014_000000521797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in dark clothes leans against a street light while texting\"."}], "task": "refering", "answer_template": "The \"a woman in dark clothes leans against a street light while texting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 314, 315, 316, 338, 360, 361, 383, 384, 406, 407, 428, 429, 430], "bbox": [0.644140625, 0.22184374999999998, 0.84684375, 0.832203125], "iscrowd": 0, "area": 21752.8488, "rle": {"size": [640, 640], "counts": "WPR81mc03M3L4M3M3M4L3M3L4M3M3M1O000O1O1N2_IiNXJZ1g5jNSJY1m5iNnIZ1T6gNfI_1X6cNoEH_1i1b8`NhEMa1h1f8]NcE0b1i1j8WN^E7b1g1o8TNYE:c1h1S9oMSE?f1f1W9lMmDd0f1e1\\9hMdDm0?bN:n2b:dMVDZ1h0ZN<l2e:ChDgM>k2i:_OdDkM?i2m:]O^DPN`0f2S;[OWDUNb0b2X;<dDF^;:]DHe;9WDIj;8RDKn;T4O1N2N1O2N2M2O2N2ND<_Oa0_Oa0H8L3M4L4L4M3LXORJhCj5\\<ZJ`Cb5c<dJXC\\5h<iJRCW5o<nJlBQ5U=l0O0100O1\\Oc02O2N101O0010O01O3N2M2N3M101N1O100O1O1O2O0O1O1000001O00000000001O000000000000I7M3O3L3N3M3M2N3L4M2N3M2N3L4M2N3M3lMS2mMT2M3M2M4M>BURm1"}, "label": "a woman in dark clothes leans against a street light while texting"}]}
{"id": 521797, "image": "COCO_train2014_000000521797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman leaning against a pole looking at her phone\"."}], "task": "refering", "answer_template": "The \"a woman leaning against a pole looking at her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 314, 315, 316, 338, 360, 361, 383, 384, 406, 407, 428, 429, 430], "bbox": [0.644140625, 0.22184374999999998, 0.84684375, 0.832203125], "iscrowd": 0, "area": 21752.8488, "rle": {"size": [640, 640], "counts": "WPR81mc03M3L4M3M3M4L3M3L4M3M3M1O000O1O1N2_IiNXJZ1g5jNSJY1m5iNnIZ1T6gNfI_1X6cNoEH_1i1b8`NhEMa1h1f8]NcE0b1i1j8WN^E7b1g1o8TNYE:c1h1S9oMSE?f1f1W9lMmDd0f1e1\\9hMdDm0?bN:n2b:dMVDZ1h0ZN<l2e:ChDgM>k2i:_OdDkM?i2m:]O^DPN`0f2S;[OWDUNb0b2X;<dDF^;:]DHe;9WDIj;8RDKn;T4O1N2N1O2N2M2O2N2ND<_Oa0_Oa0H8L3M4L4L4M3LXORJhCj5\\<ZJ`Cb5c<dJXC\\5h<iJRCW5o<nJlBQ5U=l0O0100O1\\Oc02O2N101O0010O01O3N2M2N3M101N1O100O1O1O2O0O1O1000001O00000000001O000000000000I7M3O3L3N3M3M2N3L4M2N3M2N3L4M2N3M3lMS2mMT2M3M2M4M>BURm1"}, "label": "a woman leaning against a pole looking at her phone"}]}
{"id": 128581, "image": "COCO_train2014_000000128581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a projector screen on a wall behind people at a meeting\"."}], "task": "refering", "answer_template": "The \"a projector screen on a wall behind people at a meeting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.85059375, 0.0005647058823529411, 1.0, 0.4230352941176471], "iscrowd": 0, "area": 16951.20315, "rle": {"size": [425, 640], "counts": "PiQ7:o<S3mLV2jM00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000"}, "label": "a projector screen on a wall behind people at a meeting"}]}
{"id": 128581, "image": "COCO_train2014_000000128581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large projector screen on wall\"."}], "task": "refering", "answer_template": "The \"large projector screen on wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.85059375, 0.0005647058823529411, 1.0, 0.4230352941176471], "iscrowd": 0, "area": 16951.20315, "rle": {"size": [425, 640], "counts": "PiQ7:o<S3mLV2jM00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000"}, "label": "large projector screen on wall"}]}
{"id": 267851, "image": "COCO_train2014_000000267851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"picture of an eldery woman sitting on a bench in the middle of two others\"."}], "task": "refering", "answer_template": "The \"picture of an eldery woman sitting on a bench in the middle of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 172, 173, 195, 196, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 355, 356, 357, 358, 379, 380, 381, 402, 403, 404], "bbox": [0.411640625, 0.29775, 0.6420625, 0.7765], "iscrowd": 0, "area": 25928.198050000003, "rle": {"size": [640, 640], "counts": "eYU57gc07Y]OD^a0c0\\^O@`a0g0Y^O]Oca0j0W^OYOea0n0U^OUOga0R1S^OPOja0f1M3L4L4L4L4L4L4L3N3L4L4L4L4L4L4L4N1N3N2M3N2N2M5L4K5L3L5L4K5L4L4K5L4K5L4K5ZOf0ZOf0K5J6K5J6K5K1N101N2O1N101O1N2O0O2O\\NXERJh:j4aFQK_9d4QGVKo8k4WGoJi8R5[GiJe8W5aGbJ`8^5eG]JZ8d5kGWJU8i5QHQJo7o5Q30O10001O0O100000000O101O0001O1n0TDoHl8X7gFmHZ9Y7[FlHe9[7nEkHR:\\7bEiH^:]7WEhHi:X8000\\HVEQ6j:cIbE\\6_:WImEj6S:jHXFW7h9]HdF`7o:O2N1O1O2N1O1O2N1N3N1O1O2L3M4M2M3N3L3M4M2M3M4M2lNT1K6J5L5J5K8I7H7I8H8I6I8I7I6J7I7I7I6J7I7I6J7I7IZj^4"}, "label": "picture of an eldery woman sitting on a bench in the middle of two others"}]}
{"id": 267851, "image": "COCO_train2014_000000267851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a old women sitting the chair\"."}], "task": "refering", "answer_template": "The \"a old women sitting the chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 172, 173, 195, 196, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 355, 356, 357, 358, 379, 380, 381, 402, 403, 404], "bbox": [0.411640625, 0.29775, 0.6420625, 0.7765], "iscrowd": 0, "area": 25928.198050000003, "rle": {"size": [640, 640], "counts": "eYU57gc07Y]OD^a0c0\\^O@`a0g0Y^O]Oca0j0W^OYOea0n0U^OUOga0R1S^OPOja0f1M3L4L4L4L4L4L4L3N3L4L4L4L4L4L4L4N1N3N2M3N2N2M5L4K5L3L5L4K5L4L4K5L4K5L4K5ZOf0ZOf0K5J6K5J6K5K1N101N2O1N101O1N2O0O2O\\NXERJh:j4aFQK_9d4QGVKo8k4WGoJi8R5[GiJe8W5aGbJ`8^5eG]JZ8d5kGWJU8i5QHQJo7o5Q30O10001O0O100000000O101O0001O1n0TDoHl8X7gFmHZ9Y7[FlHe9[7nEkHR:\\7bEiH^:]7WEhHi:X8000\\HVEQ6j:cIbE\\6_:WImEj6S:jHXFW7h9]HdF`7o:O2N1O1O2N1O1O2N1N3N1O1O2L3M4M2M3N3L3M4M2M3M4M2lNT1K6J5L5J5K8I7H7I8H8I6I8I7I6J7I7I7I6J7I7I6J7I7IZj^4"}, "label": "a old women sitting the chair"}]}
{"id": 267851, "image": "COCO_train2014_000000267851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person sitting on the right side of a bench\"."}], "task": "refering", "answer_template": "The \"a person sitting on the right side of a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 176, 177, 178, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 336, 337, 338, 339, 359, 360, 362, 382, 383, 385, 386, 405, 406, 408, 409], "bbox": [0.6250312499999999, 0.28925, 0.8400000000000001, 0.761015625], "iscrowd": 0, "area": 15178.275050000002, "rle": {"size": [640, 640], "counts": "l[j71nc02N3fLO`B3JM_:6eE0D0a:7hEL_O4h:3eD@SO;Y18o:MaDFTO7o0b0\\;B]DJUO4a0P1m;SOZDMUO19Z1X<iNjC_OZO?81;Y1X<iNcCDC:516YOCn1k<lN[CJL2425ZOEl1k<mNZCIM1347YODl1k<nNZCGM1367XODl1k<nN[CFL1387WODl1k<mN]CFJ1487XOCl1k<lN_CFH1685YOCl1k<kNSCXO6?N02AIg0`0ZOCl1k<kNSCYO6=N02CJd0`0\\OBl1j<kNTC[O6:N02FJa0`0]OBl1j<kNTC\\O68N11GM>>_OBl1j<jNUC_O36001JL<?_OBl1j<jNTCA43001ML8`0AAl0OKk<NUCC31001OL7`0AAj02Lh<OTCE4N0003L4a0AAi04Lf<OUCG3L0005M1a0C@g06Nc<0UCI2I2O08LOb0C@e09O`<0VCK1G3NO<LLb0D@c0;0_<1TCM2D4MN?LJc0E_Oa0>0i;_OSDc0EO0B7LLc0KHd0E_O?`02c;CWD?E1OZO=1Fg0LDd0F_O=c02];IZD:E4j01TOBd0H_O:f03V;N^D6E5h02UO@e0H@7g05Q;3`D2G6f01WO_Od0I@5j05k:9cDNG6f02VO^Oe0I@3l07e:>fDHI7f02UO\\Of0K_O1o07`:c0hDEI7f02UOZOh0M]O0P18\\:f0kDAI9g01SOYOj0N\\ONS18W:k0mD]OJ9g02SOVOk01g06j9n0UETOF?g02QOUOg26c8T1_EhN@g0n0Jd1B]8X1PF[O:]O^1OW8Z1PFZOf3ZO[6\\1oElNT4Hl5]1jKaNW4_1iK`NW4b1S70O100j@hNc;Y1fA]NU2`0P<S1kAbN4b0>L\\ONV=R1kBhN0=a0K_ONT=S1lBlNK9d0K@NU=R1lBnNI7e0LAMT=T1lBlNK5f0M_ONT=T1lBmNJ4h0L]O0T=T1mBlNJ4h0L]O0T=T1mBlNJ4i0K\\O1T=T1mBmNJ2j0K[O2S=U1nBlNJ1l0KXO4T=T1nBmNIOo0KVO5S=U1oBlNINc0@@;26S=V1oBjNIN5AL42538Q=W1PCjNIJ6EH7502:S=U1oBkNII6GG86M2;T=T1nBmNIF6JG77L1<T=T1nBmNIE6LG77J1=U=T1mBlNJD5OG77H0?W=R1lBnNIB61G77F0`0X=Q1kBoNJ@54G77D0a05ZO]1OY9g1XEQOI^O57H67B0a05\\O\\10Z9d1XEROJ\\O4:I56A0b05\\O[12[9a1XETOJYO4=I5d02G\\OI3S10k9`1WESOJXO5`0H4e01F]OI5Q1Om9^1WEUO5HB4d00G]OI8m0MQ:]1WEUO4IC4c0OH]OI:j0MT:Z1WEWO2IE4c0MH^OI<g0MW:W1XEWO0KF4b0LI^OI>d0L[:V1VEXOOKH4b0JI_OI`0b0K]:T1VEYOOKH5a0IJ_OIb0?Ja:R1UE[O0HG7a0IJ_OId0<Jd:o0VE[O1GE8c0IH@If0:Hg:o0TEZO5EC:d0IFAIi06Gj:m0UEYO8D@9g0LCAIm02CR;i0QE]O9C_O87^O5?NBIT1J]O`;c0kD@KZO89E79C3<NCIk1^;TOgDDK]O95D7:G1:OCIk1d;nNaDIJA;0C6=KN8ODIl1g;jN]DKKD<LB6>OM5OFHk1j;hN\\DLIG>HA5`04K2OGHk1l;gNYDMIIe0IL8I00GHl1m;dNXDNJKc0E0<FN0HHl1o;bNWDOIL]13lNK0IHl1P<bNUDNKM0ER1`0VOH0JHm1Q<_NUDOJO0Fn0a0YOD1KHm1S<]NSD0K0OJi0?^OA1LIl1T<]NQD0K1OM>H@e09_O1MIm1U<ZNPD1L3M01ON:9^O2MHm1X<YNnC2K4NO00N::]O3L6^1j;jNkC1M5MO01N8<[OV1P1o:UOhC2L7LM22M7>[O\\1h0i:^OeC2M7LM14N4e2KQ:FcC3K7NMO4O4m2Al9O`C4K8ML05O2R`0AW@4I9OK06N0U`0AU@5I:NJ16OOU`0AU@5H;OI08NNW`0AS@5H<0H08OMX`0@R@5H?NG19NKj`0FY_O?OF0:OJj`0FX_O`0OF1:MJk`0FW_Ob0OD2;MHl`0EW_Od0NE1:NGm`0FV_Od0OD1;MFn`0FT_Og0OB2;MFn`0EU_Oh0NB2<MBXa0`0i^OB3;LAZa0c0f^OA4;M_OYa0Lf^Oe01D3<M]Ooa04Q^O33<N[ORb03m]O64;Yb0\\Oc]O3;`0nb0N1O1O2O0O1O2N1O2N1O1O2N1O1Nn\\o1"}, "label": "a person sitting on the right side of a bench"}]}
{"id": 267851, "image": "COCO_train2014_000000267851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with dark hair sitting to the right of an elderly woman\"."}], "task": "refering", "answer_template": "The \"a girl with dark hair sitting to the right of an elderly woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 176, 177, 178, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 336, 337, 338, 339, 359, 360, 362, 382, 383, 385, 386, 405, 406, 408, 409], "bbox": [0.6250312499999999, 0.28925, 0.8400000000000001, 0.761015625], "iscrowd": 0, "area": 15178.275050000002, "rle": {"size": [640, 640], "counts": "l[j71nc02N3fLO`B3JM_:6eE0D0a:7hEL_O4h:3eD@SO;Y18o:MaDFTO7o0b0\\;B]DJUO4a0P1m;SOZDMUO19Z1X<iNjC_OZO?81;Y1X<iNcCDC:516YOCn1k<lN[CJL2425ZOEl1k<mNZCIM1347YODl1k<nNZCGM1367XODl1k<nN[CFL1387WODl1k<mN]CFJ1487XOCl1k<lN_CFH1685YOCl1k<kNSCXO6?N02AIg0`0ZOCl1k<kNSCYO6=N02CJd0`0\\OBl1j<kNTC[O6:N02FJa0`0]OBl1j<kNTC\\O68N11GM>>_OBl1j<jNUC_O36001JL<?_OBl1j<jNTCA43001ML8`0AAl0OKk<NUCC31001OL7`0AAj02Lh<OTCE4N0003L4a0AAi04Lf<OUCG3L0005M1a0C@g06Nc<0UCI2I2O08LOb0C@e09O`<0VCK1G3NO<LLb0D@c0;0_<1TCM2D4MN?LJc0E_Oa0>0i;_OSDc0EO0B7LLc0KHd0E_O?`02c;CWD?E1OZO=1Fg0LDd0F_O=c02];IZD:E4j01TOBd0H_O:f03V;N^D6E5h02UO@e0H@7g05Q;3`D2G6f01WO_Od0I@5j05k:9cDNG6f02VO^Oe0I@3l07e:>fDHI7f02UO\\Of0K_O1o07`:c0hDEI7f02UOZOh0M]O0P18\\:f0kDAI9g01SOYOj0N\\ONS18W:k0mD]OJ9g02SOVOk01g06j9n0UETOF?g02QOUOg26c8T1_EhN@g0n0Jd1B]8X1PF[O:]O^1OW8Z1PFZOf3ZO[6\\1oElNT4Hl5]1jKaNW4_1iK`NW4b1S70O100j@hNc;Y1fA]NU2`0P<S1kAbN4b0>L\\ONV=R1kBhN0=a0K_ONT=S1lBlNK9d0K@NU=R1lBnNI7e0LAMT=T1lBlNK5f0M_ONT=T1lBmNJ4h0L]O0T=T1mBlNJ4h0L]O0T=T1mBlNJ4i0K\\O1T=T1mBmNJ2j0K[O2S=U1nBlNJ1l0KXO4T=T1nBmNIOo0KVO5S=U1oBlNINc0@@;26S=V1oBjNIN5AL42538Q=W1PCjNIJ6EH7502:S=U1oBkNII6GG86M2;T=T1nBmNIF6JG77L1<T=T1nBmNIE6LG77J1=U=T1mBlNJD5OG77H0?W=R1lBnNIB61G77F0`0X=Q1kBoNJ@54G77D0a05ZO]1OY9g1XEQOI^O57H67B0a05\\O\\10Z9d1XEROJ\\O4:I56A0b05\\O[12[9a1XETOJYO4=I5d02G\\OI3S10k9`1WESOJXO5`0H4e01F]OI5Q1Om9^1WEUO5HB4d00G]OI8m0MQ:]1WEUO4IC4c0OH]OI:j0MT:Z1WEWO2IE4c0MH^OI<g0MW:W1XEWO0KF4b0LI^OI>d0L[:V1VEXOOKH4b0JI_OI`0b0K]:T1VEYOOKH5a0IJ_OIb0?Ja:R1UE[O0HG7a0IJ_OId0<Jd:o0VE[O1GE8c0IH@If0:Hg:o0TEZO5EC:d0IFAIi06Gj:m0UEYO8D@9g0LCAIm02CR;i0QE]O9C_O87^O5?NBIT1J]O`;c0kD@KZO89E79C3<NCIk1^;TOgDDK]O95D7:G1:OCIk1d;nNaDIJA;0C6=KN8ODIl1g;jN]DKKD<LB6>OM5OFHk1j;hN\\DLIG>HA5`04K2OGHk1l;gNYDMIIe0IL8I00GHl1m;dNXDNJKc0E0<FN0HHl1o;bNWDOIL]13lNK0IHl1P<bNUDNKM0ER1`0VOH0JHm1Q<_NUDOJO0Fn0a0YOD1KHm1S<]NSD0K0OJi0?^OA1LIl1T<]NQD0K1OM>H@e09_O1MIm1U<ZNPD1L3M01ON:9^O2MHm1X<YNnC2K4NO00N::]O3L6^1j;jNkC1M5MO01N8<[OV1P1o:UOhC2L7LM22M7>[O\\1h0i:^OeC2M7LM14N4e2KQ:FcC3K7NMO4O4m2Al9O`C4K8ML05O2R`0AW@4I9OK06N0U`0AU@5I:NJ16OOU`0AU@5H;OI08NNW`0AS@5H<0H08OMX`0@R@5H?NG19NKj`0FY_O?OF0:OJj`0FX_O`0OF1:MJk`0FW_Ob0OD2;MHl`0EW_Od0NE1:NGm`0FV_Od0OD1;MFn`0FT_Og0OB2;MFn`0EU_Oh0NB2<MBXa0`0i^OB3;LAZa0c0f^OA4;M_OYa0Lf^Oe01D3<M]Ooa04Q^O33<N[ORb03m]O64;Yb0\\Oc]O3;`0nb0N1O1O2O0O1O2N1O2N1O1O2N1O1Nn\\o1"}, "label": "a girl with dark hair sitting to the right of an elderly woman"}]}
{"id": 267851, "image": "COCO_train2014_000000267851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman shielding her eyes on a bench next to two other women\"."}], "task": "refering", "answer_template": "The \"a woman shielding her eyes on a bench next to two other women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 302, 303, 304, 305, 327, 328, 352, 353, 354, 375, 376, 377, 398, 399, 400], "bbox": [0.1460625, 0.2661875, 0.432125, 0.75921875], "iscrowd": 0, "area": 19513.73055, "rle": {"size": [640, 640], "counts": "gYj11kc0401\\\\OK[c05c\\O0a2No=3]_O5W20\\NFZ?f0mAMiNDT?b0iA1SO^OS?c0fA0XOHg>;kAO_OJb>:jAN@PO@k0S?<eAKHoNAl0o>?bAFNPOAn0f0\\Oo<Q1jBF0PO@P1?_OV=k0iBG2PO@R18A\\=g0hBF4RO@R13B`=e0gBF5ROAR13A_=e0fBG7SO@P14B]=f0fBE9TO@o04C\\=f0eBE:UOAl04D\\=g0bBE=UOAk04D\\=S2oB_MAj04E[=R2PC_MAk03D\\=Q2PCaMBi03EZ=P2QCcMBh03EZ=P2PCdMCg03FX=o1RCeMCf03FW=o1SCfMCe03GV=n1SCgMDe02FV=n1TChMEc02GS=n1VCiMEb02GR=o1WChMEb02HQ=m1XCjMEa02HP=m1YCkME`02HP=m1XClMF`01GQ=l1XCnMF?1GQ=k1XCPNF>2Fo<m1YCnMG?1Fo<l1ZCoMF?1Fo<k1[CPND`01FP=i1\\CQNC`01FP=i1]CoMBc01Eo<i1_CoMAb01Go<g1`CPN@b01GP=f1_CPN@d01Go<e1aCoM_Od02Hn<d1ZCZMIf0Le02GP=d1XC\\MId0ME1`017P=c1YC]MHc0NF1>19o<c1XC_MI`0MG5<N;o<c1XC_MI`0MG5<M<P=a1YC`MH?MI5:N=o<a1XCaMI>MI5:N=NmNl<c2^CcMH<MK49O<OnNk<c2]CdMI:MM47O>NnNm<a2\\CeMI:MM47O>NnNm<`2\\ChMH7N0341>MPOn<^2[CiMH7N0242?MoNo<\\2ZClMI5M1241`0MoNQ=Z2YCnMH3N3222`0MPOQ=X2XCPNI3M3212b0LPOS=2lB\\1OkN<NI2N4102d0LoNT=OnB\\1NnN:NI2M52O1d0NoNT=LPC]1LoN90I1M52N1f0MoN`=U1cBQO71I0M72L2f0MPO`=S1cBRO63IOM71L3g0LPOb=P1cBTO5ODI4;N61L2g0NPOb=n0cBVO3MIJ0;N81J2i0MPOe=i0dBXO1OIIO<N81I2;KE3Je=g0dBZOO0O4I80I290DMNg=d0dB[OO0O5H80H2:3AK1h=0UB8>FM205G35KN<3_OK3j=MUB8<HM2O6IL;NG?3@K2k=LVB6;JM3N6IFi0LXOh03_OL3o=GTB89LL6N2JER1e0QOAL3[>LQBNN9KNLDU1e0nNCL4]>FQB4K:LK=F4`0_OGK5P?JYA:LI9K:9]OJL5P?HYA=KG50LK58JLL6\\=DQD3kN?LD33NM31LOL7\\=EPDOmNb0K@38NM3LL2L8\\=EeB7i0F_Od0K\\O3=NN2EN6L8\\=FaBa0b0[OIe06JFN1AO8L9\\=a0RCnNKh02KIM9HC:\\=a0`CE^OLJM7GE;]=a0\\CF_OKLM4HH9^=a0ZCG^OKNM2GJ:^=b0XCF_OJON1EK<_=b0UCEBIN01BK?`=a0SCDEIK13AJa0a=a0PCCHII06@Hd0b=a0lBBMHG08@Ef0d=`0jBA0HE09@Dh0e=`0gB@HVO2a0M0;_OBk0f=`0dB^OJYO1?NO`0MAId=T1aB^OK\\O1<NO4B072Na=S1^B_OK^O29NO4EO34Mg=R1WBCL_O17NO4HN0d>k0_AGLB14NO4JOMm>d0UA0KD20O03MNIR?d0QA1LF1NO12O]?9e@2LH2LN121_?5d@5JJ2JO024_?2d@6JK2EHL7726X3H`87ZD9JL3A8OH?Q3Lj8NXD;JM3^O:OFb0P3Ll8LWD<JNP1NR2Nn8HWD=I1n0JU22n8EVD>I2l0GY25m8CTD`0J2j0FQ1DUN5a0`0a;^OTDa0J3h0ET1a1W:UNSDb0J3h0DU1a1X:TNQDd0K3h0BU1c1X:SNPDe0J3j0AU1d1e:hN\\C2k0AU1e1d:hN\\C2k0AU1e1c:iN]C1l0@T1f1c:iN]C1l0@T1f1c:hN]C3l0_OT1f1b:iN^CI\\1Ad0l1c:jN]C@m3f1f8UN]He1R=M3Mh]OgNoa0W1Q^OmNf<:\\CKd3k0ZLTOe<6]C@L;`3j0aLWOg<3]C_OL=_3i0`LZOj<O[C@N=]3LoL0A:j<MZC_O0=]3KVMKYOb0l<JXC_O2?[3I^M`0n;H`CMU3H_Md0m;F_CNU3FaMf0m;D`CLS3G^Mn0P<_O`CIS3KWMR1Y<WO`CHQ31hL\\1k<gNUGk1`=00000000N200001O00010O00001O00SOW^OCia08\\^OIca02c^OM]a0Nh^O2Xa0Jm^O4Ta0HP_O8Pa0FS_O9m`0EV_O:j`0DX_O;i`0CZ_O<f`0B\\_O>d`0@__O?Qb000O101O000O101O00000O101O0O100O2O0O101N5LT]R7"}, "label": "a woman shielding her eyes on a bench next to two other women"}]}
{"id": 267851, "image": "COCO_train2014_000000267851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a dark dress wearing sunglasses sitting on a bench\"."}], "task": "refering", "answer_template": "The \"a woman in a dark dress wearing sunglasses sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 302, 303, 304, 305, 327, 328, 352, 353, 354, 375, 376, 377, 398, 399, 400], "bbox": [0.1460625, 0.2661875, 0.432125, 0.75921875], "iscrowd": 0, "area": 19513.73055, "rle": {"size": [640, 640], "counts": "gYj11kc0401\\\\OK[c05c\\O0a2No=3]_O5W20\\NFZ?f0mAMiNDT?b0iA1SO^OS?c0fA0XOHg>;kAO_OJb>:jAN@PO@k0S?<eAKHoNAl0o>?bAFNPOAn0f0\\Oo<Q1jBF0PO@P1?_OV=k0iBG2PO@R18A\\=g0hBF4RO@R13B`=e0gBF5ROAR13A_=e0fBG7SO@P14B]=f0fBE9TO@o04C\\=f0eBE:UOAl04D\\=g0bBE=UOAk04D\\=S2oB_MAj04E[=R2PC_MAk03D\\=Q2PCaMBi03EZ=P2QCcMBh03EZ=P2PCdMCg03FX=o1RCeMCf03FW=o1SCfMCe03GV=n1SCgMDe02FV=n1TChMEc02GS=n1VCiMEb02GR=o1WChMEb02HQ=m1XCjMEa02HP=m1YCkME`02HP=m1XClMF`01GQ=l1XCnMF?1GQ=k1XCPNF>2Fo<m1YCnMG?1Fo<l1ZCoMF?1Fo<k1[CPND`01FP=i1\\CQNC`01FP=i1]CoMBc01Eo<i1_CoMAb01Go<g1`CPN@b01GP=f1_CPN@d01Go<e1aCoM_Od02Hn<d1ZCZMIf0Le02GP=d1XC\\MId0ME1`017P=c1YC]MHc0NF1>19o<c1XC_MI`0MG5<N;o<c1XC_MI`0MG5<M<P=a1YC`MH?MI5:N=o<a1XCaMI>MI5:N=NmNl<c2^CcMH<MK49O<OnNk<c2]CdMI:MM47O>NnNm<a2\\CeMI:MM47O>NnNm<`2\\ChMH7N0341>MPOn<^2[CiMH7N0242?MoNo<\\2ZClMI5M1241`0MoNQ=Z2YCnMH3N3222`0MPOQ=X2XCPNI3M3212b0LPOS=2lB\\1OkN<NI2N4102d0LoNT=OnB\\1NnN:NI2M52O1d0NoNT=LPC]1LoN90I1M52N1f0MoN`=U1cBQO71I0M72L2f0MPO`=S1cBRO63IOM71L3g0LPOb=P1cBTO5ODI4;N61L2g0NPOb=n0cBVO3MIJ0;N81J2i0MPOe=i0dBXO1OIIO<N81I2;KE3Je=g0dBZOO0O4I80I290DMNg=d0dB[OO0O5H80H2:3AK1h=0UB8>FM205G35KN<3_OK3j=MUB8<HM2O6IL;NG?3@K2k=LVB6;JM3N6IFi0LXOh03_OL3o=GTB89LL6N2JER1e0QOAL3[>LQBNN9KNLDU1e0nNCL4]>FQB4K:LK=F4`0_OGK5P?JYA:LI9K:9]OJL5P?HYA=KG50LK58JLL6\\=DQD3kN?LD33NM31LOL7\\=EPDOmNb0K@38NM3LL2L8\\=EeB7i0F_Od0K\\O3=NN2EN6L8\\=FaBa0b0[OIe06JFN1AO8L9\\=a0RCnNKh02KIM9HC:\\=a0`CE^OLJM7GE;]=a0\\CF_OKLM4HH9^=a0ZCG^OKNM2GJ:^=b0XCF_OJON1EK<_=b0UCEBIN01BK?`=a0SCDEIK13AJa0a=a0PCCHII06@Hd0b=a0lBBMHG08@Ef0d=`0jBA0HE09@Dh0e=`0gB@HVO2a0M0;_OBk0f=`0dB^OJYO1?NO`0MAId=T1aB^OK\\O1<NO4B072Na=S1^B_OK^O29NO4EO34Mg=R1WBCL_O17NO4HN0d>k0_AGLB14NO4JOMm>d0UA0KD20O03MNIR?d0QA1LF1NO12O]?9e@2LH2LN121_?5d@5JJ2JO024_?2d@6JK2EHL7726X3H`87ZD9JL3A8OH?Q3Lj8NXD;JM3^O:OFb0P3Ll8LWD<JNP1NR2Nn8HWD=I1n0JU22n8EVD>I2l0GY25m8CTD`0J2j0FQ1DUN5a0`0a;^OTDa0J3h0ET1a1W:UNSDb0J3h0DU1a1X:TNQDd0K3h0BU1c1X:SNPDe0J3j0AU1d1e:hN\\C2k0AU1e1d:hN\\C2k0AU1e1c:iN]C1l0@T1f1c:iN]C1l0@T1f1c:hN]C3l0_OT1f1b:iN^CI\\1Ad0l1c:jN]C@m3f1f8UN]He1R=M3Mh]OgNoa0W1Q^OmNf<:\\CKd3k0ZLTOe<6]C@L;`3j0aLWOg<3]C_OL=_3i0`LZOj<O[C@N=]3LoL0A:j<MZC_O0=]3KVMKYOb0l<JXC_O2?[3I^M`0n;H`CMU3H_Md0m;F_CNU3FaMf0m;D`CLS3G^Mn0P<_O`CIS3KWMR1Y<WO`CHQ31hL\\1k<gNUGk1`=00000000N200001O00010O00001O00SOW^OCia08\\^OIca02c^OM]a0Nh^O2Xa0Jm^O4Ta0HP_O8Pa0FS_O9m`0EV_O:j`0DX_O;i`0CZ_O<f`0B\\_O>d`0@__O?Qb000O101O000O101O00000O101O0O100O2O0O101N5LT]R7"}, "label": "a woman in a dark dress wearing sunglasses sitting on a bench"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass cup with water\"."}], "task": "refering", "answer_template": "The \"the glass cup with water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 133, 134, 135], "bbox": [0.7339375, 0.0057083333333333335, 0.9828749999999999, 0.3333333333333333], "iscrowd": 0, "area": 17928.917649999996, "rle": {"size": [480, 640], "counts": "g]l67_>;E;E:F;E;E:F;E:F;E;E:K6N2O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O0010O01O1O001O1O001O001O1O001O1O0000O10000O10000O10000O10000O10000O1000000O2O000O10000O10000O10000O10000O10000O1000000O10000O10000M3N2M3N2N2M3N2M3N3L3N2M3N2N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2N2M3N2M3N3L3N2M3N2N2M3N2M3N2M3NkT5"}, "label": "the glass cup with water"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass far right\"."}], "task": "refering", "answer_template": "The \"glass far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 133, 134, 135], "bbox": [0.7339375, 0.0057083333333333335, 0.9828749999999999, 0.3333333333333333], "iscrowd": 0, "area": 17928.917649999996, "rle": {"size": [480, 640], "counts": "g]l67_>;E;E:F;E;E:F;E:F;E;E:K6N2O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O0010O01O1O001O1O001O001O1O001O1O0000O10000O10000O10000O10000O10000O1000000O2O000O10000O10000O10000O10000O10000O1000000O10000O10000M3N2M3N2N2M3N2M3N3L3N2M3N2N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2N2M3N2M3N3L3N2M3N2N2M3N2M3N2M3NkT5"}, "label": "glass far right"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich that is not cut\"."}], "task": "refering", "answer_template": "The \"sandwich that is not cut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0, 0.3940208333333333, 0.586359375, 0.9862291666666666], "iscrowd": 0, "area": 81848.49995, "rle": {"size": [480, 640], "counts": "Y8m0Q>h1WNd1]N3L4M3M3O0O2O1N2O1N101O1N2O1N101N2O1N2O0O2O1N2O1O0O2O1N2O1N101N2O1N2O0O2O1O1N2O0O2O1N2O1N101N2O1N2O001N2O1N2O0O2O1N2O1N101N2O1O1N101N2O1N2O0O2O1N2O1N101O1N101N100O2O0O2O0O101N101O0O2O0O101N101N100O2O0O2O001N100O2O0O2O0O101N100O10000O100O100O100O100O100O10000O100O100O100O100O10000000000001O000000000000000000000000000000001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O000000000000000000000000001O000000000000000000000000000000000000001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O0O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O001O001O002N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2L3N3M3L3N3L4M3M3L4M3L4M2N3L4POP1L4M3L4L3N3L4L4M`0_OWmk3"}, "label": "sandwich that is not cut"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bread inn the left side of the plate\"."}], "task": "refering", "answer_template": "The \"bread inn the left side of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0, 0.3940208333333333, 0.586359375, 0.9862291666666666], "iscrowd": 0, "area": 81848.49995, "rle": {"size": [480, 640], "counts": "Y8m0Q>h1WNd1]N3L4M3M3O0O2O1N2O1N101O1N2O1N101N2O1N2O0O2O1N2O1O0O2O1N2O1N101N2O1N2O0O2O1O1N2O0O2O1N2O1N101N2O1N2O001N2O1N2O0O2O1N2O1N101N2O1O1N101N2O1N2O0O2O1N2O1N101O1N101N100O2O0O2O0O101N101O0O2O0O101N101N100O2O0O2O001N100O2O0O2O0O101N100O10000O100O100O100O100O100O10000O100O100O100O100O10000000000001O000000000000000000000000000000001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O000000000000000000000000001O000000000000000000000000000000000000001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O0O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O001O001O002N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2L3N3M3L3N3L4M3M3L4M3L4M2N3L4POP1L4M3L4L3N3L4L4M`0_OWmk3"}, "label": "bread inn the left side of the plate"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden handled knife\"."}], "task": "refering", "answer_template": "The \"wooden handled knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 242, 243, 244, 245, 266, 267, 268, 269, 270, 290, 291, 292, 293, 294, 295, 315, 316, 317, 318, 319, 320, 340, 341, 342, 343, 344, 365, 366, 367, 389, 390], "bbox": [0.5309062499999999, 0.5528125, 1.0, 0.9797708333333334], "iscrowd": 0, "area": 16968.099400000003, "rle": {"size": [480, 640], "counts": "`do42m>1O2N2M3N2N200O010O10O010O01O10O010O0100O0010O0100O010O010O3M6K5J6K4K6K0O1O100O100O100O1O100O100O100O2O0O1O100O010O010O0010O010O010O01O010O01O010O010O01O010O010N2N1O2N1O2N1O2N1100O100O100O100O100O100O100O100O100O100O10000O010O010O10O010O0100O010O010O010O10O010O0100O010O010O01O10O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O10O010O010O010O010O010O010O010O010O010O010O010O010O010O010O10O10O010O010O010O010O010O010O010O010O010O010O010O010O010O01000O010O010O010O010O010O010O010O010O010O010O010O010O010O01000O010O010O010O010O010O010O010O010O10O010O010O010O010O010O10O10O010O010O010O010ON3WOh0YOoC"}, "label": "wooden handled knife"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a knife with wooden handle\"."}], "task": "refering", "answer_template": "The \"a knife with wooden handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 242, 243, 244, 245, 266, 267, 268, 269, 270, 290, 291, 292, 293, 294, 295, 315, 316, 317, 318, 319, 320, 340, 341, 342, 343, 344, 365, 366, 367, 389, 390], "bbox": [0.5309062499999999, 0.5528125, 1.0, 0.9797708333333334], "iscrowd": 0, "area": 16968.099400000003, "rle": {"size": [480, 640], "counts": "`do42m>1O2N2M3N2N200O010O10O010O01O10O010O0100O0010O0100O010O010O3M6K5J6K4K6K0O1O100O100O100O1O100O100O100O2O0O1O100O010O010O0010O010O010O01O010O01O010O010O01O010O010N2N1O2N1O2N1O2N1100O100O100O100O100O100O100O100O100O100O10000O010O010O10O010O0100O010O010O010O10O010O0100O010O010O01O10O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O10O010O010O010O010O010O010O010O010O010O010O010O010O010O010O10O10O010O010O010O010O010O010O010O010O010O010O010O010O010O01000O010O010O010O010O010O010O010O010O010O010O010O010O010O01000O010O010O010O010O010O010O010O010O10O010O010O010O010O010O10O10O010O010O010O010ON3WOh0YOoC"}, "label": "a knife with wooden handle"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one half of a sandwich leaning against a whole sandwich\"."}], "task": "refering", "answer_template": "The \"one half of a sandwich leaning against a whole sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 217, 218, 219], "bbox": [0.136515625, 0.1460625, 0.67246875, 0.5707916666666667], "iscrowd": 0, "area": 48423.44145, "rle": {"size": [480, 640], "counts": "_mX11i><D<E;D<D<E;D8H6K6I6J6M3O100000001O000000000000001N100000000000001O0000000O1000001O000000000000001O0O1000000000001O000000000O10001O000000000000001O000O10001O1O1O001O1O001O1O1N101O1O1O001O1O1O00100O1O001O1O1O001O1O1O001O1O1O001O1O1O001O10O01O1O1O001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O00001O00001O00001O00001O000010O000001O00001O00001O00001O00001O0000001O01O01O00001O00001O00001O00001O0000001O00010O00001O000010O0001O01O01O0000010O00001O01O01O001O010O001O010O001O0010O01O001OO2O0O1O100O2N100O1O2O0O1O100O2N1O100O1O2O0O1O100O2N100O1O101N1O1O101N1O100O1O2O0O1O100O2N100O2N2N101N1O2N1O2N1O2O0O2N1O2N1O2N2O0O2N1O2N1O2N101N1O2N1O2N2N101M2N3M2N3M2N3M2N3M2N3N1N3M3M2I8G8H9G8I8I6L5L3L5L3LYZR3"}, "label": "one half of a sandwich leaning against a whole sandwich"}]}
{"id": 398924, "image": "COCO_train2014_000000398924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a meat sandwich laying on top of the other sandwich\"."}], "task": "refering", "answer_template": "The \"half of a meat sandwich laying on top of the other sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 217, 218, 219], "bbox": [0.136515625, 0.1460625, 0.67246875, 0.5707916666666667], "iscrowd": 0, "area": 48423.44145, "rle": {"size": [480, 640], "counts": "_mX11i><D<E;D<D<E;D8H6K6I6J6M3O100000001O000000000000001N100000000000001O0000000O1000001O000000000000001O0O1000000000001O000000000O10001O000000000000001O000O10001O1O1O001O1O001O1O1N101O1O1O001O1O1O00100O1O001O1O1O001O1O1O001O1O1O001O1O1O001O10O01O1O1O001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O00001O00001O00001O00001O000010O000001O00001O00001O00001O00001O0000001O01O01O00001O00001O00001O00001O0000001O00010O00001O000010O0001O01O01O0000010O00001O01O01O001O010O001O010O001O0010O01O001OO2O0O1O100O2N100O1O2O0O1O100O2N1O100O1O2O0O1O100O2N100O1O101N1O1O101N1O100O1O2O0O1O100O2N100O2N2N101N1O2N1O2N1O2O0O2N1O2N1O2N2O0O2N1O2N1O2N101N1O2N1O2N2N101M2N3M2N3M2N3M2N3M2N3N1N3M3M2I8G8H9G8I8I6L5L3L5L3LYZR3"}, "label": "half of a meat sandwich laying on top of the other sandwich"}]}
{"id": 120400, "image": "COCO_train2014_000000120400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back top of a white empty chair\"."}], "task": "refering", "answer_template": "The \"the back top of a white empty chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0, 0.7963333333333333, 0.353984375, 0.9881458333333334], "iscrowd": 0, "area": 18096.881450000004, "rle": {"size": [480, 640], "counts": "n;l2T<0000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O0000000000000000000000001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O2N2N3M2N3M2N3M2N3M2N3K4M4L3L5L3O2M2O]cQ6"}, "label": "the back top of a white empty chair"}]}
{"id": 120400, "image": "COCO_train2014_000000120400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair\"."}], "task": "refering", "answer_template": "The \"a white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0, 0.7963333333333333, 0.353984375, 0.9881458333333334], "iscrowd": 0, "area": 18096.881450000004, "rle": {"size": [480, 640], "counts": "n;l2T<0000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O0000000000000000000000001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O2N2N3M2N3M2N3M2N3M2N3K4M4L3L5L3O2M2O]cQ6"}, "label": "a white chair"}]}
{"id": 439889, "image": "COCO_train2014_000000439889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the blue wind breakers\"."}], "task": "refering", "answer_template": "The \"the man in the blue wind breakers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 23, 24, 25, 41, 42, 43, 44, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 113, 114, 115, 116, 117, 118, 119, 120, 133, 134, 135, 136, 137, 138, 149, 150, 151, 152, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 203, 204, 205, 206, 207, 221, 222, 223, 224, 225, 226, 239, 240, 242, 243, 244, 257, 258, 260, 261, 262, 275, 276, 278, 279, 280, 293, 294, 297, 298, 310, 311, 315, 316, 317, 328, 329, 333, 334, 335, 336, 347, 348, 365, 366], "bbox": [0.25296442687747034, 0.00225, 0.6963636363636364, 0.8809062500000001], "iscrowd": 0, "area": 52408.93795, "rle": {"size": [640, 506], "counts": "Uh`23mc01O2N1O1O2TGJiM8S2NjM3R24jMMR29kMHQ2>lMCP2e0jM]OR2j0jMWOoK@S2`1nM^NMc00Dn1c1PN\\NL>4Ih1k1mMSN2:6Nd1n1nMPN24;3^1o1RNPNOO?7Y1P2VNmMNN`0`0P1k1_NhMOO`0i0=m1RO\\M0O?S1WO`2:_LVNZO=f0i1W1POa2c0YLTNB8?n0OlM[1n1a2l0ULRNI29U1LlMa1h1`2R1SLPNK27W1^2]Ob1Y1oKnMO23Y1`2WOc1`1lKlM112[1a2QOd1g1iKiM50O^1c2iNf1o1dKeM;1Ja1c2dNh1U2aK^Md03Be1c2_Nj1[2bM^MRNg1d2[Nl1_2bMZMPNk1d2VNn1e2QOTOW5j0kJUOU5k0kJUOU5k0lJTOT3]N`I^2]3TOP3kNYIQ2g3TOm2ROXIj1l3SOi2YOYIc1o3SOe2@YI]1R4SOb2FZIV1U4SO^2NYIn0[4SOY25ZIg0]4TOV2;[I`0`4TOR2c0YI:f4ROn1j0YI4i4ROk1S4VNlKg1W4ZNhKc1[4]NfK_1]4bNbK[1`4gN_KT1f4lNZKn0l4SOSKg0S5ZOlJ`0Z5@fJ;_5F`J4f5MZJLl54TJFR6;mI@W6b0hIXO^6h0bISOc6n0\\InNh6S1WIhNn6Y1QIdNR7]1nH`NT7b1jH[NY7g1eHVN]7l1bHRNa7o1]HnMi7P2VHmMQ8Q2mGmMY8P2fGmMa8Q2^GkMi8VMdFf3b0ROo8mLhFP49QOU9hLgFU44TOX9aLgFZ41VOW:i0iEWOY:g0gEZOW:g0jEYOU:g0kEZOT:f0lE[OS:e0mE[OS:e0mE\\OR:d0nE]OQ:d0nE]OQ:c0oE]OQ:c0PF]OP:b0PF_Oo9a0QF@n9`0RF@n9`0RFAm9?SFBl9>TFCk9=UFEi9;XFEh9:XFHf98ZFJe96ZFKf94ZFNe91[F1d9N\\F3c9M^F4a9K_F7`9H`F9`9F`F<^9DbF>]9AcF`0]9_OcFc0\\9\\OdFf0[9YOfFi0W9XOhFk0V9TOjFo0T9POlFS1R9lNnFW1o8iNQGZ1m8eNSG^1k8kLUGaN0g4i8cL[GfNMi4e8^LaGiNJl4c8WLfGmNGP5`8TLhGkNJR5\\8SLiGkNMS5X8RLjGkN0T5T8RLjGkN3T5Q8QLkGkN6U5m7PLlGkN=R5e7SLmGkNb0Q5_7ULnGiNg0Q5Y7VLoGiNl0P5R7XLQHhNQ1o4l6XLSHiNT1o4g6VLVHkNW1n4a6VLXHlNZ1n4\\6TL[HnN]1n4U6SL_HnN_1o4P6QLbHPOb1n4aN\\JU6e1hIROd1m4ZN_JZ6`1iITOg1l4QNcJ_6\\1iIUOj1l4hMgJe6W1iIVOm1l4`MiJk6S1hIYOP2k4XMlJP7o0hIZOR2l4QMnJU7j0iI\\OT2`5T4SKhI]OW2^5Q4SKiI_OY2\\5m3SKkIB[2Y5j3nJQJIX2V5g3QKQJI[2T5c3SKRJI]2S5a3SKQJKa2P5]3UKRJKd2m4Z3YKQJJh2k4V3[KRJIl2j4R3\\KQJKP3f4o2`KPJJT3d4k2bKQJJV3c4i2bKQJKY3a4e2dKRJJ]3_4a2hKPJJb3\\4]2jKQJJe3Z4Z2kKQJKh3W4W2oKPJJl3U4S2QLPJKo3R4R2RLoILR4n3P2VLnIKV4l3l1ZLmIJZ4i3j1\\LlIK]4f3g1_LlIK`4c3d1cLkIJj4Z3[1lLjIKb5b2e0bMiILZ6i1N[NhILg6\\1AiNgIKk6Y1_OkNfILm6W1]OmNfILP7T1ZOQOeIKT7Q1WOTOeIKV7o0VOUOcIMZ7j0TOYObIM\\7h0RO\\OaIL`7e0PO^O`IMc7b0mNA`IMe7`0kND_ILg7?jNE_ILi7=iNF^IMk7;gNH]INm78gNKZINP86gNKXI0R84fNLWI1S83fNLVI2U81eNNUI1W80eNNTI2X8OdNOTI2X8NeN0SI2Y8MeN0QI4[8KdN1QI4\\8JcN2QI4]8IbN2RI5\\8IcN1QI6]8HbN2QI6^8FbN4PI5`8FaN4oH6`8FaN3PI7`8E`N4QI6`8E_N5QI6a8D_N5QI6`8E_N5QIYOCa0n8O_N7^IHT80_N6^IJT8O^N7^IJT8OaN4\\ILT8OdN1XI0U8NgNNTI3W8NiNJQI7W8NmNGmH6[815F58k9N101N2N2NPho2"}, "label": "the man in the blue wind breakers"}]}
{"id": 439889, "image": "COCO_train2014_000000439889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the guy on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 23, 24, 25, 41, 42, 43, 44, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 113, 114, 115, 116, 117, 118, 119, 120, 133, 134, 135, 136, 137, 138, 149, 150, 151, 152, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 203, 204, 205, 206, 207, 221, 222, 223, 224, 225, 226, 239, 240, 242, 243, 244, 257, 258, 260, 261, 262, 275, 276, 278, 279, 280, 293, 294, 297, 298, 310, 311, 315, 316, 317, 328, 329, 333, 334, 335, 336, 347, 348, 365, 366], "bbox": [0.25296442687747034, 0.00225, 0.6963636363636364, 0.8809062500000001], "iscrowd": 0, "area": 52408.93795, "rle": {"size": [640, 506], "counts": "Uh`23mc01O2N1O1O2TGJiM8S2NjM3R24jMMR29kMHQ2>lMCP2e0jM]OR2j0jMWOoK@S2`1nM^NMc00Dn1c1PN\\NL>4Ih1k1mMSN2:6Nd1n1nMPN24;3^1o1RNPNOO?7Y1P2VNmMNN`0`0P1k1_NhMOO`0i0=m1RO\\M0O?S1WO`2:_LVNZO=f0i1W1POa2c0YLTNB8?n0OlM[1n1a2l0ULRNI29U1LlMa1h1`2R1SLPNK27W1^2]Ob1Y1oKnMO23Y1`2WOc1`1lKlM112[1a2QOd1g1iKiM50O^1c2iNf1o1dKeM;1Ja1c2dNh1U2aK^Md03Be1c2_Nj1[2bM^MRNg1d2[Nl1_2bMZMPNk1d2VNn1e2QOTOW5j0kJUOU5k0kJUOU5k0lJTOT3]N`I^2]3TOP3kNYIQ2g3TOm2ROXIj1l3SOi2YOYIc1o3SOe2@YI]1R4SOb2FZIV1U4SO^2NYIn0[4SOY25ZIg0]4TOV2;[I`0`4TOR2c0YI:f4ROn1j0YI4i4ROk1S4VNlKg1W4ZNhKc1[4]NfK_1]4bNbK[1`4gN_KT1f4lNZKn0l4SOSKg0S5ZOlJ`0Z5@fJ;_5F`J4f5MZJLl54TJFR6;mI@W6b0hIXO^6h0bISOc6n0\\InNh6S1WIhNn6Y1QIdNR7]1nH`NT7b1jH[NY7g1eHVN]7l1bHRNa7o1]HnMi7P2VHmMQ8Q2mGmMY8P2fGmMa8Q2^GkMi8VMdFf3b0ROo8mLhFP49QOU9hLgFU44TOX9aLgFZ41VOW:i0iEWOY:g0gEZOW:g0jEYOU:g0kEZOT:f0lE[OS:e0mE[OS:e0mE\\OR:d0nE]OQ:d0nE]OQ:c0oE]OQ:c0PF]OP:b0PF_Oo9a0QF@n9`0RF@n9`0RFAm9?SFBl9>TFCk9=UFEi9;XFEh9:XFHf98ZFJe96ZFKf94ZFNe91[F1d9N\\F3c9M^F4a9K_F7`9H`F9`9F`F<^9DbF>]9AcF`0]9_OcFc0\\9\\OdFf0[9YOfFi0W9XOhFk0V9TOjFo0T9POlFS1R9lNnFW1o8iNQGZ1m8eNSG^1k8kLUGaN0g4i8cL[GfNMi4e8^LaGiNJl4c8WLfGmNGP5`8TLhGkNJR5\\8SLiGkNMS5X8RLjGkN0T5T8RLjGkN3T5Q8QLkGkN6U5m7PLlGkN=R5e7SLmGkNb0Q5_7ULnGiNg0Q5Y7VLoGiNl0P5R7XLQHhNQ1o4l6XLSHiNT1o4g6VLVHkNW1n4a6VLXHlNZ1n4\\6TL[HnN]1n4U6SL_HnN_1o4P6QLbHPOb1n4aN\\JU6e1hIROd1m4ZN_JZ6`1iITOg1l4QNcJ_6\\1iIUOj1l4hMgJe6W1iIVOm1l4`MiJk6S1hIYOP2k4XMlJP7o0hIZOR2l4QMnJU7j0iI\\OT2`5T4SKhI]OW2^5Q4SKiI_OY2\\5m3SKkIB[2Y5j3nJQJIX2V5g3QKQJI[2T5c3SKRJI]2S5a3SKQJKa2P5]3UKRJKd2m4Z3YKQJJh2k4V3[KRJIl2j4R3\\KQJKP3f4o2`KPJJT3d4k2bKQJJV3c4i2bKQJKY3a4e2dKRJJ]3_4a2hKPJJb3\\4]2jKQJJe3Z4Z2kKQJKh3W4W2oKPJJl3U4S2QLPJKo3R4R2RLoILR4n3P2VLnIKV4l3l1ZLmIJZ4i3j1\\LlIK]4f3g1_LlIK`4c3d1cLkIJj4Z3[1lLjIKb5b2e0bMiILZ6i1N[NhILg6\\1AiNgIKk6Y1_OkNfILm6W1]OmNfILP7T1ZOQOeIKT7Q1WOTOeIKV7o0VOUOcIMZ7j0TOYObIM\\7h0RO\\OaIL`7e0PO^O`IMc7b0mNA`IMe7`0kND_ILg7?jNE_ILi7=iNF^IMk7;gNH]INm78gNKZINP86gNKXI0R84fNLWI1S83fNLVI2U81eNNUI1W80eNNTI2X8OdNOTI2X8NeN0SI2Y8MeN0QI4[8KdN1QI4\\8JcN2QI4]8IbN2RI5\\8IcN1QI6]8HbN2QI6^8FbN4PI5`8FaN4oH6`8FaN3PI7`8E`N4QI6`8E_N5QI6a8D_N5QI6`8E_N5QIYOCa0n8O_N7^IHT80_N6^IJT8O^N7^IJT8OaN4\\ILT8OdN1XI0U8NgNNTI3W8NiNJQI7W8NmNGmH6[815F58k9N101N2N2NPho2"}, "label": "the guy on the right in the right hand picture"}]}
{"id": 439889, "image": "COCO_train2014_000000439889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a pink skirt\"."}], "task": "refering", "answer_template": "The \"a girl wearing a pink skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 38, 39, 40, 41, 55, 56, 57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 272, 273, 274, 275, 289, 290, 291, 292, 307, 308, 309, 310, 325, 326, 327, 328, 342, 343, 344, 345, 346, 347, 360, 361, 363, 364, 365, 366, 378, 379, 380, 382, 383, 384, 396, 397, 398], "bbox": [0.008517786561264822, 0.0558125, 0.6537351778656126, 0.9906406249999999], "iscrowd": 0, "area": 61892.989349999996, "rle": {"size": [640, 506], "counts": "bf35hc04K5L4K5K5L4K5M3N2M3N2N2M3^J_NbHc1Z7aNdHa1X7cNeH`1X7cNfH_1V7eNhH]1T7gNiH\\1S7hNkHZ1Q7jN]Hh1_7\\NkGZ2Q8jMmGX2o7lMnGV2P8mMnGU2n7oMPHS2l7PNSHR2i7RNTHQ2h7RNWHo1f7TNYHn1c7VN[Hl1]2SLAS2PNm1Y2VLCP2SNl1T2YLFo1TNj1P2]LHl1WNi1k1aLJi1YNh1g1fLMe1ZNg1c1iL[MQOf0b2;f1^1mLSMZOo0V2?d1Z1RMiLEX1i1b0b1Y1UM_LOa1]1f0_1X1XMUL:i1Q1i0^1V1[MVL:e1Q1m0[1U1FiMROP1Z1T1IgMPOT1SOaMDa3_2eMmNS1QOnM@X3f2bMlNS1nN[N\\On2o2_MjNR1mNhNWOd2X3\\MhNR1lNRORO]2_3ZMfNn0oNAiNT2h3XMbNj0SO0`Nk1P4VMaNe0VO?UNd1Y4SMiN7oNY1kMZ1j<mNUBl1k=Q22N3L3N30O2N3N1N3N2M3M201O1O00100O1O00100O001O10O01O1O10O01O1O010O1O0101O004]LXBe0n=SOUBm0Q>jNSBU1R>cNQB\\1U>\\NoAc1W>TNmAk1X>mMkAR2[>_MPB`2h?01O1N101O0O2O001N1000000O10000O1000aAVMh;i2ZDXMd;h2\\DYMc;g2]DZMb;f2^D[Ma;d2`D]M_;c2aD^M^;b2bD_M];`2dDbMf9GSEg2W1cM_9MYE`2Y1cMX93^EZ2Z1dMQ99cES2]1eMi8?iEl1^1fMc8d0nEi1\\1eMW8P1\\Fa1W1`Mh7d1PGR1R1[MZ7V2dGd0m0WMR7b2PH=h0RMS7e2UH?b0mLU7g2YH?`0kL\\2_O:\\3iL<?jLZ2E6Y3QM:=iLW2K5V3VM7=iLT211S3^M5;hLR27MP3fM39gLo1=Kn2lM08gLl1b0Ij2TNM6hLi1h0Eh2[NJ4hLg1m0Ce2bNH1hLf1R1@b2hNE0iLc1X1^O]2oNDNhLa1]1\\OZ2VOBJiL`1b1YOW2\\O@HjL]1g1XOR2C^OFkL[1k1VOP2H\\OEjLS1eNWN\\3R1l1O[OBjLS1dNWNc3n0g16YO@kLS1]2ROb1;WO]OlLU1_2lNa1b0VO[OlLU1a2gNa1h0TOYOlLW1b2bNa1n0ROWOmLX1a8@dJUOmLZ1_8AfJROmL]1]8AfJQOnL]1\\8BhJnNnL_1[8BjL<V3DkL;V3DkL;U3ElL;T3ElL:T3FmL9T3FnL8S3GmL9S3GnL8S3HmL8R3HnL8S3FoL9Q3FQM9P3FPM:P3ERM:o2DRM=m2BUM=k2CVM<k2BVM>j2BWM=i2BXM>h2AZM?e2A\\M>d2A]M?c2\\OcMc0]2[OeMe0[2XOiMg0X2VOjMk0U2ROnMn0b:0001O0001O0001O000001O01O00000001O01O000000010O000000001O01O00000001O01O000001O0001O000001O0001O0001O0O10000O2O000O10000O2O000O10001N10000O10001N10000O1000O010000O10000O01000O100001O0O101O0000001O0O101O00001O000O2O00001O00000O01O100O1O1O10O01O1O100O1O2N2N1N3M2N3M_U]3"}, "label": "a girl wearing a pink skirt"}]}
{"id": 439889, "image": "COCO_train2014_000000439889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a tennis outfit\"."}], "task": "refering", "answer_template": "The \"a girl wearing a tennis outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 38, 39, 40, 41, 55, 56, 57, 58, 59, 73, 74, 75, 76, 77, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 272, 273, 274, 275, 289, 290, 291, 292, 307, 308, 309, 310, 325, 326, 327, 328, 342, 343, 344, 345, 346, 347, 360, 361, 363, 364, 365, 366, 378, 379, 380, 382, 383, 384, 396, 397, 398], "bbox": [0.008517786561264822, 0.0558125, 0.6537351778656126, 0.9906406249999999], "iscrowd": 0, "area": 61892.989349999996, "rle": {"size": [640, 506], "counts": "bf35hc04K5L4K5K5L4K5M3N2M3N2N2M3^J_NbHc1Z7aNdHa1X7cNeH`1X7cNfH_1V7eNhH]1T7gNiH\\1S7hNkHZ1Q7jN]Hh1_7\\NkGZ2Q8jMmGX2o7lMnGV2P8mMnGU2n7oMPHS2l7PNSHR2i7RNTHQ2h7RNWHo1f7TNYHn1c7VN[Hl1]2SLAS2PNm1Y2VLCP2SNl1T2YLFo1TNj1P2]LHl1WNi1k1aLJi1YNh1g1fLMe1ZNg1c1iL[MQOf0b2;f1^1mLSMZOo0V2?d1Z1RMiLEX1i1b0b1Y1UM_LOa1]1f0_1X1XMUL:i1Q1i0^1V1[MVL:e1Q1m0[1U1FiMROP1Z1T1IgMPOT1SOaMDa3_2eMmNS1QOnM@X3f2bMlNS1nN[N\\On2o2_MjNR1mNhNWOd2X3\\MhNR1lNRORO]2_3ZMfNn0oNAiNT2h3XMbNj0SO0`Nk1P4VMaNe0VO?UNd1Y4SMiN7oNY1kMZ1j<mNUBl1k=Q22N3L3N30O2N3N1N3N2M3M201O1O00100O1O00100O001O10O01O1O10O01O1O010O1O0101O004]LXBe0n=SOUBm0Q>jNSBU1R>cNQB\\1U>\\NoAc1W>TNmAk1X>mMkAR2[>_MPB`2h?01O1N101O0O2O001N1000000O10000O1000aAVMh;i2ZDXMd;h2\\DYMc;g2]DZMb;f2^D[Ma;d2`D]M_;c2aD^M^;b2bD_M];`2dDbMf9GSEg2W1cM_9MYE`2Y1cMX93^EZ2Z1dMQ99cES2]1eMi8?iEl1^1fMc8d0nEi1\\1eMW8P1\\Fa1W1`Mh7d1PGR1R1[MZ7V2dGd0m0WMR7b2PH=h0RMS7e2UH?b0mLU7g2YH?`0kL\\2_O:\\3iL<?jLZ2E6Y3QM:=iLW2K5V3VM7=iLT211S3^M5;hLR27MP3fM39gLo1=Kn2lM08gLl1b0Ij2TNM6hLi1h0Eh2[NJ4hLg1m0Ce2bNH1hLf1R1@b2hNE0iLc1X1^O]2oNDNhLa1]1\\OZ2VOBJiL`1b1YOW2\\O@HjL]1g1XOR2C^OFkL[1k1VOP2H\\OEjLS1eNWN\\3R1l1O[OBjLS1dNWNc3n0g16YO@kLS1]2ROb1;WO]OlLU1_2lNa1b0VO[OlLU1a2gNa1h0TOYOlLW1b2bNa1n0ROWOmLX1a8@dJUOmLZ1_8AfJROmL]1]8AfJQOnL]1\\8BhJnNnL_1[8BjL<V3DkL;V3DkL;U3ElL;T3ElL:T3FmL9T3FnL8S3GmL9S3GnL8S3HmL8R3HnL8S3FoL9Q3FQM9P3FPM:P3ERM:o2DRM=m2BUM=k2CVM<k2BVM>j2BWM=i2BXM>h2AZM?e2A\\M>d2A]M?c2\\OcMc0]2[OeMe0[2XOiMg0X2VOjMk0U2ROnMn0b:0001O0001O0001O000001O01O00000001O01O000000010O000000001O01O00000001O01O000001O0001O000001O0001O0001O0O10000O2O000O10000O2O000O10001N10000O10001N10000O1000O010000O10000O01000O100001O0O101O0000001O0O101O00001O000O2O00001O00000O01O100O1O1O10O01O1O100O1O2N2N1N3M2N3M_U]3"}, "label": "a girl wearing a tennis outfit"}]}
{"id": 243292, "image": "COCO_train2014_000000243292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train that is in the lead\"."}], "task": "refering", "answer_template": "The \"the train that is in the lead\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 282, 283, 284, 285], "bbox": [0.2991875, 0.4834192037470726, 0.8661406250000001, 0.869063231850117], "iscrowd": 0, "area": 14845.684600000008, "rle": {"size": [427, 640], "counts": "ho_21T=7SOm0L5K4M3L4L4L3M3N3M2O1O2N1O1O2N100O2N1O1O2N1O1O1O1O101N1O1O100O2O00000O01O1O00100O001O10O01O1O1O010O1O00100O001O10O01O100O001O10O01O10O01O1O010O1O10O01O10O01O1O010O1O010O1O1O010O1O010O1O00100O00100O1O00100O00100O001O10O01O100O001O10O0100O00100O00100O1O010O1O010O1O010O10O01O100O00100O010O1O010O10O01O100O010O10O01O10O0100O00100O10O01O10O0100O00100O010O1O10O0100O00100O010O1O010O100O00100O010O1O01O0100O1O1]OgC4Y<GRD2o;H]D2d;Nl07I7I8H5J10O100O1000O010000O100O01000O100O01000O100O10O10O10000O10O010000O100O01000O100O10000O010O100O100O1000O0100O100O100O100O01000O100O100O10O010000O100O100O0100000O10000000O01000000O1000000O010000000O100000O01000000O100000O0100000000O1000O10O100000000O1000O100000O10000000O0100000000O1000O1000O100000000O01TeS1"}, "label": "the train that is in the lead"}]}
{"id": 243292, "image": "COCO_train2014_000000243292.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train with the red engine\"."}], "task": "refering", "answer_template": "The \"the train with the red engine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 282, 283, 284, 285], "bbox": [0.2991875, 0.4834192037470726, 0.8661406250000001, 0.869063231850117], "iscrowd": 0, "area": 14845.684600000008, "rle": {"size": [427, 640], "counts": "ho_21T=7SOm0L5K4M3L4L4L3M3N3M2O1O2N1O1O2N100O2N1O1O2N1O1O1O1O101N1O1O100O2O00000O01O1O00100O001O10O01O1O1O010O1O00100O001O10O01O100O001O10O01O10O01O1O010O1O10O01O10O01O1O010O1O010O1O1O010O1O010O1O00100O00100O1O00100O00100O001O10O01O100O001O10O0100O00100O00100O1O010O1O010O1O010O10O01O100O00100O010O1O010O10O01O100O010O10O01O10O0100O00100O10O01O10O0100O00100O010O1O10O0100O00100O010O1O010O100O00100O010O1O01O0100O1O1]OgC4Y<GRD2o;H]D2d;Nl07I7I8H5J10O100O1000O010000O100O01000O100O01000O100O10O10O10000O10O010000O100O01000O100O10000O010O100O100O1000O0100O100O100O100O01000O100O100O10O010000O100O100O0100000O10000000O01000000O1000000O010000000O100000O01000000O100000O0100000000O1000O10O100000000O1000O100000O10000000O0100000000O1000O1000O100000000O01TeS1"}, "label": "the train with the red engine"}]}
{"id": 382559, "image": "COCO_train2014_000000382559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white couch against the back wall\"."}], "task": "refering", "answer_template": "The \"white couch against the back wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 260, 261, 262, 284, 285, 307], "bbox": [0.337359375, 0.47014084507042253, 0.7745468750000001, 0.9392018779342723], "iscrowd": 0, "area": 22148.77145000001, "rle": {"size": [426, 640], "counts": "]Tj22S=5K5L4K5SD@l:d0dDn0Y:S2mNc0]O2O1N2N2O1N2N2N^NmHQMR7l2f12N2O1kN]MaGe2_8^1E0000000000001OD<00000000000000001O0000000000000000001O00000000000000001OM3I7I7H8I7H8I7I7H8I7K5000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000CYEWNg:h1>00000000000000001O001O00=C1O0000_Oa00000O1000000O11O00001O001O0cDbNn:m10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000O10000000[1eN2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1OlQl1"}, "label": "white couch against the back wall"}]}
{"id": 382559, "image": "COCO_train2014_000000382559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"maroon seats below windows\"."}], "task": "refering", "answer_template": "The \"maroon seats below windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 260, 261, 262, 284, 285, 307], "bbox": [0.337359375, 0.47014084507042253, 0.7745468750000001, 0.9392018779342723], "iscrowd": 0, "area": 22148.77145000001, "rle": {"size": [426, 640], "counts": "]Tj22S=5K5L4K5SD@l:d0dDn0Y:S2mNc0]O2O1N2N2O1N2N2N^NmHQMR7l2f12N2O1kN]MaGe2_8^1E0000000000001OD<00000000000000001O0000000000000000001O00000000000000001OM3I7I7H8I7H8I7I7H8I7K5000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000CYEWNg:h1>00000000000000001O001O00=C1O0000_Oa00000O1000000O11O00001O001O0cDbNn:m10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000O10000000[1eN2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1OlQl1"}, "label": "maroon seats below windows"}]}
{"id": 382559, "image": "COCO_train2014_000000382559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white chair is pushed out\"."}], "task": "refering", "answer_template": "The \"the white chair is pushed out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [275, 296, 297, 298, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.748484375, 0.7365727699530515, 0.9998125, 1.0], "iscrowd": 0, "area": 5224.807499999999, "rle": {"size": [426, 640], "counts": "neW61U=4L4M3L4M3L40000000000000001O0O1000000000000000000O1001O0000001O000000001O000000001O0000001O000000001O0000001O00001O00001O0000001O00001O00001O00001O001O001O001O001O001OO1M3L4M3L4M3L4M3L4M3M3L4M3L3N3L4M3L10O3M4M3L3M4L4L4M3L4LREROh9j0XFZOh9c0YF_Og9=[FEf96[FMe92YF1g9OVF3l9LQF7P:HlE<T:CjE`0V:@gEc0Y:]OdEf0[:[OaEi0_:WO^El0b:TO[Eo0d:ROYEQ1g:oNUEU1j:kNTEX1l:hNQEZ1P;`0OK6O0O1O2N1O000O00010O00010O00010O00011N3M3N2M3M3N2M3M4M7]O^F"}, "label": "the white chair is pushed out"}]}
{"id": 382559, "image": "COCO_train2014_000000382559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair closer to the wine\"."}], "task": "refering", "answer_template": "The \"the chair closer to the wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [275, 296, 297, 298, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.748484375, 0.7365727699530515, 0.9998125, 1.0], "iscrowd": 0, "area": 5224.807499999999, "rle": {"size": [426, 640], "counts": "neW61U=4L4M3L4M3L40000000000000001O0O1000000000000000000O1001O0000001O000000001O000000001O0000001O000000001O0000001O00001O00001O0000001O00001O00001O00001O001O001O001O001O001OO1M3L4M3L4M3L4M3L4M3M3L4M3L3N3L4M3L10O3M4M3L3M4L4L4M3L4LREROh9j0XFZOh9c0YF_Og9=[FEf96[FMe92YF1g9OVF3l9LQF7P:HlE<T:CjE`0V:@gEc0Y:]OdEf0[:[OaEi0_:WO^El0b:TO[Eo0d:ROYEQ1g:oNUEU1j:kNTEX1l:hNQEZ1P;`0OK6O0O1O2N1O000O00010O00010O00010O00011N3M3N2M3M3N2M3M4M7]O^F"}, "label": "the chair closer to the wine"}]}
{"id": 79456, "image": "COCO_train2014_000000079456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with pigtails\"."}], "task": "refering", "answer_template": "The \"a girl with pigtails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 160, 161, 162, 163, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305], "bbox": [0.3967019027484143, 0.2892, 0.9644397463002115, 0.9768399999999999], "iscrowd": 0, "area": 67647.09015, "rle": {"size": [500, 473], "counts": "gVl27S>MPC7m<OmB3Q=3iBOU=7fBJY=;aBH\\=>_BC`=b0ZB@d=Z1N1N2N2K6G8I7M4M2N2N2M4M2N2M3N2N2M3N3ZMXM\\Hj2b7^MWHc2g7eMSH[2k7mMoGS2o7SNmGl1R8WNmGi1Q8ZNmGg1Q8\\NnGd1P8_NoGa1o7bNPH^1n7dNRH\\1l7gNRHZ1h7mNWHR1e7TOZHl0e7VOYHk0e7XOZHh0e7YO[Hg0d7[OZHf0d7]O[Hc0d7_OZHb0e7_O[H`0d7C[H=d7EZH<e7E[H;c7H\\H8c7J[H7d7K[H5e7K[H5e7l3101O001O001O1N101O01O001O010O001O001O001O001O001M2N3M3M2N5K7I7G9A<E;M2O1O1O1O1O1O1O2NO100000000O1000000O1000LPFdJP:\\5QFcJo9]5RFbJm9_54001O000O1000001O0O100001O2N101N1O1O000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0001O001O01O01dJQFQ5o9jJVFV5T:00001O00000000001O00000O1O1O100O1O100O1O1UFdJ^9\\5]FjJa9f5N3M2O1N3M2O1N3N1O2N1O1O2N1O2N1O1O2N101N2N3M2N2N2N3M2N2N2N3M2N2N2O2M2N2gH`I[5h9dN1O2N100O2N100O8H:F9H8G9G9H8G9G9G9H8G9G9G:Gej7"}, "label": "a girl with pigtails"}]}
{"id": 79456, "image": "COCO_train2014_000000079456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with blonde hair wearing white top\"."}], "task": "refering", "answer_template": "The \"a girl with blonde hair wearing white top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 150, 151, 160, 161, 162, 163, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305], "bbox": [0.3967019027484143, 0.2892, 0.9644397463002115, 0.9768399999999999], "iscrowd": 0, "area": 67647.09015, "rle": {"size": [500, 473], "counts": "gVl27S>MPC7m<OmB3Q=3iBOU=7fBJY=;aBH\\=>_BC`=b0ZB@d=Z1N1N2N2K6G8I7M4M2N2N2M4M2N2M3N2N2M3N3ZMXM\\Hj2b7^MWHc2g7eMSH[2k7mMoGS2o7SNmGl1R8WNmGi1Q8ZNmGg1Q8\\NnGd1P8_NoGa1o7bNPH^1n7dNRH\\1l7gNRHZ1h7mNWHR1e7TOZHl0e7VOYHk0e7XOZHh0e7YO[Hg0d7[OZHf0d7]O[Hc0d7_OZHb0e7_O[H`0d7C[H=d7EZH<e7E[H;c7H\\H8c7J[H7d7K[H5e7K[H5e7l3101O001O001O1N101O01O001O010O001O001O001O001O001M2N3M3M2N5K7I7G9A<E;M2O1O1O1O1O1O1O2NO100000000O1000000O1000LPFdJP:\\5QFcJo9]5RFbJm9_54001O000O1000001O0O100001O2N101N1O1O000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0001O001O01O01dJQFQ5o9jJVFV5T:00001O00000000001O00000O1O1O100O1O100O1O1UFdJ^9\\5]FjJa9f5N3M2O1N3M2O1N3N1O2N1O1O2N1O2N1O1O2N101N2N3M2N2N2N3M2N2N2N3M2N2N2O2M2N2gH`I[5h9dN1O2N100O2N100O8H:F9H8G9G9H8G9G9G9H8G9G9G:Gej7"}, "label": "a girl with blonde hair wearing white top"}]}
{"id": 79456, "image": "COCO_train2014_000000079456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the photograph\"."}], "task": "refering", "answer_template": "The \"the boy in the photograph\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 94, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 273, 274, 275, 276, 277, 278, 279, 280, 281, 296, 297, 298], "bbox": [0.007167019027484144, 0.10102, 0.5846088794926004, 0.9746], "iscrowd": 0, "area": 65688.30705000002, "rle": {"size": [500, 473], "counts": "mj11`?4K5L3L5K5L4K5L4K5L4K4L5L4K5L4K5L4K5M6J7I6J6J7I6J6K6I6J6J7I3M2N3M2N2N2N2N2N2N3M2N2N2N2N2N201N100O1O100O101N1O100O100O1O2O0O100O1O100O2O0O1O100O100O2N100O1O100O101N1O100O100O1O2O0O1mKdInN\\6f0UJfL_Oo0]6U2^JeLAl0R6^2cJ]L@T1m5^2iJVL^OZ1k5^2TLbMl3]2VLaMk3^2WL`Mj3_2XL`Mh3_2YL`Mh3_2ZL`Mf3_2\\L_Mf3`2[L^Mf3a2[L_Me3`2]L^Md3a2^L^Mb3a2_L^Mb3a2`L^M`3a2bL^M_3`2bL`M^3_2dL`M\\3_2fL`MZ3`2fL`MZ3_2hL`MX3_2jL`MV3_2kLaMV3]2lLbMT3]2nLbMR3]2oLcMQ3\\2QMcMo2]2RMbMn2]2SMaMo2^2RMaMP3^2PMaMQ3^2PM`MR3_2oL`MR3`2nL_MS3`2nL^MT3a2mL^MT3b2lL]MV3a2o4O1O100O1O101N1O100O1O1O101O0000000000001O0000000000001O0000000001O01O0000000000001O000000000000eNbMaE]2^:dMbE\\2\\:fMdEZ2Z:hMfEX2X:jMhEV2V:lMjET2U:mMkET2R:nMoEQ2P9PNUF1k0o1h8cNPF@X1m1e8[O[Ge0c8^O\\Gb0c8_O]Ga0a8A_G`0^8BcG>Z8DfG=V8FjG;Q8IoG8m7KSH6h7NXH3d70\\H1_73aHN[75eHLV78jHIR7:nHGm6=SIDh6a0WI@e6c0[I^O`6f0`I[O\\6h0dIYOX6j0hIWOW6i0iIXOV6h0jIYOU6g0kIZOa3eMPMQ3_O\\O\\3_N\\LU28]OY3hNVLl1a0]OU3kNYLh1b0^OQ3nN\\Ld1d0^Om2QO^La1e0_Oi2TOaL]1f0@f2VOcLZ1g0Ab2XOfLY1g0@`2ZOhLV1h0A\\2]OkLR1i0BY2_OmLo0j0CU2BPMk0k0DQ2ESMg0l0En1GUMd0m0Fl1HUMc0o0Fi1JWM`0P1Gf1LYM=Q1Hd1MZM;R1Ia1O\\M8S1J^11^M5T1K\\12^M4W1LV13bM1Y14j0NlMM\\1=>HUNK]1e04`3MaL0`31aLKa35`LHb39_LDa3>_L_Oc3b0^LZOd3f0]LXOd3i0]LSOe3n0[LoNg3R1ZLkNf3V1[LgNg3Y1[LdNf3[1\\LbNf3^1\\L_Nf3_1\\L^Ng3a1X5M2N3N2M2O2M3N1N3N2M2N3N2M2O3L4M3L4M2M4MgWo2"}, "label": "the boy in the photograph"}]}
{"id": 79456, "image": "COCO_train2014_000000079456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy to the left of the girl\"."}], "task": "refering", "answer_template": "The \"the boy to the left of the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 88, 89, 90, 91, 92, 93, 94, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 273, 274, 275, 276, 277, 278, 279, 280, 281, 296, 297, 298], "bbox": [0.007167019027484144, 0.10102, 0.5846088794926004, 0.9746], "iscrowd": 0, "area": 65688.30705000002, "rle": {"size": [500, 473], "counts": "mj11`?4K5L3L5K5L4K5L4K5L4K4L5L4K5L4K5L4K5M6J7I6J6J7I6J6K6I6J6J7I3M2N3M2N2N2N2N2N2N3M2N2N2N2N2N201N100O1O100O101N1O100O100O1O2O0O100O1O100O2O0O1O100O100O2N100O1O100O101N1O100O100O1O2O0O1mKdInN\\6f0UJfL_Oo0]6U2^JeLAl0R6^2cJ]L@T1m5^2iJVL^OZ1k5^2TLbMl3]2VLaMk3^2WL`Mj3_2XL`Mh3_2YL`Mh3_2ZL`Mf3_2\\L_Mf3`2[L^Mf3a2[L_Me3`2]L^Md3a2^L^Mb3a2_L^Mb3a2`L^M`3a2bL^M_3`2bL`M^3_2dL`M\\3_2fL`MZ3`2fL`MZ3_2hL`MX3_2jL`MV3_2kLaMV3]2lLbMT3]2nLbMR3]2oLcMQ3\\2QMcMo2]2RMbMn2]2SMaMo2^2RMaMP3^2PMaMQ3^2PM`MR3_2oL`MR3`2nL_MS3`2nL^MT3a2mL^MT3b2lL]MV3a2o4O1O100O1O101N1O100O1O1O101O0000000000001O0000000000001O0000000001O01O0000000000001O000000000000eNbMaE]2^:dMbE\\2\\:fMdEZ2Z:hMfEX2X:jMhEV2V:lMjET2U:mMkET2R:nMoEQ2P9PNUF1k0o1h8cNPF@X1m1e8[O[Ge0c8^O\\Gb0c8_O]Ga0a8A_G`0^8BcG>Z8DfG=V8FjG;Q8IoG8m7KSH6h7NXH3d70\\H1_73aHN[75eHLV78jHIR7:nHGm6=SIDh6a0WI@e6c0[I^O`6f0`I[O\\6h0dIYOX6j0hIWOW6i0iIXOV6h0jIYOU6g0kIZOa3eMPMQ3_O\\O\\3_N\\LU28]OY3hNVLl1a0]OU3kNYLh1b0^OQ3nN\\Ld1d0^Om2QO^La1e0_Oi2TOaL]1f0@f2VOcLZ1g0Ab2XOfLY1g0@`2ZOhLV1h0A\\2]OkLR1i0BY2_OmLo0j0CU2BPMk0k0DQ2ESMg0l0En1GUMd0m0Fl1HUMc0o0Fi1JWM`0P1Gf1LYM=Q1Hd1MZM;R1Ia1O\\M8S1J^11^M5T1K\\12^M4W1LV13bM1Y14j0NlMM\\1=>HUNK]1e04`3MaL0`31aLKa35`LHb39_LDa3>_L_Oc3b0^LZOd3f0]LXOd3i0]LSOe3n0[LoNg3R1ZLkNf3V1[LgNg3Y1[LdNf3[1\\LbNf3^1\\L_Nf3_1\\L^Ng3a1X5M2N3N2M2O2M3N1N3N2M2N3N2M2O3L4M3L4M2M4MgWo2"}, "label": "the boy to the left of the girl"}]}
{"id": 112226, "image": "COCO_train2014_000000112226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pure green umbrella towards the right\"."}], "task": "refering", "answer_template": "The \"the pure green umbrella towards the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 36, 37, 38, 39, 40, 56, 74], "bbox": [0.02348, 0.08298666666666667, 0.28020000000000006, 0.3805333333333333], "iscrowd": 0, "area": 3795.9451999999997, "rle": {"size": [375, 500], "counts": "g^41e;1O1O100O1O1O1O1O1O2N1O1O1O1O1O010O001O001O2N3M3M1O1O1O1O1O100O1O1O2N10000000000O100000001O000O1000000000000O101O000cGiNZ6W1mHCR7=TH=k7i1OTOl0iNW1L5O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O00001O001O1O001O1N1O2N2N2N101O11N2O000O01O1O001O1O1O011N1O1O2N1O1OghS4"}, "label": "the pure green umbrella towards the right"}]}
{"id": 112226, "image": "COCO_train2014_000000112226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green umbrella with no one under it\"."}], "task": "refering", "answer_template": "The \"a green umbrella with no one under it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 36, 37, 38, 39, 40, 56, 74], "bbox": [0.02348, 0.08298666666666667, 0.28020000000000006, 0.3805333333333333], "iscrowd": 0, "area": 3795.9451999999997, "rle": {"size": [375, 500], "counts": "g^41e;1O1O100O1O1O1O1O1O2N1O1O1O1O1O010O001O001O2N3M3M1O1O1O1O1O100O1O1O2N10000000000O100000001O000O1000000000000O101O000cGiNZ6W1mHCR7=TH=k7i1OTOl0iNW1L5O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O00001O001O1O001O1N1O2N2N2N101O11N2O000O01O1O001O1O1O011N1O1O2N1O1OghS4"}, "label": "a green umbrella with no one under it"}]}
{"id": 112226, "image": "COCO_train2014_000000112226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white lg umbrella\"."}], "task": "refering", "answer_template": "The \"a white lg umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 53, 68, 69, 86], "bbox": [0.64888, 0.03146666666666667, 0.9994400000000001, 0.4], "iscrowd": 0, "area": 8515.302900000002, "rle": {"size": [375, 500], "counts": "_ff31f;4K5K5L4K100O1O100O100O1O100O2O0O1O100O1O100O100O1O100O100O1O100O100O1O100O10M200101O1O1N2O1O1O000O1M3N20000O10000O1000000O10000O1000000000000000000000000000000000000000001O00000PHgNb5Y1]IHc6f20K5^Oa0_Ob0_Oa0^Ob0N11ON5L2N3O000002N1O1O1O2N00001O01O0000000000000000000000001O01O00000001O00001O00001O00000010O0001O00001O001O001O1N101O1O1O0O200O00100O1O001O1O1O001N2O001O1OgN"}, "label": "a white lg umbrella"}]}
{"id": 112226, "image": "COCO_train2014_000000112226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white umbrella that says lg\"."}], "task": "refering", "answer_template": "The \"a white umbrella that says lg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 53, 68, 69, 86], "bbox": [0.64888, 0.03146666666666667, 0.9994400000000001, 0.4], "iscrowd": 0, "area": 8515.302900000002, "rle": {"size": [375, 500], "counts": "_ff31f;4K5K5L4K100O1O100O100O1O100O2O0O1O100O1O100O100O1O100O100O1O100O100O1O100O10M200101O1O1N2O1O1O000O1M3N20000O10000O1000000O10000O1000000000000000000000000000000000000000001O00000PHgNb5Y1]IHc6f20K5^Oa0_Ob0_Oa0^Ob0N11ON5L2N3O000002N1O1O1O2N00001O01O0000000000000000000000001O01O00000001O00001O00001O00000010O0001O00001O001O001O1N101O1O1O0O200O00100O1O001O1O1O001N2O001O1OgN"}, "label": "a white umbrella that says lg"}]}
{"id": 112226, "image": "COCO_train2014_000000112226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green and while striped umbrella in the background closest to the woman eating\"."}], "task": "refering", "answer_template": "The \"the green and while striped umbrella in the background closest to the woman eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 21, 22, 23, 26, 27, 28, 29, 30], "bbox": [0.17568, 0.007893333333333334, 0.7618199999999999, 0.13626666666666667], "iscrowd": 0, "area": 8039.220450000001, "rle": {"size": [375, 500], "counts": "[WP12e;2N3M3M3M2N3M3M2N3M3M1O00000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000000000000000O1N2N2N2N2001O00001O001O00001O00001O000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O100000000000000000000006J8H8H1OO1O1O1[OfEL[:4iEGX:9kECV:=nE^OS:c0`000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1003M2N3M2N3M1O0000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O2O3M2N2N2N3L3N2N2N3Mha[1"}, "label": "the green and while striped umbrella in the background closest to the woman eating"}]}
{"id": 538210, "image": "COCO_train2014_000000538210.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stack of papers on the farthest left\"."}], "task": "refering", "answer_template": "The \"the stack of papers on the farthest left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 162, 163, 180, 181, 182, 198, 199, 200, 216, 217], "bbox": [0.0, 0.6502133333333334, 0.17488, 0.9888], "iscrowd": 0, "area": 5905.663349999998, "rle": {"size": [375, 500], "counts": "[:X1P8_2H81O001N101O00001O001O0O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2M2]NZFT1h9jNZFT1g9kN[FS1g9jN\\FT1e9kN^FR1d9lN^FR1c9lN`FR1S:O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1Ofnf4"}, "label": "the stack of papers on the farthest left"}]}
{"id": 374374, "image": "COCO_train2014_000000374374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl skateboarding outside\"."}], "task": "refering", "answer_template": "The \"a young girl skateboarding outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 56, 57, 58, 78, 79, 80, 81, 82, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 287, 288], "bbox": [0.341328125, 0.12266355140186916, 0.67125, 0.844929906542056], "iscrowd": 0, "area": 37250.10755000002, "rle": {"size": [428, 640], "counts": "bZk26m<j0XC@`;i1[O6K4K4M3K5J6I8I6K6J6L2M4L4M3M3M3N1N3N2O1N1O2N2N2N1O2O1N1O2O0O2N101N101_Oa0N1O2M3N1N2O2N1ONPIjIo6V631O1O10O0100O101K4O2O1N2O1lI[I\\5h6bJ]IX5f6gJ[IX5f6eJ^IY5c6dJ`I[5`6bJdIW5c6fJaIX5`6fJbIS5\\OmJm7R58M2O1N2O1O1M3J6O101O00000O1O100O100000000O2O000O1O1O1000000000000001O1O0000O2O0O11O1O001O1O00000000O100001O0O1OnKRGi3l8WLUGi3k8VLWGi3OQLd85^Gj3LSLg81_GR4a8lK`GU4_8kKbGT4_8kK`GW4^8jKaGZ4\\8eKjGW4`1iKd41lIW4V1gKbN7X6KPJX4R1lKbN4Z6HQJZ4m0^Lo4XOTJ\\4a0eKZOg0c0XO_40RK^4a0hLGVOm4U4ZKgLGWOj4W4\\KdLHYOg4T4aKeLFZOe4R4dKgLDYOf4R4dKlL@SOg1VOT1m4dMnL^OPOf1ZOU1i4_MQKKl2]1_MV1e4`MSKKk2^1^MS1g4bMTKJi2`1]MP1]8nNeGP1_8jNdGV1`8bNdG^1_8\\NdGc1T:O00000O2O0O101N10000O2O0O1O100O1O100O1O100O0011N1001O000O101O00001O000N2O2N12N01M2O1O1O2N1O1O2N2N2M3N1N2N2O3K5J6J5H9HS`g2"}, "label": "a young girl skateboarding outside"}]}
{"id": 374374, "image": "COCO_train2014_000000374374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with a white hoodie riding a skateboard on the street\"."}], "task": "refering", "answer_template": "The \"a girl with a white hoodie riding a skateboard on the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 56, 57, 58, 78, 79, 80, 81, 82, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 287, 288], "bbox": [0.341328125, 0.12266355140186916, 0.67125, 0.844929906542056], "iscrowd": 0, "area": 37250.10755000002, "rle": {"size": [428, 640], "counts": "bZk26m<j0XC@`;i1[O6K4K4M3K5J6I8I6K6J6L2M4L4M3M3M3N1N3N2O1N1O2N2N2N1O2O1N1O2O0O2N101N101_Oa0N1O2M3N1N2O2N1ONPIjIo6V631O1O10O0100O101K4O2O1N2O1lI[I\\5h6bJ]IX5f6gJ[IX5f6eJ^IY5c6dJ`I[5`6bJdIW5c6fJaIX5`6fJbIS5\\OmJm7R58M2O1N2O1O1M3J6O101O00000O1O100O100000000O2O000O1O1O1000000000000001O1O0000O2O0O11O1O001O1O00000000O100001O0O1OnKRGi3l8WLUGi3k8VLWGi3OQLd85^Gj3LSLg81_GR4a8lK`GU4_8kKbGT4_8kK`GW4^8jKaGZ4\\8eKjGW4`1iKd41lIW4V1gKbN7X6KPJX4R1lKbN4Z6HQJZ4m0^Lo4XOTJ\\4a0eKZOg0c0XO_40RK^4a0hLGVOm4U4ZKgLGWOj4W4\\KdLHYOg4T4aKeLFZOe4R4dKgLDYOf4R4dKlL@SOg1VOT1m4dMnL^OPOf1ZOU1i4_MQKKl2]1_MV1e4`MSKKk2^1^MS1g4bMTKJi2`1]MP1]8nNeGP1_8jNdGV1`8bNdG^1_8\\NdGc1T:O00000O2O0O101N10000O2O0O1O100O1O100O1O100O0011N1001O000O101O00001O000N2O2N12N01M2O1O1O2N1O1O2N2N2M3N1N2N2O3K5J6J5H9HS`g2"}, "label": "a girl with a white hoodie riding a skateboard on the street"}]}
{"id": 374374, "image": "COCO_train2014_000000374374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing on skateboard\"."}], "task": "refering", "answer_template": "The \"man standing on skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 39, 40, 61, 62, 63, 83, 84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 154, 155, 156, 178, 179, 201, 202, 203, 224, 225, 226], "bbox": [0.6477031249999999, 0.029205607476635514, 0.864109375, 0.647196261682243], "iscrowd": 0, "area": 15153.264000000008, "rle": {"size": [428, 640], "counts": "]a]5`0i<6L4L4L4K4M3M4L3L5L3M3L4M3M2M4M3M2N3L3N3M3M2M4M3M2N3L3N3M2N2M4M2N2O2M2O1O2N1N2O1O1N2O1O1O1O1O001O1N1O2N2N1O2N2M2M4M3K4O2M3M2O2N2N2O0O2N2N1bJ^JQ3W1PKi2h1WLV3n0WKe0JOb1dN]3f0ZK`030P1POc3?\\K`021h0VOj39^K?01b0]OP42`K`0M2<CW4JcK?J36J]4CgK>F5N1f4YOjK`0A6I7m4QOlK`0_O8A=T5jNUL:VOf2f4mLYL;QOg2f4mL[L;oNh2g4kL\\L=mNh2f4iL`L>kNh2f4hLaL`0iNg2f4gLdLa0gNg2f4fLeLc0eNf2f4fLgLd0cNe2h4cLhLg0aNe2i4_LjLl0]Nd2Y7[MhHd2Y7\\MgHc2Z7\\MgHc2Z7]MgHa2Z7_MfHa2Z7^MgHa2Z7_MfH`2[7_MgH_2Z7aMfH^2[7bMeH]2\\7bMfH\\2\\7cMeH[2\\7dMeH[2\\7eMeHY2]7eMdHZ2]7fMdHX2]7hMeHU2\\7jMjHP2X7oMlHl1U7SNoHi1R7WNRId1o6\\NUI_1l6`NZIZ1g6fN^IT1c6lNbI0oMO`80gIHoM5[83cJK^54cJK^55aJK`55`JJa55_JKb55^JJd55[JKf54[JKf55ZJJg56XJJj54WJKj55UJKl55TJJm55TJJn55QJKQ63PJLR63mIMU61kINW60jIOX6NkI0i90SST1"}, "label": "man standing on skateboard"}]}
{"id": 374374, "image": "COCO_train2014_000000374374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was skating the road ajnd with in girl friend\"."}], "task": "refering", "answer_template": "The \"a man was skating the road ajnd with in girl friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 39, 40, 61, 62, 63, 83, 84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 154, 155, 156, 178, 179, 201, 202, 203, 224, 225, 226], "bbox": [0.6477031249999999, 0.029205607476635514, 0.864109375, 0.647196261682243], "iscrowd": 0, "area": 15153.264000000008, "rle": {"size": [428, 640], "counts": "]a]5`0i<6L4L4L4K4M3M4L3L5L3M3L4M3M2M4M3M2N3L3N3M3M2M4M3M2N3L3N3M2N2M4M2N2O2M2O1O2N1N2O1O1N2O1O1O1O1O001O1N1O2N2N1O2N2M2M4M3K4O2M3M2O2N2N2O0O2N2N1bJ^JQ3W1PKi2h1WLV3n0WKe0JOb1dN]3f0ZK`030P1POc3?\\K`021h0VOj39^K?01b0]OP42`K`0M2<CW4JcK?J36J]4CgK>F5N1f4YOjK`0A6I7m4QOlK`0_O8A=T5jNUL:VOf2f4mLYL;QOg2f4mL[L;oNh2g4kL\\L=mNh2f4iL`L>kNh2f4hLaL`0iNg2f4gLdLa0gNg2f4fLeLc0eNf2f4fLgLd0cNe2h4cLhLg0aNe2i4_LjLl0]Nd2Y7[MhHd2Y7\\MgHc2Z7\\MgHc2Z7]MgHa2Z7_MfHa2Z7^MgHa2Z7_MfH`2[7_MgH_2Z7aMfH^2[7bMeH]2\\7bMfH\\2\\7cMeH[2\\7dMeH[2\\7eMeHY2]7eMdHZ2]7fMdHX2]7hMeHU2\\7jMjHP2X7oMlHl1U7SNoHi1R7WNRId1o6\\NUI_1l6`NZIZ1g6fN^IT1c6lNbI0oMO`80gIHoM5[83cJK^54cJK^55aJK`55`JJa55_JKb55^JJd55[JKf54[JKf55ZJJg56XJJj54WJKj55UJKl55TJJm55TJJn55QJKQ63PJLR63mIMU61kINW60jIOX6NkI0i90SST1"}, "label": "a man was skating the road ajnd with in girl friend"}]}
{"id": 71271, "image": "COCO_train2014_000000071271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the right not wearing glasses\"."}], "task": "refering", "answer_template": "The \"the man on the right not wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 110, 111, 112, 133, 134, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 339, 340, 341, 342, 343, 362, 363, 364, 365, 366, 385, 386, 387, 388, 389], "bbox": [0.6912812500000001, 0.1713333333333333, 0.9662656250000001, 1.0], "iscrowd": 0, "area": 45410.09645, "rle": {"size": [480, 640], "counts": "e]_61i><D;_Oa0C<J3M2N2O1N2N3M2N2N2M3N2RFfMT7\\2hHjMT7X2gHnMU7U2fHQNW7P2eHVNY7k1cHZN[7g1bH^NZ7d1bHaN]7_1`HfN]7Z1aHkN]7U1_HRO^7n0_HXO^7h0^HEY7;dHNV72gH3V7NfH9W7GaHc0]7b3N3M2N2N3L3N3M2N2M4M4L3L4M3M3L5L3M3L4L4M4K4M3L4L4kMhFPMKk0`9R2PGdLKR1Y9V2[GhMi8T2YGkMi8R2YGlMh8S2ZGlMf8S2\\GkMe8T2]GiMe8V2_GfMb8Y2bGbM`8]2lGWMU8i2Q2O1O100O1O100O1O100O100O100O100000000O1000000001mN^MfEb2Z:fM^EZ2b:nMVER2j:VNnDk1Q;]NgDc1Y;R10001O000lF]Lk6d3mHdLS7[3jHjLZ7P3eHRM`7i2^HZMf7`2YHbMl7X2SHjMR8Q2kGQNX8l1gGVNX8j1fGXNZ8i1dGYN[8i1aGYN_8h1^G[Na8g1\\GZNd8h1XG[Ng8g1lFdNT9_1fFeNY9^1aFeN_9_3O001O001O001O00001O00001O0000001O00001O00001O00001O001O2YM]JmKd5l3nJhKS5Q4^KdKc4X4jK[KZ4d4oKgJ_4W5kKTJc4h5^2M2N3lMcF_N`9]1cFaN`9\\1cF`Na9\\1bFbN`9\\1cFaN`9[1cFcN`9Y1dFcN`9X1dFfN`9T1eFhN`9S1cFkNb9n0cFlNd9n0aFlNg9l0^FoNi9i0^FROi9f0]FTOj9e0\\FVOk9b0[FYOl9>b2HdQ:"}, "label": "the man on the right not wearing glasses"}]}
{"id": 71271, "image": "COCO_train2014_000000071271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man without glasses playing the wii\"."}], "task": "refering", "answer_template": "The \"a man without glasses playing the wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 110, 111, 112, 133, 134, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 339, 340, 341, 342, 343, 362, 363, 364, 365, 366, 385, 386, 387, 388, 389], "bbox": [0.6912812500000001, 0.1713333333333333, 0.9662656250000001, 1.0], "iscrowd": 0, "area": 45410.09645, "rle": {"size": [480, 640], "counts": "e]_61i><D;_Oa0C<J3M2N2O1N2N3M2N2N2M3N2RFfMT7\\2hHjMT7X2gHnMU7U2fHQNW7P2eHVNY7k1cHZN[7g1bH^NZ7d1bHaN]7_1`HfN]7Z1aHkN]7U1_HRO^7n0_HXO^7h0^HEY7;dHNV72gH3V7NfH9W7GaHc0]7b3N3M2N2N3L3N3M2N2M4M4L3L4M3M3L5L3M3L4L4M4K4M3L4L4kMhFPMKk0`9R2PGdLKR1Y9V2[GhMi8T2YGkMi8R2YGlMh8S2ZGlMf8S2\\GkMe8T2]GiMe8V2_GfMb8Y2bGbM`8]2lGWMU8i2Q2O1O100O1O100O1O100O100O100O100000000O1000000001mN^MfEb2Z:fM^EZ2b:nMVER2j:VNnDk1Q;]NgDc1Y;R10001O000lF]Lk6d3mHdLS7[3jHjLZ7P3eHRM`7i2^HZMf7`2YHbMl7X2SHjMR8Q2kGQNX8l1gGVNX8j1fGXNZ8i1dGYN[8i1aGYN_8h1^G[Na8g1\\GZNd8h1XG[Ng8g1lFdNT9_1fFeNY9^1aFeN_9_3O001O001O001O00001O00001O0000001O00001O00001O00001O001O2YM]JmKd5l3nJhKS5Q4^KdKc4X4jK[KZ4d4oKgJ_4W5kKTJc4h5^2M2N3lMcF_N`9]1cFaN`9\\1cF`Na9\\1bFbN`9\\1cFaN`9[1cFcN`9Y1dFcN`9X1dFfN`9T1eFhN`9S1cFkNb9n0cFlNd9n0aFlNg9l0^FoNi9i0^FROi9f0]FTOj9e0\\FVOk9b0[FYOl9>b2HdQ:"}, "label": "a man without glasses playing the wii"}]}
{"id": 71271, "image": "COCO_train2014_000000071271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a button - up shirt wearing glasses is holding a wiimote\"."}], "task": "refering", "answer_template": "The \"a man in a button - up shirt wearing glasses is holding a wiimote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351, 352, 353, 371, 372, 373, 375, 376], "bbox": [0.145265625, 0.03491666666666667, 0.4054062500000001, 0.989875], "iscrowd": 0, "area": 54123.94729999999, "rle": {"size": [480, 640], "counts": "cQ\\15c>8G9cF\\OU4m0jKYOi3o0WLVO[3S1dLTOm2U1RMQOa2W1^MoNT2Z1kMlNZ1j1eN\\N>\\2AjMBn2=XMQOU3n0QMeNW3Z1oLYNZ3e1mLnMZ3Q2lLbM\\3]2jLVM^3i2hLjL`3U3fL^Lb3a3dLRLd3m3bLfKf3Y4a3O100O1O100O1000000O101O0O10000O10000O1000000O10000O10000O10001O0O10000O10000hM]E2d:_OPF8T:FTF0P:0UFGo98WF^On9b0XFTOi9n0n17I6J7I6J6J7K4L5K4L5K4L5K4M4K4N34K4M4K4M4L3M4L3M3M4L3M4L3N3L3M4L3M4L3M4L3M4L3M4L2OO00000000001O00100O1iDQO[8o0cGTO\\8m0`GVO`8k0]GXOb8i0[GYOe8g0YG\\Of8e0WG^Oh8c0UG_Ok8b0RGAn8>PGDP9=mFFR9;kFGU9:hFIW97fFLZ95cFNc2kNb2X1hJOf2lN_2U1hJ3h2lN\\2R1iJ4j2oNY2n0jJ6l2oNX2k0iJ8o2QOT2_3jMeLS2\\3lMhLP2X3PNkLm1V3QNoLk1Q3UNRMh1o2VNVMf1j2ZNZMb1g2]N]M_1c2`NbMYNaNg0m3o0fMPNUO?U3a1jMgMH6_2Q2nM`M:Mh1c2RNVMn0EQ1T3UNnLj5Q3[JeLl5Z3m204K5L4L4K5L4J6I7J5J7I7Ic]b5"}, "label": "a man in a button - up shirt wearing glasses is holding a wiimote"}]}
{"id": 71271, "image": "COCO_train2014_000000071271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in glasses is playing with his wii\"."}], "task": "refering", "answer_template": "The \"the man in glasses is playing with his wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351, 352, 353, 371, 372, 373, 375, 376], "bbox": [0.145265625, 0.03491666666666667, 0.4054062500000001, 0.989875], "iscrowd": 0, "area": 54123.94729999999, "rle": {"size": [480, 640], "counts": "cQ\\15c>8G9cF\\OU4m0jKYOi3o0WLVO[3S1dLTOm2U1RMQOa2W1^MoNT2Z1kMlNZ1j1eN\\N>\\2AjMBn2=XMQOU3n0QMeNW3Z1oLYNZ3e1mLnMZ3Q2lLbM\\3]2jLVM^3i2hLjL`3U3fL^Lb3a3dLRLd3m3bLfKf3Y4a3O100O1O100O1000000O101O0O10000O10000O1000000O10000O10000O10001O0O10000O10000hM]E2d:_OPF8T:FTF0P:0UFGo98WF^On9b0XFTOi9n0n17I6J7I6J6J7K4L5K4L5K4L5K4M4K4N34K4M4K4M4L3M4L3M3M4L3M4L3N3L3M4L3M4L3M4L3M4L3M4L2OO00000000001O00100O1iDQO[8o0cGTO\\8m0`GVO`8k0]GXOb8i0[GYOe8g0YG\\Of8e0WG^Oh8c0UG_Ok8b0RGAn8>PGDP9=mFFR9;kFGU9:hFIW97fFLZ95cFNc2kNb2X1hJOf2lN_2U1hJ3h2lN\\2R1iJ4j2oNY2n0jJ6l2oNX2k0iJ8o2QOT2_3jMeLS2\\3lMhLP2X3PNkLm1V3QNoLk1Q3UNRMh1o2VNVMf1j2ZNZMb1g2]N]M_1c2`NbMYNaNg0m3o0fMPNUO?U3a1jMgMH6_2Q2nM`M:Mh1c2RNVMn0EQ1T3UNnLj5Q3[JeLl5Z3m204K5L4L4K5L4J6I7J5J7I7Ic]b5"}, "label": "the man in glasses is playing with his wii"}]}
{"id": 194154, "image": "COCO_train2014_000000194154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue table that people are sitting at\"."}], "task": "refering", "answer_template": "The \"a blue table that people are sitting at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 264, 265, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 372, 373, 374, 375, 376, 377, 378, 379, 384, 385, 386], "bbox": [0.170609375, 0.6745416666666666, 0.8243281249999999, 0.989875], "iscrowd": 0, "area": 35133.94709999999, "rle": {"size": [480, 640], "counts": "bac11m>3M3M3M2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2O1O1O1O1N2O1O1O1O2N1000000000000O10000000000O10000000000O100O100O1O1O1O100O1J6M3O2N100O1O1O100O1O100O1O1O100003M2N2N2N1O001O1O001O1O001O1O001O1O010O1O001O1O1O001O1O001O1O001O1OO100000000000000000000O10000000001O000000000O100000000000000000000O10000O100O100O010O0100O010O010O010O010O010O010O10O010I6I80O010O001O001O0O2O1O001O001O001O8H1O2N1O1O1O1O1O1O1O1O1N2O1O2N1O1O1O1O1bMQDj1P<TNSDj1n;TNUDj1l;TNWDj1j;TNXDk1i;SNZDk1g;SN\\Dl1e;PN_Do1Y<10O1O1O1O001O100O001O001O0001O01O00001O0000010O00001O0000010O00010O0001O01O01O01O00010O00010O0001O01O01O01O01O01O01O0001O01O00010O00010O000010O0001O01O0001O01O01O01O01O00010O0000010O002O1N2N2N2N1O2N2N2N2N1O2N2N2OO000000000000000000000000000000001O000000009H9F1O00000000001O0000000000001O00000000001O0000000TORCFn<O_COa<FkC9_=O1O2N1OM4M2N21O1O1O2N1O1O2N1O1O2N1O1O2NYQd1"}, "label": "a blue table that people are sitting at"}]}
{"id": 194154, "image": "COCO_train2014_000000194154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"four girls having their breakfast in the dining table\"."}], "task": "refering", "answer_template": "The \"four girls having their breakfast in the dining table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 264, 265, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 372, 373, 374, 375, 376, 377, 378, 379, 384, 385, 386], "bbox": [0.170609375, 0.6745416666666666, 0.8243281249999999, 0.989875], "iscrowd": 0, "area": 35133.94709999999, "rle": {"size": [480, 640], "counts": "bac11m>3M3M3M2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2O1O1O1O1N2O1O1O1O2N1000000000000O10000000000O10000000000O100O100O1O1O1O100O1J6M3O2N100O1O1O100O1O100O1O1O100003M2N2N2N1O001O1O001O1O001O1O001O1O010O1O001O1O1O001O1O001O1O001O1OO100000000000000000000O10000000001O000000000O100000000000000000000O10000O100O100O010O0100O010O010O010O010O010O010O10O010I6I80O010O001O001O0O2O1O001O001O001O8H1O2N1O1O1O1O1O1O1O1O1N2O1O2N1O1O1O1O1bMQDj1P<TNSDj1n;TNUDj1l;TNWDj1j;TNXDk1i;SNZDk1g;SN\\Dl1e;PN_Do1Y<10O1O1O1O001O100O001O001O0001O01O00001O0000010O00001O0000010O00010O0001O01O01O01O00010O00010O0001O01O01O01O01O01O01O0001O01O00010O00010O000010O0001O01O0001O01O01O01O01O00010O0000010O002O1N2N2N2N1O2N2N2N2N1O2N2N2OO000000000000000000000000000000001O000000009H9F1O00000000001O0000000000001O00000000001O0000000TORCFn<O_COa<FkC9_=O1O2N1OM4M2N21O1O1O2N1O1O2N1O1O2N1O1O2NYQd1"}, "label": "four girls having their breakfast in the dining table"}]}
{"id": 325229, "image": "COCO_train2014_000000325229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing to the right of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe standing to the right of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 33, 34, 46, 47, 59, 60, 72, 73, 85, 86, 97, 98, 99, 110, 111, 112, 122, 123, 124, 125, 134, 135, 136, 137, 138, 147, 148, 149, 150, 151, 160, 161, 162, 163, 164, 173, 174, 175, 176, 177, 186, 187, 188, 189, 199, 200, 201, 202, 212, 213, 214, 215, 225, 226, 227, 228], "bbox": [0.31090666666666666, 0.04932, 0.70552, 0.9686], "iscrowd": 0, "area": 31482.881550000002, "rle": {"size": [500, 375], "counts": "Sci15T;d0XHBf7k0hG\\OV8R1WGUOf8Y1hFnNV9`1XFfNe9h1iE_NU:a3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2N2O1O10O01O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O010O1O1O100N2N2M3N2M3N2M3N2M3N1NO1O2N1O2N1O2N1oNnGbJT8_5WHSJj7n5m00O2eKgI[OY6d0jIYOW6f0lIXOU6e0oIXOR6i0oITOS6l0oIPOR6bNcI0<Z1S6XNQJ=MV1V6oM\\Jk0^OR1X6eMjJX1oNo0Y6\\MVKe1cNi0[6SMbKS2TNf0T9ZOlFa0Y9^OhF>\\9AeF:`9F`F6d9I^F1g9NZFNj91WFJn96RFFR:9oE@`9W1aFaN]9h1eFPNX9Z2hF^MV9k2mFkLQ9^3Z18G:N1O1O2N1O1N2O2N1O1O1fKdK\\M^4V2TLcMo3j1gLQNZ3[1_M_Nc2k0WNoNk1a0hNZOY1`0ROZOP1`0ZOYOh0a0CYO?a0JZO8?3[ON`0<ZOF?d0\\O]O?^1iNeNQ1R8L3M4M3L5K5K5K6J5L5J5K5L\\Uf1"}, "label": "a giraffe standing to the right of another giraffe"}]}
{"id": 325229, "image": "COCO_train2014_000000325229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe next to a shorter giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe next to a shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 33, 34, 46, 47, 59, 60, 72, 73, 85, 86, 97, 98, 99, 110, 111, 112, 122, 123, 124, 125, 134, 135, 136, 137, 138, 147, 148, 149, 150, 151, 160, 161, 162, 163, 164, 173, 174, 175, 176, 177, 186, 187, 188, 189, 199, 200, 201, 202, 212, 213, 214, 215, 225, 226, 227, 228], "bbox": [0.31090666666666666, 0.04932, 0.70552, 0.9686], "iscrowd": 0, "area": 31482.881550000002, "rle": {"size": [500, 375], "counts": "Sci15T;d0XHBf7k0hG\\OV8R1WGUOf8Y1hFnNV9`1XFfNe9h1iE_NU:a3N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2N2O1O10O01O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O010O1O1O100N2N2M3N2M3N2M3N2M3N1NO1O2N1O2N1O2N1oNnGbJT8_5WHSJj7n5m00O2eKgI[OY6d0jIYOW6f0lIXOU6e0oIXOR6i0oITOS6l0oIPOR6bNcI0<Z1S6XNQJ=MV1V6oM\\Jk0^OR1X6eMjJX1oNo0Y6\\MVKe1cNi0[6SMbKS2TNf0T9ZOlFa0Y9^OhF>\\9AeF:`9F`F6d9I^F1g9NZFNj91WFJn96RFFR:9oE@`9W1aFaN]9h1eFPNX9Z2hF^MV9k2mFkLQ9^3Z18G:N1O1O2N1O1N2O2N1O1O1fKdK\\M^4V2TLcMo3j1gLQNZ3[1_M_Nc2k0WNoNk1a0hNZOY1`0ROZOP1`0ZOYOh0a0CYO?a0JZO8?3[ON`0<ZOF?d0\\O]O?^1iNeNQ1R8L3M4M3L5K5K5K6J5L5J5K5L\\Uf1"}, "label": "a giraffe next to a shorter giraffe"}]}
{"id": 325229, "image": "COCO_train2014_000000325229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter giraffe\"."}], "task": "refering", "answer_template": "The \"the shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 58, 59, 71, 72, 84, 85, 97, 98, 109, 110, 121, 122, 123, 132, 133, 134, 144, 145, 146, 147, 157, 158, 159, 160, 170, 171, 172, 173, 183, 184, 185, 196, 197, 209, 210, 222, 223], "bbox": [0.08389333333333333, 0.19832, 0.5932533333333333, 1.0], "iscrowd": 0, "area": 18410.362150000004, "rle": {"size": [500, 375], "counts": "jc?4\\>T1PNQ2ROm0F:M3N3L3N1N2O2M2O1O2O00001O00M3POP1L5M2N2O1N2N2O1N2O1N2N2O1NRDQM\\;n2]DdMY;d3G5K5K4L5K4L5K5K4L5K4M4K5L00O010O1ZNhFPMW9n2UGhLl8T3`GdL`8X3mG_LR8_3ZHWLg7e3fHRLZ7j3Y2N3L4L4L3N3L4L3N3O10O10000O10000O1000O01000000O10000O100O1O00aNjCBU<=cDmN];P1d1O1O1N2O2M2O1O1N2O2M2O1O1N2O1N3N00O2N1N2O1O1100O2N1O2N101O000010O01O0010O0O1N2O1O2M2O1O1N3N1O1N2O2N1N2O1O2F9VOj0WOi0L5O0nN_LQFb3P:aLkE`3U:dLgE\\3Z:kL]EV3d:k02N2N2M3N2N2N2N1O2M3N2N2N2N2M3N1O2N2N2M3N2N1O2N2M3N2N2N1N3J6J6J6J6I6K6J:Fb0^OWcZ2"}, "label": "the shorter giraffe"}]}
{"id": 472686, "image": "COCO_train2014_000000472686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man reading his phone\"."}], "task": "refering", "answer_template": "The \"the man reading his phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 47, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 81, 91, 92, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.10711864406779661, 0.003390625, 0.9947669491525425, 1.0], "iscrowd": 0, "area": 189778.27730000002, "rle": {"size": [640, 472], "counts": "_WP1a8\\;3M3M3M3N2M3M3M3N2M3M3M3M3L4L4L4M3L4L4M3C=oNQ1I7K5L4K5L4L4O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10O10O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O1N2O1O1O1N2SNRBgMo=n1jBcMW=R2\\CdMf<Q2nCeMS<S2Q3N2M3N2M3N2M3N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N200O100O100O10O0100O100O100O100O10000O100O10000O100O10000O100O100000000000000000000000000000000000000000000000000000000000000001O001O1O001O1O001O1O1O001O1O00000000001O0000000000000000000000000000001O00000000001O1O1O2N1O1O1O1UB@T8a0hGCW8?eGDZ8=dGE[8<bGG]8:`GI_88_GJ`87]GLb85\\GMc85YGNf83WGOi82UG0i82TG1k80RG3m8NQG4n8NnF5Q9LlF6T9KjF7U9JhF8X9IeF:Z9GdF:\\9H`F;_9F^F<b9E[F>d9CYF?g9BUFb0j9_OSFc0m9_OoEd0P:]OmEe0S:\\OiEg0W:ZOfEi0Y:XOdEj0\\:WOaEl0^:VO]Em0c:TOZEo0e:ROXEP1h:QOTES1k:QOoDQ1Q;UOgDn0X;XO`Dj0`;\\OWDh0h;^OPDd0P<BhCa0W<E`C>`<b3O00001O001O001O001O001O001O001O001O001O1O1O1O1O1O001O1O1jGYHQ4h7lK^HP4c7nKbHn3_7oKhHl3Y7QLmHk3T7RLRIj3n6ULXIf3i6WL]Ie3d6XLbId3_6ZLfIb3[6[LlI`3U6]LQJ_3P6^LVJ^3k5`L[J[3f5bL`JZ3`5dLfJX3[5eLlJV3U5hLPKT3Q5iLTKT3m4iLYKS3h4kL]KQ3d4lLaKQ3`4lLfKP3Z4nLkKo2V4oLoKm2R4PMTLl2T4jLQLS3X4bLnKZ3Z4]LkK_3^4VLgKg3b4nKdKn3Z:M4L4L3M4L4L3M4L3M4L4L3M4L4L3M4L3M4L4L3M4L4L3M4L3M4K5L3M4L4L3M4L3M4L4L3MXX1"}, "label": "the man reading his phone"}]}
{"id": 472686, "image": "COCO_train2014_000000472686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man who is holding his stuff and looking down at his phone\"."}], "task": "refering", "answer_template": "The \"the man who is holding his stuff and looking down at his phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 47, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 81, 91, 92, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.10711864406779661, 0.003390625, 0.9947669491525425, 1.0], "iscrowd": 0, "area": 189778.27730000002, "rle": {"size": [640, 472], "counts": "_WP1a8\\;3M3M3M3N2M3M3M3N2M3M3M3M3L4L4L4M3L4L4M3C=oNQ1I7K5L4K5L4L4O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10O10O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O100O100O10000O100O1N2O1O1O1N2SNRBgMo=n1jBcMW=R2\\CdMf<Q2nCeMS<S2Q3N2M3N2M3N2M3N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N200O100O100O10O0100O100O100O100O10000O100O10000O100O10000O100O100000000000000000000000000000000000000000000000000000000000000001O001O1O001O1O001O1O1O001O1O00000000001O0000000000000000000000000000001O00000000001O1O1O2N1O1O1O1UB@T8a0hGCW8?eGDZ8=dGE[8<bGG]8:`GI_88_GJ`87]GLb85\\GMc85YGNf83WGOi82UG0i82TG1k80RG3m8NQG4n8NnF5Q9LlF6T9KjF7U9JhF8X9IeF:Z9GdF:\\9H`F;_9F^F<b9E[F>d9CYF?g9BUFb0j9_OSFc0m9_OoEd0P:]OmEe0S:\\OiEg0W:ZOfEi0Y:XOdEj0\\:WOaEl0^:VO]Em0c:TOZEo0e:ROXEP1h:QOTES1k:QOoDQ1Q;UOgDn0X;XO`Dj0`;\\OWDh0h;^OPDd0P<BhCa0W<E`C>`<b3O00001O001O001O001O001O001O001O001O001O1O1O1O1O1O001O1O1jGYHQ4h7lK^HP4c7nKbHn3_7oKhHl3Y7QLmHk3T7RLRIj3n6ULXIf3i6WL]Ie3d6XLbId3_6ZLfIb3[6[LlI`3U6]LQJ_3P6^LVJ^3k5`L[J[3f5bL`JZ3`5dLfJX3[5eLlJV3U5hLPKT3Q5iLTKT3m4iLYKS3h4kL]KQ3d4lLaKQ3`4lLfKP3Z4nLkKo2V4oLoKm2R4PMTLl2T4jLQLS3X4bLnKZ3Z4]LkK_3^4VLgKg3b4nKdKn3Z:M4L4L3M4L4L3M4L3M4L4L3M4L4L3M4L3M4L4L3M4L4L3M4L3M4K5L3M4L4L3M4L3M4L4L3MXX1"}, "label": "the man who is holding his stuff and looking down at his phone"}]}
{"id": 472686, "image": "COCO_train2014_000000472686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with the blond curly hair\"."}], "task": "refering", "answer_template": "The \"person with the blond curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 32, 33, 49, 50, 66, 67, 81, 82, 83, 84, 97, 98, 99, 100, 101, 114, 115], "bbox": [0.7452330508474576, 0.0, 1.0, 0.302953125], "iscrowd": 0, "area": 7266.121099999999, "rle": {"size": [640, 472], "counts": "^Ul64jc06J5K6J5J3N2N2N3M2N2N2N2N2N3M10J7J5K5O2O000O10000O100O100O2O0O100O2O1N2O1N2O1N2O0O2O1N101O0100O1m\\OLZb03e]O2Wb0Oh]O2Xb0Nf]O4Yb0Mf]O4Zb0Le]O5Zb0Le]O5[b0Kd]O6[b0Jd]O8\\b0Hd]O8[b0Ie]O7[b0Ie]O7[b0Ie]O7[b0Ie]O7[b0Ie]O7[b0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Zb0Hf]O8Yb0Ig]O7Yb0Ig]O7Yb0Ig]O7Yb0Ig]O7[`0Nf@KoN7Y`02f@GQO7W`06f@CSO7U`0:e@@VO6R`0?f@[OXO6P`0b0g@XOYO6n?e0g@VO[O5l?h0h@SO\\O5j?j0i@RO]O4g?n0k@nN^O4e?P1m@lN^O4c?S1m@jN@3a?V1m@hNB2^?Y1o@fNC1\\?\\1PAcND1Z?^1QAbNE0W?a1SA`NFOU?c1TA_NGNS?e1UA^NHMQ?g1VA]NILo>i1XA[NILn>j1XA[NJKm>k1YAZNJKk>m1ZAYNKJj>o1YAXNMIi>P2ZAWNMIh>Z3XAfLh>Z3XAfLg>[3YAeLg>[3YAeLg>[3YAeLf>\\3ZAdLf>\\3ZAdLf>\\3ZAdLe>]3[AcLe>]3[AcLd>^3\\AbLd>^3\\AbLd>^3\\AbLc>_3]AaLc>_3]AaLc>_3]AaLc>_3iAULW>"}, "label": "person with the blond curly hair"}]}
{"id": 448115, "image": "COCO_train2014_000000448115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of a black train car meant to carry liquid sitting still on the train tracks\"."}], "task": "refering", "answer_template": "The \"back of a black train car meant to carry liquid sitting still on the train tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 240, 241, 242, 243], "bbox": [0.409859375, 0.24112499999999998, 0.595234375, 0.6107291666666667], "iscrowd": 0, "area": 18077.671199999997, "rle": {"size": [480, 640], "counts": "`Pk31j>9H8G9H8H9F9H8G9H8G:G8F:C=C=C;E;M4M2N2N2N2N3M2N2O1N2N000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000M3F:F:E;Ff0ZOf0ZOg0XOh0YOg0YO_Wi3"}, "label": "back of a black train car meant to carry liquid sitting still on the train tracks"}]}
{"id": 448115, "image": "COCO_train2014_000000448115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the oil car\"."}], "task": "refering", "answer_template": "The \"the back of the oil car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 240, 241, 242, 243], "bbox": [0.409859375, 0.24112499999999998, 0.595234375, 0.6107291666666667], "iscrowd": 0, "area": 18077.671199999997, "rle": {"size": [480, 640], "counts": "`Pk31j>9H8G9H8H9F9H8G9H8G:G8F:C=C=C;E;M4M2N2N2N2N3M2N2O1N2N000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000M3F:F:E;Ff0ZOf0ZOg0XOh0YOg0YO_Wi3"}, "label": "the back of the oil car"}]}
{"id": 448115, "image": "COCO_train2014_000000448115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old looking boxcar\"."}], "task": "refering", "answer_template": "The \"an old looking boxcar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 249, 250, 251, 252], "bbox": [0.62696875, 0.0050625, 1.0, 0.64775], "iscrowd": 0, "area": 46529.13225000001, "rle": {"size": [480, 640], "counts": "_Tl56i>f0ZOe0[Od0]O5J5K6K1N1O2N100O1O2N100O1O2N100O1O1O2O0O1O1O2O0O1O101N1O1O101N1O1O100O2N1O100O2N1O100O2N100O1O2N100O1O2N100O1O1O2O0O1O1O2O0O1O1O2O0O1O101N1O1O100O2N1O100O2N1O100O2N1O100O2N1O100O2N100O1O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O1O101N1O100O2N1O100O2N1O100O2N1O100O2N1O100O1O2N100O1O2O0O1O1O2O0O1O1O2O0O1O1O101N1O1O101N1O1O101N1O100O2N1O100O2N1O100O1O2N100O1O2N100O1O2O0O1O1O2O0O1O1O101N1O1O101N1O1O101N1O1O101N1O100O2N1O100O1O2N100O1O2N100O1O2N100O1O2N100O1O101N1O100O2O0000001N1N"}, "label": "an old looking boxcar"}]}
{"id": 448115, "image": "COCO_train2014_000000448115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large red train on a track\"."}], "task": "refering", "answer_template": "The \"a large red train on a track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 249, 250, 251, 252], "bbox": [0.62696875, 0.0050625, 1.0, 0.64775], "iscrowd": 0, "area": 46529.13225000001, "rle": {"size": [480, 640], "counts": "_Tl56i>f0ZOe0[Od0]O5J5K6K1N1O2N100O1O2N100O1O2N100O1O1O2O0O1O1O2O0O1O101N1O1O101N1O1O100O2N1O100O2N1O100O2N100O1O2N100O1O2N100O1O1O2O0O1O1O2O0O1O1O2O0O1O101N1O1O100O2N1O100O2N1O100O2N1O100O2N1O100O2N100O1O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O1O101N1O100O2N1O100O2N1O100O2N1O100O2N1O100O1O2N100O1O2O0O1O1O2O0O1O1O2O0O1O1O101N1O1O101N1O1O101N1O100O2N1O100O2N1O100O1O2N100O1O2N100O1O2O0O1O1O2O0O1O1O101N1O1O101N1O1O101N1O1O101N1O100O2N1O100O1O2N100O1O2N100O1O2N100O1O2N100O1O101N1O100O2O0000001N1N"}, "label": "a large red train on a track"}]}
{"id": 374391, "image": "COCO_train2014_000000374391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bear near a soda bottle\"."}], "task": "refering", "answer_template": "The \"a brown bear near a soda bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 341, 342], "bbox": [0.613140625, 0.39122916666666663, 0.9016875000000001, 0.8352708333333332], "iscrowd": 0, "area": 28369.03025000001, "rle": {"size": [480, 640], "counts": "Z_h54i>3M4M2M4L3O2N1N3N1O2M4M4K4L4N3M2N2N2N2O1N2N2O1N1O2O1N2PEPNl8Q2mFVNS9k1fF\\NX9f1aF`N_9a1ZFfNd9\\1UFiNk9\\1lEiNT:Z1cEkN]:Y1[EkNe:h210O0001O010O000010O0XOQL_Fo3W9]LhFb3T9cLkF^3R9fLmFY3P9kLoFU3o8oLoFR3n8QMQGo2o8RMPGn2o8UMoFl2P9b101N1000000O1000000000000O1000000O101O0O101O0O10001N1000001O0O101O0000001N100O2O000O101N10000O10O100O100000000000000000000000000000000000001O000010O000001O001O010O001O1O001O00100O001O1O1aMXGdNi8Z1YGeNg8Z1[GeNe8Z1^GdNb8Z1aGeN_8Z1cGeN^8Y1dGfN\\8X1fGhNZ8R1mGmNS8Q1PHnNP8P1RHPOn7n0UHQOk7l0XHTOh7j0[HVOe7e0_H[Oa7`0eH_O[79nHFR72WIMi6O\\I0d6O^I0b6O`I0`6NcI1]6NeI1[6NgI1Y6MkI1V6MnI0S6NPJ0P6NUJ0k5NYJOh5N\\J0e5M`J0a5KeJ3_5EPK0T5G]5DRZm0"}, "label": "a brown bear near a soda bottle"}]}
{"id": 374391, "image": "COCO_train2014_000000374391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a without hairy brown color teddy bear\"."}], "task": "refering", "answer_template": "The \"a without hairy brown color teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 341, 342], "bbox": [0.613140625, 0.39122916666666663, 0.9016875000000001, 0.8352708333333332], "iscrowd": 0, "area": 28369.03025000001, "rle": {"size": [480, 640], "counts": "Z_h54i>3M4M2M4L3O2N1N3N1O2M4M4K4L4N3M2N2N2N2O1N2N2O1N1O2O1N2PEPNl8Q2mFVNS9k1fF\\NX9f1aF`N_9a1ZFfNd9\\1UFiNk9\\1lEiNT:Z1cEkN]:Y1[EkNe:h210O0001O010O000010O0XOQL_Fo3W9]LhFb3T9cLkF^3R9fLmFY3P9kLoFU3o8oLoFR3n8QMQGo2o8RMPGn2o8UMoFl2P9b101N1000000O1000000000000O1000000O101O0O101O0O10001N1000001O0O101O0000001N100O2O000O101N10000O10O100O100000000000000000000000000000000000001O000010O000001O001O010O001O1O001O00100O001O1O1aMXGdNi8Z1YGeNg8Z1[GeNe8Z1^GdNb8Z1aGeN_8Z1cGeN^8Y1dGfN\\8X1fGhNZ8R1mGmNS8Q1PHnNP8P1RHPOn7n0UHQOk7l0XHTOh7j0[HVOe7e0_H[Oa7`0eH_O[79nHFR72WIMi6O\\I0d6O^I0b6O`I0`6NcI1]6NeI1[6NgI1Y6MkI1V6MnI0S6NPJ0P6NUJ0k5NYJOh5N\\J0e5M`J0a5KeJ3_5EPK0T5G]5DRZm0"}, "label": "a without hairy brown color teddy bear"}]}
{"id": 374391, "image": "COCO_train2014_000000374391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear with a brown ribbon around the neck\"."}], "task": "refering", "answer_template": "The \"a teddy bear with a brown ribbon around the neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 233, 234, 235, 236, 237, 253, 256, 257, 258, 259, 276, 277, 279, 280, 281, 282, 299, 300], "bbox": [0.018453125, 0.24406250000000002, 0.47146875, 0.7854583333333333], "iscrowd": 0, "area": 33805.97545, "rle": {"size": [480, 640], "counts": "hl54j>7H7J6J6J7H7J6J7I6[DRNo9S2gESNX:R2^EUN`:n0XFXOg9e0lEhNSOi0P;<UFLi90YF3f9J^F7`9FdF:[9CiF>V9_OnF`0Q9@PGa0n8@RG>o8BRG9R9HnF4U9LmFNW93iFH[98fFC^9>bF]Ob9c0_FXOe9i0X20010O01O001O001O0100O10O10O100O10O0100O100O010O100000O01000000O01000O10O1000O1000O10O100000O010000O11N9H9G8H9G6I6K6K5J6J6J5K6kMjLTH\\3d7QMQHU3g7XMnGm2k7XMRHh2l7\\MRHd2l7aMQH`2l7dMRH\\2l7iMQHW2m7mMQHS2m7RNPHn1n7^NiGb1T8f2O1N2N3M2O1N2N2O2M2N2O1O101O00000O2O000000001N100000001N100001O01O0001O01O0001O0001O01O000001O0000001O000000001O000000001O00O1L4L4L4L4M3L4L4L4L4M3O1N1O2N2N2O0O1O1O2N3N1N2N2N2O1N3M2N2O1O1O1O2N1O1O1O1O2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2M3M3N2M3N2M3M3N3L4L4M3L4M3L3M4M3L4L4M3N1O2N2N2N3N1N3M2N3M2N000001O000000000000001O0000000000001O000000000O100103L4L3M4L3M4I7I6IVWn4"}, "label": "a teddy bear with a brown ribbon around the neck"}]}
{"id": 374391, "image": "COCO_train2014_000000374391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear on the left\"."}], "task": "refering", "answer_template": "The \"teddy bear on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 233, 234, 235, 236, 237, 253, 256, 257, 258, 259, 276, 277, 279, 280, 281, 282, 299, 300], "bbox": [0.018453125, 0.24406250000000002, 0.47146875, 0.7854583333333333], "iscrowd": 0, "area": 33805.97545, "rle": {"size": [480, 640], "counts": "hl54j>7H7J6J6J7H7J6J7I6[DRNo9S2gESNX:R2^EUN`:n0XFXOg9e0lEhNSOi0P;<UFLi90YF3f9J^F7`9FdF:[9CiF>V9_OnF`0Q9@PGa0n8@RG>o8BRG9R9HnF4U9LmFNW93iFH[98fFC^9>bF]Ob9c0_FXOe9i0X20010O01O001O001O0100O10O10O100O10O0100O100O010O100000O01000000O01000O10O1000O1000O10O100000O010000O11N9H9G8H9G6I6K6K5J6J6J5K6kMjLTH\\3d7QMQHU3g7XMnGm2k7XMRHh2l7\\MRHd2l7aMQH`2l7dMRH\\2l7iMQHW2m7mMQHS2m7RNPHn1n7^NiGb1T8f2O1N2N3M2O1N2N2O2M2N2O1O101O00000O2O000000001N100000001N100001O01O0001O01O0001O0001O01O000001O0000001O000000001O000000001O00O1L4L4L4L4M3L4L4L4L4M3O1N1O2N2N2O0O1O1O2N3N1N2N2N2O1N3M2N2O1O1O1O2N1O1O1O1O2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2M3M3N2M3N2M3M3N3L4L4M3L4M3L3M4M3L4L4M3N1O2N2N2N3N1N3M2N3M2N000001O000000000000001O0000000000001O000000000O100103L4L3M4L3M4I7I6IVWn4"}, "label": "teddy bear on the left"}]}
{"id": 521847, "image": "COCO_train2014_000000521847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue plaid looking at a horse\"."}], "task": "refering", "answer_template": "The \"a man wearing blue plaid looking at a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 114, 115, 116, 131, 132, 133, 134, 150, 151, 168, 169, 186, 187], "bbox": [0.32245999999999997, 0.34656000000000003, 0.45812, 0.7942933333333334], "iscrowd": 0, "area": 5942.771499999997, "rle": {"size": [375, 500], "counts": "^Uk11d;2N2N2WOJoE9n9IoE;o9GlE<S:FiE<X:FcE=\\:?O0000O10000OO21O2N1N2OaEL_9\\1VHnMb5X2WJkMg5Z2SJhMY5NdI^2n0fM[51cI^2l0bMb5i3gIXLX6d400O10O0100001O0O2O001O1mNhIeLZ6S3QJgLW6l2VJmLo5h2d10O8I7H9G8I;D;GO0N2O2N1O1N3N1PMQIV1P7WNdHF`1Q1n5XOgH_O81X1:o55jHN[1CQ6?WH:m9f02O2M2N3L5kNRF9W\\S3"}, "label": "a man wearing blue plaid looking at a horse"}]}
{"id": 521847, "image": "COCO_train2014_000000521847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a plaid shirt next to the horse\"."}], "task": "refering", "answer_template": "The \"a man in a plaid shirt next to the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 114, 115, 116, 131, 132, 133, 134, 150, 151, 168, 169, 186, 187], "bbox": [0.32245999999999997, 0.34656000000000003, 0.45812, 0.7942933333333334], "iscrowd": 0, "area": 5942.771499999997, "rle": {"size": [375, 500], "counts": "^Uk11d;2N2N2WOJoE9n9IoE;o9GlE<S:FiE<X:FcE=\\:?O0000O10000OO21O2N1N2OaEL_9\\1VHnMb5X2WJkMg5Z2SJhMY5NdI^2n0fM[51cI^2l0bMb5i3gIXLX6d400O10O0100001O0O2O001O1mNhIeLZ6S3QJgLW6l2VJmLo5h2d10O8I7H9G8I;D;GO0N2O2N1O1N3N1PMQIV1P7WNdHF`1Q1n5XOgH_O81X1:o55jHN[1CQ6?WH:m9f02O2M2N3L5kNRF9W\\S3"}, "label": "a man in a plaid shirt next to the horse"}]}
{"id": 251523, "image": "COCO_train2014_000000251523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a persons right hand reaching for the donut\"."}], "task": "refering", "answer_template": "The \"a persons right hand reaching for the donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 248, 249, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 352, 353, 354, 355, 356, 366, 367, 368, 369, 370, 371, 372, 373, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.500375, 0.507859375, 0.99475, 0.9887656249999999], "iscrowd": 0, "area": 37120.118650000004, "rle": {"size": [640, 480], "counts": "_Zf4f0Zc0000O10001N1[EXOR2h0mMYOS2g0lMZOT2f0kM[OT2f0kM[OU2f0iM[OW2e0hM\\OX2d0hM\\OW2e0hM\\OX2d0gM]OY2c0fM^OY2d0eM]O[2c0dM^O\\2b0cM_O]2a0bM@]2a0bM@^2a0`M@`2`0VFROa6?Y3?VFRO`6`0Y3?WFRO^6`0[3>VFSO^6`0\\3=VFTO\\6`0]3>UFSO]6`0^3=UFTO[6`0`3<UFTOZ6a0a3;TFVOY6?c3<TFUOX6`0d3<RFVOX6?f3;RFVOW6`0g3;QFUOW6a0g3;QFVOV6`0i3;PFUOV6a0j3:PFVOU6`0k3;nEVOV6`0m39mEXOT6`0o39kEXOV6?o39kEYOT6?Q49jEXOU6?R48hE[OU6=S49gEZOV6=S49fE\\OU6<U49eE\\OU6;W48dE^OT6:X49bE_OU68Y4:aE_OT68[49aE@S67\\4W1cKiN^4W1aKiN_4X1_KiNa4X1^KhNb4Y1]KgNd4Y1[KgNe4Z1YKgNg4[1WKeNi4\\1VKdNk4\\1TKdNl4]1SKcNm4^1QKcNo4^1PKbNP5_1oJaNR5_1mJaNS5`1kJaNU5`1jJ`NV5a1iJ_NX5a1gJ_NY5b1eJ_N[5b1dJ^N\\5b1dJ^N]5a1cJ_N]5a1bJ`N^5`1bJ`N^5`1bJ`N^5a1aJ_N`5`1_JaNa5_1_JaNa5_1_JaNa5_1_JaNb5^1]JcNc5]1]JcNc5]1]JcNc5]1]JcNd5]1[JcNe5]1ZJdNf5\\1ZJdNf5\\1ZJdNf5\\1ZJdNg5[1XJfNh5Z1XJfNi5Y1WJgNj5Y1UJgNk5Y1TJhNm5W1SJiNn5V1RJjNo5U1QJkNP6T1oImNR6R1nInNS6Q1mIoNT6P1lIPOU6P1iIQOW6o0iIQOX6n0hIROY6m0fITO[6k0dIVO]6i0bIXO_6i0_IWOb6i0\\IXOe6i0XIXOh6i0VIWOl6i0RIXOo6i0nHXOS7h0kHYOV7h0gHYOZ7g0eHYO\\7h0aHYO`7g0^HZOc7g0ZHZOf7g0XHZOi7f0UH[Ol7f0QH[OP8e0nG\\OS8e0kG[OV8e0hG\\OY8e0dG\\O]8d0aG]O_8d0_G]Ob8d0[G]Of8c0XG^Oi8c0TG^Om8b0RG^Oo8e0lF\\OU9X1UFiNl9l1]EUNc:`2gDaMZ;T3oCmLR<`51O1O001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O001O00001N10001O00001O001O00001O00001O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1Oo1QNV2jMYZ1"}, "label": "a persons right hand reaching for the donut"}]}
{"id": 251523, "image": "COCO_train2014_000000251523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand holding the donut\"."}], "task": "refering", "answer_template": "The \"the hand holding the donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 248, 249, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 339, 352, 353, 354, 355, 356, 366, 367, 368, 369, 370, 371, 372, 373, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.500375, 0.507859375, 0.99475, 0.9887656249999999], "iscrowd": 0, "area": 37120.118650000004, "rle": {"size": [640, 480], "counts": "_Zf4f0Zc0000O10001N1[EXOR2h0mMYOS2g0lMZOT2f0kM[OT2f0kM[OU2f0iM[OW2e0hM\\OX2d0hM\\OW2e0hM\\OX2d0gM]OY2c0fM^OY2d0eM]O[2c0dM^O\\2b0cM_O]2a0bM@]2a0bM@^2a0`M@`2`0VFROa6?Y3?VFRO`6`0Y3?WFRO^6`0[3>VFSO^6`0\\3=VFTO\\6`0]3>UFSO]6`0^3=UFTO[6`0`3<UFTOZ6a0a3;TFVOY6?c3<TFUOX6`0d3<RFVOX6?f3;RFVOW6`0g3;QFUOW6a0g3;QFVOV6`0i3;PFUOV6a0j3:PFVOU6`0k3;nEVOV6`0m39mEXOT6`0o39kEXOV6?o39kEYOT6?Q49jEXOU6?R48hE[OU6=S49gEZOV6=S49fE\\OU6<U49eE\\OU6;W48dE^OT6:X49bE_OU68Y4:aE_OT68[49aE@S67\\4W1cKiN^4W1aKiN_4X1_KiNa4X1^KhNb4Y1]KgNd4Y1[KgNe4Z1YKgNg4[1WKeNi4\\1VKdNk4\\1TKdNl4]1SKcNm4^1QKcNo4^1PKbNP5_1oJaNR5_1mJaNS5`1kJaNU5`1jJ`NV5a1iJ_NX5a1gJ_NY5b1eJ_N[5b1dJ^N\\5b1dJ^N]5a1cJ_N]5a1bJ`N^5`1bJ`N^5`1bJ`N^5a1aJ_N`5`1_JaNa5_1_JaNa5_1_JaNa5_1_JaNb5^1]JcNc5]1]JcNc5]1]JcNc5]1]JcNd5]1[JcNe5]1ZJdNf5\\1ZJdNf5\\1ZJdNf5\\1ZJdNg5[1XJfNh5Z1XJfNi5Y1WJgNj5Y1UJgNk5Y1TJhNm5W1SJiNn5V1RJjNo5U1QJkNP6T1oImNR6R1nInNS6Q1mIoNT6P1lIPOU6P1iIQOW6o0iIQOX6n0hIROY6m0fITO[6k0dIVO]6i0bIXO_6i0_IWOb6i0\\IXOe6i0XIXOh6i0VIWOl6i0RIXOo6i0nHXOS7h0kHYOV7h0gHYOZ7g0eHYO\\7h0aHYO`7g0^HZOc7g0ZHZOf7g0XHZOi7f0UH[Ol7f0QH[OP8e0nG\\OS8e0kG[OV8e0hG\\OY8e0dG\\O]8d0aG]O_8d0_G]Ob8d0[G]Of8c0XG^Oi8c0TG^Om8b0RG^Oo8e0lF\\OU9X1UFiNl9l1]EUNc:`2gDaMZ;T3oCmLR<`51O1O001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O001O00001N10001O00001O001O00001O00001O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1Oo1QNV2jMYZ1"}, "label": "the hand holding the donut"}]}
{"id": 251523, "image": "COCO_train2014_000000251523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the persons legs\"."}], "task": "refering", "answer_template": "The \"the persons legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 238, 239, 240, 241, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 323, 324, 340, 357], "bbox": [0.0, 0.21945312499999997, 0.6920208333333334, 0.951], "iscrowd": 0, "area": 55435.66295, "rle": {"size": [640, 480], "counts": "R?m3P9Q72N2N2N1O2N2M3N2N2N2N2N1O2M3N2N2N2N2N1O2M3N2N2N2N2N2N1N3N2N2N2N2N2N2M3N3M2N2N2N2N3L3N2N2N2N2N3M2M3N2N2N2N3M2N2M3N2N3M2N2N2N2M3N3M2N2N2N2M4L3L4M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3L4M3N2O1O1O100O1O1O1O10fNUL_Bj3a=bLTB]3k=oLkAP3U>VMfAj2Y>XMfAg2Y>[MgAd2Y>^MfAa2Y>aMgA_2X>cMgA\\2Y>eMgA[2X>gMgAY2Y>gMgAY2X>iMgAW2X>jMhAV2X>kMgAU2Y>kMgAU2Z>kMeAU2[>kMeAU2\\>kMcAU2]>kMcAU2^>kMaAU2_>kMaAU2`>kM_AU2a>kM_AU2b>kM]AU2d>jM\\AV2d>kM[AU2f>jMZAV2f>b11O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1OdMcKUE]4i:WLfDh3W;]LgDc3W;aLhD^3V;fLhDZ3X;iLeDW3\\;jLcDU3];mLaDS3_;oL_DQ3a;RM]Dm2c;UM[Dk2e;WMYDi2g;YMXDe2i;]MUDc2k;`MRD`2n;bMQD\\2P<fMnCZ2R<hMmCW2S<lMjCT2V<nMhCQ2Y<QNfCn1Z<TNdCl1\\<VNbCj1^<YN`Ce1a<]N]Cc1c<_N[Ca1e<aNZC^1f<m22O000O1000000O2O1O1O1N2O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O105J;E;E;E8H100O1O1O1O1O100O1O1O1O1O100O1O1O1\\NUAQNl>n1UAQNl>o1SARNm>m1SASNn>l1SASNn>m1QASNP?l1PATNQ?k1PAUNP?j1PAVNQ?j1n@VNS?i1n@VNS?i1m@WNT?h1l@XNU?h1j@YNV?f1k@YNV?f1j@ZNW?e1i@[NX?e1h@ZNY?e1g@[NZ?d1f@]NZ?c1f@\\N[?c1e@]N\\?b1d@^N]?a1d@^N]?b1b@^N_?a1a@`N_?_1b@`N_?_1a@aN`?_1_@aNb?^1^@bNc?]1^@aNd?^1\\@bNe?^1Z@aNh?^1Y@aNh?^1X@aNj?_1U@`Nm?_1T@`Nm?_1S@`No?_1Q@aNP`0_1P@_NR`0`1n_O_NT`0`1l_O`NU`0_1l_O_NW`0`1i_O_NX`0`1j_O]NX`0b1i_O]NX`0c1S1O1O2N1O100O1O2N1O1O1O100O2N1O1O1O1O2O0O1O1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O1O2O3L4L^Ul2"}, "label": "the persons legs"}]}
{"id": 251523, "image": "COCO_train2014_000000251523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blck stripped pajama\"."}], "task": "refering", "answer_template": "The \"a man wearing blck stripped pajama\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 238, 239, 240, 241, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 323, 324, 340, 357], "bbox": [0.0, 0.21945312499999997, 0.6920208333333334, 0.951], "iscrowd": 0, "area": 55435.66295, "rle": {"size": [640, 480], "counts": "R?m3P9Q72N2N2N1O2N2M3N2N2N2N2N1O2M3N2N2N2N2N1O2M3N2N2N2N2N2N1N3N2N2N2N2N2N2M3N3M2N2N2N2N3L3N2N2N2N2N3M2M3N2N2N2N3M2N2M3N2N3M2N2N2N2M3N3M2N2N2N2M4L3L4M3M3L4M3M3M3L4M3M3L4M3M3M3L4M3M3L4M3N2O1O1O100O1O1O1O10fNUL_Bj3a=bLTB]3k=oLkAP3U>VMfAj2Y>XMfAg2Y>[MgAd2Y>^MfAa2Y>aMgA_2X>cMgA\\2Y>eMgA[2X>gMgAY2Y>gMgAY2X>iMgAW2X>jMhAV2X>kMgAU2Y>kMgAU2Z>kMeAU2[>kMeAU2\\>kMcAU2]>kMcAU2^>kMaAU2_>kMaAU2`>kM_AU2a>kM_AU2b>kM]AU2d>jM\\AV2d>kM[AU2f>jMZAV2f>b11O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1OdMcKUE]4i:WLfDh3W;]LgDc3W;aLhD^3V;fLhDZ3X;iLeDW3\\;jLcDU3];mLaDS3_;oL_DQ3a;RM]Dm2c;UM[Dk2e;WMYDi2g;YMXDe2i;]MUDc2k;`MRD`2n;bMQD\\2P<fMnCZ2R<hMmCW2S<lMjCT2V<nMhCQ2Y<QNfCn1Z<TNdCl1\\<VNbCj1^<YN`Ce1a<]N]Cc1c<_N[Ca1e<aNZC^1f<m22O000O1000000O2O1O1O1N2O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O105J;E;E;E8H100O1O1O1O1O100O1O1O1O1O100O1O1O1\\NUAQNl>n1UAQNl>o1SARNm>m1SASNn>l1SASNn>m1QASNP?l1PATNQ?k1PAUNP?j1PAVNQ?j1n@VNS?i1n@VNS?i1m@WNT?h1l@XNU?h1j@YNV?f1k@YNV?f1j@ZNW?e1i@[NX?e1h@ZNY?e1g@[NZ?d1f@]NZ?c1f@\\N[?c1e@]N\\?b1d@^N]?a1d@^N]?b1b@^N_?a1a@`N_?_1b@`N_?_1a@aN`?_1_@aNb?^1^@bNc?]1^@aNd?^1\\@bNe?^1Z@aNh?^1Y@aNh?^1X@aNj?_1U@`Nm?_1T@`Nm?_1S@`No?_1Q@aNP`0_1P@_NR`0`1n_O_NT`0`1l_O`NU`0_1l_O_NW`0`1i_O_NX`0`1j_O]NX`0b1i_O]NX`0c1S1O1O2N1O100O1O2N1O1O1O100O2N1O1O1O1O2O0O1O1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O1O2O3L4L^Ul2"}, "label": "a man wearing blck stripped pajama"}]}
{"id": 448131, "image": "COCO_train2014_000000448131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman cross country skiing in a orange jacket and brown ski pants\"."}], "task": "refering", "answer_template": "The \"a young woman cross country skiing in a orange jacket and brown ski pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 189, 190, 211, 212, 213, 214, 234, 235, 236, 237, 258, 259, 281, 282, 304, 305], "bbox": [0.20746875, 0.46880000000000005, 0.322796875, 0.978875], "iscrowd": 0, "area": 8497.345299999997, "rle": {"size": [400, 640], "counts": "_Wd15W<5K6K3M4M2N2N2H8H6J421N2PHaNc45dH\\1g2bNc4k1]KUNj32kIi1[2VNe39lIb1^2VNe3;jI_1a2VNe3`0eI[1f2UNd3h2nI\\Me1L]4R4QKoKn4a4bJ_K_5h4YJYKg5l4SJTKn5o4nIRKR6Q5jIPKV6e5hIRJn5m5RJTJn5l5QJUJo5k5PJWJo5i5PJWJQ6h5oIYJR6f5mIZJU6d5jI^JV6b5jI]JX6b5gI_JY6k53N2N3L;F1O1O1N10001N100O2eJnHV5V7O2O0O100O2O000O5L4nMQ2_Ob0L=E:D5K5K6J9ROZD`0n;L5L2M2O2MdWY5"}, "label": "a young woman cross country skiing in a orange jacket and brown ski pants"}]}
{"id": 448131, "image": "COCO_train2014_000000448131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with an orange coat wearing skis\"."}], "task": "refering", "answer_template": "The \"person with an orange coat wearing skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 189, 190, 211, 212, 213, 214, 234, 235, 236, 237, 258, 259, 281, 282, 304, 305], "bbox": [0.20746875, 0.46880000000000005, 0.322796875, 0.978875], "iscrowd": 0, "area": 8497.345299999997, "rle": {"size": [400, 640], "counts": "_Wd15W<5K6K3M4M2N2N2H8H6J421N2PHaNc45dH\\1g2bNc4k1]KUNj32kIi1[2VNe39lIb1^2VNe3;jI_1a2VNe3`0eI[1f2UNd3h2nI\\Me1L]4R4QKoKn4a4bJ_K_5h4YJYKg5l4SJTKn5o4nIRKR6Q5jIPKV6e5hIRJn5m5RJTJn5l5QJUJo5k5PJWJo5i5PJWJQ6h5oIYJR6f5mIZJU6d5jI^JV6b5jI]JX6b5gI_JY6k53N2N3L;F1O1O1N10001N100O2eJnHV5V7O2O0O100O2O000O5L4nMQ2_Ob0L=E:D5K5K6J9ROZD`0n;L5L2M2O2MdWY5"}, "label": "person with an orange coat wearing skis"}]}
{"id": 407173, "image": "COCO_train2014_000000407173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wine glass sitting on the left side of the table by the white cup\"."}], "task": "refering", "answer_template": "The \"the wine glass sitting on the left side of the table by the white cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 257, 258, 274, 275, 291, 292, 307, 308, 309, 324, 325, 326], "bbox": [0.056937499999999995, 0.503515625, 0.2846458333333333, 0.84959375], "iscrowd": 0, "area": 11901.949500000002, "rle": {"size": [640, 480], "counts": "WWa01kc09F;F9F;I6J6J7I2N2N2N2N2nAlMU:V2jEkMT:W2jEkMT:W2jEkMT:W2kEjMT:W2jEkMU:V2iEmMU:S2jEoMU:R2jEoMT:S2jEoMU:R2jEoMU:R2jEnMV:S2iEnMV:S2hEnMW:T2hElMX:U2gElMX:U2gEkMY:V2fEjMZ:W2eEjMY:X2fEhMZ:X2fEiMY:X2fEhMZ:X2eEiM[:X2dEiMZ:Y2eEgM[:Y2eEhMZ:Y2eEgM[:Z2dEfM\\:Z2dEgMZ:[2eEdM\\:]2cEcM]:]2cEbM^:_2aE`M`:a2_E_M`:k3WDTLj;]4eCbK\\<n4TCRKl<P610O01O100O2N100O100\\OeBbJ\\=R5PCnJP=g4\\CXKd<U4oCjKS<R3PEnLP;n2TERMl:n2TERMl:m2UESMk:l2VESMk:l2VETMj:l2VETMj:k2WEUMi:j2XEUMj:i2WEWMi:i2XEVMh:i2YEWMg:h2ZEWMg:h2ZEXMf:h2ZEXMf:g2\\EXMd:g2]EYMc:f2_EXMb:g2_EYMa:f2aEYM_:g2aEYM_:f2cEXM^:g2cEYM]:f2eEYM[:f2fEZMZ:e2hEZMX:e2iEZMX:e2b3O1O1O1O100O1N3M4L4L5K4L4M3L5K4L4L4L5KSQf6"}, "label": "the wine glass sitting on the left side of the table by the white cup"}]}
{"id": 407173, "image": "COCO_train2014_000000407173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rounded wineglass behind the plate\"."}], "task": "refering", "answer_template": "The \"rounded wineglass behind the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 205, 206, 207, 208, 222, 223, 224, 225, 239, 240, 241, 242, 257, 258, 274, 275, 291, 292, 307, 308, 309, 324, 325, 326], "bbox": [0.056937499999999995, 0.503515625, 0.2846458333333333, 0.84959375], "iscrowd": 0, "area": 11901.949500000002, "rle": {"size": [640, 480], "counts": "WWa01kc09F;F9F;I6J6J7I2N2N2N2N2nAlMU:V2jEkMT:W2jEkMT:W2jEkMT:W2kEjMT:W2jEkMU:V2iEmMU:S2jEoMU:R2jEoMT:S2jEoMU:R2jEoMU:R2jEnMV:S2iEnMV:S2hEnMW:T2hElMX:U2gElMX:U2gEkMY:V2fEjMZ:W2eEjMY:X2fEhMZ:X2fEiMY:X2fEhMZ:X2eEiM[:X2dEiMZ:Y2eEgM[:Y2eEhMZ:Y2eEgM[:Z2dEfM\\:Z2dEgMZ:[2eEdM\\:]2cEcM]:]2cEbM^:_2aE`M`:a2_E_M`:k3WDTLj;]4eCbK\\<n4TCRKl<P610O01O100O2N100O100\\OeBbJ\\=R5PCnJP=g4\\CXKd<U4oCjKS<R3PEnLP;n2TERMl:n2TERMl:m2UESMk:l2VESMk:l2VETMj:l2VETMj:k2WEUMi:j2XEUMj:i2WEWMi:i2XEVMh:i2YEWMg:h2ZEWMg:h2ZEXMf:h2ZEXMf:g2\\EXMd:g2]EYMc:f2_EXMb:g2_EYMa:f2aEYM_:g2aEYM_:f2cEXM^:g2cEYM]:f2eEYM[:f2fEZMZ:e2hEZMX:e2iEZMX:e2b3O1O1O1O100O1N3M4L4L5K4L4M3L5K4L4L4L5KSQf6"}, "label": "rounded wineglass behind the plate"}]}
{"id": 407173, "image": "COCO_train2014_000000407173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass with red wine in it , directly in front of a dark skinned woman with black hair in a black dress\"."}], "task": "refering", "answer_template": "The \"a wine glass with red wine in it , directly in front of a dark skinned woman with black hair in a black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 213, 214, 215, 216, 230, 231, 232, 233, 247, 248, 249, 250, 264, 265, 266, 267, 282, 283, 299, 316, 333, 350, 367, 384], "bbox": [0.5423125, 0.501125, 0.7340833333333333, 0.9842656250000001], "iscrowd": 0, "area": 12889.258749999999, "rle": {"size": [640, 480], "counts": "YkR5k1ha0>J7L3N3L3M3N3L3N3L3N2M3N2M3N2O1O1O0O2O1O1O1N2O2N2N2N2N1O2N2N2SESLc5o3jIaLW6a3UIQMk6Q3bH_M_7c2nGmMS8[2SGVNn8_2lEQNU:m50O1O1O1O100O1YOg0ZOf0ZOf0ZOf0ZOf0ZOf0ZOf0E;O100O100O100O1O2O0O100O2O0O100O101N100O101N100O101N1O1M3M4L6J6K5J7I6J6I8G8H8HQe_2"}, "label": "a wine glass with red wine in it , directly in front of a dark skinned woman with black hair in a black dress"}]}
{"id": 407173, "image": "COCO_train2014_000000407173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman facing the camera\"."}], "task": "refering", "answer_template": "The \"the woman facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 287, 288, 304, 305], "bbox": [0.29958333333333337, 0.19924999999999998, 0.9744791666666667, 0.759734375], "iscrowd": 0, "area": 66866.44685000001, "rle": {"size": [640, 480], "counts": "Q^j25kc06I7J3L100O010O100O100O100O100O100O010O100O100O1O100O100O010O100O100O100O100O010O010N101N101N100O2N101N101N100O2O0M4E:F;E:U_O\\MQ`0o2j_OVMo?P3l_OUMR`0m2j_OXMS`0\\3L4M3M3M2M4M3M3M3L4M2N3L4M3M3M3O100O100O100O1O100O100O100O1dLaLWF_3h9cLVF^3i9dLVF\\3i9fLUF\\3i9fLUF[3j9gLUFY3k9hLSFY3l9iLSFW3l9kLRFV3m9lLRFT3m9nLQFS3n9oLQFQ3n9QMPFP3P:QMnEP3Q:RMnEn2Q:TMmEm2R:UMmEk2R:WMmEi2R:YMmEg2R:[MmEf2R:[MmEe2R:]MmEc2R:_MmEa2R:aMmE_2R:cMmE]2R:eMmE\\2R:eMmE[2R:gMmEY2R:iMmEW2R:kMmEU2R:mMmES2R:o3N100O10000O2O0O100O100O2O000O100O101N10000O101N100O100O101O0O100O10O010O001O10O01O0010O01O1O010O001O010O010001O0001O01O01O00010O000010O01O000010O0001O01O01O00010O001O01O01O00010O00001O01O01O0010O0003M4M3L4L5K4M3L4L4L4L4M3L4L4K6J5K7I8H8H7I8H8H7I8H8H5D9@`0J6K4L5K5J6K5K5K5001O1O1O001O1N2O1O001O1O1O1O0O2N2N2O1N1O2N2N;F;D=C4L01O0000001O0000001O0000001O00000010O000007I=C<D;E5K5K5K6J5K5K5F:F:G:Ec0]OX1iNWR7"}, "label": "the woman facing the camera"}]}
{"id": 407173, "image": "COCO_train2014_000000407173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young attractive hispanic woman\"."}], "task": "refering", "answer_template": "The \"a young attractive hispanic woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 177, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 287, 288, 304, 305], "bbox": [0.29958333333333337, 0.19924999999999998, 0.9744791666666667, 0.759734375], "iscrowd": 0, "area": 66866.44685000001, "rle": {"size": [640, 480], "counts": "Q^j25kc06I7J3L100O010O100O100O100O100O100O010O100O100O1O100O100O010O100O100O100O100O010O010N101N101N100O2N101N101N100O2O0M4E:F;E:U_O\\MQ`0o2j_OVMo?P3l_OUMR`0m2j_OXMS`0\\3L4M3M3M2M4M3M3M3L4M2N3L4M3M3M3O100O100O100O1O100O100O100O1dLaLWF_3h9cLVF^3i9dLVF\\3i9fLUF\\3i9fLUF[3j9gLUFY3k9hLSFY3l9iLSFW3l9kLRFV3m9lLRFT3m9nLQFS3n9oLQFQ3n9QMPFP3P:QMnEP3Q:RMnEn2Q:TMmEm2R:UMmEk2R:WMmEi2R:YMmEg2R:[MmEf2R:[MmEe2R:]MmEc2R:_MmEa2R:aMmE_2R:cMmE]2R:eMmE\\2R:eMmE[2R:gMmEY2R:iMmEW2R:kMmEU2R:mMmES2R:o3N100O10000O2O0O100O100O2O000O100O101N10000O101N100O100O101O0O100O10O010O001O10O01O0010O01O1O010O001O010O010001O0001O01O01O00010O000010O01O000010O0001O01O01O00010O001O01O01O00010O00001O01O01O0010O0003M4M3L4L5K4M3L4L4L4L4M3L4L4K6J5K7I8H8H7I8H8H7I8H8H5D9@`0J6K4L5K5J6K5K5K5001O1O1O001O1N2O1O001O1O1O1O0O2N2N2O1N1O2N2N;F;D=C4L01O0000001O0000001O0000001O00000010O000007I=C<D;E5K5K5K6J5K5K5F:F:G:Ec0]OX1iNWR7"}, "label": "a young attractive hispanic woman"}]}
{"id": 128647, "image": "COCO_train2014_000000128647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with glasses and middle length blond hair standing with arms crossed\"."}], "task": "refering", "answer_template": "The \"person with glasses and middle length blond hair standing with arms crossed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 219, 220, 242, 265, 288, 310, 311, 333, 334, 356, 357], "bbox": [0.500953125, 0.2957916666666666, 0.6555625, 0.9021666666666666], "iscrowd": 0, "area": 8789.727850000003, "rle": {"size": [480, 640], "counts": "[kf4:e>;F3L3M2O2M3K4L5`MfNSG]1Y8eNnE2b1]1[8dNTF2X1`1\\7ZNcG`0JJo0a1`7YNdG>MLm0[1_7^NeG=02e0R1b7bNgG;3f00<S8fNhG:5d02WO@9]8FiG88b04SOI4o72jG7:`05PO3O`7=lG6<=8nNh8h1QGWO<mNc8n1QGRO`0nN^8R2RGmNY:U1WFXNj9i1`FkMb9V2c1100O1O1O1O100O1O1N2O1N1O2O1N2O1N2O1O1O00001O0O2M2O2N2N1O2L2KO02033L4M2N10020000000O10000000000000000000000000000O1000000O3N3M2M4M3L4L3M4L6I6F:UOU1SOanV3"}, "label": "person with glasses and middle length blond hair standing with arms crossed"}]}
{"id": 128647, "image": "COCO_train2014_000000128647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing wearing a black t - shirt\"."}], "task": "refering", "answer_template": "The \"a woman standing wearing a black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 219, 220, 242, 265, 288, 310, 311, 333, 334, 356, 357], "bbox": [0.500953125, 0.2957916666666666, 0.6555625, 0.9021666666666666], "iscrowd": 0, "area": 8789.727850000003, "rle": {"size": [480, 640], "counts": "[kf4:e>;F3L3M2O2M3K4L5`MfNSG]1Y8eNnE2b1]1[8dNTF2X1`1\\7ZNcG`0JJo0a1`7YNdG>MLm0[1_7^NeG=02e0R1b7bNgG;3f00<S8fNhG:5d02WO@9]8FiG88b04SOI4o72jG7:`05PO3O`7=lG6<=8nNh8h1QGWO<mNc8n1QGRO`0nN^8R2RGmNY:U1WFXNj9i1`FkMb9V2c1100O1O1O1O100O1O1N2O1N1O2O1N2O1N2O1O1O00001O0O2M2O2N2N1O2L2KO02033L4M2N10020000000O10000000000000000000000000000O1000000O3N3M2M4M3L4L3M4L6I6F:UOU1SOanV3"}, "label": "a woman standing wearing a black t - shirt"}]}
{"id": 128647, "image": "COCO_train2014_000000128647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was sitting\"."}], "task": "refering", "answer_template": "The \"a man was sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 204, 205, 206, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.606734375, 0.3344583333333333, 1.0, 0.9883958333333333], "iscrowd": 0, "area": 39609.783500000005, "rle": {"size": [480, 640], "counts": "ZZf52b>?I8H7L5K1O100O1O1O100O1N2N2N200O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000O1N2O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2O1N2N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2M3N2N2N2N2jL`JjLb5U3aJgLa5Y3`JdLb5[3aJaLa5^3aJ_La5`3bJ\\L`5c3bJ\\L^5c3cJ\\L^5d3aJ]L_5b3bJ]L_5c3aJ]L_5c3aJ\\L`5d3`J[La5d3_J]La5c3_J\\Lb5d3^J\\Lb5d3^J[Lc5e3]J[Lc5d3_JZLb5f3aJVL`5j3cJSL]5m3fJoK[5Q4UK_Kk4`4\\KYKe4g4`KTK`4l4fKmJ[4S5gKjJZ4U5iKiJW4W5kKfJV4Z5lKdJT4\\5nKaJS4_5oK_JQ4`5QL^JP4b5RL[Jo3d5SL[Jm3d5VLYJk3f5WLYJi3f5i200O1O1O1O1O1O100O1O1O1O1O100O1O1O1000000O1000000000000000000000000O1000O100000000000000000O100000000000000000B>eK[4dK[O"}, "label": "a man was sitting"}]}
{"id": 128647, "image": "COCO_train2014_000000128647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sitting man in an orange hat\"."}], "task": "refering", "answer_template": "The \"a sitting man in an orange hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 204, 205, 206, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.606734375, 0.3344583333333333, 1.0, 0.9883958333333333], "iscrowd": 0, "area": 39609.783500000005, "rle": {"size": [480, 640], "counts": "ZZf52b>?I8H7L5K1O100O1O1O100O1N2N2N200O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000O1N2O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2O1N2N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2M3N2N2N2N2jL`JjLb5U3aJgLa5Y3`JdLb5[3aJaLa5^3aJ_La5`3bJ\\L`5c3bJ\\L^5c3cJ\\L^5d3aJ]L_5b3bJ]L_5c3aJ]L_5c3aJ\\L`5d3`J[La5d3_J]La5c3_J\\Lb5d3^J\\Lb5d3^J[Lc5e3]J[Lc5d3_JZLb5f3aJVL`5j3cJSL]5m3fJoK[5Q4UK_Kk4`4\\KYKe4g4`KTK`4l4fKmJ[4S5gKjJZ4U5iKiJW4W5kKfJV4Z5lKdJT4\\5nKaJS4_5oK_JQ4`5QL^JP4b5RL[Jo3d5SL[Jm3d5VLYJk3f5WLYJi3f5i200O1O1O1O1O1O100O1O1O1O1O100O1O1O1000000O1000000000000000000000000O1000O100000000000000000O100000000000000000B>eK[4dK[O"}, "label": "a sitting man in an orange hat"}]}
{"id": 128647, "image": "COCO_train2014_000000128647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a gray shirt talking to two other people\"."}], "task": "refering", "answer_template": "The \"a man in a gray shirt talking to two other people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.023458333333333335, 0.351515625, 0.9836041666666666], "iscrowd": 0, "area": 47937.301200000016, "rle": {"size": [480, 640], "counts": "Z11i:a2jF`MQ4e7nK\\HnNh<Q1d0N2O1O1N2O1O1N2O1O1N2O1O1O2N100O1O100O1O100O1O100O1O100O1O100000000001O000000000000000000000000001nE[Og5e0XJCa5=]JGa5:]JHb58]JKa55]JOa52]J1a5O^J4`5L`J6^5K`J7_5IaJ9]5GbJ<\\5DdJ>Z5CdJ`0Z5@fJa0Y5_OfJd0X5]OgJd0X5\\OhJe0X5ZOhJh0V5YOhJi0W5WOiJk0U5UOkJl0T5UOkJm0S5SOmJn0R5ROmJP1R5POnJZ1h4gNWK^1d4cN[K_1c4bN\\K`1b4aN\\Kb1b4_N]Kc1a4]N^Kf1`4[N_Kg1_4ZN_Kj1^4WN`Ko1[4RNdKS2W4nMgKY2S4hMkK^2P4bMPLa2m3`MQLa2o3`MoKa2Q4`MmKa2S4`MlK`2T4aMaKi2_4XM^Kj2b4VM]Kk2c4VM[Kk2e4VMYKk2g4VMVKl2gNeKV5`1QLl2iNhKl1e0Dh0El2kNjKg1Z1SO28k2mNmKd1h1fNDe0h2POnKa1T2^NYOm0e2UOnK_1[2YNVO7TNK]45oK^1\\2TN@6nM0V49PL[1]2mMO6dM5Q4=nK=ECc6@iM`0PL<J]Oa6]OlMk0iK;NWOV;>lD:o;GQD9o;GQD9n;HRD7o;IQD7n;JRD5o;LPD4o;MQD2o;OQD1o;OQD0o;1QDOo;1PDOP<2PDNP<2PDMQ<3oCMP<4PDLP<5oCJQ<7oCIQ<7oCHR<8nCHQ<9oCFR<:nCFQ<;oCDR<<nCDR<_OcC`0:1S<_OdCa080T<]OgCb051U<[OgCd040W<ZOgCd032X<VOgCh011f<NZC2f<M[C2g<MYC3g<LZC3h<LWC5i<JXC5i<KWC5j<JVC5k<JUC7k<IUC7j<IWC6j<JUC7j<IVC7k<IUC7k<ITC7l<IUC7k<ITC8l<GTC9l<HTC8l<GTC:l<FSC;m<ESC;m<DSC<n<CSC=n<@SCa0m<^OTCb0l<]OUCc0l<[OUCd0l<[OTCf0l<YOUCg0k<WOWCi0j<UOWCk0i<TOWCl0_=00000O010O1O0O2O1O1N101O1N101O11O0O2O1O0O10O100O010O10O101O0O100O0000001O1O1O2O0O1O1O1O1O1^OiA:Y>CiA;Y>CjA9a>M2N3LgXR6"}, "label": "a man in a gray shirt talking to two other people"}]}
{"id": 128647, "image": "COCO_train2014_000000128647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a grey shirt\"."}], "task": "refering", "answer_template": "The \"a man in a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.023458333333333335, 0.351515625, 0.9836041666666666], "iscrowd": 0, "area": 47937.301200000016, "rle": {"size": [480, 640], "counts": "Z11i:a2jF`MQ4e7nK\\HnNh<Q1d0N2O1O1N2O1O1N2O1O1N2O1O1O2N100O1O100O1O100O1O100O1O100O1O100000000001O000000000000000000000000001nE[Og5e0XJCa5=]JGa5:]JHb58]JKa55]JOa52]J1a5O^J4`5L`J6^5K`J7_5IaJ9]5GbJ<\\5DdJ>Z5CdJ`0Z5@fJa0Y5_OfJd0X5]OgJd0X5\\OhJe0X5ZOhJh0V5YOhJi0W5WOiJk0U5UOkJl0T5UOkJm0S5SOmJn0R5ROmJP1R5POnJZ1h4gNWK^1d4cN[K_1c4bN\\K`1b4aN\\Kb1b4_N]Kc1a4]N^Kf1`4[N_Kg1_4ZN_Kj1^4WN`Ko1[4RNdKS2W4nMgKY2S4hMkK^2P4bMPLa2m3`MQLa2o3`MoKa2Q4`MmKa2S4`MlK`2T4aMaKi2_4XM^Kj2b4VM]Kk2c4VM[Kk2e4VMYKk2g4VMVKl2gNeKV5`1QLl2iNhKl1e0Dh0El2kNjKg1Z1SO28k2mNmKd1h1fNDe0h2POnKa1T2^NYOm0e2UOnK_1[2YNVO7TNK]45oK^1\\2TN@6nM0V49PL[1]2mMO6dM5Q4=nK=ECc6@iM`0PL<J]Oa6]OlMk0iK;NWOV;>lD:o;GQD9o;GQD9n;HRD7o;IQD7n;JRD5o;LPD4o;MQD2o;OQD1o;OQD0o;1QDOo;1PDOP<2PDNP<2PDMQ<3oCMP<4PDLP<5oCJQ<7oCIQ<7oCHR<8nCHQ<9oCFR<:nCFQ<;oCDR<<nCDR<_OcC`0:1S<_OdCa080T<]OgCb051U<[OgCd040W<ZOgCd032X<VOgCh011f<NZC2f<M[C2g<MYC3g<LZC3h<LWC5i<JXC5i<KWC5j<JVC5k<JUC7k<IUC7j<IWC6j<JUC7j<IVC7k<IUC7k<ITC7l<IUC7k<ITC8l<GTC9l<HTC8l<GTC:l<FSC;m<ESC;m<DSC<n<CSC=n<@SCa0m<^OTCb0l<]OUCc0l<[OUCd0l<[OTCf0l<YOUCg0k<WOWCi0j<UOWCk0i<TOWCl0_=00000O010O1O0O2O1O1N101O1N101O11O0O2O1O0O10O100O010O10O101O0O100O0000001O1O1O2O0O1O1O1O1O1^OiA:Y>CiA;Y>CjA9a>M2N3LgXR6"}, "label": "a man in a grey shirt"}]}
{"id": 505479, "image": "COCO_train2014_000000505479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse being led by it ' s trainer presenting the horse with a saddle\"."}], "task": "refering", "answer_template": "The \"a horse being led by it ' s trainer presenting the horse with a saddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 176, 177, 178, 180], "bbox": [0.26527083333333334, 0.27166144200626957, 0.6816875, 0.990564263322884], "iscrowd": 0, "area": 29325.70665, "rle": {"size": [319, 480], "counts": "`oW16e95K5N3O001O001N10001O001O000O2M6K7H8kIbNZ3f1YLaNf3e1oKbNn3f1fK`NY4g1ZKaNc4f1RK`Nl4d1kJdNS5_1eJgNY5]1fJcNW5a1hJ^NW5e1gJ\\NV5X3N3N1N3M2O2M3N1N3M2O2N1O2M4M3M2N3L4M2N3M3L4M2N3M3L3N3M2N0O1000000O1000000O11O001\\NQMVLo2j3RMULn2k3SMTLn2k3RMULn2k3QMVLo2j3QMVLP3i3PMWLP3h3PMYLQ3f3oLZLQ3f3nL[LR3e3nL[LS3d3mL\\LS3V2mLaM2:Q3U2SMWM0d0n2T2XMmLNP1j2T2]MfNc2Z1_MdNb2[1^MeNb2[1^MeNb2[1^MeNb2[1^MeNa2\\1_MdN`2]1`McN`2]1`McN_2_1`MaN_2`1aM`N^2a1bM_N^2a1cM^N\\2c1dM]N[2d1eM\\N[2d1eM\\NZ2e1fM[NY2g1fMUN]2l1cMoMc2P2]MlMh2S2XMhMn2W2RMdMT3[2lLaMY3^2gL^M^3a2bL^M`3b2`L]Mb3a2^L_Mc3`2]L`Md3_2\\L`Mf3_2ZLaMg3^2YLbMh3]2XLcMi3\\2WLcMi3^2WLbMi3_2VLaMi3`2WL`Mh3a2XL_Mh3a2XL^Mh3c2YL\\Mg3d2YL\\Mf3e2ZL[Mf3e2ZLZMf3h2YLXMg3h2YLXMf3i2ZLWMf3i2ZLWMe3j2[LUMe3l2[LTMe3l2[LTMd3m2\\LSMd3m2\\LRMd3P3\\LoLd3Q3\\LoLc3R3]LnLc3R3]LmLc3S3^LmLb3R3_LnL`3R3aLnL^3S3bLmL^3R3cLmL]3S3dLmL\\3R3eLnLZ3R3gLnLY3Q3hLPMW3o2kLQMT3n2mLSMS3k2nLWMP3h2QMYMo2e2RM\\Mm2c2TM^Ml2`2UMbMi2]2XMdMh2Z2YMgMf2X2[MjMc2T2^MnMb2n1`MTN_2i1dMXN\\2d1fM_NX2^1kMcNU2Y1mMiNS2S1oMPOT2h0oM[OS2=oMFR24QNNP2LRN7k1GWN<g1CZN?d1A[Nc0b1\\O^Nd0c1\\O]Nc0d1\\O\\NPOTN=h3d0WNkNTN3Q4R1mMhNUNHZ4_1eMdNS3\\1PM`NQ3_1RM^Nn2c1UMYNl2f1g200O1000O01QOo0J6N2N2N1O2N2N2N2N1O2NPb_1"}, "label": "a horse being led by it ' s trainer presenting the horse with a saddle"}]}
{"id": 505479, "image": "COCO_train2014_000000505479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse being led by an equestrian\"."}], "task": "refering", "answer_template": "The \"a horse being led by an equestrian\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 176, 177, 178, 180], "bbox": [0.26527083333333334, 0.27166144200626957, 0.6816875, 0.990564263322884], "iscrowd": 0, "area": 29325.70665, "rle": {"size": [319, 480], "counts": "`oW16e95K5N3O001O001N10001O001O000O2M6K7H8kIbNZ3f1YLaNf3e1oKbNn3f1fK`NY4g1ZKaNc4f1RK`Nl4d1kJdNS5_1eJgNY5]1fJcNW5a1hJ^NW5e1gJ\\NV5X3N3N1N3M2O2M3N1N3M2O2N1O2M4M3M2N3L4M2N3M3L4M2N3M3L3N3M2N0O1000000O1000000O11O001\\NQMVLo2j3RMULn2k3SMTLn2k3RMULn2k3QMVLo2j3QMVLP3i3PMWLP3h3PMYLQ3f3oLZLQ3f3nL[LR3e3nL[LS3d3mL\\LS3V2mLaM2:Q3U2SMWM0d0n2T2XMmLNP1j2T2]MfNc2Z1_MdNb2[1^MeNb2[1^MeNb2[1^MeNb2[1^MeNa2\\1_MdN`2]1`McN`2]1`McN_2_1`MaN_2`1aM`N^2a1bM_N^2a1cM^N\\2c1dM]N[2d1eM\\N[2d1eM\\NZ2e1fM[NY2g1fMUN]2l1cMoMc2P2]MlMh2S2XMhMn2W2RMdMT3[2lLaMY3^2gL^M^3a2bL^M`3b2`L]Mb3a2^L_Mc3`2]L`Md3_2\\L`Mf3_2ZLaMg3^2YLbMh3]2XLcMi3\\2WLcMi3^2WLbMi3_2VLaMi3`2WL`Mh3a2XL_Mh3a2XL^Mh3c2YL\\Mg3d2YL\\Mf3e2ZL[Mf3e2ZLZMf3h2YLXMg3h2YLXMf3i2ZLWMf3i2ZLWMe3j2[LUMe3l2[LTMe3l2[LTMd3m2\\LSMd3m2\\LRMd3P3\\LoLd3Q3\\LoLc3R3]LnLc3R3]LmLc3S3^LmLb3R3_LnL`3R3aLnL^3S3bLmL^3R3cLmL]3S3dLmL\\3R3eLnLZ3R3gLnLY3Q3hLPMW3o2kLQMT3n2mLSMS3k2nLWMP3h2QMYMo2e2RM\\Mm2c2TM^Ml2`2UMbMi2]2XMdMh2Z2YMgMf2X2[MjMc2T2^MnMb2n1`MTN_2i1dMXN\\2d1fM_NX2^1kMcNU2Y1mMiNS2S1oMPOT2h0oM[OS2=oMFR24QNNP2LRN7k1GWN<g1CZN?d1A[Nc0b1\\O^Nd0c1\\O]Nc0d1\\O\\NPOTN=h3d0WNkNTN3Q4R1mMhNUNHZ4_1eMdNS3\\1PM`NQ3_1RM^Nn2c1UMYNl2f1g200O1000O01QOo0J6N2N2N1O2N2N2N2N1O2NPb_1"}, "label": "a horse being led by an equestrian"}]}
{"id": 505479, "image": "COCO_train2014_000000505479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back end of a horse behind another force\"."}], "task": "refering", "answer_template": "The \"the back end of a horse behind another force\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 88, 89, 90, 104, 105, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 155, 158], "bbox": [0.1486458333333333, 0.42166144200626954, 0.313, 0.8452978056426331], "iscrowd": 0, "area": 5676.608800000004, "rle": {"size": [319, 480], "counts": "]df02k91hF1_81[G5d8LWG9h8IQG=n8:N2lG[On6f0PI]On6d0RI]Ol6d0TI\\Oi6k0SIVO^6W1cIjNY6X1hIiNV6W1kIiNS6X1nIiNo5X1RJhNl5X1VJgNj5Y1UJiNi5X1SJmNk5T1QJQOm5P1oITOR6j0iI]OX6a0bIF^69dIF\\69fIE\\68gIGY68iIGW68kIGU67nIEV69lI^O\\6a0fIVOb6`0iI[O\\68QJGo5=nIBR6b0kI]OU6g0eI[O\\6h0_IYOa6m0WIUOj6P1nHROS7T1eHmN\\7j11O100O1O100O100O100O1O010O100O1O100O1OnHWNP6h1jIhNm5V1QJWOf5h0WJDa5;[JJe54YJ1f5NVJ7j5o1002O4K4]N_J`Nf5U1dJgNk5h0\\JSOn5a0WJ[Oo5>UJ@P69UJCQ65UJHP61beV3"}, "label": "the back end of a horse behind another force"}]}
{"id": 505479, "image": "COCO_train2014_000000505479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rear end of a horse with a person riding on it\"."}], "task": "refering", "answer_template": "The \"the rear end of a horse with a person riding on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 88, 89, 90, 104, 105, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 155, 158], "bbox": [0.1486458333333333, 0.42166144200626954, 0.313, 0.8452978056426331], "iscrowd": 0, "area": 5676.608800000004, "rle": {"size": [319, 480], "counts": "]df02k91hF1_81[G5d8LWG9h8IQG=n8:N2lG[On6f0PI]On6d0RI]Ol6d0TI\\Oi6k0SIVO^6W1cIjNY6X1hIiNV6W1kIiNS6X1nIiNo5X1RJhNl5X1VJgNj5Y1UJiNi5X1SJmNk5T1QJQOm5P1oITOR6j0iI]OX6a0bIF^69dIF\\69fIE\\68gIGY68iIGW68kIGU67nIEV69lI^O\\6a0fIVOb6`0iI[O\\68QJGo5=nIBR6b0kI]OU6g0eI[O\\6h0_IYOa6m0WIUOj6P1nHROS7T1eHmN\\7j11O100O1O100O100O100O1O010O100O1O100O1OnHWNP6h1jIhNm5V1QJWOf5h0WJDa5;[JJe54YJ1f5NVJ7j5o1002O4K4]N_J`Nf5U1dJgNk5h0\\JSOn5a0WJ[Oo5>UJ@P69UJCQ65UJHP61beV3"}, "label": "the rear end of a horse with a person riding on it"}]}
{"id": 562826, "image": "COCO_train2014_000000562826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back view of man in black shirts head\"."}], "task": "refering", "answer_template": "The \"back view of man in black shirts head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 38, 39, 40, 41, 42, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 365, 366, 367], "bbox": [0.62190625, 0.035958333333333335, 0.9977499999999999, 0.9213541666666667], "iscrowd": 0, "area": 54349.049649999986, "rle": {"size": [480, 640], "counts": "Xhj51n>1O2N100O2N1O2N1O2]MF]F;a9I\\F8b9L\\F5a9O\\F2b92[F0b94[FMb98[FJa9;\\FFa9?\\FC`9b0]F_O_9g0_FYO^9m0^FUO^9P1_FQO^9T1_FnN]9X1_FiN]9]1`FeN\\9a1`F`N]9e1`F\\N]9i1aFXN[9n1aFSN[9S2bFoMZ9V2cFkMZ9[2bFgMZ9^2cFcM\\9a2`FaM^9Q400O2N1O2O0O2N1O2N100O2N1O2O0O2N1O2N100O2N1O2O0O2N1O2N10001O001O001N10001O001O001O0O2O00001O001O0O2O0010O00010O010O010O01O01O010O010O01O01O010O0101N2N3N2M3N1N3N2M2N3N2M3N1N3N2M2N3N2M2O2M3N2N101O1O001N2O001O1O10O5L4N2N2N2N2O1N2N2N00M3M3L4M3M3M3K5L4K5K6K4K1O01O00001O00001O001N10001O00001O00001O00001O00001N10001O00001O00001O00001O000O2O001O00001O00001O00001O00001N10001O00001O00001O00001O000O2O001O00001O00001O00k9"}, "label": "back view of man in black shirts head"}]}
{"id": 562826, "image": "COCO_train2014_000000562826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt sits behind a man in a green shirt\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt sits behind a man in a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 38, 39, 40, 41, 42, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 365, 366, 367], "bbox": [0.62190625, 0.035958333333333335, 0.9977499999999999, 0.9213541666666667], "iscrowd": 0, "area": 54349.049649999986, "rle": {"size": [480, 640], "counts": "Xhj51n>1O2N100O2N1O2N1O2]MF]F;a9I\\F8b9L\\F5a9O\\F2b92[F0b94[FMb98[FJa9;\\FFa9?\\FC`9b0]F_O_9g0_FYO^9m0^FUO^9P1_FQO^9T1_FnN]9X1_FiN]9]1`FeN\\9a1`F`N]9e1`F\\N]9i1aFXN[9n1aFSN[9S2bFoMZ9V2cFkMZ9[2bFgMZ9^2cFcM\\9a2`FaM^9Q400O2N1O2O0O2N1O2N100O2N1O2O0O2N1O2N100O2N1O2O0O2N1O2N10001O001O001N10001O001O001O0O2O00001O001O0O2O0010O00010O010O010O01O01O010O010O01O01O010O0101N2N3N2M3N1N3N2M2N3N2M3N1N3N2M2N3N2M2O2M3N2N101O1O001N2O001O1O10O5L4N2N2N2N2O1N2N2N00M3M3L4M3M3M3K5L4K5K6K4K1O01O00001O00001O001N10001O00001O00001O00001O00001N10001O00001O00001O00001O000O2O001O00001O00001O00001O00001N10001O00001O00001O00001O000O2O001O00001O00001O00k9"}, "label": "a man in a black shirt sits behind a man in a green shirt"}]}
{"id": 562826, "image": "COCO_train2014_000000562826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"closed umbrella second from the left\"."}], "task": "refering", "answer_template": "The \"closed umbrella second from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 24, 25, 26, 27, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 141, 164, 187, 210], "bbox": [0.07614062499999999, 0.0010625, 0.242, 0.5400416666666666], "iscrowd": 0, "area": 14167.875150000002, "rle": {"size": [480, 640], "counts": "nof0=S>c0@a0L3M4L4L3M4L3M4L4L4L5K4UEWMW9n2aFWM]9m2]FVMb9o2WFUMg9P3QFTMn9P3lETMR:Q3fETMX:k3O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O00O1O1O100O1O100O1h2XM000000O100000000WOZIWIg6m5[JmIe5X5[KcJe4Z5b2O100O100O1O1O1O1O1O1O1O1O1N2O1N2O1O1N2K5\\KkEW4[:iKkEm3l:I7I7J6I7I7I6J5K5L4K6K4L4K5L5K4K5L4L4K6K4L4L4K5LRWS7"}, "label": "closed umbrella second from the left"}]}
{"id": 562826, "image": "COCO_train2014_000000562826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second umbrella from the left\"."}], "task": "refering", "answer_template": "The \"the second umbrella from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 24, 25, 26, 27, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 141, 164, 187, 210], "bbox": [0.07614062499999999, 0.0010625, 0.242, 0.5400416666666666], "iscrowd": 0, "area": 14167.875150000002, "rle": {"size": [480, 640], "counts": "nof0=S>c0@a0L3M4L4L3M4L3M4L4L4L5K4UEWMW9n2aFWM]9m2]FVMb9o2WFUMg9P3QFTMn9P3lETMR:Q3fETMX:k3O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O00O1O1O100O1O100O1h2XM000000O100000000WOZIWIg6m5[JmIe5X5[KcJe4Z5b2O100O100O1O1O1O1O1O1O1O1O1N2O1N2O1O1N2K5\\KkEW4[:iKkEm3l:I7I7J6I7I7I6J5K5L4K6K4L4K5L5K4K5L4L4K6K4L4L4K5LRWS7"}, "label": "the second umbrella from the left"}]}
{"id": 358029, "image": "COCO_train2014_000000358029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet on the right\"."}], "task": "refering", "answer_template": "The \"the toilet on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 163, 164, 165, 166, 175, 176, 177, 178, 187, 188, 189, 199, 200, 201], "bbox": [0.4948948948948949, 0.49238, 0.908978978978979, 0.9318399999999999], "iscrowd": 0, "area": 21378.18195, "rle": {"size": [500, 333], "counts": "Pj`24`?5J6J6K5J6J5L5J6J5L4L4L4L4L4L4L3M4L4L4L4L3M001O001O6J8H7I7I7I6J1O1O1O1O1O1O1O1O2N1O1O1[F[KY8f4fG_KU8b4jGcKQ8^4mGhKn7Y4QHkKk7V4SHnKj7S4UHmKk7T4SHmKm7U4QHkKo7V4oGkKQ8V4hGPLX8Q4aGUL_8l3ZGZLf8Z500000O100000000000000000000000000000000000000000000O100000000000000000000000000000000000000O100O1aNVGTLj8k3[GPLf8o3`GlK`8S4fGhKZ8V4mGdKT8[4QHaKo7^4WH\\Kj7c4\\HXKd7g4bHSK_7l4e1N2O1O1O1O1N2N2N2N2O1N2N2N2QNUESOm:a0cEYO_::RFXOX:<WFSOW:a0d2F:O4K5Kbl>"}, "label": "the toilet on the right"}]}
{"id": 358029, "image": "COCO_train2014_000000358029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a commode on the right side at the bottom of steps\"."}], "task": "refering", "answer_template": "The \"this is a commode on the right side at the bottom of steps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 163, 164, 165, 166, 175, 176, 177, 178, 187, 188, 189, 199, 200, 201], "bbox": [0.4948948948948949, 0.49238, 0.908978978978979, 0.9318399999999999], "iscrowd": 0, "area": 21378.18195, "rle": {"size": [500, 333], "counts": "Pj`24`?5J6J6K5J6J5L5J6J5L4L4L4L4L4L4L3M4L4L4L4L3M001O001O6J8H7I7I7I6J1O1O1O1O1O1O1O1O2N1O1O1[F[KY8f4fG_KU8b4jGcKQ8^4mGhKn7Y4QHkKk7V4SHnKj7S4UHmKk7T4SHmKm7U4QHkKo7V4oGkKQ8V4hGPLX8Q4aGUL_8l3ZGZLf8Z500000O100000000000000000000000000000000000000000000O100000000000000000000000000000000000000O100O1aNVGTLj8k3[GPLf8o3`GlK`8S4fGhKZ8V4mGdKT8[4QHaKo7^4WH\\Kj7c4\\HXKd7g4bHSK_7l4e1N2O1O1O1O1N2N2N2N2O1N2N2N2QNUESOm:a0cEYO_::RFXOX:<WFSOW:a0d2F:O4K5Kbl>"}, "label": "this is a commode on the right side at the bottom of steps"}]}
{"id": 358029, "image": "COCO_train2014_000000358029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left side toilet\"."}], "task": "refering", "answer_template": "The \"left side toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 157, 158, 159, 160, 161, 170, 171, 172, 173, 182, 183, 184, 194, 195, 196], "bbox": [0.06762762762762763, 0.50226, 0.503873873873874, 0.9369400000000001], "iscrowd": 0, "area": 22077.7693, "rle": {"size": [500, 333], "counts": "\\_;`0T?9G9F:G:F9G7I2N1O2N2N1O2N2N1O2N2N2N1O2N2iCZM\\;g2YDeMe;W3N3M2M3N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1aFbKf7_4UHkKe7V4WHmKi7T4RHPLm7R4oGQLQ8P4jGTLV8m3eGWL[8j3aGYL_8g3]G]Lc8d3XG`Lh8a3TGbLl8S50000000000000000000O1000000000O100000000000000000000000000000O100000000000O1000000000000000000000TNZGjLf8V3cGaL]8^3mGYLS8f3VHRLj7m3ZHoKg7Q4[HmKe7R4_HkKa7T4bHjK_7T4dHjK\\7V4fHhKZ7W4iHgKW7X4mHeKS7Z4PIdKP7\\4RIbKn6]4UIaKk6^4V2O1N2M3L4L4M3L4L4K5^ORDXMT<g2=N2N2O1N2O1N2J6H9G:F9G:F:F9G:FUZ`2"}, "label": "left side toilet"}]}
{"id": 358029, "image": "COCO_train2014_000000358029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white toilet sitting to the left of another white toilet\"."}], "task": "refering", "answer_template": "The \"a white toilet sitting to the left of another white toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 157, 158, 159, 160, 161, 170, 171, 172, 173, 182, 183, 184, 194, 195, 196], "bbox": [0.06762762762762763, 0.50226, 0.503873873873874, 0.9369400000000001], "iscrowd": 0, "area": 22077.7693, "rle": {"size": [500, 333], "counts": "\\_;`0T?9G9F:G:F9G7I2N1O2N2N1O2N2N1O2N2N2N1O2N2iCZM\\;g2YDeMe;W3N3M2M3N2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1aFbKf7_4UHkKe7V4WHmKi7T4RHPLm7R4oGQLQ8P4jGTLV8m3eGWL[8j3aGYL_8g3]G]Lc8d3XG`Lh8a3TGbLl8S50000000000000000000O1000000000O100000000000000000000000000000O100000000000O1000000000000000000000TNZGjLf8V3cGaL]8^3mGYLS8f3VHRLj7m3ZHoKg7Q4[HmKe7R4_HkKa7T4bHjK_7T4dHjK\\7V4fHhKZ7W4iHgKW7X4mHeKS7Z4PIdKP7\\4RIbKn6]4UIaKk6^4V2O1N2M3L4L4M3L4L4K5^ORDXMT<g2=N2N2O1N2O1N2J6H9G:F9G:F:F9G:FUZ`2"}, "label": "a white toilet sitting to the left of another white toilet"}]}
{"id": 546447, "image": "COCO_train2014_000000546447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman jackie who is riding a horse of brown color\"."}], "task": "refering", "answer_template": "The \"a woman jackie who is riding a horse of brown color\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 39, 40, 41, 55, 56, 57, 58, 70, 71, 72, 73, 74, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 151, 152, 153, 154, 169, 170, 185, 186, 201, 202, 203, 218, 219, 234, 235, 250], "bbox": [0.37203125, 0.05859375, 0.7105580357142857, 0.657546875], "iscrowd": 0, "area": 25766.283350000005, "rle": {"size": [640, 448], "counts": "o`X34ic05L5K3M1N3N1O1\\]O@ha0b0U^ODfa0=X^OIba09\\^OKaa07]^OL`a05^^OO\\a06a^OOTa09j^OKTa05j^OOTa0`1O1N2O2Me0[O3N2M4M2M3M3N2M3N2M2N1_NcK_C]4`<fK]C[4a<iK\\CX4c<jK[CW4d<lKYCU4f<nKXCR4g<QLVCQ4g<SLVCn3i<VLSCk3l<XLQCi3n<ZLPCf3o<^LmBc3Q=aLlB`3S=cLjB^3U=fLgB[3X=hLfBX3Y=h101O000000000000000O10001O000000000000001O2N2N2N2N2N2N1O2N2N2N2N2OmI`Cf4^<ZKhCb4V<]KnCb4P<^KPDd4n;\\KSDe4k;ZKVDg4i;YKWDi4h;UKYDm4g;QKYDQ5h;lJXDV5h;gJYD[5g;cJZD^5f;`JZDb5f;[J[Dg5f;kIeDW6a<2N1O2O1N2N2N2N2N2N2O1N3M4L4L3M4L3NhJmCe2o;[MVDc2g;]M_Da2];^MiDa2S;_MRE_2l:`M[E]2a:bMgEZ2W:eMPFX2m9fM[FV2c9iMdFT2X9lMQGo1l8oMbGg1\\8XNQHa1j7]NcH\\1X7_NPId1k6UN^Im1^6mMiIW2R6hMQJV2o5hMUJU2k5kMXJR2P6fMSJW2[6[MhIb2f6oL^In2m;M3M3M2N4L4L4L3M4L4`Nc^O3ca0Ai^O9]a0[Oo^O`0\\b0K6J6K[k`2"}, "label": "a woman jackie who is riding a horse of brown color"}]}
{"id": 546447, "image": "COCO_train2014_000000546447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with the number 21 on his shirt\"."}], "task": "refering", "answer_template": "The \"a man with the number 21 on his shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 39, 40, 41, 55, 56, 57, 58, 70, 71, 72, 73, 74, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 151, 152, 153, 154, 169, 170, 185, 186, 201, 202, 203, 218, 219, 234, 235, 250], "bbox": [0.37203125, 0.05859375, 0.7105580357142857, 0.657546875], "iscrowd": 0, "area": 25766.283350000005, "rle": {"size": [640, 448], "counts": "o`X34ic05L5K3M1N3N1O1\\]O@ha0b0U^ODfa0=X^OIba09\\^OKaa07]^OL`a05^^OO\\a06a^OOTa09j^OKTa05j^OOTa0`1O1N2O2Me0[O3N2M4M2M3M3N2M3N2M2N1_NcK_C]4`<fK]C[4a<iK\\CX4c<jK[CW4d<lKYCU4f<nKXCR4g<QLVCQ4g<SLVCn3i<VLSCk3l<XLQCi3n<ZLPCf3o<^LmBc3Q=aLlB`3S=cLjB^3U=fLgB[3X=hLfBX3Y=h101O000000000000000O10001O000000000000001O2N2N2N2N2N2N1O2N2N2N2N2OmI`Cf4^<ZKhCb4V<]KnCb4P<^KPDd4n;\\KSDe4k;ZKVDg4i;YKWDi4h;UKYDm4g;QKYDQ5h;lJXDV5h;gJYD[5g;cJZD^5f;`JZDb5f;[J[Dg5f;kIeDW6a<2N1O2O1N2N2N2N2N2N2O1N3M4L4L3M4L3NhJmCe2o;[MVDc2g;]M_Da2];^MiDa2S;_MRE_2l:`M[E]2a:bMgEZ2W:eMPFX2m9fM[FV2c9iMdFT2X9lMQGo1l8oMbGg1\\8XNQHa1j7]NcH\\1X7_NPId1k6UN^Im1^6mMiIW2R6hMQJV2o5hMUJU2k5kMXJR2P6fMSJW2[6[MhIb2f6oL^In2m;M3M3M2N4L4L4L3M4L4`Nc^O3ca0Ai^O9]a0[Oo^O`0\\b0K6J6K[k`2"}, "label": "a man with the number 21 on his shirt"}]}
{"id": 358033, "image": "COCO_train2014_000000358033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"double decker bus to far right\"."}], "task": "refering", "answer_template": "The \"double decker bus to far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 292], "bbox": [0.6686093750000001, 0.37078299776286355, 0.87265625, 0.791006711409396], "iscrowd": 0, "area": 17253.1107, "rle": {"size": [447, 640], "counts": "]Tk51l=3M2N2O1N3N1O2M2O2N2O001O1O2N2Nj0UO101cMjNmGX1l7@bGa0X8JbG8V85bGMX8b0^G_OZ8R1\\GnN\\8_3L3001O10O01E;E;0O2O000O1000000O10000O10000O1000000O10000000000000001O00000001O000000000000000000001O00000000001O00000000001O00000000001O000000O100000000000000O10000000000003M2N3M3M2M4M3N101K5J5D=I7J5N3M3M2dMREd0>VOc:L\\GGg80ZWS1"}, "label": "double decker bus to far right"}]}
{"id": 358033, "image": "COCO_train2014_000000358033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a double - decker bus , visible directly from the front , without anyone near it\"."}], "task": "refering", "answer_template": "The \"a double - decker bus , visible directly from the front , without anyone near it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 292], "bbox": [0.6686093750000001, 0.37078299776286355, 0.87265625, 0.791006711409396], "iscrowd": 0, "area": 17253.1107, "rle": {"size": [447, 640], "counts": "]Tk51l=3M2N2O1N3N1O2M2O2N2O001O1O2N2Nj0UO101cMjNmGX1l7@bGa0X8JbG8V85bGMX8b0^G_OZ8R1\\GnN\\8_3L3001O10O01E;E;0O2O000O1000000O10000O10000O1000000O10000000000000001O00000001O000000000000000000001O00000000001O00000000001O00000000001O000000O100000000000000O10000000000003M2N3M3M2M4M3N101K5J5D=I7J5N3M3M2dMREd0>VOc:L\\GGg80ZWS1"}, "label": "a double - decker bus , visible directly from the front , without anyone near it"}]}
{"id": 358033, "image": "COCO_train2014_000000358033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a double decker bus second on the right\"."}], "task": "refering", "answer_template": "The \"a double decker bus second on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264], "bbox": [0.34809375, 0.4853691275167785, 0.6111093750000001, 0.7191498881431768], "iscrowd": 0, "area": 12871.542049999996, "rle": {"size": [447, 640], "counts": "maQ3P1o<^1bNg0YO0O1000000O100000000O10000000000000000000000000000000000000000000000000000001N10001O00001O00001O00001O0000001N1000001O00001O0000001O0000001O0000001O00001O0000001O000O101O00000O2O0000001N10001O000O101O000O101O0O1N2XOQE_NQ;]1j0000001O00001O000002O6I001O010O000001O0O101N100O1N3N1O1N2O2N1O11O010O001O01O00001N1004L4MN1M3N2N3O0000001VOTC?l<^O[C>e<@aCc0X<ZOoCe0h<Ck]\\3"}, "label": "a double decker bus second on the right"}]}
{"id": 358033, "image": "COCO_train2014_000000358033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two buses behind the cinzano bus\"."}], "task": "refering", "answer_template": "The \"two buses behind the cinzano bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264], "bbox": [0.34809375, 0.4853691275167785, 0.6111093750000001, 0.7191498881431768], "iscrowd": 0, "area": 12871.542049999996, "rle": {"size": [447, 640], "counts": "maQ3P1o<^1bNg0YO0O1000000O100000000O10000000000000000000000000000000000000000000000000000001N10001O00001O00001O00001O0000001N1000001O00001O0000001O0000001O0000001O00001O0000001O000O101O00000O2O0000001N10001O000O101O000O101O0O1N2XOQE_NQ;]1j0000001O00001O000002O6I001O010O000001O0O101N100O1N3N1O1N2O2N1O11O010O001O01O00001N1004L4MN1M3N2N3O0000001VOTC?l<^O[C>e<@aCc0X<ZOoCe0h<Ck]\\3"}, "label": "two buses behind the cinzano bus"}]}
{"id": 358033, "image": "COCO_train2014_000000358033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck in the front of the line with cinzano written on it\"."}], "task": "refering", "answer_template": "The \"truck in the front of the line with cinzano written on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277], "bbox": [0.0015625, 0.3578523489932886, 0.34921875, 0.7771364653243847], "iscrowd": 0, "area": 32498.2738, "rle": {"size": [447, 640], "counts": "ob0U1j<Z2gMk1TN001O010O0010O01O0O2O0O2O1O0O2O1O00100O00100O001O10O01O00O101N100O100O2N100O100O2O0O1O100O2O0O1O100O1O2O0O1O100O2N100O10001O000000001O000000001O0O100000001O000000001O000000001O0O1000001O000000001O00000000001O0O1000001O000000001O000000001O0O1000001O00000000001O000000001O0O1000001O0000001O0000001O00001O0O10001O00001O0000001O00001O00000O2O01O010O001O010O00010O0001N100O101N100O2O0O100O2O0O101N1000001O00001N1000001O00001O0O10001O001O2N1O2Mg2SMmYe5"}, "label": "truck in the front of the line with cinzano written on it"}]}
{"id": 358033, "image": "COCO_train2014_000000358033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"side of a cinzano double decker bus with many windows\"."}], "task": "refering", "answer_template": "The \"side of a cinzano double decker bus with many windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277], "bbox": [0.0015625, 0.3578523489932886, 0.34921875, 0.7771364653243847], "iscrowd": 0, "area": 32498.2738, "rle": {"size": [447, 640], "counts": "ob0U1j<Z2gMk1TN001O010O0010O01O0O2O0O2O1O0O2O1O00100O00100O001O10O01O00O101N100O100O2N100O100O2O0O1O100O2O0O1O100O1O2O0O1O100O2N100O10001O000000001O000000001O0O100000001O000000001O000000001O0O1000001O000000001O00000000001O0O1000001O000000001O000000001O0O1000001O00000000001O000000001O0O1000001O0000001O0000001O00001O0O10001O00001O0000001O00001O00000O2O01O010O001O010O00010O0001N100O101N100O2O0O100O2O0O101N1000001O00001N1000001O00001O0O10001O001O2N1O2Mg2SMmYe5"}, "label": "side of a cinzano double decker bus with many windows"}]}
{"id": 5782, "image": "COCO_train2014_000000005782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and brown dog walking through the ocean water\"."}], "task": "refering", "answer_template": "The \"a black and brown dog walking through the ocean water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 166, 167, 189, 190, 191, 192, 193, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 281, 282, 283, 284, 286, 287, 288, 289, 290, 303, 304, 307, 311, 312, 334, 335], "bbox": [0.191296875, 0.35222916666666665, 0.79403125, 0.864375], "iscrowd": 0, "area": 32317.930449999993, "rle": {"size": [480, 640], "counts": "Sbi11l>3M3N2N2N2O1N2O1O001O1O001O001O001N10000O2O000N2N2M31O101N1O]K1WJNh5:RJEl5b0PJ]Oo5g0PJXOo5j0PJUOP6n0nIQOR6Q1mInNS6U1lIiNT6Y1kIgNT6[1kIeNT6^1jIbNU6`1jI`NU6b1jI^NT6f1iI[NV6h1gIYNW6k1gIUNY6m1dITN[6o1bIQN_6P2`IPN_6R2_IoMa6S2\\InMc6T2\\IlMd6U2RHaMH;V88`G<:ZON2W83kG?JA2NY81nG?BG5JZ80QH?[OZ1d8VNSH`0VO\\1g8RNVHc0oN]1j8PNZH\\3f7cL\\H\\3c7dL_H[3a7cLbH\\3^7dLcH[3]7dLfHZ3Z7fLgHY3Y7fLjHX3V7hLkHW3U7hLmHW3S7iLoHU3Q7jLQIU3o6kLRIT3n6kLUIS3k6mLVIR3j6mLYIQ3g6oLZIP3g6nL[IP3g6lL\\IT3d6iL_IW3T9010O010O10O010O010O10O10O0103M4L2N1O1O2N1O1O3M3MO10000O100O100O01000O100O100O0^OVEPMj:i2_EUMa:i2cETM^:j2fETMZ:k2iERMX:l2o0N2N2O1O100000000O1O2N1O1O1000000000000010O001O001O010O0010O01O010O1O101N1O101N0010O01O010O00100O00100O1O100O00100O10O01O010O1O010O010O1O0001O00000000000000000000001O00001O001O00001O0000WD[MX;d2fDaMW;_2hDfMU;Y2jDkMS;U2kDQNQ;n1oDVNn:j1PE[Nn:o2L3M3M4L10N1O100O101N1O100O100O2N100O100XOUEZMl:d2bEPM^:o2j0O2N1O1O1O1O100O1O1O1O1O1O1O1O1N2N2N2N2O0O2O1O1O001O1O001O1O1O001O0000000000000000000O1000000000000000000100000000O010000000O0100O100O100O100O100O100O1O1O1N2O1O1O1O1N200000000000000001O00001O01O010O00100O001O001O0CkBoNV=P1jBPOW=o0iBQOX=m0iBSOW=m0iBSOX=l0hBSOZ=l0fBTO[=j0fBVOZ=j0fBUO\\=j0eBUO[=j0?O00000O100O100O100O2O0O1O1O1N2O1O2M2N2N2N2Klcm1"}, "label": "a black and brown dog walking through the ocean water"}]}
{"id": 5782, "image": "COCO_train2014_000000005782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bog in watar\"."}], "task": "refering", "answer_template": "The \"a bog in watar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 166, 167, 189, 190, 191, 192, 193, 199, 200, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 281, 282, 283, 284, 286, 287, 288, 289, 290, 303, 304, 307, 311, 312, 334, 335], "bbox": [0.191296875, 0.35222916666666665, 0.79403125, 0.864375], "iscrowd": 0, "area": 32317.930449999993, "rle": {"size": [480, 640], "counts": "Sbi11l>3M3N2N2N2O1N2O1O001O1O001O001O001N10000O2O000N2N2M31O101N1O]K1WJNh5:RJEl5b0PJ]Oo5g0PJXOo5j0PJUOP6n0nIQOR6Q1mInNS6U1lIiNT6Y1kIgNT6[1kIeNT6^1jIbNU6`1jI`NU6b1jI^NT6f1iI[NV6h1gIYNW6k1gIUNY6m1dITN[6o1bIQN_6P2`IPN_6R2_IoMa6S2\\InMc6T2\\IlMd6U2RHaMH;V88`G<:ZON2W83kG?JA2NY81nG?BG5JZ80QH?[OZ1d8VNSH`0VO\\1g8RNVHc0oN]1j8PNZH\\3f7cL\\H\\3c7dL_H[3a7cLbH\\3^7dLcH[3]7dLfHZ3Z7fLgHY3Y7fLjHX3V7hLkHW3U7hLmHW3S7iLoHU3Q7jLQIU3o6kLRIT3n6kLUIS3k6mLVIR3j6mLYIQ3g6oLZIP3g6nL[IP3g6lL\\IT3d6iL_IW3T9010O010O10O010O010O10O10O0103M4L2N1O1O2N1O1O3M3MO10000O100O100O01000O100O100O0^OVEPMj:i2_EUMa:i2cETM^:j2fETMZ:k2iERMX:l2o0N2N2O1O100000000O1O2N1O1O1000000000000010O001O001O010O0010O01O010O1O101N1O101N0010O01O010O00100O00100O1O100O00100O10O01O010O1O010O010O1O0001O00000000000000000000001O00001O001O00001O0000WD[MX;d2fDaMW;_2hDfMU;Y2jDkMS;U2kDQNQ;n1oDVNn:j1PE[Nn:o2L3M3M4L10N1O100O101N1O100O100O2N100O100XOUEZMl:d2bEPM^:o2j0O2N1O1O1O1O100O1O1O1O1O1O1O1O1N2N2N2N2O0O2O1O1O001O1O001O1O1O001O0000000000000000000O1000000000000000000100000000O010000000O0100O100O100O100O100O100O1O1O1N2O1O1O1O1N200000000000000001O00001O01O010O00100O001O001O0CkBoNV=P1jBPOW=o0iBQOX=m0iBSOW=m0iBSOX=l0hBSOZ=l0fBTO[=j0fBVOZ=j0fBUO\\=j0eBUO[=j0?O00000O100O100O100O2O0O1O1O1N2O1O2M2N2N2N2Klcm1"}, "label": "a bog in watar"}]}
{"id": 267927, "image": "COCO_train2014_000000267927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a traffic signal in side view\"."}], "task": "refering", "answer_template": "The \"a traffic signal in side view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 38, 39, 40, 41, 56, 57, 58, 72, 73, 74, 75, 90, 91, 92, 107, 108, 109, 125, 126], "bbox": [0.2544166666666667, 0.078453125, 0.4709166666666667, 0.34059375000000003], "iscrowd": 0, "area": 8673.510449999998, "rle": {"size": [640, 480], "counts": "^n\\21oc00O2O001O0O2O001d]OJl`06S_OKl`07S_OJl`06T_OJl`07S_OJ2M[?:b@J3MZ?:b@J3L[?:b@J2N[?9b@J2NZ?9c@J3MZ?:b@J3MZ?9c@J3MZ?:b@J3MY?:d@I3NX?:c@J3MZ?9c@J3NX?:d@I3MY?:d@I3NX?:c@J4MX?9d@J4MW?;d@I3NX?9e@I3OW?9d@J4MX?9d@J4NV?:e@I4MW?:e@I4NV?:e@I4NU?:f@I4NV?:e@I4NV?9f@I4OU?9f@I4NU?:f@I5NT?:f@I5MU?:f@I5NT?:f@I4OT?9g@I5NT?:f@I5Nj>c0QA_O5Nj>d0PA_O5Nj>c0QA_O5Oi>c0PA_O6Oj>b0PA@5Oi>c0QA^O60h>b0RA_O5Oi>Q2WAPNh>Q2WAoMi>Q2WAPNg>R2WAPNh>P2XAPNh>U2SAlMl>[2m@fMR?d301N3N2N1O2N2M2O00O1000O10O10000O10O1006I7J7I6J7H7J7I6I8I6J7I6I8I6J6J7H7J7I6I8I6J7I6I8Ican4"}, "label": "a traffic signal in side view"}]}
{"id": 267927, "image": "COCO_train2014_000000267927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stoplight faces to the left at an intersection\"."}], "task": "refering", "answer_template": "The \"a stoplight faces to the left at an intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 38, 39, 40, 41, 56, 57, 58, 72, 73, 74, 75, 90, 91, 92, 107, 108, 109, 125, 126], "bbox": [0.2544166666666667, 0.078453125, 0.4709166666666667, 0.34059375000000003], "iscrowd": 0, "area": 8673.510449999998, "rle": {"size": [640, 480], "counts": "^n\\21oc00O2O001O0O2O001d]OJl`06S_OKl`07S_OJl`06T_OJl`07S_OJ2M[?:b@J3MZ?:b@J3L[?:b@J2N[?9b@J2NZ?9c@J3MZ?:b@J3MZ?9c@J3MZ?:b@J3MY?:d@I3NX?:c@J3MZ?9c@J3NX?:d@I3MY?:d@I3NX?:c@J4MX?9d@J4MW?;d@I3NX?9e@I3OW?9d@J4MX?9d@J4NV?:e@I4MW?:e@I4NV?:e@I4NU?:f@I4NV?:e@I4NV?9f@I4OU?9f@I4NU?:f@I5NT?:f@I5MU?:f@I5NT?:f@I4OT?9g@I5NT?:f@I5Nj>c0QA_O5Nj>d0PA_O5Nj>c0QA_O5Oi>c0PA_O6Oj>b0PA@5Oi>c0QA^O60h>b0RA_O5Oi>Q2WAPNh>Q2WAoMi>Q2WAPNg>R2WAPNh>P2XAPNh>U2SAlMl>[2m@fMR?d301N3N2N1O2N2M2O00O1000O10O10000O10O1006I7J7I6J7H7J7I6I8I6J7I6I8I6J6J7H7J7I6I8I6J7I6I8Ican4"}, "label": "a stoplight faces to the left at an intersection"}]}
{"id": 267927, "image": "COCO_train2014_000000267927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white bus with a woman in white inside the door\"."}], "task": "refering", "answer_template": "The \"a blue and white bus with a woman in white inside the door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 139, 153, 154, 155, 156, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259], "bbox": [0.003, 0.37528125, 0.27266666666666667, 0.6831406250000001], "iscrowd": 0, "area": 20821.777399999995, "rle": {"size": [640, 480], "counts": "ik0<dc0h3XLU1jN2O1O1O1O2N1O1O1O1O1N2O2N1O1O1O1O1O00000O10O10000000000O10O10000000O1000000000O10O10000000000O1000O10000000O10000000O10O1000000000000O10O10000000O1000000000O01000001O001O0O2O001O00001N1oJfA\\4[>UKWBg4d>M4L4L3M4K4M4L4L3M4L3L5L4L3M4L3M4L4K4M4L3M4L4L3L5L3M4L4L3M4K4M4L4Lkgi6"}, "label": "a blue and white bus with a woman in white inside the door"}]}
{"id": 267927, "image": "COCO_train2014_000000267927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and blue bus with tinted windows\"."}], "task": "refering", "answer_template": "The \"a white and blue bus with tinted windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 139, 153, 154, 155, 156, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259], "bbox": [0.003, 0.37528125, 0.27266666666666667, 0.6831406250000001], "iscrowd": 0, "area": 20821.777399999995, "rle": {"size": [640, 480], "counts": "ik0<dc0h3XLU1jN2O1O1O1O2N1O1O1O1O1N2O2N1O1O1O1O1O00000O10O10000000000O10O10000000O1000000000O10O10000000000O1000O10000000O10000000O10O1000000000000O10O10000000O1000000000O01000001O001O0O2O001O00001N1oJfA\\4[>UKWBg4d>M4L4L3M4K4M4L4L3M4L3L5L4L3M4L3M4L4K4M4L3M4L4L3L5L3M4L4L3M4K4M4L4Lkgi6"}, "label": "a white and blue bus with tinted windows"}]}
{"id": 177817, "image": "COCO_train2014_000000177817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcycle with large back tire\"."}], "task": "refering", "answer_template": "The \"motorcycle with large back tire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58, 59, 60, 61, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 213], "bbox": [0.077890625, 0.0, 0.75021875, 0.6722976501305483], "iscrowd": 0, "area": 29922.578100000002, "rle": {"size": [383, 640], "counts": "Skb0`0n:e0]Od0G8I8H8G7J5J6K51O2NN2N2N2N1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O001O1O1O1O1O1N101O1O1O1O1O1O001O1O1O1O1O001O1O1N2O1O1O001O1O000000001O000000000O101O0000000000001O0000000000001N10000000]LQIV2o6gM]IP2c6lMeIQ2Z6lMjIS2V6jMnIU2R6hMQJX2o5gMSJX2m5gMTJY2l5fMVJY2j5fMWJ[2h5dMZJ[2f5dM[J\\2e5cM]J\\2b5eM^J[2b5dM`J[2`5eM`J[2`5eMaJ[2^5dMdJ[2\\5eMeJZ2[5fMgJX2Y5gMkJV2U5jMmJT2S5lMPKQ2P5nMSKQ2l4oMVKo1j4QNXKU1mMQOk6I[KS1lMSOi6J]KP1mMTOf6L^Kn0nMUOc6M`Kl0PNVO`6N`Kk0SNVO\\6ObKi0TNWOZ6OcKh0UNYOX6OcKg0WNYOV60cKf0YNYOT60eKd0ZN[OQ61eKc0\\N[Oo52eKb0^N[Om53dKa0cN[Oh54eK`0eN[Of54fK`0gNZOc56fK>kNZO_58eK>nNYO]59eK=l5CTJ;n5DSJ;m5FRJ:o5FQJ8Q6HoI7R6HoI7R6ImI6U6JkI5V6KjI3X6MgI2[6MfI0]60cIN_62`IMb63^IJe66[IHg67a2000O1000O1000000O100000000O01N2O1O1N101O1N2O00O1N101O1N101O1N2O001O0O12N2O1N3M2N2O1N2N2O1N2N2N2O1N2M4M2N2M3N2M3N2M3N2M3N2N2M3N2M30O100000000000000000000O1000000000000000000O100000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000001O0000000000000000000000000000000YNZFU1f9hNaFT1_9iNgFT1Y9iNnFT1n90SOkNPGU1P9mNnFS1R9nNmFR1S9oNmFP1S9QOlFo0T9SOjFm0V9VOjFd0Y9^OgF>[9EdF6_9M`FNc94]FGf9<XFAj9a0i03M2N3M2N2N2N2N2N2N2N2N2N2N2N1O2N2N1O1OO1000000O1000000O1000000O1000000O10000O1000000O1000000O10000O10000O1000000O1000000O1000000O101O4L4L4L3M4L4L4L4L3L4L4M3Ldik1"}, "label": "motorcycle with large back tire"}]}
{"id": 177817, "image": "COCO_train2014_000000177817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle behind the blue motorcycle\"."}], "task": "refering", "answer_template": "The \"a black motorcycle behind the blue motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58, 59, 60, 61, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 213], "bbox": [0.077890625, 0.0, 0.75021875, 0.6722976501305483], "iscrowd": 0, "area": 29922.578100000002, "rle": {"size": [383, 640], "counts": "Skb0`0n:e0]Od0G8I8H8G7J5J6K51O2NN2N2N2N1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O001O1O1O1O1O1N101O1O1O1O1O1O001O1O1O1O1O001O1O1N2O1O1O001O1O000000001O000000000O101O0000000000001O0000000000001N10000000]LQIV2o6gM]IP2c6lMeIQ2Z6lMjIS2V6jMnIU2R6hMQJX2o5gMSJX2m5gMTJY2l5fMVJY2j5fMWJ[2h5dMZJ[2f5dM[J\\2e5cM]J\\2b5eM^J[2b5dM`J[2`5eM`J[2`5eMaJ[2^5dMdJ[2\\5eMeJZ2[5fMgJX2Y5gMkJV2U5jMmJT2S5lMPKQ2P5nMSKQ2l4oMVKo1j4QNXKU1mMQOk6I[KS1lMSOi6J]KP1mMTOf6L^Kn0nMUOc6M`Kl0PNVO`6N`Kk0SNVO\\6ObKi0TNWOZ6OcKh0UNYOX6OcKg0WNYOV60cKf0YNYOT60eKd0ZN[OQ61eKc0\\N[Oo52eKb0^N[Om53dKa0cN[Oh54eK`0eN[Of54fK`0gNZOc56fK>kNZO_58eK>nNYO]59eK=l5CTJ;n5DSJ;m5FRJ:o5FQJ8Q6HoI7R6HoI7R6ImI6U6JkI5V6KjI3X6MgI2[6MfI0]60cIN_62`IMb63^IJe66[IHg67a2000O1000O1000000O100000000O01N2O1O1N101O1N2O00O1N101O1N101O1N2O001O0O12N2O1N3M2N2O1N2N2O1N2N2N2O1N2M4M2N2M3N2M3N2M3N2M3N2N2M3N2M30O100000000000000000000O1000000000000000000O100000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000001O0000000000000000000000000000000YNZFU1f9hNaFT1_9iNgFT1Y9iNnFT1n90SOkNPGU1P9mNnFS1R9nNmFR1S9oNmFP1S9QOlFo0T9SOjFm0V9VOjFd0Y9^OgF>[9EdF6_9M`FNc94]FGf9<XFAj9a0i03M2N3M2N2N2N2N2N2N2N2N2N2N2N1O2N2N1O1OO1000000O1000000O1000000O1000000O10000O1000000O1000000O10000O10000O1000000O1000000O1000000O101O4L4L4L3M4L4L4L4L3L4L4M3Ldik1"}, "label": "a black motorcycle behind the blue motorcycle"}]}
{"id": 177817, "image": "COCO_train2014_000000177817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue motorcycles with the orange handled saw on it\"."}], "task": "refering", "answer_template": "The \"the blue motorcycles with the orange handled saw on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 284, 285, 286, 287, 288, 289, 290, 291, 311], "bbox": [0.21956250000000002, 0.003864229765013055, 0.9987343750000001, 0.979242819843342], "iscrowd": 0, "area": 121886.93725000003, "rle": {"size": [383, 640], "counts": "Pld14X5MV18hNKS18lNKn0:oNJk0:TOIf0<XOGe0:YOJd08ZOLb05]ON`04^OO?2_O2>0@3=NB3=OA2>O@2`00^O1a00^O1b00[O2d0O[O1e01YO0f01YO0f02WOOj01UO0j02TOOk03RONn05oNLP17mNJS18iNIW1:SMRNEe1W3=PMPNGd1X3a0lLnMHb1\\3e0gLkMKa1]3h0dLjML_1`3i0aLjML^1c3i0`LlMJ\\1e3i0`LmMI[1f3h0aLPNEY1k3g0_LRNDW1n3g0]LUNBT1R4g0[LXN_OR1V4g0YLZN_Oo0Y4g0WL]N]Ol0]4f0VL`NZOk0a4e0TLcNXOh0e4e0RLeNWOf0h4e0PLj0Q4VOnKj0S4UOlKl0T4UOiKm0X4SOfKn0[4RObKP1_4PO_KQ1b4nN\\KT1e4lNYKU1h4kNUKW1l4iNRKX1n4iNoJY1R5fNmJ[1T5g21O1O1O1N2O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O16J8H8H8E:D=D<C6J50001O0000001O000O101O0000000000000000000000000000000000000000000000000000000000000000000000000001O00000000O1000000O101O000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O100001O00000000kMVGV1i8P10iMWGX1i8o00hMWGZ1i8fNWG[1h8n00fMXG]1h8cNXG]1h8bNXG_1h8aNXG`1g8_NYGb1g8^NYGb1g8]NZGc1f8]NYGd1g8[NZGf1e8ZNZGg1f8XN[Gh1e8XN[Gh1e8WN[Gj1e8VN[Gk1d8UN[Gl1e8SN\\Gm1d8SN\\Gm1d8RN\\Go1d8QN\\GP2c8oM]GR2c8nM]GR2c8mM^GS2b8mM]GT2c8kM^GV2a8jM^GW2b8hM_GX2a8hM_GX2a8hM^GY2b8fM_GZ2a8fM^G\\2a8cM`G]2`8cM`G]2`8bM`G_2`8aM`G_2`8`M`Gb2_8^MaGb2_8]MbGc2^8]MaGd2_8[MbGe2^8[MaGf2_8ZMaGf2_8YMbGg2^8YMaGh2_8WMbGi2^8WMaGj2_8UMbGk2^8UMbGk2^8TMbGl2_8TMaGl2_8SMaGn2`8QM`Go2`8PMaGP3_8PM`GQ3`8nLaGR3_8nL`GS3`8mL`GS3`8lLaGT3_8lL`GT3f80O10000O100O100O10O0100O100O100O100O10O10O100O100O100O10O0100O100O100O100O10O0100O100O010O100O010O100O010O1000O0100O100O010O100O010O100O010O100O010O100O010O100O00100O10O0H9C<O20O0010O0100O010O0010O0100O010O010O1O010O010O10O01O010O010O10O010O01O10O010O010O0100O0010O010O0100O0010O0100O0100O010O0100O2O0O100O100O100O100O10000O100O100O100O100O101N100O100O10000O100O100O1SOmKmIT4S6nKjIS4V6nKhIS4X6oKdIS4\\6nKbIS4^6nK`IS4`6oK]IR4c6oKZIS4f6oKWIR4i6oKUIS4j6nKTIS4k6a00000000000000000000000000000000000O10001N100O100O1O100O100O100O100O100O100O1O100O100O2O0O100O1E;_LZ?"}, "label": "the blue motorcycles with the orange handled saw on it"}]}
{"id": 177817, "image": "COCO_train2014_000000177817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle with a saw attached to it\"."}], "task": "refering", "answer_template": "The \"a motorcycle with a saw attached to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 284, 285, 286, 287, 288, 289, 290, 291, 311], "bbox": [0.21956250000000002, 0.003864229765013055, 0.9987343750000001, 0.979242819843342], "iscrowd": 0, "area": 121886.93725000003, "rle": {"size": [383, 640], "counts": "Pld14X5MV18hNKS18lNKn0:oNJk0:TOIf0<XOGe0:YOJd08ZOLb05]ON`04^OO?2_O2>0@3=NB3=OA2>O@2`00^O1a00^O1b00[O2d0O[O1e01YO0f01YO0f02WOOj01UO0j02TOOk03RONn05oNLP17mNJS18iNIW1:SMRNEe1W3=PMPNGd1X3a0lLnMHb1\\3e0gLkMKa1]3h0dLjML_1`3i0aLjML^1c3i0`LlMJ\\1e3i0`LmMI[1f3h0aLPNEY1k3g0_LRNDW1n3g0]LUNBT1R4g0[LXN_OR1V4g0YLZN_Oo0Y4g0WL]N]Ol0]4f0VL`NZOk0a4e0TLcNXOh0e4e0RLeNWOf0h4e0PLj0Q4VOnKj0S4UOlKl0T4UOiKm0X4SOfKn0[4RObKP1_4PO_KQ1b4nN\\KT1e4lNYKU1h4kNUKW1l4iNRKX1n4iNoJY1R5fNmJ[1T5g21O1O1O1N2O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O16J8H8H8E:D=D<C6J50001O0000001O000O101O0000000000000000000000000000000000000000000000000000000000000000000000000001O00000000O1000000O101O000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O100001O00000000kMVGV1i8P10iMWGX1i8o00hMWGZ1i8fNWG[1h8n00fMXG]1h8cNXG]1h8bNXG_1h8aNXG`1g8_NYGb1g8^NYGb1g8]NZGc1f8]NYGd1g8[NZGf1e8ZNZGg1f8XN[Gh1e8XN[Gh1e8WN[Gj1e8VN[Gk1d8UN[Gl1e8SN\\Gm1d8SN\\Gm1d8RN\\Go1d8QN\\GP2c8oM]GR2c8nM]GR2c8mM^GS2b8mM]GT2c8kM^GV2a8jM^GW2b8hM_GX2a8hM_GX2a8hM^GY2b8fM_GZ2a8fM^G\\2a8cM`G]2`8cM`G]2`8bM`G_2`8aM`G_2`8`M`Gb2_8^MaGb2_8]MbGc2^8]MaGd2_8[MbGe2^8[MaGf2_8ZMaGf2_8YMbGg2^8YMaGh2_8WMbGi2^8WMaGj2_8UMbGk2^8UMbGk2^8TMbGl2_8TMaGl2_8SMaGn2`8QM`Go2`8PMaGP3_8PM`GQ3`8nLaGR3_8nL`GS3`8mL`GS3`8lLaGT3_8lL`GT3f80O10000O100O100O10O0100O100O100O100O10O10O100O100O100O10O0100O100O100O100O10O0100O100O010O100O010O100O010O1000O0100O100O010O100O010O100O010O100O010O100O010O100O00100O10O0H9C<O20O0010O0100O010O0010O0100O010O010O1O010O010O10O01O010O010O10O010O01O10O010O010O0100O0010O010O0100O0010O0100O0100O010O0100O2O0O100O100O100O100O10000O100O100O100O100O101N100O100O10000O100O100O1SOmKmIT4S6nKjIS4V6nKhIS4X6oKdIS4\\6nKbIS4^6nK`IS4`6oK]IR4c6oKZIS4f6oKWIR4i6oKUIS4j6nKTIS4k6a00000000000000000000000000000000000O10001N100O100O1O100O100O100O100O100O100O1O100O100O2O0O100O1E;_LZ?"}, "label": "a motorcycle with a saw attached to it"}]}
{"id": 38558, "image": "COCO_train2014_000000038558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bus\"."}], "task": "refering", "answer_template": "The \"a yellow bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 190, 191], "bbox": [0.40072599531615927, 0.336890625, 0.8127868852459016, 0.5586875], "iscrowd": 0, "area": 16456.319200000005, "rle": {"size": [640, 427], "counts": "XU[33ac0=D<C<D=G801N1O2O1N1O2O0O3N1N2N3N1N2O1N2O0O10000O10000O101OKj^O`MUa0`2n^O^MRa0a2700O001O1O001O10O10000O1000000O101O0Q_O_M^`0b2^_ObMb`0^2Y_OgMg`0Y2U_OkMj`0g2O00000O2ON2O0O2O1N101N2O0O2O1O1O0O3N1O1M3L4L4O1000000001O000001O01O2N4L4L3M2N2N2N3M2N1O0010O00000001O1O1O001O1O1O001O100O1O001O1O1O1O001O1O1O1O1O100O1O1O2N3M2N2N3M2N2NO1O1N2O2N1N2O1O1N2O1N2O1O1N10O010000O3O1N2O1O1N2O2M2O1O1N2O1N2O1O1N2O:FP1oNQ1PO]fa1"}, "label": "a yellow bus"}]}
{"id": 38558, "image": "COCO_train2014_000000038558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow bus with 707 on it\"."}], "task": "refering", "answer_template": "The \"the yellow bus with 707 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 190, 191], "bbox": [0.40072599531615927, 0.336890625, 0.8127868852459016, 0.5586875], "iscrowd": 0, "area": 16456.319200000005, "rle": {"size": [640, 427], "counts": "XU[33ac0=D<C<D=G801N1O2O1N1O2O0O3N1N2N3N1N2O1N2O0O10000O10000O101OKj^O`MUa0`2n^O^MRa0a2700O001O1O001O10O10000O1000000O101O0Q_O_M^`0b2^_ObMb`0^2Y_OgMg`0Y2U_OkMj`0g2O00000O2ON2O0O2O1N101N2O0O2O1O1O0O3N1O1M3L4L4O1000000001O000001O01O2N4L4L3M2N2N2N3M2N1O0010O00000001O1O1O001O1O1O001O100O1O001O1O1O1O001O1O1O1O1O100O1O1O2N3M2N2N3M2N2NO1O1N2O2N1N2O1O1N2O1N2O1O1N10O010000O3O1N2O1O1N2O2M2O1O1N2O1N2O1O1N2O:FP1oNQ1PO]fa1"}, "label": "the yellow bus with 707 on it"}]}
{"id": 210604, "image": "COCO_train2014_000000210604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"frisbee player in neon green jersey jumping to attempt catching the frisbee\"."}], "task": "refering", "answer_template": "The \"frisbee player in neon green jersey jumping to attempt catching the frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 142, 157, 158, 172, 173, 186, 187, 188, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 231, 232, 233, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 290, 291, 292, 293, 304, 305, 307, 308, 319, 320, 322, 337], "bbox": [0.2218735362997658, 0.35840625, 0.5901639344262295, 0.96596875], "iscrowd": 0, "area": 17215.436300000005, "rle": {"size": [640, 427], "counts": "j\\l11nc01O1O1O001O1O1O1O1O1O010000002N1O1OI_\\ONbc02600O100000000O1000O1QBKS84mGMR84nGMP84oGMQ83nGNQ82oG0o71QHOo71PH0o71PH0P80oG2o7NQH2o7OQH1o7OPH2o7OPH2o7NQH3o7MoG5P8LoG5Q8KnG6Q8JoG7P8JoG6R8JmG7R8JmG7R8JmG7S8HmG9R8HmG9S8GlG:S8GkG;T8EjG>V8BgG`0Y8AeGa0Z8@cGc0]8\\OaGg0^8ZO_Gi0`8WO_Gk0`8VO\\Gn0d8QOVGV1i8kNoF]1P9cNjFd1V9[NeFk1Z9VN_FQ2`9oM[FW2e9iMTFk1lLQNo<4QFm1d:SNYEn1j:RNREn1Q;RNlDo1V;QNgDP2\\;QN^DQ2d;QNWDP2k;SNoCo1S<RNhCo1Z<TN`Cn1a<TN[Cm1g<RNWCo1j<RNTCm1n<SNPCn1Q=RNnBl1U=TNiBe1_=`N\\BY1l=QOhAg0a>^OYAe0T;iMeGf1SMc0W;hMeGg1PMd0R;dMfD3U3i1oLb0V;cMdD3V3k1kLb0[;aMbD3V3m1jLa0^;_MaD5U3d4Y8bKeG_4[8aKdG`4[8bKcG_4^8`KaGa4_8_K`Gb4b0PK[4>RKc4a0QK]4<QKd4`0RK`49oJf4>TKc47mJf4>UKe45lJg4>VKf42kJi4=WKh40jJj4=WKj4NhJl4<XKm4LfJl4<YKo4KcJm4<ZKR5L]Jk4`0ZKU5OTJh4d0\\KY51bI[O_NY5`2`Kb50XIk4S1XKg51oHh4Y1WKk52hHh4]1UKm54cHm4Z1PKU63]HU5X1gJ^64WH\\5T1`Jg68oGd5o0TJT7>dGQ8_8VHVGk7m8c11N3K6K4L5K5K4M4K4M4WORFXGS:a6lEeJb0cNh9c6iE\\Je;_5^DTJQ<g5RDkI[<S6j0N3TMYB[Oh=a0^BXOh=c0^BoMGKT>Q2\\BnMK@V>[2UBnM]?]1i@]N]?Q1VAhNP?S1VAgNo>U1Q2K5L4K6J5L6I7JPQ]3"}, "label": "frisbee player in neon green jersey jumping to attempt catching the frisbee"}]}
{"id": 210604, "image": "COCO_train2014_000000210604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow shirt jumping from grass reaching in the air\"."}], "task": "refering", "answer_template": "The \"man in yellow shirt jumping from grass reaching in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 142, 157, 158, 172, 173, 186, 187, 188, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 231, 232, 233, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 290, 291, 292, 293, 304, 305, 307, 308, 319, 320, 322, 337], "bbox": [0.2218735362997658, 0.35840625, 0.5901639344262295, 0.96596875], "iscrowd": 0, "area": 17215.436300000005, "rle": {"size": [640, 427], "counts": "j\\l11nc01O1O1O001O1O1O1O1O1O010000002N1O1OI_\\ONbc02600O100000000O1000O1QBKS84mGMR84nGMP84oGMQ83nGNQ82oG0o71QHOo71PH0o71PH0P80oG2o7NQH2o7OQH1o7OPH2o7OPH2o7NQH3o7MoG5P8LoG5Q8KnG6Q8JoG7P8JoG6R8JmG7R8JmG7R8JmG7S8HmG9R8HmG9S8GlG:S8GkG;T8EjG>V8BgG`0Y8AeGa0Z8@cGc0]8\\OaGg0^8ZO_Gi0`8WO_Gk0`8VO\\Gn0d8QOVGV1i8kNoF]1P9cNjFd1V9[NeFk1Z9VN_FQ2`9oM[FW2e9iMTFk1lLQNo<4QFm1d:SNYEn1j:RNREn1Q;RNlDo1V;QNgDP2\\;QN^DQ2d;QNWDP2k;SNoCo1S<RNhCo1Z<TN`Cn1a<TN[Cm1g<RNWCo1j<RNTCm1n<SNPCn1Q=RNnBl1U=TNiBe1_=`N\\BY1l=QOhAg0a>^OYAe0T;iMeGf1SMc0W;hMeGg1PMd0R;dMfD3U3i1oLb0V;cMdD3V3k1kLb0[;aMbD3V3m1jLa0^;_MaD5U3d4Y8bKeG_4[8aKdG`4[8bKcG_4^8`KaGa4_8_K`Gb4b0PK[4>RKc4a0QK]4<QKd4`0RK`49oJf4>TKc47mJf4>UKe45lJg4>VKf42kJi4=WKh40jJj4=WKj4NhJl4<XKm4LfJl4<YKo4KcJm4<ZKR5L]Jk4`0ZKU5OTJh4d0\\KY51bI[O_NY5`2`Kb50XIk4S1XKg51oHh4Y1WKk52hHh4]1UKm54cHm4Z1PKU63]HU5X1gJ^64WH\\5T1`Jg68oGd5o0TJT7>dGQ8_8VHVGk7m8c11N3K6K4L5K5K4M4K4M4WORFXGS:a6lEeJb0cNh9c6iE\\Je;_5^DTJQ<g5RDkI[<S6j0N3TMYB[Oh=a0^BXOh=c0^BoMGKT>Q2\\BnMK@V>[2UBnM]?]1i@]N]?Q1VAhNP?S1VAgNo>U1Q2K5L4K6J5L6I7JPQ]3"}, "label": "man in yellow shirt jumping from grass reaching in the air"}]}
{"id": 210604, "image": "COCO_train2014_000000210604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white shirt trying to catch a frisbee\"."}], "task": "refering", "answer_template": "The \"a man with a white shirt trying to catch a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 129, 130, 144, 145, 159, 160, 161, 162, 174, 175, 176, 177, 188, 189, 190, 191, 203, 204, 205, 206, 218, 219, 220, 221, 233, 234, 235, 236, 247, 248, 249, 250, 251, 263, 264, 265, 266, 278, 279, 280, 281, 294, 295, 296, 309, 310, 311, 312, 324, 325], "bbox": [0.5288056206088994, 0.3049375, 0.8454098360655737, 0.945375], "iscrowd": 0, "area": 24708.990600000005, "rle": {"size": [640, 427], "counts": "`e]4l0Pc09H7I8H7I8G9@>[Oc0J6L4M4L3M3M3M4L3M3L5L4L4L5K4L4L5K4L5K4L4L:F8H7I6J7I6J6J7IcHkCW7Y<8I5J6J6J6hLX3^Ob0O1O1O001N2O1O1OO1O1\\L_HhJa7V5gHcJY7]5nH\\JS7b5UIWJl6h5\\IPJe6n5aImI`6Q6eIkI\\6T6hIhIY6X6jIeIU6]6mI_IT6a6PJZIS6e6SJSIP6k6UJoHo5o6RJnHQ6P7QJmHR6R7TJgHn5W7[J_Hh5`7aJUHc5h7dJPH`5m7]3N1N3hHVDT6l;cI_D7ZO\\5Z<TJbD3@k5`=2O2M2OeM\\B]N`=c1dB\\NZ=b1lB]NP=c1TC]Ni<a1]C]N`<c1dC]NX<b1nC]No;b1UD]Nh;b1^D]N_;b1eD^NW;a1oD]No:b1UE^Ng:a1_E^N^:a1aFdM[9[2nFaMn8_2ZG^Mb8l0UDC]3BZ8j0cDXOX30R8g0nDlNX3=g7g0YE_NW3k0^7d0mI]OP6c0SJ]Ok5b0WJ_Of5a0]J_Oa5?bJB[5>hJBU5>mJCQ5;SKEj4;XKFf49]KEb4:aKE_4:dKD[4<gKCY4;jKDU4<nKBQ4=RLBn3<VLBm3:ULEX4LlK2a4_OcK?P<O2N2N1O2N2N1O2N1O2N2N1O2NheX1"}, "label": "a man with a white shirt trying to catch a frisbee"}]}
{"id": 210604, "image": "COCO_train2014_000000210604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white and blue shirt and black shorts reaching for a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in white and blue shirt and black shorts reaching for a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 129, 130, 144, 145, 159, 160, 161, 162, 174, 175, 176, 177, 188, 189, 190, 191, 203, 204, 205, 206, 218, 219, 220, 221, 233, 234, 235, 236, 247, 248, 249, 250, 251, 263, 264, 265, 266, 278, 279, 280, 281, 294, 295, 296, 309, 310, 311, 312, 324, 325], "bbox": [0.5288056206088994, 0.3049375, 0.8454098360655737, 0.945375], "iscrowd": 0, "area": 24708.990600000005, "rle": {"size": [640, 427], "counts": "`e]4l0Pc09H7I8H7I8G9@>[Oc0J6L4M4L3M3M3M4L3M3L5L4L4L5K4L4L5K4L5K4L4L:F8H7I6J7I6J6J7IcHkCW7Y<8I5J6J6J6hLX3^Ob0O1O1O001N2O1O1OO1O1\\L_HhJa7V5gHcJY7]5nH\\JS7b5UIWJl6h5\\IPJe6n5aImI`6Q6eIkI\\6T6hIhIY6X6jIeIU6]6mI_IT6a6PJZIS6e6SJSIP6k6UJoHo5o6RJnHQ6P7QJmHR6R7TJgHn5W7[J_Hh5`7aJUHc5h7dJPH`5m7]3N1N3hHVDT6l;cI_D7ZO\\5Z<TJbD3@k5`=2O2M2OeM\\B]N`=c1dB\\NZ=b1lB]NP=c1TC]Ni<a1]C]N`<c1dC]NX<b1nC]No;b1UD]Nh;b1^D]N_;b1eD^NW;a1oD]No:b1UE^Ng:a1_E^N^:a1aFdM[9[2nFaMn8_2ZG^Mb8l0UDC]3BZ8j0cDXOX30R8g0nDlNX3=g7g0YE_NW3k0^7d0mI]OP6c0SJ]Ok5b0WJ_Of5a0]J_Oa5?bJB[5>hJBU5>mJCQ5;SKEj4;XKFf49]KEb4:aKE_4:dKD[4<gKCY4;jKDU4<nKBQ4=RLBn3<VLBm3:ULEX4LlK2a4_OcK?P<O2N2N1O2N2N1O2N1O2N2N1O2NheX1"}, "label": "a man in white and blue shirt and black shorts reaching for a frisbee"}]}
{"id": 54957, "image": "COCO_train2014_000000054957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man nearest to the black dog\"."}], "task": "refering", "answer_template": "The \"the man nearest to the black dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 219, 220, 221, 222, 223, 224, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 310, 311, 312, 313, 314, 333, 334, 335, 336, 356, 357, 358, 359, 378, 379, 380, 381], "bbox": [0.45, 0.48089583333333336, 0.810671875, 1.0], "iscrowd": 0, "area": 20650.334100000004, "rle": {"size": [480, 640], "counts": "j^W41m>2O1O1N2O1N2O2M2N2O1O1O1O1O1O1O2M2O1N2N2N2L4L4VOk0M2N2M3N2N2N2O1N3N1N2O1N2O1N2O1N3N1O1O1O1O0eLkLmJT3T5lLkJT3U5mLjJS3W5mLhJT3W5mLhJS3Y5nLeJR3[5oLdJQ3]5oLbJQ3^5PMaJP3`5PM_JQ3`5PM_JP3b5QM\\JP3c5QM\\Jo2e5QM[Jo2e5QMZJo2f5RMZJn2f5RMYJo2h5QMWJn2j5RMUJo2k5QMUJn2l5RMSJo2m5QMSJn2o5QMPJP3P6QMoIo2Q6QMoIn2R6RMnIn2R6RMnIm2S6SMmIm2T6SMkIl2V6TMjIl2V6TMiIl2X6UMgIk2Y6UMgIj2[6UMeIk2[6VMdIj2\\6VMdIi2]6WMcIi2]6XMbIg2`6XM`Ih2`6XM_Ih2b6YM]Ig2c6YM]If2d6ZM\\If2e6YM[If2f6[MYId2h6\\MWIe2i6[MWId2j6]MUIc2l6\\MSId2n6\\MRId2n6]MQIb2P7^MPIb2P7^MoHb2R7_MmHa2T7^MlHa2U7`MiH`2X7`MgHa2Y7`MeH`2\\7`McHa2^7_M`Ha2a7_M^Hb2b7_M\\Ha2e7_MZHb2f7_MXHa2j7^MTHb2n7_MoGa2S8_MkGb2V8_MgGa2[8_McGa2`8_M^G_2e8aMYG_2i8bMTG]2o8cMPG\\2R9fMjFZ2Y9gMdFW2_9kM]FU2e9lMXFS2k9oMPFQ2S:QNfEP2_:QNZEQ2i:R11O100O2N1O1O1O1O2N10000N2N3M2M3N2M3N2N2N3N1N2N2N2O1N2O2N1O1O1O10ZOPN\\Do1];]N_Db1[;U1M21O001O1O001OM_LhDa3Y;_LgD`3Z;`LfD_3Z;aLfD_3[;aLeD^3\\;bLcD^3^;bLbD^3];21`LbD^3^;bLbD]3_;2O0100O100O1O1O1O1O1O1O100O100000001O1O1O2N3L3N3L6QNSD`0Y=N1O0000001O00000010O0010O000000100O10000O1O1N3N2M8[Od]h1"}, "label": "the man nearest to the black dog"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor in the back on the left\"."}], "task": "refering", "answer_template": "The \"the monitor in the back on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 92], "bbox": [0.003109375, 0.0025848563968668406, 0.254921875, 0.322088772845953], "iscrowd": 0, "area": 14639.507450000001, "rle": {"size": [383, 640], "counts": "mi0l1W8l10O10000O10000O10000O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O10000000000000O1000000000000000000000000000000000000000000000000000000000O100000000O10000000000O100000000O100000000O100000000O100000000O10000000000O10000000O01000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O2O00001N3N6J7I6J7I6J7I6J7I6K6J5K6K^Zb5"}, "label": "the monitor in the back on the left"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a screen above a white power strip\"."}], "task": "refering", "answer_template": "The \"a screen above a white power strip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 92], "bbox": [0.003109375, 0.0025848563968668406, 0.254921875, 0.322088772845953], "iscrowd": 0, "area": 14639.507450000001, "rle": {"size": [383, 640], "counts": "mi0l1W8l10O10000O10000O10000O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O10000000000000O1000000000000000000000000000000000000000000000000000000000O100000000O10000000000O100000000O100000000O100000000O100000000O10000000000O10000000O01000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O2O00001N3N6J7I6J7I6J7I6J7I6K6J5K6K^Zb5"}, "label": "a screen above a white power strip"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"keyboard all the way to right\"."}], "task": "refering", "answer_template": "The \"keyboard all the way to right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 227, 228, 229, 252], "bbox": [0.68315625, 0.3752741514360313, 1.0, 0.7348302872062663], "iscrowd": 0, "area": 13679.486599999997, "rle": {"size": [383, 640], "counts": "[dS53k;1O2O0O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2O0O1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1101N01O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O01O01O01O010O001O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O01O01O01O010O0010O0001O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O00010O0ZNbK"}, "label": "keyboard all the way to right"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black keyboard to the right of the laptops\"."}], "task": "refering", "answer_template": "The \"the black keyboard to the right of the laptops\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 227, 228, 229, 252], "bbox": [0.68315625, 0.3752741514360313, 1.0, 0.7348302872062663], "iscrowd": 0, "area": 13679.486599999997, "rle": {"size": [383, 640], "counts": "[dS53k;1O2O0O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2O0O1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1101N01O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O01O01O01O010O001O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O01O01O01O010O0010O0001O010O0010O00010O01O010O00010O0010O01O01O01O010O0010O00010O0ZNbK"}, "label": "the black keyboard to the right of the laptops"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mini laptop sitting on a big laptop\"."}], "task": "refering", "answer_template": "The \"a mini laptop sitting on a big laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242], "bbox": [0.23937499999999998, 0.16143603133159268, 0.59171875, 0.777154046997389], "iscrowd": 0, "area": 40838.68555, "rle": {"size": [383, 640], "counts": "e_i11m;1N2N2N3M2N2O1N2N2N2N2O2M2N2N2N2O1N2N3M2N2O1N2N2N2N2N3N1N2N2N2N2O1N3M2_KlMUOV2k0XNeNj1[1cNWN^1i1oNhMS1X2[OXMg0h2FiL<W32YL1f3<kKFU4h0\\KYOd4T1mJnNS5o3000001O0000000001O0001O0000000000001O000000000000001O0000000000001O0001O0000000001O0000000000001O000000000000001O00000000000010O00000000000001O0000000000001O000000000000001O00000001O0001O000000000000001O0000000000001O000000000000001O01O0000000001O000000000000001O0000000000001O000000000001O01O0000000000001OL4K5L4K5K5K5K5L4K5K5K5K5L4K5K5K5K5L5J5A?\\Od0\\Od0]Oc0\\Od0\\Od0]OYbQ3"}, "label": "a mini laptop sitting on a big laptop"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small ibm thinkpad laptop sitting on top of a larger laptop both powered off\"."}], "task": "refering", "answer_template": "The \"a small ibm thinkpad laptop sitting on top of a larger laptop both powered off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242], "bbox": [0.23937499999999998, 0.16143603133159268, 0.59171875, 0.777154046997389], "iscrowd": 0, "area": 40838.68555, "rle": {"size": [383, 640], "counts": "e_i11m;1N2N2N3M2N2O1N2N2N2N2O2M2N2N2N2O1N2N3M2N2O1N2N2N2N2N3N1N2N2N2N2O1N3M2_KlMUOV2k0XNeNj1[1cNWN^1i1oNhMS1X2[OXMg0h2FiL<W32YL1f3<kKFU4h0\\KYOd4T1mJnNS5o3000001O0000000001O0001O0000000000001O000000000000001O0000000000001O0001O0000000001O0000000000001O000000000000001O00000000000010O00000000000001O0000000000001O000000000000001O00000001O0001O000000000000001O0000000000001O000000000000001O01O0000000001O000000000000001O0000000000001O000000000001O01O0000000000001OL4K5L4K5K5K5K5L4K5K5K5K5L4K5K5K5K5L5J5A?\\Od0\\Od0]Oc0\\Od0\\Od0]OYbQ3"}, "label": "a small ibm thinkpad laptop sitting on top of a larger laptop both powered off"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"computer monitor with silver frame\"."}], "task": "refering", "answer_template": "The \"computer monitor with silver frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 111, 112, 113, 114], "bbox": [0.79209375, 0.0, 1.0, 0.359556135770235], "iscrowd": 0, "area": 15253.061900000002, "rle": {"size": [383, 640], "counts": "Ugm56];<E;E<C<E;D<E<C<O10001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001N1000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001N1000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001N10001O0000001O0000001O00001O0000001O"}, "label": "computer monitor with silver frame"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor with the silver border around the outside to the right of all the others\"."}], "task": "refering", "answer_template": "The \"the monitor with the silver border around the outside to the right of all the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 111, 112, 113, 114], "bbox": [0.79209375, 0.0, 1.0, 0.359556135770235], "iscrowd": 0, "area": 15253.061900000002, "rle": {"size": [383, 640], "counts": "Ugm56];<E;E<C<E;D<E<C<O10001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001N1000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001N1000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001N10001O0000001O0000001O00001O0000001O"}, "label": "the monitor with the silver border around the outside to the right of all the others"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black laptop computer underneath a smaller laptop\"."}], "task": "refering", "answer_template": "The \"a black laptop computer underneath a smaller laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 174, 175, 176, 197, 198, 199, 212, 213, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 313, 314], "bbox": [0.188265625, 0.013472584856396868, 0.71678125, 0.9483028720626633], "iscrowd": 0, "area": 47491.21285000001, "rle": {"size": [383, 640], "counts": "Sb]13k;1N2N2N2N2N2O2M2N2N2N2O1N2N3M2N2O1N2N2N2N3N1N2N2N2N2O2N10000001O0001O00000001O0001O000001O000001O0001O00000001O01O00000001O01OdGTOe5l0[JOj41VKOj41VKOj41VKOj42UKNk42UKNl41TKOl41UKNk42UKNk42UKNk42UKNk43TKMm42SKNm42SKNm42SKNm42SKNm42SKNm42SKNn42QKNo42QKNo42QKNo42QKNo42QKNo42QKNP51PKOP52oJNQ52oJNQ52oJNQ52oJNQ52oJOQ50oJ0Q51nJOR51nJOR51nJOR51oJNQ52oJNQ52oJNR51nJOR52mJNS52mJNS52mJNS52mJNS52mJNT51lJOT51lJOT52kJNU52kJNU52kJNU52kJNV51jJOV51jJOV51jJOV52iJNW52iJNW52iJNX51hJOX51hJOX51hJOX51hJOX52gJNY52gJNZ51gJOX51hJOX51hJOX51hJOX52gJNY52gJNZ51fJOZ51fJOZ51fJOZ51fJOZ51fJOZ52eJN[52eJN\\51dJO\\51dJO\\51dJO\\51dJO\\51dJO\\52cJN^51bJO^51bJO^51bJO^51bJO^51bJO^51bJO_51`JO`51`JO`51aJN_52aJN_52aJN_52aJN`51`JO`52_JNa52_JO`51`JO`51`JO`51`JOa50_J0a51^JOb51^JOb51^JOb51^JOb51^JOb51^JOc50]J0c51\\JOd51\\JOd51\\JOd51\\JOd51\\JOe50[J0e50[J0e51ZJOf51ZJOf51[JNe52[JNf51ZJOf51ZJOf51ZJOf52YJNg52YJNg52YJNh51XJOh51XJOh51XJOh51XJ0g51XJOh51XJOi50WJ0i50WJ0i50WJ0i50WJ0i51VJOj51VJOj51VJOk50UJ0k50UJ0k50UJ0k50UJ0k51TJOl51UJNl51TJOl51TJOl51TJOl51TJOl51TJOl52SJNn51RJOn51RJOn51RJOn51RJOn51RJOn51RJOo51PJOP61PJOP61PJOP61PJOP61PJ0o50QJ0P6OPJ1P60oI0Q60oI0Q60oI0Q60oI0Q60oI0R6OnI1o52QJNk57UJHg5<YJDb5a0^J_O^5e0bJ[OY5j0gJVOU5n0kJROQ5S1nJmNm4X1SKhNi4\\1WKdNd4a1\\K_N`4e1`K[N\\4i1dKWNW4n1iKRNS4S2lKmMo3X2QLhMk3\\2ULdMg3`2YL`Mb3e2^L[M^3i2bLWMZ3m2fLSMU3S3jLmLa2h3_MXLi1`4WNaKQ1V5oNjJ:m5FSJBe6>[IkN\\7U1h10O000000000001O00000000001O0000000000001O00000000000010O000000000001O0000000000001O000000M3G9F:G9F:G9F:G9F:G:E:G9F:G9F:G9E;E;E;E;D<E;E;E;E;E;E;E;E;E;E;E;E;EWfS2"}, "label": "a black laptop computer underneath a smaller laptop"}]}
{"id": 530097, "image": "COCO_train2014_000000530097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop with another smaller laptop on top\"."}], "task": "refering", "answer_template": "The \"a laptop with another smaller laptop on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 174, 175, 176, 197, 198, 199, 212, 213, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 313, 314], "bbox": [0.188265625, 0.013472584856396868, 0.71678125, 0.9483028720626633], "iscrowd": 0, "area": 47491.21285000001, "rle": {"size": [383, 640], "counts": "Sb]13k;1N2N2N2N2N2O2M2N2N2N2O1N2N3M2N2O1N2N2N2N3N1N2N2N2N2O2N10000001O0001O00000001O0001O000001O000001O0001O00000001O01O00000001O01OdGTOe5l0[JOj41VKOj41VKOj41VKOj42UKNk42UKNl41TKOl41UKNk42UKNk42UKNk42UKNk43TKMm42SKNm42SKNm42SKNm42SKNm42SKNm42SKNn42QKNo42QKNo42QKNo42QKNo42QKNo42QKNP51PKOP52oJNQ52oJNQ52oJNQ52oJNQ52oJOQ50oJ0Q51nJOR51nJOR51nJOR51oJNQ52oJNQ52oJNR51nJOR52mJNS52mJNS52mJNS52mJNS52mJNT51lJOT51lJOT52kJNU52kJNU52kJNU52kJNV51jJOV51jJOV51jJOV52iJNW52iJNW52iJNX51hJOX51hJOX51hJOX51hJOX52gJNY52gJNZ51gJOX51hJOX51hJOX51hJOX52gJNY52gJNZ51fJOZ51fJOZ51fJOZ51fJOZ51fJOZ52eJN[52eJN\\51dJO\\51dJO\\51dJO\\51dJO\\51dJO\\52cJN^51bJO^51bJO^51bJO^51bJO^51bJO^51bJO_51`JO`51`JO`51aJN_52aJN_52aJN_52aJN`51`JO`52_JNa52_JO`51`JO`51`JO`51`JOa50_J0a51^JOb51^JOb51^JOb51^JOb51^JOb51^JOc50]J0c51\\JOd51\\JOd51\\JOd51\\JOd51\\JOe50[J0e50[J0e51ZJOf51ZJOf51[JNe52[JNf51ZJOf51ZJOf51ZJOf52YJNg52YJNg52YJNh51XJOh51XJOh51XJOh51XJ0g51XJOh51XJOi50WJ0i50WJ0i50WJ0i50WJ0i51VJOj51VJOj51VJOk50UJ0k50UJ0k50UJ0k50UJ0k51TJOl51UJNl51TJOl51TJOl51TJOl51TJOl51TJOl52SJNn51RJOn51RJOn51RJOn51RJOn51RJOn51RJOo51PJOP61PJOP61PJOP61PJOP61PJ0o50QJ0P6OPJ1P60oI0Q60oI0Q60oI0Q60oI0Q60oI0R6OnI1o52QJNk57UJHg5<YJDb5a0^J_O^5e0bJ[OY5j0gJVOU5n0kJROQ5S1nJmNm4X1SKhNi4\\1WKdNd4a1\\K_N`4e1`K[N\\4i1dKWNW4n1iKRNS4S2lKmMo3X2QLhMk3\\2ULdMg3`2YL`Mb3e2^L[M^3i2bLWMZ3m2fLSMU3S3jLmLa2h3_MXLi1`4WNaKQ1V5oNjJ:m5FSJBe6>[IkN\\7U1h10O000000000001O00000000001O0000000000001O00000000000010O000000000001O0000000000001O000000M3G9F:G9F:G9F:G9F:G:E:G9F:G9F:G9E;E;E;E;D<E;E;E;E;E;E;E;E;E;E;E;E;EWfS2"}, "label": "a laptop with another smaller laptop on top"}]}
{"id": 382643, "image": "COCO_train2014_000000382643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on skis with a red jacket tied on her waist\"."}], "task": "refering", "answer_template": "The \"the woman on skis with a red jacket tied on her waist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 142, 143, 165, 166, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304, 305, 326, 327], "bbox": [0.176171875, 0.29552083333333334, 0.302015625, 0.86375], "iscrowd": 0, "area": 12318.300350000003, "rle": {"size": [480, 640], "counts": "QWe13l>5J5L5K5RN[OWEj0d:\\OWEi0c:]OXEh0b:_OWEg0c:_OXEf0c1VOe5:cHe0o0HZ6JaHc0;;P7YO^Hb0EP1i7eN\\Hb3`7eLZH_3b7Q2N1O1O2N1O1N3N1N2I8I501O001O0010O01O001O00]Od0UOk0XNh1001O000001RLWFd2h9ZM]Fc2c9[MaFc2_9[MeFi2U9UMPGS3e8lL_GZ3Y8dLlG[3R8cLSH\\3l7aL`HW3^7gLfHY3X7eLlH\\3Q7bLSI^3k6`LXIa3f6]L^Ic3g6TL]Il3l81O1O2M2O1O11OMUMXE_1h:`NeET1Z:kNRFl0l9TOYFh0f9WO_Ff0`9aN]G]1a8bNdGY1^8eNhGT1Z8kNkGP1V8nNPHl0S8RORHh0P8UOXHd0j7YO^H?Z;GaUa6"}, "label": "the woman on skis with a red jacket tied on her waist"}]}
{"id": 382643, "image": "COCO_train2014_000000382643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing black pants with a red jacket tied around her waist\"."}], "task": "refering", "answer_template": "The \"a lady wearing black pants with a red jacket tied around her waist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 142, 143, 165, 166, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304, 305, 326, 327], "bbox": [0.176171875, 0.29552083333333334, 0.302015625, 0.86375], "iscrowd": 0, "area": 12318.300350000003, "rle": {"size": [480, 640], "counts": "QWe13l>5J5L5K5RN[OWEj0d:\\OWEi0c:]OXEh0b:_OWEg0c:_OXEf0c1VOe5:cHe0o0HZ6JaHc0;;P7YO^Hb0EP1i7eN\\Hb3`7eLZH_3b7Q2N1O1O2N1O1N3N1N2I8I501O001O0010O01O001O00]Od0UOk0XNh1001O000001RLWFd2h9ZM]Fc2c9[MaFc2_9[MeFi2U9UMPGS3e8lL_GZ3Y8dLlG[3R8cLSH\\3l7aL`HW3^7gLfHY3X7eLlH\\3Q7bLSI^3k6`LXIa3f6]L^Ic3g6TL]Il3l81O1O2M2O1O11OMUMXE_1h:`NeET1Z:kNRFl0l9TOYFh0f9WO_Ff0`9aN]G]1a8bNdGY1^8eNhGT1Z8kNkGP1V8nNPHl0S8RORHh0P8UOXHd0j7YO^H?Z;GaUa6"}, "label": "a lady wearing black pants with a red jacket tied around her waist"}]}
{"id": 374453, "image": "COCO_train2014_000000374453.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic signal\"."}], "task": "refering", "answer_template": "The \"traffic signal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 108, 109, 121, 122, 134, 135, 147, 148, 149, 160, 161], "bbox": [0.3082933333333333, 0.39036000000000004, 0.47381333333333336, 0.69596], "iscrowd": 0, "area": 6945.344950000001, "rle": {"size": [500, 375], "counts": "]lh16X?6K5O1_A1[=1dB2=Ha;7PD5:Ic;3RD85Hl:^OZEc0D:3Fn:BUE`0Hb0L]OW;DQEd0E[1X;UNoDo2P;UMlDl2S;k000O2O0O100O1O2O0O100O100O2O0O1O11O010O1O1O00100O1O1O0O2O1N2O1N101N2O1N108G;F:E;E:G:E;E;F:E9D8H:G9FWlo2"}, "label": "traffic signal"}]}
{"id": 374453, "image": "COCO_train2014_000000374453.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic signal\"."}], "task": "refering", "answer_template": "The \"traffic signal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 108, 109, 121, 122, 134, 135, 147, 148, 149, 160, 161], "bbox": [0.3082933333333333, 0.39036000000000004, 0.47381333333333336, 0.69596], "iscrowd": 0, "area": 6945.344950000001, "rle": {"size": [500, 375], "counts": "]lh16X?6K5O1_A1[=1dB2=Ha;7PD5:Ic;3RD85Hl:^OZEc0D:3Fn:BUE`0Hb0L]OW;DQEd0E[1X;UNoDo2P;UMlDl2S;k000O2O0O100O1O2O0O100O100O2O0O1O11O010O1O1O00100O1O1O0O2O1N2O1N101N2O1N108G;F:E;E:G:E;E;F:E9D8H:G9FWlo2"}, "label": "traffic signal"}]}
{"id": 579255, "image": "COCO_train2014_000000579255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table in the middle with the vase on it\"."}], "task": "refering", "answer_template": "The \"the table in the middle with the vase on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 239, 240, 241, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 332, 333, 335, 336, 355, 356, 378, 379], "bbox": [0.357046875, 0.5293958333333334, 0.636265625, 0.9790833333333334], "iscrowd": 0, "area": 14801.937299999998, "rle": {"size": [480, 640], "counts": "^e[3:a><D9J6K4L5L3L4L5K4M7H7I2O1N2NWOj00O2N100O101O0000001O0O2O00001O00001O001N1000000000001O00000O100000001O0000001N101O1O2N3M3M2N3M2N3YOTMSEP3h:TMVEn2f:VMXEm2c:WM[El2`:WM_Ek2i9kLZF>Kk2g9oLTFT4h9e0L5TGlJd7W5PHRKP8P6O1O1O1O1O1OF:YNg1N1O2N2N2O2O1O2N1O4L6J1N101O1O1O0XMPE`1Q;XNVEh1k:PN\\EP2j;000001O0000000001O000001O001O001O00010O001O00001O001O00001O001O00001O001O001O00001O0O2O000O2O001O0O101O001OTCWNa<i1ZC\\Ng<m110001O001N10001O0N3L3N5J8H7J7H7J6I5L4K5Kck\\3"}, "label": "the table in the middle with the vase on it"}]}
{"id": 579255, "image": "COCO_train2014_000000579255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small brown coffee table\"."}], "task": "refering", "answer_template": "The \"a small brown coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 239, 240, 241, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 332, 333, 335, 336, 355, 356, 378, 379], "bbox": [0.357046875, 0.5293958333333334, 0.636265625, 0.9790833333333334], "iscrowd": 0, "area": 14801.937299999998, "rle": {"size": [480, 640], "counts": "^e[3:a><D9J6K4L5L3L4L5K4M7H7I2O1N2NWOj00O2N100O101O0000001O0O2O00001O00001O001N1000000000001O00000O100000001O0000001N101O1O2N3M3M2N3M2N3YOTMSEP3h:TMVEn2f:VMXEm2c:WM[El2`:WM_Ek2i9kLZF>Kk2g9oLTFT4h9e0L5TGlJd7W5PHRKP8P6O1O1O1O1O1OF:YNg1N1O2N2N2O2O1O2N1O4L6J1N101O1O1O0XMPE`1Q;XNVEh1k:PN\\EP2j;000001O0000000001O000001O001O001O00010O001O00001O001O00001O001O00001O001O001O00001O0O2O000O2O001O0O101O001OTCWNa<i1ZC\\Ng<m110001O001N10001O0N3L3N5J8H7J7H7J6I5L4K5Kck\\3"}, "label": "a small brown coffee table"}]}
{"id": 579255, "image": "COCO_train2014_000000579255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair in front of a window\"."}], "task": "refering", "answer_template": "The \"a chair in front of a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 236, 237, 238, 239, 240, 259, 260, 261, 282, 283, 284], "bbox": [0.23529687500000002, 0.4451875, 0.46328125, 0.7614583333333333], "iscrowd": 0, "area": 12611.0989, "rle": {"size": [480, 640], "counts": "WPW25j>6J5K5K3M3M4M2N2N2N2N3M2N2N2N3M2N2N2N8H6I6K6I6K6I7I6K6I3N0O2O1N2O0O2O1N3N2N2N3M00001O00001O0O101N1000101N1O2O1N1O2NO1O1N1O2O101O0001O0001O000000001O0000000QO^E`Mc:]2dE_M[:^2U1N2O101N10000O1000O0010O0100O010O100O010001N2O001O1]NWDLj;1YDOh;L]D3d;I_D7b;F`D9b;DaD;`;AcD?^;^OeD`0^;\\OdDd0];YOeDg0k<O001O001O01O01O001O001O00001O001O001O00001O0^OTB0_>O2O0O1O2N1OZnP5"}, "label": "a chair in front of a window"}]}
{"id": 579255, "image": "COCO_train2014_000000579255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch is front of a table\"."}], "task": "refering", "answer_template": "The \"a couch is front of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 234, 253, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 355, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.016125, 0.526875, 0.47096875000000005, 0.9892500000000001], "iscrowd": 0, "area": 33303.05820000001, "rle": {"size": [480, 640], "counts": "Vn41f>:[AFb>:^AJ^>`0L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4M4K4L4L4M3L4L4L4L4M3L2N2N2N2O2M2N23M5K4M3L4L4L4L4M1NN2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N000000000001O0000000000000000000O1000000000000000001O000000000000O1000000000000O1000000000000O100000000000aKWFg3i9VLXFl3g9RL[Fo3e9nK]FS4c9hKaFY4_9aKfF`4Z9_KfFb4Z9]KfFd4Z9[KfFf4j9000000000000000000O2O00001O0000001O00001O1O001O1O001O001O1O001O1O001O001O1O001O1O1O001O1O3M4L1O1O1O1O00001O000000001O0000002N1O1N3N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1O001O00001O001O001O001O1O1O1O1O1O1O1O1O1O001O001O001O001O001O001O001O001O000O2O001O001O1O001O1O1O1O001O1O1O001O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O1O2M2J7IY_n4"}, "label": "a couch is front of a table"}]}
{"id": 579255, "image": "COCO_train2014_000000579255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a love seat to the left of the coffee table\"."}], "task": "refering", "answer_template": "The \"a love seat to the left of the coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 230, 231, 232, 234, 253, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 355, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.016125, 0.526875, 0.47096875000000005, 0.9892500000000001], "iscrowd": 0, "area": 33303.05820000001, "rle": {"size": [480, 640], "counts": "Vn41f>:[AFb>:^AJ^>`0L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4M4K4L4L4M3L4L4L4L4M3L2N2N2N2O2M2N23M5K4M3L4L4L4L4M1NN2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N000000000001O0000000000000000000O1000000000000000001O000000000000O1000000000000O1000000000000O100000000000aKWFg3i9VLXFl3g9RL[Fo3e9nK]FS4c9hKaFY4_9aKfF`4Z9_KfFb4Z9]KfFd4Z9[KfFf4j9000000000000000000O2O00001O0000001O00001O1O001O1O001O001O1O001O1O001O001O1O001O1O1O001O1O3M4L1O1O1O1O00001O000000001O0000002N1O1N3N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1O001O00001O001O001O001O1O1O1O1O1O1O1O1O1O001O001O001O001O001O001O001O001O000O2O001O001O1O001O1O1O1O001O1O1O001O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O1O2M2J7IY_n4"}, "label": "a love seat to the left of the coffee table"}]}
{"id": 87737, "image": "COCO_train2014_000000087737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cylindrical toilet in a dirty restroom between toilet paper and a urinal\"."}], "task": "refering", "answer_template": "The \"a cylindrical toilet in a dirty restroom between toilet paper and a urinal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.40528125, 0.4639583333333333, 0.65428125, 0.848125], "iscrowd": 0, "area": 23950.24495, "rle": {"size": [480, 640], "counts": "igi31n>6I7J6I7J6I3N3L3N2M4M2M3N3L3N2YOh0WOh0WOi0L5M2N2N2N1N3N1O2N100O2O0O2O0O2O0O101N101N10001N101N101N100O2O0O2O0O101O0O2O0O100O1000000O100000000O100000000O1000000O100000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000O100O2O0O101O0O101N100O2O0O101O0O101N101N101O1N2O0O2O1N2O2M2O2N1M4L3N3lMdE^Oc:3kEG\\:IRF2U:_OXF=n9SO`Fh0o;J7H7I8GQXW3"}, "label": "a cylindrical toilet in a dirty restroom between toilet paper and a urinal"}]}
{"id": 87737, "image": "COCO_train2014_000000087737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a completely cylindrical toilet\"."}], "task": "refering", "answer_template": "The \"a completely cylindrical toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.40528125, 0.4639583333333333, 0.65428125, 0.848125], "iscrowd": 0, "area": 23950.24495, "rle": {"size": [480, 640], "counts": "igi31n>6I7J6I7J6I3N3L3N2M4M2M3N3L3N2YOh0WOh0WOi0L5M2N2N2N1N3N1O2N100O2O0O2O0O2O0O101N101N10001N101N101N100O2O0O2O0O101O0O2O0O100O1000000O100000000O100000000O1000000O100000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000O100O2O0O101O0O101N100O2O0O101O0O101N101N101O1N2O0O2O1N2O2M2O2N1M4L3N3lMdE^Oc:3kEG\\:IRF2U:_OXF=n9SO`Fh0o;J7H7I8GQXW3"}, "label": "a completely cylindrical toilet"}]}
{"id": 87737, "image": "COCO_train2014_000000087737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cream diamond shape urinal\"."}], "task": "refering", "answer_template": "The \"a cream diamond shape urinal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 47, 48, 49, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 258, 278, 279], "bbox": [0.04390625, 0.09625, 0.29859375, 0.7402708333333333], "iscrowd": 0, "area": 35273.76635, "rle": {"size": [480, 640], "counts": "nX=3i>9G9F;F9F:G9G9F:G9F:G:F9F:G9F:G9G9F:G:E:G9G9F:G9F:I7K5K6J5L4K5L4L4K5L4K6K4L3L01O1N101O0O2O1N101O0O2O1O0O2O001N2O001N2O0O2O001N2O001N101O1N101O0O2O1O0O2O0O2O1O0O2O1O0O2O001N2O001N101N2O001N101O1N101O1N101O0O2O1N101O0O2O1O0O2O0O2N2O0O2N2O0O2O0O2N2O0O2N101N2O0O2N101N2N101N101N2N101N2N101N101N2N2O<C<D<E;D=C<E;D=D;D<D=D;D<D<E<CPVb6"}, "label": "a cream diamond shape urinal"}]}
{"id": 87737, "image": "COCO_train2014_000000087737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a urinal\"."}], "task": "refering", "answer_template": "The \"a urinal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 47, 48, 49, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 258, 278, 279], "bbox": [0.04390625, 0.09625, 0.29859375, 0.7402708333333333], "iscrowd": 0, "area": 35273.76635, "rle": {"size": [480, 640], "counts": "nX=3i>9G9F;F9F:G9G9F:G9F:G:F9F:G9F:G9G9F:G:E:G9G9F:G9F:I7K5K6J5L4K5L4L4K5L4K6K4L3L01O1N101O0O2O1N101O0O2O1O0O2O001N2O001N2O0O2O001N2O001N101O1N101O0O2O1O0O2O0O2O1O0O2O1O0O2O001N2O001N101N2O001N101O1N101O1N101O0O2O1N101O0O2O1O0O2O0O2N2O0O2N2O0O2O0O2N2O0O2N101N2O0O2N101N2N101N101N2N101N2N101N101N2N2O<C<D<E;D=C<E;D=D;D<D=D;D<D<E<CPVb6"}, "label": "a urinal"}]}
{"id": 218809, "image": "COCO_train2014_000000218809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cooked carrots on a plate next to green beans\"."}], "task": "refering", "answer_template": "The \"cooked carrots on a plate next to green beans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 264, 265, 266], "bbox": [0.427359375, 0.5531024930747922, 0.6211249999999999, 0.8778116343490304], "iscrowd": 0, "area": 8490.755899999998, "rle": {"size": [361, 640], "counts": "bjP33S;4L3L4M3O2O0O1O2O0O1O2O0O1O2O0O1O1O2N1O1O2N1O1O2N1O1O1O1O2N1O100O101N3M4M2M4M3MO1001aGQNb7o1\\HWN`7i1^H\\N`7c1^HcNa7Z1^HkN`7T1]HQOb7n0[HVOe7j0WHWOl7i0PHXOS8m13M1O1O1O1O1O11O1O100N2O1O1O1O0100000000O1O1O1O1N10O1O100O1O1O100G9O100O100O11O001N2O001O001O1M2N3N1N3N2M2O2M3M2O3L4M3L4M3L4L4M3L5L3L4M3L5K4M;D?BcQe2"}, "label": "cooked carrots on a plate next to green beans"}]}
{"id": 218809, "image": "COCO_train2014_000000218809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pile of carrots on a plate\"."}], "task": "refering", "answer_template": "The \"a pile of carrots on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 264, 265, 266], "bbox": [0.427359375, 0.5531024930747922, 0.6211249999999999, 0.8778116343490304], "iscrowd": 0, "area": 8490.755899999998, "rle": {"size": [361, 640], "counts": "bjP33S;4L3L4M3O2O0O1O2O0O1O2O0O1O2O0O1O1O2N1O1O2N1O1O2N1O1O1O1O2N1O100O101N3M4M2M4M3MO1001aGQNb7o1\\HWN`7i1^H\\N`7c1^HcNa7Z1^HkN`7T1]HQOb7n0[HVOe7j0WHWOl7i0PHXOS8m13M1O1O1O1O1O11O1O100N2O1O1O1O0100000000O1O1O1O1N10O1O100O1O1O100G9O100O100O11O001N2O001O001O1M2N3N1N3N2M2O2M3M2O3L4M3L4M3L4L4M3L5L3L4M3L5K4M;D?BcQe2"}, "label": "a pile of carrots on a plate"}]}
{"id": 218809, "image": "COCO_train2014_000000218809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clean plate with fruit design\"."}], "task": "refering", "answer_template": "The \"a clean plate with fruit design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 24, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74], "bbox": [0.053359375, 0.055180055401662056, 0.3265, 0.24886426592797786], "iscrowd": 0, "area": 9776.011649999999, "rle": {"size": [361, 640], "counts": "fQ<2Q;7I7M2N3L4M3M3M2N2O2N101N101O0O2O0O2O0O101N101N101N101N100O2O0O10000O100O100O100O100000000O10000000000000000000001N10000000000000000000000O10000000001O00000000000O1000000000000000001O00000000000000000O10000000001O0000000000O10000O10000O10000O101O000O10000O10000O10000O10000O10001O0O100O100O100O100O1O100O2O0O100O100O100O1O2O1N2O2M2N2M3M4L3M3M4L3M:Fklg4"}, "label": "a clean plate with fruit design"}]}
{"id": 218809, "image": "COCO_train2014_000000218809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate with no food on it\"."}], "task": "refering", "answer_template": "The \"a plate with no food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 24, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74], "bbox": [0.053359375, 0.055180055401662056, 0.3265, 0.24886426592797786], "iscrowd": 0, "area": 9776.011649999999, "rle": {"size": [361, 640], "counts": "fQ<2Q;7I7M2N3L4M3M3M2N2O2N101N101O0O2O0O2O0O101N101N101N101N100O2O0O10000O100O100O100O100000000O10000000000000000000001N10000000000000000000000O10000000001O00000000000O1000000000000000001O00000000000000000O10000000001O0000000000O10000O10000O10000O101O000O10000O10000O10000O10000O10001O0O100O100O100O100O1O100O2O0O100O100O100O1O2O1N2O2M2N2M3M4L3M3M4L3M:Fklg4"}, "label": "a plate with no food on it"}]}
{"id": 284348, "image": "COCO_train2014_000000284348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a collection of books visible over the right shoulder of a stuffed teddy bear\"."}], "task": "refering", "answer_template": "The \"a collection of books visible over the right shoulder of a stuffed teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164], "bbox": [0.003, 0.011238317757009345, 0.296046875, 0.4898831775700935], "iscrowd": 0, "area": 24615.312550000002, "rle": {"size": [428, 640], "counts": "bk0h5d700O2O000000000000000000000000000000000O100000000O10O1000O10O1000O010000O10O100000O01000000O1000O1000O10000000O1000000000001O0O1000000000000O1000000000000O10001O0000000O100000000000000O10000000001O0O100000000000[M]IaNb6;eJB\\5;iJBX5=kJ@V5`0mJ\\OT5d0nJXOT5g0oJVOR5j0PKSOP5n0RKoNo4P1UKlNl4T1VKiNk4W1WKfNj4Z1XKcNi4\\1[K`Nf4`1\\K\\Nf4d1\\KYNd4g1_KVNb4j1`KSNa4m1_300O1000O01000000O1000O010000O100000O010000O1000000O01000O1000000O01000O10000O10O1000O10000O101O1O0O2O003L3N3M3M3L3N3M3L3N3MQ_l5"}, "label": "a collection of books visible over the right shoulder of a stuffed teddy bear"}]}
{"id": 284348, "image": "COCO_train2014_000000284348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a series of childrens books , propped in a small cubby next to a teddy bear\"."}], "task": "refering", "answer_template": "The \"a series of childrens books , propped in a small cubby next to a teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 38, 54, 55, 56, 57, 58, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 150, 151, 152, 153, 175], "bbox": [0.377984375, 0.10651869158878506, 0.67321875, 0.4786915887850467], "iscrowd": 0, "area": 15277.692150000004, "rle": {"size": [428, 640], "counts": "`WU35m;7jD4R;:YD0d;P1L3M4L4L3M4L3M4L4L0000000000000000000000000101N2N1O2O1N2N2N2O1N2N0010O0001O01O0001O01O01O01O01O01O0002N;F;D8H01N1@iNkDW1T;mNiDT1V;oNfDR1Z;QOcDo0\\;UO`Dm0_;`000O2O000O2O0:F01N1POZNZFf1e9_NVFb1j9bNRF_1l9fNoE[1Q:iNkEW1T:nNgET1X:POdEP1\\:P1O2O0O1010O10001O010O001O001O001O001O001O00001O001O001O010O010O0100O010O00100O010OO2N1O2O0O2N1O2N1O2O0O1O2N1O2N101N1O2N1O2N101N1O2N1O2N1O101N1O1OK5L4K6J5K5K5L4K5K6J5L4K5K5K5K6K4K5K5K5K5L5JHcGkK^8n3c0I7L4N3M2N2N2O1N3N1N2OjYg2"}, "label": "a series of childrens books , propped in a small cubby next to a teddy bear"}]}
{"id": 284348, "image": "COCO_train2014_000000284348.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle of the three compartments of books behind the bear ' s arm\"."}], "task": "refering", "answer_template": "The \"middle of the three compartments of books behind the bear ' s arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 38, 54, 55, 56, 57, 58, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 150, 151, 152, 153, 175], "bbox": [0.377984375, 0.10651869158878506, 0.67321875, 0.4786915887850467], "iscrowd": 0, "area": 15277.692150000004, "rle": {"size": [428, 640], "counts": "`WU35m;7jD4R;:YD0d;P1L3M4L4L3M4L3M4L4L0000000000000000000000000101N2N1O2O1N2N2N2O1N2N0010O0001O01O0001O01O01O01O01O01O0002N;F;D8H01N1@iNkDW1T;mNiDT1V;oNfDR1Z;QOcDo0\\;UO`Dm0_;`000O2O000O2O0:F01N1POZNZFf1e9_NVFb1j9bNRF_1l9fNoE[1Q:iNkEW1T:nNgET1X:POdEP1\\:P1O2O0O1010O10001O010O001O001O001O001O001O00001O001O001O010O010O0100O010O00100O010OO2N1O2O0O2N1O2N1O2O0O1O2N1O2N101N1O2N1O2N101N1O2N1O2N1O101N1O1OK5L4K6J5K5K5L4K5K6J5L4K5K5K5K6K4K5K5K5K5L5JHcGkK^8n3c0I7L4N3M2N2N2O1N3N1N2OjYg2"}, "label": "middle of the three compartments of books behind the bear ' s arm"}]}
{"id": 104126, "image": "COCO_train2014_000000104126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vase on right\"."}], "task": "refering", "answer_template": "The \"vase on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 60, 61, 64, 65, 66, 67, 77, 78, 79, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 129, 130, 131, 133, 134, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 184, 196, 197, 198, 199, 200, 214, 215, 216, 217, 231, 232, 233, 234, 248, 249, 250, 265, 266, 267, 282, 283, 284, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 351, 352], "bbox": [0.5273333333333333, 0.11842187500000001, 0.9647916666666667, 0.886953125], "iscrowd": 0, "area": 45590.148550000005, "rle": {"size": [640, 480], "counts": "ocn42lc05J7J5K6I6K6J5\\HjNVL\\1POdNVN5c2[1VOfNQN3c2Y1\\OjNjM2d2W1AgNkM6_2T1FfNkM:[2Q1HfNlM>X2l0LfNlMb0T2h00fNkMg0P2e04dNlMk0l1a08dNlMo0h1=<dNkMT1e19>dNmMW1a15b0dNlM\\1]12f0bNmM`1Y1Nj0bNmMd1U1Jn0bNlMh1S1Go0bNnMk1n0DT1aNmMP2k0_OX1aNmMT2g0\\O[1`NnMX2c0XO_1`NmM]2?UOb1_NoM`2;QOf1_NnMe28lNj1_NnMi24iNm1^NnMn21dNQ2^NnMR3LaNU2^NoMU3H^NX2]NoMZ3EYN\\2]NoM^3AUN`2]NnMc3]ORNd2[NoMg3YOnMg2\\NPNj3UOkMj2[NPNn3SOgMm2[NPNR4nNdMR3ZNoMW4kN`MU3YNPN[4gN\\MY3POYMh3ZOXM]3QOWMl3WOUMb3oNVMP4TOQMg3oNSMU4ROmLk3nNRMY4oNiLP4oNoL\\4mNeLU4oNlLa4jNbLZ4mNnKlNjNi5j0^L_4mNkKg5B\\Kd4mNgKm5AWKh4lNeKe9\\4\\F`Kg9`4YF^Kh9c4XFZKj9g4VFWKl9i4TFTKn9m4b2100O100O1O100O1O100O100O1O100O1O1O1O1O1O1N2O1O1O1N2O1O1N0100000O100000O0100QLcI[J]6b5iI[JW6b5PJYJQ6e5TJXJk5f5[JWJe5f5bJUJ_5h5gJUJY5i5lJTJT5gN\\I[5g1jKm4hNcIW5f1mKh4iNiIS5e1QLa4kNPJm4c1WL]4iNWJj4_1\\LY4hN_Je4\\1aLV4gNeJa4Y1gLQ4gNlJ[4W1mLm3eNSKW4T1SMh3dN[KS4P1WMf3fN^Kl3P1]Mb3fNbKg3P1bM]3hNcKc3T1dMY3iNdK_3V1gMV3jNfK\\3W1iMS3kNgKX3Z1lMn2mNhKT3^1mMP2TNlJk0g1P3a1PNk1XNjJh0l1m2b1RNg1\\NiJe0o1i2e1UNb1_NiJc0Q2e2h1WN^1dNfJ`0U2b2k1YNZ1gNdJ>X2^2n1\\NU1kNbJ<\\2Z2P2^NQ1oNaJ8`2W2R2`Nm0TO^J6c2S2V2bNh0XO]J3c0SO[1l2Q3eNc0[O]J1c0UO\\1f2U3hN>=QKWN\\1a2X3iN;`0PKWN]1]2\\3kN7a0oJYN^1W2`3nN2b0QKZN]1R2d3PONe0PKZN_1m1g3SOIg0PK[N_1i1k3TOEh0QK]Nn0CoNP2`5XOAi0QK]Nm0FPOh1e5ZO]Ol0PK^Nl0FROb1_<;cC^Ni0HTO]1b<\\OQCL`0Dg0JVOV1d<_OQCM=Ff0KWOo0g<BRCM9Id0GnNF<W1i<DTCM5Lb0GROE9T1l<FTCN2Na0EUOG7o0n<HVCNO0>EYOG6j0P=KWCMM2<E\\OG4f0R=MXCNO04I@G2b0V=NXCOOONLDH0<Y=1XCO2LH0GHN9Z=2[C01KB2LIL4\\=5\\CO3JZO60JIOa=6[C14GQO>8DHLd=7ZC26d0UOVOIJd=;XCO6f0VOVOIFf=?UCO6e0XOUOk=7gBM6g0YOUOj=8fBL6g0[OTOk=9dBJ7h0\\OTOi=;dBH6h0_OTOi=<aBG8h0_OUOh==aBE7i0BTOg=>`BC7k0CSOg=`0^BA9k0CTOf=a0^B_O8l0FROf=c0[B^O:l0FSOe=d0[B\\O:l0GSOf=e0eC8`=H`B6c=J[B6f=KYB4i=LUB4l=MSB1o=0PBOR>1lAOU>2jAMX>3fAL\\>5_ANc>2YA0h>1SA1o>0l@3V?Me@6\\?Ja@8a?HZ@:h?GS@<o?Dm_O>T`0Cg_O`0[`0@`_Ob0b`0_OZ_Oc0g`0^OT_Oe0n`0[Om^Oh0Ta0P13N1O2M2O2N1O1N3N1O2N1N3NV1jNi_:"}, "label": "vase on right"}]}
{"id": 104126, "image": "COCO_train2014_000000104126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase in the left\"."}], "task": "refering", "answer_template": "The \"the vase in the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 54, 55, 56, 57, 58, 59, 60, 71, 72, 73, 74, 75, 76, 77, 88, 89, 90, 91, 92, 93, 94, 105, 106, 107, 108, 109, 110, 111, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 294, 295, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364], "bbox": [0.16779166666666667, 0.094, 0.5573125, 0.9434531249999999], "iscrowd": 0, "area": 63954.22169999998, "rle": {"size": [640, 480], "counts": "]lb17hc07I7J6I7I8H7cKgNfE`1Q3^NE>VM[1n2^N^Oa0aMX1i2_NWOe0lMS1e2]NWOh0PNS1a2YN[Oh0PNU1^2XN]Oh0QNU1]2WN^Og0SNU1Z2YN^Og0TNT1Y2ZN_Of0TNT1Y2XN@i0SNS1X2WNCj0QNS1X2VNCk0RNS1V2UNFl0PNT1T2SNIn0oMS1T2QNKP1mMS1T2PNKR1mMR1S2oMNS1lMQ1R2nMOV1kMP1R2lM1X1iMP1Q2kM3[1gMn0R2jM4\\1fMo0Q2gM6_1eMn0Q2dM9b1aMo0Q2bM<c1^MP1R2^M?f1YMR1U2YM`0i1VMR1W2VMc0k1QMS1Z2SMc0n1oLQ1]2QMd0Q2kLQ1^2PMe0T2hLn0b2nLe0Y8XOhGh0Z8VOgGh0[8VOfGj0\\8TOeGj0^8TObGk0`8SObGl0`8RO`Gm0b8QO`Go0a8oN_GR1b8lN_GT1a8jN`GW1a8gN`GX1b8fN^G[1b8dN^G]1c8`N]Gb1d8\\N\\Ge1d8ZN\\Gg1e8WN[Gj1e8TN\\Go1c8mLVDe0W3c2`8cL^Dh0R3g2_8^LbDj0o2j2k8UMUGm2k8PMVGR3i8hL[G[3d8`L`Ga3`8YLeGi3[8PLjGV4m;5KO10000O01000O100O10000O10O10O10000O10000O10O10O1001O00001N10000[LjK`GW4]8lKbGT4\\8oKcGQ4[8RLcGP4Z8TLdGl3Y8XLfGh3X8[LgGe3l7iLSHX3j7kLUHU3i7nLUHS3i7PMVHQ3g7RMXHn2f7UMYHk2e7XMZHh2e7YM[Hg2c7\\M[He2c7]M]Hc2a7_M_Ha2_7bM`H^2]7eMcH[2[7hMdHX2Z7jMeHW2X7mMgHS2W7oMiHQ2U7RNjHn1T7TNlHl1Q7XNnHh1P7ZNoHg1o6[NQIe1m6^NRIa1l6bNTI^1j6eNUI[1i6gNWIY1g6jNWIV1g6mNYIS1c6RO[Io0`6VO_Ik0]6ZOaIg0[6]OdIc0Y6AfI`0U6FhI<T6HkI9Q6LmI5o50oI0m56QJKj5;TJFf5a0XJ@b5g0\\JZO_5l0_JTO\\5R1cJoNW5X1gJhNU5^1hJcNh2`K^MT6H]Nn1aLnMY52VNn1iLeMX5;PNn1kLaM[5?jMo1oL\\M\\5c0fMo1QMYM]5g0bMP2TMTM^5j0_MQ2UMPMa5m0ZMS2XMkLb5P1WMU2Z5iMfJW2\\5gMdJY2^5eMcJZ2_5cMcJ]2]5bMeJ\\2f5YM[Jf2`;O2N1O2N1O1N3N1O1O2M2O2_Ok^OTNVa0h1c0L5eNh^OI[a04Z1M4L3MUQU4"}, "label": "the vase in the left"}]}
{"id": 104126, "image": "COCO_train2014_000000104126.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the container on the left of the table when looking straight at the picture\"."}], "task": "refering", "answer_template": "The \"the container on the left of the table when looking straight at the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 54, 55, 56, 57, 58, 59, 60, 71, 72, 73, 74, 75, 76, 77, 88, 89, 90, 91, 92, 93, 94, 105, 106, 107, 108, 109, 110, 111, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 294, 295, 309, 310, 311, 312, 313, 314, 326, 327, 328, 329, 330, 331, 343, 344, 345, 346, 347, 348, 360, 361, 362, 363, 364], "bbox": [0.16779166666666667, 0.094, 0.5573125, 0.9434531249999999], "iscrowd": 0, "area": 63954.22169999998, "rle": {"size": [640, 480], "counts": "]lb17hc07I7J6I7I8H7cKgNfE`1Q3^NE>VM[1n2^N^Oa0aMX1i2_NWOe0lMS1e2]NWOh0PNS1a2YN[Oh0PNU1^2XN]Oh0QNU1]2WN^Og0SNU1Z2YN^Og0TNT1Y2ZN_Of0TNT1Y2XN@i0SNS1X2WNCj0QNS1X2VNCk0RNS1V2UNFl0PNT1T2SNIn0oMS1T2QNKP1mMS1T2PNKR1mMR1S2oMNS1lMQ1R2nMOV1kMP1R2lM1X1iMP1Q2kM3[1gMn0R2jM4\\1fMo0Q2gM6_1eMn0Q2dM9b1aMo0Q2bM<c1^MP1R2^M?f1YMR1U2YM`0i1VMR1W2VMc0k1QMS1Z2SMc0n1oLQ1]2QMd0Q2kLQ1^2PMe0T2hLn0b2nLe0Y8XOhGh0Z8VOgGh0[8VOfGj0\\8TOeGj0^8TObGk0`8SObGl0`8RO`Gm0b8QO`Go0a8oN_GR1b8lN_GT1a8jN`GW1a8gN`GX1b8fN^G[1b8dN^G]1c8`N]Gb1d8\\N\\Ge1d8ZN\\Gg1e8WN[Gj1e8TN\\Go1c8mLVDe0W3c2`8cL^Dh0R3g2_8^LbDj0o2j2k8UMUGm2k8PMVGR3i8hL[G[3d8`L`Ga3`8YLeGi3[8PLjGV4m;5KO10000O01000O100O10000O10O10O10000O10000O10O10O1001O00001N10000[LjK`GW4]8lKbGT4\\8oKcGQ4[8RLcGP4Z8TLdGl3Y8XLfGh3X8[LgGe3l7iLSHX3j7kLUHU3i7nLUHS3i7PMVHQ3g7RMXHn2f7UMYHk2e7XMZHh2e7YM[Hg2c7\\M[He2c7]M]Hc2a7_M_Ha2_7bM`H^2]7eMcH[2[7hMdHX2Z7jMeHW2X7mMgHS2W7oMiHQ2U7RNjHn1T7TNlHl1Q7XNnHh1P7ZNoHg1o6[NQIe1m6^NRIa1l6bNTI^1j6eNUI[1i6gNWIY1g6jNWIV1g6mNYIS1c6RO[Io0`6VO_Ik0]6ZOaIg0[6]OdIc0Y6AfI`0U6FhI<T6HkI9Q6LmI5o50oI0m56QJKj5;TJFf5a0XJ@b5g0\\JZO_5l0_JTO\\5R1cJoNW5X1gJhNU5^1hJcNh2`K^MT6H]Nn1aLnMY52VNn1iLeMX5;PNn1kLaM[5?jMo1oL\\M\\5c0fMo1QMYM]5g0bMP2TMTM^5j0_MQ2UMPMa5m0ZMS2XMkLb5P1WMU2Z5iMfJW2\\5gMdJY2^5eMcJZ2_5cMcJ]2]5bMeJ\\2f5YM[Jf2`;O2N1O2N1O1N3N1O1O2M2O2_Ok^OTNVa0h1c0L5eNh^OI[a04Z1M4L3MUQU4"}, "label": "the container on the left of the table when looking straight at the picture"}]}
{"id": 440002, "image": "COCO_train2014_000000440002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chef holding a silver bowl\"."}], "task": "refering", "answer_template": "The \"chef holding a silver bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 139, 140, 141, 142, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 300, 301, 302, 303, 318, 319, 320, 336, 337, 338, 354, 355, 356, 372, 373, 374, 390, 391, 409], "bbox": [0.6331781376518218, 0.28153125, 0.9774898785425101, 0.97296875], "iscrowd": 0, "area": 36957.62215000001, "rle": {"size": [640, 494], "counts": "elS67gc0;F:E:G:E;E3N2M3f_OoMb>S2XAdNT>_1fAXOf=l0SBI\\=9^B0\\=3^B6[=N_B;X=KbB>T=IfBa0P=^3F:K5L4L4L4K5gEZIP7l9O1O0O2O1O1O001N2O1O001N2O1O001N2O1O1O0O2O1O1O0O2O1O1O0O2O1O1O001N2O1O001N2O1O001N2O1O1O0O2OdN\\1L4L3M4M3L4L4M2MVOj04L4M2M4L4M3L4L4L3N3L4L4M3M3M2N3M3M3M3M3M3M2\\IVEQ4m:oKVEk3m:TLVEg3m:YLVEa3m:^LVE8d0^OY::VE5o=JTB4l=KXB1h=0ZBNf=1^BLb=4`BIa=6cBG]=9eBE[=9jBCW=<lBBT==PC@P=?SC^On<`0WC]Oi<b0ZC\\Of<c0^CYOc<f0`CXO`<f0eCWO[<h0hCUOY<j0kCSOU<l0nCROR<l0oCTOR<k0oCUOQ<j0oCWOQ<h0PDWOQ<h0PDXOP<f0QD[Oo;d0RD[Oo;d0RD\\On;c0RD^On;`0TD@l;?TDAl;<XDDh;7]DIc;2aDN`;MeD3[;GjD:V;AoD>R;]OSEc0m:XOWEi0P?O10O10O01000O0100O01000O01000O010O10O10O100OK6D<Eib6"}, "label": "chef holding a silver bowl"}]}
{"id": 440002, "image": "COCO_train2014_000000440002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a purple bandanna on her head\"."}], "task": "refering", "answer_template": "The \"a woman wearing a purple bandanna on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 130, 131, 132, 149, 150, 151, 167, 168, 169, 185, 186, 187, 188, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 275, 276, 277, 278, 293, 294, 295, 296, 311, 312, 313, 314, 329, 330, 331, 332, 347, 348, 349, 350, 365, 366, 367], "bbox": [0.15398785425101214, 0.27156250000000004, 0.4854048582995951, 0.886546875], "iscrowd": 0, "area": 30746.094199999996, "rle": {"size": [640, 494], "counts": "ik_14fc06J6J6L4L4L4L3M41O00001O001O000O2O00001O001O0O101O0O2O00001N101O000O2O00000O01O1O0O2O1O1E:01O001O00010O00001O0YLTO[Dl0];_O`Da0\\;EbD;];I`D7_;L_D4_;0^D0b;3\\DMc;6ZDKe;8YDHg;:VDGk;:RDFP4fN_1e1[JIl3nNh1Z1VJLm3POm1U1PJOR4oNm1S1kI1W4oNn1Q1SIf0n4[No1o0PIh0o4]NP2l0mHj0P5]NS2j0hHn0R5[NU2h0eHP1S5[NX2f0`HT1U5XN[2e0YHZ1X5UN^2n3_MULa2l3\\MVLd2j3YMYLf2i3WMYLi2h3SM[Lm2e3PM^LP3c3dIH\\68\\I0d61SI7m6JiH?W7AaHg0_7m50O100000001O1O001O1O1O1O1N2O1iDPH\\:P8`EWH]:j7^E\\H`:e7ZEbHd:_7WEgHg:Z8O1O1O2N1N2O1O1O2N1O1O1N2O2N1N2O1N2O2N1N2O1N2O2M2O1O1N2O2M2O1N2O4K6J5L5J5K6K4K6J5K6K2M4L3N3L3M3eKbCP1a<gNmCP1V<gNYDP1k;fNdDa0`NgMn<_1QE<h;\\OeD7Q`0^Ol0TOiXn4"}, "label": "a woman wearing a purple bandanna on her head"}]}
{"id": 440002, "image": "COCO_train2014_000000440002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef wearing a purple bandana\"."}], "task": "refering", "answer_template": "The \"a chef wearing a purple bandana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 130, 131, 132, 149, 150, 151, 167, 168, 169, 185, 186, 187, 188, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 275, 276, 277, 278, 293, 294, 295, 296, 311, 312, 313, 314, 329, 330, 331, 332, 347, 348, 349, 350, 365, 366, 367], "bbox": [0.15398785425101214, 0.27156250000000004, 0.4854048582995951, 0.886546875], "iscrowd": 0, "area": 30746.094199999996, "rle": {"size": [640, 494], "counts": "ik_14fc06J6J6L4L4L4L3M41O00001O001O000O2O00001O001O0O101O0O2O00001N101O000O2O00000O01O1O0O2O1O1E:01O001O00010O00001O0YLTO[Dl0];_O`Da0\\;EbD;];I`D7_;L_D4_;0^D0b;3\\DMc;6ZDKe;8YDHg;:VDGk;:RDFP4fN_1e1[JIl3nNh1Z1VJLm3POm1U1PJOR4oNm1S1kI1W4oNn1Q1SIf0n4[No1o0PIh0o4]NP2l0mHj0P5]NS2j0hHn0R5[NU2h0eHP1S5[NX2f0`HT1U5XN[2e0YHZ1X5UN^2n3_MULa2l3\\MVLd2j3YMYLf2i3WMYLi2h3SM[Lm2e3PM^LP3c3dIH\\68\\I0d61SI7m6JiH?W7AaHg0_7m50O100000001O1O001O1O1O1O1N2O1iDPH\\:P8`EWH]:j7^E\\H`:e7ZEbHd:_7WEgHg:Z8O1O1O2N1N2O1O1O2N1O1O1N2O2N1N2O1N2O2N1N2O1N2O2M2O1O1N2O2M2O1N2O4K6J5L5J5K6K4K6J5K6K2M4L3N3L3M3eKbCP1a<gNmCP1V<gNYDP1k;fNdDa0`NgMn<_1QE<h;\\OeD7Q`0^Ol0TOiXn4"}, "label": "a chef wearing a purple bandana"}]}
{"id": 521923, "image": "COCO_train2014_000000521923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a green hat is walking a dog\"."}], "task": "refering", "answer_template": "The \"a woman with a green hat is walking a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 60, 61, 62, 77, 78, 79, 80, 95, 96, 97, 98, 114, 115, 116, 132, 133, 134, 150, 151], "bbox": [0.31207999999999997, 0.15010666666666667, 0.48153999999999997, 0.6870133333333334], "iscrowd": 0, "area": 9478.93225, "rle": {"size": [375, 500], "counts": "\\Xi12d;4L4L3M4L4L4L3M2N2N2M300O01000O1000O10O9He0[O9A;D<E:\\OhLYHd3a7?J6I7J6J2M01O010O1O001O1O01lNcKnJ\\4R5T11O1O1O100O1O001O1O1O100O1O001O1O100O1O001O1O3nKeId2^6SMhIl2\\6lLjIR3Y6fLnIX3V6_LPJ`3T6XLRJf3Q7O1N1O2YNTH@n7>_HUOb7j0jHjNX7U1TI_Nn6^1h1M3M3M3M3M3NYgn2"}, "label": "a woman with a green hat is walking a dog"}]}
{"id": 521923, "image": "COCO_train2014_000000521923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black snow suit walking with her dog\"."}], "task": "refering", "answer_template": "The \"a woman in a black snow suit walking with her dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 60, 61, 62, 77, 78, 79, 80, 95, 96, 97, 98, 114, 115, 116, 132, 133, 134, 150, 151], "bbox": [0.31207999999999997, 0.15010666666666667, 0.48153999999999997, 0.6870133333333334], "iscrowd": 0, "area": 9478.93225, "rle": {"size": [375, 500], "counts": "\\Xi12d;4L4L3M4L4L4L3M2N2N2M300O01000O1000O10O9He0[O9A;D<E:\\OhLYHd3a7?J6I7J6J2M01O010O1O001O1O01lNcKnJ\\4R5T11O1O1O100O1O001O1O1O100O1O001O1O100O1O001O1O3nKeId2^6SMhIl2\\6lLjIR3Y6fLnIX3V6_LPJ`3T6XLRJf3Q7O1N1O2YNTH@n7>_HUOb7j0jHjNX7U1TI_Nn6^1h1M3M3M3M3M3NYgn2"}, "label": "a woman in a black snow suit walking with her dog"}]}
{"id": 521923, "image": "COCO_train2014_000000521923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in all black\"."}], "task": "refering", "answer_template": "The \"a man in all black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 48, 49, 50, 66, 67, 68, 69, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 191, 192, 193, 194, 209, 210, 211, 212, 227, 228, 229, 230], "bbox": [0.61218, 0.10029333333333333, 0.88966, 1.0], "iscrowd": 0, "area": 25353.41949999999, "rle": {"size": [375, 500], "counts": "YV`3:X;7M2O1O1O1I5010O13O30OM1N101N1WJnNWKg0j4<\\O:c0HROb0l0@POd0l0@QOb0m0AiNaNjLQ2Y4BfNdNmLP2X4^OfN]1i0hNoLVNY1X3g1cNnLZNS1X3m1`NnLSO4f2k2YNoLb3P3_LmLd3Q3^LeLl3Y3Z2O100O1O100O1O1LcGnL]8P36OCoLVHP3j7TMSHk2m7YMPHf2P8]MnG[17TOk7EkGT1`0SOd7^1`H`NZ7e1jHkMZO4c7Y2bIaMV6i2V15VOQMkHR3o6VMlHl2m6S1J5N3L4M2M4N16K6kKQIT131R7cNSIo04=l6^NWIh08g0X9N1O2N100O1O010O102M01O1O1O2N1O1O2N1O2N7I7I3M1JkMhFV2V9mMhFS2W98O1I6N3PN^MhJd2j4jMfJ@QOh2R6PNiJa2n4kMkJX2n4UNiJo1o4^NgJg1P5d2L5L3L5XK\\IW4T7K6I6K6J5L4L5^OfGaM^8_2>0O00010O01O001N1O2O2M2O2M5K4M3L5L3L5L3L4M2M3N2M3N3L3K5K5Knoc0"}, "label": "a man in all black"}]}
{"id": 521923, "image": "COCO_train2014_000000521923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man dressed in black ski gear\"."}], "task": "refering", "answer_template": "The \"man dressed in black ski gear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 48, 49, 50, 66, 67, 68, 69, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 191, 192, 193, 194, 209, 210, 211, 212, 227, 228, 229, 230], "bbox": [0.61218, 0.10029333333333333, 0.88966, 1.0], "iscrowd": 0, "area": 25353.41949999999, "rle": {"size": [375, 500], "counts": "YV`3:X;7M2O1O1O1I5010O13O30OM1N101N1WJnNWKg0j4<\\O:c0HROb0l0@POd0l0@QOb0m0AiNaNjLQ2Y4BfNdNmLP2X4^OfN]1i0hNoLVNY1X3g1cNnLZNS1X3m1`NnLSO4f2k2YNoLb3P3_LmLd3Q3^LeLl3Y3Z2O100O1O100O1O1LcGnL]8P36OCoLVHP3j7TMSHk2m7YMPHf2P8]MnG[17TOk7EkGT1`0SOd7^1`H`NZ7e1jHkMZO4c7Y2bIaMV6i2V15VOQMkHR3o6VMlHl2m6S1J5N3L4M2M4N16K6kKQIT131R7cNSIo04=l6^NWIh08g0X9N1O2N100O1O010O102M01O1O1O2N1O1O2N1O2N7I7I3M1JkMhFV2V9mMhFS2W98O1I6N3PN^MhJd2j4jMfJ@QOh2R6PNiJa2n4kMkJX2n4UNiJo1o4^NgJg1P5d2L5L3L5XK\\IW4T7K6I6K6J5L4L5^OfGaM^8_2>0O00010O01O001N1O2O2M2O2M5K4M3L5L3L5L3L4M2M3N2M3N3L3K5K5Knoc0"}, "label": "man dressed in black ski gear"}]}
{"id": 227012, "image": "COCO_train2014_000000227012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man looks to his friends mobile\"."}], "task": "refering", "answer_template": "The \"a man looks to his friends mobile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 66, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5366875, 0.015889145496535795, 0.9943125, 0.9686836027713626], "iscrowd": 0, "area": 81787.13215000003, "rle": {"size": [433, 640], "counts": "T^a42o<`0N2O1N2O1N3M2O1O1N2O1O1O2N1N2O1O1O1O2M200O100O100O100O100O10000O100O100O10000]E[NZ9e1eFoNh8Q1XGAV8?iG5e7JZHj0S7VOlHX1g6hNXI[1f6eNYI]1f6dNXI^1g6bNXI`1g6`NYIb1e6^NZId1e6\\NZIf1e6ZNZIh1f6WNYIl1e6TN[Im1d6SN[Io1d6QN\\IP2c6PN]IR2a6nM^IT2a6lM_IU2`6kM`IW2^6iMbIX2]6hMbIZ2]6fMcI[2\\6eMdI]2Z6cMeI_2Z6aMfI`2Y6`MgIa2X6_MhIc2V6]MiIe2V6[MjIf2U6ZMkIg2S6ZMmIh2Q6XMoIi2o5XMQJi2m5XMSJj2k5VMUJk2i5VMWJj2i5VMWJj2h5WMXJi2g5XMYJh2g5XMYJh2f5YMZJg2e5ZM[Jg2d5YM\\Jg2c5ZM]Jf2c5ZM]Jf2b5[M^Je2a5\\M_Jd2a5\\M_Jd2`5]M`Jc2_5^MaJb2_5^MaJc2]5^McJb2]5^McJb2U5eMlJ[2i4PNWKP2^4[NbKe1S4fNmKZ1h3QOXLo0]3\\OcLd0R3GnL:f21ZMO[2<eMDQ2f0oMZOf1Q1ZNoN[1\\1eNdNP1g1POYNe0R2[OnM:]2FcMOh21XMDS3<Q500000000000O1000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O100000000000001O4SEVOk8n0lF[OP9h0hFAT9c0cFFY9>^FK^99YFOd95SF4i9l1NO1O1O1N2O1O1O001O1O1O1O1O1O1O1000000O10000000000000000001O0O2O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1N2O001O1O001O1O1O001O1O000000001O00000000000O1000000000001O0000000000:FT3lLT3lLU\\1"}, "label": "a man looks to his friends mobile"}]}
{"id": 227012, "image": "COCO_train2014_000000227012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing black coat\"."}], "task": "refering", "answer_template": "The \"man wearing black coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 66, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5366875, 0.015889145496535795, 0.9943125, 0.9686836027713626], "iscrowd": 0, "area": 81787.13215000003, "rle": {"size": [433, 640], "counts": "T^a42o<`0N2O1N2O1N3M2O1O1N2O1O1O2N1N2O1O1O1O2M200O100O100O100O100O10000O100O100O10000]E[NZ9e1eFoNh8Q1XGAV8?iG5e7JZHj0S7VOlHX1g6hNXI[1f6eNYI]1f6dNXI^1g6bNXI`1g6`NYIb1e6^NZId1e6\\NZIf1e6ZNZIh1f6WNYIl1e6TN[Im1d6SN[Io1d6QN\\IP2c6PN]IR2a6nM^IT2a6lM_IU2`6kM`IW2^6iMbIX2]6hMbIZ2]6fMcI[2\\6eMdI]2Z6cMeI_2Z6aMfI`2Y6`MgIa2X6_MhIc2V6]MiIe2V6[MjIf2U6ZMkIg2S6ZMmIh2Q6XMoIi2o5XMQJi2m5XMSJj2k5VMUJk2i5VMWJj2i5VMWJj2h5WMXJi2g5XMYJh2g5XMYJh2f5YMZJg2e5ZM[Jg2d5YM\\Jg2c5ZM]Jf2c5ZM]Jf2b5[M^Je2a5\\M_Jd2a5\\M_Jd2`5]M`Jc2_5^MaJb2_5^MaJc2]5^McJb2]5^McJb2U5eMlJ[2i4PNWKP2^4[NbKe1S4fNmKZ1h3QOXLo0]3\\OcLd0R3GnL:f21ZMO[2<eMDQ2f0oMZOf1Q1ZNoN[1\\1eNdNP1g1POYNe0R2[OnM:]2FcMOh21XMDS3<Q500000000000O1000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O100000000000001O4SEVOk8n0lF[OP9h0hFAT9c0cFFY9>^FK^99YFOd95SF4i9l1NO1O1O1N2O1O1O001O1O1O1O1O1O1O1000000O10000000000000000001O0O2O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1N2O001O1O001O1O1O001O1O000000001O00000000000O1000000000001O0000000000:FT3lLT3lLU\\1"}, "label": "man wearing black coat"}]}
{"id": 227012, "image": "COCO_train2014_000000227012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white man with a blue and black checkered hoodie holding a cellphone by another man\"."}], "task": "refering", "answer_template": "The \"a white man with a blue and black checkered hoodie holding a cellphone by another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.07145312499999999, 0.06066974595842956, 0.59903125, 0.9887528868360277], "iscrowd": 0, "area": 87723.13475, "rle": {"size": [433, 640], "counts": "Xic08Q=:K5K5K6J5K5K5K5K6J5K5K5K5K6J5K5M3L5L3M3L4M3L4M3L4M3M3L3N1N3N2N1N3N2M2O2M3N2N1N2O1N2O1N2O1O1N2O1N2O1M3N2M3M3N3L:F3N2M3N2M3M3N2M3M3N2M3]Oc0L4L4L4L4L4L4M3M3N2M3M3M3N2M3M3M3N2M3M3M3N2M3O1O1O1O1N2O1O001O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O001N2O1O100O100O100O100O100O1001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O002N2N1O2N1O2M3N1O2N1O2N1O2N2N1O2N1O2N1O2N6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J3M1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3M3L4M3L4M3L4M3L4M2M4M3M3L4M3L4M3N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N6K8G8H9H7H9G8H9H7F;E:E<EmZ\\3"}, "label": "a white man with a blue and black checkered hoodie holding a cellphone by another man"}]}
{"id": 227012, "image": "COCO_train2014_000000227012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a buy in a blue plaid jacket\"."}], "task": "refering", "answer_template": "The \"a buy in a blue plaid jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.07145312499999999, 0.06066974595842956, 0.59903125, 0.9887528868360277], "iscrowd": 0, "area": 87723.13475, "rle": {"size": [433, 640], "counts": "Xic08Q=:K5K5K6J5K5K5K5K6J5K5K5K5K6J5K5M3L5L3M3L4M3L4M3L4M3M3L3N1N3N2N1N3N2M2O2M3N2N1N2O1N2O1N2O1O1N2O1N2O1M3N2M3M3N3L:F3N2M3N2M3M3N2M3M3N2M3]Oc0L4L4L4L4L4L4M3M3N2M3M3M3N2M3M3M3N2M3M3M3N2M3O1O1O1O1N2O1O001O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O001N2O1O100O100O100O100O100O1001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O002N2N1O2N1O2M3N1O2N1O2N1O2N2N1O2N1O2N1O2N6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J3M1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3M3L4M3L4M3L4M3L4M2M4M3M3L4M3L4M3N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N6K8G8H9H7H9G8H9H7F;E:E<EmZ\\3"}, "label": "a buy in a blue plaid jacket"}]}
{"id": 227012, "image": "COCO_train2014_000000227012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white car parked behind the person in the blue jacket\"."}], "task": "refering", "answer_template": "The \"the white car parked behind the person in the blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 148, 149, 162, 163, 164, 165, 172, 173, 184, 185, 186, 187, 196, 197, 207, 208, 209, 220, 221, 222, 230, 231, 232, 243, 244, 245, 253, 254, 266, 267, 268, 276, 277, 289, 290, 291, 299, 300, 301, 322, 323, 324, 325], "bbox": [0.0, 0.4022401847575057, 0.6887187499999999, 1.0], "iscrowd": 0, "area": 21651.480649999994, "rle": {"size": [433, 640], "counts": "U8a1i;P4PL7K5O1O1O100O1O1O2N1O100O100O100O100O100O2O0O100O100O100O100O100O100O10000O10000O1000000O10iLXIEh69^ICa6=dI_O\\6?kI\\OU6b0QJZOn5f0WJVOi5h0]JTOc5k0cJPO]5n0hJoNW5P1nJmNR5R1QKlNo4R1VKkNi4T1\\KiNd4V1`KgN`4V1fKgNY4W1mKfNS4W1RLgNn3W1WLfNh3X1^LeNb3X1dLeN\\3X1jLeNV3X1nLgNQ3W1RMiNn2T1VMkNj2S1YMlNf2R1^MmNb2P1aMPO_2n0dMQO[2n0hMQOX2n0jMQOV2m0lMSOT2k0oMTOP2k0SNTOm1j0VNUOj1i0XNWOg1i0[NVOe1h0^NWOb1g0aNXO^1g0dNYO\\1e0gNZOY1e0iNZOW1d0lN[OS1d0oN\\OQ1b0QO^Oo0`0SO@l0`0UO@k0>WOBi0<YODf0;\\OEd0:]OFc09^OCe0<]OAf0>[O_Oh0`0YO\\Ok0c0Q601N2O001O1O0O2O1O1O0O2O1O001O1N101O1O1O001O1O001Og`Q20[_nM3M1N1000000O10000O10000O1000000O10O10O1000000O1000O01000O010O100O010O10O01000O0100O10O0100O01000O010O100O010O10O10O1000O10001N10001N1000001N10000O2O000O2O0001O0101O1N101N101N2O0O2O1O0O2O0O2hD2i8OoEW1Q:T12N2O0O2OO0010O0010O010O0010O0010O0010O010O00001N101N1O2N100O2N1O2O0O1O2N101N1O2O0O1O2N101N1O1O2O0O2N1O2\\Oc0O4L6J5K3M4L4L3H9DZkc2"}, "label": "the white car parked behind the person in the blue jacket"}]}
{"id": 227012, "image": "COCO_train2014_000000227012.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car is back of blue jacket man\"."}], "task": "refering", "answer_template": "The \"a white car is back of blue jacket man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 148, 149, 162, 163, 164, 165, 172, 173, 184, 185, 186, 187, 196, 197, 207, 208, 209, 220, 221, 222, 230, 231, 232, 243, 244, 245, 253, 254, 266, 267, 268, 276, 277, 289, 290, 291, 299, 300, 301, 322, 323, 324, 325], "bbox": [0.0, 0.4022401847575057, 0.6887187499999999, 1.0], "iscrowd": 0, "area": 21651.480649999994, "rle": {"size": [433, 640], "counts": "U8a1i;P4PL7K5O1O1O100O1O1O2N1O100O100O100O100O100O2O0O100O100O100O100O100O100O10000O10000O1000000O10iLXIEh69^ICa6=dI_O\\6?kI\\OU6b0QJZOn5f0WJVOi5h0]JTOc5k0cJPO]5n0hJoNW5P1nJmNR5R1QKlNo4R1VKkNi4T1\\KiNd4V1`KgN`4V1fKgNY4W1mKfNS4W1RLgNn3W1WLfNh3X1^LeNb3X1dLeN\\3X1jLeNV3X1nLgNQ3W1RMiNn2T1VMkNj2S1YMlNf2R1^MmNb2P1aMPO_2n0dMQO[2n0hMQOX2n0jMQOV2m0lMSOT2k0oMTOP2k0SNTOm1j0VNUOj1i0XNWOg1i0[NVOe1h0^NWOb1g0aNXO^1g0dNYO\\1e0gNZOY1e0iNZOW1d0lN[OS1d0oN\\OQ1b0QO^Oo0`0SO@l0`0UO@k0>WOBi0<YODf0;\\OEd0:]OFc09^OCe0<]OAf0>[O_Oh0`0YO\\Ok0c0Q601N2O001O1O0O2O1O1O0O2O1O001O1N101O1O1O001O1O001Og`Q20[_nM3M1N1000000O10000O10000O1000000O10O10O1000000O1000O01000O010O100O010O10O01000O0100O10O0100O01000O010O100O010O10O10O1000O10001N10001N1000001N10000O2O000O2O0001O0101O1N101N101N2O0O2O1O0O2O0O2hD2i8OoEW1Q:T12N2O0O2OO0010O0010O010O0010O0010O0010O010O00001N101N1O2N100O2N1O2O0O1O2N101N1O2O0O1O2N101N1O1O2O0O2N1O2\\Oc0O4L6J5K3M4L4L3H9DZkc2"}, "label": "a white car is back of blue jacket man"}]}
{"id": 120524, "image": "COCO_train2014_000000120524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep interacting with a dog\"."}], "task": "refering", "answer_template": "The \"a sheep interacting with a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 252, 269, 270, 271, 272, 275, 292, 293, 294, 295, 315, 317, 318], "bbox": [0.523546875, 0.20158333333333334, 0.9909062500000001, 0.8311458333333334], "iscrowd": 0, "area": 46173.1567, "rle": {"size": [480, 640], "counts": "_Um42m>3M4L2N100O1O1O1O2N1O100O1O1O100O1O100O1O1O2O1N2N3N1N2O3L4M3L5L3M3L4M:Eb0_O9F2O1O2M2O1N3N1N2O2N1N2O1N2O1O1N2O1N10000O1000O0100O0100000O0100000O0100000O0100000O01N2N2O0O2N2N23L3N3M2N3M4K4M3M3M4K2O1O1O1N2O1O1O1N2O1O2N1O1O100O1O1O2O0O1O1ZIcJg3]5WLfJh3[5ULgJl3Y5QLiJo3X5nKkJQ4V5lKlJT4T5jKoJV4Q5gKQKY4P5bKUK]4m4]KWKc4R5PKSKP5U5cJoJ]5X5XJlJh5Z5lIlJT6Z5_IlJb6i61O0O101O1O0O2O001N101O1O0O2O001N101N1jMnIYKX6a4W2I6L5N1O101N1O100O2N100O10001O01O001O010O001O0010OcHQLi3P4mJYMS5g2jJ\\MV5e2eJ`M[5`2`JdM`5]2[JgMe5[2SJkMm5V2mIPNR6R2gISNZ6Y5000001O01O0001O01O0001O00O2_N`1I7I7M3M3M3M3M3M3M3M3N2O1O2M2O1N2O1N2O1N2O1N2O1O1N2O1N2O2N1O1O1O100000000000001O00000000000O100000001O000000000000000000001O000000000000000000001O000000000000]FoKW8Q4[G[Lg8l43N2M3M4L3M3M2N3M2N3F901N101Nc0^OS1lNT1mNS1mNZb2"}, "label": "a sheep interacting with a dog"}]}
{"id": 120524, "image": "COCO_train2014_000000120524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep being sniffed by a dog\"."}], "task": "refering", "answer_template": "The \"a sheep being sniffed by a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 252, 269, 270, 271, 272, 275, 292, 293, 294, 295, 315, 317, 318], "bbox": [0.523546875, 0.20158333333333334, 0.9909062500000001, 0.8311458333333334], "iscrowd": 0, "area": 46173.1567, "rle": {"size": [480, 640], "counts": "_Um42m>3M4L2N100O1O1O1O2N1O100O1O1O100O1O100O1O1O2O1N2N3N1N2O3L4M3L5L3M3L4M:Eb0_O9F2O1O2M2O1N3N1N2O2N1N2O1N2O1O1N2O1N10000O1000O0100O0100000O0100000O0100000O0100000O01N2N2O0O2N2N23L3N3M2N3M4K4M3M3M4K2O1O1O1N2O1O1O1N2O1O2N1O1O100O1O1O2O0O1O1ZIcJg3]5WLfJh3[5ULgJl3Y5QLiJo3X5nKkJQ4V5lKlJT4T5jKoJV4Q5gKQKY4P5bKUK]4m4]KWKc4R5PKSKP5U5cJoJ]5X5XJlJh5Z5lIlJT6Z5_IlJb6i61O0O101O1O0O2O001N101O1O0O2O001N101N1jMnIYKX6a4W2I6L5N1O101N1O100O2N100O10001O01O001O010O001O0010OcHQLi3P4mJYMS5g2jJ\\MV5e2eJ`M[5`2`JdM`5]2[JgMe5[2SJkMm5V2mIPNR6R2gISNZ6Y5000001O01O0001O01O0001O00O2_N`1I7I7M3M3M3M3M3M3M3M3N2O1O2M2O1N2O1N2O1N2O1N2O1O1N2O1N2O2N1O1O1O100000000000001O00000000000O100000001O000000000000000000001O000000000000000000001O000000000000]FoKW8Q4[G[Lg8l43N2M3M4L3M3M2N3M2N3F901N101Nc0^OS1lNT1mNS1mNZb2"}, "label": "a sheep being sniffed by a dog"}]}
{"id": 63182, "image": "COCO_train2014_000000063182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small , white lamb eating grass\"."}], "task": "refering", "answer_template": "The \"a small , white lamb eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 302, 303], "bbox": [0.0726875, 0.31723004694835677, 0.389109375, 0.9424882629107981], "iscrowd": 0, "area": 37420.09235, "rle": {"size": [426, 640], "counts": "khc0P1U2YO`7Q1ZHYOY7P1aHWOT7R1gHROT7U1fHnNV7Y1dHjNW7]1]HnN^7Y1lGZOo7S3L4M4M2M3N3M2M3O1O1O010ON2M4M2M3_Ob0M2N2N3NlH\\Ke5b4[JbKc5\\4]JgKc5V4\\JoKb5o3^JTLa5j3_JYLa5e3^J_L`5b3^J`La5b3\\JaLc5_3ZJdLe5]3YJfLe5\\3XJfLh5Z3UJiLj5X3TJkLj5V3TJlLl5U3QJmLn5T3oIPMo5Q3nIRMR6o2iIWMT6j2iI[MU6e2hI`MU6m4N2M3N2N6I7J6I4M0O101N101OO010O10O10O100O10O010000O100O010O10000O1M2N3L4M3M3N2O100O100O100O100O100O100O100O100O1O1O1O100O1O1O1O2N1O2O0N3M2N3O0O2N101N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2M2O2N1N3N101O001O0010O01003M2M4M2N2N3L2OM3N2M2O2N2K5K5J6K4L5J6K5K5J6L4M3L3N3M3N3M2N3M2M4M2N3M2N2N3M2N3M2M4L5K4M3L5K4L4M4J9Eg0XOm]R5"}, "label": "a small , white lamb eating grass"}]}
{"id": 63182, "image": "COCO_train2014_000000063182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb eating some grass to the left of another lamb\"."}], "task": "refering", "answer_template": "The \"a lamb eating some grass to the left of another lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 302, 303], "bbox": [0.0726875, 0.31723004694835677, 0.389109375, 0.9424882629107981], "iscrowd": 0, "area": 37420.09235, "rle": {"size": [426, 640], "counts": "khc0P1U2YO`7Q1ZHYOY7P1aHWOT7R1gHROT7U1fHnNV7Y1dHjNW7]1]HnN^7Y1lGZOo7S3L4M4M2M3N3M2M3O1O1O010ON2M4M2M3_Ob0M2N2N3NlH\\Ke5b4[JbKc5\\4]JgKc5V4\\JoKb5o3^JTLa5j3_JYLa5e3^J_L`5b3^J`La5b3\\JaLc5_3ZJdLe5]3YJfLe5\\3XJfLh5Z3UJiLj5X3TJkLj5V3TJlLl5U3QJmLn5T3oIPMo5Q3nIRMR6o2iIWMT6j2iI[MU6e2hI`MU6m4N2M3N2N6I7J6I4M0O101N101OO010O10O10O100O10O010000O100O010O10000O1M2N3L4M3M3N2O100O100O100O100O100O100O100O100O1O1O1O100O1O1O1O2N1O2O0N3M2N3O0O2N101N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1O2M2O2N1N3N101O001O0010O01003M2M4M2N2N3L2OM3N2M2O2N2K5K5J6K4L5J6K5K5J6L4M3L3N3M3N3M2N3M2M4M2N3M2N2N3M2N3M2M4L5K4M3L5K4L4M4J9Eg0XOm]R5"}, "label": "a lamb eating some grass to the left of another lamb"}]}
{"id": 63182, "image": "COCO_train2014_000000063182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep that is looking into the camera\"."}], "task": "refering", "answer_template": "The \"the sheep that is looking into the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 243, 244, 245, 246, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 312, 313, 314, 335, 336], "bbox": [0.45421875, 0.16009389671361504, 0.7741875, 0.9733802816901408], "iscrowd": 0, "area": 41632.052350000005, "rle": {"size": [426, 640], "counts": "mVi3<l<c0^Ob0^Oa0_O4K5L3M4L4K5L4L4L3L5L4L4L4K4M5K6J6I7J6J6J8Ge0\\Od0\\O>B1O1O1O2N1OlLPJPOP6Q1UJiNj5Y1ZJbNe5`1`JZN_5g1gJSNY5n1lJlMS5V2RKeMl4]2YK]Mf4e2_KUMa4l2dKnL[4S3kKgLT4[3oKaLP4a3RL\\Ln3e3ULXLi3j3YLSLf3o3]LmKb3T4aLiK_3X4dLdK[3^4gL_KX3c4kLYKU3h4nLTKR3m4PMQKn2P5f21N2O1O1N2O001N2O1O1N2O0O2O1O1N2O1O0O2O1N2M6J5K5K5J7J5K5K6K4M3L5YL`H\\1c7XNVIX1m6\\N]Ic1f6QNdIm1a6eMiI[2Z6YMPJf2[8N101O1O0O2O1OO010ZLhMYLY2]3TN`Ll1`3XN\\Lg1e3]NVLd1j3_NSLa1m3cNoK]1Q4gNjKY1W4kNVKd1j4_NkJi1U5[N_Jl1b5XNRJP2n5SNgIU2Y6oM\\IX2d6V210000O100001O00000000mKaI[1_6]NPJ\\1P6\\NYJd1f5TNcJk1]5lMmJS2S5fMUKY2k4bM[K]2e4^M`Kb2`4ZMeKf2[4TMkKk2U4QMPLn2P4mLTLT3l3hLXLX3l61O1YObLaG`3[8gL`GZ3]8mL_GS3^8TM]Gm2`8ZM[Gg2b8R1M4L3L4M3M3M3M3M4L3M3M3M3M3M4L4L4K5L4L4L4M3N3M2N2N2O1N2N3M2N2N2N2N3M4L3M4L3L5K4L5K4M4K4L5K4M4K4L5K4M4K4L5K4M3L5K4LbUl1"}, "label": "the sheep that is looking into the camera"}]}
{"id": 63182, "image": "COCO_train2014_000000063182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheep with a black head and legs staring at the camera with one eye\"."}], "task": "refering", "answer_template": "The \"a white sheep with a black head and legs staring at the camera with one eye\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 243, 244, 245, 246, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 312, 313, 314, 335, 336], "bbox": [0.45421875, 0.16009389671361504, 0.7741875, 0.9733802816901408], "iscrowd": 0, "area": 41632.052350000005, "rle": {"size": [426, 640], "counts": "mVi3<l<c0^Ob0^Oa0_O4K5L3M4L4K5L4L4L3L5L4L4L4K4M5K6J6I7J6J6J8Ge0\\Od0\\O>B1O1O1O2N1OlLPJPOP6Q1UJiNj5Y1ZJbNe5`1`JZN_5g1gJSNY5n1lJlMS5V2RKeMl4]2YK]Mf4e2_KUMa4l2dKnL[4S3kKgLT4[3oKaLP4a3RL\\Ln3e3ULXLi3j3YLSLf3o3]LmKb3T4aLiK_3X4dLdK[3^4gL_KX3c4kLYKU3h4nLTKR3m4PMQKn2P5f21N2O1O1N2O001N2O1O1N2O0O2O1O1N2O1O0O2O1N2M6J5K5K5J7J5K5K6K4M3L5YL`H\\1c7XNVIX1m6\\N]Ic1f6QNdIm1a6eMiI[2Z6YMPJf2[8N101O1O0O2O1OO010ZLhMYLY2]3TN`Ll1`3XN\\Lg1e3]NVLd1j3_NSLa1m3cNoK]1Q4gNjKY1W4kNVKd1j4_NkJi1U5[N_Jl1b5XNRJP2n5SNgIU2Y6oM\\IX2d6V210000O100001O00000000mKaI[1_6]NPJ\\1P6\\NYJd1f5TNcJk1]5lMmJS2S5fMUKY2k4bM[K]2e4^M`Kb2`4ZMeKf2[4TMkKk2U4QMPLn2P4mLTLT3l3hLXLX3l61O1YObLaG`3[8gL`GZ3]8mL_GS3^8TM]Gm2`8ZM[Gg2b8R1M4L3L4M3M3M3M3M4L3M3M3M3M3M4L4L4K5L4L4L4M3N3M2N2N2O1N2N3M2N2N2N2N3M4L3M4L3L5K4L5K4M4K4L5K4M4K4L5K4M4K4L5K4M3L5K4LbUl1"}, "label": "a white sheep with a black head and legs staring at the camera with one eye"}]}
{"id": 292558, "image": "COCO_train2014_000000292558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor for the apple brand computer . it is to the left of the other computer monitor\"."}], "task": "refering", "answer_template": "The \"the monitor for the apple brand computer . it is to the left of the other computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215], "bbox": [0.0831875, 0.12791569086651053, 0.542421875, 0.6577985948477751], "iscrowd": 0, "area": 53935.556249999994, "rle": {"size": [427, 640], "counts": "aUf02W=7I7K5K4L5K5K5K5K5K5K4L5K5K5J6K5K5K4L5K5K5K5K5K5K5K4L5K5K5K5K5K5K4K6K5K5K5K5K5K4L5K5K1O00O100000000000O01000000000000O10000000000O10000000O1000O1000000000000O1000000000000O1000O100000O1000000000000O1000000000000O01000000000O1000000000000O100000000000O010000000000O1000000000000O10000000O1000O1000000000000O10000000000O100000O100000O1000000000000O10000000000O10O1000000000O1000000000000O1000000000O10O10000000000O1000000000000O10000000O1000O10000000000O1000000000000O1000O10000000O10000000000O1000000000000O10O1000000000O1000000000000O1000000000O01000000000000O1G9B>B>B>C=B>B>B>C>A?A`0@?BXTj3"}, "label": "the monitor for the apple brand computer . it is to the left of the other computer monitor"}]}
{"id": 292558, "image": "COCO_train2014_000000292558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple computer\"."}], "task": "refering", "answer_template": "The \"an apple computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215], "bbox": [0.0831875, 0.12791569086651053, 0.542421875, 0.6577985948477751], "iscrowd": 0, "area": 53935.556249999994, "rle": {"size": [427, 640], "counts": "aUf02W=7I7K5K4L5K5K5K5K5K5K4L5K5K5J6K5K5K4L5K5K5K5K5K5K5K4L5K5K5K5K5K5K4K6K5K5K5K5K5K4L5K5K1O00O100000000000O01000000000000O10000000000O10000000O1000O1000000000000O1000000000000O1000O100000O1000000000000O1000000000000O01000000000O1000000000000O100000000000O010000000000O1000000000000O10000000O1000O1000000000000O10000000000O100000O100000O1000000000000O10000000000O10O1000000000O1000000000000O1000000000O10O10000000000O1000000000000O10000000O1000O10000000000O1000000000000O1000O10000000O10000000000O1000000000000O10O1000000000O1000000000000O1000000000O01000000000000O1G9B>B>B>C=B>B>B>C>A?A`0@?BXTj3"}, "label": "an apple computer"}]}
{"id": 292558, "image": "COCO_train2014_000000292558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small black rim monitor\"."}], "task": "refering", "answer_template": "The \"small black rim monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 225], "bbox": [0.5445156250000001, 0.14035128805620609, 0.939453125, 0.6134426229508196], "iscrowd": 0, "area": 36211.78615, "rle": {"size": [427, 640], "counts": "QZa41i<a0@`0_Oa0@`0_Oa0@`0bNmLTHc3l7]LcGT4]8>O000000000001O000000000000WH_Kh6a4WI`Ki6`4WIaKh6_4XIbKg6^4XIcKh6]4XIdKg6\\4YIeKf6[4YIfKg6Z4YIgKf6Y4YIiKf6W4ZIiKf6W4ZIjKe6V4ZIlKe6T4[ImKe6R4[InKe6R4ZIPLe6P4[IQLd6o3\\IQLd6o3[ISLd6m3\\ITLc6l3]ITLc6l3\\IVLc6j3]IVLc6j3]IULd6k3[IVLe6j3[IVLe6j3ZIVLg6j3YIVLg6j3YIVLg6j3XIVLi6j3WIVLi6j3WIVLi6j3VIVLk6j3UIVLl6i3TIWLl6i3SIWLn6i3RIWLn6i3RIVLo6j3PIWLP7i3PIWLP7i3PIVLQ7j3nHWLR7i3nHWLR7i3mHWLT7i3lHWLT7i3lHWLT7i3kHWLV7i3jHWLV7i3jHVLW7j3hHWLX7i3hHWLX7i3hHVLY7j3fHWLZ7i3fHWL[7h3dHXL]7h3cHXL]7h3cHXL]7h3bHXL_7h3aHXL_7h3`HYL`7g3`HXLa7h3^HYLb7g3^HXLc7h400000001O0000000000010O00000000001O0000O100O10000O1UO_HlKa7T4aHjK_7V4aHkK^7U4cHjK^7U4bHlK]7T4cHlK]7T4dHlK[7T4eHlK[7T4eHlK[7T4fHlKY7T4gHlKY7T4gHmKX7S4iHlKW7T4iHmKV7S4jHmKV7S4kHlKU7T4kHmKT7S4lHmKT7S4mHmKR7S4nHmKR7S4nHnKQ7R4PImKP7S4PInKo6R4QInKP7Q4QInKo6R4QIoKn6Q4RIoKn6Q4SIoKl6Q4TIoKl6Q4TIPLk6P4VIoKj6R4UInKk6R4UIoKj6Q4VIoKj6Q4WIoKh6Q4XIoKh6Q4XIPLg6P4ZIoKf6Q4ZIPLe6P4[IPLe6P4\\IoKd6Q4\\IPLc6P4]IPLc6P4^IPLb6o3^IQLb6o3^IRLa6n3`IQL`6o3`IQL`6o3`IRL_6n3bIQL^6o3bIRL]6n3cIRL]6n3dIRL[6n3eIQL\\6o3dIPL]6P4dInK]6R4cIlK_6T4aIkK`6U4aIhKa6X4_IgKb6Y4^IeKd6[4]IcKd6]4\\IaKf6_4ZI`Kh6_4R100000000000000000000000000000000000001O000000000000000000000000000000000000001M2M3L4M3L4L4M3L4M3L4L4M3L4M3L4M3L4L4M4K4M3L4L4M3L4M3L4L4M3L4M3L4M3L4L4M3L5LWV`0"}, "label": "small black rim monitor"}]}
{"id": 292558, "image": "COCO_train2014_000000292558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor on the right\"."}], "task": "refering", "answer_template": "The \"the monitor on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 225], "bbox": [0.5445156250000001, 0.14035128805620609, 0.939453125, 0.6134426229508196], "iscrowd": 0, "area": 36211.78615, "rle": {"size": [427, 640], "counts": "QZa41i<a0@`0_Oa0@`0_Oa0@`0bNmLTHc3l7]LcGT4]8>O000000000001O000000000000WH_Kh6a4WI`Ki6`4WIaKh6_4XIbKg6^4XIcKh6]4XIdKg6\\4YIeKf6[4YIfKg6Z4YIgKf6Y4YIiKf6W4ZIiKf6W4ZIjKe6V4ZIlKe6T4[ImKe6R4[InKe6R4ZIPLe6P4[IQLd6o3\\IQLd6o3[ISLd6m3\\ITLc6l3]ITLc6l3\\IVLc6j3]IVLc6j3]IULd6k3[IVLe6j3[IVLe6j3ZIVLg6j3YIVLg6j3YIVLg6j3XIVLi6j3WIVLi6j3WIVLi6j3VIVLk6j3UIVLl6i3TIWLl6i3SIWLn6i3RIWLn6i3RIVLo6j3PIWLP7i3PIWLP7i3PIVLQ7j3nHWLR7i3nHWLR7i3mHWLT7i3lHWLT7i3lHWLT7i3kHWLV7i3jHWLV7i3jHVLW7j3hHWLX7i3hHWLX7i3hHVLY7j3fHWLZ7i3fHWL[7h3dHXL]7h3cHXL]7h3cHXL]7h3bHXL_7h3aHXL_7h3`HYL`7g3`HXLa7h3^HYLb7g3^HXLc7h400000001O0000000000010O00000000001O0000O100O10000O1UO_HlKa7T4aHjK_7V4aHkK^7U4cHjK^7U4bHlK]7T4cHlK]7T4dHlK[7T4eHlK[7T4eHlK[7T4fHlKY7T4gHlKY7T4gHmKX7S4iHlKW7T4iHmKV7S4jHmKV7S4kHlKU7T4kHmKT7S4lHmKT7S4mHmKR7S4nHmKR7S4nHnKQ7R4PImKP7S4PInKo6R4QInKP7Q4QInKo6R4QIoKn6Q4RIoKn6Q4SIoKl6Q4TIoKl6Q4TIPLk6P4VIoKj6R4UInKk6R4UIoKj6Q4VIoKj6Q4WIoKh6Q4XIoKh6Q4XIPLg6P4ZIoKf6Q4ZIPLe6P4[IPLe6P4\\IoKd6Q4\\IPLc6P4]IPLc6P4^IPLb6o3^IQLb6o3^IRLa6n3`IQL`6o3`IQL`6o3`IRL_6n3bIQL^6o3bIRL]6n3cIRL]6n3dIRL[6n3eIQL\\6o3dIPL]6P4dInK]6R4cIlK_6T4aIkK`6U4aIhKa6X4_IgKb6Y4^IeKd6[4]IcKd6]4\\IaKf6_4ZI`Kh6_4R100000000000000000000000000000000000001O000000000000000000000000000000000000001M2M3L4M3L4L4M3L4M3L4L4M3L4M3L4M3L4L4M4K4M3L4L4M3L4M3L4L4M3L4M3L4M3L4L4M3L5LWV`0"}, "label": "the monitor on the right"}]}
{"id": 407246, "image": "COCO_train2014_000000407246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red skiboard\"."}], "task": "refering", "answer_template": "The \"the red skiboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126], "bbox": [0.1528125, 0.22471875, 0.7790208333333333, 0.32809374999999996], "iscrowd": 0, "area": 10688.72615, "rle": {"size": [640, 480], "counts": "li]1=ac03N2M3M2N3M3M3O001O1N2O000O1000O01000000O100O1O1O1O0011O00000000O100000O10000000000000000O10000000000000O1000O10000000000000000O1000000000000000O10O10000000000000000O100000000000000000O010000000000000000O1000000000000000000O0100000000000000000O10000000000000000O0100000000000000000O10000000000000000O10O1000000000000000O10000000000000000O10O10000000000O1000O100000O1000000000O10O10000000000O1000O10000000O1000000000O01000000000000O10O10000000O1000000000O10O10000000000O10O1000000000O10000000O10O1000000000000O01000000000O10000000O1000O100000000001N1O2O0O2O0O101N1O2O0O2O0O2O0O2N100O2O0O2N101N102M5L4KlRR2"}, "label": "the red skiboard"}]}
{"id": 407246, "image": "COCO_train2014_000000407246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red snowboard on the feet of the man doing a headstand\"."}], "task": "refering", "answer_template": "The \"a red snowboard on the feet of the man doing a headstand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126], "bbox": [0.1528125, 0.22471875, 0.7790208333333333, 0.32809374999999996], "iscrowd": 0, "area": 10688.72615, "rle": {"size": [640, 480], "counts": "li]1=ac03N2M3M2N3M3M3O001O1N2O000O1000O01000000O100O1O1O1O0011O00000000O100000O10000000000000000O10000000000000O1000O10000000000000000O1000000000000000O10O10000000000000000O100000000000000000O010000000000000000O1000000000000000000O0100000000000000000O10000000000000000O0100000000000000000O10000000000000000O10O1000000000000000O10000000000000000O10O10000000000O1000O100000O1000000000O10O10000000000O1000O10000000O1000000000O01000000000000O10O10000000O1000000000O10O10000000000O10O1000000000O10000000O10O1000000000000O01000000000O10000000O1000O100000000001N1O2O0O2O0O101N1O2O0O2O0O2O0O2N100O2O0O2N101N102M5L4KlRR2"}, "label": "a red snowboard on the feet of the man doing a headstand"}]}
{"id": 22222, "image": "COCO_train2014_000000022222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white station wagon to the left of the lady\"."}], "task": "refering", "answer_template": "The \"the white station wagon to the left of the lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211], "bbox": [0.001265625, 0.2941111111111111, 0.33342187500000003, 0.7669722222222223], "iscrowd": 0, "area": 28364.83320000001, "rle": {"size": [360, 640], "counts": "d>j4^6000000000000000000000000000000000000000000000000000000000000001O2N1O1O1O2N1O1O1N2O0000001O000000001O000000000000000000000000000000000000O100000000000000O100000000000000O1O010O1O100O100O1O100O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1000000000000000001N100000001O000000001O0O1000001O0000001O00000O101O000000001O0000000O2O000000001O0000001O0O1000001O0000001gNnHZNS7a1QI_NP7\\1UIbNl6Z1YIeNh6V1\\IjNe6Q1`InNa6m0cISO^6h0gIWOY6e0lIZOU6a0oI_OR6=RJBo59VJFk55YJKg51^JMd5O_J1b5LaJ3`5KbJ4_5IcJ7]5GfJ8[5EhJ:Y5CiJ=X5AjJ>W5_OkJa0U5]OnJb0m7N101O1N1O2N101N2N101N2N101N^le4"}, "label": "the white station wagon to the left of the lady"}]}
{"id": 30418, "image": "COCO_train2014_000000030418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white train with a black top\"."}], "task": "refering", "answer_template": "The \"a white train with a black top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0045, 0.29213114754098357, 0.99703125, 0.9887587822014051], "iscrowd": 0, "area": 114927.92295, "rle": {"size": [427, 640], "counts": "S_1o0P;]1cN\\1C=010O00010O0001O01O00010O00010J5A?@a0@?I701O01O0010O00010O00010O01O01O01O0010O000104K;E:G9F;E:G:E:F;F9F;E:F:G0O1O2O0O1O2O0O1O2O0O1O2N100O2N100O2N100O2N100O2N100O2N1O101N1O101N1O101N1O101N1O1O2O0O1O2O0O1O2O0O1O2O0O1O2O0O1O2N100O1O01O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O000O2O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0O101O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001N1000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0O10001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001OO100000000000000O1000000000000O100000000000000O10000000000001O001O001O001O00001O0O2O001O001O001O00001O0_Ob0UNRc0"}, "label": "a white train with a black top"}]}
{"id": 30418, "image": "COCO_train2014_000000030418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a group of trains\"."}], "task": "refering", "answer_template": "The \"a group of trains\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0045, 0.29213114754098357, 0.99703125, 0.9887587822014051], "iscrowd": 0, "area": 114927.92295, "rle": {"size": [427, 640], "counts": "S_1o0P;]1cN\\1C=010O00010O0001O01O00010O00010J5A?@a0@?I701O01O0010O00010O00010O01O01O01O0010O000104K;E:G9F;E:G:E:F;F9F;E:F:G0O1O2O0O1O2O0O1O2O0O1O2N100O2N100O2N100O2N100O2N100O2N1O101N1O101N1O101N1O101N1O1O2O0O1O2O0O1O2O0O1O2O0O1O2O0O1O2N100O1O01O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O000O2O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0O101O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001N1000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0O10001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001OO100000000000000O1000000000000O100000000000000O10000000000001O001O001O001O00001O0O2O001O001O001O00001O0_Ob0UNRc0"}, "label": "a group of trains"}]}
{"id": 30418, "image": "COCO_train2014_000000030418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pale yellowish train next to white train\"."}], "task": "refering", "answer_template": "The \"pale yellowish train next to white train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 156, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206], "bbox": [0.202984375, 0.22840749414519906, 1.0, 0.5930679156908666], "iscrowd": 0, "area": 23495.45480000001, "rle": {"size": [427, 640], "counts": "RZf12X=5K2N1O100O1O001O100O1O0100000000O101O0000000O100000001O0O10000000000O101O0000000O10000000000O101O0000000O10000000000O2O000000000001O0000000001O000000000000000001O01O000000000000_Cf0Q;ZOoDe0R;\\OlDe0T;[OlDd0U;\\OkDd0U;]OjDc0W;\\OhDd0Y;\\OgDd0Y;\\OgDd0Y;]OfDb0[;^OeDb0[;^OdDc0\\;]OdDb0];_ObDa0^;_ObD`0_;@aD`0_;A_D`0b;_O]Da0d;_OYDd0g;\\OUDh0k;YOoCk0R<70O100O100O101N100O100O100O100O100O101N100O100O100O100O100O2N1O100O1O1O1O100O1O1O2N100O1O1O100O1O1O\\CMR<2nCOR<1mC1S<MmC4R<LnC5R<JmC8S<HlC:a<1CEiC<X<DfC=Z<CeC?[<AbCa0^<_OaCb0_<_O_Cb0a<^O^Cd0a<3000001O01O0001O0000001O000010O000001O0000006J9G9H8G4L3M1O000000001O0001O0001O000000001O0001OO101N100001O10O01O00000001O01O000001O0001O0001O00000O101O0001O000010O00000001O000001O00000000000010O0N2L4K5L4L4K5N2O1O1O2O0O1O1O100O1O1O11O00100O001O001O001O001O001O010O001M3N3N1OL4L4L5\\Oe0N01O0000001O01O0001O00001O00100O1O001O1O1O00L4F;0O1O1O1O1O1O2N100O1O1O1O1O1N2O1O1O1O1O1O1O100O1O0010O01O001O1O0010O01O001O001O10O01O1O1O001O100O1O1N101O1O1N2O001O1O1O100O00100O1O1O00103L6J6J6J6K5J0000001O01O000001O01O0001O000001O01O0000001O00000O2O000000001O01O0010O0000001O0000000010O000001O0000001O01O000001O0000001O0001O0001O0000001O0000001O0000000O20O000010O0001O01O01O001O01O01O00O2O0O1\\K"}, "label": "pale yellowish train next to white train"}]}
{"id": 22223, "image": "COCO_train2014_000000022223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow nearest to the camera\"."}], "task": "refering", "answer_template": "The \"the cow nearest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 219, 220, 221, 222, 223, 236, 237, 238, 241, 242, 244, 245, 259, 260, 261], "bbox": [0.22690625, 0.20608899297423888, 0.764859375, 0.7803981264637003], "iscrowd": 0, "area": 46114.09185, "rle": {"size": [427, 640], "counts": "Ydl1;l<4M3M3M3N2M3N1O2N2N@`00O100000000O01000000000O10000hH^Oi1b0TNDi1=SNIj19QNLn14nM2P2OkM6U2IgM<X2EcMa0[2C]Mb0b2FjLh0U3]OTLU1k3oNlKX1R4mNiKT1V4POfKQ1Z4RObKP1\\4UO_Kl0`4XO\\Kj0b4[OYKf0g4]OUKe0i4_OSKb0l4CoJ?o4FlJoNClNa5Z2fJjNLjN]5a2bJeN4hNY5i2]J^N>dNX5S3TJYNh0^NW5]3lIUN_7Q2[HnMg7j31O1N2O1O1N23L6K5K5K5K4L5K3L5L3M3M3M3M3M3M3L4M00nNRJhJm5o4_JnJa5i4jJVKU5a4WK\\Ki4]4`KbK^4[4iKbKW4[4nKdKQ4X4VLeKj3W4\\LgKd3U4cLiK\\3W4fLgKZ3Y4hLfKW3[4iLdKW3\\4kLcKT3^4mL`KT3`4lL`KS3`4oL^KQ3[4WMeKh2[4ZMcKf2^4[MaKd2_4T31O1O001O00N3M2N2O1N2N2001O1O1O1O1O010OBXGZLg8g3`GRL_8o3>001O001O1O001O10O01O001O000O2O001O001O00001O001N101O00001O001O1O001N101O001O1O00O1000000O10000O10000000000000000000000000001N1000oH`KW5`4gJdKW5\\4gJgKX5Y4fJjKY5U4fJnKY5R4eJRLY5n3eJULZ5k3\\J`Lc5`3RJkLn5U3hIUMX6j2cI\\M]6d2^IaMb6d4000C^IQJb6n5dImI\\6S6>O1O1O1O100O2N1O100O1O2N100N2N2N3M2N2\\Od0O2N1O1O1O1O2N1O11O1O010O1O1O0010O01O0001O01O01O01O01O01O8fHXKb5o4SJUKl5Z5aIkJ_6Q6O2N100O0001O000O2O001O00001O00001N101O2N2N1N3L3N3M2M4M2N3M2M4M2N3M2M4M2N3N1N3M3M2O2M2N3N3L4L4L4M3L4L4M3J6F:F;F9F:oNiDD^;5gDJ\\;OiD0b;AcD>W<N1N2O1O00O1O101N1O2O0O1O20O001O001O0O101O001N1O2N1N2O2N1O2N1N2Obgn1"}, "label": "the cow nearest to the camera"}]}
{"id": 22223, "image": "COCO_train2014_000000022223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow closest to the camera\"."}], "task": "refering", "answer_template": "The \"cow closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 219, 220, 221, 222, 223, 236, 237, 238, 241, 242, 244, 245, 259, 260, 261], "bbox": [0.22690625, 0.20608899297423888, 0.764859375, 0.7803981264637003], "iscrowd": 0, "area": 46114.09185, "rle": {"size": [427, 640], "counts": "Ydl1;l<4M3M3M3N2M3N1O2N2N@`00O100000000O01000000000O10000hH^Oi1b0TNDi1=SNIj19QNLn14nM2P2OkM6U2IgM<X2EcMa0[2C]Mb0b2FjLh0U3]OTLU1k3oNlKX1R4mNiKT1V4POfKQ1Z4RObKP1\\4UO_Kl0`4XO\\Kj0b4[OYKf0g4]OUKe0i4_OSKb0l4CoJ?o4FlJoNClNa5Z2fJjNLjN]5a2bJeN4hNY5i2]J^N>dNX5S3TJYNh0^NW5]3lIUN_7Q2[HnMg7j31O1N2O1O1N23L6K5K5K5K4L5K3L5L3M3M3M3M3M3M3L4M00nNRJhJm5o4_JnJa5i4jJVKU5a4WK\\Ki4]4`KbK^4[4iKbKW4[4nKdKQ4X4VLeKj3W4\\LgKd3U4cLiK\\3W4fLgKZ3Y4hLfKW3[4iLdKW3\\4kLcKT3^4mL`KT3`4lL`KS3`4oL^KQ3[4WMeKh2[4ZMcKf2^4[MaKd2_4T31O1O001O00N3M2N2O1N2N2001O1O1O1O1O010OBXGZLg8g3`GRL_8o3>001O001O1O001O10O01O001O000O2O001O001O00001O001N101O00001O001O1O001N101O001O1O00O1000000O10000O10000000000000000000000000001N1000oH`KW5`4gJdKW5\\4gJgKX5Y4fJjKY5U4fJnKY5R4eJRLY5n3eJULZ5k3\\J`Lc5`3RJkLn5U3hIUMX6j2cI\\M]6d2^IaMb6d4000C^IQJb6n5dImI\\6S6>O1O1O1O100O2N1O100O1O2N100N2N2N3M2N2\\Od0O2N1O1O1O1O2N1O11O1O010O1O1O0010O01O0001O01O01O01O01O01O8fHXKb5o4SJUKl5Z5aIkJ_6Q6O2N100O0001O000O2O001O00001O00001N101O2N2N1N3L3N3M2M4M2N3M2M4M2N3M2M4M2N3N1N3M3M2O2M2N3N3L4L4L4M3L4L4M3J6F:F;F9F:oNiDD^;5gDJ\\;OiD0b;AcD>W<N1N2O1O00O1O101N1O2O0O1O20O001O001O0O101O001N1O2N1N2O2N1O2N1N2Obgn1"}, "label": "cow closest to the camera"}]}
{"id": 399064, "image": "COCO_train2014_000000399064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pitcher wearing a black jersey just pitched the ball\"."}], "task": "refering", "answer_template": "The \"a pitcher wearing a black jersey just pitched the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 197, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290], "bbox": [0.5562343750000001, 0.32583138173302106, 0.701671875, 0.8449414519906322], "iscrowd": 0, "area": 11326.403850000002, "rle": {"size": [427, 640], "counts": "Zdd4l0^<3M4L3M3M4L3L5L3RHYNR4k1lKUNS4o1jKQNT4S2jKmMU4V2hKkMV4Z2gKgM[32RJZ2a2dM\\36nIZ2c2aM_38jI[2d2]Ma3=eIZ2f2[Md3?bIY2g2YMf3`3ZLaLf3^3ZLcLf3]3XLeLg3[3YLfLg3Y3VLkLj3T3PLTMo3l2jK[MV4d2cKdM]4[2]KlMc4T2VKSNj4l1hJcNX5\\1UJXOk5g0cIL]6a300000001O0000000000000003Mg0YO2O0O2N2N2\\Od0D;L5M3O0O100O100O1O100O1nMVGBj87]GHd81dGM]8MiG2X8MiG1Y8MjG1W8MkG2V8MkG2V8LmG2S8NnG1S8MoG1S8MPH1Q8NPH1Q8MRHOQ80PHLT82nGKU83nGHX85iGH[85gGG^87dGDb88`GEd88^GDh89ZGCj8:XGBn8;S2M4L3N3L[]_2"}, "label": "a pitcher wearing a black jersey just pitched the ball"}]}
{"id": 399064, "image": "COCO_train2014_000000399064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pitcher following through after pitch\"."}], "task": "refering", "answer_template": "The \"pitcher following through after pitch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 197, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290], "bbox": [0.5562343750000001, 0.32583138173302106, 0.701671875, 0.8449414519906322], "iscrowd": 0, "area": 11326.403850000002, "rle": {"size": [427, 640], "counts": "Zdd4l0^<3M4L3M3M4L3L5L3RHYNR4k1lKUNS4o1jKQNT4S2jKmMU4V2hKkMV4Z2gKgM[32RJZ2a2dM\\36nIZ2c2aM_38jI[2d2]Ma3=eIZ2f2[Md3?bIY2g2YMf3`3ZLaLf3^3ZLcLf3]3XLeLg3[3YLfLg3Y3VLkLj3T3PLTMo3l2jK[MV4d2cKdM]4[2]KlMc4T2VKSNj4l1hJcNX5\\1UJXOk5g0cIL]6a300000001O0000000000000003Mg0YO2O0O2N2N2\\Od0D;L5M3O0O100O100O1O100O1nMVGBj87]GHd81dGM]8MiG2X8MiG1Y8MjG1W8MkG2V8MkG2V8LmG2S8NnG1S8MoG1S8MPH1Q8NPH1Q8MRHOQ80PHLT82nGKU83nGHX85iGH[85gGG^87dGDb88`GEd88^GDh89ZGCj8:XGBn8;S2M4L3N3L[]_2"}, "label": "pitcher following through after pitch"}]}
{"id": 480991, "image": "COCO_train2014_000000480991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red part of the scooter or bike\"."}], "task": "refering", "answer_template": "The \"the red part of the scooter or bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 255, 256, 257, 260, 261, 262, 263, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0016875000000000002, 0.5867083333333334, 1.0, 0.989875], "iscrowd": 0, "area": 81709.04019999999, "rle": {"size": [480, 640], "counts": "ek0n0R1UOT:g2`E`MY9m3[FYLa9f4K4[OmJYGW5g8nJPGV5Q9=O01O1O001O10O01O001O1O001O1O001O00001O001O001O00001O001O00001O001O00001O001O001O001O001O001O001O001O001O1O1O2N1O1O2N1O2N001O1O001O1O001O1O1O0000001O000000001O000000001O000000001O000000001O00000000001O0000001O000000001O0fK]EW4c:300001O00000gK\\EV4d:jK\\EV4e:iK[EW4e:20000001gKZEV4f:jKZEV4f:jKZEV4f:jKZEV4h:0O100002N0000000000001O000000000000000000000000001O00000000000000000000000000000000000000000000O100000000O100000000O1000000O10000O100O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000000000001O000000000000001O000000001O00001O00001O0000001O00001O00001O00001O0000001O00001O001O1O001O001O1O001O1O001O001O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O001O00001O001O001O001O001O001O1O1O1O1O1O1O1O1O001O1O001O001O0000O1O100O1O100O1O100O1O1O100000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000O10000O10000O10000O100O10000O10000O10000O10000O100O10000O100O100O100O100O10000O100O100O100O100M3M3M3M3K5K5J6O10000001O000000001O00001O001O1O1O1O1O1O1O1O1O1O001N101O1O001O001O1O001O001O1O001O001N2O001O001O1O000000001O00000000001N100000001O00000000001O00000000001N100000001O00001O00001O001O00001O0O2O00001O00001O1O1O2N1O1_Ob0]Ob0@iD"}, "label": "the red part of the scooter or bike"}]}
{"id": 480991, "image": "COCO_train2014_000000480991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red scooty\"."}], "task": "refering", "answer_template": "The \"red scooty\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 255, 256, 257, 260, 261, 262, 263, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0016875000000000002, 0.5867083333333334, 1.0, 0.989875], "iscrowd": 0, "area": 81709.04019999999, "rle": {"size": [480, 640], "counts": "ek0n0R1UOT:g2`E`MY9m3[FYLa9f4K4[OmJYGW5g8nJPGV5Q9=O01O1O001O10O01O001O1O001O1O001O00001O001O001O00001O001O00001O001O00001O001O001O001O001O001O001O001O001O1O1O2N1O1O2N1O2N001O1O001O1O001O1O1O0000001O000000001O000000001O000000001O000000001O00000000001O0000001O000000001O0fK]EW4c:300001O00000gK\\EV4d:jK\\EV4e:iK[EW4e:20000001gKZEV4f:jKZEV4f:jKZEV4f:jKZEV4h:0O100002N0000000000001O000000000000000000000000001O00000000000000000000000000000000000000000000O100000000O100000000O1000000O10000O100O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000000000001O000000000000001O000000001O00001O00001O0000001O00001O00001O00001O0000001O00001O001O1O001O001O1O001O1O001O001O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O001O00001O001O001O001O001O001O1O1O1O1O1O1O1O1O001O1O001O001O0000O1O100O1O100O1O100O1O1O100000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000O10000O10000O10000O100O10000O10000O10000O10000O100O10000O100O100O100O100O10000O100O100O100O100M3M3M3M3K5K5J6O10000001O000000001O00001O001O1O1O1O1O1O1O1O1O1O001N101O1O001O001O1O001O001O1O001O001N2O001O001O1O000000001O00000000001N100000001O00000000001O00000000001N100000001O00001O00001O001O00001O0O2O00001O00001O1O1O2N1O1_Ob0]Ob0@iD"}, "label": "red scooty"}]}
{"id": 480991, "image": "COCO_train2014_000000480991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wheel in the background\"."}], "task": "refering", "answer_template": "The \"the wheel in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 84, 85, 86, 87, 88, 89, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204], "bbox": [0.647421875, 0.05964583333333333, 0.9265625, 0.48341666666666666], "iscrowd": 0, "area": 13442.542800000003, "rle": {"size": [480, 640], "counts": "lcR65h>3L4M3L4L5L3L4M3N2O1O2O0O1O100O2N100O1O100O2N100O1O101N1O100O1O2O0O1O100O1O2O0O1O100O2O0O1O100O2N100O1O100O2N100O1O1012M4L5K4L4L4M3L4L4L5K4M3L4L4L1O01O01O00000010O000001O01O2N2O1N2N3M2O1N3M2N2N2O2M2N2N3N1Nf0ZOh0XOi0XO;D000001O0000001O00000O101O0000001O0000001O0000001O0000001O000000001N1000001O0000001O00001O2N1O1O1O1O1M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M>Ca0^Ob0_Oa0^OUke0"}, "label": "the wheel in the background"}]}
{"id": 480991, "image": "COCO_train2014_000000480991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a barey visible tire in the background behind the luggage\"."}], "task": "refering", "answer_template": "The \"a barey visible tire in the background behind the luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 84, 85, 86, 87, 88, 89, 110, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204], "bbox": [0.647421875, 0.05964583333333333, 0.9265625, 0.48341666666666666], "iscrowd": 0, "area": 13442.542800000003, "rle": {"size": [480, 640], "counts": "lcR65h>3L4M3L4L5L3L4M3N2O1O2O0O1O100O2N100O1O100O2N100O1O101N1O100O1O2O0O1O100O1O2O0O1O100O2O0O1O100O2N100O1O100O2N100O1O1012M4L5K4L4L4M3L4L4L5K4M3L4L4L1O01O01O00000010O000001O01O2N2O1N2N3M2O1N3M2N2N2O2M2N2N3N1Nf0ZOh0XOi0XO;D000001O0000001O00000O101O0000001O0000001O0000001O0000001O000000001N1000001O0000001O00001O2N1O1O1O1O1M3N2M3N2M3N2M3N2M3N2M3N2M3N2M3N2M>Ca0^Ob0_Oa0^OUke0"}, "label": "a barey visible tire in the background behind the luggage"}]}
{"id": 464615, "image": "COCO_train2014_000000464615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person standing to the lesft of the person holding ball\"."}], "task": "refering", "answer_template": "The \"the person standing to the lesft of the person holding ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 33, 34, 35, 46, 47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 138, 139, 141, 142, 153, 154, 157, 158, 168, 169], "bbox": [0.07747072599531615, 0.065171875, 0.5422716627634661, 0.49214062499999994], "iscrowd": 0, "area": 21635.29990000001, "rle": {"size": [640, 427], "counts": "kgd01nc02N2N1N3N2N2N2N1O2N2M3N2N1O2N1O000O20O010O010O00010O010O010O0010O010O0010O11N2O1N101O1N2O001N2O1N2O1O5J7J6J5J7J5J7J8H9l_O[MZ>g4_Ob0]O\\1eN000O10000L4M3M3H8I7I700O100O10000O100O10000O10cM`DZL`;b3iDXLW;e3RETLP;h3YERLg:k3aEPL`:l3iEnKX:n3QFlKP:P4YFiKi9S4T3N200000000000002N2N2N1O2N2N2N1O2N2O1O00>CO001N101O00001O001O001O001N10001OmM\\@Id?Nf@2Y?EQA:o>\\O[Ae0d>ROfAm0Z>POjAo0V>oNmAQ1Q>oNQBP1o=oNSBP1m=oNTBQ1l=POTBo0l=QOUBn0k=SOUBm0j=TOVBk0j=UOWBj0i=WOVBi0j=XOVBg0i=\\OVBc0j=_OUB`0k=BTB>k=DTB;l=HQB8o=JPB5P>MoA2Q>0nAOR>3mALS>6lAIS>:kAGT>;kADU>>jAAV>a0iA^OW>d0hA\\OW>f0hAYOX>i0fAXOY>j0fAUOZ>k0gATOY>k0iAUOU>k0lAUOT>j0nAUOR>j0PBVOo=h0SBXOm=g0UBYOl=d0VB[Oo=?SB@R>:oAGV>2lAMY>MiA3\\>EfA;_>_OcA`0c>YO_Ag0e>SO]Al0i>mNXAT1k`0O1O010O10O01N2K5L3M4L4L5J6KUWj3"}, "label": "the person standing to the lesft of the person holding ball"}]}
{"id": 464615, "image": "COCO_train2014_000000464615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"runner in the back\"."}], "task": "refering", "answer_template": "The \"runner in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 33, 34, 35, 46, 47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 138, 139, 141, 142, 153, 154, 157, 158, 168, 169], "bbox": [0.07747072599531615, 0.065171875, 0.5422716627634661, 0.49214062499999994], "iscrowd": 0, "area": 21635.29990000001, "rle": {"size": [640, 427], "counts": "kgd01nc02N2N1N3N2N2N2N1O2N2M3N2N1O2N1O000O20O010O010O00010O010O010O0010O010O0010O11N2O1N101O1N2O001N2O1N2O1O5J7J6J5J7J5J7J8H9l_O[MZ>g4_Ob0]O\\1eN000O10000L4M3M3H8I7I700O100O10000O100O10000O10cM`DZL`;b3iDXLW;e3RETLP;h3YERLg:k3aEPL`:l3iEnKX:n3QFlKP:P4YFiKi9S4T3N200000000000002N2N2N1O2N2N2N1O2N2O1O00>CO001N101O00001O001O001O001N10001OmM\\@Id?Nf@2Y?EQA:o>\\O[Ae0d>ROfAm0Z>POjAo0V>oNmAQ1Q>oNQBP1o=oNSBP1m=oNTBQ1l=POTBo0l=QOUBn0k=SOUBm0j=TOVBk0j=UOWBj0i=WOVBi0j=XOVBg0i=\\OVBc0j=_OUB`0k=BTB>k=DTB;l=HQB8o=JPB5P>MoA2Q>0nAOR>3mALS>6lAIS>:kAGT>;kADU>>jAAV>a0iA^OW>d0hA\\OW>f0hAYOX>i0fAXOY>j0fAUOZ>k0gATOY>k0iAUOU>k0lAUOT>j0nAUOR>j0PBVOo=h0SBXOm=g0UBYOl=d0VB[Oo=?SB@R>:oAGV>2lAMY>MiA3\\>EfA;_>_OcA`0c>YO_Ag0e>SO]Al0i>mNXAT1k`0O1O010O10O01N2K5L3M4L4L5J6KUWj3"}, "label": "runner in the back"}]}
{"id": 464615, "image": "COCO_train2014_000000464615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man throwing baseball\"."}], "task": "refering", "answer_template": "The \"man throwing baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 98, 99, 100, 102, 103, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 173, 174, 175, 176, 187, 188, 189, 190, 191, 203, 204, 205, 206, 218, 219, 220, 221, 232, 233, 234, 235, 236, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 292, 293, 295, 296, 297, 307, 308, 321, 322, 323, 336], "bbox": [0.3982201405152225, 0.219234375, 0.9435128805620608, 0.96359375], "iscrowd": 0, "area": 43964.93424999999, "rle": {"size": [640, 427], "counts": "RkZ37`c09O1O010O100O1O100O100O10O10O10000O1jKJbD6X;<\\DDc;?[DAd;a0ZD@e;c0YD]Oe;f0YDZOg;i0UDYOj;i0SDYOl;i0RDXOl;l0PDVOo;l0oCUOP<n0lCTOS<P1hCROV<S1fCnNY<W1aCkN^<Z1\\CgNc<^1XCdNg<a1SCaNl<d1nB^NQ=f1hB^NV=h1bB\\N]=i1[B[Nd=i1UB[Nj=j1nAZNQ>k3O1O001O1O1O1O1O1O100\\HlI[1U6dNlI_O3RLQ6`4mIXO9ULk5c4lIRO`0YLe5d4lImNe0^L`5e4lIhNh0bL]5f4kIfNj0cL\\5g4kInLL^NP1LZ5g4kIZLd0mN91Y5h4jIfK]1ZOB7X5i4jI[Kj1^OWO>U5h4mITKR2_OSO?S3lNWMR6cNoJW2YOZOb0\\2GWM_5WOQJ@7_2EAi01CROe03P5KXJj3<VM4\\Nd0g0b41XJk38PMa2ROn26XJ^6m2ZIi2=WJW6T3ZId2c0UJR6Y3ZI`2d8`MZG_2k8_MSG`2Q9_MmF_2Z9^MdFb2`9[M_Fd2g9XMXFg2S:PMlEP3^:gL`EY3c:fLZE\\3g:dLVE]3k:dLQE^3Q;bLlD`3U;`LhDa3Y;c21N2O0O2O1O000O10001N10000O101N100001O00100O1O1O1O00100O1O001O1O0mEiG]8X8_GkG`8W8\\GlGd8T8ZGoGe8R8WGQHi8P8SGSHm8o7nFUHP9o7kFSHU9P8cFUH]9n7[FXHd9R900]NQFTIQ:b6ZF\\If9j5VGTJcNVOi9e6eGdJZ8[5hGdJX8[5iGeJX8Y5jGgJU8Y5lGfJT8Y5mGgJS8X5oGgJQ8X5QHhJn7W5THhJm7V5UHiJk7V5WHiJi7V5YHiJg7V5[HiJe7V5]HiJc7V5_HiJb7U5_HkJa7T5aHlJ^7T5cHkJ]7T5eHkJ[7W3^FUMY2CY7R3jFTMo1IX7c2^G]M\\1OV7l1\\HnM?6U7a1nHPNO?T7R1aISNAl0m6j0SMUOm2j0TMVOm2h0TMXOl2g0UMYOk2f0WMYOj2e0WM[Oi2d0XM\\Oh2c0YM]Oh2a0ZM_Oe2`0\\M_Oe2a0[M_Of2?[M@f2?\\M@d2`0\\M@e2>[MBf2=[MCe2<\\MCf2<ZMDf2;[MDf2;[MEf2:ZMFf29[MGe28\\MHe27[MIe26\\MKc24^MLc23^MLb23_MNa20`M0P1POhHo0Z60l0SOjHm0Z60k0TOkHk0[62g0VOmHg0]63e0WOnHf0^62b0[OoHb0`62b0]OoGF=l0R71b0^OlGH?i0S70b0AjGH`0g0U7Oa0ChGHb0f0U7Na0GeGGc0e0W7Mb05VHNX7Lc07THMZ7Kc08RHM\\7Ic0:PHN^7Ee0<mGO_7Ce0>lGO_7Ah0?iG0`7]Oj0c0fG0a7WOo0i0^G1d7QOR1R1VGMQa07c^O1]a0Y100001O001O1O001O001O1O2N3M4L8H7H5L4K5K9H8G9G9HZh>"}, "label": "man throwing baseball"}]}
{"id": 464615, "image": "COCO_train2014_000000464615.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player wearing white uniform and a cap throwing ball and holding gloves on his left hand\"."}], "task": "refering", "answer_template": "The \"baseball player wearing white uniform and a cap throwing ball and holding gloves on his left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 98, 99, 100, 102, 103, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 173, 174, 175, 176, 187, 188, 189, 190, 191, 203, 204, 205, 206, 218, 219, 220, 221, 232, 233, 234, 235, 236, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 292, 293, 295, 296, 297, 307, 308, 321, 322, 323, 336], "bbox": [0.3982201405152225, 0.219234375, 0.9435128805620608, 0.96359375], "iscrowd": 0, "area": 43964.93424999999, "rle": {"size": [640, 427], "counts": "RkZ37`c09O1O010O100O1O100O100O10O10O10000O1jKJbD6X;<\\DDc;?[DAd;a0ZD@e;c0YD]Oe;f0YDZOg;i0UDYOj;i0SDYOl;i0RDXOl;l0PDVOo;l0oCUOP<n0lCTOS<P1hCROV<S1fCnNY<W1aCkN^<Z1\\CgNc<^1XCdNg<a1SCaNl<d1nB^NQ=f1hB^NV=h1bB\\N]=i1[B[Nd=i1UB[Nj=j1nAZNQ>k3O1O001O1O1O1O1O1O100\\HlI[1U6dNlI_O3RLQ6`4mIXO9ULk5c4lIRO`0YLe5d4lImNe0^L`5e4lIhNh0bL]5f4kIfNj0cL\\5g4kInLL^NP1LZ5g4kIZLd0mN91Y5h4jIfK]1ZOB7X5i4jI[Kj1^OWO>U5h4mITKR2_OSO?S3lNWMR6cNoJW2YOZOb0\\2GWM_5WOQJ@7_2EAi01CROe03P5KXJj3<VM4\\Nd0g0b41XJk38PMa2ROn26XJ^6m2ZIi2=WJW6T3ZId2c0UJR6Y3ZI`2d8`MZG_2k8_MSG`2Q9_MmF_2Z9^MdFb2`9[M_Fd2g9XMXFg2S:PMlEP3^:gL`EY3c:fLZE\\3g:dLVE]3k:dLQE^3Q;bLlD`3U;`LhDa3Y;c21N2O0O2O1O000O10001N10000O101N100001O00100O1O1O1O00100O1O001O1O0mEiG]8X8_GkG`8W8\\GlGd8T8ZGoGe8R8WGQHi8P8SGSHm8o7nFUHP9o7kFSHU9P8cFUH]9n7[FXHd9R900]NQFTIQ:b6ZF\\If9j5VGTJcNVOi9e6eGdJZ8[5hGdJX8[5iGeJX8Y5jGgJU8Y5lGfJT8Y5mGgJS8X5oGgJQ8X5QHhJn7W5THhJm7V5UHiJk7V5WHiJi7V5YHiJg7V5[HiJe7V5]HiJc7V5_HiJb7U5_HkJa7T5aHlJ^7T5cHkJ]7T5eHkJ[7W3^FUMY2CY7R3jFTMo1IX7c2^G]M\\1OV7l1\\HnM?6U7a1nHPNO?T7R1aISNAl0m6j0SMUOm2j0TMVOm2h0TMXOl2g0UMYOk2f0WMYOj2e0WM[Oi2d0XM\\Oh2c0YM]Oh2a0ZM_Oe2`0\\M_Oe2a0[M_Of2?[M@f2?\\M@d2`0\\M@e2>[MBf2=[MCe2<\\MCf2<ZMDf2;[MDf2;[MEf2:ZMFf29[MGe28\\MHe27[MIe26\\MKc24^MLc23^MLb23_MNa20`M0P1POhHo0Z60l0SOjHm0Z60k0TOkHk0[62g0VOmHg0]63e0WOnHf0^62b0[OoHb0`62b0]OoGF=l0R71b0^OlGH?i0S70b0AjGH`0g0U7Oa0ChGHb0f0U7Na0GeGGc0e0W7Mb05VHNX7Lc07THMZ7Kc08RHM\\7Ic0:PHN^7Ee0<mGO_7Ce0>lGO_7Ah0?iG0`7]Oj0c0fG0a7WOo0i0^G1d7QOR1R1VGMQa07c^O1]a0Y100001O001O1O001O001O1O2N3M4L8H7H5L4K5K9H8G9G9HZh>"}, "label": "baseball player wearing white uniform and a cap throwing ball and holding gloves on his left hand"}]}
{"id": 145129, "image": "COCO_train2014_000000145129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"jean paul sartre ' s book nausea\"."}], "task": "refering", "answer_template": "The \"jean paul sartre ' s book nausea\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 170], "bbox": [0.26790625, 0.16879166666666665, 0.455296875, 0.4379375], "iscrowd": 0, "area": 11349.90685, "rle": {"size": [480, 640], "counts": "Vj`26g>5J6J6K3L4M3L4L4M3L4M2M4L4M3L4M3L4L4M3L4M3L3M4M3L4M3L4L4M3M300000O1000000000000000000O100000000000000000000O01000000000000O100000O1000O1000000000000O10O10000000O1000000000O10O10000000000O100000O100000O1000J6I7J6I6J7J6I7J6I7I7J5J7J6I9G:G``S5"}, "label": "jean paul sartre ' s book nausea"}]}
{"id": 145129, "image": "COCO_train2014_000000145129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"nausea book by jean paul sarte . ( white book with green drawing .)\"."}], "task": "refering", "answer_template": "The \"nausea book by jean paul sarte . ( white book with green drawing .)\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 170], "bbox": [0.26790625, 0.16879166666666665, 0.455296875, 0.4379375], "iscrowd": 0, "area": 11349.90685, "rle": {"size": [480, 640], "counts": "Vj`26g>5J6J6K3L4M3L4L4M3L4M2M4L4M3L4M3L4L4M3L4M3L3M4M3L4M3L4L4M3M300000O1000000000000000000O100000000000000000000O01000000000000O100000O1000O1000000000000O10O10000000O1000000000O10O10000000000O100000O100000O1000J6I7J6I6J7J6I7J6I7I7J5J7J6I9G:G``S5"}, "label": "nausea book by jean paul sarte . ( white book with green drawing .)"}]}
{"id": 366313, "image": "COCO_train2014_000000366313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver compact car driving down the road in the snow\"."}], "task": "refering", "answer_template": "The \"a silver compact car driving down the road in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 32, 33, 34, 35, 49, 50, 51, 52, 53, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 122, 123, 124], "bbox": [0.74382, 0.03975975975975976, 1.0, 0.5566066066066067], "iscrowd": 0, "area": 17090.655100000004, "rle": {"size": [333, 500], "counts": "`Ri34X:8H7eFA[8f0RGHl8m0N2L4L5K4L4K5L5K4L4O2O4K6J6J3M4L2O1N2N2N2N2N2O1N1O1O2N100O1O1O1O1O2O0O1O1O1O1O101N1O1O1O1O100O1O2N1O100O1O1O1O1O100O1O1O1O00100O1O1O1O00100O001O0010O1O100O1O100O1O100O1N1O2N2M3N200001O00000000001O00000O1000001O00000000001O00000000001N100000000j1WNP2oMhK"}, "label": "a silver compact car driving down the road in the snow"}]}
{"id": 366313, "image": "COCO_train2014_000000366313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a white car\"."}], "task": "refering", "answer_template": "The \"the back of a white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 32, 33, 34, 35, 49, 50, 51, 52, 53, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 122, 123, 124], "bbox": [0.74382, 0.03975975975975976, 1.0, 0.5566066066066067], "iscrowd": 0, "area": 17090.655100000004, "rle": {"size": [333, 500], "counts": "`Ri34X:8H7eFA[8f0RGHl8m0N2L4L5K4L4K5L5K4L4O2O4K6J6J3M4L2O1N2N2N2N2N2O1N1O1O2N100O1O1O1O1O2O0O1O1O1O1O101N1O1O1O1O100O1O2N1O100O1O1O1O1O100O1O1O1O00100O1O1O1O00100O001O0010O1O100O1O100O1O100O1N1O2N2M3N200001O00000000001O00000O1000001O00000000001O00000000001N100000000j1WNP2oMhK"}, "label": "the back of a white car"}]}
{"id": 366313, "image": "COCO_train2014_000000366313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red car\"."}], "task": "refering", "answer_template": "The \"a red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 18, 19, 20, 21, 22, 23, 24, 25, 27, 36, 37, 38, 39, 40, 41, 42, 43, 45, 46, 54, 55, 56, 57, 58, 59, 60, 63, 64, 72, 73, 75, 76, 77, 78, 81, 82, 96], "bbox": [0.0, 0.002972972972972973, 0.6043, 0.43984984984984987], "iscrowd": 0, "area": 26045.043749999997, "rle": {"size": [333, 500], "counts": "m1d0g9X1iN6VOf0\\Od0M4L3N2M3M3M2N3M2O2O001O001O001OO100000000000000O10000O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N200001O00000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O00000000001O00000O1000001O0000000000001O00000000001O000000001O001O001O_McIR1]6kNgIU1X6jNiIV1W6jNiIW1V6iNjIW1V6hNkIY1T6gNlIY1T6fNmI[1R6eNnI[1R6eNnI\\1Q6cNPJ]1P6cNPJ^1o5aNRJ`1m5`NSJa1l5_NTJb1k5]NWJc1h5]NXJc1h5]NXJd1g5[NZJf1e5ZN[Jg1d5XN]Ji1b5WN^Jj1a5VN_Jk1`5TNaJm1^5SNbJn1]5QNdJP2[5PNeJQ2Z5oMfJR2Y5mMhJ[1@]Ng58iJP1LgN[58kJg03QOR58kJe05SOP58kJc08TOm48lJc08UOl48lJa0;VOi49lJ`0<VOi49lJ?=XOg49lJ>?XOe49mJ=`0ZOc49mJ<b0ZOa4:mJ:d0\\O_49nJ:e0\\O]4:nJ8g0^O[49oJ8g0^O[4:oJ5i0@X4;oJ4j0AW4:PK3l0BT4;PK2U7MlH2U7NkH1V7OjH0W7OjH1V7OjH0W70iHOX70iHOX71hHNY71hHNY72gHMZ73gHKZ74gHKZ75fHI\\76eHI\\77dHH]78cHF_79bHF_7:Z1O10000O100000000001O000000000000000010O001O1O1O001O1O10O01O1O_GKk64TINk62TI0l6ORI4n6KQI7o6InH9R7GmH<R7CmH`0R7_OlHd0]83L4M8H7H7K4L4L4L5K4L4L3M001O1O0O101O0001O1O1O1O2N1O1O1O1O1O2N1O1N2M3N2N3M2N2N2M3N2M4M2M3N2M3BYZP2"}, "label": "a red car"}]}
{"id": 366313, "image": "COCO_train2014_000000366313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red car\"."}], "task": "refering", "answer_template": "The \"red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 18, 19, 20, 21, 22, 23, 24, 25, 27, 36, 37, 38, 39, 40, 41, 42, 43, 45, 46, 54, 55, 56, 57, 58, 59, 60, 63, 64, 72, 73, 75, 76, 77, 78, 81, 82, 96], "bbox": [0.0, 0.002972972972972973, 0.6043, 0.43984984984984987], "iscrowd": 0, "area": 26045.043749999997, "rle": {"size": [333, 500], "counts": "m1d0g9X1iN6VOf0\\Od0M4L3N2M3M3M2N3M2O2O001O001O001OO100000000000000O10000O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N200001O00000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O00000000001O00000O1000001O0000000000001O00000000001O000000001O001O001O_McIR1]6kNgIU1X6jNiIV1W6jNiIW1V6iNjIW1V6hNkIY1T6gNlIY1T6fNmI[1R6eNnI[1R6eNnI\\1Q6cNPJ]1P6cNPJ^1o5aNRJ`1m5`NSJa1l5_NTJb1k5]NWJc1h5]NXJc1h5]NXJd1g5[NZJf1e5ZN[Jg1d5XN]Ji1b5WN^Jj1a5VN_Jk1`5TNaJm1^5SNbJn1]5QNdJP2[5PNeJQ2Z5oMfJR2Y5mMhJ[1@]Ng58iJP1LgN[58kJg03QOR58kJe05SOP58kJc08TOm48lJc08UOl48lJa0;VOi49lJ`0<VOi49lJ?=XOg49lJ>?XOe49mJ=`0ZOc49mJ<b0ZOa4:mJ:d0\\O_49nJ:e0\\O]4:nJ8g0^O[49oJ8g0^O[4:oJ5i0@X4;oJ4j0AW4:PK3l0BT4;PK2U7MlH2U7NkH1V7OjH0W7OjH1V7OjH0W70iHOX70iHOX71hHNY71hHNY72gHMZ73gHKZ74gHKZ75fHI\\76eHI\\77dHH]78cHF_79bHF_7:Z1O10000O100000000001O000000000000000010O001O1O1O001O1O10O01O1O_GKk64TINk62TI0l6ORI4n6KQI7o6InH9R7GmH<R7CmH`0R7_OlHd0]83L4M8H7H7K4L4L4L5K4L4L3M001O1O0O101O0001O1O1O1O2N1O1O1O1O1O2N1O1N2M3N2N3M2N2N2M3N2M4M2M3N2M3BYZP2"}, "label": "red car"}]}
{"id": 325362, "image": "COCO_train2014_000000325362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady drinking coffee with her boy friend\"."}], "task": "refering", "answer_template": "The \"a lady drinking coffee with her boy friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 223, 224, 225, 226, 227, 242, 243, 244, 245, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.52696875, 0.22935714285714284, 0.9539375, 1.0], "iscrowd": 0, "area": 56700.47359999999, "rle": {"size": [420, 640], "counts": "^bZ44Q=8G101O7I00000000000N2N2kL_OXI2bMb0U9FmHh0Q7ZOlHi0R7YOlHi0S7XOmHg0R7[OmHf0g6FYI:e6H[I8d6I\\I6c6L]I4b6M]I3b6N_I2`6O_I1a60]I1a62^IN^66aIK\\69cIG[6<dIE[6;dIF]6:bIF^6;`IF`6;_IEb6:]IGc6:\\IFe6:YIGg6:YIEh6:XIFh6<VIEi6`0RI@o6d0lH\\Oe6aNcHX2d0VOh6eNcHY2`0ROm6fNbH]2<mNS7fNaHa27iNX7gN`He23dN]7iN^Hh21^Nb7jN]H_4c7bK\\H_4c7cK\\H\\4d7eK[H[4f7eKZHZ4f7b0001O000010YNiJ_KY5\\4kJeKU5U4PKkKQ5P4SKQLm4j3WKVLj4d3[K\\Lf4_3^KbLa4Y3dKhLZ4U3jKnLR4P3QLSMk3l2WLVMf3i2\\LZM`3e2dL[MY3d2jL^MR3a2QMaMk2_2XMaMe2^2^MdM^2[2fMeMW2Z2lMhMP2W2TNjMh1U2[NlMb1S2bNnMZ1Q2iNPNT1o1SOoMi0P2XOTNd0k1]OXN`0g1B\\N:c1G`N6_1KeN1Z10jNLU16mNGR1:ROBm0?VO^Oi0d0ZOXOe0i0_OSO`0n0CoN=R1FjN9W1JfN5^1L^N3g1LVN3o1HRN7Q2EQN:Q2DPN;R2BPN=Q2BPN=R2AoM>S2_LlKj2S2f0R2]LnKl2Q2f0Y2WOiMh0X2WOiMh0Y2UOiMk0W2TOjMk0X2ROjMn0V2QOkMo0Z2kNgMT1^2fNdMZ1^2cNcM\\1`2`NbM`1_2^NbMa1_2]NcMc1]2\\NdMc1]2[NeMe1[2ZNfMe1[2ZNfMf1Z2XNhMh1X2WNiMh1Y2UNiMk1W2TNjMk1W2SNkMm1U2RNlMm1T2SNmMm1S2QNoMn1R2QNoMo1P2PNRNP2n1oMSNP2n1nMTNR2k1nMVNQ2k1mMWNS2h1mMYNR2h1mMYNS2g1kM[NT2e1kM]NU2c1jM^NU2c1iM_NW2`1iMaNW2_1gMcNX2^1fMdNZ2[1fMfNY2[1eMgN\\2W1dMjN\\2W1aMkN_2V1^MlNb2V1[MkNe2W1WMkNj2V1SMkNm2V1PMlNP3d500001O000000001O000000001O00000000001O000000001O00001O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M2N2N2N2N3M2N2_JZK`2h4[M\\Kd2f4XM]Kg2e4WM\\Kh2g4UMZKj2h4SMZKl2h4RMYKm2i4PMZKn2h4PMYKo2j4mLXKR3j4lLWKS3k4jLWKU3k4iLVKV3l4hLUKW3m4fLUKY3n4dLSK[3o4bLSK]3o4`LSK_3o4^LSKa3o4\\LSKc3P5YLRKf3P5XLQKg3Q5WLQKg3T5TLmJk3X5PLiJo3\\5lKeJS4`5hKaJW4e5cK\\J\\4W7O1O1O1O1O1O1O1O1O2N1O2N4L3M3M3M3M3M3J6H8H8H9G8G9Ha1_NQa;"}, "label": "a lady drinking coffee with her boy friend"}]}
{"id": 325362, "image": "COCO_train2014_000000325362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman drinking out of cup\"."}], "task": "refering", "answer_template": "The \"woman drinking out of cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 223, 224, 225, 226, 227, 242, 243, 244, 245, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.52696875, 0.22935714285714284, 0.9539375, 1.0], "iscrowd": 0, "area": 56700.47359999999, "rle": {"size": [420, 640], "counts": "^bZ44Q=8G101O7I00000000000N2N2kL_OXI2bMb0U9FmHh0Q7ZOlHi0R7YOlHi0S7XOmHg0R7[OmHf0g6FYI:e6H[I8d6I\\I6c6L]I4b6M]I3b6N_I2`6O_I1a60]I1a62^IN^66aIK\\69cIG[6<dIE[6;dIF]6:bIF^6;`IF`6;_IEb6:]IGc6:\\IFe6:YIGg6:YIEh6:XIFh6<VIEi6`0RI@o6d0lH\\Oe6aNcHX2d0VOh6eNcHY2`0ROm6fNbH]2<mNS7fNaHa27iNX7gN`He23dN]7iN^Hh21^Nb7jN]H_4c7bK\\H_4c7cK\\H\\4d7eK[H[4f7eKZHZ4f7b0001O000010YNiJ_KY5\\4kJeKU5U4PKkKQ5P4SKQLm4j3WKVLj4d3[K\\Lf4_3^KbLa4Y3dKhLZ4U3jKnLR4P3QLSMk3l2WLVMf3i2\\LZM`3e2dL[MY3d2jL^MR3a2QMaMk2_2XMaMe2^2^MdM^2[2fMeMW2Z2lMhMP2W2TNjMh1U2[NlMb1S2bNnMZ1Q2iNPNT1o1SOoMi0P2XOTNd0k1]OXN`0g1B\\N:c1G`N6_1KeN1Z10jNLU16mNGR1:ROBm0?VO^Oi0d0ZOXOe0i0_OSO`0n0CoN=R1FjN9W1JfN5^1L^N3g1LVN3o1HRN7Q2EQN:Q2DPN;R2BPN=Q2BPN=R2AoM>S2_LlKj2S2f0R2]LnKl2Q2f0Y2WOiMh0X2WOiMh0Y2UOiMk0W2TOjMk0X2ROjMn0V2QOkMo0Z2kNgMT1^2fNdMZ1^2cNcM\\1`2`NbM`1_2^NbMa1_2]NcMc1]2\\NdMc1]2[NeMe1[2ZNfMe1[2ZNfMf1Z2XNhMh1X2WNiMh1Y2UNiMk1W2TNjMk1W2SNkMm1U2RNlMm1T2SNmMm1S2QNoMn1R2QNoMo1P2PNRNP2n1oMSNP2n1nMTNR2k1nMVNQ2k1mMWNS2h1mMYNR2h1mMYNS2g1kM[NT2e1kM]NU2c1jM^NU2c1iM_NW2`1iMaNW2_1gMcNX2^1fMdNZ2[1fMfNY2[1eMgN\\2W1dMjN\\2W1aMkN_2V1^MlNb2V1[MkNe2W1WMkNj2V1SMkNm2V1PMlNP3d500001O000000001O000000001O00000000001O000000001O00001O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M2N2N2N2N3M2N2_JZK`2h4[M\\Kd2f4XM]Kg2e4WM\\Kh2g4UMZKj2h4SMZKl2h4RMYKm2i4PMZKn2h4PMYKo2j4mLXKR3j4lLWKS3k4jLWKU3k4iLVKV3l4hLUKW3m4fLUKY3n4dLSK[3o4bLSK]3o4`LSK_3o4^LSKa3o4\\LSKc3P5YLRKf3P5XLQKg3Q5WLQKg3T5TLmJk3X5PLiJo3\\5lKeJS4`5hKaJW4e5cK\\J\\4W7O1O1O1O1O1O1O1O1O2N1O2N4L3M3M3M3M3M3J6H8H8H9G8G9Ha1_NQa;"}, "label": "woman drinking out of cup"}]}
{"id": 325362, "image": "COCO_train2014_000000325362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young adult making food in his kitchen , taken in blank and white\"."}], "task": "refering", "answer_template": "The \"a young adult making food in his kitchen , taken in blank and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 332, 333], "bbox": [0.07703125, 0.11795238095238095, 0.5125625, 0.9892857142857143], "iscrowd": 0, "area": 54532.740049999986, "rle": {"size": [420, 640], "counts": "_[d01m<<D<D<D<D<D<E;_H_MQ4i2cK]MZ4l2[KYMb4\\4fIjKW6c5N2M3M3N2M3M3M3N2M3M3M3NoHXJj6o52N2N2N2N2N2N2M3N2N2N2N2N2O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O2N1L4I7H8K5N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1OoHmMS3R2mLPNR3o1nLTNP3k1PMXNn2f1SM\\Nm2a1TMbNj2]1VMfNh2X1YMjNf2T1[MoNk1h1UN[Nh1e1XN^Ne1c1ZN_Nd1b1[NaNb1`1]NcNoL@n2o12cNcL7m2W1?eNVLe0V3f0d0j0XOUOi0o0SOQOn0R1nNnNR1S1mNmNT1R1lNnNT1S1kNmNV1R1jNnNV1S1iNmNX1R1iNmNX1S1gNmNY1T1fNlN[1S1eNnNZ1S1eNmN\\1R1dNnN\\1S1cNmN^1R1bNoN^1Q1aNoN_1Q1aNoN_1R1`NoN_1R1`NnN_1S1aNmN_1T1`NlN`1T1`NmN^1U1aNkN_1U1aNkN_1V1`NkN^1V1bNjN^1W1aNiN_1W1aNjN]1X1bNhN^1Y1aNgN_1Y1aNhN]1Z1bNfN^1Z1bNgN]1Z1bNfN]1[1cNeN]1\\1bNeN]1[1cNeN\\1]1dNbN\\1^1dNcN[1^1dNbN[1`1dN`N\\1`1dNaN[1_1eNaNZ1a1eN`NZ1`1fN`NZ1a1eN`NY1a1gN_NY1b1fN_NY1a1gN_NX1b1hN_NW1b1hN^NX1b1hN_NQ1h1nNXNl0n1TOSNj0n1VORNi0P2VOQNh0P2XOPNg0R2XOoMf0R2ZOoMd0R2\\OnMc0T2\\OmMb0T2^OlMa0V2^OlM?U2AnM;T2DPN7Q2IRN3o1MTNOn11TNKm15WNFk19VNEk1;VNCk1=VNAl1>UN@l1`0UN^Om1a0TN\\On1d0SNXOP2h0QNTOS2k0nMROT2n0mMnNW2Q1jMkNY2U1hMhN[2W1fMeN]2[1dMaN_2`1aM]Nb2b1_MZNd2f1^40O00010O010O00010O001O00010O00001O0010O00100O100O00100O1O1O001O1O1O1N2O1O1O1O2N1O1N2O1O2N1O1O1O1Naco3"}, "label": "a young adult making food in his kitchen , taken in blank and white"}]}
{"id": 325362, "image": "COCO_train2014_000000325362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man makind coffee\"."}], "task": "refering", "answer_template": "The \"man makind coffee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 332, 333], "bbox": [0.07703125, 0.11795238095238095, 0.5125625, 0.9892857142857143], "iscrowd": 0, "area": 54532.740049999986, "rle": {"size": [420, 640], "counts": "_[d01m<<D<D<D<D<D<E;_H_MQ4i2cK]MZ4l2[KYMb4\\4fIjKW6c5N2M3M3N2M3M3M3N2M3M3M3NoHXJj6o52N2N2N2N2N2N2M3N2N2N2N2N2O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O2N1L4I7H8K5N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1OoHmMS3R2mLPNR3o1nLTNP3k1PMXNn2f1SM\\Nm2a1TMbNj2]1VMfNh2X1YMjNf2T1[MoNk1h1UN[Nh1e1XN^Ne1c1ZN_Nd1b1[NaNb1`1]NcNoL@n2o12cNcL7m2W1?eNVLe0V3f0d0j0XOUOi0o0SOQOn0R1nNnNR1S1mNmNT1R1lNnNT1S1kNmNV1R1jNnNV1S1iNmNX1R1iNmNX1S1gNmNY1T1fNlN[1S1eNnNZ1S1eNmN\\1R1dNnN\\1S1cNmN^1R1bNoN^1Q1aNoN_1Q1aNoN_1R1`NoN_1R1`NnN_1S1aNmN_1T1`NlN`1T1`NmN^1U1aNkN_1U1aNkN_1V1`NkN^1V1bNjN^1W1aNiN_1W1aNjN]1X1bNhN^1Y1aNgN_1Y1aNhN]1Z1bNfN^1Z1bNgN]1Z1bNfN]1[1cNeN]1\\1bNeN]1[1cNeN\\1]1dNbN\\1^1dNcN[1^1dNbN[1`1dN`N\\1`1dNaN[1_1eNaNZ1a1eN`NZ1`1fN`NZ1a1eN`NY1a1gN_NY1b1fN_NY1a1gN_NX1b1hN_NW1b1hN^NX1b1hN_NQ1h1nNXNl0n1TOSNj0n1VORNi0P2VOQNh0P2XOPNg0R2XOoMf0R2ZOoMd0R2\\OnMc0T2\\OmMb0T2^OlMa0V2^OlM?U2AnM;T2DPN7Q2IRN3o1MTNOn11TNKm15WNFk19VNEk1;VNCk1=VNAl1>UN@l1`0UN^Om1a0TN\\On1d0SNXOP2h0QNTOS2k0nMROT2n0mMnNW2Q1jMkNY2U1hMhN[2W1fMeN]2[1dMaN_2`1aM]Nb2b1_MZNd2f1^40O00010O010O00010O001O00010O00001O0010O00100O100O00100O1O1O001O1O1O1N2O1O1O1O2N1O1N2O1O2N1O1O1O1Naco3"}, "label": "man makind coffee"}]}
{"id": 358134, "image": "COCO_train2014_000000358134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"that giraffe that is not facing towards the camera\"."}], "task": "refering", "answer_template": "The \"that giraffe that is not facing towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 130, 151, 152, 153, 154, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362, 363, 382, 383, 384, 385, 386], "bbox": [0.543390625, 0.2311875, 0.8237343749999999, 0.9619583333333334], "iscrowd": 0, "area": 23728.088, "rle": {"size": [480, 640], "counts": "mXS59e>5J4_OA]B`0a=a0O10O01N1N2O1G9O101002N1O2N2N3M0O010O0010L301O2N3N2O1NcNlBo0U=mNnBT1_=0O2O010N1O101N1O2N1O2N1O1O1ObNiBT1V=kNmBU1R=hNQCX1n<gNUCX1j<fNZCY1W=000O100O100O10000O100]ClNVOKg;Y1REZOm:f0PE^OS5ZOFX1UKAR5YOJV1QKEP5XOOT1nJ1g4lN;T1iJ:d4dNc0T1dJ?0SNV4;W1S1\\JU1R4QNb1g3YN[Lg1g3TN\\Ll1g3nM\\LS2e3gK^K>P1l3e3`KcK9n0W4a3ZKkK1l0e4\\3TKPLKk0Q5U5^JVKb5T70O100O1O1O1G9O1N2lNT1[Of0H7L4L4[MYImLk6S3a201000001N2O03N2N1N2O1N2N2N2N2nEVLi8T5M3M4L3M3M7^HaJT5l5WJ[Jj5k5lIXJU6n5`IVJa6T71O001O0bKPJLQ64QJSMGX1X6g1SJfL2]1l5m1fKnMZ4T2iKgMX4Z2lK`MT4a2QLYMP4h2ULQMl3P3YLYLiL6n6c3dLYL^3h3cLTL^3m3eLnK]3S4g3100O2O0jHkK_3W4[LnKe3S4ULQLk3R4oKQLR4P4jKRLV4P4eKSL\\4n3_KULa4n3XKVLi4R4jJSLU5W4]JmKd5[4kImKV6^61O1O1O1O2N1O1O2N1O2O0O2N2N1O2N1hKWHoM6S2e7GbHkM8R2W7LoHgM5U2m6MbJN`5JkJ1V5GVK4n4A^K;d4[OhK`0[4UORLf0Q4PO[Lk0R9J5]OVBGm=0]BKVod1"}, "label": "that giraffe that is not facing towards the camera"}]}
{"id": 358134, "image": "COCO_train2014_000000358134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe standing on right side\"."}], "task": "refering", "answer_template": "The \"giraffe standing on right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 130, 151, 152, 153, 154, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362, 363, 382, 383, 384, 385, 386], "bbox": [0.543390625, 0.2311875, 0.8237343749999999, 0.9619583333333334], "iscrowd": 0, "area": 23728.088, "rle": {"size": [480, 640], "counts": "mXS59e>5J4_OA]B`0a=a0O10O01N1N2O1G9O101002N1O2N2N3M0O010O0010L301O2N3N2O1NcNlBo0U=mNnBT1_=0O2O010N1O101N1O2N1O2N1O1O1ObNiBT1V=kNmBU1R=hNQCX1n<gNUCX1j<fNZCY1W=000O100O100O10000O100]ClNVOKg;Y1REZOm:f0PE^OS5ZOFX1UKAR5YOJV1QKEP5XOOT1nJ1g4lN;T1iJ:d4dNc0T1dJ?0SNV4;W1S1\\JU1R4QNb1g3YN[Lg1g3TN\\Ll1g3nM\\LS2e3gK^K>P1l3e3`KcK9n0W4a3ZKkK1l0e4\\3TKPLKk0Q5U5^JVKb5T70O100O1O1O1G9O1N2lNT1[Of0H7L4L4[MYImLk6S3a201000001N2O03N2N1N2O1N2N2N2N2nEVLi8T5M3M4L3M3M7^HaJT5l5WJ[Jj5k5lIXJU6n5`IVJa6T71O001O0bKPJLQ64QJSMGX1X6g1SJfL2]1l5m1fKnMZ4T2iKgMX4Z2lK`MT4a2QLYMP4h2ULQMl3P3YLYLiL6n6c3dLYL^3h3cLTL^3m3eLnK]3S4g3100O2O0jHkK_3W4[LnKe3S4ULQLk3R4oKQLR4P4jKRLV4P4eKSL\\4n3_KULa4n3XKVLi4R4jJSLU5W4]JmKd5[4kImKV6^61O1O1O1O2N1O1O2N1O2O0O2N2N1O2N1hKWHoM6S2e7GbHkM8R2W7LoHgM5U2m6MbJN`5JkJ1V5GVK4n4A^K;d4[OhK`0[4UORLf0Q4PO[Lk0R9J5]OVBGm=0]BKVod1"}, "label": "giraffe standing on right side"}]}
{"id": 358134, "image": "COCO_train2014_000000358134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe facing camera\"."}], "task": "refering", "answer_template": "The \"giraffe facing camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 147, 170, 192, 193, 215, 216, 217, 237, 238, 239, 240, 261, 262, 263, 284, 285, 307, 308, 330, 331, 332, 353, 354, 355], "bbox": [0.340609375, 0.28210416666666666, 0.44967187500000005, 0.9420416666666666], "iscrowd": 0, "area": 12077.913300000002, "rle": {"size": [480, 640], "counts": "T_V35h>;F9F;F:E;F:E=D=\\EcMm7j2ZGhM_8f2gFmMS9R4N001N2O1N2OmNRGULn8]2SGiM=A^8]2dGjM<_Om7a2UHfMY9_2jFXMW9m2kFjLW9[3S16K4K:dMTLSIX4c6lKWI^4@ZLj4YObKe4UO\\LX5PO_Km4jN^Lf5fNZKP7e4QIVKT7Z4cHmJ;c0W7[4eHPK4`0]7Y4hHSK0:^7e4bHoJ3Nf7T5^11N2O1O1N2O1`KdJfM5dNX5b3gJdM>aNl4k3fJ_Mk0^N`4R4eJ[MV1\\NV4Y4dJVMb1ZNk3_4cJRMh7n2XHUMd7@YHg14nN\\7ZOZIP1ZOKU7TO]J8^Nj0McNh59TL6WN\\2_5^M^L3SNd2Z5WMgL3oMl2U5oLoL3mMn2S5RLgJ<]2b0iMP3T5PLiJ;^2b0eM?0e0U5lMmJ:\\2d0bM9;j0_5TNhLg0^M3l0i0P5YNkLh0YMO\\1g0c4]NlLk0VMKi1d0^7_OiFHW2?S7GfFEe2;h6McFCP3:_61aF@\\38U66aJCc54gJG\\51ikT5"}, "label": "giraffe facing camera"}]}
{"id": 358134, "image": "COCO_train2014_000000358134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left that is facing foward\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left that is facing foward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 147, 170, 192, 193, 215, 216, 217, 237, 238, 239, 240, 261, 262, 263, 284, 285, 307, 308, 330, 331, 332, 353, 354, 355], "bbox": [0.340609375, 0.28210416666666666, 0.44967187500000005, 0.9420416666666666], "iscrowd": 0, "area": 12077.913300000002, "rle": {"size": [480, 640], "counts": "T_V35h>;F9F;F:E;F:E=D=\\EcMm7j2ZGhM_8f2gFmMS9R4N001N2O1N2OmNRGULn8]2SGiM=A^8]2dGjM<_Om7a2UHfMY9_2jFXMW9m2kFjLW9[3S16K4K:dMTLSIX4c6lKWI^4@ZLj4YObKe4UO\\LX5PO_Km4jN^Lf5fNZKP7e4QIVKT7Z4cHmJ;c0W7[4eHPK4`0]7Y4hHSK0:^7e4bHoJ3Nf7T5^11N2O1O1N2O1`KdJfM5dNX5b3gJdM>aNl4k3fJ_Mk0^N`4R4eJ[MV1\\NV4Y4dJVMb1ZNk3_4cJRMh7n2XHUMd7@YHg14nN\\7ZOZIP1ZOKU7TO]J8^Nj0McNh59TL6WN\\2_5^M^L3SNd2Z5WMgL3oMl2U5oLoL3mMn2S5RLgJ<]2b0iMP3T5PLiJ;^2b0eM?0e0U5lMmJ:\\2d0bM9;j0_5TNhLg0^M3l0i0P5YNkLh0YMO\\1g0c4]NlLk0VMKi1d0^7_OiFHW2?S7GfFEe2;h6McFCP3:_61aF@\\38U66aJCc54gJG\\51ikT5"}, "label": "the giraffe on the left that is facing foward"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 51, 52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.04090625, 0.08312499999999999, 0.448046875, 0.9844166666666666], "iscrowd": 0, "area": 77912.02315, "rle": {"size": [480, 640], "counts": "d_<2j>9H7H8I7H8I8G8I7I6K6I6K6J6J5J7J5K6J6I6K6J6J5J7J5K5K4L3M3M3M4L3M3M3M4L2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2cIUIb4n6RK_Ik4g6jJ`IT5h6cJ[IZ5o6]JSIa5W8M201O100O100O10000O100O100O100O100O10000O100O100O100O10000O1000000O1000000O1000000O100000000O1000000O0100000O10000aNoEoMQ:m1UFPNl9l1[FQNe9m1_FPNb9n1cFnM^9P2fFmM[9Q2iFlMX9S2kFjMV9k1VGSNk8h1\\GUNe8j1]GTNd8j1aGRN`8l1lGiMU8U2\\2O1N2N2O1N2N2N2O1N2N2O1000000000000O1000000000000O1000000000O0100000000000000000000000000000000001O001O1O001O001O1O001O001O1O0dEYN_7h1WHdNf7^1WHfNf7]1UHfNj7\\1SHeNm7^1nGeNQ8]1lGdNT8_1gGdNX8^1dGdN\\8_1`GcN_8`1\\GbNd8a1XGaNg8b1TG`Nl8c1nF`NR9c1gFbNX9a1bFbN^9a1\\FcNc9`1VFdNi9[3O001O001O001O001O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001WN\\JTJe5e5dJXJ]5a5lJ\\JU5^5SK_JP5X5ZKdJh4T5aKRJY5c5X2E<C<D<J6I7J6J6J6^Ob0^Ob0@`0B>C=BbXU5"}, "label": "a man in an orange shirt"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 51, 52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.04090625, 0.08312499999999999, 0.448046875, 0.9844166666666666], "iscrowd": 0, "area": 77912.02315, "rle": {"size": [480, 640], "counts": "d_<2j>9H7H8I7H8I8G8I7I6K6I6K6J6J5J7J5K6J6I6K6J6J5J7J5K5K4L3M3M3M4L3M3M3M4L2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2cIUIb4n6RK_Ik4g6jJ`IT5h6cJ[IZ5o6]JSIa5W8M201O100O100O10000O100O100O100O100O10000O100O100O100O10000O1000000O1000000O1000000O100000000O1000000O0100000O10000aNoEoMQ:m1UFPNl9l1[FQNe9m1_FPNb9n1cFnM^9P2fFmM[9Q2iFlMX9S2kFjMV9k1VGSNk8h1\\GUNe8j1]GTNd8j1aGRN`8l1lGiMU8U2\\2O1N2N2O1N2N2N2O1N2N2O1000000000000O1000000000000O1000000000O0100000000000000000000000000000000001O001O1O001O001O1O001O001O1O0dEYN_7h1WHdNf7^1WHfNf7]1UHfNj7\\1SHeNm7^1nGeNQ8]1lGdNT8_1gGdNX8^1dGdN\\8_1`GcN_8`1\\GbNd8a1XGaNg8b1TG`Nl8c1nF`NR9c1gFbNX9a1bFbN^9a1\\FcNc9`1VFdNi9[3O001O001O001O001O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001WN\\JTJe5e5dJXJ]5a5lJ\\JU5^5SK_JP5X5ZKdJh4T5aKRJY5c5X2E<C<D<J6I7J6J6J6^Ob0^Ob0@`0B>C=BbXU5"}, "label": "a man in an orange shirt"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with curly hair and glasses posing with a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man with curly hair and glasses posing with a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.43820312499999997, 0.13483333333333333, 0.7938124999999999, 0.9865208333333333], "iscrowd": 0, "area": 63059.00755, "rle": {"size": [480, 640], "counts": "]bS42a>c0]Oc0]Od0\\Oc0]Oc0]Oc0]Oc0B?H7H8H8I8G8H9H9F9G:G8G:F:G4K2N2O1N2N2N2O1N2N20000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O1_MSGoNm8P1VGmNk8R1YGkNg8T1\\GjNd8U1`GgNa8X1cGeN]8Z1fGcN[8\\1iGaNW8^1lG`NT8_1PH]NQ8b1RH\\Nn7c1VHZNj7e1YHXNh7g1\\HVNd7i1`HSNa7k1cHSN]7l1gHQNY7n1jHoMW7P2mHmMS7R2QIjMP7U2VIfMj6Y2\\IbMd6]2bI]M_6b2gIYMY6f2Q3O1O1O1O1O1O1O1O1O1O1O1O10O1000O0100O100O100O100O100O100O010O100O100O10000O101N101N101N2O0O2O0O2O0O2O0O2O0TGRMe5?\\Gk1l2iMg5;^Gl1i2lMh57`Gn1e2mMj55bGo1b2nMk52dGP2_2QNk5OfGR2\\2QNn5LgGS2Y2TNn5IjGT2V2TNP6GkGU2S2WNP6CoGW2n1YNR6_OQHX2k1[NS6]OSHY2g1]NU6YOUHZ2e1_NT6WOXH[2^1cNZ6QOYH\\2W1jN^6jN\\H]2P1oNc6cN^H^2i0UOh6]N`H_2b0ZOm6WNaH`2;@R7QNcH_26EW7lMcH`2OK]7fMcH_2K0a7bMdH_2D6g7[MeHh5Y7YJgHg5W7oI[H4>n5U7nIaH1:R6R7nIgHN7T6P7nIkHN5U6n6mInHN4V6k6mIRIN2U6j6nIUIM1V6h6mIXIM0W6e6mI\\ILOW6c6nI_IKNX6a6mIbIKMX6_6nIeIJLY6\\6nIiIIKZ6Z6mIlIIJZ6X6nIoIIHZ6V6nISJHG[6T6lIWJIE[6R6mIZJHD\\6o5mI^JGC]6k5nIcJEB]6h5PJgJCA^6e5PJkJB@_6a5RJoJ_O@_6_5VJoJ\\OA_6\\5ZJQKWOC`6Y5]JRKSOE`6V5bJSKnNGa6S5eJTKjNIa6P5jJUKeNKb6m4mJVKaNMc6i4QKXK\\NOc6f4VKYKWN1d6c4YKZKTN2d6a4\\K[KPN4d6a4^KYKnM6e6_4`KYKkM8f6]4bKYKhM:f6\\4eKXKeM<g6Z4nKeKT4Y4mKgKS4X4nKhKS4V4nKjKX4o3jKPL[4j3PLlKV4n3d3I6J7I6J7I7I6J7I6J7I6K6I7I6J7I6J7I7I6J\\am1"}, "label": "a man with curly hair and glasses posing with a tennis racket"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white and red shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a white and red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.43820312499999997, 0.13483333333333333, 0.7938124999999999, 0.9865208333333333], "iscrowd": 0, "area": 63059.00755, "rle": {"size": [480, 640], "counts": "]bS42a>c0]Oc0]Od0\\Oc0]Oc0]Oc0]Oc0B?H7H8H8I8G8H9H9F9G:G8G:F:G4K2N2O1N2N2N2O1N2N20000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O1_MSGoNm8P1VGmNk8R1YGkNg8T1\\GjNd8U1`GgNa8X1cGeN]8Z1fGcN[8\\1iGaNW8^1lG`NT8_1PH]NQ8b1RH\\Nn7c1VHZNj7e1YHXNh7g1\\HVNd7i1`HSNa7k1cHSN]7l1gHQNY7n1jHoMW7P2mHmMS7R2QIjMP7U2VIfMj6Y2\\IbMd6]2bI]M_6b2gIYMY6f2Q3O1O1O1O1O1O1O1O1O1O1O1O10O1000O0100O100O100O100O100O100O010O100O100O10000O101N101N101N2O0O2O0O2O0O2O0O2O0TGRMe5?\\Gk1l2iMg5;^Gl1i2lMh57`Gn1e2mMj55bGo1b2nMk52dGP2_2QNk5OfGR2\\2QNn5LgGS2Y2TNn5IjGT2V2TNP6GkGU2S2WNP6CoGW2n1YNR6_OQHX2k1[NS6]OSHY2g1]NU6YOUHZ2e1_NT6WOXH[2^1cNZ6QOYH\\2W1jN^6jN\\H]2P1oNc6cN^H^2i0UOh6]N`H_2b0ZOm6WNaH`2;@R7QNcH_26EW7lMcH`2OK]7fMcH_2K0a7bMdH_2D6g7[MeHh5Y7YJgHg5W7oI[H4>n5U7nIaH1:R6R7nIgHN7T6P7nIkHN5U6n6mInHN4V6k6mIRIN2U6j6nIUIM1V6h6mIXIM0W6e6mI\\ILOW6c6nI_IKNX6a6mIbIKMX6_6nIeIJLY6\\6nIiIIKZ6Z6mIlIIJZ6X6nIoIIHZ6V6nISJHG[6T6lIWJIE[6R6mIZJHD\\6o5mI^JGC]6k5nIcJEB]6h5PJgJCA^6e5PJkJB@_6a5RJoJ_O@_6_5VJoJ\\OA_6\\5ZJQKWOC`6Y5]JRKSOE`6V5bJSKnNGa6S5eJTKjNIa6P5jJUKeNKb6m4mJVKaNMc6i4QKXK\\NOc6f4VKYKWN1d6c4YKZKTN2d6a4\\K[KPN4d6a4^KYKnM6e6_4`KYKkM8f6]4bKYKhM:f6\\4eKXKeM<g6Z4nKeKT4Y4mKgKS4X4nKhKS4V4nKjKX4o3jKPL[4j3PLlKV4n3d3I6J7I6J7I7I6J7I6J7I6K6I7I6J7I6J7I7I6J\\am1"}, "label": "a man wearing a white and red shirt"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis racket with the black rim that the man in white shirt is holding\"."}], "task": "refering", "answer_template": "The \"the tennis racket with the black rim that the man in white shirt is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.5851875, 0.6788541666666668, 0.9883124999999999, 1.0], "iscrowd": 0, "area": 19901.657050000005, "rle": {"size": [480, 640], "counts": "mW`53j>300O100O100O100O100O100O100O100O100O100O100O100O10O10O10O010O0100O010O010O00100O010O0010O01O010O00100O0010O01O010O001O01J51O010O001O001O1O0O2O1O1O1O1O1N3N2M3N2M3N1N2O2M2N2N1O101N2N2M4M2N2M3N3M2N3M3N1N2N2N2O1O1O0O011O1O1N2O1O1O1N2O1O100O1O1O1O1O1O1O1O1O100O2N1O1O1O100O2O0O100O100O2O0O100O100O2O0O100O100O100O10000O100O10000O10O010000O100O10000O100O10000000000000O100O10000O100O100O10000O100O10000O100O100O101O0O100O101O0O100O101N1O2O1N101N2O0O2O0O2O1N101N2O0O2N2N1O2N1O2N2N1O2N2N3N1N3M2N2N3M2N3M2N2L5K4L5J5L4L=Cd0[Okl2"}, "label": "the tennis racket with the black rim that the man in white shirt is holding"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis racket in the hand of the man with the white shirt\"."}], "task": "refering", "answer_template": "The \"a tennis racket in the hand of the man with the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.5851875, 0.6788541666666668, 0.9883124999999999, 1.0], "iscrowd": 0, "area": 19901.657050000005, "rle": {"size": [480, 640], "counts": "mW`53j>300O100O100O100O100O100O100O100O100O100O100O100O10O10O10O010O0100O010O010O00100O010O0010O01O010O00100O0010O01O010O001O01J51O010O001O001O1O0O2O1O1O1O1O1N3N2M3N2M3N1N2O2M2N2N1O101N2N2M4M2N2M3N3M2N3M3N1N2N2N2O1O1O0O011O1O1N2O1O1O1N2O1O100O1O1O1O1O1O1O1O1O100O2N1O1O1O100O2O0O100O100O2O0O100O100O2O0O100O100O100O10000O100O10000O10O010000O100O10000O100O10000000000000O100O10000O100O100O10000O100O10000O100O100O101O0O100O101O0O100O101N1O2O1N101N2O0O2O0O2O1N101N2O0O2N2N1O2N1O2N2N1O2N2N3N1N3M2N2N3M2N3M2N2L5K4L5J5L4L=Cd0[Okl2"}, "label": "a tennis racket in the hand of the man with the white shirt"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis racket held by the hand with the man wearing orange\"."}], "task": "refering", "answer_template": "The \"the tennis racket held by the hand with the man wearing orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 325, 326, 327, 328, 329, 330, 331, 345, 351, 352], "bbox": [0.0050625, 0.6337083333333333, 0.45337500000000003, 0.896625], "iscrowd": 0, "area": 17353.7853, "rle": {"size": [480, 640], "counts": "Xj13e>8H8O0100000O0100000O10O1000O10O1000O10O1000O10O1000O10O1000O1000O10O1000O10O1000O3N2N2N1N3N2N2N1O2MUZa01ke^O1N2N2N2O2M2N2N2O1N1O2N1O2O0O2N1O2O0O2N1O2O0O2N1O2O0O2N1O1O2O0O2N1O2O0O2N1O2M2O2N1O2N1O2N1O2N1O1N3N1O1O1O1O2N1O1O1O2M2O1O1O1O2N1O1O1O1O2O0O100O101N10000O101N100O100O2O0O10000O2O0O100O2O0O10000O2O0O100O101N100O10001N100O100O2O01O000000001N1000001N10000O2O00000O101O0O10001O0O10000O2O00000O2O000O101O000O10001N1O100O2O0O1O100O2N100O101N1O100O2N100O102M2N3N1N3M2O2M2O2M2N3N2M2N3N1K6J5K6J5K6J5K6J5L5JQgS5"}, "label": "the tennis racket held by the hand with the man wearing orange"}]}
{"id": 177915, "image": "COCO_train2014_000000177915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver tennis racket\"."}], "task": "refering", "answer_template": "The \"a silver tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 325, 326, 327, 328, 329, 330, 331, 345, 351, 352], "bbox": [0.0050625, 0.6337083333333333, 0.45337500000000003, 0.896625], "iscrowd": 0, "area": 17353.7853, "rle": {"size": [480, 640], "counts": "Xj13e>8H8O0100000O0100000O10O1000O10O1000O10O1000O10O1000O10O1000O1000O10O1000O10O1000O3N2N2N1N3N2N2N1O2MUZa01ke^O1N2N2N2O2M2N2N2O1N1O2N1O2O0O2N1O2O0O2N1O2O0O2N1O2O0O2N1O1O2O0O2N1O2O0O2N1O2M2O2N1O2N1O2N1O2N1O1N3N1O1O1O1O2N1O1O1O2M2O1O1O1O2N1O1O1O1O2O0O100O101N10000O101N100O100O2O0O10000O2O0O100O2O0O10000O2O0O100O101N100O10001N100O100O2O01O000000001N1000001N10000O2O00000O101O0O10001O0O10000O2O00000O2O000O101O000O10001N1O100O2O0O1O100O2N100O101N1O100O2N100O102M2N3N1N3M2O2M2O2M2N3N2M2N3N1K6J5K6J5K6J5K6J5L5JQgS5"}, "label": "a silver tennis racket"}]}
{"id": 153340, "image": "COCO_train2014_000000153340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey color bike parked between 2 bikes\"."}], "task": "refering", "answer_template": "The \"grey color bike parked between 2 bikes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312], "bbox": [0.40190625, 0.2255625, 0.654140625, 0.8107708333333332], "iscrowd": 0, "area": 31322.808000000008, "rle": {"size": [480, 640], "counts": "Qfh3g0V>4K5K5L3L5K5L4L3N3M3N2[OPNVDR2i;QNSDR2l;QNoCR2o;QNmCR2P;eMmEj2n9^MmEc2P:dMlE]2P:kMkEV2P:oMnET2m9QNQFP2l9SNRFo1m9RNRFP2k9RNTFo1j9TNSFo1j9SNUFn1i9TNUFn1j9TNTFn1i9TNVFm1h9g1N3N1O2N1O1O2N1N2O2N1O2N1O1N3N1O1O2RGdJU8W6L3M3M3N2M3M3N1N3M3M2O2N1O2N1O1O0000000O1000001O00000006J00O1O1O2N000O1000000000000000000O10000000000000000O100000O1000000O1000O10O1000000O1000001N101O000O2O000O2PLjH3W7MjH0X70jHMX71kHKW75jHIX76jHFX79kHDV7<kHAX7>jH_OW7`0lH\\OV7d0kHYOX7f0jHUOY7k0iHoN\\7o0iHjNZ7V1jHcNZ7\\1jH]NZ7a1kHWNZ7h1jHQNZ7n1jHkMZ7S2kH]Mb7b2bHnLk7Q3T2M5K5J6I6K6J6I6K6I7J5KX]W3"}, "label": "grey color bike parked between 2 bikes"}]}
{"id": 153340, "image": "COCO_train2014_000000153340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray motorcycle between two other motorcycles\"."}], "task": "refering", "answer_template": "The \"a gray motorcycle between two other motorcycles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312], "bbox": [0.40190625, 0.2255625, 0.654140625, 0.8107708333333332], "iscrowd": 0, "area": 31322.808000000008, "rle": {"size": [480, 640], "counts": "Qfh3g0V>4K5K5L3L5K5L4L3N3M3N2[OPNVDR2i;QNSDR2l;QNoCR2o;QNmCR2P;eMmEj2n9^MmEc2P:dMlE]2P:kMkEV2P:oMnET2m9QNQFP2l9SNRFo1m9RNRFP2k9RNTFo1j9TNSFo1j9SNUFn1i9TNUFn1j9TNTFn1i9TNVFm1h9g1N3N1O2N1O1O2N1N2O2N1O2N1O1N3N1O1O2RGdJU8W6L3M3M3N2M3M3N1N3M3M2O2N1O2N1O1O0000000O1000001O00000006J00O1O1O2N000O1000000000000000000O10000000000000000O100000O1000000O1000O10O1000000O1000001N101O000O2O000O2PLjH3W7MjH0X70jHMX71kHKW75jHIX76jHFX79kHDV7<kHAX7>jH_OW7`0lH\\OV7d0kHYOX7f0jHUOY7k0iHoN\\7o0iHjNZ7V1jHcNZ7\\1jH]NZ7a1kHWNZ7h1jHQNZ7n1jHkMZ7S2kH]Mb7b2bHnLk7Q3T2M5K5J6I6K6J6I6K6I7J5KX]W3"}, "label": "a gray motorcycle between two other motorcycles"}]}
{"id": 153340, "image": "COCO_train2014_000000153340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle what is painted red\"."}], "task": "refering", "answer_template": "The \"a motorcycle what is painted red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 94, 95, 99, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 233, 234, 235, 255, 256, 257, 258, 259, 278, 279, 280, 281, 301, 302, 303, 304, 325, 326, 327, 349], "bbox": [0.041953125, 0.17247916666666668, 0.34395312499999997, 0.8995416666666667], "iscrowd": 0, "area": 25575.252800000013, "rle": {"size": [480, 640], "counts": "ki<1m>3N2M2UNIlD9R;KlD4S;0jD0nNHd;<[EMmNJg;=YEIPOJg;`0VEFSOJf;b0VEDSOJg;d0UEBTOJf;f0UE@UOJe;h0VE]OUOKe;i0UE]OUOJf;j0TE\\OVOIg;l0RE[OVOJh;k0RE[OVOJh;k0RE[OVOJi;j0QE\\OVOIj;l0PE[OUOIk;l0PE[OTOJl;k0PE[OTOJl;k0PE[OTOJl;k0PE[OTOIm;l0oD[OTOIm;l0oD[OTOIn;l0mD\\OSOIP<k0nD[OROJP<k0nD[OROIQ<l0mD[OROIQ<l0mD[OROIR<l0kD[OROJT<k0iD\\OROIV<k0gD\\OSOHX<l0dD\\OTOHY<l0bD\\OUOHZ<a0bCFn01UOI\\<:gCKg03SOK`<0mC2>=e;[OSD86>]<BaC?`<A^C`0c<@]C?g1YO_88iE`0OXOS11U96jE`0NZOR11U96kE>M^OQ1OW95jE>N_OP1OX94jE=NAo0OY93jE=LCP1M[93hE=MDo0L]93gE<MEo0LQ1Mi56ZH;LHo0JmNKe1S1\\5TOWI<LHo0JmNLa1[1[5kN\\I;LJb0ITO06M\\1d1[5aNaI:KLb05[OAX1m1Y5WNdI<LLc04]O@S1V2X5nMfI>LNd01_O@o0^2W5eMhI`0LOe00A_Ok0c2Y5_MgIb0L1f0MC^OJ6=_2m5\\MdIf0L1h0LC=Jf1Z6ZMbIh1e0lNDb3T6jMUJgNCd3X6dMnIoNFc3j6nLYIc4e6m101O1N101O0O0KhGcIW8]6kGbIU8]6lGbIT8^6nGaIQ8^6:O001O0O1100O010O0001N1O1O2N1O2N100O2N0010O10O0100O1O100O010O100O100O100O2M2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N3nMdH\\L_7c3fHUL^7i2^HPLL3>l0\\7i2cHVLF1e0d0U7Q3fHXLCNS17g6_3hH\\L_OKa9e3UG_LW9]3lFdLS9[3nFeLR9j1mE]OR1iNP9h1QF^OQ1jNl8e1WF@n0kNj8c1[FAl0mNg8`1`FWOITOQ15e8]1fFVOHVOn07d8j0WF[Oe07C]On08a81VF=7Ak02YOGo07a80VF?8@m0NUOLP15a80XF`05Am0NUOMP13`81ZF`03Cn0KUONP12a81ZFa02Cn0KUONP12`80]Fb00Dn0IVONo02a8O_Fd0MCn0Kc0NS8OaFe0KDn0Id0NR8OcF<CI65n0Id0MR81eF9AK66n0Gd0NQ82gF8_OK76n0Ga:OlD9_OJ67h<GcC8_OK57g<GfC6^OL66f<HfC7^OJ66f<JgC5]OK66f<IhC6\\OK66e<JiC5]OJ65d<MjC4[OK73d<NjC4\\OJ82a<1lC2[OL9N`<5lC2[OK9M`<5nC2YOLW=3`C1ZOLP=8fCK[ONl<9jCG\\OOi<:lCF]O0d<<PDB]O3`<=SD_O_O3]<>dDCY;?kD^OS;c0e1O10000O1000O010000O1O2M4K5L3MjgT6"}, "label": "a motorcycle what is painted red"}]}
{"id": 153340, "image": "COCO_train2014_000000153340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and black classic motorcycle\"."}], "task": "refering", "answer_template": "The \"a red and black classic motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 94, 95, 99, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 233, 234, 235, 255, 256, 257, 258, 259, 278, 279, 280, 281, 301, 302, 303, 304, 325, 326, 327, 349], "bbox": [0.041953125, 0.17247916666666668, 0.34395312499999997, 0.8995416666666667], "iscrowd": 0, "area": 25575.252800000013, "rle": {"size": [480, 640], "counts": "ki<1m>3N2M2UNIlD9R;KlD4S;0jD0nNHd;<[EMmNJg;=YEIPOJg;`0VEFSOJf;b0VEDSOJg;d0UEBTOJf;f0UE@UOJe;h0VE]OUOKe;i0UE]OUOJf;j0TE\\OVOIg;l0RE[OVOJh;k0RE[OVOJh;k0RE[OVOJi;j0QE\\OVOIj;l0PE[OUOIk;l0PE[OTOJl;k0PE[OTOJl;k0PE[OTOJl;k0PE[OTOIm;l0oD[OTOIm;l0oD[OTOIn;l0mD\\OSOIP<k0nD[OROJP<k0nD[OROIQ<l0mD[OROIQ<l0mD[OROIR<l0kD[OROJT<k0iD\\OROIV<k0gD\\OSOHX<l0dD\\OTOHY<l0bD\\OUOHZ<a0bCFn01UOI\\<:gCKg03SOK`<0mC2>=e;[OSD86>]<BaC?`<A^C`0c<@]C?g1YO_88iE`0OXOS11U96jE`0NZOR11U96kE>M^OQ1OW95jE>N_OP1OX94jE=NAo0OY93jE=LCP1M[93hE=MDo0L]93gE<MEo0LQ1Mi56ZH;LHo0JmNKe1S1\\5TOWI<LHo0JmNLa1[1[5kN\\I;LJb0ITO06M\\1d1[5aNaI:KLb05[OAX1m1Y5WNdI<LLc04]O@S1V2X5nMfI>LNd01_O@o0^2W5eMhI`0LOe00A_Ok0c2Y5_MgIb0L1f0MC^OJ6=_2m5\\MdIf0L1h0LC=Jf1Z6ZMbIh1e0lNDb3T6jMUJgNCd3X6dMnIoNFc3j6nLYIc4e6m101O1N101O0O0KhGcIW8]6kGbIU8]6lGbIT8^6nGaIQ8^6:O001O0O1100O010O0001N1O1O2N1O2N100O2N0010O10O0100O1O100O010O100O100O100O2M2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N3nMdH\\L_7c3fHUL^7i2^HPLL3>l0\\7i2cHVLF1e0d0U7Q3fHXLCNS17g6_3hH\\L_OKa9e3UG_LW9]3lFdLS9[3nFeLR9j1mE]OR1iNP9h1QF^OQ1jNl8e1WF@n0kNj8c1[FAl0mNg8`1`FWOITOQ15e8]1fFVOHVOn07d8j0WF[Oe07C]On08a81VF=7Ak02YOGo07a80VF?8@m0NUOLP15a80XF`05Am0NUOMP13`81ZF`03Cn0KUONP12a81ZFa02Cn0KUONP12`80]Fb00Dn0IVONo02a8O_Fd0MCn0Kc0NS8OaFe0KDn0Id0NR8OcF<CI65n0Id0MR81eF9AK66n0Gd0NQ82gF8_OK76n0Ga:OlD9_OJ67h<GcC8_OK57g<GfC6^OL66f<HfC7^OJ66f<JgC5]OK66f<IhC6\\OK66e<JiC5]OJ65d<MjC4[OK73d<NjC4\\OJ82a<1lC2[OL9N`<5lC2[OK9M`<5nC2YOLW=3`C1ZOLP=8fCK[ONl<9jCG\\OOi<:lCF]O0d<<PDB]O3`<=SD_O_O3]<>dDCY;?kD^OS;c0e1O10000O1000O010000O1O2M4K5L3MjgT6"}, "label": "a red and black classic motorcycle"}]}
{"id": 153340, "image": "COCO_train2014_000000153340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black motorcycle with low handlebars\"."}], "task": "refering", "answer_template": "The \"black motorcycle with low handlebars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 317, 318], "bbox": [0.6505625, 0.2674166666666667, 0.9455, 0.7887708333333333], "iscrowd": 0, "area": 31115.78139999998, "rle": {"size": [480, 640], "counts": "UUS6o2Q<01O01O01O00001O01O000BWMfDh2m:fMREZ2a:SN_En1`:SN`El1`:UN_Ek1a:VN^Ej1b:WN^Eh1b:YN]Eh1b:XN^Eh1b:YN^Ef1b:[N]Ee1c:\\N\\Ed1d:]N\\Eb1d:_N[Eb1d:^N\\Eb1d:_N[Ea1e:`N[E_1e:bNZE^1f:_1O01O00001O01O0001O00010N10001O000O101O00001N10001O000O2O00000O2O00001O0O101OXGWLY6i3bI]L]6c3^IcLa6\\3[IjLe6U3VIRMh6n2RIYMm6g2nH_MR7`2iHfMV7Z2eHlMZ7S2bHTN]7k1_HZN`7f1]H^Nb7b1[HbNd7^1YHfNf7Y1YHkNe7U1XHoNf7R1WHROh7n0VHUOi7k0THYOj7h0TH\\Oj7c0UH@i7a0UHBj7>THEk7;THGj7i3O10O00010O0010O0010O00010O01O01O010O01O01O01O01O001O0O101N100O2O0O2O0O10O0100O10O0100O10O0100O10O0100O10O0100O10OYLjGe0W8[OiGe0W8ZOjGe0W8ZOjGf0V8YOkGf0V8YOkGf0V8YOlGf0U8YOkGf0V8YOkGg0U8XOlGg0U8XOmGg0S8XOoGf0S8XOoGf0R8ZOoGd0R8[OPHc0Q8\\OQHa0R8]OPHa0Q8^OQH`0P8_ORH?o7ARH<Q8BQH<P8ATH<n7@WH>j7_O\\H<g7_OaH<`7AgH9[7CmH8T7ESI5P7IUI2l6LZINh61_IFh64V4L4K5L5J5L4K6K4K\\W`0"}, "label": "black motorcycle with low handlebars"}]}
{"id": 153340, "image": "COCO_train2014_000000153340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike rightside\"."}], "task": "refering", "answer_template": "The \"bike rightside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296, 317, 318], "bbox": [0.6505625, 0.2674166666666667, 0.9455, 0.7887708333333333], "iscrowd": 0, "area": 31115.78139999998, "rle": {"size": [480, 640], "counts": "UUS6o2Q<01O01O01O00001O01O000BWMfDh2m:fMREZ2a:SN_En1`:SN`El1`:UN_Ek1a:VN^Ej1b:WN^Eh1b:YN]Eh1b:XN^Eh1b:YN^Ef1b:[N]Ee1c:\\N\\Ed1d:]N\\Eb1d:_N[Eb1d:^N\\Eb1d:_N[Ea1e:`N[E_1e:bNZE^1f:_1O01O00001O01O0001O00010N10001O000O101O00001N10001O000O2O00000O2O00001O0O101OXGWLY6i3bI]L]6c3^IcLa6\\3[IjLe6U3VIRMh6n2RIYMm6g2nH_MR7`2iHfMV7Z2eHlMZ7S2bHTN]7k1_HZN`7f1]H^Nb7b1[HbNd7^1YHfNf7Y1YHkNe7U1XHoNf7R1WHROh7n0VHUOi7k0THYOj7h0TH\\Oj7c0UH@i7a0UHBj7>THEk7;THGj7i3O10O00010O0010O0010O00010O01O01O010O01O01O01O01O001O0O101N100O2O0O2O0O10O0100O10O0100O10O0100O10O0100O10O0100O10OYLjGe0W8[OiGe0W8ZOjGe0W8ZOjGf0V8YOkGf0V8YOkGf0V8YOlGf0U8YOkGf0V8YOkGg0U8XOlGg0U8XOmGg0S8XOoGf0S8XOoGf0R8ZOoGd0R8[OPHc0Q8\\OQHa0R8]OPHa0Q8^OQH`0P8_ORH?o7ARH<Q8BQH<P8ATH<n7@WH>j7_O\\H<g7_OaH<`7AgH9[7CmH8T7ESI5P7IUI2l6LZINh61_IFh64V4L4K5L5J5L4K6K4K\\W`0"}, "label": "bike rightside"}]}
{"id": 169725, "image": "COCO_train2014_000000169725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe being fed by person\"."}], "task": "refering", "answer_template": "The \"giraffe being fed by person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 26, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 194, 195, 196, 218, 219], "bbox": [0.002984375, 0.0, 0.557984375, 0.6251053864168619], "iscrowd": 0, "area": 38589.37945, "rle": {"size": [427, 640], "counts": "fj0h2c:f1ZN01O000000010O00000dKSGZ4m8fKSGZ4P90O01O001O001O010O001O001O010O001O001O0010O01O001O1O0010O01O001O0010O01O001O001O010O001O001O010O001O001O0010O01O001O1O010O001O001O0010O01O001O0010O01O001O001O010O0010O010O010O010O010O0100O010O010O010O010O1O10O0100O10O0100O010O01M2O2M2N3M2N3N110O001O001O010O001O001O010O001O0010O010O10O01000O01000O01000O010000O010000O01000O01000O0100O010O01O00001O001O00001O001O00001O0001O010O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O010O1O1O1O01O00000000001O000000001O00000000001O00000O1010O1O1O001O100O1O1O100O10O0100O100O100O10O010000O101N101N2O0O2O1N2O0O2O1N2O0O2O1O1N101O1O1O001O1O0O2O1O1O001O1O00000O101O00000000000O10001O00000O100O1O100O2N1O100O1O100O1O101N1O100O1O1N2O1N2O2N1N2O1N2O3K4K6K4L5K4L5K4Kjhe3"}, "label": "giraffe being fed by person"}]}
{"id": 169725, "image": "COCO_train2014_000000169725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eats some food through a fence\"."}], "task": "refering", "answer_template": "The \"a giraffe eats some food through a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 26, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 194, 195, 196, 218, 219], "bbox": [0.002984375, 0.0, 0.557984375, 0.6251053864168619], "iscrowd": 0, "area": 38589.37945, "rle": {"size": [427, 640], "counts": "fj0h2c:f1ZN01O000000010O00000dKSGZ4m8fKSGZ4P90O01O001O001O010O001O001O010O001O001O0010O01O001O1O0010O01O001O0010O01O001O001O010O001O001O010O001O001O0010O01O001O1O010O001O001O0010O01O001O0010O01O001O001O010O0010O010O010O010O010O0100O010O010O010O010O1O10O0100O10O0100O010O01M2O2M2N3M2N3N110O001O001O010O001O001O010O001O0010O010O10O01000O01000O01000O010000O010000O01000O01000O0100O010O01O00001O001O00001O001O00001O0001O010O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O010O1O1O1O01O00000000001O000000001O00000000001O00000O1010O1O1O001O100O1O1O100O10O0100O100O100O10O010000O101N101N2O0O2O1N2O0O2O1N2O0O2O1O1N101O1O1O001O1O0O2O1O1O001O1O00000O101O00000000000O10001O00000O100O1O100O2N1O100O1O100O1O101N1O100O1O1N2O1N2O2N1N2O1N2O3K4K6K4L5K4L5K4Kjhe3"}, "label": "a giraffe eats some food through a fence"}]}
{"id": 169725, "image": "COCO_train2014_000000169725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe looking down and not sticking head through fence\"."}], "task": "refering", "answer_template": "The \"giraffe looking down and not sticking head through fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 32, 33, 34, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 105, 117, 118, 119, 123, 124, 125, 126, 127, 128, 140, 141, 142, 143, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 193, 194, 195, 196, 197, 210, 211, 212, 216, 217, 218, 219, 233, 234, 235, 239, 240, 241, 242, 256, 257, 262, 263, 264, 265, 279, 285, 286, 287, 288, 308, 310, 311, 312, 330, 331, 334], "bbox": [0.11492187499999999, 0.047728337236533955, 0.583609375, 1.0], "iscrowd": 0, "area": 46573.26914999999, "rle": {"size": [427, 640], "counts": "gRo06g:h0dF^OZ9o0TFYOj9V2M3N101N1[HRMl4o2lJXMS5i2fJ^MZ5c2^JdMa5]2RJPNm5]2UIUNk6V4O10O001O3M2M3N2NF:1O1O1N2O1O010O100O10000O100O101M2M0000O1011N3M2N3M2L5J6K4K6L3O2N1O2N2N1O2N1O2N2N1N3N1O2N1O2M3M2O2M2N3M3M2N3M2O2M2N3M3M2O2N1O2N1O2N2N1N3N1O2N2N1O2N1O2M2M4M3L3M4L3M4M3F9F;FVm>2QSA4K4L4M4K4L4M4K4L5L3L4L5L3L4L5L3L4L5L3L5K4M3L5K4M3L5K4M3\\MaLUKd3c4`L\\Kd3[4cLdKa3f3QMYLS3`3QM_LS3Z3QMfLS3R3RMlLR3m2RMRMR3X2aMhMd2W1YNgNn1l0XNTOo1`0UNBQ22TNNS2GPN:V2\\OjMi0\\2lNbMY1R76J6J6I7I6K61O001N101O001O0O2O1O001NK6I7H8I7G9H7H9H8F:]Oc0_Oa02N1O2N2N1O2N1O2O11N2O2N1N2O1O1N2O1N25K6J5K5K2N2N2N1O2TM_HGd71`H1a7IbH8a7@cHb0_7VOeHk0]7nNgHT1[7YMTJh2W82N1O2N2N1O2N1O2N1O1N2J6I7F:J69G=\\LgF[2U:0O2N1O1O1O1O1O1iKWMkMi2o1eMhM\\2R2RNQKVOa1i2Y3XOgLi0S3]OlLd0n2BQM`0i2EVM<f2HYM9c2KhLjLVO[3n3OgLoLVOS3o32fLTMVOk2P44gLXMUOe2P47jLj0R3ZOmLg0P3\\OoLe0l2@SMa0i2CVMa0c2C]Ma0Z2HaM=W2N^M:Z20\\M8[2g5D<C=DZh^3"}, "label": "giraffe looking down and not sticking head through fence"}]}
{"id": 169725, "image": "COCO_train2014_000000169725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe in the back\"."}], "task": "refering", "answer_template": "The \"the giraffe in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 32, 33, 34, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 105, 117, 118, 119, 123, 124, 125, 126, 127, 128, 140, 141, 142, 143, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 193, 194, 195, 196, 197, 210, 211, 212, 216, 217, 218, 219, 233, 234, 235, 239, 240, 241, 242, 256, 257, 262, 263, 264, 265, 279, 285, 286, 287, 288, 308, 310, 311, 312, 330, 331, 334], "bbox": [0.11492187499999999, 0.047728337236533955, 0.583609375, 1.0], "iscrowd": 0, "area": 46573.26914999999, "rle": {"size": [427, 640], "counts": "gRo06g:h0dF^OZ9o0TFYOj9V2M3N101N1[HRMl4o2lJXMS5i2fJ^MZ5c2^JdMa5]2RJPNm5]2UIUNk6V4O10O001O3M2M3N2NF:1O1O1N2O1O010O100O10000O100O101M2M0000O1011N3M2N3M2L5J6K4K6L3O2N1O2N2N1O2N1O2N2N1N3N1O2N1O2M3M2O2M2N3M3M2N3M2O2M2N3M3M2O2N1O2N1O2N2N1N3N1O2N2N1O2N1O2M2M4M3L3M4L3M4M3F9F;FVm>2QSA4K4L4M4K4L4M4K4L5L3L4L5L3L4L5L3L4L5L3L5K4M3L5K4M3L5K4M3\\MaLUKd3c4`L\\Kd3[4cLdKa3f3QMYLS3`3QM_LS3Z3QMfLS3R3RMlLR3m2RMRMR3X2aMhMd2W1YNgNn1l0XNTOo1`0UNBQ22TNNS2GPN:V2\\OjMi0\\2lNbMY1R76J6J6I7I6K61O001N101O001O0O2O1O001NK6I7H8I7G9H7H9H8F:]Oc0_Oa02N1O2N2N1O2N1O2O11N2O2N1N2O1O1N2O1N25K6J5K5K2N2N2N1O2TM_HGd71`H1a7IbH8a7@cHb0_7VOeHk0]7nNgHT1[7YMTJh2W82N1O2N2N1O2N1O2N1O1N2J6I7F:J69G=\\LgF[2U:0O2N1O1O1O1O1O1iKWMkMi2o1eMhM\\2R2RNQKVOa1i2Y3XOgLi0S3]OlLd0n2BQM`0i2EVM<f2HYM9c2KhLjLVO[3n3OgLoLVOS3o32fLTMVOk2P44gLXMUOe2P47jLj0R3ZOmLg0P3\\OoLe0l2@SMa0i2CVMa0c2C]Ma0Z2HaM=W2N^M:Z20\\M8[2g5D<C=DZh^3"}, "label": "the giraffe in the back"}]}
{"id": 390908, "image": "COCO_train2014_000000390908.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller zebra standing on the grass\"."}], "task": "refering", "answer_template": "The \"the taller zebra standing on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330, 331, 351, 352, 353, 354], "bbox": [0.258296875, 0.41179166666666667, 0.439046875, 0.9382291666666667], "iscrowd": 0, "area": 17776.222999999998, "rle": {"size": [480, 640], "counts": "id]21h>:F:G9K5K5L4K5L4K5K6K8G8I7`DaMQ:g2cEbMY:f2ZEcMa:a3L5K5aF]KZ8g4QGkK\\7]OjIh5T6[JTI0Ig5Q7]JRI0Ie5S7^JPI3Gb5W7^JoHY6P7jIlHW6S7lIjHU6U7nIhHT6V7k0O10hN_ISJa6a5e1D<H84L3M3M5K4L4L3M3M5K8H8H4L4L4L4L2N000000000000000000SOhHRJX7n5PIjIP7W6VIbIj6_6ZI\\Ig6j6e07I7I2NO2]O\\HeIf7U6hHaIZ7Z6i0J6kNfG^K_8a4R1M2O1O1N2O2N1N2mLhE\\1Y:aNkE]1V:`NmE_1U:]NnEa1X:XNlEf1\\:PNgEo1a:gMcEX2d:^M_Ef2];2M2N2O0O2O0O<E0O100O2N102M7J6JO1N2ZMUER1o:]NiEY1Y:ZN]F[1n;E9H4L4K5L4L;DfkW5"}, "label": "the taller zebra standing on the grass"}]}
{"id": 390908, "image": "COCO_train2014_000000390908.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one of the zebras has its backside to the camera\"."}], "task": "refering", "answer_template": "The \"one of the zebras has its backside to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330, 331, 351, 352, 353, 354], "bbox": [0.258296875, 0.41179166666666667, 0.439046875, 0.9382291666666667], "iscrowd": 0, "area": 17776.222999999998, "rle": {"size": [480, 640], "counts": "id]21h>:F:G9K5K5L4K5L4K5K6K8G8I7`DaMQ:g2cEbMY:f2ZEcMa:a3L5K5aF]KZ8g4QGkK\\7]OjIh5T6[JTI0Ig5Q7]JRI0Ie5S7^JPI3Gb5W7^JoHY6P7jIlHW6S7lIjHU6U7nIhHT6V7k0O10hN_ISJa6a5e1D<H84L3M3M5K4L4L3M3M5K8H8H4L4L4L4L2N000000000000000000SOhHRJX7n5PIjIP7W6VIbIj6_6ZI\\Ig6j6e07I7I2NO2]O\\HeIf7U6hHaIZ7Z6i0J6kNfG^K_8a4R1M2O1O1N2O2N1N2mLhE\\1Y:aNkE]1V:`NmE_1U:]NnEa1X:XNlEf1\\:PNgEo1a:gMcEX2d:^M_Ef2];2M2N2O0O2O0O<E0O100O2N102M7J6JO1N2ZMUER1o:]NiEY1Y:ZN]F[1n;E9H4L4K5L4L;DfkW5"}, "label": "one of the zebras has its backside to the camera"}]}
{"id": 554756, "image": "COCO_train2014_000000554756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck carring two pink pass the pig pieces\"."}], "task": "refering", "answer_template": "The \"a white truck carring two pink pass the pig pieces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 75, 76, 77, 78, 79, 80, 83, 84, 85, 86, 87, 88, 89, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 285], "bbox": [0.13203125, 0.15652452025586353, 1.0, 0.7150959488272921], "iscrowd": 0, "area": 97347.03464999997, "rle": {"size": [469, 640], "counts": "oVW12`>4L3VNj1I7J7H7J6I7K5L5K4L4L4M4K4L4L4L5K4M3L4L5K4L4L4M3L4N201N100000000O101O000O100000001N1000000O10001O000O1000000O2O001O2N2M2O2N2N2N1N1000000O100000000O1000000O2O00000000000000000000000000001O00O1O1O100O1O100O1O1O100O1O100O100001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O001O1O1O1O1O1O1O1O1O1O00001O001O001O00001O001O000000O10000O1000O0100O10O10O10O10O100O01000O10O10O100O01000O10O10O100O01000O10O10O100O01000O1M2N3M3N2M4O1O1N101O1O001O1O1O002N1O1O1O1O1O1O1O1O1O2N1O1O1N2M3M3N2M3N2O00000000O10O1000O100000000O1000O1000O1000000O1000O1000O100000000O1000O1000O1000000O100000O10O1000001O1N2O1O1O2N1O100O2N000001O000000ZO[J]Hd5b7_J\\Ha5c7aJ\\H_5b7dJ]H]5a7eJ^H[5a7gJ_HX5`7jJ_HV5`7lJ_HT5`7nJ_HR5_7QKaHn4^7UK`Hk4_7WK`Hi4_7YK_Hh4a7YK^Hg4b7ZK\\Hg4d7ZKZHg4e7[KZHf4e7[KYHe4h7S11O100O001O1O100O1O001O100O1O1O010O1O100O10O10000000O10O10000000O10O10000000O1000O100000O1000O1000001O001N200O100O1O100O100O1O100O010O1O01O001O00010O001O00010O001O000001O0001O0001OO1O100O100O2O0O1O100O100O1O100O2O0O100O10000000001O0eLeG<[8AhG?X8]OmGb0S8[OPHe0P8XOSHh0m7TNlG\\O:`1k8ROWHm0k7QOVHo0j7POWHo0k7oNVHP1l7nNUHR1k7mNVHR1l7lNUHS1m7kNTHU1l7jNUHU1m7iNTHV1n7hNSHX1m7gNTHX1n7fNSHY1o7eNRHZ1o7eNRH[1o7bNSH]1n7bNSH]1o7`NSH`1n7^NSHa1n7]NUHa1m7]NTHc1m7ZNUHe1l7ZNUHe1^:00O1O1001O000000000000001O000000000001O01O000000000000001O00000000000000001O00000001O000001O000000000000001O00000000000000001O00L5H7H[J"}, "label": "a white truck carring two pink pass the pig pieces"}]}
{"id": 554756, "image": "COCO_train2014_000000554756.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck carrying inflated pigs\"."}], "task": "refering", "answer_template": "The \"the truck carrying inflated pigs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 75, 76, 77, 78, 79, 80, 83, 84, 85, 86, 87, 88, 89, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 285], "bbox": [0.13203125, 0.15652452025586353, 1.0, 0.7150959488272921], "iscrowd": 0, "area": 97347.03464999997, "rle": {"size": [469, 640], "counts": "oVW12`>4L3VNj1I7J7H7J6I7K5L5K4L4L4M4K4L4L4L5K4M3L4L5K4L4L4M3L4N201N100000000O101O000O100000001N1000000O10001O000O1000000O2O001O2N2M2O2N2N2N1N1000000O100000000O1000000O2O00000000000000000000000000001O00O1O1O100O1O100O1O1O100O1O100O100001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O001O1O1O1O1O1O1O1O1O1O00001O001O001O00001O001O000000O10000O1000O0100O10O10O10O10O100O01000O10O10O100O01000O10O10O100O01000O10O10O100O01000O1M2N3M3N2M4O1O1N101O1O001O1O1O002N1O1O1O1O1O1O1O1O1O2N1O1O1N2M3M3N2M3N2O00000000O10O1000O100000000O1000O1000O1000000O1000O1000O100000000O1000O1000O1000000O100000O10O1000001O1N2O1O1O2N1O100O2N000001O000000ZO[J]Hd5b7_J\\Ha5c7aJ\\H_5b7dJ]H]5a7eJ^H[5a7gJ_HX5`7jJ_HV5`7lJ_HT5`7nJ_HR5_7QKaHn4^7UK`Hk4_7WK`Hi4_7YK_Hh4a7YK^Hg4b7ZK\\Hg4d7ZKZHg4e7[KZHf4e7[KYHe4h7S11O100O001O1O100O1O001O100O1O1O010O1O100O10O10000000O10O10000000O10O10000000O1000O100000O1000O1000001O001N200O100O1O100O100O1O100O010O1O01O001O00010O001O00010O001O000001O0001O0001OO1O100O100O2O0O1O100O100O1O100O2O0O100O10000000001O0eLeG<[8AhG?X8]OmGb0S8[OPHe0P8XOSHh0m7TNlG\\O:`1k8ROWHm0k7QOVHo0j7POWHo0k7oNVHP1l7nNUHR1k7mNVHR1l7lNUHS1m7kNTHU1l7jNUHU1m7iNTHV1n7hNSHX1m7gNTHX1n7fNSHY1o7eNRHZ1o7eNRH[1o7bNSH]1n7bNSH]1o7`NSH`1n7^NSHa1n7]NUHa1m7]NTHc1m7ZNUHe1l7ZNUHe1^:00O1O1001O000000000000001O000000000001O01O000000000000001O00000000000000001O00000001O000001O000000000000001O00000000000000001O00L5H7H[J"}, "label": "the truck carrying inflated pigs"}]}
{"id": 87813, "image": "COCO_train2014_000000087813.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection in the window of a white chair\"."}], "task": "refering", "answer_template": "The \"a reflection in the window of a white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 76, 77, 78], "bbox": [0.2705, 0.03300254452926209, 0.495140625, 0.28534351145038167], "iscrowd": 0, "area": 7951.391899999998, "rle": {"size": [393, 640], "counts": "e]R22W<5J4M2N3M3M2N3L4M2N3M3M2M4M3M2N3M3M2M4M3M2N3M3M2M4M3M2N2N0O10000000000O2O000000000O100000000O10000000000O10000000000O10001O00000O1000000007I:F:F:F:F:F:F:F:F_e5LcZJ;G8H9F:G9G9G9G8H00O100O010O1O100O100O1000O010000O100O10000O010O10000O5L5K5J6K5M3M3L4K5K5L4K5K5K5KYlk3"}, "label": "a reflection in the window of a white chair"}]}
{"id": 87813, "image": "COCO_train2014_000000087813.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white patterned fence outside\"."}], "task": "refering", "answer_template": "The \"the white patterned fence outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 76, 77, 78], "bbox": [0.2705, 0.03300254452926209, 0.495140625, 0.28534351145038167], "iscrowd": 0, "area": 7951.391899999998, "rle": {"size": [393, 640], "counts": "e]R22W<5J4M2N3M3M2N3L4M2N3M3M2M4M3M2N3M3M2M4M3M2N3M3M2M4M3M2N2N0O10000000000O2O000000000O100000000O10000000000O10000000000O10001O00000O1000000007I:F:F:F:F:F:F:F:F_e5LcZJ;G8H9F:G9G9G9G8H00O100O010O1O100O100O1000O010000O100O10000O010O10000O5L5K5J6K5M3M3L4K5K5L4K5K5K5KYlk3"}, "label": "the white patterned fence outside"}]}
{"id": 456453, "image": "COCO_train2014_000000456453.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"side of a white truck\"."}], "task": "refering", "answer_template": "The \"side of a white truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 183], "bbox": [0.409265625, 0.004342723004694836, 0.9969374999999999, 0.48408450704225353], "iscrowd": 0, "area": 54401.28705, "rle": {"size": [426, 640], "counts": "bP]3Q1e;e0O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O01O0001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000aj0"}, "label": "side of a white truck"}]}
{"id": 456453, "image": "COCO_train2014_000000456453.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a white truck\"."}], "task": "refering", "answer_template": "The \"the side of a white truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 183], "bbox": [0.409265625, 0.004342723004694836, 0.9969374999999999, 0.48408450704225353], "iscrowd": 0, "area": 54401.28705, "rle": {"size": [426, 640], "counts": "bP]3Q1e;e0O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O01O0001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000aj0"}, "label": "the side of a white truck"}]}
{"id": 415499, "image": "COCO_train2014_000000415499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"soup in a white bowl with vegetables , nuts and broth\"."}], "task": "refering", "answer_template": "The \"soup in a white bowl with vegetables , nuts and broth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 376, 377, 378, 379, 380, 381], "bbox": [0.07571875, 0.08541666666666667, 0.8819218750000001, 0.9743125], "iscrowd": 0, "area": 154390.65545000005, "rle": {"size": [480, 640], "counts": "ogf0g1W=5K5K5K5J6K5K5K5K5K5J6K5K4L4L3L5L3M4L3M4L3L5L3M4L3M4M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M1O2N2N1O2N1N3N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O100O2O0000000O2O00000O2O00000O2O00000O101O00000O2O00000O2O00000O101O000O101O00000O2O00000O101O000O101O0\\O[EmLe:Q3^EnLb:Q3`EmLb:P3aEoL_:o2dEPM\\:n2gEQMY:m2jEQMW:n2kEQMV:l2nERMR:l2QFSMo9k2TFSMm9k2VFTMk9j2WFUMi9k2V100O2O00000O101O0O10001N10000O2O00000O101O0O10001N10000O2O0000000010O001O01O01O001O010O000010OO2O00001O001O00001N101O00001O001O000O2O001O00001O001O0O101O001O00001O001OjNTDXOk;k0XDPOh;R1[DjNd;X1_DcNa;_1bD]N];f1eDUN[;m1fDPNZ;R2fDmMY;U2gDiMY;Y2gDfMW;]2jD`MV;b2c02N2N2N2N2N1O0O10O1000000000000O0100000000000O1000O10000000O10000000O1000O100000000000O010000000000O1000O100O01000O100O01000O10O010000O10O0100O10O10O100O102N2M4M3L3N3M2M4M2M4M3M2M4M2M4M2N3L4M2M4M2N1N2O0O2O001N101N10001N101N101N101O0O2O0O101O0O2O0O2O001N100O2O001N101N101O0O101N101O0O2O0O2N1O2N1O1O2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N4L4J7H7I8H7I8H7I8H7I7I8H7I8H7I8H7I8H7Ie[S1"}, "label": "soup in a white bowl with vegetables , nuts and broth"}]}
{"id": 415499, "image": "COCO_train2014_000000415499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large white bowl of curry , with carrot and kale garnishes\"."}], "task": "refering", "answer_template": "The \"a large white bowl of curry , with carrot and kale garnishes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 376, 377, 378, 379, 380, 381], "bbox": [0.07571875, 0.08541666666666667, 0.8819218750000001, 0.9743125], "iscrowd": 0, "area": 154390.65545000005, "rle": {"size": [480, 640], "counts": "ogf0g1W=5K5K5K5J6K5K5K5K5K5J6K5K4L4L3L5L3M4L3M4L3L5L3M4L3M4M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M1O2N2N1O2N1N3N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2O1N1O2O1N101N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O0O2O1N1O100O2O0000000O2O00000O2O00000O2O00000O101O00000O2O00000O2O00000O101O000O101O00000O2O00000O101O000O101O0\\O[EmLe:Q3^EnLb:Q3`EmLb:P3aEoL_:o2dEPM\\:n2gEQMY:m2jEQMW:n2kEQMV:l2nERMR:l2QFSMo9k2TFSMm9k2VFTMk9j2WFUMi9k2V100O2O00000O101O0O10001N10000O2O00000O101O0O10001N10000O2O0000000010O001O01O01O001O010O000010OO2O00001O001O00001N101O00001O001O000O2O001O00001O001O0O101O001O00001O001OjNTDXOk;k0XDPOh;R1[DjNd;X1_DcNa;_1bD]N];f1eDUN[;m1fDPNZ;R2fDmMY;U2gDiMY;Y2gDfMW;]2jD`MV;b2c02N2N2N2N2N1O0O10O1000000000000O0100000000000O1000O10000000O10000000O1000O100000000000O010000000000O1000O100O01000O100O01000O10O010000O10O0100O10O10O100O102N2M4M3L3N3M2M4M2M4M3M2M4M2M4M2N3L4M2M4M2N1N2O0O2O001N101N10001N101N101N101O0O2O0O101O0O2O0O2O001N100O2O001N101N101O0O101N101O0O2O0O2N1O2N1O1O2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N4L4J7H7I8H7I8H7I8H7I7I8H7I8H7I8H7I8H7Ie[S1"}, "label": "a large white bowl of curry , with carrot and kale garnishes"}]}
{"id": 382742, "image": "COCO_train2014_000000382742.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a scatting board holded by a man\"."}], "task": "refering", "answer_template": "The \"a scatting board holded by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 205, 217, 218, 232, 247], "bbox": [0.4563529411764706, 0.516578125, 0.7481882352941177, 0.7015625], "iscrowd": 0, "area": 1251.5958000000019, "rle": {"size": [640, 425], "counts": "jei32mc03L3N3L2O1O1N2O1O1O1O1O1OM3N2O2O0O1000000001O3M4L100O1O2N00M3N2H80000010O000000000000000001O0000000000000000001O00001O6JTSb0Mol]O4L4L2N1O001O001O00O1000000O1000000O1000000O1000000O1000000O2O001OfaR2"}, "label": "a scatting board holded by a man"}]}
{"id": 382742, "image": "COCO_train2014_000000382742.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surf board that the man is holding\"."}], "task": "refering", "answer_template": "The \"the surf board that the man is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 205, 217, 218, 232, 247], "bbox": [0.4563529411764706, 0.516578125, 0.7481882352941177, 0.7015625], "iscrowd": 0, "area": 1251.5958000000019, "rle": {"size": [640, 425], "counts": "jei32mc03L3N3L2O1O1N2O1O1O1O1O1OM3N2O2O0O1000000001O3M4L100O1O2N00M3N2H80000010O000000000000000001O0000000000000000001O00001O6JTSb0Mol]O4L4L2N1O001O001O00O1000000O1000000O1000000O1000000O1000000O2O001OfaR2"}, "label": "the surf board that the man is holding"}]}
{"id": 505625, "image": "COCO_train2014_000000505625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"asparagus in the middle of the plate\"."}], "task": "refering", "answer_template": "The \"asparagus in the middle of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 148, 149, 150, 151, 152, 169, 170, 171, 172, 191, 192, 193, 194, 214, 215, 237, 238, 260, 261], "bbox": [0.33785937499999996, 0.29952083333333335, 0.6367968749999999, 0.67925], "iscrowd": 0, "area": 6056.3278999999975, "rle": {"size": [480, 640], "counts": "PaU3j0V=WOhCj0T<\\OiCf0Q<AkCb0o;FmC<n;JoC7l;0QD2i;5TDMg;9UDJe;=XDDd;o1O10O0100O1O01M2H9POQDeNW<Y1WDSNQ<k1d0L3N3M2N2N3M1N3N2N1O2N2I6E<E;1N100O2O000O2O000O2O000O2O000O101N10001N10000O10000O1000O0100O10000O010O0O101N101O0O2O0O101O0O2O0O1010O100O010O1O010O10O01O1O001O1O001O1O1O00100O001O1O001O1O001O1O001O1O1O010O10O010O010O010O010O010O0010O010O010O010O010O0100O010O010O010O1O100O100O100O100O100O100O7J3L100O00100O010O10O0100O0100N2N3M2O1N3M2M4I7H7Jbc\\3"}, "label": "asparagus in the middle of the plate"}]}
{"id": 505625, "image": "COCO_train2014_000000505625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of green brocolli in a stir fry\"."}], "task": "refering", "answer_template": "The \"a piece of green brocolli in a stir fry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 148, 149, 150, 151, 152, 169, 170, 171, 172, 191, 192, 193, 194, 214, 215, 237, 238, 260, 261], "bbox": [0.33785937499999996, 0.29952083333333335, 0.6367968749999999, 0.67925], "iscrowd": 0, "area": 6056.3278999999975, "rle": {"size": [480, 640], "counts": "PaU3j0V=WOhCj0T<\\OiCf0Q<AkCb0o;FmC<n;JoC7l;0QD2i;5TDMg;9UDJe;=XDDd;o1O10O0100O1O01M2H9POQDeNW<Y1WDSNQ<k1d0L3N3M2N2N3M1N3N2N1O2N2I6E<E;1N100O2O000O2O000O2O000O2O000O101N10001N10000O10000O1000O0100O10000O010O0O101N101O0O2O0O101O0O2O0O1010O100O010O1O010O10O01O1O001O1O001O1O1O00100O001O1O001O1O001O1O001O1O1O010O10O010O010O010O010O010O0010O010O010O010O010O0100O010O010O010O1O100O100O100O100O100O100O7J3L100O00100O010O10O0100O0100N2N3M2O1N3M2M4I7H7Jbc\\3"}, "label": "a piece of green brocolli in a stir fry"}]}
{"id": 145178, "image": "COCO_train2014_000000145178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green jacket holding a silver laptop\"."}], "task": "refering", "answer_template": "The \"a man in a green jacket holding a silver laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 188, 189, 190, 202, 203, 204, 205, 206, 207], "bbox": [0.07415999999999999, 0.47559420289855076, 0.5812, 1.0], "iscrowd": 0, "area": 22878.51215, "rle": {"size": [345, 500], "counts": "ag<2d:3M4L4M3N2O1N2N2N2N2O1N2N2N2N101N2N2N2O1N2N1O2N101N2N1O2N101N1O1O2N101N1O1O2O0O1O2N1O101N1O1O2N100O0010O100O1N1O2N2M3N2N2N110000000000O100bHPNS6P2mIQNR6o1nIRNQ6n1oISNo5n1QJSNn5l1SJVNk5j1UJWNj5i1VJXNi5h1VJZNi5f1WJ[Ng5f1YJZNg5e1ZJ[Nf5e1ZJ[Nf5e1ZJ[Nf5e1ZJ[Ne5f1[JZNe5f1ZJ[Nf5e1ZJ[Nf5d1[J\\Ne5d1[J\\Ne5d1[J\\Nd5e1\\J[Nc5f1]JZNb5g1^JYNa5h1_JXN`5i1_JXNi4CbJW2e0VNd4IeJR2g0UN_4OhJm1i0TN]42hJk1k0SN\\44fJk1n0QN[46eJj1P1PNY49eJh1R1PNW49fJh1S1oMV4;dJg1V1oMU4;dJg1W1nMT4=cJf1Y1mMR4S3nKmLQ4T3oKlLP4U3PLkLo3V3QLjLn3W3RLiLm3X3RLiLn3W3RLiLm3X3SLhLm3X3SLhLm3X3SLhLm3X3SLhLl3Y3TLgLl3Y3TLgLl3Y3TLgLl3Y3SLhLm3W3TLiLl3W3TLiLm3V3SLjLm3V3SLjLm3V3SLjLm3V3SLjLm3V3SLjLm3V3SLjLm3V3RLkLo3T3QLlLo3T3QLlLo3T3QLlLo3T3QLlLo3T3QLlLo3S3RLmLn3S3RLmLn3S3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3PLoLP4Q3PLoLP4Q3PLoLP4Q3PLoLP4P3QLPMo3P3QLPMP4o2PLQMP4o2PLQMP4o2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMR4m2nKSMR4m2mKTMS4l2mKTMS4k2nKUMR4k2nKUMR4k2nKUMS4j2mKVMT4i2lKWMU4h2kKXMU4h2kKXMV4g2iKZMX4e2hK[MY4d2gK\\MZ4KQKd1e0bN]4EPKh1c0cN`4@oJm1a0cNd4[OmJQ2?dN`5Z1aJfN`5Y1_JhNa5X1_JhNb5W1^JiNc5V1]JjNc5V1]JjNd5U1\\JkNd5U1\\JkNe5T1[JlNe5T1[JlNe5T1ZJmNf5S1ZJmNf5S1YJnNh5Q1WJPOi5P1WJPOi5o0WJROi5n0VJSOj5m0UJTOk5`1`IaNa6k2000000001O000000001O0000001N2O0O2]OPIZMQ7d2VIWMk6f2[IUMg6h2f0N3M3N1O2O1N1O2N2N2O1N2N2N1O2O2M3M2N3N2M3M2N3M3N2M2N3M3M3N1N3M3M4M8G9GZTV2"}, "label": "a man in a green jacket holding a silver laptop"}]}
{"id": 145178, "image": "COCO_train2014_000000145178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing the green t shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing the green t shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 188, 189, 190, 202, 203, 204, 205, 206, 207], "bbox": [0.07415999999999999, 0.47559420289855076, 0.5812, 1.0], "iscrowd": 0, "area": 22878.51215, "rle": {"size": [345, 500], "counts": "ag<2d:3M4L4M3N2O1N2N2N2N2O1N2N2N2N101N2N2N2O1N2N1O2N101N2N1O2N101N1O1O2N101N1O1O2O0O1O2N1O101N1O1O2N100O0010O100O1N1O2N2M3N2N2N110000000000O100bHPNS6P2mIQNR6o1nIRNQ6n1oISNo5n1QJSNn5l1SJVNk5j1UJWNj5i1VJXNi5h1VJZNi5f1WJ[Ng5f1YJZNg5e1ZJ[Nf5e1ZJ[Nf5e1ZJ[Nf5e1ZJ[Ne5f1[JZNe5f1ZJ[Nf5e1ZJ[Nf5d1[J\\Ne5d1[J\\Ne5d1[J\\Nd5e1\\J[Nc5f1]JZNb5g1^JYNa5h1_JXN`5i1_JXNi4CbJW2e0VNd4IeJR2g0UN_4OhJm1i0TN]42hJk1k0SN\\44fJk1n0QN[46eJj1P1PNY49eJh1R1PNW49fJh1S1oMV4;dJg1V1oMU4;dJg1W1nMT4=cJf1Y1mMR4S3nKmLQ4T3oKlLP4U3PLkLo3V3QLjLn3W3RLiLm3X3RLiLn3W3RLiLm3X3SLhLm3X3SLhLm3X3SLhLm3X3SLhLl3Y3TLgLl3Y3TLgLl3Y3TLgLl3Y3SLhLm3W3TLiLl3W3TLiLm3V3SLjLm3V3SLjLm3V3SLjLm3V3SLjLm3V3SLjLm3V3SLjLm3V3RLkLo3T3QLlLo3T3QLlLo3T3QLlLo3T3QLlLo3T3QLlLo3S3RLmLn3S3RLmLn3S3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3QLnLo3R3PLoLP4Q3PLoLP4Q3PLoLP4Q3PLoLP4P3QLPMo3P3QLPMP4o2PLQMP4o2PLQMP4o2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMQ4n2oKRMR4m2nKSMR4m2mKTMS4l2mKTMS4k2nKUMR4k2nKUMR4k2nKUMS4j2mKVMT4i2lKWMU4h2kKXMU4h2kKXMV4g2iKZMX4e2hK[MY4d2gK\\MZ4KQKd1e0bN]4EPKh1c0cN`4@oJm1a0cNd4[OmJQ2?dN`5Z1aJfN`5Y1_JhNa5X1_JhNb5W1^JiNc5V1]JjNc5V1]JjNd5U1\\JkNd5U1\\JkNe5T1[JlNe5T1[JlNe5T1ZJmNf5S1ZJmNf5S1YJnNh5Q1WJPOi5P1WJPOi5o0WJROi5n0VJSOj5m0UJTOk5`1`IaNa6k2000000001O000000001O0000001N2O0O2]OPIZMQ7d2VIWMk6f2[IUMg6h2f0N3M3N1O2O1N1O2N2N2O1N2N2N1O2O2M3M2N3N2M3M2N3M3N2M2N3M3M3N1N3M3M4M8G9GZTV2"}, "label": "a man wearing the green t shirt"}]}
{"id": 145178, "image": "COCO_train2014_000000145178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man dall wearing the black color dress peace design tshirt\"."}], "task": "refering", "answer_template": "The \"a man dall wearing the black color dress peace design tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 65, 66, 67, 83, 84, 85, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212], "bbox": [0.509, 0.2107536231884058, 0.83542, 1.0], "iscrowd": 0, "area": 28663.879, "rle": {"size": [345, 500], "counts": "RTf23d:3M2N3M3N2M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3^JRNU2Q2fMSNX2P2dMSNZ2o1bMUN\\2n1_MVN_2m1\\MWNb2l1YMXNe2j1WMZNg2i1UMZNi2i1RM[Nm2g1nL]NP3e1lL_NR3d1iL`NU3c1fLaNX3b1dLaNZ3a1bLcN]3_1^LeNa3]1ZLgNf3Z1ULjNk3V1QLnNn3T1mKPOS4Q1iKROW4o0dKUO\\4d3O10000O1000000O10000O10000O1000000O10000O1UOhIUMX6b2SJ\\Mm5[2^JcMb5U2hJhMY5W2jJgMV5Y2lJeMT5Z2oJcMR5]2PKaMP5^2f100O1O100O100O100O1000000000000000000000000O10000000000000000000000002XIfMm4\\2PKeMP5\\2nJfMQ5\\2lJeMT5]2iJeMV5]2gJdMY5]2eJeMZ5]2dJcM\\5_2aJcM^5_2_JbMa5`2\\JaMd5a2YJaMf5`2XJaMh5e2QJ]Mn5k2iIVMW6g3O001O00001O001O001O001O001O0YMRLcNn3\\1WL`Nj3]1\\L_Nd3_1bL]N_3a1gLZNY3d1mLXNT3e1RMWNn2h1WMTNj2i1]MRNc2l1cMPNb2j1cMRNb2g1dMUNb2c1dMYNa2a1eMZNa2^1eM^N`2[1fMaN`2X1eMdN`2U1i3H7J7H7I7JX]k0"}, "label": "a man dall wearing the black color dress peace design tshirt"}]}
{"id": 145178, "image": "COCO_train2014_000000145178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a digital image of a male wearing a john lennon tee shirt\"."}], "task": "refering", "answer_template": "The \"a digital image of a male wearing a john lennon tee shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 65, 66, 67, 83, 84, 85, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212], "bbox": [0.509, 0.2107536231884058, 0.83542, 1.0], "iscrowd": 0, "area": 28663.879, "rle": {"size": [345, 500], "counts": "RTf23d:3M2N3M3N2M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3^JRNU2Q2fMSNX2P2dMSNZ2o1bMUN\\2n1_MVN_2m1\\MWNb2l1YMXNe2j1WMZNg2i1UMZNi2i1RM[Nm2g1nL]NP3e1lL_NR3d1iL`NU3c1fLaNX3b1dLaNZ3a1bLcN]3_1^LeNa3]1ZLgNf3Z1ULjNk3V1QLnNn3T1mKPOS4Q1iKROW4o0dKUO\\4d3O10000O1000000O10000O10000O1000000O10000O1UOhIUMX6b2SJ\\Mm5[2^JcMb5U2hJhMY5W2jJgMV5Y2lJeMT5Z2oJcMR5]2PKaMP5^2f100O1O100O100O100O1000000000000000000000000O10000000000000000000000002XIfMm4\\2PKeMP5\\2nJfMQ5\\2lJeMT5]2iJeMV5]2gJdMY5]2eJeMZ5]2dJcM\\5_2aJcM^5_2_JbMa5`2\\JaMd5a2YJaMf5`2XJaMh5e2QJ]Mn5k2iIVMW6g3O001O00001O001O001O001O001O0YMRLcNn3\\1WL`Nj3]1\\L_Nd3_1bL]N_3a1gLZNY3d1mLXNT3e1RMWNn2h1WMTNj2i1]MRNc2l1cMPNb2j1cMRNb2g1dMUNb2c1dMYNa2a1eMZNa2^1eM^N`2[1fMaN`2X1eMdN`2U1i3H7J7H7I7JX]k0"}, "label": "a digital image of a male wearing a john lennon tee shirt"}]}
{"id": 505626, "image": "COCO_train2014_000000505626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"butter knife laid on plate\"."}], "task": "refering", "answer_template": "The \"butter knife laid on plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [313, 314, 315, 335, 336, 337, 356, 357, 358, 359, 378, 379, 380, 399, 400, 401, 402], "bbox": [0.358640625, 0.7272619047619048, 0.74034375, 1.0], "iscrowd": 0, "area": 6361.445400000003, "rle": {"size": [504, 640], "counts": "Wfa31g?0O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O10O000010O01O000010O00010O01O01O01O01O01O010O00010O00010O0010O00010O00010O01O01O01O010O0010O01O010O0010O01O010O0010O01O00O1N2O100O1O1001O001O000100O00100O1O010O1O1O010O1O100O001O10O00010O00010O00010O0001O01O01O01O00010O00010O00010O0001O01O01O01O01O01O0010O01O010O0010O01O010O0010O01O010O001O010O00100O00100O00100O00100O001O10O01O100O1O10O01O100O1O10O01O100O1O1O1O1O2N1O101N1O1O2N1O1O2N1O10mZa2"}, "label": "butter knife laid on plate"}]}
{"id": 505626, "image": "COCO_train2014_000000505626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife on the left of the plate\"."}], "task": "refering", "answer_template": "The \"the knife on the left of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [313, 314, 315, 335, 336, 337, 356, 357, 358, 359, 378, 379, 380, 399, 400, 401, 402], "bbox": [0.358640625, 0.7272619047619048, 0.74034375, 1.0], "iscrowd": 0, "area": 6361.445400000003, "rle": {"size": [504, 640], "counts": "Wfa31g?0O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O10O000010O01O000010O00010O01O01O01O01O01O010O00010O00010O0010O00010O00010O01O01O01O010O0010O01O010O0010O01O010O0010O01O00O1N2O100O1O1001O001O000100O00100O1O010O1O1O010O1O100O001O10O00010O00010O00010O0001O01O01O01O00010O00010O00010O0001O01O01O01O01O01O0010O01O010O0010O01O010O0010O01O010O001O010O00100O00100O00100O00100O001O10O01O100O1O10O01O100O1O10O01O100O1O1O1O1O2N1O101N1O1O2N1O1O2N1O10mZa2"}, "label": "the knife on the left of the plate"}]}
{"id": 505626, "image": "COCO_train2014_000000505626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown chair with no one in it\"."}], "task": "refering", "answer_template": "The \"brown chair with no one in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 94, 95, 98, 99, 100, 121, 122, 123, 124], "bbox": [0.100875, 0.07640873015873015, 0.39640624999999996, 0.31684523809523807], "iscrowd": 0, "area": 9855.124000000002, "rle": {"size": [504, 640], "counts": "lQP13d?4M3M3L4M3M3L5L3M3J6M3M3N2N3L3N2NO001O1O010O000000O1000@PBXOP>h0SBUOm=k0VBROj=n0YBoNf=R1\\BlNd=T1`0000000O10O1000000000000O10O1000000000O100000O100000O100000000000O01000000000000O1000O10000000O1000000000O01000000000000O1000O1000000000000000_BjN`<V1`CoN[<P1fCTOV<l0jCXOR<h0nC]Om;c0SDAi;?WDAi;?WDAi;?WDAi;?WDAi;?WDAi;?WDAi;>WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;<XDDh;<XDDh;<XDDh;<XDDh;<XDDh;<XDDh;<XDDh;<XDEg;;XDFh;:XDFh;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;8ZDHf;8ZDHf;8ZDHf;8ZDHf;8ZDHf;8ZDHf;8YDIg;7YDIg;7YDIg;7YDIg;6ZDJf;6ZDJf;6ZDJf;6ZDJf;6ZDJf;6ZDJf;6ZDJf;b0nC^Oo;S1cCmNZ<Z2L4M3M4L30000000nMTD9k;@_D=a;\\OiDa0W;XOSEe0m:TO]Ei0c:POgEn0X:kNRFR1n9gN\\FV1V<M3M3N2M3M4L:F:Gbjm5"}, "label": "brown chair with no one in it"}]}
{"id": 505626, "image": "COCO_train2014_000000505626.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair in the left on which no one is sitting\"."}], "task": "refering", "answer_template": "The \"the chair in the left on which no one is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 94, 95, 98, 99, 100, 121, 122, 123, 124], "bbox": [0.100875, 0.07640873015873015, 0.39640624999999996, 0.31684523809523807], "iscrowd": 0, "area": 9855.124000000002, "rle": {"size": [504, 640], "counts": "lQP13d?4M3M3L4M3M3L5L3M3J6M3M3N2N3L3N2NO001O1O010O000000O1000@PBXOP>h0SBUOm=k0VBROj=n0YBoNf=R1\\BlNd=T1`0000000O10O1000000000000O10O1000000000O100000O100000O100000000000O01000000000000O1000O10000000O1000000000O01000000000000O1000O1000000000000000_BjN`<V1`CoN[<P1fCTOV<l0jCXOR<h0nC]Om;c0SDAi;?WDAi;?WDAi;?WDAi;?WDAi;?WDAi;?WDAi;>WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;=WDCi;<XDDh;<XDDh;<XDDh;<XDDh;<XDDh;<XDDh;<XDDh;<XDEg;;XDFh;:XDFh;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;9YDGg;8ZDHf;8ZDHf;8ZDHf;8ZDHf;8ZDHf;8ZDHf;8YDIg;7YDIg;7YDIg;7YDIg;6ZDJf;6ZDJf;6ZDJf;6ZDJf;6ZDJf;6ZDJf;6ZDJf;b0nC^Oo;S1cCmNZ<Z2L4M3M4L30000000nMTD9k;@_D=a;\\OiDa0W;XOSEe0m:TO]Ei0c:POgEn0X:kNRFR1n9gN\\FV1V<M3M3N2M3M4L:F:Gbjm5"}, "label": "the chair in the left on which no one is sitting"}]}
{"id": 145180, "image": "COCO_train2014_000000145180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a suit standing next to the fridge\"."}], "task": "refering", "answer_template": "The \"a man wearing a suit standing next to the fridge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 74, 75, 76, 96, 97, 98, 119, 120, 121, 122, 141, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328], "bbox": [0.157375, 0.18202409638554218, 0.4036406250000001, 0.9865060240963857], "iscrowd": 0, "area": 30740.99455, "rle": {"size": [415, 640], "counts": "UWY17c<:E;F:cFQOX3JXM`1nNSOg3DQMR4l2ULkLP4Q3XLeLn3X3YL_Ll3^3[LXLk3e3\\LRLi3k3^LkKh3R4_LeKf3Z4_L\\Kg3c4^LSKh3l4V200O100O100O100O100O1O100O100O100O100O1I7[Oe0C=L4K5L4K5L4O10001O00000000000000000000000000000000000000000000000001O00001O00001O001O00001O00001O0O2E:PHmLj5^3QJfLe5e3WJ_L^5l3]JYLX5R4dJQLR5Z4iJkKl4`4oJdKg4g4UK]Kg4g4TK^Kk4b4QKbKo4_4mJeKR5\\4iJiKV5X4eJlK[5m5O1eN_HVMa7h2hHQMY7l2oHlLT7Q3UIeLn6Y3[I\\Lj6a3\\1N5K8H7I8H7J7cN`ELj;00O10O1000000000O100000O100000O10000000O10O101O001O001O0O2O001O001O0O2O001O001O0O3N`bj4"}, "label": "a man wearing a suit standing next to the fridge"}]}
{"id": 145180, "image": "COCO_train2014_000000145180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man black suit\"."}], "task": "refering", "answer_template": "The \"a man black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 74, 75, 76, 96, 97, 98, 119, 120, 121, 122, 141, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328], "bbox": [0.157375, 0.18202409638554218, 0.4036406250000001, 0.9865060240963857], "iscrowd": 0, "area": 30740.99455, "rle": {"size": [415, 640], "counts": "UWY17c<:E;F:cFQOX3JXM`1nNSOg3DQMR4l2ULkLP4Q3XLeLn3X3YL_Ll3^3[LXLk3e3\\LRLi3k3^LkKh3R4_LeKf3Z4_L\\Kg3c4^LSKh3l4V200O100O100O100O100O1O100O100O100O100O1I7[Oe0C=L4K5L4K5L4O10001O00000000000000000000000000000000000000000000000001O00001O00001O001O00001O00001O0O2E:PHmLj5^3QJfLe5e3WJ_L^5l3]JYLX5R4dJQLR5Z4iJkKl4`4oJdKg4g4UK]Kg4g4TK^Kk4b4QKbKo4_4mJeKR5\\4iJiKV5X4eJlK[5m5O1eN_HVMa7h2hHQMY7l2oHlLT7Q3UIeLn6Y3[I\\Lj6a3\\1N5K8H7I8H7J7cN`ELj;00O10O1000000000O100000O100000O10000000O10O101O001O001O0O2O001O001O0O2O001O001O0O3N`bj4"}, "label": "a man black suit"}]}
{"id": 145180, "image": "COCO_train2014_000000145180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady in a black frock in the left side of the image\"."}], "task": "refering", "answer_template": "The \"lady in a black frock in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 70, 71, 72, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325], "bbox": [0.00134375, 0.19231325301204819, 0.1781875, 1.0], "iscrowd": 0, "area": 25040.1394, "rle": {"size": [415, 640], "counts": "ib07c<:F:G9F:F:Ff0\\O6L4M1N2O0O2O0O101N100O2O0O2Ok1TN6K4L5K5K5J6K4L=CY1fNd0\\O1O100O1O1O100O1O1O1oNZHiLf7U3T1M2N2N2N2N2N2O1O1O100O1O1O100O1O100O1000000001O01O0000000001O00000000001O1O1O1O1O1oI\\Me1e2UNbMi1_2PNiMn1X2lMPNQ2Q2iMVNU2k1hMYNV2h1hM[NV2g1fM^NX2b1fMaNX2a1dMcNZ2`1bMdN[2^1aMfN]2]1aMdN]2_1aMcN\\2T2mLSNl2Q2QM?>EaNiK1f4Y1F]NiK:d4U1HXNhKd0d4o09POLh08XOL>9CK3;MHJ>6F@d0;U6L4L5K4LXod6"}, "label": "lady in a black frock in the left side of the image"}]}
{"id": 145180, "image": "COCO_train2014_000000145180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black dress and holding a champagne flute\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black dress and holding a champagne flute\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 70, 71, 72, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325], "bbox": [0.00134375, 0.19231325301204819, 0.1781875, 1.0], "iscrowd": 0, "area": 25040.1394, "rle": {"size": [415, 640], "counts": "ib07c<:F:G9F:F:Ff0\\O6L4M1N2O0O2O0O101N100O2O0O2Ok1TN6K4L5K5K5J6K4L=CY1fNd0\\O1O100O1O1O100O1O1O1oNZHiLf7U3T1M2N2N2N2N2N2O1O1O100O1O1O100O1O100O1000000001O01O0000000001O00000000001O1O1O1O1O1oI\\Me1e2UNbMi1_2PNiMn1X2lMPNQ2Q2iMVNU2k1hMYNV2h1hM[NV2g1fM^NX2b1fMaNX2a1dMcNZ2`1bMdN[2^1aMfN]2]1aMdN]2_1aMcN\\2T2mLSNl2Q2QM?>EaNiK1f4Y1F]NiK:d4U1HXNhKd0d4o09POLh08XOL>9CK3;MHJ>6F@d0;U6L4L5K4LXod6"}, "label": "a woman wearing a black dress and holding a champagne flute"}]}
{"id": 333599, "image": "COCO_train2014_000000333599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a skateboard at the top of the ramp\"."}], "task": "refering", "answer_template": "The \"a man on a skateboard at the top of the ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 35, 36, 49, 50, 51, 64, 65, 66, 67, 68, 69, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 107, 108, 109, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 146, 147, 148], "bbox": [0.1313583138173302, 0.062921875, 0.9161358313817332, 0.42021875], "iscrowd": 0, "area": 24428.947200000006, "rle": {"size": [640, 427], "counts": "fUS15`c0<M3O1N2N2O0O2N2O1N2O1N2N101N2N2O1N2O1N1O2O1N2N2O1N2O0O2N2O1N1O010O100O1O010O1O100O10000000000000000000000000000000O10000O100OO2N2N1O2N1O2N2N1O2N1O2O1N1O2N1O2N1O2N1O2N101N1O2N1O2N2N2N1O2O1N2O1N1O2O1N2O1N2OiMm_Ob0R`0_On_Ob0P`0@P@`0o?@Q@a0m?@T@`0k?@V@?i?CV@>i?BX@>f?CZ@>e?B\\@>b?D]@=b?C_@<a?D_@=_?Db@<]?Dc@=\\?Dd@<[?De@<[?Df@<Y?Dg@=X?Dh@<W?Dj@<T?Em@:S?Eo@;P?EQA;o>CSA=n>ASA?m>_OUAa0k>^OVAa0k>]OWAc0j>[OWAe0i>XOZAh0f>UO]Ak0c>RO`Am0a>PObAP1_>lNdAT1\\>iNgAW1Y>fNjAZ1V>cNmA]1T>_NoA`1R>]NQBc1o=ZNTBf1l=WNWBi1P`000O100000000000000000000000000000000O100000000O100000000O1000000002N1O3M4L3M1O1O1O1O1O001O1O1O1O001000000O01000000O11N1000001O000O2O01O1O2N1N2O1N2O3M4K4M4K4M3M3L4M3M3L2O1O1N2O1OO001O10O01O1O010O1O00100O001O1O10O01O0010O01O001O010O001O0010O01O001O001O001O001O0010O01O001O001O00001O00000010O000001O00000000000001O0000001O00001O00000O2O0O101N1O100O2O0O1N3N1N2O2M2O1N2O2K4K5J7J5L4L5K\\Yf0"}, "label": "a man on a skateboard at the top of the ramp"}]}
{"id": 333599, "image": "COCO_train2014_000000333599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in bluejeans and a white shirt riding on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man in bluejeans and a white shirt riding on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 35, 36, 49, 50, 51, 64, 65, 66, 67, 68, 69, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 107, 108, 109, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 146, 147, 148], "bbox": [0.1313583138173302, 0.062921875, 0.9161358313817332, 0.42021875], "iscrowd": 0, "area": 24428.947200000006, "rle": {"size": [640, 427], "counts": "fUS15`c0<M3O1N2N2O0O2N2O1N2O1N2N101N2N2O1N2O1N1O2O1N2N2O1N2O0O2N2O1N1O010O100O1O010O1O100O10000000000000000000000000000000O10000O100OO2N2N1O2N1O2N2N1O2N1O2O1N1O2N1O2N1O2N1O2N101N1O2N1O2N2N2N1O2O1N2O1N1O2O1N2O1N2OiMm_Ob0R`0_On_Ob0P`0@P@`0o?@Q@a0m?@T@`0k?@V@?i?CV@>i?BX@>f?CZ@>e?B\\@>b?D]@=b?C_@<a?D_@=_?Db@<]?Dc@=\\?Dd@<[?De@<[?Df@<Y?Dg@=X?Dh@<W?Dj@<T?Em@:S?Eo@;P?EQA;o>CSA=n>ASA?m>_OUAa0k>^OVAa0k>]OWAc0j>[OWAe0i>XOZAh0f>UO]Ak0c>RO`Am0a>PObAP1_>lNdAT1\\>iNgAW1Y>fNjAZ1V>cNmA]1T>_NoA`1R>]NQBc1o=ZNTBf1l=WNWBi1P`000O100000000000000000000000000000000O100000000O100000000O1000000002N1O3M4L3M1O1O1O1O1O001O1O1O1O001000000O01000000O11N1000001O000O2O01O1O2N1N2O1N2O3M4K4M4K4M3M3L4M3M3L2O1O1N2O1OO001O10O01O1O010O1O00100O001O1O10O01O0010O01O001O010O001O0010O01O001O001O001O001O0010O01O001O001O00001O00000010O000001O00000000000001O0000001O00001O00000O2O0O101N1O100O2O0O1N3N1N2O2M2O1N2O2K4K5J7J5L4L5K\\Yf0"}, "label": "a man in bluejeans and a white shirt riding on a skateboard"}]}
{"id": 423711, "image": "COCO_train2014_000000423711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tabletop with the man resting his elbow\"."}], "task": "refering", "answer_template": "The \"the tabletop with the man resting his elbow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 231, 232, 233, 234, 255, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305], "bbox": [0.049703125, 0.6450351288056206, 0.30121875, 0.9159250585480094], "iscrowd": 0, "area": 8140.241700000001, "rle": {"size": [427, 640], "counts": "gc=1Z=1O1N2O2N1O000O2O00001O0O2O001O001N110O0010O010O010O010O010O01O010O011N100O101N100O1O2O0O2O1N2O0O2O1N2N101N2SD]Oo:d0jDIn:]1N3L3N3L3N3L3M01O001O001O0O2O001O00001O001O0O2O001O00001O001O0O2O001O001O00001O0O2O001O001O001O000O2O001O001O001O000O2O001O001O001O001N10001O001O001O001N101O00001O001O001O0O2O00001O001O1O4L4K5L4L4L4L4L4L4L4L4KkPj5"}, "label": "the tabletop with the man resting his elbow"}]}
{"id": 423711, "image": "COCO_train2014_000000423711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a white tablecloth seats a guy smiling at a standing woman in white\"."}], "task": "refering", "answer_template": "The \"a table with a white tablecloth seats a guy smiling at a standing woman in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 231, 232, 233, 234, 255, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305], "bbox": [0.049703125, 0.6450351288056206, 0.30121875, 0.9159250585480094], "iscrowd": 0, "area": 8140.241700000001, "rle": {"size": [427, 640], "counts": "gc=1Z=1O1N2O2N1O000O2O00001O0O2O001O001N110O0010O010O010O010O010O01O010O011N100O101N100O1O2O0O2O1N2O0O2O1N2N101N2SD]Oo:d0jDIn:]1N3L3N3L3N3L3M01O001O001O0O2O001O00001O001O0O2O001O00001O001O0O2O001O001O00001O0O2O001O001O001O000O2O001O001O001O000O2O001O001O001O001N10001O001O001O001N101O00001O001O001O0O2O00001O001O1O4L4K5L4L4L4L4L4L4L4L4KkPj5"}, "label": "a table with a white tablecloth seats a guy smiling at a standing woman in white"}]}
{"id": 366373, "image": "COCO_train2014_000000366373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in white shirt sleeping on a bed\"."}], "task": "refering", "answer_template": "The \"boy in white shirt sleeping on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 279, 280, 281, 282, 283, 284, 285, 287, 288, 289, 304, 307, 308, 330, 331, 353, 354, 376], "bbox": [0.14865625, 0.5202708333333333, 0.69425, 0.9527083333333333], "iscrowd": 0, "area": 27850.78680000001, "rle": {"size": [480, 640], "counts": "mj\\15c>c0]Ob0E2N2N2N3M2N2M3N2O1N3N1N2O1O1N2O1N101O1O1O1O1O001O000000000000001O000000000010O000000010O001O00010O00001O010O0O100O100O2N100O100O100O2O1N2N01O001N1DTNeCm1[<;010O01O010O00100O0100O10OL5J6N1101O1O0O2O1O001N101O1O0O2O001O0O2O00001O001N10001O001O001N10001O1O1O2N1N2O2N1O1O1O2N4K;mEXLZ8R4TGXLj8Z4]FPL`9k4M1N3N2N7I4L0010O001OOWLUG_1k8aNXG\\1g8eN\\GX1d8hN`GT1_8mNcGP1^8POdGn0\\8ROfGk0Z8VOiGg0W8YOkGe0T8\\OnGa0S8_OoG?P8BSH;m7EVHROnNFl8Y1gIeNY6[1hIbNY6_1hI_NY6a1iI[NX6e1jIYNW6g1jIVNW6k1kIQNV6P2lImMU6S2nIhMS6Y2PJcMQ6]2X3100O1O100000000000000000000000000000O10000000000000000000000002N1O2N1O2N1O2N1O0010O0001O00001O00010O00001O00000001O000000001O1O001N2O001N101O1N101O1O0O2O1O0O2O001N2O00001O0010O0001O000000001O0000001O0000000000001O00000000001OO1O1O1O1O1O1OVOgCoNY<Q1iClNW<U1kChNU<Y1nCbNT<a0_C2?ZOR<d0`C2U=NmB0T=OnB0R=OoB1Q=NQC1o<ORC0o<OQC1o<ORC0n<0SCOm<1SCOm<1TCNm<1SCOm<1SCOm<1TCNl<2TCMm<3SCMn<2SCLn<3SCMm<3SCMm<3TCKm<5SCKn<4RCKo<5RCJn<6RCIo<7QCIo<7RCHn<8TCEm<;UCCl<<l0000000O101O0000001O0O101O00001N1000001O0O10_ck2"}, "label": "boy in white shirt sleeping on a bed"}]}
{"id": 366373, "image": "COCO_train2014_000000366373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy sleeping on the end of the bed\"."}], "task": "refering", "answer_template": "The \"a small boy sleeping on the end of the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 279, 280, 281, 282, 283, 284, 285, 287, 288, 289, 304, 307, 308, 330, 331, 353, 354, 376], "bbox": [0.14865625, 0.5202708333333333, 0.69425, 0.9527083333333333], "iscrowd": 0, "area": 27850.78680000001, "rle": {"size": [480, 640], "counts": "mj\\15c>c0]Ob0E2N2N2N3M2N2M3N2O1N3N1N2O1O1N2O1N101O1O1O1O1O001O000000000000001O000000000010O000000010O001O00010O00001O010O0O100O100O2N100O100O100O2O1N2N01O001N1DTNeCm1[<;010O01O010O00100O0100O10OL5J6N1101O1O0O2O1O001N101O1O0O2O001O0O2O00001O001N10001O001O001N10001O1O1O2N1N2O2N1O1O1O2N4K;mEXLZ8R4TGXLj8Z4]FPL`9k4M1N3N2N7I4L0010O001OOWLUG_1k8aNXG\\1g8eN\\GX1d8hN`GT1_8mNcGP1^8POdGn0\\8ROfGk0Z8VOiGg0W8YOkGe0T8\\OnGa0S8_OoG?P8BSH;m7EVHROnNFl8Y1gIeNY6[1hIbNY6_1hI_NY6a1iI[NX6e1jIYNW6g1jIVNW6k1kIQNV6P2lImMU6S2nIhMS6Y2PJcMQ6]2X3100O1O100000000000000000000000000000O10000000000000000000000002N1O2N1O2N1O2N1O0010O0001O00001O00010O00001O00000001O000000001O1O001N2O001N101O1N101O1O0O2O1O0O2O001N2O00001O0010O0001O000000001O0000001O0000000000001O00000000001OO1O1O1O1O1O1OVOgCoNY<Q1iClNW<U1kChNU<Y1nCbNT<a0_C2?ZOR<d0`C2U=NmB0T=OnB0R=OoB1Q=NQC1o<ORC0o<OQC1o<ORC0n<0SCOm<1SCOm<1TCNm<1SCOm<1SCOm<1TCNl<2TCMm<3SCMn<2SCLn<3SCMm<3SCMm<3TCKm<5SCKn<4RCKo<5RCJn<6RCIo<7QCIo<7RCHn<8TCEm<;UCCl<<l0000000O101O0000001O0O101O00001N1000001O0O10_ck2"}, "label": "a small boy sleeping on the end of the bed"}]}
{"id": 145192, "image": "COCO_train2014_000000145192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stool the cat is lying on\"."}], "task": "refering", "answer_template": "The \"the stool the cat is lying on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 200, 201, 202, 214, 215, 223, 224, 225, 237, 238, 246, 247, 248, 260, 261, 270, 271], "bbox": [0.306734375, 0.4123055555555556, 0.817421875, 0.9044444444444445], "iscrowd": 0, "area": 18982.975850000006, "rle": {"size": [360, 640], "counts": "W]U21W;001O1O001N2O1O001O1O008Hh0XOh0XOh0XOh0XOg0YOh0WO8I000000000000O1000000O1O1N1L5F:N2O1N2O1[Oe0gNY1lNT10000000000000000000000000000000000O10000000000000000000O10000000000000000000000000000000000000000000000000O100000000000O1000000000000000000000000000000000000000000000000000000000O1000O100000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000O10000000000000000000000000000000O100000000000000000000000000000000000000000000000O100000000000003Mo0QOo0QOo0QO3M00000000O10000YNg1eN[100000000000O100000000000000000k3UL1O1O1N010000O100A?^Ob0_Oa0_Oa0^Ob0_Oa0_Oa0^Oa0B?O2O0O2O0O2O0O5L4KZoX1"}, "label": "the stool the cat is lying on"}]}
{"id": 145192, "image": "COCO_train2014_000000145192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden piano bench with a gray cat sitting on it\"."}], "task": "refering", "answer_template": "The \"a wooden piano bench with a gray cat sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 200, 201, 202, 214, 215, 223, 224, 225, 237, 238, 246, 247, 248, 260, 261, 270, 271], "bbox": [0.306734375, 0.4123055555555556, 0.817421875, 0.9044444444444445], "iscrowd": 0, "area": 18982.975850000006, "rle": {"size": [360, 640], "counts": "W]U21W;001O1O001N2O1O001O1O008Hh0XOh0XOh0XOh0XOg0YOh0WO8I000000000000O1000000O1O1N1L5F:N2O1N2O1[Oe0gNY1lNT10000000000000000000000000000000000O10000000000000000000O10000000000000000000000000000000000000000000000000O100000000000O1000000000000000000000000000000000000000000000000000000000O1000O100000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000O10000000000000000000000000000000O100000000000000000000000000000000000000000000000O100000000000003Mo0QOo0QOo0QO3M00000000O10000YNg1eN[100000000000O100000000000000000k3UL1O1O1N010000O100A?^Ob0_Oa0_Oa0^Ob0_Oa0_Oa0^Oa0B?O2O0O2O0O2O0O5L4KZoX1"}, "label": "a wooden piano bench with a gray cat sitting on it"}]}
{"id": 145192, "image": "COCO_train2014_000000145192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown piano with lid closed\"."}], "task": "refering", "answer_template": "The \"brown piano with lid closed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 19, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 97, 98, 99, 100, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 133, 134, 144, 156, 157, 167, 179, 180, 190, 202, 203, 213, 225, 226], "bbox": [0.237640625, 0.06291666666666666, 0.8507031250000001, 0.7550555555555556], "iscrowd": 0, "area": 37939.28019999997, "rle": {"size": [360, 640], "counts": "]ae1>o9l0SOl0WOi001O0O2O001O0O102N>B=C>A>C>B=C>A>C>B=C7H1000000O13M3M3N2M3M000J6D<E;D<D<E;D<D<D<E;D<D<H800001O00000000000000000000000O10000000000000N2N2N2M3N2M3N2M3N2O1O1O100O1O1O100O1O1O100O1000000000000000000000000000000000000000000000000001O00000000001N10001O00001O00001O00001O0000001O0000001O00001O0000001O0000001O0000001O000000001O00000000001O00000000001O00000000001O000O1000001O00000000001O00000000001O000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N4L4L4L4L4L5K3M0000000O2O001kHaL_6_3`IoLS6R3kI[Mi5f2UJgM^5Z2aJSNR5o1lJ^Ng4c1XKjNZ4_4C5K1N201O000000G9SOm0ROn0SOm0ROo0YOf0A?@`0A?A?AggQ1"}, "label": "brown piano with lid closed"}]}
{"id": 309034, "image": "COCO_train2014_000000309034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person travelling with a suitcase and a handbag\"."}], "task": "refering", "answer_template": "The \"a person travelling with a suitcase and a handbag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 68, 69, 70, 71, 72, 73, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 188, 189, 190, 191, 192, 205, 206, 207, 208, 209, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 257, 258, 259, 260], "bbox": [0.014916666666666667, 0.013640625, 0.65325, 0.6825468750000001], "iscrowd": 0, "area": 64063.11665, "rle": {"size": [640, 480], "counts": "fc4=Vb0^1bN^1cN\\1cN^1bN^1B=01O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O00o0QO3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M00O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000000000O100000000000000O1000000000000oJ[IgLe6X3bIaL_6]3iI]LW6b3QJWLo5g3YJSLg5l3`JnK`5P4iJiKW5V4PKdKP5Z4XK_Ki4`4_KYKa4e4gKUKY4\\OWIT3g2ZMR4BeId2b2cMi3JQJU2]2kMb30_Je1W2UNZ36mJV1Q2]NR3=[Kf0k1fNk2e0gK6`1TOi2f0ULFT1Dg2f0bLXOh02f2f0PMhN<b0d2g0]MXN0Q1c2g0Y2YOgMg0Y2YOgMh0X2WOiMi0W2WOiMi0W2WOiMi0W2WOiMj0V2VOjMj0V2VOjMj0V2VOjMj0V2UOkMl0T2ROnMn0R2QOoMo0Q2oNQNQ1o1nNRNS1m1kNUNU1k1iNWNW1i1hNXNX1h1fNZN[1e1dN\\N\\1e1aN]N_1c1`N^Na1a1]NaNc1_1\\NbNd1^1ZNdNf1\\1XNfNi1Y1VNhNj1X1TNjNl1V1SNkNm1U1QNnNo1Q1PNPOP2P1nMROR2n0mMTOR2l0lMVOU2i0iMZOV2f0iM[OW2e0gM^OY2a0fM@Z2S901O010O001O01O01O0010O01O0010O00010O01O0010O0001O010O001O01O01O0010O01O01O01O0010O01O00010O001O010O00010O001O001N101O000O2O001N101O000O2O001N101O000O2O001N101O000O2O001N101O000O2O001N101O000O2O001N101O001N10001N101O001N10001N101J5@afW3"}, "label": "a person travelling with a suitcase and a handbag"}]}
{"id": 309034, "image": "COCO_train2014_000000309034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a light coat and patterned pants in a black and white photo with luggage\"."}], "task": "refering", "answer_template": "The \"a person with a light coat and patterned pants in a black and white photo with luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 68, 69, 70, 71, 72, 73, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 188, 189, 190, 191, 192, 205, 206, 207, 208, 209, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 257, 258, 259, 260], "bbox": [0.014916666666666667, 0.013640625, 0.65325, 0.6825468750000001], "iscrowd": 0, "area": 64063.11665, "rle": {"size": [640, 480], "counts": "fc4=Vb0^1bN^1cN\\1cN^1bN^1B=01O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O00o0QO3M3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M3M4L3M3M4L3M4L3M3M4L3M3M4L3M00O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O100000000000000O100000000000000O1000000000000oJ[IgLe6X3bIaL_6]3iI]LW6b3QJWLo5g3YJSLg5l3`JnK`5P4iJiKW5V4PKdKP5Z4XK_Ki4`4_KYKa4e4gKUKY4\\OWIT3g2ZMR4BeId2b2cMi3JQJU2]2kMb30_Je1W2UNZ36mJV1Q2]NR3=[Kf0k1fNk2e0gK6`1TOi2f0ULFT1Dg2f0bLXOh02f2f0PMhN<b0d2g0]MXN0Q1c2g0Y2YOgMg0Y2YOgMh0X2WOiMi0W2WOiMi0W2WOiMi0W2WOiMj0V2VOjMj0V2VOjMj0V2VOjMj0V2UOkMl0T2ROnMn0R2QOoMo0Q2oNQNQ1o1nNRNS1m1kNUNU1k1iNWNW1i1hNXNX1h1fNZN[1e1dN\\N\\1e1aN]N_1c1`N^Na1a1]NaNc1_1\\NbNd1^1ZNdNf1\\1XNfNi1Y1VNhNj1X1TNjNl1V1SNkNm1U1QNnNo1Q1PNPOP2P1nMROR2n0mMTOR2l0lMVOU2i0iMZOV2f0iM[OW2e0gM^OY2a0fM@Z2S901O010O001O01O01O0010O01O0010O00010O01O0010O0001O010O001O01O01O0010O01O01O01O0010O01O00010O001O010O00010O001O001N101O000O2O001N101O000O2O001N101O000O2O001N101O000O2O001N101O000O2O001N101O000O2O001N101O001N10001N101O001N10001N101J5@afW3"}, "label": "a person with a light coat and patterned pants in a black and white photo with luggage"}]}
{"id": 309034, "image": "COCO_train2014_000000309034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with the luggage handle on their butt\"."}], "task": "refering", "answer_template": "The \"the person with the luggage handle on their butt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 25, 26, 27, 28, 29, 30, 31, 32, 33, 44, 45, 46, 47, 48, 49, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 179, 180, 181, 182, 199], "bbox": [0.4744791666666667, 0.020265625000000002, 0.9519583333333334, 0.5112656250000001], "iscrowd": 0, "area": 38223.71534999999, "rle": {"size": [640, 480], "counts": "c`^42mc05L3M3M2N1N101O00001O00001O0g^O]OU?d0j@BP?>PAGk>9TANf>3YA2b>N^A8\\>HdAa0S>@lAl0h=TOXBX1\\=iNcB_1U=aNkBd1P=]NnBj1l<VNTCP2f<QNYCT2b<lM^CZ2\\<gMcC^2X<bMhCf2P<[MoCn2g;SMYDV3^;kLaD^3V;bLiDj3l:WLSE\\4Z:eKeEd4R:\\KnEh4n9YKQFk4k9UKUFo4g9RKXFQ5e9oJZFR5f9oJYFQ5g9oJYFQ5g9PKXFP5h9PKXFP5h9QKWFo4i9QKVFP5j9QKUFo4k9QKUFo4k9RKTFn4l9RKTFn4l9RKTFn4l9RKSFo4m9RKRFn4n9RKRFn4n9RKRFn4n9RKRFn4n9RKQFo4o9RKPFn4o9SKQFm4o9SKPFn4P:RKPFn4P:RKPFn4P:SKoEm4Q:SKnEn4R:RKnEn4R:RKnEn4R:RKmEm4U:SKkEl4V:UKiEj4X:VKhEh4Z:XKeEh4\\:XKdEg4]:YKcEf4^:[K`Eo18QNZ:0^Ei0R2bNa8e0^EEh3QOk6Z1_EoNQ`0Q1_12Nd0\\Od0\\Od0\\Od0\\Oc0]Od0\\Od0[Oe0\\Oc0]Od0\\Od0\\O<D2N1O2N1O000000000000000000001O001O001O001O002N3M3M3M3M3M2N001O1O1O1O1O001OK5L4L4L4K5J6K5K5K5K5J6I6G:E;F:F:F:J6M3M3M3M3M3M3L4M3M3N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1I7H8I7H8K5O1N2N2N2N2O1N2N2N2O1N2N2M3M3L4M3M3L4O1000000O101O00000O2O00001O000O101J5E;An[>"}, "label": "the person with the luggage handle on their butt"}]}
{"id": 309034, "image": "COCO_train2014_000000309034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"handle of a trolley\"."}], "task": "refering", "answer_template": "The \"handle of a trolley\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 25, 26, 27, 28, 29, 30, 31, 32, 33, 44, 45, 46, 47, 48, 49, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 179, 180, 181, 182, 199], "bbox": [0.4744791666666667, 0.020265625000000002, 0.9519583333333334, 0.5112656250000001], "iscrowd": 0, "area": 38223.71534999999, "rle": {"size": [640, 480], "counts": "c`^42mc05L3M3M2N1N101O00001O00001O0g^O]OU?d0j@BP?>PAGk>9TANf>3YA2b>N^A8\\>HdAa0S>@lAl0h=TOXBX1\\=iNcB_1U=aNkBd1P=]NnBj1l<VNTCP2f<QNYCT2b<lM^CZ2\\<gMcC^2X<bMhCf2P<[MoCn2g;SMYDV3^;kLaD^3V;bLiDj3l:WLSE\\4Z:eKeEd4R:\\KnEh4n9YKQFk4k9UKUFo4g9RKXFQ5e9oJZFR5f9oJYFQ5g9oJYFQ5g9PKXFP5h9PKXFP5h9QKWFo4i9QKVFP5j9QKUFo4k9QKUFo4k9RKTFn4l9RKTFn4l9RKTFn4l9RKSFo4m9RKRFn4n9RKRFn4n9RKRFn4n9RKRFn4n9RKQFo4o9RKPFn4o9SKQFm4o9SKPFn4P:RKPFn4P:RKPFn4P:SKoEm4Q:SKnEn4R:RKnEn4R:RKnEn4R:RKmEm4U:SKkEl4V:UKiEj4X:VKhEh4Z:XKeEh4\\:XKdEg4]:YKcEf4^:[K`Eo18QNZ:0^Ei0R2bNa8e0^EEh3QOk6Z1_EoNQ`0Q1_12Nd0\\Od0\\Od0\\Od0\\Oc0]Od0\\Od0[Oe0\\Oc0]Od0\\Od0\\O<D2N1O2N1O000000000000000000001O001O001O001O002N3M3M3M3M3M2N001O1O1O1O1O001OK5L4L4L4K5J6K5K5K5K5J6I6G:E;F:F:F:J6M3M3M3M3M3M3L4M3M3N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1I7H8I7H8K5O1N2N2N2N2O1N2N2N2O1N2N2M3M3L4M3M3L4O1000000O101O00000O2O00001O000O101J5E;An[>"}, "label": "handle of a trolley"}]}
{"id": 538411, "image": "COCO_train2014_000000538411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lemon between a pear and a lime\"."}], "task": "refering", "answer_template": "The \"the lemon between a pear and a lime\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 188, 189, 190, 191], "bbox": [0.51, 0.368828125, 0.790281030444965, 0.5584374999999999], "iscrowd": 0, "area": 11812.723549999997, "rle": {"size": [640, 427], "counts": "o_X4k0Sc0T1lN3M2N3M3N1N3N2O1O001O1O001O1N2O001O1O001O1O1O001O1O001O1O1N101O1O001O1O1O001O1O001O1O1O0O2O1O001O1O1O001O1O001O1O1O00000O10000000000000000000000000000O10000000000000O2O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1N3M2N2N3M2N2N2N3M2N2N3TOW^O\\Oka0`0^^OYOda0c0e^OVO]a0g0l^OQOWa0j0YPh1"}, "label": "the lemon between a pear and a lime"}]}
{"id": 538411, "image": "COCO_train2014_000000538411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow lemon in front of the orange between the green fruit\"."}], "task": "refering", "answer_template": "The \"the yellow lemon in front of the orange between the green fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 188, 189, 190, 191], "bbox": [0.51, 0.368828125, 0.790281030444965, 0.5584374999999999], "iscrowd": 0, "area": 11812.723549999997, "rle": {"size": [640, 427], "counts": "o_X4k0Sc0T1lN3M2N3M3N1N3N2O1O001O1O001O1N2O001O1O001O1O1O001O1O001O1O1N101O1O001O1O1O001O1O001O1O1O0O2O1O001O1O1O001O1O001O1O1O00000O10000000000000000000000000000O10000000000000O2O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1N3M2N2N3M2N2N2N3M2N2N3TOW^O\\Oka0`0^^OYOda0c0e^OVO]a0g0l^OQOWa0j0YPh1"}, "label": "the yellow lemon in front of the orange between the green fruit"}]}
{"id": 259884, "image": "COCO_train2014_000000259884.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large bus\"."}], "task": "refering", "answer_template": "The \"the large bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.2561875, 0.16179166666666667, 0.9960625000000001, 0.7528125], "iscrowd": 0, "area": 84780.40490000002, "rle": {"size": [480, 640], "counts": "jS]2d0U=W1UOk0O1O2M2OO1O10O02M3M2N3N2M2O2N2N101O001O1N101O1O0B?L4M210O001O10O011O0O101O0O101O0O2O001O0O2O001N101O7I2M2O1O1O0O2O1O1O0O01O100O00100O01000O10O0100O12M3N1O2N1N3N1O1O0O10000O1000000O1000O10O1000000O1000O10O10001N2O1O1O1N2O1O1O1N3N1O1O1N01WO^DRNd;m1bDgMe;X2a001O1O001O1O10O2O1O1O1N2O1O1O1N2O2N2N5J2O001N101O001N101O001N3N2N000O0100000O01000O10O1000O10O1000O10O1000O1000O10O10000O10O1001N10001O0O101O2N1N3N1O2N1N5L3M2M2O1O2N1N2O2N1O2bGgKP6[4eIPLZ6S4aIPL^6S4]IQLa6R4ZIQLd6S4XIoKg6T4TIoKk6T4PIoKo6T4lHoKR7]4aHfK^7o5O1N101O001N101O000O2O001O0O101O001N10000O1000000O100000O01000000O10O100O001N2O001N2000O1000000O10001O0O1000000O10001N1000000O1000000O2O00000O1000000O2O00000O100000000O2O0000000O1000001O0O100000000O10001O000O100000000O10O100000O10O1000O1000O1000O100N1O2O1N2O1O010000000001N100000000O101O00000O10001O000O1000001O0O100000000O2O0000000O101O00000O1000001O0O100000001N100000000O2O0000000O10001O000O1000001O0O100000000O2O0000000O101O00000O10001O000O100000001N10000000001N100000000O101O00000O10001O000O1000001O0O10oY1"}, "label": "the large bus"}]}
{"id": 259884, "image": "COCO_train2014_000000259884.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a local mass transit buss\"."}], "task": "refering", "answer_template": "The \"a local mass transit buss\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.2561875, 0.16179166666666667, 0.9960625000000001, 0.7528125], "iscrowd": 0, "area": 84780.40490000002, "rle": {"size": [480, 640], "counts": "jS]2d0U=W1UOk0O1O2M2OO1O10O02M3M2N3N2M2O2N2N101O001O1N101O1O0B?L4M210O001O10O011O0O101O0O101O0O2O001O0O2O001N101O7I2M2O1O1O0O2O1O1O0O01O100O00100O01000O10O0100O12M3N1O2N1N3N1O1O0O10000O1000000O1000O10O1000000O1000O10O10001N2O1O1O1N2O1O1O1N3N1O1O1N01WO^DRNd;m1bDgMe;X2a001O1O001O1O10O2O1O1O1N2O1O1O1N2O2N2N5J2O001N101O001N101O001N3N2N000O0100000O01000O10O1000O10O1000O10O1000O1000O10O10000O10O1001N10001O0O101O2N1N3N1O2N1N5L3M2M2O1O2N1N2O2N1O2bGgKP6[4eIPLZ6S4aIPL^6S4]IQLa6R4ZIQLd6S4XIoKg6T4TIoKk6T4PIoKo6T4lHoKR7]4aHfK^7o5O1N101O001N101O000O2O001O0O101O001N10000O1000000O100000O01000000O10O100O001N2O001N2000O1000000O10001O0O1000000O10001N1000000O1000000O2O00000O1000000O2O00000O100000000O2O0000000O1000001O0O100000000O10001O000O100000000O10O100000O10O1000O1000O1000O100N1O2O1N2O1O010000000001N100000000O101O00000O10001O000O1000001O0O100000000O2O0000000O101O00000O1000001O0O100000001N100000000O2O0000000O10001O000O1000001O0O100000000O2O0000000O101O00000O10001O000O100000001N10000000001N100000000O101O00000O10001O000O1000001O0O10oY1"}, "label": "a local mass transit buss"}]}
{"id": 235316, "image": "COCO_train2014_000000235316.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a us army officer , named wilkins , cutting a cake with two other officers\"."}], "task": "refering", "answer_template": "The \"a us army officer , named wilkins , cutting a cake with two other officers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 241, 242, 243, 262, 263, 264, 265, 285, 286, 287, 288], "bbox": [0.379953125, 0.16441964285714286, 0.610140625, 0.7815401785714285], "iscrowd": 0, "area": 24281.574900000007, "rle": {"size": [448, 640], "counts": "^_Z31n=4L3L5L4L3bG]OZ3g0^L^O`3g0WL_Oh3d0QLAn3c0jKBU4a0dKDZ4a0`KB_4b0^K^Oa4e0]K[Ob4i0\\KVOd4l0ZKTOe4P1XKPOg4T1WKkNh4X1VKhNi4\\1TKdNl4^1SKaNl4c1QK]Nn4g1oJYNP5j1oJUNQ5n1lJRNS5Q2kJoMT5U2jJjMU5Z2hJfMW5]2gJcMY5`2eJ_MZ5e2cJ[M\\5h2bJXM]5l2`JTM`5n2_JQM`5S3]JmLb5V3\\JjLc5X3]JgLb5Z3^JfLb5[3]JeLb5]3^JbLa5_3_JaLe4_OWJR4S1_Lf4DRJm3Y1^Le4JlIj3^1\\Le40gIe3e1ZLd4a4\\K^Kd4b4^K\\K_4h4aKWK]4l4cKSKZ4P5hKnJU4V5kKiJS4Y5oKeJn3_5RL`Jk3c5VL\\Jh3g5YLWJd3l5]LSJb3o5^LPJb3P6`LnI_3T6V21O0O2O0O2O0O2O001RIdIU6]6kIdIT6]6kIeIR6]6lIeIS6[6mIfIQ6\\6nIeIQ6[6nIgIP6[6oIfIP6Z6oIhIP6Y6oIhIo5Y6QJhIn5Y6PJiIo5X6PJiIo5W6QJjIo5V6oIlI8^OT5f6dJmI5_OX5d6aJoI4@Z5a6bJPJ1B]5^6aJQJOD_5\\6`JRJNEa5Y6aJfJ_5Z5`JgJ_5Z5_JhJa5W5_JjJ`5k61O001N101O2N5J6K5K5J5L7H8I8H5J1O2L3M3L4M3K6E:I7O1O1N3N1N2O1O1N2O2M2O1O1N3N2N2M3N2M3L4L4M3L4L4L4L4L5K<D<D;E<E>AjU]3"}, "label": "a us army officer , named wilkins , cutting a cake with two other officers"}]}
{"id": 235316, "image": "COCO_train2014_000000235316.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a military man is cutting a cake with military men on both sides of him\"."}], "task": "refering", "answer_template": "The \"a military man is cutting a cake with military men on both sides of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 241, 242, 243, 262, 263, 264, 265, 285, 286, 287, 288], "bbox": [0.379953125, 0.16441964285714286, 0.610140625, 0.7815401785714285], "iscrowd": 0, "area": 24281.574900000007, "rle": {"size": [448, 640], "counts": "^_Z31n=4L3L5L4L3bG]OZ3g0^L^O`3g0WL_Oh3d0QLAn3c0jKBU4a0dKDZ4a0`KB_4b0^K^Oa4e0]K[Ob4i0\\KVOd4l0ZKTOe4P1XKPOg4T1WKkNh4X1VKhNi4\\1TKdNl4^1SKaNl4c1QK]Nn4g1oJYNP5j1oJUNQ5n1lJRNS5Q2kJoMT5U2jJjMU5Z2hJfMW5]2gJcMY5`2eJ_MZ5e2cJ[M\\5h2bJXM]5l2`JTM`5n2_JQM`5S3]JmLb5V3\\JjLc5X3]JgLb5Z3^JfLb5[3]JeLb5]3^JbLa5_3_JaLe4_OWJR4S1_Lf4DRJm3Y1^Le4JlIj3^1\\Le40gIe3e1ZLd4a4\\K^Kd4b4^K\\K_4h4aKWK]4l4cKSKZ4P5hKnJU4V5kKiJS4Y5oKeJn3_5RL`Jk3c5VL\\Jh3g5YLWJd3l5]LSJb3o5^LPJb3P6`LnI_3T6V21O0O2O0O2O0O2O001RIdIU6]6kIdIT6]6kIeIR6]6lIeIS6[6mIfIQ6\\6nIeIQ6[6nIgIP6[6oIfIP6Z6oIhIP6Y6oIhIo5Y6QJhIn5Y6PJiIo5X6PJiIo5W6QJjIo5V6oIlI8^OT5f6dJmI5_OX5d6aJoI4@Z5a6bJPJ1B]5^6aJQJOD_5\\6`JRJNEa5Y6aJfJ_5Z5`JgJ_5Z5_JhJa5W5_JjJ`5k61O001N101O2N5J6K5K5J5L7H8I8H5J1O2L3M3L4M3K6E:I7O1O1N3N1N2O1O1N2O2M2O1O1N3N2N2M3N2M3L4L4M3L4L4L4L4L5K<D<D;E<E>AjU]3"}, "label": "a military man is cutting a cake with military men on both sides of him"}]}
{"id": 235316, "image": "COCO_train2014_000000235316.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man furthest to the left\"."}], "task": "refering", "answer_template": "The \"the man furthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 32, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 214, 215, 216, 217, 232, 233, 234, 235, 238, 239, 240, 241, 255, 256, 257, 258, 262, 263, 264, 278, 279, 280, 281, 286, 287, 301, 302, 303], "bbox": [0.08021875, 0.047410714285714285, 0.51909375, 0.8159375000000001], "iscrowd": 0, "area": 46365.895800000006, "rle": {"size": [448, 640], "counts": "Rbf01m=8G9G9G8H9H8H8H7J7H8I7H7J7H8H8TFZLQ9m3hFTLV9S4dFnK[9\\4N2O1N2N2O1N2O1N2O1N2N2M3M3L4M3L4L4M3L4M3L4M3L4M3L4M3L4M3L4M3L4M3L4L4M3O1O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100OWOi0VOk0O001O0O101O000O2O001O0POlHTKU7l4TIkJm6S5^IcJb6]5Q1O0O101O000O2O001N10001N101O0O101O0O101O0O20O2N2M4L3N2M4L3N2M3M4L3N2M4L3M3N3L3M3N2N3M2N2N3N1N2N2N3M2N2N3M2N2N2N3M2N2N2N2O10O01O100O1O100O1O10O0fJeJP2Z5lMQKn1n4nM^Kk1b4SNgKf1X4YNPLa1o3^NZL[1f3dNaLV1^3iNfLU1Y3jNkLS1V3kNmLT1R3lNQMR1n2mNUMR1k2lNYMR1f2mN]MR1b2mNbMQ1_2mNcMR1^2kNeMT1\\2iNhMT1Z2iNiMV1Y2fNkMX1V2fNlMY1U2dNoMZ1R2cNQN\\1P2aNTN]1m1aNUN^1m1^NWN`1j1]NYNa1i1XN_Nf1c1oMgNP2Z1gMoNX2S1^MWO`2V6O2M2O2M2N3M2O2M2N3M2O0O1O1O010O1O00100O001O10O01O10O2O001O1N101O1N101NhNoCb0Q<]OPDd0n;\\ORDd0o;[ORDd0n;\\OSDd0m;[OTDd0l;]OTDb0m;]OTDc0k;]OVDb0k;]OVDb0j;^OWDa0j;^OWD?k;AVD=l;BUD<l;DUD:m;ETD9m;GTD6o;IRD5o;KRD3P<LQD1Q<OPDOR<OPDNR<1PDLS<2nCMS<3nCJU<4l0O2N1OjlV4"}, "label": "the man furthest to the left"}]}
{"id": 235316, "image": "COCO_train2014_000000235316.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall man in glasses to the left of two other men\"."}], "task": "refering", "answer_template": "The \"a tall man in glasses to the left of two other men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 32, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 214, 215, 216, 217, 232, 233, 234, 235, 238, 239, 240, 241, 255, 256, 257, 258, 262, 263, 264, 278, 279, 280, 281, 286, 287, 301, 302, 303], "bbox": [0.08021875, 0.047410714285714285, 0.51909375, 0.8159375000000001], "iscrowd": 0, "area": 46365.895800000006, "rle": {"size": [448, 640], "counts": "Rbf01m=8G9G9G8H9H8H8H7J7H8I7H7J7H8H8TFZLQ9m3hFTLV9S4dFnK[9\\4N2O1N2N2O1N2O1N2O1N2N2M3M3L4M3L4L4M3L4M3L4M3L4M3L4M3L4M3L4M3L4M3L4L4M3O1O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100OWOi0VOk0O001O0O101O000O2O001O0POlHTKU7l4TIkJm6S5^IcJb6]5Q1O0O101O000O2O001N10001N101O0O101O0O101O0O20O2N2M4L3N2M4L3N2M3M4L3N2M4L3M3N3L3M3N2N3M2N2N3N1N2N2N3M2N2N3M2N2N2N3M2N2N2N2O10O01O100O1O100O1O10O0fJeJP2Z5lMQKn1n4nM^Kk1b4SNgKf1X4YNPLa1o3^NZL[1f3dNaLV1^3iNfLU1Y3jNkLS1V3kNmLT1R3lNQMR1n2mNUMR1k2lNYMR1f2mN]MR1b2mNbMQ1_2mNcMR1^2kNeMT1\\2iNhMT1Z2iNiMV1Y2fNkMX1V2fNlMY1U2dNoMZ1R2cNQN\\1P2aNTN]1m1aNUN^1m1^NWN`1j1]NYNa1i1XN_Nf1c1oMgNP2Z1gMoNX2S1^MWO`2V6O2M2O2M2N3M2O2M2N3M2O0O1O1O010O1O00100O001O10O01O10O2O001O1N101O1N101NhNoCb0Q<]OPDd0n;\\ORDd0o;[ORDd0n;\\OSDd0m;[OTDd0l;]OTDb0m;]OTDc0k;]OVDb0k;]OVDb0j;^OWDa0j;^OWD?k;AVD=l;BUD<l;DUD:m;ETD9m;GTD6o;IRD5o;KRD3P<LQD1Q<OPDOR<OPDNR<1PDLS<2nCMS<3nCJU<4l0O2N1OjlV4"}, "label": "a tall man in glasses to the left of two other men"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown and colorful luggage\"."}], "task": "refering", "answer_template": "The \"brown and colorful luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [294, 309, 310, 311, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 346, 347, 348, 358, 359, 360, 361, 362, 363, 364, 365, 366, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.07272916666666666, 0.75584375, 0.658, 0.9870156250000001], "iscrowd": 0, "area": 19592.346550000002, "rle": {"size": [640, 480], "counts": "a]f02nc02M3N2M3N2M3M3N3L3N2M3N2M3M3N2M3N2M4M2M3M3N2M3N2M3N2M4L3N2M3N2M3N2M1O100O100O100O1O100O100O100O1O010O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O10jL]@o1b?oMb@P2^?mMf@R2Y?mMi@S2V?kMn@T2R?iMRAV2m>hMVAX2j>eMZAZ2e>dM^A\\2b>aMbA^2]>`MgA_2X>_MkAa2U>\\MoAc2P>\\MRBd2n=YMVBf2i=XM[Bg2_?01O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001N2O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O00[`V3"}, "label": "brown and colorful luggage"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red piece of luggage with it ' s handle extended , next to a black trash bag , and with a colorful cloth partially draped on it\"."}], "task": "refering", "answer_template": "The \"a red piece of luggage with it ' s handle extended , next to a black trash bag , and with a colorful cloth partially draped on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 93, 94, 95, 110, 111, 112, 127, 128, 129, 144, 145, 146, 161, 162, 163, 178, 179, 180, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 235, 246, 247, 248, 251, 252, 263, 264, 280], "bbox": [0.494375, 0.17528125, 0.8898958333333333, 0.7078593750000001], "iscrowd": 0, "area": 17563.24759999999, "rle": {"size": [640, 480], "counts": "TXd41`84h23dL]1Y2jNSMo1a2XNnLS2Q3TNiLl1V3[NdLf1Z3aN_L`1`3gNZLZ1d3mNULT1j3@cKa0[4`0eJ@]5]1cIdN_6Z2aHgM_7X3`GiLb8U4^FlKc9W7101N2O1oNVFiGk9V8XFgGj9X8WFgGj9W8XFgGi9X8ZFfGg9X8[FfGg9X8\\FfGe9Y8\\FeGe9Z8]FdGe9Z8^FdGc9Z8_FdGc9Z8_FeGa9Z8bFcG`9\\8aFcG`9[8cFbG^9]8dFbG]9\\8U1O1O0O2O1N3N3L4M3M3L4M3Lg5ZJ0O10000O10O0100O1000O0100O100O10O100000000O10000000000O100000000O10000000000O100000000O1000000001O001O001O00001O001O00001O001O001O000010O01O00001O001O101N1O1O1O1O1O1O1O2O0O1O1O10000O101N100O10000O100O101N10000O100O10001N100O100O10000O2O0O100O101O0O101N101O0O2O0O101N101O0O2O0O101N10kfP1"}, "label": "a red piece of luggage with it ' s handle extended , next to a black trash bag , and with a colorful cloth partially draped on it"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red piece of luggage in an airport with a bagtag and telescoping handle fully extended\"."}], "task": "refering", "answer_template": "The \"a red piece of luggage in an airport with a bagtag and telescoping handle fully extended\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 93, 94, 95, 110, 111, 112, 127, 128, 129, 144, 145, 146, 161, 162, 163, 178, 179, 180, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 235, 246, 247, 248, 251, 252, 263, 264, 280], "bbox": [0.494375, 0.17528125, 0.8898958333333333, 0.7078593750000001], "iscrowd": 0, "area": 17563.24759999999, "rle": {"size": [640, 480], "counts": "TXd41`84h23dL]1Y2jNSMo1a2XNnLS2Q3TNiLl1V3[NdLf1Z3aN_L`1`3gNZLZ1d3mNULT1j3@cKa0[4`0eJ@]5]1cIdN_6Z2aHgM_7X3`GiLb8U4^FlKc9W7101N2O1oNVFiGk9V8XFgGj9X8WFgGj9W8XFgGi9X8ZFfGg9X8[FfGg9X8\\FfGe9Y8\\FeGe9Z8]FdGe9Z8^FdGc9Z8_FdGc9Z8_FeGa9Z8bFcG`9\\8aFcG`9[8cFbG^9]8dFbG]9\\8U1O1O0O2O1N3N3L4M3M3L4M3Lg5ZJ0O10000O10O0100O1000O0100O100O10O100000000O10000000000O100000000O10000000000O100000000O1000000001O001O001O00001O001O00001O001O001O000010O01O00001O001O101N1O1O1O1O1O1O1O2O0O1O1O10000O101N100O10000O100O101N10000O100O10001N100O100O10000O2O0O100O101O0O101N101O0O2O0O101N101O0O2O0O101N10kfP1"}, "label": "a red piece of luggage in an airport with a bagtag and telescoping handle fully extended"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bag to the right of another red bag\"."}], "task": "refering", "answer_template": "The \"a red bag to the right of another red bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 101, 118, 134, 135, 151, 152, 168, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 234, 235, 236, 237, 252, 253, 254, 270, 271, 288], "bbox": [0.781875, 0.1831875, 0.9999375, 0.7136250000000001], "iscrowd": 0, "area": 14868.4882, "rle": {"size": [640, 480], "counts": "XgZ71ec0;XOh0G9L4L4L3M4O1N2O1N2N2O1N2O1N2N3N1N2N2O2M2O1N2N3N1N2O2M2O1N3N1O1N3N1N2O2N1N2O1N3N1O1N3N1N2O2N1N2O2M101N2O001N2O0O2O1N101O1O0M4F:F9G:F:F9G:G9H7H9HXKjCU2m;lMZDT2_;lMiDT2n:mMYET2^:mMjER2o9nMXFS2_9oMhFQ2P9oMWGQ2`8QNgGP2o7RNYHn1c7oMcHQ2[7jMkHX2T7aMTI_2k6[M[Ie2d6UMcIl2\\6nLiIT3V6eLoI]3Q6]LUJd3j5UL]Jl3b5nKdJR4\\5gKkJZ4T5bKPK_4o4bKoJ_4Q5aKoJ`4Q5_KPKa4o4`KPK`4P5`KQK`4n4`KRKa4m4`KSK_4n4`KRKa4m4_KGNaL"}, "label": "a red bag to the right of another red bag"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"luggage that is rust in color , with the handle raised . that is not beside the black bag\"."}], "task": "refering", "answer_template": "The \"luggage that is rust in color , with the handle raised . that is not beside the black bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 101, 118, 134, 135, 151, 152, 168, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 234, 235, 236, 237, 252, 253, 254, 270, 271, 288], "bbox": [0.781875, 0.1831875, 0.9999375, 0.7136250000000001], "iscrowd": 0, "area": 14868.4882, "rle": {"size": [640, 480], "counts": "XgZ71ec0;XOh0G9L4L4L3M4O1N2O1N2N2O1N2O1N2N3N1N2N2O2M2O1N2N3N1N2O2M2O1N3N1O1N3N1N2O2N1N2O1N3N1O1N3N1N2O2N1N2O2M101N2O001N2O0O2O1N101O1O0M4F:F9G:F:F9G:G9H7H9HXKjCU2m;lMZDT2_;lMiDT2n:mMYET2^:mMjER2o9nMXFS2_9oMhFQ2P9oMWGQ2`8QNgGP2o7RNYHn1c7oMcHQ2[7jMkHX2T7aMTI_2k6[M[Ie2d6UMcIl2\\6nLiIT3V6eLoI]3Q6]LUJd3j5UL]Jl3b5nKdJR4\\5gKkJZ4T5bKPK_4o4bKoJ_4Q5aKoJ`4Q5_KPKa4o4`KPK`4P5`KQK`4n4`KRKa4m4`KSK_4n4`KRKa4m4_KGNaL"}, "label": "luggage that is rust in color , with the handle raised . that is not beside the black bag"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black plastic bag in front that it tied on top\"."}], "task": "refering", "answer_template": "The \"the black plastic bag in front that it tied on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 110, 111, 112, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 227, 228, 229, 244, 245, 246, 262, 263, 279, 280], "bbox": [0.3791041666666667, 0.23690625, 0.792375, 0.7254375000000001], "iscrowd": 0, "area": 23619.3119, "rle": {"size": [640, 480], "counts": "foa3i0Uc0?@<E7I7I7I7I7I7I6J5K6K4L4L4j@WMn<P5M2N3M3M2N3M3M3M2N4L4L3M4M3L3N2N1O2M2O2N1N3mC\\Hf;n710O10OO2N2N1O2O0O2N2N1O2N2N1O1O1O0OiMYFZJf9X5QG`Jn8^5]GZJb8f5fGRJZ8m5l20O10001>Bc0^Oc0]Oc0\\Od0]Oc0\\Od0]Ob0^Oc0\\OjJ1PG\\OP9j1Q4T1lNU1jNa0@7H8I7H9H2OK4M4L3M4L3M4Lf0ZOY1hNW1iNX1hNYl4HQTKW1hNX1hNX1hNl0TO00O2N100O1O100O2O0O1O100O101PLg@V3[?hLg@V3Z?iLh@V3Z?gLg@Y3[?cLh@]3Y?`Li@`3X?]Li@d3X?YLj@h3g?000O12O3L4L4K2O2M2O1N3M2N2N3M2N3M2N2M4M2N2iN\\^ONWb0YOP^Of0eb001O00001O00001O0O2O00001O00001O001O001O1O100O1O010O100O1O010O1O100O10O01O100O1O010O1O2N4L5KRUn1"}, "label": "the black plastic bag in front that it tied on top"}]}
{"id": 382784, "image": "COCO_train2014_000000382784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"luggage wrapped in a black plastic bag closest to 2 red suitcases\"."}], "task": "refering", "answer_template": "The \"luggage wrapped in a black plastic bag closest to 2 red suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 110, 111, 112, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 227, 228, 229, 244, 245, 246, 262, 263, 279, 280], "bbox": [0.3791041666666667, 0.23690625, 0.792375, 0.7254375000000001], "iscrowd": 0, "area": 23619.3119, "rle": {"size": [640, 480], "counts": "foa3i0Uc0?@<E7I7I7I7I7I7I6J5K6K4L4L4j@WMn<P5M2N3M3M2N3M3M3M2N4L4L3M4M3L3N2N1O2M2O2N1N3mC\\Hf;n710O10OO2N2N1O2O0O2N2N1O2N2N1O1O1O0OiMYFZJf9X5QG`Jn8^5]GZJb8f5fGRJZ8m5l20O10001>Bc0^Oc0]Oc0\\Od0]Oc0\\Od0]Ob0^Oc0\\OjJ1PG\\OP9j1Q4T1lNU1jNa0@7H8I7H9H2OK4M4L3M4L3M4Lf0ZOY1hNW1iNX1hNYl4HQTKW1hNX1hNX1hNl0TO00O2N100O1O100O2O0O1O100O101PLg@V3[?hLg@V3Z?iLh@V3Z?gLg@Y3[?cLh@]3Y?`Li@`3X?]Li@d3X?YLj@h3g?000O12O3L4L4K2O2M2O1N3M2N2N3M2N3M2N2M4M2N2iN\\^ONWb0YOP^Of0eb001O00001O00001O0O2O00001O00001O001O001O1O100O1O010O100O1O010O1O100O10O01O100O1O010O1O2N4L5KRUn1"}, "label": "luggage wrapped in a black plastic bag closest to 2 red suitcases"}]}
{"id": 202567, "image": "COCO_train2014_000000202567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teen in beige pants standing behind a skateboard\"."}], "task": "refering", "answer_template": "The \"teen in beige pants standing behind a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 124, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 216, 217, 218, 240, 241, 263, 264, 286, 287, 309, 310, 332, 333, 355, 356], "bbox": [0.381546875, 0.18583044982698962, 0.5175156249999999, 0.7274048442906573], "iscrowd": 0, "area": 13144.803950000005, "rle": {"size": [578, 640], "counts": "V]Z41Pb0001O00=B3N2N2N2gBYOT9i0gF[OX9g0dF]OZ9e0bF^O^9c0]FBb9?ZFDf9=VFFi9=RFGm9:oEIQ:8lEJT:7hELW:6fEK[:6bEL^:5^EMc:5YELg:6UEMk:5PEMQ;4jD0V;2dD1\\;1_D2b;0XD2j;Q32MeGYK`3e4oK]KVLNk7c4oKbKTLMm7]4PLhKRLMm7Z4PLlKQLLn7Z4mKmKSLJP8[4iKmKVLIQ8]4dKmKYLGS8^4`KnK[LEU8_4[KoK_LCV8`4VKQLbL@X8b4PKRLgL\\OY8d4kJaLV5TOfGf3n2YM\\5hNRHV4T2VMj5[N_He4Z1SMX7U3YHoLj7]3dGfL`8R4`FRL`9X61O1O001O1O001O001O1O001O001O1O0O2O001O1O001OV1jN`0@1O1N2M3UN[DUMh;o1[EhMg:\\1\\FZNg9V1oFaNU9R1^GeNd8P1]4E;E;E^V^5"}, "label": "teen in beige pants standing behind a skateboard"}]}
{"id": 202567, "image": "COCO_train2014_000000202567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder wearing a light - colored hoodie and light - colored pants\"."}], "task": "refering", "answer_template": "The \"a skateboarder wearing a light - colored hoodie and light - colored pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 124, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 216, 217, 218, 240, 241, 263, 264, 286, 287, 309, 310, 332, 333, 355, 356], "bbox": [0.381546875, 0.18583044982698962, 0.5175156249999999, 0.7274048442906573], "iscrowd": 0, "area": 13144.803950000005, "rle": {"size": [578, 640], "counts": "V]Z41Pb0001O00=B3N2N2N2gBYOT9i0gF[OX9g0dF]OZ9e0bF^O^9c0]FBb9?ZFDf9=VFFi9=RFGm9:oEIQ:8lEJT:7hELW:6fEK[:6bEL^:5^EMc:5YELg:6UEMk:5PEMQ;4jD0V;2dD1\\;1_D2b;0XD2j;Q32MeGYK`3e4oK]KVLNk7c4oKbKTLMm7]4PLhKRLMm7Z4PLlKQLLn7Z4mKmKSLJP8[4iKmKVLIQ8]4dKmKYLGS8^4`KnK[LEU8_4[KoK_LCV8`4VKQLbL@X8b4PKRLgL\\OY8d4kJaLV5TOfGf3n2YM\\5hNRHV4T2VMj5[N_He4Z1SMX7U3YHoLj7]3dGfL`8R4`FRL`9X61O1O001O1O001O001O1O001O001O1O0O2O001O1O001OV1jN`0@1O1N2M3UN[DUMh;o1[EhMg:\\1\\FZNg9V1oFaNU9R1^GeNd8P1]4E;E;E^V^5"}, "label": "a skateboarder wearing a light - colored hoodie and light - colored pants"}]}
{"id": 202567, "image": "COCO_train2014_000000202567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blurry skateboard with the guy in black pants\"."}], "task": "refering", "answer_template": "The \"the blurry skateboard with the guy in black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [400, 401, 402, 403, 404, 405, 406, 407, 408, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 446, 447, 451, 452], "bbox": [0.299015625, 0.8190138408304498, 0.747546875, 0.9308650519031141], "iscrowd": 0, "area": 6479.427950000003, "rle": {"size": [578, 640], "counts": "dk\\31Qb00000000O10001O0O100000000O10001O000O1000001N10001O001O00001O001O00001O000000000000001O000000000000001O000000000000001N100000000Q_OCY`0=__OL``04__OMa`04]_ONb`02]_OOc`01]_O0b`00]_O1c`0O\\_O1e`0OZ_O2f`0NY_O3g`0d0000000000O10O1000000000O10000_OZ_OCi`0=[_O\\Oh`0c0<0000O10000O10000O10000O10O10O0100O010O10O0100O010O10O0100O010O100O11O00000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O000000000000000000O1000000000000O1000000000000000000000000000001O00000000000m^O_Oe`0a0V_OFh`0:S_OLk`0g0O2N1O1N3N1O001N101O1O0O2O001N101OO10O0100O1UOe_OOZ`02i_OJX`06k_OFV`0:k_OEU`0;l_ODS`0=n_OAS`0?n_O^OT`0b0f00O10000O100000O010O100O010O1O10O0100O10O01O100O010O2O0O1O2O0O2O0O101N1O1O2M2O1O2M2O1NWoj2"}, "label": "the blurry skateboard with the guy in black pants"}]}
{"id": 202567, "image": "COCO_train2014_000000202567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black skateboard with white wheels being ridden by a rider with black pants and white shoes\"."}], "task": "refering", "answer_template": "The \"a black skateboard with white wheels being ridden by a rider with black pants and white shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [400, 401, 402, 403, 404, 405, 406, 407, 408, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 446, 447, 451, 452], "bbox": [0.299015625, 0.8190138408304498, 0.747546875, 0.9308650519031141], "iscrowd": 0, "area": 6479.427950000003, "rle": {"size": [578, 640], "counts": "dk\\31Qb00000000O10001O0O100000000O10001O000O1000001N10001O001O00001O001O00001O000000000000001O000000000000001O000000000000001N100000000Q_OCY`0=__OL``04__OMa`04]_ONb`02]_OOc`01]_O0b`00]_O1c`0O\\_O1e`0OZ_O2f`0NY_O3g`0d0000000000O10O1000000000O10000_OZ_OCi`0=[_O\\Oh`0c0<0000O10000O10000O10000O10O10O0100O010O10O0100O010O10O0100O010O100O11O00000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O000000000000000000O1000000000000O1000000000000000000000000000001O00000000000m^O_Oe`0a0V_OFh`0:S_OLk`0g0O2N1O1N3N1O001N101O1O0O2O001N101OO10O0100O1UOe_OOZ`02i_OJX`06k_OFV`0:k_OEU`0;l_ODS`0=n_OAS`0?n_O^OT`0b0f00O10000O100000O010O100O010O1O10O0100O10O01O100O010O2O0O1O2O0O2O0O101N1O1O2M2O1O2M2O1NWoj2"}, "label": "a black skateboard with white wheels being ridden by a rider with black pants and white shoes"}]}
{"id": 202567, "image": "COCO_train2014_000000202567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blurry person on the skateboard with jeans and white shoes\"."}], "task": "refering", "answer_template": "The \"the blurry person on the skateboard with jeans and white shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 35, 36, 37, 38, 57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 356, 357, 358, 359, 360, 378, 379, 380, 381, 382, 383, 400, 401, 402, 403, 404, 405, 406, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.39778125000000003, 0.011228373702422146, 0.71640625, 0.8719031141868513], "iscrowd": 0, "area": 59022.04725, "rle": {"size": [578, 640], "counts": "[]`42oa01O1N2O2N1N2O1O1N2O2N100O100O100O2O0O100O100O2O0O100O100O101N100O100O101N100O100O101M2O1O1N2O1O1N3N1O1O1N2O1O2M2O1O1eHWN[Nk1a0ZO]Og09DE=3LK5N3OOJ:3GEb09_O@i0=YO[OP1c0QOVOW1h0iNQO`1l0bNoNe1n0\\NmNj1Q1WNjNo1m0YNnNm1RO2j03VMl1e2ZNnKR3m3d5K5L4K5K5K5L4K5K5K5L4K5K5M2010000O100001O0000000000000000000000000O1000000000000000000000000000001O000O10000000000000000000000000000000000O10000000000000001O00000000000000000O10000000006J6J6J6J5K6J6J6J6J5K6J6J6I7J5ROo0fNZ1eN[1fNZ1fNY1fN[1fNZ1fNZ1F:J5K6J6J6J6J5K6I7J6J6J5KfnU3"}, "label": "the blurry person on the skateboard with jeans and white shoes"}]}
{"id": 202567, "image": "COCO_train2014_000000202567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy in skateboard who seems to be moving\"."}], "task": "refering", "answer_template": "The \"the guy in skateboard who seems to be moving\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 35, 36, 37, 38, 57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 356, 357, 358, 359, 360, 378, 379, 380, 381, 382, 383, 400, 401, 402, 403, 404, 405, 406, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.39778125000000003, 0.011228373702422146, 0.71640625, 0.8719031141868513], "iscrowd": 0, "area": 59022.04725, "rle": {"size": [578, 640], "counts": "[]`42oa01O1N2O2N1N2O1O1N2O2N100O100O100O2O0O100O100O2O0O100O100O101N100O100O101N100O100O101M2O1O1N2O1O1N3N1O1O1N2O1O2M2O1O1eHWN[Nk1a0ZO]Og09DE=3LK5N3OOJ:3GEb09_O@i0=YO[OP1c0QOVOW1h0iNQO`1l0bNoNe1n0\\NmNj1Q1WNjNo1m0YNnNm1RO2j03VMl1e2ZNnKR3m3d5K5L4K5K5K5L4K5K5K5L4K5K5M2010000O100001O0000000000000000000000000O1000000000000000000000000000001O000O10000000000000000000000000000000000O10000000000000001O00000000000000000O10000000006J6J6J6J5K6J6J6J6J5K6J6J6I7J5ROo0fNZ1eN[1fNZ1fNY1fN[1fNZ1fNZ1F:J5K6J6J6J6J5K6I7J6J6J5KfnU3"}, "label": "the guy in skateboard who seems to be moving"}]}
{"id": 431947, "image": "COCO_train2014_000000431947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a umbrella near the green flag\"."}], "task": "refering", "answer_template": "The \"a umbrella near the green flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 167, 186], "bbox": [0.0, 0.3948036253776435, 0.307625, 0.7532326283987916], "iscrowd": 0, "area": 10184.692649999997, "rle": {"size": [331, 640], "counts": "e5o0W95K5N2O1O100O100O1O100O1O100O100O2N10000O100O100O10000O100O100O2O0O10000O100O100O1O100O1O2N100O1O1O1O100O1O1O00100O1O1VIbMf5_2RJiMm5X2jIQNV6n1cIZN\\6g1[IbNe6b200O1000000001YO\\IeMe6l1jISNV6n1jIQNW6n1jIQNW6n1jImM[6R2fIgM`6Y2j0O1O1N101O1O1O001O1O1O001O100O1O1O1O1O1O1O1O1O1O1O001O0000000000001O00000000001O00001O100O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O10O01O1O001O001O001O001O001O001O001O001O0010O01O001O0O2O001O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O2Nbo^4"}, "label": "a umbrella near the green flag"}]}
{"id": 431947, "image": "COCO_train2014_000000431947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left shade cover\"."}], "task": "refering", "answer_template": "The \"the left shade cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 167, 186], "bbox": [0.0, 0.3948036253776435, 0.307625, 0.7532326283987916], "iscrowd": 0, "area": 10184.692649999997, "rle": {"size": [331, 640], "counts": "e5o0W95K5N2O1O100O100O1O100O1O100O100O2N10000O100O100O10000O100O100O2O0O10000O100O100O1O100O1O2N100O1O1O1O100O1O1O00100O1O1VIbMf5_2RJiMm5X2jIQNV6n1cIZN\\6g1[IbNe6b200O1000000001YO\\IeMe6l1jISNV6n1jIQNW6n1jIQNW6n1jImM[6R2fIgM`6Y2j0O1O1N101O1O1O001O1O1O001O100O1O1O1O1O1O1O1O1O1O1O001O0000000000001O00000000001O00001O100O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O10O01O1O001O001O001O001O001O001O001O001O0010O01O001O0O2O001O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O2Nbo^4"}, "label": "the left shade cover"}]}
{"id": 431947, "image": "COCO_train2014_000000431947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large canopy in the middle of two other canopys next to a tall orange flag\"."}], "task": "refering", "answer_template": "The \"large canopy in the middle of two other canopys next to a tall orange flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 194, 217], "bbox": [0.326484375, 0.46447129909365564, 0.599328125, 0.7861631419939578], "iscrowd": 0, "area": 5098.337699999998, "rle": {"size": [331, 640], "counts": "WhS21X:200O100O100O100O100O100O101N100O1O100O100O100O100O100O100O100O100O100O100O1O2O0O1L4O1000000O1000000O100O100O100O100O100O100O100O100O100O100O2O0O100O100O100O1O100O1O100O1WHQOU6a22O0O2C>mNR101O1O010O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O010O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O001O001O002N2O0O1O001O1O1O1O1O1Ojkb2"}, "label": "large canopy in the middle of two other canopys next to a tall orange flag"}]}
{"id": 431947, "image": "COCO_train2014_000000431947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a straw umbrella between two others\"."}], "task": "refering", "answer_template": "The \"a straw umbrella between two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 194, 217], "bbox": [0.326484375, 0.46447129909365564, 0.599328125, 0.7861631419939578], "iscrowd": 0, "area": 5098.337699999998, "rle": {"size": [331, 640], "counts": "WhS21X:200O100O100O100O100O100O101N100O1O100O100O100O100O100O100O100O100O100O100O1O2O0O1L4O1000000O1000000O100O100O100O100O100O100O100O100O100O100O2O0O100O100O100O1O100O1O100O1WHQOU6a22O0O2C>mNR101O1O010O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O010O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O001O001O001O002N2O0O1O001O1O1O1O1O1Ojkb2"}, "label": "a straw umbrella between two others"}]}
{"id": 358223, "image": "COCO_train2014_000000358223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cake missing some slices\"."}], "task": "refering", "answer_template": "The \"a cake missing some slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 79, 80, 81, 82, 83, 85, 86, 87, 88, 89, 95, 96, 97, 98, 100, 101, 102, 103, 115, 116, 117], "bbox": [0.27791469194312796, 0.0095, 0.9501184834123224, 0.319765625], "iscrowd": 0, "area": 42375.02165, "rle": {"size": [640, 422], "counts": "[WY24fc0`0A?@`0A7I6I7J6K5N2M2O2N2M3N2N2M3N2M3N1O2M3N2N2M3N2N2O1N2N101N2O1N2O0O2N100O2O0O2N101N100O2N101N101N1O101N101O0O2O000O2O0O2O001N100O2O001N101O0O101N101O0O2O0O101O0O2O001N100O2O001N101O00000O10000000000O100000O10O0100O100O00100O100O0_NjAYMW>g2SBmLo=S3XBcLk=]3VB_Lm=a3TBZLP>e3RBWLQ>i3PBSLS>m3P100000000000001N1000000000000000000O1000000000000O2O000O100O10000O101N10000O16J:F9G9H8G9G8H2N1O2N1O2N101N1O2N1O2NO1000001O0000000O10000000000000000O100000000000000O10000000000O2O0000000O101O00001N10001O000O101O00001N10001O0O101N100O2O1N101O0O2O1N101N2O0O2O0O2O2M3N2N2M3N1N3M3K5K5L4K5K5L4L5K4M4K5K4M4K5K4M4K4M4K5K<E<C>B=Dmo<"}, "label": "a cake missing some slices"}]}
{"id": 358223, "image": "COCO_train2014_000000358223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cake missing two slices\"."}], "task": "refering", "answer_template": "The \"a cake missing two slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 79, 80, 81, 82, 83, 85, 86, 87, 88, 89, 95, 96, 97, 98, 100, 101, 102, 103, 115, 116, 117], "bbox": [0.27791469194312796, 0.0095, 0.9501184834123224, 0.319765625], "iscrowd": 0, "area": 42375.02165, "rle": {"size": [640, 422], "counts": "[WY24fc0`0A?@`0A7I6I7J6K5N2M2O2N2M3N2N2M3N2M3N1O2M3N2N2M3N2N2O1N2N101N2O1N2O0O2N100O2O0O2N101N100O2N101N101N1O101N101O0O2O000O2O0O2O001N100O2O001N101O0O101N101O0O2O0O101O0O2O001N100O2O001N101O00000O10000000000O100000O10O0100O100O00100O100O0_NjAYMW>g2SBmLo=S3XBcLk=]3VB_Lm=a3TBZLP>e3RBWLQ>i3PBSLS>m3P100000000000001N1000000000000000000O1000000000000O2O000O100O10000O101N10000O16J:F9G9H8G9G8H2N1O2N1O2N101N1O2N1O2NO1000001O0000000O10000000000000000O100000000000000O10000000000O2O0000000O101O00001N10001O000O101O00001N10001O0O101N100O2O1N101O0O2O1N101N2O0O2O0O2O2M3N2N2M3N1N3M3K5K5L4K5K5L4L5K4M4K5K4M4K5K4M4K4M4K5K<E<C>B=Dmo<"}, "label": "a cake missing two slices"}]}
{"id": 358223, "image": "COCO_train2014_000000358223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of cake on its side in front of another plate\"."}], "task": "refering", "answer_template": "The \"a piece of cake on its side in front of another plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 275, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 307, 308, 309, 310], "bbox": [0.39002369668246445, 0.6931093749999999, 0.7277962085308056, 0.877390625], "iscrowd": 0, "area": 12312.05935, "rle": {"size": [640, 422], "counts": "^cW3<cc03L5L3M3L3N3M2N2N3L3N2N2N3M2N2M4M2N2N3M2N2N3N1N2O1N3M2O1N3N1N2N3N1N2O1N2N2O0O2O1N1O2O0O1O100O10O01O100O100O100O10000O1000000O100000000O100000000O100000000001O001O0010O01O001O010O001O1O010O0O2O001O001O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1N2O1O2M4M3J6L4L4M4K5K6J4M2M3M3M3N2M3AR\\W2"}, "label": "a piece of cake on its side in front of another plate"}]}
{"id": 358223, "image": "COCO_train2014_000000358223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pastry sitting on a plate next to a blue fork\"."}], "task": "refering", "answer_template": "The \"a slice of pastry sitting on a plate next to a blue fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 275, 276, 277, 278, 279, 280, 291, 292, 293, 294, 295, 307, 308, 309, 310], "bbox": [0.39002369668246445, 0.6931093749999999, 0.7277962085308056, 0.877390625], "iscrowd": 0, "area": 12312.05935, "rle": {"size": [640, 422], "counts": "^cW3<cc03L5L3M3L3N3M2N2N3L3N2N2N3M2N2M4M2N2N3M2N2N3N1N2O1N3M2O1N3N1N2N3N1N2O1N2N2O0O2O1N1O2O0O1O100O10O01O100O100O100O10000O1000000O100000000O100000000O100000000001O001O0010O01O001O010O001O1O010O0O2O001O001O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1N2O1O2M4M3J6L4L4M4K5K6J4M2M3M3M3N2M3AR\\W2"}, "label": "a slice of pastry sitting on a plate next to a blue fork"}]}
{"id": 358223, "image": "COCO_train2014_000000358223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork is resting on the napkin\"."}], "task": "refering", "answer_template": "The \"the fork is resting on the napkin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 270, 271, 286, 287, 301, 302, 303, 317, 318], "bbox": [0.003933649289099526, 0.727265625, 0.24817535545023695, 0.9428593749999999], "iscrowd": 0, "area": 2999.684150000002, "rle": {"size": [640, 422], "counts": "bf1l0Uc02O0O00001O001O010O001O001O00010O000000001O01O0000000001O01O0000001O010O001O0O2O0O2O0O2O0O2O0O20O11O01X]OYOSb0g0g]O_OYb0a0a]OE`b0:Z]OLfb0h0001O1O1O2M2M3N0O010O01O010O0010O01O010O0010O010O00002M2O1O1N2O1O2M2O1N2O1O1N2O2N1N2O1O1N2O2N^UV6"}, "label": "the fork is resting on the napkin"}]}
{"id": 358223, "image": "COCO_train2014_000000358223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork with a blue handle sitting on a napkin\"."}], "task": "refering", "answer_template": "The \"a fork with a blue handle sitting on a napkin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 270, 271, 286, 287, 301, 302, 303, 317, 318], "bbox": [0.003933649289099526, 0.727265625, 0.24817535545023695, 0.9428593749999999], "iscrowd": 0, "area": 2999.684150000002, "rle": {"size": [640, 422], "counts": "bf1l0Uc02O0O00001O001O010O001O001O00010O000000001O01O0000000001O01O0000001O010O001O0O2O0O2O0O2O0O2O0O20O11O01X]OYOSb0g0g]O_OYb0a0a]OE`b0:Z]OLfb0h0001O1O1O2M2M3N0O010O01O010O0010O01O010O0010O010O00002M2O1O1N2O1O2M2O1N2O1O1N2O2N1N2O1O1N2O2N^UV6"}, "label": "a fork with a blue handle sitting on a napkin"}]}
{"id": 14160, "image": "COCO_train2014_000000014160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very dirty elephant with some sort of headgear on\"."}], "task": "refering", "answer_template": "The \"a very dirty elephant with some sort of headgear on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 45, 58, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 292, 293, 294, 295, 296, 297, 298, 309, 310, 317, 318, 319, 320, 321, 332, 342, 343, 344], "bbox": [0.379328125, 0.09213114754098362, 1.0, 0.9865105386416863], "iscrowd": 0, "area": 109355.08034999997, "rle": {"size": [427, 640], "counts": "caU32X=3L3M4M2M4L3N2M4L3N3L3M4M2M4L3^GaNU5b1aJgN\\5_1XJjNf5_1jIjNl5g1eIaNl5U2dITNl5c2dIfMm5P3dIXMl5\\5J6N2N2N2N2N2N2N2N2N2N2N4L3M4L3M4L3M4L3L4M4L3M4L3M4L3M4L3M1O1O1O1O1O1OM3L4M3M3L5M3O1N2O1O1O1N101O1O1N2O1O1N2O1O001N2O1O1O1N2O1000O10O10000O10000O10O1000O10000O100aM\\KWLd4h3dKPL]4o3kKjKU4U4SLdKm3[4[L^Ke3a4bLXK_3g4eLVK[3i4jLSKV3l4nLPKS3o4RMmJo2Q5UMlJk2S5ZMhJg2W5]MfJc2Y5bMcJ^2\\5fMaJZ2^5kM]JV2b5j2O1O2N1O1O1O1O1O1O1O1O2N11O0000000000001O01O000000000001O0001O00000001O000000001O001O1N2O001O1O1N101O1O1O001N2O1O001O1O2M2O2M3N2M2N3N2M2O2M3N2M2O2M3M2O2M3N2M2O2M3N2M3M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M2O1N101N101N100O2N101N101N10001O001N101O001N10001O0O2O001O0O101O1O001N101O001N2O001O0O2O1O001O0O2O001O1N101O001N2O001O001N101O1O0O2O001O0O2O1O001N101O001O0O2O00100O001O001O001O001O001O1O001O001O001O001O001O1O001O010O001O1O2N2N2N2N1O2N2N0000001O000000000000000000000001O000000000000000000000000000000000000010O000000001O000000001O000000001O000000001O000UJ\\4"}, "label": "a very dirty elephant with some sort of headgear on"}]}
{"id": 14160, "image": "COCO_train2014_000000014160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant whose ear is visible\"."}], "task": "refering", "answer_template": "The \"the elephant whose ear is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 45, 58, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 292, 293, 294, 295, 296, 297, 298, 309, 310, 317, 318, 319, 320, 321, 332, 342, 343, 344], "bbox": [0.379328125, 0.09213114754098362, 1.0, 0.9865105386416863], "iscrowd": 0, "area": 109355.08034999997, "rle": {"size": [427, 640], "counts": "caU32X=3L3M4M2M4L3N2M4L3N3L3M4M2M4L3^GaNU5b1aJgN\\5_1XJjNf5_1jIjNl5g1eIaNl5U2dITNl5c2dIfMm5P3dIXMl5\\5J6N2N2N2N2N2N2N2N2N2N2N4L3M4L3M4L3M4L3L4M4L3M4L3M4L3M4L3M1O1O1O1O1O1OM3L4M3M3L5M3O1N2O1O1O1N101O1O1N2O1O1N2O1O001N2O1O1O1N2O1000O10O10000O10000O10O1000O10000O100aM\\KWLd4h3dKPL]4o3kKjKU4U4SLdKm3[4[L^Ke3a4bLXK_3g4eLVK[3i4jLSKV3l4nLPKS3o4RMmJo2Q5UMlJk2S5ZMhJg2W5]MfJc2Y5bMcJ^2\\5fMaJZ2^5kM]JV2b5j2O1O2N1O1O1O1O1O1O1O1O2N11O0000000000001O01O000000000001O0001O00000001O000000001O001O1N2O001O1O1N101O1O1O001N2O1O001O1O2M2O2M3N2M2N3N2M2O2M3N2M2O2M3M2O2M3N2M2O2M3N2M3M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M2O1N101N101N100O2N101N101N10001O001N101O001N10001O0O2O001O0O101O1O001N101O001N2O001O0O2O1O001O0O2O001O1N101O001N2O001O001N101O1O0O2O001O0O2O1O001N101O001O0O2O00100O001O001O001O001O001O1O001O001O001O001O001O1O001O010O001O1O2N2N2N2N1O2N2N0000001O000000000000000000000001O000000000000000000000000000000000000010O000000001O000000001O000000001O000000001O000UJ\\4"}, "label": "the elephant whose ear is visible"}]}
{"id": 14160, "image": "COCO_train2014_000000014160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small part of a elephant visible with group of people posing back side along with another elephant\"."}], "task": "refering", "answer_template": "The \"small part of a elephant visible with group of people posing back side along with another elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 216, 238, 239, 261, 262, 285], "bbox": [0.24112499999999998, 0.2981030444964871, 0.41662499999999997, 0.833512880562061], "iscrowd": 0, "area": 10175.94285, "rle": {"size": [427, 640], "counts": "Q_P21W=2N3M3M2M4M3M3M2N3M3M3L3N3N2O1N2O1O1O1N2O1O1N2O1O1O2M2O1O1N3N2N2O0O2N2O1N010O1O010O10O01O010O101N100O2N100O2O0O2N100O2O0O1O2O0O101N101N1O2O0O2^GSMd6m2ZIVMd6l2[ITMd6n2ZITMe6l2[IUMc6m2[ITMd6m2\\ITMa6o2]IRMb6o2]ISMa6o2^IQMa6P3^IRM`6P3]IRMb6P3ZITMc6n2ZIUMe6m2XIVMf6m2UIWMi6l2RIWMm6l2nHXMP7P1dHV1Z7jNiHT1U7lNnHT1P7kNTIT1j6lNXIS1g6lN]IS1a6mNaIS1]6lNgIS1V6mNmIR1R6nNPJR1n5mNVJR1h5nNZJR1d5mN`JR1^5nNdJR1Z5mNjJS1S5jNRKV1l4gN[KW1c4fNiKo0Y4mNSL6b4CmKOY4E\\LLj3IoLBX32V5De]k4"}, "label": "small part of a elephant visible with group of people posing back side along with another elephant"}]}
{"id": 14160, "image": "COCO_train2014_000000014160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mostly obscured elephant\"."}], "task": "refering", "answer_template": "The \"the mostly obscured elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 189, 190, 191, 192, 193, 213, 214, 215, 216, 238, 239, 261, 262, 285], "bbox": [0.24112499999999998, 0.2981030444964871, 0.41662499999999997, 0.833512880562061], "iscrowd": 0, "area": 10175.94285, "rle": {"size": [427, 640], "counts": "Q_P21W=2N3M3M2M4M3M3M2N3M3M3L3N3N2O1N2O1O1O1N2O1O1N2O1O1O2M2O1O1N3N2N2O0O2N2O1N010O1O010O10O01O010O101N100O2N100O2O0O2N100O2O0O1O2O0O101N101N1O2O0O2^GSMd6m2ZIVMd6l2[ITMd6n2ZITMe6l2[IUMc6m2[ITMd6m2\\ITMa6o2]IRMb6o2]ISMa6o2^IQMa6P3^IRM`6P3]IRMb6P3ZITMc6n2ZIUMe6m2XIVMf6m2UIWMi6l2RIWMm6l2nHXMP7P1dHV1Z7jNiHT1U7lNnHT1P7kNTIT1j6lNXIS1g6lN]IS1a6mNaIS1]6lNgIS1V6mNmIR1R6nNPJR1n5mNVJR1h5nNZJR1d5mN`JR1^5nNdJR1Z5mNjJS1S5jNRKV1l4gN[KW1c4fNiKo0Y4mNSL6b4CmKOY4E\\LLj3IoLBX32V5De]k4"}, "label": "the mostly obscured elephant"}]}
{"id": 415570, "image": "COCO_train2014_000000415570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pile of brocolli on a blue plate\"."}], "task": "refering", "answer_template": "The \"a pile of brocolli on a blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 36, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 174, 175], "bbox": [0.368046875, 0.1011737089201878, 0.684421875, 0.5364788732394365], "iscrowd": 0, "area": 19895.416149999997, "rle": {"size": [426, 640], "counts": "\\XR32W=2N2N3M2M3N2N2N2O1N3M2N2N2N2N2N3M2O1N2N1O1O1O1O1O100O1O1O1O1O1O2O0O100O100O100O100O100O1_FUNc7k1[HXNc7i1\\HYNc7g1[H[Nd7f1[H\\Nd7d1ZH_Nd7b1ZHaNe7_1[HaNd7`1[HbNd7^1[HdNc7]1]HdNc7[1\\HgNc7Y1\\HhNe7W1ZHkNf7T1XHoNg7Q1VHSOj7l0SHXOm7h0PHZOP8h0kG\\OU8e0fG]O[8e0aG\\Oa8e0ZG\\Oi8e0RG]Oo8e0mF[OV9f0eF\\O]9V23N1N3N2M3N2O1O1O010O100O1O01O001O001N10001O001O00001O001O00001O001O00001O001O0N2O2N101O0O101N102M4M2M4M3M2M4M3L3N3L4M1N2O1O1N2O1N2O1N2001O001O1O001O0O2N1O2N2N1O2N1O2N2N1N3M2M4M2N3M3L3N3M2N3M3L4M4L3M3L4M3M2N1N2O1O1O1O001O1O1O1O1O1O10O01O3M4L4L4K5K5K5K5K5K5K5K5K5J`0AYlc2"}, "label": "a pile of brocolli on a blue plate"}]}
{"id": 415570, "image": "COCO_train2014_000000415570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cooked broccoli on a plate next to chips and rice\"."}], "task": "refering", "answer_template": "The \"cooked broccoli on a plate next to chips and rice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 36, 54, 55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 174, 175], "bbox": [0.368046875, 0.1011737089201878, 0.684421875, 0.5364788732394365], "iscrowd": 0, "area": 19895.416149999997, "rle": {"size": [426, 640], "counts": "\\XR32W=2N2N3M2M3N2N2N2O1N3M2N2N2N2N2N3M2O1N2N1O1O1O1O1O100O1O1O1O1O1O2O0O100O100O100O100O100O1_FUNc7k1[HXNc7i1\\HYNc7g1[H[Nd7f1[H\\Nd7d1ZH_Nd7b1ZHaNe7_1[HaNd7`1[HbNd7^1[HdNc7]1]HdNc7[1\\HgNc7Y1\\HhNe7W1ZHkNf7T1XHoNg7Q1VHSOj7l0SHXOm7h0PHZOP8h0kG\\OU8e0fG]O[8e0aG\\Oa8e0ZG\\Oi8e0RG]Oo8e0mF[OV9f0eF\\O]9V23N1N3N2M3N2O1O1O010O100O1O01O001O001N10001O001O00001O001O00001O001O00001O001O0N2O2N101O0O101N102M4M2M4M3M2M4M3L3N3L4M1N2O1O1N2O1N2O1N2001O001O1O001O0O2N1O2N2N1O2N1O2N2N1N3M2M4M2N3M3L3N3M2N3M3L4M4L3M3L4M3M2N1N2O1O1O1O001O1O1O1O1O1O10O01O3M4L4L4K5K5K5K5K5K5K5K5K5J`0AYlc2"}, "label": "cooked broccoli on a plate next to chips and rice"}]}
{"id": 210773, "image": "COCO_train2014_000000210773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lines man\"."}], "task": "refering", "answer_template": "The \"lines man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 138, 139, 140, 162, 163, 185, 186, 187, 208, 209, 210], "bbox": [0.006500000000000001, 0.015575, 0.175328125, 0.7142750000000001], "iscrowd": 0, "area": 13989.6877, "rle": {"size": [400, 640], "counts": "ae1k0h9P2M2N3M2M4M3SOjLZHX3b7lL\\HW3`7lL^HV3_7mL_HV3]7nL`HU3\\7nLaHU3\\7nLbHU3Z7nLdHU3W7oLgHS3V7PMhHS3T7PMjHR3S7Y1L4L3N3M3M2N3L3N3M3M2N3M3L3N3M2N3M3M2M4M3M2N3M2N3L4M2O2O1OcL\\KlNd4eM\\KY243_4`M`KZ257Z4ZMeK\\25:V4UMiK^25>Q4oLnK`25a0m3kLQLa27d0g3fLVLc27h0]4UOgKk0X4ROlKl0U4QOoKm0R4PORLn0o3oNULo0l3oNXLn0i3oN[Lo0f3nN^LP1c3mNaLQ1_3hNjLV1W3aNSM]1n2[N[Mc1f2TNeMi1\\2oMmMo1T2hMVNV2k1bM^N\\2m5N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2O1N2N3M2N2N2O1NTo]6"}, "label": "lines man"}]}
{"id": 210773, "image": "COCO_train2014_000000210773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the green shirt\"."}], "task": "refering", "answer_template": "The \"the man with the green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 138, 139, 140, 162, 163, 185, 186, 187, 208, 209, 210], "bbox": [0.006500000000000001, 0.015575, 0.175328125, 0.7142750000000001], "iscrowd": 0, "area": 13989.6877, "rle": {"size": [400, 640], "counts": "ae1k0h9P2M2N3M2M4M3SOjLZHX3b7lL\\HW3`7lL^HV3_7mL_HV3]7nL`HU3\\7nLaHU3\\7nLbHU3Z7nLdHU3W7oLgHS3V7PMhHS3T7PMjHR3S7Y1L4L3N3M3M2N3L3N3M3M2N3M3L3N3M2N3M3M2M4M3M2N3M2N3L4M2O2O1OcL\\KlNd4eM\\KY243_4`M`KZ257Z4ZMeK\\25:V4UMiK^25>Q4oLnK`25a0m3kLQLa27d0g3fLVLc27h0]4UOgKk0X4ROlKl0U4QOoKm0R4PORLn0o3oNULo0l3oNXLn0i3oN[Lo0f3nN^LP1c3mNaLQ1_3hNjLV1W3aNSM]1n2[N[Mc1f2TNeMi1\\2oMmMo1T2hMVNV2k1bM^N\\2m5N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2O1N2N3M2N2N2O1NTo]6"}, "label": "the man with the green shirt"}]}
{"id": 210773, "image": "COCO_train2014_000000210773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in white with a baseball cap on leaning down to the ground\"."}], "task": "refering", "answer_template": "The \"the man in white with a baseball cap on leaning down to the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 242, 243, 244, 245, 246, 247, 260, 261, 265, 266, 268, 269, 270, 271, 272, 273, 282, 283, 288, 289, 294, 295, 296], "bbox": [0.289328125, 0.41572499999999996, 0.9072968749999999, 0.9258500000000001], "iscrowd": 0, "area": 25620.01525, "rle": {"size": [400, 640], "counts": "`cX21]<5L3L5L4K2O1N2O1N2O1N2O1N2O1N2O1N2N2N2M3M3M3N1N3M[O[EEb:>^EB`:`0aE_O\\:d0eE[OX:h0iEWOT:l0lETOR:n0oEQOn9R1SFmNk9U1VFjNh9X1XFhNf9Z1[FbNf9`1[F\\Nf9g1ZFUNg9m1YFoMi9S291N2NF:2N1O2O001O001O0O101O001O001O001O0O101O001O001O00000O101N100O100O100O2O0O100O100O2O0O100O1XOWMjGi2V8\\MdGe2\\8`M_G_2a8gMYGZ2g8jMUGU2l8g00O100O100O10100O1O010O100O010O100O00100O10O0101N2O1N2O1N2N2O1N101O1O1O1O1O1O00O10001O0000000O1000000000000O1000000YIXMf3h2WL\\Mh3d2VL^Mj3b2TLaMj3a2SLaMm3_2QLdMn3\\2PLfMP4Z2mKiMS4W2hKoMW4R2cKSN]4S2XKSNf4T2nJRNR5S2dJRN\\5T2YJRNf5T2nIRNS6R2cISN]6P400000O101O0000000AaIlJ`6P5eIoJ\\6l4iISKW6i4mIWKS6e4RJZKn5b4WJ\\Kj5`4[J_Ke5]4_JcKa5_2mIdNg0lN]5_2^JnM;BW5`2_L_Mb3`2^L_Mc3a2^L^Mc3a2^L^Mc3a2W3O001O1O1O10O101O1O001N2O001O0O2O1O001O1N101O1O0O01O1O1O1O001O2N3M4N1N2N3M2O2M2N3M2N2O2M2N3M10O01O00100O001O10O01O001O1N1O2N2N101N2N1O2N100O1O2N100O1O1O1O100O00001O01O000001O01O000001O01O0001O01O000001O01O000001O01O0001O0001O01O0000010O000001O01O0000010O0000010O0001O01O010O001O010O1O010O0010O01O010O001O10O010O01O010O010O0001O001O001N10001O0O2O00001O0O2O001O000O2O001O0O101O001O0O21N2N2O1M3N2M3M3M3M2N3M3M3M3MQSg0"}, "label": "the man in white with a baseball cap on leaning down to the ground"}]}
{"id": 210773, "image": "COCO_train2014_000000210773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt playing tennis\"."}], "task": "refering", "answer_template": "The \"man in white shirt playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 242, 243, 244, 245, 246, 247, 260, 261, 265, 266, 268, 269, 270, 271, 272, 273, 282, 283, 288, 289, 294, 295, 296], "bbox": [0.289328125, 0.41572499999999996, 0.9072968749999999, 0.9258500000000001], "iscrowd": 0, "area": 25620.01525, "rle": {"size": [400, 640], "counts": "`cX21]<5L3L5L4K2O1N2O1N2O1N2O1N2O1N2O1N2N2N2M3M3M3N1N3M[O[EEb:>^EB`:`0aE_O\\:d0eE[OX:h0iEWOT:l0lETOR:n0oEQOn9R1SFmNk9U1VFjNh9X1XFhNf9Z1[FbNf9`1[F\\Nf9g1ZFUNg9m1YFoMi9S291N2NF:2N1O2O001O001O0O101O001O001O001O0O101O001O001O00000O101N100O100O100O2O0O100O100O2O0O100O1XOWMjGi2V8\\MdGe2\\8`M_G_2a8gMYGZ2g8jMUGU2l8g00O100O100O10100O1O010O100O010O100O00100O10O0101N2O1N2O1N2N2O1N101O1O1O1O1O1O00O10001O0000000O1000000000000O1000000YIXMf3h2WL\\Mh3d2VL^Mj3b2TLaMj3a2SLaMm3_2QLdMn3\\2PLfMP4Z2mKiMS4W2hKoMW4R2cKSN]4S2XKSNf4T2nJRNR5S2dJRN\\5T2YJRNf5T2nIRNS6R2cISN]6P400000O101O0000000AaIlJ`6P5eIoJ\\6l4iISKW6i4mIWKS6e4RJZKn5b4WJ\\Kj5`4[J_Ke5]4_JcKa5_2mIdNg0lN]5_2^JnM;BW5`2_L_Mb3`2^L_Mc3a2^L^Mc3a2^L^Mc3a2W3O001O1O1O10O101O1O001N2O001O0O2O1O001O1N101O1O0O01O1O1O1O001O2N3M4N1N2N3M2O2M2N3M2N2O2M2N3M10O01O00100O001O10O01O001O1N1O2N2N101N2N1O2N100O1O2N100O1O1O1O100O00001O01O000001O01O000001O01O0001O01O000001O01O000001O01O0001O0001O01O0000010O000001O01O0000010O0000010O0001O01O010O001O010O1O010O0010O01O010O001O10O010O01O010O010O0001O001O001N10001O0O2O00001O0O2O001O000O2O001O0O101O001O0O21N2N2O1M3N2M3M3M3M2N3M3M3M3MQSg0"}, "label": "man in white shirt playing tennis"}]}
{"id": 79701, "image": "COCO_train2014_000000079701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman is a rose colored gown standing next to a man\"."}], "task": "refering", "answer_template": "The \"a woman is a rose colored gown standing next to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 124, 125, 126, 127, 141, 142, 143, 144, 158, 159, 160, 175, 176, 177, 192, 193, 194, 209, 210, 211, 226, 227, 228, 242, 243, 244, 245, 260, 261, 262, 263, 277, 278, 279, 280, 294, 295, 296, 297, 311, 312, 313, 314, 328, 329, 330, 331, 345, 346, 347, 362, 363, 364, 380, 381], "bbox": [0.2844166666666667, 0.2708125, 0.49935416666666665, 0.9726874999999999], "iscrowd": 0, "area": 33180.242000000006, "rle": {"size": [640, 480], "counts": "]Pf27ac08I7N3gBYOV7h0SH0j72_Ge0`8]OhF[1V9hNTFm1j9XNbEX2\\:kMSEf2k:[MfDT3k8cKSH[1aNb3Z9WKRFm0c0<AQ4g9XLbFf5\\9]JaFe5]9hJVFZ5h9a2SOQFeGQ:o7]FmGe9T8\\FhGf9m7mEkG=5h9n7oElG:2j9o7QFmG6Nm9U8oElGf:c7lDiHa0Bb:b7REjHc;T7_DkHa;T7bDiH_;W7bDgH_;X7cDfH_;n6nDoHS;m6R1M2O02O0O2O10O4M2N2N1O1O2N1O1N3J5O1N2O2N1N2O1O000O01000O100000UOQJeCo5\\<TJ`Cl5b<UJ[Cl5e<VJXCj5i<XJRCk5n<WJlBm5T=?mN\\IkDd6S;`IkD`6T;dIiD[6W;hIgDX6W;kIgDV6X;lIfDT6Z;nIdDS6Z;QJbDQ6];QJaDo5^;TJ`Dm5_;UJ_Dn5^;UJ^Do5^;UJ_Dl5W2WIc6V8[IPH^6S8`ISHY6o7eIWHlNVOV6g8lJXHjNTOU6g8oJXHjNTOR6i8PKVHlNV1m3h6TMUHkN]1g3c6ZMSHhNg1e3Y6`MSHdNP2c3R6eMmKT2U4gMSLS2Q4hMRLV2S4cMRLk1VJbMj9HYF>j5U2ZJ_Ml9<fLZ1j3ZNZE5k7b0k2VObE3bc001N_\\f4"}, "label": "a woman is a rose colored gown standing next to a man"}]}
{"id": 79701, "image": "COCO_train2014_000000079701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing pink gown and black scraf standing near a person\"."}], "task": "refering", "answer_template": "The \"a lady wearing pink gown and black scraf standing near a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 124, 125, 126, 127, 141, 142, 143, 144, 158, 159, 160, 175, 176, 177, 192, 193, 194, 209, 210, 211, 226, 227, 228, 242, 243, 244, 245, 260, 261, 262, 263, 277, 278, 279, 280, 294, 295, 296, 297, 311, 312, 313, 314, 328, 329, 330, 331, 345, 346, 347, 362, 363, 364, 380, 381], "bbox": [0.2844166666666667, 0.2708125, 0.49935416666666665, 0.9726874999999999], "iscrowd": 0, "area": 33180.242000000006, "rle": {"size": [640, 480], "counts": "]Pf27ac08I7N3gBYOV7h0SH0j72_Ge0`8]OhF[1V9hNTFm1j9XNbEX2\\:kMSEf2k:[MfDT3k8cKSH[1aNb3Z9WKRFm0c0<AQ4g9XLbFf5\\9]JaFe5]9hJVFZ5h9a2SOQFeGQ:o7]FmGe9T8\\FhGf9m7mEkG=5h9n7oElG:2j9o7QFmG6Nm9U8oElGf:c7lDiHa0Bb:b7REjHc;T7_DkHa;T7bDiH_;W7bDgH_;X7cDfH_;n6nDoHS;m6R1M2O02O0O2O10O4M2N2N1O1O2N1O1N3J5O1N2O2N1N2O1O000O01000O100000UOQJeCo5\\<TJ`Cl5b<UJ[Cl5e<VJXCj5i<XJRCk5n<WJlBm5T=?mN\\IkDd6S;`IkD`6T;dIiD[6W;hIgDX6W;kIgDV6X;lIfDT6Z;nIdDS6Z;QJbDQ6];QJaDo5^;TJ`Dm5_;UJ_Dn5^;UJ^Do5^;UJ_Dl5W2WIc6V8[IPH^6S8`ISHY6o7eIWHlNVOV6g8lJXHjNTOU6g8oJXHjNTOR6i8PKVHlNV1m3h6TMUHkN]1g3c6ZMSHhNg1e3Y6`MSHdNP2c3R6eMmKT2U4gMSLS2Q4hMRLV2S4cMRLk1VJbMj9HYF>j5U2ZJ_Ml9<fLZ1j3ZNZE5k7b0k2VObE3bc001N_\\f4"}, "label": "a lady wearing pink gown and black scraf standing near a person"}]}
{"id": 79701, "image": "COCO_train2014_000000079701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black suit\"."}], "task": "refering", "answer_template": "The \"a man wearing a black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 127, 128, 129, 144, 145, 146, 147, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 280, 281, 282, 283, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 348, 349, 350, 351, 352, 365, 366, 367, 368, 369, 382, 383, 385, 386], "bbox": [0.45447916666666666, 0.272015625, 0.7464166666666666, 0.985328125], "iscrowd": 0, "area": 42779.39659999999, "rle": {"size": [640, 480], "counts": "\\`X41mc05L5K4L5J5L5K5n]OQ1j>TOk@[1k>lNi@b1o>R2E=C6K1N3M2O1N3M2O1N3N8bGeIf2d6oL_IP3j6eLYI[3o6ZLTIe3U7PLnHP4Z7eKiHZ4a7YKcHg4e7nJ^HS3XOnKc8c0XH_3XOjKi8<RHj3WOhKP92lGU4LRKT;n4REkJo:U5VEfJj:Z5YEcJg:\\5]E`Jc:a5bEZJ]:g5hETJW:m5n1O01O1O010O1O1O0O2N2M3N1O2N2M2O2N2N3L4N2M3N2N2N2M3N1O2N2M3N2gLY3<D<D;E5K5K4L5K5K4L5O1O0100O0O2O00N1L5L4L1N2O1O0O2O1O1N101O1N10O10O010O1O0010O01O001O00fKUH`Lk7Y2QH_J>U3a7[2_HkI;g3V7]2fI`MZ6_2jI^MV6a2nI\\MR6d2QJYMQ6d2]JoLf5m2QK]LS5^3eKkK^4Q4QL`KS4\\4TL\\KQ4`4ULYKd4o3cKiKa4T4fKeK]4X4jK`KZ4^4g5K4L4K5K5L4K:G?@?A`0Aol[2"}, "label": "a man wearing a black suit"}]}
{"id": 79701, "image": "COCO_train2014_000000079701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald man in a suit\"."}], "task": "refering", "answer_template": "The \"a bald man in a suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 127, 128, 129, 144, 145, 146, 147, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 280, 281, 282, 283, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 348, 349, 350, 351, 352, 365, 366, 367, 368, 369, 382, 383, 385, 386], "bbox": [0.45447916666666666, 0.272015625, 0.7464166666666666, 0.985328125], "iscrowd": 0, "area": 42779.39659999999, "rle": {"size": [640, 480], "counts": "\\`X41mc05L5K4L5J5L5K5n]OQ1j>TOk@[1k>lNi@b1o>R2E=C6K1N3M2O1N3M2O1N3N8bGeIf2d6oL_IP3j6eLYI[3o6ZLTIe3U7PLnHP4Z7eKiHZ4a7YKcHg4e7nJ^HS3XOnKc8c0XH_3XOjKi8<RHj3WOhKP92lGU4LRKT;n4REkJo:U5VEfJj:Z5YEcJg:\\5]E`Jc:a5bEZJ]:g5hETJW:m5n1O01O1O010O1O1O0O2N2M3N1O2N2M2O2N2N3L4N2M3N2N2N2M3N1O2N2M3N2gLY3<D<D;E5K5K4L5K5K4L5O1O0100O0O2O00N1L5L4L1N2O1O0O2O1O1N101O1N10O10O010O1O0010O01O001O00fKUH`Lk7Y2QH_J>U3a7[2_HkI;g3V7]2fI`MZ6_2jI^MV6a2nI\\MR6d2QJYMQ6d2]JoLf5m2QK]LS5^3eKkK^4Q4QL`KS4\\4TL\\KQ4`4ULYKd4o3cKiKa4T4fKeK]4X4jK`KZ4^4g5K4L4K5K5L4K:G?@?A`0Aol[2"}, "label": "a bald man in a suit"}]}
{"id": 522074, "image": "COCO_train2014_000000522074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman with brown hair holding a donut\"."}], "task": "refering", "answer_template": "The \"a young woman with brown hair holding a donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 54, 55, 72, 73, 90, 91, 92, 93, 108, 109, 110, 126, 127, 128, 144, 145, 162, 163, 180, 198, 199, 216, 217], "bbox": [0.0, 0.18816, 0.18364, 1.0], "iscrowd": 0, "area": 12879.938900000001, "rle": {"size": [375, 500], "counts": "Z8]3W2S60000000000000000000000001O0000000000000000001O000RNYNlIh1l5aNRJ_1f5kNXJV1o0gMX2\\1gLm0P1oMk1]1SMe0Q1UN^1_1_M<S1\\NT1^1fM7U1bNl0[1mM3V1gNf0[1RNOW1kN`0Z1WNLX1oN9Z1]NGZ1SO3Z1aND[1WOM]41dKL_44`KKc44[KLh43VKLo42PKMS51lJOX5MhJ2\\5ClJ=_5YO`Jg0n5lNQJT1Y6bNgI]1`6]N`Ic1Y800O10000O1000O10O1000001N2O1O1O0O2O3L5L3L5E:O1N1O10O0100O1O010O1O100O1O1N2O1O1N5L]Xe4"}, "label": "a young woman with brown hair holding a donut"}]}
{"id": 522074, "image": "COCO_train2014_000000522074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with a donut\"."}], "task": "refering", "answer_template": "The \"a women with a donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 54, 55, 72, 73, 90, 91, 92, 93, 108, 109, 110, 126, 127, 128, 144, 145, 162, 163, 180, 198, 199, 216, 217], "bbox": [0.0, 0.18816, 0.18364, 1.0], "iscrowd": 0, "area": 12879.938900000001, "rle": {"size": [375, 500], "counts": "Z8]3W2S60000000000000000000000001O0000000000000000001O000RNYNlIh1l5aNRJ_1f5kNXJV1o0gMX2\\1gLm0P1oMk1]1SMe0Q1UN^1_1_M<S1\\NT1^1fM7U1bNl0[1mM3V1gNf0[1RNOW1kN`0Z1WNLX1oN9Z1]NGZ1SO3Z1aND[1WOM]41dKL_44`KKc44[KLh43VKLo42PKMS51lJOX5MhJ2\\5ClJ=_5YO`Jg0n5lNQJT1Y6bNgI]1`6]N`Ic1Y800O10000O1000O10O1000001N2O1O1O0O2O3L5L3L5E:O1N1O10O0100O1O010O1O100O1O1N2O1O1N5L]Xe4"}, "label": "a women with a donut"}]}
{"id": 522074, "image": "COCO_train2014_000000522074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady with black jacket\"."}], "task": "refering", "answer_template": "The \"the lady with black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 58, 59, 60, 76, 77, 78, 94, 95, 96, 97, 112, 113, 114, 115, 131, 132, 146, 147, 148, 149, 150, 165, 166, 167, 168, 183, 184, 185, 186, 201, 202, 203, 204, 219, 220, 221, 222], "bbox": [0.15780000000000002, 0.20258666666666666, 0.41298, 0.9844266666666666], "iscrowd": 0, "area": 21165.406699999996, "rle": {"size": [375, 500], "counts": "jUm03_;7H7M3O1O1O1O1O1N2O1O1RHVOm4k0nJYOR5h0iJ\\OW5e0dJ_O\\5c0^JAb5`0YJDf5?TJEl5<oIHQ6:jIIV69dIK\\67^IMb65XIOh63RI1m62mH2S70gH4Y7NaH6_7L\\H7d7n1XNTLeKl3Z4ULfKk3Y4WLeKj3S2VLZN2Ch3i1cL`NGFf3e1kLaN@Jf3c1nL^N_ONc3b1SM\\N[O2b3a1WMYNYO5`3a1[MUNWO:^3`1^MSNVO;]3a1aMPNTO>[3`1eMnMROb0Y3_1fMnMSOb0W3_1hMmMROd0V3^1iMmMSOd0T3^1kMkMSOg0R3\\1mMlMSOf0Q3]1nMkMROh0P3\\1oMjMTOi0m2]1PNhMTOk0l2]1QNfMUOl0j2^1QNdMWOn0h2]1SNcMWOo0f2^1SNbMYOo0d2_1TN_MZOQ1c2`1SN^M\\OQ1a2a1H_N8a1I^N7a1J_N6a1K^N5b1K^N5b1L\\N5d1K\\N5d1L[N4e1L[N4d1N[N2e1NYN4g1MQN:o1GiM`0X2_OaMh0`2W4001O1O1O001O1O1O001O1O001O1O1O001O1M3_O`0L5M3M3M5K5K5K6J5K5K6J5K5K6[N`JiLe5n2iJhL\\5o2QKhLT5P3YKfLm4P3aKfLd4Q3W2G9H6I8I6L5K4L5J5L5K4L5K4K6K4L5K4K6Ko_[3"}, "label": "the lady with black jacket"}]}
{"id": 522074, "image": "COCO_train2014_000000522074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with blonde hair and a black shirt is standing near a display of donuts\"."}], "task": "refering", "answer_template": "The \"a woman with blonde hair and a black shirt is standing near a display of donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 58, 59, 60, 76, 77, 78, 94, 95, 96, 97, 112, 113, 114, 115, 131, 132, 146, 147, 148, 149, 150, 165, 166, 167, 168, 183, 184, 185, 186, 201, 202, 203, 204, 219, 220, 221, 222], "bbox": [0.15780000000000002, 0.20258666666666666, 0.41298, 0.9844266666666666], "iscrowd": 0, "area": 21165.406699999996, "rle": {"size": [375, 500], "counts": "jUm03_;7H7M3O1O1O1O1O1N2O1O1RHVOm4k0nJYOR5h0iJ\\OW5e0dJ_O\\5c0^JAb5`0YJDf5?TJEl5<oIHQ6:jIIV69dIK\\67^IMb65XIOh63RI1m62mH2S70gH4Y7NaH6_7L\\H7d7n1XNTLeKl3Z4ULfKk3Y4WLeKj3S2VLZN2Ch3i1cL`NGFf3e1kLaN@Jf3c1nL^N_ONc3b1SM\\N[O2b3a1WMYNYO5`3a1[MUNWO:^3`1^MSNVO;]3a1aMPNTO>[3`1eMnMROb0Y3_1fMnMSOb0W3_1hMmMROd0V3^1iMmMSOd0T3^1kMkMSOg0R3\\1mMlMSOf0Q3]1nMkMROh0P3\\1oMjMTOi0m2]1PNhMTOk0l2]1QNfMUOl0j2^1QNdMWOn0h2]1SNcMWOo0f2^1SNbMYOo0d2_1TN_MZOQ1c2`1SN^M\\OQ1a2a1H_N8a1I^N7a1J_N6a1K^N5b1K^N5b1L\\N5d1K\\N5d1L[N4e1L[N4d1N[N2e1NYN4g1MQN:o1GiM`0X2_OaMh0`2W4001O1O1O001O1O1O001O1O001O1O1O001O1M3_O`0L5M3M3M5K5K5K6J5K5K6J5K5K6[N`JiLe5n2iJhL\\5o2QKhLT5P3YKfLm4P3aKfLd4Q3W2G9H6I8I6L5K4L5J5L5K4L5K4K6K4L5K4K6Ko_[3"}, "label": "a woman with blonde hair and a black shirt is standing near a display of donuts"}]}
{"id": 522074, "image": "COCO_train2014_000000522074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green purse\"."}], "task": "refering", "answer_template": "The \"a green purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 129, 130, 131, 147, 148, 149, 150, 167], "bbox": [0.15936, 0.4991466666666667, 0.37304000000000004, 0.7426400000000001], "iscrowd": 0, "area": 2745.2581499999997, "rle": {"size": [375, 500], "counts": "j`m02c;4L3M3M3M2N2N2N2O1N2N2L4L4M3N2O1N2N101N2O1N2N21O001O001O00001O001O001O001O0O2O001O001O001O001O001O001O001O0O2O00001O001N[FQOd8o0\\GROd8m0[GUOd8k0\\GVOd8=hGDW83RHNn7KXH6d9001O000O1000O1O1O1N2N2O_P41^oK3M3M3M3M3M2N0000M7L4M2N3N1N2Nchb3"}, "label": "a green purse"}]}
{"id": 522074, "image": "COCO_train2014_000000522074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green leather purse under woman ' s arm\"."}], "task": "refering", "answer_template": "The \"a green leather purse under woman ' s arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 129, 130, 131, 147, 148, 149, 150, 167], "bbox": [0.15936, 0.4991466666666667, 0.37304000000000004, 0.7426400000000001], "iscrowd": 0, "area": 2745.2581499999997, "rle": {"size": [375, 500], "counts": "j`m02c;4L3M3M3M2N2N2N2O1N2N2L4L4M3N2O1N2N101N2O1N2N21O001O001O00001O001O001O001O0O2O001O001O001O001O001O001O001O0O2O00001O001N[FQOd8o0\\GROd8m0[GUOd8k0\\GVOd8=hGDW83RHNn7KXH6d9001O000O1000O1O1O1N2N2O_P41^oK3M3M3M3M3M2N0000M7L4M2N3N1N2Nchb3"}, "label": "a green leather purse under woman ' s arm"}]}
{"id": 440154, "image": "COCO_train2014_000000440154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop the gray haired woman is looking at\"."}], "task": "refering", "answer_template": "The \"the laptop the gray haired woman is looking at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 185, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 260, 280, 281, 282], "bbox": [0.065734375, 0.43370833333333336, 0.342140625, 0.7303333333333334], "iscrowd": 0, "area": 13788.144400000001, "rle": {"size": [480, 640], "counts": "`lc02n>4L5L3L4L4L5K4L4L4L5K4M3L4L5K4L4L4L4L5K4M3L4L5K4L4L4L3M01O01O00001O00000010O0001O00001O000010O000001O00001O00010O0000001O00001O01O01O00001O00000010O0001O00001O0000010O00001O00001O00010O000002O3L3M4M3L4L3N3L4L4M2M4L4L4M2M4L4M3L3M010O001O001O001N101O0O2O00001N101O0O2O001N101O0O2O001N10001N101O0O2O001O0O2O001N10001N101O0O2O001N101O0O2O001O0O101O3L5L4L4KgPU6"}, "label": "the laptop the gray haired woman is looking at"}]}
{"id": 440154, "image": "COCO_train2014_000000440154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black labtop\"."}], "task": "refering", "answer_template": "The \"a black labtop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 185, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 260, 280, 281, 282], "bbox": [0.065734375, 0.43370833333333336, 0.342140625, 0.7303333333333334], "iscrowd": 0, "area": 13788.144400000001, "rle": {"size": [480, 640], "counts": "`lc02n>4L5L3L4L4L5K4L4L4L5K4M3L4L5K4L4L4L4L5K4M3L4L5K4L4L4L3M01O01O00001O00000010O0001O00001O000010O000001O00001O00010O0000001O00001O01O01O00001O00000010O0001O00001O0000010O00001O00001O00010O000002O3L3M4M3L4L3N3L4L4M2M4L4L4M2M4L4M3L3M010O001O001O001N101O0O2O00001N101O0O2O001N101O0O2O001N10001N101O0O2O001O0O2O001N10001N101O0O2O001N101O0O2O001O0O101O3L5L4L4KgPU6"}, "label": "a black labtop"}]}
{"id": 137052, "image": "COCO_train2014_000000137052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant stands beside another\"."}], "task": "refering", "answer_template": "The \"an elephant stands beside another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267], "bbox": [0.612453125, 0.3323541666666667, 0.8844531250000001, 0.6665833333333334], "iscrowd": 0, "area": 17559.772449999997, "rle": {"size": [480, 640], "counts": "\\Qh55e>:G:I5hLSOUHQ1h7QOUHR1i7POTHR1k7PORHR1n7oNnGT1Q8oNkGR1U8POhGe0fMWOa:6fGb0nMVO\\:9cG:]NZOo9?`GMoNB`9c0^GJi99SFEQ:=kECV:?gEAZ:b0bE^O_:d0]E]Oe:e0UE_Oj:d0RECi:>VEGd:g0oDBh:k0kD^Ok:c0UEGa::^EJ]:8bEH^:8bEH]::bEF^::bEF^::aEG^::bEE_:;aEE^:<bED^:<bEC^:>bEA_:>bEA^:?cE@^:?cEA\\:`0dE_O]:`0dE_O\\:b0dE]O]:b0cEkNK[Ob:k1cEfN3XO]:Q2aEdN8UOY:X2_EcNU;\\1kDoNj:R1VEWOa:h0`E_OY:a0gE@X:`0gE@Z:`0fE@Z:`0fEAY:`0fE@Z:`0fE@[:?dEB\\:>dEC[:=dED\\:<dED\\:<cEE]:<aEF_:<^EDb:V1aEmM_:Q2jEhMV:X2`FSM_9m2Z10001O00000000001N100000000000VEUM`9j2oEhMP:X2^EZNb:S3O001O1O01O00000O100O1M3M4M2M3N3L3N2M4M[E_Lo9c3mEaLR:e3fE\\LZ:k3^EVLb:T41O01O01O00001O01O0001O0kM[EOe:N_E2`:KeE3[:NfE0Z:0gEeNg0Bb9j1hE[No0IY9m1iEQNW11o8n1dGPN]8P2Y2001O1O001O001O10OO2O001O0O2O001O0O2O1O001O0010O0I7J75J6K7H9H?B:E1O1001O4L4L4K5L3ROSD`NR<S1_DdNf;n0hDkN_;h0lDTOZ;?m1Cb^R1"}, "label": "an elephant stands beside another"}]}
{"id": 137052, "image": "COCO_train2014_000000137052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the darker colored elephant facing to the right\"."}], "task": "refering", "answer_template": "The \"the darker colored elephant facing to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267], "bbox": [0.612453125, 0.3323541666666667, 0.8844531250000001, 0.6665833333333334], "iscrowd": 0, "area": 17559.772449999997, "rle": {"size": [480, 640], "counts": "\\Qh55e>:G:I5hLSOUHQ1h7QOUHR1i7POTHR1k7PORHR1n7oNnGT1Q8oNkGR1U8POhGe0fMWOa:6fGb0nMVO\\:9cG:]NZOo9?`GMoNB`9c0^GJi99SFEQ:=kECV:?gEAZ:b0bE^O_:d0]E]Oe:e0UE_Oj:d0RECi:>VEGd:g0oDBh:k0kD^Ok:c0UEGa::^EJ]:8bEH^:8bEH]::bEF^::bEF^::aEG^::bEE_:;aEE^:<bED^:<bEC^:>bEA_:>bEA^:?cE@^:?cEA\\:`0dE_O]:`0dE_O\\:b0dE]O]:b0cEkNK[Ob:k1cEfN3XO]:Q2aEdN8UOY:X2_EcNU;\\1kDoNj:R1VEWOa:h0`E_OY:a0gE@X:`0gE@Z:`0fE@Z:`0fEAY:`0fE@Z:`0fE@[:?dEB\\:>dEC[:=dED\\:<dED\\:<cEE]:<aEF_:<^EDb:V1aEmM_:Q2jEhMV:X2`FSM_9m2Z10001O00000000001N100000000000VEUM`9j2oEhMP:X2^EZNb:S3O001O1O01O00000O100O1M3M4M2M3N3L3N2M4M[E_Lo9c3mEaLR:e3fE\\LZ:k3^EVLb:T41O01O01O00001O01O0001O0kM[EOe:N_E2`:KeE3[:NfE0Z:0gEeNg0Bb9j1hE[No0IY9m1iEQNW11o8n1dGPN]8P2Y2001O1O001O001O10OO2O001O0O2O001O0O2O1O001O0010O0I7J75J6K7H9H?B:E1O1001O4L4L4K5L3ROSD`NR<S1_DdNf;n0hDkN_;h0lDTOZ;?m1Cb^R1"}, "label": "the darker colored elephant facing to the right"}]}
{"id": 137052, "image": "COCO_train2014_000000137052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant that the man is walking and guiding\"."}], "task": "refering", "answer_template": "The \"the elephant that the man is walking and guiding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 266, 267, 286, 287, 289, 290], "bbox": [0.3595625, 0.34225, 0.702375, 0.7674583333333334], "iscrowd": 0, "area": 22263.756249999995, "rle": {"size": [480, 640], "counts": "QP\\34j>;D<E9G4M2O1N2N3M2O1N2Nd0]O7H7I8I6I8I6J5J5L3M4K4N3LhM^E5_:JdE5[:JhE4Y:JiE6V:HnE6R:JoE6P:JQF5P:JRF3n9NSFSOk:m0VEZNb;g1j01000000O101O1N2O1O1O1N2O001O1O1O1O1O1O1O1O001O1O1O1O1O1O3M5K6J5K6Jo0nEdKb7b4THcKk7d4kG_KU8g4aG]K^8j4YGZKf8`501O001O0O2O000000O01000O010O10O01000O0100G9UOk0fNZ1001O001O001O00001O0O2O001O001O0001O000O10000O100O101O0O10000O1001O100O1O1O100O2gFTLZ7m3`HXLa7h3YH]Lg7d3RHbLo7^3kGgLU8R51N10001O001O001O0O2O1O1N2O1O1O1N2O1O1eNZFYMh9^2dF^M]9Z2nFbMS9U2ZGeMh8U2aGgM`8W2eGcM^8Z2hG_M]8]2iG[M]8b2R2M4L3N3L4L3N3L3M4L3N3L4L3M4Mg0XO4L4L3O000000001O0000O1000O10O10O10O1O1O001O1O1O001O11O00001N1O1O1O2MfYi2"}, "label": "the elephant that the man is walking and guiding"}]}
{"id": 137052, "image": "COCO_train2014_000000137052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant in the sun\"."}], "task": "refering", "answer_template": "The \"elephant in the sun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 266, 267, 286, 287, 289, 290], "bbox": [0.3595625, 0.34225, 0.702375, 0.7674583333333334], "iscrowd": 0, "area": 22263.756249999995, "rle": {"size": [480, 640], "counts": "QP\\34j>;D<E9G4M2O1N2N3M2O1N2Nd0]O7H7I8I6I8I6J5J5L3M4K4N3LhM^E5_:JdE5[:JhE4Y:JiE6V:HnE6R:JoE6P:JQF5P:JRF3n9NSFSOk:m0VEZNb;g1j01000000O101O1N2O1O1O1N2O001O1O1O1O1O1O1O1O001O1O1O1O1O1O3M5K6J5K6Jo0nEdKb7b4THcKk7d4kG_KU8g4aG]K^8j4YGZKf8`501O001O0O2O000000O01000O010O10O01000O0100G9UOk0fNZ1001O001O001O00001O0O2O001O001O0001O000O10000O100O101O0O10000O1001O100O1O1O100O2gFTLZ7m3`HXLa7h3YH]Lg7d3RHbLo7^3kGgLU8R51N10001O001O001O0O2O1O1N2O1O1O1N2O1O1eNZFYMh9^2dF^M]9Z2nFbMS9U2ZGeMh8U2aGgM`8W2eGcM^8Z2hG_M]8]2iG[M]8b2R2M4L3N3L4L3N3L3M4L3N3L4L3M4Mg0XO4L4L3O000000001O0000O1000O10O10O10O1O1O001O1O1O001O11O00001N1O1O1O2MfYi2"}, "label": "elephant in the sun"}]}
{"id": 366430, "image": "COCO_train2014_000000366430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car where three persons are standing near it\"."}], "task": "refering", "answer_template": "The \"a black car where three persons are standing near it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 175, 177, 178, 179, 200, 201, 202, 203, 220, 223, 224, 225, 226, 246, 247, 248], "bbox": [0.55378125, 0.37116666666666664, 0.8552500000000001, 0.6194791666666667], "iscrowd": 0, "area": 6897.0140999999985, "rle": {"size": [480, 640], "counts": "cSV5<d>4L4L4L00O1O2O0O1O100O2N3K4L4LXn1NlQN6K5K4L1N10O1000jMDgE8^:K_EMi:5P22N2N2N2N2N4L3M4L3MB>O100O1O1O1O1O1N6JlT5MYkJ2M3N2M4N1N5KO2N100O2M2O1N3N1N2O2M2Ol`89e^G7N2N2N2O1dN]OSDk0m;A[Ck0e<h0O100000000]O[NkCe1W<a01O1O2N1O1O1O2N1O1O2N1O1O2N1O1O0000000010O1O1O2N1O101N1O1O1O0001O0001O00000001O000000000O2O000000000O101O00000N2N2N3M2O1N2N2O1O2O0O10000O105K9F:G2N1N2O1L4I7IW[[1"}, "label": "a black car where three persons are standing near it"}]}
{"id": 366430, "image": "COCO_train2014_000000366430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car that the woman in purple is standing in front of\"."}], "task": "refering", "answer_template": "The \"the car that the woman in purple is standing in front of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 175, 177, 178, 179, 200, 201, 202, 203, 220, 223, 224, 225, 226, 246, 247, 248], "bbox": [0.55378125, 0.37116666666666664, 0.8552500000000001, 0.6194791666666667], "iscrowd": 0, "area": 6897.0140999999985, "rle": {"size": [480, 640], "counts": "cSV5<d>4L4L4L00O1O2O0O1O100O2N3K4L4LXn1NlQN6K5K4L1N10O1000jMDgE8^:K_EMi:5P22N2N2N2N2N4L3M4L3MB>O100O1O1O1O1O1N6JlT5MYkJ2M3N2M4N1N5KO2N100O2M2O1N3N1N2O2M2Ol`89e^G7N2N2N2O1dN]OSDk0m;A[Ck0e<h0O100000000]O[NkCe1W<a01O1O2N1O1O1O2N1O1O2N1O1O2N1O1O0000000010O1O1O2N1O101N1O1O1O0001O0001O00000001O000000000O2O000000000O101O00000N2N2N3M2O1N2N2O1O2O0O10000O105K9F:G2N1N2O1L4I7IW[[1"}, "label": "the car that the woman in purple is standing in front of"}]}
{"id": 366430, "image": "COCO_train2014_000000366430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ford truck parked next to a dark blue chevy truck\"."}], "task": "refering", "answer_template": "The \"a ford truck parked next to a dark blue chevy truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 250, 252], "bbox": [0.725375, 0.3996875, 0.99975, 0.6051041666666667], "iscrowd": 0, "area": 8777.717699999997, "rle": {"size": [480, 640], "counts": "RUj61o>1O1O1O1O1O1O1O1O1O1O1O1O1N100000O100002N1O2N2N1O2N0O10001O0000000000000001O10O01O1O001OO10000000001O00000O1000000000001O1O1N101O1O1O001O1O1O1N101O1Od0\\O0001OTCZN^<f1_CcN[<S2N2N2N2N2N1O1O2O0O1O1O1O000000000001O00001O001N101O0O101O1O0O2O1O0O2O1M3N1O2M3L3M4000O10O1O1O001O1O001O1O001O1O1O000000000000001O0000000001O2N2O0O2N1O1O001O001O010O0000000000O101N1O100_Oa0\\Od0\\OdI"}, "label": "a ford truck parked next to a dark blue chevy truck"}]}
{"id": 366430, "image": "COCO_train2014_000000366430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ford suv next to a chevy truck , both are backed in\"."}], "task": "refering", "answer_template": "The \"the ford suv next to a chevy truck , both are backed in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 250, 252], "bbox": [0.725375, 0.3996875, 0.99975, 0.6051041666666667], "iscrowd": 0, "area": 8777.717699999997, "rle": {"size": [480, 640], "counts": "RUj61o>1O1O1O1O1O1O1O1O1O1O1O1O1N100000O100002N1O2N2N1O2N0O10001O0000000000000001O10O01O1O001OO10000000001O00000O1000000000001O1O1N101O1O1O001O1O1O1N101O1Od0\\O0001OTCZN^<f1_CcN[<S2N2N2N2N2N1O1O2O0O1O1O1O000000000001O00001O001N101O0O101O1O0O2O1O0O2O1M3N1O2M3L3M4000O10O1O1O001O1O001O1O001O1O1O000000000000001O0000000001O2N2O0O2N1O1O001O001O010O0000000000O101N1O100_Oa0\\Od0\\OdI"}, "label": "the ford suv next to a chevy truck , both are backed in"}]}
{"id": 366430, "image": "COCO_train2014_000000366430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ford suv\"."}], "task": "refering", "answer_template": "The \"the ford suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 252], "bbox": [0.72809375, 0.39325, 1.0, 0.6112291666666666], "iscrowd": 0, "area": 9656.678100000006, "rle": {"size": [480, 640], "counts": "PSk62n>1O1O1N2O1O2N1O1O1O1O1O2N1O1N2O1O1O2N0000001O000O10001O0000001O0001O010O001O0010O0000001O0000001O000000001O0000001O0000001O0000001O000000001O000O102Nc0]Om0SO2N3M2N3M2N3M1O000000000000000000000000O10000000000000001O1O001O1O1N2O1O001O1N2M3N2N2M2O2N2M3N2N2O1O001O1O100O1O001O00000000001O00000000000001O100O1O1O001O1O1O1O100O1O00000000000001O00000M3UOk0VOmI"}, "label": "the ford suv"}]}
{"id": 366430, "image": "COCO_train2014_000000366430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray ford suv\"."}], "task": "refering", "answer_template": "The \"the gray ford suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 252], "bbox": [0.72809375, 0.39325, 1.0, 0.6112291666666666], "iscrowd": 0, "area": 9656.678100000006, "rle": {"size": [480, 640], "counts": "PSk62n>1O1O1N2O1O2N1O1O1O1O1O2N1O1N2O1O1O2N0000001O000O10001O0000001O0001O010O001O0010O0000001O0000001O000000001O0000001O0000001O0000001O000000001O000O102Nc0]Om0SO2N3M2N3M2N3M1O000000000000000000000000O10000000000000001O1O001O1O1N2O1O001O1N2M3N2N2M2O2N2M3N2N2O1O001O1O100O1O001O00000000001O00000000000001O100O1O1O001O1O1O1O100O1O00000000000001O00000M3UOk0VOmI"}, "label": "the gray ford suv"}]}
{"id": 46942, "image": "COCO_train2014_000000046942.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown hooded jacket and jeans pulling a black suitcase\"."}], "task": "refering", "answer_template": "The \"a man in a brown hooded jacket and jeans pulling a black suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 24, 25, 40, 41, 42, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 89, 90, 91, 92, 93, 94, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 158, 159, 160, 161, 162, 175, 176, 177, 178, 179, 192, 193, 194, 209, 210, 226, 227, 243, 244], "bbox": [0.27368750000000003, 0.022640625, 0.580875, 0.65421875], "iscrowd": 0, "area": 33618.4028, "rle": {"size": [640, 480], "counts": "ePb21lc04L5K5J6K4L5K5K5J6K4L5K5J6Km0SOa0_O7J5M4K4oBRLi9S4mEYLm9k3iEaLP:c3fEiLT:^3`EnLZ:Z3WERMc:V3PEVMi:S3iDYMQ;P6J7I6J7J6I6J6K1N1O2O0O2N101N1O2N101N1O2O0O2N101N1O2O0O2N1O2O0O2N10G8ROn0ROo0ZOe00010O00N3M2N3M2O2M2N3M2N3N1N3O0010O010O010O010O0010O010O010O0010O010O010O010O0100O2OlHSEU5l:jJVEV5j:iJXEV5g:iJ\\EV5g:eJ[E[5h:aJZE^5i:]JYEc5j:XJYEg5j:UJXEj5k:QJWEP6k:kIXET6k:hIVEX6m:cIVE\\6m:`IUE_6n:\\ITEd6o:WITEh6R<01O0O2J6J5K6J5L5J6J5QNdAUNb>c1]B`Mh=Y2T2H8H7I8I6I8H7I8H8H7J7H7I8HPam3"}, "label": "a man in a brown hooded jacket and jeans pulling a black suitcase"}]}
{"id": 46942, "image": "COCO_train2014_000000046942.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person carring a luggage\"."}], "task": "refering", "answer_template": "The \"person carring a luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 24, 25, 40, 41, 42, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 89, 90, 91, 92, 93, 94, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 158, 159, 160, 161, 162, 175, 176, 177, 178, 179, 192, 193, 194, 209, 210, 226, 227, 243, 244], "bbox": [0.27368750000000003, 0.022640625, 0.580875, 0.65421875], "iscrowd": 0, "area": 33618.4028, "rle": {"size": [640, 480], "counts": "ePb21lc04L5K5J6K4L5K5K5J6K4L5K5J6Km0SOa0_O7J5M4K4oBRLi9S4mEYLm9k3iEaLP:c3fEiLT:^3`EnLZ:Z3WERMc:V3PEVMi:S3iDYMQ;P6J7I6J7J6I6J6K1N1O2O0O2N101N1O2N101N1O2O0O2N101N1O2O0O2N1O2O0O2N10G8ROn0ROo0ZOe00010O00N3M2N3M2O2M2N3M2N3N1N3O0010O010O010O010O0010O010O010O0010O010O010O010O0100O2OlHSEU5l:jJVEV5j:iJXEV5g:iJ\\EV5g:eJ[E[5h:aJZE^5i:]JYEc5j:XJYEg5j:UJXEj5k:QJWEP6k:kIXET6k:hIVEX6m:cIVE\\6m:`IUE_6n:\\ITEd6o:WITEh6R<01O0O2J6J5K6J5L5J6J5QNdAUNb>c1]B`Mh=Y2T2H8H7I8I6I8H7I8H8H7J7H7I8HPam3"}, "label": "person carring a luggage"}]}
{"id": 325472, "image": "COCO_train2014_000000325472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and a brown horse standing in a field\"."}], "task": "refering", "answer_template": "The \"a white and a brown horse standing in a field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 219, 220, 221, 222, 223, 224, 225, 243, 246, 247, 248, 249, 266, 269, 270, 271, 289], "bbox": [0.41360937499999995, 0.44927107061503413, 0.836578125, 0.7626195899772209], "iscrowd": 0, "area": 13807.781400000002, "rle": {"size": [439, 640], "counts": "Tka37^=3M3M3M3M3N2M3M2N1O2N2N2N0000O11O1O2N1O2N1O20O01O010O010O0011N3N1N2O1N100O100O100O2O0O100O11O1O100O1O1O1O1O100O001O1O1O1O01N100000000O101O00000O1000001O000001O0001O01O00001O0001O01O00000iDkNl9V1TFnNg9R1YFSOc9l0]FZO]9g0bF^OY9b0hFBS9>mFHm88SGMh84WG1d8O]G0c80]GOd81\\GOe81ZGOf81[GNe81\\GNe82[GNe82[GNe82nG[OR8e0`201N100000000O10001O000O10000000001N01000000O0100000O1000O01000000O1000O10O1000000O10000O1000000O1000000O1000000O2O2N2M2O2N2N1N3N2N1O1N1000O010000001O05K6`ElMR9X2eFlM[9Z2[FjMe9\\2QFhMo9Q30O000010ON2M4M2M3N3L3NiEWMd9^30010O0001WNjEOV:ORFLm91\\FIe93eFF[97nFQOd9m0d1M3M3M3N2M3M3N22N4L4L4L5[DnN[:V1]EoNb:U1VEPOi:T1nDROQ;k1O10O1M3N2M4M2MVElMV:R2_EZNb:c1TEiNm:P21O1O1N3N1O1O1O1O1O1O2N1O1O1O1O1N3N4L3M4L3M4L3M4L3I8I6I8H7Jhg\\1"}, "label": "a white and a brown horse standing in a field"}]}
{"id": 325472, "image": "COCO_train2014_000000325472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse and a foal\"."}], "task": "refering", "answer_template": "The \"a horse and a foal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 219, 220, 221, 222, 223, 224, 225, 243, 246, 247, 248, 249, 266, 269, 270, 271, 289], "bbox": [0.41360937499999995, 0.44927107061503413, 0.836578125, 0.7626195899772209], "iscrowd": 0, "area": 13807.781400000002, "rle": {"size": [439, 640], "counts": "Tka37^=3M3M3M3M3N2M3M2N1O2N2N2N0000O11O1O2N1O2N1O20O01O010O010O0011N3N1N2O1N100O100O100O2O0O100O11O1O100O1O1O1O1O100O001O1O1O1O01N100000000O101O00000O1000001O000001O0001O01O00001O0001O01O00000iDkNl9V1TFnNg9R1YFSOc9l0]FZO]9g0bF^OY9b0hFBS9>mFHm88SGMh84WG1d8O]G0c80]GOd81\\GOe81ZGOf81[GNe81\\GNe82[GNe82[GNe82nG[OR8e0`201N100000000O10001O000O10000000001N01000000O0100000O1000O01000000O1000O10O1000000O10000O1000000O1000000O1000000O2O2N2M2O2N2N1N3N2N1O1N1000O010000001O05K6`ElMR9X2eFlM[9Z2[FjMe9\\2QFhMo9Q30O000010ON2M4M2M3N3L3NiEWMd9^30010O0001WNjEOV:ORFLm91\\FIe93eFF[97nFQOd9m0d1M3M3M3N2M3M3N22N4L4L4L5[DnN[:V1]EoNb:U1VEPOi:T1nDROQ;k1O10O1M3N2M4M2MVElMV:R2_EZNb:c1TEiNm:P21O1O1N3N1O1O1O1O1O1O2N1O1O1O1O1N3N4L3M4L3M4L3M4L3I8I6I8H7Jhg\\1"}, "label": "a horse and a foal"}]}
{"id": 325472, "image": "COCO_train2014_000000325472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small horse\"."}], "task": "refering", "answer_template": "The \"a small horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 264, 266, 287, 289], "bbox": [0.4715625, 0.5480410022779043, 0.695609375, 0.7733712984054669], "iscrowd": 0, "area": 5825.999349999999, "rle": {"size": [439, 640], "counts": "kgQ43a=3N3L3M2O2M3O1N2O0O01L3I8NkC9`:F_E=_:E_DIi0l0YOUOn:`0bEg0^:\\O]Ee0d:]OXEe0h:T1OYO]E^Nb:a1hEXNV:g1jE[NU:U2REmM10m:b2O10O1HREgMm:m1dEoM\\:n1hEQNW:m1mERNS:l1m0N3M2N2N2N2N2O1N2N2O1O101O00010O00010O00010O00010O00001O00010O0000000001O00001O1O1O1lDeNj9]1PFgNP:[1iEjNV:X1eElN[:X2O100O100O2O000O]Od0O1\\OfDiN[;U1nDaNV;]1c0N2N101N2O1O1O1O0O2O1O1O1O1O100O001N2L4M3M3M4N11O1O001O3M5K1N2L4N13N00O1O1O1001O1OWCTOg<m02N1O2N1O1O6J4L2N2M4M2LaZc2"}, "label": "a small horse"}]}
{"id": 325472, "image": "COCO_train2014_000000325472.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller brown baby horse\"."}], "task": "refering", "answer_template": "The \"smaller brown baby horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 264, 266, 287, 289], "bbox": [0.4715625, 0.5480410022779043, 0.695609375, 0.7733712984054669], "iscrowd": 0, "area": 5825.999349999999, "rle": {"size": [439, 640], "counts": "kgQ43a=3N3L3M2O2M3O1N2O0O01L3I8NkC9`:F_E=_:E_DIi0l0YOUOn:`0bEg0^:\\O]Ee0d:]OXEe0h:T1OYO]E^Nb:a1hEXNV:g1jE[NU:U2REmM10m:b2O10O1HREgMm:m1dEoM\\:n1hEQNW:m1mERNS:l1m0N3M2N2N2N2N2O1N2N2O1O101O00010O00010O00010O00010O00001O00010O0000000001O00001O1O1O1lDeNj9]1PFgNP:[1iEjNV:X1eElN[:X2O100O100O2O000O]Od0O1\\OfDiN[;U1nDaNV;]1c0N2N101N2O1O1O1O0O2O1O1O1O1O100O001N2L4M3M3M4N11O1O001O3M5K1N2L4N13N00O1O1O1001O1OWCTOg<m02N1O2N1O1O6J4L2N2M4M2LaZc2"}, "label": "smaller brown baby horse"}]}
{"id": 63334, "image": "COCO_train2014_000000063334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a blue jacket with a white stipe and grey pants standing on a baseball field\"."}], "task": "refering", "answer_template": "The \"a man with a blue jacket with a white stipe and grey pants standing on a baseball field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 155, 156, 178, 179, 180, 201, 202, 203, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 316, 317, 318, 339, 340, 341], "bbox": [0.73809375, 0.3423963133640553, 0.87340625, 0.9342396313364056], "iscrowd": 0, "area": 12979.867500000002, "rle": {"size": [434, 640], "counts": "ekX62^=2N2N2N2M3N2N2nHBXKO]4a08GWKMZ4>;MVKJY4;>2TKI<Ca2d0o14oJJK1T33R26jJP1P3lNU29fJm0R3lNX2e2dM^M[2e2aM]M_2k2WMWMh2e3[L]Le3h60000000000000000000000001O1O2N1O1O1O1O8Fe0ZO4L4K4_NQI\\LS7f3QIPLS7R4QIdKS7^4QIWKS7l4QIjJS7X5c02M3N3M2M=D<D2M3NO1O1O1O1N3N1QMgHCZ71WIIj6KfI0[6FTJ3n5LWJOk5O[JKf54_JGb58cJC^5;kJ\\OW5b0VKROk4l0bKhN`4U1mK_NU4^1T4K5K5K5K5J6K5K5J6K5JbRR1"}, "label": "a man with a blue jacket with a white stipe and grey pants standing on a baseball field"}]}
{"id": 63334, "image": "COCO_train2014_000000063334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue jacket\"."}], "task": "refering", "answer_template": "The \"a man in a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 155, 156, 178, 179, 180, 201, 202, 203, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 316, 317, 318, 339, 340, 341], "bbox": [0.73809375, 0.3423963133640553, 0.87340625, 0.9342396313364056], "iscrowd": 0, "area": 12979.867500000002, "rle": {"size": [434, 640], "counts": "ekX62^=2N2N2N2M3N2N2nHBXKO]4a08GWKMZ4>;MVKJY4;>2TKI<Ca2d0o14oJJK1T33R26jJP1P3lNU29fJm0R3lNX2e2dM^M[2e2aM]M_2k2WMWMh2e3[L]Le3h60000000000000000000000001O1O2N1O1O1O1O8Fe0ZO4L4K4_NQI\\LS7f3QIPLS7R4QIdKS7^4QIWKS7l4QIjJS7X5c02M3N3M2M=D<D2M3NO1O1O1O1N3N1QMgHCZ71WIIj6KfI0[6FTJ3n5LWJOk5O[JKf54_JGb58cJC^5;kJ\\OW5b0VKROk4l0bKhN`4U1mK_NU4^1T4K5K5K5K5J6K5K5J6K5JbRR1"}, "label": "a man in a blue jacket"}]}
{"id": 63334, "image": "COCO_train2014_000000063334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball pitcher after he just threw a ball\"."}], "task": "refering", "answer_template": "The \"a baseball pitcher after he just threw a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 309, 310, 311, 333, 334, 356, 357], "bbox": [0.3949375, 0.2962672811059908, 0.69190625, 0.9825115207373272], "iscrowd": 0, "area": 22996.197200000006, "rle": {"size": [434, 640], "counts": "Sa[33_=3SOn0I5J5L5K4L3L5L4L3M4K4M4L4L3M4K4M4L4K4M4M3M4L4K7J9G9F:G7I000O101O1O1NUN]HlMb7T2`HkM_7T2dHkM[7U2gHjMW7W2lHgMS7Y2oHeMR7Z2PIeMP7Y2SIeMm6[2VI_Mn6`2TIXMS7^2XI`Mi6a2XI]Mh6d2XI[Mh6e2YI[Me6g2[IXMe6i2\\IUMd6k2]IUMb6l2_IRMa6o2_IPM\\6U3fIjLV6Z3kIdLS6_3mI`LP6Y4\\IfKc6j4oHTKQ7[5<6I5K6[ITJY5Q6^J[JZ5U7I5L4K5K5K5La0_O1N3N1N2O000O1000O0100O100000XLmKPNT4h1cLiM]3o1[MaMe2X2RNXMn1`2UOeLk0d23UMMg2?PMAo2b500000000O0100000O2N100O100O100O1O100M37I7I6J1O001O002N00O101O0000000O10001O00000O1000000`MQG=n8DTG:k8FVG9j8HUG9j8HVG8j8HVG6k8KUG3m8MSG1o8OQGOQ91nFMU93kFMU93kFMU93kFMU93kFLW93iFMW9CQFAh0l0X9BQFAg0m0X9ATF@c0P1Z9_OSFAc0o0[9_OTFAa0P1\\9^OSFAb0Q1[9\\OVFB?R1\\9YOWFD?R1Z9XO\\Gg0g8UO[Gk0g8QO[Gn0e:00O1O2N100O1O2N2N2O1N2N2N101N1N3N1N2O2N1N2O2M2O1Oifc2"}, "label": "a baseball pitcher after he just threw a ball"}]}
{"id": 63334, "image": "COCO_train2014_000000063334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pitcher with glove on hand throwing the ball\"."}], "task": "refering", "answer_template": "The \"pitcher with glove on hand throwing the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 309, 310, 311, 333, 334, 356, 357], "bbox": [0.3949375, 0.2962672811059908, 0.69190625, 0.9825115207373272], "iscrowd": 0, "area": 22996.197200000006, "rle": {"size": [434, 640], "counts": "Sa[33_=3SOn0I5J5L5K4L3L5L4L3M4K4M4L4L3M4K4M4L4K4M4M3M4L4K7J9G9F:G7I000O101O1O1NUN]HlMb7T2`HkM_7T2dHkM[7U2gHjMW7W2lHgMS7Y2oHeMR7Z2PIeMP7Y2SIeMm6[2VI_Mn6`2TIXMS7^2XI`Mi6a2XI]Mh6d2XI[Mh6e2YI[Me6g2[IXMe6i2\\IUMd6k2]IUMb6l2_IRMa6o2_IPM\\6U3fIjLV6Z3kIdLS6_3mI`LP6Y4\\IfKc6j4oHTKQ7[5<6I5K6[ITJY5Q6^J[JZ5U7I5L4K5K5K5La0_O1N3N1N2O000O1000O0100O100000XLmKPNT4h1cLiM]3o1[MaMe2X2RNXMn1`2UOeLk0d23UMMg2?PMAo2b500000000O0100000O2N100O100O100O1O100M37I7I6J1O001O002N00O101O0000000O10001O00000O1000000`MQG=n8DTG:k8FVG9j8HUG9j8HVG8j8HVG6k8KUG3m8MSG1o8OQGOQ91nFMU93kFMU93kFMU93kFMU93kFLW93iFMW9CQFAh0l0X9BQFAg0m0X9ATF@c0P1Z9_OSFAc0o0[9_OTFAa0P1\\9^OSFAb0Q1[9\\OVFB?R1\\9YOWFD?R1Z9XO\\Gg0g8UO[Gk0g8QO[Gn0e:00O1O2N100O1O2N2N2O1N2N2N101N1N3N1N2O2N1N2O2M2O1Oifc2"}, "label": "pitcher with glove on hand throwing the ball"}]}
{"id": 63334, "image": "COCO_train2014_000000063334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player shown running behind the pitcher\"."}], "task": "refering", "answer_template": "The \"the player shown running behind the pitcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 211, 212, 213, 214, 233, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 282, 301, 302, 303, 304, 305, 325, 327], "bbox": [0.124953125, 0.5529493087557603, 0.33678125, 0.9057603686635946], "iscrowd": 0, "area": 8362.619500000003, "rle": {"size": [434, 640], "counts": "hXR19S=6O110O0O2N101N101N1O2O000O01O1O010O1O10O01O1O001O1O1OZMHSH8n7IQH7n7KQH5o7LoG4R8MmG3R8OmG1S80lGOU80kG1T80lG0T80lGOT82kGOU81kGOT81mGNT82lGNS83lGNT81mGNS83iG1V8OeG6[8J`G<`8DZGa0f8_OUGg0j8YOiFT1W9mNcFY1]9fNcF[1\\9eNdF[1\\9eNdF\\1\\9cNdF]1\\9cNdF^1[9bNdF_1\\9aNdF`1\\9_NdF`1]9`NcF`1]9_NdF`1GRNS9>VG`1@\\NX93XG`1A`NT90[G`1BaNQ9O\\Ga1CbNP9L]Ga1DdNn8K^Ga1DfNm8H_Ga1EhNl8F^G`1IlNh8D^GW13WO^8C]GQ1:]OY8C[GQ1;^OY8BYGS1<\\OZ8CXGS1<\\O[8BWGS1=]O[8AVGT1=\\O]8ASGV1=\\O_8_ORGV1=]Oa8^OPGW1=\\Oc8^OnFX1<\\Oe8]1YGeNg8\\1VGfNj8[1SGWNN\\Oo8]2RGUN5ZOj8b2nFTN=VOf8c3`GYLa8e3g0M3N2N2O1N2O1N2O1N2O1fNhEYOY:b0nEZOT:e0nEYOS:f0PFVOR:h0QFVOP:i0SFTOn9k0TFROn9m0TFQOm9n0\\11N2Nc0^O0O1O010O100O1O1000O0100000000OeNVEOk:0]EIc:6aEDc::`E@e:=Z1L3N3M3M3L4M]lc5"}, "label": "the player shown running behind the pitcher"}]}
{"id": 63334, "image": "COCO_train2014_000000063334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player running to a base\"."}], "task": "refering", "answer_template": "The \"a baseball player running to a base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 211, 212, 213, 214, 233, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 282, 301, 302, 303, 304, 305, 325, 327], "bbox": [0.124953125, 0.5529493087557603, 0.33678125, 0.9057603686635946], "iscrowd": 0, "area": 8362.619500000003, "rle": {"size": [434, 640], "counts": "hXR19S=6O110O0O2N101N101N1O2O000O01O1O010O1O10O01O1O001O1O1OZMHSH8n7IQH7n7KQH5o7LoG4R8MmG3R8OmG1S80lGOU80kG1T80lG0T80lGOT82kGOU81kGOT81mGNT82lGNS83lGNT81mGNS83iG1V8OeG6[8J`G<`8DZGa0f8_OUGg0j8YOiFT1W9mNcFY1]9fNcF[1\\9eNdF[1\\9eNdF\\1\\9cNdF]1\\9cNdF^1[9bNdF_1\\9aNdF`1\\9_NdF`1]9`NcF`1]9_NdF`1GRNS9>VG`1@\\NX93XG`1A`NT90[G`1BaNQ9O\\Ga1CbNP9L]Ga1DdNn8K^Ga1DfNm8H_Ga1EhNl8F^G`1IlNh8D^GW13WO^8C]GQ1:]OY8C[GQ1;^OY8BYGS1<\\OZ8CXGS1<\\O[8BWGS1=]O[8AVGT1=\\O]8ASGV1=\\O_8_ORGV1=]Oa8^OPGW1=\\Oc8^OnFX1<\\Oe8]1YGeNg8\\1VGfNj8[1SGWNN\\Oo8]2RGUN5ZOj8b2nFTN=VOf8c3`GYLa8e3g0M3N2N2O1N2O1N2O1N2O1fNhEYOY:b0nEZOT:e0nEYOS:f0PFVOR:h0QFVOP:i0SFTOn9k0TFROn9m0TFQOm9n0\\11N2Nc0^O0O1O010O100O1O1000O0100000000OeNVEOk:0]EIc:6aEDc::`E@e:=Z1L3N3M3M3L4M]lc5"}, "label": "a baseball player running to a base"}]}
{"id": 538470, "image": "COCO_train2014_000000538470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging at a ball\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging at a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 177, 178, 179, 194, 195, 196, 211, 212, 213, 229, 230, 246, 247, 263, 264, 280, 281], "bbox": [0.4115163934426229, 0.41051562500000005, 0.5729713114754099, 0.7376093750000001], "iscrowd": 0, "area": 9229.285550000006, "rle": {"size": [640, 488], "counts": "lmm32mc03L3N2N101N2N2O0O2O1N2N2RO@Z^Oa0_a0I\\^O9_a0O\\^O3ca0O[^O3ba0N]^O4ba0L]^O7`a0J`^O7_a0I`^O:]a0Hb^O:[a0Hc^O:\\a0Ga^O<^a0T1O1O1O>B8^@iLk=^3SAdMR1oNT<S5dCnJ\\<U5`ClJ`<U5^ClJc<U5ZCmJe<T5YCmJh<T5UCmJk<T5TClJm<S5RCnJo<R5oBoJQ=R5mBoJT=P5kBQKV=P5fBRKZ=o4cBSK_=l4_BUKe=h4XBZKk=\\53M1O1O1O001N2O001O1O001O1O001O0000001O0O2N1DhAoJZ>n4=aKkAV2BgMf>LoAX2AiMa>EXB]2^OkMh?P2\\@nMf?6^_OZ1Q1[Ng?8Z_OY1na0eNU^Oj0jb0J6H9EldQ4"}, "label": "a baseball player swinging at a ball"}]}
{"id": 538470, "image": "COCO_train2014_000000538470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 177, 178, 179, 194, 195, 196, 211, 212, 213, 229, 230, 246, 247, 263, 264, 280, 281], "bbox": [0.4115163934426229, 0.41051562500000005, 0.5729713114754099, 0.7376093750000001], "iscrowd": 0, "area": 9229.285550000006, "rle": {"size": [640, 488], "counts": "lmm32mc03L3N2N101N2N2O0O2O1N2N2RO@Z^Oa0_a0I\\^O9_a0O\\^O3ca0O[^O3ba0N]^O4ba0L]^O7`a0J`^O7_a0I`^O:]a0Hb^O:[a0Hc^O:\\a0Ga^O<^a0T1O1O1O>B8^@iLk=^3SAdMR1oNT<S5dCnJ\\<U5`ClJ`<U5^ClJc<U5ZCmJe<T5YCmJh<T5UCmJk<T5TClJm<S5RCnJo<R5oBoJQ=R5mBoJT=P5kBQKV=P5fBRKZ=o4cBSK_=l4_BUKe=h4XBZKk=\\53M1O1O1O001N2O001O1O001O1O001O0000001O0O2N1DhAoJZ>n4=aKkAV2BgMf>LoAX2AiMa>EXB]2^OkMh?P2\\@nMf?6^_OZ1Q1[Ng?8Z_OY1na0eNU^Oj0jb0J6H9EldQ4"}, "label": "a baseball player swinging a bat"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hands of a man holding a piece of food that has white sauce on it\"."}], "task": "refering", "answer_template": "The \"the hands of a man holding a piece of food that has white sauce on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185], "bbox": [0.0, 0.135125, 0.1300625, 0.5315416666666667], "iscrowd": 0, "area": 11748.095699999998, "rle": {"size": [480, 640], "counts": "e6Z1Q;e2TNl1000O10000000O10O10000000000O010000000O100O1O010O1O100O100O1O101N1O101N100O2N1O1O2M2O1O2M2O1N2O2M2O1N3N1N2O1RMfET1]:aNmE]1U:`NnE_1S:]NRF`1Q:[NSFd1n9YNUFf1l9YNUFf1l9ZNUFd1m9[NSFd1n9\\NRFc1P:\\NQFb1P:^NQFa1P:^NRF_1o9`NUF\\1l9dNVFY1l9fNXFU1i9kN[FP1g9oN]Fl0d9TO`Fg0a9YOcFb0_9]OfF=[9CkF7V9GPG3Q9LZGIh83U3M3L_PU8"}, "label": "the hands of a man holding a piece of food that has white sauce on it"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white hands holding a barbeque wing\"."}], "task": "refering", "answer_template": "The \"white hands holding a barbeque wing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185], "bbox": [0.0, 0.135125, 0.1300625, 0.5315416666666667], "iscrowd": 0, "area": 11748.095699999998, "rle": {"size": [480, 640], "counts": "e6Z1Q;e2TNl1000O10000000O10O10000000000O010000000O100O1O010O1O100O100O1O101N1O101N100O2N1O1O2M2O1O2M2O1N2O2M2O1N3N1N2O1RMfET1]:aNmE]1U:`NnE_1S:]NRF`1Q:[NSFd1n9YNUFf1l9YNUFf1l9ZNUFd1m9[NSFd1n9\\NRFc1P:\\NQFb1P:^NQFa1P:^NRF_1o9`NUF\\1l9dNVFY1l9fNXFU1i9kN[FP1g9oN]Fl0d9TO`Fg0a9YOcFb0_9]OfF=[9CkF7V9GPG3Q9LZGIh83U3M3L_PU8"}, "label": "white hands holding a barbeque wing"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half drink in glass in middle front behind french fry and food plate\"."}], "task": "refering", "answer_template": "The \"half drink in glass in middle front behind french fry and food plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 146, 147, 148, 170, 171, 193, 194], "bbox": [0.35148437499999996, 0.057708333333333334, 0.49157812500000003, 0.5212916666666667], "iscrowd": 0, "area": 15488.32815, "rle": {"size": [480, 640], "counts": "S`Y39g>a0_Oa0_Oa0^Oa0@5K5K5K4K6K5K5K5K5J6K5K4L5K5J6K5K5K5K5J6K6J5K5K5K6I6K1O00000O10001O000000001O00000000001O00000000001O000000001O00000000001O0001O000001O000000SOnGlJS8i4YHVKf7_4hH^KX7Y4TIdKm6T4\\IjKd6o3fInKZ6k3PJQLQ6g3ZJTLi5d3`JYLe5\\3Z3_O`0@`0@a0WOh0nNhYh4"}, "label": "half drink in glass in middle front behind french fry and food plate"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pilsner glass is half full of beer next to a fuller glass and full plate of food\"."}], "task": "refering", "answer_template": "The \"a pilsner glass is half full of beer next to a fuller glass and full plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 146, 147, 148, 170, 171, 193, 194], "bbox": [0.35148437499999996, 0.057708333333333334, 0.49157812500000003, 0.5212916666666667], "iscrowd": 0, "area": 15488.32815, "rle": {"size": [480, 640], "counts": "S`Y39g>a0_Oa0_Oa0^Oa0@5K5K5K4K6K5K5K5K5J6K5K4L5K5J6K5K5K5K5J6K6J5K5K5K6I6K1O00000O10001O000000001O00000000001O00000000001O000000001O00000000001O0001O000001O000000SOnGlJS8i4YHVKf7_4hH^KX7Y4TIdKm6T4\\IjKd6o3fInKZ6k3PJQLQ6g3ZJTLi5d3`JYLe5\\3Z3_O`0@`0@a0WOh0nNhYh4"}, "label": "a pilsner glass is half full of beer next to a fuller glass and full plate of food"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass of beer , not full or empty , next to man wearing a ring on his finger\"."}], "task": "refering", "answer_template": "The \"glass of beer , not full or empty , next to man wearing a ring on his finger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 38, 39, 40, 41, 61, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 130, 131, 132, 153, 154, 155, 176, 177, 178, 199, 200, 201, 223], "bbox": [0.67078125, 0.035958333333333335, 0.812359375, 0.5460625], "iscrowd": 0, "area": 14690.3031, "rle": {"size": [480, 640], "counts": "ciY65h>3L5L3PKZ1bIkNj4_O`He3g1PMh5P4dITL\\6V4WInKh6\\4jHhKV7b4]HbKb7n501O1N101O001O1O001N101O1O001O00O1O001N2O1O1O1O1O001O1O1O1O1O1N1O2M3N2M3M4M2M3N2M3M3N2M3M3N3I6K5K5J6K5J6K5K6I6K5K5J6K5K5J6K6J5J6K5K5J6K5K5J6K6I6K5K5J6K5J6K5K6I_Wh1"}, "label": "glass of beer , not full or empty , next to man wearing a ring on his finger"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in gray shirt in background of photo\"."}], "task": "refering", "answer_template": "The \"person in gray shirt in background of photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 89], "bbox": [0.585078125, 0.0, 0.9871874999999999, 0.19691666666666666], "iscrowd": 0, "area": 11612.104599999995, "rle": {"size": [480, 640], "counts": "XZ_52f>b0F001O001O1O001O1O001O1O001O0000001O2N000000O10000O1O1O1O1O100O10000O100O10000O10000O10000O1000000O10000000000000000O1000000000000000000O100000000O100000000O100000000O100000000000000000000000000000000000000000000001O0000000000001O000000001O000000001O000000001O0000001O001O001O001O0aC_O_:b0UEKi:5lD6T;K_D`0b;BQDh0P<T10O1O1M3N2001O00001O00001O00001O001O001O001O001O00O1O1O1O100000000001O000000004L2N1O2N1O0000000000O1O1O1O1N2M3N20000O10000O1000000000000001O000000001O0000001O0000001O000000O1N2O1O1O100O10001O3M2N3M2M4M5K6J6I6K6J6J5J7J6J5J3Mie3"}, "label": "person in gray shirt in background of photo"}]}
{"id": 399208, "image": "COCO_train2014_000000399208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on top right in grey\"."}], "task": "refering", "answer_template": "The \"person on top right in grey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 89], "bbox": [0.585078125, 0.0, 0.9871874999999999, 0.19691666666666666], "iscrowd": 0, "area": 11612.104599999995, "rle": {"size": [480, 640], "counts": "XZ_52f>b0F001O001O1O001O1O001O1O001O0000001O2N000000O10000O1O1O1O1O100O10000O100O10000O10000O10000O1000000O10000000000000000O1000000000000000000O100000000O100000000O100000000O100000000000000000000000000000000000000000000001O0000000000001O000000001O000000001O000000001O0000001O001O001O001O0aC_O_:b0UEKi:5lD6T;K_D`0b;BQDh0P<T10O1O1M3N2001O00001O00001O00001O001O001O001O001O00O1O1O1O100000000001O000000004L2N1O2N1O0000000000O1O1O1O1N2M3N20000O10000O1000000000000001O000000001O0000001O0000001O000000O1N2O1O1O100O10001O3M2N3M2M4M5K6J6I6K6J6J5J7J6J5J3Mie3"}, "label": "person on top right in grey"}]}
{"id": 464744, "image": "COCO_train2014_000000464744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sofa with back turned\"."}], "task": "refering", "answer_template": "The \"sofa with back turned\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.001421875, 0.7582394366197183, 0.514546875, 0.9809389671361503], "iscrowd": 0, "area": 14369.295050000002, "rle": {"size": [426, 640], "counts": "fg0f2e:O00001O00001O0000001O00001O00000000O100O100O1O100000000001O00000000O100O10000O10000O100O10000O10000O100O10000O10000O100O100001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O1O3M3M4L3M3M3M1O001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O00000000_\\Q4"}, "label": "sofa with back turned"}]}
{"id": 464744, "image": "COCO_train2014_000000464744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black leather couch facing the tv\"."}], "task": "refering", "answer_template": "The \"black leather couch facing the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.001421875, 0.7582394366197183, 0.514546875, 0.9809389671361503], "iscrowd": 0, "area": 14369.295050000002, "rle": {"size": [426, 640], "counts": "fg0f2e:O00001O00001O0000001O00001O00000000O100O100O1O100000000001O00000000O100O10000O10000O100O10000O10000O100O10000O10000O100O100001O1O001O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O1O3M3M4L3M3M3M1O001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O00000000_\\Q4"}, "label": "black leather couch facing the tv"}]}
{"id": 464744, "image": "COCO_train2014_000000464744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black couch with stuff on it\"."}], "task": "refering", "answer_template": "The \"black couch with stuff on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 331], "bbox": [0.14254687500000002, 0.4998122065727699, 0.5224062500000001, 0.9446478873239436], "iscrowd": 0, "area": 28660.561049999997, "rle": {"size": [426, 640], "counts": "dSV12W=8H9BB`Ce0W<`0G8I8H7H8I7I7H7J3M4K7J7I0O101O00001M2O2O00001O00001N11O01O0100O102M1O01M2N3N100O10001O0O10000000001O0O10001O001O001N101O1O001N100O100O2O000O101O000O101O0O101N10001N10001O002N2M010O0100O1000001N1O3N1OL30100O10000O1000YMcGb0[8]OhGb0X8\\OkGRNLZ2X8DnGoML]2V8CVH=i7BXH>g7CYH=g7CZH<f7C[H>c7D\\H<d7E[H;d7G\\H9c7H\\H8d7H\\H8c7H_H7a7I_H8`7H`H8_7HcH7]7IcH7\\7JeH6Z7JfH6Z7JeH7Z7JfH7Y7HhH8X7FjH:V7DlH<S7EmH<R7ElH<T7DmH;S7EnH:Q7GPI9o6GRI8n6HSI7m6ISI7k6KVI5g6MYI3g6NXI2h6NXI3g6MXI4h6TO]H`N3=i0n1f6UOiHjNb0Q2d6TOlHjNa0Q2c6TOnHjN?S2c6ROnHkN?S2c6QOPIlN<S2e6oNQInN9T2g6SOiHjN1HL[2[7^N_HG6BJK94MZ2]7\\N^H@NK6OMLO1O55X2`7ZN\\HBOFN27O0OH4325Y2b7VNZHANJ61OO30F3255W2e7TNXHC9J74^O45W2h7QNTHGQ1LnN63W2m7mMPHLR16QOQ2o7kMmGNV15mNS2T8dMjG3U17lNS2Y8_MeG7X17gNU2\\8]MeG6Y1IfNT3Z8nLgG4Z1P3m6mLiG5Z1m2m6nLiG5Z1IeNQ3X8PMiG7Z1k2o6lLhG9Y1i2T7YMkHHfNi2`8^MkHf2U7ZMlHe2U7ZMlHLcNb2b8aMlHe2U7ZMkHf2W7XMjHg2X7YMgHf2Z7WMiHh2W7XMjHg2V7XMlHg2U7XMlHg2T7YMmHf2l8O1O1O1O2N1O1N2O001O1O1O1O1O1O0jNoD4R;0nDKR;9nDCS;`0mD\\OS;h0mDTOS;Q1mDjNT;Y1lDcNi:F`ES2`:nM_ER2a:oM]ER2m9oM_F0BR2P:nM[F2CP2U:mMWF\\2j9eMSF\\2n9dMRF[2o9eMkE1KY2[:gMgE`2Z:71O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M2O2N1fN`EIb:BXFLiN1]\\o3"}, "label": "black couch with stuff on it"}]}
{"id": 464744, "image": "COCO_train2014_000000464744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather chair near a television\"."}], "task": "refering", "answer_template": "The \"a black leather chair near a television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 331], "bbox": [0.14254687500000002, 0.4998122065727699, 0.5224062500000001, 0.9446478873239436], "iscrowd": 0, "area": 28660.561049999997, "rle": {"size": [426, 640], "counts": "dSV12W=8H9BB`Ce0W<`0G8I8H7H8I7I7H7J3M4K7J7I0O101O00001M2O2O00001O00001N11O01O0100O102M1O01M2N3N100O10001O0O10000000001O0O10001O001O001N101O1O001N100O100O2O000O101O000O101O0O101N10001N10001O002N2M010O0100O1000001N1O3N1OL30100O10000O1000YMcGb0[8]OhGb0X8\\OkGRNLZ2X8DnGoML]2V8CVH=i7BXH>g7CYH=g7CZH<f7C[H>c7D\\H<d7E[H;d7G\\H9c7H\\H8d7H\\H8c7H_H7a7I_H8`7H`H8_7HcH7]7IcH7\\7JeH6Z7JfH6Z7JeH7Z7JfH7Y7HhH8X7FjH:V7DlH<S7EmH<R7ElH<T7DmH;S7EnH:Q7GPI9o6GRI8n6HSI7m6ISI7k6KVI5g6MYI3g6NXI2h6NXI3g6MXI4h6TO]H`N3=i0n1f6UOiHjNb0Q2d6TOlHjNa0Q2c6TOnHjN?S2c6ROnHkN?S2c6QOPIlN<S2e6oNQInN9T2g6SOiHjN1HL[2[7^N_HG6BJK94MZ2]7\\N^H@NK6OMLO1O55X2`7ZN\\HBOFN27O0OH4325Y2b7VNZHANJ61OO30F3255W2e7TNXHC9J74^O45W2h7QNTHGQ1LnN63W2m7mMPHLR16QOQ2o7kMmGNV15mNS2T8dMjG3U17lNS2Y8_MeG7X17gNU2\\8]MeG6Y1IfNT3Z8nLgG4Z1P3m6mLiG5Z1m2m6nLiG5Z1IeNQ3X8PMiG7Z1k2o6lLhG9Y1i2T7YMkHHfNi2`8^MkHf2U7ZMlHe2U7ZMlHLcNb2b8aMlHe2U7ZMkHf2W7XMjHg2X7YMgHf2Z7WMiHh2W7XMjHg2V7XMlHg2U7XMlHg2T7YMmHf2l8O1O1O1O2N1O1N2O001O1O1O1O1O1O0jNoD4R;0nDKR;9nDCS;`0mD\\OS;h0mDTOS;Q1mDjNT;Y1lDcNi:F`ES2`:nM_ER2a:oM]ER2m9oM_F0BR2P:nM[F2CP2U:mMWF\\2j9eMSF\\2n9dMRF[2o9eMkE1KY2[:gMgE`2Z:71O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M2O2N1fN`EIb:BXFLiN1]\\o3"}, "label": "a black leather chair near a television"}]}
{"id": 63347, "image": "COCO_train2014_000000063347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue kids ' chair\"."}], "task": "refering", "answer_template": "The \"the blue kids ' chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 197, 217, 218, 219, 223, 240, 241, 245, 246, 247, 263, 264, 265, 287, 288, 311, 314, 315], "bbox": [0.47037500000000004, 0.44908333333333333, 0.7550156250000001, 0.7990625], "iscrowd": 0, "area": 5057.963649999996, "rle": {"size": [480, 640], "counts": "g[]47`><nA@]=d0XB7T=m0M4M2M3N1O2N1O1N2O1O1O2M2O1O1O1N2O_NlC8R<@VDa0i;ZO]Df0b;YO_Dh0_;XOcDi0[;VOgDj0X;UOiDm0U;SOlDm0R;SOoDo0o:POSEP1l:oNVER1h:mNYET1e:lN\\EV1b:jN^EW1a:hN`EZ1^:eNcE\\1\\:cNeE_1X:aNiEa0mNLY;BkE;SO5P;@mE3[O>g:^OoEMAg0]:\\OSFEHj0SOVOQ;:XG>P9AQG=Q9CmF=U9BcFe0^9[OZFl0h9SOPFd0b:[O[E=Q;BQE2Y;Nm1O1O1O0O2O1OdR62WmI4L4L3N3M2N3O010O0000L5N2O1N2N2O1N2O001O1O1N2O1OW]a01hb^O2N101N1O2N2UDFn8:lFKT97fFNY94`F1`90ZF6f9KYF5f9MZF1f90[FOd93\\FKe96[FI[9b0eF]OW9i0hFUOU9P1kFnNT9V1kFhNV9Y1jFeNW9\\1iFbNZ9]1eFbN]9\\1eFbN]9]1T2N2O1N2O1N2N2N3M2O1N2N2N2N2O1O1OlA_Oj=b0QBCo=g001OO1K5K5L4K6KTZY2"}, "label": "the blue kids ' chair"}]}
{"id": 63347, "image": "COCO_train2014_000000063347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tiny chair behind the little boy\"."}], "task": "refering", "answer_template": "The \"the tiny chair behind the little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 197, 217, 218, 219, 223, 240, 241, 245, 246, 247, 263, 264, 265, 287, 288, 311, 314, 315], "bbox": [0.47037500000000004, 0.44908333333333333, 0.7550156250000001, 0.7990625], "iscrowd": 0, "area": 5057.963649999996, "rle": {"size": [480, 640], "counts": "g[]47`><nA@]=d0XB7T=m0M4M2M3N1O2N1O1N2O1O1O2M2O1O1O1N2O_NlC8R<@VDa0i;ZO]Df0b;YO_Dh0_;XOcDi0[;VOgDj0X;UOiDm0U;SOlDm0R;SOoDo0o:POSEP1l:oNVER1h:mNYET1e:lN\\EV1b:jN^EW1a:hN`EZ1^:eNcE\\1\\:cNeE_1X:aNiEa0mNLY;BkE;SO5P;@mE3[O>g:^OoEMAg0]:\\OSFEHj0SOVOQ;:XG>P9AQG=Q9CmF=U9BcFe0^9[OZFl0h9SOPFd0b:[O[E=Q;BQE2Y;Nm1O1O1O0O2O1OdR62WmI4L4L3N3M2N3O010O0000L5N2O1N2N2O1N2O001O1O1N2O1OW]a01hb^O2N101N1O2N2UDFn8:lFKT97fFNY94`F1`90ZF6f9KYF5f9MZF1f90[FOd93\\FKe96[FI[9b0eF]OW9i0hFUOU9P1kFnNT9V1kFhNV9Y1jFeNW9\\1iFbNZ9]1eFbN]9\\1eFbN]9]1T2N2O1N2O1N2N2N3M2O1N2N2N2N2O1O1OlA_Oj=b0QBCo=g001OO1K5K5L4K6KTZY2"}, "label": "the tiny chair behind the little boy"}]}
{"id": 63347, "image": "COCO_train2014_000000063347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a jeans and gray t shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a jeans and gray t shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 51, 52, 53, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330], "bbox": [0.133375, 0.07429166666666666, 0.466859375, 0.8800624999999999], "iscrowd": 0, "area": 46681.91215, "rle": {"size": [480, 640], "counts": "gRX11m>4L4M3O1N2O00O1O010O1O1O3NO0]OH^B8_=L`B5]=MdB3X=1gB0V=3iBOR=4nBMn<7QCK^<e0aC[OY<l0fCTOW<o0iCQOS<T1lClNl;]1SDcNk2IZ5m1hG\\Nj2NT5h2gJ_Mo4i2lJ^Mo4e2lJbMo4a2lJgMo4[2lJlMP5U2mJQNn4R2oJSNm4a1QIlLP2i1j4^1VIfLn1Q2j4Z1YIcLk1V2k4X1ZKlNf4U1VKnNi4S1SKQOm4m0SKUOl4j0TKXOl4e0UK]Oj4b0VK@i4?VKDj49WKIi45WKNg40ZK2f4K[K7d4G]K;c4A_Ka0`4]OaKe0_4XObKj0^4ROcKQ1\\4mNeKU1[4gNgK[1X4fNfK\\1Z4eNaK_1^4dN]K_1c4cNeJ[M_OU4k5cNTJRNC_3Z4iK^Ma8^2dG`M\\8^2hG_MX8_2nGVLDf0]8S3SHQLO`0o7\\3TI`Lk6^3[I^Lf6a3^I\\La6d3cIYL\\6g3hIULY6j3^21O0BWLgEh3V:\\LiEe3S:`LlE_3S:dLkE\\3S:gLlEZ3Q:jLnEU3P:oLnEQ3Q:QMoEo2n9R1N3M3M3N2M3M3N2N2N2N2N2N2O1N2NhKVG^2k8^MhGR2[8hMjGV2Y8cMlG\\2V8SM[Hk2n9O2N1O2N2N1O2N10O0001O001O001O01O100O100O2N100O1O1O1N2O1O1N2]FbLb7h3oG^LP8P4]GWL`8R4TGSLk8o3oFVLP9l3jFYLT9k4O001N101N10001N101N1O1O2N1O2N1O1O2N101N2DcIRH^6g7lIUHV6d7f0J7I7I6I8G9G8I8H8H7L5M3N1N3M3N1N3N2M3M4M2M3N3L3N2M4M3L3N3L3N3M3L5L4K5L5J5L3L3N3L3N3L3N3L3N3L3N5I;F;D8I6I7Jjco4"}, "label": "a man wearing a jeans and gray t shirt"}]}
{"id": 63347, "image": "COCO_train2014_000000063347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the grey shirt\"."}], "task": "refering", "answer_template": "The \"the man in the grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 51, 52, 53, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330], "bbox": [0.133375, 0.07429166666666666, 0.466859375, 0.8800624999999999], "iscrowd": 0, "area": 46681.91215, "rle": {"size": [480, 640], "counts": "gRX11m>4L4M3O1N2O00O1O010O1O1O3NO0]OH^B8_=L`B5]=MdB3X=1gB0V=3iBOR=4nBMn<7QCK^<e0aC[OY<l0fCTOW<o0iCQOS<T1lClNl;]1SDcNk2IZ5m1hG\\Nj2NT5h2gJ_Mo4i2lJ^Mo4e2lJbMo4a2lJgMo4[2lJlMP5U2mJQNn4R2oJSNm4a1QIlLP2i1j4^1VIfLn1Q2j4Z1YIcLk1V2k4X1ZKlNf4U1VKnNi4S1SKQOm4m0SKUOl4j0TKXOl4e0UK]Oj4b0VK@i4?VKDj49WKIi45WKNg40ZK2f4K[K7d4G]K;c4A_Ka0`4]OaKe0_4XObKj0^4ROcKQ1\\4mNeKU1[4gNgK[1X4fNfK\\1Z4eNaK_1^4dN]K_1c4cNeJ[M_OU4k5cNTJRNC_3Z4iK^Ma8^2dG`M\\8^2hG_MX8_2nGVLDf0]8S3SHQLO`0o7\\3TI`Lk6^3[I^Lf6a3^I\\La6d3cIYL\\6g3hIULY6j3^21O0BWLgEh3V:\\LiEe3S:`LlE_3S:dLkE\\3S:gLlEZ3Q:jLnEU3P:oLnEQ3Q:QMoEo2n9R1N3M3M3N2M3M3N2N2N2N2N2N2O1N2NhKVG^2k8^MhGR2[8hMjGV2Y8cMlG\\2V8SM[Hk2n9O2N1O2N2N1O2N10O0001O001O001O01O100O100O2N100O1O1O1N2O1O1N2]FbLb7h3oG^LP8P4]GWL`8R4TGSLk8o3oFVLP9l3jFYLT9k4O001N101N10001N101N1O1O2N1O2N1O1O2N101N2DcIRH^6g7lIUHV6d7f0J7I7I6I8G9G8I8H8H7L5M3N1N3M3N1N3N2M3M4M2M3N3L3N2M4M3L3N3L3N3M3L5L4K5L5J5L3L3N3L3N3L3N3L3N3L3N5I;F;D8I6I7Jjco4"}, "label": "the man in the grey shirt"}]}
{"id": 202617, "image": "COCO_train2014_000000202617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella with multiple colors\"."}], "task": "refering", "answer_template": "The \"umbrella with multiple colors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 127, 129, 130, 150, 173, 196], "bbox": [0.3226875, 0.18243559718969557, 0.73721875, 0.576440281030445], "iscrowd": 0, "area": 11688.230999999996, "rle": {"size": [427, 640], "counts": "d^f21Y=1O2N2N1O2O0O1O1O1O1O100O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O00100O1O1O1O1O100O1O100O1O100O1O100O100O1O100O00100O100O1O100O1O100O1O100O100000000O100000O01000000O100000000O1000000O1000O10O100000000O100000000O100000O10000000000000000000O10000000O1000O1N2fD`Nh:a1YEU1R8iNnGW1S8hNmGX1T8gNnGX1S8fN_Hh0a7XOQI6P7JaID_6<SJSOl5m0fJaNZ5_1Y3000000000001O000000000000001O0000001O0001O0001O000000001O000000000001O00000000000000000000001O000000000000000000001O01O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O001O1O1O001O1O1O1N101O1O1O1O001O1O1O001O1O1O1O001O1N2O1O001O1O1O1O001N2O1O001O1O001O1N10hZV2"}, "label": "umbrella with multiple colors"}]}
{"id": 202617, "image": "COCO_train2014_000000202617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella of various colors\"."}], "task": "refering", "answer_template": "The \"umbrella of various colors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 127, 129, 130, 150, 173, 196], "bbox": [0.3226875, 0.18243559718969557, 0.73721875, 0.576440281030445], "iscrowd": 0, "area": 11688.230999999996, "rle": {"size": [427, 640], "counts": "d^f21Y=1O2N2N1O2O0O1O1O1O1O100O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O00100O1O1O1O1O100O1O100O1O100O1O100O100O1O100O00100O100O1O100O1O100O1O100O100000000O100000O01000000O100000000O1000000O1000O10O100000000O100000000O100000O10000000000000000000O10000000O1000O1N2fD`Nh:a1YEU1R8iNnGW1S8hNmGX1T8gNnGX1S8fN_Hh0a7XOQI6P7JaID_6<SJSOl5m0fJaNZ5_1Y3000000000001O000000000000001O0000001O0001O0001O000000001O000000000001O00000000000000000000001O000000000000000000001O01O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O001O1O1O001O1O1O1N101O1O1O1O001O1O1O001O1O1O1O001O1N2O1O001O1O1O1O001N2O1O001O1O001O1N10hZV2"}, "label": "umbrella of various colors"}]}
{"id": 202617, "image": "COCO_train2014_000000202617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bag being worn by a woman in a green coat\"."}], "task": "refering", "answer_template": "The \"a black bag being worn by a woman in a green coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 217, 239, 240, 262, 283, 284, 285, 306, 307, 308, 329, 330], "bbox": [0.316359375, 0.5573067915690867, 0.485765625, 0.9842622950819672], "iscrowd": 0, "area": 4939.3162999999995, "rle": {"size": [427, 640], "counts": "Tdd21W=3M4K4M4K4M4L3L5L3L5L3M4K4M3N3N1O2N1O2M2O2N1O1O1O1O1N3N100000001O00000001O01O000000M3L4L5K2N1O2O1N2O1O1N101O1O1O100O10O0100O10000O010O1O100O1O10O0100O1O100O00100O1O100O00100O10000O2O1O1N2O2M2O1O1N2O1O2M3N2M3N2N2M3N2NgnX4"}, "label": "a black bag being worn by a woman in a green coat"}]}
{"id": 202617, "image": "COCO_train2014_000000202617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black purse a woman in carrying\"."}], "task": "refering", "answer_template": "The \"a black purse a woman in carrying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 217, 239, 240, 262, 283, 284, 285, 306, 307, 308, 329, 330], "bbox": [0.316359375, 0.5573067915690867, 0.485765625, 0.9842622950819672], "iscrowd": 0, "area": 4939.3162999999995, "rle": {"size": [427, 640], "counts": "Tdd21W=3M4K4M4K4M4L3L5L3L5L3M4K4M3N3N1O2N1O2M2O2N1O1O1O1O1N3N100000001O00000001O01O000000M3L4L5K2N1O2O1N2O1O1N101O1O1O100O10O0100O10000O010O1O100O1O10O0100O1O100O00100O1O100O00100O10000O2O1O1N2O2M2O1O1N2O1O2M3N2M3N2N2M3N2NgnX4"}, "label": "a black purse a woman in carrying"}]}
{"id": 227202, "image": "COCO_train2014_000000227202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse is standing near a man\"."}], "task": "refering", "answer_template": "The \"a brown horse is standing near a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 58, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 371, 372, 373, 374, 375, 376, 377, 378, 379, 394, 395, 396, 397, 398, 399, 400, 401, 402, 417, 418, 419, 420, 421, 422, 423, 424, 425, 440, 441, 442, 443, 444, 445, 446, 447, 448, 463, 464, 465, 466, 467, 468, 469, 470, 471, 487, 488, 489, 490, 491, 492, 493, 494, 510, 511, 512, 515, 516, 517], "bbox": [0.1345, 0.10753125, 0.6473125, 1.0], "iscrowd": 0, "area": 121016.2656, "rle": {"size": [640, 640], "counts": "ZTf1<`c0h0YOg0XOi0XOg0XOi0XO8G7J7H7J7H7J6I8I6I>Ci0VO9H5K5J7J5J6K6I6K6I6K5J7J5J6N3L3M4M2M3M4L3N2M4L3N3L3M3N2M1O1O100O1O100O1O1O100O1O100O1O1O100O1K5K5L4K5L4K5K5L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4N2O1N2O1O1OQO_KaB`4_=hKZBW4f=QLSBn3n=ZLjAe3V>Q1100000001O000O100000001O001O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O10000O100O100O100O10000O100O100O100O2O1N2O001N2O1N101N2O1N2O001N2O1N2O0O2O1N2O1O0O2O1N2O1N4M3L4M3L5L3M3L100O100O100O10000O100O100O2O01O6J5K6J4L00001O00001O00001O00001O00010O00001O000XMSM]Dm2[;\\MTDIWNj2^=fMnCNbN[2Y=PNhC4lNl1T=YNcCZ3V<oL\\C]3]<V2I8H7I7I7I8H7I7I7K5L5J5L4L4K6K4L4K5M40O1O1OhGeEm6Z:PIlEn6T:nHSFo6l9nH[Fo6d9nHcFo6\\9nHkFn6V9oHPGn6o8oHWGo6h8nH_Go6`8nHgGo6Y8mHnGP7Q8mHVHP7i7mH^HP7a7mHeHQ7X:M3M3dJmBe3V=TLoBk3T=nKPCR4R=hKSCW4P=bKTC^4o<[KUCe4n<UKVCj4m<oJWCQ5l=01O001O00001O00001O0]LSAS2m>lMXAP2i>oMZAm1g>RN^Aj1b>VNbAf1_>XNeAc1]>]NgA[1]>dNhAU1\\>jNgAo0]>POhAh0\\>XOhA`0]>_OfA:^>EgA3]>MgAK^>3fAF^>:fA^O^>a0gAXO]>g0^2O101N100O2O000O2O0O101N3N3L5L3L4MTP]4"}, "label": "a brown horse is standing near a man"}]}
{"id": 227202, "image": "COCO_train2014_000000227202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse standing next to a person\"."}], "task": "refering", "answer_template": "The \"a brown horse standing next to a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 58, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 371, 372, 373, 374, 375, 376, 377, 378, 379, 394, 395, 396, 397, 398, 399, 400, 401, 402, 417, 418, 419, 420, 421, 422, 423, 424, 425, 440, 441, 442, 443, 444, 445, 446, 447, 448, 463, 464, 465, 466, 467, 468, 469, 470, 471, 487, 488, 489, 490, 491, 492, 493, 494, 510, 511, 512, 515, 516, 517], "bbox": [0.1345, 0.10753125, 0.6473125, 1.0], "iscrowd": 0, "area": 121016.2656, "rle": {"size": [640, 640], "counts": "ZTf1<`c0h0YOg0XOi0XOg0XOi0XO8G7J7H7J7H7J6I8I6I>Ci0VO9H5K5J7J5J6K6I6K6I6K5J7J5J6N3L3M4M2M3M4L3N2M4L3N3L3M3N2M1O1O100O1O100O1O1O100O1O100O1O1O100O1K5K5L4K5L4K5K5L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4N2O1N2O1O1OQO_KaB`4_=hKZBW4f=QLSBn3n=ZLjAe3V>Q1100000001O000O100000001O001O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O10000O100O100O100O10000O100O100O100O2O1N2O001N2O1N101N2O1N2O001N2O1N2O0O2O1N2O1O0O2O1N2O1N4M3L4M3L5L3M3L100O100O100O10000O100O100O2O01O6J5K6J4L00001O00001O00001O00001O00010O00001O000XMSM]Dm2[;\\MTDIWNj2^=fMnCNbN[2Y=PNhC4lNl1T=YNcCZ3V<oL\\C]3]<V2I8H7I7I7I8H7I7I7K5L5J5L4L4K6K4L4K5M40O1O1OhGeEm6Z:PIlEn6T:nHSFo6l9nH[Fo6d9nHcFo6\\9nHkFn6V9oHPGn6o8oHWGo6h8nH_Go6`8nHgGo6Y8mHnGP7Q8mHVHP7i7mH^HP7a7mHeHQ7X:M3M3dJmBe3V=TLoBk3T=nKPCR4R=hKSCW4P=bKTC^4o<[KUCe4n<UKVCj4m<oJWCQ5l=01O001O00001O00001O0]LSAS2m>lMXAP2i>oMZAm1g>RN^Aj1b>VNbAf1_>XNeAc1]>]NgA[1]>dNhAU1\\>jNgAo0]>POhAh0\\>XOhA`0]>_OfA:^>EgA3]>MgAK^>3fAF^>:fA^O^>a0gAXO]>g0^2O101N100O2O000O2O0O101N3N3L5L3L4MTP]4"}, "label": "a brown horse standing next to a person"}]}
{"id": 227205, "image": "COCO_train2014_000000227205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a green and red train\"."}], "task": "refering", "answer_template": "The \"the side of a green and red train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244], "bbox": [0.1516875, 0.5004375, 0.6690937499999999, 0.6105416666666666], "iscrowd": 0, "area": 10727.452150000001, "rle": {"size": [480, 640], "counts": "Wg]11n>2N2N1O2N2N1O2M3A?0O01O1O00100O001O10O01O1O10O01O<D0001O0000001O0000001O0000000O100000000000000000000000001O0000000000000000000000000000000O10001O0000000000000000000000000000000000001O000000000000000O100000000000000000001O0000000000000000000000000000000000001O0000000000000000000O10000000000000001O000000000000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000O10001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000O101N1O1O100O1O1O100O1O1001O1O1O001O1O1O010O18H<CUl8LPTG;D=C2N001O01O01OO1000001O000O1000001O00000O1K5EXTS3"}, "label": "the side of a green and red train"}]}
{"id": 227205, "image": "COCO_train2014_000000227205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train with the dark burgundy painting around the windows\"."}], "task": "refering", "answer_template": "The \"the train with the dark burgundy painting around the windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244], "bbox": [0.1516875, 0.5004375, 0.6690937499999999, 0.6105416666666666], "iscrowd": 0, "area": 10727.452150000001, "rle": {"size": [480, 640], "counts": "Wg]11n>2N2N1O2N2N1O2M3A?0O01O1O00100O001O10O01O1O10O01O<D0001O0000001O0000001O0000000O100000000000000000000000001O0000000000000000000000000000000O10001O0000000000000000000000000000000000001O000000000000000O100000000000000000001O0000000000000000000000000000000000001O0000000000000000000O10000000000000001O000000000000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000O10001O0000000000000000000000000000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000O101N1O1O100O1O1O100O1O1001O1O1O001O1O1O010O18H<CUl8LPTG;D=C2N001O01O01OO1000001O000O1000001O00000O1K5EXTS3"}, "label": "the train with the dark burgundy painting around the windows"}]}
{"id": 227205, "image": "COCO_train2014_000000227205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train cart on the left in front of the train in the back\"."}], "task": "refering", "answer_template": "The \"train cart on the left in front of the train in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299], "bbox": [0.0, 0.37977083333333334, 0.20224999999999999, 0.7775208333333333], "iscrowd": 0, "area": 16060.455699999999, "rle": {"size": [480, 640], "counts": "m5m0S>j1oMZMSD`4T:`KkEi5R90LWJPGj5P9VJoFj5R9200000O101O00000O1000000O101O000O100000000O2O00000O1000000O101O00000O1000000O2O00000O100000000O2O00000O1000000O2O0000000O1000001N100H8eM[2dM\\Q84mPH`0@`0@`0@`0A`0_O`0@`0@SNnD;c:@aEo0P:mNTFR1m9jNVFV1k9fNYFY1i9bN[F]1f9`N\\F`1e9]N^Fb1c9]N^Fb1c9\\N_Fc1c9ZN^Ff1c9YN^Ff1c9XN_Fh1a9WN_Fi1c9TN_Fk1b9SN`Fl1a9SN`Fl1a9RN`Fn1a9QN`Fn1b9oM`FP2a9nM`F[1X:dNiEP1c:nNaEo0a:oNeEj0]:TOjEe0X:YOoEa0]<J7JWW_7"}, "label": "train cart on the left in front of the train in the back"}]}
{"id": 227205, "image": "COCO_train2014_000000227205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train right beind a yellow and black striped pole\"."}], "task": "refering", "answer_template": "The \"train right beind a yellow and black striped pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299], "bbox": [0.0, 0.37977083333333334, 0.20224999999999999, 0.7775208333333333], "iscrowd": 0, "area": 16060.455699999999, "rle": {"size": [480, 640], "counts": "m5m0S>j1oMZMSD`4T:`KkEi5R90LWJPGj5P9VJoFj5R9200000O101O00000O1000000O101O000O100000000O2O00000O1000000O101O00000O1000000O2O00000O100000000O2O00000O1000000O2O0000000O1000001N100H8eM[2dM\\Q84mPH`0@`0@`0@`0A`0_O`0@`0@SNnD;c:@aEo0P:mNTFR1m9jNVFV1k9fNYFY1i9bN[F]1f9`N\\F`1e9]N^Fb1c9]N^Fb1c9\\N_Fc1c9ZN^Ff1c9YN^Ff1c9XN_Fh1a9WN_Fi1c9TN_Fk1b9SN`Fl1a9SN`Fl1a9RN`Fn1a9QN`Fn1b9oM`FP2a9nM`F[1X:dNiEP1c:nNaEo0a:oNeEj0]:TOjEe0X:YOoEa0]<J7JWW_7"}, "label": "train right beind a yellow and black striped pole"}]}
{"id": 194438, "image": "COCO_train2014_000000194438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a army ship is in the sea which is back to the standing two person ' s\"."}], "task": "refering", "answer_template": "The \"a army ship is in the sea which is back to the standing two person ' s\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 34, 35, 37, 57, 58, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 121, 122, 123, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 139, 140, 145, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183], "bbox": [0.034765625, 0.055808080808080814, 1.0, 0.5164898989898989], "iscrowd": 0, "area": 38822.69414999998, "rle": {"size": [396, 640], "counts": "Pe81W<6M2N3N1O1O1O1O1N2O1O1O1O1O1O1N2DWOSEj0l:\\OnDS1c:<O1O1O1O1O1O000O0100000000O100O1O1O1O00100O1O1O1O1O1O1O100O1O00100O100O1O100O100DUEVOl:i0;1000O10000O10000O1000O010000O10000O10000000O10000001O1O1O1O1N2N2M3N1O2N2O1N2N2O1Nhjf04UUYO0O2O00000ZOJUE7i:LVE54Hl95oE34Il96oE13Km96nE21JP:5oE2OJR:5nE3MJS:5oE2LJU:5mE4KHW:6iE7NDX:d100O1O100000001O0001O00000000000001O00@jEjNV:V1nEfNR:Z1oEeNQ:[1PFdNP:`0mF_OT9>oFAQ9>PGBP9<SGCn8;TGDl8:YGDf8;g1M2O2M3N2Mfm69SRI>B4L2N10O6J5L5J2N101N1O10N101N100M4K4M3M4O0O1O2O0O2O2M2N3N1N201N1O1UFjN^8X1]GPO_8P1]GSOc8o0WGUOi8l0kF_OU9l1001O@UM_Gk2a8WM]Gi2c8YM[Gg2e8[MYGe2g8\\MYGc2g8_MVGb2j8_MTGb2l8^MSGc2m8]MRGd2o8[MPGf2P9;0000001O000000AgLkGY3U8gLlGX3T8hLlGX3T8hLlGX3T8iLkGW3U8iLkGW3U8iLlGV3T8kLkGU3[7RMgHINU3[7SMfHINT3\\7UMdHG0T3\\7VMcHF2T3[7WMaHF3S3\\7YM_HI0n2a7ZM^HOJg2h7[10000N2I7H8H8N20ZLiIW1W6dNoI[1Q6`NUJ_1k5\\NZJd1e5ZN_Je1a5eMZIMR1b2d5TMgI:;l2f6SMXIP3h6oLYIQ3g6oLYIQ3g6nLZIR3f6mL[IS3e6mL[IS3e6lL\\IT3n700000000000002N4L4L3M4L4L3NO0000000000000O100000000000000G9B>1O001O00001O001O001O001O1OgNlL_Io2a6TMaIg2_6\\MaIa2_6bM_I]2b6dM\\I\\2d6dM]I[2c6eM]I[2c6eM^IZ2b6gM]IZ2b6fMeHD7e2T7gMaHL7]2T7lM`HO8U2R7RNbH19l1S7WOiHi0W7@aH?_7W20000004L4PLbHj2^7PMiHo2X7oLiHQ3W7nLjHR3W7mLfHV3Z7iLdHZ3\\7eLbH^3_7aL_Ha3d7[L[Hg3i7TLVHn3R82001O1O1O1[LdGX3\\8dLjGZ3W8bLmG]3b8O2N1O1O000000000010O0000000001O000000001O00000000001O00000000001O006J7I5K000000000000001O00000001O00000001O000000001O6J00000000000000000000O10000000000000000000000000000000000O2O0000000000000000000000000000000002N1O1O2N1O4L3M000000001O0000000000001O000001O01O000000001O000000001O000000O100O1O100O100O1O100O100O10000O10000O10000O10001NnMQFi1o9QNWFo1S:000000000000000000000000000000000000000O100O10000O100O10000O100O10000O101N10000lK"}, "label": "a army ship is in the sea which is back to the standing two person ' s"}]}
{"id": 194438, "image": "COCO_train2014_000000194438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the war ship in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the war ship in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 34, 35, 37, 57, 58, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 121, 122, 123, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 139, 140, 145, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183], "bbox": [0.034765625, 0.055808080808080814, 1.0, 0.5164898989898989], "iscrowd": 0, "area": 38822.69414999998, "rle": {"size": [396, 640], "counts": "Pe81W<6M2N3N1O1O1O1O1N2O1O1O1O1O1O1N2DWOSEj0l:\\OnDS1c:<O1O1O1O1O1O000O0100000000O100O1O1O1O00100O1O1O1O1O1O1O100O1O00100O100O1O100O100DUEVOl:i0;1000O10000O10000O1000O010000O10000O10000000O10000001O1O1O1O1N2N2M3N1O2N2O1N2N2O1Nhjf04UUYO0O2O00000ZOJUE7i:LVE54Hl95oE34Il96oE13Km96nE21JP:5oE2OJR:5nE3MJS:5oE2LJU:5mE4KHW:6iE7NDX:d100O1O100000001O0001O00000000000001O00@jEjNV:V1nEfNR:Z1oEeNQ:[1PFdNP:`0mF_OT9>oFAQ9>PGBP9<SGCn8;TGDl8:YGDf8;g1M2O2M3N2Mfm69SRI>B4L2N10O6J5L5J2N101N1O10N101N100M4K4M3M4O0O1O2O0O2O2M2N3N1N201N1O1UFjN^8X1]GPO_8P1]GSOc8o0WGUOi8l0kF_OU9l1001O@UM_Gk2a8WM]Gi2c8YM[Gg2e8[MYGe2g8\\MYGc2g8_MVGb2j8_MTGb2l8^MSGc2m8]MRGd2o8[MPGf2P9;0000001O000000AgLkGY3U8gLlGX3T8hLlGX3T8hLlGX3T8iLkGW3U8iLkGW3U8iLlGV3T8kLkGU3[7RMgHINU3[7SMfHINT3\\7UMdHG0T3\\7VMcHF2T3[7WMaHF3S3\\7YM_HI0n2a7ZM^HOJg2h7[10000N2I7H8H8N20ZLiIW1W6dNoI[1Q6`NUJ_1k5\\NZJd1e5ZN_Je1a5eMZIMR1b2d5TMgI:;l2f6SMXIP3h6oLYIQ3g6oLYIQ3g6nLZIR3f6mL[IS3e6mL[IS3e6lL\\IT3n700000000000002N4L4L3M4L4L3NO0000000000000O100000000000000G9B>1O001O00001O001O001O001O1OgNlL_Io2a6TMaIg2_6\\MaIa2_6bM_I]2b6dM\\I\\2d6dM]I[2c6eM]I[2c6eM^IZ2b6gM]IZ2b6fMeHD7e2T7gMaHL7]2T7lM`HO8U2R7RNbH19l1S7WOiHi0W7@aH?_7W20000004L4PLbHj2^7PMiHo2X7oLiHQ3W7nLjHR3W7mLfHV3Z7iLdHZ3\\7eLbH^3_7aL_Ha3d7[L[Hg3i7TLVHn3R82001O1O1O1[LdGX3\\8dLjGZ3W8bLmG]3b8O2N1O1O000000000010O0000000001O000000001O00000000001O00000000001O006J7I5K000000000000001O00000001O00000001O000000001O6J00000000000000000000O10000000000000000000000000000000000O2O0000000000000000000000000000000002N1O1O2N1O4L3M000000001O0000000000001O000001O01O000000001O000000001O000000O100O1O100O100O1O100O100O10000O10000O10000O10001NnMQFi1o9QNWFo1S:000000000000000000000000000000000000000O100O10000O100O10000O100O10000O101N10000lK"}, "label": "the war ship in the right hand picture"}]}
{"id": 194438, "image": "COCO_train2014_000000194438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female sailor wearing a blue hard hat\"."}], "task": "refering", "answer_template": "The \"female sailor wearing a blue hard hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 217, 218, 219, 240, 241, 242], "bbox": [0.310765625, 0.32499999999999996, 0.567421875, 0.7891161616161616], "iscrowd": 0, "area": 14575.803599999997, "rle": {"size": [396, 640], "counts": "aV]21Y<10O01O1O001]OOkD2U;4dDM\\;`021N3N1N2O1N3M2O1N2O2lNlNkFV1Z8FcG<Z8IcG8Z8LcG7Y8NdG6V8NgG8R8MkG8m7MPH9i7LSH9g7KWH5g7Z2N1O10000O101N10000O101O0ON3N1N3M3N2M3N2M3M3N2M3N2M3M300O100000000000001O1O0100000O01000000O0100000O1SH[LR7d3nH^LR7a3mHdLP7[3nHkLo6T3PIQMn6l2RIWMm6g0_Hb0c0gNP72_HGNa1b0fNS7LeHIFf1a0eNV7GjHJ_Ok1?eNZ7AnHK[On1=fN]7[OUIP2]OfNb8[1]GeNc8\\1\\GdNd8\\1\\GdNd8]1ZGdNg8\\1XGdNh8]1WGcNi8^1VGbNj8^1VGbNj8_1UGaNk8`1TG`Nm8`1RG`Nn8`1RG`Nn8a1RG^Nn8c1QG]No8d1PG\\NQ9d1oF[NQ9e1oF[NQ9f1nFZNR9g1nFYNQ9h1nFXNR9h1nFXNS9h1mFWNS9a20000001O00003M3M4M2M3M4[MYF]2j9aMWF]2n9N2N2O1N1O2N2N2N2O1N2N2N2N2N2N101N2N2N2N2N2N2M3M3N2M2N3M3N2M3M3M3N2M3M3MajZ3"}, "label": "female sailor wearing a blue hard hat"}]}
{"id": 194438, "image": "COCO_train2014_000000194438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an women wearing blue helmet driving boat in the ocean\"."}], "task": "refering", "answer_template": "The \"an women wearing blue helmet driving boat in the ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 217, 218, 219, 240, 241, 242], "bbox": [0.310765625, 0.32499999999999996, 0.567421875, 0.7891161616161616], "iscrowd": 0, "area": 14575.803599999997, "rle": {"size": [396, 640], "counts": "aV]21Y<10O01O1O001]OOkD2U;4dDM\\;`021N3N1N2O1N3M2O1N2O2lNlNkFV1Z8FcG<Z8IcG8Z8LcG7Y8NdG6V8NgG8R8MkG8m7MPH9i7LSH9g7KWH5g7Z2N1O10000O101N10000O101O0ON3N1N3M3N2M3N2M3M3N2M3N2M3M300O100000000000001O1O0100000O01000000O0100000O1SH[LR7d3nH^LR7a3mHdLP7[3nHkLo6T3PIQMn6l2RIWMm6g0_Hb0c0gNP72_HGNa1b0fNS7LeHIFf1a0eNV7GjHJ_Ok1?eNZ7AnHK[On1=fN]7[OUIP2]OfNb8[1]GeNc8\\1\\GdNd8\\1\\GdNd8]1ZGdNg8\\1XGdNh8]1WGcNi8^1VGbNj8^1VGbNj8_1UGaNk8`1TG`Nm8`1RG`Nn8`1RG`Nn8a1RG^Nn8c1QG]No8d1PG\\NQ9d1oF[NQ9e1oF[NQ9f1nFZNR9g1nFYNQ9h1nFXNR9h1nFXNS9h1mFWNS9a20000001O00003M3M4M2M3M4[MYF]2j9aMWF]2n9N2N2O1N1O2N2N2N2O1N2N2N2N2N2N101N2N2N2N2N2N2M3M3N2M2N3M3N2M3M3M3N2M3M3MajZ3"}, "label": "an women wearing blue helmet driving boat in the ocean"}]}
{"id": 194438, "image": "COCO_train2014_000000194438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in yellow hard hat on the ocean\"."}], "task": "refering", "answer_template": "The \"a man in yellow hard hat on the ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236], "bbox": [0.055625, 0.21573232323232325, 0.311453125, 0.7640404040404041], "iscrowd": 0, "area": 24657.054400000005, "rle": {"size": [396, 640], "counts": "cV>8i;9I9M3M3M4L3L4M3M3M3M3M3fNWNmGm1P8WNlGl1Q8WNlGl1Q8WNlGl1R8VNkGn1S8SNiGQ2W8oMfGT2Y8nMcGU2]8kM`GX2_8iM^GZ2a8gM\\G\\2d8f0O100O101N100O100O100O100O1O100O2O0O1AVL\\Hj3d7?00000000O1000001O00000O10000000000O10001O00O1000O10000000000RO_KPJb4o5aKoI_4P6cKoI]4P6fKnIY4Q6kKkIW4T6kKhIX4W6kKdIX4[6kK`IX4_6k0O1O1O1O1O1O1O1O001O1O1O1O1O1O100O1O1O1O10O01O1O100O100O100O100O100O010O100O100O11O00001O0000001N10001O000O101N100O2M2N2N3M2N2O1N3[LaIY1c6eN^IX1f6eN]IW1g6gNZIV1j6gNXIjNL`1P7DVIgN1_1m6GTIgN5\\1k6JTIeN7[1i6MVJNn5OWJLl51ZJHl53YJGn53XJCQ68TJ]OX64^`Z5"}, "label": "a man in yellow hard hat on the ocean"}]}
{"id": 194438, "image": "COCO_train2014_000000194438.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the yellow helmet next to the man in the blue helmet\"."}], "task": "refering", "answer_template": "The \"the man in the yellow helmet next to the man in the blue helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236], "bbox": [0.055625, 0.21573232323232325, 0.311453125, 0.7640404040404041], "iscrowd": 0, "area": 24657.054400000005, "rle": {"size": [396, 640], "counts": "cV>8i;9I9M3M3M4L3L4M3M3M3M3M3fNWNmGm1P8WNlGl1Q8WNlGl1Q8WNlGl1R8VNkGn1S8SNiGQ2W8oMfGT2Y8nMcGU2]8kM`GX2_8iM^GZ2a8gM\\G\\2d8f0O100O101N100O100O100O100O1O100O2O0O1AVL\\Hj3d7?00000000O1000001O00000O10000000000O10001O00O1000O10000000000RO_KPJb4o5aKoI_4P6cKoI]4P6fKnIY4Q6kKkIW4T6kKhIX4W6kKdIX4[6kK`IX4_6k0O1O1O1O1O1O1O1O001O1O1O1O1O1O100O1O1O1O10O01O1O100O100O100O100O100O010O100O100O11O00001O0000001N10001O000O101N100O2M2N2N3M2N2O1N3[LaIY1c6eN^IX1f6eN]IW1g6gNZIV1j6gNXIjNL`1P7DVIgN1_1m6GTIgN5\\1k6JTIeN7[1i6MVJNn5OWJLl51ZJHl53YJGn53XJCQ68TJ]OX64^`Z5"}, "label": "the man in the yellow helmet next to the man in the blue helmet"}]}
{"id": 186246, "image": "COCO_train2014_000000186246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman waring a lather jacket and holding an umrella\"."}], "task": "refering", "answer_template": "The \"the woman waring a lather jacket and holding an umrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 127, 128, 129, 130, 144, 145, 146, 147, 161, 162, 163, 164, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 228, 229, 230, 231, 232, 246, 247, 248, 249, 263, 264, 265, 266, 280, 281, 282, 283, 297, 298, 299, 314, 315, 316, 331, 332, 333, 348, 349, 350, 365, 366, 367], "bbox": [0.45739583333333333, 0.2711875, 0.7413958333333334, 0.946078125], "iscrowd": 0, "area": 35727.15295, "rle": {"size": [640, 480], "counts": "`jY49dc0:F:F:F:F:F:F:]BhMm8b2jFhMk8b2lFiMh8a2PGhMf8a2QGfMg8d2PG_Ml8k2lFXMo8R3hFQMT9Y3cFjLY9`3_FbL]9d3^F_L]7aNdHU5K^L[7L\\Gk3V1[LY7l5dHWJW7o5eHTJV7Q6hHQJU7S6gHQJV7R6fHQJY7P6eHRJZ7P6bHSJ]7o5_HWJ]7j5aH\\JZ7f5bHaJY7`5dHhJV7Z5gHjJU7Q900000000O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O]N[ImEd6o9bIPF]6l9iISFW6h9oIWFP6e9VJZFi5b9]J]Fc5^9cJaF\\5[9kJcFT5Y9RKfFn4V9WKiFi4S9\\KlFd4R9_KmFa4R9aKmF_4Q9dKnF\\4P9hKnFX4Q9jKmFW4Q9lKmFV4Q9lKmFU4Q9nKnFR4Q9PLmFQ4Q9RLmFo3R9SLlFn3R9VLlFj3R9YLlFi3R9YLlFh3R9[LlFf3S9\\LlFd3R9_LlFb3S9`LkFa3S9cLjF^3T9eLkF[3P9kLnFW3k8R4K5K5J6K5J6K5K6I6KU1jNU2lMk0TO9H6I6N2N2N2N2N2N2N3M2N200O1O1O1O1O1O2N1O1O100O3M7I7E;E:F;EiU]2"}, "label": "the woman waring a lather jacket and holding an umrella"}]}
{"id": 186246, "image": "COCO_train2014_000000186246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black jacket holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a woman in a black jacket holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 127, 128, 129, 130, 144, 145, 146, 147, 161, 162, 163, 164, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 228, 229, 230, 231, 232, 246, 247, 248, 249, 263, 264, 265, 266, 280, 281, 282, 283, 297, 298, 299, 314, 315, 316, 331, 332, 333, 348, 349, 350, 365, 366, 367], "bbox": [0.45739583333333333, 0.2711875, 0.7413958333333334, 0.946078125], "iscrowd": 0, "area": 35727.15295, "rle": {"size": [640, 480], "counts": "`jY49dc0:F:F:F:F:F:F:]BhMm8b2jFhMk8b2lFiMh8a2PGhMf8a2QGfMg8d2PG_Ml8k2lFXMo8R3hFQMT9Y3cFjLY9`3_FbL]9d3^F_L]7aNdHU5K^L[7L\\Gk3V1[LY7l5dHWJW7o5eHTJV7Q6hHQJU7S6gHQJV7R6fHQJY7P6eHRJZ7P6bHSJ]7o5_HWJ]7j5aH\\JZ7f5bHaJY7`5dHhJV7Z5gHjJU7Q900000000O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O]N[ImEd6o9bIPF]6l9iISFW6h9oIWFP6e9VJZFi5b9]J]Fc5^9cJaF\\5[9kJcFT5Y9RKfFn4V9WKiFi4S9\\KlFd4R9_KmFa4R9aKmF_4Q9dKnF\\4P9hKnFX4Q9jKmFW4Q9lKmFV4Q9lKmFU4Q9nKnFR4Q9PLmFQ4Q9RLmFo3R9SLlFn3R9VLlFj3R9YLlFi3R9YLlFh3R9[LlFf3S9\\LlFd3R9_LlFb3S9`LkFa3S9cLjF^3T9eLkF[3P9kLnFW3k8R4K5K5J6K5J6K5K6I6KU1jNU2lMk0TO9H6I6N2N2N2N2N2N2N3M2N200O1O1O1O1O1O2N1O1O100O3M7I7E;E:F;EiU]2"}, "label": "a woman in a black jacket holding an umbrella"}]}
{"id": 186246, "image": "COCO_train2014_000000186246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing a white shirt with her back turned\"."}], "task": "refering", "answer_template": "The \"woman wearing a white shirt with her back turned\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 132, 133, 134, 148, 149, 150, 151, 165, 166, 167, 168, 182, 183, 184, 185, 199, 200, 201, 202, 216, 217, 218, 219, 233, 234, 235, 251, 252, 268, 269, 285, 286], "bbox": [0.7230833333333333, 0.264890625, 0.9436874999999999, 0.7147343749999999], "iscrowd": 0, "area": 15572.106150000003, "rle": {"size": [640, 480], "counts": "`Ti61lc04M3L5L3M3R_O_OX>e0bAIS>:hAKV>8cANZ>5aAO^>4[A2c>0VA7g>M_@l0_?VO\\@n0b?VOX@o0f?SOV@Q1g?e1O2M3N2N1O2N2N2N1O2N2M7J5K5K6J5K5K4I7J7H7J6J6I8I7F:G:G8J3L4M2N3N101O0O101O0O2O0OL5F:UOj0I8K5J6K5I7H8L4O1O1N4M4L3L5L3M4L3L4M3M4J5I7J7I8G8^Ob0UOk0E9H8G9L4M3L4M3L4M2M4M2M4M2M4M2M4M2N3L3N3Lod`0"}, "label": "woman wearing a white shirt with her back turned"}]}
{"id": 186246, "image": "COCO_train2014_000000186246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white sweater and black skirt with her hair in a ponytail\"."}], "task": "refering", "answer_template": "The \"a woman in a white sweater and black skirt with her hair in a ponytail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 132, 133, 134, 148, 149, 150, 151, 165, 166, 167, 168, 182, 183, 184, 185, 199, 200, 201, 202, 216, 217, 218, 219, 233, 234, 235, 251, 252, 268, 269, 285, 286], "bbox": [0.7230833333333333, 0.264890625, 0.9436874999999999, 0.7147343749999999], "iscrowd": 0, "area": 15572.106150000003, "rle": {"size": [640, 480], "counts": "`Ti61lc04M3L5L3M3R_O_OX>e0bAIS>:hAKV>8cANZ>5aAO^>4[A2c>0VA7g>M_@l0_?VO\\@n0b?VOX@o0f?SOV@Q1g?e1O2M3N2N1O2N2N2N1O2N2M7J5K5K6J5K5K4I7J7H7J6J6I8I7F:G:G8J3L4M2N3N101O0O101O0O2O0OL5F:UOj0I8K5J6K5I7H8L4O1O1N4M4L3L5L3M4L3L4M3M4J5I7J7I8G8^Ob0UOk0E9H8G9L4M3L4M3L4M2M4M2M4M2M4M2M4M2N3L3N3Lod`0"}, "label": "a woman in a white sweater and black skirt with her hair in a ponytail"}]}
{"id": 71558, "image": "COCO_train2014_000000071558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large sheep under a tree\"."}], "task": "refering", "answer_template": "The \"a large sheep under a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 261, 264, 284], "bbox": [0.327140625, 0.4732552693208431, 0.594015625, 0.8173536299765808], "iscrowd": 0, "area": 13863.31235, "rle": {"size": [427, 640], "counts": "V]g23l<i0XO<C>I6N2O1O2M2O1O2N1N2O1nEmMj8U2`FaN]9i2N3M2N2M2O000OTOfFiMZ9U2kFhMU9U2QGhMo8V2WGfMh8X2^GeMc8X2bGeM^8X2hGeMX8Y2lGeMT8Z2oGdMQ8\\2RHaMn7_2_100000001O00000000000000001O00000000000000001O0000000000000001O01O00000000000000001O0000000000000000001O00000000000O0100000O1000000O19G:F9G4L3M4L2M100000000FbLgF^3X9cLhFo2MWMZ9JjFf26_MP9KjF]2a0fMd8MlFV2i0kMZ80mFU2j0jMY80nFV2k0hMV83oFT2m0hMS84QGT2n0fMP86SGT2U:lMjEU2U:kMlEU2S:lMmES2S:mMnES2R:mMmET2R:mMkEV2T:lMhEV2Y:>0O1O1O100O1O2N1O1O100O1O1O1O1O1O10O01N2O1O1O1O1O1O001O1O1O2N1L5L3M4L3jN\\Db0h;]O\\D<g;C_D1i;N[DEn;:c0M2M4MmV\\3"}, "label": "a large sheep under a tree"}]}
{"id": 71558, "image": "COCO_train2014_000000071558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ram with a black and white face looking at the camera , standing next to a large tree trunk\"."}], "task": "refering", "answer_template": "The \"a ram with a black and white face looking at the camera , standing next to a large tree trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 261, 264, 284], "bbox": [0.327140625, 0.4732552693208431, 0.594015625, 0.8173536299765808], "iscrowd": 0, "area": 13863.31235, "rle": {"size": [427, 640], "counts": "V]g23l<i0XO<C>I6N2O1O2M2O1O2N1N2O1nEmMj8U2`FaN]9i2N3M2N2M2O000OTOfFiMZ9U2kFhMU9U2QGhMo8V2WGfMh8X2^GeMc8X2bGeM^8X2hGeMX8Y2lGeMT8Z2oGdMQ8\\2RHaMn7_2_100000001O00000000000000001O00000000000000001O0000000000000001O01O00000000000000001O0000000000000000001O00000000000O0100000O1000000O19G:F9G4L3M4L2M100000000FbLgF^3X9cLhFo2MWMZ9JjFf26_MP9KjF]2a0fMd8MlFV2i0kMZ80mFU2j0jMY80nFV2k0hMV83oFT2m0hMS84QGT2n0fMP86SGT2U:lMjEU2U:kMlEU2S:lMmES2S:mMnES2R:mMmET2R:mMkEV2T:lMhEV2Y:>0O1O1O100O1O2N1O1O100O1O1O1O1O1O10O01N2O1O1O1O1O1O001O1O1O2N1L5L3M4L3jN\\Db0h;]O\\D<g;C_D1i;N[DEn;:c0M2M4MmV\\3"}, "label": "a ram with a black and white face looking at the camera , standing next to a large tree trunk"}]}
{"id": 366469, "image": "COCO_train2014_000000366469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl having a pizza picnic\"."}], "task": "refering", "answer_template": "The \"a girl having a pizza picnic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 26, 27, 28, 29, 38, 39, 40, 41, 50, 51, 52, 53, 54, 61, 62, 63, 64, 65, 66, 73, 74, 75, 76, 77, 78, 79, 86, 87, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 108, 109, 110, 111, 112, 113, 115, 116, 120, 121, 122, 123, 124, 125, 126, 127, 128, 132, 133, 134, 135, 136, 137, 138, 139, 140, 144, 145, 146], "bbox": [0.0, 0.08988, 0.7321921921921922, 0.72134], "iscrowd": 0, "area": 40196.074, "rle": {"size": [500, 333], "counts": "U81X2d0^:^OXDj1g;S1N101N1O2N100O2N101N1O2OO01O1O10O01O100O1O00100O1O10O01O100O001O100O1O010O1[N]LXGd3e8aLXG_3f8eLYG[3c8kLZGV3c8nL[GS3a8RM^Gn2_8TMaGl2[8XMeGi2U8\\MlGd2m7bMSH_2k7bMUH^2j7cMUH_2i7bMWH_2g7bMYH_2e7bM[H^2d7cM\\H^2b7cM]H_2a7cM^H^2`7cM`H]2_7dMaH]2\\7eMdH\\2Y7fMgHZ2W7jMfHX2W7mMfHT2X7PNeHQ2X7VNcHj1\\7aNZH`1d7Y3N2N1N3N2N1O2M3L3L4K5K6N1001O01O00000010O01O001O010O010O10O010O0100O010O010O010000O10O0100O1O100O2O0O101N100O2O0O1O2O0O101N101N2O0O2O1N2N101N2O0O2O1N10n0QO4M3L4M2N1O1O001O1O001O1O001QOeFdK[9R4nFoKQ9h3YGWLh8d3\\G]Lc8_3bGaL^8]2nFWMi0=X8Z2ZGlLc0i0S8Y2iHhMW7T2mHkMS7S2oHmMR7o1RIPNn6n1UIRNk6k1WIUNj6g1ZIXNf6f1\\IZNe6e1\\IZNd6h1ZIYNf6i1XIVNh6l1WISNj6n1TIRNl6Q2RInMo6S2oHnMR7R2mHmMU7S2jHlMW7U2gHkM[7V2bHjM`7V2]HlMd7T2ZHlMh7T2VHlMl7U2QHkMQ8U2mGkMU8U2iGlMX8T2eGmM]8T2`GlMc8S2[GmMk8o1SGPNT9l1kFSN[9i1dFVNc9e1\\FZNk9a1SF^NQ:`1oE_NT:`1kE_NZ:^1eE`Na:]1^EbNh:Y1XEfNm:V1UEfNQ;W1oDgNS;X1nDfNT;Y1nDcNU;]1kDaNW;^1a1O101N2O0O2O0O2O1N12O2M2O?A7HO1O1N2O1O1N2O1O1N2O1O1L4L4L4L5K5J6K5K5K5K5KWU[1"}, "label": "a girl having a pizza picnic"}]}
{"id": 366469, "image": "COCO_train2014_000000366469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with long dark hair sitting on the ground eating pizza\"."}], "task": "refering", "answer_template": "The \"girl with long dark hair sitting on the ground eating pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 26, 27, 28, 29, 38, 39, 40, 41, 50, 51, 52, 53, 54, 61, 62, 63, 64, 65, 66, 73, 74, 75, 76, 77, 78, 79, 86, 87, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 108, 109, 110, 111, 112, 113, 115, 116, 120, 121, 122, 123, 124, 125, 126, 127, 128, 132, 133, 134, 135, 136, 137, 138, 139, 140, 144, 145, 146], "bbox": [0.0, 0.08988, 0.7321921921921922, 0.72134], "iscrowd": 0, "area": 40196.074, "rle": {"size": [500, 333], "counts": "U81X2d0^:^OXDj1g;S1N101N1O2N100O2N101N1O2OO01O1O10O01O100O1O00100O1O10O01O100O001O100O1O010O1[N]LXGd3e8aLXG_3f8eLYG[3c8kLZGV3c8nL[GS3a8RM^Gn2_8TMaGl2[8XMeGi2U8\\MlGd2m7bMSH_2k7bMUH^2j7cMUH_2i7bMWH_2g7bMYH_2e7bM[H^2d7cM\\H^2b7cM]H_2a7cM^H^2`7cM`H]2_7dMaH]2\\7eMdH\\2Y7fMgHZ2W7jMfHX2W7mMfHT2X7PNeHQ2X7VNcHj1\\7aNZH`1d7Y3N2N1N3N2N1O2M3L3L4K5K6N1001O01O00000010O01O001O010O010O10O010O0100O010O010O010000O10O0100O1O100O2O0O101N100O2O0O1O2O0O101N101N2O0O2O1N2N101N2O0O2O1N10n0QO4M3L4M2N1O1O001O1O001O1O001QOeFdK[9R4nFoKQ9h3YGWLh8d3\\G]Lc8_3bGaL^8]2nFWMi0=X8Z2ZGlLc0i0S8Y2iHhMW7T2mHkMS7S2oHmMR7o1RIPNn6n1UIRNk6k1WIUNj6g1ZIXNf6f1\\IZNe6e1\\IZNd6h1ZIYNf6i1XIVNh6l1WISNj6n1TIRNl6Q2RInMo6S2oHnMR7R2mHmMU7S2jHlMW7U2gHkM[7V2bHjM`7V2]HlMd7T2ZHlMh7T2VHlMl7U2QHkMQ8U2mGkMU8U2iGlMX8T2eGmM]8T2`GlMc8S2[GmMk8o1SGPNT9l1kFSN[9i1dFVNc9e1\\FZNk9a1SF^NQ:`1oE_NT:`1kE_NZ:^1eE`Na:]1^EbNh:Y1XEfNm:V1UEfNQ;W1oDgNS;X1nDfNT;Y1nDcNU;]1kDaNW;^1a1O101N2O0O2O0O2O1N12O2M2O?A7HO1O1N2O1O1N2O1O1N2O1O1L4L4L4L5K5J6K5K5K5K5KWU[1"}, "label": "girl with long dark hair sitting on the ground eating pizza"}]}
{"id": 6026, "image": "COCO_train2014_000000006026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby girl with red hair , wearing a pink shirt , being held by an adult male in front of a lake\"."}], "task": "refering", "answer_template": "The \"a baby girl with red hair , wearing a pink shirt , being held by an adult male in front of a lake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 283, 284, 285, 286, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 357, 377, 378, 379, 380, 381, 382], "bbox": [0.22921874999999997, 0.5460625, 0.632015625, 0.9887708333333334], "iscrowd": 0, "area": 23169.005650000006, "rle": {"size": [480, 640], "counts": "oYU22m>1O1O2N1O1N3N1O1O2N1O1N3N1O1O2N1N2O2N1O1O2N1N2O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1O1N3N1O1O2O0O2O00001O00001O0O101O0oN^MdEb2Y:eMcE\\2Y:jMdEV2X:RNdEn1Y:YNcEh1Y:_NcEa1Z:fNaE[1[:h1L3M3L5L300O2N1O1O2N1O1O2N1O101N1O1O2N1O1O2N1O1O2O0O101O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0100O1O10O01O100O1O010O1O100O1O0RMPGMP90RG0o8NRG2o8KSG6m8HTG8m8EVG:j8EWG<i8AYG?h8_OYGa0h8]OYGd0g8YO[Gg0f8gNkGY1U8cNoG^1Q8\\NTHd1m7VNXHj1i7QN[HP2Y8VMlGj2W:N2UNPDc0Q<ZORDd0P<YOSDe0o;XOTDg0m;UOVDj0m;ROVDl0P<mNSDR1Q<hNQDW1n<N2N1O2N2N2N1O1O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O000000O1O1O100O1O100O1O1O100O1O100O1O1O1001O1O2N1O1O1O1N3M2N2N2N3M2N_U^3"}, "label": "a baby girl with red hair , wearing a pink shirt , being held by an adult male in front of a lake"}]}
{"id": 6026, "image": "COCO_train2014_000000006026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby\"."}], "task": "refering", "answer_template": "The \"baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 283, 284, 285, 286, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 357, 377, 378, 379, 380, 381, 382], "bbox": [0.22921874999999997, 0.5460625, 0.632015625, 0.9887708333333334], "iscrowd": 0, "area": 23169.005650000006, "rle": {"size": [480, 640], "counts": "oYU22m>1O1O2N1O1N3N1O1O2N1O1N3N1O1O2N1N2O2N1O1O2N1N2O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1O1N3N1O1O2O0O2O00001O00001O0O101O0oN^MdEb2Y:eMcE\\2Y:jMdEV2X:RNdEn1Y:YNcEh1Y:_NcEa1Z:fNaE[1[:h1L3M3L5L300O2N1O1O2N1O1O2N1O101N1O1O2N1O1O2N1O1O2O0O101O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0100O1O10O01O100O1O010O1O100O1O0RMPGMP90RG0o8NRG2o8KSG6m8HTG8m8EVG:j8EWG<i8AYG?h8_OYGa0h8]OYGd0g8YO[Gg0f8gNkGY1U8cNoG^1Q8\\NTHd1m7VNXHj1i7QN[HP2Y8VMlGj2W:N2UNPDc0Q<ZORDd0P<YOSDe0o;XOTDg0m;UOVDj0m;ROVDl0P<mNSDR1Q<hNQDW1n<N2N1O2N2N2N1O1O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O000000O1O1O100O1O100O1O1O100O1O100O1O1O1001O1O2N1O1O1O1N3M2N2N2N3M2N_U^3"}, "label": "baby"}]}
{"id": 6026, "image": "COCO_train2014_000000006026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with child , catching a frisby\"."}], "task": "refering", "answer_template": "The \"man with child , catching a frisby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0, 0.07400000000000001, 0.49607812500000004, 0.9887916666666666], "iscrowd": 0, "area": 57586.74465, "rle": {"size": [480, 640], "counts": "_1e1g8X1fI3\\1j2UOdLl0\\3ZO]Lg0c3^OXLb0h3CRL>n3GlK:T4KgK5Y40aK1_44[KMe44ZKLf43ZKMg43XKNh42XKNh42WKOi41VK0k4OTK2l4NTK3k4MTK4l4LSK5m4KSK5m4KRK6n4JQK8n4HRK8n4GRK:n4FQK;o4EQK;o4EPK=o4CPK>P5BPKiNYOjNg5]2oJhN]OjNd5^2oJfNAjN`5`2oJ`MZOB;=[5a2oJ]ME^O3c0Y5b2oJZMM[ONg0V5d2oJWM6WOHl0S5f2nJTMa0SO@R1Q5g2nJQMi0PO[OW1m4g2PKoLh18X3i2PKkLm1;S3j2oJiLR2<o2k2oJfLU2?l2k2oJcLX2c0h2j2oJ`L]2f0d2j2oJ^L`2g0a2k2oJ]La2h0`2k2nJ]Lc2h0_2k2nJ\\Ld2j0]2j2oJ[Le2k0\\2j2oJ[Le2k0\\2j2nJ[Lg2k0[2j2nJZLi2l0X2j2oJYLj2m0W2j2nJYLl2m0V2j2nJYLl2m0V2j2nJXLm2o0T2i2nJXLP3n0R2i2oJXLQ3n0P2j2nJXLT3n0m1l2mJULX3m0l1U3dJoKa3j0l1c3VNZLl1e3VNYLk1f3XNWLi1h3]NQLe1o3aNjK`1U4gNdKZ1[4f4O1O100O1O1O1O1O100O100001O00000000001O000000001O00000000001O00000000001O000000001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O1O2N1O2aJjJP1W5iNoJW1S5aNTK^1m4[NYKe1h4TN_Kk1c4mMdK;oNTN^5Y1jKa0kNUN]5R1oKf0hNVNZ5m0TLl0eNUN`5`0RLX1aNWNg55jKb1cNWNm5MaKj1eNWNo5J\\Km1hNXNm5J\\Kk1kNYNj5K[Kk1nNXNh5L[Ki1POZNf5LZKh1TOZNc5MZKg1UO\\Na5K[Kh1UO]Na5J[Kg1UO_Na5I[Ke1WOaN_5IZKe1XObNa8\\1`GdN`8[1aGeN_8Z1bGfN_8X1bGhN^8W1cGiN]8V1dGjN\\8T1fGlNZ8S1gGmNY8R1iGmNW8R1jGnNW8P1jGPOV8o0kGQOU8n0lGROT8m0mGSOS8l0nGTOR8j0PHVOP8i0QHWOo7h0RHXOo7f0QH[Oo7d0QH]Oo7b0QH_Oo7`0QHAo7>QHCo7;RHFn79RHHn77RHJo74WGhNZOU1_92XGkNVOT1b90YGmNSOT1d9NYGQOPOR1g9LZGTOmNP1i9KZGXOjNn0l9H\\G[OfNn0n9F]G^ObNm0R:C\\GC_Nk0U:A]GU1c8jN]GW1c8hN^GX1b8gN^GZ1b8eN_G[1a8eN_G[1b8dN]G]1c8cN]G]1c8cN\\G^1d8bN\\G^1d8bN\\G^1d8bN[G_1e8aN[G_1e8aNZG`1g8_NYGa1g8_NXGb1i8]NWGc1l8ZNTGf1o8WNRGh1Q9UNoFk1T9RNlFn1W9oMiFQ2Z9lMgFS2\\9jMdFV2_9gMaFY2S;01O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O00002N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O2O0O2N1O2N2N1O2N2N1O2N2N1O2NWnf4"}, "label": "man with child , catching a frisby"}]}
{"id": 6026, "image": "COCO_train2014_000000006026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a t - shirt holding a baby and catching a blue frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a t - shirt holding a baby and catching a blue frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0, 0.07400000000000001, 0.49607812500000004, 0.9887916666666666], "iscrowd": 0, "area": 57586.74465, "rle": {"size": [480, 640], "counts": "_1e1g8X1fI3\\1j2UOdLl0\\3ZO]Lg0c3^OXLb0h3CRL>n3GlK:T4KgK5Y40aK1_44[KMe44ZKLf43ZKMg43XKNh42XKNh42WKOi41VK0k4OTK2l4NTK3k4MTK4l4LSK5m4KSK5m4KRK6n4JQK8n4HRK8n4GRK:n4FQK;o4EQK;o4EPK=o4CPK>P5BPKiNYOjNg5]2oJhN]OjNd5^2oJfNAjN`5`2oJ`MZOB;=[5a2oJ]ME^O3c0Y5b2oJZMM[ONg0V5d2oJWM6WOHl0S5f2nJTMa0SO@R1Q5g2nJQMi0PO[OW1m4g2PKoLh18X3i2PKkLm1;S3j2oJiLR2<o2k2oJfLU2?l2k2oJcLX2c0h2j2oJ`L]2f0d2j2oJ^L`2g0a2k2oJ]La2h0`2k2nJ]Lc2h0_2k2nJ\\Ld2j0]2j2oJ[Le2k0\\2j2oJ[Le2k0\\2j2nJ[Lg2k0[2j2nJZLi2l0X2j2oJYLj2m0W2j2nJYLl2m0V2j2nJYLl2m0V2j2nJXLm2o0T2i2nJXLP3n0R2i2oJXLQ3n0P2j2nJXLT3n0m1l2mJULX3m0l1U3dJoKa3j0l1c3VNZLl1e3VNYLk1f3XNWLi1h3]NQLe1o3aNjK`1U4gNdKZ1[4f4O1O100O1O1O1O1O100O100001O00000000001O000000001O00000000001O00000000001O000000001O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O1O2N1O2aJjJP1W5iNoJW1S5aNTK^1m4[NYKe1h4TN_Kk1c4mMdK;oNTN^5Y1jKa0kNUN]5R1oKf0hNVNZ5m0TLl0eNUN`5`0RLX1aNWNg55jKb1cNWNm5MaKj1eNWNo5J\\Km1hNXNm5J\\Kk1kNYNj5K[Kk1nNXNh5L[Ki1POZNf5LZKh1TOZNc5MZKg1UO\\Na5K[Kh1UO]Na5J[Kg1UO_Na5I[Ke1WOaN_5IZKe1XObNa8\\1`GdN`8[1aGeN_8Z1bGfN_8X1bGhN^8W1cGiN]8V1dGjN\\8T1fGlNZ8S1gGmNY8R1iGmNW8R1jGnNW8P1jGPOV8o0kGQOU8n0lGROT8m0mGSOS8l0nGTOR8j0PHVOP8i0QHWOo7h0RHXOo7f0QH[Oo7d0QH]Oo7b0QH_Oo7`0QHAo7>QHCo7;RHFn79RHHn77RHJo74WGhNZOU1_92XGkNVOT1b90YGmNSOT1d9NYGQOPOR1g9LZGTOmNP1i9KZGXOjNn0l9H\\G[OfNn0n9F]G^ObNm0R:C\\GC_Nk0U:A]GU1c8jN]GW1c8hN^GX1b8gN^GZ1b8eN_G[1a8eN_G[1b8dN]G]1c8cN]G]1c8cN\\G^1d8bN\\G^1d8bN\\G^1d8bN[G_1e8aN[G_1e8aNZG`1g8_NYGa1g8_NXGb1i8]NWGc1l8ZNTGf1o8WNRGh1Q9UNoFk1T9RNlFn1W9oMiFQ2Z9lMgFS2\\9jMdFV2_9gMaFY2S;01O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O00002N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O2O0O2N1O2N2N1O2N2N1O2N2N1O2NWnf4"}, "label": "a man in a t - shirt holding a baby and catching a blue frisbee"}]}
{"id": 276365, "image": "COCO_train2014_000000276365.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"multi - colored circles on poles to the left of a little girl wearing a pink shirt\"."}], "task": "refering", "answer_template": "The \"multi - colored circles on poles to the left of a little girl wearing a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 90, 91, 92, 93, 109, 126, 127, 145], "bbox": [0.0019399999999999999, 0.31333333333333335, 0.18847999999999998, 0.6423466666666666], "iscrowd": 0, "area": 5322.99495, "rle": {"size": [375, 500], "counts": "k?>n:?L4M3L4M0O1O001O001O0O2O001O001O0O2O001O001N101O0ZGmNh6R3O000O2O001O00001N101O001OA?ZOf0[Oe0J600O10000000000000000O101O001O00001N101O001O001O001N101O001O001O001N101O001N2N2N2N2N2N3M2N2O1N2N2N2N[ad4"}, "label": "multi - colored circles on poles to the left of a little girl wearing a pink shirt"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a knife on the right of the plate\"."}], "task": "refering", "answer_template": "The \"a knife on the right of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 127, 143, 158, 159, 173, 174, 175, 189, 190, 191, 205, 206, 221, 222, 237, 238, 253, 254], "bbox": [0.8101746724890829, 0.28090625, 1.0, 0.689890625], "iscrowd": 0, "area": 7993.02195, "rle": {"size": [640, 458], "counts": "QXX73ic07I7J6I7I7I7J6I7I7I7J6I7I7I7J6I6J4M4K4L5K5L3L5K4N3N1O2M2OO0N3M3N1N3M2O2M2N3N2M2N3M2O2M3M2O2M2N3M3N1N3M2O2M2N3M2O1N2N101N2N101N2N1O2O1N1O2O1N1O2N2O0O2N2O0O2N2O0O2N2N101N2N9HXJ"}, "label": "a knife on the right of the plate"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife to the right of two forks and a knife\"."}], "task": "refering", "answer_template": "The \"the knife to the right of two forks and a knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 127, 143, 158, 159, 173, 174, 175, 189, 190, 191, 205, 206, 221, 222, 237, 238, 253, 254], "bbox": [0.8101746724890829, 0.28090625, 1.0, 0.689890625], "iscrowd": 0, "area": 7993.02195, "rle": {"size": [640, 458], "counts": "QXX73ic07I7J6I7I7I7J6I7I7I7J6I7I7I7J6I6J4M4K4L5K5L3L5K4N3N1O2M2OO0N3M3N1N3M2O2M2N3N2M2N3M2O2M3M2O2M2N3M3N1N3M2O2M2N3M2O1N2N101N2N101N2N1O2O1N1O2O1N1O2N2O0O2N2O0O2N2O0O2N2N101N2N9HXJ"}, "label": "the knife to the right of two forks and a knife"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a steel knife which is inbetween two forks\"."}], "task": "refering", "answer_template": "The \"a steel knife which is inbetween two forks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 144, 145, 160, 161, 162, 177, 178, 179, 194, 195, 196, 211, 212, 213, 228, 229, 230, 245, 246, 247, 262, 263, 278, 279, 280, 295, 296], "bbox": [0.0, 0.37809375, 0.5470742358078603, 0.8287968749999999], "iscrowd": 0, "area": 9254.391349999996, "rle": {"size": [640, 458], "counts": "b7d0]c000000000000000001O0000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000000000000001O0000000O101O000000001O00000000001O000O10001O000O1000001N100000001N100000000O2O0000000O2O00000O10001O0001O0000010O000000010O0000000001O000000001O00000000001O0O1000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O000000001O01O000000010O00000010O00000001O01O0001O0001O0001O0000010O0000000010O000001O01O0001O2N3M2N3L3M4K4M4K4M4L1N4M3M2M4M2NckP4"}, "label": "a steel knife which is inbetween two forks"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dirty knife sitting by a slice of pizza on a round white plate\"."}], "task": "refering", "answer_template": "The \"a dirty knife sitting by a slice of pizza on a round white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 144, 145, 160, 161, 162, 177, 178, 179, 194, 195, 196, 211, 212, 213, 228, 229, 230, 245, 246, 247, 262, 263, 278, 279, 280, 295, 296], "bbox": [0.0, 0.37809375, 0.5470742358078603, 0.8287968749999999], "iscrowd": 0, "area": 9254.391349999996, "rle": {"size": [640, 458], "counts": "b7d0]c000000000000000001O0000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000000000000001O0000000O101O000000001O00000000001O000O10001O000O1000001N100000001N100000000O2O0000000O2O00000O10001O0001O0000010O000000010O0000000001O000000001O00000000001O0O1000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O000000001O01O000000010O00000010O00000001O01O0001O0001O0001O0000010O0000000010O000001O01O0001O2N3M2N3L3M4K4M4K4M4L1N4M3M2M4M2NckP4"}, "label": "a dirty knife sitting by a slice of pizza on a round white plate"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pizza bigger than the others\"."}], "task": "refering", "answer_template": "The \"a slice of pizza bigger than the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 85, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 179, 180, 181, 182, 183, 184, 185, 186], "bbox": [0.18526200873362444, 0.2, 0.8352838427947598, 0.51909375], "iscrowd": 0, "area": 37437.75545, "rle": {"size": [640, 458], "counts": "PXe1i0Xc0R1mNS1mNS1nNR1mNT1mN2M01O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001N10001O001O0O101O001O0O101O001O0O101O001O0O101O001O0O2O00001O0O2O00001O001N10001O001N10001O001N10001O0O2O00001N101O0O2O00001N101O000O2O001N10001O0O2O00001N101O0O101O001N101O0O101O0O2O0O101O0O2O0O101O0O2O0O101N101O0O101N101O0O2O0O101O0O2O0O101O0O2O0O101O0O2O0O101O0O2N1O101N1O2N101N1O1O2N101N1O1O2O0O2N1O101N1O2N1O101N1O2N2O0O2N1O2O0O2N1O2N1O2M2N3M2N3M2OVd^1"}, "label": "a slice of pizza bigger than the others"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"last full slice of pizza on plate\"."}], "task": "refering", "answer_template": "The \"last full slice of pizza on plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 85, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 179, 180, 181, 182, 183, 184, 185, 186], "bbox": [0.18526200873362444, 0.2, 0.8352838427947598, 0.51909375], "iscrowd": 0, "area": 37437.75545, "rle": {"size": [640, 458], "counts": "PXe1i0Xc0R1mNS1mNS1nNR1mNT1mN2M01O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001N10001O001O0O101O001O0O101O001O0O101O001O0O101O001O0O2O00001O0O2O00001O001N10001O001N10001O001N10001O0O2O00001N101O0O2O00001N101O000O2O001N10001O0O2O00001N101O0O101O001N101O0O101O0O2O0O101O0O2O0O101O0O2O0O101N101O0O101N101O0O2O0O101O0O2O0O101O0O2O0O101O0O2O0O101O0O2N1O101N1O2N101N1O1O2N101N1O1O2O0O2N1O101N1O2N1O101N1O2N2O0O2N1O2O0O2N1O2N1O2M2N3M2N3M2OVd^1"}, "label": "last full slice of pizza on plate"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of mostly eaten pizza on a white plate\"."}], "task": "refering", "answer_template": "The \"a piece of mostly eaten pizza on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 37, 38, 39, 40, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 87, 88], "bbox": [0.3148471615720524, 0.075953125, 0.6721615720524018, 0.24303125000000003], "iscrowd": 0, "area": 9324.6149, "rle": {"size": [640, 458], "counts": "eQj21oc02N1O2N2N2N3M2N3M3M2N3M3L4M2N3M3M3M2N3M3M2N3M3M3M2N3M3L3N3M00000000000000000000000O10001O00000000000000000000000O100000000000000000000000000000000010O0002N3M3N2M4L8H2NO2N1O1O1O1O2N1O1O1O1O2N1oMk^OV1Va0hNl^OX1Ta0fNn^OY1Sa0fNo^OX1Sa0eNo^OZ1Ra0dNQ_OZ1Pa0dNR_O[1o`0dNR_O[1o`0dNS_OZ1o`0dNR_O[1o`0eNR_OY1o`0gNQ_OX1Pa0gNQ_OX1Pa0hNQ_OV1Pa0jNP_OU1Ra0iNP_OU1Qa0kNo^OT1Ra0lNn^OS1Sa0lNo^OR1Ra0nNn^OR1Sa0lNo^OS1Qa0mNo^OT1Pa0kNQ_OU1o`0kNR_OT1ma010O00010O0010O00010O010O00010O00010O01O01O01O010O00010O0O2N1N2O2N1N3M2N2N3M3M2O2M3M3M2Nfdm2"}, "label": "a piece of mostly eaten pizza on a white plate"}]}
{"id": 464784, "image": "COCO_train2014_000000464784.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bitten piece of pizza sitting on white plate\"."}], "task": "refering", "answer_template": "The \"bitten piece of pizza sitting on white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 37, 38, 39, 40, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 87, 88], "bbox": [0.3148471615720524, 0.075953125, 0.6721615720524018, 0.24303125000000003], "iscrowd": 0, "area": 9324.6149, "rle": {"size": [640, 458], "counts": "eQj21oc02N1O2N2N2N3M2N3M3M2N3M3L4M2N3M3M3M2N3M3M2N3M3M3M2N3M3L3N3M00000000000000000000000O10001O00000000000000000000000O100000000000000000000000000000000010O0002N3M3N2M4L8H2NO2N1O1O1O1O2N1O1O1O1O2N1oMk^OV1Va0hNl^OX1Ta0fNn^OY1Sa0fNo^OX1Sa0eNo^OZ1Ra0dNQ_OZ1Pa0dNR_O[1o`0dNR_O[1o`0dNS_OZ1o`0dNR_O[1o`0eNR_OY1o`0gNQ_OX1Pa0gNQ_OX1Pa0hNQ_OV1Pa0jNP_OU1Ra0iNP_OU1Qa0kNo^OT1Ra0lNn^OS1Sa0lNo^OR1Ra0nNn^OR1Sa0lNo^OS1Qa0mNo^OT1Pa0kNQ_OU1o`0kNR_OT1ma010O00010O0010O00010O010O00010O00010O01O01O01O010O00010O0O2N1N2O2N1N3M2N2N3M3M2O2M3M3M2Nfdm2"}, "label": "bitten piece of pizza sitting on white plate"}]}
{"id": 194448, "image": "COCO_train2014_000000194448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman skiing with her child\"."}], "task": "refering", "answer_template": "The \"a woman skiing with her child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 105, 106, 107, 108, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 268, 269, 288, 292, 311, 315], "bbox": [0.5270312500000001, 0.14029166666666668, 0.73709375, 0.7913958333333333], "iscrowd": 0, "area": 24142.140450000006, "rle": {"size": [480, 640], "counts": "kYn41i>:E:F:H8H9F8J6J5N3M2M3N1SNgMRGZ2f8PNVGR2a8ZN[Gg1V8jNgGW1o7TOnGn0k7XOUHc0j7DUH4m73RHJn79QHEo7?oG_OQ8e0mGZOQ8k0mGTOR8o0mGPOR8T1mGkNQ8Y1mGfNR8^1jGdNV8]1hGdNZ8]1bGdN_8_1[GdNe8^1WGcNj8`1QGbNo8`1mFaNT9e3000O1000000O10000O1000]OhJ_GW5W8VKhGi4U8\\KiGd4U8T1N1010O2N1O1O2N1O1O2O0O1O2M2O1N3M2O1N3M3N2N2N2N3M2L41O4L5L3L4L4L5L3L4M3M3L5L3M3L3N001N101O0d0\\O;FN101N1O1O2O0O2O2M8I8G5L4K5L3L2O2N1N2O2M2O10N2N2N2N2aLoHkNS7h0ZIVOi6=bIBg63]IKl6IYI5P7@TI;[8WNhGe1m:L4K4L5K5Li^^2"}, "label": "a woman skiing with her child"}]}
{"id": 194448, "image": "COCO_train2014_000000194448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady teaching a child to ski\"."}], "task": "refering", "answer_template": "The \"a lady teaching a child to ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 105, 106, 107, 108, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 268, 269, 288, 292, 311, 315], "bbox": [0.5270312500000001, 0.14029166666666668, 0.73709375, 0.7913958333333333], "iscrowd": 0, "area": 24142.140450000006, "rle": {"size": [480, 640], "counts": "kYn41i>:E:F:H8H9F8J6J5N3M2M3N1SNgMRGZ2f8PNVGR2a8ZN[Gg1V8jNgGW1o7TOnGn0k7XOUHc0j7DUH4m73RHJn79QHEo7?oG_OQ8e0mGZOQ8k0mGTOR8o0mGPOR8T1mGkNQ8Y1mGfNR8^1jGdNV8]1hGdNZ8]1bGdN_8_1[GdNe8^1WGcNj8`1QGbNo8`1mFaNT9e3000O1000000O10000O1000]OhJ_GW5W8VKhGi4U8\\KiGd4U8T1N1010O2N1O1O2N1O1O2O0O1O2M2O1N3M2O1N3M3N2N2N2N3M2L41O4L5L3L4L4L5L3L4M3M3L5L3M3L3N001N101O0d0\\O;FN101N1O1O2O0O2O2M8I8G5L4K5L3L2O2N1N2O2M2O10N2N2N2N2aLoHkNS7h0ZIVOi6=bIBg63]IKl6IYI5P7@TI;[8WNhGe1m:L4K4L5K5Li^^2"}, "label": "a lady teaching a child to ski"}]}
{"id": 194448, "image": "COCO_train2014_000000194448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two black skis covered in snow of a person with a blue jacket on next to a child with a black and blue outfit on\"."}], "task": "refering", "answer_template": "The \"two black skis covered in snow of a person with a blue jacket on next to a child with a black and blue outfit on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 311, 312, 315, 334, 335, 338, 339, 358, 359, 361, 362], "bbox": [0.505875, 0.6452500000000001, 0.76790625, 0.9378333333333334], "iscrowd": 0, "area": 3146.92655, "rle": {"size": [480, 640], "counts": "fUh41o>3M2N3M2N2N3M2N3M201O001O00J6BdA1Pg4MkgK2M3N2M3N2M3M4K4L4M3N2N2O1O1O2N1O1O1O1N2O1O1O00O1N2O0O2O1N2N2N2N2N2N101N2N2N2N2N2N101N2N2N2N2N2O1N1O2N2NeWi07UhVO8G:F9H9F9G9E;E2003N2M3M2N3M3N2M3M3M3M3M3M3M3M3M2N3L4M3M3M3M3M3M3M3LT]U2"}, "label": "two black skis covered in snow of a person with a blue jacket on next to a child with a black and blue outfit on"}]}
{"id": 366480, "image": "COCO_train2014_000000366480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a camera standing behind another man with a camera\"."}], "task": "refering", "answer_template": "The \"a man with a camera standing behind another man with a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 44, 45, 67, 68, 90, 91, 113, 114, 136, 137, 159, 160, 182, 183, 205, 206, 228, 229], "bbox": [0.914578125, 0.0022482435597189696, 1.0, 0.6584309133489461], "iscrowd": 0, "area": 11257.851349999992, "rle": {"size": [427, 640], "counts": "fPd75T=5K3]HLj18TN0c12XKAj0`0l30UKMf04T40SK8>J]41RKb05^Oh41PKR1InNV51oJV1GjNX53nJX1FfNU59RKW1EaNU5=SKV1E^NV5g0iJo0O[NV5i0hJP1OXNX5j0fJQ10VNX5l0cJS13RNX5n0`JV15mMY5o0_JX16jMY5Q1\\J[18eMZ5j4dJWKZ5l4dJUKZ5n4dJSK[5o4cJRK[5Q5cJPK]4hNZL[6WOmJ`4iNVL]6XOkJb4iNSLj7m3WHPLk7Q4UHlKm7T4THiKn7W4SHgKn7Y4SHdKo7\\4RHaKP8_4PH_KR8a4oG\\KR8e4:1N1O100O1O1O1O1O2N100O1O1O1O1O1K6VOi0mKn3"}, "label": "a man with a camera standing behind another man with a camera"}]}
{"id": 366480, "image": "COCO_train2014_000000366480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with long hair wearing plaid shorts and a helmet and elbow pads\"."}], "task": "refering", "answer_template": "The \"a child with long hair wearing plaid shorts and a helmet and elbow pads\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 125, 126, 127, 148, 149, 150, 171, 172, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 287, 288, 289, 290, 312, 313, 314, 335, 336, 337], "bbox": [0.40931249999999997, 0.32332552693208433, 0.701671875, 0.9570491803278689], "iscrowd": 0, "area": 19373.51475000001, "rle": {"size": [427, 640], "counts": "Z`]39j<`0_Ob0C=F3L3M2N2N2O1N3M2N2O1N2N3M2O1N2N2O2N1N2]NmMmGT2Q8RNjGo1T8WNfGl1W8ZNdGg1Z8_NaGb1^8f1N3M2N2O101O1N101O0O2O1N1\\HhJU7Y5hHkJV7U5iHnJT7T5iHPKU7Q5iHSKT7m4jHWKS7c5N2M3N1O001O1O001O010O001O1O001O01000O0100O1bKcIm1\\6PNRJe1o5XNUJg1k5VNYJi1g5SN^Jk1c5RNaJm1^5QNfJm1\\5nMiJQ2X5kMlJT2U5hMPKU2R5fMSKY2n4cMVK[2l4_MYKa2i4XM\\Kg2f4SM`Kl2a4nLdKR3\\7N100O2O0O101N1O101N001O100O001O100O00100O1O11O010O10O01O001N2O0O2N2O0O3M3N2M3M3N0O0001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1O1VNnE9S:WO^Fg0c9iNmFV1f:N1N3N1O1O2M2O2N1N101O000O2O00001N10001N10001O0O101O000O2O0O2O1N2O1N2O1N2O1Nge_2"}, "label": "a child with long hair wearing plaid shorts and a helmet and elbow pads"}]}
{"id": 366480, "image": "COCO_train2014_000000366480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy sitting closest to the camera\"."}], "task": "refering", "answer_template": "The \"the boy sitting closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 125, 126, 127, 148, 149, 150, 171, 172, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 287, 288, 289, 290, 312, 313, 314, 335, 336, 337], "bbox": [0.40931249999999997, 0.32332552693208433, 0.701671875, 0.9570491803278689], "iscrowd": 0, "area": 19373.51475000001, "rle": {"size": [427, 640], "counts": "Z`]39j<`0_Ob0C=F3L3M2N2N2O1N3M2N2O1N2N3M2O1N2N2O2N1N2]NmMmGT2Q8RNjGo1T8WNfGl1W8ZNdGg1Z8_NaGb1^8f1N3M2N2O101O1N101O0O2O1N1\\HhJU7Y5hHkJV7U5iHnJT7T5iHPKU7Q5iHSKT7m4jHWKS7c5N2M3N1O001O1O001O010O001O1O001O01000O0100O1bKcIm1\\6PNRJe1o5XNUJg1k5VNYJi1g5SN^Jk1c5RNaJm1^5QNfJm1\\5nMiJQ2X5kMlJT2U5hMPKU2R5fMSKY2n4cMVK[2l4_MYKa2i4XM\\Kg2f4SM`Kl2a4nLdKR3\\7N100O2O0O101N1O101N001O100O001O100O00100O1O11O010O10O01O001N2O0O2N2O0O3M3N2M3M3N0O0001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1O1VNnE9S:WO^Fg0c9iNmFV1f:N1N3N1O1O2M2O2N1N101O000O2O00001N10001N10001O0O101O000O2O0O2O1N2O1N2O1N2O1Nge_2"}, "label": "the boy sitting closest to the camera"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with cheese and soup on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the bowl with cheese and soup on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87], "bbox": [0.56003125, 0.04635416666666667, 0.9520625, 0.2341875], "iscrowd": 0, "area": 11359.20295, "rle": {"size": [480, 640], "counts": "`kW51k>5L4M3M2O2O1N2O1N2O1N101O1O1N3N2N2N2M3N2N2N1N2O2N1O1N2O1O2N1O1O1O1O2N1N2O1O1O1O1O2N1O1O1N2O1O2N1O1OO10001N2O1O0O2O1O1N101O1N2O1O1N2O1O1N2O1O]OWNPDg1Q<[NnCd1R<]NmCb1T<_NlC`1S<bNmC]1S<dNmCZ1T<gNlCX1T<iNlCV1T<lNkCR1V<oNjCP1U<ROkCm0U<TOjCk0W<VOhCj0X<WOfCj0Z<WOeCh0\\<ZObCf0^<[OaCe0_<\\O_Cd0b<]O]Cc0c<^O\\Cb0d<_O[C`0g<@XC`0h<l00O10000O10000O10000O100O10000O100O1O10`CSNm;l1TDVNj;j1UDYNi;f1XD\\Nf;c1ZD`Nd;`1\\D`Ne;^1[DcNe;\\1\\DdNd;\\1[DeNe;Z1\\DfNd;Z1\\DfNd;Y1\\DhNd;X1\\DhNd;W1\\DjNd;V1\\DjNd;V1[DkNe;T1\\DlNd;T1[DmNe;R1\\DnNd;R1\\DmNe;R1[DoNe;Q1[DoNe;P1[DQOe;o0[DQOe;o0ZDROf;m0[DROf;n0YDSOg;l0ZDSOg;m0XDTOh;k0YDTOh;l0WDUOi;j0XDVOh;j0XDVOi;i0VDWOk;h0VDXOj;h0UDYOk;f0VDZOj;f0UDZOl;e0UD[Ok;e0TD\\Ol;c0UD]Ok;c0TD]Om;c0SD]Om;c0RD^On;b0RD^On;b0QD]OQ<c0oCZOT<f0kCWOY<j0fCRO^<n0bCoNb<P1_CkNe<U1c0001O2N1O2N00O1O1O1O2N1O1O1O1O1O1O100O1O1O1O2N1O100O1O1O1O100O1O1O101N1O10000O10000O10000O101O00000001O000001O000O2O0O101N100O101N10010O00010O01O00010O00010O0001O0O100O2O0O10h_>"}, "label": "the bowl with cheese and soup on the right in the right hand picture"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl with a spoon facing west in it\"."}], "task": "refering", "answer_template": "The \"bowl with a spoon facing west in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87], "bbox": [0.56003125, 0.04635416666666667, 0.9520625, 0.2341875], "iscrowd": 0, "area": 11359.20295, "rle": {"size": [480, 640], "counts": "`kW51k>5L4M3M2O2O1N2O1N2O1N101O1O1N3N2N2N2M3N2N2N1N2O2N1O1N2O1O2N1O1O1O1O2N1N2O1O1O1O1O2N1O1O1N2O1O2N1O1OO10001N2O1O0O2O1O1N101O1N2O1O1N2O1O1N2O1O]OWNPDg1Q<[NnCd1R<]NmCb1T<_NlC`1S<bNmC]1S<dNmCZ1T<gNlCX1T<iNlCV1T<lNkCR1V<oNjCP1U<ROkCm0U<TOjCk0W<VOhCj0X<WOfCj0Z<WOeCh0\\<ZObCf0^<[OaCe0_<\\O_Cd0b<]O]Cc0c<^O\\Cb0d<_O[C`0g<@XC`0h<l00O10000O10000O10000O100O10000O100O1O10`CSNm;l1TDVNj;j1UDYNi;f1XD\\Nf;c1ZD`Nd;`1\\D`Ne;^1[DcNe;\\1\\DdNd;\\1[DeNe;Z1\\DfNd;Z1\\DfNd;Y1\\DhNd;X1\\DhNd;W1\\DjNd;V1\\DjNd;V1[DkNe;T1\\DlNd;T1[DmNe;R1\\DnNd;R1\\DmNe;R1[DoNe;Q1[DoNe;P1[DQOe;o0[DQOe;o0ZDROf;m0[DROf;n0YDSOg;l0ZDSOg;m0XDTOh;k0YDTOh;l0WDUOi;j0XDVOh;j0XDVOi;i0VDWOk;h0VDXOj;h0UDYOk;f0VDZOj;f0UDZOl;e0UD[Ok;e0TD\\Ol;c0UD]Ok;c0TD]Om;c0SD]Om;c0RD^On;b0RD^On;b0QD]OQ<c0oCZOT<f0kCWOY<j0fCRO^<n0bCoNb<P1_CkNe<U1c0001O2N1O2N00O1O1O1O2N1O1O1O1O1O1O100O1O1O1O2N1O100O1O1O1O100O1O1O101N1O10000O10000O10000O101O00000001O000001O000O2O0O101N100O101N10010O00010O01O00010O00010O0001O0O100O2O0O10h_>"}, "label": "bowl with a spoon facing west in it"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich on a plate in front of a bowl of soup\"."}], "task": "refering", "answer_template": "The \"a sandwich on a plate in front of a bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.0, 0.34472916666666664, 0.555375, 0.9353333333333335], "iscrowd": 0, "area": 80547.58224999999, "rle": {"size": [480, 640], "counts": "S;Z2S:c2^Ob0]Oc0^Oc0\\Oc0@`0O2O0000001N1000001O000O101O00001N1000001O005J10001O001N10001O001O0O101O001O0O101O001O000O10000000O010000000000O100000O10O10000000000O100000O10O10000000000O100000O10O10000000000O11O000O10001O000000001O000O10001O000000001N100000001O000O10001O000000001N100000001O000O10001O000000001N100000001O000O10001O000000001N100000001O000O10001O00001O000O2O001O001O00001N101O00001O001N10001O001O00001N101O001O00001N101O00001O0O2O00001O001O0O101O001O001N10001O001O0O101O001O00001N101O00001O0O2O001O00001N101O00001O001N10001O001O0O101O001O001N10001O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O001O1N2O1O1O1O001N2O1O1O2N3M2M3N3M200O2O0M3K5L5J5L4L5K4L4L4LmXU4"}, "label": "a sandwich on a plate in front of a bowl of soup"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"soup with shredded cheese on table\"."}], "task": "refering", "answer_template": "The \"soup with shredded cheese on table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 138, 139, 143, 144, 145, 146, 147, 148, 149, 150, 169, 170, 171, 193], "bbox": [0.008421874999999999, 0.1460625, 0.640453125, 0.4831458333333334], "iscrowd": 0, "area": 39205.146400000005, "rle": {"size": [480, 640], "counts": "d`21d>;E;E:N3N2N1O2N2N2N101N2N2N1O2O1N1O2O1N2HTN[Cl1e<WNWCk1i<6000O01000O10O1000O10O10O1000O10O1000O10O1000O0100000O0100000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O2O00000O10000000000O101O000000000O10000000001O0O10001O0000001N1000001O0000001N100000001O000O101O0000001O0O10001O0000001N1000001O00000O101O0000001O000O10001O0000001N1000001O00000O101O0000001O001O1O1O001O1O1O1O001O1O1O001O1O1O1O010O1O1O1O001O1O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O001ON2O1O1O1O1O1O1N2O1O1O1O2N1O1O1N2O1O2N1O1O1O1N2O101N100O100O101N100O100O101N100O100O101N100O100O100O2O000O100O101N100O100O101N100O100O101N100O100O100O2O0O10001N100O2O0O2N100O2N101N100O2N101N1O101N1O2O0O101N1O2O0O101N100O2O0O2O0O101N101N101N2O2M2O2M2O1N3N1N5L4K5L4Ka0@be[3"}, "label": "soup with shredded cheese on table"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl of soup next to sandwich halves\"."}], "task": "refering", "answer_template": "The \"bowl of soup next to sandwich halves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 138, 139, 143, 144, 145, 146, 147, 148, 149, 150, 169, 170, 171, 193], "bbox": [0.008421874999999999, 0.1460625, 0.640453125, 0.4831458333333334], "iscrowd": 0, "area": 39205.146400000005, "rle": {"size": [480, 640], "counts": "d`21d>;E;E:N3N2N1O2N2N2N101N2N2N1O2O1N1O2O1N2HTN[Cl1e<WNWCk1i<6000O01000O10O1000O10O10O1000O10O1000O10O1000O0100000O0100000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O2O00000O10000000000O101O000000000O10000000001O0O10001O0000001N1000001O0000001N100000001O000O101O0000001O0O10001O0000001N1000001O00000O101O0000001O000O10001O0000001N1000001O00000O101O0000001O001O1O1O001O1O1O1O001O1O1O001O1O1O1O010O1O1O1O001O1O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O001ON2O1O1O1O1O1O1N2O1O1O1O2N1O1O1N2O1O2N1O1O1O1N2O101N100O100O101N100O100O101N100O100O101N100O100O100O2O000O100O101N100O100O101N100O100O101N100O100O100O2O0O10001N100O2O0O2N100O2N101N100O2N101N1O101N1O2O0O101N1O2O0O101N100O2O0O2O0O101N101N101N2O2M2O2M2O1N3N1N5L4K5L4Ka0@be[3"}, "label": "bowl of soup next to sandwich halves"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one half of a sandwich on a plate with various foods\"."}], "task": "refering", "answer_template": "The \"one half of a sandwich on a plate with various foods\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 366, 367], "bbox": [0.43553125000000004, 0.3264583333333333, 1.0, 0.8892291666666666], "iscrowd": 0, "area": 72903.30709999999, "rle": {"size": [480, 640], "counts": "XPS41n>2M2O1N3N2M3N2N2O1N2N2O1N2N2O0O2N2O1N2N101N2N2O0O2N2O1N1O2N2O1N101N2N2O0O2O1N2N2O2M2O1N2N2O1N2N2O1N101N2O0O2N2O0O2O1N101N2N102M4M3L4M2M4L4M3L4QF]L[8f3`GhLS8\\3gGSMl7Q3oGWMk7l2oGWMP8m2kGUMU8n2eGVMY8i400O1O100O2O0O100O100O2O00000O10001O000O1000001N1000000O10001O000O1000001N1000000O101O00000O10001O0O1000000O2O00000O10001O000O1000001N100000001O0O100000001O00000000001O0O1000001O00000000001O0001O000001O000001O0001O000001O0000000001O01O000000000000010O0000000000000001O01O00000000000001O0001O000000000001O000001O000000000001O0001O000000000001O000001O0000000001O00000001O00001O01O0001O0000001O0001O01O00001O0001O01O0000010O00000010O0001O01O0001O0001O01O0000010O00000010O000000001O0000001O0000000010O001O0010O01O1N101O001O001O0O2O1O001O001O0O2O001O1O001O0O2O001O001O0\\M\\L"}, "label": "one half of a sandwich on a plate with various foods"}]}
{"id": 46997, "image": "COCO_train2014_000000046997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwich that is on the right of the dish\"."}], "task": "refering", "answer_template": "The \"the half of the sandwich that is on the right of the dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 366, 367], "bbox": [0.43553125000000004, 0.3264583333333333, 1.0, 0.8892291666666666], "iscrowd": 0, "area": 72903.30709999999, "rle": {"size": [480, 640], "counts": "XPS41n>2M2O1N3N2M3N2N2O1N2N2O1N2N2O0O2N2O1N2N101N2N2O0O2N2O1N1O2N2O1N101N2N2O0O2O1N2N2O2M2O1N2N2O1N2N2O1N101N2O0O2N2O0O2O1N101N2N102M4M3L4M2M4L4M3L4QF]L[8f3`GhLS8\\3gGSMl7Q3oGWMk7l2oGWMP8m2kGUMU8n2eGVMY8i400O1O100O2O0O100O100O2O00000O10001O000O1000001N1000000O10001O000O1000001N1000000O101O00000O10001O0O1000000O2O00000O10001O000O1000001N100000001O0O100000001O00000000001O0O1000001O00000000001O0001O000001O000001O0001O000001O0000000001O01O000000000000010O0000000000000001O01O00000000000001O0001O000000000001O000001O000000000001O0001O000000000001O000001O0000000001O00000001O00001O01O0001O0000001O0001O01O00001O0001O01O0000010O00000010O0001O01O0001O0001O01O0000010O00000010O000000001O0000001O0000000010O001O0010O01O1N101O001O001O0O2O1O001O001O0O2O001O1O001O0O2O001O001O0\\M\\L"}, "label": "the half of the sandwich that is on the right of the dish"}]}
{"id": 391063, "image": "COCO_train2014_000000391063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwitch which is behind\"."}], "task": "refering", "answer_template": "The \"sandwitch which is behind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 47, 48, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 75, 76, 77, 78, 79, 92, 93, 94, 95, 110, 111, 112, 113, 129, 130], "bbox": [0.14262, 0.08757333333333334, 0.7432799999999999, 0.56408], "iscrowd": 0, "area": 21543.292200000004, "rle": {"size": [375, 500], "counts": "eUj02Y;?L4L5L1N1O1O2O0O2L3M4L4\\OoN]FU1\\9g0I7K5M3M3N2M3M2O1N1O2N102M2O2N1O2N2O0O2N1O2O0O2N1O2O0O1O1O010O1O0010O01O1O0010O01O10O010O01OO2M2N2N3M2O101N101N10001N100O2O001N101N101O0O2O0O2O00001O001O0010O01O001O001O001O1O2O0O1O1O1O1O1O1O1O101N1O1O1O2N2N2O2M2N10N101N1O2N1N201O010O00010O001O0100O1O010O100O10O01O100O10O0100O1O10O0100O10O01O100O10O01O1O1O100O1O1O1O10O01O11O001O1O001O1O0O2O001O1O001O1O001O10O01O10O01O1O010O1O00100OO1O1O2M2O1O2N1N2O1O2M2O1O1O2O0000001O0O10001O001O001O001O1O001O1O001O001O1O010O1O001O001O1O001O001O1O010O1O1O001O1O1O001O1O1O00100O1O001O1O1O1O010O1O10O01O100O1O100O1O100O1O100O1O100O2N2LgT_1"}, "label": "sandwitch which is behind"}]}
{"id": 391063, "image": "COCO_train2014_000000391063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwhich furthest from the edge of the counter\"."}], "task": "refering", "answer_template": "The \"the sandwhich furthest from the edge of the counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 148, 149, 150, 151, 152, 153, 154, 157, 167, 168, 169, 170], "bbox": [0.25786000000000003, 0.25842666666666664, 0.86124, 0.7685333333333333], "iscrowd": 0, "area": 37247.56940000001, "rle": {"size": [375, 500], "counts": "d^_15];8G8H9I7K5J6K7I6I8I7I5L5J6K4K4M2M4M2M4M3L3N1N2O1N3N1N2O1N2O0O2O1N101N2O010O1O10O0100O00100O10O0100O1O010O10O01O100O010O100O00100O010O1O10O0100O00100O10O0100O1O010O10O01O100O010O100O00100O010O1O10O0100O00100O10O0100O1O010O10O01O010O01O010O01O01O2O1N101N2O0O2N2O1N101N2N101N2O1N101N2N101N2O1N1O2O1N10O0O2N100O2M2O2M21O100O100O100O1O100O10000000000O10001O000000000000000000000O101O00000000000000000000000O2O00000000000001O0000000001O000000000001O00000000000000000000001O0000000000000001O001O010O00001O001O001O001O001M3M2N3M2N3M3M2O2N2M2O2N1O2N2M2O2N2N1N3N1O2N2N1N3N2N2N3M2M4M3M3M3L3N3M3L4M2M4M3L3L5K5J5L__i0"}, "label": "the sandwhich furthest from the edge of the counter"}]}
{"id": 391063, "image": "COCO_train2014_000000391063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the loaf of bread , more bent on the middle , and closest to the middle of the wooden board\"."}], "task": "refering", "answer_template": "The \"the loaf of bread , more bent on the middle , and closest to the middle of the wooden board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 148, 149, 150, 151, 152, 153, 154, 157, 167, 168, 169, 170], "bbox": [0.25786000000000003, 0.25842666666666664, 0.86124, 0.7685333333333333], "iscrowd": 0, "area": 37247.56940000001, "rle": {"size": [375, 500], "counts": "d^_15];8G8H9I7K5J6K7I6I8I7I5L5J6K4K4M2M4M2M4M3L3N1N2O1N3N1N2O1N2O0O2O1N101N2O010O1O10O0100O00100O10O0100O1O010O10O01O100O010O100O00100O010O1O10O0100O00100O10O0100O1O010O10O01O100O010O100O00100O010O1O10O0100O00100O10O0100O1O010O10O01O010O01O010O01O01O2O1N101N2O0O2N2O1N101N2N101N2O1N101N2N101N2O1N1O2O1N10O0O2N100O2M2O2M21O100O100O100O1O100O10000000000O10001O000000000000000000000O101O00000000000000000000000O2O00000000000001O0000000001O000000000001O00000000000000000000001O0000000000000001O001O010O00001O001O001O001O001M3M2N3M2N3M3M2O2N2M2O2N1O2N2M2O2N2N1N3N1O2N2N1N3N2N2N3M2M4M3M3M3L3N3M3L4M2M4M3L3L5K5J5L__i0"}, "label": "the loaf of bread , more bent on the middle , and closest to the middle of the wooden board"}]}
{"id": 210847, "image": "COCO_train2014_000000210847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy riding a very old motorcycle\"."}], "task": "refering", "answer_template": "The \"a guy riding a very old motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 99, 100, 121, 122, 123, 144, 145, 146, 147, 167, 168, 190, 191, 214, 237], "bbox": [0.25451562499999997, 0.26031400966183577, 0.407015625, 0.7214009661835749], "iscrowd": 0, "area": 6555.849349999999, "rle": {"size": [414, 640], "counts": "lRR23h<4M4K4M4K5K4L5K4L5L4L3M4L4L3N2M4L3M3M3N2M4L3M3M3N3L3O1N2N2O2M2`FRMn8f3H8Ic0\\OYMPHc0n7^OSHb0l7_OTHa0j7@XH?g7BYH=f7D\\H:d7G\\H8c7I_H5`7M`H3_7McH1\\70eHOX7F`HXN8Q2V7GeHWN7[1JhNW7g1kHVN4T11POk6g1SITN3l07XOa6h1WITN2d0=AV6h1]IRN3<b0Il5i1bIQN15h01d5h1eIQN2Mm08]5h1gIQNi16`4g1iISNg15a4f1kISNf16_4h1jIQNh16_4l0nHHl0UOh16^4m0PIG`3[O`3n0RIE^3\\Oa3o0SID\\3\\Oa3P1UIBZ3]Ob3P1WIBW3]Ob3Q1YIA[:>hE@X:?kE@T:`0nE^OR:a0QF^On9a0UF]Ok9c0VF]Oi9b0YF]Og9b0ZF_Og9>[FAf9>[FBf9;\\FEe98]FG];10O100O010O10O010O000001O0000001O000000c^i4"}, "label": "a guy riding a very old motorcycle"}]}
{"id": 210847, "image": "COCO_train2014_000000210847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 99, 100, 121, 122, 123, 144, 145, 146, 147, 167, 168, 190, 191, 214, 237], "bbox": [0.25451562499999997, 0.26031400966183577, 0.407015625, 0.7214009661835749], "iscrowd": 0, "area": 6555.849349999999, "rle": {"size": [414, 640], "counts": "lRR23h<4M4K4M4K5K4L5K4L5L4L3M4L4L3N2M4L3M3M3N2M4L3M3M3N3L3O1N2N2O2M2`FRMn8f3H8Ic0\\OYMPHc0n7^OSHb0l7_OTHa0j7@XH?g7BYH=f7D\\H:d7G\\H8c7I_H5`7M`H3_7McH1\\70eHOX7F`HXN8Q2V7GeHWN7[1JhNW7g1kHVN4T11POk6g1SITN3l07XOa6h1WITN2d0=AV6h1]IRN3<b0Il5i1bIQN15h01d5h1eIQN2Mm08]5h1gIQNi16`4g1iISNg15a4f1kISNf16_4h1jIQNh16_4l0nHHl0UOh16^4m0PIG`3[O`3n0RIE^3\\Oa3o0SID\\3\\Oa3P1UIBZ3]Ob3P1WIBW3]Ob3Q1YIA[:>hE@X:?kE@T:`0nE^OR:a0QF^On9a0UF]Ok9c0VF]Oi9b0YF]Og9b0ZF_Og9>[FAf9>[FBf9;\\FEe98]FG];10O100O010O10O010O000001O0000001O000000c^i4"}, "label": "a man riding a motorcycle"}]}
{"id": 481185, "image": "COCO_train2014_000000481185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork between the bread plate and dinner plate\"."}], "task": "refering", "answer_template": "The \"the fork between the bread plate and dinner plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 164, 165, 186, 187, 208, 209, 231, 232], "bbox": [0.058390624999999995, 0.32449648711943796, 0.29035937500000003, 0.7223419203747072], "iscrowd": 0, "area": 2994.2087500000002, "rle": {"size": [427, 640], "counts": "[T`03W=2M4M2M3N2N3L2O1N2O1N2O0O2O1O1N2O1N101N10O010O001O10O01O1O1O10O01O1O100O1O1O1O010O1O100O100O100O10O0100O100O100O100O100O010O100O10000O10000O100O10000O10000O10000O10O10000001O00000000001O000000000001O10O01O1O1O001N2O1O1O1O1O1O@]C:b<F_C9`<GaC:]<GbC;\\<EdC=Y<DgC>W<CiC=U<DkC>T<AlC0O6T<KlC015R<JnC016Q<InC118P<GnC227P<FnC338n;DPD419o;BPD529n;APD72NL1R<IPD82NM0R<IPD91NM0V<2mCNM0V<2nCMM0V<2mCNM0W<2lCNL0Z<0jC0M0Z<NjC1L1\\<LiC2L1]<KgC5K0h<OXC1h<OYC0h<OXC1h<OXC2h<NXC1h<OXC1i<NWC2Q=NcUm5"}, "label": "the fork between the bread plate and dinner plate"}]}
{"id": 481185, "image": "COCO_train2014_000000481185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a somewhat vertical fork near a horizontal fork\"."}], "task": "refering", "answer_template": "The \"a somewhat vertical fork near a horizontal fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 164, 165, 186, 187, 208, 209, 231, 232], "bbox": [0.058390624999999995, 0.32449648711943796, 0.29035937500000003, 0.7223419203747072], "iscrowd": 0, "area": 2994.2087500000002, "rle": {"size": [427, 640], "counts": "[T`03W=2M4M2M3N2N3L2O1N2O1N2O0O2O1O1N2O1N101N10O010O001O10O01O1O1O10O01O1O100O1O1O1O010O1O100O100O100O10O0100O100O100O100O100O010O100O10000O10000O100O10000O10000O10000O10O10000001O00000000001O000000000001O10O01O1O1O001N2O1O1O1O1O1O@]C:b<F_C9`<GaC:]<GbC;\\<EdC=Y<DgC>W<CiC=U<DkC>T<AlC0O6T<KlC015R<JnC016Q<InC118P<GnC227P<FnC338n;DPD419o;BPD529n;APD72NL1R<IPD82NM0R<IPD91NM0V<2mCNM0V<2nCMM0V<2mCNM0W<2lCNL0Z<0jC0M0Z<NjC1L1\\<LiC2L1]<KgC5K0h<OXC1h<OYC0h<OXC1h<OXC2h<NXC1h<OXC1i<NWC2Q=NcUm5"}, "label": "a somewhat vertical fork near a horizontal fork"}]}
{"id": 522146, "image": "COCO_train2014_000000522146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cruise ship\"."}], "task": "refering", "answer_template": "The \"cruise ship\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.30915625, 0.36346604215456674, 0.7289218749999999, 0.601990632318501], "iscrowd": 0, "area": 16893.717350000003, "rle": {"size": [427, 640], "counts": "Xib2:1Hl<=O2M2Ab0L4L1O000O100000000000000000000O10000000000000000000000O100000000000000000000O10000O1O1N2O1N2O1N2N2O1N2O1N2O1YO]NdEe1[:f0000000O1000000O100000000O100HUMPFk2o9WMPFi2P:XMnEi2R:WMmEj2T:VMjEk2V:50000001O0PMkEj2U:VMlEi2T:WMnEg2R:YMoEf2R:ZMnEe2^:N1O2N2N2N1OO1O1O12N1O2N001O00L4000000001O00000000000000000000000000002N0000O100O10000O12N2N001O0000000000000000000000000000O1O100O1O1OgMUET2l:jMUEV2k:iMVEW2j:hMWEY2m:01O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O00001O0000000000000000001O0000001O001O00001OI7L40000O10000000000000000000000O100000000000000000000O100000000000000000000O10000000WNkD\\1U;bNmD\\1U;bNmD\\1U;cNlD[1V;cNlD[1V;dNkDZ1d;N2O1N2N2N2N2N2N2O1N2N3M2N3M2N3M5L4KlmW2"}, "label": "cruise ship"}]}
{"id": 522146, "image": "COCO_train2014_000000522146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat waiting at the harbour\"."}], "task": "refering", "answer_template": "The \"boat waiting at the harbour\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.30915625, 0.36346604215456674, 0.7289218749999999, 0.601990632318501], "iscrowd": 0, "area": 16893.717350000003, "rle": {"size": [427, 640], "counts": "Xib2:1Hl<=O2M2Ab0L4L1O000O100000000000000000000O10000000000000000000000O100000000000000000000O10000O1O1N2O1N2O1N2N2O1N2O1N2O1YO]NdEe1[:f0000000O1000000O100000000O100HUMPFk2o9WMPFi2P:XMnEi2R:WMmEj2T:VMjEk2V:50000001O0PMkEj2U:VMlEi2T:WMnEg2R:YMoEf2R:ZMnEe2^:N1O2N2N2N1OO1O1O12N1O2N001O00L4000000001O00000000000000000000000000002N0000O100O10000O12N2N001O0000000000000000000000000000O1O100O1O1OgMUET2l:jMUEV2k:iMVEW2j:hMWEY2m:01O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O00001O0000000000000000001O0000001O001O00001OI7L40000O10000000000000000000000O100000000000000000000O100000000000000000000O10000000WNkD\\1U;bNmD\\1U;bNmD\\1U;cNlD[1V;cNlD[1V;dNkDZ1d;N2O1N2N2N2N2N2N2O1N2N3M2N3M2N3M5L4KlmW2"}, "label": "boat waiting at the harbour"}]}
{"id": 522146, "image": "COCO_train2014_000000522146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange and black rubber boat\"."}], "task": "refering", "answer_template": "The \"orange and black rubber boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 207, 208, 209, 210, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.003296875, 0.5897189695550351, 0.883, 0.9872365339578455], "iscrowd": 0, "area": 48312.44484999999, "rle": {"size": [427, 640], "counts": "YV1c1Q8g300001O00001O0000001O00001O00001O0000001O00001O000O101O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0O101O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0O10001O00001O00001O00001O0000001O00001O0000hSP1"}, "label": "orange and black rubber boat"}]}
{"id": 522146, "image": "COCO_train2014_000000522146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange and black railing\"."}], "task": "refering", "answer_template": "The \"the orange and black railing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 207, 208, 209, 210, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.003296875, 0.5897189695550351, 0.883, 0.9872365339578455], "iscrowd": 0, "area": 48312.44484999999, "rle": {"size": [427, 640], "counts": "YV1c1Q8g300001O00001O0000001O00001O00001O0000001O00001O000O101O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0O101O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0O10001O00001O00001O00001O0000001O00001O0000hSP1"}, "label": "the orange and black railing"}]}
{"id": 407460, "image": "COCO_train2014_000000407460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing wii sports boxing next to a woman\"."}], "task": "refering", "answer_template": "The \"a man playing wii sports boxing next to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 74, 75, 98, 121, 142, 143, 144, 165, 166, 167, 187, 188, 189, 211, 233, 234, 235, 256, 257, 258, 279, 280, 281, 282, 303, 304, 305, 326, 327, 328, 350, 351, 352, 373, 374, 375], "bbox": [0.15337499999999998, 0.1546875, 0.33539062499999994, 0.9883958333333334], "iscrowd": 0, "area": 15016.18195, "rle": {"size": [480, 640], "counts": "]X^11l>4K5L4L4K6K4K5L4L4YNROXEQ1W:_OiEc0e9MZF6Y94hFMV94iFOT92lF0Q91oF0P90oF3n8NRG4k8MTG5k8KUG7h8JWG9f8G[G;c8E\\G=b8D^G>_8C`G`0^8@bGa0\\8@cGc0Z8^OfGd0X8\\OgGg0V8ZOjGg0T8YOlGj0Q8WOoGk0o7UOPHm0o7RORHP1l7POSHS1k7mNUHU1i7jNWH\\1d7dN\\Hb1^7^NcHh1V7WNlHn1n6RNTIS2Y3PLlNm1lMY2P3lKSOj1PN_2e2hK[Oi1RNc2]2dKAi1SNi2T2`KHf1WNo2i1\\K0e1XNU3`1WK8d1ZNT3]1YK8b1^NT3X1[K;`1^NU3U1\\K=_1`NT3Q1_K>]1bNT3n0`K`0[1eNT3j0aKa0[1gNS3f0dKb0Y1iNS3d0dKc0X1lNR3a0fKc0X1mNQ3`0hKb0T1SOS3:iKc0R1WOS36kKc0P1ZOT34lKa0n0_OT30nK?n0ER3LPL=m0KR3HRL:l02P3DTL7m08n2AUL5m0>l2]OXL2k0f0l2WOYL1k0k0k2TOZLNl0R1h2PO\\LLl0Y1e2kNaLHj0b1b2fNTMSO=\\2\\2aNhMbMk0Q4Z1]Na0f1\\OZNc0j1[OUNe0n1XOSNQO`MnNa4n1PNQOfMlN]4P2oMQOiMjN[4S2mMPOnMiNX4T2kMQORNgNV4;\\K3]2i0VNdNU4GQLi0e1i0[NcNi5a0mKj0_N`Nh5d0jKi0dN^Ne5f0hK8TMUOf1Gb5i0eK8[MQOe1H^5l0cK7Q7IPI4R7LPIOR72oHKS75nHGU79lHAX7?jHWO_7i0\\3O010O010O10O0100O010O10O0100O010O1O7J6I7JiWW6"}, "label": "a man playing wii sports boxing next to a woman"}]}
{"id": 407460, "image": "COCO_train2014_000000407460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing wii game with a woman\"."}], "task": "refering", "answer_template": "The \"a man playing wii game with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 74, 75, 98, 121, 142, 143, 144, 165, 166, 167, 187, 188, 189, 211, 233, 234, 235, 256, 257, 258, 279, 280, 281, 282, 303, 304, 305, 326, 327, 328, 350, 351, 352, 373, 374, 375], "bbox": [0.15337499999999998, 0.1546875, 0.33539062499999994, 0.9883958333333334], "iscrowd": 0, "area": 15016.18195, "rle": {"size": [480, 640], "counts": "]X^11l>4K5L4L4K6K4K5L4L4YNROXEQ1W:_OiEc0e9MZF6Y94hFMV94iFOT92lF0Q91oF0P90oF3n8NRG4k8MTG5k8KUG7h8JWG9f8G[G;c8E\\G=b8D^G>_8C`G`0^8@bGa0\\8@cGc0Z8^OfGd0X8\\OgGg0V8ZOjGg0T8YOlGj0Q8WOoGk0o7UOPHm0o7RORHP1l7POSHS1k7mNUHU1i7jNWH\\1d7dN\\Hb1^7^NcHh1V7WNlHn1n6RNTIS2Y3PLlNm1lMY2P3lKSOj1PN_2e2hK[Oi1RNc2]2dKAi1SNi2T2`KHf1WNo2i1\\K0e1XNU3`1WK8d1ZNT3]1YK8b1^NT3X1[K;`1^NU3U1\\K=_1`NT3Q1_K>]1bNT3n0`K`0[1eNT3j0aKa0[1gNS3f0dKb0Y1iNS3d0dKc0X1lNR3a0fKc0X1mNQ3`0hKb0T1SOS3:iKc0R1WOS36kKc0P1ZOT34lKa0n0_OT30nK?n0ER3LPL=m0KR3HRL:l02P3DTL7m08n2AUL5m0>l2]OXL2k0f0l2WOYL1k0k0k2TOZLNl0R1h2PO\\LLl0Y1e2kNaLHj0b1b2fNTMSO=\\2\\2aNhMbMk0Q4Z1]Na0f1\\OZNc0j1[OUNe0n1XOSNQO`MnNa4n1PNQOfMlN]4P2oMQOiMjN[4S2mMPOnMiNX4T2kMQORNgNV4;\\K3]2i0VNdNU4GQLi0e1i0[NcNi5a0mKj0_N`Nh5d0jKi0dN^Ne5f0hK8TMUOf1Gb5i0eK8[MQOe1H^5l0cK7Q7IPI4R7LPIOR72oHKS75nHGU79lHAX7?jHWO_7i0\\3O010O010O10O0100O010O10O0100O010O1O7J6I7JiWW6"}, "label": "a man playing wii game with a woman"}]}
{"id": 407460, "image": "COCO_train2014_000000407460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with glasses playing nintendo wi boxing game\"."}], "task": "refering", "answer_template": "The \"a woman with glasses playing nintendo wi boxing game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 71, 72, 73, 74, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 184, 185, 186, 187, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.14529166666666665, 0.327765625, 1.0], "iscrowd": 0, "area": 45386.42980000001, "rle": {"size": [480, 640], "counts": "[6]5b9Y3gL1E^GVJd8h5^GWJc8g5`GWJa8g5=N2N2N2N2N2N2O1000000O100000000000000O100O100O100O100O100O100O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3M3M3M3M3M3M3M3N2N2O1O1N2O1O1O1O1O1O1OkLXMTJg2h5gMoIY2m5TNjIk1R6\\NlIc1P6bNPJ]1m5hNRJX1i5nNVJQ1f5TOZJk0c5ZO\\Je0`5A_J?\\5FdJ9Z5KgH^Mf1f2a50gH\\Mg1d2`52hH\\Mg1a2_57hHZMh1_2^5c0aJ\\O]5h0bJXO\\5k0cJUOZ5o0eJPOZ5S1eJmNX5W1gJiNW5Z1hJeNV2`MTOo3eNaNW2hMkNi3mN_NW2QNdNa3TO]NX2[N^LJP1`39TNeNhMd3g1^M^28PNc2CUM^27nMe2DSM`27kMf2ETMa25iMh2FRMc25gMh2GSMc24fMi2GRMe24iMd2CXMd24nM_2]O]Mg23mM^2]O_Mg22lM^2^O_Mh22kM^2]O`Mi21RNV2UOiMk20QNV2TOjMl2OQNU2TOkMm2OPNU2SOlMn2NnMV2UOkMo2NlMW2TOlMQ3LkMW2UOlMR3LiMX2UOlMS3KgMZ2VOkMT3JfM\\2TOjMX3IdM]2TOjMY3HcM^2TOiM[3H`M`2UOhM\\3G_Mb2TOfM_3G]Mc2SOgMa3E\\Md2SOfMc3EYMf2TOeMd3DXMh2SOcMg3DVMi2SOcMh3CUMj2ROcMk3BSMk2ROcMl3AQMm2SOaMn3AoLo2RO`MP4@oLo2POaMS4_OmLP3POaMT4^OmLP3oNaMV4^OlLQ3mNaMX4]OlLQ3lNbMY4\\OlLQ3jNcM\\4[OkLQ3iNdM]4ZOjLS3hNbM`4ZOiLS3gNcMa4YOiLf2ROQNW4XOhLf2QORNX4WOhLe2QOSN\\7k1eHUN\\7i1eHVN]7f1eH[N]7a1eH^N]7`1dH`N^7]1cHbN_7\\1aHeNa7X1`HgNb7W1_HiNc7T1^HkNd7R1^HnNd7`0kHAV7;mHDb7JbH6T;000O0100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O4M3M3M3L5L3MlYY6"}, "label": "a woman with glasses playing nintendo wi boxing game"}]}
{"id": 407460, "image": "COCO_train2014_000000407460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"people using wii controller\"."}], "task": "refering", "answer_template": "The \"people using wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 71, 72, 73, 74, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 184, 185, 186, 187, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.14529166666666665, 0.327765625, 1.0], "iscrowd": 0, "area": 45386.42980000001, "rle": {"size": [480, 640], "counts": "[6]5b9Y3gL1E^GVJd8h5^GWJc8g5`GWJa8g5=N2N2N2N2N2N2O1000000O100000000000000O100O100O100O100O100O100O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3M3M3M3M3M3M3M3N2N2O1O1N2O1O1O1O1O1O1OkLXMTJg2h5gMoIY2m5TNjIk1R6\\NlIc1P6bNPJ]1m5hNRJX1i5nNVJQ1f5TOZJk0c5ZO\\Je0`5A_J?\\5FdJ9Z5KgH^Mf1f2a50gH\\Mg1d2`52hH\\Mg1a2_57hHZMh1_2^5c0aJ\\O]5h0bJXO\\5k0cJUOZ5o0eJPOZ5S1eJmNX5W1gJiNW5Z1hJeNV2`MTOo3eNaNW2hMkNi3mN_NW2QNdNa3TO]NX2[N^LJP1`39TNeNhMd3g1^M^28PNc2CUM^27nMe2DSM`27kMf2ETMa25iMh2FRMc25gMh2GSMc24fMi2GRMe24iMd2CXMd24nM_2]O]Mg23mM^2]O_Mg22lM^2^O_Mh22kM^2]O`Mi21RNV2UOiMk20QNV2TOjMl2OQNU2TOkMm2OPNU2SOlMn2NnMV2UOkMo2NlMW2TOlMQ3LkMW2UOlMR3LiMX2UOlMS3KgMZ2VOkMT3JfM\\2TOjMX3IdM]2TOjMY3HcM^2TOiM[3H`M`2UOhM\\3G_Mb2TOfM_3G]Mc2SOgMa3E\\Md2SOfMc3EYMf2TOeMd3DXMh2SOcMg3DVMi2SOcMh3CUMj2ROcMk3BSMk2ROcMl3AQMm2SOaMn3AoLo2RO`MP4@oLo2POaMS4_OmLP3POaMT4^OmLP3oNaMV4^OlLQ3mNaMX4]OlLQ3lNbMY4\\OlLQ3jNcM\\4[OkLQ3iNdM]4ZOjLS3hNbM`4ZOiLS3gNcMa4YOiLf2ROQNW4XOhLf2QORNX4WOhLe2QOSN\\7k1eHUN\\7i1eHVN]7f1eH[N]7a1eH^N]7`1dH`N^7]1cHbN_7\\1aHeNa7X1`HgNb7W1_HiNc7T1^HkNd7R1^HnNd7`0kHAV7;mHDb7JbH6T;000O0100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O4M3M3M3L5L3MlYY6"}, "label": "people using wii controller"}]}
{"id": 79783, "image": "COCO_train2014_000000079783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the women in pink is playing video games\"."}], "task": "refering", "answer_template": "The \"the women in pink is playing video games\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 168, 187, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281], "bbox": [0.072125, 0.0, 0.32596875000000003, 0.7350416666666667], "iscrowd": 0, "area": 28417.591450000004, "rle": {"size": [480, 640], "counts": "Yfe01o>001N101O0O2O001N2O001N2O1N101O0O2O1O2M2N2O2N1O1O2N3M1N2O0O101N1N2O2M3M2O2J5O2N2N3N1N2N3M2GlMcCW2Z<;L4HYMQDj2f;?iDgLc91dF^3XOmLS:FdFm3Z9ZL`Fh3]9[LaFh3l8QLlF:6g3k8WLgF4<h3j8jLRGZ3l8hLoF]3n8\\1N2N3M2O1N2N3M4L4L3N2M3M3M4L4L5K5K5RK^HV2f7dKXHc15h2Q8UMQHj2V8mLlGS3X8hLjGW3Z8dLgG\\3]8_LeG_3`8[LbGe3d8TL^Gj3h8PLZGo3l8jKVGT4l9N2O1O100O100O00100O1000001O0000000000O10O10000000000000000000O1K5J6J6K5L4L4K5D<M3N2O1O1O1nKYJ^KMS2:gMb5h4XJWK2Q2a0gMZ5i4nI[Kh0g0ZO5T7i3lHWKh0j0a7n3eHQL[7f3nHZLR7d3PI\\LP7b3RI^Ln6a3TI_Lk6`3VIoK`N7Z8i3XIPL_N6X8R2lGPO_1gN_N5V8c1gG`N:k0\\1mN]N5U8a1PH\\N7k0Y1RO\\N5T8b1RHZN9c0_OWOh14\\N4R8a1VH\\N:<Z1DTN3Q8[1kG^N=2>7Y1JQN3P8\\1^H`N<0Y10nM2P8^1]H`Nm20fL0Q8a1[H_Nn20gLOP8b1[H_No2OfL0P8b1ZH_NQ30eLMQ8T1hGWOa0HR3OdLNQ8S1nGQO=NQ30cLNS8P1iHROa20cLNT8o0hHROc2ObLOU8h0iLYOd3?cL@`3=e5I9JbPZ6"}, "label": "the women in pink is playing video games"}]}
{"id": 79783, "image": "COCO_train2014_000000079783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a magenta dress with a green sash\"."}], "task": "refering", "answer_template": "The \"a woman wearing a magenta dress with a green sash\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 168, 187, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281], "bbox": [0.072125, 0.0, 0.32596875000000003, 0.7350416666666667], "iscrowd": 0, "area": 28417.591450000004, "rle": {"size": [480, 640], "counts": "Yfe01o>001N101O0O2O001N2O001N2O1N101O0O2O1O2M2N2O2N1O1O2N3M1N2O0O101N1N2O2M3M2O2J5O2N2N3N1N2N3M2GlMcCW2Z<;L4HYMQDj2f;?iDgLc91dF^3XOmLS:FdFm3Z9ZL`Fh3]9[LaFh3l8QLlF:6g3k8WLgF4<h3j8jLRGZ3l8hLoF]3n8\\1N2N3M2O1N2N3M4L4L3N2M3M3M4L4L5K5K5RK^HV2f7dKXHc15h2Q8UMQHj2V8mLlGS3X8hLjGW3Z8dLgG\\3]8_LeG_3`8[LbGe3d8TL^Gj3h8PLZGo3l8jKVGT4l9N2O1O100O100O00100O1000001O0000000000O10O10000000000000000000O1K5J6J6K5L4L4K5D<M3N2O1O1O1nKYJ^KMS2:gMb5h4XJWK2Q2a0gMZ5i4nI[Kh0g0ZO5T7i3lHWKh0j0a7n3eHQL[7f3nHZLR7d3PI\\LP7b3RI^Ln6a3TI_Lk6`3VIoK`N7Z8i3XIPL_N6X8R2lGPO_1gN_N5V8c1gG`N:k0\\1mN]N5U8a1PH\\N7k0Y1RO\\N5T8b1RHZN9c0_OWOh14\\N4R8a1VH\\N:<Z1DTN3Q8[1kG^N=2>7Y1JQN3P8\\1^H`N<0Y10nM2P8^1]H`Nm20fL0Q8a1[H_Nn20gLOP8b1[H_No2OfL0P8b1ZH_NQ30eLMQ8T1hGWOa0HR3OdLNQ8S1nGQO=NQ30cLNS8P1iHROa20cLNT8o0hHROc2ObLOU8h0iLYOd3?cL@`3=e5I9JbPZ6"}, "label": "a woman wearing a magenta dress with a green sash"}]}
{"id": 79783, "image": "COCO_train2014_000000079783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a flowered shirt with a yellow scarf\"."}], "task": "refering", "answer_template": "The \"a woman in a flowered shirt with a yellow scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372], "bbox": [0.0, 0.0015833333333333333, 0.261234375, 0.9858333333333333], "iscrowd": 0, "area": 53305.83379999999, "rle": {"size": [480, 640], "counts": "^;[3o4f6ZKf4O1000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000]EI]67cIJ\\66dIJ\\66cIL\\64dIM[63eINZ62fIOY61fI1Y6OgI2X6NhI3W6MiI4V6LiI6V6JjI7U6IkI8T6HlI9S6GlI;S6EmI<R6DnI=Q6CoI>P6BoI`0P6@PJa0o5_OQJb0n5^OQJd0n5\\ORJe0m5[OSJf0l5ZOTJg0k5YOTJi0k5WOUJi0k5WOUJj0j5VOVJk0i5UOVJm0i5SOWJn0h5ROXJo0g5QOXJQ1g5oNWJS1i5mNVJU1i5kNUJX1j5hNUJZ1j5gNSJ\\1l5dNRJ^1n5bNPJa1o5_NPJc1o5]NPJe1o5[NPJg1o5YNQJg1o5YNPJn1j5RNVJT2d5lMUJa2e5_MVJk2e5UMZJR3`5nL_JY3[5gL[Jh3`5XL`Jj3^5VLbJl3\\5TLcI3TOj3W7SL]I=ZOb3W7QLWIh0_OZ3W7nKXIm0ZO[3\\7hKXIQ1UO^3a7aKYIo5e6QJYIS6e6mIlHd6R7g0N2E;M4M2M3M3M3M3N2M3M3M4L3M3iLkF?W9[OmFd0V9VOnFi0W9oNlFQ1\\9eNgFZ1a9[NcFd1f9oM^FR2i9RMlFm2d:O001O1O1O001O10O01O1O001nNjCYOW<f0YDjNh;T1hD]NY;b1S1O1O1N2O1O1O1N2O1O1O1000000UOYB=g=C\\B:d=F_B7a=IaB5b=HaB5d=F_B5Y>KY[m6"}, "label": "a woman in a flowered shirt with a yellow scarf"}]}
{"id": 79783, "image": "COCO_train2014_000000079783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman dressed in a floral shirt and yellow scarf\"."}], "task": "refering", "answer_template": "The \"a woman dressed in a floral shirt and yellow scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372], "bbox": [0.0, 0.0015833333333333333, 0.261234375, 0.9858333333333333], "iscrowd": 0, "area": 53305.83379999999, "rle": {"size": [480, 640], "counts": "^;[3o4f6ZKf4O1000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000]EI]67cIJ\\66dIJ\\66cIL\\64dIM[63eINZ62fIOY61fI1Y6OgI2X6NhI3W6MiI4V6LiI6V6JjI7U6IkI8T6HlI9S6GlI;S6EmI<R6DnI=Q6CoI>P6BoI`0P6@PJa0o5_OQJb0n5^OQJd0n5\\ORJe0m5[OSJf0l5ZOTJg0k5YOTJi0k5WOUJi0k5WOUJj0j5VOVJk0i5UOVJm0i5SOWJn0h5ROXJo0g5QOXJQ1g5oNWJS1i5mNVJU1i5kNUJX1j5hNUJZ1j5gNSJ\\1l5dNRJ^1n5bNPJa1o5_NPJc1o5]NPJe1o5[NPJg1o5YNQJg1o5YNPJn1j5RNVJT2d5lMUJa2e5_MVJk2e5UMZJR3`5nL_JY3[5gL[Jh3`5XL`Jj3^5VLbJl3\\5TLcI3TOj3W7SL]I=ZOb3W7QLWIh0_OZ3W7nKXIm0ZO[3\\7hKXIQ1UO^3a7aKYIo5e6QJYIS6e6mIlHd6R7g0N2E;M4M2M3M3M3M3N2M3M3M4L3M3iLkF?W9[OmFd0V9VOnFi0W9oNlFQ1\\9eNgFZ1a9[NcFd1f9oM^FR2i9RMlFm2d:O001O1O1O001O10O01O1O001nNjCYOW<f0YDjNh;T1hD]NY;b1S1O1O1N2O1O1O1N2O1O1O1000000UOYB=g=C\\B:d=F_B7a=IaB5b=HaB5d=F_B5Y>KY[m6"}, "label": "a woman dressed in a floral shirt and yellow scarf"}]}
{"id": 538537, "image": "COCO_train2014_000000538537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady pouring wine in a glass\"."}], "task": "refering", "answer_template": "The \"a lady pouring wine in a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 72, 73, 74, 75, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258], "bbox": [0.0015, 0.15955399061032863, 0.29915625, 0.7999999999999999], "iscrowd": 0, "area": 27059.757800000007, "rle": {"size": [426, 640], "counts": "ig06g<=D<D<D<C=D<D<D<L3N3M300001O^MTMfJl2Z5iMQJW2n5jMSJU2m5kMSJU2m5kMSJU2m5lMSJS2m5mMSJS2m5mMSJS2m5mMTJR2l5nMTJR2l5oMSJQ2m5oMSJQ2m5oMZJj1f5VNcJa1]5`N^Kd0b4\\O]LEc3;\\MfNd2Z1f41O0000000000O100O100O2N100O100O100O2O0O100O1000000001O0O10001O00000O101O00000N2M4L3M3WFiN\\7Z1dHQOo6Q1QIZOa6j0^IAT6b0lIJe59ZJ3Y5OgJ<k4HTK`0a4C_K=^4FbK:[4IeK6Z4LfK4X4OgK1W41iKOV42iKOV42jKMU45kKKT47kKIS49mKGR4:mKGR4:nKEQ4>mKCR4>mKCR4>nKBP4`0oKAP4a0nK_OR4d0lK\\OR4h0kKYOT4j0iKWOW4l0fKTOY4o0dKQO\\4R1cKmN\\4W1aKiN^4Z1`KfN`4\\1^KdNa4_1^K_Nb4b1^K^Na4c1_K]N`4e1`KZN`4f1`KZN_4g1aKXN_4j1`KVN`4j1aKUN_4l1`KTN`4l1`KTN`4l1`KSN`4o1`KPN`4P2`KPN`4P2`KPN`4Q2_KoMa4Q2`KmMa4S2_KmMa4T2^KlMb4T2^KlMb4U2^KjMb4V2^KiMc4W2]KiMc4X2\\KhMc4Y2^KfMb4Z2^KfMb4[2]KdMd4\\2\\KdMd4]2\\KbMd4^2\\KbMd4^2\\KbMd4_2\\K_Mf4`2ZK`Mf4`2ZK`Mf4a2YK_Mh4`2YK_Mg4a2YK_Mg4b2XK_Mh4`2XK`Mh4a2XK_Mg4a2YK`Mg4_2YKbMf4_2YKaMg4_2ZKaMf4^2ZKcMe4^2ZKbMf4^2ZKcMf4]2ZKcMe4]2[KkK^On0X5V3ZKlKAl0U5X3YKkKEl0R5X3ZKkKGj0o4[3ZKjKKi0k4]3YKjKN>S5g3oJjK2N^5X4`JiK4_Ok5g4oIkKk6U4SIlKo6T4mHoKT7P4jHQLX7n3fHTL[7l3bHUL_7k3_HWLb7g42fL\\Hl0e7lNdHQ1^7gNkHV1W7bNRI\\1R7YNWId1S7lMVIR2S7^MVI_2m8\\M_E[2j:M3N3L3N2M4L3B>A`0_OZcj5"}, "label": "a lady pouring wine in a glass"}]}
{"id": 538537, "image": "COCO_train2014_000000538537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman pouring wine from a bottle\"."}], "task": "refering", "answer_template": "The \"woman pouring wine from a bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 72, 73, 74, 75, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258], "bbox": [0.0015, 0.15955399061032863, 0.29915625, 0.7999999999999999], "iscrowd": 0, "area": 27059.757800000007, "rle": {"size": [426, 640], "counts": "ig06g<=D<D<D<C=D<D<D<L3N3M300001O^MTMfJl2Z5iMQJW2n5jMSJU2m5kMSJU2m5kMSJU2m5lMSJS2m5mMSJS2m5mMSJS2m5mMTJR2l5nMTJR2l5oMSJQ2m5oMSJQ2m5oMZJj1f5VNcJa1]5`N^Kd0b4\\O]LEc3;\\MfNd2Z1f41O0000000000O100O100O2N100O100O100O2O0O100O1000000001O0O10001O00000O101O00000N2M4L3M3WFiN\\7Z1dHQOo6Q1QIZOa6j0^IAT6b0lIJe59ZJ3Y5OgJ<k4HTK`0a4C_K=^4FbK:[4IeK6Z4LfK4X4OgK1W41iKOV42iKOV42jKMU45kKKT47kKIS49mKGR4:mKGR4:nKEQ4>mKCR4>mKCR4>nKBP4`0oKAP4a0nK_OR4d0lK\\OR4h0kKYOT4j0iKWOW4l0fKTOY4o0dKQO\\4R1cKmN\\4W1aKiN^4Z1`KfN`4\\1^KdNa4_1^K_Nb4b1^K^Na4c1_K]N`4e1`KZN`4f1`KZN_4g1aKXN_4j1`KVN`4j1aKUN_4l1`KTN`4l1`KTN`4l1`KSN`4o1`KPN`4P2`KPN`4P2`KPN`4Q2_KoMa4Q2`KmMa4S2_KmMa4T2^KlMb4T2^KlMb4U2^KjMb4V2^KiMc4W2]KiMc4X2\\KhMc4Y2^KfMb4Z2^KfMb4[2]KdMd4\\2\\KdMd4]2\\KbMd4^2\\KbMd4^2\\KbMd4_2\\K_Mf4`2ZK`Mf4`2ZK`Mf4a2YK_Mh4`2YK_Mg4a2YK_Mg4b2XK_Mh4`2XK`Mh4a2XK_Mg4a2YK`Mg4_2YKbMf4_2YKaMg4_2ZKaMf4^2ZKcMe4^2ZKbMf4^2ZKcMf4]2ZKcMe4]2[KkK^On0X5V3ZKlKAl0U5X3YKkKEl0R5X3ZKkKGj0o4[3ZKjKKi0k4]3YKjKN>S5g3oJjK2N^5X4`JiK4_Ok5g4oIkKk6U4SIlKo6T4mHoKT7P4jHQLX7n3fHTL[7l3bHUL_7k3_HWLb7g42fL\\Hl0e7lNdHQ1^7gNkHV1W7bNRI\\1R7YNWId1S7lMVIR2S7^MVI_2m8\\M_E[2j:M3N3L3N2M4L3B>A`0_OZcj5"}, "label": "woman pouring wine from a bottle"}]}
{"id": 538537, "image": "COCO_train2014_000000538537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman whos face is hidden behind the flowers\"."}], "task": "refering", "answer_template": "The \"woman whos face is hidden behind the flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 83, 84, 85, 106, 107, 108, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 219, 220], "bbox": [0.53175, 0.17856807511737088, 0.7425312500000001, 0.6708215962441314], "iscrowd": 0, "area": 12527.958, "rle": {"size": [426, 640], "counts": "be]4>i<6K4K6K4K6K4K4M3L4M4K4M3L4M4K4M3M3L5L3L4M3L4N2M3M3N1NROZF\\Ne9c1^F\\N`9e1bFZN]9f1eFYN[9e1hFZNX9e1jFZNU9e1nFZNR9e1PGZNo8f1SGZNl8d1XGZNg8f1[GYNe8e1^GZNa8g1hGoMY8W2eGeMZ8c2S16I8IO1M2O2NoN[MaGc2_8dM[G[2f8gMYGW2g8kMYGS2h8oMVGo1l8SNSGj1n8YNQGd1Q9]NnFb1R9aNmF\\1U9fNjFW1W9lNhFP1Z9TOcFi0`9YO_Fa0e9CYF7k9LSFOR:P1c0I700100000000000001N2O2N1O2O1N1O2O0O2O4K5K5L4K6J3N2M2O2N1O2N010O001O001O1O010O0010O0100O01O100O1O1O1O1O1O1O001O1O1K5L4ZOf0K6I8I5TOmD\\OX;>mD_OV;<oD@V;;mDDU;7PEFT;4REIP;2UELn:OVENnkT2"}, "label": "woman whos face is hidden behind the flowers"}]}
{"id": 538537, "image": "COCO_train2014_000000538537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman hidden behind flowers\"."}], "task": "refering", "answer_template": "The \"woman hidden behind flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 83, 84, 85, 106, 107, 108, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 219, 220], "bbox": [0.53175, 0.17856807511737088, 0.7425312500000001, 0.6708215962441314], "iscrowd": 0, "area": 12527.958, "rle": {"size": [426, 640], "counts": "be]4>i<6K4K6K4K6K4K4M3L4M4K4M3L4M4K4M3M3L5L3L4M3L4N2M3M3N1NROZF\\Ne9c1^F\\N`9e1bFZN]9f1eFYN[9e1hFZNX9e1jFZNU9e1nFZNR9e1PGZNo8f1SGZNl8d1XGZNg8f1[GYNe8e1^GZNa8g1hGoMY8W2eGeMZ8c2S16I8IO1M2O2NoN[MaGc2_8dM[G[2f8gMYGW2g8kMYGS2h8oMVGo1l8SNSGj1n8YNQGd1Q9]NnFb1R9aNmF\\1U9fNjFW1W9lNhFP1Z9TOcFi0`9YO_Fa0e9CYF7k9LSFOR:P1c0I700100000000000001N2O2N1O2O1N1O2O0O2O4K5K5L4K6J3N2M2O2N1O2N010O001O001O1O010O0010O0100O01O100O1O1O1O1O1O1O001O1O1K5L4ZOf0K6I8I5TOmD\\OX;>mD_OV;<oD@V;;mDDU;7PEFT;4REIP;2UELn:OVENnkT2"}, "label": "woman hidden behind flowers"}]}
{"id": 538537, "image": "COCO_train2014_000000538537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black shirt and gold hoop earring\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black shirt and gold hoop earring\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 219, 224, 225, 226, 227, 228, 229, 241, 242, 243, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.501453125, 0.008849765258215963, 0.9969374999999999, 0.9842018779342723], "iscrowd": 0, "area": 80209.68554999997, "rle": {"size": [426, 640], "counts": "_jU48Q=?B?@`0A5J3N1N2O2M2O1N3N1N2O1N3N1N2O2M2O1N3M2O1N2O1O10000O10000O1000000O10000O10000O10000bFjLf8W3WGkLj8U3SGmLn8S3PGnLQ9R3lFPMT9Q3iFQMX9o2eFSM\\9m2aFUM_9[31O1O001O1O1O001O1O001O1O001O1O1O001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O3M3M[FWNk7g1SH\\No7a1PH`NR8^1lGeNU8Y1iGiNU8Y1iGiNV8X1iGjNU8W1XG\\Of8f0ZG[Od8f0\\GZOc8g0]GZO`8h0`GXO_8i0aGWO]8k0cGVO[8k0eGUOZ8l0fGUOW8m0iGSOV8n0jGTOR8n0nGTOo7m0QHUOl7l0THUOi7m0WHUOf7l0ZHVOc7k0]HWO_7k0aHWO\\7j0dHXOX7j0hHXOU7i0kHYOR7h0nHYOo6i0QIYOl6h0TIZOh6h0XIZOe6g0[I[Ob6f0^I\\O^6f0bI\\O[6e0eI]OX6d0hI^OT6d0lI]OR6d0nI^On5d0RJ^Ok5c0UJ_Oh5b0XJ_Oe5c0[J_Ob5b0^J_O_5c0aJ_O\\5b0dJ_OZ5b0fJ_OW5c0iJ_OT5b0lJ_OR5b0nJ@n4b0RK_Ol4b0TK_Oi4c0WK_Of4b0ZK_Od4a0]KA_4a0aK@]4a0cK@Z4b0fK@W4a0iK@U4a0kKAQ4a0oK_OP4b0PL^On3d0RL\\Om3e0SL[Ok3g0ULYOj3h0VLXOh3j0XLWOf3j0ZLVOd3l0\\LTOc3m0]LSOb3n0^LRO`3P1`LPO_3Q1aLoN]3S1cLmN\\3T1dLlNZ3V1fLjNY3W1gLjNV3X1jLhNU3Y1kLgNS3[1mLeNR3\\1nLdNP3^1PMbNo2_1QMaNn2`1RM`Nl2b1TM^Nk2c1UM]Ni2e1WM\\Ng2e1YM[Ne2g1[MYNd2h1\\MXNb2j1^MVNa2k1_MUN_2m1aMSN^2n1bMRN\\2P2dMPN[2Q2eMoMZ2R2fMoMW2S2iMmMV2T2jMlMT2V2lMjMS2W2mMiMQ2Y2oMgMP2Z2PNfMn1\\2RNdMm1]2SNcMk1_2UNbMi1_2WNaMg1a2YN_Mf1b2ZN^Me1c2[N]Mc1e2]N[Ma1g2_NYM`1h2`NXM^1j2bNVM\\1l2dNTM[1m2eNSMY1o2gNQMW1Q3iNoLV1R3jNnLT1T3lNlLR1V3nNjLQ1W3oNiLo0Y3QOgLm0[3SOeLl0\\3TOdLj0^3VObLi0_3WOaLg0a3YO_Le0c3[O]Ld0d3\\O\\Lb0f3^OZL`0h3@XL?i3AWL=k3CUL;m3ESL:n3FRL8P4HPL6R4JnK5S4KmK3U4MkK1W4OiK0X40hKNZ42fKL\\44dKK]45cKI_47aKGa49_KFb4:W400000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000_j0"}, "label": "a woman wearing a black shirt and gold hoop earring"}]}
{"id": 538537, "image": "COCO_train2014_000000538537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with her fingers pinched\"."}], "task": "refering", "answer_template": "The \"a lady with her fingers pinched\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 219, 224, 225, 226, 227, 228, 229, 241, 242, 243, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.501453125, 0.008849765258215963, 0.9969374999999999, 0.9842018779342723], "iscrowd": 0, "area": 80209.68554999997, "rle": {"size": [426, 640], "counts": "_jU48Q=?B?@`0A5J3N1N2O2M2O1N3N1N2O1N3N1N2O2M2O1N3M2O1N2O1O10000O10000O1000000O10000O10000O10000bFjLf8W3WGkLj8U3SGmLn8S3PGnLQ9R3lFPMT9Q3iFQMX9o2eFSM\\9m2aFUM_9[31O1O001O1O1O001O1O001O1O001O1O1O001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O3M3M[FWNk7g1SH\\No7a1PH`NR8^1lGeNU8Y1iGiNU8Y1iGiNV8X1iGjNU8W1XG\\Of8f0ZG[Od8f0\\GZOc8g0]GZO`8h0`GXO_8i0aGWO]8k0cGVO[8k0eGUOZ8l0fGUOW8m0iGSOV8n0jGTOR8n0nGTOo7m0QHUOl7l0THUOi7m0WHUOf7l0ZHVOc7k0]HWO_7k0aHWO\\7j0dHXOX7j0hHXOU7i0kHYOR7h0nHYOo6i0QIYOl6h0TIZOh6h0XIZOe6g0[I[Ob6f0^I\\O^6f0bI\\O[6e0eI]OX6d0hI^OT6d0lI]OR6d0nI^On5d0RJ^Ok5c0UJ_Oh5b0XJ_Oe5c0[J_Ob5b0^J_O_5c0aJ_O\\5b0dJ_OZ5b0fJ_OW5c0iJ_OT5b0lJ_OR5b0nJ@n4b0RK_Ol4b0TK_Oi4c0WK_Of4b0ZK_Od4a0]KA_4a0aK@]4a0cK@Z4b0fK@W4a0iK@U4a0kKAQ4a0oK_OP4b0PL^On3d0RL\\Om3e0SL[Ok3g0ULYOj3h0VLXOh3j0XLWOf3j0ZLVOd3l0\\LTOc3m0]LSOb3n0^LRO`3P1`LPO_3Q1aLoN]3S1cLmN\\3T1dLlNZ3V1fLjNY3W1gLjNV3X1jLhNU3Y1kLgNS3[1mLeNR3\\1nLdNP3^1PMbNo2_1QMaNn2`1RM`Nl2b1TM^Nk2c1UM]Ni2e1WM\\Ng2e1YM[Ne2g1[MYNd2h1\\MXNb2j1^MVNa2k1_MUN_2m1aMSN^2n1bMRN\\2P2dMPN[2Q2eMoMZ2R2fMoMW2S2iMmMV2T2jMlMT2V2lMjMS2W2mMiMQ2Y2oMgMP2Z2PNfMn1\\2RNdMm1]2SNcMk1_2UNbMi1_2WNaMg1a2YN_Mf1b2ZN^Me1c2[N]Mc1e2]N[Ma1g2_NYM`1h2`NXM^1j2bNVM\\1l2dNTM[1m2eNSMY1o2gNQMW1Q3iNoLV1R3jNnLT1T3lNlLR1V3nNjLQ1W3oNiLo0Y3QOgLm0[3SOeLl0\\3TOdLj0^3VObLi0_3WOaLg0a3YO_Le0c3[O]Ld0d3\\O\\Lb0f3^OZL`0h3@XL?i3AWL=k3CUL;m3ESL:n3FRL8P4HPL6R4JnK5S4KmK3U4MkK1W4OiK0X40hKNZ42fKL\\44dKK]45cKI_47aKGa49_KFb4:W400000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000_j0"}, "label": "a lady with her fingers pinched"}]}
{"id": 473003, "image": "COCO_train2014_000000473003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch with no pillows on it\"."}], "task": "refering", "answer_template": "The \"a brown couch with no pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.002859375, 0.41075117370892017, 0.293453125, 0.9892488262910797], "iscrowd": 0, "area": 35759.942749999995, "rle": {"size": [426, 640], "counts": "YP1Q2T;d5bJO1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000O100000<Dd0\\O000000O10000000O100000O100000000000000O10O2O1O2N2N1O2N1O2N2N1O2N1O2N2M2O2N1O2N2N1O2N1O2N2N1O2N1N3N2N1O2N1O2N2N1O2N1O2N2N1N3N1O2Nd0\\OW1iNV1jNfdk5"}, "label": "a brown couch with no pillows on it"}]}
{"id": 473003, "image": "COCO_train2014_000000473003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large brown sofa\"."}], "task": "refering", "answer_template": "The \"the large brown sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.002859375, 0.41075117370892017, 0.293453125, 0.9892488262910797], "iscrowd": 0, "area": 35759.942749999995, "rle": {"size": [426, 640], "counts": "YP1Q2T;d5bJO1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000O100000<Dd0\\O000000O10000000O100000O100000000000000O10O2O1O2N2N1O2N1O2N2N1O2N1O2N2M2O2N1O2N2N1O2N1O2N2N1O2N1N3N2N1O2N1O2N2N1O2N1O2N2N1N3N1O2Nd0\\OW1iNV1jNfdk5"}, "label": "the large brown sofa"}]}
{"id": 473003, "image": "COCO_train2014_000000473003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color sofa in the living room\"."}], "task": "refering", "answer_template": "The \"white color sofa in the living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 224, 225, 226, 227], "bbox": [0.624171875, 0.4094131455399061, 0.9304218750000001, 0.650281690140845], "iscrowd": 0, "area": 14147.69255, "rle": {"size": [426, 640], "counts": "]VV5;m;S1M2N201O00001O00001O00001O00001O0M[NaDe1`;3O00001O01O01O00001O01O01O000010O00O2O00001O001O00001O4L0O101O00001O00001O000O101O00001O00001O0001O01O00001O00001O0001O01O00001O00001O01O0001O00001O000010O000001O00001O00010O0000001O00001O019F0000001O000000I8L3O100000000000001O0O100000000000001O0O100000000000001O0O100000000000001O000O1000000000001O000O100000000JSEoMm:m1;O00000O10000000001N1000000000000O1]Oc0[Ojab0"}, "label": "white color sofa in the living room"}]}
{"id": 473003, "image": "COCO_train2014_000000473003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white couch in a room\"."}], "task": "refering", "answer_template": "The \"white couch in a room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 224, 225, 226, 227], "bbox": [0.624171875, 0.4094131455399061, 0.9304218750000001, 0.650281690140845], "iscrowd": 0, "area": 14147.69255, "rle": {"size": [426, 640], "counts": "]VV5;m;S1M2N201O00001O00001O00001O00001O0M[NaDe1`;3O00001O01O01O00001O01O01O000010O00O2O00001O001O00001O4L0O101O00001O00001O000O101O00001O00001O0001O01O00001O00001O0001O01O00001O00001O01O0001O00001O000010O000001O00001O00010O0000001O00001O019F0000001O000000I8L3O100000000000001O0O100000000000001O0O100000000000001O0O100000000000001O000O1000000000001O000O100000000JSEoMm:m1;O00000O10000000001N1000000000000O1]Oc0[Ojab0"}, "label": "white couch in a room"}]}
{"id": 325548, "image": "COCO_train2014_000000325548.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bag of lemons\"."}], "task": "refering", "answer_template": "The \"a green bag of lemons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 260, 261], "bbox": [0.093390625, 0.3524766355140187, 0.43529687499999997, 0.7488785046728972], "iscrowd": 0, "area": 26925.680650000002, "rle": {"size": [428, 640], "counts": "PSj02Y=0\\D`0]9AbFb0[9^OeFd0Y9]OfFe0X9[OhFh0U9YOkFh0S9XOmFj0Q9WOnFk0P9UOPGm0n8TOQGn0m8ROSGP1k8QOUGQ1h8oNXGS1g8mNYGT1f8mNkEKW1Z1m8lNkE5l0Q1X9E]F<b9ORF3m9g1O1O1O2N001O001O001O00001O001O001O001O00001L3C>O010O010O00010O010O010O010O0100O010O0010O001O001O01O0001O0K5J60000O100O2O000O100O101O00001O001O00001O001O00001O0001O01O0000010O000000010O000001O010O1O1O1O101N1O1O1O2O1N2N1O01K4M3O101N10000O2O0O11O010O001O0010O01O001O0000001N1000000O10000O3N2N2N2M3N001O1O1N101O1O001O1O0O2O1O2N2N1O2N2N2M3N4L1O1O1O1O1O1N2O2N1O1O1N2N2M3M3N2M3M4M3L4L4M3L5K4M3L41OA?_Oa0@Sef4"}, "label": "a green bag of lemons"}]}
{"id": 325548, "image": "COCO_train2014_000000325548.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag of lemons\"."}], "task": "refering", "answer_template": "The \"the bag of lemons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 260, 261], "bbox": [0.093390625, 0.3524766355140187, 0.43529687499999997, 0.7488785046728972], "iscrowd": 0, "area": 26925.680650000002, "rle": {"size": [428, 640], "counts": "PSj02Y=0\\D`0]9AbFb0[9^OeFd0Y9]OfFe0X9[OhFh0U9YOkFh0S9XOmFj0Q9WOnFk0P9UOPGm0n8TOQGn0m8ROSGP1k8QOUGQ1h8oNXGS1g8mNYGT1f8mNkEKW1Z1m8lNkE5l0Q1X9E]F<b9ORF3m9g1O1O1O2N001O001O001O00001O001O001O001O00001L3C>O010O010O00010O010O010O010O0100O010O0010O001O001O01O0001O0K5J60000O100O2O000O100O101O00001O001O00001O001O00001O0001O01O0000010O000000010O000001O010O1O1O1O101N1O1O1O2O1N2N1O01K4M3O101N10000O2O0O11O010O001O0010O01O001O0000001N1000000O10000O3N2N2N2M3N001O1O1N101O1O001O1O0O2O1O2N2N1O2N2N2M3N4L1O1O1O1O1O1N2O2N1O1O1N2N2M3M3N2M3M4M3L4L4M3L5K4M3L41OA?_Oa0@Sef4"}, "label": "the bag of lemons"}]}
{"id": 96177, "image": "COCO_train2014_000000096177.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person that is holding the sprayer\"."}], "task": "refering", "answer_template": "The \"the person that is holding the sprayer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 64, 65, 66, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 104, 116, 121, 122, 139, 140, 157, 158, 175, 176, 193, 194, 211, 212, 229], "bbox": [0.42906, 0.16327999999999998, 0.8429, 0.9448266666666667], "iscrowd": 0, "area": 12994.904999999999, "rle": {"size": [375, 500], "counts": "Tl^22d;3M3M2N3N1N3M2N3M2N3M2N3MQO[Eo0e:3M2O2M2N1O2O0O1O1O100O1O2N10GmEnNR:S1oElNP:T1SFjNl9W1VFgNj9X1XFhNf9Y1\\FeNc9[1<10O1O100O1O100O1O1O1M3N2N2M31O001O001O001O1O001O001O001O001O1O0O10000000O1000000O10O1O1O001O1O1O00100O1O001O1O1O001O1O1O001O1O1O1O001O1O1O2N1N2M3N2N3L3N2N2N2M3N1O2N2M23N1N3N2M2O2M2N3N2M2N3M3M2N3M3M2N3M3M2N3N2N101N21N5LN2O0O2NRImNY3S1]LCX3;`L:U3DhLl0l2SOSMZ1c2dNmKSO1f2g3UNjKC?b2_3iMSLD=e2a3eMRLF=f2c3aMPLJ<f2e3^MPLK:i2g3ZMoKN9i2i3WMnK09j2j3TMmK37k2m3PMmK46m2n3mLlK75m2P4jLkK94o2R4fLkK;2P3T4cLjK=2Q3T4aLjKb0Mo2X4_LkKh0Gj2]4^LmKm0@f2b4]LnKS1YOb2h4[LoKY1SO[2n4]LPL]1kNS2Y5`LlKm4X4SKhKh4]4XKcKd4`4]KaK_4b4aK^K\\4e4dK[KY4h4gKYKU4i4lKWKQ4l4oKTKn3P5QLPKl3S5TLnJh3U5XLkJe3X5[LhJa3\\5_LeJ]3^5cLbJZ3a5fL_JW3d5iL\\JT3g5lLZJP3i5PMWJm2l5SMTJi2P6WMQJa2V6_MjI[2]6dMcIU2d6kM\\In1k6RNVIa1V7_NjHV1a7jN_Hj0m7VOTH>W8BiG9]8FcG8`8G`G7R:M2N2Ohgl0"}, "label": "the person that is holding the sprayer"}]}
{"id": 96177, "image": "COCO_train2014_000000096177.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person wearing black shirt\"."}], "task": "refering", "answer_template": "The \"person wearing black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 64, 65, 66, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 104, 116, 121, 122, 139, 140, 157, 158, 175, 176, 193, 194, 211, 212, 229], "bbox": [0.42906, 0.16327999999999998, 0.8429, 0.9448266666666667], "iscrowd": 0, "area": 12994.904999999999, "rle": {"size": [375, 500], "counts": "Tl^22d;3M3M2N3N1N3M2N3M2N3M2N3MQO[Eo0e:3M2O2M2N1O2O0O1O1O100O1O2N10GmEnNR:S1oElNP:T1SFjNl9W1VFgNj9X1XFhNf9Y1\\FeNc9[1<10O1O100O1O100O1O1O1M3N2N2M31O001O001O001O1O001O001O001O001O1O0O10000000O1000000O10O1O1O001O1O1O00100O1O001O1O1O001O1O1O001O1O1O1O001O1O1O2N1N2M3N2N3L3N2N2N2M3N1O2N2M23N1N3N2M2O2M2N3N2M2N3M3M2N3M3M2N3M3M2N3N2N101N21N5LN2O0O2NRImNY3S1]LCX3;`L:U3DhLl0l2SOSMZ1c2dNmKSO1f2g3UNjKC?b2_3iMSLD=e2a3eMRLF=f2c3aMPLJ<f2e3^MPLK:i2g3ZMoKN9i2i3WMnK09j2j3TMmK37k2m3PMmK46m2n3mLlK75m2P4jLkK94o2R4fLkK;2P3T4cLjK=2Q3T4aLjKb0Mo2X4_LkKh0Gj2]4^LmKm0@f2b4]LnKS1YOb2h4[LoKY1SO[2n4]LPL]1kNS2Y5`LlKm4X4SKhKh4]4XKcKd4`4]KaK_4b4aK^K\\4e4dK[KY4h4gKYKU4i4lKWKQ4l4oKTKn3P5QLPKl3S5TLnJh3U5XLkJe3X5[LhJa3\\5_LeJ]3^5cLbJZ3a5fL_JW3d5iL\\JT3g5lLZJP3i5PMWJm2l5SMTJi2P6WMQJa2V6_MjI[2]6dMcIU2d6kM\\In1k6RNVIa1V7_NjHV1a7jN_Hj0m7VOTH>W8BiG9]8FcG8`8G`G7R:M2N2Ohgl0"}, "label": "person wearing black shirt"}]}
{"id": 96177, "image": "COCO_train2014_000000096177.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a white shirt with a red bicycle logo , cleaning a large hunk of meat in a sink\"."}], "task": "refering", "answer_template": "The \"a person in a white shirt with a red bicycle logo , cleaning a large hunk of meat in a sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 159, 160, 161, 176, 177, 178, 179, 192, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.56124, 0.008986666666666667, 0.99438, 0.9842666666666667], "iscrowd": 0, "area": 47460.447550000004, "rle": {"size": [375, 500], "counts": "eRW35];5K6J6J5O10000O100O10000O10000O100O10000O100O10000O100O2O0000OgJPOMP1OVONj0N^OOa0NE0;B3>LA6?J_O8a0G_O:a0E^O=b0C\\O?d0@[Ob0e0^OZOc0f0\\O[Od0e0[O\\Oe0d0[O[Of0VLbNk3g00g0TLPO_39=g0RL]OV3Ki0h0PLAT3Fl0j0oKDR3Bo0j0oKGo2^OS1k0nKIm2\\OT1l0oKKj2XOX1m0nKNf2UO]1m0mK1c2RO_1n0nK2a2oNb1o0mK5^2lNe1o0mK8[2iNh1o0mK;X2fNj1P1nK=U2cNm1P1nK?S2aNo1P1nKb0P2^NQ2Q1oKd0m1[NT2Q1oKg0j1XNW2Q1oKj0g1UN[2P1nKm0e1SN_2n0lKo0e1TN_2l0lKQ1d1SNb2j0jKT1b1SNe2h0iKV1a1RNh2f0gKX1a1RNi2e0fKZ1`1QNl2c0dK]1_1PNo2a0bK_1_1PNP3`0aKa1^1oMQ3`0aKb1]1oMR3>aKd1\\1nMT3=`Ke1[1PNT3;aKf1Z1oMV3:`Kh1Y1nMX39_Kj1X1nMX37aKm1U1lM[36`KP2S1jM^35_KS2Q1iM_34`KU2o0gMb33_KX2m0fMd31_K[2j0eMg30_K]2h0cMj3O^Kd2b0^Mo3N_Kl2:VMX4M^KT33oL`4L]KU33PM_4K^KU33PM`4J]KW32PMa4H]KX32PMa4H]KX32PMb4G\\KY31RMc4D\\K[30QMd4D\\K[30QMe4C[K\\30RMe4A[K]30RMe4A[K]30RMf4@ZK_3ORMh4]OYKa3ORMm4XOTKf3OSMl4WOUKf3NTMj4YOXKc3NTMg4\\O[Ka3MRMf4@]K^3MRMc4C`K[3MRM`4FcKX3MQM^4JeKV3LPM\\4MhKS3LPMY40kKP3LPMV43nKm2KPMU46PLj2KPMR49SLh2JoLP4<VLe2JnLn3`0XLb2JnLk3c0[L_2JnLh3f0^L\\2JmLf3j0`LZ2IlLd3m0cLW2IlL`3Q1gLS2HlL^3U1jLo1HlLZ3Y1nLl1GkLW3]1RMn2k2VMUMj2g2ZMYMf2b2_M^Ma2^2cMbM\\2Z2iMfMV2W2nMiMQ2S2TNmMk1o1ZNQNe1l1_NTN`1h1eNXNZ1e1jN[NU1a1PO_Ni1c0\\N]Oo6000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000n3RL\\4dKhj0"}, "label": "a person in a white shirt with a red bicycle logo , cleaning a large hunk of meat in a sink"}]}
{"id": 96177, "image": "COCO_train2014_000000096177.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the white shirt with the red bike rider logo\"."}], "task": "refering", "answer_template": "The \"the person in the white shirt with the red bike rider logo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 159, 160, 161, 176, 177, 178, 179, 192, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.56124, 0.008986666666666667, 0.99438, 0.9842666666666667], "iscrowd": 0, "area": 47460.447550000004, "rle": {"size": [375, 500], "counts": "eRW35];5K6J6J5O10000O100O10000O10000O100O10000O100O10000O100O2O0000OgJPOMP1OVONj0N^OOa0NE0;B3>LA6?J_O8a0G_O:a0E^O=b0C\\O?d0@[Ob0e0^OZOc0f0\\O[Od0e0[O\\Oe0d0[O[Of0VLbNk3g00g0TLPO_39=g0RL]OV3Ki0h0PLAT3Fl0j0oKDR3Bo0j0oKGo2^OS1k0nKIm2\\OT1l0oKKj2XOX1m0nKNf2UO]1m0mK1c2RO_1n0nK2a2oNb1o0mK5^2lNe1o0mK8[2iNh1o0mK;X2fNj1P1nK=U2cNm1P1nK?S2aNo1P1nKb0P2^NQ2Q1oKd0m1[NT2Q1oKg0j1XNW2Q1oKj0g1UN[2P1nKm0e1SN_2n0lKo0e1TN_2l0lKQ1d1SNb2j0jKT1b1SNe2h0iKV1a1RNh2f0gKX1a1RNi2e0fKZ1`1QNl2c0dK]1_1PNo2a0bK_1_1PNP3`0aKa1^1oMQ3`0aKb1]1oMR3>aKd1\\1nMT3=`Ke1[1PNT3;aKf1Z1oMV3:`Kh1Y1nMX39_Kj1X1nMX37aKm1U1lM[36`KP2S1jM^35_KS2Q1iM_34`KU2o0gMb33_KX2m0fMd31_K[2j0eMg30_K]2h0cMj3O^Kd2b0^Mo3N_Kl2:VMX4M^KT33oL`4L]KU33PM_4K^KU33PM`4J]KW32PMa4H]KX32PMa4H]KX32PMb4G\\KY31RMc4D\\K[30QMd4D\\K[30QMe4C[K\\30RMe4A[K]30RMe4A[K]30RMf4@ZK_3ORMh4]OYKa3ORMm4XOTKf3OSMl4WOUKf3NTMj4YOXKc3NTMg4\\O[Ka3MRMf4@]K^3MRMc4C`K[3MRM`4FcKX3MQM^4JeKV3LPM\\4MhKS3LPMY40kKP3LPMV43nKm2KPMU46PLj2KPMR49SLh2JoLP4<VLe2JnLn3`0XLb2JnLk3c0[L_2JnLh3f0^L\\2JmLf3j0`LZ2IlLd3m0cLW2IlL`3Q1gLS2HlL^3U1jLo1HlLZ3Y1nLl1GkLW3]1RMn2k2VMUMj2g2ZMYMf2b2_M^Ma2^2cMbM\\2Z2iMfMV2W2nMiMQ2S2TNmMk1o1ZNQNe1l1_NTN`1h1eNXNZ1e1jN[NU1a1PO_Ni1c0\\N]Oo6000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000n3RL\\4dKhj0"}, "label": "the person in the white shirt with the red bike rider logo"}]}
{"id": 112577, "image": "COCO_train2014_000000112577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two bananas hanging from a toilet paper holder\"."}], "task": "refering", "answer_template": "The \"two bananas hanging from a toilet paper holder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 55, 56, 57, 79, 80, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 192, 193, 194, 195, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 305, 306, 307, 308, 328, 329, 330, 331, 352, 353], "bbox": [0.298234375, 0.07058333333333335, 0.55059375, 0.9247083333333334], "iscrowd": 0, "area": 26804.099499999997, "rle": {"size": [480, 640], "counts": "]mi2c0V>d0\\O;E;E:H6K4K5K5L4L4M4M2M3N2M3M3N3L3N2M3M3N2M4M2M2N2O1N2O1NJ63N2M2N3N1N3N2M2N3N2M2O2M2N3N2M2O2M3MFiFPKV9Q5jFPKS9R5nFmJQ9T5oFlJo8W5QGhJm8Z5SGfJl8[5UGdJi8^5WGbJh8_5YG_Jf8c5ZG]Jd8e5]GZJb8g5^GYJ`8i5aGVJ^8k5cGSJ\\8P6cGPJ[8R6fGmIY8T6gGlIW8V681N2N2O0O1O10OQLcI^O^6=kI_OT6=UJ^Ol5d0VJWOk5k0XJPOg5T1ZJgNg5[1[JaNe5a1]JZNc5i1`JRN`5Q2aJkM^5X2cJeM]5]2eJ_M[5c2gJYMX5k2iJQMW5Q3kJkLU5W3mJeLR5_3nJ_LQ5c3QKYLo4i3SKSLl4P4VKlKj4X4VKdKi4a4WK[Ki4h4YKQKi4S5XKeJj4`5XKWJk4m5VKlIl4W6k15L4L3L5L4K4M4L4K4M4L4K5L3L5L4L3L5L4L3L5L4L3L5L3L5L4L3L5L3M4K4M3M3L3N3L4M3M3L3N3M3L4M3L3N3M3L4M3M4K4M4K5L3L5L4K4L5L4K4ZOg0BUlV4"}, "label": "two bananas hanging from a toilet paper holder"}]}
{"id": 112577, "image": "COCO_train2014_000000112577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"banana with less black dots\"."}], "task": "refering", "answer_template": "The \"banana with less black dots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 55, 56, 57, 79, 80, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 192, 193, 194, 195, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 305, 306, 307, 308, 328, 329, 330, 331, 352, 353], "bbox": [0.298234375, 0.07058333333333335, 0.55059375, 0.9247083333333334], "iscrowd": 0, "area": 26804.099499999997, "rle": {"size": [480, 640], "counts": "]mi2c0V>d0\\O;E;E:H6K4K5K5L4L4M4M2M3N2M3M3N3L3N2M3M3N2M4M2M2N2O1N2O1NJ63N2M2N3N1N3N2M2N3N2M2O2M2N3N2M2O2M3MFiFPKV9Q5jFPKS9R5nFmJQ9T5oFlJo8W5QGhJm8Z5SGfJl8[5UGdJi8^5WGbJh8_5YG_Jf8c5ZG]Jd8e5]GZJb8g5^GYJ`8i5aGVJ^8k5cGSJ\\8P6cGPJ[8R6fGmIY8T6gGlIW8V681N2N2O0O1O10OQLcI^O^6=kI_OT6=UJ^Ol5d0VJWOk5k0XJPOg5T1ZJgNg5[1[JaNe5a1]JZNc5i1`JRN`5Q2aJkM^5X2cJeM]5]2eJ_M[5c2gJYMX5k2iJQMW5Q3kJkLU5W3mJeLR5_3nJ_LQ5c3QKYLo4i3SKSLl4P4VKlKj4X4VKdKi4a4WK[Ki4h4YKQKi4S5XKeJj4`5XKWJk4m5VKlIl4W6k15L4L3L5L4K4M4L4K4M4L4K5L3L5L4L3L5L4L3L5L4L3L5L3L5L4L3L5L3M4K4M3M3L3N3L4M3M3L3N3M3L4M3L3N3M3L4M3M4K4M4K5L3L5L4K4L5L4K4ZOg0BUlV4"}, "label": "banana with less black dots"}]}
{"id": 112577, "image": "COCO_train2014_000000112577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two bananas hang on a rack\"."}], "task": "refering", "answer_template": "The \"two bananas hang on a rack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 33, 34, 56, 57, 58, 59, 81, 82, 83, 104, 105, 106, 127, 128, 129, 150, 151, 152, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 285, 286, 287, 288, 308, 309, 310, 331, 332, 333, 354, 355], "bbox": [0.40235937499999996, 0.018979166666666665, 0.6547031249999999, 0.9248541666666666], "iscrowd": 0, "area": 27269.603199999994, "rle": {"size": [480, 640], "counts": "aZi3<b>9F:G9F6K3L3N3M2M4M2M4M2N3L3N2M4M2M4WE^MP9d2PGdMf8]2ZGkM\\8W2eGmMU8U2jGmMR8U2oGmMm7T2THmMi7U2VHnMe7T2\\HmMa7U2^HnM]7T2dHmMY7T2iHmMS7U2mHlMo6V2RIlMj6U2XIkMd6W2]IkM_6V2cIjMY6X2hIjMT6W2mIjMP6W2RJjMi5X2XJiMe5X2\\JjM_5X2cJhMZ5Y2gJiMU5X2lJiMP5Y2RKhMj4Y2YKfMc4\\2aKaM]4`2fK_MV4c2mK[MQ4f2QLZMl3g2VLXMg3j2[LVMb3k2`LUM\\3l2gLSMW3m2lLSMQ3m2RMRMk2P3WMPMf2P3]MoL`2R3dMmLY2T3iMlLS2U3PNjLn1V3UNjLh1W3ZNhLc1Y3`NgL]1Y3fNfLW1\\3kNdLR1\\3QOdLl0\\3WOcLf0_3\\OaLa0_3B`L;a3H_L5a3N^L0c32]LJd39\\LCe3`0ZL]Oh3e0XLWOi3l0VLQOk3R1WLhNk3Z1WL`Nj3c1XLVNj3m1WLnMk3T2WLfMj3]2XL]Mi3e2YLVMh3m2ZLmLh3T3[LfLf3]3\\L\\Lf3g3[LTLg3m3\\LmKe3V4^LcKc3_4bLYK`3i4dLoJ]3T5gLcJ[3_5jLYJX3i5kLPJV3R6oLfIR3]6g22M2N3M2O2M3M2O2M2N3N2M2N3N1N3M3N1N3M2N3N2M2N3M2N3M3M2N3L3N3M3M2N3M3M2N3M2N3M3M2N3M4L4L3L5L4L4L3M4L4L4L3M4L4L4L3M4I7A?B=B?B>A?B=CX_W3"}, "label": "two bananas hang on a rack"}]}
{"id": 112577, "image": "COCO_train2014_000000112577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana with the most brown spots\"."}], "task": "refering", "answer_template": "The \"the banana with the most brown spots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 33, 34, 56, 57, 58, 59, 81, 82, 83, 104, 105, 106, 127, 128, 129, 150, 151, 152, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 285, 286, 287, 288, 308, 309, 310, 331, 332, 333, 354, 355], "bbox": [0.40235937499999996, 0.018979166666666665, 0.6547031249999999, 0.9248541666666666], "iscrowd": 0, "area": 27269.603199999994, "rle": {"size": [480, 640], "counts": "aZi3<b>9F:G9F6K3L3N3M2M4M2M4M2N3L3N2M4M2M4WE^MP9d2PGdMf8]2ZGkM\\8W2eGmMU8U2jGmMR8U2oGmMm7T2THmMi7U2VHnMe7T2\\HmMa7U2^HnM]7T2dHmMY7T2iHmMS7U2mHlMo6V2RIlMj6U2XIkMd6W2]IkM_6V2cIjMY6X2hIjMT6W2mIjMP6W2RJjMi5X2XJiMe5X2\\JjM_5X2cJhMZ5Y2gJiMU5X2lJiMP5Y2RKhMj4Y2YKfMc4\\2aKaM]4`2fK_MV4c2mK[MQ4f2QLZMl3g2VLXMg3j2[LVMb3k2`LUM\\3l2gLSMW3m2lLSMQ3m2RMRMk2P3WMPMf2P3]MoL`2R3dMmLY2T3iMlLS2U3PNjLn1V3UNjLh1W3ZNhLc1Y3`NgL]1Y3fNfLW1\\3kNdLR1\\3QOdLl0\\3WOcLf0_3\\OaLa0_3B`L;a3H_L5a3N^L0c32]LJd39\\LCe3`0ZL]Oh3e0XLWOi3l0VLQOk3R1WLhNk3Z1WL`Nj3c1XLVNj3m1WLnMk3T2WLfMj3]2XL]Mi3e2YLVMh3m2ZLmLh3T3[LfLf3]3\\L\\Lf3g3[LTLg3m3\\LmKe3V4^LcKc3_4bLYK`3i4dLoJ]3T5gLcJ[3_5jLYJX3i5kLPJV3R6oLfIR3]6g22M2N3M2O2M3M2O2M2N3N2M2N3N1N3M3N1N3M2N3N2M2N3M2N3M3M2N3L3N3M3M2N3M3M2N3M2N3M3M2N3M4L4L3L5L4L4L3M4L4L4L3M4L4L4L3M4I7A?B=B?B>A?B=CX_W3"}, "label": "the banana with the most brown spots"}]}
{"id": 14283, "image": "COCO_train2014_000000014283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball catcher\"."}], "task": "refering", "answer_template": "The \"a baseball catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 108, 109, 110, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 223, 224, 225, 226, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319, 340, 341, 363, 364, 386, 387], "bbox": [0.665734375, 0.1505625, 0.9842656250000001, 0.9842708333333333], "iscrowd": 0, "area": 32453.559600000008, "rle": {"size": [480, 640], "counts": "alW63h>8I6I8I6I7I8I6I8I6I7K6L3L5K3N1N2N2N2O0O2N2N2O1O1O1O1O1O100O00O100O2N10SORMiEm2S:`MeE_2W:mMcEZ2S:f1D;E<E8G4L5K4M3L4L5L3L4L5K4M3L4L5L3L4N3L3M3N2M4L3N3L3M3N3L3M4M2M3M4M2M3N3L3M4M2M3QMSFe0Q:QOUFP1m9fNZFZ1j9\\N[Fe1h9RN^Fo1d9hMbFX2b9^McFc2m:1O0O101YEXMT9h2jF`MP9c2jFbMS9d2`FdM`9`2QFkMo9c3O2O00000O2O000O101O0O10001O0O1WOPKbGP5[8TKdGm4X8WKhGh4U8\\KkGc4T8`KjG`4U8bKkG^4S8dKmG[4Q8hKoGW4P8kKPHU4n7nKPHR4R1XKR5h0lIo3S1cKe4`0WJm3U1mKW49cJk3W1SLl36jJg3[1XLd35oJb3^1^L\\35RK_3a1aLV34VK\\3d1eLn2X5QMlJi2V5WMoJb2T5]MQK\\2Q5dMTKR2Q5oMSKg1S5XNRK_1R5aNUKS1Q5lNWKg0n4YOZK:l4E\\KOh41`KBf4=bKVOc4j0eKiN`4V1g4N2N1O2N2N2O0O2N2N2N1O1O2N1O1O1O1O1O0000O10O01000O010O010O010O01000O010O010O010O010O00100O001O000000001O0000001O1O001O1O010O^`4"}, "label": "a baseball catcher"}]}
{"id": 14283, "image": "COCO_train2014_000000014283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher stands during a game\"."}], "task": "refering", "answer_template": "The \"a catcher stands during a game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 108, 109, 110, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 223, 224, 225, 226, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319, 340, 341, 363, 364, 386, 387], "bbox": [0.665734375, 0.1505625, 0.9842656250000001, 0.9842708333333333], "iscrowd": 0, "area": 32453.559600000008, "rle": {"size": [480, 640], "counts": "alW63h>8I6I8I6I7I8I6I8I6I7K6L3L5K3N1N2N2N2O0O2N2N2O1O1O1O1O1O100O00O100O2N10SORMiEm2S:`MeE_2W:mMcEZ2S:f1D;E<E8G4L5K4M3L4L5L3L4L5K4M3L4L5L3L4N3L3M3N2M4L3N3L3M3N3L3M4M2M3M4M2M3N3L3M4M2M3QMSFe0Q:QOUFP1m9fNZFZ1j9\\N[Fe1h9RN^Fo1d9hMbFX2b9^McFc2m:1O0O101YEXMT9h2jF`MP9c2jFbMS9d2`FdM`9`2QFkMo9c3O2O00000O2O000O101O0O10001O0O1WOPKbGP5[8TKdGm4X8WKhGh4U8\\KkGc4T8`KjG`4U8bKkG^4S8dKmG[4Q8hKoGW4P8kKPHU4n7nKPHR4R1XKR5h0lIo3S1cKe4`0WJm3U1mKW49cJk3W1SLl36jJg3[1XLd35oJb3^1^L\\35RK_3a1aLV34VK\\3d1eLn2X5QMlJi2V5WMoJb2T5]MQK\\2Q5dMTKR2Q5oMSKg1S5XNRK_1R5aNUKS1Q5lNWKg0n4YOZK:l4E\\KOh41`KBf4=bKVOc4j0eKiN`4V1g4N2N1O2N2N2O0O2N2N2N1O1O2N1O1O1O1O1O0000O10O01000O010O010O010O01000O010O010O010O010O00100O001O000000001O0000001O1O001O1O010O^`4"}, "label": "a catcher stands during a game"}]}
{"id": 14283, "image": "COCO_train2014_000000014283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a bat wearing a red helmet\"."}], "task": "refering", "answer_template": "The \"a man with a bat wearing a red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 78, 79, 80, 81, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 308, 309, 310, 311, 312, 329, 330, 331, 332, 334, 335, 336, 352, 353, 354, 357, 358, 359], "bbox": [0.32865625, 0.16591666666666666, 0.637078125, 0.9456875], "iscrowd": 0, "area": 30639.226799999993, "rle": {"size": [480, 640], "counts": "TkR31o>5K001N2O001O0O2O001O1N1010O01O1O010O0010O01O1O010O0010O01O1O001O001O1O0O2M2N3M2N3M3M2N3M2N3M2N3WI`NNc1JdN4^1EiN9Y1@nN>T1ZOTOd0m0VOZOh0h0QO_Om0c0kNES1=fNJX16cN1[1O_N:^1FcN;[1EgN;W1EkN;S1EoN;o0ERO<l0DVO<h0EYO;3_MgKU2W4=0bMeKQ2]4=LeMdKn1b4=HiMbKj1g4>EkMaKg1l4>AnM`Kd1Q5>]O5d0LZO4h0LWO4j0LTO4n0kNlJYOT4l1R1eNPK_Ol3l1W1_NQKFf3k1\\1XNTKM^3k1`1RNXK3V3k1e1kMZK;o2j1j1dM]Kb0g2k1n1\\MaKi0_2k1b3VN\\Lj1f3VNXLj1]2dMkJc0f2i1n1XN[KOe2i1Q2ZNYKMd2i1T2\\NVKLd2h1V2_NUKIc2h1Y2bNQKGd2h1[2cNPKEc2h1]2fNoJBb2h1`2hNlJAb2g1b2mNiJEZ2^1n2nNgJf0X1<R4oNfJb19_OR5QOdJa17^OW5QOaJb16]O[5QO_Jc13]O_5PO]Je11[Od5PO[Jf1NZOi5POYJf1LZOm5QOUJg1KXOR6QOSJh1HYOU6oNRJi4m5WKSJj4l5VKTJl4j5UKTJm4k5SKUJn4j5SKTJo4k5QKUJQ5j5oJUJR5j5nJUJT5j5mJUJU5i5R2O100O2N1O1O2O0O1O1O2N1O101N1TL]FY2d9bM`F^2a9_MbF`2`9^M`Fc2`9[MaFe2`9ZM`Fg2`9WMaFl2^9RMbFP3]9nLeFT3Y9kLgFX3X9eLiF^3U9aLkFb3S9]LmFf30gK_8a0bGk3IhKc6KUK`0_No3DjKe6JXK;`Na4U6XKZK6aNc4R6ZK^K0aNg4n5]K`KKaNj4l5^KcKFbNm4h5aKeKAcNo4e5cKiK[OcNS5a5fKkKVOdNU5^5hKnKQOdNY5[5iKeLX4Y3kKeLV4X3mKgLT4V3PLhLQ4U3RLjLo3S3ULjLm3S3VLlLk3Q3YLmLh3P3[LoLf3n2^LPMc3m2`LRMa3k2bLSM`3k2cLSM^3j2eLUM\\3h2hLVMY3f2kLYMX3b2lL[MY3^2kLaMY35nI2m2GY32SJLh21Y3NXJGc29Y3K]JA^2c0Z3EbJ^OW2k0\\3BfJXOS2T1[3_OdN`0`1[OcNc0a1XOaNg0c1TO`Ni0e1QO^Nn0g1lN\\NR1j7O2N1O2M2O2N1O1O2N1N3N1O2N1OhZ\\3"}, "label": "a man with a bat wearing a red helmet"}]}
{"id": 14283, "image": "COCO_train2014_000000014283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player with white dress and red helmet\"."}], "task": "refering", "answer_template": "The \"a player with white dress and red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 78, 79, 80, 81, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 308, 309, 310, 311, 312, 329, 330, 331, 332, 334, 335, 336, 352, 353, 354, 357, 358, 359], "bbox": [0.32865625, 0.16591666666666666, 0.637078125, 0.9456875], "iscrowd": 0, "area": 30639.226799999993, "rle": {"size": [480, 640], "counts": "TkR31o>5K001N2O001O0O2O001O1N1010O01O1O010O0010O01O1O010O0010O01O1O001O001O1O0O2M2N3M2N3M3M2N3M2N3M2N3WI`NNc1JdN4^1EiN9Y1@nN>T1ZOTOd0m0VOZOh0h0QO_Om0c0kNES1=fNJX16cN1[1O_N:^1FcN;[1EgN;W1EkN;S1EoN;o0ERO<l0DVO<h0EYO;3_MgKU2W4=0bMeKQ2]4=LeMdKn1b4=HiMbKj1g4>EkMaKg1l4>AnM`Kd1Q5>]O5d0LZO4h0LWO4j0LTO4n0kNlJYOT4l1R1eNPK_Ol3l1W1_NQKFf3k1\\1XNTKM^3k1`1RNXK3V3k1e1kMZK;o2j1j1dM]Kb0g2k1n1\\MaKi0_2k1b3VN\\Lj1f3VNXLj1]2dMkJc0f2i1n1XN[KOe2i1Q2ZNYKMd2i1T2\\NVKLd2h1V2_NUKIc2h1Y2bNQKGd2h1[2cNPKEc2h1]2fNoJBb2h1`2hNlJAb2g1b2mNiJEZ2^1n2nNgJf0X1<R4oNfJb19_OR5QOdJa17^OW5QOaJb16]O[5QO_Jc13]O_5PO]Je11[Od5PO[Jf1NZOi5POYJf1LZOm5QOUJg1KXOR6QOSJh1HYOU6oNRJi4m5WKSJj4l5VKTJl4j5UKTJm4k5SKUJn4j5SKTJo4k5QKUJQ5j5oJUJR5j5nJUJT5j5mJUJU5i5R2O100O2N1O1O2O0O1O1O2N1O101N1TL]FY2d9bM`F^2a9_MbF`2`9^M`Fc2`9[MaFe2`9ZM`Fg2`9WMaFl2^9RMbFP3]9nLeFT3Y9kLgFX3X9eLiF^3U9aLkFb3S9]LmFf30gK_8a0bGk3IhKc6KUK`0_No3DjKe6JXK;`Na4U6XKZK6aNc4R6ZK^K0aNg4n5]K`KKaNj4l5^KcKFbNm4h5aKeKAcNo4e5cKiK[OcNS5a5fKkKVOdNU5^5hKnKQOdNY5[5iKeLX4Y3kKeLV4X3mKgLT4V3PLhLQ4U3RLjLo3S3ULjLm3S3VLlLk3Q3YLmLh3P3[LoLf3n2^LPMc3m2`LRMa3k2bLSM`3k2cLSM^3j2eLUM\\3h2hLVMY3f2kLYMX3b2lL[MY3^2kLaMY35nI2m2GY32SJLh21Y3NXJGc29Y3K]JA^2c0Z3EbJ^OW2k0\\3BfJXOS2T1[3_OdN`0`1[OcNc0a1XOaNg0c1TO`Ni0e1QO^Nn0g1lN\\NR1j7O2N1O2M2O2N1O1O2N1N3N1O2N1OhZ\\3"}, "label": "a player with white dress and red helmet"}]}
{"id": 235468, "image": "COCO_train2014_000000235468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman with long hair wearing black and white striped pants sitting in the snow near a road\"."}], "task": "refering", "answer_template": "The \"a young woman with long hair wearing black and white striped pants sitting in the snow near a road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260], "bbox": [0.18334375, 0.2696728971962617, 0.47339062500000006, 0.7685280373831775], "iscrowd": 0, "area": 21441.833250000007, "rle": {"size": [428, 640], "counts": "hVa11Q=:YOg0L3O2O1O1O010O1O00100O1O00100O001TDeNf;\\1ZDfNd;Z1\\DhNa;X1_DlN];U1cDlN[;U1eDlNY;T1hDmNV;T1iDnNU;S1kDnNR;T1nDlNQ;T1oDmNP;T1PElNo:U1QEkNm:V1TEjNk:W1TEjNk:W1UEiNj:m1O1O1O1O1O1O1O1O1O1O1O1O1O10O01O10O01O10O010O10000O101O0O10001N10001N01000O01000O0100E;G9H8N1O2N2O1O1O1O1O001O1O1O1O1O1N101O00001O0H9G8J7O001N101N101N1N3M2M4N1O2N1O2O1N101O1N1100O01000O01000O10O010000O10O11N10001N101N10001N2O1O1N2O1O2N1O2N2N1O2N3M2N2N3M2N2N2N3M2N2N2N3N3L3M3M4L2N2N2N2XNSFMo9LYF1h9J`F2c9GeF5^9CkF9Y9_OPGCiN=[:GcG3a8EgG7j:K4L^b\\4"}, "label": "a young woman with long hair wearing black and white striped pants sitting in the snow near a road"}]}
{"id": 235468, "image": "COCO_train2014_000000235468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman in blue shirt and striped pants sitting in the snow\"."}], "task": "refering", "answer_template": "The \"a young woman in blue shirt and striped pants sitting in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260], "bbox": [0.18334375, 0.2696728971962617, 0.47339062500000006, 0.7685280373831775], "iscrowd": 0, "area": 21441.833250000007, "rle": {"size": [428, 640], "counts": "hVa11Q=:YOg0L3O2O1O1O010O1O00100O1O00100O001TDeNf;\\1ZDfNd;Z1\\DhNa;X1_DlN];U1cDlN[;U1eDlNY;T1hDmNV;T1iDnNU;S1kDnNR;T1nDlNQ;T1oDmNP;T1PElNo:U1QEkNm:V1TEjNk:W1TEjNk:W1UEiNj:m1O1O1O1O1O1O1O1O1O1O1O1O1O10O01O10O01O10O010O10000O101O0O10001N10001N01000O01000O0100E;G9H8N1O2N2O1O1O1O1O001O1O1O1O1O1N101O00001O0H9G8J7O001N101N101N1N3M2M4N1O2N1O2O1N101O1N1100O01000O01000O10O010000O10O11N10001N101N10001N2O1O1N2O1O2N1O2N2N1O2N3M2N2N3M2N2N2N3M2N2N2N3N3L3M3M4L2N2N2N2XNSFMo9LYF1h9J`F2c9GeF5^9CkF9Y9_OPGCiN=[:GcG3a8EgG7j:K4L^b\\4"}, "label": "a young woman in blue shirt and striped pants sitting in the snow"}]}
{"id": 235468, "image": "COCO_train2014_000000235468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a red coat leaning over\"."}], "task": "refering", "answer_template": "The \"a person in a red coat leaning over\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 269, 270, 271], "bbox": [0.4989375, 0.16869158878504673, 0.8355625, 0.7529672897196261], "iscrowd": 0, "area": 33417.635200000004, "rle": {"size": [428, 640], "counts": "X^U43Y=f0YO2O1N2O1N2O1N2O1N3M2O1O1N2O1O1O1O1N2O1O2N1O1O1N2O1O1O1N2O1O2M10000O01000O10O10O1000O1000O10O10000O1000O10O1000O10O01WH\\Nj3d1TK^Ol4b0SJ`0k5@TIc1k6_2O2O0O2O1N101O1N101N2O0O2O0OO2F:F9G:H7K_IgJ[5S5kJmJT5n4QKSKo4g4VKZKi4a4]K`Kb4Z4cKgK]4S4iKmKV4T4iKmKW4U4fKlKY4W4eKiK[4Z4aKgK^4\\4_KeKa4]4]KdKa4_4\\KbKd4a4YK_Kg4c4VK^Ki4e4TK\\Kl4f4RKZKm4j4oJWKQ5k4mJUKR5n4kJSKU5o4hJSKV5Q5gJoJY5S5dJnJ\\5T5bJlJ]5W5`JjJ`5Y5\\JhJc5[5[JeJe5]5XJdJg5_5WJaJi5a5TJaJj5c5RJ^Jn5d5PJ\\JP6f5mI[JR6`60000000001O00000000000000001OkLQJmNo5o0[JkNe5R1dJhN\\5U1lJfNT5Y1SKaNm4`1XKZNh4g1]KSNc4o1aKlM^4U2hKdMX4]2mK]MS4d2RLiL[4X3jKPLh4R4X21O1O1O2_IiKi3X4RLaLY3`3bLdL^3]3]LhLb3Y3ZLjLf3X3TLlLl3U3oKoLP4S3kKQMU4P3dKVM\\4k2YK_Mg4c5000000000001O000000000001O00001O00001O010O00001O00001O001O00001O010O00001O0WLRKTOn4j0SLWNn3e1^LRNc3k1eLoM\\3n1lLlMU3P2TMgMo2W2YMaMj2\\2^M\\Me2`2dMWM`2f2U4K5L4L4L4L4K6K8H7B?_O`0@Ui[1"}, "label": "a person in a red coat leaning over"}]}
{"id": 235468, "image": "COCO_train2014_000000235468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in an orange jacket bending over\"."}], "task": "refering", "answer_template": "The \"a skier in an orange jacket bending over\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 269, 270, 271], "bbox": [0.4989375, 0.16869158878504673, 0.8355625, 0.7529672897196261], "iscrowd": 0, "area": 33417.635200000004, "rle": {"size": [428, 640], "counts": "X^U43Y=f0YO2O1N2O1N2O1N2O1N3M2O1O1N2O1O1O1O1N2O1O2N1O1O1N2O1O1O1N2O1O2M10000O01000O10O10O1000O1000O10O10000O1000O10O1000O10O01WH\\Nj3d1TK^Ol4b0SJ`0k5@TIc1k6_2O2O0O2O1N101O1N101N2O0O2O0OO2F:F9G:H7K_IgJ[5S5kJmJT5n4QKSKo4g4VKZKi4a4]K`Kb4Z4cKgK]4S4iKmKV4T4iKmKW4U4fKlKY4W4eKiK[4Z4aKgK^4\\4_KeKa4]4]KdKa4_4\\KbKd4a4YK_Kg4c4VK^Ki4e4TK\\Kl4f4RKZKm4j4oJWKQ5k4mJUKR5n4kJSKU5o4hJSKV5Q5gJoJY5S5dJnJ\\5T5bJlJ]5W5`JjJ`5Y5\\JhJc5[5[JeJe5]5XJdJg5_5WJaJi5a5TJaJj5c5RJ^Jn5d5PJ\\JP6f5mI[JR6`60000000001O00000000000000001OkLQJmNo5o0[JkNe5R1dJhN\\5U1lJfNT5Y1SKaNm4`1XKZNh4g1]KSNc4o1aKlM^4U2hKdMX4]2mK]MS4d2RLiL[4X3jKPLh4R4X21O1O1O2_IiKi3X4RLaLY3`3bLdL^3]3]LhLb3Y3ZLjLf3X3TLlLl3U3oKoLP4S3kKQMU4P3dKVM\\4k2YK_Mg4c5000000000001O000000000001O00001O00001O010O00001O00001O001O00001O010O00001O0WLRKTOn4j0SLWNn3e1^LRNc3k1eLoM\\3n1lLlMU3P2TMgMo2W2YMaMj2\\2^M\\Me2`2dMWM`2f2U4K5L4L4L4L4K6K8H7B?_O`0@Ui[1"}, "label": "a skier in an orange jacket bending over"}]}
{"id": 235468, "image": "COCO_train2014_000000235468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing in the backround watching people getting ready to ski\"."}], "task": "refering", "answer_template": "The \"a man standing in the backround watching people getting ready to ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 79, 80, 81, 102, 103, 125, 126, 127, 149, 150, 172, 173, 195, 196, 218, 219], "bbox": [0.44982812499999997, 0.12182242990654206, 0.562296875, 0.6150934579439252], "iscrowd": 0, "area": 7167.874450000001, "rle": {"size": [428, 640], "counts": "m`h37l<:F;E:F8H7I8O0100O1O100O1O100O11O5K4L4L5K4K^1cN?A>B?A7I2N1OdK^HQ3`7oLeHn2Y7RMmHi2n6[MXIa2`6eMfIW2T6mMRJo1l5QNYJk1f5UN`Je1a5ZN`Je1`5[NbJd1^5\\NbJc1^5]NdJa1\\5`NeJ^1\\5bNdJ_1Z5aNhJ^1W5cNiJ^1W5`NkJ_1V5`NjJa1V5^NkJa1V5]NkJd1U5[NlJd1U5ZNmJf1S5YNmJg1T5WNnJe1V5ZNjJc1[5[NfJ`1j5SNWJi1m5WNSJc1T6[NmI_1Z6`NfIY1b6fN^IT1h6kNZIm0n6RORIh0U7VOlHc0\\7\\OdH>b7A_H9h7FYH2o7MQHMU82mGF[89]2N2O0O[nd3"}, "label": "a man standing in the backround watching people getting ready to ski"}]}
{"id": 235468, "image": "COCO_train2014_000000235468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the background who is walking up to the snowboarders\"."}], "task": "refering", "answer_template": "The \"the man in the background who is walking up to the snowboarders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 79, 80, 81, 102, 103, 125, 126, 127, 149, 150, 172, 173, 195, 196, 218, 219], "bbox": [0.44982812499999997, 0.12182242990654206, 0.562296875, 0.6150934579439252], "iscrowd": 0, "area": 7167.874450000001, "rle": {"size": [428, 640], "counts": "m`h37l<:F;E:F8H7I8O0100O1O100O1O100O11O5K4L4L5K4K^1cN?A>B?A7I2N1OdK^HQ3`7oLeHn2Y7RMmHi2n6[MXIa2`6eMfIW2T6mMRJo1l5QNYJk1f5UN`Je1a5ZN`Je1`5[NbJd1^5\\NbJc1^5]NdJa1\\5`NeJ^1\\5bNdJ_1Z5aNhJ^1W5cNiJ^1W5`NkJ_1V5`NjJa1V5^NkJa1V5]NkJd1U5[NlJd1U5ZNmJf1S5YNmJg1T5WNnJe1V5ZNjJc1[5[NfJ`1j5SNWJi1m5WNSJc1T6[NmI_1Z6`NfIY1b6fN^IT1h6kNZIm0n6RORIh0U7VOlHc0\\7\\OdH>b7A_H9h7FYH2o7MQHMU82mGF[89]2N2O0O[nd3"}, "label": "the man in the background who is walking up to the snowboarders"}]}
{"id": 538574, "image": "COCO_train2014_000000538574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the asian sign\"."}], "task": "refering", "answer_template": "The \"the asian sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 158, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229], "bbox": [0.31485416666666666, 0.38264062499999996, 0.6988333333333333, 0.5860624999999999], "iscrowd": 0, "area": 17597.92275, "rle": {"size": [640, 480], "counts": "adn24ic0:H8I7I7I7H8I7I7I7I7I7H8I7I5K00001N1000001O000O1000O100000O10O100000O1000O100000O10O10000000O10O10000000O10O10000000O01000000000O01000000000O0100000000O01000000000O010000000O10O10000000O10O10000000O10O100002M3N2N2N2N3M2M3N0000000O10O1001O000O10000000000O100000O10O10O1HZMV_Of2h`0_MT_Ob2i`0<L3N3000000000O10O100000O10000000000O10O100000O100000003M5J6K4L5K5K5K5J6K4L5K5K5J6K4L5K5J6I7HdYj2"}, "label": "the asian sign"}]}
{"id": 538574, "image": "COCO_train2014_000000538574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tablet screen with korean characters\"."}], "task": "refering", "answer_template": "The \"a tablet screen with korean characters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 158, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229], "bbox": [0.31485416666666666, 0.38264062499999996, 0.6988333333333333, 0.5860624999999999], "iscrowd": 0, "area": 17597.92275, "rle": {"size": [640, 480], "counts": "adn24ic0:H8I7I7I7H8I7I7I7I7I7H8I7I5K00001N1000001O000O1000O100000O10O100000O1000O100000O10O10000000O10O10000000O10O10000000O01000000000O01000000000O0100000000O01000000000O010000000O10O10000000O10O10000000O10O100002M3N2N2N2N3M2M3N0000000O10O1001O000O10000000000O100000O10O10O1HZMV_Of2h`0_MT_Ob2i`0<L3N3000000000O10O100000O10000000000O10O100000O100000003M5J6K4L5K5K5K5J6K4L5K5K5J6K4L5K5J6I7HdYj2"}, "label": "a tablet screen with korean characters"}]}
{"id": 538574, "image": "COCO_train2014_000000538574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white laptop on a white desk running windows operating system\"."}], "task": "refering", "answer_template": "The \"a white laptop on a white desk running windows operating system\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [282, 283, 284, 285, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 337, 347, 348, 349, 350, 351, 352, 353, 354, 365, 366, 367, 368, 369, 370], "bbox": [0.42685416666666665, 0.7059375, 0.8945208333333333, 0.957796875], "iscrowd": 0, "area": 20623.71065, "rle": {"size": [640, 480], "counts": "mYQ41nc02O0O101N1O101N100O1O2O0O101N100O2N100O2O0O1O101N100O2O0O1O2O0O100O2N100O2O0O1O2O0O100O2O0O1O2O0O101N1O100O2O0O101L3M3M4K4M3M3M4L3L4M4L3M3L5L3M3M3M4K4M3M4L3L4M3M4L3M3L4O2O0000001O00001O01O0001O01O0001O01O01O0000010O0000010O0000010O00001O01O0001O01O0001O01O0001O01O01O0000O2N100O101N1O100O2N100O100O2N100O1O2O0O100O2N100O101N1O100O1O2O0O100O2N100O1O2O0O100O2N100O100O2N100O1O2O0O11O100O2NN2K5L5J5K5L4K6K4K5K5L5L3N2M3M4L3M3M3N2M4L3M3M3M4M2M3M3M4LP]o0"}, "label": "a white laptop on a white desk running windows operating system"}]}
{"id": 538574, "image": "COCO_train2014_000000538574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop on the table\"."}], "task": "refering", "answer_template": "The \"a laptop on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [282, 283, 284, 285, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 337, 347, 348, 349, 350, 351, 352, 353, 354, 365, 366, 367, 368, 369, 370], "bbox": [0.42685416666666665, 0.7059375, 0.8945208333333333, 0.957796875], "iscrowd": 0, "area": 20623.71065, "rle": {"size": [640, 480], "counts": "mYQ41nc02O0O101N1O101N100O1O2O0O101N100O2N100O2O0O1O101N100O2O0O1O2O0O100O2N100O2O0O1O2O0O100O2O0O1O2O0O101N1O100O2O0O101L3M3M4K4M3M3M4L3L4M4L3M3L5L3M3M3M4K4M3M4L3L4M3M4L3M3L4O2O0000001O00001O01O0001O01O0001O01O01O0000010O0000010O0000010O00001O01O0001O01O0001O01O0001O01O01O0000O2N100O101N1O100O2N100O100O2N100O1O2O0O100O2N100O101N1O100O1O2O0O100O2N100O1O2O0O100O2N100O100O2N100O1O2O0O11O100O2NN2K5L5J5K5L4K6K4K5K5L5L3N2M3M4L3M3M3N2M4L3M3M3M4M2M3M3M4LP]o0"}, "label": "a laptop on the table"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one out of four giraffes , facing away from the camera and behind a tree\"."}], "task": "refering", "answer_template": "The \"one out of four giraffes , facing away from the camera and behind a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 122, 123, 124, 138, 139, 153, 154, 168, 169, 170, 184, 185, 200, 215, 230], "bbox": [0.18072599531615927, 0.2984375, 0.40117096018735365, 0.666140625], "iscrowd": 0, "area": 5143.058400000001, "rle": {"size": [640, 427], "counts": "[[`19fc02M3N2M2N2O1N2O2M2O1N2N2N2N2N2M4M2N2N2M3N2N2N20O2O1O2NY^O^NQa0`1P_OcNn`0\\1R_OgNm`0W1S_OlNk`0R1V_OROg`0m0Y_O\\O``0b0`_OHW`07i_O3n?KS@:j?DU@?j?@V@a0j?^OV@d0j?YOW@h0i?WOW@k0h?TOX@m0i?POX@R1g?mNY@T1h?jNX@X1g?gNX@[1i?bNX@`1g?_NY@b1h?\\NX@e1h?ZN\\@d1d?YNa@d1_?[Nd@d1[?[Ni@b1X?[Nm@c1R?\\NQAb1P?\\NTAb1k>]NYA`1h>]N]Aa1c`0O2M3N00O10O10000O10000O1001N4M2N2N2N2N3M2N=C3LN2O2M2O1N2O2M3N1G:E;F:E;Egdo4"}, "label": "one out of four giraffes , facing away from the camera and behind a tree"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is the giraffe that is second from the left\"."}], "task": "refering", "answer_template": "The \"this is the giraffe that is second from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 122, 123, 124, 138, 139, 153, 154, 168, 169, 170, 184, 185, 200, 215, 230], "bbox": [0.18072599531615927, 0.2984375, 0.40117096018735365, 0.666140625], "iscrowd": 0, "area": 5143.058400000001, "rle": {"size": [640, 427], "counts": "[[`19fc02M3N2M2N2O1N2O2M2O1N2N2N2N2N2M4M2N2N2M3N2N2N20O2O1O2NY^O^NQa0`1P_OcNn`0\\1R_OgNm`0W1S_OlNk`0R1V_OROg`0m0Y_O\\O``0b0`_OHW`07i_O3n?KS@:j?DU@?j?@V@a0j?^OV@d0j?YOW@h0i?WOW@k0h?TOX@m0i?POX@R1g?mNY@T1h?jNX@X1g?gNX@[1i?bNX@`1g?_NY@b1h?\\NX@e1h?ZN\\@d1d?YNa@d1_?[Nd@d1[?[Ni@b1X?[Nm@c1R?\\NQAb1P?\\NTAb1k>]NYA`1h>]N]Aa1c`0O2M3N00O10O10000O10000O1001N4M2N2N2N2N3M2N=C3LN2O2M2O1N2O2M3N1G:E;F:E;Egdo4"}, "label": "this is the giraffe that is second from the left"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two girafees on the left hand corner next to other girafees in tall grass\"."}], "task": "refering", "answer_template": "The \"two girafees on the left hand corner next to other girafees in tall grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 120, 121, 122, 136, 137, 151, 152, 153, 167, 168, 169, 182, 183, 184, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 286], "bbox": [0.029250585480093676, 0.316046875, 0.3689461358313818, 0.8667343750000001], "iscrowd": 0, "area": 20004.074000000004, "rle": {"size": [640, 427], "counts": "b[82mc01O2N2N1O2O0O2O1N101N2O0ZOAj]Oa0Tb0N^]O6]b0k0J7J6K5K3M1O2N2N2N0002N\\EfMj3R2QLUNT4d1gKdNV4\\1fKkNU4W1dKROW4Q1`K[OZ4h0^KB]4a0ZKd0g3_OQLi0k3[OlKn0P4VOhKR1Y4mN^K]1a4eNPKd1T5_N\\Jj1i5XNfIR2^6QNRIX2R7jM_H]2h7eMoG[2a8^MZG^2Z9VMbFf2d9]MVF`2P:cMkE^1V;eNeDZ1^;dLeC_1i0l1d;eLeCa1b0j1k;eLeCb1;h1S<eLeCl5^<SJdCi5_<VJdCe5_<[JcCa5_<^JcC^5`<aJcCZ5`<fJaCV5b<iJaCS5a<lJbCo4`<QKbCk4a<UKaCh4`<WKbCh4^<WKeCf4\\<YKgCe4Y<[KhCc4Y<\\KiCc4W<\\KlCa4U<_KlC`4T<_KnC`4Q<`KRD_4h=O101O0O101O0O1015J6J6KQ2nMN2O2N1N2O1N3N1N2O2N1N2O2H7\\Od0\\Od000:G9F:F1O0000O101O0N2J6K5K5J6K6J5K5J6K5L4M3L5L3M3M4L3M4L4L3L5L4L3M4L4L3M6J6J5I8G9G8G:G9G8GfgW5"}, "label": "two girafees on the left hand corner next to other girafees in tall grass"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 120, 121, 122, 136, 137, 151, 152, 153, 167, 168, 169, 182, 183, 184, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 286], "bbox": [0.029250585480093676, 0.316046875, 0.3689461358313818, 0.8667343750000001], "iscrowd": 0, "area": 20004.074000000004, "rle": {"size": [640, 427], "counts": "b[82mc01O2N2N1O2O0O2O1N101N2O0ZOAj]Oa0Tb0N^]O6]b0k0J7J6K5K3M1O2N2N2N0002N\\EfMj3R2QLUNT4d1gKdNV4\\1fKkNU4W1dKROW4Q1`K[OZ4h0^KB]4a0ZKd0g3_OQLi0k3[OlKn0P4VOhKR1Y4mN^K]1a4eNPKd1T5_N\\Jj1i5XNfIR2^6QNRIX2R7jM_H]2h7eMoG[2a8^MZG^2Z9VMbFf2d9]MVF`2P:cMkE^1V;eNeDZ1^;dLeC_1i0l1d;eLeCa1b0j1k;eLeCb1;h1S<eLeCl5^<SJdCi5_<VJdCe5_<[JcCa5_<^JcC^5`<aJcCZ5`<fJaCV5b<iJaCS5a<lJbCo4`<QKbCk4a<UKaCh4`<WKbCh4^<WKeCf4\\<YKgCe4Y<[KhCc4Y<\\KiCc4W<\\KlCa4U<_KlC`4T<_KnC`4Q<`KRD_4h=O101O0O101O0O1015J6J6KQ2nMN2O2N1N2O1N3N1N2O2N1N2O2H7\\Od0\\Od000:G9F:F1O0000O101O0N2J6K5K5J6K6J5K5J6K5L4M3L5L3M3M4L3M4L4L3L5L4L3M4L4L3M6J6J5I8G9G8G:G9G8GfgW5"}, "label": "the giraffe on the left"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe with butt facing this way\"."}], "task": "refering", "answer_template": "The \"giraffe with butt facing this way\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 112, 113, 127, 142, 157, 171, 172, 186, 187, 188, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 248, 260, 261, 262, 275, 276, 277, 291, 292, 306, 307, 321], "bbox": [0.3788758782201405, 0.260390625, 0.5585480093676815, 0.932078125], "iscrowd": 0, "area": 17205.129849999998, "rle": {"size": [640, 427], "counts": "ngU36dc08gML\\@<\\?Y2I7H8H8I7H7JL3N3L3N22O<G9F:G9F;F9F:G9F:G9F;F8G6J7I6J6J6J5K2N1N3L3kKUGTMo8m2SGmLo8T3TGoIMe1R9^4SGjINa1S9e4RGhIMP1KaNX9i6RGcINl0NbNc78[Ig6YO`INg0T8Y7PHnGNb0Y8]7lGnG2:Z8e7fGnG;NW8Q8`GoGm9o7UFnGo9o7SFnGP:P8RFnGQ:P8R1M2N\\OZHQEc7`:QI]En6_:VI_Ek6^:YI`Ei6\\:[IcEe6Z:_IdEc6Y:aIdEa6X:cIfE_6V:h1]I^Fc2f9RKXFe0>Q4]9WK]F<d0R4S9`K_F3P1n3e8lKcFJY1l3X8WLlI[3X6cLnIT3W6iLmIm2[6RMhI`2f6\\M_IU2U7bMnHP2e7gM_Hk1S=A?A`0@?AjTe3"}, "label": "giraffe with butt facing this way"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe on far right\"."}], "task": "refering", "answer_template": "The \"giraffe on far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 88, 89, 103, 104, 117, 118, 119, 132, 133, 134, 147, 148, 161, 162, 163, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 208, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 248, 249, 252, 253, 263, 264, 267, 268, 278, 279, 282, 283, 293, 294, 297, 298, 308, 309, 312, 313], "bbox": [0.5422716627634661, 0.1983125, 1.0, 0.9039375], "iscrowd": 0, "area": 34937.944649999976, "rle": {"size": [640, 427], "counts": "T`a48ea0h0Y_OCi>W2QARNg>R2SAXNe>l1VA^Na>f1ZAcN^>b1[AhN^>d3G8H8H5K4M2M4L4L3N2N20O10O100O100O100O10L4K4M4K4L5K4M4K5K4L5L3L5K4L5K4L5K5K4L5K4K6K4L5K5K4L4O11O001O001N101O001O001O001O001N101O001O001O001O001N101O001O1O1O1O1O0O2O1O1O1O1O1O001O1O1N2O1O1O002N1O2N1O2M3N<D;E;C>C<_DdIg8i6[FlIa9b6aERJ\\:k7L4M3L3N3L4M3L4L4M3L3N3L4M3L4L3N2M2O1N2O1N2N2O1hM_DPLc;Q4^DlKc;U4^DgKe;Y4\\DdKe;]4\\D_Kh7`8XH^Gg7e8ZHWGf7l8[HPGd7T9]HiFb7Z9_HbFa7a9`H\\F^7h9cHTF]7o9dHoEY7U:hHhEW7Z:kHdES7_:nH^EP7f:QIXEh4J_LQ;jNREb4`0ULa:ZOmD\\4T1jKS:KgDU4h1bKd99bDP4T>QLjAi3]>XL`Ad3f>]LXA]3o>c07I8I6I7I7J6I7J6I8I6I7N2M3N2M3N2N3L3N2N2M4M2N2M4M2M3N3G8I7I8H7IQL"}, "label": "giraffe on far right"}]}
{"id": 317391, "image": "COCO_train2014_000000317391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing to the right of other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe standing to the right of other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 88, 89, 103, 104, 117, 118, 119, 132, 133, 134, 147, 148, 161, 162, 163, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 208, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 248, 249, 252, 253, 263, 264, 267, 268, 278, 279, 282, 283, 293, 294, 297, 298, 308, 309, 312, 313], "bbox": [0.5422716627634661, 0.1983125, 1.0, 0.9039375], "iscrowd": 0, "area": 34937.944649999976, "rle": {"size": [640, 427], "counts": "T`a48ea0h0Y_OCi>W2QARNg>R2SAXNe>l1VA^Na>f1ZAcN^>b1[AhN^>d3G8H8H5K4M2M4L4L3N2N20O10O100O100O100O10L4K4M4K4L5K4M4K5K4L5L3L5K4L5K4L5K5K4L5K4K6K4L5K5K4L4O11O001O001N101O001O001O001O001N101O001O001O001O001N101O001O1O1O1O1O0O2O1O1O1O1O1O001O1O1N2O1O1O002N1O2N1O2M3N<D;E;C>C<_DdIg8i6[FlIa9b6aERJ\\:k7L4M3L3N3L4M3L4L4M3L3N3L4M3L4L3N2M2O1N2O1N2N2O1hM_DPLc;Q4^DlKc;U4^DgKe;Y4\\DdKe;]4\\D_Kh7`8XH^Gg7e8ZHWGf7l8[HPGd7T9]HiFb7Z9_HbFa7a9`H\\F^7h9cHTF]7o9dHoEY7U:hHhEW7Z:kHdES7_:nH^EP7f:QIXEh4J_LQ;jNREb4`0ULa:ZOmD\\4T1jKS:KgDU4h1bKd99bDP4T>QLjAi3]>XL`Ad3f>]LXA]3o>c07I8I6I7I7J6I7J6I8I6I7N2M3N2M3N2N3L3N2N2M4M2N2M4M2M3N3G8I7I8H7IQL"}, "label": "a giraffe standing to the right of other giraffes"}]}
{"id": 522191, "image": "COCO_train2014_000000522191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giant red bus drives down the road\"."}], "task": "refering", "answer_template": "The \"the giant red bus drives down the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 291, 292, 293, 294, 295], "bbox": [0.217390625, 0.20606557377049178, 0.87859375, 0.8510070257611241], "iscrowd": 0, "area": 96290.74709999998, "rle": {"size": [427, 640], "counts": "WUj15R=>A`0@?B>A`0@?A?B?@?A?B6M101N101O001N101O001N10000O2O00000O2O00000O2O002N2M2O2N2N2M3N2N2M2O00000O100000O01000000O10000O1000O10O1000000O1000000O01000O1000000O1000O10O1000000O100000O010000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O101O000O10001O0O10001O0O10000O2O00000O101O000O101O000O101O001N101O0O2O001O1N101O001N101O001N101O0O2O1O001N100000O010O100O010O100O010O10O0100O1O010O100O010O100O010O10O0100O10O1000000O1000000O10000O1000000O10000000000001O00000000000000001O00000000000000001O000000000000001O00000001O1O1O00100O1O1O1O001O1O1O10O01O1O1O1O001O1O01O01O00001O00001O01O01O0000000000O101O0O10000O10000O101O0O10000O10000O101O0O10000O10000O101O0O1000000000000O2O00000000000000001O000000000O1000001O000000000000001O00000O10000000001O000O1O1O1O1O1O1O2N1O1O1O1D<\\LTaP1"}, "label": "the giant red bus drives down the road"}]}
{"id": 522191, "image": "COCO_train2014_000000522191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"double decker bus with 109 busaras on it\"."}], "task": "refering", "answer_template": "The \"double decker bus with 109 busaras on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 291, 292, 293, 294, 295], "bbox": [0.217390625, 0.20606557377049178, 0.87859375, 0.8510070257611241], "iscrowd": 0, "area": 96290.74709999998, "rle": {"size": [427, 640], "counts": "WUj15R=>A`0@?B>A`0@?A?B?@?A?B6M101N101O001N101O001N10000O2O00000O2O00000O2O002N2M2O2N2N2M3N2N2M2O00000O100000O01000000O10000O1000O10O1000000O1000000O01000O1000000O1000O10O1000000O100000O010000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O101O000O10001O0O10001O0O10000O2O00000O101O000O101O000O101O001N101O0O2O001O1N101O001N101O001N101O0O2O1O001N100000O010O100O010O100O010O10O0100O1O010O100O010O100O010O10O0100O10O1000000O1000000O10000O1000000O10000000000001O00000000000000001O00000000000000001O000000000000001O00000001O1O1O00100O1O1O1O001O1O1O10O01O1O1O1O001O1O01O01O00001O00001O01O01O0000000000O101O0O10000O10000O101O0O10000O10000O101O0O10000O10000O101O0O1000000000000O2O00000000000000001O000000000O1000001O000000000000001O00000O10000000001O000O1O1O1O1O1O1O2N1O1O1O1D<\\LTaP1"}, "label": "double decker bus with 109 busaras on it"}]}
{"id": 522191, "image": "COCO_train2014_000000522191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus is very close to hitting car\"."}], "task": "refering", "answer_template": "The \"bus is very close to hitting car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258], "bbox": [0.0015, 0.3700936768149883, 0.23689062500000002, 0.8015690866510539], "iscrowd": 0, "area": 19188.791349999996, "rle": {"size": [427, 640], "counts": "bb0>l<l0TOk0UO:F1O1O2N1O2N10001O00001O001O00001O00001O001O00001O00001O001O00001O00010O001O00001O001O00001O00001O001O1O2N2N2N2N1O2N2N2N2N2N1O2N2N0lGQLl6P4RIQLn6o3PISLP7m3oHTLQ7m3lHULT7k3\\HoKO7e7k3[HnKO8f7j3[HnKM;g7h3[HmKM<h7g3ZHnKL=j7e3ZHnKK>k7e3YHmKJ`0m7c3YHmKIa0n7c3WHmKIb0P8a3WHmKHc0Q8a3VHcLj7]3VHcLj7]3VHcLj7^3THcLl7]3THcLl7^3SHcLl7]3THcLl7^3RHcLn7]3RHcLn7]3RHcLn7^3QHbLo7^3PHcLQ8[4000000000001O000000001O1O1O1O001O1O1O1O001O1O1O1OO1O1N2O3M3L4M3L5L3L4M4K4M3L4M4M2N2N3M2N2N2N3M2N2N3L3N2N2N3M2UMd[[6"}, "label": "bus is very close to hitting car"}]}
{"id": 522191, "image": "COCO_train2014_000000522191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a 107 an uaimh bus running on the road\"."}], "task": "refering", "answer_template": "The \"a 107 an uaimh bus running on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258], "bbox": [0.0015, 0.3700936768149883, 0.23689062500000002, 0.8015690866510539], "iscrowd": 0, "area": 19188.791349999996, "rle": {"size": [427, 640], "counts": "bb0>l<l0TOk0UO:F1O1O2N1O2N10001O00001O001O00001O00001O001O00001O00001O001O00001O00010O001O00001O001O00001O00001O001O1O2N2N2N2N1O2N2N2N2N2N1O2N2N0lGQLl6P4RIQLn6o3PISLP7m3oHTLQ7m3lHULT7k3\\HoKO7e7k3[HnKO8f7j3[HnKM;g7h3[HmKM<h7g3ZHnKL=j7e3ZHnKK>k7e3YHmKJ`0m7c3YHmKIa0n7c3WHmKIb0P8a3WHmKHc0Q8a3VHcLj7]3VHcLj7]3VHcLj7^3THcLl7]3THcLl7^3SHcLl7]3THcLl7^3RHcLn7]3RHcLn7]3RHcLn7^3QHbLo7^3PHcLQ8[4000000000001O000000001O1O1O1O001O1O1O1O001O1O1O1OO1O1N2O3M3L4M3L5L3L4M4K4M3L4M4M2N2N3M2N2N2N3M2N2N3L3N2N2N3M2UMd[[6"}, "label": "a 107 an uaimh bus running on the road"}]}
{"id": 415697, "image": "COCO_train2014_000000415697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a circular artwork on the outside of a church and under a clock\"."}], "task": "refering", "answer_template": "The \"a circular artwork on the outside of a church and under a clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 171, 172, 173], "bbox": [0.34753521126760567, 0.31153125, 0.6667136150234741, 0.504], "iscrowd": 0, "area": 13103.2965, "rle": {"size": [640, 426], "counts": "ggl2;cc0e0[O6J3M4L3M4L3M3M4K4M4L3M4M2O2M2N3M2O1N1O2N2O0O2N1O2O0O2N2N101N1O2O1O0O2O001O1O0O2O0000001N1000001O000O101O00000O101O00000O2OO10001O00000O10000000000O10000000000O100000000O100O101N101N101N101N101N101N101N1O2O0O2O0O2O0O2O0O2O1N2N2N3M2N2O1N2N3M2N2N2N2N3L3N2M5K5K5L5J5K5J7^O``h2"}, "label": "a circular artwork on the outside of a church and under a clock"}]}
{"id": 415697, "image": "COCO_train2014_000000415697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lower clock\"."}], "task": "refering", "answer_template": "The \"the lower clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 155, 156, 157, 158, 159, 171, 172, 173], "bbox": [0.34753521126760567, 0.31153125, 0.6667136150234741, 0.504], "iscrowd": 0, "area": 13103.2965, "rle": {"size": [640, 426], "counts": "ggl2;cc0e0[O6J3M4L3M4L3M3M4K4M4L3M4M2O2M2N3M2O1N1O2N2O0O2N1O2O0O2N2N101N1O2O1O0O2O001O1O0O2O0000001N1000001O000O101O00000O101O00000O2OO10001O00000O10000000000O10000000000O100000000O100O101N101N101N101N101N101N101N1O2O0O2O0O2O0O2O0O2O1N2N2N3M2N2O1N2N3M2N2N2N2N3L3N2M5K5K5L5J5K5J7^O``h2"}, "label": "the lower clock"}]}
{"id": 415697, "image": "COCO_train2014_000000415697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clock , at the top of the building\"."}], "task": "refering", "answer_template": "The \"the clock , at the top of the building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 34, 35, 36, 37, 38, 39, 40, 49, 50, 51, 52, 53, 54, 55, 65, 66, 67, 68, 69, 70], "bbox": [0.31396713615023475, 0.042703124999999995, 0.6920892018779343, 0.20224999999999999], "iscrowd": 0, "area": 13794.299949999997, "rle": {"size": [640, 426], "counts": "njc27`c0;E:L4N3L3M3N3L3M3N3L3M3N3L3M3N3L3M3N3L3M3N1N2O1O100000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1001O000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O1O2N1O2N2N2N1O2N2N2N1N3M3M2N3M3L4M2N3M3M3L4M2N3M3M3K4ZOTja2"}, "label": "the clock , at the top of the building"}]}
{"id": 415697, "image": "COCO_train2014_000000415697.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clock face at the top of the tower\"."}], "task": "refering", "answer_template": "The \"the clock face at the top of the tower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 34, 35, 36, 37, 38, 39, 40, 49, 50, 51, 52, 53, 54, 55, 65, 66, 67, 68, 69, 70], "bbox": [0.31396713615023475, 0.042703124999999995, 0.6920892018779343, 0.20224999999999999], "iscrowd": 0, "area": 13794.299949999997, "rle": {"size": [640, 426], "counts": "njc27`c0;E:L4N3L3M3N3L3M3N3L3M3N3L3M3N3L3M3N3L3M3N1N2O1O100000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1001O000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O1O2N1O2N2N2N1O2N2N2N1N3M3M2N3M3L4M2N3M3M3L4M2N3M3M3K4ZOTja2"}, "label": "the clock face at the top of the tower"}]}
{"id": 178131, "image": "COCO_train2014_000000178131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich half on the left\"."}], "task": "refering", "answer_template": "The \"the sandwich half on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 259, 260, 261, 262], "bbox": [0.0438125, 0.19775, 0.461796875, 0.6943750000000001], "iscrowd": 0, "area": 45124.84105000001, "rle": {"size": [480, 640], "counts": "iZ=3j>5K5J6K5K6J5J6M3L4L4L4M3L4L4M3L4L4L4M4K4L4L4M2M2N2O2M2N2N3N1N2N3N1N2O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O0O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O001O001O001O00001O0O2O001O001O00001O001O001N101O00001O001O001O001O0O101O001O001O001O00001O001N101O00001O001O001O001O0O101O001O001O01O01O01O00010O00010O00010O0010O00010O00010O00010O01O01O01O01O01O01O01O010O00010O00010O00010O0010O01O010O001O001O001O001O001O001O0O2O001O0O2M2N3M2N3M2N3M2O2M2N3M2N3M2N3M2N3M2N3M2M4M2M4M2M4L3N3L4M3L5K4L5K4L5K4L5J6K4L5K4L5K4K6K4J7E;D;E<E:EPSQ5"}, "label": "the sandwich half on the left"}]}
{"id": 178131, "image": "COCO_train2014_000000178131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left half of the sandwich\"."}], "task": "refering", "answer_template": "The \"the left half of the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 259, 260, 261, 262], "bbox": [0.0438125, 0.19775, 0.461796875, 0.6943750000000001], "iscrowd": 0, "area": 45124.84105000001, "rle": {"size": [480, 640], "counts": "iZ=3j>5K5J6K5K6J5J6M3L4L4L4M3L4L4M3L4L4L4M4K4L4L4M2M2N2O2M2N2N3N1N2N3N1N2O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O0O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O001O001O001O00001O0O2O001O001O00001O001O001N101O00001O001O001O001O0O101O001O001O001O00001O001N101O00001O001O001O001O0O101O001O001O01O01O01O00010O00010O00010O0010O00010O00010O00010O01O01O01O01O01O01O01O010O00010O00010O00010O0010O01O010O001O001O001O001O001O001O0O2O001O0O2M2N3M2N3M2N3M2O2M2N3M2N3M2N3M2N3M2N3M2M4M2M4M2M4L3N3L4M3L5K4L5K4L5K4L5J6K4L5K4L5K4K6K4J7E;D;E<E:EPSQ5"}, "label": "the left half of the sandwich"}]}
{"id": 178131, "image": "COCO_train2014_000000178131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right half of the sandwich\"."}], "task": "refering", "answer_template": "The \"the right half of the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 357], "bbox": [0.41965624999999995, 0.20897916666666666, 0.8477499999999999, 0.898875], "iscrowd": 0, "area": 64594.291150000005, "rle": {"size": [480, 640], "counts": "b\\n36f>5J6K5J6J6K6I6K5J6J6K6I6K5J6J6K5J7J5J6J6K5J7J5J6K5K5K5QNUKcIQ5Z6TK^IQ5a6SKWIR5h6RKPIS5o6RKiHR5U7SKcHS5[7QK]HT5b7V1O1N3N2N2N2M3N2N2N2N2M3N2N2N2M2O2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N1N3N2N2N2M3N2N2N2M3N1O00O10O0100O10O0100O10O01O100O010O100O10O0100O10O0100O10O0100O10O0100O10O0100O100O002O0O101N100O101N100O101N100O2O0O100O2O0O100O2O0O101N100O101N1O100O2O0O101N100O101N100O101N100O101N100O2O0O100O2O0O100O2N100O2O0O100O2O0O100O2O0O101N100O101N101N2O1N2O0O2O1N2O1N2O0O2N2O1N2O1N101N2O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O0O2O1N2O1N2N101N2O1N2O1N2O0O2O1N2O1N2O0O2O1N2O1N2O0O2O1N2O1N2O8G<E;D<E;B>\\Od0]Od0[Od0]Oc0]OZV]1"}, "label": "the right half of the sandwich"}]}
{"id": 178131, "image": "COCO_train2014_000000178131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of sandwich on the right side of the plate\"."}], "task": "refering", "answer_template": "The \"the half of sandwich on the right side of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 357], "bbox": [0.41965624999999995, 0.20897916666666666, 0.8477499999999999, 0.898875], "iscrowd": 0, "area": 64594.291150000005, "rle": {"size": [480, 640], "counts": "b\\n36f>5J6K5J6J6K6I6K5J6J6K6I6K5J6J6K5J7J5J6J6K5J7J5J6K5K5K5QNUKcIQ5Z6TK^IQ5a6SKWIR5h6RKPIS5o6RKiHR5U7SKcHS5[7QK]HT5b7V1O1N3N2N2N2M3N2N2N2N2M3N2N2N2M2O2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N1N3N2N2N2M3N2N2N2M3N1O00O10O0100O10O0100O10O01O100O010O100O10O0100O10O0100O10O0100O10O0100O10O0100O100O002O0O101N100O101N100O101N100O2O0O100O2O0O100O2O0O101N100O101N1O100O2O0O101N100O101N100O101N100O101N100O2O0O100O2O0O100O2N100O2O0O100O2O0O100O2O0O101N100O101N101N2O1N2O0O2O1N2O1N2O0O2N2O1N2O1N101N2O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O0O2O1N2O1N2N101N2O1N2O1N2O0O2O1N2O1N2O0O2O1N2O1N2O0O2O1N2O1N2O8G<E;D<E;B>\\Od0]Od0[Od0]Oc0]OZV]1"}, "label": "the half of sandwich on the right side of the plate"}]}
{"id": 251868, "image": "COCO_train2014_000000251868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a black scarf sitting on a park bench\"."}], "task": "refering", "answer_template": "The \"a woman with a black scarf sitting on a park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 172, 173, 174, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 285, 286, 287, 288], "bbox": [0.419015625, 0.41477777777777775, 0.6196406250000001, 0.9858888888888888], "iscrowd": 0, "area": 14729.677850000002, "rle": {"size": [360, 640], "counts": "kan21T;3M3N2M3N2M4L3N2M3M3N2M3M4M2M3O100O100O2N100O100O101N1mN^N`Hb1_7`N_Ha1`7aN_H_1`7cN^H^1a7dN^H\\1`7gN^H[1`7gN_HY1`7iN^HX1a7jN^HV1`7mN_HS1`7oN^HR1a7PO^HP1a7RO]HP1`7SO_Hm0`7UO^Hl0a7UO_Hk0`7WO^Hj0`7YO_Hg0`7[O^Hf0a7\\O^He0`7]O^Hd0a7^O^Hb0`7A_H?`7e1O1O1N2O1L4D<K5M3N2O1O1O1SO`KQKa4j4hKQKY4j4PLQKQ4k4XLPKh3l4\\1O10O000O2O0O2O001N110O011N1O100O100O2N100O100O100O2N100O2O0O2N2O0O2O0O2O1N1dKTJ[3n5dLSJ[3m5dLUJZ3m5dLUJX3o5fLSJW3o5hLSJU3P6iLRJT3T6gLnIV3_6^LcI^3V7M3L4L4L4L4M3La0_O?A9G9E;D<E;DiUe2"}, "label": "a woman with a black scarf sitting on a park bench"}]}
{"id": 251868, "image": "COCO_train2014_000000251868.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing a black scarf\"."}], "task": "refering", "answer_template": "The \"a women wearing a black scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 172, 173, 174, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 285, 286, 287, 288], "bbox": [0.419015625, 0.41477777777777775, 0.6196406250000001, 0.9858888888888888], "iscrowd": 0, "area": 14729.677850000002, "rle": {"size": [360, 640], "counts": "kan21T;3M3N2M3N2M4L3N2M3M3N2M3M4M2M3O100O100O2N100O100O101N1mN^N`Hb1_7`N_Ha1`7aN_H_1`7cN^H^1a7dN^H\\1`7gN^H[1`7gN_HY1`7iN^HX1a7jN^HV1`7mN_HS1`7oN^HR1a7PO^HP1a7RO]HP1`7SO_Hm0`7UO^Hl0a7UO_Hk0`7WO^Hj0`7YO_Hg0`7[O^Hf0a7\\O^He0`7]O^Hd0a7^O^Hb0`7A_H?`7e1O1O1N2O1L4D<K5M3N2O1O1O1SO`KQKa4j4hKQKY4j4PLQKQ4k4XLPKh3l4\\1O10O000O2O0O2O001N110O011N1O100O100O2N100O100O100O2N100O2O0O2N2O0O2O0O2O1N1dKTJ[3n5dLSJ[3m5dLUJZ3m5dLUJX3o5fLSJW3o5hLSJU3P6iLRJT3T6gLnIV3_6^LcI^3V7M3L4L4L4L4M3La0_O?A9G9E;D<E;DiUe2"}, "label": "a women wearing a black scarf"}]}
{"id": 161757, "image": "COCO_train2014_000000161757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"animal on the right of the two kissing\"."}], "task": "refering", "answer_template": "The \"animal on the right of the two kissing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 319, 320, 340, 342, 343], "bbox": [0.551078125, 0.4344262295081967, 0.9721875000000001, 0.9902576112412179], "iscrowd": 0, "area": 27735.949399999983, "rle": {"size": [427, 640], "counts": "]^c47S=2M2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1M4L3N2M3M3M3N3L3M300O1O2N100O1O1O101N1O1O100O2N1O100O1O1O2O0O1O1O2O1N101O1O010O001O10O01O010O1O0010O01O10O01O2N2O1N1O21O1N101O1O00001O2O1N1OO2N1O1O2N1O1O1O2N2N1O2N2N1O2N2M2N3M3M3M2M4M3M3M4L4L4L1OO1O100O1O1O1O1O1K5H8G9G9H81N10001O001O00001O001O00001O001O00003M8H7I8H7I6J5K5K5K2NfNPHYMP8e2THXMl7h2XHUMh7j2[HTMe7j2`HRMa7m2cHPM]7n2gHPMY7o2kHmLV7R3mHlLS7R3RIkLn6T3UIiLl6V3XIgLh6W3\\IgLd6X3`IdLa6Z3cIdL]6[3j1O1N200001O00001O00001O00001N10001O01O02N100O2N1O101N1O101N1O101N1O1O2O0O3M2O2M2O2M3M2O2M2O2M2O2M2N2O1N101N2N101N2O0O2O1N1O01M2N3N2M3N1N3L4@?@a0@`0I7L3N3L4M3L3M4M3L4M2M4M4K5L4K6J5L4K6K4K6K4K5L5JRW7"}, "label": "animal on the right of the two kissing"}]}
{"id": 161757, "image": "COCO_train2014_000000161757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan goat facing to the left\"."}], "task": "refering", "answer_template": "The \"tan goat facing to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 319, 320, 340, 342, 343], "bbox": [0.551078125, 0.4344262295081967, 0.9721875000000001, 0.9902576112412179], "iscrowd": 0, "area": 27735.949399999983, "rle": {"size": [427, 640], "counts": "]^c47S=2M2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1M4L3N2M3M3M3N3L3M300O1O2N100O1O1O101N1O1O100O2N1O100O1O1O2O0O1O1O2O1N101O1O010O001O10O01O010O1O0010O01O10O01O2N2O1N1O21O1N101O1O00001O2O1N1OO2N1O1O2N1O1O1O2N2N1O2N2N1O2N2M2N3M3M3M2M4M3M3M4L4L4L1OO1O100O1O1O1O1O1K5H8G9G9H81N10001O001O00001O001O00001O001O00003M8H7I8H7I6J5K5K5K2NfNPHYMP8e2THXMl7h2XHUMh7j2[HTMe7j2`HRMa7m2cHPM]7n2gHPMY7o2kHmLV7R3mHlLS7R3RIkLn6T3UIiLl6V3XIgLh6W3\\IgLd6X3`IdLa6Z3cIdL]6[3j1O1N200001O00001O00001O00001N10001O01O02N100O2N1O101N1O101N1O101N1O1O2O0O3M2O2M2O2M3M2O2M2O2M2O2M2N2O1N101N2N101N2O0O2O1N1O01M2N3N2M3N1N3L4@?@a0@`0I7L3N3L4M3L3M4M3L4M2M4M4K5L4K6J5L4K6K4K6K4K5L5JRW7"}, "label": "tan goat facing to the left"}]}
{"id": 161757, "image": "COCO_train2014_000000161757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an animal with no horns\"."}], "task": "refering", "answer_template": "The \"an animal with no horns\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 258, 259, 260, 261, 281, 282, 283, 304, 305, 306, 327, 329], "bbox": [0.1414375, 0.4092037470725995, 0.55146875, 0.9873536299765808], "iscrowd": 0, "area": 30449.975549999985, "rle": {"size": [427, 640], "counts": "iSV1;P=<D2N2N2N2N2N2N2O1N2N2N2N2N2O1N2Ni0eDnMk9n2K4L4M1N2N1O1O2N101N1O2N1O2N101N1O00000001O0O100000001OSJRL]2n3bMSL^2m3`MUL`2k3^MWLb2i3]MXLc2h3[MZLe2f3ZM[Lf2e3XM]Li2a3WM`Li2`3UMbLk2^3TMcLl2]3QMfLo2\\3mLfLS3^3gLdLY3a3`LaLa3c3ULbLk3c3jKcLV4b3`KcL`4g3PK_LP5n500001O0N2L4K5K5J5O2N2N2O1N2N2N2O1N2N2N2aNPIbLR7]3^1O1O2O0O2N100O2O04L7J5J7I6K5J4M3L4L4M3L3M3N3L3N6I;FO1O1O1O1O1O001O1O1O1O1O1O001O1XOh0[Od0\\Oe0G9F:E:0000O100O10000O100O10000O100O1O1O1N2N2O1N2O1N2O1O1O1O1N101O000O2O001O0O2O1O1O1N2O1O1O001N200O100O10O10O100O100OYOnMUFT2k9oMPFS2o9PNmEQ2S:SNhEo1W:b02O1O0O01O010O1O001O10O10000O2O000O10001N1001O01O0001O00010O00001O01O01O01O010O1O1O100O2N2O1N2N2O2M4L4M3L3M4M3L2N2N101M2M4L4L3M4L6J5L5J6J5Kh]g3"}, "label": "an animal with no horns"}]}
{"id": 161757, "image": "COCO_train2014_000000161757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female goat touching noses with a male goat\"."}], "task": "refering", "answer_template": "The \"a female goat touching noses with a male goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 258, 259, 260, 261, 281, 282, 283, 304, 305, 306, 327, 329], "bbox": [0.1414375, 0.4092037470725995, 0.55146875, 0.9873536299765808], "iscrowd": 0, "area": 30449.975549999985, "rle": {"size": [427, 640], "counts": "iSV1;P=<D2N2N2N2N2N2N2O1N2N2N2N2N2O1N2Ni0eDnMk9n2K4L4M1N2N1O1O2N101N1O2N1O2N101N1O00000001O0O100000001OSJRL]2n3bMSL^2m3`MUL`2k3^MWLb2i3]MXLc2h3[MZLe2f3ZM[Lf2e3XM]Li2a3WM`Li2`3UMbLk2^3TMcLl2]3QMfLo2\\3mLfLS3^3gLdLY3a3`LaLa3c3ULbLk3c3jKcLV4b3`KcL`4g3PK_LP5n500001O0N2L4K5K5J5O2N2N2O1N2N2N2O1N2N2N2aNPIbLR7]3^1O1O2O0O2N100O2O04L7J5J7I6K5J4M3L4L4M3L3M3N3L3N6I;FO1O1O1O1O1O001O1O1O1O1O1O001O1XOh0[Od0\\Oe0G9F:E:0000O100O10000O100O10000O100O1O1O1N2N2O1N2O1N2O1O1O1O1N101O000O2O001O0O2O1O1O1N2O1O1O001N200O100O10O10O100O100OYOnMUFT2k9oMPFS2o9PNmEQ2S:SNhEo1W:b02O1O0O01O010O1O001O10O10000O2O000O10001N1001O01O0001O00010O00001O01O01O01O010O1O1O100O2N2O1N2N2O2M4L4M3L3M4M3L2N2N101M2M4L4L3M4L6J5L5J6J5Kh]g3"}, "label": "a female goat touching noses with a male goat"}]}
{"id": 194531, "image": "COCO_train2014_000000194531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man about to throw a frisbee\"."}], "task": "refering", "answer_template": "The \"a man about to throw a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 138, 139, 140, 141, 142, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.70638, 0.48568, 0.98994, 0.9890399999999999], "iscrowd": 0, "area": 16544.977750000002, "rle": {"size": [375, 500], "counts": "R_Q41d;5K5L4K5L4K6J7J3M1O2O1N2O1N1SFfN]9[1aFjNZ9X1cFnNY9S1dFTOV9m0iFXOS9k1J8H8I7J6I2O0O2N101N101iITL_4m3[KXLe4j3SK\\Ln4d3mJaLR5a3jJaLW5_3fJcLZ5_3bJcL^5^3_JeLa5\\3[JfLe5[3XJgLi5Z3SJhLm5Y3PJjLP6W3lIkLT6\\4004L4M3L5K4L10O01O0010O01M3N1N3N1N2O2M2O1N2O1N3N1N2O100O2O0O100O2O0O100O101O01O001O010O1O001O010O1O7I6K2M0001O001O00001O001O001O001O001O001O00001O001O001O001O0D=B=G9N3N1O2N1N3N1O2M2L5L3M4L3M3M<Do0QOla1"}, "label": "a man about to throw a frisbee"}]}
{"id": 194531, "image": "COCO_train2014_000000194531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue button down shirt by a lake preparing to throw a teal flying disc to a man far away from him\"."}], "task": "refering", "answer_template": "The \"a man in a blue button down shirt by a lake preparing to throw a teal flying disc to a man far away from him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 138, 139, 140, 141, 142, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.70638, 0.48568, 0.98994, 0.9890399999999999], "iscrowd": 0, "area": 16544.977750000002, "rle": {"size": [375, 500], "counts": "R_Q41d;5K5L4K5L4K6J7J3M1O2O1N2O1N1SFfN]9[1aFjNZ9X1cFnNY9S1dFTOV9m0iFXOS9k1J8H8I7J6I2O0O2N101N101iITL_4m3[KXLe4j3SK\\Ln4d3mJaLR5a3jJaLW5_3fJcLZ5_3bJcL^5^3_JeLa5\\3[JfLe5[3XJgLi5Z3SJhLm5Y3PJjLP6W3lIkLT6\\4004L4M3L5K4L10O01O0010O01M3N1N3N1N2O2M2O1N2O1N3N1N2O100O2O0O100O2O0O100O101O01O001O010O1O001O010O1O7I6K2M0001O001O00001O001O001O001O001O001O00001O001O001O001O0D=B=G9N3N1O2N1N3N1O2M2L5L3M4L3M3M<Do0QOla1"}, "label": "a man in a blue button down shirt by a lake preparing to throw a teal flying disc to a man far away from him"}]}
{"id": 268260, "image": "COCO_train2014_000000268260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with open mouth\"."}], "task": "refering", "answer_template": "The \"man with open mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 157, 158, 174, 175, 176, 177, 178, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342, 360, 361], "bbox": [0.500375, 0.25166666666666665, 0.9318125, 0.8915208333333333], "iscrowd": 0, "area": 47887.65645000002, "rle": {"size": [480, 640], "counts": "kXf41m>3L3M3N3M2O1O2N1O2N1O1O2N1O1O1O1O1O1N3M2N2M3N2N2N3ON2N1O2N2N2N1O2O100O01000O10O10O10O10O1000O011O0O101O000M4C<I7J7iNV1M3N3L3M3N2M4M2O1O2N1O2M2O2N1O2hGTLY5m3`J[L^5g3\\J_Lb5b3XJeLf5]3XJeLe5]3YJeLf5\\3YJeLf5]3WJfLg5[3XJfLg5\\3VJfLi5[3UJhLi5Z3TJhLk5Y3SJiLl5X3SJjLk5X3nInLR6R3kIRMT6o2iISMW6m2hIUMV6l2hIVMX6l2eIVMZ6k2cIWM]6j2aIXM]6k2]IZMb6g2WI_Mi6]5nImGU5S8iJRHT5o7iJTHU5m7jJTHV5o7eJTHZ5R8^JPHb5h8N101O001N101O001N101O001O0O1000000001O00000010O0001O00001O0000001O00001O00001O1O1N101N2O0O2O1O1N101N2O1N102N3L3N2M4M2M3N3M5J8I7H=D4K6K3M1O2N2N1O2N9G2N2N2N001O000000001O01O000010O000010O000010O001O010O001O010O000001O0001O0O10000000N1N3M2ZORJYHP6g7d00001O001O1O010O1O010O0010O01O10O010O010O1000O1000[K_Hj1b7TNWIU1h6jNjIf0V6XORJb0n5\\OZJ>f5@bJ9`5DhJ6Z5FmJ5U5GQK5P5HWK3k4I[K2h4K^KOd4ObKKa42dKJ]44iKFZ46lKEV49PLAS4<SL_Oo3>VL^Ok3`0[LZOh3c0\\5M4M2N2M3N2N2MY]d0"}, "label": "man with open mouth"}]}
{"id": 268260, "image": "COCO_train2014_000000268260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue striped shirt\"."}], "task": "refering", "answer_template": "The \"man in blue striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 157, 158, 174, 175, 176, 177, 178, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342, 360, 361], "bbox": [0.500375, 0.25166666666666665, 0.9318125, 0.8915208333333333], "iscrowd": 0, "area": 47887.65645000002, "rle": {"size": [480, 640], "counts": "kXf41m>3L3M3N3M2O1O2N1O2N1O1O2N1O1O1O1O1O1N3M2N2M3N2N2N3ON2N1O2N2N2N1O2O100O01000O10O10O10O10O1000O011O0O101O000M4C<I7J7iNV1M3N3L3M3N2M4M2O1O2N1O2M2O2N1O2hGTLY5m3`J[L^5g3\\J_Lb5b3XJeLf5]3XJeLe5]3YJeLf5\\3YJeLf5]3WJfLg5[3XJfLg5\\3VJfLi5[3UJhLi5Z3TJhLk5Y3SJiLl5X3SJjLk5X3nInLR6R3kIRMT6o2iISMW6m2hIUMV6l2hIVMX6l2eIVMZ6k2cIWM]6j2aIXM]6k2]IZMb6g2WI_Mi6]5nImGU5S8iJRHT5o7iJTHU5m7jJTHV5o7eJTHZ5R8^JPHb5h8N101O001N101O001N101O001O0O1000000001O00000010O0001O00001O0000001O00001O00001O1O1N101N2O0O2O1O1N101N2O1N102N3L3N2M4M2M3N3M5J8I7H=D4K6K3M1O2N2N1O2N9G2N2N2N001O000000001O01O000010O000010O000010O001O010O001O010O000001O0001O0O10000000N1N3M2ZORJYHP6g7d00001O001O1O010O1O010O0010O01O10O010O010O1000O1000[K_Hj1b7TNWIU1h6jNjIf0V6XORJb0n5\\OZJ>f5@bJ9`5DhJ6Z5FmJ5U5GQK5P5HWK3k4I[K2h4K^KOd4ObKKa42dKJ]44iKFZ46lKEV49PLAS4<SL_Oo3>VL^Ok3`0[LZOh3c0\\5M4M2N2M3N2N2MY]d0"}, "label": "man in blue striped shirt"}]}
{"id": 268260, "image": "COCO_train2014_000000268260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person on the left\"."}], "task": "refering", "answer_template": "The \"the person on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.27416666666666667, 0.5545, 0.9865208333333333], "iscrowd": 0, "area": 72926.27560000001, "rle": {"size": [480, 640], "counts": "f>2i>5K5K5K5K6K4K5K5K5K5K5K5M3O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O2M2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O10000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000O10O1000000000000000000000000000000O100O100O100O100O010O10O0100O010OnNYImIg6Q6^IkIb6n4hIiJM5[6k4YJdJB=T6o4[KmJe4Q5`KkJ`4U5dKhJ\\4V5iKgJV4Y5i2N101N101N110O010O1O010O00100O010O1O01000O010000O010O1000O010000O10000O10000O100O10000O10000O10000O10000O100O10001N101O0O101O0O101O0PK^Ja0d5]ObJ>^5BfJ9[5FfJ:[5EfJ:Z5FfJ9[5FfJ:[5EfJ:Z5EgJ:[5EeJ;\\5DdJ<]5BeJ<]5CcJ=]5BdJ>]5AdJ=^5AcJ?^5@bJ?`5@aJ?`5_OaJa0`5^O`Ja0c5\\O^Jd0c5XOaJg0`5TOdJk0^5QOeJo0\\5lNjJR1W5jNnJS1U5hNQKU1P5fNVKX1k4dNZKY1h4bN_KZ1c4_MeL\\2_3]MiL^2[3[MnLh0PL:V7gNQMi0SL7P7iNVMi0TL5j6lNYMi0WL3];HlDOZ;JPEMU;MUEKo:OiZU4"}, "label": "the person on the left"}]}
{"id": 268260, "image": "COCO_train2014_000000268260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman leaning toward a man\"."}], "task": "refering", "answer_template": "The \"a woman leaning toward a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.27416666666666667, 0.5545, 0.9865208333333333], "iscrowd": 0, "area": 72926.27560000001, "rle": {"size": [480, 640], "counts": "f>2i>5K5K5K5K6K4K5K5K5K5K5K5M3O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O2M2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O10000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000O10O1000000000000000000000000000000O100O100O100O100O010O10O0100O010OnNYImIg6Q6^IkIb6n4hIiJM5[6k4YJdJB=T6o4[KmJe4Q5`KkJ`4U5dKhJ\\4V5iKgJV4Y5i2N101N101N110O010O1O010O00100O010O1O01000O010000O010O1000O010000O10000O10000O100O10000O10000O10000O10000O100O10001N101O0O101O0O101O0PK^Ja0d5]ObJ>^5BfJ9[5FfJ:[5EfJ:Z5FfJ9[5FfJ:[5EfJ:Z5EgJ:[5EeJ;\\5DdJ<]5BeJ<]5CcJ=]5BdJ>]5AdJ=^5AcJ?^5@bJ?`5@aJ?`5_OaJa0`5^O`Ja0c5\\O^Jd0c5XOaJg0`5TOdJk0^5QOeJo0\\5lNjJR1W5jNnJS1U5hNQKU1P5fNVKX1k4dNZKY1h4bN_KZ1c4_MeL\\2_3]MiL^2[3[MnLh0PL:V7gNQMi0SL7P7iNVMi0TL5j6lNYMi0WL3];HlDOZ;JPEMU;MUEKo:OiZU4"}, "label": "a woman leaning toward a man"}]}
{"id": 481257, "image": "COCO_train2014_000000481257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of the white car parked to the left of the front of the bus\"."}], "task": "refering", "answer_template": "The \"the front of the white car parked to the left of the front of the bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351], "bbox": [0.00146875, 0.670958904109589, 0.3031875, 0.9871004566210047], "iscrowd": 0, "area": 23779.667200000007, "rle": {"size": [438, 640], "counts": "nf0h0n<a1_Na1_N>B0000000O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000001O001O1O1O001O1O1O001O1O1O001O1_OlE^MT:`2VFXMk9f2d0M3M2O2N2N2O2M4M4K4L5L4K4I8H6I7J6J6I7J6I]]n5"}, "label": "the front of the white car parked to the left of the front of the bus"}]}
{"id": 481257, "image": "COCO_train2014_000000481257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white vehicle close to the front of the bus\"."}], "task": "refering", "answer_template": "The \"the white vehicle close to the front of the bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351], "bbox": [0.00146875, 0.670958904109589, 0.3031875, 0.9871004566210047], "iscrowd": 0, "area": 23779.667200000007, "rle": {"size": [438, 640], "counts": "nf0h0n<a1_Na1_N>B0000000O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000001O001O1O1O001O1O1O001O1O1O001O1_OlE^MT:`2VFXMk9f2d0M3M2O2N2N2O2M4M4K4L5L4K4I8H6I7J6J6I7J6I]]n5"}, "label": "the white vehicle close to the front of the bus"}]}
{"id": 350191, "image": "COCO_train2014_000000350191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"very back right , leg of a person\"."}], "task": "refering", "answer_template": "The \"very back right , leg of a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 25, 26, 27, 40, 41, 42, 56, 57, 58, 72, 73, 87, 88], "bbox": [0.6998594847775175, 0.0215, 0.9408899297423887, 0.24784375], "iscrowd": 0, "area": 6558.2823, "rle": {"size": [640, 427], "counts": "omj59Zc0=C=C=J600O010O10000O100O1000O2O001O1O1O007I00000O10000000001O00000O14L7I6J7l_OoMk=X2kAkMU>\\2bAfM^>a2XAbMh>e2n@^MR?h2e@[M[?l2[@WMe?b30000001O00000000000000004L4L3M4L4L4L3M4L4L3M4L4L4L3M4L4L6J5K6J6J5K6J6J2N1O1O001O1O1O1O001O100O1O1O001O1M3M3M2N3M3M`_?"}, "label": "very back right , leg of a person"}]}
{"id": 350191, "image": "COCO_train2014_000000350191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leg of a man in tan pants sitting on a bench\"."}], "task": "refering", "answer_template": "The \"the leg of a man in tan pants sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 25, 26, 27, 40, 41, 42, 56, 57, 58, 72, 73, 87, 88], "bbox": [0.6998594847775175, 0.0215, 0.9408899297423887, 0.24784375], "iscrowd": 0, "area": 6558.2823, "rle": {"size": [640, 427], "counts": "omj59Zc0=C=C=J600O010O10000O100O1000O2O001O1O1O007I00000O10000000001O00000O14L7I6J7l_OoMk=X2kAkMU>\\2bAfM^>a2XAbMh>e2n@^MR?h2e@[M[?l2[@WMe?b30000001O00000000000000004L4L3M4L4L4L3M4L4L3M4L4L4L3M4L4L6J5K6J6J5K6J6J2N1O1O001O1O1O1O001O100O1O1O001O1M3M3M2N3M3M`_?"}, "label": "the leg of a man in tan pants sitting on a bench"}]}
{"id": 473072, "image": "COCO_train2014_000000473072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half top of sandwich bread with cut edge touching the fried potatoes\"."}], "task": "refering", "answer_template": "The \"half top of sandwich bread with cut edge touching the fried potatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 336, 337], "bbox": [0.41046875, 0.3017916666666667, 0.7297343749999999, 0.8333333333333334], "iscrowd": 0, "area": 28813.419, "rle": {"size": [480, 640], "counts": "Z`k38b><D=C<D=C<D<E<C<D<I8H7I7I3M2N2N3M2N2M3N2N2N3M2N2N2N2N3M2M3N2N2N3M2N2N2N2N3M2N2M3N2N2O2O000000010O00000000O1O1O1O1O1O1O100O1O1O2N1O1O100O1O1O1O1O1O1O100O1O1O1O2N1O100O1O1O100O1O100O1O100O1O100O20O000000000000000000000000001OO2O0O1O1O1O100O1O1O1O100O1O1O1O101N1O1O2N101N1O2N1O2O0O2N1O2N101N2N1O2N101N1O2N1O2O1N3M6J5K3M2N2N2M3L4M2M3M3N2M3N2M3M3N2M2N3N2M3N2M3M3N0O2O0O2N101N101N1O2O1N2N3N2M2O2M3M3N2M3N2M2N2K5JVi`2"}, "label": "half top of sandwich bread with cut edge touching the fried potatoes"}]}
{"id": 473072, "image": "COCO_train2014_000000473072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of sandwich without the cross facing the fries\"."}], "task": "refering", "answer_template": "The \"the piece of sandwich without the cross facing the fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 336, 337], "bbox": [0.41046875, 0.3017916666666667, 0.7297343749999999, 0.8333333333333334], "iscrowd": 0, "area": 28813.419, "rle": {"size": [480, 640], "counts": "Z`k38b><D=C<D=C<D<E<C<D<I8H7I7I3M2N2N3M2N2M3N2N2N3M2N2N2N2N3M2M3N2N2N3M2N2N2N2N3M2N2M3N2N2O2O000000010O00000000O1O1O1O1O1O1O100O1O1O2N1O1O100O1O1O1O1O1O1O100O1O1O1O2N1O100O1O1O100O1O100O1O100O1O100O20O000000000000000000000000001OO2O0O1O1O1O100O1O1O1O100O1O1O1O101N1O1O2N101N1O2N1O2O0O2N1O2N101N2N1O2N101N1O2N1O2O1N3M6J5K3M2N2N2M3L4M2M3M3N2M3N2M3M3N2M2N3N2M3N2M3M3N0O2O0O2N101N101N1O2O1N2N3N2M2O2M3M3N2M3N2M2N2K5JVi`2"}, "label": "the piece of sandwich without the cross facing the fries"}]}
{"id": 473072, "image": "COCO_train2014_000000473072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half a sandwich facing away from french fries\"."}], "task": "refering", "answer_template": "The \"half a sandwich facing away from french fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 258, 259, 260, 261, 283, 284], "bbox": [0.10795312500000001, 0.20158333333333334, 0.39471875, 0.7618541666666666], "iscrowd": 0, "area": 33191.33725, "rle": {"size": [480, 640], "counts": "c_P12m>5J5L5K4K6K4K6K4L4K6K4L5K4L5K4L5K4L4M4K4L3M3M2N3M2N3M2O2M2N2N3M2N3M2N3M2O2M2N3M2N3N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1N3N1O2N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1O1N2O1O1O001O1O1O1O1O1O010O1O1O1O1O1O001O1O1O0010O01O1O001O001O001O1O010O001O001O1O001O00100O001O001O00100O001O001O00100O001O1O1O100O100O10000O100O100O100O10000O100O100O100O01N2M2O2N1N3N1N3N1H9D<G8K6J5J:G:A?^Ob0bLdE`1o:mMfE_1[<\\Oe0ZOe0\\OgVe5"}, "label": "half a sandwich facing away from french fries"}]}
{"id": 473072, "image": "COCO_train2014_000000473072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a grilled turkey sandwich touching a green pickle\"."}], "task": "refering", "answer_template": "The \"half of a grilled turkey sandwich touching a green pickle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 258, 259, 260, 261, 283, 284], "bbox": [0.10795312500000001, 0.20158333333333334, 0.39471875, 0.7618541666666666], "iscrowd": 0, "area": 33191.33725, "rle": {"size": [480, 640], "counts": "c_P12m>5J5L5K4K6K4K6K4L4K6K4L5K4L5K4L5K4L4M4K4L3M3M2N3M2N3M2O2M2N2N3M2N3M2N3M2O2M2N3M2N3N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1N3N1O2N1O2M2O2N1O2M2O2N1O2N1N3N1O2N1O1N2O1O1O001O1O1O1O1O1O010O1O1O1O1O1O001O1O1O0010O01O1O001O001O001O1O010O001O001O1O001O00100O001O001O00100O001O001O00100O001O1O1O100O100O10000O100O100O100O10000O100O100O100O01N2M2O2N1N3N1N3N1H9D<G8K6J5J:G:A?^Ob0bLdE`1o:mMfE_1[<\\Oe0ZOe0\\OgVe5"}, "label": "half of a grilled turkey sandwich touching a green pickle"}]}
{"id": 440310, "image": "COCO_train2014_000000440310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an electric wheelchair\"."}], "task": "refering", "answer_template": "The \"an electric wheelchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 112, 113, 135, 136, 151, 152, 153, 157, 158, 159, 174, 175, 176, 180, 181, 182, 198, 202, 203, 204, 220, 221, 224, 225, 226, 227, 242, 243, 246, 247, 248, 249, 250, 265, 268, 269, 270, 271, 272, 273, 288, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 361, 362, 363], "bbox": [0.531609375, 0.19073170731707317, 0.9474218750000001, 0.8376547842401502], "iscrowd": 0, "area": 28848.225800000004, "rle": {"size": [533, 640], "counts": "mYa51``04M2N3L4M2M4M3L3N3M3L4MJ53N1N3N2O1N101OCh@@W?>SA[Ol>e0`02O1O1O1O110aMSOUEm0l:WOoDi0R;YOlDf0U;\\OhDe0X;]OeDd0[;_OaDa0`;@_D`0a;@^Da0a;@^D`0c;A\\D>e;BZD?f;A[D=f;D[D9f;G[D8e;H\\D6e;K[D4e;L\\D2j1ZOV6d0QHOi1@T6c0SHJk1DQ6b0THHm1Fo5b0THFn1Jm5a0gG\\OlN7b3Mi5a0gG_OkN1h3Oa5g0hG^OkNLm30[5k0jGDn2AX5k0hGCS3BV5j0dGDX3CU5h0aGD]3EQ5g0`GDa3EP5g0]GBe3Hn4f0\\GAh3Im4e0ZGAk3Jk4e0ZG_Ol3Nj4b0ZG_Ok31k4`0\\G\\Oh36m4=\\GZOh3:U54SG_Oh3?U52SG^Og3b0V50SG\\Og3f0V5NSG[Og3h0V5MSGZOf3k0V5MbK5^4KaK6_4J^FXOl4P1f4H^FXOj4R1g4KYK6g4IYKJUKDb9`0ZKLUKDa9?ZKMWKB`9`0YKNXKB^9`0ZKNZKA\\9`0YK1\\K]O\\9`0XK6\\KVO_9c0TK9W5HgJ;X5DgJ>Y5AfJb0Y5^OeJd0[5[OdJh0Z5YOcJj0]5UObJn0]5ROaJQ1^5nNaJT1^5mN`JV1_5iN_JY1b5fN]J\\1c5bN\\Jb1c5]NYJh1f5XNUJo1j5PNnIY2R6fMoI[2P6cMQJ^2n5bMRJ`2m5^MTJc2l5\\MUJd2k5ZMVJg2k900000000M201O100O1O00100O1O1O01aCdMf:]2UEjMh:W2UEnMh:S2TERNj:o1SEVNj:k1aDWM5R1W;i1_D^M4l0[;g1\\DdM5g0^;]2\\DeMc;h3N1N2O2N1O1O2M2O1O1O100O01000000O0100000O0100000O0100O100O00\\NoDeMR;[2oDcMQ;^2PE`MP;a2QE^Mo:b2SE[Mm:f2TEXMl:j2TERMn:o2TElLm:V3UEdLn:]3TE^Lm:c3VEWLm:j3UEiKT;Y4c01N2O1M3M3M3K5K5O1O100O100OO2F91O101N4L5L3L3M2O1N1O2M2N3M2O1N3M2O1kLRJaKP6_4SJ\\Ko5d4SJWKY2ZMV1_7cLSKX2aMS1\\7gLoJW2hMP1Z7hLlJY2mMm0W7kLhJZ2TNi0T7mLfJ[2WNh0S7nLcJ[2\\Nf0Q7oLaJ\\2_Nd0R7PM[J_2cNa0R7QMXJ_2gN?R7RMUJ`2kN=P7TMRJ`2oNhM[On1d7jMoI]2XOjMXOl1c7mMjI]2^OgMYOm1_7oMhI\\2CgMYOk1]7UNaI[2JdMYOk1]7[NYIY21^M[On1[7_NRIX29XM\\OP2Z7\\2[IbK\\OR2Z7[2\\I_K\\OU2Y7\\2QJcMP6]2PJcMP6]2PJbMQ6h1VHdLi1d1R6d1YHgLe1d1S6b1]HeLc1h1Q6_1aHfLa1h1P6_1dHdL`1k1m5^1RK`No4]1TKaNo4[1UKcNl4Z1XKcNj4Y1[KdNi4X1[KbNj4\\1XK_Nn4^1UK\\NP5b1RKZNS5OUG=m3]OS5K^Gc0l3ROl4OeGk0d>M2N1N3N2M2OO100000O100000001N1N2N2O4KY`a0"}, "label": "an electric wheelchair"}]}
{"id": 440310, "image": "COCO_train2014_000000440310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorized black wheelchair\"."}], "task": "refering", "answer_template": "The \"a motorized black wheelchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 112, 113, 135, 136, 151, 152, 153, 157, 158, 159, 174, 175, 176, 180, 181, 182, 198, 202, 203, 204, 220, 221, 224, 225, 226, 227, 242, 243, 246, 247, 248, 249, 250, 265, 268, 269, 270, 271, 272, 273, 288, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 361, 362, 363], "bbox": [0.531609375, 0.19073170731707317, 0.9474218750000001, 0.8376547842401502], "iscrowd": 0, "area": 28848.225800000004, "rle": {"size": [533, 640], "counts": "mYa51``04M2N3L4M2M4M3L3N3M3L4MJ53N1N3N2O1N101OCh@@W?>SA[Ol>e0`02O1O1O1O110aMSOUEm0l:WOoDi0R;YOlDf0U;\\OhDe0X;]OeDd0[;_OaDa0`;@_D`0a;@^Da0a;@^D`0c;A\\D>e;BZD?f;A[D=f;D[D9f;G[D8e;H\\D6e;K[D4e;L\\D2j1ZOV6d0QHOi1@T6c0SHJk1DQ6b0THHm1Fo5b0THFn1Jm5a0gG\\OlN7b3Mi5a0gG_OkN1h3Oa5g0hG^OkNLm30[5k0jGDn2AX5k0hGCS3BV5j0dGDX3CU5h0aGD]3EQ5g0`GDa3EP5g0]GBe3Hn4f0\\GAh3Im4e0ZGAk3Jk4e0ZG_Ol3Nj4b0ZG_Ok31k4`0\\G\\Oh36m4=\\GZOh3:U54SG_Oh3?U52SG^Og3b0V50SG\\Og3f0V5NSG[Og3h0V5MSGZOf3k0V5MbK5^4KaK6_4J^FXOl4P1f4H^FXOj4R1g4KYK6g4IYKJUKDb9`0ZKLUKDa9?ZKMWKB`9`0YKNXKB^9`0ZKNZKA\\9`0YK1\\K]O\\9`0XK6\\KVO_9c0TK9W5HgJ;X5DgJ>Y5AfJb0Y5^OeJd0[5[OdJh0Z5YOcJj0]5UObJn0]5ROaJQ1^5nNaJT1^5mN`JV1_5iN_JY1b5fN]J\\1c5bN\\Jb1c5]NYJh1f5XNUJo1j5PNnIY2R6fMoI[2P6cMQJ^2n5bMRJ`2m5^MTJc2l5\\MUJd2k5ZMVJg2k900000000M201O100O1O00100O1O1O01aCdMf:]2UEjMh:W2UEnMh:S2TERNj:o1SEVNj:k1aDWM5R1W;i1_D^M4l0[;g1\\DdM5g0^;]2\\DeMc;h3N1N2O2N1O1O2M2O1O1O100O01000000O0100000O0100000O0100O100O00\\NoDeMR;[2oDcMQ;^2PE`MP;a2QE^Mo:b2SE[Mm:f2TEXMl:j2TERMn:o2TElLm:V3UEdLn:]3TE^Lm:c3VEWLm:j3UEiKT;Y4c01N2O1M3M3M3K5K5O1O100O100OO2F91O101N4L5L3L3M2O1N1O2M2N3M2O1N3M2O1kLRJaKP6_4SJ\\Ko5d4SJWKY2ZMV1_7cLSKX2aMS1\\7gLoJW2hMP1Z7hLlJY2mMm0W7kLhJZ2TNi0T7mLfJ[2WNh0S7nLcJ[2\\Nf0Q7oLaJ\\2_Nd0R7PM[J_2cNa0R7QMXJ_2gN?R7RMUJ`2kN=P7TMRJ`2oNhM[On1d7jMoI]2XOjMXOl1c7mMjI]2^OgMYOm1_7oMhI\\2CgMYOk1]7UNaI[2JdMYOk1]7[NYIY21^M[On1[7_NRIX29XM\\OP2Z7\\2[IbK\\OR2Z7[2\\I_K\\OU2Y7\\2QJcMP6]2PJcMP6]2PJbMQ6h1VHdLi1d1R6d1YHgLe1d1S6b1]HeLc1h1Q6_1aHfLa1h1P6_1dHdL`1k1m5^1RK`No4]1TKaNo4[1UKcNl4Z1XKcNj4Y1[KdNi4X1[KbNj4\\1XK_Nn4^1UK\\NP5b1RKZNS5OUG=m3]OS5K^Gc0l3ROl4OeGk0d>M2N1N3N2M2OO100000O100000001N1N2N2O4KY`a0"}, "label": "a motorized black wheelchair"}]}
{"id": 440310, "image": "COCO_train2014_000000440310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table full of pastries\"."}], "task": "refering", "answer_template": "The \"a table full of pastries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423], "bbox": [0.00375, 0.2809005628517824, 0.5539531249999999, 0.9887617260787992], "iscrowd": 0, "area": 96514.67360000001, "rle": {"size": [533, 640], "counts": "bZ1b0R`0m5SJd0\\O1O1O1O1O100O1O1O100000000000000O1000001O0000000000000O100000000000000O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O2N100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O2N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O2N1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O2N1O1O1000000000000000000000000001O000000000000000000000000000000000000001O00000001O0000000000000000000000000001O00000000000N2M3K5L4L4L4L4L4L4L4K5L4L5K4L4L4L4L4K5L4L4L4L4L4L4L4L4K5L4L4L4L5K4L4L4K5L4L4L4L4L4L4L4K5L4L4L4L4L4L4L5J5N200001O00001O00001O00N2L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L4L4L4Llfd4"}, "label": "a table full of pastries"}]}
{"id": 440310, "image": "COCO_train2014_000000440310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table tops with dessert on them\"."}], "task": "refering", "answer_template": "The \"the table tops with dessert on them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423], "bbox": [0.00375, 0.2809005628517824, 0.5539531249999999, 0.9887617260787992], "iscrowd": 0, "area": 96514.67360000001, "rle": {"size": [533, 640], "counts": "bZ1b0R`0m5SJd0\\O1O1O1O1O100O1O1O100000000000000O1000001O0000000000000O100000000000000O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O2N100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O2N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O2N1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O2N1O1O1000000000000000000000000001O000000000000000000000000000000000000001O00000001O0000000000000000000000000001O00000000000N2M3K5L4L4L4L4L4L4L4K5L4L5K4L4L4L4L4K5L4L4L4L4L4L4L4L4K5L4L4L4L5K4L4L4K5L4L4L4L4L4L4L4K5L4L4L4L4L4L4L5J5N200001O00001O00001O00N2L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4L4L4L4L4L4L4Llfd4"}, "label": "the table tops with dessert on them"}]}
{"id": 219127, "image": "COCO_train2014_000000219127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laamb with a black tail eats it ' s meal\"."}], "task": "refering", "answer_template": "The \"a laamb with a black tail eats it ' s meal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 264, 266, 267, 268, 287, 290, 291, 313, 314, 336, 337], "bbox": [0.43618750000000006, 0.18763466042154567, 0.7705, 0.9897892271662763], "iscrowd": 0, "area": 40526.30965, "rle": {"size": [427, 640], "counts": "Vbd33S=:F:E;F9G9G8H9G9G8H8H8H7J7J6J6J5K6J6K5J5K6J6J6J5K6K5aIUJV5o5fJSJY5Q6aJRJ^5R6]JPJa5U6YJnIf5U6VJmIi5W6RJkIm5j6O1O1N2O1O001O1O0O2O001O00O1O10O01O2N2N1@a0_Oa0J5L5K5K4N3O1000O0100O010O10O10O10O0100O10O01000O0100O0101O0O101O000O2O001N10001N10001N10002N6I6K6J5J6K6J5J;F:F:E;F:F9G:E7J2N2M3N1O2M3N2N2N1O2N2N2N2N1O2N2N2N00O1000001O00000000000000000O1000000]LZOaJf0U5HgJ8P55lJKk4b0QK^Of4P1UKPOh4W1TKiNf4a1VK_Nf4j1UKVNj4o1RKQNl4U2PKkMo4X2nJiMQ5Y2nJgMP5\\2oJdMP5^2oJcMo4_2PKbMm4a2RK`Mk4c2TK]Mi4g2VKZMf4j2YKWMb4l5L4L4K5L4K5L3M4G9J6N2N2N2N2N2M3N1O2N2N2N2N2N2N2M3N2N1O2N2N2N2N2N2M3N2N2N1O2N2N5J7H7J6F]Um1"}, "label": "a laamb with a black tail eats it ' s meal"}]}
{"id": 219127, "image": "COCO_train2014_000000219127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back end of the animal with the biggest blackest tail\"."}], "task": "refering", "answer_template": "The \"back end of the animal with the biggest blackest tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 264, 266, 267, 268, 287, 290, 291, 313, 314, 336, 337], "bbox": [0.43618750000000006, 0.18763466042154567, 0.7705, 0.9897892271662763], "iscrowd": 0, "area": 40526.30965, "rle": {"size": [427, 640], "counts": "Vbd33S=:F:E;F9G9G8H9G9G8H8H8H7J7J6J6J5K6J6K5J5K6J6J6J5K6K5aIUJV5o5fJSJY5Q6aJRJ^5R6]JPJa5U6YJnIf5U6VJmIi5W6RJkIm5j6O1O1N2O1O001O1O0O2O001O00O1O10O01O2N2N1@a0_Oa0J5L5K5K4N3O1000O0100O010O10O10O10O0100O10O01000O0100O0101O0O101O000O2O001N10001N10001N10002N6I6K6J5J6K6J5J;F:F:E;F:F9G:E7J2N2M3N1O2M3N2N2N1O2N2N2N2N1O2N2N2N00O1000001O00000000000000000O1000000]LZOaJf0U5HgJ8P55lJKk4b0QK^Of4P1UKPOh4W1TKiNf4a1VK_Nf4j1UKVNj4o1RKQNl4U2PKkMo4X2nJiMQ5Y2nJgMP5\\2oJdMP5^2oJcMo4_2PKbMm4a2RK`Mk4c2TK]Mi4g2VKZMf4j2YKWMb4l5L4L4K5L4K5L3M4G9J6N2N2N2N2N2M3N1O2N2N2N2N2N2N2M3N2N1O2N2N2N2N2N2M3N2N2N1O2N2N5J7H7J6F]Um1"}, "label": "back end of the animal with the biggest blackest tail"}]}
{"id": 219127, "image": "COCO_train2014_000000219127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gray sheep farthest on left\"."}], "task": "refering", "answer_template": "The \"gray sheep farthest on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 47, 48, 49, 50, 51, 52, 53, 54, 71, 72, 94, 95, 117, 139, 140, 162, 163, 186, 209], "bbox": [0.067140625, 0.0002107728337236534, 0.5000312499999999, 0.61903981264637], "iscrowd": 0, "area": 18834.346550000002, "rle": {"size": [427, 640], "counts": "doa05l<;K5N2M3N2N2M3N2N1O2bGYOT4i0XK4^4NZKc0[4^ObKo0S4SOjKh1[3ZNcLj1Z3XNbLk1]3VNaLl1^3VN_Lk1a3VN\\Lm1c3UNXLo1h3SNPLS2o3RNgKT2Y4QN\\KT2d4QNRKU2n4^NYJh1f5o20O0100O10jMcJ^L\\5P3\\KjLe4d2QLXMn3a2_LZMa3c2fLYMY3e2nLVMS3g2UMTMj2j2PN_LP2^3o3N3N1O2O1N2O0O2O1N2O1N2O1N101N2O1N2O1O001O1O1O001O001O1O001O1O001O001O1O001O010O100O10O10O100O100O10O0100O100O100O101N100O10000O10000O2O000O100O10000O2O000O10000O100O100000000000000000001O0001O000000000000000O1000000000000000000000000001N1000001O00000O2O0000001O0O10001O0000000O2O0000001O000O101O0000001N1000001O00001N10001O001O000O2O001O00001O001O0000001O0000000000001O0000000000001O0000000000001O00000000001O00000000001O000000001O00000000001O2N101N2N2N2N2N1O2N2N2N2N2N2N2NX\\U4"}, "label": "gray sheep farthest on left"}]}
{"id": 219127, "image": "COCO_train2014_000000219127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep closest to the wall on the end\"."}], "task": "refering", "answer_template": "The \"the sheep closest to the wall on the end\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 47, 48, 49, 50, 51, 52, 53, 54, 71, 72, 94, 95, 117, 139, 140, 162, 163, 186, 209], "bbox": [0.067140625, 0.0002107728337236534, 0.5000312499999999, 0.61903981264637], "iscrowd": 0, "area": 18834.346550000002, "rle": {"size": [427, 640], "counts": "doa05l<;K5N2M3N2N2M3N2N1O2bGYOT4i0XK4^4NZKc0[4^ObKo0S4SOjKh1[3ZNcLj1Z3XNbLk1]3VNaLl1^3VN_Lk1a3VN\\Lm1c3UNXLo1h3SNPLS2o3RNgKT2Y4QN\\KT2d4QNRKU2n4^NYJh1f5o20O0100O10jMcJ^L\\5P3\\KjLe4d2QLXMn3a2_LZMa3c2fLYMY3e2nLVMS3g2UMTMj2j2PN_LP2^3o3N3N1O2O1N2O0O2O1N2O1N2O1N101N2O1N2O1O001O1O1O001O001O1O001O1O001O001O1O001O010O100O10O10O100O100O10O0100O100O100O101N100O10000O10000O2O000O100O10000O2O000O10000O100O100000000000000000001O0001O000000000000000O1000000000000000000000000001N1000001O00000O2O0000001O0O10001O0000000O2O0000001O000O101O0000001N1000001O00001N10001O001O000O2O001O00001O001O0000001O0000000000001O0000000000001O0000000000001O00000000001O00000000001O000000001O00000000001O2N101N2N2N2N2N1O2N2N2N2N2N2N2NX\\U4"}, "label": "the sheep closest to the wall on the end"}]}
{"id": 219127, "image": "COCO_train2014_000000219127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep from second left\"."}], "task": "refering", "answer_template": "The \"sheep from second left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 211, 212, 213, 235, 236, 258, 259, 282], "bbox": [0.11765624999999999, 0.1311943793911007, 0.58253125, 0.8344028103044495], "iscrowd": 0, "area": 37740.544700000006, "rle": {"size": [427, 640], "counts": "k^o01R=`0A`0_O`0A?A?@`0A?@9H7O001O1O0O2O001O1O0O2O1O001O001N2O001O1O0O2O001O1OM2E<D<E;D=N13M3N2M3N2M3M3N2M3N2M4SHWLQ6l3kIWLT6k3iIXLT6k3hIYLV6i3fI[LW6h3eI[LY6h3dI[LZ6g3bI]L[6f3aI^L]6d3`I^L^6e3^I_L_6d3]I`La6b3[IbLb6a3[IaLd6a3XIcLe6T5M3N1N3N2M3M2O2M3N2M2O2M3M3N1N3N2M3M2O2N2N2N1O2N2N2N1O2N2N2NmM]K_Ka4`4cK`K[4^4iKbKU4]4oKbKo3\\4ULcKj3[4ZLeKd3Y4aLfK]3X4gLgKX3W4lLfKU3Y4oLcKR3[4RMbKo2\\4UMaKl2]4XM`Ki2^4\\M^Ke2a4^M]Kb2a4bM\\K_2b4fMZK[2d4iMYKW2f4mMWKT2g4PNVKQ2i4SNSKn1k4VNRKk1l4YNQKh1m4]NoJd1o4`NoJ`1P5cNmJ^1Q5fNlJ[1R5`3N2N2N2O1N2N2N2N101N2N2N2N2N2O1N2N2O10000000000O10O10000000O10000000000O10O10000000O10000000000O01000000000O1000000000O010000000000O1000000000O010000000000O10000000O10O10000000000O100000ON3I7J6J6J6I7N2N2O2M2N2N2O1N2N3M2O1N2N2N2O2M2N2O1N2N2N2O2M2N2O1O100O2O000O100O101O0O2O1N2O001N2O1N2O1O0O2O1N2O1O0O2O1O1N2O0O101O0O101N10001N100O2O000O101N10001N10`e_3"}, "label": "sheep from second left"}]}
{"id": 219127, "image": "COCO_train2014_000000219127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle sheep with black legs\"."}], "task": "refering", "answer_template": "The \"the middle sheep with black legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 211, 212, 213, 235, 236, 258, 259, 282], "bbox": [0.11765624999999999, 0.1311943793911007, 0.58253125, 0.8344028103044495], "iscrowd": 0, "area": 37740.544700000006, "rle": {"size": [427, 640], "counts": "k^o01R=`0A`0_O`0A?A?@`0A?@9H7O001O1O0O2O001O1O0O2O1O001O001N2O001O1O0O2O001O1OM2E<D<E;D=N13M3N2M3N2M3M3N2M3N2M4SHWLQ6l3kIWLT6k3iIXLT6k3hIYLV6i3fI[LW6h3eI[LY6h3dI[LZ6g3bI]L[6f3aI^L]6d3`I^L^6e3^I_L_6d3]I`La6b3[IbLb6a3[IaLd6a3XIcLe6T5M3N1N3N2M3M2O2M3N2M2O2M3M3N1N3N2M3M2O2N2N2N1O2N2N2N1O2N2N2NmM]K_Ka4`4cK`K[4^4iKbKU4]4oKbKo3\\4ULcKj3[4ZLeKd3Y4aLfK]3X4gLgKX3W4lLfKU3Y4oLcKR3[4RMbKo2\\4UMaKl2]4XM`Ki2^4\\M^Ke2a4^M]Kb2a4bM\\K_2b4fMZK[2d4iMYKW2f4mMWKT2g4PNVKQ2i4SNSKn1k4VNRKk1l4YNQKh1m4]NoJd1o4`NoJ`1P5cNmJ^1Q5fNlJ[1R5`3N2N2N2O1N2N2N2N101N2N2N2N2N2O1N2N2O10000000000O10O10000000O10000000000O10O10000000O10000000000O01000000000O1000000000O010000000000O1000000000O010000000000O10000000O10O10000000000O100000ON3I7J6J6J6I7N2N2O2M2N2N2O1N2N3M2O1N2N2N2O2M2N2O1N2N2N2O2M2N2O1O100O2O000O100O101O0O2O1N2O001N2O1N2O1O0O2O1N2O1O0O2O1O1N2O0O101O0O101N10001N100O2O000O101N10001N10`e_3"}, "label": "the middle sheep with black legs"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the entire picture\"."}], "task": "refering", "answer_template": "The \"the entire picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.004313725490196079, 0.9974019607843136, 0.9861764705882352], "iscrowd": 0, "area": 365835.4296, "rle": {"size": [612, 612], "counts": "P=l5`1h;kNU100000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000UV1"}, "label": "the entire picture"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fruit bagel dessert\"."}], "task": "refering", "answer_template": "The \"a fruit bagel dessert\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.004313725490196079, 0.9974019607843136, 0.9861764705882352], "iscrowd": 0, "area": 365835.4296, "rle": {"size": [612, 612], "counts": "P=l5`1h;kNU100000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000UV1"}, "label": "a fruit bagel dessert"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee kept near the breakfast plate\"."}], "task": "refering", "answer_template": "The \"a cup of coffee kept near the breakfast plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149], "bbox": [0.5793137254901961, 0.03565359477124183, 0.8908333333333335, 0.3133986928104575], "iscrowd": 0, "area": 23351.624299999992, "rle": {"size": [612, 612], "counts": "nWd6R1Rb05K2N1O2N2N1O2N1O001O1O001O1O001O1O001O1O10O01O1O001O1lNYNi@g1U?cNc@^1[?jN^@W1`?ROX@n0g?Y1M3N1O2N2M3N1O2M3O001N2O001O1N101O1O1N4M3M3M3L3N3M3M1N10001O0O101O000O2O00000O2O00001N10001O0O101O00001O0000000000001O0000000000000000001O000000O1000000O1000000O1000000O1000000O1000000O1000000O100O1O1N2N2O1N2O1N2N3N1O1N3N1O1O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2M2O2M2N2O2M2N2O2M2N2O2M2N2O2M2L5K4K5L5K4K5L6I6K6J5J7J5K5M4K9GhoW1"}, "label": "a cup of coffee kept near the breakfast plate"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffiee in the table\"."}], "task": "refering", "answer_template": "The \"a cup of coffiee in the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149], "bbox": [0.5793137254901961, 0.03565359477124183, 0.8908333333333335, 0.3133986928104575], "iscrowd": 0, "area": 23351.624299999992, "rle": {"size": [612, 612], "counts": "nWd6R1Rb05K2N1O2N2N1O2N1O001O1O001O1O001O1O001O1O10O01O1O001O1lNYNi@g1U?cNc@^1[?jN^@W1`?ROX@n0g?Y1M3N1O2N2M3N1O2M3O001N2O001O1N101O1O1N4M3M3M3L3N3M3M1N10001O0O101O000O2O00000O2O00001N10001O0O101O00001O0000000000001O0000000000000000001O000000O1000000O1000000O1000000O1000000O1000000O1000000O100O1O1N2N2O1N2O1N2N3N1O1N3N1O1O2N1O1O2N1O2N1N2O2N1O1O2N1O1O2M2O2M2N2O2M2N2O2M2N2O2M2N2O2M2L5K4K5L5K4K5L6I6K6J5J7J5K5M4K9GhoW1"}, "label": "a cup of coffiee in the table"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"crack on a wooden table\"."}], "task": "refering", "answer_template": "The \"crack on a wooden table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 108, 109, 130, 131, 152, 153, 171, 172, 173, 174, 175, 194, 195, 196, 197, 218, 219, 241], "bbox": [0.8044934640522876, 0.13933006535947712, 1.0, 0.46965686274509805], "iscrowd": 0, "area": 8945.26565, "rle": {"size": [612, 612], "counts": "YkV91Sc00O2O001N2O001O0O2O001N2O001O0O2O001N2O001O0O2O001N101O1N101O001N101O1N101O001N1fLSO]Dn0];XObDi0X;\\OgDe0Z;[OeDf0[;ZOdDf0\\;ZOdDg0\\;YObDh0_;XO`Di0`;VO`Dj0`;WO^Dk0b;TO^Dm0b;SO]Dm0d;SO[Dn0d;RO[Do0f;QOYDP1g;POXDP1h;POWDR1i;nNVDS1j;mNUDS1l;lNTDU1k;lNSDU1n;kNQDV1o;iNQDW1P<iNnCY1Q<hNnCY1R<fNnCZ1S<fNlC[1S<fNkC[1V<dNjC]1V<cNiC]1X<cNeC`1Z<`NeCb1[<^NbCd1_<\\N_Cf1a<YN]Cj1b<WN[Cl1e<SNZCo1f<QNWCR2h<oMUCS2l<lMSCV2l<kMQCX2n<iMPCY2o<gMoB\\2P=eMmB]2S=jMeBX2Z=QN[BR2d=VNSBl1l=\\NjAg1V>d1O001O1O1O1O1O00100O1O1O001O1O1O1O1O010O1O1O1OI7ZOf0YOg0YOg0YOh0XOQ1oNiL"}, "label": "crack on a wooden table"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small cup of syrup\"."}], "task": "refering", "answer_template": "The \"a small cup of syrup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.576764705882353, 0.4073366013071895, 0.8476797385620914, 0.6639052287581699], "iscrowd": 0, "area": 19573.209500000004, "rle": {"size": [612, 612], "counts": "^Xc6=db0e0[O6I7J6J6J5M4L4L2N2N1O2N1O2N1O2N1O1O2N100O10001N100O10001N2O1N2O1O1N2Oa0^O2O1O2M2O1N2O2N1N2O2M2O1O2M2O1N3N00001O0000001N10001O0000001O00000O2O00000000000000000O10O1000001O00000000000O2O00000000001O000O100000001O000O100O100O2O0O1O101N2O0O2O0O2O0O2O1N101N1O2O0O2O1N101N101M3M2N3L3N3M3M3M2N3L4M3M3M2N3M3L4M3M2N3M3L4M3M5K4L4L5J5L4L5K;EQYg1"}, "label": "a small cup of syrup"}]}
{"id": 342011, "image": "COCO_train2014_000000342011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white bowl of syrup\"."}], "task": "refering", "answer_template": "The \"a small white bowl of syrup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.576764705882353, 0.4073366013071895, 0.8476797385620914, 0.6639052287581699], "iscrowd": 0, "area": 19573.209500000004, "rle": {"size": [612, 612], "counts": "^Xc6=db0e0[O6I7J6J6J5M4L4L2N2N1O2N1O2N1O2N1O1O2N100O10001N100O10001N2O1N2O1O1N2Oa0^O2O1O2M2O1N2O2N1N2O2M2O1O2M2O1N3N00001O0000001N10001O0000001O00000O2O00000000000000000O10O1000001O00000000000O2O00000000001O000O100000001O000O100O100O2O0O1O101N2O0O2O0O2O0O2O1N101N1O2O0O2O1N101N101M3M2N3L3N3M3M3M2N3L4M3M3M2N3M3L4M3M2N3M3L4M3M5K4L4L5J5L4L5K;EQYg1"}, "label": "a small white bowl of syrup"}]}
{"id": 169986, "image": "COCO_train2014_000000169986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brunette woman with a dumb hat , holding pizza\"."}], "task": "refering", "answer_template": "The \"a brunette woman with a dumb hat , holding pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.0016875000000000002, 1.0, 0.985890625], "iscrowd": 0, "area": 252779.84045, "rle": {"size": [640, 480], "counts": "X9^:a9100000000O100000000O1000000O100000000O100000000O1000000O1000000SIgES4Y:VKQGV4P9YKTHT4l7`KPIP4P7oKYIi3g6ULcIc3]6\\LlI[3U6dLVJR3j5lLlJ`2T5_MYKU2g4iMaKP2`4oMfKl1Z4SNlKh1T4VNTLd1l3[NZL`1f3`NRM>X3Bc80000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O2O1N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2a_ONl<e3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O^F"}, "label": "a brunette woman with a dumb hat , holding pizza"}]}
{"id": 169986, "image": "COCO_train2014_000000169986.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with a slice of pizza\"."}], "task": "refering", "answer_template": "The \"the person with a slice of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.0016875000000000002, 1.0, 0.985890625], "iscrowd": 0, "area": 252779.84045, "rle": {"size": [640, 480], "counts": "X9^:a9100000000O100000000O1000000O100000000O100000000O1000000O1000000SIgES4Y:VKQGV4P9YKTHT4l7`KPIP4P7oKYIi3g6ULcIc3]6\\LlI[3U6dLVJR3j5lLlJ`2T5_MYKU2g4iMaKP2`4oMfKl1Z4SNlKh1T4VNTLd1l3[NZL`1f3`NRM>X3Bc80000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O2O1N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2a_ONl<e3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O^F"}, "label": "the person with a slice of pizza"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with an orange t - shirt with pizza\"."}], "task": "refering", "answer_template": "The \"a man with an orange t - shirt with pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 21, 22, 23, 24, 25, 26, 37, 38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 68, 69, 70, 71, 72, 73, 74, 75, 76, 85, 86, 87, 88, 89, 90, 91, 92, 93, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241], "bbox": [0.0, 0.0010781249999999999, 0.7011458333333334, 0.6433125], "iscrowd": 0, "area": 89435.31075, "rle": {"size": [640, 480], "counts": "h4i2Va0o4PK3N2N2N2N2N2M3N2N2N2N2N2N2M1010000000O10000000000O100000000O1000O1000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10001O000O10001O000O1000001O000O10001O000O10O100000O10000000O0VNPHfGP8W8VHfGj7T7THRH9g0c7S7[HPH8j0]7R7bHnG7m0i6eNgH]8`0kG6P1`6lNiHU8h0hG5T1[6oNhHQ8Y2mHo4SOgHl7a2nHh4WOfHg7i2oHa4ZOeHd7Q3nH[4_OcHb7V3lHW4CbHa7Z3hHU4GaHa7]3eHR4K`H`7a3aHQ4O\\Ha7f3]Hn32\\Ha7i3ZHk36[H`7l3WHj39ZH`7f;aHYD_7g;=0000000000001O00000000000000000000001O0000000000000000000000001O000000000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O100000001O000000000O100000000000000ZO_GmEa8l9fGTFZ8e9nGZFR8_9UHaFk7Y9[HgFe7R9cHmF]7l8jHTGV7e8QIZGP7_8XI`Gh6Y8_IgGa6R8fInGZ6l7mISHS6o7kIQHU6Q8iIoGW6S8hIlGX6V8fIjGZ6W8fIhGZ6Z8dIfG\\6`4^GmNT2cL^6]4bGPOP2bL^6Z4gGUOi1aL`6W4kGYOc1`Lb6S4PH]O^1_Lb6Q4THAX1]Le6o3WHER1\\Lg6k3]HHl0\\Lg6i3aHLf0VLXNMa8m3fH1?UL[NL`8k3jH4;SLQ7e3iH94QLT7c3lH=NPLV7\\3TId0b8SOiGm0U8jNUHW1i7eN\\H[1c7aNbH`1\\7\\NiHe1U7WNPIi1T7oMQIQ2R7fMSI[2P7\\MUId2o6SMVIn2n6hLWIY3l6_LXIa3l6VLYIj3k6mK[IR4Z;0003M2N3M3M3M2M4M3M0000000000001O00000000000000000O1O1O1O1O2O0O1O1O1O1O1O1FT_O^Mm`0a2:O1O1O1O2O0O1O1O1O1O1O1O1O100O1O1O1O2N1O1O100M3M3M8QOk]OE]b02i]OI^b0Oh]OL`b0Kf]O0kdi2"}, "label": "a man with an orange t - shirt with pizza"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in orange shirt\"."}], "task": "refering", "answer_template": "The \"a man in orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 21, 22, 23, 24, 25, 26, 37, 38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 68, 69, 70, 71, 72, 73, 74, 75, 76, 85, 86, 87, 88, 89, 90, 91, 92, 93, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241], "bbox": [0.0, 0.0010781249999999999, 0.7011458333333334, 0.6433125], "iscrowd": 0, "area": 89435.31075, "rle": {"size": [640, 480], "counts": "h4i2Va0o4PK3N2N2N2N2N2M3N2N2N2N2N2N2M1010000000O10000000000O100000000O1000O1000O100000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O10001O000O10001O000O1000001O000O10001O000O10O100000O10000000O0VNPHfGP8W8VHfGj7T7THRH9g0c7S7[HPH8j0]7R7bHnG7m0i6eNgH]8`0kG6P1`6lNiHU8h0hG5T1[6oNhHQ8Y2mHo4SOgHl7a2nHh4WOfHg7i2oHa4ZOeHd7Q3nH[4_OcHb7V3lHW4CbHa7Z3hHU4GaHa7]3eHR4K`H`7a3aHQ4O\\Ha7f3]Hn32\\Ha7i3ZHk36[H`7l3WHj39ZH`7f;aHYD_7g;=0000000000001O00000000000000000000001O0000000000000000000000001O000000000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O100000001O000000000O100000000000000ZO_GmEa8l9fGTFZ8e9nGZFR8_9UHaFk7Y9[HgFe7R9cHmF]7l8jHTGV7e8QIZGP7_8XI`Gh6Y8_IgGa6R8fInGZ6l7mISHS6o7kIQHU6Q8iIoGW6S8hIlGX6V8fIjGZ6W8fIhGZ6Z8dIfG\\6`4^GmNT2cL^6]4bGPOP2bL^6Z4gGUOi1aL`6W4kGYOc1`Lb6S4PH]O^1_Lb6Q4THAX1]Le6o3WHER1\\Lg6k3]HHl0\\Lg6i3aHLf0VLXNMa8m3fH1?UL[NL`8k3jH4;SLQ7e3iH94QLT7c3lH=NPLV7\\3TId0b8SOiGm0U8jNUHW1i7eN\\H[1c7aNbH`1\\7\\NiHe1U7WNPIi1T7oMQIQ2R7fMSI[2P7\\MUId2o6SMVIn2n6hLWIY3l6_LXIa3l6VLYIj3k6mK[IR4Z;0003M2N3M3M3M2M4M3M0000000000001O00000000000000000O1O1O1O1O2O0O1O1O1O1O1O1FT_O^Mm`0a2:O1O1O1O2O0O1O1O1O1O1O1O1O100O1O1O1O2N1O1O100M3M3M8QOk]OE]b02i]OI^b0Oh]OL`b0Kf]O0kdi2"}, "label": "a man in orange shirt"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large piece of pizza in front\"."}], "task": "refering", "answer_template": "The \"large piece of pizza in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338, 339, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.3428541666666666, 0.65121875, 1.0, 0.996671875], "iscrowd": 0, "area": 40083.1268, "rle": {"size": [640, 480], "counts": "`gW31nc01O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O2M2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O2M2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N3N1N2N2O1N2001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O10O01O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O00100O1O1O1O001O1O1O1O1O1O001O1O1O1O001O001O001O010O001O001O001O1O001O001O001O001O001O001O001O1O010O001O001O001O001O001O1O001O001O001O0000001O01O0001O000000001O0000001O0000001O000000001O0001O01O0000001O0J6_Oa0_Ob0]Ob0_Oa0^Ob0_Oa0^Om@"}, "label": "large piece of pizza in front"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of pizza on paper plate closest to camera\"."}], "task": "refering", "answer_template": "The \"slice of pizza on paper plate closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 315, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338, 339, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.3428541666666666, 0.65121875, 1.0, 0.996671875], "iscrowd": 0, "area": 40083.1268, "rle": {"size": [640, 480], "counts": "`gW31nc01O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O2M2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O2M2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N3N1N2N2O1N2001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O10O01O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O00100O1O1O1O001O1O1O1O1O1O001O1O1O1O001O001O001O010O001O001O001O1O001O001O001O001O001O001O001O1O010O001O001O001O001O001O1O001O001O001O0000001O01O0001O000000001O0000001O0000001O000000001O0001O01O0000001O0J6_Oa0_Ob0]Ob0_Oa0^Ob0_Oa0^Om@"}, "label": "slice of pizza on paper plate closest to camera"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of pizzxa that is ont he plate that has cheese and vegetables on it\"."}], "task": "refering", "answer_template": "The \"a piece of pizzxa that is ont he plate that has cheese and vegetables on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 306, 307, 308, 323], "bbox": [0.0, 0.62528125, 0.33558333333333334, 0.8455], "iscrowd": 0, "area": 12780.028149999998, "rle": {"size": [640, 480], "counts": "T`0i0ea0a1^Nb1B?1N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O101N100O10001N100O100O2O0O100O100O2O0O100O101N100O100O2O0O100O101N100O100O2O0O100O100O2O0O100O101N100O100O2O0O100O101N100O100O2O0O1O2N2M3N2N1N3N2N2M3N2N2M2O2N2M3N2NSnV6"}, "label": "a piece of pizzxa that is ont he plate that has cheese and vegetables on it"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of pizza in the left corner is almost cut from the picture\"."}], "task": "refering", "answer_template": "The \"the piece of pizza in the left corner is almost cut from the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 306, 307, 308, 323], "bbox": [0.0, 0.62528125, 0.33558333333333334, 0.8455], "iscrowd": 0, "area": 12780.028149999998, "rle": {"size": [640, 480], "counts": "T`0i0ea0a1^Nb1B?1N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O101N100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O101N100O10001N100O100O2O0O100O100O2O0O100O101N100O100O2O0O100O101N100O100O2O0O100O100O2O0O100O101N100O100O2O0O100O101N100O100O2O0O1O2N2M3N2N1N3N2N2M3N2N2M2O2N2M3N2NSnV6"}, "label": "the piece of pizza in the left corner is almost cut from the picture"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of pizza in a plate before the guy with yellow t - shirt\"."}], "task": "refering", "answer_template": "The \"a piece of pizza in a plate before the guy with yellow t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.22, 0.547625, 0.742125, 0.654296875], "iscrowd": 0, "area": 7836.906600000001, "rle": {"size": [640, 480], "counts": "acR21Pd00O1O1O1O100O001O1O1O10O01O1O00100O001O1O010O1O3M4L10O000001O01O000001O01O0001O01O0001O01O0000010O0000010O01O1O1O010O1O1O001O100O001O1O01O00000000001O0000000000001O0000000000001O000000000000001O000000000000000000000000000000000000O101O000000000000001O000000000000001O000O10000000001O000000000000001O00000000000O101O0000000000000000O100000000000000000O1000000000000000000O1O1N110O10000O100O100O100O100O100O10000O100O100O010O100O1N2O1N2O1N2O1N2O1O1N2O0O20O101N100O100O1O100O100O100O100O1O1N2N2N2N2N2NSU]2"}, "label": "a piece of pizza in a plate before the guy with yellow t - shirt"}]}
{"id": 194564, "image": "COCO_train2014_000000194564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of pepperoni pizza on a paper plate\"."}], "task": "refering", "answer_template": "The \"slice of pepperoni pizza on a paper plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.22, 0.547625, 0.742125, 0.654296875], "iscrowd": 0, "area": 7836.906600000001, "rle": {"size": [640, 480], "counts": "acR21Pd00O1O1O1O100O001O1O1O10O01O1O00100O001O1O010O1O3M4L10O000001O01O000001O01O0001O01O0001O01O0000010O0000010O01O1O1O010O1O1O001O100O001O1O01O00000000001O0000000000001O0000000000001O000000000000001O000000000000000000000000000000000000O101O000000000000001O000000000000001O000O10000000001O000000000000001O00000000000O101O0000000000000000O100000000000000000O1000000000000000000O1O1N110O10000O100O100O100O100O100O10000O100O100O010O100O1N2O1N2O1N2O1N2O1O1N2O0O20O101N100O100O1O100O100O100O100O1O1N2N2N2N2N2NSU]2"}, "label": "slice of pepperoni pizza on a paper plate"}]}
{"id": 358405, "image": "COCO_train2014_000000358405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a wine glass\"."}], "task": "refering", "answer_template": "The \"a woman holding a wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 143, 144, 162, 163, 164, 166, 167, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0003125, 0.18270833333333333, 0.321125, 1.0], "iscrowd": 0, "area": 57302.45489999998, "rle": {"size": [480, 640], "counts": "m8S6]8`000O100O100O100O100O100O100O100O100O100O100O100O1O100O100O1O100O100O100_LVIkNk6R1[IkNe6R1aIjN`6R1hIjNX6S1XJ_Ni5^1]J_Nc5^1bJ`N^5^1fJ_N[5^1kJ_NU5`1nJ]NS5b1PK\\NP5c1TKZNl4e1WKgLbMZ1W7n1aKPN`4P2aKoM_4P2cKoM]4P2eKoM[4Q2gKmMY4R2mKiMS4V2RLfMn3Z2gHjL[3f0o3d2ULUMo3j2RLSMQ4l2oKSMS4m2mKQMU4n2lKPMV4o2jKPMX4o2iKoLY4Q3gKmL[4R3fKmL[4R3eKnL\\4R3dKmL]4R3dKmL]4S3bKmL_4R3bKnL^4R3bKmL_4S3`KmLa4R3g300O100O100O100O100O100O100O100O100OlFRNX5n1]JaN_5^1XJPOd5P1RJ\\Ol5d0lIFR6:nIHP68PJJn56RJLl54SJOk51UJ1h50XJ2f5NZJ4d5K\\J8b5H^J9a5G_J9a5G^J;a5E_J<`5D`J<`5D`J=_5C`J?_5AaJ?_5AaJ`0^5@bJa0]5_ObJb0^5^ObJc0]5]OcJc0]5]ObJe0]5[OcJe0]5[OcJe0]5[OcJf0\\5[OcJe0]5[OcJf0\\5ZOdJf0\\5ZOeJk0U5UOkJR1n4nNRKh1Y4WNgKi1Y4XNfKh1Z4XNfKg1[4YNeKg1[4YNeKg1[4ZNdKf1\\4ZNdKf1\\4ZNdKa1a4_N_Kc0lNfLo5h2TKb0`5^O`Ja0a5_O`J`0`5A_J>b5B^J>b5B^J=c5C]J=c5D]J:d5F\\J:d5F\\J9e5G[J9e5H[J6f5JZJ6f5J[J4f5M[J0f51\\JKe56\\JGe5:]JCc5=`J@`5a0aJ]O_5d0cJYO]5h0dJVO\\5k0fJROZ5n0jJmNW5T1nJfNR5[1RK`Nn4a1ZKVNf4k1RL\\Mn3f2PLZMP4g2nKZMR4h2kKZMT4g2kKZMT4h2iKZMV4h2gK[MW4f2hK\\MV4f2gK^MV4c2ZIbLe1n0o4d2PIiLn1c0R5_3jJcLU5_3hJcLW5_3cJeL]5]3YJlLf5U3RJRMn5P3iIXMV6k2`I]M_6b501O1O001O001gJnGb3S8[LSHa3m7^LXH^3i7`L]H[3d7bLbHZ3`7cLfHX3[7fLjHV3X7gLmHU3U7fLRIV3Q7cLWIY3T7XLTId3X9L5K4L6J5K5\\NQDMU<jNXT[6"}, "label": "a woman holding a wine glass"}]}
{"id": 358405, "image": "COCO_train2014_000000358405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman raising the wine glass\"."}], "task": "refering", "answer_template": "The \"woman raising the wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 143, 144, 162, 163, 164, 166, 167, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0003125, 0.18270833333333333, 0.321125, 1.0], "iscrowd": 0, "area": 57302.45489999998, "rle": {"size": [480, 640], "counts": "m8S6]8`000O100O100O100O100O100O100O100O100O100O100O100O1O100O100O1O100O100O100_LVIkNk6R1[IkNe6R1aIjN`6R1hIjNX6S1XJ_Ni5^1]J_Nc5^1bJ`N^5^1fJ_N[5^1kJ_NU5`1nJ]NS5b1PK\\NP5c1TKZNl4e1WKgLbMZ1W7n1aKPN`4P2aKoM_4P2cKoM]4P2eKoM[4Q2gKmMY4R2mKiMS4V2RLfMn3Z2gHjL[3f0o3d2ULUMo3j2RLSMQ4l2oKSMS4m2mKQMU4n2lKPMV4o2jKPMX4o2iKoLY4Q3gKmL[4R3fKmL[4R3eKnL\\4R3dKmL]4R3dKmL]4S3bKmL_4R3bKnL^4R3bKmL_4S3`KmLa4R3g300O100O100O100O100O100O100O100O100OlFRNX5n1]JaN_5^1XJPOd5P1RJ\\Ol5d0lIFR6:nIHP68PJJn56RJLl54SJOk51UJ1h50XJ2f5NZJ4d5K\\J8b5H^J9a5G_J9a5G^J;a5E_J<`5D`J<`5D`J=_5C`J?_5AaJ?_5AaJ`0^5@bJa0]5_ObJb0^5^ObJc0]5]OcJc0]5]ObJe0]5[OcJe0]5[OcJe0]5[OcJf0\\5[OcJe0]5[OcJf0\\5ZOdJf0\\5ZOeJk0U5UOkJR1n4nNRKh1Y4WNgKi1Y4XNfKh1Z4XNfKg1[4YNeKg1[4YNeKg1[4ZNdKf1\\4ZNdKf1\\4ZNdKa1a4_N_Kc0lNfLo5h2TKb0`5^O`Ja0a5_O`J`0`5A_J>b5B^J>b5B^J=c5C]J=c5D]J:d5F\\J:d5F\\J9e5G[J9e5H[J6f5JZJ6f5J[J4f5M[J0f51\\JKe56\\JGe5:]JCc5=`J@`5a0aJ]O_5d0cJYO]5h0dJVO\\5k0fJROZ5n0jJmNW5T1nJfNR5[1RK`Nn4a1ZKVNf4k1RL\\Mn3f2PLZMP4g2nKZMR4h2kKZMT4g2kKZMT4h2iKZMV4h2gK[MW4f2hK\\MV4f2gK^MV4c2ZIbLe1n0o4d2PIiLn1c0R5_3jJcLU5_3hJcLW5_3cJeL]5]3YJlLf5U3RJRMn5P3iIXMV6k2`I]M_6b501O1O001O001gJnGb3S8[LSHa3m7^LXH^3i7`L]H[3d7bLbHZ3`7cLfHX3[7fLjHV3X7gLmHU3U7fLRIV3Q7cLWIY3T7XLTId3X9L5K4L6J5K5\\NQDMU<jNXT[6"}, "label": "woman raising the wine glass"}]}
{"id": 358405, "image": "COCO_train2014_000000358405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in purple shirt at a winery\"."}], "task": "refering", "answer_template": "The \"a man in purple shirt at a winery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 316, 317, 318, 319, 320, 321, 335, 336, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389], "bbox": [0.5771875, 0.00020833333333333335, 1.0, 1.0], "iscrowd": 0, "area": 74690.0938, "rle": {"size": [480, 640], "counts": "^j]51n>1O2O0O2N101N2N101N1O2O0O010O0010ON3M3M2O2M2N3M2M4N101O001O0010O0000]BSOT=m0hBXOW=X1001O0000000000010O0O100000001O00O1O1O1O1O001O2N1O1O1O1O1O1N3N1N2O1N2O1O1N2O2MK6I6011N2O2M2O1N3N1N2O2M2O1K5L5J5K5`GlLP5Z3mJoLg4V3WKSM^4Q3`K^Mn3h2oKdMb3a2\\LbM^3c2_LaM[3e2bL^MX3g2fL\\MT3i2jLZMS3i2iL[MT3g2jL\\MS3f2jL^MS3e2jL^MS3d2jL`MS3b2kLaMR3b2jLbMS3`2hLfMV3\\2aLmM]3W2[LoMc3T2[LmMc3V2ZLlMe3W2WLkMg3X2VLjMh3Z2SLiMk3Z2eKUNY4Z6O1O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O10000000002N4K4M2N1O001O001O0O2OAjLTEV3l:jLUEU3k:kLUEU3k:kLVET3j:lLVET3j:lLWEb0Bk1W;cMWE?Fl1T;eMWE;Io1P;fMWE8LR2m:fMXE5N9DV1W;\\NWE4015T1d:gNXE2Y1l0_9ROXE2c1a0V9]OXE0c1c0U9]OXE0d1a0U9_ObG`0^8@cG>^8BcG=]8CcG=]8CdG;]8EdG:\\8FeG8\\8HdG8\\8HeG6\\8JeG5[8KfG3[8MfG1[8O[3000000000000000000001O000000000000000000000000000000000000000000000000000N2M3M3L4M3M3N2M3N2M3N3M2M3N2O1N2N2N2O1N2N2O1O1000000000000000000O1000000000000000000O1iKW4PJm5"}, "label": "a man in purple shirt at a winery"}]}
{"id": 358405, "image": "COCO_train2014_000000358405.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing purple shirt , jeans and many beaded necklaces\"."}], "task": "refering", "answer_template": "The \"man wearing purple shirt , jeans and many beaded necklaces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 316, 317, 318, 319, 320, 321, 335, 336, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389], "bbox": [0.5771875, 0.00020833333333333335, 1.0, 1.0], "iscrowd": 0, "area": 74690.0938, "rle": {"size": [480, 640], "counts": "^j]51n>1O2O0O2N101N2N101N1O2O0O010O0010ON3M3M2O2M2N3M2M4N101O001O0010O0000]BSOT=m0hBXOW=X1001O0000000000010O0O100000001O00O1O1O1O1O001O2N1O1O1O1O1O1N3N1N2O1N2O1O1N2O2MK6I6011N2O2M2O1N3N1N2O2M2O1K5L5J5K5`GlLP5Z3mJoLg4V3WKSM^4Q3`K^Mn3h2oKdMb3a2\\LbM^3c2_LaM[3e2bL^MX3g2fL\\MT3i2jLZMS3i2iL[MT3g2jL\\MS3f2jL^MS3e2jL^MS3d2jL`MS3b2kLaMR3b2jLbMS3`2hLfMV3\\2aLmM]3W2[LoMc3T2[LmMc3V2ZLlMe3W2WLkMg3X2VLjMh3Z2SLiMk3Z2eKUNY4Z6O1O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O10000000002N4K4M2N1O001O001O0O2OAjLTEV3l:jLUEU3k:kLUEU3k:kLVET3j:lLVET3j:lLWEb0Bk1W;cMWE?Fl1T;eMWE;Io1P;fMWE8LR2m:fMXE5N9DV1W;\\NWE4015T1d:gNXE2Y1l0_9ROXE2c1a0V9]OXE0c1c0U9]OXE0d1a0U9_ObG`0^8@cG>^8BcG=]8CcG=]8CdG;]8EdG:\\8FeG8\\8HdG8\\8HeG6\\8JeG5[8KfG3[8MfG1[8O[3000000000000000000001O000000000000000000000000000000000000000000000000000N2M3M3L4M3M3N2M3N2M3N3M2M3N2O1N2N2N2O1N2N2O1O1000000000000000000O1000000000000000000O1iKW4PJm5"}, "label": "man wearing purple shirt , jeans and many beaded necklaces"}]}
{"id": 309252, "image": "COCO_train2014_000000309252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white horse\"."}], "task": "refering", "answer_template": "The \"white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 215, 216, 217, 218, 219, 230, 231, 232, 234, 240, 241, 242, 253, 255], "bbox": [0.0, 0.17295774647887324, 0.551359375, 0.7652347417840376], "iscrowd": 0, "area": 49529.50275, "rle": {"size": [426, 640], "counts": "T6P2[;Q2oMNI8K5K4M4000fNgGbMZ8]2Z1O0O2O1N2O1O1N2O1O001O1O1N2O1O1O1O1N1011O1O1O001O1O001N2O1O001O1O001O1N2O001O1O0DnL`FU3]9oL_FZ3X9iLeF_3S9d0H9G8H3M0000000000000000000000000000000000cNTH\\Ml7c2mHeLQ7]3[12N2O1N2N2N2O1O1100O1O1O1O1O1O1O100O1O1O1010O1O1O1O1O1O1N2N2O1N101N2N2O1N2O1N6K<C=C=D3L10O01O010O010O0010O010O010O01O01I6J7I6J6M4L3N3L3N201O001O0O101O001O001O02N101N101N1O2O0O2O1N3N2N1O2N2N2N1O2N2N2NM3M3M3M3N2O1O1O1O1O1O1O1O001O000000001O00000000N2M3N2N2M310100000000000000O10000000000000N1O2M3N2N2N1N3N2N2001N1000001O0O101O00000O0100O1000O0100O100O01000O010O0100O0010O0100O0010O0100O0010O010O01O100O100O2O0O2N101N100O2O0O2O001N100O2O001N101N10001N101N101O0O101N101O0O2O0O1N3N3M2N3L4M3M2M4M3M2M4M3M3M2M4L4M2M4L4M3L3M4M3L3M4M3L4L3N3L4L3N3L4L4K4K6K5K4K6KYTg3"}, "label": "white horse"}]}
{"id": 309252, "image": "COCO_train2014_000000309252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse behind a gate\"."}], "task": "refering", "answer_template": "The \"a white horse behind a gate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 215, 216, 217, 218, 219, 230, 231, 232, 234, 240, 241, 242, 253, 255], "bbox": [0.0, 0.17295774647887324, 0.551359375, 0.7652347417840376], "iscrowd": 0, "area": 49529.50275, "rle": {"size": [426, 640], "counts": "T6P2[;Q2oMNI8K5K4M4000fNgGbMZ8]2Z1O0O2O1N2O1O1N2O1O001O1O1N2O1O1O1O1N1011O1O1O001O1O001N2O1O001O1O001O1N2O001O1O0DnL`FU3]9oL_FZ3X9iLeF_3S9d0H9G8H3M0000000000000000000000000000000000cNTH\\Ml7c2mHeLQ7]3[12N2O1N2N2N2O1O1100O1O1O1O1O1O1O100O1O1O1010O1O1O1O1O1O1N2N2O1N101N2N2O1N2O1N6K<C=C=D3L10O01O010O010O0010O010O010O01O01I6J7I6J6M4L3N3L3N201O001O0O101O001O001O02N101N101N1O2O0O2O1N3N2N1O2N2N2N1O2N2N2NM3M3M3M3N2O1O1O1O1O1O1O1O001O000000001O00000000N2M3N2N2M310100000000000000O10000000000000N1O2M3N2N2N1N3N2N2001N1000001O0O101O00000O0100O1000O0100O100O01000O010O0100O0010O0100O0010O0100O0010O010O01O100O100O2O0O2N101N100O2O0O2O001N100O2O001N101N10001N101N101O0O101N101O0O2O0O1N3N3M2N3L4M3M2M4M3M2M4M3M3M2M4L4M2M4L4M3L3M4M3L3M4M3L4L3N3L4L3N3L4L4K4K6K5K4K6KYTg3"}, "label": "a white horse behind a gate"}]}
{"id": 178181, "image": "COCO_train2014_000000178181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black jacket skateboarding\"."}], "task": "refering", "answer_template": "The \"a man in a black jacket skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 103, 104, 105, 121, 122, 123, 139, 140, 141, 157, 158, 175, 176, 193, 194], "bbox": [0.72238, 0.3817117117117117, 0.85114, 0.8666966966966967], "iscrowd": 0, "area": 5638.651499999998, "rle": {"size": [333, 500], "counts": "nbe32Y:7H8H4M2M3M3M3N2M3M3N2M5K5L4L3N3L4M3\\ISNb4Q2XKSNf4P2VKSNi4P2[JfM2<b5R2SJQN00k5g3O10000O10000000000O100000O1_ORJbLn5[3WJcLa5b3eJ[L\\5a3iJ]LW5`3nJ^LR5`3SK^Lm4^3XK`Lh4[3X1H7J7O00010O0010O01O0M5L4K5K5L3L5]NjGi0[8oNmGm0W8lNPHf0S9[Ojmg0"}, "label": "a man in a black jacket skateboarding"}]}
{"id": 178181, "image": "COCO_train2014_000000178181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy riding a skateboard in the middle of a street\"."}], "task": "refering", "answer_template": "The \"a guy riding a skateboard in the middle of a street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 103, 104, 105, 121, 122, 123, 139, 140, 141, 157, 158, 175, 176, 193, 194], "bbox": [0.72238, 0.3817117117117117, 0.85114, 0.8666966966966967], "iscrowd": 0, "area": 5638.651499999998, "rle": {"size": [333, 500], "counts": "nbe32Y:7H8H4M2M3M3M3N2M3M3N2M5K5L4L3N3L4M3\\ISNb4Q2XKSNf4P2VKSNi4P2[JfM2<b5R2SJQN00k5g3O10000O10000000000O100000O1_ORJbLn5[3WJcLa5b3eJ[L\\5a3iJ]LW5`3nJ^LR5`3SK^Lm4^3XK`Lh4[3X1H7J7O00010O0010O01O0M5L4K5K5L3L5]NjGi0[8oNmGm0W8lNPHf0S9[Ojmg0"}, "label": "a guy riding a skateboard in the middle of a street"}]}
{"id": 178181, "image": "COCO_train2014_000000178181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small silver car parked on the road side\"."}], "task": "refering", "answer_template": "The \"a small silver car parked on the road side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 135, 136, 137, 138, 139, 141, 142, 154, 155, 156, 157, 158, 159], "bbox": [0.54516, 0.5445345345345346, 0.89484, 0.7221321321321322], "iscrowd": 0, "area": 5146.7309000000005, "rle": {"size": [333, 500], "counts": "foh28S:2M5N1N1O2N2N1O2O0O2O0O2O0O101N1000001O0000001O0O1000001O000O2O0000O1000O10O11O0O100O1O2O0O1O101N101N1O100O100O100O100O10000O100O010O0100O1N1O2N200000000000O1000000000000O1000000000lNZGa0f8ZO_Gf0a8VObGk0]8ROgGn0k81004L7I2N2N2N2M3N2N2N3M2NPT8;ekG1N10000O10000M2N3N2M2N3M3N3L3M4J51N2O1O1O0000000O100O100N2O1O1O1O101O2N2N2M4M4LoPa0"}, "label": "a small silver car parked on the road side"}]}
{"id": 120836, "image": "COCO_train2014_000000120836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top half of a whole wheat hamburger bun\"."}], "task": "refering", "answer_template": "The \"the top half of a whole wheat hamburger bun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 127, 128], "bbox": [0.362171875, 0.006752336448598131, 0.82203125, 0.33932242990654204], "iscrowd": 0, "area": 31961.8696, "rle": {"size": [428, 640], "counts": "bQQ32X=3L4M3M3M3L4M3M3M3L4M3M3M2M4M3M3M^N\\Da1`;]NdDd1b;1N2O1N101N2O1N101N2O1N101N2O1N101N2N2O1N101N2O1N101N2O1N2O0O2O1N2O1N101N2O1N100O2O0O101N100O2O0O101N101N1O101N100O2O0O101N100O2O0O101O00001N10001O000O2O00001O0O101O00000O100000000O100000000O100000000O100000000O100000000O100000000O100000000000000000O100000000000000000000000000O1000000000000000001O0000000O10000000000000000000001O00000O101O000000001O000000001O000000001O000000001O000000001O000000001O00001O000O101O00001N100O2O0O100O2O0O101N100O101N100O2O0O101N100O101N100O2O0O101N1O100O2N1O2O1N2N2N2O2M2N2N2O1N2N2N2O1N2N3M2O1N2M3M3M6J5K6J5K5K6J5K6J5K[b_1"}, "label": "the top half of a whole wheat hamburger bun"}]}
{"id": 120836, "image": "COCO_train2014_000000120836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there are cucumber pieces on a white plate\"."}], "task": "refering", "answer_template": "The \"there are cucumber pieces on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343], "bbox": [0.471875, 0.32135514018691586, 1.0, 0.9707943925233645], "iscrowd": 0, "area": 76145.06280000001, "rle": {"size": [428, 640], "counts": "o`n36n<:F;E:F;E;E:F;E:F;J6K4L5J4M3M3M2M4M3M3L4M3M3L4M2N3L4M3M3L4M3M2N3L4M3N2O1N2N101N2O0O2N100O2O0O100O2N100O2O0O1O101N100O2N100O101N1O101N100O101N1O101N100O1O2O0O101N1O100O2O0O1O2O0O101N1000001O00001N10001O0000001O0O101O0000001O00001N10001O0000001O000O2O00001O0000001O0O101O00001O0000001N10001O00001O00000O2O00001O00001O000O101O00001O00001O0O10001O00001O00001O0000010O00001O01O01O0001O000000001O0O1000001O0000001O000000001O0000001O0000001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O001O001O001O001O001O001O1O001N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2O0O2N2O0O2O0O2N101N101N1O2O0O2O0O2N101N101N1O2O0O2O2M3M3N1N3N2M3N2M2N3N2M3N2M3M2O2M3N2M3TM^K"}, "label": "there are cucumber pieces on a white plate"}]}
{"id": 120836, "image": "COCO_train2014_000000120836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate filled with cucumber\"."}], "task": "refering", "answer_template": "The \"a plate filled with cucumber\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343], "bbox": [0.471875, 0.32135514018691586, 1.0, 0.9707943925233645], "iscrowd": 0, "area": 76145.06280000001, "rle": {"size": [428, 640], "counts": "o`n36n<:F;E:F;E;E:F;E:F;J6K4L5J4M3M3M2M4M3M3L4M3M3L4M2N3L4M3M3L4M3M2N3L4M3N2O1N2N101N2O0O2N100O2O0O100O2N100O2O0O1O101N100O2N100O101N1O101N100O101N1O101N100O1O2O0O101N1O100O2O0O1O2O0O101N1000001O00001N10001O0000001O0O101O0000001O00001N10001O0000001O000O2O00001O0000001O0O101O00001O0000001N10001O00001O00000O2O00001O00001O000O101O00001O00001O0O10001O00001O00001O0000010O00001O01O01O0001O000000001O0O1000001O0000001O000000001O0000001O0000001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O001O001O001O001O001O001O1O001N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2O0O2N2O0O2O0O2N101N101N1O2O0O2O0O2N101N101N1O2O0O2O2M3M3N1N3N2M3N2M2N3N2M3N2M3M2O2M3N2M3TM^K"}, "label": "a plate filled with cucumber"}]}
{"id": 47116, "image": "COCO_train2014_000000047116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train passing through the address board back side\"."}], "task": "refering", "answer_template": "The \"a train passing through the address board back side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.278015625, 0.7415654205607476, 0.846078125, 0.8988785046728971], "iscrowd": 0, "area": 15675.387450000002, "rle": {"size": [428, 640], "counts": "mfZ21=1^<1`C3[<0bC4[<McC8X<KeC9X<`0M3N2O1N1O00000000000000001OO1M3M3M3010O2N1O2N2N1O2TDiN_;X1[DoNc;_1N2N2N2N000001O00000000000000001O0000000000000O2O000000000000001O00000000000000001O000000000000000000001O0000000000000000001O000000000000000000001O0000000000000000001N100000000000000000001O000000000000000000001O0000000000000000001O0000000000000000000G:B=C=CZj8c0dTGl0M300000000000000000O101O000000000000000000000000000000001O0000000000000000000000000000001O00000O100000000000000000000000001O0000000000000000000000000000001O000000000000000O10000000000000001O0000000000000000000000000000001O00000O100O100O100O100O100O1000001O0000000000000000000000000O10001O000000000000000000000000000000001O00000000000O100000000O1FhC@X<`0:O1O100O10\\QY1"}, "label": "a train passing through the address board back side"}]}
{"id": 47116, "image": "COCO_train2014_000000047116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three olive green train cars with orangish tops\"."}], "task": "refering", "answer_template": "The \"three olive green train cars with orangish tops\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.278015625, 0.7415654205607476, 0.846078125, 0.8988785046728971], "iscrowd": 0, "area": 15675.387450000002, "rle": {"size": [428, 640], "counts": "mfZ21=1^<1`C3[<0bC4[<McC8X<KeC9X<`0M3N2O1N1O00000000000000001OO1M3M3M3010O2N1O2N2N1O2TDiN_;X1[DoNc;_1N2N2N2N000001O00000000000000001O0000000000000O2O000000000000001O00000000000000001O000000000000000000001O0000000000000000001O000000000000000000001O0000000000000000001N100000000000000000001O000000000000000000001O0000000000000000001O0000000000000000000G:B=C=CZj8c0dTGl0M300000000000000000O101O000000000000000000000000000000001O0000000000000000000000000000001O00000O100000000000000000000000001O0000000000000000000000000000001O000000000000000O10000000000000001O0000000000000000000000000000001O00000O100O100O100O100O100O1000001O0000000000000000000000000O10001O000000000000000000000000000000001O00000000000O100000000O1FhC@X<`0:O1O100O10\\QY1"}, "label": "three olive green train cars with orangish tops"}]}
{"id": 79887, "image": "COCO_train2014_000000079887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"steel mixer above box on bottom shelf\"."}], "task": "refering", "answer_template": "The \"steel mixer above box on bottom shelf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 299, 300, 301, 302], "bbox": [0.5783541666666667, 0.55325, 0.8692708333333333, 0.7818125], "iscrowd": 0, "area": 16341.40465, "rle": {"size": [640, 480], "counts": "US^57ic08H8H8H8H7I8H8G9H8H7I8H8H8H8H7I6J0O2O001O00001O001O000O2O001O00001O001O00001O001O00001O00001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000001O0O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1N2L5L3L4L5L3L4L5L3L4L5L3L4L5L3L4L5L3L6J5L5J5K6K4K6J5L4KYPW1"}, "label": "steel mixer above box on bottom shelf"}]}
{"id": 79887, "image": "COCO_train2014_000000079887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mixer with the blue , grey and white sticker on it\"."}], "task": "refering", "answer_template": "The \"the mixer with the blue , grey and white sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 299, 300, 301, 302], "bbox": [0.5783541666666667, 0.55325, 0.8692708333333333, 0.7818125], "iscrowd": 0, "area": 16341.40465, "rle": {"size": [640, 480], "counts": "US^57ic08H8H8H8H7I8H8G9H8H7I8H8H8H8H7I6J0O2O001O00001O001O000O2O001O00001O001O00001O001O00001O00001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000001O0O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1N2L5L3L4L5L3L4L5L3L4L5L3L4L5L3L4L5L3L6J5L5J5K6K4K6J5L4KYPW1"}, "label": "the mixer with the blue , grey and white sticker on it"}]}
{"id": 79887, "image": "COCO_train2014_000000079887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white kitchenaid mixer\"."}], "task": "refering", "answer_template": "The \"white kitchenaid mixer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 295, 296, 297], "bbox": [0.2803541666666666, 0.5758593750000001, 0.5619583333333333, 0.75953125], "iscrowd": 0, "area": 11264.724300000003, "rle": {"size": [640, 480], "counts": "Qid2m0ob07I6L3L4L3O2N2O0O2O0O101N100@VNi^Ol1Sa0ZNj^Of1Ua0c0O10O0100000O10O00000100O01000O1002N2N2M3N1O2N1O1O2N1O0V_OTM``0m2^_OUMd`0R310000O01O010O0010O00010O3M000001O01O00000001O0003M001O00001O0O101O001RMT_Of2Sa00O10O1000J600000O01O1H70100O01O0EZ_O^Mf`0a2\\_O]Md`0d2__OVMd`0i290O2N1001O1O1N3N1O1O1O1N2O2N1N2O2N1N3M2N2N3N1M4L3M4M2N3K5I7@`0D;N6KQlR4"}, "label": "white kitchenaid mixer"}]}
{"id": 79887, "image": "COCO_train2014_000000079887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stainless steel mixing bowl of the white mixer\"."}], "task": "refering", "answer_template": "The \"the stainless steel mixing bowl of the white mixer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 295, 296, 297], "bbox": [0.2803541666666666, 0.5758593750000001, 0.5619583333333333, 0.75953125], "iscrowd": 0, "area": 11264.724300000003, "rle": {"size": [640, 480], "counts": "Qid2m0ob07I6L3L4L3O2N2O0O2O0O101N100@VNi^Ol1Sa0ZNj^Of1Ua0c0O10O0100000O10O00000100O01000O1002N2N2M3N1O2N1O1O2N1O0V_OTM``0m2^_OUMd`0R310000O01O010O0010O00010O3M000001O01O00000001O0003M001O00001O0O101O001RMT_Of2Sa00O10O1000J600000O01O1H70100O01O0EZ_O^Mf`0a2\\_O]Md`0d2__OVMd`0i290O2N1001O1O1N3N1O1O1O1N2O2N1N2O2N1N3M2N2N3N1M4L3M4M2N3K5I7@`0D;N6KQlR4"}, "label": "the stainless steel mixing bowl of the white mixer"}]}
{"id": 448531, "image": "COCO_train2014_000000448531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange backback with a monkey print sitting on top of a peice of blue luggage\"."}], "task": "refering", "answer_template": "The \"an orange backback with a monkey print sitting on top of a peice of blue luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285], "bbox": [0.6583006535947712, 0.19864379084967318, 0.9979738562091504, 0.5943464052287581], "iscrowd": 0, "area": 39486.32585, "rle": {"size": [612, 612], "counts": "jTa75gb09H7I7H9H7H8I8H7H8I8G8I7I8G8I8G8I7I8G8I7I8N100O2O0O1O2O0O101N101N1O101N100O2N100O2O0O1O2O0O2O0O1O2O0O101N1O101N100O2N100O2O0O2O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O010O000010O0001O01O01O00010O001O01O01O000010O0001O01O01O00010O001O01O01O00010O000010O0000001N100O1O101N100O101N1O100100O1O00100000O1000000O1O2L3N2M3N2M3N2M3N2M4M2M3N2M3N2M3N2K5^M_BZOh=GWC2o<oNQDi0Z?I7H7J7I7H8I7Ia:"}, "label": "an orange backback with a monkey print sitting on top of a peice of blue luggage"}]}
{"id": 448531, "image": "COCO_train2014_000000448531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange colored backpack\"."}], "task": "refering", "answer_template": "The \"an orange colored backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285], "bbox": [0.6583006535947712, 0.19864379084967318, 0.9979738562091504, 0.5943464052287581], "iscrowd": 0, "area": 39486.32585, "rle": {"size": [612, 612], "counts": "jTa75gb09H7I7H9H7H8I8H7H8I8G8I7I8G8I8G8I7I8G8I7I8N100O2O0O1O2O0O101N101N1O101N100O2N100O2O0O1O2O0O2O0O1O2O0O101N1O101N100O2N100O2O0O2O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O010O000010O0001O01O01O00010O001O01O01O000010O0001O01O01O00010O001O01O01O00010O000010O0000001N100O1O101N100O101N1O100100O1O00100000O1000000O1O2L3N2M3N2M3N2M3N2M4M2M3N2M3N2M3N2K5^M_BZOh=GWC2o<oNQDi0Z?I7H7J7I7H8I7Ia:"}, "label": "an orange colored backpack"}]}
{"id": 448531, "image": "COCO_train2014_000000448531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black backpack sits behind a little boy\"."}], "task": "refering", "answer_template": "The \"a black backpack sits behind a little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 176, 177, 178, 179, 198, 199, 200, 201, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267], "bbox": [0.0, 0.2541666666666667, 0.23022875816993466, 0.555326797385621], "iscrowd": 0, "area": 17927.849400000003, "rle": {"size": [612, 612], "counts": "V94ib09H7H9G8I8G8H9H7H9G8I8G8H9H7H9G9H7K6L3N3M2M4M2N3L3N3M1N2O1N3N1O1N200001O001O001O1O001O001O001O001O001O1O001O001O001O00100O001O000000000000000O100000000O10000000000O100000000O100000000O10000000000XMPB;Q>CTB:l=CZB:f=C`B:`=CfB:Z=ClB9U=DQC9o<EVC8j<E\\C8d<EbC8^<EhC8X<EnC7S<FRD8n;FTD:l;CXD<h;A\\D>d;_O_D`0b;]ObDb0^;[OeDe0[;VOkDi0U;SOoDm0Q;nNUEQ1k:jNZEU1g:gN^EX1X>O001O0O2O001O001O1N101O001O0O2O003Ma0_OnUi8"}, "label": "a black backpack sits behind a little boy"}]}
{"id": 448531, "image": "COCO_train2014_000000448531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suitcase with a small blue bag placed on top\"."}], "task": "refering", "answer_template": "The \"a black suitcase with a small blue bag placed on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 176, 177, 178, 179, 198, 199, 200, 201, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267], "bbox": [0.0, 0.2541666666666667, 0.23022875816993466, 0.555326797385621], "iscrowd": 0, "area": 17927.849400000003, "rle": {"size": [612, 612], "counts": "V94ib09H7H9G8I8G8H9H7H9G8I8G8H9H7H9G9H7K6L3N3M2M4M2N3L3N3M1N2O1N3N1O1N200001O001O001O1O001O001O001O001O001O1O001O001O001O00100O001O000000000000000O100000000O10000000000O100000000O100000000O10000000000XMPB;Q>CTB:l=CZB:f=C`B:`=CfB:Z=ClB9U=DQC9o<EVC8j<E\\C8d<EbC8^<EhC8X<EnC7S<FRD8n;FTD:l;CXD<h;A\\D>d;_O_D`0b;]ObDb0^;[OeDe0[;VOkDi0U;SOoDm0Q;nNUEQ1k:jNZEU1g:gN^EX1X>O001O0O2O001O001O1N101O001O0O2O003Ma0_OnUi8"}, "label": "a black suitcase with a small blue bag placed on top"}]}
{"id": 448531, "image": "COCO_train2014_000000448531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a blue shirt and tan pants holding a drink\"."}], "task": "refering", "answer_template": "The \"a little boy in a blue shirt and tan pants holding a drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 71, 72, 73, 93, 94, 95, 115, 116, 117, 136, 137, 138, 139, 140, 158, 159, 160, 161, 162, 163, 179, 180, 181, 182, 183, 184, 185, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 335, 336, 337, 357, 358, 359, 360, 379, 380], "bbox": [0.16179738562091503, 0.12321895424836601, 0.438202614379085, 0.8243464052287581], "iscrowd": 0, "area": 38887.469849999994, "rle": {"size": [612, 612], "counts": "[]k12Pc03L4L3M4L4L3M4L3M4M3L3M4L4L3M4L4M2M4L4L3N3L4L3M4L4M2M4L3YCUMR9o2eF[MW9h2aF`M^9`2ZFjMd9W2TFRNj9o1mEZNQ:g1hEbNV:_1bEjN\\:W1[EROc:o0TE\\O8VMY9_3SFH`0nLZ9[3iE4j0eL[9U7aFPIk6a0hIb6YORIk6`0iI_6YOVIk6>iI^6XOYIl6;iI_6WO[Il6:jI\\6WO_Il67kI\\6UOaIn66iI\\6UObIo65jIn7T6UHiIl7V6WHgIk7V6XHgIj7X6YHeIh7EWF`6S2iIh7FVF^6V2hIf7JUF\\6W2hIf7KTF[6Y2gId7NTFY6^<gIcCV6`<iI`CV6a<jI`CT6j<M3N2N1N10O10O01O10O0100O00100O010O10O02N2O0O2O1N2N2O0jNW1O100O10O0100O10O0100O100O01000O10O0100O100O010O4M3]FTHn60QHP8k0SHP71UHn7e0THT70WHP8?RHW72ZHo78SH[70]HQ82QH^72`Ho7LSH`71dHP8FQHd72eHQ8@QHg71iHm8U7VGkHj8Q7YGoHg8m6^GRIc8i6aGWI_8d6gG[IZ8`6jGaIU8[6PHdIP8R4[F\\Mj1bNl7[3fGYMb0\\Oh7e2QIUM]O5c7Z2PLgMo3X2SLgMn3X2RLhMn3W2TLhMm3V2TLjMl3U2VLkMj3T2VLlMj3S2XLlMi3R2XLnMh3Q2ZLnMf3R2ZLoMf3o1\\LPNd3o1^LPNc3n1iLgMY3U2n6M4L4L3M4L4L3M4L4K4M4L3M3M2N3M2N3M2N3L3N3M2Nm[]6"}, "label": "a little boy in a blue shirt and tan pants holding a drink"}]}
{"id": 448531, "image": "COCO_train2014_000000448531.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy wearing a jacket and holding a juice box\"."}], "task": "refering", "answer_template": "The \"a little boy wearing a jacket and holding a juice box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 71, 72, 73, 93, 94, 95, 115, 116, 117, 136, 137, 138, 139, 140, 158, 159, 160, 161, 162, 163, 179, 180, 181, 182, 183, 184, 185, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 335, 336, 337, 357, 358, 359, 360, 379, 380], "bbox": [0.16179738562091503, 0.12321895424836601, 0.438202614379085, 0.8243464052287581], "iscrowd": 0, "area": 38887.469849999994, "rle": {"size": [612, 612], "counts": "[]k12Pc03L4L3M4L4L3M4L3M4M3L3M4L4L3M4L4M2M4L4L3N3L4L3M4L4M2M4L3YCUMR9o2eF[MW9h2aF`M^9`2ZFjMd9W2TFRNj9o1mEZNQ:g1hEbNV:_1bEjN\\:W1[EROc:o0TE\\O8VMY9_3SFH`0nLZ9[3iE4j0eL[9U7aFPIk6a0hIb6YORIk6`0iI_6YOVIk6>iI^6XOYIl6;iI_6WO[Il6:jI\\6WO_Il67kI\\6UOaIn66iI\\6UObIo65jIn7T6UHiIl7V6WHgIk7V6XHgIj7X6YHeIh7EWF`6S2iIh7FVF^6V2hIf7JUF\\6W2hIf7KTF[6Y2gId7NTFY6^<gIcCV6`<iI`CV6a<jI`CT6j<M3N2N1N10O10O01O10O0100O00100O010O10O02N2O0O2O1N2N2O0jNW1O100O10O0100O10O0100O100O01000O10O0100O100O010O4M3]FTHn60QHP8k0SHP71UHn7e0THT70WHP8?RHW72ZHo78SH[70]HQ82QH^72`Ho7LSH`71dHP8FQHd72eHQ8@QHg71iHm8U7VGkHj8Q7YGoHg8m6^GRIc8i6aGWI_8d6gG[IZ8`6jGaIU8[6PHdIP8R4[F\\Mj1bNl7[3fGYMb0\\Oh7e2QIUM]O5c7Z2PLgMo3X2SLgMn3X2RLhMn3W2TLhMm3V2TLjMl3U2VLkMj3T2VLlMj3S2XLlMi3R2XLnMh3Q2ZLnMf3R2ZLoMf3o1\\LPNd3o1^LPNc3n1iLgMY3U2n6M4L4L3M4L4L3M4L4K4M4L3M3M2N3M2N3M2N3L3N3M2Nm[]6"}, "label": "a little boy wearing a jacket and holding a juice box"}]}
{"id": 79893, "image": "COCO_train2014_000000079893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child with a purple backpack\"."}], "task": "refering", "answer_template": "The \"child with a purple backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 106, 107, 129, 130, 152, 153], "bbox": [0.5627500000000001, 0.0029411764705882353, 0.7045156250000001, 0.43734117647058823], "iscrowd": 0, "area": 7440.111299999998, "rle": {"size": [425, 640], "counts": "l]e43U=5K3M3M3N2M3M2O00000000000O1O100O1O1O0010O0001O000010]2bM101O00001O001O001O00001O001O001O0nG_L\\6e3`I[L`6j3ZIXLe6n3TISLl6R4nHPLQ7V4hHkKX7Z4cHgK\\7a4[HaKd7R501O0ORMYH=f7_O`H?`7\\OfHd0Y7XOmHg0R7TOUIj0k6QO\\In0c6nNbIQ1^6jNiIU1V6gNPJW1P6dNVJ\\1i5_N^J_1b5]NbJd1]5WNhJi1X5UNjJl1U5RNmJn1S5QNnJo1R5oMPKQ2P5mMRKS2n4kMTKV2k4hMVKY2j4eMXK[2h4cMZK2WMo1_7mM\\K4XMl1\\7oM]K5[Mh1X7QN_K7\\Mf1T7QNbK:\\Mb1U:^NnE]1T:cNoEW1T:iNoEQ1S:nNbF<a9BdF8^9GgF4[9JjF0Y9NQ2O_h^2"}, "label": "child with a purple backpack"}]}
{"id": 79893, "image": "COCO_train2014_000000079893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing tights and a purple backpack looking into a window\"."}], "task": "refering", "answer_template": "The \"a little girl wearing tights and a purple backpack looking into a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 106, 107, 129, 130, 152, 153], "bbox": [0.5627500000000001, 0.0029411764705882353, 0.7045156250000001, 0.43734117647058823], "iscrowd": 0, "area": 7440.111299999998, "rle": {"size": [425, 640], "counts": "l]e43U=5K3M3M3N2M3M2O00000000000O1O100O1O1O0010O0001O000010]2bM101O00001O001O001O00001O001O001O0nG_L\\6e3`I[L`6j3ZIXLe6n3TISLl6R4nHPLQ7V4hHkKX7Z4cHgK\\7a4[HaKd7R501O0ORMYH=f7_O`H?`7\\OfHd0Y7XOmHg0R7TOUIj0k6QO\\In0c6nNbIQ1^6jNiIU1V6gNPJW1P6dNVJ\\1i5_N^J_1b5]NbJd1]5WNhJi1X5UNjJl1U5RNmJn1S5QNnJo1R5oMPKQ2P5mMRKS2n4kMTKV2k4hMVKY2j4eMXK[2h4cMZK2WMo1_7mM\\K4XMl1\\7oM]K5[Mh1X7QN_K7\\Mf1T7QNbK:\\Mb1U:^NnE]1T:cNoEW1T:iNoEQ1S:nNbF<a9BdF8^9GgF4[9JjF0Y9NQ2O_h^2"}, "label": "a little girl wearing tights and a purple backpack looking into a window"}]}
{"id": 260118, "image": "COCO_train2014_000000260118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child brushing their teeth\"."}], "task": "refering", "answer_template": "The \"a young child brushing their teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 288, 289, 290, 291, 292, 293, 294, 308, 310, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.285875, 0.15545833333333334, 0.8559374999999999, 1.0], "iscrowd": 0, "area": 95729.4549, "rle": {"size": [480, 640], "counts": "kRf24g>8I8G8I8H8G8I8H8G501N101O0O2O1O1O1N2O1O1N2O1O1O1N1000000O10000O0100000O1000000O10000O1000000O01000O1000000O1000000kF_Mj5b2SJoM_5Q2_JQNa5o1]JSNb5n1[JUNe5k1YJWNg5h1XJZNh5f1UJ]Nj5d1RJ`Nn5`1oIcNQ6]1lIfNS6[1iIiNW6W1fIlNZ6T1cIoN]6Q1`IRO_6o0^ITOb6l0]IUOc6k0[IWOe6i0YIYOg6g0XIZOi6e0UI]Ol6a0SIAn6>RIBn6>RIBo6=PIDQ7;oHER7:mHGS79mHGT78kHIV76jHJW75iHKW75hHLY73gHMZ72eHO\\7OeH1\\7MdH4\\7KeH5\\7IeH7\\7GfH8[7FfH:Z7EhH:Y7]N_G0Z1c1X7XNaG5Y1b1W7TNcG;V1a1W7QNdG?W1_1Z7iMbGh0U1_1R8aNPH^1o7bNRH^1m7bNUH]1j7dNVH\\1i7dNYH[1f7eN[H[1d7eN^HZ1b7eN_H[1a7dNaH[1^7fNbHZ1^7eNdHZ1\\7eNeH[1[7eNfHZ1Y7fNhHZ1X7eNjHZ1V7fNjHZ1U7fNlHZ1T7fNmHY1S7fNnHZ1R7eNPIZ1o6gNQIY1o6gNRIX1n6hNRIX1n6hNSIW1l6kNSIU1m6kNTIT1l6lNTIT1k6mNVIR1j6oNUIQ1Z3jMjMU1lNQ1W3UNeMj0SOR1U3_N_M`0\\OQ1R3jNZM6CP1Q3TOTMLJQ1o2_OmLA4P1l2JhLVO<P1i25cLlNb0P1h2<aLdNg0P1e2d0^L]Nm0o0c2j0\\LXNo0o0b2Q1ZLPNS1P1`2W1XLkMT1Q1a2[1WLfMR1S1d2_1UL`MR1T1f2c1TL[MP1V1j2e1QLWMP1X1m2c1SLWMi0[1S3_1TLXMb0^1X3[1VLYM<a1\\3X1XLYM6c1a3U1YLYM2e1c3S1\\LZMLf1g3Q1\\L\\MHf1j3P1^LV1a3kN_LU1_3lNaLU1]3mNcLS1\\3nNcLS1[3nNeLS1Z3nNfLR1X3POgLQ1X3oNhLR1V3POjLP1T3ROkLo0T3ROkLo0S3ROnLn0Q3SOnLn0P3TOoLm0P3SOQMn0m2SORMn0n2ROQMo0n2RORMn0m2ROSMo0m2QORMP1m2QOSMo0m2QORMP1m2POSMQ1l2POTMP1l2POSMQ1l2b5O^IUMQ1k2b5O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O00001\\J^MhNc2W1^MhNc2V1_MiNa2W1`MhNa2W1aMgN`2W1bMhN_2W1bMhN_2W1bMhN_2V1cMiN^2V1cMiN^2V1cMiN^2U1dMjN]2U1dMjN]2U1dMjN\\2U1fMjN[2U1fMjN[2U1fMjN[2T1gMkNZ2T1gMkNZ2S1hMlNY2S1hMlNY2R1iMmNX2R1iMmNX2Q1jMnNW2P1kMoNU2P1mMoNT2o0nMPOS2o0nMPOS2n0oMQOR2m0PNROQ2l0QNSOP2l0QNSOP2k0RNTOo1j0SNUOn1i0TNVOm1h0UNWOl1h0UNWOk1h0WNWOk1f0WNYOj1e0XNZOj1d0WN[Oj1c0XN\\Oj1a0XN^Oi1`0YN_Oi1?XN@i1>YNAh1=[NAg1<]NAd1=_NAc1=`N@a1>cN_O_1>eN_O\\1?hN^OZ1=kNAV1:QOCQ16WOGj04]OIe01AM`0NGO=IJ4=AJ<=YOJd0=QOJl0=iNIT1?aNHR1h0cN@P1Q1eNVOn0[1gNlNl0l7D<CZgZ1"}, "label": "a young child brushing their teeth"}]}
{"id": 260118, "image": "COCO_train2014_000000260118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy brushing his teeth with a green toothbrush , wearing a striped blue shirt\"."}], "task": "refering", "answer_template": "The \"a young boy brushing his teeth with a green toothbrush , wearing a striped blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 288, 289, 290, 291, 292, 293, 294, 308, 310, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.285875, 0.15545833333333334, 0.8559374999999999, 1.0], "iscrowd": 0, "area": 95729.4549, "rle": {"size": [480, 640], "counts": "kRf24g>8I8G8I8H8G8I8H8G501N101O0O2O1O1O1N2O1O1N2O1O1O1N1000000O10000O0100000O1000000O10000O1000000O01000O1000000O1000000kF_Mj5b2SJoM_5Q2_JQNa5o1]JSNb5n1[JUNe5k1YJWNg5h1XJZNh5f1UJ]Nj5d1RJ`Nn5`1oIcNQ6]1lIfNS6[1iIiNW6W1fIlNZ6T1cIoN]6Q1`IRO_6o0^ITOb6l0]IUOc6k0[IWOe6i0YIYOg6g0XIZOi6e0UI]Ol6a0SIAn6>RIBn6>RIBo6=PIDQ7;oHER7:mHGS79mHGT78kHIV76jHJW75iHKW75hHLY73gHMZ72eHO\\7OeH1\\7MdH4\\7KeH5\\7IeH7\\7GfH8[7FfH:Z7EhH:Y7]N_G0Z1c1X7XNaG5Y1b1W7TNcG;V1a1W7QNdG?W1_1Z7iMbGh0U1_1R8aNPH^1o7bNRH^1m7bNUH]1j7dNVH\\1i7dNYH[1f7eN[H[1d7eN^HZ1b7eN_H[1a7dNaH[1^7fNbHZ1^7eNdHZ1\\7eNeH[1[7eNfHZ1Y7fNhHZ1X7eNjHZ1V7fNjHZ1U7fNlHZ1T7fNmHY1S7fNnHZ1R7eNPIZ1o6gNQIY1o6gNRIX1n6hNRIX1n6hNSIW1l6kNSIU1m6kNTIT1l6lNTIT1k6mNVIR1j6oNUIQ1Z3jMjMU1lNQ1W3UNeMj0SOR1U3_N_M`0\\OQ1R3jNZM6CP1Q3TOTMLJQ1o2_OmLA4P1l2JhLVO<P1i25cLlNb0P1h2<aLdNg0P1e2d0^L]Nm0o0c2j0\\LXNo0o0b2Q1ZLPNS1P1`2W1XLkMT1Q1a2[1WLfMR1S1d2_1UL`MR1T1f2c1TL[MP1V1j2e1QLWMP1X1m2c1SLWMi0[1S3_1TLXMb0^1X3[1VLYM<a1\\3X1XLYM6c1a3U1YLYM2e1c3S1\\LZMLf1g3Q1\\L\\MHf1j3P1^LV1a3kN_LU1_3lNaLU1]3mNcLS1\\3nNcLS1[3nNeLS1Z3nNfLR1X3POgLQ1X3oNhLR1V3POjLP1T3ROkLo0T3ROkLo0S3ROnLn0Q3SOnLn0P3TOoLm0P3SOQMn0m2SORMn0n2ROQMo0n2RORMn0m2ROSMo0m2QORMP1m2QOSMo0m2QORMP1m2POSMQ1l2POTMP1l2POSMQ1l2b5O^IUMQ1k2b5O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O00001\\J^MhNc2W1^MhNc2V1_MiNa2W1`MhNa2W1aMgN`2W1bMhN_2W1bMhN_2W1bMhN_2V1cMiN^2V1cMiN^2V1cMiN^2U1dMjN]2U1dMjN]2U1dMjN\\2U1fMjN[2U1fMjN[2U1fMjN[2T1gMkNZ2T1gMkNZ2S1hMlNY2S1hMlNY2R1iMmNX2R1iMmNX2Q1jMnNW2P1kMoNU2P1mMoNT2o0nMPOS2o0nMPOS2n0oMQOR2m0PNROQ2l0QNSOP2l0QNSOP2k0RNTOo1j0SNUOn1i0TNVOm1h0UNWOl1h0UNWOk1h0WNWOk1f0WNYOj1e0XNZOj1d0WN[Oj1c0XN\\Oj1a0XN^Oi1`0YN_Oi1?XN@i1>YNAh1=[NAg1<]NAd1=_NAc1=`N@a1>cN_O_1>eN_O\\1?hN^OZ1=kNAV1:QOCQ16WOGj04]OIe01AM`0NGO=IJ4=AJ<=YOJd0=QOJl0=iNIT1?aNHR1h0cN@P1Q1eNVOn0[1gNlNl0l7D<CZgZ1"}, "label": "a young boy brushing his teeth with a green toothbrush , wearing a striped blue shirt"}]}
{"id": 161818, "image": "COCO_train2014_000000161818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person behind the white colored shirt wearing person\"."}], "task": "refering", "answer_template": "The \"the person behind the white colored shirt wearing person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 125, 148, 171, 193, 194, 215, 216, 217, 238, 239, 240, 241, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 310, 327, 328, 329, 330, 331, 333], "bbox": [0.2095625, 0.2948946135831382, 0.5256093749999999, 1.0], "iscrowd": 0, "area": 15477.305199999995, "rle": {"size": [427, 640], "counts": "QTi11Z=0000000O10000000000O100000000O100000000O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O10000O01000O0100O01000O010O01000O01000O010O10O10OgNXOQFh0l9[OSFe0k9_OTFa0h9DWF<f9GYF9e9KZF5b90]FOa95]FL`97`FI]9;bFEZ9`0dFAY9b0gF^OV9f0hF[OT9j0kFVOR9m0nFROP9R1nFoNo8U1fFUOV9b2O1O1O1N2O1O001O1O1N2O1O1O1O1O001N100000O100000O10O10O1O1O1O1O010O1O1O1O001O1O000O0O2O1N2N2N2N1O2N2O1N2YNnJUKS5m4SKjJo4W5WK`Jn4]5YKZJl4d5ZKSJk4k5`KgIe4V6Z1N2N21O1nKSJc0n5TO_Jf0c5QOiJk0X5mNSKo0n4hN_KT1a4cNkKY1V4^NWL]1j3[NaLa1`3VNlLg1T3PNYMk1h2iMgMS2n6K4M[FlMR8R2kGPNW8o1fGSN\\8l1aGVN`8k1\\GWNf8h1WGZNk8e1RG]NP9b1mF`NU9`1gFbNZ9^1cFdN_9[1^FgNd9X1YFjNi9V1SFlNo9S1nEoNS:Q1jEQO]:i0`EYOf:c17I7I6I8I7H8I6J7H7J6I8I_[n3"}, "label": "the person behind the white colored shirt wearing person"}]}
{"id": 161818, "image": "COCO_train2014_000000161818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in green\"."}], "task": "refering", "answer_template": "The \"the person in green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 125, 148, 171, 193, 194, 215, 216, 217, 238, 239, 240, 241, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 310, 327, 328, 329, 330, 331, 333], "bbox": [0.2095625, 0.2948946135831382, 0.5256093749999999, 1.0], "iscrowd": 0, "area": 15477.305199999995, "rle": {"size": [427, 640], "counts": "QTi11Z=0000000O10000000000O100000000O100000000O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O10000O01000O0100O01000O010O01000O01000O010O10O10OgNXOQFh0l9[OSFe0k9_OTFa0h9DWF<f9GYF9e9KZF5b90]FOa95]FL`97`FI]9;bFEZ9`0dFAY9b0gF^OV9f0hF[OT9j0kFVOR9m0nFROP9R1nFoNo8U1fFUOV9b2O1O1O1N2O1O001O1O1N2O1O1O1O1O001N100000O100000O10O10O1O1O1O1O010O1O1O1O001O1O000O0O2O1N2N2N2N1O2N2O1N2YNnJUKS5m4SKjJo4W5WK`Jn4]5YKZJl4d5ZKSJk4k5`KgIe4V6Z1N2N21O1nKSJc0n5TO_Jf0c5QOiJk0X5mNSKo0n4hN_KT1a4cNkKY1V4^NWL]1j3[NaLa1`3VNlLg1T3PNYMk1h2iMgMS2n6K4M[FlMR8R2kGPNW8o1fGSN\\8l1aGVN`8k1\\GWNf8h1WGZNk8e1RG]NP9b1mF`NU9`1gFbNZ9^1cFdN_9[1^FgNd9X1YFjNi9V1SFlNo9S1nEoNS:Q1jEQO]:i0`EYOf:c17I7I6I8I7H8I6J7H7J6I8I_[n3"}, "label": "the person in green"}]}
{"id": 161818, "image": "COCO_train2014_000000161818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white shirt and black shorts is catching a frisbee\"."}], "task": "refering", "answer_template": "The \"the man in the white shirt and black shorts is catching a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 264, 265, 266, 267, 268, 287, 288, 289, 290, 310, 311, 312, 313], "bbox": [0.45428125, 0.09662763466042154, 0.6626875, 0.9236065573770492], "iscrowd": 0, "area": 28604.895899999996, "rle": {"size": [427, 640], "counts": "kbi35R=8H7I7I7I7WOhNUE_1c:kNTE\\1d:k0H8H8H8H8H8H8H8H9G6L3N2N2O4K6J6K5J6J6K4K6QOkI^J[6Z5lIbJY6V5oIdJW6T5PJgJV6Q5X1J6hKiGT3W910O01O001O001O001O001O001O001OE;D<I8H7H8J6000000000mNeMVGZ2d8RNWGn1g8YNTGg1m8]NnFc1S9bNgF^1Z9fNaFZ1`9jN[FV1g9V11N3N1O2N1O2N2N1O2M2O2N2N1O2N1O2N2M2O0>B1O2XJoLS1R3fNUMX1m2bNZM[1f2`NaM^1_2^NhMa0QObLW3h2oMd0lNcLT3e2WNe0fNeLS3c2\\Nf0cNeLQ3b2bNe0`NgLn2a2gNf0\\NhLm2_2kNh0ZNgLj2_2QOh0WNhLh2\\2UOk0TNhLl2U2UOQ1QNfLh3Z1[No1oMcLj3Y1\\NR2kMbLn3V1[NV2jM`Lo3T1]NZ2_2_MhM_2Z2ZMlMe2W2TMoMF\\Lj2g5YMSNJ\\Li2c5VMXNMZLj2W:L4K4M4L4L3M4L3M4L4K5L3M4K5J6K5K4K6K5K5Jigi2"}, "label": "the man in the white shirt and black shorts is catching a frisbee"}]}
{"id": 333855, "image": "COCO_train2014_000000333855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boston red sox player holding a black baseball bat\"."}], "task": "refering", "answer_template": "The \"a boston red sox player holding a black baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 52, 53, 54, 55, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330], "bbox": [0.07796875, 0.09662763466042154, 0.61171875, 0.9842622950819672], "iscrowd": 0, "area": 77483.5277, "rle": {"size": [427, 640], "counts": "XTe03S=6K5J6J6J6J5L5J6J6J6J6K5J6M3N2M3M3M2N3N2M3M3M3N2M3M3M3N2M3M2N3M3N2M3M3M3N2M3M3M3N2M2N3M3M3N2M3M3M3N2M3M3M3M201N2O1N2O1N2O1N2N2O1N2O0O2O1N2VJcIc4_6YKhIc4Y6ZKmIa4U6[KRJa4o5[KYJ_4i5^K]J^4d5^KcJFdNe4k6aKfJ_4[5]KhJb4Y5\\KhJd4P700O100O100O010O1O100O100O010O100O1O10O0100O100O100O00100O100O10O0100O1O100O10O0100O100O1E;C<E<M3O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N101001O0000000O2O000000001N10000000001O0O1000001O00000O101O0000000O2O000000001O0O1000001O00000O10001O00000O101O1O4L5K4L5J5L5K4L5K4L4L5J5L5K4L5K4L0O2O00001O001O0O2O00001O001N101O00001O001N101O000ZMlJUMU5j2gKZLY4g3cL\\K^3c4e200010O01O0010O00010O01O0010O0001O010O0010O0001O001O000O2O001N101N10002M2O2M2O2N2M2O2M2O2N2M2O2N1N3N2M2O2N1N3N2M2O2N1N3N1O2M3N1N3N1O2M3N1N3N1O2M3N1O2M2O2K5J5K6J5J7J6J5K6J5J7J6J5K[cW3"}, "label": "a boston red sox player holding a black baseball bat"}]}
{"id": 333855, "image": "COCO_train2014_000000333855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player holding a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player holding a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 52, 53, 54, 55, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330], "bbox": [0.07796875, 0.09662763466042154, 0.61171875, 0.9842622950819672], "iscrowd": 0, "area": 77483.5277, "rle": {"size": [427, 640], "counts": "XTe03S=6K5J6J6J6J5L5J6J6J6J6K5J6M3N2M3M3M2N3N2M3M3M3N2M3M3M3N2M3M2N3M3N2M3M3M3N2M3M3M3N2M2N3M3M3N2M3M3M3N2M3M3M3M201N2O1N2O1N2O1N2N2O1N2O0O2O1N2VJcIc4_6YKhIc4Y6ZKmIa4U6[KRJa4o5[KYJ_4i5^K]J^4d5^KcJFdNe4k6aKfJ_4[5]KhJb4Y5\\KhJd4P700O100O100O010O1O100O100O010O100O1O10O0100O100O100O00100O100O10O0100O1O100O10O0100O100O1E;C<E<M3O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N101001O0000000O2O000000001N10000000001O0O1000001O00000O101O0000000O2O000000001O0O1000001O00000O10001O00000O101O1O4L5K4L5J5L5K4L5K4L4L5J5L5K4L5K4L0O2O00001O001O0O2O00001O001N101O00001O001N101O000ZMlJUMU5j2gKZLY4g3cL\\K^3c4e200010O01O0010O00010O01O0010O0001O010O0010O0001O001O000O2O001N101N10002M2O2M2O2N2M2O2M2O2N2M2O2N1N3N2M2O2N1N3N2M2O2N1N3N1O2M3N1N3N1O2M3N1N3N1O2M3N1O2M2O2K5J5K6J5J7J6J5K6J5J7J6J5K[cW3"}, "label": "a baseball player holding a bat"}]}
{"id": 309280, "image": "COCO_train2014_000000309280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant\"."}], "task": "refering", "answer_template": "The \"a baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 293, 294, 295, 296, 307, 308, 309, 310, 313, 314, 315, 316, 317, 318, 319, 320, 333, 336, 338, 339, 340, 342], "bbox": [0.35921875000000003, 0.4414319248826291, 0.9252343749999999, 0.967488262910798], "iscrowd": 0, "area": 48092.10810000001, "rle": {"size": [426, 640], "counts": "oQP31X=1O1O1O1N2O1O1O1O1O1N2N2N2N2N2N3L3L4H8H8I7M3N2MKfD`NW;b1jD^NT;c1nD\\No:f1SEYNk:h1WEWNf:k1\\ETNa:n1aEQN]:P2eEoMX:S2>1N3M2H8H8G9M3O1N2N2N2N2N2N3M2N2N2O11O0O1N3N1N3N1N2O2M200O2O001N100O2O1O1N2O1O1O1:F:[GVMV6U3YIQMg6Z3iHkLW7d4LYJeHg5[7ZJdHf5[7\\JdHd5\\7500000000O10000000000O1000000000O01000O1B>D<D<D;H9O1O100O1O001O100O1O1O001O100O1O1O00100O1O1O1O00100O1O1O1O010O11O3M3M4K4M3M2N2M3N2N2N2M3N2N3L3N2N2N2gHRKQ6Q5kIRKT6P5hIRKX6P5dIRK[6Q5aIQK_6Q5]IQKc6Q5YIQKg6R5TIPKk6S5QIoJo6f500O1000000O2O000O1000000O10000O1000000O0100O0cMVI\\Nj6d1_IRNb6n1fIjMZ6V2oI`MQ6a2XJUMi5k2QK[Lo4f3UKTLk4m3YKoKg4Q4Y2000O10000O1000000O10001N10001O0O2O001N10001O010O00010O0010O01O00010_IbKT4^4jKdKV4]4fKfK[4Y4bKjK^4W4_KkKa4U4\\KnKe4Q4XKRLh4o3UKSLk4m3RKVLn4k3nJXLS5i3iJYLW5j3dJXL\\5k3_JWLb5k3XJXLh5l3RJVLn5o3kISLV6R4bIPL^6^5001O0000001O00001O00001O0000001O00001O00001O00001O0000001O0N3M2N3M2N3M2N3M2N3I6H9G8H9G8N30O01O001O010O001O010O001O010O001O0010O010O2O11N4M4L3M3M3M3M3L5L3M3M3M3M2N3L3N3M3M2N3M2N3M3L3M2L3N2M3M3M3N2M3M3M3N2M3M3M3N2M3M3N2M3M3M3N1N3M3Mmbc0"}, "label": "a baby elephant"}]}
{"id": 309280, "image": "COCO_train2014_000000309280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby elephant\"."}], "task": "refering", "answer_template": "The \"baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 293, 294, 295, 296, 307, 308, 309, 310, 313, 314, 315, 316, 317, 318, 319, 320, 333, 336, 338, 339, 340, 342], "bbox": [0.35921875000000003, 0.4414319248826291, 0.9252343749999999, 0.967488262910798], "iscrowd": 0, "area": 48092.10810000001, "rle": {"size": [426, 640], "counts": "oQP31X=1O1O1O1N2O1O1O1O1O1N2N2N2N2N2N3L3L4H8H8I7M3N2MKfD`NW;b1jD^NT;c1nD\\No:f1SEYNk:h1WEWNf:k1\\ETNa:n1aEQN]:P2eEoMX:S2>1N3M2H8H8G9M3O1N2N2N2N2N2N3M2N2N2O11O0O1N3N1N3N1N2O2M200O2O001N100O2O1O1N2O1O1O1:F:[GVMV6U3YIQMg6Z3iHkLW7d4LYJeHg5[7ZJdHf5[7\\JdHd5\\7500000000O10000000000O1000000000O01000O1B>D<D<D;H9O1O100O1O001O100O1O1O001O100O1O1O00100O1O1O1O00100O1O1O1O010O11O3M3M4K4M3M2N2M3N2N2N2M3N2N3L3N2N2N2gHRKQ6Q5kIRKT6P5hIRKX6P5dIRK[6Q5aIQK_6Q5]IQKc6Q5YIQKg6R5TIPKk6S5QIoJo6f500O1000000O2O000O1000000O10000O1000000O0100O0cMVI\\Nj6d1_IRNb6n1fIjMZ6V2oI`MQ6a2XJUMi5k2QK[Lo4f3UKTLk4m3YKoKg4Q4Y2000O10000O1000000O10001N10001O0O2O001N10001O010O00010O0010O01O00010_IbKT4^4jKdKV4]4fKfK[4Y4bKjK^4W4_KkKa4U4\\KnKe4Q4XKRLh4o3UKSLk4m3RKVLn4k3nJXLS5i3iJYLW5j3dJXL\\5k3_JWLb5k3XJXLh5l3RJVLn5o3kISLV6R4bIPL^6^5001O0000001O00001O00001O0000001O00001O00001O00001O0000001O0N3M2N3M2N3M2N3M2N3I6H9G8H9G8N30O01O001O010O001O010O001O010O001O0010O010O2O11N4M4L3M3M3M3M3L5L3M3M3M3M2N3L3N3M3M2N3M2N3M3L3M2L3N2M3M3M3N2M3M3M3N2M3M3M3N2M3M3N2M3M3M3N1N3M3Mmbc0"}, "label": "baby elephant"}]}
{"id": 309280, "image": "COCO_train2014_000000309280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult elephant with its trunk through a fence\"."}], "task": "refering", "answer_template": "The \"an adult elephant with its trunk through a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 305, 306, 307, 330], "bbox": [0.24231250000000001, 0.2900469483568075, 0.670109375, 0.9439671361502348], "iscrowd": 0, "area": 34483.67095, "rle": {"size": [426, 640], "counts": "eiP26m<;oKCeJj0Q5_OdJm0S5[ObJP1U5YO`JS1W5UO^JW1X5SO\\JX1[5QOZJ[1]5mNWJ`1_5iNVJ^1i5gNlI`1R6fNcIa1\\6g2N2O0O2O1N2O1N10^OoImIo5P6WJoIh5m5_JQJ`5k5gJSJY5h5mJWJS5d5TKYJl4d5ZKZJf4b5`K\\J`4a5dK^J\\4^5jK`JV4\\5n1L3M4L4L4M3L3O2O1N2O1N2N10]JPLn1n3iL]MW3b2_KiNa4P501O0O10001O00000O2O0000001N1000001O000O101O00000O2O0000001O0O10hNX1^Ob0B>A>C>B>F:00001N1000001O0O10001O0000001O0000001O0000001O0000O1M3L4M3L4M3L4M3L4M3L4M3L4M2M4M3L4M3000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O10001O000O1000000O101O00000O2O00000O101O00000O2O0000001N10001O0O2O1O1O1O1N2O1O1O1N101O1O1O1N2O1O1O1O1N101O1O1N2O1O1O2N2M3N2N2N2M3N2N2N2N2M2O2N2N2N2M2O2N2Nobg2"}, "label": "an adult elephant with its trunk through a fence"}]}
{"id": 309280, "image": "COCO_train2014_000000309280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult elephant\"."}], "task": "refering", "answer_template": "The \"the adult elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 305, 306, 307, 330], "bbox": [0.24231250000000001, 0.2900469483568075, 0.670109375, 0.9439671361502348], "iscrowd": 0, "area": 34483.67095, "rle": {"size": [426, 640], "counts": "eiP26m<;oKCeJj0Q5_OdJm0S5[ObJP1U5YO`JS1W5UO^JW1X5SO\\JX1[5QOZJ[1]5mNWJ`1_5iNVJ^1i5gNlI`1R6fNcIa1\\6g2N2O0O2O1N2O1N10^OoImIo5P6WJoIh5m5_JQJ`5k5gJSJY5h5mJWJS5d5TKYJl4d5ZKZJf4b5`K\\J`4a5dK^J\\4^5jK`JV4\\5n1L3M4L4L4M3L3O2O1N2O1N2N10]JPLn1n3iL]MW3b2_KiNa4P501O0O10001O00000O2O0000001N1000001O000O101O00000O2O0000001O0O10hNX1^Ob0B>A>C>B>F:00001N1000001O0O10001O0000001O0000001O0000001O0000O1M3L4M3L4M3L4M3L4M3L4M3L4M2M4M3L4M3000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O10001O000O1000000O101O00000O2O00000O101O00000O2O0000001N10001O0O2O1O1O1O1N2O1O1O1N101O1O1O1N2O1O1O1O1N101O1O1N2O1O1O2N2M3N2N2N2M3N2N2N2N2M2O2N2N2N2M2O2N2Nobg2"}, "label": "the adult elephant"}]}
{"id": 260129, "image": "COCO_train2014_000000260129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person holding green skateboard\"."}], "task": "refering", "answer_template": "The \"person holding green skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 240, 241, 260, 261, 263, 264, 283, 284, 286, 287, 306, 307, 309, 310], "bbox": [0.30203125000000003, 0.15088992974238877, 0.5364531250000001, 0.9144028103044496], "iscrowd": 0, "area": 23927.525800000003, "rle": {"size": [427, 640], "counts": "mj`25U=1N2O1N2O1O1N2O1oL:\\HGb7o1hFSNW9Q3N2O1N2O1N1O2O1N2O1N2O1N2O1N2O1N2O1N2N101N2O1N2O1N2O1N2jMUKkKl4V4VKeKl4Z4XKbKi4]4[K^Kg4`4]K]Kd4a4aKZKS3POXMf5IUKi2FUMS55UK`28RMb4?TKY2j0PMP4i0UKP2V7QNhHj1]7YNaHg1]7\\NaHd1]7`NbH_1\\7eNcH[1[7hNcHX1\\7kNcHU1]7lNbHS1`7lN_HT1b7mN]HS1c7nN\\HQ1f7nNYHS1g7mNYH2h8NUG1P9NmF1W9OeF4\\9KaF9_9G^F<c9e10O2O1O1O001N20\\OdLXG_3c8lLUGW3d8l0J6L2N2O2N1O1O01O10O0100000O10000000000000000]L`K[Na4c1`K\\Nf4]1\\KcNi4W1XKjNm4KjJkM:Z2Q5SO_KZNBc2T5ZNTLlNiNk2]5ZMeLCoMS3^9001O001O1O1O001O2N4L3M3M3M?A1O1O1N2O1O1N2O1N2M3M3M3M3M3N2XLVKkNm4n0mK\\NV4Q1oLZNT3o0UN]Nm1o0ZO]Ng0[13RNNj1o5J6I8I6I7J7F9G`fk3"}, "label": "person holding green skateboard"}]}
{"id": 260129, "image": "COCO_train2014_000000260129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl holding a green skateboard\"."}], "task": "refering", "answer_template": "The \"a girl holding a green skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 240, 241, 260, 261, 263, 264, 283, 284, 286, 287, 306, 307, 309, 310], "bbox": [0.30203125000000003, 0.15088992974238877, 0.5364531250000001, 0.9144028103044496], "iscrowd": 0, "area": 23927.525800000003, "rle": {"size": [427, 640], "counts": "mj`25U=1N2O1N2O1O1N2O1oL:\\HGb7o1hFSNW9Q3N2O1N2O1N1O2O1N2O1N2O1N2O1N2O1N2O1N2N101N2O1N2O1N2O1N2jMUKkKl4V4VKeKl4Z4XKbKi4]4[K^Kg4`4]K]Kd4a4aKZKS3POXMf5IUKi2FUMS55UK`28RMb4?TKY2j0PMP4i0UKP2V7QNhHj1]7YNaHg1]7\\NaHd1]7`NbH_1\\7eNcH[1[7hNcHX1\\7kNcHU1]7lNbHS1`7lN_HT1b7mN]HS1c7nN\\HQ1f7nNYHS1g7mNYH2h8NUG1P9NmF1W9OeF4\\9KaF9_9G^F<c9e10O2O1O1O001N20\\OdLXG_3c8lLUGW3d8l0J6L2N2O2N1O1O01O10O0100000O10000000000000000]L`K[Na4c1`K\\Nf4]1\\KcNi4W1XKjNm4KjJkM:Z2Q5SO_KZNBc2T5ZNTLlNiNk2]5ZMeLCoMS3^9001O001O1O1O001O2N4L3M3M3M?A1O1O1N2O1O1N2O1N2M3M3M3M3M3N2XLVKkNm4n0mK\\NV4Q1oLZNT3o0UN]Nm1o0ZO]Ng0[13RNNj1o5J6I8I6I7J7F9G`fk3"}, "label": "a girl holding a green skateboard"}]}
{"id": 505895, "image": "COCO_train2014_000000505895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with leaves on top that is below another pizza\"."}], "task": "refering", "answer_template": "The \"a pizza with leaves on top that is below another pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0016875000000000002, 0.017979166666666668, 0.52246875, 0.9887708333333334], "iscrowd": 0, "area": 130382.08090000002, "rle": {"size": [480, 640], "counts": "``0Z=a15000000000000000000000000000000000000O1000000000000O100000000O1000000O10000O1000000O1000000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O100O1000000O10000O10000O1000000O1000000000000000000O1000000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O1000000000000000000000000000000000000001O00000000000000001O000000001O00001O0000001O000010O000001O00001O0000001O001O00001O000000001O0O1O1O100O2N1O100O1O2N100O1O1O2O0O1O1O100O2N1O100O1O2N100O1O1O2O0O1O100O2N1O100O1O2N100O1O2N1O2N1N3N2N1N3N1N3N2N2M3O0O2N2O1N2O0O2N2WKlMdLU2S3ZNeLh1S3fNfLZ1T3SOdLo0U3_OcLb0W3L_L6[3a0RLAi3l0nKUOm3Q1PLPOm3V1PLkNj3]1RLeNh3c1UL^Ng3g1oKbNn3Z6M3M2N3L4M3M3M2N3M3M3M2M4M3M3M3M3M3M3M3M3N2M3N2M3N2M3M3M3K5L4K5L5K4L4L5I6J7L3M3L5IT[_4"}, "label": "a pizza with leaves on top that is below another pizza"}]}
{"id": 505895, "image": "COCO_train2014_000000505895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fresh pizza topped with veggies\"."}], "task": "refering", "answer_template": "The \"a fresh pizza topped with veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0016875000000000002, 0.017979166666666668, 0.52246875, 0.9887708333333334], "iscrowd": 0, "area": 130382.08090000002, "rle": {"size": [480, 640], "counts": "``0Z=a15000000000000000000000000000000000000O1000000000000O100000000O1000000O10000O1000000O1000000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O100O1000000O10000O10000O1000000O1000000000000000000O1000000000000O100000000O100000000O100000000O10000000000O100000000O100000000O100000000O1000000000000000000000000000000000000001O00000000000000001O000000001O00001O0000001O000010O000001O00001O0000001O001O00001O000000001O0O1O1O100O2N1O100O1O2N100O1O1O2O0O1O1O100O2N1O100O1O2N100O1O1O2O0O1O100O2N1O100O1O2N100O1O2N1O2N1N3N2N1N3N1N3N2N2M3O0O2N2O1N2O0O2N2WKlMdLU2S3ZNeLh1S3fNfLZ1T3SOdLo0U3_OcLb0W3L_L6[3a0RLAi3l0nKUOm3Q1PLPOm3V1PLkNj3]1RLeNh3c1UL^Ng3g1oKbNn3Z6M3M2N3L4M3M3M2N3M3M3M2M4M3M3M3M3M3M3M3M3N2M3N2M3N2M3M3M3K5L4K5L5K4L4L5I6J7L3M3L5IT[_4"}, "label": "a fresh pizza topped with veggies"}]}
{"id": 505895, "image": "COCO_train2014_000000505895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with cheese , basil and sauce next to another pizza with greens and olives\"."}], "task": "refering", "answer_template": "The \"a pizza with cheese , basil and sauce next to another pizza with greens and olives\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.458421875, 0.0022500000000000003, 1.0, 1.0], "iscrowd": 0, "area": 149363.0201, "rle": {"size": [480, 640], "counts": "dcY42g>?A`0@?A?A?A`0@?B>D=D;D<D<D<E7H7I8H8H7J7H6J6K5J6J5L5J6K5J6J6K5J6J5L5J6K5J6J6K5J5L5J6J6K5J6J6K4K6J6K5J6J6K4L4N2O1N2N2N2O1N2N2O1N2N2O0O1O1O100O1O100O1O010O1O1O100O1O100O1O1O100O1O100O1O100O1O1O010O1O100O1O100O1O1000000O100000000O100000O01000000O1000000O1000000O1000000O1000000O0100000O1000000O100000000O1000000O1000000O0100000O1000000O1000000O1000000O1000000O1000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000OoEQ:"}, "label": "a pizza with cheese , basil and sauce next to another pizza with greens and olives"}]}
{"id": 505895, "image": "COCO_train2014_000000505895.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza on the right with mozarella\"."}], "task": "refering", "answer_template": "The \"pizza on the right with mozarella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.458421875, 0.0022500000000000003, 1.0, 1.0], "iscrowd": 0, "area": 149363.0201, "rle": {"size": [480, 640], "counts": "dcY42g>?A`0@?A?A?A`0@?B>D=D;D<D<D<E7H7I8H8H7J7H6J6K5J6J5L5J6K5J6J6K5J6J5L5J6K5J6J6K5J5L5J6J6K5J6J6K4K6J6K5J6J6K4L4N2O1N2N2N2O1N2N2O1N2N2O0O1O1O100O1O100O1O010O1O1O100O1O100O1O1O100O1O100O1O100O1O1O010O1O100O1O100O1O1000000O100000000O100000O01000000O1000000O1000000O1000000O1000000O0100000O1000000O100000000O1000000O1000000O0100000O1000000O1000000O1000000O1000000O1000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000OoEQ:"}, "label": "pizza on the right with mozarella"}]}
{"id": 399408, "image": "COCO_train2014_000000399408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hotdog with the mustard in the middle of it\"."}], "task": "refering", "answer_template": "The \"the hotdog with the mustard in the middle of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219], "bbox": [0.10040625, 0.5311944444444444, 0.6169374999999999, 0.7161388888888889], "iscrowd": 0, "area": 17548.80875, "rle": {"size": [360, 640], "counts": "^gf0;m:1O1O1O1O1O1N2N1O1O1O100O2N1O1O1001O000000001O000000001O0001O0001O000000001O000000000000000000000000A?M3L4L4K5N2000000O1000000000000O1000000000000000000O100000000000000000000000000O1000000000000000000000000O100000000O10000000000O10000000000O10000000000O100000000O100000000000000001O0000000000000000000000000000000000000O10001O0000000000000000000000000000000000000O100000000000000000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000000O101O0000000000000000000000001O0O100000000000001O00000000000000001O0000000000000O101O0000000O100000001O0O10000000001O000000001O000000001O000000001O000001O01O0000001O2N3M3M3K6K4K6K4KTme2"}, "label": "the hotdog with the mustard in the middle of it"}]}
{"id": 399408, "image": "COCO_train2014_000000399408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog with ketchup on top\"."}], "task": "refering", "answer_template": "The \"a hot dog with ketchup on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 164, 165, 166, 173], "bbox": [0.13093749999999998, 0.35047222222222224, 0.5846093750000001, 0.6321944444444445], "iscrowd": 0, "area": 16905.9814, "rle": {"size": [360, 640], "counts": "Wfm07o:3[EMR:5jE1R:1jE4S:c0M4L4M3L3N1O2M2O2F[NmFf1P9<L2O2N2N101N1O2N2O0O2N2N101N10I7I7H8N200O001000000O10O100000O100000O1000O100000O10O10000000O010000000000O010000000O10000000000O10000000O01000000000000000000000000O10000000O100000000000000000000000O1000000000O10000000000000000000O10000000000000O1000000000001O01O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O0000000001O000001O000000000000001O0000000000000010O0000000000000001O0000000000001O000000010O000000001O000000001O0000000010O000001O000000001O01O000001O00001O001O010O001O001O1O1N4K6J2O1N2GkE]OX:b07O1O1O100N2O1N3N2Lb[m2"}, "label": "a hot dog with ketchup on top"}]}
{"id": 399408, "image": "COCO_train2014_000000399408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hot dog furthest from camera\"."}], "task": "refering", "answer_template": "The \"hot dog furthest from camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 164, 165, 166, 173], "bbox": [0.13093749999999998, 0.35047222222222224, 0.5846093750000001, 0.6321944444444445], "iscrowd": 0, "area": 16905.9814, "rle": {"size": [360, 640], "counts": "Wfm07o:3[EMR:5jE1R:1jE4S:c0M4L4M3L3N1O2M2O2F[NmFf1P9<L2O2N2N101N1O2N2O0O2N2N101N10I7I7H8N200O001000000O10O100000O100000O1000O100000O10O10000000O010000000000O010000000O10000000000O10000000O01000000000000000000000000O10000000O100000000000000000000000O1000000000O10000000000000000000O10000000000000O1000000000001O01O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O0000000001O000001O000000000000001O0000000000000010O0000000000000001O0000000000001O000000010O000000001O000000001O0000000010O000001O000000001O01O000001O00001O001O010O001O001O1O1N4K6J2O1N2GkE]OX:b07O1O1O100N2O1N3N2Lb[m2"}, "label": "hot dog furthest from camera"}]}
{"id": 571441, "image": "COCO_train2014_000000571441.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big elephant\"."}], "task": "refering", "answer_template": "The \"a big elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208], "bbox": [0.050671875, 0.2837777777777778, 0.32559375, 0.7071944444444445], "iscrowd": 0, "area": 20053.648549999994, "rle": {"size": [360, 640], "counts": "T];3Q;j0UOc0^O;_FZOn7S2J5M4N2M2O2N2N0O2O1O1O1N101O1O1O1O00I72N3M2NnN[IeMb6[2bIcM\\6\\2jIaMU6]2PJ`Mo5`2VJZMk5e2[JUMg5j2[1100O100O100O100O100O100O2O001O0O2O1O3M2N3M3M2N3M2N2N2N2N2N2N2N2M3N2N2N001O1O001N100000cNbIPN^6n1mIiMS6U2VJdMj5Z2YJeMg5[2[JbMf5]2]JaMc5_2^J`Mb5`2`J^M`5a2bJ^M^5b2dJ\\M\\5d2eJ[M[5e2fJZMZ5e2hJZMX5f2e100001O1O6K6I6Ja0_O<D6J001O000000000001O000O100O100O10OhNY1N2N2N22N2lHlLP6W3kIlLT6V3iIlLV6W3fIkLY6W3cIlL\\6W3`IkL_6W3ZInLf6j300000000000001OO1TNYIWOg6:jIDV6<mIAS6?PJ]OQ6c0QJ[Oo5e0TJnNV6R1mIiNU6W1mIdNV6\\1mI_NU6a1nIYNU6h1e1001O001O1O010O004M:F?A3M2N1O2N2N1010<C6K1OL5K5Lc0\\Of0ZO>Bj0RObeg4"}, "label": "a big elephant"}]}
{"id": 571441, "image": "COCO_train2014_000000571441.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large gray elephant in the grass in an animal pen\"."}], "task": "refering", "answer_template": "The \"a large gray elephant in the grass in an animal pen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208], "bbox": [0.050671875, 0.2837777777777778, 0.32559375, 0.7071944444444445], "iscrowd": 0, "area": 20053.648549999994, "rle": {"size": [360, 640], "counts": "T];3Q;j0UOc0^O;_FZOn7S2J5M4N2M2O2N2N0O2O1O1O1N101O1O1O1O00I72N3M2NnN[IeMb6[2bIcM\\6\\2jIaMU6]2PJ`Mo5`2VJZMk5e2[JUMg5j2[1100O100O100O100O100O100O2O001O0O2O1O3M2N3M3M2N3M2N2N2N2N2N2N2N2M3N2N2N001O1O001N100000cNbIPN^6n1mIiMS6U2VJdMj5Z2YJeMg5[2[JbMf5]2]JaMc5_2^J`Mb5`2`J^M`5a2bJ^M^5b2dJ\\M\\5d2eJ[M[5e2fJZMZ5e2hJZMX5f2e100001O1O6K6I6Ja0_O<D6J001O000000000001O000O100O100O10OhNY1N2N2N22N2lHlLP6W3kIlLT6V3iIlLV6W3fIkLY6W3cIlL\\6W3`IkL_6W3ZInLf6j300000000000001OO1TNYIWOg6:jIDV6<mIAS6?PJ]OQ6c0QJ[Oo5e0TJnNV6R1mIiNU6W1mIdNV6\\1mI_NU6a1nIYNU6h1e1001O001O1O010O004M:F?A3M2N1O2N2N1010<C6K1OL5K5Lc0\\Of0ZO>Bj0RObeg4"}, "label": "a large gray elephant in the grass in an animal pen"}]}
{"id": 423988, "image": "COCO_train2014_000000423988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white pants\"."}], "task": "refering", "answer_template": "The \"a woman in white pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 212, 213, 214, 228, 229, 230, 231, 244, 245, 246, 247, 260, 261, 262, 263, 276, 277, 278, 292, 293, 308, 309, 324, 325, 340, 341], "bbox": [0.2458823529411765, 0.521875, 0.47800904977375563, 0.93353125], "iscrowd": 0, "area": 13759.45375, "rle": {"size": [640, 442], "counts": "`eT26fc08H8H7I8H8H8H8I7H4bMVNbBn1Y=gNSB]1g=XOfAl0U>HXA<c>c2K5J6K5K5K5J6K5K5J6K5K5L4O1N2O1O1O1O1N2O1O1O1N2N2N2N2M300O1O1O1O1O100O1O1001hNW1YOg0@a0@?A?M4M2dLd@R2^?mMf@o1[?PNi@k1Z?TNi@g1Y?XNk@d1V?\\Nl@`1W?^Nm@]1U?cNn@Y1S?fNPAV1S?iNPAS1Q?lNRAQ1o>nNUAo0l>oNWAn0j>POYAn0h>QOYAo0h>nN[AP1f>nN\\AR1d>mN]AS1d>jN^AV1b>hN`AX1b>eN`AY1c>cN_A]1c>_N_Aa1``001O0O10000G:L3M3N2N3M2N2O2M4M2M4M2M4M2Min_4"}, "label": "a woman in white pants"}]}
{"id": 423988, "image": "COCO_train2014_000000423988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white pants and brown shoes with brown bag looking at a book\"."}], "task": "refering", "answer_template": "The \"a woman in white pants and brown shoes with brown bag looking at a book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 212, 213, 214, 228, 229, 230, 231, 244, 245, 246, 247, 260, 261, 262, 263, 276, 277, 278, 292, 293, 308, 309, 324, 325, 340, 341], "bbox": [0.2458823529411765, 0.521875, 0.47800904977375563, 0.93353125], "iscrowd": 0, "area": 13759.45375, "rle": {"size": [640, 442], "counts": "`eT26fc08H8H7I8H8H8H8I7H4bMVNbBn1Y=gNSB]1g=XOfAl0U>HXA<c>c2K5J6K5K5K5J6K5K5J6K5K5L4O1N2O1O1O1O1N2O1O1O1N2N2N2N2M300O1O1O1O1O100O1O1001hNW1YOg0@a0@?A?M4M2dLd@R2^?mMf@o1[?PNi@k1Z?TNi@g1Y?XNk@d1V?\\Nl@`1W?^Nm@]1U?cNn@Y1S?fNPAV1S?iNPAS1Q?lNRAQ1o>nNUAo0l>oNWAn0j>POYAn0h>QOYAo0h>nN[AP1f>nN\\AR1d>mN]AS1d>jN^AV1b>hN`AX1b>eN`AY1c>cN_A]1c>_N_Aa1``001O0O10000G:L3M3N2N3M2N2O2M4M2M4M2M4M2Min_4"}, "label": "a woman in white pants and brown shoes with brown bag looking at a book"}]}
{"id": 317495, "image": "COCO_train2014_000000317495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the man in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 374, 375, 376, 377, 397, 398, 399, 400, 401, 420, 421, 422, 423, 424, 443, 444, 445, 446, 466, 467, 468, 469, 489, 490, 491, 492, 493, 512, 513, 514, 515, 516, 517], "bbox": [0.185265625, 0.121203125, 0.490921875, 0.9869375000000001], "iscrowd": 0, "area": 59482.61839999999, "rle": {"size": [640, 640], "counts": "fbZ2>Uc0a0E<I6J6J6J6J6K5L5J5L4L4L4K5L4L5J5L1O1N2O1O001N2O1O0O2O1O001N2hBnK`:S4SEZLk:g3hDfLW;\\3[DQMd;Q3oC\\Mo;[5oMQIPGP7o8YIhFi6U9bIbFa6Y9jI]FZ6^9PJYFT6c9UJTFo5g9YJRFk5i9\\JQFi5j9\\2I8H9G8I7H8J6I7J6J6I7aIdCn4b<QJZEn4m=nN=D4L4L4K5L4L2M10000O1000000O1000000000000000000010O000000000000001O0n^OVMn`0j2Q_OWMo`0i2P_OYMo`0l201O0VBPMd:Q3ZEPMf:P3XESMg:n2VETMj:l2TEWMk:i2REZMn:g2oD[MQ;e2mD^MR;c2kD_MU;c2gD`MX;a2eDaM[;a2aDbM^;_2_DcMa;_2[DdMd;]2ZDdMf;^2WDcMi;^2UDdMj;^2SDcMm;_2QDbMn;b2mC_MS<e2YCWLAV1U=g2UCVLCX1U=e2TCULDZ1U=e2RCTLF\\1T=c2RCnMl<U2PClMP=W2mBiMS=b4VDWI]:i6bE[I[:e6cE`IZ:a6dEcIY:]6fEgIW:Y6iEjIT:V6kEnIR:R6mESJo9m5QFVJl9j5SFZJj9f5UF]Ji9d5VF_Jg9b5WFbJf9_5XFdJf9^5XFeJe9_5VFeJg9[5XFiJe9W5[FkJc9U5\\FoJX2QNY4P7^ISKT2RN[4k6`IWKT2oMY4j6cIYKS2oMW4h6eI]KR2nMU4e6hIaKQ2lMT4c6kIcKQ2kMQ4b6mIfKQ2jMo3`6oIhKR2jMl3^6QJkKR2iMj3\\6RJoKT2fMg3[6SJSLU2eMd3X6TJXLW2bMc3U6TJ\\LY2aM`3h9`LYF]3i9bLYF[3i9dL[FW3h9gL\\F:1kNf9j0\\FO;QO]9m0\\FE`0]OW9l0]FZOe0IR9i0^FROh04m8g0`FhNl0`0h8d0`FaNP1j0\\9IRI6\\7\\OeHc0]=O001O1O1O1O1O1O1O00001O0000001O00001O000010O0001O0H8JQU[6"}, "label": "the man in the right hand picture"}]}
{"id": 317495, "image": "COCO_train2014_000000317495.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing in an alley\"."}], "task": "refering", "answer_template": "The \"a man standing in an alley\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 374, 375, 376, 377, 397, 398, 399, 400, 401, 420, 421, 422, 423, 424, 443, 444, 445, 446, 466, 467, 468, 469, 489, 490, 491, 492, 493, 512, 513, 514, 515, 516, 517], "bbox": [0.185265625, 0.121203125, 0.490921875, 0.9869375000000001], "iscrowd": 0, "area": 59482.61839999999, "rle": {"size": [640, 640], "counts": "fbZ2>Uc0a0E<I6J6J6J6J6K5L5J5L4L4L4K5L4L5J5L1O1N2O1O001N2O1O0O2O1O001N2hBnK`:S4SEZLk:g3hDfLW;\\3[DQMd;Q3oC\\Mo;[5oMQIPGP7o8YIhFi6U9bIbFa6Y9jI]FZ6^9PJYFT6c9UJTFo5g9YJRFk5i9\\JQFi5j9\\2I8H9G8I7H8J6I7J6J6I7aIdCn4b<QJZEn4m=nN=D4L4L4K5L4L2M10000O1000000O1000000000000000000010O000000000000001O0n^OVMn`0j2Q_OWMo`0i2P_OYMo`0l201O0VBPMd:Q3ZEPMf:P3XESMg:n2VETMj:l2TEWMk:i2REZMn:g2oD[MQ;e2mD^MR;c2kD_MU;c2gD`MX;a2eDaM[;a2aDbM^;_2_DcMa;_2[DdMd;]2ZDdMf;^2WDcMi;^2UDdMj;^2SDcMm;_2QDbMn;b2mC_MS<e2YCWLAV1U=g2UCVLCX1U=e2TCULDZ1U=e2RCTLF\\1T=c2RCnMl<U2PClMP=W2mBiMS=b4VDWI]:i6bE[I[:e6cE`IZ:a6dEcIY:]6fEgIW:Y6iEjIT:V6kEnIR:R6mESJo9m5QFVJl9j5SFZJj9f5UF]Ji9d5VF_Jg9b5WFbJf9_5XFdJf9^5XFeJe9_5VFeJg9[5XFiJe9W5[FkJc9U5\\FoJX2QNY4P7^ISKT2RN[4k6`IWKT2oMY4j6cIYKS2oMW4h6eI]KR2nMU4e6hIaKQ2lMT4c6kIcKQ2kMQ4b6mIfKQ2jMo3`6oIhKR2jMl3^6QJkKR2iMj3\\6RJoKT2fMg3[6SJSLU2eMd3X6TJXLW2bMc3U6TJ\\LY2aM`3h9`LYF]3i9bLYF[3i9dL[FW3h9gL\\F:1kNf9j0\\FO;QO]9m0\\FE`0]OW9l0]FZOe0IR9i0^FROh04m8g0`FhNl0`0h8d0`FaNP1j0\\9IRI6\\7\\OeHc0]=O001O1O1O1O1O1O1O00001O0000001O00001O000010O0001O0H8JQU[6"}, "label": "a man standing in an alley"}]}
{"id": 358462, "image": "COCO_train2014_000000358462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white truck on the right top corner\"."}], "task": "refering", "answer_template": "The \"white truck on the right top corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 182, 183], "bbox": [0.833234375, 0.22421545667447307, 1.0, 0.5291569086651053], "iscrowd": 0, "area": 9461.822549999995, "rle": {"size": [427, 640], "counts": "X\\n65P=d0\\Od0^O5K5J5N20000000000000000000000O1000000000000000000000000O101O001O1O001O1O001N3N1O1O1O2N1O1O2N1O1O1N3N1O1O1O1O001O1O1O0O2O1O001O3M2N3M2N3M2M3N2N1O2N2N1O2N1O1O01O000O2O00001O01O0001O01O01N1O1N3N1O1N9H:G9G2N101NlK"}, "label": "white truck on the right top corner"}]}
{"id": 358462, "image": "COCO_train2014_000000358462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white semi truck parked facing a car\"."}], "task": "refering", "answer_template": "The \"a white semi truck parked facing a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 182, 183], "bbox": [0.833234375, 0.22421545667447307, 1.0, 0.5291569086651053], "iscrowd": 0, "area": 9461.822549999995, "rle": {"size": [427, 640], "counts": "X\\n65P=d0\\Od0^O5K5J5N20000000000000000000000O1000000000000000000000000O101O001O1O001O1O001N3N1O1O1O2N1O1O2N1O1O1N3N1O1O1O1O001O1O1O0O2O1O001O3M2N3M2N3M2M3N2N1O2N2N1O2N1O1O01O000O2O00001O01O0001O01O01N1O1N3N1O1N9H:G9G2N101NlK"}, "label": "a white semi truck parked facing a car"}]}
{"id": 112707, "image": "COCO_train2014_000000112707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire at home base\"."}], "task": "refering", "answer_template": "The \"an umpire at home base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 324, 325, 347, 348], "bbox": [0.003296875, 0.476625, 0.17565625, 0.9051041666666666], "iscrowd": 0, "area": 11991.562599999994, "rle": {"size": [480, 640], "counts": "bX11m>3L4gBIe;9YDKb;9[DK_;:^DJ^;9^DL];7aDM[;6bDNZ;5cDMZ;7cDJ\\;8bDJ[;:bDH\\;;aDG\\;=aDE];=aDD^;?^DD_;`0^DB`;a0]DA`;Q2E:01O1O1N2O001O1O1O1O001O1O1O1O0O2OAYEgLf:X3^EfLb:W3cEgL]:W3gEfLZ:W3kEgLU:V3PFhLP:V3TFhLl9U3Q1O11O00001O0000001O000O2L3N2M4L3M3N3L3jFYLP7j3cHcLZ7`3YHmLf7U3kGYMS8i2`GdM_8_400000000001O00000WLcGo0]8oNPHf0P8WO[Ha0e7]O]Hc0d7ZO_He0a7XOcHg0]7WOeHi0[7UOhHj0X7SOlHl0T7ROnHn0R7POQIo0o6nNUIQ1l6lNVIT1j6iNZIV1f6hN]IW1c6gN_IY1a6dNcIZ1^6cNmIU1T6fN_J7nLUOf8=oKCU43RLLR4HVL8Z9O00001O001O00001O1O5KcRg7"}, "label": "an umpire at home base"}]}
{"id": 112707, "image": "COCO_train2014_000000112707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire of a baseball game\"."}], "task": "refering", "answer_template": "The \"the umpire of a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 324, 325, 347, 348], "bbox": [0.003296875, 0.476625, 0.17565625, 0.9051041666666666], "iscrowd": 0, "area": 11991.562599999994, "rle": {"size": [480, 640], "counts": "bX11m>3L4gBIe;9YDKb;9[DK_;:^DJ^;9^DL];7aDM[;6bDNZ;5cDMZ;7cDJ\\;8bDJ[;:bDH\\;;aDG\\;=aDE];=aDD^;?^DD_;`0^DB`;a0]DA`;Q2E:01O1O1N2O001O1O1O1O001O1O1O1O0O2OAYEgLf:X3^EfLb:W3cEgL]:W3gEfLZ:W3kEgLU:V3PFhLP:V3TFhLl9U3Q1O11O00001O0000001O000O2L3N2M4L3M3N3L3jFYLP7j3cHcLZ7`3YHmLf7U3kGYMS8i2`GdM_8_400000000001O00000WLcGo0]8oNPHf0P8WO[Ha0e7]O]Hc0d7ZO_He0a7XOcHg0]7WOeHi0[7UOhHj0X7SOlHl0T7ROnHn0R7POQIo0o6nNUIQ1l6lNVIT1j6iNZIV1f6hN]IW1c6gN_IY1a6dNcIZ1^6cNmIU1T6fN_J7nLUOf8=oKCU43RLLR4HVL8Z9O00001O001O00001O1O5KcRg7"}, "label": "the umpire of a baseball game"}]}
{"id": 112707, "image": "COCO_train2014_000000112707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wicket keeper who is sitting and ready to catch the ball\"."}], "task": "refering", "answer_template": "The \"a wicket keeper who is sitting and ready to catch the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 238, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 326, 327, 328, 350, 351], "bbox": [0.16109374999999998, 0.5897708333333332, 0.375, 0.9002708333333334], "iscrowd": 0, "area": 10505.457, "rle": {"size": [480, 640], "counts": "`d`14b>>C<L5J5L4L5K4L5K4L2N2N2N2N2N1O2N2N2N2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O10O01O1eDPMe:Q3XERMh:n2UEUMj:l2SEWMl:j2REXMn:h2PEZMo:f2QE[Mn:e2QE]Mn:b2RE`Mn:_2QEcMn:\\2REfMc:\\O_En2MhMb:b2_E_M_:b2`E`M]:a2cEaM[:`2dEbM[:]2eEeM[:Z2eEgM[:W2eEkM[:]3O1000000000000000000000000000O1ZOfEfLZ:X3iEgLW:W3mEgLS:X3oEgLQ:W3SFgLn9V3VFhLk9U3XFjLi9V1cE;g0]Ng9V1fE8f0`Ng9V1eE6h0bNd9W1gE2i0eNa9W1iE1h0gN`9W1iEOb;0^DNe;0[DO[<]OdCc0]<]ObCb0_<^OaCa0a<_O^C`0d<@[C`0f<_OZC`0g<AXC>j<BVC<l<DSC<n<CRC<o<ERC8P=HQC5Q=JQC3Q=MoB0S=1nBLT=3nBJT=6lBGW=9jBDW=<kBAW=?d010O010000O100O10000001O00000001O00O1O2N1O1O1O1O1O1O2M4M3MYVk5"}, "label": "a wicket keeper who is sitting and ready to catch the ball"}]}
{"id": 112707, "image": "COCO_train2014_000000112707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher\"."}], "task": "refering", "answer_template": "The \"the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 238, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 326, 327, 328, 350, 351], "bbox": [0.16109374999999998, 0.5897708333333332, 0.375, 0.9002708333333334], "iscrowd": 0, "area": 10505.457, "rle": {"size": [480, 640], "counts": "`d`14b>>C<L5J5L4L5K4L5K4L2N2N2N2N2N1O2N2N2N2N1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O10O01O1eDPMe:Q3XERMh:n2UEUMj:l2SEWMl:j2REXMn:h2PEZMo:f2QE[Mn:e2QE]Mn:b2RE`Mn:_2QEcMn:\\2REfMc:\\O_En2MhMb:b2_E_M_:b2`E`M]:a2cEaM[:`2dEbM[:]2eEeM[:Z2eEgM[:W2eEkM[:]3O1000000000000000000000000000O1ZOfEfLZ:X3iEgLW:W3mEgLS:X3oEgLQ:W3SFgLn9V3VFhLk9U3XFjLi9V1cE;g0]Ng9V1fE8f0`Ng9V1eE6h0bNd9W1gE2i0eNa9W1iE1h0gN`9W1iEOb;0^DNe;0[DO[<]OdCc0]<]ObCb0_<^OaCa0a<_O^C`0d<@[C`0f<_OZC`0g<AXC>j<BVC<l<DSC<n<CRC<o<ERC8P=HQC5Q=JQC3Q=MoB0S=1nBLT=3nBJT=6lBGW=9jBDW=<kBAW=?d010O010000O100O10000001O00000001O00O1O2N1O1O1O1O1O1O2M4M3MYVk5"}, "label": "the catcher"}]}
{"id": 391237, "image": "COCO_train2014_000000391237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 54, 76, 77, 99, 100, 101, 121, 122, 124], "bbox": [0.293015625, 0.13568548387096774, 0.404296875, 0.5913709677419354], "iscrowd": 0, "area": 3101.7276499999975, "rle": {"size": [248, 640], "counts": "]e]12e72O0O10001N1N2L4M3M3N2G9N2NL50O0aN^OjKb0V4FbK;]4L\\K5c44UKKm4:lJGT5;iJFo4ROUKY1IEQ5XORKS2l4PNTKo1k4TNTKj1n4?0O2O1N2O1N2O2M2O3M<C2O2N1O101O01000O1O1N2N2O2;FN4M6I6KO0O36I5L1N21XIQO`6T11M3M3M3M3N3K5L4L8HXTl2"}, "label": "a man in a black shirt"}]}
{"id": 391237, "image": "COCO_train2014_000000391237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing black\"."}], "task": "refering", "answer_template": "The \"a man wearing black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 54, 76, 77, 99, 100, 101, 121, 122, 124], "bbox": [0.293015625, 0.13568548387096774, 0.404296875, 0.5913709677419354], "iscrowd": 0, "area": 3101.7276499999975, "rle": {"size": [248, 640], "counts": "]e]12e72O0O10001N1N2L4M3M3N2G9N2NL50O0aN^OjKb0V4FbK;]4L\\K5c44UKKm4:lJGT5;iJFo4ROUKY1IEQ5XORKS2l4PNTKo1k4TNTKj1n4?0O2O1N2O1N2O2M2O3M<C2O2N1O101O01000O1O1N2N2O2;FN4M6I6KO0O36I5L1N21XIQO`6T11M3M3M3M3N3K5L4L8HXTl2"}, "label": "a man wearing black"}]}
{"id": 243782, "image": "COCO_train2014_000000243782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a gray hoodie and black shorts , sitting in the bleachers while watching a tennis match\"."}], "task": "refering", "answer_template": "The \"a man wearing a gray hoodie and black shorts , sitting in the bleachers while watching a tennis match\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 150, 151, 152, 153, 175, 176], "bbox": [0.5012187499999999, 0.09121212121212122, 0.6899374999999999, 0.5033799533799533], "iscrowd": 0, "area": 13263.934200000002, "rle": {"size": [429, 640], "counts": "`cV47n<:F;J5K5K5K6J5K5L5jDWNa:]2L3M3M3N2M3N2N2N2N2N1O2M3N2N2N2M2N3N2M3M3M3M2M4L3M4L3N3N1O2N1O1O2N1O1O100000000000O1O100O1O1O2O0O2N2N2N2O010O2O6J6I2O000O2O0001O001O000000O100O100O2O0O1N3M2N2O2M2N3N3L3M5K5L3L5L4XNZEd0g:hNlEX1U;1N00001O00O1O1O1O1O1O1O1O1O0O2N5K5K5J6Kehb2"}, "label": "a man wearing a gray hoodie and black shorts , sitting in the bleachers while watching a tennis match"}]}
{"id": 243782, "image": "COCO_train2014_000000243782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a dark hoodie sitting in the stands\"."}], "task": "refering", "answer_template": "The \"a man wearing a dark hoodie sitting in the stands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 150, 151, 152, 153, 175, 176], "bbox": [0.5012187499999999, 0.09121212121212122, 0.6899374999999999, 0.5033799533799533], "iscrowd": 0, "area": 13263.934200000002, "rle": {"size": [429, 640], "counts": "`cV47n<:F;J5K5K5K6J5K5L5jDWNa:]2L3M3M3N2M3N2N2N2N2N1O2M3N2N2N2M2N3N2M3M3M3M2M4L3M4L3N3N1O2N1O1O2N1O1O100000000000O1O100O1O1O2O0O2N2N2N2O010O2O6J6I2O000O2O0001O001O000000O100O100O2O0O1N3M2N2O2M2N3N3L3M5K5L3L5L4XNZEd0g:hNlEX1U;1N00001O00O1O1O1O1O1O1O1O1O0O2N5K5K5J6Kehb2"}, "label": "a man wearing a dark hoodie sitting in the stands"}]}
{"id": 464965, "image": "COCO_train2014_000000464965.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue sirt about to eat lunch\"."}], "task": "refering", "answer_template": "The \"a man in a blue sirt about to eat lunch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.28314062500000003, 0.0, 0.9488749999999999, 0.9887708333333334], "iscrowd": 0, "area": 123553.46654999997, "rle": {"size": [480, 640], "counts": "[Ye21n>1O1N2O2N1O1O1N3N1000001_JFZL:^3OaL1[33eLMW38hLIS3;mLEQ3>nLBQ3?oLAo2a0QM@l2a0UM_Oi2c0WM]Oh2c0YM]Oe2d0\\M]Oa2e0_M[O_2f0bMZO]2f0cM[O[2g0dMZO[2f0eMZO[2f0eM[OY2g0fMZOY2f0gM[OX2f0gM[OW2f0iMZOW2f0iM[OV2f0iM[OU2f0kM[OS2f0nMZOQ2g0nMYOQ2h0oMYOP2g0QNYOm1i0RNXOm1h0SNYOk1i0UNVOk1j0UNWOi1j0WNWOh1j0XNVOf1k0ZNVOd1k0\\NUOd1l0\\NTOb1m0^NTO`1n0_NSO_1n0bNRO\\1o0dNQO[1Q1dNPOZ1Q1gNoNX1Q1hNPOV1R1iNnNV1S1kNmNS1T1mNmNQ1U1nNkNQ1V1POjNn0X1QOiNn0W1ROiNm0X1ROjNl0X1SOiNk0X1UOhNj0Y1VOhNh0Z1WOgNg0Z1YOfNg0[1XOfNf0[1ZOfNe0[1YOfNg0[1XOfNg0[1XOfNg0[1XOeNh0\\1WOeNh0\\1VOfNi0[1VOeNj0\\1UOeNi0]1VOdNi0]1WObNi0_1WOaNh0`1WOaNh0_1YO`Ng0a1YO_Nf0b1ZO^Ne0c1[O^Nb0d1]O_N`0b1@_N>b1B`N;a1EaN8`1HaN6`1IcN4^1LdN1]1OdNO]11eNK]14fNI[17fNG[19gNDZ1<hNAY1?hN_OY1`0iN^OX1b0hN]OY1c0gN[O[1e0eNZO\\1f0dNYO]1f0dNYO]1g0cNXO^1h0bNWO^1j0bNUO_1j0bNUO_1k0aNTO`1l0`NROb1n0^NQOc1o0]NPOd1o0]NPOd1P1\\NoNe1Q1[NnNf1R1ZNmNg1S1YNlNh1S1YNkNi1U1WNjNj1V1VNiNk1W1UNhNl1X1TNgNm1X1TNhNl1X1TNhNl1X1TNhNl1X1TNhNm1V1TNjNl1V1TNjNl1V1SNkNm1U1SNkNn1T1RNlNn1T1RNlNn1T1RNlNn1T1RNlNo1S1QNmNo1S1QNmNo1S1QNmNo1S1QNmNP2R1[KjK_2T3V2X1jMhNV2X1jMhN^2P1bMPOj2d0VM\\OQ3>nLBU3;kLDY39gLG[37eLI^34bLL`32`LNc3O]L1f3MYL3i3KWL5l3GUL9m3DTL<o3@RL`0P4]OQLc0Q4ZOPLf0R4WOoKi0T4SOmKm0U4POlKP1V4mNkKS1W4jNjKV1X4gNiKY1Z4cNgK]1[4`NfK`1\\4]NeKc1]4ZNdKf1^4WNcKi1_4TNbKl1`4QNaKo1a4nM`KR2a4lM`KT2`4jMbKV2_4hMbKX2^4gMcKY2]4fMdKZ2]4cMeK\\2\\4cMeK]2\\4aMeK_2[4`MfK`2Z4^MhKb2Y4\\MhKd2X4[MiKe2X4YMiKg2W4WMkKi2W4TMjKl2Y4PMhKP3[4lLfKT3S800000000L4jN\\DmNg;R1\\DhNh;W1[DcNi;\\1YD_Nk;`1XDZNl;f1g0O1O1O1O1O1O1O1L4M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000kEL^54`JN`52^J1a5O]J4b5L\\J7c5I[J:d5F[J<d5DZJ>f5BXJa0g5_OWJd0h5\\OVJg0i5YOUJj0j5VOUJl0j5TOTJo0k5QOUILoMU1l8oNQI1PNS1n8lNoH4QNS1o8iNlH8SNR1P9fNiH=UNo0Q9dNgH`0VNo0R9aNdHe0WNm0T9^NaHi0YNk0V9\\N^Hm0YNj0X9YN[HQ1[Ni0Y9VNXHU1^Ng0Y9TNVHY1^Nf0[9QNSH]1`Ne0\\9nMPHk3o7ULnGn3R8RLjGS4U8mKgGX4X8hKeG]4Y8iK]G\\4b8S1N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N4L4L4L3M4L4L3M4L4L4L3K6J6J6I6K6J6I6K6J6I7J5K6I7YNnDVOX;<ZEZOk::fE[Oa:7PF@V:2hV?"}, "label": "a man in a blue sirt about to eat lunch"}]}
{"id": 464965, "image": "COCO_train2014_000000464965.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man wearing a blue shirt with yellow stripes holding a sandwich in foil paper\"."}], "task": "refering", "answer_template": "The \"a young man wearing a blue shirt with yellow stripes holding a sandwich in foil paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.28314062500000003, 0.0, 0.9488749999999999, 0.9887708333333334], "iscrowd": 0, "area": 123553.46654999997, "rle": {"size": [480, 640], "counts": "[Ye21n>1O1N2O2N1O1O1N3N1000001_JFZL:^3OaL1[33eLMW38hLIS3;mLEQ3>nLBQ3?oLAo2a0QM@l2a0UM_Oi2c0WM]Oh2c0YM]Oe2d0\\M]Oa2e0_M[O_2f0bMZO]2f0cM[O[2g0dMZO[2f0eMZO[2f0eM[OY2g0fMZOY2f0gM[OX2f0gM[OW2f0iMZOW2f0iM[OV2f0iM[OU2f0kM[OS2f0nMZOQ2g0nMYOQ2h0oMYOP2g0QNYOm1i0RNXOm1h0SNYOk1i0UNVOk1j0UNWOi1j0WNWOh1j0XNVOf1k0ZNVOd1k0\\NUOd1l0\\NTOb1m0^NTO`1n0_NSO_1n0bNRO\\1o0dNQO[1Q1dNPOZ1Q1gNoNX1Q1hNPOV1R1iNnNV1S1kNmNS1T1mNmNQ1U1nNkNQ1V1POjNn0X1QOiNn0W1ROiNm0X1ROjNl0X1SOiNk0X1UOhNj0Y1VOhNh0Z1WOgNg0Z1YOfNg0[1XOfNf0[1ZOfNe0[1YOfNg0[1XOfNg0[1XOfNg0[1XOeNh0\\1WOeNh0\\1VOfNi0[1VOeNj0\\1UOeNi0]1VOdNi0]1WObNi0_1WOaNh0`1WOaNh0_1YO`Ng0a1YO_Nf0b1ZO^Ne0c1[O^Nb0d1]O_N`0b1@_N>b1B`N;a1EaN8`1HaN6`1IcN4^1LdN1]1OdNO]11eNK]14fNI[17fNG[19gNDZ1<hNAY1?hN_OY1`0iN^OX1b0hN]OY1c0gN[O[1e0eNZO\\1f0dNYO]1f0dNYO]1g0cNXO^1h0bNWO^1j0bNUO_1j0bNUO_1k0aNTO`1l0`NROb1n0^NQOc1o0]NPOd1o0]NPOd1P1\\NoNe1Q1[NnNf1R1ZNmNg1S1YNlNh1S1YNkNi1U1WNjNj1V1VNiNk1W1UNhNl1X1TNgNm1X1TNhNl1X1TNhNl1X1TNhNl1X1TNhNm1V1TNjNl1V1TNjNl1V1SNkNm1U1SNkNn1T1RNlNn1T1RNlNn1T1RNlNn1T1RNlNo1S1QNmNo1S1QNmNo1S1QNmNo1S1QNmNP2R1[KjK_2T3V2X1jMhNV2X1jMhN^2P1bMPOj2d0VM\\OQ3>nLBU3;kLDY39gLG[37eLI^34bLL`32`LNc3O]L1f3MYL3i3KWL5l3GUL9m3DTL<o3@RL`0P4]OQLc0Q4ZOPLf0R4WOoKi0T4SOmKm0U4POlKP1V4mNkKS1W4jNjKV1X4gNiKY1Z4cNgK]1[4`NfK`1\\4]NeKc1]4ZNdKf1^4WNcKi1_4TNbKl1`4QNaKo1a4nM`KR2a4lM`KT2`4jMbKV2_4hMbKX2^4gMcKY2]4fMdKZ2]4cMeK\\2\\4cMeK]2\\4aMeK_2[4`MfK`2Z4^MhKb2Y4\\MhKd2X4[MiKe2X4YMiKg2W4WMkKi2W4TMjKl2Y4PMhKP3[4lLfKT3S800000000L4jN\\DmNg;R1\\DhNh;W1[DcNi;\\1YD_Nk;`1XDZNl;f1g0O1O1O1O1O1O1O1L4M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000kEL^54`JN`52^J1a5O]J4b5L\\J7c5I[J:d5F[J<d5DZJ>f5BXJa0g5_OWJd0h5\\OVJg0i5YOUJj0j5VOUJl0j5TOTJo0k5QOUILoMU1l8oNQI1PNS1n8lNoH4QNS1o8iNlH8SNR1P9fNiH=UNo0Q9dNgH`0VNo0R9aNdHe0WNm0T9^NaHi0YNk0V9\\N^Hm0YNj0X9YN[HQ1[Ni0Y9VNXHU1^Ng0Y9TNVHY1^Nf0[9QNSH]1`Ne0\\9nMPHk3o7ULnGn3R8RLjGS4U8mKgGX4X8hKeG]4Y8iK]G\\4b8S1N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N4L4L4L3M4L4L3M4L4L4L3K6J6J6I6K6J6I6K6J6I7J5K6I7YNnDVOX;<ZEZOk::fE[Oa:7PF@V:2hV?"}, "label": "a young man wearing a blue shirt with yellow stripes holding a sandwich in foil paper"}]}
{"id": 481355, "image": "COCO_train2014_000000481355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tabletop covered in food and drink\"."}], "task": "refering", "answer_template": "The \"a tabletop covered in food and drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199], "bbox": [0.1668125, 0.3983177570093458, 0.739375, 0.5623598130841121], "iscrowd": 0, "area": 20217.909600000006, "rle": {"size": [428, 640], "counts": "Zm\\14W=4L4L4L2N2M200O2N1O1O2N101O0O101O001N10001O001N10001O001N10001O001N10001O001N10001O001N10001O001N10001O001N1000000O10001O000O100000000O10001O000O100000000O1000001O0000000000000000001O000000000000000000001O000000000000000000001O0000000000000O10001O000001O00000000000010O00000000000001O0001O00000000O2O0000000000001O0000000000000000000000O1000000000000000000O100000000000O100000O10000000000000000O1000000000000000000O1000000000000000000O1001O0O1000000000000000001O00000000000000000000001O000000000000000000001N100000000000000000001O0000000000000000001O00000000001O0O100000001O00000000001O00000000001O00000000001O000000000O2O00000000001O00000000001O000O1O2O0O2O0O2N101N101N101N1O2O0O2O0O2N101N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2MZcU2"}, "label": "a tabletop covered in food and drink"}]}
{"id": 481355, "image": "COCO_train2014_000000481355.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black table is full of flatware\"."}], "task": "refering", "answer_template": "The \"the black table is full of flatware\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199], "bbox": [0.1668125, 0.3983177570093458, 0.739375, 0.5623598130841121], "iscrowd": 0, "area": 20217.909600000006, "rle": {"size": [428, 640], "counts": "Zm\\14W=4L4L4L2N2M200O2N1O1O2N101O0O101O001N10001O001N10001O001N10001O001N10001O001N10001O001N10001O001N10001O001N1000000O10001O000O100000000O10001O000O100000000O1000001O0000000000000000001O000000000000000000001O000000000000000000001O0000000000000O10001O000001O00000000000010O00000000000001O0001O00000000O2O0000000000001O0000000000000000000000O1000000000000000000O100000000000O100000O10000000000000000O1000000000000000000O1000000000000000000O1001O0O1000000000000000001O00000000000000000000001O000000000000000000001N100000000000000000001O0000000000000000001O00000000001O0O100000001O00000000001O00000000001O00000000001O000000000O2O00000000001O00000000001O000O1O2O0O2O0O2N101N101N101N1O2O0O2O0O2N101N101N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2MZcU2"}, "label": "the black table is full of flatware"}]}
{"id": 96338, "image": "COCO_train2014_000000096338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue umbrella\"."}], "task": "refering", "answer_template": "The \"blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 167, 168, 169, 170, 171], "bbox": [0.193640625, 0.24944444444444444, 0.47768750000000004, 0.5004888888888889], "iscrowd": 0, "area": 14310.8167, "rle": {"size": [450, 640], "counts": "mcf13o=4K6J5L5J5K5L5K4K6K2N1N2O001N2O001N2O001N2O001N101O1N101O1N101N2O001N2O001N101O1N101O1N101O1N101O0O2O1O0O2O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O0000000000001O00000O100000000000000000000000000010O001O1O001O1O001O1O001O1O001O1O00100O001O1O001O1O001O1O001O001O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1N2N2N3M2O1N3M2N2N2O2M2N2N3N1N2O2M2N2O1N3M2O1N3M2O1N2N3N1N2N3N5J7Jebb4"}, "label": "blue umbrella"}]}
{"id": 96338, "image": "COCO_train2014_000000096338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue umbrella\"."}], "task": "refering", "answer_template": "The \"the blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 167, 168, 169, 170, 171], "bbox": [0.193640625, 0.24944444444444444, 0.47768750000000004, 0.5004888888888889], "iscrowd": 0, "area": 14310.8167, "rle": {"size": [450, 640], "counts": "mcf13o=4K6J5L5J5K5L5K4K6K2N1N2O001N2O001N2O001N2O001N101O1N101O1N101N2O001N2O001N101O1N101O1N101O1N101O0O2O1O0O2O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O0000000000001O00000O100000000000000000000000000010O001O1O001O1O001O1O001O1O001O1O00100O001O1O001O1O001O1O001O001O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1N2N2N3M2O1N3M2N2N2O2M2N2N3N1N2O2M2N2O1N3M2O1N3M2O1N2N3N1N2N3N5J7Jebb4"}, "label": "the blue umbrella"}]}
{"id": 276564, "image": "COCO_train2014_000000276564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a white t - shirt who ' s in the middle of a skateboard trick\"."}], "task": "refering", "answer_template": "The \"a boy wearing a white t - shirt who ' s in the middle of a skateboard trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 86, 87, 89, 90, 91, 92, 101, 102, 103, 104, 113, 114, 115, 116, 124, 125, 126, 127, 128], "bbox": [0.19118694362017807, 0.21830000000000002, 0.9243026706231454, 0.59902], "iscrowd": 0, "area": 14876.8205, "rle": {"size": [500, 337], "counts": "_^o01\\?9M4M2M3N0O10O001O010O1O10O01O1O10O01O100O001O10O01O10O010O10O10O100O100O01000O100O1000O0100O01000O10O0100O01000O10O010O10O010O0100OdCE`9<`FE_9:aFG^9:bFG\\99dFH\\98dFH[98fFIX98gFIX97iFIW97iFJU96kFKT95mFLQ95nFLR93oFMP94PGMn84RGLn7LlE8V2Nl73fEO^2Nl75cENa2Mk78aELR1?VO^OU::_EJo0h0ZOUOW:R1\\FM\\OQOW:S1[F0ZOPOY:Q1]F1VOPO]:n0]Fn0b9RO^Fn0a9SO_Fm0a9RO_FP1`9PO`FP1`9PO`FQ1^9POaFQ1_9nNbFR1^9nNbFS1\\9nNcFS1]9mNcFS1]9lNdFU1[9kNdFV1[9kNeFT1\\9lNdFT1[9lNeFU1[9kNeFS1\\9nNdFJZO_OT:g0aFKCSOn9S1_FKCmNR:X1[FKDhNU:]1VFMEaNY:b1RFMF]N[:f1oENc:2[EOe:1[E0d:1[EOf:0YE2f:OYE2g:MYE3g:NWE4i:KWE5i:KWE6i:JUE7l:HTE9k:HSE9n:FRE;m:QNgD_1;`0U;@iDb0U;@jD`0V;AiD`0V;AhD`0X;AgD`0X;AgD?Y;BfD?X;ChD<X;EgD<g0lMa9j1gE:f0nMc9i1fE@1XOb0Ce9W3jEUM>Fh9U3kEUM;Hj9S3lEUM7Jm9P3nEUM4Kn9Q3nEUM2KP:P3oETM0MQ:o2PFVMKKU:P3PFlMP:T2QFkMo9T2SFjMn9V2RFjMm9W2TFhMl9X2TFhMl9X2UFfMl9Z2TFfMk9Z2WFdMj9\\2VFcMk9]2UFbMl9^2UF`Mk9a2UF^Ml9a2VF]Mk9c2UF\\Mk9HcEU2c0QNk9JcES2d0QNj9LcER2d0PNj9NdEo1c0RNi90gEk1a0SNi92hEi1`0RNj94jEf1=UNi96lEc1<UNi98lEb1<TNi9:lE`1<UNh9<mE^1<SNi9?lE\\1=RNh9b0lE[1>PNh9e0kEY1?oMg9?UEIf0g1\\;^NQE_1P;`NRE_1n:aNTE]1m:cNUEZ1k:fNXEW1i:iNZEP1j:oNYEj0k:VOXEc0m:]OWE;m:EYE6f:JbEK`:5`20000000O1000O10000000000000001O000O100000001O00000000000000000O10O100O10000O10000O100001O0000000000001O001O1O2M2O1N2Obb<"}, "label": "a boy wearing a white t - shirt who ' s in the middle of a skateboard trick"}]}
{"id": 276564, "image": "COCO_train2014_000000276564.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in mid air above skateboard\"."}], "task": "refering", "answer_template": "The \"boy in mid air above skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 86, 87, 89, 90, 91, 92, 101, 102, 103, 104, 113, 114, 115, 116, 124, 125, 126, 127, 128], "bbox": [0.19118694362017807, 0.21830000000000002, 0.9243026706231454, 0.59902], "iscrowd": 0, "area": 14876.8205, "rle": {"size": [500, 337], "counts": "_^o01\\?9M4M2M3N0O10O001O010O1O10O01O1O10O01O100O001O10O01O10O010O10O10O100O100O01000O100O1000O0100O01000O10O0100O01000O10O010O10O010O0100OdCE`9<`FE_9:aFG^9:bFG\\99dFH\\98dFH[98fFIX98gFIX97iFIW97iFJU96kFKT95mFLQ95nFLR93oFMP94PGMn84RGLn7LlE8V2Nl73fEO^2Nl75cENa2Mk78aELR1?VO^OU::_EJo0h0ZOUOW:R1\\FM\\OQOW:S1[F0ZOPOY:Q1]F1VOPO]:n0]Fn0b9RO^Fn0a9SO_Fm0a9RO_FP1`9PO`FP1`9PO`FQ1^9POaFQ1_9nNbFR1^9nNbFS1\\9nNcFS1]9mNcFS1]9lNdFU1[9kNdFV1[9kNeFT1\\9lNdFT1[9lNeFU1[9kNeFS1\\9nNdFJZO_OT:g0aFKCSOn9S1_FKCmNR:X1[FKDhNU:]1VFMEaNY:b1RFMF]N[:f1oENc:2[EOe:1[E0d:1[EOf:0YE2f:OYE2g:MYE3g:NWE4i:KWE5i:KWE6i:JUE7l:HTE9k:HSE9n:FRE;m:QNgD_1;`0U;@iDb0U;@jD`0V;AiD`0V;AhD`0X;AgD`0X;AgD?Y;BfD?X;ChD<X;EgD<g0lMa9j1gE:f0nMc9i1fE@1XOb0Ce9W3jEUM>Fh9U3kEUM;Hj9S3lEUM7Jm9P3nEUM4Kn9Q3nEUM2KP:P3oETM0MQ:o2PFVMKKU:P3PFlMP:T2QFkMo9T2SFjMn9V2RFjMm9W2TFhMl9X2TFhMl9X2UFfMl9Z2TFfMk9Z2WFdMj9\\2VFcMk9]2UFbMl9^2UF`Mk9a2UF^Ml9a2VF]Mk9c2UF\\Mk9HcEU2c0QNk9JcES2d0QNj9LcER2d0PNj9NdEo1c0RNi90gEk1a0SNi92hEi1`0RNj94jEf1=UNi96lEc1<UNi98lEb1<TNi9:lE`1<UNh9<mE^1<SNi9?lE\\1=RNh9b0lE[1>PNh9e0kEY1?oMg9?UEIf0g1\\;^NQE_1P;`NRE_1n:aNTE]1m:cNUEZ1k:fNXEW1i:iNZEP1j:oNYEj0k:VOXEc0m:]OWE;m:EYE6f:JbEK`:5`20000000O1000O10000000000000001O000O100000001O00000000000000000O10O100O10000O10000O100001O0000000000001O001O1O2M2O1N2Obb<"}, "label": "boy in mid air above skateboard"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two patchwork throw pillows on a dark brown couch\"."}], "task": "refering", "answer_template": "The \"two patchwork throw pillows on a dark brown couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291], "bbox": [0.010354166666666666, 0.48734374999999996, 0.3626875, 0.798234375], "iscrowd": 0, "area": 22028.919200000004, "rle": {"size": [640, 480], "counts": "V_3h4X?M3N1O2M3O010O100O01000O01000O01000O10O01000O010000O010O10O10O10O10O1000O0100O01000O01000O10O01000O010000O010O10O10O10O10O1000O0100O01000O10O10OI8@?01000O10000O100O10001N10000O10000O100O10000O10000O100O10001N10000O10000O100O10000O10000O100O10000O2O000O100O10000O10000O10000O100O10000O101O0O100O10000O10000dM\\2nNR1N2O1O1O1O1N2O1O2N1N2O1O1OTbo5"}, "label": "two patchwork throw pillows on a dark brown couch"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two throw pillows on the couch\"."}], "task": "refering", "answer_template": "The \"two throw pillows on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291], "bbox": [0.010354166666666666, 0.48734374999999996, 0.3626875, 0.798234375], "iscrowd": 0, "area": 22028.919200000004, "rle": {"size": [640, 480], "counts": "V_3h4X?M3N1O2M3O010O100O01000O01000O01000O10O01000O010000O010O10O10O10O10O1000O0100O01000O01000O10O01000O010000O010O10O10O10O10O1000O0100O01000O10O10OI8@?01000O10000O100O10001N10000O10000O100O10000O10000O100O10001N10000O10000O100O10000O10000O100O10000O2O000O100O10000O10000O10000O100O10000O101O0O100O10000O10000dM\\2nNR1N2O1O1O1O1N2O1O2N1N2O1O1OTbo5"}, "label": "two throw pillows on the couch"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black pot with a plant in it\"."}], "task": "refering", "answer_template": "The \"a black pot with a plant in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 203, 220, 236, 237, 252, 253, 254, 269, 270, 271, 286, 287, 288, 304, 305, 321, 322, 339], "bbox": [0.8410416666666667, 0.46820312499999994, 1.0, 0.8405156249999999], "iscrowd": 0, "area": 6354.853799999998, "rle": {"size": [640, 480], "counts": "ill74jc07J7H7J6J7H8I7I7H8I7H3CcMT_OZ1KDQa0SOR_OV17Bf`0YOR_OQ1b0@\\`0_OQ_On0m0]OS`0Eo^Oj0X1[Oi?Kn^Og0c1XO_?2m^Ob0Pc0^Oo\\O?Uc0Bj\\O:Zc0Ff\\O6^c025K4N20n]OKY`05g_O3Q`0Nn_O:j?FU@c0d?\\O\\@k0]?UOc@S1U?mNj@\\1n>dNRAd1f>\\NZAl1^>TNbAS2W>mMhA\\2Q>cMoA_2o=bMPB^2P>bMoA_2Q>cMmA^2R>cMmA]2S>eMjA\\2V>eMiA\\2W>eMgA[2Y>fMeA\\2Z>fMdAZ2\\>gMcAY2]>iMaAX2^>jM_AW2b>iM]AX2b>jM\\AV2d>kMZAV2f>UMc@a0e0[2g>TMf@`0b0\\2h>SMi@a0<^2j>PMl@b09^2X?bMf@^2\\?bMa@`2_?`M`@`2b?`M\\@a2e?^MY@c2i?]MU@c2l?]MS@d2n?\\MP@d2R`0[Ml_Og2a<jLnFh3R9]LkFb3U9bLhFX3\\9mLXFT3n9QMhCAQ1`3_;SMTCGP1c2jNaMX=Q1aCg1^OlLW=a1mBl1g=VNlAS2U>j13NaF"}, "label": "a black pot with a plant in it"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch with comfortable pillows\"."}], "task": "refering", "answer_template": "The \"a brown couch with comfortable pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 233, 248, 249, 250, 265, 266, 267, 275, 276, 277, 282, 283, 290, 291, 292, 293, 294, 299, 300, 306, 307, 308, 309, 310, 311, 312, 314, 316, 317, 323, 324, 325, 326, 327, 328, 329, 331, 332, 333, 334, 340, 341, 342, 343, 344, 345, 346, 348, 349], "bbox": [0.0, 0.565515625, 0.7451875, 0.8995625], "iscrowd": 0, "area": 25097.436800000007, "rle": {"size": [640, 480], "counts": "j?;ec0h0XOg0YO8G100000000000000000000000000KnM\\^OR2da0PNY^OQ2fa050000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O1O1O1O1O5K[1eN=C1O00001O001O001O00001O001O00001O001O001O0000h0XOb[h0<ocWO3M2O2M3M3N2M3O100Q^OSNla0l1T^OTNla0l1T^OTNma0k1R^OVNna0j1R^OVNoa0i1Q^OWNoa0i1Q^OWNPb0h1P^OXNPb0h1P^OXNQb0g1n]OZNRb0f1n]OZNSb0e1m]O[NSb0e1m]O[NTb0d1l]O\\NTb0d1l]O\\NUb0g11Jj]O`NWb0_1i]OaNXb0^1h]ObNYb0]1g]OcNZb0\\1f]OdN[b0[1e]OeN\\b0_1100O6J<D<D;Ecm<JdRC2M4L3M3N3L3M3N3L3M3N3L3M3o@nNo:U1aBW1_=DaAa0_>`21O0000001O0000bAkJR>T5kAPKT>P5iATKV>l4fAXK[>W5O1O100O1O100O1O1O10oMXChLh<n2Z2E<E:L4M4M2N2N3L3N2N3M2M3N2O2N1O1O2N1O1O2N100O2N1O1O1O2N1O1O2L4L5K7Jb0]O\\k[2"}, "label": "a brown couch with comfortable pillows"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown couch in the livingroom\"."}], "task": "refering", "answer_template": "The \"a brown couch in the livingroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 233, 248, 249, 250, 265, 266, 267, 275, 276, 277, 282, 283, 290, 291, 292, 293, 294, 299, 300, 306, 307, 308, 309, 310, 311, 312, 314, 316, 317, 323, 324, 325, 326, 327, 328, 329, 331, 332, 333, 334, 340, 341, 342, 343, 344, 345, 346, 348, 349], "bbox": [0.0, 0.565515625, 0.7451875, 0.8995625], "iscrowd": 0, "area": 25097.436800000007, "rle": {"size": [640, 480], "counts": "j?;ec0h0XOg0YO8G100000000000000000000000000KnM\\^OR2da0PNY^OQ2fa050000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O1O1O1O1O5K[1eN=C1O00001O001O001O00001O001O00001O001O001O0000h0XOb[h0<ocWO3M2O2M3M3N2M3O100Q^OSNla0l1T^OTNla0l1T^OTNma0k1R^OVNna0j1R^OVNoa0i1Q^OWNoa0i1Q^OWNPb0h1P^OXNPb0h1P^OXNQb0g1n]OZNRb0f1n]OZNSb0e1m]O[NSb0e1m]O[NTb0d1l]O\\NTb0d1l]O\\NUb0g11Jj]O`NWb0_1i]OaNXb0^1h]ObNYb0]1g]OcNZb0\\1f]OdN[b0[1e]OeN\\b0_1100O6J<D<D;Ecm<JdRC2M4L3M3N3L3M3N3L3M3N3L3M3o@nNo:U1aBW1_=DaAa0_>`21O0000001O0000bAkJR>T5kAPKT>P5iATKV>l4fAXK[>W5O1O100O1O100O1O1O10oMXChLh<n2Z2E<E:L4M4M2N2N3L3N2N3M2M3N2O2N1O1O2N1O1O2N100O2N1O1O1O2N1O1O2L4L5K7Jb0]O\\k[2"}, "label": "a brown couch in the livingroom"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a four light decorative lamp\"."}], "task": "refering", "answer_template": "The \"a four light decorative lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 16, 25, 26, 27, 28, 30, 31, 32, 33, 43, 44, 45, 46, 47, 48, 50, 62, 63, 64, 65, 66, 79, 80, 81, 82, 96, 113, 130, 147, 163, 164, 165, 166, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 215, 216, 217, 232], "bbox": [0.4986041666666667, 0.0052499999999999995, 1.0, 0.5861875000000001], "iscrowd": 0, "area": 18206.6917, "rle": {"size": [640, 480], "counts": "m\\e41oc00O101N101N1O2O0O101N1O2O0O5L4K5L4K5K3N1N2O0O2N101N2O0O2O0O1O100O1nD\\Nh3c1XLbNc3_1\\LfN_3[1`LjN\\3V1cLoNX3Q1hLTOT3l0kLYOP3h0oL]Om2c0RM@k2`0UMAj2`0UMAk2?TMBl2>SMCn2<QMEP3:oLGR38mLIT36kLKV34iLMX32gLO[3OdL2]3MbL4_3K`L6a3H_L9b3F]L;d3B]L?d3_O\\Lb0e3[O\\Lf0f3WOZLj0g3SOZLn0g3POYLQ1h3lNYLU1h3iNXLX1i3eNXL\\1i3dNUL]1l3jNkKW1W4oN`KR1a4UOQKQ1P5VObJP1_5lM\\Ek0g4_1n5eMbEm0R4d1]6^MhEo0X3n1Q7RMnEQ1[2[2i7bLSFT1^1h2`8SLVFY1d0R3W9cKXF`1L[3m9TKWFi1WO`3k;[20000O100^DeH^:\\7XEPIf:P7oD]In:l7O2M^KZEMe:HhE8V:]OVFd0g9ROfFi0]9lNPG0Q:NSF1m9KYF3f9J`F4`9IdF6\\9FjF8U9FPG7Q9ETG:k8C[G;e8B`G<_8AfG>Z8^OlG`0T8]OPHb0o7[OWHc0i7ZO[He0d7XOaHg0_7UOgHi0Z7UOhHj0X7WOhHh0Y7XOhHf0X7[OhHd0Y7[OhHd0X7]OiHa0X7_OhH`0X7AgH?Y7BgH=Z7@iH`0V7_OlH`0U7]OmHd0R7ZOQIf0o6WOTIh0l6VOWIj0i6TOcHnNWLm1V;SOfHROQLl1V;ROmHSOkKk1V;QORIUOgKi1V;QOVIXObKg1U;TOZIVO^Kg1V;UO\\IVO\\Kd1W;XO^IUOYKP1j;MmHTOXK?X<`0`HSOVK7^<h0]HBg7?ZHZOj7i0VHQOn7Q1SHeNT8[1nG]NW8:iBm0Zb0SOg]Ol0Yb0TOh]Oj0Xb0WOi]Oh0Wb0XOj]Og0Vb0YOk]Of0Tb0[Om]Oc0Tb0]Om]Ob0Tb0]Om]Ob0Tb0]Om]Ob0Sb0^Om]Ob0Tb0]Om]Oh0ma0WOU^Om0ga0ROZ^OS1aa0lN`^OX1[a0hNf^OY1Ya0fNh^OZ1Wa0gNi^OY1Wa0fNj^OZ1Va0fNj^OZ1Ua0fNl^OZ1Ta0fNk^O[1Ta0eNm^OZ1Ta0fNl^OZ1Ta0eNm^O[1Ra0fNn^OZ1Ra0eNP_OZ1o`0gNQ_OY1o`0fNS_OX1n`0hNR_OX1m`0hNU_OW1k`0iNU_OW1j`0iNX_OV1h`0jNX_OV1h`0iNZ_OU1f`0lNZ_OT1f`0kN\\_OT1c`0mN]_OS1c`0lN__OS1a`0mN__OS1``0mNb_OQ1_`0nNb_OR1]`0oNd_OP1\\`0oNe_OQ1[`0oNf_OP1Y`0POh_Oo0Y`0QOh_On0W`0SOi_Om0W`0SOj_Ol0V`0SOk_Om0T`0TOm_Oj0T`0VOl_Oj0T`0VOm_Oc0X`0]Oi_O4f`0L[_O3d`0N\\_O2d`0M^_O2b`0N^_O1b`00__OOa`01__OO``01b_ON^`02b_ON^`02c_OM\\`04d_OK]`04g1100000O10O1000O10O10OR]OLRb04k]ONVb02g]O1Yb0Od]O3\\b0Mb]O5_b0Ka]O4`b0L_]O4bb0L]]O5cb0J^]O5bb0L]]O4db0L\\]O3eb0MZ]O4fb0K[]O4fb0LY]O4gb0MY]O3gb0MX]O:bb0E^]Oa0]b0_Oc]Og0Wb0YOh]On0Qb0ROP^OS1ka0mNT^OV1ja0jNV^OU1ka0kNR^OV1Pb0iNl]O[1Tb081K5oN="}, "label": "a four light decorative lamp"}]}
{"id": 260181, "image": "COCO_train2014_000000260181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large , ornamental lamp\"."}], "task": "refering", "answer_template": "The \"the large , ornamental lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 16, 25, 26, 27, 28, 30, 31, 32, 33, 43, 44, 45, 46, 47, 48, 50, 62, 63, 64, 65, 66, 79, 80, 81, 82, 96, 113, 130, 147, 163, 164, 165, 166, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 215, 216, 217, 232], "bbox": [0.4986041666666667, 0.0052499999999999995, 1.0, 0.5861875000000001], "iscrowd": 0, "area": 18206.6917, "rle": {"size": [640, 480], "counts": "m\\e41oc00O101N101N1O2O0O101N1O2O0O5L4K5L4K5K3N1N2O0O2N101N2O0O2O0O1O100O1nD\\Nh3c1XLbNc3_1\\LfN_3[1`LjN\\3V1cLoNX3Q1hLTOT3l0kLYOP3h0oL]Om2c0RM@k2`0UMAj2`0UMAk2?TMBl2>SMCn2<QMEP3:oLGR38mLIT36kLKV34iLMX32gLO[3OdL2]3MbL4_3K`L6a3H_L9b3F]L;d3B]L?d3_O\\Lb0e3[O\\Lf0f3WOZLj0g3SOZLn0g3POYLQ1h3lNYLU1h3iNXLX1i3eNXL\\1i3dNUL]1l3jNkKW1W4oN`KR1a4UOQKQ1P5VObJP1_5lM\\Ek0g4_1n5eMbEm0R4d1]6^MhEo0X3n1Q7RMnEQ1[2[2i7bLSFT1^1h2`8SLVFY1d0R3W9cKXF`1L[3m9TKWFi1WO`3k;[20000O100^DeH^:\\7XEPIf:P7oD]In:l7O2M^KZEMe:HhE8V:]OVFd0g9ROfFi0]9lNPG0Q:NSF1m9KYF3f9J`F4`9IdF6\\9FjF8U9FPG7Q9ETG:k8C[G;e8B`G<_8AfG>Z8^OlG`0T8]OPHb0o7[OWHc0i7ZO[He0d7XOaHg0_7UOgHi0Z7UOhHj0X7WOhHh0Y7XOhHf0X7[OhHd0Y7[OhHd0X7]OiHa0X7_OhH`0X7AgH?Y7BgH=Z7@iH`0V7_OlH`0U7]OmHd0R7ZOQIf0o6WOTIh0l6VOWIj0i6TOcHnNWLm1V;SOfHROQLl1V;ROmHSOkKk1V;QORIUOgKi1V;QOVIXObKg1U;TOZIVO^Kg1V;UO\\IVO\\Kd1W;XO^IUOYKP1j;MmHTOXK?X<`0`HSOVK7^<h0]HBg7?ZHZOj7i0VHQOn7Q1SHeNT8[1nG]NW8:iBm0Zb0SOg]Ol0Yb0TOh]Oj0Xb0WOi]Oh0Wb0XOj]Og0Vb0YOk]Of0Tb0[Om]Oc0Tb0]Om]Ob0Tb0]Om]Ob0Tb0]Om]Ob0Sb0^Om]Ob0Tb0]Om]Oh0ma0WOU^Om0ga0ROZ^OS1aa0lN`^OX1[a0hNf^OY1Ya0fNh^OZ1Wa0gNi^OY1Wa0fNj^OZ1Va0fNj^OZ1Ua0fNl^OZ1Ta0fNk^O[1Ta0eNm^OZ1Ta0fNl^OZ1Ta0eNm^O[1Ra0fNn^OZ1Ra0eNP_OZ1o`0gNQ_OY1o`0fNS_OX1n`0hNR_OX1m`0hNU_OW1k`0iNU_OW1j`0iNX_OV1h`0jNX_OV1h`0iNZ_OU1f`0lNZ_OT1f`0kN\\_OT1c`0mN]_OS1c`0lN__OS1a`0mN__OS1``0mNb_OQ1_`0nNb_OR1]`0oNd_OP1\\`0oNe_OQ1[`0oNf_OP1Y`0POh_Oo0Y`0QOh_On0W`0SOi_Om0W`0SOj_Ol0V`0SOk_Om0T`0TOm_Oj0T`0VOl_Oj0T`0VOm_Oc0X`0]Oi_O4f`0L[_O3d`0N\\_O2d`0M^_O2b`0N^_O1b`00__OOa`01__OO``01b_ON^`02b_ON^`02c_OM\\`04d_OK]`04g1100000O10O1000O10O10OR]OLRb04k]ONVb02g]O1Yb0Od]O3\\b0Mb]O5_b0Ka]O4`b0L_]O4bb0L]]O5cb0J^]O5bb0L]]O4db0L\\]O3eb0MZ]O4fb0K[]O4fb0LY]O4gb0MY]O3gb0MX]O:bb0E^]Oa0]b0_Oc]Og0Wb0YOh]On0Qb0ROP^OS1ka0mNT^OV1ja0jNV^OU1ka0kNR^OV1Pb0iNl]O[1Tb081K5oN="}, "label": "the large , ornamental lamp"}]}
{"id": 284765, "image": "COCO_train2014_000000284765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is walking away\"."}], "task": "refering", "answer_template": "The \"a giraffe that is walking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 153, 154, 155, 156, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270], "bbox": [0.60403125, 0.28881249999999997, 0.8204687500000001, 0.7004375], "iscrowd": 0, "area": 9774.188600000005, "rle": {"size": [480, 640], "counts": "Vge55f>5K6J5K5K6J5K5K6I6K5J7H7I7J6TOm0K40J5I8H7I8G8I8K43N2M3N3L3N2M3N2M3M3N2M3N2MhN[DZOb;e0bDZO[;f0jDXOT;f0REXOk:h0YEWOe:h0`EVO]:i0iESOW:l0mEROQ:o0SFmNl9T1XFhNg9X1]FeNb9\\1bF`N]9a1fF\\NY9e1kFWNT9i1QGSNn8n1Q21O1O0O2O1O1O00100TEoMh8R2QGUNn8k1lF\\NT9d1eFcNZ9^1_FjN_9W1YFQOg9n0SFYOl9h0mE_OR:b0gEFX:a2O1O101NE;D<E<C<L44M3L4L4M3L4L4M3L[MZF5b9KcF3[9LkF2Q9MUG2g8N_G0]80iGNT80SHNi72^HKn6POlFU1]2Ge6VOlFQ1g2C\\6@iFm0R3^OT6GgFk0\\3XOm5OeFg0f3TOd5[1Q4O1O0O2O1O1N101O1N101O1N101O1N10100000100O1O1O1O2O0O1O1O1O1O100O1O1O1L4J6JWhe1"}, "label": "a giraffe that is walking away"}]}
{"id": 284765, "image": "COCO_train2014_000000284765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe walking away from the other giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe walking away from the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 153, 154, 155, 156, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270], "bbox": [0.60403125, 0.28881249999999997, 0.8204687500000001, 0.7004375], "iscrowd": 0, "area": 9774.188600000005, "rle": {"size": [480, 640], "counts": "Vge55f>5K6J5K5K6J5K5K6I6K5J7H7I7J6TOm0K40J5I8H7I8G8I8K43N2M3N3L3N2M3N2M3M3N2M3N2MhN[DZOb;e0bDZO[;f0jDXOT;f0REXOk:h0YEWOe:h0`EVO]:i0iESOW:l0mEROQ:o0SFmNl9T1XFhNg9X1]FeNb9\\1bF`N]9a1fF\\NY9e1kFWNT9i1QGSNn8n1Q21O1O0O2O1O1O00100TEoMh8R2QGUNn8k1lF\\NT9d1eFcNZ9^1_FjN_9W1YFQOg9n0SFYOl9h0mE_OR:b0gEFX:a2O1O101NE;D<E<C<L44M3L4L4M3L4L4M3L[MZF5b9KcF3[9LkF2Q9MUG2g8N_G0]80iGNT80SHNi72^HKn6POlFU1]2Ge6VOlFQ1g2C\\6@iFm0R3^OT6GgFk0\\3XOm5OeFg0f3TOd5[1Q4O1O0O2O1O1N101O1N101O1N101O1N10100000100O1O1O1O2O0O1O1O1O1O100O1O1O1L4J6JWhe1"}, "label": "a giraffe walking away from the other giraffe"}]}
{"id": 284765, "image": "COCO_train2014_000000284765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe craning its neck down to the ground\"."}], "task": "refering", "answer_template": "The \"a giraffe craning its neck down to the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 236, 237, 238, 253, 254, 255, 256, 257, 260, 261, 262, 263, 276, 277, 278, 279, 280, 284, 285, 286, 299, 300, 301, 302, 303, 309, 322, 323, 324, 325, 326, 327], "bbox": [0.0033750000000000004, 0.4207916666666666, 0.46854687500000003, 0.8814583333333333], "iscrowd": 0, "area": 23110.021650000017, "rle": {"size": [480, 640], "counts": "aZ14e><_MAPFj0l9YOjES1Q:QOcE[1Y:hN]E^1d:fNQE`1P;cNfDc1[;aNYDf1h;h01O1O00ON2H8G9H8G9H8H8001O1O1O1N3N1YNjNdEW1\\:f1000001O3M2O2M9G9G9G:F9G:F9G9G8I7H1O1OYMVHZNi7e1nHfMR7X2eISMZ6l2PJlLP6R3TJlLl5R3XJlLh5R3\\JlLc5S3aJkL_5S3eJkL[5S3iJkLW5R3mJmLR5R3RKlLn4R3a3N2N2N1O2N2N2N1O2N2N2N1O200O10VGlMg4T2PKVNP5i1gJbNX5^1]JmNc5T1RJVOn5k0gI@X6b0\\IHd69PI2P7OeH:[7IYH;m7FhG>^8m25J6J6L4L4L4L4L3M4L4WOi0L4M3L4L4M3L3M4M3L43M5K5K5K5K6J7H7J7I6J7I6J7IVNeFUNU9g1UGZNd8a1fG]NW8^1SH_Nk7]1^H_Na7\\1jH`NU7Z1UIcNi6Z1_IbN`6^1cIaN[6_1iI_NU6a1nI]NQ6c1SJ[Nk5d1ZJZNd5f1_JWNa5i1cJTN\\5l1hJPNX5P2kJmMU5S2nJkMQ5U2QKiMP5V2RKiMm4V2VKhMj4X2WKgMi4Y2YKfMf4Z2\\KdMd4X2bKfM^4W2gKhMX4U2[4L4M4L4M300O1O1O100O1O1O1O100O10000000000000000O2O0000000000000000000000000O1000000000000000000000000O1000000000001O0000000000000O1000000000000000000000000O2O00000000000000000000000O100000001O00000000000000000O100000000000001O000O10000O10000OI8G9O11N2O1O0O2O1O1N2O0O2O1N2O1N2O1N4M4K6K2M101N100O101M2N2N3M2N]Po4"}, "label": "a giraffe craning its neck down to the ground"}]}
{"id": 284765, "image": "COCO_train2014_000000284765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large giraffe eating grass\"."}], "task": "refering", "answer_template": "The \"a large giraffe eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 236, 237, 238, 253, 254, 255, 256, 257, 260, 261, 262, 263, 276, 277, 278, 279, 280, 284, 285, 286, 299, 300, 301, 302, 303, 309, 322, 323, 324, 325, 326, 327], "bbox": [0.0033750000000000004, 0.4207916666666666, 0.46854687500000003, 0.8814583333333333], "iscrowd": 0, "area": 23110.021650000017, "rle": {"size": [480, 640], "counts": "aZ14e><_MAPFj0l9YOjES1Q:QOcE[1Y:hN]E^1d:fNQE`1P;cNfDc1[;aNYDf1h;h01O1O00ON2H8G9H8G9H8H8001O1O1O1N3N1YNjNdEW1\\:f1000001O3M2O2M9G9G9G:F9G:F9G9G8I7H1O1OYMVHZNi7e1nHfMR7X2eISMZ6l2PJlLP6R3TJlLl5R3XJlLh5R3\\JlLc5S3aJkL_5S3eJkL[5S3iJkLW5R3mJmLR5R3RKlLn4R3a3N2N2N1O2N2N2N1O2N2N2N1O200O10VGlMg4T2PKVNP5i1gJbNX5^1]JmNc5T1RJVOn5k0gI@X6b0\\IHd69PI2P7OeH:[7IYH;m7FhG>^8m25J6J6L4L4L4L4L3M4L4WOi0L4M3L4L4M3L3M4M3L43M5K5K5K5K6J7H7J7I6J7I6J7IVNeFUNU9g1UGZNd8a1fG]NW8^1SH_Nk7]1^H_Na7\\1jH`NU7Z1UIcNi6Z1_IbN`6^1cIaN[6_1iI_NU6a1nI]NQ6c1SJ[Nk5d1ZJZNd5f1_JWNa5i1cJTN\\5l1hJPNX5P2kJmMU5S2nJkMQ5U2QKiMP5V2RKiMm4V2VKhMj4X2WKgMi4Y2YKfMf4Z2\\KdMd4X2bKfM^4W2gKhMX4U2[4L4M4L4M300O1O1O100O1O1O1O100O10000000000000000O2O0000000000000000000000000O1000000000000000000000000O1000000000001O0000000000000O1000000000000000000000000O2O00000000000000000000000O100000001O00000000000000000O100000000000001O000O10000O10000OI8G9O11N2O1O0O2O1O1N2O0O2O1N2O1N2O1N4M4K6K2M101N100O101M2N2N3M2N]Po4"}, "label": "a large giraffe eating grass"}]}
{"id": 350302, "image": "COCO_train2014_000000350302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting on bench with face in hand wearing hat\"."}], "task": "refering", "answer_template": "The \"man sitting on bench with face in hand wearing hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 244, 245], "bbox": [0.350640625, 0.1805152224824356, 0.6734218750000001, 0.7048243559718969], "iscrowd": 0, "area": 19317.998799999994, "rle": {"size": [427, 640], "counts": "Yam2f0c<6K4K6J5K6K4K5K8H9G9H3M100O100O1O10O0100O1O100O10O0100O1O100O1O0011O00O1O1O1N2O11O00001N10O1N2O1O1001O001O1O00_OUMbFk2]9b0O1O10010O001O010O001O1N101N10000O100O1001O010O001O0010O2OWMgF`1W9_NkFb1T9\\NoFd1o8[NSGf1l8XNWGh1g8XNZGT2[8iMgG`2P8[MUHg2j7kLcHX3[7fLfH\\3X7cLhHa3V7^LkHc3S7\\LoHf3Z82O1O2M2O1O2M2N2N3L3N2NfLYGn1h8PNZGo1h8nMYGR2h8mMYGR2h8lMZGS2h8jMYGV2h8iMYGV2i8gMXGY2i8eMYGZ2h8aM]G^2e8\\M_Gd2b8WMcGh2_8QMfGo2[9000000000001O00000YNjE9V:GZFIf97jFYOV9g0ZGiNf8W1g100000000000O1000000000000O1000000000O10O0100O100O100OUEoN^9Q1oEBP:?aE1^:^101O00001O001O00000000O101O00000000000O10000000001O0O2O1O001O1O002N2M4M2N:Fa0_Oa0_Ob0^OUlf2"}, "label": "man sitting on bench with face in hand wearing hat"}]}
{"id": 350302, "image": "COCO_train2014_000000350302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the bench resting his head in his hand\"."}], "task": "refering", "answer_template": "The \"the man on the bench resting his head in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 244, 245], "bbox": [0.350640625, 0.1805152224824356, 0.6734218750000001, 0.7048243559718969], "iscrowd": 0, "area": 19317.998799999994, "rle": {"size": [427, 640], "counts": "Yam2f0c<6K4K6J5K6K4K5K8H9G9H3M100O100O1O10O0100O1O100O10O0100O1O100O1O0011O00O1O1O1N2O11O00001N10O1N2O1O1001O001O1O00_OUMbFk2]9b0O1O10010O001O010O001O1N101N10000O100O1001O010O001O0010O2OWMgF`1W9_NkFb1T9\\NoFd1o8[NSGf1l8XNWGh1g8XNZGT2[8iMgG`2P8[MUHg2j7kLcHX3[7fLfH\\3X7cLhHa3V7^LkHc3S7\\LoHf3Z82O1O2M2O1O2M2N2N3L3N2NfLYGn1h8PNZGo1h8nMYGR2h8mMYGR2h8lMZGS2h8jMYGV2h8iMYGV2i8gMXGY2i8eMYGZ2h8aM]G^2e8\\M_Gd2b8WMcGh2_8QMfGo2[9000000000001O00000YNjE9V:GZFIf97jFYOV9g0ZGiNf8W1g100000000000O1000000000000O1000000000O10O0100O100O100OUEoN^9Q1oEBP:?aE1^:^101O00001O001O00000000O101O00000000000O10000000001O0O2O1O001O1O002N2M4M2N:Fa0_Oa0_Ob0^OUlf2"}, "label": "the man on the bench resting his head in his hand"}]}
{"id": 350302, "image": "COCO_train2014_000000350302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a green sweater and gray pants with his hand resting on the bench ' s armrest\"."}], "task": "refering", "answer_template": "The \"a man with a green sweater and gray pants with his hand resting on the bench ' s armrest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 122, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 242, 261, 262, 264, 265], "bbox": [0.16134375, 0.09238875878220142, 0.558765625, 0.7917564402810304], "iscrowd": 0, "area": 26299.779350000004, "rle": {"size": [427, 640], "counts": "eR[14n<?@`0_Oa0I8G9H6J4L4L3M3M4L4M3SOPMXGT3`8RM^GS3Y8SMfGP3R8WMkGm2n7XMPHl2l7VMSHn2h7UMUHo2g7SMWHQ3d7RM[HQ3a7QM]HR3`7QM]HP3b7^1N2N2O001O1O001O1O1O0O2O1O001O1O1O001O1O1O0O2O010O010O0010O010O01O010O02O0O2N101N1O2N2N4L6J2N1O2SK_Hl3b7PLcHn3_8N2N2N001O1O1O1O1O1O1O2N2N2N2N1O3M2N3M3M3M3M2N3M3M2N3M3M2N3M2N2N1O1O2N1O1O2N1O2N1O1O2N1O2M2O1O2N2N2N2N2N3M1O1OQFcNP8\\1oGgNP8Y1nGjNQ8V1mGmNS8Q1lGROS8n0jGVOU8i0jGZOU8f0iG]OV8c0hG@W8?hGDW8<fGHY88eGKZ84eGOZ81dG1[8OcG5\\8KbG8]8HaG;^8DaG>_8B_Ga0`8_O]Gd0c8[O\\Gh0c8XO[Gj0e8UOZGn0e8TOVGo0j8ROQGS1n8oNmFT1S9mNhFX1W9^101O1O0O2O00lNlFlMY9T2SG\\MQ9c2l0O100O1O100O1O100O1O100O1O1O100O101O1N2O001O1O1O001O1OO1000000000001O000O101O00000012M4L5L3L5L3L4M4L3L5N1N3N1O1N3N11M8H6KM3M3M3M2N1O2N1O2N1O2N1O2N1N2O1O1O001O1N2N2N2N2O1N2N1O2N2NXfe3"}, "label": "a man with a green sweater and gray pants with his hand resting on the bench ' s armrest"}]}
{"id": 350302, "image": "COCO_train2014_000000350302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man without a hat , who is sitting on the bench\"."}], "task": "refering", "answer_template": "The \"a man without a hat , who is sitting on the bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 122, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 242, 261, 262, 264, 265], "bbox": [0.16134375, 0.09238875878220142, 0.558765625, 0.7917564402810304], "iscrowd": 0, "area": 26299.779350000004, "rle": {"size": [427, 640], "counts": "eR[14n<?@`0_Oa0I8G9H6J4L4L3M3M4L4M3SOPMXGT3`8RM^GS3Y8SMfGP3R8WMkGm2n7XMPHl2l7VMSHn2h7UMUHo2g7SMWHQ3d7RM[HQ3a7QM]HR3`7QM]HP3b7^1N2N2O001O1O001O1O1O0O2O1O001O1O1O001O1O1O0O2O010O010O0010O010O01O010O02O0O2N101N1O2N2N4L6J2N1O2SK_Hl3b7PLcHn3_8N2N2N001O1O1O1O1O1O1O2N2N2N2N1O3M2N3M3M3M3M2N3M3M2N3M3M2N3M2N2N1O1O2N1O1O2N1O2N1O1O2N1O2M2O1O2N2N2N2N2N3M1O1OQFcNP8\\1oGgNP8Y1nGjNQ8V1mGmNS8Q1lGROS8n0jGVOU8i0jGZOU8f0iG]OV8c0hG@W8?hGDW8<fGHY88eGKZ84eGOZ81dG1[8OcG5\\8KbG8]8HaG;^8DaG>_8B_Ga0`8_O]Gd0c8[O\\Gh0c8XO[Gj0e8UOZGn0e8TOVGo0j8ROQGS1n8oNmFT1S9mNhFX1W9^101O1O0O2O00lNlFlMY9T2SG\\MQ9c2l0O100O1O100O1O100O1O100O1O1O100O101O1N2O001O1O1O001O1OO1000000000001O000O101O00000012M4L5L3L5L3L4M4L3L5N1N3N1O1N3N11M8H6KM3M3M3M2N1O2N1O2N1O2N1O2N1N2O1O1O001O1N2N2N2N2O1N2N1O2N2NXfe3"}, "label": "a man without a hat , who is sitting on the bench"}]}
{"id": 317537, "image": "COCO_train2014_000000317537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small 1 person chair in the corner\"."}], "task": "refering", "answer_template": "The \"the small 1 person chair in the corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 261, 262, 263, 264, 284], "bbox": [0.290296875, 0.47010416666666666, 0.48817187500000003, 0.7339166666666667], "iscrowd": 0, "area": 8331.188500000002, "rle": {"size": [480, 640], "counts": "`]g2;`>?E:G9F9H4K10000O10O10O10000000000000000000000000O1000O1000000000000000000000O1000000002M4M2N5jCjMX;[2\\DmMc;i2O1O1O1O002N2N1N6K;EO1N2O1N1O2N2N2N2N3L4M3O1O1N2O1O1O1N2O1O1Od0[O5L4L1O1N2O1O001O1N2O1O000O101O000O2O00001N100002O1N2N101N2N2O1N2N101N2O1NN3ZNQCW1Q=iNSCQ1o<nN\\Cg0f<WOkC7W<Ia1N1O2OR_i4"}, "label": "the small 1 person chair in the corner"}]}
{"id": 317537, "image": "COCO_train2014_000000317537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the only red chair made for one person that can entirely be seen\"."}], "task": "refering", "answer_template": "The \"the only red chair made for one person that can entirely be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 261, 262, 263, 264, 284], "bbox": [0.290296875, 0.47010416666666666, 0.48817187500000003, 0.7339166666666667], "iscrowd": 0, "area": 8331.188500000002, "rle": {"size": [480, 640], "counts": "`]g2;`>?E:G9F9H4K10000O10O10O10000000000000000000000000O1000O1000000000000000000000O1000000002M4M2N5jCjMX;[2\\DmMc;i2O1O1O1O002N2N1N6K;EO1N2O1N1O2N2N2N2N3L4M3O1O1N2O1O1O1N2O1O1Od0[O5L4L1O1N2O1O001O1N2O1O000O101O000O2O00001N100002O1N2N101N2N2O1N2N101N2O1NN3ZNQCW1Q=iNSCQ1o<nN\\Cg0f<WOkC7W<Ia1N1O2OR_i4"}, "label": "the only red chair made for one person that can entirely be seen"}]}
{"id": 317537, "image": "COCO_train2014_000000317537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather and wood sofa between two chairs\"."}], "task": "refering", "answer_template": "The \"a leather and wood sofa between two chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 305], "bbox": [0.049234375000000004, 0.49822916666666667, 0.370046875, 0.797], "iscrowd": 0, "area": 16147.444150000001, "rle": {"size": [480, 640], "counts": "SX?3k>4L3O1O1O1O1N2O001O1O5K7I7Hg0ZO4L1O001O000O2O001O00001N101O00001O000O10001O0000000O100000O100000O10000000001N100000001O000O101jC`Mj;`2VDbMh;^2WDeMg;[2YDgMe;Z2ZDhMd;X2[DkMc;U2]DmMa;S2YDSNg;c2000000000O100O1O1JVDUMk;k250000O10O1000000000000000000O0100001O1O001O1O1O1N10000000000000001O0O100000000000001O0gDiLi:W3WEkLf:V3YEkLg:U3YEkLg:U3XEkLi:U3WEkLi:U3VEkLk:U3UEkLk:V3SEjLn:V3REjLn:V3QEkLn:V3REiLo:W3PEjLP;Y3mDfLS;d31N1011Nd0]OO010O1O1O1O5RO^ETMl:b2`ERMk:c2o0I5L4L3L2O000O101N10000O10000O1000O0100O10000O10000O10000O100O10O10O10000O10000O100O11O;E6J00VN^CW1i<aNYC]1V=O100O2Mj0WO>Anal5"}, "label": "a leather and wood sofa between two chairs"}]}
{"id": 317537, "image": "COCO_train2014_000000317537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two seater chair near the mirror\"."}], "task": "refering", "answer_template": "The \"the two seater chair near the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 305], "bbox": [0.049234375000000004, 0.49822916666666667, 0.370046875, 0.797], "iscrowd": 0, "area": 16147.444150000001, "rle": {"size": [480, 640], "counts": "SX?3k>4L3O1O1O1O1N2O001O1O5K7I7Hg0ZO4L1O001O000O2O001O00001N101O00001O000O10001O0000000O100000O100000O10000000001N100000001O000O101jC`Mj;`2VDbMh;^2WDeMg;[2YDgMe;Z2ZDhMd;X2[DkMc;U2]DmMa;S2YDSNg;c2000000000O100O1O1JVDUMk;k250000O10O1000000000000000000O0100001O1O001O1O1O1N10000000000000001O0O100000000000001O0gDiLi:W3WEkLf:V3YEkLg:U3YEkLg:U3XEkLi:U3WEkLi:U3VEkLk:U3UEkLk:V3SEjLn:V3REjLn:V3QEkLn:V3REiLo:W3PEjLP;Y3mDfLS;d31N1011Nd0]OO010O1O1O1O5RO^ETMl:b2`ERMk:c2o0I5L4L3L2O000O101N10000O10000O1000O0100O10000O10000O10000O100O10O10O10000O10000O100O11O;E6J00VN^CW1i<aNYC]1V=O100O2Mj0WO>Anal5"}, "label": "the two seater chair near the mirror"}]}
{"id": 333922, "image": "COCO_train2014_000000333922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ambulance behind the woman\"."}], "task": "refering", "answer_template": "The \"the ambulance behind the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 81, 82, 96, 97, 98, 99, 100, 101, 102, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 150, 151, 152, 153, 165, 166, 167, 168, 169, 174, 175, 176, 177, 188, 189, 190, 191, 197, 198, 199, 200, 211, 212, 213, 214, 221, 222, 223, 224, 234, 235, 236, 237, 243, 244, 245, 246, 247, 248, 257, 258, 259, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294], "bbox": [0.174625, 0.19931873479318735, 0.802375, 0.8689781021897811], "iscrowd": 0, "area": 49327.006649999996, "rle": {"size": [411, 640], "counts": "jW]13e<3L4aJP4oITLP6n3lIULT6l3gIXLX6j3dIXL]6i3^I[Lb6f3YI^Lf6d3VI_Lj6b3QIbLo6e4O100O10000O100O104L:E;F3M0O100O10000O100O10000O10O1O010O001O10O01O00100O00100O001O010O1O00100O001O010O1O0100O10O10O10O10O100O01000O10O010O1L3N3L4M2M4M2N3L4M2M4M3L4M3L4M3M3L4M3L4M3L4N2N2M3N2M3N2M3N2M3N3L3N2N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2O100O100O100O100O100O100O100O100O1O2O0O100O100O100O100kNfDh0[;UOkDh0U;UOREf0o:WOXEe0b;L3N2M2O1N2O2M2O1N3N^ea03]Z^O3L4M3M3M3L4M3M3M3K6K5K5K5K5L4N2O1N2N2O1N2N2O1N2N2O1nJdMA^29fMG\\22iMOX2JmM6U2DoM<S2]ORNd0o1mM_OR2c0lM]OU2d0jM[OV2g0hMYOX2i0gMVOZ2k0dMUO\\2m0cMRO^2o0`MQO`2Q1_MnNb2T1\\MkNd2W1ZMiNg2X1XMgNh2[1VMeNk2\\1TMcNl2_1RMaNn2a1QM^NP3c1nL]NR3e1mLZNT3g1jLYNV3i1iLVNX3k1fLUNZ3m1eLRN\\3o1bLQN^3Q2aLnM`3S2^LmMb3U2]LjMc3X2\\LgMe3Z2YLfMg3\\2XLcMi3^2ULbMk3`2TL_Mm3a2RL_Mn3b2RL]Mo3c2PL]MP4d2PL[MQ4e2nK[MR4g2mKXMS4i2lKWMU4i2kKVMU4k2jKUMW4k2iKTMW4m2iKRMX4n2gKRMY4P3fKoL[4Q3dKoL\\4R3dKmL]4S3bKmL^4T3bKkL^4V3aKjL`4V3`KiL`4X3_KhLb4Y3]KfLc4[3\\KeLe4[3[KdLe4]3ZKcLg4]3YKbLg4_3XKaLh4`3WK`Lj4`3SKbLm4`3oJbLR5^3lJcLT5^3iJdLX5\\3eJfL[5[3cJfL^5Z3_JhLa5Y3\\JjLd5W3XJkLh5T5001O001O001O001O001O0VJUJd4k5[KjJQ4W5mKYKd3g4ZL\\Ke3e4YL]Kf3c4XL`Kg3a4WLbKg3^4XLdKg3]4VLfKi3Z4VLhKi3Y4ULiKj3W4TLmKj3S4ULoKj3R4SLQLl3o3SLSLl3n3RLULl3k3RLXLm3i3QLYLn3g3PL\\Lo3[6M3M3M4L3M3M4L3M3M3M2N1O1O1O1N2N101N2N2O1N2O0O2N2O1N2N2O0O2N2O1N2O1N1O2O1N2N2O1N1O2O1N2N4J6J5J7J6J5K6J5KfXb1"}, "label": "the ambulance behind the woman"}]}
{"id": 333922, "image": "COCO_train2014_000000333922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the medical truck directly behind the woman\"."}], "task": "refering", "answer_template": "The \"the medical truck directly behind the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 81, 82, 96, 97, 98, 99, 100, 101, 102, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 150, 151, 152, 153, 165, 166, 167, 168, 169, 174, 175, 176, 177, 188, 189, 190, 191, 197, 198, 199, 200, 211, 212, 213, 214, 221, 222, 223, 224, 234, 235, 236, 237, 243, 244, 245, 246, 247, 248, 257, 258, 259, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294], "bbox": [0.174625, 0.19931873479318735, 0.802375, 0.8689781021897811], "iscrowd": 0, "area": 49327.006649999996, "rle": {"size": [411, 640], "counts": "jW]13e<3L4aJP4oITLP6n3lIULT6l3gIXLX6j3dIXL]6i3^I[Lb6f3YI^Lf6d3VI_Lj6b3QIbLo6e4O100O10000O100O104L:E;F3M0O100O10000O100O10000O10O1O010O001O10O01O00100O00100O001O010O1O00100O001O010O1O0100O10O10O10O10O100O01000O10O010O1L3N3L4M2M4M2N3L4M2M4M3L4M3L4M3M3L4M3L4M3L4N2N2M3N2M3N2M3N2M3N3L3N2N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2O100O100O100O100O100O100O100O100O1O2O0O100O100O100O100kNfDh0[;UOkDh0U;UOREf0o:WOXEe0b;L3N2M2O1N2O2M2O1N3N^ea03]Z^O3L4M3M3M3L4M3M3M3K6K5K5K5K5L4N2O1N2N2O1N2N2O1N2N2O1nJdMA^29fMG\\22iMOX2JmM6U2DoM<S2]ORNd0o1mM_OR2c0lM]OU2d0jM[OV2g0hMYOX2i0gMVOZ2k0dMUO\\2m0cMRO^2o0`MQO`2Q1_MnNb2T1\\MkNd2W1ZMiNg2X1XMgNh2[1VMeNk2\\1TMcNl2_1RMaNn2a1QM^NP3c1nL]NR3e1mLZNT3g1jLYNV3i1iLVNX3k1fLUNZ3m1eLRN\\3o1bLQN^3Q2aLnM`3S2^LmMb3U2]LjMc3X2\\LgMe3Z2YLfMg3\\2XLcMi3^2ULbMk3`2TL_Mm3a2RL_Mn3b2RL]Mo3c2PL]MP4d2PL[MQ4e2nK[MR4g2mKXMS4i2lKWMU4i2kKVMU4k2jKUMW4k2iKTMW4m2iKRMX4n2gKRMY4P3fKoL[4Q3dKoL\\4R3dKmL]4S3bKmL^4T3bKkL^4V3aKjL`4V3`KiL`4X3_KhLb4Y3]KfLc4[3\\KeLe4[3[KdLe4]3ZKcLg4]3YKbLg4_3XKaLh4`3WK`Lj4`3SKbLm4`3oJbLR5^3lJcLT5^3iJdLX5\\3eJfL[5[3cJfL^5Z3_JhLa5Y3\\JjLd5W3XJkLh5T5001O001O001O001O001O0VJUJd4k5[KjJQ4W5mKYKd3g4ZL\\Ke3e4YL]Kf3c4XL`Kg3a4WLbKg3^4XLdKg3]4VLfKi3Z4VLhKi3Y4ULiKj3W4TLmKj3S4ULoKj3R4SLQLl3o3SLSLl3n3RLULl3k3RLXLm3i3QLYLn3g3PL\\Lo3[6M3M3M4L3M3M4L3M3M3M2N1O1O1O1N2N101N2N2O1N2O0O2N2O1N2N2O0O2N2O1N2O1N1O2O1N2N2O1N1O2O1N2N4J6J5J7J6J5K6J5KfXb1"}, "label": "the medical truck directly behind the woman"}]}
{"id": 333922, "image": "COCO_train2014_000000333922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a medical truck to the right of another truck\"."}], "task": "refering", "answer_template": "The \"a medical truck to the right of another truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321], "bbox": [0.671109375, 0.306301703163017, 1.0, 0.9279318734793187], "iscrowd": 0, "area": 35447.550449999995, "rle": {"size": [411, 640], "counts": "kS]51j<000O101O000000001N1000001O000SDT1k9lNTFY1g9hNXF^1c9cN\\F_1b9aN]Fb1a9^N^Fd1a9\\N_Fe1`9\\N^Ff1a9ZN^Fi1_9XNaFi1^9WNaFk1^9VN`Fm1^9SNbFo1\\9QNcFS2Z9mMeFW2X9iMhFY2U9nMdFV2Y9i0M2N3M1O2M2O2N1O1O2N1O2M2O1O2N1O2N1O1N3N1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O100O1O1O1O1O0000001O0000001O0000000010O000001O000000001O00001O1O1O1O1O1O100O1O1O001O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O001O1O1O1O1O1O1O1O1O5K5L4K6J5K5K5K2N001O001O001O001O1O001O010O001O0000000000000000000000000001O1O001O001O001O1O001O001O001O1O001O001N101O1O0O2O1O1O1N2O2XJ_JR4b5dKPKS4Q5dK`KR4b4dK_Nd1b1SNd0O_J"}, "label": "a medical truck to the right of another truck"}]}
{"id": 333922, "image": "COCO_train2014_000000333922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck that is behind another truck , we can only see the front\"."}], "task": "refering", "answer_template": "The \"a truck that is behind another truck , we can only see the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321], "bbox": [0.671109375, 0.306301703163017, 1.0, 0.9279318734793187], "iscrowd": 0, "area": 35447.550449999995, "rle": {"size": [411, 640], "counts": "kS]51j<000O101O000000001N1000001O000SDT1k9lNTFY1g9hNXF^1c9cN\\F_1b9aN]Fb1a9^N^Fd1a9\\N_Fe1`9\\N^Ff1a9ZN^Fi1_9XNaFi1^9WNaFk1^9VN`Fm1^9SNbFo1\\9QNcFS2Z9mMeFW2X9iMhFY2U9nMdFV2Y9i0M2N3M1O2M2O2N1O1O2N1O2M2O1O2N1O2N1O1N3N1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O100O1O1O1O1O0000001O0000001O0000000010O000001O000000001O00001O1O1O1O1O1O100O1O1O001O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O001O1O1O1O1O1O1O1O1O5K5L4K6J5K5K5K2N001O001O001O001O1O001O010O001O0000000000000000000000000001O1O001O001O001O1O001O001O001O1O001O001N101O1O0O2O1O1O1N2O2XJ_JR4b5dKPKS4Q5dK`KR4b4dK_Nd1b1SNd0O_J"}, "label": "a truck that is behind another truck , we can only see the front"}]}
{"id": 63587, "image": "COCO_train2014_000000063587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle giraffe in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the middle giraffe in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 177, 178, 195, 196, 212, 213, 214, 229, 230, 231, 232, 246, 247, 248, 263, 264, 265, 280, 281, 282], "bbox": [0.4036458333333333, 0.428296875, 0.6566666666666666, 0.7372343750000001], "iscrowd": 0, "area": 8575.367600000001, "rle": {"size": [640, 480], "counts": "aai32lc02N1N3N1N3H8F9100O10000O01000O100O101O0O2O001001N101N101N101N101O001O0O2O00001O0O2O001O001N101O001g]OiNea0Y1R^OROka0f1L3M4SAgMj;]2mChMR<a2aCcM`<d2UC_Ml<h2hB\\MY=k2[BXMg=_41K6K5J6[Oe0K401O1N101N11O2N1O1O2N1O2N101NQOU@kMm?k1]@TNc?d1f@ZN[?`1l@^NU?]1RA`No>[1WAcNg>[1_AcN`>Z1gAcNW>[1oAcNP>^1oAcNo=_1PBaNP>_1PBbNo=_1PBbNn=`1QB`No=a1PB`Nn=b1QB[NR>f1mAWNU>k1jAQNZ>P2eAlM_>T2f12j_OmM\\>T2\\ARNd>P2TAWNl>j1g@aNY?a1\\@hNe?Y1W@jNi?V1S@mNm?c210O01O10O02cNn_OhNT`0P1X@iNk?o0`@kNc?n0h@kN[?<bA]Oa>:kA@X>9SB@P>9o2JmhV3"}, "label": "the middle giraffe in the right hand picture"}]}
{"id": 63587, "image": "COCO_train2014_000000063587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 giraffes standing near each other , looking in opposite directions\"."}], "task": "refering", "answer_template": "The \"2 giraffes standing near each other , looking in opposite directions\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 177, 178, 195, 196, 212, 213, 214, 229, 230, 231, 232, 246, 247, 248, 263, 264, 265, 280, 281, 282], "bbox": [0.4036458333333333, 0.428296875, 0.6566666666666666, 0.7372343750000001], "iscrowd": 0, "area": 8575.367600000001, "rle": {"size": [640, 480], "counts": "aai32lc02N1N3N1N3H8F9100O10000O01000O100O101O0O2O001001N101N101N101N101O001O0O2O00001O0O2O001O001N101O001g]OiNea0Y1R^OROka0f1L3M4SAgMj;]2mChMR<a2aCcM`<d2UC_Ml<h2hB\\MY=k2[BXMg=_41K6K5J6[Oe0K401O1N101N11O2N1O1O2N1O2N101NQOU@kMm?k1]@TNc?d1f@ZN[?`1l@^NU?]1RA`No>[1WAcNg>[1_AcN`>Z1gAcNW>[1oAcNP>^1oAcNo=_1PBaNP>_1PBbNo=_1PBbNn=`1QB`No=a1PB`Nn=b1QB[NR>f1mAWNU>k1jAQNZ>P2eAlM_>T2f12j_OmM\\>T2\\ARNd>P2TAWNl>j1g@aNY?a1\\@hNe?Y1W@jNi?V1S@mNm?c210O01O10O02cNn_OhNT`0P1X@iNk?o0`@kNc?n0h@kN[?<bA]Oa>:kA@X>9SB@P>9o2JmhV3"}, "label": "2 giraffes standing near each other , looking in opposite directions"}]}
{"id": 63587, "image": "COCO_train2014_000000063587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffes with crossed necks\"."}], "task": "refering", "answer_template": "The \"two giraffes with crossed necks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 196, 197, 211, 212, 228, 229, 245, 262], "bbox": [0.424375, 0.439046875, 0.7012708333333334, 0.691875], "iscrowd": 0, "area": 3514.420900000001, "rle": {"size": [640, 480], "counts": "Ylo3[1ab06K5J6K4K3N3L4M3L3OL5H7I7M4O001O1O1N3N1O2N1010O010O01O010O010O0010O02N101N101N1O2N1OnRb0ROPn]O2N1O2O1N2N2N2N2N3M2N2N001O0001O00000O10000O100O10000O100O2O1O1N2O1N21O0O10001O00001O00010O0000001O01O100O1O1O1O1O2L5K4KkRi2"}, "label": "two giraffes with crossed necks"}]}
{"id": 63587, "image": "COCO_train2014_000000063587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe that seems to be facing the tallest giraffe\"."}], "task": "refering", "answer_template": "The \"the giraffe that seems to be facing the tallest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 196, 197, 211, 212, 228, 229, 245, 262], "bbox": [0.424375, 0.439046875, 0.7012708333333334, 0.691875], "iscrowd": 0, "area": 3514.420900000001, "rle": {"size": [640, 480], "counts": "Ylo3[1ab06K5J6K4K3N3L4M3L3OL5H7I7M4O001O1O1N3N1O2N1010O010O01O010O010O0010O02N101N101N1O2N1OnRb0ROPn]O2N1O2O1N2N2N2N2N3M2N2N001O0001O00000O10000O100O10000O100O2O1O1N2O1N21O0O10001O00001O00010O0000001O01O100O1O1O1O1O2L5K4KkRi2"}, "label": "the giraffe that seems to be facing the tallest giraffe"}]}
{"id": 63587, "image": "COCO_train2014_000000063587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest giraffe\"."}], "task": "refering", "answer_template": "The \"the tallest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 165, 166, 167, 168, 169, 184, 185, 186, 202, 203, 220, 237, 254, 271, 288], "bbox": [0.7262083333333333, 0.360796875, 1.0, 0.73615625], "iscrowd": 0, "area": 7060.341049999999, "rle": {"size": [640, 480], "counts": "a\\j65hc06J4M2M2N3N2N1N3N1O2M3DoNg]OR1Yb0:010O2N2N2N2O2M2N3M2N1O100O1O2N100001O001O1O001O1O001O001O1O0010O01O10O0100O010O010O10O01O010O10O0100O010O010O1O010O10O010O0100O0010O0100O010O1O100O100O00100O100O10O01O100O100o]OfNZa0[1a^OmN[a0T1a^ORO]a0l1L4M3TB\\MT:g2gE[MY:i2aEZM_:i2[EYMe:k2VEVMk:m2oDUMQ;n2jDTMW;o2cDSM];Q3]DQMc;Y3QDiLP<b3]CeLc<f3_BmLb=h400001O001O0jNV1kMSAQOT?P1`AUNe>l1PBXMW>i2_12OXOg0bNgH"}, "label": "the tallest giraffe"}]}
{"id": 63587, "image": "COCO_train2014_000000063587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe to the far right of the group\"."}], "task": "refering", "answer_template": "The \"the giraffe to the far right of the group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 165, 166, 167, 168, 169, 184, 185, 186, 202, 203, 220, 237, 254, 271, 288], "bbox": [0.7262083333333333, 0.360796875, 1.0, 0.73615625], "iscrowd": 0, "area": 7060.341049999999, "rle": {"size": [640, 480], "counts": "a\\j65hc06J4M2M2N3N2N1N3N1O2M3DoNg]OR1Yb0:010O2N2N2N2O2M2N3M2N1O100O1O2N100001O001O1O001O1O001O001O1O0010O01O10O0100O010O010O10O01O010O10O0100O010O010O1O010O10O010O0100O0010O0100O010O1O100O100O00100O100O10O01O100O100o]OfNZa0[1a^OmN[a0T1a^ORO]a0l1L4M3TB\\MT:g2gE[MY:i2aEZM_:i2[EYMe:k2VEVMk:m2oDUMQ;n2jDTMW;o2cDSM];Q3]DQMc;Y3QDiLP<b3]CeLc<f3_BmLb=h400001O001O0jNV1kMSAQOT?P1`AUNe>l1PBXMW>i2_12OXOg0bNgH"}, "label": "the giraffe to the far right of the group"}]}
{"id": 522342, "image": "COCO_train2014_000000522342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the parking meter on the right\"."}], "task": "refering", "answer_template": "The \"the parking meter on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.44157812500000004, 0.16141666666666668, 0.9893281250000001, 0.9906458333333333], "iscrowd": 0, "area": 68549.45039999999, "rle": {"size": [480, 640], "counts": "hQU48f>8G9G9H8G8H9H8G9G9H7dMWMTHl2T7oMfHT2W7QNeHQ2Z7SN`HP2_7TN[Ho1d7WNUHk1k7\\NkGg1T8aNbGb1]8fNZG\\1e8`2O100O2N1O1O1O100O1O1O1O100O1O1O1O100O1O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2M3N2M3N2M3M3N2M3M3N2O1O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100000000001O000000001O000000001O000000001O000000001O0000001O001O1O1O1O001O100O1O001O1O1O1O001O1O1O1O0M2N2N2O1N2oMhIYKZ6f4\\JcJf5[5jJVJW5i5o101O1O1O1O1O1N101O1O1O1O1O1O1N101O001O001O001O0O2O1O001O001O001N101O001O001O001N101O00000011N2N2O1N1O2O1N2N2O1TO`JVHb5j7hJjGY5W8h00000O100000O10000O001O1N2O001N101O001N101N101N101N102M2N3N1N2O2M2O1O2N101N1O101N100O2N101N100O2N100O2N101N100O2N1O1O2N1O2M2O1O2N1O1N3N1O2N1N2O2N1O1O2M2O2N1O1O2M2O1O2N1M4M2N2N3M2M3N3M2N3L3N2N3XOg0O1O2N1N3N1O1O2M2O1O2N1O2M2O1O2N3L5L4L3M4L4K4MQU3"}, "label": "the parking meter on the right"}]}
{"id": 522342, "image": "COCO_train2014_000000522342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the parking meter on the right\"."}], "task": "refering", "answer_template": "The \"the parking meter on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.44157812500000004, 0.16141666666666668, 0.9893281250000001, 0.9906458333333333], "iscrowd": 0, "area": 68549.45039999999, "rle": {"size": [480, 640], "counts": "hQU48f>8G9G9H8G8H9H8G9G9H7dMWMTHl2T7oMfHT2W7QNeHQ2Z7SN`HP2_7TN[Ho1d7WNUHk1k7\\NkGg1T8aNbGb1]8fNZG\\1e8`2O100O2N1O1O1O100O1O1O1O100O1O1O1O100O1O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2M3N2M3N2M3M3N2M3M3N2O1O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100000000001O000000001O000000001O000000001O000000001O0000001O001O1O1O1O001O100O1O001O1O1O1O001O1O1O1O0M2N2N2O1N2oMhIYKZ6f4\\JcJf5[5jJVJW5i5o101O1O1O1O1O1N101O1O1O1O1O1O1N101O001O001O001O0O2O1O001O001O001N101O001O001O001N101O00000011N2N2O1N1O2O1N2N2O1TO`JVHb5j7hJjGY5W8h00000O100000O10000O001O1N2O001N101O001N101N101N101N102M2N3N1N2O2M2O1O2N101N1O101N100O2N101N100O2N100O2N101N100O2N1O1O2N1O2M2O1O2N1O1N3N1O2N1N2O2N1O1O2M2O2N1O1O2M2O1O2N1M4M2N2N3M2M3N3M2N3L3N2N3XOg0O1O2N1N3N1O1O2M2O1O2N1O2M2O1O2N3L5L4L3M4L4K4MQU3"}, "label": "the parking meter on the right"}]}
{"id": 522342, "image": "COCO_train2014_000000522342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a parking meter to the left of another parking meter\"."}], "task": "refering", "answer_template": "The \"a parking meter to the left of another parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0169375, 0.1168125, 0.55021875, 0.981375], "iscrowd": 0, "area": 77270.5363, "rle": {"size": [480, 640], "counts": "mk51o>011N3M7I`0@`0@`0@a0_O`0@3N1N2N2N2N2N2N2N2O0O2N2N2N2N2N2N2N2O1N2N1O2N2N2N2N2N2N2N2N2N2L4K4M4K5K5M3M3M3M3M3L4M3M3M2N3L4M3M3M3M3M3L4M3N1N3N2N2M3N2N2M10000O10000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O1000000000001O00000000000001O0000000000000000000000001O0001O00000000000000000001O0000000000001O00001O001O010O00001O001O001O00001O001OlKWHQ1i7nNYHR1f7nNZHR1f7mN\\HR1d7nN\\HS1c7lN_HS1a7mN`HS1_7lNbHT1^7lNcHS1]7lNeHT1Z7lNfHT1[7jNgHV1X7jNiHU1W7jNjHW1U7iNlHV1T7iNmHW1S7hNoHX1P7hNQIW1o6hNRIY1m6gNTIX1l6gNVIY1i6cN[I]1e6`N_I_1a6^NcIb1\\6[NnI^1R6^NZJ[1e5bNeJW1[5iNlJQ1S5oNTKj0l4VO[Kc0e4]OcK<\\4EjK4V4LQLNn32XLHi37^LCa3=fL\\OZ3d0lLVOT3j0SMPOl2P1[MjNd2V1bMfNc1]OTJn1_4_NY1JTJg1j4ZNn05TJa1T5UNd0`0UJ[1^5PN9k0UJU1b;QO[Do0b;Y1N2M4L3N2M3N2M3M3N2M3N3L3M3N2M3N2M3M3N3L3N3L3M3N3L3N3L3M3N3L3M4M2M3N3L3M4M2M3N3L3M4M2M4M2M3O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2N1N3N1O1O2N1N3N1O2N1O1N3N1O2N1O1N3N1O2N1O1N4M2N2N3M2M3N2N3L3M3M4K4M3M4K4M3M3L5L3M3L5L3L4M3M4K4M3M4K4M;E?@`0A`0@?@YjV4"}, "label": "a parking meter to the left of another parking meter"}]}
{"id": 522342, "image": "COCO_train2014_000000522342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a opened meter with the top opened at a 90 degree angle\"."}], "task": "refering", "answer_template": "The \"a opened meter with the top opened at a 90 degree angle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0169375, 0.1168125, 0.55021875, 0.981375], "iscrowd": 0, "area": 77270.5363, "rle": {"size": [480, 640], "counts": "mk51o>011N3M7I`0@`0@`0@a0_O`0@3N1N2N2N2N2N2N2N2O0O2N2N2N2N2N2N2N2O1N2N1O2N2N2N2N2N2N2N2N2N2L4K4M4K5K5M3M3M3M3M3L4M3M3M2N3L4M3M3M3M3M3L4M3N1N3N2N2M3N2N2M10000O10000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O1000000000001O00000000000001O0000000000000000000000001O0001O00000000000000000001O0000000000001O00001O001O010O00001O001O001O00001O001OlKWHQ1i7nNYHR1f7nNZHR1f7mN\\HR1d7nN\\HS1c7lN_HS1a7mN`HS1_7lNbHT1^7lNcHS1]7lNeHT1Z7lNfHT1[7jNgHV1X7jNiHU1W7jNjHW1U7iNlHV1T7iNmHW1S7hNoHX1P7hNQIW1o6hNRIY1m6gNTIX1l6gNVIY1i6cN[I]1e6`N_I_1a6^NcIb1\\6[NnI^1R6^NZJ[1e5bNeJW1[5iNlJQ1S5oNTKj0l4VO[Kc0e4]OcK<\\4EjK4V4LQLNn32XLHi37^LCa3=fL\\OZ3d0lLVOT3j0SMPOl2P1[MjNd2V1bMfNc1]OTJn1_4_NY1JTJg1j4ZNn05TJa1T5UNd0`0UJ[1^5PN9k0UJU1b;QO[Do0b;Y1N2M4L3N2M3N2M3M3N2M3N3L3M3N2M3N2M3M3N3L3N3L3M3N3L3N3L3M3N3L3M4M2M3N3L3M4M2M3N3L3M4M2M4M2M3O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2N1N3N1O1O2N1N3N1O2N1O1N3N1O2N1O1N3N1O2N1O1N4M2N2N3M2M3N2N3L3M3M4K4M3M4K4M3M3L5L3M3L5L3L4M3M4K4M3M4K4M;E?@`0A`0@?@YjV4"}, "label": "a opened meter with the top opened at a 90 degree angle"}]}
{"id": 55402, "image": "COCO_train2014_000000055402.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe at a beautiful open place\"."}], "task": "refering", "answer_template": "The \"giraffe at a beautiful open place\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 173, 174, 175, 196, 197, 218, 219, 220, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334, 354, 355, 356, 357, 379, 380], "bbox": [0.40418750000000003, 0.3526666666666667, 0.633578125, 0.9781666666666666], "iscrowd": 0, "area": 13175.675850000023, "rle": {"size": [480, 640], "counts": "Rii38^=2`C1\\OG_<j0jCACGb<Q1cCCZ<d0aC\\O]<e1N3L5bD^MX:f2ZEiMa:\\3N2N2OAUEkLj:j2iEoLV:l2R1L3N2HhD^M\\ONQ;W3lDPM1KQ;a3mD`LS;g300O01^OWEmLELP;X3\\EgLF3n:k3000000O3N5K8HkNeEeMZ:R2UFiMk9S2]FiMc9S2hFfMW9W2TGcMj8[2]G`Md8]2P2N2N2N3M2N2010O1O1O1O1O1O1O1O2N3M;E8H7If0ZO5VFoK\\8^4gFSLX9k4O2M2O2NfNVGiKL8m8l3RHlKm7S3cGoLe0Hi7U3iGkLk9R3R1N2K4102N1O1N2gDPMa:P3mDRNc:c3A4[GVKo6P5kHQKS7^5kGlJNHU8T6hGnIV8_6L4O1OXNUH]Lh7Q3^HbK;o0U7U3VJ^Lh5^3bJ^L\\5`3oJYLo4f3UKYLi4f3[KYLc4V2_HoNR3jN[4S2lHnNn2mNT4Q2UIhNb:S1hEbNR9^O[Go1IXNPOKm9OWGl1_:VN]Ej1f:XNUEh1m:YNQEf1R;[NjDe1Y;\\NcDd1`;]N\\Db1h;^NUDa1o;_NoCa1T<^NiCb1Z<^NdCa1`<]N_C\\1j<dNTCZ1P=fNnBX1W=43N2M4K4L3M03002N3M5L3NO2J6O02N100O3M3NO100O1O[OoA;Q>EQB:m=GUB8j=HWB7i=IXB6h=IZB6f=J[B5f=I_B1c=O^BOc=Oh0O\\o]3"}, "label": "giraffe at a beautiful open place"}]}
{"id": 55402, "image": "COCO_train2014_000000055402.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"majestic and beautiful giraffe\"."}], "task": "refering", "answer_template": "The \"majestic and beautiful giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 173, 174, 175, 196, 197, 218, 219, 220, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334, 354, 355, 356, 357, 379, 380], "bbox": [0.40418750000000003, 0.3526666666666667, 0.633578125, 0.9781666666666666], "iscrowd": 0, "area": 13175.675850000023, "rle": {"size": [480, 640], "counts": "Rii38^=2`C1\\OG_<j0jCACGb<Q1cCCZ<d0aC\\O]<e1N3L5bD^MX:f2ZEiMa:\\3N2N2OAUEkLj:j2iEoLV:l2R1L3N2HhD^M\\ONQ;W3lDPM1KQ;a3mD`LS;g300O01^OWEmLELP;X3\\EgLF3n:k3000000O3N5K8HkNeEeMZ:R2UFiMk9S2]FiMc9S2hFfMW9W2TGcMj8[2]G`Md8]2P2N2N2N3M2N2010O1O1O1O1O1O1O1O2N3M;E8H7If0ZO5VFoK\\8^4gFSLX9k4O2M2O2NfNVGiKL8m8l3RHlKm7S3cGoLe0Hi7U3iGkLk9R3R1N2K4102N1O1N2gDPMa:P3mDRNc:c3A4[GVKo6P5kHQKS7^5kGlJNHU8T6hGnIV8_6L4O1OXNUH]Lh7Q3^HbK;o0U7U3VJ^Lh5^3bJ^L\\5`3oJYLo4f3UKYLi4f3[KYLc4V2_HoNR3jN[4S2lHnNn2mNT4Q2UIhNb:S1hEbNR9^O[Go1IXNPOKm9OWGl1_:VN]Ej1f:XNUEh1m:YNQEf1R;[NjDe1Y;\\NcDd1`;]N\\Db1h;^NUDa1o;_NoCa1T<^NiCb1Z<^NdCa1`<]N_C\\1j<dNTCZ1P=fNnBX1W=43N2M4K4L3M03002N3M5L3NO2J6O02N100O3M3NO100O1O[OoA;Q>EQB:m=GUB8j=HWB7i=IXB6h=IZB6f=J[B5f=I_B1c=O^BOc=Oh0O\\o]3"}, "label": "majestic and beautiful giraffe"}]}
{"id": 284778, "image": "COCO_train2014_000000284778.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with a broken left tusk\"."}], "task": "refering", "answer_template": "The \"an elephant with a broken left tusk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 305, 306, 307, 308, 309, 310, 311, 312, 313, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 344], "bbox": [0.196515625, 0.0045047169811320755, 0.9989531249999999, 0.9865094339622642], "iscrowd": 0, "area": 131987.93145, "rle": {"size": [424, 640], "counts": "Xbd14R=200O2O0O10001O0O101O000000000O100000001O0O1000000000000O10000000000O10000000000O1000000000000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1000000O10O1O001O1O010O1O001O1O0014L5J2O1O0O2C=F9K5J6J6K5J6K5K5M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3G9M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3N2M3M3M3N2M3M3N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2N2O1N2N2N2N101N2N2N2O1N2N2N2N2O1N2N2N2N2N`NfMVHY2e7f1003M2N2N2O2M2N2N2N6J6K4K6J6J5L5J6K3M001O1O1N2O1O1O1O001O1N2O1O1O1O1O001N2O1O1O1O01N2O1N2O1N2O1N2O1O001O1000O1000O1000O10O1000O1000O1000O1000O1000O1000O100000000O100000000O100000000O100000000O100000000O10O100000O10000N2N2O1N2O1004L4L4L5K4L4L4L4L2N00O10000000000000000O1000000000000000000001O00000001O0000000000000O1000000000000000000000000000O10000O100O100O10000O100O10000O1000000O10000O101O000O10000O1000000O10000O1000000O2O000O1000000O1000000O1000000O10001O0O1000000O10000O1000000O10001O0O3N2N2N3L3N2N3M2M3N3M2N1N10000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000O01000f0ZO_6aIQ6"}, "label": "an elephant with a broken left tusk"}]}
{"id": 284778, "image": "COCO_train2014_000000284778.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two elephants\"."}], "task": "refering", "answer_template": "The \"two elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 305, 306, 307, 308, 309, 310, 311, 312, 313, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 342, 343, 344], "bbox": [0.196515625, 0.0045047169811320755, 0.9989531249999999, 0.9865094339622642], "iscrowd": 0, "area": 131987.93145, "rle": {"size": [424, 640], "counts": "Xbd14R=200O2O0O10001O0O101O000000000O100000001O0O1000000000000O10000000000O10000000000O1000000000000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O1000000O10O1O001O1O010O1O001O1O0014L5J2O1O0O2C=F9K5J6J6K5J6K5K5M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3G9M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3N2M3M3M3N2M3M3N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2N2O1N2N2N2N101N2N2N2O1N2N2N2N2O1N2N2N2N2N`NfMVHY2e7f1003M2N2N2O2M2N2N2N6J6K4K6J6J5L5J6K3M001O1O1N2O1O1O1O001O1N2O1O1O1O1O001N2O1O1O1O01N2O1N2O1N2O1N2O1O001O1000O1000O1000O10O1000O1000O1000O1000O1000O1000O100000000O100000000O100000000O100000000O100000000O10O100000O10000N2N2O1N2O1004L4L4L5K4L4L4L4L2N00O10000000000000000O1000000000000000000001O00000001O0000000000000O1000000000000000000000000000O10000O100O100O10000O100O10000O1000000O10000O101O000O10000O1000000O10000O1000000O2O000O1000000O1000000O1000000O10001O0O1000000O10000O1000000O10001O0O3N2N2N3L3N2N3M2M3N3M2N1N10000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000O01000f0ZO_6aIQ6"}, "label": "two elephants"}]}
{"id": 284778, "image": "COCO_train2014_000000284778.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large grey african elephants standing behind the first elephant\"."}], "task": "refering", "answer_template": "The \"a large grey african elephants standing behind the first elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 55, 56, 57, 58, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 278, 279, 281, 282, 283, 304, 305, 306, 327], "bbox": [0.093578125, 0.0, 0.7219218749999999, 0.9461792452830189], "iscrowd": 0, "area": 64465.63610000001, "rle": {"size": [424, 640], "counts": "[Ui03S=5L5K4L2M3N1O2M2O1O2N1N3N1O2M2O1O2M2O2N000O10000O1000000O10000O10000O10O1000O10000O10000O101O001N2SL[NVLe1f3gNRLZ1j3QOmKP1o3\\OiKe0R4GeK:X4OaK3Z44cKMX49eKHW4?fKBV4d0fK]OV4j0gKWOT4P1hKROT4T1hKmNT4Z1hKhNS4U2SKlMi4\\5L4K4L4M4K4M310O3M3M4M5J6J7I6K5J6J6J7J5J6J6J6K5J7SNgD^1b;O10O01O1O1O010O1O1O010O1O1O010O1O1O010O1O1O010O1000O1000000O10O1000O100000000O10O1000O1000000O1000O10O1000000O100000O01000000O1000000O0100000O1POP1M3M3L4M2M4M3L4M3M3L4M3M3N1N3N2N2\\Od0L4L4L4L4L4M3L4L4L4L4L4N2O1O1O1O1O1O1O1N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2M2O2N2N2N2M3N2N2N2M3N2N1O2M3N2N2N2N2M3N2N2N2N101N2O1N2O1N2O1N2O1N2O1O0O2O1N2O1N2O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2O2N1N2O1N2O1N2O1N2O1N2O1N3N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N20001O000000000000000000001O000000000000000000001O000000001O001O001O001N101O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O00aRZ2"}, "label": "a large grey african elephants standing behind the first elephant"}]}
{"id": 284778, "image": "COCO_train2014_000000284778.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 elephants with white tusks standing next to each other\"."}], "task": "refering", "answer_template": "The \"2 elephants with white tusks standing next to each other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 55, 56, 57, 58, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 278, 279, 281, 282, 283, 304, 305, 306, 327], "bbox": [0.093578125, 0.0, 0.7219218749999999, 0.9461792452830189], "iscrowd": 0, "area": 64465.63610000001, "rle": {"size": [424, 640], "counts": "[Ui03S=5L5K4L2M3N1O2M2O1O2N1N3N1O2M2O1O2M2O2N000O10000O1000000O10000O10000O10O1000O10000O10000O101O001N2SL[NVLe1f3gNRLZ1j3QOmKP1o3\\OiKe0R4GeK:X4OaK3Z44cKMX49eKHW4?fKBV4d0fK]OV4j0gKWOT4P1hKROT4T1hKmNT4Z1hKhNS4U2SKlMi4\\5L4K4L4M4K4M310O3M3M4M5J6J7I6K5J6J6J7J5J6J6J6K5J7SNgD^1b;O10O01O1O1O010O1O1O010O1O1O010O1O1O010O1O1O010O1000O1000000O10O1000O100000000O10O1000O1000000O1000O10O1000000O100000O01000000O1000000O0100000O1POP1M3M3L4M2M4M3L4M3M3L4M3M3N1N3N2N2\\Od0L4L4L4L4L4M3L4L4L4L4L4N2O1O1O1O1O1O1O1N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2M2O2N2N2N2M3N2N2N2M3N2N1O2M3N2N2N2N2M3N2N2N2N101N2O1N2O1N2O1N2O1N2O1O0O2O1N2O1N2O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2O2N1N2O1N2O1N2O1N2O1N2O1N3N1O1N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N20001O000000000000000000001O000000000000000000001O000000001O001O001O001N101O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O00aRZ2"}, "label": "2 elephants with white tusks standing next to each other"}]}
{"id": 194669, "image": "COCO_train2014_000000194669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rv is white\"."}], "task": "refering", "answer_template": "The \"the rv is white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 50, 51, 52, 53, 54, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89], "bbox": [0.183078125, 0.0, 0.92909375, 0.20938967136150236], "iscrowd": 0, "area": 32794.8382, "rle": {"size": [426, 640], "counts": "jg`11U=4M3L5K4E;O1N2O1N2O1N2N200O100O100O100O100O1O100O100O100O100O100O1O2N1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1000000000001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1J6K5J6J6K5J6J6K5J6J6K5J`Y1KffN5K5L4K5L4K5L4K5N20O10000000000000000000000000000000O100000O10000000000001O0000001O0000001O0000001O00001N1000001O0000001O0000001O0000001O0000001O0000001O000O101O00001O2N2N2N2N1O2N2N2N2N1O2N1O0000000000001O000O100000001O00000000000000001O00000000000000001O00000O10000000001O00000000000000001O0000000000000000001O000O1000000000001O00000000000000001O00000000000000001O0O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000I7F:F:F:F:E;F:F:F\\gb0"}, "label": "the rv is white"}]}
{"id": 194669, "image": "COCO_train2014_000000194669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white camper with blue stripes\"."}], "task": "refering", "answer_template": "The \"white camper with blue stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 50, 51, 52, 53, 54, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89], "bbox": [0.183078125, 0.0, 0.92909375, 0.20938967136150236], "iscrowd": 0, "area": 32794.8382, "rle": {"size": [426, 640], "counts": "jg`11U=4M3L5K4E;O1N2O1N2O1N2N200O100O100O100O100O1O100O100O100O100O100O1O2N1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1000000000001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1J6K5J6J6K5J6J6K5J6J6K5J`Y1KffN5K5L4K5L4K5L4K5N20O10000000000000000000000000000000O100000O10000000000001O0000001O0000001O0000001O00001N1000001O0000001O0000001O0000001O0000001O0000001O000O101O00001O2N2N2N2N1O2N2N2N2N1O2N1O0000000000001O000O100000001O00000000000000001O00000000000000001O00000O10000000001O00000000000000001O0000000000000000001O000O1000000000001O00000000000000001O00000000000000001O0O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000I7F:F:F:F:E;F:F:F\\gb0"}, "label": "white camper with blue stripes"}]}
{"id": 194669, "image": "COCO_train2014_000000194669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"water truck watering stuff\"."}], "task": "refering", "answer_template": "The \"water truck watering stuff\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 231, 232, 233, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 289, 290, 291, 292], "bbox": [0.07615625000000001, 0.11471830985915492, 0.8174531249999999, 0.8568544600938967], "iscrowd": 0, "area": 107994.53155, "rle": {"size": [426, 640], "counts": "ebd09n0<X:OUER1V:UOcET1U:V1J5K4M3M3M3N2M101N1]O`L\\G`3Z8m001N100O2O0O10001N10001O0O10001bNnJbJR5^5QKoI_5Q6eJ\\Ik5c6>000O100O10000O100O1N2M3N2N2N2N2N3K4M3L4O1000O1000000000000000O1000000000000000000000000000000O1000O1000000O10000E;N2N2_JQJe3Q6YLRJd3P6lKaJQ4a5oK_JQ4`5oKaJQ4_5oK`JR4`5nKVJkNMW5m5nKQJSONP5Q6mKQJUOLn4S6mKQJa4o5_KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJc4o5]KQJc4n5^KRJb4n5^KRJb4n5^KRJb4n5^KRJb4n5^KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5\\KRJd4n5\\KRJd4n5\\KRJd4m5]KSJc4m5]KSJc4m5]KSJc4m5]KSJd4l5\\KTJd4l5\\KTJd4l5\\KTJd4l5\\KTJd4l5\\KTJd4l5\\KTJUO2k4j5PLUJjN<U5b7O00001O1O1O2N1O1O1O001O1O1O001O1N100000000000000000O1O1O1O1000000000000O12N1O2N2N2N2N1O0O2O001O1O001O000000000000000000001O00000000O1000000O100000O10000J6F:H800000000000O10O100000000000000000000O01000000N2N2N2M3N2J6J5I8L4J600002N1O0000000000000001O0000000000000O10]KbGZ4^8dKdG]4[8`KhG`4W8_KkGa4_800000O10000F:01O0000000000000000001O000000000000001O5K6J6J5K3M001O0004L3M4M2M5K5K3NO0001O00010O0000010O00000001O0000001O00000O101O0O100O2O0O100100O2N2N2N010O0000010O00000010O0001O01O000000001O00000O2O000000001O0000001O000000001O0000001O00000O101O0000001O0000001O0O10000O2O000O2O1N1N3K4M4K4M4K4L5[Od001O1O001O001O001N101O001O1O001O001O001O001O001O1O001O5KY1gNW1`NjE^MZ;n0VEROa<000000000000000001O0O10000000000000002NS\\`1"}, "label": "water truck watering stuff"}]}
{"id": 194669, "image": "COCO_train2014_000000194669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red truck\"."}], "task": "refering", "answer_template": "The \"a red truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 59, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 231, 232, 233, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 289, 290, 291, 292], "bbox": [0.07615625000000001, 0.11471830985915492, 0.8174531249999999, 0.8568544600938967], "iscrowd": 0, "area": 107994.53155, "rle": {"size": [426, 640], "counts": "ebd09n0<X:OUER1V:UOcET1U:V1J5K4M3M3M3N2M101N1]O`L\\G`3Z8m001N100O2O0O10001N10001O0O10001bNnJbJR5^5QKoI_5Q6eJ\\Ik5c6>000O100O10000O100O1N2M3N2N2N2N2N3K4M3L4O1000O1000000000000000O1000000000000000000000000000000O1000O1000000O10000E;N2N2_JQJe3Q6YLRJd3P6lKaJQ4a5oK_JQ4`5oKaJQ4_5oK`JR4`5nKVJkNMW5m5nKQJSONP5Q6mKQJUOLn4S6mKQJa4o5_KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJb4P6^KPJc4o5]KQJc4n5^KRJb4n5^KRJb4n5^KRJb4n5^KRJb4n5^KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5]KQJc4o5\\KRJd4n5\\KRJd4n5\\KRJd4m5]KSJc4m5]KSJc4m5]KSJc4m5]KSJd4l5\\KTJd4l5\\KTJd4l5\\KTJd4l5\\KTJd4l5\\KTJd4l5\\KTJUO2k4j5PLUJjN<U5b7O00001O1O1O2N1O1O1O001O1O1O001O1N100000000000000000O1O1O1O1000000000000O12N1O2N2N2N2N1O0O2O001O1O001O000000000000000000001O00000000O1000000O100000O10000J6F:H800000000000O10O100000000000000000000O01000000N2N2N2M3N2J6J5I8L4J600002N1O0000000000000001O0000000000000O10]KbGZ4^8dKdG]4[8`KhG`4W8_KkGa4_800000O10000F:01O0000000000000000001O000000000000001O5K6J6J5K3M001O0004L3M4M2M5K5K3NO0001O00010O0000010O00000001O0000001O00000O101O0O100O2O0O100100O2N2N2N010O0000010O00000010O0001O01O000000001O00000O2O000000001O0000001O000000001O0000001O00000O101O0000001O0000001O0O10000O2O000O2O1N1N3K4M4K4M4K4L5[Od001O1O001O001O001N101O001O1O001O001O001O001O001O1O001O5KY1gNW1`NjE^MZ;n0VEROa<000000000000000001O0O10000000000000002NS\\`1"}, "label": "a red truck"}]}
{"id": 260206, "image": "COCO_train2014_000000260206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mid sized vase\"."}], "task": "refering", "answer_template": "The \"the mid sized vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 119, 120, 121, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 282, 303, 304, 305], "bbox": [0.12978125000000001, 0.251, 0.357296875, 0.8105416666666667], "iscrowd": 0, "area": 24493.29555, "rle": {"size": [480, 640], "counts": "aTW15e>8H9G9F9H9G8G:G8M4M3L3N3L3N3L3N3L4M2M4M2M4M2M4M3L3N3L3N3L3N3L4M2M4M2M4M3Lc0^Oh0WO9H0O100O100O1WOXIZIh6f6\\IUIe6j6`IRI`6m6eInH[6R7jIjHV6V7nIeHS6Z7j0O100O1O1O100O1O10000000000_OXHjIh7V6\\HfId7Z6_HcIa7]6cH_I\\7b6a00000000000000WH\\IV7d6iH]IW7c6hH^IX7b6gH`IX7`6gHaIY7a6dH`I\\7c6aH]I_7g6\\H[Ic7Q700000001N1000001O0000001BcHWI]7f6hHXIX7f6lHXIU7d6PIZIY7Z6lHdI]7P6hHnIY7n5o0M3L4L4L4M3L4L4L4M3L4L4L4M3L3M4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L3M4L6J9H7H8H8H8I7H8HilP6"}, "label": "the mid sized vase"}]}
{"id": 260206, "image": "COCO_train2014_000000260206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase of medium height , neither tall nor short\"."}], "task": "refering", "answer_template": "The \"the vase of medium height , neither tall nor short\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 119, 120, 121, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 282, 303, 304, 305], "bbox": [0.12978125000000001, 0.251, 0.357296875, 0.8105416666666667], "iscrowd": 0, "area": 24493.29555, "rle": {"size": [480, 640], "counts": "aTW15e>8H9G9F9H9G8G:G8M4M3L3N3L3N3L3N3L4M2M4M2M4M2M4M3L3N3L3N3L3N3L4M2M4M2M4M3Lc0^Oh0WO9H0O100O100O1WOXIZIh6f6\\IUIe6j6`IRI`6m6eInH[6R7jIjHV6V7nIeHS6Z7j0O100O1O1O100O1O10000000000_OXHjIh7V6\\HfId7Z6_HcIa7]6cH_I\\7b6a00000000000000WH\\IV7d6iH]IW7c6hH^IX7b6gH`IX7`6gHaIY7a6dH`I\\7c6aH]I_7g6\\H[Ic7Q700000001N1000001O0000001BcHWI]7f6hHXIX7f6lHXIU7d6PIZIY7Z6lHdI]7P6hHnIY7n5o0M3L4L4L4M3L4L4L4M3L4L4L4M3L3M4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L3M4L6J9H7H8H8H8I7H8HilP6"}, "label": "the vase of medium height , neither tall nor short"}]}
{"id": 260206, "image": "COCO_train2014_000000260206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall vase with trees painted on it\"."}], "task": "refering", "answer_template": "The \"a tall vase with trees painted on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 40, 41, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 316, 317], "bbox": [0.694375, 0.0513125, 0.87978125, 0.801875], "iscrowd": 0, "area": 28270.02675, "rle": {"size": [480, 640], "counts": "gW`61m><E;D<D;E<D<E;D;E<D<D;F;E;E;E:F:F9G9G9G9G9G9G9G9H8G9G9G9G9G2N1O1O1O1O1O1bN`F[Ma9e2`FYMa9g2`FXMa9g2`FWMa9h2aFVM`9j2bFSM_9m2bFQM_9o2bFoL_9Q3bFmL_9S3W1000001N100000000000000O101O0000000000000O10001O00000000\\EoLY9Q3]FZMb9f2]F\\Mb9d2]F]Mc9c2\\F_Mc9a2]F`Mc9_2\\FbMd9^2[FdMd9\\2[FfMd9[2ZFfMe9[2[FfMV9Z4B=C>C=E:F;H8I6K6J6J5K6J6J5K6J6J5K6J6I7J5I8I8G;E;F:E;F:E;F:E;F:E;F:EonS1"}, "label": "a tall vase with trees painted on it"}]}
{"id": 260206, "image": "COCO_train2014_000000260206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest vase in arrangement with skinny neck\"."}], "task": "refering", "answer_template": "The \"tallest vase in arrangement with skinny neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 40, 41, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 316, 317], "bbox": [0.694375, 0.0513125, 0.87978125, 0.801875], "iscrowd": 0, "area": 28270.02675, "rle": {"size": [480, 640], "counts": "gW`61m><E;D<D;E<D<E;D;E<D<D;F;E;E;E:F:F9G9G9G9G9G9G9G9H8G9G9G9G9G2N1O1O1O1O1O1bN`F[Ma9e2`FYMa9g2`FXMa9g2`FWMa9h2aFVM`9j2bFSM_9m2bFQM_9o2bFoL_9Q3bFmL_9S3W1000001N100000000000000O101O0000000000000O10001O00000000\\EoLY9Q3]FZMb9f2]F\\Mb9d2]F]Mc9c2\\F_Mc9a2]F`Mc9_2\\FbMd9^2[FdMd9\\2[FfMd9[2ZFfMe9[2[FfMV9Z4B=C>C=E:F;H8I6K6J6J5K6J6J5K6J6J5K6J6I7J5I8I8G;E;F:E;F:E;F:E;F:E;F:EonS1"}, "label": "tallest vase in arrangement with skinny neck"}]}
{"id": 260206, "image": "COCO_train2014_000000260206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small vase in the middle of the other 2 vases\"."}], "task": "refering", "answer_template": "The \"the small vase in the middle of the other 2 vases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335], "bbox": [0.399546875, 0.4794375, 0.627140625, 0.8623958333333334], "iscrowd": 0, "area": 23944.8356, "rle": {"size": [480, 640], "counts": "YXh3Z1b=l0UO<C=D<C=F:F4M2M4M2N2M4M2N2M4M1N2O1O0O2O1O1N2O0O2O1O1N2O001N2O1O1O1O0O100000000000000O100000000000000O100000000000O10O100000000000000O100000000000000O1000000001O00000000000000001O000000000000001O000000000O1000001O0000001O00001O00000O2N1O1O2N1O1O2N1O1O1O2N2N2M3N2N2L4K5L4K5K4M4K5K6K8G9G:aNli_3"}, "label": "the small vase in the middle of the other 2 vases"}]}
{"id": 260206, "image": "COCO_train2014_000000260206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small urn with red , green and brown flowers on it with a blue background\"."}], "task": "refering", "answer_template": "The \"a small urn with red , green and brown flowers on it with a blue background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335], "bbox": [0.399546875, 0.4794375, 0.627140625, 0.8623958333333334], "iscrowd": 0, "area": 23944.8356, "rle": {"size": [480, 640], "counts": "YXh3Z1b=l0UO<C=D<C=F:F4M2M4M2N2M4M2N2M4M1N2O1O0O2O1O1N2O0O2O1O1N2O001N2O1O1O1O0O100000000000000O100000000000000O100000000000O10O100000000000000O100000000000000O1000000001O00000000000000001O000000000000001O000000000O1000001O0000001O00001O00000O2N1O1O2N1O1O2N1O1O1O2N2N2M3N2N2L4K5L4K5K4M4K5K6K8G9G:aNli_3"}, "label": "a small urn with red , green and brown flowers on it with a blue background"}]}
{"id": 366702, "image": "COCO_train2014_000000366702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white striped chair in a living room that is half on top of the rug\"."}], "task": "refering", "answer_template": "The \"a black and white striped chair in a living room that is half on top of the rug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 289, 290, 293, 312, 313], "bbox": [0.5208437499999999, 0.535480093676815, 0.771921875, 0.922576112412178], "iscrowd": 0, "area": 13654.495599999997, "rle": {"size": [427, 640], "counts": "^S[45V=\\1dNo0QO1O001O00001O04L1E:E<I6010O1000O0100O010O1O1O001O1O1O0O2O1O1O001O1O1O001N2O1O001O1O001N2M3M2N3M3L4fFaMd7a2RHiMl7Z2hGRNU8Q2`GYN^8[3N3O001O001OO001O1O1O1O1G9[Nd110O10000000O0100000000O010000000O1000O10O10000000O0100000000O10O100000O1000O10O100000000O010000000O10O100000O100000O01001O001O1N101O1O1O001N2O001O[1eN000001O2N2N1O2N1H9A>B=C=C>C<E<D;D<Endl1"}, "label": "a black and white striped chair in a living room that is half on top of the rug"}]}
{"id": 366702, "image": "COCO_train2014_000000366702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white striped chair near the television that is half on area rug and half on the hard wood floor\"."}], "task": "refering", "answer_template": "The \"a black and white striped chair near the television that is half on area rug and half on the hard wood floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 289, 290, 293, 312, 313], "bbox": [0.5208437499999999, 0.535480093676815, 0.771921875, 0.922576112412178], "iscrowd": 0, "area": 13654.495599999997, "rle": {"size": [427, 640], "counts": "^S[45V=\\1dNo0QO1O001O00001O04L1E:E<I6010O1000O0100O010O1O1O001O1O1O0O2O1O1O001O1O1O001N2O1O001O1O001N2M3M2N3M3L4fFaMd7a2RHiMl7Z2hGRNU8Q2`GYN^8[3N3O001O001OO001O1O1O1O1G9[Nd110O10000000O0100000000O010000000O1000O10O10000000O0100000000O10O100000O1000O10O100000000O010000000O10O100000O100000O01001O001O1N101O1O1O001N2O001O[1eN000001O2N2N1O2N1H9A>B=C=C>C<E<D;D<Endl1"}, "label": "a black and white striped chair near the television that is half on area rug and half on the hard wood floor"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on left\"."}], "task": "refering", "answer_template": "The \"man on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 234, 235, 237, 238, 239, 240, 241, 242, 253, 254, 255, 257, 258, 260, 261, 262, 263, 264, 276, 283, 284], "bbox": [0.0033750000000000004, 0.09120833333333334, 0.53209375, 0.753375], "iscrowd": 0, "area": 47681.865, "rle": {"size": [480, 640], "counts": "lW13k><D<E;D101N101N1O2O0O2O0014K4L4M4K4L4M4K3N2M3M2O2M3N2MZNOf14L5K4L4L4L5K4L4L4L5K4M3L4L5K4L4L4L5K4L3M2N3M2N2N3M2O1N3MbNYDGe;:`DB^;?gD^OV;c0PEXOn:i0WESOg:n0^EoN_:R1gEiNW:X1nEdNP:]1UF`Nh9b1]FYNa9i1cFTNZ9n1jFnMT9S2l11N2N2N2N2N2N2N2N1O2N2N2N2O1N2O1N1O2O1N2O1N2N1fMgLnHZ3R7RM`HP3_7]MSHe2l7hMgGY2X8TNZGn1e8_NnFa1R9W2O1O100O1O005K8H9G8H9H7H8H9G2N1O1O2N100O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O100O2O0000^NnIVJR6i5PJWJo5h5TJVJm5h5UJWJk5h5XJVJh5i5ZJWJe5h5^JVJb5i5`JVJ`5h5dJUJ]5j5eJRJ_5k5gJnI\\5P6ZKZIh4d6k1N2M3N20001O000000000000001O00000001O01O1O010O010O00100O010O010O00100O010O010O101N1O101N100OaGfJZ7Q7IIY1hNK5L3L5L4K4L5L0O0001N101O000O2O000O2O001O0O101QNYHcLg7R3iHhLX7X3mHcLT7j2mG^LU1c0n6l2RH^LW1?i6o2TH`LY1<c6Q3YHaLY19^6U3\\H_L\\17Y6Y3\\H^La13T6_3hJ\\LX5c3T3O0O2O00001O00001O001O2N1O1O1O2N1003M2O1N0000010O0000001O1O2N2N2N1O2M3N2N2N?A2N2N1O2M2O1O2N1O1ON2N1O2O1N2N2O1N2N2O0O2N2O2M3M3N2M3M3N2M3M3N1N3M3M3M3M3M3M3M3M3L3N3M3M3M3M3M3M3N2M3M3N1N3M4M4K4L4M3L5K4M3L4M4K4L4Mkk[4"}, "label": "man on left"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a checked jacket\"."}], "task": "refering", "answer_template": "The \"a man in a checked jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 234, 235, 237, 238, 239, 240, 241, 242, 253, 254, 255, 257, 258, 260, 261, 262, 263, 264, 276, 283, 284], "bbox": [0.0033750000000000004, 0.09120833333333334, 0.53209375, 0.753375], "iscrowd": 0, "area": 47681.865, "rle": {"size": [480, 640], "counts": "lW13k><D<E;D101N101N1O2O0O2O0014K4L4M4K4L4M4K3N2M3M2O2M3N2MZNOf14L5K4L4L4L5K4L4L4L5K4M3L4L5K4L4L4L5K4L3M2N3M2N2N3M2O1N3MbNYDGe;:`DB^;?gD^OV;c0PEXOn:i0WESOg:n0^EoN_:R1gEiNW:X1nEdNP:]1UF`Nh9b1]FYNa9i1cFTNZ9n1jFnMT9S2l11N2N2N2N2N2N2N2N1O2N2N2N2O1N2O1N1O2O1N2O1N2N1fMgLnHZ3R7RM`HP3_7]MSHe2l7hMgGY2X8TNZGn1e8_NnFa1R9W2O1O100O1O005K8H9G8H9H7H8H9G2N1O1O2N100O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O100O2O0000^NnIVJR6i5PJWJo5h5TJVJm5h5UJWJk5h5XJVJh5i5ZJWJe5h5^JVJb5i5`JVJ`5h5dJUJ]5j5eJRJ_5k5gJnI\\5P6ZKZIh4d6k1N2M3N20001O000000000000001O00000001O01O1O010O010O00100O010O010O00100O010O010O101N1O101N100OaGfJZ7Q7IIY1hNK5L3L5L4K4L5L0O0001N101O000O2O000O2O001O0O101QNYHcLg7R3iHhLX7X3mHcLT7j2mG^LU1c0n6l2RH^LW1?i6o2TH`LY1<c6Q3YHaLY19^6U3\\H_L\\17Y6Y3\\H^La13T6_3hJ\\LX5c3T3O0O2O00001O00001O001O2N1O1O1O2N1003M2O1N0000010O0000001O1O2N2N2N1O2M3N2N2N?A2N2N1O2M2O1O2N1O1ON2N1O2O1N2N2O1N2N2O0O2N2O2M3M3N2M3M3N2M3M3N1N3M3M3M3M3M3M3M3M3L3N3M3M3M3M3M3M3N2M3M3N1N3M4M4K4L4M3L5K4M3L4M4K4L4Mkk[4"}, "label": "a man in a checked jacket"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife that a woman is holding\"."}], "task": "refering", "answer_template": "The \"the knife that a woman is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290], "bbox": [0.42978125, 0.3236041666666667, 0.67415625, 0.7528125], "iscrowd": 0, "area": 19631.094000000012, "rle": {"size": [480, 640], "counts": "gWQ49g>0O1000000O1000000O1jJEgK;Y4FfK:Y4HfK8Y4IfK8Z4KcK5\\43]KMc4:UKGj4<TKDk4?SKBl4`0RK@m4c0PK^OP5d0nJ\\OQ5g0mJYOR5j0lJVOT5l0iJUOV5n0hJROX5P1fJPOY5R1fJnNY5U1dJlN\\5U1cJkN\\5W1cJiN\\5Z1bJgN]5Z1aJgN^5[1aJeN_5\\1`JdN_5_1_JaNa5`1]JaNc5`1\\J`Nd5b1ZJ^Nf5c1YJ]Ng5d1WJ]Ni5e1TJ\\Nk5f1PJ^NP6c1kIaNU6`1fIeNY6]1aIgN_6[1[IiNe6X1VIlNj6V1SIkNm6V1RIjNm6Y1PIhNP7Y1oHgNQ7[1mHeNS7]1jHdNV7]1iHcNW7_1gHaNY7`1eHaNZ7a1eH_N[7c1cH^N\\7c1bH^N^7d1`H\\N`7e1_H[Na7g1\\HZNd7g1[HYNd7j1UH[Nl7e1mGaNS8a1dGfN\\8[1\\GlNd8h31O00010O0000001O00001O0000001O01O01O00001O0000001O00001O0001O01O00001O0000001O00001O01O0001^JoFS5Q9gJWGW5i8cJ`GZ5U9N2N3M2N3M2N3M2O2M2N3M2N3M2N3M2N3M2N3[OWEmLk:Q3e0L3M4L3M4L3N2M4L3M4L3M4M2M4L3M4L3M8HRfQ3"}, "label": "the knife that a woman is holding"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"in hand there is a knife\"."}], "task": "refering", "answer_template": "The \"in hand there is a knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290], "bbox": [0.42978125, 0.3236041666666667, 0.67415625, 0.7528125], "iscrowd": 0, "area": 19631.094000000012, "rle": {"size": [480, 640], "counts": "gWQ49g>0O1000000O1000000O1jJEgK;Y4FfK:Y4HfK8Y4IfK8Z4KcK5\\43]KMc4:UKGj4<TKDk4?SKBl4`0RK@m4c0PK^OP5d0nJ\\OQ5g0mJYOR5j0lJVOT5l0iJUOV5n0hJROX5P1fJPOY5R1fJnNY5U1dJlN\\5U1cJkN\\5W1cJiN\\5Z1bJgN]5Z1aJgN^5[1aJeN_5\\1`JdN_5_1_JaNa5`1]JaNc5`1\\J`Nd5b1ZJ^Nf5c1YJ]Ng5d1WJ]Ni5e1TJ\\Nk5f1PJ^NP6c1kIaNU6`1fIeNY6]1aIgN_6[1[IiNe6X1VIlNj6V1SIkNm6V1RIjNm6Y1PIhNP7Y1oHgNQ7[1mHeNS7]1jHdNV7]1iHcNW7_1gHaNY7`1eHaNZ7a1eH_N[7c1cH^N\\7c1bH^N^7d1`H\\N`7e1_H[Na7g1\\HZNd7g1[HYNd7j1UH[Nl7e1mGaNS8a1dGfN\\8[1\\GlNd8h31O00010O0000001O00001O0000001O01O01O00001O0000001O00001O0001O01O00001O0000001O00001O01O0001^JoFS5Q9gJWGW5i8cJ`GZ5U9N2N3M2N3M2N3M2O2M2N3M2N3M2N3M2N3M2N3[OWEmLk:Q3e0L3M4L3M4L3N2M4L3M4L3M4M2M4L3M4L3M8HRfQ3"}, "label": "in hand there is a knife"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza in front of a woman with a gray sweatshirt\"."}], "task": "refering", "answer_template": "The \"a pizza in front of a woman with a gray sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5123593750000001, 0.7348333333333333, 0.9674218750000001, 0.9865208333333334], "iscrowd": 0, "area": 27478.369250000003, "rle": {"size": [480, 640], "counts": "^Uj4`0^>5J5L5K4L4K6K4L4M4K2O1N3N1O1N2O1N2O1N2O1O2O0O100O100O100O100O10000O100O100O10O0100O100O1000000O100000000O100000000O010000000O10000000000O100000000O1000O1000O10000O100O100O100O100O2O0O100O10000O100O101N100O100O100O10000O2O0O100O100O100O100O2O0O10000O100O100O101O00000001O0000000001O01O0000000000001O01O0000000001N1000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000001O0000001O0000001O00000O2O0000001O0000001O0000001O0000001O0000001O0000001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O1O1O1N2O1O1O1O1O1N2N1O2M3N2N2M3N5J5L4L4K5L4L7H:G9FX]9"}, "label": "a pizza in front of a woman with a gray sweatshirt"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza\"."}], "task": "refering", "answer_template": "The \"pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5123593750000001, 0.7348333333333333, 0.9674218750000001, 0.9865208333333334], "iscrowd": 0, "area": 27478.369250000003, "rle": {"size": [480, 640], "counts": "^Uj4`0^>5J5L5K4L4K6K4L4M4K2O1N3N1O1N2O1N2O1N2O1O2O0O100O100O100O100O10000O100O100O10O0100O100O1000000O100000000O100000000O010000000O10000000000O100000000O1000O1000O10000O100O100O100O100O2O0O100O10000O100O101N100O100O100O10000O2O0O100O100O100O100O2O0O10000O100O100O101O00000001O0000000001O01O0000000000001O01O0000000001N1000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000001O0000001O0000001O00000O2O0000001O0000001O0000001O0000001O0000001O0000001O001O001O00001O001O001O001O001O001O00001O001O001O001O001O1O1O1N2O1O1O1O1O1N2N1O2M3N2N2M3N5J5L4L4K5L4L7H:G9FX]9"}, "label": "pizza"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a knife and a fork in her hands at the table\"."}], "task": "refering", "answer_template": "The \"a woman holding a knife and a fork in her hands at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 60, 61, 62, 63, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321], "bbox": [0.496625, 0.10360416666666666, 0.9983124999999999, 0.7815416666666668], "iscrowd": 0, "area": 68985.2213, "rle": {"size": [480, 640], "counts": "o[e44j>7I7I7I7I7I6J201O000000000000000000001O0001O001O00001O001O0001M2N2N2O1N2N2N2N2I7F:G9I7K5L4L4K5L4L4K5L4L4K5L4L4L4K5L4L4K4M4L3@a0oMP201N101O0O2O0O10O10O1000O0100O10O10O10000O010O100O10O0100O100OO2L4L5L4K5K5L4L3M4L4L4L45K8H8H8H8H8H8H8H9G8H8H10O0000001O00001O00001O0000001O00001O0000001O00001O001O00001O010O001O00001O001O1O2N1O2N2N1O2N101N1O2N2N1O2N1O2N101N1O2N1O2N1O2N1O100O1O1O1O1O1O1O1O010O001O001O001O01O01O001O001O000001O0000000001O000001O000000000000000010O0000000000001O100O1O1O1O001O1O100O1O1O1O1O1O001O100O001O1O1O001O1O010O1O1O001O1O1O00100O1O001O1O1O001O100O001O1O001O1O10O01O1O1O001O1O1O10O01O1O1O1O1O001O100O1O1O001O1O1O1O10O01O1O1O1OS9"}, "label": "a woman holding a knife and a fork in her hands at the table"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person holding a knife and fork\"."}], "task": "refering", "answer_template": "The \"a person holding a knife and fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 60, 61, 62, 63, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321], "bbox": [0.496625, 0.10360416666666666, 0.9983124999999999, 0.7815416666666668], "iscrowd": 0, "area": 68985.2213, "rle": {"size": [480, 640], "counts": "o[e44j>7I7I7I7I7I6J201O000000000000000000001O0001O001O00001O001O0001M2N2N2O1N2N2N2N2I7F:G9I7K5L4L4K5L4L4K5L4L4K5L4L4L4K5L4L4K4M4L3@a0oMP201N101O0O2O0O10O10O1000O0100O10O10O10000O010O100O10O0100O100OO2L4L5L4K5K5L4L3M4L4L4L45K8H8H8H8H8H8H8H9G8H8H10O0000001O00001O00001O0000001O00001O0000001O00001O001O00001O010O001O00001O001O1O2N1O2N2N1O2N101N1O2N2N1O2N1O2N101N1O2N1O2N1O2N1O100O1O1O1O1O1O1O1O010O001O001O001O01O01O001O001O000001O0000000001O000001O000000000000000010O0000000000001O100O1O1O1O001O1O100O1O1O1O1O1O001O100O001O1O1O001O1O010O1O1O001O1O1O00100O1O001O1O1O001O100O001O1O001O1O10O01O1O1O001O1O1O10O01O1O1O1O1O001O100O1O1O001O1O1O1O10O01O1O1O1OS9"}, "label": "a person holding a knife and fork"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of salad dressing with a pour - spout sitting on the table next to a pepper grinder\"."}], "task": "refering", "answer_template": "The \"a bottle of salad dressing with a pour - spout sitting on the table next to a pepper grinder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 238, 239, 261, 262, 284, 285, 307, 308, 309, 331, 332, 354, 355, 377, 378], "bbox": [0.353140625, 0.5417083333333333, 0.470859375, 0.9744375000000001], "iscrowd": 0, "area": 5686.276099999999, "rle": {"size": [480, 640], "counts": "UWZ33l>4K6nNFTC`0i<KiB;T=m0L4jDZNi8j1nFZNS9j1eFYN\\9k1\\FYNd9j1UFYNl9k1kEYNV:k1bEXN^:m1ZEWNf:m1REVNo:m2100O1O100O1O2N100O2N2O1N3M2N5L:E;F:E;F:E3N2N1N2O1N2O2d0ZO9H8G9G7IN3N1N2O1N3N1N2O1N3N1N2O1N3N1N2O2M2O1O1N3N1N2L4J7J5J6J6G_`n4"}, "label": "a bottle of salad dressing with a pour - spout sitting on the table next to a pepper grinder"}]}
{"id": 243824, "image": "COCO_train2014_000000243824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottle of oil\"."}], "task": "refering", "answer_template": "The \"the bottle of oil\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 238, 239, 261, 262, 284, 285, 307, 308, 309, 331, 332, 354, 355, 377, 378], "bbox": [0.353140625, 0.5417083333333333, 0.470859375, 0.9744375000000001], "iscrowd": 0, "area": 5686.276099999999, "rle": {"size": [480, 640], "counts": "UWZ33l>4K6nNFTC`0i<KiB;T=m0L4jDZNi8j1nFZNS9j1eFYN\\9k1\\FYNd9j1UFYNl9k1kEYNV:k1bEXN^:m1ZEWNf:m1REVNo:m2100O1O100O1O2N100O2N2O1N3M2N5L:E;F:E;F:E3N2N1N2O1N2O2d0ZO9H8G9G7IN3N1N2O1N3N1N2O1N3N1N2O1N3N1N2O2M2O1O1N3N1N2L4J7J5J6J6G_`n4"}, "label": "the bottle of oil"}]}
{"id": 342130, "image": "COCO_train2014_000000342130.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color chair in front of laptop and monitor\"."}], "task": "refering", "answer_template": "The \"a black color chair in front of laptop and monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 233], "bbox": [0.70976, 0.7630933333333334, 1.0, 0.9884000000000001], "iscrowd": 0, "area": 8038.37635, "rle": {"size": [375, 500], "counts": "a[R47X;8N2O1N2O1N2O1N2N2O1N2O1O1000000000O10000000000000000000000000000000000000000000000000000000000000^NZOSHf0l7[OTHe0l7[OTHe0l7\\OSHd0m7\\OSHd0m7]ORHc0m7^OSHb0m7^OSHb0m7`0QG@o8d1001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00000RNYFl1l9M3M4L3F:UOcF"}, "label": "a black color chair in front of laptop and monitor"}]}
{"id": 194679, "image": "COCO_train2014_000000194679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"standing giraffe in the background\"."}], "task": "refering", "answer_template": "The \"standing giraffe in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 61, 62, 63, 77, 78, 92, 93, 94, 108, 109, 110, 111, 123, 124, 125, 126, 138, 139, 140, 141, 153, 168, 170, 171, 183, 185, 186], "bbox": [0.11981264637002341, 0.12765625, 0.4310070257611241, 0.55140625], "iscrowd": 0, "area": 10981.3553, "rle": {"size": [640, 427], "counts": "eoo06gc03N3N1O1O1N2O1C_OZ]Ob0gb0_OU]Oc0kb0@Q]O`0Pc080000M3M3M2O20000O10kN_]Oj0jb02N2M3O3N4K69FL2N0_]OPOQb0P1o]O[Oga0d0Y^OG\\1SO\\>V1X@2P1nNd>o0]@8h0nNh>K^@b01i0d0oNj>Df@`0LQ1`0oNh2[Oc84]D>IW1:ROd2ZOl8MbD<D`14ZOZ2RO\\9FfD;@g1O_OR2oNd:H[Cm1K_Oo1POo:@WCl3b1iLY;XOUCR4:ZL1b0i<V4oBoKP=V4gBoKY=Y56E:E<E^BgJo<Q5nBXKU=a4fBgK]=T4]BULf=d3UBeLm=f43L3N3M2N3L2D<YOg0E:I8H8^Ob0]Oc0L4O1O1O1O1O1O1O1O2N1O2N2N1O1O1O1O1O1O1O1O1O1O0O10001Oc]O_N[b0a1e]O`NZb0`1f]ObNXb0]1i]OdNVb0\\1j]OfNTb0Z1k]OhNTb0X1l]OjNRb0V1n]OkNQb0T1P^OnNna0R1Q^OPOna0R1R^OnNla0e1N1O2NZOY^OTOfa0l0Z^OVO[2YOl<a1i@WOX2ZOo<_1i@WOV2\\OQ=]1i@VOl1H[=R1i@UOb14e=^1PBmNR>:g@VOn0j0\\>Of@VOo0k0]>Md@WOo0m0^>Kc@WOP1n0^>Jb@WOQ1n0_>J`@WOQ1n0b>J]@WOR1n0c>J[@WOS1I[O>Y?b0Y@UOS12d?i0Y_OSOV19\\?b0__OSOW1a0T?<e_OQOY1i0UOQOV?U1\\@oN[1P1eNTOe?m0[@mN]1\\1Y>F[@lN_1\\1W>BUB;k`0M4K4M8Hgag4"}, "label": "standing giraffe in the background"}]}
{"id": 194679, "image": "COCO_train2014_000000194679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing up behind the tree\"."}], "task": "refering", "answer_template": "The \"a giraffe standing up behind the tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 61, 62, 63, 77, 78, 92, 93, 94, 108, 109, 110, 111, 123, 124, 125, 126, 138, 139, 140, 141, 153, 168, 170, 171, 183, 185, 186], "bbox": [0.11981264637002341, 0.12765625, 0.4310070257611241, 0.55140625], "iscrowd": 0, "area": 10981.3553, "rle": {"size": [640, 427], "counts": "eoo06gc03N3N1O1O1N2O1C_OZ]Ob0gb0_OU]Oc0kb0@Q]O`0Pc080000M3M3M2O20000O10kN_]Oj0jb02N2M3O3N4K69FL2N0_]OPOQb0P1o]O[Oga0d0Y^OG\\1SO\\>V1X@2P1nNd>o0]@8h0nNh>K^@b01i0d0oNj>Df@`0LQ1`0oNh2[Oc84]D>IW1:ROd2ZOl8MbD<D`14ZOZ2RO\\9FfD;@g1O_OR2oNd:H[Cm1K_Oo1POo:@WCl3b1iLY;XOUCR4:ZL1b0i<V4oBoKP=V4gBoKY=Y56E:E<E^BgJo<Q5nBXKU=a4fBgK]=T4]BULf=d3UBeLm=f43L3N3M2N3L2D<YOg0E:I8H8^Ob0]Oc0L4O1O1O1O1O1O1O1O2N1O2N2N1O1O1O1O1O1O1O1O1O1O0O10001Oc]O_N[b0a1e]O`NZb0`1f]ObNXb0]1i]OdNVb0\\1j]OfNTb0Z1k]OhNTb0X1l]OjNRb0V1n]OkNQb0T1P^OnNna0R1Q^OPOna0R1R^OnNla0e1N1O2NZOY^OTOfa0l0Z^OVO[2YOl<a1i@WOX2ZOo<_1i@WOV2\\OQ=]1i@VOl1H[=R1i@UOb14e=^1PBmNR>:g@VOn0j0\\>Of@VOo0k0]>Md@WOo0m0^>Kc@WOP1n0^>Jb@WOQ1n0_>J`@WOQ1n0b>J]@WOR1n0c>J[@WOS1I[O>Y?b0Y@UOS12d?i0Y_OSOV19\\?b0__OSOW1a0T?<e_OQOY1i0UOQOV?U1\\@oN[1P1eNTOe?m0[@mN]1\\1Y>F[@lN_1\\1W>BUB;k`0M4K4M8Hgag4"}, "label": "a giraffe standing up behind the tree"}]}
{"id": 194679, "image": "COCO_train2014_000000194679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe in the middle looking to the left\"."}], "task": "refering", "answer_template": "The \"the giraffe in the middle looking to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 140, 141, 142, 143, 155, 156, 157, 158, 172, 173, 187, 188, 202, 203, 217, 218, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 288, 289, 290, 291], "bbox": [0.026955503512880563, 0.3479375, 0.6466744730679157, 0.8423281250000001], "iscrowd": 0, "area": 23603.443300000003, "rle": {"size": [640, 427], "counts": "UP81oc0000O10000O10000O10000O10000O1\\\\OK]c05^\\O0bc07O10000O010000000000000000000000000000000000000000O100000001O0001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O000010O000O10000O1000000O10000O10N2N1O2N2N1O2M3N1O2N2N2N1O2M3N1O2N2N1O2N2M3N2N2N2N2N2N2M3N3M2N2N2N2^K]MfGd2T8bMlG^2R8dMmG]2R8dMnG\\2P8fMPHZ2o7hMPHX2n7jMQHW2n7jMRHV2l7lMTHT2k7mMTHT2j7oMUHQ2j7PNVHP2h7RNXHn1g7SNXHn1f7TNZHl1e7VNZHj1d7XN\\Hh1c7YN\\Hh1b7ZN^Hf1a7[N_He1_7]N`Hd1_7]NaHc1]7_NcHa1\\7_NeHa1Y7aNfH`1Y7`NhH`1V7bNjH^1U7cNjH^1T7cNmH]1Q7eNoH[1P7eNQI[1m6gNRIZ1m6fNTIZ1j6hNVIX1i6iNWIW1f6kNZIV1d6lN[IU1b6mN]IU1a6mN]IU1a6mN^IT1_6nN`IT1^6nN`IT1`6kN`IV1`6jN^IX1b6hN\\IZ1d6eN\\I\\1e6cNYI_1g6`NXIb1h6^NVId1j6\\NUIe1l6YNSIi1m6WNQIk1o6TNQIl1P7TNnHn1S7QNkHQ2U7nMkHS2V7lMhHV2Y7hMfHZ2[7eMcH\\2_7]OfFd0[9P5001O1N2O1O1O1O001N2O1O1O1O1O2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N3cISD`4P<PKhDf4Z;kJ^Ej4V=F;Dd0]O4L5K4K5L5K4L4K6K4hMP_OX1Ta0_NT_O^1ka0K4M4L4L3L5L3M2N3L4M2N3M3L3N3M3Mokm2"}, "label": "the giraffe in the middle looking to the left"}]}
{"id": 252025, "image": "COCO_train2014_000000252025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe running in front of the other two\"."}], "task": "refering", "answer_template": "The \"the giraffe running in front of the other two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 189, 190, 191, 192, 193, 207, 208, 209, 224, 225, 226, 241, 242, 243, 258, 259, 260, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 325, 326, 327, 328, 329, 342, 343, 344, 345, 346, 359, 360, 361, 362, 376, 377, 378, 379], "bbox": [0.13864583333333333, 0.44121875, 0.41375000000000006, 1.0], "iscrowd": 0, "area": 27882.9912, "rle": {"size": [640, 480], "counts": "cUZ11Pd01N1O2O0O1XCn0`4TOSJ`2e5`MTIo3c6SLZIW4]6jK`Ia4W6aKfIi4R6WKkIT5l5mJRJW5k5kJQJZ5m5fJQJ]5m5dJPJ^5P6cJnI^5S6aJkIa5U6_JjIb5V6_JhIb5X6^JfId5[6[JdIf5\\6ZJbIh5^6YJ`Ih5a6WJ]Ik5c6UJ\\Il5d6TJZIn5g6RJWIo5i6QJUIQ6k6oITIR6m6mIQIU6o6lIoHU6Q7kImHW6S7iIlHX6U7gIiH[6W7iIdHX6\\7lI_HU6b7oIWHS6i7QJRHP6n7UJkGm5V8VJeGk5[8ZJ^Gh5b8\\JYGe5b7eKWH]4i7hKQHY4o7kKkGW4U8nKeGS4[8QL_GQ4a8TLYGm3h8e3000000001lMS20003M6J6J7I2NM4Md0\\Od0[Od0]O?A2N2M3N2N21O00001O00001M2N2ZHTGU4P9eKWGV4l8]IPGY1=T5d9gJcFT5`9gJhFS5[9gJmFU5T9gJSGT5P9fJYGT5j8gJ]GT5f8gJbGS5a8kJcGP5_8PKeGk4]8UKeGf4^8YKeGb4^8^KeG\\4^8cKeGX4^8hKeGR4]8nKfGn3\\8RLeGl3\\8TLeGk3[8TLfGl3Z8TLgGk3X8ULjGj3W8ULiGk3W8TLkGj3W8ULiGk3W8TLkGk3V8TLkGk3U8TLlGS3^LWMh;EkGQ3aLWMe;GkGP3cLWMd;HjGm2fLYMa;JjGj2iLYM_;KiGj2kLYM];MiGf2oLZMZ;NhGf2QMYMY;1gGc2TMYMW;2gGa2]9_McF_2`9_MbF^2`9bM`F[2e=M3M2M4M2N3L4M<D=B?B=C[T_5"}, "label": "the giraffe running in front of the other two"}]}
{"id": 252025, "image": "COCO_train2014_000000252025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is on the bottom of the picture\"."}], "task": "refering", "answer_template": "The \"a giraffe that is on the bottom of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 189, 190, 191, 192, 193, 207, 208, 209, 224, 225, 226, 241, 242, 243, 258, 259, 260, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 325, 326, 327, 328, 329, 342, 343, 344, 345, 346, 359, 360, 361, 362, 376, 377, 378, 379], "bbox": [0.13864583333333333, 0.44121875, 0.41375000000000006, 1.0], "iscrowd": 0, "area": 27882.9912, "rle": {"size": [640, 480], "counts": "cUZ11Pd01N1O2O0O1XCn0`4TOSJ`2e5`MTIo3c6SLZIW4]6jK`Ia4W6aKfIi4R6WKkIT5l5mJRJW5k5kJQJZ5m5fJQJ]5m5dJPJ^5P6cJnI^5S6aJkIa5U6_JjIb5V6_JhIb5X6^JfId5[6[JdIf5\\6ZJbIh5^6YJ`Ih5a6WJ]Ik5c6UJ\\Il5d6TJZIn5g6RJWIo5i6QJUIQ6k6oITIR6m6mIQIU6o6lIoHU6Q7kImHW6S7iIlHX6U7gIiH[6W7iIdHX6\\7lI_HU6b7oIWHS6i7QJRHP6n7UJkGm5V8VJeGk5[8ZJ^Gh5b8\\JYGe5b7eKWH]4i7hKQHY4o7kKkGW4U8nKeGS4[8QL_GQ4a8TLYGm3h8e3000000001lMS20003M6J6J7I2NM4Md0\\Od0[Od0]O?A2N2M3N2N21O00001O00001M2N2ZHTGU4P9eKWGV4l8]IPGY1=T5d9gJcFT5`9gJhFS5[9gJmFU5T9gJSGT5P9fJYGT5j8gJ]GT5f8gJbGS5a8kJcGP5_8PKeGk4]8UKeGf4^8YKeGb4^8^KeG\\4^8cKeGX4^8hKeGR4]8nKfGn3\\8RLeGl3\\8TLeGk3[8TLfGl3Z8TLgGk3X8ULjGj3W8ULiGk3W8TLkGj3W8ULiGk3W8TLkGk3V8TLkGk3U8TLlGS3^LWMh;EkGQ3aLWMe;GkGP3cLWMd;HjGm2fLYMa;JjGj2iLYM_;KiGj2kLYM];MiGf2oLZMZ;NhGf2QMYMY;1gGc2TMYMW;2gGa2]9_McF_2`9_MbF^2`9bM`F[2e=M3M2M4M2N3L4M<D=B?B=C[T_5"}, "label": "a giraffe that is on the bottom of the picture"}]}
{"id": 252025, "image": "COCO_train2014_000000252025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe standing in the middle of two other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe standing in the middle of two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 125, 126, 142, 143, 144, 160, 161, 162, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 228, 229, 230, 231, 245, 246, 247, 262, 263, 264, 279, 280, 281, 296, 297, 313, 314], "bbox": [0.3865208333333333, 0.2535625, 0.6142291666666666, 0.82209375], "iscrowd": 0, "area": 16604.203950000006, "rle": {"size": [640, 480], "counts": "S_d38_c0:G:F9G9D<_Oa0K50O1O1O1O1O1O1OWAoMd;P2RD`Ni;_1lCQO[<c0[CM`<2YC;a<E^Cb0]<]ObCj0X<VOgCQ1T<nNkCY1o;gNPD_1l;cNPDd1k;aNnCf1l;`NmCg1n;^NkCi1o;]NjCj1Q<[NhCk1_;oNZDa1V;W4\\Od0]Oc0\\O<E23N2UFlFc8W9PGTGo8o8cF]G]9d9cNgEPIZ:k6QFoHo9l6]FlHc9o6iFjHX9d4^FYLf0lNl8l4cFmKl0POb8T5gFaKa;`4dDUKa;l4j11N2O1O1O1O1O1O1dBPKc;R5iC`KX<T61N101N101L4K4K6K4L5K5K4K6K4K6J6K4K6K4K6K4L5J5L5K4L5K4K5L5K4L5K4M4L3L4M4L3M4L3L5L3M4L3L8I7I7I7H8I7I7IVYc3"}, "label": "a giraffe standing in the middle of two other giraffes"}]}
{"id": 252025, "image": "COCO_train2014_000000252025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in the middle of two other griaffes\"."}], "task": "refering", "answer_template": "The \"a giraffe in the middle of two other griaffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 125, 126, 142, 143, 144, 160, 161, 162, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 228, 229, 230, 231, 245, 246, 247, 262, 263, 264, 279, 280, 281, 296, 297, 313, 314], "bbox": [0.3865208333333333, 0.2535625, 0.6142291666666666, 0.82209375], "iscrowd": 0, "area": 16604.203950000006, "rle": {"size": [640, 480], "counts": "S_d38_c0:G:F9G9D<_Oa0K50O1O1O1O1O1O1OWAoMd;P2RD`Ni;_1lCQO[<c0[CM`<2YC;a<E^Cb0]<]ObCj0X<VOgCQ1T<nNkCY1o;gNPD_1l;cNPDd1k;aNnCf1l;`NmCg1n;^NkCi1o;]NjCj1Q<[NhCk1_;oNZDa1V;W4\\Od0]Oc0\\O<E23N2UFlFc8W9PGTGo8o8cF]G]9d9cNgEPIZ:k6QFoHo9l6]FlHc9o6iFjHX9d4^FYLf0lNl8l4cFmKl0POb8T5gFaKa;`4dDUKa;l4j11N2O1O1O1O1O1O1dBPKc;R5iC`KX<T61N101N101L4K4K6K4L5K5K4K6K4K6J6K4K6K4K6K4L5J5L5K4L5K4K5L5K4L5K4M4L3L4M4L3M4L3L5L3M4L3L8I7I7I7H8I7I7IVYc3"}, "label": "a giraffe in the middle of two other griaffes"}]}
{"id": 252025, "image": "COCO_train2014_000000252025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in third place in a line of giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe in third place in a line of giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 77, 78, 79, 94, 95, 96, 112, 113, 129, 130, 131, 146, 147, 148, 163, 164, 165, 166, 180, 181, 182, 183, 197, 198, 199, 200, 214, 215, 216, 232, 233, 249, 250], "bbox": [0.5602916666666666, 0.087640625, 0.7910208333333333, 0.62921875], "iscrowd": 0, "area": 18188.77505, "rle": {"size": [640, 480], "counts": "XVX5;Uc0f0J6J6J6K4K6J6J6J6J5L5J6J6J5K4L4M3L4L4L4L4L4M3L4L4L3SCdKP:`4`EQL[:S4VE]Lf:f3jDkLQ;Y3`DXMZ;l2WDdMd;a5K5K5K5L4K5M3M00O1N2N2O1N2cGVF`6g;O1N2O1N2N2O1N2N3N11O1O1O2N1O1O2N1O1O1ON2N2M3N2N3L3N2M3N2M3N2M3N2K5K5K5K5K5K5K5J7J5K5K5K5K5K5K4L5K5K4L5K5K4L5K5K4L5L4K5K4L5K5K;Em0SOlVn1"}, "label": "a giraffe in third place in a line of giraffes"}]}
{"id": 252025, "image": "COCO_train2014_000000252025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe walking behind two other giraffes in an outside enclosure\"."}], "task": "refering", "answer_template": "The \"a tall giraffe walking behind two other giraffes in an outside enclosure\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 77, 78, 79, 94, 95, 96, 112, 113, 129, 130, 131, 146, 147, 148, 163, 164, 165, 166, 180, 181, 182, 183, 197, 198, 199, 200, 214, 215, 216, 232, 233, 249, 250], "bbox": [0.5602916666666666, 0.087640625, 0.7910208333333333, 0.62921875], "iscrowd": 0, "area": 18188.77505, "rle": {"size": [640, 480], "counts": "XVX5;Uc0f0J6J6J6K4K6J6J6J6J5L5J6J6J5K4L4M3L4L4L4L4L4M3L4L4L3SCdKP:`4`EQL[:S4VE]Lf:f3jDkLQ;Y3`DXMZ;l2WDdMd;a5K5K5K5L4K5M3M00O1N2N2O1N2cGVF`6g;O1N2O1N2N2O1N2N3N11O1O1O2N1O1O2N1O1O1ON2N2M3N2N3L3N2M3N2M3N2M3N2K5K5K5K5K5K5K5J7J5K5K5K5K5K5K4L5K5K4L5K5K4L5K5K4L5L4K5K4L5K5K;Em0SOlVn1"}, "label": "a tall giraffe walking behind two other giraffes in an outside enclosure"}]}
{"id": 505980, "image": "COCO_train2014_000000505980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing the toilet room front of the mirrior\"."}], "task": "refering", "answer_template": "The \"a man standing the toilet room front of the mirrior\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 102, 103, 124, 125, 126, 147, 148, 149, 170, 171, 193, 194, 216, 217, 239], "bbox": [0.395359375, 0.23536423841059603, 0.526078125, 0.6362693156732893], "iscrowd": 0, "area": 5054.182050000001, "rle": {"size": [453, 640], "counts": "ed`31S>1O1O1O1POP1N2]MoNoGT1P8TOgGn0X8YO`Gi0_8^OYGd0f8@UGb0j8BQG?o8DmF>R9EjF=U9FgF<X9GdF;[9IaF7_9X2O1O1N2N101N2N100O2O0O101N101NnMXGjNg8m0iGmNW8n0RHoNZ7SObGl1[1mNT7[OVGK1h1l1oNn6FSGY1S2nNl6`1XI]Ni6a1ZI^Ng6_1]I^Ne6^1`I_Nb6]1cI`N^6_1eI_N\\6a1fI[N\\6e1eIXN]6h1Q3O1O100O2O0ROiCNV<KTD3k;KYD4f;J^D5b;H`D9b;BaD>e<0N000001N100O10000O2N1O1O1O10O01O1O1O1O1O2N1OeYV4"}, "label": "a man standing the toilet room front of the mirrior"}]}
{"id": 505980, "image": "COCO_train2014_000000505980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man looking to a mirror\"."}], "task": "refering", "answer_template": "The \"a man looking to a mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 102, 103, 124, 125, 126, 147, 148, 149, 170, 171, 193, 194, 216, 217, 239], "bbox": [0.395359375, 0.23536423841059603, 0.526078125, 0.6362693156732893], "iscrowd": 0, "area": 5054.182050000001, "rle": {"size": [453, 640], "counts": "ed`31S>1O1O1O1POP1N2]MoNoGT1P8TOgGn0X8YO`Gi0_8^OYGd0f8@UGb0j8BQG?o8DmF>R9EjF=U9FgF<X9GdF;[9IaF7_9X2O1O1N2N101N2N100O2O0O101N101NnMXGjNg8m0iGmNW8n0RHoNZ7SObGl1[1mNT7[OVGK1h1l1oNn6FSGY1S2nNl6`1XI]Ni6a1ZI^Ng6_1]I^Ne6^1`I_Nb6]1cI`N^6_1eI_N\\6a1fI[N\\6e1eIXN]6h1Q3O1O100O2O0ROiCNV<KTD3k;KYD4f;J^D5b;H`D9b;BaD>e<0N000001N100O10000O2N1O1O1O10O01O1O1O1O1O2N1OeYV4"}, "label": "a man looking to a mirror"}]}
{"id": 350335, "image": "COCO_train2014_000000350335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard laying on its side , looks to be very used\"."}], "task": "refering", "answer_template": "The \"a skateboard laying on its side , looks to be very used\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 266, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 311, 312, 313], "bbox": [0.5725235849056604, 0.7521875, 0.9225943396226415, 0.8967968749999999], "iscrowd": 0, "area": 8743.541650000003, "rle": {"size": [640, 424], "counts": "Y[h45gc07J7L8H8J7I6H3M1O00010O00001O00001O01O0001O00001O01O01O00001O000010O000001O000010O0001O00001O00010O0000001O00010O00001O000010O0001O0000010O00001O001O2O1N2N2N1O2N100O0000001O0001O0001O0000O2O0O10000O100O2O0000000000001O00000000001O0000000000001O0O1000002N1O2N2N1O2M2O2N2M2O2M3N1N3N1O2M3N1N3N1N3NWSd0"}, "label": "a skateboard laying on its side , looks to be very used"}]}
{"id": 350335, "image": "COCO_train2014_000000350335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard on its side\"."}], "task": "refering", "answer_template": "The \"a skateboard on its side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 266, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 311, 312, 313], "bbox": [0.5725235849056604, 0.7521875, 0.9225943396226415, 0.8967968749999999], "iscrowd": 0, "area": 8743.541650000003, "rle": {"size": [640, 424], "counts": "Y[h45gc07J7L8H8J7I6H3M1O00010O00001O00001O01O0001O00001O01O01O00001O000010O000001O000010O0001O00001O00010O0000001O00010O00001O000010O0001O0000010O00001O001O2O1N2N2N1O2N100O0000001O0001O0001O0000O2O0O10000O100O2O0000000000001O00000000001O0000000000001O0O1000002N1O2N2N1O2M2O2N2M2O2M3N1N3N1O2M3N1N3N1N3NWSd0"}, "label": "a skateboard on its side"}]}
{"id": 546944, "image": "COCO_train2014_000000546944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair , hardly visible behind the cat ' s left and right shoulder\"."}], "task": "refering", "answer_template": "The \"the chair , hardly visible behind the cat ' s left and right shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 127, 128, 145, 195, 218], "bbox": [0.31710937499999997, 0.29252083333333334, 0.609984375, 0.5454791666666666], "iscrowd": 0, "area": 1598.9685999999997, "rle": {"size": [480, 640], "counts": "bYo23m>8H7I8H3M00000000000000O101O00000000008H6I2O0O1O1O1O1O1O1O2N1OQeT12lZkN2M3N2N2M3N2M2O2M3N2M3N2M3N27Ig0YO1O1O1O1O001Okg93RXF4L0000000000000^Ob00001O000000000000000000001O00000O1005K5K5K5K5LjPe3"}, "label": "the chair , hardly visible behind the cat ' s left and right shoulder"}]}
{"id": 546944, "image": "COCO_train2014_000000546944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair behind the cat ' s head\"."}], "task": "refering", "answer_template": "The \"chair behind the cat ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 127, 128, 145, 195, 218], "bbox": [0.31710937499999997, 0.29252083333333334, 0.609984375, 0.5454791666666666], "iscrowd": 0, "area": 1598.9685999999997, "rle": {"size": [480, 640], "counts": "bYo23m>8H7I8H3M00000000000000O101O00000000008H6I2O0O1O1O1O1O1O1O2N1OQeT12lZkN2M3N2N2M3N2M2O2M3N2M3N2M3N27Ig0YO1O1O1O1O001Okg93RXF4L0000000000000^Ob00001O000000000000000000001O00000O1005K5K5K5K5LjPe3"}, "label": "chair behind the cat ' s head"}]}
{"id": 546944, "image": "COCO_train2014_000000546944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair behind cat\"."}], "task": "refering", "answer_template": "The \"chair behind cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 139, 140, 141, 162, 163, 164, 165, 185, 186, 187, 188, 209, 210, 211, 232, 233, 234, 255, 256, 278, 301, 324], "bbox": [0.06331250000000001, 0.3268333333333333, 0.1915625, 0.8657916666666666], "iscrowd": 0, "area": 11014.936649999998, "rle": {"size": [480, 640], "counts": "^\\c0:`><E;D<J6K5K5J6K5K5J6K5J6K5K5J6K4L3N1YHbLi3^3aKVM^4m2\\KWMc4m2UKYMi4j2QKXMP5k2iJYMW5j2bJYM_5k2ZJXMf5k2SJYMm5j2mIXMT6l2fHVNZ7h40O1O1O1N2O1N2O1K5fNZ11OM3J6J6J6J6J6L4000000O1000O10000O01000O1L33L4L5K4L4M4K4L4L5K4L4Lj0VO4J5L5J6L3L5L3L5K4DXTb7"}, "label": "chair behind cat"}]}
{"id": 546944, "image": "COCO_train2014_000000546944.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair with a light - colored cushioned and positioned behind a white cat\"."}], "task": "refering", "answer_template": "The \"a wooden chair with a light - colored cushioned and positioned behind a white cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 139, 140, 141, 162, 163, 164, 165, 185, 186, 187, 188, 209, 210, 211, 232, 233, 234, 255, 256, 278, 301, 324], "bbox": [0.06331250000000001, 0.3268333333333333, 0.1915625, 0.8657916666666666], "iscrowd": 0, "area": 11014.936649999998, "rle": {"size": [480, 640], "counts": "^\\c0:`><E;D<J6K5K5J6K5K5J6K5J6K5K5J6K4L3N1YHbLi3^3aKVM^4m2\\KWMc4m2UKYMi4j2QKXMP5k2iJYMW5j2bJYM_5k2ZJXMf5k2SJYMm5j2mIXMT6l2fHVNZ7h40O1O1O1N2O1N2O1K5fNZ11OM3J6J6J6J6J6L4000000O1000O10000O01000O1L33L4L5K4L4M4K4L4L5K4L4Lj0VO4J5L5J6L3L5L3L5K4DXTb7"}, "label": "a wooden chair with a light - colored cushioned and positioned behind a white cat"}]}
{"id": 202881, "image": "COCO_train2014_000000202881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow with both ears tagged\"."}], "task": "refering", "answer_template": "The \"the cow with both ears tagged\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 220, 221, 222, 236, 237, 238, 252, 253, 254], "bbox": [0.6208764044943821, 0.3425, 0.9526741573033709, 0.88612], "iscrowd": 0, "area": 25196.62435, "rle": {"size": [500, 445], "counts": "`oV43[?9G:E:PCn0V9XO[FT1a9BaEi0]:Y2M2N3L4M2N3M2O2O0OZNRFiMm9W2YFcMf9]2bFZM`9f2fFSMZ9m2jFnLX9S3jFiLV9Y3mFbLT9_3`11N2O1N2O1N2O1N101N2O1N3N2N2M3N0O2O0O2O001_EZKR:f4iEaKU:R5M2N2N1O00001O001O00001O010fHZJd4g5WK]Jj4c5PKbJQ5]5oH]J^1;c5Y5lH`JZ1=j5S5hHdJW1?R6\\5fIjJZ6W5TIZKm6g4kH_KU7a60O1O001O0010O01O0O2N101`NjH_JW7_5QI[JP7c5VIWJk6_5cIUJf6h5d1N3O00001O001O010O001O0001aGZJo6e5kHbJT7^5lGbKT8^4iGeKW8l51O00001O000010OO101O1O1O2M3N2N2YNdHRK]7k4iHPKY7l4nHPKT7l4l1M4L3M3M4L3N2M4L3N3L3M3N3L3M3N3L3J6J7I7I6J7J6K4L5K4K6K?ARP:"}, "label": "the cow with both ears tagged"}]}
{"id": 202881, "image": "COCO_train2014_000000202881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white adult cow with tags in its ears in front of a calf\"."}], "task": "refering", "answer_template": "The \"a black and white adult cow with tags in its ears in front of a calf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 220, 221, 222, 236, 237, 238, 252, 253, 254], "bbox": [0.6208764044943821, 0.3425, 0.9526741573033709, 0.88612], "iscrowd": 0, "area": 25196.62435, "rle": {"size": [500, 445], "counts": "`oV43[?9G:E:PCn0V9XO[FT1a9BaEi0]:Y2M2N3L4M2N3M2O2O0OZNRFiMm9W2YFcMf9]2bFZM`9f2fFSMZ9m2jFnLX9S3jFiLV9Y3mFbLT9_3`11N2O1N2O1N2O1N101N2O1N3N2N2M3N0O2O0O2O001_EZKR:f4iEaKU:R5M2N2N1O00001O001O00001O010fHZJd4g5WK]Jj4c5PKbJQ5]5oH]J^1;c5Y5lH`JZ1=j5S5hHdJW1?R6\\5fIjJZ6W5TIZKm6g4kH_KU7a60O1O001O0010O01O0O2N101`NjH_JW7_5QI[JP7c5VIWJk6_5cIUJf6h5d1N3O00001O001O010O001O0001aGZJo6e5kHbJT7^5lGbKT8^4iGeKW8l51O00001O000010OO101O1O1O2M3N2N2YNdHRK]7k4iHPKY7l4nHPKT7l4l1M4L3M3M4L3N2M4L3N3L3M3N3L3M3N3L3J6J7I7I6J7J6K4L5K4K6K?ARP:"}, "label": "a black and white adult cow with tags in its ears in front of a calf"}]}
{"id": 14468, "image": "COCO_train2014_000000014468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a black tee shirt\"."}], "task": "refering", "answer_template": "The \"a person in a black tee shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 59, 60, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 311, 312, 313, 336, 337], "bbox": [0.49887499999999996, 0.103375, 0.6960624999999999, 0.8651666666666668], "iscrowd": 0, "area": 27789.663449999996, "rle": {"size": [480, 640], "counts": "Zee4m2Q<6K4L5K5J6K5K5K4K6K5K5K5J7J;E;E;D<E;E;E;D<E4L2N2M3N2N2N2M3N2O1ON2O1N2N2N2N2N2N2O1NQORHPKl7n4bHhJ\\7V5RI`Jl6]5cIYJ[6e5UOnJjJT5V5cKQJ_4o5Y21O0000001O000000001O00R3nL4L3M4L4L3M4L3M4L3M4L0000001O0000001O01O000001O000000bKlLQMS3eM]Ld3T1UN_2TN`Ld3g1eMi1cNeLe3W2UMV1QOfLg3j2eLb0^OiLj3R3^L7BjLm3U3^L3_OlLQ4W3\\L0]OmLS4Z3\\LLZOnLX4[3\\Lo0`3XO\\Lh0b3]O[Ld0a3C[L=c3HZL9b3N[L1c34ZLLc3;YLFd3l5M4M2M4M2M4M2M4M2M3N3L3N3L3N3L3N3L3N2M4J5SOVZk2"}, "label": "a person in a black tee shirt"}]}
{"id": 14468, "image": "COCO_train2014_000000014468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man with a shirt that has a giant musical note on it\"."}], "task": "refering", "answer_template": "The \"a young man with a shirt that has a giant musical note on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 59, 60, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 311, 312, 313, 336, 337], "bbox": [0.49887499999999996, 0.103375, 0.6960624999999999, 0.8651666666666668], "iscrowd": 0, "area": 27789.663449999996, "rle": {"size": [480, 640], "counts": "Zee4m2Q<6K4L5K5J6K5K5K4K6K5K5K5J7J;E;E;D<E;E;E;D<E4L2N2M3N2N2N2M3N2O1ON2O1N2N2N2N2N2N2O1NQORHPKl7n4bHhJ\\7V5RI`Jl6]5cIYJ[6e5UOnJjJT5V5cKQJ_4o5Y21O0000001O000000001O00R3nL4L3M4L4L3M4L3M4L3M4L0000001O0000001O01O000001O000000bKlLQMS3eM]Ld3T1UN_2TN`Ld3g1eMi1cNeLe3W2UMV1QOfLg3j2eLb0^OiLj3R3^L7BjLm3U3^L3_OlLQ4W3\\L0]OmLS4Z3\\LLZOnLX4[3\\Lo0`3XO\\Lh0b3]O[Ld0a3C[L=c3HZL9b3N[L1c34ZLLc3;YLFd3l5M4M2M4M2M4M2M4M2M3N3L3N3L3N3L3N3L3N2M4J5SOVZk2"}, "label": "a young man with a shirt that has a giant musical note on it"}]}
{"id": 14468, "image": "COCO_train2014_000000014468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a striped shirt\"."}], "task": "refering", "answer_template": "The \"a boy wearing a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 214, 215, 216, 237, 238, 260, 261, 283, 284, 306, 307, 330, 331], "bbox": [0.248875, 0.1280625, 0.459078125, 0.8769374999999999], "iscrowd": 0, "area": 24380.736549999998, "rle": {"size": [480, 640], "counts": "ghZ23f>7H8H9H7H8H8I8G8H8I8G8H8H7L4O1O100O1OO2O000O2O000O2O002O03N3L3M4M3L3N9F;E<E;C=D;E<D<D;E4L5K4A`0G8H9J5K6K4K5K6K4L5L3L5K3M3M3N2M3M3M3N2M3N1N2O1N101N20O01O1O1O001OoLRLQKm3m4XLPKi3m4\\LQKc3m4bLQK]3m4hLQKW3m4nL`IUO<m3R6TM\\IZO:a3X6ZMZIXO=]3V6bMWIUOa0Y3V6gMUISOd0V3T6mMRIPOj0S3P6SNRImNl0Q3m5YOPJj0k5\\ORJe0i5CSJ?h5HSJ;h5LTJ5g5f4K4Kl0UO3L5L3L4M3K5K5K6J5K5K5K5K6J5K5K5M3L4L5K4L4M3L4L4L5L4K6J5K6J5L5J5K6J5L\\RR5"}, "label": "a boy wearing a striped shirt"}]}
{"id": 14468, "image": "COCO_train2014_000000014468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue jeans\"."}], "task": "refering", "answer_template": "The \"a man wearing blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 214, 215, 216, 237, 238, 260, 261, 283, 284, 306, 307, 330, 331], "bbox": [0.248875, 0.1280625, 0.459078125, 0.8769374999999999], "iscrowd": 0, "area": 24380.736549999998, "rle": {"size": [480, 640], "counts": "ghZ23f>7H8H9H7H8H8I8G8H8I8G8H8H7L4O1O100O1OO2O000O2O000O2O002O03N3L3M4M3L3N9F;E<E;C=D;E<D<D;E4L5K4A`0G8H9J5K6K4K5K6K4L5L3L5K3M3M3N2M3M3M3N2M3N1N2O1N101N20O01O1O1O001OoLRLQKm3m4XLPKi3m4\\LQKc3m4bLQK]3m4hLQKW3m4nL`IUO<m3R6TM\\IZO:a3X6ZMZIXO=]3V6bMWIUOa0Y3V6gMUISOd0V3T6mMRIPOj0S3P6SNRImNl0Q3m5YOPJj0k5\\ORJe0i5CSJ?h5HSJ;h5LTJ5g5f4K4Kl0UO3L5L3L4M3K5K5K6J5K5K5K5K6J5K5K5M3L4L5K4L4M3L4L4L5L4K6J5K6J5L5J5K6J5L\\RR5"}, "label": "a man wearing blue jeans"}]}
{"id": 317573, "image": "COCO_train2014_000000317573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person to the right\"."}], "task": "refering", "answer_template": "The \"person to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160], "bbox": [0.7550625, 0.008979166666666667, 1.0, 0.38877083333333334], "iscrowd": 0, "area": 24984.134499999993, "rle": {"size": [480, 640], "counts": "T`R72k>7I6I8I6I7J7I6I8I6aDRNb9V2WFmMf9Y2QFlMm9[2iEkMT:[2cEkMY:]2]EiM`:]3M3M3M3L4M3M3M3N2M3M3M3N2M3N2O010O10000O100O10000O100O100000000001O0O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000010O001O001O001O1O001O001O001O001O1O00000000O100O100O100O1O1O1O1N2O1O1N2O1O1O1N2000000O1000000O10005K\\1dN\\1cN]1dNaK"}, "label": "person to the right"}]}
{"id": 317573, "image": "COCO_train2014_000000317573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person who can ' t really be seen eating\"."}], "task": "refering", "answer_template": "The \"the person who can ' t really be seen eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160], "bbox": [0.7550625, 0.008979166666666667, 1.0, 0.38877083333333334], "iscrowd": 0, "area": 24984.134499999993, "rle": {"size": [480, 640], "counts": "T`R72k>7I6I8I6I7J7I6I8I6aDRNb9V2WFmMf9Y2QFlMm9[2iEkMT:[2cEkMY:]2]EiM`:]3M3M3M3L4M3M3M3N2M3M3M3N2M3N2O010O10000O100O10000O100O100000000001O0O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000010O001O001O001O1O001O001O001O001O1O00000000O100O100O100O1O1O1O1N2O1O1N2O1O1O1N2000000O1000000O10005K\\1dN\\1cN]1dNaK"}, "label": "the person who can ' t really be seen eating"}]}
{"id": 317573, "image": "COCO_train2014_000000317573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is eating dry fruits\"."}], "task": "refering", "answer_template": "The \"a man is eating dry fruits\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167], "bbox": [0.168515625, 0.0, 0.703703125, 0.43704166666666666], "iscrowd": 0, "area": 59629.162999999986, "rle": {"size": [480, 640], "counts": "Pdb16j>8H8H8H8H8H9G8H8H8H8H8H9G8H8H8H8H8H9G3M2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N1O2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O1OO100000000O1000000O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O2N1O100O10000O10000O1000000O10000O10000O10000O10000O100000000000000O10000000000000000000000000000000000O10000000001O000000000000000000000O10000000000000000000000000000000000O10000000000001O001O00001O00001O00001O00001O00001O000010O0001O00001O00001O0000000000000000O100000000000000000000000000000000000000000000000000000000O100000000000000000001O00000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O1000000000000000001O00000000000000000000000000000000000O10000000000000000eNUSi2"}, "label": "a man is eating dry fruits"}]}
{"id": 317573, "image": "COCO_train2014_000000317573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man taking food with spoon in a bowl\"."}], "task": "refering", "answer_template": "The \"a man taking food with spoon in a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 166, 167], "bbox": [0.168515625, 0.0, 0.703703125, 0.43704166666666666], "iscrowd": 0, "area": 59629.162999999986, "rle": {"size": [480, 640], "counts": "Pdb16j>8H8H8H8H8H9G8H8H8H8H8H9G8H8H8H8H8H9G3M2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N1O2N2N2N1O1O1O1O1O1O1O1O1O1O1O1O1OO100000000O1000000O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O2N1O100O10000O10000O1000000O10000O10000O10000O10000O100000000000000O10000000000000000000000000000000000O10000000001O000000000000000000000O10000000000000000000000000000000000O10000000000001O001O00001O00001O00001O00001O00001O000010O0001O00001O00001O0000000000000000O100000000000000000000000000000000000000000000000000000000O100000000000000000001O00000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O1000000000000000001O00000000000000000000000000000000000O10000000000000000eNUSi2"}, "label": "a man taking food with spoon in a bowl"}]}
{"id": 88200, "image": "COCO_train2014_000000088200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball batter\"."}], "task": "refering", "answer_template": "The \"a baseball batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 248, 249, 265, 266, 267, 271, 272, 288, 289], "bbox": [0.523578125, 0.13004683840749415, 0.88559375, 0.8296018735362998], "iscrowd": 0, "area": 28170.281150000006, "rle": {"size": [427, 640], "counts": "ij[42X=2M3M3M3O1N2O1dHAi1a0UNAj1`0TNBk1>TNCk1=TNFk1:TNHk17UNKi16VNLi13WNOh11VN2h1OWN3h1LXN6g1JWN9g1GYN;f1EXN>g1AYN`0f1AXNb0g1^OXNd0f1]OYNe0f1ZOYNi0f1WOYNk0e1VOYNl0g1SOYNo0e1ROZNo0f1QOYNP1f1QOXNQ1h1nNXNS1g1nNXNS1h1mNWNT1h1mNWNT1i1kNWNV1h1kNVNX1i1hNVNY1i1hNVNc0eKZOU62VNe0fKXOS64VNd0iKWOQ65UNd0lKVOn57TNd0PLTOl57TNf0PLSOk58TNe0SLROi59SNe0VLPOg5;SNe0XLoNe5<RNf0ZLmNc5>QNf0^LkNa5?PNf0aLiN_5a0PNf0bLiN^5a0oMg0dLgN\\5c0oMf0gLfNZ5d0nMf0jLdNX5g0mMe0mLcNV5h0lMf0oLaNT5j0lMe0RM`NR5k0kMe0U3\\OjLd0W3\\OhLe0W3\\OhLd0Y3\\OgLc0Y3^OfLb0[3^OdLc0[3^OdLb0]3^ObLb0^3_OaLa0`3_O_Lb0`3_O_La0b3_O]La0c3A[L?f3CWL>h3EVL:k3GSL9m3JPL6Q4LlK5S4NgK5Q3\\NYKa1`16W3]NUK`1^16\\3^NTK^1Z18a3^NQK\\1Y19f3^NoJ[1T1;m3^NkJZ1R1;S4^NiJY1n0=X4^NfJX1g0b0b4ZNeJV12T1Z5iM`Jn4a5TK]Jl4c5WKYJi4h5YKVJf4k5\\11O100O1O1O1O1O1O1O1O2O0O1O1O1O1O1O1O1N2O1N2N2O1N2N2O1N3M2O1K5L4K51O1O1O1O1O1O1O001O11O1O2N1O2N2N1O2N1O2N001O00001O00001O001\\LVHd1j7[NXHd1h7\\NYHb1i7\\NYHc1g7\\N[Hc1e7\\N^Hb1c7]N^Hb1b7]N`Hb1`7]NbHb1_7\\NdHb1\\7^NeHa1[7^NgHa1Z7]NhHb1X7]NkHa1V7^NkHa1T7_NnHa1Q7^NRIb1l6]NVIc1h6^NYIb1f6]N\\Id1a6\\NbIc1]6\\NeId1Z6\\NgIe1V6[NmId1R6[NPJe1o5ZNSJg1j5ZNXJe1g5ZN[Jg1c5XN_Ji1^5WNdJl1X5TNiJn1S5RNoJQ2m4nMUKT2h4kMZKU2d4lM]KR2d4mM^KQ2c4nM^KQ2b4oM`Ko1a4QN`Kn1`4QNbKm1^4SNdKk1]4TNeKj1\\4VNeKh1\\4WNfKg1[4XNiKd1X4[NlKb1T4]NPL_1Q4`NSL\\1n3cNVLY1k3fNYLV1h3iN]LR1d3mN`Lo0a3POcLk0_3TOeLh0\\3WOgLf0Z3YOiLd0X3[OkLb0V3]OnL?S3@PM=Q3BRM;o2DTM9m2EWM8j2GYM6h2I\\M3e2L^M1c2N`MMc22_MKb25^MJd25]MIe26[MIg26YMIi26W5O1O1O1O1O1OkXn0"}, "label": "a baseball batter"}]}
{"id": 88200, "image": "COCO_train2014_000000088200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player weaing a white uniform and blue helmet and holding a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player weaing a white uniform and blue helmet and holding a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 248, 249, 265, 266, 267, 271, 272, 288, 289], "bbox": [0.523578125, 0.13004683840749415, 0.88559375, 0.8296018735362998], "iscrowd": 0, "area": 28170.281150000006, "rle": {"size": [427, 640], "counts": "ij[42X=2M3M3M3O1N2O1dHAi1a0UNAj1`0TNBk1>TNCk1=TNFk1:TNHk17UNKi16VNLi13WNOh11VN2h1OWN3h1LXN6g1JWN9g1GYN;f1EXN>g1AYN`0f1AXNb0g1^OXNd0f1]OYNe0f1ZOYNi0f1WOYNk0e1VOYNl0g1SOYNo0e1ROZNo0f1QOYNP1f1QOXNQ1h1nNXNS1g1nNXNS1h1mNWNT1h1mNWNT1i1kNWNV1h1kNVNX1i1hNVNY1i1hNVNc0eKZOU62VNe0fKXOS64VNd0iKWOQ65UNd0lKVOn57TNd0PLTOl57TNf0PLSOk58TNe0SLROi59SNe0VLPOg5;SNe0XLoNe5<RNf0ZLmNc5>QNf0^LkNa5?PNf0aLiN_5a0PNf0bLiN^5a0oMg0dLgN\\5c0oMf0gLfNZ5d0nMf0jLdNX5g0mMe0mLcNV5h0lMf0oLaNT5j0lMe0RM`NR5k0kMe0U3\\OjLd0W3\\OhLe0W3\\OhLd0Y3\\OgLc0Y3^OfLb0[3^OdLc0[3^OdLb0]3^ObLb0^3_OaLa0`3_O_Lb0`3_O_La0b3_O]La0c3A[L?f3CWL>h3EVL:k3GSL9m3JPL6Q4LlK5S4NgK5Q3\\NYKa1`16W3]NUK`1^16\\3^NTK^1Z18a3^NQK\\1Y19f3^NoJ[1T1;m3^NkJZ1R1;S4^NiJY1n0=X4^NfJX1g0b0b4ZNeJV12T1Z5iM`Jn4a5TK]Jl4c5WKYJi4h5YKVJf4k5\\11O100O1O1O1O1O1O1O1O2O0O1O1O1O1O1O1O1N2O1N2N2O1N2N2O1N3M2O1K5L4K51O1O1O1O1O1O1O001O11O1O2N1O2N2N1O2N1O2N001O00001O00001O001\\LVHd1j7[NXHd1h7\\NYHb1i7\\NYHc1g7\\N[Hc1e7\\N^Hb1c7]N^Hb1b7]N`Hb1`7]NbHb1_7\\NdHb1\\7^NeHa1[7^NgHa1Z7]NhHb1X7]NkHa1V7^NkHa1T7_NnHa1Q7^NRIb1l6]NVIc1h6^NYIb1f6]N\\Id1a6\\NbIc1]6\\NeId1Z6\\NgIe1V6[NmId1R6[NPJe1o5ZNSJg1j5ZNXJe1g5ZN[Jg1c5XN_Ji1^5WNdJl1X5TNiJn1S5RNoJQ2m4nMUKT2h4kMZKU2d4lM]KR2d4mM^KQ2c4nM^KQ2b4oM`Ko1a4QN`Kn1`4QNbKm1^4SNdKk1]4TNeKj1\\4VNeKh1\\4WNfKg1[4XNiKd1X4[NlKb1T4]NPL_1Q4`NSL\\1n3cNVLY1k3fNYLV1h3iN]LR1d3mN`Lo0a3POcLk0_3TOeLh0\\3WOgLf0Z3YOiLd0X3[OkLb0V3]OnL?S3@PM=Q3BRM;o2DTM9m2EWM8j2GYM6h2I\\M3e2L^M1c2N`MMc22_MKb25^MJd25]MIe26[MIg26YMIi26W5O1O1O1O1O1OkXn0"}, "label": "a baseball player weaing a white uniform and blue helmet and holding a bat"}]}
{"id": 88200, "image": "COCO_train2014_000000088200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player playing the catcher position\"."}], "task": "refering", "answer_template": "The \"the baseball player playing the catcher position\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 234, 235, 236, 256, 257, 258, 259, 280, 281, 282], "bbox": [0.1424375, 0.3730444964871194, 0.367328125, 0.8471896955503513], "iscrowd": 0, "area": 15004.5776, "rle": {"size": [427, 640], "counts": "SVV11X=2M4M3M3M3M2N3M3M3M2O2N2M3N1O1N2eEPOZ8Q1aGXOZ8i0aG@Y8a0eGEW8<fGIW88fGLX85eG0W82gG2V8OgG6V8KgG:U8HhG<V8EgG`0V8AhGc0T8^OjGg0S8ZOiGk0U8VOgGU1P8mNlGW1Q8lNkGX1S8P2N3G9K5N2O1O1O1O1N2O1O1O1O0O20000O10000O10000O1000O10O100000000O1001O2N1O1N3N1O1O2N3M3M3SOZHoKh7n3ZHQL^8V3eGhL_8S3cGlLa8n2bGQMb8j2aGSMd8g2_GXMa8g2bGWM]8h2fGWMY8i2jGUMV8i2mGVMR8j2QHTMn7k2UHSMk7m2XHPMh7o2Z10O00_NkEOT:2oEJQ:7QFGn9:UF\\Oo9g0SFVOk9m0XFPOi9o0ZFmNg9S1[FkNe9U1^FgNc9Y1_FeNb9Z1aFcN_9]1dF_N]9a1eF]N\\9b1V1000001O002N1O1O2N1O1O2M2N2O2M2O1O2O0O1O2O0O1O2O1N3N1O2M2O2N1N3N1N3L3LheX5"}, "label": "the baseball player playing the catcher position"}]}
{"id": 88200, "image": "COCO_train2014_000000088200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher wearing a red and white uniform\"."}], "task": "refering", "answer_template": "The \"a catcher wearing a red and white uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 234, 235, 236, 256, 257, 258, 259, 280, 281, 282], "bbox": [0.1424375, 0.3730444964871194, 0.367328125, 0.8471896955503513], "iscrowd": 0, "area": 15004.5776, "rle": {"size": [427, 640], "counts": "SVV11X=2M4M3M3M3M2N3M3M3M2O2N2M3N1O1N2eEPOZ8Q1aGXOZ8i0aG@Y8a0eGEW8<fGIW88fGLX85eG0W82gG2V8OgG6V8KgG:U8HhG<V8EgG`0V8AhGc0T8^OjGg0S8ZOiGk0U8VOgGU1P8mNlGW1Q8lNkGX1S8P2N3G9K5N2O1O1O1O1N2O1O1O1O0O20000O10000O10000O1000O10O100000000O1001O2N1O1N3N1O1O2N3M3M3SOZHoKh7n3ZHQL^8V3eGhL_8S3cGlLa8n2bGQMb8j2aGSMd8g2_GXMa8g2bGWM]8h2fGWMY8i2jGUMV8i2mGVMR8j2QHTMn7k2UHSMk7m2XHPMh7o2Z10O00_NkEOT:2oEJQ:7QFGn9:UF\\Oo9g0SFVOk9m0XFPOi9o0ZFmNg9S1[FkNe9U1^FgNc9Y1_FeNb9Z1aFcN_9]1dF_N]9a1eF]N\\9b1V1000001O002N1O1O2N1O1O2M2N2O2M2O1O2O0O1O2O0O1O2O1N3N1O2M2O2N1N3N1N3L3LheX5"}, "label": "a catcher wearing a red and white uniform"}]}
{"id": 219274, "image": "COCO_train2014_000000219274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green , white and gray bus\"."}], "task": "refering", "answer_template": "The \"a green , white and gray bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 78, 79, 91, 92, 93, 104, 105, 106, 107, 117, 118, 119, 120, 130, 131, 132, 133, 143, 144, 145, 146, 156, 157, 158, 159, 169, 170, 171, 172, 182, 183, 184, 195, 196, 208], "bbox": [0.00045333333333333337, 0.29902, 0.27192, 0.94216], "iscrowd": 0, "area": 21733.121999999996, "rle": {"size": [500, 375], "counts": "f4e1P>b5^Jg2YMN2N2N2N2N2N2N2N2N2N3M2M3N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N3M2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N3M2N2N2N2N2M3N2N2N2N2N2N2N2N2O1000001O0O10000000000000000O1M3L4J6G9G9K5M3N3L3N2N2M3N2N2M3N2M3N2N2M3F:ROn0QOo0ROaRU4"}, "label": "a green , white and gray bus"}]}
{"id": 219274, "image": "COCO_train2014_000000219274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and green bus driving in front of the other bus\"."}], "task": "refering", "answer_template": "The \"a white and green bus driving in front of the other bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 78, 79, 91, 92, 93, 104, 105, 106, 107, 117, 118, 119, 120, 130, 131, 132, 133, 143, 144, 145, 146, 156, 157, 158, 159, 169, 170, 171, 172, 182, 183, 184, 195, 196, 208], "bbox": [0.00045333333333333337, 0.29902, 0.27192, 0.94216], "iscrowd": 0, "area": 21733.121999999996, "rle": {"size": [500, 375], "counts": "f4e1P>b5^Jg2YMN2N2N2N2N2N2N2N2N2N3M2M3N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N3M2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N3M2N2N2N2N2M3N2N2N2N2N2N2N2N2O1000001O0O10000000000000000O1M3L4J6G9G9K5M3N3L3N2N2M3N2N2M3N2M3N2N2M3F:ROn0QOo0ROaRU4"}, "label": "a white and green bus driving in front of the other bus"}]}
{"id": 309400, "image": "COCO_train2014_000000309400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vertical computer monitor to the right of other monitor\"."}], "task": "refering", "answer_template": "The \"vertical computer monitor to the right of other monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 225], "bbox": [0.58428125, 0.08600000000000001, 0.85415625, 0.5372708333333333], "iscrowd": 0, "area": 30731.683599999993, "rle": {"size": [480, 640], "counts": "Pa_58Y>`0@`0_O`0A`0_O`0A?A`0_O`0A?@`0A?B>00010O0000000001O00000001O01O0000000000001O01O00000001O00000000010O0000000000001O0001O000001O000000000010O000000000001O000001O0001O00000000001O01O0000000001O00000001O01O00000000001O01O0000000001O00000001O01O00000000001O0001O000001O0000000000010O00000000001O000001O0001O00M3H8H8H8G9H9G8H8H8G9H8H8H8G9H8H9G8H8G9H8H8H8H8G9H_a[1"}, "label": "vertical computer monitor to the right of other monitor"}]}
{"id": 309400, "image": "COCO_train2014_000000309400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the screen closest to the telephone and mouse\"."}], "task": "refering", "answer_template": "The \"the screen closest to the telephone and mouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 225], "bbox": [0.58428125, 0.08600000000000001, 0.85415625, 0.5372708333333333], "iscrowd": 0, "area": 30731.683599999993, "rle": {"size": [480, 640], "counts": "Pa_58Y>`0@`0_O`0A`0_O`0A?A`0_O`0A?@`0A?B>00010O0000000001O00000001O01O0000000000001O01O00000001O00000000010O0000000000001O0001O000001O000000000010O000000000001O000001O0001O00000000001O01O0000000001O00000001O01O00000000001O01O0000000001O00000001O01O00000000001O0001O000001O0000000000010O00000000001O000001O0001O00M3H8H8H8G9H9G8H8H8G9H8H8H8G9H8H9G8H8G9H8H8H8H8G9H_a[1"}, "label": "the screen closest to the telephone and mouse"}]}
{"id": 309400, "image": "COCO_train2014_000000309400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pc monitor that is inbetween a laptop and another monitor\"."}], "task": "refering", "answer_template": "The \"a pc monitor that is inbetween a laptop and another monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197], "bbox": [0.35415625, 0.07920833333333334, 0.59640625, 0.4899791666666667], "iscrowd": 0, "area": 29174.354750000002, "rle": {"size": [480, 640], "counts": "XaZ3R3l8R30001O000000000000000000000001O00000000000000000000000000000000000000000000000000000O2O000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000O10001O0000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000M3ROn0ROn0SOm0ROn0SOm0ROcmh3"}, "label": "a pc monitor that is inbetween a laptop and another monitor"}]}
{"id": 309400, "image": "COCO_train2014_000000309400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tall monitor closest to the laptop\"."}], "task": "refering", "answer_template": "The \"the tall monitor closest to the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197], "bbox": [0.35415625, 0.07920833333333334, 0.59640625, 0.4899791666666667], "iscrowd": 0, "area": 29174.354750000002, "rle": {"size": [480, 640], "counts": "XaZ3R3l8R30001O000000000000000000000001O00000000000000000000000000000000000000000000000000000O2O000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000O10001O0000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000M3ROn0ROn0SOm0ROn0SOm0ROcmh3"}, "label": "the tall monitor closest to the laptop"}]}
{"id": 170139, "image": "COCO_train2014_000000170139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall plant in the kitchen , near to table\"."}], "task": "refering", "answer_template": "The \"a tall plant in the kitchen , near to table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 15, 16, 17, 30, 31, 32, 46, 47, 61, 62, 75, 76, 77, 90, 91, 92, 105, 106, 120, 121, 122, 135, 136, 150], "bbox": [0.0033723653395784543, 0.006750000000000001, 0.20929742388758782, 0.4482031249999999], "iscrowd": 0, "area": 11406.050050000003, "rle": {"size": [640, 427], "counts": "`f02Tc0j0H710TAnNj:Q1REWOk:i0PE_Om:`0PEGm:9REKk:4TE2b9\\OcCb0j26`9XOgC`0i2<]MhNo:<kD`0h2`0]MdNm:<oD?g2f0ZMaNm::SE>e2k0YM^Nl::VE<e2P1XMZNk::YE;c2U1WMXNi:8\\E<c1VObNn2]:`MiDQ1U2gNaNl2^:QOPGYN]Ni2a:nNQG]NZNg2d:kNRGcNUNc2j:jNoFgNSN_2P;jNkFlNoM[2W;jNhFY2Z9gMcF[2^9eM`F[2b9eM\\F[2e9fMYFZ2i9gMTFZ2m9iMnEW2T:kMgEV2[:TNYEl1i:bNgD^1[;oNVDR1k;`32N2N2O1nMPDQLR<n3UDjKm;T4[DdKg;Z4\\DcKf;[4\\DdKe;Z4\\DeKf;Y4\\DeKf;Y4[DfKg;X4[DgKf;W4[DhKg;V4[DhKg;W4YDhKi;V4YDhKi;CYCn3o0_Li;@^Cl3k0bLi;A`Ch3h0fLj;BaCd3f0hLk;CcC`3c0mLk;CeC[3b0PMk;DgCX3>SMm;EhCS3Z>lLjAo2X>QMlAj2U>UMoAf2R>[MPBa2R>^MoAa2R>_MnA_2T>`MlAZ2[>fMeAT2a>kM_Ao1h>QNXAi1n>VNRAh1P?YNPAd1S?[Nm@b1W?^Ni@^1[?aNe@]1^?cNb@l0cN\\On`0G`@k0gNYOk`0L]@j0lNUOj`00[@i0S`0XOl_Og0V`0XOk_Of0W`0ZOh_Od0[`0[Of_Ob0]`0^Ob_O`0Qb0L4M2N3M2NaTc6"}, "label": "a tall plant in the kitchen , near to table"}]}
{"id": 170139, "image": "COCO_train2014_000000170139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant in a white pot\"."}], "task": "refering", "answer_template": "The \"plant in a white pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 15, 16, 17, 30, 31, 32, 46, 47, 61, 62, 75, 76, 77, 90, 91, 92, 105, 106, 120, 121, 122, 135, 136, 150], "bbox": [0.0033723653395784543, 0.006750000000000001, 0.20929742388758782, 0.4482031249999999], "iscrowd": 0, "area": 11406.050050000003, "rle": {"size": [640, 427], "counts": "`f02Tc0j0H710TAnNj:Q1REWOk:i0PE_Om:`0PEGm:9REKk:4TE2b9\\OcCb0j26`9XOgC`0i2<]MhNo:<kD`0h2`0]MdNm:<oD?g2f0ZMaNm::SE>e2k0YM^Nl::VE<e2P1XMZNk::YE;c2U1WMXNi:8\\E<c1VObNn2]:`MiDQ1U2gNaNl2^:QOPGYN]Ni2a:nNQG]NZNg2d:kNRGcNUNc2j:jNoFgNSN_2P;jNkFlNoM[2W;jNhFY2Z9gMcF[2^9eM`F[2b9eM\\F[2e9fMYFZ2i9gMTFZ2m9iMnEW2T:kMgEV2[:TNYEl1i:bNgD^1[;oNVDR1k;`32N2N2O1nMPDQLR<n3UDjKm;T4[DdKg;Z4\\DcKf;[4\\DdKe;Z4\\DeKf;Y4\\DeKf;Y4[DfKg;X4[DgKf;W4[DhKg;V4[DhKg;W4YDhKi;V4YDhKi;CYCn3o0_Li;@^Cl3k0bLi;A`Ch3h0fLj;BaCd3f0hLk;CcC`3c0mLk;CeC[3b0PMk;DgCX3>SMm;EhCS3Z>lLjAo2X>QMlAj2U>UMoAf2R>[MPBa2R>^MoAa2R>_MnA_2T>`MlAZ2[>fMeAT2a>kM_Ao1h>QNXAi1n>VNRAh1P?YNPAd1S?[Nm@b1W?^Ni@^1[?aNe@]1^?cNb@l0cN\\On`0G`@k0gNYOk`0L]@j0lNUOj`00[@i0S`0XOl_Og0V`0XOk_Of0W`0ZOh_Od0[`0[Of_Ob0]`0^Ob_O`0Qb0L4M2N3M2NaTc6"}, "label": "plant in a white pot"}]}
{"id": 227484, "image": "COCO_train2014_000000227484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black pickup van\"."}], "task": "refering", "answer_template": "The \"a black pickup van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 245, 246, 247, 257, 258, 259, 260, 261, 262, 280, 281, 282], "bbox": [0.16679738562091503, 0.34522875816993465, 0.9459640522875818, 0.5740849673202615], "iscrowd": 0, "area": 28488.680900000003, "rle": {"size": [612, 612], "counts": "PXm1?]b09G8J7N101N1O2O0O1O00100O1O100O00100O1O1000O01000000000O10000000000O01000000000O10000000000O10O10000000O10000000000O1000O100000O10000000000O1000O100000O10000000000O10000O00100O100O100O100O1O100O10O0100O10O0000001O00001O000O2O00001O00001O00001N1000O1O1O100O11O001O1N2O1O001O1000O0100000O1000000O10000000000000000000000000000000O1000000000001O000000000000000000000000000000000O1000001O0000000000000000000000000000001O00001O0O101O00001O00001O001O00001O00001O00001O001O000O101O00001O00001O00001O00001O00001O00001O00001O0O10001O0000001O02N2N102M2N3M2O2M2N3M2O2M2N3N1N3M0001O0000001O000O101O00001O0000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O00001O00000O2O0000001O0000001O0000001O00001O00002N3M3M3M2N3M3M3M1O1O001N10001O001O001O001O001O00000000001O0000000000000000000000000O10001O00000000000000000000000000O1N2N2O1N2N2N2O1N2M3F:O1O100O1O100O100O1O10000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000000003M6J6I7J6J6J6J5K1O1O1M3M3M2N3Mo]c0"}, "label": "a black pickup van"}]}
{"id": 227484, "image": "COCO_train2014_000000227484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black truck\"."}], "task": "refering", "answer_template": "The \"a black truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 223, 224, 225, 226, 227, 232, 233, 234, 235, 236, 237, 238, 239, 240, 245, 246, 247, 257, 258, 259, 260, 261, 262, 280, 281, 282], "bbox": [0.16679738562091503, 0.34522875816993465, 0.9459640522875818, 0.5740849673202615], "iscrowd": 0, "area": 28488.680900000003, "rle": {"size": [612, 612], "counts": "PXm1?]b09G8J7N101N1O2O0O1O00100O1O100O00100O1O1000O01000000000O10000000000O01000000000O10000000000O10O10000000O10000000000O1000O100000O10000000000O1000O100000O10000000000O10000O00100O100O100O100O1O100O10O0100O10O0000001O00001O000O2O00001O00001O00001N1000O1O1O100O11O001O1N2O1O001O1000O0100000O1000000O10000000000000000000000000000000O1000000000001O000000000000000000000000000000000O1000001O0000000000000000000000000000001O00001O0O101O00001O00001O001O00001O00001O00001O001O000O101O00001O00001O00001O00001O00001O00001O00001O0O10001O0000001O02N2N102M2N3M2O2M2N3M2O2M2N3N1N3M0001O0000001O000O101O00001O0000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O00001O00000O2O0000001O0000001O0000001O00001O00002N3M3M3M2N3M3M3M1O1O001N10001O001O001O001O001O00000000001O0000000000000000000000000O10001O00000000000000000000000000O1N2N2O1N2N2N2O1N2M3F:O1O100O1O100O100O1O10000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000000003M6J6I7J6J6J6J5K1O1O1M3M3M2N3Mo]c0"}, "label": "a black truck"}]}
{"id": 227484, "image": "COCO_train2014_000000227484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red sports car with black convertible top\"."}], "task": "refering", "answer_template": "The \"red sports car with black convertible top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 209, 226, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.11730392156862746, 0.4388725490196078, 0.6953921568627451, 0.6296078431372549], "iscrowd": 0, "area": 26227.608350000002, "rle": {"size": [612, 612], "counts": "g[[13ob04L5K4L4L5K4M3M2N3M2N1O10000O10001N10000O10001N101N2O0O2O1N102N1O2N1N3N1O1O0O2O001O001N1000000O100000000O1000000O1000O1000O100000O01000O100O01000O10O010000O010O100O100O10O010000O100O1000O1000000000000000O0100000000O1O100O100O1O100O010O1O100O100O1O100O100O1O100O010O100O1O100O100O1O100O100O1O010O100O1O100O100O1O100O100O0100000000000O1000000000000000O01000000000000000000O10000000O10000000000000000000000000000O100000000000000000001O00000O1000000000000000000000000003M3M1O1O001O1O00100O010O10O01000O001O001O001O001O001O001O001O001N2O0O2O001N101O0O2O001N1O2N1O2N1O2N1O1O1O1O101N100O100O1O100O100O2O0O10000O100000000O2O0000000O1000000O101O00000O1000000O10001O000O1000000O100000001N2O1N2N2N9G<D2N2N1O3M4L3M3MQ^_3"}, "label": "red sports car with black convertible top"}]}
{"id": 227484, "image": "COCO_train2014_000000227484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red sports car parked in the parking lot\"."}], "task": "refering", "answer_template": "The \"a red sports car parked in the parking lot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 209, 226, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.11730392156862746, 0.4388725490196078, 0.6953921568627451, 0.6296078431372549], "iscrowd": 0, "area": 26227.608350000002, "rle": {"size": [612, 612], "counts": "g[[13ob04L5K4L4L5K4M3M2N3M2N1O10000O10001N10000O10001N101N2O0O2O1N102N1O2N1N3N1O1O0O2O001O001N1000000O100000000O1000000O1000O1000O100000O01000O100O01000O10O010000O010O100O100O10O010000O100O1000O1000000000000000O0100000000O1O100O100O1O100O010O1O100O100O1O100O100O1O100O010O100O1O100O100O1O100O100O1O010O100O1O100O100O1O100O100O0100000000000O1000000000000000O01000000000000000000O10000000O10000000000000000000000000000O100000000000000000001O00000O1000000000000000000000000003M3M1O1O001O1O00100O010O10O01000O001O001O001O001O001O001O001O001N2O0O2O001N101O0O2O001N1O2N1O2N1O2N1O1O1O1O101N100O100O1O100O100O2O0O10000O100000000O2O0000000O1000000O101O00000O1000000O10001O000O1000000O100000001N2O1N2N2N9G<D2N2N1O3M4L3M3MQ^_3"}, "label": "a red sports car parked in the parking lot"}]}
{"id": 374946, "image": "COCO_train2014_000000374946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"near zebra\"."}], "task": "refering", "answer_template": "The \"near zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 196, 197, 214, 216, 217, 218, 219, 220], "bbox": [0.31784375, 0.3310208333333333, 0.6034375, 0.5955416666666667], "iscrowd": 0, "area": 13914.815400000003, "rle": {"size": [480, 640], "counts": "Xjo25l>4L4L3cBEm;h100O1O1O0O010N2M3N2N2M3N2N100O01O10O01O10O01O00100_O]CeNc<\\1cC]N^<b1iCXNW<h1a0O001O001N101O001O00001O0101O1O0]DeNa9\\1]FdNd9]1ZFdNf9\\1YFeNg9\\1WFeNj9[1TFfNl9Z1SFgNm9Z1\\E[Oe:`20000000N2M3L4M3K5I7I7J6L3M4L4O13M2N2N3M2N2aD_M[:d2`E^M`:g2YE[Mg:j2SEWMm:n2lDTMT;c1gD_OOPOZ;a1lDFT;:QEAo:`0UE[Ok:e0YEWOg:i0^EROb:n0cEmN]:S1hEfNY:\\1jE]NY:c1jEVNZ:j1^100001O000000001O0000001O0000000000O1000000O10000O1000000O1000bDUNg9k1XFXNf9g1ZF[Ne9e1ZF[Ng9e1XF\\Nh9c1XF^Nh9b1WF_Ni9a1VF_Nk9a1TF`Nl9_1TFaNm9_1RFbNn9`1mEbNT:c1eE^N\\:g1\\E[Nf:i1REZNn:j1lDWNU;n1cDTN^;g20O1O1N2M3N2M3M3N2M3M3NiD\\MW:`2kEaMU:^2kEdMU:[2iEgMW:Y2hEiMX:V2fElM[:S2dEnM\\:U2`EmM`:U2[EmMe:V2WElMi:U310O01O010I7J5O2O0K6^Ob0J56K5J6001K5B=Ce0[Of0ZORkf3"}, "label": "near zebra"}]}
{"id": 374946, "image": "COCO_train2014_000000374946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra was eating grass\"."}], "task": "refering", "answer_template": "The \"zebra was eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 196, 197, 214, 216, 217, 218, 219, 220], "bbox": [0.31784375, 0.3310208333333333, 0.6034375, 0.5955416666666667], "iscrowd": 0, "area": 13914.815400000003, "rle": {"size": [480, 640], "counts": "Xjo25l>4L4L3cBEm;h100O1O1O0O010N2M3N2N2M3N2N100O01O10O01O10O01O00100_O]CeNc<\\1cC]N^<b1iCXNW<h1a0O001O001N101O001O00001O0101O1O0]DeNa9\\1]FdNd9]1ZFdNf9\\1YFeNg9\\1WFeNj9[1TFfNl9Z1SFgNm9Z1\\E[Oe:`20000000N2M3L4M3K5I7I7J6L3M4L4O13M2N2N3M2N2aD_M[:d2`E^M`:g2YE[Mg:j2SEWMm:n2lDTMT;c1gD_OOPOZ;a1lDFT;:QEAo:`0UE[Ok:e0YEWOg:i0^EROb:n0cEmN]:S1hEfNY:\\1jE]NY:c1jEVNZ:j1^100001O000000001O0000001O0000000000O1000000O10000O1000000O1000bDUNg9k1XFXNf9g1ZF[Ne9e1ZF[Ng9e1XF\\Nh9c1XF^Nh9b1WF_Ni9a1VF_Nk9a1TF`Nl9_1TFaNm9_1RFbNn9`1mEbNT:c1eE^N\\:g1\\E[Nf:i1REZNn:j1lDWNU;n1cDTN^;g20O1O1N2M3N2M3M3N2M3M3NiD\\MW:`2kEaMU:^2kEdMU:[2iEgMW:Y2hEiMX:V2fElM[:S2dEnM\\:U2`EmM`:U2[EmMe:V2WElMi:U310O01O010I7J5O2O0K6^Ob0J56K5J6001K5B=Ce0[Of0ZORkf3"}, "label": "zebra was eating grass"}]}
{"id": 374946, "image": "COCO_train2014_000000374946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 151, 152, 174, 175, 197, 198, 199, 220], "bbox": [0.3705625, 0.2905208333333333, 0.6761874999999999, 0.5612499999999999], "iscrowd": 0, "area": 3830.045100000003, "rle": {"size": [480, 640], "counts": "oW_37a>80000O1000O0100O10000O10000O010O10000O100O1000O0100O1000O010000O11O0000000000000000001N1000000000000000001O00000000000000000O101O0000000000000000001O0000000000000000001O0O10000000000000001O000000000000000001O00010O00001O001O00001O00001O00001O0010O0001O00001O00001O002N1O1O2N1O101N100O2O0O2N103L6ZDfN[9^1YFhNh9^1lEgNV:]1YEnNh:j0hD`NMS1\\;:QEHR;2YECn:7\\EXOQ;d0i1M3N2N1O2N2N2N2N2N1O1N2O1O1O1O1O1O1O1001O1O1O2N1ON2O001O001O0O101O001O0O2O00`XQ3"}, "label": "a zebra behind another zebra"}]}
{"id": 374946, "image": "COCO_train2014_000000374946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra can be seen behind the other one\"."}], "task": "refering", "answer_template": "The \"a zebra can be seen behind the other one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 151, 152, 174, 175, 197, 198, 199, 220], "bbox": [0.3705625, 0.2905208333333333, 0.6761874999999999, 0.5612499999999999], "iscrowd": 0, "area": 3830.045100000003, "rle": {"size": [480, 640], "counts": "oW_37a>80000O1000O0100O10000O10000O010O10000O100O1000O0100O1000O010000O11O0000000000000000001N1000000000000000001O00000000000000000O101O0000000000000000001O0000000000000000001O0O10000000000000001O000000000000000001O00010O00001O001O00001O00001O00001O0010O0001O00001O00001O002N1O1O2N1O101N100O2O0O2N103L6ZDfN[9^1YFhNh9^1lEgNV:]1YEnNh:j0hD`NMS1\\;:QEHR;2YECn:7\\EXOQ;d0i1M3N2N1O2N2N2N2N2N1O1N2O1O1O1O1O1O1O1001O1O1O2N1ON2O001O001O0O101O001O0O2O00`XQ3"}, "label": "a zebra can be seen behind the other one"}]}
{"id": 137378, "image": "COCO_train2014_000000137378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the beige loveseat next to the window\"."}], "task": "refering", "answer_template": "The \"the beige loveseat next to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 250, 251, 273, 274], "bbox": [0.777296875, 0.4177708333333333, 0.9764531250000001, 0.7047083333333333], "iscrowd": 0, "area": 8871.160049999995, "rle": {"size": [480, 640], "counts": "TfY71m>3N1O1N2O1O1N3N1B>J600O2O000O100O101O0O100O101O0O100O10001N101O0O2O0O2O000O2O0O2O001N101N100N3L3M4L3M4O001O000010O01O001O001O01O01O001O1O10O01O1O1O1O100O1SDhMP;Y2nDjMP;V2PEkMo:V2oDkMQ;V2nDkMR;U2lDlMT;U2jDmMU;T2jDlMV;P3O1O010O001O001O010O001O001O00100O001O001OO2F9G9G9G9G9J7J5K5K5K5K5K6J5K5K5K5J6K6J5K`j6"}, "label": "the beige loveseat next to the window"}]}
{"id": 137378, "image": "COCO_train2014_000000137378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sofa on the right with blue pillows\"."}], "task": "refering", "answer_template": "The \"the sofa on the right with blue pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 250, 251, 273, 274], "bbox": [0.777296875, 0.4177708333333333, 0.9764531250000001, 0.7047083333333333], "iscrowd": 0, "area": 8871.160049999995, "rle": {"size": [480, 640], "counts": "TfY71m>3N1O1N2O1O1N3N1B>J600O2O000O100O101O0O100O101O0O100O10001N101O0O2O0O2O000O2O0O2O001N101N100N3L3M4L3M4O001O000010O01O001O001O01O01O001O1O10O01O1O1O1O100O1SDhMP;Y2nDjMP;V2PEkMo:V2oDkMQ;V2nDkMR;U2lDlMT;U2jDmMU;T2jDlMV;P3O1O010O001O001O010O001O001O00100O001O001OO2F9G9G9G9G9J7J5K5K5K5K5K6J5K5K5K5J6K6J5K`j6"}, "label": "the sofa on the right with blue pillows"}]}
{"id": 137378, "image": "COCO_train2014_000000137378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather loveseat\"."}], "task": "refering", "answer_template": "The \"a brown leather loveseat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 262], "bbox": [0.306453125, 0.36343749999999997, 0.772578125, 0.6537708333333333], "iscrowd": 0, "area": 28412.354799999994, "rle": {"size": [480, 640], "counts": "fRl23j><E;E9I000000000O100000000_OPORCP1l<UOQCk0n<^OiBc0V=d0O3L5L4M3M4L3M3M3M3M3M2N1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O00O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O2O000000000O100000000000O010000000000O10000000000000000000000000000@jDVMV;i2lDVMS;k2nDTMR;k2oDUMQ;j2QEUMo:k2RESMo:m2QESMo:m2REQMo:n2SEPMn:P3REnLP;R3QEkLQ;U3=0000000O10O10000000000000000O1000O10000000000000O100000000000O1000O1000000000000000000O10O10000000000000O1000000000O100000O1000000000000000000O011O2N2N2N2N2N2N2M3N2N00001O00001O000O10000000001O0B_C]Na<]1iC_NW<a1mCZNT<f1c000000000000001O9G2N000000001O000000001O00000000001O000O1M3N3L4M4K4M3L5K4M4KUWT2"}, "label": "a brown leather loveseat"}]}
{"id": 137378, "image": "COCO_train2014_000000137378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a maroon leather couch with two striped throw pillows\"."}], "task": "refering", "answer_template": "The \"a maroon leather couch with two striped throw pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 262], "bbox": [0.306453125, 0.36343749999999997, 0.772578125, 0.6537708333333333], "iscrowd": 0, "area": 28412.354799999994, "rle": {"size": [480, 640], "counts": "fRl23j><E;E9I000000000O100000000_OPORCP1l<UOQCk0n<^OiBc0V=d0O3L5L4M3M4L3M3M3M3M3M2N1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O00O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O2O000000000O100000000000O010000000000O10000000000000000000000000000@jDVMV;i2lDVMS;k2nDTMR;k2oDUMQ;j2QEUMo:k2RESMo:m2QESMo:m2REQMo:n2SEPMn:P3REnLP;R3QEkLQ;U3=0000000O10O10000000000000000O1000O10000000000000O100000000000O1000O1000000000000000000O10O10000000000000O1000000000O100000O1000000000000000000O011O2N2N2N2N2N2N2M3N2N00001O00001O000O10000000001O0B_C]Na<]1iC_NW<a1mCZNT<f1c000000000000001O9G2N000000001O000000001O00000000001O000O1M3N3L4M4K4M3L5K4M4KUWT2"}, "label": "a maroon leather couch with two striped throw pillows"}]}
{"id": 137378, "image": "COCO_train2014_000000137378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan coach with 5 pillows\"."}], "task": "refering", "answer_template": "The \"tan coach with 5 pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.11015625, 0.46814583333333337, 0.526453125, 0.9864999999999999], "iscrowd": 0, "area": 47577.24659999999, "rle": {"size": [480, 640], "counts": "W`Q1g0[>i0WOj0SOl0TOl0TOm0SO:F2N3M2N3M2N3N1O2N2N101N1O2O001N101N101N101N101N101N101O1N101N101N101N101N101N101N101N1O2N1O000000001O01O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000010O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O01O0000000000000001O002N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N100O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2O0O2N1O1O2N1O2N1O1O2N1O2N1O1O2N001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O002N2N2N1O2N2N2N2O0O2NQ1oNVc]4"}, "label": "tan coach with 5 pillows"}]}
{"id": 137378, "image": "COCO_train2014_000000137378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sofa closest to the wines\"."}], "task": "refering", "answer_template": "The \"a white sofa closest to the wines\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.11015625, 0.46814583333333337, 0.526453125, 0.9864999999999999], "iscrowd": 0, "area": 47577.24659999999, "rle": {"size": [480, 640], "counts": "W`Q1g0[>i0WOj0SOl0TOl0TOm0SO:F2N3M2N3M2N3N1O2N2N101N1O2O001N101N101N101N101N101N101O1N101N101N101N101N101N101N101N1O2N1O000000001O01O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000010O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O01O0000000000000001O002N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N100O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2O0O2N1O1O2N1O2N1O1O2N1O2N1O1O2N001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O002N2N2N1O2N2N2N2O0O2NQ1oNVc]4"}, "label": "a white sofa closest to the wines"}]}
{"id": 47267, "image": "COCO_train2014_000000047267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in an orange shirt\"."}], "task": "refering", "answer_template": "The \"a woman in an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 46, 47, 48, 59, 60, 61, 72, 73, 74, 85, 86, 87, 98, 99, 100, 110, 111, 112, 113, 123, 124, 125, 126, 137], "bbox": [0.5243466666666666, 0.13932, 0.7640533333333333, 0.58876], "iscrowd": 0, "area": 11908.562249999999, "rle": {"size": [500, 375], "counts": "a]P3:W?:oM^O^Dm0X;VOgDo0R;SOnDm0o:WOQEi0l:YOnDDhNS1X<\\OeDLSOg0U<@]D4^O=T<@SD>H2T<AQD`0LN9POY;c0aD`0LN9SOV;?fD?LN8YOP;<kD>LM9]Ol:8oD>MM8Af:6UE<LM9Eb:2YE<LN8Ea:3[E:LM9E`:4[E:LM9E`:5[E8MN7F_:6\\E7MM8F_:6\\E7NL8Fa9]OiFl0@5MM8E_9KaF>K5MM8F\\9McF<L4ML9FX92dF9N3MM8EV95gF7N2ML8GS97hF501MM7FQ9:kF3O1NK8Fn8>kF12OML8Gj8`0oFO1OML8Fh8d0RGJ2OML8Ff8f0SGJ1OMK9Gf8e0SGI2OML7Gg8f0RGI2OMK8Gh8e0QGJ2ONJ8Hg8e0QGJ2OMK8Gh8e0QGJ2OMK8Gi8e0oFK3MNL7Gi8d0PGK2OMK8Gi8e0oFK3MNK7Hj8d0oFK2OMK7Hk8c0nFL2NOJ7Hk8d0lFL4NMJ8Hl8c0lFM2NNK7Gn8c0jFM4NMJ8Hm8b0kFN2NNJ8Hn8b0iFO3MNJ7HP9a0iFO2NNJ7IP9`0hF02NOI7HQ9a0fF04MNJ7HR9`0fF12NNI8IR9?eF14MNJ7HS9?eF22NNI8HT9>dF33MNJ6IU9=dF32MOJ6IV9<cF42NOI6HW9<cF51MOJ6IW9:cF61NNJ6HY99bF81LOJ6HY99bF80NNJ6HZ98aF91LOJ5I[97aF90NNJ5H^95_F<0LOJ5I^93_F=ONNJ5Ha91]F`0NMOJ5Hb90]F`0NNNJ4He9OZFb0NMOJ5He9MZFc0NNNJ4Gi9LXFe0LNOJ4Bo90SFe0LONJ4\\OV:4mE5JH0i0OJ4WO]:7fE6LHNk0NJ4nNg:>^E5NIKk0OJR;[OWE50IHm0OKS;WOXE60KFm0OIV;VOWE80JDo0OGY;TOXE:NKCP1NEP<@aDIAR1NCS<@aDH_OV1M@Y<]O\\DL_OV1L_Oa=a0^B]Of=a0[B]Og=c0XB[Ol=c0UBZOn=f0QBXOS>f0nAVOV>j0iATO[>j0eASO_>m0ajZ1"}, "label": "a woman in an orange shirt"}]}
{"id": 47267, "image": "COCO_train2014_000000047267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting on a blue chair next to a board\"."}], "task": "refering", "answer_template": "The \"a woman sitting on a blue chair next to a board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 46, 47, 48, 59, 60, 61, 72, 73, 74, 85, 86, 87, 98, 99, 100, 110, 111, 112, 113, 123, 124, 125, 126, 137], "bbox": [0.5243466666666666, 0.13932, 0.7640533333333333, 0.58876], "iscrowd": 0, "area": 11908.562249999999, "rle": {"size": [500, 375], "counts": "a]P3:W?:oM^O^Dm0X;VOgDo0R;SOnDm0o:WOQEi0l:YOnDDhNS1X<\\OeDLSOg0U<@]D4^O=T<@SD>H2T<AQD`0LN9POY;c0aD`0LN9SOV;?fD?LN8YOP;<kD>LM9]Ol:8oD>MM8Af:6UE<LM9Eb:2YE<LN8Ea:3[E:LM9E`:4[E:LM9E`:5[E8MN7F_:6\\E7MM8F_:6\\E7NL8Fa9]OiFl0@5MM8E_9KaF>K5MM8F\\9McF<L4ML9FX92dF9N3MM8EV95gF7N2ML8GS97hF501MM7FQ9:kF3O1NK8Fn8>kF12OML8Gj8`0oFO1OML8Fh8d0RGJ2OML8Ff8f0SGJ1OMK9Gf8e0SGI2OML7Gg8f0RGI2OMK8Gh8e0QGJ2ONJ8Hg8e0QGJ2OMK8Gh8e0QGJ2OMK8Gi8e0oFK3MNL7Gi8d0PGK2OMK8Gi8e0oFK3MNK7Hj8d0oFK2OMK7Hk8c0nFL2NOJ7Hk8d0lFL4NMJ8Hl8c0lFM2NNK7Gn8c0jFM4NMJ8Hm8b0kFN2NNJ8Hn8b0iFO3MNJ7HP9a0iFO2NNJ7IP9`0hF02NOI7HQ9a0fF04MNJ7HR9`0fF12NNI8IR9?eF14MNJ7HS9?eF22NNI8HT9>dF33MNJ6IU9=dF32MOJ6IV9<cF42NOI6HW9<cF51MOJ6IW9:cF61NNJ6HY99bF81LOJ6HY99bF80NNJ6HZ98aF91LOJ5I[97aF90NNJ5H^95_F<0LOJ5I^93_F=ONNJ5Ha91]F`0NMOJ5Hb90]F`0NNNJ4He9OZFb0NMOJ5He9MZFc0NNNJ4Gi9LXFe0LNOJ4Bo90SFe0LONJ4\\OV:4mE5JH0i0OJ4WO]:7fE6LHNk0NJ4nNg:>^E5NIKk0OJR;[OWE50IHm0OKS;WOXE60KFm0OIV;VOWE80JDo0OGY;TOXE:NKCP1NEP<@aDIAR1NCS<@aDH_OV1M@Y<]O\\DL_OV1L_Oa=a0^B]Of=a0[B]Og=c0XB[Ol=c0UBZOn=f0QBXOS>f0nAVOV>j0iATO[>j0eASO_>m0ajZ1"}, "label": "a woman sitting on a blue chair next to a board"}]}
{"id": 514213, "image": "COCO_train2014_000000514213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white sheep grazing behind a black sheep\"."}], "task": "refering", "answer_template": "The \"white sheep grazing behind a black sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 255, 256, 257, 272, 273, 274, 275, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 360], "bbox": [0.0, 0.619765625, 0.2965360824742268, 0.9223281249999999], "iscrowd": 0, "area": 16615.15575, "rle": {"size": [640, 485], "counts": "]<X4h?01N101N1011N2N2N1O7H9H7I4L4M3L3M100O001O1O1O1O001O01O0RN_B]Ma=d2`BZM`=g2bBVM^=k2cBSM]=n2dBPM\\=P3gBnLX=S3iBkLW=V3kBgLU=Z3lBdLT=]3nB`LS=_3oB_LQ=b3QC[LP=e3QCXLP=i3RCTLn<m3c10O010O1O100O00100O100O01000O100000001N2O2N1O100O2N1O^N_@fN_?Y1c@hN[?X1g@gNX?Y1j@gNU?W1n@hNQ?X1PAhNo>X1YA`Ng>_1[AaNd>_1]A`Nc>`1]A`Nc>_1_A`Na>`1`A^Nb>a1^A_Nb>a1_A]Nb>c1_A\\Na>c1`A\\Na>d1`A[N`>e1`AZNa>f1`AYN`>g1aAWN`>i1`AVNb>i1_AUNb>k1_ARNc>n1]AQNd>o1]AoMd>R2\\AkMf>U2ZAjMg>V2[AgMf>Z2[AbMh>]2ZA`Mg>a2ZA\\Mg>e2Z11N2O000O100000O100O10000O100O101N1O100O1O1O6K5J1O1O1O100QOV^OFja08\\^ODea09b^OB_a0<g^O_OZa0>l^O^OUa0?R_O]On`0`0X_O\\Oi`0b0\\1M3M2O2M3M3N2MVdd6"}, "label": "white sheep grazing behind a black sheep"}]}
{"id": 514213, "image": "COCO_train2014_000000514213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep to the left of all the other sheep\"."}], "task": "refering", "answer_template": "The \"a sheep to the left of all the other sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 255, 256, 257, 272, 273, 274, 275, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 360], "bbox": [0.0, 0.619765625, 0.2965360824742268, 0.9223281249999999], "iscrowd": 0, "area": 16615.15575, "rle": {"size": [640, 485], "counts": "]<X4h?01N101N1011N2N2N1O7H9H7I4L4M3L3M100O001O1O1O1O001O01O0RN_B]Ma=d2`BZM`=g2bBVM^=k2cBSM]=n2dBPM\\=P3gBnLX=S3iBkLW=V3kBgLU=Z3lBdLT=]3nB`LS=_3oB_LQ=b3QC[LP=e3QCXLP=i3RCTLn<m3c10O010O1O100O00100O100O01000O100000001N2O2N1O100O2N1O^N_@fN_?Y1c@hN[?X1g@gNX?Y1j@gNU?W1n@hNQ?X1PAhNo>X1YA`Ng>_1[AaNd>_1]A`Nc>`1]A`Nc>_1_A`Na>`1`A^Nb>a1^A_Nb>a1_A]Nb>c1_A\\Na>c1`A\\Na>d1`A[N`>e1`AZNa>f1`AYN`>g1aAWN`>i1`AVNb>i1_AUNb>k1_ARNc>n1]AQNd>o1]AoMd>R2\\AkMf>U2ZAjMg>V2[AgMf>Z2[AbMh>]2ZA`Mg>a2ZA\\Mg>e2Z11N2O000O100000O100O10000O100O101N1O100O1O1O6K5J1O1O1O100QOV^OFja08\\^ODea09b^OB_a0<g^O_OZa0>l^O^OUa0?R_O]On`0`0X_O\\Oi`0b0\\1M3M2O2M3M3N2MVdd6"}, "label": "a sheep to the left of all the other sheep"}]}
{"id": 514213, "image": "COCO_train2014_000000514213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third sheep whose head is not visible\"."}], "task": "refering", "answer_template": "The \"the third sheep whose head is not visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 218, 219, 220, 234, 235, 236, 237, 250, 251, 252, 253, 254, 267, 268, 269, 270, 271, 285, 286, 287, 288, 304, 305], "bbox": [0.7499381443298969, 0.47659375, 1.0, 0.7717031249999999], "iscrowd": 0, "area": 14262.21635, "rle": {"size": [640, 485], "counts": "SmS72lc04K4M4L3M3M4L3L4M4L5K4L4L5K4L4L4L5K4L3M4L3M3M3M3M3M3M2O2O0O2N1O2N1O1O2N1O1O1O1OO100000000001O1O1O1O10O01O1O1O1O1O1O1O001O1O1O1O1O1O100O1O010O1O1O100O1O100O1O1O103L4L8I<C2N2O1N2N2N2O1N2N2O1N2N0100O100O100O100YNlBbLT=]3h1O1O1000000000000O10000O1O1N2[OTG"}, "label": "the third sheep whose head is not visible"}]}
{"id": 514213, "image": "COCO_train2014_000000514213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheep standing between two other white sheep\"."}], "task": "refering", "answer_template": "The \"a white sheep standing between two other white sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 344, 345, 346, 347, 350, 351, 352, 353, 362, 363, 364, 368, 369, 380, 381, 382, 385, 386], "bbox": [0.2501443298969072, 0.6216875, 0.8669072164948454, 1.0], "iscrowd": 0, "area": 46047.72355, "rle": {"size": [640, 485], "counts": "eR\\22lc0d0]O2N2M3N1N3N2N2X^OPO\\`0S1a_OUOX`0l0g_OZOS`0h0j_O@n?c0o_ODk?>S@Fh?=U@Gh?:V@Jg?8V@Lf?7X@Le?6X@Ne?5W@Oe?4Y@Od?3Y@0d?Y2M3M3L4M3M3L4M3L4M4L2M201N2O1N101N2O1O0O2O1N2O0O1000O0100O100O10O10O010O012M4M3M3L4M2M4M3L4M3M3L3N1N100O10000O100000001O00000O100000000000000O1000fN_CVK`<h4dCVK\\<h4hCVKX<h4lCVKT<h4PDVKP<e4WDYKh;a4aD]K_;]4iDaKW;X4REfKn:T4ZEjKf:o3cEoK]:j3lETLT:h3RFVLn9k3SFSLl9o3UFPLj9P4YFmKg9T4ZFjKf9W4[FgKe9Y4]FeKc9\\4^FbKb9_4o21N101O1O001O0000000000O10000000O10000000O100000000000000O100000000000000O10001O000000001N100000001O000O10001O00000O17Je0ZO7I001O010O001O010O001O0010O01O001O010O001O000001O000000001O04L6J6J6K3L00001O001O00001O010O001O00001O0010O00100O10O01O1L4L3N3L4M2M4M3H8F9H9F:G8G:O3N1O1O1O1O2N1O1N2N3M2N2N2N3M2N2N2N2M2O1O00001N10001O0L4L4M4K4bNQ_OGSa09[1N10001N100O101N100O2O000O2O3L4M3L4MhTX1"}, "label": "a white sheep standing between two other white sheep"}]}
{"id": 514213, "image": "COCO_train2014_000000514213.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white lamb in between two other white lambs\"."}], "task": "refering", "answer_template": "The \"the white lamb in between two other white lambs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 344, 345, 346, 347, 350, 351, 352, 353, 362, 363, 364, 368, 369, 380, 381, 382, 385, 386], "bbox": [0.2501443298969072, 0.6216875, 0.8669072164948454, 1.0], "iscrowd": 0, "area": 46047.72355, "rle": {"size": [640, 485], "counts": "eR\\22lc0d0]O2N2M3N1N3N2N2X^OPO\\`0S1a_OUOX`0l0g_OZOS`0h0j_O@n?c0o_ODk?>S@Fh?=U@Gh?:V@Jg?8V@Lf?7X@Le?6X@Ne?5W@Oe?4Y@Od?3Y@0d?Y2M3M3L4M3M3L4M3L4M4L2M201N2O1N101N2O1O0O2O1N2O0O1000O0100O100O10O10O010O012M4M3M3L4M2M4M3L4M3M3L3N1N100O10000O100000001O00000O100000000000000O1000fN_CVK`<h4dCVK\\<h4hCVKX<h4lCVKT<h4PDVKP<e4WDYKh;a4aD]K_;]4iDaKW;X4REfKn:T4ZEjKf:o3cEoK]:j3lETLT:h3RFVLn9k3SFSLl9o3UFPLj9P4YFmKg9T4ZFjKf9W4[FgKe9Y4]FeKc9\\4^FbKb9_4o21N101O1O001O0000000000O10000000O10000000O100000000000000O100000000000000O10001O000000001N100000001O000O10001O00000O17Je0ZO7I001O010O001O010O001O0010O01O001O010O001O000001O000000001O04L6J6J6K3L00001O001O00001O010O001O00001O0010O00100O10O01O1L4L3N3L4M2M4M3H8F9H9F:G8G:O3N1O1O1O1O2N1O1N2N3M2N2N2N3M2N2N2N2M2O1O00001N10001O0L4L4M4K4bNQ_OGSa09[1N10001N100O101N100O2O000O2O3L4M3L4MhTX1"}, "label": "the white lamb in between two other white lambs"}]}
{"id": 227490, "image": "COCO_train2014_000000227490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train ' s from bnsf\"."}], "task": "refering", "answer_template": "The \"the train ' s from bnsf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.0, 0.17956268221574345, 1.0, 0.7265889212827988], "iscrowd": 0, "area": 91831.53524999997, "rle": {"size": [343, 640], "counts": "U65W:;E;J6N2N2N2N3M2N2M3N2N2N2N2M300001O0000010O00003M9G3M00oNhNcHX1\\7VOWHj0i7DiG<W8R100000001O0000000O1000000000000000001N10O1000\\Od0N2N2N1O2O1O1O1O1O001O100O1M3L31000001O000O100000M30F9010000000000000000000000000000000000000O100000001O000000000M3000000000000000000O1000001O001O00001L30001O00001O001O0O101O00001O00001O001N101O001O1O001O1OmKdIl3[6TLgIj3Y6VLgIk3Y6TLgIl3Y6TLgIl3Z6SLfIn3Y6RLgIn3Y6RLgIn3Z6QLfIo3Z6QLfIo3Z6QLfIo3[6PLeIP4[6nKgIR4_600000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000000000000XNVJcNj5U1^JkNb5U1^JkNb5U1^JkNb5U1^JlNa5T1_JlNa5T1_JlNa5T1`JkN`5U1`JkN`5U1`JkN`5U1`JkN`5U1`JWNT6i1]100000000000O1001O]HkMh6Q2XIgNP6Y1PJROd5o0\\JQOd5o0\\JROc5n0]JROc5n0]JROc5n0]JRO]5T1cJlN]5T1cJlN]5a1VJ_Nj5o1hIRNW6W30000000001Od0\\O000O10000000000000000000001O000000000000000000000000000O10000000001O000000000000000000000000000000000000001N100000000000000J6O1000000000000001O00000000000iJiJn4W5QKmJl4S5TKmJm4R5SKnJm4U5oJlJQ5]50000000LnJcJR5]5400000000000000000000000000000000000000000001OhJdJU5_5000000000000000O10000000000000000000000G90000000000000001O0000000000000000000000000000000000000000000000000000000000001O00000000000000_JRKY5V50000000000000000000O1000000000000K5L4000001O000000000000000000000000000O1000000000000000000000000000000000001O00000000000000000008H0000000000000000000000000000000000F:0000000001O0000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000001O000000001]JSKZ5m4aJXK_5P500000000000L4000O10000000000_JoJ\\5Q5bJXKW5Y5XKeJj3]5SLPKa3Q5]L]KU3e4jLfKk2[4SMfKm2\\4PMeKP3\\4oLdKQ3]4mLdKS3P600000000O10000000000000000O100000000000001O4L4K6K4L5K4LcJ"}, "label": "the train ' s from bnsf"}]}
{"id": 227490, "image": "COCO_train2014_000000227490.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange and black train with a bnsf logo sitting on some train tracks\"."}], "task": "refering", "answer_template": "The \"an orange and black train with a bnsf logo sitting on some train tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.0, 0.17956268221574345, 1.0, 0.7265889212827988], "iscrowd": 0, "area": 91831.53524999997, "rle": {"size": [343, 640], "counts": "U65W:;E;J6N2N2N2N3M2N2M3N2N2N2N2M300001O0000010O00003M9G3M00oNhNcHX1\\7VOWHj0i7DiG<W8R100000001O0000000O1000000000000000001N10O1000\\Od0N2N2N1O2O1O1O1O1O001O100O1M3L31000001O000O100000M30F9010000000000000000000000000000000000000O100000001O000000000M3000000000000000000O1000001O001O00001L30001O00001O001O0O101O00001O00001O001N101O001O1O001O1OmKdIl3[6TLgIj3Y6VLgIk3Y6TLgIl3Y6TLgIl3Z6SLfIn3Y6RLgIn3Y6RLgIn3Z6QLfIo3Z6QLfIo3Z6QLfIo3[6PLeIP4[6nKgIR4_600000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000000000000XNVJcNj5U1^JkNb5U1^JkNb5U1^JkNb5U1^JlNa5T1_JlNa5T1_JlNa5T1`JkN`5U1`JkN`5U1`JkN`5U1`JkN`5U1`JWNT6i1]100000000000O1001O]HkMh6Q2XIgNP6Y1PJROd5o0\\JQOd5o0\\JROc5n0]JROc5n0]JROc5n0]JRO]5T1cJlN]5T1cJlN]5a1VJ_Nj5o1hIRNW6W30000000001Od0\\O000O10000000000000000000001O000000000000000000000000000O10000000001O000000000000000000000000000000000000001N100000000000000J6O1000000000000001O00000000000iJiJn4W5QKmJl4S5TKmJm4R5SKnJm4U5oJlJQ5]50000000LnJcJR5]5400000000000000000000000000000000000000000001OhJdJU5_5000000000000000O10000000000000000000000G90000000000000001O0000000000000000000000000000000000000000000000000000000000001O00000000000000_JRKY5V50000000000000000000O1000000000000K5L4000001O000000000000000000000000000O1000000000000000000000000000000000001O00000000000000000008H0000000000000000000000000000000000F:0000000001O0000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000001O000000001]JSKZ5m4aJXK_5P500000000000L4000O10000000000_JoJ\\5Q5bJXKW5Y5XKeJj3]5SLPKa3Q5]L]KU3e4jLfKk2[4SMfKm2\\4PMeKP3\\4oLdKQ3]4mLdKS3P600000000O10000000000000000O100000000000001O4L4K6K4L5K4LcJ"}, "label": "an orange and black train with a bnsf logo sitting on some train tracks"}]}
{"id": 325806, "image": "COCO_train2014_000000325806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue chair\"."}], "task": "refering", "answer_template": "The \"a blue chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 129, 130, 131, 132, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 249, 250, 251], "bbox": [0.5964166666666666, 0.264515625, 1.0, 0.6473125], "iscrowd": 0, "area": 25773.445899999995, "rle": {"size": [640, 480], "counts": "\\gc52nc0001N2O001SKa0VEA]:m0aESOR:\\1_BjN\\2Kn:c1dBcN\\2Lo:c1cBcN\\2JQ;e1aBbN]2JQ;f1_BbN^2JQ;g1_BaN^2HS;i1]B`N^2IT;i1\\B`N^2HU;j1[B_N_2GV;h2gDZMX;g2gDZMX;g2gDYMY;i2dDYM[;h2dDYM[;i2cDWM];j2aDXM^;i2aDXM];j2bDVM^;j2aDXM^;h2bDYM];h2bDXM^;h2aDZM^;g2aDZM^;i2_DWMa;Q3VDQMi;V3PDkLo;]3iCcLW<d3aC^L^<i3[CXLd<h3\\CXLc<i3\\CYLc<h3\\CXLd<h3[CZLd<f3\\C[Lc<f3\\CZLd<f3[C\\Ld<d3\\C]Lc<d3\\C\\Ld<d3[C^Ld<b3\\C_Lc<b3\\C^Ld<b3[C`Lc<a3]C`Lb<a3]C_Lc<a3\\CaLc<_3]CbLb<^3^CbLb<_3\\CcLc<]3]CdLb<\\3^CdLb<]3\\CeLc<[3]CfLb<Z3^CfLb<[3\\CgLb<Z3^CgLa<Y3_CgLa<Z3]ChLb<X3^CiLa<[501O1O001O1O001N10000000000000000000O10O10000001O3M3J6J_1aN6J6J7I6J7I6J7J5J6J5K3M3M3O10000O10000O10000O2O1O0O2O001N101O0O2O1O0O2O0O2O00U_OQMf`0m270U_OSMd`0l2]_OTMc`0j2^_OWMa`0i2__OWMb`0h2__OXM``0g2a_OZM_`0e2b_O[M]`0e2c_O[M^`0c2d_O]M[`0c2e_O^M[`0a2f_O^M[`0`2f_OaMY`0_2h_OaMX`0^2g_OdMX`0[2i_OeMX`0Z2h_OgMW`0Y2h_OiMX`0U2i_OlMV`0T2i_OmMX`0Q2i_OPNV`0P2j_OoMX`0P2g_OPN[`0n1f_ORNZ`0n1f_OQN\\`0n1c_ORN^`0m1c_ORN_`0m1a_ORN``0n1__ORNc`0l1\\_OVNd`0j1Z_OWNh`0h1U_OZNm`0d1R_O]No`0c1o^O^NSa0a1j^OaNWa0^1h^OdNYa0[1e^OfN\\a0n12N1L5K5K4L5K4L5L3L7K5L4L4L4K5L2N2N3L3N2N3M2NTF"}, "label": "a blue chair"}]}
{"id": 325806, "image": "COCO_train2014_000000325806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue fabric chair with white frame\"."}], "task": "refering", "answer_template": "The \"blue fabric chair with white frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 129, 130, 131, 132, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 249, 250, 251], "bbox": [0.5964166666666666, 0.264515625, 1.0, 0.6473125], "iscrowd": 0, "area": 25773.445899999995, "rle": {"size": [640, 480], "counts": "\\gc52nc0001N2O001SKa0VEA]:m0aESOR:\\1_BjN\\2Kn:c1dBcN\\2Lo:c1cBcN\\2JQ;e1aBbN]2JQ;f1_BbN^2JQ;g1_BaN^2HS;i1]B`N^2IT;i1\\B`N^2HU;j1[B_N_2GV;h2gDZMX;g2gDZMX;g2gDYMY;i2dDYM[;h2dDYM[;i2cDWM];j2aDXM^;i2aDXM];j2bDVM^;j2aDXM^;h2bDYM];h2bDXM^;h2aDZM^;g2aDZM^;i2_DWMa;Q3VDQMi;V3PDkLo;]3iCcLW<d3aC^L^<i3[CXLd<h3\\CXLc<i3\\CYLc<h3\\CXLd<h3[CZLd<f3\\C[Lc<f3\\CZLd<f3[C\\Ld<d3\\C]Lc<d3\\C\\Ld<d3[C^Ld<b3\\C_Lc<b3\\C^Ld<b3[C`Lc<a3]C`Lb<a3]C_Lc<a3\\CaLc<_3]CbLb<^3^CbLb<_3\\CcLc<]3]CdLb<\\3^CdLb<]3\\CeLc<[3]CfLb<Z3^CfLb<[3\\CgLb<Z3^CgLa<Y3_CgLa<Z3]ChLb<X3^CiLa<[501O1O001O1O001N10000000000000000000O10O10000001O3M3J6J_1aN6J6J7I6J7I6J7J5J6J5K3M3M3O10000O10000O10000O2O1O0O2O001N101O0O2O1O0O2O0O2O00U_OQMf`0m270U_OSMd`0l2]_OTMc`0j2^_OWMa`0i2__OWMb`0h2__OXM``0g2a_OZM_`0e2b_O[M]`0e2c_O[M^`0c2d_O]M[`0c2e_O^M[`0a2f_O^M[`0`2f_OaMY`0_2h_OaMX`0^2g_OdMX`0[2i_OeMX`0Z2h_OgMW`0Y2h_OiMX`0U2i_OlMV`0T2i_OmMX`0Q2i_OPNV`0P2j_OoMX`0P2g_OPN[`0n1f_ORNZ`0n1f_OQN\\`0n1c_ORN^`0m1c_ORN_`0m1a_ORN``0n1__ORNc`0l1\\_OVNd`0j1Z_OWNh`0h1U_OZNm`0d1R_O]No`0c1o^O^NSa0a1j^OaNWa0^1h^OdNYa0[1e^OfN\\a0n12N1L5K5K4L5K4L5L3L7K5L4L4L4K5L2N2N3L3N2N3M2NTF"}, "label": "blue fabric chair with white frame"}]}
{"id": 325806, "image": "COCO_train2014_000000325806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer chair in front of a white desk near a computer screen\"."}], "task": "refering", "answer_template": "The \"a white computer chair in front of a white desk near a computer screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 267, 268, 269, 270, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 334, 335, 336, 337, 338, 352, 353, 354, 355, 369, 370, 371, 386, 387, 388], "bbox": [0.6681041666666666, 0.63440625, 0.9519791666666666, 0.98925], "iscrowd": 0, "area": 20079.244199999997, "rle": {"size": [640, 480], "counts": "dSY61nc02O1N4L5K4L5K5L3L5K4L5K5K4M4K4K6F:G8H9G8H9I6L5J6K4K6K4cAcKo<c4kB_KT=f4fB\\KX=i4bBZK]=k4]BWKa=n4ZBTKd=R5UBQKj=c5N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O100O10000O10O010000O100001O001O001O001O001O00001O001O001O001O001O001O000nNoBUKR=i4fC`JZ<^5V1N1N3N2M3N2M3M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3M4L3M3M3M4L3M3M5K5K5K5K6J5K5K6J5K5K6J5K5K5K?AUl="}, "label": "a white computer chair in front of a white desk near a computer screen"}]}
{"id": 325806, "image": "COCO_train2014_000000325806.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white chair\"."}], "task": "refering", "answer_template": "The \"white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 267, 268, 269, 270, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 334, 335, 336, 337, 338, 352, 353, 354, 355, 369, 370, 371, 386, 387, 388], "bbox": [0.6681041666666666, 0.63440625, 0.9519791666666666, 0.98925], "iscrowd": 0, "area": 20079.244199999997, "rle": {"size": [640, 480], "counts": "dSY61nc02O1N4L5K4L5K5L3L5K4L5K5K4M4K4K6F:G8H9G8H9I6L5J6K4K6K4cAcKo<c4kB_KT=f4fB\\KX=i4bBZK]=k4]BWKa=n4ZBTKd=R5UBQKj=c5N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O100O10000O10O010000O100001O001O001O001O001O00001O001O001O001O001O001O000nNoBUKR=i4fC`JZ<^5V1N1N3N2M3N2M3M4M2M3N2M3N3L3N2M3N2M4M2M3N2M3M4L3M3M3M4L3M3M5K5K5K5K6J5K5K6J5K5K6J5K5K5K?AUl="}, "label": "white chair"}]}
{"id": 448689, "image": "COCO_train2014_000000448689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a striped shirt eating a very long hot dog on a bun\"."}], "task": "refering", "answer_template": "The \"a woman with a striped shirt eating a very long hot dog on a bun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 290, 291, 295, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.0, 0.17695833333333333, 0.849, 0.9890208333333333], "iscrowd": 0, "area": 118759.32005000001, "rle": {"size": [480, 640], "counts": "Y>b0\\=R1mNS1XOh0O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100SNkKnHV4Q7QLhHP4W7WLbHj3^7\\L[He3d7bLUH_3j7hLoGY3P8nLiGS3V8TMcGm2\\8l100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1cNeHUK[7f4TIoJm6l4bIiJ_6R5oIdJR6W5]J^Jd5^5Q2K5K5K5M3O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O100O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O100000000000000000000000000000000001O000000000000000000000000000000001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O0kKcLgL^3X3cLgL]3Y3dLfL]3X3eLgL[3Y3fLfL[3X3gLgLY3Y3hLfLY3X3jLfLV3Z3kLeLV3Y3lLfLT3X3oLgLR3W3PMhLQ3V3QMiLo2V3SMiLn2U3TMjLl2U3VMjLk2T3WMkLi2T3YMkLh2S3[MkLe2T3]MkLd2S3^MlLb2S3`MlLa2R3aMmL_2R3cMmL^2Q3dMnL]2P3dMPM\\2P3eMoL]2n2eMQM]2m2dMRM_2k2bMTMa2i2_MWMd2e2^MZMd2d2]M[Mf2b2[M]Mh2_2ZM`Mi2]2WMcMl2Z2UMeMm2X2UMgMn2V2SMiMP3T2PMlMS3Q2nLnMT3o1nLPNU3m1kLSNX3j1iLUNZ3g1hLXNZ3f1gLYN\\3d1dL\\N`3`1aL_Nc3\\1_LcNf3X1ZLhNj3T1WLkNm3o0VLPOn3j0TLVOP4c0TL\\OP4>SLAQ48SLGQ41SLOQ4HUL7o3AVL>n3ZOWLe0m3oN\\LP1j801O1O1O1O1O00DnMkCQ2U<=00001O000000001O00000000001O0000QN^M`Gb2^8aMaG_2]8dMcG[2[8hMdGX2Z8jMfGV2W8nMhGR2V8QNiGo1U8SNkGm1S8UNnGj1P8XNPHh1n7[NQHe1n7\\NRHd1m7]NSHc1m7]NSHc1m7]NTHb1k7_NUHa1k7_NUHa1j7`NVH`1j7`NVH`1j7^NXHb1h7]NYHc1g7[N[He1e7ZN[Hg1e7WN]Hi1c7VN^Hj1b7TN`Hl1`7RNbHn1`7oMaHQ2`7lMaHU2`7iMaHW2`7fMbHZ2_7cMcH]2^7aMcH_2o90O10000O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2L4L4M3L4L4L4L4L4N2O100O100O1O100O100O1O100O100O1O100O100O10_LTFn1l9mM\\FP2e9jMbFT2_9fMiFW2X9cMoF[2R9_MVG^2k8\\M\\Gb2e8XMcGe2]8UMkGh2Y:M1O1O1O1O1O1O1O2O0O100O100O100O101O0O100O100O100O2O0O10O010O10O010O010O01O010O010O010O010O1O010O010O01O000010O01O1O0010O01M2J7J6I601O001O1O001O001O001O1O001O001O001O3M4L3M4L4L3M4L4K4MQS]1"}, "label": "a woman with a striped shirt eating a very long hot dog on a bun"}]}
{"id": 448689, "image": "COCO_train2014_000000448689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman eating a footlong hot dog\"."}], "task": "refering", "answer_template": "The \"a woman eating a footlong hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 290, 291, 295, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.0, 0.17695833333333333, 0.849, 0.9890208333333333], "iscrowd": 0, "area": 118759.32005000001, "rle": {"size": [480, 640], "counts": "Y>b0\\=R1mNS1XOh0O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100SNkKnHV4Q7QLhHP4W7WLbHj3^7\\L[He3d7bLUH_3j7hLoGY3P8nLiGS3V8TMcGm2\\8l100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1cNeHUK[7f4TIoJm6l4bIiJ_6R5oIdJR6W5]J^Jd5^5Q2K5K5K5M3O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O100O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O100000000000000000000000000000000001O000000000000000000000000000000001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O0kKcLgL^3X3cLgL]3Y3dLfL]3X3eLgL[3Y3fLfL[3X3gLgLY3Y3hLfLY3X3jLfLV3Z3kLeLV3Y3lLfLT3X3oLgLR3W3PMhLQ3V3QMiLo2V3SMiLn2U3TMjLl2U3VMjLk2T3WMkLi2T3YMkLh2S3[MkLe2T3]MkLd2S3^MlLb2S3`MlLa2R3aMmL_2R3cMmL^2Q3dMnL]2P3dMPM\\2P3eMoL]2n2eMQM]2m2dMRM_2k2bMTMa2i2_MWMd2e2^MZMd2d2]M[Mf2b2[M]Mh2_2ZM`Mi2]2WMcMl2Z2UMeMm2X2UMgMn2V2SMiMP3T2PMlMS3Q2nLnMT3o1nLPNU3m1kLSNX3j1iLUNZ3g1hLXNZ3f1gLYN\\3d1dL\\N`3`1aL_Nc3\\1_LcNf3X1ZLhNj3T1WLkNm3o0VLPOn3j0TLVOP4c0TL\\OP4>SLAQ48SLGQ41SLOQ4HUL7o3AVL>n3ZOWLe0m3oN\\LP1j801O1O1O1O1O00DnMkCQ2U<=00001O000000001O00000000001O0000QN^M`Gb2^8aMaG_2]8dMcG[2[8hMdGX2Z8jMfGV2W8nMhGR2V8QNiGo1U8SNkGm1S8UNnGj1P8XNPHh1n7[NQHe1n7\\NRHd1m7]NSHc1m7]NSHc1m7]NTHb1k7_NUHa1k7_NUHa1j7`NVH`1j7`NVH`1j7^NXHb1h7]NYHc1g7[N[He1e7ZN[Hg1e7WN]Hi1c7VN^Hj1b7TN`Hl1`7RNbHn1`7oMaHQ2`7lMaHU2`7iMaHW2`7fMbHZ2_7cMcH]2^7aMcH_2o90O10000O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2L4L4M3L4L4L4L4L4N2O100O100O1O100O100O1O100O100O1O100O100O10_LTFn1l9mM\\FP2e9jMbFT2_9fMiFW2X9cMoF[2R9_MVG^2k8\\M\\Gb2e8XMcGe2]8UMkGh2Y:M1O1O1O1O1O1O1O2O0O100O100O100O101O0O100O100O100O2O0O10O010O10O010O010O01O010O010O010O010O1O010O010O01O000010O01O1O0010O01M2J7J6I601O001O1O001O001O001O1O001O001O001O3M4L3M4L4L3M4L4K4MQS]1"}, "label": "a woman eating a footlong hot dog"}]}
{"id": 161970, "image": "COCO_train2014_000000161970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarding boy wearing a blue knit hat\"."}], "task": "refering", "answer_template": "The \"a skateboarding boy wearing a blue knit hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 39, 40, 51, 52, 63, 64, 65, 67, 75, 76, 77, 78, 79, 80, 87, 88, 89, 90, 91, 92, 93, 94, 99, 100, 101, 102, 104, 105, 114, 115, 116, 126, 127, 128, 139, 140], "bbox": [0.26435435435435434, 0.15084, 0.8560660660660661, 0.64034], "iscrowd": 0, "area": 13749.385249999987, "rle": {"size": [500, 333], "counts": "^S[1`0S?3L3N2M4M3N2N2M3QOYOmBi0h<BoBXO0Y1m<JoB;a0Bg:f1PEkNf:X1VEkNh:W1VEjNf:U3N2O001O00L]EXKb:h45POgEdL[:U3PFfLP:Y3UF`Ln9`3T1N2012M3N2McMTDT1_;ROeDo0o:YOSEh0m:UOVEk0j:ROXEo0k:kNXET1f:lN]ER1`:PObEP1]:nNgER1]:eNhE\\1Z:^NhEb1V:_NlEa1Q:`NPFa1m9`NUF`1g9oMnFP2R9nMPGS2o8aM[E2j1]2T;10O101N2N2O1N2N3M100O1O001O1O101N5K1001O0O1O1N3M3N1N3O1O1N2O0O2M2N2OfNPC1n<OTC1j<OXC1f<O\\C2b<MaC2]<NeC2Z<MhC3V<MmC2R<MPD3n;MTD4j;KXD6e;JbD1];NeD2Y;NiD2V;MlD3S;LoD4o:LRE5m:KTE5j:KXE5g:J[E6c:J_E7_:HcE8[:HgE8X:GkE9R:GPF9o9GRF9m9FVF:g9F[F:d9F^F9a9FaF;]9EdF<Y9DjF=S9CnF=Q9CQG;o8DTG;j8HVG9g8KXG5e8L\\G6`8K_G7^8KaG6^8L`G4`8M`G3`8M_G4a8L^G4d8L[G3f8MYG4h8KXG4i8LWG3k8MTG2m8NSGOQ90oFNS93kFLX92hFMZ93fFK\\95cFJ`95^FKd95[FJh95VFKl96QFJR:6kEJW:6hEI[:6cEI`:7^EIe:6ZEIh:7VEIm:6QEJQ;6mDJV;6hDIZ;7dDI^;7`DIc;7[DGh;9VDGm;8QDHQ<8nCGU<8iCG]<6aCHc<8[CFh<;WCBl<?RC_OR=a0lB]OZ=a0eB[O^=g0`BWOc=j0[BTOh=m0?1OO1O1O1O001O1O1O1O1O001O000O2OO1O11N1O1O2N1O000010O01O1O3M6JgWg0"}, "label": "a skateboarding boy wearing a blue knit hat"}]}
{"id": 161970, "image": "COCO_train2014_000000161970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy on skateboard on ramp\"."}], "task": "refering", "answer_template": "The \"guy on skateboard on ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 39, 40, 51, 52, 63, 64, 65, 67, 75, 76, 77, 78, 79, 80, 87, 88, 89, 90, 91, 92, 93, 94, 99, 100, 101, 102, 104, 105, 114, 115, 116, 126, 127, 128, 139, 140], "bbox": [0.26435435435435434, 0.15084, 0.8560660660660661, 0.64034], "iscrowd": 0, "area": 13749.385249999987, "rle": {"size": [500, 333], "counts": "^S[1`0S?3L3N2M4M3N2N2M3QOYOmBi0h<BoBXO0Y1m<JoB;a0Bg:f1PEkNf:X1VEkNh:W1VEjNf:U3N2O001O00L]EXKb:h45POgEdL[:U3PFfLP:Y3UF`Ln9`3T1N2012M3N2McMTDT1_;ROeDo0o:YOSEh0m:UOVEk0j:ROXEo0k:kNXET1f:lN]ER1`:PObEP1]:nNgER1]:eNhE\\1Z:^NhEb1V:_NlEa1Q:`NPFa1m9`NUF`1g9oMnFP2R9nMPGS2o8aM[E2j1]2T;10O101N2N2O1N2N3M100O1O001O1O101N5K1001O0O1O1N3M3N1N3O1O1N2O0O2M2N2OfNPC1n<OTC1j<OXC1f<O\\C2b<MaC2]<NeC2Z<MhC3V<MmC2R<MPD3n;MTD4j;KXD6e;JbD1];NeD2Y;NiD2V;MlD3S;LoD4o:LRE5m:KTE5j:KXE5g:J[E6c:J_E7_:HcE8[:HgE8X:GkE9R:GPF9o9GRF9m9FVF:g9F[F:d9F^F9a9FaF;]9EdF<Y9DjF=S9CnF=Q9CQG;o8DTG;j8HVG9g8KXG5e8L\\G6`8K_G7^8KaG6^8L`G4`8M`G3`8M_G4a8L^G4d8L[G3f8MYG4h8KXG4i8LWG3k8MTG2m8NSGOQ90oFNS93kFLX92hFMZ93fFK\\95cFJ`95^FKd95[FJh95VFKl96QFJR:6kEJW:6hEI[:6cEI`:7^EIe:6ZEIh:7VEIm:6QEJQ;6mDJV;6hDIZ;7dDI^;7`DIc;7[DGh;9VDGm;8QDHQ<8nCGU<8iCG]<6aCHc<8[CFh<;WCBl<?RC_OR=a0lB]OZ=a0eB[O^=g0`BWOc=j0[BTOh=m0?1OO1O1O1O001O1O1O1O1O001O000O2OO1O11N1O1O2N1O000010O01O1O3M6JgWg0"}, "label": "guy on skateboard on ramp"}]}
{"id": 383154, "image": "COCO_train2014_000000383154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right table in which a fat man is taking sun bath\"."}], "task": "refering", "answer_template": "The \"right table in which a fat man is taking sun bath\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 295, 298], "bbox": [0.649046875, 0.5574166666666667, 0.9825624999999999, 0.7671458333333334], "iscrowd": 0, "area": 9923.250450000001, "rle": {"size": [480, 640], "counts": "ijR68h>1N2O1N2O1Ng0SNkNcDU1];oN_DQ1`;SO\\Dn0d;UOYDk0f;WOXDk0g;WOWDi0i;XOVDh0i;ZOUDg0k;ZOTDf0l;ZOSDa0@oN]<`0SDb0R<^OnCb0R<_OlCb0T<^OlCb0T<^OkCd0U<[OkCe0U<\\OjCd0V<\\OiCe0W<[OiCf0V<ZOiCg0W<ZOhCf0X<ZOhCf0Y<YOfCi0Y<WOgCi0Y<XOeCi0[<WOeCi0[<WOdCk0[<UOeCk0[<VOdCj0]<UObCl0^<TObCm0]<SObCn0^<SOaCm0_<TO`Cl0`<k0O01O0001O0000003M3N2M3M3M2N000001N1000000SOPDjNP<V1TDfNm;Y1WDbNj;^1ZD^Ng;`1`DZN`;f1n0O1O001O001N2O001O001O1O001N101O1O001O001N101O1O001O001O1N101O001O1O001O0O2O1O001O001O1N101O001O0000000O10000000000003M7I6J4LO100O100O100O1[O]BLd=4aBG_=:eB@\\=`0b0000000001O0O100O101O0O10000O2O000O10000O2O0N2J6O1010O0000000001O000001O01O000000000010O00000001O00002O5J10O0O101O000O2O001O0O101O000WObBO`=0oBAQ=>j0N100O8H]k4"}, "label": "right table in which a fat man is taking sun bath"}]}
{"id": 358580, "image": "COCO_train2014_000000358580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suv closest to the camera\"."}], "task": "refering", "answer_template": "The \"the suv closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.67246875, 0.46291666666666664, 0.9994375, 0.9820208333333333], "iscrowd": 0, "area": 36447.3201, "rle": {"size": [480, 640], "counts": "Q\\Z68f>7I6I4M2N2M3N2M4N11O00000000001O000001O000001O0000000000010O000000iDWOW8i0XGHh88mF3T9LhF8X9HdF<\\9D`F`0`9@\\Fd0d9]OWFg0i9YOSFk0m9UOoEo0R:POjET1V:lNfEX1Z:hNbE\\1^:dN^E`1b:`NYEe1g:[NUEi1l:U10000001ON2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O100000000000000000000000000O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O100O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O1008H^2bM^2bM^2bM`A"}, "label": "the suv closest to the camera"}]}
{"id": 358580, "image": "COCO_train2014_000000358580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green van in the forefront of the shot , waiting for a sign to turn from stop to slow\"."}], "task": "refering", "answer_template": "The \"a green van in the forefront of the shot , waiting for a sign to turn from stop to slow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.67246875, 0.46291666666666664, 0.9994375, 0.9820208333333333], "iscrowd": 0, "area": 36447.3201, "rle": {"size": [480, 640], "counts": "Q\\Z68f>7I6I4M2N2M3N2M4N11O00000000001O000001O000001O0000000000010O000000iDWOW8i0XGHh88mF3T9LhF8X9HdF<\\9D`F`0`9@\\Fd0d9]OWFg0i9YOSFk0m9UOoEo0R:POjET1V:lNfEX1Z:hNbE\\1^:dN^E`1b:`NYEe1g:[NUEi1l:U10000001ON2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O100000000000000000000000000O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O100O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O1008H^2bM^2bM^2bM`A"}, "label": "a green van in the forefront of the shot , waiting for a sign to turn from stop to slow"}]}
{"id": 358580, "image": "COCO_train2014_000000358580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black truck stopped in traffic with some white bags in the bed\"."}], "task": "refering", "answer_template": "The \"a black truck stopped in traffic with some white bags in the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 270, 289, 290, 291, 312, 313, 314, 315, 316, 335, 336, 337, 338, 359, 360], "bbox": [0.605078125, 0.5141875, 0.847703125, 0.90325], "iscrowd": 0, "area": 14501.203900000006, "rle": {"size": [480, 640], "counts": "_he5\\1Q=f0\\Oc0K2N10O01O001O1O011aD[Mb:i2ZE]M`:f2^E`M\\:c2`EcMZ:a2bEbM\\:`2aEbM^:a2_EaM_:a3N2N2M3N2N2N2N2N2N2N2N2M3N2N2N20O1000000000000YOdFjK\\9V4PG^KP9b4g0SN[F_Ne9_1aF]N_9b1bF^N^9b1cF]N]9b1dF^N\\9b1dF^N\\9a1eF_N[9a1eF_N[9`1fF`NZ9`1gF^NZ9a1gF_NY9a1gF_NY9`1hF`NX9`1hF`NX9_1iFaNW9_1jF`NV9_1kFaNU9_1kFaNU9_1kFaNU9_1kFaNU9_1kFaNU9_1lF`NT9`1lF`NT9`1lF`NT9`1lF`NT9`1lF`NT9`1lF`NT9`1mF_NS9a1mF_NS9a1mF_NS9a1mF_NS9a1mF_NS9a1mF_NS9a1nF^NR9b1nF^NR9b1nF^NR9c1mF]NR9d1nF\\NP9f1PGZNm8i1SGWNk8k1UGUNi8m1WGSNi8m1WGSNi8m1WG^M^9b2bF]M_9c2aF\\M`9d2`F[Ma9e2_FZMb9f2^FYMc9g2]FXMd9h2\\FWMe9j2ZFUMg9k2YFTMh9m2V1O01O00O1N4L7H3N2N2N2N3M2N2N2O1N2O1O1N2O1N3M2N2N2N2N2N2M3N2N2N3N1O1N2O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1Ome]1"}, "label": "a black truck stopped in traffic with some white bags in the bed"}]}
{"id": 358580, "image": "COCO_train2014_000000358580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bog grey coloured car behind a red coloured is stopped by the traffic man\"."}], "task": "refering", "answer_template": "The \"a bog grey coloured car behind a red coloured is stopped by the traffic man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 270, 289, 290, 291, 312, 313, 314, 315, 316, 335, 336, 337, 338, 359, 360], "bbox": [0.605078125, 0.5141875, 0.847703125, 0.90325], "iscrowd": 0, "area": 14501.203900000006, "rle": {"size": [480, 640], "counts": "_he5\\1Q=f0\\Oc0K2N10O01O001O1O011aD[Mb:i2ZE]M`:f2^E`M\\:c2`EcMZ:a2bEbM\\:`2aEbM^:a2_EaM_:a3N2N2M3N2N2N2N2N2N2N2N2M3N2N2N20O1000000000000YOdFjK\\9V4PG^KP9b4g0SN[F_Ne9_1aF]N_9b1bF^N^9b1cF]N]9b1dF^N\\9b1dF^N\\9a1eF_N[9a1eF_N[9`1fF`NZ9`1gF^NZ9a1gF_NY9a1gF_NY9`1hF`NX9`1hF`NX9_1iFaNW9_1jF`NV9_1kFaNU9_1kFaNU9_1kFaNU9_1kFaNU9_1kFaNU9_1lF`NT9`1lF`NT9`1lF`NT9`1lF`NT9`1lF`NT9`1lF`NT9`1mF_NS9a1mF_NS9a1mF_NS9a1mF_NS9a1mF_NS9a1mF_NS9a1nF^NR9b1nF^NR9b1nF^NR9c1mF]NR9d1nF\\NP9f1PGZNm8i1SGWNk8k1UGUNi8m1WGSNi8m1WGSNi8m1WG^M^9b2bF]M_9c2aF\\M`9d2`F[Ma9e2_FZMb9f2^FYMc9g2]FXMd9h2\\FWMe9j2ZFUMg9k2YFTMh9m2V1O01O00O1N4L7H3N2N2N2N3M2N2N2O1N2O1O1N2O1N3M2N2N2N2N2N2M3N2N2N3N1O1N2O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1Ome]1"}, "label": "a bog grey coloured car behind a red coloured is stopped by the traffic man"}]}
{"id": 514230, "image": "COCO_train2014_000000514230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing behind the other zebra\"."}], "task": "refering", "answer_template": "The \"zebra standing behind the other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 43, 44, 45, 46, 47, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 99, 100, 101, 102, 113, 114, 117, 118, 119, 120, 121, 130, 131, 132, 133, 137, 138, 139, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204], "bbox": [0.00102, 0.10474474474474475, 0.738, 1.0], "iscrowd": 0, "area": 31517.56875, "rle": {"size": [333, 500], "counts": "cd06V:100000000000001O0O10000000000000000O10000000000000000O1000000000000O1000000O1000000O1000000O1000000O0100O010O10O0O2O0O2N2O0O2N1O2O0O2N1O2O0O2N2O0O2N1O2O0O2N1O2O0O2O1N2O1N2O1O1N2O1O1N2O1N2O1O3L:G4L1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N3N1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O100000000000010O00000000000001O000000000[OZJcLf5[3bJ_L^5`3jJYLV5f3j0N1O1O2N2N2N1O2N2N2M3N1O2N2N2N2N2N1O2N2N2M3N`K`MU1_2kNcMU1Z2mNgMS1V2oNlMP1R2POQNo0l1SOUNl0j1TOYNk0d1WO]Ni0`1YObNf0\\1ZOfNf0W1\\OkNc0R1^OPOa0o0@SO?j0CWO=f0D]O;a0F@:=GF78KJ44LN4ON31J17OG1<MB5?K^O6e0IYO8h0HUO:m0EPO<R1DlN=V1AhN`0Z1@cNb0_1]O_Nd0b1\\O[Ne0h1ZOUNh0m1VORNj0P2i32N1N3N1N3N1N3N1O2M2O2M3N1O2M2O2M2O2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O000O2O001O001O001O001O1O1O001O1O001O1O1O010000O01000O10O1000O10O10O10O1000O10O10O1000O0100RLUJY3k5fLVJZ3j5dLYJZ3j5cLXJ\\3j5aLWJ_3l5]LVJa3MZLg52^Jc3F`L\\6h32O1N101N2O001N2O0O2O1O0O2O1N101N2N1N3N2M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N3L4M2K6J5K6J6J5K6J5Kc]Z1"}, "label": "zebra standing behind the other zebra"}]}
{"id": 514230, "image": "COCO_train2014_000000514230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second zebra from the front\"."}], "task": "refering", "answer_template": "The \"the second zebra from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 43, 44, 45, 46, 47, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 99, 100, 101, 102, 113, 114, 117, 118, 119, 120, 121, 130, 131, 132, 133, 137, 138, 139, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204], "bbox": [0.00102, 0.10474474474474475, 0.738, 1.0], "iscrowd": 0, "area": 31517.56875, "rle": {"size": [333, 500], "counts": "cd06V:100000000000001O0O10000000000000000O10000000000000000O1000000000000O1000000O1000000O1000000O1000000O0100O010O10O0O2O0O2N2O0O2N1O2O0O2N1O2O0O2N2O0O2N1O2O0O2N1O2O0O2O1N2O1N2O1O1N2O1O1N2O1N2O1O3L:G4L1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N3N1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O100000000000010O00000000000001O000000000[OZJcLf5[3bJ_L^5`3jJYLV5f3j0N1O1O2N2N2N1O2N2N2M3N1O2N2N2N2N2N1O2N2N2M3N`K`MU1_2kNcMU1Z2mNgMS1V2oNlMP1R2POQNo0l1SOUNl0j1TOYNk0d1WO]Ni0`1YObNf0\\1ZOfNf0W1\\OkNc0R1^OPOa0o0@SO?j0CWO=f0D]O;a0F@:=GF78KJ44LN4ON31J17OG1<MB5?K^O6e0IYO8h0HUO:m0EPO<R1DlN=V1AhN`0Z1@cNb0_1]O_Nd0b1\\O[Ne0h1ZOUNh0m1VORNj0P2i32N1N3N1N3N1N3N1O2M2O2M3N1O2M2O2M2O2N1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O000O2O001O001O001O001O1O1O001O1O001O1O1O010000O01000O10O1000O10O10O10O1000O10O10O1000O0100RLUJY3k5fLVJZ3j5dLYJZ3j5cLXJ\\3j5aLWJ_3l5]LVJa3MZLg52^Jc3F`L\\6h32O1N101N2O001N2O0O2O1O0O2O1N101N2N1N3N2M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N3L4M2K6J5K6J6J5K6J5Kc]Z1"}, "label": "the second zebra from the front"}]}
{"id": 514230, "image": "COCO_train2014_000000514230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra closest to camera\"."}], "task": "refering", "answer_template": "The \"zebra closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 21, 22, 23, 24, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 132, 133, 134, 135, 144, 145, 146, 147, 148, 152, 162, 163, 164, 165, 180, 181, 182, 198, 199, 200], "bbox": [0.0, 0.057777777777777775, 0.5505800000000001, 0.9851651651651652], "iscrowd": 0, "area": 46858.46265000001, "rle": {"size": [333, 500], "counts": "W:1\\:0dI\\61O00O010O1O1O1O1B>N2O1N101N2O1N2N2O1N2O1N10100O1O1O1O1O1O100O001O1O1O1O100O1O1O001O1O100O1O1O1O1O010O1O1O1O1O10N101N2N2N2N2N2N1O2O1N2N2N2N2N1O2N2O1N2N2N1O2N2N2N2O1N2N1O2N2N2N2N2O1N1O2N2O1N2O1O1O0O2O1O1N2O1O1O0O2O1O1O1N2O001N2O1O1O1N2O001NXOjJ]LU5d3PKWLo4j3WKPLh4Q4]KjKa4X4i01O001N2O1O1N2O001O2M2O2N2M2O2N2N1N3N2N2N1N3N2N1N300O010O10O0100O010O100O010O10O010O010O0100O0100O0100O010O010O010O10O010O010O05L0O10O0100O010O10O10O10O1000O01000O010000O02O001N2O001O0O2O1O0O2O001N101O1O0O2O001N1N3N2N1N3N1O2M3N1N3N1O2M2O2N2M3N2N2M3N2M3N2N2M3N1O2M3N2M3N2L4K5L4L4L4K5L4L4K5L4L4L4K5LhoX2"}, "label": "zebra closest to camera"}]}
{"id": 514230, "image": "COCO_train2014_000000514230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zebras standing in a grassland\"."}], "task": "refering", "answer_template": "The \"two zebras standing in a grassland\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 21, 22, 23, 24, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 132, 133, 134, 135, 144, 145, 146, 147, 148, 152, 162, 163, 164, 165, 180, 181, 182, 198, 199, 200], "bbox": [0.0, 0.057777777777777775, 0.5505800000000001, 0.9851651651651652], "iscrowd": 0, "area": 46858.46265000001, "rle": {"size": [333, 500], "counts": "W:1\\:0dI\\61O00O010O1O1O1O1B>N2O1N101N2O1N2N2O1N2O1N10100O1O1O1O1O1O100O001O1O1O1O100O1O1O001O1O100O1O1O1O1O010O1O1O1O1O10N101N2N2N2N2N2N1O2O1N2N2N2N2N1O2N2O1N2N2N1O2N2N2N2O1N2N1O2N2N2N2N2O1N1O2N2O1N2O1O1O0O2O1O1N2O1O1O0O2O1O1O1N2O001N2O1O1O1N2O001NXOjJ]LU5d3PKWLo4j3WKPLh4Q4]KjKa4X4i01O001N2O1O1N2O001O2M2O2N2M2O2N2N1N3N2N2N1N3N2N1N300O010O10O0100O010O100O010O10O010O010O0100O0100O0100O010O010O010O10O010O010O05L0O10O0100O010O10O10O10O1000O01000O010000O02O001N2O001O0O2O1O0O2O001N101O1O0O2O001N1N3N2N1N3N1O2M3N1N3N1O2M2O2N2M3N2N2M3N2M3N2N2M3N1O2M3N2M3N2L4K5L4L4L4K5L4L4K5L4L4L4K5LhoX2"}, "label": "two zebras standing in a grassland"}]}
{"id": 522423, "image": "COCO_train2014_000000522423.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with long curly hair on the pros : poster\"."}], "task": "refering", "answer_template": "The \"girl with long curly hair on the pros : poster\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 238, 239, 240], "bbox": [0.0, 0.2749, 0.36624, 0.73722], "iscrowd": 0, "area": 25361.01035, "rle": {"size": [500, 500], "counts": "^9?W>n0SOm0ROn0SOm0[Oe00000O10000O10000O10VKZEf4f:4O10TK[Ei4e:VK[Ek4d:30RK\\El4e:1NSK]Em4d:1O100O1O100O100O1O100O1O100O100O1O100O100O2O0O2O0O101N101N100O2O0O2O0O2O0O1010O01O00001O001O001O00001O2N1O2N101N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2NlNdGQK\\8j4RHnJl7n4bHiJ]7R5QIgJm6U5k1L4K5L4L4L5L4N1O2M2O2NkEkKn8S4gFXLZ9g3[FdLg9Z3TFlLm9R3UFlLl9S3VFkLl9S3UFlLl9S3VFkLl9R3VFmLl9R3UFmLk9R3WFlLk9S3UFlLm9S3TFkLm9U3SFjLo9U3RFiLP:V3PFjLR:T3Y1N2N1N3N2N1OeFmLW6Q3^IYMf6d2]HWM_N`0W9V2ZH\\Nj7`1QHXM]NX1e9]1`GiMjNk0j9W1[GPNkNi0m9T1XGSNkNi0Q:P1TGWNkNi0T:m0QGZNkNh0Y:j0lF^NkNh0\\:g0iFaNlNg0_:d0eFeNlNg0b:a0bFhNlNg0f:=^FkNnNg0g::\\FoNmNg0k:6XFSOnNf0m:4UFVOnNe0R;1PFZOoNd0S;0nE\\OoNd0U;NlE^OPOc0W;LiEAPOc0Y;JgECQOb0Z;IeEEQOb0];EdEHoNc0_;CdEImNb0b;CbEKkNb0f;@`ENjNa0h;_O`ENhNc0j;]O_E0gNb0m;[O^E2dNc0P<YO]E0gNf0n;XO\\ENiNj0n;TOlEk0`<O001O00001O001N10001O0O101O000O;FT^j4"}, "label": "girl with long curly hair on the pros : poster"}]}
{"id": 522423, "image": "COCO_train2014_000000522423.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on a billboard\"."}], "task": "refering", "answer_template": "The \"a woman on a billboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 238, 239, 240], "bbox": [0.0, 0.2749, 0.36624, 0.73722], "iscrowd": 0, "area": 25361.01035, "rle": {"size": [500, 500], "counts": "^9?W>n0SOm0ROn0SOm0[Oe00000O10000O10000O10VKZEf4f:4O10TK[Ei4e:VK[Ek4d:30RK\\El4e:1NSK]Em4d:1O100O1O100O100O1O100O1O100O100O1O100O100O2O0O2O0O101N101N100O2O0O2O0O2O0O1010O01O00001O001O001O00001O2N1O2N101N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2NlNdGQK\\8j4RHnJl7n4bHiJ]7R5QIgJm6U5k1L4K5L4L4L5L4N1O2M2O2NkEkKn8S4gFXLZ9g3[FdLg9Z3TFlLm9R3UFlLl9S3VFkLl9S3UFlLl9S3VFkLl9R3VFmLl9R3UFmLk9R3WFlLk9S3UFlLm9S3TFkLm9U3SFjLo9U3RFiLP:V3PFjLR:T3Y1N2N1N3N2N1OeFmLW6Q3^IYMf6d2]HWM_N`0W9V2ZH\\Nj7`1QHXM]NX1e9]1`GiMjNk0j9W1[GPNkNi0m9T1XGSNkNi0Q:P1TGWNkNi0T:m0QGZNkNh0Y:j0lF^NkNh0\\:g0iFaNlNg0_:d0eFeNlNg0b:a0bFhNlNg0f:=^FkNnNg0g::\\FoNmNg0k:6XFSOnNf0m:4UFVOnNe0R;1PFZOoNd0S;0nE\\OoNd0U;NlE^OPOc0W;LiEAPOc0Y;JgECQOb0Z;IeEEQOb0];EdEHoNc0_;CdEImNb0b;CbEKkNb0f;@`ENjNa0h;_O`ENhNc0j;]O_E0gNb0m;[O^E2dNc0P<YO]E0gNf0n;XO\\ENiNj0n;TOlEk0`<O001O00001O001N10001O0O101O000O;FT^j4"}, "label": "a woman on a billboard"}]}
{"id": 522423, "image": "COCO_train2014_000000522423.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy riding a bike\"."}], "task": "refering", "answer_template": "The \"a boy riding a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 43, 44, 45, 46, 61, 62, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 223, 224, 225, 226, 227, 228, 229, 230, 231, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 295, 296, 297, 298, 299, 300, 313, 314, 315, 316, 317], "bbox": [0.27294, 0.09437999999999999, 0.868, 0.99818], "iscrowd": 0, "area": 79533.46545, "rle": {"size": [500, 500], "counts": "_US22b?1N2N2N2O1N1O2N2O1N1O1O2O0O10001O0O101O00001N100000001TDVOl8j0RGZOk8g0SG]Ok8c0RGBl8>RGEm8;PGJm88PGLn84oF1o8OPG4n8LQG7m8IRG:k8GUG:j8GTG<j8DUG?i8AWGa0g8_OXGd0e8]O[Ge0c8[O\\Gh0b8XO]Gk0a8UO_Gm0_8SO`GP1]8QObGS1[8mNeGU1Y8kNfGX1X8hNhGZ1V8fNiG]1T8dNbFPOb0^2]3aMg0Q1XKWO>Z2`3aMg0o0XK^O<T2b3bMh0l0WKA?S2^3dMi0i0WKCb0S2Z3dMk0k0RKAh0R2X3eMl0f1kKg0V3gMl0c1nKg0T3iMl0a1oKf0S3mMk0^1RLf0Q3oMk0[1TLg0o2RNi0Y1VLg0n2TNj0U1XLg0m2XNg0S1[Lf0l2[Nf0o0^Lg0j2_Nc0k0cLg0g2cNc0g0eLg0f2fNb0c0gLh0f2jN>?lLh0d2mN=<nLh0b2QO<8RMh0`2TO;4UMh0_2YO80WMi0`2XO70YMi0^2XO90XMi0^2XO9OYMi0]2YO:NYMj0\\2YO9OZMi0;_LU1j2V1NYMk0:^LW1j2U1MZMk08`LX1i2V1MYMk06aL[1h2T1MZMl05`L]1g2T1MYMn03`L_1i2R1JYMo05_L`1k2n0H\\MR11]Ld1m2l0E]MV1MZLj1m2j0C^Mg1g1hNj0A^Mh1]1[KjN]3Y2B^Mg1_1ZKiN^3Z2X4nNZHhN_3Y2W4oNZHhN`3W2W4QOZHfN`3Y2V4QOZHfNa3X2U4ROZHeNc3X2S4TOZHcNc3X2T4UOYHbNe3X2R4VOYHbNf3W2Q4WOYHaNg3W2Q4XOYH`Ng3W2P4YOYH`Nh3U2P4[OXH_Ni3U2P4\\OWH_Nj3T2o3]OVLa0k3_OUL`0l3@UL?k3AUL>l3BTL=l3DSL=m3CSL<n3DQL<P4DPL<P4DPL7U4IjK2\\4NdKMa43^KFj4:VKZOV5f0kJPO^5Q1dJiN_5W1dJcN_5]1dJ]N_5d1X40000001N10001O0000001O0000001O00001O001O1O2M2O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1N3M3M4L3M3M1O1O1O1O1O2N100O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O2O0O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2O2M2N2N2N2N3M2N2N2N3M2N2N3M2O2N2M3N1N3N2M3N2M2O2N2M3N1N3N2M3N2M2O2N2M3N1N3N4K5L4K5L4L5J5L4K5L4K5L5K4K5J6G9F;F9G9F:G9F;F9G9F:G9F:G:F9F:Ggmo0"}, "label": "a boy riding a bike"}]}
{"id": 522423, "image": "COCO_train2014_000000522423.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the asian man on a bicycle\"."}], "task": "refering", "answer_template": "The \"the asian man on a bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 43, 44, 45, 46, 61, 62, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 85, 97, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 223, 224, 225, 226, 227, 228, 229, 230, 231, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 295, 296, 297, 298, 299, 300, 313, 314, 315, 316, 317], "bbox": [0.27294, 0.09437999999999999, 0.868, 0.99818], "iscrowd": 0, "area": 79533.46545, "rle": {"size": [500, 500], "counts": "_US22b?1N2N2N2O1N1O2N2O1N1O1O2O0O10001O0O101O00001N100000001TDVOl8j0RGZOk8g0SG]Ok8c0RGBl8>RGEm8;PGJm88PGLn84oF1o8OPG4n8LQG7m8IRG:k8GUG:j8GTG<j8DUG?i8AWGa0g8_OXGd0e8]O[Ge0c8[O\\Gh0b8XO]Gk0a8UO_Gm0_8SO`GP1]8QObGS1[8mNeGU1Y8kNfGX1X8hNhGZ1V8fNiG]1T8dNbFPOb0^2]3aMg0Q1XKWO>Z2`3aMg0o0XK^O<T2b3bMh0l0WKA?S2^3dMi0i0WKCb0S2Z3dMk0k0RKAh0R2X3eMl0f1kKg0V3gMl0c1nKg0T3iMl0a1oKf0S3mMk0^1RLf0Q3oMk0[1TLg0o2RNi0Y1VLg0n2TNj0U1XLg0m2XNg0S1[Lf0l2[Nf0o0^Lg0j2_Nc0k0cLg0g2cNc0g0eLg0f2fNb0c0gLh0f2jN>?lLh0d2mN=<nLh0b2QO<8RMh0`2TO;4UMh0_2YO80WMi0`2XO70YMi0^2XO90XMi0^2XO9OYMi0]2YO:NYMj0\\2YO9OZMi0;_LU1j2V1NYMk0:^LW1j2U1MZMk08`LX1i2V1MYMk06aL[1h2T1MZMl05`L]1g2T1MYMn03`L_1i2R1JYMo05_L`1k2n0H\\MR11]Ld1m2l0E]MV1MZLj1m2j0C^Mg1g1hNj0A^Mh1]1[KjN]3Y2B^Mg1_1ZKiN^3Z2X4nNZHhN_3Y2W4oNZHhN`3W2W4QOZHfN`3Y2V4QOZHfNa3X2U4ROZHeNc3X2S4TOZHcNc3X2T4UOYHbNe3X2R4VOYHbNf3W2Q4WOYHaNg3W2Q4XOYH`Ng3W2P4YOYH`Nh3U2P4[OXH_Ni3U2P4\\OWH_Nj3T2o3]OVLa0k3_OUL`0l3@UL?k3AUL>l3BTL=l3DSL=m3CSL<n3DQL<P4DPL<P4DPL7U4IjK2\\4NdKMa43^KFj4:VKZOV5f0kJPO^5Q1dJiN_5W1dJcN_5]1dJ]N_5d1X40000001N10001O0000001O0000001O00001O001O1O2M2O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1N3M3M4L3M3M1O1O1O1O1O2N100O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O2O0O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2O2M2N2N2N2N3M2N2N2N3M2N2N3M2O2N2M3N1N3N2M3N2M2O2N2M3N1N3N2M3N2M2O2N2M3N1N3N4K5L4K5L4L5J5L4K5L4K5L5K4K5J6G9F;F9G9F:G9F;F9G9F:G9F:G:F9F:Ggmo0"}, "label": "the asian man on a bicycle"}]}
{"id": 440511, "image": "COCO_train2014_000000440511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leaning elephant\"."}], "task": "refering", "answer_template": "The \"the leaning elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 242, 243, 259, 260, 261, 262, 263, 265, 266, 284, 285, 286], "bbox": [0.22771875000000003, 0.3476666666666667, 0.585890625, 0.7417916666666666], "iscrowd": 0, "area": 23741.460100000004, "rle": {"size": [480, 640], "counts": "gfT22l>4M4K4M3L5L3L4L5L3L4M3L5L2M3N2M2O2M3M2O2M3N1N3N2M2O2M3N1N3M3N1N3N1N3N2M2O2M3N1N2O1N2O10O01O100O100O1O10O0100O1O100O10O01O100O100O1O010O100O1O10O0100O1O100O10O0100OO2O0O2O0O2O^EZLl9d3QFbLm9^3nEiLQ:U3lERMR:m2iEZMU:j3O1O2M2O1N2O1O2M10000O2O000O100O10000O100000J`FlJ`9P5eFPKZ9k4?K6H8H8KfEQLm9m3TF[Le9c3YFhL`9V3YFZM`9T4N2N101N1O2N101N1O2N1O2O0O2O000000O1^MlFXOT9f0PGXOP9g0SGWOm8g0WGVOj8h0ZGVOf8i0]GUOb8j0bG_NZOUOT9Z2fG\\N[OXOo8[2iGXN]OZOk8\\2lGUN]O^Og8[2oGTN^O_Oc8[2RHRN@A^8Z2WHPN_OD[8Z2YHnMAFV8Z2]HkMBIR8Y2_HjMDKn7X2bHhMDOk7W2_3O1N101N2O1N2O1N200001O1N101O1O001O1O001O10O0100O010O10O0100O010O10O0100O1O0100O3N3L4M3L3M3N1N2O2M2O1N3N1NfBiNm<U1PCnNQ=P1lBTOU=j0iBYOY=e0cB_O^=R11O2J5L4K5L4K5K5L4K5L4KSmk3"}, "label": "the leaning elephant"}]}
{"id": 6338, "image": "COCO_train2014_000000006338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the heads of two horses walking on the beach\"."}], "task": "refering", "answer_template": "The \"the heads of two horses walking on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 144, 145, 167, 168, 170, 192, 193, 194, 215, 216, 217, 238, 240], "bbox": [0.288703125, 0.29691666666666666, 0.45806250000000004, 0.6173541666666668], "iscrowd": 0, "area": 3744.0776, "rle": {"size": [480, 640], "counts": "Ulf23j>5K5K5\\BE^<>WCKh<8lB2S=o0N1O1O101N1O1O2N100O1O2N10001O00N2N2N2N2N2N2N2C=K5O100OhElNf6S1YIPOf6o0XITOh6k0WIWOi6i0UIYOk6f0TI\\OV;1OO1O01M2N2010O1O1O100O0001O002N2O100O100O11O00001O001O00001O00001O001O01O3N1N3N2M3M3N1O2O1O1N101O1N2O1OkN\\Bm0c=SO`Bl0_=SOcBm0>oNX<3]Cm07RO\\<1_Cn00SOb<NaCn0JUOf<LcCn0CXOj<JeCn0]OYOP=HfCV1T=N3L3M4M2M4L3M4If\\R5"}, "label": "the heads of two horses walking on the beach"}]}
{"id": 6338, "image": "COCO_train2014_000000006338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse closest to the waves\"."}], "task": "refering", "answer_template": "The \"horse closest to the waves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 144, 145, 167, 168, 170, 192, 193, 194, 215, 216, 217, 238, 240], "bbox": [0.288703125, 0.29691666666666666, 0.45806250000000004, 0.6173541666666668], "iscrowd": 0, "area": 3744.0776, "rle": {"size": [480, 640], "counts": "Ulf23j>5K5K5\\BE^<>WCKh<8lB2S=o0N1O1O101N1O1O2N100O1O2N10001O00N2N2N2N2N2N2N2C=K5O100OhElNf6S1YIPOf6o0XITOh6k0WIWOi6i0UIYOk6f0TI\\OV;1OO1O01M2N2010O1O1O100O0001O002N2O100O100O11O00001O001O00001O00001O001O01O3N1N3N2M3M3N1O2O1O1N101O1N2O1OkN\\Bm0c=SO`Bl0_=SOcBm0>oNX<3]Cm07RO\\<1_Cn00SOb<NaCn0JUOf<LcCn0CXOj<JeCn0]OYOP=HfCV1T=N3L3M4M2M4L3M4If\\R5"}, "label": "horse closest to the waves"}]}
{"id": 6338, "image": "COCO_train2014_000000006338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman riding the horse that is closest to the camera . she is wearing a pink , purple , green , and yellow bandeau bikini\"."}], "task": "refering", "answer_template": "The \"the woman riding the horse that is closest to the camera . she is wearing a pink , purple , green , and yellow bandeau bikini\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 152, 153, 175, 176, 198, 199, 221, 222, 244, 245], "bbox": [0.559671875, 0.0134375, 0.701609375, 0.6005416666666666], "iscrowd": 0, "area": 12267.64685, "rle": {"size": [480, 640], "counts": "]nW51n>3M3L4M3M3M1N2O1O1O1N2O1O1O1O1N2O1N2N2N3M2O1N2N3M2N2N2oM_NbFc1U9hNeF[1V9lNdFX1Z9lN`FX1_9kN[FZ1c9iNWF[1h9hNRF\\1m9eNPF^1o9cNPF_1n9cNPF^1EYNU9:UGh2e8YMYGn2a8TM]GR3]8oLbGW3X8kLfG[3e1]KZ4Y1PJ`3[1]Ke4T1nIe3R1]KP5n0lIl3g0^K]5g0jIQ4<`Kj5Y6iIoIW6]700O10O10000000O10PNkIUKU6i4RJRKn5l4XJPKh5m4`JnJ`5P5fJlJY5S5kJkJV5R5oJkJQ5S5TKiJn4T5WKiJj4S5\\KjJe4S5`KjJa4S5dKjJ]4T5fKgJ^4W5fKcJ]4[5hK_J\\4_5d2N2M3SKdFS4^9hKhFU4`9_KgF]4Q:M3M3L4M5K6I6K6J7cNYDSOn;c0ZDYOm;<ZDAP<1WDKX<CPD1kii2"}, "label": "the woman riding the horse that is closest to the camera . she is wearing a pink , purple , green , and yellow bandeau bikini"}]}
{"id": 6338, "image": "COCO_train2014_000000006338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a bikini and white headband rides a horse on the beach\"."}], "task": "refering", "answer_template": "The \"a woman with a bikini and white headband rides a horse on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 152, 153, 175, 176, 198, 199, 221, 222, 244, 245], "bbox": [0.559671875, 0.0134375, 0.701609375, 0.6005416666666666], "iscrowd": 0, "area": 12267.64685, "rle": {"size": [480, 640], "counts": "]nW51n>3M3L4M3M3M1N2O1O1O1N2O1O1O1O1N2O1N2N2N3M2O1N2N3M2N2N2oM_NbFc1U9hNeF[1V9lNdFX1Z9lN`FX1_9kN[FZ1c9iNWF[1h9hNRF\\1m9eNPF^1o9cNPF_1n9cNPF^1EYNU9:UGh2e8YMYGn2a8TM]GR3]8oLbGW3X8kLfG[3e1]KZ4Y1PJ`3[1]Ke4T1nIe3R1]KP5n0lIl3g0^K]5g0jIQ4<`Kj5Y6iIoIW6]700O10O10000000O10PNkIUKU6i4RJRKn5l4XJPKh5m4`JnJ`5P5fJlJY5S5kJkJV5R5oJkJQ5S5TKiJn4T5WKiJj4S5\\KjJe4S5`KjJa4S5dKjJ]4T5fKgJ^4W5fKcJ]4[5hK_J\\4_5d2N2M3SKdFS4^9hKhFU4`9_KgF]4Q:M3M3L4M5K6I6K6J7cNYDSOn;c0ZDYOm;<ZDAP<1WDKX<CPD1kii2"}, "label": "a woman with a bikini and white headband rides a horse on the beach"}]}
{"id": 6338, "image": "COCO_train2014_000000006338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown horse with a woman in a striped swimsuit riding it\"."}], "task": "refering", "answer_template": "The \"brown horse with a woman in a striped swimsuit riding it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 123, 124, 125, 126, 127, 128, 129, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 192, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 247, 248, 249, 250, 251, 263, 264, 265, 266, 270, 271, 272, 273, 274, 286, 289, 293, 295, 309, 312, 315, 316, 318, 332, 335, 339], "bbox": [0.337984375, 0.25235416666666666, 0.9544062499999999, 0.8875624999999999], "iscrowd": 0, "area": 43184.490800000014, "rle": {"size": [480, 640], "counts": "P_U31g>;D;F9F:G8G:F:fNnM\\EZ2e:kMmD^2R;hM`Da2`;?10000O0100O001O1N10O1L3M4L4N10100O010O1O10O0100O010O1O010O100O010O2N100O100O1O2O000000010O000001O000jIYM4f2HbM4^2HjM4W2HPN4P2HWN6h1G[N9e1C_N>a1^OcNa0]1[OgNe0Y1XOjNi0U1SOoNm0R1oNROQ1m0kNWOU1i0hNZOX1f0dN\\O_1c0^N^Od1c0WN@k1?dM0^22^MOd22YMMi25SMKo27nLoN\\M]Nh5f2hLbNRN`NX5P3bLVNo4k1nJjM^5X2^JfMf5\\2WJdMk5^2PJcMQ6_2lI`MW6a2eI`M\\6b2aI^M`6d2\\I]Mf6e2VI[Mk6g2QIZMQ7g2lHXMV7j2fHWM\\7j2aHVM`7m2[HTMf7m2XHSMj7n2RHRMP8o2nGQMT8o2iGRMX8o2fGQM[8P3bGQM`8o2^GQMc8Q3YGoLj8Q3TGoLm8R3PGoLR9Q3lFoLU9R3hFnLZ9S3dFmL^9S3_FnLb9W42N1N2O2N1O2M2O2N1O1O2M2O2N1O1012M3N2M2N3M3L3N3M3M2N3M3iH^Jj4e5RK\\Jm4g5PKZJP5i5mJWJS5S4RI_Mg1_NW5Q4VI_M`1aNZ5P4YI_MZ1aN]5o3]I`MR1bNa5m3aI`Mk0dNe5j3dIbMd0dNh5i3hIcM<dNm5i3jIcM5eNQ6g3nIcMOfNS6g3PJdMIfNW6f3SJdMBgN[6e3VJcM\\OiN_6c3XJdMVOiNb6c3[JdMoNjNf6b3^JcMiNlNi6a3aJcMcNlNl6a3dJcM\\NmNP7`3gJcMUNnNT7`3iJoMW5Q2lJkMV5T2mJhMT5X2oJcMS5]2oJ_MS5a2PKZMS5e2PKVMS5i2PKRMT5l2oJoLT5P3oJkLT5T3oJgLT5X3oJbLU5]3nJ^LU5b3U3_MfDZ1\\;aNkD[1V;`NRE\\1o:`NXE\\1i:_N_E]1b:_NfE\\1[:`NlE\\1U:_NSF]1n9_NYF]1i;L4L7I7J6I7I7Ide3M^ZL4YCMZ:6_ENa:6VE0i:3PE1P;3gD3X;0aD5^;OYD6g;MRD8m;LjC:U<X101O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O0100O10000O10000O010O10000OlG]Mm3d2PL_Mo3a2PLaMo3`2oKaMQ4`2mKaMS4_2kKdMS4^2kKcMU4^2iKdMV4\\2iKeMW4\\2fKfMZ4Z2eKhMZ4Y2dKhM\\4Y2`KjM`4V2\\KkMg4V2SKhMT5Z2fJfM]5^2\\JeMe5^2SJfMn5^2jIeMW6^2aIeMa6^2VIgMk6]2dHoM]7T2THWNm7\\42N2M2M4M4K5L4K4M4K5L4K4M4K5L4K4D=^Ob0001N1000hNQE_Nn:]1VEcNi:Z1\\EeNd:V1aEkN]:R1gEnNX:P1kEQOS:m0QFROn9l0UFTOj9k0XFVOe9i0^FWOa9g0cFXO\\9f0gF[OW9c0lF]OS9b0oF_Oo8?UGBh8<[GGa87bGK[84gGOU8OoG3m7KVH7g7G\\H<`7CeH<X7BoH=m6AYI<g6FZI7g6KYI2h6OYINh64XIIi68YIDh6>XI_Oi6b0XI\\Oh6f0XIWOi6j0XISOi6o0^H]N`Na0S9S1UHbNlN7P9Y1lGeNWOOn8^1cGiN_OGP9a1YGnNH@P9d1PGQO0ZOR9f1gFUO7UOS9h1]FjNo0]Of8j1TFnNV1XOg8P3ZGnLh8Q3XGoLi8P3WGoLk8P3UGPMl8o2TGPMn8n2TGQMm8n2WFPM71d9n2kEYMc0Ic9^3_FaLc9\\3_FdLb9Y3`FfLb9W3V1M3L4M3M3M3M3M3M3M3M3L4O1N2O0O01000O100M3K5J5N30000oN`Ba0`=^OaBa0`=^OaBMN:a=IaBL28\\=LcBK47Y=McBL74W=OcBL93T=0dBL<1P=2eBL>0m<3fBL`0Nj<7fBJU>5lAJU>5kAKV>4kAKV>4<O1O010O1O1Omh="}, "label": "brown horse with a woman in a striped swimsuit riding it"}]}
{"id": 6338, "image": "COCO_train2014_000000006338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse walking on the beach\"."}], "task": "refering", "answer_template": "The \"a horse walking on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 123, 124, 125, 126, 127, 128, 129, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 192, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 247, 248, 249, 250, 251, 263, 264, 265, 266, 270, 271, 272, 273, 274, 286, 289, 293, 295, 309, 312, 315, 316, 318, 332, 335, 339], "bbox": [0.337984375, 0.25235416666666666, 0.9544062499999999, 0.8875624999999999], "iscrowd": 0, "area": 43184.490800000014, "rle": {"size": [480, 640], "counts": "P_U31g>;D;F9F:G8G:F:fNnM\\EZ2e:kMmD^2R;hM`Da2`;?10000O0100O001O1N10O1L3M4L4N10100O010O1O10O0100O010O1O010O100O010O2N100O100O1O2O000000010O000001O000jIYM4f2HbM4^2HjM4W2HPN4P2HWN6h1G[N9e1C_N>a1^OcNa0]1[OgNe0Y1XOjNi0U1SOoNm0R1oNROQ1m0kNWOU1i0hNZOX1f0dN\\O_1c0^N^Od1c0WN@k1?dM0^22^MOd22YMMi25SMKo27nLoN\\M]Nh5f2hLbNRN`NX5P3bLVNo4k1nJjM^5X2^JfMf5\\2WJdMk5^2PJcMQ6_2lI`MW6a2eI`M\\6b2aI^M`6d2\\I]Mf6e2VI[Mk6g2QIZMQ7g2lHXMV7j2fHWM\\7j2aHVM`7m2[HTMf7m2XHSMj7n2RHRMP8o2nGQMT8o2iGRMX8o2fGQM[8P3bGQM`8o2^GQMc8Q3YGoLj8Q3TGoLm8R3PGoLR9Q3lFoLU9R3hFnLZ9S3dFmL^9S3_FnLb9W42N1N2O2N1O2M2O2N1O1O2M2O2N1O1012M3N2M2N3M3L3N3M3M2N3M3iH^Jj4e5RK\\Jm4g5PKZJP5i5mJWJS5S4RI_Mg1_NW5Q4VI_M`1aNZ5P4YI_MZ1aN]5o3]I`MR1bNa5m3aI`Mk0dNe5j3dIbMd0dNh5i3hIcM<dNm5i3jIcM5eNQ6g3nIcMOfNS6g3PJdMIfNW6f3SJdMBgN[6e3VJcM\\OiN_6c3XJdMVOiNb6c3[JdMoNjNf6b3^JcMiNlNi6a3aJcMcNlNl6a3dJcM\\NmNP7`3gJcMUNnNT7`3iJoMW5Q2lJkMV5T2mJhMT5X2oJcMS5]2oJ_MS5a2PKZMS5e2PKVMS5i2PKRMT5l2oJoLT5P3oJkLT5T3oJgLT5X3oJbLU5]3nJ^LU5b3U3_MfDZ1\\;aNkD[1V;`NRE\\1o:`NXE\\1i:_N_E]1b:_NfE\\1[:`NlE\\1U:_NSF]1n9_NYF]1i;L4L7I7J6I7I7Ide3M^ZL4YCMZ:6_ENa:6VE0i:3PE1P;3gD3X;0aD5^;OYD6g;MRD8m;LjC:U<X101O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O0100O10000O10000O010O10000OlG]Mm3d2PL_Mo3a2PLaMo3`2oKaMQ4`2mKaMS4_2kKdMS4^2kKcMU4^2iKdMV4\\2iKeMW4\\2fKfMZ4Z2eKhMZ4Y2dKhM\\4Y2`KjM`4V2\\KkMg4V2SKhMT5Z2fJfM]5^2\\JeMe5^2SJfMn5^2jIeMW6^2aIeMa6^2VIgMk6]2dHoM]7T2THWNm7\\42N2M2M4M4K5L4K4M4K5L4K4M4K5L4K4D=^Ob0001N1000hNQE_Nn:]1VEcNi:Z1\\EeNd:V1aEkN]:R1gEnNX:P1kEQOS:m0QFROn9l0UFTOj9k0XFVOe9i0^FWOa9g0cFXO\\9f0gF[OW9c0lF]OS9b0oF_Oo8?UGBh8<[GGa87bGK[84gGOU8OoG3m7KVH7g7G\\H<`7CeH<X7BoH=m6AYI<g6FZI7g6KYI2h6OYINh64XIIi68YIDh6>XI_Oi6b0XI\\Oh6f0XIWOi6j0XISOi6o0^H]N`Na0S9S1UHbNlN7P9Y1lGeNWOOn8^1cGiN_OGP9a1YGnNH@P9d1PGQO0ZOR9f1gFUO7UOS9h1]FjNo0]Of8j1TFnNV1XOg8P3ZGnLh8Q3XGoLi8P3WGoLk8P3UGPMl8o2TGPMn8n2TGQMm8n2WFPM71d9n2kEYMc0Ic9^3_FaLc9\\3_FdLb9Y3`FfLb9W3V1M3L4M3M3M3M3M3M3M3M3L4O1N2O0O01000O100M3K5J5N30000oN`Ba0`=^OaBa0`=^OaBMN:a=IaBL28\\=LcBK47Y=McBL74W=OcBL93T=0dBL<1P=2eBL>0m<3fBL`0Nj<7fBJU>5lAJU>5kAKV>4kAKV>4<O1O010O1O1Omh="}, "label": "a horse walking on the beach"}]}
{"id": 514243, "image": "COCO_train2014_000000514243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman in a blue shirt and white shorts talking on her cell phone\"."}], "task": "refering", "answer_template": "The \"a blonde woman in a blue shirt and white shorts talking on her cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 57, 58, 59, 60, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 289, 290, 309, 310, 313, 314, 332, 333, 335, 336, 337], "bbox": [0.44578125, 0.012552693208430913, 0.67915625, 0.9856440281030445], "iscrowd": 0, "area": 37162.4497, "rle": {"size": [427, 640], "counts": "[Wg33S=e0ZO7J6L3M4K5L4VOi0aN_1A?N2N2N2N2N2N2N2N2oJSLNc40dKFa47gK@]4=kKYOZ4g0kKoNZ4P1lKfNY4Y1mK]NY4b1lKVNW4i1oKPNS4o1QLlMZ2XObLl2V1iMW2E[La2`1gMU22SLW2j1cMS2`0kKl1T2bMP2l0dKa1_2_Mm1Y1]KX1g2\\Ml1f1UK75WNm2Ih1P4fKPLd2Mf1Q4PLlKY7Q4Q1M3M3M3N2M3M3N2001O001O1J7D;E<D;M43M4K4M3M4K4M3L5L3M3L5M0O1O100O1O1O101N1O100O1O100O10001O0O1000000O10001N100002NPOnEgNQ:W1PFkNo9R1SFQOk9l0WFVOg0oNn7i1\\GB2POc8Z1]Gb0b8[O`Gf0`8WObGj0_6bMRKb1`Nn0\\6bMSK]1cNR1Y6bMTKY1eNU1V6dMUKU1fNX1T6dMVKQ1hN[1Q6fMZKi0gNa1o5gM\\K=mNl1g5iM^KOSOX2_5jMoLW2Q3iMnLW2R3jMmLV2S3kMmLT2T3mMjLS2V3nMiLS2W3mMhLS2W3PNgLP2Y3TNcLl1]3XN_Lh1a3[N\\Lf1c3^NYLb1g3bNUL^1^1`MPNV1>Z1_1gMoMS1>V1`1mMoMQ1<T1a1RNoMn0;Q1c1XNnMj0;o0d1^NmMg0:l0g1cNkMe09i0i1iNgMe0;d0j1nNeMb0>a0j1TObM>a0`0j1YO^M;e0>j1^O[M8h0;j1DWM5l0:i1HUM1o09i1n2SNUMi1o2TNUMh1n2VNUMe1P3WNTMd1P3YNTMb1R3YNRMb1U3XNoLc1Y3WNjLd1i7D>B=C>CW[e2"}, "label": "a blonde woman in a blue shirt and white shorts talking on her cell phone"}]}
{"id": 514243, "image": "COCO_train2014_000000514243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde headed girl wearing a blue shirt and white shorts , talking on the phone\"."}], "task": "refering", "answer_template": "The \"a blonde headed girl wearing a blue shirt and white shorts , talking on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 57, 58, 59, 60, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 289, 290, 309, 310, 313, 314, 332, 333, 335, 336, 337], "bbox": [0.44578125, 0.012552693208430913, 0.67915625, 0.9856440281030445], "iscrowd": 0, "area": 37162.4497, "rle": {"size": [427, 640], "counts": "[Wg33S=e0ZO7J6L3M4K5L4VOi0aN_1A?N2N2N2N2N2N2N2N2oJSLNc40dKFa47gK@]4=kKYOZ4g0kKoNZ4P1lKfNY4Y1mK]NY4b1lKVNW4i1oKPNS4o1QLlMZ2XObLl2V1iMW2E[La2`1gMU22SLW2j1cMS2`0kKl1T2bMP2l0dKa1_2_Mm1Y1]KX1g2\\Ml1f1UK75WNm2Ih1P4fKPLd2Mf1Q4PLlKY7Q4Q1M3M3M3N2M3M3N2001O001O1J7D;E<D;M43M4K4M3M4K4M3L5L3M3L5M0O1O100O1O1O101N1O100O1O100O10001O0O1000000O10001N100002NPOnEgNQ:W1PFkNo9R1SFQOk9l0WFVOg0oNn7i1\\GB2POc8Z1]Gb0b8[O`Gf0`8WObGj0_6bMRKb1`Nn0\\6bMSK]1cNR1Y6bMTKY1eNU1V6dMUKU1fNX1T6dMVKQ1hN[1Q6fMZKi0gNa1o5gM\\K=mNl1g5iM^KOSOX2_5jMoLW2Q3iMnLW2R3jMmLV2S3kMmLT2T3mMjLS2V3nMiLS2W3mMhLS2W3PNgLP2Y3TNcLl1]3XN_Lh1a3[N\\Lf1c3^NYLb1g3bNUL^1^1`MPNV1>Z1_1gMoMS1>V1`1mMoMQ1<T1a1RNoMn0;Q1c1XNnMj0;o0d1^NmMg0:l0g1cNkMe09i0i1iNgMe0;d0j1nNeMb0>a0j1TObM>a0`0j1YO^M;e0>j1^O[M8h0;j1DWM5l0:i1HUM1o09i1n2SNUMi1o2TNUMh1n2VNUMe1P3WNTMd1P3YNTMb1R3YNRMb1U3XNoLc1Y3WNjLd1i7D>B=C>CW[e2"}, "label": "a blonde headed girl wearing a blue shirt and white shorts , talking on the phone"}]}
{"id": 514243, "image": "COCO_train2014_000000514243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bag held by a woman on a cell phone sitting on a public bench\"."}], "task": "refering", "answer_template": "The \"a white bag held by a woman on a cell phone sitting on a public bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 83, 130, 131, 152, 153, 154, 175, 176, 177, 178, 198, 199, 200, 201, 222, 223, 224], "bbox": [0.624109375, 0.1748711943793911, 0.774421875, 0.6638173302107728], "iscrowd": 0, "area": 7477.994149999999, "rle": {"size": [427, 640], "counts": "n^V5<m<>B4L2OA>I7O3N1O1O_G8b3GmKn0Q4POiKY1P86K5K6J7I6J7I7I5L3L4L4L4L4L4M4K3N3L4M2O2N2N1O2O11N2O0O2N1O2N1O2N2N1O2N1O1O2N1N2O2N1O1O2N1O2M4M3L3N3L4M2M4M3M2M4M1N3N1N2O2M2O1N3N1N2N3M3N1N3M3M4L3M3N1N3M3M2N3N1N3M3M2N3M2Ogik1"}, "label": "a white bag held by a woman on a cell phone sitting on a public bench"}]}
{"id": 514243, "image": "COCO_train2014_000000514243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bag that is being worn by a woman wearing a blue shirt that says pink\"."}], "task": "refering", "answer_template": "The \"a white bag that is being worn by a woman wearing a blue shirt that says pink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 83, 130, 131, 152, 153, 154, 175, 176, 177, 178, 198, 199, 200, 201, 222, 223, 224], "bbox": [0.624109375, 0.1748711943793911, 0.774421875, 0.6638173302107728], "iscrowd": 0, "area": 7477.994149999999, "rle": {"size": [427, 640], "counts": "n^V5<m<>B4L2OA>I7O3N1O1O_G8b3GmKn0Q4POiKY1P86K5K6J7I6J7I7I5L3L4L4L4L4L4M4K3N3L4M2O2N2N1O2O11N2O0O2N1O2N1O2N2N1O2N1O1O2N1N2O2N1O1O2N1O2M4M3L3N3L4M2M4M3M2M4M1N3N1N2O2M2O1N3N1N2N3M3N1N3M3M4L3M3N1N3M3M2N3N1N3M3M2N3M2Ogik1"}, "label": "a white bag that is being worn by a woman wearing a blue shirt that says pink"}]}
{"id": 514243, "image": "COCO_train2014_000000514243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with black sitting and wearing a striped shirt\"."}], "task": "refering", "answer_template": "The \"a woman with black sitting and wearing a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 130, 131, 132, 153, 154, 155, 156, 177, 178, 179, 180, 181, 201, 202, 203, 204, 227, 250, 272, 273, 295, 296, 319, 320], "bbox": [0.682140625, 0.09812646370023419, 0.951421875, 0.9282435597189694], "iscrowd": 0, "area": 14435.755899999995, "rle": {"size": [427, 640], "counts": "dZf5f1c;5K5J6K5K5K5J5L5J5L5J5L5J5L5J5L5J5L3N3L3M4M2M4L3N3M2N2N1O101N1O2N1O1O2O001O0O10jMQIQNn6e1\\I[Nd6c1_I\\Na6b1aI_N]6`1eI`N[6^1hIbNW6[1lIeNT6Y1oIfM`N`0`7i1RJgM_N`0^7g1UJiM_N>\\7g1WJjM_N?Y7e1YJmM_N=X7d1[JoM^N<V7d1^JPN]N<T7c1_JRN^N:S7c1`JRN_N;Q7a1aJTN_N:Q7`1aJVN_N9Q7_1`JYNaN7n6_1bJZNaN6n6]1cJ\\NaN7l6[1cJ_NbN5l6Z1cJaNaN5l6Y1dJbNaN5k6W1eJdN`N5l6T1eJgN`N6k6P1gJkN]N5l6m0jJnNZN5m6j0kJQOXN6m6f0lJUOWN5n6c0mJYOTN5n6`0QKZORN5n6?QK\\OQN5o6<RK@nM5P78TKClM5R75SKFkM5T71TKIhM7U7LUKNeM6W7IVK1cM6Y7EVK4cM6g:H[E8f:F[E:e:D]E=b:A`E?`:@bE?^:AbE?_:_ObEa0^:_ObEa0^:_ObEb0]:]OeEb0\\:]OdEc0\\:\\OeEd0[:\\OeEd0[:\\OeEd0\\:ZOfEf0Y:ZOgEf0Y:YOhEg0^;00000O2O0000000O101O000O1000001O0O10001O2eFJ]57fIg0X6ZOaIo0\\6RO]IV1a6lNXI\\1e6eNUIc1h6^NTIg1j6ZNRIk1k6XNPIk1R7V23M4M2M3M?A2O1N1O2N4Lj0mMWFQOR<O1O1O100O1O1O100O1O1O1O1O1O1N2O001N2O1O0O2O1O001N2O001N2O1O1N2O2N1Mea<"}, "label": "a woman with black sitting and wearing a striped shirt"}]}
{"id": 514243, "image": "COCO_train2014_000000514243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with black hair wearing jeans , a striped gray shirt and flip flops sitting on a bench\"."}], "task": "refering", "answer_template": "The \"a lady with black hair wearing jeans , a striped gray shirt and flip flops sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 130, 131, 132, 153, 154, 155, 156, 177, 178, 179, 180, 181, 201, 202, 203, 204, 227, 250, 272, 273, 295, 296, 319, 320], "bbox": [0.682140625, 0.09812646370023419, 0.951421875, 0.9282435597189694], "iscrowd": 0, "area": 14435.755899999995, "rle": {"size": [427, 640], "counts": "dZf5f1c;5K5J6K5K5K5J5L5J5L5J5L5J5L5J5L5J5L3N3L3M4M2M4L3N3M2N2N1O101N1O2N1O1O2O001O0O10jMQIQNn6e1\\I[Nd6c1_I\\Na6b1aI_N]6`1eI`N[6^1hIbNW6[1lIeNT6Y1oIfM`N`0`7i1RJgM_N`0^7g1UJiM_N>\\7g1WJjM_N?Y7e1YJmM_N=X7d1[JoM^N<V7d1^JPN]N<T7c1_JRN^N:S7c1`JRN_N;Q7a1aJTN_N:Q7`1aJVN_N9Q7_1`JYNaN7n6_1bJZNaN6n6]1cJ\\NaN7l6[1cJ_NbN5l6Z1cJaNaN5l6Y1dJbNaN5k6W1eJdN`N5l6T1eJgN`N6k6P1gJkN]N5l6m0jJnNZN5m6j0kJQOXN6m6f0lJUOWN5n6c0mJYOTN5n6`0QKZORN5n6?QK\\OQN5o6<RK@nM5P78TKClM5R75SKFkM5T71TKIhM7U7LUKNeM6W7IVK1cM6Y7EVK4cM6g:H[E8f:F[E:e:D]E=b:A`E?`:@bE?^:AbE?_:_ObEa0^:_ObEa0^:_ObEb0]:]OeEb0\\:]OdEc0\\:\\OeEd0[:\\OeEd0[:\\OeEd0\\:ZOfEf0Y:ZOgEf0Y:YOhEg0^;00000O2O0000000O101O000O1000001O0O10001O2eFJ]57fIg0X6ZOaIo0\\6RO]IV1a6lNXI\\1e6eNUIc1h6^NTIg1j6ZNRIk1k6XNPIk1R7V23M4M2M3M?A2O1N1O2N4Lj0mMWFQOR<O1O1O100O1O1O100O1O1O1O1O1O1N2O001N2O1O0O2O1O001N2O001N2O1O1N2O2N1Mea<"}, "label": "a lady with black hair wearing jeans , a striped gray shirt and flip flops sitting on a bench"}]}
{"id": 366795, "image": "COCO_train2014_000000366795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby bib is fish printed\"."}], "task": "refering", "answer_template": "The \"the baby bib is fish printed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 260, 261, 262, 263, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.089609375, 0.6207708333333334, 0.46168750000000003, 0.9870208333333335], "iscrowd": 0, "area": 22851.645399999994, "rle": {"size": [480, 640], "counts": "eSk0g0X>2N3M3M2N3M3M2N3M3M2N3L5L4L4L4L5K4L4L4L4L1O1O1O1O1O1O100001O00001O0O101O00001O00001O00001O00001O00000000000000000000000000000000000000000000000000O10000000O100000000000000000000000000000000000000000000000000000000000O1O1O100O1O1O1O1O1O100O1O1O1O00100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100L4J6I7J6J6J6L43L3N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M4L4L4L4L4L5K4K5J6J6J6K5J7I5K2N100OXN_NUF_1k9hNoEW1R:oNhEP1V:YOeEf0X:DaE;]:N^E0`:9ZEFd:c0VE]Og:l0TEROj:W1PEhNn:e2N2N31N3M2N3N2M2N3M2N3N1N3M2J7H7H9H7H9H7H9H7H9H6I7J6If]Q5"}, "label": "the baby bib is fish printed"}]}
{"id": 366795, "image": "COCO_train2014_000000366795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bag sitting on a bench\"."}], "task": "refering", "answer_template": "The \"a black bag sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 316, 317, 318, 319], "bbox": [0.6990000000000001, 0.47920833333333335, 0.915375, 0.7817916666666667], "iscrowd": 0, "area": 14226.262949999993, "rle": {"size": [480, 640], "counts": "gka61o>0O1O2YO2YBOe=:TBFj==TBEi==VBDh=?VBCh=?WBAg=a0XB@f=o0M2N2O2M2N2N3M2N3O2N1O1O2CVN`Ck1]<?M2M3M4M2O1O2N1O2N1O00000000000K5N3N1O1N2O1^OhL[EY3`:f0K5O2N10000000000000kNaEjM^:U2eEiM\\:V2fEgM[:X2hEfMX:Z2iEdMX:[2kEbMV:^2lE_MU:`2nE\\MT:e2mEYMS:h2nEVMR:l2nERMR:o2oEoLQ:S3oEkLQ:V3PFhLP:Z3PFdLP:^3PF_LQ:b3f02O0O2N1O2N1O2N1O2N2N1O2N1O2N1O2N101N1N100O10000000000000000000000000000000O1N2N3M2N2O2L3N2N2M4K4K5K6J5K5L4I8F9H8H9aNPCc0Y=\\OoB5Y=JnBG\\=7g0O102MaRi0"}, "label": "a black bag sitting on a bench"}]}
{"id": 366795, "image": "COCO_train2014_000000366795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is bag beside the banana eating boy\"."}], "task": "refering", "answer_template": "The \"there is bag beside the banana eating boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 316, 317, 318, 319], "bbox": [0.6990000000000001, 0.47920833333333335, 0.915375, 0.7817916666666667], "iscrowd": 0, "area": 14226.262949999993, "rle": {"size": [480, 640], "counts": "gka61o>0O1O2YO2YBOe=:TBFj==TBEi==VBDh=?VBCh=?WBAg=a0XB@f=o0M2N2O2M2N2N3M2N3O2N1O1O2CVN`Ck1]<?M2M3M4M2O1O2N1O2N1O00000000000K5N3N1O1N2O1^OhL[EY3`:f0K5O2N10000000000000kNaEjM^:U2eEiM\\:V2fEgM[:X2hEfMX:Z2iEdMX:[2kEbMV:^2lE_MU:`2nE\\MT:e2mEYMS:h2nEVMR:l2nERMR:o2oEoLQ:S3oEkLQ:V3PFhLP:Z3PFdLP:^3PF_LQ:b3f02O0O2N1O2N1O2N1O2N2N1O2N1O2N1O2N101N1N100O10000000000000000000000000000000O1N2N3M2N2O2L3N2N2M4K4K5K6J5K5L4I8F9H8H9aNPCc0Y=\\OoB5Y=JnBG\\=7g0O102MaRi0"}, "label": "there is bag beside the banana eating boy"}]}
{"id": 366795, "image": "COCO_train2014_000000366795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue coat behind the baby\"."}], "task": "refering", "answer_template": "The \"blue coat behind the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 197, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268], "bbox": [0.510390625, 0.015583333333333334, 0.707140625, 0.6753333333333333], "iscrowd": 0, "area": 29783.579400000006, "rle": {"size": [480, 640], "counts": "g]i4Z1Z<_1aN_1fNZ1L4M3M3M2NO1O1O1O001O1O1O1O100O001O1O1O1O11O001N101O1O001O001N102N2N2N1O2N2XITJX4R6ZKXJc4P6nJZJP5m7N2N2N2N2N2M3N2N1O1O0000000O101O000000001N100000001O00000O101O000000001N100000001O00000O101O000000001O0O1000001O00000OK6G9H8G9G9G9H8G8H9H8G9G9F:YOg0ZOe0ZOg0ZOf0YOg0ZOf0YOg0B>E;F9FTeg2"}, "label": "blue coat behind the baby"}]}
{"id": 366795, "image": "COCO_train2014_000000366795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long blue bag behind a kid ' s hand\"."}], "task": "refering", "answer_template": "The \"a long blue bag behind a kid ' s hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 197, 198, 199, 220, 221, 222, 243, 244, 245, 266, 267, 268], "bbox": [0.510390625, 0.015583333333333334, 0.707140625, 0.6753333333333333], "iscrowd": 0, "area": 29783.579400000006, "rle": {"size": [480, 640], "counts": "g]i4Z1Z<_1aN_1fNZ1L4M3M3M2NO1O1O1O001O1O1O1O100O001O1O1O1O11O001N101O1O001O001N102N2N2N1O2N2XITJX4R6ZKXJc4P6nJZJP5m7N2N2N2N2N2M3N2N1O1O0000000O101O000000001N100000001O00000O101O000000001N100000001O00000O101O000000001O0O1000001O00000OK6G9H8G9G9G9H8G8H9H8G9G9F:YOg0ZOe0ZOg0ZOf0YOg0ZOf0YOg0B>E;F9FTeg2"}, "label": "a long blue bag behind a kid ' s hand"}]}
{"id": 6352, "image": "COCO_train2014_000000006352.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse to the right\"."}], "task": "refering", "answer_template": "The \"horse to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 155, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 224, 244, 247], "bbox": [0.568953125, 0.45647058823529413, 0.8391562499999999, 0.7097647058823529], "iscrowd": 0, "area": 6631.801249999997, "rle": {"size": [425, 640], "counts": "dXg41Y=2SC0\\<2^C1c<:0010O00000O101O0O10001N10000O101N100O10001N100O101N10000O100000001O00001O000YDQOT;o0lDUOP;l0oDXOm:h0SE[Og0Fi8o0aF^O`0Eo8m0aFE7AW9j0bF1HYOe9f0cFf0]9[ObFf0]9ZOdFd0]9\\OcFd0]9\\OcFd0^9[ObFe0^9[ObFTOo:l0g010O000001O01O0001O01O0001O0001O01O000001O000000001O0001O0001O0001O0000001O0001O01O0000001O00000O2O000O101N100O2O000O2O0O1dEcNl8^1RGdNl8]1SGeNl8\\1RGfNl8[1RGgNn8Z1PGfNQ9Z1nFeNT9[1kFeNV9[1iFeNX9[1YFiN^OL[:Z1VFTOk9l0UFROm9n0SFQOo9n0PFROQ:n0oEQOR:o0nEoNU:Q1jEmNX:S1P1O01O000000O100O2M2O1N2N2N2O1O1N2M3N3ZOf0M3O^Cf0V;\\OcDh0^;@UDd0l;<1O1O001O010O001O001O1O7J5J7I6Fd`Z1"}, "label": "horse to the right"}]}
{"id": 6352, "image": "COCO_train2014_000000006352.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white hose with brown spots\"."}], "task": "refering", "answer_template": "The \"a white hose with brown spots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 155, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 224, 244, 247], "bbox": [0.568953125, 0.45647058823529413, 0.8391562499999999, 0.7097647058823529], "iscrowd": 0, "area": 6631.801249999997, "rle": {"size": [425, 640], "counts": "dXg41Y=2SC0\\<2^C1c<:0010O00000O101O0O10001N10000O101N100O10001N100O101N10000O100000001O00001O000YDQOT;o0lDUOP;l0oDXOm:h0SE[Og0Fi8o0aF^O`0Eo8m0aFE7AW9j0bF1HYOe9f0cFf0]9[ObFf0]9ZOdFd0]9\\OcFd0]9\\OcFd0^9[ObFe0^9[ObFTOo:l0g010O000001O01O0001O01O0001O0001O01O000001O000000001O0001O0001O0001O0000001O0001O01O0000001O00000O2O000O101N100O2O000O2O0O1dEcNl8^1RGdNl8]1SGeNl8\\1RGfNl8[1RGgNn8Z1PGfNQ9Z1nFeNT9[1kFeNV9[1iFeNX9[1YFiN^OL[:Z1VFTOk9l0UFROm9n0SFQOo9n0PFROQ:n0oEQOR:o0nEoNU:Q1jEmNX:S1P1O01O000000O100O2M2O1N2N2N2O1O1N2M3N3ZOf0M3O^Cf0V;\\OcDh0^;@UDd0l;<1O1O001O010O001O001O1O7J5J7I6Fd`Z1"}, "label": "a white hose with brown spots"}]}
{"id": 112849, "image": "COCO_train2014_000000112849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow car lifted in the air\"."}], "task": "refering", "answer_template": "The \"the yellow car lifted in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 201, 202, 203, 204, 205], "bbox": [0.63428125, 0.39685446009389674, 0.9969374999999999, 0.6008920187793427], "iscrowd": 0, "area": 11157.081400000005, "rle": {"size": [426, 640], "counts": "^SY54S=6K;F2M3N2N6H5M3M2O1N3N00001O001[D_N_;a1`D_N`;b1^D_Nc;d10O1O1O001N101O001K4Glj:]OoUE3M4M2M3M4I6N3M2O2M2N3N2M3M3M3N8G2N3N1N1O2N101N1O010O1O010O1O00100O01N2N2O1N2O1O1O1O1010O001O00O100000000000000000000000001N10000000000000000000000000000O1000001O000000000000000000000O1000000000001O0000000000000O100000000000000000001O0000000O101O1O1O1O1O1O1O1N2N2N2N2N2N2M3N2N2N2N2N2N2O1N2O1O1O1O1O0001O001O1O1O001O1O1O001O1ON2N200O100O101O000O4L3L5HWd0"}, "label": "the yellow car lifted in the air"}]}
{"id": 112849, "image": "COCO_train2014_000000112849.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow car with the front prompted up\"."}], "task": "refering", "answer_template": "The \"a yellow car with the front prompted up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 201, 202, 203, 204, 205], "bbox": [0.63428125, 0.39685446009389674, 0.9969374999999999, 0.6008920187793427], "iscrowd": 0, "area": 11157.081400000005, "rle": {"size": [426, 640], "counts": "^SY54S=6K;F2M3N2N6H5M3M2O1N3N00001O001[D_N_;a1`D_N`;b1^D_Nc;d10O1O1O001N101O001K4Glj:]OoUE3M4M2M3M4I6N3M2O2M2N3N2M3M3M3N8G2N3N1N1O2N101N1O010O1O010O1O00100O01N2N2O1N2O1O1O1O1010O001O00O100000000000000000000000001N10000000000000000000000000000O1000001O000000000000000000000O1000000000001O0000000000000O100000000000000000001O0000000O101O1O1O1O1O1O1O1N2N2N2N2N2N2M3N2N2N2N2N2N2O1N2O1O1O1O1O0001O001O1O1O001O1O1O001O1ON2N200O100O101O000O4L3L5HWd0"}, "label": "a yellow car with the front prompted up"}]}
{"id": 350421, "image": "COCO_train2014_000000350421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dining chair that is closest to window\"."}], "task": "refering", "answer_template": "The \"dining chair that is closest to window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245], "bbox": [0.464875, 0.3462295081967213, 0.671484375, 0.7419437939110071], "iscrowd": 0, "area": 17723.0705, "rle": {"size": [427, 640], "counts": "V^l3`0Q2l0\\7WO_Hl0a7UO[Hm0f7UOUHn0R7R31N101O1O0O2O1O1N:G8H6I10O10O0100@_LUGa3k8`LUG`3k8aLSG`3l8aLTG^3m8cLQG^3n8cLRG]3m8eLQG\\3o8dLQG[3o8gLoFZ3P9gLPGY3P9gLoFY3Q9a000O010O1O010O100O101N101N101N101N101N10000O100O10000O100O100001O001N101O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001TLlF^3U9`LmF`3T9^LnFa3R9^LPGa3Q9]LPGc3P9[LSGd3[9O001O1O001O11O2M3H7I8H7I7K210O1000O01000O10S1mNn0ROl0SOe0TOiYg2"}, "label": "dining chair that is closest to window"}]}
{"id": 350421, "image": "COCO_train2014_000000350421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fancy chair away from the window\"."}], "task": "refering", "answer_template": "The \"a fancy chair away from the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 250, 251, 252, 270], "bbox": [0.75390625, 0.3491569086651054, 0.967875, 0.7472131147540984], "iscrowd": 0, "area": 15264.591600000005, "rle": {"size": [427, 640], "counts": "nbY6<m<6^OGiCa0`1m0S70_H0_7`0SH@j7Q1hGnNW8T3M31O2UKfGa4c8hKXGj3j8ULVGk3U902N1AiFhLW9T3oFjLQ9V3QGgLP9X3SGfLm8Y3VGeLj8Z3c0100O100AcLoF^3P9dLnF]3R9cLnF]3Q9dLlF_3T9aLkF`3R9bLnF_3Q9>O01000O100O1M3N21O00001O1O0ROQG`Mn8`2UG^Mi8c2]GXMb8h2hGoLW8Q3Q100100O10O10O1O100O1001O00O1001O000000001O00000000001O00100O001O1O001O2N1OO2O0O2O000100O3NO001O010aFQM`8P3TG[Ml8f2PG]MQ9c2kF`MV9_OlF_3U9`LmF^3T9aLnF]3S9bLnF]3R9bLPG\\3a9O1O001O001O00001O0000BiLgFX3V9nLfFS3X9QMeFP3T9XMkFh2U9f01101N3MkKPGo3X9N1O1YLfF[3[9aLiFb0KT2h9iMZF_O11M[2i9SNfF8@k0c;]N_D19S1l;M103TOd`8"}, "label": "a fancy chair away from the window"}]}
{"id": 424152, "image": "COCO_train2014_000000424152.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind another\"."}], "task": "refering", "answer_template": "The \"a zebra behind another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 59, 70, 71, 72, 73, 74, 75, 76, 77, 78, 80, 81, 82, 83, 84, 93, 94, 95, 96, 97, 98, 99, 100, 101, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 148, 166, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 213, 214, 215, 237, 238, 260, 261, 282, 283, 284, 306, 307], "bbox": [0.038984375, 0.13482435597189696, 0.692671875, 0.8988758782201405], "iscrowd": 0, "area": 35784.98249999999, "rle": {"size": [427, 640], "counts": "mb:7k<?A?G9H4M1O1N2O001N2O1O1N2N2M3N2M3N2N2M2O2M3N2M3M3M3M3M3M3L3N3L3N2O2O0O101N100O2O0O101O0O20O1O100O1O010O1O100O10O0100O100O10O0100O100O100O101N101N100O2O0O2O0O2O0O101O010O001O01O100O100O010O100O1O100O010O100O1O100O10O0100O100O1O10O02O0O101N1O101N100O2O0O1O2O0O101N100O2O0O1O2O0O2O1N2O1N2O1O1O100O000001O00000O101OO1O2N3M3M4L3M4L3M4L3M5K4SJWKP3m4dL]KZ3g4ZLcKe3`4PLiKn3l6N1O00000000O100000000000001O00000000000O10000O100O10000O1hLZK\\LOGg4Y3]NPLc1j3lNlKU1m3X4K50000TOiFdMW9W2PGgMP9S2XGkMh8o1`GoM`8l1gGRNY8j1mGTNS8h1SHVNm7f1YHXNg7d1_HZNa7b1eH\\N[7`1kH^NU7^1PIaNP7Z1UIfNk6V1XIkNh6Q1\\IoNd6m0`ISO`6j0cIVO]6g0eIZO[6c0hI]OX6`0kI@U6=nICR6:PJGP66SJJm53VJMj51XJOg50ZJ1f5MZJ5f5I[J8e5F\\J;X92O1O1N2O1N2O1N2O1N200O2O0O1O100O100O100O100O1O100O100O100O100O100O1O2O0O100O10000001O00001O000O101O0O10001O1N101O001NcLPObJo0^5QObJo0^5RObJm0^5SOcJk0^5VOaJj0^5WOcJh0]5YObJf0_5ZObJe0^5\\ObJc0^5]ObJc0^5^ObJ`0^5AcJ>]5CbJ=^5CcJ<]5DcJ;^5FbJ7`5IaJ4`5M`J1b50^JLe54\\JIf57ZJGh5:d30O101O00001O0O101O001O00001N10001O00001N101O001O001N101O001O001N101O1O1O1O1O1O1O100O1O010O0001O001O00001N10001O00001O00001O00001O00001O00001O00001O00001O000O2O00001O001O1O001O1O001O1O001O001O1N1L5L4K_Qb2"}, "label": "a zebra behind another"}]}
{"id": 424152, "image": "COCO_train2014_000000424152.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the partially obscured zebra\"."}], "task": "refering", "answer_template": "The \"the partially obscured zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 59, 70, 71, 72, 73, 74, 75, 76, 77, 78, 80, 81, 82, 83, 84, 93, 94, 95, 96, 97, 98, 99, 100, 101, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 148, 166, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 213, 214, 215, 237, 238, 260, 261, 282, 283, 284, 306, 307], "bbox": [0.038984375, 0.13482435597189696, 0.692671875, 0.8988758782201405], "iscrowd": 0, "area": 35784.98249999999, "rle": {"size": [427, 640], "counts": "mb:7k<?A?G9H4M1O1N2O001N2O1O1N2N2M3N2M3N2N2M2O2M3N2M3M3M3M3M3M3L3N3L3N2O2O0O101N100O2O0O101O0O20O1O100O1O010O1O100O10O0100O100O10O0100O100O100O101N101N100O2O0O2O0O2O0O101O010O001O01O100O100O010O100O1O100O010O100O1O100O10O0100O100O1O10O02O0O101N1O101N100O2O0O1O2O0O101N100O2O0O1O2O0O2O1N2O1N2O1O1O100O000001O00000O101OO1O2N3M3M4L3M4L3M4L3M5K4SJWKP3m4dL]KZ3g4ZLcKe3`4PLiKn3l6N1O00000000O100000000000001O00000000000O10000O100O10000O1hLZK\\LOGg4Y3]NPLc1j3lNlKU1m3X4K50000TOiFdMW9W2PGgMP9S2XGkMh8o1`GoM`8l1gGRNY8j1mGTNS8h1SHVNm7f1YHXNg7d1_HZNa7b1eH\\N[7`1kH^NU7^1PIaNP7Z1UIfNk6V1XIkNh6Q1\\IoNd6m0`ISO`6j0cIVO]6g0eIZO[6c0hI]OX6`0kI@U6=nICR6:PJGP66SJJm53VJMj51XJOg50ZJ1f5MZJ5f5I[J8e5F\\J;X92O1O1N2O1N2O1N2O1N200O2O0O1O100O100O100O100O1O100O100O100O100O100O1O2O0O100O10000001O00001O000O101O0O10001O1N101O001NcLPObJo0^5QObJo0^5RObJm0^5SOcJk0^5VOaJj0^5WOcJh0]5YObJf0_5ZObJe0^5\\ObJc0^5]ObJc0^5^ObJ`0^5AcJ>]5CbJ=^5CcJ<]5DcJ;^5FbJ7`5IaJ4`5M`J1b50^JLe54\\JIf57ZJGh5:d30O101O00001O0O101O001O00001N10001O00001N101O001O001N101O001O001N101O1O1O1O1O1O1O100O1O010O0001O001O00001N10001O00001O00001O00001O00001O00001O00001O00001O000O2O00001O001O1O001O1O001O1O001O001O1N1L5L4K_Qb2"}, "label": "the partially obscured zebra"}]}
{"id": 424152, "image": "COCO_train2014_000000424152.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra has full view\"."}], "task": "refering", "answer_template": "The \"a zebra has full view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 78, 79, 80, 81, 82, 87, 88, 89, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 224, 226, 227, 228, 243, 244, 245, 249, 250, 251, 266, 267, 268, 272, 273, 274, 289, 290, 291, 294, 295, 296, 312, 313, 314, 319], "bbox": [0.37482812499999996, 0.190327868852459, 0.953546875, 0.9341451990632318], "iscrowd": 0, "area": 58309.41905000001, "rle": {"size": [427, 640], "counts": "bXT3b0d<k0UO;F4K4L4L5K4M3L5K4M4K4M3L5K4M4L300O2O0O100O100O00100O100O010001O000O2O001O000010O01O01O010O010O0010O00101N101N100O2N100O2O0O101N1O101O01O010O1O010O1O10O01O10O01O10O0100O1O010O10O01O100O010O10O01O10O02O1N2N2O1N2O1N2N2O1N2O2M2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N101N3N4iJeJT2_5dMeJ]2^5ZMgJf2^5QMfJP3]5gLiJX3\\5_LhJb3[5ULjJk3[5kKkJU4X5cKlJ]4Y5YKlJh4W5PKmJP5h60O001O0O2N1O2N1O2N1O2M2O1O1O1]NhJRKY5a4]KUKd4h4iKmJXOAW4g5fL]JPO0Z4i5VM\\Ji2j5mLZJS3k5dLYJ[3n5[LVJe3o5RLVJm3P6iKTJW4R6_KRJa4S6VKQJj4Y7O0000001N1000001ObK]K3d4MhKD[4<RLTOQ4l0\\LeNf3[1lMPMW2P3QNeLQ2\\3XNXLk1h3^NmKd1S4j30000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000O101O000O1000000O100000000O1000000O100000O01000O10O010000O10O10O1000O0100O10O10O1000O0100OjJdK_1\\4_NfKa1Z4]NhKb1X4\\NkKd1V4YNlKg1U4UNnKk1S4RNoKn1R4nMRLQ2o3lMSLT2n3iMTLW2m3eMVL[2l3aMVL_2k3]MXLb2j3[MXLe2i3XMYLh2h3TM[Ll2f3QM\\Lo2e3mL^LS3c3jL_LV3c3eL`L[3a3bLaL^3`3_LbLa3_3[LaLh3`3UL_Ln3b3nK^LU4c3hK]LY4e3cK[L`4f3\\KZLg4h3UKWLn4j3nJVLU5k3gJUL\\5l3aJSLa0`NR4^5YKRLf0bNQ4]5UKQLk0dNP4\\5RKoKo0hNn3h6SLZIm3c6TL_Il3_6TLcIl3\\6TLeIk3Z6ULiIj3V6ULlIk3S6TLoIl3o5TLSJl3l5SLVJm3i5RLYJn3e5RL^Jm3a5RLaJn3^5QLdJo3Z5QLhJo3W5PLkJP4T5oKoJP4P5PLQKP4m4PLTKP4l4iK\\KX4b4ZKlKh4Z62N2N2N101N2N2N2N2J6H8I7I7H8I7I7`MSIZNT7b1VIQNQ7k1YIhMn6T2\\I`Mj6\\2`IWMf6f2cInLd6o2n1L4L4L4L4L4K5L7I7I8H7I8G8I7I8H7IS[<"}, "label": "a zebra has full view"}]}
{"id": 424152, "image": "COCO_train2014_000000424152.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in front and closer to the right side of picture\"."}], "task": "refering", "answer_template": "The \"zebra in front and closer to the right side of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 78, 79, 80, 81, 82, 87, 88, 89, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 224, 226, 227, 228, 243, 244, 245, 249, 250, 251, 266, 267, 268, 272, 273, 274, 289, 290, 291, 294, 295, 296, 312, 313, 314, 319], "bbox": [0.37482812499999996, 0.190327868852459, 0.953546875, 0.9341451990632318], "iscrowd": 0, "area": 58309.41905000001, "rle": {"size": [427, 640], "counts": "bXT3b0d<k0UO;F4K4L4L5K4M3L5K4M4K4M3L5K4M4L300O2O0O100O100O00100O100O010001O000O2O001O000010O01O01O010O010O0010O00101N101N100O2N100O2O0O101N1O101O01O010O1O010O1O10O01O10O01O10O0100O1O010O10O01O100O010O10O01O10O02O1N2N2O1N2O1N2N2O1N2O2M2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N101N3N4iJeJT2_5dMeJ]2^5ZMgJf2^5QMfJP3]5gLiJX3\\5_LhJb3[5ULjJk3[5kKkJU4X5cKlJ]4Y5YKlJh4W5PKmJP5h60O001O0O2N1O2N1O2N1O2M2O1O1O1]NhJRKY5a4]KUKd4h4iKmJXOAW4g5fL]JPO0Z4i5VM\\Ji2j5mLZJS3k5dLYJ[3n5[LVJe3o5RLVJm3P6iKTJW4R6_KRJa4S6VKQJj4Y7O0000001N1000001ObK]K3d4MhKD[4<RLTOQ4l0\\LeNf3[1lMPMW2P3QNeLQ2\\3XNXLk1h3^NmKd1S4j30000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000O101O000O1000000O100000000O1000000O100000O01000O10O010000O10O10O1000O0100O10O10O1000O0100OjJdK_1\\4_NfKa1Z4]NhKb1X4\\NkKd1V4YNlKg1U4UNnKk1S4RNoKn1R4nMRLQ2o3lMSLT2n3iMTLW2m3eMVL[2l3aMVL_2k3]MXLb2j3[MXLe2i3XMYLh2h3TM[Ll2f3QM\\Lo2e3mL^LS3c3jL_LV3c3eL`L[3a3bLaL^3`3_LbLa3_3[LaLh3`3UL_Ln3b3nK^LU4c3hK]LY4e3cK[L`4f3\\KZLg4h3UKWLn4j3nJVLU5k3gJUL\\5l3aJSLa0`NR4^5YKRLf0bNQ4]5UKQLk0dNP4\\5RKoKo0hNn3h6SLZIm3c6TL_Il3_6TLcIl3\\6TLeIk3Z6ULiIj3V6ULlIk3S6TLoIl3o5TLSJl3l5SLVJm3i5RLYJn3e5RL^Jm3a5RLaJn3^5QLdJo3Z5QLhJo3W5PLkJP4T5oKoJP4P5PLQKP4m4PLTKP4l4iK\\KX4b4ZKlKh4Z62N2N2N101N2N2N2N2J6H8I7I7H8I7I7`MSIZNT7b1VIQNQ7k1YIhMn6T2\\I`Mj6\\2`IWMf6f2cInLd6o2n1L4L4L4L4L4K5L7I7I8H7I8G8I7I8H7IS[<"}, "label": "zebra in front and closer to the right side of picture"}]}
{"id": 235744, "image": "COCO_train2014_000000235744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee on a table next to a birthday cake\"."}], "task": "refering", "answer_template": "The \"a cup of coffee on a table next to a birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210], "bbox": [0.008015625, 0.28089715536105037, 0.197375, 0.597746170678337], "iscrowd": 0, "area": 14035.27725, "rle": {"size": [457, 640], "counts": "V\\28o=d0]Od0\\Od0\\Od0\\O?A3M3M3M3M3M3L4N2M3N0O10001N100O10000O101N10000O10001N1000000000001O000O100000001O00000000000O10000000O100000000000O11O0O101O00001O00001O0O101O00001O0O2N100O2N100O2O1N2O1N3N1O1N3N1O1PMeEo1^:bMPF]2o:O1N2O1O1N2O1O1N101N2N2N2N2N2N2N1N4M5K6J5I7E;DPVU7"}, "label": "a cup of coffee on a table next to a birthday cake"}]}
{"id": 235744, "image": "COCO_train2014_000000235744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the coffee mug to the left of the cake\"."}], "task": "refering", "answer_template": "The \"the coffee mug to the left of the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210], "bbox": [0.008015625, 0.28089715536105037, 0.197375, 0.597746170678337], "iscrowd": 0, "area": 14035.27725, "rle": {"size": [457, 640], "counts": "V\\28o=d0]Od0\\Od0\\Od0\\O?A3M3M3M3M3M3L4N2M3N0O10001N100O10000O101N10000O10001N1000000000001O000O100000001O00000000000O10000000O100000000000O11O0O101O00001O00001O0O101O00001O0O2N100O2N100O2O1N2O1N3N1O1N3N1O1PMeEo1^:bMPF]2o:O1N2O1O1N2O1O1N101N2N2N2N2N2N2N1N4M5K6J5I7E;DPVU7"}, "label": "the coffee mug to the left of the cake"}]}
{"id": 522465, "image": "COCO_train2014_000000522465.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table behind pizza box\"."}], "task": "refering", "answer_template": "The \"table behind pizza box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 64, 65, 66, 84, 85, 86, 87, 88, 89, 110, 111, 112], "bbox": [0.598625, 0.07345833333333333, 0.916796875, 0.2983125], "iscrowd": 0, "area": 12297.151499999998, "rle": {"size": [480, 640], "counts": "VRd53j>3N2M3M3N2M3O1O1000000000000000000O100000001O001O001O0O101O001O001O001O001N10001O001O001O001O0O2O00001O001O001O001N101O00001O001O001N101O001O001O00001O0O2O001O001O001O00001N101O001O001O001O000O2O001O001O001O001O0O101O001O001O001O001N101O00001O001O001O0O2O001O00001O001O001N101O001O00001O001O0O2O001O001O00001O001N101O001O001O001O000O2O00000000001O00000O100001O2O0O2N1O2N1O1FSD_Mm;\\2YDdMg;U2aDiM`;Q2fDnM[;l1lDRNT;i1U1J6J6J5K6J6J5K6J5K5K^ih0"}, "label": "table behind pizza box"}]}
{"id": 522465, "image": "COCO_train2014_000000522465.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table to the right\"."}], "task": "refering", "answer_template": "The \"the table to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 64, 65, 66, 84, 85, 86, 87, 88, 89, 110, 111, 112], "bbox": [0.598625, 0.07345833333333333, 0.916796875, 0.2983125], "iscrowd": 0, "area": 12297.151499999998, "rle": {"size": [480, 640], "counts": "VRd53j>3N2M3M3N2M3O1O1000000000000000000O100000001O001O001O0O101O001O001O001O001N10001O001O001O001O0O2O00001O001O001O001N101O00001O001O001N101O001O001O00001O0O2O001O001O001O00001N101O001O001O001O000O2O001O001O001O001O0O101O001O001O001O001N101O00001O001O001O0O2O001O00001O001O001N101O001O00001O001O0O2O001O001O00001O001N101O001O001O001O000O2O00000000001O00000O100001O2O0O2N1O2N1O1FSD_Mm;\\2YDdMg;U2aDiM`;Q2fDnM[;l1lDRNT;i1U1J6J6J5K6J6J5K6J5K5K^ih0"}, "label": "the table to the right"}]}
{"id": 522465, "image": "COCO_train2014_000000522465.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picnic table\"."}], "task": "refering", "answer_template": "The \"a picnic table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 138, 161, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 300, 320, 321, 322, 323, 341, 342, 343, 344, 345, 346, 362, 363, 364, 365, 366, 367, 368, 369, 370, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.24414583333333334, 0.989609375, 0.9870208333333333], "iscrowd": 0, "area": 21948.177699999997, "rle": {"size": [480, 640], "counts": "^4U2i<X4iKj3UL2N2N2N2N2N2N3M2N2O1N25K7I8H8H8H8H8H8H7I8H8H8H8H8H7I8H8H8I7H8H8H7I8H8H8H8H8H7I8H8H8H8H8H8H7I8H8H8H8I7H8H7I8H8H8HgPk5:onTJ4K100O100O100O100O100O100O00100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O10O0100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O010O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O10O0100O100O100O100O100O100O1O100O100O100O100OG:C<C>C<C>B=Dnm2"}, "label": "a picnic table"}]}
{"id": 522465, "image": "COCO_train2014_000000522465.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tablecloth\"."}], "task": "refering", "answer_template": "The \"tablecloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 138, 161, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 300, 320, 321, 322, 323, 341, 342, 343, 344, 345, 346, 362, 363, 364, 365, 366, 367, 368, 369, 370, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.24414583333333334, 0.989609375, 0.9870208333333333], "iscrowd": 0, "area": 21948.177699999997, "rle": {"size": [480, 640], "counts": "^4U2i<X4iKj3UL2N2N2N2N2N2N3M2N2O1N25K7I8H8H8H8H8H8H7I8H8H8H8H8H7I8H8H8I7H8H8H7I8H8H8H8H8H7I8H8H8H8H8H8H7I8H8H8H8I7H8H7I8H8H8HgPk5:onTJ4K100O100O100O100O100O100O00100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O10O0100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O010O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O10O0100O100O100O100O100O100O1O100O100O100O100OG:C<C>C<C>B=Dnm2"}, "label": "tablecloth"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown vinyl chair on wooden floor\"."}], "task": "refering", "answer_template": "The \"brown vinyl chair on wooden floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291], "bbox": [0.5065000000000001, 0.4259791666666667, 0.68353125, 0.7690625], "iscrowd": 0, "area": 13257.443650000001, "rle": {"size": [480, 640], "counts": "WVh44\\>?B>A?Bl0ROT1hNe0[Od0O20O1O2K4M3M3M3N2000000000000000000000000001O000000000000000000000001O0001O00000000000000000000000000001O00000000000000000000000001O01O00000000000000000000000eEgKo9Y4oEkKo9U4nEQL^:a3^EdLP;`3?A4H3UOhDkMZ;T2iDgM[;W2h0O1O001O01O01O000TNm1O00001O0O3N2M2N]ln2"}, "label": "brown vinyl chair on wooden floor"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan , leather rocking chair\"."}], "task": "refering", "answer_template": "The \"a tan , leather rocking chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291], "bbox": [0.5065000000000001, 0.4259791666666667, 0.68353125, 0.7690625], "iscrowd": 0, "area": 13257.443650000001, "rle": {"size": [480, 640], "counts": "WVh44\\>?B>A?Bl0ROT1hNe0[Od0O20O1O2K4M3M3M3N2000000000000000000000000001O000000000000000000000001O0001O00000000000000000000000000001O00000000000000000000000001O01O00000000000000000000000eEgKo9Y4oEkKo9U4nEQL^:a3^EdLP;`3?A4H3UOhDkMZ;T2iDgM[;W2h0O1O001O01O01O000TNm1O00001O0O3N2M2N]ln2"}, "label": "a tan , leather rocking chair"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray couch with a gray pillow with multicolors\"."}], "task": "refering", "answer_template": "The \"a gray couch with a gray pillow with multicolors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 320, 321, 342, 343, 344, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.761765625, 0.6965833333333333, 1.0, 0.9998333333333334], "iscrowd": 0, "area": 9903.560200000002, "rle": {"size": [480, 640], "counts": "mVU72l>2N2O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O2N1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O10000001O0000001O00000000L4O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1\\OhL`EX3`:iL^EX3a:iL^EX3a:jL]EW3b:kL\\EV3c:lL[EU3e:kLZEV3e:lLYEU3f:mLXET3f:nLYES3d:QMZEP3d:SM[Em2c:i0O100O1O100O1O100"}, "label": "a gray couch with a gray pillow with multicolors"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the edge of a couch with polka dotted pillows\"."}], "task": "refering", "answer_template": "The \"the edge of a couch with polka dotted pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 320, 321, 342, 343, 344, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.761765625, 0.6965833333333333, 1.0, 0.9998333333333334], "iscrowd": 0, "area": 9903.560200000002, "rle": {"size": [480, 640], "counts": "mVU72l>2N2O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O2N1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O10000001O0000001O00000000L4O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1\\OhL`EX3`:iL^EX3a:iL^EX3a:jL]EW3b:kL\\EV3c:lL[EU3e:kLZEV3e:lLYEU3f:mLXET3f:nLYES3d:QMZEP3d:SM[Em2c:i0O100O1O100O1O100"}, "label": "the edge of a couch with polka dotted pillows"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather chair beside a fireplace , at the corner of an area rug\"."}], "task": "refering", "answer_template": "The \"a brown leather chair beside a fireplace , at the corner of an area rug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291], "bbox": [0.5103125000000001, 0.43389583333333337, 0.67690625, 0.7573541666666667], "iscrowd": 0, "area": 12885.677799999996, "rle": {"size": [480, 640], "counts": "ibi4c0h=f0ROm0jCoMT;[2cDlMT;^2eDgMS;W3H7H9J52O1O1N2O10000001O000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000002N1O2N2N1O2NgLQFc1m9aMoF`2i:K5I6K6I7O0100O10O10OI8G9F9nNXCDQ=9l0M4M3L_hP3"}, "label": "a brown leather chair beside a fireplace , at the corner of an area rug"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty orange chair facing this way\"."}], "task": "refering", "answer_template": "The \"an empty orange chair facing this way\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291], "bbox": [0.5103125000000001, 0.43389583333333337, 0.67690625, 0.7573541666666667], "iscrowd": 0, "area": 12885.677799999996, "rle": {"size": [480, 640], "counts": "ibi4c0h=f0ROm0jCoMT;[2cDlMT;^2eDgMS;W3H7H9J52O1O1N2O10000001O000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000002N1O2N2N1O2NgLQFc1m9aMoF`2i:K5I6K6I7O0100O10O10OI8G9F9nNXCDQ=9l0M4M3L_hP3"}, "label": "an empty orange chair facing this way"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden rocking chair\"."}], "task": "refering", "answer_template": "The \"a wooden rocking chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 204, 205, 206, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 364], "bbox": [0.6979375, 0.44333333333333336, 0.9999375, 0.9157083333333333], "iscrowd": 0, "area": 19766.78965, "rle": {"size": [480, 640], "counts": "Qma66j>1O1O001O001O0O2M2M4L3L4M4L3L5L3M4L3L4M4L3M3L4M3L4M3M4TObMPEb2Q;_MjDd2V;^M_Dk2a;WM\\Dj2d;91O01O0001O00000010O000001O0000010O0000001O000001O01O0O1O101N1O1O110O01O1O010O001O10O01O00100O001kDmL[:S3cEoL^:Q3_ERM`:o2]ESMc:m2[EUMe:l2XEWMg:i2WEYMj:g2SE[Mm:f2PE]Mo:Z3001M2M3L4M4K4M3M0011N3M3M2N3M3M3M2N3M3M2N3M3M3M2N3N2M2N3M3M3M2N3M3M2N3M3M3M2N3M3N2M2011O0O2O001O`0_O100O2L3M3M3M3M3M3L5L3M3M3M3M3M3M4L3M3M3M3M3N2N3M2N2M3N2N2N3M2N2M3N3M2N2N3M2N2M3N3M2N2N3M2N2M3N2N2N2N2N2N2M3ClH"}, "label": "a wooden rocking chair"}]}
{"id": 424161, "image": "COCO_train2014_000000424161.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown wicker rocking chair\"."}], "task": "refering", "answer_template": "The \"a light brown wicker rocking chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 204, 205, 206, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 364], "bbox": [0.6979375, 0.44333333333333336, 0.9999375, 0.9157083333333333], "iscrowd": 0, "area": 19766.78965, "rle": {"size": [480, 640], "counts": "Qma66j>1O1O001O001O0O2M2M4L3L4M4L3L5L3M4L3L4M4L3M3L4M3L4M3M4TObMPEb2Q;_MjDd2V;^M_Dk2a;WM\\Dj2d;91O01O0001O00000010O000001O0000010O0000001O000001O01O0O1O101N1O1O110O01O1O010O001O10O01O00100O001kDmL[:S3cEoL^:Q3_ERM`:o2]ESMc:m2[EUMe:l2XEWMg:i2WEYMj:g2SE[Mm:f2PE]Mo:Z3001M2M3L4M4K4M3M0011N3M3M2N3M3M3M2N3M3M2N3M3M3M2N3N2M2N3M3M3M2N3M3M2N3M3M3M2N3M3N2M2011O0O2O001O`0_O100O2L3M3M3M3M3M3L5L3M3M3M3M3M3M4L3M3M3M3M3N2N3M2N2M3N2N2N3M2N2M3N3M2N2N3M2N2M3N3M2N2N3M2N2M3N2N2N2N2N2N2M3ClH"}, "label": "a light brown wicker rocking chair"}]}
{"id": 153827, "image": "COCO_train2014_000000153827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in glasses chasing someone\"."}], "task": "refering", "answer_template": "The \"woman in glasses chasing someone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 65, 66, 67, 77, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 106, 107, 118, 119, 120, 121, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 173, 174, 175, 176, 192, 193, 194, 210, 211, 212], "bbox": [0.28136, 0.21123123123123125, 0.9832799999999999, 0.9887687687687687], "iscrowd": 0, "area": 24344.00925, "rle": {"size": [333, 500], "counts": "kY^11\\:101N1O100O1O100O1O100O1O2O0O001O010O001O001O010O001O0000O1O100O1O100O101N1O100O11O00001O00001O001O01O01O001O00001O001O00001O01O01O00O10000000000O010000000O1000O100000O100000O1000O10000000O10O100000000O10O100O100O010O100O100O100O2O0O101N100O2O0O2O0O101N100O2O0O101O0O3N1N2O2N1N2O2M2O1O1N3N1N2O2N1SO[NoHg1h6gNRI[1j6mNoHU1P7QOjHP1T7W1M2O1N3RJfLe4[3SKnLj4T3oJSMP5o2hJYMU5i2dJ^M[5l3N1N3N2M2O2N1N3N2M2O2N1O2N101N2O0O2O0O2O0O2O1N101N101N2O0O2O0O2O15J<EN2N1O2O0O2N2N1O2N2O0O2N1O2N2N101NlJXLd3g3\\L[Ld3c3[L_Le3`3[LaLf3]3YLeLg3[3XLfLh3Y3WLjLh3U3XLlLh3S3WLoLj3o2VLSMi3l2VLWMi3i2VLZMg3f2XL]Mf3c2ZL`Mc3`2\\LcMb3]2^LeM`3[2`LgM^3Y2bLhM]3Y2aLiM^3W2bLkM\\3U2dLlM[3T2dLoMZ3Q2fLPNY3P2gLQNX3P2gLRNW3n1hLTNW3l1iLUNV3k1jLWNT3i1lLXNS3h1lLZNS3g1lL[NR3e1nL\\NQ3d1oL^No2b1PM`No2`1QMaNn2`1QMbNm2^1SMcNl2]1SMeNl2\\1SMfNj2[1VMfNi2[1VMfNi2Z1VMiNh2W1XMjNg2W1XMkNf2U1ZMlNe2U1ZMoNb2Q1^MRO_2n0aMUO\\2l0cMWOZ2i0fMZOW2g0hM\\OU2d0kM_OR2b0mMAP2?PNDm1<SNMd14[N6[1JeN`0Q1AnNj0g0VOYOT1=lNC^13cNLc1N]N2f1K[N4h1IXN6m1FSN:P2CQN<P600000001O000000000001O0010O01O001O001O0010O01O001O001O001O010O001O001O001O01O010O001O001O00010O001O001O000010OO2O001O001O00001O001N101O0O1M4L3Nm^2"}, "label": "woman in glasses chasing someone"}]}
{"id": 153827, "image": "COCO_train2014_000000153827.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black shirt and glasses\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black shirt and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 65, 66, 67, 77, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 106, 107, 118, 119, 120, 121, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 173, 174, 175, 176, 192, 193, 194, 210, 211, 212], "bbox": [0.28136, 0.21123123123123125, 0.9832799999999999, 0.9887687687687687], "iscrowd": 0, "area": 24344.00925, "rle": {"size": [333, 500], "counts": "kY^11\\:101N1O100O1O100O1O100O1O2O0O001O010O001O001O010O001O0000O1O100O1O100O101N1O100O11O00001O00001O001O01O01O001O00001O001O00001O01O01O00O10000000000O010000000O1000O100000O100000O1000O10000000O10O100000000O10O100O100O010O100O100O100O2O0O101N100O2O0O2O0O101N100O2O0O101O0O3N1N2O2N1N2O2M2O1O1N3N1N2O2N1SO[NoHg1h6gNRI[1j6mNoHU1P7QOjHP1T7W1M2O1N3RJfLe4[3SKnLj4T3oJSMP5o2hJYMU5i2dJ^M[5l3N1N3N2M2O2N1N3N2M2O2N1O2N101N2O0O2O0O2O0O2O1N101N101N2O0O2O0O2O15J<EN2N1O2O0O2N2N1O2N2O0O2N1O2N2N101NlJXLd3g3\\L[Ld3c3[L_Le3`3[LaLf3]3YLeLg3[3XLfLh3Y3WLjLh3U3XLlLh3S3WLoLj3o2VLSMi3l2VLWMi3i2VLZMg3f2XL]Mf3c2ZL`Mc3`2\\LcMb3]2^LeM`3[2`LgM^3Y2bLhM]3Y2aLiM^3W2bLkM\\3U2dLlM[3T2dLoMZ3Q2fLPNY3P2gLQNX3P2gLRNW3n1hLTNW3l1iLUNV3k1jLWNT3i1lLXNS3h1lLZNS3g1lL[NR3e1nL\\NQ3d1oL^No2b1PM`No2`1QMaNn2`1QMbNm2^1SMcNl2]1SMeNl2\\1SMfNj2[1VMfNi2[1VMfNi2Z1VMiNh2W1XMjNg2W1XMkNf2U1ZMlNe2U1ZMoNb2Q1^MRO_2n0aMUO\\2l0cMWOZ2i0fMZOW2g0hM\\OU2d0kM_OR2b0mMAP2?PNDm1<SNMd14[N6[1JeN`0Q1AnNj0g0VOYOT1=lNC^13cNLc1N]N2f1K[N4h1IXN6m1FSN:P2CQN<P600000001O000000000001O0010O01O001O001O0010O01O001O001O001O010O001O001O001O01O010O001O001O00010O001O001O000010OO2O001O001O00001O001N101O0O1M4L3Nm^2"}, "label": "a woman wearing a black shirt and glasses"}]}
{"id": 276711, "image": "COCO_train2014_000000276711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a glass of orange juice\"."}], "task": "refering", "answer_template": "The \"a woman holding a glass of orange juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 144, 145, 166, 167, 168, 189, 190, 191, 278, 301, 303, 325, 326, 348, 349], "bbox": [0.08445312499999999, 0.239875, 0.356421875, 0.9245416666666666], "iscrowd": 0, "area": 6238.285600000002, "rle": {"size": [480, 640], "counts": "Uei05h>:E:G4L4N2N3M2N2M3N2N3M2O0010O010O010Od1]Nhk92lSF;M3L3N3M3L3N3M3M3L3N3L4L3N3L3O1O0000gNPCb0P=YOXCd0h<WO`Cf0`<VOgCg0\\=M3M_Ta0_OQl^O4K4L4L5L3L4L5K4M3L5K4L4M4K4L4L2O2K5G8H9H8G7I8N2N2O0O2O1O1N101N2O0O2O1O1O010O1O00100O1OlM[ENe:BoE;Q:TOcFi0\\9gNXGU1];M3L4M3L4M3L3N3L4M3L4M3L101N2O001N101N2O0O2OblP6"}, "label": "a woman holding a glass of orange juice"}]}
{"id": 276711, "image": "COCO_train2014_000000276711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl that is holding a class of orange juice\"."}], "task": "refering", "answer_template": "The \"the girl that is holding a class of orange juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 144, 145, 166, 167, 168, 189, 190, 191, 278, 301, 303, 325, 326, 348, 349], "bbox": [0.08445312499999999, 0.239875, 0.356421875, 0.9245416666666666], "iscrowd": 0, "area": 6238.285600000002, "rle": {"size": [480, 640], "counts": "Uei05h>:E:G4L4N2N3M2N2M3N2N3M2O0010O010O010Od1]Nhk92lSF;M3L3N3M3L3N3M3M3L3N3L4L3N3L3O1O0000gNPCb0P=YOXCd0h<WO`Cf0`<VOgCg0\\=M3M_Ta0_OQl^O4K4L4L5L3L4L5K4M3L5K4L4M4K4L4L2O2K5G8H9H8G7I8N2N2O0O2O1O1N101N2O0O2O1O1O010O1O00100O1OlM[ENe:BoE;Q:TOcFi0\\9gNXGU1];M3L4M3L4M3L3N3L4M3L4M3L101N2O001N101N2O0O2OblP6"}, "label": "the girl that is holding a class of orange juice"}]}
{"id": 276711, "image": "COCO_train2014_000000276711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with the curly hair\"."}], "task": "refering", "answer_template": "The \"the girl with the curly hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.18556250000000002, 0.11654166666666667, 0.638625, 1.0], "iscrowd": 0, "area": 81421.81100000002, "rle": {"size": [480, 640], "counts": "^Uh19d><D<D;E<D<Dl0TO3M3M3M4L3M3M3M3L4M3L4L4L4L4K5L5K4L4K5L4L4L4K5L4L4L4K6L300O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O2O0O10000O100O100O100O100dLRIeNo6U1cI_N]6[1UJXNl5b1fJRNZ5h1nJUNS5i1RKTNn4j1WKRNj4l1ZKRNf4l1_KPNb4m1dKPN\\4n1iKnMX4P2lKnMT4P2QLlMP4R2ULkMk3S2YLjMh3T2]LiMc3T2cLhM^3V2gLgMY3W2kLfMU3Y2nLeMS3Y2oLgMQ3W2QMhMP3X2PMhMP3W2RMgMo2Y2QMgMo2Y2QMfMP3Y2QMgMo2Y2RMeMo2Z2RMfMn2Z2RMeMo2[2QMeMo2Z2SMdMn2\\2RMdMn2[2SMdMn2\\2RMdMn2\\2RMcMo2\\2SMcMm2]2SMbMn2]2SMcMm2]2SMbMn2^2j4O10O0100O10000O100O10000O100O10000O100O10000O100O10O10O100O100O10000O100O10000O100O10000O100O10000O2O0O10000001O00000000001O000000001O00000000001O000001O01O00000ZHSNY2m1eMVNZ2j1eMYNY2h1eMZNZ2f1eM]NY2c1eM`NZ2`1fMbNX2^1gMeNW2\\1hMeNW2[1iMgNU2Y1kMhNT2X1kMkNS2U1mMlNR2T1nMnNP2S1oMnNP2S1oMnNP2S1nMoNQ2R1nMoNQ2R1nMoNQ2R1nMoNQ2R1mMPOR2Q1lMQOS2P1lMQOS2P1kMROT2o0jMSOU2n0iMTOV2m0hMUOW2l0gMVOX2k0gMVOX2k0fMWOh0[N_L_2g2YOf0[NbL]2f2ZOe0[NeL\\2d2[Oc0]NhLY2c2\\Oc0]NiLX2b2]Od0[NjLZ2h14]1cMjLZ2f16^1bMkLY2d18`1_MlLZ2a1:b1]MlLZ2_1<c1\\MmLY2]1>e1YMnLZ2Y1a0h1VMnLZ2W1c0j1TMnLZ2U1e0k1RMoL[2S1f0m1PMoL[2Q1h0o1nLoL[2o0j0P2lLQM[2l0m0Q2iLRM[2i0T1o1cLWMZ2NRM@Z4V3ZL\\M[2EdM[On3_3UL`Mi8[2WGeMn8U2RGlMR9P2nFPNW9j1iFWN\\9d1dF\\Na9^1`FbNd9Z1[FgNk9R1VFnNR:j0mEWO\\:?eEAc:7\\EIm:MUE3S;CoD=Y;YOhDh0`;nNbDR1e<10000O1O1O1O1O]O\\BLc=4`BJ_=6cBI]=6fBHY=8jBFU=:mBES=:PCCR=:RCDP=9SCEo<8SCGo<6TCHm<6Q1M3MWk[3"}, "label": "the girl with the curly hair"}]}
{"id": 276711, "image": "COCO_train2014_000000276711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with curly hair playing wii\"."}], "task": "refering", "answer_template": "The \"a woman with curly hair playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.18556250000000002, 0.11654166666666667, 0.638625, 1.0], "iscrowd": 0, "area": 81421.81100000002, "rle": {"size": [480, 640], "counts": "^Uh19d><D<D;E<D<Dl0TO3M3M3M4L3M3M3M3L4M3L4L4L4L4K5L5K4L4K5L4L4L4K5L4L4L4K6L300O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O2O0O10000O100O100O100O100dLRIeNo6U1cI_N]6[1UJXNl5b1fJRNZ5h1nJUNS5i1RKTNn4j1WKRNj4l1ZKRNf4l1_KPNb4m1dKPN\\4n1iKnMX4P2lKnMT4P2QLlMP4R2ULkMk3S2YLjMh3T2]LiMc3T2cLhM^3V2gLgMY3W2kLfMU3Y2nLeMS3Y2oLgMQ3W2QMhMP3X2PMhMP3W2RMgMo2Y2QMgMo2Y2QMfMP3Y2QMgMo2Y2RMeMo2Z2RMfMn2Z2RMeMo2[2QMeMo2Z2SMdMn2\\2RMdMn2[2SMdMn2\\2RMdMn2\\2RMcMo2\\2SMcMm2]2SMbMn2]2SMcMm2]2SMbMn2^2j4O10O0100O10000O100O10000O100O10000O100O10000O100O10O10O100O100O10000O100O10000O100O10000O100O10000O2O0O10000001O00000000001O000000001O00000000001O000001O01O00000ZHSNY2m1eMVNZ2j1eMYNY2h1eMZNZ2f1eM]NY2c1eM`NZ2`1fMbNX2^1gMeNW2\\1hMeNW2[1iMgNU2Y1kMhNT2X1kMkNS2U1mMlNR2T1nMnNP2S1oMnNP2S1oMnNP2S1nMoNQ2R1nMoNQ2R1nMoNQ2R1nMoNQ2R1mMPOR2Q1lMQOS2P1lMQOS2P1kMROT2o0jMSOU2n0iMTOV2m0hMUOW2l0gMVOX2k0gMVOX2k0fMWOh0[N_L_2g2YOf0[NbL]2f2ZOe0[NeL\\2d2[Oc0]NhLY2c2\\Oc0]NiLX2b2]Od0[NjLZ2h14]1cMjLZ2f16^1bMkLY2d18`1_MlLZ2a1:b1]MlLZ2_1<c1\\MmLY2]1>e1YMnLZ2Y1a0h1VMnLZ2W1c0j1TMnLZ2U1e0k1RMoL[2S1f0m1PMoL[2Q1h0o1nLoL[2o0j0P2lLQM[2l0m0Q2iLRM[2i0T1o1cLWMZ2NRM@Z4V3ZL\\M[2EdM[On3_3UL`Mi8[2WGeMn8U2RGlMR9P2nFPNW9j1iFWN\\9d1dF\\Na9^1`FbNd9Z1[FgNk9R1VFnNR:j0mEWO\\:?eEAc:7\\EIm:MUE3S;CoD=Y;YOhDh0`;nNbDR1e<10000O1O1O1O1O]O\\BLc=4`BJ_=6cBI]=6fBHY=8jBFU=:mBES=:PCCR=:RCDP=9SCEo<8SCGo<6TCHm<6Q1M3MWk[3"}, "label": "a woman with curly hair playing wii"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male legs with jean shorts and black and white tennis shoes\"."}], "task": "refering", "answer_template": "The \"male legs with jean shorts and black and white tennis shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 25, 26, 27, 47, 48, 49, 50, 71, 72, 94, 95, 117, 118, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166], "bbox": [0.07946875, 0.0022482435597189696, 0.254875, 0.5213583138173302], "iscrowd": 0, "area": 14815.620650000003, "rle": {"size": [427, 640], "counts": "^Ze06o34h45nJ3l45nJMm4<mJMg4;SKIh4?RKDg4f0SK]Oe4n0UKUOd4V1TKmN\\4k1ZKWNf4R2oJPNQ5X2dJkM\\5^2YJdMg5e2nI^MQ6n400000000000001O00000O100000001O0000000000001O000000000000001O0000000000001O[MXJgMh5d1RKWNn4;dLA\\3:nL@S3`0nL_OR3b0mL^OS3b0nL]OR3c0nL]OR3c0nL]OR3c0nL]OR3c0oL\\OQ3d0oL\\OQ3d0oL\\OQ3e0oLZOQ3f0oLZOQ3f0oLZOQ3f0PMYOP3g0PMYOP3g0PMYOP3g0QMXOo2h0QMXOo2i0PMWOP3i0QMVOo2j0QMVOo2j0QMVOo2j0RMUOm2l0SMTOm2l0SMTOm2l0SMTOm2m0RMSOn2m0RMSOn2m0RMSOn2g0cIVN_3S1n2c0\\M]Od2;dME\\2NQN2o1LSN4m1IVN7j1FYN:g1C\\N=d1@_N?b1_O`Na0`1\\OcNd0]1YOfNg0Z1VOiNj0W1SOlNm0T1QOoNm0R1POROo0n0nNVOQ1j0nNXOQ1h0oNYOP1a601[OfC7[<HeC7]<GeC8\\<GeC8]<EeC:^<CcC<g<M3IffV6"}, "label": "male legs with jean shorts and black and white tennis shoes"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"legs in blue shorts and black shoes\"."}], "task": "refering", "answer_template": "The \"legs in blue shorts and black shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 25, 26, 27, 47, 48, 49, 50, 71, 72, 94, 95, 117, 118, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166], "bbox": [0.07946875, 0.0022482435597189696, 0.254875, 0.5213583138173302], "iscrowd": 0, "area": 14815.620650000003, "rle": {"size": [427, 640], "counts": "^Ze06o34h45nJ3l45nJMm4<mJMg4;SKIh4?RKDg4f0SK]Oe4n0UKUOd4V1TKmN\\4k1ZKWNf4R2oJPNQ5X2dJkM\\5^2YJdMg5e2nI^MQ6n400000000000001O00000O100000001O0000000000001O000000000000001O0000000000001O[MXJgMh5d1RKWNn4;dLA\\3:nL@S3`0nL_OR3b0mL^OS3b0nL]OR3c0nL]OR3c0nL]OR3c0nL]OR3c0oL\\OQ3d0oL\\OQ3d0oL\\OQ3e0oLZOQ3f0oLZOQ3f0oLZOQ3f0PMYOP3g0PMYOP3g0PMYOP3g0QMXOo2h0QMXOo2i0PMWOP3i0QMVOo2j0QMVOo2j0QMVOo2j0RMUOm2l0SMTOm2l0SMTOm2l0SMTOm2m0RMSOn2m0RMSOn2m0RMSOn2g0cIVN_3S1n2c0\\M]Od2;dME\\2NQN2o1LSN4m1IVN7j1FYN:g1C\\N=d1@_N?b1_O`Na0`1\\OcNd0]1YOfNg0Z1VOiNj0W1SOlNm0T1QOoNm0R1POROo0n0nNVOQ1j0nNXOQ1h0oNYOP1a601[OfC7[<HeC7]<GeC8\\<GeC8]<EeC:^<CcC<g<M3IffV6"}, "label": "legs in blue shorts and black shoes"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"legs in red shoes standing on a skate board\"."}], "task": "refering", "answer_template": "The \"legs in red shoes standing on a skate board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199], "bbox": [0.47271875, 0.004473067915690867, 0.7404843750000001, 0.5515690866510539], "iscrowd": 0, "area": 22808.111050000003, "rle": {"size": [427, 640], "counts": "Pbn32X=1O2M3M2N3L3N3N2O001O1O001O001O1O001N100O1O100O1O100O1O100mKSOgKn0k3BRL?_33^LMT3c0iL^Oh2S1VMmNc2\\1ZMeNf2]1WMdNi2^1UMbNk2`1RMaNn2`1PMbNo2_1PMaNP3`1nLaNR3`1lLaNT3_1lLaNT3`1jLaNV3a1hL_NX3k1]LVNc3T2RLmMn3\\2iKdMW4f2^K[Mb4o2SKRMm4X3iJhLW5b3^J_Lb5k3SJVLm5d50000000000000001O0000000000000000000000000000000000000000\\N[JbKe5m3PKoKP5`3eK]LZ4S3YLjLg3e2nLWMR3X2cMRMo2h2m3J6K5J6K5J6M3001O001O0lHWNf2j1XM]Nb2c1\\MeN^2\\1`MkNZ2V1cMSOV2m0iMVOU2k0iMVOW2j0gMXOY2i0eMXO[2h0cMZO]2g0aMZO_2j1[LXNe3Q3QKPMo4g5O00000000000000001O00000000000000001O0000000000gNoIXKQ6`4XJ_Kh5Y4`JgK`5Q4iJoKV5h3SKVLo4b3ZKWLl4a3\\KXLk4`3^KYLh4^3aK\\Le4\\3dK]Lb4[3k2G9H8H8H8G9N2O10000O1000000O10000O10000O;Fd0\\Od0[O_UU2"}, "label": "legs in red shoes standing on a skate board"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing red shoes is riding a skate broad\"."}], "task": "refering", "answer_template": "The \"a man wearing red shoes is riding a skate broad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 35, 36, 37, 38, 39, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199], "bbox": [0.47271875, 0.004473067915690867, 0.7404843750000001, 0.5515690866510539], "iscrowd": 0, "area": 22808.111050000003, "rle": {"size": [427, 640], "counts": "Pbn32X=1O2M3M2N3L3N3N2O001O1O001O001O1O001N100O1O100O1O100O1O100mKSOgKn0k3BRL?_33^LMT3c0iL^Oh2S1VMmNc2\\1ZMeNf2]1WMdNi2^1UMbNk2`1RMaNn2`1PMbNo2_1PMaNP3`1nLaNR3`1lLaNT3_1lLaNT3`1jLaNV3a1hL_NX3k1]LVNc3T2RLmMn3\\2iKdMW4f2^K[Mb4o2SKRMm4X3iJhLW5b3^J_Lb5k3SJVLm5d50000000000000001O0000000000000000000000000000000000000000\\N[JbKe5m3PKoKP5`3eK]LZ4S3YLjLg3e2nLWMR3X2cMRMo2h2m3J6K5J6K5J6M3001O001O0lHWNf2j1XM]Nb2c1\\MeN^2\\1`MkNZ2V1cMSOV2m0iMVOU2k0iMVOW2j0gMXOY2i0eMXO[2h0cMZO]2g0aMZO_2j1[LXNe3Q3QKPMo4g5O00000000000000001O00000000000000001O0000000000gNoIXKQ6`4XJ_Kh5Y4`JgK`5Q4iJoKV5h3SKVLo4b3ZKWLl4a3\\KXLk4`3^KYLh4^3aK\\Le4\\3dK]Lb4[3k2G9H8H8H8G9N2O10000O1000000O10000O10000O;Fd0\\Od0[O_UU2"}, "label": "a man wearing red shoes is riding a skate broad"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing grey gym shoes in the sand\"."}], "task": "refering", "answer_template": "The \"a person wearing grey gym shoes in the sand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 51, 52, 54, 55, 74, 75, 77, 78, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148], "bbox": [0.23803125, 0.0013114754098360658, 0.45695312500000007, 0.4674473067915691], "iscrowd": 0, "area": 14945.069849999998, "rle": {"size": [427, 640], "counts": "Pbo11Q=:E;E;F9M4O1ZMjNUIW1f6YOnHg0m6JgH7S7:aHGZ7i0ZHXO`7Y1THhNg7`3K6K5]Ob0M30000000000000000000000000000000000000SMcIlN]6n0iIROW6h0PJWOP6c0WJ\\Oi5>]JBc58dJG\\53kJLU5NQK2o4HXK7h4C_K<a4^OfKa0Z4[OkKd0U4ZOoKd0Q4ZOPMGP37VMEj29]MBc2<dM_O\\2?jM]OV2a0[5N2N2O1000TH^Oo2b0mLCR3=kLGT39hLLX33dL2[3LdL8[3GbL>]3@aLe0^3YO`Ll0_3SO_LR1_3lNbLU1^3iNbLZ1]3gN`L[1`3eN^L]1b3dN[L^1e3bNZL_1f3bNWL`1i3aNTLa1l3_NnIK`0h1b5^NfI7;b1o5XN_Ic0Hh1i6iMUIU4j6oKlHW4T7mKbHY4^7j0000000000001O0000001O1O001OPNhHmMX7o1nHPNQ7m1UIQNj6k1]IRNc6j1cIUN\\6h1jIUNV6g1PJXNo5d1XJZNg5c1^J[Nb5a1eJ]NZ5_1lJ^NU5_1QK^No4^1WK`Ni4]1[KbNe4Z1`KeN`4W1eKhN[4o0nKQOR4e0YLYOi3=aLB_34lLKT3KWM4i2E_M:a2F`M9`2FaM:_2FbM8_2GcM8]2HdM7\\2HfM7Z2IgM6Y2IiM6W2JjM5V2JlM4U2KmM4e701O1O1O1O1N2O1O1M3Ngm`4"}, "label": "a person wearing grey gym shoes in the sand"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the shadows with gray shoes\"."}], "task": "refering", "answer_template": "The \"a man in the shadows with gray shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 51, 52, 54, 55, 74, 75, 77, 78, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148], "bbox": [0.23803125, 0.0013114754098360658, 0.45695312500000007, 0.4674473067915691], "iscrowd": 0, "area": 14945.069849999998, "rle": {"size": [427, 640], "counts": "Pbo11Q=:E;E;F9M4O1ZMjNUIW1f6YOnHg0m6JgH7S7:aHGZ7i0ZHXO`7Y1THhNg7`3K6K5]Ob0M30000000000000000000000000000000000000SMcIlN]6n0iIROW6h0PJWOP6c0WJ\\Oi5>]JBc58dJG\\53kJLU5NQK2o4HXK7h4C_K<a4^OfKa0Z4[OkKd0U4ZOoKd0Q4ZOPMGP37VMEj29]MBc2<dM_O\\2?jM]OV2a0[5N2N2O1000TH^Oo2b0mLCR3=kLGT39hLLX33dL2[3LdL8[3GbL>]3@aLe0^3YO`Ll0_3SO_LR1_3lNbLU1^3iNbLZ1]3gN`L[1`3eN^L]1b3dN[L^1e3bNZL_1f3bNWL`1i3aNTLa1l3_NnIK`0h1b5^NfI7;b1o5XN_Ic0Hh1i6iMUIU4j6oKlHW4T7mKbHY4^7j0000000000001O0000001O1O001OPNhHmMX7o1nHPNQ7m1UIQNj6k1]IRNc6j1cIUN\\6h1jIUNV6g1PJXNo5d1XJZNg5c1^J[Nb5a1eJ]NZ5_1lJ^NU5_1QK^No4^1WK`Ni4]1[KbNe4Z1`KeN`4W1eKhN[4o0nKQOR4e0YLYOi3=aLB_34lLKT3KWM4i2E_M:a2F`M9`2FaM:_2FbM8_2GcM8]2HdM7\\2HfM7Z2IgM6Y2IiM6W2JjM5V2JlM4U2KmM4e701O1O1O1O1N2O1O1M3Ngm`4"}, "label": "a man in the shadows with gray shoes"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bare footed lega on a scatter board\"."}], "task": "refering", "answer_template": "The \"bare footed lega on a scatter board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 39, 40, 41, 62, 63, 64, 85, 86, 87, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180], "bbox": [0.694171875, 0.0, 0.848609375, 0.532576112412178], "iscrowd": 0, "area": 15435.934500000008, "rle": {"size": [427, 640], "counts": "mZi54l<;H9O001N2O001N2O0O2N2PMGjH;m65jHNT7:cHI[7?\\HCd7S3N3M4mN[KVIk4d6bKnHf4k6Q1J4M3L4L4L4L3O2000000O10O10000000000O100000O10O1O1O1O1O1O1000000O0100000000000000000O1000O100000000000XJ]Ih4c6UKcIh4]6TKjIh4W6UKnIi4R6TKTJi4k5UK[Jh4e5UKaJh4_5UKfJi4Z5TKlJi4T5TKRKi4n4TKWKj4W7^Oi0VO=D7I8H7I7I7I7I7I7I7H9H7I7I7IkWX1"}, "label": "bare footed lega on a scatter board"}]}
{"id": 186605, "image": "COCO_train2014_000000186605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man that is not wearing any shoes\"."}], "task": "refering", "answer_template": "The \"a man that is not wearing any shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 39, 40, 41, 62, 63, 64, 85, 86, 87, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180], "bbox": [0.694171875, 0.0, 0.848609375, 0.532576112412178], "iscrowd": 0, "area": 15435.934500000008, "rle": {"size": [427, 640], "counts": "mZi54l<;H9O001N2O001N2O0O2N2PMGjH;m65jHNT7:cHI[7?\\HCd7S3N3M4mN[KVIk4d6bKnHf4k6Q1J4M3L4L4L4L3O2000000O10O10000000000O100000O10O1O1O1O1O1O1000000O0100000000000000000O1000O100000000000XJ]Ih4c6UKcIh4]6TKjIh4W6UKnIi4R6TKTJi4k5UK[Jh4e5UKaJh4_5UKfJi4Z5TKlJi4T5TKRKi4n4TKWKj4W7^Oi0VO=D7I8H7I7I7I7I7I7I7H9H7I7I7IkWX1"}, "label": "a man that is not wearing any shoes"}]}
{"id": 47347, "image": "COCO_train2014_000000047347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white suv\"."}], "task": "refering", "answer_template": "The \"a white suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 201, 202, 203, 204, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 318], "bbox": [0.716984375, 0.46093555093555094, 0.9990781249999999, 0.7785239085239086], "iscrowd": 0, "area": 17690.56384999999, "rle": {"size": [481, 640], "counts": "^mg6e0S><M001O001O00001N1O2O0O1mMkNYFW1f9nNTFS1k9SO^EF[OY1U;SO`EHUOV1Z;TOaEKoNQ1`;UOaEX1^:jNcET1\\:nNdEQ1[:QOeEn0[:SOdEm0[:UOdEk0[:WOeEh0Z:YOfEg0P:DPF;m9JQF6o9KPF5Q:KoE4Q:MnE3R:NnE1R:0nEOS:1nELS:6lEIT:9kEEW:<iEAX:a0gE^OY:d0fEZO\\:g0cEWO^:k0bESO^:m0bERO_:l0bETO`:h0cEXO]:d0gE[OZ:`0jE@X::mEET:4SFLm9K[F5f9@eF?\\9@dFa0\\9_OdF`0]9@bFa0^9_O[Fi0d9WOUFQ1j9oNSFU1l9kNRFW1n9iNPFZ1o9eNPF^1o9bNoEa1P:_NnEc1R:]NlEf1S:n1F001O1O001O1O001O1O001O1O001O1O1O1O1O10O0100O010O0001O0O2O001N101N100O2N2O0O2N2N101O1O001O1O0O2O001O1O001O1O001O1O001N101O001O001O001O0O2O001O001O001O0O10001O000O2O0000001N1O1N2O2N1O100O2O00001O0000001O000O2O001O0O101O001N101N100O2O0oNb7"}, "label": "a white suv"}]}
{"id": 47347, "image": "COCO_train2014_000000047347.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light colored car behind a woman ' s bare arm holding up an umbrella\"."}], "task": "refering", "answer_template": "The \"a light colored car behind a woman ' s bare arm holding up an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 201, 202, 203, 204, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 318], "bbox": [0.716984375, 0.46093555093555094, 0.9990781249999999, 0.7785239085239086], "iscrowd": 0, "area": 17690.56384999999, "rle": {"size": [481, 640], "counts": "^mg6e0S><M001O001O00001N1O2O0O1mMkNYFW1f9nNTFS1k9SO^EF[OY1U;SO`EHUOV1Z;TOaEKoNQ1`;UOaEX1^:jNcET1\\:nNdEQ1[:QOeEn0[:SOdEm0[:UOdEk0[:WOeEh0Z:YOfEg0P:DPF;m9JQF6o9KPF5Q:KoE4Q:MnE3R:NnE1R:0nEOS:1nELS:6lEIT:9kEEW:<iEAX:a0gE^OY:d0fEZO\\:g0cEWO^:k0bESO^:m0bERO_:l0bETO`:h0cEXO]:d0gE[OZ:`0jE@X::mEET:4SFLm9K[F5f9@eF?\\9@dFa0\\9_OdF`0]9@bFa0^9_O[Fi0d9WOUFQ1j9oNSFU1l9kNRFW1n9iNPFZ1o9eNPF^1o9bNoEa1P:_NnEc1R:]NlEf1S:n1F001O1O001O1O001O1O001O1O001O1O1O1O1O10O0100O010O0001O0O2O001N101N100O2N2O0O2N2N101O1O001O1O0O2O001O1O001O1O001O1O001N101O001O001O001O0O2O001O001O001O0O10001O000O2O0000001N1O1N2O2N1O100O2O00001O0000001O000O2O001O0O101O001N101N100O2O0oNb7"}, "label": "a light colored car behind a woman ' s bare arm holding up an umbrella"}]}
{"id": 39155, "image": "COCO_train2014_000000039155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cow looking to the right\"."}], "task": "refering", "answer_template": "The \"a black and white cow looking to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 214, 215, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 302, 303, 304, 326, 327], "bbox": [0.15140625000000002, 0.5865227272727273, 0.373875, 0.9258409090909091], "iscrowd": 0, "area": 9173.2857, "rle": {"size": [440, 640], "counts": "eQZ12d=2M4L4[Od0H9J5M2N2N1O1OK3N2O2O3]Ob02N3M5L3L5\\EiN`8[1[GiNa8\\1[GhNa8]1ZGfNc8^1YGfNc8]1ZGgNd8[1XGhNh8Y1PGoNn8R1jFVOV9k0aF^O]9d0ZFDe9W21N1O2O0O1O2N1L4L5L4K4L5K4@a0001N1107H0010O0010O01QEkMY:U2dEgNd9X1ZFlN]9lN^FY22oN^9lN]FT22UO_9T1^FPO`9e2O01O0O100O100cN_GcMa8[2bGcM_8\\2cGcM]8\\2eGcM\\8Z2gGdMZ8[2iGcMW8\\2kGcMU8\\2d1N2O1O1O2L3M3L4M3M3N2O1N2N3M2O1N100010O00001D;4L3N2M3N1O00O2O001O1N101N101N1O2O0O100OgNPDl0P<SORDl0m;UOUDi0k;WOVDi0i;WOXDh0g;XO[Dg0e;YO\\Df0c;[O_Dc0a;]O`Db0e;XO]Dh0f;TO[Dl0Y<O2O0O1N2N2N3M3M2N3L4M3M3MUP\\5"}, "label": "a black and white cow looking to the right"}]}
{"id": 39155, "image": "COCO_train2014_000000039155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow seeing right side\"."}], "task": "refering", "answer_template": "The \"a cow seeing right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 214, 215, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 302, 303, 304, 326, 327], "bbox": [0.15140625000000002, 0.5865227272727273, 0.373875, 0.9258409090909091], "iscrowd": 0, "area": 9173.2857, "rle": {"size": [440, 640], "counts": "eQZ12d=2M4L4[Od0H9J5M2N2N1O1OK3N2O2O3]Ob02N3M5L3L5\\EiN`8[1[GiNa8\\1[GhNa8]1ZGfNc8^1YGfNc8]1ZGgNd8[1XGhNh8Y1PGoNn8R1jFVOV9k0aF^O]9d0ZFDe9W21N1O2O0O1O2N1L4L5L4K4L5K4@a0001N1107H0010O0010O01QEkMY:U2dEgNd9X1ZFlN]9lN^FY22oN^9lN]FT22UO_9T1^FPO`9e2O01O0O100O100cN_GcMa8[2bGcM_8\\2cGcM]8\\2eGcM\\8Z2gGdMZ8[2iGcMW8\\2kGcMU8\\2d1N2O1O1O2L3M3L4M3M3N2O1N2N3M2O1N100010O00001D;4L3N2M3N1O00O2O001O1N101N101N1O2O0O100OgNPDl0P<SORDl0m;UOUDi0k;WOVDi0i;WOXDh0g;XO[Dg0e;YO\\Df0c;[O_Dc0a;]O`Db0e;XO]Dh0f;TO[Dl0Y<O2O0O1N2N2N3M3M2N3L4M3M3MUP\\5"}, "label": "a cow seeing right side"}]}
{"id": 178421, "image": "COCO_train2014_000000178421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a polka dotted tie\"."}], "task": "refering", "answer_template": "The \"a polka dotted tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 194, 210, 211, 227, 228, 244, 245, 261, 262, 278, 279, 294, 295, 296, 311, 312, 328, 329, 345, 362], "bbox": [0.3281041666666667, 0.449875, 0.4593541666666667, 0.9465312499999999], "iscrowd": 0, "area": 8130.356299999997, "rle": {"size": [640, 480], "counts": "beR31fc0a0@`0@?@;F:E;F:F:E;F:EYO`@bMV?c2ZARM\\>T3P16K6I6K6N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2O0O2N1O2N2N1O2N1O2N1O2N1O2N1O23M4aN_1J6K4K6K5J5L5J6Km0ROQ1POkTR5"}, "label": "a polka dotted tie"}]}
{"id": 178421, "image": "COCO_train2014_000000178421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the purple tie with circles on it\"."}], "task": "refering", "answer_template": "The \"the purple tie with circles on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 194, 210, 211, 227, 228, 244, 245, 261, 262, 278, 279, 294, 295, 296, 311, 312, 328, 329, 345, 362], "bbox": [0.3281041666666667, 0.449875, 0.4593541666666667, 0.9465312499999999], "iscrowd": 0, "area": 8130.356299999997, "rle": {"size": [640, 480], "counts": "beR31fc0a0@`0@?@;F:E;F:F:E;F:EYO`@bMV?c2ZARM\\>T3P16K6I6K6N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2O0O2N1O2N2N1O2N1O2N1O2N1O2N1O23M4aN_1J6K4K6K5J5L5J6Km0ROQ1POkTR5"}, "label": "the purple tie with circles on it"}]}
{"id": 178421, "image": "COCO_train2014_000000178421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the blue tie with circles on it\"."}], "task": "refering", "answer_template": "The \"the man with the blue tie with circles on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 57, 58, 59, 60, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 125, 126, 127, 128, 142, 143, 144, 145, 146, 159, 160, 161, 162, 163, 164, 165, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.19518749999999999, 0.109234375, 0.9639583333333334, 0.989859375], "iscrowd": 0, "area": 131873.81970000002, "rle": {"size": [640, 480], "counts": "bXk1?`c0a0_Oa0_O7I5K5K6I6K5K6J5K5K6J5K9G1O1O1N2O1O001O1O1O1O1O1O1O1N2@`0TOl0M3M3M3M3L4M3M3M3M3L4M3M3M3M3L4M3M3M3M3L4M3M3L4I7I7I7I7I7I7J6I7I7I7I7I7I7O1_K\\EEe:7bED`:8gEDZ:8mEDT:9RFBP::WFBj9:]FBd9;bF@_9=hF_OY9=nF_OS9=TG^On8?XG]Oi8?^G]Oc8?dG\\O^8`0iG\\OX8a0nG[OS8a0TH[Om7b0YHZOh7f0[HVOf7j0]HROd7m0`HoNa7Q1bHkN_7U1dHgN]7X1gHdNZ7\\1iH`NX7`1kH\\NV7c1nHYNS7g1PIUNQ7j1SIRNn6n1UInMl6R2WIkMi6T2[IhMf6X2]IdMd6\\2_I`Mb6_2`50000O10000O10000O10000O10000O10000000000000000000000000O100000001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N2dBTL`:P4[ESLc:Q4XESLe:Q4VERLg:S4TEQLi:S4QEQLm:S4nDQLo:R4mDQLQ;P4mDRLR;o3lDTLR;m3lDULS;l3kDWLS;j3kDYLS;h3kDZLT;g3jD\\LT;e3gD`LX;a3bDfL\\;[3_DjL`;W3ZDPMd;Q3WDTMh;m2RDZMl;g2oC^MP<[5N1O1O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O0O2O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001N101O001O001O001O00001O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M6J5K6J5K6J5K6I6I8H7H8I8H7H9H7H9H7I7H8I7I7H9H7H8I7I7H8I8H7H8I7H8I7I7H9H7I7HiR:"}, "label": "the man with the blue tie with circles on it"}]}
{"id": 178421, "image": "COCO_train2014_000000178421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black suit and blue tie\"."}], "task": "refering", "answer_template": "The \"a man in a black suit and blue tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 57, 58, 59, 60, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 125, 126, 127, 128, 142, 143, 144, 145, 146, 159, 160, 161, 162, 163, 164, 165, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.19518749999999999, 0.109234375, 0.9639583333333334, 0.989859375], "iscrowd": 0, "area": 131873.81970000002, "rle": {"size": [640, 480], "counts": "bXk1?`c0a0_Oa0_O7I5K5K6I6K5K6J5K5K6J5K9G1O1O1N2O1O001O1O1O1O1O1O1O1N2@`0TOl0M3M3M3M3L4M3M3M3M3L4M3M3M3M3L4M3M3M3M3L4M3M3L4I7I7I7I7I7I7J6I7I7I7I7I7I7O1_K\\EEe:7bED`:8gEDZ:8mEDT:9RFBP::WFBj9:]FBd9;bF@_9=hF_OY9=nF_OS9=TG^On8?XG]Oi8?^G]Oc8?dG\\O^8`0iG\\OX8a0nG[OS8a0TH[Om7b0YHZOh7f0[HVOf7j0]HROd7m0`HoNa7Q1bHkN_7U1dHgN]7X1gHdNZ7\\1iH`NX7`1kH\\NV7c1nHYNS7g1PIUNQ7j1SIRNn6n1UInMl6R2WIkMi6T2[IhMf6X2]IdMd6\\2_I`Mb6_2`50000O10000O10000O10000O10000O10000000000000000000000000O100000001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N2dBTL`:P4[ESLc:Q4XESLe:Q4VERLg:S4TEQLi:S4QEQLm:S4nDQLo:R4mDQLQ;P4mDRLR;o3lDTLR;m3lDULS;l3kDWLS;j3kDYLS;h3kDZLT;g3jD\\LT;e3gD`LX;a3bDfL\\;[3_DjL`;W3ZDPMd;Q3WDTMh;m2RDZMl;g2oC^MP<[5N1O1O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O0O2O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001N101O001O001O001O00001O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M6J5K6J5K6J5K6I6I8H7H8I8H7H9H7H9H7I7H8I7I7H9H7H8I7I7H8I8H7H8I7H8I7I7H9H7I7HiR:"}, "label": "a man in a black suit and blue tie"}]}
{"id": 243959, "image": "COCO_train2014_000000243959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing giants baseball jersey\"."}], "task": "refering", "answer_template": "The \"man wearing giants baseball jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 151, 152, 153, 166, 167, 168, 180, 181, 182, 183, 184, 196, 197, 198, 199, 211, 212, 213, 214, 226, 227, 228, 229, 241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 287, 288, 289], "bbox": [0.0573943661971831, 0.406734375, 0.30384976525821594, 0.8764062499999999], "iscrowd": 0, "area": 18929.254300000004, "rle": {"size": [640, 426], "counts": "_[?1ec0:E;F:E;L4O1O1O1O1N2O1O1O1O1O1N2O1O9G;E<C=D;E<D<D;E<C=D<D;E<D<D;D=D<D6J2N2M2O2N1O2N1N3N1O2N2N1N3N1O2NE<A>G9H9G8H9G8G:G8H9Y1fN]1dN08H=C=C2N1O001O1O1O1O010O1O1O001O1O1O1O001O1O1O3M3M4L4L4QLRCd0R=UObC:b<_ORD0R<IbDFb;4QE[OR;?bEQOa:h0SFgNQ:R1d4I7J6I7I7I6JWXi5"}, "label": "man wearing giants baseball jersey"}]}
{"id": 243959, "image": "COCO_train2014_000000243959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man was speaking in a mouth peace\"."}], "task": "refering", "answer_template": "The \"a man was speaking in a mouth peace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 151, 152, 153, 166, 167, 168, 180, 181, 182, 183, 184, 196, 197, 198, 199, 211, 212, 213, 214, 226, 227, 228, 229, 241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 287, 288, 289], "bbox": [0.0573943661971831, 0.406734375, 0.30384976525821594, 0.8764062499999999], "iscrowd": 0, "area": 18929.254300000004, "rle": {"size": [640, 426], "counts": "_[?1ec0:E;F:E;L4O1O1O1O1N2O1O1O1O1O1N2O1O9G;E<C=D;E<D<D;E<C=D<D;E<D<D;D=D<D6J2N2M2O2N1O2N1N3N1O2N2N1N3N1O2NE<A>G9H9G8H9G8G:G8H9Y1fN]1dN08H=C=C2N1O001O1O1O1O010O1O1O001O1O1O1O001O1O1O3M3M4L4L4QLRCd0R=UObC:b<_ORD0R<IbDFb;4QE[OR;?bEQOa:h0SFgNQ:R1d4I7J6I7I7I6JWXi5"}, "label": "a man was speaking in a mouth peace"}]}
{"id": 243959, "image": "COCO_train2014_000000243959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt holding a baseball bat\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt holding a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 160, 161, 162, 175, 176, 177, 178, 190, 191, 192, 193, 205, 206, 207, 208, 220, 221, 222, 223, 235, 236, 237, 238, 251, 252, 253, 266, 267, 268, 281, 282, 283, 296, 297, 298, 311, 312, 313, 326, 327, 328], "bbox": [0.6742957746478874, 0.40006250000000004, 0.9329342723004695, 0.926296875], "iscrowd": 0, "area": 21366.85425000001, "rle": {"size": [640, 426], "counts": "afc5=\\c0?F:nMWO\\@P1Z?WOa@o0Y?VOc@o0X?UOd@P1V?TOf@S1T?oNj@V1Q?mNl@S1T?POi@Q1W?POg@Q1X?ROf@o0Y?SOd@R1X?ROd@Q1X?UOb@o0;ZMa>c4\\AaKa>`4^AbK`>^4_AeK_>\\4`AfK^>[4^AhKb>l4O1O001O1O0010000O010O100aE`J\\6`5bIbJ_6^5_IcJb6]5\\IdJd6]5[IcJhLL[9`5lIeJhLM\\9^5kIeJhLO^9[5iFYK_1^OZNO_9a5SFfKl1jNaN1o9a6]G^IdN2Q:_6XG`IfN3S:^6TG_IiN4U:\\6PG`IkN4U:i6cFSIWO5W:X7PFdHI4W:Z7nEbHK4X:Z7lEaHL6X:[7jE_HN6Y:\\7fE_H15Y:]7eE^H16[:]7bE]H36[:_7`E[H56[:o7eEQH\\:n7cESH]:m7cESH]:m7aEUH_:k7^EXHb:i7YE[Hh:^8000000001O001O1O1O1O1O1O1O2N1O2N2N1O2N3M3M2N3M4L5K4L1cLQDgNP<U1[DcNg;Y1cD^N_;^1lDXNW;d1SESNP;i1[EmMh:P2XFiLR:S3`3L5L5I7I6I8I7J6J5K6J5K5H9Gg0YOhga0"}, "label": "a man in a black shirt holding a baseball bat"}]}
{"id": 243959, "image": "COCO_train2014_000000243959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man swinging a baseball bat\"."}], "task": "refering", "answer_template": "The \"a man swinging a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 160, 161, 162, 175, 176, 177, 178, 190, 191, 192, 193, 205, 206, 207, 208, 220, 221, 222, 223, 235, 236, 237, 238, 251, 252, 253, 266, 267, 268, 281, 282, 283, 296, 297, 298, 311, 312, 313, 326, 327, 328], "bbox": [0.6742957746478874, 0.40006250000000004, 0.9329342723004695, 0.926296875], "iscrowd": 0, "area": 21366.85425000001, "rle": {"size": [640, 426], "counts": "afc5=\\c0?F:nMWO\\@P1Z?WOa@o0Y?VOc@o0X?UOd@P1V?TOf@S1T?oNj@V1Q?mNl@S1T?POi@Q1W?POg@Q1X?ROf@o0Y?SOd@R1X?ROd@Q1X?UOb@o0;ZMa>c4\\AaKa>`4^AbK`>^4_AeK_>\\4`AfK^>[4^AhKb>l4O1O001O1O0010000O010O100aE`J\\6`5bIbJ_6^5_IcJb6]5\\IdJd6]5[IcJhLL[9`5lIeJhLM\\9^5kIeJhLO^9[5iFYK_1^OZNO_9a5SFfKl1jNaN1o9a6]G^IdN2Q:_6XG`IfN3S:^6TG_IiN4U:\\6PG`IkN4U:i6cFSIWO5W:X7PFdHI4W:Z7nEbHK4X:Z7lEaHL6X:[7jE_HN6Y:\\7fE_H15Y:]7eE^H16[:]7bE]H36[:_7`E[H56[:o7eEQH\\:n7cESH]:m7cESH]:m7aEUH_:k7^EXHb:i7YE[Hh:^8000000001O001O1O1O1O1O1O1O2N1O2N2N1O2N3M3M2N3M4L5K4L1cLQDgNP<U1[DcNg;Y1cD^N_;^1lDXNW;d1SESNP;i1[EmMh:P2XFiLR:S3`3L5L5I7I6I8I7J6J5K6J5K5H9Gg0YOhga0"}, "label": "a man swinging a baseball bat"}]}
{"id": 243959, "image": "COCO_train2014_000000243959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a microphone\"."}], "task": "refering", "answer_template": "The \"a woman holding a microphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 187, 188, 201, 202, 203, 216, 217, 218, 231, 232, 233, 246, 247, 248, 261, 262, 263, 276, 278], "bbox": [0.3985211267605634, 0.51096875, 0.5979107981220657, 0.8164374999999999], "iscrowd": 0, "area": 7629.401399999999, "rle": {"size": [640, 426], "counts": "QhZ32lc02M4M2N2M3N3L3M3M1O101N1O1O100O2N1O100O1O2N100]MQOfA`1V>iNTAh1k>^Nm@e1T?_Nd@e1]?]N\\@f1e?]NT@f1m?U11O@YLj@g3V?bLa@]3_?iL]@U3d?b01O100O003M3J6J5L5J6003M2N2N3M2N3M2N8H;E:F;E7IO100O1O1O100O1O100O1O1oN^AbLc>g2^BoLb=Z2UCeMl<j1eCUN\\<g1hCYNX<c1lC\\NU<`1oC_NR<]1QDcNo;Z1UDeNl;W1XDhNi;T1[DjNi;P1[DnNh;m0j3J6J5Bb`Z3"}, "label": "a woman holding a microphone"}]}
{"id": 243959, "image": "COCO_train2014_000000243959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in black pants and jean jacket with a pink shirt standing behind a man with a bat\"."}], "task": "refering", "answer_template": "The \"a lady in black pants and jean jacket with a pink shirt standing behind a man with a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 187, 188, 201, 202, 203, 216, 217, 218, 231, 232, 233, 246, 247, 248, 261, 262, 263, 276, 278], "bbox": [0.3985211267605634, 0.51096875, 0.5979107981220657, 0.8164374999999999], "iscrowd": 0, "area": 7629.401399999999, "rle": {"size": [640, 426], "counts": "QhZ32lc02M4M2N2M3N3L3M3M1O101N1O1O100O2N1O100O1O2N100]MQOfA`1V>iNTAh1k>^Nm@e1T?_Nd@e1]?]N\\@f1e?]NT@f1m?U11O@YLj@g3V?bLa@]3_?iL]@U3d?b01O100O003M3J6J5L5J6003M2N2N3M2N3M2N8H;E:F;E7IO100O1O1O100O1O100O1O1oN^AbLc>g2^BoLb=Z2UCeMl<j1eCUN\\<g1hCYNX<c1lC\\NU<`1oC_NR<]1QDcNo;Z1UDeNl;W1XDhNi;T1[DjNi;P1[DnNh;m0j3J6J5Bb`Z3"}, "label": "a lady in black pants and jean jacket with a pink shirt standing behind a man with a bat"}]}
{"id": 538872, "image": "COCO_train2014_000000538872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in purple\"."}], "task": "refering", "answer_template": "The \"girl in purple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 78, 79, 80, 81, 101, 102, 103, 124, 125, 126, 147, 148, 149, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264], "bbox": [0.41178125000000004, 0.10464373464373465, 0.538140625, 0.7842997542997544], "iscrowd": 0, "area": 10764.12075, "rle": {"size": [407, 640], "counts": "^QY3<T<?Fe0[O?A4L3bGTOm4n0TJ8d5I\\J=]5DcJc0UOYMa5T2[KR1d4oN\\KQ1d4oN\\KQ1d4oN\\KP1d4QO\\Ko0d4QO\\Ko0c4RO]Kn0c4RO]Kn0T4fLTK[2j0m0P4EPL;n3GRL9k3JULClNiMm4f2WLYOZOkM\\4o2ZLUNPOTOg0Bo3V3ZLQNb5Q2^JlMc5V2]JhMc5Z2^JbMc5`2]J`Mb5a2^JbM]5`2cJbMROIm4g2QLbMhN6o4Y2YLVO_3l0aL@R3a0nL_OQ3b0PM]Oo2d0QM\\On2e0RM[Ol2g0TMYOk2h0UMXOj2i0VMWOj2i0VMWOj2i0VMWOj2i0VMgNjNdMP4e3WMaNoNiMj3f3WM\\NUOlMe3h3VMXNZOoM`3i3VMnMQ4R2oKcLD^Og4o3eKaL^5_3bJ_L`5`3aJ^La5b3_J\\Lc5d3^JYLd5g3\\JVLh5i3c101O1O1O1O1YN`GUO`8j0eGRO\\8l0iGPOX8=WGmNe0c0V87\\GUOb0a0X8M[GDc09Y8EXG4n0Ih89i1110O1O010O11N3N2N2M2O1Khae3"}, "label": "girl in purple"}]}
{"id": 538872, "image": "COCO_train2014_000000538872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girl that is on the left wearing purple\"."}], "task": "refering", "answer_template": "The \"the little girl that is on the left wearing purple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 78, 79, 80, 81, 101, 102, 103, 124, 125, 126, 147, 148, 149, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264], "bbox": [0.41178125000000004, 0.10464373464373465, 0.538140625, 0.7842997542997544], "iscrowd": 0, "area": 10764.12075, "rle": {"size": [407, 640], "counts": "^QY3<T<?Fe0[O?A4L3bGTOm4n0TJ8d5I\\J=]5DcJc0UOYMa5T2[KR1d4oN\\KQ1d4oN\\KQ1d4oN\\KP1d4QO\\Ko0d4QO\\Ko0c4RO]Kn0c4RO]Kn0T4fLTK[2j0m0P4EPL;n3GRL9k3JULClNiMm4f2WLYOZOkM\\4o2ZLUNPOTOg0Bo3V3ZLQNb5Q2^JlMc5V2]JhMc5Z2^JbMc5`2]J`Mb5a2^JbM]5`2cJbMROIm4g2QLbMhN6o4Y2YLVO_3l0aL@R3a0nL_OQ3b0PM]Oo2d0QM\\On2e0RM[Ol2g0TMYOk2h0UMXOj2i0VMWOj2i0VMWOj2i0VMWOj2i0VMgNjNdMP4e3WMaNoNiMj3f3WM\\NUOlMe3h3VMXNZOoM`3i3VMnMQ4R2oKcLD^Og4o3eKaL^5_3bJ_L`5`3aJ^La5b3_J\\Lc5d3^JYLd5g3\\JVLh5i3c101O1O1O1O1YN`GUO`8j0eGRO\\8l0iGPOX8=WGmNe0c0V87\\GUOb0a0X8M[GDc09Y8EXG4n0Ih89i1110O1O010O11N3N2N2M2O1Khae3"}, "label": "the little girl that is on the left wearing purple"}]}
{"id": 538872, "image": "COCO_train2014_000000538872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a red and gray shirt next to the water\"."}], "task": "refering", "answer_template": "The \"a boy in a red and gray shirt next to the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 39, 40, 61, 62, 63, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 180, 199, 200, 201, 222, 223, 224, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294], "bbox": [0.6502656250000001, 0.033783783783783786, 0.8608125000000001, 0.8536117936117936], "iscrowd": 0, "area": 21304.965050000006, "rle": {"size": [407, 640], "counts": "__U53^<:G9_Oa0C<I6I7J6J6M3Lc0^O:E9G7J5K6J6J;F:E;E`0@=mJnIi2[6gLlIX3^6XLiIg3`6kKfIU4^7O1O1O1O1N2O1QObGcM_8[2hG`MX8]2QH\\MP8b2XHXMh7f2_HTMb7i2W100000000010kM\\MfIOmNe2\\7SNjHd2T7a1N2M3M3N2N20001O0000000001O2N1O1O2O8H:G;D8I2M3N2M30lG]LP7d3hHbLX7`3aHdL`7Z4d0\\O1N01N2N2N1]L\\LQNe3c1kLXNU3`1XM[Ni2]1eM[N]2]1QN\\NP2<ZOAf0MO021OL25OH1:0C1>0^O2c0a0gN@[1b0_NAb1a0YN@i1a0SNEj1i5L3M3M1O1O1O2NO100O1O100O0010O01O4M8H1OO1O2M2Oh0XO1N2O1N101O1N2O2N1N2O1O1N2O0O2LcUS1"}, "label": "a boy in a red and gray shirt next to the water"}]}
{"id": 538872, "image": "COCO_train2014_000000538872.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a black white and red shirt playing in water\"."}], "task": "refering", "answer_template": "The \"a boy in a black white and red shirt playing in water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 39, 40, 61, 62, 63, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 180, 199, 200, 201, 222, 223, 224, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294], "bbox": [0.6502656250000001, 0.033783783783783786, 0.8608125000000001, 0.8536117936117936], "iscrowd": 0, "area": 21304.965050000006, "rle": {"size": [407, 640], "counts": "__U53^<:G9_Oa0C<I6I7J6J6M3Lc0^O:E9G7J5K6J6J;F:E;E`0@=mJnIi2[6gLlIX3^6XLiIg3`6kKfIU4^7O1O1O1O1N2O1QObGcM_8[2hG`MX8]2QH\\MP8b2XHXMh7f2_HTMb7i2W100000000010kM\\MfIOmNe2\\7SNjHd2T7a1N2M3M3N2N20001O0000000001O2N1O1O2O8H:G;D8I2M3N2M30lG]LP7d3hHbLX7`3aHdL`7Z4d0\\O1N01N2N2N1]L\\LQNe3c1kLXNU3`1XM[Ni2]1eM[N]2]1QN\\NP2<ZOAf0MO021OL25OH1:0C1>0^O2c0a0gN@[1b0_NAb1a0YN@i1a0SNEj1i5L3M3M1O1O1O2NO100O1O100O0010O01O4M8H1OO1O2M2Oh0XO1N2O1N101O1N2O2N1N2O1O1N2O0O2LcUS1"}, "label": "a boy in a black white and red shirt playing in water"}]}
{"id": 47357, "image": "COCO_train2014_000000047357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a virgin train with a yellow front\"."}], "task": "refering", "answer_template": "The \"a virgin train with a yellow front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 161, 162, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227], "bbox": [0.0, 0.30972568578553616, 0.9053437499999999, 0.6805236907730673], "iscrowd": 0, "area": 33817.89905, "rle": {"size": [401, 640], "counts": "j4o1b:000000001O000000000000000010O0000000000000001O0000000000000000001O00000000000000001O01O00000000000N2XOh0XOcRS3F_nlL7I6J6J6J6jEBl7e0lF=T9`1000000000000000000000000000000000000O01000000000000000000000000000000000000000O100000000000000000O1000000000000000000000O100000000000000000000000000000000000O1000O10000000000000000O100000000O1000000O100000000O1000000O010000000O1000000O100000000000000000000000001O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O000000O01000000000000000000001O001N101O001O001O001O00001O001N101O001O001O001O001O001O0O2O1O001O001O1O001O1O001O0O2O1O001O1O001O001O1O1O2N2M3N2N1O2N2N2N2N2N2N2N2N2N2M3N2N1O2N2N2N2N2N2N2N3M4J5G:F9G9G:Faeg0"}, "label": "a virgin train with a yellow front"}]}
{"id": 47357, "image": "COCO_train2014_000000047357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train that is to the right of another train\"."}], "task": "refering", "answer_template": "The \"a train that is to the right of another train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 109, 110, 115, 116, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 161, 162, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227], "bbox": [0.0, 0.30972568578553616, 0.9053437499999999, 0.6805236907730673], "iscrowd": 0, "area": 33817.89905, "rle": {"size": [401, 640], "counts": "j4o1b:000000001O000000000000000010O0000000000000001O0000000000000000001O00000000000000001O01O00000000000N2XOh0XOcRS3F_nlL7I6J6J6J6jEBl7e0lF=T9`1000000000000000000000000000000000000O01000000000000000000000000000000000000000O100000000000000000O1000000000000000000000O100000000000000000000000000000000000O1000O10000000000000000O100000000O1000000O100000000O1000000O010000000O1000000O100000000000000000000000001O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O000000O01000000000000000000001O001N101O001O001O001O00001O001N101O001O001O001O001O001O0O2O1O001O001O1O001O1O001O0O2O1O001O1O001O001O1O1O2N2M3N2N1O2N2N2N2N2N2N2N2N2N2M3N2N1O2N2N2N2N2N2N2N3M4J5G:F9G9G:Faeg0"}, "label": "a train that is to the right of another train"}]}
{"id": 47357, "image": "COCO_train2014_000000047357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow , red , and gray train below another train\"."}], "task": "refering", "answer_template": "The \"a yellow , red , and gray train below another train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218], "bbox": [0.07603125, 0.2675561097256858, 0.49703125000000004, 0.683291770573566], "iscrowd": 0, "area": 34518.0859, "rle": {"size": [401, 640], "counts": "j\\c0e0T;h0XOh0_Oa0O100O1O1O101N1O100O1O1O100O1O100O1O1O100O1O1O100O1O2O000000000O100000000O10000000000O10001O00000O10000000000O100000000O10000000001N10000000000O10000000000O10000000000O2O0000000O10000000000O10000000000O1000001O000O10000000000O100000000O10000000000O1000O100000O10000000000O100000O1000O100000000O1lGXL_7h3^H\\La7d3[HaLd7_3YHeLf7S401O1N2O001O1O1O001O1N10000000000000000O10O100000000000O10000000000000000O10000001O000O10001O000000001O0000001O001O001O001O001O0O101O001O001O001O001O1O001O1O001O001O1O001O1N101O1O001O001O1O001O1O001O3M3M4L3M3M4K5L4K6K4K6RNWGIl:M3L3L5H[km3"}, "label": "a yellow , red , and gray train below another train"}]}
{"id": 47357, "image": "COCO_train2014_000000047357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train that is not pulling any other cars\"."}], "task": "refering", "answer_template": "The \"the train that is not pulling any other cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218], "bbox": [0.07603125, 0.2675561097256858, 0.49703125000000004, 0.683291770573566], "iscrowd": 0, "area": 34518.0859, "rle": {"size": [401, 640], "counts": "j\\c0e0T;h0XOh0_Oa0O100O1O1O101N1O100O1O1O100O1O100O1O1O100O1O1O100O1O2O000000000O100000000O10000000000O10001O00000O10000000000O100000000O10000000001N10000000000O10000000000O10000000000O2O0000000O10000000000O10000000000O1000001O000O10000000000O100000000O10000000000O1000O100000O10000000000O100000O1000O100000000O1lGXL_7h3^H\\La7d3[HaLd7_3YHeLf7S401O1N2O001O1O1O001O1N10000000000000000O10O100000000000O10000000000000000O10000001O000O10001O000000001O0000001O001O001O001O001O0O101O001O001O001O001O1O001O1O001O001O1O001O1N101O1O001O001O1O001O1O001O3M3M4L3M3M4K5L4K6K4K6RNWGIl:M3L3L5H[km3"}, "label": "the train that is not pulling any other cars"}]}
{"id": 424193, "image": "COCO_train2014_000000424193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black horse with a blue cover tied to a horse trailer\"."}], "task": "refering", "answer_template": "The \"a black horse with a blue cover tied to a horse trailer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 227, 228, 229, 245, 246, 247, 249, 250, 251, 252, 273], "bbox": [0.5646093750000001, 0.32739583333333333, 0.9825781250000001, 0.6667291666666666], "iscrowd": 0, "area": 22370.641699999993, "rle": {"size": [480, 640], "counts": "dlY53j>3M4L3M4M2M4L3M4L3M4J5M4N1O1N3N1O2NO1O0O2O1O1O11N100002N1O101N1O1O2N1O2O0O1O20O100O1O100O1O100O100O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O0_CVNl;k1oCYNQ<h1jC]NU<d1gC`NX<a1cCdN\\<P201VE_MS9b2gFdMX9]2cFhM\\9X2`FmM_9T2[FRNd9o1WFVNh9j1SF\\Nl9e1oE`NP:a1kEdNT:]1fEiNY:P3O1O10M2N2N2N2N2K5H8H8H8H8I7O100O100O100O2O0O101N100O21N2N101N2N101N2N1O2O1N1QElLP:V3iEoLX:R3aETM^:U2UEUN7Kd:o1YEUNL2k:g1]EmNc:R1bEkN]:S1hEjNY:T1kEiNU:V1PFgNo9X1UFeNl9Y1XFeNg9Y1_FcNa9]1bF^Na9a1bFVNd9j1i1O1000000000000000000000000000000000000000000000000000O10O10000000O10000000000O100mDmMY9S2fFoMY9Q2eFRNZ9n1eFTNY9m1eFVNZ9j1eFWN[9i1cFYN]9g1bFZN^9f1`F]N_9c1`F^N_9c1_F_Na9a1^F`Nb9`1\\FbNd9^1[FcNe9]1YFfNf9Z1YFgNf9Z1XFhNh9W1XFiNi9W1iEUOY:k0YEBh:Z21N10dEQLi9n3WFSLi9k3WFWLi9h3WFYLi9f3WF[Li9c3WF_Lh9a3XFaLg9^3YFcLh9Z3XFhLh9W3XFjLh9U3WFmLj9m2ZFTMf9h2]FYMd9b2^F_Mc9_2^FaMc9_2[FcMf9]2XFcMi9^2TFcMn9]2lEgMU:Y2fEkM[:V2`EmMb:W3100O2N1O2N1O1[Of0J5oNYDdNh0YOX:Q2[EWN=NW:j1\\F\\Nd9c1WFdNi9Y1TFlNl9S1oETOQ:j0jE\\OV:b0gEDX:;cEK^:3]E4b:KZE:f:DVEc0j:[OQEk0o:c11O00001O0O2O00001O001N10006J;E;D=D;E;E;E:F;D;F;E`k4"}, "label": "a black horse with a blue cover tied to a horse trailer"}]}
{"id": 424193, "image": "COCO_train2014_000000424193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse with the blue cover on its back\"."}], "task": "refering", "answer_template": "The \"the horse with the blue cover on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 227, 228, 229, 245, 246, 247, 249, 250, 251, 252, 273], "bbox": [0.5646093750000001, 0.32739583333333333, 0.9825781250000001, 0.6667291666666666], "iscrowd": 0, "area": 22370.641699999993, "rle": {"size": [480, 640], "counts": "dlY53j>3M4L3M4M2M4L3M4L3M4J5M4N1O1N3N1O2NO1O0O2O1O1O11N100002N1O101N1O1O2N1O2O0O1O20O100O1O100O1O100O100O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O0_CVNl;k1oCYNQ<h1jC]NU<d1gC`NX<a1cCdN\\<P201VE_MS9b2gFdMX9]2cFhM\\9X2`FmM_9T2[FRNd9o1WFVNh9j1SF\\Nl9e1oE`NP:a1kEdNT:]1fEiNY:P3O1O10M2N2N2N2N2K5H8H8H8H8I7O100O100O100O2O0O101N100O21N2N101N2N101N2N1O2O1N1QElLP:V3iEoLX:R3aETM^:U2UEUN7Kd:o1YEUNL2k:g1]EmNc:R1bEkN]:S1hEjNY:T1kEiNU:V1PFgNo9X1UFeNl9Y1XFeNg9Y1_FcNa9]1bF^Na9a1bFVNd9j1i1O1000000000000000000000000000000000000000000000000000O10O10000000O10000000000O100mDmMY9S2fFoMY9Q2eFRNZ9n1eFTNY9m1eFVNZ9j1eFWN[9i1cFYN]9g1bFZN^9f1`F]N_9c1`F^N_9c1_F_Na9a1^F`Nb9`1\\FbNd9^1[FcNe9]1YFfNf9Z1YFgNf9Z1XFhNh9W1XFiNi9W1iEUOY:k0YEBh:Z21N10dEQLi9n3WFSLi9k3WFWLi9h3WFYLi9f3WF[Li9c3WF_Lh9a3XFaLg9^3YFcLh9Z3XFhLh9W3XFjLh9U3WFmLj9m2ZFTMf9h2]FYMd9b2^F_Mc9_2^FaMc9_2[FcMf9]2XFcMi9^2TFcMn9]2lEgMU:Y2fEkM[:V2`EmMb:W3100O2N1O2N1O1[Of0J5oNYDdNh0YOX:Q2[EWN=NW:j1\\F\\Nd9c1WFdNi9Y1TFlNl9S1oETOQ:j0jE\\OV:b0gEDX:;cEK^:3]E4b:KZE:f:DVEc0j:[OQEk0o:c11O00001O0O2O00001O001N10006J;E;D=D;E;E;E:F;D;F;E`k4"}, "label": "the horse with the blue cover on its back"}]}
{"id": 424193, "image": "COCO_train2014_000000424193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown and white horse\"."}], "task": "refering", "answer_template": "The \"the brown and white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 235, 236, 237, 240, 253, 254, 255, 258, 259, 260, 276, 277, 278, 279, 281, 282, 283, 299, 300, 301, 302, 322, 323], "bbox": [0.025750000000000002, 0.3948541666666667, 0.4731718750000001, 0.8626666666666668], "iscrowd": 0, "area": 29018.112750000008, "rle": {"size": [480, 640], "counts": "Tk7`0V>b0^BQOn;h1lCZN`:^3ZEeLb:d3WE^Ld:R4L3M4L4K5L6J4L1N3N3M3L4M3M3N2N2O1N2O1O1NSOn0O1O001O1O2N1LRFVKP:g45N3N1N3M2N3M2O2N1O2M3N1O25K3L4M3M2N2M4M2N2M3N1O2N1N3N2N3M`NiFZMT9a2TG`Mh8\\2`GcM]8X2lGfMQ8W2]HaMa7]2cHaM]7]2fHbMZ7]2hHbMX7\\2kHcMT7]2nHaMS7]2PIbMP7\\2SIcMm6\\2UIcMk6[2XIdMg6\\2[IcMe6[2_IcMa6\\2aIcM_6\\2cIcM]6]2dIbM[6_2fIaMY6_2hIaMW6_2jIaMU6_2lI`MT6`2oI^MP6b2VJYMi5g2X3000000000000000001O001O001O001O001O001O001O001O000O1000000000000000000O10000000000000000O1000O100^ElMX8T2dGRNY8o1dGUN[8k1bGYN]8f1_G_N`8b1[GcNe8]1WGfNi8[1RGjNn8V1nFmNR9U1iFoNW9R1cFSO\\9P1^FSOc9U3ODZF_Kg9]4?O2N2N2N1O2M3K5G8H9F:I7000000O100000000O11bDWMc:m2hDlMP;X3J:VESLm9c4M2N2N10O0001O01O0YNdFkM[9S2TG`Mm8]2cGUM]8f2PHRMP8f2ZHYMe7`2cH_M^7[2iHcMW7\\2lHcMS7\\2PIbMP7]2RIbMo6\\2k2O1O2M201N1O1O2O0O1O2N101N1O1O2O0O2O00001O00001O00001O01O01O00010O0000010O000010O0001O01O01O00010O001O010O02N2O1N2O1N101N100O2O0O101O0O101O2N2N1O1O1ON2N3N1N2N2O2M2O1N2N2O1N2O1N5K6K:E<E_fm4"}, "label": "the brown and white horse"}]}
{"id": 424193, "image": "COCO_train2014_000000424193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that isn ' t covered\"."}], "task": "refering", "answer_template": "The \"the horse that isn ' t covered\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 235, 236, 237, 240, 253, 254, 255, 258, 259, 260, 276, 277, 278, 279, 281, 282, 283, 299, 300, 301, 302, 322, 323], "bbox": [0.025750000000000002, 0.3948541666666667, 0.4731718750000001, 0.8626666666666668], "iscrowd": 0, "area": 29018.112750000008, "rle": {"size": [480, 640], "counts": "Tk7`0V>b0^BQOn;h1lCZN`:^3ZEeLb:d3WE^Ld:R4L3M4L4K5L6J4L1N3N3M3L4M3M3N2N2O1N2O1O1NSOn0O1O001O1O2N1LRFVKP:g45N3N1N3M2N3M2O2N1O2M3N1O25K3L4M3M2N2M4M2N2M3N1O2N1N3N2N3M`NiFZMT9a2TG`Mh8\\2`GcM]8X2lGfMQ8W2]HaMa7]2cHaM]7]2fHbMZ7]2hHbMX7\\2kHcMT7]2nHaMS7]2PIbMP7\\2SIcMm6\\2UIcMk6[2XIdMg6\\2[IcMe6[2_IcMa6\\2aIcM_6\\2cIcM]6]2dIbM[6_2fIaMY6_2hIaMW6_2jIaMU6_2lI`MT6`2oI^MP6b2VJYMi5g2X3000000000000000001O001O001O001O001O001O001O001O000O1000000000000000000O10000000000000000O1000O100^ElMX8T2dGRNY8o1dGUN[8k1bGYN]8f1_G_N`8b1[GcNe8]1WGfNi8[1RGjNn8V1nFmNR9U1iFoNW9R1cFSO\\9P1^FSOc9U3ODZF_Kg9]4?O2N2N2N1O2M3K5G8H9F:I7000000O100000000O11bDWMc:m2hDlMP;X3J:VESLm9c4M2N2N10O0001O01O0YNdFkM[9S2TG`Mm8]2cGUM]8f2PHRMP8f2ZHYMe7`2cH_M^7[2iHcMW7\\2lHcMS7\\2PIbMP7]2RIbMo6\\2k2O1O2M201N1O1O2O0O1O2N101N1O1O2O0O2O00001O00001O00001O01O01O00010O0000010O000010O0001O01O01O00010O001O010O02N2O1N2O1N101N100O2O0O101O0O101O2N2N1O1O1ON2N3N1N2N2O2M2O1N2N2O1N2O1N5K6K:E<E_fm4"}, "label": "the horse that isn ' t covered"}]}
{"id": 276740, "image": "COCO_train2014_000000276740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chili dog with slices of cheese visible under the chili\"."}], "task": "refering", "answer_template": "The \"a chili dog with slices of cheese visible under the chili\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0973125, 0.29702083333333335, 0.41785937500000003, 1.0], "iscrowd": 0, "area": 51131.7325, "rle": {"size": [480, 640], "counts": "d_m0:`>7K6K4K6J5L4K5L4K5L5J5L4L4K5L4K5L4K5L4L4K4M3M4L3M4L3M3M4L3L4L5K4L4L5K4L5K4L4L5K4L4L5K4L4L4L5L3N2M3M4M2M3M3N3L3M3N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2O1O1O1N2O1O1O1O1N2O1O1O100O1000000O100000000O100000000000000001O00000000001O00001O1O1O1O1O1O2N1O001O1O1O1O001O00O100O100O100O100000000000000000000001O1O001O1O001O000N2O1O1O1N2O2N1O1_MgJYK[5e4jJVKY5X2SK]MH6W5W2[K^MD4S5X2dK]MB1m4`2gKZMDMh4f2hKZMIFa4n2jKXMKD^4Q3lKWMHF^4P3nKVMHF^4P3nKWMFF^4Q3QLUMDF]4S3SLSMCG\\4l2^LYMYOHZ4d2jLaMnNHZ4Z2WMiMS5k0W5D<ETf^5"}, "label": "a chili dog with slices of cheese visible under the chili"}]}
{"id": 276740, "image": "COCO_train2014_000000276740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog with chili on top\"."}], "task": "refering", "answer_template": "The \"a hot dog with chili on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0973125, 0.29702083333333335, 0.41785937500000003, 1.0], "iscrowd": 0, "area": 51131.7325, "rle": {"size": [480, 640], "counts": "d_m0:`>7K6K4K6J5L4K5L4K5L5J5L4L4K5L4K5L4K5L4L4K4M3M4L3M4L3M3M4L3L4L5K4L4L5K4L5K4L4L5K4L4L5K4L4L4L5L3N2M3M4M2M3M3N3L3M3N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2O1O1O1N2O1O1O1O1N2O1O1O100O1000000O100000000O100000000000000001O00000000001O00001O1O1O1O1O1O2N1O001O1O1O1O001O00O100O100O100O100000000000000000000001O1O001O1O001O000N2O1O1O1N2O2N1O1_MgJYK[5e4jJVKY5X2SK]MH6W5W2[K^MD4S5X2dK]MB1m4`2gKZMDMh4f2hKZMIFa4n2jKXMKD^4Q3lKWMHF^4P3nKVMHF^4P3nKWMFF^4Q3QLUMDF]4S3SLSMCG\\4l2^LYMYOHZ4d2jLaMnNHZ4Z2WMiMS5k0W5D<ETf^5"}, "label": "a hot dog with chili on top"}]}
{"id": 276740, "image": "COCO_train2014_000000276740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a container of french fries covered with cheese and chili\"."}], "task": "refering", "answer_template": "The \"a container of french fries covered with cheese and chili\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 62, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 247, 248, 249, 250, 251, 270, 271, 272, 273, 293, 294, 295, 317, 318], "bbox": [0.720109375, 0.07952083333333333, 0.9799374999999999, 0.7896875], "iscrowd": 0, "area": 33258.85770000001, "rle": {"size": [480, 640], "counts": "bTh61o>2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N3M2N2N2XG2Z21aM:V2IeM>V2EdMc0W2@dMg0W2]OcMj0X2YObMo0Y2TObMP1\\2SO_MP1`2SO[MQ1c2ROWMR1h2QOSMR1l2ROnLQ1Q3ROiLS1U3POfLS1Y3PObLS1]3PO]LT1b3oNYLU1e3oNULT1j3oNQLT1n3P5N3N1N2O1O2M2O1N3N1O1N2O2M2O1O1N3N1N2OlNYKPHf4P8[KQHb4P8_KPH`4P8aKPH^4P8cKPH[4P8hKPHV4P8kKPHS4Q8nKnGR4R8oKlGR4T8oKkGP4V8QLhGQ4W8PLhGQ4V8QLiGo3W8SLfGo3Y8RLfGo3Y8RLfGo3Y8RLeGP4Z8QLeGo3Z8SLdGo3[8RLdGo3]8QLaGP4`8oK^GR4d8mK[GT4f8kKYGV4h8iKVGY4k8fKTGZ4l8gKSGZ4l8hKQGZ4n8gKQGZ4n8gKPGZ4P9gKoFZ4P9k00O1O1O1OAYJgGf5Z8\\JdGc5\\8`JbG_5^8eJ_GZ5b8hJ\\GW5d8lJZGS5f8c010O1O101N1O100O2O0O1O2O0O1O2O0O1O2`M^I`Lb6`3dIYL]6g3hISLZ6l3lInKT6R4RJgKo5Y4WJ`Kk5_4ZJ\\Ki5a4]JXKi5c4]JVKh5f4^JTKf5h4`JQKe5k4`JoJf5l4[2L4L5K4L4L4L5K4L4L5K4L4L4L5K4L4L4L5K2N3M3M3M3M2N3M3N2M3M2N3J6H8H8G9H7Inl5"}, "label": "a container of french fries covered with cheese and chili"}]}
{"id": 276740, "image": "COCO_train2014_000000276740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chili and cheese on fries\"."}], "task": "refering", "answer_template": "The \"chili and cheese on fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 62, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 247, 248, 249, 250, 251, 270, 271, 272, 273, 293, 294, 295, 317, 318], "bbox": [0.720109375, 0.07952083333333333, 0.9799374999999999, 0.7896875], "iscrowd": 0, "area": 33258.85770000001, "rle": {"size": [480, 640], "counts": "bTh61o>2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N3M2N2N2XG2Z21aM:V2IeM>V2EdMc0W2@dMg0W2]OcMj0X2YObMo0Y2TObMP1\\2SO_MP1`2SO[MQ1c2ROWMR1h2QOSMR1l2ROnLQ1Q3ROiLS1U3POfLS1Y3PObLS1]3PO]LT1b3oNYLU1e3oNULT1j3oNQLT1n3P5N3N1N2O1O2M2O1N3N1O1N2O2M2O1O1N3N1N2OlNYKPHf4P8[KQHb4P8_KPH`4P8aKPH^4P8cKPH[4P8hKPHV4P8kKPHS4Q8nKnGR4R8oKlGR4T8oKkGP4V8QLhGQ4W8PLhGQ4V8QLiGo3W8SLfGo3Y8RLfGo3Y8RLfGo3Y8RLeGP4Z8QLeGo3Z8SLdGo3[8RLdGo3]8QLaGP4`8oK^GR4d8mK[GT4f8kKYGV4h8iKVGY4k8fKTGZ4l8gKSGZ4l8hKQGZ4n8gKQGZ4n8gKPGZ4P9gKoFZ4P9k00O1O1O1OAYJgGf5Z8\\JdGc5\\8`JbG_5^8eJ_GZ5b8hJ\\GW5d8lJZGS5f8c010O1O101N1O100O2O0O1O2O0O1O2O0O1O2`M^I`Lb6`3dIYL]6g3hISLZ6l3lInKT6R4RJgKo5Y4WJ`Kk5_4ZJ\\Ki5a4]JXKi5c4]JVKh5f4^JTKf5h4`JQKe5k4`JoJf5l4[2L4L5K4L4L4L5K4L4L5K4L4L4L5K4L4L4L5K2N3M3M3M3M2N3M3N2M3M2N3J6H8H8G9H7Inl5"}, "label": "chili and cheese on fries"}]}
{"id": 276740, "image": "COCO_train2014_000000276740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"delicious food\"."}], "task": "refering", "answer_template": "The \"delicious food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 378, 379, 380, 381, 382], "bbox": [0.3944375, 0.1728333333333333, 0.6555625, 0.9876458333333333], "iscrowd": 0, "area": 53441.75939999999, "rle": {"size": [480, 640], "counts": "S^f31f><E;^CKn9=nEHk9>PFHh9`0SFDg9b0TFDf9b0UFCe9c0UFBf9d0TF@h9f0nECl9l2nNS1B>B=J7J5L3L4J6\\Od0J7K4K5L4K6K3\\NfG^L_8[3jG_LY8^3mG\\LW8b3mGWLW8h3b1M2O2N1O2N2O0O2N2O1N2O0O2O1N2N2O0O100O1O100O100O100O1O100O100O1O10000000001O00000001O000000000000000000000000000000001O0000000000000000000000000000001O001O001O000O2O001O001O001N101O001O001N101O001O001N1O2N1O2N1O2N1O2N1O2O2dFYLZ7i3cHZL[7i3aH[L\\7g3`H]L]7f3`H^L\\7f3`H]L]7f3`H^L\\7e3aH^L\\7f3`H`LZ7c3cHdLV7_3gHhLR7\\3kHhLP7]3kHhLo6^3lHgL\\3dNfL<T1^4fNgLS3;oMT3jNeLo2a0nM]5l1jJmM`5j1Z4_O6J6K5J6J6J6G9GZlV3"}, "label": "delicious food"}]}
{"id": 276740, "image": "COCO_train2014_000000276740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"onion rings with sauce\"."}], "task": "refering", "answer_template": "The \"onion rings with sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 378, 379, 380, 381, 382], "bbox": [0.3944375, 0.1728333333333333, 0.6555625, 0.9876458333333333], "iscrowd": 0, "area": 53441.75939999999, "rle": {"size": [480, 640], "counts": "S^f31f><E;^CKn9=nEHk9>PFHh9`0SFDg9b0TFDf9b0UFCe9c0UFBf9d0TF@h9f0nECl9l2nNS1B>B=J7J5L3L4J6\\Od0J7K4K5L4K6K3\\NfG^L_8[3jG_LY8^3mG\\LW8b3mGWLW8h3b1M2O2N1O2N2O0O2N2O1N2O0O2O1N2N2O0O100O1O100O100O100O1O100O100O1O10000000001O00000001O000000000000000000000000000000001O0000000000000000000000000000001O001O001O000O2O001O001O001N101O001O001N101O001O001N1O2N1O2N1O2N1O2N1O2O2dFYLZ7i3cHZL[7i3aH[L\\7g3`H]L]7f3`H^L\\7f3`H]L]7f3`H^L\\7e3aH^L\\7f3`H`LZ7c3cHdLV7_3gHhLR7\\3kHhLP7]3kHhLo6^3lHgL\\3dNfL<T1^4fNgLS3;oMT3jNeLo2a0nM]5l1jJmM`5j1Z4_O6J6K5J6J6J6G9GZlV3"}, "label": "onion rings with sauce"}]}
{"id": 178436, "image": "COCO_train2014_000000178436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor facing up\"."}], "task": "refering", "answer_template": "The \"the monitor facing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 155, 156, 157], "bbox": [0.66875, 0.24060747663551402, 0.9302343750000001, 0.4473364485981309], "iscrowd": 0, "area": 8194.94275, "rle": {"size": [428, 640], "counts": "WQc51[=0O2N1O2N1O1O2N100O2N1O1O2N1O1O2O0O1O2N1O1O2N1O2O0O1O2N1O1O2N1O1O2O0O1O2N1O1O2N1O101N1O2N1O1O2N1O101N1O1O2N1O101O00010O00001O010O00001O01O01O000010O0001O000010O0001O0010O0001O00010O00001O01O01O00001O01O01O001O01O01O000010O000O1O1O2N1N2O1O1O1N3N1O1O1N2O2N1O1O1N2O2N1O1N2O1O2N1O1N2O1O1O2M2O1O1O1O2M2O1O1O1N3N1O1O1N2O1O2N1O1N2O1O2N1N2OQfb0"}, "label": "the monitor facing up"}]}
{"id": 178436, "image": "COCO_train2014_000000178436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer laying down flat\"."}], "task": "refering", "answer_template": "The \"a computer laying down flat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 155, 156, 157], "bbox": [0.66875, 0.24060747663551402, 0.9302343750000001, 0.4473364485981309], "iscrowd": 0, "area": 8194.94275, "rle": {"size": [428, 640], "counts": "WQc51[=0O2N1O2N1O1O2N100O2N1O1O2N1O1O2O0O1O2N1O1O2N1O2O0O1O2N1O1O2N1O1O2O0O1O2N1O1O2N1O101N1O2N1O1O2N1O101N1O1O2N1O101O00010O00001O010O00001O01O01O000010O0001O000010O0001O0010O0001O00010O00001O01O01O00001O01O01O001O01O01O000010O000O1O1O2N1N2O1O1O1N3N1O1O1N2O2N1O1O1N2O2N1O1N2O1O2N1O1N2O1O1O2M2O1O1O1O2M2O1O1O1N3N1O1O1N2O1O2N1O1N2O1O2N1N2OQfb0"}, "label": "a computer laying down flat"}]}
{"id": 284934, "image": "COCO_train2014_000000284934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in a blue dress\"."}], "task": "refering", "answer_template": "The \"a little girl in a blue dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 87, 88, 89, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318], "bbox": [0.662875, 0.17527522935779816, 0.92771875, 0.8472018348623853], "iscrowd": 0, "area": 28915.322750000003, "rle": {"size": [436, 640], "counts": "mjd51b=2M3N2N2N3L3N2N2M3N3M1N2O1N2O2N1N2O1N2O1O1N2O1O1N2YDaNX;`1hDaNU;a1kDaNR;`1nDaNP;`1PEbNl:a1SE`Nk:a1UEaNg:a1YE`N\\:j1dEXNo9S2QFnMd9\\2\\FeMZ9d2fF\\MQ9m2oFRMQ9o2oFQMP9P3PGPMo8R3PGnLo8T3PGlLn8W3QGiLn8Z3PGfLo8\\3PGdLo8^3PGbLo8`3PG`Lo8b3PG^Ln8e3QG[Ln8g3QGYLn8W4O1O1YOZK^Hh4`7\\K]He4b7^K[Hc4c7aKZH`4e7cKXH^4g7eKVH\\4h7hKTH^4g7l0J2O2M2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N3QN[IlLf6T3fI^L\\6a3PJSLQ6m3[JeKg5Z4j10000O10000O10000O10000O2O000O11O0000000000001O000000000L4G9N3O0000000O1000000000001O1O2N1O1O1O1O1O1O2N1cKbJb0_5ZOgJb0[5ZOlJb0U5ZOQKc0P5YOVKd0k4XO\\Kd0e4XOaKe0a4VOeKg0\\4UOkKg0V4UOPLh0U4POQLm0\\4dNjKX1c4XNcKe1k4jM[KS2R5]MUK_2S8M4L3M3L5L3M3M4L3M3M4L3M4L3M4L3M4B=_OTYc0"}, "label": "a little girl in a blue dress"}]}
{"id": 284934, "image": "COCO_train2014_000000284934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in a blue dress near a man\"."}], "task": "refering", "answer_template": "The \"a little girl in a blue dress near a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 87, 88, 89, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318], "bbox": [0.662875, 0.17527522935779816, 0.92771875, 0.8472018348623853], "iscrowd": 0, "area": 28915.322750000003, "rle": {"size": [436, 640], "counts": "mjd51b=2M3N2N2N3L3N2N2M3N3M1N2O1N2O2N1N2O1N2O1O1N2O1O1N2YDaNX;`1hDaNU;a1kDaNR;`1nDaNP;`1PEbNl:a1SE`Nk:a1UEaNg:a1YE`N\\:j1dEXNo9S2QFnMd9\\2\\FeMZ9d2fF\\MQ9m2oFRMQ9o2oFQMP9P3PGPMo8R3PGnLo8T3PGlLn8W3QGiLn8Z3PGfLo8\\3PGdLo8^3PGbLo8`3PG`Lo8b3PG^Ln8e3QG[Ln8g3QGYLn8W4O1O1YOZK^Hh4`7\\K]He4b7^K[Hc4c7aKZH`4e7cKXH^4g7eKVH\\4h7hKTH^4g7l0J2O2M2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N3QN[IlLf6T3fI^L\\6a3PJSLQ6m3[JeKg5Z4j10000O10000O10000O10000O2O000O11O0000000000001O000000000L4G9N3O0000000O1000000000001O1O2N1O1O1O1O1O1O2N1cKbJb0_5ZOgJb0[5ZOlJb0U5ZOQKc0P5YOVKd0k4XO\\Kd0e4XOaKe0a4VOeKg0\\4UOkKg0V4UOPLh0U4POQLm0\\4dNjKX1c4XNcKe1k4jM[KS2R5]MUK_2S8M4L3M3L5L3M3M4L3M3M4L3M4L3M4L3M4B=_OTYc0"}, "label": "a little girl in a blue dress near a man"}]}
{"id": 284934, "image": "COCO_train2014_000000284934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man carrying a bow\"."}], "task": "refering", "answer_template": "The \"a man carrying a bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 33, 34, 37, 56, 57, 59, 60, 61, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 288, 289, 290], "bbox": [0.41334375, 0.03332568807339449, 0.7363593749999999, 0.8063761467889908], "iscrowd": 0, "area": 40673.32935, "rle": {"size": [436, 640], "counts": "hTa32a=3N3L3WLFTJ:k55fILZ6b0XI^Oh6P1jHoNW7_1ZHbNf7l1lGTNT8Z2^GfMb8h2oFYMQ9Q3eFoL[9Q3eFnL\\9R3cFoL]9Q3cFoL]9Q3bFPMj4CD]3bKPMi4HAX3fKPMi4M\\OS3jKPMj42XOn2nKPMj47SOi2SLPMi4<POd2VLQMj4`0lN^2ZLRMi4e0iNY2^LRMi4j0dNT2bLSMi4n0aNo1fLRMj4T1[Nj1kLRMg4[1ZNc1nLSMd4c1YNZ1SMSM_4l1ZNb1W1TLWLa2`2[1Y1VLQLc2f2V1Y1ROgNm0Y1SOmNg0R1[OmNe0S1\\OlNd0S1]OmNc0S1^OlNb0T1^OkNb0U1@jN`0V1@jN`0U1BjN>V1BjN>U1DjN<V1D_Nf0a1UMYLl0U2o1b1RM^Ln0o1P2c1nLeLo0h1S2d1jLiLR1b1T2f1eLnLU1\\1V2g1aLSMW1U1W2j1\\LXM[1n0Y2k1VL_M_1e0[2e3cM[L]2g3bMXL^2k3`MTL`2n3^MRLb2P4]MoKb2T4\\MlKd2V4[MiKe2X4ZMhKf2X4[MgKe2Z4ZMgKe2Y4\\MfKc2[4^MdKb2]4]McKc2]NVL`5X1SL`2XN_Le5P1SLa2RNfLj5j0cKgNPO\\4\\5mLcKiNPO[4]5lLaKkNROY4\\5lL`KoNROV4]5lL_KPOSOU4]5lL^KQOoNEN^4d5lLUK\\OWO^OMZ4h5lL`J0H^3g5S3O1O100OO100000000000000000000O0100O1O100O100000000000O100000000000001O000O100OL4N2M4M2N2M4M2N2M4N11O0010O00010O01O01O100O100O100O1O010001O001O001O0O2N1O2N1O2O0M4K4K6J7J6M4eInIm4U6oJmIP5V6mJlIQ5W6kJlIS5W6jJjIU5\\6dJgIZ5`6^JcI`5e6XJ]If5Y7O2N2N1O2M3M2N3\\Od0SOl0G:M3M2N3M3M2N3M3M3M2O2N2N1O2N2N1O2N2H7N3M3M2N3DWDiNl;f0i0_OnjW2"}, "label": "a man carrying a bow"}]}
{"id": 284934, "image": "COCO_train2014_000000284934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a great man with bow and arrow\"."}], "task": "refering", "answer_template": "The \"a great man with bow and arrow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 33, 34, 37, 56, 57, 59, 60, 61, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 288, 289, 290], "bbox": [0.41334375, 0.03332568807339449, 0.7363593749999999, 0.8063761467889908], "iscrowd": 0, "area": 40673.32935, "rle": {"size": [436, 640], "counts": "hTa32a=3N3L3WLFTJ:k55fILZ6b0XI^Oh6P1jHoNW7_1ZHbNf7l1lGTNT8Z2^GfMb8h2oFYMQ9Q3eFoL[9Q3eFnL\\9R3cFoL]9Q3cFoL]9Q3bFPMj4CD]3bKPMi4HAX3fKPMi4M\\OS3jKPMj42XOn2nKPMj47SOi2SLPMi4<POd2VLQMj4`0lN^2ZLRMi4e0iNY2^LRMi4j0dNT2bLSMi4n0aNo1fLRMj4T1[Nj1kLRMg4[1ZNc1nLSMd4c1YNZ1SMSM_4l1ZNb1W1TLWLa2`2[1Y1VLQLc2f2V1Y1ROgNm0Y1SOmNg0R1[OmNe0S1\\OlNd0S1]OmNc0S1^OlNb0T1^OkNb0U1@jN`0V1@jN`0U1BjN>V1BjN>U1DjN<V1D_Nf0a1UMYLl0U2o1b1RM^Ln0o1P2c1nLeLo0h1S2d1jLiLR1b1T2f1eLnLU1\\1V2g1aLSMW1U1W2j1\\LXM[1n0Y2k1VL_M_1e0[2e3cM[L]2g3bMXL^2k3`MTL`2n3^MRLb2P4]MoKb2T4\\MlKd2V4[MiKe2X4ZMhKf2X4[MgKe2Z4ZMgKe2Y4\\MfKc2[4^MdKb2]4]McKc2]NVL`5X1SL`2XN_Le5P1SLa2RNfLj5j0cKgNPO\\4\\5mLcKiNPO[4]5lLaKkNROY4\\5lL`KoNROV4]5lL_KPOSOU4]5lL^KQOoNEN^4d5lLUK\\OWO^OMZ4h5lL`J0H^3g5S3O1O100OO100000000000000000000O0100O1O100O100000000000O100000000000001O000O100OL4N2M4M2N2M4M2N2M4N11O0010O00010O01O01O100O100O100O1O010001O001O001O0O2N1O2N1O2O0M4K4K6J7J6M4eInIm4U6oJmIP5V6mJlIQ5W6kJlIS5W6jJjIU5\\6dJgIZ5`6^JcI`5e6XJ]If5Y7O2N2N1O2M3M2N3\\Od0SOl0G:M3M2N3M3M2N3M3M3M2O2N2N1O2N2N1O2N2H7N3M3M2N3DWDiNl;f0i0_OnjW2"}, "label": "a great man with bow and arrow"}]}
{"id": 284934, "image": "COCO_train2014_000000284934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black dress holding an apple\"."}], "task": "refering", "answer_template": "The \"a woman in a black dress holding an apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 70, 71, 72, 73, 93, 94, 95, 96, 97, 117, 118, 119, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.030625000000000003, 0.12809633027522935, 0.54040625, 1.0], "iscrowd": 0, "area": 55671.617699999995, "rle": {"size": [436, 640], "counts": "ji8j0e<n0RO9H5J7I6K6I6L5J5L4K5L4K5L4L4K4L5K5K5gK]K[Oh4?aKYOd4b0eKUO`4i0fKoNo4?VKXOP5e0VKROo4l0VKmNo4Q1TKiNQ5T1RKgNR5W1RKdNR5Z1QK`NR5_1RKZNR5e1QKVNR5j1nJSNU5l1mJoMW5P2kJlMW5U2jJgMY5X2iJcM[5\\2fJaM]5_2dJ]M_5b2cJ[M_5d2cJYMa5e2`JYMd5c2]J[Mg5b2[J\\Mh5a2ZJ\\Mk5a2f2M2O1O1N2O1O1N2O1O1N101N2O1O1O1O010O1O10O01O10O01O01O00001O001O00001O001O00001O001O00001O001O001O001O1O1O1O1O1bJfM@[2iLeM_11h1[2hLmMV1IR2[2gLlMV1JS2[2eLmMV1IU2[2dLlMV1KU2Z2dLlMU1KW2Z2cLkMV1KW2[2aLlMV1JY2Z2aLlMV1JY2]2^LjMW1J[2a2YLeM[1L[2d2UL`M_1M\\2h2oK]Mb1M_2h2mK[Mb1Oa2h2jK[Mb1Od2h2gKZMc11e2h2dKYMd11h2h2`KYMf11j2k2ZKVMi11m2T3nJlLS22o2U4kLQLU3P4`L[L_3e3\\L`Ld3a3[L_Le3a3ZL`Lf3`3ZL`Lf3a3YL`Lf3`3ZL`Lf3a3YL_Lg3a3YL_Lg3a3YL_Lg3b3XL^Lh3b3XL_Lg3a3YL_Lg3b3XL^Lh3b3XL^Lh3c3WL]Li3c3WL]Li3c3WL^Lg3d3XL\\Lh3d3WL]Li3d3VL\\Lj3d3VL\\Lj3d3VL\\Lj3e3UL\\Lj3d3VL\\Lj3d3VL\\Lj3e3UL[Lk3e3UL[Lk3f3TLZLl3f3TL[Lk3e3UL[Lk3f3TLZLl3f3TL[Lk3f3TLZLl3f3TLZLl3f3TL[Lk3f3SL[Lm3e3SL[Lm3e3SL\\Ll3e3SL[Lm3e3SL[Lm3e3SL[Lm3e3SL\\Ll3e3SL[Lm3e3SL[Lm3e3SL\\Ll3d3TL\\Ll3e3SL[Lm3e3SL\\Ll3d3TL]Lk3c3UL]Lk3c3UL^Lj3b3UL`Lj3`3VL`Lj3`3VLaLi3_3WLbLh3^3XLbLh3^3XLcLg3]3YLdLf3\\3ZLeLe3[3[LeLe3Z3\\LgLc3Y3]LhLa3Y3_LgL`3Z3`LgL_3Y3aLhL]3Y3cLgL]3Y3cLhL[3Y3dLjLZ3V3fLlLW3U3iLmLT3T3lLoLQ3Q3oLQMn2P3RMRMl2n2TMTMi2m2WMUMg2k2YMWMd2i2]MYM`2h2`MZM^2f2bM\\M[2e2eM]MY2c2gM_MV2b2jM`MQ2aLQMo5n0cMf1h2YN]MZ1l2fNXMU1i2kN[MP1f2PO^Mk0c2UOaMf0`2ZObMc0_2]OdM?]2AfM:\\2FgMJ_KAk6e0iMG\\KCk6g0lMCZKDk6i0nM_OXKGk6i0QN]OUKHk6k0TNYOQKKl6l0XNSOQ2m0[5000000000000000O010000000000000000000000O01001O00000000001O00000O101O00000000001O001O000O2O001O00001O00001O001N10001O0000000000000000O01000000000000000O10000000000O1000O10O10000O1000000O11O3M2N2N3M2N2N3M2N_ol3"}, "label": "a woman in a black dress holding an apple"}]}
{"id": 284934, "image": "COCO_train2014_000000284934.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older lady in polka dot skirt sits with an apple on her hand\"."}], "task": "refering", "answer_template": "The \"an older lady in polka dot skirt sits with an apple on her hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 70, 71, 72, 73, 93, 94, 95, 96, 97, 117, 118, 119, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.030625000000000003, 0.12809633027522935, 0.54040625, 1.0], "iscrowd": 0, "area": 55671.617699999995, "rle": {"size": [436, 640], "counts": "ji8j0e<n0RO9H5J7I6K6I6L5J5L4K5L4K5L4L4K4L5K5K5gK]K[Oh4?aKYOd4b0eKUO`4i0fKoNo4?VKXOP5e0VKROo4l0VKmNo4Q1TKiNQ5T1RKgNR5W1RKdNR5Z1QK`NR5_1RKZNR5e1QKVNR5j1nJSNU5l1mJoMW5P2kJlMW5U2jJgMY5X2iJcM[5\\2fJaM]5_2dJ]M_5b2cJ[M_5d2cJYMa5e2`JYMd5c2]J[Mg5b2[J\\Mh5a2ZJ\\Mk5a2f2M2O1O1N2O1O1N2O1O1N101N2O1O1O1O010O1O10O01O10O01O01O00001O001O00001O001O00001O001O00001O001O001O001O1O1O1O1O1bJfM@[2iLeM_11h1[2hLmMV1IR2[2gLlMV1JS2[2eLmMV1IU2[2dLlMV1KU2Z2dLlMU1KW2Z2cLkMV1KW2[2aLlMV1JY2Z2aLlMV1JY2]2^LjMW1J[2a2YLeM[1L[2d2UL`M_1M\\2h2oK]Mb1M_2h2mK[Mb1Oa2h2jK[Mb1Od2h2gKZMc11e2h2dKYMd11h2h2`KYMf11j2k2ZKVMi11m2T3nJlLS22o2U4kLQLU3P4`L[L_3e3\\L`Ld3a3[L_Le3a3ZL`Lf3`3ZL`Lf3a3YL`Lf3`3ZL`Lf3a3YL_Lg3a3YL_Lg3a3YL_Lg3b3XL^Lh3b3XL_Lg3a3YL_Lg3b3XL^Lh3b3XL^Lh3c3WL]Li3c3WL]Li3c3WL^Lg3d3XL\\Lh3d3WL]Li3d3VL\\Lj3d3VL\\Lj3d3VL\\Lj3e3UL\\Lj3d3VL\\Lj3d3VL\\Lj3e3UL[Lk3e3UL[Lk3f3TLZLl3f3TL[Lk3e3UL[Lk3f3TLZLl3f3TL[Lk3f3TLZLl3f3TLZLl3f3TL[Lk3f3SL[Lm3e3SL[Lm3e3SL\\Ll3e3SL[Lm3e3SL[Lm3e3SL[Lm3e3SL\\Ll3e3SL[Lm3e3SL[Lm3e3SL\\Ll3d3TL\\Ll3e3SL[Lm3e3SL\\Ll3d3TL]Lk3c3UL]Lk3c3UL^Lj3b3UL`Lj3`3VL`Lj3`3VLaLi3_3WLbLh3^3XLbLh3^3XLcLg3]3YLdLf3\\3ZLeLe3[3[LeLe3Z3\\LgLc3Y3]LhLa3Y3_LgL`3Z3`LgL_3Y3aLhL]3Y3cLgL]3Y3cLhL[3Y3dLjLZ3V3fLlLW3U3iLmLT3T3lLoLQ3Q3oLQMn2P3RMRMl2n2TMTMi2m2WMUMg2k2YMWMd2i2]MYM`2h2`MZM^2f2bM\\M[2e2eM]MY2c2gM_MV2b2jM`MQ2aLQMo5n0cMf1h2YN]MZ1l2fNXMU1i2kN[MP1f2PO^Mk0c2UOaMf0`2ZObMc0_2]OdM?]2AfM:\\2FgMJ_KAk6e0iMG\\KCk6g0lMCZKDk6i0nM_OXKGk6i0QN]OUKHk6k0TNYOQKKl6l0XNSOQ2m0[5000000000000000O010000000000000000000000O01001O00000000001O00000O101O00000000001O001O000O2O001O00001O00001O001N10001O0000000000000000O01000000000000000O10000000000O1000O10O10000O1000000O11O3M2N2N3M2N2N3M2N_ol3"}, "label": "an older lady in polka dot skirt sits with an apple on her hand"}]}
{"id": 6407, "image": "COCO_train2014_000000006407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bananas sitting in front of the woman\"."}], "task": "refering", "answer_template": "The \"the bananas sitting in front of the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.05378666666666667, 0.7209, 0.68896, 1.0], "iscrowd": 0, "area": 22505.914899999992, "rle": {"size": [500, 375], "counts": "UW:1a?5K4L5J5L5K4L4L5K4L2N3N2M3N1O2M3N2M2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O100O100O1O100O100O1O100O100O100O1000000O10000O1000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000O1O1O1N2O1L4L4L4N2O100O100O100000000100O1O1O1O2N1O1O1OO100O100O100O1O100O10000O10000O1000000001O000000001O01O000001O2NL4M3M300O1O1O1O1O12N2N1O2K4L5K5K4L5K4L5K5K4O0ON2M4M2N21O100O1O1O100O1O1O10000000000000001O0O101N100O2O000O2O4K6K5J5La0^O5L1N2N1O2N2N1O2M3N1O2N1O2N2L3M4K5L3MQgh1"}, "label": "the bananas sitting in front of the woman"}]}
{"id": 6407, "image": "COCO_train2014_000000006407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"larger bunch of bananas\"."}], "task": "refering", "answer_template": "The \"larger bunch of bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.05378666666666667, 0.7209, 0.68896, 1.0], "iscrowd": 0, "area": 22505.914899999992, "rle": {"size": [500, 375], "counts": "UW:1a?5K4L5J5L5K4L4L5K4L2N3N2M3N1O2M3N2M2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O100O100O1O100O100O1O100O100O100O1000000O10000O1000000O1000000000000O10000000000O10000000000O10000000000O1000000000000O10000O1O1O1N2O1L4L4L4N2O100O100O100000000100O1O1O1O2N1O1O1OO100O100O100O1O100O10000O10000O1000000001O000000001O01O000001O2NL4M3M300O1O1O1O1O12N2N1O2K4L5K5K4L5K4L5K5K4O0ON2M4M2N21O100O1O1O100O1O1O10000000000000001O0O101N100O2O000O2O4K6K5J5La0^O5L1N2N1O2N2N1O2M3N1O2N1O2N2L3M4K5L3MQgh1"}, "label": "larger bunch of bananas"}]}
{"id": 6407, "image": "COCO_train2014_000000006407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with an orange skirt and sandals on looks on at her friend\"."}], "task": "refering", "answer_template": "The \"a woman with an orange skirt and sandals on looks on at her friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 51, 63, 64, 75, 76, 77, 88, 89, 90, 101, 102, 103, 114, 115, 116, 127, 128, 129, 140, 141, 142, 153, 154, 155, 166, 167, 168, 180], "bbox": [0.7952800000000001, 0.10204, 0.9961333333333333, 0.73766], "iscrowd": 0, "area": 15805.119550000001, "rle": {"size": [500, 375], "counts": "cfa42]?9H7I8ZDWOT8R1_GXO\\8o0YGYOc8o0PG[Oj8n0jF[OQ9l0cF^OX9j0[F@_9i0UF@i9d0kEF?dMn8Z5PGjJm8X5PGlJm8V5QGlJm8V5PGnJm8T5QGoJm8R5PGRKm8o4RGRKm8P5PGRKo8k5N1O1O2N1O100O2N1O1O1O2N1O101N100O101N100O101O0O101N01L3N3M2NG8C>D<001N101000O100O010O100O5L3M4M201N101N101N10\\H`e0"}, "label": "a woman with an orange skirt and sandals on looks on at her friend"}]}
{"id": 6407, "image": "COCO_train2014_000000006407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman to the right of the woman holding money\"."}], "task": "refering", "answer_template": "The \"the woman to the right of the woman holding money\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 51, 63, 64, 75, 76, 77, 88, 89, 90, 101, 102, 103, 114, 115, 116, 127, 128, 129, 140, 141, 142, 153, 154, 155, 166, 167, 168, 180], "bbox": [0.7952800000000001, 0.10204, 0.9961333333333333, 0.73766], "iscrowd": 0, "area": 15805.119550000001, "rle": {"size": [500, 375], "counts": "cfa42]?9H7I8ZDWOT8R1_GXO\\8o0YGYOc8o0PG[Oj8n0jF[OQ9l0cF^OX9j0[F@_9i0UF@i9d0kEF?dMn8Z5PGjJm8X5PGlJm8V5QGlJm8V5PGnJm8T5QGoJm8R5PGRKm8o4RGRKm8P5PGRKo8k5N1O1O2N1O100O2N1O1O1O2N1O101N100O101N100O101O0O101N01L3N3M2NG8C>D<001N101000O100O010O100O5L3M4M201N101N101N10\\H`e0"}, "label": "the woman to the right of the woman holding money"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book entitled understanding the linux kernel\"."}], "task": "refering", "answer_template": "The \"a book entitled understanding the linux kernel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.8004460093896714, 0.359765625, 1.0], "iscrowd": 0, "area": 11293.386649999999, "rle": {"size": [426, 640], "counts": "o<;Z<e0J600O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O010000000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O10O10000O0100000O10O1000O1000O01000000O01000O1000O10O10000O0100000O0100000O10O10O100000O010000O10O1000O1000O01000000O0100000O10O10O1000O10O10000O0100000O10O10O100000O01000000O01000O1000O10O10000O0100000O10O1000O10O10O100000O010000O10O1000O1000O01000000O0100000O1N1@a0@oWZ5"}, "label": "a book entitled understanding the linux kernel"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"linux kernel book\"."}], "task": "refering", "answer_template": "The \"linux kernel book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.8004460093896714, 0.359765625, 1.0], "iscrowd": 0, "area": 11293.386649999999, "rle": {"size": [426, 640], "counts": "o<;Z<e0J600O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O010000000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O10O10000O0100000O10O1000O1000O01000000O01000O1000O10O10000O0100000O0100000O10O10O100000O010000O10O1000O1000O01000000O0100000O10O10O1000O10O10000O0100000O10O10O100000O01000000O01000O1000O10O10000O0100000O10O1000O10O10O100000O010000O10O1000O1000O01000000O0100000O1N1@a0@oWZ5"}, "label": "linux kernel book"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leftmost computer monitor that is being blocked by the white lamp\"."}], "task": "refering", "answer_template": "The \"the leftmost computer monitor that is being blocked by the white lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196], "bbox": [0.08040625, 0.006924882629107981, 0.55096875, 0.5773943661971831], "iscrowd": 0, "area": 59585.67655, "rle": {"size": [426, 640], "counts": "[de01Y=1O010O1O1Og0YOZ2fMY2gMU1lN0O0000000001O00000001O0001O0000000000001O01O00000001O0000000001O01O0000000000001O01O0000000001O00000000010O0000000000001O0001O00000001O0000000001O01O00000000001O0001O00000001O0000000000010O00000000001O000001O000001O0000000000010O0000000000001O000001O0001O00000000000010O000000000001O000001O0001O0000000000001O01O0000000001O00000001O1O3N1N3M3M2N0001O00000O1000001O00000000000O2O0000000000001O00000O10001O0000000000001O0O10000000001O000000000O2O0000000000001O00000O1000001O00000000000O2O00000000001O0000000O10001O0000000000001O0O10000000001O000000000O2O0000000000001O00000O1UOk0iNW1iNX1hNW1iN]\\g3"}, "label": "the leftmost computer monitor that is being blocked by the white lamp"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"monitor that is closer to the books\"."}], "task": "refering", "answer_template": "The \"monitor that is closer to the books\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196], "bbox": [0.08040625, 0.006924882629107981, 0.55096875, 0.5773943661971831], "iscrowd": 0, "area": 59585.67655, "rle": {"size": [426, 640], "counts": "[de01Y=1O010O1O1Og0YOZ2fMY2gMU1lN0O0000000001O00000001O0001O0000000000001O01O00000001O0000000001O01O0000000000001O01O0000000001O00000000010O0000000000001O0001O00000001O0000000001O01O00000000001O0001O00000001O0000000000010O00000000001O000001O000001O0000000000010O0000000000001O000001O0001O00000000000010O000000000001O000001O0001O0000000000001O01O0000000001O00000001O1O3N1N3M3M2N0001O00000O1000001O00000000000O2O0000000000001O00000O10001O0000000000001O0O10000000001O000000000O2O0000000000001O00000O1000001O00000000000O2O00000000001O0000000O10001O0000000000001O0O10000000001O000000000O2O0000000000001O00000O1UOk0iNW1iNX1hNW1iN]\\g3"}, "label": "monitor that is closer to the books"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one black color computer keyboard on the table\"."}], "task": "refering", "answer_template": "The \"there is one black color computer keyboard on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335], "bbox": [0.381421875, 0.7573004694835681, 0.7718281250000001, 0.9865258215962441], "iscrowd": 0, "area": 11890.847149999998, "rle": {"size": [426, 640], "counts": "[mU31T=5L4K5L4K5K5L4K5O1O10000O10000O1000000O10O10O10000O10000O10000O10000O10000O10000O10000O1000000O10000O01000O10000O10O10O10000O01000O10O10O1000O10O10000O01000O10O10O1000O010000O10O10O1000O010000O0100000O01000O1000O010000O01000O10O10O10000O01000O10O10O1000O10O10000O01000O10O10O1000O010000O10O10O1000O010000O0100000O01000O1000O010000O01000O10O10O1000O010O10O0100O010O100O010O10O0100O100O2O0O2O0O2O0O2O0O2N101N101N101N101N101N101N101N101N101N101N101N100O2O0O2N101N101N101N101N101N101N101N101N101N101N10e\\l1"}, "label": "there is one black color computer keyboard on the table"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black keyboard nearest the table edge\"."}], "task": "refering", "answer_template": "The \"the black keyboard nearest the table edge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335], "bbox": [0.381421875, 0.7573004694835681, 0.7718281250000001, 0.9865258215962441], "iscrowd": 0, "area": 11890.847149999998, "rle": {"size": [426, 640], "counts": "[mU31T=5L4K5L4K5K5L4K5O1O10000O10000O1000000O10O10O10000O10000O10000O10000O10000O10000O10000O1000000O10000O01000O10000O10O10O10000O01000O10O10O1000O10O10000O01000O10O10O1000O010000O10O10O1000O010000O0100000O01000O1000O010000O01000O10O10O10000O01000O10O10O1000O10O10000O01000O10O10O1000O010000O10O10O1000O010000O0100000O01000O1000O010000O01000O10O10O1000O010O10O0100O010O100O010O10O0100O100O2O0O2O0O2O0O2O0O2N101N101N101N101N101N101N101N101N101N101N101N100O2O0O2N101N101N101N101N101N101N101N101N101N101N10e\\l1"}, "label": "the black keyboard nearest the table edge"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white book called red hat linux certification\"."}], "task": "refering", "answer_template": "The \"a white book called red hat linux certification\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255], "bbox": [0.0015, 0.5470892018779343, 0.36415625, 0.7668075117370892], "iscrowd": 0, "area": 16062.792449999997, "rle": {"size": [426, 640], "counts": "Re0_2k:O10000000000O1000000000O10O1000000000000O100000000000O0100000000000000O100000000000O01000000000000O1000000000000O0100000000000O100000000000000O0100000000000O1000000000000O0100000000000O1000000000000O10O100000000000O1000000000000O10O1000000000O1000000000000O1000O10000000O1000000000000O100000O10000000O1000000000000O1000O10000000O1000000000000O100000O100000O100000000000000O100000O100000O1000000000000O10000000O1000O1000000000000O1000000000O1000O100000005K`0@`0_Oa0@XQY5"}, "label": "a white book called red hat linux certification"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book entitled red hat linux certification study guide\"."}], "task": "refering", "answer_template": "The \"a book entitled red hat linux certification study guide\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255], "bbox": [0.0015, 0.5470892018779343, 0.36415625, 0.7668075117370892], "iscrowd": 0, "area": 16062.792449999997, "rle": {"size": [426, 640], "counts": "Re0_2k:O10000000000O1000000000O10O1000000000000O100000000000O0100000000000000O100000000000O01000000000000O1000000000000O0100000000000O100000000000000O0100000000000O1000000000000O0100000000000O1000000000000O10O100000000000O1000000000000O10O1000000000O1000000000000O1000O10000000O1000000000000O100000O10000000O1000000000000O1000O10000000O1000000000000O100000O100000O100000000000000O100000O100000O1000000000000O10000000O1000O1000000000000O1000000000O1000O100000005K`0@`0_Oa0@XQY5"}, "label": "a book entitled red hat linux certification study guide"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book of ubuntu linux\"."}], "task": "refering", "answer_template": "The \"a book of ubuntu linux\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301], "bbox": [0.0, 0.6841784037558685, 0.362765625, 0.9105164319248826], "iscrowd": 0, "area": 12374.252400000005, "rle": {"size": [426, 640], "counts": "Z:j1`;O10O100000O1000O1000O100000O10O100000O10O1000000O010000000O10O100000O1000O1000O100000O10O1000000O010000000O10O100000O1000O1000O1000O1000O100000O0100000000O010000000O10O100000O1000O10O10000000O0100000000O010000000O10O100000O10O1000O100000O10O10000000O0100000000O010000000O10O1000O100000O10O10000000O010000000O0100000000O0100000O1000O1000O100000O10O10000000O0100000000O0100000O1000O1000O100000O10O100000O10O10000000O01000000O10O100000O1000O1000O100000O11N2O1O001OY^Y5"}, "label": "a book of ubuntu linux"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red book\"."}], "task": "refering", "answer_template": "The \"red book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301], "bbox": [0.0, 0.6841784037558685, 0.362765625, 0.9105164319248826], "iscrowd": 0, "area": 12374.252400000005, "rle": {"size": [426, 640], "counts": "Z:j1`;O10O100000O1000O1000O100000O10O100000O10O1000000O010000000O10O100000O1000O1000O100000O10O1000000O010000000O10O100000O1000O1000O1000O1000O100000O0100000000O010000000O10O100000O1000O10O10000000O0100000000O010000000O10O100000O10O1000O100000O10O10000000O0100000000O010000000O10O1000O100000O10O10000000O010000000O0100000000O0100000O1000O1000O100000O10O10000000O0100000000O0100000O1000O1000O100000O10O100000O10O10000000O01000000O10O100000O1000O1000O100000O11N2O1O001OY^Y5"}, "label": "red book"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first monitor from the right\"."}], "task": "refering", "answer_template": "The \"the first monitor from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224], "bbox": [0.5551875, 0.10448356807511737, 0.985, 0.6582863849765258], "iscrowd": 0, "area": 50948.74985, "rle": {"size": [426, 640], "counts": "]mc42d<d0\\Od0[Oe0\\Od0\\Od0\\Od0\\Od0[Oe0\\Od0I7N2N2O100000000001O000000000000001O000000000000000000000000000O2O00000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O000000000000000000000O1000001O0002N>B?A:F000000000000000001O00000000000000000000000000001O000000000000000000000iNkJYJU5g5W10001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O0000000000000000000000000000001O00000000000000000O100000000000001O0000000000000H8_Oa0_Oa0^Ob0_Oa0_Oa0_Oa0_Oa0_Ob0^OfS4"}, "label": "the first monitor from the right"}]}
{"id": 260360, "image": "COCO_train2014_000000260360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"screen to right of other screen\"."}], "task": "refering", "answer_template": "The \"screen to right of other screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 223, 224], "bbox": [0.5551875, 0.10448356807511737, 0.985, 0.6582863849765258], "iscrowd": 0, "area": 50948.74985, "rle": {"size": [426, 640], "counts": "]mc42d<d0\\Od0[Oe0\\Od0\\Od0\\Od0\\Od0[Oe0\\Od0I7N2N2O100000000001O000000000000001O000000000000000000000000000O2O00000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O000000000000000000000O1000001O0002N>B?A:F000000000000000001O00000000000000000000000000001O000000000000000000000iNkJYJU5g5W10001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O0000000000000000000000000000001O00000000000000000O100000000000001O0000000000000H8_Oa0_Oa0^Ob0_Oa0_Oa0_Oa0_Oa0_Ob0^OfS4"}, "label": "screen to right of other screen"}]}
{"id": 555273, "image": "COCO_train2014_000000555273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"these are housplants in the left side of the room\"."}], "task": "refering", "answer_template": "The \"these are housplants in the left side of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 348, 349, 350], "bbox": [0.03975, 0.46620833333333334, 0.314828125, 0.9242083333333333], "iscrowd": 0, "area": 21964.245600000006, "rle": {"size": [480, 640], "counts": "c_<5h>5L4K5K5K5L3L5K5L4K5K5L4K4M4K5K5M2OO10O10O10000O01000O10000O010001O001N101O1O001N101O001O0O2O001O001N101O1O001N1O2O0O2gDPM_:Q3[EXMa:j2YE^Mc:c2WEfMf:]3L4L4L3M3N2M3M4L3N2O13N2M2N3M3NeNUK^Hj4`7[KZHh4d7\\KWHg4g7]KUHf4g7_KTHd4j7`KQHc4m7aKoGa4o7dKkG`4R8dKiG^4V8W1N01000000000O10O10000000O1000O10001O1O2N1O1O1N2O2N1O1O1O2N1N2O1O1O2N1O1OO0O2O1N2O1N2O1N2O00L5L5J6J6[KUGo2P9PMWGd2o8ZMRGa2T9^MmF]2X9aMiF[2[9dMgFZ2[9eMeFY2^9eMdFY2^9eMcFk1WOiMX:;bFk1n9SNSFn1m9QNTFo1l9oMUFR2j9mMXFR2i9mMWFT2i9jMYFV2g9hMZFX2g9gMZFY2f9eM[F\\2e9cM\\F\\2e9bM\\FS1iNOl:mN]Fn0oN2e:oN]Fh0WO5m;F[D6g;DaD9_;FeD6];IeD4\\;KhD2Y;LkD0V;0mDMT;1PEKQ;4o1O0O2N101N[_]6"}, "label": "these are housplants in the left side of the room"}]}
{"id": 555273, "image": "COCO_train2014_000000555273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bushy houseplant sitting next to a desk\"."}], "task": "refering", "answer_template": "The \"a bushy houseplant sitting next to a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 348, 349, 350], "bbox": [0.03975, 0.46620833333333334, 0.314828125, 0.9242083333333333], "iscrowd": 0, "area": 21964.245600000006, "rle": {"size": [480, 640], "counts": "c_<5h>5L4K5K5K5L3L5K5L4K5K5L4K4M4K5K5M2OO10O10O10000O01000O10000O010001O001N101O1O001N101O001O0O2O001O001N101O1O001N1O2O0O2gDPM_:Q3[EXMa:j2YE^Mc:c2WEfMf:]3L4L4L3M3N2M3M4L3N2O13N2M2N3M3NeNUK^Hj4`7[KZHh4d7\\KWHg4g7]KUHf4g7_KTHd4j7`KQHc4m7aKoGa4o7dKkG`4R8dKiG^4V8W1N01000000000O10O10000000O1000O10001O1O2N1O1O1N2O2N1O1O1O2N1N2O1O1O2N1O1OO0O2O1N2O1N2O1N2O00L5L5J6J6[KUGo2P9PMWGd2o8ZMRGa2T9^MmF]2X9aMiF[2[9dMgFZ2[9eMeFY2^9eMdFY2^9eMcFk1WOiMX:;bFk1n9SNSFn1m9QNTFo1l9oMUFR2j9mMXFR2i9mMWFT2i9jMYFV2g9hMZFX2g9gMZFY2f9eM[F\\2e9cM\\F\\2e9bM\\FS1iNOl:mN]Fn0oN2e:oN]Fh0WO5m;F[D6g;DaD9_;FeD6];IeD4\\;KhD2Y;LkD0V;0mDMT;1PEKQ;4o1O0O2N101N[_]6"}, "label": "a bushy houseplant sitting next to a desk"}]}
{"id": 63754, "image": "COCO_train2014_000000063754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player watching batter\"."}], "task": "refering", "answer_template": "The \"baseball player watching batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 225, 226, 227, 228, 240, 241, 242, 243, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 331, 332], "bbox": [0.0033723653395784543, 0.613734375, 0.2565573770491803, 0.9876093750000001], "iscrowd": 0, "area": 14303.1526, "rle": {"size": [640, 427], "counts": "mP1g4V?4M2N2M4M2N3L3N2O20O0010O01O010O12M3N2L4M3M3M3M3M3M4L3M3M3M3M3LRJoBR5n<cJ`C]5]<XJPDi5T=O2N1N2N2N3M2N2OO00000000001O01O00000001O00000000001N1O100O100O1O2O0O1OI8H7H8lLPB6T>DaBIb=1jBFY=5SCAQ=<WC[Ol<d0\\CSOg<k0`ClNc<S1U3O2O0O1O2N1O101N1O1O2N1O2O0O1O2N1O101N1O2N2N2O1N2N2N2NgZV6"}, "label": "baseball player watching batter"}]}
{"id": 63754, "image": "COCO_train2014_000000063754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy behind the guy batting\"."}], "task": "refering", "answer_template": "The \"guy behind the guy batting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 225, 226, 227, 228, 240, 241, 242, 243, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 331, 332], "bbox": [0.0033723653395784543, 0.613734375, 0.2565573770491803, 0.9876093750000001], "iscrowd": 0, "area": 14303.1526, "rle": {"size": [640, 427], "counts": "mP1g4V?4M2N2M4M2N3L3N2O20O0010O01O010O12M3N2L4M3M3M3M3M3M4L3M3M3M3M3LRJoBR5n<cJ`C]5]<XJPDi5T=O2N1N2N2N3M2N2OO00000000001O01O00000001O00000000001N1O100O100O1O2O0O1OI8H7H8lLPB6T>DaBIb=1jBFY=5SCAQ=<WC[Ol<d0\\CSOg<k0`ClNc<S1U3O2O0O1O2N1O101N1O1O2N1O2O0O1O2N1O101N1O2N2N2O1N2N2N2NgZV6"}, "label": "guy behind the guy batting"}]}
{"id": 375054, "image": "COCO_train2014_000000375054.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white sink\"."}], "task": "refering", "answer_template": "The \"the white sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [330, 331, 344, 345, 346, 347, 348, 349, 350, 351, 360, 361, 362, 363, 364, 365, 366, 367, 368, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.16682403433476395, 0.8424687499999999, 0.7158154506437768, 0.999875], "iscrowd": 0, "area": 16325.915499999988, "rle": {"size": [640, 466], "counts": "l[a14gc05K5O1O1O1CBW]O?gb0DW]O=fb0GY]O9fb0IX]O8gb0JW]O7hb0JW]O7ib0`0O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100_O[Nf^Of1Wa0]Ni^Oc1Va0^Ni^Oc1Va0^Nj^Ob1Va0^Nj^Ob1Va0^Ni^Oc1Wa0]Ni^Oc1Wa0]Ni^Oc1Wa0]Ni^Oc1Va0^Nk^Oa1Ua0_Nk^Oa1Ua0^Nl^Ob1Ta0^Nl^Ob1Ta0^Nm^Oa1Sa0_Nm^Oa1Sa0_Nn^O`1Qa0`NQ_O_1o`0aNQ_O_1o`0aNR_O^1n`0bNR_O^1n`0bNP_O`1Pa0_Nn^Od1Qa0]Nm^Oe1Sa0[Ni^Oi1Wa0WNc^Oo1]a0:000000O1001O1O1O2N1O1O1O1O2N1O2N2N00O1O1O1O1O1O1O100O10000O100000000O1000000O10000O100E;O1O1O1O1O100O1000YMP_O^2o`0cMT_OZ2l`0eMX_OX2h`0hM\\_OT2c`0mM^_OR2b`0mM__OS2a`0mM__OS2``0nM__OS2a`0mM__OS2a`0lM`_OT2``0lM`_OT2_`0mMa_OS2_`0lMb_OT2^`0lMb_OT2]`0mMc_OS2]`0mMc_OS2]`0lMd_OT2[`0mMe_OS2[`0mMe_OS2[`0lMf_OT2Y`0mMg_OS2Y`0mMg_OS2\\`0jMd_OV2^`0gMc_OY2Ra000000000000001O0000000000000000000000000000001O0000000000000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O0000001O001O1O001O001O001O1O001O001O001O1O001O002N1O1N2O1N2O1N2O1L6I8H7J6I7Inla2"}, "label": "the white sink"}]}
{"id": 375054, "image": "COCO_train2014_000000375054.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bathroom sink with silver handles\"."}], "task": "refering", "answer_template": "The \"white bathroom sink with silver handles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [330, 331, 344, 345, 346, 347, 348, 349, 350, 351, 360, 361, 362, 363, 364, 365, 366, 367, 368, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.16682403433476395, 0.8424687499999999, 0.7158154506437768, 0.999875], "iscrowd": 0, "area": 16325.915499999988, "rle": {"size": [640, 466], "counts": "l[a14gc05K5O1O1O1CBW]O?gb0DW]O=fb0GY]O9fb0IX]O8gb0JW]O7hb0JW]O7ib0`0O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100_O[Nf^Of1Wa0]Ni^Oc1Va0^Ni^Oc1Va0^Nj^Ob1Va0^Nj^Ob1Va0^Ni^Oc1Wa0]Ni^Oc1Wa0]Ni^Oc1Wa0]Ni^Oc1Va0^Nk^Oa1Ua0_Nk^Oa1Ua0^Nl^Ob1Ta0^Nl^Ob1Ta0^Nm^Oa1Sa0_Nm^Oa1Sa0_Nn^O`1Qa0`NQ_O_1o`0aNQ_O_1o`0aNR_O^1n`0bNR_O^1n`0bNP_O`1Pa0_Nn^Od1Qa0]Nm^Oe1Sa0[Ni^Oi1Wa0WNc^Oo1]a0:000000O1001O1O1O2N1O1O1O1O2N1O2N2N00O1O1O1O1O1O1O100O10000O100000000O1000000O10000O100E;O1O1O1O1O100O1000YMP_O^2o`0cMT_OZ2l`0eMX_OX2h`0hM\\_OT2c`0mM^_OR2b`0mM__OS2a`0mM__OS2``0nM__OS2a`0mM__OS2a`0lM`_OT2``0lM`_OT2_`0mMa_OS2_`0lMb_OT2^`0lMb_OT2]`0mMc_OS2]`0mMc_OS2]`0lMd_OT2[`0mMe_OS2[`0mMe_OS2[`0lMf_OT2Y`0mMg_OS2Y`0mMg_OS2\\`0jMd_OV2^`0gMc_OY2Ra000000000000001O0000000000000000000000000000001O0000000000000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O0000001O001O1O001O001O001O1O001O001O001O1O001O002N1O1N2O1N2O1N2O1L6I8H7J6I7Inla2"}, "label": "white bathroom sink with silver handles"}]}
{"id": 55571, "image": "COCO_train2014_000000055571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the behind of the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the behind of the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 203, 204, 205, 226, 227, 228, 249, 250, 251, 272, 273, 274, 295, 296, 297, 298], "bbox": [0.815625, 0.4988758782201405, 0.965546875, 0.8763934426229509], "iscrowd": 0, "area": 8082.980850000001, "rle": {"size": [427, 640], "counts": "Sli61Z=2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M7I7I7I7I7J6I7I7I7I7I7I7I7TGWLh7o3fG[L\\8X42O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N3M3N2M3010O001O1O010O1O001O010O1O001O10O01O00100OB>\\Od0K;J9G8I8G9G8H8I2M3M2N3L4M3L3N3L4M3L3Nbj8"}, "label": "the behind of the zebra on the right"}]}
{"id": 55571, "image": "COCO_train2014_000000055571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the butts and tails of two zebras\"."}], "task": "refering", "answer_template": "The \"the butts and tails of two zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 181, 203, 204, 205, 226, 227, 228, 249, 250, 251, 272, 273, 274, 295, 296, 297, 298], "bbox": [0.815625, 0.4988758782201405, 0.965546875, 0.8763934426229509], "iscrowd": 0, "area": 8082.980850000001, "rle": {"size": [427, 640], "counts": "Sli61Z=2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M7I7I7I7I7J6I7I7I7I7I7I7I7TGWLh7o3fG[L\\8X42O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N3M3N2M3010O001O1O010O1O001O010O1O001O10O01O00100OB>\\Od0K;J9G8I8G9G8H8I2M3M2N3L4M3L3N3L4M3L3Nbj8"}, "label": "the butts and tails of two zebras"}]}
{"id": 55571, "image": "COCO_train2014_000000055571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra closest to the camera\"."}], "task": "refering", "answer_template": "The \"the zebra closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 148, 149, 150, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 292, 293, 294, 295, 311, 316, 317, 318, 339, 340], "bbox": [0.44479687500000004, 0.3344496487119438, 0.8595312500000001, 0.9583372365339577], "iscrowd": 0, "area": 37297.17750000002, "rle": {"size": [427, 640], "counts": "^Pg3]1n;;D6K6J6I6K6YO[M\\Fk2]9[M]Fl2\\9[M^Fj2\\9h0I4M3M4L3M3M3L5L3N2O1O2M10001N10000O1O2N1O1O1N2O2N1O1N2O1OYKYHl3e7TL[Hm3d7RL]Ho3a7QLaHo3^7QLbHQ4\\7nKeHS4Z7lKhHT4V7mKjHT4U7kKlHV4S7iKoHX4X7^KiHR5S7bJmHV6h64M0O1009G9G8HO010O1O100O1O00N3M2O1N3M2N2N2N3M2O1N3I6I7J6I8I6I7J70O0000010O0001O0001O01O0001O01O000010O0000010O0001OO100000001N100000000O100000000O1000001N100000000O100000000O1000001O0O100000000O10O100O10O10O10000O010O10000O010O10000O010O10000O010O100005J5L4L4L4L3L5L4L3M4L3L5L4L3^HPKc6k5001N2O1O1N3N1O1N2O1O1N2O1O1N3N1O1bNXHWMj7U2mHgMT7X2PIdMQ7\\2QI`MQ7`2RI\\Mo6c2UIYMl6g2WITMk6l2]IkLe6T3i1O1002N6J5K5K5K5_HkKc5Z4UJhKk5]4nIeKR6`4hIaKX6d4aI^K_6f4[I]Kd6g4VI[Kj6i4QIXKP7k4jHWKW7^51O100O1O1O1O010O1O1O1L4I7H8I7H8I7H8I72M5L4L4M39F:^Oa0_Ob0]Ob0_Ob0^Oa0_Ob0^OZVU1"}, "label": "the zebra closest to the camera"}]}
{"id": 55571, "image": "COCO_train2014_000000055571.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra looking over his shoulder\"."}], "task": "refering", "answer_template": "The \"a zebra looking over his shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 148, 149, 150, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 292, 293, 294, 295, 311, 316, 317, 318, 339, 340], "bbox": [0.44479687500000004, 0.3344496487119438, 0.8595312500000001, 0.9583372365339577], "iscrowd": 0, "area": 37297.17750000002, "rle": {"size": [427, 640], "counts": "^Pg3]1n;;D6K6J6I6K6YO[M\\Fk2]9[M]Fl2\\9[M^Fj2\\9h0I4M3M4L3M3M3L5L3N2O1O2M10001N10000O1O2N1O1O1N2O2N1O1N2O1OYKYHl3e7TL[Hm3d7RL]Ho3a7QLaHo3^7QLbHQ4\\7nKeHS4Z7lKhHT4V7mKjHT4U7kKlHV4S7iKoHX4X7^KiHR5S7bJmHV6h64M0O1009G9G8HO010O1O100O1O00N3M2O1N3M2N2N2N3M2O1N3I6I7J6I8I6I7J70O0000010O0001O0001O01O0001O01O000010O0000010O0001OO100000001N100000000O100000000O1000001N100000000O100000000O1000001O0O100000000O10O100O10O10O10000O010O10000O010O10000O010O10000O010O100005J5L4L4L4L3L5L4L3M4L3L5L4L3^HPKc6k5001N2O1O1N3N1O1N2O1O1N2O1O1N3N1O1bNXHWMj7U2mHgMT7X2PIdMQ7\\2QI`MQ7`2RI\\Mo6c2UIYMl6g2WITMk6l2]IkLe6T3i1O1002N6J5K5K5K5_HkKc5Z4UJhKk5]4nIeKR6`4hIaKX6d4aI^K_6f4[I]Kd6g4VI[Kj6i4QIXKP7k4jHWKW7^51O100O1O1O1O010O1O1O1L4I7H8I7H8I7H8I72M5L4L4M39F:^Oa0_Ob0]Ob0_Ob0^Oa0_Ob0^OZVU1"}, "label": "a zebra looking over his shoulder"}]}
{"id": 121110, "image": "COCO_train2014_000000121110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table closest to the trash can\"."}], "task": "refering", "answer_template": "The \"the table closest to the trash can\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 69, 70, 79, 80, 81, 82, 83, 84, 85, 97, 112], "bbox": [0.3189411764705883, 0.17806249999999998, 0.7077176470588235, 0.337265625], "iscrowd": 0, "area": 6747.885649999998, "rle": {"size": [640, 425], "counts": "`Te25ic06K1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O10000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000o1QN00000O10000000000000O1000000iNW1ZOf000000000000000000000000000000000000000000000000000O100000000000000000O100000000000000000000000000000000000000000000002N3M2N2N3M2N2N3M2N2N3M2N2N`[]2"}, "label": "the table closest to the trash can"}]}
{"id": 121110, "image": "COCO_train2014_000000121110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a checkerboard on it in the foreground\"."}], "task": "refering", "answer_template": "The \"a table with a checkerboard on it in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 232, 233, 247, 248, 262, 263, 277, 278], "bbox": [0.018541176470588235, 0.359890625, 0.9903294117647059, 0.8229531250000001], "iscrowd": 0, "area": 61583.708600000005, "rle": {"size": [640, 425], "counts": "`[54jc05K6J6J5K6J5K2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N20000000001O0000000000000000000001O00000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000=Cj0VOj0VOj0VOj0VO8H00001O0000000000001O00000000001O0000000000001O00000000001OjKV40000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2I7F:GeT2"}, "label": "a table with a checkerboard on it in the foreground"}]}
{"id": 424222, "image": "COCO_train2014_000000424222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat on a lift with a duck crossing sign on its left\"."}], "task": "refering", "answer_template": "The \"a boat on a lift with a duck crossing sign on its left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296], "bbox": [0.667109375, 0.6244835680751173, 0.987265625, 0.8593896713615022], "iscrowd": 0, "area": 10349.586200000009, "rle": {"size": [426, 640], "counts": "bna5d0f<a0^O10000O10O010000O10000PDlNi;[100001O00000O1000000000000000000O100000000000eNeNfF\\1Y9eNaFa1^9aN^Fb1b9_N]Fa1c9_N]F`1d9aN[F_1d9bN\\F^1d9cN[F]1e9dNZF\\1f9dNZF\\1f9eNYF[1f9gNYFY1g9gNYFY1g9hNXFX1h9iNWFW1i9iNWFV1i9lNVFT1j9mNUFS1k9mNUFS1k9nNTFR1k9POTFP1l9POTFP1l9QOSFo0m9QOSFo0m9QOSFo0l9ROTFm0m9SOSFm0m9SOSFm0m9SOSFm0m9SOSFm0l9SOUFm0k9QOWFo0i9POXFP1h9nNXFT1g9lNXFV1h9jNVFX1j9hNTFY1m9gNQF[1o9eNoE]1P:dNnE^1R:bNlE`1T:`NjEb1V:^NhEd1X:\\NfEf1Y:[NeEg1[:d00O1000001O0000000000000O101O00000000000000001N100dMcEk1]:UNmEa1T:^NVFX1j9gNXFX1i9gNXFX1h9hNXFX1i9gNWFY1i9gNWFY1j9eNWF[1i9eNWF[1j9dNVF\\1j9dNVF\\1k9bNVF^1j9bNVF^1k9aNUF`1j9`NVF`1k9_NUFa1j:0000001O0000000000001O000000000000000000000000J_DcNa;\\17O100O1O1O100O1O1O100O1O1O100N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O100O100O10000O100O100O100O10000O2O0O100O100O1000U^3"}, "label": "a boat on a lift with a duck crossing sign on its left"}]}
{"id": 424222, "image": "COCO_train2014_000000424222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat to the right of the pole\"."}], "task": "refering", "answer_template": "The \"a boat to the right of the pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296], "bbox": [0.667109375, 0.6244835680751173, 0.987265625, 0.8593896713615022], "iscrowd": 0, "area": 10349.586200000009, "rle": {"size": [426, 640], "counts": "bna5d0f<a0^O10000O10O010000O10000PDlNi;[100001O00000O1000000000000000000O100000000000eNeNfF\\1Y9eNaFa1^9aN^Fb1b9_N]Fa1c9_N]F`1d9aN[F_1d9bN\\F^1d9cN[F]1e9dNZF\\1f9dNZF\\1f9eNYF[1f9gNYFY1g9gNYFY1g9hNXFX1h9iNWFW1i9iNWFV1i9lNVFT1j9mNUFS1k9mNUFS1k9nNTFR1k9POTFP1l9POTFP1l9QOSFo0m9QOSFo0m9QOSFo0l9ROTFm0m9SOSFm0m9SOSFm0m9SOSFm0m9SOSFm0l9SOUFm0k9QOWFo0i9POXFP1h9nNXFT1g9lNXFV1h9jNVFX1j9hNTFY1m9gNQF[1o9eNoE]1P:dNnE^1R:bNlE`1T:`NjEb1V:^NhEd1X:\\NfEf1Y:[NeEg1[:d00O1000001O0000000000000O101O00000000000000001N100dMcEk1]:UNmEa1T:^NVFX1j9gNXFX1i9gNXFX1h9hNXFX1i9gNWFY1i9gNWFY1j9eNWF[1i9eNWF[1j9dNVF\\1j9dNVF\\1k9bNVF^1j9bNVF^1k9aNUF`1j9`NVF`1k9_NUFa1j:0000001O0000000000001O000000000000000000000000J_DcNa;\\17O100O1O1O100O1O1O100O1O1O100N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O100O100O10000O100O100O100O10000O2O0O100O100O1000U^3"}, "label": "a boat to the right of the pole"}]}
{"id": 424222, "image": "COCO_train2014_000000424222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white boat to the left of two other boats\"."}], "task": "refering", "answer_template": "The \"a white boat to the left of two other boats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 280], "bbox": [0.0, 0.5679107981220658, 0.21365625000000002, 0.8592488262910799], "iscrowd": 0, "area": 6174.471050000004, "rle": {"size": [426, 640], "counts": "_:1Y=2N2N2N2N2M3N2N2N2N2M3N2N1O2N2N1N100000000O100000000O01000000000O100iNmNWFS1j9nNTFR1l9PORFP1n9QOPFP1P:ROnEn0R:TOhEP1X:QO`EV1`:lNVE\\1j:b00000001O0000000O1000000000001O000000000000000000001O000000000O10000000001O00000000000000000000001OJ6I7J6J6I6K6I7JhMfGLn:N3M2N2N001O001O001O001O1O001O001O001O0O2O001O1O001O001O002N1O2N2N1O2N1Onl`6"}, "label": "a white boat to the left of two other boats"}]}
{"id": 424222, "image": "COCO_train2014_000000424222.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat with white pole sticking out from the top\"."}], "task": "refering", "answer_template": "The \"boat with white pole sticking out from the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 280], "bbox": [0.0, 0.5679107981220658, 0.21365625000000002, 0.8592488262910799], "iscrowd": 0, "area": 6174.471050000004, "rle": {"size": [426, 640], "counts": "_:1Y=2N2N2N2N2M3N2N2N2N2M3N2N1O2N2N1N100000000O100000000O01000000000O100iNmNWFS1j9nNTFR1l9PORFP1n9QOPFP1P:ROnEn0R:TOhEP1X:QO`EV1`:lNVE\\1j:b00000001O0000000O1000000000001O000000000000000000001O000000000O10000000001O00000000000000000000001OJ6I7J6J6I6K6I7JhMfGLn:N3M2N2N001O001O001O001O1O001O001O001O0O2O001O1O001O001O002N1O2N2N1O2N1Onl`6"}, "label": "boat with white pole sticking out from the top"}]}
{"id": 47391, "image": "COCO_train2014_000000047391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in striped pajamas pointing a remote at the camera\"."}], "task": "refering", "answer_template": "The \"a boy in striped pajamas pointing a remote at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 270, 271, 272, 294, 295, 296, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410], "bbox": [0.393703125, 0.18427734375, 0.9671875, 0.9865234375], "iscrowd": 0, "area": 73667.57390000002, "rle": {"size": [512, 640], "counts": "h_n31l?3M3M3M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M3M30000000000O100ZM_MmGa2P8bMPH^2n7dMRH\\2k7gMUHY2h7iMYHW2e7kMZHV2c7mM]HS2`7PN`HP2]7SNcHm1[7UNeHk1X7XNhHh1U7ZNlHf1R7\\NmHe1R2^M5P1iMb1Q2]NUO4jN_1o1eNnNOSO\\1m1mNhNI[OZ1l1TObNCBY1j1[O^N\\OIY1g1CZNTONZ1g1IUNmN4Z1e11PNfN;Y1c19lM^Na0Y1b1=iM[Ne0X1`1c0gMUNi0W1_1j0dMnMn0X1]1n0aMkMR1W1[1T1_MeMU1X1Z1X1^M`MX1X1Y1]1ZM\\M]1W1W1c1XMVMa1W1W1d1WMUMb1W1V1g1UMRMf1W1U1i1SMPMh1W1U1j1QMPMi1W1U1l1PMmLk1W1U1m1oLlLl1W1T1P2mLjLo1V1T1R2kLhLQ2V1T1S2jLfLS2W1R1V2hLdLV2V1R1W2gLcLV2W1S1X2eLaLX2W1R1[2cL_L[2V1R1\\2bL^L\\2V1Q1_2`L\\L_2U1Q1`2_LZLa2V1P1b2]LXLc2V1o0W4POjKP1V4POjKP1V4POjKo0W4QOiKo0W4QOiKn0W4SOiKm0W4SOiKm0W4SOiKl0X4SOiKm0W4SOiKm0W4SOiKl0X4TOhKl0W4UOiKj0X4VOhKj0X4VOhKj0X4UOiKj0X4VOhKj0X4VOhKj0X4VOhKi0X4XOhKh0X4XOhKh0X4XOhKh0X4WOiKi0W4WOiKi0W4WOiKh0X4XOhKh0W4YOiKg0W4YOiKg0W4YOiKg0W4XOjKh0V4XOjKh0V4XOjKh0V4XOjKg0V4ZOjKf0V4ZOjKf0V4ZOjKf0V4ZOjKf0V4YOkKg0U4YOkKg0T4ZOlKf0T4ZOlKe0U4[OkKe0U4[OkKe0T4\\OlKd0T4[OmKe0S4[OmKe0S4[OmKe0R4\\OnKd0R4\\OnKc0S4]OmKc0R4^OnKb0R4]OoKc0Q4]OoKc0Q4]OoKc0P4^OPLb0P4^OPLb0P4^OPLb0P4^OPLa0P4_OQLa0o3_OQLa0o3_OPLb0o3_OQLa0o3_OQLa0o3_OQLa0o3_OQLa0n3@RL?o3@RLa0m3_OSLa0m3_OSLb0k3_OULb0j3^OVLb0j3^OVLc0h3^OXLb0h3]OYLd0f3\\OZLd0f3\\OZLe0d3\\O\\Ld0d3\\O\\Le0c3[O]Lf0b3ZO^Lf0a3ZO`Lg0_3YOaLg0_3YOaLh0^3XObLh0]3YOcLh0\\3XOdLh0\\3XOdLi0Z3WOgLi0Y3WOgLj0X3VOhLk0W3UOiLk0V3VOjLk0U3UOkLk0U3UOkLl0T3TOlLl0S3TOnLm0Q3SOoLm0Q3SOoLo0n2RORMP1l2POTMQ1k2oNUMS1i2mNWMU1f2kN[MV1d2jN\\MX1b2hN^MZ1`2fN`M\\1]2eNcM\\1\\2dNdM^1Z2bNfM`1W2`NjMa1U2_NkMc1S2]NmMe1Q2[NoMi1l1XNTNn1f1QN[NT2`1lM`NY2[1gMeN^2U1cMkNb2P1^MPOg2k0XMVOk2g0UMYOn2e0QM[OR3b0nL^OV3>iLCZ3;eLE^38bLHb35]LKf32YLOj3NVL2m3LRL4R4HnK8W4DgK=a4[O_Ke0i4SOWKm0P5mNoJS1Y5eNgJ[1[5dNcJ]1^5bNbJ^1_5aNaJ_1`5aN_J_1a5aN^J`1c5_N]Ja1d5_N[Ja1f5^NZJb1g5^NWJb1k5]NUJc1l5\\NTJd1m5\\NRJd1o5[NPJf1Q6ZNnIf1S6YNmIg1T6XNlIh1U6XNiIi1X6VNhIj1Y6VNfIj1[6UNeIk1[6UNeIk1\\6UNbIl1_6SNaIm1`6SN_Im1b6RN]Io1d6PNZIR2g6nMWIS2j6lMUIU2l6kMQIW2P7hMoHY2R7fMmH[2T7eMiH]2X7bMgH_2Z7aMdH`2]7aM`H`2`7bM\\H`2e7aMXH`2i7aMTH`2m7aMoGa2R8`MkGa2V8`MgGa2Z8`MbGb2_8_M^Gb2c8_MZGb2g8_MUGc2l8^MQGc2P9]MnFd2S9]MiFe2X9\\MeFe2\\9o11O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001XOTFZKm9d4UF[Kl9d4VFZKk9e4WFYKj9f4XFXKj9f4WFYKj9f4XFXKj9f4WFYKj9f4WFYKk9e4VFZKk9e4VFZKl9d4UF[Kl9d4UF[Kl9d4UF[Km9b4UF]Kl9b4UF]Km9a4TF^Km9a4k0N1O2N1O1O2N1O2N1O2M2O2N1O2N1O1O2N1O1O2N1O1O2M2O1O2N1O2N1O1O2N1O1O2N1O1N3N1O1O2N1O1O2M2N3M2N2N3M2N2N3M5K5K5K5K5K6J5L4K5K5K5K5KnQ:"}, "label": "a boy in striped pajamas pointing a remote at the camera"}]}
{"id": 47391, "image": "COCO_train2014_000000047391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling child holding a device\"."}], "task": "refering", "answer_template": "The \"a smiling child holding a device\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 270, 271, 272, 294, 295, 296, 315, 316, 317, 318, 319, 320, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410], "bbox": [0.393703125, 0.18427734375, 0.9671875, 0.9865234375], "iscrowd": 0, "area": 73667.57390000002, "rle": {"size": [512, 640], "counts": "h_n31l?3M3M3M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M3M30000000000O100ZM_MmGa2P8bMPH^2n7dMRH\\2k7gMUHY2h7iMYHW2e7kMZHV2c7mM]HS2`7PN`HP2]7SNcHm1[7UNeHk1X7XNhHh1U7ZNlHf1R7\\NmHe1R2^M5P1iMb1Q2]NUO4jN_1o1eNnNOSO\\1m1mNhNI[OZ1l1TObNCBY1j1[O^N\\OIY1g1CZNTONZ1g1IUNmN4Z1e11PNfN;Y1c19lM^Na0Y1b1=iM[Ne0X1`1c0gMUNi0W1_1j0dMnMn0X1]1n0aMkMR1W1[1T1_MeMU1X1Z1X1^M`MX1X1Y1]1ZM\\M]1W1W1c1XMVMa1W1W1d1WMUMb1W1V1g1UMRMf1W1U1i1SMPMh1W1U1j1QMPMi1W1U1l1PMmLk1W1U1m1oLlLl1W1T1P2mLjLo1V1T1R2kLhLQ2V1T1S2jLfLS2W1R1V2hLdLV2V1R1W2gLcLV2W1S1X2eLaLX2W1R1[2cL_L[2V1R1\\2bL^L\\2V1Q1_2`L\\L_2U1Q1`2_LZLa2V1P1b2]LXLc2V1o0W4POjKP1V4POjKP1V4POjKo0W4QOiKo0W4QOiKn0W4SOiKm0W4SOiKm0W4SOiKl0X4SOiKm0W4SOiKm0W4SOiKl0X4TOhKl0W4UOiKj0X4VOhKj0X4VOhKj0X4UOiKj0X4VOhKj0X4VOhKj0X4VOhKi0X4XOhKh0X4XOhKh0X4XOhKh0X4WOiKi0W4WOiKi0W4WOiKh0X4XOhKh0W4YOiKg0W4YOiKg0W4YOiKg0W4XOjKh0V4XOjKh0V4XOjKh0V4XOjKg0V4ZOjKf0V4ZOjKf0V4ZOjKf0V4ZOjKf0V4YOkKg0U4YOkKg0T4ZOlKf0T4ZOlKe0U4[OkKe0U4[OkKe0T4\\OlKd0T4[OmKe0S4[OmKe0S4[OmKe0R4\\OnKd0R4\\OnKc0S4]OmKc0R4^OnKb0R4]OoKc0Q4]OoKc0Q4]OoKc0P4^OPLb0P4^OPLb0P4^OPLb0P4^OPLa0P4_OQLa0o3_OQLa0o3_OPLb0o3_OQLa0o3_OQLa0o3_OQLa0o3_OQLa0n3@RL?o3@RLa0m3_OSLa0m3_OSLb0k3_OULb0j3^OVLb0j3^OVLc0h3^OXLb0h3]OYLd0f3\\OZLd0f3\\OZLe0d3\\O\\Ld0d3\\O\\Le0c3[O]Lf0b3ZO^Lf0a3ZO`Lg0_3YOaLg0_3YOaLh0^3XObLh0]3YOcLh0\\3XOdLh0\\3XOdLi0Z3WOgLi0Y3WOgLj0X3VOhLk0W3UOiLk0V3VOjLk0U3UOkLk0U3UOkLl0T3TOlLl0S3TOnLm0Q3SOoLm0Q3SOoLo0n2RORMP1l2POTMQ1k2oNUMS1i2mNWMU1f2kN[MV1d2jN\\MX1b2hN^MZ1`2fN`M\\1]2eNcM\\1\\2dNdM^1Z2bNfM`1W2`NjMa1U2_NkMc1S2]NmMe1Q2[NoMi1l1XNTNn1f1QN[NT2`1lM`NY2[1gMeN^2U1cMkNb2P1^MPOg2k0XMVOk2g0UMYOn2e0QM[OR3b0nL^OV3>iLCZ3;eLE^38bLHb35]LKf32YLOj3NVL2m3LRL4R4HnK8W4DgK=a4[O_Ke0i4SOWKm0P5mNoJS1Y5eNgJ[1[5dNcJ]1^5bNbJ^1_5aNaJ_1`5aN_J_1a5aN^J`1c5_N]Ja1d5_N[Ja1f5^NZJb1g5^NWJb1k5]NUJc1l5\\NTJd1m5\\NRJd1o5[NPJf1Q6ZNnIf1S6YNmIg1T6XNlIh1U6XNiIi1X6VNhIj1Y6VNfIj1[6UNeIk1[6UNeIk1\\6UNbIl1_6SNaIm1`6SN_Im1b6RN]Io1d6PNZIR2g6nMWIS2j6lMUIU2l6kMQIW2P7hMoHY2R7fMmH[2T7eMiH]2X7bMgH_2Z7aMdH`2]7aM`H`2`7bM\\H`2e7aMXH`2i7aMTH`2m7aMoGa2R8`MkGa2V8`MgGa2Z8`MbGb2_8_M^Gb2c8_MZGb2g8_MUGc2l8^MQGc2P9]MnFd2S9]MiFe2X9\\MeFe2\\9o11O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001XOTFZKm9d4UF[Kl9d4VFZKk9e4WFYKj9f4XFXKj9f4WFYKj9f4XFXKj9f4WFYKj9f4WFYKk9e4VFZKk9e4VFZKl9d4UF[Kl9d4UF[Kl9d4UF[Km9b4UF]Kl9b4UF]Km9a4TF^Km9a4k0N1O2N1O1O2N1O2N1O2M2O2N1O2N1O1O2N1O1O2N1O1O2M2O1O2N1O2N1O1O2N1O1O2N1O1N3N1O1O2N1O1O2M2N3M2N2N3M2N2N3M5K5K5K5K5K6J5L4K5K5K5K5KnQ:"}, "label": "a smiling child holding a device"}]}
{"id": 47391, "image": "COCO_train2014_000000047391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an arm resting on a couch back\"."}], "task": "refering", "answer_template": "The \"an arm resting on a couch back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 107, 108, 109, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 131, 132, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164], "bbox": [0.0, 0.10337890625, 0.7928124999999999, 0.438203125], "iscrowd": 0, "area": 42340.01430000001, "rle": {"size": [512, 640], "counts": "]3a0_?R1nNR1nNn0RO0O100O01000O10O10O100O01000O10O10O100O01000O10O10O100O01000O10O10O10O010000O01000O10O010000O01000O10O010000O01000O10O1000000O0100000O100000O10O1000000O10O100000O10000000O01000000O1000O1000O1000000O010000000O1000O1000O1000000O010000000O1000O1000O1000000O010000000O100000000O2O00000O100000000O100000000O2O00000O100000000O100000001N1000000O100000000O1000001O0O1000000O100000000O10001O0O10O100000O1000000O1000O1000O1000000O10000000O01000000O100000000O10O1000O100000000O100000O0100000000O1000000O10O100000O1000000O010000000O10O1000O100O10O0100O100O10O0100O100O10O0100O100O10O0100O100O010O100O100N1O2M3M3N2M3M2N3N200O1O10O0100O1O100O00100O100O1O010O1O100O1O010O100O1O010O1O100O1O100O100O1O100O1000000000000000000000000000000000000O1000000000000000000000000000000000000000O100O10000O010O10000O100O100O10000O100O2O001O0010O01O001O001O00100O001O001O001O010O001O001O1O2N2O0O2N1O2N1O2NBWBgNg=Z1ZBeNe=[1]BeNa=\\1`BcN_=^1`BdN^=\\1cBdN\\=]1dBbN\\=^1eBbNZ=_1eBbN[=c0aBG5FY=d0bBF5FX=d0cBG5DX=f0cBE5FW=e0eBE4FV=f0fBC5GT=g0fBC5GT=f0hBB5GS=h0hBA5GR=h0jBA4GQ=i0jB@5HP=h0lB@4Ho<i0mB^O5Hn<j0nB^O3In<j0nB]O5Im<i0oB^O4Il<j0PC]O4Ik<k0QC[O4Jk<k0RC[O3Jj<l0RCZO5Jh<l0TCZO4Jg<NRC?3H4Lf<JWCa00I3Kf<IZCc0LI5Kd<F_Ce0IJ3Ke<DaCh0GH4If<EbCi0EH4G\\=a0_BI5D^=b0^BI4C`=e0\\BH4Ab=f0[BH4_Od=i0WBHW>7jAIV>7jAHV>8kAHU>8jAHW>7jAHW>8iAHV>8jAHW>8iAHW>7jAHW>8iAGW>9iAHW>8iAGX>8iAGX>9gAF\\>8eAF^>9bAF_>9aAFb>9^AEe>9\\AFe>:[ADh>9YAFi>8YAGi>6c0M2O`\\R2"}, "label": "an arm resting on a couch back"}]}
{"id": 47391, "image": "COCO_train2014_000000047391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm on the back of the couch\"."}], "task": "refering", "answer_template": "The \"the arm on the back of the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 107, 108, 109, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 131, 132, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164], "bbox": [0.0, 0.10337890625, 0.7928124999999999, 0.438203125], "iscrowd": 0, "area": 42340.01430000001, "rle": {"size": [512, 640], "counts": "]3a0_?R1nNR1nNn0RO0O100O01000O10O10O100O01000O10O10O100O01000O10O10O100O01000O10O10O10O010000O01000O10O010000O01000O10O010000O01000O10O1000000O0100000O100000O10O1000000O10O100000O10000000O01000000O1000O1000O1000000O010000000O1000O1000O1000000O010000000O1000O1000O1000000O010000000O100000000O2O00000O100000000O100000000O2O00000O100000000O100000001N1000000O100000000O1000001O0O1000000O100000000O10001O0O10O100000O1000000O1000O1000O1000000O10000000O01000000O100000000O10O1000O100000000O100000O0100000000O1000000O10O100000O1000000O010000000O10O1000O100O10O0100O100O10O0100O100O10O0100O100O10O0100O100O010O100O100N1O2M3M3N2M3M2N3N200O1O10O0100O1O100O00100O100O1O010O1O100O1O010O100O1O010O1O100O1O100O100O1O100O1000000000000000000000000000000000000O1000000000000000000000000000000000000000O100O10000O010O10000O100O100O10000O100O2O001O0010O01O001O001O00100O001O001O001O010O001O001O1O2N2O0O2N1O2N1O2NBWBgNg=Z1ZBeNe=[1]BeNa=\\1`BcN_=^1`BdN^=\\1cBdN\\=]1dBbN\\=^1eBbNZ=_1eBbN[=c0aBG5FY=d0bBF5FX=d0cBG5DX=f0cBE5FW=e0eBE4FV=f0fBC5GT=g0fBC5GT=f0hBB5GS=h0hBA5GR=h0jBA4GQ=i0jB@5HP=h0lB@4Ho<i0mB^O5Hn<j0nB^O3In<j0nB]O5Im<i0oB^O4Il<j0PC]O4Ik<k0QC[O4Jk<k0RC[O3Jj<l0RCZO5Jh<l0TCZO4Jg<NRC?3H4Lf<JWCa00I3Kf<IZCc0LI5Kd<F_Ce0IJ3Ke<DaCh0GH4If<EbCi0EH4G\\=a0_BI5D^=b0^BI4C`=e0\\BH4Ab=f0[BH4_Od=i0WBHW>7jAIV>7jAHV>8kAHU>8jAHW>7jAHW>8iAHV>8jAHW>8iAHW>7jAHW>8iAGW>9iAHW>8iAGX>8iAGX>9gAF\\>8eAF^>9bAF_>9aAFb>9^AEe>9\\AFe>:[ADh>9YAFi>8YAGi>6c0M2O`\\R2"}, "label": "the arm on the back of the couch"}]}
{"id": 432417, "image": "COCO_train2014_000000432417.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the left by the camera\"."}], "task": "refering", "answer_template": "The \"the chair on the left by the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330], "bbox": [0.118359375, 0.6504449648711944, 0.35678125, 1.0], "iscrowd": 0, "area": 15170.172199999997, "rle": {"size": [427, 640], "counts": "\\PP12Y=2N3M2N3L3N3M2N3M2N3M2M4M2N3M2N3M2N3L3N3M3M2N3M2N3L3N3M2N3M2N3M2N3L3N3M2N3M2N2NO00100000000000O1000000000000O10000O10000O10000O1000000O10000O10000O10001N10000O10000O1000000O10000O10000O10000O10000O1000000O10000O10000O100O100O100O1O100O100O1O100O100O1O2O0O100O1O100O100O1O100O100O1O100O100O1O10WM_HH`7MhIWOW6>aJVO_5=m3EZa[5"}, "label": "the chair on the left by the camera"}]}
{"id": 350500, "image": "COCO_train2014_000000350500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white paper bag on top of the red and black suitcases\"."}], "task": "refering", "answer_template": "The \"the white paper bag on top of the red and black suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 171, 172, 173, 174, 175, 195, 196], "bbox": [0.367671875, 0.0032426778242677827, 0.67509375, 0.48604602510460254], "iscrowd": 0, "area": 31141.884349999997, "rle": {"size": [478, 640], "counts": "^Y^32k>3M2K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5L4K6J5K5K5K5K5K5K5K5K5K5K5K5K5K5K5O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O2N1O000000O100O10000O10000O100O10000O10000O100O1000O010000O100O10000O10000O100O10000O10000O100O10000O10000O2O0O2O1O0O2O001VOcGPK^8l4hGRKY8j4lGSKV8i4oGUKQ8g4UHWKl7e4YHXKh7e4^HXKc7c4cH[K^7a4a1M4K4M4L4K4M4L3L5L4L3L5L3M4K4M4L4K4M4K4M4L4K4M4L3L5L4L3L5L3M4K4M4L4K4M4L3LSRQ3"}, "label": "the white paper bag on top of the red and black suitcases"}]}
{"id": 350500, "image": "COCO_train2014_000000350500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bag on the suitcases\"."}], "task": "refering", "answer_template": "The \"white bag on the suitcases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 32, 33, 34, 35, 36, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 171, 172, 173, 174, 175, 195, 196], "bbox": [0.367671875, 0.0032426778242677827, 0.67509375, 0.48604602510460254], "iscrowd": 0, "area": 31141.884349999997, "rle": {"size": [478, 640], "counts": "^Y^32k>3M2K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5L4K6J5K5K5K5K5K5K5K5K5K5K5K5K5K5K5O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O2N1O000000O100O10000O10000O100O10000O10000O100O1000O010000O100O10000O10000O100O10000O10000O100O10000O10000O2O0O2O1O0O2O001VOcGPK^8l4hGRKY8j4lGSKV8i4oGUKQ8g4UHWKl7e4YHXKh7e4^HXKc7c4cH[K^7a4a1M4K4M4L4K4M4L3L5L4L3L5L3M4K4M4L4K4M4K4M4L4K4M4L3L5L4L3L5L3M4K4M4L4K4M4L3LSRQ3"}, "label": "white bag on the suitcases"}]}
{"id": 350500, "image": "COCO_train2014_000000350500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suitcase with red strap on the floor\"."}], "task": "refering", "answer_template": "The \"black suitcase with red strap on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 172, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332, 333, 334, 335, 353, 354, 355, 356, 357, 358, 377, 378, 379, 380], "bbox": [0.218671875, 0.2421129707112971, 0.603890625, 0.9898744769874478], "iscrowd": 0, "area": 57218.7613, "rle": {"size": [478, 640], "counts": "Z`Q23i>6J6J5K6J6L4L3L5L4K5L4L3L5L4L4K4M4K5L4L4K4M4K5L4L3L5L4_EZLb9j3YF\\La9i3ZF]La9f3[F`L_9e3\\FaL_9d4K5J101N101O0O2RHUJg6l5WIWJg6m5TIUJj6Q6PIRJn6T6kHnIS7X6gHkIW7[6cHfI\\7P7M2O1O1N2O0O2O1O0O2O0O1000O0101O1N2O2M2O1O1N2O2M2O1O1N2O1O2M2O1N2O1O2M2O1O1N2O2M2O1O1N2O2M2O1O1N2O1O2M2O1N2O1?BO000000001O00000000001O0001O0001O000000001O00000000001O000000001O00000000O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2OO010O0100O010O010O001O001O10O01O001O001O010O1O001O001O010O001O1O001O0010O01O001[I\\Ho5S9VOj0VOi0XOh0WOh0XOg0YOg0YOeUf3"}, "label": "black suitcase with red strap on the floor"}]}
{"id": 350500, "image": "COCO_train2014_000000350500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suitcase with red strap\"."}], "task": "refering", "answer_template": "The \"black suitcase with red strap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 172, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332, 333, 334, 335, 353, 354, 355, 356, 357, 358, 377, 378, 379, 380], "bbox": [0.218671875, 0.2421129707112971, 0.603890625, 0.9898744769874478], "iscrowd": 0, "area": 57218.7613, "rle": {"size": [478, 640], "counts": "Z`Q23i>6J6J5K6J6L4L3L5L4K5L4L3L5L4L4K4M4K5L4L4K4M4K5L4L3L5L4_EZLb9j3YF\\La9i3ZF]La9f3[F`L_9e3\\FaL_9d4K5J101N101O0O2RHUJg6l5WIWJg6m5TIUJj6Q6PIRJn6T6kHnIS7X6gHkIW7[6cHfI\\7P7M2O1O1N2O0O2O1O0O2O0O1000O0101O1N2O2M2O1O1N2O2M2O1O1N2O1O2M2O1N2O1O2M2O1O1N2O2M2O1O1N2O2M2O1O1N2O1O2M2O1N2O1?BO000000001O00000000001O0001O0001O000000001O00000000001O000000001O00000000O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2OO010O0100O010O010O001O001O10O01O001O001O010O1O001O001O010O001O1O001O0010O01O001[I\\Ho5S9VOj0VOi0XOh0WOh0XOg0YOg0YOeUf3"}, "label": "black suitcase with red strap"}]}
{"id": 350500, "image": "COCO_train2014_000000350500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue purse\"."}], "task": "refering", "answer_template": "The \"a blue purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 39, 61, 62, 63, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154], "bbox": [0.6497499999999999, 0.0, 0.82390625, 0.3600836820083682], "iscrowd": 0, "area": 10975.0355, "rle": {"size": [478, 640], "counts": "U[R67\\>;F:F:E;F:E;F:E;F:F:O100000000000000O1O100O1O1O1O1ZN\\LRHd3n7^LoGc3Q8_LlGa3U8`LiGa3W8aLfG`3Z8bLdG^3\\8iL\\GX3d8RMQGo2o8\\MeFe2[9_10000000000000000000ULcFU2]9fMjFX2W9aMQG]2Q9\\MVGb2l8WMYGi2i8oL]GP3f8iL_GW3T:02N001N10001O001N10001O001N10001O001N2O1O1O1O1N2O1O1O2N1N2O1O1O1N2O1O1O1O1N2O1O1O1O2M2O1O1O1O2M2N2N2N2N3L3N2N2N2N7I`0_Olcd1"}, "label": "a blue purse"}]}
{"id": 350500, "image": "COCO_train2014_000000350500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark colored bag is sitting on top of a suitcase that is next to a wall\"."}], "task": "refering", "answer_template": "The \"a dark colored bag is sitting on top of a suitcase that is next to a wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 39, 61, 62, 63, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154], "bbox": [0.6497499999999999, 0.0, 0.82390625, 0.3600836820083682], "iscrowd": 0, "area": 10975.0355, "rle": {"size": [478, 640], "counts": "U[R67\\>;F:F:E;F:E;F:E;F:F:O100000000000000O1O100O1O1O1O1ZN\\LRHd3n7^LoGc3Q8_LlGa3U8`LiGa3W8aLfG`3Z8bLdG^3\\8iL\\GX3d8RMQGo2o8\\MeFe2[9_10000000000000000000ULcFU2]9fMjFX2W9aMQG]2Q9\\MVGb2l8WMYGi2i8oL]GP3f8iL_GW3T:02N001N10001O001N10001O001N10001O001N2O1O1O1O1N2O1O1O2N1N2O1O1O1N2O1O1O1O1N2O1O1O1O2M2O1O1O1O2M2N2N2N2N3L3N2N2N2N7I`0_Olcd1"}, "label": "a dark colored bag is sitting on top of a suitcase that is next to a wall"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bread vegetable sandwich with tooth pick next to white bowl\"."}], "task": "refering", "answer_template": "The \"brown bread vegetable sandwich with tooth pick next to white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 238, 239, 240], "bbox": [0.243609375, 0.3829375, 0.500625, 0.597125], "iscrowd": 0, "area": 11542.970800000001, "rle": {"size": [480, 640], "counts": "o[Y22j>4L5K4L5M20001N101O00001O001O0O101N101N101N100O2O0O2O0O101bNZO_Dg0a;ZO^Dg0a;YO^Dh0b;XO^Dh0b;YOQDT1n;lNoCW1R<hNlC[1S<fNiC]1W<f001O001O010O00001O000010O000001O00001O01O0001O00000010O000001O00000010O0001O0000001O00000010O00000000000000002N6JO100O1O1O1O10000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00000010O0001O1O1O1O1O001O1VOj0O0O2O0O2O001XOiBDX=<nB^OS=b0e001O010O0O2K4L5L3LYie4"}, "label": "brown bread vegetable sandwich with tooth pick next to white bowl"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich in the plate of the man in black\"."}], "task": "refering", "answer_template": "The \"the sandwich in the plate of the man in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 238, 239, 240], "bbox": [0.243609375, 0.3829375, 0.500625, 0.597125], "iscrowd": 0, "area": 11542.970800000001, "rle": {"size": [480, 640], "counts": "o[Y22j>4L5K4L5M20001N101O00001O001O0O101N101N101N100O2O0O2O0O101bNZO_Dg0a;ZO^Dg0a;YO^Dh0b;XO^Dh0b;YOQDT1n;lNoCW1R<hNlC[1S<fNiC]1W<f001O001O010O00001O000010O000001O00001O01O0001O00000010O000001O00000010O0001O0000001O00000010O00000000000000002N6JO100O1O1O1O10000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00000010O0001O1O1O1O1O001O1VOj0O0O2O0O2O001XOiBDX=<nB^OS=b0e001O010O0O2K4L5L3LYie4"}, "label": "the sandwich in the plate of the man in black"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with a blue toothpick sitting next to a bowl of soup\"."}], "task": "refering", "answer_template": "The \"a sandwich with a blue toothpick sitting next to a bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.474421875, 0.5350625, 1.0, 1.0], "iscrowd": 0, "area": 47330.04665000001, "rle": {"size": [480, 640], "counts": "Sh^4i0W>9G00001O00001O0000001O0000001O0000001O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O000lGTMT4m2hKVMX4j2fKXMZ4h2cK[M\\4g2aK[M_4h2[K[Me4i2TKZMl4j2nJXMR5l2gJWMY5m2aJUM_5o2ZJTMf5P3TJRMl5R3mIQMS6S3gIoLY6U3`InL`6V3ZIlLf6X3SIkLm6Y3mHiLS7[3fHhLZ7\\3`HfL`7^3YHeLg7_3SHcLm7X50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000"}, "label": "a sandwich with a blue toothpick sitting next to a bowl of soup"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman eating soup\"."}], "task": "refering", "answer_template": "The \"woman eating soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 161, 162, 163, 164, 165, 166, 167, 171, 172, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 201, 202, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 253], "bbox": [0.0, 0.0, 0.841015625, 0.6561875], "iscrowd": 0, "area": 104341.91660000001, "rle": {"size": [480, 640], "counts": "i92i>5K5L3_OB]Bc0R=MjB8e<8WCMY<b0dCBk;R2I7J5L5J6J6J6J6K4K6J6J6J5L5J6J6J5M4M3M3M2N3M3M3M2N3M3M3M2N3M3L4M2M4J6I6K6J6J6N1O2N2N1O20O0100O0100O010O10O10O10O0100O0100O0100O010O10O10O10O0100O010000O10O01000000O010000000000O010000000O1000O100000O100000O10O1000000000O0100000000O10O010000O100O10O010000O100O010O10000O10O0100O100L4K4N3M3N2N2N2O1O001O1N2O1O1O001O100O100O1O100O100O100O1O1O1O1O1N2O1O1O1O1N2O1XOQH]JQ8a5SHZJP8d5UHWJm7g5XHTJj7j5g0N200000000000000000000000000000000000000000000000000001O00000000001O000000001O000000001O000000001O00000000001O000000001O000001O01O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O000000O100000000O100000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000001O0000001O0000001O00010O0000001O00001O0000001O00001O0000001OM3L4K5L4K5L41O1O1O1O1O1O2N1O1O1O1O1O1O1OjKSGZ2l8cMXG]2g8`M^G_2a8^McGc2\\8YMhGg2W8WMmGh2R8WMPHi2o7UMUHj2j7TMYHl2f7SM]Hm2a7QMbHo2]7oLfHQ3Y7nLiHR3V7mLlHS3S7kLPIU3o6jLSIU3m6jLTIW3k6hLWIW3i6hLYIX3f6gL\\IX3d6gL^IY3a6fLaIY3_6fLcIZ3\\6eLfIZ3Z6eLhI[3W6dLkI[3U6dLmI\\3R6bLQJ]3o5aLTJ_3k5_LXJ`3h5^L[Jb3d5\\L^Jd3b5ZL`Jf3`5XLcJg3[8000O1O1O1O1O1O1O1N2O1O1O1O1O1O1N20000O100000000O101O1O001O1N2O001O1O001O1O001O1O001O1O1O00100O001O1O001O1O1O1O1O101N1O1O1O1O1OO100O100O101N100O1jMV2O2N1O2N1N2O2N1O2M2O1O2N1N3N1O2M2O2N4K6J5L4K6J5K5L4J6H9G\\f_1"}, "label": "woman eating soup"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman\"."}], "task": "refering", "answer_template": "The \"woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 161, 162, 163, 164, 165, 166, 167, 171, 172, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 201, 202, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 253], "bbox": [0.0, 0.0, 0.841015625, 0.6561875], "iscrowd": 0, "area": 104341.91660000001, "rle": {"size": [480, 640], "counts": "i92i>5K5L3_OB]Bc0R=MjB8e<8WCMY<b0dCBk;R2I7J5L5J6J6J6J6K4K6J6J6J5L5J6J6J5M4M3M3M2N3M3M3M2N3M3M3M2N3M3L4M2M4J6I6K6J6J6N1O2N2N1O20O0100O0100O010O10O10O10O0100O0100O0100O010O10O10O10O0100O010000O10O01000000O010000000000O010000000O1000O100000O100000O10O1000000000O0100000000O10O010000O100O10O010000O100O010O10000O10O0100O100L4K4N3M3N2N2N2O1O001O1N2O1O1O001O100O100O1O100O100O100O1O1O1O1O1N2O1O1O1O1N2O1XOQH]JQ8a5SHZJP8d5UHWJm7g5XHTJj7j5g0N200000000000000000000000000000000000000000000000000001O00000000001O000000001O000000001O000000001O00000000001O000000001O000001O01O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O000000O100000000O100000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000001O0000001O0000001O00010O0000001O00001O0000001O00001O0000001OM3L4K5L4K5L41O1O1O1O1O1O2N1O1O1O1O1O1O1OjKSGZ2l8cMXG]2g8`M^G_2a8^McGc2\\8YMhGg2W8WMmGh2R8WMPHi2o7UMUHj2j7TMYHl2f7SM]Hm2a7QMbHo2]7oLfHQ3Y7nLiHR3V7mLlHS3S7kLPIU3o6jLSIU3m6jLTIW3k6hLWIW3i6hLYIX3f6gL\\IX3d6gL^IY3a6fLaIY3_6fLcIZ3\\6eLfIZ3Z6eLhI[3W6dLkI[3U6dLmI\\3R6bLQJ]3o5aLTJ_3k5_LXJ`3h5^L[Jb3d5\\L^Jd3b5ZL`Jf3`5XLcJg3[8000O1O1O1O1O1O1O1N2O1O1O1O1O1O1N20000O100000000O101O1O001O1N2O001O1O001O1O001O1O001O1O1O00100O001O1O001O1O1O1O1O101N1O1O1O1O1OO100O100O101N100O1jMV2O2N1O2N1N2O2N1O2M2O1O2N1N3N1O2M2O2N4K6J5L4K6J5K5L4J6H9G\\f_1"}, "label": "woman"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of red colored soup\"."}], "task": "refering", "answer_template": "The \"a cup of red colored soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.003359375, 0.6360208333333334, 0.7516875, 0.9857916666666667], "iscrowd": 0, "area": 71022.04830000001, "rle": {"size": [480, 640], "counts": "SY1e3Z;100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10001O00000O100000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O1000000000000000000000000O1000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O1O001O1O1O2N1O2J5H9G8H9G8G:G8H9I7J5J7J5J7J5JUUZ2"}, "label": "a cup of red colored soup"}]}
{"id": 440614, "image": "COCO_train2014_000000440614.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big bowl of vegetable soup on a table in front of a lady who is eating soup from her bowl\"."}], "task": "refering", "answer_template": "The \"a big bowl of vegetable soup on a table in front of a lady who is eating soup from her bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.003359375, 0.6360208333333334, 0.7516875, 0.9857916666666667], "iscrowd": 0, "area": 71022.04830000001, "rle": {"size": [480, 640], "counts": "SY1e3Z;100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10001O00000O100000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O1000000000000000000000000O1000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O1O001O1O1O2N1O2J5H9G8H9G8G:G8H9I7J5J7J5J7J5JUUZ2"}, "label": "a big bowl of vegetable soup on a table in front of a lady who is eating soup from her bowl"}]}
{"id": 80168, "image": "COCO_train2014_000000080168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of dark wood\"."}], "task": "refering", "answer_template": "The \"a piece of dark wood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 41, 42, 43, 64, 65, 66, 87, 88, 89, 90, 111, 112, 113, 134, 135, 136, 137, 158, 159, 160], "bbox": [0.792140625, 0.004125, 0.9994375, 0.35916666666666663], "iscrowd": 0, "area": 9975.841199999995, "rle": {"size": [480, 640], "counts": "Re]73m>6J7I6J6J7I6J6J6K6I6J4L3M3M4L3M3M3M3M3M4L3M3M3M3M3M5K5L5J5K6J4L2N1O2N2N1O2N2N2N110001N2O1O1N1N3M3M3M2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2QOUCEl<O_C1c<CgC=^=000001O00000000001O00000000001N10000000001O00000000001O00000000001O0000000O101O001O7I7IgJ"}, "label": "a piece of dark wood"}]}
{"id": 80168, "image": "COCO_train2014_000000080168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden object coming out right behind the farther right side of the cat\"."}], "task": "refering", "answer_template": "The \"the wooden object coming out right behind the farther right side of the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 41, 42, 43, 64, 65, 66, 87, 88, 89, 90, 111, 112, 113, 134, 135, 136, 137, 158, 159, 160], "bbox": [0.792140625, 0.004125, 0.9994375, 0.35916666666666663], "iscrowd": 0, "area": 9975.841199999995, "rle": {"size": [480, 640], "counts": "Re]73m>6J7I6J6J7I6J6J6K6I6J4L3M3M4L3M3M3M3M3M4L3M3M3M3M3M5K5L5J5K6J4L2N1O2N2N1O2N2N2N110001N2O1O1N1N3M3M3M2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2QOUCEl<O_C1c<CgC=^=000001O00000000001O00000000001N10000000001O00000000001O00000000001O0000000O101O001O7I7IgJ"}, "label": "the wooden object coming out right behind the farther right side of the cat"}]}
{"id": 80168, "image": "COCO_train2014_000000080168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bambo stool with a white cushion on top\"."}], "task": "refering", "answer_template": "The \"a bambo stool with a white cushion on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 80, 81, 82, 83, 106], "bbox": [0.38709375, 0.0018750000000000001, 0.651078125, 0.24195833333333333], "iscrowd": 0, "area": 13096.449849999999, "rle": {"size": [480, 640], "counts": "_Yd39f>2N2N2N2N2O1N2jN[OnCg0Q<[OlCg0S<[OjCh0T<ZOjCg0U<\\OgCf0X<\\OeCf0Z<\\OcCf0\\<\\OaCe0_<]O_Cc0a<@[Ca0e<AXCa0g<AVC`0j<k0O10O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O1000000O10000O1000000O]DZNn9f1RFZNn9f1QF[No9d1QF]No9c1PF^No9c1iDmN8Ao:e2lD\\MT;i2gDWMY;n2aDSM_;W3000000000000000O10000000000000000000000O10000000000000000000000O1000000000000000O10000000001O00000000000000001O00aM`D^1`;]NeDc1[;XNjDh1V;SNPEl1P;oMUEQ2k:jMZEW2g;01O00000010O0001O0001O01O00001O01O0001O000010O000001O1N3N1N3M3\\Oc0B?B>Ae_X3"}, "label": "a bambo stool with a white cushion on top"}]}
{"id": 80168, "image": "COCO_train2014_000000080168.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cushion on a chair\"."}], "task": "refering", "answer_template": "The \"a white cushion on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 36, 37, 55, 56, 57, 58, 59, 60, 80, 81, 82, 83, 106], "bbox": [0.38709375, 0.0018750000000000001, 0.651078125, 0.24195833333333333], "iscrowd": 0, "area": 13096.449849999999, "rle": {"size": [480, 640], "counts": "_Yd39f>2N2N2N2N2O1N2jN[OnCg0Q<[OlCg0S<[OjCh0T<ZOjCg0U<\\OgCf0X<\\OeCf0Z<\\OcCf0\\<\\OaCe0_<]O_Cc0a<@[Ca0e<AXCa0g<AVC`0j<k0O10O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O1000000O10000O1000000O]DZNn9f1RFZNn9f1QF[No9d1QF]No9c1PF^No9c1iDmN8Ao:e2lD\\MT;i2gDWMY;n2aDSM_;W3000000000000000O10000000000000000000000O10000000000000000000000O1000000000000000O10000000001O00000000000000001O00aM`D^1`;]NeDc1[;XNjDh1V;SNPEl1P;oMUEQ2k:jMZEW2g;01O00000010O0001O0001O01O00001O01O0001O000010O000001O1N3N1N3M3\\Oc0B?B>Ae_X3"}, "label": "a white cushion on a chair"}]}
{"id": 530729, "image": "COCO_train2014_000000530729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table nearest the window\"."}], "task": "refering", "answer_template": "The \"table nearest the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 231, 232, 254, 255, 277, 278], "bbox": [0.021718750000000002, 0.4562083333333333, 0.288015625, 0.7708125], "iscrowd": 0, "area": 7373.075000000003, "rle": {"size": [480, 640], "counts": "ai61o>1O1O2N1O1N2O2N8H8H8H7H9H8H8H8H8H8G9H8H7I001O1N101O001O001N101O1O001O0O2O001O1O001N101O001O1O0O2O00eLYH^Of7b0\\300O010000000O10O100000O1000O1000O100000O10O10000000O01000000000O0100000000O010000000O10O100000O1000O1000O100000O10O10000000O0100000000O010000000O10O100000O1000O1000O101O001O1N101O001O0O2O001O001N2O001O001O0O2O001O001N101O1O001N101O001Ob`e6"}, "label": "table nearest the window"}]}
{"id": 530729, "image": "COCO_train2014_000000530729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stove top near the window\"."}], "task": "refering", "answer_template": "The \"the stove top near the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 231, 232, 254, 255, 277, 278], "bbox": [0.021718750000000002, 0.4562083333333333, 0.288015625, 0.7708125], "iscrowd": 0, "area": 7373.075000000003, "rle": {"size": [480, 640], "counts": "ai61o>1O1O2N1O1N2O2N8H8H8H7H9H8H8H8H8H8G9H8H7I001O1N101O001O001N101O1O001O0O2O001O1O001N101O001O1O0O2O00eLYH^Of7b0\\300O010000000O10O100000O1000O1000O100000O10O10000000O01000000000O0100000000O010000000O10O100000O1000O1000O100000O10O10000000O0100000000O010000000O10O100000O1000O1000O101O001O1N101O001O0O2O001O001N2O001O001O0O2O001O001N101O1O001N101O001Ob`e6"}, "label": "the stove top near the window"}]}
{"id": 465196, "image": "COCO_train2014_000000465196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two skies\"."}], "task": "refering", "answer_template": "The \"the two skies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338], "bbox": [0.2767291666666667, 0.699296875, 0.9339583333333334, 0.8573125000000001], "iscrowd": 0, "area": 6198.271099999999, "rle": {"size": [640, 480], "counts": "fbc24gc05L5N2O1ON2001O00000010O0001O01O00010O0001O01OHBn\\O>ob0EQ]O<jb0IV]O6ib0KW]O5ib0KW]O5ib0KX]O5gb0LX]O4hb0LX]O4hb0LY]O4gb0KY]O5gb0KZ]O4fb0MY]O3gb0MY]O4fb0L[]O3eb0M[]O3eb0M[]O4eb0L[]O3eb0M[]O3eb0M\\]O2db0N\\]O3cb0M]]O3cb0N]]O1cb0O]]O1db0N\\]O3cb0M^]O2bb0O]]O1cb0O]]O2bb0N_]O1ab0O_]O1ab0O`]O0ab00^]O1ab0O_]O1ab0O`]O0`b00`]O1`b0N`]O2`b0O`]O0ab0O_]O1ab0O`]O1`b0N`]O2`b0N`]O2ab0N_]O1ab0O_]O2ab0M_]O3ab0M`]O2ab0M_]O4`b0M`]O2ab0M_]O3ab0M_]O3bb0L_]O4`b0M_]O3bb0Hb]O8Tc00O0000010O0000010O000010O0000010O0001O0001O01O0001O01O00010O0000010O000010O0000010O0000010O0001O01O0001O01O01O0001O01O0001O01O000TODU^O<ja0HR^O8oa0Io]O7Qb0Io]O7Qb0Io]O8Pb0HQ^O7oa0JP^O6Pb0JP^O7Pb0HP^O8Pb0HQ^O7oa0IQ^O7oa0JP^O7oa0IQ^O7oa0IR^O6oa0IQ^O7oa0IQ^O8na0HR^O8na0IR^O6na0JR^O7na0HR^O8na0HS^O7ma0IS^O7ma0JR^O7ma0IS^O7ma0IT^O6ma0IS^O8la0HT^O8la0IS^O7ma0IT^O6la0JT^O7ka0IU^O7la0HT^O8la0HU^O7ka0JT^O7ka0IU^O7ka0IU^O7ka0IV^O7ja0HV^O8ja0IU^O7ka0IU^O7ka0IV^O7ia0IW^O7ia0IW^O7ja0HW^O8ha0IW^O7ia0IW^O7ia0IW^O7ia0IX^O7ha0HX^O8ha0IW^O7ia0IW^O7ia0IX^O7ga0IY^O7ga0IY^O7ha0HX^O9ga0HY^O7ga0IY^O7ga0IY^O7ga0IY^O8fa0H[^O7fa0IY^O7ga0IY^O8fa0HZ^O8fa0H[^O7ea0I[^O7ea0JZ^O7fa0HZ^O8fa0H[^O7ea0I[^O7ea0I[^O8da0H]^O7da0I[^O7ea0I[^O8da0H\\^O8da0H]^O7ca0I]^O7ca0J\\^O7da0H\\^O8da0H]^O7ca0I]^O8ba0H^^O8ba0H^^O8ba0I^^O6ca0I]^O8ba0H^^O8ba0H^^O8ba0H_^O7aa0J^^O7aa0I_^O7ba0H^^O8ba0H_^O8`a0H`^O8`a0H`^O8`a0I`^O6`a0J`^O7`a0H`^O8`a0H`^O8`a0Ha^O8^a0Ia^O7_a0Ia^O7`a0H`^O8`a0Ha^O8^a0Hb^O8^a0Hb^O8^a0Ia^O7_a0Ib^O7^a0Hb^O8^a0Hb^O8^a0Hb^O9]a0Hc^O7]a0Ic^O7]a0Ic^O7^a0Hb^O9]a0Gd^O8\\a0Ic^O7]a0Ic^O8\\a0Hd^O8\\a0He^O7\\a0Hd^O8\\a0Hd^O9[a0He^O7[a0Ie^O7[a0Ie^O7\\a0Hd^O9[a0Gf^O8Za0Ie^O7[a0Ie^O8Za0Hf^O8Za0Hg^O7Za0Hf^O8Za0Hf^O8Za0Ie^O7[a0If^O5[a0Ke^O5[a0Ke^O4]a0Kc^O4^a0Ma^O3_a0Ma^O2`a0N`^O2`a0N`^O1aa0O^^O2ca0M]^O2da0O[^O0fa00Y^O1ga0OY^O0ha00X^O0ha00W^O0ka00T^O0la00T^OOma01R^O0ma01S^ONna02R^OMoa03Q^OMoa03Q^OLPb04P^OLPb04Q^OJoa06R^OJna06R^OIoa07Q^OHPb08P^OHPb08P^OGQb09j0100000000000O010N2N2N2N2Nf`c0"}, "label": "the two skies"}]}
{"id": 465196, "image": "COCO_train2014_000000465196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis of the woman with the red jacket\"."}], "task": "refering", "answer_template": "The \"the skis of the woman with the red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338], "bbox": [0.2767291666666667, 0.699296875, 0.9339583333333334, 0.8573125000000001], "iscrowd": 0, "area": 6198.271099999999, "rle": {"size": [640, 480], "counts": "fbc24gc05L5N2O1ON2001O00000010O0001O01O00010O0001O01OHBn\\O>ob0EQ]O<jb0IV]O6ib0KW]O5ib0KW]O5ib0KX]O5gb0LX]O4hb0LX]O4hb0LY]O4gb0KY]O5gb0KZ]O4fb0MY]O3gb0MY]O4fb0L[]O3eb0M[]O3eb0M[]O4eb0L[]O3eb0M[]O3eb0M\\]O2db0N\\]O3cb0M]]O3cb0N]]O1cb0O]]O1db0N\\]O3cb0M^]O2bb0O]]O1cb0O]]O2bb0N_]O1ab0O_]O1ab0O`]O0ab00^]O1ab0O_]O1ab0O`]O0`b00`]O1`b0N`]O2`b0O`]O0ab0O_]O1ab0O`]O1`b0N`]O2`b0N`]O2ab0N_]O1ab0O_]O2ab0M_]O3ab0M`]O2ab0M_]O4`b0M`]O2ab0M_]O3ab0M_]O3bb0L_]O4`b0M_]O3bb0Hb]O8Tc00O0000010O0000010O000010O0000010O0001O0001O01O0001O01O00010O0000010O000010O0000010O0000010O0001O01O0001O01O01O0001O01O0001O01O000TODU^O<ja0HR^O8oa0Io]O7Qb0Io]O7Qb0Io]O8Pb0HQ^O7oa0JP^O6Pb0JP^O7Pb0HP^O8Pb0HQ^O7oa0IQ^O7oa0JP^O7oa0IQ^O7oa0IR^O6oa0IQ^O7oa0IQ^O8na0HR^O8na0IR^O6na0JR^O7na0HR^O8na0HS^O7ma0IS^O7ma0JR^O7ma0IS^O7ma0IT^O6ma0IS^O8la0HT^O8la0IS^O7ma0IT^O6la0JT^O7ka0IU^O7la0HT^O8la0HU^O7ka0JT^O7ka0IU^O7ka0IU^O7ka0IV^O7ja0HV^O8ja0IU^O7ka0IU^O7ka0IV^O7ia0IW^O7ia0IW^O7ja0HW^O8ha0IW^O7ia0IW^O7ia0IW^O7ia0IX^O7ha0HX^O8ha0IW^O7ia0IW^O7ia0IX^O7ga0IY^O7ga0IY^O7ha0HX^O9ga0HY^O7ga0IY^O7ga0IY^O7ga0IY^O8fa0H[^O7fa0IY^O7ga0IY^O8fa0HZ^O8fa0H[^O7ea0I[^O7ea0JZ^O7fa0HZ^O8fa0H[^O7ea0I[^O7ea0I[^O8da0H]^O7da0I[^O7ea0I[^O8da0H\\^O8da0H]^O7ca0I]^O7ca0J\\^O7da0H\\^O8da0H]^O7ca0I]^O8ba0H^^O8ba0H^^O8ba0I^^O6ca0I]^O8ba0H^^O8ba0H^^O8ba0H_^O7aa0J^^O7aa0I_^O7ba0H^^O8ba0H_^O8`a0H`^O8`a0H`^O8`a0I`^O6`a0J`^O7`a0H`^O8`a0H`^O8`a0Ha^O8^a0Ia^O7_a0Ia^O7`a0H`^O8`a0Ha^O8^a0Hb^O8^a0Hb^O8^a0Ia^O7_a0Ib^O7^a0Hb^O8^a0Hb^O8^a0Hb^O9]a0Hc^O7]a0Ic^O7]a0Ic^O7^a0Hb^O9]a0Gd^O8\\a0Ic^O7]a0Ic^O8\\a0Hd^O8\\a0He^O7\\a0Hd^O8\\a0Hd^O9[a0He^O7[a0Ie^O7[a0Ie^O7\\a0Hd^O9[a0Gf^O8Za0Ie^O7[a0Ie^O8Za0Hf^O8Za0Hg^O7Za0Hf^O8Za0Hf^O8Za0Ie^O7[a0If^O5[a0Ke^O5[a0Ke^O4]a0Kc^O4^a0Ma^O3_a0Ma^O2`a0N`^O2`a0N`^O1aa0O^^O2ca0M]^O2da0O[^O0fa00Y^O1ga0OY^O0ha00X^O0ha00W^O0ka00T^O0la00T^OOma01R^O0ma01S^ONna02R^OMoa03Q^OMoa03Q^OLPb04P^OLPb04Q^OJoa06R^OJna06R^OIoa07Q^OHPb08P^OHPb08P^OGQb09j0100000000000O010N2N2N2N2Nf`c0"}, "label": "the skis of the woman with the red jacket"}]}
{"id": 440623, "image": "COCO_train2014_000000440623.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass with juice\"."}], "task": "refering", "answer_template": "The \"a glass with juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126], "bbox": [0.35339062499999996, 0.06774117647058824, 0.519578125, 0.3927058823529412], "iscrowd": 0, "area": 11869.7487, "rle": {"size": [425, 640], "counts": "Tkm22V=8H7I8H7I8H7I6J6J5K5K6J5K5K6J5L4L5K4K5L5K4L4K1000000O100000000O1000000O1000000O1000000000000000001O0000000000000000000000001O000000O1O100O100O101N1O101N100O20O1O1O100O1O101N1O100OM3L4M4J5L4K6J6J6J5L5J6J6J6J6K5J6JR\\o3"}, "label": "a glass with juice"}]}
{"id": 440623, "image": "COCO_train2014_000000440623.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass with a drink in it by the plate of food\"."}], "task": "refering", "answer_template": "The \"the glass with a drink in it by the plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126], "bbox": [0.35339062499999996, 0.06774117647058824, 0.519578125, 0.3927058823529412], "iscrowd": 0, "area": 11869.7487, "rle": {"size": [425, 640], "counts": "Tkm22V=8H7I8H7I8H7I6J6J5K5K6J5K5K6J5L4L5K4K5L5K4L4K1000000O100000000O1000000O1000000O1000000000000000001O0000000000000000000000001O000000O1O100O100O101N1O101N100O20O1O1O100O1O101N1O100OM3L4M4J5L4K6J6J6J5L5J6J6J6J6K5J6JR\\o3"}, "label": "the glass with a drink in it by the plate of food"}]}
{"id": 440623, "image": "COCO_train2014_000000440623.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli stems that are partially hanging off the plate\"."}], "task": "refering", "answer_template": "The \"broccoli stems that are partially hanging off the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 145, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197], "bbox": [0.33596875000000004, 0.3647764705882353, 0.5994843750000001, 0.5624470588235294], "iscrowd": 0, "area": 6896.624199999999, "rle": {"size": [425, 640], "counts": "a]i26R=101N2N1O2N1O2O1N2N2O1N2N1O2N2N1O2O00000000000002N2N2N2O10O0O2OO1O0000000000000001O0O100O101O0O2O1O1O2M2O2N1O1N4M3M3M0O10O1000O10O01O1O1B[NVEg1i:\\NTEe1k:^NREc1m:=O1O1OI610O1001O2N1000001O0O100010O2N2N2O0O1O010O1O0010O01O01O01O0001O01O0000010O00001nNjDOV;NoD1P;MSE2m:MUE2l:MTE3l:MTE3l:MUE2k:MVE4j:KVE5j:KWE4i:LWE4j:KVE5j:KVE6i:JXE5h:KXE5i:JWE6i:IYE6g:JYE7f:IZE7g:HZE7f:IZE7f:IZE6h:IYE6g:JYE6g:JYE6g:JYE5i:JWE6i:JVE7j:IVE7k:HUE8k:HUE7l:ITE7l:ITE7m:HSE8m:HSE8m:HRE9n:GRE9o:FQE:o:FQE:o:FSE8n:GTE7l:IVE5j:KWE3j:MXE0j:OZEMf:2Y10001O000O3N1O]_Z3"}, "label": "broccoli stems that are partially hanging off the plate"}]}
{"id": 244016, "image": "COCO_train2014_000000244016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant with its head down\"."}], "task": "refering", "answer_template": "The \"the elephant with its head down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 312, 313, 322, 323, 324, 325, 326, 327, 329, 330, 331, 332], "bbox": [0.0015, 0.33932084309133487, 0.6566875, 0.9865105386416863], "iscrowd": 0, "area": 90600.37495000001, "rle": {"size": [427, 640], "counts": "gf0d3R6f3N100O100O100O100O100O100O2O0O100O100O100O100O100O101N100O100O100O100O100O101N100O100O100O100O100O100O2O0O1000000O10000000O010O1O10O0100O100O010O100O010O100O0100000000O0100000O100000O10O1000000O0100000O100000O10000000000O10000000000O10001O0000000O10000000000O10000000000O100000001O000O10000000000O10000000000O1000000O100O2O0O100O100O100O100O10000O100O100O100O100O2O0O100O100O100O10000O100O100O100O100O100O100O100O100O10000O10]JhIS4h70000000000000000000O1000000000000000001O000000000O1000001O00001O00001O00001O001O0O101O00001O00001O001O00001O00001N10001O001O00001O00001O00001O0dL^KnMc4S2WMRLj2m3Z31O001O001O00001N101O001O001O001O001O001O001O0O2O001O001O001O00001O001O001O001N101O1O1O1O1O001O1O1O1O1O001O1N2O1O1O001O11O000N2I7J6J6I7J6M3O2O0O1O100O11O1001N101O00001O000O2O001O00001NaGlKg7T4UHoKl7Q4PHQLS8n3iGTLY8l3cGWL_8h3]GZLe8U42O2M6J8I6I7H8I3L5L4K4M4K4L5L4K4M4K5L3L5L3L5K5L3L5L3L5L4K4M4K4L5LZ\\k2"}, "label": "the elephant with its head down"}]}
{"id": 244016, "image": "COCO_train2014_000000244016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with with no tusks\"."}], "task": "refering", "answer_template": "The \"an elephant with with no tusks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 312, 313, 322, 323, 324, 325, 326, 327, 329, 330, 331, 332], "bbox": [0.0015, 0.33932084309133487, 0.6566875, 0.9865105386416863], "iscrowd": 0, "area": 90600.37495000001, "rle": {"size": [427, 640], "counts": "gf0d3R6f3N100O100O100O100O100O100O2O0O100O100O100O100O100O101N100O100O100O100O100O101N100O100O100O100O100O100O2O0O1000000O10000000O010O1O10O0100O100O010O100O010O100O0100000000O0100000O100000O10O1000000O0100000O100000O10000000000O10000000000O10001O0000000O10000000000O10000000000O100000001O000O10000000000O10000000000O1000000O100O2O0O100O100O100O100O10000O100O100O100O100O2O0O100O100O100O10000O100O100O100O100O100O100O100O100O10000O10]JhIS4h70000000000000000000O1000000000000000001O000000000O1000001O00001O00001O00001O001O0O101O00001O00001O001O00001O00001N10001O001O00001O00001O00001O0dL^KnMc4S2WMRLj2m3Z31O001O001O00001N101O001O001O001O001O001O001O0O2O001O001O001O00001O001O001O001N101O1O1O1O1O001O1O1O1O1O001O1N2O1O1O001O11O000N2I7J6J6I7J6M3O2O0O1O100O11O1001N101O00001O000O2O001O00001NaGlKg7T4UHoKl7Q4PHQLS8n3iGTLY8l3cGWL_8h3]GZLe8U42O2M6J8I6I7H8I3L5L4K4M4K4L5L4K4M4K5L3L5L3L5K5L3L5L3L5L4K4M4K4L5LZ\\k2"}, "label": "an elephant with with no tusks"}]}
{"id": 244016, "image": "COCO_train2014_000000244016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant walking toward the camera\"."}], "task": "refering", "answer_template": "The \"an elephant walking toward the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 269, 270, 271, 293, 294, 316, 317, 339, 340], "bbox": [0.581734375, 0.09494145199063232, 0.8725937500000001, 0.9601170960187354], "iscrowd": 0, "area": 41898.45069999999, "rle": {"size": [427, 640], "counts": "hek48o<7I8H7I7I8H7J7H7I8H7I8J5L5K4M3L5K4L5K4L5K4L5K4`HcKk5b4QJaKk5c4RJ`Kj5d4RJ`Kk5d4QJ_Kk5e4lIeKo5`4aIoK[6d5L4L2N10O01O010O010O00100O0010O01O010O010O0010O01O10O01O100O010O1O10O01O10O0100O1O010O1O10O01O100O00101N3M3N1N3M3M3N2M2N3N2M3M2O2M3N2N2N101N2N2O0O2N2N2O1N1O5L5J6J6[M`FT1e9[NgFd1`:0100O1O10O01OO101O0O101O0O10000O2O0O10000SOiMeFX2[9jMcFV2\\9mMbFS2^9oM_FR2a9PN]FQ2a9SN[Fn1e9TNYFl1g9VNWFj1h9YNUFh1k9ZNSFf1m9\\NQFd1n9m00001TOk0J6K5J6K6J6G9G9F:G9F:K5K6J5K5K5J6K5K5K5K5G:[Od0XOh0@`0M3N2M3M3N2M3M4M9B`0A>A`0AbRR1"}, "label": "an elephant walking toward the camera"}]}
{"id": 244016, "image": "COCO_train2014_000000244016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with a broken tusk\"."}], "task": "refering", "answer_template": "The \"an elephant with a broken tusk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 269, 270, 271, 293, 294, 316, 317, 339, 340], "bbox": [0.581734375, 0.09494145199063232, 0.8725937500000001, 0.9601170960187354], "iscrowd": 0, "area": 41898.45069999999, "rle": {"size": [427, 640], "counts": "hek48o<7I8H7I7I8H7J7H7I8H7I8J5L5K4M3L5K4L5K4L5K4L5K4`HcKk5b4QJaKk5c4RJ`Kj5d4RJ`Kk5d4QJ_Kk5e4lIeKo5`4aIoK[6d5L4L2N10O01O010O010O00100O0010O01O010O010O0010O01O10O01O100O010O1O10O01O10O0100O1O010O1O10O01O100O00101N3M3N1N3M3M3N2M2N3N2M3M2O2M3N2N2N101N2N2O0O2N2N2O1N1O5L5J6J6[M`FT1e9[NgFd1`:0100O1O10O01OO101O0O101O0O10000O2O0O10000SOiMeFX2[9jMcFV2\\9mMbFS2^9oM_FR2a9PN]FQ2a9SN[Fn1e9TNYFl1g9VNWFj1h9YNUFh1k9ZNSFf1m9\\NQFd1n9m00001TOk0J6K5J6K6J6G9G9F:G9F:K5K6J5K5K5J6K5K5K5K5G:[Od0XOh0@`0M3N2M3M3N2M3M4M9B`0A>A`0AbRR1"}, "label": "an elephant with a broken tusk"}]}
{"id": 6447, "image": "COCO_train2014_000000006447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe ' s head\"."}], "task": "refering", "answer_template": "The \"giraffe ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 57, 58, 59, 60, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 308, 309, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.10067187500000001, 0.11824355971896956, 0.9707343749999999, 1.0], "iscrowd": 0, "area": 107148.82454999996, "rle": {"size": [427, 640], "counts": "mXk03W=3M3L4M3M3M3L4M3M3M2M4M3M2N1O2M3N2N2N2M3N1O2N2N2M3N2N2N1N3N2N2N2N2N1O2O1O1N2O1N2O0O2O1O1N2O1N101O1N2O1N2O001N2O0O10000O2O0O100O10000O100O101O0O100O10000O10001O0001O00000000001O01O000001O000000000010O00002N2N2O1N2N2N2N3N1N2N2N2N2N2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N3M2N2N1O1O2N1O1O1N2O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N100O100O101N100O1O101N1O1O1O1N101O001O1N101O001N2O001O001N2O001O0O2O1O001N101O001O1N101O001N2M2O2M2O2M3M2O2M2O2M3N1N2O2M2N2O1N3N1O1O2O0O100O2O0O101N100O101N3N1N2O1N3N1N2O1O2N1O1O1O2N1O1O1O2N1O1O1O00000O1O1O1O1O100O1O1O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O100001O000VFUNT8l1dGZN]8f1cGZN]8f1cGYN^8g1cGXN]8i1bGVN_8j1bGUN^8k1bGTN_8m1_GTNa8l1^GTNc8l1\\GUNd8l1ZGTNg8l1XGUNh8m1VGRNk8o1SGRNm8P2PGPNQ9R2lFoMT9S2iFmMX9T2fFmMZ9U2cFlM]9V2`FjMa9X2\\FiMd9X2ZFhMg9Z2VFgMj9[2SFeMn9l20O100O100O100O100O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1OVHbMi4^2VKiMd4V2\\KQN^4n1bKXNY4h1fK^NU4a1jKcNT4\\1lKgNR4X1nKkNP4T1PLoNn3o0SLTOk3k0TLYOj3f0VL]Oh3b0XLAf3=[LFc39]LHc37]LKb34]LO7nM:Q2_O36PN8l1_O98nM6h1@>9mM4d1Ab0;lM2`1Ai0;kM1[1Bn0<jMOW1CR1>jMLS1CX1`0hMJn0E^1?gMJj0Eb1a0gMGf0Eh1c0eMEb0Fl1d0fMC<GS2d0eMB7HX2e0cMA6FZ2h0dM_O2I[2g0fM]O0K[2f0iM\\OLN[2e0mMZOH1[2e0oMXOG2Z2e0SNVOC4[2e0VNTO@6Z2f0YNQO]O9Z2e0]NoNZO;Y2f0_NmNXO=Y2e0cNkNTO`0Y2d0fNjNROa0X2e0gNiNQOa0Y2e0iNhNoNb0X2f0kNfNmNd0X2f0mNdNlNd0X2h0mNcNmNb0W2k0nNaNlNb0W2n0nN^NmNa0V2Q1oN\\NmN?V2U1nN[NmN>V2X1nNXNnN=U2[1oNVNnN<T2^1POTNmN;U2b1nNRNoN9T2e1oNPNoN8S2h1POnMnN8R2l1QOjMoN7Q2o1QOiMoN5R2R2QOgMoN4Q2V2QOdMPO3P2Y2e1eM[N\\2Z61O001O000O2O00001O0O101O00001N10001O00001N10001O000O2O000000000O100000O10O100000O10O10000000O1000000O1O1O2N1O1O100O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2M2N2N3L4M2NSf7"}, "label": "giraffe ' s head"}]}
{"id": 6447, "image": "COCO_train2014_000000006447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the face of a giraffe looking right at the camera\"."}], "task": "refering", "answer_template": "The \"the face of a giraffe looking right at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 57, 58, 59, 60, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 308, 309, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.10067187500000001, 0.11824355971896956, 0.9707343749999999, 1.0], "iscrowd": 0, "area": 107148.82454999996, "rle": {"size": [427, 640], "counts": "mXk03W=3M3L4M3M3M3L4M3M3M2M4M3M2N1O2M3N2N2N2M3N1O2N2N2M3N2N2N1N3N2N2N2N2N1O2O1O1N2O1N2O0O2O1O1N2O1N101O1N2O1N2O001N2O0O10000O2O0O100O10000O100O101O0O100O10000O10001O0001O00000000001O01O000001O000000000010O00002N2N2O1N2N2N2N3N1N2N2N2N2N2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N3M2N2N1O1O2N1O1O1N2O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N100O100O101N100O1O101N1O1O1O1N101O001O1N101O001N2O001O001N2O001O0O2O1O001N101O001O1N101O001N2M2O2M2O2M3M2O2M2O2M3N1N2O2M2N2O1N3N1O1O2O0O100O2O0O101N100O101N3N1N2O1N3N1N2O1O2N1O1O1O2N1O1O1O2N1O1O1O00000O1O1O1O1O100O1O1O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O100001O000VFUNT8l1dGZN]8f1cGZN]8f1cGYN^8g1cGXN]8i1bGVN_8j1bGUN^8k1bGTN_8m1_GTNa8l1^GTNc8l1\\GUNd8l1ZGTNg8l1XGUNh8m1VGRNk8o1SGRNm8P2PGPNQ9R2lFoMT9S2iFmMX9T2fFmMZ9U2cFlM]9V2`FjMa9X2\\FiMd9X2ZFhMg9Z2VFgMj9[2SFeMn9l20O100O100O100O100O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1OVHbMi4^2VKiMd4V2\\KQN^4n1bKXNY4h1fK^NU4a1jKcNT4\\1lKgNR4X1nKkNP4T1PLoNn3o0SLTOk3k0TLYOj3f0VL]Oh3b0XLAf3=[LFc39]LHc37]LKb34]LO7nM:Q2_O36PN8l1_O98nM6h1@>9mM4d1Ab0;lM2`1Ai0;kM1[1Bn0<jMOW1CR1>jMLS1CX1`0hMJn0E^1?gMJj0Eb1a0gMGf0Eh1c0eMEb0Fl1d0fMC<GS2d0eMB7HX2e0cMA6FZ2h0dM_O2I[2g0fM]O0K[2f0iM\\OLN[2e0mMZOH1[2e0oMXOG2Z2e0SNVOC4[2e0VNTO@6Z2f0YNQO]O9Z2e0]NoNZO;Y2f0_NmNXO=Y2e0cNkNTO`0Y2d0fNjNROa0X2e0gNiNQOa0Y2e0iNhNoNb0X2f0kNfNmNd0X2f0mNdNlNd0X2h0mNcNmNb0W2k0nNaNlNb0W2n0nN^NmNa0V2Q1oN\\NmN?V2U1nN[NmN>V2X1nNXNnN=U2[1oNVNnN<T2^1POTNmN;U2b1nNRNoN9T2e1oNPNoN8S2h1POnMnN8R2l1QOjMoN7Q2o1QOiMoN5R2R2QOgMoN4Q2V2QOdMPO3P2Y2e1eM[N\\2Z61O001O000O2O00001O0O101O00001N10001O00001N10001O000O2O000000000O100000O10O100000O10O10000000O1000000O1O1O2N1O1O100O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2M2N2N3L4M2NSf7"}, "label": "the face of a giraffe looking right at the camera"}]}
{"id": 506162, "image": "COCO_train2014_000000506162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coffee mug labeled madison concourse hotel\"."}], "task": "refering", "answer_template": "The \"a coffee mug labeled madison concourse hotel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 37, 38, 39, 40, 41, 42, 43, 44, 60, 61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 108, 109], "bbox": [0.613640625, 0.0, 0.915578125, 0.24675], "iscrowd": 0, "area": 16768.458349999994, "rle": {"size": [480, 640], "counts": "]Yh5c0c=j0VOj0XOi0O001O001O1O001O001O001O1O001O001O001O1O001O001O001O00001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O0000000000000000000000000000000000000O01000000000000000000000000000000O10000000000000000O1O1O100O1O1O1O1O100O1O1O1O1O1N2M3M3M3M3M3O1O1XOeCkN[<Q1iCoNW<l0nCSOS<i0QDWOo;d0UD]Ok;?YDAg;:^DEc;7aDI_;2fDNZ;LlD3U;ESE;m:\\O\\Ed0_<10O10000O1000O10O10000O100000O010000O100000O01000\\N_OiD`0X;BeD?[;CcD=\\;FaD;_;G_D8b;J[D7e;KYD5g;MTD5m;MiC;W<GVCj0j<f00O100O1M3N1N5L5J7J5Jg0ZOPYi0"}, "label": "a coffee mug labeled madison concourse hotel"}]}
{"id": 506162, "image": "COCO_train2014_000000506162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white coffee mug from madison concourse hotel\"."}], "task": "refering", "answer_template": "The \"a white coffee mug from madison concourse hotel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 37, 38, 39, 40, 41, 42, 43, 44, 60, 61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 108, 109], "bbox": [0.613640625, 0.0, 0.915578125, 0.24675], "iscrowd": 0, "area": 16768.458349999994, "rle": {"size": [480, 640], "counts": "]Yh5c0c=j0VOj0XOi0O001O001O1O001O001O001O1O001O001O001O1O001O001O001O00001O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O0000000000000000000000000000000000000O01000000000000000000000000000000O10000000000000000O1O1O100O1O1O1O1O100O1O1O1O1O1N2M3M3M3M3M3O1O1XOeCkN[<Q1iCoNW<l0nCSOS<i0QDWOo;d0UD]Ok;?YDAg;:^DEc;7aDI_;2fDNZ;LlD3U;ESE;m:\\O\\Ed0_<10O10000O1000O10O10000O100000O010000O100000O01000\\N_OiD`0X;BeD?[;CcD=\\;FaD;_;G_D8b;J[D7e;KYD5g;MTD5m;MiC;W<GVCj0j<f00O100O1M3N1N5L5J7J5Jg0ZOPYi0"}, "label": "a white coffee mug from madison concourse hotel"}]}
{"id": 153909, "image": "COCO_train2014_000000153909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coffee kept on the table without untouched\"."}], "task": "refering", "answer_template": "The \"a coffee kept on the table without untouched\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 351], "bbox": [0.1955, 0.5206666666666666, 0.42978125, 0.8959374999999999], "iscrowd": 0, "area": 19739.02975, "rle": {"size": [480, 640], "counts": "Tlj12k>5K5K6J5K6J6L4M3L4M3L4M3L5L4K5L4K5L3L5L4K5L4K5L4K5L3L5L4K5L4K5L4K5L3L5L4K5L4K3N0O10O01000000O0100000O10O1000O100000O10O100000O01000000001N100000001N100000001N100000001O0O1000001O0O1000001O000O10001O000O10001O000O10001O00000O101O00001N10001O00001N10001N1K5J7I6K5J7I6K5J7I6J6I9H7I7H9H7H8I7I8G8K5L4L5KXjZ5"}, "label": "a coffee kept on the table without untouched"}]}
{"id": 153909, "image": "COCO_train2014_000000153909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cup of coffee near the croissant\"."}], "task": "refering", "answer_template": "The \"the cup of coffee near the croissant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 351], "bbox": [0.1955, 0.5206666666666666, 0.42978125, 0.8959374999999999], "iscrowd": 0, "area": 19739.02975, "rle": {"size": [480, 640], "counts": "Tlj12k>5K5K6J5K6J6L4M3L4M3L4M3L5L4K5L4K5L3L5L4K5L4K5L4K5L3L5L4K5L4K5L4K5L3L5L4K5L4K3N0O10O01000000O0100000O10O1000O100000O10O100000O01000000001N100000001N100000001N100000001O0O1000001O0O1000001O000O10001O000O10001O000O10001O00000O101O00001N10001O00001N10001N1K5J7I6K5J7I6K5J7I6J6I9H7I7H9H7H8I7I8G8K5L4L5KXjZ5"}, "label": "the cup of coffee near the croissant"}]}
{"id": 153909, "image": "COCO_train2014_000000153909.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair behind a man\"."}], "task": "refering", "answer_template": "The \"a white chair behind a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 84, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 155, 156, 157, 179, 180], "bbox": [0.49921875, 0.1684166666666667, 0.86371875, 0.4452291666666667], "iscrowd": 0, "area": 12496.588399999999, "rle": {"size": [480, 640], "counts": "ZUf43m>1N3M2N3UNFkD<T;EjD>S;EjD=T;EjD:X;HdD:a:_OkD:a08d:@hD:b07f:BeD9b06i:CbD9a07m:C^D:?6S;B[D<;6Z;AXD;97_;b0ZDBf;l100000000000O101O00000000000000000000000000000000000000000000001O00000000000000000O100000000000000000000000001O00000O10000000000000000000000000000000000O1I7N3L3N2N2N2N2N2M3N2N200001O1O1O1O1O1O;E;E4L000O1O10SNZD`0e;QOnDl0Q;fN`EX1W<M2N0O2E:F;N101O010O000010O01O0010O01O0010O01O001O010O010O010O010O01O010O010O01O010O010O010O010O01O010O00001O010O001O0010O01O001O010O00001O010O001O0WB^OT=c0eBC[==_BJa=j001O001O010O001O1O010O0000L4L5K4M3L4M4K4O2N2O001O1O0O2O001O1O0M4M`cX1"}, "label": "a white chair behind a man"}]}
{"id": 571702, "image": "COCO_train2014_000000571702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a balding man with glasses wearing a wedding ring and holding a baby\"."}], "task": "refering", "answer_template": "The \"a balding man with glasses wearing a wedding ring and holding a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 120, 121, 122, 125, 126, 127, 128, 129, 135, 136, 139, 140, 141, 142, 143, 144, 145, 150, 155, 156, 157, 158, 159, 160, 165, 170, 171, 172, 173, 174, 180, 186, 187, 188, 189, 201, 202, 216, 217, 218, 225, 226, 227, 228, 231, 232, 233, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 270, 271, 272, 273, 274, 276, 277, 278, 286, 287, 288], "bbox": [0.002897196261682243, 0.33464062499999997, 0.6845093457943926, 0.8624375000000001], "iscrowd": 0, "area": 42196.33870000001, "rle": {"size": [640, 428], "counts": "Zk0Q1ob0U3jL0hBnMc8R2]GoMc8P2^GQN`8o1aGQN^8o1cGRN[8o1dGRN\\8m1eGTNY8l1hGTNW8l1jGUNU8j1kGWNT8i1mGXNQ8<^E_Na2U1Q8;dEZN[2]1P89jETNW2c1o78oEQNR2h1n77SFmMP2l1m77UFkMm1P2m74YFiMk1S2m73ZFgMj1W2k72^FdMf1[2l71`FaMe1_2j70cF_Mc1a2j70`I1_6O`I2`6O_I2`60^I0b6eNhJ]1W5`NlJ`1T5]NoJd1P5YNSKg1m4WNTKk1k4TNVKl1j4RNXKo1g4PNZKP2f4oMZKS2e4lM\\KT2d4kM]KV2b4iM^KX2b4gM_KZ2a4dM`K\\2`4cMaK^2^4aMbK`2^4_McKb2]4\\MdKd2\\4\\McKf2\\4YMeKg2\\4XMdKi2[4VMfKj2Z4VMeKl2Z4SMgKm2Z4RMeKP3Z4PMfKP3Z4oLfKS3Z4lLfKT3Z4lLeKV3[4iLeKW3\\4hLcKZ3]4dLdK\\3]4cLbK_3^4`LbK`3_4_LaKb3_4]L`Kd3a4[L_Kf3]:001O001O001O001O0000_Lm_OU3S`0hLP@X3P`0fLS@Z3l?cLW@]3V`0000000O2O0O10000O2O000O100O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O100O1O1O1O1O1O1N2N\\JXNTIg1l6\\NRIb1o6`NPI^1Q7eNmHY1S7kNkHT1U7oNiHo0X7TOfHj0[7YO_D^Om2W1d8^OXD@T3P1e8DPD@\\3j0c8KiCAd3c0b81aCAm3>a8P1`GnN_8S1dGjN[8X1gGeNX8\\1PH\\Nn7f1SHZNk7h1THYNj7h1VHYNh7i1WHXNg7i1YHXNd7k1[HVNc7k1\\HWNb7k1\\HWNb7j1XH]Nf7e1THaNj7`1SHdNk7^1QHfNn7Z1QHhNm7Z1RHgNl7Z1THhNi7Z1VHgNh7Z1XHgNg7Y1YHhNe7Z1ZHgNd7Z1\\HgNc7Z1\\HgNc7Y1\\HiNb7Y1]HhNb7X1^HhNb7X1^HhNb7Y1]HgNb7Z1^HgNa7Z1^HfNb7Z1^HfNa7[1_HfN`7[1^HfNb7Z1^HfNb7[1]HfNa7[1_HeNa7\\1^HdNb7\\1^HdNa7^1]F^Lf0U2l8]1\\FaLg0R2m8^1XFdLi0o1o8^1UFfLk0m1n8^1UFhLk0k1P9^1RFjLK@:X2i9^1PFnLBFd0n1i9`1oEdM7m0j9`1lEfM9j0k9`1jEhM:i0l9`1hEiM;g0l9d1dEgM`0e0l9g1`EfMc0c0m9j1[EfMf0a0o9h3PFWLQ:j3nEVLQ:k3oEULQ:l3mETLT:m3kESLU:n3jERLU:P4jEPLV:Q4hEoKY:Q4gEoKY:R4fEnKY:T4fEkK[:V4cEkK]:W4aEiK_:X4`EhK_:Z4`EeKa:\\4^EdKb:]4\\EdKd:]4[EbKe:`4ZE`Kf:a4YE_Kg:m6000XKZE6f:I_E3a:McEO]:1gEKY:5kEGV:7oEEQ:;UF_Ok9a0gFmNY9S1ZGZNf8e1`GVN`8j1cGSN]8m1fGPNZ8P2iGmMW8R2lGlMT8T2nGjMR8U2QHiMo7W2SHgMm7X2VHfMj7Z2XHdMh7[2[HdMd7\\2fH[MY7d2TIQMk6o2ZImLe6R3_IlLa6S3bIkL]6U3fIiLZ6U3jIjLU6U3nIiLQ6V3TJgLl5X3WJfLh5Y3`JaL`5^3gJ\\LX5c3lJ[LT5d3nJ[LQ5d3SKZLm4e3UKZLj4e3ZKYLf4f3\\KYLc4f3m5M2N3N1N3M3M3N2M3M3M2O2M3M3M3N2M3M3M3N2M3M3M3N2L6K6I6K6I6K?@aSd2"}, "label": "a balding man with glasses wearing a wedding ring and holding a baby"}]}
{"id": 571702, "image": "COCO_train2014_000000571702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man playing with infant holding remote control\"."}], "task": "refering", "answer_template": "The \"man playing with infant holding remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 120, 121, 122, 125, 126, 127, 128, 129, 135, 136, 139, 140, 141, 142, 143, 144, 145, 150, 155, 156, 157, 158, 159, 160, 165, 170, 171, 172, 173, 174, 180, 186, 187, 188, 189, 201, 202, 216, 217, 218, 225, 226, 227, 228, 231, 232, 233, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 270, 271, 272, 273, 274, 276, 277, 278, 286, 287, 288], "bbox": [0.002897196261682243, 0.33464062499999997, 0.6845093457943926, 0.8624375000000001], "iscrowd": 0, "area": 42196.33870000001, "rle": {"size": [640, 428], "counts": "Zk0Q1ob0U3jL0hBnMc8R2]GoMc8P2^GQN`8o1aGQN^8o1cGRN[8o1dGRN\\8m1eGTNY8l1hGTNW8l1jGUNU8j1kGWNT8i1mGXNQ8<^E_Na2U1Q8;dEZN[2]1P89jETNW2c1o78oEQNR2h1n77SFmMP2l1m77UFkMm1P2m74YFiMk1S2m73ZFgMj1W2k72^FdMf1[2l71`FaMe1_2j70cF_Mc1a2j70`I1_6O`I2`6O_I2`60^I0b6eNhJ]1W5`NlJ`1T5]NoJd1P5YNSKg1m4WNTKk1k4TNVKl1j4RNXKo1g4PNZKP2f4oMZKS2e4lM\\KT2d4kM]KV2b4iM^KX2b4gM_KZ2a4dM`K\\2`4cMaK^2^4aMbK`2^4_McKb2]4\\MdKd2\\4\\McKf2\\4YMeKg2\\4XMdKi2[4VMfKj2Z4VMeKl2Z4SMgKm2Z4RMeKP3Z4PMfKP3Z4oLfKS3Z4lLfKT3Z4lLeKV3[4iLeKW3\\4hLcKZ3]4dLdK\\3]4cLbK_3^4`LbK`3_4_LaKb3_4]L`Kd3a4[L_Kf3]:001O001O001O001O0000_Lm_OU3S`0hLP@X3P`0fLS@Z3l?cLW@]3V`0000000O2O0O10000O2O000O100O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O100O1O1O1O1O1O1N2N\\JXNTIg1l6\\NRIb1o6`NPI^1Q7eNmHY1S7kNkHT1U7oNiHo0X7TOfHj0[7YO_D^Om2W1d8^OXD@T3P1e8DPD@\\3j0c8KiCAd3c0b81aCAm3>a8P1`GnN_8S1dGjN[8X1gGeNX8\\1PH\\Nn7f1SHZNk7h1THYNj7h1VHYNh7i1WHXNg7i1YHXNd7k1[HVNc7k1\\HWNb7k1\\HWNb7j1XH]Nf7e1THaNj7`1SHdNk7^1QHfNn7Z1QHhNm7Z1RHgNl7Z1THhNi7Z1VHgNh7Z1XHgNg7Y1YHhNe7Z1ZHgNd7Z1\\HgNc7Z1\\HgNc7Y1\\HiNb7Y1]HhNb7X1^HhNb7X1^HhNb7Y1]HgNb7Z1^HgNa7Z1^HfNb7Z1^HfNa7[1_HfN`7[1^HfNb7Z1^HfNb7[1]HfNa7[1_HeNa7\\1^HdNb7\\1^HdNa7^1]F^Lf0U2l8]1\\FaLg0R2m8^1XFdLi0o1o8^1UFfLk0m1n8^1UFhLk0k1P9^1RFjLK@:X2i9^1PFnLBFd0n1i9`1oEdM7m0j9`1lEfM9j0k9`1jEhM:i0l9`1hEiM;g0l9d1dEgM`0e0l9g1`EfMc0c0m9j1[EfMf0a0o9h3PFWLQ:j3nEVLQ:k3oEULQ:l3mETLT:m3kESLU:n3jERLU:P4jEPLV:Q4hEoKY:Q4gEoKY:R4fEnKY:T4fEkK[:V4cEkK]:W4aEiK_:X4`EhK_:Z4`EeKa:\\4^EdKb:]4\\EdKd:]4[EbKe:`4ZE`Kf:a4YE_Kg:m6000XKZE6f:I_E3a:McEO]:1gEKY:5kEGV:7oEEQ:;UF_Ok9a0gFmNY9S1ZGZNf8e1`GVN`8j1cGSN]8m1fGPNZ8P2iGmMW8R2lGlMT8T2nGjMR8U2QHiMo7W2SHgMm7X2VHfMj7Z2XHdMh7[2[HdMd7\\2fH[MY7d2TIQMk6o2ZImLe6R3_IlLa6S3bIkL]6U3fIiLZ6U3jIjLU6U3nIiLQ6V3TJgLl5X3WJfLh5Y3`JaL`5^3gJ\\LX5c3lJ[LT5d3nJ[LQ5d3SKZLm4e3UKZLj4e3ZKYLf4f3\\KYLc4f3m5M2N3N1N3M3M3N2M3M3M2O2M3M3M3N2M3M3M3N2M3M3M3N2L6K6I6K6I6K?@aSd2"}, "label": "man playing with infant holding remote control"}]}
{"id": 571702, "image": "COCO_train2014_000000571702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby holding the remote controller\"."}], "task": "refering", "answer_template": "The \"the baby holding the remote controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 229, 230, 231, 232], "bbox": [0.0033644859813084112, 0.36148437499999997, 0.5119158878504673, 0.6925625], "iscrowd": 0, "area": 25196.395900000007, "rle": {"size": [640, 428], "counts": "PP13lc0P1PO;E2N1O2N2N2N2N1O2O1N2N1O2YNQNZAQ2d>RNYAP2e>TNWAn1h>UNTAl1k>WNRAk1l>YNPAi1n>[Nn@g1P?]Nl@d1S?_Nj@a1]>dMdAn0K_1_>gMcAn0J[1b>kMcAk0H[1c>nMdAi0EZ1d>RNeAg0CW1g>VNeAW3Y>mLeAR3Z>RMeAm2Y>WMfAg2X>^MfAb2X>bMgA\\2X>hMfAX2X>lMgAR2X>RNgAm1V>XNhAg1X>Q2O010O0100O010O01000O010O10O010O0100O010O10O01N1O2N2N1O2N2N1O2N1O2N2O0100O2O000O10000O100O10000O10000O100O11O001O00001O001O00001O001O01O01O001O001O0PNUAkNk>S1VAnNk>n0XAROh>l0YAVOg>g0[AYOe>d0]A^Ob>`0`A@a><bAE]>9dAI\\>4fALZ>3gANX>2hANY>1hAOW>1iAOGkMS>V2VBOEoMR>R2YBOEQNP>P2\\BgN[O87UOP>k1^B0AXNo=i1`BgN[O85ZOo=f1aBiN[O64_Om=b1eBhN[O71Bn=^1fBjNZO70Eo=Y1gBjN[O7OFS>U1cBoN[OR1U>L`BRO[OR1X>I^BTO[OR1Z>HZBWO\\OP1]>EYBZO[OP1`>@WBAXOo0d>[OWBEVOo0f>WOVBKTOm0i>SOVBOQOn0l>mNVB5oNm0o>hNUB;lNl0R?dNTB`0kNk0T?`NTBe0gNk0X?[NSB4bN;4U1m`0]NR_O>1T1m`0\\NU_O?OT1m`0ZNW_Oa0LU1Za0kNf^OT1[a0kNf^OT1[a0lNd^OT1]a0kNd^OT1\\a0mNd^OR1]a0mNd^OR1]a0nNb^OR1_a0mNb^OR1_a0mNb^OR1_a0lNb^OT1_a0jNc^OU1^a0hNe^OW1\\a0gNe^OY1\\a0eNd^O\\1oa01O100O1O1O100O1O1O100O1O1O100O100000000O1000000001O6J5K6J5K2N1O2N2N1O2N1O1O2N1O1O1O1N2N3M2N2N2O1N3M2N_VR4"}, "label": "the baby holding the remote controller"}]}
{"id": 571702, "image": "COCO_train2014_000000571702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby sitting on mans lap , holding remote control\"."}], "task": "refering", "answer_template": "The \"baby sitting on mans lap , holding remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 229, 230, 231, 232], "bbox": [0.0033644859813084112, 0.36148437499999997, 0.5119158878504673, 0.6925625], "iscrowd": 0, "area": 25196.395900000007, "rle": {"size": [640, 428], "counts": "PP13lc0P1PO;E2N1O2N2N2N2N1O2O1N2N1O2YNQNZAQ2d>RNYAP2e>TNWAn1h>UNTAl1k>WNRAk1l>YNPAi1n>[Nn@g1P?]Nl@d1S?_Nj@a1]>dMdAn0K_1_>gMcAn0J[1b>kMcAk0H[1c>nMdAi0EZ1d>RNeAg0CW1g>VNeAW3Y>mLeAR3Z>RMeAm2Y>WMfAg2X>^MfAb2X>bMgA\\2X>hMfAX2X>lMgAR2X>RNgAm1V>XNhAg1X>Q2O010O0100O010O01000O010O10O010O0100O010O10O01N1O2N2N1O2N2N1O2N1O2N2O0100O2O000O10000O100O10000O10000O100O11O001O00001O001O00001O001O01O01O001O001O0PNUAkNk>S1VAnNk>n0XAROh>l0YAVOg>g0[AYOe>d0]A^Ob>`0`A@a><bAE]>9dAI\\>4fALZ>3gANX>2hANY>1hAOW>1iAOGkMS>V2VBOEoMR>R2YBOEQNP>P2\\BgN[O87UOP>k1^B0AXNo=i1`BgN[O85ZOo=f1aBiN[O64_Om=b1eBhN[O71Bn=^1fBjNZO70Eo=Y1gBjN[O7OFS>U1cBoN[OR1U>L`BRO[OR1X>I^BTO[OR1Z>HZBWO\\OP1]>EYBZO[OP1`>@WBAXOo0d>[OWBEVOo0f>WOVBKTOm0i>SOVBOQOn0l>mNVB5oNm0o>hNUB;lNl0R?dNTB`0kNk0T?`NTBe0gNk0X?[NSB4bN;4U1m`0]NR_O>1T1m`0\\NU_O?OT1m`0ZNW_Oa0LU1Za0kNf^OT1[a0kNf^OT1[a0lNd^OT1]a0kNd^OT1\\a0mNd^OR1]a0mNd^OR1]a0nNb^OR1_a0mNb^OR1_a0mNb^OR1_a0lNb^OT1_a0jNc^OU1^a0hNe^OW1\\a0gNe^OY1\\a0eNd^O\\1oa01O100O1O1O100O1O1O100O1O1O100O100000000O1000000001O6J5K6J5K2N1O2N2N1O2N1O1O2N1O1O1O1N2N3M2N2N2O1N3M2N_VR4"}, "label": "baby sitting on mans lap , holding remote control"}]}
{"id": 448824, "image": "COCO_train2014_000000448824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white suv parked near field\"."}], "task": "refering", "answer_template": "The \"white suv parked near field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219], "bbox": [0.0799375, 0.40939866369710465, 0.5595156250000001, 0.5934298440979955], "iscrowd": 0, "area": 18452.570099999997, "rle": {"size": [449, 640], "counts": "PPg04l=2N2M3N2N2M3N2N1N3N1O2O001O001O001N101O001O0000001O001O1N2O1N2N101O1O1O1_CeNZ<b1O2N0000001O001N101N2O1O010O1O1O02OO1O001O1N2O5K6J1O000O10000000000O100000000000000000000000000000000000000000000000000000000001O00000000000000000000000O1O1O1O100000000O100000000001O0000000000001O001O2N0000000000000000000000000000000000000000000O1000000000000000000002N1O1O00000000000000001O000000000000000000000000001O0000000000000000001O000000001O001O001N2O001O001O002N2N2N2N1ON2O11O1WNmCb1T<\\NnCc1Y<O1O1O2N1N2O1O1O1O1O1O2N001O1O001O001O00O1N2N2M3O1000000000000000000000000000SO[C>f<AZC?f<AZC?g<@YC`0g<@ZC?g<@YC`0h<^OYCb0i<\\OWCd0k<YOVCg0l<VOUCj0R=O000000000000000000001O0001O000000000001O00002N3M4IT]k3"}, "label": "white suv parked near field"}]}
{"id": 448824, "image": "COCO_train2014_000000448824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a parked white car\"."}], "task": "refering", "answer_template": "The \"a parked white car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219], "bbox": [0.0799375, 0.40939866369710465, 0.5595156250000001, 0.5934298440979955], "iscrowd": 0, "area": 18452.570099999997, "rle": {"size": [449, 640], "counts": "PPg04l=2N2M3N2N2M3N2N1N3N1O2O001O001O001N101O001O0000001O001O1N2O1N2N101O1O1O1_CeNZ<b1O2N0000001O001N101N2O1O010O1O1O02OO1O001O1N2O5K6J1O000O10000000000O100000000000000000000000000000000000000000000000000000000001O00000000000000000000000O1O1O1O100000000O100000000001O0000000000001O001O2N0000000000000000000000000000000000000000000O1000000000000000000002N1O1O00000000000000001O000000000000000000000000001O0000000000000000001O000000001O001O001N2O001O001O002N2N2N2N1ON2O11O1WNmCb1T<\\NnCc1Y<O1O1O2N1N2O1O1O1O1O1O2N001O1O001O001O00O1N2N2M3O1000000000000000000000000000SO[C>f<AZC?f<AZC?g<@YC`0g<@ZC?g<@YC`0h<^OYCb0i<\\OWCd0k<YOVCg0l<VOUCj0R=O000000000000000000001O0001O000000000001O00002N3M4IT]k3"}, "label": "a parked white car"}]}
{"id": 473403, "image": "COCO_train2014_000000473403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a navy blue rolling suitcase holding a black bag\"."}], "task": "refering", "answer_template": "The \"a navy blue rolling suitcase holding a black bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 119, 120, 121, 122, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 193], "bbox": [0.61314, 0.44050666666666666, 0.78772, 0.7829866666666667], "iscrowd": 0, "area": 9419.420650000002, "rle": {"size": [375, 500], "counts": "Rf`3;Q;<C<D<D<E;D=C<E;00O1O2O0O1O100O2N100O1O101N1O100O1010O00000001O0001O01O000000010O000000001O01O000001O0001O0001O000000010O0000000001O0O1O1O100O1O2N100O1O1O101N1O1O3Nh0WOh0XOi0WOh0YOhaV1"}, "label": "a navy blue rolling suitcase holding a black bag"}]}
{"id": 473403, "image": "COCO_train2014_000000473403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large grey luggage with black bag on top\"."}], "task": "refering", "answer_template": "The \"large grey luggage with black bag on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 119, 120, 121, 122, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 193], "bbox": [0.61314, 0.44050666666666666, 0.78772, 0.7829866666666667], "iscrowd": 0, "area": 9419.420650000002, "rle": {"size": [375, 500], "counts": "Rf`3;Q;<C<D<D<E;D=C<E;00O1O2O0O1O100O2N100O1O101N1O100O1010O00000001O0001O01O000000010O000000001O01O000001O0001O0001O000000010O0000000001O0O1O1O100O1O2N100O1O1O101N1O1O3Nh0WOh0XOi0WOh0YOhaV1"}, "label": "large grey luggage with black bag on top"}]}
{"id": 473403, "image": "COCO_train2014_000000473403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black duffle bag lying on the floor\"."}], "task": "refering", "answer_template": "The \"a black duffle bag lying on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 188, 189, 190, 191, 192, 205, 206, 207, 208, 209, 224, 225, 226, 227], "bbox": [0.41778, 0.73032, 0.6854199999999999, 0.9892], "iscrowd": 0, "area": 8175.30565, "rle": {"size": [375, 500], "counts": "`k\\23c;4K5K5K5L2M2N3N1N3M2O2M2N2N3N1N3M2O2M2N3N1N2N3M201O001O000O2O001O001O001O00001N101O001N1O2N1O1O2N1O2N1O2N1O2N10010O010O0010O010O001N101O001N10000O2O00001N100O101N100O2N100O101N100O2O0O1O101N100O2O0O101N1O100O101N100O101N1O100O2O0O100O2gNXFa0h9_OX100O01N2O0O12N2N1O2N2N1O2N^bi1"}, "label": "a black duffle bag lying on the floor"}]}
{"id": 473403, "image": "COCO_train2014_000000473403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black small bag on the floor by itself\"."}], "task": "refering", "answer_template": "The \"black small bag on the floor by itself\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 188, 189, 190, 191, 192, 205, 206, 207, 208, 209, 224, 225, 226, 227], "bbox": [0.41778, 0.73032, 0.6854199999999999, 0.9892], "iscrowd": 0, "area": 8175.30565, "rle": {"size": [375, 500], "counts": "`k\\23c;4K5K5K5L2M2N3N1N3M2O2M2N2N3N1N3M2O2M2N3N1N2N3M201O001O000O2O001O001O001O00001N101O001N1O2N1O1O2N1O2N1O2N1O2N10010O010O0010O010O001N101O001N10000O2O00001N100O101N100O2N100O101N100O2O0O1O101N100O2O0O101N1O100O101N100O101N1O100O2O0O100O2gNXFa0h9_OX100O01N2O0O12N2N1O2N2N1O2N^bi1"}, "label": "black small bag on the floor by itself"}]}
{"id": 473403, "image": "COCO_train2014_000000473403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black duffelbag sitting on a brown tile floor\"."}], "task": "refering", "answer_template": "The \"a black duffelbag sitting on a brown tile floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 188, 189, 190, 191, 192, 205, 206, 207, 208, 209, 224, 225, 226, 227], "bbox": [0.41546, 0.72784, 0.68644, 0.99336], "iscrowd": 0, "area": 8534.4885, "rle": {"size": [375, 500], "counts": "R`\\23`;4L5K4O2M2N3N2M2N3M2O2M2N3N1N3M2O2M3M2O2M2N3M2O2M2N3N1O2O1O0O2O001O001O0O2O001O001O1N101O001O001N101N1O2M3N1O2N1O2O00000000O100O10001N2O0O2O001N2O0O2O001N2O0O2O001N2O0O2O000O100O101O0O100O101N10000O100O2O0O10000O2O0O100O101O0O100O101N1gNWFd0i9[O^F?c9_OdF:]9EkF4U9LQGMP92XGGh88[1001O00000101N3M2N3NfVi1"}, "label": "a black duffelbag sitting on a brown tile floor"}]}
{"id": 473403, "image": "COCO_train2014_000000473403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black carry bag laying in front of two women ' s luggage\"."}], "task": "refering", "answer_template": "The \"a black carry bag laying in front of two women ' s luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 188, 189, 190, 191, 192, 205, 206, 207, 208, 209, 224, 225, 226, 227], "bbox": [0.41546, 0.72784, 0.68644, 0.99336], "iscrowd": 0, "area": 8534.4885, "rle": {"size": [375, 500], "counts": "R`\\23`;4L5K4O2M2N3N2M2N3M2O2M2N3N1N3M2O2M3M2O2M2N3M2O2M2N3N1O2O1O0O2O001O001O0O2O001O001O1N101O001O001N101N1O2M3N1O2N1O2O00000000O100O10001N2O0O2O001N2O0O2O001N2O0O2O001N2O0O2O000O100O101O0O100O101N10000O100O2O0O10000O2O0O100O101O0O100O101N1gNWFd0i9[O^F?c9_OdF:]9EkF4U9LQGMP92XGGh88[1001O00000101N3M2N3NfVi1"}, "label": "a black carry bag laying in front of two women ' s luggage"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with green shirt on\"."}], "task": "refering", "answer_template": "The \"woman with green shirt on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 373, 374, 376], "bbox": [0.117984375, 0.15437499999999998, 0.37753125, 0.9858333333333335], "iscrowd": 0, "area": 37629.764849999985, "rle": {"size": [480, 640], "counts": "RiS1k0n=;D;J7J5K6J5K6J5L4M4K4L4L3M3M3M2N3N1N2nDRMo9Q3iEXMS:n3L5K4L4L3L5L3^Ob0J7K4K5L4L5K4L4L5K4N2N2N2M3N2N2N2M4M2N3M3M2M4M2N3M2M4M3M2N3L5L5L4K5K5L4K5K5K4M4K5K4M4K4L4M4K4M4K4M3M4K4MO2O1N2N2O1N2O1N3M2O1N1O2N2N2M3N2N3M3M4K3hLTLXKm3^4dLZK]3\\4SM]Ko2\\4]M]Kd2c4aMVKa2j4aMSKa2l4`MRKa2m4bMPK_2P5cMmJ^2S5cMjJ`2U5bMhJ_2W5dMfJ]2\\5cMaJ_2a5`MYJe2h5[MQJk2P6UMiIR3W6nLcIW3]6jL\\I\\3e6a2O010O3N200000O2N2N2N01L3N1N3oIeGV5^8eJnGR5Q8kJ[Hl4e7QK^Ho4b7nJaHS5^7jJfHm0nNT3e9iLaFU3_9gLfFW3Z9fLkFX3V9dLoF[3R9`LSG`3o8YLVGj3o93N1ON2N1N3N2N2M2O2N6I6K2M3N1N3N4Lf0YOf0[Of0YOf0[O^Sj5"}, "label": "woman with green shirt on"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green sweater digging in her purse for change\"."}], "task": "refering", "answer_template": "The \"a woman in a green sweater digging in her purse for change\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 373, 374, 376], "bbox": [0.117984375, 0.15437499999999998, 0.37753125, 0.9858333333333335], "iscrowd": 0, "area": 37629.764849999985, "rle": {"size": [480, 640], "counts": "RiS1k0n=;D;J7J5K6J5K6J5L4M4K4L4L3M3M3M2N3N1N2nDRMo9Q3iEXMS:n3L5K4L4L3L5L3^Ob0J7K4K5L4L5K4L4L5K4N2N2N2M3N2N2N2M4M2N3M3M2M4M2N3M2M4M3M2N3L5L5L4K5K5L4K5K5K4M4K5K4M4K4L4M4K4M4K4M3M4K4MO2O1N2N2O1N2O1N3M2O1N1O2N2N2M3N2N3M3M4K3hLTLXKm3^4dLZK]3\\4SM]Ko2\\4]M]Kd2c4aMVKa2j4aMSKa2l4`MRKa2m4bMPK_2P5cMmJ^2S5cMjJ`2U5bMhJ_2W5dMfJ]2\\5cMaJ_2a5`MYJe2h5[MQJk2P6UMiIR3W6nLcIW3]6jL\\I\\3e6a2O010O3N200000O2N2N2N01L3N1N3oIeGV5^8eJnGR5Q8kJ[Hl4e7QK^Ho4b7nJaHS5^7jJfHm0nNT3e9iLaFU3_9gLfFW3Z9fLkFX3V9dLoF[3R9`LSG`3o8YLVGj3o93N1ON2N1N3N2N2M2O2N6I6K2M3N1N3N4Lf0YOf0[Of0YOf0[O^Sj5"}, "label": "a woman in a green sweater digging in her purse for change"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large black leather purse\"."}], "task": "refering", "answer_template": "The \"a large black leather purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 242, 243, 265, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.54775, 0.5528125, 0.7483125, 0.8719166666666667], "iscrowd": 0, "area": 10650.221099999999, "rle": {"size": [480, 640], "counts": "fjT51n>2N2N2N2N2N2N2N2O1NFdA2[>OfA0Y>0hA1V>OlA0S>1nANQ>2QBNm=2UBMn07V;MmCKh0l0k:YO_DKa0_1`:fNQEJ:l1]:ZN[EI3R2a:VN]EHLW2f:QN`EGF]2i:kMcEG_Oc2m:gMeE_2Z:aMgE_2X:aMjE^2U:bMlE^2S:bMnE^2Q:bMPF^2o9cMQF^2m9bMTF^2k9bMVF^2i9bMWF_2h9aMYF_2f9aM[F_2d9aM]F_2c9aM]F_2b9aM_F_2`9aM`Fa2^9_McFa2\\9_MeFa2Z9`MfF`2Y9`MhF`2X9_MiFa2W9^MiFc2V9^MjFb2V9]MkFd2T9[MmFe2S9ZMnFf2Q9[MoFe2Q9ZMoFg2Q9XMPGh2o8XMRGh2n8XMQGi2o8VMRGj2n8UMSGl2k8UMUGk2k8VMSGk2m8UMSGk2m8VMRGj2n8VMQGk2o8UMQGk2P9UMoFk2Q9UMnFl2S9TMlFm2S9SMmFl2U9TMiFm2W9SMiFm2X9RMhFn2X9SMgFm2Z9RMeFn2\\9RMdFn2]9QMcFo2]9PMcFQ3^9nLbFR3^9nLbFQ3`9mL`FT3a9kL_FU3a9jL`FV3a9iL^FX3b9hL^FW3f:0O10000O10O10O10000O01000O12N1N3N1O1O2N1N3N1O2N1O2N1N3N2N3M3M3L5L3M3M3UNkBe1]=J6K5K5K5K5J6K5K5K5K\\R[2"}, "label": "a large black leather purse"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big black purse the woman in the purple vest is holding\"."}], "task": "refering", "answer_template": "The \"the big black purse the woman in the purple vest is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 242, 243, 265, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.54775, 0.5528125, 0.7483125, 0.8719166666666667], "iscrowd": 0, "area": 10650.221099999999, "rle": {"size": [480, 640], "counts": "fjT51n>2N2N2N2N2N2N2N2O1NFdA2[>OfA0Y>0hA1V>OlA0S>1nANQ>2QBNm=2UBMn07V;MmCKh0l0k:YO_DKa0_1`:fNQEJ:l1]:ZN[EI3R2a:VN]EHLW2f:QN`EGF]2i:kMcEG_Oc2m:gMeE_2Z:aMgE_2X:aMjE^2U:bMlE^2S:bMnE^2Q:bMPF^2o9cMQF^2m9bMTF^2k9bMVF^2i9bMWF_2h9aMYF_2f9aM[F_2d9aM]F_2c9aM]F_2b9aM_F_2`9aM`Fa2^9_McFa2\\9_MeFa2Z9`MfF`2Y9`MhF`2X9_MiFa2W9^MiFc2V9^MjFb2V9]MkFd2T9[MmFe2S9ZMnFf2Q9[MoFe2Q9ZMoFg2Q9XMPGh2o8XMRGh2n8XMQGi2o8VMRGj2n8UMSGl2k8UMUGk2k8VMSGk2m8UMSGk2m8VMRGj2n8VMQGk2o8UMQGk2P9UMoFk2Q9UMnFl2S9TMlFm2S9SMmFl2U9TMiFm2W9SMiFm2X9RMhFn2X9SMgFm2Z9RMeFn2\\9RMdFn2]9QMcFo2]9PMcFQ3^9nLbFR3^9nLbFQ3`9mL`FT3a9kL_FU3a9jL`FV3a9iL^FX3b9hL^FW3f:0O10000O10O10O10000O01000O12N1N3N1O1O2N1N3N1O2N1O2N1N3N2N3M3M3L5L3M3M3UNkBe1]=J6K5K5K5K5J6K5K5K5K\\R[2"}, "label": "the big black purse the woman in the purple vest is holding"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with long sleeves and a maroon top looking in her purse\"."}], "task": "refering", "answer_template": "The \"a woman with long sleeves and a maroon top looking in her purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 266, 267, 269, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 382, 383, 384, 385, 386], "bbox": [0.5937812499999999, 0.1226875, 0.911796875, 0.9920833333333333], "iscrowd": 0, "area": 51355.20499999999, "rle": {"size": [480, 640], "counts": "g]b51o>1N2N3N1N2N2O1UFFm4<oJGP5;lJHT58jJJU57iJKV57gJKY55eJMZ54dJN[53cJO]51bJ0]52`J0_51_J1`50^J2a5O]J3c5M[J5d5LZJ6e5JZJ8e5HZJ:e5FZJ<e5DZJ>f5AYJa0f5_OYJc0f5^OXJd0g5\\OXJf0f5[OYJg0]5BbJ`0U5GkJ;3fNT1P1hN<1kNQ1j0mN=OPOn0c0SO?KVOl0<XO`0I[Oi06]Oa0H_Oe01Bb0GCa0KJb0DF>INb0CI:F4a0AM7B9a0_O02lMVMc1[3a0\\O4N^Of0?[O6J]Ol0=YO9GZOR1<WO<DYOU1<VO=AYOZ1:TO>@\\OZ16UO`0^O_OZ11VOc0\\OBZ1LYOd0ZOEZ1G[Of0XOE]1EZOh0UOFa1BYOj0SOFc1AYOj0ROGe1_OXOl0oNHi1\\OXOm0lNIk1[OXOn0iNJo1XOWOP1gNJR2VOWOQ1dNKU2TOVOR1aNNX2QOWOR1\\N1]2mNVOZ1SNLg2jNVOT5i0mJVOT5j0lJVOT5j0lJUOU5j0lJVOT5j0lJUOV5j0jJVOV5j0jJUOW5j0jJVOV5j0jJUOX5j0hJVOZ5g0gJXO\\5f0dJZO^5d0bJ\\O`5a0aJ_O`5`0`J@b5>^JBd5<\\JDf59[JFg59YJGi57WJIk54VJLk53UJMm51SJOo5NRJ2P6LPJ4U6GkI9[6@fI`0`6ZO`If0e6UO[Ik0k6oNUIQ1P7iNQIW1\\7\\NdHd1\\:0000001O1O001N2O001O1O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1\\Od0YOg0L4N2O1N1O2N2N2N1O2N2N2N1O2N2N1O2N1O2N2N1O2N2N3L4M3M3M3M3L4M3M3M3M3L4M3M2N2N3L3L4L5K4M3L5K6K5J6PN[H\\Lk7a3fHlK_7Q4j1N2N2N2M3N2N1O2N2N2L3M4K5L4K5XOl0lN]1cN`Rj0"}, "label": "a woman with long sleeves and a maroon top looking in her purse"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the long - sleeved white shirt\"."}], "task": "refering", "answer_template": "The \"the woman in the long - sleeved white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 266, 267, 269, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 382, 383, 384, 385, 386], "bbox": [0.5937812499999999, 0.1226875, 0.911796875, 0.9920833333333333], "iscrowd": 0, "area": 51355.20499999999, "rle": {"size": [480, 640], "counts": "g]b51o>1N2N3N1N2N2O1UFFm4<oJGP5;lJHT58jJJU57iJKV57gJKY55eJMZ54dJN[53cJO]51bJ0]52`J0_51_J1`50^J2a5O]J3c5M[J5d5LZJ6e5JZJ8e5HZJ:e5FZJ<e5DZJ>f5AYJa0f5_OYJc0f5^OXJd0g5\\OXJf0f5[OYJg0]5BbJ`0U5GkJ;3fNT1P1hN<1kNQ1j0mN=OPOn0c0SO?KVOl0<XO`0I[Oi06]Oa0H_Oe01Bb0GCa0KJb0DF>INb0CI:F4a0AM7B9a0_O02lMVMc1[3a0\\O4N^Of0?[O6J]Ol0=YO9GZOR1<WO<DYOU1<VO=AYOZ1:TO>@\\OZ16UO`0^O_OZ11VOc0\\OBZ1LYOd0ZOEZ1G[Of0XOE]1EZOh0UOFa1BYOj0SOFc1AYOj0ROGe1_OXOl0oNHi1\\OXOm0lNIk1[OXOn0iNJo1XOWOP1gNJR2VOWOQ1dNKU2TOVOR1aNNX2QOWOR1\\N1]2mNVOZ1SNLg2jNVOT5i0mJVOT5j0lJVOT5j0lJUOU5j0lJVOT5j0lJUOV5j0jJVOV5j0jJUOW5j0jJVOV5j0jJUOX5j0hJVOZ5g0gJXO\\5f0dJZO^5d0bJ\\O`5a0aJ_O`5`0`J@b5>^JBd5<\\JDf59[JFg59YJGi57WJIk54VJLk53UJMm51SJOo5NRJ2P6LPJ4U6GkI9[6@fI`0`6ZO`If0e6UO[Ik0k6oNUIQ1P7iNQIW1\\7\\NdHd1\\:0000001O1O001N2O001O1O001O1O001O1O001O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1\\Od0YOg0L4N2O1N1O2N2N2N1O2N2N2N1O2N2N1O2N1O2N2N1O2N2N3L4M3M3M3M3L4M3M3M3M3L4M3M2N2N3L3L4L5K4M3L5K6K5J6PN[H\\Lk7a3fHlK_7Q4j1N2N2N2M3N2N1O2N2N2L3M4K5L4K5XOl0lN]1cN`Rj0"}, "label": "the woman in the long - sleeved white shirt"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a light purple top with short sleeves\"."}], "task": "refering", "answer_template": "The \"a woman in a light purple top with short sleeves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.42303125, 0.199625, 0.638765625, 0.9883958333333334], "iscrowd": 0, "area": 32339.731099999964, "rle": {"size": [480, 640], "counts": "dWo32m>2N3M2N2gA2\\=1\\Bf0S=k0H8H7H2O1O1O1N2WG\\NU4d1dKfNY4[1`KnN^4R1[KXOa4i0RJf0k5[OeIW1Y6iNZIg1b6YNTIT2i6T3Nl0iMUG`Ll8]3YG`Lh8^3\\G`Ld8]3aG`L`8]3eG`L\\8]3iGaLW8\\3nGaLS8[3SHbLn7Z3XHcLi7[3\\HaLe7^3aH\\L`7b3fHZLZ7d3R2O1N2O1N2N2O1N2O1N2O1O10000O100O10oM_MbGa2[8cNeF]1W9kNeFU1V9UOeFk0W9a2L4L41O6J6J5K3M3M2N3M2O2dK]FV3e9bLoFo2e:MO1O001O00001O001O00001O1O2N1O1O1O1O001O2bMXL`Ij3[6[LcIg3W6_LgId3]ORLP5?aKa3\\ORLn4c0dK^3ZOQLm4h0fKY3YORLn4i0gKX3WOQLP5j0gKW3VOPLe4Y1SLi2TOQLZ3CgLS2h0Z2UORLV3GjLR2h0W2TOSLX2CeM75Q2f0S2VOTLY2H[M4>R2e0P2UOUL\\2b0gM^1e0l1UOWL^2`0^MP2e0[1[OWLb2>TM^2k0n0\\OYLd2Z4lM_O\\OZLg2j9VMYFi2P;O1O1O001O1O2N1O2N1O2M2O2N1N2O3M3L4M3M3L5L3L4K6J6I6I8G9HZQ\\3"}, "label": "a woman in a light purple top with short sleeves"}]}
{"id": 55618, "image": "COCO_train2014_000000055618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a purple shirt feeding a parking meter\"."}], "task": "refering", "answer_template": "The \"a woman in a purple shirt feeding a parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.42303125, 0.199625, 0.638765625, 0.9883958333333334], "iscrowd": 0, "area": 32339.731099999964, "rle": {"size": [480, 640], "counts": "dWo32m>2N3M2N2gA2\\=1\\Bf0S=k0H8H7H2O1O1O1N2WG\\NU4d1dKfNY4[1`KnN^4R1[KXOa4i0RJf0k5[OeIW1Y6iNZIg1b6YNTIT2i6T3Nl0iMUG`Ll8]3YG`Lh8^3\\G`Ld8]3aG`L`8]3eG`L\\8]3iGaLW8\\3nGaLS8[3SHbLn7Z3XHcLi7[3\\HaLe7^3aH\\L`7b3fHZLZ7d3R2O1N2O1N2N2O1N2O1N2O1O10000O100O10oM_MbGa2[8cNeF]1W9kNeFU1V9UOeFk0W9a2L4L41O6J6J5K3M3M2N3M2O2dK]FV3e9bLoFo2e:MO1O001O00001O001O00001O1O2N1O1O1O1O001O2bMXL`Ij3[6[LcIg3W6_LgId3]ORLP5?aKa3\\ORLn4c0dK^3ZOQLm4h0fKY3YORLn4i0gKX3WOQLP5j0gKW3VOPLe4Y1SLi2TOQLZ3CgLS2h0Z2UORLV3GjLR2h0W2TOSLX2CeM75Q2f0S2VOTLY2H[M4>R2e0P2UOUL\\2b0gM^1e0l1UOWL^2`0^MP2e0[1[OWLb2>TM^2k0n0\\OYLd2Z4lM_O\\OZLg2j9VMYFi2P;O1O1O001O1O2N1O2N1O2M2O2N1N2O3M3L4M3M3L5L3L4K6J6I6I8G9HZQ\\3"}, "label": "a woman in a purple shirt feeding a parking meter"}]}
{"id": 391492, "image": "COCO_train2014_000000391492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plateful of greens next to a plate of pasta\"."}], "task": "refering", "answer_template": "The \"a plateful of greens next to a plate of pasta\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 226, 227, 238, 239, 240, 241, 242, 248, 249, 262, 263, 264], "bbox": [0.292515625, 0.1716588785046729, 0.942171875, 0.7449065420560748], "iscrowd": 0, "area": 59021.76000000001, "rle": {"size": [428, 640], "counts": "V\\^27U=;D2N2O1N2N3M2O1N2N2O2M2N2O1N2N3N1N2N2O1N3M2N2O1N2O1N2O001N2O1O1N101O1N2O1O0O2O1O1N2O1O0O2O1N2O1N1O2O1N2O8G2O1N2N3N1N2O1N3M2O1N2O2M100O1O100O101N100O100O100O100O100O100O100O1lNlKPIT4h6ULVIl3c6\\L\\Id3^6cL`I^3`6cL_I]3a6dL]I]3c6dL\\I]3c6dL[I]3f6cLYI]3g6dLWI]3i6dLVI\\3j6eLTI\\3l6eLSI[3m6fLQI[3o6eLQI[3o6fLoH[3Q7fLnHZ3R7gLlHY3U7hLjHX3V7iLhHX3X7^10000000000000000000O1000O1O1O100O1O1O100O1O1O010O1O00010O00010O010O01O010O1O010N101O001O001O1N101O0010O02O0O10001N10001N10000O2O0O10001N10001N10001N10001N101N10001N101O0O101O0O2O000O2O0O101OO001O10O01O10O01O10O01O010O1O1O100O1O100O1O101N1O10000O100000000O10O1O1N2N101N2N2O1N2O0O2N2O1N2N101N2O1N2N2O0O2N2O1N2O0O2N2O1N2N101N2O1N2N2O0XOi00000000O100000000O10001O001O0O2O001O1O0O2O001O0O2O001O001N2O00000O10O100O10O0100000O100000001N2O001O1N2ZGSLk7m3SHVLl7k3QHYLm7h3oG\\Lo7e3oG^LP8c3mG_LS8b3jGaLU8`3hGbLX8^3fGdLY8Y4O0bIoJa4Q5\\KRKd4o4YKTKe4m4XKVKi4i4UKYKk4h4RKZKo4e4nJ^KS5a4kJaKV5_4fJcK[5]4cJdK_5[4^JgKd5X4ZJiKg5W4VJkKl5T4RJmKP6R4nIoKS6Q4jIQLX6n3fIRL]6m3`IULa6l3\\IULf6j3WIXLk6g3SIZLn6f3PI[LR7d3kH^LW7j42N1O2N2N2N2N2N2N2N2M3N2N1O2N2N2N2N2N2N2N2N2N2N2N2M3L4K5K5K4M4K5K5K5L3L5K5K5L4K4L5K5J6J5M4M3M3L3N3L4M2N3L4M2M4G8FQ[?"}, "label": "a plateful of greens next to a plate of pasta"}]}
{"id": 391492, "image": "COCO_train2014_000000391492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"collard greens\"."}], "task": "refering", "answer_template": "The \"collard greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 226, 227, 238, 239, 240, 241, 242, 248, 249, 262, 263, 264], "bbox": [0.292515625, 0.1716588785046729, 0.942171875, 0.7449065420560748], "iscrowd": 0, "area": 59021.76000000001, "rle": {"size": [428, 640], "counts": "V\\^27U=;D2N2O1N2N3M2O1N2N2O2M2N2O1N2N3N1N2N2O1N3M2N2O1N2O1N2O001N2O1O1N101O1N2O1O0O2O1O1N2O1O0O2O1N2O1N1O2O1N2O8G2O1N2N3N1N2O1N3M2O1N2O2M100O1O100O101N100O100O100O100O100O100O100O1lNlKPIT4h6ULVIl3c6\\L\\Id3^6cL`I^3`6cL_I]3a6dL]I]3c6dL\\I]3c6dL[I]3f6cLYI]3g6dLWI]3i6dLVI\\3j6eLTI\\3l6eLSI[3m6fLQI[3o6eLQI[3o6fLoH[3Q7fLnHZ3R7gLlHY3U7hLjHX3V7iLhHX3X7^10000000000000000000O1000O1O1O100O1O1O100O1O1O010O1O00010O00010O010O01O010O1O010N101O001O001O1N101O0010O02O0O10001N10001N10000O2O0O10001N10001N10001N10001N101N10001N101O0O101O0O2O000O2O0O101OO001O10O01O10O01O10O01O010O1O1O100O1O100O1O101N1O10000O100000000O10O1O1N2N101N2N2O1N2O0O2N2O1N2N101N2O1N2N2O0O2N2O1N2O0O2N2O1N2N101N2O1N2N2O0XOi00000000O100000000O10001O001O0O2O001O1O0O2O001O0O2O001O001N2O00000O10O100O10O0100000O100000001N2O001O1N2ZGSLk7m3SHVLl7k3QHYLm7h3oG\\Lo7e3oG^LP8c3mG_LS8b3jGaLU8`3hGbLX8^3fGdLY8Y4O0bIoJa4Q5\\KRKd4o4YKTKe4m4XKVKi4i4UKYKk4h4RKZKo4e4nJ^KS5a4kJaKV5_4fJcK[5]4cJdK_5[4^JgKd5X4ZJiKg5W4VJkKl5T4RJmKP6R4nIoKS6Q4jIQLX6n3fIRL]6m3`IULa6l3\\IULf6j3WIXLk6g3SIZLn6f3PI[LR7d3kH^LW7j42N1O2N2N2N2N2N2N2N2M3N2N1O2N2N2N2N2N2N2N2N2N2N2N2M3L4K5K5K4M4K5K5K5L3L5K5K5L4K4L5K5J6J5M4M3M3L3N3L4M2N3L4M2M4G8FQ[?"}, "label": "collard greens"}]}
{"id": 391492, "image": "COCO_train2014_000000391492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccoli at the very bottom at the plate\"."}], "task": "refering", "answer_template": "The \"the broccoli at the very bottom at the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 311, 312], "bbox": [0.50825, 0.6416355140186916, 0.7191093749999999, 0.8839953271028038], "iscrowd": 0, "area": 9748.12355, "rle": {"size": [428, 640], "counts": "cUX42W=6I7J7H7J6J6I7J3L5L3M3O2N100O1O2O0O1O101N100O2N100O1O2O0O1O101N1O101N1O100O2N100O101N1O101N1000000O10O100O01000O10O0100O10O10O10O010000O010O1000O0100O10O010000O10001N101N2O001N101N101O0O2O0O2O001N101N101O0O2O001N101N101O0O2O0O2O1O0O2O0O2O001N101N1O2M2M4M2N3K4K6I6K6I6KWnZ2"}, "label": "the broccoli at the very bottom at the plate"}]}
{"id": 481609, "image": "COCO_train2014_000000481609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red color bucket holding white color mug with the pink color tooth brush\"."}], "task": "refering", "answer_template": "The \"red color bucket holding white color mug with the pink color tooth brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.6, 0.04944, 0.99776, 1.0], "iscrowd": 0, "area": 49273.051750000006, "rle": {"size": [375, 500], "counts": "\\S^32e;3L4M2M4M3L4M3L4M2M4M3L4M3L4M2M4M3L4M3L4M2M3N1N2O1N101N2O1N2O0O2O1N2O1N101N2O1SKQMo0Q3oNQMP1o2oNSMo0o2PORMo0o2oNSMo0o2oNSMP1n2oNTMn0m2QOUMn0l2QOUMm0m2QOUMn0l2POVMn0l2QOUMn0k2QOWMm0k2QOWMn0j2QOWMm0k2QOWMn0i2QOYMm0i2ROYMl0g2SO\\Mj0f2TO]Mj0d2UO_Mg0b2XOaMf0`2XOcMe0^2[OdMc0]2[OfMb0[2]OhMa0Y2^OiM>Z2@iM<Y2DjM7Y2GjM5X2JkM3W2LkM0W2OlMMW21lMKW24kMHW27lMEW29mMBU2>mM^OV2`0mM\\OU2c0nMYOU2f0mMVOV2h0mMTOU2k0nM`Nf2_1[M]Nh2b1YM^Nh2`1YM`Ng2`1YM`Nh2^1YMbNh2]1XMcNh2\\1YMcNi2[1XMeNh2[1XMeNi2Y1XMgNh2X1YMhNh2W1XMiNi2U1XMkNh2T1YMlNh2S1XMlNi2S1XMmNi2Q1XMoNh2Q1YMnNh2P1YMPOh2n0YMROg2n0YMROh2l0YMTOg2k0ZMTOh2k0XMUOh2j0YMVOh2h0YMXOh2g0XMYOh2f0YMZOh2e0XM[Oh2d0YM\\Oh2b0YM]Oh2c0XM]Oh2b0YM^Oh2`0YM@g2`0YM@g2?\\M_Oe2?_M^Oa2b0bM[O^2d0[4O2O00001O0000001O00000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000O10000000000000001O00000000000000000000000000001O0UIja0"}, "label": "red color bucket holding white color mug with the pink color tooth brush"}]}
{"id": 481609, "image": "COCO_train2014_000000481609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large red bucket\"."}], "task": "refering", "answer_template": "The \"a large red bucket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.6, 0.04944, 0.99776, 1.0], "iscrowd": 0, "area": 49273.051750000006, "rle": {"size": [375, 500], "counts": "\\S^32e;3L4M2M4M3L4M3L4M2M4M3L4M3L4M2M4M3L4M3L4M2M3N1N2O1N101N2O1N2O0O2O1N2O1N101N2O1SKQMo0Q3oNQMP1o2oNSMo0o2PORMo0o2oNSMo0o2oNSMP1n2oNTMn0m2QOUMn0l2QOUMm0m2QOUMn0l2POVMn0l2QOUMn0k2QOWMm0k2QOWMn0j2QOWMm0k2QOWMn0i2QOYMm0i2ROYMl0g2SO\\Mj0f2TO]Mj0d2UO_Mg0b2XOaMf0`2XOcMe0^2[OdMc0]2[OfMb0[2]OhMa0Y2^OiM>Z2@iM<Y2DjM7Y2GjM5X2JkM3W2LkM0W2OlMMW21lMKW24kMHW27lMEW29mMBU2>mM^OV2`0mM\\OU2c0nMYOU2f0mMVOV2h0mMTOU2k0nM`Nf2_1[M]Nh2b1YM^Nh2`1YM`Ng2`1YM`Nh2^1YMbNh2]1XMcNh2\\1YMcNi2[1XMeNh2[1XMeNi2Y1XMgNh2X1YMhNh2W1XMiNi2U1XMkNh2T1YMlNh2S1XMlNi2S1XMmNi2Q1XMoNh2Q1YMnNh2P1YMPOh2n0YMROg2n0YMROh2l0YMTOg2k0ZMTOh2k0XMUOh2j0YMVOh2h0YMXOh2g0XMYOh2f0YMZOh2e0XM[Oh2d0YM\\Oh2b0YM]Oh2c0XM]Oh2b0YM^Oh2`0YM@g2`0YM@g2?\\M_Oe2?_M^Oa2b0bM[O^2d0[4O2O00001O0000001O00000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000O10000000000000001O00000000000000000000000000001O0UIja0"}, "label": "a large red bucket"}]}
{"id": 481609, "image": "COCO_train2014_000000481609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bucket containing dirty water\"."}], "task": "refering", "answer_template": "The \"a white bucket containing dirty water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220], "bbox": [0.0, 0.05808, 0.43522000000000005, 1.0], "iscrowd": 0, "area": 58815.94934999999, "rle": {"size": [375, 500], "counts": "n0i:n00O100O100O100O100O100000000000000O100000000000000O10000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000001O000000001O000000001O000000001N10000O2O000O10001O0O10000O2O000O10001N101O0O101O0O2N101N100O2N101N100O2O0O2N101N100O2N101N101N100O2N101N2O1N2N101N2O1N2N101N2N2O0O2N2WIaMW4a2eKYNc3i1YL`N`3a1]LfN^3\\1^LkN]3W1^LPO^3R1^LUO]3m0_LYO]3h0`L]O]3e0_L@^3b0`LA]3a0aLA]3a0aLB\\3`0bLC[3?bLE\\3<bLF\\3<bLG[3;cLHZ3:dLIX3:eLJX38fLJX38fLKV38hLKU37iLLS37jLLT3^4M3L3N1N3N2N2M3N2N2M3N2M3N2N2M2O2N2M3N1N3M3M2N3M3M3M2N2N3M2M3N3M2N2M4K4M4KP]W3"}, "label": "a white bucket containing dirty water"}]}
{"id": 481609, "image": "COCO_train2014_000000481609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color bucket\"."}], "task": "refering", "answer_template": "The \"white color bucket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220], "bbox": [0.0, 0.05808, 0.43522000000000005, 1.0], "iscrowd": 0, "area": 58815.94934999999, "rle": {"size": [375, 500], "counts": "n0i:n00O100O100O100O100O100000000000000O100000000000000O10000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000001O000000001O000000001O000000001N10000O2O000O10001O0O10000O2O000O10001N101O0O101O0O2N101N100O2N101N100O2O0O2N101N100O2N101N101N100O2N101N2O1N2N101N2O1N2N101N2N2O0O2N2WIaMW4a2eKYNc3i1YL`N`3a1]LfN^3\\1^LkN]3W1^LPO^3R1^LUO]3m0_LYO]3h0`L]O]3e0_L@^3b0`LA]3a0aLA]3a0aLB\\3`0bLC[3?bLE\\3<bLF\\3<bLG[3;cLHZ3:dLIX3:eLJX38fLJX38fLKV38hLKU37iLLS37jLLT3^4M3L3N1N3N2N2M3N2N2M3N2M3N2N2M2O2N2M3N1N3M3M2N3M3M3M2N2N3M2M3N3M2N2M4K4M4KP]W3"}, "label": "white color bucket"}]}
{"id": 481609, "image": "COCO_train2014_000000481609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear lid resting on two cups and a spoon on it\"."}], "task": "refering", "answer_template": "The \"a clear lid resting on two cups and a spoon on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 42, 43, 44, 45, 46, 47, 48, 59, 60, 61, 62, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 119, 120, 121, 132, 133, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155], "bbox": [0.2949, 0.11597333333333334, 0.7696799999999999, 0.6727733333333333], "iscrowd": 0, "area": 34512.0584, "rle": {"size": [375, 500], "counts": "]ne11d;=D:E5K5K6K4K5K6K4K5K5L5J5K5L3L4L4M2N3N2M3M3M3M2O2M3M3M3N2M2N3M2N3N1N2O1N2O1N3N1N2O1N2O1N2N2O1N2O1N2O1N2O1NeMcI^O\\6b0hI\\OV6e0lIYOS6g0QJVOn5j0SJWOk5j0UJVOj5j0WJVOh5j0YJWOe5j0\\JUOc5k0^JUOa5l0_JUO`5k0`JUO`5k0`JUO_5m0`JTO_5l0aJTO_5l0aJTO^5m0bJTO]5l0cJTO]5m0bJSO]5n0cJSO\\5m0dJSO\\5m0eJROZ5o0fJROY5n0gJROY5o0fJQOY5P1gJQOX5o0hJQOX5o0hJQOW5P1iJQOV5P1iJPOW5P1iJPOW5P1iJQOU5P1kJPOU5P1kJPOU5P1jJQOU5P1kJQOT5o0lJQOT5o0lJQOS5P1mJPOS5P1lJQOT5o0lJROR5o0nJQOR5o0nJQOR5o0nJQOQ5P1nJROQ5n0oJROQ5n0oJROP5o0PKQOP5o0PKQOP5o0oJSOo4n0QKROo4n0QKROo4m0RKSOm4m0TKSOl4m0SKUOm4i0TKWOl4h0UKXOk4g0UKZOk4f0UKZOk4e0VK\\Oi4c0WK^Oi4a0XK_Oh4`0YK@g4`0XKAh4>YKBg4>YKBg4>XKCh4=XKCh4=XKCi4;WKFi4:WKFi4:WKEj4;VKEj4;VKEj4;VKEj4:WKFi4:WKFi4:WKFi4:WKFi4:WKFi4:WKFi4:WKFi4:WKFj49VKFk4:UKFk4:UKFk4:UKFk4:UKFk4:UKEl4;TKEl4;TKDn4;RKEn4;RKEo4:QKEP5<oJDQ5<oJCS5<mJDS5<mJDT5;mJCT5=lJCU5<kJCV5=jJCW5<iJCX5=hJCY5<gJDY5<gJC[5<eJD[5<eJC]5<cJD]5<cJD]5<cJC_5<aJD_5<aJCa5<_JDa5<_JCc5<]JDc5<]JCe5<[JDe5<[JCg5<YJDg5<YJDh5;XJDi5<VJEk5:UJEl5;TJEm5:SJEn5;RJEn5;RJDP6;PJEP6;PJDR6;nIES6:mIEU6:kIFV69jIFX69hIGZ67fII[66eII]66cIJ^65bIJ`65`IKa64_IKc64]ILd63\\ILf63ZIMg62YINi60WIOk60UI0l6OTI0n6ORI1o6NQI1Q7NoH2R7NmH1U7NkH2V7MjH2X7MhH3Z7KfH5[7JeH5]7JcH6^7IbH6`7I`H7a7H_H7e7F[H:g7DYH;j7CUH<P8APH>S8@mG?W8^OiG`0[8^OeGa0^8]ObGa0c8\\O]Gc0f8[OZGd0j8YOVGe0n8YORGf0Q9XOoFf0V9WOjFg0Z9WOfFh0]9VOcFh0a9VO_Fh0e9VO[Fi0h9UOXFg0^:L4M3L7IfnY1"}, "label": "a clear lid resting on two cups and a spoon on it"}]}
{"id": 481609, "image": "COCO_train2014_000000481609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plastic cup with a straw\"."}], "task": "refering", "answer_template": "The \"a plastic cup with a straw\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 42, 43, 44, 45, 46, 47, 48, 59, 60, 61, 62, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 119, 120, 121, 132, 133, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155], "bbox": [0.2949, 0.11597333333333334, 0.7696799999999999, 0.6727733333333333], "iscrowd": 0, "area": 34512.0584, "rle": {"size": [375, 500], "counts": "]ne11d;=D:E5K5K6K4K5K6K4K5K5L5J5K5L3L4L4M2N3N2M3M3M3M2O2M3M3M3N2M2N3M2N3N1N2O1N2O1N3N1N2O1N2O1N2N2O1N2O1N2O1N2O1NeMcI^O\\6b0hI\\OV6e0lIYOS6g0QJVOn5j0SJWOk5j0UJVOj5j0WJVOh5j0YJWOe5j0\\JUOc5k0^JUOa5l0_JUO`5k0`JUO`5k0`JUO_5m0`JTO_5l0aJTO_5l0aJTO^5m0bJTO]5l0cJTO]5m0bJSO]5n0cJSO\\5m0dJSO\\5m0eJROZ5o0fJROY5n0gJROY5o0fJQOY5P1gJQOX5o0hJQOX5o0hJQOW5P1iJQOV5P1iJPOW5P1iJPOW5P1iJQOU5P1kJPOU5P1kJPOU5P1jJQOU5P1kJQOT5o0lJQOT5o0lJQOS5P1mJPOS5P1lJQOT5o0lJROR5o0nJQOR5o0nJQOR5o0nJQOQ5P1nJROQ5n0oJROQ5n0oJROP5o0PKQOP5o0PKQOP5o0oJSOo4n0QKROo4n0QKROo4m0RKSOm4m0TKSOl4m0SKUOm4i0TKWOl4h0UKXOk4g0UKZOk4f0UKZOk4e0VK\\Oi4c0WK^Oi4a0XK_Oh4`0YK@g4`0XKAh4>YKBg4>YKBg4>XKCh4=XKCh4=XKCi4;WKFi4:WKFi4:WKEj4;VKEj4;VKEj4;VKEj4:WKFi4:WKFi4:WKFi4:WKFi4:WKFi4:WKFi4:WKFi4:WKFj49VKFk4:UKFk4:UKFk4:UKFk4:UKFk4:UKEl4;TKEl4;TKDn4;RKEn4;RKEo4:QKEP5<oJDQ5<oJCS5<mJDS5<mJDT5;mJCT5=lJCU5<kJCV5=jJCW5<iJCX5=hJCY5<gJDY5<gJC[5<eJD[5<eJC]5<cJD]5<cJD]5<cJC_5<aJD_5<aJCa5<_JDa5<_JCc5<]JDc5<]JCe5<[JDe5<[JCg5<YJDg5<YJDh5;XJDi5<VJEk5:UJEl5;TJEm5:SJEn5;RJEn5;RJDP6;PJEP6;PJDR6;nIES6:mIEU6:kIFV69jIFX69hIGZ67fII[66eII]66cIJ^65bIJ`65`IKa64_IKc64]ILd63\\ILf63ZIMg62YINi60WIOk60UI0l6OTI0n6ORI1o6NQI1Q7NoH2R7NmH1U7NkH2V7MjH2X7MhH3Z7KfH5[7JeH5]7JcH6^7IbH6`7I`H7a7H_H7e7F[H:g7DYH;j7CUH<P8APH>S8@mG?W8^OiG`0[8^OeGa0^8]ObGa0c8\\O]Gc0f8[OZGd0j8YOVGe0n8YORGf0Q9XOoFf0V9WOjFg0Z9WOfFh0]9VOcFh0a9VO_Fh0e9VO[Fi0h9UOXFg0^:L4M3L7IfnY1"}, "label": "a plastic cup with a straw"}]}
{"id": 80207, "image": "COCO_train2014_000000080207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown vase that looks like a pineapple\"."}], "task": "refering", "answer_template": "The \"brown vase that looks like a pineapple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.45740625, 0.49774999999999997, 0.7331875, 0.9843125], "iscrowd": 0, "area": 29216.942600000006, "rle": {"size": [480, 640], "counts": "diY44f>6J6J6J6J6J6J6J6J6N2M3M3N2M3N2M3N2M3N2M3N2M3N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2M3N2N2K5iNW1K5K5J6K5M3O1O1N2O1O1N200O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000000O1000000000000000000OdK[H_1e7^N`H`1`7]NfH`1Z7\\NlHb1T7[NQIc1o6ZNWIc1i6ZN\\Id1d6XNbIe1_6XNgIe1Y6XNlIf1S6YNQJe1o5XNWJe1i5YN[Je1e5XN`Jf1`5WNfJf1Z5XNjJf1V5WNTKb1l4\\NbKX1^4eNQLo0o3oN\\Lh0c3VOeLe0[3YOlLb0T3[OUM?k2_O\\M<d2AdM:\\2DkM7U2FTN3m1K[NOe1NeNK[13mNFU16l6N2M3Nbm_2"}, "label": "brown vase that looks like a pineapple"}]}
{"id": 80207, "image": "COCO_train2014_000000080207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pineapple - shaped vase filled with wilting pink roses\"."}], "task": "refering", "answer_template": "The \"a pineapple - shaped vase filled with wilting pink roses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 338, 355, 356, 357, 358, 359, 360, 361, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.45740625, 0.49774999999999997, 0.7331875, 0.9843125], "iscrowd": 0, "area": 29216.942600000006, "rle": {"size": [480, 640], "counts": "diY44f>6J6J6J6J6J6J6J6J6N2M3M3N2M3N2M3N2M3N2M3N2M3N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2M3N2N2K5iNW1K5K5J6K5M3O1O1N2O1O1N200O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000000O1000000000000000000OdK[H_1e7^N`H`1`7]NfH`1Z7\\NlHb1T7[NQIc1o6ZNWIc1i6ZN\\Id1d6XNbIe1_6XNgIe1Y6XNlIf1S6YNQJe1o5XNWJe1i5YN[Je1e5XN`Jf1`5WNfJf1Z5XNjJf1V5WNTKb1l4\\NbKX1^4eNQLo0o3oN\\Lh0c3VOeLe0[3YOlLb0T3[OUM?k2_O\\M<d2AdM:\\2DkM7U2FTN3m1K[NOe1NeNK[13mNFU16l6N2M3Nbm_2"}, "label": "a pineapple - shaped vase filled with wilting pink roses"}]}
{"id": 80207, "image": "COCO_train2014_000000080207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a globular design on a lamp post next to a vase filled with flowers\"."}], "task": "refering", "answer_template": "The \"a globular design on a lamp post next to a vase filled with flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.1735833333333333, 0.46753125000000006, 0.9839791666666666], "iscrowd": 0, "area": 92161.74325000001, "rle": {"size": [480, 640], "counts": "n5P1n=Q2oMR2nMQ2PNm1RN2N2N2O1N2J6N2O1O1O1O1O1O1O1O1O1O1nMVFmNk9S1WFiNk9V1YFfNh9Z1ZFcNg9]1[F`Nf9`1\\F]Ne9c1]FZNd9f1^FWNc9i1`FSNa9m1aFPN`9P2bFmM_9S2cFjM^9V2dFgM]9Y2eFdM\\9\\2gF`MZ9`2jF[MW9e2mFVMT9j2PGQMQ9o2`10000000000000000000000O10000000000000000000000000000000000000000000000000000O100000000000000000O100000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000000000000UETMb9l2\\FWMc9i2ZF\\Md9d2ZF`Mc9a2[FbMd9^2ZFfMd9Z2ZFjMd9V2ZFmMe9S2YFPNf9P2XFSNg9m1VFVNj9j1RF[Nm9d1PF`NP:`1lEeNS:[1iEiNW:Q3O1O001O001O001O001O001O001^FWKk8i4PG]Ko8d4jFcKU9U5O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O2N2L4J7J5K5K5K5K6J5K5K5K5K6J5K5L4N2M4L3M3M3M3M3M4L3MU2kM2N1O2N2N1]N[Cj0g<PO_Co0a<kNfCS1U=N2N3M2N2N2N2N2N3M2N2N2N2N2N3N1N2N2N2Nk^o4"}, "label": "a globular design on a lamp post next to a vase filled with flowers"}]}
{"id": 80207, "image": "COCO_train2014_000000080207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom of a vase beside flowers\"."}], "task": "refering", "answer_template": "The \"the bottom of a vase beside flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.1735833333333333, 0.46753125000000006, 0.9839791666666666], "iscrowd": 0, "area": 92161.74325000001, "rle": {"size": [480, 640], "counts": "n5P1n=Q2oMR2nMQ2PNm1RN2N2N2O1N2J6N2O1O1O1O1O1O1O1O1O1O1nMVFmNk9S1WFiNk9V1YFfNh9Z1ZFcNg9]1[F`Nf9`1\\F]Ne9c1]FZNd9f1^FWNc9i1`FSNa9m1aFPN`9P2bFmM_9S2cFjM^9V2dFgM]9Y2eFdM\\9\\2gF`MZ9`2jF[MW9e2mFVMT9j2PGQMQ9o2`10000000000000000000000O10000000000000000000000000000000000000000000000000000O100000000000000000O100000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000000000000000000000000000UETMb9l2\\FWMc9i2ZF\\Md9d2ZF`Mc9a2[FbMd9^2ZFfMd9Z2ZFjMd9V2ZFmMe9S2YFPNf9P2XFSNg9m1VFVNj9j1RF[Nm9d1PF`NP:`1lEeNS:[1iEiNW:Q3O1O001O001O001O001O001O001^FWKk8i4PG]Ko8d4jFcKU9U5O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O2N2L4J7J5K5K5K5K6J5K5K5K5K6J5K5L4N2M4L3M3M3M3M3M4L3MU2kM2N1O2N2N1]N[Cj0g<PO_Co0a<kNfCS1U=N2N3M2N2N2N2N2N3M2N2N2N2N2N3N1N2N2N2Nk^o4"}, "label": "the bottom of a vase beside flowers"}]}
{"id": 342353, "image": "COCO_train2014_000000342353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green shirt cutting vegetables\"."}], "task": "refering", "answer_template": "The \"a woman in a green shirt cutting vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 82, 102, 103, 104, 105, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 330, 331, 332, 333, 353, 354], "bbox": [0.3036875, 0.15458333333333335, 0.651, 0.9085], "iscrowd": 0, "area": 47931.85124999999, "rle": {"size": [480, 640], "counts": "_Vk2f1W=4L3M3M3L5L3M3M3M4L3L4M3M3M4L3L4M3M4L3M3L4M4L3N2000001N10000O2O000aFnKo7S4oGVLj7k3RH`Lf7`3WHjLb7V3\\HSM\\7n2aH\\MX7e2dHfMT7Z2iHPNo6Q2oHVNl6j1QIYNo6g1nH[NS7f1jH\\NU7e1hH^NX7e40O010000O010000O0100\\O]HkIc7U6bHeI^7[6iH_IW7`6oH[IQ7e6e0O00100O1O010O1O10O01O10O01O100O00100O1O010M3L4K4M4K5LF[JaG`5d8bJYG[5k8gJSGT5Q9oJmFl4X9VKfFf4]9]K`F_4e9:4K4M4N101000O10O01000O0100O10O10O100O01000O010O1000O0100O01000O10O0101O0O101N10001N100O2O001N100O2mNeJ_H[5a7fJ^HZ5b7gJ\\H[5c7fJ\\HZ5d7gJZHZ5f7gJYHZ5f7gJYHY5g7hJWHY5i7hJVHY5i7hJUHY5k7hJTHY5k7hJTHX5l7iJRHX5n7iJQHX5n7iJPHX5P8jJnGV5R8kJmGV5m7oJRHR5i7SKWHm4e7XKYHj4f7VKZHj4g7UKYHk4h7TKWHn4e2XJT2j0VKn4a2_JW2c0WKP5\\2eJ[2:YKQ5W2lJ_23YKR5R2RKd2KYKU5m1WKn1iNlLj0YOX5f1\\KQ2jNkLi6n0dJS2iNkLd6k0jJW2iNhL_6j0oJ[2iNfLY6h0UK^2Q6\\MWJ`2l5YM[Jd2h5UM_Jh2d5QMcJk2l8MVEWMa9f2_F\\Mb9a2]FaMd9`2YFaMi9c2oE_MS:f2eE\\M[:i2]EYMe:l2SEUMo:]31O2O1N2N1O2O1N1O2N2O1N1O2N2O0O2N2N2N101N2N1O2O1J6G8G:G9G8E<D<D<E[XX3"}, "label": "a woman in a green shirt cutting vegetables"}]}
{"id": 342353, "image": "COCO_train2014_000000342353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in green shirt cutting cucumbers\"."}], "task": "refering", "answer_template": "The \"woman in green shirt cutting cucumbers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 82, 102, 103, 104, 105, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 330, 331, 332, 333, 353, 354], "bbox": [0.3036875, 0.15458333333333335, 0.651, 0.9085], "iscrowd": 0, "area": 47931.85124999999, "rle": {"size": [480, 640], "counts": "_Vk2f1W=4L3M3M3L5L3M3M3M4L3L4M3M3M4L3L4M3M4L3M3L4M4L3N2000001N10000O2O000aFnKo7S4oGVLj7k3RH`Lf7`3WHjLb7V3\\HSM\\7n2aH\\MX7e2dHfMT7Z2iHPNo6Q2oHVNl6j1QIYNo6g1nH[NS7f1jH\\NU7e1hH^NX7e40O010000O010000O0100\\O]HkIc7U6bHeI^7[6iH_IW7`6oH[IQ7e6e0O00100O1O010O1O10O01O10O01O100O00100O1O010M3L4K4M4K5LF[JaG`5d8bJYG[5k8gJSGT5Q9oJmFl4X9VKfFf4]9]K`F_4e9:4K4M4N101000O10O01000O0100O10O10O100O01000O010O1000O0100O01000O10O0101O0O101N10001N100O2O001N100O2mNeJ_H[5a7fJ^HZ5b7gJ\\H[5c7fJ\\HZ5d7gJZHZ5f7gJYHZ5f7gJYHY5g7hJWHY5i7hJVHY5i7hJUHY5k7hJTHY5k7hJTHX5l7iJRHX5n7iJQHX5n7iJPHX5P8jJnGV5R8kJmGV5m7oJRHR5i7SKWHm4e7XKYHj4f7VKZHj4g7UKYHk4h7TKWHn4e2XJT2j0VKn4a2_JW2c0WKP5\\2eJ[2:YKQ5W2lJ_23YKR5R2RKd2KYKU5m1WKn1iNlLj0YOX5f1\\KQ2jNkLi6n0dJS2iNkLd6k0jJW2iNhL_6j0oJ[2iNfLY6h0UK^2Q6\\MWJ`2l5YM[Jd2h5UM_Jh2d5QMcJk2l8MVEWMa9f2_F\\Mb9a2]FaMd9`2YFaMi9c2oE_MS:f2eE\\M[:i2]EYMe:l2SEUMo:]31O2O1N2N1O2O1N1O2N2O1N1O2N2O0O2N2N2N101N2N1O2O1J6G8G:G9G8E<D<D<E[XX3"}, "label": "woman in green shirt cutting cucumbers"}]}
{"id": 342353, "image": "COCO_train2014_000000342353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white chef coat and white chef hat\"."}], "task": "refering", "answer_template": "The \"a man in a white chef coat and white chef hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0033750000000000004, 0.0060625, 0.293265625, 0.9813333333333334], "iscrowd": 0, "area": 67604.47304999999, "rle": {"size": [480, 640], "counts": "dS1g2Q6dMKh81O0100O01000O100O1000O0100O1O1O1N2O1O1O1O1N2O1O2N1O1N2O1O1O1O1O100O1O1YN]FPNd9g1kFSNV9c1ZGUNg8b1iGVNX8a1WHYNi7^1gHZNZ7]1Y3G9H8M3M3L4O1O1O2O0O1O1O100O1O100O100000001O0000000O100000000000000000000000000000O0100000000000000000000001O000000000000001O00000000000O101O0000000000001O000000000000001O0000000gG@Q2a0mMDP2<nMIo18nMLP24nM1o10nM4P2LdM\\OlJm0]7GfMg0V2[OhMh0V2XOjMk0S2VOkMm0S2SOmMP1P2QOnMS1o1mNVMFPL`1g6lNXMKjK\\1k6iN[MS2c2nM[MV2b2jM^MY2_2hM_M[2_2eMaM^2\\2cM]LTOSN]3]5`M_LZOiM^3e5YM_LBaM\\3m5RM`LP4^3QL_LT4^3lKWLa4g3`KkKP5Q4QKbK_5Z4cJXKl5e4VJnJY6n4hIeJh6X5b1M4L3M4L4M:E;E;E;F:E6J3M4L3N2M4L3M3M4_OZCbNh<X1aCbNc<W1fCcN]<W1j0JV`c6"}, "label": "a man in a white chef coat and white chef hat"}]}
{"id": 342353, "image": "COCO_train2014_000000342353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a manin a chef hat and shirt is watching two women cut vegetables\"."}], "task": "refering", "answer_template": "The \"a manin a chef hat and shirt is watching two women cut vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0033750000000000004, 0.0060625, 0.293265625, 0.9813333333333334], "iscrowd": 0, "area": 67604.47304999999, "rle": {"size": [480, 640], "counts": "dS1g2Q6dMKh81O0100O01000O100O1000O0100O1O1O1N2O1O1O1O1N2O1O2N1O1N2O1O1O1O1O100O1O1YN]FPNd9g1kFSNV9c1ZGUNg8b1iGVNX8a1WHYNi7^1gHZNZ7]1Y3G9H8M3M3L4O1O1O2O0O1O1O100O1O100O100000001O0000000O100000000000000000000000000000O0100000000000000000000001O000000000000001O00000000000O101O0000000000001O000000000000001O0000000gG@Q2a0mMDP2<nMIo18nMLP24nM1o10nM4P2LdM\\OlJm0]7GfMg0V2[OhMh0V2XOjMk0S2VOkMm0S2SOmMP1P2QOnMS1o1mNVMFPL`1g6lNXMKjK\\1k6iN[MS2c2nM[MV2b2jM^MY2_2hM_M[2_2eMaM^2\\2cM]LTOSN]3]5`M_LZOiM^3e5YM_LBaM\\3m5RM`LP4^3QL_LT4^3lKWLa4g3`KkKP5Q4QKbK_5Z4cJXKl5e4VJnJY6n4hIeJh6X5b1M4L3M4L4M:E;E;E;F:E6J3M4L3N2M4L3M3M4_OZCbNh<X1aCbNc<W1fCcN]<W1j0JV`c6"}, "label": "a manin a chef hat and shirt is watching two women cut vegetables"}]}
{"id": 342353, "image": "COCO_train2014_000000342353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a purple shirt and black apron cuttinga cucumber and wearing a chef ' s hat\"."}], "task": "refering", "answer_template": "The \"a lady in a purple shirt and black apron cuttinga cucumber and wearing a chef ' s hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 41, 42, 43, 44, 64, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343], "bbox": [0.746625, 0.0449375, 1.0, 0.8786458333333333], "iscrowd": 0, "area": 44299.44045, "rle": {"size": [480, 640], "counts": "oYP7<\\>b0oAZOl<e1D<D<D;F8M3L4M4K4M3M3L4M4L3M5L6J5K5K2N2N3M2mKoJTNS5j1RKQNP5m1VKnMk4P2ZKkMh4S2]KiMd4U2aKgMa4V2dKfM]4X2hKcMZ4Z2mKaMT4]2PL`Mo3`2TL]Ml3c2XLZMg3f2\\LXMd3g2`LVM_3j2fLcKkMf0_5f3lL_KhMi0\\5g3QM[KeMm0Y5h3WMVKbMQ1V5i3]MQK`MS1S5l3mMSLd1[4gN[K[1b4gN]K\\1_4fN_K]1^4fN`K]1]4dNbK^1\\4cNbK`1\\4aNcK_1]4cNaK]1_4dN`K[1a4fN]K[1b4gN]K]1_4eN_Ka1[4`NdKf1V4[NhKl1R4h4I6J4L1O00001O001O001aLhDT3X;lLoDm2R;QMSEk2m:UMSEk2n:SMREn2n:RMREn2`;00000000000001O1O1O1O1O1O002N1O1O1O1O1O00001O00001O00O10000O1O1N2O1O1L8H9G:F:D<H7J7J6I7I6K6UHgJb5`5TJeJi5c5nIaJP6f5gI^JX6g5`I^J^6g5TIcJk6j6O2N001O0O100O2O0O2N101N101O000O1000000O1000000O101O000O1O1M3UNaGSMa8j2eGRM\\8k2kGPMV8m2PHaL`8[3gGkKV9k3eK"}, "label": "a lady in a purple shirt and black apron cuttinga cucumber and wearing a chef ' s hat"}]}
{"id": 342353, "image": "COCO_train2014_000000342353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chef in purple\"."}], "task": "refering", "answer_template": "The \"chef in purple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 41, 42, 43, 44, 64, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343], "bbox": [0.746625, 0.0449375, 1.0, 0.8786458333333333], "iscrowd": 0, "area": 44299.44045, "rle": {"size": [480, 640], "counts": "oYP7<\\>b0oAZOl<e1D<D<D;F8M3L4M4K4M3M3L4M4L3M5L6J5K5K2N2N3M2mKoJTNS5j1RKQNP5m1VKnMk4P2ZKkMh4S2]KiMd4U2aKgMa4V2dKfM]4X2hKcMZ4Z2mKaMT4]2PL`Mo3`2TL]Ml3c2XLZMg3f2\\LXMd3g2`LVM_3j2fLcKkMf0_5f3lL_KhMi0\\5g3QM[KeMm0Y5h3WMVKbMQ1V5i3]MQK`MS1S5l3mMSLd1[4gN[K[1b4gN]K\\1_4fN_K]1^4fN`K]1]4dNbK^1\\4cNbK`1\\4aNcK_1]4cNaK]1_4dN`K[1a4fN]K[1b4gN]K]1_4eN_Ka1[4`NdKf1V4[NhKl1R4h4I6J4L1O00001O001O001aLhDT3X;lLoDm2R;QMSEk2m:UMSEk2n:SMREn2n:RMREn2`;00000000000001O1O1O1O1O1O002N1O1O1O1O1O00001O00001O00O10000O1O1N2O1O1L8H9G:F:D<H7J7J6I7I6K6UHgJb5`5TJeJi5c5nIaJP6f5gI^JX6g5`I^J^6g5TIcJk6j6O2N001O0O100O2O0O2N101N101O000O1000000O1000000O101O000O1O1M3UNaGSMa8j2eGRM\\8k2kGPMV8m2PHaL`8[3gGkKV9k3eK"}, "label": "chef in purple"}]}
{"id": 334162, "image": "COCO_train2014_000000334162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy riding the skateboard\"."}], "task": "refering", "answer_template": "The \"a boy riding the skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 34, 56, 57, 78, 79, 101, 102, 103, 124, 125, 126, 147, 148, 149, 169, 170, 171, 172, 192, 193, 195, 215], "bbox": [0.361546875, 0.029925373134328356, 0.525125, 0.6524129353233831], "iscrowd": 0, "area": 8010.299399999998, "rle": {"size": [402, 640], "counts": "mmj21`<6I2O10000O1O1O2M2O10000J6J6J5K5K5N3N1OO1K5N2OM40O2O1OSNkNWHOaNX1R9UOUHS1k7TOnGl0n7AQGnNf0`1Y8IdG4_80VGfNGY1V99lFIV9g13OHPMoFP3h8[MVGf2h8c0N2O100O10O1O1N2N2O1O1O1N2N2O1O2O1L4ONNjFlLT9R37O1003M2M3N2N3M2M4M4K4L5JXMhGk1g7nMcHW2T7gMSI[2g6aMaI`2[6\\MlIf2n5XMoHYOV1c3mNbL^6b0iJW3S5jLoJY3l4gLYKX3b4hLhKQ3V4nLWLg2g3XM\\Lh2c3WM`Lh2^3WMeLe0eMc0e5fNkL`0hMg0[5hNQMCZM;f0>VOHh5L`MJnNOa359IYOc0i0]OXO?l0@QO`0S1^OjNd0X1\\OiNa0Z1^OfN`0e601EUDOP<IVD6j;JVD4m;KRD5T<O2O101N_Zg3"}, "label": "a boy riding the skateboard"}]}
{"id": 334162, "image": "COCO_train2014_000000334162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboarder\"."}], "task": "refering", "answer_template": "The \"skateboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 34, 56, 57, 78, 79, 101, 102, 103, 124, 125, 126, 147, 148, 149, 169, 170, 171, 172, 192, 193, 195, 215], "bbox": [0.361546875, 0.029925373134328356, 0.525125, 0.6524129353233831], "iscrowd": 0, "area": 8010.299399999998, "rle": {"size": [402, 640], "counts": "mmj21`<6I2O10000O1O1O2M2O10000J6J6J5K5K5N3N1OO1K5N2OM40O2O1OSNkNWHOaNX1R9UOUHS1k7TOnGl0n7AQGnNf0`1Y8IdG4_80VGfNGY1V99lFIV9g13OHPMoFP3h8[MVGf2h8c0N2O100O10O1O1N2N2O1O1O1N2N2O1O2O1L4ONNjFlLT9R37O1003M2M3N2N3M2M4M4K4L5JXMhGk1g7nMcHW2T7gMSI[2g6aMaI`2[6\\MlIf2n5XMoHYOV1c3mNbL^6b0iJW3S5jLoJY3l4gLYKX3b4hLhKQ3V4nLWLg2g3XM\\Lh2c3WM`Lh2^3WMeLe0eMc0e5fNkL`0hMg0[5hNQMCZM;f0>VOHh5L`MJnNOa359IYOc0i0]OXO?l0@QO`0S1^OjNd0X1\\OiNa0Z1^OfN`0e601EUDOP<IVD6j;JVD4m;KRD5T<O2O101N_Zg3"}, "label": "skateboarder"}]}
{"id": 121172, "image": "COCO_train2014_000000121172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in red pants jumping on skis\"."}], "task": "refering", "answer_template": "The \"a man in red pants jumping on skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 37, 38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226], "bbox": [0.6370625000000001, 0.0522196261682243, 0.891515625, 0.6539485981308413], "iscrowd": 0, "area": 26331.06890000001, "rle": {"size": [428, 640], "counts": "jeZ5:o<>B5K5dLSOQJR1f5ZOSJi0k5]OnIg0o5_OPHEf0P1W7AmGDd0P1]7BhGEb0o0e7b0SHCm7a0jGEU8i201O0000O100O011O0O10001M2N2O2M2N2O2M2N2N3M2M3N3L4K4NQG_LX8^3iGmLm7P3TH\\Mc7a2^HjM^1jNc3Z3oJUNV1eNi3T3QKYNT1dNl3Q3PK]NS1bNm3P3oJbNQ1_NQ4m2nJfNo0^NS4k2mJjNo0[NU4j2kJnNm0YNY4g2jJSOj0WN\\4e2iJYOg0SNa4b2hJ^Oe0PNd4b2dJCd0lMi4a2aJFc0jMm4a2]JJb0eMQ5c2YJO`0_MX5c2UJ3>[M^5o5XJVJi5f61O001O0000000000O01000000000O1000000000O0100000000O100000000001N1000001O000O101O000O101O000O10001O0O101O1O0O2PMYJYNg5f1[JYNf5d1^JZNc5c1aJZN`5e1bJZN_5c1eJ[N\\5b1gJ^NX5a1kJ[NeNUNa6]3nJ\\NU5a1nJ_NR5^1RK`No4^1TK^NWN]Nf6S3UK_NUN^Nh6o2WKbNk4\\1XKbNi4[1\\K`NiMfNm6g2`K_Nb4^1bK`N_4^1cK`N`4\\1cKcN^4[1dKdN]4Z1eKdN^4Y1dKfN]4Y1dKeN^4Z1dKdN^4Z1cKdN_4[1bKdN_4[1bKdN`4Z1aKdNa4\\1_KcNb4\\1_KcNc4[1_KbNc4]1_KaNb4^1aK_Na4_1bK[Nb4d1e3M3N2M2N3M3N2M2N3M3N4K5I6G:GWgl0"}, "label": "a man in red pants jumping on skis"}]}
{"id": 121172, "image": "COCO_train2014_000000121172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in maroon pants is skiing\"."}], "task": "refering", "answer_template": "The \"a man in maroon pants is skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 37, 38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226], "bbox": [0.6370625000000001, 0.0522196261682243, 0.891515625, 0.6539485981308413], "iscrowd": 0, "area": 26331.06890000001, "rle": {"size": [428, 640], "counts": "jeZ5:o<>B5K5dLSOQJR1f5ZOSJi0k5]OnIg0o5_OPHEf0P1W7AmGDd0P1]7BhGEb0o0e7b0SHCm7a0jGEU8i201O0000O100O011O0O10001M2N2O2M2N2O2M2N2N3M2M3N3L4K4NQG_LX8^3iGmLm7P3TH\\Mc7a2^HjM^1jNc3Z3oJUNV1eNi3T3QKYNT1dNl3Q3PK]NS1bNm3P3oJbNQ1_NQ4m2nJfNo0^NS4k2mJjNo0[NU4j2kJnNm0YNY4g2jJSOj0WN\\4e2iJYOg0SNa4b2hJ^Oe0PNd4b2dJCd0lMi4a2aJFc0jMm4a2]JJb0eMQ5c2YJO`0_MX5c2UJ3>[M^5o5XJVJi5f61O001O0000000000O01000000000O1000000000O0100000000O100000000001N1000001O000O101O000O101O000O10001O0O101O1O0O2PMYJYNg5f1[JYNf5d1^JZNc5c1aJZN`5e1bJZN_5c1eJ[N\\5b1gJ^NX5a1kJ[NeNUNa6]3nJ\\NU5a1nJ_NR5^1RK`No4^1TK^NWN]Nf6S3UK_NUN^Nh6o2WKbNk4\\1XKbNi4[1\\K`NiMfNm6g2`K_Nb4^1bK`N_4^1cK`N`4\\1cKcN^4[1dKdN]4Z1eKdN^4Y1dKfN]4Y1dKeN^4Z1dKdN^4Z1cKdN_4[1bKdN_4[1bKdN`4Z1aKdNa4\\1_KcNb4\\1_KcNc4[1_KbNc4]1_KaNb4^1aK_Na4_1bK[Nb4d1e3M3N2M2N3M3N2M2N3M3N4K5I6G:GWgl0"}, "label": "a man in maroon pants is skiing"}]}
{"id": 121172, "image": "COCO_train2014_000000121172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a persn snowboarding\"."}], "task": "refering", "answer_template": "The \"a persn snowboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 80, 81, 102, 103, 104, 124, 125, 126, 127, 148, 149, 150, 171, 172, 173, 174, 194, 195], "bbox": [0.41967187499999997, 0.16873831775700934, 0.578171875, 0.5795794392523365], "iscrowd": 0, "area": 8155.669549999998, "rle": {"size": [428, 640], "counts": "]b`32Z=1O1O1O1N2O1O2N2N1N3N1O2N1O2N2M2O2^EZOZ8g0]GE_8=RG3k8NmF8S9KiF7W9KfF6Z9KcF6^9L^F6b9K[F7d9LXF5i9MTF4l9MQF5o9c101O010O1O010O1O0102N1N2O1N3N1O1N10N2N1O2N2N1O2N2N1O2N1O2H8G8H9M3L3N3L4M2M4N2001O0O1000000O2O000jL\\Hd0d7XNbIf1^6mMQJP2h8N2N2N2N2M3N2N2N2N2M3N2N2N3M4K6K5K4L5J6K5KZd`3"}, "label": "a persn snowboarding"}]}
{"id": 121172, "image": "COCO_train2014_000000121172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ski boarder in a pink striped top and black pants\"."}], "task": "refering", "answer_template": "The \"a ski boarder in a pink striped top and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 80, 81, 102, 103, 104, 124, 125, 126, 127, 148, 149, 150, 171, 172, 173, 174, 194, 195], "bbox": [0.41967187499999997, 0.16873831775700934, 0.578171875, 0.5795794392523365], "iscrowd": 0, "area": 8155.669549999998, "rle": {"size": [428, 640], "counts": "]b`32Z=1O1O1O1N2O1O2N2N1N3N1O2N1O2N2M2O2^EZOZ8g0]GE_8=RG3k8NmF8S9KiF7W9KfF6Z9KcF6^9L^F6b9K[F7d9LXF5i9MTF4l9MQF5o9c101O010O1O010O1O0102N1N2O1N3N1O1N10N2N1O2N2N1O2N2N1O2N1O2H8G8H9M3L3N3L4M2M4N2001O0O1000000O2O000jL\\Hd0d7XNbIf1^6mMQJP2h8N2N2N2N2M3N2N2N2N2M3N2N2N3M4K6K5K4L5J6K5KZd`3"}, "label": "a ski boarder in a pink striped top and black pants"}]}
{"id": 14676, "image": "COCO_train2014_000000014676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person ' s leg underneath the table below the pizza\"."}], "task": "refering", "answer_template": "The \"the person ' s leg underneath the table below the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6348906249999999, 0.6983891213389121, 1.0, 1.0], "iscrowd": 0, "area": 25994.562600000005, "rle": {"size": [478, 640], "counts": "R]n53j>3N2M2N3M3M3M2O2M2N3M2N3M2O2M2N3M2N3M3N1N3M1O2N1O2O0O2N2N1O2N101N1O2N2N1O2O0O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O100O1O100O1O100N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O10000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O00001O00001O0000001O00001O00001O0000000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000000000000000"}, "label": "the person ' s leg underneath the table below the pizza"}]}
{"id": 14676, "image": "COCO_train2014_000000014676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girls legs with lack shorts on underneath the table\"."}], "task": "refering", "answer_template": "The \"a girls legs with lack shorts on underneath the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6348906249999999, 0.6983891213389121, 1.0, 1.0], "iscrowd": 0, "area": 25994.562600000005, "rle": {"size": [478, 640], "counts": "R]n53j>3N2M2N3M3M3M2O2M2N3M2N3M2O2M2N3M2N3M3N1N3M1O2N1O2O0O2N2N1O2N101N1O2N2N1O2O0O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O100O1O100O1O100N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O10000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O00001O00001O0000001O00001O00001O0000000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O100000000000000000000"}, "label": "a girls legs with lack shorts on underneath the table"}]}
{"id": 14676, "image": "COCO_train2014_000000014676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden black colored chair with a black padded cushion to the upper right\"."}], "task": "refering", "answer_template": "The \"a wooden black colored chair with a black padded cushion to the upper right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 39, 40, 41, 42, 43, 44, 62, 63, 64, 65, 66, 67, 88, 89, 90], "bbox": [0.7286406249999999, 0.006150627615062762, 0.9602656249999999, 0.246255230125523], "iscrowd": 0, "area": 10016.611850000001, "rle": {"size": [478, 640], "counts": "Rbi61l>`0`ABb=U1XBiNX=c1L000O2O0O100OO2I6101N100O100O101N100O100O2O000O100O101N100O100O100O100O100O100O100O2O0O1000000O101OO02O0000001O00001O0O101O000001O000001O000001O01O00001O01O0001O012M9G4M1N000010O0001O01L3AVDgMm;Y2<0L3M4K4N30O100O2O00000000000000001O0O1000000O100O2d0@2MgNZDROi;f0^DYOd;a0aD]Oc;>`D@a;>aD_Oa;?bD_O_;`0cD]O_;b0bD\\O`;b0cD[O_;d0eDUO`;i0[1O1O1O1O2M2H9M2M3N3L3N`S<"}, "label": "a wooden black colored chair with a black padded cushion to the upper right"}]}
{"id": 14676, "image": "COCO_train2014_000000014676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with a topping of olives and a garnish\"."}], "task": "refering", "answer_template": "The \"a pizza with a topping of olives and a garnish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266], "bbox": [0.25007812500000004, 0.26081589958158996, 0.8207187499999999, 0.6630753138075314], "iscrowd": 0, "area": 48377.208799999986, "rle": {"size": [478, 640], "counts": "ilZ2;Y>b0H8H5J2O1N2O1O1N2O1O0O2O1O1N2O1N2O1O0O2O1O1N2O1N2N1N3N2N2N2N2M3N1O2N2N2M3N3M2N2N3L3N3M2N2N3L3N2N2N1O1N3O00001N10001O0O101O000O2O00000O2O00001N10001O0O101O000O1000001N1000000O1000001N1000000O1000001N1000000O101O000O2O00001N10001O0O101O000O2O00001N10001O0O101OO100000000000000000000000000000O10000000000000000000000O1O1O1O1N2O1O1O1N200O1O100O1O100O100O1O100O1O100001O001O1O001O1O1O001O1O000000000001O000000000001O00000O10000O10001N10000O10000O101O0O10000TLbFW2_9hMbFX2_9fMcFY2]9fMeFY2[9fMgFY2Y9fMiFY2W9fMkFX2V9gMlFX2T9gMnFX2S9fMoFY2Q9fMPGZ2P9eMQGZ2P9eMPG\\2P9bMQG_2o8_MRGb2o8ZMTGf2l8XMUGi2k8UMVGk2k8RMWGo2i8oLYGQ3g8mLZGT3f8iL\\GX3e8eL\\G[3e8cL]G]3c8aL^G`3c8\\L_Ge3l91O1O001O1O1O1O1O001O1O1O1O1O1O10O01O1O1O1O1O1O0011O0000001N100O101N1O100O100O100O100O100O100O010O1O100O100O100O10O0100O101N101N2O0O2O1N101N2N101N2B^D^Mb;a2_D^Mc;`2^D`Mc;^2^DaMd;\\2^DdMb;[2_DdMc;Z2_DeMb;Y2_DfMc;Y2]DgMc;X2^DgMd;W2]DiMd;U2^DiMd;U2]DkMc;T2^DkMd;S2]DmMd;R2\\DmMf;Q2d0O2N2N2N101N2N2dNiBn0X=QOjBm0X=POkBn0W=POkBn0W=oNlBo0c=N2N2N2N2N2N2N2N4L7Hn^e1"}, "label": "a pizza with a topping of olives and a garnish"}]}
{"id": 14676, "image": "COCO_train2014_000000014676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plate of pizza with spinach leaves on pizza\"."}], "task": "refering", "answer_template": "The \"plate of pizza with spinach leaves on pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266], "bbox": [0.25007812500000004, 0.26081589958158996, 0.8207187499999999, 0.6630753138075314], "iscrowd": 0, "area": 48377.208799999986, "rle": {"size": [478, 640], "counts": "ilZ2;Y>b0H8H5J2O1N2O1O1N2O1O0O2O1O1N2O1N2O1O0O2O1O1N2O1N2N1N3N2N2N2N2M3N1O2N2N2M3N3M2N2N3L3N3M2N2N3L3N2N2N1O1N3O00001N10001O0O101O000O2O00000O2O00001N10001O0O101O000O1000001N1000000O1000001N1000000O1000001N1000000O101O000O2O00001N10001O0O101O000O2O00001N10001O0O101OO100000000000000000000000000000O10000000000000000000000O1O1O1O1N2O1O1O1N200O1O100O1O100O100O1O100O1O100001O001O1O001O1O1O001O1O000000000001O000000000001O00000O10000O10001N10000O10000O101O0O10000TLbFW2_9hMbFX2_9fMcFY2]9fMeFY2[9fMgFY2Y9fMiFY2W9fMkFX2V9gMlFX2T9gMnFX2S9fMoFY2Q9fMPGZ2P9eMQGZ2P9eMPG\\2P9bMQG_2o8_MRGb2o8ZMTGf2l8XMUGi2k8UMVGk2k8RMWGo2i8oLYGQ3g8mLZGT3f8iL\\GX3e8eL\\G[3e8cL]G]3c8aL^G`3c8\\L_Ge3l91O1O001O1O1O1O1O001O1O1O1O1O1O10O01O1O1O1O1O1O0011O0000001N100O101N1O100O100O100O100O100O100O010O1O100O100O100O10O0100O101N101N2O0O2O1N101N2N101N2B^D^Mb;a2_D^Mc;`2^D`Mc;^2^DaMd;\\2^DdMb;[2_DdMc;Z2_DeMb;Y2_DfMc;Y2]DgMc;X2^DgMd;W2]DiMd;U2^DiMd;U2]DkMc;T2^DkMd;S2]DmMd;R2\\DmMf;Q2d0O2N2N2N101N2N2dNiBn0X=QOjBm0X=POkBn0W=POkBn0W=oNlBo0c=N2N2N2N2N2N2N2N4L7Hn^e1"}, "label": "plate of pizza with spinach leaves on pizza"}]}
{"id": 170327, "image": "COCO_train2014_000000170327.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra laying down\"."}], "task": "refering", "answer_template": "The \"zebra laying down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 143, 144, 150, 151, 164, 166, 167, 168, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267], "bbox": [0.12442187499999999, 0.34539906103286383, 0.6698593749999999, 0.7743661971830986], "iscrowd": 0, "area": 29682.3648, "rle": {"size": [426, 640], "counts": "[`Q13V=3M3L4M2M2O1O2M2O2N1N201O0O2O1O2N2O1N2O0O2N2O1N2N2O1N3jDcN[:^1^EfNc:\\1UEiNl:k10O1O1N3N1O1O1O1O1O1N3N00001O00001N10001O001O01O010O1O100O1O1RM^NSJb1k5aNSJ`1j5cNVJ\\1i5fNWJZ1g5iNWJX1h5kNVJT1j5nNUJR1i5QOVJn0k5SOSJn0l5SOTJl0m5UORJk0n5UORJj0o5VOQJj0o5VOQJi0o5XOQJg0P6YOPJg0P6]OlIb0V6H_I8a6L\\I2e60[INf64XIJi69UIFl6:SIEn6;RIEP78PIHS76mHJU73lHLW72iHNZ7NfH2^7KaH6b7F^H:d7E[H<g7BXH>j7@VHa0l7]OSHc0n7]OQHc0P8]OnGe0R8ZOlGh0U8XOhGj0Y8VOdGl0]8SObGo0_8PO_GQ1b8oN[GS1f8mNXGU1h8kNVGV1k8jNSGW1n8iNPGX1Q9hNlF[1S9fNkF[1V9eNhF\\1X9eNgF[1Z9eNdF]1\\9cNbF^1^9cNaF]1`9cN^F_1a9bN]F_1d9aN[F_1f9aNXF`1h9aNWF`1i9P1001O001O1O001O1O001O1N101O1O1O001O1O001O1O001O1O001O1O1O001O1O001O00001O0O2O00000000000000000000000000000000000000O10O10O1000000O10000O10000O1000000O1000000000000000000001O00000000000000001O000O10000O10001N10000O101O0O10001N10000O2O000000010O000000000000001O0000O100O101N1[OdM\\F\\2^9lM`FT2[9TNaFm1]9WNaFi1]9\\N_Fe1_9_N_Fa1_9cN_F]1_9X1M3N2N2O1O1O1O100000O10000000000O101O1O1O2N2N2M4M2N1O1O1O1O1O1O1O1O1O001O1O1O1O001O00001O001O00001O001O00001O00001O001O0O101O001O000O2O1O001O001O0O2nM]E[1c:eN_EX1c:fN`EW1a:iNaET1a:jNbET1_:kNeEP1\\:oNjEk0X:TOmEe0X:TOPFd0d;F<Dj_g2"}, "label": "zebra laying down"}]}
{"id": 170327, "image": "COCO_train2014_000000170327.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on it ' s back\"."}], "task": "refering", "answer_template": "The \"the zebra on it ' s back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 143, 144, 150, 151, 164, 166, 167, 168, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267], "bbox": [0.12442187499999999, 0.34539906103286383, 0.6698593749999999, 0.7743661971830986], "iscrowd": 0, "area": 29682.3648, "rle": {"size": [426, 640], "counts": "[`Q13V=3M3L4M2M2O1O2M2O2N1N201O0O2O1O2N2O1N2O0O2N2O1N2N2O1N3jDcN[:^1^EfNc:\\1UEiNl:k10O1O1N3N1O1O1O1O1O1N3N00001O00001N10001O001O01O010O1O100O1O1RM^NSJb1k5aNSJ`1j5cNVJ\\1i5fNWJZ1g5iNWJX1h5kNVJT1j5nNUJR1i5QOVJn0k5SOSJn0l5SOTJl0m5UORJk0n5UORJj0o5VOQJj0o5VOQJi0o5XOQJg0P6YOPJg0P6]OlIb0V6H_I8a6L\\I2e60[INf64XIJi69UIFl6:SIEn6;RIEP78PIHS76mHJU73lHLW72iHNZ7NfH2^7KaH6b7F^H:d7E[H<g7BXH>j7@VHa0l7]OSHc0n7]OQHc0P8]OnGe0R8ZOlGh0U8XOhGj0Y8VOdGl0]8SObGo0_8PO_GQ1b8oN[GS1f8mNXGU1h8kNVGV1k8jNSGW1n8iNPGX1Q9hNlF[1S9fNkF[1V9eNhF\\1X9eNgF[1Z9eNdF]1\\9cNbF^1^9cNaF]1`9cN^F_1a9bN]F_1d9aN[F_1f9aNXF`1h9aNWF`1i9P1001O001O1O001O1O001O1N101O1O1O001O1O001O1O001O1O001O1O1O001O1O001O00001O0O2O00000000000000000000000000000000000000O10O10O1000000O10000O10000O1000000O1000000000000000000001O00000000000000001O000O10000O10001N10000O101O0O10001N10000O2O000000010O000000000000001O0000O100O101N1[OdM\\F\\2^9lM`FT2[9TNaFm1]9WNaFi1]9\\N_Fe1_9_N_Fa1_9cN_F]1_9X1M3N2N2O1O1O1O100000O10000000000O101O1O1O2N2N2M4M2N1O1O1O1O1O1O1O1O1O001O1O1O1O001O00001O001O00001O001O00001O00001O001O0O101O001O000O2O1O001O001O0O2nM]E[1c:eN_EX1c:fN`EW1a:iNaET1a:jNbET1_:kNeEP1\\:oNjEk0X:TOmEe0X:TOPFd0d;F<Dj_g2"}, "label": "the zebra on it ' s back"}]}
{"id": 170327, "image": "COCO_train2014_000000170327.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on far right\"."}], "task": "refering", "answer_template": "The \"zebra on far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 201, 202, 203, 224, 225, 226, 247, 248], "bbox": [0.72096875, 0.26384976525821596, 0.9297656249999999, 0.7082629107981221], "iscrowd": 0, "area": 14530.731499999998, "rle": {"size": [426, 640], "counts": "kmo51V=<E;D:G7I6J6K6I<D<E;D;F8lFgLh7b3lGeLP8d3cGeLX8X4N10000O10E;WOh0N3VOdFbM`9[2mFXMU9e2j0L3N3N20O2O1O1O1N1]F]M_8d2_GdMZ8]2dG^Na7d1]HSOm6n0QISOP7l0lHWOU7j0dH\\O\\7e0]H@d7R3100O1QO]HWLc7f3RIgKo6V4T1M3M34M:E:F4L2N1OO1L4L4L4L4^N[G[Ni8e1gGdMa8[2W1O10O100O100O100O010N2O1O1O10000TGjLd7X3YHkLe7Z3THiLk7[3PHhLn7]3kGgL2D\\7j3[HfLKMi7b3VHVMj7V4O001O001O\\Od0VOj0H8G8H9G9G9G9N21O1O1O1O001O1O1O1O1O001O1O1O1O4L3NO2N1O2NO100OkMcE_1]:_NgE_1Y:^NmE^1T:`NQF\\1P:bNTF[1n9aNXF5ZOd0c:QOWF1Cg0S<J5K6I7JQab0"}, "label": "zebra on far right"}]}
{"id": 170327, "image": "COCO_train2014_000000170327.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra walking to back side in the right side of the image\"."}], "task": "refering", "answer_template": "The \"zebra walking to back side in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 201, 202, 203, 224, 225, 226, 247, 248], "bbox": [0.72096875, 0.26384976525821596, 0.9297656249999999, 0.7082629107981221], "iscrowd": 0, "area": 14530.731499999998, "rle": {"size": [426, 640], "counts": "kmo51V=<E;D:G7I6J6K6I<D<E;D;F8lFgLh7b3lGeLP8d3cGeLX8X4N10000O10E;WOh0N3VOdFbM`9[2mFXMU9e2j0L3N3N20O2O1O1O1N1]F]M_8d2_GdMZ8]2dG^Na7d1]HSOm6n0QISOP7l0lHWOU7j0dH\\O\\7e0]H@d7R3100O1QO]HWLc7f3RIgKo6V4T1M3M34M:E:F4L2N1OO1L4L4L4L4^N[G[Ni8e1gGdMa8[2W1O10O100O100O100O010N2O1O1O10000TGjLd7X3YHkLe7Z3THiLk7[3PHhLn7]3kGgL2D\\7j3[HfLKMi7b3VHVMj7V4O001O001O\\Od0VOj0H8G8H9G9G9G9N21O1O1O1O001O1O1O1O1O001O1O1O1O4L3NO2N1O2NO100OkMcE_1]:_NgE_1Y:^NmE^1T:`NQF\\1P:bNTF[1n9aNXF5ZOd0c:QOWF1Cg0S<J5K6I7JQab0"}, "label": "zebra walking to back side in the right side of the image"}]}
{"id": 170327, "image": "COCO_train2014_000000170327.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra between two other zebras standing\"."}], "task": "refering", "answer_template": "The \"zebra between two other zebras standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 107, 108, 109, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 199, 200, 222], "bbox": [0.603078125, 0.22392018779342723, 0.76390625, 0.6579342723004694], "iscrowd": 0, "area": 6857.505749999998, "rle": {"size": [426, 640], "counts": "mhP5<V<h0G:M2010O10000O10O0100O100O00100O1O1O100O1O1O1O1O1O001O1OhMoNXHo0h7SOXHl0h7WOVHh0j7\\OSHc0m7AoGa0P8BlG?S8LbG6\\8M`G5_8N^G4`8O]G2b81ZG1e82XGMi86PGLR97cFO_94WF0l9d13M3NdGUM]6h2^IZM]NKU8h2XIhMcN^OU8j2gH[Ni6X2PIoMo6Q2jHRNZ7o1^HTNf7d34L4L4L4L4K5L4L4L3L5L4L4M2M4L4M2M5M2N2fNnEZOT:c0TFVOn9g0ZFROh9k0aFmNa9P1gFjNZ9S1nFfNU9V1a1N3M2N2O2M2O1N3N1N;F00000000000000001O002N2N5Kmfn1"}, "label": "zebra between two other zebras standing"}]}
{"id": 358744, "image": "COCO_train2014_000000358744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby giraffe eating\"."}], "task": "refering", "answer_template": "The \"baby giraffe eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 272, 273, 274, 275, 287, 288, 289, 290, 301, 302, 303, 304, 305, 316, 317, 318, 319, 320, 331, 332, 334, 335], "bbox": [0.07387706855791962, 0.6779843750000001, 0.6024113475177305, 1.0], "iscrowd": 0, "area": 17654.3343, "rle": {"size": [640, 423], "counts": "boc03jc08H7I5K3M3M3M3L4M3M3M3M3M3M3M3L2O000000000O1M3L4L5L3L4L4M3L4L42N2N2:GW1hN5K1O100O1O1O100O1O1O10O0L4L4L5K4L4K5L5K4L4L4L5J5L4L4L5M22N1O100O1O1O2O0O1O1O101N1O1O100O1O2N100O1O1O10k0TOn0ROf0ZO100O1O100O100000000O100000000O10000000L3K6K5J6J6K4K6J6J6J5K6J6J6J6J5K6J6J6J6J6J6J6I6K6J6J6J6L400O1000O10O10000O1000000O10O1000O1000000O10000O0100000O1000000O10O1000O10000O1000000O0100000O1000000O10O10O1000000O1000000O01000O1000000O1O100O010O1O100O2O2M2N2O1N3N1N2N2OSVY3"}, "label": "baby giraffe eating"}]}
{"id": 358744, "image": "COCO_train2014_000000358744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shorter giraffe\"."}], "task": "refering", "answer_template": "The \"shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 272, 273, 274, 275, 287, 288, 289, 290, 301, 302, 303, 304, 305, 316, 317, 318, 319, 320, 331, 332, 334, 335], "bbox": [0.07387706855791962, 0.6779843750000001, 0.6024113475177305, 1.0], "iscrowd": 0, "area": 17654.3343, "rle": {"size": [640, 423], "counts": "boc03jc08H7I5K3M3M3M3L4M3M3M3M3M3M3M3L2O000000000O1M3L4L5L3L4L4M3L4L42N2N2:GW1hN5K1O100O1O1O100O1O1O10O0L4L4L5K4L4K5L5K4L4L4L5J5L4L4L5M22N1O100O1O1O2O0O1O1O101N1O1O100O1O2N100O1O1O10k0TOn0ROf0ZO100O1O100O100000000O100000000O10000000L3K6K5J6J6K4K6J6J6J5K6J6J6J6J5K6J6J6J6J6J6J6I6K6J6J6J6L400O1000O10O10000O1000000O10O1000O1000000O10000O0100000O1000000O10O1000O10000O1000000O0100000O1000000O10O10O1000000O1000000O01000O1000000O1O100O010O1O100O2O2M2N2O1N3N1N2N2OSVY3"}, "label": "shorter giraffe"}]}
{"id": 358744, "image": "COCO_train2014_000000358744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe with neck upright facing left\"."}], "task": "refering", "answer_template": "The \"giraffe with neck upright facing left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 47, 48, 49, 50, 51, 65, 66, 67, 80, 81, 82, 96, 97, 98, 111, 112, 113, 114, 127, 128, 129, 130, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 219, 220, 221, 222, 223, 224, 233, 234, 235, 236, 237, 238, 239, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 325, 326, 340, 341], "bbox": [0.14626477541371158, 0.018328125, 1.0, 1.0], "iscrowd": 0, "area": 58514.481599999985, "rle": {"size": [640, 423], "counts": "UkV11oc01O000G2_\\OO^c05a\\OK]c08a\\OI^c0=O1O100O2N001O002O1N1O1O1O2N2N2N1O1O1O1N2N2O1O001N110O00001O1O1O1O1O1O100O1O1O1M3M3M5M2N1N3O0000000O100000000F:00O1O100O110O0001O3M5K1O1O1O00O1O1N2O1O1O2N2N1K6N2N2N2M4L3N3N2N2NSMj_OS2S`0iMT@V2k?eMZ@]2d`02N1O2N1O2M3N1N3M2N2O1N3MUM`_OX2^`0gMe_OY2Y`0fMi_O[2T`0fMm_O[2P`0fMR@[2k?eMV@\\2g?eM[@[2b?fM_@[2_?eMb@\\2[?eMg@\\2U?dMm@^2P?bMRA_2k>aMWA`2f>`M\\A`2b>`MaA`2\\>`MhA_2U>aMPB\\2n=dMSB]2k=bMWB_2g=aMZB`2d=_M^Bb2`=^MaBc2]=\\MfBd2X=[MjBf2T=ZMnBf2Q=XMRCh2l<WMWCi2g<VM^Ch2`<WMdCh2[<VMjCh2T<WMnCk2P<SMSDm2k;QMYDo2f;oL\\DS3b;kLaDU3V>0O101N100O2O0O100O0010O0010O0010O01O010O100O1O10O010O01O01O01O01O0010O0001O01O010O010O0dBmLj9T3RFPMn9o2PFSMP:n2lEVMS:j2jEYMV:h2gE]MW:c2eEbMY:^2dEhMY:Y2cEmM[:S2`ESN^:n1^EXN_:i1\\E^Na:c1[EcNc:\\1YEkNd:V1XEPOe:Q1SEZOi:i0mDFl:?gD0Q;g4I100O2N1O001O1O001O001O1O1O002N2O0O2N1hGkFa5W9YJ^GW5c8cJaG^5`8]JcGd5^8WJeGj5\\8QJgGo5[8lIhGU6Y8eIkG\\6V8ZIRHg6o7nHZHS7g7cHaH]7b7XHfHi7h91O1O1O001O1O1O1O001O001O001O1O001O001O001O1O001O1aMkGQIU8k6UHoHl7m6]HmHd7o6eHkH]7o6lHlHU7P7TIjHm6Q7[IkHf6P7`InHa6l6fIRI[6i6kIUIU6g6QJWIP6d6VJZIk5a6m3lH[Cf6l<VIXCf6S=01O1O00100O001O001O010O001O001O0010O01O0010O01O00010O0001O0000O1O1O1O1O1O1O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OSD"}, "label": "giraffe with neck upright facing left"}]}
{"id": 358744, "image": "COCO_train2014_000000358744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with it ' s tongue sticking out\"."}], "task": "refering", "answer_template": "The \"a giraffe with it ' s tongue sticking out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 47, 48, 49, 50, 51, 65, 66, 67, 80, 81, 82, 96, 97, 98, 111, 112, 113, 114, 127, 128, 129, 130, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 219, 220, 221, 222, 223, 224, 233, 234, 235, 236, 237, 238, 239, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 325, 326, 340, 341], "bbox": [0.14626477541371158, 0.018328125, 1.0, 1.0], "iscrowd": 0, "area": 58514.481599999985, "rle": {"size": [640, 423], "counts": "UkV11oc01O000G2_\\OO^c05a\\OK]c08a\\OI^c0=O1O100O2N001O002O1N1O1O1O2N2N2N1O1O1O1N2N2O1O001N110O00001O1O1O1O1O1O100O1O1O1M3M3M5M2N1N3O0000000O100000000F:00O1O100O110O0001O3M5K1O1O1O00O1O1N2O1O1O2N2N1K6N2N2N2M4L3N3N2N2NSMj_OS2S`0iMT@V2k?eMZ@]2d`02N1O2N1O2M3N1N3M2N2O1N3MUM`_OX2^`0gMe_OY2Y`0fMi_O[2T`0fMm_O[2P`0fMR@[2k?eMV@\\2g?eM[@[2b?fM_@[2_?eMb@\\2[?eMg@\\2U?dMm@^2P?bMRA_2k>aMWA`2f>`M\\A`2b>`MaA`2\\>`MhA_2U>aMPB\\2n=dMSB]2k=bMWB_2g=aMZB`2d=_M^Bb2`=^MaBc2]=\\MfBd2X=[MjBf2T=ZMnBf2Q=XMRCh2l<WMWCi2g<VM^Ch2`<WMdCh2[<VMjCh2T<WMnCk2P<SMSDm2k;QMYDo2f;oL\\DS3b;kLaDU3V>0O101N100O2O0O100O0010O0010O0010O01O010O100O1O10O010O01O01O01O01O0010O0001O01O010O010O0dBmLj9T3RFPMn9o2PFSMP:n2lEVMS:j2jEYMV:h2gE]MW:c2eEbMY:^2dEhMY:Y2cEmM[:S2`ESN^:n1^EXN_:i1\\E^Na:c1[EcNc:\\1YEkNd:V1XEPOe:Q1SEZOi:i0mDFl:?gD0Q;g4I100O2N1O001O1O001O001O1O1O002N2O0O2N1hGkFa5W9YJ^GW5c8cJaG^5`8]JcGd5^8WJeGj5\\8QJgGo5[8lIhGU6Y8eIkG\\6V8ZIRHg6o7nHZHS7g7cHaH]7b7XHfHi7h91O1O1O001O1O1O1O001O001O001O1O001O001O001O1O001O1aMkGQIU8k6UHoHl7m6]HmHd7o6eHkH]7o6lHlHU7P7TIjHm6Q7[IkHf6P7`InHa6l6fIRI[6i6kIUIU6g6QJWIP6d6VJZIk5a6m3lH[Cf6l<VIXCf6S=01O1O00100O001O001O010O001O001O0010O01O0010O01O00010O0001O0000O1O1O1O1O1O1O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OSD"}, "label": "a giraffe with it ' s tongue sticking out"}]}
{"id": 514391, "image": "COCO_train2014_000000514391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow banana above the other fruit\"."}], "task": "refering", "answer_template": "The \"the yellow banana above the other fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285], "bbox": [0.019754901960784315, 0.024281045751633985, 0.9802450980392158, 0.5823039215686274], "iscrowd": 0, "area": 197350.56795, "rle": {"size": [612, 612], "counts": "T`7`0da0P1POP1POP1POP1POP1POP1POP1POP1POP1POP1L400000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000[Oe0]Id[7"}, "label": "the yellow banana above the other fruit"}]}
{"id": 514391, "image": "COCO_train2014_000000514391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a bruised yellow banana\"."}], "task": "refering", "answer_template": "The \"this is a bruised yellow banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285], "bbox": [0.019754901960784315, 0.024281045751633985, 0.9802450980392158, 0.5823039215686274], "iscrowd": 0, "area": 197350.56795, "rle": {"size": [612, 612], "counts": "T`7`0da0P1POP1POP1POP1POP1POP1POP1POP1POP1POP1L400000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000[Oe0]Id[7"}, "label": "this is a bruised yellow banana"}]}
{"id": 514391, "image": "COCO_train2014_000000514391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small orange\"."}], "task": "refering", "answer_template": "The \"the small orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [293, 294, 295, 296, 297, 298, 299, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 403, 404, 405, 406, 407, 408, 409, 410, 425, 426, 427, 428, 429, 430, 431, 432, 447, 448, 449, 450, 451, 452, 453, 454, 469, 470, 471, 472, 473, 474, 475], "bbox": [0.3391013071895425, 0.6126307189542484, 0.6410947712418301, 0.9715359477124184], "iscrowd": 0, "area": 39217.0214, "rle": {"size": [612, 612], "counts": "Tkl3]1V`0a1^Nb1_Na1WOi0000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000001O00000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000001O00000000000007IX2hMX2hMnnR4"}, "label": "the small orange"}]}
{"id": 514391, "image": "COCO_train2014_000000514391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shiny orange\"."}], "task": "refering", "answer_template": "The \"a shiny orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [293, 294, 295, 296, 297, 298, 299, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 403, 404, 405, 406, 407, 408, 409, 410, 425, 426, 427, 428, 429, 430, 431, 432, 447, 448, 449, 450, 451, 452, 453, 454, 469, 470, 471, 472, 473, 474, 475], "bbox": [0.3391013071895425, 0.6126307189542484, 0.6410947712418301, 0.9715359477124184], "iscrowd": 0, "area": 39217.0214, "rle": {"size": [612, 612], "counts": "Tkl3]1V`0a1^Nb1_Na1WOi0000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000001O00000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000001O00000000000007IX2hMX2hMnnR4"}, "label": "a shiny orange"}]}
{"id": 514391, "image": "COCO_train2014_000000514391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a strawberry among other fruit\"."}], "task": "refering", "answer_template": "The \"a strawberry among other fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 344, 345, 346, 347, 348, 349, 350, 351, 366, 367, 368, 369, 370, 371, 372, 373, 388, 389, 390, 391, 392, 393, 394, 395, 410, 411, 412, 413, 414, 415, 416, 417, 432, 433, 434, 435, 436, 437, 438, 439, 454, 455, 456, 457, 458, 459, 460, 461, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.6568954248366012, 0.6045424836601307, 0.9855392156862745, 0.986388888888889], "iscrowd": 0, "area": 46239.42739999999, "rle": {"size": [612, 612], "counts": "\\d`7]2^`0X4PL4K2O2N2N2N2N101O1O1O1O1O00000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000N2N2N4M3L3M4L4SOl0iN[Q5"}, "label": "a strawberry among other fruit"}]}
{"id": 514391, "image": "COCO_train2014_000000514391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the strawberry fruit is kept on the table\"."}], "task": "refering", "answer_template": "The \"the strawberry fruit is kept on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 344, 345, 346, 347, 348, 349, 350, 351, 366, 367, 368, 369, 370, 371, 372, 373, 388, 389, 390, 391, 392, 393, 394, 395, 410, 411, 412, 413, 414, 415, 416, 417, 432, 433, 434, 435, 436, 437, 438, 439, 454, 455, 456, 457, 458, 459, 460, 461, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.6568954248366012, 0.6045424836601307, 0.9855392156862745, 0.986388888888889], "iscrowd": 0, "area": 46239.42739999999, "rle": {"size": [612, 612], "counts": "\\d`7]2^`0X4PL4K2O2N2N2N2N101O1O1O1O1O00000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000N2N2N4M3L3M4L4SOl0iN[Q5"}, "label": "the strawberry fruit is kept on the table"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"popcorn in a silver bowl\"."}], "task": "refering", "answer_template": "The \"popcorn in a silver bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 283], "bbox": [0.54996, 0.46223849372384934, 0.9624400000000001, 0.8869456066945606], "iscrowd": 0, "area": 33415.175749999995, "rle": {"size": [478, 500], "counts": "\\fP4a0l<h1C8H8J5N3M3M2N3M3M3M2N3M3M2N3M2N3M2N2N3M2N2N3M2O2M2N2N3N1N1O1O2O0O1O2N100O2O0O1O101N100O2O0O101N100O2O0O101N10001N10000O2O000O2O001N10001O0O2O000O2O001N10001N10000O10001N10000O1000001O00000000000000000000000000000000000000000O1000O100000O10000000O10000000001N100O100O100O2O1N2O1N101N2O1N2O1N2O001N2O0O2O1O0O2O1N101N2O0O2N2O0O2N2O0O2N2O1N2N1O2N2N2N101N2N2N2N1O3M2O2L3M3M4L3N3L4L3M4L4M3L3M4L4L4K4M4L4L6J6H8F;F9F[b8"}, "label": "popcorn in a silver bowl"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a popcorn in a plate and its near a pizza in a bowl\"."}], "task": "refering", "answer_template": "The \"a popcorn in a plate and its near a pizza in a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 283], "bbox": [0.54996, 0.46223849372384934, 0.9624400000000001, 0.8869456066945606], "iscrowd": 0, "area": 33415.175749999995, "rle": {"size": [478, 500], "counts": "\\fP4a0l<h1C8H8J5N3M3M2N3M3M3M2N3M3M2N3M2N3M2N2N3M2N2N3M2O2M2N2N3N1N1O1O2O0O1O2N100O2O0O1O101N100O2O0O101N100O2O0O101N10001N10000O2O000O2O001N10001O0O2O000O2O001N10001N10000O10001N10000O1000001O00000000000000000000000000000000000000000O1000O100000O10000000O10000000001N100O100O100O2O1N2O1N101N2O1N2O1N2O001N2O0O2O1O0O2O1N101N2O0O2N2O0O2N2O0O2N2O1N2N1O2N2N2N101N2N2N2N1O3M2O2L3M3M4L3N3L4L3M4L4M3L3M4L4L4K4M4L4L6J6H8F;F9F[b8"}, "label": "a popcorn in a plate and its near a pizza in a bowl"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with a test tube in it\"."}], "task": "refering", "answer_template": "The \"the bowl with a test tube in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 22, 23, 37, 38, 40, 41, 42, 54, 55, 56, 57, 59, 60, 61, 72, 73, 74, 75, 76, 78, 79, 90, 91, 92, 93, 94, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 165], "bbox": [0.0128, 0.08089958158995816, 0.43974, 0.535397489539749], "iscrowd": 0, "area": 26211.7829, "rle": {"size": [478, 500], "counts": "Pn22i>?A>B=B6K6J6I6K6J5J7J6J5L5K4M4K4L4L3M3M2N3M3M201N2N1O2N2N1O2N2O1N1O2N2N1O2N1O2O0O2N1O2N2O0O2N101N1O2O0O2N101N1O2O0O2O0O1O2O01O2O1N2N2O1N101N2O1O1N2O1N2O1O0O2O1N2O1O1O1O1O010O1O1O1N2N2O0O2N2N2N1O2N2N2N2N1O2M3M3L4M2O2M3N2M3N1N3N2M3NcLnLPKo2Q5SMnJl2R5VMnJg2S5[MmJc2S5^MnJ_2T5bMlJ\\2T5fMkJX2V5jMjJT2V5nMjJP2V5RNjJk1W5VNjJh1W5YNhJBgLk1a8eNhJ]OkLj1^8kNRKR1n4POQKo0o4RORKk0P5VOPKh0P5ZOoJe0Q5]OoJ`0R5BnJ<R5FmJ9T5GmJ6T5LkJ3U5OhJ1Y51eJO[53cJM^53`JN`54^JLb56[JJf58XJHi59TJHm59QJGo5:PJFQ6;lIFU6;iIDX6>eIC\\6>bIB^6?aIA`6`0]IAd6`0ZI@f6b0WI_Oj6b0TI^Ol6d0QI\\OQ7d0nH\\OS7e0jH\\OV7f0gH[OZ7f0cH[O]7g0`HZOa7f0]H[Od7f0YH[Og7g0VHYOl7h0QHYOo7i0nGXOS8i0jGXOW8h0gGYOY8i0dGXO]8i0`GYO_8i0^GXOd8h0ZGXOh8h0UGYOl8f0TGYOo8e0PG\\OR9a0nF@S9?mF@V9?hFA[9>cFA`9`0^F_Oe9`0YF@j9`0SF@o9`0nEAU:>hEC[:=`EEb:Z23M3L3N3M3M3M3M3M4L3M3L4M2N3L4L3L5K5K4M4Kj0VO>C7H9G8IUaR4"}, "label": "the bowl with a test tube in it"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a container of salad\"."}], "task": "refering", "answer_template": "The \"a container of salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 22, 23, 37, 38, 40, 41, 42, 54, 55, 56, 57, 59, 60, 61, 72, 73, 74, 75, 76, 78, 79, 90, 91, 92, 93, 94, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 165], "bbox": [0.0128, 0.08089958158995816, 0.43974, 0.535397489539749], "iscrowd": 0, "area": 26211.7829, "rle": {"size": [478, 500], "counts": "Pn22i>?A>B=B6K6J6I6K6J5J7J6J5L5K4M4K4L4L3M3M2N3M3M201N2N1O2N2N1O2N2O1N1O2N2N1O2N1O2O0O2N1O2N2O0O2N101N1O2O0O2N101N1O2O0O2O0O1O2O01O2O1N2N2O1N101N2O1O1N2O1N2O1O0O2O1N2O1O1O1O1O010O1O1O1N2N2O0O2N2N2N1O2N2N2N2N1O2M3M3L4M2O2M3N2M3N1N3N2M3NcLnLPKo2Q5SMnJl2R5VMnJg2S5[MmJc2S5^MnJ_2T5bMlJ\\2T5fMkJX2V5jMjJT2V5nMjJP2V5RNjJk1W5VNjJh1W5YNhJBgLk1a8eNhJ]OkLj1^8kNRKR1n4POQKo0o4RORKk0P5VOPKh0P5ZOoJe0Q5]OoJ`0R5BnJ<R5FmJ9T5GmJ6T5LkJ3U5OhJ1Y51eJO[53cJM^53`JN`54^JLb56[JJf58XJHi59TJHm59QJGo5:PJFQ6;lIFU6;iIDX6>eIC\\6>bIB^6?aIA`6`0]IAd6`0ZI@f6b0WI_Oj6b0TI^Ol6d0QI\\OQ7d0nH\\OS7e0jH\\OV7f0gH[OZ7f0cH[O]7g0`HZOa7f0]H[Od7f0YH[Og7g0VHYOl7h0QHYOo7i0nGXOS8i0jGXOW8h0gGYOY8i0dGXO]8i0`GYO_8i0^GXOd8h0ZGXOh8h0UGYOl8f0TGYOo8e0PG\\OR9a0nF@S9?mF@V9?hFA[9>cFA`9`0^F_Oe9`0YF@j9`0SF@o9`0nEAU:>hEC[:=`EEb:Z23M3L3N3M3M3M3M3M4L3M3L4M2N3L4L3L5K5K4M4Kj0VO>C7H9G8IUaR4"}, "label": "a container of salad"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of pizza under a big piece of pizza in a vessel\"."}], "task": "refering", "answer_template": "The \"a piece of pizza under a big piece of pizza in a vessel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 205, 206, 207, 223, 224, 256, 257, 258, 275, 276, 277], "bbox": [0.2449, 0.6067364016736402, 0.5113, 0.9325941422594142], "iscrowd": 0, "area": 3817.959399999998, "rle": {"size": [478, 500], "counts": "Qji12m>0O1O2N1O101N1O1O1O100O1O1O1O010O1O1O1O100O001O1O100O1O1O00100O1O1O00010O001O00010O00001O010O0000O101O000000001N1O1O1O1O2N1O1O1O1O2O0OnKEcI:_6G_I8b6J\\I5e6MZI1h60VIOk63SILn66PIIQ79mHFU7;jHCW7>hHAY7a0eH_O\\7b0_32N2O1N2N2N2O1N1O2O1N2N2N2O1N2N2O0O2N2O1N2N2N2O1N2N2O0O2N10M2O1N2O2N1N2O2N1N2O1O2M2O1N3N1O1N3N1O1N2H9H7I7IZca3"}, "label": "a piece of pizza under a big piece of pizza in a vessel"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pizza underneath another slice\"."}], "task": "refering", "answer_template": "The \"a slice of pizza underneath another slice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 205, 206, 207, 223, 224, 256, 257, 258, 275, 276, 277], "bbox": [0.2449, 0.6067364016736402, 0.5113, 0.9325941422594142], "iscrowd": 0, "area": 3817.959399999998, "rle": {"size": [478, 500], "counts": "Qji12m>0O1O2N1O101N1O1O1O100O1O1O1O010O1O1O1O100O001O1O100O1O1O00100O1O1O00010O001O00010O00001O010O0000O101O000000001N1O1O1O1O2N1O1O1O1O2O0OnKEcI:_6G_I8b6J\\I5e6MZI1h60VIOk63SILn66PIIQ79mHFU7;jHCW7>hHAY7a0eH_O\\7b0_32N2O1N2N2N2O1N1O2O1N2N2N2O1N2N2O0O2N2O1N2N2N2O1N2N2O0O2N10M2O1N2O2N1N2O2N1N2O1O2M2O1N3N1O1N3N1O1N2H9H7I7IZca3"}, "label": "a slice of pizza underneath another slice"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza on the plate\"."}], "task": "refering", "answer_template": "The \"pizza on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 184, 185, 186, 187, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260], "bbox": [0.17996, 0.5411715481171548, 0.45888, 0.8870502092050209], "iscrowd": 0, "area": 15456.162399999997, "rle": {"size": [478, 500], "counts": "j\\Z12h>4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L5K4L4K5L5K4L4L5J5L4L5K4L4K5L5K4L4O2O01O0010O00010O01O010O000010O01O0010O01O000010O01O0010O0001O0010O01O0010O01O001O010O001O0010O01O001O001O0010O01O001O001O100O2N1O1O2N1O1O101N1O1O2N1O1O2N100O1O2N1O1O2N100O2N1O1O1O2N2N3N2M4L3M3M3M3M3N2M3M3M3F:UOZUn3"}, "label": "pizza on the plate"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza piece kept on the table along with few more snacks\"."}], "task": "refering", "answer_template": "The \"a pizza piece kept on the table along with few more snacks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 184, 185, 186, 187, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260], "bbox": [0.17996, 0.5411715481171548, 0.45888, 0.8870502092050209], "iscrowd": 0, "area": 15456.162399999997, "rle": {"size": [478, 500], "counts": "j\\Z12h>4L4L4L4K5L4L4L4L4K5L4L4L4L4K5L4L5K4L4K5L5K4L4L5J5L4L5K4L4K5L5K4L4O2O01O0010O00010O01O010O000010O01O0010O01O000010O01O0010O0001O0010O01O0010O01O001O010O001O0010O01O001O001O0010O01O001O001O100O2N1O1O2N1O1O101N1O1O2N1O1O2N100O1O2N1O1O2N100O2N1O1O1O2N2N3N2M4L3M3M3M3M3N2M3M3M3F:UOZUn3"}, "label": "a pizza piece kept on the table along with few more snacks"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza in white bowl\"."}], "task": "refering", "answer_template": "The \"pizza in white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 182, 183, 184, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 278], "bbox": [0.12029999999999999, 0.4898953974895397, 0.5370799999999999, 0.9415690376569038], "iscrowd": 0, "area": 34743.92444999999, "rle": {"size": [478, 500], "counts": "lZl03h>7H9G8H8H9G8H8I8G8H9G8H8H9H7L3L2N2N2N1O2N2O0O2N2N2N1O2N2O1N1O2N2N1O2N2N2O0O2N2N1O2N2N2O0O2N2N2N1O2N2O0O2N2N2N1O2N2O1N1O2N2N1O2N2N2O0O2N2O1O001O1N101O0000000O100000000O100000000O1000O1000O10000000000O100000000O100001O00001O0000001O00001O0000001O00001N101O00001O001O00001O001O00001O00001N1O2N1O101N1O2N100O2N1O2O0O1O2N1O101N1O2N100O2N1O2O0O1O2N1O101N3M2N3N1N2N3M2O2M2N3M2N3N1N3M2N2O2M2N3K4L5K4L5L3L5K7I;E;E:F;F:E;E:F;E[_[3"}, "label": "pizza in white bowl"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white container with pizza in it\"."}], "task": "refering", "answer_template": "The \"a white container with pizza in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 182, 183, 184, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 205, 206, 207, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 278], "bbox": [0.12029999999999999, 0.4898953974895397, 0.5370799999999999, 0.9415690376569038], "iscrowd": 0, "area": 34743.92444999999, "rle": {"size": [478, 500], "counts": "lZl03h>7H9G8H8H9G8H8I8G8H9G8H8H9H7L3L2N2N2N1O2N2O0O2N2N2N1O2N2O1N1O2N2N1O2N2N2O0O2N2N1O2N2N2O0O2N2N2N1O2N2O0O2N2N2N1O2N2O1N1O2N2N1O2N2N2O0O2N2O1O001O1N101O0000000O100000000O100000000O1000O1000O10000000000O100000000O100001O00001O0000001O00001O0000001O00001N101O00001O001O00001O001O00001O00001N1O2N1O101N1O2N100O2N1O2O0O1O2N1O101N1O2N100O2N1O2O0O1O2N1O101N3M2N3N1N2N3M2O2M2N3M2N3N1N3M2N2O2M2N3K4L5K4L5L3L5K7I;E;E:F;F:E;E:F;E[_[3"}, "label": "a white container with pizza in it"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gravy in a bowl near popcorn\"."}], "task": "refering", "answer_template": "The \"a gravy in a bowl near popcorn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 44, 45, 46, 47, 48, 49, 50, 61, 62, 63, 64, 65, 66, 67, 68, 69, 79, 80, 81, 82, 83, 84, 85, 86, 87, 97, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 139, 153, 154, 155, 156], "bbox": [0.43316000000000004, 0.07742677824267782, 0.86172, 0.49198744769874475], "iscrowd": 0, "area": 32379.79975, "rle": {"size": [478, 500], "counts": "k\\U3:b>=D;D=D:E:G5J5L5J5L5J5L3M4K4M4K4M4K4M2M3N2M3N2N2N1O2N2N2O1N2N1O2N2N2O1N1O2N2N2O1N2N1O2N2N2O0O1O2N1O10001O0O10001O0O10001O0O10001O000O2O00000O1000000O100000000O1000000O1000000O1000000O100000000O100000O01000000O10O100000O10O1000O100000O2O0000001N10001O00000O2O000O2O000O101N10000O2O000O2O000O101N10001N10000O2O000O1O2N1O1N3N2N1O2M3N2N2M2O2N2N2N2N1O2N2N2N101N2N2N2N1O2N2N2N2N1O2O1M3N2M3N4K5K5L4K5K5L4K5L4K5K5L4K5L4K5K5L4K7I6K5J7JWRP1"}, "label": "a gravy in a bowl near popcorn"}]}
{"id": 203098, "image": "COCO_train2014_000000203098.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of chili with corn in it\"."}], "task": "refering", "answer_template": "The \"a bowl of chili with corn in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 44, 45, 46, 47, 48, 49, 50, 61, 62, 63, 64, 65, 66, 67, 68, 69, 79, 80, 81, 82, 83, 84, 85, 86, 87, 97, 98, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 134, 135, 136, 137, 138, 139, 153, 154, 155, 156], "bbox": [0.43316000000000004, 0.07742677824267782, 0.86172, 0.49198744769874475], "iscrowd": 0, "area": 32379.79975, "rle": {"size": [478, 500], "counts": "k\\U3:b>=D;D=D:E:G5J5L5J5L5J5L3M4K4M4K4M4K4M2M3N2M3N2N2N1O2N2N2O1N2N1O2N2N2O1N1O2N2N2O1N2N1O2N2N2O0O1O2N1O10001O0O10001O0O10001O0O10001O000O2O00000O1000000O100000000O1000000O1000000O1000000O100000000O100000O01000000O10O100000O10O1000O100000O2O0000001N10001O00000O2O000O2O000O101N10000O2O000O2O000O101N10001N10000O2O000O1O2N1O1N3N2N1O2M3N2N2M2O2N2N2N2N1O2N2N2N101N2N2N2N1O2N2N2N2N1O2O1M3N2M3N4K5K5L4K5K5L4K5L4K5K5L4K5L4K5K5L4K7I6K5J7JWRP1"}, "label": "a bowl of chili with corn in it"}]}
{"id": 563545, "image": "COCO_train2014_000000563545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older man in the back\"."}], "task": "refering", "answer_template": "The \"older man in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252, 274, 275, 297, 298, 320, 321, 343, 344], "bbox": [0.811125, 0.13707259953161594, 1.0, 0.9797658079625293], "iscrowd": 0, "area": 23373.56759999999, "rle": {"size": [427, 640], "counts": "iah61X=4K5L4L4L4M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3XOh0\\Od0B>N1O1O2N100O2N1O2N102M3M3M2N3N2M2N2N1O2^HPKf6P5XISKf6o4WISKh6n4VITKi6n4TITKk6h5O0O1O1O1O2O0O1O1O1O100O1O1O1O101N2N3M3N1_MYI[Ni6\\1]IeNf6S1^InNh6g0\\IZOU9QNnFo1[9eMhF\\2S:2N3M3M3M3M3M3M3M4L3M3M3M2N1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O001nLVLaLk3Z3hLWLX3e3ZMmKg2n3cMlK^2o3lMkKT2Q4UNjKl1o3`NjKb1l3mNnKT1h3[ORLR1\\3YO^LR2Q2YNjMR3e0aM"}, "label": "older man in the back"}]}
{"id": 563545, "image": "COCO_train2014_000000563545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on the right with no hat and jacket on\"."}], "task": "refering", "answer_template": "The \"man on the right with no hat and jacket on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252, 274, 275, 297, 298, 320, 321, 343, 344], "bbox": [0.811125, 0.13707259953161594, 1.0, 0.9797658079625293], "iscrowd": 0, "area": 23373.56759999999, "rle": {"size": [427, 640], "counts": "iah61X=4K5L4L4L4M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3XOh0\\Od0B>N1O1O2N100O2N1O2N102M3M3M2N3N2M2N2N1O2^HPKf6P5XISKf6o4WISKh6n4VITKi6n4TITKk6h5O0O1O1O1O2O0O1O1O1O100O1O1O1O101N2N3M3N1_MYI[Ni6\\1]IeNf6S1^InNh6g0\\IZOU9QNnFo1[9eMhF\\2S:2N3M3M3M3M3M3M3M4L3M3M3M2N1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O001nLVLaLk3Z3hLWLX3e3ZMmKg2n3cMlK^2o3lMkKT2Q4UNjKl1o3`NjKb1l3mNnKT1h3[ORLR1\\3YO^LR2Q2YNjMR3e0aM"}, "label": "man on the right with no hat and jacket on"}]}
{"id": 563545, "image": "COCO_train2014_000000563545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black sweatshirt between man in suit and woman\"."}], "task": "refering", "answer_template": "The \"man in black sweatshirt between man in suit and woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 220, 221, 241, 242, 243, 244, 264, 265, 287, 288, 289, 310, 311, 312, 333, 334], "bbox": [0.495796875, 0.14304449648711942, 0.665453125, 1.0], "iscrowd": 0, "area": 20292.5081, "rle": {"size": [427, 640], "counts": "^^T4j0`;V1^Oc0lGaMd4o2VKVM]4T3bKQMo3[3nKkLh3[3VLjLa3S1TJ;X2hNY3S1bJ3l1WOX3k0nJM_1F\\3`0XKHR12b39_K@g0b0e32fKPNAe0k0f1j3GnKgMK?4`2o3^OTL\\M66IV3j3\\OQNf0o1ZOoMh0P2YOmMj0S2VOkMl0T2UOjMm0V2TOgMn0X2UOlMe0T2]OmM`0R2DmM:R2InM5R2MoM0P23PNKP28PNEo1=SN@m1`0VN]Oi1d0ZNYOf1g0]NVOb1k0aNRO_1n0dNoN[1R1hNkN7_MlMf3o1iN4dMjMd3S2gN3hMgMa3X2eN1lMdMa3[2bN\\OeNVNj2^2`NXOlNWNh2_2WN]OSOSNh2`2PN_OZOoMi2a2hMCAkMj2`2`MHHfMk2b2WMJ1bMl2b2nLN8`Ml2a2gL2a0YMl2d2]L6j0TMk2e2VL;i5ERJ>Q6AkIb0W6X32M2O2N2N2N2M3N1O2N2N2N2O1N2N1O10N1\\K]J_1d5_NcJ\\1^5bNfJZ1[5eNgJY1[5eNfJZ1[5fNfJW1]5gNdJX1]5hNdJV1]5iNcJW1_5hN`JX1b5gN]JX1f5fN[JY1g5fNXJZ1j5dNVJ\\1l5cNSJ]1P6`NQJ^1R6aNmI\\1X6cNgIZ1_6cNbIY1g6bNXI[1P7`NPI^1X7`MZH0>\\2Z8cMfGZ2^8dMcG[2^8eMcGX2`8fMaGX2b8fM`GX2a8hM_GV2d8hM^GV2d8hM]GW2d8iM]GU2e8iMnGd1X8WNlGd1Q:L5K4SObDFb;4o0K5K5KPQi2"}, "label": "man in black sweatshirt between man in suit and woman"}]}
{"id": 563545, "image": "COCO_train2014_000000563545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green sweater\"."}], "task": "refering", "answer_template": "The \"a man in a green sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 220, 221, 241, 242, 243, 244, 264, 265, 287, 288, 289, 310, 311, 312, 333, 334], "bbox": [0.495796875, 0.14304449648711942, 0.665453125, 1.0], "iscrowd": 0, "area": 20292.5081, "rle": {"size": [427, 640], "counts": "^^T4j0`;V1^Oc0lGaMd4o2VKVM]4T3bKQMo3[3nKkLh3[3VLjLa3S1TJ;X2hNY3S1bJ3l1WOX3k0nJM_1F\\3`0XKHR12b39_K@g0b0e32fKPNAe0k0f1j3GnKgMK?4`2o3^OTL\\M66IV3j3\\OQNf0o1ZOoMh0P2YOmMj0S2VOkMl0T2UOjMm0V2TOgMn0X2UOlMe0T2]OmM`0R2DmM:R2InM5R2MoM0P23PNKP28PNEo1=SN@m1`0VN]Oi1d0ZNYOf1g0]NVOb1k0aNRO_1n0dNoN[1R1hNkN7_MlMf3o1iN4dMjMd3S2gN3hMgMa3X2eN1lMdMa3[2bN\\OeNVNj2^2`NXOlNWNh2_2WN]OSOSNh2`2PN_OZOoMi2a2hMCAkMj2`2`MHHfMk2b2WMJ1bMl2b2nLN8`Ml2a2gL2a0YMl2d2]L6j0TMk2e2VL;i5ERJ>Q6AkIb0W6X32M2O2N2N2N2M3N1O2N2N2N2O1N2N1O10N1\\K]J_1d5_NcJ\\1^5bNfJZ1[5eNgJY1[5eNfJZ1[5fNfJW1]5gNdJX1]5hNdJV1]5iNcJW1_5hN`JX1b5gN]JX1f5fN[JY1g5fNXJZ1j5dNVJ\\1l5cNSJ]1P6`NQJ^1R6aNmI\\1X6cNgIZ1_6cNbIY1g6bNXI[1P7`NPI^1X7`MZH0>\\2Z8cMfGZ2^8dMcG[2^8eMcGX2`8fMaGX2b8fM`GX2a8hM_GV2d8hM^GV2d8hM]GW2d8iM]GU2e8iMnGd1X8WNlGd1Q:L5K4SObDFb;4o0K5K5KPQi2"}, "label": "a man in a green sweater"}]}
{"id": 563545, "image": "COCO_train2014_000000563545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the green sweater vest and red turtleneck\"."}], "task": "refering", "answer_template": "The \"the woman in the green sweater vest and red turtleneck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 196, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.51359375, 0.23824355971896957, 0.9208124999999999, 0.9995784543325527], "iscrowd": 0, "area": 55215.475800000015, "rle": {"size": [427, 640], "counts": "X^Y41;5\\<3ZC4a<>L3N3O0O3MWOjC?U<\\ORDd07XOS;0jDj0M\\Oe;g0UD]Oj;T1N2M3N2N2N4\\OVN_Em1[:XNcEk1Z:WNcEm1Z:UNdEo1X:SNfEo1X:SNeEo1Y:d0O10O101O2N2M3M0000O10000O1000000004L4L4K2O001N2O001O1N101OYFkMW8T2bGSN_8m1]FlMg0>m8k1kF\\NT9e1fFaNg4ZOWOX2jKeNm4VOWO^3g0eLVO]3i0eLUO\\3j0eLTO]3k0dLTO]3l0cLRO_3m0bLRO_3n0aLPOa3o0_LQOb3o0^LPOc3o0]LPOe3P1ZLPOg3o0ZLoNh3Q1WLoNj3P1VLoNl3Q1SLnNo3Q1RLlNQ4T1nKiNV4V1jKgNZ4Y1fKdN]4[1cKeN^4[1aKdNa4[1`KcNb4]1_K`Nc4_1^K_Nd4a1\\K]Nf4b1[K\\Ng4d1ZKZNg4e1ZKXL\\OX1\\5`2YKVL^OV1[5c2[KRL^OW1Y5\\2aJRLn05[O[1X5\\2aJPLS11YOa1U5]2\\LbMe3]2^LmLUN\\O_5e3`LlLSN]O^5f3cLhLRN@\\5g3fLdLQNCZ5g3kL`LmMHY5g3nL\\Li3c3[LYLf3e3Q3O1O1N200O100O100O100O100O10000O100O100O10000000000O1000000000000O100000000000000O10000000000001O00001O000010O0001O001O001O001O1O001O001O001O1O1O1O1O001O1O1O1O1O1O003M2N3M2N3M3M2N1O2N1O2N1O2N1O1O2N1O2N1O2N2N2N2N2Nb0^O001O1O1O1O1O1O1O1O1O001O1O1O4L7I3M2N1O1O1O2VJWIn4j6PKXIo4i6oJYIP5h6oJZIo4Q7eJQIZ5b7N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N3]Ob0L5I7\\MiFj0]9oNRGa0X9WOUG:^Xe0"}, "label": "the woman in the green sweater vest and red turtleneck"}]}
{"id": 563545, "image": "COCO_train2014_000000563545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a baseball cap\"."}], "task": "refering", "answer_template": "The \"a woman wearing a baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 196, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.51359375, 0.23824355971896957, 0.9208124999999999, 0.9995784543325527], "iscrowd": 0, "area": 55215.475800000015, "rle": {"size": [427, 640], "counts": "X^Y41;5\\<3ZC4a<>L3N3O0O3MWOjC?U<\\ORDd07XOS;0jDj0M\\Oe;g0UD]Oj;T1N2M3N2N2N4\\OVN_Em1[:XNcEk1Z:WNcEm1Z:UNdEo1X:SNfEo1X:SNeEo1Y:d0O10O101O2N2M3M0000O10000O1000000004L4L4K2O001N2O001O1N101OYFkMW8T2bGSN_8m1]FlMg0>m8k1kF\\NT9e1fFaNg4ZOWOX2jKeNm4VOWO^3g0eLVO]3i0eLUO\\3j0eLTO]3k0dLTO]3l0cLRO_3m0bLRO_3n0aLPOa3o0_LQOb3o0^LPOc3o0]LPOe3P1ZLPOg3o0ZLoNh3Q1WLoNj3P1VLoNl3Q1SLnNo3Q1RLlNQ4T1nKiNV4V1jKgNZ4Y1fKdN]4[1cKeN^4[1aKdNa4[1`KcNb4]1_K`Nc4_1^K_Nd4a1\\K]Nf4b1[K\\Ng4d1ZKZNg4e1ZKXL\\OX1\\5`2YKVL^OV1[5c2[KRL^OW1Y5\\2aJRLn05[O[1X5\\2aJPLS11YOa1U5]2\\LbMe3]2^LmLUN\\O_5e3`LlLSN]O^5f3cLhLRN@\\5g3fLdLQNCZ5g3kL`LmMHY5g3nL\\Li3c3[LYLf3e3Q3O1O1N200O100O100O100O100O10000O100O100O10000000000O1000000000000O100000000000000O10000000000001O00001O000010O0001O001O001O001O1O001O001O001O1O1O1O1O001O1O1O1O1O1O003M2N3M2N3M3M2N1O2N1O2N1O2N1O1O2N1O2N1O2N2N2N2N2Nb0^O001O1O1O1O1O1O1O1O1O001O1O1O4L7I3M2N1O1O1O2VJWIn4j6PKXIo4i6oJYIP5h6oJZIo4Q7eJQIZ5b7N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N3]Ob0L5I7\\MiFj0]9oNRGa0X9WOUG:^Xe0"}, "label": "a woman wearing a baseball cap"}]}
{"id": 260443, "image": "COCO_train2014_000000260443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sofa\"."}], "task": "refering", "answer_template": "The \"the sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 183, 184, 185], "bbox": [0.1635, 0.615, 0.54926, 0.8979518072289157], "iscrowd": 0, "area": 11404.021700000001, "rle": {"size": [332, 500], "counts": "jij0h0b9?B6I7I7I8I6K5K1O001O001O0000O1O100O100001O001O00001O001O0000000000000000000000O010000000000O1000000000000O1000O10000000O10C[HSNe7k1^HTNb7k1`0O1N3N1O1O1N2O2N1O1N2O1O2O0O10000O2O0O100O100O100O010O10O0100O010O100O010O100O01000001N102N1O1N2O1O001N101O001O1O001O001N1000O10000001O001O001N10001O001OO1O1O100O1O10O0100O10000O100O10000O1000O1000000001O1O1O1O1O1O1O1O100O1O1O1O1O2O0O2O1N101N14M2NmNWG?^9L4M3M3LVfX2"}, "label": "the sofa"}]}
{"id": 260443, "image": "COCO_train2014_000000260443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green three person couch with three pillows on it\"."}], "task": "refering", "answer_template": "The \"a green three person couch with three pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 183, 184, 185], "bbox": [0.1635, 0.615, 0.54926, 0.8979518072289157], "iscrowd": 0, "area": 11404.021700000001, "rle": {"size": [332, 500], "counts": "jij0h0b9?B6I7I7I8I6K5K1O001O001O0000O1O100O100001O001O00001O001O0000000000000000000000O010000000000O1000000000000O1000O10000000O10C[HSNe7k1^HTNb7k1`0O1N3N1O1O1N2O2N1O1N2O1O2O0O10000O2O0O100O100O100O010O10O0100O010O100O010O100O01000001N102N1O1N2O1O001N101O001O1O001O001N1000O10000001O001O001N10001O001OO1O1O100O1O10O0100O10000O100O10000O1000O1000000001O1O1O1O1O1O1O1O100O1O1O1O1O2O0O2O1N101N14M2NmNWG?^9L4M3M3LVfX2"}, "label": "a green three person couch with three pillows on it"}]}
{"id": 375133, "image": "COCO_train2014_000000375133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in orange ski pants\"."}], "task": "refering", "answer_template": "The \"a man in orange ski pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 84, 85, 86, 87, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193, 194, 195, 210, 211, 212, 213, 214], "bbox": [0.64216, 0.24852852852852855, 0.89842, 0.9867867867867869], "iscrowd": 0, "area": 20517.873700000004, "rle": {"size": [333, 500], "counts": "XaX34W:9H8G9H8G8I8G9UNWNbJl1\\5\\N\\Jf1b5^NYJd1g5^NVJc1i5_NTJd1j5^NTJc1l5_NPJd1n5^NoId1P6_NmIb1R6`NkIc1T6^NjIc1U6`NgIb1Y6_NeIc1Z6X1UMQLWOX4>kKB^42eKNd4F^K;k4YOXKg0R5mNPKS1Y5bNiJ^1\\5[NfJe1`5TNaJl1d5mM^JS2h5fMYJZ2l5_MVJa2k6000O101O001O001O001O001O0O2O001O003M=C001O001O0O2O001O0M4L3N3L3B?^Oa0O2O001O001O0O2O0013L3N3L3M3N3L4M2M4M3L4L4M1N01O001O00001O001O0010O00TMSLXOn3a0\\L[Oe3>dL_O^39lLCU3<oL@S3=RM_OP3>TM_Om2?WM^Ok2?YM^Oi2`0ZM]Og2a0^M\\Oc2a0aM\\Oa2a0cM\\O^2b0fM[O\\2b0hM[OZ2c0jMYOX2d0lMYOV2d0mMZOU2c0oMZOS2c0PN[OR2`0TN]On1>WN@k1:[NEf15aNFc14cNGb13eNG`13ha`0"}, "label": "a man in orange ski pants"}]}
{"id": 375133, "image": "COCO_train2014_000000375133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier wearing black jacket and brown colour trouser\"."}], "task": "refering", "answer_template": "The \"a skier wearing black jacket and brown colour trouser\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 84, 85, 86, 87, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193, 194, 195, 210, 211, 212, 213, 214], "bbox": [0.64216, 0.24852852852852855, 0.89842, 0.9867867867867869], "iscrowd": 0, "area": 20517.873700000004, "rle": {"size": [333, 500], "counts": "XaX34W:9H8G9H8G8I8G9UNWNbJl1\\5\\N\\Jf1b5^NYJd1g5^NVJc1i5_NTJd1j5^NTJc1l5_NPJd1n5^NoId1P6_NmIb1R6`NkIc1T6^NjIc1U6`NgIb1Y6_NeIc1Z6X1UMQLWOX4>kKB^42eKNd4F^K;k4YOXKg0R5mNPKS1Y5bNiJ^1\\5[NfJe1`5TNaJl1d5mM^JS2h5fMYJZ2l5_MVJa2k6000O101O001O001O001O001O0O2O001O003M=C001O001O0O2O001O0M4L3N3L3B?^Oa0O2O001O001O0O2O0013L3N3L3M3N3L4M2M4M3L4L4M1N01O001O00001O001O0010O00TMSLXOn3a0\\L[Oe3>dL_O^39lLCU3<oL@S3=RM_OP3>TM_Om2?WM^Ok2?YM^Oi2`0ZM]Og2a0^M\\Oc2a0aM\\Oa2a0cM\\O^2b0fM[O\\2b0hM[OZ2c0jMYOX2d0lMYOV2d0mMZOU2c0oMZOS2c0PN[OR2`0TN]On1>WN@k1:[NEf15aNFc14cNGb13eNG`13ha`0"}, "label": "a skier wearing black jacket and brown colour trouser"}]}
{"id": 375133, "image": "COCO_train2014_000000375133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling girl in a gray snow jacket and black gloves\"."}], "task": "refering", "answer_template": "The \"a smiling girl in a gray snow jacket and black gloves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 80, 81, 97, 98, 99, 115, 116, 117, 132, 133, 134, 135, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 186, 187, 188, 189, 205, 206, 207], "bbox": [0.363, 0.3164564564564564, 0.564, 0.9876276276276276], "iscrowd": 0, "area": 15835.005, "rle": {"size": [333, 500], "counts": "U]k1m1R8?I7N2M3M3M3N2M3M3M3N1N3M3M3N1N3M3M3N1N3M3M2N2M3M3L3NK502N1O12N5L4K6K4K5WKnJP4V5iKnJV4V5cKnJ[4c5O1N2L4K5L4M3N2O1N2N2N2N2N2O1000000O1000000O10O1000O100000000001O2N1O1O2N1O1UO[LlJg3Q5\\LmJe3Q5^LmJk3i4XLUKR4_4SL^KP4^4SL`KQ4[4RLcKQ4Y4RLeKR4V4QLhKT4R4oKlKV4n3mKPLX4j3kKTL[4e3^1I8G8H8H8I8G8H8Fa0]Of0ZOcdV2"}, "label": "a smiling girl in a gray snow jacket and black gloves"}]}
{"id": 375133, "image": "COCO_train2014_000000375133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the white ski jacket\"."}], "task": "refering", "answer_template": "The \"the person in the white ski jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 80, 81, 97, 98, 99, 115, 116, 117, 132, 133, 134, 135, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 186, 187, 188, 189, 205, 206, 207], "bbox": [0.363, 0.3164564564564564, 0.564, 0.9876276276276276], "iscrowd": 0, "area": 15835.005, "rle": {"size": [333, 500], "counts": "U]k1m1R8?I7N2M3M3M3N2M3M3M3N1N3M3M3N1N3M3M3N1N3M3M2N2M3M3L3NK502N1O12N5L4K6K4K5WKnJP4V5iKnJV4V5cKnJ[4c5O1N2L4K5L4M3N2O1N2N2N2N2N2O1000000O1000000O10O1000O100000000001O2N1O1O2N1O1UO[LlJg3Q5\\LmJe3Q5^LmJk3i4XLUKR4_4SL^KP4^4SL`KQ4[4RLcKQ4Y4RLeKR4V4QLhKT4R4oKlKV4n3mKPLX4j3kKTL[4e3^1I8G8H8H8I8G8H8Fa0]Of0ZOcdV2"}, "label": "the person in the white ski jacket"}]}
{"id": 375133, "image": "COCO_train2014_000000375133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"snowboarder in dark green jacket holding a snowboard\"."}], "task": "refering", "answer_template": "The \"snowboarder in dark green jacket holding a snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 76, 77, 78, 92, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167, 168, 183, 184, 185, 186, 201, 202, 203, 204], "bbox": [0.12122, 0.3168468468468469, 0.3981, 0.9865165165165165], "iscrowd": 0, "area": 15418.213500000002, "rle": {"size": [333, 500], "counts": "loc04V:4K5K401N2O0O2O1N1O2OL3L5J5010O1O11O01O00000001O000001O01O00000000VKFA:=HD88KH56MJ330M012ONN52KL75HH;8EF<;EA>?IcMn1]2SN^MQ2c2oMYMT2g2mMTMW2l2jMoLZ2Q3gMkL\\2U3eMfL_2Z3bMaLb2`3]M]Lf2c3[MXLi2h3XMSLl2m3UMoKn2Q4SMjKQ3W4b10000O1O1O1N2O1N2O1O1N2O1N2L4^Ob0O1O1O1O1N2O1O1O10000000000001O0000000000001O0000001O00001O001O00001O5J8C=C<D=B=D8H7I7I7H8E;[Oe0\\Oe0M2O1N2N2N2N2M3M3L4M3M3L4M3L5L5K6I7JafQ3"}, "label": "snowboarder in dark green jacket holding a snowboard"}]}
{"id": 375133, "image": "COCO_train2014_000000375133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person holding snow board\"."}], "task": "refering", "answer_template": "The \"person holding snow board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 76, 77, 78, 92, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167, 168, 183, 184, 185, 186, 201, 202, 203, 204], "bbox": [0.12122, 0.3168468468468469, 0.3981, 0.9865165165165165], "iscrowd": 0, "area": 15418.213500000002, "rle": {"size": [333, 500], "counts": "loc04V:4K5K401N2O0O2O1N1O2OL3L5J5010O1O11O01O00000001O000001O01O00000000VKFA:=HD88KH56MJ330M012ONN52KL75HH;8EF<;EA>?IcMn1]2SN^MQ2c2oMYMT2g2mMTMW2l2jMoLZ2Q3gMkL\\2U3eMfL_2Z3bMaLb2`3]M]Lf2c3[MXLi2h3XMSLl2m3UMoKn2Q4SMjKQ3W4b10000O1O1O1N2O1N2O1O1N2O1N2L4^Ob0O1O1O1O1N2O1O1O10000000000001O0000000000001O0000001O00001O001O00001O5J8C=C<D=B=D8H7I7I7H8E;[Oe0\\Oe0M2O1N2N2N2N2M3M3L4M3M3L4M3L5L5K6I7JafQ3"}, "label": "person holding snow board"}]}
{"id": 260448, "image": "COCO_train2014_000000260448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman bottom left corner\"."}], "task": "refering", "answer_template": "The \"woman bottom left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 91, 92, 93, 109, 110, 126, 127, 128, 129, 130, 145, 146, 147, 148, 163, 164, 165, 181, 182, 183, 199, 200, 201, 217, 218, 219], "bbox": [0.04064, 0.3327466666666667, 0.24101999999999998, 0.9887733333333333], "iscrowd": 0, "area": 12927.574049999996, "rle": {"size": [375, 500], "counts": "ll74a;3L4L4M3L4L4L4RHUOk4P1QKTOl4Q1nJROo4S1nJnNo4V1nJlNo4W1PKjNn4X1PKjNm4Y1RKhNk4[1SKgNj4\\1UKeNh4^1VKdNg4c1TK_Nh4h1RK[Nk4j1QKXNk4m1QK[Nf4i1WK_Na4e1YKeNP4m1kK\\Ni3m1QLYNf3o1TLUNj3m1SLWNj3k1SLXNl3i1QLZNn3f1PL]No3d1nK_NR4a1kKbNT4_1iKdNW4^4O100O1O100O100O100O1O100O100001O000O2O0ZJoKf3R4XLRLe3n3YLXLc3i3[L\\La3d3^LaL^3_3aLfL[3[3bLkLZ3U3eLoLX3Q3gLTMU3m2hLYMT3g2kL^MQ3c2mLbM^OYNe2U4lMgMZO_N`2j3TNlMYO^O_1g2VOjM]OFV1`2[OjM@Om0Y2_OiME3g0V2@gMJ6d0[2XO_M68`0^2SOiLLgNc0T2<a2nNgL8\\N?^2:Y44iKJW45lKIT47nKGR49PLEP4:SLDm3<ULBk3=WLBi3>YL@Q46QLHQ45RLIQ44QLJn36ULHk38WLFi3:XLEh3:[LDd3=^LAb3>aL@_3`0cL^O]3b0k301L4M3L3M4M`bZ4"}, "label": "woman bottom left corner"}]}
{"id": 260448, "image": "COCO_train2014_000000260448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in bluejeans is carrying a purse\"."}], "task": "refering", "answer_template": "The \"the woman in bluejeans is carrying a purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 91, 92, 93, 109, 110, 126, 127, 128, 129, 130, 145, 146, 147, 148, 163, 164, 165, 181, 182, 183, 199, 200, 201, 217, 218, 219], "bbox": [0.04064, 0.3327466666666667, 0.24101999999999998, 0.9887733333333333], "iscrowd": 0, "area": 12927.574049999996, "rle": {"size": [375, 500], "counts": "ll74a;3L4L4M3L4L4L4RHUOk4P1QKTOl4Q1nJROo4S1nJnNo4V1nJlNo4W1PKjNn4X1PKjNm4Y1RKhNk4[1SKgNj4\\1UKeNh4^1VKdNg4c1TK_Nh4h1RK[Nk4j1QKXNk4m1QK[Nf4i1WK_Na4e1YKeNP4m1kK\\Ni3m1QLYNf3o1TLUNj3m1SLWNj3k1SLXNl3i1QLZNn3f1PL]No3d1nK_NR4a1kKbNT4_1iKdNW4^4O100O1O100O100O100O1O100O100001O000O2O0ZJoKf3R4XLRLe3n3YLXLc3i3[L\\La3d3^LaL^3_3aLfL[3[3bLkLZ3U3eLoLX3Q3gLTMU3m2hLYMT3g2kL^MQ3c2mLbM^OYNe2U4lMgMZO_N`2j3TNlMYO^O_1g2VOjM]OFV1`2[OjM@Om0Y2_OiME3g0V2@gMJ6d0[2XO_M68`0^2SOiLLgNc0T2<a2nNgL8\\N?^2:Y44iKJW45lKIT47nKGR49PLEP4:SLDm3<ULBk3=WLBi3>YL@Q46QLHQ45RLIQ44QLJn36ULHk38WLFi3:XLEh3:[LDd3=^LAb3>aL@_3`0cL^O]3b0k301L4M3L3M4M`bZ4"}, "label": "the woman in bluejeans is carrying a purse"}]}
{"id": 260448, "image": "COCO_train2014_000000260448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light colored wooden bench in front of two people\"."}], "task": "refering", "answer_template": "The \"a light colored wooden bench in front of two people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.41524, 0.6930133333333334, 1.0, 0.9891733333333333], "iscrowd": 0, "area": 23126.27145, "rle": {"size": [375, 500], "counts": "Q_\\2=Y;>B;F5J6J5K00100O1O001N2`FVNQ9l1gFaNR9S2O1000000O1000000O1002N2N2N2N2N000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000O10O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O12N2N2N1O1O00001O00001O00001O0000O100XO\\M]Hd2c7bMVH_2j7gMoGZ2Q8f000001O1O001O1O0_NeGF\\88gGFZ89gGFY8:gGFZ88gGHZ87fGIZ87fGI[85fGK[84dGM\\83dGM\\82eGN[82dGO\\81dGO\\80eG0[80eG0[80dG0^8NcG2^8MbG3`8K`G5b8I]G8d8H[G8n8_ORGa0f9hNYFX1R:1000000O10000O1000000O1000000O1000000O10000O0100000O1000000O1000000O1000000O10000O100000F;QOWG"}, "label": "a light colored wooden bench in front of two people"}]}
{"id": 260448, "image": "COCO_train2014_000000260448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light wooden back rest of a bench\"."}], "task": "refering", "answer_template": "The \"a light wooden back rest of a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.41524, 0.6930133333333334, 1.0, 0.9891733333333333], "iscrowd": 0, "area": 23126.27145, "rle": {"size": [375, 500], "counts": "Q_\\2=Y;>B;F5J6J5K00100O1O001N2`FVNQ9l1gFaNR9S2O1000000O1000000O1002N2N2N2N2N000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000O10O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000000000O1000000000000O12N2N2N1O1O00001O00001O00001O0000O100XO\\M]Hd2c7bMVH_2j7gMoGZ2Q8f000001O1O001O1O0_NeGF\\88gGFZ89gGFY8:gGFZ88gGHZ87fGIZ87fGI[85fGK[84dGM\\83dGM\\82eGN[82dGO\\81dGO\\80eG0[80eG0[80dG0^8NcG2^8MbG3`8K`G5b8I]G8d8H[G8n8_ORGa0f9hNYFX1R:1000000O10000O1000000O1000000O1000000O10000O0100000O1000000O1000000O1000000O10000O100000F;QOWG"}, "label": "a light wooden back rest of a bench"}]}
{"id": 260448, "image": "COCO_train2014_000000260448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red headed woman in striped shirt and grey pants\"."}], "task": "refering", "answer_template": "The \"red headed woman in striped shirt and grey pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 136, 151, 152, 153, 154, 169, 170, 171, 172, 188, 189, 190, 206, 207, 208, 224, 225, 226], "bbox": [0.39814, 0.32552, 0.60556, 0.9897066666666666], "iscrowd": 0, "area": 12809.690449999998, "rle": {"size": [375, 500], "counts": "YRY23`;7J6I6K4O2O01O01O01O01O0101N2N2O1N1OjEAU9>fFHY97cFNZ94`F3]9V1WOUNaGP2Y8UNcGU2R8QNiGT2Q8RNjGP2S8VNhGk1V8l0N2O100O1O101N3hNYL\\Ji3a5[L[Jf3d5^LYJb3f5bLVJ^3j5fLRJ[37XLe4a0QKW38ZLf4c0nJT3<YLe4h0bJX3h0QLe4P5[KPKN1f3o4]LPKM:\\3g4gLnJN;Z3h4gLnJO:Y3i4hLmJO:Y3i4hLmJN;Z3g4iLnJM;Z3g4iLnJM;Y3h4iLnJN:Y3h4iLmJO;X3h4iLmJN<X3h4jLlJN<X3g4jLnJN;X3g4jLnJN;W3h4kLmJN;W3h4kLmJM<X3g4kLlJN=X3e4jLoJN<X3e4jLoJN<Y3d4iLPKN<Y3d4iLPKN<Z3c4gLRKN<[3b4gLRKN<\\3`4gLTKM<\\3`4gLSKN=\\3_4eLUKO<\\3_4eLUKN=^3]4dLVKN=^3]4dLVKN=_3\\4cLWKN=_3[4cLYKN<`3Z4bLYKN=a3Z4aLYKN=b3Y4`LZKN=b3Y4_L[KO<c3X4^L\\KO<f3U4[L_KN=l3n3WLeKM=Q4DcK`3?^LN>f4S3[K`LO=g4R3ZKaLOL[O9]5Y3YKbLNLC3W5^3XKcLNKJNQ5c3VKeLOI`5`3bJgLNGc5a3_JgLOGd5a3]JhLNGg5`3ZJjL^6U3bIkL_6T3aIaLk6h2^1L4K5K5F;E=C=C=D<CdnW2"}, "label": "red headed woman in striped shirt and grey pants"}]}
{"id": 260448, "image": "COCO_train2014_000000260448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older woman wearing a striped shirt\"."}], "task": "refering", "answer_template": "The \"an older woman wearing a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 98, 99, 100, 115, 116, 117, 118, 133, 134, 135, 136, 151, 152, 153, 154, 169, 170, 171, 172, 188, 189, 190, 206, 207, 208, 224, 225, 226], "bbox": [0.39814, 0.32552, 0.60556, 0.9897066666666666], "iscrowd": 0, "area": 12809.690449999998, "rle": {"size": [375, 500], "counts": "YRY23`;7J6I6K4O2O01O01O01O01O0101N2N2O1N1OjEAU9>fFHY97cFNZ94`F3]9V1WOUNaGP2Y8UNcGU2R8QNiGT2Q8RNjGP2S8VNhGk1V8l0N2O100O1O101N3hNYL\\Ji3a5[L[Jf3d5^LYJb3f5bLVJ^3j5fLRJ[37XLe4a0QKW38ZLf4c0nJT3<YLe4h0bJX3h0QLe4P5[KPKN1f3o4]LPKM:\\3g4gLnJN;Z3h4gLnJO:Y3i4hLmJO:Y3i4hLmJN;Z3g4iLnJM;Z3g4iLnJM;Y3h4iLnJN:Y3h4iLmJO;X3h4iLmJN<X3h4jLlJN<X3g4jLnJN;X3g4jLnJN;W3h4kLmJN;W3h4kLmJM<X3g4kLlJN=X3e4jLoJN<X3e4jLoJN<Y3d4iLPKN<Y3d4iLPKN<Z3c4gLRKN<[3b4gLRKN<\\3`4gLTKM<\\3`4gLSKN=\\3_4eLUKO<\\3_4eLUKN=^3]4dLVKN=^3]4dLVKN=_3\\4cLWKN=_3[4cLYKN<`3Z4bLYKN=a3Z4aLYKN=b3Y4`LZKN=b3Y4_L[KO<c3X4^L\\KO<f3U4[L_KN=l3n3WLeKM=Q4DcK`3?^LN>f4S3[K`LO=g4R3ZKaLOL[O9]5Y3YKbLNLC3W5^3XKcLNKJNQ5c3VKeLOI`5`3bJgLNGc5a3_JgLOGd5a3]JhLNGg5`3ZJjL^6U3bIkL_6T3aIaLk6h2^1L4K5K5F;E=C=C=D<CdnW2"}, "label": "an older woman wearing a striped shirt"}]}
{"id": 579937, "image": "COCO_train2014_000000579937.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman walking in the rain with a red umbrella\"."}], "task": "refering", "answer_template": "The \"a woman walking in the rain with a red umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 308, 309, 310, 311, 312, 333, 334, 356, 357, 379, 380, 402, 403, 425, 426], "bbox": [0.419953125, 0.441625, 0.5984062499999999, 0.84734375], "iscrowd": 0, "area": 12264.147099999996, "rle": {"size": [640, 640], "counts": "n_X52mc03Z\\OL[c06a\\OL^c0<O001O0ZNCk_O=S`0Fk_O;R`0Il_O7S`0Km_O5Q`0Mn_O3R`0Nm_O3Q`0Oo_O0Q`01n_O0Q`01n_O0Q`01n_OOR`02n_ONQ`02o_ONP`04o_OMP`04P@Lo?5P@KP`06o_OKP`06n_OKR`06m_OKR`05n_OKR`06m_OKS`05l_OKU`05j_OLU`04j_OMW`03h_ONX`02g_ONY`03f_ONZ`01e_O0\\`00c_O1\\`00c_O0^`00`_O2``0M^_O5b`0L[_O6f`0JW_O5l`0LP_O4Ta0Kg^O7]a0J]^O7ga0m0_@_Nk<]1\\BF^=6iAm0Q>h2H7J5K5K4L3M4K5L2N2N2N2N2M3N2N00N2O001N2O1N101O1N2O1N101O1N2c0]O5K5K5K4L5K5KD<\\Md2^LPAl1X?d0TAYLT?U1`@m0[`0T18nM__Oc0d`0\\O^_Oa0c`0^O`_O`0``0_Ob_O?_`0Ac_O=^`0Ae_O<\\`0Dm_O3S`0LT@Mn?2U@Jl?6X@Ei?:\\@Af?>^@]Od?b0c@WO^?h0n1N3N1N2N2N2N2N2N2N3M2NbZP5"}, "label": "a woman walking in the rain with a red umbrella"}]}
{"id": 579937, "image": "COCO_train2014_000000579937.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person walking down the street with the red umbrella\"."}], "task": "refering", "answer_template": "The \"the person walking down the street with the red umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 289, 308, 309, 310, 311, 312, 333, 334, 356, 357, 379, 380, 402, 403, 425, 426], "bbox": [0.419953125, 0.441625, 0.5984062499999999, 0.84734375], "iscrowd": 0, "area": 12264.147099999996, "rle": {"size": [640, 640], "counts": "n_X52mc03Z\\OL[c06a\\OL^c0<O001O0ZNCk_O=S`0Fk_O;R`0Il_O7S`0Km_O5Q`0Mn_O3R`0Nm_O3Q`0Oo_O0Q`01n_O0Q`01n_O0Q`01n_OOR`02n_ONQ`02o_ONP`04o_OMP`04P@Lo?5P@KP`06o_OKP`06n_OKR`06m_OKR`05n_OKR`06m_OKS`05l_OKU`05j_OLU`04j_OMW`03h_ONX`02g_ONY`03f_ONZ`01e_O0\\`00c_O1\\`00c_O0^`00`_O2``0M^_O5b`0L[_O6f`0JW_O5l`0LP_O4Ta0Kg^O7]a0J]^O7ga0m0_@_Nk<]1\\BF^=6iAm0Q>h2H7J5K5K4L3M4K5L2N2N2N2N2M3N2N00N2O001N2O1N101O1N2O1N101O1N2c0]O5K5K5K4L5K5KD<\\Md2^LPAl1X?d0TAYLT?U1`@m0[`0T18nM__Oc0d`0\\O^_Oa0c`0^O`_O`0``0_Ob_O?_`0Ac_O=^`0Ae_O<\\`0Dm_O3S`0LT@Mn?2U@Jl?6X@Ei?:\\@Af?>^@]Od?b0c@WO^?h0n1N3N1N2N2N2N2N2N2N3M2NbZP5"}, "label": "the person walking down the street with the red umbrella"}]}
{"id": 514404, "image": "COCO_train2014_000000514404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby wearing a red sweater\"."}], "task": "refering", "answer_template": "The \"a baby wearing a red sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 173, 174, 175, 176, 177, 178, 179, 180, 184, 185, 190, 191, 192, 193, 194, 202, 208, 209, 210, 211, 212, 227, 228, 229, 245, 246], "bbox": [0.17379166666666668, 0.0045, 0.9498125000000001, 0.62021875], "iscrowd": 0, "area": 71731.50940000002, "rle": {"size": [640, 480], "counts": "hgd18hc06I7I8I6I7J7H7J6I7I3N1N3N1N3N2M2O2M2N3N1N3N2M2O2M2N3YOgLc@Z3\\?hLa@[3\\?gLc@[3Z?gLd@Z3\\?gLc@Y3\\?iLb@X3]?jLa@X3^?iLa@W3^?lL_@U3a?lL]@U3b?mL\\@T3c?f0O0O1O101N100O1O101N1O100O1O2O0O100O2N100O1N2O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1N2O2N1O1OL41SBYJ`=h5^B\\J_=e5_B`J]=b5_BcJ^=S6L3N2M2O0O2O1N2O1N2O0WMVI[Hl6a7[I[Hf6a7bIYH`6c7hIXHY6d7oIVHR6g7VJTHk5h7\\JSHf5i7bJRH_5i7jJQHW5l7QKnGQ5n7bKbG_4Z8f3L3M4L3M3L4M3M4O0O1O100O1O1O2O0O1O1000001O00000000\\M\\HlHd7R7`HmH_7R7cHmH]7R7eHmH[7S7gHkHY7T7iHkHW7T7lHiHU7V7mHiHS7V7PIhHP7W7RIgHo6Y7RIfHn6Y7UIeHk6[7VIcHk6]7WIaHi6_7XI^Hj6b7WIZHl6f7VIUHm6k7TIQHo6o7SIlGP7T8QIhGR7X8oHdGT7\\8nH_GU7a8lH[GW7e8kHVGX7j8iHQG[7o8gHlF\\7T9eHhF^7X9e10000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000ZO\\EPHd:P8^EnGb:Q8bElG^:S8eEkG[:U8hEhGX:W8kEgGU:X8oEeGQ:[8QFcGo9\\8UFaGk9^8l000O100O1O100O1O100O1O1hL^DRNb;j1bDUN_;f1fDYN[;c1jD\\NV;_1oD`NR;\\1REcNo:Y1VEfNj:U1[EjNf:R1^EnNb:m0dERO\\:j0hEVOX:e0mE[OS:a0RF^On9>VFBn95WFKn9LWF3o9BVF>o9YOUFh0P:mNVFR1o9eNUF[1Q:ZNTFf1Q:QNTFn1R:hMRFX2S:^MRFb2g=O001O000010O0001O00001O00001O0000100O2N1O1O1O2N1O1O2N1O1O2O0O1O1O1O0010O10000O10000O10000O1000O0100O0100O01a^OoMn`0P2m^OVNSa0j1f^O\\NZa0V2000O010O10O10O010O10O0100O0100O100O10O001N1O2O0O2O0O2O0O2O0O2N101N101N100O2O0^OU^OnNma0m0Y^OPOha0l0^^OROca0i0c^OTO^a0g0i^OWOXa0d0T1K6J5L5Jbh>"}, "label": "a baby wearing a red sweater"}]}
{"id": 514404, "image": "COCO_train2014_000000514404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby boy wearing a red shirt and gray bib\"."}], "task": "refering", "answer_template": "The \"the baby boy wearing a red shirt and gray bib\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 173, 174, 175, 176, 177, 178, 179, 180, 184, 185, 190, 191, 192, 193, 194, 202, 208, 209, 210, 211, 212, 227, 228, 229, 245, 246], "bbox": [0.17379166666666668, 0.0045, 0.9498125000000001, 0.62021875], "iscrowd": 0, "area": 71731.50940000002, "rle": {"size": [640, 480], "counts": "hgd18hc06I7I8I6I7J7H7J6I7I3N1N3N1N3N2M2O2M2N3N1N3N2M2O2M2N3YOgLc@Z3\\?hLa@[3\\?gLc@[3Z?gLd@Z3\\?gLc@Y3\\?iLb@X3]?jLa@X3^?iLa@W3^?lL_@U3a?lL]@U3b?mL\\@T3c?f0O0O1O101N100O1O101N1O100O1O2O0O100O2N100O1N2O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1N2O2N1O1OL41SBYJ`=h5^B\\J_=e5_B`J]=b5_BcJ^=S6L3N2M2O0O2O1N2O1N2O0WMVI[Hl6a7[I[Hf6a7bIYH`6c7hIXHY6d7oIVHR6g7VJTHk5h7\\JSHf5i7bJRH_5i7jJQHW5l7QKnGQ5n7bKbG_4Z8f3L3M4L3M3L4M3M4O0O1O100O1O1O2O0O1O1000001O00000000\\M\\HlHd7R7`HmH_7R7cHmH]7R7eHmH[7S7gHkHY7T7iHkHW7T7lHiHU7V7mHiHS7V7PIhHP7W7RIgHo6Y7RIfHn6Y7UIeHk6[7VIcHk6]7WIaHi6_7XI^Hj6b7WIZHl6f7VIUHm6k7TIQHo6o7SIlGP7T8QIhGR7X8oHdGT7\\8nH_GU7a8lH[GW7e8kHVGX7j8iHQG[7o8gHlF\\7T9eHhF^7X9e10000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000ZO\\EPHd:P8^EnGb:Q8bElG^:S8eEkG[:U8hEhGX:W8kEgGU:X8oEeGQ:[8QFcGo9\\8UFaGk9^8l000O100O1O100O1O100O1O1hL^DRNb;j1bDUN_;f1fDYN[;c1jD\\NV;_1oD`NR;\\1REcNo:Y1VEfNj:U1[EjNf:R1^EnNb:m0dERO\\:j0hEVOX:e0mE[OS:a0RF^On9>VFBn95WFKn9LWF3o9BVF>o9YOUFh0P:mNVFR1o9eNUF[1Q:ZNTFf1Q:QNTFn1R:hMRFX2S:^MRFb2g=O001O000010O0001O00001O00001O0000100O2N1O1O1O2N1O1O2N1O1O2O0O1O1O1O0010O10000O10000O10000O1000O0100O0100O01a^OoMn`0P2m^OVNSa0j1f^O\\NZa0V2000O010O10O10O010O10O0100O0100O100O10O001N1O2O0O2O0O2O0O2O0O2N101N101N100O2O0^OU^OnNma0m0Y^OPOha0l0^^OROca0i0c^OTO^a0g0i^OWOXa0d0T1K6J5L5Jbh>"}, "label": "the baby boy wearing a red shirt and gray bib"}]}
{"id": 514404, "image": "COCO_train2014_000000514404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white unoccupied chair\"."}], "task": "refering", "answer_template": "The \"a white unoccupied chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 85, 102, 119, 136, 137, 153, 154, 171, 188, 205], "bbox": [0.0, 0.155671875, 0.14247916666666666, 0.5360625], "iscrowd": 0, "area": 4543.634650000001, "rle": {"size": [640, 480], "counts": "U3U2ka0[1eN2N3M2N2N2N3M2N2NfL^@^2Z?aMk@d2l>[MYAb2f>]M`A_2_>`MfA]2Y>cMkAY2U>fMPBW2o=hMVBT2j=kM[BR2d=mMbBn1^=RNfBk1Y=VNjBg1U=YNoBc1Q=]NSC`1l<`NXC^1f<bN_C_1[<aNiC`1R<`NRDa1i;_N[Db1`;^NdDc1V;^NnDc1m:]NXEc1c:]NaEe1Y:[NkEf1P:ZNTFg1g9ZN\\F0^Ll0Q=TOeFKbLo0d<VOoFEeLS1W<XOXG@iLV1j;ZOgHd0T7\\OSIc0g6]O`I=_6ChI6X6JoI0P60WJIi57^JBb5>P7001O000000000000000000000002N3N2M2N3M3M2NZ\\Q8"}, "label": "a white unoccupied chair"}]}
{"id": 514404, "image": "COCO_train2014_000000514404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white chair in the background that can be seen more of that ' s close to the cooking pot\"."}], "task": "refering", "answer_template": "The \"white chair in the background that can be seen more of that ' s close to the cooking pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 85, 102, 119, 136, 137, 153, 154, 171, 188, 205], "bbox": [0.0, 0.155671875, 0.14247916666666666, 0.5360625], "iscrowd": 0, "area": 4543.634650000001, "rle": {"size": [640, 480], "counts": "U3U2ka0[1eN2N3M2N2N2N3M2N2NfL^@^2Z?aMk@d2l>[MYAb2f>]M`A_2_>`MfA]2Y>cMkAY2U>fMPBW2o=hMVBT2j=kM[BR2d=mMbBn1^=RNfBk1Y=VNjBg1U=YNoBc1Q=]NSC`1l<`NXC^1f<bN_C_1[<aNiC`1R<`NRDa1i;_N[Db1`;^NdDc1V;^NnDc1m:]NXEc1c:]NaEe1Y:[NkEf1P:ZNTFg1g9ZN\\F0^Ll0Q=TOeFKbLo0d<VOoFEeLS1W<XOXG@iLV1j;ZOgHd0T7\\OSIc0g6]O`I=_6ChI6X6JoI0P60WJIi57^JBb5>P7001O000000000000000000000002N3N2M2N3M3M2NZ\\Q8"}, "label": "white chair in the background that can be seen more of that ' s close to the cooking pot"}]}
{"id": 514404, "image": "COCO_train2014_000000514404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a beige chair\"."}], "task": "refering", "answer_template": "The \"the back of a beige chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 72, 73, 74, 85, 86, 87, 88, 89, 90, 91, 92, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 137, 138, 139, 140, 154, 155, 171, 172, 173, 188, 189, 190, 205, 206], "bbox": [0.0193125, 0.1806875, 0.4305833333333334, 0.5388437500000001], "iscrowd": 0, "area": 19812.73835, "rle": {"size": [640, 480], "counts": "ng52mc06J6I7J6J6J7H7K5K5K5K4L3L4M3M2N3M3M3M2N3M3M3M2N3M3M3M3M2N3M6J6J7I7J5J7cAXKT=n5L4L5K4L4M4K4L3M2N2N2NO101N1O100O100O100O2O0O100O1O100O2O0O100O100O1O100O100O1hMgBiMZ=Q2QChMP=T2YCfMh<U2cCeM]<V2mCdMT<X2UDcMk;]2XD_Mi;a2ZD\\Mf;c2]DZMd;f2_DWMa;i2bDSM_;m2dDPM\\;P3fDmL[;S3hDjLX;U3lDhLT;X3nDeLS;[3i2000000O101O00000O1000000000000O10000_O`_ObM``0X2j_OdMV`0\\2n_O`MR`0_2h000O1O101N100O100O100O1O100O1N2O1O1O1O1N2O1O2N1N2O1O1O1O1N200O1O1O1O1O100O1O1O2N1O100O1O1O1O1O100O100O10000O100O2O0O100O101N1000001O0000001O001O2N2N2N2M2O2N2N2N1O2M3NYS[5"}, "label": "the back of a beige chair"}]}
{"id": 514404, "image": "COCO_train2014_000000514404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dining chair with biege cover on which a child in red dress is sitting\"."}], "task": "refering", "answer_template": "The \"a dining chair with biege cover on which a child in red dress is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 72, 73, 74, 85, 86, 87, 88, 89, 90, 91, 92, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 137, 138, 139, 140, 154, 155, 171, 172, 173, 188, 189, 190, 205, 206], "bbox": [0.0193125, 0.1806875, 0.4305833333333334, 0.5388437500000001], "iscrowd": 0, "area": 19812.73835, "rle": {"size": [640, 480], "counts": "ng52mc06J6I7J6J6J7H7K5K5K5K4L3L4M3M2N3M3M3M2N3M3M3M2N3M3M3M3M2N3M6J6J7I7J5J7cAXKT=n5L4L5K4L4M4K4L3M2N2N2NO101N1O100O100O100O2O0O100O1O100O2O0O100O100O1O100O100O1hMgBiMZ=Q2QChMP=T2YCfMh<U2cCeM]<V2mCdMT<X2UDcMk;]2XD_Mi;a2ZD\\Mf;c2]DZMd;f2_DWMa;i2bDSM_;m2dDPM\\;P3fDmL[;S3hDjLX;U3lDhLT;X3nDeLS;[3i2000000O101O00000O1000000000000O10000_O`_ObM``0X2j_OdMV`0\\2n_O`MR`0_2h000O1O101N100O100O100O1O100O1N2O1O1O1O1N2O1O2N1N2O1O1O1O1N200O1O1O1O1O100O1O1O2N1O100O1O1O1O1O100O100O10000O100O2O0O100O101N1000001O0000001O001O2N2N2N2M2O2N2N2N1O2M3NYS[5"}, "label": "a dining chair with biege cover on which a child in red dress is sitting"}]}
{"id": 203108, "image": "COCO_train2014_000000203108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train that has a b on the front of it\"."}], "task": "refering", "answer_template": "The \"a train that has a b on the front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 300, 301, 315, 316, 317, 318, 319, 320, 321, 322, 323, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382, 383, 384, 385, 386, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 423, 424, 425, 426, 427, 428, 429, 445, 446, 447], "bbox": [0.0, 0.638203125, 0.4466723549488055, 0.9438124999999999], "iscrowd": 0, "area": 36808.675100000015, "rle": {"size": [640, 586], "counts": "\\?6fc09F;F:F:E7I8G8H9G8K6M2M4M1O1N101O1N101O0O2O001N101N101O0O2O1O0O2O001N100O01000O10000O10000O01000O100O10000O10O10O10000O10000O10O100000O10000000O0101O00000O10001O0O1000001O0O1000001O0O2O001O1O0O2O001O0O2O001O001N101O1O0bA`K\\=a4lAVLS>m4O001O001N10001O001N101O001N10O100O10O10O100O10O010000O10O0100O100O01001N100O10000O10000000000O1000000000000O101O000000000O10000000000O1000000000000O101O0000000O100000000O10000000000O10001O000O10001O000O10001O000O101O00000O101O0000000O2O00004L5J6K4L5K5K7H8I7I7I7I7H8I7I7I7I7H8I6J4L4L4L4K5Ll\\O@lb0<o\\OHVc0>5L5J5K5KmmY6"}, "label": "a train that has a b on the front of it"}]}
{"id": 203108, "image": "COCO_train2014_000000203108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green passenger train car\"."}], "task": "refering", "answer_template": "The \"a green passenger train car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 300, 301, 315, 316, 317, 318, 319, 320, 321, 322, 323, 336, 337, 338, 339, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382, 383, 384, 385, 386, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 423, 424, 425, 426, 427, 428, 429, 445, 446, 447], "bbox": [0.0, 0.638203125, 0.4466723549488055, 0.9438124999999999], "iscrowd": 0, "area": 36808.675100000015, "rle": {"size": [640, 586], "counts": "\\?6fc09F;F:F:E7I8G8H9G8K6M2M4M1O1N101O1N101O0O2O001N101N101O0O2O1O0O2O001N100O01000O10000O10000O01000O100O10000O10O10O10000O10000O10O100000O10000000O0101O00000O10001O0O1000001O0O1000001O0O2O001O1O0O2O001O0O2O001O001N101O1O0bA`K\\=a4lAVLS>m4O001O001N10001O001N101O001N10O100O10O10O100O10O010000O10O0100O100O01001N100O10000O10000000000O1000000000000O101O000000000O10000000000O1000000000000O101O0000000O100000000O10000000000O10001O000O10001O000O10001O000O101O00000O101O0000000O2O00004L5J6K4L5K5K7H8I7I7I7I7H8I7I7I7I7H8I6J4L4L4L4K5Ll\\O@lb0<o\\OHVc0>5L5J5K5KmmY6"}, "label": "a green passenger train car"}]}
{"id": 203108, "image": "COCO_train2014_000000203108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and red train with the number ' 69621 ' painted on the front\"."}], "task": "refering", "answer_template": "The \"a black and red train with the number ' 69621 ' painted on the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 432, 433, 434, 435, 436, 437, 438, 439, 440, 454, 455, 456, 457, 458, 459, 460], "bbox": [0.37059726962457334, 0.511984375, 0.9669795221843003, 0.959171875], "iscrowd": 0, "area": 68684.24935, "rle": {"size": [640, 586], "counts": "nPX44lc01O1O001O001O001N2O001O001O1O001O4LW1iN>B2N2N3M2N2M4c_O]MW?e2c@_M]?c2]@aMc?b2V@cMi?_2Q@eMo?W3O00001O00001O000000000O1000000000000000000O10000O10000O100O10000O10000O10000O10000O10000O10000000000000000O100000001O001O001N101O001O001O001O0O2O0O2O0O2N101N102M3N2M3M3N2M2O1N3N1M3N2N2N3L3N2N2N2M4M1O2N2M201N2N1O2O1N1O2N2O1N2N3M3M2O0O2N100O1O100O100O1O100O1O2O0O1O10000O2O000O2O00001N2O1O1N2O00001N10000O2O000O10001N10000O2O000O1000O00100O010O1O010O01O100O1O1O1O1O2N1O1N2000001O000010O01O1O001O1O010O0000001O01O000001O0000001O000001O01O00000000O2O1O007I0O10000000000O101O0001O000000000001O0001O000000000001O01O0001O0000001O0001O01O000000001O011N1O1O1O1O1O1O1O101N1M3K5K5K5K5M3M3M4L3N2M3N2O1N2N2N2O1N2N2OW1hN101N2O001O1N101OX1hNW2hM101O000O10001OV[;"}, "label": "a black and red train with the number ' 69621 ' painted on the front"}]}
{"id": 203108, "image": "COCO_train2014_000000203108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train number 69621\"."}], "task": "refering", "answer_template": "The \"train number 69621\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 432, 433, 434, 435, 436, 437, 438, 439, 440, 454, 455, 456, 457, 458, 459, 460], "bbox": [0.37059726962457334, 0.511984375, 0.9669795221843003, 0.959171875], "iscrowd": 0, "area": 68684.24935, "rle": {"size": [640, 586], "counts": "nPX44lc01O1O001O001O001N2O001O001O1O001O4LW1iN>B2N2N3M2N2M4c_O]MW?e2c@_M]?c2]@aMc?b2V@cMi?_2Q@eMo?W3O00001O00001O000000000O1000000000000000000O10000O10000O100O10000O10000O10000O10000O10000O10000000000000000O100000001O001O001N101O001O001O001O0O2O0O2O0O2N101N102M3N2M3M3N2M2O1N3N1M3N2N2N3L3N2N2N2M4M1O2N2M201N2N1O2O1N1O2N2O1N2N3M3M2O0O2N100O1O100O100O1O100O1O2O0O1O10000O2O000O2O00001N2O1O1N2O00001N10000O2O000O10001N10000O2O000O1000O00100O010O1O010O01O100O1O1O1O1O2N1O1N2000001O000010O01O1O001O1O010O0000001O01O000001O0000001O000001O01O00000000O2O1O007I0O10000000000O101O0001O000000000001O0001O000000000001O01O0001O0000001O0001O01O000000001O011N1O1O1O1O1O1O1O101N1M3K5K5K5K5M3M3M4L3N2M3N2O1N2N2N2O1N2N2OW1hN101N2O001O1N101OX1hNW2hM101O000O10001OV[;"}, "label": "train number 69621"}]}
{"id": 268644, "image": "COCO_train2014_000000268644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an horse which goes straight in the road\"."}], "task": "refering", "answer_template": "The \"an horse which goes straight in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351, 352, 353, 373, 374, 375, 376, 396, 397, 398, 399, 419, 420, 421, 422, 441, 442, 444, 445], "bbox": [0.13082812500000002, 0.37757053291536047, 0.37854687499999995, 0.8598746081504701], "iscrowd": 0, "area": 26551.963199999995, "rle": {"size": [638, 640], "counts": "\\gd13hc04K5K5L4K5K4M4K5K5L4K5L3L5K5L4L4N2M2O2M3N2M3N2M2O2M3N1N01O001O0O2O001O001O00001O0O10000000hCXM\\7h2dHZMZ7f2eH]MY7c2gH_MW7a2hHbMW7\\2jHfMT7Z2kHiMS7W2mHkMQ7U2nHnMQ7\\2dHfMZ7g2XH\\Mg7o2mGSMR8Y3`GjL^8a3VGbLi8f3TFWMj9P3kEUMS:T3aEQM^:W3UEoLj:X3kDmLS;P6O1O1N2O1O1N2O1O1N]MTEXLk:g3`EPL^:P4mEgKR:X4YF_Ke9a4fFVKY9g3]EgKi16i8o3fEfKS2LT8\\4PFcK^2Aa7h4YFbKl<\\4ZC`Kg<\\4aC^K`<_4a14L5L3M4L3M4L3M4L3N3QD]J[9h5XFhJ`9]5RFPKj9U5hEUKW:P5PEdKP;e600000000001O0_LlDXNT;e1REXNn:e1XEXNh:e1]EYNc:d1cEYN]:d1iEZNV:d1oESLQO<NDR;j3UFmK[O0L6d:k3YFfKb0<U9o3bHnK^7R4eHkK\\7T4gHiKY7W4jHfKV7[4V400000kDhKd6Y4mFUNS9k1iFYNW9g1eF]N[9g1^F]Na9Q2mESNT:Z2ZEjMf:h5O0002N100O2N100O2RIQEi4P;TKWEg4j:VK]Ef4c:VKeEe4]:WKkEc4V:ZKQFb4n9\\KYF_4g9^KaF]4_9`KiFd3oMnKX;<oFb3mMoKT;;VGa3kMPLo:=[G`3jMoKk:>bGb1YM]O>aNf:>iG]1[M_O:bNb:?PHY1\\M@`;UOZGU1]M@Z;XOUH<iL7S;YO]H7kL7m?@_@7la0G__g7"}, "label": "an horse which goes straight in the road"}]}
{"id": 268644, "image": "COCO_train2014_000000268644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black horse which is ahead of the brown horse\"."}], "task": "refering", "answer_template": "The \"the black horse which is ahead of the brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330, 348, 349, 350, 351, 352, 353, 373, 374, 375, 376, 396, 397, 398, 399, 419, 420, 421, 422, 441, 442, 444, 445], "bbox": [0.13082812500000002, 0.37757053291536047, 0.37854687499999995, 0.8598746081504701], "iscrowd": 0, "area": 26551.963199999995, "rle": {"size": [638, 640], "counts": "\\gd13hc04K5K5L4K5K4M4K5K5L4K5L3L5K5L4L4N2M2O2M3N2M3N2M2O2M3N1N01O001O0O2O001O001O00001O0O10000000hCXM\\7h2dHZMZ7f2eH]MY7c2gH_MW7a2hHbMW7\\2jHfMT7Z2kHiMS7W2mHkMQ7U2nHnMQ7\\2dHfMZ7g2XH\\Mg7o2mGSMR8Y3`GjL^8a3VGbLi8f3TFWMj9P3kEUMS:T3aEQM^:W3UEoLj:X3kDmLS;P6O1O1N2O1O1N2O1O1N]MTEXLk:g3`EPL^:P4mEgKR:X4YF_Ke9a4fFVKY9g3]EgKi16i8o3fEfKS2LT8\\4PFcK^2Aa7h4YFbKl<\\4ZC`Kg<\\4aC^K`<_4a14L5L3M4L3M4L3M4L3N3QD]J[9h5XFhJ`9]5RFPKj9U5hEUKW:P5PEdKP;e600000000001O0_LlDXNT;e1REXNn:e1XEXNh:e1]EYNc:d1cEYN]:d1iEZNV:d1oESLQO<NDR;j3UFmK[O0L6d:k3YFfKb0<U9o3bHnK^7R4eHkK\\7T4gHiKY7W4jHfKV7[4V400000kDhKd6Y4mFUNS9k1iFYNW9g1eF]N[9g1^F]Na9Q2mESNT:Z2ZEjMf:h5O0002N100O2N100O2RIQEi4P;TKWEg4j:VK]Ef4c:VKeEe4]:WKkEc4V:ZKQFb4n9\\KYF_4g9^KaF]4_9`KiFd3oMnKX;<oFb3mMoKT;;VGa3kMPLo:=[G`3jMoKk:>bGb1YM]O>aNf:>iG]1[M_O:bNb:?PHY1\\M@`;UOZGU1]M@Z;XOUH<iL7S;YO]H7kL7m?@_@7la0G__g7"}, "label": "the black horse which is ahead of the brown horse"}]}
{"id": 268644, "image": "COCO_train2014_000000268644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse who ' s face can be seen\"."}], "task": "refering", "answer_template": "The \"a horse who ' s face can be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 249, 250, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 335, 336, 337, 354, 355, 356, 359, 360, 377, 378, 379, 382, 383, 400, 401, 402, 404, 405, 406, 423, 424, 425, 427, 428, 429, 446, 447, 448, 450, 451, 452], "bbox": [0.350828125, 0.2545141065830721, 0.8984687499999999, 0.8703918495297804], "iscrowd": 0, "area": 52904.8677, "rle": {"size": [638, 640], "counts": "`_\\47dc09H7I8H8H8G9H8H7I8H5J7J6J6J5J7J6J6J5K6I7K5K4M4L4L4L3M4L5K7H8I7I8H7I7I7I8H7H8[DnHP:Y7lEdHT:c7hEYHY:o7bEnG^:Y8^EdGb:k81000O010000O10000O1000O01jN]EQIc:n6cEmH]:R7iEhHX:W7nEdHR:\\7SF_Hm9`7YFZHh9e7^FVHa9j7Z1O100O1O2N1O100O\\DaHk:^7QEiHm:V7nDRIP;m6lD[IQ;d6kDcIT;e7N2N2N2N3M2NO1O2N100O1dLUEdMl:X2eEZM\\:c2TFoLm9m2_FjLb9R3fFiL\\9S3lFgLU9U3TGfLl8V3]GdLd8Y3cGbL^8Z3kG`LV8\\3SH^Ln7_3ZH[Lg7a3aHZLa7b3gHYLY7c3PIWLQ7f3e40000000001O0000000000000O2O000000000000001O000000000O101O001O001O001O001O001O001N101O1O001O001O001OlBUNU8k1jGVNW8i1gGYNY8g1eG[N\\8d1cG]N]8b1bG_N`8a1]GaNc8c1XG^Ni8e1QG]NP9g1iF[NW9i1dFWN^9l1\\FVNd9o1UFSNl9P2oEQNQ:S2iEoMX:\\2ZEeMg:g2lDZMU;R3\\DPMd;\\3nCfLS<a5101O00f2ZM0O10000J`KVKYJb4m5cKoIU4X6oKeIS4Y6QLcIQ4Z6SLdIo3X6VLdIm3Y6VLdIm3X6XLdIk3X6YLeIj3W6ZLgIg3W6\\LfIf3W6^LfIb3Y6cLcI]3\\6gLaIY3_6jL_IU3`6oL]IQ3b6SM[In2d6UMYIi1`McLV9h1XIR1UNUMb8m1VI=iNfMP8Q2TIG_OWN]7U2QI]OJ]NT7Z2oHUO2aNm6^2oHmN8dNh6c2mHfN?fNd6h2iH_Ng0hN_6Q3cHTNR1jNZ6c3UH_Md1nNV6U4eGiLZ2ROP6_4\\G\\Lh2TOk5d4ZGULo2VOf5i4XGnKV3XOb5m4VGgK\\3[O]5R5TG_Kd3^OW5W5RGXKk3AQ5[5QGQKR4Cm4_5oFjJW4Gi4c5mFcJ^4Id4o5`KPJ_4m5fKRJZ4k5jKUJT4i5Q5N3M2M3N2N3M2N2M3N3M2N2N3M2M3N3M3M2N3L4M3M3M2N3M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1O2N2M3N2N1N3N2N2M22O1O1O2N1N2O1O1O1N2O1O2N1O1N2O1O1O1N3N1O1O1N2O1O1N3N1O1N2O1O1O1N2O2N1N2O1O1002N2N2N2M4M2M3N2M3N2N2M3N3L3N2M3N2M3N2N2M3L4M3L4L4M3L3M4M3L4L4M3L4L4M3L3N3L4L4M3LSWX1"}, "label": "a horse who ' s face can be seen"}]}
{"id": 268644, "image": "COCO_train2014_000000268644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse looking at the cycles\"."}], "task": "refering", "answer_template": "The \"the horse looking at the cycles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 249, 250, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 335, 336, 337, 354, 355, 356, 359, 360, 377, 378, 379, 382, 383, 400, 401, 402, 404, 405, 406, 423, 424, 425, 427, 428, 429, 446, 447, 448, 450, 451, 452], "bbox": [0.350828125, 0.2545141065830721, 0.8984687499999999, 0.8703918495297804], "iscrowd": 0, "area": 52904.8677, "rle": {"size": [638, 640], "counts": "`_\\47dc09H7I8H8H8G9H8H7I8H5J7J6J6J5J7J6J6J5K6I7K5K4M4L4L4L3M4L5K7H8I7I8H7I7I7I8H7H8[DnHP:Y7lEdHT:c7hEYHY:o7bEnG^:Y8^EdGb:k81000O010000O10000O1000O01jN]EQIc:n6cEmH]:R7iEhHX:W7nEdHR:\\7SF_Hm9`7YFZHh9e7^FVHa9j7Z1O100O1O2N1O100O\\DaHk:^7QEiHm:V7nDRIP;m6lD[IQ;d6kDcIT;e7N2N2N2N3M2NO1O2N100O1dLUEdMl:X2eEZM\\:c2TFoLm9m2_FjLb9R3fFiL\\9S3lFgLU9U3TGfLl8V3]GdLd8Y3cGbL^8Z3kG`LV8\\3SH^Ln7_3ZH[Lg7a3aHZLa7b3gHYLY7c3PIWLQ7f3e40000000001O0000000000000O2O000000000000001O000000000O101O001O001O001O001O001O001N101O1O001O001O001OlBUNU8k1jGVNW8i1gGYNY8g1eG[N\\8d1cG]N]8b1bG_N`8a1]GaNc8c1XG^Ni8e1QG]NP9g1iF[NW9i1dFWN^9l1\\FVNd9o1UFSNl9P2oEQNQ:S2iEoMX:\\2ZEeMg:g2lDZMU;R3\\DPMd;\\3nCfLS<a5101O00f2ZM0O10000J`KVKYJb4m5cKoIU4X6oKeIS4Y6QLcIQ4Z6SLdIo3X6VLdIm3Y6VLdIm3X6XLdIk3X6YLeIj3W6ZLgIg3W6\\LfIf3W6^LfIb3Y6cLcI]3\\6gLaIY3_6jL_IU3`6oL]IQ3b6SM[In2d6UMYIi1`McLV9h1XIR1UNUMb8m1VI=iNfMP8Q2TIG_OWN]7U2QI]OJ]NT7Z2oHUO2aNm6^2oHmN8dNh6c2mHfN?fNd6h2iH_Ng0hN_6Q3cHTNR1jNZ6c3UH_Md1nNV6U4eGiLZ2ROP6_4\\G\\Lh2TOk5d4ZGULo2VOf5i4XGnKV3XOb5m4VGgK\\3[O]5R5TG_Kd3^OW5W5RGXKk3AQ5[5QGQKR4Cm4_5oFjJW4Gi4c5mFcJ^4Id4o5`KPJ_4m5fKRJZ4k5jKUJT4i5Q5N3M2M3N2N3M2N2M3N3M2N2N3M2M3N3M3M2N3L4M3M3M2N3M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1O2N2M3N2N1N3N2N2M22O1O1O2N1N2O1O1O1N2O1O2N1O1N2O1O1O1N3N1O1O1N2O1O1N3N1O1N2O1O1O1N2O2N1N2O1O1002N2N2N2M4M2M3N2M3N2N2M3N3L3N2M3N2M3N2N2M3L4M3L4L4M3L3M4M3L4L4M3L4L4M3L3N3L4L4M3LSWX1"}, "label": "the horse looking at the cycles"}]}
{"id": 112996, "image": "COCO_train2014_000000112996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink on the left\"."}], "task": "refering", "answer_template": "The \"the sink on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 379, 380, 381, 382, 383, 384, 385, 404, 405], "bbox": [0.22774509803921567, 0.6835947712418301, 0.6050980392156863, 0.8270915032679739], "iscrowd": 0, "area": 10822.705649999998, "rle": {"size": [612, 612], "counts": "STd21Sc01O0O2O0O2O001N101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O2O001N101N101O0O2O001N101N101O0O2O001N100O10000O100O101O0O10000O100O10001N10000O100O10000O2O0O10000O10000O101N10000O10000O010O10O10O0100O010O01000O010O010O10O11N100O2O0O100O2O000O101N100O2O0O10001N100O2O0O100O2O000O2O0O100O2O0O101O0O100O2O0O100O2O000O2O0O100O2O0O101O0O100O2O0O101N10000O2O0O101N100O101O0O101N100O101N10000O2O0O101N100O101O0O101N2O1N2O1N2O1O1NUV`4"}, "label": "the sink on the left"}]}
{"id": 112996, "image": "COCO_train2014_000000112996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a square sink to the left of another sink\"."}], "task": "refering", "answer_template": "The \"a square sink to the left of another sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 379, 380, 381, 382, 383, 384, 385, 404, 405], "bbox": [0.22774509803921567, 0.6835947712418301, 0.6050980392156863, 0.8270915032679739], "iscrowd": 0, "area": 10822.705649999998, "rle": {"size": [612, 612], "counts": "STd21Sc01O0O2O0O2O001N101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O2O001N101N101O0O2O001N101N101O0O2O001N100O10000O100O101O0O10000O100O10001N10000O100O10000O2O0O10000O10000O101N10000O10000O010O10O10O0100O010O01000O010O010O10O11N100O2O0O100O2O000O101N100O2O0O10001N100O2O0O100O2O000O2O0O100O2O0O101O0O100O2O0O100O2O000O2O0O100O2O0O101O0O100O2O0O101N10000O2O0O101N100O101O0O101N100O101N10000O2O0O101N100O101O0O101N2O1N2O1N2O1O1NUV`4"}, "label": "a square sink to the left of another sink"}]}
{"id": 301413, "image": "COCO_train2014_000000301413.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with light brown hair bends over a counter while filling out a form at a wine tasting\"."}], "task": "refering", "answer_template": "The \"a man with light brown hair bends over a counter while filling out a form at a wine tasting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 11, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 82, 83, 84, 85, 86, 87, 88, 89, 90, 97, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 127, 128, 129, 140, 141, 142, 154, 155, 167, 168, 177, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233], "bbox": [0.26066666666666666, 0.00898, 1.0, 0.97978], "iscrowd": 0, "area": 78866.42999999998, "rle": {"size": [500, 375], "counts": "hm_1i1h=6I6J6J6J6K6I6J6J6J6K5K3O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2nMUKXIl4f6WKXIk4d6YKTIFcNR5W8[KUICcNT5T8]KXI@bNT5n6SKmI<b0f4_5cL_J]3a5dL^J\\3a5fL^J[3a5fL^JZ3b5gL]JZ3a5hL^JX3b5jL\\JW3b5jL^JV3b5jL^JW3`5kL_JU3a5kL_JV3`5jL`JV3_5kLaJV3^5jLbJV3]5lLbJT3^5lLbJU3\\5lLdJT3\\5lLdJU3[5kLeJU3Z5mLeJT3Z5lLfJT3Z5lLfJU3Y5kLgJU3Y5kLgJV3X5kLgJU3Y5kLgJU3Y5kLgJV3X5jLhJV3X5jLhJW3W5jLhJV3X5jLhJW3W5hLjJX3V5hLjJY3U5fLlJZ3T5eLmJ\\3R5dLnJ]3Q5bLPK`3n4`LRKc3k4\\LVKf3i4YLWKj3f4UL[Km3c4RL^KQ4_4oKaKS4]4lKdKW4Y4iKgKY4W4fKjK]4S4bKnK^4R4bKnK^4S4`KnK`4R4`KnK`4S4^KnKb4R4]KoKc4R4\\KnKd4R4[KoKe4Q4[KoKe4R4YKoKg4Q4YKoKg4R4XKnKh4R4XKnKh4S4WKmKi4S4WKmKi4T4VKlKj4T4VKlKj4U4TKlKl4T4TKlKl4T4TKlKl4U4SKkKm4U4SKkKm4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4QKkKo4U4QKkKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKQ5R4oJoKQ5Q4oJoKQ5P4PKPLP5P4PKPLP5o3QKQLo4o3QKQLo4n3RKRLn4n3RKRLn4m3SKSLm4m3SKSLm4l3TKTLl4l3TKTLl4k3TKVLl4j3TKVLl4i3UKWLk4i3UKWLk4h3VKXLj4h3VKXLk4f3VKZLj4f3VKZLj4e3WK[Li4e3WK[Li4d3XK\\Lh4d3WK]Li4b3XK^Lh4b3XK^Lh4a3YK_Lg4a3YK_Lg4`3ZK`Lf4`3ZK`Lf4_3[KaLe4_3[KaLe4^3\\KbLd4^3\\KbLd4]3]KcLc4]3\\KdLd4[3]KeLc4[3]KeLc4Z3^KfLb4Z3^KfLb4Z3^KfLb4Y3_KgLa4Y3_KgLa4X3`KhL`4X3`KhL`4W3aKiL`4V3_KkLc4S3]KmLe4P3\\KQMe4m2[KTMf4i2[KXMf4f2ZK[Mg4c2YK^Mg4`2ZKaMg4]2YKeMg4X2ZKiMg4U2YKlMh4R2XKoMi4n1WKTNj4j1VKWNk4f1VK\\Ni4c1WK^Nj4_1WKbNj4\\1VKeNk4Y1UKhNl4U1UKlNl4R1TKPOl4m0UKTOk4k0UKVOl4h0SKZOn4c0SK^On4`0RKAo4<RKEo49QKIo44RKMo41QK0o4OPK3Q5JPK7Q5GoJ:R5CoJ?Q5_OoJb0R5\\OmJf0T5WOmJj0S5UOmJl0T5QOmJP1T5nNmJR1T5lNmJR1V5kNlJR1Y5kNhJS1[5jNgJT1]5iNdJU1`5UNjG]Oh2\\2b5QNmG_Ob2]2e5nMQHA[2_2h5iMVKU2n4eMVKY2T9N2N2M3N2N2N2N2M3N2N2N2N2M210O1000000O10O10O100000O01000000O01000O10O1000O1000O1O001O1[Od0hNY1fNZ1gNY1gNX1hNY1gNY1gNY1gNX1gNZ1gNH"}, "label": "a man with light brown hair bends over a counter while filling out a form at a wine tasting"}]}
{"id": 301413, "image": "COCO_train2014_000000301413.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man looking down at a paper and writing with a pen\"."}], "task": "refering", "answer_template": "The \"a man looking down at a paper and writing with a pen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 11, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 82, 83, 84, 85, 86, 87, 88, 89, 90, 97, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 127, 128, 129, 140, 141, 142, 154, 155, 167, 168, 177, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233], "bbox": [0.26066666666666666, 0.00898, 1.0, 0.97978], "iscrowd": 0, "area": 78866.42999999998, "rle": {"size": [500, 375], "counts": "hm_1i1h=6I6J6J6J6K6I6J6J6J6K5K3O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2nMUKXIl4f6WKXIk4d6YKTIFcNR5W8[KUICcNT5T8]KXI@bNT5n6SKmI<b0f4_5cL_J]3a5dL^J\\3a5fL^J[3a5fL^JZ3b5gL]JZ3a5hL^JX3b5jL\\JW3b5jL^JV3b5jL^JW3`5kL_JU3a5kL_JV3`5jL`JV3_5kLaJV3^5jLbJV3]5lLbJT3^5lLbJU3\\5lLdJT3\\5lLdJU3[5kLeJU3Z5mLeJT3Z5lLfJT3Z5lLfJU3Y5kLgJU3Y5kLgJV3X5kLgJU3Y5kLgJU3Y5kLgJV3X5jLhJV3X5jLhJW3W5jLhJV3X5jLhJW3W5hLjJX3V5hLjJY3U5fLlJZ3T5eLmJ\\3R5dLnJ]3Q5bLPK`3n4`LRKc3k4\\LVKf3i4YLWKj3f4UL[Km3c4RL^KQ4_4oKaKS4]4lKdKW4Y4iKgKY4W4fKjK]4S4bKnK^4R4bKnK^4S4`KnK`4R4`KnK`4S4^KnKb4R4]KoKc4R4\\KnKd4R4[KoKe4Q4[KoKe4R4YKoKg4Q4YKoKg4R4XKnKh4R4XKnKh4S4WKmKi4S4WKmKi4T4VKlKj4T4VKlKj4U4TKlKl4T4TKlKl4T4TKlKl4U4SKkKm4U4SKkKm4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4RKjKn4V4QKkKo4U4QKkKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKP5T4PKlKQ5R4oJoKQ5Q4oJoKQ5P4PKPLP5P4PKPLP5o3QKQLo4o3QKQLo4n3RKRLn4n3RKRLn4m3SKSLm4m3SKSLm4l3TKTLl4l3TKTLl4k3TKVLl4j3TKVLl4i3UKWLk4i3UKWLk4h3VKXLj4h3VKXLk4f3VKZLj4f3VKZLj4e3WK[Li4e3WK[Li4d3XK\\Lh4d3WK]Li4b3XK^Lh4b3XK^Lh4a3YK_Lg4a3YK_Lg4`3ZK`Lf4`3ZK`Lf4_3[KaLe4_3[KaLe4^3\\KbLd4^3\\KbLd4]3]KcLc4]3\\KdLd4[3]KeLc4[3]KeLc4Z3^KfLb4Z3^KfLb4Z3^KfLb4Y3_KgLa4Y3_KgLa4X3`KhL`4X3`KhL`4W3aKiL`4V3_KkLc4S3]KmLe4P3\\KQMe4m2[KTMf4i2[KXMf4f2ZK[Mg4c2YK^Mg4`2ZKaMg4]2YKeMg4X2ZKiMg4U2YKlMh4R2XKoMi4n1WKTNj4j1VKWNk4f1VK\\Ni4c1WK^Nj4_1WKbNj4\\1VKeNk4Y1UKhNl4U1UKlNl4R1TKPOl4m0UKTOk4k0UKVOl4h0SKZOn4c0SK^On4`0RKAo4<RKEo49QKIo44RKMo41QK0o4OPK3Q5JPK7Q5GoJ:R5CoJ?Q5_OoJb0R5\\OmJf0T5WOmJj0S5UOmJl0T5QOmJP1T5nNmJR1T5lNmJR1V5kNlJR1Y5kNhJS1[5jNgJT1]5iNdJU1`5UNjG]Oh2\\2b5QNmG_Ob2]2e5nMQHA[2_2h5iMVKU2n4eMVKY2T9N2N2M3N2N2N2N2M3N2N2N2N2M210O1000000O10O10O100000O01000000O01000O10O1000O1000O1O001O1[Od0hNY1fNZ1gNY1gNX1hNY1gNY1gNY1gNX1gNZ1gNH"}, "label": "a man looking down at a paper and writing with a pen"}]}
{"id": 301413, "image": "COCO_train2014_000000301413.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person with less hair in the picture\"."}], "task": "refering", "answer_template": "The \"the person with less hair in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 13, 14, 15, 16, 26, 27, 28, 29, 39, 40, 41, 52, 53, 54, 55, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 91, 92, 93, 104, 105, 106, 117, 118, 119, 130, 131, 143], "bbox": [0.0, 0.0028599999999999997, 0.3360533333333333, 0.65964], "iscrowd": 0, "area": 22658.60185, "rle": {"size": [500, 375], "counts": "o1U3^<\\2dMe0\\Oe0ZOe0\\Od0[O:F01O00010O001O00001O00001O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1OH8E;I7N2N2N2N2O10010O000001O001O1O1O1O1O1O100O0D=_Oa0_Oa0UOk0D<J600cLZFl0f9SOeFc0[9]OhF`0X9@jF=W9ClF:T9FnF8R9HQG5o8KSG3m8MVG0j8OYGOf82]GJd86_GGa89aGE_8;dGB\\8>fG_O[8a0hGXO\\8h0fGRO^8n0cGmNa8S1_GiNe8W1[GhNf8X1ZGhNf8X1ZGhNf8X1ZGhNf8X1ZGhNf8W1[GiNe8W1[GjNd8V1\\GjNd8T1^GlNb8R1`GnN`8Q1aGoN_8o0cGQO]8m0eGTO\\8g0hGXOY8e0kGYOV8c0nG\\OS8a0PH^OR8>RH@o7<UHDk74^HJc7HkH7V7[OXId0n:N1O2O0O1O2N1O2M2O1O2N1N3N1O1O2M2O\\]i3"}, "label": "the person with less hair in the picture"}]}
{"id": 301413, "image": "COCO_train2014_000000301413.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt behind a red vase\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt behind a red vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 13, 14, 15, 16, 26, 27, 28, 29, 39, 40, 41, 52, 53, 54, 55, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 91, 92, 93, 104, 105, 106, 117, 118, 119, 130, 131, 143], "bbox": [0.0, 0.0028599999999999997, 0.3360533333333333, 0.65964], "iscrowd": 0, "area": 22658.60185, "rle": {"size": [500, 375], "counts": "o1U3^<\\2dMe0\\Oe0ZOe0\\Od0[O:F01O00010O001O00001O00001O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1OH8E;I7N2N2N2N2O10010O000001O001O1O1O1O1O1O100O0D=_Oa0_Oa0UOk0D<J600cLZFl0f9SOeFc0[9]OhF`0X9@jF=W9ClF:T9FnF8R9HQG5o8KSG3m8MVG0j8OYGOf82]GJd86_GGa89aGE_8;dGB\\8>fG_O[8a0hGXO\\8h0fGRO^8n0cGmNa8S1_GiNe8W1[GhNf8X1ZGhNf8X1ZGhNf8X1ZGhNf8X1ZGhNf8W1[GiNe8W1[GjNd8V1\\GjNd8T1^GlNb8R1`GnN`8Q1aGoN_8o0cGQO]8m0eGTO\\8g0hGXOY8e0kGYOV8c0nG\\OS8a0PH^OR8>RH@o7<UHDk74^HJc7HkH7V7[OXId0n:N1O2O0O1O2N1O2M2O1O2N1N3N1O1O2M2O\\]i3"}, "label": "a man in a black shirt behind a red vase"}]}
{"id": 22890, "image": "COCO_train2014_000000022890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a nicely made bed next two a window\"."}], "task": "refering", "answer_template": "The \"a nicely made bed next two a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 339, 340, 341, 342, 343, 364], "bbox": [0.2118300653594771, 0.4715032679738562, 0.7056862745098039, 0.7459967320261438], "iscrowd": 0, "area": 24050.047049999994, "rle": {"size": [612, 612], "counts": "bP^2Q1Rb03N3L3N3L3N000001O1O001O00O2N1O1O1O1O1O1O1O100M4FXNS_Ok1l`07U_OmMe`0S2[_OUN]`0k1b_O[NX`0f1g_O[NY`0e1g_O[NY`0e1f_O\\NY`0X200O2O000O1O1O1O1N2O2N1O1O1O1O1O1O100O2N1000000000000010O001O1O001O00100O001O001O001O00100O001O000000010O0000000000O2O00000O10000O101O000O100O10001O01O001O001O00100O001O1O6J3M01N100O10000O1003N2M3M2N3M3N2N2O1N2N1O10N1O1O1O2O0O1O101N1O1O1O0000000000001O01O00000000000001O00000000000001N10000O10000O100O10000O10000O101N100001O1O1O1O1O001O1O1O1O00100O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10O01O1O1O00O1N2O1N2N2O1N2O1mNQ@oNQ`0o0T@mNm?R1X@hNj?X1Y@dNh?\\1\\@^Nf?a1o000O10000O100O10000O100O2O0O10000O100O10000O100O10000O100O100O10001O0O2O001O0O2O1O001O0O2O001O001N3N1O2N1O1N3N1O3M3LiV[3"}, "label": "a nicely made bed next two a window"}]}
{"id": 22890, "image": "COCO_train2014_000000022890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bed near to the window\"."}], "task": "refering", "answer_template": "The \"bed near to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 339, 340, 341, 342, 343, 364], "bbox": [0.2118300653594771, 0.4715032679738562, 0.7056862745098039, 0.7459967320261438], "iscrowd": 0, "area": 24050.047049999994, "rle": {"size": [612, 612], "counts": "bP^2Q1Rb03N3L3N3L3N000001O1O001O00O2N1O1O1O1O1O1O1O100M4FXNS_Ok1l`07U_OmMe`0S2[_OUN]`0k1b_O[NX`0f1g_O[NY`0e1g_O[NY`0e1f_O\\NY`0X200O2O000O1O1O1O1N2O2N1O1O1O1O1O1O100O2N1000000000000010O001O1O001O00100O001O001O001O00100O001O000000010O0000000000O2O00000O10000O101O000O100O10001O01O001O001O00100O001O1O6J3M01N100O10000O1003N2M3M2N3M3N2N2O1N2N1O10N1O1O1O2O0O1O101N1O1O1O0000000000001O01O00000000000001O00000000000001N10000O10000O100O10000O10000O101N100001O1O1O1O1O001O1O1O1O00100O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10O01O1O1O00O1N2O1N2N2O1N2O1mNQ@oNQ`0o0T@mNm?R1X@hNj?X1Y@dNh?\\1\\@^Nf?a1o000O10000O100O10000O100O2O0O10000O100O10000O100O10000O100O100O10001O0O2O001O0O2O1O001O0O2O001O001N3N1O2N1O1N3N1O3M3LiV[3"}, "label": "bed near to the window"}]}
{"id": 22890, "image": "COCO_train2014_000000022890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed on the left that you can ' t see fully\"."}], "task": "refering", "answer_template": "The \"the bed on the left that you can ' t see fully\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 264, 265, 266, 286, 287, 288, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 396, 397, 398, 399, 400, 419, 420, 421, 422, 442, 443, 444, 465, 466], "bbox": [0.03895424836601307, 0.534281045751634, 0.4701307189542483, 0.9836437908496732], "iscrowd": 0, "area": 30928.721349999993, "rle": {"size": [612, 612], "counts": "^e>i2X`0c3]L4N101O001O1O001O001O010O001O00001O00001O00001O001O00001O001O\\InCV6S<eISDZ6l;cIYD[6[<M2O2M2O1N3M2O1O2N1N2O2N1O1N2O00O1O10O0100O1O10O0100O100O00100O10O0100O1O100O100O100O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1OK5H8H8I7H8I7H9H7H8H8I7H8I7O1O100O100O100O101N10000O100O100O100O10000O100O2O0O100O100O10000O100O100O101N100O10000O100O100O100O10000O2O0O100O100O100O10000O101N100O101N100O101O0O100O2O0O100O2O000O101N100O101N100O101O0O100O2O0O100O2O0O10000O2O0O100O2O0O10001N100O101N100O101N10000O2O0O100O2O0001O01O1O2N1O2N2N101N2N1O2N2N1O2O1N1O2N1O2NeiQ6"}, "label": "the bed on the left that you can ' t see fully"}]}
{"id": 22890, "image": "COCO_train2014_000000022890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed on the left\"."}], "task": "refering", "answer_template": "The \"the bed on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 264, 265, 266, 286, 287, 288, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 396, 397, 398, 399, 400, 419, 420, 421, 422, 442, 443, 444, 465, 466], "bbox": [0.03895424836601307, 0.534281045751634, 0.4701307189542483, 0.9836437908496732], "iscrowd": 0, "area": 30928.721349999993, "rle": {"size": [612, 612], "counts": "^e>i2X`0c3]L4N101O001O1O001O001O010O001O00001O00001O00001O001O00001O001O\\InCV6S<eISDZ6l;cIYD[6[<M2O2M2O1N3M2O1O2N1N2O2N1O1N2O00O1O10O0100O1O10O0100O100O00100O10O0100O1O100O100O100O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1OK5H8H8I7H8I7H9H7H8H8I7H8I7O1O100O100O100O101N10000O100O100O100O10000O100O2O0O100O100O10000O100O100O101N100O10000O100O100O100O10000O2O0O100O100O100O10000O101N100O101N100O101O0O100O2O0O100O2O000O101N100O101N100O101O0O100O2O0O100O2O0O10000O2O0O100O2O0O10001N100O101N100O101N10000O2O0O100O2O0001O01O1O2N1O2N2N101N2N1O2N2N1O2O1N1O2N1O2NeiQ6"}, "label": "the bed on the left"}]}
{"id": 530796, "image": "COCO_train2014_000000530796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a base ball player ready to throw the ball to the batsmen\"."}], "task": "refering", "answer_template": "The \"a base ball player ready to throw the ball to the batsmen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 334], "bbox": [0.34045312499999997, 0.13272916666666668, 0.5494375, 0.8743124999999999], "iscrowd": 0, "area": 24580.680049999995, "rle": {"size": [480, 640], "counts": "Q\\V32l>2M2O2M3M3N1N3M3N1N3N2M2N3N2M2N3N2M2N3N2M2O2M3M201O10O01O1O101N2N2O1N2N2N3N1N2N2N2O1Nl2TM7I7J5J7I7I7J6I7_On2SM:F:G9K500000O1000O010000O100O10000O100O10000O10001N1B?@`0@`0_Oa0@`0@`0@`0_Oa0F:5K4L5J5M4K4L6J5K5K7J6I8H5K2N2N2O0h0YOO1O1N101O00000O1000000ObN\\I`Jd6i4SJWKm5l0RJ=a0fN]5i0mJ]O4GQ5h0gK[NGk0b4g0^NVOd1f0aNWOa1e0dNYO]1=mNAT10]ONd0AN<^7M3N2M3N2M2NYcV4"}, "label": "a base ball player ready to throw the ball to the batsmen"}]}
{"id": 530796, "image": "COCO_train2014_000000530796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in a white uniform preparing to throw the ball\"."}], "task": "refering", "answer_template": "The \"a baseball player in a white uniform preparing to throw the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 334], "bbox": [0.34045312499999997, 0.13272916666666668, 0.5494375, 0.8743124999999999], "iscrowd": 0, "area": 24580.680049999995, "rle": {"size": [480, 640], "counts": "Q\\V32l>2M2O2M3M3N1N3M3N1N3N2M2N3N2M2N3N2M2N3N2M2O2M3M201O10O01O1O101N2N2O1N2N2N3N1N2N2N2O1Nl2TM7I7J5J7I7I7J6I7_On2SM:F:G9K500000O1000O010000O100O10000O100O10000O10001N1B?@`0@`0_Oa0@`0@`0@`0_Oa0F:5K4L5J5M4K4L6J5K5K7J6I8H5K2N2N2O0h0YOO1O1N101O00000O1000000ObN\\I`Jd6i4SJWKm5l0RJ=a0fN]5i0mJ]O4GQ5h0gK[NGk0b4g0^NVOd1f0aNWOa1e0dNYO]1=mNAT10]ONd0AN<^7M3N2M3N2M2NYcV4"}, "label": "a baseball player in a white uniform preparing to throw the ball"}]}
{"id": 358767, "image": "COCO_train2014_000000358767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and silver motorbike with the number 7 on it\"."}], "task": "refering", "answer_template": "The \"a red and silver motorbike with the number 7 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 113, 122, 123, 124, 125, 126, 134, 135, 136, 137, 138, 149, 150], "bbox": [0.1681081081081081, 0.52992, 0.5860060060060059, 0.71274], "iscrowd": 0, "area": 7431.863399999999, "rle": {"size": [500, 333], "counts": "]ck07\\?3RAKX>7fANV>4eA3V>OgA<o=g0N2N2N2N1O1O00001O01O01O01O100O1O1O100O1O0O101O001O0O10001OO100000000000001O0O10001O001O001O010O001O0000001O000000010O00001O000001O0O1000000O1O1O1N2N2O2O0O10O0100O100O1002M5L4K5K5K3N1N2N2O2O0O100O100O10O0010O01O01O01_NiBc0W=\\OkBc0V=ZOlBf0U=WOnBh0S=VOnBk0S=POQCo0Q=jNTCV1n<dNWC[1a=01O001O001O001O002N1N3M2N3N1N3M2K6EWRS2"}, "label": "a red and silver motorbike with the number 7 on it"}]}
{"id": 358767, "image": "COCO_train2014_000000358767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike which is to take a ride\"."}], "task": "refering", "answer_template": "The \"a bike which is to take a ride\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 113, 122, 123, 124, 125, 126, 134, 135, 136, 137, 138, 149, 150], "bbox": [0.1681081081081081, 0.52992, 0.5860060060060059, 0.71274], "iscrowd": 0, "area": 7431.863399999999, "rle": {"size": [500, 333], "counts": "]ck07\\?3RAKX>7fANV>4eA3V>OgA<o=g0N2N2N2N1O1O00001O01O01O01O100O1O1O100O1O0O101O001O0O10001OO100000000000001O0O10001O001O001O010O001O0000001O000000010O00001O000001O0O1000000O1O1O1N2N2O2O0O10O0100O100O1002M5L4K5K5K3N1N2N2O2O0O100O100O10O0010O01O01O01_NiBc0W=\\OkBc0V=ZOlBf0U=WOnBh0S=VOnBk0S=POQCo0Q=jNTCV1n<dNWC[1a=01O001O001O001O002N1N3M2N3N1N3M2K6EWRS2"}, "label": "a bike which is to take a ride"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby carrier behind the little girl\"."}], "task": "refering", "answer_template": "The \"a baby carrier behind the little girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 125, 126, 127, 128, 129, 149, 150, 151, 152, 173, 174, 175, 196, 197, 198, 219, 220, 221, 242, 243], "bbox": [0.44934375, 0.305607476635514, 0.6206562499999999, 0.7146028037383179], "iscrowd": 0, "area": 9256.5596, "rle": {"size": [428, 640], "counts": "T`h33Y=4L4L3M1O1O001O1O001iC_Oa;b0\\D@c;a0\\D@d;a0YDAg;`0WDAi;?UDCk;>RDDn;=PDCQ<l0000O100O10000O10000O11O1O1O1O1UDkN];c1O1O1O1O1O1O1O2O0O101N3N2M3N2M3N2M3M3N2M3N2M3N1N3N2M3N2M3N2M3M3Nm0RO2O2M3N1N3N1N10N101O0N2O2N1N3N1O2M2O2M2O2N1N3N1N2O2N1N3N1O2M2O2M2O2N1N2O2N1N3N1K6kNT1lNV^U3"}, "label": "a baby carrier behind the little girl"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green baby seat\"."}], "task": "refering", "answer_template": "The \"a green baby seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 125, 126, 127, 128, 129, 149, 150, 151, 152, 173, 174, 175, 196, 197, 198, 219, 220, 221, 242, 243], "bbox": [0.44934375, 0.305607476635514, 0.6206562499999999, 0.7146028037383179], "iscrowd": 0, "area": 9256.5596, "rle": {"size": [428, 640], "counts": "T`h33Y=4L4L3M1O1O001O1O001iC_Oa;b0\\D@c;a0\\D@d;a0YDAg;`0WDAi;?UDCk;>RDDn;=PDCQ<l0000O100O10000O10000O11O1O1O1O1UDkN];c1O1O1O1O1O1O1O2O0O101N3N2M3N2M3N2M3M3N2M3N2M3N1N3N2M3N2M3N2M3M3Nm0RO2O2M3N1N3N1N10N101O0N2O2N1N3N1O2M2O2M2O2N1N3N1N2O2N1N3N1O2M2O2M2O2N1N2O2N1N3N1K6kNT1lNV^U3"}, "label": "a green baby seat"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl licking cake off her fingers\"."}], "task": "refering", "answer_template": "The \"a little girl licking cake off her fingers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263], "bbox": [0.212921875, 0.08322429906542056, 0.5547968750000001, 0.758107476635514], "iscrowd": 0, "area": 42860.343550000005, "rle": {"size": [428, 640], "counts": "YSi1h0a<8G9H7I8H8H7H9H5K3M3L4N2N2N2N2N2N2N2N2N2N2O1O1N2O1QOYLXHh3d7^LXHd3e7aLXH`3d7fLYHZ3e7jLXHX3d7nLYHS3c7SMZHn2d7VMXHl2f7XMWHi2h7YMVHh2m6oKjI`5S6eJjI\\5S6hJjIZ5S6jJkIW5Q6oJkIS5R6QKlIP5Q6W1M3M3M2N3L4K5J6K5J6K5K5M3O100O100O1O100O010O1O100O100O10000O100000000000O1000000000O1000000000000000000000O1000000000001O000000000000001O0O1000000000001O00001O1O1O001O1O001N2O1O001O1O001O1O1O2N2N1O2N1O2N2M2O2N1O3M5K5K4L5K5K8H:F:F:F8H9G6J6I2O1O1O1O1O1O1O001O1O1O1O1O2N1O1O1N2O1O1O2N1O1O1O1O2N1O1O1O1O2N1O3L7J6J4L3M2N3M2N3M2N3M3M2N3M2N3M2N3L4[O^DZOe;?bD^Oa;<eDA^;8hDF[;3lDJW;OoDOZ<MUjf3"}, "label": "a little girl licking cake off her fingers"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl wearing a blue tank top licking her thumb\"."}], "task": "refering", "answer_template": "The \"a young girl wearing a blue tank top licking her thumb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263], "bbox": [0.212921875, 0.08322429906542056, 0.5547968750000001, 0.758107476635514], "iscrowd": 0, "area": 42860.343550000005, "rle": {"size": [428, 640], "counts": "YSi1h0a<8G9H7I8H8H7H9H5K3M3L4N2N2N2N2N2N2N2N2N2N2O1O1N2O1QOYLXHh3d7^LXHd3e7aLXH`3d7fLYHZ3e7jLXHX3d7nLYHS3c7SMZHn2d7VMXHl2f7XMWHi2h7YMVHh2m6oKjI`5S6eJjI\\5S6hJjIZ5S6jJkIW5Q6oJkIS5R6QKlIP5Q6W1M3M3M2N3L4K5J6K5J6K5K5M3O100O100O1O100O010O1O100O100O10000O100000000000O1000000000O1000000000000000000000O1000000000001O000000000000001O0O1000000000001O00001O1O1O001O1O001N2O1O001O1O001O1O1O2N2N1O2N1O2N2M2O2N1O3M5K5K4L5K5K8H:F:F:F8H9G6J6I2O1O1O1O1O1O1O001O1O1O1O1O2N1O1O1N2O1O1O2N1O1O1O1O2N1O1O1O1O2N1O3L7J6J4L3M2N3M2N3M2N3M3M2N3M2N3M2N3L4[O^DZOe;?bD^Oa;<eDA^;8hDF[;3lDJW;OoDOZ<MUjf3"}, "label": "a young girl wearing a blue tank top licking her thumb"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black stool with a bottled water and paper plate sitting on it\"."}], "task": "refering", "answer_template": "The \"a black stool with a bottled water and paper plate sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 37, 38, 39, 40, 41, 42, 61, 62, 63, 64, 65, 84, 87, 88, 107, 108, 109, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 227, 246], "bbox": [0.641828125, 0.0002102803738317757, 0.9170937499999999, 0.7138551401869159], "iscrowd": 0, "area": 25825.41595, "rle": {"size": [428, 640], "counts": "bQ\\54W51_25`M2T24lM2h14WN3]13bN4R13lN4g03YO4:2E5O017CM?9UOJm0=gNGZ1`0ZNCh1d0lM_OW2g0]M]Od2h0RMZOQ3g0iL[O[3e0_L\\Oe3e0UL]On3d0lK]OY4c0aK_Oc4a0WKYOU5g0eJQOi5o0UJnNP6Q1QJnNP6P1RJPOn5o0SJPOn5n0TJROl5m0UJROl5m0UJSOk5l0VJSOk5m0UJROl5m0TJTOl5k0UJTOl5l0TJTOl5k0UJTOl5l0TJTOl5k0UJUOo1TN>f2dMXOh1VNd0b2dMZOb1WNk0^2dM]O[1XNR1Z2dM@>POn1`1dMB7lN\\2a1^ME0eNk2f1UMGI_N[3h1mLLAXNk3j1fL_1c3_N_LX1j3fNXLQ1Q4mNSLh0V4VOnK?[4@iK=W4BlK<T4CoK;Q4DSL:l3FTL:l3ETL=k3BVL>j3AWL?i3@XLa0g3^OZLb0f3]O[Ld0d3[O]Le0c3ZO^Lg0a3YO_Lg0a3AWL?i3CUL>j3DTL<l3FRL;m3GQL9o3IoK8P4JnK6R4LlK4T4LkK6T4KhK8X4HeK;[4EbK>_4A^Kb0b4^O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4\\OZKoMcNR2S6O\\KlMcNT2Q61^KfMcNY2o52`KaMbN]2n53bK[MbNb2l53kKMU44kKKU46jKJV47iKIW48hKHX48hKHX49gKGY4:fKFZ4;eKE[4;fKDZ4=eKC[4>dKB\\4?cKB\\4>dKB\\4?cKA]4`0bK@^4a0aK_O_4b0`K^O`4b0`K^O`4c0_K]Oa4d0^K\\Ob4e0]K[Oc4f0\\KZOd4f0\\K[Oc4f0\\KZOd4g0[KYOe4h0ZKXOf4i0YKWOg4i0YKWOg4j0XKVOh4j0XKVOh4j0XKVOh4j0XKVOh4j0XKVOh4j0XKWOg4i0YKWOg4i0YKVOh4j0XKVOh4j0XKVOh4j0XKVOh4i0YKWOg4i0YKVOh4j0XKVOk4g0UKYOo4c0QK]OS57UKIo4JZK6k4\\O^Kc0g4QOaKo0c4dNfK\\1^4WNkKi1Y4kMoKT2e7O1O1O1O100O1O1O1O1O1O11O2N2N2N2N2N2N2N2N3Mb0^O2N1OhNWFoNg9i0gFSOW9e0WGWOg8a0gG[OW89[HCe7MoHOR7IYI3g6F_Sf0"}, "label": "a black stool with a bottled water and paper plate sitting on it"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black bar stool with a plate and water bottle in it\"."}], "task": "refering", "answer_template": "The \"black bar stool with a plate and water bottle in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 37, 38, 39, 40, 41, 42, 61, 62, 63, 64, 65, 84, 87, 88, 107, 108, 109, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 227, 246], "bbox": [0.641828125, 0.0002102803738317757, 0.9170937499999999, 0.7138551401869159], "iscrowd": 0, "area": 25825.41595, "rle": {"size": [428, 640], "counts": "bQ\\54W51_25`M2T24lM2h14WN3]13bN4R13lN4g03YO4:2E5O017CM?9UOJm0=gNGZ1`0ZNCh1d0lM_OW2g0]M]Od2h0RMZOQ3g0iL[O[3e0_L\\Oe3e0UL]On3d0lK]OY4c0aK_Oc4a0WKYOU5g0eJQOi5o0UJnNP6Q1QJnNP6P1RJPOn5o0SJPOn5n0TJROl5m0UJROl5m0UJSOk5l0VJSOk5m0UJROl5m0TJTOl5k0UJTOl5l0TJTOl5k0UJTOl5l0TJTOl5k0UJUOo1TN>f2dMXOh1VNd0b2dMZOb1WNk0^2dM]O[1XNR1Z2dM@>POn1`1dMB7lN\\2a1^ME0eNk2f1UMGI_N[3h1mLLAXNk3j1fL_1c3_N_LX1j3fNXLQ1Q4mNSLh0V4VOnK?[4@iK=W4BlK<T4CoK;Q4DSL:l3FTL:l3ETL=k3BVL>j3AWL?i3@XLa0g3^OZLb0f3]O[Ld0d3[O]Le0c3ZO^Lg0a3YO_Lg0a3AWL?i3CUL>j3DTL<l3FRL;m3GQL9o3IoK8P4JnK6R4LlK4T4LkK6T4KhK8X4HeK;[4EbK>_4A^Kb0b4^O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4[O[Ke0e4\\OZKoMcNR2S6O\\KlMcNT2Q61^KfMcNY2o52`KaMbN]2n53bK[MbNb2l53kKMU44kKKU46jKJV47iKIW48hKHX48hKHX49gKGY4:fKFZ4;eKE[4;fKDZ4=eKC[4>dKB\\4?cKB\\4>dKB\\4?cKA]4`0bK@^4a0aK_O_4b0`K^O`4b0`K^O`4c0_K]Oa4d0^K\\Ob4e0]K[Oc4f0\\KZOd4f0\\K[Oc4f0\\KZOd4g0[KYOe4h0ZKXOf4i0YKWOg4i0YKWOg4j0XKVOh4j0XKVOh4j0XKVOh4j0XKVOh4j0XKVOh4j0XKWOg4i0YKWOg4i0YKVOh4j0XKVOh4j0XKVOh4j0XKVOh4i0YKWOg4i0YKVOh4j0XKVOk4g0UKYOo4c0QK]OS57UKIo4JZK6k4\\O^Kc0g4QOaKo0c4dNfK\\1^4WNkKi1Y4kMoKT2e7O1O1O1O100O1O1O1O1O1O11O2N2N2N2N2N2N2N2N3Mb0^O2N1OhNWFoNg9i0gFSOW9e0WGWOg8a0gG[OW89[HCe7MoHOR7IYI3g6F_Sf0"}, "label": "black bar stool with a plate and water bottle in it"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with a baby on her lap\"."}], "task": "refering", "answer_template": "The \"a women with a baby on her lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 210, 211, 212, 232, 233, 234, 255, 256, 257], "bbox": [0.0, 0.0002102803738317757, 0.302203125, 0.7653271028037383], "iscrowd": 0, "area": 46917.335999999996, "rle": {"size": [428, 640], "counts": "0`3l9a3_L1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O101N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O00O10000000000000000000000000000000000000000000000O100000000000000gKmLYNS3f1QMWNo2h1UMUNk2j1YMSNg2l1]MQNc2n1aMoM_2P2fMlMZ2S2jMjMV2U2nMhMR2X2QNeMo1Z2UNcMk1]2XN`Mh1_2\\N^Md1b2aNXM`1g2eNTM\\1l2hNoLY1P3mNkLT1T3oNhLR1W3POgLQ1Y3QOhKoMYOQ3n4SO_K_NWO_2Z5TOVKV2i4lMnJ\\2R5fMeJa2Z5i200O100O100O100O100O100O100O1O100O1O1O100O1O1O100O1N2I7I7I7I7H8J6N2M3N2N2G9POR1K6J7J5L4L5L3L4L5K5Kbij5"}, "label": "a women with a baby on her lap"}]}
{"id": 358770, "image": "COCO_train2014_000000358770.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing blue jeans sitting on a chair with a baby sitting in her lap\"."}], "task": "refering", "answer_template": "The \"a woman wearing blue jeans sitting on a chair with a baby sitting in her lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 210, 211, 212, 232, 233, 234, 255, 256, 257], "bbox": [0.0, 0.0002102803738317757, 0.302203125, 0.7653271028037383], "iscrowd": 0, "area": 46917.335999999996, "rle": {"size": [428, 640], "counts": "0`3l9a3_L1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O101N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O00O10000000000000000000000000000000000000000000000O100000000000000gKmLYNS3f1QMWNo2h1UMUNk2j1YMSNg2l1]MQNc2n1aMoM_2P2fMlMZ2S2jMjMV2U2nMhMR2X2QNeMo1Z2UNcMk1]2XN`Mh1_2\\N^Md1b2aNXM`1g2eNTM\\1l2hNoLY1P3mNkLT1T3oNhLR1W3POgLQ1Y3QOhKoMYOQ3n4SO_K_NWO_2Z5TOVKV2i4lMnJ\\2R5fMeJa2Z5i200O100O100O100O100O100O100O1O100O1O1O100O1O1O100O1N2I7I7I7I7H8J6N2M3N2N2G9POR1K6J7J5L4L5L3L4L5K5Kbij5"}, "label": "a woman wearing blue jeans sitting on a chair with a baby sitting in her lap"}]}
{"id": 506226, "image": "COCO_train2014_000000506226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man helping to cut a cake\"."}], "task": "refering", "answer_template": "The \"an older man helping to cut a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 68, 69, 70, 86, 87, 88, 104, 105, 106, 107, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 318, 319, 320, 321, 322, 323, 338, 339, 340, 341, 357, 358, 359, 375, 376, 377, 393, 394, 395, 411, 412, 413], "bbox": [0.46140625, 0.103125, 1.0, 0.990859375], "iscrowd": 0, "area": 75324.57585000001, "rle": {"size": [640, 512], "counts": "Tkc41oc01N2O1O1N2O2M2O1O1N2O1N2O2N1N2O1N2O1O1N3N1N2O1OoNU]On0ib0ROW]OQ1gb0nNY]OT1eb0lN\\]OU1cb0jN]]OX1ab0iN^]OY1ab0fN`]O\\1`b02O1M3L2N2N3M2N2O1N2N2O1N2O1N2O1N2N2001N2O1O1O1O1OR_OPNn?o1S@YNe?f1[@bN^?]1c@eN[?Z1f@fNZ?Y1f@hNY?X1h@iNV?W1j@jNT?W1m@jNQ?U1QAkNn>U1RAmNl>S1UAmNj>S1WAnNf>R1[AnNe>R1\\AnNc>R1^AnNa>R1_AoN`>Q1aAoN]>Q1dAPO[>P1fAoNZ>Q1gAoNX>Q1hAPOW>o0kAQOS>P1nAPOQ>P1oAQOP>o0QBQOn=n0SBROm=n0TBROj=o0WBQOh=o0XBROg=n0ZBROe=m0]BSOb=m0^BSOa=n0`BRO_=n0bBRO]=n0cBSO\\=m0eBSOZ=m0gBSOW=n0iBSOV=n0jBROU=n0lBROS=n0mBSOR=m0oBROP=o0QCQOn<o0RCROm<n0TCROk<o0UCSOh<m0XCUOf<k0[CVOc<j0^CXO_<h0aCZO]<f0dC[O[<d0fC^OW<c0hC_OV<a0kC@S<`0nCBo;>QDDm;<TDGi;8XDLc;4]D1^;0bD4Y;LhD8S;HmD=l:EUES1P:oNQFQ1m9POSFQ1j9QOWFP1e9RO\\Fn0a9UO^Fl0_9VObFj0[9XOfFh0W9ZOiFg0T9[OmFe0P9]OQGc0l8_OnFh0Q9YO[F[1d9a4O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1N2O1O1O2N1O1eMYF\\Jh9c5bF[IW:d6TFbH\\:]7T1O2N1O100O1O2N2N2N3M2N2N2O1N2N3N1N2O1N2O1N3N1N5K7J5K6I7J5J7J6I6K6J6I7J7H8I7H8I7I8G8I7H3N0001O001O001O00001O0O2O00001O001O00001O001O001O001O1N3N1O1O2N1O1O2N1O1O2N1OQ2oM1O1O1O1N101O1O1O1O1O1O001O1O1O1O1O001O1N2O1O1O1RNoKkCR4`9d4XMg2ZMg2XMiL"}, "label": "an older man helping to cut a cake"}]}
{"id": 506226, "image": "COCO_train2014_000000506226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the far right of the photo\"."}], "task": "refering", "answer_template": "The \"the man on the far right of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 68, 69, 70, 86, 87, 88, 104, 105, 106, 107, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 318, 319, 320, 321, 322, 323, 338, 339, 340, 341, 357, 358, 359, 375, 376, 377, 393, 394, 395, 411, 412, 413], "bbox": [0.46140625, 0.103125, 1.0, 0.990859375], "iscrowd": 0, "area": 75324.57585000001, "rle": {"size": [640, 512], "counts": "Tkc41oc01N2O1O1N2O2M2O1O1N2O1N2O2N1N2O1N2O1O1N3N1N2O1OoNU]On0ib0ROW]OQ1gb0nNY]OT1eb0lN\\]OU1cb0jN]]OX1ab0iN^]OY1ab0fN`]O\\1`b02O1M3L2N2N3M2N2O1N2N2O1N2O1N2O1N2N2001N2O1O1O1O1OR_OPNn?o1S@YNe?f1[@bN^?]1c@eN[?Z1f@fNZ?Y1f@hNY?X1h@iNV?W1j@jNT?W1m@jNQ?U1QAkNn>U1RAmNl>S1UAmNj>S1WAnNf>R1[AnNe>R1\\AnNc>R1^AnNa>R1_AoN`>Q1aAoN]>Q1dAPO[>P1fAoNZ>Q1gAoNX>Q1hAPOW>o0kAQOS>P1nAPOQ>P1oAQOP>o0QBQOn=n0SBROm=n0TBROj=o0WBQOh=o0XBROg=n0ZBROe=m0]BSOb=m0^BSOa=n0`BRO_=n0bBRO]=n0cBSO\\=m0eBSOZ=m0gBSOW=n0iBSOV=n0jBROU=n0lBROS=n0mBSOR=m0oBROP=o0QCQOn<o0RCROm<n0TCROk<o0UCSOh<m0XCUOf<k0[CVOc<j0^CXO_<h0aCZO]<f0dC[O[<d0fC^OW<c0hC_OV<a0kC@S<`0nCBo;>QDDm;<TDGi;8XDLc;4]D1^;0bD4Y;LhD8S;HmD=l:EUES1P:oNQFQ1m9POSFQ1j9QOWFP1e9RO\\Fn0a9UO^Fl0_9VObFj0[9XOfFh0W9ZOiFg0T9[OmFe0P9]OQGc0l8_OnFh0Q9YO[F[1d9a4O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1N2O1O1O2N1O1eMYF\\Jh9c5bF[IW:d6TFbH\\:]7T1O2N1O100O1O2N2N2N3M2N2N2O1N2N3N1N2O1N2O1N3N1N5K7J5K6I7J5J7J6I6K6J6I7J7H8I7H8I7I8G8I7H3N0001O001O001O00001O0O2O00001O001O00001O001O001O001O1N3N1O1O2N1O1O2N1O1O2N1OQ2oM1O1O1O1N101O1O1O1O1O1O001O1O1O1O1O001O1N2O1O1O1RNoKkCR4`9d4XMg2ZMg2XMiL"}, "label": "the man on the far right of the photo"}]}
{"id": 506226, "image": "COCO_train2014_000000506226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl cutting a cake with a sword near some men\"."}], "task": "refering", "answer_template": "The \"a girl cutting a cake with a sword near some men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 113, 114, 115, 131, 132, 133, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 203, 204, 205, 206, 207, 222, 223, 224, 225, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314], "bbox": [0.23876953125, 0.22393749999999998, 0.5414257812500001, 0.7528125], "iscrowd": 0, "area": 29697.938550000006, "rle": {"size": [640, 512], "counts": "`f\\2;db0Q1ROn0M3M3M3M3M3kKXMoFk2o8cMcF_2\\9nMUFU2i9SNnEP2P:SNlEP2S:RNiEQ2U:RNgEQ2W:RNeEQ2Z:QNcER2Z:QNbER2\\:PNcEQ2\\:QNcEo1[:TNeEk1Y:XNfEh1Y:ZNgEe1X:]NgEc1m8^L[FQ2h0`1i8eL\\Fm1j0^1e8lL^Fg1n0\\1_8TM`Fb1P1Z1[8\\MaF\\1T1W1V8dMcFW1W1T1V8gM`FW1Y1R1V8kM^FT1\\1P1U8oM\\FS1^1n0U8RNZFR1a1k0T8VNXFP1d1k0R8YNVFn0h1h0Q8]NTFm0j1f0R8Y1nGfNQ8\\1nGdNQ8^1oGaNP8a1oG_NP8c1PH\\No7e1QH[No7f1QHYNn7i1RHVNm7l1RHTNm7n1SHQNm7P2RHPNm7R2SHmMm7T2RHlMn7T2SHkMm7V2RHkMm7V2SHiMm7X2RHhMm7Z2SHeMm7\\2RHdMn7]2RHbMn7_2RH`Mn7`2RH`Mm7b2SH]Mm7d2RH\\Mn7e2RHZMn7g2QHYMo7h2QHWMo7j2PHVMP8k2PHTMP8l2PHTMQ8l2oGSMR8m2mGSMT8m2lGSMT8m2kGSMU8n2kGQMV8o2jGPMW8P3hGPMY8o2hGPMY8P3fGPMZ8Q3fGnL[8R3cGoL^8Q3aGoL`8Q3^GPMc8P3\\GPMd8Q3ZGPMg8o2YGQMh8o2VGRMl8m2SGSMo8hNnE\\2Q1mNR9cNQF`2l0nNT9^NRFe2i0mNW9YNTFj2c0nN[9TNUFn2?nN^9oMWFR3:PO`9kMYFU36POc9fM[FY31ROm9ZMUFc3NSOW;n0gDSOZ;l0fDTO[;k0dDVO];j0dDTO];k0dDTO];l0dDRO\\;n0fDPO[;o0fDPO[;P1fDnN[;Q1gDmNZ;S1fDmNZ;R1hDlNY;S1hDlNX;U1iDiNX;V1jDhNW;X1iDgNX;X1jDfNW;Y1kDeNV;[1jDdNV;\\1lDbNU;^1kDaNV;^1jDbNW;]1hDdNY;\\1fDdN[;[1eDeN\\;Y1dDhN\\;X1dDhN];V1dDjN];U1bDmN^;Q1cDoN^;o0cDQO^;n0aDiMhN2l<o1]DmMGVOQ<h2XDPNZ=k1gBSN`=g1`BXNg=c1YB[Nn=_1SB_NS>]1lAbN[>X1fAfNd>Q1]AmNS?e0l@ZOe?6\\@Hj?4V@Jm?4T@Jn?5R@JQ`05o_OIS`06n_OHT`08k_OGX`07i_OGY`08h_OF[`09d_OF^`09c_OE``09d1O2N1O2N1O101N^ab4"}, "label": "a girl cutting a cake with a sword near some men"}]}
{"id": 506226, "image": "COCO_train2014_000000506226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady holding the sword\"."}], "task": "refering", "answer_template": "The \"a lady holding the sword\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 113, 114, 115, 131, 132, 133, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 203, 204, 205, 206, 207, 222, 223, 224, 225, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314], "bbox": [0.23876953125, 0.22393749999999998, 0.5414257812500001, 0.7528125], "iscrowd": 0, "area": 29697.938550000006, "rle": {"size": [640, 512], "counts": "`f\\2;db0Q1ROn0M3M3M3M3M3kKXMoFk2o8cMcF_2\\9nMUFU2i9SNnEP2P:SNlEP2S:RNiEQ2U:RNgEQ2W:RNeEQ2Z:QNcER2Z:QNbER2\\:PNcEQ2\\:QNcEo1[:TNeEk1Y:XNfEh1Y:ZNgEe1X:]NgEc1m8^L[FQ2h0`1i8eL\\Fm1j0^1e8lL^Fg1n0\\1_8TM`Fb1P1Z1[8\\MaF\\1T1W1V8dMcFW1W1T1V8gM`FW1Y1R1V8kM^FT1\\1P1U8oM\\FS1^1n0U8RNZFR1a1k0T8VNXFP1d1k0R8YNVFn0h1h0Q8]NTFm0j1f0R8Y1nGfNQ8\\1nGdNQ8^1oGaNP8a1oG_NP8c1PH\\No7e1QH[No7f1QHYNn7i1RHVNm7l1RHTNm7n1SHQNm7P2RHPNm7R2SHmMm7T2RHlMn7T2SHkMm7V2RHkMm7V2SHiMm7X2RHhMm7Z2SHeMm7\\2RHdMn7]2RHbMn7_2RH`Mn7`2RH`Mm7b2SH]Mm7d2RH\\Mn7e2RHZMn7g2QHYMo7h2QHWMo7j2PHVMP8k2PHTMP8l2PHTMQ8l2oGSMR8m2mGSMT8m2lGSMT8m2kGSMU8n2kGQMV8o2jGPMW8P3hGPMY8o2hGPMY8P3fGPMZ8Q3fGnL[8R3cGoL^8Q3aGoL`8Q3^GPMc8P3\\GPMd8Q3ZGPMg8o2YGQMh8o2VGRMl8m2SGSMo8hNnE\\2Q1mNR9cNQF`2l0nNT9^NRFe2i0mNW9YNTFj2c0nN[9TNUFn2?nN^9oMWFR3:PO`9kMYFU36POc9fM[FY31ROm9ZMUFc3NSOW;n0gDSOZ;l0fDTO[;k0dDVO];j0dDTO];k0dDTO];l0dDRO\\;n0fDPO[;o0fDPO[;P1fDnN[;Q1gDmNZ;S1fDmNZ;R1hDlNY;S1hDlNX;U1iDiNX;V1jDhNW;X1iDgNX;X1jDfNW;Y1kDeNV;[1jDdNV;\\1lDbNU;^1kDaNV;^1jDbNW;]1hDdNY;\\1fDdN[;[1eDeN\\;Y1dDhN\\;X1dDhN];V1dDjN];U1bDmN^;Q1cDoN^;o0cDQO^;n0aDiMhN2l<o1]DmMGVOQ<h2XDPNZ=k1gBSN`=g1`BXNg=c1YB[Nn=_1SB_NS>]1lAbN[>X1fAfNd>Q1]AmNS?e0l@ZOe?6\\@Hj?4V@Jm?4T@Jn?5R@JQ`05o_OIS`06n_OHT`08k_OGX`07i_OGY`08h_OF[`09d_OF^`09c_OE``09d1O2N1O2N1O101N^ab4"}, "label": "a lady holding the sword"}]}
{"id": 506226, "image": "COCO_train2014_000000506226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glass , cutting the cake and standing second from right\"."}], "task": "refering", "answer_template": "The \"a man with glass , cutting the cake and standing second from right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 45, 46, 47, 63, 64, 65, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 152, 153, 154, 155, 156, 171, 172, 173, 189, 190, 207, 208, 225], "bbox": [0.446640625, 0.0793125, 0.7858984375, 0.532421875], "iscrowd": 0, "area": 23953.232599999996, "rle": {"size": [640, 512], "counts": "SZ_47fc06J6L4K:F8H3M3N3L3M3M3N2M3M4M2N2O1N2O1N3N1O2M3N2M2O2M3N2l@RMQ=P3fBbMP=`2eBjMY=Y2XBUNg=S4N101N2O1O1N101ZM\\JmFf5o8eJiF\\5R9mJhFS5V9hKRFV4P:nKkER4U:QLiEn3W:ULfEk3Z:XLcEh3]:[L`Ed3a:_L]E`3c:cLZE\\3h:fLUEY3m:iLQEU3Q;lLmDS3T;oLjDP3X;RMfDl2\\;UMbDj2`;WM^Dh2d;ZMZDd2g;^MWD`2l;aMRD^2P<g22N1O2O100O10O0100O100O100O100O100O1O1N2O2M3N2N2M3N2M2O2N2N2N2N2N2M4M2N2M3N2M3M4L5L3L5K4L3N2eK\\AW3g>cL_AZ3c>`LcA_3_>XLhAg3\\>oKlAo3S?O2O0O1O2N101N1O2N1O1M4M2N3L3N3L3N2N3M2M4I6M4M2O1N3M2O2M2O1O2N1O2N1O2M2O1O2N1O2M2O2O0O1O2N100O1O100O2N1O101N1O2N2L4L4L4L7JZcT2"}, "label": "a man with glass , cutting the cake and standing second from right"}]}
{"id": 506226, "image": "COCO_train2014_000000506226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man with glasses , standing between another man and two women , cutting a cake\"."}], "task": "refering", "answer_template": "The \"an older man with glasses , standing between another man and two women , cutting a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 45, 46, 47, 63, 64, 65, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 152, 153, 154, 155, 156, 171, 172, 173, 189, 190, 207, 208, 225], "bbox": [0.446640625, 0.0793125, 0.7858984375, 0.532421875], "iscrowd": 0, "area": 23953.232599999996, "rle": {"size": [640, 512], "counts": "SZ_47fc06J6L4K:F8H3M3N3L3M3M3N2M3M4M2N2O1N2O1N3N1O2M3N2M2O2M3N2l@RMQ=P3fBbMP=`2eBjMY=Y2XBUNg=S4N101N2O1O1N101ZM\\JmFf5o8eJiF\\5R9mJhFS5V9hKRFV4P:nKkER4U:QLiEn3W:ULfEk3Z:XLcEh3]:[L`Ed3a:_L]E`3c:cLZE\\3h:fLUEY3m:iLQEU3Q;lLmDS3T;oLjDP3X;RMfDl2\\;UMbDj2`;WM^Dh2d;ZMZDd2g;^MWD`2l;aMRD^2P<g22N1O2O100O10O0100O100O100O100O100O1O1N2O2M3N2N2M3N2M2O2N2N2N2N2N2M4M2N2M3N2M3M4L5L3L5K4L3N2eK\\AW3g>cL_AZ3c>`LcA_3_>XLhAg3\\>oKlAo3S?O2O0O1O2N101N1O2N1O1M4M2N3L3N3L3N2N3M2M4I6M4M2O1N3M2O2M2O1O2N1O2N1O2M2O1O2N1O2M2O2O0O1O2N100O1O100O2N1O101N1O2N2L4L4L4L7JZcT2"}, "label": "an older man with glasses , standing between another man and two women , cutting a cake"}]}
{"id": 416117, "image": "COCO_train2014_000000416117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with her hands in the air enjoying herself playing a video game\"."}], "task": "refering", "answer_template": "The \"a lady with her hands in the air enjoying herself playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 83, 84, 104, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.5195937500000001, 0.07843091334894614, 0.67471875, 0.9881498829039813], "iscrowd": 0, "area": 24439.40729999999, "rle": {"size": [427, 640], "counts": "TR[4:o<;E<D:F3fGcNb4a1UKkNd4X1SKTOh2n2QM\\Mf2j2XM\\M^2j2_M\\MW2k2fM[MQ2k2mM[Mi1k2TN[Mc1k2[N[MZ1l2cNZMT1l2SMcK;f1Y2m2PMlKa0^1U2]4fMfKT2c4gM^KS2j4iMXKP2P5nMQKk1X5QNjJg1a8H8H8O1000000OPFJm66iH9Q7HeHf0W7ZOZHY1b7gNSHd1m7\\NSHd1m7\\NRHe1n7[NRHf1m7ZNRHg1n7YNRHg1n7YNQHh1o7XNQHh1o7YNoGh1Q8ZNlGg1T8ZNkGf1U8\\NhGe1X8\\NfGe1Z8\\NdGe1\\8]NaGd1_8]N_Gd1a8b110000000000000O011O001O001O001O00001O1O001iISKn3n4PLUKn3l4PLXKm3i4QLZKm3g4QL]Kl3d4SL^Kk3c4SLaKj3]2eJfN_1POj3X2RK\\NV1@e3R2aKPNm01`3o1_NUN^1k1aNXN]1g1cN]NZ1c1fN_NW1b1hNbNU1]1kNhNQ1X1kN]OTOhMbNk2W21^NaMVO]2Z2f0hMYMIQ2\\2[1RMoL>f1]2[3^MeL^2d3]M\\L_2l3]MTL_2U4\\MlK`2\\4[MdKa2e4ZM\\Ki0[Mc0^:XOWF4m9FiFFTnf2"}, "label": "a lady with her hands in the air enjoying herself playing a video game"}]}
{"id": 416117, "image": "COCO_train2014_000000416117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a stripped shirt and jeans with her hands above her head\"."}], "task": "refering", "answer_template": "The \"a woman in a stripped shirt and jeans with her hands above her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 83, 84, 104, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.5195937500000001, 0.07843091334894614, 0.67471875, 0.9881498829039813], "iscrowd": 0, "area": 24439.40729999999, "rle": {"size": [427, 640], "counts": "TR[4:o<;E<D:F3fGcNb4a1UKkNd4X1SKTOh2n2QM\\Mf2j2XM\\M^2j2_M\\MW2k2fM[MQ2k2mM[Mi1k2TN[Mc1k2[N[MZ1l2cNZMT1l2SMcK;f1Y2m2PMlKa0^1U2]4fMfKT2c4gM^KS2j4iMXKP2P5nMQKk1X5QNjJg1a8H8H8O1000000OPFJm66iH9Q7HeHf0W7ZOZHY1b7gNSHd1m7\\NSHd1m7\\NRHe1n7[NRHf1m7ZNRHg1n7YNRHg1n7YNQHh1o7XNQHh1o7YNoGh1Q8ZNlGg1T8ZNkGf1U8\\NhGe1X8\\NfGe1Z8\\NdGe1\\8]NaGd1_8]N_Gd1a8b110000000000000O011O001O001O001O00001O1O001iISKn3n4PLUKn3l4PLXKm3i4QLZKm3g4QL]Kl3d4SL^Kk3c4SLaKj3]2eJfN_1POj3X2RK\\NV1@e3R2aKPNm01`3o1_NUN^1k1aNXN]1g1cN]NZ1c1fN_NW1b1hNbNU1]1kNhNQ1X1kN]OTOhMbNk2W21^NaMVO]2Z2f0hMYMIQ2\\2[1RMoL>f1]2[3^MeL^2d3]M\\L_2l3]MTL_2U4\\MlK`2\\4[MdKa2e4ZM\\Ki0[Mc0^:XOWF4m9FiFFTnf2"}, "label": "a woman in a stripped shirt and jeans with her hands above her head"}]}
{"id": 506231, "image": "COCO_train2014_000000506231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table all the people are sitting on\"."}], "task": "refering", "answer_template": "The \"the table all the people are sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 71, 72, 83, 84, 85, 86, 87, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 133, 134, 135, 137, 138, 139, 146, 147, 148, 151, 152, 159, 160, 161, 164, 165, 171, 172, 177, 178, 184, 185, 190, 191, 197, 203, 204, 210, 216, 217, 229], "bbox": [0.16861333333333334, 0.25902, 0.7042933333333333, 0.9610599999999999], "iscrowd": 0, "area": 23290.36325000001, "rle": {"size": [500, 375], "counts": "gVo02[?8H7I7I8G8I8H7I7I8H7I8M2M3N3M2N3M2M4MaNhCFV<8]DYOa;f0PEmNm:Q1fE`NW:_1o1N2N2O11N101O00001O0O2O001O001N101O00001N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2O02N100O1O2O0O1O2O0O1O2O0O1O2O0O1O2O0O1O2GmNWMkEY2d:jMZEf1T;^NkDS1a;RO\\Dn0b;VO\\Dj0c;YO\\De0c;@ZD`0d;DZD<d;H[D7d;MYD3e;o1N1O2O1N2N2N2O0O2N2N2O1N2N2N101N2N2N2O1N1O2N22N4L4L4K5L4L4L4L4L4K5L4L4L4L4L4K5L1OO3M4M3L4L5L3L4L4M4K4L4M3L5K4M2M3M4RHZMh2h2]LQNb3R2ULTNk3o1kKYNT4k1bK\\N^4f1YK`Ng4c1PKdNo4`1gJfNZ5\\1\\JlNc5W1TJPOk5S1lISOU6V5O00100O0010O01O010O1O010O05D<B>C=BmPf1"}, "label": "the table all the people are sitting on"}]}
{"id": 506231, "image": "COCO_train2014_000000506231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large white table with people seated around\"."}], "task": "refering", "answer_template": "The \"a large white table with people seated around\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 71, 72, 83, 84, 85, 86, 87, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 133, 134, 135, 137, 138, 139, 146, 147, 148, 151, 152, 159, 160, 161, 164, 165, 171, 172, 177, 178, 184, 185, 190, 191, 197, 203, 204, 210, 216, 217, 229], "bbox": [0.16861333333333334, 0.25902, 0.7042933333333333, 0.9610599999999999], "iscrowd": 0, "area": 23290.36325000001, "rle": {"size": [500, 375], "counts": "gVo02[?8H7I7I8G8I8H7I7I8H7I8M2M3N3M2N3M2M4MaNhCFV<8]DYOa;f0PEmNm:Q1fE`NW:_1o1N2N2O11N101O00001O0O2O001O001N101O00001N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2O02N100O1O2O0O1O2O0O1O2O0O1O2O0O1O2O0O1O2GmNWMkEY2d:jMZEf1T;^NkDS1a;RO\\Dn0b;VO\\Dj0c;YO\\De0c;@ZD`0d;DZD<d;H[D7d;MYD3e;o1N1O2O1N2N2N2O0O2N2N2O1N2N2N101N2N2N2O1N1O2N22N4L4L4K5L4L4L4L4L4K5L4L4L4L4L4K5L1OO3M4M3L4L5L3L4L4M4K4L4M3L5K4M2M3M4RHZMh2h2]LQNb3R2ULTNk3o1kKYNT4k1bK\\N^4f1YK`Ng4c1PKdNo4`1gJfNZ5\\1\\JlNc5W1TJPOk5S1lISOU6V5O00100O0010O01O010O1O010O05D<B>C=BmPf1"}, "label": "a large white table with people seated around"}]}
{"id": 506231, "image": "COCO_train2014_000000506231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty table in the back\"."}], "task": "refering", "answer_template": "The \"the empty table in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 60, 61, 68, 69, 70, 71, 73, 81, 84], "bbox": [0.2857866666666667, 0.24716, 0.75744, 0.37856], "iscrowd": 0, "area": 2522.5245000000014, "rle": {"size": [500, 375], "counts": "f]d13Y?8G8H9G9G8O20O0DnAXOQ>g0UBUOk=j0YBROg=m0`001N10000010O0001O0100000O1000000O100N2N2N2O1N2D\\Y96jfF2N3M3M3M2N3M3M2M4M3M3M1ON2N2N2N2N2N1O2N2N2N2N2N2N2O1N2N1O2N2N2N2N2N2Onc?1R\\@2N2N2N001O001O0O2O001O001O1O006J7I1N101O[O]A:3\\OZ>7hA`0X>^OlA?U>AlA=U>BnA;S>DoA:R>FPB7P>IRB4P>LRB1o=Nl0O100000000000O010000000000003M4M001O2O0O22L5I`Y\\1"}, "label": "the empty table in the back"}]}
{"id": 506231, "image": "COCO_train2014_000000506231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small table near the window\"."}], "task": "refering", "answer_template": "The \"a small table near the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 60, 61, 68, 69, 70, 71, 73, 81, 84], "bbox": [0.2857866666666667, 0.24716, 0.75744, 0.37856], "iscrowd": 0, "area": 2522.5245000000014, "rle": {"size": [500, 375], "counts": "f]d13Y?8G8H9G9G8O20O0DnAXOQ>g0UBUOk=j0YBROg=m0`001N10000010O0001O0100000O1000000O100N2N2N2O1N2D\\Y96jfF2N3M3M3M2N3M3M2M4M3M3M1ON2N2N2N2N2N1O2N2N2N2N2N2N2O1N2N1O2N2N2N2N2N2Onc?1R\\@2N2N2N001O001O0O2O001O001O1O006J7I1N101O[O]A:3\\OZ>7hA`0X>^OlA?U>AlA=U>BnA;S>DoA:R>FPB7P>IRB4P>LRB1o=Nl0O100000000000O010000000000003M4M001O2O0O22L5I`Y\\1"}, "label": "a small table near the window"}]}
{"id": 252283, "image": "COCO_train2014_000000252283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red motorcycle closest to camera\"."}], "task": "refering", "answer_template": "The \"red motorcycle closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266], "bbox": [0.3493125, 0.5632084309133489, 0.584671875, 0.8017330210772834], "iscrowd": 0, "area": 9768.136950000004, "rle": {"size": [427, 640], "counts": "nem21Y=2N2[OMnC6P<LlC8S<IkC9T<HjCl0c;UOZDQ1b;POZDV1c;;N1N3M2N3M2N100N3K4N3M2O2M2O2N001O1O100001O0O2O001O0O2N2gM]Ef1U;O2N1N3M3O0O2O00001O00000000O101O000O100000000O2O000O100O100O1O100O1O1O2N1O10001N1000000O2O00001N100000000O10ESEWNn:j19001O1O1O001O1O1O001O1O100O1O1N2O1N2O1N3M3N1N2N2O1N10100O1ORNfEk0Z:ROmEj0T:UOlEk0U:TOkEm0T:ROmEn0T:QOmEn0U:POkEP1V:oNjEQ1W:nNiER1Z:iNhEW1[:cNhE]1S;O2O00001O001O2L6K4K4M4K5L4K4M4Kgc^3"}, "label": "red motorcycle closest to camera"}]}
{"id": 252283, "image": "COCO_train2014_000000252283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red & black color bike in ftont of the three guys\"."}], "task": "refering", "answer_template": "The \"a red & black color bike in ftont of the three guys\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266], "bbox": [0.3493125, 0.5632084309133489, 0.584671875, 0.8017330210772834], "iscrowd": 0, "area": 9768.136950000004, "rle": {"size": [427, 640], "counts": "nem21Y=2N2[OMnC6P<LlC8S<IkC9T<HjCl0c;UOZDQ1b;POZDV1c;;N1N3M2N3M2N100N3K4N3M2O2M2O2N001O1O100001O0O2O001O0O2N2gM]Ef1U;O2N1N3M3O0O2O00001O00000000O101O000O100000000O2O000O100O100O1O100O1O1O2N1O10001N1000000O2O00001N100000000O10ESEWNn:j19001O1O1O001O1O1O001O1O100O1O1N2O1N2O1N3M3N1N2N2O1N10100O1ORNfEk0Z:ROmEj0T:UOlEk0U:TOkEm0T:ROmEn0T:QOmEn0U:POkEP1V:oNjEQ1W:nNiER1Z:iNhEW1[:cNhE]1S;O2O00001O001O2L6K4K4M4K5L4K4M4Kgc^3"}, "label": "a red & black color bike in ftont of the three guys"}]}
{"id": 252283, "image": "COCO_train2014_000000252283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"all black motorcycle\"."}], "task": "refering", "answer_template": "The \"all black motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321], "bbox": [0.60421875, 0.582271662763466, 0.9985312499999999, 0.9125995316159251], "iscrowd": 0, "area": 21746.524899999993, "rle": {"size": [427, 640], "counts": "odQ58S=1Oc0]Of0ZO3M2N1O1O2N1N3N1O2N1N3N1O2N1N3N1O2N1N3N1M4L3M2N2N2N2N20000000O100000001O0000000O1000000000001O0O10000000000000000O101O0000002N1L4L4K6K4K5L4K6N1O0010O0001O00010O000010O0001O00O2O0O1O100O2N100O1O2O0O1O100O2N100O1O101N1O100O1O2O0O1O101N10000O101O0O10000O2O0O10001N10000O101O0O100O101O0O10001N10000O101N1H8000001O01O0001O0001O0001O000001O01O0000010O000000010O000000010O0000000010O000001O02N2N2N2O1N2N1O2N3N1O1O1O1O2N1O1O1O1O1O2N1O1OO100O1O1O1O1O000001O000000000000001O000000000000001O000000007Je0ZOf0ZOf0ZO`1"}, "label": "all black motorcycle"}]}
{"id": 252283, "image": "COCO_train2014_000000252283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a black colored bike on the right side of the picture\"."}], "task": "refering", "answer_template": "The \"there is a black colored bike on the right side of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321], "bbox": [0.60421875, 0.582271662763466, 0.9985312499999999, 0.9125995316159251], "iscrowd": 0, "area": 21746.524899999993, "rle": {"size": [427, 640], "counts": "odQ58S=1Oc0]Of0ZO3M2N1O1O2N1N3N1O2N1N3N1O2N1N3N1O2N1N3N1M4L3M2N2N2N2N20000000O100000001O0000000O1000000000001O0O10000000000000000O101O0000002N1L4L4K6K4K5L4K6N1O0010O0001O00010O000010O0001O00O2O0O1O100O2N100O1O2O0O1O100O2N100O1O101N1O100O1O2O0O1O101N10000O101O0O10000O2O0O10001N10000O101O0O100O101O0O10001N10000O101N1H8000001O01O0001O0001O0001O000001O01O0000010O000000010O000000010O0000000010O000001O02N2N2N2O1N2N1O2N3N1O1O1O1O2N1O1O1O1O1O2N1O1OO100O1O1O1O1O000001O000000000000001O000000000000001O000000007Je0ZOf0ZOf0ZO`1"}, "label": "there is a black colored bike on the right side of the picture"}]}
{"id": 63867, "image": "COCO_train2014_000000063867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding an umbrella on a bench\"."}], "task": "refering", "answer_template": "The \"a woman holding an umbrella on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 80, 81, 82, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 154, 155, 156, 172, 173, 174, 191, 192, 209, 210], "bbox": [0.45168, 0.2561866666666667, 0.72808, 0.9011199999999999], "iscrowd": 0, "area": 14580.390049999998, "rle": {"size": [375, 500], "counts": "`mb23a;5B>K5K5K5K5XOh0M3N2M3M3N2M3M3N2M3N2Ma0_O8I4K1O101N100O2N100O1O2O0O100O20O1O1O100O1O100O1O1O100O1O1O010O1O3M3N2M3M4L4M4K4L2N100O1O1O1001O2O1N1O2N2N2N2N4L4L4M3L4K4M4K4L3M3M3M3N2M4L2N2N1O2O0O1O2N101N1O1O2N100O1O000001O000iKfIQ3Z6mLlIo2T6PMRJj2P6SMVJi2j5VM[Jf2e5XMbJc2^5\\MgJ`2Y5^MjJj0gN6`6kNnJm0fN5\\6kNSKl0fN6X6iNWKo0eN5U6hNZKQ1eN1W8LYHDk79`HZOd7d0ZHYOm7e0QHVOY8h0\\100O1O2M2N2N2N2N3M2N2N2N2Meaa1"}, "label": "a woman holding an umbrella on a bench"}]}
{"id": 63867, "image": "COCO_train2014_000000063867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding a polka dotted umbrella\"."}], "task": "refering", "answer_template": "The \"woman holding a polka dotted umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 80, 81, 82, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 154, 155, 156, 172, 173, 174, 191, 192, 209, 210], "bbox": [0.45168, 0.2561866666666667, 0.72808, 0.9011199999999999], "iscrowd": 0, "area": 14580.390049999998, "rle": {"size": [375, 500], "counts": "`mb23a;5B>K5K5K5K5XOh0M3N2M3M3N2M3M3N2M3N2Ma0_O8I4K1O101N100O2N100O1O2O0O100O20O1O1O100O1O100O1O1O100O1O1O010O1O3M3N2M3M4L4M4K4L2N100O1O1O1001O2O1N1O2N2N2N2N4L4L4M3L4K4M4K4L3M3M3M3N2M4L2N2N1O2O0O1O2N101N1O1O2N100O1O000001O000iKfIQ3Z6mLlIo2T6PMRJj2P6SMVJi2j5VM[Jf2e5XMbJc2^5\\MgJ`2Y5^MjJj0gN6`6kNnJm0fN5\\6kNSKl0fN6X6iNWKo0eN5U6hNZKQ1eN1W8LYHDk79`HZOd7d0ZHYOm7e0QHVOY8h0\\100O1O2M2N2N2N2N3M2N2N2N2Meaa1"}, "label": "woman holding a polka dotted umbrella"}]}
{"id": 63867, "image": "COCO_train2014_000000063867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a woman wearing black shoes on the left\"."}], "task": "refering", "answer_template": "The \"this is a woman wearing black shoes on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 114, 115, 116, 132, 133, 134, 150, 151, 152, 168, 169, 170, 187, 188, 205, 206, 224], "bbox": [0.35224, 0.37767999999999996, 0.50224, 0.9394666666666666], "iscrowd": 0, "area": 10634.30195, "rle": {"size": [375, 500], "counts": "XdP2j0j:`1aN8G9H8F:C=D8H501O1000O01000O0100O01000O10O10O10O10O1000O0101O8H9H8G5L3L5L4L4M2M4M3M1NO2N2O0O2N1O2O0O2N2N3M2O1N3M2N2N5L5J6J6J6I7D=C=D<F:E9G6J7I7J6Immj2"}, "label": "this is a woman wearing black shoes on the left"}]}
{"id": 63867, "image": "COCO_train2014_000000063867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman who is not holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a woman who is not holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 114, 115, 116, 132, 133, 134, 150, 151, 152, 168, 169, 170, 187, 188, 205, 206, 224], "bbox": [0.35224, 0.37767999999999996, 0.50224, 0.9394666666666666], "iscrowd": 0, "area": 10634.30195, "rle": {"size": [375, 500], "counts": "XdP2j0j:`1aN8G9H8F:C=D8H501O1000O01000O0100O01000O10O10O10O10O1000O0101O8H9H8G5L3L5L4L4M2M4M3M1NO2N2O0O2N1O2O0O2N2N3M2O1N3M2N2N5L5J6J6J6I7D=C=D<F:E9G6J7I7J6Immj2"}, "label": "a woman who is not holding an umbrella"}]}
{"id": 244095, "image": "COCO_train2014_000000244095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man in red shorts catches air on his skateboard\"."}], "task": "refering", "answer_template": "The \"a young man in red shorts catches air on his skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 194, 195, 197, 198, 199, 200, 216, 217, 221, 222, 239, 240, 241], "bbox": [0.322546875, 0.10241304347826087, 0.7026562500000001, 0.6864565217391305], "iscrowd": 0, "area": 26438.577549999987, "rle": {"size": [460, 640], "counts": "oTm28Q>8H7J7H4M3M3N2M3N2N2M3N2M3N2N2M3N1N3N1N2O2N1N2O2M2O2M2O1O2N1O1N3N1O2N1N2O2N1O1O2M2O1O1O1N3N1O1O1O2M2VOZLmFg3l1ZL`42aIe3j1^Ld41^Ib3l1`Ld41]Ia3l1aLe45XI[3R2bLd4X4[KiKd4X4\\KiKb4X4]KiKc4X4\\KiKc4W4\\KjKd4V4[KlKd4T4ZKnKg4Q4WKQLi4P4UKRLj4n3TKTLl4l3SKVLl4j3RKXLn4h3QKZLo4f3nJ\\LR5d3mJ^LR5b3lJ`LT5`3kJaLU5_3iJdLW5\\3fJfLZ5Z3eJhLZ5X3dJjL\\5V3cJlL\\5T3bJnL^5S3_JnLc5Q3[JQMe5o2ZJRMf5n2XJTMh5l2VJVMk5j2RJWMo5i2PJXMP6h2nIZMR6f2lIaL^O^Og6P4iIcLDZOc6T4fIcLJWO`6V4eIdLLTO`6X4bIeL1QO^6Y4_IeL7QOZ6Z4]IeL=oNV6]4[IcLb0POS6\\4ZIbLh0POn5^4YIaLl0POk5_4XI_LR1POf5a4WI^LU1QOe5`4UI^LY1QOb5g5_JYJa5f5aJYJ_5f5bJZJ^5d5eJ[JZ5e5hJZJX5e5`1XJkGh5V86001N2O001O001O1O1N2O1O1O1O1O2NN1OQLZG]2d8fM_GV2a8jMnGg1R8\\NQH^1P8eNoGX1Q8kNPHQ1P8SOPHh0R8YOoGd0Q8^OPH?P8BRH;o7ERH9n7GUH5m7KUH1m7NVHOk70XHLj74WHIk76XHFj7:XHCj7=VHAk7?VHAi7>XHCg7=ZHCe7=\\HCc7>\\HCc7=^HCa7<`HF^7:cHF\\79fHGY79hHGW79kHFT7:oHDP7;SIDm69XIEg69]IGa67cIH]65gIJX66kIHU66nIIQ67QJHn57UJHk57WJHh56]JHc56_JK`52cJN_5NdJ1_5JcJ6_5FcJ:_5BdJ=]5@eJb0Z5ZOiJh0V5SOoJm0Q5oNRKS1m4jNUKW1j4fNYKZ1n80O1O001M3N4L3M2N3M3M2O0O10O001O0O2O0hNgC=Z<@iC>X<AiC?W<VOdCO7j0T<YOfCI7o0S<XOTDg0m;YOSDg0m;POfC3<n0U<ROjCn0W<ROhCm0Z<ROfCn0[<ROdCm0]<TObCl0`<SO_Cl0d<RO\\Cn0Q=O1O100O100O100O1O1O1O1O001O1O001O1O1O001O1O1O001O1O1O1O1O2N2N2MdSe2"}, "label": "a young man in red shorts catches air on his skateboard"}]}
{"id": 244095, "image": "COCO_train2014_000000244095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skater in red shorts riding a skateboard in the air\"."}], "task": "refering", "answer_template": "The \"a skater in red shorts riding a skateboard in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 194, 195, 197, 198, 199, 200, 216, 217, 221, 222, 239, 240, 241], "bbox": [0.322546875, 0.10241304347826087, 0.7026562500000001, 0.6864565217391305], "iscrowd": 0, "area": 26438.577549999987, "rle": {"size": [460, 640], "counts": "oTm28Q>8H7J7H4M3M3N2M3N2N2M3N2M3N2N2M3N1N3N1N2O2N1N2O2M2O2M2O1O2N1O1N3N1O2N1N2O2N1O1O2M2O1O1O1N3N1O1O1O2M2VOZLmFg3l1ZL`42aIe3j1^Ld41^Ib3l1`Ld41]Ia3l1aLe45XI[3R2bLd4X4[KiKd4X4\\KiKb4X4]KiKc4X4\\KiKc4W4\\KjKd4V4[KlKd4T4ZKnKg4Q4WKQLi4P4UKRLj4n3TKTLl4l3SKVLl4j3RKXLn4h3QKZLo4f3nJ\\LR5d3mJ^LR5b3lJ`LT5`3kJaLU5_3iJdLW5\\3fJfLZ5Z3eJhLZ5X3dJjL\\5V3cJlL\\5T3bJnL^5S3_JnLc5Q3[JQMe5o2ZJRMf5n2XJTMh5l2VJVMk5j2RJWMo5i2PJXMP6h2nIZMR6f2lIaL^O^Og6P4iIcLDZOc6T4fIcLJWO`6V4eIdLLTO`6X4bIeL1QO^6Y4_IeL7QOZ6Z4]IeL=oNV6]4[IcLb0POS6\\4ZIbLh0POn5^4YIaLl0POk5_4XI_LR1POf5a4WI^LU1QOe5`4UI^LY1QOb5g5_JYJa5f5aJYJ_5f5bJZJ^5d5eJ[JZ5e5hJZJX5e5`1XJkGh5V86001N2O001O001O1O1N2O1O1O1O1O2NN1OQLZG]2d8fM_GV2a8jMnGg1R8\\NQH^1P8eNoGX1Q8kNPHQ1P8SOPHh0R8YOoGd0Q8^OPH?P8BRH;o7ERH9n7GUH5m7KUH1m7NVHOk70XHLj74WHIk76XHFj7:XHCj7=VHAk7?VHAi7>XHCg7=ZHCe7=\\HCc7>\\HCc7=^HCa7<`HF^7:cHF\\79fHGY79hHGW79kHFT7:oHDP7;SIDm69XIEg69]IGa67cIH]65gIJX66kIHU66nIIQ67QJHn57UJHk57WJHh56]JHc56_JK`52cJN_5NdJ1_5JcJ6_5FcJ:_5BdJ=]5@eJb0Z5ZOiJh0V5SOoJm0Q5oNRKS1m4jNUKW1j4fNYKZ1n80O1O001M3N4L3M2N3M3M2O0O10O001O0O2O0hNgC=Z<@iC>X<AiC?W<VOdCO7j0T<YOfCI7o0S<XOTDg0m;YOSDg0m;POfC3<n0U<ROjCn0W<ROhCm0Z<ROfCn0[<ROdCm0]<TObCl0`<SO_Cl0d<RO\\Cn0Q=O1O100O100O100O1O1O1O1O001O1O001O1O1O001O1O1O001O1O1O1O1O2N2N2MdSe2"}, "label": "a skater in red shorts riding a skateboard in the air"}]}
{"id": 129407, "image": "COCO_train2014_000000129407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall blue vase\"."}], "task": "refering", "answer_template": "The \"a tall blue vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194], "bbox": [0.31440625, 0.15738317757009346, 0.470421875, 0.5721261682242991], "iscrowd": 0, "area": 14012.311249999999, "rle": {"size": [428, 640], "counts": "_Rd23Y=8I7H9G8H8H9G8H8H9G8H8H9G8H8H8H9H7H8H9G4L2N1O1O1O1O1O1O1O001O0001O0000000000000001O00000000000001O000001O0000000000000000000010O00000O1O100O100O1O100O100O1O101N100O100K5J6J6J6J6G9F:E;F:E;F;D;F:D<A?@Zd]4"}, "label": "a tall blue vase"}]}
{"id": 129407, "image": "COCO_train2014_000000129407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue vase with a plant\"."}], "task": "refering", "answer_template": "The \"a blue vase with a plant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194], "bbox": [0.31440625, 0.15738317757009346, 0.470421875, 0.5721261682242991], "iscrowd": 0, "area": 14012.311249999999, "rle": {"size": [428, 640], "counts": "_Rd23Y=8I7H9G8H8H9G8H8H9G8H8H9G8H8H8H9H7H8H9G4L2N1O1O1O1O1O1O1O001O0001O0000000000000001O00000000000001O000001O0000000000000000000010O00000O1O100O100O1O100O100O1O101N100O100K5J6J6J6J6G9F:E;F:E;F;D;F:D<A?@Zd]4"}, "label": "a blue vase with a plant"}]}
{"id": 350596, "image": "COCO_train2014_000000350596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of the grey sleeve person\"."}], "task": "refering", "answer_template": "The \"the arm of the grey sleeve person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273], "bbox": [0.8022812500000001, 0.5308665105386418, 1.0, 0.8018266978922717], "iscrowd": 0, "area": 4394.53905, "rle": {"size": [427, 640], "counts": "idf63X=O10O100O10000O100O10000O010O10O10O10O010001N100O100O100O100O1O2N2N2N2N2N1O2N1O2hCUOm;l0QDWOm;k0QDWOm;j0QDXOo;S1N0O010000O010O1O010O101N100O1001O001O01O01O001O001O00010O0O2N1O1O20O100O1O100O01O0O101O0010O0010O0100O010O10O100O001O1O001O1O1O001O1O1O00001O00010O01O01O01O01N1R1nNlG"}, "label": "the arm of the grey sleeve person"}]}
{"id": 350596, "image": "COCO_train2014_000000350596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of someone wearing a gray sweatshirt\"."}], "task": "refering", "answer_template": "The \"the arm of someone wearing a gray sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273], "bbox": [0.8022812500000001, 0.5308665105386418, 1.0, 0.8018266978922717], "iscrowd": 0, "area": 4394.53905, "rle": {"size": [427, 640], "counts": "idf63X=O10O100O10000O100O10000O010O10O10O10O010001N100O100O100O100O1O2N2N2N2N2N1O2N1O2hCUOm;l0QDWOm;k0QDWOm;j0QDXOo;S1N0O010000O010O1O010O101N100O1001O001O01O01O001O001O00010O0O2N1O1O20O100O1O100O01O0O101O0010O0010O0100O010O10O100O001O1O001O1O1O001O1O1O00001O00010O01O01O01O01N1R1nNlG"}, "label": "the arm of someone wearing a gray sweatshirt"}]}
{"id": 350596, "image": "COCO_train2014_000000350596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with black pants carrying a back pack\"."}], "task": "refering", "answer_template": "The \"a person with black pants carrying a back pack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 136, 137, 159, 160, 182, 183, 205, 206], "bbox": [0.91034375, 0.0008665105386416862, 1.0, 0.6079625292740047], "iscrowd": 0, "area": 9879.712949999996, "rle": {"size": [427, 640], "counts": "]Sc74W=5K6J5K5K2N3M3M2N3M2N2N1O1SIiNd1X1eMNk13oM=f1DfLiNdNV2X4ROoLSO]NV2^4hNPMh2Q3YMjLi2X3XMdLj2]3WM^Ll2c3UMXLo2h3RMSLQ3n3QMlKR3U4PMeKT3[4nL`KT3a4j2100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1"}, "label": "a person with black pants carrying a back pack"}]}
{"id": 350596, "image": "COCO_train2014_000000350596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person leaning on a brick wall wearing a tan jacket and white tee shirt\"."}], "task": "refering", "answer_template": "The \"a person leaning on a brick wall wearing a tan jacket and white tee shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 130, 131, 132, 133, 153, 154, 155, 156, 177, 178, 179, 201, 202, 225, 248, 249], "bbox": [0.49515624999999996, 0.003583138173302108, 0.834734375, 0.7121077283372366], "iscrowd": 0, "area": 28662.53410000001, "rle": {"size": [427, 640], "counts": "YZT41X=4M2N2M4M1N2O1O2M2O1O1N2O1O1N2O11O00001O01O01O0000001O000O2O00001O0O1000O1000O0100O1O001N2N1N3N1O1N3N1N2cNQO^FQ1^9YOZFh0c9CSF?j9KnE7n9d1M4N1O1N3N1O1O2N1O1N2O3M4L4L4L3O2N2N2N00000000O1000000000000O100000000000000O1000000000000O12N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N00O10000O100O100O12N2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2NZNRKQKl4m4YKQKf4g0WKf19bM^4d0_Ke19fMV4a0gKd19iMo3?nKc19mMg3<WLa18RN_39_La16VNY36fL_17YNR34mL^17]Nj21VM\\16bNb2N^M[16fNZ2KfMZ16iNS2InMY13nNm1TOgNi1BROQ5i0UKUOl4f0ZKYOg4a0_K^Of48`KGh4M]K1l4BYK>o4UOWKj0g8L4L4L4K4M4LYm[1"}, "label": "a person leaning on a brick wall wearing a tan jacket and white tee shirt"}]}
{"id": 350596, "image": "COCO_train2014_000000350596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a brown jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing a brown jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 130, 131, 132, 133, 153, 154, 155, 156, 177, 178, 179, 201, 202, 225, 248, 249], "bbox": [0.49515624999999996, 0.003583138173302108, 0.834734375, 0.7121077283372366], "iscrowd": 0, "area": 28662.53410000001, "rle": {"size": [427, 640], "counts": "YZT41X=4M2N2M4M1N2O1O2M2O1O1N2O1O1N2O11O00001O01O01O0000001O000O2O00001O0O1000O1000O0100O1O001N2N1N3N1O1N3N1N2cNQO^FQ1^9YOZFh0c9CSF?j9KnE7n9d1M4N1O1N3N1O1O2N1O1N2O3M4L4L4L3O2N2N2N00000000O1000000000000O100000000000000O1000000000000O12N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N00O10000O100O100O12N2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2NZNRKQKl4m4YKQKf4g0WKf19bM^4d0_Ke19fMV4a0gKd19iMo3?nKc19mMg3<WLa18RN_39_La16VNY36fL_17YNR34mL^17]Nj21VM\\16bNb2N^M[16fNZ2KfMZ16iNS2InMY13nNm1TOgNi1BROQ5i0UKUOl4f0ZKYOg4a0_K^Of48`KGh4M]K1l4BYK>o4UOWKj0g8L4L4L4K4M4LYm[1"}, "label": "a man wearing a brown jacket"}]}
{"id": 285064, "image": "COCO_train2014_000000285064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in her thirtees wearing light ash color dress making announcement\"."}], "task": "refering", "answer_template": "The \"a woman in her thirtees wearing light ash color dress making announcement\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 123, 141, 142, 143, 145, 146, 147, 164, 165, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332], "bbox": [0.15592187500000002, 0.2949414519906323, 0.47978125, 0.9915690866510538], "iscrowd": 0, "area": 34945.947199999995, "rle": {"size": [427, 640], "counts": "dnY1k0P<a0J6I7nM[NXHj1e7[NTHi1k7[NnGj1P8ZNjGj1S8[NfGj1X8\\N_Gh1`8_NWGd1g8cNPGa1o8^1O1N2O1O1N2C=0000O10000O11O2N1O2NN2M3M3N2M3M3N3N1mKdJ7m3RL\\Me3jN9[3aLkMS3mN<i2oL[Na2PO`0V2]MjNQ2SOb0j1fMSOd1WOf0=PL6g17`1YOi07SL6g19[1]Ok02UL5g1=U1@P1JWL6f1a0P1BS1DZL6f1c0k0FU1_O]L5d1g0f0IY1XO`L7a1i0c0K\\1ROcL9_1i0?0_1kNfL:]1l0;2b1fNhL;\\1l086d1`NlL;Y1P139j2fNSMR10;m2aNUMS1Ja0Q3[NUMU1[Oo0`3jMWMW1VOR1d3eMVMY1eNd1U4QMWM\\1bNe1W4nLXM]1^Nh1Z4iLYM_1\\Nj1k6XNPIk1o6\\NiHf1V7bN`Ha1`7U2O100O1O100001O000000001O000000001O000O1000001O000000001O000000001O00000000001O0000O100O1O1O100O1O100O1O100O1ZNVKjJk4T5^KeJb4Y5eKaJ\\4]5j1O1N2N200O100O10000O100O10O10000000000000000000000000001O00001O00001O00001O00001fJbJe2^5VMhJi2X5RMnJm2S5mLRKS3n4hLXKW3h4hLZKW3g4hLZKW3f4iL[KV3g4hLYKX3i4fLXKY3k4dLVK[3l4cLUKT2SO_Mk5:SKU2WO]Mh5<QKW2[OYMf5>PKW2_OWMd5?nJY2BTMb5a0mJZ2EQMa5b0kJ[2IoL^5d0iJ]2MkLa7S3dHiL]7V3]1N2O001M3L4L4L3M4L5K5K6J5L4K6J5K5K6J5K5K6J5K5K6JiaZ4"}, "label": "a woman in her thirtees wearing light ash color dress making announcement"}]}
{"id": 285064, "image": "COCO_train2014_000000285064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in brown is holding two wii remotes as she moves to play the game\"."}], "task": "refering", "answer_template": "The \"a woman in brown is holding two wii remotes as she moves to play the game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 123, 141, 142, 143, 145, 146, 147, 164, 165, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332], "bbox": [0.15592187500000002, 0.2949414519906323, 0.47978125, 0.9915690866510538], "iscrowd": 0, "area": 34945.947199999995, "rle": {"size": [427, 640], "counts": "dnY1k0P<a0J6I7nM[NXHj1e7[NTHi1k7[NnGj1P8ZNjGj1S8[NfGj1X8\\N_Gh1`8_NWGd1g8cNPGa1o8^1O1N2O1O1N2C=0000O10000O11O2N1O2NN2M3M3N2M3M3N3N1mKdJ7m3RL\\Me3jN9[3aLkMS3mN<i2oL[Na2PO`0V2]MjNQ2SOb0j1fMSOd1WOf0=PL6g17`1YOi07SL6g19[1]Ok02UL5g1=U1@P1JWL6f1a0P1BS1DZL6f1c0k0FU1_O]L5d1g0f0IY1XO`L7a1i0c0K\\1ROcL9_1i0?0_1kNfL:]1l0;2b1fNhL;\\1l086d1`NlL;Y1P139j2fNSMR10;m2aNUMS1Ja0Q3[NUMU1[Oo0`3jMWMW1VOR1d3eMVMY1eNd1U4QMWM\\1bNe1W4nLXM]1^Nh1Z4iLYM_1\\Nj1k6XNPIk1o6\\NiHf1V7bN`Ha1`7U2O100O1O100001O000000001O000000001O000O1000001O000000001O000000001O00000000001O0000O100O1O1O100O1O100O1O100O1ZNVKjJk4T5^KeJb4Y5eKaJ\\4]5j1O1N2N200O100O10000O100O10O10000000000000000000000000001O00001O00001O00001O00001fJbJe2^5VMhJi2X5RMnJm2S5mLRKS3n4hLXKW3h4hLZKW3g4hLZKW3f4iL[KV3g4hLYKX3i4fLXKY3k4dLVK[3l4cLUKT2SO_Mk5:SKU2WO]Mh5<QKW2[OYMf5>PKW2_OWMd5?nJY2BTMb5a0mJZ2EQMa5b0kJ[2IoL^5d0iJ]2MkLa7S3dHiL]7V3]1N2O001M3L4L4L3M4L5K5K6J5L4K6J5K5K6J5K5K6J5K5K6JiaZ4"}, "label": "a woman in brown is holding two wii remotes as she moves to play the game"}]}
{"id": 285064, "image": "COCO_train2014_000000285064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white polo shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a white polo shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 175, 176, 177, 178, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.60821875, 0.4073536299765808, 0.8995, 0.9873536299765808], "iscrowd": 0, "area": 30207.54455, "rle": {"size": [427, 640], "counts": "[`R54U=:F:F:F;D;F:F:F6K3L5L4L4L3L5L4L3M4L4K5L3M1]NPLjIQ4S5QLVJ6c0k3V5[LfI2Q1d3Y5VMdJk2[5XMbJi2^5YM^Ji2a5ZM\\Jh2b5[MUJl2k5VMkIR3T6S2O100O1O1RJQI`5o6]JTIb5m6[JVId5k6YJXIg5V7O1O100O1O100O10000000000000000000000000000001O001O001O001O1O001O001O001O1O001O1O2N1O2N1O2N2N1O2N1O2N2iIdI[5]6\\JkIe5V6QJRJo5f6O000000001O000000001O000000001O000000001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001_LRHc1o7]NTH_1m7aNVH[1j7fNYHV1h7jN\\HQ1e7oN^Hm0c7SO`Hi0`7XOcHd0^7\\OeH`0\\7@gH<Z7DjH7V7JmH2T7NoHNR72PIKQ74QIJo66SIHn68SIFn69TIEm6:UIDk6<VICl6;VICk6<WIBj6>WI@k6>WI@j6?XI_Oj6?XI_Oi6`0XI_Oj6?XI_Oi6a0XI]Oj6a0XI]Oi6b0YI\\Oh6c0ZI[Oh6c0YI\\Oh6c0ZI[Oh6d0YI[Og6d0[IZOg6d0[IZOf6e0\\IYOf6e0[IZOf6e0\\IYOf6f0[IXOf6e0^IYOc6d0aIYOb6d0T3M4L3N3L3N3Lhij0"}, "label": "a woman in a white polo shirt"}]}
{"id": 285064, "image": "COCO_train2014_000000285064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 175, 176, 177, 178, 199, 200, 201, 202, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.60821875, 0.4073536299765808, 0.8995, 0.9873536299765808], "iscrowd": 0, "area": 30207.54455, "rle": {"size": [427, 640], "counts": "[`R54U=:F:F:F;D;F:F:F6K3L5L4L4L3L5L4L3M4L4K5L3M1]NPLjIQ4S5QLVJ6c0k3V5[LfI2Q1d3Y5VMdJk2[5XMbJi2^5YM^Ji2a5ZM\\Jh2b5[MUJl2k5VMkIR3T6S2O100O1O1RJQI`5o6]JTIb5m6[JVId5k6YJXIg5V7O1O100O1O100O10000000000000000000000000000001O001O001O001O1O001O001O001O1O001O1O2N1O2N1O2N2N1O2N1O2N2iIdI[5]6\\JkIe5V6QJRJo5f6O000000001O000000001O000000001O000000001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001_LRHc1o7]NTH_1m7aNVH[1j7fNYHV1h7jN\\HQ1e7oN^Hm0c7SO`Hi0`7XOcHd0^7\\OeH`0\\7@gH<Z7DjH7V7JmH2T7NoHNR72PIKQ74QIJo66SIHn68SIFn69TIEm6:UIDk6<VICl6;VICk6<WIBj6>WI@k6>WI@j6?XI_Oj6?XI_Oi6`0XI_Oj6?XI_Oi6a0XI]Oj6a0XI]Oi6b0YI\\Oh6c0ZI[Oh6c0YI\\Oh6c0ZI[Oh6d0YI[Og6d0[IZOg6d0[IZOf6e0\\IYOf6e0[IZOf6e0\\IYOf6f0[IXOf6e0^IYOc6d0aIYOb6d0T3M4L3N3L3N3Lhij0"}, "label": "a woman wearing a white shirt"}]}
{"id": 301461, "image": "COCO_train2014_000000301461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue back end of a large truck behind a horse\"."}], "task": "refering", "answer_template": "The \"the blue back end of a large truck behind a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140], "bbox": [0.0, 0.003371559633027523, 0.220953125, 0.4222935779816514], "iscrowd": 0, "area": 16244.596550000002, "rle": {"size": [436, 640], "counts": "1f5n7000001O01O00000000000000000000000000000000000000000000O100000000000000000001O0000000000000O10000000000O100[M[HZOf7f0gHmNY7R1kHjNV7U1nHhNR7W1PIgNQ7X1QIfNP7Z1PIfNP7Y1QIfNP7Y1RIdNP7[1QIbNR7^1nH_NU7`1kH]NY7b1hHYN]7f1dHVN`7j1`HQNe7n1\\HoMg7Q2XHoMi7P2XHnMj7R2YHiMi7V2[HeMh7Z2i100O100O100O100O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100000000000000O2O000000000000000O10000000000000000O10000000000000000O1O1C=B>CT_d6"}, "label": "the blue back end of a large truck behind a horse"}]}
{"id": 301461, "image": "COCO_train2014_000000301461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue truck with a horse standing next to it\"."}], "task": "refering", "answer_template": "The \"a blue truck with a horse standing next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140], "bbox": [0.0, 0.003371559633027523, 0.220953125, 0.4222935779816514], "iscrowd": 0, "area": 16244.596550000002, "rle": {"size": [436, 640], "counts": "1f5n7000001O01O00000000000000000000000000000000000000000000O100000000000000000001O0000000000000O10000000000O100[M[HZOf7f0gHmNY7R1kHjNV7U1nHhNR7W1PIgNQ7X1QIfNP7Z1PIfNP7Y1QIfNP7Y1RIdNP7[1QIbNR7^1nH_NU7`1kH]NY7b1hHYN]7f1dHVN`7j1`HQNe7n1\\HoMg7Q2XHoMi7P2XHnMj7R2YHiMi7V2[HeMh7Z2i100O100O100O100O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100000000000000O2O000000000000000O10000000000000000O10000000000000000O1O1C=B>CT_d6"}, "label": "a blue truck with a horse standing next to it"}]}
{"id": 293272, "image": "COCO_train2014_000000293272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red umbrella\"."}], "task": "refering", "answer_template": "The \"the red umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 122, 127], "bbox": [0.0013124999999999999, 0.04494652406417112, 0.53840625, 0.39999999999999997], "iscrowd": 0, "area": 27925.615149999998, "rle": {"size": [374, 640], "counts": "c<U3`82O00000000001O0O1000000000001O000O10000000001O3M3M1N100O1N2M3N200000000000000001O1O001O1O002N1O1O00001O00001O01O010O010O00100O00N3M2N210O00000000001O00000000000001O01O00000000000000001O0001O000000000001O00000000000000001O00000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000001O00001O00001O0000001O00001O01O0001O00001O000010O000001O00001O0010O0001O00001O001O01O01O00001O00001O010O00001O000O2O001O000O2O00001N101O000O2O00001N2O001O1N2O001O1N2O001O1N101O1O1N101O1O1N101O1O1O001O1O1O0O2O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O0O2O1O1O001O1O01000O10O10O10O10O1000O01000O01O1O1O0O2O1O001O1N2O001O1O0O2O1O1O001N2O001O1O1O0O2O10O10000O01000O100O01O1N1OPo[3"}, "label": "the red umbrella"}]}
{"id": 293272, "image": "COCO_train2014_000000293272.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food truck with the safeway logo on it on a rainy street\"."}], "task": "refering", "answer_template": "The \"a food truck with the safeway logo on it on a rainy street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 185, 186, 187, 188, 209, 210, 211], "bbox": [0.005734375, 0.36994652406417117, 0.23246875, 0.7665508021390376], "iscrowd": 0, "area": 11673.433349999998, "rle": {"size": [374, 640], "counts": "[e13`;4L4L4M3L4M3L4L4M3L4L4M3L4M3L4N2O00100O1O1O1O1O1O1ROXNlGi1R8o0N2N2O10000000000000000002N3M1OO1O00100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O000000000001O000000000001O00000001O00000001O000001O001O1O00100O001O1O0010O100O1O1O100O1O1O0105J5K=C4LON3M2N3H8G81000O10000dN`Fb0`9[OfFa0\\9_OgF;[9EhF5[9KTGDn8=V1O00010O04M1O0O10O101N2M3N3M2M4LiSc5"}, "label": "a food truck with the safeway logo on it on a rainy street"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a birthday cake with two carrots on it\"."}], "task": "refering", "answer_template": "The \"a birthday cake with two carrots on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 360, 374, 375, 376, 377, 378, 379, 380, 381, 382, 396, 397, 398, 399, 400, 401, 402, 403, 404, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 440, 441, 442, 443, 444, 445, 446, 447, 448, 464, 465, 466, 467, 468], "bbox": [0.02696078431372549, 0.7168464052287581, 0.4134803921568628, 0.9707843137254902], "iscrowd": 0, "area": 31601.018000000007, "rle": {"size": [612, 612], "counts": "cd:n1Ua03L3N3L3N3L4M2N3L4M2M4M2M4M1N3N1O1N3N1N3N1O2O0O2O0O101N101N101N101N101N100O2O0O2O0O2O0O2O0O101N101N100O100O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O10000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000O10000O10000O101O0O10000O100O10000O10000O10000O10000O10001N10001O0O101O000O2O00001N10001O0O10001N10001O0O102N2N2M2O2N2N1N3N2N1O2M2O1O2N1N3N1O1N3N1N2]Mi_OU2[`0hMi_OT2Y`0jMk_OR2W`0lMm_Oo1h`0L4L4K5L4K5L4WOYRf6"}, "label": "a birthday cake with two carrots on it"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cake with carrots on it\"."}], "task": "refering", "answer_template": "The \"the cake with carrots on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 360, 374, 375, 376, 377, 378, 379, 380, 381, 382, 396, 397, 398, 399, 400, 401, 402, 403, 404, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 440, 441, 442, 443, 444, 445, 446, 447, 448, 464, 465, 466, 467, 468], "bbox": [0.02696078431372549, 0.7168464052287581, 0.4134803921568628, 0.9707843137254902], "iscrowd": 0, "area": 31601.018000000007, "rle": {"size": [612, 612], "counts": "cd:n1Ua03L3N3L3N3L4M2N3L4M2M4M2M4M1N3N1O1N3N1N3N1O2O0O2O0O101N101N101N101N101N100O2O0O2O0O2O0O2O0O101N101N100O100O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O10000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000O10000O10000O101O0O10000O100O10000O10000O10000O10000O10001N10001O0O101O000O2O00001N10001O0O10001N10001O0O102N2N2M2O2N2N1N3N2N1O2M2O1O2N1N3N1O1N3N1N2]Mi_OU2[`0hMi_OT2Y`0jMk_OR2W`0lMm_Oo1h`0L4L4K5L4K5L4WOYRf6"}, "label": "the cake with carrots on it"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child in a pink shirt\"."}], "task": "refering", "answer_template": "The \"a small child in a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281], "bbox": [0.5681699346405229, 0.13748366013071894, 0.8719607843137255, 0.5680718954248366], "iscrowd": 0, "area": 35001.732500000006, "rle": {"size": [612, 612], "counts": "]X`6:gb0?A?A7J2M4L3M4L3M4L3ROoMg@T2U?^NY@e1e?o0O0O2N101N101N3N3L4L4M0O2O0O101N1O100O2O0O101N100lMnKUET4h:\\LiDe3T;cLgD]3V;kLdDV3Z;QMaDo2\\;XM_Di2_;[M_De2^;`M_D`2`;dM]D]2`;hM^DX2`;lM^DS2`;RN]Do1a;TN^Dk1`;ZN^Df1_;_N_Da1_;R3O200O010O100O2O1N100O100O2O0O100O100O101N100O100O101N100O100000001OO1O1O1O1O1O1O1O1O1O01000000000000000000000000000O0100000000001O00001O00001O00001O0O101O00001O00001O00001O00001O0jKaDg0_;XOdDe0^;YOeDe0[;ZOgDe0Y;[OhDc0Z;[OgDd0[;ZOgDd0[;ZOfDe0];YOdDe0_;XObDg0`;WObDg0a;VOaDh0a;WO_Dh0d;UO^Dj0c;TO^Dk0e;RO]Dl0f;oN]DP1e;kN`DS1c;fNcDX1_;cNeD\\1^;^NgD`1[;]NgDb1\\;ZNgDd1\\;XNfDh1c;mM`DQ2j;dMZDY2P<\\MSDb2Z>E:F;E:E<E:K6L4L3M4L3M2N3L3N3M[o^1"}, "label": "a small child in a pink shirt"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a pink shirt\"."}], "task": "refering", "answer_template": "The \"a girl in a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281], "bbox": [0.5681699346405229, 0.13748366013071894, 0.8719607843137255, 0.5680718954248366], "iscrowd": 0, "area": 35001.732500000006, "rle": {"size": [612, 612], "counts": "]X`6:gb0?A?A7J2M4L3M4L3M4L3ROoMg@T2U?^NY@e1e?o0O0O2N101N101N3N3L4L4M0O2O0O101N1O100O2O0O101N100lMnKUET4h:\\LiDe3T;cLgD]3V;kLdDV3Z;QMaDo2\\;XM_Di2_;[M_De2^;`M_D`2`;dM]D]2`;hM^DX2`;lM^DS2`;RN]Do1a;TN^Dk1`;ZN^Df1_;_N_Da1_;R3O200O010O100O2O1N100O100O2O0O100O100O101N100O100O101N100O100000001OO1O1O1O1O1O1O1O1O1O01000000000000000000000000000O0100000000001O00001O00001O00001O0O101O00001O00001O00001O00001O0jKaDg0_;XOdDe0^;YOeDe0[;ZOgDe0Y;[OhDc0Z;[OgDd0[;ZOgDd0[;ZOfDe0];YOdDe0_;XObDg0`;WObDg0a;VOaDh0a;WO_Dh0d;UO^Dj0c;TO^Dk0e;RO]Dl0f;oN]DP1e;kN`DS1c;fNcDX1_;cNeD\\1^;^NgD`1[;]NgDb1\\;ZNgDd1\\;XNfDh1c;mM`DQ2j;dMZDY2P<\\MSDb2Z>E:F;E:E<E:K6L4L3M4L3M2N3L3N3M[o^1"}, "label": "a girl in a pink shirt"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of an off colored green chair with a black haired child sitting on it\"."}], "task": "refering", "answer_template": "The \"the back of an off colored green chair with a black haired child sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 210, 216, 217, 218, 219, 232, 238, 239, 240, 241, 261, 262, 263], "bbox": [0.5406209150326797, 0.3665849673202614, 0.9830555555555556, 0.5132843137254902], "iscrowd": 0, "area": 4655.117799999999, "rle": {"size": [612, 612], "counts": "VSV65jb07I7J6I8H6J5L4K6J5K5L4K6O00I[_OoMe`0m1;M3M3L5L3M3L4M3N3M4L3M3N3L3M4L3M4L3M[Sl2OdlSM4L4ROGl^O>Ta0O[^O5da0f0000001O01O0000000000000010O000000000001O0000000001O0001O000000000000001O0001O00000001O000000000000O2M2M3N2N200001O000000001O01O0001O0000001O0N2I7H9H7I7H8ITg5"}, "label": "the back of an off colored green chair with a black haired child sitting on it"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green chair on which the kid with pink dress sitting\"."}], "task": "refering", "answer_template": "The \"the green chair on which the kid with pink dress sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 210, 216, 217, 218, 219, 232, 238, 239, 240, 241, 261, 262, 263], "bbox": [0.5406209150326797, 0.3665849673202614, 0.9830555555555556, 0.5132843137254902], "iscrowd": 0, "area": 4655.117799999999, "rle": {"size": [612, 612], "counts": "VSV65jb07I7J6I8H6J5L4K6J5K5L4K6O00I[_OoMe`0m1;M3M3L5L3M3L4M3N3M4L3M3N3L3M4L3M4L3M[Sl2OdlSM4L4ROGl^O>Ta0O[^O5da0f0000001O01O0000000000000010O000000000001O0000000001O0001O000000000000001O0001O00000001O000000000000O2M2M3N2N200001O000000001O01O0001O0000001O0N2I7H9H7I7H8ITg5"}, "label": "the green chair on which the kid with pink dress sitting"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl with white shirt with a design on it sitting at the table with a little girl with a pink shirt\"."}], "task": "refering", "answer_template": "The \"little girl with white shirt with a design on it sitting at the table with a little girl with a pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 71, 72, 73, 74, 92, 93, 94, 95, 96, 115, 116, 117, 118, 136, 137, 138, 139, 140, 141, 158, 159, 160, 161, 162, 163, 164, 180, 181, 182, 183, 184, 185, 186, 202, 203, 204, 205, 206, 207, 208, 209, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.19101307189542485, 0.09594771241830065, 0.5348366013071896, 0.4981862745098039], "iscrowd": 0, "area": 31334.271549999998, "rle": {"size": [612, 612], "counts": "oSV23m14o>3`@;\\?LS@b0i?Fd_Oi0X`0T1M4L4L4L3M4L1O100O1O1O1O1QNiLoCX3m;oLPDR3l;UMoCm2l;[MPDf2k;aMQDa2k;fMQD[2j;lMRDV2i;RNSDo1k;UNRDl1m;XNoCh1P<\\NmCe1R<_NjCb1T<bNiC_1U<fNgC[1X<hNeCY1Y<j2O1N101N2O1O1O100O1O1O010O1O1O1O10000O10000000000O10000000000O10000001O001O1O001O001O001O1O0000O10000O1000000O10000O10000O1000000O10001N10000O2O0000010O01nKZDd0f;\\OZDd0g;ZOZDg0e;XO\\Dh0e;WO[Di0f;UO[Dk0e;TO\\Dl0e;SO[Dn0d;QO]Do0d;oN]DQ1f;kN[DU1g;iNYDW1j;eNWD[1k;bNVD^1m;XN[Dg1h;PN^DP2d;gMcDY2`;^MgDa2[;\\MfDd2];WMfDh2\\;UMeDk2^;PMdDP3_;lLcDS3`;dLfD\\3b=O01O01kNn@VNR?`1YA_Nh>W1aAiN_>n0kAQOU>i0RBUOP>k0PBTOP>m0PBROQ>n0oAPOR>R1mAmNS>T1mAkNT>U1lAiNU>X1kAgNV>Y1jAfNV>[1jAdNW>]1hAaNY>`1gA_NY>b1gA]N^>_1bA_Ne>\\1ZAdNk>X1UAgNl>Y1TAeNn>[1RAdNP?[1PAdNQ?\\1o@cNR?]1n@aNU?_1j@`NW?`1X1N10001N10001N10001N101O0O10]^OmNTa0S1g^OROYa0m0b^OZO^a0W1O0010000O0100000O010000O0O2N2N2M2O2N2N2N1O1O1O2N1O1O1N3M2M3NQRZ5"}, "label": "little girl with white shirt with a design on it sitting at the table with a little girl with a pink shirt"}]}
{"id": 170398, "image": "COCO_train2014_000000170398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl in a white shirt\"."}], "task": "refering", "answer_template": "The \"a little girl in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 71, 72, 73, 74, 92, 93, 94, 95, 96, 115, 116, 117, 118, 136, 137, 138, 139, 140, 141, 158, 159, 160, 161, 162, 163, 164, 180, 181, 182, 183, 184, 185, 186, 202, 203, 204, 205, 206, 207, 208, 209, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.19101307189542485, 0.09594771241830065, 0.5348366013071896, 0.4981862745098039], "iscrowd": 0, "area": 31334.271549999998, "rle": {"size": [612, 612], "counts": "oSV23m14o>3`@;\\?LS@b0i?Fd_Oi0X`0T1M4L4L4L3M4L1O100O1O1O1O1QNiLoCX3m;oLPDR3l;UMoCm2l;[MPDf2k;aMQDa2k;fMQD[2j;lMRDV2i;RNSDo1k;UNRDl1m;XNoCh1P<\\NmCe1R<_NjCb1T<bNiC_1U<fNgC[1X<hNeCY1Y<j2O1N101N2O1O1O100O1O1O010O1O1O1O10000O10000000000O10000000000O10000001O001O1O001O001O001O1O0000O10000O1000000O10000O10000O1000000O10001N10000O2O0000010O01nKZDd0f;\\OZDd0g;ZOZDg0e;XO\\Dh0e;WO[Di0f;UO[Dk0e;TO\\Dl0e;SO[Dn0d;QO]Do0d;oN]DQ1f;kN[DU1g;iNYDW1j;eNWD[1k;bNVD^1m;XN[Dg1h;PN^DP2d;gMcDY2`;^MgDa2[;\\MfDd2];WMfDh2\\;UMeDk2^;PMdDP3_;lLcDS3`;dLfD\\3b=O01O01kNn@VNR?`1YA_Nh>W1aAiN_>n0kAQOU>i0RBUOP>k0PBTOP>m0PBROQ>n0oAPOR>R1mAmNS>T1mAkNT>U1lAiNU>X1kAgNV>Y1jAfNV>[1jAdNW>]1hAaNY>`1gA_NY>b1gA]N^>_1bA_Ne>\\1ZAdNk>X1UAgNl>Y1TAeNn>[1RAdNP?[1PAdNQ?\\1o@cNR?]1n@aNU?_1j@`NW?`1X1N10001N10001N10001N101O0O10]^OmNTa0S1g^OROYa0m0b^OZO^a0W1O0010000O0100000O010000O0O2N2N2M2O2N2N2N1O1O1O2N1O1O1N3M2M3NQRZ5"}, "label": "a little girl in a white shirt"}]}
{"id": 498079, "image": "COCO_train2014_000000498079.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bicycle\"."}], "task": "refering", "answer_template": "The \"a blue bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 198, 201, 220, 221, 222, 268, 269, 290, 291, 292, 293, 307, 308, 309, 312, 313, 314, 315, 316, 317, 318, 326, 331, 333, 334, 335, 336, 338, 339, 340, 341, 342, 348, 349, 356, 357, 359, 361, 362, 363, 364, 365, 366, 371, 372, 377, 378, 379, 380, 381, 382, 384, 385, 386, 387, 388, 389, 394, 395, 401, 402, 403, 404, 407, 408, 409, 410, 411, 412, 417, 418, 423, 424, 425, 426, 431, 432, 433, 434, 435, 436, 445, 446, 447, 454, 455, 456, 457, 458, 469, 478, 479, 480, 481], "bbox": [0.147140625, 0.3611168384879725, 0.9604687499999999, 0.9858247422680413], "iscrowd": 0, "area": 38235.99669999998, "rle": {"size": [582, 640], "counts": "\\le13ka0;E:I8J5J7J5K6J5J6K5K5K5J6K5K5K5J6L43M3M2nMSAe0P?SOWAk0l>mNZAR1i>eN^AZ1e>^NbA`1d?M4B=B?AlSP23]loM0O100O100O010O100RCDP8=nGGo7:nGKP84oG0n71oG4o7LoG8n7IPH;n7EoG<R8FkG:W8GeG:\\8HaG8`8K\\G6c8MYG4h8OTG1l82PGOQ93lFMU91kF0U9OlF2T9LmF4S9KmF6T9GnF9R9FnF;S9CnF=R9APG?Q9_OoFc0Q9\\OoD]O;W1f:[OQE@5V1k:YOQEC1T1n:YOREDMT1R;WOREGIR1U;VOTEJDQ1X;TOUEM_Oo0^;SOTE0[Ok0c;TOTE3UOh0j;TORE6POd0Q<VOPE8kNa0X<VOnD;fN>^<VOnD<bN=c<VOlD=^N=i<UOjD>[N;n<WOhD>XN:S=VOgD`0TN9W=WOfD`0QN8\\=WOdDa0mM8a=WOcDa0jM7f=VObDc0iM2i=ZO_Dd0hMOk=]O^Dd0hMK_`05a_OHb`08__ODd`0<\\_OAg`0?Z_O\\Oj`0d0<00000lN[Og@f0X?\\Of@e0Y?]Oe@d0Y?@d@a0[?Ac@T1h>nNVAT1h>nNVAT1g>POVAR1h>POVAQ1i>QOUAQ1h>ROVAP1h>SOUAo0i>SOUAn0j>TOTAn0i>VOTAl0m>SOQAm0S?Q13M3M3M1OO1O100O1O1O100O1O100O1O1OfMVAZ1j>eNXAZ1g>gNYAY1f>gN\\AX1c>hN^AX1b>hN_AW1`>iNaAW1^>iNdAV1\\>jNeAU1Z>kNgAU1X>kNjAT1U>mNkAS1U>kNnAT1Q>lNPBT1o=lNSBS1l=mNVBR1i=mNYBS1f=mN\\BR1c=nN^BR1a=nNaBQ1]=oNfBo0Z=QOgBn0Z9^NZId0^Ml0V9bN\\Ib0_Mk0T9eN[Ia0cMh0Q9iN[I?dMJeN:[:_O[I=dMKjN5V:D\\I<dMLnNOQ:K\\I:dMMSOJm90ZI:fMLXOEi95XI:fMM^O_Oe99WI;fMNP:GYH;fMOQ:GWH;jMLP:IUH;mMKn9IUH;oMKm9JSH:SNJk9LQH:VNIh9NPH9ZNHg9NoG:[NGg9OmG9^NGf90kG8`NIe9OiG9cNGe9OhG9eNGc91gG8gNGb91fG7jNGa92dG6lNHa91bG8nNG`91aG7QOG^93`G6SOF^94^G6UOF]93]G7TOI`90[G7ROLd9MYG6QO0f9KXG5nN5j9EXG5lN9m9BUG6kN;Q:_OSG5iNa0T:ZOSG4hNc0U:ZORG2iNe0V:YOPG2iNg0W:WOoF1jNi0X:VOnF0iNk0Z:VOkFNkNm0[:UOiFNkNm0]:VOhFKkNn0`:WOdFKkNl0e:ZO^FImNl0h:[OZFImNk0l:\\OWFGmNk0o:_OSFFlNk0U;_OmEFmNi0Z;AhEFmNh0_;AdEFlNg0d;C_EGkNe0i;EZEFlNc0o;FUEFkNc0S<GQEGjN`0Z<HkDHjN?^<IgDHkN<c<LaDHjN;h<M]DHkN:j<NZDIkN9m<MXDIjN:P=NTDHlN:R=MQDJlN9T=MoCJlN:V=LnCJkN:Y=LjCJmN:Z=LhCJmN;]=JfCKlN:`=KcCKmN:b=J`CMlN:g=H[CNnN:j=FXCOnN;n=CSC3mN;S>@oB5nN;V>_OkB6mN;R`0En_O;R`0En_O;R`0Fl_O;U`0Dl_O;T`0Fk_O:U`0Fl_O9U`0Fk_O:U`0Gk_O7V`0Ij_O7V`0Ji_O6X`0Ii_O6W`0Ji_O6W`0Ki_O4X`0Kh_O5X`0Lh_O2Y`0Ng_O2Y`0Ng_O2Z`0Nf_O1Z`0Of_O1Z`0Og_O0[`0Od_O1]`0Nd_O0^`00a_O0a`0N__O2b`0M__O2b`0N]_O2e`0L\\_O3e`0MZ_O3]2KZ;0ZB3Z25l;KTB0k0=m<CVB5f0`0P=\\OXB8a0d0S=UOYB<=g0V=nN[B?8k0Y=gN\\Bc04n0\\=`N^Bf0OR1_=YN_Bj0LT1a=SNaBm0GX1d=lMbBd3Z=]LdBh3X=YLeBm3V=ULhBP4T=f0K4L3O0PMdJTI[5i6fJZIZ5b6gJaIX5\\6iJgIV5V6kJmIU5\\5gJkG5m2R5W5jJiG5S3Q5R5kJhG4Z3P5m4mJfG4_3P5j4lJdG5c3P5h4lJbG5h3o4e4XK]Ki4b4UK_Kl4`4SKbKn4\\4PKfKQ5Z4mJhKS5W4lJkKU5S4jJnKW5R4hJoKY5o3gJRLY5m3hJRLY5n3fJSL[5k3fJTL[5k3eJVL\\5i3cJXL]5g3dJXL^5f3bJ[L^5d3aJ^L_5T82N1O2N1nCUJh;k5WDUJi;m5TDTJl;m5SDSJm;o5PDSJo;R6000001O00000010O001O10O01O00100O001O10O01O1O010O1O00100O001O10O01O1O010O1O00100O1O00100O001O10O01N2O1N101O1N101N2O0O2O1O2M2O2M3N2M2O2M3N2N1N3N2M3N1N3N2N2M2O2M3N2M2M4K5L4L3L8I8H7H8I8H7H8I8H7H8I8H7H8Iff="}, "label": "a blue bicycle"}]}
{"id": 498079, "image": "COCO_train2014_000000498079.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue bicycle that the woman is riding\"."}], "task": "refering", "answer_template": "The \"the blue bicycle that the woman is riding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 198, 201, 220, 221, 222, 268, 269, 290, 291, 292, 293, 307, 308, 309, 312, 313, 314, 315, 316, 317, 318, 326, 331, 333, 334, 335, 336, 338, 339, 340, 341, 342, 348, 349, 356, 357, 359, 361, 362, 363, 364, 365, 366, 371, 372, 377, 378, 379, 380, 381, 382, 384, 385, 386, 387, 388, 389, 394, 395, 401, 402, 403, 404, 407, 408, 409, 410, 411, 412, 417, 418, 423, 424, 425, 426, 431, 432, 433, 434, 435, 436, 445, 446, 447, 454, 455, 456, 457, 458, 469, 478, 479, 480, 481], "bbox": [0.147140625, 0.3611168384879725, 0.9604687499999999, 0.9858247422680413], "iscrowd": 0, "area": 38235.99669999998, "rle": {"size": [582, 640], "counts": "\\le13ka0;E:I8J5J7J5K6J5J6K5K5K5J6K5K5K5J6L43M3M2nMSAe0P?SOWAk0l>mNZAR1i>eN^AZ1e>^NbA`1d?M4B=B?AlSP23]loM0O100O100O010O100RCDP8=nGGo7:nGKP84oG0n71oG4o7LoG8n7IPH;n7EoG<R8FkG:W8GeG:\\8HaG8`8K\\G6c8MYG4h8OTG1l82PGOQ93lFMU91kF0U9OlF2T9LmF4S9KmF6T9GnF9R9FnF;S9CnF=R9APG?Q9_OoFc0Q9\\OoD]O;W1f:[OQE@5V1k:YOQEC1T1n:YOREDMT1R;WOREGIR1U;VOTEJDQ1X;TOUEM_Oo0^;SOTE0[Ok0c;TOTE3UOh0j;TORE6POd0Q<VOPE8kNa0X<VOnD;fN>^<VOnD<bN=c<VOlD=^N=i<UOjD>[N;n<WOhD>XN:S=VOgD`0TN9W=WOfD`0QN8\\=WOdDa0mM8a=WOcDa0jM7f=VObDc0iM2i=ZO_Dd0hMOk=]O^Dd0hMK_`05a_OHb`08__ODd`0<\\_OAg`0?Z_O\\Oj`0d0<00000lN[Og@f0X?\\Of@e0Y?]Oe@d0Y?@d@a0[?Ac@T1h>nNVAT1h>nNVAT1g>POVAR1h>POVAQ1i>QOUAQ1h>ROVAP1h>SOUAo0i>SOUAn0j>TOTAn0i>VOTAl0m>SOQAm0S?Q13M3M3M1OO1O100O1O1O100O1O100O1O1OfMVAZ1j>eNXAZ1g>gNYAY1f>gN\\AX1c>hN^AX1b>hN_AW1`>iNaAW1^>iNdAV1\\>jNeAU1Z>kNgAU1X>kNjAT1U>mNkAS1U>kNnAT1Q>lNPBT1o=lNSBS1l=mNVBR1i=mNYBS1f=mN\\BR1c=nN^BR1a=nNaBQ1]=oNfBo0Z=QOgBn0Z9^NZId0^Ml0V9bN\\Ib0_Mk0T9eN[Ia0cMh0Q9iN[I?dMJeN:[:_O[I=dMKjN5V:D\\I<dMLnNOQ:K\\I:dMMSOJm90ZI:fMLXOEi95XI:fMM^O_Oe99WI;fMNP:GYH;fMOQ:GWH;jMLP:IUH;mMKn9IUH;oMKm9JSH:SNJk9LQH:VNIh9NPH9ZNHg9NoG:[NGg9OmG9^NGf90kG8`NIe9OiG9cNGe9OhG9eNGc91gG8gNGb91fG7jNGa92dG6lNHa91bG8nNG`91aG7QOG^93`G6SOF^94^G6UOF]93]G7TOI`90[G7ROLd9MYG6QO0f9KXG5nN5j9EXG5lN9m9BUG6kN;Q:_OSG5iNa0T:ZOSG4hNc0U:ZORG2iNe0V:YOPG2iNg0W:WOoF1jNi0X:VOnF0iNk0Z:VOkFNkNm0[:UOiFNkNm0]:VOhFKkNn0`:WOdFKkNl0e:ZO^FImNl0h:[OZFImNk0l:\\OWFGmNk0o:_OSFFlNk0U;_OmEFmNi0Z;AhEFmNh0_;AdEFlNg0d;C_EGkNe0i;EZEFlNc0o;FUEFkNc0S<GQEGjN`0Z<HkDHjN?^<IgDHkN<c<LaDHjN;h<M]DHkN:j<NZDIkN9m<MXDIjN:P=NTDHlN:R=MQDJlN9T=MoCJlN:V=LnCJkN:Y=LjCJmN:Z=LhCJmN;]=JfCKlN:`=KcCKmN:b=J`CMlN:g=H[CNnN:j=FXCOnN;n=CSC3mN;S>@oB5nN;V>_OkB6mN;R`0En_O;R`0En_O;R`0Fl_O;U`0Dl_O;T`0Fk_O:U`0Fl_O9U`0Fk_O:U`0Gk_O7V`0Ij_O7V`0Ji_O6X`0Ii_O6W`0Ji_O6W`0Ki_O4X`0Kh_O5X`0Lh_O2Y`0Ng_O2Y`0Ng_O2Z`0Nf_O1Z`0Of_O1Z`0Og_O0[`0Od_O1]`0Nd_O0^`00a_O0a`0N__O2b`0M__O2b`0N]_O2e`0L\\_O3e`0MZ_O3]2KZ;0ZB3Z25l;KTB0k0=m<CVB5f0`0P=\\OXB8a0d0S=UOYB<=g0V=nN[B?8k0Y=gN\\Bc04n0\\=`N^Bf0OR1_=YN_Bj0LT1a=SNaBm0GX1d=lMbBd3Z=]LdBh3X=YLeBm3V=ULhBP4T=f0K4L3O0PMdJTI[5i6fJZIZ5b6gJaIX5\\6iJgIV5V6kJmIU5\\5gJkG5m2R5W5jJiG5S3Q5R5kJhG4Z3P5m4mJfG4_3P5j4lJdG5c3P5h4lJbG5h3o4e4XK]Ki4b4UK_Kl4`4SKbKn4\\4PKfKQ5Z4mJhKS5W4lJkKU5S4jJnKW5R4hJoKY5o3gJRLY5m3hJRLY5n3fJSL[5k3fJTL[5k3eJVL\\5i3cJXL]5g3dJXL^5f3bJ[L^5d3aJ^L_5T82N1O2N1nCUJh;k5WDUJi;m5TDTJl;m5SDSJm;o5PDSJo;R6000001O00000010O001O10O01O00100O001O10O01O1O010O1O00100O001O10O01O1O010O1O00100O1O00100O001O10O01N2O1N101O1N101N2O0O2O1O2M2O2M3N2M2O2M3N2N1N3N2M3N1N3N2N2M2O2M3N2M2M4K5L4L3L8I8H7H8I8H7H8I8H7H8I8H7H8Iff="}, "label": "the blue bicycle that the woman is riding"}]}
{"id": 203175, "image": "COCO_train2014_000000203175.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the backround that is looking away\"."}], "task": "refering", "answer_template": "The \"the horse in the backround that is looking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 238, 239, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 261, 262, 267, 268, 269, 270, 271, 272, 273, 274, 278, 279, 280, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.073640625, 0.0, 0.999375, 0.9865186915887851], "iscrowd": 0, "area": 106800.16375000004, "rle": {"size": [428, 640], "counts": "emc02W=6J5K6J5L5K4M4K5K4L5L3L5K4L5L4K3M4L3N3L3M4M2L5J5K6I5L5K5K5J5L5K5K5J5L5K5K5L4K4M4L4K5L3L5L4K4PNcInL`6P3jIdLZ6h0\\I:b0dNU6R1_I2e0aNo5^1cIHg0`Ni5g1hIAg0]Ne5R2jIYOk7g0YHTOg7m0[HoNg7Q1[HkNg7U1[HhNe7Y1]HcNe7]1]H`Nd7`1^H\\Nd7d1^HXNc7i1_HTNb7k1aHQNa7o1aHnM_7S2cHiM_7W2k1000O1000000O10O10O100O1N3L4M3M4L3L4M4L3M4K4M3N3M2N2N3M2N3M2O1N3M2N001O001O1O010O001O10O01O010O1O010O1O01N100O2O0O100O2N100O2O000100O1O100O1O010O100O100O3N?@10O01000O010O0100O010O10O010O0100O0100O0100O010O10O0100O101N1101O2M2O2M2O2N1N3N1N3N1N3N1O2M2O2M2O2M2O2N1N3N1N3N1O2M2nNWF^Nk9_1UFaNn9[1SFfNn9X1RFhNQ:Q1SFPOn9g0YFZOi9<^FDd93cFN_9IgF7[9@lFa0V9UOQGl0P9kNWGU1`:1O0YLoNSKR1i4SOUKm0g4YOWKh0d4^OZKc0a4C]K=`4H^K8^4N`K3[43cKMZ48dKHX4>fKCU4c0iK]OT4h0jKXOT4l0jKUOS4o0kKQOS4S1kKmNT4V1jKkNS4Y1kKgNS4]1kKcNS4a1kK`NR4d1lK\\NR4h1lKXNS4i1mKXNP4k1oKUNo3m1QLSNm3P2RLQNk3Q2ULoMj3S2ULmMi3V2VLkMg3W2YLiMe3Z2ZLfMd3\\2\\LeMa3^2^LbMa3_2_LaM_3b2`L_M]3d2bL\\M\\3f2dL[MY3h2fLXMX3j2hLVMW3l2hLUMU3m2kLSMS3P3lLPMR3R3mLPMP3S3oLmLo2V3PMjLo2W3QMjLl2Y3SMgLk2[3UMeLi2^3VMcLg2_3YMaLf2a3YM_Le2d3ZM\\Ld2f3\\MZLc2h3\\MXLb2j3^MVLa2l3^MTL`2n3`MRL^2Q4aMoK^2S4aMmK]2U4cMkK[2X4dMhK[2Y4eMgKY2[4gMeKX2]4gMcKW2_4iMaKU2a4kM_KT2b4lM^KR2d4nM\\KQ2e4oM[Ko1g4QNYKm1i4SNWKl1k4SNUKk1m4UNSKi1o4WNQKh1P5XNPKf1R5ZNnJe1S5[NmJc1U5]NkJa1X5^NhJa1Y5_NgJ_1[5aNeJ^1\\5bNeJ[1]5eNdJX1^5hNdJU1]5kNdJR1_5mNbJP1`5POaJn0`5ROaJk0a5UO`Ji0a5WOaJe0a5[O`Jb0b5^O_J`0c5_O^J>d5B]J;e5E]J8d5H]J5e5K\\J3e5M\\J0f50[JMh52YJLh54ZJHh58YJFh5:YJCh5>YJ_Og5c0YJ\\Og5e0ZJXOf5j0ZJTOg5m0ZJQOe5Q1[JmNf5T1[JjNd5X1\\JfNe5[1\\JbNd5`1\\J_Nd5b1]J[Nc5g1]JXNc5i1^JTNb5n1^JPNc5Q2j22O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10O0100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000001O001O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2M3N2N2N2N2N2N2N="}, "label": "the horse in the backround that is looking away"}]}
{"id": 367020, "image": "COCO_train2014_000000367020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sky blue toyoda\"."}], "task": "refering", "answer_template": "The \"a sky blue toyoda\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 42, 43, 44, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 177, 178], "bbox": [0.6028504672897196, 0.046171875, 1.0, 0.49887500000000007], "iscrowd": 0, "area": 30524.159700000004, "rle": {"size": [640, 428], "counts": "i_Q54cc0:F:G8G:G9F9B?SOm0I6L5L4L3L5L4K4M4K5L3M4K5L3O2O1N101N2O001N2O0O2O1O0O2O1O0O2O1N101O1N101N2O001N2O1N101O1N101O1N101N2O001N2O0O2O1O0O2O1N10001N100O2O000O2O000O2O0O101O0O101N10001N100O2O000O2O000O2O0O2O000O2O0O101O0O101N10001N100O2O0O1O2O0O1O2N100O2N1O10kLoC\\No;W1aDgN^;m0PEROo:c0_E\\O`:c0cE[O\\:f0eEVO]:j0fEQO\\:P1fEjN^:U1dEfN_:[1cE_N`:a1cEYN`:h1aESNc:l1`EnMc:S2_EhMc:X2_EcMd:^2^E\\Mf:c2]EWMf:j2[EQMh:o2[EkLh:V3ZEeLi:[3XE`Lk:`3XEZLk:g3g2O01O0010O01O0010O01O00100O001O010O001O010O^Me1"}, "label": "a sky blue toyoda"}]}
{"id": 367020, "image": "COCO_train2014_000000367020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light blue camry sitting along the road\"."}], "task": "refering", "answer_template": "The \"light blue camry sitting along the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 42, 43, 44, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 177, 178], "bbox": [0.6028504672897196, 0.046171875, 1.0, 0.49887500000000007], "iscrowd": 0, "area": 30524.159700000004, "rle": {"size": [640, 428], "counts": "i_Q54cc0:F:G8G:G9F9B?SOm0I6L5L4L3L5L4K4M4K5L3M4K5L3O2O1N101N2O001N2O0O2O1O0O2O1O0O2O1N101O1N101N2O001N2O1N101O1N101O1N101N2O001N2O0O2O1O0O2O1N10001N100O2O000O2O000O2O0O101O0O101N10001N100O2O000O2O000O2O0O2O000O2O0O101O0O101N10001N100O2O0O1O2O0O1O2N100O2N1O10kLoC\\No;W1aDgN^;m0PEROo:c0_E\\O`:c0cE[O\\:f0eEVO]:j0fEQO\\:P1fEjN^:U1dEfN_:[1cE_N`:a1cEYN`:h1aESNc:l1`EnMc:S2_EhMc:X2_EcMd:^2^E\\Mf:c2]EWMf:j2[EQMh:o2[EkLh:V3ZEeLi:[3XE`Lk:`3XEZLk:g3g2O01O0010O01O0010O01O00100O001O010O001O010O^Me1"}, "label": "light blue camry sitting along the road"}]}
{"id": 367020, "image": "COCO_train2014_000000367020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car standing at the side parking bay\"."}], "task": "refering", "answer_template": "The \"a white car standing at the side parking bay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 24, 25, 26, 27, 28, 29, 39, 40, 41], "bbox": [0.5752803738317757, 0.0, 0.9994626168224299, 0.119125], "iscrowd": 0, "area": 8563.12895, "rle": {"size": [640, 428], "counts": "_\\j42nc01O1O1O1O1N2O001O1O1O1N2O1O2N2N2N2M3N2N3M2N2N2M3N3M2N2N1O001N2O1O1O1O2N2M3N2N2N1O000O10000000000O01000000000O10000000000O1000O100O1N101O1N01O11N2O1N2O1O1N2O1O1O1O1000000000000O100000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O1000000000000000000O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O10f0YO[O"}, "label": "a white car standing at the side parking bay"}]}
{"id": 367020, "image": "COCO_train2014_000000367020.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck on a street behind a green convertable\"."}], "task": "refering", "answer_template": "The \"a white truck on a street behind a green convertable\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 24, 25, 26, 27, 28, 29, 39, 40, 41], "bbox": [0.5752803738317757, 0.0, 0.9994626168224299, 0.119125], "iscrowd": 0, "area": 8563.12895, "rle": {"size": [640, 428], "counts": "_\\j42nc01O1O1O1O1N2O001O1O1O1N2O1O2N2N2N2M3N2N3M2N2N2M3N3M2N2N1O001N2O1O1O1O2N2M3N2N2N1O000O10000000000O01000000000O10000000000O1000O100O1N101O1N01O11N2O1N2O1O1N2O1O1O1O1000000000000O100000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O1000000000000000000O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O10f0YO[O"}, "label": "a white truck on a street behind a green convertable"}]}
{"id": 72111, "image": "COCO_train2014_000000072111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark gray couch\"."}], "task": "refering", "answer_template": "The \"a dark gray couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.566921875, 0.31283333333333335, 1.0, 0.9895], "iscrowd": 0, "area": 68777.30374999999, "rle": {"size": [480, 640], "counts": "k_Z54g>6K5J6J6J6K5J6J6J6J6K5J6J6J6J6J6J6J6J6I7[Oe0O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2J6O00100O1O1O1N2O1O1O1O1N101O1O1O1N2O1O1O1O1O1N101O1O1O1N2O1O1O1O1N2O001O1O1N2O1O1O1O1N2O1O001O1O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O0O2O1O1O1O1O1N2O00000001O00L4L4L4K5L4L4O1N2O1O1N2O1O1N2O1O1N2O1O2M2O1O100O1O100O100O100O100O1001O00000000001O000000001O00000000001O0001O0001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O000010O01O00001O001O00001O0000000000001O00001O00001O00001O00001O00001O000010O0001O001O00001O00001O00001O00001_N`1oJXO"}, "label": "a dark gray couch"}]}
{"id": 72111, "image": "COCO_train2014_000000072111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black sofa has a black pillow and a green pillow on top of it along with some bags of goodies\"."}], "task": "refering", "answer_template": "The \"the black sofa has a black pillow and a green pillow on top of it along with some bags of goodies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.566921875, 0.31283333333333335, 1.0, 0.9895], "iscrowd": 0, "area": 68777.30374999999, "rle": {"size": [480, 640], "counts": "k_Z54g>6K5J6J6J6K5J6J6J6J6K5J6J6J6J6J6J6J6J6I7[Oe0O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2J6O00100O1O1O1N2O1O1O1O1N101O1O1O1N2O1O1O1O1O1N101O1O1O1N2O1O1O1O1N2O001O1O1N2O1O1O1O1N2O1O001O1O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O0O2O1O1O1O1O1N2O00000001O00L4L4L4K5L4L4O1N2O1O1N2O1O1N2O1O1N2O1O2M2O1O100O1O100O100O100O100O1001O00000000001O000000001O00000000001O0001O0001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O000010O01O00001O001O00001O0000000000001O00001O00001O00001O00001O00001O000010O0001O001O00001O00001O00001O00001_N`1oJXO"}, "label": "the black sofa has a black pillow and a green pillow on top of it along with some bags of goodies"}]}
{"id": 72111, "image": "COCO_train2014_000000072111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower patterned couch with pillows\"."}], "task": "refering", "answer_template": "The \"a flower patterned couch with pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 261, 262, 263, 264, 265, 266, 277, 278, 284, 285, 286, 287, 288], "bbox": [0.0, 0.21720833333333334, 0.664515625, 0.7677499999999999], "iscrowd": 0, "area": 78586.73364999998, "rle": {"size": [480, 640], "counts": "l8l0^<f1YNg1jNV1000001N100000000O1000001O000O100000000O10001O000O10000000000O2O001O2N3L4M3M3M3M3L3N3M3M3M00O010000000O100000000O1000O1000O10000000000O100000000O10000000O0100000000O100000000O100000000O1N2cNfGSL[8k3lGnKU8R4RHgKo7W4XHbKj7\\4^H]Kc7a4dHXK^7g4^10000O10000000000000000O10O100000000000O10000000000000000O01000000000000000O100000O1000000000O100000000000O1000O1000000000000000O010000000000000000O1000O1000000000O1000000000O100000O11O1O1O1O2N1O1O1N2O2N1O1O1O1O2N1O1O1O2M2O2N3M3M2N3M3M3M2N3M3M3L3N3M3M3M2N3M3L4M2N3M3M3M3M4K5L0000000000O0100000O100000O10O1000000O10O100000O100000O0100000000O10O100000O100000O0100000000O10O1000O10000000O01000000O1000O1000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O10000O101O001N2O1O1O1N101O1N2O1O1N1L5K5K5K5K5K9G:F:F:F9F;F:F6J5K4L5K4dN^DVOg;i0`DlNd;S1cDcNb;\\1eDYN_;g1n0O0O2O001O0O2O001N101O001N101O000O2O001N101O001N101O0O2O001M2N3M4K4M3L5L3L5L3MRcT3"}, "label": "a flower patterned couch with pillows"}]}
{"id": 72111, "image": "COCO_train2014_000000072111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and brown sofa in a dimly lit room\"."}], "task": "refering", "answer_template": "The \"a white and brown sofa in a dimly lit room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 261, 262, 263, 264, 265, 266, 277, 278, 284, 285, 286, 287, 288], "bbox": [0.0, 0.21720833333333334, 0.664515625, 0.7677499999999999], "iscrowd": 0, "area": 78586.73364999998, "rle": {"size": [480, 640], "counts": "l8l0^<f1YNg1jNV1000001N100000000O1000001O000O100000000O10001O000O10000000000O2O001O2N3L4M3M3M3M3L3N3M3M3M00O010000000O100000000O1000O1000O10000000000O100000000O10000000O0100000000O100000000O100000000O1N2cNfGSL[8k3lGnKU8R4RHgKo7W4XHbKj7\\4^H]Kc7a4dHXK^7g4^10000O10000000000000000O10O100000000000O10000000000000000O01000000000000000O100000O1000000000O100000000000O1000O1000000000000000O010000000000000000O1000O1000000000O1000000000O100000O11O1O1O1O2N1O1O1N2O2N1O1O1O1O2N1O1O1O2M2O2N3M3M2N3M3M3M2N3M3M3L3N3M3M3M2N3M3L4M2N3M3M3M3M4K5L0000000000O0100000O100000O10O1000000O10O100000O100000O0100000000O10O100000O100000O0100000000O10O1000O10000000O01000000O1000O1000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O10000O101O001N2O1O1O1N101O1N2O1O1N1L5K5K5K5K5K9G:F:F:F9F;F:F6J5K4L5K4dN^DVOg;i0`DlNd;S1cDcNb;\\1eDYN_;g1n0O0O2O001O0O2O001N101O001N101O000O2O001N101O001N101O0O2O001M2N3M4K4M3L5L3L5L3MRcT3"}, "label": "a white and brown sofa in a dimly lit room"}]}
{"id": 391600, "image": "COCO_train2014_000000391600.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat to the left of a dark building\"."}], "task": "refering", "answer_template": "The \"a boat to the left of a dark building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 187, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257], "bbox": [0.0, 0.405, 0.376234375, 0.66], "iscrowd": 0, "area": 9601.462450000003, "rle": {"size": [480, 640], "counts": "a94k>2O001O1O001O1O001O001O000DFRB:l=HTB8i=KWB5[=JcB125[=M`BN55[=0]BK85[=9eBG[=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=;cBE]=;cBE]=;cBE]=;\\BLd=h000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000dMhNPGX1o8oNkFQ1T9VOfFj0Y9]OaFc0`9A[F?j9BPF>U:CeE=`:DZE<l:DnD<W;EcD;b;FXD:m;GmC9X<HbC8d<o05K5K5K5K5L3L00001O00000000000000001O0000000O100000001O000O10000000000O2O000000000O1000001O00000O10M3YOPO[CS1d<d01O100O1O1M2K6001O001O001O1O4L1OO1O100O100M3M31O4L00001O0000001O0000001O00000000000000O101O000O1M3M3N2M32N2N7I3M0001bNSCg0d=O1O1O1O1O1O001O0O10O100004L001O1O000H8L4O1O10O10000001O00000000001N1000000O1NTWk5"}, "label": "a boat to the left of a dark building"}]}
{"id": 391600, "image": "COCO_train2014_000000391600.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and red building sitting on the left side of marina\"."}], "task": "refering", "answer_template": "The \"white and red building sitting on the left side of marina\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 187, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257], "bbox": [0.0, 0.405, 0.376234375, 0.66], "iscrowd": 0, "area": 9601.462450000003, "rle": {"size": [480, 640], "counts": "a94k>2O001O1O001O1O001O001O000DFRB:l=HTB8i=KWB5[=JcB125[=M`BN55[=0]BK85[=9eBG[=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=:dBF\\=;cBE]=;cBE]=;cBE]=;\\BLd=h000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000dMhNPGX1o8oNkFQ1T9VOfFj0Y9]OaFc0`9A[F?j9BPF>U:CeE=`:DZE<l:DnD<W;EcD;b;FXD:m;GmC9X<HbC8d<o05K5K5K5K5L3L00001O00000000000000001O0000000O100000001O000O10000000000O2O000000000O1000001O00000O10M3YOPO[CS1d<d01O100O1O1M2K6001O001O001O1O4L1OO1O100O100M3M31O4L00001O0000001O0000001O00000000000000O101O000O1M3M3N2M32N2N7I3M0001bNSCg0d=O1O1O1O1O1O001O0O10O100004L001O1O000H8L4O1O10O10000001O00000000001N1000000O1NTWk5"}, "label": "white and red building sitting on the left side of marina"}]}
{"id": 55733, "image": "COCO_train2014_000000055733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green ski board that the woman is holding\"."}], "task": "refering", "answer_template": "The \"the green ski board that the woman is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 211, 212, 213, 235, 236, 237, 259, 260, 282, 283, 306, 329, 352, 375, 376], "bbox": [0.10925, 0.4812708333333333, 0.393078125, 1.0], "iscrowd": 0, "area": 6637.8196000000025, "rle": {"size": [480, 640], "counts": "dQQ11o>0O1O100O1O2O0O100000000O101O000O1000000O10001O0O100000000O10001O0O100000000000001O0000000O101O1O001O1O001O1O001O1O001O0O10000000000000001O0010O0001O001O0010O01O0YBVOX=j0gB[OU=f0iB[OW=e0gB^OY=b0eB@Z=T1O001O1O10O1000O1000000O2O00001O0O101O000O2O001O001O1O2O0O2N2O0O2O0O3M2O2M2N30O01O001O1O100O1O1O11O6J5K5M4M1N01O01N1N3M2N2N2N2N3N1^NoBm0c=00N2O1M3M3M3M3M2N001O001O001O001O001O00001O001O001O0O2O2NTle5"}, "label": "the green ski board that the woman is holding"}]}
{"id": 55733, "image": "COCO_train2014_000000055733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue snowboard\"."}], "task": "refering", "answer_template": "The \"a blue snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 211, 212, 213, 235, 236, 237, 259, 260, 282, 283, 306, 329, 352, 375, 376], "bbox": [0.10925, 0.4812708333333333, 0.393078125, 1.0], "iscrowd": 0, "area": 6637.8196000000025, "rle": {"size": [480, 640], "counts": "dQQ11o>0O1O100O1O2O0O100000000O101O000O1000000O10001O0O100000000O10001O0O100000000000001O0000000O101O1O001O1O001O1O001O1O001O0O10000000000000001O0010O0001O001O0010O01O0YBVOX=j0gB[OU=f0iB[OW=e0gB^OY=b0eB@Z=T1O001O1O10O1000O1000000O2O00001O0O101O000O2O001O001O1O2O0O2N2O0O2O0O3M2O2M2N30O01O001O1O100O1O1O11O6J5K5M4M1N01O01N1N3M2N2N2N2N3N1^NoBm0c=00N2O1M3M3M3M3M2N001O001O001O001O001O00001O001O001O0O2O2NTle5"}, "label": "a blue snowboard"}]}
{"id": 55733, "image": "COCO_train2014_000000055733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue jacket and orange glasses next to a woman with a white hood\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue jacket and orange glasses next to a woman with a white hood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 355, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.10785937500000001, 0.20014583333333333, 0.48034374999999996, 0.9866666666666666], "iscrowd": 0, "area": 62842.554100000016, "rle": {"size": [480, 640], "counts": "QiP1i0e<b1^Nb1^Na1dN]1L4L4L4L4K5L4L4L4L4M3N2^MnGUNS8k1SHmMo7R2YHfMh7Y2`H^Mb7a2fHWM[7i2kHoLW7P3nHjLT7U3PIgLQ7X3SIcLo6]3TI_Lm6`3WI[Lk6d3ZIWLg6i3\\IRLf6m3W2O1O100O1O001O100O1O1O1O100O1O1O100O1O1O1O100O1O10000O10000O100O10000O1000O010000O1000000000000000000000000000000000000000000000000001O1N101O1O002N1O1O2N1O1O2N1O2N1O1O2N1O1O2N2N3M2N3M2N3M2N3M2N3M2N3M3lFoJR8T5VGeKg8]5M3M3M4L3M3M3M3M3M1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N2N3M2N2N3M2N3M2N2N3M3M2N3M2N3M2N3M3M2N3M2M3N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N3M3L4M3M3M2N3M3MS2mM8H3M3M2N3M3M2Nadk4"}, "label": "a woman wearing a blue jacket and orange glasses next to a woman with a white hood"}]}
{"id": 55733, "image": "COCO_train2014_000000055733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with orange ski goggles on her head and a black hat\"."}], "task": "refering", "answer_template": "The \"a woman with orange ski goggles on her head and a black hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 355, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.10785937500000001, 0.20014583333333333, 0.48034374999999996, 0.9866666666666666], "iscrowd": 0, "area": 62842.554100000016, "rle": {"size": [480, 640], "counts": "QiP1i0e<b1^Nb1^Na1dN]1L4L4L4L4K5L4L4L4L4M3N2^MnGUNS8k1SHmMo7R2YHfMh7Y2`H^Mb7a2fHWM[7i2kHoLW7P3nHjLT7U3PIgLQ7X3SIcLo6]3TI_Lm6`3WI[Lk6d3ZIWLg6i3\\IRLf6m3W2O1O100O1O001O100O1O1O1O100O1O1O100O1O1O1O100O1O10000O10000O100O10000O1000O010000O1000000000000000000000000000000000000000000000000001O1N101O1O002N1O1O2N1O1O2N1O2N1O1O2N1O1O2N2N3M2N3M2N3M2N3M2N3M2N3M3lFoJR8T5VGeKg8]5M3M3M4L3M3M3M3M3M1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O2N1O2N1N3N1O2N1O2N1O1O2N1O2N2N3M2N2N3M2N3M2N2N3M3M2N3M2N3M2N3M3M2N3M2M3N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N3M3L4M3M3M2N3M3MS2mM8H3M3M2N3M3M2Nadk4"}, "label": "a woman with orange ski goggles on her head and a black hat"}]}
{"id": 55733, "image": "COCO_train2014_000000055733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman snow boarder coming inside from the snow\"."}], "task": "refering", "answer_template": "The \"a woman snow boarder coming inside from the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 334, 335, 336, 357, 358, 359, 380, 381, 382, 383], "bbox": [0.538265625, 0.1616875, 0.8655781250000001, 1.0], "iscrowd": 0, "area": 49247.87705, "rle": {"size": [480, 640], "counts": "^oQ58d>7I8I6I6J7oGiNo1]1fMlNW2Z1\\MPO`2W1TMROj2U1gLUOW3R1WL[Of3m0hK@V4g0XKFf4b0QJa0m5Z4M3N2N2N2M3N2N2M3N2O1O1O1O1O1O1O1O1O1O1O10000O1000000O1000000O10000O1000000N2N2N2N2N2N2N2M3N2M3M3iNcFhL`9V3oF[LS9c3jMZL^Kd3c4gLSKV3o4VMfJh2[5cM[J[2g5oMoIo1T6ZNbId1b6dNTIZ1P7nNfHP1]7ZOXHe0k7CkG;W8O_G0c8T32N3N1N3M2O2M2O2M2O2M2O2N101O001O0O2O001O0O2O001O001N101O001N010000O1000000O1000000O2O000O10000O1N2O2N1N2O1O1N2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O2M2O1N3M2O1N3M2N3N1N2RKmHm1V7PNlHo1U7nMnHn1V7PNlHi1\\7SNgHe1a7YNaH`1f7]N]H^1i7_NZH^1i7^NZHa1h7\\NZHc1h7YN[He1h7XNZHg1i7UNYHj1l7PNVHo1P8kMQHU1oN\\NX97kG[1PO\\N[92gGa1QOZN_9MbGg1SOXNb9K]Gk1TOWNf9FYGQ2TOVNY;g1kDVNW;h1kDUNY;g1kDVNW;h1R1O21N1O2O1N1O2N2O0ON2N3N1N2N3M2O1N3M2N2N3N1J6I>BTSX1"}, "label": "a woman snow boarder coming inside from the snow"}]}
{"id": 55733, "image": "COCO_train2014_000000055733.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rightmost woman\"."}], "task": "refering", "answer_template": "The \"rightmost woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 334, 335, 336, 357, 358, 359, 380, 381, 382, 383], "bbox": [0.538265625, 0.1616875, 0.8655781250000001, 1.0], "iscrowd": 0, "area": 49247.87705, "rle": {"size": [480, 640], "counts": "^oQ58d>7I8I6I6J7oGiNo1]1fMlNW2Z1\\MPO`2W1TMROj2U1gLUOW3R1WL[Of3m0hK@V4g0XKFf4b0QJa0m5Z4M3N2N2N2M3N2N2M3N2O1O1O1O1O1O1O1O1O1O1O10000O1000000O1000000O10000O1000000N2N2N2N2N2N2N2M3N2M3M3iNcFhL`9V3oF[LS9c3jMZL^Kd3c4gLSKV3o4VMfJh2[5cM[J[2g5oMoIo1T6ZNbId1b6dNTIZ1P7nNfHP1]7ZOXHe0k7CkG;W8O_G0c8T32N3N1N3M2O2M2O2M2O2M2O2N101O001O0O2O001O0O2O001O001N101O001N010000O1000000O1000000O2O000O10000O1N2O2N1N2O1O1N2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O2M2O1N3M2O1N3M2N3N1N2RKmHm1V7PNlHo1U7nMnHn1V7PNlHi1\\7SNgHe1a7YNaH`1f7]N]H^1i7_NZH^1i7^NZHa1h7\\NZHc1h7YN[He1h7XNZHg1i7UNYHj1l7PNVHo1P8kMQHU1oN\\NX97kG[1PO\\N[92gGa1QOZN_9MbGg1SOXNb9K]Gk1TOWNf9FYGQ2TOVNY;g1kDVNW;h1kDUNY;g1kDVNW;h1R1O21N1O2O1N1O2N2O0ON2N3N1N2N3M2O1N3M2N2N3N1J6I>BTSX1"}, "label": "rightmost woman"}]}
{"id": 268726, "image": "COCO_train2014_000000268726.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a sofa with some blanket tossed upon it\"."}], "task": "refering", "answer_template": "The \"the back of a sofa with some blanket tossed upon it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 66, 67, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193], "bbox": [0.048859375, 0.08809145129224652, 0.952890625, 0.4974155069582505], "iscrowd": 0, "area": 78467.57010000001, "rle": {"size": [503, 640], "counts": "i[?2d?8I7H7I8H7I8H7I8I6I8H7I5K1O1O100O001O1O1O1O10O01O1O1O1O1O010O1O1O1O1O1O100O1O1O1O1O1O100O1O100O100O1O100O100O100O2O0O1O100O100O100O100O1000000000000O100000000000000000000000000000000000001O0000001O0000001O000000001O0000001O0000001O0000001O00001O00001O00001O001N10001O00001O00001O001O00001O00001O001O00001O00001O00001O001O0O101O00001O00001O001O00001O00001O001O000O2O0000O100O100O100O100O1O100O100O100O10O0100O100O10000O100O100O100O100O10000O10O0100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10O10O10000O100O10000O100O10000000000000000001O0000000000000000001O000000000O10000O100O2O0O100O100O100O10001N100O100O100O101N100O10000O100O2O0O100O100O10001N100O100O100O101N100O10000O101N100O100O100O101N1000000O101O1O001N2O1O001N2O001O1O0O2O1O1O000O101O000O100000001N100000000O101O00000O1000001N100000000O2O0000000O11O01O0000000001O0000000O10001O00000O1000001O0O100000000O2O0000000O10001O000O100000001N100000000O101O00000O100O1O2N100O1O1O100O2N1O100O1O1O101N1O1O100O1O1O2O0O10001O00001O0010O01O001O00001O001O010O00001O001O001O0010O0001O001O001O000010OF;D;F;D;F;D;G9I8H7I8H7I8H7I[e>"}, "label": "the back of a sofa with some blanket tossed upon it"}]}
{"id": 268726, "image": "COCO_train2014_000000268726.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flowered quilt on back of couch\"."}], "task": "refering", "answer_template": "The \"flowered quilt on back of couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 66, 67, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193], "bbox": [0.048859375, 0.08809145129224652, 0.952890625, 0.4974155069582505], "iscrowd": 0, "area": 78467.57010000001, "rle": {"size": [503, 640], "counts": "i[?2d?8I7H7I8H7I8H7I8I6I8H7I5K1O1O100O001O1O1O1O10O01O1O1O1O1O010O1O1O1O1O1O100O1O1O1O1O1O100O1O100O100O1O100O100O100O2O0O1O100O100O100O100O1000000000000O100000000000000000000000000000000000001O0000001O0000001O000000001O0000001O0000001O0000001O00001O00001O00001O001N10001O00001O00001O001O00001O00001O001O00001O00001O00001O001O0O101O00001O00001O001O00001O00001O001O000O2O0000O100O100O100O100O1O100O100O100O10O0100O100O10000O100O100O100O100O10000O10O0100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10O10O10000O100O10000O100O10000000000000000001O0000000000000000001O000000000O10000O100O2O0O100O100O100O10001N100O100O100O101N100O10000O100O2O0O100O100O10001N100O100O100O101N100O10000O101N100O100O100O101N1000000O101O1O001N2O1O001N2O001O1O0O2O1O1O000O101O000O100000001N100000000O101O00000O1000001N100000000O2O0000000O11O01O0000000001O0000000O10001O00000O1000001O0O100000000O2O0000000O10001O000O100000001N100000000O101O00000O100O1O2N100O1O1O100O2N1O100O1O1O101N1O1O100O1O1O2O0O10001O00001O0010O01O001O00001O001O010O00001O001O001O0010O0001O001O001O000010OF;D;F;D;F;D;G9I8H7I8H7I8H7I[e>"}, "label": "flowered quilt on back of couch"}]}
{"id": 424376, "image": "COCO_train2014_000000424376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three zebras standing on a rock\"."}], "task": "refering", "answer_template": "The \"three zebras standing on a rock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 53, 54, 55, 56, 57, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 86, 87, 88, 89, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 136, 137, 138, 139, 140, 141, 142, 148, 151, 152, 153, 154, 161, 162, 165, 166, 167, 174, 175, 179, 180, 181, 187, 188, 189, 192, 193, 194, 205, 206, 218, 219], "bbox": [0.026959999999999998, 0.21086000000000002, 0.9887733333333334, 0.9434400000000001], "iscrowd": 0, "area": 49501.63305, "rle": {"size": [500, 375], "counts": "fS53Y?9G:F:K5K5L3L5K5K5K4L3K5J6J6K5J6L4M3N2M3N2M3WOdLXE^3f:fLUE]3j:gLPE\\3n:iLmDY3R;e00000O10M3N1O2M3N1O2N2N1O2N2N1O2O1N1O2N2N1O2O1N101O1O0O2O1O002M2O1O1O1N2O1O1O2M2O1O1N2001O0010O01O1O1O1O001O100O1O1O001O1O1O100O001O1O1O1O010O1O1O1O1O2O0O100O1O2O0O100jD`LV:a3[EmLe:Q41N101O0O101O0O2O000O2O000O2O0O2O00010O0100O0018H;EO001fF\\Lb6e3oHkLo6V3nHnLQ7Q3oHQMo6P3PIQMo6P3oHSMo6n2PITMXOXOY6d3_JVMnN@a6[3`JVMfNHj6R3^JYM^NOR7i2_JZMUN6[7`2`JhN^5Y1aJhN^5Y1`JjN^5W1aJjN^5V1eJiNZ5W1kJcNU5^1PK[NQ5e1VKSNk4n1ZK^LeMh0S7k2]KWLdMi0P7Q3[KVLjMc0m6V3YKVLoM>j6]3VKULTN9g6c3TKTLZN3c6j3QKSLaNM`6P4oJSLW6n3hIRLX6o3fIQL\\6n3dIRL\\6n3dIRL\\6n3cIRL^6n3bIRL^6n3bIRLnN4V6^3kJ^LoN4S7b2nIZMmN5W7`2lI[MlN5Z7_2jI[MkN7\\7^2hI\\MkN6_7]2fI]MiN8b7Z2eI]MiN9d7Y2bI_MhN9g7X2aIfM`N2Q8W2^IUNRNEb8U2\\IQOf6o0XIQOj6o0UIQOk6o0TIQOn6o0QIPOQ7P1mHPOU7P1iHPOY7P1fHoN[7Q1iHjNY7V1VIZNk6f1dIiM^6W2_3010O1N2O0O2O1N2N101N2N2O1N1O101N100O1O2O0O1O101N100O1O2N1O1O1O2N1O1O1O2N1000001N1000001O0000001O00001N2O1O1O1O1O1O001O1O1O1N2O1O1O001O1O1O1O1O1O1O1N101O1O1O1O2N2N2N2N3M2M3N2N2N2N2N2N2N2\\IWJY3k5bLXJ^3j5\\LZJd3h5WL[Jh3h5TLZJl3h5QLYJo3i5mKYJS4i5iKYJX4h5eKYJ[4i5aKYJ_4i5]KYJd4h5XKZJh4h5UKZJk4h5PKZJP5h5lJZJU5g5gJ[JY5g5dJZJ\\5h5`JYJb5f5\\JZJf5g5WJZJk5e5TJ[Jm5f5PJ[JQ6e5mI\\JU6c5iI^JX6b5gI^J[6a5cI`J^6`5`IaJb6^5]IbJd6^5ZIcJg6]5WIdJk6[5SIeJo6[5PIdJS7[5kHeJW7\\5fHdJ]7[5cHcJ_7]5bH`J`7`5`H^Jc7a5]H]Je7c5[H[Jh7d3ZHVMNTOj7b3^HXMHUOk7_3cHZMBUOn7\\3fH]M\\OUOP8Z3kH^MUOVOS8W3nHaMoNVOU8W3PIaMkNVOX8V3RIaMfNWOZ8W3SI`McNWO\\8W3UI`M_NWO_8V3VIaM[NWOa8V3YI`MVNXOd8U3ZIaMRNXOf8V3[I`MoMXOi8U3]I`MjMYOk8U3_I`MgMXOl8V3aI`Me6]2_IaMa6^2cI_M^6^2fI`MZ6^2jI`MV6^2oI_MR6^2RJ`Mn5_2UJ_Ml5^2ZJ^Mh5^2]J_Mf5\\2m3K4L5L4K5K5K4L5L^g1"}, "label": "three zebras standing on a rock"}]}
{"id": 424376, "image": "COCO_train2014_000000424376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing between two fighting zebras\"."}], "task": "refering", "answer_template": "The \"a zebra standing between two fighting zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 53, 54, 55, 56, 57, 66, 67, 68, 69, 70, 71, 78, 79, 80, 81, 82, 83, 84, 86, 87, 88, 89, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 136, 137, 138, 139, 140, 141, 142, 148, 151, 152, 153, 154, 161, 162, 165, 166, 167, 174, 175, 179, 180, 181, 187, 188, 189, 192, 193, 194, 205, 206, 218, 219], "bbox": [0.026959999999999998, 0.21086000000000002, 0.9887733333333334, 0.9434400000000001], "iscrowd": 0, "area": 49501.63305, "rle": {"size": [500, 375], "counts": "fS53Y?9G:F:K5K5L3L5K5K5K4L3K5J6J6K5J6L4M3N2M3N2M3WOdLXE^3f:fLUE]3j:gLPE\\3n:iLmDY3R;e00000O10M3N1O2M3N1O2N2N1O2N2N1O2O1N1O2N2N1O2O1N101O1O0O2O1O002M2O1O1O1N2O1O1O2M2O1O1N2001O0010O01O1O1O1O001O100O1O1O001O1O1O100O001O1O1O1O010O1O1O1O1O2O0O100O1O2O0O100jD`LV:a3[EmLe:Q41N101O0O101O0O2O000O2O000O2O0O2O00010O0100O0018H;EO001fF\\Lb6e3oHkLo6V3nHnLQ7Q3oHQMo6P3PIQMo6P3oHSMo6n2PITMXOXOY6d3_JVMnN@a6[3`JVMfNHj6R3^JYM^NOR7i2_JZMUN6[7`2`JhN^5Y1aJhN^5Y1`JjN^5W1aJjN^5V1eJiNZ5W1kJcNU5^1PK[NQ5e1VKSNk4n1ZK^LeMh0S7k2]KWLdMi0P7Q3[KVLjMc0m6V3YKVLoM>j6]3VKULTN9g6c3TKTLZN3c6j3QKSLaNM`6P4oJSLW6n3hIRLX6o3fIQL\\6n3dIRL\\6n3dIRL\\6n3cIRL^6n3bIRL^6n3bIRLnN4V6^3kJ^LoN4S7b2nIZMmN5W7`2lI[MlN5Z7_2jI[MkN7\\7^2hI\\MkN6_7]2fI]MiN8b7Z2eI]MiN9d7Y2bI_MhN9g7X2aIfM`N2Q8W2^IUNRNEb8U2\\IQOf6o0XIQOj6o0UIQOk6o0TIQOn6o0QIPOQ7P1mHPOU7P1iHPOY7P1fHoN[7Q1iHjNY7V1VIZNk6f1dIiM^6W2_3010O1N2O0O2O1N2N101N2N2O1N1O101N100O1O2O0O1O101N100O1O2N1O1O1O2N1O1O1O2N1000001N1000001O0000001O00001N2O1O1O1O1O1O001O1O1O1N2O1O1O001O1O1O1O1O1O1O1N101O1O1O1O2N2N2N2N3M2M3N2N2N2N2N2N2N2\\IWJY3k5bLXJ^3j5\\LZJd3h5WL[Jh3h5TLZJl3h5QLYJo3i5mKYJS4i5iKYJX4h5eKYJ[4i5aKYJ_4i5]KYJd4h5XKZJh4h5UKZJk4h5PKZJP5h5lJZJU5g5gJ[JY5g5dJZJ\\5h5`JYJb5f5\\JZJf5g5WJZJk5e5TJ[Jm5f5PJ[JQ6e5mI\\JU6c5iI^JX6b5gI^J[6a5cI`J^6`5`IaJb6^5]IbJd6^5ZIcJg6]5WIdJk6[5SIeJo6[5PIdJS7[5kHeJW7\\5fHdJ]7[5cHcJ_7]5bH`J`7`5`H^Jc7a5]H]Je7c5[H[Jh7d3ZHVMNTOj7b3^HXMHUOk7_3cHZMBUOn7\\3fH]M\\OUOP8Z3kH^MUOVOS8W3nHaMoNVOU8W3PIaMkNVOX8V3RIaMfNWOZ8W3SI`McNWO\\8W3UI`M_NWO_8V3VIaM[NWOa8V3YI`MVNXOd8U3ZIaMRNXOf8V3[I`MoMXOi8U3]I`MjMYOk8U3_I`MgMXOl8V3aI`Me6]2_IaMa6^2cI_M^6^2fI`MZ6^2jI`MV6^2oI_MR6^2RJ`Mn5_2UJ_Ml5^2ZJ^Mh5^2]J_Mf5\\2m3K4L5L4K5K5K4L5L^g1"}, "label": "a zebra standing between two fighting zebras"}]}
{"id": 424376, "image": "COCO_train2014_000000424376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra cuddling another zebra while a baby zebra feeds\"."}], "task": "refering", "answer_template": "The \"zebra cuddling another zebra while a baby zebra feeds\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 32, 33, 34, 35, 36, 45, 46, 47, 48, 49, 50, 58, 59, 60, 61, 62, 63, 64, 71, 72, 73, 74, 75, 76, 77, 84, 85, 86, 87, 88, 89, 90, 97, 98, 103], "bbox": [0.47341333333333335, 0.05998, 1.0, 0.4173], "iscrowd": 0, "area": 20505.529599999998, "rle": {"size": [500, 375], "counts": "gnf21c?1O1O2N1O1O1bDOW73THa0l7_OmGh0R8YOhGm0W8TObGS1]8oNXG[1g8fNmFf1R9[NfFm1Y9TNaFR2^9PN[FV2d9kMWFZ2h9gMRF_2m9gMiE^2W:gMYEd2f:V1O1O1O1O1O1O0000O100O1O100O1O100O1O010O1O100O1O100O100O1O1O1O00M3M2N3M3002M2O2N1O2M2O2N1O2M2O2N1O2N2M2O2N1010O100O010O10O0100O10O010O001O001O001O0O2O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O00100O2O0O1O2O0O101N1O101N101N1O101N100O1O10O0100O0010O0100O00100O010O1O010O10O01O10O0100O00100O010O1O010O1O00100O001O1O010O1O001O1M2N3M3M2N3M3MgK"}, "label": "zebra cuddling another zebra while a baby zebra feeds"}]}
{"id": 424376, "image": "COCO_train2014_000000424376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with its head on the back of another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra with its head on the back of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 32, 33, 34, 35, 36, 45, 46, 47, 48, 49, 50, 58, 59, 60, 61, 62, 63, 64, 71, 72, 73, 74, 75, 76, 77, 84, 85, 86, 87, 88, 89, 90, 97, 98, 103], "bbox": [0.47341333333333335, 0.05998, 1.0, 0.4173], "iscrowd": 0, "area": 20505.529599999998, "rle": {"size": [500, 375], "counts": "gnf21c?1O1O2N1O1O1bDOW73THa0l7_OmGh0R8YOhGm0W8TObGS1]8oNXG[1g8fNmFf1R9[NfFm1Y9TNaFR2^9PN[FV2d9kMWFZ2h9gMRF_2m9gMiE^2W:gMYEd2f:V1O1O1O1O1O1O0000O100O1O100O1O100O1O010O1O100O1O100O100O1O1O1O00M3M2N3M3002M2O2N1O2M2O2N1O2M2O2N1O2N2M2O2N1010O100O010O10O0100O10O010O001O001O001O0O2O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O00100O2O0O1O2O0O101N1O101N101N1O101N100O1O10O0100O0010O0100O00100O010O1O010O10O01O10O0100O00100O010O1O010O1O00100O001O1O010O1O001O1M2N3M3M2N3M3MgK"}, "label": "a zebra with its head on the back of another zebra"}]}
{"id": 293311, "image": "COCO_train2014_000000293311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady holding food for the swooping lady\"."}], "task": "refering", "answer_template": "The \"a lady holding food for the swooping lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 177, 189, 190, 191, 192, 193, 194, 195, 196, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.22140624999999997, 0.10896955503512881, 0.876625, 1.0], "iscrowd": 0, "area": 94493.44815000001, "rle": {"size": [427, 640], "counts": "P]k1f0e;R1H7L4K5mJkMjNZ2?bNXOb11TOKQ1@C=a0@A<d0@_O=e0@]O<h0@[O<j0AXO<l0@WO=m0_OVO?n0]OTO`0P1\\OSOb0P1[OROc0Q1YOROe0Q1XOQOe0S1WOPOg0S1VOoNh0T1TOoNj0T1ROoNk0U1ROmNl0V1POmNn0V1oNlNo0W1mNlNP1X1mNjNQ1Z1jNiNT1Z1hNiNV1Z1fNiNW1[1eNhNY1[1cNhN[1[1aNhN]1[1_NhN_1[1]NhNc1Y1YNjNg1W1UNlNj1V1RNmNn1T1nMoNR2U6O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100000000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000002N3M3M2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M3M2PI\\LU4g3fK_LV4c3gKcLT4`3gKgLT4Y4kJlKQ5V4kJQLP5P4nJaLb4a3[KeL`4\\3]KkL^4V3`KPM[4R3aKUMZ4l2dKWMZ4k2cKYMZ4h2eKZMY4g2eK]MX4e2eK^MY4c2fK`MW4a2gKbMW4`2fKdMW4]2gKfMW4\\2gKfMW4j5M2N3M2N3M2N3M2N3M2N2N3M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O00001O00001O00001O0000O1M3M3M3M3M3M3M3M3M3N2M3M3XN]MeHf2U7jM^HY2[7XNXHk1b7R2N2O1UOcJfI_5Z6gJ^I[5b6lJVIU5i6g00000O100O10000O1002N2N2N2N2N2N2N2N2N3M3M3M3M4LZOdJ`IY5`6iJ\\IY5c6kJTI[5l6gJlH_5S7=0000O100O10000O12N2N1O2N1ON2O1O1N2O1O1NZMQITOo6k0RIUOn6i0UIVOl6g0VIYOk6d0XI[Oh6c0ZI]Og6`0\\I_Od6?^IAc6<`ICa6;`IE`69cIF^67dII]64fIKZ63hIMY60jIOV6_NhHa0T1P1U6ZNkHe0R1P1T6WNnHf0P1S1R6TNQIg0P1T1P6PNUIi0m0V1o5nMVIk0n0V1m5kMVIo0o0V1l5gMVIS1Q1T1_6mNbIS1\\6mNgIQ1X6oNjIQ1T6oNoIP1o5POSJo0l5ROVJm0h5SOZJl0c5VO`Ji0\\5YOgJf0V5[OlJd0Q5_OQK`0k4BWK<m4@VK>l4_OVK?n4\\OVKb0l4ZOWKd0l4YOWKe0l4VOWKh0l4TOXKj0l4POWKn0S5hNPKV1c8N3N1N3N1N2O2M2O2M2O1N3N1N3N1N2O2M2O2M2O1N3NbbP1"}, "label": "a lady holding food for the swooping lady"}]}
{"id": 293311, "image": "COCO_train2014_000000293311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding the food\"."}], "task": "refering", "answer_template": "The \"the woman holding the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 177, 189, 190, 191, 192, 193, 194, 195, 196, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 266, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.22140624999999997, 0.10896955503512881, 0.876625, 1.0], "iscrowd": 0, "area": 94493.44815000001, "rle": {"size": [427, 640], "counts": "P]k1f0e;R1H7L4K5mJkMjNZ2?bNXOb11TOKQ1@C=a0@A<d0@_O=e0@]O<h0@[O<j0AXO<l0@WO=m0_OVO?n0]OTO`0P1\\OSOb0P1[OROc0Q1YOROe0Q1XOQOe0S1WOPOg0S1VOoNh0T1TOoNj0T1ROoNk0U1ROmNl0V1POmNn0V1oNlNo0W1mNlNP1X1mNjNQ1Z1jNiNT1Z1hNiNV1Z1fNiNW1[1eNhNY1[1cNhN[1[1aNhN]1[1_NhN_1[1]NhNc1Y1YNjNg1W1UNlNj1V1RNmNn1T1nMoNR2U6O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100000000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000002N3M3M2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M3M2PI\\LU4g3fK_LV4c3gKcLT4`3gKgLT4Y4kJlKQ5V4kJQLP5P4nJaLb4a3[KeL`4\\3]KkL^4V3`KPM[4R3aKUMZ4l2dKWMZ4k2cKYMZ4h2eKZMY4g2eK]MX4e2eK^MY4c2fK`MW4a2gKbMW4`2fKdMW4]2gKfMW4\\2gKfMW4j5M2N3M2N3M2N3M2N3M2N2N3M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O00001O00001O00001O0000O1M3M3M3M3M3M3M3M3M3N2M3M3XN]MeHf2U7jM^HY2[7XNXHk1b7R2N2O1UOcJfI_5Z6gJ^I[5b6lJVIU5i6g00000O100O10000O1002N2N2N2N2N2N2N2N2N3M3M3M3M4LZOdJ`IY5`6iJ\\IY5c6kJTI[5l6gJlH_5S7=0000O100O10000O12N2N1O2N1ON2O1O1N2O1O1NZMQITOo6k0RIUOn6i0UIVOl6g0VIYOk6d0XI[Oh6c0ZI]Og6`0\\I_Od6?^IAc6<`ICa6;`IE`69cIF^67dII]64fIKZ63hIMY60jIOV6_NhHa0T1P1U6ZNkHe0R1P1T6WNnHf0P1S1R6TNQIg0P1T1P6PNUIi0m0V1o5nMVIk0n0V1m5kMVIo0o0V1l5gMVIS1Q1T1_6mNbIS1\\6mNgIQ1X6oNjIQ1T6oNoIP1o5POSJo0l5ROVJm0h5SOZJl0c5VO`Ji0\\5YOgJf0V5[OlJd0Q5_OQK`0k4BWK<m4@VK>l4_OVK?n4\\OVKb0l4ZOWKd0l4YOWKe0l4VOWKh0l4TOXKj0l4POWKn0S5hNPKV1c8N3N1N3N1N2O2M2O2M2O1N3N1N3N1N2O2M2O2M2O1N3NbbP1"}, "label": "the woman holding the food"}]}
{"id": 293311, "image": "COCO_train2014_000000293311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with her mouth wide open ready to take a bite\"."}], "task": "refering", "answer_template": "The \"a woman with her mouth wide open ready to take a bite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297, 298, 321], "bbox": [0.664828125, 0.18838407494145198, 0.99734375, 0.8973536299765809], "iscrowd": 0, "area": 35630.24774999999, "rle": {"size": [427, 640], "counts": "P[a52X=d0\\Od0\\Ob0^O3M3M3M3M2M4M3M3M2N3M3M2N2N3M2O1N2N2O1N3M100O1O2N100O1O101N1O100O1O101N1O100O1O1O100O2N100O2N101N1O2O0O101O001O00001O00000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000001O001O0O2O001O001O001O0010O0010O010O0100O010O010>A`0A?@4M1N3N1N2O1N3N1N2O1N3N1N100O100O2O0O100O100O10001N100O100O1000001O00000000000000001O1O00000000O1000001O000O1000000000000O1000000000000O01N2N2O1N2N2N101N2N2^f0"}, "label": "a woman with her mouth wide open ready to take a bite"}]}
{"id": 293311, "image": "COCO_train2014_000000293311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman opening her mouth for some food\"."}], "task": "refering", "answer_template": "The \"a woman opening her mouth for some food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297, 298, 321], "bbox": [0.664828125, 0.18838407494145198, 0.99734375, 0.8973536299765809], "iscrowd": 0, "area": 35630.24774999999, "rle": {"size": [427, 640], "counts": "P[a52X=d0\\Od0\\Ob0^O3M3M3M3M2M4M3M3M2N3M3M2N2N3M2O1N2N2O1N3M100O1O2N100O1O101N1O100O1O101N1O100O1O1O100O2N100O2N101N1O2O0O101O001O00001O00000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000001O001O0O2O001O001O001O0010O0010O010O0100O010O010>A`0A?@4M1N3N1N2O1N3N1N2O1N3N1N100O100O2O0O100O100O10001N100O100O1000001O00000000000000001O1O00000000O1000001O000O1000000000000O1000000000000O01N2N2O1N2N2N101N2N2^f0"}, "label": "a woman opening her mouth for some food"}]}
{"id": 522687, "image": "COCO_train2014_000000522687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green boat and trailer on wheels sitting on the grass in front of a building\"."}], "task": "refering", "answer_template": "The \"a green boat and trailer on wheels sitting on the grass in front of a building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 228, 229, 230, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 298, 299, 300, 301, 302, 303, 316, 317, 318, 319, 335, 336, 353, 354, 370, 371, 388], "bbox": [0.3955416666666667, 0.550203125, 0.8782708333333334, 0.9741875], "iscrowd": 0, "area": 30258.5892, "rle": {"size": [640, 480], "counts": "aUg3V1ea0U100000000O1000000O2N1O10000001O1O00001O0_Oa0B>003M3M2N2N1O101N1O2N1SOhNP_OY1_`0YO__Oi0_`0XO`_Oh0``0]1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O0010O01O001O1O001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O0001O00001O00001O0O101O01O100O2O0O2N101N1O2O1KfJ`A[5_>5O1N2O001O1O1O1N101O1N2O0O2O1N2O0O2O1N2O1O0O2O1N2O0O2O1N2O0O2O1OPMnCRNP<m1`DeM`;Y2cDgM[;X2hDhMW;V2kDkMS;T2PElMo:S2SElMl:S2VEnMi:P2ZEPNd:o1^ERNa:l1bESN^:k1eEUNY:j1iEWNV:h1lEXNR:g1QFXNo9f1SF[Nk9e1WF[Nh9c1ZF]Ne9SO^Cm1n2POe9QOjCb1b2^Oc9POWDP1\\20^9nN]DNS3T1a8lNcDJl2Z1b8jNmHU1T7iNmHX1S7fNnHZ1l<00000000O102N6I7J6J:F:EfdS1"}, "label": "a green boat and trailer on wheels sitting on the grass in front of a building"}]}
{"id": 522687, "image": "COCO_train2014_000000522687.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green color boat in the lawn\"."}], "task": "refering", "answer_template": "The \"a green color boat in the lawn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 228, 229, 230, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 298, 299, 300, 301, 302, 303, 316, 317, 318, 319, 335, 336, 353, 354, 370, 371, 388], "bbox": [0.3955416666666667, 0.550203125, 0.8782708333333334, 0.9741875], "iscrowd": 0, "area": 30258.5892, "rle": {"size": [640, 480], "counts": "aUg3V1ea0U100000000O1000000O2N1O10000001O1O00001O0_Oa0B>003M3M2N2N1O101N1O2N1SOhNP_OY1_`0YO__Oi0_`0XO`_Oh0``0]1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O0010O01O001O1O001O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O0001O00001O00001O0O101O01O100O2O0O2N101N1O2O1KfJ`A[5_>5O1N2O001O1O1O1N101O1N2O0O2O1N2O0O2O1N2O1O0O2O1N2O0O2O1N2O0O2O1OPMnCRNP<m1`DeM`;Y2cDgM[;X2hDhMW;V2kDkMS;T2PElMo:S2SElMl:S2VEnMi:P2ZEPNd:o1^ERNa:l1bESN^:k1eEUNY:j1iEWNV:h1lEXNR:g1QFXNo9f1SF[Nk9e1WF[Nh9c1ZF]Ne9SO^Cm1n2POe9QOjCb1b2^Oc9POWDP1\\20^9nN]DNS3T1a8lNcDJl2Z1b8jNmHU1T7iNmHX1S7fNnHZ1l<00000000O102N6I7J6J:F:EfdS1"}, "label": "a green color boat in the lawn"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone is driving a motorcycle with someone on the back\"."}], "task": "refering", "answer_template": "The \"someone is driving a motorcycle with someone on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 148, 149, 150, 172, 173, 174, 195, 196, 197, 219, 220, 242, 265], "bbox": [0.44529687500000004, 0.2786416861826698, 0.61321875, 0.7618032786885246], "iscrowd": 0, "area": 7229.314950000003, "rle": {"size": [427, 640], "counts": "UPg32\\<7UDO_;<\\DJ`;:[DJc;9WDLf;o0M4L3M3N2M3M3M3N2M2O1O2N101N1O2N1O2N1O1O2O0O1TFRMY9o2eFRM[9P3bFQM^9`2XFkM8G_9_2ZFiM5Ia9_2ZFiM3Hd9^2[FiMOKe9]2]FgMLMg9]2^FPNa9P2`FoM`9R2aFlM_9U2aFkM^99[Fn08hN^99[Fo08fN]9:^Fo05gN\\9:`Fo04fN\\9:cFo01gN[9:eFI36NFY9:hF^O=b0AHX98jG1nNGW97mG2kNHX95mG4jNHX93oG6hNH4Am0Oi64_Id0gNHKJo0Mg8b0^F7e0^Om8;]F8?DS95]F7<IW90\\F89LZ9K_F841]9F_F:13_9C`F:N7b9^O`Fl1_9TNbFl1^9SNbFn1]9RNcFo1]9PNcFQ2\\9oMeFQ2[9nMeFS2Z9mMgFc1HUN`98hFa1LVN\\98hFa1NVNZ99iF^11SN[9>dF]14QN[9a0cF[1Q:dNoEZ1S:fNnEW1U:gNlEW1V:iNkET1X:kNhES1Z:mNfEQ1]:mNeEP1]:POcEn0_:RObEk0`:TOaEi0b:WO_Ef0b:[OaE`0a:_OdE:_:FdE5^:KfEN]:1hEIY:8]10000O10O10O10000O011O1N2O1O1N101O1O1NfVW3"}, "label": "someone is driving a motorcycle with someone on the back"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with a white jacket in the front of the motorcycle\"."}], "task": "refering", "answer_template": "The \"person with a white jacket in the front of the motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 148, 149, 150, 172, 173, 174, 195, 196, 197, 219, 220, 242, 265], "bbox": [0.44529687500000004, 0.2786416861826698, 0.61321875, 0.7618032786885246], "iscrowd": 0, "area": 7229.314950000003, "rle": {"size": [427, 640], "counts": "UPg32\\<7UDO_;<\\DJ`;:[DJc;9WDLf;o0M4L3M3N2M3M3M3N2M2O1O2N101N1O2N1O2N1O1O2O0O1TFRMY9o2eFRM[9P3bFQM^9`2XFkM8G_9_2ZFiM5Ia9_2ZFiM3Hd9^2[FiMOKe9]2]FgMLMg9]2^FPNa9P2`FoM`9R2aFlM_9U2aFkM^99[Fn08hN^99[Fo08fN]9:^Fo05gN\\9:`Fo04fN\\9:cFo01gN[9:eFI36NFY9:hF^O=b0AHX98jG1nNGW97mG2kNHX95mG4jNHX93oG6hNH4Am0Oi64_Id0gNHKJo0Mg8b0^F7e0^Om8;]F8?DS95]F7<IW90\\F89LZ9K_F841]9F_F:13_9C`F:N7b9^O`Fl1_9TNbFl1^9SNbFn1]9RNcFo1]9PNcFQ2\\9oMeFQ2[9nMeFS2Z9mMgFc1HUN`98hFa1LVN\\98hFa1NVNZ99iF^11SN[9>dF]14QN[9a0cF[1Q:dNoEZ1S:fNnEW1U:gNlEW1V:iNkET1X:kNhES1Z:mNfEQ1]:mNeEP1]:POcEn0_:RObEk0`:TOaEi0b:WO_Ef0b:[OaE`0a:_OdE:_:FdE5^:KfEN]:1hEIY:8]10000O10O10O10000O011O1N2O1O1N101O1O1NfVW3"}, "label": "person with a white jacket in the front of the motorcycle"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a silver helmet riding on the back of a motorcycle\"."}], "task": "refering", "answer_template": "The \"a person with a silver helmet riding on the back of a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 146, 147, 148, 169, 170, 171, 172, 192, 193, 194, 195, 196, 216, 217, 218, 240, 241, 263, 264], "bbox": [0.35990625, 0.2799297423887588, 0.52853125, 0.7575409836065574], "iscrowd": 0, "area": 10600.163900000009, "rle": {"size": [427, 640], "counts": "UQP33h1Z1Z8XORGo0i8SOTGQ1i8RORGQ1l8SOoFQ1o8ROmFQ1Q9QOkFR1V9oNeFU1Y9nNcFT1\\9oN`FT1_9]1N101N1O2O00001O001O001O0010O0001O001O001O001O01O01O00001O00001O01O01O010O01XLPG2O\\2Q9^MWG0Lb2m8WM`G1Hf2h8TMhG0Dk2Q:O2N1O1O2N2N1O2NbFdMo7Z2oGjMP8U2nGQNP8m1mGVNU8h1iG[NW8d1gG_NZ8_1dGcN^8[1`GhN`8W1]GlNe8S1XGPOi8n0UGTOl8k0SGVOo8h0oFZOS9e0jF]OW9b0gFAZ9=dFE]9:aFHa97\\FKf93XFOi90UF3l9KRF7P:a11O2WMmEV2U:hMkEX2V:gMjEY2X:dMiE\\2Y:bMgEg12dNY:BfEh14eN^:X1eEhN[:U1iEiNY:T1jEkNV:R1mEnNT:o0oEPOQ:m0SFQOo9l0SFROo9l0UFQOm9l0XFPOi9n0\\1N1N3M2N3M2N3M3L5L4LRfm3"}, "label": "a person with a silver helmet riding on the back of a motorcycle"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"passenger riding on the back of a motorcycle\"."}], "task": "refering", "answer_template": "The \"passenger riding on the back of a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 146, 147, 148, 169, 170, 171, 172, 192, 193, 194, 195, 196, 216, 217, 218, 240, 241, 263, 264], "bbox": [0.35990625, 0.2799297423887588, 0.52853125, 0.7575409836065574], "iscrowd": 0, "area": 10600.163900000009, "rle": {"size": [427, 640], "counts": "UQP33h1Z1Z8XORGo0i8SOTGQ1i8RORGQ1l8SOoFQ1o8ROmFQ1Q9QOkFR1V9oNeFU1Y9nNcFT1\\9oN`FT1_9]1N101N1O2O00001O001O001O0010O0001O001O001O001O01O01O00001O00001O01O01O010O01XLPG2O\\2Q9^MWG0Lb2m8WM`G1Hf2h8TMhG0Dk2Q:O2N1O1O2N2N1O2NbFdMo7Z2oGjMP8U2nGQNP8m1mGVNU8h1iG[NW8d1gG_NZ8_1dGcN^8[1`GhN`8W1]GlNe8S1XGPOi8n0UGTOl8k0SGVOo8h0oFZOS9e0jF]OW9b0gFAZ9=dFE]9:aFHa97\\FKf93XFOi90UF3l9KRF7P:a11O2WMmEV2U:hMkEX2V:gMjEY2X:dMiE\\2Y:bMgEg12dNY:BfEh14eN^:X1eEhN[:U1iEiNY:T1jEkNV:R1mEnNT:o0oEPOQ:m0SFQOo9l0SFROo9l0UFQOm9l0XFPOi9n0\\1N1N3M2N3M2N3M3L5L4LRfm3"}, "label": "passenger riding on the back of a motorcycle"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car is just a blur in front of a speeding motorcycle\"."}], "task": "refering", "answer_template": "The \"a car is just a blur in front of a speeding motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.539453125, 0.273231850117096, 0.9387187499999999, 0.4610070257611242], "iscrowd": 0, "area": 12705.380999999998, "rle": {"size": [427, 640], "counts": "a]`42Y=0O2O0O1O100O101N1O100O100O1O2O0O100O1O100O2O0O1O100O100O2N100O100O1O101N100000000O101O0000000O10001O000O10003M4L4K5L4L4L4L0O101O0000000O100000001N100000000O1000001O00000000000000001O00000000000000001O00000000000000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O00O100O1O100O100O100O100O1O100O100O100O100O1O100O100O102N2N2N2N2N2N2N2N3M2N2N000000000001O0001O00000000000000000000000000000000001O000001O00000000000000000000000001O00000000000000000001O000000000001N1N2N4L3M3M3N3L3M3M3M4L3M3MaS`0"}, "label": "a car is just a blur in front of a speeding motorcycle"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black car in the front\"."}], "task": "refering", "answer_template": "The \"the black car in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.539453125, 0.273231850117096, 0.9387187499999999, 0.4610070257611242], "iscrowd": 0, "area": 12705.380999999998, "rle": {"size": [427, 640], "counts": "a]`42Y=0O2O0O1O100O101N1O100O100O1O2O0O100O1O100O2O0O1O100O100O2N100O100O1O101N100000000O101O0000000O10001O000O10003M4L4K5L4L4L4L0O101O0000000O100000001N100000000O1000001O00000000000000001O00000000000000001O00000000000000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O00O100O1O100O100O100O100O1O100O100O100O100O1O100O100O102N2N2N2N2N2N2N2N3M2N2N000000000001O0001O00000000000000000000000000000000001O000001O00000000000000000000000001O00000000000000000001O000000000001N1N2N4L3M3M3N3L3M3M3M4L3M3MaS`0"}, "label": "the black car in the front"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry blue car behind a man and woman riding a motorbike\"."}], "task": "refering", "answer_template": "The \"a blurry blue car behind a man and woman riding a motorbike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 127, 128, 129, 141, 142, 143, 144, 145, 146, 150, 151, 165, 166, 167, 168], "bbox": [0.14154687500000002, 0.27096018735363, 0.63003125, 0.5179625292740048], "iscrowd": 0, "area": 12011.09935, "rle": {"size": [427, 640], "counts": "hSV17m<8H9H7O2M2O2N1O2N1O1O2M2O00000000001N1000000000000O100000001O000O1000000000000O101O000000000O1000000000001N100O1O1O1O1O1O1O1O1O3M3M3M3M3M3M3M3M3M1O1O1O1O1O1O1O1O100000000000000O10000000000000000O010000O1000000O100000O010000O1000000O1000000O1000000O10O10O1N2L4M3L4M3M3N2N2UNPEY1R;gNSER1o:mNVEm0l:ROZEh0g:WO^Ec0d:]O_E>c:A_E<c:C_E;a:EbE7`:IaE4a:KaE3`:LbE1`:NbE0_:0cEL_:3]1O1O100OjTU19njjN3M001O1O001O1O001O001O1O100O1O1O1O2N1O1O1kCTOi;m0PDYOP<R1O000000000000000000O1000000N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N201O0000001O00000O101O000000aeR3"}, "label": "a blurry blue car behind a man and woman riding a motorbike"}]}
{"id": 47554, "image": "COCO_train2014_000000047554.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue blurry car\"."}], "task": "refering", "answer_template": "The \"blue blurry car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 127, 128, 129, 141, 142, 143, 144, 145, 146, 150, 151, 165, 166, 167, 168], "bbox": [0.14154687500000002, 0.27096018735363, 0.63003125, 0.5179625292740048], "iscrowd": 0, "area": 12011.09935, "rle": {"size": [427, 640], "counts": "hSV17m<8H9H7O2M2O2N1O2N1O1O2M2O00000000001N1000000000000O100000001O000O1000000000000O101O000000000O1000000000001N100O1O1O1O1O1O1O1O1O3M3M3M3M3M3M3M3M3M1O1O1O1O1O1O1O1O100000000000000O10000000000000000O010000O1000000O100000O010000O1000000O1000000O1000000O10O10O1N2L4M3L4M3M3N2N2UNPEY1R;gNSER1o:mNVEm0l:ROZEh0g:WO^Ec0d:]O_E>c:A_E<c:C_E;a:EbE7`:IaE4a:KaE3`:LbE1`:NbE0_:0cEL_:3]1O1O100OjTU19njjN3M001O1O001O1O001O001O1O100O1O1O1O2N1O1O1kCTOi;m0PDYOP<R1O000000000000000000O1000000N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N201O0000001O00000O101O000000aeR3"}, "label": "blue blurry car"}]}
{"id": 170439, "image": "COCO_train2014_000000170439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black rashguard on a beach , sitting on a concrete object , with his hands on his thighs\"."}], "task": "refering", "answer_template": "The \"a man wearing a black rashguard on a beach , sitting on a concrete object , with his hands on his thighs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 178, 179, 201, 202, 203, 204, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 318], "bbox": [0.68803125, 0.37381422924901186, 0.948046875, 0.7339920948616601], "iscrowd": 0, "area": 14195.62975, "rle": {"size": [506, 640], "counts": "YXj61h?2N1O1O1O1O1O1O2N1O2N100O1O2N1O1O2N100O1O2N1O1O2N100O1O2N1O101O00000O2O00001O00000]OjNbBW1j<DnB<n<0jB0d:_OQGo1o8YNeFl1Y9WNbFl1^9WN[Fm1d9\\NPFh1P:Q201N10O0O2O1O0O2O1O0O2O001O11N100O101O0O11O010O001O010O001O010O1O1O6K48HK6J5K5K3M1O001O001O1O001O001O001N101N1O2N101N1O2N101N1O2N1M4K4L5K4L5K4I8I6I8I6I8I7J600OUBjNS=U1kBnNU=R1hBQOX=n0fBTO\\=j0bBYO^=g0^B]Ob=b0\\BAd=>ZBEf=:XBHi=U11N1N2O1N2O1N3M2O1N2O2M2O1N3N1N2O2M2O1N3M2O1N4M4K5L4K5LSP`0"}, "label": "a man wearing a black rashguard on a beach , sitting on a concrete object , with his hands on his thighs"}]}
{"id": 170439, "image": "COCO_train2014_000000170439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in wet suit beside surfboard\"."}], "task": "refering", "answer_template": "The \"man in wet suit beside surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 178, 179, 201, 202, 203, 204, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 318], "bbox": [0.68803125, 0.37381422924901186, 0.948046875, 0.7339920948616601], "iscrowd": 0, "area": 14195.62975, "rle": {"size": [506, 640], "counts": "YXj61h?2N1O1O1O1O1O1O2N1O2N100O1O2N1O1O2N100O1O2N1O1O2N100O1O2N1O101O00000O2O00001O00000]OjNbBW1j<DnB<n<0jB0d:_OQGo1o8YNeFl1Y9WNbFl1^9WN[Fm1d9\\NPFh1P:Q201N10O0O2O1O0O2O1O0O2O001O11N100O101O0O11O010O001O010O001O010O1O1O6K48HK6J5K5K3M1O001O001O1O001O001O001N101N1O2N101N1O2N101N1O2N1M4K4L5K4L5K4I8I6I8I6I8I7J600OUBjNS=U1kBnNU=R1hBQOX=n0fBTO\\=j0bBYO^=g0^B]Ob=b0\\BAd=>ZBEf=:XBHi=U11N1N2O1N2O1N3M2O1N2O2M2O1N3N1N2O2M2O1N3M2O1N4M4K5L4K5LSP`0"}, "label": "man in wet suit beside surfboard"}]}
{"id": 375245, "image": "COCO_train2014_000000375245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the pipe in his mouth\"."}], "task": "refering", "answer_template": "The \"the man with the pipe in his mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 33, 34, 35, 36, 47, 48, 49, 50, 51, 61, 62, 63, 64, 75, 76, 77, 78, 79, 88, 89, 90, 91, 92, 93, 94, 101, 102, 103, 104, 105, 106, 107, 108, 115, 116, 117, 118, 119, 120, 121, 122, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 200, 201, 202, 203, 204, 205, 206, 207, 213, 214, 215, 216, 217, 218, 219, 220, 221, 227, 228, 229, 231, 234, 235, 241, 242, 243, 248, 249, 255, 256, 257, 262, 263, 269, 270, 271, 277, 283, 284, 285, 297, 298, 311, 312], "bbox": [0.19578534031413614, 0.07640625, 0.8471204188481676, 0.97303125], "iscrowd": 0, "area": 75814.59850000001, "rle": {"size": [640, 382], "counts": "TU_14ic07I8I6I7J6I7dFfNHa10bNMd1L`N2b1KaN2b1KaN3a1JbN3a1JbN4`1IcN4`1HdN6^1GeN6^1GeN7]1FfN7\\1GgN7[1FhN7[1EiN9Y1DjN9Y1DjN:X1CkN:X1CkN;X1@lN>W1^OlN?Y1\\OjNb0Y1ZOjNb0[1hK]MSN6nNU1S7]1dKJXMlNP7_1aKI]MlNn6`1]K_2_4fMZK`2b4eMWKa2f4cMRKd2j4aMoJe2m4cMhJd2T5hM^J^2a5kMQJ[2o5k5O100O100O10000O100O01L3N3M2N3L4M2N3M2M4M2N3L3QMnHoHU7o6oHlHT7R7QIhHQ7W7TIbHP7\\7TI_H_4RNVM^9`NZH\\4\\NQMX9hNVHX4gNlLQ9QORHV4QOfLk8YOnGS4[O`Lf8BiGo3F\\L_8IeGn30VLY81aGj3:RLT88]Gh3c0mKn7`0YGe3m0gKi7i0SGb3Y1bKa7Q1QG_3b1\\K\\7P7iHmHU7T7nHiHQ7W7TIfHj6[7YIbHf6_7[IaHc6_7_IaH^6a7cI^H\\6b7fI^HX6c7jI[HU6e7mI[HQ6f7QJXHn5i7ZJPHd5P8eJhGZ5Y8T3101O001O2N2N2M3N3M2N2N21O3M3M3M3M2N3N2M3M3M3M2N3M3M3M3N2M2N3M3M3M3M2N3M2OM2O1O1N2O1O1N3N2M2O2N1N3N2M2O2N2M2O2N2mFlGV6V8i22M3SGeGR6\\8nIdGQ6^8f23XG_Gn5b8RJ^Gl5e8SJ\\Gk5f8TJZGk5g8VJXGh5k8WJVGg5l8XJTGg5m8YJTGd5o8\\JPGc5Q9cInFoN1\\7U9bIkFRO1Z7Z9^IfFXO0Y7_9XIeFnN<h7U9TI]Gj6i8oHYGo6m8jHWGT7n8fHTGY7R9_HQG_7U9[HnFc7h:O2M2O1N3N1O1N30O4M2M4L3N3L3N3L3M4M2M4L4M2M4M2M4L3N3L3N3L3M4M2M4L3N3L3N3L3M4M2M4M2M4O01SJaFU1_9gNcF[1]9aNdFa1]9ZNfFe1]9WNeFg1_9UNbFj1b9YKZF^16W3d9PK_Fg1NX3f9hJdFm1IY3g9_JiFV2BY3l:fLUEY3m:eLUEY3m:eLUEY3l:eLVEZ3l:dLVEZ3l:dLVEZ3l:dLUE[3l:cLWEDVOn1f;\\NVEROM[2o:aNUEaNd0h2Y:eN[GU1f8jNaGo0b8mNeGm0^8POhGj0[8SOlGf0W8WOoGc0S8[OSH?P8[O[H=h7VOmH?V7UO^I`0\\=E<D^gS1"}, "label": "the man with the pipe in his mouth"}]}
{"id": 375245, "image": "COCO_train2014_000000375245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man smiling with a pipe in his mouth\"."}], "task": "refering", "answer_template": "The \"a man smiling with a pipe in his mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 33, 34, 35, 36, 47, 48, 49, 50, 51, 61, 62, 63, 64, 75, 76, 77, 78, 79, 88, 89, 90, 91, 92, 93, 94, 101, 102, 103, 104, 105, 106, 107, 108, 115, 116, 117, 118, 119, 120, 121, 122, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 189, 190, 191, 192, 193, 200, 201, 202, 203, 204, 205, 206, 207, 213, 214, 215, 216, 217, 218, 219, 220, 221, 227, 228, 229, 231, 234, 235, 241, 242, 243, 248, 249, 255, 256, 257, 262, 263, 269, 270, 271, 277, 283, 284, 285, 297, 298, 311, 312], "bbox": [0.19578534031413614, 0.07640625, 0.8471204188481676, 0.97303125], "iscrowd": 0, "area": 75814.59850000001, "rle": {"size": [640, 382], "counts": "TU_14ic07I8I6I7J6I7dFfNHa10bNMd1L`N2b1KaN2b1KaN3a1JbN3a1JbN4`1IcN4`1HdN6^1GeN6^1GeN7]1FfN7\\1GgN7[1FhN7[1EiN9Y1DjN9Y1DjN:X1CkN:X1CkN;X1@lN>W1^OlN?Y1\\OjNb0Y1ZOjNb0[1hK]MSN6nNU1S7]1dKJXMlNP7_1aKI]MlNn6`1]K_2_4fMZK`2b4eMWKa2f4cMRKd2j4aMoJe2m4cMhJd2T5hM^J^2a5kMQJ[2o5k5O100O100O10000O100O01L3N3M2N3L4M2N3M2M4M2N3L3QMnHoHU7o6oHlHT7R7QIhHQ7W7TIbHP7\\7TI_H_4RNVM^9`NZH\\4\\NQMX9hNVHX4gNlLQ9QORHV4QOfLk8YOnGS4[O`Lf8BiGo3F\\L_8IeGn30VLY81aGj3:RLT88]Gh3c0mKn7`0YGe3m0gKi7i0SGb3Y1bKa7Q1QG_3b1\\K\\7P7iHmHU7T7nHiHQ7W7TIfHj6[7YIbHf6_7[IaHc6_7_IaH^6a7cI^H\\6b7fI^HX6c7jI[HU6e7mI[HQ6f7QJXHn5i7ZJPHd5P8eJhGZ5Y8T3101O001O2N2N2M3N3M2N2N21O3M3M3M3M2N3N2M3M3M3M2N3M3M3M3N2M2N3M3M3M3M2N3M2OM2O1O1N2O1O1N3N2M2O2N1N3N2M2O2N2M2O2N2mFlGV6V8i22M3SGeGR6\\8nIdGQ6^8f23XG_Gn5b8RJ^Gl5e8SJ\\Gk5f8TJZGk5g8VJXGh5k8WJVGg5l8XJTGg5m8YJTGd5o8\\JPGc5Q9cInFoN1\\7U9bIkFRO1Z7Z9^IfFXO0Y7_9XIeFnN<h7U9TI]Gj6i8oHYGo6m8jHWGT7n8fHTGY7R9_HQG_7U9[HnFc7h:O2M2O1N3N1O1N30O4M2M4L3N3L3N3L3M4M2M4L4M2M4M2M4L3N3L3N3L3M4M2M4L3N3L3N3L3M4M2M4M2M4O01SJaFU1_9gNcF[1]9aNdFa1]9ZNfFe1]9WNeFg1_9UNbFj1b9YKZF^16W3d9PK_Fg1NX3f9hJdFm1IY3g9_JiFV2BY3l:fLUEY3m:eLUEY3m:eLUEY3l:eLVEZ3l:dLVEZ3l:dLVEZ3l:dLUE[3l:cLWEDVOn1f;\\NVEROM[2o:aNUEaNd0h2Y:eN[GU1f8jNaGo0b8mNeGm0^8POhGj0[8SOlGf0W8WOoGc0S8[OSH?P8[O[H=h7VOmH?V7UO^I`0\\=E<D^gS1"}, "label": "a man smiling with a pipe in his mouth"}]}
{"id": 375245, "image": "COCO_train2014_000000375245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man whose left leg and arm are only seen\"."}], "task": "refering", "answer_template": "The \"the man whose left leg and arm are only seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 56, 70, 84, 85, 86, 98, 99, 100, 112, 113, 114, 126, 127, 128, 140, 141, 142, 154, 155, 156, 168, 169, 170, 182, 183, 184, 196, 197, 198, 211, 212, 225, 226, 239, 240, 241, 253, 254, 255, 267, 268, 269], "bbox": [0.0, 0.13443750000000002, 0.23986910994764396, 0.8503281250000001], "iscrowd": 0, "area": 21993.22655, "rle": {"size": [640, 382], "counts": "i2m167i?_O\\@[4o;_KQDb4S<XKnCi4W<QKiCP5W<oJjCQ5V<mJkCT5T<lJlCV5S<hJnCZ5Q<eJPD\\5o;cJQD^5o;`JRDb5m;]JTDd5k;ZJVDP6a;mIaD_6T;^InDe6o:ZISEg6R<3M2N2O1N2N1O2O1N2N2O1N4L4L4M3L4L4L4M1N2N2O1O1O1O1Oe0[O2N1O1O2N1O2NY1gNO100O100O1O2O2M2N2O>A5L4K5K5L4K5L4K5K5L4K6J5SIbEo3b:jKhEn3^:kKkEn3Z:jKPFn3U:kKUFm3Q:lKXFm3l9lK^Fl3h9lKSHZ2k<N2N3L3N3L3J7I6I7J7I6I7JUid5"}, "label": "the man whose left leg and arm are only seen"}]}
{"id": 375245, "image": "COCO_train2014_000000375245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right side of a man who is mostly out of view\"."}], "task": "refering", "answer_template": "The \"the right side of a man who is mostly out of view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 68, 69, 83, 96, 97, 109, 110, 111, 123, 124, 125, 137, 138, 139, 151, 152, 153, 165, 166, 167, 179, 180, 181, 194, 195, 207, 208, 209, 222, 223, 236, 237, 250, 251, 264, 265, 278, 279], "bbox": [0.8264397905759162, 0.12421875, 1.0, 0.834984375], "iscrowd": 0, "area": 20975.3856, "rle": {"size": [640, 382], "counts": "cjU67`c0:F:RM9h@b0?\\NS=l3nBULk<P4TCRLg<R4VCSLc<S4YCRL`<S4]CQL^<S4^CSL]<Q4_CTL]<n3`CWL\\<k3aCZL[<h3aC^L[<e3aCaLZ<a3cCdLY<c3_CcL\\<d5L4L5K4L4L5K4UFQH_7S8WHSHi7P8nGVHQ8n7fGXHZ8l7\\GZHd8i7SG]Hm8g7iF_HV9V900O1O1O100XMoE`KR:_4QF^KP:a4SF\\Kn9c4VFZKj9f4XFWKi9h4[FTKf9k4]FRKd9m4_FQKa9o4bFmJ_9R5dFkJ]9T5gFhJZ9W5iFgJW9Y5kFdJV9[5nFaJS9^5QG^JP9a5UG[Jk8d5YGXJh8h5\\GSJe8l5h2O1O100O1O1O10O100A?[Oe0[OUN"}, "label": "the right side of a man who is mostly out of view"}]}
{"id": 375245, "image": "COCO_train2014_000000375245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person on the far right\"."}], "task": "refering", "answer_template": "The \"the person on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 68, 69, 83, 96, 97, 109, 110, 111, 123, 124, 125, 137, 138, 139, 151, 152, 153, 165, 166, 167, 179, 180, 181, 194, 195, 207, 208, 209, 222, 223, 236, 237, 250, 251, 264, 265, 278, 279], "bbox": [0.8264397905759162, 0.12421875, 1.0, 0.834984375], "iscrowd": 0, "area": 20975.3856, "rle": {"size": [640, 382], "counts": "cjU67`c0:F:RM9h@b0?\\NS=l3nBULk<P4TCRLg<R4VCSLc<S4YCRL`<S4]CQL^<S4^CSL]<Q4_CTL]<n3`CWL\\<k3aCZL[<h3aC^L[<e3aCaLZ<a3cCdLY<c3_CcL\\<d5L4L5K4L4L5K4UFQH_7S8WHSHi7P8nGVHQ8n7fGXHZ8l7\\GZHd8i7SG]Hm8g7iF_HV9V900O1O1O100XMoE`KR:_4QF^KP:a4SF\\Kn9c4VFZKj9f4XFWKi9h4[FTKf9k4]FRKd9m4_FQKa9o4bFmJ_9R5dFkJ]9T5gFhJZ9W5iFgJW9Y5kFdJV9[5nFaJS9^5QG^JP9a5UG[Jk8d5YGXJh8h5\\GSJe8l5h2O1O100O1O1O10O100A?[Oe0[OUN"}, "label": "the person on the far right"}]}
{"id": 522704, "image": "COCO_train2014_000000522704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black and white keyboard next to the silver spoon\"."}], "task": "refering", "answer_template": "The \"the black and white keyboard next to the silver spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137], "bbox": [0.6623593750000001, 0.0, 1.0, 0.33864583333333337], "iscrowd": 0, "area": 31874.943149999996, "rle": {"size": [480, 640], "counts": "ShV6g1V=g1\\NO1000000000000O1000000000000O1000000000000O100000001O000O1000000000000O15K6J5K6J5K6J5K6J5K5K00000000000000000000001O00000000000000000000000000000000000000000001O000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000001O0000000000000000000000000000000000000000000001O00000L"}, "label": "the black and white keyboard next to the silver spoon"}]}
{"id": 522704, "image": "COCO_train2014_000000522704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a keyboard and backside of a spoon\"."}], "task": "refering", "answer_template": "The \"a keyboard and backside of a spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137], "bbox": [0.6623593750000001, 0.0, 1.0, 0.33864583333333337], "iscrowd": 0, "area": 31874.943149999996, "rle": {"size": [480, 640], "counts": "ShV6g1V=g1\\NO1000000000000O1000000000000O1000000000000O100000001O000O1000000000000O15K6J5K6J5K6J5K6J5K5K00000000000000000000001O00000000000000000000000000000000000000000001O000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000001O0000000000000000000000000000000000000000000001O00000L"}, "label": "a keyboard and backside of a spoon"}]}
{"id": 244179, "image": "COCO_train2014_000000244179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rinocerose on top of another one\"."}], "task": "refering", "answer_template": "The \"a rinocerose on top of another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 211, 212, 213, 214, 215, 216, 234, 235, 237, 238, 239, 257, 258, 260, 261, 279, 280, 281, 284, 285, 302, 303, 307, 308, 309], "bbox": [0.149921875, 0.009086651053864168, 0.7913125, 0.9362763466042155], "iscrowd": 0, "area": 72953.70485000002, "rle": {"size": [427, 640], "counts": "m\\X12V=4K5L4L4L4^JBQNc0g1IQN<g1NSN4b1:YNIc1=YNDc1d0WN^Of1l0RNUOk1P1QNSOk1S1QNnNl1W1PNkNm1[1oMfNn1`1mMcNo1b1nM_No1g1lM[NQ2P2fMRNV2X2aMjM\\2\\2_MfM\\2b2_M`MX2Q3VLVK?j1W3a3gKQKj0_1\\3]4\\LeK_3d4XLaKd3T7M4M201NM4K50O0100O01000O1000O1000O1iNTLPIm3o6WLmHi3R7\\LiHe3V7`LfH`3Z7dLVHANl3j7Z1ZNQKnJP5h4_KPKd4k4fKlJ^4S5hKdJ]4Z5iK]J[4b5g10000100O001O00100O0010O0100O100O4Mh0WO10001N10O0100O010O01O001O1O0010O01O0010O01O10O010O2O1N2O1N2O1O1N2O1`KRI`2P7VM\\Ig2d6VMbIf2`6WMcIh2`6RMdIm2`6lLdIR3`6hLdIW3`6`LgI]3a6XLdIg3S8L4M2N3L3N2M3N2M2O1O1N2O0O2O2M3N2N3L3N2M2O2M2O2N1N3N2M3N1N3N2N2M3N2MVNdNiGZ1P8oNoGo0k7ZOUHd0g7CWH<g7HWH7h7MVH3h72UHNi76UHJi7:UHEj7?THAi7e0TH[Oh7k0VHTOh7Q1VHoNf7X1WHhNf7]1YHbNc7d1[H[Nc7j1[HVNa7Q2\\HoMa7U2^HkM^7Z2aHeM]7^2cHbMY7c2fH]MW7g2hHXMU7m2jHSMS7Q3lHoLP7V3oHjLn6Z3QIeLl6`3SI`Lj6d3UI\\Li6g3VIYLi6i3VIWLj6j3UIVLj6l3UITLj6n3UIRLk6o3TIQLk6Q4TIoKk6S4UIlKk6U4TIkKk6W4TIiKk6Y4TIgKk6[4TIeKl6\\4SIdKl6^4SIbKl6_500O1O100O1O100O1O2N100O1O2O0O1O2O0O1O1N3M3M3N2M3M2N3M3M3M3N2M3N2N2N101N101N2O001N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1O1O1O1O1O2O0O2N1O1O2N101N1O1O001O010O001O001O01O0N2O1O1O1OWLdFf3Z9ZLfFg3Y9YLhFg3W9YLjFg3V9XLjFi3U9WLlFi3S9WLoFY2KSOT9dNSGX2ITOS9dNUGW2JTOP9eNXGV2HUOo8eN[GU2GUOm8fN]GU2FUOl8fN_GU2FUOi8fNaGU2GUOg8eNbGX2HROf8eNaG[2IPOT9Q1mFnNR9R1PGmNo8T1QGmNn8S1SGlNm8T1SGlNm8T1TGkNl8T1UGlNk8T1VGlNi8T1WGlNi8S1XGmNh8S1YGlNg8T1YGlNg8T1ZGkNf8T1[GmNe8R1\\GmNd8S1\\GmNd8S1]GlNc8S1^GmNb8S1^GmNb8S1_GmN`8R1aGnN_8R1bGmN^8S1bGnN]8R1cGnN]8Q1eGnN\\8Q1dGoN\\8Q1dGPO[8P1fGoNZ8P1gGPOY8P1hGPOX8o0hGQOb7nNTHP2;ROa7TOnGj1b0QO`7[OiGc1g0SO_7_OeG^1m0SO^7c1cH]N^7`1dH`N\\7_1eHaN\\7]1eHcN\\7[1dHfN]7W1eHiN[7V1fHjN[7T1fHlN[7R1fHnN\\7o0dHRO]7k0eHTO]7j0dHVO^7g0cHZO]7d0dH\\O]7b0gH\\OZ7`0hHAX7=iHCX7;iHFW78jHIV75kHKU74mHLS72nHOS7MoH3R7XOTGNk1k0R7ROVG2i1l0X7hNRG;g1m0c7RO^Hn0c7PO^HP1b7oN_HQ1a7nN`HR1`7mNaHS1_7lNbHT1^7kNcHU1^7iNcHW1]7hNcHY1]7fNdHZ1g900O100000O10000000000000000001O0O100O101N1O100O101N100O1O101N2O1N2M2N2N2O1N2N2N2N2Nadg1"}, "label": "a rinocerose on top of another one"}]}
{"id": 244179, "image": "COCO_train2014_000000244179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink cow on top of another cow\"."}], "task": "refering", "answer_template": "The \"a pink cow on top of another cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 211, 212, 213, 214, 215, 216, 234, 235, 237, 238, 239, 257, 258, 260, 261, 279, 280, 281, 284, 285, 302, 303, 307, 308, 309], "bbox": [0.149921875, 0.009086651053864168, 0.7913125, 0.9362763466042155], "iscrowd": 0, "area": 72953.70485000002, "rle": {"size": [427, 640], "counts": "m\\X12V=4K5L4L4L4^JBQNc0g1IQN<g1NSN4b1:YNIc1=YNDc1d0WN^Of1l0RNUOk1P1QNSOk1S1QNnNl1W1PNkNm1[1oMfNn1`1mMcNo1b1nM_No1g1lM[NQ2P2fMRNV2X2aMjM\\2\\2_MfM\\2b2_M`MX2Q3VLVK?j1W3a3gKQKj0_1\\3]4\\LeK_3d4XLaKd3T7M4M201NM4K50O0100O01000O1000O1000O1iNTLPIm3o6WLmHi3R7\\LiHe3V7`LfH`3Z7dLVHANl3j7Z1ZNQKnJP5h4_KPKd4k4fKlJ^4S5hKdJ]4Z5iK]J[4b5g10000100O001O00100O0010O0100O100O4Mh0WO10001N10O0100O010O01O001O1O0010O01O0010O01O10O010O2O1N2O1N2O1O1N2O1`KRI`2P7VM\\Ig2d6VMbIf2`6WMcIh2`6RMdIm2`6lLdIR3`6hLdIW3`6`LgI]3a6XLdIg3S8L4M2N3L3N2M3N2M2O1O1N2O0O2O2M3N2N3L3N2M2O2M2O2N1N3N2M3N1N3N2N2M3N2MVNdNiGZ1P8oNoGo0k7ZOUHd0g7CWH<g7HWH7h7MVH3h72UHNi76UHJi7:UHEj7?THAi7e0TH[Oh7k0VHTOh7Q1VHoNf7X1WHhNf7]1YHbNc7d1[H[Nc7j1[HVNa7Q2\\HoMa7U2^HkM^7Z2aHeM]7^2cHbMY7c2fH]MW7g2hHXMU7m2jHSMS7Q3lHoLP7V3oHjLn6Z3QIeLl6`3SI`Lj6d3UI\\Li6g3VIYLi6i3VIWLj6j3UIVLj6l3UITLj6n3UIRLk6o3TIQLk6Q4TIoKk6S4UIlKk6U4TIkKk6W4TIiKk6Y4TIgKk6[4TIeKl6\\4SIdKl6^4SIbKl6_500O1O100O1O100O1O2N100O1O2O0O1O2O0O1O1N3M3M3N2M3M2N3M3M3M3N2M3N2N2N101N101N2O001N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1O1O1O1O1O2O0O2N1O1O2N101N1O1O001O010O001O001O01O0N2O1O1O1OWLdFf3Z9ZLfFg3Y9YLhFg3W9YLjFg3V9XLjFi3U9WLlFi3S9WLoFY2KSOT9dNSGX2ITOS9dNUGW2JTOP9eNXGV2HUOo8eN[GU2GUOm8fN]GU2FUOl8fN_GU2FUOi8fNaGU2GUOg8eNbGX2HROf8eNaG[2IPOT9Q1mFnNR9R1PGmNo8T1QGmNn8S1SGlNm8T1SGlNm8T1TGkNl8T1UGlNk8T1VGlNi8T1WGlNi8S1XGmNh8S1YGlNg8T1YGlNg8T1ZGkNf8T1[GmNe8R1\\GmNd8S1\\GmNd8S1]GlNc8S1^GmNb8S1^GmNb8S1_GmN`8R1aGnN_8R1bGmN^8S1bGnN]8R1cGnN]8Q1eGnN\\8Q1dGoN\\8Q1dGPO[8P1fGoNZ8P1gGPOY8P1hGPOX8o0hGQOb7nNTHP2;ROa7TOnGj1b0QO`7[OiGc1g0SO_7_OeG^1m0SO^7c1cH]N^7`1dH`N\\7_1eHaN\\7]1eHcN\\7[1dHfN]7W1eHiN[7V1fHjN[7T1fHlN[7R1fHnN\\7o0dHRO]7k0eHTO]7j0dHVO^7g0cHZO]7d0dH\\O]7b0gH\\OZ7`0hHAX7=iHCX7;iHFW78jHIV75kHKU74mHLS72nHOS7MoH3R7XOTGNk1k0R7ROVG2i1l0X7hNRG;g1m0c7RO^Hn0c7PO^HP1b7oN_HQ1a7nN`HR1`7mNaHS1_7lNbHT1^7kNcHU1^7iNcHW1]7hNcHY1]7fNdHZ1g900O100000O10000000000000000001O0O100O101N1O100O101N100O1O101N2O1N2M2N2N2O1N2N2N2N2Nadg1"}, "label": "a pink cow on top of another cow"}]}
{"id": 55764, "image": "COCO_train2014_000000055764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umpire\"."}], "task": "refering", "answer_template": "The \"umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 179, 201, 202, 203, 224, 225, 226, 247, 248, 270, 271, 272, 293, 294, 295], "bbox": [0.714, 0.36000000000000004, 0.8559999999999999, 0.7466666666666666], "iscrowd": 0, "area": 8051.916799999997, "rle": {"size": [480, 640], "counts": "`\\f6>b>d0\\O000O1000000O2O000O1000000001O07I7I7J6I6J3N0O1O100O1jDlM^9U2[FUNb9k1WF_Ne9b1TFiNg9X1WFmNf9S1WFTOd9m0ZFXOb9i0\\F]Oe9?XFFNoM_9U2aFOMnMe9P2\\F3OmMf9P2YF40mMh9o1VF51mMj9n1SF63lMk9n1oE94jMn9Z4QFgKP:X4oEiKR:V4nEjKS:U4lElKU:_41O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1jM\\ENe:0^ENc:0`ENa:OcEO^:OeEO\\:OgEXO_ODj:R1kEXO]ODi:R1mEXO\\ODh:R1oEXO\\OCf:R1RFYOZOCe:R1TFYOYOCd:R1VFYOXOCc:Q1YFZOWOBa:R1[FZOVOB`:R1]FZOUOB_:R1_FYOZ:d0jEZOW:d0lEZOU:d0nEZOS:c0QF[OP:c0SF[On9c0Y2M3N4L3L4M4LTlZ1"}, "label": "umpire"}]}
{"id": 55764, "image": "COCO_train2014_000000055764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a squatting baseball umpire watching closely to judge a play\"."}], "task": "refering", "answer_template": "The \"a squatting baseball umpire watching closely to judge a play\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 179, 201, 202, 203, 224, 225, 226, 247, 248, 270, 271, 272, 293, 294, 295], "bbox": [0.714, 0.36000000000000004, 0.8559999999999999, 0.7466666666666666], "iscrowd": 0, "area": 8051.916799999997, "rle": {"size": [480, 640], "counts": "`\\f6>b>d0\\O000O1000000O2O000O1000000001O07I7I7J6I6J3N0O1O100O1jDlM^9U2[FUNb9k1WF_Ne9b1TFiNg9X1WFmNf9S1WFTOd9m0ZFXOb9i0\\F]Oe9?XFFNoM_9U2aFOMnMe9P2\\F3OmMf9P2YF40mMh9o1VF51mMj9n1SF63lMk9n1oE94jMn9Z4QFgKP:X4oEiKR:V4nEjKS:U4lElKU:_41O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1jM\\ENe:0^ENc:0`ENa:OcEO^:OeEO\\:OgEXO_ODj:R1kEXO]ODi:R1mEXO\\ODh:R1oEXO\\OCf:R1RFYOZOCe:R1TFYOYOCd:R1VFYOXOCc:Q1YFZOWOBa:R1[FZOVOB`:R1]FZOUOB_:R1_FYOZ:d0jEZOW:d0lEZOU:d0nEZOS:c0QF[OP:c0SF[On9c0Y2M3N4L3L4M4LTlZ1"}, "label": "a squatting baseball umpire watching closely to judge a play"}]}
{"id": 55764, "image": "COCO_train2014_000000055764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man swinging a bat\"."}], "task": "refering", "answer_template": "The \"a man swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 168, 169, 170, 191, 192, 193, 214, 215, 216, 235, 236, 237, 238, 239, 240, 258, 259, 262, 263], "bbox": [0.23479687500000002, 0.25675, 0.4493281250000001, 0.6756666666666666], "iscrowd": 0, "area": 10953.954100000003, "rle": {"size": [480, 640], "counts": "b_V21m>2O2N2M3YEN]64bIL]66aIK_66`IJ`66`IJ_68_IIa67_IIa67_II`68_IIa67_IIa67^IJa66`IJ`66_IKa65^ILa65_IKa65^ILb64^ILa65^ILb64]IMb64^ILb64]IMc63]IMb64]IMc63]INb61^I0a61^I0b60^I0b60]I1b60]I1c6O[I3e6MZI4e6MYI5g6KXI6h6JWI7h6JVI8j6HUI9k6GSI;l6ETI<l6DRI>m6CRI>n6BPI?Q7AnH?R7BmH>T7BjH?W7AhH;\\7FbH6d7J[H4h7LSH8n7HjG>W8CcGb0^8]O_Gf0b8ZO[Gg0f8ZOVGi0k8WORGk0o8UOnFl0S9UOiFn0X9ROeFP1\\9POaFQ1a9oN\\FS1d9nNYFS1h9nNUFT1l9lNQFV1i9kM_ES1a0T1Q:^OaEd0_:n111N2N1O2M2N3M2N3M2N3O001O0O2O1O001O1N2O001O1O0O2O10O0jNZFSMf9i2bFSM]9k2jFPMW9l2RGoLm8o2VGQMj8k2ZGUMe8i2]GWMc8g2`GYM`8d2cG\\M\\8c2fG\\M[8a2gG`MX8_2jGaMV8\\2mGdMU8X2mGgMU8U2mGlMU8P2mGPNT8X1YF`Nf18T8S1[F`Nb1=V8o0\\F_Na1b0U8k0^F_N_1f0V8f0_HZOc7;eHE_7FSI:j:000O01000O014L4LO0O2N2N101N2N2O1N6J7I6KUfT5"}, "label": "a man swinging a bat"}]}
{"id": 55764, "image": "COCO_train2014_000000055764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in blue and grey\"."}], "task": "refering", "answer_template": "The \"a baseball player in blue and grey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 168, 169, 170, 191, 192, 193, 214, 215, 216, 235, 236, 237, 238, 239, 240, 258, 259, 262, 263], "bbox": [0.23479687500000002, 0.25675, 0.4493281250000001, 0.6756666666666666], "iscrowd": 0, "area": 10953.954100000003, "rle": {"size": [480, 640], "counts": "b_V21m>2O2N2M3YEN]64bIL]66aIK_66`IJ`66`IJ_68_IIa67_IIa67_II`68_IIa67_IIa67^IJa66`IJ`66_IKa65^ILa65_IKa65^ILb64^ILa65^ILb64]IMb64^ILb64]IMc63]IMb64]IMc63]INb61^I0a61^I0b60^I0b60]I1b60]I1c6O[I3e6MZI4e6MYI5g6KXI6h6JWI7h6JVI8j6HUI9k6GSI;l6ETI<l6DRI>m6CRI>n6BPI?Q7AnH?R7BmH>T7BjH?W7AhH;\\7FbH6d7J[H4h7LSH8n7HjG>W8CcGb0^8]O_Gf0b8ZO[Gg0f8ZOVGi0k8WORGk0o8UOnFl0S9UOiFn0X9ROeFP1\\9POaFQ1a9oN\\FS1d9nNYFS1h9nNUFT1l9lNQFV1i9kM_ES1a0T1Q:^OaEd0_:n111N2N1O2M2N3M2N3M2N3O001O0O2O1O001O1N2O001O1O0O2O10O0jNZFSMf9i2bFSM]9k2jFPMW9l2RGoLm8o2VGQMj8k2ZGUMe8i2]GWMc8g2`GYM`8d2cG\\M\\8c2fG\\M[8a2gG`MX8_2jGaMV8\\2mGdMU8X2mGgMU8U2mGlMU8P2mGPNT8X1YF`Nf18T8S1[F`Nb1=V8o0\\F_Na1b0U8k0^F_N_1f0V8f0_HZOc7;eHE_7FSI:j:000O01000O014L4LO0O2N2N101N2N2O1N6J7I6KUfT5"}, "label": "a baseball player in blue and grey"}]}
{"id": 55764, "image": "COCO_train2014_000000055764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher in red\"."}], "task": "refering", "answer_template": "The \"a catcher in red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 174, 175, 194, 195, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 292], "bbox": [0.439765625, 0.4060625, 0.73109375, 0.7559583333333334], "iscrowd": 0, "area": 11717.739300000003, "rle": {"size": [480, 640], "counts": "m\\T44j>3L5L3M4K4M4M2O1O1O00001O00001O000011N1O101O001N101N4M3L4M2M10O0100O010O10O0100O0000000001O000000001O00000000001O000000001O1O1O1O001O0000001O000000000000O1000lB^Oj;a0VDEe;;[DIa;7^DN^;2bD2Z;NeD4Z;LeD6Z;KeD73WOm9b0oE91WOP:a0nE9NXOU:?lE;LXO`90RF>b0<HXOZ99ZF5c0;FYOY9=[F0f0;B[O]9;YF1g0:@\\O_9;XF0i0;\\O\\Oc9[1QGZO[OZOe9]1oFZOZOZOf9]1PGZOXOZOh9]1oF[OWOXOj9^1nF[OVOXOl9]1nF\\OTOXOm9^1nFOS91lF0T91kFOU92jFNU93kFMU96hFJX9:dFE\\9`0`F@`9j0VFVOj9k0UFkMNa0m9e1TFkM0=l9j1SFjM2:k9g2WFWMi9j2XFTMh9l2ZFQMf9Q3[FmLe9T3\\FjLd9V3^FhLc9X3^FeLc9[3_FbLc9^3_F^Lb9b3`F[La9e3`FYLb9f3`FWLa9i3`FULb9j3`FTL`9l3j001O0kLQEY2Q;eMoD[2T;bMlD^2X;^MhDb2];YMdDf2b;TM^Dl2c;SM]Dm2d;RM\\Dn2l;O0000001O00000O2O001O1N101O1O1N101O1001O1N3N1SDYM^;h2^D[Mb;g2YD\\Mf;P31O1O0cMVDc1k;[NWDc1k;[NWDc1k;[NWDc1k;[NWDc1j;\\NXDb1j;[NYDb1j;\\NXDb1j;TN`Dj1^<M2M4eNoB`0T=[OPCc0g=O1O2N1_OhA9Y>ElA7a>M4L^Y`2"}, "label": "a catcher in red"}]}
{"id": 55764, "image": "COCO_train2014_000000055764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red cather with the number 36 is standing behind a batter and in front of a umpire\"."}], "task": "refering", "answer_template": "The \"a red cather with the number 36 is standing behind a batter and in front of a umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 174, 175, 194, 195, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 292], "bbox": [0.439765625, 0.4060625, 0.73109375, 0.7559583333333334], "iscrowd": 0, "area": 11717.739300000003, "rle": {"size": [480, 640], "counts": "m\\T44j>3L5L3M4K4M4M2O1O1O00001O00001O000011N1O101O001N101N4M3L4M2M10O0100O010O10O0100O0000000001O000000001O00000000001O000000001O1O1O1O001O0000001O000000000000O1000lB^Oj;a0VDEe;;[DIa;7^DN^;2bD2Z;NeD4Z;LeD6Z;KeD73WOm9b0oE91WOP:a0nE9NXOU:?lE;LXO`90RF>b0<HXOZ99ZF5c0;FYOY9=[F0f0;B[O]9;YF1g0:@\\O_9;XF0i0;\\O\\Oc9[1QGZO[OZOe9]1oFZOZOZOf9]1PGZOXOZOh9]1oF[OWOXOj9^1nF[OVOXOl9]1nF\\OTOXOm9^1nFOS91lF0T91kFOU92jFNU93kFMU96hFJX9:dFE\\9`0`F@`9j0VFVOj9k0UFkMNa0m9e1TFkM0=l9j1SFjM2:k9g2WFWMi9j2XFTMh9l2ZFQMf9Q3[FmLe9T3\\FjLd9V3^FhLc9X3^FeLc9[3_FbLc9^3_F^Lb9b3`F[La9e3`FYLb9f3`FWLa9i3`FULb9j3`FTL`9l3j001O0kLQEY2Q;eMoD[2T;bMlD^2X;^MhDb2];YMdDf2b;TM^Dl2c;SM]Dm2d;RM\\Dn2l;O0000001O00000O2O001O1N101O1O1N101O1001O1N3N1SDYM^;h2^D[Mb;g2YD\\Mf;P31O1O0cMVDc1k;[NWDc1k;[NWDc1k;[NWDc1k;[NWDc1j;\\NXDb1j;[NYDb1j;\\NXDb1j;TN`Dj1^<M2M4eNoB`0T=[OPCc0g=O1O2N1_OhA9Y>ElA7a>M4L^Y`2"}, "label": "a red cather with the number 36 is standing behind a batter and in front of a umpire"}]}
{"id": 195027, "image": "COCO_train2014_000000195027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair which a boy is sitting in\"."}], "task": "refering", "answer_template": "The \"a chair which a boy is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 161, 162, 163, 176, 177, 178, 191, 192, 193, 194, 207, 208, 209, 222, 223, 224, 236, 237, 238, 251, 252, 253, 266, 267, 268, 281, 282, 295, 296, 297, 311], "bbox": [0.7157276995305164, 0.388765625, 0.9554225352112676, 0.9033750000000002], "iscrowd": 0, "area": 17106.743250000003, "rle": {"size": [640, 426], "counts": "WUo52lc03M3M4L3M3M4L3M3M3dHZOYKk0X4E`Ka0]4D[Kb0c4CTKd0i4@oJf0o4\\OkJj0S5YOfJn0W5UObJQ1\\5QO^JU1`5nNZJW1d5lNUJY1j5iNPJZ1P6jNiIY1X6jNaIZ1^6iN[I[1_6oNZIT1d6ROUIQ1i6TOQIo0m6XOkHk0W7WOcHk0`7VOZHl0i7UOQHn0P8UOiGm0Z8XO\\Gj0g8_OjFd0X9B^F`0e9FPF<R:g43M3N1N3M3M2N3M3M2N3N2M2N3M3M2N3M2N3N2M2N3M3M2N3M3M2O2M3M2N3M3M2N3M2O2M3M4L5K4L4L4L5L3L4L4L4L5K4L4L4M4K4E;_Oh0B?C<D6K5J5K6J6F:E;D<Ei`;"}, "label": "a chair which a boy is sitting in"}]}
{"id": 195027, "image": "COCO_train2014_000000195027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the little boy is sitting on\"."}], "task": "refering", "answer_template": "The \"the chair the little boy is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 161, 162, 163, 176, 177, 178, 191, 192, 193, 194, 207, 208, 209, 222, 223, 224, 236, 237, 238, 251, 252, 253, 266, 267, 268, 281, 282, 295, 296, 297, 311], "bbox": [0.7157276995305164, 0.388765625, 0.9554225352112676, 0.9033750000000002], "iscrowd": 0, "area": 17106.743250000003, "rle": {"size": [640, 426], "counts": "WUo52lc03M3M4L3M3M4L3M3M3dHZOYKk0X4E`Ka0]4D[Kb0c4CTKd0i4@oJf0o4\\OkJj0S5YOfJn0W5UObJQ1\\5QO^JU1`5nNZJW1d5lNUJY1j5iNPJZ1P6jNiIY1X6jNaIZ1^6iN[I[1_6oNZIT1d6ROUIQ1i6TOQIo0m6XOkHk0W7WOcHk0`7VOZHl0i7UOQHn0P8UOiGm0Z8XO\\Gj0g8_OjFd0X9B^F`0e9FPF<R:g43M3N1N3M3M2N3M3M2N3N2M2N3M3M2N3M2N3N2M2N3M3M2N3M3M2O2M3M2N3M3M2N3M2O2M3M4L5K4L4L4L5L3L4L4L4L5K4L4L4M4K4E;_Oh0B?C<D6K5J5K6J6F:E;D<Ei`;"}, "label": "the chair the little boy is sitting on"}]}
{"id": 195027, "image": "COCO_train2014_000000195027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a blue and white striped shirt\"."}], "task": "refering", "answer_template": "The \"a boy wearing a blue and white striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 248, 249, 250, 251, 264, 265, 266, 279, 280, 295], "bbox": [0.4391549295774648, 0.24176562499999998, 0.8386619718309859, 0.848359375], "iscrowd": 0, "area": 39380.4165, "rle": {"size": [640, 426], "counts": "WWe34ic06I7I7J5100O2N1NN3L3M4L3M4L3N3L3M3M4L3L5K4L5K4dMiMjB\\2m<QNiBT2o<YNhBl1o<bNgBb1R=mNbBX1Z=[OQBj0l=g2L5L3M4L3L5L3M4L3M4L3N3L3N3M2M3N3L3N3M2M4M2M4M2N3M2N3M2O1N3M2N3M2N3M2N3M2N3N1N3N1O2M2O1N3N1N3N1N3N1N3N1N3N1N3N1N2O2N1N3N1N3N1N3N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1OO1TOZH]Eg7a:`HYEa7d:hHVEX7h:oHRER7l:R1N2XImF`2V9`MkF]2W9cMiF[2Z9cMhFUMNl3\\9nNhFgL;W4Q9POoGm0S8SOnGj0U8TOmGh0V8WOlGf0W8YOjGc0Y8\\OiGa0Z8]OhG?\\8@fG;]8DeG9^8EdG7_8IbG5`8IbG4`8KbG3`8KcG2_8MbG1_8YOXHd0k7[OWH?m7AUH9P8FRH4R8LPHNV80kGlMQMi0e;Y1\\GfMXMd0h;d1oGTNX8j1kGnM]8o1eGkMa8S2h4N2N2N2N2N2M3L4L4L4L5L3Lb0^OUmZ1"}, "label": "a boy wearing a blue and white striped shirt"}]}
{"id": 195027, "image": "COCO_train2014_000000195027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy sitting at table smiling\"."}], "task": "refering", "answer_template": "The \"boy sitting at table smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 248, 249, 250, 251, 264, 265, 266, 279, 280, 295], "bbox": [0.4391549295774648, 0.24176562499999998, 0.8386619718309859, 0.848359375], "iscrowd": 0, "area": 39380.4165, "rle": {"size": [640, 426], "counts": "WWe34ic06I7I7J5100O2N1NN3L3M4L3M4L3N3L3M3M4L3L5K4L5K4dMiMjB\\2m<QNiBT2o<YNhBl1o<bNgBb1R=mNbBX1Z=[OQBj0l=g2L5L3M4L3L5L3M4L3M4L3N3L3N3M2M3N3L3N3M2M4M2M4M2N3M2N3M2O1N3M2N3M2N3M2N3M2N3N1N3N1O2M2O1N3N1N3N1N3N1N3N1N3N1N3N1N2O2N1N3N1N3N1N3N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1OO1TOZH]Eg7a:`HYEa7d:hHVEX7h:oHRER7l:R1N2XImF`2V9`MkF]2W9cMiF[2Z9cMhFUMNl3\\9nNhFgL;W4Q9POoGm0S8SOnGj0U8TOmGh0V8WOlGf0W8YOjGc0Y8\\OiGa0Z8]OhG?\\8@fG;]8DeG9^8EdG7_8IbG5`8IbG4`8KbG3`8KcG2_8MbG1_8YOXHd0k7[OWH?m7AUH9P8FRH4R8LPHNV80kGlMQMi0e;Y1\\GfMXMd0h;d1oGTNX8j1kGnM]8o1eGkMa8S2h4N2N2N2N2N2M3L4L4L4L5L3Lb0^OUmZ1"}, "label": "boy sitting at table smiling"}]}
{"id": 96728, "image": "COCO_train2014_000000096728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty chair in front of the laptop\"."}], "task": "refering", "answer_template": "The \"the empty chair in front of the laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436], "bbox": [0.521859375, 0.8580478821362799, 0.999671875, 1.0], "iscrowd": 0, "area": 21460.8794, "rle": {"size": [543, 640], "counts": "bba5d0[`0W1hN5L0O10000O100O100O1000000000000001O000000000000000000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00"}, "label": "the empty chair in front of the laptop"}]}
{"id": 96728, "image": "COCO_train2014_000000096728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown and black chair sitting in front of the woman\"."}], "task": "refering", "answer_template": "The \"a brown and black chair sitting in front of the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436], "bbox": [0.521859375, 0.8580478821362799, 0.999671875, 1.0], "iscrowd": 0, "area": 21460.8794, "rle": {"size": [543, 640], "counts": "bba5d0[`0W1hN5L0O10000O100O100O1000000000000001O000000000000000000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00"}, "label": "a brown and black chair sitting in front of the woman"}]}
{"id": 96728, "image": "COCO_train2014_000000096728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red color couch in the right side of the image\"."}], "task": "refering", "answer_template": "The \"red color couch in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 251, 252, 266, 274, 275, 297, 298, 320, 321, 343, 344, 367], "bbox": [0.4671875, 0.2649355432780847, 1.0, 0.7948066298342542], "iscrowd": 0, "area": 45371.92585, "rle": {"size": [543, 640], "counts": "ehn42l`04L4M3L4L4L2O0O2N2N2O0O2N2O1N1O2O1N101N2N2O0O2N2O0O2O1N2N101N2N101N2O1N3M4M2M3M3N3L3N2M4L3N2M3M4M2M3N3L3M3N2M4L3N2M4M2M3M3N3L3M3NN1N3L3N2M4M2M4M2M4M3L3N3L3N3L4M210000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O10000O100O10001N10000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O100000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000001O00000000000000000O10000000000000000000000000000000000000000000000001O000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O0000000000bGTLi3l3RLYLn3h3kK^LU4b3fKcLZ4]3`KiL`4W3ZKPMe4P3VKUMj4k2PK[MP5e2kJ`MU5`2eJfM[5Z2`JkM`5U2ZJQNf5o1UJVNk5j1oI\\NQ6d1jIaNV6_1dIhN[6X1`ImN`6S1ZISOf6m0UIXOk6h0oH^OQ7b0jHCV7=dHI\\77^HOb71YH4g7LSH:m7FnG`0Q8@iGf0W8Q40000000000000010O01O00001O001O000010O01O00001O001O004L4MiJ"}, "label": "red color couch in the right side of the image"}]}
{"id": 96728, "image": "COCO_train2014_000000096728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"high back of a red chair\"."}], "task": "refering", "answer_template": "The \"high back of a red chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 251, 252, 266, 274, 275, 297, 298, 320, 321, 343, 344, 367], "bbox": [0.4671875, 0.2649355432780847, 1.0, 0.7948066298342542], "iscrowd": 0, "area": 45371.92585, "rle": {"size": [543, 640], "counts": "ehn42l`04L4M3L4L4L2O0O2N2N2O0O2N2O1N1O2O1N101N2N2O0O2N2O0O2O1N2N101N2N101N2O1N3M4M2M3M3N3L3N2M4L3N2M3M4M2M3N3L3M3N2M4L3N2M4M2M3M3N3L3M3NN1N3L3N2M4M2M4M2M4M3L3N3L3N3L4M210000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O10000O100O10001N10000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O100000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000001O00000000000000000O10000000000000000000000000000000000000000000000001O000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O0000000000bGTLi3l3RLYLn3h3kK^LU4b3fKcLZ4]3`KiL`4W3ZKPMe4P3VKUMj4k2PK[MP5e2kJ`MU5`2eJfM[5Z2`JkM`5U2ZJQNf5o1UJVNk5j1oI\\NQ6d1jIaNV6_1dIhN[6X1`ImN`6S1ZISOf6m0UIXOk6h0oH^OQ7b0jHCV7=dHI\\77^HOb71YH4g7LSH:m7FnG`0Q8@iGf0W8Q40000000000000010O01O00001O001O000010O01O00001O001O004L4MiJ"}, "label": "high back of a red chair"}]}
{"id": 96728, "image": "COCO_train2014_000000096728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady working on laptop\"."}], "task": "refering", "answer_template": "The \"lady working on laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357], "bbox": [0.071234375, 0.056795580110497235, 0.5781875, 0.8296316758747697], "iscrowd": 0, "area": 80102.9118, "rle": {"size": [543, 640], "counts": "Xjh04e`07J5K5J6K5K5K5J6K5K5K5J6K5K5J6I7I7J6J7H7J6J6I7J6J6I7J6J6I7K5O100O1O100O2O0O1O100O100O1O100O100O100O1O100O2O0O1O100O100O100O1O100O100O1O100O2O0O100O1O100O100O1O100O100O100O1O101N100O1O100O100O1O100O100O100O1O101N100O1O100O100O100O1O100O100O100O2O[M]GnL^5[NfLf4RNkLV5oNZLS4gNjLm4CmKb3]OfLd48aKP31dL\\4m0SK^2g0aLU4`1fJl1\\1`Lm3S2XJ\\1Q2]Lf3m6`LoH_3P7gLlHX3R7oLjHP3U7WMfHh2Z7]MbHa2`7cM\\H\\2f7]31O1O1O1O1O2N1N2O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1N2O1O2N1O1O1O1O1O2M2O1O1O1O101N10000O101O0_KcF_1^9\\NgFd1Y9VNmFj1S9PNSGo1n8kMXGU2i8eM[G\\2e8^MaGa2`8YMfGg2Z8TMkGk2V8oLPHQ3P8iLVHW3k7_L]Ha3d7RLiHn3W7eKVI[4g9O1000O010000O10000O01000O10000O01000O100O10O10O1000000001N101O00001O0O101O00001O0O101O00001O000O2O001O0eI]K:c4D`K;a4BbK<`4AcK:a4CbK8d4E_K5g4H\\K2j4KXK0n4MUKMP51SKIS54PKFV57lJDZ58jJDY5:jJEW58kJHV55mJJT53oJLQ52QKNP5OSK0n4LVK3k4JXK5h4IZK;c4B`Kg1W3VNlLU2i2hMYMY2g2dM\\M[2f2`M^M_2c2^M`Ma2a2\\MaMd2a2XMbMh2^2UMeMj2\\2SMfMm2\\2oLgMP3Z2lLjMS3W2jLkMW3V2eLmMY3f2SL]Mj3]4\\JfKa5m7N3N2M2N3N2M3M2N3N2M2N3N3L4L4L4M3J7H7J6I7J6J6I7J6I7J7I6I7J6J6I7J6I7J6J7H7Jek^4"}, "label": "lady working on laptop"}]}
{"id": 96728, "image": "COCO_train2014_000000096728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman behind laptop\"."}], "task": "refering", "answer_template": "The \"woman behind laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 301, 302, 303, 304, 305, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357], "bbox": [0.071234375, 0.056795580110497235, 0.5781875, 0.8296316758747697], "iscrowd": 0, "area": 80102.9118, "rle": {"size": [543, 640], "counts": "Xjh04e`07J5K5J6K5K5K5J6K5K5K5J6K5K5J6I7I7J6J7H7J6J6I7J6J6I7J6J6I7K5O100O1O100O2O0O1O100O100O1O100O100O100O1O100O2O0O1O100O100O100O1O100O100O1O100O2O0O100O1O100O100O1O100O100O100O1O101N100O1O100O100O1O100O100O100O1O101N100O1O100O100O100O1O100O100O100O2O[M]GnL^5[NfLf4RNkLV5oNZLS4gNjLm4CmKb3]OfLd48aKP31dL\\4m0SK^2g0aLU4`1fJl1\\1`Lm3S2XJ\\1Q2]Lf3m6`LoH_3P7gLlHX3R7oLjHP3U7WMfHh2Z7]MbHa2`7cM\\H\\2f7]31O1O1O1O1O2N1N2O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1N2O1O2N1O1O1O1O1O2M2O1O1O1O101N10000O101O0_KcF_1^9\\NgFd1Y9VNmFj1S9PNSGo1n8kMXGU2i8eM[G\\2e8^MaGa2`8YMfGg2Z8TMkGk2V8oLPHQ3P8iLVHW3k7_L]Ha3d7RLiHn3W7eKVI[4g9O1000O010000O10000O01000O10000O01000O100O10O10O1000000001N101O00001O0O101O00001O0O101O00001O000O2O001O0eI]K:c4D`K;a4BbK<`4AcK:a4CbK8d4E_K5g4H\\K2j4KXK0n4MUKMP51SKIS54PKFV57lJDZ58jJDY5:jJEW58kJHV55mJJT53oJLQ52QKNP5OSK0n4LVK3k4JXK5h4IZK;c4B`Kg1W3VNlLU2i2hMYMY2g2dM\\M[2f2`M^M_2c2^M`Ma2a2\\MaMd2a2XMbMh2^2UMeMj2\\2SMfMm2\\2oLgMP3Z2lLjMS3W2jLkMW3V2eLmMY3f2SL]Mj3]4\\JfKa5m7N3N2M2N3N2M3M2N3N2M2N3N3L4L4L4M3J7H7J6I7J6J6I7J6I7J7I6I7J6J6I7J6I7J6J7H7Jek^4"}, "label": "woman behind laptop"}]}
{"id": 424408, "image": "COCO_train2014_000000424408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with back to others\"."}], "task": "refering", "answer_template": "The \"person with back to others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 98, 99, 100, 112, 113, 127, 128, 130, 144, 145, 146, 160, 161, 162], "bbox": [0.49453051643192486, 0.246015625, 0.8410328638497652, 0.4809374999999999], "iscrowd": 0, "area": 4788.410500000001, "rle": {"size": [640, 426], "counts": "jST46ac08J6O2O0O1O100O100O100O1O2O0O100O100O100O2N101N10001N101O0O2O001O0O2O001O001O00N2L4K5L4K5N2O002M6K4LLEj\\O6Wc0Kk\\O0Wc01;2M2O1O2N1O1N3N1O100010On^OBb>>]AFa>:]AK`>4_A0^>0aA4]>KcA8[>HcA<Z>DeA`0Y>_OfAe0W>\\OgAh0W>WOhAm0V>ROiAR1T>nNkAV1T>iNkAZ1T>dNkA_1V>^NiAc1Y>\\NeAf1\\>XNcAi1_>UN`Am1X`01O2O0O2N2N1O2N11J9L3M3L4M3M4L3N2M3M4M2M4L3M3NO01O0001O01O0001O01O01O01O0001O01O0001O01O0001O01O00010N100O1O2O0O101N1O2O0O105JlVZ1"}, "label": "person with back to others"}]}
{"id": 424408, "image": "COCO_train2014_000000424408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue and yellow plaid\"."}], "task": "refering", "answer_template": "The \"a man in blue and yellow plaid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 98, 99, 100, 112, 113, 127, 128, 130, 144, 145, 146, 160, 161, 162], "bbox": [0.49453051643192486, 0.246015625, 0.8410328638497652, 0.4809374999999999], "iscrowd": 0, "area": 4788.410500000001, "rle": {"size": [640, 426], "counts": "jST46ac08J6O2O0O1O100O100O100O1O2O0O100O100O100O2N101N10001N101O0O2O001O0O2O001O001O00N2L4K5L4K5N2O002M6K4LLEj\\O6Wc0Kk\\O0Wc01;2M2O1O2N1O1N3N1O100010On^OBb>>]AFa>:]AK`>4_A0^>0aA4]>KcA8[>HcA<Z>DeA`0Y>_OfAe0W>\\OgAh0W>WOhAm0V>ROiAR1T>nNkAV1T>iNkAZ1T>dNkA_1V>^NiAc1Y>\\NeAf1\\>XNcAi1_>UN`Am1X`01O2O0O2N2N1O2N11J9L3M3L4M3M4L3N2M3M4M2M4L3M3NO01O0001O01O0001O01O01O01O0001O01O0001O01O0001O01O00010N100O1O2O0O101N1O2O0O105JlVZ1"}, "label": "a man in blue and yellow plaid"}]}
{"id": 424408, "image": "COCO_train2014_000000424408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a plaid shirt and black beret\"."}], "task": "refering", "answer_template": "The \"a man in a plaid shirt and black beret\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 113, 114, 115, 116, 128, 129, 130, 142, 143, 144, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 201, 202, 203, 216, 217], "bbox": [0.4136619718309859, 0.2714375, 0.8159624413145541, 0.64528125], "iscrowd": 0, "area": 17592.901700000002, "rle": {"size": [640, 426], "counts": "g[^31lc05K5L4K5L4K5L4K5J6K5K5J6J6J6J6J6J6K5J4L4L3M3M3M3N3M2M3N2N201OCW@dLj?Z3\\@aLZ?Oc@a3j?cLR@^3m?eLQ@[3n?<O1O1O1N101O1O1O1O1O1O1O1O1O1O1O001O1O1O100O1O00100O0010O01O1O001O0O1O100L3J7M3O1N2N2O1N2N1O2O000100O11O1O1O001O1O1O1O10O100eLiAm0W>SOlAj0U>VOlAh0T>XOPBd0P>\\OUB>l=BXB:h=G\\B4d=L`B0a=OcBM]=3gBHZ=7kBEU=:oBCQ=<SCAn<=VC@j<?XC@h<>ZCAg<>[CAf<=\\CAf<>ZCAg<>[C@g<>ZC@i<?XC_Oi<`0ZC]Oh<b0YC[Oj<c0YCZOh<f0YCXOi<f0YCXOi<;cCE_<8bCH_<6bCJ`<3aCM`<0bC9V<ElC9W<DjC<W<BjC>W<@jC`0X<^OhCa0Z<^OfCb0[<]OeCc0]<ZOdCf0]<YOcCf0_<YOaCg0a<WO_Ci0b<VO^Cj0c<TO_Ck0c<SO]Cm0d<RO\\Cn0l?N101O0O1O2N101N2N1O2O1N2N101O10O01O1O0N3L4L31000O1O0Ng]`1"}, "label": "a man in a plaid shirt and black beret"}]}
{"id": 424408, "image": "COCO_train2014_000000424408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a plaid shirt with a black cap holding a pan\"."}], "task": "refering", "answer_template": "The \"a man in a plaid shirt with a black cap holding a pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 113, 114, 115, 116, 128, 129, 130, 142, 143, 144, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 201, 202, 203, 216, 217], "bbox": [0.4136619718309859, 0.2714375, 0.8159624413145541, 0.64528125], "iscrowd": 0, "area": 17592.901700000002, "rle": {"size": [640, 426], "counts": "g[^31lc05K5L4K5L4K5L4K5J6K5K5J6J6J6J6J6J6K5J4L4L3M3M3M3N3M2M3N2N201OCW@dLj?Z3\\@aLZ?Oc@a3j?cLR@^3m?eLQ@[3n?<O1O1O1N101O1O1O1O1O1O1O1O1O1O1O001O1O1O100O1O00100O0010O01O1O001O0O1O100L3J7M3O1N2N2O1N2N1O2O000100O11O1O1O001O1O1O1O10O100eLiAm0W>SOlAj0U>VOlAh0T>XOPBd0P>\\OUB>l=BXB:h=G\\B4d=L`B0a=OcBM]=3gBHZ=7kBEU=:oBCQ=<SCAn<=VC@j<?XC@h<>ZCAg<>[CAf<=\\CAf<>ZCAg<>[C@g<>ZC@i<?XC_Oi<`0ZC]Oh<b0YC[Oj<c0YCZOh<f0YCXOi<f0YCXOi<;cCE_<8bCH_<6bCJ`<3aCM`<0bC9V<ElC9W<DjC<W<BjC>W<@jC`0X<^OhCa0Z<^OfCb0[<]OeCc0]<ZOdCf0]<YOcCf0_<YOaCg0a<WO_Ci0b<VO^Cj0c<TO_Ck0c<SO]Cm0d<RO\\Cn0l?N101O0O1O2N101N2N1O2O1N2N101O10O01O1O0N3L4L31000O1O0Ng]`1"}, "label": "a man in a plaid shirt with a black cap holding a pan"}]}
{"id": 399835, "image": "COCO_train2014_000000399835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult zebra behind the other zebras facing right\"."}], "task": "refering", "answer_template": "The \"the adult zebra behind the other zebras facing right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 62, 63, 64, 73, 74, 75, 76, 77, 80, 81, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 119, 120, 121, 122, 123, 128, 129, 130, 131, 132, 133, 134, 142, 143, 144, 145, 146, 151, 152, 153, 154, 156, 157, 158, 174, 175, 179, 180, 181, 197, 198, 220, 221, 244, 267, 268], "bbox": [0.18740625, 0.15056206088992977, 0.88009375, 0.7865105386416863], "iscrowd": 0, "area": 34054.836500000005, "rle": {"size": [427, 640], "counts": "nUb12W=4L5K5K4L5K4M4K5K4L5K4L0J6J6K52N1O1O2M2N2N3L3N2N3N1O13N4K5K4L5L4K5Ka0_O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O100O1000000000000O1000000000000O10000000000O10000000001O0O100000001O0O10005K7IN2N2N1O2M3N2N2N2N2M2O2N2010O000010O00010O01O00iLPFS3Q:lLSFQ3l9oLWFn2i9SMYFk2f9UM]Fh2d9WM_Ff2a9ZMcFc2\\9]MgF`2Y9`MjF^2U9cMmFZ2S9fMPGW2P9iMTGU1j9kNYF0i:0Y1001O000001O01O00000000001O000000000O2O000O10000O1000O010O10O0100O00100O010O10O01001N1100O1O100O100O0010O0100O010O001000O01000O100O01000O10O010000O10O010000O0010O01O00001O00010O001O000010O0001O00001O01O01O00001O00010O00001O00010O2N1O1O1O2N100O1O2N1O1O2N101N1O1O2N1\\F@[6a0kHe0i6\\OeHb1m6nNUHm1^7W2I5K4L5K3N1N2N3M2N3M2O1N3M2M4M2N1OYMPJTNn5g1bJoM^5k1QKmMn4m1`KkM_4P2oKiMP4Q2^LgMb3W2gLcMX3\\2QM\\Mn2d2VMYMj2g2WMYMh2f2[MXMd2h2^MXMa2h2`MWM`2h2cMWM\\2h2fMWMY2j2hMVMW2i2kMVMU2i2nMVMP2k2QNTMo1k2SNTMm1k2VNSMj1l2XNRMh1o2ZNmLh1R3ZNlLg1S3[NkLe1V3]NiLb1V3`NjL_1U3\\4N101N2O0O2O1N2O0O2O1O00100000O01000O1000O10O10000O0100000O10O10O1000O10O10000O0100000O10O10O1000O10O10000O0100001O00100O001O100O001O10O01O1O10O01O2N100O2N1O101N1O1O2O0O1O2N100O2N1O101N1O2N2O1N2N2N2O1N1O2N2O1N2N1O02N1N2O1O1N2O2N1N2O1O1N2O2N1N2SMeFh1\\9RNnFi1S9PNXGj1j8TN[Gh1g8TN`Gg1a8WNeGa1_8]NfG]1]8`NhGZ1Z:J5J6K6Iimo0"}, "label": "the adult zebra behind the other zebras facing right"}]}
{"id": 399835, "image": "COCO_train2014_000000399835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"biggest zebra among the three zebras\"."}], "task": "refering", "answer_template": "The \"biggest zebra among the three zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 62, 63, 64, 73, 74, 75, 76, 77, 80, 81, 82, 83, 84, 85, 86, 87, 88, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 119, 120, 121, 122, 123, 128, 129, 130, 131, 132, 133, 134, 142, 143, 144, 145, 146, 151, 152, 153, 154, 156, 157, 158, 174, 175, 179, 180, 181, 197, 198, 220, 221, 244, 267, 268], "bbox": [0.18740625, 0.15056206088992977, 0.88009375, 0.7865105386416863], "iscrowd": 0, "area": 34054.836500000005, "rle": {"size": [427, 640], "counts": "nUb12W=4L5K5K4L5K4M4K5K4L5K4L0J6J6K52N1O1O2M2N2N3L3N2N3N1O13N4K5K4L5L4K5Ka0_O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O100O1000000000000O1000000000000O10000000000O10000000001O0O100000001O0O10005K7IN2N2N1O2M3N2N2N2N2M2O2N2010O000010O00010O01O00iLPFS3Q:lLSFQ3l9oLWFn2i9SMYFk2f9UM]Fh2d9WM_Ff2a9ZMcFc2\\9]MgF`2Y9`MjF^2U9cMmFZ2S9fMPGW2P9iMTGU1j9kNYF0i:0Y1001O000001O01O00000000001O000000000O2O000O10000O1000O010O10O0100O00100O010O10O01001N1100O1O100O100O0010O0100O010O001000O01000O100O01000O10O010000O10O010000O0010O01O00001O00010O001O000010O0001O00001O01O01O00001O00010O00001O00010O2N1O1O1O2N100O1O2N1O1O2N101N1O1O2N1\\F@[6a0kHe0i6\\OeHb1m6nNUHm1^7W2I5K4L5K3N1N2N3M2N3M2O1N3M2M4M2N1OYMPJTNn5g1bJoM^5k1QKmMn4m1`KkM_4P2oKiMP4Q2^LgMb3W2gLcMX3\\2QM\\Mn2d2VMYMj2g2WMYMh2f2[MXMd2h2^MXMa2h2`MWM`2h2cMWM\\2h2fMWMY2j2hMVMW2i2kMVMU2i2nMVMP2k2QNTMo1k2SNTMm1k2VNSMj1l2XNRMh1o2ZNmLh1R3ZNlLg1S3[NkLe1V3]NiLb1V3`NjL_1U3\\4N101N2O0O2O1N2O0O2O1O00100000O01000O1000O10O10000O0100000O10O10O1000O10O10000O0100000O10O10O1000O10O10000O0100001O00100O001O100O001O10O01O1O10O01O2N100O2N1O101N1O1O2O0O1O2N100O2N1O101N1O2N2O1N2N2N2O1N1O2N2O1N2N1O02N1N2O1O1N2O2N1N2O1O1N2O2N1N2SMeFh1\\9RNnFi1S9PNXGj1j8TN[Gh1g8TN`Gg1a8WNeGa1_8]NfG]1]8`NhGZ1Z:J5J6K6Iimo0"}, "label": "biggest zebra among the three zebras"}]}
{"id": 399835, "image": "COCO_train2014_000000399835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra facing towards the camera that ' s not a baby\"."}], "task": "refering", "answer_template": "The \"the zebra facing towards the camera that ' s not a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334], "bbox": [0.36965625, 0.1475175644028103, 0.6055781250000001, 0.9515690866510539], "iscrowd": 0, "area": 31619.156150000006, "rle": {"size": [427, 640], "counts": "llR37R=8H7I8[FSO`6m0[I[Ob6f0WICf6>SIKj66PI3l6MnH<P7DbGTOb0_1j7^O`GE1U1]8WO_GQ2`8oM^GU2`8lM]GW2b8jMZGY2g8U10O100O101N100O100O101N04M4Kj0WO4L4hL\\J`Ni5V1]JjNg5n0^JPOg5i0]JVOg5b0]J^Oi59[JEl51YJNl5JWJ5P6AUJ>_6kNdIU1Y9O100O10000O100O1000L3L5K4M4K7I6G:^Oa0_Ob0YNg1F9G:G8K50O1O001O001O1O0`KZJ[1g5bN\\J\\1e5aN`J]1a5_NbJb1^5[NeJd1[5YNhJg1Y5UNjJk1^5iMgJV2`5`MfJ]2b5XMeJd2d5PMbJm2n5`LYJ\\3n7M4L3M^FeLW9f34M1N0010O0010O0001O001N100_M[M\\Jf2d5\\MYJd2l5ZMQJg2R6WMkIj2Z6TMcIl2a6SM\\In2c7ULZHk3j7TLRHn3P8RLmGn3T8TLiGl3W8WLfGi3Z8YLcGi3\\8ZLaGf3_8]L^Gc3b8`L[G`3e8c0kMmKRKS4b4]LZKc3a4eL\\K[3_4nL\\KS3`4TM]Kl2_4\\M]Kd2`4bM]K^2`4hM]KX2`4mM]KT2`4RN\\KQ2^4UN_Km1]4YN^Kk1]4ZN]Kl1^4c3L5K4K6J6J5L5L3M4L4M2M3M2N3N1N3M3M2O2MZ1fN[mX3"}, "label": "the zebra facing towards the camera that ' s not a baby"}]}
{"id": 399835, "image": "COCO_train2014_000000399835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"medium size zebra facing forward\"."}], "task": "refering", "answer_template": "The \"medium size zebra facing forward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334], "bbox": [0.36965625, 0.1475175644028103, 0.6055781250000001, 0.9515690866510539], "iscrowd": 0, "area": 31619.156150000006, "rle": {"size": [427, 640], "counts": "llR37R=8H7I8[FSO`6m0[I[Ob6f0WICf6>SIKj66PI3l6MnH<P7DbGTOb0_1j7^O`GE1U1]8WO_GQ2`8oM^GU2`8lM]GW2b8jMZGY2g8U10O100O101N100O100O101N04M4Kj0WO4L4hL\\J`Ni5V1]JjNg5n0^JPOg5i0]JVOg5b0]J^Oi59[JEl51YJNl5JWJ5P6AUJ>_6kNdIU1Y9O100O10000O100O1000L3L5K4M4K7I6G:^Oa0_Ob0YNg1F9G:G8K50O1O001O001O1O0`KZJ[1g5bN\\J\\1e5aN`J]1a5_NbJb1^5[NeJd1[5YNhJg1Y5UNjJk1^5iMgJV2`5`MfJ]2b5XMeJd2d5PMbJm2n5`LYJ\\3n7M4L3M^FeLW9f34M1N0010O0010O0001O001N100_M[M\\Jf2d5\\MYJd2l5ZMQJg2R6WMkIj2Z6TMcIl2a6SM\\In2c7ULZHk3j7TLRHn3P8RLmGn3T8TLiGl3W8WLfGi3Z8YLcGi3\\8ZLaGf3_8]L^Gc3b8`L[G`3e8c0kMmKRKS4b4]LZKc3a4eL\\K[3_4nL\\KS3`4TM]Kl2_4\\M]Kd2`4bM]K^2`4hM]KX2`4mM]KT2`4RN\\KQ2^4UN_Km1]4YN^Kk1]4ZN]Kl1^4c3L5K4K6J6J5L5L3M4L4M2M3M2N3N1N3M3M2O2MZ1fN[mX3"}, "label": "medium size zebra facing forward"}]}
{"id": 399835, "image": "COCO_train2014_000000399835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby zebra\"."}], "task": "refering", "answer_template": "The \"a baby zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307], "bbox": [0.22715625, 0.41426229508196716, 0.41498437499999996, 0.9094613583138172], "iscrowd": 0, "area": 14695.8167, "rle": {"size": [427, 640], "counts": "Zfl11U=<E:E:F6M2O2N2N2N2M2O3M5K5VOSNmEQ2P:RNjEU2Q:nMjEX2S:d0L3M4L3L3N2N2N2cHZLQ5h3kJ[LT5g3hJ\\LW5f3fJ\\LY5f3cJ]L\\5g3^J\\La5U4lImKT6d4YI^Kg6c5000000000N2O1VNkH\\MW7e2oHRMS7o2SIiLo6X3WI^Lk6d3YISLj6o3YIhKk6Z4XI_Kj6c4n0`0@2O1N2N3N1N4LSN^HXN[7^1SIeNf6T1fIkNV6o0WJlNh5Q1eJeNZ5Z1hJfNW5X1mJfNS5Z1oJcNR5]1oJbNZ5U1hJhNZ5W1gJhNY5X1iJfNX5Y1jJdNW5\\1kJbNU5_1nJ\\NT5c1PKTNU5m1nJjMW5W2i25K6M3M4L3N2M3N2N3M2O1N2O2M2O2N100^HZL]5h3^J_L_5a3]JeL`5]3[JiLb5X3TJQMm5o2bI`M_6QOSIh24_Ne7_1SHhNP8X33M3M2N0O101O0O2N101N1N101N2O1N1eLdG]1h8_N`Gf0W9VORG5^9HQG^O\\9a0h1M3N2F9GRhk4"}, "label": "a baby zebra"}]}
{"id": 399835, "image": "COCO_train2014_000000399835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small zebra standing next to two adults\"."}], "task": "refering", "answer_template": "The \"a small zebra standing next to two adults\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307], "bbox": [0.22715625, 0.41426229508196716, 0.41498437499999996, 0.9094613583138172], "iscrowd": 0, "area": 14695.8167, "rle": {"size": [427, 640], "counts": "Zfl11U=<E:E:F6M2O2N2N2N2M2O3M5K5VOSNmEQ2P:RNjEU2Q:nMjEX2S:d0L3M4L3L3N2N2N2cHZLQ5h3kJ[LT5g3hJ\\LW5f3fJ\\LY5f3cJ]L\\5g3^J\\La5U4lImKT6d4YI^Kg6c5000000000N2O1VNkH\\MW7e2oHRMS7o2SIiLo6X3WI^Lk6d3YISLj6o3YIhKk6Z4XI_Kj6c4n0`0@2O1N2N3N1N4LSN^HXN[7^1SIeNf6T1fIkNV6o0WJlNh5Q1eJeNZ5Z1hJfNW5X1mJfNS5Z1oJcNR5]1oJbNZ5U1hJhNZ5W1gJhNY5X1iJfNX5Y1jJdNW5\\1kJbNU5_1nJ\\NT5c1PKTNU5m1nJjMW5W2i25K6M3M4L3N2M3N2N3M2O1N2O2M2O2N100^HZL]5h3^J_L_5a3]JeL`5]3[JiLb5X3TJQMm5o2bI`M_6QOSIh24_Ne7_1SHhNP8X33M3M2N0O101O0O2N101N1N101N2O1N1eLdG]1h8_N`Gf0W9VORG5^9HQG^O\\9a0h1M3N2F9GRhk4"}, "label": "a small zebra standing next to two adults"}]}
{"id": 432603, "image": "COCO_train2014_000000432603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in a blue and green striped shirt playing wii\"."}], "task": "refering", "answer_template": "The \"a young boy in a blue and green striped shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 73, 74, 75, 76, 89, 90, 91, 92, 93, 107, 108, 109, 110, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 344, 345, 346, 347, 348, 349, 361, 362, 363, 364, 365], "bbox": [0.21320833333333333, 0.16215625, 0.5976041666666667, 0.952703125], "iscrowd": 0, "area": 65553.89809999999, "rle": {"size": [640, 480], "counts": "\\RP23cc0j0UOa0@;E<H7J7I6J7I6K5M4L3M4L3M6J7I8H5K6L3L5L3L5K4M3L4M4K4M8RGVJU3]6YKcJf4]7bH[I_7c9O100O1TM\\D\\Mc;c2XEcLi:[3aE]L_:`3fE]L[:a3hE]LY:`3kE_LT:`3PF]LQ:`3SF_Lm9_3VF_Lk9^3ZF`Lf9^3]F`Lc9a3_F\\Lb9c3`F\\L`9d3aFZL`9e3cFWL_9i3bFTL_9m3bFPL`9o3cFoK]9Q4dFnK\\9Q4gFmKY9S4hFkKX9U4kFiKU9W4nFfKR9Z4PGcKQ9\\4SGaKm8_4UG_Kj8a4[300O1000O0100O100O01000O100O10O100000000O100000000O010000000O100000000O1000001O0O2O001O001N101O001O0O2O001O001N101O001O0O2\\DnJX8R5eGRKZ8o4cGTK\\8m4`GXK^8j4^GYKa8h4]GZKa8i4[G[Kc8f4[G\\Kd8f4YG\\Kf8f4TG_Kk8b4PGcKn8`4lFdKT9^4fFgKY9Z4aFkK_9X4YFnKf9W4QFmKn9Z4gElKX:o6O001O0O2O001O001N2O001O001N101O001^LcF_L]9]3UGTLm8h3`GnK`8m3kGmKV8k3XHnKh7k3cHnKa7l3hHlK\\7T4gHfK]7Z4eH_K_7a4dHXKa7h4aHQKc7o4k3L8H7I8H7I8H7I8G8I8HS1lN8I8H:E>C=CUWh3"}, "label": "a young boy in a blue and green striped shirt playing wii"}]}
{"id": 432603, "image": "COCO_train2014_000000432603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a blue and green striped shirt\"."}], "task": "refering", "answer_template": "The \"a little boy in a blue and green striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 73, 74, 75, 76, 89, 90, 91, 92, 93, 107, 108, 109, 110, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 225, 226, 227, 228, 229, 230, 231, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 344, 345, 346, 347, 348, 349, 361, 362, 363, 364, 365], "bbox": [0.21320833333333333, 0.16215625, 0.5976041666666667, 0.952703125], "iscrowd": 0, "area": 65553.89809999999, "rle": {"size": [640, 480], "counts": "\\RP23cc0j0UOa0@;E<H7J7I6J7I6K5M4L3M4L3M6J7I8H5K6L3L5L3L5K4M3L4M4K4M8RGVJU3]6YKcJf4]7bH[I_7c9O100O1TM\\D\\Mc;c2XEcLi:[3aE]L_:`3fE]L[:a3hE]LY:`3kE_LT:`3PF]LQ:`3SF_Lm9_3VF_Lk9^3ZF`Lf9^3]F`Lc9a3_F\\Lb9c3`F\\L`9d3aFZL`9e3cFWL_9i3bFTL_9m3bFPL`9o3cFoK]9Q4dFnK\\9Q4gFmKY9S4hFkKX9U4kFiKU9W4nFfKR9Z4PGcKQ9\\4SGaKm8_4UG_Kj8a4[300O1000O0100O100O01000O100O10O100000000O100000000O010000000O100000000O1000001O0O2O001O001N101O001O0O2O001O001N101O001O0O2\\DnJX8R5eGRKZ8o4cGTK\\8m4`GXK^8j4^GYKa8h4]GZKa8i4[G[Kc8f4[G\\Kd8f4YG\\Kf8f4TG_Kk8b4PGcKn8`4lFdKT9^4fFgKY9Z4aFkK_9X4YFnKf9W4QFmKn9Z4gElKX:o6O001O0O2O001O001N2O001O001N101O001^LcF_L]9]3UGTLm8h3`GnK`8m3kGmKV8k3XHnKh7k3cHnKa7l3hHlK\\7T4gHfK]7Z4eH_K_7a4dHXKa7h4aHQKc7o4k3L8H7I8H7I8H7I8G8I8HS1lN8I8H:E>C=CUWh3"}, "label": "a little boy in a blue and green striped shirt"}]}
{"id": 432603, "image": "COCO_train2014_000000432603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sofa with the guitar\"."}], "task": "refering", "answer_template": "The \"the sofa with the guitar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 225, 230, 238, 239, 240, 241, 242, 247, 248, 249, 250, 255, 256, 257, 264, 265, 266, 267, 272, 281, 282, 283, 289, 290, 291, 292, 293, 298, 299, 300, 301, 306, 307, 308, 309, 310, 315, 316, 317, 318, 323, 324, 325, 326, 327, 332, 333, 334, 335, 340, 341, 342, 343, 344, 349, 350, 351, 352, 357, 358, 359, 360, 361, 374, 375, 376, 377], "bbox": [0.0, 0.46021875, 0.7457499999999999, 1.0], "iscrowd": 0, "area": 46579.702999999994, "rle": {"size": [640, 480], "counts": "o<h0k1f3e8jN[GU1\\75bHL^75aHK_76`HIa78^HHb78^HHb78^HHb79]HGc79]HFd7:\\HFd7;ZHEg7;YHDh7=VHCk7=UHCk7>SHBn7>RHBn7>QHCo7=QHCo7=QHCo7=QHBP8?oGAQ8?oGAQ87fFcKX1V4R81nFgKP1W4S81]HOc70^H0b7O_H1a7O_H1a7O_H0b70^H0b70^H0b70^H0b70]H1c70\\HOe72ZHNf72ZHNf73YHMg74XHKi76VHJj79SHGm7;QHEo7>mGCS8>lGAU8>lGBT8>lGBT8>lGBT8=mGBT8>kGCU8=kGCU8<kGEU8;kGEU8;jGEW8:jGFV8:jGFV8:iGGW88jGHV88jGGW89iGGW88jGHV88jGHV88jGGW88jGHV88jGHV88iGIW86jGJV86jGIW87iGIW86jGJV86jGJW8H^FVL[1Q4X8AgF\\LR1R4W8AiF[LR1S4U8BjFZLS1R4S8CcH=]7CcH<^7CcH=]7CcH=]7CbH>^7AcH?]7AcH>^7AcH?]7AcH?]7@dH`0\\7@dH?]7AcH?]7AcH?]7AcH?]7AcH?]7@dH?]7AcH?]7AcH?]7AcH?]7AcH>^7BbH>^7BbH>^7AcH?]7AcH?]7AcH>^7BbH>^7BbH>^7BbH>^7AcH?]7AcH>^7BbH>^7BbH>^7_MhFNj1c2d9\\M\\Fc2f9\\MZFd2g9[MXFf2i9YMWFg2i9YMWFg2j9WMWFh2k9VMVFj2k9TMVFl2k9RMVFn2k9PMVFo2l9oLUFQ3k9nLUFS3l9jLVFV3k9hLVFX3k9fLVFY3l9eLUF[3l9cLUF]3l9aLUF_3m9^LTFb3S:VLnEi3U=000000O012N1Od0\\O=fNh^OQOga0>]^O@Sb00P^OB`b09f0M4CTkU29aTjM<_^OZOh?g0V@\\Oi?e0U@^OR2IQ:i0lC@o1HW:d0mCGh1F\\:`0mCNb1Db::oC4[1Dg:5PDf0h0WOY;OQDe1J]NX<JPDa2nNgMU?[2h@fMi>Dk@h28fMm>Dh@i27eMQ?Ee@i25cMW?Fb@l20`M^?F_@e3a?=0O100000000O01O100O1O001O1O1O001L4iNcKgBa4X=aKcBc4]=eKSBb4n=h01O00001N1000001N10001O000O101O000O2O00001O0O10001O000O2O00001N1000001O0O101O000QMZAd0g>ZO[Ad0f>ZO]Ae0c>YO`Af0`>UOfAj0[>QOjAm0W>oNmAQ1S>QNVA3j0k1Q>QNWA0k0n1n=QNZAMk0Q2l=PN]AJj0T2k=PN^AGk0X2i=nM_AFj0\\2h=lMoBS2R=jMQCS2`?K4K6J6lNm]O4Yb0Jj]O0Zb0Oi]OG`b07c]O^Ogb0a0>M3L3L5L[h[2"}, "label": "the sofa with the guitar"}]}
{"id": 432603, "image": "COCO_train2014_000000432603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green couch behind a little boy\"."}], "task": "refering", "answer_template": "The \"a green couch behind a little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 225, 230, 238, 239, 240, 241, 242, 247, 248, 249, 250, 255, 256, 257, 264, 265, 266, 267, 272, 281, 282, 283, 289, 290, 291, 292, 293, 298, 299, 300, 301, 306, 307, 308, 309, 310, 315, 316, 317, 318, 323, 324, 325, 326, 327, 332, 333, 334, 335, 340, 341, 342, 343, 344, 349, 350, 351, 352, 357, 358, 359, 360, 361, 374, 375, 376, 377], "bbox": [0.0, 0.46021875, 0.7457499999999999, 1.0], "iscrowd": 0, "area": 46579.702999999994, "rle": {"size": [640, 480], "counts": "o<h0k1f3e8jN[GU1\\75bHL^75aHK_76`HIa78^HHb78^HHb78^HHb79]HGc79]HFd7:\\HFd7;ZHEg7;YHDh7=VHCk7=UHCk7>SHBn7>RHBn7>QHCo7=QHCo7=QHCo7=QHBP8?oGAQ8?oGAQ87fFcKX1V4R81nFgKP1W4S81]HOc70^H0b7O_H1a7O_H1a7O_H0b70^H0b70^H0b70^H0b70]H1c70\\HOe72ZHNf72ZHNf73YHMg74XHKi76VHJj79SHGm7;QHEo7>mGCS8>lGAU8>lGBT8>lGBT8>lGBT8=mGBT8>kGCU8=kGCU8<kGEU8;kGEU8;jGEW8:jGFV8:jGFV8:iGGW88jGHV88jGGW89iGGW88jGHV88jGHV88jGGW88jGHV88jGHV88iGIW86jGJV86jGIW87iGIW86jGJV86jGJW8H^FVL[1Q4X8AgF\\LR1R4W8AiF[LR1S4U8BjFZLS1R4S8CcH=]7CcH<^7CcH=]7CcH=]7CbH>^7AcH?]7AcH>^7AcH?]7AcH?]7@dH`0\\7@dH?]7AcH?]7AcH?]7AcH?]7AcH?]7@dH?]7AcH?]7AcH?]7AcH?]7AcH>^7BbH>^7BbH>^7AcH?]7AcH?]7AcH>^7BbH>^7BbH>^7BbH>^7AcH?]7AcH>^7BbH>^7BbH>^7_MhFNj1c2d9\\M\\Fc2f9\\MZFd2g9[MXFf2i9YMWFg2i9YMWFg2j9WMWFh2k9VMVFj2k9TMVFl2k9RMVFn2k9PMVFo2l9oLUFQ3k9nLUFS3l9jLVFV3k9hLVFX3k9fLVFY3l9eLUF[3l9cLUF]3l9aLUF_3m9^LTFb3S:VLnEi3U=000000O012N1Od0\\O=fNh^OQOga0>]^O@Sb00P^OB`b09f0M4CTkU29aTjM<_^OZOh?g0V@\\Oi?e0U@^OR2IQ:i0lC@o1HW:d0mCGh1F\\:`0mCNb1Db::oC4[1Dg:5PDf0h0WOY;OQDe1J]NX<JPDa2nNgMU?[2h@fMi>Dk@h28fMm>Dh@i27eMQ?Ee@i25cMW?Fb@l20`M^?F_@e3a?=0O100000000O01O100O1O001O1O1O001L4iNcKgBa4X=aKcBc4]=eKSBb4n=h01O00001N1000001N10001O000O101O000O2O00001O0O10001O000O2O00001N1000001O0O101O000QMZAd0g>ZO[Ad0f>ZO]Ae0c>YO`Af0`>UOfAj0[>QOjAm0W>oNmAQ1S>QNVA3j0k1Q>QNWA0k0n1n=QNZAMk0Q2l=PN]AJj0T2k=PN^AGk0X2i=nM_AFj0\\2h=lMoBS2R=jMQCS2`?K4K6J6lNm]O4Yb0Jj]O0Zb0Oi]OG`b07c]O^Ogb0a0>M3L3L5L[h[2"}, "label": "a green couch behind a little boy"}]}
{"id": 432603, "image": "COCO_train2014_000000432603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand and leg of an adult\"."}], "task": "refering", "answer_template": "The \"the hand and leg of an adult\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 237, 253, 254, 270, 271, 287, 288, 303, 304, 305, 320, 321, 322, 336, 337, 338, 353, 354], "bbox": [0.7987916666666667, 0.5236562499999999, 1.0, 0.8817499999999999], "iscrowd": 0, "area": 8488.9023, "rle": {"size": [640, 480], "counts": "Sm_71mc0?A5L1N2O1N1O2O1N2N2O0O2O1N2N2O0O2N2O1N2O1N1O2N2O1N2N1O2N2O1N2N1O2N2N2O1N1O2N2N2N2O0O2N2N2N2N01OO2J5K5K6L3N2M3M4M2N2O2O0O101N1O2O0O2O0O2N1100O010O100O010O10O01O100O010O10N101N101N101N101NB?^Na1_NTG"}, "label": "the hand and leg of an adult"}]}
{"id": 432603, "image": "COCO_train2014_000000432603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone sitting by a coffee table with hand on knee with a ring on right hand\"."}], "task": "refering", "answer_template": "The \"someone sitting by a coffee table with hand on knee with a ring on right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 237, 253, 254, 270, 271, 287, 288, 303, 304, 305, 320, 321, 322, 336, 337, 338, 353, 354], "bbox": [0.7987916666666667, 0.5236562499999999, 1.0, 0.8817499999999999], "iscrowd": 0, "area": 8488.9023, "rle": {"size": [640, 480], "counts": "Sm_71mc0?A5L1N2O1N1O2O1N2N2O0O2O1N2N2O0O2N2O1N2O1N1O2N2O1N2N1O2N2O1N2N1O2N2N2O1N1O2N2N2N2O0O2N2N2N2N01OO2J5K5K6L3N2M3M4M2N2O2O0O101N1O2O0O2O0O2N1100O010O100O010O10O01O100O010O10N101N101N101N101NB?^Na1_NTG"}, "label": "someone sitting by a coffee table with hand on knee with a ring on right hand"}]}
{"id": 39395, "image": "COCO_train2014_000000039395.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra next to a fence\"."}], "task": "refering", "answer_template": "The \"a zebra next to a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 250, 251, 252, 253, 254, 255, 258, 259, 260, 261, 262, 263, 268, 269, 270, 271, 274, 275, 276, 277, 278, 285, 286, 287, 291, 292, 293, 302, 303, 319], "bbox": [0.16329694323144106, 0.193265625, 1.0, 0.8786562499999999], "iscrowd": 0, "area": 111902.57224999997, "rle": {"size": [640, 458], "counts": "eQ_12nc02M3N3M2]EF\\1<]NJa19YNKg18QNNn14lM1R22gM4X2O`M7_2K[M:d2HUM=j2FoL`0P3BjLc0U3^OeLh0Y3[O`Lk0_3VO\\Ln0d3SOVLS1i3nNRLW1m3jNmK\\1R4eNhKa1X4_NcKf1]4ZN]Kk1d4UNVKQ2i4PNRKU2n4kMoJX2Q5hMlJ[2T5eMiJ^2V5cMhJ^2Y5bMdJa2\\5_MaJd2^5]M_Jf2a5ZM]Jh2c5XMZJj2g5WMUJl2j5UMSJn2m5RMQJn2Q6RMlIP3U6PMhIR3X6oLfIR3[6nLbIT3_6lL^IV3b6kL[IW3f6iLXIX3i6hLTIZ3k6hLRIZ3l6iLQIY3n6iLPIX3n6kLoHW3P7kLmHW3Q7lLlHV3m5iJ^HT2c1T3m5mJ\\HQ2c1U3P6nJXHP2d1T3T6PKTHo1b1U3Y6PKPHn1c1U3[6QKnGm1c1U3^6RKkGk1b1W3b6RKgGj1c1W3d6oMXIT2g6nMTIV2k6lMQIW2o6jMmHY2S7hMhH\\2Y7dMcH_2]7bM_Ha2b7_MYHe2g7\\MUHg2l7YMoGk2Q8VMkGl2W8TMeGo2\\8Z4001O001O001O00001O0000000N2O1O001N2O001O1N101O1O0O2O001O1O0O2O1O001N2O001O0O2O1O001N2O001O1N101O001N2O001O1O0O2O1O001N101O1O0O2O1O1O1O100O100O2N100O1O100O2N100O1O100O1O2O0O100O1O101N1O100O1O101N1O100O100O1O2O0O1O100O1O2O0O1O100O1O101N100O1O100O2N100O1O100O2N100001O1O1O010O1O1O001O100O001O1O1O00100O1O010O1O10O01O100O00100O1O00100O1O010O1O10O01O100O00100O1O010O1O10O01O100O001O100O00100O1O010O1O10O01O100O00100O1O010O1O10O01O1O10O01O100O00100O1O010O1O10O01O100O00100O1O010O1O1O010O2N100O1O100O1O2O0O1O100O1O100O2N100O1O100O1O2N100O1O100O1O101N1O100O1O100O2N100O1O100O1O101N1O1O100O1O100O2N100O1O100O1O2O0O1O100O1O100O2N1SJ"}, "label": "a zebra next to a fence"}]}
{"id": 39395, "image": "COCO_train2014_000000039395.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra which is close to fence\"."}], "task": "refering", "answer_template": "The \"a zebra which is close to fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 250, 251, 252, 253, 254, 255, 258, 259, 260, 261, 262, 263, 268, 269, 270, 271, 274, 275, 276, 277, 278, 285, 286, 287, 291, 292, 293, 302, 303, 319], "bbox": [0.16329694323144106, 0.193265625, 1.0, 0.8786562499999999], "iscrowd": 0, "area": 111902.57224999997, "rle": {"size": [640, 458], "counts": "eQ_12nc02M3N3M2]EF\\1<]NJa19YNKg18QNNn14lM1R22gM4X2O`M7_2K[M:d2HUM=j2FoL`0P3BjLc0U3^OeLh0Y3[O`Lk0_3VO\\Ln0d3SOVLS1i3nNRLW1m3jNmK\\1R4eNhKa1X4_NcKf1]4ZN]Kk1d4UNVKQ2i4PNRKU2n4kMoJX2Q5hMlJ[2T5eMiJ^2V5cMhJ^2Y5bMdJa2\\5_MaJd2^5]M_Jf2a5ZM]Jh2c5XMZJj2g5WMUJl2j5UMSJn2m5RMQJn2Q6RMlIP3U6PMhIR3X6oLfIR3[6nLbIT3_6lL^IV3b6kL[IW3f6iLXIX3i6hLTIZ3k6hLRIZ3l6iLQIY3n6iLPIX3n6kLoHW3P7kLmHW3Q7lLlHV3m5iJ^HT2c1T3m5mJ\\HQ2c1U3P6nJXHP2d1T3T6PKTHo1b1U3Y6PKPHn1c1U3[6QKnGm1c1U3^6RKkGk1b1W3b6RKgGj1c1W3d6oMXIT2g6nMTIV2k6lMQIW2o6jMmHY2S7hMhH\\2Y7dMcH_2]7bM_Ha2b7_MYHe2g7\\MUHg2l7YMoGk2Q8VMkGl2W8TMeGo2\\8Z4001O001O001O00001O0000000N2O1O001N2O001O1N101O1O0O2O001O1O0O2O1O001N2O001O0O2O1O001N2O001O1N101O001N2O001O1O0O2O1O001N101O1O0O2O1O1O1O100O100O2N100O1O100O2N100O1O100O1O2O0O100O1O101N1O100O1O101N1O100O100O1O2O0O1O100O1O2O0O1O100O1O101N100O1O100O2N100O1O100O2N100001O1O1O010O1O1O001O100O001O1O1O00100O1O010O1O10O01O100O00100O1O00100O1O010O1O10O01O100O00100O1O010O1O10O01O100O001O100O00100O1O010O1O10O01O100O00100O1O010O1O10O01O1O10O01O100O00100O1O010O1O10O01O100O00100O1O010O1O1O010O2N100O1O100O1O2O0O1O100O1O100O2N100O1O100O1O2N100O1O100O1O101N1O100O1O100O2N100O1O100O1O101N1O1O100O1O100O2N100O1O100O1O2O0O1O100O1O100O2N1SJ"}, "label": "a zebra which is close to fence"}]}
{"id": 39395, "image": "COCO_train2014_000000039395.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the back with its mouth on the ground\"."}], "task": "refering", "answer_template": "The \"the zebra in the back with its mouth on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 28, 29, 30, 31, 43, 44, 45, 46, 47, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 110, 111, 232, 233, 234, 248, 249, 250, 251, 264, 265, 266], "bbox": [0.5314410480349345, 0.009171875, 1.0, 0.734], "iscrowd": 0, "area": 24092.077050000004, "rle": {"size": [640, 458], "counts": "SYh41oc09F:G9F:G3L100O2O0O2N100O2O0O101N101N100O2N100O2O0O2O0O101N101N1O101N100O100O10O01001O00fGjMkNU2S1nMmNQ2Q1QNPOn1n0TNROl1l0WNTOh1j0ZNWOe1h0\\NXOd1f0_NZO`1d0bN]O]1a0eN_O[1?hNAW1=kNDT1:nNFQ19QOHn07TOIk05WOLe05]OKa05BK;5GL55MK152KK57LF4<LA6a0J\\O6f0KWO5k0KRO6P1KmN5V1KgN5[1LbN4`1L]N5f1KWN6j1KSN5o1KnM6U2JhM6Z2f82O2O0O1O2N100O1O2N100O2N1O101N1O101N1O1O2O0O1O1O2O0O1O2O0O1O2N100O2N1O100O2N1O101N1O101N1O1O2O0O1O2N100O1O2O0O1O2N100O2N1O101N1O1O2O0O1O101N1O1O2O0O101N10001O00001O0000001N10001O00001O00001O0O10001O00001O00001O0O101O00001O0000001O0O101O00001O00001O00001N1000001O00001X3gLaL"}, "label": "the zebra in the back with its mouth on the ground"}]}
{"id": 39395, "image": "COCO_train2014_000000039395.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in back\"."}], "task": "refering", "answer_template": "The \"zebra in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 28, 29, 30, 31, 43, 44, 45, 46, 47, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 110, 111, 232, 233, 234, 248, 249, 250, 251, 264, 265, 266], "bbox": [0.5314410480349345, 0.009171875, 1.0, 0.734], "iscrowd": 0, "area": 24092.077050000004, "rle": {"size": [640, 458], "counts": "SYh41oc09F:G9F:G3L100O2O0O2N100O2O0O101N101N100O2N100O2O0O2O0O101N101N1O101N100O100O10O01001O00fGjMkNU2S1nMmNQ2Q1QNPOn1n0TNROl1l0WNTOh1j0ZNWOe1h0\\NXOd1f0_NZO`1d0bN]O]1a0eN_O[1?hNAW1=kNDT1:nNFQ19QOHn07TOIk05WOLe05]OKa05BK;5GL55MK152KK57LF4<LA6a0J\\O6f0KWO5k0KRO6P1KmN5V1KgN5[1LbN4`1L]N5f1KWN6j1KSN5o1KnM6U2JhM6Z2f82O2O0O1O2N100O1O2N100O2N1O101N1O101N1O1O2O0O1O1O2O0O1O2O0O1O2N100O2N1O100O2N1O101N1O101N1O1O2O0O1O2N100O1O2O0O1O2N100O2N1O101N1O1O2O0O1O101N1O1O2O0O101N10001O00001O0000001N10001O00001O00001O0O10001O00001O00001O0O101O00001O0000001O0O101O00001O00001O00001N1000001O00001X3gLaL"}, "label": "zebra in back"}]}
{"id": 137706, "image": "COCO_train2014_000000137706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest cow on the grass\"."}], "task": "refering", "answer_template": "The \"the biggest cow on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 240], "bbox": [0.33965625, 0.32728337236533955, 0.490421875, 0.6883138173302107], "iscrowd": 0, "area": 8828.685300000003, "rle": {"size": [427, 640], "counts": "cQk22X=2O2N2N2N2N2N1O000000001lC`0[:@dEf0W:ZOgEk0V:UOiEn0U:ROiER1U:nNjEV1S:jNkEZ1R:gNmE]1P:fNkE^1R:fNjE]1T:P1L3N3L3N3eFjLZ8T40O0100O001000O10000O100ZOg0N1O1N3f0[ON1O2N1O2K4C<E;6J5K5K5K6J5K5K4LO101N2N2O1N2N3N1N2O1N1O2O0O1K6J5J6K6J5J7J5K5J7J5K5M4N1N4L7I6J6K5J6J6J6J7IigW4"}, "label": "the biggest cow on the grass"}]}
{"id": 137706, "image": "COCO_train2014_000000137706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow with white speckled legs\"."}], "task": "refering", "answer_template": "The \"a brown cow with white speckled legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 240], "bbox": [0.33965625, 0.32728337236533955, 0.490421875, 0.6883138173302107], "iscrowd": 0, "area": 8828.685300000003, "rle": {"size": [427, 640], "counts": "cQk22X=2O2N2N2N2N2N1O000000001lC`0[:@dEf0W:ZOgEk0V:UOiEn0U:ROiER1U:nNjEV1S:jNkEZ1R:gNmE]1P:fNkE^1R:fNjE]1T:P1L3N3L3N3eFjLZ8T40O0100O001000O10000O100ZOg0N1O1N3f0[ON1O2N1O2K4C<E;6J5K5K5K6J5K5K4LO101N2N2O1N2N3N1N2O1N1O2O0O1K6J5J6K6J5J7J5K5J7J5K5M4N1N4L7I6J6K5J6J6J6J7IigW4"}, "label": "a brown cow with white speckled legs"}]}
{"id": 137715, "image": "COCO_train2014_000000137715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a marines military lady standing backside of the two ladies\"."}], "task": "refering", "answer_template": "The \"a marines military lady standing backside of the two ladies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 33, 34, 35, 36, 57, 58, 59, 80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 217, 218, 219, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 268, 269, 270, 271, 292, 293, 316], "bbox": [0.11490625000000002, 0.002191142191142191, 0.8294218750000001, 0.8747319347319347], "iscrowd": 0, "area": 56756.796299999995, "rle": {"size": [429, 640], "counts": "hWo0U1n:Z1kNT1101N2O001N2O0O2O0O2O1N101N2O1N3N1O1N2O1N2O1N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1N2O1N2O2N1N2O1N2O1N2O4K6K5J5L5J6K5K4K6KY[`1Nid_N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1`JRO_No0Z1XOeNi0T1]OlNd0n0ARO`0g0FYO;a0J_O7RMSOa0l0]22QMZO4j0k2MPMBGf0Y3InLKZOc0g3CnL2nN`0T4_OmL:`N=c4ZOkLc0TN8Q5VOjLS3V3nLiLR3W3oLgLR3Y3oLfLR3Y3oLfLQ3Z3PMdLQ3\\3PMcLP3]3QMbLo2^3RM`Lo2`3QM`Lo2`3RM_Ln2a3SM]Lo2b3^MQLb2o3lMcKT2]4ZNTKg1l4]300000001O0000000000001O000000000000001O000000000000001O000000000000001O0O10000000001O0000001O00001O001O00001O001O00001O001O001O00001O001O00001O001O08H<E;D<E;D<D<E:E9H1N101N2N101N2O0O2N2O0O2O1N1O2O1N101N2O001OO2N1N2O2M2O2M2O1N3M2O1N3N1N2O2M2N2O2O001N101N101O1N101O1N2O0O2O1O0O2O1N2N1O2N2M2O2N1N101O1N101O0O2O001O1N101O0O2O001N2O03N1N100O010O1001O0O01O1N2O1O001O1O1N2O001O1O1N2O001\\ObIWJ_6g5hISJY6j5g0N1N3N2M3N2M2O2M3N2M3N2M2O2M3N2`KeGj3^8QLgGl3\\8oKjGn3W8nKnGm2JfMQ9=iF<e0mNd8e0nF2a:JgEJa:2hEA_:=[1L3M4M2M4L4Mg^]1"}, "label": "a marines military lady standing backside of the two ladies"}]}
{"id": 137715, "image": "COCO_train2014_000000137715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in camoflage shirt with the word \" marines \" stitched on it\"."}], "task": "refering", "answer_template": "The \"person in camoflage shirt with the word \" marines \" stitched on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 33, 34, 35, 36, 57, 58, 59, 80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 171, 172, 173, 174, 175, 176, 177, 178, 179, 186, 187, 188, 194, 195, 196, 197, 198, 199, 200, 201, 202, 209, 210, 211, 217, 218, 219, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 268, 269, 270, 271, 292, 293, 316], "bbox": [0.11490625000000002, 0.002191142191142191, 0.8294218750000001, 0.8747319347319347], "iscrowd": 0, "area": 56756.796299999995, "rle": {"size": [429, 640], "counts": "hWo0U1n:Z1kNT1101N2O001N2O0O2O0O2O1N101N2O1N3N1O1N2O1N2O1N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1N2O1N2O2N1N2O1N2O1N2O4K6K5J5L5J6K5K4K6KY[`1Nid_N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N101O1O1O1O1O1O1O1O1O1`JRO_No0Z1XOeNi0T1]OlNd0n0ARO`0g0FYO;a0J_O7RMSOa0l0]22QMZO4j0k2MPMBGf0Y3InLKZOc0g3CnL2nN`0T4_OmL:`N=c4ZOkLc0TN8Q5VOjLS3V3nLiLR3W3oLgLR3Y3oLfLR3Y3oLfLQ3Z3PMdLQ3\\3PMcLP3]3QMbLo2^3RM`Lo2`3QM`Lo2`3RM_Ln2a3SM]Lo2b3^MQLb2o3lMcKT2]4ZNTKg1l4]300000001O0000000000001O000000000000001O000000000000001O000000000000001O0O10000000001O0000001O00001O001O00001O001O00001O001O001O00001O001O00001O001O08H<E;D<E;D<D<E:E9H1N101N2N101N2O0O2N2O0O2O1N1O2O1N101N2O001OO2N1N2O2M2O2M2O1N3M2O1N3N1N2O2M2N2O2O001N101N101O1N101O1N2O0O2O1O0O2O1N2N1O2N2M2O2N1N101O1N101O0O2O001O1N101O0O2O001N2O03N1N100O010O1001O0O01O1N2O1O001O1O1N2O001O1O1N2O001\\ObIWJ_6g5hISJY6j5g0N1N3N2M3N2M2O2M3N2M3N2M2O2M3N2`KeGj3^8QLgGl3\\8oKjGn3W8nKnGm2JfMQ9=iF<e0mNd8e0nF2a:JgEJa:2hEA_:=[1L3M4M2M4L4Mg^]1"}, "label": "person in camoflage shirt with the word \" marines \" stitched on it"}]}
{"id": 137715, "image": "COCO_train2014_000000137715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white woman on the right\"."}], "task": "refering", "answer_template": "The \"the white woman on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 204, 205, 206, 217, 218, 226, 227, 228, 229, 240, 241, 242, 243, 244, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.463953125, 0.0, 1.0, 0.9820279720279721], "iscrowd": 0, "area": 77215.97310000002, "rle": {"size": [429, 640], "counts": "ffl31Y=4K4M4O00001[ECY8<eGGZ89eGH[89bGJ]86aGL`84^GNa82]G0c80\\G1e8OXG4g8LWG6i8KUG7j8ITG9m8FQG=n8DPG=P9CnFa0Q9_OlFe0R9[OmFh0Q9XOmFl0Q9UOmFm0S9ROkFR1S9oNjFU1T9kNlFW1S9hNmFY1R9hNmFZ1Q9fNoF\\1o8eNoF^1P9aNPGa1n8_NRGb1m8_NRGc1m8\\NSGf1k8ZNUGh1i8YNVGi1h8WNXGj1h8VNWGl1g8TNYGn1e8RN[GP2c8QN\\GP2d8oM\\GQ2d8PN[GP2e8PN[GP2f8oMZGP2g8QNXGo1h8QNXGo1h8RNWGm1k8SNTGm1l8TNSGl1m8UNRGk1o8UNPGj1Q9WNnFi1R9XNmFh1S9YNlFg1U9ZNiFe1X9\\NgFd1Y9]NfFc1[9]NdFb1]9_NbFa1^9`NaF`1_9aN`F_1a9aN^F^1c9cN\\F]1d9dN[F\\1f9dNYF\\1g9fNWFY1j9hNUFX1k9iNTFW1m9S110001O0O10001O0O10001N10001O0O10001O0O101O001O100O^FRMf8m2ZGUMd8k2\\GXMb8g2^G[Ma8d2_G_M_8`2aGbM^8^2aGeM]8Z2cGhM[8X2eGkMY8T2gGnMX8Q2hGRNV8m1jGVNT8j1kGXNS8h1mG[NQ8d1oG^NP8a1PHbNn7]1RHeNm7Z1SHiNk7V1UHlNi7U1VHnNh7Q1XHQOg7n0YHUOe7j0[HXOd7g0\\H\\Oa7d0_H^O`7a0`HB^7>aHE]7:cHH\\77dHLZ73fH0W70iH2V7MjH6T7JkH9S7FmH<R7CnH`0o6@PId0n6[ORIh0l6WOTIk0k6TOUIU1c6kN\\Ia1X6_NhIm1m5RNSJR2j5mMVJS2k5lMUJT2l5kMTJU2m5jMSJV2m5kMRJU2o5jMQJV2P6iMPJW2Q6hMoIX2R6gMnIY2R6gMnIY2S6fMmIZ2T6fMkIZ2V6eMjI[2W6dMiI\\2X6cMhI]2X6cMhI]2Y6bMgI^2Z6bMeI^2\\6aMdI_2]6`McI`2]6`McI`2^6_MbIa2_6^MaIb2`6]M`Ic2a6]M^Ic2c6\\M]Id2c6\\M]Id2d6[M\\Ie2e6ZM[If2f6YMZIg2g6XMYIh2g6YMXIg2i6XMWIh2j6WMVIi2j6WMVIi2j6WMVIi2j6WMVIi2j6XMUIh2j6YMVIh2i6XMWIi2h6WMXIj2g6VMYIk2f6UMZIl2e6TM[Im2d6TM[Im2c6TM]Im2b6SM^In2a6RM_Io2`6QM`IP3_6PMaIQ3^6oLbIR3]6oLbIR3\\6oLdIR3[6nLeIS3Z6mLfIT3Y6lLgIT3Y6lLgIU3X6kLhIV3W6jLiIW3U6jLkIW3T6iLmIW3R6jLmIW3R6iLnIX3Q6hLoIY3P6gLPJZ3o5fLQJ[3m5fLSJ[3k5fLUJ[3g5hLYJY3d5iL\\JX3`5kL`JV3\\5mLdJT3X5oLhJR3T5QMlJQ3o4RMQKR3h4QMXKR3b4QM^KS3[4PMeKS3V4oLjKU3P4mLPLW3j3kLVLX3d3kL\\LY3^3iLbLZ3Y3hLgL\\3S3fLmLZ3Q3hLoLX3n2kLRMU3l2mLTMR3k2PMUMP3i2RMWMn2f2UMZMk2d2WM\\Mh2c2ZM]Mf2a2\\M_Md2^2_MbMa2]2`McM_2\\2cMdM]2[2dMeM\\2Z2eMfM[2Y2fMgMY2X2iMhMW2W2jMiMV2V2kMjMU2U2lMkMT2S2nMmMQ2S2PNmMP2R2QNnMo1Q2SNnMm1P2UNPNj1P2WNPNi1o1XNQNh1n1YNRNg1l1[NTNd1l1]NTNc1k1^NUNb1j1_NVNa1h1aNXN_1g1bNYN]1g1dNYN\\1d1gN\\NY1b1iN^NW1`1kN`NT1^1oNbNQ1\\1QOeNn0X1UOhNk0V1WOjNf70000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000O2N1O1O100O1O1O1O100O1O1O1O1O1G9XMh2XMh2XMh2XMe2"}, "label": "the white woman on the right"}]}
{"id": 137715, "image": "COCO_train2014_000000137715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white lady with phone\"."}], "task": "refering", "answer_template": "The \"white lady with phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 204, 205, 206, 217, 218, 226, 227, 228, 229, 240, 241, 242, 243, 244, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.463953125, 0.0, 1.0, 0.9820279720279721], "iscrowd": 0, "area": 77215.97310000002, "rle": {"size": [429, 640], "counts": "ffl31Y=4K4M4O00001[ECY8<eGGZ89eGH[89bGJ]86aGL`84^GNa82]G0c80\\G1e8OXG4g8LWG6i8KUG7j8ITG9m8FQG=n8DPG=P9CnFa0Q9_OlFe0R9[OmFh0Q9XOmFl0Q9UOmFm0S9ROkFR1S9oNjFU1T9kNlFW1S9hNmFY1R9hNmFZ1Q9fNoF\\1o8eNoF^1P9aNPGa1n8_NRGb1m8_NRGc1m8\\NSGf1k8ZNUGh1i8YNVGi1h8WNXGj1h8VNWGl1g8TNYGn1e8RN[GP2c8QN\\GP2d8oM\\GQ2d8PN[GP2e8PN[GP2f8oMZGP2g8QNXGo1h8QNXGo1h8RNWGm1k8SNTGm1l8TNSGl1m8UNRGk1o8UNPGj1Q9WNnFi1R9XNmFh1S9YNlFg1U9ZNiFe1X9\\NgFd1Y9]NfFc1[9]NdFb1]9_NbFa1^9`NaF`1_9aN`F_1a9aN^F^1c9cN\\F]1d9dN[F\\1f9dNYF\\1g9fNWFY1j9hNUFX1k9iNTFW1m9S110001O0O10001O0O10001N10001O0O10001O0O101O001O100O^FRMf8m2ZGUMd8k2\\GXMb8g2^G[Ma8d2_G_M_8`2aGbM^8^2aGeM]8Z2cGhM[8X2eGkMY8T2gGnMX8Q2hGRNV8m1jGVNT8j1kGXNS8h1mG[NQ8d1oG^NP8a1PHbNn7]1RHeNm7Z1SHiNk7V1UHlNi7U1VHnNh7Q1XHQOg7n0YHUOe7j0[HXOd7g0\\H\\Oa7d0_H^O`7a0`HB^7>aHE]7:cHH\\77dHLZ73fH0W70iH2V7MjH6T7JkH9S7FmH<R7CnH`0o6@PId0n6[ORIh0l6WOTIk0k6TOUIU1c6kN\\Ia1X6_NhIm1m5RNSJR2j5mMVJS2k5lMUJT2l5kMTJU2m5jMSJV2m5kMRJU2o5jMQJV2P6iMPJW2Q6hMoIX2R6gMnIY2R6gMnIY2S6fMmIZ2T6fMkIZ2V6eMjI[2W6dMiI\\2X6cMhI]2X6cMhI]2Y6bMgI^2Z6bMeI^2\\6aMdI_2]6`McI`2]6`McI`2^6_MbIa2_6^MaIb2`6]M`Ic2a6]M^Ic2c6\\M]Id2c6\\M]Id2d6[M\\Ie2e6ZM[If2f6YMZIg2g6XMYIh2g6YMXIg2i6XMWIh2j6WMVIi2j6WMVIi2j6WMVIi2j6WMVIi2j6XMUIh2j6YMVIh2i6XMWIi2h6WMXIj2g6VMYIk2f6UMZIl2e6TM[Im2d6TM[Im2c6TM]Im2b6SM^In2a6RM_Io2`6QM`IP3_6PMaIQ3^6oLbIR3]6oLbIR3\\6oLdIR3[6nLeIS3Z6mLfIT3Y6lLgIT3Y6lLgIU3X6kLhIV3W6jLiIW3U6jLkIW3T6iLmIW3R6jLmIW3R6iLnIX3Q6hLoIY3P6gLPJZ3o5fLQJ[3m5fLSJ[3k5fLUJ[3g5hLYJY3d5iL\\JX3`5kL`JV3\\5mLdJT3X5oLhJR3T5QMlJQ3o4RMQKR3h4QMXKR3b4QM^KS3[4PMeKS3V4oLjKU3P4mLPLW3j3kLVLX3d3kL\\LY3^3iLbLZ3Y3hLgL\\3S3fLmLZ3Q3hLoLX3n2kLRMU3l2mLTMR3k2PMUMP3i2RMWMn2f2UMZMk2d2WM\\Mh2c2ZM]Mf2a2\\M_Md2^2_MbMa2]2`McM_2\\2cMdM]2[2dMeM\\2Z2eMfM[2Y2fMgMY2X2iMhMW2W2jMiMV2V2kMjMU2U2lMkMT2S2nMmMQ2S2PNmMP2R2QNnMo1Q2SNnMm1P2UNPNj1P2WNPNi1o1XNQNh1n1YNRNg1l1[NTNd1l1]NTNc1k1^NUNb1j1_NVNa1h1aNXN_1g1bNYN]1g1dNYN\\1d1gN\\NY1b1iN^NW1`1kN`NT1^1oNbNQ1\\1QOeNn0X1UOhNk0V1WOjNf70000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000O2N1O1O100O1O1O1O100O1O1O1O1O1G9XMh2XMh2XMh2XMe2"}, "label": "white lady with phone"}]}
{"id": 137715, "image": "COCO_train2014_000000137715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"a girl wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0015, 0.033846153846153845, 0.70646875, 0.9866666666666666], "iscrowd": 0, "area": 88968.21669999996, "rle": {"size": [429, 640], "counts": "Zg0=o<k0VOj0UOl0UO8G100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100SJXLT2i3mMXLQ2h3oMYLo1h3QNZLm1f3TN[Li1f3WN\\Lg1d3YN]Le1d3\\N]La1d3_N^L_1b3`NaL]1`3aNcL^1^3^NgL_1Z3_NjL_1V3_NnL^1S3`NPM_1P3_NTM^1m2`NWM]1j2`NZM_1f2_N^M^1c2`NaM^1_2`NeM]1\\2aNgM^1Y2`NkM]1V2`NoM^1Q2`NSN]1n1aNUN]1l1aNXN]1h1aN\\N\\1e1bN^N]1b1`NcN]1^1aNfN]1Z1aNjN\\1W1dNjN[1V1dNlNZ1U1eNnNY1R1fNPOX1Q1gNQOW1P1iNQOV1o0iNTOT1m0kNUOT1k0kNWOS1j0mNWOR1i0mNZOP1g0oN[OP1e0oN]Oo0d0PO^On0c0RO^Om0b0ROAk0`0TOBk0>TODj0=UOEj0;VOGg0:XOHg08XOJf07YOKe06ZOLe04[ONb03]OOb01]O1a00^O2a0N^O5?LA5>L@6>KA7=JB8=HB;;FE;:EE=9DF>9BFa07@Ia05@Jb05^OKc03^OMc02]ONe0O\\O0f0OZO1g0MZO3g0LYO3j0JWO6j0HWO8m0DSO<Q1_OPO`0U1\\OkNd0X1XOiNh0[1TOeNk0`1POaNP1n60000O10000O1000000O1000000O1000000O100000000O100000001N1000000O100000000O1000000O1000000000000000000000000000000000000000TL[OPKe0P5AjJ?U5HeJ8[5N_J2a54YJLg5:SJFl5b0mI^OS6h0gIXOY6n0aIRO_6T1[IlNd6[1VIeNj6\\1UIdNk6\\1UIdNk6\\1UIdNj6]1VIcNj6]1VIcNj6]1VIcNj6]1VIcNi6^1XIaNh6_1XIaNh6_1XIaNh6_1XIaNg6`1YI`Ng6`1YI`Ng6`1YI`Ng6a1XI_Ng6b1YI^Ng6b1YI^Ng6b1YI^Ng6b1YI^Nf6c1ZI]Nf6c1ZI^Ne6b1[I^Ne6b1[I^Nd6c1\\I]Nd6c1\\I^Nc6b1]I^Nc6b1]I^Nb6c1^I^Na6b1_I^Na6c1^I^Na6b1_I^N`6c1`I^N_6b1aI^N_6b1aI_N^6a1bI_N]6b1cI_N\\6a1dI_N\\6a1dI`N[6`1eI`NZ6a1fI`NY6`1gI`NY6`1gIaNX6_1hIaNW6`1iIaNV6_1jIbNU6_1jIaNV6_1jIbNT6_1lIaNT6_1lIbNS6^1mIbNS6^1mIcNQ6^1oIbNQ6^1oIcNP6]1PJcNP6]1PJdNn5]1RJcNn5]1RJdNm5\\1SJdNm5\\1SJeNk5\\1VJcNj5^1UJcNj5]1VJcNj5]1VJdNh5]1XJcNh5]1XJdNg5\\1YJdNg5\\1YJeNe5\\1[JdNe5\\1[JfNc5Z1]JiN`5W1`JmN[5T1eJoNX5Q1hJROU5n0kJUOR5k0nJXOn4i0RK[Oj4f0UK]Oh4c0XK@e4`0[KBb4?^KC`4=`KE^4;bKG\\49dKIZ47fKKX45hKMV43PKTMKk2S51PKUMNl2P5OQKVMOm2o4LPKYM1m2m4JQKYM3o2j4HRKZM4P3h4FRK\\M6P3f4ERK[M9R3c4CRK]M;U3^4^OVK]M=]3f2aLeMd2VO`M?b3R2lLTNR2ZOaM`0i3\\1TMhNb1[OaMb0Q4e0[M]OS1ZOcMd0V4OcM3d0YOdMe0]4YOjMh05XOeMh0Z70QKWOfMi0Y70QKVOfMk0Z7OoJTOiMm0X7OoJSOjMn0X7NdK3\\4MdK3]4LcK4]4LbK5_4JaK6_4JaK6`4J^K8a4H_K8a4H^K9c4F]K:c4F]K:d4E[K<e4D[K<f4DYK<g4DXK=i4BWK>i4BVK?j4AVK?k4@UK`0k4@TKa0m4^OSKb0m4_ORKa0o4^OPKc0P5]OPKc0Q5\\OPKc0P5]ORKa0o4^OSK`0m4@TK?l4BUK<l4CVK;j4EXK9i4FYK8g4H[K6f4I[K6e4J]K4d4K`K1`40cKL^43fKIZ47iKFW4:kKDV4;mKBS4>oK@R4?^4010O01O00001O0010O01O001O001O00010O001O001O0010O0001O001O0010O100O100O1O010O100O100O100O100O1O100O100O100O100O00100O100O100O100O11O1O001O1O1O1O1O001OGAdC=^<DaC;`<F_C9b<H\\C8d<J[C5f<LYC3h<NWC1j<0UCOl<9001O1O1O1O001O1O1O1O001N2O1O001O1O1O1O00Y]^2"}, "label": "a girl wearing a white shirt"}]}
{"id": 137715, "image": "COCO_train2014_000000137715.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white t shirt smilingly looking at a man with military t shirt showing something in mobile phone to her\"."}], "task": "refering", "answer_template": "The \"a woman in white t shirt smilingly looking at a man with military t shirt showing something in mobile phone to her\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0015, 0.033846153846153845, 0.70646875, 0.9866666666666666], "iscrowd": 0, "area": 88968.21669999996, "rle": {"size": [429, 640], "counts": "Zg0=o<k0VOj0UOl0UO8G100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100SJXLT2i3mMXLQ2h3oMYLo1h3QNZLm1f3TN[Li1f3WN\\Lg1d3YN]Le1d3\\N]La1d3_N^L_1b3`NaL]1`3aNcL^1^3^NgL_1Z3_NjL_1V3_NnL^1S3`NPM_1P3_NTM^1m2`NWM]1j2`NZM_1f2_N^M^1c2`NaM^1_2`NeM]1\\2aNgM^1Y2`NkM]1V2`NoM^1Q2`NSN]1n1aNUN]1l1aNXN]1h1aN\\N\\1e1bN^N]1b1`NcN]1^1aNfN]1Z1aNjN\\1W1dNjN[1V1dNlNZ1U1eNnNY1R1fNPOX1Q1gNQOW1P1iNQOV1o0iNTOT1m0kNUOT1k0kNWOS1j0mNWOR1i0mNZOP1g0oN[OP1e0oN]Oo0d0PO^On0c0RO^Om0b0ROAk0`0TOBk0>TODj0=UOEj0;VOGg0:XOHg08XOJf07YOKe06ZOLe04[ONb03]OOb01]O1a00^O2a0N^O5?LA5>L@6>KA7=JB8=HB;;FE;:EE=9DF>9BFa07@Ia05@Jb05^OKc03^OMc02]ONe0O\\O0f0OZO1g0MZO3g0LYO3j0JWO6j0HWO8m0DSO<Q1_OPO`0U1\\OkNd0X1XOiNh0[1TOeNk0`1POaNP1n60000O10000O1000000O1000000O1000000O100000000O100000001N1000000O100000000O1000000O1000000000000000000000000000000000000000TL[OPKe0P5AjJ?U5HeJ8[5N_J2a54YJLg5:SJFl5b0mI^OS6h0gIXOY6n0aIRO_6T1[IlNd6[1VIeNj6\\1UIdNk6\\1UIdNk6\\1UIdNj6]1VIcNj6]1VIcNj6]1VIcNj6]1VIcNi6^1XIaNh6_1XIaNh6_1XIaNh6_1XIaNg6`1YI`Ng6`1YI`Ng6`1YI`Ng6a1XI_Ng6b1YI^Ng6b1YI^Ng6b1YI^Ng6b1YI^Nf6c1ZI]Nf6c1ZI^Ne6b1[I^Ne6b1[I^Nd6c1\\I]Nd6c1\\I^Nc6b1]I^Nc6b1]I^Nb6c1^I^Na6b1_I^Na6c1^I^Na6b1_I^N`6c1`I^N_6b1aI^N_6b1aI_N^6a1bI_N]6b1cI_N\\6a1dI_N\\6a1dI`N[6`1eI`NZ6a1fI`NY6`1gI`NY6`1gIaNX6_1hIaNW6`1iIaNV6_1jIbNU6_1jIaNV6_1jIbNT6_1lIaNT6_1lIbNS6^1mIbNS6^1mIcNQ6^1oIbNQ6^1oIcNP6]1PJcNP6]1PJdNn5]1RJcNn5]1RJdNm5\\1SJdNm5\\1SJeNk5\\1VJcNj5^1UJcNj5]1VJcNj5]1VJdNh5]1XJcNh5]1XJdNg5\\1YJdNg5\\1YJeNe5\\1[JdNe5\\1[JfNc5Z1]JiN`5W1`JmN[5T1eJoNX5Q1hJROU5n0kJUOR5k0nJXOn4i0RK[Oj4f0UK]Oh4c0XK@e4`0[KBb4?^KC`4=`KE^4;bKG\\49dKIZ47fKKX45hKMV43PKTMKk2S51PKUMNl2P5OQKVMOm2o4LPKYM1m2m4JQKYM3o2j4HRKZM4P3h4FRK\\M6P3f4ERK[M9R3c4CRK]M;U3^4^OVK]M=]3f2aLeMd2VO`M?b3R2lLTNR2ZOaM`0i3\\1TMhNb1[OaMb0Q4e0[M]OS1ZOcMd0V4OcM3d0YOdMe0]4YOjMh05XOeMh0Z70QKWOfMi0Y70QKVOfMk0Z7OoJTOiMm0X7OoJSOjMn0X7NdK3\\4MdK3]4LcK4]4LbK5_4JaK6_4JaK6`4J^K8a4H_K8a4H^K9c4F]K:c4F]K:d4E[K<e4D[K<f4DYK<g4DXK=i4BWK>i4BVK?j4AVK?k4@UK`0k4@TKa0m4^OSKb0m4_ORKa0o4^OPKc0P5]OPKc0Q5\\OPKc0P5]ORKa0o4^OSK`0m4@TK?l4BUK<l4CVK;j4EXK9i4FYK8g4H[K6f4I[K6e4J]K4d4K`K1`40cKL^43fKIZ47iKFW4:kKDV4;mKBS4>oK@R4?^4010O01O00001O0010O01O001O001O00010O001O001O0010O0001O001O0010O100O100O1O010O100O100O100O100O1O100O100O100O100O00100O100O100O100O11O1O001O1O1O1O1O001OGAdC=^<DaC;`<F_C9b<H\\C8d<J[C5f<LYC3h<NWC1j<0UCOl<9001O1O1O1O001O1O1O1O001N2O1O001O1O1O1O00Y]^2"}, "label": "a woman in white t shirt smilingly looking at a man with military t shirt showing something in mobile phone to her"}]}
{"id": 301558, "image": "COCO_train2014_000000301558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis of the person in black who ' s lifting their leg up\"."}], "task": "refering", "answer_template": "The \"the skis of the person in black who ' s lifting their leg up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 271, 272, 273, 292, 293, 311, 312, 313, 314, 315, 316, 334, 335], "bbox": [0.5270625, 0.6150416666666667, 0.89471875, 0.8380416666666668], "iscrowd": 0, "area": 1898.339949999999, "rle": {"size": [480, 640], "counts": "Zjn42m>3N2N2M100000O1000O100000O1000O100000O1000O1000O100000O1000O1000O1000O100000O1000O100000O01000000O10O1000O1000O10O1000000O01000O1000000O1000000O1000000O1000O10O10O10O010000O01000O10O10O1000O010000O01000O010O1000O010000O01000O10O10O1000O010000O010O10O10O1000O010000O01000001N10001O0O101O0O101O00QYa0<bf^O1O2N1O1O1N2O10010O0001O00001O00010O1O001O1O00100O1O001O1O1O01O01O0010O010O010O0010O010M2N3M2O1NWbo0"}, "label": "the skis of the person in black who ' s lifting their leg up"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black iphone\"."}], "task": "refering", "answer_template": "The \"black iphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 240, 241, 242, 243, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 331, 332, 333, 334, 355, 356], "bbox": [0.40449999999999997, 0.5842708333333333, 0.5983125, 0.9123541666666666], "iscrowd": 0, "area": 11767.126500000004, "rle": {"size": [480, 640], "counts": "^ii3<b>3M2N2M4M2N3L3N3M2M4M2N2M4M2N3L3N3M2N3L3N3M2M3N3M2M4M2N3L3N3M2M3N3M2N3L3N3M2M4M2N3L3N2N3L3N3M2M4M2N3M2O1010O01O001O001O001O0000O1000000O1000000O100O1N2M3N3L3M3N2M4M2M3N3L3M3N3L3N2M4L3N2M4M2M3M4M2M3N3L3N3L3M3N3L3N2M4L3N2M4M2M3M4M2M3N3L3N3L5K5L5J5LaUh3"}, "label": "black iphone"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"telephone underneath a tablet\"."}], "task": "refering", "answer_template": "The \"telephone underneath a tablet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 240, 241, 242, 243, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 331, 332, 333, 334, 355, 356], "bbox": [0.40449999999999997, 0.5842708333333333, 0.5983125, 0.9123541666666666], "iscrowd": 0, "area": 11767.126500000004, "rle": {"size": [480, 640], "counts": "^ii3<b>3M2N2M4M2N3L3N3M2M4M2N2M4M2N3L3N3M2N3L3N3M2M3N3M2M4M2N3L3N3M2M3N3M2N3L3N3M2M4M2N3L3N2N3L3N3M2M4M2N3M2O1010O01O001O001O001O0000O1000000O1000000O100O1N2M3N3L3M3N2M4M2M3N3L3M3N3L3N2M4L3N2M4M2M3M4M2M3N3L3N3L3M3N3L3N2M4L3N2M4M2M3M4M2M3N3L3N3L5K5L5J5LaUh3"}, "label": "telephone underneath a tablet"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tablet between two other tablets\"."}], "task": "refering", "answer_template": "The \"a tablet between two other tablets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 242, 243, 244], "bbox": [0.41974999999999996, 0.1826875, 0.691015625, 0.6132708333333333], "iscrowd": 0, "area": 27664.5741, "rle": {"size": [480, 640], "counts": "kZn38c>7J5K6J5K6J5K6I6K6J5K5K4L4L4L4L4L5J5L4L4L4L4L5K4L4L4K5L4L4L5K4L4L4L4L4K6K4M3O1O1O1O2N1O1O1000001O00000000001O00000000001O00000000001O0001O00000001O0001O000001O000001O0001O000001O0001O00000001O01O00000000010O0000000000010O000000000010O00000000010O0000000001O01O0000000O1O101O0O100O10000O2O000N2L4L4J7J5J6J6K5J6J6K6I6I7H8I7I8H7H9H7I8H7H9H8H7I9G:Ehfl2"}, "label": "a tablet between two other tablets"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tablet in between two other tablets\"."}], "task": "refering", "answer_template": "The \"a tablet in between two other tablets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 242, 243, 244], "bbox": [0.41974999999999996, 0.1826875, 0.691015625, 0.6132708333333333], "iscrowd": 0, "area": 27664.5741, "rle": {"size": [480, 640], "counts": "kZn38c>7J5K6J5K6J5K6I6K6J5K5K4L4L4L4L4L5J5L4L4L4L4L5K4L4L4K5L4L4L5K4L4L4L4L4K6K4M3O1O1O1O2N1O1O1000001O00000000001O00000000001O00000000001O0001O00000001O0001O000001O000001O0001O000001O0001O00000001O01O00000000010O0000000000010O000000000010O00000000010O0000000001O01O0000000O1O101O0O100O10000O2O000N2L4L4J7J5J6J6K5J6J6K6I6I7H8I7I8H7H9H7I8H7H9H8H7I9G:Ehfl2"}, "label": "a tablet in between two other tablets"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big i pad is place first from left\"."}], "task": "refering", "answer_template": "The \"a big i pad is place first from left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.015578125000000002, 0.12208333333333334, 0.43442187500000007, 0.6337708333333334], "iscrowd": 0, "area": 52231.5937, "rle": {"size": [480, 640], "counts": "Yn4h0U>3N2M3M4L3M3M3N2M3M3M3M3N2M3M3M3M3M4M2M3M3M3M3M3N2M3M3M3M3M3N2M4L3M3M3M3N2M3M3M3M3M3N2M3M3M4L3N2M3M3M3M3M3N2M3M3M3M3M3N3L3M3M3M3M3N2M3M3M3M3M3N2M301O000000000000000000000000001O000000000000000000000000001O0000000000000000000000000010O00000000000000000000000001O0000000001O0000000000000001O000000000000000001O00000001O000000000000000000000000010O000000000000000000000000001O000001O00000000000000000001O00000000000001O000000000O100000000O100000000O10001O00000O100000000O100J6J6J6J6I7J6J6J7H7J6J6I7J6J6J6I7J6J6J6I7J6J6I7J6J9G9F;F:F:F;D;FXcY5"}, "label": "a big i pad is place first from left"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ipad\"."}], "task": "refering", "answer_template": "The \"an ipad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.015578125000000002, 0.12208333333333334, 0.43442187500000007, 0.6337708333333334], "iscrowd": 0, "area": 52231.5937, "rle": {"size": [480, 640], "counts": "Yn4h0U>3N2M3M4L3M3M3N2M3M3M3M3N2M3M3M3M3M4M2M3M3M3M3M3N2M3M3M3M3M3N2M4L3M3M3M3N2M3M3M3M3M3N2M3M3M4L3N2M3M3M3M3M3N2M3M3M3M3M3N3L3M3M3M3M3N2M3M3M3M3M3N2M301O000000000000000000000000001O000000000000000000000000001O0000000000000000000000000010O00000000000000000000000001O0000000001O0000000000000001O000000000000000001O00000001O000000000000000000000000010O000000000000000000000000001O000001O00000000000000000001O00000000000001O000000000O100000000O100000000O10001O00000O100000000O100J6J6J6J6I7J6J6J7H7J6J6I7J6J6J6I7J6J6J6I7J6J6I7J6J9G9F;F:F:F;D;FXcY5"}, "label": "an ipad"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rightmost kindle\"."}], "task": "refering", "answer_template": "The \"rightmost kindle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 341, 342], "bbox": [0.6368750000000001, 0.33379166666666665, 0.9385468750000001, 0.8421875], "iscrowd": 0, "area": 35752.39794999999, "rle": {"size": [480, 640], "counts": "[co55f>6J5K5K6J5K6J5K5J7J5K5K6J5K5K6J5K6J5J6K6J5K5K6J5K5K6J5K6I6K5K6J5K5K6J5K6J5K5J7N1O10000O2O000O1001O01O000001O0001O01O000000010O0000000010O000001O0001O0001O0000010O0000000010O00000001O01O0001O000001O01O00000010O00000001O01O000001O0001O01O000000010O0000000010O000001O0001O0001O000001O01O00000010O00000001O01O0001O000001O01O000000010O0000001J5K5J6K5J7H7G9F:G9F;F9F:G9F;F9G:E;F:E;F:E<E:E;FRSb0"}, "label": "rightmost kindle"}]}
{"id": 375294, "image": "COCO_train2014_000000375294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tablet with keyboard\"."}], "task": "refering", "answer_template": "The \"tablet with keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 341, 342], "bbox": [0.6368750000000001, 0.33379166666666665, 0.9385468750000001, 0.8421875], "iscrowd": 0, "area": 35752.39794999999, "rle": {"size": [480, 640], "counts": "[co55f>6J5K5K6J5K6J5K5J7J5K5K6J5K5K6J5K6J5J6K6J5K5K6J5K5K6J5K6I6K5K6J5K5K6J5K6J5K5J7N1O10000O2O000O1001O01O000001O0001O01O000000010O0000000010O000001O0001O0001O0000010O0000000010O00000001O01O0001O000001O01O00000010O00000001O01O000001O0001O01O000000010O0000000010O000001O0001O0001O000001O01O00000010O00000001O01O0001O000001O01O000000010O0000001J5K5J6K5J7H7G9F:G9F;F9F:G9F;F9G:E;F:E;F:E<E:E;FRSb0"}, "label": "tablet with keyboard"}]}
{"id": 514559, "image": "COCO_train2014_000000514559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man crossing a stream of water\"."}], "task": "refering", "answer_template": "The \"a man crossing a stream of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 26, 27, 28, 29, 30, 31, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 150, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 230, 231, 232, 235, 236, 237, 253, 254, 255, 258, 259, 260, 276, 277, 281, 282, 299, 305, 306], "bbox": [0.0, 0.04395989974937343, 0.548078125, 1.0], "iscrowd": 0, "area": 45657.0199, "rle": {"size": [399, 640], "counts": "P46j67mM4R2f0QM^Om2k0kLVOS3l0lLUOS3k0nLUOP3k0QMVOm2i0UMXOi2h0XMXOh2g0ZMYOd2g0]MYOb2g0_MZO_2f0cMYO[2h0fMYOX2g0iMYOV2g0lMYOR2g0oMYOP2h0PNYOn1h0SNWOk1k0UNVOi1k0WNUOh1l0XNUOf1l0[NTOc1m0]NSOa1n0`NSO^1n0cNQO[1Q1eNPOY1P1hNQOV1P1kNoNS1S1mNnNQ1R1POnNo0S1QOnNn0R1SOmNm0S1SOnNl0R1TOnNl0R1UOmNk0S1SOPOl0P1QOSOo0m0QOTOn0l0SOTOl0l0UOTOmNlNgNQ2]2SOhNVOcNg1g2SOcN_O^N_1P3RO^NIYNW1Z3POZN2TNo0c3nNVNl0`M7[4mNRNX3o1hLnM[3R2eLkM^3U2bLiM_3X2aLeMb3Z2_LcMc3^2]L_Mf3a2YL]Mj3b2WLVMP4k2PLnLV4R3kKlLV4U3jKhLX4Y3hKeLY4[3hKcLY4^3fKaL[4_3fK_L[4b3eK\\L\\4e3dKZL\\4f3eKYL[4h3eKXLZ4h3gKWLY4j3fKVLY4k3hKTLX4l3iKSLW4n3hKRLW4o3jKnKX4R4iKlKW4V4iKgKY4Y4gKeK[4[4fKbK[4`4eK^K\\4b4eKZK]4g4cKTKa4n4^11O1O2M2O1O2N1O1O2N1O1N3N1N3N1N3N1O2O1O001O1O001O11N101O1N2O1O1O001O1N2O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001OQMSK]Nl4b1VK^Ni4a1YK_Nf4`1\\K`Nc4^1aK`N_4_1cKaNOVMo3W4TLcNI[MQ4Q4XLdNB`MU4j3\\LWOb3h0`LXO_3g0cLZO[3e0gL\\OW3b0lL^OS3a0PM_On2`0TMFe29]MOU24nMOk13WNOb14`NNY15iNNP13SOOf04]ON=3E07OK4POaK6Z4m0n0ROROP1n0nNROT1m0lNQOW1n0hNROZ1m0eNSO_1j0`NVOc1h0]NVOh1g0WNYOm1d0RN\\OQ2b0nM^OV2?jMBX2;hMC\\2;dMCa2:_MDf2:YMEj29VMEo28QMGS36mLIW34iLJ[34eLK_32aLNb3O^L1d3M\\L3g3JYL6i3HWL7l3kMdLSOAQ3l3dMlLYOXOS3n3_MoL\\OSOU3o3\\MRM^OoNU3Q4XMUMAjNV3S4cLbKOe16gNW3o6hLQIX3Q7fLoHZ3S7dLnH[3S7dLmH\\3T7cLlH]3T7cLmH\\3T7cLlH]3U7bLlH]3U7bLlH]3T7cLlHd1WOOn7]NkHb1YO1l7\\NkHb1[O1k7\\NkH`1]O3h7]NkH^1@4f7]NnHZ1^O8e7]NnHY1^O:f8D\\G;e8C]G<d8C]G=d8@^G?d8^O^Ga0d8\\O]Gd0f8XO\\Gg0f8WO[Gh0h8SO[Gl0h8oNZGR1R:O1O010O1O1O100O1O100O1O1O010O001O01O01O001O000O2O001O000010O0001O01O01O00010O0010O0001O010O000010O01O00010O001O010O001O010O001O010O0010O01O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O010O01O01O010O010O01O001O001O001O010O001O1O001O002N4L5K4MR^`3"}, "label": "a man crossing a stream of water"}]}
{"id": 514559, "image": "COCO_train2014_000000514559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing jeans and a green vest trying to cross creek walking on stones\"."}], "task": "refering", "answer_template": "The \"a man wearing jeans and a green vest trying to cross creek walking on stones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 26, 27, 28, 29, 30, 31, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 150, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 230, 231, 232, 235, 236, 237, 253, 254, 255, 258, 259, 260, 276, 277, 281, 282, 299, 305, 306], "bbox": [0.0, 0.04395989974937343, 0.548078125, 1.0], "iscrowd": 0, "area": 45657.0199, "rle": {"size": [399, 640], "counts": "P46j67mM4R2f0QM^Om2k0kLVOS3l0lLUOS3k0nLUOP3k0QMVOm2i0UMXOi2h0XMXOh2g0ZMYOd2g0]MYOb2g0_MZO_2f0cMYO[2h0fMYOX2g0iMYOV2g0lMYOR2g0oMYOP2h0PNYOn1h0SNWOk1k0UNVOi1k0WNUOh1l0XNUOf1l0[NTOc1m0]NSOa1n0`NSO^1n0cNQO[1Q1eNPOY1P1hNQOV1P1kNoNS1S1mNnNQ1R1POnNo0S1QOnNn0R1SOmNm0S1SOnNl0R1TOnNl0R1UOmNk0S1SOPOl0P1QOSOo0m0QOTOn0l0SOTOl0l0UOTOmNlNgNQ2]2SOhNVOcNg1g2SOcN_O^N_1P3RO^NIYNW1Z3POZN2TNo0c3nNVNl0`M7[4mNRNX3o1hLnM[3R2eLkM^3U2bLiM_3X2aLeMb3Z2_LcMc3^2]L_Mf3a2YL]Mj3b2WLVMP4k2PLnLV4R3kKlLV4U3jKhLX4Y3hKeLY4[3hKcLY4^3fKaL[4_3fK_L[4b3eK\\L\\4e3dKZL\\4f3eKYL[4h3eKXLZ4h3gKWLY4j3fKVLY4k3hKTLX4l3iKSLW4n3hKRLW4o3jKnKX4R4iKlKW4V4iKgKY4Y4gKeK[4[4fKbK[4`4eK^K\\4b4eKZK]4g4cKTKa4n4^11O1O2M2O1O2N1O1O2N1O1N3N1N3N1N3N1O2O1O001O1O001O11N101O1N2O1O1O001O1N2O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001OQMSK]Nl4b1VK^Ni4a1YK_Nf4`1\\K`Nc4^1aK`N_4_1cKaNOVMo3W4TLcNI[MQ4Q4XLdNB`MU4j3\\LWOb3h0`LXO_3g0cLZO[3e0gL\\OW3b0lL^OS3a0PM_On2`0TMFe29]MOU24nMOk13WNOb14`NNY15iNNP13SOOf04]ON=3E07OK4POaK6Z4m0n0ROROP1n0nNROT1m0lNQOW1n0hNROZ1m0eNSO_1j0`NVOc1h0]NVOh1g0WNYOm1d0RN\\OQ2b0nM^OV2?jMBX2;hMC\\2;dMCa2:_MDf2:YMEj29VMEo28QMGS36mLIW34iLJ[34eLK_32aLNb3O^L1d3M\\L3g3JYL6i3HWL7l3kMdLSOAQ3l3dMlLYOXOS3n3_MoL\\OSOU3o3\\MRM^OoNU3Q4XMUMAjNV3S4cLbKOe16gNW3o6hLQIX3Q7fLoHZ3S7dLnH[3S7dLmH\\3T7cLlH]3T7cLmH\\3T7cLlH]3U7bLlH]3U7bLlH]3T7cLlHd1WOOn7]NkHb1YO1l7\\NkHb1[O1k7\\NkH`1]O3h7]NkH^1@4f7]NnHZ1^O8e7]NnHY1^O:f8D\\G;e8C]G<d8C]G=d8@^G?d8^O^Ga0d8\\O]Gd0f8XO\\Gg0f8WO[Gh0h8SO[Gl0h8oNZGR1R:O1O010O1O1O100O1O100O1O1O010O001O01O01O001O000O2O001O000010O0001O01O01O00010O0010O0001O010O000010O01O00010O001O010O001O010O001O010O0010O01O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O010O01O01O010O010O01O001O001O001O010O001O1O001O002N4L5K4MR^`3"}, "label": "a man wearing jeans and a green vest trying to cross creek walking on stones"}]}
{"id": 514559, "image": "COCO_train2014_000000514559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in white dress\"."}], "task": "refering", "answer_template": "The \"woman in white dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320], "bbox": [0.5001562500000001, 0.1595488721804511, 1.0, 0.9887719298245614], "iscrowd": 0, "area": 31721.381799999977, "rle": {"size": [399, 640], "counts": "khn31^<1O1O100O1O1O1O1O001O100O1O1O1O1O1O1O1O100O00000000O100O2O0O100000000000000000O10000000000000000000000000000O100000001O00000O1000000O10001O000O100000000O1000001N1000000O10000O0100000O1000O1000001O000O10000000001N10000000000O101O0000O010000000O0100000O10O100000O10O1000\\NJVFNO9j9c0VF]Oi9d0WF[Oi9f0WFZOh9g0XFYOg9h0XFYOf9j0YFVOf9k0ZFTOf9m0ZFSOe9o0YFROf9P1YFPOf9R1YFnNe9U1ZFjNf9W1ZFiNe9Y1YFhNf9[1XFeNg9]1XFbNh9Y2J:F;E;E1O1N101O00001O001N10001O001O001O01O010O010O0010O0010O010O001O001O001N3N3K5YOg0YOg0D<O0O2N2O0O1004J5K3M1O00001O0O101O0^JgMn0Z2mNjMS1V2jNmMV1S2gNPNY1Q2bNTN]1l1`NWN`1j1\\NYNd1g1XN]Nh1c1UN`Nk1`1QNdNo1]1mMfNS2Z1jMiNV2W1fMmNY2T1dMoN\\2Q1aMRO_2o0\\MUOd2k0YMXOg2h0VM[Oj2e0RM_On2a0oLBQ3?kLDU3<gLHY38dLK\\35aLN_32]L2c3OYL4g3LUL8k3HRL;m3FPL=P4ClKa0U4^OhKe0X4\\OdKg0]4XO_Kl0a4TO\\Ko0d4QOYKR1h4mNTKW1l4jNPKY1P5gNmJ\\1T5cNhJa1X5_NeJb3^3]L_Lm3Z3d2001N101O00001O001O00001O001O00001O001O00001O00001O001O00001O001J5J6J7I6J7I6K5J7I6K6J500O2O001O000O2O00001O0O2O001O1O1N3B=VLC"}, "label": "woman in white dress"}]}
{"id": 514559, "image": "COCO_train2014_000000514559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"women in white dress\"."}], "task": "refering", "answer_template": "The \"women in white dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320], "bbox": [0.5001562500000001, 0.1595488721804511, 1.0, 0.9887719298245614], "iscrowd": 0, "area": 31721.381799999977, "rle": {"size": [399, 640], "counts": "khn31^<1O1O100O1O1O1O1O001O100O1O1O1O1O1O1O1O100O00000000O100O2O0O100000000000000000O10000000000000000000000000000O100000001O00000O1000000O10001O000O100000000O1000001N1000000O10000O0100000O1000O1000001O000O10000000001N10000000000O101O0000O010000000O0100000O10O100000O10O1000\\NJVFNO9j9c0VF]Oi9d0WF[Oi9f0WFZOh9g0XFYOg9h0XFYOf9j0YFVOf9k0ZFTOf9m0ZFSOe9o0YFROf9P1YFPOf9R1YFnNe9U1ZFjNf9W1ZFiNe9Y1YFhNf9[1XFeNg9]1XFbNh9Y2J:F;E;E1O1N101O00001O001N10001O001O001O01O010O010O0010O0010O010O001O001O001N3N3K5YOg0YOg0D<O0O2N2O0O1004J5K3M1O00001O0O101O0^JgMn0Z2mNjMS1V2jNmMV1S2gNPNY1Q2bNTN]1l1`NWN`1j1\\NYNd1g1XN]Nh1c1UN`Nk1`1QNdNo1]1mMfNS2Z1jMiNV2W1fMmNY2T1dMoN\\2Q1aMRO_2o0\\MUOd2k0YMXOg2h0VM[Oj2e0RM_On2a0oLBQ3?kLDU3<gLHY38dLK\\35aLN_32]L2c3OYL4g3LUL8k3HRL;m3FPL=P4ClKa0U4^OhKe0X4\\OdKg0]4XO_Kl0a4TO\\Ko0d4QOYKR1h4mNTKW1l4jNPKY1P5gNmJ\\1T5cNhJa1X5_NeJb3^3]L_Lm3Z3d2001N101O00001O001O00001O001O00001O001O00001O00001O001O00001O001J5J6J7I6J7I6K5J7I6K6J500O2O001O000O2O00001O0O2O001O1O1N3B=VLC"}, "label": "women in white dress"}]}
{"id": 342532, "image": "COCO_train2014_000000342532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a police officer in a green vest riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a police officer in a green vest riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 176, 177, 199, 200, 222, 223, 245, 246, 269], "bbox": [0.47865624999999995, 0.07191666666666667, 0.7247187499999999, 0.6876458333333333], "iscrowd": 0, "area": 17433.236650000003, "rle": {"size": [480, 640], "counts": "gR`41o>2N2O1N2L4K5K5K5K3M1O2N1O1O2M2O2N1O1O2N1O2O01O100001O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1iNiMdEX2[:kMbEV2]:mM`ET2_:oM^EQ2b:QN\\EP2d:RNYEo1f:TNWEn1g:UNVEl1i:WNTEk1j:XNSEi1l:ZNQEh1m:[NPEf1o:W1N1O2N1O2N1O2N1O2N1O101O001O001O001O001O001O001O001O0010O01O00001O001O001O001O3M2NWLVF_2g9\\M^Fg2_9SMhFn2V9lLQGV3l8dL[G^3c8[LdGf3Z8TLmGn3b92N1O2M3N1O\\2dM0000000000000O10O10000000000000O13M4L4K5nKXGc1l8UN[Gh1j8oM\\Go1h8jM]GT2g8dM`GY2d8_MbG_2b8YMeGd2_8TMgGj2]8oLiGj1_9nMgF]1k;[Od0]ORYb2"}, "label": "a police officer in a green vest riding a motorcycle"}]}
{"id": 342532, "image": "COCO_train2014_000000342532.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"policeman riding motorcycle\"."}], "task": "refering", "answer_template": "The \"policeman riding motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 176, 177, 199, 200, 222, 223, 245, 246, 269], "bbox": [0.47865624999999995, 0.07191666666666667, 0.7247187499999999, 0.6876458333333333], "iscrowd": 0, "area": 17433.236650000003, "rle": {"size": [480, 640], "counts": "gR`41o>2N2O1N2L4K5K5K5K3M1O2N1O1O2M2O2N1O1O2N1O2O01O100001O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1iNiMdEX2[:kMbEV2]:mM`ET2_:oM^EQ2b:QN\\EP2d:RNYEo1f:TNWEn1g:UNVEl1i:WNTEk1j:XNSEi1l:ZNQEh1m:[NPEf1o:W1N1O2N1O2N1O2N1O2N1O101O001O001O001O001O001O001O001O0010O01O00001O001O001O001O3M2NWLVF_2g9\\M^Fg2_9SMhFn2V9lLQGV3l8dL[G^3c8[LdGf3Z8TLmGn3b92N1O2M3N1O\\2dM0000000000000O10O10000000000000O13M4L4K5nKXGc1l8UN[Gh1j8oM\\Go1h8jM]GT2g8dM`GY2d8_MbG_2b8YMeGd2_8TMgGj2]8oLiGj1_9nMgF]1k;[Od0]ORYb2"}, "label": "policeman riding motorcycle"}]}
{"id": 104973, "image": "COCO_train2014_000000104973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with dark hair and bangs\"."}], "task": "refering", "answer_template": "The \"a woman with dark hair and bangs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 55, 56, 57, 58, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 217, 218], "bbox": [0.303375, 0.07879166666666666, 0.615171875, 0.5911666666666666], "iscrowd": 0, "area": 27360.958150000006, "rle": {"size": [480, 640], "counts": "eTk22m>4K6J5L5J5L5J5L2M2O0O2O1N101N101N2N101N2O0O2O0O2O1N101N101N2O0O2O0O2N2O0O2O0O2O1N101N101N2O0O2O0N3N2N1O2M2O3M2M4M3M3L4M3M2M4M3M3L4M3M2M4M2N2N3L3N3M2M4M2N3L3N2N3L3N3M2N3N100O2N101N2N2O1N2N2O1N3N1N2N2O1N2N2O1N3M2O=C9G0O01O1O1O010O1O1O1O010O1O1O00100O1O00100O1O1O1O100O1O1O2N100O1O1O1O100O2N1O1O101N2N2N3N1N2N3M5K6K6I7I7I4L4M2M4L3M4L3N3L3M4L3N3L3M4L3M4M2M4L3M4M2M4L3M4M2M4L3M4M3L4L4L4M3L3MT1lN3N2M4L3M4M]Sc3"}, "label": "a woman with dark hair and bangs"}]}
{"id": 104973, "image": "COCO_train2014_000000104973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young lady\"."}], "task": "refering", "answer_template": "The \"a young lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 55, 56, 57, 58, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 217, 218], "bbox": [0.303375, 0.07879166666666666, 0.615171875, 0.5911666666666666], "iscrowd": 0, "area": 27360.958150000006, "rle": {"size": [480, 640], "counts": "eTk22m>4K6J5L5J5L5J5L2M2O0O2O1N101N101N2N101N2O0O2O0O2O1N101N101N2O0O2O0O2N2O0O2O0O2O1N101N101N2O0O2O0N3N2N1O2M2O3M2M4M3M3L4M3M2M4M3M3L4M3M2M4M2N2N3L3N3M2M4M2N3L3N2N3L3N3M2N3N100O2N101N2N2O1N2N2O1N3N1N2N2O1N2N2O1N3M2O=C9G0O01O1O1O010O1O1O1O010O1O1O00100O1O00100O1O1O1O100O1O1O2N100O1O1O1O100O2N1O1O101N2N2N3N1N2N3M5K6K6I7I7I4L4M2M4L3M4L3N3L3M4L3N3L3M4L3M4M2M4L3M4M2M4L3M4M2M4L3M4M3L4L4L4M3L3MT1lN3N2M4L3M4M]Sc3"}, "label": "a young lady"}]}
{"id": 104973, "image": "COCO_train2014_000000104973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the back in the blue shirt\"."}], "task": "refering", "answer_template": "The \"the man in the back in the blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 209, 210, 211, 212, 232, 233, 234, 255, 256, 257, 279], "bbox": [0.10304687500000001, 0.0033750000000000004, 0.52871875, 0.7849166666666666], "iscrowd": 0, "area": 31147.76675, "rle": {"size": [480, 640], "counts": "YWo01m>4L3M3M4L3M3M4L3M3L4L5K4L4M3L3M3dDXN]9j1^FZNb9h1XF]Nf9f1TF^Nk9e1oE_NP:c1kEaNT:b1fEcNY:_1bEdN]:_1]EeNb:]1YEgNf:\\1TEiNj:f200O1O2N100O1O2N100O2O0O10001N100O101O0O100O2O0O10001NM4J5J7K50O100O100O10000O2O2M2O2N2M2O2M3N1N3N2N1N3N2M2O2M2O2N2M2O2M3N1O2M3N1N3N2M2O2N2M2O2M2O2N2M2O2M3N1N3N2N1N3N2M2O2N2M2O2M3N1N30O100O100O1O100O1O100UOaHTJ`7j5fHRJZ7l5lHoIT7P6RIkIo6S6WIiIi6U6R1N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N1O2M3N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2M3N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O10000O1000O10O1000000O10000O1000000O1000000O1000000O10000000000000000000000000000000000000002N3M3M2N3M2N3M2N3M1O2NTa]4"}, "label": "the man in the back in the blue shirt"}]}
{"id": 375311, "image": "COCO_train2014_000000375311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark red ford van\"."}], "task": "refering", "answer_template": "The \"a dark red ford van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279], "bbox": [0.0013281249999999999, 0.2141111111111111, 0.17206249999999998, 0.9811666666666666], "iscrowd": 0, "area": 23694.27495, "rle": {"size": [360, 640], "counts": "e=b7g3Q1oNO001O1O2N0000001O000000001O0000001N1000001O0000001O0000001O0000001O0000001O0000001O0O2O00001O001O001O3M2N3M3M3M2N3M3M2N3M3M3M2N3M3M3M2N3L4M2N2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N5K5K5K5K5K5J7J6J6J6J6J6J6J6J6SOfG`N`8j0RHTOT87\\HHd9Nioi5"}, "label": "a dark red ford van"}]}
{"id": 375311, "image": "COCO_train2014_000000375311.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red van with half the rear window visible\"."}], "task": "refering", "answer_template": "The \"a red van with half the rear window visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279], "bbox": [0.0013281249999999999, 0.2141111111111111, 0.17206249999999998, 0.9811666666666666], "iscrowd": 0, "area": 23694.27495, "rle": {"size": [360, 640], "counts": "e=b7g3Q1oNO001O1O2N0000001O000000001O0000001N1000001O0000001O0000001O0000001O0000001O0000001O0O2O00001O001O001O3M2N3M3M3M2N3M3M2N3M3M3M2N3M3M3M2N3L4M2N2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N5K5K5K5K5K5J7J6J6J6J6J6J6J6J6SOfG`N`8j0RHTOT87\\HHd9Nioi5"}, "label": "a red van with half the rear window visible"}]}
{"id": 539158, "image": "COCO_train2014_000000539158.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a urinal to the right of the sink\"."}], "task": "refering", "answer_template": "The \"a urinal to the right of the sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 187, 188, 189, 204, 205, 206, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 256, 257, 258, 259, 273, 274, 275], "bbox": [0.027020833333333334, 0.46621875, 0.3183125, 0.72521875], "iscrowd": 0, "area": 11702.365549999997, "rle": {"size": [640, 480], "counts": "g^83ic07I7L4M4K4M3L4M3L5L3L4M3L4M4K4M3L4M3L5L3L4M3M4K4M3L4M3L5L3L4M3L4M4K4M3F:L4L1N2O1O001N2O1O1O0O2001O1O1N2O1O2N2M4M2N2N2N2M6K4K6K3M4K5L3M3L3N2M3N2N2M3N1O1N101N101O0O2O001N101N2O001N101N101O0O2O001N2O0O2O0O2O0O2N1O2N101N2N1O2N101N1O2N101N2N1O2N101N2N2N2N3N1N2N2N3M2N2M3N3M2N2N2N3M2NXo[6"}, "label": "a urinal to the right of the sink"}]}
{"id": 539158, "image": "COCO_train2014_000000539158.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the urinal on the wall\"."}], "task": "refering", "answer_template": "The \"the urinal on the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 187, 188, 189, 204, 205, 206, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 256, 257, 258, 259, 273, 274, 275], "bbox": [0.027020833333333334, 0.46621875, 0.3183125, 0.72521875], "iscrowd": 0, "area": 11702.365549999997, "rle": {"size": [640, 480], "counts": "g^83ic07I7L4M4K4M3L4M3L5L3L4M3L4M4K4M3L4M3L5L3L4M3M4K4M3L4M3L5L3L4M3L4M4K4M3F:L4L1N2O1O001N2O1O1O0O2001O1O1N2O1O2N2M4M2N2N2N2M6K4K6K3M4K5L3M3L3N2M3N2N2M3N1O1N101N101O0O2O001N101N2O001N101N101O0O2O001N2O0O2O0O2O0O2N1O2N101N2N1O2N101N1O2N101N2N1O2N101N2N2N2N3N1N2N2N3M2N2M3N3M2N2N2N3M2NXo[6"}, "label": "the urinal on the wall"}]}
{"id": 539158, "image": "COCO_train2014_000000539158.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toilet bowl\"."}], "task": "refering", "answer_template": "The \"the toilet bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 161, 162, 178, 179, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 244, 245, 246, 247], "bbox": [0.3301875, 0.37734375, 0.6022083333333333, 0.637], "iscrowd": 0, "area": 10659.055599999998, "rle": {"size": [640, 480], "counts": "URS31oc03L4M3L4M3L4L4M1N2O1N2O1N2N2O1N2\\^OiN]`0X1__OlN_`0V1\\_OoNb`0S1Y_OQOg`0P1T_OUOj`0m0Q_OXOn`0i0n^O[OPa0g0k^O^OTa0g1000001N10000000001O000000001O0O100000001O00000000001O000O10001O00000000001O000000001OO10001N10000ON3M3L4M2N3L4M2N3M3L3N3M3M2N3M3M2N3M3M2N3MoMPAXOP?c0UA]Ok>?YA@h>>ZABe><^ADb>:`AEa>8cAG]>7eAIZ>5iAJX>4jALV>2lANT>OPBOT>LnA4U>FnA9U>BnA<W>[OoAc0V>TOPBj0U>mNRBP1S>gNSBW1\\`0M3N2N3M2N2N2N2N3M2N3M2N3M2N3M2M4M2NYQg3"}, "label": "the toilet bowl"}]}
{"id": 539158, "image": "COCO_train2014_000000539158.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toilet , except for tank at top\"."}], "task": "refering", "answer_template": "The \"toilet , except for tank at top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 161, 162, 178, 179, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 227, 228, 229, 230, 244, 245, 246, 247], "bbox": [0.3301875, 0.37734375, 0.6022083333333333, 0.637], "iscrowd": 0, "area": 10659.055599999998, "rle": {"size": [640, 480], "counts": "URS31oc03L4M3L4M3L4L4M1N2O1N2O1N2N2O1N2\\^OiN]`0X1__OlN_`0V1\\_OoNb`0S1Y_OQOg`0P1T_OUOj`0m0Q_OXOn`0i0n^O[OPa0g0k^O^OTa0g1000001N10000000001O000000001O0O100000001O00000000001O000O10001O00000000001O000000001OO10001N10000ON3M3L4M2N3L4M2N3M3L3N3M3M2N3M3M2N3M3M2N3MoMPAXOP?c0UA]Ok>?YA@h>>ZABe><^ADb>:`AEa>8cAG]>7eAIZ>5iAJX>4jALV>2lANT>OPBOT>LnA4U>FnA9U>BnA<W>[OoAc0V>TOPBj0U>mNRBP1S>gNSBW1\\`0M3N2N3M2N2N2N2N3M2N3M2N3M2N3M2M4M2NYQg3"}, "label": "toilet , except for tank at top"}]}
{"id": 47639, "image": "COCO_train2014_000000047639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman putting her hands on the giant teddy bears\"."}], "task": "refering", "answer_template": "The \"the woman putting her hands on the giant teddy bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 192, 212, 213, 235, 236, 258, 259, 282, 283], "bbox": [0.215375, 0.2720140515222483, 0.45934375000000005, 0.8566510538641685], "iscrowd": 0, "area": 11498.120750000002, "rle": {"size": [427, 640], "counts": "Thi1a0g<;E6K4L5L3M3L4M3M3L4M5K5J6ZFcMU8a2fGbMW8d2dG_MW8l2aGUM\\8[3VGfLg8]3XGdLe8`3YGaLd8b3[G_Lb8e3\\G[Lb8h3]GYL_8l3_GTL_8]4M4M2M3N3L3N4M5K4M3L4L4MUOhHfKT7V4RIkKl6P4[IoKd6o3`IQL]6o3eIQLZ6n3iIRLU6l3oISLo5m3SJTLk5j3YJULf5j3]JVL`5j3bJVL]5h3gJXLV5Y1TJFh0QOR5U1_JDc0UOm4V1fJ_O`0ZOi4Y1jJWO`0_Oe4]1mJnNb0C`4a1XK_N;OT4l1oKdM0`0k3Q2\\NmMc1U2]NjMc1W2]NhMb1Y2`NeM_1]2aNbM_1_2aN`M^1a2dN]M[1k1eJYNP4KZ1m1gJXNP4IY1n1hJZNj9f1XFYNh9g1XFZNg9e1[F[Nc9f1]FZNc9e1_F[N`9d1bF\\N]9d1cF\\N\\9d1fF\\NY9c1hF^NW9a1kF^NU9b1kF_NT9`1nF`NP9a1QG^NP9a1PG`NP9`1PG`NP9`1oF`NS9^1nFbNR9^1mFcNT9[1mFeNS9[1mFdNT9[1mFeNT9Z1lFfNT9Y1lFgNU9Y1kFgNV9<nEMl0GV9M_F9<KW:3iENV:1jE1V:NjE3U:LkE6T:JlE5U:JkE6W:IhE6Z:JfE5[:JeE6\\:JcE6^:IcE5`:J_E6b:I_E4d:L^E0c:1^ELc:4_EHc:9_ECc:<V11O001O010O0000010O00010O000010O000010O00010O01O010O001O010O0010O01O010O001O010O0010O010O0104K]R`4"}, "label": "the woman putting her hands on the giant teddy bears"}]}
{"id": 47639, "image": "COCO_train2014_000000047639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young lady wearing t shirt with yellow over coat and blue colored jeans wearing sandals showing some toys is present in the image\"."}], "task": "refering", "answer_template": "The \"a young lady wearing t shirt with yellow over coat and blue colored jeans wearing sandals showing some toys is present in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 192, 212, 213, 235, 236, 258, 259, 282, 283], "bbox": [0.215375, 0.2720140515222483, 0.45934375000000005, 0.8566510538641685], "iscrowd": 0, "area": 11498.120750000002, "rle": {"size": [427, 640], "counts": "Thi1a0g<;E6K4L5L3M3L4M3M3L4M5K5J6ZFcMU8a2fGbMW8d2dG_MW8l2aGUM\\8[3VGfLg8]3XGdLe8`3YGaLd8b3[G_Lb8e3\\G[Lb8h3]GYL_8l3_GTL_8]4M4M2M3N3L3N4M5K4M3L4L4MUOhHfKT7V4RIkKl6P4[IoKd6o3`IQL]6o3eIQLZ6n3iIRLU6l3oISLo5m3SJTLk5j3YJULf5j3]JVL`5j3bJVL]5h3gJXLV5Y1TJFh0QOR5U1_JDc0UOm4V1fJ_O`0ZOi4Y1jJWO`0_Oe4]1mJnNb0C`4a1XK_N;OT4l1oKdM0`0k3Q2\\NmMc1U2]NjMc1W2]NhMb1Y2`NeM_1]2aNbM_1_2aN`M^1a2dN]M[1k1eJYNP4KZ1m1gJXNP4IY1n1hJZNj9f1XFYNh9g1XFZNg9e1[F[Nc9f1]FZNc9e1_F[N`9d1bF\\N]9d1cF\\N\\9d1fF\\NY9c1hF^NW9a1kF^NU9b1kF_NT9`1nF`NP9a1QG^NP9a1PG`NP9`1PG`NP9`1oF`NS9^1nFbNR9^1mFcNT9[1mFeNS9[1mFdNT9[1mFeNT9Z1lFfNT9Y1lFgNU9Y1kFgNV9<nEMl0GV9M_F9<KW:3iENV:1jE1V:NjE3U:LkE6T:JlE5U:JkE6W:IhE6Z:JfE5[:JeE6\\:JcE6^:IcE5`:J_E6b:I_E4d:L^E0c:1^ELc:4_EHc:9_ECc:<V11O001O010O0000010O00010O000010O000010O00010O01O010O001O010O0010O01O010O001O010O0010O010O0104K]R`4"}, "label": "a young lady wearing t shirt with yellow over coat and blue colored jeans wearing sandals showing some toys is present in the image"}]}
{"id": 301591, "image": "COCO_train2014_000000301591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hellman ' s ketchup\"."}], "task": "refering", "answer_template": "The \"hellman ' s ketchup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 28, 29, 30, 31, 32, 51, 52, 53, 54, 74, 75, 76, 97, 98, 99, 120, 121, 122], "bbox": [0.21862499999999999, 0.0, 0.40715625, 0.3572916666666667], "iscrowd": 0, "area": 13511.280550000003, "rle": {"size": [480, 640], "counts": "`dQ25h>:G9F:G9F:G9H8I7I7I7I7I7I8H8H7I8H8H7I8H9G00000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001OO1_MfFYO]9e0XGfNj8Y1^G]Ne8a1cGWN_8g1hGRNZ8l1^2O1N2N2N2O1N2N2N2O1N2N200O1O1O1O1O1O1O1O1O100O1O1O100O1O100O1O100O1O101N4M2M4L3N3L3M4M]da5"}, "label": "hellman ' s ketchup"}]}
{"id": 301591, "image": "COCO_train2014_000000301591.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hellmans ketchup\"."}], "task": "refering", "answer_template": "The \"hellmans ketchup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 28, 29, 30, 31, 32, 51, 52, 53, 54, 74, 75, 76, 97, 98, 99, 120, 121, 122], "bbox": [0.21862499999999999, 0.0, 0.40715625, 0.3572916666666667], "iscrowd": 0, "area": 13511.280550000003, "rle": {"size": [480, 640], "counts": "`dQ25h>:G9F:G9F:G9H8I7I7I7I7I7I8H8H7I8H8H7I8H9G00000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001OO1_MfFYO]9e0XGfNj8Y1^G]Ne8a1cGWN_8g1hGRNZ8l1^2O1N2N2N2O1N2N2N2O1N2N200O1O1O1O1O1O1O1O1O100O1O1O100O1O100O1O100O1O101N4M2M4L3N3L3M4M]da5"}, "label": "hellmans ketchup"}]}
{"id": 121372, "image": "COCO_train2014_000000121372.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white shirt and the brown hat\"."}], "task": "refering", "answer_template": "The \"the man in the white shirt and the brown hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 197, 198], "bbox": [0.3918125, 0.004375, 0.815578125, 0.4797083333333333], "iscrowd": 0, "area": 32816.6114, "rle": {"size": [480, 640], "counts": "Yje31m>3N1N3M2O2M2O2M2O2M3M2O2M2O2M2O2M2N3N1N3N1N3M2N3M2O2M2N3M2N3M3M2N3M2O2M2N3M2N3M2O2N1O2N1O2N1O2M2O1OM3M3N2M3N2001O1O1O001O1N2O1O1O001O1O1O1O001O4K7J5K6I6K6J5J7J1O2M2O2M2O2N1N3N2N1N3N1O2M2O2N1N3N2N1N3N1O2M201N101O001N2O001O0O2O00000O10001N1000000O1000000O2O00000O1000000000001N10000000000000001N1000000000000000O0100000O10000000O01000000O100QOo0J6O001O1O1N2O1O1O1O1N101O1O1N2O1O1O1O0O2O1O1O1O1N2O001O1O1I7E;D<E;D;F;F:O100O100O10O0100O100O100O10O10O100O100O100O001O1O1O1O1N2O1O001O1N2O1O1O1O001N2O1O100O1O1O001O1O2N3N2M2N3M3M2N3M2N3N2M2N3M3M2N3MUYg1"}, "label": "the man in the white shirt and the brown hat"}]}
{"id": 121372, "image": "COCO_train2014_000000121372.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the worker at the donut shop wears white and attends to the donut machinery\"."}], "task": "refering", "answer_template": "The \"the worker at the donut shop wears white and attends to the donut machinery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 197, 198], "bbox": [0.3918125, 0.004375, 0.815578125, 0.4797083333333333], "iscrowd": 0, "area": 32816.6114, "rle": {"size": [480, 640], "counts": "Yje31m>3N1N3M2O2M2O2M2O2M3M2O2M2O2M2O2M2N3N1N3N1N3M2N3M2O2M2N3M2N3M3M2N3M2O2M2N3M2N3M2O2N1O2N1O2N1O2M2O1OM3M3N2M3N2001O1O1O001O1N2O1O1O001O1O1O1O001O4K7J5K6I6K6J5J7J1O2M2O2M2O2N1N3N2N1N3N1O2M2O2N1N3N2N1N3N1O2M201N101O001N2O001O0O2O00000O10001N1000000O1000000O2O00000O1000000000001N10000000000000001N1000000000000000O0100000O10000000O01000000O100QOo0J6O001O1O1N2O1O1O1O1N101O1O1N2O1O1O1O0O2O1O1O1O1N2O001O1O1I7E;D<E;D;F;F:O100O100O10O0100O100O100O10O10O100O100O100O001O1O1O1O1N2O1O001O1N2O1O1O1O001N2O1O100O1O1O001O1O2N3N2M2N3M3M2N3M2N3N2M2N3M3M2N3MUYg1"}, "label": "the worker at the donut shop wears white and attends to the donut machinery"}]}
{"id": 416286, "image": "COCO_train2014_000000416286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty part of the blue plate on the left\"."}], "task": "refering", "answer_template": "The \"the empty part of the blue plate on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 94, 95, 96, 113, 114], "bbox": [0.23642, 0.19909638554216866, 0.62624, 0.539487951807229], "iscrowd": 0, "area": 11646.415250000002, "rle": {"size": [332, 500], "counts": "YgV12Y:1O2N1O2N2N1O2N1O2N1O2N2lM\\O]Je0c5\\O[Jf0d5[OZJf0f5ZOYJh0f5YOXJh0h5YOVJi0h5YOVJi0i5WOVJj0j5WOTJk0k5VOSJk0m5VOQJl0n5TOQJm0o5TOnIo0Q6ROmIP1Q6ROmIo0S6QOkIR1T6oNhIT1X6mNdIW1[6jNaIZ1^6gN_I[1a6eN\\I_1c6bNYIa1f6bNUIb1j6_NSId1l6]NPIf1P7l0O1O001N2O1O001O1OnNVIbNj6^1XI`Nh6_1\\I^Nd6b1_I[Na6e1aIYN`6e1dIXN\\6h1fIVNZ6j1iISNW6m1kIQNU6o1nInMR6R2PJlMP6S2TJiMm5W2VJfMj5Z2T10000000000O1000000000000O101O00000000000O1000000000000O100000000000000O101O00000000000O100000001O00000000001O00000O10001O00000000001O000000000O2O00000000001O0O100000001O0000000010O0001O00010O000010O00O2O0O1O2O0O101N1O101N101N1O101N2O1N2O0O2N2O1N2O0O2N2O1N2O0O2N2O1EZal1"}, "label": "the empty part of the blue plate on the left"}]}
{"id": 416286, "image": "COCO_train2014_000000416286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of tomato soup\"."}], "task": "refering", "answer_template": "The \"a bowl of tomato soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215], "bbox": [0.49389999999999995, 0.2824698795180723, 1.0, 0.9521084337349398], "iscrowd": 0, "area": 43279.7491, "rle": {"size": [332, 500], "counts": "PX`22W:4M4L4L4L3L5L4L4L4L3N3M3N2M2N3N2M2O2M2N3N1N3M2O2M2N3N1N3N1N3N2M2O2M2O2M2O2M2O2N1N3N1N3N2M2O2M2O2M2O2N1N3N1N2O1N2O0O2O1O1N101N2O1N2O0O2O1N2O001N2O1N101N2O1O0O2O1O1O0O2O1O1O0O2O1O1O0O2O1O1N101O1O1N101O001N10001O0O101O000O2O000O101O000O2O00001N10001O0O101O0O10001O0O101O000O2O00001N10001O0000001O00001O000O2O00001O00001O000000001O00000000000000000O1000001O0000000000000000000000001O0000000O10000000000000001O000000000000000000000O10O100000000O10000000000O100000000O100000O1000O100000000O10000000000O10_1aNS2mMS2mM]G"}, "label": "a bowl of tomato soup"}]}
{"id": 416286, "image": "COCO_train2014_000000416286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl of soup\"."}], "task": "refering", "answer_template": "The \"the bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215], "bbox": [0.49389999999999995, 0.2824698795180723, 1.0, 0.9521084337349398], "iscrowd": 0, "area": 43279.7491, "rle": {"size": [332, 500], "counts": "PX`22W:4M4L4L4L3L5L4L4L4L3N3M3N2M2N3N2M2O2M2N3N1N3M2O2M2N3N1N3N1N3N2M2O2M2O2M2O2M2O2N1N3N1N3N2M2O2M2O2M2O2N1N3N1N2O1N2O0O2O1O1N101N2O1N2O0O2O1N2O001N2O1N101N2O1O0O2O1O1O0O2O1O1O0O2O1O1O0O2O1O1N101O1O1N101O001N10001O0O101O000O2O000O101O000O2O00001N10001O0O101O0O10001O0O101O000O2O00001N10001O0000001O00001O000O2O00001O00001O000000001O00000000000000000O1000001O0000000000000000000000001O0000000O10000000000000001O000000000000000000000O10O100000000O10000000000O100000000O100000O1000O100000000O10000000000O10_1aNS2mMS2mM]G"}, "label": "the bowl of soup"}]}
{"id": 416286, "image": "COCO_train2014_000000416286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the foil far above the bowl of soup\"."}], "task": "refering", "answer_template": "The \"the foil far above the bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 30, 31, 32, 33, 34, 35], "bbox": [0.6809400000000001, 0.0, 1.0, 0.16515060240963855], "iscrowd": 0, "area": 7907.140149999999, "rle": {"size": [332, 500], "counts": "`W^31[::F4L5K5K5K4L2N1O001O001O00001O001O001O00001O001O001O001O001O001O1O000000001O00000000000000001O0000000000000000001O00000000000000001O00000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1"}, "label": "the foil far above the bowl of soup"}]}
{"id": 416286, "image": "COCO_train2014_000000416286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a emty blue plate\"."}], "task": "refering", "answer_template": "The \"a emty blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 36, 37], "bbox": [0.00346, 0.011716867469879518, 0.40115999999999996, 0.18879518072289156], "iscrowd": 0, "area": 7110.597949999999, "rle": {"size": [332, 500], "counts": "Pf0>a9>O1N2O0O2O001N2O0O10001N100O101O0O100O101OO010O100O10O0100O10000O010O100O10O010000000O10000000000O100000000O10000000000O100000000O1000001O000O100000000O100000000O2O000000000O100000000O10001O00O10O01O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O1O100O10000O1000000000000000000000000O100000001O0000001O00001O0O10001O0000001O0000001O0O101O0000001O0000001O00005JoWQ3"}, "label": "a emty blue plate"}]}
{"id": 416286, "image": "COCO_train2014_000000416286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the edge of an empty blue plate\"."}], "task": "refering", "answer_template": "The \"the edge of an empty blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 36, 37], "bbox": [0.00346, 0.011716867469879518, 0.40115999999999996, 0.18879518072289156], "iscrowd": 0, "area": 7110.597949999999, "rle": {"size": [332, 500], "counts": "Pf0>a9>O1N2O0O2O001N2O0O10001N100O101O0O100O101OO010O100O10O0100O10000O010O100O10O010000000O10000000000O100000000O10000000000O100000000O1000001O000O100000000O100000000O2O000000000O100000000O10001O00O10O01O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O1O100O10000O1000000000000000000000000O100000001O0000001O00001O0O10001O0000001O0000001O0O101O0000001O0000001O00005JoWQ3"}, "label": "the edge of an empty blue plate"}]}
{"id": 539167, "image": "COCO_train2014_000000539167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in red on first base\"."}], "task": "refering", "answer_template": "The \"a baseball player in red on first base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 72, 73, 74, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 214, 215, 233, 234, 237, 238, 239, 256, 257, 261, 262, 263, 278, 279, 280, 284, 285, 286], "bbox": [0.08565625, 0.16779859484777518, 0.45981250000000007, 0.8569789227166276], "iscrowd": 0, "area": 25839.815950000004, "rle": {"size": [427, 640], "counts": "SYg01Y=101N1O1O1O1O1O2O0O1O1O1O100O2O0O100O100O101N100O100O1[H[Od2d0\\M^Ob2c0\\M@b2a0]MAi0j0PKFU4CGo1oK^NY4NZOl1XLVN]41YOo1TLQNb43WOl2h0VMVOk2h0XMVOi2i0ZMULnNT2i3f1[MSLSOQ2b3l1]MPLUOQ2_3m1_MoKVOQ2\\3o1aMlKWOR2Y3R2aMiKZOR2V3U2aMfK\\OS2T3W2bMbK^OT2P3[2fNbM[1^2gN_MZ1`2iN]MX1c2jNZMW1f2kNfLSMYOS4Q4lNdLSMXOQ4U4mNaLVMVOn3Y4nN^LWMVOl3[4PO\\LWMVOj3^4QOYL`1g3eNSL\\1m3iNmKW1T4oNeKR1[4l32N3M3hMaKaKb4^4V2O1N101O1N2O1O1N2O2N2M3O1N3M2O1N2N2O1N2N2O1N2aKoGb3S8YLPHf3R8VLQHi3Q8SLRHm3o7oKTHP4n7lKUHS4b800O1O100O1O100O2N100O1O100O1O100O101N20O010O10O0100O01000O010O10O010000O10O0100O1000O0100O100O10O10O100O10000O2O1O1N2O1O0O2O1O1N2O1O1N2O1O1N2O1O0O2O000O2O001NYLRHR2n7iMXIW1g6dNoIk0P6QOUJo0k5lN[JT1d5fNbJZ1]5bNhJ_1W5\\NoJc1]81O010O01O01O010O0010OSE\\NR:e1kE\\NW:c1gE_NY:a1dEaN]:`1`EaNb:^1\\EcNe:^1WEeNi:[1UEfNm:Z1oDhNR;X1lDiNV;f10O1O2N100O2N1N3N1N2O2N1N2O2M2O2N1N2O2M2O1O2M2O2N1N3N1N3N1O2M2O2M2O2N1N3N1N3N1Oom_4"}, "label": "a baseball player in red on first base"}]}
{"id": 539167, "image": "COCO_train2014_000000539167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy on defensive team in red playing baseball\"."}], "task": "refering", "answer_template": "The \"boy on defensive team in red playing baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 72, 73, 74, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 214, 215, 233, 234, 237, 238, 239, 256, 257, 261, 262, 263, 278, 279, 280, 284, 285, 286], "bbox": [0.08565625, 0.16779859484777518, 0.45981250000000007, 0.8569789227166276], "iscrowd": 0, "area": 25839.815950000004, "rle": {"size": [427, 640], "counts": "SYg01Y=101N1O1O1O1O1O2O0O1O1O1O100O2O0O100O100O101N100O100O1[H[Od2d0\\M^Ob2c0\\M@b2a0]MAi0j0PKFU4CGo1oK^NY4NZOl1XLVN]41YOo1TLQNb43WOl2h0VMVOk2h0XMVOi2i0ZMULnNT2i3f1[MSLSOQ2b3l1]MPLUOQ2_3m1_MoKVOQ2\\3o1aMlKWOR2Y3R2aMiKZOR2V3U2aMfK\\OS2T3W2bMbK^OT2P3[2fNbM[1^2gN_MZ1`2iN]MX1c2jNZMW1f2kNfLSMYOS4Q4lNdLSMXOQ4U4mNaLVMVOn3Y4nN^LWMVOl3[4PO\\LWMVOj3^4QOYL`1g3eNSL\\1m3iNmKW1T4oNeKR1[4l32N3M3hMaKaKb4^4V2O1N101O1N2O1O1N2O2N2M3O1N3M2O1N2N2O1N2N2O1N2aKoGb3S8YLPHf3R8VLQHi3Q8SLRHm3o7oKTHP4n7lKUHS4b800O1O100O1O100O2N100O1O100O1O100O101N20O010O10O0100O01000O010O10O010000O10O0100O1000O0100O100O10O10O100O10000O2O1O1N2O1O0O2O1O1N2O1O1N2O1O1N2O1O0O2O000O2O001NYLRHR2n7iMXIW1g6dNoIk0P6QOUJo0k5lN[JT1d5fNbJZ1]5bNhJ_1W5\\NoJc1]81O010O01O01O010O0010OSE\\NR:e1kE\\NW:c1gE_NY:a1dEaN]:`1`EaNb:^1\\EcNe:^1WEeNi:[1UEfNm:Z1oDhNR;X1lDiNV;f10O1O2N100O2N1N3N1N2O2N1N2O2M2O2N1N2O2M2O1O2M2O2N1N3N1N3N1O2M2O2M2O2N1N3N1N3N1Oom_4"}, "label": "boy on defensive team in red playing baseball"}]}
{"id": 539167, "image": "COCO_train2014_000000539167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in a green shirt standing at a base\"."}], "task": "refering", "answer_template": "The \"a baseball player in a green shirt standing at a base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 36, 56, 57, 58, 59, 60, 80, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 271, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318], "bbox": [0.45326562499999995, 0.049320843091334894, 0.858671875, 0.9304918032786885], "iscrowd": 0, "area": 35720.29785000002, "rle": {"size": [427, 640], "counts": "Yoh32V=5J6M3M3M3M2N101N1O2O1N2N3M2O1N3M4M3L4M2M2N2O1N2N2O1N2N101N1O100O2O0O2O1O1O1O1O1O2N1O1O2N1O1O1O2N2N2N2N1O2N2N2O1`GUM\\6m2`IUM`6l2]IWMc6j2YIXMg6j2UIXMl6i2RIXMm6j2QIVMo6l2oHUMQ7k2mHVMS7l2kHUMT7m2jHSMW7n2gHSMX7P3eHPM\\7Q3aHPM`7Q3^HnLe7R3YHmLj7T3SHkLP8U3mGlLU82gGQ23lMY8LiGY2MkM\\8BoGd2DiM`8YOTHo2ZOiM\\9W2dFiM\\9X2cFhM]9Y2bFfM_9S3001N100O100O1O1001O00100O10O01O10O010O010O010O010O010O01O0QIYL[4g3]KbLa4^3XKkLg4U3RKSMl4m2mJ\\MR5d2fJeMX5[2aJoM]5Q2[JXNc5h1VJaNi5_1oIjNo5V1jISOT6T4O1N2O1N2O2M2O1N2O1N2O1N2N2O1N3N1N2O1N2RLRH\\2o7aMSH^2o7_MRHb2o7[MRHe2P8XMQHi2P8TMQHl2R8PMoGP3S8mLnGT3S8jLmGV3U8gLlGZ3U8cLlG^3U8_LlGb3U8[LlGf3U8WLlGk3T8RLmGo3T8oKlGR4f81O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O001O1O1O1gKSK2n4IXK5h4H]K6d4FaK8`4CfK:\\4BiK<X4@mK>T4^OQL`0o3\\OWLb0j3ZO[Lc0f3ZO_Ld0b3WOdLg0\\3VOhLi0Y3SOkLl0U3QOPMl0R3oNSMP1m2mNWMR1j2jNZMU1f2gN_MX1b2dNbMZ1a2aNcM^1^2]NgMb1[2YNjMe1W2WNmMh1U2SNoMl1R2oMSNP2o1jMVNU2k1fMZNX2i1bM\\N\\2f1_M_N`2X6M2N2N2O1N2O1N2O2N1N2O1O1O1N2O1N3N1N2O1N2O1N2N2N3M2O1N2O1N2O1O1N2O1N2O1N2O1N3N1N2O1O1N2O1N2N2N4L4LdUU1"}, "label": "a baseball player in a green shirt standing at a base"}]}
{"id": 539167, "image": "COCO_train2014_000000539167.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first baseman stands with his foot on the bag as the runner leads off\"."}], "task": "refering", "answer_template": "The \"the first baseman stands with his foot on the bag as the runner leads off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 36, 56, 57, 58, 59, 60, 80, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 271, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318], "bbox": [0.45326562499999995, 0.049320843091334894, 0.858671875, 0.9304918032786885], "iscrowd": 0, "area": 35720.29785000002, "rle": {"size": [427, 640], "counts": "Yoh32V=5J6M3M3M3M2N101N1O2O1N2N3M2O1N3M4M3L4M2M2N2O1N2N2O1N2N101N1O100O2O0O2O1O1O1O1O1O2N1O1O2N1O1O1O2N2N2N2N1O2N2N2O1`GUM\\6m2`IUM`6l2]IWMc6j2YIXMg6j2UIXMl6i2RIXMm6j2QIVMo6l2oHUMQ7k2mHVMS7l2kHUMT7m2jHSMW7n2gHSMX7P3eHPM\\7Q3aHPM`7Q3^HnLe7R3YHmLj7T3SHkLP8U3mGlLU82gGQ23lMY8LiGY2MkM\\8BoGd2DiM`8YOTHo2ZOiM\\9W2dFiM\\9X2cFhM]9Y2bFfM_9S3001N100O100O1O1001O00100O10O01O10O010O010O010O010O010O01O0QIYL[4g3]KbLa4^3XKkLg4U3RKSMl4m2mJ\\MR5d2fJeMX5[2aJoM]5Q2[JXNc5h1VJaNi5_1oIjNo5V1jISOT6T4O1N2O1N2O2M2O1N2O1N2O1N2N2O1N3N1N2O1N2RLRH\\2o7aMSH^2o7_MRHb2o7[MRHe2P8XMQHi2P8TMQHl2R8PMoGP3S8mLnGT3S8jLmGV3U8gLlGZ3U8cLlG^3U8_LlGb3U8[LlGf3U8WLlGk3T8RLmGo3T8oKlGR4f81O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O001O1O1O1gKSK2n4IXK5h4H]K6d4FaK8`4CfK:\\4BiK<X4@mK>T4^OQL`0o3\\OWLb0j3ZO[Lc0f3ZO_Ld0b3WOdLg0\\3VOhLi0Y3SOkLl0U3QOPMl0R3oNSMP1m2mNWMR1j2jNZMU1f2gN_MX1b2dNbMZ1a2aNcM^1^2]NgMb1[2YNjMe1W2WNmMh1U2SNoMl1R2oMSNP2o1jMVNU2k1fMZNX2i1bM\\N\\2f1_M_N`2X6M2N2N2O1N2O1N2O2N1N2O1O1O1N2O1N3N1N2O1N2O1N2N2N3M2O1N2O1N2O1O1N2O1N2O1N2O1N3N1N2O1O1N2O1N2N2N4L4LdUU1"}, "label": "the first baseman stands with his foot on the bag as the runner leads off"}]}
{"id": 285214, "image": "COCO_train2014_000000285214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a closed gray suitcase standing next to an open suitcase\"."}], "task": "refering", "answer_template": "The \"a closed gray suitcase standing next to an open suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0, 0.42695833333333333, 0.581453125, 0.8808958333333333], "iscrowd": 0, "area": 74395.4185, "rle": {"size": [480, 640], "counts": "i74k>8H8H8G9H8H8H8H9G8H8H8H8H8H8H8H9G8H8H8H8H8H8H8H4L1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000001O01O0000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000001O0000000001O000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000001O0000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000001O00000000000000000000000000000001O0000003M7I7I7F:@`0A?@`0A?@?A`0A?@`0@`0A?@_Zm3"}, "label": "a closed gray suitcase standing next to an open suitcase"}]}
{"id": 285214, "image": "COCO_train2014_000000285214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey closed suitcase\"."}], "task": "refering", "answer_template": "The \"a grey closed suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0, 0.42695833333333333, 0.581453125, 0.8808958333333333], "iscrowd": 0, "area": 74395.4185, "rle": {"size": [480, 640], "counts": "i74k>8H8H8G9H8H8H8H9G8H8H8H8H8H8H8H9G8H8H8H8H8H8H8H4L1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000001O01O0000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000001O0000000001O000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000001O0000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000001O00000000000000000000000000000001O0000003M7I7I7F:@`0A?@`0A?@?A`0A?@`0@`0A?@_Zm3"}, "label": "a grey closed suitcase"}]}
{"id": 285214, "image": "COCO_train2014_000000285214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey open suitcase\"."}], "task": "refering", "answer_template": "The \"a grey open suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 335, 336, 337], "bbox": [0.41965624999999995, 0.06029166666666667, 0.9977499999999999, 0.8602916666666667], "iscrowd": 0, "area": 85583.53864999999, "rle": {"size": [480, 640], "counts": "RUn3e0Z>g0XOh0YOg0XOg0ZOg0XO=C2O1N2O1N3O00000000000000000O2O0000000000000000001N1000000000000000001O000O100000000000001O00000O1000000000001O000000000O100000001O00000000000O1000002N1O1O1O1O1O1O2N1N2O1O1O1O1O2N8H9`KbILg6BZIf0o6hNRI`1W7nMkHSN1X3^7TNaH`N>[3[7bMYHoNh0^3X7QMQH]OT1a3T7`LjGK^1d3V7XLWIg3R9O1O1O1O1O1O1N2O1O1O1O1O1O1O0000000O1000O1000000000000000O100000O1000000000000000O10O100O100O10000O100O010O10000O100O100O100O10O10O100O100O100O10000O010O100O100O10000O100O10O0100O10000O100O100O10O010000O100O100O100O1000O0100O100O10000O100O10O0100O10000O100O100O100O01000O100O100bJeK6]4GeK9[4EgK:[4CfK>Z4@hK`0Y4]OiKb0X4\\OjKd0W4VOnKi0S4nNVLR1k3dN]L\\1d3[NeLe1\\3RNlLm1U3jMTMV2m2`M\\M`2d2XMeMf2\\2QMmMo2T2gLVNW3k1aL]N_3d1WLfNg3[1PLnNP4U1eKTOY4S1XKVOh4b5O0O2O0O2O0O2O0O101O0O2O0O2O0O2O001N101N101N101N10001N101N101N101N101O0O2O0O2O0O2O0O101O0O2O0O2O0O2O0O2O001N101N101N100O2O001N101N101N101N101O0O2O0O2O0O101O0O2O0O2O0O2O0O2O001N101N101N100O2O0L5I6I8I6I8I6I8I_7"}, "label": "a grey open suitcase"}]}
{"id": 285214, "image": "COCO_train2014_000000285214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"open gray suitcase with red interior\"."}], "task": "refering", "answer_template": "The \"open gray suitcase with red interior\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 335, 336, 337], "bbox": [0.41965624999999995, 0.06029166666666667, 0.9977499999999999, 0.8602916666666667], "iscrowd": 0, "area": 85583.53864999999, "rle": {"size": [480, 640], "counts": "RUn3e0Z>g0XOh0YOg0XOg0ZOg0XO=C2O1N2O1N3O00000000000000000O2O0000000000000000001N1000000000000000001O000O100000000000001O00000O1000000000001O000000000O100000001O00000000000O1000002N1O1O1O1O1O1O2N1N2O1O1O1O1O2N8H9`KbILg6BZIf0o6hNRI`1W7nMkHSN1X3^7TNaH`N>[3[7bMYHoNh0^3X7QMQH]OT1a3T7`LjGK^1d3V7XLWIg3R9O1O1O1O1O1O1N2O1O1O1O1O1O1O0000000O1000O1000000000000000O100000O1000000000000000O10O100O100O10000O100O010O10000O100O100O100O10O10O100O100O100O10000O010O100O100O10000O100O10O0100O10000O100O100O10O010000O100O100O100O1000O0100O100O10000O100O10O0100O10000O100O100O100O01000O100O100bJeK6]4GeK9[4EgK:[4CfK>Z4@hK`0Y4]OiKb0X4\\OjKd0W4VOnKi0S4nNVLR1k3dN]L\\1d3[NeLe1\\3RNlLm1U3jMTMV2m2`M\\M`2d2XMeMf2\\2QMmMo2T2gLVNW3k1aL]N_3d1WLfNg3[1PLnNP4U1eKTOY4S1XKVOh4b5O0O2O0O2O0O2O0O101O0O2O0O2O0O2O001N101N101N101N10001N101N101N101N101O0O2O0O2O0O2O0O101O0O2O0O2O0O2O0O2O001N101N101N100O2O001N101N101N101N101O0O2O0O2O0O101O0O2O0O2O0O2O0O2O001N101N101N100O2O0L5I6I8I6I8I6I8I_7"}, "label": "open gray suitcase with red interior"}]}
{"id": 432673, "image": "COCO_train2014_000000432673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat in the bowl\"."}], "task": "refering", "answer_template": "The \"the cat in the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 26, 27, 28, 29, 30, 31, 46, 47, 48, 49], "bbox": [0.44904, 0.004293333333333334, 0.7661, 0.18669333333333335], "iscrowd": 0, "area": 5481.01415, "rle": {"size": [375, 500], "counts": "b]b22d;3M2N3XO7]ELa:6\\EM9Dl9R1lESOU:W11O1O001O1O1O1O010O1O100O01O01O000001O0000000000000000000000001O000O2O000O1O100O100O2N2O0O2N2O1N2O0O10000O2O00000000POWF3g9L[F4e9I^F7b9F`F;_9_OhFa0X9\\OkFd0S:10000O010000000000000000000000001O001O01O0001O01O00010O0000001O00001O00001O000010O01O001O001O001O001O001O001O001O001O001O001O001O00001N101O00001O001O001O1N101O1O001O001N3N3MRiZ1"}, "label": "the cat in the bowl"}]}
{"id": 432673, "image": "COCO_train2014_000000432673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a striped gray cat sitting in a bowl\"."}], "task": "refering", "answer_template": "The \"a striped gray cat sitting in a bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 26, 27, 28, 29, 30, 31, 46, 47, 48, 49], "bbox": [0.44904, 0.004293333333333334, 0.7661, 0.18669333333333335], "iscrowd": 0, "area": 5481.01415, "rle": {"size": [375, 500], "counts": "b]b22d;3M2N3XO7]ELa:6\\EM9Dl9R1lESOU:W11O1O001O1O1O1O010O1O100O01O01O000001O0000000000000000000000001O000O2O000O1O100O100O2N2O0O2N2O1N2O0O10000O2O00000000POWF3g9L[F4e9I^F7b9F`F;_9_OhFa0X9\\OkFd0S:10000O010000000000000000000000001O001O01O0001O01O00010O0000001O00001O00001O000010O01O001O001O001O001O001O001O001O001O001O001O001O00001N101O00001O001O001O1N101O1O001O001N3N3MRiZ1"}, "label": "a striped gray cat sitting in a bowl"}]}
{"id": 432673, "image": "COCO_train2014_000000432673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat is taking rest on a towel\"."}], "task": "refering", "answer_template": "The \"a cat is taking rest on a towel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 176, 177, 178, 179, 188, 189, 190, 191, 192, 207, 208, 209, 210, 226, 227], "bbox": [0.4348, 0.3480533333333334, 1.0, 0.9601333333333334], "iscrowd": 0, "area": 40079.06990000001, "rle": {"size": [375, 500], "counts": "^e_2k0k:5J4M3L4M3L4L3N3M3M3L4M3M3M4K4M3M4L4K5L4L3M4K5L4L3M4K5L3M4M2M3M3M2N3N1N2N2N2N2N2O1N2N3M2N2O1N2N2N2N2O1N2N2N2N101N2N2N2N1O2O1N2N1O2N2O0O2N2N2N102N1O2M3N1O2N1N3N000O1000000O1000001O000000000000000000000000000001O0O1O100O1O1O100O1O100O1O1O100O1O2N100O1N2O1N2O1O1M3N2M3N2M3N2M3M3M4L3ZNmIcMX6Z2oI]MV6_2QJZMS6c2TJUMP6g2XJQMl5l2ZJnLi5o2^JiLf5S3WIPMm7m2QHYMn7e2PH_MP8^2nGhMQ8U2mGPNS8k2O000010O01O010O010O01O010O010O10O010O010O010O010O010O010O10O10O01O001O001O001O001O001O000000000000001O000000000000000000001O0000000000001O1O1O2N1O1O1O2N1O1O2O1N2N2N2O1NO1000001N10001O1N2O1O1O1N20O01O010O1O0010O01O010O001O10O01O010O0010O01O00100O0010O0001O01J5I\\J"}, "label": "a cat is taking rest on a towel"}]}
{"id": 432673, "image": "COCO_train2014_000000432673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat laying down on a white towel next to some keys\"."}], "task": "refering", "answer_template": "The \"a cat laying down on a white towel next to some keys\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 176, 177, 178, 179, 188, 189, 190, 191, 192, 207, 208, 209, 210, 226, 227], "bbox": [0.4348, 0.3480533333333334, 1.0, 0.9601333333333334], "iscrowd": 0, "area": 40079.06990000001, "rle": {"size": [375, 500], "counts": "^e_2k0k:5J4M3L4M3L4L3N3M3M3L4M3M3M4K4M3M4L4K5L4L3M4K5L4L3M4K5L3M4M2M3M3M2N3N1N2N2N2N2N2O1N2N3M2N2O1N2N2N2N2O1N2N2N2N101N2N2N2N1O2O1N2N1O2N2O0O2N2N2N102N1O2M3N1O2N1N3N000O1000000O1000001O000000000000000000000000000001O0O1O100O1O1O100O1O100O1O1O100O1O2N100O1N2O1N2O1O1M3N2M3N2M3N2M3M3M4L3ZNmIcMX6Z2oI]MV6_2QJZMS6c2TJUMP6g2XJQMl5l2ZJnLi5o2^JiLf5S3WIPMm7m2QHYMn7e2PH_MP8^2nGhMQ8U2mGPNS8k2O000010O01O010O010O01O010O010O10O010O010O010O010O010O010O10O10O01O001O001O001O001O001O000000000000001O000000000000000000001O0000000000001O1O1O2N1O1O1O2N1O1O2O1N2N2N2O1NO1000001N10001O1N2O1O1O1N20O01O010O1O0010O01O010O001O10O01O010O0010O01O00100O0010O0001O01J5I\\J"}, "label": "a cat laying down on a white towel next to some keys"}]}
{"id": 432673, "image": "COCO_train2014_000000432673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"salad bowl\"."}], "task": "refering", "answer_template": "The \"salad bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 28, 29, 30, 31, 32, 44, 45, 46, 47, 48, 49, 50, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85], "bbox": [0.46854, 0.09901333333333334, 0.83596, 0.35519999999999996], "iscrowd": 0, "area": 11823.6088, "rle": {"size": [375, 500], "counts": "oge23c;8H8I8G8H8H2O2N1O1O1O1O100O2O0O100O10001N100O100O100O2O0O1O1O1O101N1O1O1O1O01O0000000000001O000000000000ZOlMiGT2W8lMiGT2W8lMjGS2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2T8oMmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1S8RNmGn1S8RNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1Q8TNoGl1Q8TNPHk1P8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1P8WNPHi1P8WNPHi1P8WNQHh1o7XNQHh1o7XNQHh1o7XNQHh1P8WNPHi1P8WNPHi1P8WNPHi1P8WNPHi1P8WNPHi1P8WNPHi1Q8WNnGi1R8WNnGi1R8WNnGi1R8WNnGi1R8XNmGh1S8XNnGg1S8YNlGg1T8YNlGf1U8[NjGe1V8\\NiGd1W8\\NiGd1X8\\NgGd1Y8\\NgGd1Y8]NeGd1[8\\NeGd1[8]NcGd1]8]NaGd1`8[N_Gf1a8[N^Ge1b8\\N\\Ge1d8]NYGd1g8^NVGc1j8_NTGa1m8d00O1O1N2N3M2N2N2O1N2N3M2N2O1N2N2N3M2O1N2N2N3M3N2M3M2N3M3N2M3M2N5K6K4K5Kmnm0"}, "label": "salad bowl"}]}
{"id": 432673, "image": "COCO_train2014_000000432673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl that a cat is in\"."}], "task": "refering", "answer_template": "The \"a bowl that a cat is in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 28, 29, 30, 31, 32, 44, 45, 46, 47, 48, 49, 50, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85], "bbox": [0.46854, 0.09901333333333334, 0.83596, 0.35519999999999996], "iscrowd": 0, "area": 11823.6088, "rle": {"size": [375, 500], "counts": "oge23c;8H8I8G8H8H2O2N1O1O1O1O100O2O0O100O10001N100O100O100O2O0O1O1O1O101N1O1O1O1O01O0000000000001O000000000000ZOlMiGT2W8lMiGT2W8lMjGS2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2U8nMkGR2T8oMmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8PNmGP2S8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1T8QNlGo1S8RNmGn1S8RNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1R8SNnGm1Q8TNoGl1Q8TNPHk1P8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1Q8VNoGj1P8WNPHi1P8WNPHi1P8WNQHh1o7XNQHh1o7XNQHh1o7XNQHh1P8WNPHi1P8WNPHi1P8WNPHi1P8WNPHi1P8WNPHi1P8WNPHi1Q8WNnGi1R8WNnGi1R8WNnGi1R8WNnGi1R8XNmGh1S8XNnGg1S8YNlGg1T8YNlGf1U8[NjGe1V8\\NiGd1W8\\NiGd1X8\\NgGd1Y8\\NgGd1Y8]NeGd1[8\\NeGd1[8]NcGd1]8]NaGd1`8[N_Gf1a8[N^Ge1b8\\N\\Ge1d8]NYGd1g8^NVGc1j8_NTGa1m8d00O1O1N2N3M2N2N2O1N2N3M2N2O1N2N2N3M2O1N2N2N3M3N2M3M2N3M3N2M3M2N5K6K4K5Kmnm0"}, "label": "a bowl that a cat is in"}]}
{"id": 88609, "image": "COCO_train2014_000000088609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden chair behind the boy ' s head closest to the window\"."}], "task": "refering", "answer_template": "The \"the wooden chair behind the boy ' s head closest to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 35, 36, 37, 38, 40, 41, 42, 58, 59, 60, 64, 65, 81, 82, 83, 88, 105], "bbox": [0.535234375, 0.002458333333333333, 0.855703125, 0.25302083333333336], "iscrowd": 0, "area": 11330.473799999994, "rle": {"size": [480, 640], "counts": "jjP5a1U=?@;F:K6M2O1N2N2N3M2N2N2N2N3O000000001O000000001O000000001O000000001O000000001O0000001O00000000C=B>I7I7I7I7J6O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O10O0100O100O100O10000O100O10000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O1O1N2O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N1O2N2N2N200O100O1O101N3N2M3M3NM2J8G<E:E<E;D<E:E_R[1"}, "label": "the wooden chair behind the boy ' s head closest to the window"}]}
{"id": 88609, "image": "COCO_train2014_000000088609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair placed the farthest end of the table\"."}], "task": "refering", "answer_template": "The \"a wooden chair placed the farthest end of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 35, 36, 37, 38, 40, 41, 42, 58, 59, 60, 64, 65, 81, 82, 83, 88, 105], "bbox": [0.535234375, 0.002458333333333333, 0.855703125, 0.25302083333333336], "iscrowd": 0, "area": 11330.473799999994, "rle": {"size": [480, 640], "counts": "jjP5a1U=?@;F:K6M2O1N2N2N3M2N2N2N2N3O000000001O000000001O000000001O000000001O000000001O0000001O00000000C=B>I7I7I7I7J6O1N2O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O10O0100O100O100O10000O100O10000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O1O1N2O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N1O2N2N2N200O100O1O101N3N2M3M3NM2J8G<E:E<E;D<E:E_R[1"}, "label": "a wooden chair placed the farthest end of the table"}]}
{"id": 88609, "image": "COCO_train2014_000000088609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a wooden chair where a young sits in a booster seat eating\"."}], "task": "refering", "answer_template": "The \"the back of a wooden chair where a young sits in a booster seat eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 228, 229, 250, 251, 252, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 362, 363, 364, 365, 366, 384, 385, 386, 387, 388, 389], "bbox": [0.714234375, 0.47550000000000003, 1.0, 1.0], "iscrowd": 0, "area": 19284.99600000001, "rle": {"size": [480, 640], "counts": "oef61n>1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2OBPGnJo8P5a0O0O2N2O1DXF_Ki9]4\\FbKe9\\4^FaKd9]4`0M2O200O10O0100O10O010000O010O100O010O100O001O1O1O001O1O1O001O100O001O1O1O001M3M3M2N3M3cM]2[OaI"}, "label": "the back of a wooden chair where a young sits in a booster seat eating"}]}
{"id": 88609, "image": "COCO_train2014_000000088609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair on which a kid is sitting\"."}], "task": "refering", "answer_template": "The \"a brown chair on which a kid is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 228, 229, 250, 251, 252, 273, 274, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 362, 363, 364, 365, 366, 384, 385, 386, 387, 388, 389], "bbox": [0.714234375, 0.47550000000000003, 1.0, 1.0], "iscrowd": 0, "area": 19284.99600000001, "rle": {"size": [480, 640], "counts": "oef61n>1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2OBPGnJo8P5a0O0O2N2O1DXF_Ki9]4\\FbKe9\\4^FaKd9]4`0M2O200O10O0100O10O010000O010O100O010O100O001O1O1O001O1O1O001O100O001O1O1O001M3M3M2N3M3cM]2[OaI"}, "label": "a brown chair on which a kid is sitting"}]}
{"id": 563764, "image": "COCO_train2014_000000563764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe walking towards the right with its head bent down\"."}], "task": "refering", "answer_template": "The \"the giraffe walking towards the right with its head bent down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 309, 310, 311, 313, 314, 315, 332, 333, 334, 335, 336, 338, 355, 356, 357, 358, 359, 361, 378, 380, 381, 384, 385], "bbox": [0.43764062499999995, 0.48804166666666665, 0.9758749999999999, 1.0], "iscrowd": 0, "area": 25351.03314999999, "rle": {"size": [480, 640], "counts": "SeS42k>5J5L4L5K4L5K4M4K4XCgNd;^1WDgNe;[1XDiNd;Q1aDUO[;c0lDAP;>QEEm:9TEJi:5XENf:0[E3d:L[E5f:Q21N2O1O1N100ON10O0O2O1O0O2O0OG^M\\Da2h;65M30001O000000001O000000001O0000WDdMo:\\2gDPNW;o2N2M2N3M3N3L4L4L5L3M3M3L5L3M3L4M4L`N^FeM_9V2QGbMk8Z2dG_MY8\\2PHdMl7W2\\HgMb7U2gHgMY7T2oHjMP7Q2YIlMf6o1cImM\\6P2kInMT6P2RJmMm5Q2XJlMh5S2ZJmMd5R2`JmM_5Q2eJmM[5R2gJnMW5R2lJlMT5T2mJlMR5T2PKkMo4U2RKjMm4W2n3000O1aFkMS6U2jIPNT6P2iIUNU6k1hIZNU6h1gI^NU6c1hIbNV6^1gIgNV6Z1fImNV6T1gIoNY6Q1dIRO[6o0bITO]6m0`IVO`6j0^IgNoMXOb8R2\\IfNYNTOZ8W2ZIdNcNROS8Z2WIcNnNoNj7_2UIbNi7_1TH`No7a1nG_NU8b1fG^N]8c1_G^Nc8c1YG]Nk8c1QG^NQ9c1kF]NX9d1dF]N_9c1\\F^Ng9c1UF]Nn9d1nE^NT:b1hE`NY:a1cEaN^:o22N1O1O1O2N1O2N2N1O2O1N2N2N2N2N2N2N1O2N2N2O100O100O010O100O14L3M4L3L5L3M4L3M4L4L3M4K4M4L3M4L4L3M4LWNiFlMT9j1ZGVNb8`1lG_NQ8X1]HgN_78mFUOl1c0S78SGoNR2h0h68ZGhNX2o0Z68aGcN]2T1o58gG^Nc2Y1b59PKFn49WKFf49^KF`49eKEY4;jKES4:QLHj37[LEe3;^LBb3=cL@\\3?hL^OX3a0lL\\OT3d0PMYOo2f0UMWOk2h0ZMTOf2k0^MSOa2m0`MRO`2m0bMRO^2n0cMQO\\2o0gMPOX2P1iMoNW2Q1jMnNV2Q1lMoNS2Q1oMmNQ2S1n5O1000000O10000O10O10O10000O10O1000O10000O01000O100000O010000O1000O010000O10O1000O10000O01000O100000O010000O1000O01000000O10O10O10000O10O100000000000000000000000000001O000000000000000000000000000010O001O001O0O2N1O2N1O2N1O2N10000000lN]Bk0c=QO`BP1h=1N100O1004L3M4M2N100O10O01N1O2N1O2M2O2N2M2O1O\\h6"}, "label": "the giraffe walking towards the right with its head bent down"}]}
{"id": 563764, "image": "COCO_train2014_000000563764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe walking to the right\"."}], "task": "refering", "answer_template": "The \"a giraffe walking to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 309, 310, 311, 313, 314, 315, 332, 333, 334, 335, 336, 338, 355, 356, 357, 358, 359, 361, 378, 380, 381, 384, 385], "bbox": [0.43764062499999995, 0.48804166666666665, 0.9758749999999999, 1.0], "iscrowd": 0, "area": 25351.03314999999, "rle": {"size": [480, 640], "counts": "SeS42k>5J5L4L5K4L5K4M4K4XCgNd;^1WDgNe;[1XDiNd;Q1aDUO[;c0lDAP;>QEEm:9TEJi:5XENf:0[E3d:L[E5f:Q21N2O1O1N100ON10O0O2O1O0O2O0OG^M\\Da2h;65M30001O000000001O000000001O0000WDdMo:\\2gDPNW;o2N2M2N3M3N3L4L4L5L3M3M3L5L3M3L4M4L`N^FeM_9V2QGbMk8Z2dG_MY8\\2PHdMl7W2\\HgMb7U2gHgMY7T2oHjMP7Q2YIlMf6o1cImM\\6P2kInMT6P2RJmMm5Q2XJlMh5S2ZJmMd5R2`JmM_5Q2eJmM[5R2gJnMW5R2lJlMT5T2mJlMR5T2PKkMo4U2RKjMm4W2n3000O1aFkMS6U2jIPNT6P2iIUNU6k1hIZNU6h1gI^NU6c1hIbNV6^1gIgNV6Z1fImNV6T1gIoNY6Q1dIRO[6o0bITO]6m0`IVO`6j0^IgNoMXOb8R2\\IfNYNTOZ8W2ZIdNcNROS8Z2WIcNnNoNj7_2UIbNi7_1TH`No7a1nG_NU8b1fG^N]8c1_G^Nc8c1YG]Nk8c1QG^NQ9c1kF]NX9d1dF]N_9c1\\F^Ng9c1UF]Nn9d1nE^NT:b1hE`NY:a1cEaN^:o22N1O1O1O2N1O2N2N1O2O1N2N2N2N2N2N2N1O2N2N2O100O100O010O100O14L3M4L3L5L3M4L3M4L4L3M4K4M4L3M4L4L3M4LWNiFlMT9j1ZGVNb8`1lG_NQ8X1]HgN_78mFUOl1c0S78SGoNR2h0h68ZGhNX2o0Z68aGcN]2T1o58gG^Nc2Y1b59PKFn49WKFf49^KF`49eKEY4;jKES4:QLHj37[LEe3;^LBb3=cL@\\3?hL^OX3a0lL\\OT3d0PMYOo2f0UMWOk2h0ZMTOf2k0^MSOa2m0`MRO`2m0bMRO^2n0cMQO\\2o0gMPOX2P1iMoNW2Q1jMnNV2Q1lMoNS2Q1oMmNQ2S1n5O1000000O10000O10O10O10000O10O1000O10000O01000O100000O010000O1000O010000O10O1000O10000O01000O100000O010000O1000O01000000O10O10O10000O10O100000000000000000000000000001O000000000000000000000000000010O001O001O0O2N1O2N1O2N1O2N10000000lN]Bk0c=QO`BP1h=1N100O1004L3M4M2N100O10O01N1O2N1O2M2O2N2M2O1O\\h6"}, "label": "a giraffe walking to the right"}]}
{"id": 563764, "image": "COCO_train2014_000000563764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe walking next to a wooden fence and tall pole , in the background behind another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe walking next to a wooden fence and tall pole , in the background behind another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 118, 119, 120, 121, 143, 144, 145, 146, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 263, 283, 286], "bbox": [0.122484375, 0.25077083333333333, 0.47315625, 0.7295208333333334], "iscrowd": 0, "area": 13934.273949999995, "rle": {"size": [480, 640], "counts": "hfT11m>2O1N2N2N2N2N2M3M2O2O010O001H81N2O0O2O0O2O1N1010O001L3M41N10O0100O00NWOnAi0R>2010O10O1000O10O10000O10O10O10000O0100000O1000O010000O10O10O10O10O10O0100O01000O010O100O01000O010O10O10O10O0100O10O10ZDR1X7oNeHS1\\7lNcHU1]7lN`HV1a7jN\\HY1c7hNZHZ1g7gNVHZ1j7gNSH[1n7fNnGE`Ni0b9CkGEPO<V90fGDB0h8=dGC:[OR8d1QGPOS1XOm7l1jFkN^1VOh7Q2gFiNZ;X1cDhN^;Z1_DfNc;[1ZDeNg;\\1WDdNj;X22N1O2N1O1O2N1O1O2NWF\\MX7c2hH^MX7a2gHaMZ7]2fHdMZ7_2aHdM^7d2XH^Mh7j2oGWMR8Q3[GYMe8a401O01O0000010PM^F`0b9@gF3]9MmFCY9=f20O0001O00fEjMj7V2THlMl7T2RHoMn7P2PHRNQ8m1lGWNU8g1iG[NX8d1fG_NZ8`1cGcN^8]1_GfNa8Y1]GhNe8W1YGiNk8U1RGlNQ9S1mFnNU9Q1iFoNZ9P1cFQO`9n0^FROe9m0VFVOm9i0nEZOU:e0fE]O^:b0^E@d:n0iDUOZ;V22N3M2N3N1N3M2N3N1N3M2M4M2@a0000O100O1000000iCoM^;Q2^DTNa;l1[DZNb;g1ZD[Nh;e1TD^Nm;Y22N2O1N2N2O1N2N2O1N2N2O1N5L5N1O1gNSC:o<[OZCf0`0POT;1dDP12QOX<Q1aCROf<k0RCXOV=e0cB^O^=Q13M3N1N3N2M3M3N2M3J6Ijdm4"}, "label": "a giraffe walking next to a wooden fence and tall pole , in the background behind another giraffe"}]}
{"id": 563764, "image": "COCO_train2014_000000563764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 118, 119, 120, 121, 143, 144, 145, 146, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 263, 283, 286], "bbox": [0.122484375, 0.25077083333333333, 0.47315625, 0.7295208333333334], "iscrowd": 0, "area": 13934.273949999995, "rle": {"size": [480, 640], "counts": "hfT11m>2O1N2N2N2N2N2M3M2O2O010O001H81N2O0O2O0O2O1N1010O001L3M41N10O0100O00NWOnAi0R>2010O10O1000O10O10000O10O10O10000O0100000O1000O010000O10O10O10O10O10O0100O01000O010O100O01000O010O10O10O10O0100O10O10ZDR1X7oNeHS1\\7lNcHU1]7lN`HV1a7jN\\HY1c7hNZHZ1g7gNVHZ1j7gNSH[1n7fNnGE`Ni0b9CkGEPO<V90fGDB0h8=dGC:[OR8d1QGPOS1XOm7l1jFkN^1VOh7Q2gFiNZ;X1cDhN^;Z1_DfNc;[1ZDeNg;\\1WDdNj;X22N1O2N1O1O2N1O1O2NWF\\MX7c2hH^MX7a2gHaMZ7]2fHdMZ7_2aHdM^7d2XH^Mh7j2oGWMR8Q3[GYMe8a401O01O0000010PM^F`0b9@gF3]9MmFCY9=f20O0001O00fEjMj7V2THlMl7T2RHoMn7P2PHRNQ8m1lGWNU8g1iG[NX8d1fG_NZ8`1cGcN^8]1_GfNa8Y1]GhNe8W1YGiNk8U1RGlNQ9S1mFnNU9Q1iFoNZ9P1cFQO`9n0^FROe9m0VFVOm9i0nEZOU:e0fE]O^:b0^E@d:n0iDUOZ;V22N3M2N3N1N3M2N3N1N3M2M4M2@a0000O100O1000000iCoM^;Q2^DTNa;l1[DZNb;g1ZD[Nh;e1TD^Nm;Y22N2O1N2N2O1N2N2O1N2N2O1N5L5N1O1gNSC:o<[OZCf0`0POT;1dDP12QOX<Q1aCROf<k0RCXOV=e0cB^O^=Q13M3N1N3N2M3M3N2M3J6Ijdm4"}, "label": "the giraffe on the left in the right hand picture"}]}
{"id": 326209, "image": "COCO_train2014_000000326209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe in front of another giraffe\"."}], "task": "refering", "answer_template": "The \"giraffe in front of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 129, 130, 131, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 286, 287, 288, 289], "bbox": [0.44825, 0.18477083333333333, 0.76078125, 0.7422916666666667], "iscrowd": 0, "area": 17258.034649999994, "rle": {"size": [480, 640], "counts": "WjV43k>3L4M3L4M4UN]OREg0l:CeDd0Y;GnCi0P<U1O1Nb0^O;E4M4K4L5L3L1O10O1000O10000O1000000O100000_NaEaN^:^1lEZNT:d1UFUNk9i1_FoM`9P2iFiMW9U2RGdMn8Z2m1N1O2O1N1011O0O2O000O2O0O101N101O0O101N100O2O001N1\\EaMg8_2mElNS:Q300O01000O10O01000O0100O10O0O2K4K6J5L5J5K5L54K7J5J7I6K6I7I4M3L4M3L1ON_FfJb9S58J7POZF_Ll9^3l0M4L3M3M4L3M4L3M3PORDlNR<R1TDeNP<[1TD]NQ<a1UDWNn;i1d0O10O0100O100O010O100O010O100O10O10O100O10O0100O100O010OZO^NQDb1P<cNjC\\1Y<eNcC\\1^<dNaC[1a<eN^C[1d<eNZCZ1h<>2M2O2M10N1AYNgCh1Z<[NaCe1`<=0O2O0O101N2O3L4M3O1O2OM3CYN\\Ci1b<:1O1O001O0O2K;B`0@<J6K0O2O1N2O1N2O1O1N2O0O2O1N2O1NeRX2"}, "label": "giraffe in front of another giraffe"}]}
{"id": 326209, "image": "COCO_train2014_000000326209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe in the foreground standing in front of the other giraffe\"."}], "task": "refering", "answer_template": "The \"the giraffe in the foreground standing in front of the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 107, 108, 109, 129, 130, 131, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 263, 264, 265, 266, 286, 287, 288, 289], "bbox": [0.44825, 0.18477083333333333, 0.76078125, 0.7422916666666667], "iscrowd": 0, "area": 17258.034649999994, "rle": {"size": [480, 640], "counts": "WjV43k>3L4M3L4M4UN]OREg0l:CeDd0Y;GnCi0P<U1O1Nb0^O;E4M4K4L5L3L1O10O1000O10000O1000000O100000_NaEaN^:^1lEZNT:d1UFUNk9i1_FoM`9P2iFiMW9U2RGdMn8Z2m1N1O2O1N1011O0O2O000O2O0O101N101O0O101N100O2O001N1\\EaMg8_2mElNS:Q300O01000O10O01000O0100O10O0O2K4K6J5L5J5K5L54K7J5J7I6K6I7I4M3L4M3L1ON_FfJb9S58J7POZF_Ll9^3l0M4L3M3M4L3M4L3M3PORDlNR<R1TDeNP<[1TD]NQ<a1UDWNn;i1d0O10O0100O100O010O100O010O100O10O10O100O10O0100O100O010OZO^NQDb1P<cNjC\\1Y<eNcC\\1^<dNaC[1a<eN^C[1d<eNZCZ1h<>2M2O2M10N1AYNgCh1Z<[NaCe1`<=0O2O0O101N2O3L4M3O1O2OM3CYN\\Ci1b<:1O1O001O0O2K;B`0@<J6K0O2O1N2O1N2O1O1N2O0O2O1N2O1NeRX2"}, "label": "the giraffe in the foreground standing in front of the other giraffe"}]}
{"id": 326209, "image": "COCO_train2014_000000326209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraff standing behind another giraff\"."}], "task": "refering", "answer_template": "The \"the giraff standing behind another giraff\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 102, 103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 239, 259, 260, 262, 285], "bbox": [0.267984375, 0.20860416666666665, 0.5780937500000001, 0.7367083333333333], "iscrowd": 0, "area": 14505.302799999996, "rle": {"size": [480, 640], "counts": "hk`2e1m<a0J5M8H8G7J3M3L3N3M3M2O1N1O101N101N100O2N100OhNSE`Nl:^1eESNZ:l1^1N1O1O2N1O2N1O2N1O1O2N1O2N12N1O1O1O1O2O0O1O1O1O1O2N1O1O1O101N001O1O00100O001O1O00100O001O1O004M4K5K4L5L4K5K5K6K;D=C=D8G8H3N0O001O010O00`NQGQMo8g2h1I7I8H7I7N2O2O0O101N1gNSD@m;?YD[Oh;e0]DVOd;i0aDQO`;o0eDlN[;S1kDgNW;X1nDcNR;]1SE^Nm:b1Y1N1000001O0O101O000O2O0000001N10001O0O10001O0O101O00001N1000001N10001O000O2O00000O2O00001N102N1O1O110O01O1O0I8I71N2O1O1O1N01BTOfBm0Y==10O1O11N4M2N1OOPO\\Be0c=ZOaBc0_=\\OdBc0[=\\OiBb0V=\\OoBb0P=]OSCa0m<]OXCa0f=M2N3L4M2N3MT]n3"}, "label": "the giraff standing behind another giraff"}]}
{"id": 326209, "image": "COCO_train2014_000000326209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left giraffe\"."}], "task": "refering", "answer_template": "The \"the left giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 102, 103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 239, 259, 260, 262, 285], "bbox": [0.267984375, 0.20860416666666665, 0.5780937500000001, 0.7367083333333333], "iscrowd": 0, "area": 14505.302799999996, "rle": {"size": [480, 640], "counts": "hk`2e1m<a0J5M8H8G7J3M3L3N3M3M2O1N1O101N101N100O2N100OhNSE`Nl:^1eESNZ:l1^1N1O1O2N1O2N1O2N1O1O2N1O2N12N1O1O1O1O2O0O1O1O1O1O2N1O1O1O101N001O1O00100O001O1O00100O001O1O004M4K5K4L5L4K5K5K6K;D=C=D8G8H3N0O001O010O00`NQGQMo8g2h1I7I8H7I7N2O2O0O101N1gNSD@m;?YD[Oh;e0]DVOd;i0aDQO`;o0eDlN[;S1kDgNW;X1nDcNR;]1SE^Nm:b1Y1N1000001O0O101O000O2O0000001N10001O0O10001O0O101O00001N1000001N10001O000O2O00000O2O00001N102N1O1O110O01O1O0I8I71N2O1O1O1N01BTOfBm0Y==10O1O11N4M2N1OOPO\\Be0c=ZOaBc0_=\\OdBc0[=\\OiBb0V=\\OoBb0P=]OSCa0m<]OXCa0f=M2N3L4M2N3MT]n3"}, "label": "the left giraffe"}]}
{"id": 105026, "image": "COCO_train2014_000000105026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a grey shirt and black dress pants\"."}], "task": "refering", "answer_template": "The \"a man with a grey shirt and black dress pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 39, 40, 60, 61, 62, 63, 64, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 382, 383, 384, 385], "bbox": [0.582390625, 0.050229166666666665, 0.8329531250000001, 0.9712291666666667], "iscrowd": 0, "area": 55612.91664999999, "rle": {"size": [480, 640], "counts": "ZR_5h1T=9G[2eM6J6J7I6J7I6K5J7I6J7I7I8H8H7J7I7I6J7J5J7I6I8H7I8H7I8G8I6J6PMjEn0]:oNgEi0_:UOdEd0b:ZObE>e:_O_E9g:E\\E3i:M[EKi:5[ECi:=\\EZOh:f0_EoNe:Q1e10000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O00000000000000000000002N2N2N2N2N2N2N2N2N2N2M3M3M3M3M3M3M3M3M3M2N2O1N2N2N2N3M2N2N2L4K5K5K5J6K6J5K5K5K5K6J5K5I7J6I8H7I7J6aL_3\\Oe0ZOe0ZOf0WOi0G9I8H7F:E;EiQb1"}, "label": "a man with a grey shirt and black dress pants"}]}
{"id": 105026, "image": "COCO_train2014_000000105026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man standing closest to the pillows\"."}], "task": "refering", "answer_template": "The \"the man standing closest to the pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 37, 38, 39, 40, 60, 61, 62, 63, 64, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 382, 383, 384, 385], "bbox": [0.582390625, 0.050229166666666665, 0.8329531250000001, 0.9712291666666667], "iscrowd": 0, "area": 55612.91664999999, "rle": {"size": [480, 640], "counts": "ZR_5h1T=9G[2eM6J6J7I6J7I6K5J7I6J7I7I8H8H7J7I7I6J7J5J7I6I8H7I8H7I8G8I6J6PMjEn0]:oNgEi0_:UOdEd0b:ZObE>e:_O_E9g:E\\E3i:M[EKi:5[ECi:=\\EZOh:f0_EoNe:Q1e10000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O00000000000000000000002N2N2N2N2N2N2N2N2N2N2M3M3M3M3M3M3M3M3M3M2N2O1N2N2N2N3M2N2N2L4K5K5K5J6K6J5K5K5K5K6J5K5I7J6I8H7I7J6aL_3\\Oe0ZOe0ZOf0WOi0G9I8H7F:E;EiQb1"}, "label": "the man standing closest to the pillows"}]}
{"id": 105026, "image": "COCO_train2014_000000105026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black suit\"."}], "task": "refering", "answer_template": "The \"man in black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 54, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 350, 351, 352, 353, 373, 374, 375, 376], "bbox": [0.198875, 0.08952083333333333, 0.385953125, 0.9906458333333333], "iscrowd": 0, "area": 36582.86575, "rle": {"size": [480, 640], "counts": "Vek1n1o<;E;F:E;E;F:E;G9H8G8I8G<E`0@?@a0@`0_Oa0@`0@`0_OU1lN2M3N2M3N2M3N3L8I7H2O0O2O001O1O0O2O0000000O10O10000O01VO]DWNc;h1_DUNc;i1`DTNa;k1bDQN`;o1cDmM];S2hDgMX;Z2cNgMaGZ2W8VN`Gj1Z8fN\\GY1^8WOXGi0b8HTG7f8T3]Ob01O1dLeG6Z8VNbId1R:K3M3M4L2N1O00001O001O001O001O010O00001O000010O00100bDWNc9i1ZF\\Nd9e1YF]N[8H_Gl11`NU80hGe1K^NS89PH]1F^NQ8`0WHV1A\\NR8f0[HU1YOYNU8j0`HT2Z7UNdHj1V7^NhHc1Q7eNmH[1m6nNQIR1h6VOVIk0c6^OZIc0`6F]I;[6OaI3W66fILR6>kIDl5o4F9G<Bo0RO=B?B>B=D=B=@j0VOdPh5"}, "label": "man in black suit"}]}
{"id": 105026, "image": "COCO_train2014_000000105026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a black suit\"."}], "task": "refering", "answer_template": "The \"a guy in a black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 52, 53, 54, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 350, 351, 352, 353, 373, 374, 375, 376], "bbox": [0.198875, 0.08952083333333333, 0.385953125, 0.9906458333333333], "iscrowd": 0, "area": 36582.86575, "rle": {"size": [480, 640], "counts": "Vek1n1o<;E;F:E;E;F:E;G9H8G8I8G<E`0@?@a0@`0_Oa0@`0@`0_OU1lN2M3N2M3N2M3N3L8I7H2O0O2O001O1O0O2O0000000O10O10000O01VO]DWNc;h1_DUNc;i1`DTNa;k1bDQN`;o1cDmM];S2hDgMX;Z2cNgMaGZ2W8VN`Gj1Z8fN\\GY1^8WOXGi0b8HTG7f8T3]Ob01O1dLeG6Z8VNbId1R:K3M3M4L2N1O00001O001O001O001O010O00001O000010O00100bDWNc9i1ZF\\Nd9e1YF]N[8H_Gl11`NU80hGe1K^NS89PH]1F^NQ8`0WHV1A\\NR8f0[HU1YOYNU8j0`HT2Z7UNdHj1V7^NhHc1Q7eNmH[1m6nNQIR1h6VOVIk0c6^OZIc0`6F]I;[6OaI3W66fILR6>kIDl5o4F9G<Bo0RO=B?B>B=D=B=@j0VOdPh5"}, "label": "a guy in a black suit"}]}
{"id": 105026, "image": "COCO_train2014_000000105026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a long black dress\"."}], "task": "refering", "answer_template": "The \"a woman wearing a long black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 81, 82, 103, 104, 105, 106, 126, 127, 128, 149, 150, 151, 172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 332, 333, 334, 335, 336, 355, 356, 357, 358, 359, 379, 380, 381, 382], "bbox": [0.46753125000000006, 0.15325, 0.6331093750000001, 0.9870208333333333], "iscrowd": 0, "area": 27018.17775, "rle": {"size": [480, 640], "counts": "\\b\\43g>;D;E;F:XJdN]Mg1[2`N[Mk1]2[N[Ml1`2[NVMl1e2[NSMl1h2[NnLl1n2[NhLm1S3YNeLm1W3ZN_Lm1]3ZNZLm1a3YNVLn1f3YNPLo1k3XNlKn1o3YNhKn1T4YNcKn1X4YN^Kn1^4XNYKP2b4WNUKo1g4XNoJP2l4T4L4L5L3L5L3M4L3M3M4L3M4L3M3M4L3M4L3M4L3M3L5M201O00O1000O100000O10000000000O10000000000O101O1O2N1O2bHjNm0X1nN_O?c0ZOGa0;XO0b01YO8b0JWO`0d0AVOi0d0ZOUOP1f0QOTOY1g0iNROa1i0aNQOh1j0YNPOQ2j0RNoNX2l0iMnNa2m0aMlNi2o0XMkNR3P1PMiNZ3R1gLhNc3R1`LgNj3T1XLeNR4V1oKdN\\4V1fKdNc4W1^KcNk4Y1WK`NS5Z1_4K5K5K4L5K5J6K5K4Ldf]3"}, "label": "a woman wearing a long black dress"}]}
{"id": 105026, "image": "COCO_train2014_000000105026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the long black dress without a scarf\"."}], "task": "refering", "answer_template": "The \"the woman in the long black dress without a scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 81, 82, 103, 104, 105, 106, 126, 127, 128, 149, 150, 151, 172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 332, 333, 334, 335, 336, 355, 356, 357, 358, 359, 379, 380, 381, 382], "bbox": [0.46753125000000006, 0.15325, 0.6331093750000001, 0.9870208333333333], "iscrowd": 0, "area": 27018.17775, "rle": {"size": [480, 640], "counts": "\\b\\43g>;D;E;F:XJdN]Mg1[2`N[Mk1]2[N[Ml1`2[NVMl1e2[NSMl1h2[NnLl1n2[NhLm1S3YNeLm1W3ZN_Lm1]3ZNZLm1a3YNVLn1f3YNPLo1k3XNlKn1o3YNhKn1T4YNcKn1X4YN^Kn1^4XNYKP2b4WNUKo1g4XNoJP2l4T4L4L5L3L5L3M4L3M3M4L3M4L3M3M4L3M4L3M4L3M3L5M201O00O1000O100000O10000000000O10000000000O101O1O2N1O2bHjNm0X1nN_O?c0ZOGa0;XO0b01YO8b0JWO`0d0AVOi0d0ZOUOP1f0QOTOY1g0iNROa1i0aNQOh1j0YNPOQ2j0RNoNX2l0iMnNa2m0aMlNi2o0XMkNR3P1PMiNZ3R1gLhNc3R1`LgNj3T1XLeNR4V1oKdN\\4V1fKdNc4W1^KcNk4Y1WK`NS5Z1_4K5K5K4L5K5J6K5K4Ldf]3"}, "label": "the woman in the long black dress without a scarf"}]}
{"id": 457286, "image": "COCO_train2014_000000457286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of a person in a window\"."}], "task": "refering", "answer_template": "The \"the reflection of a person in a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 327, 328, 329, 342, 343, 344], "bbox": [0.8347072599531616, 0.66975, 1.0, 1.0], "iscrowd": 0, "area": 12730.078050000002, "rle": {"size": [640, 427], "counts": "oSo61dc0;]MJc@o0[?WOQ@Y1l?nNl_OV1]>XN^Bi0lNT1e>XNXBj0lNR1l>XNQBS3n=QMlAR3S>SMfAP3Y>UM`An2_>WMZAl2e>YMTAj2k>V1O1O1O1O1N2M3M3N2M3M3N2M3M3N2M3O1O1O1O1O1O1O1N2O1O1O100O100O100O100O10000000000001O000000001O000000001O000000001O00"}, "label": "the reflection of a person in a window"}]}
{"id": 457286, "image": "COCO_train2014_000000457286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the camera person in the reflection near the woman ' s racquet\"."}], "task": "refering", "answer_template": "The \"the camera person in the reflection near the woman ' s racquet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314, 327, 328, 329, 342, 343, 344], "bbox": [0.8347072599531616, 0.66975, 1.0, 1.0], "iscrowd": 0, "area": 12730.078050000002, "rle": {"size": [640, 427], "counts": "oSo61dc0;]MJc@o0[?WOQ@Y1l?nNl_OV1]>XN^Bi0lNT1e>XNXBj0lNR1l>XNQBS3n=QMlAR3S>SMfAP3Y>UM`An2_>WMZAl2e>YMTAj2k>V1O1O1O1O1N2M3M3N2M3M3N2M3M3N2M3O1O1O1O1O1O1O1N2O1O1O100O100O100O100O10000000000001O000000001O000000001O000000001O00"}, "label": "the camera person in the reflection near the woman ' s racquet"}]}
{"id": 252492, "image": "COCO_train2014_000000252492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white and black portion of the bag to the right of the skateboard\"."}], "task": "refering", "answer_template": "The \"the white and black portion of the bag to the right of the skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 165, 166, 186, 187, 207, 208, 227, 228, 229, 230, 249, 250, 251, 270, 271, 272, 273, 292, 293, 294, 314, 315, 316, 336, 337, 338, 359, 360, 381, 382], "bbox": [0.2821078431372549, 0.26598039215686275, 0.6519444444444444, 0.7980392156862745], "iscrowd": 0, "area": 12775.4208, "rle": {"size": [612, 612], "counts": "ZiW34mb05K4L5K5K4L5K4L5K4L5K5K4L5K4L5K4L5K5K4L5J5L5K4L5K5K4L5K4L5K4O2M3M2N3M2N3M2N3M3N1N3M2N3M2N3M3M2N2O1N1O1O2N1O1O2N10M2E;E;E;E;E;E;E;E<D;H8O1O1O100O1O1O2O0O1O100O1O1O100O2N100O1O1O100O1O1O2O0O1O100O1O1O101N1O1O100O1O100O1O2N100O1O1O100O1O101N1O1O100O1O1O100O2N100O1O100O100O2O000O2O0O101O0O101N10001N101N10001N100O2O000O2O0O101O0O101N10001N11O01O00010O000010O0001O01O01O00010O000010O00010O01O00010O000010O0001O01O01O00010O000010O000Tc00l\\O0Tc00l\\O0WV10bTo3"}, "label": "the white and black portion of the bag to the right of the skateboard"}]}
{"id": 252492, "image": "COCO_train2014_000000252492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white and blue inflatable part between the purple stripe and green skateboard\"."}], "task": "refering", "answer_template": "The \"the white and blue inflatable part between the purple stripe and green skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 165, 166, 186, 187, 207, 208, 227, 228, 229, 230, 249, 250, 251, 270, 271, 272, 273, 292, 293, 294, 314, 315, 316, 336, 337, 338, 359, 360, 381, 382], "bbox": [0.2821078431372549, 0.26598039215686275, 0.6519444444444444, 0.7980392156862745], "iscrowd": 0, "area": 12775.4208, "rle": {"size": [612, 612], "counts": "ZiW34mb05K4L5K5K4L5K4L5K4L5K5K4L5K4L5K4L5K5K4L5J5L5K4L5K5K4L5K4L5K4O2M3M2N3M2N3M2N3M3N1N3M2N3M2N3M3M2N2O1N1O1O2N1O1O2N10M2E;E;E;E;E;E;E;E<D;H8O1O1O100O1O1O2O0O1O100O1O1O100O2N100O1O1O100O1O1O2O0O1O100O1O1O101N1O1O100O1O100O1O2N100O1O1O100O1O101N1O1O100O1O1O100O2N100O1O100O100O2O000O2O0O101O0O101N10001N101N10001N100O2O000O2O0O101O0O101N10001N11O01O00010O000010O0001O01O01O00010O000010O00010O01O00010O000010O0001O01O01O00010O000010O000Tc00l\\O0Tc00l\\O0WV10bTo3"}, "label": "the white and blue inflatable part between the purple stripe and green skateboard"}]}
{"id": 252492, "image": "COCO_train2014_000000252492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfboard covered in a white carrier to the right of other surfboards\"."}], "task": "refering", "answer_template": "The \"a surfboard covered in a white carrier to the right of other surfboards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 298, 299, 300, 301, 302, 303, 304, 320, 321, 322, 323, 324, 325, 342, 343, 344, 345, 346, 347, 364, 365, 366, 367, 368, 387, 388, 389, 390, 410], "bbox": [0.561797385620915, 0.24718954248366012, 0.8876470588235295, 0.8314542483660131], "iscrowd": 0, "area": 45926.8962, "rle": {"size": [612, 612], "counts": "Ro]6:fb0=C=C=C=C>A>C=C9G5K5K5K5J6K5K5K5K5K5K5J6K5K5K5K5K5K5K5M3N2N2M3N2M3N2M3N2N2M3N2M3N1O2M2O2M2O2M2O2N1N3N1N3N1O2M2O2M2O2M2O2N1N3N1N3N000O2O0O2O000O2O0O101N101O0O101N10001O0010O00010O0010O0010O0010O01O0O101O0O2O000O2O001N101O0O101O0O2O000O2O0O2O000O2O001N101O0O101O0O2O000O2O001N101O0O2O1O1N101O1M3M3L3N3M3M3M3M3M2N3M3L4M3M2N3M3M3M3M2N3L4M3M3M3M2N3M3M3M3L3N3M3M3M3M2N3M3L4M3M3M2N3M3M3M3M2M4M3M3M3I6hNY1hNUTY1"}, "label": "a surfboard covered in a white carrier to the right of other surfboards"}]}
{"id": 252492, "image": "COCO_train2014_000000252492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the board that is in a bag\"."}], "task": "refering", "answer_template": "The \"the board that is in a bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 298, 299, 300, 301, 302, 303, 304, 320, 321, 322, 323, 324, 325, 342, 343, 344, 345, 346, 347, 364, 365, 366, 367, 368, 387, 388, 389, 390, 410], "bbox": [0.561797385620915, 0.24718954248366012, 0.8876470588235295, 0.8314542483660131], "iscrowd": 0, "area": 45926.8962, "rle": {"size": [612, 612], "counts": "Ro]6:fb0=C=C=C=C>A>C=C9G5K5K5K5J6K5K5K5K5K5K5J6K5K5K5K5K5K5K5M3N2N2M3N2M3N2M3N2N2M3N2M3N1O2M2O2M2O2M2O2N1N3N1N3N1O2M2O2M2O2M2O2N1N3N1N3N000O2O0O2O000O2O0O101N101O0O101N10001O0010O00010O0010O0010O0010O01O0O101O0O2O000O2O001N101O0O101O0O2O000O2O0O2O000O2O001N101O0O101O0O2O000O2O001N101O0O2O1O1N101O1M3M3L3N3M3M3M3M3M2N3M3L4M3M2N3M3M3M3M2N3L4M3M3M3M2N3M3M3M3L3N3M3M3M3M2N3M3L4M3M3M2N3M3M3M3M2M4M3M3M3I6hNY1hNUTY1"}, "label": "the board that is in a bag"}]}
{"id": 88653, "image": "COCO_train2014_000000088653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with beard wearing blue shirt with his friend\"."}], "task": "refering", "answer_template": "The \"a man with beard wearing blue shirt with his friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.210203125, 0.327625, 0.8391249999999999, 0.9845624999999999], "iscrowd": 0, "area": 79975.6963, "rle": {"size": [480, 640], "counts": "dao1<^>=C=B>C=B=D=C=B>C=B>C<D=H8H3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2O1O1N2O1O1N3N1O1N2O1O1O1N2O1O1N101O0O2O1O0O2O001N2O001N101O1O0O2O001N2O001N101O0O2O1O0O2O001O1O001O001O1O001O001O1O001O001O1O001O001O1O0fJkFi4V9UKkFk4U9SKmFm4T9QKmFo4T9nJnFR5R9mJoFS5R9jJPGV5P9iJQGW5\\90000000001O00000000000000000000000000000000000000000O1000000000000000000000000000000000000001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O001O1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O3M4L3M3M4L3M3M4L3M3M4L3M3M4L3N2M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3Ma0_Oh0XO<D00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O001O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N101N1O2N1O1O2N1O1O2N1O1O2NXj_1"}, "label": "a man with beard wearing blue shirt with his friend"}]}
{"id": 88653, "image": "COCO_train2014_000000088653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a beard\"."}], "task": "refering", "answer_template": "The \"a man with a beard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.210203125, 0.327625, 0.8391249999999999, 0.9845624999999999], "iscrowd": 0, "area": 79975.6963, "rle": {"size": [480, 640], "counts": "dao1<^>=C=B>C=B=D=C=B>C=B>C<D=H8H3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2O1O1N2O1O1N3N1O1N2O1O1O1N2O1O1N101O0O2O1O0O2O001N2O001N101O1O0O2O001N2O001N101O0O2O1O0O2O001O1O001O001O1O001O001O1O001O001O1O001O001O1O0fJkFi4V9UKkFk4U9SKmFm4T9QKmFo4T9nJnFR5R9mJoFS5R9jJPGV5P9iJQGW5\\90000000001O00000000000000000000000000000000000000000O1000000000000000000000000000000000000001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O001O1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O3M4L3M3M4L3M3M4L3M3M4L3M3M4L3N2M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3Ma0_Oh0XO<D00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O001O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N101N1O2N1O1O2N1O1O2N1O1O2NXj_1"}, "label": "a man with a beard"}]}
{"id": 88653, "image": "COCO_train2014_000000088653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses and a toothbrush in his mouth\"."}], "task": "refering", "answer_template": "The \"a man with glasses and a toothbrush in his mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.580703125, 0.26579166666666665, 1.0, 0.9881458333333334], "iscrowd": 0, "area": 58781.86455000001, "rle": {"size": [480, 640], "counts": "\\b^57e>7H8I7K5K5J6K5K5J5L5K5K5J6bCiMl;h2L3N2M4M2N3L3N3L3N3M2N2N2N2O1N2N2N2O1N2WGcKl6_4SIeKh6]4WIgKe6Z4YIkKb6W4]ImK^6U4aIoK[6S4bIRLY6P4fITLV6n3hITLV6m3iISLV6o3hISLW6o3gIQLX6Q4gIoKY6S4dInK[6T4dImK[6T4dIlK\\6V4aIkK^6W4aIjK^6X4`IhK_6Y4`IhK`6Y4_IhK_6Z4`IfK`6[4_IeK`6\\4_IeKa6\\4^IeKa6\\4^IdKa6^4]IcKc6]4]IdKa6^4^IbKb6_4\\IbKc6`4\\IaKc6`4\\I`Kc6a4]I_Kc6b4[I_Ke6a4[I`Kc6b4\\I^Kd6b4[I_Kd6c4[I^Kd6b4\\I^Kd6c4ZI^Kf6b4ZI^Kf6c4YI^Kg6a4XI`Kh6a4WI_Kj6`4VIaKi6`4VI`Kj6`4UIaKl6_4SIbKl6^4TIbKm6X600010O0010O00010O0001O010O00010O0010O0001O010O00010O00010O001O02O0O2N100O2O0O2N100O2O0O2O0O1O2O0O2O0O1O2O0O2O0O101N1O2O0O2O0O1O2O0O2O0O101N1101N1O101N100O1O2O0O101N1O100O3L5K4L4L4L4L2N2N2N2N3M2N2N2N2N3M2N3M3M3M3M2N3M3M00O1000000O10000O1000000O1000000O10000O10000O100O10000O10000O100O1001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O00001O0000000000000nNR1^LVK"}, "label": "a man with glasses and a toothbrush in his mouth"}]}
{"id": 88653, "image": "COCO_train2014_000000088653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with short hair and glasses holding something in his mouth\"."}], "task": "refering", "answer_template": "The \"a man with short hair and glasses holding something in his mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.580703125, 0.26579166666666665, 1.0, 0.9881458333333334], "iscrowd": 0, "area": 58781.86455000001, "rle": {"size": [480, 640], "counts": "\\b^57e>7H8I7K5K5J6K5K5J5L5K5K5J6bCiMl;h2L3N2M4M2N3L3N3L3N3M2N2N2N2O1N2N2N2O1N2WGcKl6_4SIeKh6]4WIgKe6Z4YIkKb6W4]ImK^6U4aIoK[6S4bIRLY6P4fITLV6n3hITLV6m3iISLV6o3hISLW6o3gIQLX6Q4gIoKY6S4dInK[6T4dImK[6T4dIlK\\6V4aIkK^6W4aIjK^6X4`IhK_6Y4`IhK`6Y4_IhK_6Z4`IfK`6[4_IeK`6\\4_IeKa6\\4^IeKa6\\4^IdKa6^4]IcKc6]4]IdKa6^4^IbKb6_4\\IbKc6`4\\IaKc6`4\\I`Kc6a4]I_Kc6b4[I_Ke6a4[I`Kc6b4\\I^Kd6b4[I_Kd6c4[I^Kd6b4\\I^Kd6c4ZI^Kf6b4ZI^Kf6c4YI^Kg6a4XI`Kh6a4WI_Kj6`4VIaKi6`4VI`Kj6`4UIaKl6_4SIbKl6^4TIbKm6X600010O0010O00010O0001O010O00010O0010O0001O010O00010O00010O001O02O0O2N100O2O0O2N100O2O0O2O0O1O2O0O2O0O1O2O0O2O0O101N1O2O0O2O0O1O2O0O2O0O101N1101N1O101N100O1O2O0O101N1O100O3L5K4L4L4L4L2N2N2N2N3M2N2N2N2N3M2N3M3M3M3M2N3M3M00O1000000O10000O1000000O1000000O10000O10000O100O10000O10000O100O1001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O00001O0000000000000nNR1^LVK"}, "label": "a man with short hair and glasses holding something in his mouth"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue tray with white stuff\"."}], "task": "refering", "answer_template": "The \"blue tray with white stuff\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169], "bbox": [0.150109375, 0.18451764705882354, 0.54659375, 0.5186588235294117], "iscrowd": 0, "area": 26289.636349999997, "rle": {"size": [425, 640], "counts": "gnW15R=:F:F9H8H7J7I6J6I8I6J4K4M2N3M3L4M1O1N101O1O0O2O001N101O1O0O2O001O0O10000O10O1000O1000O01000000O10O100000O10O10000000O0100000000O01000000000O010000000O10O100000O1000O100000O1000O1000O100000O10O10000000O01000000000O0100000000O010000000O010000000O10O10000000O10O100000O1000O1000O100000O1000O100000O10O10000000O0100000000O01000000000O010000000O10O100000O10000000000O100000000O100000000O100000000O101O0000000O2O0000000O2O0000000O2O2N2N1O2M3N1O2N2N1O2M3N1O2N2N1N3N2N1O2N2M2O2N2N1O2N2M2O2N2N1N3M3M2N3M3L7J6J7I6J6J\\Vh3"}, "label": "blue tray with white stuff"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl with some food and white cream in it\"."}], "task": "refering", "answer_template": "The \"a bowl with some food and white cream in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169], "bbox": [0.150109375, 0.18451764705882354, 0.54659375, 0.5186588235294117], "iscrowd": 0, "area": 26289.636349999997, "rle": {"size": [425, 640], "counts": "gnW15R=:F:F9H8H7J7I6J6I8I6J4K4M2N3M3L4M1O1N101O1O0O2O001N101O1O0O2O001O0O10000O10O1000O1000O01000000O10O100000O10O10000000O0100000000O01000000000O010000000O10O100000O1000O100000O1000O1000O100000O10O10000000O01000000000O0100000000O010000000O010000000O10O10000000O10O100000O1000O1000O100000O1000O100000O10O10000000O0100000000O01000000000O010000000O10O100000O10000000000O100000000O100000000O100000000O101O0000000O2O0000000O2O0000000O2O2N2N1O2M3N1O2N2N1O2M3N1O2N2N1N3N2N1O2N2M2O2N2N1O2N2M2O2N2N1N3M3M2N3M3L7J6J7I6J6J\\Vh3"}, "label": "a bowl with some food and white cream in it"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small tomato section on tray\"."}], "task": "refering", "answer_template": "The \"small tomato section on tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151], "bbox": [0.5017343750000001, 0.13258823529411765, 0.741125, 0.4412235294117647], "iscrowd": 0, "area": 12183.831699999999, "rle": {"size": [425, 640], "counts": "`YU41W=4L4L4L4L<E3M2M3N3M2N2N3M2N3L3N2N3M2N2N3M2N2M4M2N3M2N2N3M3M>A=D001O000000000O101O001O00002M2O1O1OO1N2O1O1N101N2O1N2O1O1N2O0O2O100000000O10O100000O100000000O100000000O1000O1000O2O001O000O2O00001O0O101O000O2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1N3N1O2N2M2N3N1N3M3M2N3MZhT2"}, "label": "small tomato section on tray"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue tub of red tomatoes\"."}], "task": "refering", "answer_template": "The \"a blue tub of red tomatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151], "bbox": [0.5017343750000001, 0.13258823529411765, 0.741125, 0.4412235294117647], "iscrowd": 0, "area": 12183.831699999999, "rle": {"size": [425, 640], "counts": "`YU41W=4L4L4L4L<E3M2M3N3M2N2N3M2N3L3N2N3M2N2N3M2N2M4M2N3M2N2N3M3M>A=D001O000000000O101O001O00002M2O1O1OO1N2O1O1N101N2O1N2O1O1N2O0O2O100000000O10O100000O100000000O100000000O1000O1000O2O001O000O2O00001O0O101O000O2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1N3N1O2N2M2N3N1N3M3M2N3MZhT2"}, "label": "a blue tub of red tomatoes"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"largest portion of food to the left of the carrots\"."}], "task": "refering", "answer_template": "The \"largest portion of food to the left of the carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 327], "bbox": [0.1673125, 0.41818823529411764, 0.6640625, 0.9506588235294117], "iscrowd": 0, "area": 47798.9004, "rle": {"size": [425, 640], "counts": "id\\14Q=9GHZC9^<EfCa0Q<]OUDi0f;XOXDm0d;TOYDR1b;PO\\DU1`;lN^DZ1];gNaD^1[;<K5L5K4L4K6K4L4L4L5J5L4L4L5K4K5L5K4L4K5L5K4L3M1N3N1O1O2N1N3N1O1O2M2O2N1O00O0100000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O010000000O0100000O10O1000O10O1000O1000O10O100000O0100000O01000000O0100000O10O1000O10O1000O1000O10O100000O0100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O100000O0100000O01000000O0100000H8I61000O10O1000O10000O2O000O2O001O0O101O0O2O2N2N2M3N2N2M2O2N2N2M3N2N2M3N2N2N2M2O2N2M3N2N2N2M3N2N2M3N2N1O2M3N2N2M3N2N2N2M3N2N1N3N2N2N2M3N2N2M3N2N2N2M2O2N2M3N2N2M3N2K5J6K5JSnh2"}, "label": "largest portion of food to the left of the carrots"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food item made with red chillies\"."}], "task": "refering", "answer_template": "The \"a food item made with red chillies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 303, 304, 305, 306, 307, 308, 309, 327], "bbox": [0.1673125, 0.41818823529411764, 0.6640625, 0.9506588235294117], "iscrowd": 0, "area": 47798.9004, "rle": {"size": [425, 640], "counts": "id\\14Q=9GHZC9^<EfCa0Q<]OUDi0f;XOXDm0d;TOYDR1b;PO\\DU1`;lN^DZ1];gNaD^1[;<K5L5K4L4K6K4L4L4L5J5L4L4L5K4K5L5K4L4K5L5K4L3M1N3N1O1O2N1N3N1O1O2M2O2N1O00O0100000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O010000000O0100000O10O1000O10O1000O1000O10O100000O0100000O01000000O0100000O10O1000O10O1000O1000O10O100000O0100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O100000O01000000O0100000O0100000O10O1000O1000O10O1000O10O100000O0100000O01000000O0100000H8I61000O10O1000O10000O2O000O2O001O0O101O0O2O2N2N2M3N2N2M2O2N2N2M3N2N2M3N2N2N2M2O2N2M3N2N2N2M3N2N2M3N2N1O2M3N2N2M3N2N2N2M3N2N1N3N2N2N2M3N2N2M3N2N2N2M2O2N2M3N2N2M3N2K5J6K5JSnh2"}, "label": "a food item made with red chillies"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrots\"."}], "task": "refering", "answer_template": "The \"carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269], "bbox": [0.579234375, 0.36272941176470586, 0.8567968749999999, 0.756164705882353], "iscrowd": 0, "area": 19341.7538, "rle": {"size": [425, 640], "counts": "[Uj42T=4F;K6L4L3M4L3M4M2N3M2M4M2N2N1N3N2N2M3N2N2N2M3N2N2N2N2M3N2N1O2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2M3N;E1O1O0O2O1O1O001N100000000O10O100000O100000001N2O1O1O1O2M2O000000O10O10O10000O10000O01000O10000O2O001N101O0O101O0O2O001N2O1O1N2O1O1N3N1O2M2O1O2M2O2N1N3N2N1N3N1O2M3N1O2M3N1O2M2O2N2M2O2N1N3N1O2M2O2N1N3N1O2L4M2N3M2N3M2N3M3M2N3M2N3M3N1N3M3N1N3M3M2N3L4M5IimU1"}, "label": "carrots"}]}
{"id": 547411, "image": "COCO_train2014_000000547411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange food lying in it ' s own container next to the green leafy vegetables\"."}], "task": "refering", "answer_template": "The \"an orange food lying in it ' s own container next to the green leafy vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269], "bbox": [0.579234375, 0.36272941176470586, 0.8567968749999999, 0.756164705882353], "iscrowd": 0, "area": 19341.7538, "rle": {"size": [425, 640], "counts": "[Uj42T=4F;K6L4L3M4L3M4M2N3M2M4M2N2N1N3N2N2M3N2N2N2M3N2N2N2N2M3N2N1O2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2M3N;E1O1O0O2O1O1O001N100000000O10O100000O100000001N2O1O1O1O2M2O000000O10O10O10000O10000O01000O10000O2O001N101O0O101O0O2O001N2O1O1N2O1O1N3N1O2M2O1O2M2O2N1N3N2N1N3N1O2M3N1O2M3N1O2M2O2N2M2O2N1N3N1O2M2O2N1N3N1O2L4M2N3M2N3M2N3M3M2N3M2N3M3N1N3M3N1N3M3M2N3L4M5IimU1"}, "label": "an orange food lying in it ' s own container next to the green leafy vegetables"}]}
{"id": 375380, "image": "COCO_train2014_000000375380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and dark gray pants\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and dark gray pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 97, 98, 99, 114, 115, 116, 117, 118, 132, 133, 134, 135, 150, 151, 152, 153, 168, 169, 170, 171, 186, 187, 188, 189, 205, 206, 207, 223, 224, 225], "bbox": [0.34374, 0.31565333333333334, 0.5704600000000001, 0.9935466666666668], "iscrowd": 0, "area": 17076.311, "rle": {"size": [375, 500], "counts": "hWo1`0`:i0G9I6J7I7I7J5J7I6J6J7I6N2N2O2M4M3L4M4K6K5J6K5J6K4K6K5K5K5J6K5K2N000O10O10O100O10000O100@gI^K]6Y4jIbK]6T4h0M3O1O1O1O1O100O1O1OjNSMdIl2h6k0=D;D2O00000001O01O0001O001O001O1O1aInKX5[4ZJiKf5U5O001O1O001O0nNbKkJ_4i4nKVKR4d4TL[Kl3a4h1XMPJBU6;^JPOh5m0jJ_NZ5^1b2L4L3M4L3M4M3N1O2N1O2M3N1O2N2N2N1O2N2N2N2Nga^2"}, "label": "a man in a white shirt and dark gray pants"}]}
{"id": 375380, "image": "COCO_train2014_000000375380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in tie and white shirt\"."}], "task": "refering", "answer_template": "The \"man in tie and white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 97, 98, 99, 114, 115, 116, 117, 118, 132, 133, 134, 135, 150, 151, 152, 153, 168, 169, 170, 171, 186, 187, 188, 189, 205, 206, 207, 223, 224, 225], "bbox": [0.34374, 0.31565333333333334, 0.5704600000000001, 0.9935466666666668], "iscrowd": 0, "area": 17076.311, "rle": {"size": [375, 500], "counts": "hWo1`0`:i0G9I6J7I7I7J5J7I6J6J7I6N2N2O2M4M3L4M4K6K5J6K5J6K4K6K5K5K5J6K5K2N000O10O10O100O10000O100@gI^K]6Y4jIbK]6T4h0M3O1O1O1O1O100O1O1OjNSMdIl2h6k0=D;D2O00000001O01O0001O001O001O1O1aInKX5[4ZJiKf5U5O001O1O001O0nNbKkJ_4i4nKVKR4d4TL[Kl3a4h1XMPJBU6;^JPOh5m0jJ_NZ5^1b2L4L3M4L3M4M3N1O2N1O2M3N1O2N2N2N1O2N2N2N2Nga^2"}, "label": "man in tie and white shirt"}]}
{"id": 375380, "image": "COCO_train2014_000000375380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a person wearing blue shirt posing for a photo along with another person\"."}], "task": "refering", "answer_template": "The \"there is a person wearing blue shirt posing for a photo along with another person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 100, 101, 102, 117, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 156, 171, 172, 173, 174, 189, 190, 191, 192, 207, 208, 209, 210, 226, 227, 228], "bbox": [0.52538, 0.31808000000000003, 0.7165799999999999, 0.9888533333333334], "iscrowd": 0, "area": 17068.473899999997, "rle": {"size": [375, 500], "counts": "[aP35^;6J7I6J6K5J7I7I8nF^Nc7i1VHeN\\7b1]HlNW7X1cHVOQ7n0jH_Oj6e0PIJc6P3D;E2O1O1O100O1O1O1O1O1O1O1lNTJ\\Lm5a3XJ\\Li5a3\\J\\Le5_3cJ^L]5a3fJ\\L\\5b3hJZLY5d3kJYLV5f3nJVLS5i3PKTLQ5k3^100000000000000000000001O001O001O001O002N1O2N1O1O2N2N3mHmK\\6W4\\ImKe6c4O1O1O001O1O001O1O0L5J6J5J7J6J6J6I7J6J6J7I6RNhHUO`7b0QIoNV7h0ZIjNm6n0cIcNc6U1U2G9F9E`jc1"}, "label": "there is a person wearing blue shirt posing for a photo along with another person"}]}
{"id": 375380, "image": "COCO_train2014_000000375380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt\"."}], "task": "refering", "answer_template": "The \"man in blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 100, 101, 102, 117, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 156, 171, 172, 173, 174, 189, 190, 191, 192, 207, 208, 209, 210, 226, 227, 228], "bbox": [0.52538, 0.31808000000000003, 0.7165799999999999, 0.9888533333333334], "iscrowd": 0, "area": 17068.473899999997, "rle": {"size": [375, 500], "counts": "[aP35^;6J7I6J6K5J7I7I8nF^Nc7i1VHeN\\7b1]HlNW7X1cHVOQ7n0jH_Oj6e0PIJc6P3D;E2O1O1O100O1O1O1O1O1O1O1lNTJ\\Lm5a3XJ\\Li5a3\\J\\Le5_3cJ^L]5a3fJ\\L\\5b3hJZLY5d3kJYLV5f3nJVLS5i3PKTLQ5k3^100000000000000000000001O001O001O001O002N1O2N1O1O2N2N3mHmK\\6W4\\ImKe6c4O1O1O001O1O001O1O0L5J6J5J7J6J6J6I7J6J6J7I6RNhHUO`7b0QIoNV7h0ZIjNm6n0cIcNc6U1U2G9F9E`jc1"}, "label": "man in blue shirt"}]}
{"id": 383576, "image": "COCO_train2014_000000383576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black couch positioned in front of the tv\"."}], "task": "refering", "answer_template": "The \"a black couch positioned in front of the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326], "bbox": [0.0014375, 0.4473004694835681, 0.323515625, 0.9849530516431926], "iscrowd": 0, "area": 37092.55795, "rle": {"size": [426, 640], "counts": "gd0d5d72O1O1O1O10000O100O100O10000O100O10000O100O100O10001N100O100O10000O100O10000O100O100O10000O100O10000O1000000000000000000O1O1O2N1O1O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O2O000O01O010O1O1O1O1O1O1O2N1O1O1O1O2N1O2N101O1N101O002M5L3M4L4K2O00000O101O0O100O100O1O10000001OO1N2N2N2N2N3M2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N3M2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N6J8H8H8H9G>B>Blkc5"}, "label": "a black couch positioned in front of the tv"}]}
{"id": 383576, "image": "COCO_train2014_000000383576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather sofa facing the television\"."}], "task": "refering", "answer_template": "The \"a leather sofa facing the television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326], "bbox": [0.0014375, 0.4473004694835681, 0.323515625, 0.9849530516431926], "iscrowd": 0, "area": 37092.55795, "rle": {"size": [426, 640], "counts": "gd0d5d72O1O1O1O10000O100O100O10000O100O10000O100O100O10001N100O100O10000O100O10000O100O100O10000O100O10000O1000000000000000000O1O1O2N1O1O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O2O000O01O010O1O1O1O1O1O1O2N1O1O1O1O2N1O2N101O1N101O002M5L3M4L4K2O00000O101O0O100O100O1O10000001OO1N2N2N2N2N3M2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N3M2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N6J8H8H8H9G>B>Blkc5"}, "label": "a leather sofa facing the television"}]}
{"id": 383576, "image": "COCO_train2014_000000383576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather couch near two windows\"."}], "task": "refering", "answer_template": "The \"a black leather couch near two windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 238], "bbox": [0.3120625, 0.404906103286385, 0.649875, 0.675868544600939], "iscrowd": 0, "area": 15131.249400000002, "rle": {"size": [426, 640], "counts": "W]c25T=6J7H7J7I6J6I8I6J6I8I6J7H2O1O1O2M2O1O1N2O1O1AYMYFi2h9<01O0000000000000000010O000000O1O1O1N2O1N2O2M2M3M3M4L3M3L4M4L3M30001O0000001O0000001O0000001O0000001O0000001O0000001O000000000000000000000000O100000000000000000000O1000000000000000000000000O100000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N1O1O2N1O2N1O2N1O2N1O2N1G:F9F;K4O2M3N1O2M2O2M3NVol2"}, "label": "a black leather couch near two windows"}]}
{"id": 383576, "image": "COCO_train2014_000000383576.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather loveseat against the wall with windows\"."}], "task": "refering", "answer_template": "The \"a black leather loveseat against the wall with windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 238], "bbox": [0.3120625, 0.404906103286385, 0.649875, 0.675868544600939], "iscrowd": 0, "area": 15131.249400000002, "rle": {"size": [426, 640], "counts": "W]c25T=6J7H7J7I6J6I8I6J6I8I6J7H2O1O1O2M2O1O1N2O1O1AYMYFi2h9<01O0000000000000000010O000000O1O1O1N2O1N2O2M2M3M3M4L3M3L4M4L3M30001O0000001O0000001O0000001O0000001O0000001O0000001O000000000000000000000000O100000000000000000000O1000000000000000000000000O100000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002N1O1O2N1O2N1O2N1O2N1O2N1G:F9F;K4O2M3N1O2M2O2M3NVol2"}, "label": "a black leather loveseat against the wall with windows"}]}
{"id": 96859, "image": "COCO_train2014_000000096859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white sweatshirt\"."}], "task": "refering", "answer_template": "The \"a man in a white sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 213, 214, 215, 216, 217, 230, 231, 232, 233, 238, 239], "bbox": [0.0031406249999999998, 0.17509345794392522, 0.476078125, 0.6954906542056074], "iscrowd": 0, "area": 28733.30415, "rle": {"size": [428, 640], "counts": "PT12V=3O200O1O100O1O010O1O1O100O1O10O01M3M3L4L4L4M2M4L4M3L4L4O1O1O001O1O1O1O100O1O00100O10000O100O10000O10O0100O10000O100O10000000000O1000O1000O100000000O100O100O100O1O10O0100O100O1O100O10eNmE]OR:`0[FWOe9i0cFoN\\9R1^1010DnN]DS1b;<0O10O1000001N10000eNbNgF_1Q9mNkFT1S9QOiFo0V9VOfFj0Y9XOeFi0Z9YOeFg0Z9[OeFf0\\8lNYG`09e0\\8SOSG;`0b0\\8=cGC[8`0cGA[8b0dG_OZ8c0dG^OZ8e0eG[OY8h0eGYOZ8j0dGVOZ8m0eGTOX8o0fGROY8P1fGPOZ8R1cGoN]8P3O2O000O10001N10000O101O000O2O000O101O0O101O0O101O0O10001N10001N10000O2O00010O0001O010O00010O001O01O01O01O01O010O01O010000O10000O1000000O10000O10001O1N3N1O2kJbHW4X8[LWGj2g8UM`Gg2^8WMjGe2T8ZMRHc2l7[MZHc2d7\\M^He2`7YMdHg2[7WMgHj2]7nLgHR3i8100O2O0O2O1N01O001O0O2O001O001N10001O001O0O2O001O001N101O001O001N101O001O001N101O001O0O2O001O001O001N101O001O001O0O2QNnDb1R;UNZEh1W;N3M3M3M3M3M3L4M3M3M3M3M2N3M1O1O1O1O1O001O1O1M2O2N[U\\4"}, "label": "a man in a white sweatshirt"}]}
{"id": 96859, "image": "COCO_train2014_000000096859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy without glasses\"."}], "task": "refering", "answer_template": "The \"boy without glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 76, 96, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 213, 214, 215, 216, 217, 230, 231, 232, 233, 238, 239], "bbox": [0.0031406249999999998, 0.17509345794392522, 0.476078125, 0.6954906542056074], "iscrowd": 0, "area": 28733.30415, "rle": {"size": [428, 640], "counts": "PT12V=3O200O1O100O1O010O1O1O100O1O10O01M3M3L4L4L4M2M4L4M3L4L4O1O1O001O1O1O1O100O1O00100O10000O100O10000O10O0100O10000O100O10000000000O1000O1000O100000000O100O100O100O1O10O0100O100O1O100O10eNmE]OR:`0[FWOe9i0cFoN\\9R1^1010DnN]DS1b;<0O10O1000001N10000eNbNgF_1Q9mNkFT1S9QOiFo0V9VOfFj0Y9XOeFi0Z9YOeFg0Z9[OeFf0\\8lNYG`09e0\\8SOSG;`0b0\\8=cGC[8`0cGA[8b0dG_OZ8c0dG^OZ8e0eG[OY8h0eGYOZ8j0dGVOZ8m0eGTOX8o0fGROY8P1fGPOZ8R1cGoN]8P3O2O000O10001N10000O101O000O2O000O101O0O101O0O101O0O10001N10001N10000O2O00010O0001O010O00010O001O01O01O01O01O010O01O010000O10000O1000000O10000O10001O1N3N1O2kJbHW4X8[LWGj2g8UM`Gg2^8WMjGe2T8ZMRHc2l7[MZHc2d7\\M^He2`7YMdHg2[7WMgHj2]7nLgHR3i8100O2O0O2O1N01O001O0O2O001O001N10001O001O0O2O001O001N101O001O001N101O001O001N101O001O0O2O001O001O001N101O001O001O0O2QNnDb1R;UNZEh1W;N3M3M3M3M3M3L4M3M3M3M3M2N3M1O1O1O1O1O001O1O1M2O2N[U\\4"}, "label": "boy without glasses"}]}
{"id": 96859, "image": "COCO_train2014_000000096859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses and a very blurry hand\"."}], "task": "refering", "answer_template": "The \"a man with glasses and a very blurry hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.48910937499999996, 0.1055373831775701, 0.9736874999999999, 0.9813785046728972], "iscrowd": 0, "area": 74757.04905, "rle": {"size": [428, 640], "counts": "fSS43S=6I7J6J6J6I7O11OO1O1N2N2N2O1N2N2N2O1N2N2O1N2N2N2O1^KnMhMT2T2VNdMl1X2^NaMc1[2gN]M[1_2oNYMS1c2VOVMl0f2[OVMf0h2@RMb0n2AmLa0R3DhL>X3FcL;]3H^L:a3KYL7g3LUL5j30PL2P44hKNW49aKI_4>YKDf4b0RKAl4g0kJ\\OS3gL]NS4XNYOX3fL_Ng7`1ZH`Ng7]1[HcNf7Z1\\HeNf7Y1[HgNf7V1\\HiNf7T1\\HlNe7R1\\HmNf7P1\\HPOd7n0]HSOd7k0]HTOe7i0]HWOd7f0^HYOd7e0]H[Od7b0^H^Oc7?_H@c7>^HBb7<`HCb7:`HFe74\\HKk7MWH3o7FRH:T8^OnGa0X8XOjGh0\\8QOeGo0a8iNaGW1e8aN]G_1X:00000O10000000000000000000000000000000001O1O1O1N2O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O2N2N2N2N1O2N2N2N2N4L4L4L4L3M4L4L4K5L3M2N2N2N2N1O2N2N1O000000000000000000000O100000O1000000O10000O1000000O10000O100000O010000001O00001O0000001O00001O0000001O00000O2O00001O0000001O00001O0000001O000000001O0O10001O001O001O001O001O001O1O001O001O001O0O2O001O001O001O1O2N2N2N2N2N2N3M2N3M4L4L4L4K5L4L4L4L4L4L4L4L4L4L3M3M3M4L3M3M3M4L4L5K5J6K5K5K6J5K5K5K5K5K5K5K7I6J7I6J7I6J6J6J6J6I7Jif6"}, "label": "a man with glasses and a very blurry hand"}]}
{"id": 96859, "image": "COCO_train2014_000000096859.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing eyeglass cut the pizza with his friend\"."}], "task": "refering", "answer_template": "The \"a man wearing eyeglass cut the pizza with his friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.48910937499999996, 0.1055373831775701, 0.9736874999999999, 0.9813785046728972], "iscrowd": 0, "area": 74757.04905, "rle": {"size": [428, 640], "counts": "fSS43S=6I7J6J6J6I7O11OO1O1N2N2N2O1N2N2N2O1N2N2O1N2N2N2O1^KnMhMT2T2VNdMl1X2^NaMc1[2gN]M[1_2oNYMS1c2VOVMl0f2[OVMf0h2@RMb0n2AmLa0R3DhL>X3FcL;]3H^L:a3KYL7g3LUL5j30PL2P44hKNW49aKI_4>YKDf4b0RKAl4g0kJ\\OS3gL]NS4XNYOX3fL_Ng7`1ZH`Ng7]1[HcNf7Z1\\HeNf7Y1[HgNf7V1\\HiNf7T1\\HlNe7R1\\HmNf7P1\\HPOd7n0]HSOd7k0]HTOe7i0]HWOd7f0^HYOd7e0]H[Od7b0^H^Oc7?_H@c7>^HBb7<`HCb7:`HFe74\\HKk7MWH3o7FRH:T8^OnGa0X8XOjGh0\\8QOeGo0a8iNaGW1e8aN]G_1X:00000O10000000000000000000000000000000001O1O1O1N2O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O2N2N2N2N1O2N2N2N2N4L4L4L4L3M4L4L4K5L3M2N2N2N2N1O2N2N1O000000000000000000000O100000O1000000O10000O1000000O10000O100000O010000001O00001O0000001O00001O0000001O00000O2O00001O0000001O00001O0000001O000000001O0O10001O001O001O001O001O001O1O001O001O001O0O2O001O001O001O1O2N2N2N2N2N2N3M2N3M4L4L4L4K5L4L4L4L4L4L4L4L4L4L3M3M3M4L3M3M3M4L4L5K5J6K5K5K6J5K5K5K5K5K5K5K7I6J7I6J7I6J6J6J6J6I7Jif6"}, "label": "a man wearing eyeglass cut the pizza with his friend"}]}
{"id": 326237, "image": "COCO_train2014_000000326237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"backpack that police officer is wearing\"."}], "task": "refering", "answer_template": "The \"backpack that police officer is wearing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 122, 123, 130, 131, 137, 138, 145, 146, 152, 153, 160, 161, 167, 168, 176, 182, 183, 197, 198, 212, 213], "bbox": [0.15555035128805622, 0.32243750000000004, 0.7729976580796253, 0.6302031250000001], "iscrowd": 0, "area": 6267.6039999999975, "rle": {"size": [640, 427], "counts": "Z`Y13Yc0g0K5K5J6K5K5K5K4N3M3L4M3M=B?B=C>j_OUM[>X4O10000O10000O0100O001O1O010]LeA`1\\>\\NSBX1m=dNbBQ1_=kNoBk0P=QOQ4M4L3L4M4K4M3M3M4O0000001O0000001O01O01O003M8HaPY31_ofL7I4M3L4L4L4M3L3M4L4M3L4L4M1N3M2N3N2M201N101O001N2O001N10L4K5J5L5lNT1J5L5Jj[l1"}, "label": "backpack that police officer is wearing"}]}
{"id": 326237, "image": "COCO_train2014_000000326237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag of the woman in front\"."}], "task": "refering", "answer_template": "The \"the bag of the woman in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 122, 123, 130, 131, 137, 138, 145, 146, 152, 153, 160, 161, 167, 168, 176, 182, 183, 197, 198, 212, 213], "bbox": [0.15555035128805622, 0.32243750000000004, 0.7729976580796253, 0.6302031250000001], "iscrowd": 0, "area": 6267.6039999999975, "rle": {"size": [640, 427], "counts": "Z`Y13Yc0g0K5K5J6K5K5K5K4N3M3L4M3M=B?B=C>j_OUM[>X4O10000O10000O0100O001O1O010]LeA`1\\>\\NSBX1m=dNbBQ1_=kNoBk0P=QOQ4M4L3L4M4K4M3M3M4O0000001O0000001O01O01O003M8HaPY31_ofL7I4M3L4L4L4M3L3M4L4M3L4L4M1N3M2N3N2M201N101O001N2O001N10L4K5J5L5lNT1J5L5Jj[l1"}, "label": "the bag of the woman in front"}]}
{"id": 80480, "image": "COCO_train2014_000000080480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male tennis player\"."}], "task": "refering", "answer_template": "The \"male tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 116, 117, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 164, 176, 177, 191, 192], "bbox": [0.7705373831775701, 0.280015625, 1.0, 0.54853125], "iscrowd": 0, "area": 7501.528399999997, "rle": {"size": [640, 428], "counts": "n]^62b4Of:5PELoK0n>8o@Ij32WLNn>`0iDDWLNo>?hDFWLKP?c0fDIY;8eDI[;8cDJ[;<_DEa;:_DH_;9aDG_;9`DI_;<[DEd;>YDDf;=XDDh;=WDDg;>WDCj;=oCIQ<8lCJT<7iCJY<6eCK[<6cCK^<5ZC2f<OVBbNGi1S>ETBfNDh1W>CRBhNEg1\\>^OmAmNEg1^>^OgAoNHd1c>_O^AQONa1e>_OXATO1]1h>BPATO8l0L`NQ?Z2RASO3`Nn>[2n@SOd?l0[@QOk?l0U@QOo?o0P@oNS`0f0W@WOm?c0X@ZOl?b0W@[On?c0R@[OQ`0e0n_OYOU`0g0j_OYOW`0g0h_OYOX`0i0g_OWOZ`0h0e_OXO\\`0h0c_OXO_`0h0__OXOb`0h0]_OXOd`0h0[_OYOf`0f0Y_OZOh`0f0W_OZOi`0h0U_OXOl`0h0S_OXOn`0h0Q_OXOPa0h0o^OXORa0h0m^OXOTa0i0j^OWOVa0i11O1N200O0PNd^O]1]a0c00oMc^O^1^a0a01PNb^O^1^a0bNb^O]1^a0cNc^O]1]a0cNd^O^1Za0aNi^O^1Va0bNl^O^1Ra0bNP_O^1n`0aNU_O^1j`0bNW_O_1g`0aNZ_O_1e`0aN\\_O`1b`0_N`_Oa1_`0_Nb_Ob1\\`0^Ne_Oc1Y`0]Nh_Oc1W`0]Nj_Od1T`0\\NR@_1m?aNV@^1h?bN[@\\1d?dN^@[1a?eNa@Y1_?gNd@HmNS1^`0UOVAf0j>ZOXAe0g>[OYAe0g>[OYAf0f>YO\\Af0e>YO[Ah0d>XO\\Ah0e>VO]Aj0b>VO^Aj0d>TOjA?n7"}, "label": "male tennis player"}]}
{"id": 80480, "image": "COCO_train2014_000000080480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in a white shirt and black pants\"."}], "task": "refering", "answer_template": "The \"the woman in a white shirt and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 116, 117, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 164, 176, 177, 191, 192], "bbox": [0.7705373831775701, 0.280015625, 1.0, 0.54853125], "iscrowd": 0, "area": 7501.528399999997, "rle": {"size": [640, 428], "counts": "n]^62b4Of:5PELoK0n>8o@Ij32WLNn>`0iDDWLNo>?hDFWLKP?c0fDIY;8eDI[;8cDJ[;<_DEa;:_DH_;9aDG_;9`DI_;<[DEd;>YDDf;=XDDh;=WDDg;>WDCj;=oCIQ<8lCJT<7iCJY<6eCK[<6cCK^<5ZC2f<OVBbNGi1S>ETBfNDh1W>CRBhNEg1\\>^OmAmNEg1^>^OgAoNHd1c>_O^AQONa1e>_OXATO1]1h>BPATO8l0L`NQ?Z2RASO3`Nn>[2n@SOd?l0[@QOk?l0U@QOo?o0P@oNS`0f0W@WOm?c0X@ZOl?b0W@[On?c0R@[OQ`0e0n_OYOU`0g0j_OYOW`0g0h_OYOX`0i0g_OWOZ`0h0e_OXO\\`0h0c_OXO_`0h0__OXOb`0h0]_OXOd`0h0[_OYOf`0f0Y_OZOh`0f0W_OZOi`0h0U_OXOl`0h0S_OXOn`0h0Q_OXOPa0h0o^OXORa0h0m^OXOTa0i0j^OWOVa0i11O1N200O0PNd^O]1]a0c00oMc^O^1^a0a01PNb^O^1^a0bNb^O]1^a0cNc^O]1]a0cNd^O^1Za0aNi^O^1Va0bNl^O^1Ra0bNP_O^1n`0aNU_O^1j`0bNW_O_1g`0aNZ_O_1e`0aN\\_O`1b`0_N`_Oa1_`0_Nb_Ob1\\`0^Ne_Oc1Y`0]Nh_Oc1W`0]Nj_Od1T`0\\NR@_1m?aNV@^1h?bN[@\\1d?dN^@[1a?eNa@Y1_?gNd@HmNS1^`0UOVAf0j>ZOXAe0g>[OYAe0g>[OYAf0f>YO\\Af0e>YO[Ah0d>XO\\Ah0e>VO]Aj0b>VO^Aj0d>TOjA?n7"}, "label": "the woman in a white shirt and black pants"}]}
{"id": 80480, "image": "COCO_train2014_000000080480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blond woman playing tennis up at the net\"."}], "task": "refering", "answer_template": "The \"a blond woman playing tennis up at the net\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 97, 98, 99, 112, 113, 114, 115, 127, 128, 129, 130, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236], "bbox": [0.44018691588785047, 0.24734375000000003, 0.7627803738317758, 0.676546875], "iscrowd": 0, "area": 26158.66415, "rle": {"size": [640, 428], "counts": "jie31ic07H7J7H7L5K4L5K4L5K4L5L3ZOg0J5`NfMUA`2\\>UNXAP2h>WNl@o1R?`101N101O0O2O0O2O1N3N2lCjJZ9Y5_FlJ_9W5\\FmJc9U5WFoJh9U5QFoJo9S5kEQKT:R5fESKY:P5aESK_:o4[EUKd:n4VEVKj:m4oDWKP;l4kDWKV;j4dD[K[;h4^D\\Kb;a6000001O00000000010O00000000001O00000001O01O00000000001O0001O000001O00000000001O01O1O1O2N2O0O2N2N1O2N2N1O2N2N1TIoCg5S<QJWDk5k;lI_DQ6b;hIgDU6f<L3M4L3M4L3M3M4L3M4L3M4L3M3M4L3M4L3L5L3L4M4K4L5L3L5L3L5L3L4M4K4M4K4L5L3Lj0WOT\\o1"}, "label": "a blond woman playing tennis up at the net"}]}
{"id": 80480, "image": "COCO_train2014_000000080480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a gray sweatshirt ' s upper body\"."}], "task": "refering", "answer_template": "The \"a woman wearing a gray sweatshirt ' s upper body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 97, 98, 99, 112, 113, 114, 115, 127, 128, 129, 130, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236], "bbox": [0.44018691588785047, 0.24734375000000003, 0.7627803738317758, 0.676546875], "iscrowd": 0, "area": 26158.66415, "rle": {"size": [640, 428], "counts": "jie31ic07H7J7H7L5K4L5K4L5K4L5L3ZOg0J5`NfMUA`2\\>UNXAP2h>WNl@o1R?`101N101O0O2O0O2O1N3N2lCjJZ9Y5_FlJ_9W5\\FmJc9U5WFoJh9U5QFoJo9S5kEQKT:R5fESKY:P5aESK_:o4[EUKd:n4VEVKj:m4oDWKP;l4kDWKV;j4dD[K[;h4^D\\Kb;a6000001O00000000010O00000000001O00000001O01O00000000001O0001O000001O00000000001O01O1O1O2N2O0O2N2N1O2N2N1O2N2N1TIoCg5S<QJWDk5k;lI_DQ6b;hIgDU6f<L3M4L3M4L3M3M4L3M4L3M4L3M3M4L3M4L3L5L3L4M4K4L5L3L5L3L5L3L4M4K4M4K4L5L3Lj0WOT\\o1"}, "label": "a woman wearing a gray sweatshirt ' s upper body"}]}
{"id": 80481, "image": "COCO_train2014_000000080481.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy skateboarding off a wooden ramp\"."}], "task": "refering", "answer_template": "The \"a boy skateboarding off a wooden ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 62, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 129, 152, 153, 176], "bbox": [0.490265625, 0.07864168618266978, 0.7421562500000001, 0.5236065573770492], "iscrowd": 0, "area": 8118.827099999998, "rle": {"size": [427, 640], "counts": "YQS4?e<7O101N1O1O1O2O0O1O1O2N100O1O2N1O100O1O2N1O010O1O1O1O100O1O1O1O100OO10001O000O1^Ob02O1N2N1O2N2H8K5M3N1N3N2M3N2M2O2M3N2M3N2N2O2N1O2N2N2N2N2N2N2N1O2N2N2O1NhMSFT1k9lNUFV1i9iNYFX1e9hN\\FX1c9gN_FZ1_9fNbF[1\\9dNeF^1Y9bNhF_1V9`NlFa1R9_NoFb1o8]NSGW1BVNY9b0VGW1CWNV9b0XGU1DZNR9`0\\GS1E^Nn8>^GR1FaNj8<bGQ1GcNg8:bGQ1JgNd84dGS1JkNc8NdGU1KoNa8HfGW1KTO_8@hGY1LYO\\8[OhG[1O[OY9c0iF_OU9?mFCP9;SGHi87YGKc85_GM^81eG1X8NkG4Q8KQH7l7HVH:f7E]H:c7E_H1j7CbFCf1`0Q8M^F]Od1;V88VH_OR8a0\\2O10O0100O10O0100O10O0100O100O010O100O010O100O100O100O10O0100O1000000000001O1O1N2O1O001O1O1O1O1OkfT2"}, "label": "a boy skateboarding off a wooden ramp"}]}
{"id": 80481, "image": "COCO_train2014_000000080481.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a guy riding a skateboard\"."}], "task": "refering", "answer_template": "The \"this is a guy riding a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 62, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 129, 152, 153, 176], "bbox": [0.490265625, 0.07864168618266978, 0.7421562500000001, 0.5236065573770492], "iscrowd": 0, "area": 8118.827099999998, "rle": {"size": [427, 640], "counts": "YQS4?e<7O101N1O1O1O2O0O1O1O2N100O1O2N1O100O1O2N1O010O1O1O1O100O1O1O1O100OO10001O000O1^Ob02O1N2N1O2N2H8K5M3N1N3N2M3N2M2O2M3N2M3N2N2O2N1O2N2N2N2N2N2N2N1O2N2N2O1NhMSFT1k9lNUFV1i9iNYFX1e9hN\\FX1c9gN_FZ1_9fNbF[1\\9dNeF^1Y9bNhF_1V9`NlFa1R9_NoFb1o8]NSGW1BVNY9b0VGW1CWNV9b0XGU1DZNR9`0\\GS1E^Nn8>^GR1FaNj8<bGQ1GcNg8:bGQ1JgNd84dGS1JkNc8NdGU1KoNa8HfGW1KTO_8@hGY1LYO\\8[OhG[1O[OY9c0iF_OU9?mFCP9;SGHi87YGKc85_GM^81eG1X8NkG4Q8KQH7l7HVH:f7E]H:c7E_H1j7CbFCf1`0Q8M^F]Od1;V88VH_OR8a0\\2O10O0100O10O0100O10O0100O100O010O100O010O100O100O100O10O0100O1000000000001O1O1N2O1O001O1O1O1O1OkfT2"}, "label": "this is a guy riding a skateboard"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing on a sidewalk wearing a tan vest\"."}], "task": "refering", "answer_template": "The \"a man standing on a sidewalk wearing a tan vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 247, 248, 249, 250, 264, 265, 266, 267, 282, 283, 284, 285, 286, 297, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 337, 338, 348, 349, 350, 351, 352, 353, 354, 355, 365, 366, 367, 368, 369, 370, 371, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.506, 0.59146875, 0.9070833333333332, 1.0], "iscrowd": 0, "area": 35914.5323, "rle": {"size": [640, 480], "counts": "Y]h44ic05]NKT_O:g`0JU_O<I8AUOP?;^A`0J9_OQOV?:RAl02;i>nNi@R185n>d0m@AR?`0i@EV?c200O1O1O1O100O1O1O1O100O1O100O100O010O100O100TNQKfDP5k:`KSEa4l:bKRE^4m:dKQE]4n:fKPEZ4n:iKPEX4o:jKPEV4o:mKnDT4Q;nKnDj3Y;YLdD[3h;gLWD[3e;hLYD\\3c;gLZD]3b;eL]D]3`;fL]D^3_;eL_D^3\\;hL_D[3_;iL]DY3a;c2000001O0000001O00000O10000000000000000000000000000000O2O2N1O1O1O1O1O2N1O1O1O1O1O2N1O2N2N3M2N2N2N3M2N>B2N1O2N1O2N2N001O000O2O00001O00001O00001O00001O00001O00001O00001O00001O001O001O001O001O001N2O001O001O001O001O001O001O001O1O001O001O001O5K8H8H7I3M2N3M2N3`NY@aNj?Z1_@XNj?e1^@lMm?P2V1L3M4L4L3L5K4K6K5J5Li0VOSbk0"}, "label": "a man standing on a sidewalk wearing a tan vest"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man looking down with three other men talking beside a traffic sign on the side of the sidewalk curb\"."}], "task": "refering", "answer_template": "The \"man looking down with three other men talking beside a traffic sign on the side of the sidewalk curb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 247, 248, 249, 250, 264, 265, 266, 267, 282, 283, 284, 285, 286, 297, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 337, 338, 348, 349, 350, 351, 352, 353, 354, 355, 365, 366, 367, 368, 369, 370, 371, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.506, 0.59146875, 0.9070833333333332, 1.0], "iscrowd": 0, "area": 35914.5323, "rle": {"size": [640, 480], "counts": "Y]h44ic05]NKT_O:g`0JU_O<I8AUOP?;^A`0J9_OQOV?:RAl02;i>nNi@R185n>d0m@AR?`0i@EV?c200O1O1O1O100O1O1O1O100O1O100O100O010O100O100TNQKfDP5k:`KSEa4l:bKRE^4m:dKQE]4n:fKPEZ4n:iKPEX4o:jKPEV4o:mKnDT4Q;nKnDj3Y;YLdD[3h;gLWD[3e;hLYD\\3c;gLZD]3b;eL]D]3`;fL]D^3_;eL_D^3\\;hL_D[3_;iL]DY3a;c2000001O0000001O00000O10000000000000000000000000000000O2O2N1O1O1O1O1O2N1O1O1O1O1O2N1O2N2N3M2N2N2N3M2N>B2N1O2N1O2N2N001O000O2O00001O00001O00001O00001O00001O00001O00001O00001O001O001O001O001O001N2O001O001O001O001O001O001O001O1O001O001O001O5K8H8H7I3M2N3M2N3`NY@aNj?Z1_@XNj?e1^@lMm?P2V1L3M4L4L3L5K4K6K5J5Li0VOSbk0"}, "label": "man looking down with three other men talking beside a traffic sign on the side of the sidewalk curb"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing black coat and dark blue shirt\"."}], "task": "refering", "answer_template": "The \"man wearing black coat and dark blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 280, 281, 282, 297, 298, 314, 331, 347, 348, 364, 365, 381, 382], "bbox": [0.45695833333333336, 0.6610625, 0.610125, 0.999953125], "iscrowd": 0, "area": 5513.583200000002, "rle": {"size": [640, 480], "counts": "h_Y42Rc0m0[Od0J6K5J6YNmM\\AY2a>QNRAT2l>VNf@P2X?YN\\@l1a?\\1M2N2N2M3N3M2N2M3N2N3N1O1NZM]A3b>HfA7X>CPB<n=_O[B?c=\\OeBCY>@o@\\OP1l0Y=A_A8R2Nc<R1_CfNf<[1\\C\\Nj<e1h21N1000010O0001O1O100O1O1O100O100O100O100O100O2O0O2O1N101N2O0O2O0O2O1N5K6K6I7J5J7Ja]d3"}, "label": "man wearing black coat and dark blue shirt"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"of the two men whose faces we can see , he is the short man on the right\"."}], "task": "refering", "answer_template": "The \"of the two men whose faces we can see , he is the short man on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 280, 281, 282, 297, 298, 314, 331, 347, 348, 364, 365, 381, 382], "bbox": [0.45695833333333336, 0.6610625, 0.610125, 0.999953125], "iscrowd": 0, "area": 5513.583200000002, "rle": {"size": [640, 480], "counts": "h_Y42Rc0m0[Od0J6K5J6YNmM\\AY2a>QNRAT2l>VNf@P2X?YN\\@l1a?\\1M2N2N2M3N3M2N2M3N2N3N1O1NZM]A3b>HfA7X>CPB<n=_O[B?c=\\OeBCY>@o@\\OP1l0Y=A_A8R2Nc<R1_CfNf<[1\\C\\Nj<e1h21N1000010O0001O1O100O1O1O100O100O100O100O100O2O0O2O1N101N2O0O2O0O2O1N5K6K6I7J5J7Ja]d3"}, "label": "of the two men whose faces we can see , he is the short man on the right"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bald man with glasses in blue jacket\"."}], "task": "refering", "answer_template": "The \"bald man with glasses in blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 240, 241, 242, 256, 257, 258, 259, 272, 273, 274, 275, 276, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379], "bbox": [0.013208333333333332, 0.590984375, 0.3047083333333333, 1.0], "iscrowd": 0, "area": 27794.92095, "rle": {"size": [640, 480], "counts": "TX42cc0f0c]OA``0m0n^ODg`0g0l^O2]`0^2TOb0]O8K5M3N2M3M3N2M300O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1L4K5K5K5L3N3M3M3M3M3M3M3N2N20000O100000000O1000000O1000000000000O100000000000000O10000000YJVDX3j;hLYDU3g;jL\\DT3c;mL`DP3`;oLcDo2];QMeDm2\\;QMfDn2[;PMhDn2Z;oLhDP3Y;nLiDQ3X;mLjDR3W;kLlDT3U;jLnDT3T;hLoDW3R;gLPEX3Q;eLREZ3o:dLTEZ3m:cLXEZ3j:cLZEZ3g:cL_EY3b:eLbEX3_:fLeEW3\\:gLiEU3Y:hLkEU3^:aLeE]3g:UL]Ei3_=N2N3M2N8H9G8H9G9G9G8H6J5KaU`6"}, "label": "bald man with glasses in blue jacket"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald man with glasses in a navy coat\"."}], "task": "refering", "answer_template": "The \"a bald man with glasses in a navy coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 240, 241, 242, 256, 257, 258, 259, 272, 273, 274, 275, 276, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379], "bbox": [0.013208333333333332, 0.590984375, 0.3047083333333333, 1.0], "iscrowd": 0, "area": 27794.92095, "rle": {"size": [640, 480], "counts": "TX42cc0f0c]OA``0m0n^ODg`0g0l^O2]`0^2TOb0]O8K5M3N2M3M3N2M300O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1L4K5K5K5L3N3M3M3M3M3M3M3N2N20000O100000000O1000000O1000000000000O100000000000000O10000000YJVDX3j;hLYDU3g;jL\\DT3c;mL`DP3`;oLcDo2];QMeDm2\\;QMfDn2[;PMhDn2Z;oLhDP3Y;nLiDQ3X;mLjDR3W;kLlDT3U;jLnDT3T;hLoDW3R;gLPEX3Q;eLREZ3o:dLTEZ3m:cLXEZ3j:cLZEZ3g:cL_EY3b:eLbEX3_:fLeEW3\\:gLiEU3Y:hLkEU3^:aLeE]3g:UL]Ei3_=N2N3M2N8H9G8H9G9G9G8H6J5KaU`6"}, "label": "a bald man with glasses in a navy coat"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with arms folded\"."}], "task": "refering", "answer_template": "The \"man with arms folded\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 259, 260, 276, 277, 278, 293, 294, 295, 296, 310, 311, 312, 328, 329, 345, 346, 362, 363, 379, 380], "bbox": [0.2546875, 0.609546875, 0.4224791666666667, 1.0], "iscrowd": 0, "area": 11416.735250000002, "rle": {"size": [640, 480], "counts": "Yf\\22jc07J6J6I7J6J6I6K6J6I7J6J6J6K5K5K5J6K5K5K5UC[LS9j3ZFfLd9_3hESMU:R3WE_Mg:f2eDkMZ;Y2RDXNo;l40010O0001O000010O01O000010O0001`IdCZ5\\<`JkC_5V<ZJQDe5o;UJXDk5g;oI`DP6a;iIfDV6c<O1O1O101N1O1O1O100O1O1N3J6J6\\L_A`1h>]N\\AZ1l>cNXAT1P?iNTAo0S?mNQAk0W?ROm@e0[?XOi@?_?]Of@;a?Bc@5e?H^@0j?L[@Kl?3X@Eo?8U@_OS`0=k1M3M3M3L4Mdd\\5"}, "label": "man with arms folded"}]}
{"id": 408163, "image": "COCO_train2014_000000408163.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with his arms crossed on his chest\"."}], "task": "refering", "answer_template": "The \"man with his arms crossed on his chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 259, 260, 276, 277, 278, 293, 294, 295, 296, 310, 311, 312, 328, 329, 345, 346, 362, 363, 379, 380], "bbox": [0.2546875, 0.609546875, 0.4224791666666667, 1.0], "iscrowd": 0, "area": 11416.735250000002, "rle": {"size": [640, 480], "counts": "Yf\\22jc07J6J6I7J6J6I6K6J6I7J6J6J6K5K5K5J6K5K5K5UC[LS9j3ZFfLd9_3hESMU:R3WE_Mg:f2eDkMZ;Y2RDXNo;l40010O0001O000010O01O000010O0001`IdCZ5\\<`JkC_5V<ZJQDe5o;UJXDk5g;oI`DP6a;iIfDV6c<O1O1O101N1O1O1O100O1O1N3J6J6\\L_A`1h>]N\\AZ1l>cNXAT1P?iNTAo0S?mNQAk0W?ROm@e0[?XOi@?_?]Of@;a?Bc@5e?H^@0j?L[@Kl?3X@Eo?8U@_OS`0=k1M3M3M3L4Mdd\\5"}, "label": "man with his arms crossed on his chest"}]}
{"id": 121445, "image": "COCO_train2014_000000121445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white shirt and water bottle on a bench\"."}], "task": "refering", "answer_template": "The \"a white shirt and water bottle on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 161, 178, 179, 196, 197, 213, 214, 215, 232, 233], "bbox": [0.83968, 0.5989066666666667, 1.0, 1.0], "iscrowd": 0, "area": 4084.530899999997, "rle": {"size": [375, 500], "counts": "UTj46a;1O001O001O00010O010O00010O0010O00010O010O01O010O0010O010O01[EGj99RFKo94lE1T:0fE5Y:LbE9^:`0O100O100`NkN]HV1]7POcHP1V7XOhHi0T7[OkH?Z7CfH3a70^HGi7<WH_Ol7d0RH]Ok7g0THXOj7k0UHVOi7m0VHROh7R1VHoNg7U1WHlNg7W1VHjNh7Z1UHhNZ7j1cHWN^7j1_HXNa7i1\\HYNe7g1SH_Nn7e20O101O0O100O101O0O10hGjMo6_31N1000000O2ROm0POP1oNQ1POjI"}, "label": "a white shirt and water bottle on a bench"}]}
{"id": 121445, "image": "COCO_train2014_000000121445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in red shirt\"."}], "task": "refering", "answer_template": "The \"man in red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 138, 139, 140, 156, 157, 158, 173, 174, 175, 176, 190, 191, 192, 206, 207, 208, 209, 210, 224, 225], "bbox": [0.44933999999999996, 0.4607733333333333, 0.82728, 0.9979733333333334], "iscrowd": 0, "area": 8340.028550000003, "rle": {"size": [375, 500], "counts": "]gb25a;6J3N1O1O1O001O1O001N2O1O001O1O000O2O001O1O2M2O1O001O0O2O001O0O10000O010O0100O001B=01O1O0010O01O00010O001O01O001O001O001N101N101N101N1000O2M3O1000000O100OFYOfEf0Z:[OfEe0Y:\\OgEc0Y:^OgEa0Y:@gE?Y:BgE=Y:DgE<X:EhE:X:GhE8X:IhE6X:KhE5W:LiE3W:NiE1W:0iEOW:2iEMW:5hEKW:6iEIV:9jEGU::kEFT:;lEDT:=lECR:?nE@Q:b0oE_OP:a0PF@m9b0RF@k9c0TF^Ok9b0UF_Oj9a0VF@i9`0WFAg9`0YFAf9?ZFBe9>[FCd9=\\FDc9<]FDc9<]FDb9:aFG^98eFF[99gFDZ9<iF_OZ9`0hF\\OZ9d0iFXOX9i0k00O00100O10O01M3@`0K5K4L5N2O10O01O100ROoMUHR2j7QNTHn1k7VNRHk1k7YNTHg1j7\\NUHd1i7`NUH`1i7eNSH\\1l7jNoGV1Q8jNoGV1P8kNoGU1R8lNlGU1T8kNlGU1T8lNjGU1U8W1000000000000001O01O00O2M2N4]MaGh1c8RNfGi1[8SNkGj1W8PNmGo1o8M3M3N2N1O2N2N2M3M3M2N3M3M3C=L4N2M3N2M:GbXo0"}, "label": "man in red shirt"}]}
{"id": 121445, "image": "COCO_train2014_000000121445.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red sports shirt and shorts\"."}], "task": "refering", "answer_template": "The \"a man in a red sports shirt and shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 138, 139, 140, 156, 157, 158, 173, 174, 175, 176, 190, 191, 192, 206, 207, 208, 209, 210, 224, 225], "bbox": [0.44933999999999996, 0.4607733333333333, 0.82728, 0.9979733333333334], "iscrowd": 0, "area": 8340.028550000003, "rle": {"size": [375, 500], "counts": "]gb25a;6J3N1O1O1O001O1O001N2O1O001O1O000O2O001O1O2M2O1O001O0O2O001O0O10000O010O0100O001B=01O1O0010O01O00010O001O01O001O001O001N101N101N101N1000O2M3O1000000O100OFYOfEf0Z:[OfEe0Y:\\OgEc0Y:^OgEa0Y:@gE?Y:BgE=Y:DgE<X:EhE:X:GhE8X:IhE6X:KhE5W:LiE3W:NiE1W:0iEOW:2iEMW:5hEKW:6iEIV:9jEGU::kEFT:;lEDT:=lECR:?nE@Q:b0oE_OP:a0PF@m9b0RF@k9c0TF^Ok9b0UF_Oj9a0VF@i9`0WFAg9`0YFAf9?ZFBe9>[FCd9=\\FDc9<]FDc9<]FDb9:aFG^98eFF[99gFDZ9<iF_OZ9`0hF\\OZ9d0iFXOX9i0k00O00100O10O01M3@`0K5K4L5N2O10O01O100ROoMUHR2j7QNTHn1k7VNRHk1k7YNTHg1j7\\NUHd1i7`NUH`1i7eNSH\\1l7jNoGV1Q8jNoGV1P8kNoGU1R8lNlGU1T8kNlGU1T8lNjGU1U8W1000000000000001O01O00O2M2N4]MaGh1c8RNfGi1[8SNkGj1W8PNmGo1o8M3M3N2N1O2N2N2M3M3M2N3M3M3C=L4N2M3N2M:GbXo0"}, "label": "a man in a red sports shirt and shorts"}]}
{"id": 219752, "image": "COCO_train2014_000000219752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop on right that is white\"."}], "task": "refering", "answer_template": "The \"laptop on right that is white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 34, 35, 52, 53, 70, 71, 87, 88, 89, 105, 106, 107], "bbox": [0.8402999999999999, 0.0, 1.0, 0.5977659574468085], "iscrowd": 0, "area": 9444.0978, "rle": {"size": [282, 500], "counts": "Ujc34[8;O1O1O1O1O2M2O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1A?@`0@a0_O`0@`0@`0@`0E;000000000001O0000000000000000001O00000001O000000000001O0000000000000000001O0000000000000000001OTOk0"}, "label": "laptop on right that is white"}]}
{"id": 219752, "image": "COCO_train2014_000000219752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white colored laptop\"."}], "task": "refering", "answer_template": "The \"a white colored laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 34, 35, 52, 53, 70, 71, 87, 88, 89, 105, 106, 107], "bbox": [0.8402999999999999, 0.0, 1.0, 0.5977659574468085], "iscrowd": 0, "area": 9444.0978, "rle": {"size": [282, 500], "counts": "Ujc34[8;O1O1O1O1O2M2O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1A?@`0@a0_O`0@`0@`0@`0E;000000000001O0000000000000000001O00000001O000000000001O0000000000000000001O0000000000000000001OTOk0"}, "label": "a white colored laptop"}]}
{"id": 219752, "image": "COCO_train2014_000000219752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop with the keyboard covered in clutter including many scraps of paper\"."}], "task": "refering", "answer_template": "The \"a laptop with the keyboard covered in clutter including many scraps of paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160], "bbox": [0.218, 0.004113475177304964, 0.9607, 0.8243617021276596], "iscrowd": 0, "area": 64632.79905, "rle": {"size": [282, 500], "counts": "^Wn04d87J1N2O1N2O1N2O1N2O1N2O1N3N1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2kJVNY3k1gL3Z1NfNo1[OSNe0Q50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000fM^N\\Mb1d2_N[Ma1d2aN[M_1e2bNZM^1e2cN[M]1e2dNZM_6O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1N2O0O2N2N2O1N2N2O0O2N2OZY5"}, "label": "a laptop with the keyboard covered in clutter including many scraps of paper"}]}
{"id": 219752, "image": "COCO_train2014_000000219752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop computer with papers and other items behind a keyboard on a desk\"."}], "task": "refering", "answer_template": "The \"laptop computer with papers and other items behind a keyboard on a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160], "bbox": [0.218, 0.004113475177304964, 0.9607, 0.8243617021276596], "iscrowd": 0, "area": 64632.79905, "rle": {"size": [282, 500], "counts": "^Wn04d87J1N2O1N2O1N2O1N2O1N2O1N3N1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2kJVNY3k1gL3Z1NfNo1[OSNe0Q50000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000fM^N\\Mb1d2_N[Ma1d2aN[M_1e2bNZM^1e2cN[M]1e2dNZM_6O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1N2O0O2N2N2O1N2N2O0O2N2OZY5"}, "label": "laptop computer with papers and other items behind a keyboard on a desk"}]}
{"id": 449136, "image": "COCO_train2014_000000449136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfer with the blue shirt\"."}], "task": "refering", "answer_template": "The \"the surfer with the blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 197, 198, 199], "bbox": [0.38673437499999996, 0.38562893081761007, 0.66821875, 0.7743396226415095], "iscrowd": 0, "area": 7831.475950000001, "rle": {"size": [318, 640], "counts": "kU]21m91O001O1O1O0010O00000001O0000000000000000000O1000000O100000000O1000000O1^GH^78aHM[73dH1Y7OcH7[7IcH:[7GdH<Z7DhH;W7EjH;U7EmH;P7FRI9m6GUI8j6HXI7f6J\\I4d6L]I3c6L^I4b6L_I3a6M`I2`6NaI0`60aIO_61aIO_61`I0`60_I1a6O^I1c6O\\I2d6M\\I3e6M[I2f6NYI2g6OXI1i6OVI2j6MVI3k6MTI4k6MTI4l6LSI4n6LQI5o6KPI6o6JQI6P7JoH7Q7InH8R7HmH8S7IlH8T7HkH9U7FkH:V7FiH;V7FiH:X7FfH<Z7DeH=Z7DeH<[7EdH:]7GbH9^7HaH8_7I`H7_7K`H5_7M`H4Z72fHOS77lHJi6a0WI@f6b0YI@e6b0ZI^Of6b0ZI_Od6c0[I_Ob6c0]I^Ob6c0^I^O`6c0_I^O`6c0_I^O`6c0_I^O`6d0_I\\O`6e0_I\\O`6f0_IZO`6g0_IZO`6h0^IXOb6i0^IWOa6T2O1O1O1O001O1O1O0100O2O0O2O000TMdIT2T700O100O100O100O1O1O1O1O1N101O001O1O001N102N1O2L4L3M4L3M4N5L4L3N2M2N2M2N3L4M3L4M2N100O100O101N100O1O1O1N101O0O2O00YcQ2"}, "label": "the surfer with the blue shirt"}]}
{"id": 449136, "image": "COCO_train2014_000000449136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt that is surfing\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt that is surfing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 197, 198, 199], "bbox": [0.38673437499999996, 0.38562893081761007, 0.66821875, 0.7743396226415095], "iscrowd": 0, "area": 7831.475950000001, "rle": {"size": [318, 640], "counts": "kU]21m91O001O1O1O0010O00000001O0000000000000000000O1000000O100000000O1000000O1^GH^78aHM[73dH1Y7OcH7[7IcH:[7GdH<Z7DhH;W7EjH;U7EmH;P7FRI9m6GUI8j6HXI7f6J\\I4d6L]I3c6L^I4b6L_I3a6M`I2`6NaI0`60aIO_61aIO_61`I0`60_I1a6O^I1c6O\\I2d6M\\I3e6M[I2f6NYI2g6OXI1i6OVI2j6MVI3k6MTI4k6MTI4l6LSI4n6LQI5o6KPI6o6JQI6P7JoH7Q7InH8R7HmH8S7IlH8T7HkH9U7FkH:V7FiH;V7FiH:X7FfH<Z7DeH=Z7DeH<[7EdH:]7GbH9^7HaH8_7I`H7_7K`H5_7M`H4Z72fHOS77lHJi6a0WI@f6b0YI@e6b0ZI^Of6b0ZI_Od6c0[I_Ob6c0]I^Ob6c0^I^O`6c0_I^O`6c0_I^O`6c0_I^O`6d0_I\\O`6e0_I\\O`6f0_IZO`6g0_IZO`6h0^IXOb6i0^IWOa6T2O1O1O1O001O1O1O0100O2O0O2O000TMdIT2T700O100O100O100O1O1O1O1O1N101O001O1O001N102N1O2L4L3M4L3M4N5L4L3N2M2N2M2N3L4M3L4M2N100O100O101N100O1O1O1N101O0O2O00YcQ2"}, "label": "a man in a blue shirt that is surfing"}]}
{"id": 449136, "image": "COCO_train2014_000000449136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink and white surfboard being ridden by a man\"."}], "task": "refering", "answer_template": "The \"a pink and white surfboard being ridden by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.18757812499999998, 0.5145911949685534, 0.55159375, 0.7572955974842767], "iscrowd": 0, "area": 12033.741300000003, "rle": {"size": [318, 640], "counts": "k]U13k95K4L2M3N2N2M4M2N2M3N2N2N3L3N2N1N2O1O1O1N101O1N2O1O1O1O1O001O1O1O00001O0000001O00000010O000001O0000001O0000001O000000000010O00000000000000000001O00000000000001O00000001O000000000000000000001O000001O0000000000000001O000000000000000001O01O000000000000000000mMXHP2h7oMZHP2f7PN\\Ho1c7QN^Hn1b7RN`Hl1k7N1O2N1O00000000010O0000000000001O000001O00000001OO1000000OnNTH4l7LUH3k7MVH2j7MXH2h7NYH1f70\\HNd71^HNb72_HMa73`HL`74aHK^78aHG_7:aHE_7<aHC_7?`H@_7b0aH]O_7d0`H\\O`7e0_H[Oa7e0_H[Oa7e0`HZO_7g0aHYO_7g0aHYO_7g0aHXO`7h0`HXO`7h0`HXO`7h0aHWO_7i0aHWO_7j0`HVO`7j0`HVO`7i0aHWO_7h0bHXO^7g0cHYO]7f0dHZO\\7f0cH[O\\7e0eH[O[7d0eH]O[7b0fH^OZ7a0fH@Z7?gHAY7>gHCY7<hHDX7<gHEY7:hHFX79hHHX78hHHX77hHJX75iHKW74iHMW73iHLX73hHNX71iHOW70iH1W7OiH1W7NjH2V7MjH4W7JjH6V7IjH8V7GkH9U7FkH;U7DlH<U7BkH?U7BjH>V7CiH=W7DiH;W7FiH9X7FiH9W7HiH7W7JiH5a8O001O9Gglh2"}, "label": "a pink and white surfboard being ridden by a man"}]}
{"id": 449136, "image": "COCO_train2014_000000449136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and pink surfboard with a man riding it on a wave\"."}], "task": "refering", "answer_template": "The \"a white and pink surfboard with a man riding it on a wave\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.18757812499999998, 0.5145911949685534, 0.55159375, 0.7572955974842767], "iscrowd": 0, "area": 12033.741300000003, "rle": {"size": [318, 640], "counts": "k]U13k95K4L2M3N2N2M4M2N2M3N2N2N3L3N2N1N2O1O1O1N101O1N2O1O1O1O1O001O1O1O00001O0000001O00000010O000001O0000001O0000001O000000000010O00000000000000000001O00000000000001O00000001O000000000000000000001O000001O0000000000000001O000000000000000001O01O000000000000000000mMXHP2h7oMZHP2f7PN\\Ho1c7QN^Hn1b7RN`Hl1k7N1O2N1O00000000010O0000000000001O000001O00000001OO1000000OnNTH4l7LUH3k7MVH2j7MXH2h7NYH1f70\\HNd71^HNb72_HMa73`HL`74aHK^78aHG_7:aHE_7<aHC_7?`H@_7b0aH]O_7d0`H\\O`7e0_H[Oa7e0_H[Oa7e0`HZO_7g0aHYO_7g0aHYO_7g0aHXO`7h0`HXO`7h0`HXO`7h0aHWO_7i0aHWO_7j0`HVO`7j0`HVO`7i0aHWO_7h0bHXO^7g0cHYO]7f0dHZO\\7f0cH[O\\7e0eH[O[7d0eH]O[7b0fH^OZ7a0fH@Z7?gHAY7>gHCY7<hHDX7<gHEY7:hHFX79hHHX78hHHX77hHJX75iHKW74iHMW73iHLX73hHNX71iHOW70iH1W7OiH1W7NjH2V7MjH4W7JjH6V7IjH8V7GkH9U7FkH;U7DlH<U7BkH?U7BjH>V7CiH=W7DiH;W7FiH9X7FiH9W7HiH7W7JiH5a8O001O9Gglh2"}, "label": "a white and pink surfboard with a man riding it on a wave"}]}
{"id": 490097, "image": "COCO_train2014_000000490097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch behind a lamp\"."}], "task": "refering", "answer_template": "The \"a white couch behind a lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 245, 246, 249, 250, 267, 268, 269, 272, 273, 290, 291, 292, 295, 296, 318, 319], "bbox": [0.630296875, 0.615990990990991, 0.9059531249999999, 0.8347297297297297], "iscrowd": 0, "area": 5859.6505, "rle": {"size": [444, 640], "counts": "QS_51`=:N100010O1003M3M2B^O_Cc0_<A]Ca0a<C[C73B`<h0eCWOX<k0jCTOT<m0nCSOo;n0=3N1O1N2O1O2N1N3N3M2N2N2O1N201O000O100000000M4JROZD0e;1]DMc;3_DKa;5`DJ_;6bDJ^;6bDJ];7cDJ\\;6dDJ\\;6dDJ\\;5eDK[;5eDK\\;4dDM[;3eDM\\;2dDN\\;1eDO\\;0dD0\\;0dD0];OcD1];NdD3[;LfD4[;KeD5[;JgD5Z;IgD7Y;IhD6Y;HQEOo:O`EE_:6gEIY:5jEJV:5lEJU:3nELR:2QFMo91TFNl90WF0h9OZF0oQi0R1\\XWO<F;L3M3N2M301O01O0001O000000010O0000001O0000001O000000001O0O10001O00000OjMcDP2e;D<@`0_Ob0C<M3N2MhTj0"}, "label": "a white couch behind a lamp"}]}
{"id": 490097, "image": "COCO_train2014_000000490097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch against the wall\"."}], "task": "refering", "answer_template": "The \"a couch against the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 245, 246, 249, 250, 267, 268, 269, 272, 273, 290, 291, 292, 295, 296, 318, 319], "bbox": [0.630296875, 0.615990990990991, 0.9059531249999999, 0.8347297297297297], "iscrowd": 0, "area": 5859.6505, "rle": {"size": [444, 640], "counts": "QS_51`=:N100010O1003M3M2B^O_Cc0_<A]Ca0a<C[C73B`<h0eCWOX<k0jCTOT<m0nCSOo;n0=3N1O1N2O1O2N1N3N3M2N2N2O1N201O000O100000000M4JROZD0e;1]DMc;3_DKa;5`DJ_;6bDJ^;6bDJ];7cDJ\\;6dDJ\\;6dDJ\\;5eDK[;5eDK\\;4dDM[;3eDM\\;2dDN\\;1eDO\\;0dD0\\;0dD0];OcD1];NdD3[;LfD4[;KeD5[;JgD5Z;IgD7Y;IhD6Y;HQEOo:O`EE_:6gEIY:5jEJV:5lEJU:3nELR:2QFMo91TFNl90WF0h9OZF0oQi0R1\\XWO<F;L3M3N2M301O01O0001O000000010O0000001O0000001O000000001O0O10001O00000OjMcDP2e;D<@`0_Ob0C<M3N2MhTj0"}, "label": "a couch against the wall"}]}
{"id": 490097, "image": "COCO_train2014_000000490097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large white color sofa\"."}], "task": "refering", "answer_template": "The \"a large white color sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 358, 359], "bbox": [0.25078125, 0.6470495495495496, 0.6481718750000001, 0.9872297297297298], "iscrowd": 0, "area": 29807.105249999997, "rle": {"size": [444, 640], "counts": "oRV24h=5J8I;D=D;D<E<C<E;E<D3M1O00001O001O01O01O001O00001O010O00001O001O00001O010O00001O001O00001O000010O01O00001O001O00001O010O00001O010O000010O01O00010O001O01O01O001O01O01O001O01O01O00001O010O00001O010O00001O0001O000000000000001O000001O00000001O000000000000001O0000000000001O000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000001O1O1O001O1O1O001O1O1O001O1O0000001O000000000000000000000000000000000000000000001O0000000000001N1N2N2N3M2N2N2N3M2N2N2N3M2N2`NnD4T;HTE2o:IXE2j:J^EOe:MbEEi::^1M3N2M2O2M3N2M3N`WQ3"}, "label": "a large white color sofa"}]}
{"id": 490097, "image": "COCO_train2014_000000490097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sofa closet to camera\"."}], "task": "refering", "answer_template": "The \"sofa closet to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 358, 359], "bbox": [0.25078125, 0.6470495495495496, 0.6481718750000001, 0.9872297297297298], "iscrowd": 0, "area": 29807.105249999997, "rle": {"size": [444, 640], "counts": "oRV24h=5J8I;D=D;D<E<C<E;E<D3M1O00001O001O01O01O001O00001O010O00001O001O00001O010O00001O001O00001O000010O01O00001O001O00001O010O00001O010O000010O01O00010O001O01O01O001O01O01O001O01O01O00001O010O00001O010O00001O0001O000000000000001O000001O00000001O000000000000001O0000000000001O000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000001O1O1O001O1O1O001O1O1O001O1O0000001O000000000000000000000000000000000000000000001O0000000000001N1N2N2N3M2N2N2N3M2N2N2N3M2N2`NnD4T;HTE2o:IXE2j:J^EOe:MbEEi::^1M3N2M2O2M3N2M3N`WQ3"}, "label": "sofa closet to camera"}]}
{"id": 490097, "image": "COCO_train2014_000000490097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair at the right side of the table\"."}], "task": "refering", "answer_template": "The \"a chair at the right side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 345], "bbox": [0.0, 0.6277927927927928, 0.167453125, 0.966891891891892], "iscrowd": 0, "area": 8560.824350000006, "rle": {"size": [444, 640], "counts": "i:6f=5oBLP<4lC1S<OoC0P<0QDOP<1PDOo;1QD0n;0SDOn;0SD0m;0SDOm;1TDNm;2SDNR:MlF5SOMQ:7dFL[OMP:`0]FBDMo9j0UFYOLMn9S1oEQO0MP:[1hEkNJ4]:[1`E]O]:j2WO6J1O1O0000O0dNaF]N_9b1fFXN\\9g1jFPNZ9o1Z1O1000O100000O10000O010O10000O100O01000000000O10O100000O10000000O0100000000O1O010O100O1O100O010O1O10000O10O10O100O10000O01000O10000O11O0O1000000[EgMg9Y2SFnMn9Q2iEWNa:_1WEhNS;l1b1]N_hV7"}, "label": "a chair at the right side of the table"}]}
{"id": 490097, "image": "COCO_train2014_000000490097.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden chair at the table that is closest to the couch and has its back facing the camera\"."}], "task": "refering", "answer_template": "The \"the wooden chair at the table that is closest to the couch and has its back facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 345], "bbox": [0.0, 0.6277927927927928, 0.167453125, 0.966891891891892], "iscrowd": 0, "area": 8560.824350000006, "rle": {"size": [444, 640], "counts": "i:6f=5oBLP<4lC1S<OoC0P<0QDOP<1PDOo;1QD0n;0SDOn;0SD0m;0SDOm;1TDNm;2SDNR:MlF5SOMQ:7dFL[OMP:`0]FBDMo9j0UFYOLMn9S1oEQO0MP:[1hEkNJ4]:[1`E]O]:j2WO6J1O1O0000O0dNaF]N_9b1fFXN\\9g1jFPNZ9o1Z1O1000O100000O10000O010O10000O100O01000000000O10O100000O10000000O0100000000O1O010O100O1O100O010O1O10000O10O10O100O10000O01000O10000O11O0O1000000[EgMg9Y2SFnMn9Q2iEWNa:_1WEhNS;l1b1]N_hV7"}, "label": "the wooden chair at the table that is closest to the couch and has its back facing the camera"}]}
{"id": 195188, "image": "COCO_train2014_000000195188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gauge that reads 18 : 41\"."}], "task": "refering", "answer_template": "The \"a gauge that reads 18 : 41\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 103, 104, 105, 106], "bbox": [0.671, 0.009216867469879519, 0.991, 0.9700602409638555], "iscrowd": 0, "area": 19748.0, "rle": {"size": [166, 500], "counts": "Uaf19i4:G:E:F9H5J6K5J6L5L3M3M3N2M3M4L3M2N2N2O1N2N2N2N2N2N2O1N1O2N2N2N2N2N2O1N2N2O1N2O1N101N101N101N1O101N100O2O00001N100000001O000O1000001O000O1000001O0000000O101O00000000O100000000O100000000O100000000O2O00001O0O2O00001O0O2O001O000O2O0O2N100O2N101N1O2O0O2O1N2N2O1N1O2N2M3N2N2N2N1O2M3N2N2N3M3L4M2N3M3M2M4M3M3K4K6J6J:F:F;E:F;EYb0"}, "label": "a gauge that reads 18 : 41"}]}
{"id": 195188, "image": "COCO_train2014_000000195188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock with out minites and hours indicators\"."}], "task": "refering", "answer_template": "The \"a clock with out minites and hours indicators\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 66, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 103, 104, 105, 106], "bbox": [0.671, 0.009216867469879519, 0.991, 0.9700602409638555], "iscrowd": 0, "area": 19748.0, "rle": {"size": [166, 500], "counts": "Uaf19i4:G:E:F9H5J6K5J6L5L3M3M3N2M3M4L3M2N2N2O1N2N2N2N2N2N2O1N1O2N2N2N2N2N2O1N2N2O1N2O1N101N101N101N1O101N100O2O00001N100000001O000O1000001O000O1000001O0000000O101O00000000O100000000O100000000O100000000O2O00001O0O2O00001O0O2O001O000O2O0O2N100O2N101N1O2O0O2O1N2N2O1N1O2N2M3N2N2N2N1O2M3N2N2N3M3L4M2N3M3M2M4M3M3K4K6J6J:F:F;E:F;EYb0"}, "label": "a clock with out minites and hours indicators"}]}
{"id": 195188, "image": "COCO_train2014_000000195188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wach in the middle showing 18 : 34 minutes\"."}], "task": "refering", "answer_template": "The \"a wach in the middle showing 18 : 34 minutes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 24, 25, 26, 27, 28, 29, 42, 43, 44, 45, 46, 47, 60, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100], "bbox": [0.339, 0.030120481927710843, 0.659, 0.9759036144578314], "iscrowd": 0, "area": 19513.375, "rle": {"size": [166, 500], "counts": "Zdk08j49G9G9G9G9H8I6J3M4L3M3N3L3M4L3M4L3N3L3M3N3M2N2N2O0O2N101N2N1O2O0O2N2O0O2N1O2O0O2N2O0O2O001O1N101O001O0O2O00000O1000000O101O00000O1000000O10001O0O100000000001O0001O000001O00000O101O000O1000001N1000000O101O000O101O000O2O00001N101O000O2O000O2N1O2N1O101N1O1O2N2N2N2N3M2N3N1N3M2N2N3M2N3M2N3M3N1N3K5L4L3M4K5L3M4K5L3M4L:E`0A?An^k0"}, "label": "a wach in the middle showing 18 : 34 minutes"}]}
{"id": 195188, "image": "COCO_train2014_000000195188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the watch face in the middle\"."}], "task": "refering", "answer_template": "The \"the watch face in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 24, 25, 26, 27, 28, 29, 42, 43, 44, 45, 46, 47, 60, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100], "bbox": [0.339, 0.030120481927710843, 0.659, 0.9759036144578314], "iscrowd": 0, "area": 19513.375, "rle": {"size": [166, 500], "counts": "Zdk08j49G9G9G9G9H8I6J3M4L3M3N3L3M4L3M4L3N3L3M3N3M2N2N2O0O2N101N2N1O2O0O2N2O0O2N1O2O0O2N2O0O2O001O1N101O001O0O2O00000O1000000O101O00000O1000000O10001O0O100000000001O0001O000001O00000O101O000O1000001N1000000O101O000O101O000O2O00001N101O000O2O000O2N1O2N1O101N1O1O2N2N2N2N3M2N3N1N3M2N2N3M2N3M2N3M3N1N3K5L4L3M4K5L3M4K5L3M4L:E`0A?An^k0"}, "label": "the watch face in the middle"}]}
{"id": 195188, "image": "COCO_train2014_000000195188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the watch that reads 6 : 32\"."}], "task": "refering", "answer_template": "The \"the watch that reads 6 : 32\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 91, 92, 93, 94], "bbox": [0.008, 0.035240963855421685, 0.317, 0.975], "iscrowd": 0, "area": 18843.875, "rle": {"size": [166, 500], "counts": "[g06j4:F;D;I8I7H7I5L3L5K4M3L5L3N3M2N2N3M2N2N3M2N3N0O2N2N1O2N2N1O2N2O0O2N1O2O1N101N2QOkKk0U4UOkKj0W4UOiKj0[4O0O101N10001N10001O000O2O00001O000O100000001N100000000O1000001O00000O100000001O000000O101O0O10000O10001N1000000O10001N10000O2O001N101O0O2O001N101N101N101N101N101N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N3L3N3M3M3L5L3M3L4M4L3L4K6I6I7POhWg1"}, "label": "the watch that reads 6 : 32"}]}
{"id": 195188, "image": "COCO_train2014_000000195188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a watch with the seconds hand at 23 seconds\"."}], "task": "refering", "answer_template": "The \"a watch with the seconds hand at 23 seconds\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 91, 92, 93, 94], "bbox": [0.008, 0.035240963855421685, 0.317, 0.975], "iscrowd": 0, "area": 18843.875, "rle": {"size": [166, 500], "counts": "[g06j4:F;D;I8I7H7I5L3L5K4M3L5L3N3M2N2N3M2N2N3M2N3N0O2N2N1O2N2N1O2N2O0O2N1O2O1N101N2QOkKk0U4UOkKj0W4UOiKj0[4O0O101N10001N10001O000O2O00001O000O100000001N100000000O1000001O00000O100000001O000000O101O0O10000O10001N1000000O10001N10000O2O001N101O0O2O001N101N101N101N101N101N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N3L3N3M3M3L5L3M3L4M4L3L4K6I6I7POhWg1"}, "label": "a watch with the seconds hand at 23 seconds"}]}
{"id": 539255, "image": "COCO_train2014_000000539255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the largest sheep\"."}], "task": "refering", "answer_template": "The \"the largest sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 171, 172, 173, 194, 195, 196, 197, 216, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 287, 288], "bbox": [0.43089062499999997, 0.3990208333333333, 0.592875, 0.7510833333333333], "iscrowd": 0, "area": 10680.111499999997, "rle": {"size": [480, 640], "counts": "bdQ4:[>=B=C>F9J7J6K4N3M3M2N2N3N1N2N3M2N2]OXMkDk2S;b0O2N1O2N1O2N102M4Me0[O3L010O100NaNVFjMj9V2bF^M]9c2nFRMR9n2XGhLh8Y3^15UE^LP:]4L3M3M3M3M1O1O1O001O1O1000O0K6lNS1I7O1O1O2M2O7I2N2N22N5Kj0VO00O10000O10000O10000jNV1H8N2N2N2N2N2N3N3L3M3M3M3N3K4M3L4L5L3L4L4K6J5K5LiRj3"}, "label": "the largest sheep"}]}
{"id": 539255, "image": "COCO_train2014_000000539255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep with its wool standing next to a wall\"."}], "task": "refering", "answer_template": "The \"a sheep with its wool standing next to a wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 171, 172, 173, 194, 195, 196, 197, 216, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 287, 288], "bbox": [0.43089062499999997, 0.3990208333333333, 0.592875, 0.7510833333333333], "iscrowd": 0, "area": 10680.111499999997, "rle": {"size": [480, 640], "counts": "bdQ4:[>=B=C>F9J7J6K4N3M3M2N2N3N1N2N3M2N2]OXMkDk2S;b0O2N1O2N1O2N102M4Me0[O3L010O100NaNVFjMj9V2bF^M]9c2nFRMR9n2XGhLh8Y3^15UE^LP:]4L3M3M3M3M1O1O1O001O1O1000O0K6lNS1I7O1O1O2M2O7I2N2N22N5Kj0VO00O10000O10000O10000jNV1H8N2N2N2N2N2N3N3L3M3M3M3N3K4M3L4L5L3L4L4K6J5K5LiRj3"}, "label": "a sheep with its wool standing next to a wall"}]}
{"id": 170623, "image": "COCO_train2014_000000170623.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white mug with coffe\"."}], "task": "refering", "answer_template": "The \"white mug with coffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193], "bbox": [0.6591993464052288, 0.08284313725490197, 0.8946241830065359, 0.39227124183006534], "iscrowd": 0, "area": 22800.2422, "rle": {"size": [612, 612], "counts": "jQa73Sb0P1POP1POP1POP1J7L2M4M2N3L4M2M4M2N2M3N1N2O101N100O2O0O1O2O0O101N100O2O0O100O2O000O100000000O10000000O10O100000000O1000000000000000000000000000O100O100O10000O100O100O100O10000O2O0O100O100O1O1O1O1O2N1O1J6K5L4001O0001O1O001O10O01O001O1O00100O001O001O1O001O1O001O1O1PNZA@h>;]ACd>:`ADa>8cAF_>7eAF]>6hAF[>6jAGX>6lAGV>6mAGV>5oAHS>5QBHQ>5SBGP>5UBHn=4UBJY>GkA6g`0I>CRRV1"}, "label": "white mug with coffe"}]}
{"id": 170623, "image": "COCO_train2014_000000170623.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee with cream serves as a good accompaniment to this breakfast\"."}], "task": "refering", "answer_template": "The \"a cup of coffee with cream serves as a good accompaniment to this breakfast\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193], "bbox": [0.6591993464052288, 0.08284313725490197, 0.8946241830065359, 0.39227124183006534], "iscrowd": 0, "area": 22800.2422, "rle": {"size": [612, 612], "counts": "jQa73Sb0P1POP1POP1POP1J7L2M4M2N3L4M2M4M2N2M3N1N2O101N100O2O0O1O2O0O101N100O2O0O100O2O000O100000000O10000000O10O100000000O1000000000000000000000000000O100O100O10000O100O100O100O10000O2O0O100O100O1O1O1O1O2N1O1J6K5L4001O0001O1O001O10O01O001O1O00100O001O001O1O001O1O001O1O1PNZA@h>;]ACd>:`ADa>8cAF_>7eAF]>6hAF[>6jAGX>6lAGV>6mAGV>5oAHS>5QBHQ>5SBGP>5UBHn=4UBJY>GkA6g`0I>CRRV1"}, "label": "a cup of coffee with cream serves as a good accompaniment to this breakfast"}]}
{"id": 170636, "image": "COCO_train2014_000000170636.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rear motorcycle cop\"."}], "task": "refering", "answer_template": "The \"rear motorcycle cop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 289, 290], "bbox": [0.520390625, 0.3354869358669834, 0.77965625, 0.8285035629453682], "iscrowd": 0, "area": 19115.306899999996, "rle": {"size": [421, 640], "counts": "QSY4:j<3L4L4L4N1O001N10000[E\\Oj8d0UG_Oh8a0WGCg8<XG8U8IjG8U8IiG9W8HgG9X8HfG:X8HgG:W8GdFUOh0U1d81[G1c81\\G0c80]G1b80]G0c80]G1a81^G0a80_G0a81VG`NG_1Q94TGaNI\\1Q96RGcNJX1S9]2O1O100J6J6N3M`0_O8I4L4L4K5L4M3N2M3N2M1O100O100O1O100O100O100000000000O1N2N2O1N2O2L3M3M3M30000000000000000O2N2N101N101O1N101O1N101O3L3]MSHJP8KZH2j7JXH5j7GZH6k7EXH8o7ASH=S8]OPH`0T8[OPHc0P8\\ORHb0P8[OTHb0m7]OVH`0l7]OXH`0i7_OYH`0h7]O\\H`0e7^O^H?e7@[H>g7CYH;i7BdFWOc1U1k7GTH7o7HRH4Q8LoGMX83THZOT8e0V2000000101N3HROSDQ1j;QOTDW1d;g0^OX1hN[MVFQ2h9mMYFV2d9iM]FZ2a9dM`F`2]9^MdFe2Z9ZMhFg2V9YMjFh2U9XMkFb0KT1[9ZNkF`0MS1Y9]NjF`0NR1Y9\\NjFNI95[1Y9^NiFNK29a1T9^NiFNLL=f1o8_NjFNa0b1g8^NjFNa0c1g8\\NjFNa0e1g8ZNmGf1i97I01N1O100O1O1O101N1O1GbDjN`;U1bDhN`;W18kNZDh0U<J7J6J5L4K_fi1"}, "label": "rear motorcycle cop"}]}
{"id": 170636, "image": "COCO_train2014_000000170636.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a police man coming last in a motorbike\"."}], "task": "refering", "answer_template": "The \"a police man coming last in a motorbike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 289, 290], "bbox": [0.520390625, 0.3354869358669834, 0.77965625, 0.8285035629453682], "iscrowd": 0, "area": 19115.306899999996, "rle": {"size": [421, 640], "counts": "QSY4:j<3L4L4L4N1O001N10000[E\\Oj8d0UG_Oh8a0WGCg8<XG8U8IjG8U8IiG9W8HgG9X8HfG:X8HgG:W8GdFUOh0U1d81[G1c81\\G0c80]G1b80]G0c80]G1a81^G0a80_G0a81VG`NG_1Q94TGaNI\\1Q96RGcNJX1S9]2O1O100J6J6N3M`0_O8I4L4L4K5L4M3N2M3N2M1O100O100O1O100O100O100000000000O1N2N2O1N2O2L3M3M3M30000000000000000O2N2N101N101O1N101O1N101O3L3]MSHJP8KZH2j7JXH5j7GZH6k7EXH8o7ASH=S8]OPH`0T8[OPHc0P8\\ORHb0P8[OTHb0m7]OVH`0l7]OXH`0i7_OYH`0h7]O\\H`0e7^O^H?e7@[H>g7CYH;i7BdFWOc1U1k7GTH7o7HRH4Q8LoGMX83THZOT8e0V2000000101N3HROSDQ1j;QOTDW1d;g0^OX1hN[MVFQ2h9mMYFV2d9iM]FZ2a9dM`F`2]9^MdFe2Z9ZMhFg2V9YMjFh2U9XMkFb0KT1[9ZNkF`0MS1Y9]NjF`0NR1Y9\\NjFNI95[1Y9^NiFNK29a1T9^NiFNLL=f1o8_NjFNa0b1g8^NjFNa0c1g8\\NjFNa0e1g8ZNmGf1i97I01N1O100O1O1O101N1O1GbDjN`;U1bDhN`;W18kNZDh0U<J7J6J5L4K_fi1"}, "label": "a police man coming last in a motorbike"}]}
{"id": 170636, "image": "COCO_train2014_000000170636.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chubby police officer on a motorcycle numbered c - 21\"."}], "task": "refering", "answer_template": "The \"a chubby police officer on a motorcycle numbered c - 21\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.016265625, 0.3033729216152019, 0.474515625, 0.9887648456057007], "iscrowd": 0, "area": 56028.08380000001, "rle": {"size": [421, 640], "counts": "_Z42d<`0M3N2M2RDZO\\;h0aDDS;?jDHo:;oDGP;:mDHR;9lDJS;T101N101N101O0O2O001UGhMf6Y2YIPN_6Q2_IXNY6i1dIbNR6`1lImNh5T1VJWO`5j0^JAW5a0fJGT5:jJGV59iJIU59jJHU59jJGU5<iJEV5<hJEX5<]JWMYO^2X6>[JXM[O[2Y6>YJZM]OX2Z6`0UJ[M@V2Y6a0UJ\\M@S2[6b0RJ^MBQ2[6c0oI3o5OoI2Q6OmI3R6OlI1S61kI1T6a3O0O2O1N2O1O1N2O1O1M3M3M2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1O10000O100O100O10000O100O100O10000O100O100O10000O100O10000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O00000000001O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N2N2mKnJOU5NnJOT5OPKMR50SKLo42UKJm43XKIj45YKGj47ZKZOUO_Mc5T3]KUOSOhMb5Q3_KoNSOPN`5n2bKjNROXN^5l2cKeNSO_N\\5j2eK_NSOgNZ5g2hKiMC@g4e2iM\\MY2b2gM^M[2_2fMaM\\2]2dMcM^2Z2cMfM_2X2aMhMa2V2^MkMb2T2_MlM_2V2aMjM]2W2dMiMZ2X2gMhMV2Z2jMgMT2Z2mMdMS2]2nMbMR2]2PNbMR2\\2oMcMR2\\2oMcMR2\\2nMdMS2Z2oMeMR2Z2oMeMR2Z2oMeMR2Y2PNfMQ2X2iLnMjMJ^5V2bLkN]MnNR6U2\\LAd3<WLKj3PObI4_2R1P4jNcI0W2^1V4bNeILo1i1]4ZNgIIg1S2d4SNgIE`1`2i4jMjIBY1i2n4eMkI^OU1P3Q5bMlIZOP1X3U5]MnIWOj0a3X5XMXKl2i4SMSKQ3n4oLmJV3S5hLjJ\\3W5WLQKm3P74K5K4M4M2O2M2O2M201O001O02O000O101N100O1fLlGc1U8[NmGc1T8\\NoGb1Q8]NQHa1P8]NSHb1n7\\NUHb1k7]NWHa1j7^NXHa1i7\\NZHb1g7]NZHc1g7YN\\Hf1e7kMjHU2W7gMlHX2V7dMmH\\2S7aMPI^2R7^MRIa2n6\\MUId2f8O100O100O101E_EoMb:Q2`EmM`:R2cEkM^:U29O100O101N100O10000O109Fb0_Ob0]Ob0_O`iY4"}, "label": "a chubby police officer on a motorcycle numbered c - 21"}]}
{"id": 170636, "image": "COCO_train2014_000000170636.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the police bike in the very front with c - 21 on it\"."}], "task": "refering", "answer_template": "The \"the police bike in the very front with c - 21 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.016265625, 0.3033729216152019, 0.474515625, 0.9887648456057007], "iscrowd": 0, "area": 56028.08380000001, "rle": {"size": [421, 640], "counts": "_Z42d<`0M3N2M2RDZO\\;h0aDDS;?jDHo:;oDGP;:mDHR;9lDJS;T101N101N101O0O2O001UGhMf6Y2YIPN_6Q2_IXNY6i1dIbNR6`1lImNh5T1VJWO`5j0^JAW5a0fJGT5:jJGV59iJIU59jJHU59jJGU5<iJEV5<hJEX5<]JWMYO^2X6>[JXM[O[2Y6>YJZM]OX2Z6`0UJ[M@V2Y6a0UJ\\M@S2[6b0RJ^MBQ2[6c0oI3o5OoI2Q6OmI3R6OlI1S61kI1T6a3O0O2O1N2O1O1N2O1O1M3M3M2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1O10000O100O100O10000O100O100O10000O100O100O10000O100O10000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O00000000001O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N2N2mKnJOU5NnJOT5OPKMR50SKLo42UKJm43XKIj45YKGj47ZKZOUO_Mc5T3]KUOSOhMb5Q3_KoNSOPN`5n2bKjNROXN^5l2cKeNSO_N\\5j2eK_NSOgNZ5g2hKiMC@g4e2iM\\MY2b2gM^M[2_2fMaM\\2]2dMcM^2Z2cMfM_2X2aMhMa2V2^MkMb2T2_MlM_2V2aMjM]2W2dMiMZ2X2gMhMV2Z2jMgMT2Z2mMdMS2]2nMbMR2]2PNbMR2\\2oMcMR2\\2oMcMR2\\2nMdMS2Z2oMeMR2Z2oMeMR2Z2oMeMR2Y2PNfMQ2X2iLnMjMJ^5V2bLkN]MnNR6U2\\LAd3<WLKj3PObI4_2R1P4jNcI0W2^1V4bNeILo1i1]4ZNgIIg1S2d4SNgIE`1`2i4jMjIBY1i2n4eMkI^OU1P3Q5bMlIZOP1X3U5]MnIWOj0a3X5XMXKl2i4SMSKQ3n4oLmJV3S5hLjJ\\3W5WLQKm3P74K5K4M4M2O2M2O2M201O001O02O000O101N100O1fLlGc1U8[NmGc1T8\\NoGb1Q8]NQHa1P8]NSHb1n7\\NUHb1k7]NWHa1j7^NXHa1i7\\NZHb1g7]NZHc1g7YN\\Hf1e7kMjHU2W7gMlHX2V7dMmH\\2S7aMPI^2R7^MRIa2n6\\MUId2f8O100O100O101E_EoMb:Q2`EmM`:R2cEkM^:U29O100O101N100O10000O109Fb0_Ob0]Ob0_O`iY4"}, "label": "the police bike in the very front with c - 21 on it"}]}
{"id": 473745, "image": "COCO_train2014_000000473745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown couch with white and red pillow on top of it\"."}], "task": "refering", "answer_template": "The \"brown couch with white and red pillow on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.7324590163934426, 0.38125, 1.0], "iscrowd": 0, "area": 18160.67195, "rle": {"size": [427, 640], "counts": "\\;j0`<V1kN0O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1001O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O00O1O1N2O1O1O1N2O1O1N2O1O1N2O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1OO1O100O1O1O100O1O1O1O100O1O1001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O0I]gT5"}, "label": "brown couch with white and red pillow on top of it"}]}
{"id": 473745, "image": "COCO_train2014_000000473745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"seat with red pillow\"."}], "task": "refering", "answer_template": "The \"seat with red pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0, 0.7324590163934426, 0.38125, 1.0], "iscrowd": 0, "area": 18160.67195, "rle": {"size": [427, 640], "counts": "\\;j0`<V1kN0O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1001O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O00O1O1N2O1O1O1N2O1O1N2O1O1N2O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1OO1O100O1O1O100O1O1O1O100O1O1001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O0I]gT5"}, "label": "seat with red pillow"}]}
{"id": 23194, "image": "COCO_train2014_000000023194.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white horse in the middle\"."}], "task": "refering", "answer_template": "The \"the white horse in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 164, 165, 166, 187, 188, 189, 190, 191, 211, 212, 213, 214, 235, 236, 237, 238, 258, 259, 260, 261, 262, 280, 281, 282], "bbox": [0.017734375, 0.25377083333333333, 0.411296875, 0.7419375], "iscrowd": 0, "area": 19840.798600000006, "rle": {"size": [480, 640], "counts": "_[53R>k0L4gNoN`DU1^;TOVDP1i;YOkCk0T<^O`Cg0_<m0O2N1O1O2N1O2O0O1O2N1O101O0010N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2O000100O1O2O0O1O1O100O1O2O0O1O100O1O100O2N100O1O1O2O1N2N3N1N2N100O1O101N1O1O100O2N100O1O101N1O100O1O2O0O2N2N101N2N101N2N1O2O1N1O2N2O0OXNnDDP;;^EZO`1gNm6l1nGVOQ1TOm6e1UHXOi0XOo6^1[H]O`0ZOR7X1aH@7ZOY7S1dHDNZO_7Q1eHHFYOe7m0iHK]OZOj7j0kHOUOXOR8f0mH4kNXOX8c0PI6cNYO]8?SI;ZNWOe8<TI`1l6^NTId1l6[NTIe1m6YNSIi1n6UNQIm1o6QNQIP2P7mMQIU2o6fMTI\\2m6_MUIb2l6YMXIh2h6TMZIm2h6nLZIT3f6gL]I[3R91O2O10O1000O010000O10O10O1000O010000O10O10O2O1N2O1O1N2O1O1N2O1O1N101O0O2O1O0O2O1O0O2O1O0O2O1O0O2O1N101O2M3N2N2M3N2N2M3N1O2M3N2N2M3N1O2M3N2N2M3N1O2M2O2M2O2N2M2O2N0O10000O1006K6I2OO001N101O001N101O1N2O2M3N2M4Mm\\`5"}, "label": "the white horse in the middle"}]}
{"id": 23194, "image": "COCO_train2014_000000023194.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse connected to the carriage between the other two horses\"."}], "task": "refering", "answer_template": "The \"horse connected to the carriage between the other two horses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 164, 165, 166, 187, 188, 189, 190, 191, 211, 212, 213, 214, 235, 236, 237, 238, 258, 259, 260, 261, 262, 280, 281, 282], "bbox": [0.017734375, 0.25377083333333333, 0.411296875, 0.7419375], "iscrowd": 0, "area": 19840.798600000006, "rle": {"size": [480, 640], "counts": "_[53R>k0L4gNoN`DU1^;TOVDP1i;YOkCk0T<^O`Cg0_<m0O2N1O1O2N1O2O0O1O2N1O101O0010N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2O000100O1O2O0O1O1O100O1O2O0O1O100O1O100O2N100O1O1O2O1N2N3N1N2N100O1O101N1O1O100O2N100O1O101N1O100O1O2O0O2N2N101N2N101N2N1O2O1N1O2N2O0OXNnDDP;;^EZO`1gNm6l1nGVOQ1TOm6e1UHXOi0XOo6^1[H]O`0ZOR7X1aH@7ZOY7S1dHDNZO_7Q1eHHFYOe7m0iHK]OZOj7j0kHOUOXOR8f0mH4kNXOX8c0PI6cNYO]8?SI;ZNWOe8<TI`1l6^NTId1l6[NTIe1m6YNSIi1n6UNQIm1o6QNQIP2P7mMQIU2o6fMTI\\2m6_MUIb2l6YMXIh2h6TMZIm2h6nLZIT3f6gL]I[3R91O2O10O1000O010000O10O10O1000O010000O10O10O2O1N2O1O1N2O1O1N2O1O1N101O0O2O1O0O2O1O0O2O1O0O2O1O0O2O1N101O2M3N2N2M3N2N2M3N1O2M3N2N2M3N1O2M3N2N2M3N1O2M2O2M2O2N2M2O2N0O10000O1006K6I2OO001N101O001N101O1N2O2M3N2M4Mm\\`5"}, "label": "horse connected to the carriage between the other two horses"}]}
{"id": 23194, "image": "COCO_train2014_000000023194.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse in front\"."}], "task": "refering", "answer_template": "The \"horse in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 268, 269, 270, 286, 287, 291, 292, 293, 309, 310, 314, 315, 316, 332, 333, 337, 338, 339, 355, 356, 378, 379], "bbox": [0.21175000000000002, 0.25, 0.77909375, 0.9568958333333334], "iscrowd": 0, "area": 54682.72239999999, "rle": {"size": [480, 640], "counts": "]oo17a>:G8H8H9M2L5L3L3N2N2M3N1N0O1M4M2M3M4N11O1O20O2N2O1N3M2O1N2N3N1N2N3N1N2N2O2M2N2O1010O1O1O100O001O01O00001O010O1O1O1O10O10O10000O1000000O100O1O1O100O001O0010O02N4L3N3L4L4L4L3N0O1O1O10O01O1O1O010O1O1O10O01O1O1O010O1O1O1O010O1O1O10O0100O10O10O100000O010000O10O1000O1000O01000000O1000O01000000O1O0bImJP3S5]LaKa3`4nKRLP4o3hKZLV4o3[KZLc4n3mJ[LQ5o3_JYL_5P4QJWLP6\\601O01O00001O000010O000001O00001O01O01O00001O0O1C=D=B=D<C>B=D<C>C<C=C>K40001O000O101O00001N10001O0O101O00000O2O00001O0O101O000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O1000000000000000000000O100000O100000000000000000000000kG`Lj4`3TKbLl4^3SKcLm4]3QKfLn4Z3QKgLo4Y3oJiLQ5W3nJjLR5V3mJlLR5T3lJnLT5R3kJoLU5Q3iJRMV5n2iJSMW5m2gJUMY5k2fJVMZ5i2fJYMY5g2eJ[M[5e2dJ\\M\\5d2bJ^M^5b2aJ`M^5`2`JbM`5^2_JcMa5]2^JeMa5[2]JgMc5Y2\\JhMd5X2ZJjMf5V2YJlMf5T2XJnMh5R2WJoMi5c3dH^L\\7d5O00001O000000000000000000000000000000000000000000001O0000000000000000000000000001O5K4L5K4L5J5J7I7ZLjG6\\8@XH0o7EeHKa7KRIFU7O_IAg65mI[OZ6:]4FUkQ2"}, "label": "horse in front"}]}
{"id": 23194, "image": "COCO_train2014_000000023194.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse with carriage toward front of picture\"."}], "task": "refering", "answer_template": "The \"horse with carriage toward front of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 268, 269, 270, 286, 287, 291, 292, 293, 309, 310, 314, 315, 316, 332, 333, 337, 338, 339, 355, 356, 378, 379], "bbox": [0.21175000000000002, 0.25, 0.77909375, 0.9568958333333334], "iscrowd": 0, "area": 54682.72239999999, "rle": {"size": [480, 640], "counts": "]oo17a>:G8H8H9M2L5L3L3N2N2M3N1N0O1M4M2M3M4N11O1O20O2N2O1N3M2O1N2N3N1N2N3N1N2N2O2M2N2O1010O1O1O100O001O01O00001O010O1O1O1O10O10O10000O1000000O100O1O1O100O001O0010O02N4L3N3L4L4L4L3N0O1O1O10O01O1O1O010O1O1O10O01O1O1O010O1O1O1O010O1O1O10O0100O10O10O100000O010000O10O1000O1000O01000000O1000O01000000O1O0bImJP3S5]LaKa3`4nKRLP4o3hKZLV4o3[KZLc4n3mJ[LQ5o3_JYL_5P4QJWLP6\\601O01O00001O000010O000001O00001O01O01O00001O0O1C=D=B=D<C>B=D<C>C<C=C>K40001O000O101O00001N10001O0O101O00000O2O00001O0O101O000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O1000000000000000000000O100000O100000000000000000000000kG`Lj4`3TKbLl4^3SKcLm4]3QKfLn4Z3QKgLo4Y3oJiLQ5W3nJjLR5V3mJlLR5T3lJnLT5R3kJoLU5Q3iJRMV5n2iJSMW5m2gJUMY5k2fJVMZ5i2fJYMY5g2eJ[M[5e2dJ\\M\\5d2bJ^M^5b2aJ`M^5`2`JbM`5^2_JcMa5]2^JeMa5[2]JgMc5Y2\\JhMd5X2ZJjMf5V2YJlMf5T2XJnMh5R2WJoMi5c3dH^L\\7d5O00001O000000000000000000000000000000000000000000001O0000000000000000000000000001O5K4L5K4L5J5J7I7ZLjG6\\8@XH0o7EeHKa7KRIFU7O_IAg65mI[OZ6:]4FUkQ2"}, "label": "horse with carriage toward front of picture"}]}
{"id": 23194, "image": "COCO_train2014_000000023194.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse behind two other horses that are standing in line with it , being partially obscured by the middle horse\"."}], "task": "refering", "answer_template": "The \"a horse behind two other horses that are standing in line with it , being partially obscured by the middle horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 140, 161, 162, 163, 164, 184, 185, 186, 187, 207, 209, 210, 230, 232, 233, 253, 255, 256], "bbox": [0.0, 0.37254166666666666, 0.1609375, 0.7072916666666667], "iscrowd": 0, "area": 8650.578599999999, "rle": {"size": [480, 640], "counts": "S8\\2W:]200O10000000000O1000lKlE\\3T:`LPF`3P:\\LUFc3j9ZL[Fe3e9WL`FV1@o0Q:gMdFY1`0Lh94aGgN\\8X1g2O1N2O001O1O1O1O001O1N2O1OO100O1O1O1O1O001O1O100O1O1O1O1O001O1O100O1O1O3M4L3M4L3gDTNY9P2`FTN_9P2YFUNg9m1RFXNm9l1kEXNU:k1dEZN[:j1]E[Nb:h1WE]Nh:n2O1O1O1O11O1N2O1O2N1O1O2N1O1O2N1N2O2N1O1O1L5K4K5L5K4L4K6K4L4L5J:G<D;D=D<Dn^k7"}, "label": "a horse behind two other horses that are standing in line with it , being partially obscured by the middle horse"}]}
{"id": 23194, "image": "COCO_train2014_000000023194.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse to the far left of two others\"."}], "task": "refering", "answer_template": "The \"a brown horse to the far left of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 140, 161, 162, 163, 164, 184, 185, 186, 187, 207, 209, 210, 230, 232, 233, 253, 255, 256], "bbox": [0.0, 0.37254166666666666, 0.1609375, 0.7072916666666667], "iscrowd": 0, "area": 8650.578599999999, "rle": {"size": [480, 640], "counts": "S8\\2W:]200O10000000000O1000lKlE\\3T:`LPF`3P:\\LUFc3j9ZL[Fe3e9WL`FV1@o0Q:gMdFY1`0Lh94aGgN\\8X1g2O1N2O001O1O1O1O001O1N2O1OO100O1O1O1O1O001O1O100O1O1O1O1O001O1O100O1O1O3M4L3M4L3gDTNY9P2`FTN_9P2YFUNg9m1RFXNm9l1kEXNU:k1dEZN[:j1]E[Nb:h1WE]Nh:n2O1O1O1O11O1N2O1O2N1O1O2N1O1O2N1N2O2N1O1O1L5K4K5L5K4L4K6K4L4L5J:G<D;D=D<Dn^k7"}, "label": "a brown horse to the far left of two others"}]}
{"id": 326299, "image": "COCO_train2014_000000326299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black computer chair in a room with a man standing up\"."}], "task": "refering", "answer_template": "The \"black computer chair in a room with a man standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375, 376, 377], "bbox": [0.0, 0.7368125, 0.24925, 0.998875], "iscrowd": 0, "area": 17028.847299999998, "rle": {"size": [640, 480], "counts": "aa08a1:d?7Z_Oj0W`0Gg_O:R`0Ml_O4Q`00n_O0o?5o_OKo?8P@Hm?<R@Dl??S@Aj?d0T@\\Oj?g0U@YOh?k0W@UOg?c2M3N2M3N2M3N2O2N10000O10000O100O10000O10000O10000O2O000O100O10000O10000O10000O0100O01000O100000000000001O000000000O1J6Jm@_KS?e43N1N2N2O1O100O10001O000O10000000000O2O000000000O1000001O0O1000000O1O1O2M2O1O3[OR@RMT`0i2n_OUMW`0e2l_OYMY`0a2j_O\\M\\`0^2g_O`M^`0[2d_OcMa`0g1R@VNT`0n0o1TOh`P7"}, "label": "black computer chair in a room with a man standing up"}]}
{"id": 326299, "image": "COCO_train2014_000000326299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black computer chair\"."}], "task": "refering", "answer_template": "The \"a black computer chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 340, 341, 342, 343, 344, 357, 358, 359, 360, 361, 374, 375, 376, 377], "bbox": [0.0, 0.7368125, 0.24925, 0.998875], "iscrowd": 0, "area": 17028.847299999998, "rle": {"size": [640, 480], "counts": "aa08a1:d?7Z_Oj0W`0Gg_O:R`0Ml_O4Q`00n_O0o?5o_OKo?8P@Hm?<R@Dl??S@Aj?d0T@\\Oj?g0U@YOh?k0W@UOg?c2M3N2M3N2M3N2O2N10000O10000O100O10000O10000O10000O2O000O100O10000O10000O10000O0100O01000O100000000000001O000000000O1J6Jm@_KS?e43N1N2N2O1O100O10001O000O10000000000O2O000000000O1000001O0O1000000O1O1O2M2O1O3[OR@RMT`0i2n_OUMW`0e2l_OYMY`0a2j_O\\M\\`0^2g_O`M^`0[2d_OcMa`0g1R@VNT`0n0o1TOh`P7"}, "label": "a black computer chair"}]}
{"id": 326299, "image": "COCO_train2014_000000326299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with raised arms wearing a green shirt\"."}], "task": "refering", "answer_template": "The \"a man with raised arms wearing a green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 44, 45, 54, 55, 61, 62, 63, 70, 71, 72, 79, 80, 81, 87, 88, 89, 92, 93, 94, 96, 97, 98, 99, 104, 105, 106, 109, 110, 111, 113, 114, 115, 116, 121, 122, 123, 126, 127, 128, 130, 131, 132, 138, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 300, 301, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 346, 347, 348, 349, 350, 351, 352, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.13783333333333334, 0.062234374999999995, 0.8539375, 0.983578125], "iscrowd": 0, "area": 97475.48050000002, "rle": {"size": [640, 480], "counts": "i_Y18mb0l0E;G9F:G9G9F:G9G9G9F:G9I7N2N2N2O0O2N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2N101N2N2N2N2O1N2N2O1O1O1O1O1ORLVCg0h<WOhC=W<@nC?Q<^OTDa0k;\\OYDd0f;YO_Df0`;WOeDh0Z;UOkDj0T;SOPEm0o:POVEo0h:oN]Eo0c:nNaER1^:kNgET1X:iNmEV1R:gNRFY1m9dNXF[1g9bN]F]1S>100O10O0100O10O0100O1O010O100O010O100O010O10O01O100O001O102M4L4L4M3L5K4gEkL_3Y3XLbMS3a2dLWNi2m1oLWNQ3m1fLXNZ3l1]LYNd3j1TLZNl3k1jKZNV4j1aK[N_4i1YK[Nh4h1oJ]NQ5g1bJbN^5b1QJkNo5Y1`ITOa6o0oH]OQ7h0]HEc7?lGNT8o51O0000001O0000001O00_M^DdLb;[3lDYLS;f3[EmKe:Q4kEaKU:^4YFUKg9j4`FQK_9n4fFnJZ9P5mFkJS9T5SGgJm8X5XGeJg8Z5P3N2O1O1O1N2O1O1000000O1000000000000000O100000O10000000000000000000000O10000000000001O001O00001O001O001O00001O001O001O00001O001O00001O0hAmKc<k5M4L3M4LdJcCY3Y<gLgCiN0c3V<cMlCiNOg3Q<_MRDiNNj3m;\\MWDiNMn3h;YM[DiNNm3g;YM]DiNMl3g;ZM^DhNMl3f;[M^DiNMj3e;^M_DgNMi3e;_M`DgNLh3e;`MaDgNKh3d;`MbDhNKf3d;aMcDgNKf3c;cMcDeNLf3b;dMdDcNMg3`;eMdDbNOg3^;fMeDaNOh3\\;gMfD^N1i3Z;hMfD]N3i3X;iMgD\\N3i3W;jMhDZN5i3T;lMoER2R:nMoEQ2Q:nMQFP2P:PNRFm1o9TNQFj1P:VNQFh1P:YNQFd1P:\\NQFc1o9^NQF`1P:aNPF]1Q:cNQFZ1P:gNPFW1Q:iNPFU1Q:lNoES1Q:mNQFP1P:QOPFm0Q:TOoEj0R:VOoEh0R:YOoEc0S:]OnE?U:BkE;W:EkE7W:JiE3Y:OfEM]:7`EFb:>\\E^Of:f0WEWOk:l0SEPOP;T1mDiNU;[1iDaNY;c1eDYN];j1eDoM];U2dDcMn0Dm6m2VHXMn07a6d2cHnLm0i0V6Z2PIgLj0\\1k5n1gL^Nn2c1PMhNf2Z1VMSO_2n0_M^OV2c0hMHn19oM4f1MXN?n01oN;76G9[O8c0l0jMES2`1ZLQOd3g8N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3M3N2M3M4L3N2M3M3M3N2M3M3M3N2M3M4L3M4M3L3M4L3N3L4L3M4M2M4L3M4L4M2M4L3Mab[1"}, "label": "a man with raised arms wearing a green shirt"}]}
{"id": 326299, "image": "COCO_train2014_000000326299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man poses for a picture holding a video game controller\"."}], "task": "refering", "answer_template": "The \"a man poses for a picture holding a video game controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 44, 45, 54, 55, 61, 62, 63, 70, 71, 72, 79, 80, 81, 87, 88, 89, 92, 93, 94, 96, 97, 98, 99, 104, 105, 106, 109, 110, 111, 113, 114, 115, 116, 121, 122, 123, 126, 127, 128, 130, 131, 132, 138, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 267, 278, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 300, 301, 312, 313, 314, 315, 316, 317, 318, 329, 330, 331, 332, 333, 334, 335, 346, 347, 348, 349, 350, 351, 352, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.13783333333333334, 0.062234374999999995, 0.8539375, 0.983578125], "iscrowd": 0, "area": 97475.48050000002, "rle": {"size": [640, 480], "counts": "i_Y18mb0l0E;G9F:G9G9F:G9G9G9F:G9I7N2N2N2O0O2N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2N101N2N2N2N2O1N2N2O1O1O1O1O1ORLVCg0h<WOhC=W<@nC?Q<^OTDa0k;\\OYDd0f;YO_Df0`;WOeDh0Z;UOkDj0T;SOPEm0o:POVEo0h:oN]Eo0c:nNaER1^:kNgET1X:iNmEV1R:gNRFY1m9dNXF[1g9bN]F]1S>100O10O0100O10O0100O1O010O100O010O100O010O10O01O100O001O102M4L4L4M3L5K4gEkL_3Y3XLbMS3a2dLWNi2m1oLWNQ3m1fLXNZ3l1]LYNd3j1TLZNl3k1jKZNV4j1aK[N_4i1YK[Nh4h1oJ]NQ5g1bJbN^5b1QJkNo5Y1`ITOa6o0oH]OQ7h0]HEc7?lGNT8o51O0000001O0000001O00_M^DdLb;[3lDYLS;f3[EmKe:Q4kEaKU:^4YFUKg9j4`FQK_9n4fFnJZ9P5mFkJS9T5SGgJm8X5XGeJg8Z5P3N2O1O1O1N2O1O1000000O1000000000000000O100000O10000000000000000000000O10000000000001O001O00001O001O001O00001O001O001O00001O001O00001O0hAmKc<k5M4L3M4LdJcCY3Y<gLgCiN0c3V<cMlCiNOg3Q<_MRDiNNj3m;\\MWDiNMn3h;YM[DiNNm3g;YM]DiNMl3g;ZM^DhNMl3f;[M^DiNMj3e;^M_DgNMi3e;_M`DgNLh3e;`MaDgNKh3d;`MbDhNKf3d;aMcDgNKf3c;cMcDeNLf3b;dMdDcNMg3`;eMdDbNOg3^;fMeDaNOh3\\;gMfD^N1i3Z;hMfD]N3i3X;iMgD\\N3i3W;jMhDZN5i3T;lMoER2R:nMoEQ2Q:nMQFP2P:PNRFm1o9TNQFj1P:VNQFh1P:YNQFd1P:\\NQFc1o9^NQF`1P:aNPF]1Q:cNQFZ1P:gNPFW1Q:iNPFU1Q:lNoES1Q:mNQFP1P:QOPFm0Q:TOoEj0R:VOoEh0R:YOoEc0S:]OnE?U:BkE;W:EkE7W:JiE3Y:OfEM]:7`EFb:>\\E^Of:f0WEWOk:l0SEPOP;T1mDiNU;[1iDaNY;c1eDYN];j1eDoM];U2dDcMn0Dm6m2VHXMn07a6d2cHnLm0i0V6Z2PIgLj0\\1k5n1gL^Nn2c1PMhNf2Z1VMSO_2n0_M^OV2c0hMHn19oM4f1MXN?n01oN;76G9[O8c0l0jMES2`1ZLQOd3g8N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3M3N2M3M4L3N2M3M3M3N2M3M3M3N2M3M4L3M4M3L3M4L3N3L4L3M4M2M4L3M4L4M2M4L3Mab[1"}, "label": "a man poses for a picture holding a video game controller"}]}
{"id": 113310, "image": "COCO_train2014_000000113310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large airfrance plane in front while other in background\"."}], "task": "refering", "answer_template": "The \"large airfrance plane in front while other in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 93, 94, 95, 116, 117, 118, 119, 140, 141, 142, 143, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.015125, 0.24791569086651052, 0.9803593749999999, 0.746791569086651], "iscrowd": 0, "area": 44600.14874999999, "rle": {"size": [427, 640], "counts": "m\\4;P=0010O00000000000001O000000000000001O000000000000001RLDkJ<T5HiJ8W5KfJ5Y5OdJ1[52cJN\\56aJJ_59^JGa5=\\JDc5?ZJAe5b0YJ^Og5e0VJ[Oj5h0SJXOm5k0PJUOo5n0PJQOP6R1mInNS6U1jIlNU6W1hIiNX6Y1fIgNY6]1dIcN\\6`1aI`N_6c1^I]Nb6e1\\I[Nd6h1YIXNf6l1WIUNh6n1UIRNk6P2SIPNm6S2PImMQ7U2lHkMU7W2hHiMZ7W2dHiM]7Y2`HgMa7[2\\HfMd7\\2YHdMh7]2VHcMk7_2RHbMn7`2oGaMQ8a2lG`MT8b2iG_MX8a2fG`MZ8b2cG`M\\8b2aG_M_8c2^G^Mb8c2\\G^Md8d2YG]Mg8e2VG\\Mj8d3O01O00000001O00000001O0O1O1O100O1O1O1O1O2N100O1O1O1O1O1O101N1O1O1O1O1O100O1O2N1O1O1O100O1O1O1O2N1O100O1O1O1O1O1O1O2O0O1O1O1O1O1O100O2N1O1O1O1O100O1O1O2N1O1O100O1O1O1O1O2N100O1O1O1O1O1O101N1O1O1O000001O000000000000001O0000000000001O0000000000001O000001O000001O0000000000001O0000000000001O000000000000001O0000000001O01O0000000000001O0000000000001O0000000000001O0000000000001O0001O00000001O000000000000001O0000000000001O0000000000001O00000001O0001O0000000000001O0000000000001O0000000000001O0000000000000010O000000000001O00000000000000000000000000000000000000000000000001N1000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000O100000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000000O10000000000000001N10000000000000000O10000000000000000O1000001O000000000O10000000000000000O1000000000001O000O10000000000000000O10000000000000000O101O00000000000O10000000000000001N2O001O1O1O1O1O1O1N2O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1N2O1N2O1N2N2O0O2N2O1N2N2O1N2N2O1N2N]T5"}, "label": "large airfrance plane in front while other in background"}]}
{"id": 113310, "image": "COCO_train2014_000000113310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a full picture of an airplane\"."}], "task": "refering", "answer_template": "The \"a full picture of an airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 93, 94, 95, 116, 117, 118, 119, 140, 141, 142, 143, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.015125, 0.24791569086651052, 0.9803593749999999, 0.746791569086651], "iscrowd": 0, "area": 44600.14874999999, "rle": {"size": [427, 640], "counts": "m\\4;P=0010O00000000000001O000000000000001O000000000000001RLDkJ<T5HiJ8W5KfJ5Y5OdJ1[52cJN\\56aJJ_59^JGa5=\\JDc5?ZJAe5b0YJ^Og5e0VJ[Oj5h0SJXOm5k0PJUOo5n0PJQOP6R1mInNS6U1jIlNU6W1hIiNX6Y1fIgNY6]1dIcN\\6`1aI`N_6c1^I]Nb6e1\\I[Nd6h1YIXNf6l1WIUNh6n1UIRNk6P2SIPNm6S2PImMQ7U2lHkMU7W2hHiMZ7W2dHiM]7Y2`HgMa7[2\\HfMd7\\2YHdMh7]2VHcMk7_2RHbMn7`2oGaMQ8a2lG`MT8b2iG_MX8a2fG`MZ8b2cG`M\\8b2aG_M_8c2^G^Mb8c2\\G^Md8d2YG]Mg8e2VG\\Mj8d3O01O00000001O00000001O0O1O1O100O1O1O1O1O2N100O1O1O1O1O1O101N1O1O1O1O1O100O1O2N1O1O1O100O1O1O1O2N1O100O1O1O1O1O1O1O2O0O1O1O1O1O1O100O2N1O1O1O1O100O1O1O2N1O1O100O1O1O1O1O2N100O1O1O1O1O1O101N1O1O1O000001O000000000000001O0000000000001O0000000000001O000001O000001O0000000000001O0000000000001O000000000000001O0000000001O01O0000000000001O0000000000001O0000000000001O0000000000001O0001O00000001O000000000000001O0000000000001O0000000000001O00000001O0001O0000000000001O0000000000001O0000000000001O0000000000000010O000000000001O00000000000000000000000000000000000000000000000001N1000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000O100000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000000O10000000000000001N10000000000000000O10000000000000000O1000001O000000000O10000000000000000O1000000000001O000O10000000000000000O10000000000000000O101O00000000000O10000000000000001N2O001O1O1O1O1O1O1N2O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1N2O1N2O1N2N2O0O2N2O1N2N2O1N2N2O1N2N]T5"}, "label": "a full picture of an airplane"}]}
{"id": 555683, "image": "COCO_train2014_000000555683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"all tennis balls except the one touching the line and the one on the far right\"."}], "task": "refering", "answer_template": "The \"all tennis balls except the one touching the line and the one on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 89, 112, 155, 156, 178, 179, 193, 194, 224, 225, 247, 248, 290, 291, 313, 314, 337], "bbox": [0.41334375, 0.07333333333333333, 0.9176249999999999, 0.8343333333333333], "iscrowd": 0, "area": 2994.7365499999983, "rle": {"size": [480, 640], "counts": "X^l3;c>3N2N2O0O2N1O101O000000000000O2O002M2O2M3M3M3KSdd1Oi[[N;L3M4L2N2O0mA_Of=b0VBEg=l0N2N2N0000XO_B2`=g0000VO`B5`=J`B6`=e01O002M3SOZB<g=B[B=g=^O]B`0g=[O\\Bc0Q>O2N2O1N4J7Hdei02]ZVO5cHOUMOcM6R50XMMcM6S5NYMNaM7S5MZMNdM5P5L]M1^M7T5H]M3^M5VNCd65gM4_M4UNHa60kM4_M3UNL^6NmM5_M0WNM]6NmM5_M0VNN_6LmM5^M1VNO_6InM6^Mf0e4ROnM8]Mf0f4POnM:\\Mf0g4nNoM:[Mh0^9WOdFh0\\9XOdFh0]9VOeFi0_9ROcF7nM5l=JTB6m=ISB7Q>CPB>Z>N2N3M2N2N2HiXg01UgXO9N101N101N101O000000O101O001N1O2M3N`gh0"}, "label": "all tennis balls except the one touching the line and the one on the far right"}]}
{"id": 555683, "image": "COCO_train2014_000000555683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby on the ground\"."}], "task": "refering", "answer_template": "The \"the baby on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 94, 95, 96, 116, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 257, 258], "bbox": [0.03284375, 0.22014583333333335, 0.26359375, 0.7082291666666667], "iscrowd": 0, "area": 19795.154799999997, "rle": {"size": [480, 640], "counts": "YS:2T>3_B1Z=5dBOW=4fB0W=1gB3V=OhB4U=MiB7T=KjB7T=KiB8U=IiB9V=j0N2N2N2N2N2L4L4L4POP1L4L4M3L4L4M3M3O1O000000001O000O1O1O2N1O1O1M4F9H8G9L4N2N3M2N2N200O100O101N1O100O100O101N100000000000001O00000000O100O100O100O100O2O1N2O0O2O1N2O1N2O0O2OQG\\J=2Z7`5XHdJN8k7S5THaKk7^4THdKm7[4PHhKR8U4mGnKS8U2UH_Om78]HEd74eHI`70dHMd8jN\\GX1e8dN[G_1f8^NZGd1g8XNZGj1g8SN[Gl1f8TNZGl1g8SNYGm1g8SNYGm1h8QNYGo1f8RNZGm1g8SNYGm1g8SNXGn1h8QNYGo1g8QNYGo1g8QNYGn1h8QNYGo1g8QNYGo1g8QNYGn1h8QNYGo1g8QNkG\\1V8cNlG\\1S8eNnGZ1S8dNoGZ1T8cNnG\\1S8cNnG\\1T8aNnG]1T8bNmG]1U8^NoGa1R8[NRHc1Q8WNTHh1m7TNWHP1aMCP=0]COc<FhC:`=N2O1N2O0O2N1O2O0O2N1LV_l6"}, "label": "the baby on the ground"}]}
{"id": 555683, "image": "COCO_train2014_000000555683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young child crawling on a tennis court\"."}], "task": "refering", "answer_template": "The \"a young child crawling on a tennis court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 94, 95, 96, 116, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 257, 258], "bbox": [0.03284375, 0.22014583333333335, 0.26359375, 0.7082291666666667], "iscrowd": 0, "area": 19795.154799999997, "rle": {"size": [480, 640], "counts": "YS:2T>3_B1Z=5dBOW=4fB0W=1gB3V=OhB4U=MiB7T=KjB7T=KiB8U=IiB9V=j0N2N2N2N2N2L4L4L4POP1L4L4M3L4L4M3M3O1O000000001O000O1O1O2N1O1O1M4F9H8G9L4N2N3M2N2N200O100O101N1O100O100O101N100000000000001O00000000O100O100O100O100O2O1N2O0O2O1N2O1N2O0O2OQG\\J=2Z7`5XHdJN8k7S5THaKk7^4THdKm7[4PHhKR8U4mGnKS8U2UH_Om78]HEd74eHI`70dHMd8jN\\GX1e8dN[G_1f8^NZGd1g8XNZGj1g8SN[Gl1f8TNZGl1g8SNYGm1g8SNYGm1h8QNYGo1f8RNZGm1g8SNYGm1g8SNXGn1h8QNYGo1g8QNYGo1g8QNYGn1h8QNYGo1g8QNYGo1g8QNYGn1h8QNYGo1g8QNkG\\1V8cNlG\\1S8eNnGZ1S8dNoGZ1T8cNnG\\1S8cNnG\\1T8aNnG]1T8bNmG]1U8^NoGa1R8[NRHc1Q8WNTHh1m7TNWHP1aMCP=0]COc<FhC:`=N2O1N2O0O2N1O2O0O2N1LV_l6"}, "label": "a young child crawling on a tennis court"}]}
{"id": 244387, "image": "COCO_train2014_000000244387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl standing on the ground next to the men\"."}], "task": "refering", "answer_template": "The \"a girl standing on the ground next to the men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 162, 163, 164, 178, 179, 180, 194, 195, 196, 210, 211, 212, 226, 227, 228, 242, 243, 244, 258, 259, 260, 274, 275, 276, 277, 290, 291, 292, 293, 306, 307, 308, 309, 322, 323, 324, 325], "bbox": [0.14458515283842793, 0.385734375, 0.3215065502183406, 0.8921875], "iscrowd": 0, "area": 15625.447549999999, "rle": {"size": [640, 458], "counts": "SaY12i41V:7^E4\\:4VE7_NC\\9=iG:YNIj95`G>SNNY:LVGa0oM4g:CmFc0kM9T;^ObFV3Z9RMYFV3c9TMnEV3n9RMeEW3W:RMZEW3b:V3M3O1N2O0O2O1N2OkNnE]HR:[7`FZH`9`7kF\\HU9\\7SGdHl8V7\\GiHd8o6dGPI\\8j6lGSIT8g6THXIj7d6^HYIa7b6gH]IX7]6QI`In6\\6]3J5N22011N101N1O2O0O21N2O0O2O1N2O1N2O1NoJ^Be3a=XLfBe3X=ZLnBd3P=YLVCe3h<YL_Ce3^<YLiCe3U<XLRDe3l;YL[De3c;XLdDe3Z;YLlDf3R;XLTEf3i:XL^Ee3a=0`AjKo<X4hBnKY=U4]BQLd=R4RBTLo=n3hAYLX>g41O1O1`LbA\\1_>oMYBm1h=^MPC^2Q=mLfCo2j>K5L3L5K5KZ1fN^kQ6"}, "label": "a girl standing on the ground next to the men"}]}
{"id": 244387, "image": "COCO_train2014_000000244387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl reaching up to the sky with her arm\"."}], "task": "refering", "answer_template": "The \"girl reaching up to the sky with her arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 162, 163, 164, 178, 179, 180, 194, 195, 196, 210, 211, 212, 226, 227, 228, 242, 243, 244, 258, 259, 260, 274, 275, 276, 277, 290, 291, 292, 293, 306, 307, 308, 309, 322, 323, 324, 325], "bbox": [0.14458515283842793, 0.385734375, 0.3215065502183406, 0.8921875], "iscrowd": 0, "area": 15625.447549999999, "rle": {"size": [640, 458], "counts": "SaY12i41V:7^E4\\:4VE7_NC\\9=iG:YNIj95`G>SNNY:LVGa0oM4g:CmFc0kM9T;^ObFV3Z9RMYFV3c9TMnEV3n9RMeEW3W:RMZEW3b:V3M3O1N2O0O2O1N2OkNnE]HR:[7`FZH`9`7kF\\HU9\\7SGdHl8V7\\GiHd8o6dGPI\\8j6lGSIT8g6THXIj7d6^HYIa7b6gH]IX7]6QI`In6\\6]3J5N22011N101N1O2O0O21N2O0O2O1N2O1N2O1NoJ^Be3a=XLfBe3X=ZLnBd3P=YLVCe3h<YL_Ce3^<YLiCe3U<XLRDe3l;YL[De3c;XLdDe3Z;YLlDf3R;XLTEf3i:XL^Ee3a=0`AjKo<X4hBnKY=U4]BQLd=R4RBTLo=n3hAYLX>g41O1O1`LbA\\1_>oMYBm1h=^MPC^2Q=mLfCo2j>K5L3L5K5KZ1fN^kQ6"}, "label": "girl reaching up to the sky with her arm"}]}
{"id": 244387, "image": "COCO_train2014_000000244387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a blue shirt jumping\"."}], "task": "refering", "answer_template": "The \"a boy with a blue shirt jumping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 53, 54, 69, 70, 85, 86, 87, 101, 102, 103, 116, 117, 118, 119, 132, 133, 134, 148, 149, 150, 164, 165, 166, 180, 181, 182, 196, 197, 198, 212, 213, 214, 229, 230, 231, 245, 246, 247, 261, 262, 263, 264, 278, 279], "bbox": [0.28262008733624455, 0.10112499999999999, 0.5212663755458515, 0.782015625], "iscrowd": 0, "area": 22451.00985, "rle": {"size": [640, 458], "counts": "Xl`2Y2_a0=C>B=C>E:G9G:F6I8I6J7I6I8E:F:G:E:F;E:G9G:E:G:F9J7J5K6K5J5K6K4K6J6K4K6J5L5J6J5L5N12O0003N3LTF\\H]7a7WHfHQ8V7eGQI^8P7TGYIn8i6oEWJQ:k7100O2O0O100O2O0O100ObKjFUNU9e1UGWNj8e1_GWNa8c1hGZNW8ZNkFk0W1f0o7^NXG9S1T1e7cNeGHo0`1]7gNRHVOj0P2S7kN_HdNg0\\2k6oNlHRNc0j2a6TOSLh0n3WOWLd0i3\\O\\L`0d3_OaL<_3EdL7]3HhL3X3MSMHo26\\M@i2:cMYOc2c0gMRO_2h0gMRO^2h0dMWO_2f0bMYOa2d0_M\\Od2a0]M^Of2?[M@h2=XMCk2:VMEn27SMHP35PMJT33mLLV31jLOY3MiL2Z3KgL4\\3IdL7_3FbL9c;O001O1N101O1O001O1O001O00k]X4"}, "label": "a boy with a blue shirt jumping"}]}
{"id": 244387, "image": "COCO_train2014_000000244387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt jumping to catch a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt jumping to catch a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 53, 54, 69, 70, 85, 86, 87, 101, 102, 103, 116, 117, 118, 119, 132, 133, 134, 148, 149, 150, 164, 165, 166, 180, 181, 182, 196, 197, 198, 212, 213, 214, 229, 230, 231, 245, 246, 247, 261, 262, 263, 264, 278, 279], "bbox": [0.28262008733624455, 0.10112499999999999, 0.5212663755458515, 0.782015625], "iscrowd": 0, "area": 22451.00985, "rle": {"size": [640, 458], "counts": "Xl`2Y2_a0=C>B=C>E:G9G:F6I8I6J7I6I8E:F:G:E:F;E:G9G:E:G:F9J7J5K6K5J5K6K4K6J6K4K6J5L5J6J5L5N12O0003N3LTF\\H]7a7WHfHQ8V7eGQI^8P7TGYIn8i6oEWJQ:k7100O2O0O100O2O0O100ObKjFUNU9e1UGWNj8e1_GWNa8c1hGZNW8ZNkFk0W1f0o7^NXG9S1T1e7cNeGHo0`1]7gNRHVOj0P2S7kN_HdNg0\\2k6oNlHRNc0j2a6TOSLh0n3WOWLd0i3\\O\\L`0d3_OaL<_3EdL7]3HhL3X3MSMHo26\\M@i2:cMYOc2c0gMRO_2h0gMRO^2h0dMWO_2f0bMYOa2d0_M\\Od2a0]M^Of2?[M@h2=XMCk2:VMEn27SMHP35PMJT33mLLV31jLOY3MiL2Z3KgL4\\3IdL7_3FbL9c;O001O1N101O1O001O1O001O00k]X4"}, "label": "a man in a blue shirt jumping to catch a frisbee"}]}
{"id": 473773, "image": "COCO_train2014_000000473773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis the person in the fanta jacket is holding\"."}], "task": "refering", "answer_template": "The \"the skis the person in the fanta jacket is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 170, 171, 172, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221], "bbox": [0.09526562499999999, 0.40076923076923077, 0.640875, 0.6333333333333333], "iscrowd": 0, "area": 8569.087749999999, "rle": {"size": [429, 640], "counts": "dgi02W=4L5L4O001N101O0O2O1O0O2O001N101O010O1O0010O01O010O1O010O001O010O0001O00010O00001O01O0001O01O01O000010O0000010O0001O00010O000010O000001O01O01O000010O0000010O0001O00010O000010O000001O01O01O00010O0000001O01O01O00001O01O01O00000010O0001O000010O01O001O1O5LZOe0000L401O000001O01O000000001O000000010O000000001O0000000010O00000001O000000001O01O0001O000000001O0001O0001O000000001O000001O01O000000001O000000010O000000001O0000000010O00000001O000000001O01OM3K5K6KbR5NbmJ2M3M3J7I6O1000010O0001O00010O00G^OgCb0R<EoC:m;JSD7l;ITD7l;JTD5m;JSD6m;JSD6m;KSD5l;KTD5l;KTD5m;KRD5n;KSD4m;LSD5l;LSD4m;LTD3m;LSD4m;MRD3n;MSD2m;NSD3l;NSD2n;MSD2m;NSD2m;ORD1n;OSD1l;OTD1m;NSD2m;OSD0m;0SD0m;1RD0m;0TDOm;0SD0m;1RDOn;1RDOn;1SDOl;2SDMo;2QDLQ<5oCGT<9lCEV<;jCCY<=gC@[<`0eC]O^<c061O010O00001O01O01O001O01O01O001O01O01O000010O01O000010O0000001O00001O0000010O0000001N101O0O2O1N101O0N3LkSP3"}, "label": "the skis the person in the fanta jacket is holding"}]}
{"id": 473773, "image": "COCO_train2014_000000473773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis being carried by a kid\"."}], "task": "refering", "answer_template": "The \"skis being carried by a kid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 170, 171, 172, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221], "bbox": [0.09526562499999999, 0.40076923076923077, 0.640875, 0.6333333333333333], "iscrowd": 0, "area": 8569.087749999999, "rle": {"size": [429, 640], "counts": "dgi02W=4L5L4O001N101O0O2O1O0O2O001N101O010O1O0010O01O010O1O010O001O010O0001O00010O00001O01O0001O01O01O000010O0000010O0001O00010O000010O000001O01O01O000010O0000010O0001O00010O000010O000001O01O01O00010O0000001O01O01O00001O01O01O00000010O0001O000010O01O001O1O5LZOe0000L401O000001O01O000000001O000000010O000000001O0000000010O00000001O000000001O01O0001O000000001O0001O0001O000000001O000001O01O000000001O000000010O000000001O0000000010O00000001O000000001O01OM3K5K6KbR5NbmJ2M3M3J7I6O1000010O0001O00010O00G^OgCb0R<EoC:m;JSD7l;ITD7l;JTD5m;JSD6m;JSD6m;KSD5l;KTD5l;KTD5m;KRD5n;KSD4m;LSD5l;LSD4m;LTD3m;LSD4m;MRD3n;MSD2m;NSD3l;NSD2n;MSD2m;NSD2m;ORD1n;OSD1l;OTD1m;NSD2m;OSD0m;0SD0m;1RD0m;0TDOm;0SD0m;1RDOn;1RDOn;1SDOl;2SDMo;2QDLQ<5oCGT<9lCEV<;jCCY<=gC@[<`0eC]O^<c061O010O00001O01O01O001O01O01O001O01O01O000010O01O000010O0000001O00001O0000010O0000001N101O0O2O1N101O0N3LkSP3"}, "label": "skis being carried by a kid"}]}
{"id": 31411, "image": "COCO_train2014_000000031411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe standing to the left of another giraffe , interacting with the other giraffe\"."}], "task": "refering", "answer_template": "The \"giraffe standing to the left of another giraffe , interacting with the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 261, 262, 280, 281, 284, 303, 304, 307, 326, 327, 330], "bbox": [0.18953124999999998, 0.3324413145539906, 0.617046875, 1.0], "iscrowd": 0, "area": 22604.8379, "rle": {"size": [426, 640], "counts": "Shb11Q=8\\NJmE=n9MfE:U:1^E6`:1SE7j:V1N2N2N2N2QNkMeHY2X7WNWHm1e7eNiG_1T8m1N2O1O100O1O1O1O1O1O100O1O1O\\O^HbKa7S4P1F:N100O2N100O2O2M3N1N3N1O2M3N1N3N2N1O2O001N2O001N2O001O0O2O1O0O2O1O0O2O001O1O01O100O00100O1O10O01O100O0O2O1N2O1O0O200O10O0100O100O100O100O100nHVMc3k2PLcMo3]2cKQN\\4P2VK^Nj4b1gINW6c300O2O0O101N100O2O00001O00000OG:G9@`0QOo0D<C=K5N2N101N2N2N2N2M3L4L4L4L3M4M3N2M3QO^ESOd:l0aEnNb:Q1bEjN`:U1dEeN_:Z1eE`N]:_1h0O1O10O0100O10O01O100O010O100O010O10O0100O10O0100O10O0100O10O0100O10O010O001O010O0001O01O0001O01O01O0001O01O00JkNXDV1g;mNWDR1j;600010O001O010O000010_OeNRE[1o:iNlDW1U;oNeDP1\\;?0O1O1O100O1O1O1O1O1O1N2O1O2N1N2O2N1O101O00000000000O2O001O1O1O1O1O1O2N1O1O001O0O2O0O2N1O2O0O2N101N1N3M2O3L_hU3"}, "label": "giraffe standing to the left of another giraffe , interacting with the other giraffe"}]}
{"id": 31411, "image": "COCO_train2014_000000031411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the profile of a giraffe licking another giraffe\"."}], "task": "refering", "answer_template": "The \"the profile of a giraffe licking another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 261, 262, 280, 281, 284, 303, 304, 307, 326, 327, 330], "bbox": [0.18953124999999998, 0.3324413145539906, 0.617046875, 1.0], "iscrowd": 0, "area": 22604.8379, "rle": {"size": [426, 640], "counts": "Shb11Q=8\\NJmE=n9MfE:U:1^E6`:1SE7j:V1N2N2N2N2QNkMeHY2X7WNWHm1e7eNiG_1T8m1N2O1O100O1O1O1O1O1O100O1O1O\\O^HbKa7S4P1F:N100O2N100O2O2M3N1N3N1O2M3N1N3N2N1O2O001N2O001N2O001O0O2O1O0O2O1O0O2O001O1O01O100O00100O1O10O01O100O0O2O1N2O1O0O200O10O0100O100O100O100O100nHVMc3k2PLcMo3]2cKQN\\4P2VK^Nj4b1gINW6c300O2O0O101N100O2O00001O00000OG:G9@`0QOo0D<C=K5N2N101N2N2N2N2M3L4L4L4L3M4M3N2M3QO^ESOd:l0aEnNb:Q1bEjN`:U1dEeN_:Z1eE`N]:_1h0O1O10O0100O10O01O100O010O100O010O10O0100O10O0100O10O0100O10O0100O10O010O001O010O0001O01O0001O01O01O0001O01O00JkNXDV1g;mNWDR1j;600010O001O010O000010_OeNRE[1o:iNlDW1U;oNeDP1\\;?0O1O1O100O1O1O1O1O1O1N2O1O2N1N2O2N1O101O00000000000O2O001O1O1O1O1O1O2N1O1O001O0O2O0O2N1O2O0O2N101N1N3M2O3L_hU3"}, "label": "the profile of a giraffe licking another giraffe"}]}
{"id": 31411, "image": "COCO_train2014_000000031411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe that is getting licked\"."}], "task": "refering", "answer_template": "The \"the giraffe that is getting licked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 151, 152, 174, 175, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335], "bbox": [0.493609375, 0.33894366197183096, 0.6716093750000001, 1.0], "iscrowd": 0, "area": 12663.372300000008, "rle": {"size": [426, 640], "counts": "UhS44R=7I7I8H7H8I7I7L5K4eEQNT9S2fFRNV9S2dFSNX9Q2cFSNZ9R2_FTN]9P2]FYN\\9X3B?A:F2N1O2N1NUOTH]Lj7a3dHTL[7j3SIlKk6R4W1O11O4L3M4L3L4M4L3M4L3M4L3M3M4L3M4K4M3M3M1O1O1O1O^NmIoKQ6b3nJPLP5m3X2M2O1N2O2N1N2O1N3N1N24L:G8G:F9G9H6I5K6jNaJWJd5b5bJ[Jc5\\5eJ_J`5]5dJ^Ja5`5aJ[J[12k1_5lL[JW1a0i1Q5TOYKh0d4TOgKg0W4UOTLe0k3WO`Lc0_3ZO\\NoNc1m0POfNo0U1D]N;_18SNGj1n5L4L4L4M;D7I4L4M4K4L4L5O0O10001N100O101O0O100O2O0O1000bVg2"}, "label": "the giraffe that is getting licked"}]}
{"id": 31411, "image": "COCO_train2014_000000031411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe is being kissed by the larger one\"."}], "task": "refering", "answer_template": "The \"the giraffe is being kissed by the larger one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 151, 152, 174, 175, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335], "bbox": [0.493609375, 0.33894366197183096, 0.6716093750000001, 1.0], "iscrowd": 0, "area": 12663.372300000008, "rle": {"size": [426, 640], "counts": "UhS44R=7I7I8H7H8I7I7L5K4eEQNT9S2fFRNV9S2dFSNX9Q2cFSNZ9R2_FTN]9P2]FYN\\9X3B?A:F2N1O2N1NUOTH]Lj7a3dHTL[7j3SIlKk6R4W1O11O4L3M4L3L4M4L3M4L3M4L3M3M4L3M4K4M3M3M1O1O1O1O^NmIoKQ6b3nJPLP5m3X2M2O1N2O2N1N2O1N3N1N24L:G8G:F9G9H6I5K6jNaJWJd5b5bJ[Jc5\\5eJ_J`5]5dJ^Ja5`5aJ[J[12k1_5lL[JW1a0i1Q5TOYKh0d4TOgKg0W4UOTLe0k3WO`Lc0_3ZO\\NoNc1m0POfNo0U1D]N;_18SNGj1n5L4L4L4M;D7I4L4M4K4L4L5O0O10001N100O101O0O100O2O0O1000bVg2"}, "label": "the giraffe is being kissed by the larger one"}]}
{"id": 481971, "image": "COCO_train2014_000000481971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue jean shorts\"."}], "task": "refering", "answer_template": "The \"man in blue jean shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 45, 68, 91, 113, 114, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229, 251, 252, 275, 298, 320, 321, 343, 344, 366, 367], "bbox": [0.897046875, 0.020083333333333335, 1.0, 0.9080833333333334], "iscrowd": 0, "area": 14295.519649999998, "rle": {"size": [480, 640], "counts": "gW]83h>6N2M3N2N1N3fH]O4d0LANa01@Lb0PO_OmJ2P6`0QOAnJ1m5a0SO@oJ0c5DlIm0b5@mJ0d5EkIl0d5@lJOc5HjIl0e5^OnJOa5m0AUOlJOb5m0BTOlJO`5P1BROnJO^5P1DROmJNk4DPK^1X5oNlJ0k4DPK^1Y5nNlJ1g4GSKZ1Z5nNlJ1f4IRKZ1[5lNlJ2g4IQKY1\\5lNlJ2h4HoJ[1^5kNjJ3g4IPKZ1^5jNkJ3f4JPKZ1V2YNk0`0XM4g4JoJY1P2aNR17XM6e4JPKZ1e1kN]1KYM6e4JPK^1S1]Og1UO`M7e4JPKi1>CU2dNhM7d4IPK_4^2aKnM7c4o4\\MkJPN7d4P5ZMiJQN8d4R5XMgJRN9e4S5WMdJSN:f4U5SMcJUN9g4W5RM`JVN:]2ROCZ6F[JXN;[2VODX6FWJZN<W2[OFV6FUJ[N;V2^OmME`1a6NVKc0RO^OW6_OhJR1RO^O`6WO^JZ1SO_Od6QOZJ_1SO@h6jNWJe1VO\\Oi6hNSJk1XOZOm8f0WGWOh8i0ZGUOg8k0ZGTOe8l0^GQOc8n0_GPOa8P1`GmNb8S1_GkNc8U1]GiNe8V1[GiNf8X1ZGfNh8Z1XGdNj8[1WG^No8c1QGUNW9j1P20O10O01UHZN\\2\\1"}, "label": "man in blue jean shorts"}]}
{"id": 481971, "image": "COCO_train2014_000000481971.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person is standing in shorts with bag to side\"."}], "task": "refering", "answer_template": "The \"person is standing in shorts with bag to side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 45, 68, 91, 113, 114, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229, 251, 252, 275, 298, 320, 321, 343, 344, 366, 367], "bbox": [0.897046875, 0.020083333333333335, 1.0, 0.9080833333333334], "iscrowd": 0, "area": 14295.519649999998, "rle": {"size": [480, 640], "counts": "gW]83h>6N2M3N2N1N3fH]O4d0LANa01@Lb0PO_OmJ2P6`0QOAnJ1m5a0SO@oJ0c5DlIm0b5@mJ0d5EkIl0d5@lJOc5HjIl0e5^OnJOa5m0AUOlJOb5m0BTOlJO`5P1BROnJO^5P1DROmJNk4DPK^1X5oNlJ0k4DPK^1Y5nNlJ1g4GSKZ1Z5nNlJ1f4IRKZ1[5lNlJ2g4IQKY1\\5lNlJ2h4HoJ[1^5kNjJ3g4IPKZ1^5jNkJ3f4JPKZ1V2YNk0`0XM4g4JoJY1P2aNR17XM6e4JPKZ1e1kN]1KYM6e4JPK^1S1]Og1UO`M7e4JPKi1>CU2dNhM7d4IPK_4^2aKnM7c4o4\\MkJPN7d4P5ZMiJQN8d4R5XMgJRN9e4S5WMdJSN:f4U5SMcJUN9g4W5RM`JVN:]2ROCZ6F[JXN;[2VODX6FWJZN<W2[OFV6FUJ[N;V2^OmME`1a6NVKc0RO^OW6_OhJR1RO^O`6WO^JZ1SO_Od6QOZJ_1SO@h6jNWJe1VO\\Oi6hNSJk1XOZOm8f0WGWOh8i0ZGUOg8k0ZGTOe8l0^GQOc8n0_GPOa8P1`GmNb8S1_GkNc8U1]GiNe8V1[GiNf8X1ZGfNh8Z1XGdNj8[1WG^No8c1QGUNW9j1P20O10O01UHZN\\2\\1"}, "label": "person is standing in shorts with bag to side"}]}
{"id": 211641, "image": "COCO_train2014_000000211641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bench to the left of other bench\"."}], "task": "refering", "answer_template": "The \"bench to the left of other bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 285], "bbox": [0.12421875, 0.7344598337950138, 0.42590625000000004, 0.9659279778393352], "iscrowd": 0, "area": 9241.4213, "rle": {"size": [361, 640], "counts": "j^l06S;8H7I7I7I8H7I7I7I8H3M000000nNVGEj8;dGWO\\8i0SHhNm7X1R10000001O0000000000001O0000000000001O008H:F1O0000\\OmFROS9n0UGjNk8V1d000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000<Dg0YO2N00O100O5L5J3NO010O0010O01O03N4K5K7Je0ZOa0@<ChQQ4"}, "label": "bench to the left of other bench"}]}
{"id": 211641, "image": "COCO_train2014_000000211641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the whole bench that is to the left in the picture\"."}], "task": "refering", "answer_template": "The \"the whole bench that is to the left in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 285], "bbox": [0.12421875, 0.7344598337950138, 0.42590625000000004, 0.9659279778393352], "iscrowd": 0, "area": 9241.4213, "rle": {"size": [361, 640], "counts": "j^l06S;8H7I7I7I8H7I7I7I8H3M000000nNVGEj8;dGWO\\8i0SHhNm7X1R10000001O0000000000001O0000000000001O008H:F1O0000\\OmFROS9n0UGjNk8V1d000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000<Dg0YO2N00O100O5L5J3NO010O0010O01O03N4K5K7Je0ZOa0@<ChQQ4"}, "label": "the whole bench that is to the left in the picture"}]}
{"id": 6842, "image": "COCO_train2014_000000006842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat standing on the wash basin with its head up\"."}], "task": "refering", "answer_template": "The \"a cat standing on the wash basin with its head up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 135, 136, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248], "bbox": [0.528140625, 0.27743749999999995, 0.9548124999999998, 0.6428124999999999], "iscrowd": 0, "area": 18583.51204999999, "rle": {"size": [480, 640], "counts": "kSo41o>00001O001O000010O00eA0f=1WB3g=MVB9g=HVB=g=`0OO01O1O1O1O001N3M3M3N2M1O2O0O2O00001N3N3M4L2M101O00010O010O00010O010O00100O2N1O1N3N1O3M2N2N3M1O00001O0O1000000O1N201O000001O0000O100O2OO001O001O001O001O10O0100O100O100O10000O100O100O100O1001O0000001O0000001O000O101O0000001O0000001O0000002N2N2N2N2N0000000000O100000000O2O00000O2O0000001N100000001N1000001O0O100O2N1N2N2N3M2N2O1N3N1O1O0O2O1O001O0O2O001O0O2O1O001O0O2O001APCmNP=S1?O001O1O001O001O010O010O1O010O0010O00010O001O00001O000010O0001O00001O001O01O01O0L4nNS10O00100O10O01O10O01O1O001N2O001N4M2N3L4M2M4M2N3L4M2N3L3N3M3_Oc^="}, "label": "a cat standing on the wash basin with its head up"}]}
{"id": 6842, "image": "COCO_train2014_000000006842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and gray cat standing inside the right sink of a jack - and - jill style sink\"."}], "task": "refering", "answer_template": "The \"a black and gray cat standing inside the right sink of a jack - and - jill style sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 135, 136, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248], "bbox": [0.528140625, 0.27743749999999995, 0.9548124999999998, 0.6428124999999999], "iscrowd": 0, "area": 18583.51204999999, "rle": {"size": [480, 640], "counts": "kSo41o>00001O001O000010O00eA0f=1WB3g=MVB9g=HVB=g=`0OO01O1O1O1O001N3M3M3N2M1O2O0O2O00001N3N3M4L2M101O00010O010O00010O010O00100O2N1O1N3N1O3M2N2N3M1O00001O0O1000000O1N201O000001O0000O100O2OO001O001O001O001O10O0100O100O100O10000O100O100O100O1001O0000001O0000001O000O101O0000001O0000001O0000002N2N2N2N2N0000000000O100000000O2O00000O2O0000001N100000001N1000001O0O100O2N1N2N2N3M2N2O1N3N1O1O0O2O1O001O0O2O001O0O2O1O001O0O2O001APCmNP=S1?O001O1O001O001O010O010O1O010O0010O00010O001O00001O000010O0001O00001O001O01O01O0L4nNS10O00100O10O01O10O01O1O001N2O001N4M2N3L4M2M4M2N3L4M2N3L3N3M3_Oc^="}, "label": "a black and gray cat standing inside the right sink of a jack - and - jill style sink"}]}
{"id": 6842, "image": "COCO_train2014_000000006842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat in a sink\"."}], "task": "refering", "answer_template": "The \"a black cat in a sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 116, 117, 140, 141, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 258, 259], "bbox": [0.0824375, 0.2338333333333333, 0.412171875, 0.6583958333333333], "iscrowd": 0, "area": 17335.038399999998, "rle": {"size": [480, 640], "counts": "RPi0>o=e0G9N1N3M3N1N3M3N1N3M3N1N3M3N1N3M3N1N3M2O2M2N2O1N3MhMTD_1k;`N[D\\1c;dNcDX1[;iNiDU1U;jNQER1m:oNXEm0f:RO`Ek0_:UOfEh0W:XOmEf0Q:ZOQFf0ROiNU:`0kFg0dNVO^:3oF`1o8`NSGa1l8]NVGc1j8[NXGf1g8YNYGh1h8UNZGk1f8SN\\Gn1c8PN_GP2b8mM`GT2_8jMbGW2^8gMdGY2b:1O001N101O00001N101O001O0O2O0O1O2N1O2N1O2N1O2N2N2N2N1O2N2N0000000000000000001O0000000000000000000000000000000000000010O00001O001O00001O00001O00001O00001O0010O0001O00001O00001O00001N101N10001N100O2O000O2O0O101O0O2O000O2O0O101O0O101N10001N100O2O001N10001N100O2O001N2O1N2O1O1N2O1N101O1N2O1N2O1O1N2O1O1N2O1N2O1O4K4M4K5L3M4K5L3L5LRo_5"}, "label": "a black cat in a sink"}]}
{"id": 6842, "image": "COCO_train2014_000000006842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fluffy black cat sniffing around a bathroom sink\"."}], "task": "refering", "answer_template": "The \"a fluffy black cat sniffing around a bathroom sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 116, 117, 140, 141, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 258, 259], "bbox": [0.0824375, 0.2338333333333333, 0.412171875, 0.6583958333333333], "iscrowd": 0, "area": 17335.038399999998, "rle": {"size": [480, 640], "counts": "RPi0>o=e0G9N1N3M3N1N3M3N1N3M3N1N3M3N1N3M3N1N3M2O2M2N2O1N3MhMTD_1k;`N[D\\1c;dNcDX1[;iNiDU1U;jNQER1m:oNXEm0f:RO`Ek0_:UOfEh0W:XOmEf0Q:ZOQFf0ROiNU:`0kFg0dNVO^:3oF`1o8`NSGa1l8]NVGc1j8[NXGf1g8YNYGh1h8UNZGk1f8SN\\Gn1c8PN_GP2b8mM`GT2_8jMbGW2^8gMdGY2b:1O001N101O00001N101O001O0O2O0O1O2N1O2N1O2N1O2N2N2N2N1O2N2N0000000000000000001O0000000000000000000000000000000000000010O00001O001O00001O00001O00001O00001O0010O0001O00001O00001O00001N101N10001N100O2O000O2O0O101O0O2O000O2O0O101O0O101N10001N100O2O001N10001N100O2O001N2O1N2O1O1N2O1N101O1N2O1N2O1O1N2O1O1N2O1N2O1O4K4M4K5L3M4K5L3L5LRo_5"}, "label": "a fluffy black cat sniffing around a bathroom sink"}]}
{"id": 309946, "image": "COCO_train2014_000000309946.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man waxing surf board\"."}], "task": "refering", "answer_template": "The \"man waxing surf board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 293, 294, 315, 316, 317, 339, 340], "bbox": [0.4843125, 0.16393749999999999, 0.8979843750000001, 0.8679791666666666], "iscrowd": 0, "area": 33916.12320000001, "rle": {"size": [480, 640], "counts": "Qba47h>9H0O1O2O0O1O2O0O2N100O10O0100000000O1000O1000O100000000O010000O10O10O100O01000O1000O010000000O1000000SLEUI;j6JQI6n6NPI2o62nHNQ75mHKQ78nHHQ7;lHFS7<lHDS7?kHAT7a0kH_OS7e0kH[OT7g0jHZOU7i0iHWOV7k0iHTOW7m0iHSOV7o0hHROV7Q1iHoNV7S1iHmNW7S1iHmNV7U1iHkNW7V1gHjNY7X1fHhNZ7X1fHhNY7Z1fHfNZ7[1eHeNZ7]1dHdN\\7\\1dHdN[7^1dHbN\\7_1cHaN\\7a1bH_N_7a1`H`N_7b1_H_Na7b1]H_Nb7c1[H_Ne7a1ZH`Ne7b1YH_Ng7b1WH_Ni7b1UH_Nj7e1RH\\Nn7g1nGXNS8l1iGTNX8o1dGPN]8S2`GlMb8U2\\GkMd8X2YGgMi8Z2cFXN^9c3101N101N1101O2N1O2N101N1O2O0O2iG`Jj6a5SIbJl6`5oHcJR7]5kHfJT7\\5hHgJW7b6O100O00100O1O1O10O0100001O1N2O1N2O0RJoGk4S8QKPHo4Q8mJRHR5P8jJSHV5n7gJTHX5m7eJVH[5k7aJWH_5k7]JXHc5i7YJZHf5a800O100O100O010O1000O10O0100oG[Je6f5XI]Jg6d5WI^Ji6b5TI`Jl6a5RIaJn6^5PIeJo6\\5nHgJQ7Z5mHhJS7X5jHkJU7V5hHmJX7S5fHoJY7R5dHQK[7o4cHSK^7m4`HUK_7l4^HWKb7i4\\HYKc7h4ZH[Kf7P6O1O10O0100O1O100O100O1O010O100O100O1O100O010O1O100O100O100O02L3M4K5L4ZMbHcMc7V2fHdM^7W2hHdM\\7Z2hH_M\\7_2hH\\M\\7b2hHmLg7Q3V2M3N2N1N3N2N2M3N2N2M2O2N2M3N2N2M2O2M3N2N2M3N2N2M3N2N2M3M3N2M3M3N2M3N2M3M3N2M3M3M3N2M3M3N2M3M3M3NVVn0"}, "label": "man waxing surf board"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a water bottle\"."}], "task": "refering", "answer_template": "The \"a woman holding a water bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 298, 315, 316], "bbox": [0.51053125, 0.06745882352941177, 1.0, 0.8943529411764707], "iscrowd": 0, "area": 54352.14900000003, "rle": {"size": [425, 640], "counts": "PlW42U=2M3N2M3N2M3N2N3N1N2O1O10000O100O101O0O10000O100O10000O101N100O100O100O100O100O2O0O1O1O1O1N2O1O1^Ob0_Oa0D=G8K5L4K5O10000O10000O2eIULV3k3fLYLY3h3dL\\LZ3e3bL`L]3`3_LdL`3]3\\LhLa3Z3[LkLb3W3ZLmLd3U3YLoLe3R3WLTMg3l2ULZMh3h2SL^Mk3b2QLdMl3]2PLiMn3W2oKnMn3S2nKSNP4m1lKYNQ4h1kK^NS4b1iKcNU4^1gKhNW4X1fKmNX4S1dKROZ4P1aKVO]4j0_K[O`4e0\\KAb4d4O1O2N1O101N4L6J6J7I9H7H00000010O000010O0001O01O01O0001O01O0001O00010O000010O01O1O101N2N2O2M2O1N10000O1000O2N2N3N1N2N2N1O1N3N1N3M2O1N3M2N3L3POWL[Hm3d7_LmGf3Q8j0O2N2N2N2N2N2N2N2N3M2N3M2N3M1O001O1O2N1O2N2N2M4M1O1O001O001O001O1O001O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O002N1O2N2N1O2N2N1O2N2N1O2N2M2O2N2N2N2N2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N3M3M2N3M3M3M3M2N3M3L4M2N2N1O2O1N101N2O1OO1O1O1O1O2N1O1O1O2N100O2O00001N1000001N1iF"}, "label": "a woman holding a water bottle"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman talking to a man to her left\"."}], "task": "refering", "answer_template": "The \"a blonde woman talking to a man to her left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 298, 315, 316], "bbox": [0.51053125, 0.06745882352941177, 1.0, 0.8943529411764707], "iscrowd": 0, "area": 54352.14900000003, "rle": {"size": [425, 640], "counts": "PlW42U=2M3N2M3N2M3N2N3N1N2O1O10000O100O101O0O10000O100O10000O101N100O100O100O100O100O2O0O1O1O1O1N2O1O1^Ob0_Oa0D=G8K5L4K5O10000O10000O2eIULV3k3fLYLY3h3dL\\LZ3e3bL`L]3`3_LdL`3]3\\LhLa3Z3[LkLb3W3ZLmLd3U3YLoLe3R3WLTMg3l2ULZMh3h2SL^Mk3b2QLdMl3]2PLiMn3W2oKnMn3S2nKSNP4m1lKYNQ4h1kK^NS4b1iKcNU4^1gKhNW4X1fKmNX4S1dKROZ4P1aKVO]4j0_K[O`4e0\\KAb4d4O1O2N1O101N4L6J6J7I9H7H00000010O000010O0001O01O01O0001O01O0001O00010O000010O01O1O101N2N2O2M2O1N10000O1000O2N2N3N1N2N2N1O1N3N1N3M2O1N3M2N3L3POWL[Hm3d7_LmGf3Q8j0O2N2N2N2N2N2N2N2N3M2N3M2N3M1O001O1O2N1O2N2N2M4M1O1O001O001O001O1O001O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O002N1O2N2N1O2N2N1O2N2N1O2N2M2O2N2N2N2N2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N3M3M2N3M3M3M3M2N3M3L4M2N2N1O2O1N101N2O1OO1O1O1O1O2N1O1O1O2N100O2O00001N1000001N1iF"}, "label": "a blonde woman talking to a man to her left"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair holding a man\"."}], "task": "refering", "answer_template": "The \"a chair holding a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 207, 208, 230, 231, 232, 253, 254, 255, 276, 277, 300, 323, 324, 325, 326, 327], "bbox": [0.0005, 0.481364705882353, 0.252859375, 1.0], "iscrowd": 0, "area": 7968.333599999998, "rle": {"size": [425, 640], "counts": "h71Q=;I7L4L5K4M3L4L4L4L4L4L5J5K5K5J6K5K5K5TOTMoFT3k8k0M20000O100O100O1O2O0O100O100O10001N10000N2L3NO1O001O1O001O1O001O1O0oMlGmNU8f0ZHWOf7=jH_OV7=QI@o6?SI@m6?UI@k6`0VI_Oj6`0YI^Oh6`0ZI_Of6`0\\I_Od6`0_I^Oa6a0bI]O^6b0eI\\O[6d0gIZOZ6d0iIZOW6e0lIYOT6f0oIXOQ6g0RJWOn5h0UJVOk5j0UJVOk5i0VJWOk5g0VJYOk5e0VJ[Oj5d0WJ\\Oj5b0WJ^Oj5`0WJ@j5>WJBj5:YJFh55\\JKe50_J0]9000000000000000000000000000000000O1000000O1000000O1000000O1000000O10000O100O100O10O0010O01O00100O0010O01O0010O01O010O0010O01O0010O01O10O01O1O100O1O100OY]V6"}, "label": "a chair holding a man"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair man is sitting in\"."}], "task": "refering", "answer_template": "The \"chair man is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 207, 208, 230, 231, 232, 253, 254, 255, 276, 277, 300, 323, 324, 325, 326, 327], "bbox": [0.0005, 0.481364705882353, 0.252859375, 1.0], "iscrowd": 0, "area": 7968.333599999998, "rle": {"size": [425, 640], "counts": "h71Q=;I7L4L5K4M3L4L4L4L4L4L5J5K5K5J6K5K5K5TOTMoFT3k8k0M20000O100O100O1O2O0O100O100O10001N10000N2L3NO1O001O1O001O1O001O1O0oMlGmNU8f0ZHWOf7=jH_OV7=QI@o6?SI@m6?UI@k6`0VI_Oj6`0YI^Oh6`0ZI_Of6`0\\I_Od6`0_I^Oa6a0bI]O^6b0eI\\O[6d0gIZOZ6d0iIZOW6e0lIYOT6f0oIXOQ6g0RJWOn5h0UJVOk5j0UJVOk5i0VJWOk5g0VJYOk5e0VJ[Oj5d0WJ\\Oj5b0WJ^Oj5`0WJ@j5>WJBj5:YJFh55\\JKe50_J0]9000000000000000000000000000000000O1000000O1000000O1000000O1000000O10000O100O100O10O0010O01O00100O0010O01O0010O01O010O0010O01O0010O01O10O01O1O100O1O100OY]V6"}, "label": "chair man is sitting in"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt speaking to a woman\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt speaking to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 329, 330, 331], "bbox": [0.03896875, 0.010729411764705881, 0.451203125, 0.9836235294117648], "iscrowd": 0, "area": 66619.74290000001, "rle": {"size": [425, 640], "counts": "_d:5o<9^Oc0D;I4L3N2N2N2N2O1N2N1O2N2N2N2O1N2N2N2N2N2O2M3M2N3M2N3M2O2M3M2M4L3M4L3M4L3M4L4L3M4L3M4L3N3N2M2N3M2N3M2_HfKQ6]4kIoKj5T4RJXLc5j3ZJbL[5P6F3L2N2N3M2N2N3M2N2N3N10001N10001O0YM[Gi0f8VOiG<X8BoG8Q8HQHHQOSOn8T1THEQOUOm8U1UHAQOYOj8U1XH]OROD_8o0fHTOnNM[8n0oHlNiN4Y8o0kIPOU6P1lInNV6P1lIoNT6Q1mInNS6Q1nInNS6R1nImNR6R1PJmNP6S1PJlNQ6S1QJlNo5T1T3O000O10000O01000O1000O010000O10O100000000O1000000O100000000O100000000O100000000O100000000O10001O0O1bLfNmJZ1R5iNlJW1T5kNkJT1U5oNhJQ1X5QOfJo0Y5SOfJm0Z5SOfJm0Z5SOfJm0Y5SOhJm0V5UOjJl0U5TOkJl0U5TOkJm0T5SOlJm0S5TOmJm0R5SOnJm0R5SOnJn0Q5ROoJo0P5QOPKo0o4ROQKo0n4QORKo0n4QORKP1m4POSKP1m4POSKQ1k4POUKP1k4POVKP1i4POWKP1i4POWKQ1h4oNXKQ1g4POYKQ1f4oNZKQ1f4oNZKR1e4nN[KR1e4nN[KS1c4nN]KR1c4mN^KT1a4lN_KT1a4lN_KU1`4kN`KV1^4kNbKU1^4kNcKU1^OQO^1JXOQ1YOVO_1H]Oo0QO[Ob1FAk0lN@b1FFg0eNEe1DJc0_NLf1AMb0[NNh1@Na0XN1j1^ONT1_MEb2VO1V1lL3S3gN1P4OPL2P4LQL4o3LQL5n3KRL5o3JQL7o3HPL9Q4FoK:Q4FoK:R4EnK;S4CnK=S4BmK>T4AkK`0V4_OjKa0W4^OiKb0X4]OhKc0Y4\\OgKd0Z4ZOgKf0Z4YOfKg0[4XOdKi0]4VOcKj0^4UObKk0_4TOaKl0a4RO_Kn0c4PO]KP1e4mN\\KS1f4kNYKV1i4hNWKX1l4eNTK[1n4cNRK]1P5aNPK_1R5_NnJa1T5[NnJd1U5ULWJ\\1n0_2o4RLTJ\\1o0a2P5PLSJ[1P1d2P5oKQJW1T1j2m4lKRJn0[1U3o4aL[K]3g4]L^Kb3d4XLaKg3a4SLeKj3n6O1O2N1O1O2N1O2M2O1O2N1O2N1N3N1O2M2O1O1N2N2K5K5L4L4N2oMVEa1l:XNaE_1];G8Hf0ZOS[a4"}, "label": "a man in a black shirt speaking to a woman"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt and sitting with a mac laptop on his lap\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt and sitting with a mac laptop on his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 26, 27, 28, 29, 30, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 329, 330, 331], "bbox": [0.03896875, 0.010729411764705881, 0.451203125, 0.9836235294117648], "iscrowd": 0, "area": 66619.74290000001, "rle": {"size": [425, 640], "counts": "_d:5o<9^Oc0D;I4L3N2N2N2N2O1N2N1O2N2N2N2O1N2N2N2N2N2O2M3M2N3M2N3M2O2M3M2M4L3M4L3M4L3M4L4L3M4L3M4L3N3N2M2N3M2N3M2_HfKQ6]4kIoKj5T4RJXLc5j3ZJbL[5P6F3L2N2N3M2N2N3M2N2N3N10001N10001O0YM[Gi0f8VOiG<X8BoG8Q8HQHHQOSOn8T1THEQOUOm8U1UHAQOYOj8U1XH]OROD_8o0fHTOnNM[8n0oHlNiN4Y8o0kIPOU6P1lInNV6P1lIoNT6Q1mInNS6Q1nInNS6R1nImNR6R1PJmNP6S1PJlNQ6S1QJlNo5T1T3O000O10000O01000O1000O010000O10O100000000O1000000O100000000O100000000O100000000O100000000O10001O0O1bLfNmJZ1R5iNlJW1T5kNkJT1U5oNhJQ1X5QOfJo0Y5SOfJm0Z5SOfJm0Z5SOfJm0Y5SOhJm0V5UOjJl0U5TOkJl0U5TOkJm0T5SOlJm0S5TOmJm0R5SOnJm0R5SOnJn0Q5ROoJo0P5QOPKo0o4ROQKo0n4QORKo0n4QORKP1m4POSKP1m4POSKQ1k4POUKP1k4POVKP1i4POWKP1i4POWKQ1h4oNXKQ1g4POYKQ1f4oNZKQ1f4oNZKR1e4nN[KR1e4nN[KS1c4nN]KR1c4mN^KT1a4lN_KT1a4lN_KU1`4kN`KV1^4kNbKU1^4kNcKU1^OQO^1JXOQ1YOVO_1H]Oo0QO[Ob1FAk0lN@b1FFg0eNEe1DJc0_NLf1AMb0[NNh1@Na0XN1j1^ONT1_MEb2VO1V1lL3S3gN1P4OPL2P4LQL4o3LQL5n3KRL5o3JQL7o3HPL9Q4FoK:Q4FoK:R4EnK;S4CnK=S4BmK>T4AkK`0V4_OjKa0W4^OiKb0X4]OhKc0Y4\\OgKd0Z4ZOgKf0Z4YOfKg0[4XOdKi0]4VOcKj0^4UObKk0_4TOaKl0a4RO_Kn0c4PO]KP1e4mN\\KS1f4kNYKV1i4hNWKX1l4eNTK[1n4cNRK]1P5aNPK_1R5_NnJa1T5[NnJd1U5ULWJ\\1n0_2o4RLTJ\\1o0a2P5PLSJ[1P1d2P5oKQJW1T1j2m4lKRJn0[1U3o4aL[K]3g4]L^Kb3d4XLaKg3a4SLeKj3n6O1O2N1O1O2N1O2M2O1O2N1O2N1N3N1O2M2O1O1N2N2K5K5L4L4N2oMVEa1l:XNaE_1];G8Hf0ZOS[a4"}, "label": "a man wearing a black shirt and sitting with a mac laptop on his lap"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person walking behind seating\"."}], "task": "refering", "answer_template": "The \"a person walking behind seating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 35, 36, 39, 40, 41, 42, 43, 63, 64, 65, 66, 67, 87, 89, 90, 113, 114, 136, 137], "bbox": [0.5387187499999999, 0.0022588235294117645, 0.99684375, 0.3887529411764706], "iscrowd": 0, "area": 12818.1986, "rle": {"size": [425, 640], "counts": "SW_43T=4K5K5L3L5L3L5L4K4L5M3N1O2N1N10O100O10000O100O10000O100O1L4L4L4L4L4L4N200O1000000O100000000O100000O10O1000000O100000000O100000000O1000000O100000000O100000000O1000000O10O100000O100000000O1001O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1N2O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1OcMRF`1m9bNVF[1h9fN\\FV1c9lN`FP1_9QOeFk0Z9WOiFe0V9]OmF9X9HT22N1O2N1N2O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1001O2N1O1O1O2N1O1O2N1O1O1O2O00000O01O10O0100O1O010O10O0100O10O10O10O010000O01000O010O1000O01000O010000O01002M01N2O0O2O1N101N2N101N101N2O0O2O1N101N2N101N2O0O2O1N101N2O3L4L4Mee0"}, "label": "a person walking behind seating"}]}
{"id": 178874, "image": "COCO_train2014_000000178874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a grey shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 35, 36, 39, 40, 41, 42, 43, 63, 64, 65, 66, 67, 87, 89, 90, 113, 114, 136, 137], "bbox": [0.5387187499999999, 0.0022588235294117645, 0.99684375, 0.3887529411764706], "iscrowd": 0, "area": 12818.1986, "rle": {"size": [425, 640], "counts": "SW_43T=4K5K5L3L5L3L5L4K4L5M3N1O2N1N10O100O10000O100O10000O100O1L4L4L4L4L4L4N200O1000000O100000000O100000O10O1000000O100000000O100000000O1000000O100000000O100000000O1000000O10O100000O100000000O1001O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2M2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1N2O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1OcMRF`1m9bNVF[1h9fN\\FV1c9lN`FP1_9QOeFk0Z9WOiFe0V9]OmF9X9HT22N1O2N1N2O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1001O2N1O1O1O2N1O1O2N1O1O1O2O00000O01O10O0100O1O010O10O0100O10O10O10O010000O01000O010O1000O01000O010000O01002M01N2O0O2O1N101N2N101N101N2O0O2O1N101N2N101N2O0O2O1N101N2O3L4L4Mee0"}, "label": "a man wearing a grey shirt"}]}
{"id": 137918, "image": "COCO_train2014_000000137918.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white laptop sitting on the coffee table\"."}], "task": "refering", "answer_template": "The \"white laptop sitting on the coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 228, 248, 249, 250, 251], "bbox": [0.743171875, 0.27291666666666664, 0.9463125, 0.6264375], "iscrowd": 0, "area": 13780.564750000001, "rle": {"size": [480, 640], "counts": "dZo64i>5K5J6K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5K5K5J6K5K5K5N2N2O1N2N2O1N2N2N2O1N2N2N2O1N2NN201O0000001O00001O0000001O0000010O0000001O0000001O00001O0000001O0001O01O0000001O0000001O00001gMhEHX:5lEJT:2RFLn90XFNi9N\\F1c9KcF3]9JhF4X9HmF7T9EQG9o8CWG;i8B\\G<d8@bG>_8^OeGa0[8[OkGc0U8ZOPHd0P8XOVHf0k7VOZHh0f7TO`Hj0`7ROeHm0\\7oNiHo0W7mNoHQ1b:N2O0O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2O0Odd?"}, "label": "white laptop sitting on the coffee table"}]}
{"id": 137918, "image": "COCO_train2014_000000137918.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white laptop\"."}], "task": "refering", "answer_template": "The \"a white laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 228, 248, 249, 250, 251], "bbox": [0.743171875, 0.27291666666666664, 0.9463125, 0.6264375], "iscrowd": 0, "area": 13780.564750000001, "rle": {"size": [480, 640], "counts": "dZo64i>5K5J6K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5K5K5J6K5K5K5N2N2O1N2N2O1N2N2N2O1N2N2N2O1N2NN201O0000001O00001O0000001O0000010O0000001O0000001O00001O0000001O0001O01O0000001O0000001O00001gMhEHX:5lEJT:2RFLn90XFNi9N\\F1c9KcF3]9JhF4X9HmF7T9EQG9o8CWG;i8B\\G<d8@bG>_8^OeGa0[8[OkGc0U8ZOPHd0P8XOVHf0k7VOZHh0f7TO`Hj0`7ROeHm0\\7oNiHo0W7mNoHQ1b:N2O0O2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2O0Odd?"}, "label": "a white laptop"}]}
{"id": 137918, "image": "COCO_train2014_000000137918.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer currently on netgear ' s website\"."}], "task": "refering", "answer_template": "The \"a computer currently on netgear ' s website\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.167609375, 0.0784375, 0.514671875, 0.6653541666666667], "iscrowd": 0, "area": 48484.9296, "rle": {"size": [480, 640], "counts": "lZb14g>?B?@?A?B>A?A?B>A?A`0A>A<D5L4K5K5L4K5K5L4K5K5L4K5K5L4K5L5J5K5L4K5O10000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O000000010O0000001O000000001O000000001O0000001O000000010O0000001O000000001O0000000ZLlH@U7<RI@n6=YI_Og6>`I^O`6?gI]OZ6`0lI\\OT6a0SJ[Om5b0ZJZOf5c0aJYO_5d0hJYOX5c0oJYOQ5e0UKWOk4f0\\KVOd4g0cKUO]4h0jKTOW4h0PLTOP4j0VLROj3k0]LQOc3l0dLPO]3l0jLPOV3m0RMnNn2o0YMmNg2Q1_MkNa2R1fMjN[2R1lMjNT2S1SNiNm1T1ZNhNf1V1`NfNa1V1fNfNZ1W1Y6M3M3M4M2M3M3M3M4M2M3M3M4Ldfa4"}, "label": "a computer currently on netgear ' s website"}]}
{"id": 137918, "image": "COCO_train2014_000000137918.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver - colored computer\"."}], "task": "refering", "answer_template": "The \"a silver - colored computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.167609375, 0.0784375, 0.514671875, 0.6653541666666667], "iscrowd": 0, "area": 48484.9296, "rle": {"size": [480, 640], "counts": "lZb14g>?B?@?A?B>A?A?B>A?A`0A>A<D5L4K5K5L4K5K5L4K5K5L4K5K5L4K5L5J5K5L4K5O10000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O0000000010O000001O000000001O000000001O0000001O000000010O0000001O000000001O000000001O0000001O000000010O0000001O000000001O0000000ZLlH@U7<RI@n6=YI_Og6>`I^O`6?gI]OZ6`0lI\\OT6a0SJ[Om5b0ZJZOf5c0aJYO_5d0hJYOX5c0oJYOQ5e0UKWOk4f0\\KVOd4g0cKUO]4h0jKTOW4h0PLTOP4j0VLROj3k0]LQOc3l0dLPO]3l0jLPOV3m0RMnNn2o0YMmNg2Q1_MkNa2R1fMjN[2R1lMjNT2S1SNiNm1T1ZNhNf1V1`NfNa1V1fNfNZ1W1Y6M3M3M4M2M3M3M3M4M2M3M3M4Ldfa4"}, "label": "a silver - colored computer"}]}
{"id": 137918, "image": "COCO_train2014_000000137918.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small black laptop in between two other laptops on a coffee table\"."}], "task": "refering", "answer_template": "The \"a small black laptop in between two other laptops on a coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 244, 245, 246], "bbox": [0.537640625, 0.2584166666666667, 0.7432656249999999, 0.6314583333333333], "iscrowd": 0, "area": 15357.7094, "rle": {"size": [480, 640], "counts": "\\^Q5j0U>S1mN2N1N2O2N1O1N3N1O2N1N2O2N1O1ZO_MkDd2Q;dMiD]2U;h0M4L3N3L3M3N3L3M3M4M2M3M4M2010O0010O01O01O010O01O01O01O010O0010O0010O0010O01O01O01O010O01O01O010O0010O00010O010O01O01O01O010O00010O010O0010O00010O010O01O01O01O010O00010O010O0010OmMkF[NU9b1W2L3N3L3M4L3M3M4L3M4L3M3M4M2M4L3M4L3M3M4L3M4LmU]2"}, "label": "a small black laptop in between two other laptops on a coffee table"}]}
{"id": 137918, "image": "COCO_train2014_000000137918.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small black laptop between two big ones\"."}], "task": "refering", "answer_template": "The \"a small black laptop between two big ones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 244, 245, 246], "bbox": [0.537640625, 0.2584166666666667, 0.7432656249999999, 0.6314583333333333], "iscrowd": 0, "area": 15357.7094, "rle": {"size": [480, 640], "counts": "\\^Q5j0U>S1mN2N1N2O2N1O1N3N1O2N1N2O2N1O1ZO_MkDd2Q;dMiD]2U;h0M4L3N3L3M3N3L3M3M4M2M3M4M2010O0010O01O01O010O01O01O01O010O0010O0010O0010O01O01O01O010O01O01O010O0010O00010O010O01O01O01O010O00010O010O0010O00010O010O01O01O01O010O00010O010O0010OmMkF[NU9b1W2L3N3L3M4L3M3M4L3M4L3M3M4M2M4L3M4L3M3M4L3M4LmU]2"}, "label": "a small black laptop between two big ones"}]}
{"id": 96958, "image": "COCO_train2014_000000096958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of banana on the left side of a dessert\"."}], "task": "refering", "answer_template": "The \"a slice of banana on the left side of a dessert\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 143, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 311, 330, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 378, 379, 380], "bbox": [0.24934375000000003, 0.27533049040511726, 0.5824374999999999, 0.9584434968017057], "iscrowd": 0, "area": 29271.550249999993, "rle": {"size": [469, 640], "counts": "Q_Y29U><E:F:E;F;E:E;H8K6J5K5L4K6J5L4K5K6K4K5K5L3L3M4L4M2M4L3N3M3M2N3M3M2N3M3M2N3M3M2N3M3L3N3M3M2N3N2O010OO1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N2N3M2N2N2O1N2N2N2N2N2N2N1O0000000000O1000000000000O1000000000000O100000O1O1O1O001O1O1O1O1O100O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2O0O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N100O2M2N2O2M2N2O2M2N2N3N1N2N3N1N3M2O1N3M2N2O2M5K5L4K5K5L5J5Kkki3"}, "label": "a slice of banana on the left side of a dessert"}]}
{"id": 96958, "image": "COCO_train2014_000000096958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate with dried up banana on it\"."}], "task": "refering", "answer_template": "The \"a plate with dried up banana on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 143, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 311, 330, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 378, 379, 380], "bbox": [0.24934375000000003, 0.27533049040511726, 0.5824374999999999, 0.9584434968017057], "iscrowd": 0, "area": 29271.550249999993, "rle": {"size": [469, 640], "counts": "Q_Y29U><E:F:E;F;E:E;H8K6J5K5L4K6J5L4K5K6K4K5K5L3L3M4L4M2M4L3N3M3M2N3M3M2N3M3M2N3M3M2N3M3L3N3M3M2N3N2O010OO1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2O0O1O1O1O1O2N1O1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N2N3M2N2N2O1N2N2N2N2N2N2N1O0000000000O1000000000000O1000000000000O100000O1O1O1O001O1O1O1O1O100O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2O0O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N100O2M2N2O2M2N2O2M2N2N3N1N2N3N1N3M2O1N3M2N2O2M5K5L4K5K5L5J5Kkki3"}, "label": "a plate with dried up banana on it"}]}
{"id": 96958, "image": "COCO_train2014_000000096958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wedge of banana to the right of the dessert\"."}], "task": "refering", "answer_template": "The \"the wedge of banana to the right of the dessert\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 108, 109, 110, 132, 133, 134, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 251, 272, 273], "bbox": [0.627234375, 0.17302771855010662, 0.92484375, 0.6854157782515993], "iscrowd": 0, "area": 14320.189750000007, "rle": {"size": [469, 640], "counts": "cVh51c>1N2O1O1O001O100001O0000000001O001O10O01O1O1O00001O0010O00100O1O10O01O1N2O1N2O1O1O10O1O2M2O1O1N3N10O001O10O100O10000O10000O100O10000O100O10000O0100O0010O01O1O01YBBV=>fBFY=:eBH\\=8aBK_=5^BNa=d009F6K001N7[DZNn9b2UE`Mi:c2TE_Mk:X3M2O1O1N2O2M2O1N2O2N2M8I7I6J0O2O00001O0O2O00001O0O101O00001N10001O0O10000000O010000000O100000000O1000001N100O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O2M2O1N3N2M4M2N2M3N3L3N2L4K6K4K5K5K8H7J7H7I8C<C>Cnge0"}, "label": "the wedge of banana to the right of the dessert"}]}
{"id": 96958, "image": "COCO_train2014_000000096958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow piece of food onthe right\"."}], "task": "refering", "answer_template": "The \"yellow piece of food onthe right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 108, 109, 110, 132, 133, 134, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 251, 272, 273], "bbox": [0.627234375, 0.17302771855010662, 0.92484375, 0.6854157782515993], "iscrowd": 0, "area": 14320.189750000007, "rle": {"size": [469, 640], "counts": "cVh51c>1N2O1O1O001O100001O0000000001O001O10O01O1O1O00001O0010O00100O1O10O01O1N2O1N2O1O1O10O1O2M2O1O1N3N10O001O10O100O10000O10000O100O10000O100O10000O0100O0010O01O1O01YBBV=>fBFY=:eBH\\=8aBK_=5^BNa=d009F6K001N7[DZNn9b2UE`Mi:c2TE_Mk:X3M2O1O1N2O2M2O1N2O2N2M8I7I6J0O2O00001O0O2O00001O0O101O00001N10001O0O10000000O010000000O100000000O1000001N100O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O2M2O1N3N2M4M2N2M3N3L3N2L4K6K4K5K5K8H7J7H7I8C<C>Cnge0"}, "label": "yellow piece of food onthe right"}]}
{"id": 244418, "image": "COCO_train2014_000000244418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe facing the camera\"."}], "task": "refering", "answer_template": "The \"the giraffe facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 19, 20, 31, 32, 33, 44, 45, 46, 58, 71, 83, 84, 85, 96, 97, 98, 109, 110, 111, 122, 123, 124, 135, 136, 137, 148, 149, 150, 161, 162, 163, 174, 175, 176, 187, 188, 189, 200, 201, 202, 213, 214, 215, 227, 228], "bbox": [0.38565333333333335, 0.0426, 0.6158399999999999, 0.97758], "iscrowd": 0, "area": 21739.03465, "rle": {"size": [500, 375], "counts": "_QW28V?:F9H8G9G9G9N2K5K5K5L4Ke0iGnLd2m3XL^Lf3S4`KXL^4Y4dJVLZ5S7N3M2N2N2lK\\GV1f8hN`GS1a8lNdGPNJP2d8NbH0`7NcHO`7OaHOa7OaHOa70aHMa7OcHO_7LfHNT5YNnJe12IP5ROgJP1?Dk4K`J;j0Be4c0ZJGU1]Ob4[1RJZO[1QOm4_4QKYKW5h4hJoJb5R5_2:MWMnEB[:HRF8W:ROVFn0S:\\NZFc1l;01I6J6N200O105J9H8H04L5K4L5`C7j8NPG\\1h7hNQH\\1n7gNlG]1S8cNkG_1U8`NiGd1V8[NhGi1W8WNgGl1k3]LgL7P1_1\\NP2h3`LgL>l0Q1cNU2d3cLfLe0j0c0jNX2`3fLgLm0e04TO[2Z3jLgLU1b0E\\O^2V3nLgL\\1>XOD`2Q3SMgLc1:iNMd2l2VMgLk17[N5f2g2n2XMTMe2n2[MTMa2o2]MTM_2n2aMTM[2n2dMTMZ2n2eMTMW2n2hMUMT2o2jMSMR2Q3kMVMm1n2QNXMg1l2VNZMc1b8^Ob0YOYSV2"}, "label": "the giraffe facing the camera"}]}
{"id": 244418, "image": "COCO_train2014_000000244418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe that is standing on the path way\"."}], "task": "refering", "answer_template": "The \"the giraffe that is standing on the path way\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 19, 20, 31, 32, 33, 44, 45, 46, 58, 71, 83, 84, 85, 96, 97, 98, 109, 110, 111, 122, 123, 124, 135, 136, 137, 148, 149, 150, 161, 162, 163, 174, 175, 176, 187, 188, 189, 200, 201, 202, 213, 214, 215, 227, 228], "bbox": [0.38565333333333335, 0.0426, 0.6158399999999999, 0.97758], "iscrowd": 0, "area": 21739.03465, "rle": {"size": [500, 375], "counts": "_QW28V?:F9H8G9G9G9N2K5K5K5L4Ke0iGnLd2m3XL^Lf3S4`KXL^4Y4dJVLZ5S7N3M2N2N2lK\\GV1f8hN`GS1a8lNdGPNJP2d8NbH0`7NcHO`7OaHOa7OaHOa70aHMa7OcHO_7LfHNT5YNnJe12IP5ROgJP1?Dk4K`J;j0Be4c0ZJGU1]Ob4[1RJZO[1QOm4_4QKYKW5h4hJoJb5R5_2:MWMnEB[:HRF8W:ROVFn0S:\\NZFc1l;01I6J6N200O105J9H8H04L5K4L5`C7j8NPG\\1h7hNQH\\1n7gNlG]1S8cNkG_1U8`NiGd1V8[NhGi1W8WNgGl1k3]LgL7P1_1\\NP2h3`LgL>l0Q1cNU2d3cLfLe0j0c0jNX2`3fLgLm0e04TO[2Z3jLgLU1b0E\\O^2V3nLgL\\1>XOD`2Q3SMgLc1:iNMd2l2VMgLk17[N5f2g2n2XMTMe2n2[MTMa2o2]MTM_2n2aMTM[2n2dMTMZ2n2eMTMW2n2hMUMT2o2jMSMR2Q3kMVMm1n2QNXMg1l2VNZMc1b8^Ob0YOYSV2"}, "label": "the giraffe that is standing on the path way"}]}
{"id": 203459, "image": "COCO_train2014_000000203459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra showing us its butt\"."}], "task": "refering", "answer_template": "The \"the zebra showing us its butt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 194, 195, 196, 217, 218, 219, 239, 240, 241, 242, 263, 264, 265, 286, 287, 309, 310, 311, 333], "bbox": [0.41815625, 0.42833333333333334, 0.554703125, 0.8423958333333333], "iscrowd": 0, "area": 8238.4785, "rle": {"size": [480, 640], "counts": "cmm31n>100O1H0`A1^>3^ANb>800M4L3M3M3M4L3M4L4L5K4L4N3M4L4L4L5K4M4K4L5SE]MT9g2aFaM^9f2TFbMl9h2bE`M]:b3O1O1O100O010ON2K6H7I8H7I8ROm0O1O2N1O21mNeM]E^2b:eMYE^2f:dMWE^24`Mf94QF_23oM`9DXFc2N]NKXOi8I[GQ5f8nJYGS5g8nJVGT5k8lJRGV5o8=001O001O1VNZGXMf8e2mG_LUOGn8h3QH^LUOGj8i3VH\\LX8a3nG[LS8c3QHZLQ8b3g1N2N2N1O2N2O1WM\\DU2f;jM]DS2d;mM^DP1L^Oh;@_DP1M[Of;D_Do0OXOe;E`D33OM4b;H_E2SO1a;I`E3TOOV=Oo0NhZU4"}, "label": "the zebra showing us its butt"}]}
{"id": 203459, "image": "COCO_train2014_000000203459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in the middle\"."}], "task": "refering", "answer_template": "The \"zebra in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 194, 195, 196, 217, 218, 219, 239, 240, 241, 242, 263, 264, 265, 286, 287, 309, 310, 311, 333], "bbox": [0.41815625, 0.42833333333333334, 0.554703125, 0.8423958333333333], "iscrowd": 0, "area": 8238.4785, "rle": {"size": [480, 640], "counts": "cmm31n>100O1H0`A1^>3^ANb>800M4L3M3M3M4L3M4L4L5K4L4N3M4L4L4L5K4M4K4L5SE]MT9g2aFaM^9f2TFbMl9h2bE`M]:b3O1O1O100O010ON2K6H7I8H7I8ROm0O1O2N1O21mNeM]E^2b:eMYE^2f:dMWE^24`Mf94QF_23oM`9DXFc2N]NKXOi8I[GQ5f8nJYGS5g8nJVGT5k8lJRGV5o8=001O001O1VNZGXMf8e2mG_LUOGn8h3QH^LUOGj8i3VH\\LX8a3nG[LS8c3QHZLQ8b3g1N2N2N1O2N2O1WM\\DU2f;jM]DS2d;mM^DP1L^Oh;@_DP1M[Of;D_Do0OXOe;E`D33OM4b;H_E2SO1a;I`E3TOOV=Oo0NhZU4"}, "label": "zebra in the middle"}]}
{"id": 522947, "image": "COCO_train2014_000000522947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a cake that is not wearing a tie\"."}], "task": "refering", "answer_template": "The \"a man holding a cake that is not wearing a tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 109, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.49210937499999996, 0.08276995305164318, 0.928890625, 0.9996244131455398], "iscrowd": 0, "area": 66890.7373, "rle": {"size": [426, 640], "counts": "V]S41Y=0O1O2N1O1O1O1O2O0O1O1O1O101N1O1O1O100O2N1O100O1O2N1O100O2N1O2N1O1O2N100O2N1O2N1O1O2N1O1O2N1N3N1N2O2M2O1N2VHVNS4k1lK\\Nl3g1QL`Nh3b1WLeNb3\\1]LkN[3W1cLQOV3P1iLSOS3P1jLTOS3m0lLWOo2k0PMXOm2i0QM[Oi2h0WM\\Ob2g0^M_OX2g0eM_OR2j0iM[On1m0nMVOk1R1PNROi1W1RNkNh1]1SNfNi1m1fMVNT2o1jMSNR2P2lMSNP2P2oMSNl1P2SNRNi1Q2UNRNg1Q2XNRNc1Q2[NRNa1Q2^NQN^1R2aNPNZ1T2fNnMU1U2kNlMQ1W2oNjMm0Y2SOiMh0Z2WOhMe0[2[OfMa0]2_OeM<^2DcM8`2HaM4b2L_MOe21[MKi24YMGk29VMBm2?TM[OQ3e0PMVOT3j0mLQOW3o0b400O100001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O00;Ef0ZOR1nN001O1O1O001O1O1O1O001N2O1O1O001O1O1O1O001O1O000000000000O10000000000O10000000000O11O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O1O1O2N1O1O1O1N3N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O4L5K4L5K5K5K5K5K5K5K5bNWIQLn6h2UJPMcRc0"}, "label": "a man holding a cake that is not wearing a tie"}]}
{"id": 522947, "image": "COCO_train2014_000000522947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black suit coat and a white shirt without a tie holding the corner of a birthday cake\"."}], "task": "refering", "answer_template": "The \"a man in a black suit coat and a white shirt without a tie holding the corner of a birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 109, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.49210937499999996, 0.08276995305164318, 0.928890625, 0.9996244131455398], "iscrowd": 0, "area": 66890.7373, "rle": {"size": [426, 640], "counts": "V]S41Y=0O1O2N1O1O1O1O2O0O1O1O1O101N1O1O1O100O2N1O100O1O2N1O100O2N1O2N1O1O2N100O2N1O2N1O1O2N1O1O2N1N3N1N2O2M2O1N2VHVNS4k1lK\\Nl3g1QL`Nh3b1WLeNb3\\1]LkN[3W1cLQOV3P1iLSOS3P1jLTOS3m0lLWOo2k0PMXOm2i0QM[Oi2h0WM\\Ob2g0^M_OX2g0eM_OR2j0iM[On1m0nMVOk1R1PNROi1W1RNkNh1]1SNfNi1m1fMVNT2o1jMSNR2P2lMSNP2P2oMSNl1P2SNRNi1Q2UNRNg1Q2XNRNc1Q2[NRNa1Q2^NQN^1R2aNPNZ1T2fNnMU1U2kNlMQ1W2oNjMm0Y2SOiMh0Z2WOhMe0[2[OfMa0]2_OeM<^2DcM8`2HaM4b2L_MOe21[MKi24YMGk29VMBm2?TM[OQ3e0PMVOT3j0mLQOW3o0b400O100001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O00;Ef0ZOR1nN001O1O1O001O1O1O1O001N2O1O1O001O1O1O1O001O1O000000000000O10000000000O10000000000O11O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O1O1O2N1O1O1O1N3N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O4L5K4L5K5K5K5K5K5K5K5bNWIQLn6h2UJPMcRc0"}, "label": "a man in a black suit coat and a white shirt without a tie holding the corner of a birthday cake"}]}
{"id": 522947, "image": "COCO_train2014_000000522947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"john kerry looking guy in the red tie\"."}], "task": "refering", "answer_template": "The \"john kerry looking guy in the red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 50, 51, 52, 73, 74, 75, 96, 97, 98, 118, 119, 120, 121, 122, 123, 129, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 325, 326], "bbox": [0.10870312499999998, 0.08075117370892018, 0.65371875, 0.9733802816901408], "iscrowd": 0, "area": 56403.95199999998, "rle": {"size": [426, 640], "counts": "V[m0b2g:d0PNRM\\HV3]7RM[HU3_7RMYHU3`7oL[HY3^7jL_H]3[7fL`Hb3Y7aLdHf3V7]1I7I7J6I7I7J1N1O010O0010O0100O1O100O00100O1O100O1O1O1O100O1O1O00100O1O001O10O_MZH_Og7>gHVOZ7f0PISOP7k0UISOk6j0[IROf6j0`ISO`6k0eIRO\\6k0iIROX6l0lIROT6m0oIPOQ6Q1PJmNQ6R1RJkNo5T1SJjNm5W1UJgNk5X1WJfNj5Y1YJbNi5_1ZJ[Ni5d1o2O010O1O10O01O10O1000O01000O10O1000O0100000O010000O10001O1N101O1N101O1O0O2O1O0O2O1O001N2O001N2O1O001oG\\M[5d2_JbM`5^2[JiMc5X2WJoMg5R2SJUNl5k1nI\\NP6e1jI`NV6a1dIdN\\6]1nHXOS7g0hH_OW7b0bHD^7T3001O0000010O00O100000001N100000001O0SO[JTJf5j5\\JVJd5g5_JYJa5g5_JYJb5e5`JZJ`5e5aJZJ`5f5`JZJ`5e5bJZJ^5e5cJ[J^5d5bJ[J_5d5cJ[J]5e5cJ[J]5d5dJ\\J]5b5dJ]J]5c5dJ\\J\\5c5eJ]J[5c5eJ]J\\5a5fJ]J[5c5eJ]J[5b5fJ^J[5a5fJ^JZ5a5gJ^JZ5a5gJ_JZ5`5fJ`JZ5_5gJaJY5_5gJaJZ5]5gJbJZ5^5gJaJY5^5hJbJY5]5gJcJY5\\5hJcJZ5\\5fJdJ[5Z5fJeJ\\5Z5dJeJ^5Y5cJgJ^5W5dJgJ^5W5cJhJ_5V5bJjJ_5T5bJkJ`5S5aJlJa5R5`JnJa5P5`JoJb5c2ZIAU1kMc5a1gJ6g6IgIHZ68TJYOm5h0[3O001O00001O001O000010O01O00001O00001O000000001O000000000000000001O0001O00000000000000000000001O0000000000000O1000O100000000000000O100000000000000O100000000000O10O1000000000000O01000000O0100000O10O1000O1000O01000000O0100000O1000O01000000O0100000O10O10O100O1O010O1O100O1O010O100O00100O0O2O001N101O2M3N3L3N3M2M4M2N3LlUl2"}, "label": "john kerry looking guy in the red tie"}]}
{"id": 522947, "image": "COCO_train2014_000000522947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man who is wearing a red color tie and his one hand on other man ' s shoulder\"."}], "task": "refering", "answer_template": "The \"a man who is wearing a red color tie and his one hand on other man ' s shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 50, 51, 52, 73, 74, 75, 96, 97, 98, 118, 119, 120, 121, 122, 123, 129, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 325, 326], "bbox": [0.10870312499999998, 0.08075117370892018, 0.65371875, 0.9733802816901408], "iscrowd": 0, "area": 56403.95199999998, "rle": {"size": [426, 640], "counts": "V[m0b2g:d0PNRM\\HV3]7RM[HU3_7RMYHU3`7oL[HY3^7jL_H]3[7fL`Hb3Y7aLdHf3V7]1I7I7J6I7I7J1N1O010O0010O0100O1O100O00100O1O100O1O1O1O100O1O1O00100O1O001O10O_MZH_Og7>gHVOZ7f0PISOP7k0UISOk6j0[IROf6j0`ISO`6k0eIRO\\6k0iIROX6l0lIROT6m0oIPOQ6Q1PJmNQ6R1RJkNo5T1SJjNm5W1UJgNk5X1WJfNj5Y1YJbNi5_1ZJ[Ni5d1o2O010O1O10O01O10O1000O01000O10O1000O0100000O010000O10001O1N101O1N101O1O0O2O1O0O2O1O001N2O001N2O1O001oG\\M[5d2_JbM`5^2[JiMc5X2WJoMg5R2SJUNl5k1nI\\NP6e1jI`NV6a1dIdN\\6]1nHXOS7g0hH_OW7b0bHD^7T3001O0000010O00O100000001N100000001O0SO[JTJf5j5\\JVJd5g5_JYJa5g5_JYJb5e5`JZJ`5e5aJZJ`5f5`JZJ`5e5bJZJ^5e5cJ[J^5d5bJ[J_5d5cJ[J]5e5cJ[J]5d5dJ\\J]5b5dJ]J]5c5dJ\\J\\5c5eJ]J[5c5eJ]J\\5a5fJ]J[5c5eJ]J[5b5fJ^J[5a5fJ^JZ5a5gJ^JZ5a5gJ_JZ5`5fJ`JZ5_5gJaJY5_5gJaJZ5]5gJbJZ5^5gJaJY5^5hJbJY5]5gJcJY5\\5hJcJZ5\\5fJdJ[5Z5fJeJ\\5Z5dJeJ^5Y5cJgJ^5W5dJgJ^5W5cJhJ_5V5bJjJ_5T5bJkJ`5S5aJlJa5R5`JnJa5P5`JoJb5c2ZIAU1kMc5a1gJ6g6IgIHZ68TJYOm5h0[3O001O00001O001O000010O01O00001O00001O000000001O000000000000000001O0001O00000000000000000000001O0000000000000O1000O100000000000000O100000000000000O100000000000O10O1000000000000O01000000O0100000O10O1000O1000O01000000O0100000O1000O01000000O0100000O10O10O100O1O010O1O100O1O010O100O00100O0O2O001N101O2M3N3L3N3M2M4M2N3LlUl2"}, "label": "a man who is wearing a red color tie and his one hand on other man ' s shoulder"}]}
{"id": 522947, "image": "COCO_train2014_000000522947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the glasses on\"."}], "task": "refering", "answer_template": "The \"the man with the glasses on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 139, 140, 162, 163, 185, 186, 208, 209, 231, 232, 254, 255, 277, 278, 301, 324], "bbox": [0.043375000000000004, 0.356924882629108, 0.127140625, 0.9659154929577465], "iscrowd": 0, "area": 7259.353750000002, "rle": {"size": [426, 640], "counts": "Xj;8j<=D<I3N3M2N3L3N2\\EgNT9\\1iFjNR9W1kFQOo8P1oFVO`0YOi6b1eH\\O1@W7U1eHBCHe7g0fHT1V7nNfHW1W7jNfH[1W7hNdH^1X7gNaH^1\\7[2M4L4L4L4L3M4L4L4L3M4M3L4L4L3M4L4L4L3MiJTKR2k4kMRLZ1n3cNoLc0Q3[OlMJT2mNUKAV3U1e1VO^K^OW3T1\\1\\OdKXOY3U1T1AiKRO^3U1k0Gc01^OMj0LWO2R1F[OLn0LUO1T7M3MS\\X7"}, "label": "the man with the glasses on"}]}
{"id": 522947, "image": "COCO_train2014_000000522947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a tie holding his phone in a room\"."}], "task": "refering", "answer_template": "The \"a man in a tie holding his phone in a room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 139, 140, 162, 163, 185, 186, 208, 209, 231, 232, 254, 255, 277, 278, 301, 324], "bbox": [0.043375000000000004, 0.356924882629108, 0.127140625, 0.9659154929577465], "iscrowd": 0, "area": 7259.353750000002, "rle": {"size": [426, 640], "counts": "Xj;8j<=D<I3N3M2N3L3N2\\EgNT9\\1iFjNR9W1kFQOo8P1oFVO`0YOi6b1eH\\O1@W7U1eHBCHe7g0fHT1V7nNfHW1W7jNfH[1W7hNdH^1X7gNaH^1\\7[2M4L4L4L4L3M4L4L4L3M4M3L4L4L3M4L4L4L3MiJTKR2k4kMRLZ1n3cNoLc0Q3[OlMJT2mNUKAV3U1e1VO^K^OW3T1\\1\\OdKXOY3U1T1AiKRO^3U1k0Gc01^OMj0LWO2R1F[OLn0LUO1T7M3MS\\X7"}, "label": "a man in a tie holding his phone in a room"}]}
{"id": 416450, "image": "COCO_train2014_000000416450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a black suit showing a picture from his cell phone to a friend\"."}], "task": "refering", "answer_template": "The \"man in a black suit showing a picture from his cell phone to a friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 320, 321, 335, 336, 337, 338, 339, 340, 343, 344], "bbox": [0.35015625, 0.12660421545667447, 0.9948593750000001, 1.0], "iscrowd": 0, "area": 86626.50024999995, "rle": {"size": [427, 640], "counts": "mgm21X=5K4L4L3M4M2M4L3M4L3M4L3M4L2N3M2N2N2O1O2N1O100O1O2N1O100O1O2N1O100O1O2N1O1O100O2N1O1O100O2N1O1O100O1O2N1O1O100O2N1O1O100O2N1O1O100O2N1O1O100O2N1O001O1N101O0O2O001N2O001N101O0O2O1O0O2O001N101O1N1ZGiKX8Y4eGkKX8V4gGlKV8W4gGlKW8U4hGmKU8V4hGnKU8e4M3M3N3L3N2M3N3L3O1O101N100O101N100O100O2O0O100O2O0000001O0000001O0O10001O0000001O000O10001O0000001O000O101O0000001O00000O2O000000001O00000O2O0000001O0000000A`0@?@`0@`0G:M2M3N2N2M4M2M3N2M3N3N1O1O1O2N100O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2O0O1O1O1O2N1O10000O2O000O101O0O10000O101O0O100O10001N10000O10001N10000O10001N100000001O00000000001O00000000001O00000000010O00000000001O0000K5H8I7I7H8I7I7H8I7I7H9H8L4O2N100O1O1O1XKXKl0h4PO\\Ko0f4lN]KT1d4iN_KV1b4fNaK[1_4aNeK^1\\4_NfKa1[4[NiKd1X4XNkKi1V4RNnKm1S4PNoKP2R4lMQLU2o3gMULX2l3eMVL[2k3aMYL^2h3^M[Lc2f3YM]Lf2d3VM_Lj2b3RMbLm2_3PMcLQ3]3kLgLT3Z3hLiLX3X3dLlL[3S4dKoK]4_601O010O001O010O001O010O000010O01O010O001O010O001O01O01O0010O01OWIgK`4Z4\\KiKe4V4XKmKh4S4UKPLl4o3QKTLo4m3mJVLT5i3iJZLW5f3fJ]LZ5c3dJ_L]5a3_JbLa5^3\\JeLe5Z3XJiLh5W3UJlLl5T3PJoLQ6P3lISMT6m2iIVMX6j2dIYM]6f2`I]M`6c2]I`Md6_2YIdMh6\\2TIgMm6X2PIkMP7X41O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O1O?A`1`N`1`N`1`N^k0"}, "label": "man in a black suit showing a picture from his cell phone to a friend"}]}
{"id": 416450, "image": "COCO_train2014_000000416450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald man with a black suit\"."}], "task": "refering", "answer_template": "The \"a bald man with a black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 320, 321, 335, 336, 337, 338, 339, 340, 343, 344], "bbox": [0.35015625, 0.12660421545667447, 0.9948593750000001, 1.0], "iscrowd": 0, "area": 86626.50024999995, "rle": {"size": [427, 640], "counts": "mgm21X=5K4L4L3M4M2M4L3M4L3M4L3M4L2N3M2N2N2O1O2N1O100O1O2N1O100O1O2N1O100O1O2N1O1O100O2N1O1O100O2N1O1O100O1O2N1O1O100O2N1O1O100O2N1O1O100O2N1O1O100O2N1O001O1N101O0O2O001N2O001N101O0O2O1O0O2O001N101O1N1ZGiKX8Y4eGkKX8V4gGlKV8W4gGlKW8U4hGmKU8V4hGnKU8e4M3M3N3L3N2M3N3L3O1O101N100O101N100O100O2O0O100O2O0000001O0000001O0O10001O0000001O000O10001O0000001O000O101O0000001O00000O2O000000001O00000O2O0000001O0000000A`0@?@`0@`0G:M2M3N2N2M4M2M3N2M3N3N1O1O1O2N100O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2O0O1O1O1O2N1O10000O2O000O101O0O10000O101O0O100O10001N10000O10001N10000O10001N100000001O00000000001O00000000001O00000000010O00000000001O0000K5H8I7I7H8I7I7H8I7I7H9H8L4O2N100O1O1O1XKXKl0h4PO\\Ko0f4lN]KT1d4iN_KV1b4fNaK[1_4aNeK^1\\4_NfKa1[4[NiKd1X4XNkKi1V4RNnKm1S4PNoKP2R4lMQLU2o3gMULX2l3eMVL[2k3aMYL^2h3^M[Lc2f3YM]Lf2d3VM_Lj2b3RMbLm2_3PMcLQ3]3kLgLT3Z3hLiLX3X3dLlL[3S4dKoK]4_601O010O001O010O001O010O000010O01O010O001O010O001O01O01O0010O01OWIgK`4Z4\\KiKe4V4XKmKh4S4UKPLl4o3QKTLo4m3mJVLT5i3iJZLW5f3fJ]LZ5c3dJ_L]5a3_JbLa5^3\\JeLe5Z3XJiLh5W3UJlLl5T3PJoLQ6P3lISMT6m2iIVMX6j2dIYM]6f2`I]M`6c2]I`Md6_2YIdMh6\\2TIgMm6X2PIkMP7X41O1O001O1O1O001O1O001O1O1O001O1O001O1O001O1O1O?A`1`N`1`N`1`N^k0"}, "label": "a bald man with a black suit"}]}
{"id": 416450, "image": "COCO_train2014_000000416450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a businessman without a tie sitting next to another man who has a tie\"."}], "task": "refering", "answer_template": "The \"a businessman without a tie sitting next to another man who has a tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 214, 215, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0015, 0.0011241217798594848, 0.492875, 0.9785948477751757], "iscrowd": 0, "area": 81089.73374999998, "rle": {"size": [427, 640], "counts": "Re0[5e2[5001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001N1000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000lMVMgIj2Y6]M`Ic2`6dMYI\\2g6kMRIU2o6QNjHo1V7XNcHh1]7_N\\Ha1d7gNTHY1l7nNmGR1T8TOeGl0[8R20O1M4M2N2N2N2N3L3N2N2N3M2M3N2NA?hNX1K5L3L5L4K5L4K51N3N2N3M2N2M4M2N2N2N3M2M3N3M2000001O00001O01O0001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001OlLQKlMo4T2SKjMm4W2SKiMl4W2VKgMj4Y2XKeMh4\\2YKcMf4]2[KbMe4^2]K`Mc4`2_K^Ma4c2`K\\M_4d2bK[M^4e2dKYM\\4h2eKWMZ4i2hKUMX4k2iKTMW4m2jKQMV4n2mKQMR4n2QLPMo3P3RLoLn3P3ULnLk3R3WLfLo3Y3TL`LQ4`3PLYLV4f3mKSLX4l3kKlK[4T4gKdK_4[4cK_Kc4`4_KXKg4g4\\KRKi4n4j1O1O100O100O100O1O100O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2H8G;D=B?Bf`W4"}, "label": "a businessman without a tie sitting next to another man who has a tie"}]}
{"id": 416450, "image": "COCO_train2014_000000416450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a pink shirt looking at a phone\"."}], "task": "refering", "answer_template": "The \"a man in a pink shirt looking at a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 214, 215, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0015, 0.0011241217798594848, 0.492875, 0.9785948477751757], "iscrowd": 0, "area": 81089.73374999998, "rle": {"size": [427, 640], "counts": "Re0[5e2[5001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001N1000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000lMVMgIj2Y6]M`Ic2`6dMYI\\2g6kMRIU2o6QNjHo1V7XNcHh1]7_N\\Ha1d7gNTHY1l7nNmGR1T8TOeGl0[8R20O1M4M2N2N2N2N3L3N2N2N3M2M3N2NA?hNX1K5L3L5L4K5L4K51N3N2N3M2N2M4M2N2N2N3M2M3N3M2000001O00001O01O0001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001OlLQKlMo4T2SKjMm4W2SKiMl4W2VKgMj4Y2XKeMh4\\2YKcMf4]2[KbMe4^2]K`Mc4`2_K^Ma4c2`K\\M_4d2bK[M^4e2dKYM\\4h2eKWMZ4i2hKUMX4k2iKTMW4m2jKQMV4n2mKQMR4n2QLPMo3P3RLoLn3P3ULnLk3R3WLfLo3Y3TL`LQ4`3PLYLV4f3mKSLX4l3kKlK[4T4gKdK_4[4cK_Kc4`4_KXKg4g4\\KRKi4n4j1O1O100O100O100O1O100O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2H8G;D=B?Bf`W4"}, "label": "a man in a pink shirt looking at a phone"}]}
{"id": 203458, "image": "COCO_train2014_000000203458.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white shirt on the nearside of the road\"."}], "task": "refering", "answer_template": "The \"a white shirt on the nearside of the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346, 347, 368, 369, 370], "bbox": [0.0, 0.326625, 0.121015625, 1.0], "iscrowd": 0, "area": 17222.313250000003, "rle": {"size": [480, 640], "counts": "n4R:P5N2N2N000000002N2N1O2N2N1O2N1O2N2N1O2N2N1O2XGTJX8m5iGSJU8o5kGQJS8R6lGoIQ8T6nGlIP8W6oGiIo7[6oGeIo7^6PHbIn7a6QH_Im7d6RH\\Il7g6SHYIk7j6THVIj7m6UHSIi7P7VHPIh7S7WHmHg7X7N3M2N2N2N2N2N2N2N2N2N2N3M6J6J6J6J6J6J6J5K6J6J6J6J6J6J5K7I:F;E:F;E:F;E:F;E:F;E:FZ^W8"}, "label": "a white shirt on the nearside of the road"}]}
{"id": 203458, "image": "COCO_train2014_000000203458.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sleeve\"."}], "task": "refering", "answer_template": "The \"a white sleeve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346, 347, 368, 369, 370], "bbox": [0.0, 0.326625, 0.121015625, 1.0], "iscrowd": 0, "area": 17222.313250000003, "rle": {"size": [480, 640], "counts": "n4R:P5N2N2N000000002N2N1O2N2N1O2N1O2N2N1O2N2N1O2XGTJX8m5iGSJU8o5kGQJS8R6lGoIQ8T6nGlIP8W6oGiIo7[6oGeIo7^6PHbIn7a6QH_Im7d6RH\\Il7g6SHYIk7j6THVIj7m6UHSIi7P7VHPIh7S7WHmHg7X7N3M2N2N2N2N2N2N2N2N2N2N3M6J6J6J6J6J6J6J5K6J6J6J6J6J6J5K7I:F;E:F;E:F;E:F;E:F;E:FZ^W8"}, "label": "a white sleeve"}]}
{"id": 490184, "image": "COCO_train2014_000000490184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane heading down the runway\"."}], "task": "refering", "answer_template": "The \"an airplane heading down the runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 61, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 129, 132], "bbox": [0.458328125, 0.14037162162162162, 0.9436406250000001, 0.4780743243243243], "iscrowd": 0, "area": 3555.0312500000027, "rle": {"size": [296, 640], "counts": "hee21W900000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O01O000000000001O00000000000000001O000BHlG8T8HlG8T8HlG8T8HlG8b800001O0000000CGkG9U8GkG9U8GkG=Q8CoG>P8BPH?o7AQH`0n7@RHa0m7_OSHc0k7]OUHc0k7]OUHc0k7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHa0m7_OSH?o7AQH>P8BPH<R8EmG:T8FlG9U8GkG8V8HjG8V8HjG8V8HjG8V8HjG8V8HjG8V8HjG8V8HjG8V8IiG7W8IiG7W8IiG7W8IiG7W8IiG7W8=001O000000O1N2O1O1N2O1N2WNSOeJ_1[5W1005K<D<D<D<D5K000000O100001O1O001O3M2N3M000\\OcG=]8CiG7V8JjG6V8JjG6V8IkG7U8IkG7U8IjG8V8HjG8V8HjG8U8IkG7U8IkG7U8HlG8T8HlG8T8HlG8T8HlG8T8HlG8S8ImG7S8ImG7S8ImG7S8HnG<n7DRH=m7CSH>l7BSHa0j7@VHa0i7_OWHb0h7^OXHa0i7_OWHa0i7^OXHb0h7^OXHb0h7^OXHb0g7_OYHa0g7_OYHa0g7_OYH`0h7@XH?i7@XH`0h7@XH?i7AWH6Q8KoG3S8MlG5S8KmG6R8JnG7Q8IoG6R8IoG4T8LlG4S8MmG3S8MmG3S8KoG5b8000000O1000O10000000000000O1002N1O1OK500000O100000O1000000000000000000O1000O10000000000000O1000000000000000O10O1000000000000000000O100000O100000000000O1000000000000000O10O1000O1O0O1O1N01O1N101OSZ:"}, "label": "an airplane heading down the runway"}]}
{"id": 490184, "image": "COCO_train2014_000000490184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane facing away from the camera\"."}], "task": "refering", "answer_template": "The \"the plane facing away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 61, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 129, 132], "bbox": [0.458328125, 0.14037162162162162, 0.9436406250000001, 0.4780743243243243], "iscrowd": 0, "area": 3555.0312500000027, "rle": {"size": [296, 640], "counts": "hee21W900000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O01O000000000001O00000000000000001O000BHlG8T8HlG8T8HlG8T8HlG8b800001O0000000CGkG9U8GkG9U8GkG=Q8CoG>P8BPH?o7AQH`0n7@RHa0m7_OSHc0k7]OUHc0k7]OUHc0k7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHb0l7^OTHa0m7_OSH?o7AQH>P8BPH<R8EmG:T8FlG9U8GkG8V8HjG8V8HjG8V8HjG8V8HjG8V8HjG8V8HjG8V8HjG8V8IiG7W8IiG7W8IiG7W8IiG7W8IiG7W8=001O000000O1N2O1O1N2O1N2WNSOeJ_1[5W1005K<D<D<D<D5K000000O100001O1O001O3M2N3M000\\OcG=]8CiG7V8JjG6V8JjG6V8IkG7U8IkG7U8IjG8V8HjG8V8HjG8U8IkG7U8IkG7U8HlG8T8HlG8T8HlG8T8HlG8T8HlG8S8ImG7S8ImG7S8ImG7S8HnG<n7DRH=m7CSH>l7BSHa0j7@VHa0i7_OWHb0h7^OXHa0i7_OWHa0i7^OXHb0h7^OXHb0h7^OXHb0g7_OYHa0g7_OYHa0g7_OYH`0h7@XH?i7@XH`0h7@XH?i7AWH6Q8KoG3S8MlG5S8KmG6R8JnG7Q8IoG6R8IoG4T8LlG4S8MmG3S8MmG3S8KoG5b8000000O1000O10000000000000O1002N1O1OK500000O100000O1000000000000000000O1000O10000000000000O1000000000000000O10O1000000000000000000O100000O100000000000O1000000000000000O10O1000O1O0O1O1N01O1N101OSZ:"}, "label": "the plane facing away from the camera"}]}
{"id": 490184, "image": "COCO_train2014_000000490184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue and red airplane whose nose is facing towards the camera\"."}], "task": "refering", "answer_template": "The \"the blue and red airplane whose nose is facing towards the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 75, 76, 93, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148], "bbox": [0.065484375, 0.18652027027027027, 0.47809375000000004, 0.6269594594594595], "iscrowd": 0, "area": 6315.407300000002, "rle": {"size": [296, 640], "counts": "aa<2W9100O2N101N1O2O0O00O2N101N1O00000001O01O00000001O01O00000001O01O00000001O01O00000001O01O00000001O01O000000000001O000000000001O0000000000000000000000001O000000000000UON`H2`7N`H2[8000VON^H2b7N^H2b7N_H2`7N`H2[80VON^H6^7JbH6^7JbH6^7JbH6_7J`H6`7J`H6`7J`H7_7IaH7_7IaH7_7J`H6a7I_H5c7K^H2d7N\\H2d7O[H1e7O[H1f7NZH2f7NZH2f7NZH3e7NZH2f7NZH2f7NZH2g7MYH3g7NXH2h7NXH2h7NYH1g7OYH1h7NXH2h7OWH7c7I]H7c7I]H7c7I]H7c7J\\H6d7J\\H6d7J\\H5e7K[H3g7NYH0h70XH0h70XH0h70XH0h70XH0h71WHOi71WHOi71WHOi71WHOi72VHNb5NRL4\\NOa56jKKfNN`5`0aKCnNM`5i0ZKZOVOM`5Q1QKXOZOGe5Y1iJTO^OCk5d1ZJjNJF0Kc5`2\\JQNh5l1VJUNo5f1PJ[NU6`1jIaN[6[1cIfN`6W1`IiNb6n11O2M2O20ON3M2N3K4M2O100O10000O2N1N2O1O000]OgHVOZ7j0QIkNo6U1c0O01O000000001O01O0001O000000010O0000001O01O000001O0001O0001O0000010O2N3M1O0001O000O13M00O2O00001O001N101OWHiNa7^12O001O001O0IWHnNk7P1YHmNh7Q1801N3M3N2N2N2O1N2O1N2OO000O2O0O101N100001N2N2N2N2N4L5KW\\P3"}, "label": "the blue and red airplane whose nose is facing towards the camera"}]}
{"id": 490184, "image": "COCO_train2014_000000490184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue jet\"."}], "task": "refering", "answer_template": "The \"a blue jet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 75, 76, 93, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148], "bbox": [0.065484375, 0.18652027027027027, 0.47809375000000004, 0.6269594594594595], "iscrowd": 0, "area": 6315.407300000002, "rle": {"size": [296, 640], "counts": "aa<2W9100O2N101N1O2O0O00O2N101N1O00000001O01O00000001O01O00000001O01O00000001O01O00000001O01O00000001O01O000000000001O000000000001O0000000000000000000000001O000000000000UON`H2`7N`H2[8000VON^H2b7N^H2b7N_H2`7N`H2[80VON^H6^7JbH6^7JbH6^7JbH6_7J`H6`7J`H6`7J`H7_7IaH7_7IaH7_7J`H6a7I_H5c7K^H2d7N\\H2d7O[H1e7O[H1f7NZH2f7NZH2f7NZH3e7NZH2f7NZH2f7NZH2g7MYH3g7NXH2h7NXH2h7NYH1g7OYH1h7NXH2h7OWH7c7I]H7c7I]H7c7I]H7c7J\\H6d7J\\H6d7J\\H5e7K[H3g7NYH0h70XH0h70XH0h70XH0h70XH0h71WHOi71WHOi71WHOi71WHOi72VHNb5NRL4\\NOa56jKKfNN`5`0aKCnNM`5i0ZKZOVOM`5Q1QKXOZOGe5Y1iJTO^OCk5d1ZJjNJF0Kc5`2\\JQNh5l1VJUNo5f1PJ[NU6`1jIaN[6[1cIfN`6W1`IiNb6n11O2M2O20ON3M2N3K4M2O100O10000O2N1N2O1O000]OgHVOZ7j0QIkNo6U1c0O01O000000001O01O0001O000000010O0000001O01O000001O0001O0001O0000010O2N3M1O0001O000O13M00O2O00001O001N101OWHiNa7^12O001O001O0IWHnNk7P1YHmNh7Q1801N3M3N2N2N2O1N2O1N2OO000O2O0O101N100001N2N2N2N2N4L5KW\\P3"}, "label": "a blue jet"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear bowl with salad in it , next to a plate of pizza\"."}], "task": "refering", "answer_template": "The \"a clear bowl with salad in it , next to a plate of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271], "bbox": [0.51465625, 0.2731542056074766, 0.9156093750000001, 0.7697897196261683], "iscrowd": 0, "area": 42650.154800000004, "rle": {"size": [428, 640], "counts": "PgY4`0e<e0[O9H9H8G9F9K6I6L3N3M2N2M3N3M2M3N2N3M2N2N3N1N3M2N2N2OlKnFo3U93M3M3N2M3M3M3N2O1N2O1N2O1N2O1N2O1N2O2M2O0O2O001N101N2O001N101N101O0O101N101O0O101N10001N10001O001N101O0000001N1000000000001O0000001O0001O01O01O00010O0001L300O2N101N10001O000O100000000000010O0001O00001O00001O00001O0000001O0000001O0O10001N10000O101N10000O100000000000001O0O10000000000000000nIlHk5\\7O0000001O0000001N10001O00001O0O101O001O0O101N101O0O2O001N101N2O001N2O1N2O1N2O1N2N2N2O1N2N2N2O1N2O1N2O1N2N2O2M2N101N101N1O2N1O2N2L4M3M4M2M3M3N2N3N1M7J6I7J5I8G7J6I9G=C;CQ[f0"}, "label": "a clear bowl with salad in it , next to a plate of pizza"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of salad\"."}], "task": "refering", "answer_template": "The \"a bowl of salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271], "bbox": [0.51465625, 0.2731542056074766, 0.9156093750000001, 0.7697897196261683], "iscrowd": 0, "area": 42650.154800000004, "rle": {"size": [428, 640], "counts": "PgY4`0e<e0[O9H9H8G9F9K6I6L3N3M2N2M3N3M2M3N2N3M2N2N3N1N3M2N2N2OlKnFo3U93M3M3N2M3M3M3N2O1N2O1N2O1N2O1N2O1N2O2M2O0O2O001N101N2O001N101N101O0O101N101O0O101N10001N10001O001N101O0000001N1000000000001O0000001O0001O01O01O00010O0001L300O2N101N10001O000O100000000000010O0001O00001O00001O00001O0000001O0000001O0O10001N10000O101N10000O100000000000001O0O10000000000000000nIlHk5\\7O0000001O0000001N10001O00001O0O101O001O0O101N101O0O2O001N101N2O001N2O1N2O1N2O1N2N2N2O1N2N2N2O1N2O1N2O1N2N2O2M2N101N101N1O2N1O2N2L4M3M4M2M3M3N2N3N1M7J6I7J5I8G7J6I9G=C;CQ[f0"}, "label": "a bowl of salad"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small slice of thin crust , cheesy , spinach pizza on a plate with other types of pizza and some meat\"."}], "task": "refering", "answer_template": "The \"a small slice of thin crust , cheesy , spinach pizza on a plate with other types of pizza and some meat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 239], "bbox": [0.38321875, 0.3462149532710281, 0.51996875, 0.7102570093457944], "iscrowd": 0, "area": 8440.440249999996, "rle": {"size": [428, 640], "counts": "eaV3`0l<X1gNd1]NT1lN1O0O2O000O2O00001N101O0O101O00010O001O00N2M3N2N3M2N2N2M3N2N2N2N3M2N2M3N2N2N2N2N3L3N2N2N2N2N2N2M4M2N2N2N2N2N2M3N2N2N3M2N2M3N2N2N2N2N2N2M4M2N2N2N2N3L5L4L5K4L3M2N2M3N2N1OQUP4"}, "label": "a small slice of thin crust , cheesy , spinach pizza on a plate with other types of pizza and some meat"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza slice ith green topping\"."}], "task": "refering", "answer_template": "The \"pizza slice ith green topping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 239], "bbox": [0.38321875, 0.3462149532710281, 0.51996875, 0.7102570093457944], "iscrowd": 0, "area": 8440.440249999996, "rle": {"size": [428, 640], "counts": "eaV3`0l<X1gNd1]NT1lN1O0O2O000O2O00001N101O0O101O00010O001O00N2M3N2N3M2N2N2M3N2N2N2N3M2N2M3N2N2N2N2N3L3N2N2N2N2N2N2M4M2N2N2N2N2N2M3N2N2N3M2N2M3N2N2N2N2N2N2M4M2N2N2N2N3L5L4L5K4L3M2N2M3N2N1OQUP4"}, "label": "pizza slice ith green topping"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza slices on the bottom on the white plate\"."}], "task": "refering", "answer_template": "The \"pizza slices on the bottom on the white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 192, 214, 215, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283], "bbox": [0.17203125, 0.46014018691588787, 0.463578125, 0.8174299065420562], "iscrowd": 0, "area": 10238.780200000003, "rle": {"size": [428, 640], "counts": "aY^18g<?N1N3NN2001N10000O10001N10000O10001N10000O1000001O0O100000001O0000000000001O0000000000001O00000O1000001N1000000O1000001N1000000O10001O0O100O010O100O10O0100O100O010O100O10O01O1O1O100O001O1O1O1O001O100O1O001O1O1O1O00100O1O1O1O1M3L4]Nc1O100O1O1O1O100O1O1O100O1O1O1O100O1O001g2YM<D1O00001O00001O00001O001O00001O00001O00001O0000O1N2M3N2N2M3N2N2M3N2N1N3N2N2N2M33M7I7I7I7I7H8H8H8Hj^_4"}, "label": "pizza slices on the bottom on the white plate"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"some pizza and chicken wings on a plate\"."}], "task": "refering", "answer_template": "The \"some pizza and chicken wings on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307], "bbox": [0.118984375, 0.2905373831775701, 0.512109375, 0.9234345794392523], "iscrowd": 0, "area": 53096.772099999995, "rle": {"size": [428, 640], "counts": "\\QP14n<?@`0A>E<H8I6I8H7I6J5K5K4L5K5L4K5K5K4L5K5M3M3M2N3L4M3M3M3M2M4M1O2N1O2M2O2N1O2N2N1N3N1O2N1O2M201N1O2N2N1O2N101N1O2N1O2N101N2N1O2N1O2N101N1O2N1O2N1O100O2N1O1O1O101N100O10001N100O100O2O000O100O2O000O100O2O000O10001O0O100000000O01000000000O010000000O01000000000O010000000O010000000O10O100000000O10000000000O100O1O100O100O100O100O100O101N100O2N101N101N101N101N100O2O0O2O0O2O0O2N1O1O2N1O2N1O2O0O2N1O1O2N1O2N1O2N1O2O1N2N2N2N2O0O2N2O1N2N2O1N1O2O1N2N2O1N1O101N1O1N2N3L3N2M3N2M3N2M3N2M3N2M3N2M3_LUG_2m8SMcGh2`8PMjGl2^9K5H8C=C=C>F;H_VR4"}, "label": "some pizza and chicken wings on a plate"}]}
{"id": 244425, "image": "COCO_train2014_000000244425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white dish of food containing three pieces of pizza and pasta\"."}], "task": "refering", "answer_template": "The \"a white dish of food containing three pieces of pizza and pasta\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307], "bbox": [0.118984375, 0.2905373831775701, 0.512109375, 0.9234345794392523], "iscrowd": 0, "area": 53096.772099999995, "rle": {"size": [428, 640], "counts": "\\QP14n<?@`0A>E<H8I6I8H7I6J5K5K4L5K5L4K5K5K4L5K5M3M3M2N3L4M3M3M3M2M4M1O2N1O2M2O2N1O2N2N1N3N1O2N1O2M201N1O2N2N1O2N101N1O2N1O2N101N2N1O2N1O2N101N1O2N1O2N1O100O2N1O1O1O101N100O10001N100O100O2O000O100O2O000O100O2O000O10001O0O100000000O01000000000O010000000O01000000000O010000000O010000000O10O100000000O10000000000O100O1O100O100O100O100O100O101N100O2N101N101N101N101N100O2O0O2O0O2O0O2N1O1O2N1O2N1O2O0O2N1O1O2N1O2N1O2N1O2O1N2N2N2N2O0O2N2O1N2N2O1N1O2O1N2N2O1N1O101N1O1N2N3L3N2M3N2M3N2M3N2M3N2M3N2M3_LUG_2m8SMcGh2`8PMjGl2^9K5H8C=C=C>F;H_VR4"}, "label": "a white dish of food containing three pieces of pizza and pasta"}]}
{"id": 39629, "image": "COCO_train2014_000000039629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"racket in mans hand\"."}], "task": "refering", "answer_template": "The \"racket in mans hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 53, 54, 55, 56, 70, 71, 72, 73, 89, 90, 107, 125, 142], "bbox": [0.1411666666666667, 0.09668750000000001, 0.41375000000000006, 0.37253125], "iscrowd": 0, "area": 6449.229799999998, "rle": {"size": [640, 480], "counts": "dbZ1:]c0=J7J3M3N1N2N3M2N2O2M2N3M2O1O2N1N2O1O1O1O001O001O001N101O1O001O001O001O001O10O01O001O01O01O00000010O001O00010O001O01O2O0O2N100O2O0O2N101N100010O10O0100kNg^O@Za0<k^OBVa0;n^ODQa09U_ODl`08Y_OGg`05^_OIc`04__OMa`0Oc_O0]`0Mf_O3[`0Ih_O8b`0ZOb_Oe0ma010O010O001N1O2N1O2N1O2N1O2N1O2N1O2N3Mkd01S[O3N2N2N2N3M2N2N2N1O0011N2O0O1O1O1N3N1N2O1N2M3Ln\\_5"}, "label": "racket in mans hand"}]}
{"id": 39629, "image": "COCO_train2014_000000039629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis racked being held by a man to the right of a woman\"."}], "task": "refering", "answer_template": "The \"a tennis racked being held by a man to the right of a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 53, 54, 55, 56, 70, 71, 72, 73, 89, 90, 107, 125, 142], "bbox": [0.1411666666666667, 0.09668750000000001, 0.41375000000000006, 0.37253125], "iscrowd": 0, "area": 6449.229799999998, "rle": {"size": [640, 480], "counts": "dbZ1:]c0=J7J3M3N1N2N3M2N2O2M2N3M2O1O2N1N2O1O1O1O001O001O001N101O1O001O001O001O001O10O01O001O01O01O00000010O001O00010O001O01O2O0O2N100O2O0O2N101N100010O10O0100kNg^O@Za0<k^OBVa0;n^ODQa09U_ODl`08Y_OGg`05^_OIc`04__OMa`0Oc_O0]`0Mf_O3[`0Ih_O8b`0ZOb_Oe0ma010O010O001N1O2N1O2N1O2N1O2N1O2N1O2N3Mkd01S[O3N2N2N2N3M2N2N2N1O0011N2O0O1O1O1N3N1N2O1N2M3Ln\\_5"}, "label": "a tennis racked being held by a man to the right of a woman"}]}
{"id": 39629, "image": "COCO_train2014_000000039629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a tennis racquet\"."}], "task": "refering", "answer_template": "The \"a woman holding a tennis racquet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 77, 78, 79, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 166, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 246, 247, 248, 249, 250, 263, 264, 266, 267, 268, 280, 281, 283, 284, 285, 297, 298, 301, 302, 315, 318, 319, 332, 335, 336, 349, 350, 353, 354, 365, 366, 370, 371, 372], "bbox": [0.45841666666666664, 0.089890625, 0.9258333333333333, 0.9572968749999999], "iscrowd": 0, "area": 50154.73895, "rle": {"size": [640, 480], "counts": "ggY4o0ib08G:G9E:_Ob0I6N3N1N3N2M2O2M2O2WAgLg<[3jBhMa<Z2oB]Ng<f1kBiNR=X1lBmNo<V1nBnNo<S1oBROl<Q1QCSOl<n0RCVOk<R4M2N2N2O1N7I7gIXHMo7]LSHm00f2S8lK[HX1Ek2W8\\KcHc1YOo2e:kL]ET3g:gL\\EX3h:bLZE]3j:]LYEb3l:XLWEg3n:RLTEm3R;lKPEc2VOYMP<MmDf2XOZMQ<HjDk2ZO[MQ<ChDo2\\OZMn=c2WBZMj=c2[B[Mf=a2_B\\Mb=a2dB[M^=a2gB]MW=b2nB[MR=d2RCYMn<g2VCVMh<k2`CmL`<S3NhLhCY3Q<PMmCP3l;YMRDg2g;bMVD_2c;jM[DV2^;RNaDn1X;[NeDf1a96]FJ[9`0cF@T9k0iFVOo8T1oFlNR9^1cFbN^9h1VFYNk9Q2jEoMW:[2]EgMc:k501O001O001O00001O001O001O00001O001O001O001O2N2N2N2N3M2N2N3\\GXFj6j9cHhF\\7Z9]HlFb7V9WHPGh7d:2O4L5K4K]J^Dm2^;nLjDT3Q;hLTE[3h:`L\\Ee3_:WLeEm3X:oKkET4S:gKPF]4m9_KWFd4g9WK]Fl4a9PKbFS5]9gJgF]5V9^JnFe5P9WJRGm5m8mIWGV6g8eI]G^6b8]IaGf6]8UIgGn6X8lHlGW7R8eHQH]7V:5F:G9G8H9G9F:H7K6J6J7J5J7I6J6J7J5J6J6M3oHYDe5^7TIgLR1oKj5o6dIoL<VLQ6_6SJXME]LY6R6_J_MROcL`6h5hJbMaNkLh8\\3RIFPNQM]9f2oHP1_7\\NPI[1_7QNPIe1^7[MYI_2V<N1O2N1O2N2N1O2N2O0OY1gN3M3M3M3M3M2N1O2N1O1O1O1O1O1O101N100O100O100O2O000O100O100O2O0O100O100O100O10O01O1N2O0O2O1N101N2O0O2O1N101N2O0O2O1N2OZme0"}, "label": "a woman holding a tennis racquet"}]}
{"id": 39629, "image": "COCO_train2014_000000039629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman posing with a tennis racket\"."}], "task": "refering", "answer_template": "The \"a woman posing with a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 77, 78, 79, 93, 94, 95, 96, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 166, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 246, 247, 248, 249, 250, 263, 264, 266, 267, 268, 280, 281, 283, 284, 285, 297, 298, 301, 302, 315, 318, 319, 332, 335, 336, 349, 350, 353, 354, 365, 366, 370, 371, 372], "bbox": [0.45841666666666664, 0.089890625, 0.9258333333333333, 0.9572968749999999], "iscrowd": 0, "area": 50154.73895, "rle": {"size": [640, 480], "counts": "ggY4o0ib08G:G9E:_Ob0I6N3N1N3N2M2O2M2O2WAgLg<[3jBhMa<Z2oB]Ng<f1kBiNR=X1lBmNo<V1nBnNo<S1oBROl<Q1QCSOl<n0RCVOk<R4M2N2N2O1N7I7gIXHMo7]LSHm00f2S8lK[HX1Ek2W8\\KcHc1YOo2e:kL]ET3g:gL\\EX3h:bLZE]3j:]LYEb3l:XLWEg3n:RLTEm3R;lKPEc2VOYMP<MmDf2XOZMQ<HjDk2ZO[MQ<ChDo2\\OZMn=c2WBZMj=c2[B[Mf=a2_B\\Mb=a2dB[M^=a2gB]MW=b2nB[MR=d2RCYMn<g2VCVMh<k2`CmL`<S3NhLhCY3Q<PMmCP3l;YMRDg2g;bMVD_2c;jM[DV2^;RNaDn1X;[NeDf1a96]FJ[9`0cF@T9k0iFVOo8T1oFlNR9^1cFbN^9h1VFYNk9Q2jEoMW:[2]EgMc:k501O001O001O00001O001O001O00001O001O001O001O2N2N2N2N3M2N2N3\\GXFj6j9cHhF\\7Z9]HlFb7V9WHPGh7d:2O4L5K4K]J^Dm2^;nLjDT3Q;hLTE[3h:`L\\Ee3_:WLeEm3X:oKkET4S:gKPF]4m9_KWFd4g9WK]Fl4a9PKbFS5]9gJgF]5V9^JnFe5P9WJRGm5m8mIWGV6g8eI]G^6b8]IaGf6]8UIgGn6X8lHlGW7R8eHQH]7V:5F:G9G8H9G9F:H7K6J6J7J5J7I6J6J7J5J6J6M3oHYDe5^7TIgLR1oKj5o6dIoL<VLQ6_6SJXME]LY6R6_J_MROcL`6h5hJbMaNkLh8\\3RIFPNQM]9f2oHP1_7\\NPI[1_7QNPIe1^7[MYI_2V<N1O2N1O2N2N1O2N2O0OY1gN3M3M3M3M3M2N1O2N1O1O1O1O1O1O101N100O100O100O2O000O100O100O2O0O100O100O100O10O01O1N2O0O2O1N101N2O0O2O1N101N2O0O2O1N2OZme0"}, "label": "a woman posing with a tennis racket"}]}
{"id": 39629, "image": "COCO_train2014_000000039629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 39, 40, 41, 56, 57, 58, 72, 73, 74, 75, 88, 89, 90, 91, 92, 93, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 257, 258, 259, 261, 262, 263, 274, 275, 276, 278, 279, 280, 291, 292, 293, 296, 297, 308, 309, 310, 313, 314, 325, 326, 327, 329, 330, 342, 343, 344, 346, 347, 348, 358, 359, 360, 363, 364, 365], "bbox": [0.06591666666666667, 0.058421875, 0.5183541666666667, 0.9550625], "iscrowd": 0, "area": 56689.56189999999, "rle": {"size": [640, 480], "counts": "obd04hc04L4N2O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O100O1O001O100O001O100O1O001O100O0cEXOa1k0\\NYOY1Q1dNSOVN:`Kl0W6mNmMh0^Kd0b6hNjM[4T2iKeM]4X2gKbM^4[2eKaM_4\\2eKaM\\4]2hK`MZ4]2jK`MX4]2lK`MV4\\2nKbMT4[2PLcMP4[2SLcMo3Z2SLeMo3X2TLfMn3W2ULgMl3W2WLfMl3W2WLfMl3W2WLfMl3W2VLgMm3V2VLgMl3W2WLfMl3Z2TLcMo3\\2RLaMQ4^2PL_MR4a2oK\\MT4c2lK[MU4f2lKWMU4j2lKSMV4m2kKPMV4Q3kKlLV4U3kKhLV4Y3jKeLX4\\3hKaLY4`3hK\\LZ4f3fKUL]2fMRLV6b1oKV2jNcKX5W2iKP2F[Kb4f2cKj18[KW4T3SK_1P1XKn3b3eJS1i1VKc3i9fLRF\\3j9^3L5L3L5L4M2O2N101N1O2O1N1O2O0O2N2O0O2O0O2N101N2N101N1O001O1O001O10O1001NjNlGcFT8X9SHgFl7V9YHhFg7S9`HlF_7Q9fHmFZ7R9iHmFV7R9mHlFS7S9QIkFn6T9UIjFk6U9XIjFg6U9R2O1O1O002N1100O100O100O100O100O16J7ULbEfMm;;gDA^;7hDF];3iDI\\;0iDM];JjD2\\;FiD7^;@hD<_;ZOgDc0_;UOdDj0Z?O1000O1000000O10000\\@cMc=^2ZBhMb=Y2\\BkMa=V2\\BoMa=R2]BSN_=n1^BUNa=k1^BWNa=j1\\BYNc=h1[B[Nc=f1YB^Nf=c1VBaNi=`1RBeNm=\\1oAhNP>Y1lAlNR>b3O1O1O2N1N2mMQKSEP5i1fJU5n0kH^4b1YKU5e0RIS4\\1nKS5;ZIh3b0UMj2aMZNn1SN]3f0YMf2]N_MY1nNQ3l0^M`2cNYMX1VOh2n0bM\\2hNVMW1[O`2Q1fMV2mNSMn6f0]Jl1PORMi6o0`Jd1]8ZNlGY1Y8eNTH`0]8^OnGGd87c5N3N2N1N3N2M3N8HkV`4"}, "label": "a man holding a tennis racket"}]}
{"id": 39629, "image": "COCO_train2014_000000039629.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"man in white holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 39, 40, 41, 56, 57, 58, 72, 73, 74, 75, 88, 89, 90, 91, 92, 93, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 257, 258, 259, 261, 262, 263, 274, 275, 276, 278, 279, 280, 291, 292, 293, 296, 297, 308, 309, 310, 313, 314, 325, 326, 327, 329, 330, 342, 343, 344, 346, 347, 348, 358, 359, 360, 363, 364, 365], "bbox": [0.06591666666666667, 0.058421875, 0.5183541666666667, 0.9550625], "iscrowd": 0, "area": 56689.56189999999, "rle": {"size": [640, 480], "counts": "obd04hc04L4N2O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O100O1O001O100O001O100O1O001O100O0cEXOa1k0\\NYOY1Q1dNSOVN:`Kl0W6mNmMh0^Kd0b6hNjM[4T2iKeM]4X2gKbM^4[2eKaM_4\\2eKaM\\4]2hK`MZ4]2jK`MX4]2lK`MV4\\2nKbMT4[2PLcMP4[2SLcMo3Z2SLeMo3X2TLfMn3W2ULgMl3W2WLfMl3W2WLfMl3W2WLfMl3W2VLgMm3V2VLgMl3W2WLfMl3Z2TLcMo3\\2RLaMQ4^2PL_MR4a2oK\\MT4c2lK[MU4f2lKWMU4j2lKSMV4m2kKPMV4Q3kKlLV4U3kKhLV4Y3jKeLX4\\3hKaLY4`3hK\\LZ4f3fKUL]2fMRLV6b1oKV2jNcKX5W2iKP2F[Kb4f2cKj18[KW4T3SK_1P1XKn3b3eJS1i1VKc3i9fLRF\\3j9^3L5L3L5L4M2O2N101N1O2O1N1O2O0O2N2O0O2O0O2N101N2N101N1O001O1O001O10O1001NjNlGcFT8X9SHgFl7V9YHhFg7S9`HlF_7Q9fHmFZ7R9iHmFV7R9mHlFS7S9QIkFn6T9UIjFk6U9XIjFg6U9R2O1O1O002N1100O100O100O100O100O16J7ULbEfMm;;gDA^;7hDF];3iDI\\;0iDM];JjD2\\;FiD7^;@hD<_;ZOgDc0_;UOdDj0Z?O1000O1000000O10000\\@cMc=^2ZBhMb=Y2\\BkMa=V2\\BoMa=R2]BSN_=n1^BUNa=k1^BWNa=j1\\BYNc=h1[B[Nc=f1YB^Nf=c1VBaNi=`1RBeNm=\\1oAhNP>Y1lAlNR>b3O1O1O2N1N2mMQKSEP5i1fJU5n0kH^4b1YKU5e0RIS4\\1nKS5;ZIh3b0UMj2aMZNn1SN]3f0YMf2]N_MY1nNQ3l0^M`2cNYMX1VOh2n0bM\\2hNVMW1[O`2Q1fMV2mNSMn6f0]Jl1PORMi6o0`Jd1]8ZNlGY1Y8eNTH`0]8^OnGGd87c5N3N2N1N3N2M3N8HkV`4"}, "label": "man in white holding a tennis racket"}]}
{"id": 187090, "image": "COCO_train2014_000000187090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black boy riding a skateboard\"."}], "task": "refering", "answer_template": "The \"a black boy riding a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 244, 245, 246, 267, 268, 269, 270, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 340, 358, 359, 360, 361, 382, 383, 384, 405, 406, 407, 428, 429, 430, 451, 452, 474, 475], "bbox": [0.542859375, 0.4, 0.807796875, 0.9038906250000001], "iscrowd": 0, "area": 21133.5609, "rle": {"size": [640, 640], "counts": "`Xi61mc02N2N2N2N3L3N2N2N2N3M2N2N2O1O2M2O1O1N2O2N1N2O1N2O2N2M4M3M3L3N3L4M3M3L3N3M3L4M2N3L4M3L3O2M3N2M3M2O2M3M3^NULkBl3R=YLjBj3S=ZLiBi3U2PL`7;XFg3R2TLc79WFf3Q2VLe78WFe3m1YLi77UFc3m1ZLl77SFb3k1]Lo75SF`3i1_LR85QF_3h1_LV86mE_3g1_L[86hE`3h1]L_8W5[GmJd8W5VGlJh8X5RGlJm8`5eFcJZ9Q6PFRJo9P8O1O10000O100000O1000000000000ETFbFl9[9XFdFh9Y9\\FfFd9W9`FhF`9U9eFiF[9T9iFkFV9S9m0M3M3M3M3M3M3L4M9G9G9G9\\Ne1F9Gm0RO4M3M3M3L4M3M3M2N3L4M3K5K5L4K5K5L4K5K6K5L4O100O100O100O100O2O0O1O100O100O100O2O1N101N2O0O2O0O2O1N101N2O0O2O1N2O0O2O1N1L5L4L4L3L5LS`\\2"}, "label": "a black boy riding a skateboard"}]}
{"id": 187090, "image": "COCO_train2014_000000187090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black skateboarding\"."}], "task": "refering", "answer_template": "The \"a man in black skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 244, 245, 246, 267, 268, 269, 270, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 340, 358, 359, 360, 361, 382, 383, 384, 405, 406, 407, 428, 429, 430, 451, 452, 474, 475], "bbox": [0.542859375, 0.4, 0.807796875, 0.9038906250000001], "iscrowd": 0, "area": 21133.5609, "rle": {"size": [640, 640], "counts": "`Xi61mc02N2N2N2N3L3N2N2N2N3M2N2N2O1O2M2O1O1N2O2N1N2O1N2O2N2M4M3M3L3N3L4M3M3L3N3M3L4M2N3L4M3L3O2M3N2M3M2O2M3M3^NULkBl3R=YLjBj3S=ZLiBi3U2PL`7;XFg3R2TLc79WFf3Q2VLe78WFe3m1YLi77UFc3m1ZLl77SFb3k1]Lo75SF`3i1_LR85QF_3h1_LV86mE_3g1_L[86hE`3h1]L_8W5[GmJd8W5VGlJh8X5RGlJm8`5eFcJZ9Q6PFRJo9P8O1O10000O100000O1000000000000ETFbFl9[9XFdFh9Y9\\FfFd9W9`FhF`9U9eFiF[9T9iFkFV9S9m0M3M3M3M3M3M3L4M9G9G9G9\\Ne1F9Gm0RO4M3M3M3L4M3M3M2N3L4M3K5K5L4K5K5L4K5K6K5L4O100O100O100O100O2O0O1O100O100O100O2O1N101N2O0O2O0O2O1N101N2O0O2O1N2O0O2O1N1L5L4L4L3L5LS`\\2"}, "label": "a man in black skateboarding"}]}
{"id": 514771, "image": "COCO_train2014_000000514771.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane with the orange stripe and black words\"."}], "task": "refering", "answer_template": "The \"the plane with the orange stripe and black words\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.11781250000000001, 0.6228705882352942, 0.546859375, 0.7962117647058825], "iscrowd": 0, "area": 9319.025249999999, "rle": {"size": [425, 640], "counts": "Rko01X=1VC1X<0eC4X<NeC6Y<KdC9Z<HcC<Z<;O000O010O10O10OE<O1N1010O10000O10001N100O10000O10000O1O2N1N2O1O1N2O10000O101O0000000000000000001O001O2N1N01O1O1O1N2N2N2001O001O001O1O001N2O0O2N1O2N2N1N3N1O29E1O100O1O1000000000000O10000000SDmNd;S1[DoNd;Q1[DQOd;o0[DTOc;l0\\DVOc;j0\\DWOd;i0[DXOe;V100000000001I60000000000000000001O00001O001O0000O100O1O1M3H800000000001O001O001O001O1O00NnCnNQ<R1PDmNP<S1QDmNn;S140000O10000000000000000O10001O0000000000000O10000000000000000O1000000000000000000O1000000002N1O2N2N00O100000O10O0100O10O0100O100O010O100O10O01O100O1O00100O1O0HdNdD]1\\;cNdD]1\\;cNcD^1];cNaD^1_;bN`D^1a;cN^D]1b;cN]D]1d;31O1O1O1O1O1O1O1O1D<N2N2N2N2N2N2N2N2O1N2N2N2O11O0000001O0000O1M3L]Sh3"}, "label": "the plane with the orange stripe and black words"}]}
{"id": 514771, "image": "COCO_train2014_000000514771.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white plane with orange and white stripes towards the bottom of the body\"."}], "task": "refering", "answer_template": "The \"a small white plane with orange and white stripes towards the bottom of the body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264], "bbox": [0.11781250000000001, 0.6228705882352942, 0.546859375, 0.7962117647058825], "iscrowd": 0, "area": 9319.025249999999, "rle": {"size": [425, 640], "counts": "Rko01X=1VC1X<0eC4X<NeC6Y<KdC9Z<HcC<Z<;O000O010O10O10OE<O1N1010O10000O10001N100O10000O10000O1O2N1N2O1O1N2O10000O101O0000000000000000001O001O2N1N01O1O1O1N2N2N2001O001O001O1O001N2O0O2N1O2N2N1N3N1O29E1O100O1O1000000000000O10000000SDmNd;S1[DoNd;Q1[DQOd;o0[DTOc;l0\\DVOc;j0\\DWOd;i0[DXOe;V100000000001I60000000000000000001O00001O001O0000O100O1O1M3H800000000001O001O001O001O1O00NnCnNQ<R1PDmNP<S1QDmNn;S140000O10000000000000000O10001O0000000000000O10000000000000000O1000000000000000000O1000000002N1O2N2N00O100000O10O0100O10O0100O100O010O100O10O01O100O1O00100O1O0HdNdD]1\\;cNdD]1\\;cNcD^1];cNaD^1_;bN`D^1a;cN^D]1b;cN]D]1d;31O1O1O1O1O1O1O1O1D<N2N2N2N2N2N2N2N2O1N2N2N2O11O0000001O0000O1M3L]Sh3"}, "label": "a small white plane with orange and white stripes towards the bottom of the body"}]}
{"id": 277202, "image": "COCO_train2014_000000277202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the background of this image takes on a blurry appearance\"."}], "task": "refering", "answer_template": "The \"a man in the background of this image takes on a blurry appearance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 50, 66, 67, 83, 84, 100, 101, 117, 118, 134, 135, 151, 152, 168, 169, 185, 186, 202, 203, 219, 220, 236, 237, 253, 254, 270, 271, 286, 287, 288, 303, 304, 305, 320, 321, 322, 338, 339], "bbox": [0.867875, 0.039421875, 1.0, 0.8434687500000001], "iscrowd": 0, "area": 21443.8312, "rle": {"size": [640, 480], "counts": "^SU89`c0:F9G9G:E:_Ob0_O`0_Ob0_O`0@a0mMR2\\Je5ZOe0N2M4L100O001O100O1O100O1O1O010O1O1]OiB\\JX=]5QCaJo<X5YCfJh<T5`CiJa<P5gCoJY<j4PDSKk97oEKWOf4V1UK^9f0jEG_OZ4Z<\\LlCBGn3Z<mLdC\\O0b3Z<_NcC]1Y<T3N6J6K6O1N2N1O2O1N1O2N2O1N1O2N2O0O2N2N2N10XO"}, "label": "a man in the background of this image takes on a blurry appearance"}]}
{"id": 277202, "image": "COCO_train2014_000000277202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blurry image of man at far right of picture\"."}], "task": "refering", "answer_template": "The \"blurry image of man at far right of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 50, 66, 67, 83, 84, 100, 101, 117, 118, 134, 135, 151, 152, 168, 169, 185, 186, 202, 203, 219, 220, 236, 237, 253, 254, 270, 271, 286, 287, 288, 303, 304, 305, 320, 321, 322, 338, 339], "bbox": [0.867875, 0.039421875, 1.0, 0.8434687500000001], "iscrowd": 0, "area": 21443.8312, "rle": {"size": [640, 480], "counts": "^SU89`c0:F9G9G:E:_Ob0_O`0_Ob0_O`0@a0mMR2\\Je5ZOe0N2M4L100O001O100O1O100O1O1O010O1O1]OiB\\JX=]5QCaJo<X5YCfJh<T5`CiJa<P5gCoJY<j4PDSKk97oEKWOf4V1UK^9f0jEG_OZ4Z<\\LlCBGn3Z<mLdC\\O0b3Z<_NcC]1Y<T3N6J6K6O1N2N1O2O1N1O2N2O1N1O2N2O0O2N2N2N10XO"}, "label": "blurry image of man at far right of picture"}]}
{"id": 277202, "image": "COCO_train2014_000000277202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black umbrella and a black and white umbrella lean up against some cardboard boxes\"."}], "task": "refering", "answer_template": "The \"a black umbrella and a black and white umbrella lean up against some cardboard boxes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 241, 242, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 309, 310, 311, 326, 327, 328, 345], "bbox": [0.12391666666666666, 0.5845625, 0.3844166666666667, 0.91875], "iscrowd": 0, "area": 11176.49205, "rle": {"size": [640, 480], "counts": "QnU12mc03M2N2O2M2N3M2N3N1N2\\]OYOla0j0n]OZOQb0h0j]O\\OUb0g0e]O^OYb0U101N1O1O2N101N1O2N1O3N2M3M2N3M3N2M3M3M3M3M3N2M3M3M3M3N2M2N3M3M3L4L4K5K5K5L4K5J6H8J6L4K4OL6J6I8I6J6J6J6J6I5L1O001O4L4L5J6K4RM]_O[2Sa0O2N7I7H8I8Hl@eNn;T1PDSOR<f0lCBU<7iC0Y<IeC?\\<YOcCn0`<jN^C^1[?7I6J7I7I7H7I8G9QOcLRAe3i>hLf@`3V?h0L4N2N101N2O1NYOh0mNR1M3M4_OS_OkMPa0T2?O001N6Kb0^O0O2O1O1N2L4E;E;E:Fbng5"}, "label": "a black umbrella and a black and white umbrella lean up against some cardboard boxes"}]}
{"id": 277202, "image": "COCO_train2014_000000277202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white umbrella\"."}], "task": "refering", "answer_template": "The \"black and white umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 241, 242, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 309, 310, 311, 326, 327, 328, 345], "bbox": [0.12391666666666666, 0.5845625, 0.3844166666666667, 0.91875], "iscrowd": 0, "area": 11176.49205, "rle": {"size": [640, 480], "counts": "QnU12mc03M2N2O2M2N3M2N3N1N2\\]OYOla0j0n]OZOQb0h0j]O\\OUb0g0e]O^OYb0U101N1O1O2N101N1O2N1O3N2M3M2N3M3N2M3M3M3M3M3N2M3M3M3M3N2M2N3M3M3L4L4K5K5K5L4K5J6H8J6L4K4OL6J6I8I6J6J6J6J6I5L1O001O4L4L5J6K4RM]_O[2Sa0O2N7I7H8I8Hl@eNn;T1PDSOR<f0lCBU<7iC0Y<IeC?\\<YOcCn0`<jN^C^1[?7I6J7I7I7H7I8G9QOcLRAe3i>hLf@`3V?h0L4N2N101N2O1NYOh0mNR1M3M4_OS_OkMPa0T2?O001N6Kb0^O0O2O1O1N2L4E;E;E:Fbng5"}, "label": "black and white umbrella"}]}
{"id": 277202, "image": "COCO_train2014_000000277202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a black shirt sitting with a glass of wine\"."}], "task": "refering", "answer_template": "The \"woman in a black shirt sitting with a glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 295, 296, 297, 298, 299, 300, 301, 302, 303, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 347, 348, 349, 350, 351, 352, 353, 354, 355, 365, 366, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.28102083333333333, 0.07425000000000001, 0.9805624999999999, 0.991296875], "iscrowd": 0, "area": 121965.999, "rle": {"size": [640, 480], "counts": "Uhd28_c0;D<E;D<D<E;D<D<C=C=C=B>I7K5L4K4L3M3M3N1N3M3M3M3N2M3N2O100O100O10O0100O10O0100O010O1dDnJj7R5oGVKo7j4kG^KT8b4fGeKX8\\4aGlK^8S4\\GULb8l3WG\\Lh8d3QGdLm8\\3nFkLQ9U3hFSMV9n2cFZM\\9g6M2O1N2O1N2I7E;F:F:G9G9H8H9F9H8G9J6M3N2M3M3N2M3N2M3N3L3M3N2M3N2M3M3N2M3N2M3M3N3L3N2M3N2M3M3N2M3N2M3M3N2N3N1O1N2O1O1N2O1O1N2O1O1N2O1O0O2O0O2O1O0O2O1O0O2O1O0O2O000O100O10000O10001N10000000000001O0000001O0000001O0000001O0000001O00000010O000001O0000001O0000001O0000001O1O2N1O1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O1O1O3M5K5K5K5K4L5K5K5K5K4L6J:F:F:F:F:F>Cc0\\O3M1O1O1O1O1O1O2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N7I:F:F;F9F:F;E:F:F7I7I6J6J6J7I6J6J7I6J4L4L4L4L4L4L4L3M4L4L4L4L4M3L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L3M2N2N2N2N2N2N2N2N3M2O1N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N1O1O1O2N1O1OgP5"}, "label": "woman in a black shirt sitting with a glass of wine"}]}
{"id": 277202, "image": "COCO_train2014_000000277202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a glass\"."}], "task": "refering", "answer_template": "The \"a woman holding a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 141, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 183, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 295, 296, 297, 298, 299, 300, 301, 302, 303, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 347, 348, 349, 350, 351, 352, 353, 354, 355, 365, 366, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.28102083333333333, 0.07425000000000001, 0.9805624999999999, 0.991296875], "iscrowd": 0, "area": 121965.999, "rle": {"size": [640, 480], "counts": "Uhd28_c0;D<E;D<D<E;D<D<C=C=C=B>I7K5L4K4L3M3M3N1N3M3M3M3N2M3N2O100O100O10O0100O10O0100O010O1dDnJj7R5oGVKo7j4kG^KT8b4fGeKX8\\4aGlK^8S4\\GULb8l3WG\\Lh8d3QGdLm8\\3nFkLQ9U3hFSMV9n2cFZM\\9g6M2O1N2O1N2I7E;F:F:G9G9H8H9F9H8G9J6M3N2M3M3N2M3N2M3N3L3M3N2M3N2M3M3N2M3N2M3M3N3L3N2M3N2M3M3N2M3N2M3M3N2N3N1O1N2O1O1N2O1O1N2O1O1N2O1O0O2O0O2O1O0O2O1O0O2O1O0O2O000O100O10000O10001N10000000000001O0000001O0000001O0000001O0000001O00000010O000001O0000001O0000001O0000001O1O2N1O1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O1O1O3M5K5K5K5K4L5K5K5K5K4L6J:F:F:F:F:F>Cc0\\O3M1O1O1O1O1O1O2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N7I:F:F;F9F:F;E:F:F7I7I6J6J6J7I6J6J7I6J4L4L4L4L4L4L4L3M4L4L4L4L4M3L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L3M2N2N2N2N2N2N2N2N3M2O1N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N1O1O1O2N1O1OgP5"}, "label": "a woman holding a glass"}]}
{"id": 326357, "image": "COCO_train2014_000000326357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cockatoo is scratching its head\"."}], "task": "refering", "answer_template": "The \"a white cockatoo is scratching its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 34, 35, 36, 37, 38, 39, 40, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.40818750000000004, 0.009536585365853658, 0.81325, 1.0], "iscrowd": 0, "area": 53031.17345, "rle": {"size": [410, 640], "counts": "QdX32a<6N2N101UDL0MX:9ZE94H`:c1N2N2O1N7I8I6J6I6K3M3M4L3M4L3M5K5J6J5L5K5L2M4L4M3L4L3N3L4L4M3L3M4M3L4L4M2M4L4E\\IZJg6b5<M3L4L4L3M3M100O2N001O00010O00001O001`LSHS2l7nMVHo1j7RNXHk1i7TNZHi1f7WN_Hd1a7\\NcHa1]7^NgH^1Y7bNlHY1T7gNSIR1n6mN[Ik0d6UOiI>W6BPJ7Q6ISJ3l5MXJOh51]JJd56^JHa58aJF_5:bJE_5:bJE^5<bJD^5;cJE\\5;eJFY5;gJEY5:hJFW5;iJEV5;kJEU5:lJFS5:nJER5;oJDR5<nJCR5=oJBS5<nJCX57iJH]52dJMa5N`J0e5M[J2j5IWJ7l5FTJ:P6AQJ>R6@mIa0U6\\OlId0U900000000O100000000000kK]OSLb0m3DnK<Q4KiK5W40dK0[46`KJ_4<\\KDd4a0WK@h4c0UK]Oj4d0VK\\Oj4d0VK\\Oj4e0UK[Ok4e0UK[Ok4e0UK[Ok4e0UK[Oj4f0VKZOj4f0VK[Oi4f0VKZOj4f0VKZOj4f0VKZOj4f0VKZOi4g0WKYOi4g0WKYOi4g0WKYOi4f0XKZOh4e0YK[Og4e0YK[Of4e0[K[Oe4d0\\K[Oe4d0\\K[Oe4d0\\K[Oe4d0\\K[Oe4d0\\K\\Oc4d0]K]Oc4c0]K]Oc4b0^K^Ob4a0_K_Oa4`0`K@`4`0`K@_4a0aK_O_4a0aK_O_4`0bK@^4`0bK@^4`0bK@^4`0bK@]4a0cK_O]4`0dK@\\4`0dK@\\4`0dK@\\4`0dK_O]4a0cK_O\\4b0dK^OXOB\\2o0]N_OVOE[2l0_N_OUOIY2h0bN_OTOLX2e0dN_OSOOW2b0bLUOc0:c03T2>bL^Oc02e05T2;^LGf0Ig08S28mK=V1SOj0;P2X2oL]MQ1=n1^2iLUMY1`0k1c2dLnL`1b0h1k2^LcLj1d0f1R3WL[LR2f0d1l3]NVL`1k3^NWLh0b4XO_KnN[6P1gIPOY6o0iIPOW6n0kISOT6k0nIUOR6h0RJWOn5f0XJfNWNFa7b1]J[N^NMX7h1R3UOTEIm:4WEJi:5YEJh:4[EJe:5]EJd:5]EKb:5_EJa:6`EIa:6aEH_:8bEG_:8bEG^:9cEF^:9cEG\\:9eEF\\:9fEC\\:=eE_O_:a0l0010O01O10O010O01O010O10O10O10O010O0100O010O10O01000O01000O0100O10O10O10001N10000O10001N100O10001N2OXQ`1"}, "label": "a white cockatoo is scratching its head"}]}
{"id": 326357, "image": "COCO_train2014_000000326357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird with its head turned to the side\"."}], "task": "refering", "answer_template": "The \"the bird with its head turned to the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 34, 35, 36, 37, 38, 39, 40, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.40818750000000004, 0.009536585365853658, 0.81325, 1.0], "iscrowd": 0, "area": 53031.17345, "rle": {"size": [410, 640], "counts": "QdX32a<6N2N101UDL0MX:9ZE94H`:c1N2N2O1N7I8I6J6I6K3M3M4L3M4L3M5K5J6J5L5K5L2M4L4M3L4L3N3L4L4M3L3M4M3L4L4M2M4L4E\\IZJg6b5<M3L4L4L3M3M100O2N001O00010O00001O001`LSHS2l7nMVHo1j7RNXHk1i7TNZHi1f7WN_Hd1a7\\NcHa1]7^NgH^1Y7bNlHY1T7gNSIR1n6mN[Ik0d6UOiI>W6BPJ7Q6ISJ3l5MXJOh51]JJd56^JHa58aJF_5:bJE_5:bJE^5<bJD^5;cJE\\5;eJFY5;gJEY5:hJFW5;iJEV5;kJEU5:lJFS5:nJER5;oJDR5<nJCR5=oJBS5<nJCX57iJH]52dJMa5N`J0e5M[J2j5IWJ7l5FTJ:P6AQJ>R6@mIa0U6\\OlId0U900000000O100000000000kK]OSLb0m3DnK<Q4KiK5W40dK0[46`KJ_4<\\KDd4a0WK@h4c0UK]Oj4d0VK\\Oj4d0VK\\Oj4e0UK[Ok4e0UK[Ok4e0UK[Ok4e0UK[Oj4f0VKZOj4f0VK[Oi4f0VKZOj4f0VKZOj4f0VKZOj4f0VKZOi4g0WKYOi4g0WKYOi4g0WKYOi4f0XKZOh4e0YK[Og4e0YK[Of4e0[K[Oe4d0\\K[Oe4d0\\K[Oe4d0\\K[Oe4d0\\K[Oe4d0\\K\\Oc4d0]K]Oc4c0]K]Oc4b0^K^Ob4a0_K_Oa4`0`K@`4`0`K@_4a0aK_O_4a0aK_O_4`0bK@^4`0bK@^4`0bK@^4`0bK@]4a0cK_O]4`0dK@\\4`0dK@\\4`0dK@\\4`0dK_O]4a0cK_O\\4b0dK^OXOB\\2o0]N_OVOE[2l0_N_OUOIY2h0bN_OTOLX2e0dN_OSOOW2b0bLUOc0:c03T2>bL^Oc02e05T2;^LGf0Ig08S28mK=V1SOj0;P2X2oL]MQ1=n1^2iLUMY1`0k1c2dLnL`1b0h1k2^LcLj1d0f1R3WL[LR2f0d1l3]NVL`1k3^NWLh0b4XO_KnN[6P1gIPOY6o0iIPOW6n0kISOT6k0nIUOR6h0RJWOn5f0XJfNWNFa7b1]J[N^NMX7h1R3UOTEIm:4WEJi:5YEJh:4[EJe:5]EJd:5]EKb:5_EJa:6`EIa:6aEH_:8bEG_:8bEG^:9cEF^:9cEG\\:9eEF\\:9fEC\\:=eE_O_:a0l0010O01O10O010O01O010O10O10O10O010O0100O010O10O01000O01000O0100O10O10O10001N10000O10001N100O10001N2OXQ`1"}, "label": "the bird with its head turned to the side"}]}
{"id": 326357, "image": "COCO_train2014_000000326357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bird sitting with both feet on the branch\"."}], "task": "refering", "answer_template": "The \"a white bird sitting with both feet on the branch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 74, 75, 76, 97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 328], "bbox": [0.19659374999999998, 0.13948780487804877, 0.47015624999999994, 0.9785365853658536], "iscrowd": 0, "area": 36454.2214, "rle": {"size": [410, 640], "counts": "jdb17b<9F9G:G9PFgNm7b1cGiNY8a1WGhNg8a1hFiNU9`2M4M2M3M4M2M3M3N3L3G9YOh0XOg0XOh0^Ob0@h0E<F:E;F5K0O101O0O100O101O0O10001N100O101O0O100O5L4L3LnE[MU:a2kE_MY:_23JdEhM`:W23I]EQNc:S2O1O100O100O100O1K5G9G9G;M3N2N3M2N2M3N3M2N2N2N3L3N2N3M2N2N2M4M2N2N2N3M2M3N2N3M2N2N3L3N2N2N3M2N2N2M4M2N2N3M2N1N2O2N1O1O1O1N2O1O1O1O2N1N2O1O1O1O1O1O101O000O100000000O101O000O1000000O10001O2M4M3M3aNVHfMn7U2ZHeMi7W2^HcMe7X2cHbM`7Z2gH_M]7\\2kH^MX7^2oH\\MT7_2h1L4K5L4K5L4K5L4K5L4K5L4K5L4K;Fi]W4"}, "label": "a white bird sitting with both feet on the branch"}]}
{"id": 326357, "image": "COCO_train2014_000000326357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller bird at the left\"."}], "task": "refering", "answer_template": "The \"the smaller bird at the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 74, 75, 76, 97, 98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 328], "bbox": [0.19659374999999998, 0.13948780487804877, 0.47015624999999994, 0.9785365853658536], "iscrowd": 0, "area": 36454.2214, "rle": {"size": [410, 640], "counts": "jdb17b<9F9G:G9PFgNm7b1cGiNY8a1WGhNg8a1hFiNU9`2M4M2M3M4M2M3M3N3L3G9YOh0XOg0XOh0^Ob0@h0E<F:E;F5K0O101O0O100O101O0O10001N100O101O0O100O5L4L3LnE[MU:a2kE_MY:_23JdEhM`:W23I]EQNc:S2O1O100O100O100O1K5G9G9G;M3N2N3M2N2M3N3M2N2N2N3L3N2N3M2N2N2M4M2N2N2N3M2M3N2N3M2N2N3L3N2N2N3M2N2N2M4M2N2N3M2N1N2O2N1O1O1O1N2O1O1O1O2N1N2O1O1O1O1O1O101O000O100000000O101O000O1000000O10001O2M4M3M3aNVHfMn7U2ZHeMi7W2^HcMe7X2cHbM`7Z2gH_M]7\\2kH^MX7^2oH\\MT7_2h1L4K5L4K5L4K5L4K5L4K5L4K5L4K;Fi]W4"}, "label": "the smaller bird at the left"}]}
{"id": 416477, "image": "COCO_train2014_000000416477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman throwing a frisbee\"."}], "task": "refering", "answer_template": "The \"a woman throwing a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 270, 271, 272, 273, 289, 290, 294, 295, 296, 297, 311, 312, 313, 319, 320, 334, 335, 343], "bbox": [0.3146875, 0.2630444964871194, 0.951609375, 0.9750585480093676], "iscrowd": 0, "area": 37088.84710000001, "rle": {"size": [427, 640], "counts": "_Qd22f<e0M2O2N1N3N1O2M2OO10O10000O01000O1000O010000O01000O100O01000000O100000000O10000000000000000O10001O0000001O0000001O00000O2O0000000000000000000000000O1000000000O1N2O1O1N2O1N2O1N101O1N2O1N2O1N2O1O100000000O01000000000000O1000O10O100000000O1000000O100000000000O1_NaNVG_1j8nNiFR1V9\\O]Fd0c9CVF=i9DWF<o4_N]OU1dK<m4cN\\OR1gK;k4gN\\Oo0gK;k4kNZOk0kK:i4mN\\Oi0kK:i4nNZOj0lK8i4oN[Oi0lK9g4oN\\Oi0kK:h4nN]Oi0hK;j4mN]Oi0gK<l4kN]Of2b0[M]Of2b0[M^Oe2a0\\M^Oe2b0[M^Oe2a0\\M^Oe2b0[M^Oe2a0\\M^Of2a0[M^Oe2a0\\M^Oe2b0[M^Oe2a0\\M_Od2`0]M_Od2?^MAb2=`MBa2=`MC`2;bMD`2:aMF_28cMG_26cMJ^24cMK^23dMM]21dMN]2OfM1[2IjM7W2BoM>Q2[OVNe0k1TO[Nk0f1PO_NP1b1kNbNU1_1gNdNY1\\1dNgN\\1Z1VNSOj1m0TNTOl1n0QNSOP2m0mMTOU2m0hMSOZ2n0cMRO_2n0^MROe2o0XMQOi2Q1UMmNn2S1QMlNQ3U1nLiNT3X1jLgNX3Z1fLdN]3\\1bLcN`3^1^LaNe3_1ZL_Ni3a1UL]No3b1PL\\NT4d1jK[NY4e1eKYN_4g1`KWNc4h1\\KVNi4i1UKVNn4j1QKSNS5n1jJoM[5R2bJlMb5T2]JhMi5X2TJgMo5Y2oIfMT6[2jIcMY6^2dIaM_6_2_I`Md6a2YI^Mk6^43M3N2N2O2M2N2M3N2N3L3N3L3N3M2M3N3M2O2M2N3M2O2M2N3N1N2O1O001N2O1O1O0O2O1O1O1N2O001O1N2O001O1N2O001O1N2O001O1N2O001O1N2O00100O1O010O100O00100O1O010000000000O10000000000O10000000O0010000O10000000000000001O002N3M2O1N2N1O2N2N1O2N1O2N1O2M2Oc0]O2M2O0O2O001N10000O100O10000O10001O00000000000000000000001O00000000001O000001O0001O00000001O01O00000000001O000O1000000O2N1O100O1O101N1O1O100O1O2N2O2M3M2N3N1N3M2NUa<"}, "label": "a woman throwing a frisbee"}]}
{"id": 416477, "image": "COCO_train2014_000000416477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman playing frisbee at the beach\"."}], "task": "refering", "answer_template": "The \"a woman playing frisbee at the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 270, 271, 272, 273, 289, 290, 294, 295, 296, 297, 311, 312, 313, 319, 320, 334, 335, 343], "bbox": [0.3146875, 0.2630444964871194, 0.951609375, 0.9750585480093676], "iscrowd": 0, "area": 37088.84710000001, "rle": {"size": [427, 640], "counts": "_Qd22f<e0M2O2N1N3N1O2M2OO10O10000O01000O1000O010000O01000O100O01000000O100000000O10000000000000000O10001O0000001O0000001O00000O2O0000000000000000000000000O1000000000O1N2O1O1N2O1N2O1N101O1N2O1N2O1N2O1O100000000O01000000000000O1000O10O100000000O1000000O100000000000O1_NaNVG_1j8nNiFR1V9\\O]Fd0c9CVF=i9DWF<o4_N]OU1dK<m4cN\\OR1gK;k4gN\\Oo0gK;k4kNZOk0kK:i4mN\\Oi0kK:i4nNZOj0lK8i4oN[Oi0lK9g4oN\\Oi0kK:h4nN]Oi0hK;j4mN]Oi0gK<l4kN]Of2b0[M]Of2b0[M^Oe2a0\\M^Oe2b0[M^Oe2a0\\M^Oe2b0[M^Oe2a0\\M^Of2a0[M^Oe2a0\\M^Oe2b0[M^Oe2a0\\M_Od2`0]M_Od2?^MAb2=`MBa2=`MC`2;bMD`2:aMF_28cMG_26cMJ^24cMK^23dMM]21dMN]2OfM1[2IjM7W2BoM>Q2[OVNe0k1TO[Nk0f1PO_NP1b1kNbNU1_1gNdNY1\\1dNgN\\1Z1VNSOj1m0TNTOl1n0QNSOP2m0mMTOU2m0hMSOZ2n0cMRO_2n0^MROe2o0XMQOi2Q1UMmNn2S1QMlNQ3U1nLiNT3X1jLgNX3Z1fLdN]3\\1bLcN`3^1^LaNe3_1ZL_Ni3a1UL]No3b1PL\\NT4d1jK[NY4e1eKYN_4g1`KWNc4h1\\KVNi4i1UKVNn4j1QKSNS5n1jJoM[5R2bJlMb5T2]JhMi5X2TJgMo5Y2oIfMT6[2jIcMY6^2dIaM_6_2_I`Md6a2YI^Mk6^43M3N2N2O2M2N2M3N2N3L3N3L3N3M2M3N3M2O2M2N3M2O2M2N3N1N2O1O001N2O1O1O0O2O1O1O1N2O001O1N2O001O1N2O001O1N2O001O1N2O001O1N2O00100O1O010O100O00100O1O010000000000O10000000000O10000000O0010000O10000000000000001O002N3M2O1N2N1O2N2N1O2N1O2N1O2M2Oc0]O2M2O0O2O001N10000O100O10000O10001O00000000000000000000001O00000000001O000001O0001O00000001O01O00000000001O000O1000000O2N1O100O1O101N1O1O100O1O2N2O2M3M2N3N1N3M2NUa<"}, "label": "a woman playing frisbee at the beach"}]}
{"id": 309983, "image": "COCO_train2014_000000309983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slivered carrots on a plate with brocolli , grapes and a burger\"."}], "task": "refering", "answer_template": "The \"slivered carrots on a plate with brocolli , grapes and a burger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 234, 235, 236, 237, 238, 258, 259, 260, 282, 283, 284, 306, 307], "bbox": [0.159328125, 0.5309583333333333, 0.388921875, 0.7851041666666667], "iscrowd": 0, "area": 4151.417549999999, "rle": {"size": [480, 640], "counts": "WR`1?^>4M2O2M2000001O0000010O0001O01O0001O01O0001O01O0001O01O01O0001O01O0001O01O0002N100O1O100O2N1O100000001O000001O00010O00001O000010O0001O000010O0001O000010O0001O000010O0001O00001O00001O00cNVOcDj0];VOdDi0[;WOfDi0Z;XOeDi0Y;XOhDg0W;ZOiDf0W;ZOiDf0V;\\OjDc0U;^OkDb0U;^OkDb0T;@lD?S;BmD>S;BmD>R;CoD<P;FoD:Q;FoD:P;GQE8o:HQE8n:IRE7m:IUE6k:IVE7i:JWE6h:JZE5f:JYE8f:IYE8f:HYE9h:FXE:h:GVE;i:EWE:k:EUE:l:GRE9P;FPE9Q;GnD8T;HlD7V;GjD9W;GiD8X;GiD8Y;GgD8Z;GgD8Z;HgD6Z;IhD5Y;KhD2Z;NgD0Z;0gDN[;1fDM[;3fDK[;6dDI];7f100000001N1Lmof5"}, "label": "slivered carrots on a plate with brocolli , grapes and a burger"}]}
{"id": 309983, "image": "COCO_train2014_000000309983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrot slices\"."}], "task": "refering", "answer_template": "The \"carrot slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 188, 189, 190, 212, 213, 236, 237, 259, 260, 261, 283, 284], "bbox": [0.1785625, 0.44112500000000004, 0.37215625, 0.7401458333333333], "iscrowd": 0, "area": 4515.279599999999, "rle": {"size": [480, 640], "counts": "Ree11l>4M3M3M3N2N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1000000000000000000000000000000000001O00001O00010O001O00001O001O00001O001O00001O010O000010O00010O01O00010O001O01O01O0010O00010O0001O010O000010O01O00010O0010O0001O01O0N2O1N3M2N2O2M2N2O1N3M2O1N3M3N3L3M3N3L3M^Sl5"}, "label": "carrot slices"}]}
{"id": 309983, "image": "COCO_train2014_000000309983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrots on a plate\"."}], "task": "refering", "answer_template": "The \"carrots on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 188, 189, 190, 212, 213, 236, 237, 259, 260, 261, 283, 284], "bbox": [0.1785625, 0.44112500000000004, 0.37215625, 0.7401458333333333], "iscrowd": 0, "area": 4515.279599999999, "rle": {"size": [480, 640], "counts": "Ree11l>4M3M3M3N2N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1000000000000000000000000000000000001O00001O00010O001O00001O001O00001O001O00001O010O000010O00010O01O00010O001O01O01O0010O00010O0001O010O000010O01O00010O0010O0001O01O0N2O1N3M2N2O2M2N2O1N3M2O1N3M3N3L3M3N3L3M^Sl5"}, "label": "carrots on a plate"}]}
{"id": 309983, "image": "COCO_train2014_000000309983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green broccoli on a clear plate\"."}], "task": "refering", "answer_template": "The \"green broccoli on a clear plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332, 333, 353, 354, 355, 356, 357, 378, 379], "bbox": [0.318453125, 0.5444791666666667, 0.638625, 0.96675], "iscrowd": 0, "area": 23565.64905, "rle": {"size": [480, 640], "counts": "cPP35g>8G8I7J6L1O1O1O1O1O2N1O1O1O1O1O2N1O1[NaNhE`1V:bNiE_1T:eNjE\\1S:gNlE[1P:hNoEY1n9kNQFU1l9nNSFS1m9nNQFS1n9nNaEZOGi1g:nNaEZOGj1g:mN`E[OGi1h:mN`E[OGi1h:nN_EZOHi1i:mN^E\\OGh1j:mN^E\\OGh1j:nN]E\\OGh1j:mN^E\\OGh1k:lN]E]OGh1i:oN^Eb1`:`N_Ea1^:cN`E^1^:dNaE^1\\:dNcE]1\\:eNbE\\1]:eNbE\\1]:d100O2N1O1O1O1O1O2N1O1O10000O2O000O10000O2O0aFPKn8P5PGTKm8m4QGVKn8j4PGYKo8g4oF\\Ko8f4nF]KQ9c4mF`KQ9Y5O1O1N2O1O1N2O1O1N10000O11O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O00001O00001M2O1O1O2N1N2cM^F[Od9d0_FXOb9h0aFTOa9k0bFQO_9o0dFmN^9R1eFiN]9W1fFeN\\9Z1fFcN\\9\\1gF`NZ9`1iF\\NY9c1jFYNW9g1lFTNW9k1lFQNV9n1mFnMT9R2oFjMS9U2oFhMS9W2PGeMQ9Z2SGaMP9^2SG^Mn8b2i1O1O001O001O1O0000O1O1O1O1O1O1O1O00100O1O1O1O1O00000000000001O002N1O2N1O101N1O2N1O1O2N1O2N1O2N1O1O2N3M6J7I6J5K2O1N2N1O2N2N2N2N1O2N2H8GYo[3"}, "label": "green broccoli on a clear plate"}]}
{"id": 309983, "image": "COCO_train2014_000000309983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"four pieces of broccoli closest to carrots\"."}], "task": "refering", "answer_template": "The \"four pieces of broccoli closest to carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332, 333, 353, 354, 355, 356, 357, 378, 379], "bbox": [0.318453125, 0.5444791666666667, 0.638625, 0.96675], "iscrowd": 0, "area": 23565.64905, "rle": {"size": [480, 640], "counts": "cPP35g>8G8I7J6L1O1O1O1O1O2N1O1O1O1O1O2N1O1[NaNhE`1V:bNiE_1T:eNjE\\1S:gNlE[1P:hNoEY1n9kNQFU1l9nNSFS1m9nNQFS1n9nNaEZOGi1g:nNaEZOGj1g:mN`E[OGi1h:mN`E[OGi1h:nN_EZOHi1i:mN^E\\OGh1j:mN^E\\OGh1j:nN]E\\OGh1j:mN^E\\OGh1k:lN]E]OGh1i:oN^Eb1`:`N_Ea1^:cN`E^1^:dNaE^1\\:dNcE]1\\:eNbE\\1]:eNbE\\1]:d100O2N1O1O1O1O1O2N1O1O10000O2O000O10000O2O0aFPKn8P5PGTKm8m4QGVKn8j4PGYKo8g4oF\\Ko8f4nF]KQ9c4mF`KQ9Y5O1O1N2O1O1N2O1O1N10000O11O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O00001O00001M2O1O1O2N1N2cM^F[Od9d0_FXOb9h0aFTOa9k0bFQO_9o0dFmN^9R1eFiN]9W1fFeN\\9Z1fFcN\\9\\1gF`NZ9`1iF\\NY9c1jFYNW9g1lFTNW9k1lFQNV9n1mFnMT9R2oFjMS9U2oFhMS9W2PGeMQ9Z2SGaMP9^2SG^Mn8b2i1O1O001O001O1O0000O1O1O1O1O1O1O1O00100O1O1O1O1O00000000000001O002N1O2N1O101N1O2N1O1O2N1O2N1O2N1O1O2N3M6J7I6J5K2O1N2N1O2N2N2N2N1O2N2H8GYo[3"}, "label": "four pieces of broccoli closest to carrots"}]}
{"id": 318179, "image": "COCO_train2014_000000318179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair with an open checker board back with someone sitting on it\"."}], "task": "refering", "answer_template": "The \"a brown chair with an open checker board back with someone sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 132, 133, 134, 135, 150, 151, 152, 167, 168, 169, 184, 185, 201, 202, 218, 219, 234, 235, 236, 251, 252, 253, 268, 269, 285, 286, 302, 303, 319, 320], "bbox": [0.6292083333333333, 0.17753125, 0.99475, 0.82471875], "iscrowd": 0, "area": 24276.665999999994, "rle": {"size": [640, 480], "counts": "kll51nc01O2M2O2N2N1O2N1O2N1O2N2M2O2N1O2N2N1O2N1O2N1N3N2N1CiNP^OY1oa0<010O01O1O010O00100O001O010O001O10O01O010O1O0010O01O010O1O0010O01O010O1O0010O01mGmMWNT2c1RN\\Nn1^1YNbNg1X1^NhNb1R1eNmN\\1m0jNSOU1g0QOYOP1a0VO^Ok0;[OEe05BK>OH08JN63D3<M]O:b0FYO?g0BROe0m0[OnNk0R1UOgNQ1Y1oNaNX1^1iN\\N]1c1dNVNb1k1]NPNi1o1XNjMo1U2QNfMT2[2lM^M[2a2eMZM`2f2aMSMf2l2[MmLl2S3SMhLR3X3oLaLW3_3RMSLS3n3CUIb0l6X61O101K4K5K6J5K5K5K6J5L4K5K6J5K5K6J5K5K5K6J5K5K6J5K5L4K6J5K5K6J5K5K5K6J5K5K5K6J5K5K6K4K5K5K6J5K5K6J5K5K5K6J5K5K6J5L4K5K6J5K5K5K6J5K5K6J5K5K5K6J5K5L5J5K5KZg1"}, "label": "a brown chair with an open checker board back with someone sitting on it"}]}
{"id": 236260, "image": "COCO_train2014_000000236260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef that is wearing gloves and looking down\"."}], "task": "refering", "answer_template": "The \"a chef that is wearing gloves and looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 17, 18, 19, 20, 21, 22, 23, 24, 25, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 165, 166, 167, 168, 169, 170, 171, 172, 174, 175, 176, 177, 180, 181, 182, 183, 184, 185, 186, 190, 191, 192, 193, 195, 196, 197, 198, 199, 200, 201, 202, 205, 206, 207, 208, 214, 215, 216, 217, 218, 220, 223, 231, 232, 233], "bbox": [0.0023653395784543328, 0.0, 0.9424355971896956, 0.69371875], "iscrowd": 0, "area": 107502.94635, "rle": {"size": [640, 427], "counts": "Tg0Q2na0S4mKS2mM1O2N1O2N1O2M2O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O100O1O1O2O0O1O1O100O1O1O2O0O1O1O100O1O1O101N1O1O100O1O1O10001N10000O100O10001N10000O100O10000O2O000O101N10000O2O000O2O0O10001N10001N10000O2O0000001N10001O000O10001O0O10000O101O0O100O10001N10000O10001N100O10000O2O000O100O10001N100O100O101O0O2N2O0O2N2O1N2O1N2N3N1N3M2O2M2N3N1N3M2O2M2N3N1O2N1O2N1O2N1O1O\\MdIdG\\6[8fIdGZ6[8gIeGY6Z8iIeGW6Z8kIeGU6Z8mIeGS6Z8oIfGP6X8SJgGm5X8TJiGk5V8WJiGi5V8YJjGf5U8\\JjGd5U8^JjGb5U8`JjG`5U8aJjG`5T8cJkG]5T8eJjG\\5U8fJjGZ5U8hJiGY5V8iJiGW5V8jJlGT5S8nJoGo4P8SKPHl4n7WKQHi4n7YKQHg4m7[KTHd4k7^KTHb4k7`KUH_4i7dKVH\\4i7eKWH[4g7hKYHW4f7kKYHU4f7mKZHR4d7PL[HQ4d7QL[Ho3d7SLZHn3d7TL[Hm3c7VL[Hk3d7WLZHj3d7YLZHh3d7ZL[Hg3d7[LZHf3d7]LZHd3d7^L]Ha3a7bL^H^3a7dL_H[3_7hL`HX3^7jLbHV3\\7mLdHR3[7PMdHP3Z7SMfHl2X7WMgHi2X7YMgHg2W7]MfHd2Z7]MeHc2\\7^MaHc2`7]M_Hc2a7_M]Ha2d7_MZHb2g7_MWHa2j7`MSHa2n7_MQHa2P8aMlG`2U8aMiG_2X8cMdG^2\\8dM_G_2b8bMYGa2h8aMRGb2o8W41O1O1O001O1O1O010000O101O0O10000O2O001N10001O0O2O001N10001N101O0OeFhGo6[5UGQMi1RNh6j4bGVM_1QNR7f4dG\\MR1nM^7c4cG`Mi0nMk7\\4`GeM>RNY8R4]GmM0VNj8g3ZGSNB[N\\9Y3WG]NSO_Nm9ROWFR3o0j0R9jKTFX3o0j0b9SOdFi0^9SNTEeNd1T3Z9nM^EfN_1X3U9jMgEeN[1]3Q9gMnEdNX1a3m8fM_HX2f7aM_H\\2g7]M]H`2h7YM\\Hd2k7TMZHj2j7oL[Hn2k7kLXHT3\\<M3N1N3N2M2O2M3N1N3N2M2O2M3O0O2N2O0O1O1O01O00010O0001O101N2N2N2O1N2N2O0O1O1O100O00100O1O00100O1O10O01O1O10O01O1O10O10001O001O001O0O2O00001O001O001O0000O100O1000O0100O100O1N2N2N3N1N3ZOc]OD`b09f0M3M3MXX?"}, "label": "a chef that is wearing gloves and looking down"}]}
{"id": 236260, "image": "COCO_train2014_000000236260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef wearing a white uniform cutting meat on the table\"."}], "task": "refering", "answer_template": "The \"a chef wearing a white uniform cutting meat on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 17, 18, 19, 20, 21, 22, 23, 24, 25, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 165, 166, 167, 168, 169, 170, 171, 172, 174, 175, 176, 177, 180, 181, 182, 183, 184, 185, 186, 190, 191, 192, 193, 195, 196, 197, 198, 199, 200, 201, 202, 205, 206, 207, 208, 214, 215, 216, 217, 218, 220, 223, 231, 232, 233], "bbox": [0.0023653395784543328, 0.0, 0.9424355971896956, 0.69371875], "iscrowd": 0, "area": 107502.94635, "rle": {"size": [640, 427], "counts": "Tg0Q2na0S4mKS2mM1O2N1O2N1O2M2O1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O100O1O1O2O0O1O1O100O1O1O2O0O1O1O100O1O1O101N1O1O100O1O1O10001N10000O100O10001N10000O100O10000O2O000O101N10000O2O000O2O0O10001N10001N10000O2O0000001N10001O000O10001O0O10000O101O0O100O10001N10000O10001N100O10000O2O000O100O10001N100O100O101O0O2N2O0O2N2O1N2O1N2N3N1N3M2O2M2N3N1N3M2O2M2N3N1O2N1O2N1O2N1O1O\\MdIdG\\6[8fIdGZ6[8gIeGY6Z8iIeGW6Z8kIeGU6Z8mIeGS6Z8oIfGP6X8SJgGm5X8TJiGk5V8WJiGi5V8YJjGf5U8\\JjGd5U8^JjGb5U8`JjG`5U8aJjG`5T8cJkG]5T8eJjG\\5U8fJjGZ5U8hJiGY5V8iJiGW5V8jJlGT5S8nJoGo4P8SKPHl4n7WKQHi4n7YKQHg4m7[KTHd4k7^KTHb4k7`KUH_4i7dKVH\\4i7eKWH[4g7hKYHW4f7kKYHU4f7mKZHR4d7PL[HQ4d7QL[Ho3d7SLZHn3d7TL[Hm3c7VL[Hk3d7WLZHj3d7YLZHh3d7ZL[Hg3d7[LZHf3d7]LZHd3d7^L]Ha3a7bL^H^3a7dL_H[3_7hL`HX3^7jLbHV3\\7mLdHR3[7PMdHP3Z7SMfHl2X7WMgHi2X7YMgHg2W7]MfHd2Z7]MeHc2\\7^MaHc2`7]M_Hc2a7_M]Ha2d7_MZHb2g7_MWHa2j7`MSHa2n7_MQHa2P8aMlG`2U8aMiG_2X8cMdG^2\\8dM_G_2b8bMYGa2h8aMRGb2o8W41O1O1O001O1O1O010000O101O0O10000O2O001N10001O0O2O001N10001N101O0OeFhGo6[5UGQMi1RNh6j4bGVM_1QNR7f4dG\\MR1nM^7c4cG`Mi0nMk7\\4`GeM>RNY8R4]GmM0VNj8g3ZGSNB[N\\9Y3WG]NSO_Nm9ROWFR3o0j0R9jKTFX3o0j0b9SOdFi0^9SNTEeNd1T3Z9nM^EfN_1X3U9jMgEeN[1]3Q9gMnEdNX1a3m8fM_HX2f7aM_H\\2g7]M]H`2h7YM\\Hd2k7TMZHj2j7oL[Hn2k7kLXHT3\\<M3N1N3N2M2O2M3N1N3N2M2O2M3O0O2N2O0O1O1O01O00010O0001O101N2N2N2O1N2N2O0O1O1O100O00100O1O00100O1O10O01O1O10O01O1O10O10001O001O001O0O2O00001O001O001O0000O100O1000O0100O100O1N2N2N3N1N3ZOc]OD`b09f0M3M3MXX?"}, "label": "a chef wearing a white uniform cutting meat on the table"}]}
{"id": 228074, "image": "COCO_train2014_000000228074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with his head down eating the grass in front of him near the giraffe\"."}], "task": "refering", "answer_template": "The \"a zebra with his head down eating the grass in front of him near the giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 262, 265, 266, 267, 285, 288, 289, 290], "bbox": [0.386453125, 0.5124791666666667, 0.630796875, 0.7661666666666667], "iscrowd": 0, "area": 9045.429100000003, "rle": {"size": [480, 640], "counts": "`ad34j>3M4@FSB>k=GoA<o==M2N3Me0[O9H7H3MRO_C_O`<`0fC[OZ<c0mCYOR<f0UDTOk;k0[DQOd;n0bDmN^;Q1iDkNV;T1[1O1O100O2O000O1000000O10000O1002N101N2N2N1O2N2N2N1O2N2O1O1O00100O1O00001O0001O01O0000001O01O000001O00000010O000001O000001O01O0000001O4MX1gN<D00O100O101N100O100O101O0O1bNQEkNo:T1_1000001N1000001N2O1O1N101O1O1O1001N101O001O000oB`N`<b1UCjNi<i1N3M2N2N3N1N2O2M00N3N1O2N1O1O2M2O2N1O1O2N1N3N1POP1K6J9G:FYZ^3"}, "label": "a zebra with his head down eating the grass in front of him near the giraffe"}]}
{"id": 228074, "image": "COCO_train2014_000000228074.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind a giraffe eating grass\"."}], "task": "refering", "answer_template": "The \"a zebra behind a giraffe eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 262, 265, 266, 267, 285, 288, 289, 290], "bbox": [0.386453125, 0.5124791666666667, 0.630796875, 0.7661666666666667], "iscrowd": 0, "area": 9045.429100000003, "rle": {"size": [480, 640], "counts": "`ad34j>3M4@FSB>k=GoA<o==M2N3Me0[O9H7H3MRO_C_O`<`0fC[OZ<c0mCYOR<f0UDTOk;k0[DQOd;n0bDmN^;Q1iDkNV;T1[1O1O100O2O000O1000000O10000O1002N101N2N2N1O2N2N2N1O2N2O1O1O00100O1O00001O0001O01O0000001O01O000001O00000010O000001O000001O01O0000001O4MX1gN<D00O100O101N100O100O101O0O1bNQEkNo:T1_1000001N1000001N2O1O1N101O1O1O1001N101O001O000oB`N`<b1UCjNi<i1N3M2N2N3N1N2O2M00N3N1O2N1O1O2M2O2N1O1O2N1N3N1POP1K6J9G:FYZ^3"}, "label": "a zebra behind a giraffe eating grass"}]}
{"id": 187119, "image": "COCO_train2014_000000187119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver car parked behind a black car\"."}], "task": "refering", "answer_template": "The \"a silver car parked behind a black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 204, 205], "bbox": [0.716296875, 0.22247916666666667, 1.0, 0.494375], "iscrowd": 0, "area": 17148.73415, "rle": {"size": [480, 640], "counts": "fkf6b0Z>5K4L5L4K5K5M3O1O1N2O1O0O2O1O1N2O1O1N2O001O1N2O1O1N2O001N2O00O00100O010O1O10O01O10O01O10O01O10O10000O101O000O10001N10000O101O0O100O101N1O1O100O2N100O1O101N1O1O100O2N100O1O101N1O1O100O2N100O1O01O0000001O00001O00001O00ZDnL_;]3L5L4L3M4L00001O00000000000000001O000000000000000000O100O100O100O100O100O100O100O100O1O1N2O1N2N2000000000000000000000000O10000000000000000000000hMmN"}, "label": "a silver car parked behind a black car"}]}
{"id": 187119, "image": "COCO_train2014_000000187119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the light colored car directly behind the black car\"."}], "task": "refering", "answer_template": "The \"the light colored car directly behind the black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 204, 205], "bbox": [0.716296875, 0.22247916666666667, 1.0, 0.494375], "iscrowd": 0, "area": 17148.73415, "rle": {"size": [480, 640], "counts": "fkf6b0Z>5K4L5L4K5K5M3O1O1N2O1O0O2O1O1N2O1O1N2O001O1N2O1O1N2O001N2O00O00100O010O1O10O01O10O01O10O01O10O10000O101O000O10001N10000O101O0O100O101N1O1O100O2N100O1O101N1O1O100O2N100O1O101N1O1O100O2N100O1O01O0000001O00001O00001O00ZDnL_;]3L5L4L3M4L00001O00000000000000001O000000000000000000O100O100O100O100O100O100O100O100O1O1N2O1N2N2000000000000000000000000O10000000000000000000000hMmN"}, "label": "the light colored car directly behind the black car"}]}
{"id": 187119, "image": "COCO_train2014_000000187119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car beside bike\"."}], "task": "refering", "answer_template": "The \"black car beside bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 216, 217, 234, 235], "bbox": [0.033937499999999995, 0.2510833333333333, 0.61584375, 0.5980624999999999], "iscrowd": 0, "area": 41031.93550000001, "rle": {"size": [480, 640], "counts": "j`:>a>6I8I4L4L4L3L5L4L2M3M2M4L3N2M4N1O2N1O101N1O2N101N1O101N101N10001N101N100O2O001N100O2O00O0100O10000O100000000O10000000000O1000000000000O101O00O1O1O1N2O1O1N2O11O2N1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2N2O1N2N1O100O1O1O100O1O1O010O1O1O100O1O1O100O1O1O100O1O1O010OgM^E6a:KcE1\\:0hELX:3mEIR:8RFDm9<UFCj9>XF@h9`0ZF^Oe9b0]F]Ob9d0`FZO`9e0cFXO]9i0dFVO[9k0gFSOX9m0kFQOU9P1kFoNT9S1mFkNR9X1mFdNU9^1jF`NX9a1hF[NZ9g1fFVN[9l1dFQN^9Q2bFlM`9U2`FcMf9_2ZFYMl9i2T12O1N2N2N2O1N2O1O1O1O1O1O1O00001O000000000O1000001O000000000000001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O1O1O2N1O1O2N1O1O2N00O1O10000000000000000000000000000000000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O1000000O10000O101N100O2O000O101N10001O0O10001N10001O0000001O00001O00001O0000001O00001O00100O2N1O2N1O2N1O101N1O2N1O2N1O2O0O2N101N101N1O2O0M4M1N2O1O1O4M6H9G9G6I9G=B_Sc3"}, "label": "black car beside bike"}]}
{"id": 187120, "image": "COCO_train2014_000000187120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrots on a white plate with cake\"."}], "task": "refering", "answer_template": "The \"carrots on a white plate with cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 100, 101, 102, 103, 118, 119, 120, 121, 122, 137, 138, 139, 156], "bbox": [0.55788, 0.3482282282282282, 0.7913600000000001, 0.6847147147147147], "iscrowd": 0, "area": 7545.529250000001, "rle": {"size": [333, 500], "counts": "clj22Y:3N2M3M3N2\\OCVG>d8H[G:S8_OgG=55S8_ObGc09OS89lGHR8:mGGR8:mGGQ8<mGER8<mGEQ8=nGEP8=nGDQ8=nGDQ8=nGDQ8=oGCP8^1O1O1O2N1O100O1O1O1O1O2O0O100O100O10000O2O0O10000O100O101O0O100O01O1O001O010O1O001O001aN`HOa71`HM`74aHJ`76aHH_78cHF^7:cHC_7=bHA^7`0cH]O_7c0bH[O^7f0cHWO_7i0bHUO_7k0bHRO_7o0bHoN_7Q1bHlN`7T1aHjN_7W1bHfN`7Z1aHdN`7\\1aHaN`7_1bH_N_7a1f010000O01O10O01O001O1O001O10O01O001O1O001O10OO2N2N101N1O3M2N3M3M2E<DSeQ1"}, "label": "carrots on a white plate with cake"}]}
{"id": 187120, "image": "COCO_train2014_000000187120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two nearly parallel carrots\"."}], "task": "refering", "answer_template": "The \"the two nearly parallel carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 100, 101, 102, 103, 118, 119, 120, 121, 122, 137, 138, 139, 156], "bbox": [0.55788, 0.3482282282282282, 0.7913600000000001, 0.6847147147147147], "iscrowd": 0, "area": 7545.529250000001, "rle": {"size": [333, 500], "counts": "clj22Y:3N2M3M3N2\\OCVG>d8H[G:S8_OgG=55S8_ObGc09OS89lGHR8:mGGR8:mGGQ8<mGER8<mGEQ8=nGEP8=nGDQ8=nGDQ8=nGDQ8=oGCP8^1O1O1O2N1O100O1O1O1O1O2O0O100O100O10000O2O0O10000O100O101O0O100O01O1O001O010O1O001O001aN`HOa71`HM`74aHJ`76aHH_78cHF^7:cHC_7=bHA^7`0cH]O_7c0bH[O^7f0cHWO_7i0bHUO_7k0bHRO_7o0bHoN_7Q1bHlN`7T1aHjN_7W1bHfN`7Z1aHdN`7\\1aHaN`7_1bH_N_7a1f010000O01O10O01O001O1O001O10O01O001O1O001O10OO2N2N101N1O3M2N3M3M2E<DSeQ1"}, "label": "the two nearly parallel carrots"}]}
{"id": 522996, "image": "COCO_train2014_000000522996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark grey dog on a light grey round bed wearing a red collar\"."}], "task": "refering", "answer_template": "The \"a dark grey dog on a light grey round bed wearing a red collar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 43, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 85, 86, 87, 88, 89, 90, 91, 93, 94, 95, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 132, 133, 135, 136, 137, 148, 149, 150, 151, 152, 164, 165, 166, 167, 180, 181, 182, 183], "bbox": [0.2660888888888889, 0.1953550295857988, 0.9563777777777778, 0.986834319526627], "iscrowd": 0, "area": 30983.92370000002, "rle": {"size": [338, 450], "counts": "ZmW14[:5K5J6G9H8G9E:A>I7K6]JhMd2^2\\MbM_2d2`M]MY2i2gMWMT2n2lMSMQ2o2oMQMo1P3RNQMk1P3VNQMg1Q3YNPMc1R3^NnL`1S3aNnL\\1T3dNmLY1T3hNmLT1V3lNiLS1X3nNgLR1Y3oNgLo0[3QOdLn0]3TOaLk0`3VO_Li0c3WO\\Lh0e3XO\\Lg0m1iLcNU3_O1m1^MjMi29Hk1Y1VNdNi1^1\\NXNf1l1]NjMg1Y2\\N]Mg1f2\\NbLW2a3lMXLV2i3mMQLT2R4oMgKR2[4QN_KQ2c4l13N2M3N2O100O1O1O100O1O100OeLeKa0Z4\\OmKb0Q4\\OULb0j3ZO]Ld0a3ZOdLf0Y3YOjLg0T3YOmLh0R3VOPMl0m2TOTMn0j2QOWMP1h2nN[MQ1e2nN\\MS1c2lN^MT1a2kNaMU1_2jNbMV1^2iNdMV1\\2iNeMW1[2hNfMX1Z2fNhMZ1X2eNiM[1W2dNhM^1X2`NiMa1W2\\NkMe1U2YNlMh1T2UNnMl1R2SNnMn1Q2RNoMo1R2oMmMS2T2kMlMV2U2hMXMl2i2RMVMP3k2nLTMU3l2iLRMZ3o2dLPM^3Q3_LoLc3Q3[LoLg3R3VLnLl3S3QLnLP4S3mKmLU4T3hKmLY4T3eKkL]4V3`KkLa4V3\\KkLd4`42O0O2O001N101N101O0O2O1N2O001N2OPKoJ`4P5^KUKa4j4]KYKc4f4]K\\Kb4b4_K_K`4a4`KaK_4]4cKdK\\4[4dKfK\\4Y4dKhK[4_3^LaLc3_3\\LbLd3]3\\LdLd3\\3\\LcLe3\\3[LeLe3[3ZLfLf3X3[LhLe3X3ZLiLg3V3YLjLh3U3WLmLi3R3WLnLj3P3VLQMk3n2ULRMl3m2SLTMm3l2SLTMn3k2QLVMP4i2oKXMR4g2nKYMS4g2iK]MW4b2fK`M\\4_2aKbMa4^2]KbMf4^2VKbMn4]2oJdMT5[2iJgMY5X2eJiM]5d32O1N2N2N2O1N2N2O1N2N2N2L4M3L4M3L4M2N3L4M3M3O1O1N2O1N2O1N2O1O2M2O2O0O2N2O0O2N101N2N101N1O2O1N1O2O0O2N2O0O2N101N2N1O2O1N1O2N101N2N2N3N1N2N2N2N2N2K5K4L5N2O10O10000O2O001O010O010O01O100O1O10O0100O001O0O2O0O2O0O101N100O10O0010O010O01O010O01O00001O00001O0000000000000000000000000000O1000O10000O10000O10000O10000O101N101O1N10b^6"}, "label": "a dark grey dog on a light grey round bed wearing a red collar"}]}
{"id": 522996, "image": "COCO_train2014_000000522996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog lying on blue dog bed\"."}], "task": "refering", "answer_template": "The \"dog lying on blue dog bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 43, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 85, 86, 87, 88, 89, 90, 91, 93, 94, 95, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 132, 133, 135, 136, 137, 148, 149, 150, 151, 152, 164, 165, 166, 167, 180, 181, 182, 183], "bbox": [0.2660888888888889, 0.1953550295857988, 0.9563777777777778, 0.986834319526627], "iscrowd": 0, "area": 30983.92370000002, "rle": {"size": [338, 450], "counts": "ZmW14[:5K5J6G9H8G9E:A>I7K6]JhMd2^2\\MbM_2d2`M]MY2i2gMWMT2n2lMSMQ2o2oMQMo1P3RNQMk1P3VNQMg1Q3YNPMc1R3^NnL`1S3aNnL\\1T3dNmLY1T3hNmLT1V3lNiLS1X3nNgLR1Y3oNgLo0[3QOdLn0]3TOaLk0`3VO_Li0c3WO\\Lh0e3XO\\Lg0m1iLcNU3_O1m1^MjMi29Hk1Y1VNdNi1^1\\NXNf1l1]NjMg1Y2\\N]Mg1f2\\NbLW2a3lMXLV2i3mMQLT2R4oMgKR2[4QN_KQ2c4l13N2M3N2O100O1O1O100O1O100OeLeKa0Z4\\OmKb0Q4\\OULb0j3ZO]Ld0a3ZOdLf0Y3YOjLg0T3YOmLh0R3VOPMl0m2TOTMn0j2QOWMP1h2nN[MQ1e2nN\\MS1c2lN^MT1a2kNaMU1_2jNbMV1^2iNdMV1\\2iNeMW1[2hNfMX1Z2fNhMZ1X2eNiM[1W2dNhM^1X2`NiMa1W2\\NkMe1U2YNlMh1T2UNnMl1R2SNnMn1Q2RNoMo1R2oMmMS2T2kMlMV2U2hMXMl2i2RMVMP3k2nLTMU3l2iLRMZ3o2dLPM^3Q3_LoLc3Q3[LoLg3R3VLnLl3S3QLnLP4S3mKmLU4T3hKmLY4T3eKkL]4V3`KkLa4V3\\KkLd4`42O0O2O001N101N101O0O2O1N2O001N2OPKoJ`4P5^KUKa4j4]KYKc4f4]K\\Kb4b4_K_K`4a4`KaK_4]4cKdK\\4[4dKfK\\4Y4dKhK[4_3^LaLc3_3\\LbLd3]3\\LdLd3\\3\\LcLe3\\3[LeLe3[3ZLfLf3X3[LhLe3X3ZLiLg3V3YLjLh3U3WLmLi3R3WLnLj3P3VLQMk3n2ULRMl3m2SLTMm3l2SLTMn3k2QLVMP4i2oKXMR4g2nKYMS4g2iK]MW4b2fK`M\\4_2aKbMa4^2]KbMf4^2VKbMn4]2oJdMT5[2iJgMY5X2eJiM]5d32O1N2N2N2O1N2N2O1N2N2N2L4M3L4M3L4M2N3L4M3M3O1O1N2O1N2O1N2O1O2M2O2O0O2N2O0O2N101N2N101N1O2O1N1O2O0O2N2O0O2N101N2N1O2O1N1O2N101N2N2N3N1N2N2N2N2N2K5K4L5N2O10O10000O2O001O010O010O01O100O1O10O0100O001O0O2O0O2O0O101N100O10O0010O010O01O010O01O00001O00001O0000000000000000000000000000O1000O10000O10000O10000O10000O101N101O1N10b^6"}, "label": "dog lying on blue dog bed"}]}
{"id": 367357, "image": "COCO_train2014_000000367357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in white shirt\"."}], "task": "refering", "answer_template": "The \"a girl in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379, 395, 396, 397, 398, 399, 400, 401, 402, 418, 419, 420, 421, 422, 423, 424, 425, 441, 442, 443, 444, 445, 446, 447, 448, 464, 465, 466, 467, 468, 469, 470, 471], "bbox": [0.175203125, 0.37993265993265996, 0.51515625, 0.9866666666666666], "iscrowd": 0, "area": 59798.36885, "rle": {"size": [594, 640], "counts": "fSR24\\b02N2N2mJGjG<S8MeG5X84`GN]8:\\GHa8a0WGBe8g0RG\\Ok8m0mFUOP9T1hFnNV9Z1bFhN[9a1]FaN`9h1XFZNe9o1TFRNi9V2PFlMm9\\2lEfMQ:c2hE]MV:k2cEWMZ:Q3_EQM^:X3[EiLc:^3VEdLg:d3RE^Lk:j3nDXLo:Q4jDPLS;i5M3M3M3M3M3N2M3M3M3M3N2M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M4L3N2M3N2M3N2M3N2M3N2M3N2O10000O1000000O1000000O10000O1000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10001O00000O10000000000001O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1hGXGd6i8PIeGm6\\8fHSHW7g9M3M3M3N2M3M3M3M3M2N3M8H9G9G8H9G9G8H9G9G8H9G9G8H9G9G9G:F9G9G:F9G9G:F9G]Xc5"}, "label": "a girl in white shirt"}]}
{"id": 367357, "image": "COCO_train2014_000000367357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing a white coat\"."}], "task": "refering", "answer_template": "The \"woman wearing a white coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379, 395, 396, 397, 398, 399, 400, 401, 402, 418, 419, 420, 421, 422, 423, 424, 425, 441, 442, 443, 444, 445, 446, 447, 448, 464, 465, 466, 467, 468, 469, 470, 471], "bbox": [0.175203125, 0.37993265993265996, 0.51515625, 0.9866666666666666], "iscrowd": 0, "area": 59798.36885, "rle": {"size": [594, 640], "counts": "fSR24\\b02N2N2mJGjG<S8MeG5X84`GN]8:\\GHa8a0WGBe8g0RG\\Ok8m0mFUOP9T1hFnNV9Z1bFhN[9a1]FaN`9h1XFZNe9o1TFRNi9V2PFlMm9\\2lEfMQ:c2hE]MV:k2cEWMZ:Q3_EQM^:X3[EiLc:^3VEdLg:d3RE^Lk:j3nDXLo:Q4jDPLS;i5M3M3M3M3M3N2M3M3M3M3N2M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M4L3N2M3N2M3N2M3N2M3N2M3N2O10000O1000000O1000000O10000O1000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10001O00000O10000000000001O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1hGXGd6i8PIeGm6\\8fHSHW7g9M3M3M3N2M3M3M3M3M2N3M8H9G9G8H9G9G8H9G9G8H9G9G8H9G9G9G:F9G9G:F9G9G:F9G]Xc5"}, "label": "woman wearing a white coat"}]}
{"id": 367357, "image": "COCO_train2014_000000367357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bag in woman ' s lap\"."}], "task": "refering", "answer_template": "The \"brown bag in woman ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [363, 364, 384, 385, 386, 387, 388, 406, 407, 408, 409, 410, 411, 428, 429, 430, 431, 432, 433, 451, 452, 453, 454, 455, 456, 474, 475, 476, 477, 479, 480, 481], "bbox": [0.640015625, 0.7232828282828283, 0.938125, 0.9997138047138046], "iscrowd": 0, "area": 11752.559700000003, "rle": {"size": [594, 640], "counts": "[[^7`0ja0`0C:G8H8K4K5L5K9H3M000O100000000000000000001O000O100O1O10000001O000O102MaN]MQCc2o<]MQCb2P=_MoBa2Q=`MnB_2S=aMmB_2S=bMkB_2U=bMjB]2W=dMhB\\2X=dMhB\\2X=eMgBZ2Z=gMeBY2[=gMeBY2[=gMeBX2\\=hMdBX2\\=iMcBW2]=iMcBW2]=iMcBV2^=jMbBV2^=jMbBV2^=kMaBU2`=jM`BU2a=kM`BT2`=lM`BT2`=mM_BS2`=nM`BQ2a=oM_BQ2a=oM_BQ2a=oM_BQ2a=oM_BP2b=PN^BP2b=oM_BQ2a=oM_BQ2a=oM_BP2b=PN^BP2b=PN^BP2a=QN_Bo1a=PN`Bo1a=QN_Bo1a=QN_Bo1a=QN_Bo1a=QN_Bn1b=RN^Bn1b=RN^Bn1b=RN^Bm1c=SN]Bm1c=TN\\Bl1c=UN]Bk1c=UN]Bj1d=VN\\Bj1d=VN\\Bj1d=VN\\Bi1e=XNZBh1f=XNZBh1f=XNZBV1lNmNk>MYBU1nNjNl>1VBT1oNfNQ?6oAS1ROeNP?8nAR1SOfNo>8nAQ1UOeNn>:mAP1VOfNm>:mAo0VOgNn>:lAo0VOfNo>;kAn0WOgNo>:jAn0XOgNo>:jAn0XOhNn>:jAn0WOhNS>0]B:9m0h=kNmA8;m0i=lNjA7<n0k=lNgA6>n0k=nNfA2`0o0l=PObA1b0o0m=QO_A0d0o0m=SO]ANf0o0n=SOZAOh0m0P>UOUAOm0j0n=YORANQ1h0n=[On@NV1e0m=^Oj@OZ1`0m=1TBNm=1TBNm=2i@UOR1i0U>2g@WOU1f0U>:kAEV>;jADW>;jADX>;hADY>;hADX>=gACZ>=fAB[>=gAAZ>?fA@Z>`0gA_OZ>a0gA^OY>b0jAZOV>f0mAWOT>i0oASOR>l0j1100WAoNP<P1PDPOQ<P1nCPOR<P1mCQOT<o0kCROU<n0jCROV<n0iCSOX<m0gCTOY<l0fCTOZ<l0fCUOZ<k0eCUO\\<k0cCVO]<i0cCXO\\<i0cCWO^<i0aCXO_<g0aCZO^<g0aCZO_<f0_C\\O`<e0_C\\Oa<c0_C\\Ob<e0]C[Od<e0[CZOg<e0YC[Oh<e0WC[Oj<e0UCZOn<d0RC\\Oo<d0PC\\OQ=c0oB\\OS=b0nB^OS=a0mB_OT=?mB@U=>lBBV=<jBDX=9iBFZ=7gBI[=5eBK]=2dBM_=1_B1c=L]B5e=H[B8h=FWB;o?0O2O001O10O2O0O10002M3M4Lfdf0"}, "label": "brown bag in woman ' s lap"}]}
{"id": 367357, "image": "COCO_train2014_000000367357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bag in a woman ' s lap\"."}], "task": "refering", "answer_template": "The \"a brown bag in a woman ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [363, 364, 384, 385, 386, 387, 388, 406, 407, 408, 409, 410, 411, 428, 429, 430, 431, 432, 433, 451, 452, 453, 454, 455, 456, 474, 475, 476, 477, 479, 480, 481], "bbox": [0.640015625, 0.7232828282828283, 0.938125, 0.9997138047138046], "iscrowd": 0, "area": 11752.559700000003, "rle": {"size": [594, 640], "counts": "[[^7`0ja0`0C:G8H8K4K5L5K9H3M000O100000000000000000001O000O100O1O10000001O000O102MaN]MQCc2o<]MQCb2P=_MoBa2Q=`MnB_2S=aMmB_2S=bMkB_2U=bMjB]2W=dMhB\\2X=dMhB\\2X=eMgBZ2Z=gMeBY2[=gMeBY2[=gMeBX2\\=hMdBX2\\=iMcBW2]=iMcBW2]=iMcBV2^=jMbBV2^=jMbBV2^=kMaBU2`=jM`BU2a=kM`BT2`=lM`BT2`=mM_BS2`=nM`BQ2a=oM_BQ2a=oM_BQ2a=oM_BQ2a=oM_BP2b=PN^BP2b=oM_BQ2a=oM_BQ2a=oM_BP2b=PN^BP2b=PN^BP2a=QN_Bo1a=PN`Bo1a=QN_Bo1a=QN_Bo1a=QN_Bo1a=QN_Bn1b=RN^Bn1b=RN^Bn1b=RN^Bm1c=SN]Bm1c=TN\\Bl1c=UN]Bk1c=UN]Bj1d=VN\\Bj1d=VN\\Bj1d=VN\\Bi1e=XNZBh1f=XNZBh1f=XNZBV1lNmNk>MYBU1nNjNl>1VBT1oNfNQ?6oAS1ROeNP?8nAR1SOfNo>8nAQ1UOeNn>:mAP1VOfNm>:mAo0VOgNn>:lAo0VOfNo>;kAn0WOgNo>:jAn0XOgNo>:jAn0XOhNn>:jAn0WOhNS>0]B:9m0h=kNmA8;m0i=lNjA7<n0k=lNgA6>n0k=nNfA2`0o0l=PObA1b0o0m=QO_A0d0o0m=SO]ANf0o0n=SOZAOh0m0P>UOUAOm0j0n=YORANQ1h0n=[On@NV1e0m=^Oj@OZ1`0m=1TBNm=1TBNm=2i@UOR1i0U>2g@WOU1f0U>:kAEV>;jADW>;jADX>;hADY>;hADX>=gACZ>=fAB[>=gAAZ>?fA@Z>`0gA_OZ>a0gA^OY>b0jAZOV>f0mAWOT>i0oASOR>l0j1100WAoNP<P1PDPOQ<P1nCPOR<P1mCQOT<o0kCROU<n0jCROV<n0iCSOX<m0gCTOY<l0fCTOZ<l0fCUOZ<k0eCUO\\<k0cCVO]<i0cCXO\\<i0cCWO^<i0aCXO_<g0aCZO^<g0aCZO_<f0_C\\O`<e0_C\\Oa<c0_C\\Ob<e0]C[Od<e0[CZOg<e0YC[Oh<e0WC[Oj<e0UCZOn<d0RC\\Oo<d0PC\\OQ=c0oB\\OS=b0nB^OS=a0mB_OT=?mB@U=>lBBV=<jBDX=9iBFZ=7gBI[=5eBK]=2dBM_=1_B1c=L]B5e=H[B8h=FWB;o?0O2O001O10O2O0O10002M3M4Lfdf0"}, "label": "a brown bag in a woman ' s lap"}]}
{"id": 367357, "image": "COCO_train2014_000000367357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man sitting above two women\"."}], "task": "refering", "answer_template": "The \"man sitting above two women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 213, 217, 218, 219, 220, 221, 235, 240, 241, 242, 243, 244, 257, 258, 263, 264, 265, 266, 267, 278, 279, 280, 281, 286, 287, 288, 289, 301, 302, 303, 310, 311, 312, 313, 325, 326, 333, 334, 335, 336, 348, 349, 356, 357, 358, 359, 371, 372, 380, 381, 382, 394, 395, 403, 404, 405, 417, 418, 426, 427, 428, 440, 441, 449, 450, 451, 464, 473, 474], "bbox": [0.106609375, 0.143013468013468, 0.656375, 0.9898653198653199], "iscrowd": 0, "area": 56040.66169999996, "rle": {"size": [594, 640], "counts": "giW11ab03L5L3M3L4M2N1N2O1O1N3N1O1N2O1O1N3N1O1N2O1N3a_OjN^?W1]@RO^?o0]@YO^?i0]@@^?a0]@H^?k1J6K5K4K9H;D;E;F;D;E7J6I7bDhIU:_6dEaI]:e6\\E\\Ic:k6WETIi:S7PEnHP;Z7O010O00O2O0O1aKmD`1T;\\NUE]1l:`N\\EZ1e:cNdEW1]:eNlET1U:iNTFQ1l9lN]Fm0e9oNdFk0\\9ROlFh0U9UOTGe0m8WO\\Gb0e8[OdG?]8]OlG<U8BUH5l7I]H0d7MeHK\\73nHER79WI_Oj6?_IZOa6e0hIROY6l0g5N2N2N2N2N2O1N2N2N2N2N2N2O0O001O00000010O00100O1O100O00100O1O100O1O100O1O100O1O100001O001O1O001O1O001O1O0101O1N2O001N2O1O0O2O1O1N101O1N2L3M4L4M2M4L4N20O10O1000O1000O10O100000O010000O1000O10O100000O01000000000O2O1O1O1O0O2O1O1O1O0O2O1O1O1O1O0O2O1O1O1O0O2O1O1O1gNhMVBX2g=RNRBo1j=\\NnAe1o=dNiA^1S>lNfAU1V>f1L3M4L4L4N2M2N3M3N2M3M2N3N2M3N2O1O001O1O1fD\\Jd9e5oEiJo9W5dEYKY:h4ZEfKd:W6M3M2N6J:F9G9G9G:F9G9F:G9G2N1O2N2N1O2N2N1O2N2N1O2NN2N2O1N2O1N2O1N2O1N2O1N2O1N2O12N1O2N1O2N2N1O2N1O2N7I9G9I8I6J7I6J7I6I7I8H5K4K6K5K5K4L3MN2N1O2XCmKm:U4oDZLd:h3XE\\Lg:e3UE]Lk:e3PE_LZ4SOW1`4[J`L\\4QOY1a4WJ`L`4POX1c4SJ`Lc4oNY1c4PJ`Lf4oNY1d4lI`Li4mN[1f4gI_Lm4mN[1f4dI`LP5kN[1Y6dNiI[1X6cNiI]1X6aNjI_1U6`NmI_1T6`NmI_1S6`NnI`1S6ZNSJe1m5UNZJj1g5oM_JQ2b5hMeJW2[5dMkJ[2V5^MQKa2o4ZMWKe2j4VM\\Kh2d4UMbKi2^4SMhKj2X4SMnKj2S4QMTLl2m3PMYLm2g3PM_L_OPMm0b6@dLTOcMo0i5IkLjNUNo0Q53oLaN\\N[1e41UMVNbNh1Z4N:2GJ=5CI?8ADb0<^OAf0>[O_Og0a0[OZOi0e0ZOUOa13aNIX:J6J6K4K6J6KV\\o3"}, "label": "man sitting above two women"}]}
{"id": 367357, "image": "COCO_train2014_000000367357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on the bench behind a woman\"."}], "task": "refering", "answer_template": "The \"a man sitting on the bench behind a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 213, 217, 218, 219, 220, 221, 235, 240, 241, 242, 243, 244, 257, 258, 263, 264, 265, 266, 267, 278, 279, 280, 281, 286, 287, 288, 289, 301, 302, 303, 310, 311, 312, 313, 325, 326, 333, 334, 335, 336, 348, 349, 356, 357, 358, 359, 371, 372, 380, 381, 382, 394, 395, 403, 404, 405, 417, 418, 426, 427, 428, 440, 441, 449, 450, 451, 464, 473, 474], "bbox": [0.106609375, 0.143013468013468, 0.656375, 0.9898653198653199], "iscrowd": 0, "area": 56040.66169999996, "rle": {"size": [594, 640], "counts": "giW11ab03L5L3M3L4M2N1N2O1O1N3N1O1N2O1O1N3N1O1N2O1N3a_OjN^?W1]@RO^?o0]@YO^?i0]@@^?a0]@H^?k1J6K5K4K9H;D;E;F;D;E7J6I7bDhIU:_6dEaI]:e6\\E\\Ic:k6WETIi:S7PEnHP;Z7O010O00O2O0O1aKmD`1T;\\NUE]1l:`N\\EZ1e:cNdEW1]:eNlET1U:iNTFQ1l9lN]Fm0e9oNdFk0\\9ROlFh0U9UOTGe0m8WO\\Gb0e8[OdG?]8]OlG<U8BUH5l7I]H0d7MeHK\\73nHER79WI_Oj6?_IZOa6e0hIROY6l0g5N2N2N2N2N2O1N2N2N2N2N2N2O0O001O00000010O00100O1O100O00100O1O100O1O100O1O100O1O100001O001O1O001O1O001O1O0101O1N2O001N2O1O0O2O1O1N101O1N2L3M4L4M2M4L4N20O10O1000O1000O10O100000O010000O1000O10O100000O01000000000O2O1O1O1O0O2O1O1O1O0O2O1O1O1O1O0O2O1O1O1O0O2O1O1O1gNhMVBX2g=RNRBo1j=\\NnAe1o=dNiA^1S>lNfAU1V>f1L3M4L4L4N2M2N3M3N2M3M2N3N2M3N2O1O001O1O1fD\\Jd9e5oEiJo9W5dEYKY:h4ZEfKd:W6M3M2N6J:F9G9G9G:F9G9F:G9G2N1O2N2N1O2N2N1O2N2N1O2NN2N2O1N2O1N2O1N2O1N2O1N2O1N2O12N1O2N1O2N2N1O2N1O2N7I9G9I8I6J7I6J7I6I7I8H5K4K6K5K5K4L3MN2N1O2XCmKm:U4oDZLd:h3XE\\Lg:e3UE]Lk:e3PE_LZ4SOW1`4[J`L\\4QOY1a4WJ`L`4POX1c4SJ`Lc4oNY1c4PJ`Lf4oNY1d4lI`Li4mN[1f4gI_Lm4mN[1f4dI`LP5kN[1Y6dNiI[1X6cNiI]1X6aNjI_1U6`NmI_1T6`NmI_1S6`NnI`1S6ZNSJe1m5UNZJj1g5oM_JQ2b5hMeJW2[5dMkJ[2V5^MQKa2o4ZMWKe2j4VM\\Kh2d4UMbKi2^4SMhKj2X4SMnKj2S4QMTLl2m3PMYLm2g3PM_L_OPMm0b6@dLTOcMo0i5IkLjNUNo0Q53oLaN\\N[1e41UMVNbNh1Z4N:2GJ=5CI?8ADb0<^OAf0>[O_Og0a0[OZOi0e0ZOUOa13aNIX:J6J6K4K6J6KV\\o3"}, "label": "a man sitting on the bench behind a woman"}]}
{"id": 170751, "image": "COCO_train2014_000000170751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing green and white shirt with red necktie\"."}], "task": "refering", "answer_template": "The \"a person wearing green and white shirt with red necktie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0198125, 0.0012412177985948478, 1.0, 0.9863934426229508], "iscrowd": 0, "area": 233087.36239999998, "rle": {"size": [427, 640], "counts": "ef51Y=101N1O101N101N1O101N101N1O100O100O1O1O1O100O1O1O1O100O1O1O1`LXO[Jj0d5OcI3\\6f0iH\\OV7^1oGdNP8U2WGlMi8[3O1O100001O0000001O0000001O00001O0000001O0000L4K5K5K5J6K5K5J6K5nNR1J6J6I7J6H8G8L5L4L4L3M4N2M3N1N3M3M3000O0100000O1000000O100O100O100O1O100O100O1O100O10_NmGjMR8S2SHlMl7Q2XHoMh7n1\\HQNc7l1bHSN^7k1dHUN[7j1hHUNW7k1kHTNU7j1mHVNR7j1PIUNP7i1SIVNl6j1UIVNk6h1XIWNg6i1[IVNd6i1`IUN`6j1bIUN]6j1gITNY6e1oIZNP6_1XJaNg5X1aJhN_5o0jJQOU5g0TKYOl4>]KBb46gKJX4NQL2d8000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003M>B=C=CdN"}, "label": "a person wearing green and white shirt with red necktie"}]}
{"id": 170751, "image": "COCO_train2014_000000170751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the torso of the man in the green checked shirt\"."}], "task": "refering", "answer_template": "The \"the torso of the man in the green checked shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0198125, 0.0012412177985948478, 1.0, 0.9863934426229508], "iscrowd": 0, "area": 233087.36239999998, "rle": {"size": [427, 640], "counts": "ef51Y=101N1O101N101N1O101N101N1O100O100O1O1O1O100O1O1O1O100O1O1O1`LXO[Jj0d5OcI3\\6f0iH\\OV7^1oGdNP8U2WGlMi8[3O1O100001O0000001O0000001O00001O0000001O0000L4K5K5K5J6K5K5J6K5nNR1J6J6I7J6H8G8L5L4L4L3M4N2M3N1N3M3M3000O0100000O1000000O100O100O100O1O100O100O1O100O10_NmGjMR8S2SHlMl7Q2XHoMh7n1\\HQNc7l1bHSN^7k1dHUN[7j1hHUNW7k1kHTNU7j1mHVNR7j1PIUNP7i1SIVNl6j1UIVNk6h1XIWNg6i1[IVNd6i1`IUN`6j1bIUN]6j1gITNY6e1oIZNP6_1XJaNg5X1aJhN_5o0jJQOU5g0TKYOl4>]KBb46gKJX4NQL2d8000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003M>B=C=CdN"}, "label": "the torso of the man in the green checked shirt"}]}
{"id": 334596, "image": "COCO_train2014_000000334596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small monitor to the left of a bigger monitor\"."}], "task": "refering", "answer_template": "The \"a small monitor to the left of a bigger monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122], "bbox": [0.59836, 0.2981681681681682, 0.79608, 0.5559759759759759], "iscrowd": 0, "area": 7856.029999999999, "rle": {"size": [333, 500], "counts": "]\\Q3h0T8a1F:000000001O00000000000000000000000000001O0001O000000000000000000000001O0000000000000000000000000001O01O00000000000000000000000000001O000000000000000001O000000000001O00000000000000000000000C?_OR1nNoQQ1"}, "label": "a small monitor to the left of a bigger monitor"}]}
{"id": 334596, "image": "COCO_train2014_000000334596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller monitor with a picture open\"."}], "task": "refering", "answer_template": "The \"the smaller monitor with a picture open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122], "bbox": [0.59836, 0.2981681681681682, 0.79608, 0.5559759759759759], "iscrowd": 0, "area": 7856.029999999999, "rle": {"size": [333, 500], "counts": "]\\Q3h0T8a1F:000000001O00000000000000000000000000001O0001O000000000000000000000001O0000000000000000000000000001O01O00000000000000000000000000001O000000000000000001O000000000001O00000000000000000000000C?_OR1nNoQQ1"}, "label": "the smaller monitor with a picture open"}]}
{"id": 334596, "image": "COCO_train2014_000000334596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left computer monitor\"."}], "task": "refering", "answer_template": "The \"the left computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 114, 115, 116, 117, 118], "bbox": [0.34688, 0.26303303303303305, 0.6001000000000001, 0.5520720720720721], "iscrowd": 0, "area": 11319.0288, "rle": {"size": [333, 500], "counts": "n]h13`9j0WOi0VOj0C=0000000O100000000000000000000000O1000000000000000000000O10000000000000000000000000000000000000000000000O0100000000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000000000O100000000000000000000000000000000000h0XOj0VOk0UOSlP2"}, "label": "the left computer monitor"}]}
{"id": 334596, "image": "COCO_train2014_000000334596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big monitor\"."}], "task": "refering", "answer_template": "The \"the big monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 114, 115, 116, 117, 118], "bbox": [0.34688, 0.26303303303303305, 0.6001000000000001, 0.5520720720720721], "iscrowd": 0, "area": 11319.0288, "rle": {"size": [333, 500], "counts": "n]h13`9j0WOi0VOj0C=0000000O100000000000000000000000O1000000000000000000000O10000000000000000000000000000000000000000000000O0100000000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000000000O100000000000000000000000000000000000h0XOj0VOk0UOSlP2"}, "label": "the big monitor"}]}
{"id": 72454, "image": "COCO_train2014_000000072454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skiing away from the photographer\"."}], "task": "refering", "answer_template": "The \"a man skiing away from the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 86, 87, 88, 109, 110, 111, 133, 134, 156, 157, 158, 180, 181], "bbox": [0.731984375, 0.154275, 0.891046875, 0.5169], "iscrowd": 0, "area": 5559.539500000001, "rle": {"size": [400, 640], "counts": "cZg52\\<2N3N1N2N3N1N2N200O0000000000010O00001O00010O001O00001O0SO^OnEa0R:GfE8Z:0_E0m0@W8f0gFJP1EV8g0dFCT1LT8U1jGPOT8o0iGWOU8g0iG_OU8`2N2N2N2N2N2N2N2N2N2N2O03N2M3M1O00O10\\NmGiNT8Z1kGcNV8_1jG^NW8e1gGYNZ8i1fGTN[8m1fGPN\\8o1eGoM\\8R2eGkM\\8U2fGhM\\8W2fGfM[8Z2fGdM[8]2fG_M]8`2j000010O0001O0ROQF[On9>\\F_Oe98eFF[9D]G:gNCZ;h02N100O2N100O100O2N100O1O2O0O100N3M2N3N2M3M3MjTk0"}, "label": "a man skiing away from the photographer"}]}
{"id": 72454, "image": "COCO_train2014_000000072454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skiier with poles horizontal\"."}], "task": "refering", "answer_template": "The \"skiier with poles horizontal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 86, 87, 88, 109, 110, 111, 133, 134, 156, 157, 158, 180, 181], "bbox": [0.731984375, 0.154275, 0.891046875, 0.5169], "iscrowd": 0, "area": 5559.539500000001, "rle": {"size": [400, 640], "counts": "cZg52\\<2N3N1N2N3N1N2N200O0000000000010O00001O00010O001O00001O0SO^OnEa0R:GfE8Z:0_E0m0@W8f0gFJP1EV8g0dFCT1LT8U1jGPOT8o0iGWOU8g0iG_OU8`2N2N2N2N2N2N2N2N2N2N2O03N2M3M1O00O10\\NmGiNT8Z1kGcNV8_1jG^NW8e1gGYNZ8i1fGTN[8m1fGPN\\8o1eGoM\\8R2eGkM\\8U2fGhM\\8W2fGfM[8Z2fGdM[8]2fG_M]8`2j000010O0001O0ROQF[On9>\\F_Oe98eFF[9D]G:gNCZ;h02N100O2N100O100O2N100O1O2O0O100N3M2N3N2M3M3MjTk0"}, "label": "skiier with poles horizontal"}]}
{"id": 72454, "image": "COCO_train2014_000000072454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis at bottom\"."}], "task": "refering", "answer_template": "The \"skis at bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 263, 265, 266, 267, 268, 285, 286, 287, 288, 289, 307, 308, 309, 310], "bbox": [0.35154687500000004, 0.7234, 0.7042187500000001, 0.9913500000000001], "iscrowd": 0, "area": 5551.698649999996, "rle": {"size": [400, 640], "counts": "iXh21^<1O1O2G8O100O2O0O10000O100O100O1000O0100O10000O100O010O10000O100O010O10O10O10O0100O01000O010O1000O0100SOXOQFh0o9^OkEc0U:AfE`0Y:CeE<\\:FaE;_:G^E:c:F[E:f:HWE9j:f00O101O0O2O1O1N101O1N2O1O1O1N2O1N2O1O2M2O1O0O01000O100O10000O10000O100O10O10O10000O10000O100O10000O10O10O10000O10O010000O10O10O100O01000O[OEWE;i:GTE:k:HTE7m:JRE6m:LQE5o:LPE3Q;NnD2Q;NoD2R;MoD3P;MQE3o:MPE3P;MQE3o:MQE3o:MPE3P;MQE3o:MQE2o:OQE1o:NQE3n:NRE1o:OQE1n:ORE2n:NRE1n:0RE0n:ORE1n:0RE0n:0RE0m:1REOo:0RE0m:1SENn:2RENm:2SEOm:1SENm:3SEMm:2SEOl:2TEMm:2TENk:3TEMm:2TENk:2VENi:3VEMj:3WEMh:3YEMf:4YELf:5[EKd:5]EJc:7\\EJc:6^EJa:6`EI`:8_EI`:7aEI^:7cEH]:9bEH^:7cEH^:7bEJ]:6dEJ\\:6dEI]:6cEK\\:5eEK\\:4cEL^:3cEM^:1cEN^:2aEO_:0bE0_:NaE2`:MaE3`:KaE5_:JaE6`:HbE8[;O0100O01000O001O001O1O0O2O1O010O01000O02N1O1O1N3N1O2M3M3Mi`Y2"}, "label": "skis at bottom"}]}
{"id": 72454, "image": "COCO_train2014_000000072454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis without a person attached\"."}], "task": "refering", "answer_template": "The \"the skis without a person attached\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 263, 265, 266, 267, 268, 285, 286, 287, 288, 289, 307, 308, 309, 310], "bbox": [0.35154687500000004, 0.7234, 0.7042187500000001, 0.9913500000000001], "iscrowd": 0, "area": 5551.698649999996, "rle": {"size": [400, 640], "counts": "iXh21^<1O1O2G8O100O2O0O10000O100O100O1000O0100O10000O100O010O10000O100O010O10O10O10O0100O01000O010O1000O0100SOXOQFh0o9^OkEc0U:AfE`0Y:CeE<\\:FaE;_:G^E:c:F[E:f:HWE9j:f00O101O0O2O1O1N101O1N2O1O1O1N2O1N2O1O2M2O1O0O01000O100O10000O10000O100O10O10O10000O10000O100O10000O10O10O10000O10O010000O10O10O100O01000O[OEWE;i:GTE:k:HTE7m:JRE6m:LQE5o:LPE3Q;NnD2Q;NoD2R;MoD3P;MQE3o:MPE3P;MQE3o:MQE3o:MPE3P;MQE3o:MQE2o:OQE1o:NQE3n:NRE1o:OQE1n:ORE2n:NRE1n:0RE0n:ORE1n:0RE0n:0RE0m:1REOo:0RE0m:1SENn:2RENm:2SEOm:1SENm:3SEMm:2SEOl:2TEMm:2TENk:3TEMm:2TENk:2VENi:3VEMj:3WEMh:3YEMf:4YELf:5[EKd:5]EJc:7\\EJc:6^EJa:6`EI`:8_EI`:7aEI^:7cEH]:9bEH^:7cEH^:7bEJ]:6dEJ\\:6dEI]:6cEK\\:5eEK\\:4cEL^:3cEM^:1cEN^:2aEO_:0bE0_:NaE2`:MaE3`:KaE5_:JaE6`:HbE8[;O0100O01000O001O001O1O0O2O1O010O01000O02N1O1O1N3N1O2M3M3Mi`Y2"}, "label": "the skis without a person attached"}]}
{"id": 72454, "image": "COCO_train2014_000000072454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in dark clothes skiing\"."}], "task": "refering", "answer_template": "The \"a person in dark clothes skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 118, 119, 120, 121, 122, 142, 143, 144, 145], "bbox": [0.098375, 0.024325, 0.32692187500000003, 0.502475], "iscrowd": 0, "area": 12491.754000000003, "rle": {"size": [400, 640], "counts": "Rfh06W<:F7I5J`0B8XE]NV:e1eEaNX:n1L4M1O2N2M2O2N1O1O001O1O@XFZNg9f1[FXNe9h1]FWNb9n0^GhNoNNc9Z1^1N101O0004M2ZFiNn7[1SGaN88c8Y1nFgN96b8Y1PGeN;5Y8D\\GT28]NX8Q2dGSNX8R2dGQNT8QOmGU4Q86N3N1N2O2M2O1O1N3N1N2O1N2O1O1000000O10000000000SInJZ6S5cIPK]6O]I[3LiM9mNj6W3PIlM4nNm6S3RIoMOoNP7n2UIUNGmNW7j2VIbNk6]1VIbNj6`1UI_Nk6c1TI\\Nk6g1TIXNk6k1TITNl6n1SIQNl6R2TIkMm6W2RIhMR7V2mHiMV7V2iHiMW7Y2hHfMY7[2fHcM\\7^2cHaM]7`2dH^M]7c2bH\\M^7e2bHZM_7f2aHXMa7g2`HXM`7i2`HVMa7j2`HTMa7l2_HSMc7k2^HSMe7l2[HSMg7k2ZHTMh7j2YHTMk75_HH@j07XOm71fHA\\OU11WOQ8NUIi0lNWOQ8LWIk0iNWOT8IWIn0fNXOT:I\\Ei0a0]OW:d0jEYOX:g0iETOZ:m0d01O1O1O10O1000O10O10O100000O0100000O13MO1O001O1O1O1O1O1N2O1N2O1O1N2O1N2O1O1N2N2O1N2N2N2O1NQVX5"}, "label": "a person in dark clothes skiing"}]}
{"id": 72454, "image": "COCO_train2014_000000072454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier on the top left\"."}], "task": "refering", "answer_template": "The \"the skier on the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 118, 119, 120, 121, 122, 142, 143, 144, 145], "bbox": [0.098375, 0.024325, 0.32692187500000003, 0.502475], "iscrowd": 0, "area": 12491.754000000003, "rle": {"size": [400, 640], "counts": "Rfh06W<:F7I5J`0B8XE]NV:e1eEaNX:n1L4M1O2N2M2O2N1O1O001O1O@XFZNg9f1[FXNe9h1]FWNb9n0^GhNoNNc9Z1^1N101O0004M2ZFiNn7[1SGaN88c8Y1nFgN96b8Y1PGeN;5Y8D\\GT28]NX8Q2dGSNX8R2dGQNT8QOmGU4Q86N3N1N2O2M2O1O1N3N1N2O1N2O1O1000000O10000000000SInJZ6S5cIPK]6O]I[3LiM9mNj6W3PIlM4nNm6S3RIoMOoNP7n2UIUNGmNW7j2VIbNk6]1VIbNj6`1UI_Nk6c1TI\\Nk6g1TIXNk6k1TITNl6n1SIQNl6R2TIkMm6W2RIhMR7V2mHiMV7V2iHiMW7Y2hHfMY7[2fHcM\\7^2cHaM]7`2dH^M]7c2bH\\M^7e2bHZM_7f2aHXMa7g2`HXM`7i2`HVMa7j2`HTMa7l2_HSMc7k2^HSMe7l2[HSMg7k2ZHTMh7j2YHTMk75_HH@j07XOm71fHA\\OU11WOQ8NUIi0lNWOQ8LWIk0iNWOT8IWIn0fNXOT:I\\Ei0a0]OW:d0jEYOX:g0iETOZ:m0d01O1O1O10O1000O10O10O100000O0100000O13MO1O001O1O1O1O1O1N2O1N2O1O1N2O1N2O1O1N2N2O1N2N2N2O1NQVX5"}, "label": "the skier on the top left"}]}
{"id": 350984, "image": "COCO_train2014_000000350984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan colored teddy bear in the girl ' s right arm\"."}], "task": "refering", "answer_template": "The \"tan colored teddy bear in the girl ' s right arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 166, 167, 184, 185, 200, 201, 217, 218, 219, 220, 221], "bbox": [0.09612000000000001, 0.67472, 0.33488, 1.0], "iscrowd": 0, "area": 5495.8007, "rle": {"size": [375, 500], "counts": "mma0:Z;3M3M3J6O1O1O1O100O1O10000O1O100O1O1O100O100O100O10000O1000000001O0000000000O1M3N2O1O1O1O1O1O1001O002N1O1O1O1O1O2N1O2N1O2NZNWO`Hg0S7FnH9m6MSI2l6OUI0i63VIMi64XIKh66XIIh67YIHf6:YIFg6;YIDg6>WIBh6a0VI_Oj6c0UI\\Ok6f0SIZOl6i0SIVOm6n0oHROP7P1PIoNP7Q1PIoNo6R1QInNo6Q1SInNl6S1TImNl6S1TImNk6S1VImNj6S1WIlNh6T1YIlNg6T1ZIkNe6U1\\IkNd6U1]IjNc6U1_IjNa6V1_IdNeNGk7e1bIbNe6]1\\IcNd6]1]IaNd6^1]IbNb6_1_I_Nb6`1`I^Na6b1_I[Nd6d1i1M2M3N2O2M2N2N3N1O1O2N3M2N2N3M2N2M5J5KPfi3"}, "label": "tan colored teddy bear in the girl ' s right arm"}]}
{"id": 350984, "image": "COCO_train2014_000000350984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small , offwhite stuffed bear\"."}], "task": "refering", "answer_template": "The \"a small , offwhite stuffed bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 166, 167, 184, 185, 200, 201, 217, 218, 219, 220, 221], "bbox": [0.09612000000000001, 0.67472, 0.33488, 1.0], "iscrowd": 0, "area": 5495.8007, "rle": {"size": [375, 500], "counts": "mma0:Z;3M3M3J6O1O1O1O100O1O10000O1O100O1O1O100O100O100O10000O1000000001O0000000000O1M3N2O1O1O1O1O1O1001O002N1O1O1O1O1O2N1O2N1O2NZNWO`Hg0S7FnH9m6MSI2l6OUI0i63VIMi64XIKh66XIIh67YIHf6:YIFg6;YIDg6>WIBh6a0VI_Oj6c0UI\\Ok6f0SIZOl6i0SIVOm6n0oHROP7P1PIoNP7Q1PIoNo6R1QInNo6Q1SInNl6S1TImNl6S1TImNk6S1VImNj6S1WIlNh6T1YIlNg6T1ZIkNe6U1\\IkNd6U1]IjNc6U1_IjNa6V1_IdNeNGk7e1bIbNe6]1\\IcNd6]1]IaNd6^1]IbNb6_1_I_Nb6`1`I^Na6b1_I[Nd6d1i1M2M3N2O2M2N2N3N1O1O2N3M2N2N3M2N2M5J5KPfi3"}, "label": "a small , offwhite stuffed bear"}]}
{"id": 350984, "image": "COCO_train2014_000000350984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear in green shirt\"."}], "task": "refering", "answer_template": "The \"bear in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 226], "bbox": [0.26846, 0.6266133333333334, 0.6325599999999999, 1.0], "iscrowd": 0, "area": 15298.238650000005, "rle": {"size": [375, 500], "counts": "gia11e;1O1O100O1O1O1O100O1O1O1O10000000000000000001O0000000000QNG`H9\\72]HN^79`HG\\7?bHAY7e0gHZOT7m0jHSOR7S1lHmNT7U1jHkNV7V1iHjNW7X1gHhNX7[1fHeNZ7]1dHcN\\7^1kG\\N86l7^1mG^N54n7^1lG_N72m7_1lG_N72m7^1mG`N62l7_1mGaN51n7^1mGaN32Q8\\1mGbN13Q8\\1mGbN13R8[1mGcNO3T8Y1nGdNM4U8X1mGeNM4U8W1oGeNK5V8V1oGfNI5X8U1nGgNI5Y8S1oGhNG6Y8S1oGhNG6Z8R1oGiNE6\\8P1PHAo7b0nG_OR8d0kG\\OU8g0hGYOX8i0eGXOZ8S2000000BeL_H\\3a7dL_H\\3`7fL_HZ3`7gL_HZ3a7gL^HY3a7hL_HX3a7iL^HW3a7jL^HW3b7jL]HV3b7lL]HT3b7mL]HT3c7mL\\HS3c7nL]HR3c7oL\\HQ3c7PM\\HQ3c7b000O100O10000O10000000000O1000000000000O10000000000O10000000000O1000000000000O1000000000000000000000000001O0RNdHUN1i0[7R1eHTN1i0Z7S1eHTN3g0X7U1eHTN4f0W7U1fHUN5d0U7W1fHTN7d0T7W1fHTN6d0V7W1dHUN7c0V7V1dHWN7b0V7V1cHWN8c0V7U1bHXN9b0V7U1bHXN9b0V7T1bHZN8b0W7S1aH[N<>T7V1`H[Nf06j6_1`H[NP1La6g1aH\\Nn8c1RG]No8b1QG^NP9i0hFD8BR9i0fFD9CR9h0fFD9CR9h0fFD8DS9f0gFE6DU9f0fFD6FU9e0fFD5GV9d0fFD5GV9c0fFF4GX9a0eFG3GZ9a0dFF3IZ9`0dFF2J[9>eFG1J[9>eFG0J[9`0fFD0LZ9`0fFD0LZ9?hFDNMZ9?jFBMNY9`0kF@MOY9a0kF_Oh9`0h00003M4HoSS2"}, "label": "bear in green shirt"}]}
{"id": 350984, "image": "COCO_train2014_000000350984.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bear with green shirt\"."}], "task": "refering", "answer_template": "The \"brown bear with green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 221, 222, 223, 224, 225, 226], "bbox": [0.26846, 0.6266133333333334, 0.6325599999999999, 1.0], "iscrowd": 0, "area": 15298.238650000005, "rle": {"size": [375, 500], "counts": "gia11e;1O1O100O1O1O1O100O1O1O1O10000000000000000001O0000000000QNG`H9\\72]HN^79`HG\\7?bHAY7e0gHZOT7m0jHSOR7S1lHmNT7U1jHkNV7V1iHjNW7X1gHhNX7[1fHeNZ7]1dHcN\\7^1kG\\N86l7^1mG^N54n7^1lG_N72m7_1lG_N72m7^1mG`N62l7_1mGaN51n7^1mGaN32Q8\\1mGbN13Q8\\1mGbN13R8[1mGcNO3T8Y1nGdNM4U8X1mGeNM4U8W1oGeNK5V8V1oGfNI5X8U1nGgNI5Y8S1oGhNG6Y8S1oGhNG6Z8R1oGiNE6\\8P1PHAo7b0nG_OR8d0kG\\OU8g0hGYOX8i0eGXOZ8S2000000BeL_H\\3a7dL_H\\3`7fL_HZ3`7gL_HZ3a7gL^HY3a7hL_HX3a7iL^HW3a7jL^HW3b7jL]HV3b7lL]HT3b7mL]HT3c7mL\\HS3c7nL]HR3c7oL\\HQ3c7PM\\HQ3c7b000O100O10000O10000000000O1000000000000O10000000000O10000000000O1000000000000O1000000000000000000000000001O0RNdHUN1i0[7R1eHTN1i0Z7S1eHTN3g0X7U1eHTN4f0W7U1fHUN5d0U7W1fHTN7d0T7W1fHTN6d0V7W1dHUN7c0V7V1dHWN7b0V7V1cHWN8c0V7U1bHXN9b0V7U1bHXN9b0V7T1bHZN8b0W7S1aH[N<>T7V1`H[Nf06j6_1`H[NP1La6g1aH\\Nn8c1RG]No8b1QG^NP9i0hFD8BR9i0fFD9CR9h0fFD9CR9h0fFD8DS9f0gFE6DU9f0fFD6FU9e0fFD5GV9d0fFD5GV9c0fFF4GX9a0eFG3GZ9a0dFF3IZ9`0dFF2J[9>eFG1J[9>eFG0J[9`0fFD0LZ9`0fFD0LZ9?hFDNMZ9?jFBMNY9`0kF@MOY9a0kF_Oh9`0h00003M4HoSS2"}, "label": "brown bear with green shirt"}]}
{"id": 187147, "image": "COCO_train2014_000000187147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red train\"."}], "task": "refering", "answer_template": "The \"red train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 204, 205], "bbox": [0.0, 0.166296875, 0.7910208333333333, 0.557296875], "iscrowd": 0, "area": 66792.2812, "rle": {"size": [640, 480], "counts": "Z3k7V<N100000000000000O1000000O2N1O100O1O100O1O100O1O1O101N1O100O100O10000O10000O101O000O10000O10000O10000O101O000000000O100000000000001N10000000000000000O1000001O0O1000000O1000000O1000001N1000000O1000000O1000000O2O000O10000O1000000O10000O2O000O1000000O10000O1000001N1000000O1000000O1000000O2O00000O1000000O1000000O101O000O1000000O1000000O101O000O1000000O1000000O10001O0O1000000O1000000O1000001N10000O100O100O100O100O101N100O100O100O100O100O100O2O0O100O100O100O1O100O100O2O0O100O1O100O100O100O10001N1000000O100000000O1000001N1000000O1000000O100000001N1000000000000000000000001N1000000O10000O10000O10001N10000O1000000O10000O10001N10000O10000O1000000O10001N10000O10000O10000O1000001N10000O10000O10000O1000001N10000O10000O10000O10001N11O1O0000O1000002M`0A?A?A>B?@`0A?A_Xn1"}, "label": "red train"}]}
{"id": 187147, "image": "COCO_train2014_000000187147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large red train on a track\"."}], "task": "refering", "answer_template": "The \"a large red train on a track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 204, 205], "bbox": [0.0, 0.166296875, 0.7910208333333333, 0.557296875], "iscrowd": 0, "area": 66792.2812, "rle": {"size": [640, 480], "counts": "Z3k7V<N100000000000000O1000000O2N1O100O1O100O1O100O1O1O101N1O100O100O10000O10000O101O000O10000O10000O10000O101O000000000O100000000000001N10000000000000000O1000001O0O1000000O1000000O1000001N1000000O1000000O1000000O2O000O10000O1000000O10000O2O000O1000000O10000O1000001N1000000O1000000O1000000O2O00000O1000000O1000000O101O000O1000000O1000000O101O000O1000000O1000000O10001O0O1000000O1000000O1000001N10000O100O100O100O100O101N100O100O100O100O100O100O2O0O100O100O100O1O100O100O2O0O100O1O100O100O100O10001N1000000O100000000O1000001N1000000O1000000O100000001N1000000000000000000000001N1000000O10000O10000O10001N10000O1000000O10000O10001N10000O10000O1000000O10001N10000O10000O10000O1000001N10000O10000O10000O1000001N10000O10000O10000O10001N11O1O0000O1000002M`0A?A?A>B?@`0A?A_Xn1"}, "label": "a large red train on a track"}]}
{"id": 121612, "image": "COCO_train2014_000000121612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white dishwasher , under the counter to the right\"."}], "task": "refering", "answer_template": "The \"white dishwasher , under the counter to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343], "bbox": [0.787640625, 0.5659375, 0.9789218749999999, 0.8448749999999999], "iscrowd": 0, "area": 11720.696650000002, "rle": {"size": [480, 640], "counts": "Pd\\75c>8G9G9H9F9H8G:G8G9H8G:G8H8N3O0000001O01O01O000010O000001O00010O0000001O01O01O00000010O0001O000010O000001O00010O00L4I7J7I6J6J6J6I7L5NeDjMn9V2oEmMQ:R2lERNT:n1lERNT:m1mETNR:l1oESNR:n1XEQNJ1n:Q2UEnMM1n:R2TEnMM0o:R2TEnMM0o:R2TEnMM1n:Q2UEnMM1o:P2UEoML0o:Q2UEoML0o:R2TEnMM0o:R2TEnMM1n:Q2UEoML0o:Q2UEoML0P;P2TEPNMOo:R2SEoMNOo:R2SEPNMNP;R2SEPNMOo:Q2TEPNMOP;P2SEQNMOP;Q2REPNONo:R2REQNNMP;R2REQNNNo:Q2SEQNNNo:Q2SEQNNNP;P2RESNMMQ;Q2QERNNMQ;Q2QERNOLP;R2QERNOMo:Q2RESNNLP;Q2RESNNLQ;Q2QERNNMQ;Q2QERNNMQ;Q2QESNNLP;Q2RESNNLP;Q2RESNNLP;R2QERNOLQ;Q2PESNOLQ;Q2PETNNKR;Q2PETNNIT;S2nDnMh;R2XDgMo;Z2<O00000M3D<E<D;D<E;DZj5"}, "label": "white dishwasher , under the counter to the right"}]}
{"id": 121612, "image": "COCO_train2014_000000121612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"door of the dishwasher\"."}], "task": "refering", "answer_template": "The \"door of the dishwasher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343], "bbox": [0.787640625, 0.5659375, 0.9789218749999999, 0.8448749999999999], "iscrowd": 0, "area": 11720.696650000002, "rle": {"size": [480, 640], "counts": "Pd\\75c>8G9G9H9F9H8G:G8G9H8G:G8H8N3O0000001O01O01O000010O000001O00010O0000001O01O01O00000010O0001O000010O000001O00010O00L4I7J7I6J6J6J6I7L5NeDjMn9V2oEmMQ:R2lERNT:n1lERNT:m1mETNR:l1oESNR:n1XEQNJ1n:Q2UEnMM1n:R2TEnMM0o:R2TEnMM0o:R2TEnMM1n:Q2UEnMM1o:P2UEoML0o:Q2UEoML0o:R2TEnMM0o:R2TEnMM1n:Q2UEoML0o:Q2UEoML0P;P2TEPNMOo:R2SEoMNOo:R2SEPNMNP;R2SEPNMOo:Q2TEPNMOP;P2SEQNMOP;Q2REPNONo:R2REQNNMP;R2REQNNNo:Q2SEQNNNo:Q2SEQNNNP;P2RESNMMQ;Q2QERNNMQ;Q2QERNOLP;R2QERNOMo:Q2RESNNLP;Q2RESNNLQ;Q2QERNNMQ;Q2QERNNMQ;Q2QESNNLP;Q2RESNNLP;Q2RESNNLP;R2QERNOLQ;Q2PESNOLQ;Q2PETNNKR;Q2PETNNIT;S2nDnMh;R2XDgMo;Z2<O00000M3D<E<D;D<E;DZj5"}, "label": "door of the dishwasher"}]}
{"id": 367375, "image": "COCO_train2014_000000367375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting in a carriage while wearing a striped shirt\"."}], "task": "refering", "answer_template": "The \"a man sitting in a carriage while wearing a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 169, 170, 192, 193, 194], "bbox": [0.30653125000000003, 0.19995316159250584, 0.45093750000000005, 0.6084543325526932], "iscrowd": 0, "area": 8631.93305, "rle": {"size": [427, 640], "counts": "kla21V=6K4K5K6F8I7M3O1O2N1O100O2N1O1O1O1M4K4hDWNm:j1PEZNn:g1oD\\NQ;P2N2N101N1O01O0001_OdMSF`2f9iMSF\\2g9m0F:E5M4L3N3L3gGcK_7a4[HgK_7\\4\\HfKe7]4VHeKj7n40O1000000O01000oMSHkM1_On7e2WHbM2Fg7h2]HWM4La7m2bHkL63Z7Q3cIjL_6V3k12O2M5LZMWFo1d9nM]FY2]9dMeFc2V9ZMlFl2o8PMSGW3g8fL\\G`3_8]LbGj3Y8nKnGR4e81O100O100O10N100bMlFd0V9YOQGb0o8\\OXG>j8_O]GK^OROU9R1cGA^O]Oo8P1jGYO]OEk8o0oGQO\\O0e8m0SIQOo6h0WIXOk6`0\\I_Oh67^IHf6OaI0m9O2O1O0OXab4"}, "label": "a man sitting in a carriage while wearing a striped shirt"}]}
{"id": 367375, "image": "COCO_train2014_000000367375.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a grey shirt riding a carriage\"."}], "task": "refering", "answer_template": "The \"man in a grey shirt riding a carriage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147, 169, 170, 192, 193, 194], "bbox": [0.30653125000000003, 0.19995316159250584, 0.45093750000000005, 0.6084543325526932], "iscrowd": 0, "area": 8631.93305, "rle": {"size": [427, 640], "counts": "kla21V=6K4K5K6F8I7M3O1O2N1O100O2N1O1O1O1M4K4hDWNm:j1PEZNn:g1oD\\NQ;P2N2N101N1O01O0001_OdMSF`2f9iMSF\\2g9m0F:E5M4L3N3L3gGcK_7a4[HgK_7\\4\\HfKe7]4VHeKj7n40O1000000O01000oMSHkM1_On7e2WHbM2Fg7h2]HWM4La7m2bHkL63Z7Q3cIjL_6V3k12O2M5LZMWFo1d9nM]FY2]9dMeFc2V9ZMlFl2o8PMSGW3g8fL\\G`3_8]LbGj3Y8nKnGR4e81O100O100O10N100bMlFd0V9YOQGb0o8\\OXG>j8_O]GK^OROU9R1cGA^O]Oo8P1jGYO]OEk8o0oGQO\\O0e8m0SIQOo6h0WIXOk6`0\\I_Oh67^IHf6OaI0m9O2O1O0OXab4"}, "label": "man in a grey shirt riding a carriage"}]}
{"id": 375568, "image": "COCO_train2014_000000375568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third seat from the camera lens\"."}], "task": "refering", "answer_template": "The \"the third seat from the camera lens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 182, 183, 203, 204, 225, 226, 246, 247, 248, 267, 268, 269, 290], "bbox": [0.15816993464052287, 0.32826797385620915, 0.35142156862745094, 0.6046732026143791], "iscrowd": 0, "area": 5303.825399999999, "rle": {"size": [612, 612], "counts": "lYj11Rc03N2M2O2N2M2O2M3N1O2M3N2M2O2N1N010O0100O0010O010O0100O010O010O10O01O010O01M2O1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2NO1O10O01O1O100O001O100O1O00100O1O1O010O1O1O2O2M2N3M2O2M2N3M2N3N1N3M3M2O2M2N3M2O2M2N3M4M2M4L3N3M3LZR]7"}, "label": "the third seat from the camera lens"}]}
{"id": 375568, "image": "COCO_train2014_000000375568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red chair in a field or yard , next to a dog\"."}], "task": "refering", "answer_template": "The \"a red chair in a field or yard , next to a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 154, 155, 156, 157, 176, 177, 178, 179, 198, 199, 200, 201, 220, 221, 222], "bbox": [0.03820261437908497, 0.0292156862745098, 0.2808986928104575, 0.47640522875816993], "iscrowd": 0, "area": 26347.8645, "rle": {"size": [612, 612], "counts": "Pk=P6R=2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N20000000000000000000000000O1M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M`nV8"}, "label": "a red chair in a field or yard , next to a dog"}]}
{"id": 375568, "image": "COCO_train2014_000000375568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bright pink chair with a white stain on the back\"."}], "task": "refering", "answer_template": "The \"a bright pink chair with a white stain on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 154, 155, 156, 157, 176, 177, 178, 179, 198, 199, 200, 201, 220, 221, 222], "bbox": [0.03820261437908497, 0.0292156862745098, 0.2808986928104575, 0.47640522875816993], "iscrowd": 0, "area": 26347.8645, "rle": {"size": [612, 612], "counts": "Pk=P6R=2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N20000000000000000000000000O1M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M`nV8"}, "label": "a bright pink chair with a white stain on the back"}]}
{"id": 375568, "image": "COCO_train2014_000000375568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red chair that is second from the left\"."}], "task": "refering", "answer_template": "The \"the red chair that is second from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 115, 116, 117, 136, 137, 138, 139, 158, 159, 160, 179, 180, 181, 182, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268], "bbox": [0.05156862745098039, 0.18598039215686274, 0.33408496732026144, 0.5761274509803922], "iscrowd": 0, "area": 16639.2076, "rle": {"size": [612, 612], "counts": "\\]c01Sc02M2O1N2O2M2O1N2O1O2M2O1N2O2M2O1O1N2O2M2O1N2O2M2O1O1N2O2M2O1N2O2M2O1O1N3N1N2O1N2O2N0O1N2N2O0O2N2N2N101N2N1O2N2O1N1O2N2N2O0O2N2N101N2N2N1O2O1N2N1O2N2O0O2N2N2N101N2N2N101N2N1O2N10O01O1O1O010O1O001O100O001O10O0100O10O0100O10O0100O010O100O010O10O001O0O2O0O2O0O2O0O2O001N101N101N101N101N101N3N2N2M4M2M3N2M4M2M3M3N2M4L3M3M3N3L3M3M3N2M4L3N2M4M>A?B>AWfc7"}, "label": "the red chair that is second from the left"}]}
{"id": 375568, "image": "COCO_train2014_000000375568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red chairs on the left of the dog\"."}], "task": "refering", "answer_template": "The \"the red chairs on the left of the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 115, 116, 117, 136, 137, 138, 139, 158, 159, 160, 179, 180, 181, 182, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268], "bbox": [0.05156862745098039, 0.18598039215686274, 0.33408496732026144, 0.5761274509803922], "iscrowd": 0, "area": 16639.2076, "rle": {"size": [612, 612], "counts": "\\]c01Sc02M2O1N2O2M2O1N2O1O2M2O1N2O2M2O1O1N2O2M2O1N2O2M2O1O1N2O2M2O1N2O2M2O1O1N3N1N2O1N2O2N0O1N2N2O0O2N2N2N101N2N1O2N2O1N1O2N2N2O0O2N2N101N2N2N1O2O1N2N1O2N2O0O2N2N2N101N2N2N101N2N1O2N10O01O1O1O010O1O001O100O001O10O0100O10O0100O10O0100O010O100O010O10O001O0O2O0O2O0O2O0O2O001N101N101N101N101N101N3N2N2M4M2M3N2M4M2M3M3N2M4L3M3M3N3L3M3M3N2M4L3N2M4M>A?B>AWfc7"}, "label": "the red chairs on the left of the dog"}]}
{"id": 506640, "image": "COCO_train2014_000000506640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing in middle with head down\"."}], "task": "refering", "answer_template": "The \"zebra standing in middle with head down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 197, 198, 199, 200, 201, 220, 221, 222, 223, 243, 244, 246, 266, 269], "bbox": [0.57459375, 0.46273584905660375, 0.7473906250000001, 0.7716981132075471], "iscrowd": 0, "area": 6831.877800000003, "rle": {"size": [424, 640], "counts": "ndh4=^<c0^Ob0]Oc0]O?A?B>A?N2NM4M2M4M2M4M2M4I6A`0H7K6K4O2O1NcEXN\\9g1bFaNZ9^1bFfN^9Y1_FkNb9T1ZFPOf9o0WFUOj9i0TFZOl9f0PF]OR:k11G:9F101N100O100OK6H7L5L4K4M4K4M4K5I7F:O010O100O10O0100O1H8O0O2O1O1N101O1N2O1O0O2O1O1N2O001N2O1O^C:c;i0L5L3M4L2N1N101O001O1N1O2N1N2N100O2N\\NTEn0l:QO[Ei0d:VOdEe0[:ZOlE`0S:@j1O2O001O0O2O1OnjR2"}, "label": "zebra standing in middle with head down"}]}
{"id": 506640, "image": "COCO_train2014_000000506640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in the middle with its face near the ground\"."}], "task": "refering", "answer_template": "The \"the zebra in the middle with its face near the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 197, 198, 199, 200, 201, 220, 221, 222, 223, 243, 244, 246, 266, 269], "bbox": [0.57459375, 0.46273584905660375, 0.7473906250000001, 0.7716981132075471], "iscrowd": 0, "area": 6831.877800000003, "rle": {"size": [424, 640], "counts": "ndh4=^<c0^Ob0]Oc0]O?A?B>A?N2NM4M2M4M2M4M2M4I6A`0H7K6K4O2O1NcEXN\\9g1bFaNZ9^1bFfN^9Y1_FkNb9T1ZFPOf9o0WFUOj9i0TFZOl9f0PF]OR:k11G:9F101N100O100OK6H7L5L4K4M4K4M4K5I7F:O010O100O10O0100O1H8O0O2O1O1N101O1N2O1O0O2O1O1N2O001N2O1O^C:c;i0L5L3M4L2N1N101O001O1N1O2N1N2N100O2N\\NTEn0l:QO[Ei0d:VOdEe0[:ZOlE`0S:@j1O2O001O0O2O1OnjR2"}, "label": "the zebra in the middle with its face near the ground"}]}
{"id": 506640, "image": "COCO_train2014_000000506640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the furthest away of a group of three zebras , who has its head raised\"."}], "task": "refering", "answer_template": "The \"the furthest away of a group of three zebras , who has its head raised\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 199, 240, 242, 243, 263, 265], "bbox": [0.43770312499999997, 0.36405660377358495, 0.7533125, 0.7617924528301887], "iscrowd": 0, "area": 4867.276699999998, "rle": {"size": [424, 640], "counts": "VWd31W=2]CNl;5oCNP<4lCNS<6hCLX<6cCN\\<?0000O2O000012M4L3M4M3K5IYT=LnkB001O0010O01O001O001O10O01O001O001O01O000000_FLn54PJOo51oI3o5MoI6Q6ImI7U6IiI7Y6IeI8\\6HbI8`6H^I8d6HZI8h6HWI8k6HSI7o6IoH7S7IlH6V7JiH5Y7KfH1_7Om201O0000000000O10000O2O001N10001N101O0O101O0O101O0O2O000O2O001N10001N1O101N1O2N101N2N1O2O1N1O2O1N1O2O1N101N2O0O2O1N1O2O1N101N2O0O2N2O0O2O0O01N101O0O101N10001N10001N10001N101O00111O100O100OO1O2M2O1N10N1bNcDQ1_;mNdDP1];POdDn0^;QOdD>k;BWDJ\\<5;01O001O001OdXQ2"}, "label": "the furthest away of a group of three zebras , who has its head raised"}]}
{"id": 506640, "image": "COCO_train2014_000000506640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra closest to the water\"."}], "task": "refering", "answer_template": "The \"the zebra closest to the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 199, 240, 242, 243, 263, 265], "bbox": [0.43770312499999997, 0.36405660377358495, 0.7533125, 0.7617924528301887], "iscrowd": 0, "area": 4867.276699999998, "rle": {"size": [424, 640], "counts": "VWd31W=2]CNl;5oCNP<4lCNS<6hCLX<6cCN\\<?0000O2O000012M4L3M4M3K5IYT=LnkB001O0010O01O001O001O10O01O001O001O01O000000_FLn54PJOo51oI3o5MoI6Q6ImI7U6IiI7Y6IeI8\\6HbI8`6H^I8d6HZI8h6HWI8k6HSI7o6IoH7S7IlH6V7JiH5Y7KfH1_7Om201O0000000000O10000O2O001N10001N101O0O101O0O101O0O2O000O2O001N10001N1O101N1O2N101N2N1O2O1N1O2O1N1O2O1N101N2O0O2O1N1O2O1N101N2O0O2N2O0O2O0O01N101O0O101N10001N10001N10001N101O00111O100O100OO1O2M2O1N10N1bNcDQ1_;mNdDP1];POdDn0^;QOdD>k;BWDJ\\<5;01O001O001OdXQ2"}, "label": "the zebra closest to the water"}]}
{"id": 506640, "image": "COCO_train2014_000000506640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra focusing towards the camera in the front side\"."}], "task": "refering", "answer_template": "The \"zebra focusing towards the camera in the front side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 121, 122, 123, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 264, 265, 283, 285, 286, 287, 288, 306, 308, 309, 310], "bbox": [0.2724375, 0.3303301886792453, 0.5627500000000001, 0.9033726415094339], "iscrowd": 0, "area": 24165.982549999997, "rle": {"size": [424, 640], "counts": "lUX2b0f<3M3M3M3M2N4L3M3M=C4L3M4L4L3M4L4L5K5K4L2N2N2N2kImLQ2U3iMTMR2n2gMXMX2n2^MVMb2Q3RMUMm2n2kLWMU3l2dLXM\\3l2[LZMd3i2UL[Mk3h2jJaNU5i4000N2N2O1N2N2N2O1N2N3M3M3L4L4L4I7F:F:F:F:F:N2N102N2M4M2N3M3L3N3M2M4N2N1O000001O101N1O2N2N2M2N3N2M3M6K6I6Jf0[O5J1O01XMaHE_7;hH^OX7b0oHWOR7h0UIQOk6o0\\IjNc6W1bIdN]6]1gI^NY6c1kIYNT6h1PJTNo5m1UJoMj5R2\\JhMb5Z2f20O100000000O10001O0O01O100O10aHmLh4S3VKQMi4n2VKUMi4k2UKXMj4g2UK\\Mj4d2TK_Mk4`2TKcMk4\\2TKgMk4Y2SKiMm4V2RKkMo4T2PKnMQ5Q2mJQNS5n1kJUNU5k1iJVNX5i1fJYN[5f1dJZN^5f1_J[Nc5d1[J]Ng5b1SJdNo5[1hIlNZ6S1_ISOc6Q1nHXOT7m0]HROn7i2;M3N2M3N3L3NZHiK_6U4gHfLZ7W3`HRM`7[4000000001O00001O1O0O2N1O2VMVH4j7I_H0c7MeHL\\71mHHU75SIEm69[I@g6<eIZO\\6d0UJlNm5Q1V3N3M2N3N2M2N3M2N3M2N3M2O2M2N3M2Ndgc3"}, "label": "zebra focusing towards the camera in the front side"}]}
{"id": 506640, "image": "COCO_train2014_000000506640.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra looking at camera\"."}], "task": "refering", "answer_template": "The \"a zebra looking at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 121, 122, 123, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 264, 265, 283, 285, 286, 287, 288, 306, 308, 309, 310], "bbox": [0.2724375, 0.3303301886792453, 0.5627500000000001, 0.9033726415094339], "iscrowd": 0, "area": 24165.982549999997, "rle": {"size": [424, 640], "counts": "lUX2b0f<3M3M3M3M2N4L3M3M=C4L3M4L4L3M4L4L5K5K4L2N2N2N2kImLQ2U3iMTMR2n2gMXMX2n2^MVMb2Q3RMUMm2n2kLWMU3l2dLXM\\3l2[LZMd3i2UL[Mk3h2jJaNU5i4000N2N2O1N2N2N2O1N2N3M3M3L4L4L4I7F:F:F:F:F:N2N102N2M4M2N3M3L3N3M2M4N2N1O000001O101N1O2N2N2M2N3N2M3M6K6I6Jf0[O5J1O01XMaHE_7;hH^OX7b0oHWOR7h0UIQOk6o0\\IjNc6W1bIdN]6]1gI^NY6c1kIYNT6h1PJTNo5m1UJoMj5R2\\JhMb5Z2f20O100000000O10001O0O01O100O10aHmLh4S3VKQMi4n2VKUMi4k2UKXMj4g2UK\\Mj4d2TK_Mk4`2TKcMk4\\2TKgMk4Y2SKiMm4V2RKkMo4T2PKnMQ5Q2mJQNS5n1kJUNU5k1iJVNX5i1fJYN[5f1dJZN^5f1_J[Nc5d1[J]Ng5b1SJdNo5[1hIlNZ6S1_ISOc6Q1nHXOT7m0]HROn7i2;M3N2M3N3L3NZHiK_6U4gHfLZ7W3`HRM`7[4000000001O00001O1O0O2N1O2VMVH4j7I_H0c7MeHL\\71mHHU75SIEm69[I@g6<eIZO\\6d0UJlNm5Q1V3N3M2N3N2M2N3M2N3M2N3M2O2M2N3M2Ndgc3"}, "label": "a zebra looking at camera"}]}
{"id": 432914, "image": "COCO_train2014_000000432914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ferry boat to the right\"."}], "task": "refering", "answer_template": "The \"ferry boat to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.565234375, 0.7915690866510539, 1.0, 0.9219203747072598], "iscrowd": 0, "area": 12736.394650000002, "rle": {"size": [427, 640], "counts": "YYg4<n<b0_O1N2N100O1O100O1O100O1O10000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000001O000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001N10O100000000000000O100000000000001N3N2N2N2N2N2N2N2M2O2N2N2N2N2N2M3N3M4L5K4LnC"}, "label": "ferry boat to the right"}]}
{"id": 432914, "image": "COCO_train2014_000000432914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat closest to the camera\"."}], "task": "refering", "answer_template": "The \"boat closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.565234375, 0.7915690866510539, 1.0, 0.9219203747072598], "iscrowd": 0, "area": 12736.394650000002, "rle": {"size": [427, 640], "counts": "YYg4<n<b0_O1N2N100O1O100O1O100O1O10000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000001O000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001N10O100000000000000O100000000000001N3N2N2N2N2N2N2N2M2O2N2N2N2N2N2M3N3M4L5K4LnC"}, "label": "boat closest to the camera"}]}
{"id": 572179, "image": "COCO_train2014_000000572179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with glasses and wearing a green shirt serving a piece of cake\"."}], "task": "refering", "answer_template": "The \"a woman with glasses and wearing a green shirt serving a piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 303, 304, 305, 306, 326], "bbox": [0.109546875, 0.0555, 0.488765625, 0.8352708333333333], "iscrowd": 0, "area": 52011.83935000001, "rle": {"size": [480, 640], "counts": "TRQ1>d=R1I7I7J6I7I7I8J5J6K5K5J6K5J7J5J6K5K6I6WMULTJm3f5ZLVJi3d5^LYJc3b5dL[J^3`5hL\\JZ3_5nL]JT3]5SM`Jo2Z5XMcJi2Y5]McJf2Y5_MdJc2X5bMeJ_2W5gMeJ\\2W5h3M4N2N101N2N101N1kKXFQ3j9jLYFV3i9dL\\F\\3d9`L`F_3b9\\LbFc3`9XLcFi3Z:O1O00100O00100O00100O001O10O01O10O01O1O10O01O10O01O10O01000O10O10O10000O10000O10000O2O1O001N2O001N2O001N2O001N2O001N2O001N2O001N2O001O1N101O1N101O0O2O1O0O2O1O0O2O1O0O3M3M2N3M3M2jHkJ^4X5[KmJe4U5WKnJg4U5VKlJi4X5RKjJm4Z5oJhJP5[5kJgJT5Q6SJRJk5U6lInIS6a7O0O1O2N1O101N1O1O101N1O1O2O0O1O2N2O1N2N3M2N2O1N2N3M2O1N2N2N3N1N2N2N4L5L5J5K6J5L4K3M2N2[NXEoNk:o0WEoNj:P1YElNi:T1XEiNj:V1YEfNj:X1YEdNi:[1ZEbNi:\\1g1F8H1O100O010O1O10O01O10O01O010O0010O01O0100O010O1O010O100O00100O00100O0001O001O001G:M2M4M2M6K5J^Pi4"}, "label": "a woman with glasses and wearing a green shirt serving a piece of cake"}]}
{"id": 572179, "image": "COCO_train2014_000000572179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman cutting a cake\"."}], "task": "refering", "answer_template": "The \"a woman cutting a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 49, 50, 51, 52, 53, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 303, 304, 305, 306, 326], "bbox": [0.109546875, 0.0555, 0.488765625, 0.8352708333333333], "iscrowd": 0, "area": 52011.83935000001, "rle": {"size": [480, 640], "counts": "TRQ1>d=R1I7I7J6I7I7I8J5J6K5K5J6K5J7J5J6K5K6I6WMULTJm3f5ZLVJi3d5^LYJc3b5dL[J^3`5hL\\JZ3_5nL]JT3]5SM`Jo2Z5XMcJi2Y5]McJf2Y5_MdJc2X5bMeJ_2W5gMeJ\\2W5h3M4N2N101N2N101N1kKXFQ3j9jLYFV3i9dL\\F\\3d9`L`F_3b9\\LbFc3`9XLcFi3Z:O1O00100O00100O00100O001O10O01O10O01O1O10O01O10O01O10O01000O10O10O10000O10000O10000O2O1O001N2O001N2O001N2O001N2O001N2O001N2O001N2O001O1N101O1N101O0O2O1O0O2O1O0O2O1O0O3M3M2N3M3M2jHkJ^4X5[KmJe4U5WKnJg4U5VKlJi4X5RKjJm4Z5oJhJP5[5kJgJT5Q6SJRJk5U6lInIS6a7O0O1O2N1O101N1O1O101N1O1O2O0O1O2N2O1N2N3M2N2O1N2N3M2O1N2N2N3N1N2N2N4L5L5J5K6J5L4K3M2N2[NXEoNk:o0WEoNj:P1YElNi:T1XEiNj:V1YEfNj:X1YEdNi:[1ZEbNi:\\1g1F8H1O100O010O1O10O01O10O01O010O0010O01O0100O010O1O010O100O00100O00100O0001O001O001G:M2M4M2M6K5J^Pi4"}, "label": "a woman cutting a cake"}]}
{"id": 572179, "image": "COCO_train2014_000000572179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old mandecorating a cake\"."}], "task": "refering", "answer_template": "The \"an old mandecorating a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.464125, 0.05739583333333333, 0.857625, 0.6448333333333334], "iscrowd": 0, "area": 45659.72529999999, "rle": {"size": [480, 640], "counts": "^^[41P?1N2N2N101N2UOIiB7T=LlB5P=NPC4k<OVC1f<2ZCOb<3_CO[<5eCLW<7jCI8Fb:d0VEn0g:UOYEk0d:XO\\Eh0a:[O_Ee0_:]ObEb0[:AeE?Y:DgE;V:HjE8T:JmE5P:NQF1m91TFNi96WFIg99YFGd9<]FCa9?`F@^9b0cF]OZ9g0eFXOZ9j0gFUOW9m0jFROT9P1mFoNR9S1nFkNQ9W1oFiNo8Y1RGfNm8[1TGcNk8`1UG_Ni8d1WG[Nh8f1XGZNf8i1XGWNg8l1WGUNg8n1WGSNh8o1VGQNk8P2SGQNm8P2RGPNm8R2QGoMo8Q2PGoMQ9R2mFoMR9S2lFnMT9S2kFlMV9U2hFlMX9U2fFlMY9V2eFkM[9V2cFjM^9V2bFjM]9X2aFiM_9X2_FhMb9Y2\\FhMc9Z2[FhMd9k3N2ZNfJcI[5[6hJcIZ5Z6iJdIY5Y6kJdIW5Y6lJeIV5X6mJfIU5V6PKgIQ5X6QKfIQ5X6l100O2O0O1O2O0O100O2N100O2N100O101N1O101N100O1O2O0O101O0O100O101N10000O100O100O10000O100O01O100O101N100O2N1O1O2N1N3O00001O00001O00001OYISH[6Y8001O001N10001O0O102N2N1N3N2N2N1O2M3N2N1O2M3N2N4L6I6K5hImGY5X8`JjG`5[8YJhGf5]8SJeGl5h8000000O100000000O1000000O10000O1N2M3N2N2M3N31O1O1O1O1N2O1OO1O1O1O1O001O1O1O1O1O1O001O100O3M3M2N3M3M2N3M3M2N3M3M3M4L4L4K5K4L5K5K5K5K5K5K5K5K5K5K5K5K4L5J6K5I7I7H8I7Ho\\Z1"}, "label": "an old mandecorating a cake"}]}
{"id": 572179, "image": "COCO_train2014_000000572179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older man holding a spatula while sitting at a table getting ready to have cake\"."}], "task": "refering", "answer_template": "The \"older man holding a spatula while sitting at a table getting ready to have cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 59, 60, 61, 62, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.464125, 0.05739583333333333, 0.857625, 0.6448333333333334], "iscrowd": 0, "area": 45659.72529999999, "rle": {"size": [480, 640], "counts": "^^[41P?1N2N2N101N2UOIiB7T=LlB5P=NPC4k<OVC1f<2ZCOb<3_CO[<5eCLW<7jCI8Fb:d0VEn0g:UOYEk0d:XO\\Eh0a:[O_Ee0_:]ObEb0[:AeE?Y:DgE;V:HjE8T:JmE5P:NQF1m91TFNi96WFIg99YFGd9<]FCa9?`F@^9b0cF]OZ9g0eFXOZ9j0gFUOW9m0jFROT9P1mFoNR9S1nFkNQ9W1oFiNo8Y1RGfNm8[1TGcNk8`1UG_Ni8d1WG[Nh8f1XGZNf8i1XGWNg8l1WGUNg8n1WGSNh8o1VGQNk8P2SGQNm8P2RGPNm8R2QGoMo8Q2PGoMQ9R2mFoMR9S2lFnMT9S2kFlMV9U2hFlMX9U2fFlMY9V2eFkM[9V2cFjM^9V2bFjM]9X2aFiM_9X2_FhMb9Y2\\FhMc9Z2[FhMd9k3N2ZNfJcI[5[6hJcIZ5Z6iJdIY5Y6kJdIW5Y6lJeIV5X6mJfIU5V6PKgIQ5X6QKfIQ5X6l100O2O0O1O2O0O100O2N100O2N100O101N1O101N100O1O2O0O101O0O100O101N10000O100O100O10000O100O01O100O101N100O2N1O1O2N1N3O00001O00001O00001OYISH[6Y8001O001N10001O0O102N2N1N3N2N2N1O2M3N2N1O2M3N2N4L6I6K5hImGY5X8`JjG`5[8YJhGf5]8SJeGl5h8000000O100000000O1000000O10000O1N2M3N2N2M3N31O1O1O1O1N2O1OO1O1O1O1O001O1O1O1O1O1O001O100O3M3M2N3M3M2N3M3M2N3M3M3M4L4L4K5K4L5K5K5K5K5K5K5K5K5K5K5K5K4L5J6K5I7I7H8I7Ho\\Z1"}, "label": "older man holding a spatula while sitting at a table getting ready to have cake"}]}
{"id": 342807, "image": "COCO_train2014_000000342807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty red chair\"."}], "task": "refering", "answer_template": "The \"an empty red chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.621625, 0.617125, 1.0, 0.9887291666666668], "iscrowd": 0, "area": 38439.2197, "rle": {"size": [480, 640], "counts": "dPk54f>7H7J6J6I7J6J6I7J6J6I7J7I6J6I7J6J6I7J6J6I7J6J6I8N1O10000000000000000000000O1000000000000000O1000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O0100WNi1`MXI"}, "label": "an empty red chair"}]}
{"id": 342807, "image": "COCO_train2014_000000342807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a solid pink bright chair\"."}], "task": "refering", "answer_template": "The \"a solid pink bright chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.621625, 0.617125, 1.0, 0.9887291666666668], "iscrowd": 0, "area": 38439.2197, "rle": {"size": [480, 640], "counts": "dPk54f>7H7J6J6I7J6J6I7J6J6I7J7I6J6I7J6J6I7J6J6I7J6J6I8N1O10000000000000000000000O1000000000000000O1000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O0100WNi1`MXI"}, "label": "a solid pink bright chair"}]}
{"id": 228119, "image": "COCO_train2014_000000228119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small baby in pink dress having milk\"."}], "task": "refering", "answer_template": "The \"a small baby in pink dress having milk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 118, 119, 120, 135, 136, 137, 138, 152, 153, 154, 155, 156, 170, 171, 172, 173, 188, 189, 190, 191, 192, 207, 208, 209, 210], "bbox": [0.44256, 0.3997866666666666, 0.7145199999999999, 0.9065333333333333], "iscrowd": 0, "area": 13514.105150000003, "rle": {"size": [375, 500], "counts": "PWa21f;2M3N3M2M3N2M3N2M3N2M2O0O1O2N100O2N1O1bNoNUHS1i7QOTHo0l7UOoGl0Q8XOjGj0T8[OgGf0Y8^ObGc0^8_O_Gc0_8@_G`0a8C[G>e8DXG>f8FVG;j8Y101N10000O2O000O1O2O0O101N100O101N1O1O1N3N1O1O2N1N2O1O2N1O1O2O00000010O0000jNYL\\Jh3b5ZL]Jf3a5]L^Jf3^5\\LaJh3Z5ZLeJf3Z5\\LeJd3Z5^LeJb3Y5aLfJ_3Y5cLfJ`0VOg1S6jMgJ<YOj1o5lMgJ9[Oj1m5PNeJ8^Oh1l5RNcJ8Af1k5TNaJ8Dd1j5VN_J9G`1i5`NVJ21^1g52YJNg52YJNg52YJNg53XJMh53XJMh53YJLg54YJLg54YJKg56YJJg57XJIh57XJIh57XJIh57]JDc5<bJ_O^5b0fJYOZ5g0hJWOX5i0kJTOT5m0nJQOR5P1PKmNP5S1RKkNn4V1SKhNm4X1VKdNk4]1VKaNk4^1XK_Ni4_1ZK^Nh4_1]K]Nd4b1_K\\Nb4a1eKYN]4d1iKWNX4h1nKSNS4j1TLnMP4P2VLcMR4[2g2N2[OfFiN\\9V1oF_NS9_1d0O1N1O2N2O0O2N2N1O2K7J7H8I7HWVd1"}, "label": "a small baby in pink dress having milk"}]}
{"id": 228119, "image": "COCO_train2014_000000228119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby girl wearing pink clothes being fed from a bottle\"."}], "task": "refering", "answer_template": "The \"a baby girl wearing pink clothes being fed from a bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 118, 119, 120, 135, 136, 137, 138, 152, 153, 154, 155, 156, 170, 171, 172, 173, 188, 189, 190, 191, 192, 207, 208, 209, 210], "bbox": [0.44256, 0.3997866666666666, 0.7145199999999999, 0.9065333333333333], "iscrowd": 0, "area": 13514.105150000003, "rle": {"size": [375, 500], "counts": "PWa21f;2M3N3M2M3N2M3N2M3N2M2O0O1O2N100O2N1O1bNoNUHS1i7QOTHo0l7UOoGl0Q8XOjGj0T8[OgGf0Y8^ObGc0^8_O_Gc0_8@_G`0a8C[G>e8DXG>f8FVG;j8Y101N10000O2O000O1O2O0O101N100O101N1O1O1N3N1O1O2N1N2O1O2N1O1O2O00000010O0000jNYL\\Jh3b5ZL]Jf3a5]L^Jf3^5\\LaJh3Z5ZLeJf3Z5\\LeJd3Z5^LeJb3Y5aLfJ_3Y5cLfJ`0VOg1S6jMgJ<YOj1o5lMgJ9[Oj1m5PNeJ8^Oh1l5RNcJ8Af1k5TNaJ8Dd1j5VN_J9G`1i5`NVJ21^1g52YJNg52YJNg52YJNg53XJMh53XJMh53YJLg54YJLg54YJKg56YJJg57XJIh57XJIh57XJIh57]JDc5<bJ_O^5b0fJYOZ5g0hJWOX5i0kJTOT5m0nJQOR5P1PKmNP5S1RKkNn4V1SKhNm4X1VKdNk4]1VKaNk4^1XK_Ni4_1ZK^Nh4_1]K]Nd4b1_K\\Nb4a1eKYN]4d1iKWNX4h1nKSNS4j1TLnMP4P2VLcMR4[2g2N2[OfFiN\\9V1oF_NS9_1d0O1N1O2N2O0O2N2N1O2K7J7H8I7HWVd1"}, "label": "a baby girl wearing pink clothes being fed from a bottle"}]}
{"id": 228119, "image": "COCO_train2014_000000228119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old man feeding milk to baby on bed\"."}], "task": "refering", "answer_template": "The \"an old man feeding milk to baby on bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 111, 112, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 187, 188, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.10328, 0.09197333333333334, 0.8515799999999999, 0.9858933333333334], "iscrowd": 0, "area": 54692.93719999998, "rle": {"size": [375, 500], "counts": "l[c04a;3M4K5L3M4L3L5L3M4K5L3M4L3L5L4L3M4K4M2N2M3N2N200O100O10000O100]M^NPKc1o4SO\\Jm0b5AQJ`0n5BQJ>n5EPJ;n5HPJ9o5IPJ7o5KoI6P6MnI3o5l2K5K5K5E;F:M3M3M3M3L4L4O100000000000000O100O10000O10000O100O11O00001O001O00001O00001O00001O001O00001O00001O000000O10000\\OhHiLX7i2WIVMi6b2`I]M`6`2dI^M]6_2hI]MZ6`2jI]MX6a2kI\\MW6a2a1M3O10000O100O10001N1N2N2N2N2N2FZFaNh9^1YF`Ni9_1XF_Nj9_1XF_Nj9`17O1O1O10000O10000O1000000O10000O10000O1000000O1001O000000000000000000001O00001O000YIfNQ3[1kLhNU3X1jLjNh0OPNW1W1kNh04lMR1[1jNh07kMo0]1kNf09kMl0_1kNf0:jMl0_1kNe0<jMi0a1kNd0?iMf0c1kNc0b0hMd0d1kNc0c0gMc0d1kNe0c0fMc0d1jNf0c0fMd0c1jNf0c0fMd0c1iNg0c0fMe0b1iNf0c0hMe0`1iNh0c0gMe0`1hNi0c0gMf0^1iNj0b0gMf0^1hNk0b0gMg0]1gNl0c0fMh0X1jNQ1>gMi0m0SO[15hMj0f0VO4ROmNo08j0c0XO3WOoNg0;l0>YO8XOjNe0?l0:ZO=YOeNb0d0l05]Ob0YO_Na0i0k00@f0XO]N>m0l0L@j0YOYN<Q1l0J@l0ZOUN=T1l0G^OP1[ORN<W1o0B[OT1^OoM:Z1Q1_OWOGWNm0l2MJZOUOGYNR1i2LMVOROH[NV1g2L0QOoNI^NZ1d2K7jNYO[1`0Kk26UMJk25VMKj25WMJi25XMKg25ZMKf23\\MMd22^MMb21`MO`20aM0_2ObM1^2MeM2Z2NgM2Y2MhM3X2KkM4U2KlM5T2JmM6S2JmM6S2InM7Q2JoM6Q2IPN7P2IPN7P2IPN7P2HQN8o1HQN8o1GRN9n1GSN8m1HUN6k1IXN5h1KZN3f1L]N2b1O^N1c1N]N2n1BSN>n1ARN?n1@SN`0n1_ORNa0o1^OQNb0o1]ORNc0o1[ORNe0n1ZOSNf0n1WOTNi0m1UOTNk0\\6000000000000000000000000000000000000000O10000000000O10000000000O100000000O10000O1000000O10000O10000O100000000000000000000000000000001O00000000000000000000000000000000000000000001O001O001O001O001O001O1O001O001O001O1O001O2N1O1O2N1O1O2N2N2N1O2N2N2N2N2Negj0"}, "label": "an old man feeding milk to baby on bed"}]}
{"id": 228119, "image": "COCO_train2014_000000228119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man feeding the baby\"."}], "task": "refering", "answer_template": "The \"the man feeding the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 111, 112, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 187, 188, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.10328, 0.09197333333333334, 0.8515799999999999, 0.9858933333333334], "iscrowd": 0, "area": 54692.93719999998, "rle": {"size": [375, 500], "counts": "l[c04a;3M4K5L3M4L3L5L3M4K5L3M4L3L5L4L3M4K4M2N2M3N2N200O100O10000O100]M^NPKc1o4SO\\Jm0b5AQJ`0n5BQJ>n5EPJ;n5HPJ9o5IPJ7o5KoI6P6MnI3o5l2K5K5K5E;F:M3M3M3M3L4L4O100000000000000O100O10000O10000O100O11O00001O001O00001O00001O00001O001O00001O00001O000000O10000\\OhHiLX7i2WIVMi6b2`I]M`6`2dI^M]6_2hI]MZ6`2jI]MX6a2kI\\MW6a2a1M3O10000O100O10001N1N2N2N2N2N2FZFaNh9^1YF`Ni9_1XF_Nj9_1XF_Nj9`17O1O1O10000O10000O1000000O10000O10000O1000000O1001O000000000000000000001O00001O000YIfNQ3[1kLhNU3X1jLjNh0OPNW1W1kNh04lMR1[1jNh07kMo0]1kNf09kMl0_1kNf0:jMl0_1kNe0<jMi0a1kNd0?iMf0c1kNc0b0hMd0d1kNc0c0gMc0d1kNe0c0fMc0d1jNf0c0fMd0c1jNf0c0fMd0c1iNg0c0fMe0b1iNf0c0hMe0`1iNh0c0gMe0`1hNi0c0gMf0^1iNj0b0gMf0^1hNk0b0gMg0]1gNl0c0fMh0X1jNQ1>gMi0m0SO[15hMj0f0VO4ROmNo08j0c0XO3WOoNg0;l0>YO8XOjNe0?l0:ZO=YOeNb0d0l05]Ob0YO_Na0i0k00@f0XO]N>m0l0L@j0YOYN<Q1l0J@l0ZOUN=T1l0G^OP1[ORN<W1o0B[OT1^OoM:Z1Q1_OWOGWNm0l2MJZOUOGYNR1i2LMVOROH[NV1g2L0QOoNI^NZ1d2K7jNYO[1`0Kk26UMJk25VMKj25WMJi25XMKg25ZMKf23\\MMd22^MMb21`MO`20aM0_2ObM1^2MeM2Z2NgM2Y2MhM3X2KkM4U2KlM5T2JmM6S2JmM6S2InM7Q2JoM6Q2IPN7P2IPN7P2IPN7P2HQN8o1HQN8o1GRN9n1GSN8m1HUN6k1IXN5h1KZN3f1L]N2b1O^N1c1N]N2n1BSN>n1ARN?n1@SN`0n1_ORNa0o1^OQNb0o1]ORNc0o1[ORNe0n1ZOSNf0n1WOTNi0m1UOTNk0\\6000000000000000000000000000000000000000O10000000000O10000000000O100000000O10000O1000000O10000O10000O100000000000000000000000000000001O00000000000000000000000000000000000000000001O001O001O001O001O001O1O001O001O001O1O001O2N1O1O2N1O1O2N2N2N1O2N2N2N2N2Negj0"}, "label": "the man feeding the baby"}]}
{"id": 6943, "image": "COCO_train2014_000000006943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella with a hello kitty design\"."}], "task": "refering", "answer_template": "The \"an umbrella with a hello kitty design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 59, 73, 74, 75, 77, 78, 79, 80, 82, 83, 101, 102, 105], "bbox": [0.2124375, 0.07628235294117647, 0.65365625, 0.2888], "iscrowd": 0, "area": 13511.141650000001, "rle": {"size": [425, 640], "counts": "e`h1;m<=D0O101N100O101N100O100O2O0O100O1O2O0O100O101N100O100O2O0O100O101N100O100O2O0O100O0001O0O2O001O001O001O001O0O2O001O001O001O001N10001O001O1O1O0101O000000000O2O000000000O101O0000000O101O0000001O0O2O1O001O001N2O001O001O1N3N3M2N3M2N3M2M4M2N3M2N1O1N10010O00001O0000001O01O01O0000001O0000010O000000000000O101O0000000000000000001O0000000O1N2M3M3M3M4L3M3L4M3M3M3M3M4O0O101O001O00001N101O000010O01O000010O01O00010O001O0001O0001O01O0001O0001O0001O1O100O001O1003M4K4M3M3M4L1O0O01O100O1O1O001O100O1O1O001O100O1N2O1N1O2N2N2O1O1N2O001O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O2M2O1O1OSQl2"}, "label": "an umbrella with a hello kitty design"}]}
{"id": 6943, "image": "COCO_train2014_000000006943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white picture of an umbrella\"."}], "task": "refering", "answer_template": "The \"a black and white picture of an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 59, 73, 74, 75, 77, 78, 79, 80, 82, 83, 101, 102, 105], "bbox": [0.2124375, 0.07628235294117647, 0.65365625, 0.2888], "iscrowd": 0, "area": 13511.141650000001, "rle": {"size": [425, 640], "counts": "e`h1;m<=D0O101N100O101N100O100O2O0O100O1O2O0O100O101N100O100O2O0O100O101N100O100O2O0O100O0001O0O2O001O001O001O001O0O2O001O001O001O001N10001O001O1O1O0101O000000000O2O000000000O101O0000000O101O0000001O0O2O1O001O001N2O001O001O1N3N3M2N3M2N3M2M4M2N3M2N1O1N10010O00001O0000001O01O01O0000001O0000010O000000000000O101O0000000000000000001O0000000O1N2M3M3M3M4L3M3L4M3M3M3M3M4O0O101O001O00001N101O000010O01O000010O01O00010O001O0001O0001O01O0001O0001O0001O1O100O001O1003M4K4M3M3M4L1O0O01O100O1O1O001O100O1O1O001O100O1N2O1N1O2N2N2O1O1N2O001O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O2M2O1O1OSQl2"}, "label": "a black and white picture of an umbrella"}]}
{"id": 203559, "image": "COCO_train2014_000000203559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man taking car eof the elephant\"."}], "task": "refering", "answer_template": "The \"the man taking car eof the elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 206, 207, 208, 209, 210, 211, 212, 213, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 280, 281, 282, 292, 293, 294, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 344, 345, 346], "bbox": [0.1498125, 0.215046875, 0.6501875, 0.9026875000000001], "iscrowd": 0, "area": 62514.24855, "rle": {"size": [640, 480], "counts": "bZ]1g0ib0]1jNV1jNc0_O8M4M2N3L3N2M4M2N2M4M2M3N3M2M4J4L4L4L4K5gBPJ^:0kFU6bNPJQ:`0WGb5dNSJo9d0WG[5fNVJQ:d0SGX5hNZJR:c0PGQ6n8UJkFn5T9VJfFm5X9XJbFk5\\9^2N2N2N2O1N2N2N3RHbEU6_:iIeEU6\\:hIgEV6Z:gIjEW6X:fIkEX6V:fImER6^OlHg:o0nES6]OmHe:n0RFS6ZOmHf:m0SFT6T:jIoET6S:iIPFU6Q:iIRFV6o9hITFU6m9iIVFU6l9iIVFV6j9iIXFV6i9hIYFW6g9iIZFV6g9hIZFY6e9gI\\FX6e9fI]FY6c9fI^F[6b9dI_F[6b9cI`F\\6a9cI`F]6`9aIaF_6_9`IcF_6^9`IcF`6^;01O010O0000010O0001O00010O000010O000001O01O01N10001O1N101O001O0O2O001O1N101O1O0O2O1N1O2kFmHZ5U7dJnHY5T7eJoHX5R7gJQIW5P7gJSIV5n6iJUIT5m6jJVIS5l6kJWIi4S7VKPIg4R7WKQIf4P7YKSId4o6ZKTIc4m6\\KVIa4l6mIUHGR1Z6i6jI[HIP1[6f6kI[HHR1[6c6lI]HGT1Z6`6mI]HHV1Y6]6oI^HFX1Y6[6oI_HFY1Y6X6PJaHE]1V6S6SJaHFb1R6m5XJ`HFj1l5g5\\J^HIQ2f5a5`J^HJX2_5\\5eJ[HM`2X5U5jJZH0g2P5P5nJYH2l2l4k4QKXH4P3i4h4RKXH5S3g4f4RKXH6U3f4c4SKYH7V3d4a4TKZH8W3b4`4TKZH9Y3a4]4UK[H:Y3`4\\4UK\\H:V3d4_4QK[H;R3g4d4mJ\\H:n2l4f4iJ]H;j2o4j4dJ^H<e2S5m4_J`H=a2W5o4XJdH?[2]5R5oIiH`0U2c5_6ZJ_Ii5c6SJ\\IP6f6mIXIV6j6fIVI\\6T:2N3kMeIcF]6Z9nI[FT6c9VJSFm5j9YJPFi5n9^JkEe5R:aJgEb5X:cJbE`5\\:eJ^E]5a:R2N10YL`EPN^:o1eEQNY:n1kEPNT:n1QFQNm9n1VFQNi9n1[FQNc9m1aFRN^9m1fFRNX9m1kFRNS9m1RGRNl8k1ZGSNf8i1_GWNc8b1cG]N_8\\1fGcN\\8;fHD\\76hHI[70kHOW7KmH4V7EPI9V7]OoHb0V7VOoHh0W7oNnHo0Y7gNkHX1Z7aNjH]1Z7fMfDNS4[2Z7ZMVE0c3e2j<N1O2O0O2N1O2N1O1O2N1O2N1O2N1O2O0O1O2N1O2O0O2N1O2N2O1N2N1O2N2O1N2N1O2N2O1M2O2N2N2N2M2O>BSaX3"}, "label": "the man taking car eof the elephant"}]}
{"id": 203559, "image": "COCO_train2014_000000203559.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man scratching the elephant\"."}], "task": "refering", "answer_template": "The \"the man scratching the elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 206, 207, 208, 209, 210, 211, 212, 213, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 280, 281, 282, 292, 293, 294, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 344, 345, 346], "bbox": [0.1498125, 0.215046875, 0.6501875, 0.9026875000000001], "iscrowd": 0, "area": 62514.24855, "rle": {"size": [640, 480], "counts": "bZ]1g0ib0]1jNV1jNc0_O8M4M2N3L3N2M4M2N2M4M2M3N3M2M4J4L4L4L4K5gBPJ^:0kFU6bNPJQ:`0WGb5dNSJo9d0WG[5fNVJQ:d0SGX5hNZJR:c0PGQ6n8UJkFn5T9VJfFm5X9XJbFk5\\9^2N2N2N2O1N2N2N3RHbEU6_:iIeEU6\\:hIgEV6Z:gIjEW6X:fIkEX6V:fImER6^OlHg:o0nES6]OmHe:n0RFS6ZOmHf:m0SFT6T:jIoET6S:iIPFU6Q:iIRFV6o9hITFU6m9iIVFU6l9iIVFV6j9iIXFV6i9hIYFW6g9iIZFV6g9hIZFY6e9gI\\FX6e9fI]FY6c9fI^F[6b9dI_F[6b9cI`F\\6a9cI`F]6`9aIaF_6_9`IcF_6^9`IcF`6^;01O010O0000010O0001O00010O000010O000001O01O01N10001O1N101O001O0O2O001O1N101O1O0O2O1N1O2kFmHZ5U7dJnHY5T7eJoHX5R7gJQIW5P7gJSIV5n6iJUIT5m6jJVIS5l6kJWIi4S7VKPIg4R7WKQIf4P7YKSId4o6ZKTIc4m6\\KVIa4l6mIUHGR1Z6i6jI[HIP1[6f6kI[HHR1[6c6lI]HGT1Z6`6mI]HHV1Y6]6oI^HFX1Y6[6oI_HFY1Y6X6PJaHE]1V6S6SJaHFb1R6m5XJ`HFj1l5g5\\J^HIQ2f5a5`J^HJX2_5\\5eJ[HM`2X5U5jJZH0g2P5P5nJYH2l2l4k4QKXH4P3i4h4RKXH5S3g4f4RKXH6U3f4c4SKYH7V3d4a4TKZH8W3b4`4TKZH9Y3a4]4UK[H:Y3`4\\4UK\\H:V3d4_4QK[H;R3g4d4mJ\\H:n2l4f4iJ]H;j2o4j4dJ^H<e2S5m4_J`H=a2W5o4XJdH?[2]5R5oIiH`0U2c5_6ZJ_Ii5c6SJ\\IP6f6mIXIV6j6fIVI\\6T:2N3kMeIcF]6Z9nI[FT6c9VJSFm5j9YJPFi5n9^JkEe5R:aJgEb5X:cJbE`5\\:eJ^E]5a:R2N10YL`EPN^:o1eEQNY:n1kEPNT:n1QFQNm9n1VFQNi9n1[FQNc9m1aFRN^9m1fFRNX9m1kFRNS9m1RGRNl8k1ZGSNf8i1_GWNc8b1cG]N_8\\1fGcN\\8;fHD\\76hHI[70kHOW7KmH4V7EPI9V7]OoHb0V7VOoHh0W7oNnHo0Y7gNkHX1Z7aNjH]1Z7fMfDNS4[2Z7ZMVE0c3e2j<N1O2O0O2N1O2N1O1O2N1O2N1O2N1O2O0O1O2N1O2O0O2N1O2N2O1N2N1O2N2O1N2N1O2N2O1M2O2N2N2N2M2O>BSaX3"}, "label": "the man scratching the elephant"}]}
{"id": 285486, "image": "COCO_train2014_000000285486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with dark hair using a laptop\"."}], "task": "refering", "answer_template": "The \"a man with dark hair using a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 153, 154, 175, 176, 177, 197, 198, 220, 221, 225], "bbox": [0.57725, 0.3340625, 0.8177031249999999, 0.5782708333333333], "iscrowd": 0, "area": 3085.1959, "rle": {"size": [480, 640], "counts": "VW]51n>0O2O1O1N2O1N2O1N2O001N2O1N2O1N2O1O1N2O1N101N2O1O1N2O1N2O1N101O1N2O01O1O100O9G<Cab5Fh]J=C6J3M3O1O1O1O1O1O100O1O100O10000O10000O100000O101O001O001O001O001O001O001O1O3M3M3M8H6J3M3K5LhQ??im@4L4K2O1O1O12N2N1O2N2N1O2N2N1O2N2N1O2N2N_bf1"}, "label": "a man with dark hair using a laptop"}]}
{"id": 285486, "image": "COCO_train2014_000000285486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing grey long sleeved shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing grey long sleeved shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 153, 154, 175, 176, 177, 197, 198, 220, 221, 225], "bbox": [0.57725, 0.3340625, 0.8177031249999999, 0.5782708333333333], "iscrowd": 0, "area": 3085.1959, "rle": {"size": [480, 640], "counts": "VW]51n>0O2O1O1N2O1N2O1N2O001N2O1N2O1N2O1O1N2O1N101N2O1O1N2O1N2O1N101O1N2O01O1O100O9G<Cab5Fh]J=C6J3M3O1O1O1O1O1O100O1O100O10000O10000O100000O101O001O001O001O001O001O001O1O3M3M3M8H6J3M3K5LhQ??im@4L4K2O1O1O12N2N1O2N2N1O2N2N1O2N2N1O2N2N_bf1"}, "label": "a man wearing grey long sleeved shirt"}]}
{"id": 285486, "image": "COCO_train2014_000000285486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an american lady wearing pink color sleeveless t shirt watching laptop\"."}], "task": "refering", "answer_template": "The \"an american lady wearing pink color sleeveless t shirt watching laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.06496874999999999, 0.26527083333333334, 0.383296875, 0.5725416666666666], "iscrowd": 0, "area": 12374.504349999996, "rle": {"size": [480, 640], "counts": "Smc0d0Z>3N2M3N2N2M3N1N3N2N2M3^BhNY=`1N2O1O1N2O1N101O1N2O1N2O1O1N101N2N3N2M2O2M3M2O1N2O1N2O1N2N2O1N2O1N2O2M2N2M3N2M3M3N2M2O2M3N2M30000000000O1000000000I700000000000O10O2O00000L5L300000001O0000001O0000001O0001O01O000000002N2N3M2N3M2N3M2kMYD\\1Z;fMQFk1R<@>B1O2N1O2N1O2N1O10O000000000001O0N2L4M3M3M300O101O000O1000000O101O00VX;c0WgD00000N200000O100000001O001O1O1O1O1O1O00001O00001O0O1N3M2N2Nklh5"}, "label": "an american lady wearing pink color sleeveless t shirt watching laptop"}]}
{"id": 285486, "image": "COCO_train2014_000000285486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with blonde hair using a computer\"."}], "task": "refering", "answer_template": "The \"a woman with blonde hair using a computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.06496874999999999, 0.26527083333333334, 0.383296875, 0.5725416666666666], "iscrowd": 0, "area": 12374.504349999996, "rle": {"size": [480, 640], "counts": "Smc0d0Z>3N2M3N2N2M3N1N3N2N2M3^BhNY=`1N2O1O1N2O1N101O1N2O1N2O1O1N101N2N3N2M2O2M3M2O1N2O1N2O1N2N2O1N2O1N2O2M2N2M3N2M3M3N2M2O2M3N2M30000000000O1000000000I700000000000O10O2O00000L5L300000001O0000001O0000001O0001O01O000000002N2N3M2N3M2N3M2kMYD\\1Z;fMQFk1R<@>B1O2N1O2N1O2N1O10O000000000001O0N2L4M3M3M300O101O000O1000000O101O00VX;c0WgD00000N200000O100000001O001O1O1O1O1O1O00001O00001O0O1N3M2N2Nklh5"}, "label": "a woman with blonde hair using a computer"}]}
{"id": 285486, "image": "COCO_train2014_000000285486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown table near the camera\"."}], "task": "refering", "answer_template": "The \"the brown table near the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 293, 294, 295, 296, 297, 316, 317, 318, 319, 338, 339, 340, 341, 342, 361, 362, 363, 364, 384, 385, 386, 387], "bbox": [0.70621875, 0.6924583333333333, 0.9250468749999999, 1.0], "iscrowd": 0, "area": 12976.07405, "rle": {"size": [480, 640], "counts": "oZd61m>2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2gNY1N2M3N2N2M3M3M3N2M3M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3M3M3N2O1001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000O1M3L4M3L4M3M3L4M3M3L4M3K5L4L4K5L4K6K4K5L4K5L4K5L4L4K5O1N2N2N2N2O1N2N2N2N2O2M2N2N2O1N2N2N2N2O1NaUf0"}, "label": "the brown table near the camera"}]}
{"id": 285486, "image": "COCO_train2014_000000285486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of a brown table can be seen in the bottom of the picture\"."}], "task": "refering", "answer_template": "The \"a piece of a brown table can be seen in the bottom of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 293, 294, 295, 296, 297, 316, 317, 318, 319, 338, 339, 340, 341, 342, 361, 362, 363, 364, 384, 385, 386, 387], "bbox": [0.70621875, 0.6924583333333333, 0.9250468749999999, 1.0], "iscrowd": 0, "area": 12976.07405, "rle": {"size": [480, 640], "counts": "oZd61m>2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2gNY1N2M3N2N2M3M3M3N2M3M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3M3M3N2O1001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000O1M3L4M3L4M3M3L4M3M3L4M3K5L4L4K5L4K6K4K5L4K5L4K5L4L4K5O1N2N2N2N2O1N2N2N2N2O2M2N2N2O1N2N2N2N2O1NaUf0"}, "label": "a piece of a brown table can be seen in the bottom of the picture"}]}
{"id": 15151, "image": "COCO_train2014_000000015151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man singing\"."}], "task": "refering", "answer_template": "The \"man singing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 65, 66, 67, 80, 81, 82, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 234, 245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 290, 291, 292, 293, 305, 306, 307, 308, 309, 320, 321, 322, 323, 324], "bbox": [0.2733568075117371, 0.157375, 0.6592723004694836, 0.93371875], "iscrowd": 0, "area": 41537.93390000002, "rle": {"size": [640, 426], "counts": "n^Y24ic04K4M3L5L3L4cKYO^Ek0a:YOWEk0h:[OPEh0P;]OgDh0W;^OaDe0_;_OYDe0f;ARDb0n;CiCb0U<CdC`0\\<E[C?d<GTC=k<HlB<S=IfB:Z=K]B:a=LWB7i=NnA6R>NgA6W>j200O3N6I7J5J7J6I7bDRI^9S7VF]I`9j6TFeIc9a6QFnIe9Y6nEWJi9Y8G8G:G9aHQEZ5Y;XJmDg5\\;lIhDS6b;_IdD`6e;SI_Dl6\\<^NRCPLn<n3[CjKg<S4bCeK_<Y4iCaKX<\\4lCaKU<]4nCaKT<]4oC`KR<_4QD_KP<`4RD]Ko;b4SD\\Kn;d4TDZKl;e4WDXKj;g4h100O100O100O1O10VO]KWBc4h=aKUB_4f9iL[HlNlM[4\\9\\N`G_MPOU4S9FoFYLKQ4i8Z1TGfN`8g1_GYNT8U2kGjMT8Y2kGgMW8X2hGhM[8V2dGjM^8U2aGkMa8T2^GlMe8R2ZGnMh8P2XGPNk8a62N2N3M2O101O001bJkFMh1jJ^5U5lH3`1lJf5n4lH7U1QKS6c4jH>h0XK`6V4kHc0:`Kn6h3jHj0NfKZ7\\3kHo0BlKe7Q3kHe2X7WMjHh2X7TMjHl2Y7oLjHP3X7lLjHT3X7hLjHX3]9_JfF`5m;01O001O001O00PFWIj6h6gHjIV7V6kHlIS7R6oHQJm6o5SITJk6l5UIWJg6l5WIVJf6l5YIWJd6k5YIXJd6k5ZIXJc6i5\\IYJa6j5]IYJ`6h5^I[J_6h5_I[J^6g5`I[J]6g5bI[JQOWNS6`7iJ\\JQOXNR6_7kJ[JoN[NS6[7mJ\\JlN_NQ6Y7PK[JlN`NP6W7SKZJjNeNn5T7VKYJiNgNm5S7XKXJgNjNm5P7ZKYJfNlNh4hNoKV8a0XJdNoNj3BPMY7a0XJbNROW3OeMj6`0oKX1gMWN]6>QLf0SNkNn5>RL4`N\\Ob5=RLCmNOT59VLTOVOb0f45c0K@Od00^OKf04]OGf08]OBg0=[O^Oi0a0\\9O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O_ej2"}, "label": "man singing"}]}
{"id": 15151, "image": "COCO_train2014_000000015151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lead singer who is in the middle of singing on stage\"."}], "task": "refering", "answer_template": "The \"a lead singer who is in the middle of singing on stage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 65, 66, 67, 80, 81, 82, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 200, 201, 202, 203, 215, 216, 217, 218, 230, 231, 232, 233, 234, 245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 290, 291, 292, 293, 305, 306, 307, 308, 309, 320, 321, 322, 323, 324], "bbox": [0.2733568075117371, 0.157375, 0.6592723004694836, 0.93371875], "iscrowd": 0, "area": 41537.93390000002, "rle": {"size": [640, 426], "counts": "n^Y24ic04K4M3L5L3L4cKYO^Ek0a:YOWEk0h:[OPEh0P;]OgDh0W;^OaDe0_;_OYDe0f;ARDb0n;CiCb0U<CdC`0\\<E[C?d<GTC=k<HlB<S=IfB:Z=K]B:a=LWB7i=NnA6R>NgA6W>j200O3N6I7J5J7J6I7bDRI^9S7VF]I`9j6TFeIc9a6QFnIe9Y6nEWJi9Y8G8G:G9aHQEZ5Y;XJmDg5\\;lIhDS6b;_IdD`6e;SI_Dl6\\<^NRCPLn<n3[CjKg<S4bCeK_<Y4iCaKX<\\4lCaKU<]4nCaKT<]4oC`KR<_4QD_KP<`4RD]Ko;b4SD\\Kn;d4TDZKl;e4WDXKj;g4h100O100O100O1O10VO]KWBc4h=aKUB_4f9iL[HlNlM[4\\9\\N`G_MPOU4S9FoFYLKQ4i8Z1TGfN`8g1_GYNT8U2kGjMT8Y2kGgMW8X2hGhM[8V2dGjM^8U2aGkMa8T2^GlMe8R2ZGnMh8P2XGPNk8a62N2N3M2O101O001bJkFMh1jJ^5U5lH3`1lJf5n4lH7U1QKS6c4jH>h0XK`6V4kHc0:`Kn6h3jHj0NfKZ7\\3kHo0BlKe7Q3kHe2X7WMjHh2X7TMjHl2Y7oLjHP3X7lLjHT3X7hLjHX3]9_JfF`5m;01O001O001O00PFWIj6h6gHjIV7V6kHlIS7R6oHQJm6o5SITJk6l5UIWJg6l5WIVJf6l5YIWJd6k5YIXJd6k5ZIXJc6i5\\IYJa6j5]IYJ`6h5^I[J_6h5_I[J^6g5`I[J]6g5bI[JQOWNS6`7iJ\\JQOXNR6_7kJ[JoN[NS6[7mJ\\JlN_NQ6Y7PK[JlN`NP6W7SKZJjNeNn5T7VKYJiNgNm5S7XKXJgNjNm5P7ZKYJfNlNh4hNoKV8a0XJdNoNj3BPMY7a0XJbNROW3OeMj6`0oKX1gMWN]6>QLf0SNkNn5>RL4`N\\Ob5=RLCmNOT59VLTOVOb0f45c0K@Od00^OKf04]OGf08]OBg0=[O^Oi0a0\\9O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O_ej2"}, "label": "a lead singer who is in the middle of singing on stage"}]}
{"id": 15151, "image": "COCO_train2014_000000015151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man , sitting behind drums\"."}], "task": "refering", "answer_template": "The \"a man , sitting behind drums\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 139, 152, 153, 154, 167, 168, 169, 197, 198, 199, 212, 213, 214, 215, 227, 228, 229, 243], "bbox": [0.15413145539906103, 0.41937499999999994, 0.34762910798122065, 0.717375], "iscrowd": 0, "area": 7147.507750000002, "rle": {"size": [640, 426], "counts": "YeY13kc03L4M3L4oL@^Bb0b=_OZBe0e=[OXBi0b<WOkA0_1n0d<A^ABj1R1e<5aBjNoNU1]>3dBlNjNS1`>2fBlNiNS1^>3iBlNfNR1_>4kBb0R=AmB?R=BnB?P=BPC?o<AQC?o<AQC`0n<@RCa0m<_OSCc0k<]OUCd0j<\\OVCf0h<[OWCg0g<YOYCh0f<XOZCh0f<XOZCi0e<WO[Cj0d<WOZCj0f<VOZCk0e<UO[Cl0d<TO[Cm0e<SO[Cn0d<SO[Cn0d<RO[Co0e<QO[CP1d<PO\\CP1d<PO[CQ1e<POZCo0g<QOYCo0g<QOXCP1h<POXCo0i<ROVCn0j<ROUCo0k<QOUCo0k<POVCf0lNdMo=e1UCf0W=XOjBe0Z=YOgBe0\\=XOeBf0_=XObBf0a=XO`Bh0a=VO_Bj0c=QOaBo0`=kNeBU1]=eNfB\\1]=]NgBc1\\=VNhBQ1oMZO]?]OgBc0mMM10fa02Z^ON2Nda02\\^O02Laa03^^O24H^a05_^O35F\\a05a^O55CZa08b^O48_O[a08_^O9[b0^Oi]Oc0jb00000000O0100000O01N2N5K5K5Ki[]5"}, "label": "a man , sitting behind drums"}]}
{"id": 15151, "image": "COCO_train2014_000000015151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"drummer\"."}], "task": "refering", "answer_template": "The \"drummer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 139, 152, 153, 154, 167, 168, 169, 197, 198, 199, 212, 213, 214, 215, 227, 228, 229, 243], "bbox": [0.15413145539906103, 0.41937499999999994, 0.34762910798122065, 0.717375], "iscrowd": 0, "area": 7147.507750000002, "rle": {"size": [640, 426], "counts": "YeY13kc03L4M3L4oL@^Bb0b=_OZBe0e=[OXBi0b<WOkA0_1n0d<A^ABj1R1e<5aBjNoNU1]>3dBlNjNS1`>2fBlNiNS1^>3iBlNfNR1_>4kBb0R=AmB?R=BnB?P=BPC?o<AQC?o<AQC`0n<@RCa0m<_OSCc0k<]OUCd0j<\\OVCf0h<[OWCg0g<YOYCh0f<XOZCh0f<XOZCi0e<WO[Cj0d<WOZCj0f<VOZCk0e<UO[Cl0d<TO[Cm0e<SO[Cn0d<SO[Cn0d<RO[Co0e<QO[CP1d<PO\\CP1d<PO[CQ1e<POZCo0g<QOYCo0g<QOXCP1h<POXCo0i<ROVCn0j<ROUCo0k<QOUCo0k<POVCf0lNdMo=e1UCf0W=XOjBe0Z=YOgBe0\\=XOeBf0_=XObBf0a=XO`Bh0a=VO_Bj0c=QOaBo0`=kNeBU1]=eNfB\\1]=]NgBc1\\=VNhBQ1oMZO]?]OgBc0mMM10fa02Z^ON2Nda02\\^O02Laa03^^O24H^a05_^O35F\\a05a^O55CZa08b^O48_O[a08_^O9[b0^Oi]Oc0jb00000000O0100000O01N2N5K5K5Ki[]5"}, "label": "drummer"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing eyeglasses sitting on a chair while using a cellphone\"."}], "task": "refering", "answer_template": "The \"man wearing eyeglasses sitting on a chair while using a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 353, 354, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482], "bbox": [0.06468954248366014, 0.03792483660130719, 0.9773366013071896, 0.9643954248366013], "iscrowd": 0, "area": 203670.32194999998, "rle": {"size": [612, 612], "counts": "Z[h01Tc00O2O0O101N1O101N100i@H`;9]DId;8YDIg;8WDIi;8TDJm;7PDJP<7mCKT<5jCLV<6gCKZ<5cCM]<5`CL`<6\\CLe<5XCLh<6UCKl<6PCLP=5nBLS=5iBMW=5fBL[=5aBM_=5^BLc=5ZBLg=5UBMk=5RBLo=5mAMS>5jALW>5fALZ>6bAL_>4_AMb>4ZANf>4WAMj>4SAMm>5o@MR?4k@MU?5g@MZ?4c@M^?4^@Ne<_NQFe1WMMh<_NQFY3o9iLoEW3Q:kLnET3Q:nLoEQ3Q:QMmEo2S:RMmEm2S:UMkEk2U:WMjEh2V:YMiEg2W:[MhEd2W:_MhE`2X:aMgE_2Y:cMfE\\2Z:fMdEZ2\\:gMdEX2\\:hMdEX2\\:hMeEW2Z:kMfET2Z:lMfET2Z:lMgES2Y:mMgES2Y:nMgEQ2Y:oMfER2Z:nMdET2[:nMcES2]:mMaEU2_:kM_EW2a:iM^EX2b:iM[EY2e:gMZEZ2f:fMXE\\2h:eMUE]2j:dMUE]2k:cMSE_2m:aMRE`2n:aMoDa2Q;_MmDc2S;]MlDd2T;]MiDe2V;\\MhDf2X;ZMgDg2Y;YMeDi2[;XMcDi2];WMaDk2_;UM_Dm2a;TM]Dm2b;TM\\Dn2d;RMZDP3f;PMYDQ3g;PMVDR3j;nLTDT3l;lLSDU3m;lLPDV3o;kLoCW3Q<iLnCX3R<hLlCZ3T<gLiC[3W<eLhC\\3X<dLfC^3Z<cLcC_3\\<bLcC_3]<aLaCa3_<b100000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O10O01O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O1O100O1O1mG_Fb6b9^IcF]6]9cIhFW6Y9iImFP6T9PJQGj5P9VJTGf5l8ZJXGa5i8_J[G\\5f8dJ]GX5d8hJ`GT5`8lJdGo4]8QKfGk4[8UKiGf4X8ZKkGc4U8]KmG`4T8`KnG]4S8cKoGZ4R8fKQHW4o7iKSHT4n7lKTHQ4m7oKUHn3l7RLVHl3j7TLXHi3i7WLYH[OeM`3S:UMZHXOdMb3S:VM[HTOeMe3P:WM]HQOeMf3o9YM^HnNeMg3n9[M_HkNdMj3m9[MaHgNeMX1Jo0Q:RObHdNeMS12T1g9UOTIA]MW1`9XOTIYOfM\\1V9[OUIQOnMa1n8^OUIjNVNe1e8AUIcN_Ni1]8DVK:j4FWK8j4HVK7k4IUK7k4IVK5k4KUK5k4KUK4l4LTK4l4LUK2l4NTK2l4NTK1m4OSK1m4OTKOm41SKOm41RKOo41QKOo41QKNP52PKNP52PKMQ53oJMQ53oJLR54nJLR54nJKS55mJKS55mJJT56lJIU57kJHV58jJGW59iJFX5:hJDZ5<fJC[5=mJZOT5f0TKQOm4o0[KhNf4X1c6000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O1O1O1O1O1O4M3L3M4L3M4L3M4M3L3I8jNU1jNW1jNU1jNW1jNV1jNU1jNW1jNU1jNW1jNV1iNV1kNV1iNYZ8"}, "label": "man wearing eyeglasses sitting on a chair while using a cellphone"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt using blue phone\"."}], "task": "refering", "answer_template": "The \"man in blue shirt using blue phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 353, 354, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482], "bbox": [0.06468954248366014, 0.03792483660130719, 0.9773366013071896, 0.9643954248366013], "iscrowd": 0, "area": 203670.32194999998, "rle": {"size": [612, 612], "counts": "Z[h01Tc00O2O0O101N1O101N100i@H`;9]DId;8YDIg;8WDIi;8TDJm;7PDJP<7mCKT<5jCLV<6gCKZ<5cCM]<5`CL`<6\\CLe<5XCLh<6UCKl<6PCLP=5nBLS=5iBMW=5fBL[=5aBM_=5^BLc=5ZBLg=5UBMk=5RBLo=5mAMS>5jALW>5fALZ>6bAL_>4_AMb>4ZANf>4WAMj>4SAMm>5o@MR?4k@MU?5g@MZ?4c@M^?4^@Ne<_NQFe1WMMh<_NQFY3o9iLoEW3Q:kLnET3Q:nLoEQ3Q:QMmEo2S:RMmEm2S:UMkEk2U:WMjEh2V:YMiEg2W:[MhEd2W:_MhE`2X:aMgE_2Y:cMfE\\2Z:fMdEZ2\\:gMdEX2\\:hMdEX2\\:hMeEW2Z:kMfET2Z:lMfET2Z:lMgES2Y:mMgES2Y:nMgEQ2Y:oMfER2Z:nMdET2[:nMcES2]:mMaEU2_:kM_EW2a:iM^EX2b:iM[EY2e:gMZEZ2f:fMXE\\2h:eMUE]2j:dMUE]2k:cMSE_2m:aMRE`2n:aMoDa2Q;_MmDc2S;]MlDd2T;]MiDe2V;\\MhDf2X;ZMgDg2Y;YMeDi2[;XMcDi2];WMaDk2_;UM_Dm2a;TM]Dm2b;TM\\Dn2d;RMZDP3f;PMYDQ3g;PMVDR3j;nLTDT3l;lLSDU3m;lLPDV3o;kLoCW3Q<iLnCX3R<hLlCZ3T<gLiC[3W<eLhC\\3X<dLfC^3Z<cLcC_3\\<bLcC_3]<aLaCa3_<b100000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O10O01O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O1O100O1O1mG_Fb6b9^IcF]6]9cIhFW6Y9iImFP6T9PJQGj5P9VJTGf5l8ZJXGa5i8_J[G\\5f8dJ]GX5d8hJ`GT5`8lJdGo4]8QKfGk4[8UKiGf4X8ZKkGc4U8]KmG`4T8`KnG]4S8cKoGZ4R8fKQHW4o7iKSHT4n7lKTHQ4m7oKUHn3l7RLVHl3j7TLXHi3i7WLYH[OeM`3S:UMZHXOdMb3S:VM[HTOeMe3P:WM]HQOeMf3o9YM^HnNeMg3n9[M_HkNdMj3m9[MaHgNeMX1Jo0Q:RObHdNeMS12T1g9UOTIA]MW1`9XOTIYOfM\\1V9[OUIQOnMa1n8^OUIjNVNe1e8AUIcN_Ni1]8DVK:j4FWK8j4HVK7k4IUK7k4IVK5k4KUK5k4KUK4l4LTK4l4LUK2l4NTK2l4NTK1m4OSK1m4OTKOm41SKOm41RKOo41QKOo41QKNP52PKNP52PKMQ53oJMQ53oJLR54nJLR54nJKS55mJKS55mJJT56lJIU57kJHV58jJGW59iJFX5:hJDZ5<fJC[5=mJZOT5f0TKQOm4o0[KhNf4X1c6000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O1O1O1O1O1O4M3L3M4L3M4L3M4M3L3I8jNU1jNW1jNU1jNW1jNV1jNU1jNW1jNU1jNW1jNV1iNV1kNV1iNYZ8"}, "label": "man in blue shirt using blue phone"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black jerkin on chair\"."}], "task": "refering", "answer_template": "The \"a black jerkin on chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 114, 115, 116, 117, 118, 136, 137, 138, 139, 140, 141, 142, 158, 159, 160, 161, 162, 163, 164, 181, 182, 183, 184, 185, 186, 187, 204, 205, 206, 208, 209, 228], "bbox": [0.1865359477124183, 0.03398692810457517, 0.520718954248366, 0.5080718954248366], "iscrowd": 0, "area": 29465.2757, "rle": {"size": [612, 612], "counts": "XYT2Z1ea05K6J6J6J5K6J6J5K6J6J6J5K6J6I6J7I7H8I6J7I7I6J7L4N2O001O1O0O2O1O001O1N2O001O1O001N2O1O001O1N101O1O1O0O2O1O001O1N2O001O1O0O2O1O1O001N2O001O1O1N101O1O001N2O1O001O1O001OO2N1O1O1O1O1N3N1O1O1O1O1O2N1O1O1N2O1O2N1O1O1O1O1O2M2O1O1M3J7H7IjC_JW;[5jDYKg:`4[ETLV:e3mEnLd9l2]FgMV9Q2lF[Nm8^1UGfNl8T1UGPOl8i0VG[Ok8>WGFj84ZGMh8K^G5c8DbG=_8]OfGb0\\8WOjGi0W8POnGQ1S8hNSHX1o7aNVH_1k7ZN[Hf1f7SN_Hm1c7mMbHj1h7oM^Hh1l7QNYHg1^8fMgGU2f<K5K4L6K5J6M301N1O1O1O101N1O1O1O2O0O1O1O1O2O01O0001O01O0000010O0001002N3M2M3N3M2N3M2M4M2N2N1O0O22M3N3L3M3N2M3N3L3N2M3M3L4K6J4K5L4L3M4L[k^5"}, "label": "a black jerkin on chair"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back jacket around the back of a wooden chair at the table\"."}], "task": "refering", "answer_template": "The \"back jacket around the back of a wooden chair at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 114, 115, 116, 117, 118, 136, 137, 138, 139, 140, 141, 142, 158, 159, 160, 161, 162, 163, 164, 181, 182, 183, 184, 185, 186, 187, 204, 205, 206, 208, 209, 228], "bbox": [0.1865359477124183, 0.03398692810457517, 0.520718954248366, 0.5080718954248366], "iscrowd": 0, "area": 29465.2757, "rle": {"size": [612, 612], "counts": "XYT2Z1ea05K6J6J6J5K6J6J5K6J6J6J5K6J6I6J7I7H8I6J7I7I6J7L4N2O001O1O0O2O1O001O1N2O001O1O001N2O1O001O1N101O1O1O0O2O1O001O1N2O001O1O0O2O1O1O001N2O001O1O1N101O1O001N2O1O001O1O001OO2N1O1O1O1O1N3N1O1O1O1O1O2N1O1O1N2O1O2N1O1O1O1O1O2M2O1O1M3J7H7IjC_JW;[5jDYKg:`4[ETLV:e3mEnLd9l2]FgMV9Q2lF[Nm8^1UGfNl8T1UGPOl8i0VG[Ok8>WGFj84ZGMh8K^G5c8DbG=_8]OfGb0\\8WOjGi0W8POnGQ1S8hNSHX1o7aNVH_1k7ZN[Hf1f7SN_Hm1c7mMbHj1h7oM^Hh1l7QNYHg1^8fMgGU2f<K5K4L6K5J6M301N1O1O1O101N1O1O1O2O0O1O1O1O2O01O0001O01O0000010O0001002N3M2M3N3M2N3M2M4M2N2N1O0O22M3N3L3M3N2M3N3L3N2M3M3L4K6J4K5L4L3M4L[k^5"}, "label": "back jacket around the back of a wooden chair at the table"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black coat hanging on a chair\"."}], "task": "refering", "answer_template": "The \"a black coat hanging on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 113, 114, 115, 116, 117, 118, 135, 136, 137, 138, 139, 140, 157, 158, 159, 160, 161, 162, 181, 182, 183, 184, 202, 203, 204, 205, 223, 224, 225, 226, 246, 247], "bbox": [0.16367647058823528, 0.024918300653594773, 0.4103104575163399, 0.5069934640522875], "iscrowd": 0, "area": 31896.3473, "rle": {"size": [612, 612], "counts": "`mk15lb0c0]Oc0]O;aA[Nl:h1REZNk:j1REXNk:k1SEWNj:m1SEUNj:n1TETNi:P2TERNi:Q2UEQNh:S2UEoMh:T2VEnMg:V2VElMg:W2WEkMf:Y2WEiMf:Z2XEiMd:[2ZEfMc:]2]EcM_:b2_E_M\\:f2dEZMW:l2gEUMS:Q3mEoLn9W3QFiLj9\\3UFeLe9b3ZF^La9g3_FYL\\9m3bFTLX9R4hFnKS9X4kFiKo8]4QGcKj8c4UG]Kf8h4YGYKa8n4^GRK]8S5bGnJY8X5fGhJX8Z5hGfJX8[5fGeJ[8[5eGeJ[8\\5dGdJ\\8\\5cGeJ]8\\5bGdJ_8[5`GfJ`8[5_GdJb8]5]GcJc8^5[GcJe8_5YGaJg8`5WGaJi8a5UG_Jk8b5RG_Jo8c5nF^JR9c5kF_JU9c5gF_JY9b5eF_J[9c5aF_J_9b5_F^Jb9d5ZF^Jf9c5XF^Jh9d5TF^Jl9c5RF^Jn9\\70O10001O00000O10000000000O100O1O1O1O1O1O1N2O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1N2N2O1N2N2N2N2N2N3N1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2O1N4L4L4L4L4L4L4L4L4K5J6J6I7K5K6J5K5K5L4K5K5aM\\A8h>@jA2^>BSB0nSh6"}, "label": "a black coat hanging on a chair"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black jacket hanging on a chair behind a man in a blue shirt on his phone\"."}], "task": "refering", "answer_template": "The \"a black jacket hanging on a chair behind a man in a blue shirt on his phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 113, 114, 115, 116, 117, 118, 135, 136, 137, 138, 139, 140, 157, 158, 159, 160, 161, 162, 181, 182, 183, 184, 202, 203, 204, 205, 223, 224, 225, 226, 246, 247], "bbox": [0.16367647058823528, 0.024918300653594773, 0.4103104575163399, 0.5069934640522875], "iscrowd": 0, "area": 31896.3473, "rle": {"size": [612, 612], "counts": "`mk15lb0c0]Oc0]O;aA[Nl:h1REZNk:j1REXNk:k1SEWNj:m1SEUNj:n1TETNi:P2TERNi:Q2UEQNh:S2UEoMh:T2VEnMg:V2VElMg:W2WEkMf:Y2WEiMf:Z2XEiMd:[2ZEfMc:]2]EcM_:b2_E_M\\:f2dEZMW:l2gEUMS:Q3mEoLn9W3QFiLj9\\3UFeLe9b3ZF^La9g3_FYL\\9m3bFTLX9R4hFnKS9X4kFiKo8]4QGcKj8c4UG]Kf8h4YGYKa8n4^GRK]8S5bGnJY8X5fGhJX8Z5hGfJX8[5fGeJ[8[5eGeJ[8\\5dGdJ\\8\\5cGeJ]8\\5bGdJ_8[5`GfJ`8[5_GdJb8]5]GcJc8^5[GcJe8_5YGaJg8`5WGaJi8a5UG_Jk8b5RG_Jo8c5nF^JR9c5kF_JU9c5gF_JY9b5eF_J[9c5aF_J_9b5_F^Jb9d5ZF^Jf9c5XF^Jh9d5TF^Jl9c5RF^Jn9\\70O10001O00000O10000000000O100O1O1O1O1O1O1N2O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1N2N2O1N2N2N2N2N2N3N1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2O1N4L4L4L4L4L4L4L4L4K5J6J6I7K5K6J5K5K5L4K5K5aM\\A8h>@jA2^>BSB0nSh6"}, "label": "a black jacket hanging on a chair behind a man in a blue shirt on his phone"}]}
{"id": 383807, "image": "COCO_train2014_000000383807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wall\"."}], "task": "refering", "answer_template": "The \"wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 22, 23, 24, 25, 26, 44, 45, 46, 47, 48, 66, 67, 68, 69, 70, 88, 89, 90, 91, 92, 110, 111, 112, 113, 132, 133, 134, 135, 154, 155, 156, 157, 176, 177, 178, 179, 198, 199, 200, 220, 221, 242], "bbox": [0.03511437908496732, 0.030964052287581697, 0.21797385620915033, 0.5116830065359477], "iscrowd": 0, "area": 22642.788750000003, "rle": {"size": [612, 612], "counts": "[j<R1e:\\7000000O1000000O100000000O10000001O000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UNRD_Lo;_3ZDYLg;e3bDSL_;k3jDmKW;Q4REgKo:W4ZEaKh:[4aE^K`:`4S2N2N2N2N3L3M3M3M3M3M3M4L3M5K6J7J5J6J7I6J6J7I6J6J7I6J6J6K6I6JbVn8"}, "label": "wall"}]}
{"id": 179011, "image": "COCO_train2014_000000179011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of pizza with a server on it\"."}], "task": "refering", "answer_template": "The \"a piece of pizza with a server on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.38, 0.12322077922077922, 0.94665625, 0.7616103896103896], "iscrowd": 0, "area": 58236.39739999999, "rle": {"size": [385, 640], "counts": "obk21l;4M4L4L4L3M4L3M3L5N1N2O1O2M2O1N2O2N1N2O1N3N1O1N2O2M2O1O1N3N1N2O1O2M2O1N2O2N1N2O1N3N1N2O1O2M2O1N2O2M2O1O1N3N1N2O1N3N1O1N2O2M2O1N2O2N1N2M3M4M2M3M3N3L3N2N2N3L3N2N2N3N100O101O000O101O0O1O1O2N1N2O1O2M2O1N2O2N1N2O1N3N1O1N2O2M2O1O101O0000001O0O10001O00001O000O101O0000001O000O101O000003N2M2N1O01N100O101N100O101N100O101N100O101N100O101O0O100O2O0O100O2O0O100O2O0O101N11O0001O0001O01O0001O01O0000010O000001O0000001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O001O1O1O001O00001O00001O00001O00001O00001O00001O0000001O00001O00002N1O1O1O1O2O0O1O1O1O2N1O1O1O2N1O1O1O1O2N1O001O01O0001O0000001O002N1O2TK_IT4b6iKaIV4`6hKbIW4`6eKbI[4_6cKcI\\4_6`KdI`4\\6^KfIa4m601O1O1O001O1O1O1O001O1O1O1O1O1O1O1XL[HR3f7iLdHQ3]7jLlHQ3R80001O00001O00001O00001O00001O001O1O1O2N1O1O1O1O2N1O1O1O1O2ZNdFc0]9XOiFf0\\:M2N3M2N2N3M2N3M2N2N3M2N3M8HQ`<"}, "label": "a piece of pizza with a server on it"}]}
{"id": 179011, "image": "COCO_train2014_000000179011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a slice of pizza with a serving tool on top is shown\"."}], "task": "refering", "answer_template": "The \"a slice of pizza with a serving tool on top is shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.38, 0.12322077922077922, 0.94665625, 0.7616103896103896], "iscrowd": 0, "area": 58236.39739999999, "rle": {"size": [385, 640], "counts": "obk21l;4M4L4L4L3M4L3M3L5N1N2O1O2M2O1N2O2N1N2O1N3N1O1N2O2M2O1O1N3N1N2O1O2M2O1N2O2N1N2O1N3N1N2O1O2M2O1N2O2M2O1O1N3N1N2O1N3N1O1N2O2M2O1N2O2N1N2M3M4M2M3M3N3L3N2N2N3L3N2N2N3N100O101O000O101O0O1O1O2N1N2O1O2M2O1N2O2N1N2O1N3N1O1N2O2M2O1O101O0000001O0O10001O00001O000O101O0000001O000O101O000003N2M2N1O01N100O101N100O101N100O101N100O101N100O101O0O100O2O0O100O2O0O100O2O0O101N11O0001O0001O01O0001O01O0000010O000001O0000001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O001O1O1O001O00001O00001O00001O00001O00001O00001O0000001O00001O00002N1O1O1O1O2O0O1O1O1O2N1O1O1O2N1O1O1O1O2N1O001O01O0001O0000001O002N1O2TK_IT4b6iKaIV4`6hKbIW4`6eKbI[4_6cKcI\\4_6`KdI`4\\6^KfIa4m601O1O1O001O1O1O1O001O1O1O1O1O1O1O1XL[HR3f7iLdHQ3]7jLlHQ3R80001O00001O00001O00001O00001O001O1O1O2N1O1O1O1O2N1O1O1O1O2ZNdFc0]9XOiFf0\\:M2N3M2N2N3M2N3M2N2N3M2N3M8HQ`<"}, "label": "a slice of pizza with a serving tool on top is shown"}]}
{"id": 179011, "image": "COCO_train2014_000000179011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large portion of pizza leftover\"."}], "task": "refering", "answer_template": "The \"large portion of pizza leftover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.043265625, 0.11685714285714287, 0.91653125, 0.9438181818181818], "iscrowd": 0, "area": 92245.45014999998, "rle": {"size": [385, 640], "counts": "mg:c1\\9]1QNo1`IYKX3LhLP7P3V1K5L0OO1O1N2O1O1OO1O1N101O100O100O100O100O1O100O100O1O100O100O100O10000O100O10000O2O000O100O10000O10000O100O100O10000O100O100O100O100O10000O100O100O10000O100O100O100O10000O1O2O0O100O1O100O100O1O100000000O1000000000000000000O10000000001O0001O00000000001O000001O0001O00000000001O0001O0001O1O001O1O001O000001O0000001O0001O00000000000000000O10000000000000001O00001XLgMfMY2W2jMiMV2U2lMkMT2S2oMlMR2R2oMnMQ2P2QNPNo1o1RNQNn1m1TNSNl1l1VNSNj1k1XNUNh1j1YNVNh1g1ZNYNf1f1\\NYNd1f1]NZNc1e1^N[Nb1d1`N[N`1d1aN\\N_1b1cN^N]1a1dN_N]1_1dNaN\\1^1eNbN[1]1fNcNZ1\\1gNdNY1[1hNeNX1Z1iNfNW1X1kNiNU1U1lNkNT1S1nNmNQ1R1QOnNo0P1SOPOm0P1SOPOm0o0TOQOl0n0UOROk0m0WOROi0m0XOSOh0l0YOSOh0l0YOTOf0l0[OTOe0k0\\OUOd0i0^OWOb0g0@YO`0e0B[O>c0D]O<a0F_O:`0G@9>IA9=HC8<ID8:JE6:LE58NG28OH260I053JN44KL45KM34ML16OKO70IO81HN92HK:5FJ;6FH;8EG<9DE?9CE>;BC`0=A@a0`0_O^Oc0b0]O\\Oe0d0\\OXOg0h0YOVOi0j0WOTOk0l0UOQOn0o0ROnNQ1Q1POmNR1S1nNkNT1U1lNiNV1W1jNgNX1Y1hNeN[1Z1dNeN^1[1bNdN`1[1_NeNc1Z1]NdNe1\\1ZNdNh1[1XNdNj1[1UNeNm1Z1SNdNo1]1PNbNR2]1mMcNU2\\1kMbNW2^1iMaNY2^1gMaN[2^1eMaN\\2_1dM_N_2`1aM^Na2b1_M\\Nd2c1\\M[Nf2e1[MWNh2i1YMSNk2l1WMPNk2P2VMlMm2T2a3O00000001O000000001O0000000001O00001O001O00001O000000001O2N1O1O0000001O000000000000O10000000000O100000000O10000000000000000O1000000000000000000001O00001O00001O00001O00001O000O2O001O001O001O00000000000000001O001O1O1O001N2O00000000000000001O00000000O100O100O100O100O1001O000000000000001O003M000000000000000000000000000000O1001O0000001O0000001O000000000000001O000000000001O00000000000001O000000000000000000000001O00000000000001O00001O0000001O0000000001O00O1000001O000O101O0000001N1000001N1O1O101N1O1O2N1O1O2N1O1O1O2M2O1O2M2O1O2N1N2O2N1O1N3N3M4Ldoc0"}, "label": "large portion of pizza leftover"}]}
{"id": 179011, "image": "COCO_train2014_000000179011.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza shown\"."}], "task": "refering", "answer_template": "The \"the pizza shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.043265625, 0.11685714285714287, 0.91653125, 0.9438181818181818], "iscrowd": 0, "area": 92245.45014999998, "rle": {"size": [385, 640], "counts": "mg:c1\\9]1QNo1`IYKX3LhLP7P3V1K5L0OO1O1N2O1O1OO1O1N101O100O100O100O100O1O100O100O1O100O100O100O10000O100O10000O2O000O100O10000O10000O100O100O10000O100O100O100O100O10000O100O100O10000O100O100O100O10000O1O2O0O100O1O100O100O1O100000000O1000000000000000000O10000000001O0001O00000000001O000001O0001O00000000001O0001O0001O1O001O1O001O000001O0000001O0001O00000000000000000O10000000000000001O00001XLgMfMY2W2jMiMV2U2lMkMT2S2oMlMR2R2oMnMQ2P2QNPNo1o1RNQNn1m1TNSNl1l1VNSNj1k1XNUNh1j1YNVNh1g1ZNYNf1f1\\NYNd1f1]NZNc1e1^N[Nb1d1`N[N`1d1aN\\N_1b1cN^N]1a1dN_N]1_1dNaN\\1^1eNbN[1]1fNcNZ1\\1gNdNY1[1hNeNX1Z1iNfNW1X1kNiNU1U1lNkNT1S1nNmNQ1R1QOnNo0P1SOPOm0P1SOPOm0o0TOQOl0n0UOROk0m0WOROi0m0XOSOh0l0YOSOh0l0YOTOf0l0[OTOe0k0\\OUOd0i0^OWOb0g0@YO`0e0B[O>c0D]O<a0F_O:`0G@9>IA9=HC8<ID8:JE6:LE58NG28OH260I053JN44KL45KM34ML16OKO70IO81HN92HK:5FJ;6FH;8EG<9DE?9CE>;BC`0=A@a0`0_O^Oc0b0]O\\Oe0d0\\OXOg0h0YOVOi0j0WOTOk0l0UOQOn0o0ROnNQ1Q1POmNR1S1nNkNT1U1lNiNV1W1jNgNX1Y1hNeN[1Z1dNeN^1[1bNdN`1[1_NeNc1Z1]NdNe1\\1ZNdNh1[1XNdNj1[1UNeNm1Z1SNdNo1]1PNbNR2]1mMcNU2\\1kMbNW2^1iMaNY2^1gMaN[2^1eMaN\\2_1dM_N_2`1aM^Na2b1_M\\Nd2c1\\M[Nf2e1[MWNh2i1YMSNk2l1WMPNk2P2VMlMm2T2a3O00000001O000000001O0000000001O00001O001O00001O000000001O2N1O1O0000001O000000000000O10000000000O100000000O10000000000000000O1000000000000000000001O00001O00001O00001O00001O000O2O001O001O001O00000000000000001O001O1O1O001N2O00000000000000001O00000000O100O100O100O100O1001O000000000000001O003M000000000000000000000000000000O1001O0000001O0000001O000000000000001O000000000001O00000000000001O000000000000000000000001O00000000000001O00001O0000001O0000000001O00O1000001O000O101O0000001N1000001N1O1O101N1O1O2N1O1O2N1O1O1O2M2O1O2M2O1O2N1N2O2N1O1N3N3M4Ldoc0"}, "label": "the pizza shown"}]}
{"id": 531277, "image": "COCO_train2014_000000531277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey table top\"."}], "task": "refering", "answer_template": "The \"grey table top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 100, 101, 117, 118], "bbox": [0.3630543933054393, 0.0705, 0.9984728033472803, 0.29140625], "iscrowd": 0, "area": 16294.857449999996, "rle": {"size": [640, 478], "counts": "aj\\31oc01O1N2O1N2O1O1N101N2O1O1N2O1N101O1N2O1O1O1O1O0O10000000000O10O10000000O100000O10O10000000000O10000000000O1000000000O0O2O10000000O1000O11O2N1N2O2N0000000O100000000O10O10000000O100000O1000O10000000O10O1000000000O01000000000O010000000O010O10000O10O10O10000O10O10O1001O1O0O2O001O00001N101]]OhN\\b0^1000000O100000O10O10N200O100O010O1O100O1O0100000O1000O10000000000O10000000000O01000000000O5L2N010O1O1O010O0010O01O0010O01O1O010O001O010O001O010O001O010O001O10O03M3N9F1O0001O0000010O01O0010O0001O010O000010O01O00100O1O1O010O1O100O1O1O10O01O3M3N2M2N3M101N2N1O2O0O1O0010O01O00010O000010O00010O0001O01O01O0010O0001O01O01O000fa0"}, "label": "grey table top"}]}
{"id": 531277, "image": "COCO_train2014_000000531277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the corner of the gray table visible to the right of the hand\"."}], "task": "refering", "answer_template": "The \"the corner of the gray table visible to the right of the hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 100, 101, 117, 118], "bbox": [0.3630543933054393, 0.0705, 0.9984728033472803, 0.29140625], "iscrowd": 0, "area": 16294.857449999996, "rle": {"size": [640, 478], "counts": "aj\\31oc01O1N2O1N2O1O1N101N2O1O1N2O1N101O1N2O1O1O1O1O0O10000000000O10O10000000O100000O10O10000000000O10000000000O1000000000O0O2O10000000O1000O11O2N1N2O2N0000000O100000000O10O10000000O100000O1000O10000000O10O1000000000O01000000000O010000000O010O10000O10O10O10000O10O10O1001O1O0O2O001O00001N101]]OhN\\b0^1000000O100000O10O10N200O100O010O1O100O1O0100000O1000O10000000000O10000000000O01000000000O5L2N010O1O1O010O0010O01O0010O01O1O010O001O010O001O010O001O010O001O10O03M3N9F1O0001O0000010O01O0010O0001O010O000010O01O00100O1O1O010O1O100O1O1O10O01O3M3N2M2N3M101N2N1O2O0O1O0010O01O00010O000010O00010O0001O01O01O0010O0001O01O01O000fa0"}, "label": "the corner of the gray table visible to the right of the hand"}]}
{"id": 531277, "image": "COCO_train2014_000000531277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white plate which contains nut dish\"."}], "task": "refering", "answer_template": "The \"a white plate which contains nut dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 219, 220, 236, 237, 238, 250, 251, 252, 253, 254, 255, 256, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.5116875000000001, 1.0, 0.9870156250000001], "iscrowd": 0, "area": 96611.49670000002, "rle": {"size": [640, 478], "counts": "Zc0>hb0j0UOk0VOj0UOk0VOj0UOk0VOj0XOh0001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1OO1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1N2O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O11O0000001O0000001O0000001O0000001O0000f4ZKn@"}, "label": "a white plate which contains nut dish"}]}
{"id": 531277, "image": "COCO_train2014_000000531277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white round plate\"."}], "task": "refering", "answer_template": "The \"white round plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 219, 220, 236, 237, 238, 250, 251, 252, 253, 254, 255, 256, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.5116875000000001, 1.0, 0.9870156250000001], "iscrowd": 0, "area": 96611.49670000002, "rle": {"size": [640, 478], "counts": "Zc0>hb0j0UOk0VOj0UOk0VOj0UOk0VOj0XOh0001O001O001O001O001O001O001O001O001O001O001O001O001O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1OO1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1N2O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O11O0000001O0000001O0000001O0000001O0000f4ZKn@"}, "label": "white round plate"}]}
{"id": 547662, "image": "COCO_train2014_000000547662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light green plant with long leaf ' s\"."}], "task": "refering", "answer_template": "The \"a light green plant with long leaf ' s\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 99, 100, 117, 118], "bbox": [0.47254, 0.0, 1.0, 0.5336266666666667], "iscrowd": 0, "area": 25747.222350000004, "rle": {"size": [375, 500], "counts": "mjf23c;2N2ZOOcE4[:2^E0a:O_E2a:N^E3a:M`E2a:N^E2c:N\\E2e:M[E3f:MYE3h:MWE3l1Jh62\\G=^1KR7H_Gi0n0I^7_OcGW1;Oc7iNRHg3\\7YLdHe3^781N3N1N2O`ISLZ5k3cJZL\\5T4SJoKn5m4O100O0O2O1O000000a0_O1O001O1O0O2O1OO1M3M3hN_JXLe5h3fJhK_5Y4o0001O001O1O1O1O2N1O1O1O1O1O1O1O001O1OF:E;E;L43M3M3M3M3M3M3M3ML4M4K4L4M3L4N21O1O1O1O1O1O1O1O1O1O1O1O1OL4K5J6K5K52N2N3M2N2N2N3M2N2N2N3M2N2N3MZNPJgMn5Y2h1O1O1O1O1O1O1O14L>B?A>B2N1O1O1O1O1O001ON2N2O1N2N2N2iNVI`Mk6\\2Z1L4M3L5N12N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2NkNfGoNX8Q1oGhNQ8W1WHbNi7^1^H[Nb7e1eHTN[7l1X10000O1001O001O1O1O001O1O5K5K5K4L4L3M4LN2G9gNY1J60000000000000000001O1O1O1O1O1Oe0[O6J6J7I2N2N100ON2O1SOm0L4M3N300O1O2N101N1O2N1O2O0O2UObE9_:EcE;]:CdE=^:@dE?]:_OeE`0]:^OcEc0]:[OeEd0h:O10O01O1O10O01O100O001O100O004L00001O0000XN"}, "label": "a light green plant with long leaf ' s"}]}
{"id": 547662, "image": "COCO_train2014_000000547662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant next to the bicycle\"."}], "task": "refering", "answer_template": "The \"plant next to the bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 16, 17, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 99, 100, 117, 118], "bbox": [0.47254, 0.0, 1.0, 0.5336266666666667], "iscrowd": 0, "area": 25747.222350000004, "rle": {"size": [375, 500], "counts": "mjf23c;2N2ZOOcE4[:2^E0a:O_E2a:N^E3a:M`E2a:N^E2c:N\\E2e:M[E3f:MYE3h:MWE3l1Jh62\\G=^1KR7H_Gi0n0I^7_OcGW1;Oc7iNRHg3\\7YLdHe3^781N3N1N2O`ISLZ5k3cJZL\\5T4SJoKn5m4O100O0O2O1O000000a0_O1O001O1O0O2O1OO1M3M3hN_JXLe5h3fJhK_5Y4o0001O001O1O1O1O2N1O1O1O1O1O1O1O001O1OF:E;E;L43M3M3M3M3M3M3M3ML4M4K4L4M3L4N21O1O1O1O1O1O1O1O1O1O1O1O1OL4K5J6K5K52N2N3M2N2N2N3M2N2N2N3M2N2N3MZNPJgMn5Y2h1O1O1O1O1O1O1O14L>B?A>B2N1O1O1O1O1O001ON2N2O1N2N2N2iNVI`Mk6\\2Z1L4M3L5N12N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2NkNfGoNX8Q1oGhNQ8W1WHbNi7^1^H[Nb7e1eHTN[7l1X10000O1001O001O1O1O001O1O5K5K5K4L4L3M4LN2G9gNY1J60000000000000000001O1O1O1O1O1Oe0[O6J6J7I2N2N100ON2O1SOm0L4M3N300O1O2N101N1O2N1O2O0O2UObE9_:EcE;]:CdE=^:@dE?]:_OeE`0]:^OcEc0]:[OeEd0h:O10O01O1O10O01O100O001O100O004L00001O0000XN"}, "label": "plant next to the bicycle"}]}
{"id": 547662, "image": "COCO_train2014_000000547662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plant in the red pot\"."}], "task": "refering", "answer_template": "The \"the plant in the red pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 67, 68, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 197], "bbox": [0.77022, 0.16112, 0.9994400000000001, 0.7948266666666667], "iscrowd": 0, "area": 17607.98580000001, "rle": {"size": [375, 500], "counts": "fR]42e;4K5K6J5L4K4L1101O1O001O1O1N101O1O1O00N101SOZOfFh0X9\\OeFe0Y9_OcFe0Y9_OdFd0X9@dFe0?jNm7e0`Ge0<kNS8c0^Gf08kNX8V2bGoM\\8g25K6J5LWHcM[6W2WI[Ne6e1mHnNn6S3J6K2N0O100O1N2M3L4L4TOhJTK\\5k4iJnJZ5P5kJiJY5V5lJbJX5\\5b0O2O1N2O1N2N2O1N26J<De0[ON2N101N1O2N1O1O2N1O1N3]Ob0D=N10001O0O101O001N10001N110O01O10O01O010O010O1O010O00100O10O01M3N2M3M2O2M3M3N`4_KUI"}, "label": "the plant in the red pot"}]}
{"id": 547662, "image": "COCO_train2014_000000547662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a earth tone flower pot with a green bush in it\"."}], "task": "refering", "answer_template": "The \"a earth tone flower pot with a green bush in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 67, 68, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 177, 178, 179, 197], "bbox": [0.77022, 0.16112, 0.9994400000000001, 0.7948266666666667], "iscrowd": 0, "area": 17607.98580000001, "rle": {"size": [375, 500], "counts": "fR]42e;4K5K6J5L4K4L1101O1O001O1O1N101O1O1O00N101SOZOfFh0X9\\OeFe0Y9_OcFe0Y9_OdFd0X9@dFe0?jNm7e0`Ge0<kNS8c0^Gf08kNX8V2bGoM\\8g25K6J5LWHcM[6W2WI[Ne6e1mHnNn6S3J6K2N0O100O1N2M3L4L4TOhJTK\\5k4iJnJZ5P5kJiJY5V5lJbJX5\\5b0O2O1N2O1N2N2O1N26J<De0[ON2N101N1O2N1O1O2N1O1N3]Ob0D=N10001O0O101O001N10001N110O01O10O01O010O010O1O010O00100O10O01M3N2M3M2O2M3M3N`4_KUI"}, "label": "a earth tone flower pot with a green bush in it"}]}
{"id": 47949, "image": "COCO_train2014_000000047949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bunch of yellow bananas behind the plate with a gooey sandwich on it\"."}], "task": "refering", "answer_template": "The \"a bunch of yellow bananas behind the plate with a gooey sandwich on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100], "bbox": [0.05225, 0.0015625, 0.566296875, 0.27347916666666666], "iscrowd": 0, "area": 32545.715099999998, "rle": {"size": [480, 640], "counts": "d`?1j>:F:F;E:F:F:F9G6J7K5N1O2N1O2N2N1O2M3N1O2N2N1O2N1O2N2N1O2N2N001N101O00001O001O001O001O001O000O100000000000O1000000000000000001N1000000000001O00000O1000001O00000000000O2O0000000000001O000O100000001O0000001O0O10001O000000001O000O101O0000001O0000001N1000001O000000001O0O10001O0000001O000O101O000000001O00000O2O00000000001O00000O1000000000001O00000O1000000000001O00000O1000000000001O00000O1000000000001O00000O10000000001O0000000O10000000001O0000000O10000000001O000000O01000O10000O10000O10O10O10000O10000O1000O010000O10000O10000O01000O10000O10000O10O10O10000O10000O1000O010000O10000O10000O01000O10000O10000O10YOXDUNi;j1XDVNg;i1[DWNe;h1\\DXNd;g1]DYNb;f1aDYN_;f1bDZN^;d1dD\\N[;d1fD\\NZ;c1gD]NY;a1iD^N[;^1fDaN];[1fD`N`;[1cDaNd;Y1_DcNg;X1T1K5K5K5L4M3L3N3M3L4M2MQYR4"}, "label": "a bunch of yellow bananas behind the plate with a gooey sandwich on it"}]}
{"id": 47949, "image": "COCO_train2014_000000047949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three yellow bananas\"."}], "task": "refering", "answer_template": "The \"three yellow bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100], "bbox": [0.05225, 0.0015625, 0.566296875, 0.27347916666666666], "iscrowd": 0, "area": 32545.715099999998, "rle": {"size": [480, 640], "counts": "d`?1j>:F:F;E:F:F:F9G6J7K5N1O2N1O2N2N1O2M3N1O2N2N1O2N1O2N2N1O2N2N001N101O00001O001O001O001O001O000O100000000000O1000000000000000001N1000000000001O00000O1000001O00000000000O2O0000000000001O000O100000001O0000001O0O10001O000000001O000O101O0000001O0000001N1000001O000000001O0O10001O0000001O000O101O000000001O00000O2O00000000001O00000O1000000000001O00000O1000000000001O00000O1000000000001O00000O1000000000001O00000O10000000001O0000000O10000000001O0000000O10000000001O000000O01000O10000O10000O10O10O10000O10000O1000O010000O10000O10000O01000O10000O10000O10O10O10000O10000O1000O010000O10000O10000O01000O10000O10000O10YOXDUNi;j1XDVNg;i1[DWNe;h1\\DXNd;g1]DYNb;f1aDYN_;f1bDZN^;d1dD\\N[;d1fD\\NZ;c1gD]NY;a1iD^N[;^1fDaN];[1fD`N`;[1cDaNd;Y1_DcNg;X1T1K5K5K5L4M3L3N3M3L4M2MQYR4"}, "label": "three yellow bananas"}]}
{"id": 539475, "image": "COCO_train2014_000000539475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a milk bottle to a baby ' s mouth\"."}], "task": "refering", "answer_template": "The \"a woman holding a milk bottle to a baby ' s mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 66, 67, 68, 69, 70, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 113, 114, 115, 122, 123, 124, 125, 126, 127, 128, 129, 130, 137, 138, 139, 140, 141, 142, 143, 144, 145, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 313, 314, 327, 328, 329, 343, 344], "bbox": [0.1313583138173302, 0.098875, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 71076.57135, "rle": {"size": [640, 427], "counts": "fWS1W1h1ZOg>\\1f_OFT`0X2J6J6J6K5J6J2N3M3N2M3M3M2N3N2N2M2O1O2N1N2O1O2N1N2O1O2N1N3N4L4L4L4M3L2O000O100O100O100O100O010O10000O010O100O10O0100O100O01000O100O100O010O1O1O1O100O1O1O00100O1O1O1O1mMaHdG`7Z8cHeG]7X8fHgGZ7X8iHfGX7X8jHgGW7W8kHhGV7V8mHhGT7U8oHkGQ7S8RIkGo6S8SIlGn6Q8UInGl6P8VIoGk6o7WIQHi6m7XISHi6j7ZIUHg6i7[IVHf6h7[IXHf6f7\\IZHd6f7\\IYHe6g7YIZHh6f7WIZHj6f7TI[Hm6e7TIZHl6f7TIYHm6g7SIXHn6h7SIVHn6j7RITHP7l7QIRHP7n7PIPHR7P8oHnGR7R8nHlGT7T8mHjGT7W8kHeGY7[8hH`G\\7`8dH\\G`7e8_HYGd7f8]HWGe7j8ZHTGh7l8XHRGj7n8d11O001O001O00010O001O0000001O0001O01O0000001O001O001O001O00000000O1000000O10000O1000O01000O01000O010O100O011N10000O100O100O101N10000O101N101N101N2O001N101O1N101O0O2O1O0O2O1O1N101O1N2O0O2M3N1N3^IVEk3l:RLXEj3l:RLXEk3k:QLYEl3j:PLZEl3j:PLZEm3i:oK\\El3h:PL\\Em3g:oK]En3f:nK^En3g:mK]EP4f:lK^EQ4e:kK_EQ4e:kK_ER4d:jK`ER4d:eKeEX4^:^KTFW4[=^O4N2M3N2M3N3L3N2M3N2M3M3N2M3N2M3N2M3N2M3N2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M3N1O2N1O2N1O2N1O2N101N2O1N3N1N3N1N2O2M2O2M2O1N3N1N3N1N2OYg20gXM1N2O0O2N2M3N2M2O2M3M3N1N3M3N2M2N3N2M3M3N1N3M3N0O100O1O100O1O10O01O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O010O1f^O"}, "label": "a woman holding a milk bottle to a baby ' s mouth"}]}
{"id": 539475, "image": "COCO_train2014_000000539475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman feeding a baby with a bottle\"."}], "task": "refering", "answer_template": "The \"a woman feeding a baby with a bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 37, 38, 39, 50, 51, 52, 53, 54, 66, 67, 68, 69, 70, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 107, 108, 109, 110, 111, 112, 113, 114, 115, 122, 123, 124, 125, 126, 127, 128, 129, 130, 137, 138, 139, 140, 141, 142, 143, 144, 145, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 313, 314, 327, 328, 329, 343, 344], "bbox": [0.1313583138173302, 0.098875, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 71076.57135, "rle": {"size": [640, 427], "counts": "fWS1W1h1ZOg>\\1f_OFT`0X2J6J6J6K5J6J2N3M3N2M3M3M2N3N2N2M2O1O2N1N2O1O2N1N2O1O2N1N3N4L4L4L4M3L2O000O100O100O100O100O010O10000O010O100O10O0100O100O01000O100O100O010O1O1O1O100O1O1O00100O1O1O1O1mMaHdG`7Z8cHeG]7X8fHgGZ7X8iHfGX7X8jHgGW7W8kHhGV7V8mHhGT7U8oHkGQ7S8RIkGo6S8SIlGn6Q8UInGl6P8VIoGk6o7WIQHi6m7XISHi6j7ZIUHg6i7[IVHf6h7[IXHf6f7\\IZHd6f7\\IYHe6g7YIZHh6f7WIZHj6f7TI[Hm6e7TIZHl6f7TIYHm6g7SIXHn6h7SIVHn6j7RITHP7l7QIRHP7n7PIPHR7P8oHnGR7R8nHlGT7T8mHjGT7W8kHeGY7[8hH`G\\7`8dH\\G`7e8_HYGd7f8]HWGe7j8ZHTGh7l8XHRGj7n8d11O001O001O00010O001O0000001O0001O01O0000001O001O001O001O00000000O1000000O10000O1000O01000O01000O010O100O011N10000O100O100O101N10000O101N101N101N2O001N101O1N101O0O2O1O0O2O1O1N101O1N2O0O2M3N1N3^IVEk3l:RLXEj3l:RLXEk3k:QLYEl3j:PLZEl3j:PLZEm3i:oK\\El3h:PL\\Em3g:oK]En3f:nK^En3g:mK]EP4f:lK^EQ4e:kK_EQ4e:kK_ER4d:jK`ER4d:eKeEX4^:^KTFW4[=^O4N2M3N2M3N3L3N2M3N2M3M3N2M3N2M3N2M3N2M3N2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M3N1O2N1O2N1O2N1O2N101N2O1N3N1N3N1N2O2M2O2M2O1N3N1N3N1N2OYg20gXM1N2O0O2N2M3N2M2O2M3M3N1N3M3N2M2N3N2M3M3N1N3M3N0O100O1O100O1O10O01O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O010O1f^O"}, "label": "a woman feeding a baby with a bottle"}]}
{"id": 539475, "image": "COCO_train2014_000000539475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby on a blanket on a woman ' s lap\"."}], "task": "refering", "answer_template": "The \"a baby on a blanket on a woman ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 204, 205, 206, 207, 215, 216, 217, 218, 219, 220, 221, 222, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 273, 274, 275, 276, 288, 289], "bbox": [0.20498829039812647, 0.533890625, 0.8703981264637002, 0.8433125], "iscrowd": 0, "area": 23447.282949999997, "rle": {"size": [640, 427], "counts": "b^g12nc01O1O1O2N1N2O1O2N1O1O1O1O001g]ONa`04[_O0d`00[_O1e`00Z_O1d`00[_O1e`00Y_O2e`00Z_O0f`00Y_O0g`02W_OOi`01W_OOh`02W_OOi`01V_O0i`02V_ONj`02U_OOj`02U_ONl`02S_OOl`02T_ONl`02S_OOl`03R_ONn`02Q_OOn`02Q_ONPa02o^OOPa02o^OOQa01o^OOPa03n^ONPa04o^OMn`06R_OIm`09R_OHl`09T_OHj`0:V_OFh`0;X_OFf`0<Z_ODd`0=\\_OCc`0?]_OAa`0`0__O@``0b0`_O]O_`0d0a_O\\O^`0f0b_OYO^`0g0c_OYO[`0g0f_OYOY`0h0h_OWOV`0j0k_OVOT`0j0m_OVOS`0i0n_OXOR`0f0o_OZOQ`0f0o_OZOR`0e0n_O[OS`0e0l_O[OU`0e0j_O[OW`0e0h_O\\OX`0U2001O1O100O1O00100000000O100000000O100000000O1000000O100000000O100000000O10000O100O10000O10O010000O100O10000O100O1000O010O010O10O001O001O001O00010O010O010O0100O010O010O0100O10O10O100O010O10000000000000000O100000000000O100000000O100O100O1O1O1O1O100O1O001O1O1O000G:M2O100O2O0O1O101N10000O20O1O00100O001O10O01O1O010O1O001O1O001O1O1O1O1O1M2[Of000000000O010000000O100000000O100O101N101N101N100O2O0O2O001N101N101N101O2M3N2M2O2M3L3N3M3M3M3M3L5L5J:BRPR1"}, "label": "a baby on a blanket on a woman ' s lap"}]}
{"id": 539475, "image": "COCO_train2014_000000539475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby drinking a bottle\"."}], "task": "refering", "answer_template": "The \"a baby drinking a bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 204, 205, 206, 207, 215, 216, 217, 218, 219, 220, 221, 222, 229, 230, 231, 232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 273, 274, 275, 276, 288, 289], "bbox": [0.20498829039812647, 0.533890625, 0.8703981264637002, 0.8433125], "iscrowd": 0, "area": 23447.282949999997, "rle": {"size": [640, 427], "counts": "b^g12nc01O1O1O2N1N2O1O2N1O1O1O1O001g]ONa`04[_O0d`00[_O1e`00Z_O1d`00[_O1e`00Y_O2e`00Z_O0f`00Y_O0g`02W_OOi`01W_OOh`02W_OOi`01V_O0i`02V_ONj`02U_OOj`02U_ONl`02S_OOl`02T_ONl`02S_OOl`03R_ONn`02Q_OOn`02Q_ONPa02o^OOPa02o^OOQa01o^OOPa03n^ONPa04o^OMn`06R_OIm`09R_OHl`09T_OHj`0:V_OFh`0;X_OFf`0<Z_ODd`0=\\_OCc`0?]_OAa`0`0__O@``0b0`_O]O_`0d0a_O\\O^`0f0b_OYO^`0g0c_OYO[`0g0f_OYOY`0h0h_OWOV`0j0k_OVOT`0j0m_OVOS`0i0n_OXOR`0f0o_OZOQ`0f0o_OZOR`0e0n_O[OS`0e0l_O[OU`0e0j_O[OW`0e0h_O\\OX`0U2001O1O100O1O00100000000O100000000O100000000O1000000O100000000O100000000O10000O100O10000O10O010000O100O10000O100O1000O010O010O10O001O001O001O00010O010O010O0100O010O010O0100O10O10O100O010O10000000000000000O100000000000O100000000O100O100O1O1O1O1O100O1O001O1O1O000G:M2O100O2O0O1O101N10000O20O1O00100O001O10O01O1O010O1O001O1O001O1O1O1O1O1M2[Of000000000O010000000O100000000O100O101N101N101N100O2O0O2O001N101N101N101O2M3N2M2O2M3L3N3M3M3M3M3L5L5J:BRPR1"}, "label": "a baby drinking a bottle"}]}
{"id": 162645, "image": "COCO_train2014_000000162645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza on a conference room table , topped very heavily with mushrooms\"."}], "task": "refering", "answer_template": "The \"a pizza on a conference room table , topped very heavily with mushrooms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [378, 379, 380, 381, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528], "bbox": [0.0045, 0.7303437500000001, 1.0, 0.986515625], "iscrowd": 0, "area": 76195.68984999998, "rle": {"size": [640, 640], "counts": "R_2e0Vc05L4N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O10000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O100000000O100000000O10000000000O100000000O100000000O100O10000O100O100O100O100O100O10000O100O100O1O100O1O100O1O1O100O1O100O1O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000O10000000000O100000000O100000000O10000000000O100000000O10000000000O11O00001O0000001O0000001O0000001O00001O0000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000000000000000000000001O0000000000000000000000001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O00001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O00001O0000001O0000001O00001O0000001O00001O0000001O00001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00002N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1K7I:E:FU]O"}, "label": "a pizza on a conference room table , topped very heavily with mushrooms"}]}
{"id": 162645, "image": "COCO_train2014_000000162645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mushroom pizza in a box\"."}], "task": "refering", "answer_template": "The \"a mushroom pizza in a box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [378, 379, 380, 381, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528], "bbox": [0.0045, 0.7303437500000001, 1.0, 0.986515625], "iscrowd": 0, "area": 76195.68984999998, "rle": {"size": [640, 640], "counts": "R_2e0Vc05L4N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O10000000000O100000000O10000000000O100000000O10000000000O10000000000O10000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000O100000000O100000000O10000000000O100000000O100000000O100O10000O100O100O100O100O100O10000O100O100O1O100O1O100O1O1O100O1O100O1O10000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000O10000000000O100000000O100000000O10000000000O100000000O10000000000O11O00001O0000001O0000001O0000001O00001O0000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000000000000000000000001O0000000000000000000000001O00001O00001O001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O00001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O001O1O00001O0000001O0000001O00001O0000001O00001O0000001O00001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00002N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1K7I:E:FU]O"}, "label": "a mushroom pizza in a box"}]}
{"id": 162645, "image": "COCO_train2014_000000162645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is eating sandwich by sitting along with other members\"."}], "task": "refering", "answer_template": "The \"a man is eating sandwich by sitting along with other members\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 121, 142, 143, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235], "bbox": [0.079578125, 0.18815625, 0.348625, 0.45884375000000005], "iscrowd": 0, "area": 15192.81195, "rle": {"size": [640, 640], "counts": "aTP13fc08H7J6N2N2N3M2N2N2N2N2N3M2N4L3M3N2M3M3M3M3M3M4L3N2N1N2O001N1010O1O10OO2N2M3M30O1000000O2O0000000O101O0O100O101N100O100aN^M_Ad2\\>bMbA^2Z>iMbAX2Z>nMdAR2[>RNaAo1^>TN`Am1]>XN_Ai1`>ZN]Ag1a>]N\\Ad1c>^N[Ac1c>aNZAa1d>aNZA`1d>dNXA^1g>dNWA]1i>dNUA]1k>m1O000000000O10O100000000000000000000000O1001O001O001N101O0O2O0O2N2N1RMn@m0T?lNSAR1Q?iNSAT1P?hNSAW1P?dNTAY1P?bNRA^1Q?]NRAb1P?ZNRAf1Q?TNRAl1a`001O00001O00000O2O0@\\^OdNda0\\1\\^OdNda0[1^^OdNca0Z1_^OeNaa0[1`^OdN`a0[1b^OdN^a0\\1c^OcN^a0[1c^OeN]a0[1d^OdN\\a0\\1e^OcN\\a0[1f^OdNZa0\\1g^OcNYa0]1h^ObNXa0^1h^ObNYa0\\1i^OcNWa0]1j^ObNXa0\\1i^OcNYa0Z1i0N1O2MY]OoNcb0o0[]OYOab0W1L10O1O1O1O010O1O1L3I8H8K5N2N1N3N2M3N2N1N3N2MP\\T8"}, "label": "a man is eating sandwich by sitting along with other members"}]}
{"id": 162645, "image": "COCO_train2014_000000162645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the grey shirt\"."}], "task": "refering", "answer_template": "The \"the man with the grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 121, 142, 143, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235], "bbox": [0.079578125, 0.18815625, 0.348625, 0.45884375000000005], "iscrowd": 0, "area": 15192.81195, "rle": {"size": [640, 640], "counts": "aTP13fc08H7J6N2N2N3M2N2N2N2N2N3M2N4L3M3N2M3M3M3M3M3M4L3N2N1N2O001N1010O1O10OO2N2M3M30O1000000O2O0000000O101O0O100O101N100O100aN^M_Ad2\\>bMbA^2Z>iMbAX2Z>nMdAR2[>RNaAo1^>TN`Am1]>XN_Ai1`>ZN]Ag1a>]N\\Ad1c>^N[Ac1c>aNZAa1d>aNZA`1d>dNXA^1g>dNWA]1i>dNUA]1k>m1O000000000O10O100000000000000000000000O1001O001O001N101O0O2O0O2N2N1RMn@m0T?lNSAR1Q?iNSAT1P?hNSAW1P?dNTAY1P?bNRA^1Q?]NRAb1P?ZNRAf1Q?TNRAl1a`001O00001O00000O2O0@\\^OdNda0\\1\\^OdNda0[1^^OdNca0Z1_^OeNaa0[1`^OdN`a0[1b^OdN^a0\\1c^OcN^a0[1c^OeN]a0[1d^OdN\\a0\\1e^OcN\\a0[1f^OdNZa0\\1g^OcNYa0]1h^ObNXa0^1h^ObNYa0\\1i^OcNWa0]1j^ObNXa0\\1i^OcNYa0Z1i0N1O2MY]OoNcb0o0[]OYOab0W1L10O1O1O1O010O1O1L3I8H8K5N2N1N3N2M3N2N1N3N2MP\\T8"}, "label": "the man with the grey shirt"}]}
{"id": 162645, "image": "COCO_train2014_000000162645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a dark blue sweater and glasses\"."}], "task": "refering", "answer_template": "The \"a man in a dark blue sweater and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 111, 112, 133, 134, 135, 136, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.70603125, 0.14456249999999998, 0.9841562500000001, 0.43529687499999997], "iscrowd": 0, "area": 16940.256950000003, "rle": {"size": [640, 640], "counts": "hgj82lc03L3M3N2M3N2M3M3N3L3N2M3N2N2M3N2N2M4M1O2M3N1O2N1N3N2O001O1O001O1O010O10O10O10O0100O010O100O10O0100O10O0100O100O010O100O100O100O100O10O01e_ObMT?^2g@gMX?Z2e@iM[?W2d@jM[?W2e@hM\\?X2c@iM\\?X2d@hMl>La@]2b0hMW>]3iAcLU>_3jAbLT>`3lA`LQ>b3oA_Lo=c3QB]Lm=e3RB\\Lk=h3TBXLk=i3TBXLk=i3UBWLk=i3TBXLk=i3UBWLj=j3VBVLi=k3WBULh=l3XBTLg=n3XBRLh=n3XBRLg=o3ZBPLe=Q4[BoKe=Q4[BoKd=R4\\BnKd=S4[BmKd=T4\\BlKd=T4\\BlKd=T4]BkKb=V4]BkKc=V4\\BjKc=X4\\BhKd=Y4[BgKe=Y4[BgKe=Z4ZBeKg=\\4WBeKi=\\4UBeKk=[4TBfKl=[4RBfKn=\\4oAeKQ>]4kAeKU>\\4iAeKW>R5000000001O1O1O1N2O1O2ULbAP2_>kMeAU2\\>fMiAY2Y>aMkA_2X>ZMlAf2W>SMnAl2U>mLoAS3T>fLPBZ3S>_LRB`3W?0001N101O00001O00000000001O1O1O1O2N1O2M2O2N1O8H=C>B>B9G3M4L3M4L4L3M4LO1000000000O10O03N5Kgo5"}, "label": "a man in a dark blue sweater and glasses"}]}
{"id": 162645, "image": "COCO_train2014_000000162645.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male wearing glasses eats pizza at a table with soda drinks nearby\"."}], "task": "refering", "answer_template": "The \"a male wearing glasses eats pizza at a table with soda drinks nearby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 111, 112, 133, 134, 135, 136, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.70603125, 0.14456249999999998, 0.9841562500000001, 0.43529687499999997], "iscrowd": 0, "area": 16940.256950000003, "rle": {"size": [640, 640], "counts": "hgj82lc03L3M3N2M3N2M3M3N3L3N2M3N2N2M3N2N2M4M1O2M3N1O2N1N3N2O001O1O001O1O010O10O10O10O0100O010O100O10O0100O10O0100O100O010O100O100O100O100O10O01e_ObMT?^2g@gMX?Z2e@iM[?W2d@jM[?W2e@hM\\?X2c@iM\\?X2d@hMl>La@]2b0hMW>]3iAcLU>_3jAbLT>`3lA`LQ>b3oA_Lo=c3QB]Lm=e3RB\\Lk=h3TBXLk=i3TBXLk=i3UBWLk=i3TBXLk=i3UBWLj=j3VBVLi=k3WBULh=l3XBTLg=n3XBRLh=n3XBRLg=o3ZBPLe=Q4[BoKe=Q4[BoKd=R4\\BnKd=S4[BmKd=T4\\BlKd=T4\\BlKd=T4]BkKb=V4]BkKc=V4\\BjKc=X4\\BhKd=Y4[BgKe=Y4[BgKe=Z4ZBeKg=\\4WBeKi=\\4UBeKk=[4TBfKl=[4RBfKn=\\4oAeKQ>]4kAeKU>\\4iAeKW>R5000000001O1O1O1N2O1O2ULbAP2_>kMeAU2\\>fMiAY2Y>aMkA_2X>ZMlAf2W>SMnAl2U>mLoAS3T>fLPBZ3S>_LRB`3W?0001N101O00001O00000000001O1O1O1O2N1O2M2O2N1O8H=C>B>B9G3M4L3M4L4L3M4LO1000000000O10O03N5Kgo5"}, "label": "a male wearing glasses eats pizza at a table with soda drinks nearby"}]}
{"id": 260957, "image": "COCO_train2014_000000260957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing sunglasses and an identification badge\"."}], "task": "refering", "answer_template": "The \"the woman wearing sunglasses and an identification badge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.33546875, 0.2396713615023474, 0.6478906249999999, 0.9971361502347417], "iscrowd": 0, "area": 46019.96409999999, "rle": {"size": [426, 640], "counts": "ffi2a0T<g0YOh0B>M2nGaMlM<U6W2eKjMbM=h6l1]KkNb4Y1TKnNk4U1lJROS5R1dJTO[5P1_JSO`5P1ZJTOf5o0SJUOl5o0mIUOR6o0hITOW6o0cIUO\\6m3O1O1O100O1hNYIlKh6m3aIQL_6m3eIPL\\6m3iIPLX6n3lIoKU6o3oIoKQ6n3TJoKm5P4VJmKk5Q4XJnKh5P4[JnKf5Q4\\JmKe5Q4^JmKc5R4_JlKb5R4aJlK`5R4cJlK^5S4m1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O100O100O100O1O100O10000000000000000001O0000000000000000000000000O1000000000000000001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O2N1O2N2N2N1O2N5K6SH^Kf6h4QI[Ko6k4hHWKY7o4_HSKa7[50O11O000000001O000000001O001O1O002N1O2N1O2N1O2N1O1O2N1O2N2N2]LUKcNm4]1UK_Nl4a1WK[Nk4e1WKWNk4h1YKSNi4l1hKbMZ4]2WLQMk3m2[3O1N2N2O0O2O1N2N2O1N2O2M3M2O2M3N1N3M2O2M2N3N1N3N1N3K4K6J5K6J5K6J5K6J5KRZm2"}, "label": "the woman wearing sunglasses and an identification badge"}]}
{"id": 260957, "image": "COCO_train2014_000000260957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white photo of a woman with an umbrella\"."}], "task": "refering", "answer_template": "The \"a black and white photo of a woman with an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.33546875, 0.2396713615023474, 0.6478906249999999, 0.9971361502347417], "iscrowd": 0, "area": 46019.96409999999, "rle": {"size": [426, 640], "counts": "ffi2a0T<g0YOh0B>M2nGaMlM<U6W2eKjMbM=h6l1]KkNb4Y1TKnNk4U1lJROS5R1dJTO[5P1_JSO`5P1ZJTOf5o0SJUOl5o0mIUOR6o0hITOW6o0cIUO\\6m3O1O1O100O1hNYIlKh6m3aIQL_6m3eIPL\\6m3iIPLX6n3lIoKU6o3oIoKQ6n3TJoKm5P4VJmKk5Q4XJnKh5P4[JnKf5Q4\\JmKe5Q4^JmKc5R4_JlKb5R4aJlK`5R4cJlK^5S4m1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O100O100O100O1O100O10000000000000000001O0000000000000000000000000O1000000000000000001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O2N1O2N2N2N1O2N5K6SH^Kf6h4QI[Ko6k4hHWKY7o4_HSKa7[50O11O000000001O000000001O001O1O002N1O2N1O2N1O2N1O1O2N1O2N2N2]LUKcNm4]1UK_Nl4a1WK[Nk4e1WKWNk4h1YKSNi4l1hKbMZ4]2WLQMk3m2[3O1N2N2O0O2O1N2N2O1N2O2M3M2O2M3N1N3M2O2M2N3N1N3N1N3K4K6J5K6J5K6J5K6J5KRZm2"}, "label": "a black and white photo of a woman with an umbrella"}]}
{"id": 260957, "image": "COCO_train2014_000000260957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt holding a purse\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt holding a purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 109, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.662640625, 0.22840375586854458, 1.0, 0.989037558685446], "iscrowd": 0, "area": 43848.08049999999, "rle": {"size": [426, 640], "counts": "gg`52W=2M3M4L3N2O1N3N1N2O1N2O2M2O1N2O0O100O100O100O1O100O010O0M3L5K4L5L1N2N2N2N3M2O1N2N2N3M3N1O2M3N1N3N2M3N11000000O1014K`0@a0_O`0@a0_O<D7J6I7\\MSJ_MT6Z2TJ^MS6[2UJ]MR6\\2VJ\\MQ6]2VJ]Mk5a2]JWMe5g2`JUMa5i2dJSM]5k2gJRMZ5l2kJQMU5m2PKoLQ5o2SKnLn4P3WKmLi4Q3[KlLf4R3_KjLb4U3j2000000000000000000000000000000000000000000O100000000001O0000000000000000000000000000000000000000000000001O1O2N1O1O1O1jGmLS6U3iImLW6T3fInLZ6S3cIoL]6R3_IRM`6o2]ISMc6o2YISMg6n2VIUMi6l2bHhM^7R40001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O00002N1O1O1O1O2N1O1O1O2N1O1O1O2N1]NbIXL_6e3iIULX6h3PJRLR6j3UJQLl5l3\\JnKe5o3bJlK_5P4jJjKX5R4PKhKQ5U4WKeKk4W4Y2E;A?A?B>A?A?A?A?AaH"}, "label": "a woman in a white shirt holding a purse"}]}
{"id": 260957, "image": "COCO_train2014_000000260957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with glasses and a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman with glasses and a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 109, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.662640625, 0.22840375586854458, 1.0, 0.989037558685446], "iscrowd": 0, "area": 43848.08049999999, "rle": {"size": [426, 640], "counts": "gg`52W=2M3M4L3N2O1N3N1N2O1N2O2M2O1N2O0O100O100O100O1O100O010O0M3L5K4L5L1N2N2N2N3M2O1N2N2N3M3N1O2M3N1N3N2M3N11000000O1014K`0@a0_O`0@a0_O<D7J6I7\\MSJ_MT6Z2TJ^MS6[2UJ]MR6\\2VJ\\MQ6]2VJ]Mk5a2]JWMe5g2`JUMa5i2dJSM]5k2gJRMZ5l2kJQMU5m2PKoLQ5o2SKnLn4P3WKmLi4Q3[KlLf4R3_KjLb4U3j2000000000000000000000000000000000000000000O100000000001O0000000000000000000000000000000000000000000000001O1O2N1O1O1O1jGmLS6U3iImLW6T3fInLZ6S3cIoL]6R3_IRM`6o2]ISMc6o2YISMg6n2VIUMi6l2bHhM^7R40001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O00002N1O1O1O1O2N1O1O1O2N1O1O1O2N1]NbIXL_6e3iIULX6h3PJRLR6j3UJQLl5l3\\JnKe5o3bJlK_5P4jJjKX5R4PKhKQ5U4WKeKk4W4Y2E;A?A?B>A?A?A?A?AaH"}, "label": "a woman with glasses and a white shirt"}]}
{"id": 580446, "image": "COCO_train2014_000000580446.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light colored couch near a blue couch\"."}], "task": "refering", "answer_template": "The \"a light colored couch near a blue couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 134, 135, 136, 137, 138, 139, 140, 141, 155, 156, 157, 158, 159, 175, 176], "bbox": [0.45514, 0.46882165605095544, 0.8684, 0.8451592356687898], "iscrowd": 0, "area": 15577.208649999999, "rle": {"size": [314, 500], "counts": "mRV24c94N2O1N2N8I=B9G010O01000O0100O01000O10O01000O10000O101O0O1000001N10000O101O0O1000001N10000O2M2M3N2N2N3M2O1000001O01O000001O000000001O000000001O000000010O0000001O000000001O000000001O0001O0001O000000001O000WIfMR6Z2fInMZ6R2^IVNb6a201O0000010O000000001O000000001O0000010O000000001O01O000001O0001O01O0000000010O00000001O01O0001O000000010O0000000010O000001O000001O01O00000001M2O1O1N2O1N2O2N1N2O1N2O1O2M2O1N2L4I7I7H9K4N9G5K4L3M3M4L3M4LPRd0"}, "label": "a light colored couch near a blue couch"}]}
{"id": 580446, "image": "COCO_train2014_000000580446.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan loveseat topped with one floral pillow\"."}], "task": "refering", "answer_template": "The \"a tan loveseat topped with one floral pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 105, 116, 117, 118, 119, 120, 121, 122, 123, 134, 135, 136, 137, 138, 139, 140, 141, 155, 156, 157, 158, 159, 175, 176], "bbox": [0.45514, 0.46882165605095544, 0.8684, 0.8451592356687898], "iscrowd": 0, "area": 15577.208649999999, "rle": {"size": [314, 500], "counts": "mRV24c94N2O1N2N8I=B9G010O01000O0100O01000O10O01000O10000O101O0O1000001N10000O101O0O1000001N10000O2M2M3N2N2N3M2O1000001O01O000001O000000001O000000001O000000010O0000001O000000001O000000001O0001O0001O000000001O000WIfMR6Z2fInMZ6R2^IVNb6a201O0000010O000000001O000000001O0000010O000000001O01O000001O0001O01O0000000010O00000001O01O0001O000000010O0000000010O000001O000001O01O00000001M2O1O1N2O1N2O2N1N2O1N2O1O2M2O1N2L4I7I7H9K4N9G5K4L3M3M4L3M4LPRd0"}, "label": "a tan loveseat topped with one floral pillow"}]}
{"id": 564063, "image": "COCO_train2014_000000564063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck in the front\"."}], "task": "refering", "answer_template": "The \"truck in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 61, 62, 65, 66, 67, 68, 80, 81, 82, 84, 85, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 110, 111, 112, 113, 114, 125, 126, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 153, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 228], "bbox": [0.442265625, 0.07049429657794677, 0.9809218749999999, 0.4910076045627376], "iscrowd": 0, "area": 26554.856349999995, "rle": {"size": [526, 640], "counts": "Saa4:c?c0K5L2O2N1O1OO2N101N101N2N101N2OBZOUAe0m>^OPA`0Q?Cl@>S?>001Bm@@S?`0o@^OQ?b0RA\\Om>c0VA\\Oi>>n@A:1h>;SAA73e>;eAD[>:hAFW>8mAGR>7QBIo=4TBLk=2Y1Nn`1g0]^N8G1001O100O00100O1O100O0010000000O10000000000O100001O00001O00000O2O00001O00001O0002N3M3M2N3M2N3M3M2N3M2N2N3M2N00000000000B>6J:F9G9G:F9G9G0000O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1^N\\OYCe0g<]OWCc0i<_OTCb0l<@QCa0o<AnB`0R=BlB>T=CjB>V=BiB?8mNo;e0gC?9mNP<d0fC`09mNQ<d0eC?9nNR<c0dC`09nNS<c0bC`0:nNT<c0`C`0;nNU<b0^Cb0<mNV<b0\\Cb0<nNX<a0YCc0>mNY<a0VCd0`0mNY<?VCe0`0mNY<S2aCSN_<m1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<h0aCZOON`<d0gC]OHOa<a0lC^OC1a<>RD^O]O4a<;WD_OXO6a<8\\D@SO8a<5_DCPO8a<4aDCnN9a<3bDDmN9a<1dDFkN9a<0eDGjN9a<NgDIhN9a<MhDJgN9a<LjDJeN:`<KmDLcN8`<KnDMbN8`<JoDNaN8`<HQEJUNO:?`<GREJUN09?`<FTEJSN19?`<DVEKSN17`0`<CWEKSN26a0_<AYEKSN35a0_<_ObE0oMa0_<^OcE1nMa0_<]OdE2mMa0o>_OQAa0o>@PA`0P?@PA`0P?@PA`0Q?_Oo@a0Q?_OPA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?Ao@?Q?Ao@?R?@n@`0R?@n@`0R?@n@`0R?@n@`0R?@n@`0R?@n@b0P?^OPAd0n>\\ORAg0k>YOUAi0i>WOVAl0h>TOXAn0f>SOYAP1d>PO\\AR1T=TOgCJUOT1i<^OoC^OWOX1g<^OoCYOZO[1e<_OnCVO\\O^1d<]OoCUO[Ob1e<XOPDVOYOg1d<TORDT1m;kNSDV1l;jNTDV1m;jNRDV1n;jNRDW1n;iNQDW1o;iNQDX1o;hNPDX1P<hNPDX1o;jNoCW1P<jNPDV1o;lNPDT1o;mNQDS1n;QOnCP1Q<TOlCl0S<XOjCh0U<\\OhCd0W<@eCa0Z<j1O1O1O1N2N2N2N2O0O2O1000O1010O1O1O1O1O1O1O1O1O1O1O1O1O0010O01O00001O001O00001O00001O010O001O1O1O1O001O1O1O1O1O1O2N010O000000001O0000001O0000M3K5K6J5L4K5K5KkNnKXFn3k9[LmEc3T:fLdEZ3Z:PM^Eo2b:[MUEe2o:[15J6K5K5K5J6K6J2M3N1O1N3N1O2N=Bb0SOn0nNR1mNoP6"}, "label": "truck in the front"}]}
{"id": 564063, "image": "COCO_train2014_000000564063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck in front that is being followed by another truck\"."}], "task": "refering", "answer_template": "The \"the truck in front that is being followed by another truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 61, 62, 65, 66, 67, 68, 80, 81, 82, 84, 85, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 110, 111, 112, 113, 114, 125, 126, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 153, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 228], "bbox": [0.442265625, 0.07049429657794677, 0.9809218749999999, 0.4910076045627376], "iscrowd": 0, "area": 26554.856349999995, "rle": {"size": [526, 640], "counts": "Saa4:c?c0K5L2O2N1O1OO2N101N101N2N101N2OBZOUAe0m>^OPA`0Q?Cl@>S?>001Bm@@S?`0o@^OQ?b0RA\\Om>c0VA\\Oi>>n@A:1h>;SAA73e>;eAD[>:hAFW>8mAGR>7QBIo=4TBLk=2Y1Nn`1g0]^N8G1001O100O00100O1O100O0010000000O10000000000O100001O00001O00000O2O00001O00001O0002N3M3M2N3M2N3M3M2N3M2N2N3M2N00000000000B>6J:F9G9G:F9G9G0000O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1^N\\OYCe0g<]OWCc0i<_OTCb0l<@QCa0o<AnB`0R=BlB>T=CjB>V=BiB?8mNo;e0gC?9mNP<d0fC`09mNQ<d0eC?9nNR<c0dC`09nNS<c0bC`0:nNT<c0`C`0;nNU<b0^Cb0<mNV<b0\\Cb0<nNX<a0YCc0>mNY<a0VCd0`0mNY<?VCe0`0mNY<S2aCSN_<m1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<l1`CTN`<h0aCZOON`<d0gC]OHOa<a0lC^OC1a<>RD^O]O4a<;WD_OXO6a<8\\D@SO8a<5_DCPO8a<4aDCnN9a<3bDDmN9a<1dDFkN9a<0eDGjN9a<NgDIhN9a<MhDJgN9a<LjDJeN:`<KmDLcN8`<KnDMbN8`<JoDNaN8`<HQEJUNO:?`<GREJUN09?`<FTEJSN19?`<DVEKSN17`0`<CWEKSN26a0_<AYEKSN35a0_<_ObE0oMa0_<^OcE1nMa0_<]OdE2mMa0o>_OQAa0o>@PA`0P?@PA`0P?@PA`0Q?_Oo@a0Q?_OPA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?@PA`0P?Ao@?Q?Ao@?R?@n@`0R?@n@`0R?@n@`0R?@n@`0R?@n@`0R?@n@b0P?^OPAd0n>\\ORAg0k>YOUAi0i>WOVAl0h>TOXAn0f>SOYAP1d>PO\\AR1T=TOgCJUOT1i<^OoC^OWOX1g<^OoCYOZO[1e<_OnCVO\\O^1d<]OoCUO[Ob1e<XOPDVOYOg1d<TORDT1m;kNSDV1l;jNTDV1m;jNRDV1n;jNRDW1n;iNQDW1o;iNQDX1o;hNPDX1P<hNPDX1o;jNoCW1P<jNPDV1o;lNPDT1o;mNQDS1n;QOnCP1Q<TOlCl0S<XOjCh0U<\\OhCd0W<@eCa0Z<j1O1O1O1N2N2N2N2O0O2O1000O1010O1O1O1O1O1O1O1O1O1O1O1O1O0010O01O00001O001O00001O00001O010O001O1O1O1O001O1O1O1O1O1O2N010O000000001O0000001O0000M3K5K6J5L4K5K5KkNnKXFn3k9[LmEc3T:fLdEZ3Z:PM^Eo2b:[MUEe2o:[15J6K5K5K5J6K6J2M3N1O1N3N1O2N=Bb0SOn0nNR1mNoP6"}, "label": "the truck in front that is being followed by another truck"}]}
{"id": 564063, "image": "COCO_train2014_000000564063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow colored vehicle with only one man and a tire\"."}], "task": "refering", "answer_template": "The \"the yellow colored vehicle with only one man and a tire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 144, 145, 146, 147, 148, 153, 154, 155, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 364, 365, 366, 370, 371, 372, 373, 376, 377, 378, 380, 381, 382, 383, 384, 394, 395, 396, 404, 405, 406, 417, 418, 427, 428, 429], "bbox": [0.02403125, 0.2626425855513308, 0.935359375, 0.9707794676806084], "iscrowd": 0, "area": 137591.34505000003, "rle": {"size": [526, 640], "counts": "_R85h>b1]Nb1_Na1K5O1000000O2O00000O1000001N10000O1000000O2O00000O1000001N1000000O10001O3L8I7I7I7H6K1O000O101O000O2O00001N10001O0O101O000O101iFTIj8l6SG[Ii8e6TGaIh8V7M3M3M3L4M3M3M3L5L3M3M3L4M3M3M1N1000000O1000001N1000000O10000O1000000O1000000O1000000O10O1000O10O1000O010000O0100000O01000O10O1000O10O10O1000O10O1000O0100O1J5K6J6K5J5K6J6J6J5O200000O1000O10000000000O10000000000O11O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O001O1O001O1]NPHfJQ8Y5YH]Jg7d5aHSJ`7l5jHjIW7U6RIbIn6^6b100O1O0010O01O0>C?@`0@4MO00001O00001O001O00001O00001O00001O001OO010O100O1eISHU4m7hKVHW4k7gKXHX4h7eK[HZ4f7cK]H\\4d7bK^H^4b7_KaH`4`7]KcHc4]7[KeHd4\\7YKhHf4X7WKkHh4V7UKmHk4S7SKoHl4R7QKQIo4o6nJTIQ5m6mJUIS5k6jJXIU5i6hJ[IW5e6fJ^IY5b6fJ`IZ5`6cJcI\\5^6aJeI_5[6_JgI`5Z6]JiIc5W6ZJmId5T6YJoIg5\\8O100O100O2O0O100O100O101N100O100O1000M3L4K5L4K5L4K5L4K5L4K6L300000000O100000000O1000001O000O100000000O10000000001N1000002N2N2N3M2N2O1N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2M3N2M3N2M4M2M100O100O10O0100O100O010O100O100O010O100O10O0100O100O010O100O10O0100O100O0100000000O100000000O2O0000000O10000000000O1N2J6J6J6J6J6J6J6J6J6J6I7J6J600O10000000000O101N10000O100O10000O10000O100O10000O10000O100O10000O100O10000O2O3M3L4M3L4M3M3L4M3L4M3M1N10001N100O10001N100O101O0O10000O2O0O10001N100O10001N10000O101N10000104K5K5K5L4K5K5L5J01O001O00001O001O00001O001O0O2O00001O001O001O00001O001O00001O001O001O1N3N2N3M2N2N3cLeDa1];VNTE[1S;]N\\EU1l:aNeEP1b:gNoEk0W:lNZFe0n9RObF?S=Aked0"}, "label": "the yellow colored vehicle with only one man and a tire"}]}
{"id": 564063, "image": "COCO_train2014_000000564063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"military truck number l4618873\"."}], "task": "refering", "answer_template": "The \"military truck number l4618873\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 144, 145, 146, 147, 148, 153, 154, 155, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 364, 365, 366, 370, 371, 372, 373, 376, 377, 378, 380, 381, 382, 383, 384, 394, 395, 396, 404, 405, 406, 417, 418, 427, 428, 429], "bbox": [0.02403125, 0.2626425855513308, 0.935359375, 0.9707794676806084], "iscrowd": 0, "area": 137591.34505000003, "rle": {"size": [526, 640], "counts": "_R85h>b1]Nb1_Na1K5O1000000O2O00000O1000001N10000O1000000O2O00000O1000001N1000000O10001O3L8I7I7I7H6K1O000O101O000O2O00001N10001O0O101O000O101iFTIj8l6SG[Ii8e6TGaIh8V7M3M3M3L4M3M3M3L5L3M3M3L4M3M3M1N1000000O1000001N1000000O10000O1000000O1000000O1000000O10O1000O10O1000O010000O0100000O01000O10O1000O10O10O1000O10O1000O0100O1J5K6J6K5J5K6J6J6J5O200000O1000O10000000000O10000000000O11O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O001O1O001O1]NPHfJQ8Y5YH]Jg7d5aHSJ`7l5jHjIW7U6RIbIn6^6b100O1O0010O01O0>C?@`0@4MO00001O00001O001O00001O00001O00001O001OO010O100O1eISHU4m7hKVHW4k7gKXHX4h7eK[HZ4f7cK]H\\4d7bK^H^4b7_KaH`4`7]KcHc4]7[KeHd4\\7YKhHf4X7WKkHh4V7UKmHk4S7SKoHl4R7QKQIo4o6nJTIQ5m6mJUIS5k6jJXIU5i6hJ[IW5e6fJ^IY5b6fJ`IZ5`6cJcI\\5^6aJeI_5[6_JgI`5Z6]JiIc5W6ZJmId5T6YJoIg5\\8O100O100O2O0O100O100O101N100O100O1000M3L4K5L4K5L4K5L4K5L4K6L300000000O100000000O1000001O000O100000000O10000000001N1000002N2N2N3M2N2O1N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2M3N2M3N2M4M2M100O100O10O0100O100O010O100O100O010O100O10O0100O100O010O100O10O0100O100O0100000000O100000000O2O0000000O10000000000O1N2J6J6J6J6J6J6J6J6J6J6I7J6J600O10000000000O101N10000O100O10000O10000O100O10000O10000O100O10000O100O10000O2O3M3L4M3L4M3M3L4M3L4M3M1N10001N100O10001N100O101O0O10000O2O0O10001N100O10001N10000O101N10000104K5K5K5L4K5K5L5J01O001O00001O001O00001O001O0O2O00001O001O001O00001O001O00001O001O001O1N3N2N3M2N2N3cLeDa1];VNTE[1S;]N\\EU1l:aNeEP1b:gNoEk0W:lNZFe0n9RObF?S=Aked0"}, "label": "military truck number l4618873"}]}
{"id": 416612, "image": "COCO_train2014_000000416612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a nice suit cutting a large cake\"."}], "task": "refering", "answer_template": "The \"a man in a nice suit cutting a large cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 39, 40, 41, 42, 53, 54, 55, 56, 57, 66, 67, 68, 69, 70, 71, 72, 73, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 110, 111, 113, 114, 115, 116, 117, 118, 119, 125, 126, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 252, 264, 265, 266, 267, 281, 282, 297], "bbox": [0.3542622950819672, 0.051953125, 0.9771662763466042, 0.836359375], "iscrowd": 0, "area": 59509.30854999998, "rle": {"size": [640, 427], "counts": "ian21mc03M3n\\OMRb07j]O4la0OQ^O4ka0OS^O3ka00R^O2la01P^O2ma02P^O1ma02P^O0na0R1N2N3M2M3N2N3M2N2N2N3O0O100O01N1O2N2N1N3M2N3M2N3M2M4M2N3M3M2N3M2010O010O010O10O010O10O010O010O10O010O010O10O010O10O010O010O010O01000O010O010O010O0100O0100O01^@`Nd<_1UCjNi<W1PCROn<m0kB]OR=d0gBDX=<`BN]=2\\B8T=[NdA]1Q1`0XOSN[=8SBU1S1j0YOnM`=e1PCf0YOlMe=]1kBg1S=YNfBP2X=oMaBZ2^=fM[Bc2d=n1N1O2N2N2N1O2N2N2N1O2N4cFXJZ4d6TJSJk5R:0O01O1O1J6F:E;H7O2N2N2N2N2M3N2N2N1O2N2M3N2N2N2N2NVOcJTCZ5k<ZKdBe4Z=Q1N3M3QN[I`Fg6^9bIYFa6d9mInEV6o9XJcEj5Z:eJXE^5e:o1N3L4jNgFbG[9Z8QG\\GR9_8YGXGi8d8cGQGa8j8]1K4L5N1O2O1N1012O3L4L4L3N3L4hFnFZ7U9_HSG^7P9YHXGd7l8SH\\Gj7h8mGaGo7b8iGfGU8]8bGkG[8Q:L3Mk0oHjDc3j=O1O10O01O1O10O0100000O100O1O1^BVKl;k4PDZKn;g4mC_KQ<b4jCdKT<]4gCiKW<X4dCnKZ<S4bCPL^<QOkBf6Z;]IZFL\\Ni6T;cI_F]6\\9kI_E_O>h6m9RJ_EZOb0d6j9ZJ^EWOe0a6g9oJWFR5c9VKZFj4a9]K\\Fe4^9bK`F_4Z9iKbFX4X9PLfFQ4T9VLiFk3Q8PIfH\\3WOe3Q8VI]Hd:a7cETHb:k7k0O0O2N2N1O2N101O1O001O1O001O001O1O001O0O2O1O001O1O001O001O1L3J7I6J7I;G<F;F:E:F;F:E;E<Bb0VOh0YOe0ZOe0[OVP6"}, "label": "a man in a nice suit cutting a large cake"}]}
{"id": 416612, "image": "COCO_train2014_000000416612.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in dark suit cutting the cake\"."}], "task": "refering", "answer_template": "The \"man in dark suit cutting the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 39, 40, 41, 42, 53, 54, 55, 56, 57, 66, 67, 68, 69, 70, 71, 72, 73, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 110, 111, 113, 114, 115, 116, 117, 118, 119, 125, 126, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 252, 264, 265, 266, 267, 281, 282, 297], "bbox": [0.3542622950819672, 0.051953125, 0.9771662763466042, 0.836359375], "iscrowd": 0, "area": 59509.30854999998, "rle": {"size": [640, 427], "counts": "ian21mc03M3n\\OMRb07j]O4la0OQ^O4ka0OS^O3ka00R^O2la01P^O2ma02P^O1ma02P^O0na0R1N2N3M2M3N2N3M2N2N2N3O0O100O01N1O2N2N1N3M2N3M2N3M2M4M2N3M3M2N3M2010O010O010O10O010O10O010O010O10O010O010O10O010O10O010O010O010O01000O010O010O010O0100O0100O01^@`Nd<_1UCjNi<W1PCROn<m0kB]OR=d0gBDX=<`BN]=2\\B8T=[NdA]1Q1`0XOSN[=8SBU1S1j0YOnM`=e1PCf0YOlMe=]1kBg1S=YNfBP2X=oMaBZ2^=fM[Bc2d=n1N1O2N2N2N1O2N2N2N1O2N4cFXJZ4d6TJSJk5R:0O01O1O1J6F:E;H7O2N2N2N2N2M3N2N2N1O2N2M3N2N2N2N2NVOcJTCZ5k<ZKdBe4Z=Q1N3M3QN[I`Fg6^9bIYFa6d9mInEV6o9XJcEj5Z:eJXE^5e:o1N3L4jNgFbG[9Z8QG\\GR9_8YGXGi8d8cGQGa8j8]1K4L5N1O2O1N1012O3L4L4L3N3L4hFnFZ7U9_HSG^7P9YHXGd7l8SH\\Gj7h8mGaGo7b8iGfGU8]8bGkG[8Q:L3Mk0oHjDc3j=O1O10O01O1O10O0100000O100O1O1^BVKl;k4PDZKn;g4mC_KQ<b4jCdKT<]4gCiKW<X4dCnKZ<S4bCPL^<QOkBf6Z;]IZFL\\Ni6T;cI_F]6\\9kI_E_O>h6m9RJ_EZOb0d6j9ZJ^EWOe0a6g9oJWFR5c9VKZFj4a9]K\\Fe4^9bK`F_4Z9iKbFX4X9PLfFQ4T9VLiFk3Q8PIfH\\3WOe3Q8VI]Hd:a7cETHb:k7k0O0O2N2N1O2N101O1O001O1O001O001O1O001O0O2O1O001O1O001O001O1L3J7I6J7I;G<F;F:E:F;F:E;E<Bb0VOh0YOe0ZOe0[OVP6"}, "label": "man in dark suit cutting the cake"}]}
{"id": 564073, "image": "COCO_train2014_000000564073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first plane which is ahead of the other two planes\"."}], "task": "refering", "answer_template": "The \"the first plane which is ahead of the other two planes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196], "bbox": [0.341, 0.39527058823529415, 0.646109375, 0.5979764705882353], "iscrowd": 0, "area": 4224.405349999999, "rle": {"size": [425, 640], "counts": "kSk23V=2N1N10O0100O10000O10O0100O10000O10O01M3M3N21O1O2M2O1O0000O100O01000O100O100O0I8000000001N2O1O2N1O1OO10O0100O100O100O010O100O100O10O0100O100O100mN^O^Eb0a:FXE;h:KRE5o:0jD1V;5dDKf0YOk9h1UFXNl9g1SFZNm9e1TF[Nl9e1SF\\Nn9c1RF]Nn9c1RF]NP:a1oE`NS:^1mEbNU:\\1kEdNW:Z1hEgNZ:Y1dEgN^:Z1_EfNc:[1YEfNi:Z1UEfNm:[1oDfNT;Z1iDfNY;Z1eDfN];b11O001O0O2O001O001O0TO_D0b;ObDM^;3gD_OE1e;?nDAR;?nDAR;?nDAS;>mDBS;>mDBT;=lDCT;=lDCU;<kDDU;<kDDV;;jDEV;;jDEW;:iDFW;:iDFW;:iDFW;:iDFW;:iDFW;:iDFW;:iDFW;9jDGU;:kDFT;;lDER;=nDCS;<mDDS;<mDCT;=lDCT;=lDCU;<lDCT;=lDCT;=mDBT;=mDBT;=mDBU;<lDCU;<m0O1O0000000000O1000O10000000O1M3N2001O001O1O1O1O2N1O1O1O000000000O1000001O000000000000000000000000001O000000hbm2"}, "label": "the first plane which is ahead of the other two planes"}]}
{"id": 564073, "image": "COCO_train2014_000000564073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"airplane in the sky , in the middle\"."}], "task": "refering", "answer_template": "The \"airplane in the sky , in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196], "bbox": [0.341, 0.39527058823529415, 0.646109375, 0.5979764705882353], "iscrowd": 0, "area": 4224.405349999999, "rle": {"size": [425, 640], "counts": "kSk23V=2N1N10O0100O10000O10O0100O10000O10O01M3M3N21O1O2M2O1O0000O100O01000O100O100O0I8000000001N2O1O2N1O1OO10O0100O100O100O010O100O100O10O0100O100O100mN^O^Eb0a:FXE;h:KRE5o:0jD1V;5dDKf0YOk9h1UFXNl9g1SFZNm9e1TF[Nl9e1SF\\Nn9c1RF]Nn9c1RF]NP:a1oE`NS:^1mEbNU:\\1kEdNW:Z1hEgNZ:Y1dEgN^:Z1_EfNc:[1YEfNi:Z1UEfNm:[1oDfNT;Z1iDfNY;Z1eDfN];b11O001O0O2O001O001O0TO_D0b;ObDM^;3gD_OE1e;?nDAR;?nDAR;?nDAS;>mDBS;>mDBT;=lDCT;=lDCU;<kDDU;<kDDV;;jDEV;;jDEW;:iDFW;:iDFW;:iDFW;:iDFW;:iDFW;:iDFW;:iDFW;9jDGU;:kDFT;;lDER;=nDCS;<mDDS;<mDCT;=lDCT;=lDCU;<lDCT;=lDCT;=mDBT;=mDBT;=mDBU;<lDCU;<m0O1O0000000000O1000O10000000O1M3N2001O001O1O1O1O2N1O1O1O000000000O1000001O000000000000000000000000001O000000hbm2"}, "label": "airplane in the sky , in the middle"}]}
{"id": 285548, "image": "COCO_train2014_000000285548.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green leafy plant growing in a short blue vase in an office\"."}], "task": "refering", "answer_template": "The \"a green leafy plant growing in a short blue vase in an office\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 118, 119, 120, 121, 137, 138, 139, 155, 156, 157, 158, 173, 174, 175, 176, 191, 192, 193, 194, 209, 210, 211, 212], "bbox": [0.5278999999999999, 0.23071999999999998, 0.8260599999999999, 0.9036000000000001], "iscrowd": 0, "area": 18905.170000000002, "rle": {"size": [375, 500], "counts": "jjP36_;2N2N2N2O2M2N2N2N2K5L4K6L3000000000001N1000000000001O0O101O2N1O1O2N3M3L4M001O001O001O0O2O0oInM]2S2^MVN]2j1_M^NZ2f1aM`NU2i1]LPNUO8T4Q2XLYN@HV4Q2XL[N_OEX4Q2WL^N^OBZ4P2WLbN[O@]4o1VLfNYO\\O`4n1ULkNWOYOc4m1TLnNVOUOf4n1RL4m3LSL5l3LSL5l3LSL5k3LUL4k3MSL4m3LSL5l3LSL4m3MhKlLMW3Z4NgKoLMS3[40gKoLLR3\\40gKPMKP3]41gK<X4c3N100O1O2O0O1O2N100O1O0001O01O01O1O101O00010O001O1O100O1O1O101N1O1dKbK]1_4aNcK_1]4^NgK`1Z4^NiK`1X4^NjKa1W4gL`Kg0<`2U4fLgKb07h2R4dLmK>4l2P4dLSL53V3k3bLYLD<i3\\3aLkM_3U2^LnMa3S2]LPNa3Q2]LQNb3P2[LSN7UMo2h4hLUN6UMR3g4eLWN6TMU3i7hLYHX3h7eLZH[3Q81O01O000^NfGHY8OPH1P8FZH9f7]OdHc0\\7TOmHl0T7jNUIW1j6_N`I`1]8O0O100O2O0O1kNiEj0W:SOmEk0a:N1N3M3M2N3M3M6Jkao0"}, "label": "a green leafy plant growing in a short blue vase in an office"}]}
{"id": 285548, "image": "COCO_train2014_000000285548.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green plant with a blue pot\"."}], "task": "refering", "answer_template": "The \"green plant with a blue pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 118, 119, 120, 121, 137, 138, 139, 155, 156, 157, 158, 173, 174, 175, 176, 191, 192, 193, 194, 209, 210, 211, 212], "bbox": [0.5278999999999999, 0.23071999999999998, 0.8260599999999999, 0.9036000000000001], "iscrowd": 0, "area": 18905.170000000002, "rle": {"size": [375, 500], "counts": "jjP36_;2N2N2N2O2M2N2N2N2K5L4K6L3000000000001N1000000000001O0O101O2N1O1O2N3M3L4M001O001O001O0O2O0oInM]2S2^MVN]2j1_M^NZ2f1aM`NU2i1]LPNUO8T4Q2XLYN@HV4Q2XL[N_OEX4Q2WL^N^OBZ4P2WLbN[O@]4o1VLfNYO\\O`4n1ULkNWOYOc4m1TLnNVOUOf4n1RL4m3LSL5l3LSL5l3LSL5k3LUL4k3MSL4m3LSL5l3LSL4m3MhKlLMW3Z4NgKoLMS3[40gKoLLR3\\40gKPMKP3]41gK<X4c3N100O1O2O0O1O2N100O1O0001O01O01O1O101O00010O001O1O100O1O1O101N1O1dKbK]1_4aNcK_1]4^NgK`1Z4^NiK`1X4^NjKa1W4gL`Kg0<`2U4fLgKb07h2R4dLmK>4l2P4dLSL53V3k3bLYLD<i3\\3aLkM_3U2^LnMa3S2]LPNa3Q2]LQNb3P2[LSN7UMo2h4hLUN6UMR3g4eLWN6TMU3i7hLYHX3h7eLZH[3Q81O01O000^NfGHY8OPH1P8FZH9f7]OdHc0\\7TOmHl0T7jNUIW1j6_N`I`1]8O0O100O2O0O1kNiEj0W:SOmEk0a:N1N3M3M2N3M3M6Jkao0"}, "label": "green plant with a blue pot"}]}
{"id": 47983, "image": "COCO_train2014_000000047983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra closest to the camera\"."}], "task": "refering", "answer_template": "The \"the zebra closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 295, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 357, 358, 360, 361, 362, 363, 364, 381, 383, 384, 385, 386, 387], "bbox": [0.5450468749999999, 0.6842916666666666, 0.844359375, 0.9699166666666665], "iscrowd": 0, "area": 16111.79805, "rle": {"size": [480, 640], "counts": "[oS55i>m0TO=B9G4M3N1N3N0O2N2N2M3N1N2O1N2N101N2N2O0O2N2O0O0100N2O001N2O1N3N1O1N2M3L5L3L5L3O2N1N3N110O1O010O0010O01O0101O0O100O10000O101N101N101O0O2O1O0O2O1O3M4K5L5XDTNR:]3O010000000O100000O100000OkM\\E0d:0`EL`:5cEF^::dED\\:<gEAY:`0hE^OX:b0jE\\OV:e0lEWOU:i0mEUOS:l0oEPOR:P1VFhNj9X1`FYNf9f1j100001O0000000001O01O000001O0000000000000O10000O101OO0100O10mDUNQ9k1kFZNT9e1jF^NV9a1jFaNT9`1kFbNT9]1kFeNU9Z1kFhNT9X1kFiNU9V1kFjNV9X1gFiNY9\\1aFeN^9_1]FbNd9a1WFaNi9a1SF`Nn9c1lE_NU:d1eE^N\\:e1]E^Nd:f1TE]Nm:j2001N2O1L4@`0J6O1O001O1O1O1O16J1TEoLg9Q3VFRMj9o2SFTMm9l2oEWMQ:i2lE[MT:c2kE_MU:a2iEaMW:_2gEcMZ:\\2bEgM_:]31O1N2O1N2N2M3N2M4H8B>I7H8TOm0^Ob0CWa^1"}, "label": "the zebra closest to the camera"}]}
{"id": 47983, "image": "COCO_train2014_000000047983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in the front\"."}], "task": "refering", "answer_template": "The \"zebra in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 295, 311, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 357, 358, 360, 361, 362, 363, 364, 381, 383, 384, 385, 386, 387], "bbox": [0.5450468749999999, 0.6842916666666666, 0.844359375, 0.9699166666666665], "iscrowd": 0, "area": 16111.79805, "rle": {"size": [480, 640], "counts": "[oS55i>m0TO=B9G4M3N1N3N0O2N2N2M3N1N2O1N2N101N2N2O0O2N2O0O0100N2O001N2O1N3N1O1N2M3L5L3L5L3O2N1N3N110O1O010O0010O01O0101O0O100O10000O101N101N101O0O2O1O0O2O1O3M4K5L5XDTNR:]3O010000000O100000O100000OkM\\E0d:0`EL`:5cEF^::dED\\:<gEAY:`0hE^OX:b0jE\\OV:e0lEWOU:i0mEUOS:l0oEPOR:P1VFhNj9X1`FYNf9f1j100001O0000000001O01O000001O0000000000000O10000O101OO0100O10mDUNQ9k1kFZNT9e1jF^NV9a1jFaNT9`1kFbNT9]1kFeNU9Z1kFhNT9X1kFiNU9V1kFjNV9X1gFiNY9\\1aFeN^9_1]FbNd9a1WFaNi9a1SF`Nn9c1lE_NU:d1eE^N\\:e1]E^Nd:f1TE]Nm:j2001N2O1L4@`0J6O1O001O1O1O1O16J1TEoLg9Q3VFRMj9o2SFTMm9l2oEWMQ:i2lE[MT:c2kE_MU:a2iEaMW:_2gEcMZ:\\2bEgM_:]31O1N2O1N2N2M3N2M4H8B>I7H8TOm0^Ob0CWa^1"}, "label": "zebra in the front"}]}
{"id": 113521, "image": "COCO_train2014_000000113521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike which man is sitting on\"."}], "task": "refering", "answer_template": "The \"bike which man is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 235, 236, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 286, 287, 305, 306, 328, 329, 337, 338, 339, 340, 350, 351, 352, 359, 360, 361, 362, 363, 364, 373, 374, 375, 382, 383, 384, 385, 386, 387], "bbox": [0.20074999999999998, 0.5807291666666666, 0.8713124999999999, 1.0], "iscrowd": 0, "area": 19251.98775, "rle": {"size": [480, 640], "counts": "dZl11m>:F3M4L4M2M4L4M2N3N1N3N0O2O0O2O00001N101O0UOcNSD]1l;fNQD[1g2bNh56_GX1f2eNi57_GT1e2hNk57PGG5[1n2iNl5c0QGd0P3lNm5U2PJnMo5S2nIPNQ6P2mISNR6m1lIVNR6l1jIXNU6h1iI[NW6d1gI_NY6a1dIbN[6^1cIeN]6Z1aIiN_6W1]ImNc6S1ZIPOf6Q1VIROj6n0SIUOm6k0lH\\OT7e0eHA[7?aHE_7;^HHb79ZHJf76VHNj72SH1m70oG3R8LkG7U8JgG9Z8FbG>^8B^Ga0d8`MZGR1M]1n8`MVGU4Q9jKPGP4V9oKkFk3[9kK^FM8R4`9PLZFL7Q4b9RLZFK4o3f9ULXFJ3m3k9VL`Fg3^:L4L4L4M3L4M3L4M3L44L4M3L5K4L4LfNdDUOW;j0mDWOo:7gEJT:0UF3e9HcF;W9BPG?k8]O\\Gf0^8WOjGj0P8SOYHm0a7POhHn0T7nNVIl0j6QO_Ii0a6TOjId0V6ZOTJ>l5B[J7e5HdJ0\\50fJNZ52hJJ[55hJGY59iJDX5<f4000000001O000000000000001O0000000000001N1000000000001O000O100O100O2O000O10HhAHW>8jAIU>7kAIT>7mAIS>7nAHQ>8PBIo=7QBIo=6SBIl=8TBHl=7UBJi=7WBIi=6YBIf=8ZBHf=7[BIe=7\\BIb=7_BIa=7_BI`=7aBI_=6cBJ[=7eBI[=6fBJY=7hBHX=7iBJU=7kBIU=7kBJT=5nBKP=6PCJP=5QCLm<5TCKk<4VCMh<4YCKg<4\\CKb<5`CK`<3bCM^<1dCN]<0eC0\\<JiC6a=00010O001O001O1O010O001O010OO101N1Jb[Y12ZdfN:H7K3L4L4K5L4K5N2N2N2O1N2O100O100O1O100O100O100O100O1O100O100O100O100O10000001O0000000000O100O100O100O100O100O100O10000O10000000000O10000000000O100000000000000000000000000000000000000001O000000001O00000000001O0XNdCl0\\<nNjCR1V<nNkCQ1U<oNkCQ1V<mNlCR1V<lNjCT1W<jNjCV1W<gNlCX1n<01O001O00001O002N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O3MQoU1"}, "label": "bike which man is sitting on"}]}
{"id": 113521, "image": "COCO_train2014_000000113521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle without the front bar\"."}], "task": "refering", "answer_template": "The \"a bicycle without the front bar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 235, 236, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 286, 287, 305, 306, 328, 329, 337, 338, 339, 340, 350, 351, 352, 359, 360, 361, 362, 363, 364, 373, 374, 375, 382, 383, 384, 385, 386, 387], "bbox": [0.20074999999999998, 0.5807291666666666, 0.8713124999999999, 1.0], "iscrowd": 0, "area": 19251.98775, "rle": {"size": [480, 640], "counts": "dZl11m>:F3M4L4M2M4L4M2N3N1N3N0O2O0O2O00001N101O0UOcNSD]1l;fNQD[1g2bNh56_GX1f2eNi57_GT1e2hNk57PGG5[1n2iNl5c0QGd0P3lNm5U2PJnMo5S2nIPNQ6P2mISNR6m1lIVNR6l1jIXNU6h1iI[NW6d1gI_NY6a1dIbN[6^1cIeN]6Z1aIiN_6W1]ImNc6S1ZIPOf6Q1VIROj6n0SIUOm6k0lH\\OT7e0eHA[7?aHE_7;^HHb79ZHJf76VHNj72SH1m70oG3R8LkG7U8JgG9Z8FbG>^8B^Ga0d8`MZGR1M]1n8`MVGU4Q9jKPGP4V9oKkFk3[9kK^FM8R4`9PLZFL7Q4b9RLZFK4o3f9ULXFJ3m3k9VL`Fg3^:L4L4L4M3L4M3L4M3L44L4M3L5K4L4LfNdDUOW;j0mDWOo:7gEJT:0UF3e9HcF;W9BPG?k8]O\\Gf0^8WOjGj0P8SOYHm0a7POhHn0T7nNVIl0j6QO_Ii0a6TOjId0V6ZOTJ>l5B[J7e5HdJ0\\50fJNZ52hJJ[55hJGY59iJDX5<f4000000001O000000000000001O0000000000001N1000000000001O000O100O100O2O000O10HhAHW>8jAIU>7kAIT>7mAIS>7nAHQ>8PBIo=7QBIo=6SBIl=8TBHl=7UBJi=7WBIi=6YBIf=8ZBHf=7[BIe=7\\BIb=7_BIa=7_BI`=7aBI_=6cBJ[=7eBI[=6fBJY=7hBHX=7iBJU=7kBIU=7kBJT=5nBKP=6PCJP=5QCLm<5TCKk<4VCMh<4YCKg<4\\CKb<5`CK`<3bCM^<1dCN]<0eC0\\<JiC6a=00010O001O001O1O010O001O010OO101N1Jb[Y12ZdfN:H7K3L4L4K5L4K5N2N2N2O1N2O100O100O1O100O100O100O100O1O100O100O100O100O10000001O0000000000O100O100O100O100O100O100O10000O10000000000O10000000000O100000000000000000000000000000000000000001O000000001O00000000001O0XNdCl0\\<nNjCR1V<nNkCQ1U<oNkCQ1V<mNlCR1V<lNjCT1W<jNjCV1W<gNlCX1n<01O001O00001O002N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O3MQoU1"}, "label": "a bicycle without the front bar"}]}
{"id": 506740, "image": "COCO_train2014_000000506740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cat with its eyes closed . m\"."}], "task": "refering", "answer_template": "The \"an orange cat with its eyes closed . m\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363, 383, 384], "bbox": [0.45112500000000005, 0.5115555555555555, 0.82109375, 0.8471296296296297], "iscrowd": 0, "area": 24414.77205, "rle": {"size": [540, 640], "counts": "^gh49]`0:H8I7I7I7J3M3L4K5L4L3YOnMbBU2]=mM`BU2`=nMZBU2e=a002O0O2N101N000001O0O1O100O0010O00010O01O01O01O010000000000000000000001O00000000000000001O00001O00010O00001O001O001O001O0010O01O1O1O2N1OiNcBPO\\=Z2NgNfBQOY=?iBeNNn0W=;mBgNLo0V=9oBiNKP1S=UOnB;6@IS1Q=QOQC:6BHT1P=oNSC97DGT1n<oNUC77FFU1j0nNf:0dD2;K@W1c0TOl:GgDM`01[OX1=VOQ;DkENgNZ17XOW;_OlEOgNb2];^MmE0fNb2];]MnE2dNa2^;\\MoE3dN`2];[MQF5bN`2];ZMSF5`Nb2\\;XMUF6`Na2[;XMVF6`Nb2Z;WMWF5]Nh2];QMWF6[Nk2^;nLYF4[Nn2\\;mL[F3ZNP3[;lL]F1ZNT3V=kLkBU3^=000000000010O000000000000O2O0000000O10001N100O100O100O2O0000000O1O1O1O1O1O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O010O100O10000000000001O1O1O1N3N1O1O001O1O001O1O001O1O001O1O001O1O2N1O1O1O2N1O1O1O2N1N2N3N1N2O1N2O2M2N2O1N3N1N2O1N3N1N2N2N3M3M3M3M3M3M3L6I7H9H8G;F9Fofk1"}, "label": "an orange cat with its eyes closed . m"}]}
{"id": 506740, "image": "COCO_train2014_000000506740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat under man ' s left arm\"."}], "task": "refering", "answer_template": "The \"a cat under man ' s left arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 221, 240, 241, 242, 243, 263, 264, 265, 266, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 294, 309, 310, 311, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363, 383, 384], "bbox": [0.45112500000000005, 0.5115555555555555, 0.82109375, 0.8471296296296297], "iscrowd": 0, "area": 24414.77205, "rle": {"size": [540, 640], "counts": "^gh49]`0:H8I7I7I7J3M3L4K5L4L3YOnMbBU2]=mM`BU2`=nMZBU2e=a002O0O2N101N000001O0O1O100O0010O00010O01O01O01O010000000000000000000001O00000000000000001O00001O00010O00001O001O001O001O0010O01O1O1O2N1OiNcBPO\\=Z2NgNfBQOY=?iBeNNn0W=;mBgNLo0V=9oBiNKP1S=UOnB;6@IS1Q=QOQC:6BHT1P=oNSC97DGT1n<oNUC77FFU1j0nNf:0dD2;K@W1c0TOl:GgDM`01[OX1=VOQ;DkENgNZ17XOW;_OlEOgNb2];^MmE0fNb2];]MnE2dNa2^;\\MoE3dN`2];[MQF5bN`2];ZMSF5`Nb2\\;XMUF6`Na2[;XMVF6`Nb2Z;WMWF5]Nh2];QMWF6[Nk2^;nLYF4[Nn2\\;mL[F3ZNP3[;lL]F1ZNT3V=kLkBU3^=000000000010O000000000000O2O0000000O10001N100O100O100O2O0000000O1O1O1O1O1O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O010O100O10000000000001O1O1O1N3N1O1O001O1O001O1O001O1O001O1O001O1O2N1O1O1O2N1O1O1O2N1N2N3N1N2O1N2O2M2N2O1N3N1N2O1N3N1N2N2N3M3M3M3M3M3M3L6I7H9H8G;F9Fofk1"}, "label": "a cat under man ' s left arm"}]}
{"id": 506740, "image": "COCO_train2014_000000506740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat with the red leash\"."}], "task": "refering", "answer_template": "The \"the cat with the red leash\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [310, 311, 312, 329, 330, 331, 332, 333, 334, 335, 352, 353, 354, 355, 356, 357, 358, 376, 379, 380, 381, 382], "bbox": [0.30771875, 0.6953333333333334, 0.6342187499999999, 0.8969814814814815], "iscrowd": 0, "area": 11365.7987, "rle": {"size": [540, 640], "counts": "`YX37d`06J2O2M2N3M2N2O2M2N2N2O0O2O1N101O1N101O1N101N2O001N2O001N2O0O101O0O101O00001O00000O1000O1000000O1000000O1000O11N101O001N101O001N10001N1O1O2N100O2N1O1O2N1O101O000O2OO10O1000O10O1000O0100000O1000000O1000001N100N2O1O1O2N1N3N1O2N1O2N1O2XA[NV>f1hA]NW>d1eA`NY>a1eAbNY>`1eAbNY>`1fAbNW>`1hAaNV>a1jA_NT>c1kA^NS>V2O1O100O1O1O1O001000000000O10000000O02O00000O2O0000001N1000001N1000001N10001O2N1O1O1O2N1N2O1O2N1O2N1ObAUNl=^21N1O2O0O1O2N001O1O2Nd0\\O;E4L3M4L3M3M2N1O00001N101N1O1N3M2N2N3M2O2MWnj3"}, "label": "the cat with the red leash"}]}
{"id": 506740, "image": "COCO_train2014_000000506740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat with the red leash around it\"."}], "task": "refering", "answer_template": "The \"the cat with the red leash around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [310, 311, 312, 329, 330, 331, 332, 333, 334, 335, 352, 353, 354, 355, 356, 357, 358, 376, 379, 380, 381, 382], "bbox": [0.30771875, 0.6953333333333334, 0.6342187499999999, 0.8969814814814815], "iscrowd": 0, "area": 11365.7987, "rle": {"size": [540, 640], "counts": "`YX37d`06J2O2M2N3M2N2O2M2N2N2O0O2O1N101O1N101O1N101N2O001N2O001N2O0O101O0O101O00001O00000O1000O1000000O1000000O1000O11N101O001N101O001N10001N1O1O2N100O2N1O1O2N1O101O000O2OO10O1000O10O1000O0100000O1000000O1000001N100N2O1O1O2N1N3N1O2N1O2N1O2XA[NV>f1hA]NW>d1eA`NY>a1eAbNY>`1eAbNY>`1fAbNW>`1hAaNV>a1jA_NT>c1kA^NS>V2O1O100O1O1O1O001000000000O10000000O02O00000O2O0000001N1000001N1000001N10001O2N1O1O1O2N1N2O1O2N1O2N1ObAUNl=^21N1O2O0O1O2N001O1O2Nd0\\O;E4L3M4L3M3M2N1O00001N101N1O1N3M2N2N3M2O2MWnj3"}, "label": "the cat with the red leash around it"}]}
{"id": 72565, "image": "COCO_train2014_000000072565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person standing in front of a motorcycle\"."}], "task": "refering", "answer_template": "The \"person standing in front of a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 362], "bbox": [0.47424999999999995, 0.31277533039647576, 0.7565625, 1.0], "iscrowd": 0, "area": 37914.5455, "rle": {"size": [454, 640], "counts": "aSW4>f=8G5L3L4K5E;N3M2N2N2N1O2@?M3M3M4M2O1N2O2]O]M`Ed2_:\\MaEe2^:[MaEf2`:YM`Eh2_:XM`Ej2_:UMaEm2^:SMaEn2_:RM`EP3_:PM`ER3`::O001O1O1O010O1N1L5K5O1N3N1O2N1TG\\MT6g2^IfMa6[2VInMi6T2UImMc6\\2ZIfMf5m5L4L4M3L4L4N2O1O1O1N2O1O1O1O1N2O1O1O100O1O100O1O100O10000O100000000O100000000000000001O000000001O00001O001O001O001O001O1O001O1O1O001O1O1O1O2N3M2N2N2N2N2N2N3M3M4L4L5K4WHYJk6c6N2N3M2N2N2N00001O0000001O00001oLbHXO]7i0fHTOZ7l0hHROX7n0kHoNU7Q1mHmNS7S1oHkNQ7T1SIiNm6W1UIgNl6W1XIfNi6Y1YIeNh6Z1ZIdNh6Y1\\IdNe6[1^IbNc6]1`I`Nb6]1bI`N_6_1cI_N^6^1gI_N[6]1jI`NW6]1nI`NT6Z1SJcNn5o0eJkN\\5R1TK`Nn4\\1eKSN]4i1T4K4J7I7I6J7H?B^lT2"}, "label": "person standing in front of a motorcycle"}]}
{"id": 72565, "image": "COCO_train2014_000000072565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a blue jacket with writing on the back\"."}], "task": "refering", "answer_template": "The \"a person wearing a blue jacket with writing on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 362], "bbox": [0.47424999999999995, 0.31277533039647576, 0.7565625, 1.0], "iscrowd": 0, "area": 37914.5455, "rle": {"size": [454, 640], "counts": "aSW4>f=8G5L3L4K5E;N3M2N2N2N1O2@?M3M3M4M2O1N2O2]O]M`Ed2_:\\MaEe2^:[MaEf2`:YM`Eh2_:XM`Ej2_:UMaEm2^:SMaEn2_:RM`EP3_:PM`ER3`::O001O1O1O010O1N1L5K5O1N3N1O2N1TG\\MT6g2^IfMa6[2VInMi6T2UImMc6\\2ZIfMf5m5L4L4M3L4L4N2O1O1O1N2O1O1O1O1N2O1O1O100O1O100O1O100O10000O100000000O100000000000000001O000000001O00001O001O001O001O001O1O001O1O1O001O1O1O1O2N3M2N2N2N2N2N2N3M3M4L4L5K4WHYJk6c6N2N3M2N2N2N00001O0000001O00001oLbHXO]7i0fHTOZ7l0hHROX7n0kHoNU7Q1mHmNS7S1oHkNQ7T1SIiNm6W1UIgNl6W1XIfNi6Y1YIeNh6Z1ZIdNh6Y1\\IdNe6[1^IbNc6]1`I`Nb6]1bI`N_6_1cI_N^6^1gI_N[6]1jI`NW6]1nI`NT6Z1SJcNn5o0eJkN\\5R1TK`Nn4\\1eKSN]4i1T4K4J7I7I6J7H?B^lT2"}, "label": "a person wearing a blue jacket with writing on the back"}]}
{"id": 72565, "image": "COCO_train2014_000000072565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red motorcylce\"."}], "task": "refering", "answer_template": "The \"the red motorcylce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [286, 288, 303, 304, 305, 307, 308, 309, 310, 311, 316, 317, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 362, 363, 364], "bbox": [0.07810937500000001, 0.7596916299559471, 0.8671875, 0.988920704845815], "iscrowd": 0, "area": 20366.285350000002, "rle": {"size": [454, 640], "counts": "Xcf01T>1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O10000O100O100O100O10000O100O100O100O100O10000O2O0O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O1000000O10000000000000000O10000000000000000O1000000000000O1000000O10001O0O10000000000001O000000001O000000001O00001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O000000O1000000O10000O1000000O1000000O100O1O2M2O1N2O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O15K;E6J0000001O00001O00001O01O00000000000O10000000000O10000000000O1000000000000O10000000000O100000H8D<D<G8K6K5J6K5K5K4L5LXRc1Nil\\NU1D<7IO1O100O1O1O100O1O1O100O1000000000001O000000000000000000000000000001O00001O0000001O0000001O001O1O1O2N1O1O1O1O1O2N1O100O1O1O2N1O1O2N1O1O1O2N1O1O2N2N2N2N3M2NUXU1"}, "label": "the red motorcylce"}]}
{"id": 72565, "image": "COCO_train2014_000000072565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red motorcycle parked in an enclosure\"."}], "task": "refering", "answer_template": "The \"a red motorcycle parked in an enclosure\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [286, 288, 303, 304, 305, 307, 308, 309, 310, 311, 316, 317, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 362, 363, 364], "bbox": [0.07810937500000001, 0.7596916299559471, 0.8671875, 0.988920704845815], "iscrowd": 0, "area": 20366.285350000002, "rle": {"size": [454, 640], "counts": "Xcf01T>1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O10000O100O100O100O10000O100O100O100O100O10000O2O0O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O1000000O10000000000000000O10000000000000000O1000000000000O1000000O10001O0O10000000000001O000000001O000000001O00001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O000000O1000000O10000O1000000O1000000O100O1O2M2O1N2O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O15K;E6J0000001O00001O00001O01O00000000000O10000000000O10000000000O1000000000000O10000000000O100000H8D<D<G8K6K5J6K5K5K4L5LXRc1Nil\\NU1D<7IO1O100O1O1O100O1O1O100O1000000000001O000000000000000000000000000001O00001O0000001O0000001O001O1O1O2N1O1O1O1O1O2N1O100O1O1O2N1O1O2N1O1O1O2N1O1O2N2N2N2N3M2NUXU1"}, "label": "a red motorcycle parked in an enclosure"}]}
{"id": 72565, "image": "COCO_train2014_000000072565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow motorcycle parked next to a red motorcycle\"."}], "task": "refering", "answer_template": "The \"a yellow motorcycle parked next to a red motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 214, 215, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 329, 330, 346], "bbox": [0.041640625, 0.5547356828193832, 0.4771875, 0.9881497797356829], "iscrowd": 0, "area": 23342.62799999999, "rle": {"size": [454, 640], "counts": "n\\<3P>3M3M4L3N2M3N3N1O001O1O1O1O1O1O1O01O0010O01O1O001N2O0O2O0O2N2O0O2O1O0O2O1O001N101O1O00100O0100O0100O010O10SN0dE0T:9kEFS:>lEBR:a0lE_OS:e0kE[OR:i0lEWOS:l0lETOQ:P1mEQOQ:R1nEmNP:X1mEiNP:[1nEfNo9]1QFcNm9`1QF`Nm9c1SF]Nj9g1TFZNi9i1VFXNg9l1XFSNg9o1XFRNe9R2ZFnMc9U2\\FlMa9X2]FiMb9X2^FgMc9Z2[FgMd9Z2\\FfMc9\\2[FeMd9\\2[FdMe9^2ZFbMe9_2ZFbMe9`2VFdMi9]2RFhMm9Z2mEjMS:U3O1O1O0010O000001O0001O0001O000001O010O1O001O010O01003M2N2N3L3N2N0^L_E_3b:21O1O1O1O1O1O1N101O1O1O1O1O1O00001O0O101O0000001O0O101O0O101O000O2O010O010OO2O0O2N2O0O2O0O2N2O0O2O1N2O1N3M2O1M3N2M3M3N2M3M3N3L]LXFm2e9UM\\Fk2b9VM_Fj2^9XMaFi2^9XMbFh2\\9YMeFg2Z9ZMfFf2X9\\MgFe2X9\\MhFe2V9\\MjFd2W9ZMjFf2W9YMhFh2X9XMhFh2Y9WMgFj2Y9UMgFk2Z9SMfFn2[9QMeFo2\\9PMdFP3\\9oLeFQ3\\9nLcFT3]9kLcFT3_9kLaFU3`9iL`FX3a9gL_FY3a9gL_FY3b9fL^FZ3c9dL]F\\3e9cL[F]3f9bLZF^3g9`LZF`3f9`LZF`3g9_LYFa3h9^LXFa3W:N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2O1N2N3N1N3N1N3N2M2O2N1N3N1N3N1O2M3N1O2N1O2N1O2O0O2N2O0O2N<E<C2N2O2N1N2O1O1N2O2N1N2O1N2O1N3M2OTVd4"}, "label": "a yellow motorcycle parked next to a red motorcycle"}]}
{"id": 72565, "image": "COCO_train2014_000000072565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bike named cibie near by red bike\"."}], "task": "refering", "answer_template": "The \"a yellow bike named cibie near by red bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 214, 215, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 329, 330, 346], "bbox": [0.041640625, 0.5547356828193832, 0.4771875, 0.9881497797356829], "iscrowd": 0, "area": 23342.62799999999, "rle": {"size": [454, 640], "counts": "n\\<3P>3M3M4L3N2M3N3N1O001O1O1O1O1O1O1O01O0010O01O1O001N2O0O2O0O2N2O0O2O1O0O2O1O001N101O1O00100O0100O0100O010O10SN0dE0T:9kEFS:>lEBR:a0lE_OS:e0kE[OR:i0lEWOS:l0lETOQ:P1mEQOQ:R1nEmNP:X1mEiNP:[1nEfNo9]1QFcNm9`1QF`Nm9c1SF]Nj9g1TFZNi9i1VFXNg9l1XFSNg9o1XFRNe9R2ZFnMc9U2\\FlMa9X2]FiMb9X2^FgMc9Z2[FgMd9Z2\\FfMc9\\2[FeMd9\\2[FdMe9^2ZFbMe9_2ZFbMe9`2VFdMi9]2RFhMm9Z2mEjMS:U3O1O1O0010O000001O0001O0001O000001O010O1O001O010O01003M2N2N3L3N2N0^L_E_3b:21O1O1O1O1O1O1N101O1O1O1O1O1O00001O0O101O0000001O0O101O0O101O000O2O010O010OO2O0O2N2O0O2O0O2N2O0O2O1N2O1N3M2O1M3N2M3M3N2M3M3N3L]LXFm2e9UM\\Fk2b9VM_Fj2^9XMaFi2^9XMbFh2\\9YMeFg2Z9ZMfFf2X9\\MgFe2X9\\MhFe2V9\\MjFd2W9ZMjFf2W9YMhFh2X9XMhFh2Y9WMgFj2Y9UMgFk2Z9SMfFn2[9QMeFo2\\9PMdFP3\\9oLeFQ3\\9nLcFT3]9kLcFT3_9kLaFU3`9iL`FX3a9gL_FY3a9gL_FY3b9fL^FZ3c9dL]F\\3e9cL[F]3f9bLZF^3g9`LZF`3f9`LZF`3g9_LYFa3h9^LXFa3W:N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2O1N2N3N1N3N1N3N2M2O2N1N3N1N3N1O2M3N1O2N1O2N1O2O0O2N2O0O2N<E<C2N2O2N1N2O1O1N2O2N1N2O1N2O1N3M2OTVd4"}, "label": "a yellow bike named cibie near by red bike"}]}
{"id": 441205, "image": "COCO_train2014_000000441205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan recliner with a person reading a newspaper\"."}], "task": "refering", "answer_template": "The \"a tan recliner with a person reading a newspaper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.3745625, 0.13258823529411765, 0.9371562499999999, 0.9865176470588236], "iscrowd": 0, "area": 94187.11409999995, "rle": {"size": [425, 640], "counts": "hjS33W=6F9H8G9G:F9H8G:F9H8J6J7^HZMS4l2RJiNn5Y40001O000000000O100000000000000000000000000000000O100000000001O00001O001O001O00001O0010O01O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O010O001O001O00001O001O001O00001O002N;E<D8H1O1O1O001O1O1O1O001O1O1O001O1O1O00M3N2M4L3M3N2M3M3M3M3N2M3M3M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3N2N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N200O1O100O100O100O100O100O100O1O100O100O100O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O100000O1000O1001O000000000000001O0000000000o0QOn1RNn1RNn1RNo1QNn1RNhV`0"}, "label": "a tan recliner with a person reading a newspaper"}]}
{"id": 441205, "image": "COCO_train2014_000000441205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beige sofa with a newspaper on it\"."}], "task": "refering", "answer_template": "The \"beige sofa with a newspaper on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 84, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 175, 176, 177, 178, 179, 180, 181, 182, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.3745625, 0.13258823529411765, 0.9371562499999999, 0.9865176470588236], "iscrowd": 0, "area": 94187.11409999995, "rle": {"size": [425, 640], "counts": "hjS33W=6F9H8G9G:F9H8G:F9H8J6J7^HZMS4l2RJiNn5Y40001O000000000O100000000000000000000000000000000O100000000001O00001O001O001O00001O0010O01O001O00001O001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O010O001O001O00001O001O001O00001O002N;E<D8H1O1O1O001O1O1O1O001O1O1O001O1O1O00M3N2M4L3M3N2M3M3M3M3N2M3M3M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3L4M3M3M3M3M3L4M3M3M3M3N2N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N200O1O100O100O100O100O100O100O1O100O100O100O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O100000O1000O1001O000000000000001O0000000000o0QOn1RNn1RNn1RNo1QNn1RNhV`0"}, "label": "beige sofa with a newspaper on it"}]}
{"id": 441205, "image": "COCO_train2014_000000441205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green chaor with blanket on back\"."}], "task": "refering", "answer_template": "The \"green chaor with blanket on back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.37816470588235296, 0.295296875, 0.9790352941176471], "iscrowd": 0, "area": 37622.3413, "rle": {"size": [425, 640], "counts": "S9k3b5l300000000000000000000O1000000000000000000000000000000O101O000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000001O4L3M4L4L3M4L4L4L3PKjIg2Z6mLnIS3V6bLQJ^3S6VLTJk3o5jKYJV4k5_K\\Ja4V7000000000O101O00000000000O100000000000000O100000000000000O100000000000000O11O2N1O1O2N1O1O2N1O2N1O1O2N1O1^M^G;d8BkG0V8MZHEg78iHZOY7c0VInNl6o0i2M4M2M4L3N2M4L3N2M4L3N2M4L]lj5"}, "label": "green chaor with blanket on back"}]}
{"id": 441205, "image": "COCO_train2014_000000441205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"corner of couch with throw blanket on top\"."}], "task": "refering", "answer_template": "The \"corner of couch with throw blanket on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.37816470588235296, 0.295296875, 0.9790352941176471], "iscrowd": 0, "area": 37622.3413, "rle": {"size": [425, 640], "counts": "S9k3b5l300000000000000000000O1000000000000000000000000000000O101O000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000001O4L3M4L4L3M4L4L4L3PKjIg2Z6mLnIS3V6bLQJ^3S6VLTJk3o5jKYJV4k5_K\\Ja4V7000000000O101O00000000000O100000000000000O100000000000000O100000000000000O11O2N1O1O2N1O1O2N1O2N1O1O2N1O1^M^G;d8BkG0V8MZHEg78iHZOY7c0VInNl6o0i2M4M2M4L3N2M4L3N2M4L3N2M4L]lj5"}, "label": "corner of couch with throw blanket on top"}]}
{"id": 547703, "image": "COCO_train2014_000000547703.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue jacket is picking up a frisbee\"."}], "task": "refering", "answer_template": "The \"a woman in a blue jacket is picking up a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 103, 104, 105, 113, 114, 115, 116, 117, 124, 125, 126, 127, 128, 129, 130, 136, 137, 138, 139, 140, 141, 142, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 172, 173, 174, 175, 176, 177, 178, 185, 186, 197], "bbox": [0.34355882352941175, 0.42844, 0.931235294117647, 0.9477000000000001], "iscrowd": 0, "area": 27575.580349999997, "rle": {"size": [500, 340], "counts": "c]i12`?4M3L4M3L4M3M3L4M3L4M3L4M3M2M2O1O001cCaNe:_1mDVOk:j0iDJo:7PE1i:OVE9b:H]E;`:E`E>]:CbE`0\\:@cEc04oMU9^1gFf0LZNT9Q1oFh0FeNR9c0WGk0@gNX9>XGX2g8iMXGW2i8iMVGX2i8iMVGW2k8hMUGY2j8hMUGX2l8gMUGY2j8hMUGX2l8gMUGY2j8hMUGX2l8hMTGX2l8gMTGZ2l8fMTG[2k8dMUG^2j8bMVG\\2l8dMSG[2o8dMRG[2o8eMoF[2S9dMjF^2X9bMdF`2^9_M`Fb2b9f12N2N2N2N2kN]FWLe9j3_FmKe9S4`FeKc9[4k0000000001O000000000001O010O00001OH`K\\E`4c:cK\\E\\4c:gK[EZ4c:hK]EW4b:lK\\ET4d:mK[ES4d:?0000O2O000O100O1N2O2N1N2K5J6N2O1O2N1O1O1O1O1O1O1O2N1O1O10000O101N100O100O10001OK4K6L4N2N2M201O2N100O2N101N1L5O002M2O1O1O1O1O1O1N2OcESKT:S1iE^22lLl9b0UFa2MoLP:;VFe2IRMS:5WFh2DUM\\;a3O100001O1O2M2O2N1O1N3N1N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2M3N3M2L5J5L4L5K4L4L5K4L6J6J6J6J6Jll:"}, "label": "a woman in a blue jacket is picking up a frisbee"}]}
{"id": 547703, "image": "COCO_train2014_000000547703.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue jacket and holding a frisbee about to be attacked by a dog\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue jacket and holding a frisbee about to be attacked by a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 103, 104, 105, 113, 114, 115, 116, 117, 124, 125, 126, 127, 128, 129, 130, 136, 137, 138, 139, 140, 141, 142, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 172, 173, 174, 175, 176, 177, 178, 185, 186, 197], "bbox": [0.34355882352941175, 0.42844, 0.931235294117647, 0.9477000000000001], "iscrowd": 0, "area": 27575.580349999997, "rle": {"size": [500, 340], "counts": "c]i12`?4M3L4M3L4M3M3L4M3L4M3L4M3M2M2O1O001cCaNe:_1mDVOk:j0iDJo:7PE1i:OVE9b:H]E;`:E`E>]:CbE`0\\:@cEc04oMU9^1gFf0LZNT9Q1oFh0FeNR9c0WGk0@gNX9>XGX2g8iMXGW2i8iMVGX2i8iMVGW2k8hMUGY2j8hMUGX2l8gMUGY2j8hMUGX2l8gMUGY2j8hMUGX2l8hMTGX2l8gMTGZ2l8fMTG[2k8dMUG^2j8bMVG\\2l8dMSG[2o8dMRG[2o8eMoF[2S9dMjF^2X9bMdF`2^9_M`Fb2b9f12N2N2N2N2kN]FWLe9j3_FmKe9S4`FeKc9[4k0000000001O000000000001O010O00001OH`K\\E`4c:cK\\E\\4c:gK[EZ4c:hK]EW4b:lK\\ET4d:mK[ES4d:?0000O2O000O100O1N2O2N1N2K5J6N2O1O2N1O1O1O1O1O1O1O2N1O1O10000O101N100O100O10001OK4K6L4N2N2M201O2N100O2N101N1L5O002M2O1O1O1O1O1O1N2OcESKT:S1iE^22lLl9b0UFa2MoLP:;VFe2IRMS:5WFh2DUM\\;a3O100001O1O2M2O2N1O1N3N1N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2M3N3M2L5J5L4L5K4L4L5K4L6J6J6J6J6Jll:"}, "label": "a woman wearing a blue jacket and holding a frisbee about to be attacked by a dog"}]}
{"id": 260982, "image": "COCO_train2014_000000260982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white hat running towards the man in green\"."}], "task": "refering", "answer_template": "The \"a man in a white hat running towards the man in green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 280, 281], "bbox": [0.05334375, 0.3709478672985782, 0.25042187499999996, 0.8608293838862559], "iscrowd": 0, "area": 11859.670899999997, "rle": {"size": [422, 640], "counts": "Y[>1R=3M2N3L4M3M2O2N2O1N1O2O1N2O0O2N2O1N1O2O0O2N101N1O2O0O2N1O2N010O1O1O1O100O0O2O1N2N2N2N2N2N1O2N2N2N2N2N2N2O1N2N2O1N2O1N2O1O1N2O2M2O1N2O1N2O2M2O1N2O1O2N10000O10001O1O1O1O2]HXLe5i3WJ\\Lg5f3TJ^Lj5d3QJaLn5`3nIdLQ6^3jIgLT6Z3iIiLV6Y3gIiLX6X3eIlLZ6T3cIoL]6R3_IQMa6o2\\ITMd6l2YIXMf6i2VIZMj6f2SI]Mm6d2PI_Mo6\\4O10001O00001O^L\\Ib0d6XOdIf0\\6UOkIi0V6TOnIk0S6ROPJl0Q6PNQIOQ1o1P6nMRI0R1Q2l5mMSI1T1P2j5lMUI1U1Q2h5kMUI2V1Q2f5kMVI1Y1R2b5iMXI3Z1Q2a5hMWI4\\1R2b5dMWI5[1U2]6fMmIQ2m9gN6J6K6IQ_U6"}, "label": "a man in a white hat running towards the man in green"}]}
{"id": 260982, "image": "COCO_train2014_000000260982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white running\"."}], "task": "refering", "answer_template": "The \"man in white running\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 280, 281], "bbox": [0.05334375, 0.3709478672985782, 0.25042187499999996, 0.8608293838862559], "iscrowd": 0, "area": 11859.670899999997, "rle": {"size": [422, 640], "counts": "Y[>1R=3M2N3L4M3M2O2N2O1N1O2O1N2O0O2N2O1N1O2O0O2N101N1O2O0O2N1O2N010O1O1O1O100O0O2O1N2N2N2N2N2N1O2N2N2N2N2N2N2O1N2N2O1N2O1N2O1O1N2O2M2O1N2O1N2O2M2O1N2O1O2N10000O10001O1O1O1O2]HXLe5i3WJ\\Lg5f3TJ^Lj5d3QJaLn5`3nIdLQ6^3jIgLT6Z3iIiLV6Y3gIiLX6X3eIlLZ6T3cIoL]6R3_IQMa6o2\\ITMd6l2YIXMf6i2VIZMj6f2SI]Mm6d2PI_Mo6\\4O10001O00001O^L\\Ib0d6XOdIf0\\6UOkIi0V6TOnIk0S6ROPJl0Q6PNQIOQ1o1P6nMRI0R1Q2l5mMSI1T1P2j5lMUI1U1Q2h5kMUI2V1Q2f5kMVI1Y1R2b5iMXI3Z1Q2a5hMWI4\\1R2b5dMWI5[1U2]6fMmIQ2m9gN6J6K6IQ_U6"}, "label": "man in white running"}]}
{"id": 490356, "image": "COCO_train2014_000000490356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"portfolio folder on the bottom in the arrangement\"."}], "task": "refering", "answer_template": "The \"portfolio folder on the bottom in the arrangement\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 41, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 132], "bbox": [0.30128125, 0.11067757009345794, 0.8098124999999999, 0.3528738317757009], "iscrowd": 0, "area": 16689.6956, "rle": {"size": [428, 640], "counts": "ag`21[=1O1N2O1O0O2O1O1N2O1O1N2O001N2O1O0O2O000O101O0O10000O2O000O101O0O10001N10000O101O0O10001N10000O2O000O10001N10000O2O000O101O0O10000O20O000001O0001O01O00000010O00000001O01O0001O0000010O0000000010O000001O01O0001O000001O01O00000010O000001O0001O0001O0000010O00000010O00000001O01O0001O0000010O0000000010O000001O0001O01O000000010O00000010O000001O0001O0001O0000010O0000001O01O0001O000001O01O0000010O0000001O01O000001O0001O01O00000010O00000001O01O0001O0001O01O000000010O0000001O01O0001O000001O01O00000010O000001O0001O0001O0001O01O00000010O00000001O01O0001O0000010O00000O1O2N100O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O00O1N2N2O1N2O1N2O1N2M3I7I7J7H8H8H8H8I7Hanb1"}, "label": "portfolio folder on the bottom in the arrangement"}]}
{"id": 490356, "image": "COCO_train2014_000000490356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the menu on the bottom of the stack\"."}], "task": "refering", "answer_template": "The \"the menu on the bottom of the stack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 41, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 132], "bbox": [0.30128125, 0.11067757009345794, 0.8098124999999999, 0.3528738317757009], "iscrowd": 0, "area": 16689.6956, "rle": {"size": [428, 640], "counts": "ag`21[=1O1N2O1O0O2O1O1N2O1O1N2O001N2O1O0O2O000O101O0O10000O2O000O101O0O10001N10000O101O0O10001N10000O2O000O10001N10000O2O000O101O0O10000O20O000001O0001O01O00000010O00000001O01O0001O0000010O0000000010O000001O01O0001O000001O01O00000010O000001O0001O0001O0000010O00000010O00000001O01O0001O0000010O0000000010O000001O0001O01O000000010O00000010O000001O0001O0001O0000010O0000001O01O0001O000001O01O0000010O0000001O01O000001O0001O01O00000010O00000001O01O0001O0001O01O000000010O0000001O01O0001O000001O01O00000010O000001O0001O0001O0001O01O00000010O00000001O01O0001O0000010O00000O1O2N100O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O00O1N2N2O1N2O1N2O1N2M3I7I7J7H8H8H8H8I7Hanb1"}, "label": "the menu on the bottom of the stack"}]}
{"id": 7035, "image": "COCO_train2014_000000007035.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse in the back\"."}], "task": "refering", "answer_template": "The \"horse in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 198, 199, 200, 201, 202, 222, 223, 224, 243, 245, 247, 266, 268, 270], "bbox": [0.57775, 0.5270448548812665, 0.8037812499999999, 0.8178100263852243], "iscrowd": 0, "area": 4730.946849999998, "rle": {"size": [379, 640], "counts": "RWY4:`;6M3LO2AkD4`;N1O2MhZ74QeH4K6K4L5L4L3N3L41N01O1O001O001O001O00001O02O1N2N010O10O01O10O01O1O1O00]FTOX8k0bG^O\\8a0^GG[O\\On8l0aG6`8IXG0E@T9?TG1L_OR9>RGO2AP9<mF08Co86kF4;Dn81jF:<C_9;eFDZ9;iFD_70TI:BD[76mH5\\8O_G0c83YGKj87SGIo88nFHS9;jFDW9>fFA\\9b0`F^Oa9g0YFYOh9k0SFWOl9m0oETOQ:\\1O100O1O0O2O1O1N2O1O0O2OAXFlNg9R1]FmNa9R1cFmN\\9Q1gFnNY9P1kFoNS9P1QGoNn8l0aFSO;2R9e0hF\\O60o8b0kFA7Nk8`0nFC89]82[GG:7Y83YGIb02S8o0oGoN@]OV8d1^HUOb7k0aHRO_7m0fHoNZ7P1kHlNU7T1PIgNP7X1TIeNl6[1WIcNh6\\1l1O2N3L3N2N3M2N2N2N3M2N2N3M3N2M2O2N2O0OO10I600OXi^1"}, "label": "horse in the back"}]}
{"id": 7035, "image": "COCO_train2014_000000007035.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with it ' s head down in the background\"."}], "task": "refering", "answer_template": "The \"a horse with it ' s head down in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 198, 199, 200, 201, 202, 222, 223, 224, 243, 245, 247, 266, 268, 270], "bbox": [0.57775, 0.5270448548812665, 0.8037812499999999, 0.8178100263852243], "iscrowd": 0, "area": 4730.946849999998, "rle": {"size": [379, 640], "counts": "RWY4:`;6M3LO2AkD4`;N1O2MhZ74QeH4K6K4L5L4L3N3L41N01O1O001O001O001O00001O02O1N2N010O10O01O10O01O1O1O00]FTOX8k0bG^O\\8a0^GG[O\\On8l0aG6`8IXG0E@T9?TG1L_OR9>RGO2AP9<mF08Co86kF4;Dn81jF:<C_9;eFDZ9;iFD_70TI:BD[76mH5\\8O_G0c83YGKj87SGIo88nFHS9;jFDW9>fFA\\9b0`F^Oa9g0YFYOh9k0SFWOl9m0oETOQ:\\1O100O1O0O2O1O1N2O1O0O2OAXFlNg9R1]FmNa9R1cFmN\\9Q1gFnNY9P1kFoNS9P1QGoNn8l0aFSO;2R9e0hF\\O60o8b0kFA7Nk8`0nFC89]82[GG:7Y83YGIb02S8o0oGoN@]OV8d1^HUOb7k0aHRO_7m0fHoNZ7P1kHlNU7T1PIgNP7X1TIeNl6[1WIcNh6\\1l1O2N3L3N2N3M2N2N2N3M2N2N3M3N2M2O2N2O0OO10I600OXi^1"}, "label": "a horse with it ' s head down in the background"}]}
{"id": 7035, "image": "COCO_train2014_000000007035.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse closest to the camera\"."}], "task": "refering", "answer_template": "The \"the horse closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 105, 106, 107, 108, 127, 128, 129, 130, 131, 145, 146, 147, 148, 150, 151, 152, 153, 154, 168, 169, 170, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 213, 214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 239, 242, 243, 244, 260, 262, 265, 266, 267, 283, 284, 285, 288, 289, 290, 311, 312], "bbox": [0.293421875, 0.23701846965699208, 0.7360312499999999, 0.9797361477572559], "iscrowd": 0, "area": 24260.962100000004, "rle": {"size": [379, 640], "counts": "PkU27`;4L4REDZ:?cEDV:c0gE@o9j0oEXOh9R1TFSOf9d1K3N2N3L3N100O1O01O02N201N100TG[Mb8g2ZG`Ma8a2]GdM`8c3VHULT6V4^IoKa6f4N20000O10002M4MgNiIPMU6k2\\JlL`5l2mJPMS5m2QKRMn4l2WKRMi4l2ZKSMf4k2^KSMa4l2aKTM_4j2dKUM\\4h2hKWMW4g2mKXMS4e2PL[MP4b2TL^Mj3_2[L`Me3]2_LbMa3\\2bLcM]3\\2[JaMf1NS4`2WJdMP8[2QHhMm7W2THjMj7U2WHnMf7R2ZHQNd7m1]HVN`7j1`HYN^7e1dH]NY7`1jHcNT7^1jHdNU7]1hHgNU7Z1jHiNT7X1jHjNU7W1iHkNV7V1gHnNW7T1eHoNZ7R1bHSO\\7h2MYOiHnLU7P3nHQMP7m2SITMk6k2WIUMh6i2\\IWMc6g2_IZMM]Om5W3ZJ\\MEAo5R3_J]M^OEQ6j0SJ`0`0ROXOHS6b0]J?;XOPOIY6;dJ?6NU5^OnJ?02Q5[OSKb0N1Q5[ORKc00Oo4^OQKb03Mm4@QKb05Kl4APKc08Gj4FnJb0a0^Od4NlJc0e0WOc45iJc0b7]O^Hb0d7\\O\\Hd0e7[O\\Hd0f7[OYHe0h7ZOYHe0i7YOXHf0j7YOVHf0l7XOUHh0_9O1O1M4O01O1O1O00001O000010O0001O00001O001O001O01O01O001O000cj4]OeUK;D<H81O00000000000010O002M2N2hGgNl5Z1nIROj5P1oIYOm5j0lI]OQ6e0iIAV6`0bIH\\6:[I0b62UI7h6LoH>m6EjHd0T7T2]N[KjKo3iNbLY5BlKi3SOcLn4GmKa3^OgLa4JnK]3GiLY4LmKX31lLn3NoKS39oLf30mKo2c0RM\\3Y4iLfKV3W4oLhKo2W4VMgKi2W4[MhKd2V4YMdKRN4c4W4YMlKTNKb4V4]MRLQNFa4X4^MTLQNC_4\\4^MQLVNA[4`4]MoK[N^OX4e4\\MlKl2U4SMkKm2W4QMhKP3Z4PMdKP3^4oL`KR3b4nL\\KR3f4jL\\KV3f4fL\\KZ3f4_L`K`3b4^L^Kb3d4\\L\\Kd3f4ZLZKf3h4XLXKh3j4ULWKl3k4QLTKP4n4nKRKR4P5kKQKU4e3\\KkL;BY4c3cKhLNG`4_3fKPM@Bj4^3iKfMZ4Y2jK`M[4^2hK\\M\\4c2hKWM[4g2jKSMZ4j2kKRMV4l2nKQMT4k2RLQMo3j2YLRMh3k2]LRMe3j2aLSM`3j2eLRM\\3k2iLnL]3o2h2N3M3O0O2O0O2O1L4L4L4L4K5L4K5K5O10N3M2N2O1N2N3O0O1OQN^Fh1`9UNdFk1f91N2N1O2O1N1O2NVN_F\\1`9cNcF\\1\\9cNgF\\1X9cNlF\\1R9cNQG\\1n8dNTG\\1j8cNZG]1b9001O0O2O0O0:F1O2N1O2N1O2N1O2N2N3M2N1O2M3M3L4L4L4L`\\n1"}, "label": "the horse closest to the camera"}]}
{"id": 7035, "image": "COCO_train2014_000000007035.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with head up\"."}], "task": "refering", "answer_template": "The \"a horse with head up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 105, 106, 107, 108, 127, 128, 129, 130, 131, 145, 146, 147, 148, 150, 151, 152, 153, 154, 168, 169, 170, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 213, 214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 239, 242, 243, 244, 260, 262, 265, 266, 267, 283, 284, 285, 288, 289, 290, 311, 312], "bbox": [0.293421875, 0.23701846965699208, 0.7360312499999999, 0.9797361477572559], "iscrowd": 0, "area": 24260.962100000004, "rle": {"size": [379, 640], "counts": "PkU27`;4L4REDZ:?cEDV:c0gE@o9j0oEXOh9R1TFSOf9d1K3N2N3L3N100O1O01O02N201N100TG[Mb8g2ZG`Ma8a2]GdM`8c3VHULT6V4^IoKa6f4N20000O10002M4MgNiIPMU6k2\\JlL`5l2mJPMS5m2QKRMn4l2WKRMi4l2ZKSMf4k2^KSMa4l2aKTM_4j2dKUM\\4h2hKWMW4g2mKXMS4e2PL[MP4b2TL^Mj3_2[L`Me3]2_LbMa3\\2bLcM]3\\2[JaMf1NS4`2WJdMP8[2QHhMm7W2THjMj7U2WHnMf7R2ZHQNd7m1]HVN`7j1`HYN^7e1dH]NY7`1jHcNT7^1jHdNU7]1hHgNU7Z1jHiNT7X1jHjNU7W1iHkNV7V1gHnNW7T1eHoNZ7R1bHSO\\7h2MYOiHnLU7P3nHQMP7m2SITMk6k2WIUMh6i2\\IWMc6g2_IZMM]Om5W3ZJ\\MEAo5R3_J]M^OEQ6j0SJ`0`0ROXOHS6b0]J?;XOPOIY6;dJ?6NU5^OnJ?02Q5[OSKb0N1Q5[ORKc00Oo4^OQKb03Mm4@QKb05Kl4APKc08Gj4FnJb0a0^Od4NlJc0e0WOc45iJc0b7]O^Hb0d7\\O\\Hd0e7[O\\Hd0f7[OYHe0h7ZOYHe0i7YOXHf0j7YOVHf0l7XOUHh0_9O1O1M4O01O1O1O00001O000010O0001O00001O001O001O01O01O001O000cj4]OeUK;D<H81O00000000000010O002M2N2hGgNl5Z1nIROj5P1oIYOm5j0lI]OQ6e0iIAV6`0bIH\\6:[I0b62UI7h6LoH>m6EjHd0T7T2]N[KjKo3iNbLY5BlKi3SOcLn4GmKa3^OgLa4JnK]3GiLY4LmKX31lLn3NoKS39oLf30mKo2c0RM\\3Y4iLfKV3W4oLhKo2W4VMgKi2W4[MhKd2V4YMdKRN4c4W4YMlKTNKb4V4]MRLQNFa4X4^MTLQNC_4\\4^MQLVNA[4`4]MoK[N^OX4e4\\MlKl2U4SMkKm2W4QMhKP3Z4PMdKP3^4oL`KR3b4nL\\KR3f4jL\\KV3f4fL\\KZ3f4_L`K`3b4^L^Kb3d4\\L\\Kd3f4ZLZKf3h4XLXKh3j4ULWKl3k4QLTKP4n4nKRKR4P5kKQKU4e3\\KkL;BY4c3cKhLNG`4_3fKPM@Bj4^3iKfMZ4Y2jK`M[4^2hK\\M\\4c2hKWM[4g2jKSMZ4j2kKRMV4l2nKQMT4k2RLQMo3j2YLRMh3k2]LRMe3j2aLSM`3j2eLRM\\3k2iLnL]3o2h2N3M3O0O2O0O2O1L4L4L4L4K5L4K5K5O10N3M2N2O1N2N3O0O1OQN^Fh1`9UNdFk1f91N2N1O2O1N1O2NVN_F\\1`9cNcF\\1\\9cNgF\\1X9cNlF\\1R9cNQG\\1n8dNTG\\1j8cNZG]1b9001O0O2O0O0:F1O2N1O2N1O2N1O2N2N3M2N1O2M3M3L4L4L4L`\\n1"}, "label": "a horse with head up"}]}
{"id": 7035, "image": "COCO_train2014_000000007035.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rider on the larger horse\"."}], "task": "refering", "answer_template": "The \"the rider on the larger horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 102, 103, 104, 125, 126, 127, 148, 149, 171, 172, 195, 218, 221, 241], "bbox": [0.447140625, 0.155065963060686, 0.648078125, 0.7573087071240105], "iscrowd": 0, "area": 6969.606300000001, "rle": {"size": [379, 640], "counts": "aoY33`;:F:E;G9M3N2M3XOaNPGa1h8fNVG]1c8iN\\GP2j7RNTHQ2i7RNUHP2h7RNVHR2f7PNYHR2d7PNZHS2b7PN]HS2_7oM_HT2^7nMaHT2]7W1N3M2N2N3M2N7I8H9G9G8H9G6K4K5K3N00[JSKZ4n4dKTK\\4l4cKUK\\4l4bKWK]4k4\\KZKc4n4jJ^Km0Bh2S6UM_Jb1WOoM[6=jJU1SO[NS6?RKk0SOaNl5c0YKa0ROdNj5j0\\K7g5HW3O1O0O10000000000000O100000[O\\E3d:Lf0N2O1O1O[b>2a]A4M4L2HFlD=T;51O10O01N2N1O2O1N1O2O1N1O2O0O2N2N10lPc2"}, "label": "the rider on the larger horse"}]}
{"id": 7035, "image": "COCO_train2014_000000007035.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person riding the horse that is in the center front area\"."}], "task": "refering", "answer_template": "The \"the person riding the horse that is in the center front area\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 79, 80, 102, 103, 104, 125, 126, 127, 148, 149, 171, 172, 195, 218, 221, 241], "bbox": [0.447140625, 0.155065963060686, 0.648078125, 0.7573087071240105], "iscrowd": 0, "area": 6969.606300000001, "rle": {"size": [379, 640], "counts": "aoY33`;:F:E;G9M3N2M3XOaNPGa1h8fNVG]1c8iN\\GP2j7RNTHQ2i7RNUHP2h7RNVHR2f7PNYHR2d7PNZHS2b7PN]HS2_7oM_HT2^7nMaHT2]7W1N3M2N2N3M2N7I8H9G9G8H9G6K4K5K3N00[JSKZ4n4dKTK\\4l4cKUK\\4l4bKWK]4k4\\KZKc4n4jJ^Km0Bh2S6UM_Jb1WOoM[6=jJU1SO[NS6?RKk0SOaNl5c0YKa0ROdNj5j0\\K7g5HW3O1O0O10000000000000O100000[O\\E3d:Lf0N2O1O1O[b>2a]A4M4L2HFlD=T;51O10O01N2N1O2O1N1O2O1N1O2O0O2N2N10lPc2"}, "label": "the person riding the horse that is in the center front area"}]}
{"id": 473985, "image": "COCO_train2014_000000473985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaping in the air to catch a frisbee\"."}], "task": "refering", "answer_template": "The \"a man leaping in the air to catch a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 130, 131, 152, 153, 154, 175, 176, 177, 199, 200, 201, 222, 223, 224, 225, 226, 246, 247, 248, 249, 250, 270, 271, 272, 273, 294], "bbox": [0.573171875, 0.2458125, 0.8796406250000001, 0.7665208333333333], "iscrowd": 0, "area": 11965.339349999998, "rle": {"size": [480, 640], "counts": "jT\\5>b>7I5J6K3L2M30O2O001O001O001OO10O1000O100O1O100O1O100O100O100O100O100O100O2M6K4K6J5L3L3N2N2N2N2N2N2O1N2N2N2N3N1N2N2N2O0O2N100O101N100O1103L3N2NO1N2N2O1N2N2O1N2O1N2N2O1<C5K5K5K4L1O1O1O1O001O1O1O0O2O1O00001O0000001O0000001O0jN\\COd<1\\COd<0^C0a<0_C0a<O`C1`<NbC1^<ObC1^<NcC2]<NdC1\\<NeC2[<NeC2[<MgC3X<MhC3X<LiC4W<LiC4W<KkC4U<LkC4T<LmC4S<LnC3R<LoC4Q<KPD6o;JRD5n;JSD6l;KTD5l;JUD6k;JVD5j;JWD6h;KXD5h;JZD5f;KZD5e;K\\D6c;J^D5b;J_D6a;I`D7_;JbD5^;JcD6\\;KdD5[;KfD5Y;LhD3\\OUO\\;h0YE3WOXO`;f0YE2RO[Of;c0ZE<f:EZE:g:EYE;h:EWE:k:EVE:k:EUE;l:ESE;m:ETE5R;KmD1X;NiDL];4bDGd;8\\DCj;=VD]Oo;c0X100O2O0O100O1O100O102M20O00001N1000001N10O100O1O1O1O1O1O100O]iS1"}, "label": "a man leaping in the air to catch a frisbee"}]}
{"id": 473985, "image": "COCO_train2014_000000473985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one man is flying on the ground to play game\"."}], "task": "refering", "answer_template": "The \"one man is flying on the ground to play game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 130, 131, 152, 153, 154, 175, 176, 177, 199, 200, 201, 222, 223, 224, 225, 226, 246, 247, 248, 249, 250, 270, 271, 272, 273, 294], "bbox": [0.573171875, 0.2458125, 0.8796406250000001, 0.7665208333333333], "iscrowd": 0, "area": 11965.339349999998, "rle": {"size": [480, 640], "counts": "jT\\5>b>7I5J6K3L2M30O2O001O001O001OO10O1000O100O1O100O1O100O100O100O100O100O100O2M6K4K6J5L3L3N2N2N2N2N2N2O1N2N2N2N3N1N2N2N2O0O2N100O101N100O1103L3N2NO1N2N2O1N2N2O1N2O1N2N2O1<C5K5K5K4L1O1O1O1O001O1O1O0O2O1O00001O0000001O0000001O0jN\\COd<1\\COd<0^C0a<0_C0a<O`C1`<NbC1^<ObC1^<NcC2]<NdC1\\<NeC2[<NeC2[<MgC3X<MhC3X<LiC4W<LiC4W<KkC4U<LkC4T<LmC4S<LnC3R<LoC4Q<KPD6o;JRD5n;JSD6l;KTD5l;JUD6k;JVD5j;JWD6h;KXD5h;JZD5f;KZD5e;K\\D6c;J^D5b;J_D6a;I`D7_;JbD5^;JcD6\\;KdD5[;KfD5Y;LhD3\\OUO\\;h0YE3WOXO`;f0YE2RO[Of;c0ZE<f:EZE:g:EYE;h:EWE:k:EVE:k:EUE;l:ESE;m:ETE5R;KmD1X;NiDL];4bDGd;8\\DCj;=VD]Oo;c0X100O2O0O100O1O100O102M20O00001N1000001N10O100O1O1O1O1O1O100O]iS1"}, "label": "one man is flying on the ground to play game"}]}
{"id": 170882, "image": "COCO_train2014_000000170882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef in white showing off\"."}], "task": "refering", "answer_template": "The \"a chef in white showing off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 56, 57, 58, 59, 79, 80, 81, 82, 96, 97, 102, 103, 104, 105, 111, 119, 120, 121, 126, 127, 128, 129, 130, 131, 133, 134, 135, 143, 144, 145, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 267, 268, 290, 291], "bbox": [0.18426562500000002, 0.12583941605839416, 0.8918437499999999, 0.8404379562043794], "iscrowd": 0, "area": 44548.32190000002, "rle": {"size": [411, 640], "counts": "R__15d<301N1O1O100RDKT;5kD1Q;NnD6P;IoD:P;ASEb0a;2N1O2F]OgD6T;3iDMV;7gDIY;=bDA_;m01E_DYOd;0\\Dd0f;YOcDa0];\\OdDe0h;EPDNn;a0F9O200001N10000O10001O0O101O001NVEUOe9k0VF[Oj9d0RFAm9?nEFS:9iEJX:6^E3c:MZE5g:KVE7k:l0010O1O1O1O1O00100O1O001O001O0010O01O010O1O010O1O010O10O01O100O001000O01000O10O10O1000O010000O01000O1O001O1O0O1000000000000O11O2N2N1O1O00000000000000O100O100O100O100O1000O10O100000000O10000000000O010000000O1O100O1O1O100O1O010O1O1O100O1O100O100O1O3N3L3N3L3M4M3L3N3L3TMoLQJOh0V3S5PMQJLj0W3Q5RMPJIm0X3P5TMnIGP1X3n4VMmIER1Y3m4XMkIAW1Z3j4ZMjI_OY1Z3j4kMTKX2h4jMWKY2d4kMZKX2b4jM]KX2a4jM]KV2b4mM\\KS2c4oM\\KP2d4QN\\Ko1d4QN\\Ko1c4QN^Kn1b4SN^Km1a4VN]Ki1d4YNYKh1f4[NXKe1h4]NVKb1j4aNTK_1l4dNQK\\1n4gNPKX1Q5jNlJW1S5lNkJT1U5POgJo0Y5UOdJk0\\5XOaJh0^5Z310O10O1000O1000O10000000000O10000000000O2O001N101N2O0O2O001N13NO001O001O001O010O1oJSJR3o5kLTJS3m5jLVJU3h6kK\\IS4d6lK_IR4a6lKbIS4_6jKdIU4\\6iKgIW4X6hKjIW4V6gKmIX4T6aKRJ_4V7O1O1O1O010O1O1O10O01O1O100O001O10N1M3M4N100XI^KW5c4hJ]KX5c4fJ_KZ5b4cJ`K]5`4aJcK^5]4`JeK`5\\4]JfKc5Z4[JhKe5Y4XJiKh5W4VJkKj5U4TJnKk5S4RJoKn5Q4PJQLP6R4kIPLT6S4iInKW6U4dInK[6T4aInK_6W501O00000012M4L2N00001O001AWIjJi6T5ZIkJg6S5\\IjJe6T5^IkJb6T5`IkJa6S5bIkJ^6S5eIlJ\\6R5fImJZ6Q5i0O2M2YMTH8T8GUHJQ86XHZOo7f0T2N101O001O00001N101O00001O001N100010O01O010O01O010O010O01O01O010N0100O1O010O10O01O10O0100000O1000O010000O1000000O01000O10000O10O1000O10000O10000O10000O1O1N2O1O1O1O1N2O1O001O1N2OlNeEJY:7hEIW:6kEIT:8mEHQ:9PFGn9:SFFk9=TFCk9>UFBi9a0VF_Oh9c0XF\\Og9f0YFZOe9i0ZFWOf9i0ZFWOf9i0[FVOe9i0\\FUOf9@[Fe0OIh9_O\\Fh0MFi9_O^Fj0JEi9_O_Fk0JCi9_O`Fn0H@j9_ObFP1E^Oc:b0^E[Od:d0_EYOb:g0`EUOc:j0_EROc:n0e0O2O002N1O2L4K5J5L5KYak0"}, "label": "a chef in white showing off"}]}
{"id": 170882, "image": "COCO_train2014_000000170882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing behind food dishes with his arms out\"."}], "task": "refering", "answer_template": "The \"a man standing behind food dishes with his arms out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 56, 57, 58, 59, 79, 80, 81, 82, 96, 97, 102, 103, 104, 105, 111, 119, 120, 121, 126, 127, 128, 129, 130, 131, 133, 134, 135, 143, 144, 145, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 267, 268, 290, 291], "bbox": [0.18426562500000002, 0.12583941605839416, 0.8918437499999999, 0.8404379562043794], "iscrowd": 0, "area": 44548.32190000002, "rle": {"size": [411, 640], "counts": "R__15d<301N1O1O100RDKT;5kD1Q;NnD6P;IoD:P;ASEb0a;2N1O2F]OgD6T;3iDMV;7gDIY;=bDA_;m01E_DYOd;0\\Dd0f;YOcDa0];\\OdDe0h;EPDNn;a0F9O200001N10000O10001O0O101O001NVEUOe9k0VF[Oj9d0RFAm9?nEFS:9iEJX:6^E3c:MZE5g:KVE7k:l0010O1O1O1O1O00100O1O001O001O0010O01O010O1O010O1O010O10O01O100O001000O01000O10O10O1000O010000O01000O1O001O1O0O1000000000000O11O2N2N1O1O00000000000000O100O100O100O100O1000O10O100000000O10000000000O010000000O1O100O1O1O100O1O010O1O1O100O1O100O100O1O3N3L3N3L3M4M3L3N3L3TMoLQJOh0V3S5PMQJLj0W3Q5RMPJIm0X3P5TMnIGP1X3n4VMmIER1Y3m4XMkIAW1Z3j4ZMjI_OY1Z3j4kMTKX2h4jMWKY2d4kMZKX2b4jM]KX2a4jM]KV2b4mM\\KS2c4oM\\KP2d4QN\\Ko1d4QN\\Ko1c4QN^Kn1b4SN^Km1a4VN]Ki1d4YNYKh1f4[NXKe1h4]NVKb1j4aNTK_1l4dNQK\\1n4gNPKX1Q5jNlJW1S5lNkJT1U5POgJo0Y5UOdJk0\\5XOaJh0^5Z310O10O1000O1000O10000000000O10000000000O2O001N101N2O0O2O001N13NO001O001O001O010O1oJSJR3o5kLTJS3m5jLVJU3h6kK\\IS4d6lK_IR4a6lKbIS4_6jKdIU4\\6iKgIW4X6hKjIW4V6gKmIX4T6aKRJ_4V7O1O1O1O010O1O1O10O01O1O100O001O10N1M3M4N100XI^KW5c4hJ]KX5c4fJ_KZ5b4cJ`K]5`4aJcK^5]4`JeK`5\\4]JfKc5Z4[JhKe5Y4XJiKh5W4VJkKj5U4TJnKk5S4RJoKn5Q4PJQLP6R4kIPLT6S4iInKW6U4dInK[6T4aInK_6W501O00000012M4L2N00001O001AWIjJi6T5ZIkJg6S5\\IjJe6T5^IkJb6T5`IkJa6S5bIkJ^6S5eIlJ\\6R5fImJZ6Q5i0O2M2YMTH8T8GUHJQ86XHZOo7f0T2N101O001O00001N101O00001O001N100010O01O010O01O010O010O01O01O010N0100O1O010O10O01O10O0100000O1000O010000O1000000O01000O10000O10O1000O10000O10000O10000O1O1N2O1O1O1O1N2O1O001O1N2OlNeEJY:7hEIW:6kEIT:8mEHQ:9PFGn9:SFFk9=TFCk9>UFBi9a0VF_Oh9c0XF\\Og9f0YFZOe9i0ZFWOf9i0ZFWOf9i0[FVOe9i0\\FUOf9@[Fe0OIh9_O\\Fh0MFi9_O^Fj0JEi9_O_Fk0JCi9_O`Fn0H@j9_ObFP1E^Oc:b0^E[Od:d0_EYOb:g0`EUOc:j0_EROc:n0e0O2O002N1O2L4K5J5L5KYak0"}, "label": "a man standing behind food dishes with his arms out"}]}
{"id": 170882, "image": "COCO_train2014_000000170882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of green veggies\"."}], "task": "refering", "answer_template": "The \"a bowl of green veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 310, 311, 312], "bbox": [0.43759375, 0.6788564476885645, 0.654671875, 0.922287104622871], "iscrowd": 0, "area": 9082.851899999998, "rle": {"size": [411, 640], "counts": "[f`31i<2N101N1O2M4L6K4K5K3N2N2N2O1N1O2O1N2N2O0O2N2N2O0O2N2O1N2N102M3N2N2M3N1O2M3N2N2M2O0O2O000O1000000O2O000000000O2O000000000O2O000000000O10O01O100O1O100O001OUMXFe2h9XM[Fh2l900000000O1000000000000O10000O100O100O1O1N2N2M3M4M2O1O1O1O1O1O1O1O1O1O2N1O2N1O1O2N1O2N1O2N1O2N1O2M2O2M2O2M3N2M3N2M3M5J9Gl\\h2"}, "label": "a bowl of green veggies"}]}
{"id": 170882, "image": "COCO_train2014_000000170882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a greens food on the boul in front of the cook\"."}], "task": "refering", "answer_template": "The \"a greens food on the boul in front of the cook\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 310, 311, 312], "bbox": [0.43759375, 0.6788564476885645, 0.654671875, 0.922287104622871], "iscrowd": 0, "area": 9082.851899999998, "rle": {"size": [411, 640], "counts": "[f`31i<2N101N1O2M4L6K4K5K3N2N2N2O1N1O2O1N2N2O0O2N2N2O0O2N2O1N2N102M3N2N2M3N1O2M3N2N2M2O0O2O000O1000000O2O000000000O2O000000000O2O000000000O10O01O100O1O100O001OUMXFe2h9XM[Fh2l900000000O1000000000000O10000O100O100O1O1N2N2M3M4M2O1O1O1O1O1O1O1O1O1O2N1O2N1O1O2N1O2N1O2N1O2N1O2M2O2M2O2M3N2M3N2M3M5J9Gl\\h2"}, "label": "a greens food on the boul in front of the cook"}]}
{"id": 416644, "image": "COCO_train2014_000000416644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop placed near camera\"."}], "task": "refering", "answer_template": "The \"a laptop placed near camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 249], "bbox": [0.18034375, 0.0, 0.8528125, 0.6153541666666666], "iscrowd": 0, "area": 34298.46165000002, "rle": {"size": [480, 640], "counts": "Ume11o>5K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5K5K5K5K5K6J5K5J6K5K5K5K5K5K5K5K5K2N1O0000000O101O0000000000001O000000000O101O00000000001O01O001O1O00100O001O001O1O010O001O1O00100O001O001O1O2O3L5K4L4L4L5K4M3L5K4L4L4L5L3L4L5K4L4L4L5L3L4L5K4L4L4L5L3L4L5K4L4L4M4K4L4L5K4L4L4M4K4L4L5K4L4L4M4K4L4L1O00010O000000000010O00000001O0001O0001O000001O0001O000000001O01O000001O00000000010O000000001O000001O01O00000000001O01O000001O0000000010O0000000001O000001O01O000000001O0001O000001O0000000010O00000001O00000001O01O000000001O0001O0001O000000000010O00000001O000000010O00000000001O0001O0001O000000001O01O00000001O000000010O000000001O00000001O01O000000001O0001O0001O000000000010O00000001O000000010O00000000001O0001O0001O000000001O01O00000001O000000001O000O10001O00000O10001O000O10001O000O10001O000O1000001O0O1000001O0O1000001O000O1000001O0O1000001O0O100000TX\\1"}, "label": "a laptop placed near camera"}]}
{"id": 416644, "image": "COCO_train2014_000000416644.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop in the wooden stick and the bed\"."}], "task": "refering", "answer_template": "The \"laptop in the wooden stick and the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 249], "bbox": [0.18034375, 0.0, 0.8528125, 0.6153541666666666], "iscrowd": 0, "area": 34298.46165000002, "rle": {"size": [480, 640], "counts": "Ume11o>5K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5K5K5K5K5K6J5K5J6K5K5K5K5K5K5K5K5K2N1O0000000O101O0000000000001O000000000O101O00000000001O01O001O1O00100O001O001O1O010O001O1O00100O001O001O1O2O3L5K4L4L4L5K4M3L5K4L4L4L5L3L4L5K4L4L4L5L3L4L5K4L4L4L5L3L4L5K4L4L4M4K4L4L5K4L4L4M4K4L4L5K4L4L4M4K4L4L1O00010O000000000010O00000001O0001O0001O000001O0001O000000001O01O000001O00000000010O000000001O000001O01O00000000001O01O000001O0000000010O0000000001O000001O01O000000001O0001O000001O0000000010O00000001O00000001O01O000000001O0001O0001O000000000010O00000001O000000010O00000000001O0001O0001O000000001O01O00000001O000000010O000000001O00000001O01O000000001O0001O0001O000000000010O00000001O000000010O00000000001O0001O0001O000000001O01O00000001O000000001O000O10001O00000O10001O000O10001O000O10001O000O1000001O0O1000001O0O1000001O000O1000001O0O1000001O0O100000TX\\1"}, "label": "laptop in the wooden stick and the bed"}]}
{"id": 138117, "image": "COCO_train2014_000000138117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty chair closest to the table\"."}], "task": "refering", "answer_template": "The \"an empty chair closest to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 48, 69, 70, 71, 92, 93, 94, 95, 115, 116], "bbox": [0.0, 0.0002112676056338028, 0.14146875, 0.39842723004694836], "iscrowd": 0, "area": 7005.048000000001, "rle": {"size": [426, 640], "counts": "0m0]<j1VNn0RO2N2N2N2N2O1N2N2N2N2NWOQGYMm8g2VGXMh8h2ZGXMd8h2^GXM`8h2cGWM[8i2gGWMW8i2lGVMR8k2oGUMo7n2PHRMn7Q3RHnLm7U3QHkLm7X3SHgLk7\\3THdLj7_3UHaLi7b3WH]Lg7f3XHYLg7k3XHRLh7Q4WHmKj7U4UHjKk7X4UHeKm7\\4RHbKo7k2oG]N2eNQ8l2RH^NKeNS8k2YH]Nn7b1WH[Ni7b1]HXNh7c1_HVNe7g1`HTNd7g1cHRNa7k1dHoM_7n1gHlM[7R2Q2N2M3N2M3M3N2M2O2M3N2M3M4M2M3N2M4N1O10001O00000O101O0000001O0O10001O00000O2O0000001M2M3M3M3MbXT7"}, "label": "an empty chair closest to the table"}]}
{"id": 138117, "image": "COCO_train2014_000000138117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden chairs in the distance , left side\"."}], "task": "refering", "answer_template": "The \"wooden chairs in the distance , left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 48, 69, 70, 71, 92, 93, 94, 95, 115, 116], "bbox": [0.0, 0.0002112676056338028, 0.14146875, 0.39842723004694836], "iscrowd": 0, "area": 7005.048000000001, "rle": {"size": [426, 640], "counts": "0m0]<j1VNn0RO2N2N2N2N2O1N2N2N2N2NWOQGYMm8g2VGXMh8h2ZGXMd8h2^GXM`8h2cGWM[8i2gGWMW8i2lGVMR8k2oGUMo7n2PHRMn7Q3RHnLm7U3QHkLm7X3SHgLk7\\3THdLj7_3UHaLi7b3WH]Lg7f3XHYLg7k3XHRLh7Q4WHmKj7U4UHjKk7X4UHeKm7\\4RHbKo7k2oG]N2eNQ8l2RH^NKeNS8k2YH]Nn7b1WH[Ni7b1]HXNh7c1_HVNe7g1`HTNd7g1cHRNa7k1dHoM_7n1gHlM[7R2Q2N2M3N2M3M3N2M2O2M3N2M3M4M2M3N2M4N1O10001O00000O101O0000001O0O10001O00000O2O0000001M2M3M3M3MbXT7"}, "label": "wooden chairs in the distance , left side"}]}
{"id": 244616, "image": "COCO_train2014_000000244616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest giraffe\"."}], "task": "refering", "answer_template": "The \"the tallest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 43, 44, 45, 46, 47, 48, 60, 61, 62, 63, 66, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 112, 113, 114, 115, 116, 130, 131, 133, 148, 149, 151, 166, 167, 169], "bbox": [0.22813999999999998, 0.10493333333333334, 0.70004, 0.7545866666666667], "iscrowd": 0, "area": 16792.327900000004, "rle": {"size": [375, 500], "counts": "dPZ11f;00001O0O1G:lM0UH:a7LYH>]7H]Hc0Y7BaHh0U7^OfHg0T7_OiHa0W7CeH=a7@]H`0h7]OUHc0P8[OnGe0W8XOfGh0_8VO^GX1Z8Z1H9F:G2M3N2N2M3N2M3NM4B>C<C>H7N3N2M2O2M2O2M3M2O2M2O2M3N1N3N1N3000O101N101O0O101N101O0O101N101O0O10O010O0100O001O001O001O001O001O1O001O001O0VH^O`4c0lJ1S5Z3O1O1O01000O100O10000O10O0I8E;E:E<E:E<E;D;I8N1O2M3N1O2N1O2N2M2O2N1SOXGkNj8T1^GcNd8\\1dG[N]8e1k0O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O10O0100O0100O0100O010O10O10O10O010O10O10O0100O01000O010O010O10O10O10O01002M2O1O2N101N1O2O0O2N1O2O0O1O2O0O2N1O1N1O1O1O2N1004L7oNQmf1"}, "label": "the tallest giraffe"}]}
{"id": 244616, "image": "COCO_train2014_000000244616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe tilts its neck forward as another behind it appears to crane its own neck to look toward the other ' s shoulders\"."}], "task": "refering", "answer_template": "The \"a giraffe tilts its neck forward as another behind it appears to crane its own neck to look toward the other ' s shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 66, 67, 68, 69, 70, 71, 83, 84, 85, 86, 100, 101, 102, 103, 118, 119, 120, 121, 135, 136, 137, 138, 153, 154, 155, 156, 172, 174, 192], "bbox": [0.5519, 0.19944, 0.9711, 0.7785333333333333], "iscrowd": 0, "area": 12528.112749999998, "rle": {"size": [375, 500], "counts": "cXU3P2e92N2N2NSNfFS1g0iNn75eG;e0Gd7M\\I:b6EZIa0e6_OVIg0i6XORIo0n6oNiH[1W7eNcH\\1b7gNXHW1k7nNoGo0S8@_G>c8W12M4M4M2O2O001N4M2N2O1O1O1N2O1O1O1O1O1N30O2N3M2N3M5K6J6Jj0VO3M3M3N2M3M3M2N3M3M3M2N3N3L7I7Ic0]O1O1O100O1O10000O10000O100000O001gNnJmKR5g3]KULc4h3h1G9H8N1N3M3N1N3N2M3N2M2N3N2M3M3nNiFAY9>mF[OV9d0mFWOV9g0nFTOU9k0nFPOU9o0j00O0100O100O010O100O10O0100O100O10O0100O100O010O10O0100O010O100O010O10O0100O010O10O0100O010O100O010O10O0100O010O010O01O010O0010O010O01O100O100O2O0N2N2N2010O001O00010O010O010O010O2O1N2O2M4M3L100O0O2O1O1O1O0O2O1O1M2O2N2N]\\5"}, "label": "a giraffe tilts its neck forward as another behind it appears to crane its own neck to look toward the other ' s shoulders"}]}
{"id": 244616, "image": "COCO_train2014_000000244616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the right of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe to the right of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 66, 67, 68, 69, 70, 71, 83, 84, 85, 86, 100, 101, 102, 103, 118, 119, 120, 121, 135, 136, 137, 138, 153, 154, 155, 156, 172, 174, 192], "bbox": [0.5519, 0.19944, 0.9711, 0.7785333333333333], "iscrowd": 0, "area": 12528.112749999998, "rle": {"size": [375, 500], "counts": "cXU3P2e92N2N2NSNfFS1g0iNn75eG;e0Gd7M\\I:b6EZIa0e6_OVIg0i6XORIo0n6oNiH[1W7eNcH\\1b7gNXHW1k7nNoGo0S8@_G>c8W12M4M4M2O2O001N4M2N2O1O1O1N2O1O1O1O1O1N30O2N3M2N3M5K6J6Jj0VO3M3M3N2M3M3M2N3M3M3M2N3N3L7I7Ic0]O1O1O100O1O10000O10000O100000O001gNnJmKR5g3]KULc4h3h1G9H8N1N3M3N1N3N2M3N2M2N3N2M3M3nNiFAY9>mF[OV9d0mFWOV9g0nFTOU9k0nFPOU9o0j00O0100O100O010O100O10O0100O100O10O0100O100O010O10O0100O010O100O010O10O0100O010O10O0100O010O100O010O10O0100O010O010O01O010O0010O010O01O100O100O2O0N2N2N2010O001O00010O010O010O010O2O1N2O2M4M3L100O0O2O1O1O1O0O2O1O1M2O2N2N]\\5"}, "label": "a giraffe to the right of another giraffe"}]}
{"id": 170888, "image": "COCO_train2014_000000170888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table with only two chairs near it\"."}], "task": "refering", "answer_template": "The \"the table with only two chairs near it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 161, 162, 163, 164, 165, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188], "bbox": [0.04493333333333334, 0.66532, 0.7820266666666668, 0.8158800000000002], "iscrowd": 0, "area": 12026.174850000005, "rle": {"size": [500, 375], "counts": "od81c?2N2M3N2N2N2N2N2M3N1O1O0000001N1000001O000O2O0000001O0O10001O00001N1000001O00000O2O00001O000O101O0000001N10001O0000001N1000001O000O2O0000001O0O10001O00001O0O10001O00000O2O00001O000O101O0000001O0O101O000000O010000000O10O1000O1000000O010000001N1000O10O1000O1000O0100000O010000O10O1000O10O10O100000O010000O0100000O01000O1000O10O1000O100000001N1000001N100000001N1000001N1000001O0O1000000O2O0000000O1000001N1000000O100000001N1000000O100000001N1000000O1000001O0O1000000O1000001O0O1000000O1000001N100000000O1000001N100000000O10001O0O100000000O4M3M3M2M4M3M3MieW1"}, "label": "the table with only two chairs near it"}]}
{"id": 170888, "image": "COCO_train2014_000000170888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table flanked by two chairs\"."}], "task": "refering", "answer_template": "The \"a table flanked by two chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 161, 162, 163, 164, 165, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188], "bbox": [0.04493333333333334, 0.66532, 0.7820266666666668, 0.8158800000000002], "iscrowd": 0, "area": 12026.174850000005, "rle": {"size": [500, 375], "counts": "od81c?2N2M3N2N2N2N2N2M3N1O1O0000001N1000001O000O2O0000001O0O10001O00001N1000001O00000O2O00001O000O101O0000001N10001O0000001N1000001O000O2O0000001O0O10001O00001O0O10001O00000O2O00001O000O101O0000001O0O101O000000O010000000O10O1000O1000000O010000001N1000O10O1000O1000O0100000O010000O10O1000O10O10O100000O010000O0100000O01000O1000O10O1000O100000001N1000001N100000001N1000001N1000001O0O1000000O2O0000000O1000001N1000000O100000001N1000000O100000001N1000000O1000001O0O1000000O1000001O0O1000000O1000001N100000000O1000001N100000000O10001O0O100000000O4M3M3M2M4M3M3MieW1"}, "label": "a table flanked by two chairs"}]}
{"id": 261003, "image": "COCO_train2014_000000261003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white station wagon driving past a truck being unloaded\"."}], "task": "refering", "answer_template": "The \"a white station wagon driving past a truck being unloaded\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.048390625, 0.5785, 0.8741875, 0.9892500000000001], "iscrowd": 0, "area": 72529.2156, "rle": {"size": [480, 640], "counts": "io>1o>0O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000001O00000000000000000000000000000000000000000001O0000000000000000000000001O1O001O1O1O1O1O1O001O1O1O1N2O1O001O1O1O1O1O1O0O2O1O1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O1N2O2N3M2N2N2N2L4J6J6J6J6J7I6J6J6H8C=B>C=C=BRdU1"}, "label": "a white station wagon driving past a truck being unloaded"}]}
{"id": 261003, "image": "COCO_train2014_000000261003.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a station wagon in front of the delivery truck\"."}], "task": "refering", "answer_template": "The \"a station wagon in front of the delivery truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.048390625, 0.5785, 0.8741875, 0.9892500000000001], "iscrowd": 0, "area": 72529.2156, "rle": {"size": [480, 640], "counts": "io>1o>0O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O10000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000001O00000000000000000000000000000000000000000001O0000000000000000000000001O1O001O1O1O1O1O1O001O1O1O1N2O1O001O1O1O1O1O1O0O2O1O1O1O1O1O001O1O1N2O1O1O001O1O1O1O1O1N2O2N3M2N2N2N2L4J6J6J6J6J7I6J6J6H8C=B>C=C=BRdU1"}, "label": "a station wagon in front of the delivery truck"}]}
{"id": 80782, "image": "COCO_train2014_000000080782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallar giraff behind the big one\"."}], "task": "refering", "answer_template": "The \"the smallar giraff behind the big one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 264, 265, 287, 288, 289, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337], "bbox": [0.21440625, 0.38201405152224827, 0.703171875, 1.0], "iscrowd": 0, "area": 12513.463449999997, "rle": {"size": [427, 640], "counts": "Z[i11W=4M3L3M4M3N2N1O1O2N100O2N1O1O2N0000000000000000001O1O1O100O1O001O1O1O1O100O100O10O0100FfNcDZ1];:O1O1O1O1O10O01O1O100O1O1O100O`NfDQ1Y;oNiDo0Y;nNiDR1W;mNjDT1V;jNkDW1U;fNmD[1S;cNnD^1Q;`NQE`1];1O1O2O1N2N2N2N2N2N2N2N1OO2N2O2M2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N2N3M2O1N2N3M2O1Ning02QQXO7H6RMEgH`0Z7HWH>i7KiG:X8N\\G5e8LXG5h8MUG4l8LRG5o8KnF7S9JjF7V9JhF7Y9JcF8^9H`F9`9H^F9c9HYF:h9FVF;j9FTF;m9FPF;Q:ElE=T:EiE<X:[11O1O001O1O001O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O001O00001O00001O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1OPj^2"}, "label": "the smallar giraff behind the big one"}]}
{"id": 80782, "image": "COCO_train2014_000000080782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smaller giraffe behind a larger giraffe\"."}], "task": "refering", "answer_template": "The \"a smaller giraffe behind a larger giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 264, 265, 287, 288, 289, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337], "bbox": [0.21440625, 0.38201405152224827, 0.703171875, 1.0], "iscrowd": 0, "area": 12513.463449999997, "rle": {"size": [427, 640], "counts": "Z[i11W=4M3L3M4M3N2N1O1O2N100O2N1O1O2N0000000000000000001O1O1O100O1O001O1O1O1O100O100O10O0100FfNcDZ1];:O1O1O1O1O10O01O1O100O1O1O100O`NfDQ1Y;oNiDo0Y;nNiDR1W;mNjDT1V;jNkDW1U;fNmD[1S;cNnD^1Q;`NQE`1];1O1O2O1N2N2N2N2N2N2N2N1OO2N2O2M2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N2N3M2O1N2N3M2O1Ning02QQXO7H6RMEgH`0Z7HWH>i7KiG:X8N\\G5e8LXG5h8MUG4l8LRG5o8KnF7S9JjF7V9JhF7Y9JcF8^9H`F9`9H^F9c9HYF:h9FVF;j9FTF;m9FPF;Q:ElE=T:EiE<X:[11O1O001O1O001O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O001O00001O00001O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1OPj^2"}, "label": "a smaller giraffe behind a larger giraffe"}]}
{"id": 80782, "image": "COCO_train2014_000000080782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two giraffes and the ostrich\"."}], "task": "refering", "answer_template": "The \"the two giraffes and the ostrich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 60, 61, 62, 63, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 259, 260, 261, 262, 263, 264, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.08395312499999999, 0.0853864168618267, 0.799125, 0.9887587822014051], "iscrowd": 0, "area": 51790.543399999995, "rle": {"size": [427, 640], "counts": "olf0e0a<8H5K5K5L4K5K5O1N2O1N2N2O1N2O1N2O1N2O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1000000O10000OO2[Oe0[Od0\\Oe0ZOf0A?N1O2N2O0O2O10O10O10O1000000O010gE=j6CaHS1]7nNPHf1o7ZNoGh1Q8YNkGj1T8WNjGk1V8UNhGm1X8TNeGn1[8RNcGP2\\8QNdGo1\\8QNdGo1\\8RNcGn1\\8SNcGn1]8RNcGn1]8SNbGm1]8TNcGl1]8TNcGl1\\8VNbGk1]8WNbGi1^8YN`Gg1_8\\N_Gd1`8^N_Gb1a8_N]Gb1b8`N]G`1b8bN]G^1c8cN\\G]1c8eN\\G[1c8gN\\GY1d8hN[GX1d8jN[GV1d8kN\\GU1d8lN[GT1d8oNZGQ1e8QOXGQ1h8PORGU1m8e1O100O1O100O1O100O1O1O1N2N2O1N2N2O1N2N2O1N2N2O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OM3I7XMPJPNV6o1WJ^MP6a2\\JRMf5m2`JnL`5Q3cJmL]5R3fJkL[5T3iJiLW5V3lJhLT5W3oJgLQ5X3RKfLn4Y3VKdLj4[3YKcLg4\\3\\KbLd4]3_KaLa4_3bK]L_4b3dK\\L\\4d3fKZLZ4e3iKYLW4g3kKWLU4h3oKULQ4k3g2O2O0O100O100O100O100O100O100O1O100O10O0100O100O100O100O010O100O100O100O10O0100O100O100O100O010O1O100O100O100O010O100O100O100O10O0100O100O100O100O010O100O100O1O100O010O100O101N2O1N]O`M_F_2`9dM_FZ2b9iM[FW2e9lMYFR2h9QNUFo1j9UNQFl1P:WNlEj1U:d03N2M3N2M3N5J7J7H_OQNiEo1X:TNcEl1_:XN[Ej1e:>0000001O00000000002N2N3M2N3M1O1O1O1O1O1O1O1O10O01O1O1O001O001O0\\N^D_1c;^NaD`1f;O1O1O2N1O1O1O2N1O1O1O2N1O1O100O10O010000O100O001O1O1O1N101O1N2O1O0O4M3M3Imee1"}, "label": "the two giraffes and the ostrich"}]}
{"id": 80782, "image": "COCO_train2014_000000080782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two giraffes that are crossing paths\"."}], "task": "refering", "answer_template": "The \"the two giraffes that are crossing paths\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 60, 61, 62, 63, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 259, 260, 261, 262, 263, 264, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.08395312499999999, 0.0853864168618267, 0.799125, 0.9887587822014051], "iscrowd": 0, "area": 51790.543399999995, "rle": {"size": [427, 640], "counts": "olf0e0a<8H5K5K5L4K5K5O1N2O1N2N2O1N2O1N2O1N2O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1000000O10000OO2[Oe0[Od0\\Oe0ZOf0A?N1O2N2O0O2O10O10O10O1000000O010gE=j6CaHS1]7nNPHf1o7ZNoGh1Q8YNkGj1T8WNjGk1V8UNhGm1X8TNeGn1[8RNcGP2\\8QNdGo1\\8QNdGo1\\8RNcGn1\\8SNcGn1]8RNcGn1]8SNbGm1]8TNcGl1]8TNcGl1\\8VNbGk1]8WNbGi1^8YN`Gg1_8\\N_Gd1`8^N_Gb1a8_N]Gb1b8`N]G`1b8bN]G^1c8cN\\G]1c8eN\\G[1c8gN\\GY1d8hN[GX1d8jN[GV1d8kN\\GU1d8lN[GT1d8oNZGQ1e8QOXGQ1h8PORGU1m8e1O100O1O100O1O100O1O1O1N2N2O1N2N2O1N2N2O1N2N2O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OM3I7XMPJPNV6o1WJ^MP6a2\\JRMf5m2`JnL`5Q3cJmL]5R3fJkL[5T3iJiLW5V3lJhLT5W3oJgLQ5X3RKfLn4Y3VKdLj4[3YKcLg4\\3\\KbLd4]3_KaLa4_3bK]L_4b3dK\\L\\4d3fKZLZ4e3iKYLW4g3kKWLU4h3oKULQ4k3g2O2O0O100O100O100O100O100O100O1O100O10O0100O100O100O100O010O100O100O100O10O0100O100O100O100O010O1O100O100O100O010O100O100O100O10O0100O100O100O100O010O100O100O1O100O010O100O101N2O1N]O`M_F_2`9dM_FZ2b9iM[FW2e9lMYFR2h9QNUFo1j9UNQFl1P:WNlEj1U:d03N2M3N2M3N5J7J7H_OQNiEo1X:TNcEl1_:XN[Ej1e:>0000001O00000000002N2N3M2N3M1O1O1O1O1O1O1O1O10O01O1O1O001O001O0\\N^D_1c;^NaD`1f;O1O1O2N1O1O1O2N1O1O1O2N1O1O100O10O010000O100O001O1O1O1N101O1N2O1O0O4M3M3Imee1"}, "label": "the two giraffes that are crossing paths"}]}
{"id": 269199, "image": "COCO_train2014_000000269199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant standing in water to the right of another elephant\"."}], "task": "refering", "answer_template": "The \"an elephant standing in water to the right of another elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 214, 215, 216, 219, 220, 221, 237, 238, 239, 242, 243, 244, 260, 261, 265, 266, 267, 283, 288, 289, 290], "bbox": [0.316359375, 0.3483138173302108, 0.812625, 0.8359484777517564], "iscrowd": 0, "area": 26081.977249999996, "rle": {"size": [427, 640], "counts": "bad21Y=9G9G:F8H0O2O0000001O0^NjNPGV1Q9nNiFT1W9POdFP1]9TO^Fl0c9XOXFi0h9[OSFe0n9@lE`0U:DfE=Z:GaE9`:X1100O1O100O1O100O1O1O1O1OoLfMSKY2m4hMSKW2n4iMRKV2o4kMPKS2Q5nMoJQ2R5PNmJo1T5QNlJn1U5RNkJm1U5UNjJj1W5VNiJi1X5WNhJh1Y5YNfJf1[5ZNeJe1\\5[NdJc1^5^NaJa1`5_N`J`1a5aN^J^1c5bN]J]1d5cN\\J\\1e5eN[JY1f5gNZJX1g5hNYJW1h5jNWJU1j5kNWJR1k5nNUJQ1l5POSJo0n5QOSJm0n5TOQJk0P6UOPJj0Q6VOPJh0Q6YOnIf0S6ZOmIe0T6[OlId0U6]OkIa0V6_OfJC\\5>h31O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N100O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1N2O1O1O2N1O1O1O1O1N3N1O2Na0_Oa0_Oa0_Oa0_O;E1N10001O000O101O00001O0O10001O000O2O0000J6K4L5K5M31O1N3N1O2N1O1N3N1OkLeH<Y7CmH9R7GSI6j6K[I1d6NbIN]62hIKV65oIGP68VJCj5=[J]Of5b0gJROY5n0SKfNm4Z1_KYNb4f1kKnMU4R2VLcMj3]2b3O10000O100O10000O1000000001O000000000000000000000000000000000000000000000001O000000000O2O0O100O100O2O0O100O100O2O000O100O101N1000000000001O000000001O001O1O001O1O001O1O00100O001O001O1O001O1O001O1O001O1O001O2N1O1O1O1O2N100O1N2N2M4L3M3N9FSka1"}, "label": "an elephant standing in water to the right of another elephant"}]}
{"id": 269199, "image": "COCO_train2014_000000269199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with it ' s back legs slightly bent\"."}], "task": "refering", "answer_template": "The \"an elephant with it ' s back legs slightly bent\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 214, 215, 216, 219, 220, 221, 237, 238, 239, 242, 243, 244, 260, 261, 265, 266, 267, 283, 288, 289, 290], "bbox": [0.316359375, 0.3483138173302108, 0.812625, 0.8359484777517564], "iscrowd": 0, "area": 26081.977249999996, "rle": {"size": [427, 640], "counts": "bad21Y=9G9G:F8H0O2O0000001O0^NjNPGV1Q9nNiFT1W9POdFP1]9TO^Fl0c9XOXFi0h9[OSFe0n9@lE`0U:DfE=Z:GaE9`:X1100O1O100O1O100O1O1O1O1OoLfMSKY2m4hMSKW2n4iMRKV2o4kMPKS2Q5nMoJQ2R5PNmJo1T5QNlJn1U5RNkJm1U5UNjJj1W5VNiJi1X5WNhJh1Y5YNfJf1[5ZNeJe1\\5[NdJc1^5^NaJa1`5_N`J`1a5aN^J^1c5bN]J]1d5cN\\J\\1e5eN[JY1f5gNZJX1g5hNYJW1h5jNWJU1j5kNWJR1k5nNUJQ1l5POSJo0n5QOSJm0n5TOQJk0P6UOPJj0Q6VOPJh0Q6YOnIf0S6ZOmIe0T6[OlId0U6]OkIa0V6_OfJC\\5>h31O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N100O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1N2O1O1O1O2N1O1O1N2O1O1O2N1O1O1O1O1N3N1O2Na0_Oa0_Oa0_Oa0_O;E1N10001O000O101O00001O0O10001O000O2O0000J6K4L5K5M31O1N3N1O2N1O1N3N1OkLeH<Y7CmH9R7GSI6j6K[I1d6NbIN]62hIKV65oIGP68VJCj5=[J]Of5b0gJROY5n0SKfNm4Z1_KYNb4f1kKnMU4R2VLcMj3]2b3O10000O100O10000O1000000001O000000000000000000000000000000000000000000000001O000000000O2O0O100O100O2O0O100O100O2O000O100O101N1000000000001O000000001O001O1O001O1O001O1O00100O001O001O1O001O1O001O1O001O1O001O2N1O1O1O1O2N100O1N2N2M4L3M3N9FSka1"}, "label": "an elephant with it ' s back legs slightly bent"}]}
{"id": 72592, "image": "COCO_train2014_000000072592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the black shirt\"."}], "task": "refering", "answer_template": "The \"the girl in the black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 258, 259, 260, 281, 282, 283, 284, 304, 305, 306, 307, 327, 329, 330], "bbox": [0.23479687500000002, 0.3524791666666667, 0.37331250000000005, 0.8434583333333333], "iscrowd": 0, "area": 11401.041599999999, "rle": {"size": [480, 640], "counts": "oeV26_>b0^Ob0E4L4L4nLJVG:f8MSG7j8OoF5n8P3M4K4M3L4M2M4M2O2O1OfNSHbKm7^4WH_Kh7`4]1O0N3L4M3L3M4K5J5L2N23M4H8K5L4L4L4M3N2N2N2O1N2N2O2M3M3M3N1O1O2O0O1O1O2N7IeKhFU3S9iLnF^4k7`KWH`4j7\\KYHe4f7YK[Hh4f7UK\\Hk4d7PKaHo4a7iJfHV5b801O1mLgF=Z9@mF;T9ASG;n8AZG:g8B`G:`8BhG:Y8BmGVO^OXOn8S1hIlNe6`0bI@o:01O001O0000001N10002N2N^bk5"}, "label": "the girl in the black shirt"}]}
{"id": 72592, "image": "COCO_train2014_000000072592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rear view of a woman in black shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"rear view of a woman in black shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 258, 259, 260, 281, 282, 283, 284, 304, 305, 306, 307, 327, 329, 330], "bbox": [0.23479687500000002, 0.3524791666666667, 0.37331250000000005, 0.8434583333333333], "iscrowd": 0, "area": 11401.041599999999, "rle": {"size": [480, 640], "counts": "oeV26_>b0^Ob0E4L4L4nLJVG:f8MSG7j8OoF5n8P3M4K4M3L4M2M4M2O2O1OfNSHbKm7^4WH_Kh7`4]1O0N3L4M3L3M4K5J5L2N23M4H8K5L4L4L4M3N2N2N2O1N2N2O2M3M3M3N1O1O2O0O1O1O2N7IeKhFU3S9iLnF^4k7`KWH`4j7\\KYHe4f7YK[Hh4f7UK\\Hk4d7PKaHo4a7iJfHV5b801O1mLgF=Z9@mF;T9ASG;n8AZG:g8B`G:`8BhG:Y8BmGVO^OXOn8S1hIlNe6`0bI@o:01O001O0000001N10002N2N^bk5"}, "label": "rear view of a woman in black shirt and blue jeans"}]}
{"id": 72592, "image": "COCO_train2014_000000072592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 127, 128, 129, 130, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 309, 310, 311, 332, 333, 334, 356], "bbox": [0.40954687500000003, 0.2763958333333333, 0.67415625, 0.8898958333333333], "iscrowd": 0, "area": 20016.497, "rle": {"size": [480, 640], "counts": "jRk31m>3N2M3M3N2M3M3M3N1N2O1O1O1N2O1O1nD]Oe7e0XHG^7:^H2Y7OdH5Y7MdH6[7KbH8]7ZOYFBV2Y1_7VO]F_OR2]1`7TOaF[On1d1`7QOeFWOj1l1_7mNlHV1S7jNkHY1T7hNiH[1V7eNhH_1U7bNgHc1X7]NRHZ2l7gMQH^2l7dMPH`2n7`MPHd2m7^MPHf2o7ZMoGj2o7VMlGP3R8QMhGV3W8iLfG]3X8cLhG^3W8bLhG`3W8`LhGc3V8]LjGc3U8^LjG`3Y8_LgG`3[8`LeGg2T9YMkFf2W9ZMhFf2Y9[MfFc2>_Ln7n0dGa2=gLm7g0fGa2<mLk7b0iG_2<TMh7=lG]2<[Me78oG[2=`Mb75QHZ2<fM`71SHW2>lM]7LUHV2;UN^7EWH[4h7eKXH\\4g7dKXH^4g7bKYH_4g7`KXHb4g7`KVHb4i7`KTHb4k7`KRHb4n7X1O1O10O01O0010O01O10O01O010OO1O2N1N2OL3E<N12O1N2N1N3M3M3M3N2M3M3M3J6J6K5M4[KPGV3S9jLoFR3S9nLnFn2U9RMmFj2V9TMlFh2W9XMjFe2Y9ZMiFa2\\9]MeF`2_9^MbF^2c9`M^F]2g9_M[F^2i9_MYF]2l9aMUF\\2o9aMSF1^Oa1d:\\NoE2B[1c:`NlE6DS1e:dNiE8Fn0f:gNeE:Ob0a:QObE<:4X:]O`E?d0DQ:K\\E`0_=O010O1000O0100O010O1000O0100O10O001O10O01_NE\\D;b;LZD4e;1UD0l;<fCFZ<>`CCa<b0XC@h<\\10O100O2O2M3N3M3L5L7H;F6I6K4L4K5L2MSiQ3"}, "label": "man"}]}
{"id": 72592, "image": "COCO_train2014_000000072592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a sleeveless yellow shirt who is about to fly a kite\"."}], "task": "refering", "answer_template": "The \"a man in a sleeveless yellow shirt who is about to fly a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 127, 128, 129, 130, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 263, 264, 265, 266, 286, 287, 288, 309, 310, 311, 332, 333, 334, 356], "bbox": [0.40954687500000003, 0.2763958333333333, 0.67415625, 0.8898958333333333], "iscrowd": 0, "area": 20016.497, "rle": {"size": [480, 640], "counts": "jRk31m>3N2M3M3N2M3M3M3N1N2O1O1O1N2O1O1nD]Oe7e0XHG^7:^H2Y7OdH5Y7MdH6[7KbH8]7ZOYFBV2Y1_7VO]F_OR2]1`7TOaF[On1d1`7QOeFWOj1l1_7mNlHV1S7jNkHY1T7hNiH[1V7eNhH_1U7bNgHc1X7]NRHZ2l7gMQH^2l7dMPH`2n7`MPHd2m7^MPHf2o7ZMoGj2o7VMlGP3R8QMhGV3W8iLfG]3X8cLhG^3W8bLhG`3W8`LhGc3V8]LjGc3U8^LjG`3Y8_LgG`3[8`LeGg2T9YMkFf2W9ZMhFf2Y9[MfFc2>_Ln7n0dGa2=gLm7g0fGa2<mLk7b0iG_2<TMh7=lG]2<[Me78oG[2=`Mb75QHZ2<fM`71SHW2>lM]7LUHV2;UN^7EWH[4h7eKXH\\4g7dKXH^4g7bKYH_4g7`KXHb4g7`KVHb4i7`KTHb4k7`KRHb4n7X1O1O10O01O0010O01O10O01O010OO1O2N1N2OL3E<N12O1N2N1N3M3M3M3N2M3M3M3J6J6K5M4[KPGV3S9jLoFR3S9nLnFn2U9RMmFj2V9TMlFh2W9XMjFe2Y9ZMiFa2\\9]MeF`2_9^MbF^2c9`M^F]2g9_M[F^2i9_MYF]2l9aMUF\\2o9aMSF1^Oa1d:\\NoE2B[1c:`NlE6DS1e:dNiE8Fn0f:gNeE:Ob0a:QObE<:4X:]O`E?d0DQ:K\\E`0_=O010O1000O0100O010O1000O0100O10O001O10O01_NE\\D;b;LZD4e;1UD0l;<fCFZ<>`CCa<b0XC@h<\\10O100O2O2M3N3M3L5L7H;F6I6K4L4K5L2MSiQ3"}, "label": "a man in a sleeveless yellow shirt who is about to fly a kite"}]}
{"id": 301970, "image": "COCO_train2014_000000301970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra furthest to the left\"."}], "task": "refering", "answer_template": "The \"the zebra furthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 130, 131, 132, 133, 134, 148, 149, 152, 153, 167, 170, 171, 185, 189], "bbox": [0.10956, 0.22978666666666667, 0.52752, 0.8117866666666667], "iscrowd": 0, "area": 19858.25569999999, "rle": {"size": [375, 500], "counts": "\\Wd0:^;4K4hE@Q9d0YF2d9T1L4L3N3K5K4M4K3N0O1O101N100O1O20N1O2M3N2M2O2N2M2O11O0O2O001O1O002N1O1N2002N2N2O0O2N1O00O100O101N1O100O3N2M3N1N3N2M3M3N2M3N2M3N2M2O2M2O1N2N2O1N3N1N2O1N2O1N2N3dIfKV5[4dJiK\\5Y4^JkKc5V4XJmKi5T4QJPLP6Q4jITLW6e41O1O2N1O0000000100O001OaMjIZOY6g0VJaNR6_1^JjMi5V2l100001O0000000010O000001O000000001O0000001O000001O01O000000O010O10000O10O10O10000O01000O100O10O10O10000O10O10O10000O10O010000O11O2M2O1O2N1O1O2M2O11O101N2N101N1O3N2M2N3M2O24L5J7K4NMj0TO4K5L4L2M3N1O1M3L5J5K5L5J5K?BkTf2"}, "label": "the zebra furthest to the left"}]}
{"id": 301970, "image": "COCO_train2014_000000301970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large zebra in the front\"."}], "task": "refering", "answer_template": "The \"large zebra in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 130, 131, 132, 133, 134, 148, 149, 152, 153, 167, 170, 171, 185, 189], "bbox": [0.10956, 0.22978666666666667, 0.52752, 0.8117866666666667], "iscrowd": 0, "area": 19858.25569999999, "rle": {"size": [375, 500], "counts": "\\Wd0:^;4K4hE@Q9d0YF2d9T1L4L3N3K5K4M4K3N0O1O101N100O1O20N1O2M3N2M2O2N2M2O11O0O2O001O1O002N1O1N2002N2N2O0O2N1O00O100O101N1O100O3N2M3N1N3N2M3M3N2M3N2M3N2M2O2M2O1N2N2O1N3N1N2O1N2O1N2N3dIfKV5[4dJiK\\5Y4^JkKc5V4XJmKi5T4QJPLP6Q4jITLW6e41O1O2N1O0000000100O001OaMjIZOY6g0VJaNR6_1^JjMi5V2l100001O0000000010O000001O000000001O0000001O000001O01O000000O010O10000O10O10O10000O01000O100O10O10O10000O10O10O10000O10O010000O11O2M2O1O2N1O1O2M2O11O101N2N101N1O3N2M2N3M2O24L5J7K4NMj0TO4K5L4L2M3N1O1M3L5J5K5L5J5K?BkTf2"}, "label": "large zebra in the front"}]}
{"id": 301970, "image": "COCO_train2014_000000301970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small zebra in the front of the pack of zebras looking at the camera\"."}], "task": "refering", "answer_template": "The \"a small zebra in the front of the pack of zebras looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 80, 81, 82, 98, 99, 100, 101, 116, 117, 118, 119, 135, 136, 137, 153, 154, 155, 171, 172, 173, 189, 190], "bbox": [0.46516, 0.28314666666666666, 0.6623600000000001, 0.8157333333333333], "iscrowd": 0, "area": 10365.697200000002, "rle": {"size": [375, 500], "counts": "Q^e25`;5L5J6K6Ih0\\O0O2N1O2O0O2[FXNV9T2100O10O0eFoMP9R2nFUNl8k1TGZN`0Ba7T2nG_N7Eh7l1PHeNNGn7f1SHYOi7h0WH\\Oc7g0\\H\\O`7`2M3N2N2N3M3M3M3MTLUIP3h6aLYIF6k3^6ZLPJh3m5RLZJl3j6N2N2OcH\\Lh6b3RIfLn6X3lHPMS7k3N2N2O1N2N3NO000nNXIVMh6_2nIWMR6]2cJYM]5[2ZKZMf4f2V21N1O2O0O101fHWMd5i2^IWM17`6c2VIeMMNn6\\2mHgNR7o200010M2N2N2N3L3M3M3M4jNeHkM^7Q2V1MWGgMZ8U3G8G9G10M3M3M3M2N3M3L4M4L4K4F;^Nb1N1N3L3M4LYgm1"}, "label": "a small zebra in the front of the pack of zebras looking at the camera"}]}
{"id": 301970, "image": "COCO_train2014_000000301970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small zebra facing the camera\"."}], "task": "refering", "answer_template": "The \"a small zebra facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 80, 81, 82, 98, 99, 100, 101, 116, 117, 118, 119, 135, 136, 137, 153, 154, 155, 171, 172, 173, 189, 190], "bbox": [0.46516, 0.28314666666666666, 0.6623600000000001, 0.8157333333333333], "iscrowd": 0, "area": 10365.697200000002, "rle": {"size": [375, 500], "counts": "Q^e25`;5L5J6K6Ih0\\O0O2N1O2O0O2[FXNV9T2100O10O0eFoMP9R2nFUNl8k1TGZN`0Ba7T2nG_N7Eh7l1PHeNNGn7f1SHYOi7h0WH\\Oc7g0\\H\\O`7`2M3N2N2N3M3M3M3MTLUIP3h6aLYIF6k3^6ZLPJh3m5RLZJl3j6N2N2OcH\\Lh6b3RIfLn6X3lHPMS7k3N2N2O1N2N3NO000nNXIVMh6_2nIWMR6]2cJYM]5[2ZKZMf4f2V21N1O2O0O101fHWMd5i2^IWM17`6c2VIeMMNn6\\2mHgNR7o200010M2N2N2N3L3M3M3M4jNeHkM^7Q2V1MWGgMZ8U3G8G9G10M3M3M3M2N3M3L4M4L4K4F;^Nb1N1N3L3M4LYgm1"}, "label": "a small zebra facing the camera"}]}
{"id": 301970, "image": "COCO_train2014_000000301970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra facing away\"."}], "task": "refering", "answer_template": "The \"a zebra facing away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 63, 64, 81, 82, 83, 84, 85, 86, 99, 101, 102, 103, 104, 119, 120, 121, 122, 137, 138, 139, 140, 157, 158, 175, 176], "bbox": [0.5160399999999999, 0.2081066666666667, 0.80952, 0.7461866666666667], "iscrowd": 0, "area": 9860.229399999998, "rle": {"size": [375, 500], "counts": "obn21e;1N2@0TE5f:a0K5O100O1VFTOh8m0RGZOl8g0nF_OR9a0hFFW9:cFM\\9W101O1N1000O109F001O0010O01O1O001O010O0dNVFm0j9QO[Fl0d9SO_Fl0a9RObFm0]9ROfFm0Z9ROiFl0V9SOnFk0P:0?@4L00010O00010O0010O00010O00001O00001O1O1O1O1O1O2N1O1O1O1O1XFUOb8l0ZG_O^8e0\\GD]8a0[GAe8e0SG^Om8g0kF\\OU9c100000O100000O1000O100000000000O01000000O1M2O2N2N2O13L4M2N3M3M2M4M3M2N3M3N1N3M3M2O2M4L4M3L4L4M3L4L4M3L4LO2O1O1N2O1N1I8A?H8O1N2O1O1O3M3L3iMfGm0]8kNTHi0o7oNbH?g7XOkH4fkR1"}, "label": "a zebra facing away"}]}
{"id": 301970, "image": "COCO_train2014_000000301970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hind end of the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the hind end of the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 63, 64, 81, 82, 83, 84, 85, 86, 99, 101, 102, 103, 104, 119, 120, 121, 122, 137, 138, 139, 140, 157, 158, 175, 176], "bbox": [0.5160399999999999, 0.2081066666666667, 0.80952, 0.7461866666666667], "iscrowd": 0, "area": 9860.229399999998, "rle": {"size": [375, 500], "counts": "obn21e;1N2@0TE5f:a0K5O100O1VFTOh8m0RGZOl8g0nF_OR9a0hFFW9:cFM\\9W101O1N1000O109F001O0010O01O1O001O010O0dNVFm0j9QO[Fl0d9SO_Fl0a9RObFm0]9ROfFm0Z9ROiFl0V9SOnFk0P:0?@4L00010O00010O0010O00010O00001O00001O1O1O1O1O1O2N1O1O1O1O1XFUOb8l0ZG_O^8e0\\GD]8a0[GAe8e0SG^Om8g0kF\\OU9c100000O100000O1000O100000000000O01000000O1M2O2N2N2O13L4M2N3M3M2M4M3M2N3M3N1N3M3M2O2M4L4M3L4L4M3L4L4M3L4LO2O1O1N2O1N1I8A?H8O1N2O1O1O3M3L3iMfGm0]8kNTHi0o7oNbH?g7XOkH4fkR1"}, "label": "the hind end of the zebra on the right"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top right section of broccoli\"."}], "task": "refering", "answer_template": "The \"the top right section of broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 101, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150], "bbox": [0.40931249999999997, 0.23370023419203748, 0.6626875, 0.4516861826697892], "iscrowd": 0, "area": 9372.2756, "rle": {"size": [427, 640], "counts": "d]]36Q=5L3L5L3L5L3H9E:E<L3O2N1O2M2O2N1O1O10O010000O100O100O001O1O1O100O1O1O1O010O1O1O1O100000000000O1000000000001O0000000O10000001O00001O010O001O00001O001O00O10001N10000O10000O10000O100O2O0O2N101N2^OUEaNl:\\1[E`Nf:\\1`E`Nb:^1bE_N_:_1g001O1N101N2O001N2O1N101O1N101N2O001N2O0010O10000O10O001O1O001O001N101O1O001O001O1N110O01000O01001N2O1O1N2O1O1N2O1O1N2O2N1NQmi2"}, "label": "the top right section of broccoli"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl in the right\"."}], "task": "refering", "answer_template": "The \"the bowl in the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.797265625, 0.3713817330210773, 1.0, 0.7872365339578454], "iscrowd": 0, "area": 19161.19745, "rle": {"size": [427, 640], "counts": "\\ld65T=d0\\Oc0]O4L5K4L5K5K4L5K4L5K5K4L5K4L3N2N2N2N2N1O2N2N2N2N1N3N2N2N2N2N1O2N2N2N2N101N2O1O1N2O0O2O1N2O1O0O2O1N2O0O10000O2O0O1000000O2O000000000O2O000000000O2O000000000O2O00000000001O000001O01O00000001O000000001N10000O101N10000O10001N10000O101O0O2O1N101O1N2O1O1N2O1O1NbJ"}, "label": "the bowl in the right"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small silver bowl on the right\"."}], "task": "refering", "answer_template": "The \"small silver bowl on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.797265625, 0.3713817330210773, 1.0, 0.7872365339578454], "iscrowd": 0, "area": 19161.19745, "rle": {"size": [427, 640], "counts": "\\ld65T=d0\\Oc0]O4L5K4L5K5K4L5K4L5K5K4L5K4L3N2N2N2N2N1O2N2N2N2N1N3N2N2N2N2N1O2N2N2N2N101N2O1O1N2O0O2O1N2O1O0O2O1N2O0O10000O2O0O1000000O2O000000000O2O000000000O2O000000000O2O00000000001O000001O01O00000001O000000001N10000O101N10000O10001N10000O101O0O2O1N101O1N2O1O1N2O1O1NbJ"}, "label": "small silver bowl on the right"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli at 4 oclock on the plate\"."}], "task": "refering", "answer_template": "The \"broccoli at 4 oclock on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290], "bbox": [0.533875, 0.48000000000000004, 0.68309375, 0.8295550351288057], "iscrowd": 0, "area": 8454.26055, "rle": {"size": [427, 640], "counts": "Rl^42W=3N1M3M4M2N2N3M2K5M4I7lNnN\\E1KT1i:APEd0P;m0O000O2O00001N0100O10O00001O01O11]FgMQ8Y2lGkMS8U2iGoMW8Q2fGSNV8P2gGTNk7Y2SHjMk7W2SHjMm7X2PHiMP8X2nGjMQ8W2mGjMT8V2jGkMV8V2hGkMX8V2eGlM[8U2bGlM_8V2]GlMc8U2ZGlMg8\\30O100O2O0O1O2M2O1N3N1L4O4M0O100O1O100O1O1N2O1O3N2M3M3M5L3K2N2O0O2N2O0O2\\O^E]Nd:b1_EYNd:g1^ESNg:l1=N3N2M3N1N3K5kNZD?j;VObDf0T<Kf\\d2"}, "label": "broccoli at 4 oclock on the plate"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a broccoli in display\"."}], "task": "refering", "answer_template": "The \"a broccoli in display\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244, 245, 266, 267, 268, 289, 290], "bbox": [0.533875, 0.48000000000000004, 0.68309375, 0.8295550351288057], "iscrowd": 0, "area": 8454.26055, "rle": {"size": [427, 640], "counts": "Rl^42W=3N1M3M4M2N2N3M2K5M4I7lNnN\\E1KT1i:APEd0P;m0O000O2O00001N0100O10O00001O01O11]FgMQ8Y2lGkMS8U2iGoMW8Q2fGSNV8P2gGTNk7Y2SHjMk7W2SHjMm7X2PHiMP8X2nGjMQ8W2mGjMT8V2jGkMV8V2hGkMX8V2eGlM[8U2bGlM_8V2]GlMc8U2ZGlMg8\\30O100O2O0O1O2M2O1N3N1L4O4M0O100O1O100O1O1N2O1O3N2M3M3M5L3K2N2O0O2N2O0O2\\O^E]Nd:b1_EYNd:g1^ESNg:l1=N3N2M3N1N3K5kNZD?j;VObDf0T<Kf\\d2"}, "label": "a broccoli in display"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrots at top\"."}], "task": "refering", "answer_template": "The \"carrots at top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81], "bbox": [0.310125, 0.00117096018735363, 0.6039687499999999, 0.2473302107728337], "iscrowd": 0, "area": 15038.462400000002, "rle": {"size": [427, 640], "counts": "Ybb21Z=2M3N2M3M2O2M3M3M3N2N2N2N2N2N2N2N2N2N1O2N1XDcN`;^1^DeN`;]1]DeNb;d1N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N1O2N2N2N2N6J2N0000000000000000O100O1O10000001O000000O1O100L400O100O10000O10000O1000000000000O100000000001O1O001O000000000000000000000000001O00000000001O000000001O00000000001O00001O00001O00001O00001O000000DTFZMk9g2<00001O001gEVMo9j2nEZMQ:P301O000000N2N2N2N2N2O1N2N2M3N2N2N2M3N2N2L4L4K5L4L4N200000000XOiD\\OW;a0k0M3L4M3M3L4M\\mY3"}, "label": "carrots at top"}]}
{"id": 572307, "image": "COCO_train2014_000000572307.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrot slices on a vegetable tray\"."}], "task": "refering", "answer_template": "The \"carrot slices on a vegetable tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 30, 31, 32, 33, 34, 35, 36, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81], "bbox": [0.310125, 0.00117096018735363, 0.6039687499999999, 0.2473302107728337], "iscrowd": 0, "area": 15038.462400000002, "rle": {"size": [427, 640], "counts": "Ybb21Z=2M3N2M3M2O2M3M3M3N2N2N2N2N2N2N2N2N2N1O2N1XDcN`;^1^DeN`;]1]DeNb;d1N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N1O2N2N2N2N6J2N0000000000000000O100O1O10000001O000000O1O100L400O100O10000O10000O1000000000000O100000000001O1O001O000000000000000000000000001O00000000001O000000001O00000000001O00001O00001O00001O00001O000000DTFZMk9g2<00001O001gEVMo9j2nEZMQ:P301O000000N2N2N2N2N2O1N2N2M3N2N2N2M3N2N2L4L4K5L4L4N200000000XOiD\\OW;a0k0M3L4M3M3L4M\\mY3"}, "label": "carrot slices on a vegetable tray"}]}
{"id": 482195, "image": "COCO_train2014_000000482195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dish of broccoli , carrots , and cheese\"."}], "task": "refering", "answer_template": "The \"a dish of broccoli , carrots , and cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.149390625, 0.9948125, 0.9861875], "iscrowd": 0, "area": 209669.4183, "rle": {"size": [640, 480], "counts": "f>Q5m4R:oNQ100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100001O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O00001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N3M4L4L4L3M4L4L4L4L5K4L4L4L5K4L4L5K4L4L4L5K4L00001O00001O001O00001O00001O00001O001O00001O00001O0000O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O11O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O00Xm0"}, "label": "a dish of broccoli , carrots , and cheese"}]}
{"id": 482195, "image": "COCO_train2014_000000482195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the food in the black dish\"."}], "task": "refering", "answer_template": "The \"the food in the black dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 71, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.149390625, 0.9948125, 0.9861875], "iscrowd": 0, "area": 209669.4183, "rle": {"size": [640, 480], "counts": "f>Q5m4R:oNQ100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100001O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O00001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000000000001O00000000000000000000000000000000000000001O0000001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N3M4L4L4L3M4L4L4L4L5K4L4L4L5K4L4L5K4L4L4L5K4L00001O00001O001O00001O00001O00001O001O00001O00001O0000O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O11O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O00Xm0"}, "label": "the food in the black dish"}]}
{"id": 482195, "image": "COCO_train2014_000000482195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black tray holding broccoli and other vegetables\"."}], "task": "refering", "answer_template": "The \"black tray holding broccoli and other vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 34, 35, 36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 68, 71, 72, 73, 74, 75, 76, 77, 78, 79, 89, 90, 91, 92, 93, 94, 95, 96, 97, 112, 113, 114, 115, 130, 131, 132, 133, 148, 149, 150, 151, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 200, 201, 202, 203, 220], "bbox": [0.0069166666666666664, 0.04675, 1.0, 0.5506562500000001], "iscrowd": 0, "area": 36686.8247, "rle": {"size": [640, 480], "counts": "_o15Ub0f1UOk00O100O101O0O100O10000O10001N100O10000O100O101O0O10000O100O10000O2O0O10000O10000O101O00000000000000001O00000000000000000010O1O1O001O1O1O00100O1O1O001O1O10O100000O0100000000O0100000O10O100000O1000O10O1000000O010000000O10O1000O100000O010O1O1O1O001O1O1O010O0000000000000001O00000000000000001O0000000000000000001O00000000000000001O000000000O1000001O0000000000001O0O10001O0000001O00001O0O10001O0000001O0000001O0O10001O00001O0000001O000O101O0000001O0000001O000O2O0000001O0000001O0000010O01O1O010O1O010O1O00100O001O10O01O1O010O1O00100O00100O001O10O01O1O010O1O00100O001O10O01O10O01O1O010O1O010000O10000O10O1000O10000O10000O01000O10000O10000O0100000O10000O1001N101O001O001O001O0O2O001O001O004L5K5J6K6J5K5K5K5K5K4L2M4M3M2N1OO1O100O1O00100O1O100O1O1O10O01O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100N2N2N2N2N2M3N2N2N2N2N2N2M3O10000O10000000000O100000000O10000000000O10000000000O100000000O1000001O1O1N101O1O1O001N2O1O1O001N2O1O001O1O1N101O1PObG"}, "label": "black tray holding broccoli and other vegetables"}]}
{"id": 482195, "image": "COCO_train2014_000000482195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black dish\"."}], "task": "refering", "answer_template": "The \"the black dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 34, 35, 36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 68, 71, 72, 73, 74, 75, 76, 77, 78, 79, 89, 90, 91, 92, 93, 94, 95, 96, 97, 112, 113, 114, 115, 130, 131, 132, 133, 148, 149, 150, 151, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 200, 201, 202, 203, 220], "bbox": [0.0069166666666666664, 0.04675, 1.0, 0.5506562500000001], "iscrowd": 0, "area": 36686.8247, "rle": {"size": [640, 480], "counts": "_o15Ub0f1UOk00O100O101O0O100O10000O10001N100O10000O100O101O0O10000O100O10000O2O0O10000O10000O101O00000000000000001O00000000000000000010O1O1O001O1O1O00100O1O1O001O1O10O100000O0100000000O0100000O10O100000O1000O10O1000000O010000000O10O1000O100000O010O1O1O1O001O1O1O010O0000000000000001O00000000000000001O0000000000000000001O00000000000000001O000000000O1000001O0000000000001O0O10001O0000001O00001O0O10001O0000001O0000001O0O10001O00001O0000001O000O101O0000001O0000001O000O2O0000001O0000001O0000010O01O1O010O1O010O1O00100O001O10O01O1O010O1O00100O00100O001O10O01O1O010O1O00100O001O10O01O10O01O1O010O1O010000O10000O10O1000O10000O10000O01000O10000O10000O0100000O10000O1001N101O001O001O001O0O2O001O001O004L5K5J6K6J5K5K5K5K5K4L2M4M3M2N1OO1O100O1O00100O1O100O1O1O10O01O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100N2N2N2N2N2M3N2N2N2N2N2N2M3O10000O10000000000O100000000O10000000000O10000000000O100000000O1000001O1O1N101O1O1O001N2O1O1O001N2O1O001O1O1N101O1PObG"}, "label": "the black dish"}]}
{"id": 23450, "image": "COCO_train2014_000000023450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding food in his hand\"."}], "task": "refering", "answer_template": "The \"a man holding food in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 82, 83, 84, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 269, 270], "bbox": [0.502, 0.08686182669789227, 0.783703125, 0.7689929742388759], "iscrowd": 0, "area": 27727.172799999993, "rle": {"size": [427, 640], "counts": "RaV41Z=0O100O101N100O100O100O2O0O100O1O101N100O100O101N100O100O2O0O1O100O1O1O001O1N2O001lDSOi9n0TFUOk9l0UFUOi9k0WFVOg9l0YFTOf9m0ZFTOd9m0kEoNI4k8i0WG45POI3S8b1jG_O9mNI3l7k1kGZO`0hNI4j7[3^HaLH4i7\\3_HaLF4j7\\3_HaLG4i7Z3aHbLF4h7Z3cHbLD5h7Z3cHbLE4g7Z3eHbLC5g7Y3gHbLB5g7X3hHcLA5k4j0`L_2SOcLA6j4j0`L]2VOdL_O5j4m0^LZ2ZOgLZO3m4o0\\LX2]OhLXO1n4Q1\\LU2^O[MT4d0[LQ2B[MR4g0YLn1F[MP4i0XLm1F]MP4i0WLk1H]MP4k0VLj1F]MR4l0VLj1D\\MU4m0TLj1DZMW4n0SLl1BWMZ4o0RL_3n3aLRL_3m3cLQL_3n3bLPL_3o3bLPL^3P4dLoK[3R4eLmK[3T4fLjKZ3W4fLhKZ3Y4gLeKY3]4hLaKW3`4lL\\KU3d4h21O100O100O100O10000O100O100O101N10001O001O100O1O1O100O1O100O3UIhJ]5a5XJiJ`5`5UJfJh5\\5SJhJk5Y5RJjJm5X5oIjJR6V61O1002O0O2N1O101N00010O00010O010O01O010O0010O01O010O010O0100O010O100O10O01XNgH]MY7a2jH]MW7a2mH]MR7b2QI\\MP7b2TI\\Ml6b2XIVMn6h2k1O2N2N2N3M4L3N2M3L4D=D;D`0A`1_N][i1"}, "label": "a man holding food in his hand"}]}
{"id": 23450, "image": "COCO_train2014_000000023450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man preparing food in a kitchen\"."}], "task": "refering", "answer_template": "The \"a man preparing food in a kitchen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 82, 83, 84, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 269, 270], "bbox": [0.502, 0.08686182669789227, 0.783703125, 0.7689929742388759], "iscrowd": 0, "area": 27727.172799999993, "rle": {"size": [427, 640], "counts": "RaV41Z=0O100O101N100O100O100O2O0O100O1O101N100O100O101N100O100O2O0O1O100O1O1O001O1N2O001lDSOi9n0TFUOk9l0UFUOi9k0WFVOg9l0YFTOf9m0ZFTOd9m0kEoNI4k8i0WG45POI3S8b1jG_O9mNI3l7k1kGZO`0hNI4j7[3^HaLH4i7\\3_HaLF4j7\\3_HaLG4i7Z3aHbLF4h7Z3cHbLD5h7Z3cHbLE4g7Z3eHbLC5g7Y3gHbLB5g7X3hHcLA5k4j0`L_2SOcLA6j4j0`L]2VOdL_O5j4m0^LZ2ZOgLZO3m4o0\\LX2]OhLXO1n4Q1\\LU2^O[MT4d0[LQ2B[MR4g0YLn1F[MP4i0XLm1F]MP4i0WLk1H]MP4k0VLj1F]MR4l0VLj1D\\MU4m0TLj1DZMW4n0SLl1BWMZ4o0RL_3n3aLRL_3m3cLQL_3n3bLPL_3o3bLPL^3P4dLoK[3R4eLmK[3T4fLjKZ3W4fLhKZ3Y4gLeKY3]4hLaKW3`4lL\\KU3d4h21O100O100O100O10000O100O100O101N10001O001O100O1O1O100O1O100O3UIhJ]5a5XJiJ`5`5UJfJh5\\5SJhJk5Y5RJjJm5X5oIjJR6V61O1002O0O2N1O101N00010O00010O010O01O010O0010O01O010O010O0100O010O100O10O01XNgH]MY7a2jH]MW7a2mH]MR7b2QI\\MP7b2TI\\Ml6b2XIVMn6h2k1O2N2N2N3M4L3N2M3L4D=D;D`0A`1_N][i1"}, "label": "a man preparing food in a kitchen"}]}
{"id": 351134, "image": "COCO_train2014_000000351134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady is wearing a maroon shirt\"."}], "task": "refering", "answer_template": "The \"the lady is wearing a maroon shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 121, 122, 123, 137, 138, 139, 153, 154, 155, 169, 170, 171, 172, 185, 186, 187, 202, 203, 218, 219, 234, 235], "bbox": [0.5887586206896552, 0.276703125, 0.7596091954022989, 0.65290625], "iscrowd": 0, "area": 10912.543050000002, "rle": {"size": [640, 435], "counts": "WYP53fc09G:F9G:G8G:F:F;E:F;E;]Ob0J7J6J6K4L5K6I7H8H7J7H8J6K4L5J6K5J6K4K4M01O001O1O1O001O1O1O001O1O001M3M3L3N3L4F:QJ_BU5i=gJ\\Bo4n=lJWBj4S>QKRBS3GVNS?i1TAhMU?W2RAYMX?f2R1O1O1dMR_Oc1o`0UN]_Og1d`0oMh_On1Ya0L4L4L4K5L3M4L4L4L4KR[Q2"}, "label": "the lady is wearing a maroon shirt"}]}
{"id": 351134, "image": "COCO_train2014_000000351134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing a maroon shirt and black pants\"."}], "task": "refering", "answer_template": "The \"the woman wearing a maroon shirt and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 121, 122, 123, 137, 138, 139, 153, 154, 155, 169, 170, 171, 172, 185, 186, 187, 202, 203, 218, 219, 234, 235], "bbox": [0.5887586206896552, 0.276703125, 0.7596091954022989, 0.65290625], "iscrowd": 0, "area": 10912.543050000002, "rle": {"size": [640, 435], "counts": "WYP53fc09G:F9G:G8G:F:F;E:F;E;]Ob0J7J6J6K4L5K6I7H8H7J7H8J6K4L5J6K5J6K4K4M01O001O1O1O001O1O1O001O1O001M3M3L3N3L4F:QJ_BU5i=gJ\\Bo4n=lJWBj4S>QKRBS3GVNS?i1TAhMU?W2RAYMX?f2R1O1O1dMR_Oc1o`0UN]_Og1d`0oMh_On1Ya0L4L4L4K5L3M4L4L4L4KR[Q2"}, "label": "the woman wearing a maroon shirt and black pants"}]}
{"id": 351134, "image": "COCO_train2014_000000351134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man dressed in an orange safety suit , holding a stop sign\"."}], "task": "refering", "answer_template": "The \"a man dressed in an orange safety suit , holding a stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 66, 67, 68, 69, 70, 82, 83, 84, 85, 98, 99, 100, 101, 114, 115, 116, 117, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 313, 314, 315, 316, 325, 326, 327, 330, 331, 332, 340, 341, 342, 343, 346, 347, 348], "bbox": [0.14843678160919538, 0.139015625, 0.7917011494252874, 0.952921875], "iscrowd": 0, "area": 60241.630049999985, "rle": {"size": [640, 435], "counts": "PiX1`0]c0b0_Ob0]Ob0_O7H3N3L3M4L3L4M4K4M3L5L3N3N1N2O2N1N2O2N1N3N1O1N3N1N2O2N10001N101O0O101O000O2O001O0O101O000`A[Le<f3jBmLS=S5M3M3hKPIZKT7a4nH_KT7^4lHbKW7Z4jHfKY7V4hHjK[7R4gHmK\\7P4dHPL`7k3aHULb7g3`HXLc7e3]H[Lf7a3[H_Li7\\3XHdLm7V3UHiLP8R3PHoLT8iMoGf2NAW8XM]HQ3^OGY8eLmH^3jNM^8SLZIj3YN4o:FRE;S;^OoDa0V;ZOjDg0j2PLn4R3YHn0d2VLW5f2WHT1\\2[Lb5\\2RHY1W2aLl5o1nGa1P2fLV6c1kGg1j1kL`6X1hGm1c1PMj6m0cGS2_1UMR7b0`GZ2Y1XM]77\\Ga2R1]Mf7I\\Gj2j0bMn7[O\\GT3a0eMY8mNZG^39jMa8_NZGh31mMj8QNZGQ4KPNd:Q2YEQNg:o1WETNh:m1UEVNj:j1TEXNl:i1QEZNo:e1PE]No:d1nD_NQ;a1mDaNS;`1kDbNT;^1kDdNU;\\1iDfNV;Z1iDgNW;Z1gDhNX;X1hDiNW;X1gDjNX;W1fDjN[;V1cDlN\\;U1bDmN];S1bDnN];T1bDmN];T1aDmN_;T1_DmNa;T1]DnNa;T1]DmNc;T1\\DlNd;h400O101O000O10001O0O1000001N100N2O1N2O2M2O1N2O1N3N1N2O1N2O1J7J5J6J6J6K6I6J6VNVFXIP:i6]FdHh9]7^11N1O1O1O1O100O1O1O1O100O1O1O1O11O4L3M3M4L3M3M4L3M7I8H8H7J7H7I4L5K4L5K5K4L5L3L3N0O2O1O0O2O1N10100O001O10O01O10O01O1O0000001O000000001N100000001O000000001O0000001O0N2N2O1N3M2O0O2N2O0O2O1M2N3M3M2O2M3M2N3M3M2N3M3M2N3M3M2N3M3M2O2M2N2N2M3M3L4M3L4L5L3L4M3K5KnZh1"}, "label": "a man dressed in an orange safety suit , holding a stop sign"}]}
{"id": 351134, "image": "COCO_train2014_000000351134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in an orange jumpsuit\"."}], "task": "refering", "answer_template": "The \"a person in an orange jumpsuit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 66, 67, 68, 69, 70, 82, 83, 84, 85, 98, 99, 100, 101, 114, 115, 116, 117, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 313, 314, 315, 316, 325, 326, 327, 330, 331, 332, 340, 341, 342, 343, 346, 347, 348], "bbox": [0.14843678160919538, 0.139015625, 0.7917011494252874, 0.952921875], "iscrowd": 0, "area": 60241.630049999985, "rle": {"size": [640, 435], "counts": "PiX1`0]c0b0_Ob0]Ob0_O7H3N3L3M4L3L4M4K4M3L5L3N3N1N2O2N1N2O2N1N3N1O1N3N1N2O2N10001N101O0O101O000O2O001O0O101O000`A[Le<f3jBmLS=S5M3M3hKPIZKT7a4nH_KT7^4lHbKW7Z4jHfKY7V4hHjK[7R4gHmK\\7P4dHPL`7k3aHULb7g3`HXLc7e3]H[Lf7a3[H_Li7\\3XHdLm7V3UHiLP8R3PHoLT8iMoGf2NAW8XM]HQ3^OGY8eLmH^3jNM^8SLZIj3YN4o:FRE;S;^OoDa0V;ZOjDg0j2PLn4R3YHn0d2VLW5f2WHT1\\2[Lb5\\2RHY1W2aLl5o1nGa1P2fLV6c1kGg1j1kL`6X1hGm1c1PMj6m0cGS2_1UMR7b0`GZ2Y1XM]77\\Ga2R1]Mf7I\\Gj2j0bMn7[O\\GT3a0eMY8mNZG^39jMa8_NZGh31mMj8QNZGQ4KPNd:Q2YEQNg:o1WETNh:m1UEVNj:j1TEXNl:i1QEZNo:e1PE]No:d1nD_NQ;a1mDaNS;`1kDbNT;^1kDdNU;\\1iDfNV;Z1iDgNW;Z1gDhNX;X1hDiNW;X1gDjNX;W1fDjN[;V1cDlN\\;U1bDmN];S1bDnN];T1bDmN];T1aDmN_;T1_DmNa;T1]DnNa;T1]DmNc;T1\\DlNd;h400O101O000O10001O0O1000001N100N2O1N2O2M2O1N2O1N3N1N2O1N2O1J7J5J6J6J6K6I6J6VNVFXIP:i6]FdHh9]7^11N1O1O1O1O100O1O1O1O100O1O1O1O11O4L3M3M4L3M3M4L3M7I8H8H7J7H7I4L5K4L5K5K4L5L3L3N0O2O1O0O2O1N10100O001O10O01O10O01O1O0000001O000000001N100000001O000000001O0000001O0N2N2O1N3M2O0O2N2O0O2O1M2N3M3M2O2M3M2N3M3M2N3M3M2N3M3M2N3M3M2O2M2N2N2M3M3L4M3L4L5L3L4M3K5KnZh1"}, "label": "a person in an orange jumpsuit"}]}
{"id": 351134, "image": "COCO_train2014_000000351134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady dressed all in white\"."}], "task": "refering", "answer_template": "The \"lady dressed all in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 114, 115, 129, 130, 131, 145, 146, 161, 162, 177, 178, 179, 193, 194, 210, 226, 241, 242, 243], "bbox": [0.06101149425287356, 0.27349999999999997, 0.21360919540229883, 0.67775], "iscrowd": 0, "area": 7620.4437000000025, "rle": {"size": [640, 435], "counts": "VTa05ec06J5M4N2O0O2O1N2N101N2O0O2b]OFUa0;i^OLPa05n^O3k`0OR_O9h`0HU_O`0f`0_OX_Oi0c`0VO[_OR1l2lMb:R1`BZ1g2fMi:d3PE^LQ:MYDk3`1YLV:NYDn3[1ULZ:1YDn3X1RL^:3WDQ4U1mKb:5XDP1GY1Y1cMf:8XDk0M\\1o0bMk:9WDi04\\1e0cMn:;WDf0:^1;bMPOCn;i0\\Dd0>a12`MUOCm;k0\\Da0c0d1I^MZOCn;j0[Db0e0e1D`M];>TD>m0d1_OaM_;R2SE>[OaMb;Q2UE>WOaMd;Q2VE>TObMf;P2XE\\1g:fNYEY1g:gNZEX1f:gN]EW1c:iN^EV1b:eLVD^1X1m1c:bLYD_1T1o1c:`L^D]1P1R2c:_LaD]1m0S2b:_LeD[1i0V2b:^LgD[1g0W2c:[LiD\\1f0X2a:[LjD]1e0X2c;gM]DGdNU1o<TO]DDiNV1j<UO_DBkNW1g<RObDDkNY1c<POeDEUOP1W<YOfDCDBQO7f<1lBMi16f0If:1nBOf15k0Hb:2PCOc14R1G[:4SCOb13U1GW:4VC0^12\\1FQ:6WC1\\11k>Km_O2Y1Om>Mn_O1V1Ol?0T2NW`e6"}, "label": "lady dressed all in white"}]}
{"id": 351134, "image": "COCO_train2014_000000351134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a turquoise purse\"."}], "task": "refering", "answer_template": "The \"a woman with a turquoise purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 114, 115, 129, 130, 131, 145, 146, 161, 162, 177, 178, 179, 193, 194, 210, 226, 241, 242, 243], "bbox": [0.06101149425287356, 0.27349999999999997, 0.21360919540229883, 0.67775], "iscrowd": 0, "area": 7620.4437000000025, "rle": {"size": [640, 435], "counts": "VTa05ec06J5M4N2O0O2O1N2N101N2O0O2b]OFUa0;i^OLPa05n^O3k`0OR_O9h`0HU_O`0f`0_OX_Oi0c`0VO[_OR1l2lMb:R1`BZ1g2fMi:d3PE^LQ:MYDk3`1YLV:NYDn3[1ULZ:1YDn3X1RL^:3WDQ4U1mKb:5XDP1GY1Y1cMf:8XDk0M\\1o0bMk:9WDi04\\1e0cMn:;WDf0:^1;bMPOCn;i0\\Dd0>a12`MUOCm;k0\\Da0c0d1I^MZOCn;j0[Db0e0e1D`M];>TD>m0d1_OaM_;R2SE>[OaMb;Q2UE>WOaMd;Q2VE>TObMf;P2XE\\1g:fNYEY1g:gNZEX1f:gN]EW1c:iN^EV1b:eLVD^1X1m1c:bLYD_1T1o1c:`L^D]1P1R2c:_LaD]1m0S2b:_LeD[1i0V2b:^LgD[1g0W2c:[LiD\\1f0X2a:[LjD]1e0X2c;gM]DGdNU1o<TO]DDiNV1j<UO_DBkNW1g<RObDDkNY1c<POeDEUOP1W<YOfDCDBQO7f<1lBMi16f0If:1nBOf15k0Hb:2PCOc14R1G[:4SCOb13U1GW:4VC0^12\\1FQ:6WC1\\11k>Km_O2Y1Om>Mn_O1V1Ol?0T2NW`e6"}, "label": "a woman with a turquoise purse"}]}
{"id": 580511, "image": "COCO_train2014_000000580511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat was going in water\"."}], "task": "refering", "answer_template": "The \"a boat was going in water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 63, 75, 87, 98, 99, 100, 101, 110, 111, 112, 113, 122, 123, 124, 125, 126, 134, 135, 136, 137], "bbox": [0.17201807228915664, 0.24214, 0.514277108433735, 0.63286], "iscrowd": 0, "area": 8229.14125, "rle": {"size": [500, 332], "counts": "nSl03Z?:F:K4L3M3O1N3N1O1N1000000O10000O100oNQ1O1000000O1000000O1000000O1000000O1]OdMnC\\2R<jMhCV2W<PNdCP2\\<d00000000000000000TM^OeFc0Z9l2O1OTJ[Gk4m:jNS1mN00O10000000000000000000000O1000000001O4L3M3M1O1O1OS1mN[NQOgDm0Y;7dCH\\<h1000000000000000000f1ZN0000O10000000O010O1O1O101N1O2O0O2N2N6C=DkZ^2"}, "label": "a boat was going in water"}]}
{"id": 580511, "image": "COCO_train2014_000000580511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat with the orange buoys hanging off the back\"."}], "task": "refering", "answer_template": "The \"the boat with the orange buoys hanging off the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 63, 75, 87, 98, 99, 100, 101, 110, 111, 112, 113, 122, 123, 124, 125, 126, 134, 135, 136, 137], "bbox": [0.17201807228915664, 0.24214, 0.514277108433735, 0.63286], "iscrowd": 0, "area": 8229.14125, "rle": {"size": [500, 332], "counts": "nSl03Z?:F:K4L3M3O1N3N1O1N1000000O10000O100oNQ1O1000000O1000000O1000000O1000000O1]OdMnC\\2R<jMhCV2W<PNdCP2\\<d00000000000000000TM^OeFc0Z9l2O1OTJ[Gk4m:jNS1mN00O10000000000000000000000O1000000001O4L3M3M1O1O1OS1mN[NQOgDm0Y;7dCH\\<h1000000000000000000f1ZN0000O10000000O010O1O1O101N1O2O0O2N2N6C=DkZ^2"}, "label": "the boat with the orange buoys hanging off the back"}]}
{"id": 580511, "image": "COCO_train2014_000000580511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and black colour boat\"."}], "task": "refering", "answer_template": "The \"a blue and black colour boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 103, 104, 105, 106, 115, 116, 117, 118, 127, 128, 129, 130, 139, 140, 141], "bbox": [0.62, 0.41298, 0.8731626506024096, 0.62372], "iscrowd": 0, "area": 7312.015649999999, "rle": {"size": [500, 332], "counts": "ikT3=T?7J5H5I7J7H7TOl0L4L4L4L4L4L4O2O0000000000000000001O00000001O00000000000000000000000000000000000000000000000000000000000000000000000O1O100O1PNoC=Q<@TD=m;@WD?i;^O[DZOYOl0]<H]DZO[Oi0Y<J`D[O\\Of0U<LcD\\OYOg0U<JUE5l:GWE8j:FXE9j:CYE<h:A[E>g:^O\\Ea0l<H8HgWd0"}, "label": "a blue and black colour boat"}]}
{"id": 580511, "image": "COCO_train2014_000000580511.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black boat at a dock\"."}], "task": "refering", "answer_template": "The \"a black boat at a dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 103, 104, 105, 106, 115, 116, 117, 118, 127, 128, 129, 130, 139, 140, 141], "bbox": [0.62, 0.41298, 0.8731626506024096, 0.62372], "iscrowd": 0, "area": 7312.015649999999, "rle": {"size": [500, 332], "counts": "ikT3=T?7J5H5I7J7H7TOl0L4L4L4L4L4L4O2O0000000000000000001O00000001O00000000000000000000000000000000000000000000000000000000000000000000000O1O100O1PNoC=Q<@TD=m;@WD?i;^O[DZOYOl0]<H]DZO[Oi0Y<J`D[O\\Of0U<LcD\\OYOg0U<JUE5l:GWE8j:FXE9j:CYE<h:A[E>g:^O\\Ea0l<H8HgWd0"}, "label": "a black boat at a dock"}]}
{"id": 105374, "image": "COCO_train2014_000000105374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry , silver car\"."}], "task": "refering", "answer_template": "The \"a blurry , silver car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 238, 239, 253, 254, 268, 269, 283, 284], "bbox": [0.8351990632318501, 0.491609375, 1.0, 0.8197656249999999], "iscrowd": 0, "area": 11157.559799999995, "rle": {"size": [640, 427], "counts": "U^o69gc0<D<D1O001N2O:o]OeNh`0c1m^OcNQa0d1d^OcNXa0U2N2M4M2M3N2M4M4K6f_OPMZ?l3N2O1O2M2O1N2O1O1Ok0UO5K2N2N5K2N1O1O2N1O1O00001O0000001O000000001O0000001O000000001O0000001O00001O00000UF"}, "label": "a blurry , silver car"}]}
{"id": 105374, "image": "COCO_train2014_000000105374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car which is in the right\"."}], "task": "refering", "answer_template": "The \"a white car which is in the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 238, 239, 253, 254, 268, 269, 283, 284], "bbox": [0.8351990632318501, 0.491609375, 1.0, 0.8197656249999999], "iscrowd": 0, "area": 11157.559799999995, "rle": {"size": [640, 427], "counts": "U^o69gc0<D<D1O001N2O:o]OeNh`0c1m^OcNQa0d1d^OcNXa0U2N2M4M2M3N2M4M4K6f_OPMZ?l3N2O1O2M2O1N2O1O1Ok0UO5K2N2N5K2N1O1O2N1O1O00001O0000001O000000001O0000001O000000001O0000001O00001O00000UF"}, "label": "a white car which is in the right"}]}
{"id": 244646, "image": "COCO_train2014_000000244646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large stack of magazines on a wood floor\"."}], "task": "refering", "answer_template": "The \"a large stack of magazines on a wood floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314], "bbox": [0.8431934731934733, 0.586734375, 1.0, 0.9145937500000001], "iscrowd": 0, "area": 13193.735549999998, "rle": {"size": [640, 429], "counts": "ngR7^1b`0Q2oMR2H7O2N2N1O2N2N1O2N1O2O1N1O2N2N1O2N0000000001O0O10000000000O1000001O00000O10000000000O101O000000000O10000000001N1000000000000O1000SD"}, "label": "a large stack of magazines on a wood floor"}]}
{"id": 244646, "image": "COCO_train2014_000000244646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large collection of various books and magazines\"."}], "task": "refering", "answer_template": "The \"a large collection of various books and magazines\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282, 283, 284, 297, 298, 299, 312, 313, 314], "bbox": [0.8431934731934733, 0.586734375, 1.0, 0.9145937500000001], "iscrowd": 0, "area": 13193.735549999998, "rle": {"size": [640, 429], "counts": "ngR7^1b`0Q2oMR2H7O2N2N1O2N2N1O2N1O2O1N1O2N2N1O2N0000000001O0O10000000000O1000001O00000O10000000000O101O000000000O10000000001N1000000000000O1000SD"}, "label": "a large collection of various books and magazines"}]}
{"id": 244646, "image": "COCO_train2014_000000244646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white , four - legged stool\"."}], "task": "refering", "answer_template": "The \"a white , four - legged stool\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 286, 287, 288, 289, 301, 302, 303, 304, 305, 316, 317, 320, 332], "bbox": [0.061864801864801866, 0.694265625, 0.3517948717948718, 0.9914375], "iscrowd": 0, "area": 11776.481050000002, "rle": {"size": [640, 429], "counts": "\\^a03jc04K4M3M3L01ON2L4L4M3N2M3N2N6I8I7I7H9H7H8I7I7H8L4N2N2N2O1O1O1i@TMX=i410000M2L5O1O1M3J6K5J5K6K5J6J6J6O1O10O1000O10000000O01000000O1000O10O100000000O10O11O0O10001O0O2O0002O1NO2N1K6N101O000O2O001N101O0O2O;E7I6I6K6J5J7J6J5N3N1O2N1N2O1O1O1O1N2O001L4K5K5K5K5I\\U]5"}, "label": "a white , four - legged stool"}]}
{"id": 244646, "image": "COCO_train2014_000000244646.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stool made of plastic and four legs\"."}], "task": "refering", "answer_template": "The \"the stool made of plastic and four legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 286, 287, 288, 289, 301, 302, 303, 304, 305, 316, 317, 320, 332], "bbox": [0.061864801864801866, 0.694265625, 0.3517948717948718, 0.9914375], "iscrowd": 0, "area": 11776.481050000002, "rle": {"size": [640, 429], "counts": "\\^a03jc04K4M3M3L01ON2L4L4M3N2M3N2N6I8I7I7H9H7H8I7I7H8L4N2N2N2O1O1O1i@TMX=i410000M2L5O1O1M3J6K5J5K6K5J6J6J6O1O10O1000O10000000O01000000O1000O10O100000000O10O11O0O10001O0O2O0002O1NO2N1K6N101O000O2O001N101O0O2O;E7I6I6K6J5J7J6J5N3N1O2N1N2O1O1O1O1N2O001L4K5K5K5K5I\\U]5"}, "label": "the stool made of plastic and four legs"}]}
{"id": 277418, "image": "COCO_train2014_000000277418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair a little girl is eating in\"."}], "task": "refering", "answer_template": "The \"the chair a little girl is eating in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 178, 179, 180, 181, 201, 202, 203, 204, 205, 225, 226, 227, 228, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.76475, 0.4430210772833723, 1.0, 0.987096018735363], "iscrowd": 0, "area": 21666.765250000004, "rle": {"size": [427, 640], "counts": "XT\\61Q=:F:G9F:J6kIcN:^1EbN;_1CbN=_1BaN>`1@aN`0`1_O`Na0a1]O`Nc0a1\\O_Nd0b1ZO_Nf0b1YO^Ng0c1WO^Ni0c1VO]Nj0d1TO]Nm0c1RO]Nn0d1PO]NP1d1oN\\NQ1e1mN\\NS1e1lN[NT1f1jN[NV1f1iNZNW1h1fNYNZ1h1eNXN[1j1bNWN^1j1aNVN_1k1_NVNa1l1]NTNc1m1[NTNe1m1ZNSNf1o1WNRNi1o1VNQNj1P2TNQNl1Q2QNPNo1Q2PNoMP2S2mMnMS2S2lMmMU2S2iMnMW2T2fMmM[2S2dMmM\\2T2bMlM`2U2^MkMb2V2\\MkMe2V2WMlMi2W2iLTNX3Q2VL[Nj3l1cK`N]4g1PKeNQ5`1]JlNc5[1kIPOV6R4001O001O001O001O001O001O001O001O001O001O001O00001O001O001O0O2O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N3M5K4L5K5J6K5K4L1O2N2N1O2N1O2N1O2N2N1O00000000000000000000000000^F"}, "label": "the chair a little girl is eating in"}]}
{"id": 277418, "image": "COCO_train2014_000000277418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"armed wooden chair\"."}], "task": "refering", "answer_template": "The \"armed wooden chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 178, 179, 180, 181, 201, 202, 203, 204, 205, 225, 226, 227, 228, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.76475, 0.4430210772833723, 1.0, 0.987096018735363], "iscrowd": 0, "area": 21666.765250000004, "rle": {"size": [427, 640], "counts": "XT\\61Q=:F:G9F:J6kIcN:^1EbN;_1CbN=_1BaN>`1@aN`0`1_O`Na0a1]O`Nc0a1\\O_Nd0b1ZO_Nf0b1YO^Ng0c1WO^Ni0c1VO]Nj0d1TO]Nm0c1RO]Nn0d1PO]NP1d1oN\\NQ1e1mN\\NS1e1lN[NT1f1jN[NV1f1iNZNW1h1fNYNZ1h1eNXN[1j1bNWN^1j1aNVN_1k1_NVNa1l1]NTNc1m1[NTNe1m1ZNSNf1o1WNRNi1o1VNQNj1P2TNQNl1Q2QNPNo1Q2PNoMP2S2mMnMS2S2lMmMU2S2iMnMW2T2fMmM[2S2dMmM\\2T2bMlM`2U2^MkMb2V2\\MkMe2V2WMlMi2W2iLTNX3Q2VL[Nj3l1cK`N]4g1PKeNQ5`1]JlNc5[1kIPOV6R4001O001O001O001O001O001O001O001O001O001O001O00001O001O001O0O2O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N3M5K4L5K5J6K5K4L1O2N2N1O2N1O2N1O2N2N1O00000000000000000000000000^F"}, "label": "armed wooden chair"}]}
{"id": 277418, "image": "COCO_train2014_000000277418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"patio chair with black lattice style back\"."}], "task": "refering", "answer_template": "The \"patio chair with black lattice style back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303], "bbox": [0.0014375, 0.3903981264637002, 0.32854687499999996, 0.9389929742388758], "iscrowd": 0, "area": 40240.68665000002, "rle": {"size": [427, 640], "counts": "bb0]1n;j2VMj2VM9G0000O2O000000000O100000001O000O10000000000O2O000000000O100000001O00000O10000000001O0O1000000000000O101O00000000000O1000001O00000O10000000001O0O100000000000000O2O00000000000O1000001O00000O10000000001O000O1000000000000O2O000O1N2O1O1O10000O2O0O100O100O100O101O0O100O100O1000001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001N10001O001O00001O00001O001O00001O001O00001E:C<D=D;D=C=C<D=C=C<D=C<K6M3M2N3M3MjSc5"}, "label": "patio chair with black lattice style back"}]}
{"id": 277418, "image": "COCO_train2014_000000277418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark colored chair\"."}], "task": "refering", "answer_template": "The \"dark colored chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303], "bbox": [0.0014375, 0.3903981264637002, 0.32854687499999996, 0.9389929742388758], "iscrowd": 0, "area": 40240.68665000002, "rle": {"size": [427, 640], "counts": "bb0]1n;j2VMj2VM9G0000O2O000000000O100000001O000O10000000000O2O000000000O100000001O00000O10000000001O0O1000000000000O101O00000000000O1000001O00000O10000000001O0O100000000000000O2O00000000000O1000001O00000O10000000001O000O1000000000000O2O000O1N2O1O1O10000O2O0O100O100O100O101O0O100O100O1000001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001N10001O001O00001O00001O001O00001O001O00001E:C<D=D;D=C=C<D=C=C<D=C<K6M3M2N3M3MjSc5"}, "label": "dark colored chair"}]}
{"id": 547767, "image": "COCO_train2014_000000547767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right side of a dining room table\"."}], "task": "refering", "answer_template": "The \"the right side of a dining room table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 280, 281, 282], "bbox": [0.2127, 0.7268702290076337, 0.5507666666666667, 0.9978371501272265], "iscrowd": 0, "area": 9280.707299999998, "rle": {"size": [393, 600], "counts": "R^a11X<1O000O2O001N10001N101O0O101O001N10001N101O0O2O000O2O001O0O101O0O2O000O2O001N10001O0O2O001N10001N101O0O101O001N10001N101O0O101O0O2O001O0O101O0O2O001N2O1O1N2O1O1O1N3N1O3L:G?A1N0100O010O1000O0100O010O10OQOgFoNZ9P1fFQOZ9o0fFPOZ9P1fFQOZ9o0fFQOZ9n0fFRO[9n0eFROZ9n0fFRO[9n0eFRO[9m0eFSO[9n0eFRO[9m0eFSO\\9m0dFSO[9m0eFSO\\9m0dFSO\\9l0dFUO[9l0eFSO\\9l0dFUO\\9k0dFTO]9k0cFVO]9j0cFUO^9j0bFWO^9i0bFVO`9h0`FYO`9g0`FXOa9g0_FZOb9e0^F[Ob9d0_F[Ob9e0]F\\Oc9c0^F\\Od9c0[F^Oe9a0\\F^Oe9b0ZF^Oh9`0YF_Oh9a0WF_Oj9`0WF_Oj9a0UF_Om9?TF@m9`0RFAn9>SFAo9>PFBQ:=PFkNh:U1WEkNk:S1VEmNj:S1UEnNk:Q1VEnNk:R1TEoNm:o0TEQOl:n0TEROm:n0SEROn:l0REUOn:k0RETOo:k0QEVOo:j0QEVOP;h0PEXOQ;h0oDXOQ;g0PEYOQ;f0nD[OR;d0oD[OR;e0mD\\OS;c0nD\\OT;c0kD^OU;a0lD_OT;a0kD_OW;?jDAV;?iDAX;>iDBW;>hDBZ;<gDDY;<fDD[;;fDE[;:dDG\\;8eDG\\;9cDH];7dDI];6bDK^;4cDL];4bDM_;1bDO^;1aD0_;ObD1^;OaD2`;LaD4_;L`D5`;JaD5a;J_D6a;I_D8a;H_D8a;G_D9c;F]D:c;E^D;b;E]D<h;O0100O10O101O4L4KQZW3"}, "label": "the right side of a dining room table"}]}
{"id": 547767, "image": "COCO_train2014_000000547767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right half of a table with a white cloth on it\"."}], "task": "refering", "answer_template": "The \"the right half of a table with a white cloth on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 280, 281, 282], "bbox": [0.2127, 0.7268702290076337, 0.5507666666666667, 0.9978371501272265], "iscrowd": 0, "area": 9280.707299999998, "rle": {"size": [393, 600], "counts": "R^a11X<1O000O2O001N10001N101O0O101O001N10001N101O0O2O000O2O001O0O101O0O2O000O2O001N10001O0O2O001N10001N101O0O101O001N10001N101O0O101O0O2O001O0O101O0O2O001N2O1O1N2O1O1O1N3N1O3L:G?A1N0100O010O1000O0100O010O10OQOgFoNZ9P1fFQOZ9o0fFPOZ9P1fFQOZ9o0fFQOZ9n0fFRO[9n0eFROZ9n0fFRO[9n0eFRO[9m0eFSO[9n0eFRO[9m0eFSO\\9m0dFSO[9m0eFSO\\9m0dFSO\\9l0dFUO[9l0eFSO\\9l0dFUO\\9k0dFTO]9k0cFVO]9j0cFUO^9j0bFWO^9i0bFVO`9h0`FYO`9g0`FXOa9g0_FZOb9e0^F[Ob9d0_F[Ob9e0]F\\Oc9c0^F\\Od9c0[F^Oe9a0\\F^Oe9b0ZF^Oh9`0YF_Oh9a0WF_Oj9`0WF_Oj9a0UF_Om9?TF@m9`0RFAn9>SFAo9>PFBQ:=PFkNh:U1WEkNk:S1VEmNj:S1UEnNk:Q1VEnNk:R1TEoNm:o0TEQOl:n0TEROm:n0SEROn:l0REUOn:k0RETOo:k0QEVOo:j0QEVOP;h0PEXOQ;h0oDXOQ;g0PEYOQ;f0nD[OR;d0oD[OR;e0mD\\OS;c0nD\\OT;c0kD^OU;a0lD_OT;a0kD_OW;?jDAV;?iDAX;>iDBW;>hDBZ;<gDDY;<fDD[;;fDE[;:dDG\\;8eDG\\;9cDH];7dDI];6bDK^;4cDL];4bDM_;1bDO^;1aD0_;ObD1^;OaD2`;LaD4_;L`D5`;JaD5a;J_D6a;I_D8a;H_D8a;G_D9c;F]D:c;E^D;b;E]D<h;O0100O10O101O4L4KQZW3"}, "label": "the right half of a table with a white cloth on it"}]}
{"id": 547767, "image": "COCO_train2014_000000547767.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of the table that has 2 glasses with orange napkins in them\"."}], "task": "refering", "answer_template": "The \"the top of the table that has 2 glasses with orange napkins in them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 252, 253, 254, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 279, 280], "bbox": [0.0, 0.8212468193384224, 0.3509833333333333, 0.9870483460559796], "iscrowd": 0, "area": 7754.767050000005, "rle": {"size": [393, 600], "counts": "`;a0h;000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O101O0O10000O10000O10000O13M3M4L3M4L4L2N00000000000000000000001OM3J6I7I7I7K500O10000O10000O10000O10000O10001N10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000001O001O1O001O001O001O010O001O001O001O001O001O1O001O001O001O001O001O001O001O0O2O1N101N101N101N101N101N101N101N101N101N2O0O2O0O2O001N101N101N101N101N101N101N2O0O2OXZe4"}, "label": "the top of the table that has 2 glasses with orange napkins in them"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"arm with white sleeve\"."}], "task": "refering", "answer_template": "The \"arm with white sleeve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 32, 33, 46, 47, 48, 49, 50, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84], "bbox": [0.6562916666666666, 0.0, 0.9977708333333333, 0.201703125], "iscrowd": 0, "area": 10755.997399999998, "rle": {"size": [640, 480], "counts": "PdU61nc0100N2N2N2O1N2N2N1O2N2N2N2O1N2N2N2N2N2N2N2N2O1N00001O0000001O01O001L4K5J6K5J6K4O2001000000O1000000O100000000O010dNGQ_O9m`0IS_O7j`0MT_O7f`0LZ_O;\\`0Hd_O>V`0Bj_Od0P`0\\OP@k0h?VOX@P1b?QO]@U1]?kNc@\\1V?dNj@\\1V?dNj@\\1W?cNi@]1W?dNh@\\1X?dNh@\\1X?dNh@[1Y?eNg@[1Y?eNg@[1Y?fNf@Z1Z?fNf@Z1[?eNe@Z1\\?fNe@Y1[?gNe@Y1[?hNd@X1\\?hNd@X1\\?hNe@W1[?iNe@V1\\?jNd@V1]?jNb@V1^?jNb@V1^?kNb@T1^?lNb@S1_?nN`@R1`?nN`@R1`?oN_@Q1a?oN`@P1a?oN_@Q1a?PO^@o0d?PO\\@Q1d?oN[@Q1e?oN[@Q1f?oNY@Q1g?oNZ@Q1f?oNY@Q1h?nNX@R1h?oNW@Q1j?oNU@R1k?nNT@R1l?oNS@Q1n?POP@P1Q`0POn_OQ1Q`0POn_OP1S`0POm_Oo0S`0SOk_On0U`0ROj_On0W`0ROh_On0X`0SOg_Om0Z`0SOe_On0[`0ROd_On0\\`0SOc_Om0]`0TOb_Ol0[`0XOd_Oi0X`0[Oh_Od0U`0@j_O`0R`0Em_O;P`0Im_O:R`0Gk_O;U`0Ei_O=X`0Ce_O?[`0d1O00001O0000001O0000001O0000001O0000001O00001O0000001O0001O01O0001O100O1bMUf0"}, "label": "arm with white sleeve"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a white top and black shorts\"."}], "task": "refering", "answer_template": "The \"a person in a white top and black shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 32, 33, 46, 47, 48, 49, 50, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84], "bbox": [0.6562916666666666, 0.0, 0.9977708333333333, 0.201703125], "iscrowd": 0, "area": 10755.997399999998, "rle": {"size": [640, 480], "counts": "PdU61nc0100N2N2N2O1N2N2N1O2N2N2N2O1N2N2N2N2N2N2N2N2O1N00001O0000001O01O001L4K5J6K5J6K4O2001000000O1000000O100000000O010dNGQ_O9m`0IS_O7j`0MT_O7f`0LZ_O;\\`0Hd_O>V`0Bj_Od0P`0\\OP@k0h?VOX@P1b?QO]@U1]?kNc@\\1V?dNj@\\1V?dNj@\\1W?cNi@]1W?dNh@\\1X?dNh@\\1X?dNh@[1Y?eNg@[1Y?eNg@[1Y?fNf@Z1Z?fNf@Z1[?eNe@Z1\\?fNe@Y1[?gNe@Y1[?hNd@X1\\?hNd@X1\\?hNe@W1[?iNe@V1\\?jNd@V1]?jNb@V1^?jNb@V1^?kNb@T1^?lNb@S1_?nN`@R1`?nN`@R1`?oN_@Q1a?oN`@P1a?oN_@Q1a?PO^@o0d?PO\\@Q1d?oN[@Q1e?oN[@Q1f?oNY@Q1g?oNZ@Q1f?oNY@Q1h?nNX@R1h?oNW@Q1j?oNU@R1k?nNT@R1l?oNS@Q1n?POP@P1Q`0POn_OQ1Q`0POn_OP1S`0POm_Oo0S`0SOk_On0U`0ROj_On0W`0ROh_On0X`0SOg_Om0Z`0SOe_On0[`0ROd_On0\\`0SOc_Om0]`0TOb_Ol0[`0XOd_Oi0X`0[Oh_Od0U`0@j_O`0R`0Em_O;P`0Im_O:R`0Gk_O;U`0Ei_O=X`0Ce_O?[`0d1O00001O0000001O0000001O0000001O0000001O00001O0000001O0001O01O0001O100O1bMUf0"}, "label": "a person in a white top and black shorts"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table at a restaurant with a couple eating at it\"."}], "task": "refering", "answer_template": "The \"a table at a restaurant with a couple eating at it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 26, 27, 28, 29, 30, 31, 32, 44, 45, 61, 62, 78], "bbox": [0.5518958333333334, 0.003296875, 0.8955000000000001, 0.205109375], "iscrowd": 0, "area": 5277.591700000004, "rle": {"size": [640, 480], "counts": "deU52ic05L4L4K5L4L4K5L4O11O001O1O001O001O001O001O1O001O001O3M4L1OO10000O10000O100000000n\\O^1f`0l021O00001N10hMX2H80000000000000000000TOGQ^O9oa0Km]O5Sb0Oi]O1Wb0Oi]O1Wb00h]O0Xb01g]OOYb01g]OOYb02f]ONZb02f]ONZb03e]OM\\b02d]ON\\b03c]OM]b03c]OM]b04b]OL^b04b]OL^b05a]OK_b05a]OK_b06`]OJ`b06`]OJ`b0k000000000000000000000000000001O<D9G000000000001O000000000O1000000000000000000000000O10000O10000O01000O10000O100001O000O2O001O001O00001O001N10001O001O001O00001N3N2N3M2N3M2N3MbVo0"}, "label": "a table at a restaurant with a couple eating at it"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a man and a woman sitting at it\"."}], "task": "refering", "answer_template": "The \"a table with a man and a woman sitting at it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 26, 27, 28, 29, 30, 31, 32, 44, 45, 61, 62, 78], "bbox": [0.5518958333333334, 0.003296875, 0.8955000000000001, 0.205109375], "iscrowd": 0, "area": 5277.591700000004, "rle": {"size": [640, 480], "counts": "deU52ic05L4L4K5L4L4K5L4O11O001O1O001O001O001O001O1O001O001O3M4L1OO10000O10000O100000000n\\O^1f`0l021O00001N10hMX2H80000000000000000000TOGQ^O9oa0Km]O5Sb0Oi]O1Wb0Oi]O1Wb00h]O0Xb01g]OOYb01g]OOYb02f]ONZb02f]ONZb03e]OM\\b02d]ON\\b03c]OM]b03c]OM]b04b]OL^b04b]OL^b05a]OK_b05a]OK_b06`]OJ`b06`]OJ`b0k000000000000000000000000000001O<D9G000000000001O000000000O1000000000000000000000000O10000O10000O01000O10000O100001O000O2O001O001O00001O001N10001O001O001O00001N3N2N3M2N3M2N3MbVo0"}, "label": "a table with a man and a woman sitting at it"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza with pineapples\"."}], "task": "refering", "answer_template": "The \"the pizza with pineapples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 365], "bbox": [0.24325000000000002, 0.540546875, 0.8348333333333333, 0.921171875], "iscrowd": 0, "area": 54257.497550000015, "rle": {"size": [640, 480], "counts": "WcY2?mb0f0ZOe0[Of0C<M4M2N2M4M2N3L3N3L3N2N3L3N3M2M4M2M3N3M2M4M2N2M4M2M4M2N3L3N2N3M2O2M2O2M2O1N3N1N3N1O0O2O0O101N100O2O0O101N100O2O000O101N100O2O0O101N100O2O0O101N10001N100O2O0O101O0O101O000O101O000O2O00001N10001O000011N0001O00001O00001O00001N10001O0000000000000000O1000000O100000O10O100000000O1000000O100000000O100000O10O1000000O100000000O100000000O100000O0100000000O100000000O2O001O0O2O00001O0O2O001O0O2O00001O0O2O001O000O2O001O0O2O001O000O2N1O2O0O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2O0O2N3M2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3N2M2N3M3M3M2N3M3M2N3M3M2N3M3M3H7H9H8G8I8G9G8I8G9GZn`1"}, "label": "the pizza with pineapples"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with yellow seasonings in front of a bowl of food and a glass of water\"."}], "task": "refering", "answer_template": "The \"a pizza with yellow seasonings in front of a bowl of food and a glass of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 365], "bbox": [0.24325000000000002, 0.540546875, 0.8348333333333333, 0.921171875], "iscrowd": 0, "area": 54257.497550000015, "rle": {"size": [640, 480], "counts": "WcY2?mb0f0ZOe0[Of0C<M4M2N2M4M2N3L3N3L3N2N3L3N3M2M4M2M3N3M2M4M2N2M4M2M4M2N3L3N2N3M2O2M2O2M2O1N3N1N3N1O0O2O0O101N100O2O0O101N100O2O000O101N100O2O0O101N100O2O0O101N10001N100O2O0O101O0O101O000O101O000O2O00001N10001O000011N0001O00001O00001O00001N10001O0000000000000000O1000000O100000O10O100000000O1000000O100000000O100000O10O1000000O100000000O100000000O100000O0100000000O100000000O2O001O0O2O00001O0O2O001O0O2O00001O0O2O001O000O2O001O0O2O001O000O2N1O2O0O2N1O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2O0O2N3M2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3N2M2N3M3M3M2N3M3M2N3M3M2N3M3M3H7H9H8G8I8G9G8I8G9GZn`1"}, "label": "a pizza with yellow seasonings in front of a bowl of food and a glass of water"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with khaki shorts who is sitting\"."}], "task": "refering", "answer_template": "The \"the man with khaki shorts who is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 20, 21, 22, 23, 24, 25, 26, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 71, 72, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92], "bbox": [0.18791666666666668, 0.0, 0.6323541666666668, 0.23735937499999998], "iscrowd": 0, "area": 23088.97845, "rle": {"size": [640, 480], "counts": "\\Xh12lc07I7I7H7J7J6K5K4L5K5K5K4L5K5K4L5Ka1_N1O001O001O001O1O001O010O001O1O001O001O001O1O00000000O100000000000000O100000000000000O100000000000000O1000000000001O0O100000000000000O100000000000000O100000000000000O100000000hLc@m1]?TNh@f1X?ZNn@`1R?`NUAX1l>iNYAQ1g>oN_Ak0a>UO_Ak0a>VO^Aj0b>VO^Aj0b>VO_Ai0a>XO^Ah0b>XO^Ag0c>YO]Ag0c>ZO\\Af0d>ZO\\Af0d>ZO]Ae0c>\\O\\Ad0e>[O[Ae0e>[O[Ae0e>\\OZAc0g>]OYAc0g>]OZAb0f>_OYAa0g>_OYAa0g>_OYAa0g>@XA`0h>@XA`0h>@XA?i>BWA=i>CWA=i>CWA=i>DVA<j>DVA<j>DVA<j>EVA:j>FVA9k>GUA9k>HTA8l>HTA8l>HTA8m>HSA7n>HRA8P?FPA:Q?Fn@9U?Ek@;V?Dj@<X?Bh@>Y?Bg@=Z?Bf@>\\?@d@`0\\?Ac@?\\?Bd@>[?Ce@<\\?Ed@:[?Ge@9Z?Hf@8Y?Jf@6Z?Jf@6Y?Kg@5X?Mg@3Y?Mh@1X?0h@0W?2h@NW?3i@MW?3i@MV?5i@KV?4m@KR?3QAMo>1SANm>1UAOj>0XA0h>M[A3e>K^A4b>J`A6`>GcA9]>EeA;[>CgA<Z>BhA>X>_OkAa0U>]OnAb0R>\\OPBd0P>YOSBg0m=WOUBi0k=UOWBk0j=ROXBm0i=PO[Bo0e=oN]BQ1c=mN_BS1a=kNaBU1_=hNdBX1\\=fNfBZ1Q`0O0O100000002N3M4L3L5L3M4L3M4L3M4LQ`^3"}, "label": "the man with khaki shorts who is sitting"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 20, 21, 22, 23, 24, 25, 26, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 71, 72, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92], "bbox": [0.18791666666666668, 0.0, 0.6323541666666668, 0.23735937499999998], "iscrowd": 0, "area": 23088.97845, "rle": {"size": [640, 480], "counts": "\\Xh12lc07I7I7H7J7J6K5K4L5K5K5K4L5K5K4L5Ka1_N1O001O001O001O1O001O010O001O1O001O001O001O1O00000000O100000000000000O100000000000000O100000000000000O1000000000001O0O100000000000000O100000000000000O100000000000000O100000000hLc@m1]?TNh@f1X?ZNn@`1R?`NUAX1l>iNYAQ1g>oN_Ak0a>UO_Ak0a>VO^Aj0b>VO^Aj0b>VO_Ai0a>XO^Ah0b>XO^Ag0c>YO]Ag0c>ZO\\Af0d>ZO\\Af0d>ZO]Ae0c>\\O\\Ad0e>[O[Ae0e>[O[Ae0e>\\OZAc0g>]OYAc0g>]OZAb0f>_OYAa0g>_OYAa0g>_OYAa0g>@XA`0h>@XA`0h>@XA?i>BWA=i>CWA=i>CWA=i>DVA<j>DVA<j>DVA<j>EVA:j>FVA9k>GUA9k>HTA8l>HTA8l>HTA8m>HSA7n>HRA8P?FPA:Q?Fn@9U?Ek@;V?Dj@<X?Bh@>Y?Bg@=Z?Bf@>\\?@d@`0\\?Ac@?\\?Bd@>[?Ce@<\\?Ed@:[?Ge@9Z?Hf@8Y?Jf@6Z?Jf@6Y?Kg@5X?Mg@3Y?Mh@1X?0h@0W?2h@NW?3i@MW?3i@MV?5i@KV?4m@KR?3QAMo>1SANm>1UAOj>0XA0h>M[A3e>K^A4b>J`A6`>GcA9]>EeA;[>CgA<Z>BhA>X>_OkAa0U>]OnAb0R>\\OPBd0P>YOSBg0m=WOUBi0k=UOWBk0j=ROXBm0i=PO[Bo0e=oN]BQ1c=mN_BS1a=kNaBU1_=hNdBX1\\=fNfBZ1Q`0O0O100000002N3M4L3L5L3M4L3M4L3M4LQ`^3"}, "label": "a man in a blue shirt"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the man in blue is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the man in blue is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 36, 37, 54, 71, 88, 89, 90, 91, 106], "bbox": [0.14550000000000002, 0.04240625, 0.3611666666666667, 0.28504687500000003], "iscrowd": 0, "area": 3173.52105, "rle": {"size": [640, 480], "counts": "Qi[1m0Rc03N2M4L3M3N3L3M3O2O0O1O2N100O2N100O2N1O14M4K4L5L3L5K4M0O001O1O1N101O1O1N101O1N2O001O1N2O00V]O]OVb0b0e]OA^b0=^]OGdb08V]OLlb0c01O100O1O1O100O1OO1O2O0O100ON3M2N3L3N3N110000000000000001O0000000000000000001O000000000O1000001O00000000M3Gfgo5"}, "label": "the chair the man in blue is sitting in"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a man with a blue and striped shirt sitting on it\"."}], "task": "refering", "answer_template": "The \"a chair with a man with a blue and striped shirt sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 36, 37, 54, 71, 88, 89, 90, 91, 106], "bbox": [0.14550000000000002, 0.04240625, 0.3611666666666667, 0.28504687500000003], "iscrowd": 0, "area": 3173.52105, "rle": {"size": [640, 480], "counts": "Qi[1m0Rc03N2M4L3M3N3L3M3O2O0O1O2N100O2N100O2N1O14M4K4L5L3L5K4M0O001O1O1N101O1O1N101O1N2O001O1N2O00V]O]OVb0b0e]OA^b0=^]OGdb08V]OLlb0c01O100O1O1O100O1OO1O2O0O100ON3M2N3L3N3N110000000000000001O0000000000000000001O000000000O1000001O00000000M3Gfgo5"}, "label": "a chair with a man with a blue and striped shirt sitting on it"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first table of the picture on which pizzas are there\"."}], "task": "refering", "answer_template": "The \"the first table of the picture on which pizzas are there\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.006270833333333333, 0.27059375, 0.9976458333333333, 0.9858750000000001], "iscrowd": 0, "area": 204997.20330000002, "rle": {"size": [640, 480], "counts": "cV2<bc0h0XOh0hBLV5l0UHS1i7EbEZ2\\:f3O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1N2N2N2N2M3N2O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000001O0O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000002N3M3M3M3M3M3M3I7gNY1^Jb5iNW1iNW1iNW1iNW1iNV>"}, "label": "the first table of the picture on which pizzas are there"}]}
{"id": 80826, "image": "COCO_train2014_000000080826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza topped with ham and corn\"."}], "task": "refering", "answer_template": "The \"a pizza topped with ham and corn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.006270833333333333, 0.27059375, 0.9976458333333333, 0.9858750000000001], "iscrowd": 0, "area": 204997.20330000002, "rle": {"size": [640, 480], "counts": "cV2<bc0h0XOh0hBLV5l0UHS1i7EbEZ2\\:f3O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1N2N2N2N2M3N2O1000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000001O0O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000000000O100000000000000000000000000000000002N3M3M3M3M3M3M3I7gNY1^Jb5iNW1iNW1iNW1iNW1iNV>"}, "label": "a pizza topped with ham and corn"}]}
{"id": 113594, "image": "COCO_train2014_000000113594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a hat , t - shirt and jeans jumping on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man in a hat , t - shirt and jeans jumping on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 29, 30, 31, 32, 47, 48, 49, 50, 65, 66, 67, 84, 85], "bbox": [0.61512, 0.024714714714714717, 0.8186599999999999, 0.35954954954954954], "iscrowd": 0, "area": 5098.83405, "rle": {"size": [333, 500], "counts": "VWT35W:5K5K5K2N1O2N2N1O2N1O2N2M2O2N2N1nNiNmHY1R7kNiHV1V7nNfHT1Y7oNcHS1[7RO`HP1_7SO]HQ1_7TO\\HP1a7Q1L4M3L5L1N2O0O14L6I7J2N00000UO]HkNd7V1]HhNc7X1`HeN_7\\1dHaN\\7_1gH^NY7b1jHZNW7S1\\HUO`0ET7U1^HUOh0@f6Z1cHWOd0Ai6W1dHXOa0Ck6T1fHXO=Fl6R1hHXO:Im6n0jHYO7Ko6k0lHYO3NQ7g0nH[OO0R7e0PI\\OK1Q7f0VIXOG4o6g0[IUOD7P7c0]IVOA9R7`0_IVO]O<T7=`I3`6LaI4_6KcI5\\6JeI6[6JeI2_6NbINa62_IJe6JbHBi0`0h6O_HAj0<j62^HAh09n65[HCf04S78XHDf00U7<UHDf0LY7?RHFe0F]7c0oGGW9OjF2i9O0001O001O00010O001O00001O010O001O00010O0000M3N3L3M3Nbam0"}, "label": "a man in a hat , t - shirt and jeans jumping on a skateboard"}]}
{"id": 113594, "image": "COCO_train2014_000000113594.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the jumping skater boy\"."}], "task": "refering", "answer_template": "The \"the jumping skater boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 29, 30, 31, 32, 47, 48, 49, 50, 65, 66, 67, 84, 85], "bbox": [0.61512, 0.024714714714714717, 0.8186599999999999, 0.35954954954954954], "iscrowd": 0, "area": 5098.83405, "rle": {"size": [333, 500], "counts": "VWT35W:5K5K5K2N1O2N2N1O2N1O2N2M2O2N2N1nNiNmHY1R7kNiHV1V7nNfHT1Y7oNcHS1[7RO`HP1_7SO]HQ1_7TO\\HP1a7Q1L4M3L5L1N2O0O14L6I7J2N00000UO]HkNd7V1]HhNc7X1`HeN_7\\1dHaN\\7_1gH^NY7b1jHZNW7S1\\HUO`0ET7U1^HUOh0@f6Z1cHWOd0Ai6W1dHXOa0Ck6T1fHXO=Fl6R1hHXO:Im6n0jHYO7Ko6k0lHYO3NQ7g0nH[OO0R7e0PI\\OK1Q7f0VIXOG4o6g0[IUOD7P7c0]IVOA9R7`0_IVO]O<T7=`I3`6LaI4_6KcI5\\6JeI6[6JeI2_6NbINa62_IJe6JbHBi0`0h6O_HAj0<j62^HAh09n65[HCf04S78XHDf00U7<UHDf0LY7?RHFe0F]7c0oGGW9OjF2i9O0001O001O00010O001O00001O010O001O00010O0000M3N3L3M3Nbam0"}, "label": "the jumping skater boy"}]}
{"id": 269245, "image": "COCO_train2014_000000269245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog sitting in a motorcycle ' s sidecar\"."}], "task": "refering", "answer_template": "The \"a dog sitting in a motorcycle ' s sidecar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 74, 75, 76, 77, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 157, 158, 159, 162, 163, 164, 165, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.10760416666666667, 1.0, 0.9910625], "iscrowd": 0, "area": 185913.7394500001, "rle": {"size": [480, 640], "counts": "j49c4LjL0f6`0YLK`L0T79VL2VLOb73QLX1m3kNnKZ1P4hNaJoNlN`2c6bN_Jg2a5YM_Jf2b5ZM^Jf2b5[M]Je2c5[M\\Jf2d5ZM\\Jf2d5[M[Jd2f5\\MZJd2f5\\MZJd2f5]MXJe2g5[MYJh2d5XM\\Jk2a5UM_Jn2^5SMaJP3\\5PMcJT3Z5lLfJW3W5jLhJY3U5gLkJ\\3R5dLnJ_3o4bLPK`3n4`LRKb3l4^LTKd3j4]LUKe3i4[LWKg3g4ZLXKh3f4ZLXKh3f4[LWKg3g4[LWKg3f4\\LXKe3g4]LWKe3g4^LVKc3i4`LTKb3j4aLSK`3l4cLQK_3m4dLPK]3o4eLoJ]3o4dLPK]3o4eLoJ]3o4dLPK^3n4dLPK]3o4eLoJ]3o4dLPK^3n4dLPK]3o4dLPK^3m4fLPK\\3n4fLPK\\3n4gLoJZ3P5hLnJZ3P5iLmJX3Q5kLmJW3Q5lLkJV3T5lLjJV3T5\\3O2M3N1O2N2N2M2O2N2N2N1O1N101O001O0O2O001O001N101O0000hFmKc7R4]HoKc7Q4\\HPLd7P4[HRLd7n3[HSLe7m3[HSLf7l3YHVLf7i3ZHXLf7h3YHYLg7g3XH[Lg7e3YH[Lg7d3YH]Lg7c3XH_Lh7`3WHaLi7_3VHbLj7^3VHcLi7\\3WHeLi7[3VHfLj7Z3UHhLj7X3UHiLl7U3THlLl7T3THmLk7S3THnLl7R3SHPMl7P3SHQMm7n2SHSMm7m2SHTMl7l2SHUMm7k2RHWMm7h2THYMk7g2UHZMj7f2VH[Mi7e2VH^Mh7b2XH_Mg7`2ZHaMe7_2[HbMd7^2[HeMc7[2]HfMb7Y2_HhM`7X2`HiM_7W2`HlM^7T2bHmM]7R2cHPN\\7P2dHQN[7n1eHTNZ7l1fHVNX7j1gHXNX7g1iHZNV7f1iH\\NV7c1kH^NT7b1kH`NT7`1lHaNS7^1mHeNQ7[1oHfNP7Y1PIiNo6W1QIjNn6V1QIlNn6S1SInNl6R1SIPOl6o0UIRMQO[1i7c1UIRMVOY1c7f1WIQMXOY1_7e1YIRM\\OW1Z7g1YISM@T1V7i1YISMDS1R7i1ZIUMFQ1o6j1[ITMHR1k6j1]IUMIQ1i6j1]IUMLQ1f6j1]IVMNo0e6j1]IWM0o0b6j1]IXM2n0_6j1_IXM4m0]6k1^IYM6l0[6k1^IYM:j0X6l1^IYM=k0T6l1^IYM`0k0Q6k1_IYMc0k0m5m1_IXMf0k0j5l1`IXMi0k0g5m1_IXMl0k0c5m1aIWMo0l0_5m1aIWMR1k0]5m1aIWMU1l0X5n1bIUMX1n0U5l1cIVMY1n0S5m1cIUM[1o0Q5k1dIUM\\1Q1P5j1cIUM^1R1m4j1dISMa1T1j4h1fISM`1V1j4g1eISMb1W1g4f1gISMc1W1f4f1fISMd1Y1d4d1hISMe1Z1b4c1hITMe1Z1b4b1iITMf1[1_4b1jISMg1]1^4_1kIUMg1]1\\4_1lITMh1_1Z4]1nIUMh1`1X4[1oIUMi1b1V4Y1QJUMj1c1T4X1QJVMj1d1S4V1SJVMj1f1R4T1SJWMk1f1Q4R1SJYMl1g1Q4o0RJZMn1h1o3m0SJ\\Mm1i1P4j0QJ^Mo1j1o3g0RJ`Mo1j1o3e0PJbMQ2k1n3b0PJeMR2j1m3^1SLcNn3\\1QLfNn3Y1SLgNn3X1QLiNo3W1QLjNo3T1QLmNo3S1QLnNo3Q1QLoNo3P1QLQOo3o0QLROo3m0QLSOo3l0QLVOn3j0RL]NZNTNe5^3QLjMoNhNo4]3SLXMA[O]4[3RLiL4Kj3[3kNfLT1Y3nNfLS1X3nNiLQ1V3QOiLo0W3QOjLo0T3SOkLm0U3SOkLm0T3UOlLj0T3VOlLk0R3WOnLh0R3XOnLh0R3YOmLg0R3ZOoLe0Q3\\OnLe0P3\\OQMc0o2^OPMb0o2_ORM`0n2AQM?o2AQM?P3APM>R3@PM>Q3BoL>Q3APM>Q3AQM=P3CPM<Q3CQM;P3EPM:P3FRM8n2HSM7n2IRM6n2JTM5k2LUM3k2MWM1j2NWM1i20XMNh23ZMJg25]MGc2:_MCa2>aM^O`2b0cM[O^2e0dMXO\\2h0gMUOY2l0iMQOW2P1kMmNV2S1jMlNV2U1jMjNV2W1iMiNW2X1iMgNW2Z1hMfNY2Z1gMeN[2Z1dMfN_2W1bMhN`2W1_MiNc2V1]MiNd2V1\\MjNe2V1[MiNf2V1ZMjNg2V1YMiNh2V1XMjNi2U1XMjNi2V1VMjNk2U1VMjNk2V1TMjNm2U1TMjNn2U1QMkNQ3T1nLlNS3S1nLlNT3S1kLmNh2PL[LR5m0mNf2VL\\Ll4m0nNd2]L]Ld4o0nNa2dL_L\\4P1PO^2kLaLS4Q1QO[2RMdLk3P1SOY2YMfLb3Q1TOW2`MgLY3R1WOV2eMgLP3U1ZOS2jMhLi2U1]OQ2oMiLb2W1^Oo1TNjLZ2X1Bl1YNlLR2Y1Dj1^NmLj1Z1Hh1bNnLc1[1Jg1eNoL]1[1Ne1iNPMV1\\10c1oNPMm0^14a1TOPMe0`16`1YOoLUOj2b16^OoLgNR3j1NX18hNHV1;iNDW1=iNBW1`0hN@X1`0hN_OX1c0gN\\OY1e0gN[OX1g0gNXOZ1h0fNWOZ1k0eNUOZ1l0fNSO[1n0dNQO\\1P1dNPO[1Q1eNnN[1T1dNkN]1U1cNkN^1U1aNjNc1S1]NmNe1R1ZNmNj1P1VNoNn1o0QNQOR2l0nMSOV2k0iMUOZ2h0fMWO]2h0bMXOa2e0_MZOe2d0ZM[Oi2c0WM]Ol2`0TM_Oo2`0PM@S3=mLBV3<jLCY3<gLC\\3:dLE_39aLGb37]LHf36ZLJh34XLKk34TLKn34QLMP42PLMS42lKNU41jKOX40hK0Z4OeK0]4OcK0_4O`K2b4M]K2e4M[K3f4LYK4j4KUK5l4JTK5n4JQK6R5ImJ7T5HlJ7W5GhJ:Y5FfJ9]5EcJ;^5DaJ=a5B^J=d5B\\J>e5@\\J?g5^OYJc0h5[OYJd0j5YOWJg0j5WOVJi0m5SOUJm0l5QOUJo0m5nNSJR1S6hNnIX1W6aNjI`1[6ZNfIf1_6TNaIl1e6nM[IS2i90O1O10O01O1000002N2N3M2N3M2O1N3N1N3N2M3N2M3N2M3M4M3L4L3M4M3L4L4L3N3L4L4L4M3L2O2M2O1N2O0O2O001N100O2O0O2OO01O1O001O1O001O0O2O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O1O1O1O1O00100O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O001O1O001O1O001O001O00000000001O000000000O101O0000000000000000000000001O000000kN`HjJ`7U5gHeJY7Y5oHaJQ7^5Y1N2O0O2O1N2O1N2O1N20000O10000O10lH_Ji4a5UKdJh4\\5WKgJg4X5XKiJi4W5VKjJj4V5VKiJk4V5UKjJl4V5TKiJm4W5SKiJm4W5SKhJn4W5SKhJm4Y5SKfJn4Z5RKeJo4[5QKeJo4[5QKfJn4[5QKeJo4[5QKeJo4[5QKfJn4Z5RKfJn4[5PKgJo4Y5QKgJo4Y5QKgJo4Y5QKhJn4Y5PKhJP5X5PKiJo4W5PKjJP5V5PKjJP5W5oJjJP5V5oJkJQ5U5nJmJQ5S5nJnJR5S5kJoJU5Q5jJQKT5P5kJQKU5P5hJSKW5m4hJTKX5l4dJXK\\5i4^J]Ka5c4[JaKe5`4VJeKi5[4SJiKl5X4QJlKn5U4nInKR6R4kIQLT6_6O001O0O2O000O2O001O0O2O001N2O001N2O001N2O001O1N101O0O2O1O0O2O1O2M2N2O1N2O2M2N2O1lLXGMj8VOdH8\\N"}, "label": "a dog sitting in a motorcycle ' s sidecar"}]}
{"id": 269245, "image": "COCO_train2014_000000269245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vehicle and a dog in front of another vehicle\"."}], "task": "refering", "answer_template": "The \"a vehicle and a dog in front of another vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 57, 58, 59, 60, 74, 75, 76, 77, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 157, 158, 159, 162, 163, 164, 165, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.10760416666666667, 1.0, 0.9910625], "iscrowd": 0, "area": 185913.7394500001, "rle": {"size": [480, 640], "counts": "j49c4LjL0f6`0YLK`L0T79VL2VLOb73QLX1m3kNnKZ1P4hNaJoNlN`2c6bN_Jg2a5YM_Jf2b5ZM^Jf2b5[M]Je2c5[M\\Jf2d5ZM\\Jf2d5[M[Jd2f5\\MZJd2f5\\MZJd2f5]MXJe2g5[MYJh2d5XM\\Jk2a5UM_Jn2^5SMaJP3\\5PMcJT3Z5lLfJW3W5jLhJY3U5gLkJ\\3R5dLnJ_3o4bLPK`3n4`LRKb3l4^LTKd3j4]LUKe3i4[LWKg3g4ZLXKh3f4ZLXKh3f4[LWKg3g4[LWKg3f4\\LXKe3g4]LWKe3g4^LVKc3i4`LTKb3j4aLSK`3l4cLQK_3m4dLPK]3o4eLoJ]3o4dLPK]3o4eLoJ]3o4dLPK^3n4dLPK]3o4eLoJ]3o4dLPK^3n4dLPK]3o4dLPK^3m4fLPK\\3n4fLPK\\3n4gLoJZ3P5hLnJZ3P5iLmJX3Q5kLmJW3Q5lLkJV3T5lLjJV3T5\\3O2M3N1O2N2N2M2O2N2N2N1O1N101O001O0O2O001O001N101O0000hFmKc7R4]HoKc7Q4\\HPLd7P4[HRLd7n3[HSLe7m3[HSLf7l3YHVLf7i3ZHXLf7h3YHYLg7g3XH[Lg7e3YH[Lg7d3YH]Lg7c3XH_Lh7`3WHaLi7_3VHbLj7^3VHcLi7\\3WHeLi7[3VHfLj7Z3UHhLj7X3UHiLl7U3THlLl7T3THmLk7S3THnLl7R3SHPMl7P3SHQMm7n2SHSMm7m2SHTMl7l2SHUMm7k2RHWMm7h2THYMk7g2UHZMj7f2VH[Mi7e2VH^Mh7b2XH_Mg7`2ZHaMe7_2[HbMd7^2[HeMc7[2]HfMb7Y2_HhM`7X2`HiM_7W2`HlM^7T2bHmM]7R2cHPN\\7P2dHQN[7n1eHTNZ7l1fHVNX7j1gHXNX7g1iHZNV7f1iH\\NV7c1kH^NT7b1kH`NT7`1lHaNS7^1mHeNQ7[1oHfNP7Y1PIiNo6W1QIjNn6V1QIlNn6S1SInNl6R1SIPOl6o0UIRMQO[1i7c1UIRMVOY1c7f1WIQMXOY1_7e1YIRM\\OW1Z7g1YISM@T1V7i1YISMDS1R7i1ZIUMFQ1o6j1[ITMHR1k6j1]IUMIQ1i6j1]IUMLQ1f6j1]IVMNo0e6j1]IWM0o0b6j1]IXM2n0_6j1_IXM4m0]6k1^IYM6l0[6k1^IYM:j0X6l1^IYM=k0T6l1^IYM`0k0Q6k1_IYMc0k0m5m1_IXMf0k0j5l1`IXMi0k0g5m1_IXMl0k0c5m1aIWMo0l0_5m1aIWMR1k0]5m1aIWMU1l0X5n1bIUMX1n0U5l1cIVMY1n0S5m1cIUM[1o0Q5k1dIUM\\1Q1P5j1cIUM^1R1m4j1dISMa1T1j4h1fISM`1V1j4g1eISMb1W1g4f1gISMc1W1f4f1fISMd1Y1d4d1hISMe1Z1b4c1hITMe1Z1b4b1iITMf1[1_4b1jISMg1]1^4_1kIUMg1]1\\4_1lITMh1_1Z4]1nIUMh1`1X4[1oIUMi1b1V4Y1QJUMj1c1T4X1QJVMj1d1S4V1SJVMj1f1R4T1SJWMk1f1Q4R1SJYMl1g1Q4o0RJZMn1h1o3m0SJ\\Mm1i1P4j0QJ^Mo1j1o3g0RJ`Mo1j1o3e0PJbMQ2k1n3b0PJeMR2j1m3^1SLcNn3\\1QLfNn3Y1SLgNn3X1QLiNo3W1QLjNo3T1QLmNo3S1QLnNo3Q1QLoNo3P1QLQOo3o0QLROo3m0QLSOo3l0QLVOn3j0RL]NZNTNe5^3QLjMoNhNo4]3SLXMA[O]4[3RLiL4Kj3[3kNfLT1Y3nNfLS1X3nNiLQ1V3QOiLo0W3QOjLo0T3SOkLm0U3SOkLm0T3UOlLj0T3VOlLk0R3WOnLh0R3XOnLh0R3YOmLg0R3ZOoLe0Q3\\OnLe0P3\\OQMc0o2^OPMb0o2_ORM`0n2AQM?o2AQM?P3APM>R3@PM>Q3BoL>Q3APM>Q3AQM=P3CPM<Q3CQM;P3EPM:P3FRM8n2HSM7n2IRM6n2JTM5k2LUM3k2MWM1j2NWM1i20XMNh23ZMJg25]MGc2:_MCa2>aM^O`2b0cM[O^2e0dMXO\\2h0gMUOY2l0iMQOW2P1kMmNV2S1jMlNV2U1jMjNV2W1iMiNW2X1iMgNW2Z1hMfNY2Z1gMeN[2Z1dMfN_2W1bMhN`2W1_MiNc2V1]MiNd2V1\\MjNe2V1[MiNf2V1ZMjNg2V1YMiNh2V1XMjNi2U1XMjNi2V1VMjNk2U1VMjNk2V1TMjNm2U1TMjNn2U1QMkNQ3T1nLlNS3S1nLlNT3S1kLmNh2PL[LR5m0mNf2VL\\Ll4m0nNd2]L]Ld4o0nNa2dL_L\\4P1PO^2kLaLS4Q1QO[2RMdLk3P1SOY2YMfLb3Q1TOW2`MgLY3R1WOV2eMgLP3U1ZOS2jMhLi2U1]OQ2oMiLb2W1^Oo1TNjLZ2X1Bl1YNlLR2Y1Dj1^NmLj1Z1Hh1bNnLc1[1Jg1eNoL]1[1Ne1iNPMV1\\10c1oNPMm0^14a1TOPMe0`16`1YOoLUOj2b16^OoLgNR3j1NX18hNHV1;iNDW1=iNBW1`0hN@X1`0hN_OX1c0gN\\OY1e0gN[OX1g0gNXOZ1h0fNWOZ1k0eNUOZ1l0fNSO[1n0dNQO\\1P1dNPO[1Q1eNnN[1T1dNkN]1U1cNkN^1U1aNjNc1S1]NmNe1R1ZNmNj1P1VNoNn1o0QNQOR2l0nMSOV2k0iMUOZ2h0fMWO]2h0bMXOa2e0_MZOe2d0ZM[Oi2c0WM]Ol2`0TM_Oo2`0PM@S3=mLBV3<jLCY3<gLC\\3:dLE_39aLGb37]LHf36ZLJh34XLKk34TLKn34QLMP42PLMS42lKNU41jKOX40hK0Z4OeK0]4OcK0_4O`K2b4M]K2e4M[K3f4LYK4j4KUK5l4JTK5n4JQK6R5ImJ7T5HlJ7W5GhJ:Y5FfJ9]5EcJ;^5DaJ=a5B^J=d5B\\J>e5@\\J?g5^OYJc0h5[OYJd0j5YOWJg0j5WOVJi0m5SOUJm0l5QOUJo0m5nNSJR1S6hNnIX1W6aNjI`1[6ZNfIf1_6TNaIl1e6nM[IS2i90O1O10O01O1000002N2N3M2N3M2O1N3N1N3N2M3N2M3N2M3M4M3L4L3M4M3L4L4L3N3L4L4L4M3L2O2M2O1N2O0O2O001N100O2O0O2OO01O1O001O1O001O0O2O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O1O1O1O1O00100O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O001O1O001O1O001O001O00000000001O000000000O101O0000000000000000000000001O000000kN`HjJ`7U5gHeJY7Y5oHaJQ7^5Y1N2O0O2O1N2O1N2O1N20000O10000O10lH_Ji4a5UKdJh4\\5WKgJg4X5XKiJi4W5VKjJj4V5VKiJk4V5UKjJl4V5TKiJm4W5SKiJm4W5SKhJn4W5SKhJm4Y5SKfJn4Z5RKeJo4[5QKeJo4[5QKfJn4[5QKeJo4[5QKeJo4[5QKfJn4Z5RKfJn4[5PKgJo4Y5QKgJo4Y5QKgJo4Y5QKhJn4Y5PKhJP5X5PKiJo4W5PKjJP5V5PKjJP5W5oJjJP5V5oJkJQ5U5nJmJQ5S5nJnJR5S5kJoJU5Q5jJQKT5P5kJQKU5P5hJSKW5m4hJTKX5l4dJXK\\5i4^J]Ka5c4[JaKe5`4VJeKi5[4SJiKl5X4QJlKn5U4nInKR6R4kIQLT6_6O001O0O2O000O2O001O0O2O001N2O001N2O001N2O001O1N101O0O2O1O0O2O1O2M2N2O1N2O2M2N2O1lLXGMj8VOdH8\\N"}, "label": "a vehicle and a dog in front of another vehicle"}]}
{"id": 269245, "image": "COCO_train2014_000000269245.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle colored in camoflage\"."}], "task": "refering", "answer_template": "The \"a motorcycle colored in camoflage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 227, 228, 229, 252], "bbox": [0.6404687499999999, 0.0, 1.0, 0.599], "iscrowd": 0, "area": 36626.504, "rle": {"size": [480, 640], "counts": "`iP62m>2N2N2O1N2N2N2O1N1O2O1N2N2O1N2N2N2O1N1O2O1N2N2O1N2N2N2O1N2N101N2N2O1N2N2M3N2M3N2N1N3N2M3N2M3N2M3N2N2M3N101O1N2O1O1O1O1O1N2O001O1O1O1O1N2O1O1O1O1O0O2O1O1O1O1O1N2O1O1O1O001N2O1O001O001N101O00001O0O2O001O001O0O101O001O001N10001O001O001N101O00001O0O2O001O001O0O101O001N101O0O2O0O101O0O2O001N101N10001N101fMaJPK`5P5fJjJ[5U5kJeJU5[5PK_JR5`5TKZJl4f5ZKTJf4l5`KmIb4R6cKiI]4W6iKbIY4]6mK]IS4c6SLWIn3h6WLRIj3n6U2O00001O001O001O001O00001O0O2O001O001O00001O001O001O00001O001O001O001O00001O001O001N101O00001O001O001O001O00001O0101N2O1N101N2NV2kMX4gKUI"}, "label": "a motorcycle colored in camoflage"}]}
{"id": 572353, "image": "COCO_train2014_000000572353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark colored horse standing next to a wooden fence\"."}], "task": "refering", "answer_template": "The \"a dark colored horse standing next to a wooden fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 190, 191, 192, 207, 208, 209, 210, 214, 215, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 323, 324, 346, 347, 369, 370], "bbox": [0.0, 0.23010330578512397, 0.3854375, 0.9892561983471074], "iscrowd": 0, "area": 34115.93800000001, "rle": {"size": [484, 640], "counts": "a8V1a=]1dN<WNi10O10O10O010000O010O10O06K4L5J6K5J5L5K4K5L5J5L4L4K5L4K5L4L5J9H7IP1oN2O1N2O00eLXJ[Mf5e2eJRMZ5l2QKlLn4S3\\KeLb4[3dK`L\\4_3iK^LV4`3oK^Lo3b3VLeKZNWO_5S5[L^KaNZOR5X5bLUKhN\\Of4]5gLoJ_4P5fKfJ_4[5e2000O100O10000O10000O100000<C>aHPKP4V8L001N10001O0O101O00001N101OK5aN^1SOn0C=D<C<G:F:F:J5K6J6J6K4K6J6K5L4N1O2M3N2N1N3N2N2M2O2N2O1O1O001O1O1O001O1O100O001O1O1O10O0100O1O010O100O10O01O100O100O00100O100O1O10O0100O100001O0O2O001O001O0O2O0O2O0O101N2N2O1N2O1N2O2L3M3M3M3M3M3M2O2M3N2M3M2O2M3N2M3M22O001O001O0010OiK[En3l:2N2M3NUL]EX3a:^LmE`3o:N1N2N2M3M2M4M3M2N3M3M3M2N3M3M3M2N2N2N1O2N2N2N1O2N1O1N2O1N2N2O1N2N2O1N3M2O3L4M4K4Lf`i5"}, "label": "a dark colored horse standing next to a wooden fence"}]}
{"id": 572353, "image": "COCO_train2014_000000572353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a moma horse looking at her baby\"."}], "task": "refering", "answer_template": "The \"a moma horse looking at her baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 190, 191, 192, 207, 208, 209, 210, 214, 215, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 323, 324, 346, 347, 369, 370], "bbox": [0.0, 0.23010330578512397, 0.3854375, 0.9892561983471074], "iscrowd": 0, "area": 34115.93800000001, "rle": {"size": [484, 640], "counts": "a8V1a=]1dN<WNi10O10O10O010000O010O10O06K4L5J6K5J5L5K4K5L5J5L4L4K5L4K5L4L5J9H7IP1oN2O1N2O00eLXJ[Mf5e2eJRMZ5l2QKlLn4S3\\KeLb4[3dK`L\\4_3iK^LV4`3oK^Lo3b3VLeKZNWO_5S5[L^KaNZOR5X5bLUKhN\\Of4]5gLoJ_4P5fKfJ_4[5e2000O100O10000O10000O100000<C>aHPKP4V8L001N10001O0O101O00001N101OK5aN^1SOn0C=D<C<G:F:F:J5K6J6J6K4K6J6K5L4N1O2M3N2N1N3N2N2M2O2N2O1O1O001O1O1O001O1O100O001O1O1O10O0100O1O010O100O10O01O100O100O00100O100O1O10O0100O100001O0O2O001O001O0O2O0O2O0O101N2N2O1N2O1N2O2L3M3M3M3M3M3M2O2M3N2M3M2O2M3N2M3M22O001O001O0010OiK[En3l:2N2M3NUL]EX3a:^LmE`3o:N1N2N2M3M2M4M3M2N3M3M3M2N3M3M3M2N2N2N1O2N2N2N1O2N1O1N2O1N2N2O1N2N2O1N3M2O3L4M4K4Lf`i5"}, "label": "a moma horse looking at her baby"}]}
{"id": 572353, "image": "COCO_train2014_000000572353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with three legs visible facing the camera near another slightly taller horse\"."}], "task": "refering", "answer_template": "The \"a horse with three legs visible facing the camera near another slightly taller horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 241, 242, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 357, 358, 359, 381, 382], "bbox": [0.5160625, 0.36824380165289256, 0.6945781249999999, 0.9605165289256198], "iscrowd": 0, "area": 19691.03550000001, "rle": {"size": [484, 640], "counts": "^Xl42n>7I7I6J7I7H7J7I7I7I6J8H9F9H8H9G8[N[L_Gm3]8ZLXGn3e8YLoFP4l8XLgFQ4U9S1L5K4M3L5XIdI^4`6WKgIi4]6lJjIT5Z6aJnI^5W6VJPJk5S6jITJW6]70O100ZLdIbN]6]1iI]NX6a1nI[NR6c1UJWNl5g1ZJTNg5j1_JRNa5l1fJnMZ5Q2lJjMT5U2SKfMl4Y2ZKbMf4]2aK^M_4a2fKWM]4j2n31O1O1O7Jc0\\Od0\\Od0\\O`0@7I7I7J5J7I7I1O0000010O0O1N2O\\MnIXLQ6f3hJdKW5Y4]KWKa4GbIZ4]2dKo32eIW4k2[K_3>gIY4Y9hKhF_4P9aKQGf4g8ZKZGn4\\8SKeGT5S8lJnG[5k7eJUHb5i7XJYHo5c7lI^HZ6a7`I`Ha6Q8001O00001O0100K5J6K5J7J5fNZ1\\Od0QOjDkMZ;R2iDkMZ;S2hDiM\\;T2gDiM\\;U2fDhM];U2fDgM^;W2dDfM_;W2dDeM`;Y2g0N3L<E;D<E;D;F;D[lk2"}, "label": "a horse with three legs visible facing the camera near another slightly taller horse"}]}
{"id": 572353, "image": "COCO_train2014_000000572353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse standing with his face directed at the camera\"."}], "task": "refering", "answer_template": "The \"a horse standing with his face directed at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 241, 242, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 357, 358, 359, 381, 382], "bbox": [0.5160625, 0.36824380165289256, 0.6945781249999999, 0.9605165289256198], "iscrowd": 0, "area": 19691.03550000001, "rle": {"size": [484, 640], "counts": "^Xl42n>7I7I6J7I7H7J7I7I7I6J8H9F9H8H9G8[N[L_Gm3]8ZLXGn3e8YLoFP4l8XLgFQ4U9S1L5K4M3L5XIdI^4`6WKgIi4]6lJjIT5Z6aJnI^5W6VJPJk5S6jITJW6]70O100ZLdIbN]6]1iI]NX6a1nI[NR6c1UJWNl5g1ZJTNg5j1_JRNa5l1fJnMZ5Q2lJjMT5U2SKfMl4Y2ZKbMf4]2aK^M_4a2fKWM]4j2n31O1O1O7Jc0\\Od0\\Od0\\O`0@7I7I7J5J7I7I1O0000010O0O1N2O\\MnIXLQ6f3hJdKW5Y4]KWKa4GbIZ4]2dKo32eIW4k2[K_3>gIY4Y9hKhF_4P9aKQGf4g8ZKZGn4\\8SKeGT5S8lJnG[5k7eJUHb5i7XJYHo5c7lI^HZ6a7`I`Ha6Q8001O00001O0100K5J6K5J7J5fNZ1\\Od0QOjDkMZ;R2iDkMZ;S2hDiM\\;T2gDiM\\;U2fDhM];U2fDgM^;W2dDfM_;W2dDeM`;Y2g0N3L<E;D<E;D;F;D[lk2"}, "label": "a horse standing with his face directed at the camera"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with curly brown hair sitting down beside man wearing a hat\"."}], "task": "refering", "answer_template": "The \"woman with curly brown hair sitting down beside man wearing a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 322, 323], "bbox": [0.0, 0.245093896713615, 0.38590625, 0.9889201877934272], "iscrowd": 0, "area": 42564.15855, "rle": {"size": [426, 640], "counts": "n;6n<=\\CIa;W1F6J6]OTNbER2\\:RN\\ET2a:<0O01O100O1O100001N100000000000000O1O1O1O1O1N2O1O0O2O001N1O2N1010O010O1O010O10O010O10O1QKUN_Nj1j0QOSOo0k0WOPOj0o0ZOnNe0Q1@lN`0R1FiN;V1KeN4Z11bN0\\15aNK^19_NF`1?\\NBc1b0ZN]Of1g0WNYOh1j0UNWOk1l0RNSOn1Q1oMoNP2U1lMlNS2X1jMgNV2]1fMdNY2a1bM`N]2f1^MYNb2l1YMUNf2Q2UMoMk2T2QMlMo2W2oLiMQ3X2mLiMR3Z2lLeMU3]2hLdMX3^2eLcMZ3_2eL`M\\3b2aL_M^3d2XLdMh3^2jKmMW4c5O100O01000O010O1000O100O1O1O1O1O1O100O1O1O2N10000O100O10000O100O100O10000O100O101O0O101N100O2O000O2O0O101O000O101O001N10]J]H`5b7`J^H`5b7_J_Hb5`7^JaHa5^7_JcHb5\\7^JdHb5\\7]JeHd5Z7\\JfHd5Y7\\JhHe5W7[JhHg5W7YJhHh5\\72JVJiHk5Z72N1O2N101N1O2N1O2N1O2N2N2N2M3M3M3M3M3M3N2N2N2O1N2N2N2O1N2N2N101N1O2N101N1O2N101N1O2N2N100O1O1O100O1O1O100O1O1O2O1O0O2O1O1N2O1O2N1N3N2N2M3XLTGn2R9jLYGm2j8oL[Gl2m9E6J3L5L3M3L4M2N2N1N3N2N3M9F8I5J6G8GVWS5"}, "label": "woman with curly brown hair sitting down beside man wearing a hat"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with curly hair sitting next to a man holding a wii controller\"."}], "task": "refering", "answer_template": "The \"a woman with curly hair sitting next to a man holding a wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 322, 323], "bbox": [0.0, 0.245093896713615, 0.38590625, 0.9889201877934272], "iscrowd": 0, "area": 42564.15855, "rle": {"size": [426, 640], "counts": "n;6n<=\\CIa;W1F6J6]OTNbER2\\:RN\\ET2a:<0O01O100O1O100001N100000000000000O1O1O1O1O1N2O1O0O2O001N1O2N1010O010O1O010O10O010O10O1QKUN_Nj1j0QOSOo0k0WOPOj0o0ZOnNe0Q1@lN`0R1FiN;V1KeN4Z11bN0\\15aNK^19_NF`1?\\NBc1b0ZN]Of1g0WNYOh1j0UNWOk1l0RNSOn1Q1oMoNP2U1lMlNS2X1jMgNV2]1fMdNY2a1bM`N]2f1^MYNb2l1YMUNf2Q2UMoMk2T2QMlMo2W2oLiMQ3X2mLiMR3Z2lLeMU3]2hLdMX3^2eLcMZ3_2eL`M\\3b2aL_M^3d2XLdMh3^2jKmMW4c5O100O01000O010O1000O100O1O1O1O1O1O100O1O1O2N10000O100O10000O100O100O10000O100O101O0O101N100O2O000O2O0O101O000O101O001N10]J]H`5b7`J^H`5b7_J_Hb5`7^JaHa5^7_JcHb5\\7^JdHb5\\7]JeHd5Z7\\JfHd5Y7\\JhHe5W7[JhHg5W7YJhHh5\\72JVJiHk5Z72N1O2N101N1O2N1O2N1O2N2N2N2M3M3M3M3M3M3N2N2N2O1N2N2N2O1N2N2N101N1O2N101N1O2N101N1O2N2N100O1O1O100O1O1O100O1O1O2O1O0O2O1O1N2O1O2N1N3N2N2M3XLTGn2R9jLYGm2j8oL[Gl2m9E6J3L5L3M3L4M2N2N1N3N2N3M9F8I5J6G8GVWS5"}, "label": "a woman with curly hair sitting next to a man holding a wii controller"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female in jeans and a red shirt walking and eating cheese - its\"."}], "task": "refering", "answer_template": "The \"a female in jeans and a red shirt walking and eating cheese - its\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209], "bbox": [0.0, 0.0022535211267605635, 0.180984375, 0.6471830985915493], "iscrowd": 0, "area": 17507.2611, "rle": {"size": [426, 640], "counts": "_5W1i9^2bM]2WOi0M4L2N2M4M2N2N2N2N3M2N2N4L4L3M4L4L4K3N00000000000000001O0000000O10O10000000000O10000000000YLeK^N[4S1[LgNe3T1fLeN[3Z1mL_NS3_1UMZNl2e1[MUNe2i1cMPN^2o1iMkMW2S2QNfMR2W2RNfMo1X2TNeMo1W2VNfMl1W2WNfMk1W2ZNfMh1W2\\NfMf1W2^NeMd1X2bNdM`1Y2eNbM]1[2iNaMY1\\2lN_MW1^2nN^MS1_2SO\\MP1a2o4M2O2N2N1O2N2M2O2N1O2N2N1N3N2N1O2N2N1O2M3N1O2N2N1O2N1N3N2N1O2N2N1N3N2N1O2N2N1O2M2O2N2N1O2Nfli6"}, "label": "a female in jeans and a red shirt walking and eating cheese - its"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a red shirt holding a box of cheez - its\"."}], "task": "refering", "answer_template": "The \"a woman in a red shirt holding a box of cheez - its\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209], "bbox": [0.0, 0.0022535211267605635, 0.180984375, 0.6471830985915493], "iscrowd": 0, "area": 17507.2611, "rle": {"size": [426, 640], "counts": "_5W1i9^2bM]2WOi0M4L2N2M4M2N2N2N2N3M2N2N4L4L3M4L4L4K3N00000000000000001O0000000O10O10000000000O10000000000YLeK^N[4S1[LgNe3T1fLeN[3Z1mL_NS3_1UMZNl2e1[MUNe2i1cMPN^2o1iMkMW2S2QNfMR2W2RNfMo1X2TNeMo1W2VNfMl1W2WNfMk1W2ZNfMh1W2\\NfMf1W2^NeMd1X2bNdM`1Y2eNbM]1[2iNaMY1\\2lN_MW1^2nN^MS1_2SO\\MP1a2o4M2O2N2N1O2N2M2O2N1O2N2N1N3N2N1O2N2N1O2M3N1O2N2N1O2N1N3N2N1O2N2N1N3N2N1O2N2N1O2M2O2N2N1O2Nfli6"}, "label": "a woman in a red shirt holding a box of cheez - its"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brunette girl points at a man with a wii controller in his hand\"."}], "task": "refering", "answer_template": "The \"a brunette girl points at a man with a wii controller in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.7075, 0.13190140845070422, 1.0, 1.0], "iscrowd": 0, "area": 49466.80025, "rle": {"size": [426, 640], "counts": "ebl5>R<K^D8^;M]D7_;M^D6^;M`D6[;OaD5[;P1L3M4L4K6K5K5K6J5L4L5K4L3M4M2M2N1O1O1O1mFRLi8o3VGVLe8k3[GXLb8h3^G\\L]8e3bG`LY8a3gGcLT8^3lGeLP8\\3oGiLl7X3SHmLh7T3WHPMe7_4L4K6J5L4L4L4ZIbIX6k6M3M3M3M3M3M4L3M3M3M3M3M3M3M3mKjGl2Y8iLXHn2l7fLdHR3S9bMPFP1U;001O0000000000000000001O000000000000000010O0000000001O000000001O00000000001O00000000001O000000001O00101N3M3M3M2N2N2N2N2N3M2N2nFRMd7R3WHRMf7Q3VHSMg7Q3THRMj7U3nGoLo7^3bGgL[8W4M3M3M1O000000000000000001O00000000000000000000000001O0000000000001O0000000000001O000001O000000000000000000000000000"}, "label": "a brunette girl points at a man with a wii controller in his hand"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with brown hair and bangs with her mouth open pointing to a man\"."}], "task": "refering", "answer_template": "The \"a girl with brown hair and bangs with her mouth open pointing to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 85, 86, 87, 88, 89, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.7075, 0.13190140845070422, 1.0, 1.0], "iscrowd": 0, "area": 49466.80025, "rle": {"size": [426, 640], "counts": "ebl5>R<K^D8^;M]D7_;M^D6^;M`D6[;OaD5[;P1L3M4L4K6K5K5K6J5L4L5K4L3M4M2M2N1O1O1O1mFRLi8o3VGVLe8k3[GXLb8h3^G\\L]8e3bG`LY8a3gGcLT8^3lGeLP8\\3oGiLl7X3SHmLh7T3WHPMe7_4L4K6J5L4L4L4ZIbIX6k6M3M3M3M3M3M4L3M3M3M3M3M3M3M3mKjGl2Y8iLXHn2l7fLdHR3S9bMPFP1U;001O0000000000000000001O000000000000000010O0000000001O000000001O00000000001O00000000001O000000001O00101N3M3M3M2N2N2N2N2N3M2N2nFRMd7R3WHRMf7Q3VHSMg7Q3THRMj7U3nGoLo7^3bGgL[8W4M3M3M1O000000000000000001O00000000000000000000000001O0000000000001O0000000000001O000001O000000000000000000000000000"}, "label": "a girl with brown hair and bangs with her mouth open pointing to a man"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a couch with a wii remote\"."}], "task": "refering", "answer_template": "The \"a man sitting on a couch with a wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.11496875, 0.00948356807511737, 0.801796875, 1.0], "iscrowd": 0, "area": 85946.17104999999, "rle": {"size": [426, 640], "counts": "^`o01X=100O10O01O100O100O1O10O0100O100O10000O10O1000O10000O1000O10O10000O100PDEQ;;oDGo:8QEJm:7SEJl:6SELl:4TEMj:4VEMi:3VEOi:1WE0h:OYE2e:OZE2f:NZE1g:OYE1f:0YE1g:NZE1g:OXE2h:NXE1h:0XE0h:0WE1i:OWE0i:0XE0h:0WELn:4REFS;;mDES;;mDDT;<lDDT;;lDET;<_DBO2b;<^DC00c;=\\DD1Oc;=[DF1Mc;=\\DG1Kd;>ZDH1Mc;;[DI2Na;9\\DJ3O_;7]DK40\\;6_DL40[;3aDN4O[;3`DO4O\\;2_D05N\\;2^D16M[;3^D26K\\;2^D46J\\;f0dDYO];g0bDZO^;f0bDZO^;g0`DZO_;e0bD\\O^;c0cD\\O^;c0bD_O];?dDDZ;;fDFY;:gDGY;7iDIW;6iDJX;5hDLX;2iDOV;1kDOU;0kD0V;NkD3U;LkD5U;KjD5V;P100000O1000O01O100O10O01O100O100O1O100OfEbNj8]1TGiNi8W1UGnNh8Q1VGUOg8k0VG\\Of8c0XGCe8<YGJd86ZGOc80[G6b8J[G<b8C\\Gc0a8\\O`Gg0]8YObGk0[8TOeGn0Z8SOeGm0[8SOdGn0\\8ROcGo0]8QObGP1^8QOaGo0_8QO`GP1`8PO_GQ1a8oN^GR1b8oN]GQ1c8oN\\GR1d8nN[GS1f8lNYGU1g8lNXGT1h8lNWGU1j8kNTGV1l8jNTGV1l8kNRGV1o8iNPGX1P9hNoFY1Q9hNnFX1S9gNlFZ1T9^100001O0000001O0000001O000000001O01O0001O000000000000O1000000O1000000O1000000O100000000O1O1M3M3M3M3M3M3^MYK^Lj4a3fKmK]4ITKb2n0TMQ47VKa2Y1gLd3f0WK`2e1XLW3U1YK`2P7]MUI`2l6^MWI`2j6]M[I`2f6^M^I_2i5_LPJn0<MlNe1h6bMoIk0`0GoNj1c6eMmIj0c0@SOo1^6iMjIg0g0[OXOQ2X6oMhIe0j0TO\\OV2S6RNfId0i1W1b4VNdIc0k1McM=o6TObIa0n1IfM`0k6XO_I?R2CiMd0g6[O]I>T2^OlMh0d6]O[I=W2XOoMl0`6@YI;Q3Ol3GRI:V;HhD8X;IgD7X;KgD4Z;MeD3[;NdD2\\;0bD0^;1aDN`;3_DMa;4^DLb;n001O0O101O0000001O00O100O1O010O1O1O100O1O101N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O100O1000O0100O100O01O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001]FQOi6o0VIVOf6k0WI[Oe6f0YI_Oc6a0\\ID`6=^IH^68aIM[65aI1[63`I2\\62_I0`64[INd66WIKi69QIJn6:mHHR7<iHFV7>eHDZ7?bHC]7a0]HAc7c0XH_Og7e0TH]Ol7f0oG\\OP8h0kGZOT8n201O1O1O0000000000000000000000001O00000000000000001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O002N3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M4L3M3M3YKSHk3P8QLTHl3o7PLUHm3n7oKVHn3m7mKXHP4k7lKYHQ4c8M3M3M3M3M3M3M3M3M3M3M3M3M3M3XOgEYN\\jd1"}, "label": "a man sitting on a couch with a wii remote"}]}
{"id": 416723, "image": "COCO_train2014_000000416723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing wii wearing a brown shirt and brown hat , sitting between 2 women\"."}], "task": "refering", "answer_template": "The \"a man playing wii wearing a brown shirt and brown hat , sitting between 2 women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.11496875, 0.00948356807511737, 0.801796875, 1.0], "iscrowd": 0, "area": 85946.17104999999, "rle": {"size": [426, 640], "counts": "^`o01X=100O10O01O100O100O1O10O0100O100O10000O10O1000O10000O1000O10O10000O100PDEQ;;oDGo:8QEJm:7SEJl:6SELl:4TEMj:4VEMi:3VEOi:1WE0h:OYE2e:OZE2f:NZE1g:OYE1f:0YE1g:NZE1g:OXE2h:NXE1h:0XE0h:0WE1i:OWE0i:0XE0h:0WELn:4REFS;;mDES;;mDDT;<lDDT;;lDET;<_DBO2b;<^DC00c;=\\DD1Oc;=[DF1Mc;=\\DG1Kd;>ZDH1Mc;;[DI2Na;9\\DJ3O_;7]DK40\\;6_DL40[;3aDN4O[;3`DO4O\\;2_D05N\\;2^D16M[;3^D26K\\;2^D46J\\;f0dDYO];g0bDZO^;f0bDZO^;g0`DZO_;e0bD\\O^;c0cD\\O^;c0bD_O];?dDDZ;;fDFY;:gDGY;7iDIW;6iDJX;5hDLX;2iDOV;1kDOU;0kD0V;NkD3U;LkD5U;KjD5V;P100000O1000O01O100O10O01O100O100O1O100OfEbNj8]1TGiNi8W1UGnNh8Q1VGUOg8k0VG\\Of8c0XGCe8<YGJd86ZGOc80[G6b8J[G<b8C\\Gc0a8\\O`Gg0]8YObGk0[8TOeGn0Z8SOeGm0[8SOdGn0\\8ROcGo0]8QObGP1^8QOaGo0_8QO`GP1`8PO_GQ1a8oN^GR1b8oN]GQ1c8oN\\GR1d8nN[GS1f8lNYGU1g8lNXGT1h8lNWGU1j8kNTGV1l8jNTGV1l8kNRGV1o8iNPGX1P9hNoFY1Q9hNnFX1S9gNlFZ1T9^100001O0000001O0000001O000000001O01O0001O000000000000O1000000O1000000O1000000O100000000O1O1M3M3M3M3M3M3^MYK^Lj4a3fKmK]4ITKb2n0TMQ47VKa2Y1gLd3f0WK`2e1XLW3U1YK`2P7]MUI`2l6^MWI`2j6]M[I`2f6^M^I_2i5_LPJn0<MlNe1h6bMoIk0`0GoNj1c6eMmIj0c0@SOo1^6iMjIg0g0[OXOQ2X6oMhIe0j0TO\\OV2S6RNfId0i1W1b4VNdIc0k1McM=o6TObIa0n1IfM`0k6XO_I?R2CiMd0g6[O]I>T2^OlMh0d6]O[I=W2XOoMl0`6@YI;Q3Ol3GRI:V;HhD8X;IgD7X;KgD4Z;MeD3[;NdD2\\;0bD0^;1aDN`;3_DMa;4^DLb;n001O0O101O0000001O00O100O1O010O1O1O100O1O101N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O100O1000O0100O100O01O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001]FQOi6o0VIVOf6k0WI[Oe6f0YI_Oc6a0\\ID`6=^IH^68aIM[65aI1[63`I2\\62_I0`64[INd66WIKi69QIJn6:mHHR7<iHFV7>eHDZ7?bHC]7a0]HAc7c0XH_Og7e0TH]Ol7f0oG\\OP8h0kGZOT8n201O1O1O0000000000000000000000001O00000000000000001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O002N3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M4L3M3M3YKSHk3P8QLTHl3o7PLUHm3n7oKVHn3m7mKXHP4k7lKYHQ4c8M3M3M3M3M3M3M3M3M3M3M3M3M3M3XOgEYN\\jd1"}, "label": "a man playing wii wearing a brown shirt and brown hat , sitting between 2 women"}]}
{"id": 547795, "image": "COCO_train2014_000000547795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair that the man in a black shirt , hat , and glasses is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair that the man in a black shirt , hat , and glasses is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 252, 274, 275, 298, 313, 314, 315, 316, 317, 321, 336, 343, 344], "bbox": [0.606234375, 0.6374707259953162, 0.999234375, 1.0], "iscrowd": 0, "area": 4733.985349999999, "rle": {"size": [427, 640], "counts": "cVR52W=2M3N2M3N2M3N2M3M3N2M3O1O2OO0O1N3N1N3N1N3N100010O01O01O01O010O010O1O100O1O100O1O100O100O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O1O100O1O100O100O1O100Oo\\S11nblN3M3M3M3M3M4L3M3jEWOg7l0WH\\O`7g0_H[O]7h0aHZO]7h0bHYO[7k0bHWO[7l0SHdNQOa0g8P1WHbNnN`0e8T1]H_NiN?c8Y1dHZNfN>`8^1jHoNT7S1lHmNT7T1kHlNU7T1kHlNU7T1kHlNU7T1kHlNU7T1kHlNU7T1kHlNT7U1lHkNT7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7W1jHiNV7W1jHiNV7W1jHiNU7X1jHiNV7W1jHiNV7W1jHiNV7W1hIkM"}, "label": "the chair that the man in a black shirt , hat , and glasses is sitting in"}]}
{"id": 547795, "image": "COCO_train2014_000000547795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair under man in bkack shirt\"."}], "task": "refering", "answer_template": "The \"chair under man in bkack shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 252, 274, 275, 298, 313, 314, 315, 316, 317, 321, 336, 343, 344], "bbox": [0.606234375, 0.6374707259953162, 0.999234375, 1.0], "iscrowd": 0, "area": 4733.985349999999, "rle": {"size": [427, 640], "counts": "cVR52W=2M3N2M3N2M3N2M3M3N2M3O1O2OO0O1N3N1N3N1N3N100010O01O01O01O010O010O1O100O1O100O1O100O100O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O1O100O1O100O100O1O100Oo\\S11nblN3M3M3M3M3M4L3M3jEWOg7l0WH\\O`7g0_H[O]7h0aHZO]7h0bHYO[7k0bHWO[7l0SHdNQOa0g8P1WHbNnN`0e8T1]H_NiN?c8Y1dHZNfN>`8^1jHoNT7S1lHmNT7T1kHlNU7T1kHlNU7T1kHlNU7T1kHlNU7T1kHlNU7T1kHlNT7U1lHkNT7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7V1kHjNU7W1jHiNV7W1jHiNV7W1jHiNU7X1jHiNV7W1jHiNV7W1jHiNV7W1hIkM"}, "label": "chair under man in bkack shirt"}]}
{"id": 547795, "image": "COCO_train2014_000000547795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer being used by the man in front\"."}], "task": "refering", "answer_template": "The \"the computer being used by the man in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 256, 257, 258, 280, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332], "bbox": [0.1184375, 0.6763934426229508, 0.575734375, 0.9505620608899297], "iscrowd": 0, "area": 9171.059600000004, "rle": {"size": [427, 640], "counts": "Uoo09R=4L1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O2N1O1O1O1O100O1O1O1O1O1O1O1O100O1000000000000O10000000000000000000000000000O10000000000000000000000000000O1000001O000000000000000000000O1000000000O1O1O1O1O1O1O1O1O001O1O1O1O100O1O1O1O1O1O1O1O1O001O0000000000O10000000000O10000000001N10000000000O10000000O10O10000000000O100000O1000O10000000000O10O1QOPDc0P<[OSDd0m;[OUDd0k;ZOXDe0g;[O[Dd0e;ZO^De0b;ZOaDd0T<O2N1O2M2O2N1O0000000000O10O1000O100000000O100000O10O10000000000O1000O10000000O100000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O`Qa3"}, "label": "the computer being used by the man in front"}]}
{"id": 547795, "image": "COCO_train2014_000000547795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man ' s laptop\"."}], "task": "refering", "answer_template": "The \"the man ' s laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 256, 257, 258, 280, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332], "bbox": [0.1184375, 0.6763934426229508, 0.575734375, 0.9505620608899297], "iscrowd": 0, "area": 9171.059600000004, "rle": {"size": [427, 640], "counts": "Uoo09R=4L1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O2N1O1O1O1O100O1O1O1O1O1O1O1O100O1000000000000O10000000000000000000000000000O10000000000000000000000000000O1000001O000000000000000000000O1000000000O1O1O1O1O1O1O1O1O001O1O1O1O100O1O1O1O1O1O1O1O1O001O0000000000O10000000000O10000000001N10000000000O10000000O10O10000000000O100000O1000O10000000000O10O1QOPDc0P<[OSDd0m;[OUDd0k;ZOXDe0g;[O[Dd0e;ZO^De0b;ZOaDd0T<O2N1O2M2O2N1O0000000000O10O1000O100000000O100000O10O10000000000O1000O10000000O100000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O`Qa3"}, "label": "the man ' s laptop"}]}
{"id": 506837, "image": "COCO_train2014_000000506837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in the front staring at the camera\"."}], "task": "refering", "answer_template": "The \"zebra in the front staring at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 159, 160, 161, 171, 172, 178, 179, 189, 190, 195, 196, 213, 214], "bbox": [0.31583999999999995, 0.14087087087087086, 0.983, 0.9353753753753754], "iscrowd": 0, "area": 39526.51710000001, "rle": {"size": [333, 500], "counts": "j]c13Y:5K4K6K5M2M10O10O0oF\\OZ8d0cGBZ8=bGK[84aG4[8M_G=]8n0L4L3N3M3M9G0L5J5L4K3N3L300001O01O01O00001O0O101O0O101I6H8L5L3N3L3N1101N2O001N2O1O1N2O0O21O2N1O1N10O100O1O10O01O100O2N2O1N1O2O1N2O1N2N101N2N2OO01O100O00100O1O10O0100O1O010O1O100O00103L5L5J5L4K6K4K5L5K4K5L5J5SK`Kd3d5001N101O0O2O0O2O001N101O0O2N1O2N1RNQM]LQ3d2ZNnLg1R3j200O10010O00000001O0000001O01O000001O000000001O0001O01O000000001O0000001O000000001O0000001O000000001O000000001O0000001OO2O0000000O100000000O01000000O0100000O10O100000O0100000O10O1000O1000O1000O1000O1O001O100O1O001O100O001O1O1O010O1O11O2M3N2N1O2M3N2N2N2N2O0O2N2N2O1N2N2N2N101N2N2N2N2O1N1_LaKR1a4iNbKW1`4dNcK]1^4^NeKb1]4YNfKg1\\4TNgKl1[4oMhKQ2Y4iMlKX2U4aMPL_2R4ZMRLg2P4RMULn2m3kLXLU3j3cL\\L^3Y5001O1O001O1O1O0O2O1O1O001O1O1O001O1G8H9F:F:F9H9F:M3N3M4M2L5K5K5L3L5K5K5F:XNaH5h7ARIFX7Oj1Em^2"}, "label": "zebra in the front staring at the camera"}]}
{"id": 506837, "image": "COCO_train2014_000000506837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe turned left side\"."}], "task": "refering", "answer_template": "The \"giraffe turned left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 159, 160, 161, 171, 172, 178, 179, 189, 190, 195, 196, 213, 214], "bbox": [0.31583999999999995, 0.14087087087087086, 0.983, 0.9353753753753754], "iscrowd": 0, "area": 39526.51710000001, "rle": {"size": [333, 500], "counts": "j]c13Y:5K4K6K5M2M10O10O0oF\\OZ8d0cGBZ8=bGK[84aG4[8M_G=]8n0L4L3N3M3M9G0L5J5L4K3N3L300001O01O01O00001O0O101O0O101I6H8L5L3N3L3N1101N2O001N2O1O1N2O0O21O2N1O1N10O100O1O10O01O100O2N2O1N1O2O1N2O1N2N101N2N2OO01O100O00100O1O10O0100O1O010O1O100O00103L5L5J5L4K6K4K5L5K4K5L5J5SK`Kd3d5001N101O0O2O0O2O001N101O0O2N1O2N1RNQM]LQ3d2ZNnLg1R3j200O10010O00000001O0000001O01O000001O000000001O0001O01O000000001O0000001O000000001O0000001O000000001O000000001O0000001OO2O0000000O100000000O01000000O0100000O10O100000O0100000O10O1000O1000O1000O1000O1O001O100O1O001O100O001O1O1O010O1O11O2M3N2N1O2M3N2N2N2N2O0O2N2N2O1N2N2N2N101N2N2N2N2O1N1_LaKR1a4iNbKW1`4dNcK]1^4^NeKb1]4YNfKg1\\4TNgKl1[4oMhKQ2Y4iMlKX2U4aMPL_2R4ZMRLg2P4RMULn2m3kLXLU3j3cL\\L^3Y5001O1O001O1O1O0O2O1O1O001O1O1O001O1G8H9F:F:F9H9F:M3N3M4M2L5K5K5L3L5K5K5F:XNaH5h7ARIFX7Oj1Em^2"}, "label": "giraffe turned left side"}]}
{"id": 506837, "image": "COCO_train2014_000000506837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right facing zebra behind another zebra\"."}], "task": "refering", "answer_template": "The \"right facing zebra behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 26, 27, 28, 29, 37, 38, 39, 40, 41, 44, 45, 46, 47, 54, 55, 56, 57, 58, 59, 60, 64, 65, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 131, 133, 134, 144, 145, 151, 152, 162, 163, 164, 169, 170, 180, 181, 182, 187, 188], "bbox": [0.0, 0.0032432432432432435, 0.66986, 0.9147447447447447], "iscrowd": 0, "area": 35043.19265, "rle": {"size": [333, 500], "counts": "f5;m0Gg7l0iG@W1Kn4j0iI]OR12n4e0PJ[Ob0a0l4d0`Jo0[5TOcJP1W5W2K4L4L4K5L4L3M4L4K5L4L3M4LjN_LeKa3n3[MdKd2Q4jMmKU2R4oMlKP2R4VNkKi1T4[NjKd1U4`NiK_1U4hNgKW1X4_2N2O13M6J6J6J4L5K4L4L4L5K4M3M4mKZJV3_601O1O001O1N101O0000O1UNhL`LX3_3QMYLP3e3XMULh2h3bMPL_2n3iMkKX2R4oMiKR2U4SNgKn1W4VNgKi1X4\\NdKe1Z4`NbKa1\\4Y2N2N2N1O3M2N2N2N2N2N2N2N2N3M2O1N2N2N2001O00010O00001O0000M3N2010O1O1O1O1O1O001O1O1O1O1O00100O1O1O001O000000000010O00001O0001O01O00000010O0001O0000010O00001O00010O00001O0010O0001O000010O00000000000000O100mLYI`2g6]M^Ia2b6\\MdI`2]6]MhIa2X6\\MmIb2Q7N2M6K<D8H8G6K4L2N2M3N001O1N101O000O010000O010O100O001O1]IWOa3j0\\LYOb3h0[L]OX3o0eLTOR3U1kLoNR3S1jLQOT3P1hLUOX3k0aL[O`3e0iK2X4MdK7]4G`K<b4NnJ7R5b21N2M2N3K5LaJkKm4P4QKVLo4g3fJgL[5U41O0O2M3M3M\\MaK^O]4c0dK^OY4c0hK^OV4b0kK_OS4a0nK@P4a0RL_Om2dM\\Mn2H_Oh2a1XM`Nc2c1_M^N[2e1fM\\NT2i1lMYN\\1\\2fNdMU1`2kNaMS1_2oNbMm0`2SOaMk0_2VOcMf0^2\\ObMb0^2_OdM=^2CcM:^2HbM5_2LcMO_23bMI`27bMD`2=dM]O]2e0e30O1O2O0O2O0O2O0O2O0O2O0O101N1O2ROfNgHZ1Z7gNcH[1\\7gNbHY1_7hN^HZ1a7hN]HX1d7iNYHY1f7iNXHW1i7jNTHX1k7kNRHU1o7lNnGV1Q8lNmGT1T8c00O0010O010O010O010O001N1O2N1O2N1O2M2O1O2N1O2N1O2N01O2N1O2O001O3M4M2N2N2M01N2N100O1O100O1O100O10000OPN]HX1c7eNbHY1^7eNfHY1Y7eNlHMJe0b7TObIg0T8M3K5L4Hhbe1"}, "label": "right facing zebra behind another zebra"}]}
{"id": 506837, "image": "COCO_train2014_000000506837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zebras interlocked\"."}], "task": "refering", "answer_template": "The \"two zebras interlocked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 26, 27, 28, 29, 37, 38, 39, 40, 41, 44, 45, 46, 47, 54, 55, 56, 57, 58, 59, 60, 64, 65, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 131, 133, 134, 144, 145, 151, 152, 162, 163, 164, 169, 170, 180, 181, 182, 187, 188], "bbox": [0.0, 0.0032432432432432435, 0.66986, 0.9147447447447447], "iscrowd": 0, "area": 35043.19265, "rle": {"size": [333, 500], "counts": "f5;m0Gg7l0iG@W1Kn4j0iI]OR12n4e0PJ[Ob0a0l4d0`Jo0[5TOcJP1W5W2K4L4L4K5L4L3M4L4K5L4L3M4LjN_LeKa3n3[MdKd2Q4jMmKU2R4oMlKP2R4VNkKi1T4[NjKd1U4`NiK_1U4hNgKW1X4_2N2O13M6J6J6J4L5K4L4L4L5K4M3M4mKZJV3_601O1O001O1N101O0000O1UNhL`LX3_3QMYLP3e3XMULh2h3bMPL_2n3iMkKX2R4oMiKR2U4SNgKn1W4VNgKi1X4\\NdKe1Z4`NbKa1\\4Y2N2N2N1O3M2N2N2N2N2N2N2N2N3M2O1N2N2N2001O00010O00001O0000M3N2010O1O1O1O1O1O001O1O1O1O1O00100O1O1O001O000000000010O00001O0001O01O00000010O0001O0000010O00001O00010O00001O0010O0001O000010O00000000000000O100mLYI`2g6]M^Ia2b6\\MdI`2]6]MhIa2X6\\MmIb2Q7N2M6K<D8H8G6K4L2N2M3N001O1N101O000O010000O010O100O001O1]IWOa3j0\\LYOb3h0[L]OX3o0eLTOR3U1kLoNR3S1jLQOT3P1hLUOX3k0aL[O`3e0iK2X4MdK7]4G`K<b4NnJ7R5b21N2M2N3K5LaJkKm4P4QKVLo4g3fJgL[5U41O0O2M3M3M\\MaK^O]4c0dK^OY4c0hK^OV4b0kK_OS4a0nK@P4a0RL_Om2dM\\Mn2H_Oh2a1XM`Nc2c1_M^N[2e1fM\\NT2i1lMYN\\1\\2fNdMU1`2kNaMS1_2oNbMm0`2SOaMk0_2VOcMf0^2\\ObMb0^2_OdM=^2CcM:^2HbM5_2LcMO_23bMI`27bMD`2=dM]O]2e0e30O1O2O0O2O0O2O0O2O0O2O0O101N1O2ROfNgHZ1Z7gNcH[1\\7gNbHY1_7hN^HZ1a7hN]HX1d7iNYHY1f7iNXHW1i7jNTHX1k7kNRHU1o7lNnGV1Q8lNmGT1T8c00O0010O010O010O010O001N1O2N1O2N1O2M2O1O2N1O2N1O2N01O2N1O2O001O3M4M2N2N2M01N2N100O1O100O1O100O10000OPN]HX1c7eNbHY1^7eNfHY1Y7eNlHMJe0b7TObIg0T8M3K5L4Hhbe1"}, "label": "two zebras interlocked"}]}
{"id": 400343, "image": "COCO_train2014_000000400343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 104, 105, 106, 121, 122, 123, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 260, 261, 262, 277, 278, 279, 294, 295, 296, 297, 311, 312, 313, 314], "bbox": [0.0, 0.252046875, 0.5011249999999999, 0.8023906249999999], "iscrowd": 0, "area": 38248.20405000001, "rle": {"size": [640, 480], "counts": "b9?Zc0m0XOi0WOh0XOh0XOh0XO9G2N2N2N2N2O1N3M2N2N2N2N2N2N2N2O1N2N2N2N2N2N3M2N2N2N2O10000O100O10000O10001N100O10000O1000O0100O0100O010O01000O01fNY110O1O0010O01O010O1O010O0010O01O10O01O010O1O010O0010O01O10O01O010O001000O100O101O0O101N100O101O0O101N100O101O0O101N100mJgC_2Y<`MjC]2X<aMlC\\2T<bMQDZ2P<eMSDY2n;eMVDW2k;hMYDU2Q<_MSD_2\\<QMhCk2h<eL\\CX3S=XLQCd3h>N3L4M2N3L4M2M4M3M2M4M3L3N3L4M2N3O11N3K5I7J6I6K6J6iA[Lh;j0VBP2Y1jMa<2ZBX2=YNZ=[O^BZ4b=bKbB^4^=_KfB`4Z=\\KjBd4U=YKoBg4Q=UKTCj4Q>01O001O001O001O001O001N10001O003M4L4L3M4L4L4L3M4L4L4L3M4L4K5L3M4L4L4L3M4L4L4L3M4L4L4L3M4L4L2N1N101O1O1O001O1O001O1O001O1O1O001O1O0O2O1O001O1N2M2M4Ma`e4"}, "label": "a man wearing a black shirt"}]}
{"id": 400343, "image": "COCO_train2014_000000400343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt and shoes playing a video game\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt and shoes playing a video game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 104, 105, 106, 121, 122, 123, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 260, 261, 262, 277, 278, 279, 294, 295, 296, 297, 311, 312, 313, 314], "bbox": [0.0, 0.252046875, 0.5011249999999999, 0.8023906249999999], "iscrowd": 0, "area": 38248.20405000001, "rle": {"size": [640, 480], "counts": "b9?Zc0m0XOi0WOh0XOh0XOh0XO9G2N2N2N2N2O1N3M2N2N2N2N2N2N2N2O1N2N2N2N2N2N3M2N2N2N2O10000O100O10000O10001N100O10000O1000O0100O0100O010O01000O01fNY110O1O0010O01O010O1O010O0010O01O10O01O010O1O010O0010O01O10O01O010O001000O100O101O0O101N100O101O0O101N100O101O0O101N100mJgC_2Y<`MjC]2X<aMlC\\2T<bMQDZ2P<eMSDY2n;eMVDW2k;hMYDU2Q<_MSD_2\\<QMhCk2h<eL\\CX3S=XLQCd3h>N3L4M2N3L4M2M4M3M2M4M3L3N3L4M2N3O11N3K5I7J6I6K6J6iA[Lh;j0VBP2Y1jMa<2ZBX2=YNZ=[O^BZ4b=bKbB^4^=_KfB`4Z=\\KjBd4U=YKoBg4Q=UKTCj4Q>01O001O001O001O001O001N10001O003M4L4L3M4L4L4L3M4L4L4L3M4L4K5L3M4L4L4L3M4L4L4L3M4L4L4L3M4L4L2N1N101O1O1O001O1O001O1O001O1O1O001O1O0O2O1O001O1N2M2M4Ma`e4"}, "label": "a man in a black shirt and shoes playing a video game"}]}
{"id": 400343, "image": "COCO_train2014_000000400343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child playing with game controller\"."}], "task": "refering", "answer_template": "The \"child playing with game controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 292, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 340, 341, 342, 343, 344, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 364, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.002875, 0.759140625, 0.5075208333333333, 0.98925], "iscrowd": 0, "area": 25197.002750000007, "rle": {"size": [640, 480], "counts": "ST1c3\\`01O1O1O1O100O1O1O2N1O1O100O1O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000000000001O000000000000000000001O0001O0000000001O001O001O001O00001O001O2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M=CN2N2N2N2N2O1000000O100000000O1000000O1000000O1001O000000000000001O00001O1O2N1O2N2N1O1OO1000000O11O0001O0001O000000001O000000001O000000002N3M2N3M3M3M2N3M3M3M2N3M1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1OY`c4"}, "label": "child playing with game controller"}]}
{"id": 400343, "image": "COCO_train2014_000000400343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy who is holding the white controller behind the other guy\"."}], "task": "refering", "answer_template": "The \"the guy who is holding the white controller behind the other guy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 292, 306, 307, 308, 309, 310, 323, 324, 325, 326, 327, 328, 329, 340, 341, 342, 343, 344, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 364, 374, 375, 376, 377, 378, 379, 380, 381, 382], "bbox": [0.002875, 0.759140625, 0.5075208333333333, 0.98925], "iscrowd": 0, "area": 25197.002750000007, "rle": {"size": [640, 480], "counts": "ST1c3\\`01O1O1O1O100O1O1O2N1O1O100O1O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000000000001O000000000000000000001O0001O0000000001O001O001O001O00001O001O2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N3M=CN2N2N2N2N2O1000000O100000000O1000000O1000000O1001O000000000000001O00001O1O2N1O2N2N1O1OO1000000O11O0001O0001O000000001O000000001O000000002N3M2N3M3M3M2N3M3M3M2N3M1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1OY`c4"}, "label": "the guy who is holding the white controller behind the other guy"}]}
{"id": 7129, "image": "COCO_train2014_000000007129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman skiing with blue coat\"."}], "task": "refering", "answer_template": "The \"woman skiing with blue coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 131, 132, 148, 149, 150, 164, 165, 166, 167, 168, 181, 182, 183, 184, 198, 199, 200, 201, 215, 216, 217, 218, 232, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 284, 285, 286, 301, 302, 303, 317, 318, 319, 320, 321, 335, 336, 337, 338], "bbox": [0.6592083333333334, 0.257484375, 0.9157291666666667, 0.861515625], "iscrowd": 0, "area": 30008.36700000001, "rle": {"size": [640, 480], "counts": "jjU62\\c0d0\\Oc0B?K5K4K6K5K4L6J8G8I8gCgL^7`3XHgLd7`3XHaLd7g3VH[Lf7m3UHTLh7U4PHmKm7\\4lGeKS6@bHT5T1\\KY6F\\HW5S1TK_6KXHY5S1lJc61THZ5S1fJh65oGZ5V1aJi6<jGX5Z1\\Jj6b0gGV5[1YJm6f0bGV5^1TJn6l0^GU5`1oIQ7S1XGS5d1jIS7V8jHjGT7[8hHfGV7_8gHaGX7e8bH\\G^7l8ZHTGf7S9RHnFn7Y9kGgFT8a9eG^F\\8i9\\GXFd8b:fGhDl7X;SHkDj7V;VHmDg7S;XHPEf7P;ZHQEe7o:ZHREf7m:[HSEe7m:ZHTEf7l:ZHTEf7l:YHUEg7k:YHUEg7k:XHVEh7j:XHVEh7j:XHVEh7j:WHWEi7h:XHXEh7h:WHYEi7g:WHYEi7g:UH[El7\\;1O2N1O1O1O1O1O2N1O1O1O1O1O1O2NR1nN1O1O1O1O1O1O1O1O1WLWFYMj9d2ZFZMg9c2^FZMc9c2bFZM_9c2eF[M\\9b2iF[MX9b2mF[MT9a2QG]MP9`2UG]Ml8`2oEaJT1m2n8_2lEdJ]1j2i8`2]G]Me8^2aG_M`8\\2fGbM\\8Y2jGdMX8W2nGfMS8U2THhMn7R2YHkMi7n1_HoMb7l1eHQN]7<_EiN[3i0X79jE_NV3U1Q78VFTNQ3a1k67aJGa54dJJ]52hJLZ5WOcKf0`4nNlKP1V4cNWLZ1Y;O2M3N2M3N2M2O2L4K5K5K5LbSi0"}, "label": "woman skiing with blue coat"}]}
{"id": 7129, "image": "COCO_train2014_000000007129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a blue jacket is about to ski\"."}], "task": "refering", "answer_template": "The \"a boy in a blue jacket is about to ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 131, 132, 148, 149, 150, 164, 165, 166, 167, 168, 181, 182, 183, 184, 198, 199, 200, 201, 215, 216, 217, 218, 232, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 284, 285, 286, 301, 302, 303, 317, 318, 319, 320, 321, 335, 336, 337, 338], "bbox": [0.6592083333333334, 0.257484375, 0.9157291666666667, 0.861515625], "iscrowd": 0, "area": 30008.36700000001, "rle": {"size": [640, 480], "counts": "jjU62\\c0d0\\Oc0B?K5K4K6K5K4L6J8G8I8gCgL^7`3XHgLd7`3XHaLd7g3VH[Lf7m3UHTLh7U4PHmKm7\\4lGeKS6@bHT5T1\\KY6F\\HW5S1TK_6KXHY5S1lJc61THZ5S1fJh65oGZ5V1aJi6<jGX5Z1\\Jj6b0gGV5[1YJm6f0bGV5^1TJn6l0^GU5`1oIQ7S1XGS5d1jIS7V8jHjGT7[8hHfGV7_8gHaGX7e8bH\\G^7l8ZHTGf7S9RHnFn7Y9kGgFT8a9eG^F\\8i9\\GXFd8b:fGhDl7X;SHkDj7V;VHmDg7S;XHPEf7P;ZHQEe7o:ZHREf7m:[HSEe7m:ZHTEf7l:ZHTEf7l:YHUEg7k:YHUEg7k:XHVEh7j:XHVEh7j:XHVEh7j:WHWEi7h:XHXEh7h:WHYEi7g:WHYEi7g:UH[El7\\;1O2N1O1O1O1O1O2N1O1O1O1O1O1O2NR1nN1O1O1O1O1O1O1O1O1WLWFYMj9d2ZFZMg9c2^FZMc9c2bFZM_9c2eF[M\\9b2iF[MX9b2mF[MT9a2QG]MP9`2UG]Ml8`2oEaJT1m2n8_2lEdJ]1j2i8`2]G]Me8^2aG_M`8\\2fGbM\\8Y2jGdMX8W2nGfMS8U2THhMn7R2YHkMi7n1_HoMb7l1eHQN]7<_EiN[3i0X79jE_NV3U1Q78VFTNQ3a1k67aJGa54dJJ]52hJLZ5WOcKf0`4nNlKP1V4cNWLZ1Y;O2M3N2M3N2M2O2L4K5K5K5LbSi0"}, "label": "a boy in a blue jacket is about to ski"}]}
{"id": 7129, "image": "COCO_train2014_000000007129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child dressed in a black ski jacket\"."}], "task": "refering", "answer_template": "The \"a child dressed in a black ski jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 193, 194, 195, 196, 210, 211, 212, 227, 228, 229, 244, 245, 246, 247, 262, 263, 264, 278, 279, 280, 281, 282, 295, 296, 298, 299, 315, 316], "bbox": [0.3647708333333333, 0.46926562499999996, 0.6275625, 0.8069843750000001], "iscrowd": 0, "area": 13070.1843, "rle": {"size": [640, 480], "counts": "ak]32jc05N100O100O100O100O1nJ6nEJo9<mEER:=iB\\Of18`;>_BFn1Mc;>UB1U2Ae;`0lA;[2VOi;S2SC_M@?\\=V2nBaMC9_=Y2hBeMD3c=[2dBhMFMf=^2_BkMFHj=`2[BnMGCo=a2TBRNJ]OS>c2mAWNKVOZ>d2fAPO[>Y3101N2O1N101N2O00O1O1O1O2N1N2O1O2N0O2N1O2O00010OO2lNZBlKg=P4W10O2O1N2O0O2O1M3O001O100O1O010O1O1000jLWAS1i>jN\\AV1c>fNbAY1^>bNgA^1Z>YNoAf1Q>UNUBj17WMW<k0fCm1J`Mb<>hCS2BaMh<8UDk1QOnMl<2\\Dk1fNTNP=K`DP2^NVNg?j1W@WNk?h1S@YNo?g1o_OYNS`0f1k_O[NW`0d1h_O\\NZ`0c1d_O^N_`0`1`_O`Nf`0[1X_OfNn`0Ej^OU16VO]a0j0b^OVO_a0j0_^OWOba0i0\\^OXOea0h0Y^OXOja0g0T^OZOma0\\11N2O1O1N2O2N1N2O100O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1UOU]O?lb0]OY]Oa0ib0[O\\]Ob0Tc0M_\\_3"}, "label": "a child dressed in a black ski jacket"}]}
{"id": 7129, "image": "COCO_train2014_000000007129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a black coat skiing\"."}], "task": "refering", "answer_template": "The \"a boy in a black coat skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 193, 194, 195, 196, 210, 211, 212, 227, 228, 229, 244, 245, 246, 247, 262, 263, 264, 278, 279, 280, 281, 282, 295, 296, 298, 299, 315, 316], "bbox": [0.3647708333333333, 0.46926562499999996, 0.6275625, 0.8069843750000001], "iscrowd": 0, "area": 13070.1843, "rle": {"size": [640, 480], "counts": "ak]32jc05N100O100O100O100O1nJ6nEJo9<mEER:=iB\\Of18`;>_BFn1Mc;>UB1U2Ae;`0lA;[2VOi;S2SC_M@?\\=V2nBaMC9_=Y2hBeMD3c=[2dBhMFMf=^2_BkMFHj=`2[BnMGCo=a2TBRNJ]OS>c2mAWNKVOZ>d2fAPO[>Y3101N2O1N101N2O00O1O1O1O2N1N2O1O2N0O2N1O2O00010OO2lNZBlKg=P4W10O2O1N2O0O2O1M3O001O100O1O010O1O1000jLWAS1i>jN\\AV1c>fNbAY1^>bNgA^1Z>YNoAf1Q>UNUBj17WMW<k0fCm1J`Mb<>hCS2BaMh<8UDk1QOnMl<2\\Dk1fNTNP=K`DP2^NVNg?j1W@WNk?h1S@YNo?g1o_OYNS`0f1k_O[NW`0d1h_O\\NZ`0c1d_O^N_`0`1`_O`Nf`0[1X_OfNn`0Ej^OU16VO]a0j0b^OVO_a0j0_^OWOba0i0\\^OXOea0h0Y^OXOja0g0T^OZOma0\\11N2O1O1N2O2N1N2O100O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1UOU]O?lb0]OY]Oa0ib0[O\\]Ob0Tc0M_\\_3"}, "label": "a boy in a black coat skiing"}]}
{"id": 7129, "image": "COCO_train2014_000000007129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis the person in light blue is wearing\"."}], "task": "refering", "answer_template": "The \"the skis the person in light blue is wearing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [321, 322, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356], "bbox": [0.4309583333333334, 0.7895625, 1.0, 0.88946875], "iscrowd": 0, "area": 4605.71755, "rle": {"size": [640, 480], "counts": "jlQ42mc03N2M3N1O100O1O010O1O010O0010O01000O1O2O0O00010O0001O000001O00000000N201OD_\\O9ac0G_\\O9`c0400000000O1000000O100000000O1000000O1000001O0O1000000O1000Fg\\OMYc03g\\OMYc03g\\OMXc04h\\OLXc04h\\OLXc03i\\OMWc03i\\OMVc04j\\OLVc04j\\OLVc03k\\OMUc03k\\OMUc03k\\OMTc04l\\OLTc04l\\OLTc03m\\OMSc03m\\OMRc04n\\OLRc04n\\OMQc03o\\OMQc02P]ONPc02P]ONob03Q]OMob03Q]OMob02R]ONnb02R]ONmb03S]OMmb03S]OMmb03S]OMmb02T]ONlb02T]ONkb03U]OMkb03U]OMkb03U]OMkb02V]ONib03W]OMib03W]OMib03W]OMib02X]ONhb02X]ONgb03Y]OMgb03Y]OMgb03Y]ONfb01[]OOdb02\\]ONdb02\\]ONdb02\\]ONdb02\\]ONdb01]]OObb02^]ONbb02^]ONbb02^]ONbb01_]OObb00^]O0cb0O]]O1db0N\\]O2db0N\\]O2eb0L\\]O4db0L\\]O4db0L\\]O4eb0K[]O5eb0K[]O5eb0J\\]O6db0J\\]O6db0J\\]O6db0J\\]O6db0I]]O7bb0J^]O7ab0I_]O7ab0I_]O7ab0I_]O7ab0H`]O8`b0H`]O8`b0H`]O8_b0Ib]O6^b0Jb]O6^b0Ic]O7]b0Ic]O7\\b0Jd]O6\\b0Jd]O6\\b0Ie]O7Zb0Jf]O6Zb0Jf]O6Zb0Jf]O6Yb0Kg]O5Yb0Jh]O6Xb0Jh]O6Wb0Ki]O5Wb0Ki]O5Wb0Ki]O5Vb0Kk]O5Ub0Kk]O6Tb0Jl]O6Sb0HP^O8lb00000000000000000000001O0000000000000000000001O00000000000000000000000000000000000000000000001O0000000000000000000000000000000fNHl^O8Sa0Im^O7Sa0Im^O7Sa0Im^O7Ra0Jn^O6Ra0Jn^O6Ra0Io^O7Qa0Io^O7Pa0JP_O7o`0IQ_O7o`0IQ_O7n`0JQ_O7o`0IQ_O7o`0IQ_O7o`0IQ_O7n`0JR_O6n`0JR_O6n`0KQ_O5o`0KQ_O5n`0LR_O4n`0LR_O4n`0LR_O4n`0LR_O4n`0KS_O5m`0KS_O5m`0KS_O5l`0KU_O5l`0JT_O6l`0JT_O6m`0IS_O7m`0HT_O8l`0HT_O8l`0HT_O9l`0EU_O;k`0EU_O;k`0EU_O;k`0CW_O=i`0AY_O?Tb001O00000000001O00002Nc^O"}, "label": "the skis the person in light blue is wearing"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red chairs on the right behind the man\"."}], "task": "refering", "answer_template": "The \"red chairs on the right behind the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 91, 108, 109, 114, 132, 136, 137, 159, 160], "bbox": [0.715640625, 0.29746518105849584, 0.999875, 0.535125348189415], "iscrowd": 0, "area": 3816.865299999996, "rle": {"size": [359, 640], "counts": "aeP51V;2N2N2N3M2N2N2N3M2N2N2N3M2N2N2O1N3M2N2N2N3M2N2N2N2N3M2N00M3M3L4M3L4M3L4M4L3WOXmT1O]SkN5J6J6J6K5J6J6H8@`0@?0100000000001O00000000001O00000000001O0001O00000006Jb1^NiJ"}, "label": "red chairs on the right behind the man"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clear cup behind the orange juice in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the clear cup behind the orange juice in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 89, 109, 110, 132, 133, 155, 156, 178, 179, 202], "bbox": [0.77403125, 0.22002785515320333, 0.9012812499999999, 0.655933147632312], "iscrowd": 0, "area": 4910.140649999999, "rle": {"size": [359, 640], "counts": "Ve]55d9c1K5Kh2XM001O0000001O0001O01O0000001O0000001O0000001O0000dKSJ^3m5bLSK_2l4aMTL_1l3aNTM?m2@]400000000000000000000000000001O0000000001O00000000000000000000000000000000000000001O00000000005K6Jkoe0"}, "label": "the clear cup behind the orange juice in the right hand picture"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty glass stands behind a glass of orange juice\"."}], "task": "refering", "answer_template": "The \"an empty glass stands behind a glass of orange juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 89, 109, 110, 132, 133, 155, 156, 178, 179, 202], "bbox": [0.77403125, 0.22002785515320333, 0.9012812499999999, 0.655933147632312], "iscrowd": 0, "area": 4910.140649999999, "rle": {"size": [359, 640], "counts": "Ve]55d9c1K5Kh2XM001O0000001O0001O01O0000001O0000001O0000001O0000dKSJ^3m5bLSK_2l4aMTL_1l3aNTM?m2@]400000000000000000000000000001O0000000001O00000000000000000000000000000000000000001O00000000005K6Jkoe0"}, "label": "an empty glass stands behind a glass of orange juice"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"juice tumbler with istra in the table\"."}], "task": "refering", "answer_template": "The \"juice tumbler with istra in the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 227, 228], "bbox": [0.8206093750000001, 0.262924791086351, 0.9643125000000001, 0.7191086350974931], "iscrowd": 0, "area": 11807.998899999995, "rle": {"size": [359, 640], "counts": "iTh58o:i0WOj0WOh0WOi0WOj0VO8H00001O0000001O00001O00001O00001O00001O00001O00001O00010O00001O1O1O1O1O1O001O1O1O1O1O1O0000000000O100000000O100000000O2O000M3I7J6J6I7J6I7J6J6I8I6J6I7J6J7H7J6I7J7I6I7J6J7HSo7"}, "label": "juice tumbler with istra in the table"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass of juice with a straw\"."}], "task": "refering", "answer_template": "The \"glass of juice with a straw\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 227, 228], "bbox": [0.8206093750000001, 0.262924791086351, 0.9643125000000001, 0.7191086350974931], "iscrowd": 0, "area": 11807.998899999995, "rle": {"size": [359, 640], "counts": "iTh58o:i0WOj0WOh0WOi0WOj0VO8H00001O0000001O00001O00001O00001O00001O00001O00001O00010O00001O1O1O1O1O1O001O1O1O1O1O1O0000000000O100000000O100000000O2O000M3I7J6J6I7J6I7J6J6I8I6J6I7J6J7H7J6I7J7I6I7J6J7HSo7"}, "label": "glass of juice with a straw"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fork that is on top of the waffle\"."}], "task": "refering", "answer_template": "The \"a fork that is on top of the waffle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 248, 249, 250], "bbox": [0.5697656249999999, 0.6831476323119777, 0.999609375, 0.8179665738161559], "iscrowd": 0, "area": 1978.9039000000025, "rle": {"size": [359, 640], "counts": "oWP41U;0O2O0O2O0O101N100O2O001N10001N110O01O01O01O01O010O00010O010O00010O0010O0010O000100O010O1O010O10O01O10O01O10O0100O1O010000O01000O1000O0100000O100000000O10000000001O01O0000001O00001O0000001O0000001O001O00001O0100O1O010O10O01O10O0100O1O010O1O010O10O01O010O010O00010O010O0010O010O01O010O0010O010O01O010O010O001O001O00000000000000000000O100000000000O10000000000000000000000000O1000000000000000000000000000O100000000000O100000000000000000000000000000000000000O1000O1000000000000000000000000000000000O100000000000000000000000000000O1000000000O10000000000bG"}, "label": "a fork that is on top of the waffle"}]}
{"id": 89052, "image": "COCO_train2014_000000089052.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork placed on the bread\"."}], "task": "refering", "answer_template": "The \"the fork placed on the bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 248, 249, 250], "bbox": [0.5697656249999999, 0.6831476323119777, 0.999609375, 0.8179665738161559], "iscrowd": 0, "area": 1978.9039000000025, "rle": {"size": [359, 640], "counts": "oWP41U;0O2O0O2O0O101N100O2O001N10001N110O01O01O01O01O010O00010O010O00010O0010O0010O000100O010O1O010O10O01O10O01O10O0100O1O010000O01000O1000O0100000O100000000O10000000001O01O0000001O00001O0000001O0000001O001O00001O0100O1O010O10O01O10O0100O1O010O1O010O10O01O010O010O00010O010O0010O010O01O010O0010O010O01O010O010O001O001O00000000000000000000O100000000000O10000000000000000000000000O1000000000000000000000000000O100000000000O100000000000000000000000000000000000000O1000O1000000000000000000000000000000000O100000000000000000000000000000O1000000000O10000000000bG"}, "label": "the fork placed on the bread"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear umbrella with a plain black border\"."}], "task": "refering", "answer_template": "The \"a clear umbrella with a plain black border\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 210, 211, 232, 233], "bbox": [0.03546875, 0.37592592592592594, 0.19864062500000002, 0.7234567901234568], "iscrowd": 0, "area": 11002.98155, "rle": {"size": [405, 640], "counts": "QY9o1e:2N2N2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N3M2N1O1O1001O000000001O000000001O00000000001O000000001O0000000O2O0000000gMdG9\\8EfG;[8AhG?X8^OkGb0U8\\OmGd0S8YOPHg0P8VOSHj0n7TOSHl0m7SOUHl0k7ROWHn0i7QOXHo0h7POYHP1g7nN[HQ1g7mNZHR1g7mNZHR1g7lN[HR1g7mNZHR1g7mNZHR1i7jNYHT1k7iNVHV1l7hNUH;POHn8ITHKB:^8FWI9l6CWI;k6AXI=a9O2N1N2O3M3L^dZ6"}, "label": "a clear umbrella with a plain black border"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear umbrella with thin black rim held down\"."}], "task": "refering", "answer_template": "The \"clear umbrella with thin black rim held down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 210, 211, 232, 233], "bbox": [0.03546875, 0.37592592592592594, 0.19864062500000002, 0.7234567901234568], "iscrowd": 0, "area": 11002.98155, "rle": {"size": [405, 640], "counts": "QY9o1e:2N2N2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2N3M2N1O1O1001O000000001O000000001O00000000001O000000001O0000000O2O0000000gMdG9\\8EfG;[8AhG?X8^OkGb0U8\\OmGd0S8YOPHg0P8VOSHj0n7TOSHl0m7SOUHl0k7ROWHn0i7QOXHo0h7POYHP1g7nN[HQ1g7mNZHR1g7mNZHR1g7lN[HR1g7mNZHR1g7mNZHR1i7jNYHT1k7iNVHV1l7hNUH;POHn8ITHKB:^8FWI9l6CWI;k6AXI=a9O2N1N2O3M3L^dZ6"}, "label": "clear umbrella with thin black rim held down"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black coat\"."}], "task": "refering", "answer_template": "The \"a woman in a black coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 74, 75, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 234, 235, 258, 259, 281, 282], "bbox": [0.15784375, 0.18876543209876545, 0.318546875, 0.9191111111111111], "iscrowd": 0, "area": 16287.748900000004, "rle": {"size": [405, 640], "counts": "iTX12a<3M3M3M3M3L5L2N2N2N2N2N3M2M3N2N2N2N3M2XF`NY8b1]GiN`8Y1VGQOh8Q1mF[OQ9f0eFEX9m1O2M3N1O2M3Nk0UO3L4M3K5K5J6K5J6K5J6K6J6VKcIc2R7XMQIc2W7XMkHd2\\7WMeHe2c7VM_Hf2h7TM[Hl2g7PM[Ho2i7lLYHT3j7gLWHY3m7aLVH_3`800O01000O11O0O2O001O0VMaL^L_3b3gLWL[3h3lLQLT3n3SMlKn2S4XMgKh2Y4cM\\K^2b4TNlJm1T5_N`Jb1_5jNUJV1j5k201O0O2O001N104L;D8I3M2N2M3N1O0O2O1O0O2O001N1K6H7D=B=D=B=D]1bNcY\\5"}, "label": "a woman in a black coat"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dress woman holding a transparent umbrella standing with two friends\"."}], "task": "refering", "answer_template": "The \"a black dress woman holding a transparent umbrella standing with two friends\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 74, 75, 97, 98, 99, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 234, 235, 258, 259, 281, 282], "bbox": [0.15784375, 0.18876543209876545, 0.318546875, 0.9191111111111111], "iscrowd": 0, "area": 16287.748900000004, "rle": {"size": [405, 640], "counts": "iTX12a<3M3M3M3M3L5L2N2N2N2N2N3M2M3N2N2N2N3M2XF`NY8b1]GiN`8Y1VGQOh8Q1mF[OQ9f0eFEX9m1O2M3N1O2M3Nk0UO3L4M3K5K5J6K5J6K5J6K6J6VKcIc2R7XMQIc2W7XMkHd2\\7WMeHe2c7VM_Hf2h7TM[Hl2g7PM[Ho2i7lLYHT3j7gLWHY3m7aLVH_3`800O01000O11O0O2O001O0VMaL^L_3b3gLWL[3h3lLQLT3n3SMlKn2S4XMgKh2Y4cM\\K^2b4TNlJm1T5_N`Jb1_5jNUJV1j5k201O0O2O001N104L;D8I3M2N2M3N1O0O2O1O0O2O001N1K6H7D=B=D=B=D]1bNcY\\5"}, "label": "a black dress woman holding a transparent umbrella standing with two friends"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella held by a girl in red coat\"."}], "task": "refering", "answer_template": "The \"umbrella held by a girl in red coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 31, 32, 33, 34, 35, 36, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 102], "bbox": [0.358375, 0.014888888888888889, 0.608609375, 0.33859259259259256], "iscrowd": 0, "area": 10801.405649999997, "rle": {"size": [405, 640], "counts": "QQk22b<4K5L7oC_O`;l0N1O1O2N2N3M2N2N2N3M2N2N2O2N1O1O2M2O1O2N1O1O2N101N1O1O2N100O2N100O2O0O101N101N100O2O00001N102N00O01000000O1000001N100000000O1001O0N2M3N2O1NXG`M_7_2VHmMj7Q2nGYNR8d1gGaN^8]1YGdNQ9V2:D<E;D<L5N1OO2M3000O6K1O0001O000001O01O00000010O01O1O101N1O100O1O2O1N2O2M2O2N3L5L0O01O001O1N101O001O1O1O1N2O1O1O1O1O1O1N2O1O1O1O2N2N2M3N2N2N2N2M3N3M2N2M3J6I8I9I6IdiR3"}, "label": "umbrella held by a girl in red coat"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella above the middle woman ' s head\"."}], "task": "refering", "answer_template": "The \"umbrella above the middle woman ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 31, 32, 33, 34, 35, 36, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 102], "bbox": [0.358375, 0.014888888888888889, 0.608609375, 0.33859259259259256], "iscrowd": 0, "area": 10801.405649999997, "rle": {"size": [405, 640], "counts": "QQk22b<4K5L7oC_O`;l0N1O1O2N2N3M2N2N2N3M2N2N2O2N1O1O2M2O1O2N1O1O2N101N1O1O2N100O2N100O2O0O101N101N100O2O00001N102N00O01000000O1000001N100000000O1001O0N2M3N2O1NXG`M_7_2VHmMj7Q2nGYNR8d1gGaN^8]1YGdNQ9V2:D<E;D<L5N1OO2M3000O6K1O0001O000001O01O00000010O01O1O101N1O100O1O2O1N2O2M2O2N3L5L0O01O001O1N101O001O1O1O1N2O1O1O1O1O1O1N2O1O1O1O2N2N2M3N2N2N2N2M3N3M2N2M3J6I8I9I6IdiR3"}, "label": "umbrella above the middle woman ' s head"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing grey stretch pants holding a polka dot umbrella\"."}], "task": "refering", "answer_template": "The \"a woman wearing grey stretch pants holding a polka dot umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 63, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316], "bbox": [0.585765625, 0.06444444444444444, 0.8042031250000001, 0.9851358024691359], "iscrowd": 0, "area": 31983.166650000003, "rle": {"size": [405, 640], "counts": "Rad49T<=D=C=C;E7I7M4L3M3M3M3M3kNRN`GQ2[8W1J7J5K5K6M2O2M2O2N1N3N1N3N2N1N3N1N3N1M4J5K6J5K6J5N3N2N101N1O2N1O2N7lMlHZN[7U1PIjNW7g0SIWOS7<UICQ7L[I4h6[OfId0_6kNnIT1k8N1O1O1O2N1O1O101N1O1O11O010O001O010O001O010O001O001O001O1O001O001O001O001O2N4L3M4L3M3M4L3M4bLkMTLY2c3oMZLT2^3TN_LP2X3YNdLj1T3^NiLf1n2cNnL`1k2gNRM]1i2hNSM[1i2iNTMY1i2kNTMW1h2nNUMS1i2QOTMQ1i2ROUMo0h2VOUMl0h2XOUMi0i2[OTMg0i2]OPMi0m2ZOjLn0U3UOaLT1\\3POZLX1e3kNRL]1k3h3N3N2M2N3N2M5K4J7H8Gh0XO6J6J6POQ1J5L4K5LiXa1"}, "label": "a woman wearing grey stretch pants holding a polka dot umbrella"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman with spiky hair and a denim overshirt holding a polka dot umbrella\"."}], "task": "refering", "answer_template": "The \"a young woman with spiky hair and a denim overshirt holding a polka dot umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 63, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316], "bbox": [0.585765625, 0.06444444444444444, 0.8042031250000001, 0.9851358024691359], "iscrowd": 0, "area": 31983.166650000003, "rle": {"size": [405, 640], "counts": "Rad49T<=D=C=C;E7I7M4L3M3M3M3M3kNRN`GQ2[8W1J7J5K5K6M2O2M2O2N1N3N1N3N2N1N3N1N3N1M4J5K6J5K6J5N3N2N101N1O2N1O2N7lMlHZN[7U1PIjNW7g0SIWOS7<UICQ7L[I4h6[OfId0_6kNnIT1k8N1O1O1O2N1O1O101N1O1O11O010O001O010O001O010O001O001O001O1O001O001O001O001O2N4L3M4L3M3M4L3M4bLkMTLY2c3oMZLT2^3TN_LP2X3YNdLj1T3^NiLf1n2cNnL`1k2gNRM]1i2hNSM[1i2iNTMY1i2kNTMW1h2nNUMS1i2QOTMQ1i2ROUMo0h2VOUMl0h2XOUMi0i2[OTMg0i2]OPMi0m2ZOjLn0U3UOaLT1\\3POZLX1e3kNRL]1k3h3N3N2M2N3N2M5K4J7H8Gh0XO6J6J6POQ1J5L4K5LiXa1"}, "label": "a young woman with spiky hair and a denim overshirt holding a polka dot umbrella"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white polka - dotted umbrella being held by a young woman wearing a blue denim jacket and gray leggings\"."}], "task": "refering", "answer_template": "The \"a black and white polka - dotted umbrella being held by a young woman wearing a blue denim jacket and gray leggings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 182, 183, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274], "bbox": [0.71084375, 0.493283950617284, 0.9861093750000001, 0.8587407407407407], "iscrowd": 0, "area": 13321.6041, "rle": {"size": [405, 640], "counts": "Ted52b<2N1O2O0O2N2N101N2N1O2O1N1O2N1O2O1N1O2N2O0M4L4M2M4M2O2N2N1O2O1N100O101N100O100O2O0O100O1gMQNlIP2R6SNkIn1T6UNjIk1U6WNiIj1X6WNfIi1\\6WNaIk1_6VN_Ij1c6VNZIk1h6UNVIk1l6TNRIm1o6TNoHm1R7SNkHn1W7RNgHn1[7RNbHo1_7RN_Hn1c7QN[HP2g7PNWHQ2j7oMSHR2n7oMPHQ2R8oMkGR2W8nMgGR2[8mMcGU2]8lM`GU2b8kM[GV2g8jMVGW2l8iMQGX2Q9hMlFX2V9hMhFY2Z9gMcFZ2_9=2N1N3N2NO1O1O001O1O1O1O1O001O1O1O10O010000000O10O10O100O1000O0100O100O01000O100O10O10O10O01000O2O2M2OJcFVMZ9k2hFUMV9j2lFVMS9i2PGWMm8i2UGXMi8g2ZGXMd8h2^GYM`8g2bGYM[8g2j0O001O1O010O001O001O010O1O001O001O001O00001O001O1O2N1O1O1O1O2N2N2N2N2N2N2N2N2L4[Oe0[O][3"}, "label": "a black and white polka - dotted umbrella being held by a young woman wearing a blue denim jacket and gray leggings"}]}
{"id": 293853, "image": "COCO_train2014_000000293853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black - and - white polka dot umbrella held downward\"."}], "task": "refering", "answer_template": "The \"a black - and - white polka dot umbrella held downward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 182, 183, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274], "bbox": [0.71084375, 0.493283950617284, 0.9861093750000001, 0.8587407407407407], "iscrowd": 0, "area": 13321.6041, "rle": {"size": [405, 640], "counts": "Ted52b<2N1O2O0O2N2N101N2N1O2O1N1O2N1O2O1N1O2N2O0M4L4M2M4M2O2N2N1O2O1N100O101N100O100O2O0O100O1gMQNlIP2R6SNkIn1T6UNjIk1U6WNiIj1X6WNfIi1\\6WNaIk1_6VN_Ij1c6VNZIk1h6UNVIk1l6TNRIm1o6TNoHm1R7SNkHn1W7RNgHn1[7RNbHo1_7RN_Hn1c7QN[HP2g7PNWHQ2j7oMSHR2n7oMPHQ2R8oMkGR2W8nMgGR2[8mMcGU2]8lM`GU2b8kM[GV2g8jMVGW2l8iMQGX2Q9hMlFX2V9hMhFY2Z9gMcFZ2_9=2N1N3N2NO1O1O001O1O1O1O1O001O1O1O10O010000000O10O10O100O1000O0100O100O01000O100O10O10O10O01000O2O2M2OJcFVMZ9k2hFUMV9j2lFVMS9i2PGWMm8i2UGXMi8g2ZGXMd8h2^GYM`8g2bGYM[8g2j0O001O1O010O001O001O010O1O001O001O001O00001O001O1O2N1O1O1O1O2N2N2N2N2N2N2N2N2L4[Oe0[O][3"}, "label": "a black - and - white polka dot umbrella held downward"}]}
{"id": 80864, "image": "COCO_train2014_000000080864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the top hat\"."}], "task": "refering", "answer_template": "The \"the man in the top hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 133, 134, 135, 156, 157, 158, 178, 179, 180, 181, 182, 201, 202, 203, 204, 224, 225, 226, 248], "bbox": [0.767578125, 0.30028103044496485, 0.916515625, 0.6841451990632318], "iscrowd": 0, "area": 9089.511250000001, "rle": {"size": [427, 640], "counts": "`n\\64V=9F5K4[DKn9:dE2Z:`1N3M3M2O2M2N3M3M2N3M2N3M2M3N2M3N3M2M31O001N2OO0O2]OXLbGh3_8XL`Gi3_8XL`Gi3`8XL^Gh3c8XL\\Gi3c8YLWGk3[8dLaG`3_8e000010O001O01O01O2O1N2N1O2O1N2N2M4L3M3M3K5I8G:J5K5K5L3L4L4M3O1O1O1O1O1N2O1O1O1O1N2O1O1O0O2O2M3N2M3N3L3N2M3N2M3N6I7J5Jhle0"}, "label": "the man in the top hat"}]}
{"id": 80864, "image": "COCO_train2014_000000080864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"old man sitting next to his wife in a horse - drawn sleigh\"."}], "task": "refering", "answer_template": "The \"old man sitting next to his wife in a horse - drawn sleigh\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 133, 134, 135, 156, 157, 158, 178, 179, 180, 181, 182, 201, 202, 203, 204, 224, 225, 226, 248], "bbox": [0.767578125, 0.30028103044496485, 0.916515625, 0.6841451990632318], "iscrowd": 0, "area": 9089.511250000001, "rle": {"size": [427, 640], "counts": "`n\\64V=9F5K4[DKn9:dE2Z:`1N3M3M2O2M2N3M3M2N3M2N3M2M3N2M3N3M2M31O001N2OO0O2]OXLbGh3_8XL`Gi3_8XL`Gi3`8XL^Gh3c8XL\\Gi3c8YLWGk3[8dLaG`3_8e000010O001O01O01O2O1N2N1O2O1N2N2M4L3M3M3K5I8G:J5K5K5L3L4L4M3O1O1O1O1O1N2O1O1O1O1N2O1O1O0O2O2M3N2M3N3L3N2M3N2M3N6I7J5Jhle0"}, "label": "old man sitting next to his wife in a horse - drawn sleigh"}]}
{"id": 170980, "image": "COCO_train2014_000000170980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red teddy bear between 2 teddy bears on the left and 3 on the right\"."}], "task": "refering", "answer_template": "The \"a red teddy bear between 2 teddy bears on the left and 3 on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 59, 60, 61, 62, 77, 78, 79, 80, 95, 96, 97, 98, 112, 113, 114, 115, 116, 130, 131, 132, 133, 134, 150, 151, 152, 168, 169, 170, 187], "bbox": [0.21403999999999998, 0.21797333333333332, 0.50224, 0.7910133333333333], "iscrowd": 0, "area": 13017.059650000001, "rle": {"size": [375, 500], "counts": "`\\W11d;4M2M4M2N3K5M2O2M2N0001O01O01O01O0001O01O0001O01O01O01O0001O01O01OdMXOjIi0T6[OiIe0W6^OfIb0Z6AdI>\\6EaI<^6G_I9a6J]I5c6NZI3e60XI0h63UIMk65TIJl67SIJl67SIIn67RIHo68PIIQ76nHJS76lHJU76kHIW78fHI[77cHI_78_HGIZOi6P1[IGFBl6i0[IEDJP7b0YIDB3S7:YIB^O>W71XIB[Oe0\\7IWIW1g6jNWIX1g6jNVIZ1h6gNUI]1i6dNTI_1j6cNTI`1i6bNTIa1j6`NTId1h6i1M2N3M2M3N1OkMjIkNT6U1QJhNl5X1ZJdNe5\\1`JaN^5^1hJ_NV5`1kJbNR5_1nJbNQ5]1PKeNn4[1RKgNl4X1UKhNj4Y1VKhNi4W1XKiNh4W1WKkNg4U1ZKkNf4T1[KmNd4S1\\KmNc4S1^KnNa4R1_KnN`4R1aKnN^4S1bKnN\\4R1eKnNZ4S1fKnNX4R1iKnNV4S1jKnNU4Q1lKoNS4Q1nKoNQ4R1nKnNR4R1oKmNQ4T1oKjNR4V1nKjNR4W1nKhNS4W1mKiNS4X1lKgNU4Y1kKgNU4Z1kKeNV4Z1jKeNY4TObJT1U1HZ4SOdJS1Q1J]4ROdJQ1P1L^4QOfJP1k0O`4POhJn0h04`4lNlJN@7T1o0`4kNoJFE>l0S1`4gNSK^OJf0b0V1b4dNkK3C[1a4bNmK2B\\1b4`NoK1_Oa1a4]NSLO\\Oe1b4XNVL2WOh1c4mM`L8nNk1n6SNQIo1o6oMQIS2P7kMPIU2R7hMnHZ2R7eMmH\\2V817H:G:E:G5J3M3N1N3N1N3M2O2H8Hnmj2"}, "label": "a red teddy bear between 2 teddy bears on the left and 3 on the right"}]}
{"id": 170980, "image": "COCO_train2014_000000170980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red teddy bears in a row\"."}], "task": "refering", "answer_template": "The \"red teddy bears in a row\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 54, 55, 56, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 101, 102, 103, 104, 105, 106, 107, 108, 109, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 155, 156, 157, 158, 159, 160, 161, 163, 164, 165, 174, 175, 176, 177, 178, 179, 181, 182, 183, 192, 193, 194, 195, 196, 197, 214, 215], "bbox": [0.0, 0.20853333333333335, 1.0, 0.8699466666666668], "iscrowd": 0, "area": 48591.22120000001, "rle": {"size": [375, 500], "counts": "j3=`:k0L5M2bFcN`8^1_GoNS8S1lGZOg7g0XH^Ob7e0\\H[Od7f0[H[Oc7g0\\HYOd7h0[HYOc7j0[HVOe7l0YHUOe7n0YHROf7P1YHQOf7P1YHoNg7T1WHiN3QO[7W2aHeN7YOR7T2fH`N:Al6P2iH\\N>Hd6n1mHVNc01[6j1PISNi06R6\\2oIiMl5W2TJmMg5T2ZJoMb5Q2^JRN^5n1cJTN[5l1eJWNW5j1jJXNS5h1mJZNP5f1QK]Nl4c1TK`Nj4_1WKbNg4^1YKeNd4[1\\KhNa4W1`KkN_4T1aKoN\\4Q1dKROY4n0gKTOW4k0jKXOT4g0lK\\OQ4d0oK^Oo3b0QLAl3?TLDj3:WLHg38YLIf37ZLIf37YLJh34YLLg34YLLg34XLMh33XLMi32WLNi31WL0i30WL0i30WL0j3OUL1l3NUL2k3NTL2m3NSL2n3MRL2o3NPL3P4LQL4o3LQL3Q4LnK5R4SOoJgNo0R2V4VOlJgNo0n1[4SNdJ?3@m0j1a4UNeJ<NEl0h1d4UNhJ7JLi0g1h4SN[KEUOa0h0f1i4RNiK6_Of1k4RNgK8]Oe1o4PNfK;[Of1o4lMhK>YOg1n4jMiK?YOi1g6XNYIi1f6WNZIj1d6WN[Il1b6UN^IFCX1n6QO_Jo0_5RObJn0\\5SOeJl0[5TOfJl0X5TOjJl0T5UOlJl0S5TOnJk0S5TOnJl0S5ROnJn0S5oNoJQ1R5mNoJR1R5mNoJS1R5kNoJU1R5hNPKX1Q5fNPKY1R5eNPKX1S5fNoJW1T5gNnJV1l7M3M3M4L3M3M3M3M3N2Md]i19SbVN9H000O10000O10000O10O010000O10000O01000O01000O0100O0100O0100O01QNTOSIl0k6YORIf0n6]OoHd0P7@mH`0R7ClH<S7GkH:T7GlH8T7JkH6T7KkH6S7LmH3S7OlH1S70lH0T71lHOT72iH0W70hH0Y71eH0[70dH1]7NbH3^7N`H3R7VOQHi0i04T7SOTHl0d02X7ROTHn0a02Z7POUHP1>1]7oNUHR1;0`7nNUHT180b7lNVHV140f7jNWHX1OO>mN[6LXIZ1Lo0l6fMYIQ4e6PL[IP4e6PL[IP4e6PL\\IP4c6b00001O00001aITKP6]5M3M3M4L3M3M3M3M4K4M3M1O1O1O1O1O1O1N2O1O2N1O1O1O00000O11O00001O000O10000O10000O10000fN^K]Kd4d4]KWKf4i4[KTKg4m4ZKoJh4R5YKjJi4b1jJd1>fLj4^1PKm16aLm4[1TKU20\\Ln4W1ZK]2X5^MmJc2R5ZMQKg2n4VMUKk2j4RM`JVO;i3T5oLbJZO9h3T5lLcJ_O7f3V5iLcJD6d3\\5`L_JO3b3c5VL[J:2a3U6`LkIa3R6bLmI_3Q6bLoI^3o5eLPJ\\3n5eLRJ\\3m5dLSJ]3k5eLTJ\\3k5dLUJ]3j5dLUJ\\3j5eLVJ\\3i5eLVJ\\3h5eLWJ]3h5cLWJ_3h5Y1N1L[JVJf5i5[JVJf5i5ZJWJg5h5YJXJh5f5ZJYJg5f5YJZJg5f56O1N1000000O10000O1000000O1000000O10000O1000000O1000000O101O1N2O001O1N2O001O1N2O001O1N2O001O1O10O01O10O01O1O010O1O010O01O0mJ\\KS3d4hLaKX3`4aLeK`3[4QLTLn3m3QLTLo3m3oKTLQ4l3nKTLS4m3jKULV4k3hKWLX4i3fKXLY4k3dKWLZ4k3bKYL\\4i3`KZL_4i3\\K]La4d3[KcL`4\\5M2N2N2N3N10f2ZMPH"}, "label": "red teddy bears in a row"}]}
{"id": 170980, "image": "COCO_train2014_000000170980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third teddy bear from the right\"."}], "task": "refering", "answer_template": "The \"the third teddy bear from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 79, 80, 81, 82, 98, 99, 100, 101, 115, 116, 117, 118, 119, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 171, 172, 173, 174, 189, 190, 191], "bbox": [0.36278, 0.22674666666666668, 0.6829, 0.8343466666666667], "iscrowd": 0, "area": 16427.77264999999, "rle": {"size": [375, 500], "counts": "__R21f;6I7J6I7J1N100O100O100O100O100O100O100O100O100O100O100O1O01O001O010O001OhMoNkIQ1U6SOgIn0X6WOdIh0]6[O_Ie0a6@[I?e6EWI;i6ITI7k6JTI6m6JSI5m6LRI4n6MQI4n6MRI2n6OQI1P7OPI0P71oHOQ72oHNP73oHMR7`0aH_O_7d0^H\\Oc7e0[H[Oe7g0ZHYOf7h0YHWOg7k0XHDY7=gH8c6I\\I;_6G`I;^6FaI<\\6GbI;\\6FdI;Y6GfI;X6FgI<V6FiI<U6EjI>S6CkIa0Q6AmIc0P6^OnIf0n5d2N2M3N2M3N2N2M`McJjN[5V1iJhNT5X1QKfNm4Z1XKcNe4]1`KaN^4_1cKbNZ4^1gKdNW4\\1iKfNU4Y1lKiNS4V1mKlNQ4S1PLnNo3R1RLmNo3Q1RLPOm3P1SLPOm3o0TLROl3m0TLSOl3l0ULTOk3l0ULUOk3i0VLWOj3i0ULYOj3f0WLZOj3e0VL[Oj3d0VL^Oj3a0VL_Oj3`0WL@i3`0VL@k3?VLAk3>ULBk3=ULDk3<ULDk3;VLEk3:TLGl38ULGl39TLGl38ULHl37SLJm3VNSKo0Q1i0n3WNVKl0l02QOjNP56WKj0h07VOaNn4=XKh0d0;ZOYNl4c0ZK_OEf0l0P1h4iNXK7OR1i4eNZK8MT1i4bN[K9MV1h4_N]K:JY1i4[N^K;J[1h4XN`K<H]1i4TN`K>G`1k4nM`Ka0Eb1n4hM^Ke0Ed1^6[NbIf1]6YNcIi1]6VNcIk1\\6TNeIm1[6RNdIo1\\6QNdI6mNP1`7hNdI7oNP1\\7iNaJV1`5iNaJV1_5iNdJU1]5jNdJV1[5jNgJT1Y5kNiJT1X5kNjJS1V5mNkJR1V5mNlJP1V5nNlJQ1V5mNkJR1X5kNjJS1n7O1O2N1N3M2M3N3L3N2N6Ieni1"}, "label": "the third teddy bear from the right"}]}
{"id": 170980, "image": "COCO_train2014_000000170980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fourth teddy bear from the left\"."}], "task": "refering", "answer_template": "The \"the fourth teddy bear from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 79, 80, 81, 82, 98, 99, 100, 101, 115, 116, 117, 118, 119, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 171, 172, 173, 174, 189, 190, 191], "bbox": [0.36278, 0.22674666666666668, 0.6829, 0.8343466666666667], "iscrowd": 0, "area": 16427.77264999999, "rle": {"size": [375, 500], "counts": "__R21f;6I7J6I7J1N100O100O100O100O100O100O100O100O100O100O100O1O01O001O010O001OhMoNkIQ1U6SOgIn0X6WOdIh0]6[O_Ie0a6@[I?e6EWI;i6ITI7k6JTI6m6JSI5m6LRI4n6MQI4n6MRI2n6OQI1P7OPI0P71oHOQ72oHNP73oHMR7`0aH_O_7d0^H\\Oc7e0[H[Oe7g0ZHYOf7h0YHWOg7k0XHDY7=gH8c6I\\I;_6G`I;^6FaI<\\6GbI;\\6FdI;Y6GfI;X6FgI<V6FiI<U6EjI>S6CkIa0Q6AmIc0P6^OnIf0n5d2N2M3N2M3N2N2M`McJjN[5V1iJhNT5X1QKfNm4Z1XKcNe4]1`KaN^4_1cKbNZ4^1gKdNW4\\1iKfNU4Y1lKiNS4V1mKlNQ4S1PLnNo3R1RLmNo3Q1RLPOm3P1SLPOm3o0TLROl3m0TLSOl3l0ULTOk3l0ULUOk3i0VLWOj3i0ULYOj3f0WLZOj3e0VL[Oj3d0VL^Oj3a0VL_Oj3`0WL@i3`0VL@k3?VLAk3>ULBk3=ULDk3<ULDk3;VLEk3:TLGl38ULGl39TLGl38ULHl37SLJm3VNSKo0Q1i0n3WNVKl0l02QOjNP56WKj0h07VOaNn4=XKh0d0;ZOYNl4c0ZK_OEf0l0P1h4iNXK7OR1i4eNZK8MT1i4bN[K9MV1h4_N]K:JY1i4[N^K;J[1h4XN`K<H]1i4TN`K>G`1k4nM`Ka0Eb1n4hM^Ke0Ed1^6[NbIf1]6YNcIi1]6VNcIk1\\6TNeIm1[6RNdIo1\\6QNdI6mNP1`7hNdI7oNP1\\7iNaJV1`5iNaJV1_5iNdJU1]5jNdJV1[5jNgJT1Y5kNiJT1X5kNjJS1V5mNkJR1V5mNlJP1V5nNlJQ1V5mNkJR1X5kNjJS1n7O1O2N1N3M2M3N3L3N2N6Ieni1"}, "label": "the fourth teddy bear from the left"}]}
{"id": 236519, "image": "COCO_train2014_000000236519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with blonde hair in a gray jacket standing next to railroad tracks\"."}], "task": "refering", "answer_template": "The \"man with blonde hair in a gray jacket standing next to railroad tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 58, 59, 60, 71, 72, 73, 83, 84, 85, 86, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 123, 124, 125, 136, 137, 138, 149, 150, 151, 162, 163, 164, 175, 176, 177, 188, 189, 190, 201, 202, 203, 204, 217], "bbox": [0.43274666666666667, 0.19654, 0.7465066666666667, 0.89884], "iscrowd": 0, "area": 20948.501249999998, "rle": {"size": [500, 375], "counts": "VZ_28d>j0H8H8L5K4L4K5K5K5K5bCeM\\;`2ZDiMb;\\2UDlMh;R3O1O9F9H?@V1kN4K6K4K6K5J5kHUIY5P7]JVIa5o6YJTIf5P7TJSIk5P7oITIo5P7lISIS6X8N2O1N10NXORGmJn8o4YGoJe8n4aGPK_8n4fGPKY8o4jGoJU8R5mGmJR8R5UNgJkKZ5R4jKmJV4o4YLeJg3Y5fL]JZ3_5UMVJl2f5b3L5L3L5L32O1O0O200O2N3N1N3N1N3M2O1N3M2O2M4M3LVJgG\\4W8_KmGe4Q8YKPHj4n7TKRHP5m7nJRHV5l7eJXH^5g7\\J\\Hh5b7RJcHP6]7cImHe6m7;F;D:G9F:F4M0O2O0OXKRIk0m6iNfIQ1Z6dNUJX1j5gN[JV1e5hN`JV1^5iNgJT1X5kNnJQ1T5jNbKc0c4VORL8Q4BdLKa3NSM@Q39eMTO_2c0lMWOY2>mMCV23PNLV2HQN7T2_ORN`0b8O1O1O1O1O1O1O1O1O1O3M3M3M3M]n]1"}, "label": "man with blonde hair in a gray jacket standing next to railroad tracks"}]}
{"id": 236519, "image": "COCO_train2014_000000236519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a gray jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing a gray jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 58, 59, 60, 71, 72, 73, 83, 84, 85, 86, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 123, 124, 125, 136, 137, 138, 149, 150, 151, 162, 163, 164, 175, 176, 177, 188, 189, 190, 201, 202, 203, 204, 217], "bbox": [0.43274666666666667, 0.19654, 0.7465066666666667, 0.89884], "iscrowd": 0, "area": 20948.501249999998, "rle": {"size": [500, 375], "counts": "VZ_28d>j0H8H8L5K4L4K5K5K5K5bCeM\\;`2ZDiMb;\\2UDlMh;R3O1O9F9H?@V1kN4K6K4K6K5J5kHUIY5P7]JVIa5o6YJTIf5P7TJSIk5P7oITIo5P7lISIS6X8N2O1N10NXORGmJn8o4YGoJe8n4aGPK_8n4fGPKY8o4jGoJU8R5mGmJR8R5UNgJkKZ5R4jKmJV4o4YLeJg3Y5fL]JZ3_5UMVJl2f5b3L5L3L5L32O1O0O200O2N3N1N3N1N3M2O1N3M2O2M4M3LVJgG\\4W8_KmGe4Q8YKPHj4n7TKRHP5m7nJRHV5l7eJXH^5g7\\J\\Hh5b7RJcHP6]7cImHe6m7;F;D:G9F:F4M0O2O0OXKRIk0m6iNfIQ1Z6dNUJX1j5gN[JV1e5hN`JV1^5iNgJT1X5kNnJQ1T5jNbKc0c4VORL8Q4BdLKa3NSM@Q39eMTO_2c0lMWOY2>mMCV23PNLV2HQN7T2_ORN`0b8O1O1O1O1O1O1O1O1O1O3M3M3M3M]n]1"}, "label": "a man wearing a gray jacket"}]}
{"id": 498669, "image": "COCO_train2014_000000498669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a barge in the background\"."}], "task": "refering", "answer_template": "The \"a barge in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 26, 27, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 84, 85, 86, 87, 88, 89, 90, 91, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114], "bbox": [0.161796875, 0.0, 0.9994375, 0.2786458333333333], "iscrowd": 0, "area": 50202.897750000004, "rle": {"size": [480, 640], "counts": "Zi`11P?1N1O2N2O0O2N101N1O2O1N1O2O0O2N1O2O1N1O2O0O2N101N2N101N1O2N101N2N101N1O2O0O1O100O1O1O010O1O100O1O010O1O100O1O00100O1O100O00100O0001O01O0001O00001O00010O0000001ON2N2N2N2N2N3M2N2N2N2N2N2N2N2YOZMREh2n:XMPEj2P;VMnDl2S;TMjDn2V;RMhDP3X;PMfDR3Z;nLdDT3\\;mLaDU3_;kL_DX3`;40000001O000000000000000000001O01O0000000000000001O0000000000000001O0001O0000000000000000001O00000001O000000000001O00000000000000000010O00000000000000000001O000000000001O0001O000000000000001O000000000001O01O000000000000001O00000000000000001O000000000000001O000000000000001O000000000001O01O000000000000001\\OVDPNj;n1]DmMc;R2dDhM\\;W2jDdMV;Z2i0O1N2O1N3N1O1N2O1N2O1N2O2M0001O0010O010O10O0100O01000000O01000O100000O01000011O10O01O1O10O01O1O00100O1O002O6I7I8H7I7J7H7I4L01O0001O0001O0000001O01O000001O00000010O00000001O0000010O000000001O0001O01O000000001O01O0001O000000010O0000001O000001O01O0000001O0001O00O2O2N]OeLbEZ3^:iLaEU3_:nL`EP3`:SM^El2b:XM\\Ef2d:]M[Ea2e:bMZE\\2f:gMXEW2i:lMVER2j:QNUEm1k:VNSEi1l:[NSEc1m:`NRE^1n:eNPEZ1P;[100000O1000000000O10000000O10000000000000O1000O1000000000000000000O010000000000000000000O10O1000000000000001N100000001O000000001N1000001O000000001O0O10001O000000001O00000O101O0000001O000000001O00000O2O000000001O000000001O000O101O0000000YN_DOb;IgD5Y;EoD9Q;AWE;k:_O\\E<i:]O_E<f:^ObE<b:^OfE<c<J5K6J\\M"}, "label": "a barge in the background"}]}
{"id": 498669, "image": "COCO_train2014_000000498669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large boat\"."}], "task": "refering", "answer_template": "The \"large boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 26, 27, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 84, 85, 86, 87, 88, 89, 90, 91, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114], "bbox": [0.161796875, 0.0, 0.9994375, 0.2786458333333333], "iscrowd": 0, "area": 50202.897750000004, "rle": {"size": [480, 640], "counts": "Zi`11P?1N1O2N2O0O2N101N1O2O1N1O2O0O2N1O2O1N1O2O0O2N101N2N101N1O2N101N2N101N1O2O0O1O100O1O1O010O1O100O1O010O1O100O1O00100O1O100O00100O0001O01O0001O00001O00010O0000001ON2N2N2N2N2N3M2N2N2N2N2N2N2N2YOZMREh2n:XMPEj2P;VMnDl2S;TMjDn2V;RMhDP3X;PMfDR3Z;nLdDT3\\;mLaDU3_;kL_DX3`;40000001O000000000000000000001O01O0000000000000001O0000000000000001O0001O0000000000000000001O00000001O000000000001O00000000000000000010O00000000000000000001O000000000001O0001O000000000000001O000000000001O01O000000000000001O00000000000000001O000000000000001O000000000000001O000000000001O01O000000000000001\\OVDPNj;n1]DmMc;R2dDhM\\;W2jDdMV;Z2i0O1N2O1N3N1O1N2O1N2O1N2O2M0001O0010O010O10O0100O01000000O01000O100000O01000011O10O01O1O10O01O1O00100O1O002O6I7I8H7I7J7H7I4L01O0001O0001O0000001O01O000001O00000010O00000001O0000010O000000001O0001O01O000000001O01O0001O000000010O0000001O000001O01O0000001O0001O00O2O2N]OeLbEZ3^:iLaEU3_:nL`EP3`:SM^El2b:XM\\Ef2d:]M[Ea2e:bMZE\\2f:gMXEW2i:lMVER2j:QNUEm1k:VNSEi1l:[NSEc1m:`NRE^1n:eNPEZ1P;[100000O1000000000O10000000O10000000000000O1000O1000000000000000000O010000000000000000000O10O1000000000000001N100000001O000000001N1000001O000000001O0O10001O000000001O00000O101O0000001O000000001O00000O2O000000001O000000001O000O101O0000000YN_DOb;IgD5Y;EoD9Q;AWE;k:_O\\E<i:]O_E<f:^ObE<b:^OfE<c<J5K6J\\M"}, "label": "large boat"}]}
{"id": 539632, "image": "COCO_train2014_000000539632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue directly behind the food\"."}], "task": "refering", "answer_template": "The \"woman in blue directly behind the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 25, 26, 27, 28, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 75, 76, 79, 80, 92, 93, 109, 110, 126, 127, 142, 143], "bbox": [0.39639583333333334, 0.009015625, 0.7807708333333333, 0.358109375], "iscrowd": 0, "area": 15553.7144, "rle": {"size": [640, 480], "counts": "Tof31mc02N1O2N2M2O2M3_O`0@a0_Oa0_O`0@a0D<N2O0O2O1N2O1N1O2O1N2O1N2O0O2O1N2N2O0O2O1N2O1N101N2NQOd@^M[?a2o@WMo>i2ZAPMb>R3gAfLS>_3VBZLc=h1[A6f?1_@I\\?=i@]OR?i0SAQOh>U1]AeN_>`1gAZNY>f1lATNU>k1QBoMP>Q2UBiMl=W2YBcMh=]2]B]Md=c2l1000000000001O1O1O1N2O10000000000O10000000000O10000000000O1000000000000O1001O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O4L5K4L4L4L5K4L4L4L4LO1N2N2M3N2N2N2N2N2N9H:G9F;F9G9F2O0O2O001N101O0O2O0O2O1O0O2O001N101N101O0O2O001N101N102N3L4M2N3L4M2N3L4M2M4M3M2Mf`Q2"}, "label": "woman in blue directly behind the food"}]}
{"id": 539632, "image": "COCO_train2014_000000539632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a purple dress\"."}], "task": "refering", "answer_template": "The \"a girl in a purple dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 25, 26, 27, 28, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 75, 76, 79, 80, 92, 93, 109, 110, 126, 127, 142, 143], "bbox": [0.39639583333333334, 0.009015625, 0.7807708333333333, 0.358109375], "iscrowd": 0, "area": 15553.7144, "rle": {"size": [640, 480], "counts": "Tof31mc02N1O2N2M2O2M3_O`0@a0_Oa0_O`0@a0D<N2O0O2O1N2O1N1O2O1N2O1N2O0O2O1N2N2O0O2O1N2O1N101N2NQOd@^M[?a2o@WMo>i2ZAPMb>R3gAfLS>_3VBZLc=h1[A6f?1_@I\\?=i@]OR?i0SAQOh>U1]AeN_>`1gAZNY>f1lATNU>k1QBoMP>Q2UBiMl=W2YBcMh=]2]B]Md=c2l1000000000001O1O1O1N2O10000000000O10000000000O10000000000O1000000000000O1001O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O4L5K4L4L4L5K4L4L4L4LO1N2N2M3N2N2N2N2N2N9H:G9F;F9G9F2O0O2O001N101O0O2O0O2O1O0O2O001N101N101O0O2O001N101N102N3L4M2N3L4M2N3L4M2M4M3M2Mf`Q2"}, "label": "a girl in a purple dress"}]}
{"id": 539632, "image": "COCO_train2014_000000539632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing green t shirt ..\"."}], "task": "refering", "answer_template": "The \"a girl wearing green t shirt ..\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 53, 54, 55, 56, 71, 72, 73, 87, 88, 89, 104, 105, 106, 121, 122, 123, 124, 138, 139, 140, 141, 155, 156, 157, 158, 173, 174], "bbox": [0.1255625, 0.0944375, 0.3318333333333333, 0.45765625], "iscrowd": 0, "area": 15982.380649999997, "rle": {"size": [640, 480], "counts": "RfU1i0Qc09H8G<Dc0Ab0]Od0]O2N3L3N2M3O1N3M2O1jMXLRDj3m;ZLnCh3P<\\LmCe3R<_LiCd3T<aLhC`3W<cLeC_3Y<fLbC\\3\\<hLaCZ3]<jL^CX3`<lL\\CW3a<mL\\CT3b<QMYCR3f<PMWCQ3i<QMSCQ3m<R2O001O001O001O00001O000001O0001O00000001O01O01O0001O00001O0000001O00M3N2M3N3L4K5E;N1PLaB_1`=[NTCU1n<fNXCX1i<eN\\CW1f<gN^CV1c<iN`CS1b<kNbCR1_<lNdCQ1^<mNfCP1Z<oNiCn0Y<POkCl0W<ROlCk0V<SOnCi0S<VOPDh0Q<VOTDe0P<WO\\D>g;^O_D<d;@`D>c;YOeDd0^?N2M3N2M3N2M3K5Lh\\X6"}, "label": "a girl wearing green t shirt .."}]}
{"id": 539632, "image": "COCO_train2014_000000539632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing in a green tshirt\"."}], "task": "refering", "answer_template": "The \"woman standing in a green tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 53, 54, 55, 56, 71, 72, 73, 87, 88, 89, 104, 105, 106, 121, 122, 123, 124, 138, 139, 140, 141, 155, 156, 157, 158, 173, 174], "bbox": [0.1255625, 0.0944375, 0.3318333333333333, 0.45765625], "iscrowd": 0, "area": 15982.380649999997, "rle": {"size": [640, 480], "counts": "RfU1i0Qc09H8G<Dc0Ab0]Od0]O2N3L3N2M3O1N3M2O1jMXLRDj3m;ZLnCh3P<\\LmCe3R<_LiCd3T<aLhC`3W<cLeC_3Y<fLbC\\3\\<hLaCZ3]<jL^CX3`<lL\\CW3a<mL\\CT3b<QMYCR3f<PMWCQ3i<QMSCQ3m<R2O001O001O001O00001O000001O0001O00000001O01O01O0001O00001O0000001O00M3N2M3N3L4K5E;N1PLaB_1`=[NTCU1n<fNXCX1i<eN\\CW1f<gN^CV1c<iN`CS1b<kNbCR1_<lNdCQ1^<mNfCP1Z<oNiCn0Y<POkCl0W<ROlCk0V<SOnCi0S<VOPDh0Q<VOTDe0P<WO\\D>g;^O_D<d;@`D>c;YOeDd0^?N2M3N2M3N2M3K5Lh\\X6"}, "label": "woman standing in a green tshirt"}]}
{"id": 23538, "image": "COCO_train2014_000000023538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small cup of coleslaw on a plate next to a slice of pickle\"."}], "task": "refering", "answer_template": "The \"a small cup of coleslaw on a plate next to a slice of pickle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 291, 292, 293, 294], "bbox": [0.595796875, 0.4675233644859813, 0.85634375, 0.8389719626168224], "iscrowd": 0, "area": 19099.646899999996, "rle": {"size": [428, 640], "counts": "i_o41T=9G9H8G:G8L4M3N3L3N2N2M4M2N2M3N3M2M3N3L3N2N2M4M2N2M3N3M2M3N2M4M2N2M3N3M2M3N3M1N2O0O2O001N10001N101O001O001O00001O001O001O001O00001O0O2O001O001O00001O001O001O0000O10000O100000O010000O1000000O2O000O2O00000O2O000O2O00001N10000O2O00001N10001N10001O0O10001N10001N1M3M4L3M3M3M4L3M3M4L3M3M4L3M3N2M4L5K4L5K5K4L5K4M4L4L3M4M2M4L4L3M4L3M4M3L3MjVV1"}, "label": "a small cup of coleslaw on a plate next to a slice of pickle"}]}
{"id": 23538, "image": "COCO_train2014_000000023538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small plastic cup containing a side of cole slaw salad\"."}], "task": "refering", "answer_template": "The \"a small plastic cup containing a side of cole slaw salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 291, 292, 293, 294], "bbox": [0.595796875, 0.4675233644859813, 0.85634375, 0.8389719626168224], "iscrowd": 0, "area": 19099.646899999996, "rle": {"size": [428, 640], "counts": "i_o41T=9G9H8G:G8L4M3N3L3N2N2M4M2N2M3N3M2M3N3L3N2N2M4M2N2M3N3M2M3N2M4M2N2M3N3M2M3N3M1N2O0O2O001N10001N101O001O001O00001O001O001O001O00001O0O2O001O001O00001O001O001O0000O10000O100000O010000O1000000O2O000O2O00000O2O000O2O00001N10000O2O00001N10001N10001O0O10001N10001N1M3M4L3M3M3M4L3M3M4L3M3M4L3M3N2M4L5K4L5K5K4L5K4M4L4L3M4M2M4L4L3M4L3M4M3L3MjVV1"}, "label": "a small plastic cup containing a side of cole slaw salad"}]}
{"id": 23538, "image": "COCO_train2014_000000023538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"some kind of brown sauce in a small plastic cup\"."}], "task": "refering", "answer_template": "The \"some kind of brown sauce in a small plastic cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 149, 150, 151, 152, 172, 173, 174, 175], "bbox": [0.46778125, 0.26838785046728975, 0.667015625, 0.5248598130841121], "iscrowd": 0, "area": 10351.933099999997, "rle": {"size": [428, 640], "counts": "nSm31Y=4L3M4L4L4L4L3M4L4L4L6J7I7K4K6K4L5J4M1O0O2O1O0O2O0O2O001N2O001N101O1N101N10000O2O000O100O10000O2O00000000000001O01O0000000001O000001O00000001O000001O00000001O00000001OO1O2O0O1O2O0O1O101N1O101N100O2N100O1O2L3N2M4L3M3M4M2M3M4M2N2O3L3M4M3L3N3L4M2M4M3L3N3L4MWlh2"}, "label": "some kind of brown sauce in a small plastic cup"}]}
{"id": 23538, "image": "COCO_train2014_000000023538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of brown sauce in a transparent cup sitting on top of a beef steak\"."}], "task": "refering", "answer_template": "The \"a cup of brown sauce in a transparent cup sitting on top of a beef steak\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 149, 150, 151, 152, 172, 173, 174, 175], "bbox": [0.46778125, 0.26838785046728975, 0.667015625, 0.5248598130841121], "iscrowd": 0, "area": 10351.933099999997, "rle": {"size": [428, 640], "counts": "nSm31Y=4L3M4L4L4L4L3M4L4L4L6J7I7K4K6K4L5J4M1O0O2O1O0O2O0O2O001N2O001N101O1N101N10000O2O000O100O10000O2O00000000000001O01O0000000001O000001O00000001O000001O00000001O00000001OO1O2O0O1O2O0O1O101N1O101N100O2N100O1O2L3N2M4L3M3M4M2M3M4M2N2O3L3M4M3L3N3L4M2M4M3L3N3L4MWlh2"}, "label": "a cup of brown sauce in a transparent cup sitting on top of a beef steak"}]}
{"id": 23539, "image": "COCO_train2014_000000023539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first skier on the left\"."}], "task": "refering", "answer_template": "The \"first skier on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309], "bbox": [0.31179687500000003, 0.3184166666666667, 0.488765625, 0.8015625], "iscrowd": 0, "area": 14529.710850000005, "rle": {"size": [480, 640], "counts": "oSn22l>1N3N2N2N2N1O2N2L4I7I6[LPOUIW1i6XObHP1\\7]ORHj0l7CbGd0[8JSG<l81bF5^9g20O1O010O010O1O010O1O010O1O0O2O0O2O1O0O2O1N101M3N1N3M2O2M3N1cN^13L3N3L4L3M4L3M4L4L3M6J6K5H8Fc0_Ok0VO2N3L3N2N2M4M2M3N2N3L3N2N2M4O0000001OWMlHUMf7f2\\2L4L4L4L4L4L4M42M2N2N2O101N1O100O1O01N1N3M2N2O2M2Jb1^No0QOa0_OTQi4"}, "label": "first skier on the left"}]}
{"id": 23539, "image": "COCO_train2014_000000023539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing all black standing taller than the person next to him\"."}], "task": "refering", "answer_template": "The \"man wearing all black standing taller than the person next to him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309], "bbox": [0.31179687500000003, 0.3184166666666667, 0.488765625, 0.8015625], "iscrowd": 0, "area": 14529.710850000005, "rle": {"size": [480, 640], "counts": "oSn22l>1N3N2N2N2N1O2N2L4I7I6[LPOUIW1i6XObHP1\\7]ORHj0l7CbGd0[8JSG<l81bF5^9g20O1O010O010O1O010O1O010O1O0O2O0O2O1O0O2O1N101M3N1N3M2O2M3N1cN^13L3N3L4L3M4L3M4L4L3M6J6K5H8Fc0_Ok0VO2N3L3N2N2M4M2M3N2N3L3N2N2M4O0000001OWMlHUMf7f2\\2L4L4L4L4L4L4M42M2N2N2O101N1O100O1O01N1N3M2N2O2M2Jb1^No0QOa0_OTQi4"}, "label": "man wearing all black standing taller than the person next to him"}]}
{"id": 23539, "image": "COCO_train2014_000000023539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier wearing a red jacket\"."}], "task": "refering", "answer_template": "The \"a skier wearing a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 335, 337], "bbox": [0.555296875, 0.31960416666666663, 0.741484375, 0.8589166666666667], "iscrowd": 0, "area": 16006.775899999995, "rle": {"size": [480, 640], "counts": "ggV55h>\\1cN6jK]NhJh1U5IWI=f6OlH6R71cH6Y72]H3`74^GQ1Y8k2O1O100O010O1O100O1O100\\OYInHg6P7^ImHc6Q7bIlHX6X7lIeHi5d7\\JYHX5R8mJkGj4\\8i05K5K4L5J5L5K4L5J5L5L46I8I6I7K5K6I6K5J6K5K6jGWIg7R72M3N3M2M3N3L3N1OO010000O10O10O100O10O[LlG?T8\\OiHKX7OfIYOZ6a0_JlNa5n0jJlNV5o0VKkNj4o0bKjN_4P1lKkNT4P1WLjNi3V1\\LdNd3]1]LbNc3^1]LbNc3_1\\L`Ne3`1\\L_Nd3a1\\L^Nd3c1m41O00000O1100O001O10O01O1O00100O010000O1001O001K4M4L3L5L3M3L5L3M4L3L5L3MbR]2"}, "label": "a skier wearing a red jacket"}]}
{"id": 23539, "image": "COCO_train2014_000000023539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man doing ice skating with maroon jarkin\"."}], "task": "refering", "answer_template": "The \"a man doing ice skating with maroon jarkin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313, 314, 335, 337], "bbox": [0.555296875, 0.31960416666666663, 0.741484375, 0.8589166666666667], "iscrowd": 0, "area": 16006.775899999995, "rle": {"size": [480, 640], "counts": "ggV55h>\\1cN6jK]NhJh1U5IWI=f6OlH6R71cH6Y72]H3`74^GQ1Y8k2O1O100O010O1O100O1O100\\OYInHg6P7^ImHc6Q7bIlHX6X7lIeHi5d7\\JYHX5R8mJkGj4\\8i05K5K4L5J5L5K4L5J5L5L46I8I6I7K5K6I6K5J6K5K6jGWIg7R72M3N3M2M3N3L3N1OO010000O10O10O100O10O[LlG?T8\\OiHKX7OfIYOZ6a0_JlNa5n0jJlNV5o0VKkNj4o0bKjN_4P1lKkNT4P1WLjNi3V1\\LdNd3]1]LbNc3^1]LbNc3_1\\L`Ne3`1\\L_Nd3a1\\L^Nd3c1m41O00000O1100O001O10O01O1O00100O010000O1001O001K4M4L3L5L3M3L5L3M4L3L5L3MbR]2"}, "label": "a man doing ice skating with maroon jarkin"}]}
{"id": 23539, "image": "COCO_train2014_000000023539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in the middle of two people\"."}], "task": "refering", "answer_template": "The \"a skier in the middle of two people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 334], "bbox": [0.44921875, 0.3737083333333333, 0.56640625, 0.8450624999999999], "iscrowd": 0, "area": 10629.6875, "rle": {"size": [480, 640], "counts": "a[W47g0LR=<iBGR=a0hBCR=e0hB^OS=i0hBZOm:LlF`1U9aNeFd1\\9[N_Fi1b9WNYFm1i9RNRFR2o9nMlET2W:mMUEb2P;c04L5K4L5L00N1O2N2N2TNSMeGn2^8dMlF^2U9VNTFl1n9a11O2M2O2N2N1O2J5I8G8IREVMo9h2oEVNY9V4D3L4M3L4M3M4K4M3M2O001O00001O00001O00001O]OWHoIi7n5gHfIY7]4dHiLU8l2^HiLf7Q3X2N101N101N3N4K5L6I8Fh0UOk1UN]bQ4"}, "label": "a skier in the middle of two people"}]}
{"id": 23539, "image": "COCO_train2014_000000023539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier standing in the middle of two men\"."}], "task": "refering", "answer_template": "The \"a skier standing in the middle of two men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 334], "bbox": [0.44921875, 0.3737083333333333, 0.56640625, 0.8450624999999999], "iscrowd": 0, "area": 10629.6875, "rle": {"size": [480, 640], "counts": "a[W47g0LR=<iBGR=a0hBCR=e0hB^OS=i0hBZOm:LlF`1U9aNeFd1\\9[N_Fi1b9WNYFm1i9RNRFR2o9nMlET2W:mMUEb2P;c04L5K4L5L00N1O2N2N2TNSMeGn2^8dMlF^2U9VNTFl1n9a11O2M2O2N2N1O2J5I8G8IREVMo9h2oEVNY9V4D3L4M3L4M3M4K4M3M2O001O00001O00001O00001O]OWHoIi7n5gHfIY7]4dHiLU8l2^HiLf7Q3X2N101N101N3N4K5L6I8Fh0UOk1UN]bQ4"}, "label": "a skier standing in the middle of two men"}]}
{"id": 203765, "image": "COCO_train2014_000000203765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bike with yellow handlebars laying flat in a green bike lane\"."}], "task": "refering", "answer_template": "The \"a black bike with yellow handlebars laying flat in a green bike lane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 194, 195, 196, 197, 198], "bbox": [0.28498437499999996, 0.34137500000000004, 0.644171875, 0.5089166666666667], "iscrowd": 0, "area": 8559.26645, "rle": {"size": [480, 640], "counts": "k`e21n>;E101N1O2N1O100O1O1O1O100O1O1O10O01O10000O100000000O100000000O100000000O10000TOROfCm0Z<TOfCl0X<VOgCk0X<VOhCj0X<VOhCj0X<VOhCj0W<WOiCi0W<WOhCj0X<VOhCj0X<VOhCi0X<XOhCh0X<WOiCh0X<XOgCi0Y<WOgCi0X<XOhCg0Y<YOgCg0Y<YOgCg0Y<YOfCg0[<YOeCg0Z<ZOeCg0[<[OaCf0`<]O[Ce0e<^OUCe0l<]OoBd0R=e00O4M6J5J6K5K0O0100O0100O10000O10000O10000O1O1O100O1O1O1O2N100O1O1O1O1O1001O1O1O1O1O1O1O1O1O100O1O1oAUOk=l0RBWOm=o0O1O2N1O1O1O2N001O00001O00010O001O000KdNfB]1U=hNjBX1U=jNjBV1V=;O00000010O000000000000010O0000000000010O00000000000001O02N2N2N2`NdBX1^=dNdB\\1`=001O00000000000000000000000000000001O000000000000000000000001O00000000000002N2N2N2N2N2N2N2N1O001O1O1O1O0O2M3L4L4L4M2MSeZ3"}, "label": "a black bike with yellow handlebars laying flat in a green bike lane"}]}
{"id": 203765, "image": "COCO_train2014_000000203765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle that is laying in the middle of the street\"."}], "task": "refering", "answer_template": "The \"the bicycle that is laying in the middle of the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 194, 195, 196, 197, 198], "bbox": [0.28498437499999996, 0.34137500000000004, 0.644171875, 0.5089166666666667], "iscrowd": 0, "area": 8559.26645, "rle": {"size": [480, 640], "counts": "k`e21n>;E101N1O2N1O100O1O1O1O100O1O1O10O01O10000O100000000O100000000O100000000O10000TOROfCm0Z<TOfCl0X<VOgCk0X<VOhCj0X<VOhCj0X<VOhCj0W<WOiCi0W<WOhCj0X<VOhCj0X<VOhCi0X<XOhCh0X<WOiCh0X<XOgCi0Y<WOgCi0X<XOhCg0Y<YOgCg0Y<YOgCg0Y<YOfCg0[<YOeCg0Z<ZOeCg0[<[OaCf0`<]O[Ce0e<^OUCe0l<]OoBd0R=e00O4M6J5J6K5K0O0100O0100O10000O10000O10000O1O1O100O1O1O1O2N100O1O1O1O1O1001O1O1O1O1O1O1O1O1O100O1O1oAUOk=l0RBWOm=o0O1O2N1O1O1O2N001O00001O00010O001O000KdNfB]1U=hNjBX1U=jNjBV1V=;O00000010O000000000000010O0000000000010O00000000000001O02N2N2N2`NdBX1^=dNdB\\1`=001O00000000000000000000000000000001O000000000000000000000001O00000000000002N2N2N2N2N2N2N2N1O001O1O1O1O0O2M3L4L4L4M2MSeZ3"}, "label": "the bicycle that is laying in the middle of the street"}]}
{"id": 457720, "image": "COCO_train2014_000000457720.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange jeep being towed\"."}], "task": "refering", "answer_template": "The \"orange jeep being towed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247], "bbox": [0.387640625, 0.3569375, 0.798875, 0.606375], "iscrowd": 0, "area": 22647.986999999997, "rle": {"size": [480, 640], "counts": "R_d31m>3N3L3N2M4]BA`<a0PCNo<4jB3T=0fB4Y=j0O1O2N1O1O1O2N1O1O1O101O000O101O0000001O0000001O00O100O1O100O100O1O100O1O1O1O1O1O1O1O10O3N1O1O1O1O1O1O1O1O1O1O1O00000000000000000000000000000O1N2N2OM2K52O1N2N2O1N2N2O1N2N3N1N2N2O100000000001O000001O00000000000000001O0000000000O10000000000000000O10000000000000000O10000000000003M5K5K5K4L4L000000001O000000000010O0000000001O0000000000000000000000001O0000000000000000000000O1000000O1000000O100000000O10000000000000000O2O00001O001O00001O0O101O001O00001O001N100O2N104K5L5J5K3N0O1O2N1N2O2M2N2N3M2N2N3M2M3M4L3M3M4LdXl1"}, "label": "orange jeep being towed"}]}
{"id": 457720, "image": "COCO_train2014_000000457720.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange jeep sitting on the back of flat bed truck\"."}], "task": "refering", "answer_template": "The \"orange jeep sitting on the back of flat bed truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 244, 245, 246, 247], "bbox": [0.387640625, 0.3569375, 0.798875, 0.606375], "iscrowd": 0, "area": 22647.986999999997, "rle": {"size": [480, 640], "counts": "R_d31m>3N3L3N2M4]BA`<a0PCNo<4jB3T=0fB4Y=j0O1O2N1O1O1O2N1O1O1O101O000O101O0000001O0000001O00O100O1O100O100O1O100O1O1O1O1O1O1O1O10O3N1O1O1O1O1O1O1O1O1O1O1O00000000000000000000000000000O1N2N2OM2K52O1N2N2O1N2N2O1N2N3N1N2N2O100000000001O000001O00000000000000001O0000000000O10000000000000000O10000000000000000O10000000000003M5K5K5K4L4L000000001O000000000010O0000000001O0000000000000000000000001O0000000000000000000000O1000000O1000000O100000000O10000000000000000O2O00001O001O00001O0O101O001O00001O001N100O2N104K5L5J5K3N0O1O2N1N2O2M2N2N3M2N2N3M2M3M4L3M3M4LdXl1"}, "label": "orange jeep sitting on the back of flat bed truck"}]}
{"id": 64507, "image": "COCO_train2014_000000064507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow chair facing away\"."}], "task": "refering", "answer_template": "The \"a yellow chair facing away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 233, 234, 250, 251, 267, 268], "bbox": [0.6584791666666666, 0.28921874999999997, 1.0, 0.691984375], "iscrowd": 0, "area": 27228.717150000004, "rle": {"size": [640, 480], "counts": "^fU67hc0`0@`0@`0@?B?@`0@`0@6J101N1O2N2N101N1O2O1O2N3M3M3M3L3N1O1O001O1O001O00h0XOT2lM00000000000000001O000O1000000000001O00RMn2000000000000000000000000000O10001O0000000000000000000000000000000000000O100000001O000000000000000000000000000000000O100000000000001O000000000000000000000000000O10000000000000000000001O0000000000000000000O1000000000000WJ"}, "label": "a yellow chair facing away"}]}
{"id": 64507, "image": "COCO_train2014_000000064507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a yellow chair\"."}], "task": "refering", "answer_template": "The \"the back of a yellow chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 233, 234, 250, 251, 267, 268], "bbox": [0.6584791666666666, 0.28921874999999997, 1.0, 0.691984375], "iscrowd": 0, "area": 27228.717150000004, "rle": {"size": [640, 480], "counts": "^fU67hc0`0@`0@`0@?B?@`0@`0@6J101N1O2N2N101N1O2O1O2N3M3M3M3L3N1O1O001O1O001O00h0XOT2lM00000000000000001O000O1000000000001O00RMn2000000000000000000000000000O10001O0000000000000000000000000000000000000O100000001O000000000000000000000000000000000O100000000000001O000000000000000000000000000O10000000000000000000001O0000000000000000000O1000000000000WJ"}, "label": "the back of a yellow chair"}]}
{"id": 105468, "image": "COCO_train2014_000000105468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the huge boat kind of to the left that is darker in color\"."}], "task": "refering", "answer_template": "The \"the huge boat kind of to the left that is darker in color\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308], "bbox": [0.133546875, 0.5765810276679841, 0.43804687500000006, 0.7725296442687747], "iscrowd": 0, "area": 13498.604399999998, "rle": {"size": [506, 640], "counts": "YjZ16c?7Ja0_O6J1N101O001O001O00001O001O001O001O0010O0001O001O0000000000001O01O00000001O0O1000000O101O00000O10001O0O100000000O2O00000O1000001O0O10000O101N1O100O100O101N1O100O100O101N100O100O100O10000O100O10001N100O100O10000O100O100000001N1000000O100000000O1000000O1000000O1000000O10000O10000O1000000O10000O10000O10000000000O101O0000O10O10000000000O10O1O1O100O2N2N2O1N2N1O2O1N2N2N2O1N1O2N2N2OU1jNe0[OVaa5"}, "label": "the huge boat kind of to the left that is darker in color"}]}
{"id": 105468, "image": "COCO_train2014_000000105468.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large boat that is sailing behind a small one in the front\"."}], "task": "refering", "answer_template": "The \"a large boat that is sailing behind a small one in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 239, 256, 257, 258, 259, 260, 261, 262, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308], "bbox": [0.133546875, 0.5765810276679841, 0.43804687500000006, 0.7725296442687747], "iscrowd": 0, "area": 13498.604399999998, "rle": {"size": [506, 640], "counts": "YjZ16c?7Ja0_O6J1N101O001O001O00001O001O001O001O0010O0001O001O0000000000001O01O00000001O0O1000000O101O00000O10001O0O100000000O2O00000O1000001O0O10000O101N1O100O100O101N1O100O100O101N100O100O100O10000O100O10001N100O100O10000O100O100000001N1000000O100000000O1000000O1000000O1000000O10000O10000O1000000O10000O10000O10000000000O101O0000O10O10000000000O10O1O1O100O2N2N2O1N2N1O2O1N2N2N2O1N1O2N2N2OU1jNe0[OVaa5"}, "label": "a large boat that is sailing behind a small one in the front"}]}
{"id": 72701, "image": "COCO_train2014_000000072701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich in the left side of the image\"."}], "task": "refering", "answer_template": "The \"sandwich in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261], "bbox": [0.055625, 0.13933333333333334, 0.5983125, 0.6853958333333333], "iscrowd": 0, "area": 65742.07375000001, "rle": {"size": [480, 640], "counts": "[Ra02n>3M3M3M3M3K5J6I7J6I7J6I7J6J6I7J6I8K4L4M4K4M3L5L3L4M3L5L3L4M4K4M3L4M3L101N101N101N100O2O001N101N1O2O0O1O2N1O2O0O2N1O2N100O2N1O2N101N1O2N1O101N1O2N1O100O1O1O1O100O001O1O1O100O1O1O1O100O1O1O1O10O01O1O1O100O1O1O1O100O1O1O1O10O01O1O100O100O100O100O100O100O10O010000O100O100O100O100O100O10O10O100O100O100O100O100O100O1000O0100O100O100O100000000000000O0100000000000000000O100000000000O100000O100000000000000001O001N10001N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2K4M4K4L5L3M3L5L3M4L3L5A>N2M4L3N3L3N3L3M3N3L3M4M2M4M2M4L3M3M4L3Mb0^OcZh3"}, "label": "sandwich in the left side of the image"}]}
{"id": 72701, "image": "COCO_train2014_000000072701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich on left\"."}], "task": "refering", "answer_template": "The \"sandwich on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261], "bbox": [0.055625, 0.13933333333333334, 0.5983125, 0.6853958333333333], "iscrowd": 0, "area": 65742.07375000001, "rle": {"size": [480, 640], "counts": "[Ra02n>3M3M3M3M3K5J6I7J6I7J6I7J6J6I7J6I8K4L4M4K4M3L5L3L4M3L5L3L4M4K4M3L4M3L101N101N101N100O2O001N101N1O2O0O1O2N1O2O0O2N1O2N100O2N1O2N101N1O2N1O101N1O2N1O100O1O1O1O100O001O1O1O100O1O1O1O100O1O1O1O10O01O1O1O100O1O1O1O100O1O1O1O10O01O1O100O100O100O100O100O100O10O010000O100O100O100O100O100O10O10O100O100O100O100O100O100O1000O0100O100O100O100000000000000O0100000000000000000O100000000000O100000O100000000000000001O001N10001N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2O0O2O0O2O0O101N101N101N100O2K4M4K4L5L3M3L5L3M4L3L5A>N2M4L3N3L3N3L3M3N3L3M4M2M4M2M4L3M3M4L3Mb0^OcZh3"}, "label": "sandwich on left"}]}
{"id": 72701, "image": "COCO_train2014_000000072701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with the cut half facing away from the camera\"."}], "task": "refering", "answer_template": "The \"a sandwich with the cut half facing away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 220, 221, 222], "bbox": [0.39854687499999997, 0.008958333333333334, 0.852578125, 0.5470833333333334], "iscrowd": 0, "area": 40647.37955, "rle": {"size": [480, 640], "counts": "XRh32n>0O2N101N100O2N101N100O2N101N1O2O0O101N1O2O0O2O0O1O2O0O2O0O2N100O2O0O2N101N100O2N101N101N100O2N101N101N1O101N101N101N1O101N101N101N1O101N101N1O2O0O101N101N1O101N101N101N100O2O001O001N10001O001O0O2O00001WHoLa3Q3XLVMh3k2PL\\Mo3e2jKcMU4^2bKjM^4V2_KmMa4S2]KPNb4Q2[KQNd4P2YKSNg4n1UKVNj4j1SKYNm4h1oJ[NQ5e1lJ_NS5a1jJbNU5`1gJcNY5]1dJgN[5Z1aJiN_5W1^JlNb5U1ZJnNf5R1WJROg5o0VJTOj5m0RJVOn5k0nIYOQ6i0jIZOV6h0eI[OZ6g0bI]O]6e0gH5Y7i301O000O2O0000001O0O10000000000O100000000O1000lIQIW4o6dKWI[4i6`K\\I_4e6\\K`Id4_6XKgIg4Y6TKlIk4U6PKoIQ5Q6jJTJU5m5fJWJ[5i5_J\\Ja5d5[J`Jf5`5UJeJk5[5TJaJP6`5oI[JW6R74K6K5K4K6K4K6K5J5L51N101N10001N100O2O001N100O2O000O2O0O101N101O0O101N10001N101N100O2O0O1O2N1O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2M2O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1^NlEWNV:g1RFQNo9m1VFoMk9o1XFoMj9o1XFoMi9o1ZFoMg9P2ZFoMh9n1[FnMh9P2[FmMh9Q2ZFkMi9S2ZFjMh9U2YFhMk9U2XFhMj9W2`1N3M2O2M2O1N3N1N2M4I7I=C<D=CSP\\1"}, "label": "a sandwich with the cut half facing away from the camera"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bald man sitting on the green chair\"."}], "task": "refering", "answer_template": "The \"the bald man sitting on the green chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 223, 224, 225, 226, 247], "bbox": [0.54875, 0.47424882629107984, 0.8570781249999999, 0.6752347417840375], "iscrowd": 0, "area": 5173.378049999998, "rle": {"size": [426, 640], "counts": "hXb43V=8Hb0^O1O1N2O1O1O100O1O001O1O100O1K5L4N1O2C<N3N2M210O001O10O01O0010O10000O10O0100O10000O101N101N101O0O2O0O2O0O3N2N2M3N2M3Nd0[O2O1O0O2O1N2O1N2O001N2N2N2N2N2N2NWj?OeU@70O01O1O1N101O1O1O2N2N1O2N2N101N100O0100O0010O0000000000O2O00000O1000000O10000O100000000O1000000000000O100000000000000O1000000000000O1000001O1O001O0O2O1O001M2M^PV1"}, "label": "the bald man sitting on the green chair"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on the couch staring at something\"."}], "task": "refering", "answer_template": "The \"a man sitting on the couch staring at something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 223, 224, 225, 226, 247], "bbox": [0.54875, 0.47424882629107984, 0.8570781249999999, 0.6752347417840375], "iscrowd": 0, "area": 5173.378049999998, "rle": {"size": [426, 640], "counts": "hXb43V=8Hb0^O1O1N2O1O1O100O1O001O1O100O1K5L4N1O2C<N3N2M210O001O10O01O0010O10000O10O0100O10000O101N101N101O0O2O0O2O0O3N2N2M3N2M3Nd0[O2O1O0O2O1N2O1N2O001N2N2N2N2N2N2NWj?OeU@70O01O1O1N101O1O1O2N2N1O2N2N101N100O0100O0010O0000000000O2O00000O1000000O10000O100000000O1000000000000O100000000000000O1000000000000O1000001O1O001O0O2O1O001M2M^PV1"}, "label": "a man sitting on the couch staring at something"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid lying on a couch and looking at a tablet\"."}], "task": "refering", "answer_template": "The \"a kid lying on a couch and looking at a tablet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 233, 234, 235, 236, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.040921875, 0.6445539906103286, 0.42895312500000005, 1.0], "iscrowd": 0, "area": 22103.49295, "rle": {"size": [426, 640], "counts": "cU;2V=5K6I6K5J7J5J5K4N3L3N3M201N100O2O0O101N100O2O0O2N1O101N1O2O0O1O2N01O002N4L1O001O1O1O001iNlD6U;EoD;S;CnD<V;_OkDa0W;]OiDc0Y;ZOhDf0Z;XOfDh0\\;UOeDk0\\;TOeDk0];SOcDm0l;10000000000000000^O[O_De0h:VOcE7Dc0h:WObE:D?j:XO_E=E;k:YO^E`0E7m:YO[Ec1d:^NYEe1g:\\NVEf1i:>O100O1O100O1000000O100ZNhMZHX2e7mMWHS2h7RNTHn1j7XNRHh1n7ZNoGg1P8[NoGe1Q8]NmGc1S8_NkGa1U8aNiG_1V8cNhG^1X8dNeG]1[8eNaG]1_8eN^G\\1a8fN]G[1c8gNZGZ1f8e1O100O10000O10000O10000O010O10000O100000O101O001O00001O001N101O00001O001O1N101O1O1O1O2N1O1N2O5K7I4L3M3M3M3M1N100000O10000000O100000000000000O10000000O10000000O1000O1000000000O100000000000000O10000000000000000O10000000001O1O1N2O1O1O1O1O1O1O1N2O1O1O001O1O1O1O2M4M2N2N2N2N2N2N2M4M2N2N2N2N2N2N5K6I7J6J6J6G`^g4"}, "label": "a kid lying on a couch and looking at a tablet"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child sitting on the ground wearing a blue hoodie\"."}], "task": "refering", "answer_template": "The \"the child sitting on the ground wearing a blue hoodie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 233, 234, 235, 236, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.040921875, 0.6445539906103286, 0.42895312500000005, 1.0], "iscrowd": 0, "area": 22103.49295, "rle": {"size": [426, 640], "counts": "cU;2V=5K6I6K5J7J5J5K4N3L3N3M201N100O2O0O101N100O2O0O2N1O101N1O2O0O1O2N01O002N4L1O001O1O1O001iNlD6U;EoD;S;CnD<V;_OkDa0W;]OiDc0Y;ZOhDf0Z;XOfDh0\\;UOeDk0\\;TOeDk0];SOcDm0l;10000000000000000^O[O_De0h:VOcE7Dc0h:WObE:D?j:XO_E=E;k:YO^E`0E7m:YO[Ec1d:^NYEe1g:\\NVEf1i:>O100O1O100O1000000O100ZNhMZHX2e7mMWHS2h7RNTHn1j7XNRHh1n7ZNoGg1P8[NoGe1Q8]NmGc1S8_NkGa1U8aNiG_1V8cNhG^1X8dNeG]1[8eNaG]1_8eN^G\\1a8fN]G[1c8gNZGZ1f8e1O100O10000O10000O10000O010O10000O100000O101O001O00001O001N101O00001O001O1N101O1O1O1O2N1O1N2O5K7I4L3M3M3M3M1N100000O10000000O100000000000000O10000000O10000000O1000O1000000000O100000000000000O10000000000000000O10000000001O1O1N2O1O1O1O1O1O1O1N2O1O1O001O1O1O1O2M4M2N2N2N2N2N2N2M4M2N2N2N2N2N2N5K6I7J6J6J6G`^g4"}, "label": "the child sitting on the ground wearing a blue hoodie"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman getting her hair brushed\"."}], "task": "refering", "answer_template": "The \"a woman getting her hair brushed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 221, 222, 223, 224, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.5561875, 0.5912206572769954, 0.7855624999999999, 0.9830985915492958], "iscrowd": 0, "area": 16578.30775, "rle": {"size": [426, 640], "counts": "cYe44P=7J5K6J6I6K6J5N3N2N1O2N1O2O1N1O2N2N1O3M2N2N3M2N2N3M2N2N3M2N2M4L3M3M4L3L4M4L3M2N2N2N2N2M3N2N2N2N2N2N2O100O10000000000O100000000O100O1O1N2O1N2O1N2O1N201N10000000000O100000000O100000000O1000000001O1O001O1O1O001O1O001O1O1O001O1O1O001O1gM^GHb86aGI`85bGJ_83dGL^81dGN^8NeG1]8LeG3]8IfG6\\8GfG8\\8DgG;Z8CgG=Z8@[HNe70^HNc7OaHO`7OcHO^7QOkFe0k18[7POTG?d1?Y7PO[G8`1f0o7TOTHj0X:N2N2N2N2N2N2N2N2K5IVch1"}, "label": "a woman getting her hair brushed"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl who is having her hair brushed\"."}], "task": "refering", "answer_template": "The \"a girl who is having her hair brushed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 221, 222, 223, 224, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.5561875, 0.5912206572769954, 0.7855624999999999, 0.9830985915492958], "iscrowd": 0, "area": 16578.30775, "rle": {"size": [426, 640], "counts": "cYe44P=7J5K6J6I6K6J5N3N2N1O2N1O2O1N1O2N2N1O3M2N2N3M2N2N3M2N2N3M2N2M4L3M3M4L3L4M4L3M2N2N2N2N2M3N2N2N2N2N2N2O100O10000000000O100000000O100O1O1N2O1N2O1N2O1N201N10000000000O100000000O100000000O1000000001O1O001O1O1O001O1O001O1O1O001O1O1O001O1gM^GHb86aGI`85bGJ_83dGL^81dGN^8NeG1]8LeG3]8IfG6\\8GfG8\\8DgG;Z8CgG=Z8@[HNe70^HNc7OaHO`7OcHO^7QOkFe0k18[7POTG?d1?Y7PO[G8`1f0o7TOTHj0X:N2N2N2N2N2N2N2N2K5IVch1"}, "label": "a girl who is having her hair brushed"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in green dress is combing the hair of another woman\"."}], "task": "refering", "answer_template": "The \"a woman in green dress is combing the hair of another woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 261, 262, 263, 285, 286, 308, 309, 310, 311, 312, 332, 333, 334, 335], "bbox": [0.24923437499999998, 0.28833333333333333, 0.611890625, 0.9856572769953051], "iscrowd": 0, "area": 28592.679250000012, "rle": {"size": [426, 640], "counts": "ZhR2?j<2M4L3N2O2M2N2O1N2O1N1O2O1N2O1O1N2O001O1O1O1O1O1O001O1N2O1O1O1O001O1O1O1O1O001N2O1O1O1O1O4L3M2N2N2N3M2M3N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2M3N1N3N2M2N3M3M2M4L4M2M4L4M2M4M3L3M4M3L3M4M3L3M4M3L3N3L4L3N3L4L3N3N2M2N3N2M2N3N2M2N3M3N1N3M3N1N5K7J3M001N2O001N2O001O0O2O1O001NaNkKlIT4c5^L\\Jb3S5oLnJP3a4aM_K`2W4iMjKV2T4mMlKS2R4nMnKR2R4nMoKR2Q4mMPLR2P4oMoKR2Q4mMPLR2Q4mMoKT2Q4kMoKU2Q4lMnKU2R4jMnKV2S4iMmKX2R4hMnKX2S4hMlKY2S4gMmKZ2R4fMnKZ2Q4gMoKZ2U2mKoMj1KY2_1eLeNR1LZ2[1hLgNo0NZ2W1kLjNk0OZ2T1nLmNi0NZ2Q1QMPOe0O[2m0TMROb01Z2i0YMUO=2[2e0\\MXO:2Z2c0`MZO63[2?cM\\O35[2;fM_OO6[28jMAL6[25mMCI8[22oMEF:Z2NTNGB;[2KVNI@;\\2GXNL]O=]2CXN0[O=_2^OZN4WO>a2ZO[N7UO>a2XO]N8SO`0b2SO_N<oNa0d2oN`N?lNb0f2jNbNb0iNd0_6[ObId0_6[OaIe0`6ZO`If0a6XO`Ih0a6WO`Ih0a6WO_Ii0b6VO^Ij0b6UO_Ik0b6TO_Ik0b6TO^Il0c6RO^In0c6QO]Io0d6PO]Io0d6PO\\IP1e6nN\\IR1e6mN[IS1_9O00aMnNfHR1X7QOgHo0X7SOhHl0W7VOhHj0X7WOgHi0X7XOhHh0W7ZOiHe0V7]OiHc0V7_OiHa0V7@jH`0U7BkH=U7DjH<U7FjH:U7HjH8U7IlH6S7JoH5P7JRI6n6HUI7j6IXI6g6I\\I6f6F\\I;e6B]I=e6_O^I`0d6\\O_Ic0c6YO_Ig0c9O1O1O2N3M3M3M4L3M3M]hV3"}, "label": "a woman in green dress is combing the hair of another woman"}]}
{"id": 105470, "image": "COCO_train2014_000000105470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a greet sweater combing another woman ' s hair\"."}], "task": "refering", "answer_template": "The \"a woman in a greet sweater combing another woman ' s hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 261, 262, 263, 285, 286, 308, 309, 310, 311, 312, 332, 333, 334, 335], "bbox": [0.24923437499999998, 0.28833333333333333, 0.611890625, 0.9856572769953051], "iscrowd": 0, "area": 28592.679250000012, "rle": {"size": [426, 640], "counts": "ZhR2?j<2M4L3N2O2M2N2O1N2O1N1O2O1N2O1O1N2O001O1O1O1O1O1O001O1N2O1O1O1O001O1O1O1O1O001N2O1O1O1O1O4L3M2N2N2N3M2M3N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2M3N1N3N2M2N3M3M2M4L4M2M4L4M2M4M3L3M4M3L3M4M3L3M4M3L3N3L4L3N3L4L3N3N2M2N3N2M2N3N2M2N3M3N1N3M3N1N5K7J3M001N2O001N2O001O0O2O1O001NaNkKlIT4c5^L\\Jb3S5oLnJP3a4aM_K`2W4iMjKV2T4mMlKS2R4nMnKR2R4nMoKR2Q4mMPLR2P4oMoKR2Q4mMPLR2Q4mMoKT2Q4kMoKU2Q4lMnKU2R4jMnKV2S4iMmKX2R4hMnKX2S4hMlKY2S4gMmKZ2R4fMnKZ2Q4gMoKZ2U2mKoMj1KY2_1eLeNR1LZ2[1hLgNo0NZ2W1kLjNk0OZ2T1nLmNi0NZ2Q1QMPOe0O[2m0TMROb01Z2i0YMUO=2[2e0\\MXO:2Z2c0`MZO63[2?cM\\O35[2;fM_OO6[28jMAL6[25mMCI8[22oMEF:Z2NTNGB;[2KVNI@;\\2GXNL]O=]2CXN0[O=_2^OZN4WO>a2ZO[N7UO>a2XO]N8SO`0b2SO_N<oNa0d2oN`N?lNb0f2jNbNb0iNd0_6[ObId0_6[OaIe0`6ZO`If0a6XO`Ih0a6WO`Ih0a6WO_Ii0b6VO^Ij0b6UO_Ik0b6TO_Ik0b6TO^Il0c6RO^In0c6QO]Io0d6PO]Io0d6PO\\IP1e6nN\\IR1e6mN[IS1_9O00aMnNfHR1X7QOgHo0X7SOhHl0W7VOhHj0X7WOgHi0X7XOhHh0W7ZOiHe0V7]OiHc0V7_OiHa0V7@jH`0U7BkH=U7DjH<U7FjH:U7HjH8U7IlH6S7JoH5P7JRI6n6HUI7j6IXI6g6I\\I6f6F\\I;e6B]I=e6_O^I`0d6\\O_Ic0c6YO_Ig0c9O1O1O2N3M3M3M4L3M3M]hV3"}, "label": "a woman in a greet sweater combing another woman ' s hair"}]}
{"id": 211978, "image": "COCO_train2014_000000211978.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spotted goat leading\"."}], "task": "refering", "answer_template": "The \"a spotted goat leading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 106, 108, 123, 124, 125, 126, 127, 128, 129, 131, 146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 266, 286, 289, 312], "bbox": [0.36901562499999996, 0.27957894736842104, 0.7278281249999999, 0.9913947368421053], "iscrowd": 0, "area": 29847.506400000006, "rle": {"size": [380, 640], "counts": "egg23e;8H9G8H5I8E;J5M4K6L7I4L5K7I7Ia0_OQ1PO3L1O100O1OnNaIiL^6U3kIdLV6Y3W1M34L5QHfLn0Hi4h5M3N0O100O100O10ZO[JUKf5f4bJVK^5_3nJbLMJV5_3WK]LK0n4`3`L\\L`3d3]2O001O000100O1O101jJaLk1`3nMiLo1Y3kMlLT2_3^MdLb2\\4ZLiKf3Z4SLiKm3[4kKjKT4]4bKfK^4m50O01O0010O000bNTLlJl3P5c1M2O2N2SOjIQLX6c3UJYLn5f3TJVLP6g3SJULP6i3RJTLQ6k3P1O10000100O1O1O1O1O2N1O2N1N2O2N1N2O001O0N2O1N3N1N2001O0000001O002N1O2O0O2N1O2N1O2O001O010O02O1N2O1N2lJ_Ki2b4mLfKS3[4XLXLj3h3PL]LP4V2nJUN>8e4a1oJVN2b0P5W1PK9Q5DSK9n4FUK8l4FWK8i4G[K6d4J_K4a4KbK3]4MfK1Z4NjKNV42mKLR44QLJo35TLGm39QMiNP3V1iMQNW2o1kMnMW2Q2lMlMT2S2PNiMQ2W2RNfMo1Y2TNdMl1[2XNaMj1^2ZN^Mf1b2j3001aNeM\\IZ2d6_NcHb1\\7_1O2O001O001O1:E3N1O2N2M3O4K5K5K2OgGPMe7n2WHVMi7^310O1N2N2N2N2O0O11O001O1O001O000000000000000O1000000O01O001O001M2LPNmIaNQ6<]H5h9E`F9`9EdF8^9GdF7\\9IPGKn7BWHc0S9ChF=W9HeF9[9JaF6a9Q11O2O0O2N2N3M2O1N2Ne0[O?B]]P2"}, "label": "a spotted goat leading"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on the bottom bunk\"."}], "task": "refering", "answer_template": "The \"person on the bottom bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 313, 314, 315], "bbox": [0.0, 0.5507187499999999, 0.6986382113821139, 0.7754375], "iscrowd": 0, "area": 25428.89925, "rle": {"size": [640, 492], "counts": "V>2lc04M3L4M3o\\OA[b0d0^]OB_b0a0[]OEbb0o0L101N2e]OaNTb0`1l]O`NRb0b1n]O^NQb0c1n]O^NPb0c1Q^O]Nna0d1R^O\\Nna0d1Q^O]Noa0c1Q^O]Noa0c1Q^O]Noa0c1Q^O]Noa0c1P^O^NPb0a1Q^O_Noa0a1Q^O_Noa0a1P^O`NPb0`1P^O`NPb0`1P^O`NPb0`1o]OaNQb0^1P^ObNPb0^1P^ObNPb0^1o]OcNQb0]1o]OcNQb0]1o]OcNQb0]1o]ObNRb0^1m]ObNTb0]1m]OcNSb0]1m]ObNTb0^1k]OcNUb0]1k]OcNUb0]1k]OcNUb0d10Hk]OeNUb0[1k]OeNUb0c100000O10000000000O1O1N2O1O1O1O1O1O1O1O1N200001O1O1O1O2N1O00000000000000000000000000O10O00010O001O1O1O100O1O1O1O100O1O1O1O1O1N2O1N2N2O1N2O1N1O200O1O001O1O001O1O010O1O00100O10O0100O010O10O01O10O10000O1000O10O10000O10000000O100000000000000O100000000000000O10000000O1O100O001O100O10000000O01000O100O1O1O100O1O100O1O1O100O1O101N3M6J5L5J6J2N2N1O1O2N1O2N101N1O01N2N3M2O2M3M2O2M3M41N3M2N3M2M4M3L1O01O00010O0000000O1O1O1O1O1O1O1\\OS@TMn?k2W@PMj?P3c000O10000O10000O10000O10000O10000O10002M3N1O2M3M2M4L4M1N2O1O1O1O1O1O1O1O1O1O1O1O1O10000000000000001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001M2K5L5J5K6J5K5K6JhSl2"}, "label": "person on the bottom bunk"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl wearing rolled up blue jeans and a white shirt reading a book\"."}], "task": "refering", "answer_template": "The \"the girl wearing rolled up blue jeans and a white shirt reading a book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 313, 314, 315], "bbox": [0.0, 0.5507187499999999, 0.6986382113821139, 0.7754375], "iscrowd": 0, "area": 25428.89925, "rle": {"size": [640, 492], "counts": "V>2lc04M3L4M3o\\OA[b0d0^]OB_b0a0[]OEbb0o0L101N2e]OaNTb0`1l]O`NRb0b1n]O^NQb0c1n]O^NPb0c1Q^O]Nna0d1R^O\\Nna0d1Q^O]Noa0c1Q^O]Noa0c1Q^O]Noa0c1Q^O]Noa0c1P^O^NPb0a1Q^O_Noa0a1Q^O_Noa0a1P^O`NPb0`1P^O`NPb0`1P^O`NPb0`1o]OaNQb0^1P^ObNPb0^1P^ObNPb0^1o]OcNQb0]1o]OcNQb0]1o]OcNQb0]1o]ObNRb0^1m]ObNTb0]1m]OcNSb0]1m]ObNTb0^1k]OcNUb0]1k]OcNUb0]1k]OcNUb0d10Hk]OeNUb0[1k]OeNUb0c100000O10000000000O1O1N2O1O1O1O1O1O1O1O1N200001O1O1O1O2N1O00000000000000000000000000O10O00010O001O1O1O100O1O1O1O100O1O1O1O1O1N2O1N2N2O1N2O1N1O200O1O001O1O001O1O010O1O00100O10O0100O010O10O01O10O10000O1000O10O10000O10000000O100000000000000O100000000000000O10000000O1O100O001O100O10000000O01000O100O1O1O100O1O100O1O1O100O1O101N3M6J5L5J6J2N2N1O1O2N1O2N101N1O01N2N3M2O2M3M2O2M3M41N3M2N3M2M4M3L1O01O00010O0000000O1O1O1O1O1O1O1\\OS@TMn?k2W@PMj?P3c000O10000O10000O10000O10000O10000O10002M3N1O2M3M2M4L4M1N2O1O1O1O1O1O1O1O1O1O1O1O1O10000000000000001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001M2K5L5J5K6J5K5K6JhSl2"}, "label": "the girl wearing rolled up blue jeans and a white shirt reading a book"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing blue pants and a white shirt laying on a bed reading a book\"."}], "task": "refering", "answer_template": "The \"a woman wearing blue pants and a white shirt laying on a bed reading a book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 224, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 360, 361, 362, 363, 364], "bbox": [0.0, 0.483140625, 0.7951016260162601, 0.9011250000000001], "iscrowd": 0, "area": 72348.8732, "rle": {"size": [640, 492], "counts": "d;j0Vc0P1POo0QOP1POo0QOP1PO:E2O1O1O2N1O1O1O1O1O1OO10000000000O10000000O10O10000000000O10000000000O10000000000O10000000000O10000000000O1000O100000O10000000000O10000000000O10000000000O10000000000O1000000000O010000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000O100000O10000000000O10O10000000O10000000000O10O10000000O10000000000O10000000000O10000000000O01000000000O10000000000O1000000000O010O100O100O010O100O100O010O100O100O0100000O1000O1000000000O010000000000O0100000000O010O100O100O00100O10O0100O10O0100O1O010O100O010O100O010O1O10O0100O10O01O1O10O01O100O001O100O00100O001O100O00100O1O00100O1O010O100O010000O1000000O10000O10000O0100000O2O1O2M2O1O1O1N3N1O1N2O1O2M2O1O1O1N3N1O1O1N2O2N1O1N2O1O2M4M3M2N3L3N3M2N3L3bL^@Z2e?cM\\@\\2g?`M\\@^2f?`M\\@]2h?`MZ@^2h?_M]@]2f?`M]@]2e?`M_@\\2e?aM_@[2d?bM_@[2c?bMb@Z2e`0L4M3M3M3L5L3WOi0H8H8Hjen1"}, "label": "a woman wearing blue pants and a white shirt laying on a bed reading a book"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed on which the girl with the checkered shirt is lying\"."}], "task": "refering", "answer_template": "The \"the bed on which the girl with the checkered shirt is lying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 139, 140, 141, 144, 154, 155, 156, 157, 158, 159, 162, 163, 164, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 229, 234, 235, 236, 237, 238, 239, 240], "bbox": [0.002926829268292683, 0.307484375, 0.8769512195121951, 0.57940625], "iscrowd": 0, "area": 40367.6404, "rle": {"size": [640, 492], "counts": "dk02nc0`0@`0@`0@`0@`0@`0@`0@2N1O1O1O1O001O1O1O1O1O1O1O0001O1O1O001N2O001O1O001O1O1O001O1O001O1O1O001O1O001O1N101O1O001O00001O001O001O00001O001O001O001N10001O001O001O001O00001O001O001O00001O0O101O00001O00001O00001O00001O00001O0000000O01000000O10000O1000000O10000O1000000O100O010O10000O100O100O10000O100O10000O100O100O01000O100O100O10000O100O10000O100O100O1000O0100O10000O10000O10000O10000O100O10000O10O10O10000O10000O10O10O100O010001O1N2O1O1O1N2O1O1O1O1N2O2N1O1N2O1O1O1N2O1O1O1N2O1O00O010O100O01000O10O0100O10O10O01O010O010O1O010O10O01O010O10O01O100O010O100O00100O100O0010YOkMc_OT2\\`0oMc_OQ2Z`0RNe_On1Z`0UNe_Oj1Y`0YNg_Og1W`0[Ni_Od1U`0`Nj_O`1T`0bNk_O^1T`0dNl_O[1U`0fNj_OZ1V`0fNj_OY1W`0hNg_OY1Y`0gNg_OX1Z`0hNf_OX1Z`0iNe_OV1[`0kNe_OT1\\`0mNb_OT1^`0lNb_OS1_`0mNa_OS1_`0nN`_OQ1a`0oN^_OR1b`0nN^_OQ1b`0QO]_Oo0c`0QO]_On0d`0SO[_Om0e`0SOZ_Om0g`0SOY_Om0g`0TOX_Ok0i`0UOW_Ok0i`0VOU_Oj0k`0WOU_Oi0k`0WOU_Oh0l`0V10O100O100O010O100O1O100O010O1001O1O001O0O2O1O001O001O0O2O1O001O1O2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N1O1O1N2O1O1O1O1E:0C=D<G:M2M3N2N2N3M2N2N2N2M4M2N2N2N3M2N2N2M3N3M2N2N2N2N3M2N2M3N2N3M2kNX^O0ja0N_^OIca06d^OB^a0=i^O[OZa0c0R1O1O1O2N1O1O1O1O2N1O1O1O1N3N1O1O1OamU1"}, "label": "the bed on which the girl with the checkered shirt is lying"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top bunk\"."}], "task": "refering", "answer_template": "The \"the top bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 138, 139, 140, 141, 144, 154, 155, 156, 157, 158, 159, 162, 163, 164, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 229, 234, 235, 236, 237, 238, 239, 240], "bbox": [0.002926829268292683, 0.307484375, 0.8769512195121951, 0.57940625], "iscrowd": 0, "area": 40367.6404, "rle": {"size": [640, 492], "counts": "dk02nc0`0@`0@`0@`0@`0@`0@`0@2N1O1O1O1O001O1O1O1O1O1O1O0001O1O1O001N2O001O1O001O1O1O001O1O001O1O1O001O1O001O1N101O1O001O00001O001O001O00001O001O001O001N10001O001O001O001O00001O001O001O00001O0O101O00001O00001O00001O00001O00001O0000000O01000000O10000O1000000O10000O1000000O100O010O10000O100O100O10000O100O10000O100O100O01000O100O100O10000O100O10000O100O100O1000O0100O10000O10000O10000O10000O100O10000O10O10O10000O10000O10O10O100O010001O1N2O1O1O1N2O1O1O1O1N2O2N1O1N2O1O1O1N2O1O1O1N2O1O00O010O100O01000O10O0100O10O10O01O010O010O1O010O10O01O010O10O01O100O010O100O00100O100O0010YOkMc_OT2\\`0oMc_OQ2Z`0RNe_On1Z`0UNe_Oj1Y`0YNg_Og1W`0[Ni_Od1U`0`Nj_O`1T`0bNk_O^1T`0dNl_O[1U`0fNj_OZ1V`0fNj_OY1W`0hNg_OY1Y`0gNg_OX1Z`0hNf_OX1Z`0iNe_OV1[`0kNe_OT1\\`0mNb_OT1^`0lNb_OS1_`0mNa_OS1_`0nN`_OQ1a`0oN^_OR1b`0nN^_OQ1b`0QO]_Oo0c`0QO]_On0d`0SO[_Om0e`0SOZ_Om0g`0SOY_Om0g`0TOX_Ok0i`0UOW_Ok0i`0VOU_Oj0k`0WOU_Oi0k`0WOU_Oh0l`0V10O100O100O010O100O1O100O010O1001O1O001O0O2O1O001O001O0O2O1O001O1O2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N1O1O1N2O1O1O1O1E:0C=D<G:M2M3N2N2N3M2N2N2N2M4M2N2N2N3M2N2N2M3N3M2N2N2N2N3M2N2M3N2N3M2kNX^O0ja0N_^OIca06d^OB^a0=i^O[OZa0c0R1O1O1O2N1O1O1O1O2N1O1O1O1N3N1O1O1OamU1"}, "label": "the top bunk"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with check shirt on top\"."}], "task": "refering", "answer_template": "The \"a woman with check shirt on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 167, 168, 170, 171, 182, 183, 184], "bbox": [0.0, 0.197375, 0.7483333333333333, 0.45131250000000006], "iscrowd": 0, "area": 26425.03505, "rle": {"size": [640, 492], "counts": "i74dc08I7H8H8I7N200001O00001O00001O01O010O0010O01O010O0010O01O010O001O1O001O010O001O001O001O0010O01O001O001O001O0010O01O001O001O001O0010O01O001O001O001O00100O001O001O001O001O01O0O10000000000000001N100000000000TOTNd_Ol1\\`0VNb_Oi1^`0ZN`_Of1``0\\N__Oc1``0`N^_O`1a`0cN]_O]1c`0fNZ_OZ1e`0iNY_OW1g`0kNW_OU1h`0nNV_OR1j`0POT_Oo0l`0TOR_Ol0n`0VOP_Oj0o`0T100O100O10O010O0010O010O01O01O010O01O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O100O100O10000O101O0O2O1N101O1N101O1N101N10O1O001O1O010O001O1O00100O001O1O010O012M3N1O2M3N1N3N2N1N3N2N2M2O2M3N1O2M3N1N10000O100O10000O10O01000O0100O0100UO`_OVN_`0k1a_OTN``0l1a_OSN_`0m1a_ORN``0n1a_OQN_`0o1a_OQN``0n1a_OPN``0P2`_OPN``0P2`_OPN``0P2a_OnM``0R2`_OnM``0R2g001O1O1O1O001O1OM3M3M3M3M2O2M2N2N3M2N2O1010O1O10O01O100O1O00100O1O10O01O100O1O010O1O100O1O100O1O100O1O100O10000000001O1O1O1O1O2N1O1O1O1O2N3M3M3M2M4M3M3M2N3M2N3M1O1O2N1O1N2O1O1N2O2N1O1N2O1O1N2O2N1N2O1O1O1N2O2N1N2O1O1N2O1O2NcY]2"}, "label": "a woman with check shirt on top"}]}
{"id": 367630, "image": "COCO_train2014_000000367630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white shorts reading a book\"."}], "task": "refering", "answer_template": "The \"a woman in white shorts reading a book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 98, 99, 100, 101, 102, 112, 113, 114, 115, 116, 117, 118, 119, 120, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 167, 168, 170, 171, 182, 183, 184], "bbox": [0.0, 0.197375, 0.7483333333333333, 0.45131250000000006], "iscrowd": 0, "area": 26425.03505, "rle": {"size": [640, 492], "counts": "i74dc08I7H8H8I7N200001O00001O00001O01O010O0010O01O010O0010O01O010O001O1O001O010O001O001O001O0010O01O001O001O001O0010O01O001O001O001O0010O01O001O001O001O00100O001O001O001O001O01O0O10000000000000001N100000000000TOTNd_Ol1\\`0VNb_Oi1^`0ZN`_Of1``0\\N__Oc1``0`N^_O`1a`0cN]_O]1c`0fNZ_OZ1e`0iNY_OW1g`0kNW_OU1h`0nNV_OR1j`0POT_Oo0l`0TOR_Ol0n`0VOP_Oj0o`0T100O100O10O010O0010O010O01O01O010O01O100O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O100O100O10000O101O0O2O1N101O1N101O1N101N10O1O001O1O010O001O1O00100O001O1O010O012M3N1O2M3N1N3N2N1N3N2N2M2O2M3N1O2M3N1N10000O100O10000O10O01000O0100O0100UO`_OVN_`0k1a_OTN``0l1a_OSN_`0m1a_ORN``0n1a_OQN_`0o1a_OQN``0n1a_OPN``0P2`_OPN``0P2`_OPN``0P2a_OnM``0R2`_OnM``0R2g001O1O1O1O001O1OM3M3M3M3M2O2M2N2N3M2N2O1010O1O10O01O100O1O00100O1O10O01O100O1O010O1O100O1O100O1O100O1O100O10000000001O1O1O1O1O2N1O1O1O1O2N3M3M3M2M4M3M3M2N3M2N3M1O1O2N1O1N2O1O1N2O2N1O1N2O1O1N2O2N1N2O1O1O1N2O2N1N2O1O1N2O1O2NcY]2"}, "label": "a woman in white shorts reading a book"}]}
{"id": 498706, "image": "COCO_train2014_000000498706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing black and green uniform\"."}], "task": "refering", "answer_template": "The \"man wearing black and green uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 101, 102, 103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264], "bbox": [0.40142187500000004, 0.21348235294117648, 0.6013906250000001, 0.7595529411764705], "iscrowd": 0, "area": 13107.828599999995, "rle": {"size": [425, 640], "counts": "bjZ3e0a<?A3M4L4L3M4K4M4L3M4O0O2N1O2O1N1OPOXEBg:;hEYOW:d0ZFPOd9o0\\1N1N2O2N1N2O2N16J8H7J6I8H7I7QOhM`F_2Z9fMaF_2c2aM\\36jI^2m1VNW4CeI]2]1cNm4Y2kJjMU5[2cJiM\\5\\2\\JgMc5^2WJeMh5`2PJcMP6b2hIbMV6j4O001N101O1O0O10G9G:G9F:H9J5K5L4K5K5L5J5nNPHeL_8Q3l02O0O2O0O2N101N101N101N1O2O0O2OQMdFP2[9QNkFj1T9WNPGd1o8]NWG_1f8cN_GY1_8gNfGV1W8lNkGS1S8mNoGS1n7oNRHS1j7nNXHR1f7oNZHR1d7nN^HR1a7nN`HR1_7mNbHT1\\7mNeHT1Y7kNhHV1V7kNkHU1T7jNnHU1R7kNnHT1R7lNVHmN3U2h7nNhGXOb0i1f7B^H<c7B`H<a7BcH;]7DkH5V7IQI1P7MWIMi62^IHc66dID\\6;gIB`67bIGf61\\IMk6LVI3R7EPI9X7_OjH>R:O2N2N1N3N2N1O2N2M3N1OYaY3"}, "label": "man wearing black and green uniform"}]}
{"id": 498706, "image": "COCO_train2014_000000498706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier with a green vest\"."}], "task": "refering", "answer_template": "The \"a skier with a green vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 101, 102, 103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264], "bbox": [0.40142187500000004, 0.21348235294117648, 0.6013906250000001, 0.7595529411764705], "iscrowd": 0, "area": 13107.828599999995, "rle": {"size": [425, 640], "counts": "bjZ3e0a<?A3M4L4L3M4K4M4L3M4O0O2N1O2O1N1OPOXEBg:;hEYOW:d0ZFPOd9o0\\1N1N2O2N1N2O2N16J8H7J6I8H7I7QOhM`F_2Z9fMaF_2c2aM\\36jI^2m1VNW4CeI]2]1cNm4Y2kJjMU5[2cJiM\\5\\2\\JgMc5^2WJeMh5`2PJcMP6b2hIbMV6j4O001N101O1O0O10G9G:G9F:H9J5K5L4K5K5L5J5nNPHeL_8Q3l02O0O2O0O2N101N101N101N1O2O0O2OQMdFP2[9QNkFj1T9WNPGd1o8]NWG_1f8cN_GY1_8gNfGV1W8lNkGS1S8mNoGS1n7oNRHS1j7nNXHR1f7oNZHR1d7nN^HR1a7nN`HR1_7mNbHT1\\7mNeHT1Y7kNhHV1V7kNkHU1T7jNnHU1R7kNnHT1R7lNVHmN3U2h7nNhGXOb0i1f7B^H<c7B`H<a7BcH;]7DkH5V7IQI1P7MWIMi62^IHc66dID\\6;gIB`67bIGf61\\IMk6LVI3R7EPI9X7_OjH>R:O2N2N1N3N2N1O2N2M3N1OYaY3"}, "label": "a skier with a green vest"}]}
{"id": 498706, "image": "COCO_train2014_000000498706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in a professional red skiing suite looking behind\"."}], "task": "refering", "answer_template": "The \"a skier in a professional red skiing suite looking behind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 165, 166, 167, 168, 189, 190, 191, 192, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284], "bbox": [0.185046875, 0.1640470588235294, 0.419328125, 0.8494352941176471], "iscrowd": 0, "area": 29086.774050000004, "rle": {"size": [425, 640], "counts": "ega13Q=5RIJd0;ZOJ>;@J8;EK3:KKM:1KG:7KB9<M[O8c0OTO5l00kN4T12dN1\\15ZNOe17RNMm19jMKV2:aMJ^2<YMHg2=PMFP3Q1WLROi3f1[K^Nd4T500O1O100O100O100O100O1O100O100O1O1O1O1O1O1O100O001O1O1O1O1O100O1O1O1O1O1O100O1O00100O1O100O1O1O010O1O1000O100000000O10000000000O1000000000000O100000000000000O100[J[Mf0e2ZO^Mc0c2\\OaM`0_2AdM;\\2EhM7X2IlM2U2NnMOS20QNLP23TNHn18VNCl1;YN@h1?]N[Oe1d0_NXOc1f0bNUO_1j0fNPO\\1o0iNlNX1T1kNhNW1V1nNdNT1Z1TO_Nm0_1[OZNg0d1@nMh0P2A`Mh0^2S5O2M4L5K4M4K4L5L4K3M2N101N2N2N2O0O2N2O1N4L4L4L4H\\Xj4"}, "label": "a skier in a professional red skiing suite looking behind"}]}
{"id": 498706, "image": "COCO_train2014_000000498706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white hat on skiis talking\"."}], "task": "refering", "answer_template": "The \"a man in a white hat on skiis talking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 165, 166, 167, 168, 189, 190, 191, 192, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284], "bbox": [0.185046875, 0.1640470588235294, 0.419328125, 0.8494352941176471], "iscrowd": 0, "area": 29086.774050000004, "rle": {"size": [425, 640], "counts": "ega13Q=5RIJd0;ZOJ>;@J8;EK3:KKM:1KG:7KB9<M[O8c0OTO5l00kN4T12dN1\\15ZNOe17RNMm19jMKV2:aMJ^2<YMHg2=PMFP3Q1WLROi3f1[K^Nd4T500O1O100O100O100O100O1O100O100O1O1O1O1O1O1O100O001O1O1O1O1O100O1O1O1O1O1O100O1O00100O1O100O1O1O010O1O1000O100000000O10000000000O1000000000000O100000000000000O100[J[Mf0e2ZO^Mc0c2\\OaM`0_2AdM;\\2EhM7X2IlM2U2NnMOS20QNLP23TNHn18VNCl1;YN@h1?]N[Oe1d0_NXOc1f0bNUO_1j0fNPO\\1o0iNlNX1T1kNhNW1V1nNdNT1Z1TO_Nm0_1[OZNg0d1@nMh0P2A`Mh0^2S5O2M4L5K4M4K4L5L4K3M2N101N2N2N2O0O2N2O1N4L4L4L4H\\Xj4"}, "label": "a man in a white hat on skiis talking"}]}
{"id": 572439, "image": "COCO_train2014_000000572439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in red sweatshirt and green shorts\"."}], "task": "refering", "answer_template": "The \"boy in red sweatshirt and green shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 100, 101, 102, 123, 124, 125, 146, 147, 148, 169, 170, 171, 192, 193, 194, 214, 215, 216, 217, 237, 238], "bbox": [0.327578125, 0.13094142259414226, 0.474359375, 0.6188075313807531], "iscrowd": 0, "area": 10031.975800000006, "rle": {"size": [478, 640], "counts": "dYR3:a>4L4L5I7I4K600O1O10O01O1O10O010000O010O1[MdNhG\\1Q8SOgGm0S8_OcGd0]8^OZGj0d8YOWGk0h8YOoF6fN[OY:d0fF8\\OiNn9`3[FYLc9h3`FVL_9j3i00[OVLYFj3g9XLWFh3i9YLVFf3j9]LTFc3k9`LSF_3m9cLRF]32[LY9:cFZ34`LU98eFY36`LS9:cFW3:aLR9W4nFjKR9U4nFlKQ9S4PGnKP9o3RGRLm8m3SGULm8j3QGYLn8h3PGZLn8g3PG\\Lo8e3jFbLT9`3kFaLS9a3mF_LQ9b3oF_LP9b3PG_Ln8b3QG_Ln8b3QG_Ln8a3SG_Ll8b3SG_Ll8b3TG^Lk8l4N1O1O1O_O^GhJa8V5bGjJ]8S5hGlJX8Q5kGoJS8P5QHoJn7o4UHQKj7Y3`GoMj0gNe7[3eGiMg0lNc7\\3jGbMf0QO_7^3oG[Mf0UOZ7a3THUMd0YOW7n1eH`ND3c0^OT7k1jHcNn9Y1XFfNh9V1]FiNc9U1`FjN_9U1cFkN\\9S1hFmNW9P1lFPOT9n0nFROR9k0RGTOn8j0TGVOl8h0WGVOj8i0WGWOi8h0YGWOn8a0TG^Oo8>UG^On8?VG^Ol8?P3I;EThl4"}, "label": "boy in red sweatshirt and green shorts"}]}
{"id": 572439, "image": "COCO_train2014_000000572439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skater with a red shirt and green shorts\"."}], "task": "refering", "answer_template": "The \"the skater with a red shirt and green shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 100, 101, 102, 123, 124, 125, 146, 147, 148, 169, 170, 171, 192, 193, 194, 214, 215, 216, 217, 237, 238], "bbox": [0.327578125, 0.13094142259414226, 0.474359375, 0.6188075313807531], "iscrowd": 0, "area": 10031.975800000006, "rle": {"size": [478, 640], "counts": "dYR3:a>4L4L5I7I4K600O1O10O01O1O10O010000O010O1[MdNhG\\1Q8SOgGm0S8_OcGd0]8^OZGj0d8YOWGk0h8YOoF6fN[OY:d0fF8\\OiNn9`3[FYLc9h3`FVL_9j3i00[OVLYFj3g9XLWFh3i9YLVFf3j9]LTFc3k9`LSF_3m9cLRF]32[LY9:cFZ34`LU98eFY36`LS9:cFW3:aLR9W4nFjKR9U4nFlKQ9S4PGnKP9o3RGRLm8m3SGULm8j3QGYLn8h3PGZLn8g3PG\\Lo8e3jFbLT9`3kFaLS9a3mF_LQ9b3oF_LP9b3PG_Ln8b3QG_Ln8b3QG_Ln8a3SG_Ll8b3SG_Ll8b3TG^Lk8l4N1O1O1O_O^GhJa8V5bGjJ]8S5hGlJX8Q5kGoJS8P5QHoJn7o4UHQKj7Y3`GoMj0gNe7[3eGiMg0lNc7\\3jGbMf0QO_7^3oG[Mf0UOZ7a3THUMd0YOW7n1eH`ND3c0^OT7k1jHcNn9Y1XFfNh9V1]FiNc9U1`FjN_9U1cFkN\\9S1hFmNW9P1lFPOT9n0nFROR9k0RGTOn8j0TGVOl8h0WGVOj8i0WGWOi8h0YGWOn8a0TG^Oo8>UG^On8?VG^Ol8?P3I;EThl4"}, "label": "the skater with a red shirt and green shorts"}]}
{"id": 384027, "image": "COCO_train2014_000000384027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus travelling down the street\"."}], "task": "refering", "answer_template": "The \"a red bus travelling down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.27303125, 0.36854166666666666, 0.8292187500000001, 0.7662916666666667], "iscrowd": 0, "area": 56066.74260000001, "rle": {"size": [480, 640], "counts": "jYb2j0c=T1lNU1gMP2@70000000001O0O1000000000001O00000O1000001O4L1O001O0O1000O1O1O101O001O1O001O000000000000000000000000000000000O10000000000000000000O100000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000001O000000000000000000000O100000001O000000000O1O1O100000000001O0000001O2N001O000000000000001O0000000000001O0000000000001O000000000000001O0000000O10001O000000000001O3N3L1O2NK5K5002O1N001O1O1O1O01O0001O00001ON2N3M2N2N2N2N2N2N3M200000000O10001O0O100000000O2O00000O1000001O0O1000000O10001O000O1000000O2O0000000O10001O0O100000000O101O000O100000001N1000000O1000001O0O1000000O101O00000O1001O100O001O001O000010O000O1O1O1N2O2N1J6000000O100O2O0O100O100O101O0O100O105J<\\MTD]1T<\\NZDU1g;kNgD`0`;@P2010O01K6Jjkb1"}, "label": "a red bus travelling down the street"}]}
{"id": 384027, "image": "COCO_train2014_000000384027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large red and tan bus\"."}], "task": "refering", "answer_template": "The \"a large red and tan bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.27303125, 0.36854166666666666, 0.8292187500000001, 0.7662916666666667], "iscrowd": 0, "area": 56066.74260000001, "rle": {"size": [480, 640], "counts": "jYb2j0c=T1lNU1gMP2@70000000001O0O1000000000001O00000O1000001O4L1O001O0O1000O1O1O101O001O1O001O000000000000000000000000000000000O10000000000000000000O100000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000001O000000000000000000000O100000001O000000000O1O1O100000000001O0000001O2N001O000000000000001O0000000000001O0000000000001O000000000000001O0000000O10001O000000000001O3N3L1O2NK5K5002O1N001O1O1O1O01O0001O00001ON2N3M2N2N2N2N2N2N3M200000000O10001O0O100000000O2O00000O1000001O0O1000000O10001O000O1000000O2O0000000O10001O0O100000000O101O000O100000001N1000000O1000001O0O1000000O101O00000O1001O100O001O001O000010O000O1O1O1N2O2N1J6000000O100O2O0O100O100O101O0O100O105J<\\MTD]1T<\\NZDU1g;kNgD`0`;@P2010O01K6Jjkb1"}, "label": "a large red and tan bus"}]}
{"id": 564253, "image": "COCO_train2014_000000564253.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person surfing in a black and white outfit\"."}], "task": "refering", "answer_template": "The \"a person surfing in a black and white outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 177], "bbox": [0.6168125, 0.3998287671232877, 0.8063437499999999, 0.7581849315068493], "iscrowd": 0, "area": 4642.012949999999, "rle": {"size": [292, 640], "counts": "Rk`31R92N2N2O1N001O1O001O1O010O00000000O1001O00000000001O0O1000000O2O00000O10O10aNGSJ:i5LUJ3k51QJ0n55nIKS66jIJX68eIH[6:bIG_6:^IFc6;[IFf6:WIHj68TIHn68QIHP79mHHT7:iHEY7R11O2M3N2N2N2O00101N2N2N^ImNT5R1jJPOW5o0gJTOX5l0fJVO[5i0bJ[O^5d0aJ]O_5d0_J^Oa5a0^JAa5?^JBb5?\\JCd5<[JEe5;[JFd5;XJHh58UJLj54QJ1o53hIDf6]10O10000O1000O010000O2O000O100O2O1N2O1N101N2N2O1N2O1N1O2O1N100O2lNZHg0g7XO\\Hf0d7YO^He0c7ZO`Hd0`7ZOdHc0]7[OgHb0P8M3M3M3KSVS1"}, "label": "a person surfing in a black and white outfit"}]}
{"id": 564253, "image": "COCO_train2014_000000564253.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady surfing a small wave in the ocean\"."}], "task": "refering", "answer_template": "The \"lady surfing a small wave in the ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 177], "bbox": [0.6168125, 0.3998287671232877, 0.8063437499999999, 0.7581849315068493], "iscrowd": 0, "area": 4642.012949999999, "rle": {"size": [292, 640], "counts": "Rk`31R92N2N2O1N001O1O001O1O010O00000000O1001O00000000001O0O1000000O2O00000O10O10aNGSJ:i5LUJ3k51QJ0n55nIKS66jIJX68eIH[6:bIG_6:^IFc6;[IFf6:WIHj68TIHn68QIHP79mHHT7:iHEY7R11O2M3N2N2N2O00101N2N2N^ImNT5R1jJPOW5o0gJTOX5l0fJVO[5i0bJ[O^5d0aJ]O_5d0_J^Oa5a0^JAa5?^JBb5?\\JCd5<[JEe5;[JFd5;XJHh58UJLj54QJ1o53hIDf6]10O10000O1000O010000O2O000O100O2O1N2O1N101N2N2O1N2O1N1O2O1N100O2lNZHg0g7XO\\Hf0d7YO^He0c7ZO`Hd0`7ZOdHc0]7[OgHb0P8M3M3M3KSVS1"}, "label": "lady surfing a small wave in the ocean"}]}
{"id": 302116, "image": "COCO_train2014_000000302116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with black tshirt an blue jeans skateboarding on black iron rail\"."}], "task": "refering", "answer_template": "The \"man with black tshirt an blue jeans skateboarding on black iron rail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 57, 58, 59, 60, 76, 77, 78, 94, 95, 96, 97, 98, 112, 113, 114, 115, 129, 130, 131, 145, 146, 147, 148, 162, 163, 164, 179, 180, 181, 182, 197, 198, 199, 200, 215, 216, 217, 232, 233, 234, 249, 250, 251, 252], "bbox": [0.3732848232848233, 0.123921875, 0.8364033264033264, 0.65378125], "iscrowd": 0, "area": 20278.233849999997, "rle": {"size": [640, 481], "counts": "_b`3>bc000000000001O000000000000001O0000000k\\OAib0?V]OBjb0>T]ODlb0=Q]OEob0;o\\OGQc09m\\OISc0b00O1000000O101O0O1000000O10000O10000O10000O10001N10000O10000001O2O001O001O001O001O0000001O0000001O0000001O00000O2O0000001O000TAVOa:j0ZE]Od:d0WEBg:>TEJi:6RE1l:OoD7P;JkD9V;GeD>[;B`Db0a;^OZDg0f;ZOTDj0m;VOoCn0Q<ROjCR1W<nNdCW1\\<iN_C[1b<eNZC_1f<aNUCc1l<]NoBh1Q=XNjBl1W=TNeBP2[=PN`BT2a=lMZBY2f=gMUB]2l=cMoAb2Q>^MkAe2V>[MeAj2[>VM`An2a>RMZAS3f>S1100O100O100O100O1O100O100O100O010O100O100O10lCYKm8g4jFbKW9^4`FjKa9U4VFTLk9l3kE]LV:b3bEfL_:Z3WEoLj:P3mDYMT;g2dD`M\\;`2_DeMb;[2XDjMi;U2SDoMn;Q2lCTNU<k41O1O0000O100NWKSDc1j;\\NcD[1Z;dNkD[1S;cNRE\\1k:dNZEZ1d:dNaEZ1_:cNfE\\1Y:_NnE`1P:[NXFd1g9WN`Fh1_9RNiFm1U9oMSGo1l8lM[GS2d8iMbGV2\\8lMeGS2Z8nMgGP2Y8QNhGn1V8TNkGk1T8VNnGh1Q8XNQHg1o7YNRHf1n7YNTHf1m7YNTHf1m7YNTHf1l7YNVHf1k7YNWHd1j7\\NWHc1j7[NXHd1i7[NXHd1h7\\NYHc1h7[NZHd1f7\\N[Hc1f7[N\\Hd1e7[N]Hb1d7^N]Ha1d7]N^Hb1c7\\N_Hc1d7ZNbH`1a7\\NfH^1]7^NjH\\1Y7aNmHX1V7eNTIR1n6kNnI:S6D[JOg5OZJ0g5NZJ2h5KYJ5h5G[J9f5D]J:e<0001O0000001N10001O0000001N10001Kbo`1"}, "label": "man with black tshirt an blue jeans skateboarding on black iron rail"}]}
{"id": 302116, "image": "COCO_train2014_000000302116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboarder\"."}], "task": "refering", "answer_template": "The \"skateboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 57, 58, 59, 60, 76, 77, 78, 94, 95, 96, 97, 98, 112, 113, 114, 115, 129, 130, 131, 145, 146, 147, 148, 162, 163, 164, 179, 180, 181, 182, 197, 198, 199, 200, 215, 216, 217, 232, 233, 234, 249, 250, 251, 252], "bbox": [0.3732848232848233, 0.123921875, 0.8364033264033264, 0.65378125], "iscrowd": 0, "area": 20278.233849999997, "rle": {"size": [640, 481], "counts": "_b`3>bc000000000001O000000000000001O0000000k\\OAib0?V]OBjb0>T]ODlb0=Q]OEob0;o\\OGQc09m\\OISc0b00O1000000O101O0O1000000O10000O10000O10000O10001N10000O10000001O2O001O001O001O001O0000001O0000001O0000001O00000O2O0000001O000TAVOa:j0ZE]Od:d0WEBg:>TEJi:6RE1l:OoD7P;JkD9V;GeD>[;B`Db0a;^OZDg0f;ZOTDj0m;VOoCn0Q<ROjCR1W<nNdCW1\\<iN_C[1b<eNZC_1f<aNUCc1l<]NoBh1Q=XNjBl1W=TNeBP2[=PN`BT2a=lMZBY2f=gMUB]2l=cMoAb2Q>^MkAe2V>[MeAj2[>VM`An2a>RMZAS3f>S1100O100O100O100O1O100O100O100O010O100O100O10lCYKm8g4jFbKW9^4`FjKa9U4VFTLk9l3kE]LV:b3bEfL_:Z3WEoLj:P3mDYMT;g2dD`M\\;`2_DeMb;[2XDjMi;U2SDoMn;Q2lCTNU<k41O1O0000O100NWKSDc1j;\\NcD[1Z;dNkD[1S;cNRE\\1k:dNZEZ1d:dNaEZ1_:cNfE\\1Y:_NnE`1P:[NXFd1g9WN`Fh1_9RNiFm1U9oMSGo1l8lM[GS2d8iMbGV2\\8lMeGS2Z8nMgGP2Y8QNhGn1V8TNkGk1T8VNnGh1Q8XNQHg1o7YNRHf1n7YNTHf1m7YNTHf1m7YNTHf1l7YNVHf1k7YNWHd1j7\\NWHc1j7[NXHd1i7[NXHd1h7\\NYHc1h7[NZHd1f7\\N[Hc1f7[N\\Hd1e7[N]Hb1d7^N]Ha1d7]N^Hb1c7\\N_Hc1d7ZNbH`1a7\\NfH^1]7^NjH\\1Y7aNmHX1V7eNTIR1n6kNnI:S6D[JOg5OZJ0g5NZJ2h5KYJ5h5G[J9f5D]J:e<0001O0000001N10001O0000001N10001Kbo`1"}, "label": "skateboarder"}]}
{"id": 498729, "image": "COCO_train2014_000000498729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"top part of the bed , buried under personal belongings\"."}], "task": "refering", "answer_template": "The \"top part of the bed , buried under personal belongings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 305, 306], "bbox": [0.250640625, 0.5250416666666667, 0.7208593750000001, 0.7841041666666667], "iscrowd": 0, "area": 6873.542649999998, "rle": {"size": [480, 640], "counts": "fi[24k>3M3M3M3M3N2M3M3M3N2M3M3N2M3M3M100O1O010000O10000O2O00000O10001N1000000O101O0O1000000O2O000O10000O2O00000O10001N1000000O101O0O10000O2O00000O10001N1000000O101O0O1000000O2O000O1000001N1000000O0100000O0100000O01000O10O1000O1000O10O1000O10O1000O10O1000O10O1000O0100000O0100000O01000000O0100000O0100000O01000O10O1000O10O1000O10O1000O10O1000O1000O0100O10O0O2K4M40O010O10O0100O010O10O0100O010O10O0100O010O10O0100O01O0O2O001N101N101O0O2O1N1100O100O100O100O100O100O100O10ROBZC=g<DXC<h<EWC:k<EUC;k<FTC9m<HRC8n<IQC6P=KoB5Q=LnB3T=MkB3U=NjB1W=0hB0X=1gBNZ=3eBM[=3eBL]=4bBL^=5aBJ`=7_BIa=8^BGc=:\\BFd=;[BDg=<XBDh==WBBj=?UBAk=?UB@l=j00O101N100O1O100O2N100O1O100O2O0O1O100O1O2O0O1O100O1O2O0O1O100O1O2O0O1O100O1O2O0O1O10idc2"}, "label": "top part of the bed , buried under personal belongings"}]}
{"id": 498729, "image": "COCO_train2014_000000498729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"colourful bedsheet on sofa\"."}], "task": "refering", "answer_template": "The \"colourful bedsheet on sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 305, 306], "bbox": [0.250640625, 0.5250416666666667, 0.7208593750000001, 0.7841041666666667], "iscrowd": 0, "area": 6873.542649999998, "rle": {"size": [480, 640], "counts": "fi[24k>3M3M3M3M3N2M3M3M3N2M3M3N2M3M3M100O1O010000O10000O2O00000O10001N1000000O101O0O1000000O2O000O10000O2O00000O10001N1000000O101O0O10000O2O00000O10001N1000000O101O0O1000000O2O000O1000001N1000000O0100000O0100000O01000O10O1000O1000O10O1000O10O1000O10O1000O10O1000O0100000O0100000O01000000O0100000O0100000O01000O10O1000O10O1000O10O1000O10O1000O1000O0100O10O0O2K4M40O010O10O0100O010O10O0100O010O10O0100O010O10O0100O01O0O2O001N101N101O0O2O1N1100O100O100O100O100O100O100O10ROBZC=g<DXC<h<EWC:k<EUC;k<FTC9m<HRC8n<IQC6P=KoB5Q=LnB3T=MkB3U=NjB1W=0hB0X=1gBNZ=3eBM[=3eBL]=4bBL^=5aBJ`=7_BIa=8^BGc=:\\BFd=;[BDg=<XBDh==WBBj=?UBAk=?UB@l=j00O101N100O1O100O2N100O1O100O2O0O1O100O1O2O0O1O100O1O2O0O1O100O1O2O0O1O100O1O2O0O1O10idc2"}, "label": "colourful bedsheet on sofa"}]}
{"id": 498729, "image": "COCO_train2014_000000498729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of the mattress , not the top of it , but the side showing how thick it is\"."}], "task": "refering", "answer_template": "The \"the side of the mattress , not the top of it , but the side showing how thick it is\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 243, 244, 245, 246, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 329, 330, 331], "bbox": [0.302375, 0.5502708333333334, 0.740484375, 0.8771874999999999], "iscrowd": 0, "area": 7736.499, "rle": {"size": [480, 640], "counts": "iYk22m>2M3O0O2O1N2O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O1N101gBlNc<V1[CjNf<W1XCjNh<W1VCiNj<Y1TChNl<Y1RCgNo<Y1PCgNQ=Z1mBgNS=Z1kBfNU=\\1iBeNW=a1100O10O01O100O10O0100O1O010O100O10O0100O1O10O0100O100O00100O10O0100O100O00100O100O010O1O100O010O100O10O01O100O010O100O00100O10O0100O10O01O10O0100O10O01O100O010O10O0100O1O010O100O010O1O010O010O010O010O1O010O010O010O0010O010O010O010O01O010O010O010O0010O010O0010O010O0010O010O010O01O010O01O010O010O01O010O010O010O010O010O01O010O010O010O010O010O010O010O00010O010O010O010O010O010O010O0010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O0010O010O010O010O010O0010O02O1N2O0O2O1N2O1N1Kfa]2"}, "label": "the side of the mattress , not the top of it , but the side showing how thick it is"}]}
{"id": 498729, "image": "COCO_train2014_000000498729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mattress with a pink blanket over it\"."}], "task": "refering", "answer_template": "The \"a mattress with a pink blanket over it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 243, 244, 245, 246, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 329, 330, 331], "bbox": [0.302375, 0.5502708333333334, 0.740484375, 0.8771874999999999], "iscrowd": 0, "area": 7736.499, "rle": {"size": [480, 640], "counts": "iYk22m>2M3O0O2O1N2O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O1N101gBlNc<V1[CjNf<W1XCjNh<W1VCiNj<Y1TChNl<Y1RCgNo<Y1PCgNQ=Z1mBgNS=Z1kBfNU=\\1iBeNW=a1100O10O01O100O10O0100O1O010O100O10O0100O1O10O0100O100O00100O10O0100O100O00100O100O010O1O100O010O100O10O01O100O010O100O00100O10O0100O10O01O10O0100O10O01O100O010O10O0100O1O010O100O010O1O010O010O010O010O1O010O010O010O0010O010O010O010O01O010O010O010O0010O010O0010O010O0010O010O010O01O010O01O010O010O01O010O010O010O010O010O01O010O010O010O010O010O010O010O00010O010O010O010O010O010O010O0010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O0010O010O010O010O010O0010O02O1N2O0O2O1N2O1N1Kfa]2"}, "label": "a mattress with a pink blanket over it"}]}
{"id": 498729, "image": "COCO_train2014_000000498729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red chair in the foreground facing away from the camera\"."}], "task": "refering", "answer_template": "The \"a red chair in the foreground facing away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [316, 317, 318, 319, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.733515625, 0.7823541666666666, 0.999390625, 1.0], "iscrowd": 0, "area": 15100.7631, "rle": {"size": [480, 640], "counts": "lXl6b0X>h0WO7H8I7H8H8K50000000000000000O1L4K500O10000O100O1N2N2O1000000O1000000000000001O001O00001O00001O00001O001O00001O000000001O0000001O0000001O0000001O000000001O0000001O000000001O000000001O00000000001O000000001O000000001O00000000001O000000000000000000001O000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000^Ob0TN_E"}, "label": "a red chair in the foreground facing away from the camera"}]}
{"id": 498729, "image": "COCO_train2014_000000498729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red chair in the bottom right corner that is closest to you\"."}], "task": "refering", "answer_template": "The \"the red chair in the bottom right corner that is closest to you\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [316, 317, 318, 319, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.733515625, 0.7823541666666666, 0.999390625, 1.0], "iscrowd": 0, "area": 15100.7631, "rle": {"size": [480, 640], "counts": "lXl6b0X>h0WO7H8I7H8H8K50000000000000000O1L4K500O10000O100O1N2N2O1000000O1000000000000001O001O00001O00001O00001O001O00001O000000001O0000001O0000001O0000001O000000001O0000001O000000001O000000001O00000000001O000000001O000000001O00000000001O000000000000000000001O000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000^Ob0TN_E"}, "label": "the red chair in the bottom right corner that is closest to you"}]}
{"id": 498730, "image": "COCO_train2014_000000498730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman underneath an umbrella walking with black sandals\"."}], "task": "refering", "answer_template": "The \"a woman underneath an umbrella walking with black sandals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 74, 75, 76, 97, 98, 99, 120, 121, 122, 143, 144, 145, 165, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283, 304, 305, 306], "bbox": [0.1743125, 0.18805620608899296, 0.347125, 0.9222950819672131], "iscrowd": 0, "area": 17641.55785, "rle": {"size": [427, 640], "counts": "nl^1=m<2O1O1O1O10N101O00001N10001N101N10iZ3<jdL5J6K5K5J6K5L4K5oESN[8R2]GTNa8o1WGWNg8n1PGXNo8l1hF[N`1WOe3:SLY2_N\\Ng1ROT3S1^Lf2=XLi2d1cLV2a0WLh2n1_Lm1g0WLe2V2^Ld1k0WLc2`2ZL\\1Q1UL`2j2WLS1V1UL_2S3TLj0Z1TLb2\\7[MgHd2Z7ZMgHf2[7VMhHh2Z7VMgHj2Z7SMiHl2Y7QMhHo2[7lLhHS3h8O101O001O0O2O001O0O0101O000O1000001N10000O1YN\\LfIe3Y6^LdIc3Z6aLcIa3[6aLbIa3\\6cLaI^3^6eL_I\\3`6gL]I[3`6iL]IZ3`6iL^IY3_6jL^IY3^6kL`IX3[6lLbIW3Z6mLdIU3W6QMfIQ3U6TMgIR3S6RMhIV3P6oLlIX3l5Z2H7J5J6G9F;F9F:G9C>kNU1E;F:E;E;E;EaT^5"}, "label": "a woman underneath an umbrella walking with black sandals"}]}
{"id": 498730, "image": "COCO_train2014_000000498730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black dress and carrying a umbrella in the rain\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black dress and carrying a umbrella in the rain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 74, 75, 76, 97, 98, 99, 120, 121, 122, 143, 144, 145, 165, 166, 167, 168, 189, 190, 191, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283, 304, 305, 306], "bbox": [0.1743125, 0.18805620608899296, 0.347125, 0.9222950819672131], "iscrowd": 0, "area": 17641.55785, "rle": {"size": [427, 640], "counts": "nl^1=m<2O1O1O1O10N101O00001N10001N101N10iZ3<jdL5J6K5K5J6K5L4K5oESN[8R2]GTNa8o1WGWNg8n1PGXNo8l1hF[N`1WOe3:SLY2_N\\Ng1ROT3S1^Lf2=XLi2d1cLV2a0WLh2n1_Lm1g0WLe2V2^Ld1k0WLc2`2ZL\\1Q1UL`2j2WLS1V1UL_2S3TLj0Z1TLb2\\7[MgHd2Z7ZMgHf2[7VMhHh2Z7VMgHj2Z7SMiHl2Y7QMhHo2[7lLhHS3h8O101O001O0O2O001O0O0101O000O1000001N10000O1YN\\LfIe3Y6^LdIc3Z6aLcIa3[6aLbIa3\\6cLaI^3^6eL_I\\3`6gL]I[3`6iL]IZ3`6iL^IY3_6jL^IY3^6kL`IX3[6lLbIW3Z6mLdIU3W6QMfIQ3U6TMgIR3S6RMhIV3P6oLlIX3l5Z2H7J5J6G9F;F9F:G9C>kNU1E;F:E;E;E;EaT^5"}, "label": "a woman wearing a black dress and carrying a umbrella in the rain"}]}
{"id": 498730, "image": "COCO_train2014_000000498730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"un umbrella that is on the hands of a woman\"."}], "task": "refering", "answer_template": "The \"un umbrella that is on the hands of a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 55, 72, 73, 74, 76, 77, 78, 95, 96, 99, 100, 101], "bbox": [0.12798437499999998, 0.13266978922716627, 0.4255, 0.3094379391100702], "iscrowd": 0, "area": 7744.965299999999, "rle": {"size": [427, 640], "counts": "]YR11Y=1O1O1O1O1O1O1O2N1O1O3M4L4L4L4L3M4L3M1O1O2N1O1O1O1O1O1O2O0000000O1000000000001O0O100000000O100O101N100O1OO2N1N3N1O2N1O2N1N210O01000000000O10000O1N2M3N2M3N2M3N2M3N2N200O10000O10K51O2M3N0000000001O0000001N10001O00001O00001O000O3N2N2N2N2N1O2O1N2N2N2O3L6J5K6J5LO000O101O0O10001N10000O2O00001N10001N10000O2O000O2O00001N100O101O0O101O0O101N10000O2O000O2O000O2O0O100O2N1O101N1O1O2O9F:F_[i4"}, "label": "un umbrella that is on the hands of a woman"}]}
{"id": 498730, "image": "COCO_train2014_000000498730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a brown jacket and jeans holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a person in a brown jacket and jeans holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 198, 199, 221, 222, 244, 245, 267, 268, 290, 291, 292], "bbox": [0.574640625, 0.15459016393442623, 0.732859375, 0.8436065573770491], "iscrowd": 0, "area": 16486.077649999992, "rle": {"size": [427, 640], "counts": "iai4W1b;c0O1N2N2O1N2N2O1N3M2WFiM\\8X2YGSNe8o1PG]Nm8e1hFfNW9h2N3M2O1N3M2N2O2M2N2O2M2N2O2N1O1O2M2O1O2N1O1O2N1O1O2lJPKi1Q5lM`Km1a4gMRLQ2P4cM\\L[2e3ZMeLf2\\3nKoMQ4f500000000000000000000000000O100000000000000000ZOoKQHR4j7TLUHl3f7ZLYHf3b7_L^H`3]7gLbHY3Y7lLgHT3OoKS3d4mL]OMULR3`4PM[OLYLR3\\4RM[OI^LR3Y4TMYOGcLS3T4VM5h2LWM6f2LXM8e2H[M:b2H\\M;b2E]M>`2D_M>_2C_M`0^2AbMa0\\2Z5L3N2M3N2M4M7I8G9H7H9H7H9HWTW2"}, "label": "a person in a brown jacket and jeans holding an umbrella"}]}
{"id": 498730, "image": "COCO_train2014_000000498730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry person wearing pants and holding a black umbrella while it rains\"."}], "task": "refering", "answer_template": "The \"a blurry person wearing pants and holding a black umbrella while it rains\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 198, 199, 221, 222, 244, 245, 267, 268, 290, 291, 292], "bbox": [0.574640625, 0.15459016393442623, 0.732859375, 0.8436065573770491], "iscrowd": 0, "area": 16486.077649999992, "rle": {"size": [427, 640], "counts": "iai4W1b;c0O1N2N2O1N2N2O1N3M2WFiM\\8X2YGSNe8o1PG]Nm8e1hFfNW9h2N3M2O1N3M2N2O2M2N2O2M2N2O2N1O1O2M2O1O2N1O1O2N1O1O2lJPKi1Q5lM`Km1a4gMRLQ2P4cM\\L[2e3ZMeLf2\\3nKoMQ4f500000000000000000000000000O100000000000000000ZOoKQHR4j7TLUHl3f7ZLYHf3b7_L^H`3]7gLbHY3Y7lLgHT3OoKS3d4mL]OMULR3`4PM[OLYLR3\\4RM[OI^LR3Y4TMYOGcLS3T4VM5h2LWM6f2LXM8e2H[M:b2H\\M;b2E]M>`2D_M>_2C_M`0^2AbMa0\\2Z5L3N2M3N2M4M7I8G9H7H9H7H9HWTW2"}, "label": "a blurry person wearing pants and holding a black umbrella while it rains"}]}
{"id": 80943, "image": "COCO_train2014_000000080943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter giraffe\"."}], "task": "refering", "answer_template": "The \"the shorter giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 116, 117, 130, 131, 132, 145, 146, 147, 148, 160, 161, 162, 174, 175, 188, 189, 190, 203, 204, 205, 217, 218, 219, 232, 233, 234, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 280], "bbox": [0.47683098591549294, 0.28496875, 0.899906103286385, 0.832984375], "iscrowd": 0, "area": 20143.174949999997, "rle": {"size": [640, 426], "counts": "V[o32kc05J7J6J5K6J6I7J5K6J6J6J5J7eN[1L3M4L3M1O001O1O1O001O1O001O1O001O1O1O1O1O1O1O1J6N2N2M3N2M3N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O`LTCNk<N`CL_<1kCIT<3WDGh;5cDE\\;8nDBQ;:ZE@e:<iE[OS:e0YGSNd8k1h4N101O10O101N10001N10001N10N110000O110010OSO\\M`@d2b?^MZ@b2g?aMV@^2k?eMQ@Z2R`0hMj_OX2W`0kMf_OT2\\`0nM`_OR2a`0QN[_Oo1g`0b02O1O2N1O2M2O1O21N100O100O2N11O1O1O010O1O1O00100O1O1O00100O1O00100O1O001O6K6I6J7I6J2O1M2O2M3N2N1N3N2M3N2N2M4M2M3N3M2M3N2M4M2N^cj0"}, "label": "the shorter giraffe"}]}
{"id": 80943, "image": "COCO_train2014_000000080943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light brown and white giraffe leading the trail\"."}], "task": "refering", "answer_template": "The \"light brown and white giraffe leading the trail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 116, 117, 130, 131, 132, 145, 146, 147, 148, 160, 161, 162, 174, 175, 188, 189, 190, 203, 204, 205, 217, 218, 219, 232, 233, 234, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 280], "bbox": [0.47683098591549294, 0.28496875, 0.899906103286385, 0.832984375], "iscrowd": 0, "area": 20143.174949999997, "rle": {"size": [640, 426], "counts": "V[o32kc05J7J6J5K6J6I7J5K6J6J6J5J7eN[1L3M4L3M1O001O1O1O001O1O001O1O001O1O1O1O1O1O1O1J6N2N2M3N2M3N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O`LTCNk<N`CL_<1kCIT<3WDGh;5cDE\\;8nDBQ;:ZE@e:<iE[OS:e0YGSNd8k1h4N101O10O101N10001N10001N10N110000O110010OSO\\M`@d2b?^MZ@b2g?aMV@^2k?eMQ@Z2R`0hMj_OX2W`0kMf_OT2\\`0nM`_OR2a`0QN[_Oo1g`0b02O1O2N1O2M2O1O21N100O100O2N11O1O1O010O1O1O00100O1O1O00100O1O00100O1O001O6K6I6J7I6J2O1M2O2M3N2N1N3N2M3N2N2M4M2M3N3M2M3N2M4M2N^cj0"}, "label": "light brown and white giraffe leading the trail"}]}
{"id": 80943, "image": "COCO_train2014_000000080943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that ' s behind another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe that ' s behind another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 22, 23, 24, 37, 38, 39, 40, 51, 52, 53, 54, 55, 66, 67, 68, 81, 82, 83, 96, 97, 98, 110, 111, 112, 125, 126, 127, 140, 141, 142, 154, 155, 156, 157, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 275, 276, 277, 286, 287, 288, 290, 291, 292, 306], "bbox": [0.003380281690140845, 0.0060625, 0.7123474178403756, 0.87796875], "iscrowd": 0, "area": 57002.20170000001, "rle": {"size": [640, 426], "counts": "YR11mc0:G9F:G9F6K3M2M4M3L4M3L3N3L4M3I6D=B<D;E;O20O00001O0000010O0000001O000010O002N4M3L3M4M2M4L3M4M3L3M4L3N3L3M100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1ON3N1O1N2O2M2O1O1N3N1O1N3N1N2O1O2M2O1O110O01O10O01O00100O001O010O1O010O1O0O2N1O2N2N1N3N1O2N2N1N4M4L4L4L4K6K4L4L4L4K5L4L4L4L5K4K5L4L4J6K5J6J6K6I6J6K5J6K5J6J6K5J7I6K5J6J6K5J6J4M3L3M3N3L4L4M3L4L4M3L4L4M2M3M3N2M3NjIYJWNe5f1eJVNX5g1QKVNm4f1]KVN`4g1jKUNS4l1SLnMm3T2WLfMh3[2^L^Mb3c2dL^KjNlMa4g6kLnJ[OlMj3W7RMeJEdMW3i7ZM\\Jh4e5_KTJ_4n5gKkIX4X6mKaIS4`6TLXIk3j6[LPId3Q7cLgH\\3[7jL^HV3c7c4O1O1^Ob0POP1J7I8I7H8K5L4M3L3M4M3L4M3L4M3L4L4M3L4L3N3L4L4M3L4M3L4L4M3L4L3101O001O001O001O001O0O2O001O001O001O08H8H8H8H8I5J1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O2O2M2M3K5K5L4K6J5K\\i\\2"}, "label": "a giraffe that ' s behind another giraffe"}]}
{"id": 80943, "image": "COCO_train2014_000000080943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tall giraffe with dark spots\"."}], "task": "refering", "answer_template": "The \"tall giraffe with dark spots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 22, 23, 24, 37, 38, 39, 40, 51, 52, 53, 54, 55, 66, 67, 68, 81, 82, 83, 96, 97, 98, 110, 111, 112, 125, 126, 127, 140, 141, 142, 154, 155, 156, 157, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 275, 276, 277, 286, 287, 288, 290, 291, 292, 306], "bbox": [0.003380281690140845, 0.0060625, 0.7123474178403756, 0.87796875], "iscrowd": 0, "area": 57002.20170000001, "rle": {"size": [640, 426], "counts": "YR11mc0:G9F:G9F6K3M2M4M3L4M3L3N3L4M3I6D=B<D;E;O20O00001O0000010O0000001O000010O002N4M3L3M4M2M4L3M4M3L3M4L3N3L3M100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1ON3N1O1N2O2M2O1O1N3N1O1N3N1N2O1O2M2O1O110O01O10O01O00100O001O010O1O010O1O0O2N1O2N2N1N3N1O2N2N1N4M4L4L4L4K6K4L4L4L4K5L4L4L4L5K4K5L4L4J6K5J6J6K6I6J6K5J6K5J6J6K5J7I6K5J6J6K5J6J4M3L3M3N3L4L4M3L4L4M3L4L4M2M3M3N2M3NjIYJWNe5f1eJVNX5g1QKVNm4f1]KVN`4g1jKUNS4l1SLnMm3T2WLfMh3[2^L^Mb3c2dL^KjNlMa4g6kLnJ[OlMj3W7RMeJEdMW3i7ZM\\Jh4e5_KTJ_4n5gKkIX4X6mKaIS4`6TLXIk3j6[LPId3Q7cLgH\\3[7jL^HV3c7c4O1O1^Ob0POP1J7I8I7H8K5L4M3L3M4M3L4M3L4M3L4L4M3L4L3N3L4L4M3L4M3L4L4M3L4L3101O001O001O001O001O0O2O001O001O001O08H8H8H8H8I5J1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O2O2M2M3K5K5L4K6J5K\\i\\2"}, "label": "tall giraffe with dark spots"}]}
{"id": 171064, "image": "COCO_train2014_000000171064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfer on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the surfer on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 167, 168, 169, 189, 190, 191, 212, 213, 235, 236, 258, 283, 303, 306, 326, 329], "bbox": [0.1715, 0.3910647181628392, 0.35771875, 0.8622546972860125], "iscrowd": 0, "area": 5072.5310500000005, "rle": {"size": [479, 640], "counts": "Qkc19c>4L3M4M2N2N3L3N3M2N3M2N2N1=Df0ZO1O001N101O001O0O2O001O00]<0_C3N1N3M3N1N3@?M4M2M4TOPOiCP1W<SOfCn0Z<SOdCm0\\<WO`Cj0_<[OYCi0g<g0O1O1000O01000O10000O100O10000O10001O1O00100O1O001O1O1O010O1O001O001O001ElBkNS=S1QCkNn3Mi4W1\\GjNj32i4P1bGlNm27QM1^8i0iGmNh2R1]5M[K4e4DbK>Y4@lKa0o3DPL=l3FUL9k3HTL9k3GUL:k3FTL;k3EULDjKNP8>VL]OSL3g7a0VLYOVL6c7c0ULUO\\L6`7f0RLSOaL6\\7i0QLoNfL7Z7k0nKlNlL8V7k0RMUOo2i0RMWOn2h0SMXOm2g0TMYOl2f0VMYOj2f0WMZOi2f0WMZOj2d0WM\\Oi2c0XM]Oi2c0WM[Ok2d0UM\\Om2b0SM]OP3=TMBn28WMHP900O2O00iko5"}, "label": "the surfer on the left in the right hand picture"}]}
{"id": 171064, "image": "COCO_train2014_000000171064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man walking away from the beach carrying a yellow surfboard\"."}], "task": "refering", "answer_template": "The \"a man walking away from the beach carrying a yellow surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 167, 168, 169, 189, 190, 191, 212, 213, 235, 236, 258, 283, 303, 306, 326, 329], "bbox": [0.1715, 0.3910647181628392, 0.35771875, 0.8622546972860125], "iscrowd": 0, "area": 5072.5310500000005, "rle": {"size": [479, 640], "counts": "Qkc19c>4L3M4M2N2N3L3N3M2N3M2N2N1=Df0ZO1O001N101O001O0O2O001O00]<0_C3N1N3M3N1N3@?M4M2M4TOPOiCP1W<SOfCn0Z<SOdCm0\\<WO`Cj0_<[OYCi0g<g0O1O1000O01000O10000O100O10000O10001O1O00100O1O001O1O1O010O1O001O001O001ElBkNS=S1QCkNn3Mi4W1\\GjNj32i4P1bGlNm27QM1^8i0iGmNh2R1]5M[K4e4DbK>Y4@lKa0o3DPL=l3FUL9k3HTL9k3GUL:k3FTL;k3EULDjKNP8>VL]OSL3g7a0VLYOVL6c7c0ULUO\\L6`7f0RLSOaL6\\7i0QLoNfL7Z7k0nKlNlL8V7k0RMUOo2i0RMWOn2h0SMXOm2g0TMYOl2f0VMYOj2f0WMZOi2f0WMZOj2d0WM\\Oi2c0XM]Oi2c0WM[Ok2d0UM\\Om2b0SM]OP3=TMBn28WMHP900O2O00iko5"}, "label": "a man walking away from the beach carrying a yellow surfboard"}]}
{"id": 171064, "image": "COCO_train2014_000000171064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"board that the man in front is holding\"."}], "task": "refering", "answer_template": "The \"board that the man in front is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 280, 281, 282, 283, 303, 304, 305, 326], "bbox": [0.1960625, 0.5177035490605427, 0.390234375, 0.8394363256784969], "iscrowd": 0, "area": 10078.872800000003, "rle": {"size": [479, 640], "counts": "Wjj12l>c0\\Od0\\O6J3N2M3M2N3N0O2N1O2O1N1O2N101N1O2O0O2N1O2O0O2N2N101N1O2N101N1O2N101N2N1O20O01O001O0010O01O001O10O01O001O0010O01O02O1N2O1O1N2O1N2N1O2M2O00BYMeDg2\\;YMdDe2];]MaDc2`;^M^Db2b;`M\\D`2d;eMXDY2j;=001O0O2O001O0O2O001O0O2O001N101O001N101O001N101O1N1O2N1O2N2N3M2N2N3M2N2N3M2N2N2N3M4L4L3M4L4LgUf5"}, "label": "board that the man in front is holding"}]}
{"id": 171064, "image": "COCO_train2014_000000171064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow surf - board with white stars painted on it\"."}], "task": "refering", "answer_template": "The \"a yellow surf - board with white stars painted on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 280, 281, 282, 283, 303, 304, 305, 326], "bbox": [0.1960625, 0.5177035490605427, 0.390234375, 0.8394363256784969], "iscrowd": 0, "area": 10078.872800000003, "rle": {"size": [479, 640], "counts": "Wjj12l>c0\\Od0\\O6J3N2M3M2N3N0O2N1O2O1N1O2N101N1O2O0O2N1O2O0O2N2N101N1O2N101N1O2N101N2N1O20O01O001O0010O01O001O10O01O001O0010O01O02O1N2O1O1N2O1N2N1O2M2O00BYMeDg2\\;YMdDe2];]MaDc2`;^M^Db2b;`M\\D`2d;eMXDY2j;=001O0O2O001O0O2O001O0O2O001N101O001N101O001N101O1N1O2N1O2N2N3M2N2N3M2N2N3M2N2N2N3M4L4L3M4L4LgUf5"}, "label": "a yellow surf - board with white stars painted on it"}]}
{"id": 113721, "image": "COCO_train2014_000000113721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the young man is taking photo who is stting in the chair with his boy\"."}], "task": "refering", "answer_template": "The \"the young man is taking photo who is stting in the chair with his boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 306, 307, 308, 331, 332, 354], "bbox": [0.30067187500000003, 0.28379166666666666, 0.5388437500000001, 0.8873958333333334], "iscrowd": 0, "area": 21691.58395, "rle": {"size": [480, 640], "counts": "Yij28b>8H8H8H8I7H8ZC_NX:NZFk1POdNa:GXFd2d9cMSFd2h9bMPFd2l9bMmEd2n9Y1L5J5L5K4L5K4L5K4N2N1O1OEVGcJi8^5ZG_Je8b5^G[Ja8f5;1O1O1O11O1O11O2N2N1O2N2N2H8H8M2M3N2N1O2N2M3N2N2N1O2O1N2O1N101N2O1N2O1N10iNmHbJQ7^5QIbJm6^5VIaJh0kNR5c6XJbJ7YO`5U6\\JaJFGn5g5^J^Ka5a4cJ]K]5b4eJ]K[5b4hJ\\KX5c4kJ[KU5d4mJ[KS5c4QK[Ko4d4TKZKl4e4YKWKg4h4]KUKc4j4bKRK^4m4gKoJY4o4lKmJU4R5PLjJP4U5ULfJl3Y5XLcJi3\\5ZLaJg3^5\\L`Jd3_5`L]Ja3b5bL[J_3d5eLYJ\\3e5gLXJ[3o0lIb3V9\\LkFa3X9^LiF`3Z9]LhFa3[9]LgF`3e9TL]Fj3e9SL^Fk3d9SL^Ff1MIf9`N_Fd1<\\OW9mN`Fc1m0hNk8CYFc1k;\\NVDb1f<M2M4M2N32M3N1N3N2N1O2N1O2N1OO100O100O100O2N2O0O2O1N1O2N2N2N1O2N2N1O2N2N1N3N1O2N1O2N2N1O2N1O2N4L6IRQZ4"}, "label": "the young man is taking photo who is stting in the chair with his boy"}]}
{"id": 113721, "image": "COCO_train2014_000000113721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a camera\"."}], "task": "refering", "answer_template": "The \"a man holding a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 306, 307, 308, 331, 332, 354], "bbox": [0.30067187500000003, 0.28379166666666666, 0.5388437500000001, 0.8873958333333334], "iscrowd": 0, "area": 21691.58395, "rle": {"size": [480, 640], "counts": "Yij28b>8H8H8H8I7H8ZC_NX:NZFk1POdNa:GXFd2d9cMSFd2h9bMPFd2l9bMmEd2n9Y1L5J5L5K4L5K4L5K4N2N1O1OEVGcJi8^5ZG_Je8b5^G[Ja8f5;1O1O1O11O1O11O2N2N1O2N2N2H8H8M2M3N2N1O2N2M3N2N2N1O2O1N2O1N101N2O1N2O1N10iNmHbJQ7^5QIbJm6^5VIaJh0kNR5c6XJbJ7YO`5U6\\JaJFGn5g5^J^Ka5a4cJ]K]5b4eJ]K[5b4hJ\\KX5c4kJ[KU5d4mJ[KS5c4QK[Ko4d4TKZKl4e4YKWKg4h4]KUKc4j4bKRK^4m4gKoJY4o4lKmJU4R5PLjJP4U5ULfJl3Y5XLcJi3\\5ZLaJg3^5\\L`Jd3_5`L]Ja3b5bL[J_3d5eLYJ\\3e5gLXJ[3o0lIb3V9\\LkFa3X9^LiF`3Z9]LhFa3[9]LgF`3e9TL]Fj3e9SL^Fk3d9SL^Ff1MIf9`N_Fd1<\\OW9mN`Fc1m0hNk8CYFc1k;\\NVDb1f<M2M4M2N32M3N1N3N2N1O2N1O2N1OO100O100O100O2N2O0O2O1N1O2N2N2N1O2N2N1O2N2N1N3N1O2N1O2N2N1O2N1O2N4L6IRQZ4"}, "label": "a man holding a camera"}]}
{"id": 113721, "image": "COCO_train2014_000000113721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid watching something in a chair next to a man who is clicking a camera\"."}], "task": "refering", "answer_template": "The \"a kid watching something in a chair next to a man who is clicking a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 174, 175, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 264, 265, 266, 267], "bbox": [0.47128125, 0.3175625, 0.672296875, 0.6801875000000001], "iscrowd": 0, "area": 9456.214650000005, "rle": {"size": [480, 640], "counts": "Zk]41n>1O2O0O2N101N1O2O0O2N101N1O101N1O100O2N104K<E1O000010O0001O0000000000000000000001O00000000000000O100O1O1O1O100O1O1N2L4L4L5K4L4L4L4L4XOh0^NnLbFf3]9o0O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O11O001O00001O001O0O1O2M2eN_FVMc9i2Z1O101O0YOjDdMW;X2lDhMT;V2nDiMS;T2PElMQ;P2REPNo:T1lElNT:P1PFoNR:m0QFSOP:h0TFXOm9c0WF]Oj9?YF@i9<YFEh98ZFGh94\\FLj9AbF?P<0O0100O002N2M4K4MoUR3"}, "label": "a kid watching something in a chair next to a man who is clicking a camera"}]}
{"id": 113721, "image": "COCO_train2014_000000113721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little kid wearing green shirt\"."}], "task": "refering", "answer_template": "The \"little kid wearing green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 151, 152, 153, 174, 175, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 264, 265, 266, 267], "bbox": [0.47128125, 0.3175625, 0.672296875, 0.6801875000000001], "iscrowd": 0, "area": 9456.214650000005, "rle": {"size": [480, 640], "counts": "Zk]41n>1O2O0O2N101N1O2O0O2N101N1O101N1O100O2N104K<E1O000010O0001O0000000000000000000001O00000000000000O100O1O1O1O100O1O1N2L4L4L5K4L4L4L4L4XOh0^NnLbFf3]9o0O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O11O001O00001O001O0O1O2M2eN_FVMc9i2Z1O101O0YOjDdMW;X2lDhMT;V2nDiMS;T2PElMQ;P2REPNo:T1lElNT:P1PFoNR:m0QFSOP:h0TFXOm9c0WF]Oj9?YF@i9<YFEh98ZFGh94\\FLj9AbF?P<0O0100O002N2M4K4MoUR3"}, "label": "little kid wearing green shirt"}]}
{"id": 269380, "image": "COCO_train2014_000000269380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the polar bear behind the log\"."}], "task": "refering", "answer_template": "The \"the polar bear behind the log\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278], "bbox": [0.0, 0.5908578431372549, 0.365515625, 0.8233578431372549], "iscrowd": 0, "area": 10454.947000000004, "rle": {"size": [408, 640], "counts": "T91h<3M2L5J5K6J5K6L3M4L3M3M4L1OO10000O10000O100O10000O10000O10O10O10000001O001O001O001O00000000O1000000O100000000O100000000O10O1000O10001O000O100000000O100000000O10001O0O100000000O100000000O1000000O2O0000000O100000000O1000000O100000O10O100000000O10O100000O1000000O10O1L4J6K4L5O1N2O1N101N2O1O0O200O100O00100O100O010O1O100O01000000O100000000O100000000O1000000O100000000O100000000O1000000O100000O10O100000000O100000001N3N1O2N1N3N1O1O2N3M3M3M2N3M3M3M3M1O2N2N1O2N1O2N2N1O3M2NX_Q5"}, "label": "the polar bear behind the log"}]}
{"id": 269380, "image": "COCO_train2014_000000269380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear lying down relaxing\"."}], "task": "refering", "answer_template": "The \"bear lying down relaxing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278], "bbox": [0.0, 0.5908578431372549, 0.365515625, 0.8233578431372549], "iscrowd": 0, "area": 10454.947000000004, "rle": {"size": [408, 640], "counts": "T91h<3M2L5J5K6J5K6L3M4L3M3M4L1OO10000O10000O100O10000O10000O10O10O10000001O001O001O001O00000000O1000000O100000000O100000000O10O1000O10001O000O100000000O100000000O10001O0O100000000O100000000O1000000O2O0000000O100000000O1000000O100000O10O100000000O10O100000O1000000O10O1L4J6K4L5O1N2O1N101N2O1O0O200O100O00100O100O010O1O100O01000000O100000000O100000000O1000000O100000000O100000000O1000000O100000O10O100000000O100000001N3N1O2N1N3N1O1O2N3M3M3M2N3M3M3M3M1O2N2N1O2N1O2N2N1O3M2NX_Q5"}, "label": "bear lying down relaxing"}]}
{"id": 269380, "image": "COCO_train2014_000000269380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big polor bear\"."}], "task": "refering", "answer_template": "The \"a big polor bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 333, 335, 336, 337, 338, 339, 340], "bbox": [0.442671875, 0.23147058823529412, 0.925453125, 0.9865196078431373], "iscrowd": 0, "area": 55661.8142, "rle": {"size": [408, 640], "counts": "j`a33b<3N2M3O1O1O1O2N1O1O1O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2L4N2O1O1O1O1O1O1O1O1O100N2M3N2M3N2N2M3N1O2M2O1N3N1O1N3N1N3N2N2N2O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1eG^L_7c3[HcLd7^3[HcLd7^3[HcLd7^3[HcLd7^3[HbLe7_3ZHbLe7_3ZHbLd7`3[HaLd7`3\\H`Lc7a3\\H`Lc7a3\\H`Ld7`3\\H`Lc7a3\\H`Ld7`3\\H`Ld7`3[HaLd7`3\\H`Ld7`3[HaLd7`3[HaLe7Y4O100O100O100O100O1L4J6K5K5K5K5K5J6K5K5K5N2N2O1N2N2O1N2N2O1N2N2O1N2N101N10001N10001O0O101O0O101O00E:K6J6L4K4O2O1O1N2O]K_IQ3b6mLaIR3b6kL_IT3d6iL]IW3e6gL[IX3h6eLYIZ3j6cLWI]3k6aLUI^3n6_LSIa3P7[LQId3T7WLmHh3W7ULiHk3[7PLfHo3_7lKbHT4Q8O1O100O1O100O100O1O010O100O1D<00O100O100O11O0O100000002N1O2N1O0000O10O0100O100O10000000000O10O10000000O1000000000O10O1000O10000O01001O0O10000O1000000O101O0O1N2M3M3M3M4O0O101N101N100UJdKY3^4eLdKZ3\\4eLgKX3[4fLhKX3X4fLkKX3W4fLlKW3U4hLmKW3S4hLPLU3R4iLPLV3P4hLTLU3n3iLULU3k3iLXLU3j3iLYLU3g3jL[LT3g3iL]LU3c3jLcLP3_3mLgLo2Y3QMlLj2U3TMPMg2Q3XMUM0oMo0m4oNZMFSNX1c4QO_M\\OYN^1Z4TOcMRO^Nf1o3WOhMhNcNm1g3YOXOf0h0YO[Oe0f0YO\\Oe0e0ZO]Oe0d0YO^Oe0c0[O^Od0c0ZO_Oe0a0ZOAd0a0ZOBd0>[ODc0>\\ODa0=^OE`0=^OEa0;^OH?:@I;9D1O4O5E09X6O2N1O2N1O2N1LYPc0"}, "label": "a big polor bear"}]}
{"id": 269380, "image": "COCO_train2014_000000269380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large polar bear sitting in front of a log looking to it ' s left\"."}], "task": "refering", "answer_template": "The \"large polar bear sitting in front of a log looking to it ' s left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 333, 335, 336, 337, 338, 339, 340], "bbox": [0.442671875, 0.23147058823529412, 0.925453125, 0.9865196078431373], "iscrowd": 0, "area": 55661.8142, "rle": {"size": [408, 640], "counts": "j`a33b<3N2M3O1O1O1O2N1O1O1O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2L4N2O1O1O1O1O1O1O1O1O100N2M3N2M3N2N2M3N1O2M2O1N3N1O1N3N1N3N2N2N2O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1eG^L_7c3[HcLd7^3[HcLd7^3[HcLd7^3[HcLd7^3[HbLe7_3ZHbLe7_3ZHbLd7`3[HaLd7`3\\H`Lc7a3\\H`Lc7a3\\H`Ld7`3\\H`Lc7a3\\H`Ld7`3\\H`Ld7`3[HaLd7`3\\H`Ld7`3[HaLd7`3[HaLe7Y4O100O100O100O100O1L4J6K5K5K5K5K5J6K5K5K5N2N2O1N2N2O1N2N2O1N2N2O1N2N101N10001N10001O0O101O0O101O00E:K6J6L4K4O2O1O1N2O]K_IQ3b6mLaIR3b6kL_IT3d6iL]IW3e6gL[IX3h6eLYIZ3j6cLWI]3k6aLUI^3n6_LSIa3P7[LQId3T7WLmHh3W7ULiHk3[7PLfHo3_7lKbHT4Q8O1O100O1O100O100O1O010O100O1D<00O100O100O11O0O100000002N1O2N1O0000O10O0100O100O10000000000O10O10000000O1000000000O10O1000O10000O01001O0O10000O1000000O101O0O1N2M3M3M3M4O0O101N101N100UJdKY3^4eLdKZ3\\4eLgKX3[4fLhKX3X4fLkKX3W4fLlKW3U4hLmKW3S4hLPLU3R4iLPLV3P4hLTLU3n3iLULU3k3iLXLU3j3iLYLU3g3jL[LT3g3iL]LU3c3jLcLP3_3mLgLo2Y3QMlLj2U3TMPMg2Q3XMUM0oMo0m4oNZMFSNX1c4QO_M\\OYN^1Z4TOcMRO^Nf1o3WOhMhNcNm1g3YOXOf0h0YO[Oe0f0YO\\Oe0e0ZO]Oe0d0YO^Oe0c0[O^Od0c0ZO_Oe0a0ZOAd0a0ZOBd0>[ODc0>\\ODa0=^OE`0=^OEa0;^OH?:@I;9D1O4O5E09X6O2N1O2N1O2N1LYPc0"}, "label": "large polar bear sitting in front of a log looking to it ' s left"}]}
{"id": 564302, "image": "COCO_train2014_000000564302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue chair\"."}], "task": "refering", "answer_template": "The \"a blue chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 235, 236, 237, 256, 257, 258, 279, 281], "bbox": [0.1399375, 0.4015566037735849, 0.33049999999999996, 0.8150471698113207], "iscrowd": 0, "area": 8076.116349999996, "rle": {"size": [424, 640], "counts": "Q^U15R=7J5K2M3N2N1N3N3M5K5K5eGfN^4_1^HWOd1_Oj5e2TJ\\Mk5e2TJ\\Ml5d2SJ^Ml5b2TJ^Ml5b2SJ_Mm5b2QJ`Mn5`2QJ_MQ6a2nI_MS6a2mI^MT6b2kI^MV6c2hI]MY6c2fI\\M[6e2eIZM\\6f2cIZM^6f2aIZM`6f2_IZMb6f2]IZMd6f2\\IZMd6f2[IZMf6f2YIZMh6f2WI[Mi6e2WIZMj6f2UI[Mk6e2TI[Ml6f2SIZMn6f2QI[Mo6e2QIZMP7f2oH[MQ7e2nH[MS7e2lH[MU7e2kH[MU7e2jH[MW7e2hH\\MX7d2gH\\MZ7d2eH\\M[7e2eH[M[7e2dH[M]7e2bH[Mf7^2YHcMV8n1iGRNf8`1ZG`NT9R1kFnN[9m0dFSO\\9S1^FnNb9Z1VFeNk9b1mE_NS:h1eEXN\\:Y2100O13M3M3LWNmEb0P:^OUF`0h9@]F=a9BeF<X9DmF:P9EWG8f8H\\GWOSOa0^98_GWOXO?V99dGXOZO;Q9=eGXO^O7m8`0fGYOB2g8f0gGXOGMb8j0hGYOKH]8o0iGXOODX8T1iGXO\\9g0eFYO[9g0eFYO[9f0fFZOZ9f0fFZOZ9e0hFZOX9f0hFZOW9g0iFYOW9f0jFZOV9f0jFZOV9e0lFZOT9f0lFZOT9f0lFZOT9e0mF[OS9e0mF\\OR9c0oF]OR9b0oF]OR9:cE_O[17S93iEDU19S9MmEJP19T9FRF1j09k9GVF8j9GWF9i9GWF9i9GWF9\\;0001O00000000001O000000001O003MfZa5"}, "label": "a blue chair"}]}
{"id": 564302, "image": "COCO_train2014_000000564302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller desk chair that is blue in color\"."}], "task": "refering", "answer_template": "The \"the smaller desk chair that is blue in color\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 235, 236, 237, 256, 257, 258, 279, 281], "bbox": [0.1399375, 0.4015566037735849, 0.33049999999999996, 0.8150471698113207], "iscrowd": 0, "area": 8076.116349999996, "rle": {"size": [424, 640], "counts": "Q^U15R=7J5K2M3N2N1N3N3M5K5K5eGfN^4_1^HWOd1_Oj5e2TJ\\Mk5e2TJ\\Ml5d2SJ^Ml5b2TJ^Ml5b2SJ_Mm5b2QJ`Mn5`2QJ_MQ6a2nI_MS6a2mI^MT6b2kI^MV6c2hI]MY6c2fI\\M[6e2eIZM\\6f2cIZM^6f2aIZM`6f2_IZMb6f2]IZMd6f2\\IZMd6f2[IZMf6f2YIZMh6f2WI[Mi6e2WIZMj6f2UI[Mk6e2TI[Ml6f2SIZMn6f2QI[Mo6e2QIZMP7f2oH[MQ7e2nH[MS7e2lH[MU7e2kH[MU7e2jH[MW7e2hH\\MX7d2gH\\MZ7d2eH\\M[7e2eH[M[7e2dH[M]7e2bH[Mf7^2YHcMV8n1iGRNf8`1ZG`NT9R1kFnN[9m0dFSO\\9S1^FnNb9Z1VFeNk9b1mE_NS:h1eEXN\\:Y2100O13M3M3LWNmEb0P:^OUF`0h9@]F=a9BeF<X9DmF:P9EWG8f8H\\GWOSOa0^98_GWOXO?V99dGXOZO;Q9=eGXO^O7m8`0fGYOB2g8f0gGXOGMb8j0hGYOKH]8o0iGXOODX8T1iGXO\\9g0eFYO[9g0eFYO[9f0fFZOZ9f0fFZOZ9e0hFZOX9f0hFZOW9g0iFYOW9f0jFZOV9f0jFZOV9e0lFZOT9f0lFZOT9f0lFZOT9e0mF[OS9e0mF\\OR9c0oF]OR9b0oF]OR9:cE_O[17S93iEDU19S9MmEJP19T9FRF1j09k9GVF8j9GWF9i9GWF9i9GWF9\\;0001O00000000001O000000001O003MfZa5"}, "label": "the smaller desk chair that is blue in color"}]}
{"id": 564302, "image": "COCO_train2014_000000564302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of a dark colored desk chair\"."}], "task": "refering", "answer_template": "The \"back of a dark colored desk chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 317, 319, 320, 340, 343], "bbox": [0.686375, 0.23311320754716983, 0.9877812499999999, 0.9898584905660378], "iscrowd": 0, "area": 34251.19924999999, "rle": {"size": [424, 640], "counts": "_Tf52Q=6J5J7O1O0O2O1O0O2OO1BfCMZ<3mCFT<:WMJYI7f6OUI0k66PIJo6<kHDV7a0eH_OZ7g0aHYO^7m0\\HSOd7R1XHnNg7S1YHlNd7Y1ZHhN[7c1eH\\NP7P2PIPNd6\\2[IeMZ6f2fIYMo5T3oImLe5_3[J`L[5k3eJULo4W4PKjKe4a4[K^KZ4o4eKQKW4S5hKnJW4S5iKmJV4T5jKlJU4U5jKlJU4V5jKjJU4W5kKiJT4X5lKhJS4Y5eK\\JmN;^5Y5bKaJnN6_5[5_KdJPO1`5\\5^KfJQON`5]5_KeJQON_5^5`KdJQON^5_5aKdJPOM^5`5bKcJPOM]5b5bK]JUO1X5c5cKYJXO4T5d5dKXJXO4S5e5eKXJWO3S5f5fKXJVO2S5h5fKVJWO2S5h5fKWJVO1S5i5hKVJTO1S5j5jKTJSO2R5k5kKTJRO1S5l5kKSJQO1T5l5lKRJPO2e1Eg1W6eMRJnN2`1Ll1P6fM_J7HS2i5gM^JL4\\2_5hM\\JC>e2V5hM\\J[Og0l2m4iM\\JUOm0R3g4iM\\JoNT1W3`4jMXLV2i3iMXLV2h3kMVLV2j3jMVLV2j3jMVLV2W4]MhKd2j4jLeJ2\\NS3o8YMiFg2S9i0L4M3L4L4L4M3L4L4_NQKgJQ5Y5UKaJk4_5[K[Je4d5bKVJ^4j5hKPJX4P6kKmIU4S6kKmIU4S6kKnIT4R6kKoIU4Q6kKoIU4Q6kKoIU4Q6jKPJV4P6jKPJV4o5kKdIUO3P5Y6kKdIUO3P5Y6jKeIVO3o4X6kKeIWO2n4Y6kKdIXO3m4Y6jKeIYO2m4Y6jKeIZO1l4Y6kKeIZO2k4T6oKkIVO1k4P6SLoISO1i4P6TLoISO1i4P6TLnITO3g4o5TLoIVO1f4P6TLoIVO1f4o5ULoIVO3d4n5ULPJXO1c4o5ULPJXO1c4o5ULPJXO2b4n5ULPJ[O1`4o5ULPJ[O2_4n5VLPJ\\O1^4n5WLPJ\\O2]4n5WLPJ\\O3\\4m5XLPJ]O2[4n5YLoI\\O3[4n5YLnI]O5Y4m5ZLnI^O4X4n5ZLnI^O4X4m5\\LUJUOO^4l5]LVJUOM^4m5]LXJSOL^4m5_LXJSOJ^4n5`LYJPOI`4n5`LYJQOI^4m5bLYJQOJ]4m5bLXJROK\\4m5eLTJQOOY4m5jLoInN4X4m5nLjIlN9U4m5_MSJa2m5`MRJ`2m5aMTJ^2l5cMSJ]2m5cMSJ]2m5dMSJ[2m5eMSJ[2m5fMRJZ2n5fMSJY2l5iMSJW2m5iMTJV2l5jMTJV2l5kMSJU2m5kMTJT2l5mMUJQ2j5PNWJo1h5SN]Jg1a5[NeJ_1Z5cNhJZ1V5hNlJV1R5mNnJR1Q5oNPKP1n4SORKl0m4UOTKj0j4YOVKf0i4\\OWKc0g4_OXKb0g4@WKa0g4BWK?h4CVK?g4HSK:c43XKN`4?ZKC]4i4F:G9G9K5J6J7I6J6J6K5J7I6J6J6J6K6I6J6J6K5J6JVV3"}, "label": "back of a dark colored desk chair"}]}
{"id": 564302, "image": "COCO_train2014_000000564302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a velvet material office chair sitting in front of a computer on a hard wooden floor\"."}], "task": "refering", "answer_template": "The \"a velvet material office chair sitting in front of a computer on a hard wooden floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 317, 319, 320, 340, 343], "bbox": [0.686375, 0.23311320754716983, 0.9877812499999999, 0.9898584905660378], "iscrowd": 0, "area": 34251.19924999999, "rle": {"size": [424, 640], "counts": "_Tf52Q=6J5J7O1O0O2O1O0O2OO1BfCMZ<3mCFT<:WMJYI7f6OUI0k66PIJo6<kHDV7a0eH_OZ7g0aHYO^7m0\\HSOd7R1XHnNg7S1YHlNd7Y1ZHhN[7c1eH\\NP7P2PIPNd6\\2[IeMZ6f2fIYMo5T3oImLe5_3[J`L[5k3eJULo4W4PKjKe4a4[K^KZ4o4eKQKW4S5hKnJW4S5iKmJV4T5jKlJU4U5jKlJU4V5jKjJU4W5kKiJT4X5lKhJS4Y5eK\\JmN;^5Y5bKaJnN6_5[5_KdJPO1`5\\5^KfJQON`5]5_KeJQON_5^5`KdJQON^5_5aKdJPOM^5`5bKcJPOM]5b5bK]JUO1X5c5cKYJXO4T5d5dKXJXO4S5e5eKXJWO3S5f5fKXJVO2S5h5fKVJWO2S5h5fKWJVO1S5i5hKVJTO1S5j5jKTJSO2R5k5kKTJRO1S5l5kKSJQO1T5l5lKRJPO2e1Eg1W6eMRJnN2`1Ll1P6fM_J7HS2i5gM^JL4\\2_5hM\\JC>e2V5hM\\J[Og0l2m4iM\\JUOm0R3g4iM\\JoNT1W3`4jMXLV2i3iMXLV2h3kMVLV2j3jMVLV2j3jMVLV2W4]MhKd2j4jLeJ2\\NS3o8YMiFg2S9i0L4M3L4L4L4M3L4L4_NQKgJQ5Y5UKaJk4_5[K[Je4d5bKVJ^4j5hKPJX4P6kKmIU4S6kKmIU4S6kKnIT4R6kKoIU4Q6kKoIU4Q6kKoIU4Q6jKPJV4P6jKPJV4o5kKdIUO3P5Y6kKdIUO3P5Y6jKeIVO3o4X6kKeIWO2n4Y6kKdIXO3m4Y6jKeIYO2m4Y6jKeIZO1l4Y6kKeIZO2k4T6oKkIVO1k4P6SLoISO1i4P6TLoISO1i4P6TLnITO3g4o5TLoIVO1f4P6TLoIVO1f4o5ULoIVO3d4n5ULPJXO1c4o5ULPJXO1c4o5ULPJXO2b4n5ULPJ[O1`4o5ULPJ[O2_4n5VLPJ\\O1^4n5WLPJ\\O2]4n5WLPJ\\O3\\4m5XLPJ]O2[4n5YLoI\\O3[4n5YLnI]O5Y4m5ZLnI^O4X4n5ZLnI^O4X4m5\\LUJUOO^4l5]LVJUOM^4m5]LXJSOL^4m5_LXJSOJ^4n5`LYJPOI`4n5`LYJQOI^4m5bLYJQOJ]4m5bLXJROK\\4m5eLTJQOOY4m5jLoInN4X4m5nLjIlN9U4m5_MSJa2m5`MRJ`2m5aMTJ^2l5cMSJ]2m5cMSJ]2m5dMSJ[2m5eMSJ[2m5fMRJZ2n5fMSJY2l5iMSJW2m5iMTJV2l5jMTJV2l5kMSJU2m5kMTJT2l5mMUJQ2j5PNWJo1h5SN]Jg1a5[NeJ_1Z5cNhJZ1V5hNlJV1R5mNnJR1Q5oNPKP1n4SORKl0m4UOTKj0j4YOVKf0i4\\OWKc0g4_OXKb0g4@WKa0g4BWK?h4CVK?g4HSK:c43XKN`4?ZKC]4i4F:G9G9K5J6J7I6J6J6K5J7I6J6J6J6K6I6J6J6K5J6JVV3"}, "label": "a velvet material office chair sitting in front of a computer on a hard wooden floor"}]}
{"id": 23631, "image": "COCO_train2014_000000023631.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike on air which is numbered 56\"."}], "task": "refering", "answer_template": "The \"a bike on air which is numbered 56\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 243], "bbox": [0.432625, 0.2831381733021077, 0.62546875, 0.7210070257611242], "iscrowd": 0, "area": 12295.770599999998, "rle": {"size": [427, 640], "counts": "\\fc33V=4L4K4M4cDM\\96YF=]9EYFl0^9VO`Fm0^9SOaFP1\\9RObFP1W9WOgFl0V9VOhFm0V9TOhFo0V9QOiFR1T9POiFT1U9lNjFV1]9cNaF_1T:bNlEh0_:jNdEU1^:hNdEX1\\:eNfE\\1_:\\NdEd1Q;1N2O000O100O10000O0I8L4K4D<O1N1BXM[Fi2W9l0O11N2O1O0M3M3LROfLTHj2[OVMe80oGj2^8WMaGh2a8WM_Gi2b8WM]Gi2d8WMfFO=k2m8VMdF3=f2P9WMaF8;b2U9UM_F<9_2U9XMcF<4]2Z9VMaFa3`953M00M3O1000000000000O1000000000O1O1O1O12N6J9F6K5L2N1N2O1O1N2N2N1H7OWKnG:0Y3Q8]LoGN=d3c7_LkHa3T7^LnHa3Q7`LPI`3o6`LRI`3l6aLTI_3l6aLUI_3i6bLXIJoNM4m2e7\\MYIIQOH:l2\\7dMZIFCd2S7fM[IEDb2Q7jM]IBCb2Q7lM^I@Cb2n6oM_Jo1b5QN`Jm1`5SNaJk1`5UNbJh1_5XNdJe1\\5[NgJa1Z5_NiJZ1[5fNhJR1]5nNlJe0X5[ORK7S5HVKKn45U4O2N1O2N4I_nS3"}, "label": "a bike on air which is numbered 56"}]}
{"id": 23631, "image": "COCO_train2014_000000023631.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike that is in the air\"."}], "task": "refering", "answer_template": "The \"bike that is in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 243], "bbox": [0.432625, 0.2831381733021077, 0.62546875, 0.7210070257611242], "iscrowd": 0, "area": 12295.770599999998, "rle": {"size": [427, 640], "counts": "\\fc33V=4L4K4M4cDM\\96YF=]9EYFl0^9VO`Fm0^9SOaFP1\\9RObFP1W9WOgFl0V9VOhFm0V9TOhFo0V9QOiFR1T9POiFT1U9lNjFV1]9cNaF_1T:bNlEh0_:jNdEU1^:hNdEX1\\:eNfE\\1_:\\NdEd1Q;1N2O000O100O10000O0I8L4K4D<O1N1BXM[Fi2W9l0O11N2O1O0M3M3LROfLTHj2[OVMe80oGj2^8WMaGh2a8WM_Gi2b8WM]Gi2d8WMfFO=k2m8VMdF3=f2P9WMaF8;b2U9UM_F<9_2U9XMcF<4]2Z9VMaFa3`953M00M3O1000000000000O1000000000O1O1O1O12N6J9F6K5L2N1N2O1O1N2N2N1H7OWKnG:0Y3Q8]LoGN=d3c7_LkHa3T7^LnHa3Q7`LPI`3o6`LRI`3l6aLTI_3l6aLUI_3i6bLXIJoNM4m2e7\\MYIIQOH:l2\\7dMZIFCd2S7fM[IEDb2Q7jM]IBCb2Q7lM^I@Cb2n6oM_Jo1b5QN`Jm1`5SNaJk1`5UNbJh1_5XNdJe1\\5[NgJa1Z5_NiJZ1[5fNhJR1]5nNlJe0X5[ORK7S5HVKKn45U4O2N1O2N4I_nS3"}, "label": "bike that is in the air"}]}
{"id": 171086, "image": "COCO_train2014_000000171086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple that has a stick in it on the bottom left hand side\"."}], "task": "refering", "answer_template": "The \"the apple that has a stick in it on the bottom left hand side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 300, 301, 302, 303, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 369, 370, 371, 372, 373], "bbox": [0.021953125, 0.748875, 0.261828125, 0.9831041666666667], "iscrowd": 0, "area": 11638.00555, "rle": {"size": [480, 640], "counts": "Xo61n>2M4M2N3M2N3M2M4M2N3M2N3M2M4M2N2N3M2N3L3M4M2M3N2M2N3N1N3M2O2M2O2M3M2O2M2O2M2O2O001O0O2O001O1O0O2O000000O10O10000000000O10000000O010000000000O10000000O010000000000O1000000000O0100000000O10000000000O03N1O1O2N1O2N1O2N1O1O2N1O2N1O1N3N1O2N1O2N1O1O2N1O2N1O2N1O1N3N1N3M2N3M2N2M4M2N3M2M3N3M2N3L3N3M2N2N3L3N3M2N3L_Ym6"}, "label": "the apple that has a stick in it on the bottom left hand side"}]}
{"id": 293974, "image": "COCO_train2014_000000293974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant wearing saddle\"."}], "task": "refering", "answer_template": "The \"elephant wearing saddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 267, 268, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 312, 313, 314, 324, 325, 326, 328, 329, 331, 332, 333, 334, 335, 336, 347, 348, 349, 350, 351, 352, 354, 355, 356, 357, 358, 370, 371, 373, 374, 375, 377, 378, 379, 380, 396, 397, 398, 400, 401, 402], "bbox": [0.0864375, 0.11793774319066147, 0.705890625, 0.967704280155642], "iscrowd": 0, "area": 113864.50634999998, "rle": {"size": [514, 640], "counts": "bQl01n?5L4K5L4K6J5L3^OQOlAR1P>ZOeAh0^8V1nJ^On4o0eJTOX5Z1YJkNc5`1QJdNk5f1fIdNV6e1nHQOo6S5L4L4L4M300O1O1O1O1O100O1O1O1O1O100O1O1O1O2NiNWIWIi6a6cI[I\\6]6RJ^Im5[6_JaI`5X6lJdIS5W6WKeIh4Y6UKXI^N:]6[6nJeIhNMY6\\6hJRJPO@W6\\6dJ\\JWOVOT6\\6`JfJ]OmNR6\\6[JoJEcNo5\\6WJZKKXNn5[6TJcK^6[4_IkK`6S4]ISLb6k3\\I[Lb6c3[IcLd6[3ZIjLf6S3WISMh6k2UI\\Mi6o5O1O100O2N1O100O2N1O100O2N1O100O1O1O010O1O10O01OSOkIoGT6_7aJ_H_5Z7jJcHV5W7SKgHl4S7\\KlHc4n6fKPIZ4j6nKSIR4j6TLTIk3i6ZLVIe3g6aLWI_3e6gLXIZ3c6mL[IR3b6SM]Im2a6WM]Ih2d6aN\\I6e6E`I:`6AfI=[6_OjI`0V6[OPJc0Q6YOTJf0l5UOYJk0g5QO^Jm0c5QO`Jn0`5POcJo0]5nNgJP1Z5nNiJQ1W5lNlJS1U5kNmJU1S5hNPKX1P5fNRKY1o4dNTK\\1l4bNVK]1k4`NXK`1h4^NZKb1f4[N]Kd1d4ZN^Kf1b4TNdKk1^4kMkKU2U4bMTL^2l3XM^Lg2d3fLnLZ3S3RL`Mm3[70000O10000O100O10000O100O10000O100O10O10OPOXLhEh3V:\\LhEc3W:aLgE_3W:fLfEZ3X:jLfEU3Y:oLeEQ3Y:SMeEl2Z:YMcEg2[:]McEc2[:aMcE^2\\:gMaEY2]:kMaET2^:PN`EP2^:k1N1I8G8G:J6M2N3M2N3M3M2N3M2N3M2N3M3M2N3L3N3M2N3M2N3M2N3L4M2O3000O101O0O10001N10000O2O0O1O1O2M2O1N3XLTHVOn7e0UH[Om7`0VH_Om7;XHDk76XHJj71YHNj7LZH4h7G[H9g7B]H<f7^O^Hb0d7VObHj0a7fNlHY1W7XNWIg1k6iMcIV2`6nLZJR3[900O2O0001O000000001O0000001OO1O1O1O1O1O1O1O100O001O1O1O1O1O1O1O1O1O01SOVMbDi2T;dMjD[2l:QNSEn1h:YNVEg1g:^NXEa1e:dNZE[1c:jN\\EV1`:PO]EQ1_:TO`El0\\:YOcEg0Y:^OfEa0W:DgE=U:HjE8k8cMgG[2\\O2j8Y1UGgNh8]1VGcNh8a1WG_Nf8e1YG[Nd8=[GVNO]1c8?`Gl0]8VOfGh0W8YOlGe0R8]OQHa0l7@WH?Y70jHNb6d0aI[OY6j0jISOV6m0lIROT6l0PJROo5n0SJQOm5m0TJTOk5l0UJTOk5l0VJTOj5k0VJVOi5i0YJWOg5h0YJYOf5g0[JYOd5f0]J[Oc5d0^J\\Oa5d0_J]O`5b0aJ_O_5`0bJA\\5?dJB\\5<fJDY5<gJEZ59gJGY57hJKW54jJLW52hJ0X5NiJ3W5LiJ5W5JiJ7X5FiJ<V5CjJ>V5AiJa0X5]OhJd0X5ZOiJg0W5XOiJj0W5TOiJm0W5QOjJP1V5oNiJS1X5kNhJV1X5hNiJZ1V5eNjJ\\1V5cNjJ^1W5_NjJb1V5]NiJe1W5ZNiJg1X5VNiJl1V5SNjJn1V5QNjJP2W5mMiJU2W5jMiJW2W5hMiJZ2W5cMjJ_2V5_MjJc2V5XMlJk2U5nLoJT3Q5gLRK[3n4`LTKb3n4XLUKj3k4QLXKQ4h4iK[KZ4f4`K]Kb4c4YK`Ki4c4oJ_KT5e4cJ^K_5f4WJ^Kk5Z73N2N2M3N2N2M3N2N2M3N3M3L4M2N3L4M3M3L4M3M3L4M3M3L4M3M3L4M2L5K4K6K4K5K6K4K6K4K5L5J5L5]KjER3[:cLUFS3Q:`L`FW3];F:G9F:G8G:G_Tn2"}, "label": "elephant wearing saddle"}]}
{"id": 293974, "image": "COCO_train2014_000000293974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant wearing a decorative chair on its back\"."}], "task": "refering", "answer_template": "The \"an elephant wearing a decorative chair on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 267, 268, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 312, 313, 314, 324, 325, 326, 328, 329, 331, 332, 333, 334, 335, 336, 347, 348, 349, 350, 351, 352, 354, 355, 356, 357, 358, 370, 371, 373, 374, 375, 377, 378, 379, 380, 396, 397, 398, 400, 401, 402], "bbox": [0.0864375, 0.11793774319066147, 0.705890625, 0.967704280155642], "iscrowd": 0, "area": 113864.50634999998, "rle": {"size": [514, 640], "counts": "bQl01n?5L4K5L4K6J5L3^OQOlAR1P>ZOeAh0^8V1nJ^On4o0eJTOX5Z1YJkNc5`1QJdNk5f1fIdNV6e1nHQOo6S5L4L4L4M300O1O1O1O1O100O1O1O1O1O100O1O1O1O2NiNWIWIi6a6cI[I\\6]6RJ^Im5[6_JaI`5X6lJdIS5W6WKeIh4Y6UKXI^N:]6[6nJeIhNMY6\\6hJRJPO@W6\\6dJ\\JWOVOT6\\6`JfJ]OmNR6\\6[JoJEcNo5\\6WJZKKXNn5[6TJcK^6[4_IkK`6S4]ISLb6k3\\I[Lb6c3[IcLd6[3ZIjLf6S3WISMh6k2UI\\Mi6o5O1O100O2N1O100O2N1O100O2N1O100O1O1O010O1O10O01OSOkIoGT6_7aJ_H_5Z7jJcHV5W7SKgHl4S7\\KlHc4n6fKPIZ4j6nKSIR4j6TLTIk3i6ZLVIe3g6aLWI_3e6gLXIZ3c6mL[IR3b6SM]Im2a6WM]Ih2d6aN\\I6e6E`I:`6AfI=[6_OjI`0V6[OPJc0Q6YOTJf0l5UOYJk0g5QO^Jm0c5QO`Jn0`5POcJo0]5nNgJP1Z5nNiJQ1W5lNlJS1U5kNmJU1S5hNPKX1P5fNRKY1o4dNTK\\1l4bNVK]1k4`NXK`1h4^NZKb1f4[N]Kd1d4ZN^Kf1b4TNdKk1^4kMkKU2U4bMTL^2l3XM^Lg2d3fLnLZ3S3RL`Mm3[70000O10000O100O10000O100O10000O100O10O10OPOXLhEh3V:\\LhEc3W:aLgE_3W:fLfEZ3X:jLfEU3Y:oLeEQ3Y:SMeEl2Z:YMcEg2[:]McEc2[:aMcE^2\\:gMaEY2]:kMaET2^:PN`EP2^:k1N1I8G8G:J6M2N3M2N3M3M2N3M2N3M2N3M3M2N3L3N3M2N3M2N3M2N3L4M2O3000O101O0O10001N10000O2O0O1O1O2M2O1N3XLTHVOn7e0UH[Om7`0VH_Om7;XHDk76XHJj71YHNj7LZH4h7G[H9g7B]H<f7^O^Hb0d7VObHj0a7fNlHY1W7XNWIg1k6iMcIV2`6nLZJR3[900O2O0001O000000001O0000001OO1O1O1O1O1O1O1O100O001O1O1O1O1O1O1O1O1O01SOVMbDi2T;dMjD[2l:QNSEn1h:YNVEg1g:^NXEa1e:dNZE[1c:jN\\EV1`:PO]EQ1_:TO`El0\\:YOcEg0Y:^OfEa0W:DgE=U:HjE8k8cMgG[2\\O2j8Y1UGgNh8]1VGcNh8a1WG_Nf8e1YG[Nd8=[GVNO]1c8?`Gl0]8VOfGh0W8YOlGe0R8]OQHa0l7@WH?Y70jHNb6d0aI[OY6j0jISOV6m0lIROT6l0PJROo5n0SJQOm5m0TJTOk5l0UJTOk5l0VJTOj5k0VJVOi5i0YJWOg5h0YJYOf5g0[JYOd5f0]J[Oc5d0^J\\Oa5d0_J]O`5b0aJ_O_5`0bJA\\5?dJB\\5<fJDY5<gJEZ59gJGY57hJKW54jJLW52hJ0X5NiJ3W5LiJ5W5JiJ7X5FiJ<V5CjJ>V5AiJa0X5]OhJd0X5ZOiJg0W5XOiJj0W5TOiJm0W5QOjJP1V5oNiJS1X5kNhJV1X5hNiJZ1V5eNjJ\\1V5cNjJ^1W5_NjJb1V5]NiJe1W5ZNiJg1X5VNiJl1V5SNjJn1V5QNjJP2W5mMiJU2W5jMiJW2W5hMiJZ2W5cMjJ_2V5_MjJc2V5XMlJk2U5nLoJT3Q5gLRK[3n4`LTKb3n4XLUKj3k4QLXKQ4h4iK[KZ4f4`K]Kb4c4YK`Ki4c4oJ_KT5e4cJ^K_5f4WJ^Kk5Z73N2N2M3N2N2M3N2N2M3N3M3L4M2N3L4M3M3L4M3M3L4M3M3L4M3M3L4M2L5K4K6K4K5K6K4K6K4K5L5J5L5]KjER3[:cLUFS3Q:`L`FW3];F:G9F:G8G:G_Tn2"}, "label": "an elephant wearing a decorative chair on its back"}]}
{"id": 293974, "image": "COCO_train2014_000000293974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant by white truck\"."}], "task": "refering", "answer_template": "The \"elephant by white truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 247, 248, 249, 250, 272, 295], "bbox": [0.709078125, 0.3958560311284047, 0.904875, 0.6794747081712063], "iscrowd": 0, "area": 8991.314499999999, "rle": {"size": [514, 640], "counts": "^ST73m?4M4L3M3M3N4K4L5L3L2N3N2M3N2M2N3N2M3M3N2M2N3N2M3M3N1N3M3N2M3;B^1bNf^3XOSbL7Io0QOo0QO<DO1O100O1O100O1O1O100O1001O2N1O100O2N1O1O2N6J3M2N1O2N1O2N1ZDgLa:Z3]EgLc:Y3\\EhLd:Y3[EgLe:Y3ZEhLf:X3YEiLh:V3WElLh:U3WEkLi:c2`DdM`0E_O5a;_2hDbM85Q;U2oDaM0:Q;W2eEiM[:Z2bEfM^:]2_EcMb:_2[EaMe:c2WE]Mi:j2PEVMQ;k2mDUMS;j2nDVMR;j2nDUMS;k2mDUMT;j2mDUMS;k2mDUMS;k2mDUMS;k2nDTMS;k2nDTMR;l2oDSMQ;m2QEPMP;P3Q100O1000000000001N100YOg0J7I8H8I6I8M6K6I6K5J6K5Jo[n0"}, "label": "elephant by white truck"}]}
{"id": 293974, "image": "COCO_train2014_000000293974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray elephant far back behind another elephant\"."}], "task": "refering", "answer_template": "The \"a gray elephant far back behind another elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 247, 248, 249, 250, 272, 295], "bbox": [0.709078125, 0.3958560311284047, 0.904875, 0.6794747081712063], "iscrowd": 0, "area": 8991.314499999999, "rle": {"size": [514, 640], "counts": "^ST73m?4M4L3M3M3N4K4L5L3L2N3N2M3N2M2N3N2M3M3N2M2N3N2M3M3N1N3M3N2M3;B^1bNf^3XOSbL7Io0QOo0QO<DO1O100O1O100O1O1O100O1001O2N1O100O2N1O1O2N6J3M2N1O2N1O2N1ZDgLa:Z3]EgLc:Y3\\EhLd:Y3[EgLe:Y3ZEhLf:X3YEiLh:V3WElLh:U3WEkLi:c2`DdM`0E_O5a;_2hDbM85Q;U2oDaM0:Q;W2eEiM[:Z2bEfM^:]2_EcMb:_2[EaMe:c2WE]Mi:j2PEVMQ;k2mDUMS;j2nDVMR;j2nDUMS;k2mDUMT;j2mDUMS;k2mDUMS;k2mDUMS;k2nDTMS;k2nDTMR;l2oDSMQ;m2QEPMP;P3Q100O1000000000001N100YOg0J7I8H8I6I8M6K6I6K5J6K5Jo[n0"}, "label": "a gray elephant far back behind another elephant"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a red outfit\"."}], "task": "refering", "answer_template": "The \"a woman in a red outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 46, 47, 49, 50, 51, 64, 65, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 119, 120, 121, 122, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 191, 192, 193, 209, 210, 211, 227, 228, 229, 230], "bbox": [0.5662999999999999, 0.14608000000000002, 0.8545, 0.9977599999999999], "iscrowd": 0, "area": 23684.132550000006, "rle": {"size": [375, 500], "counts": "bgW35Y;<K4N2N3L3N3M2N3L3N3M2N3C<L5N101N1O2O0O2N1O2M2O1D=L3M4L3O1bJTMo1j2eMfMY2Y2\\MUNb2j1ZM]Ng2^1UMjNP3n0lLYOU3b0hLEW39eLNZ30bL7]3F`La0_3]O\\Lk0c3SOYLT1f3hNXLj1[3SNcL`2P3\\MnLR3j2jLTMY3o2dLnL_3U3]LhLg3[3ULcLn3`3nK]LV4d3hKYL[4g3eKVL_4i3bKSLb4l3^KPLg4o3YKnKk4Q4UKlKo4S4QKjKS5U4X1O1O1O001O1O1O1O1O1O1OL4K5K5N22N1O1O2N1O2N1_OkH_LV7b2bHRNR8k1WHiMn7T2m0N10001O0O101O001N10001O1O1O2N1O1OWOWNeGh1P8cNQH\\1W7]OiHb0e61\\IMc65]IJb68]IHa6:_IG_6;`IE\\5ZNoJS2DCe4WO`KY1IBd4\\O]KT1N@c4CYKo03_Oa4IUKj0:]O_4a2`K`M]4c2aK^M]4e2bK\\M[4g2dKZMY4i2fKWMX4k2hKVMT4n2jKSMT4P3kKPMS4S3lKmLZ1QNR1X5cMgLZ1UNh0]5mM^LZ1ZN>`5VNWL[1^N4g7K_HIh76^H^Oi7a0]HSOj7l0a1O100N2O1O1N3N1N2O2N1N2O1O2L5J8I[ej0"}, "label": "a woman in a red outfit"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman playing a wii in red shorts\"."}], "task": "refering", "answer_template": "The \"a woman playing a wii in red shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 46, 47, 49, 50, 51, 64, 65, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 119, 120, 121, 122, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 191, 192, 193, 209, 210, 211, 227, 228, 229, 230], "bbox": [0.5662999999999999, 0.14608000000000002, 0.8545, 0.9977599999999999], "iscrowd": 0, "area": 23684.132550000006, "rle": {"size": [375, 500], "counts": "bgW35Y;<K4N2N3L3N3M2N3L3N3M2N3C<L5N101N1O2O0O2N1O2M2O1D=L3M4L3O1bJTMo1j2eMfMY2Y2\\MUNb2j1ZM]Ng2^1UMjNP3n0lLYOU3b0hLEW39eLNZ30bL7]3F`La0_3]O\\Lk0c3SOYLT1f3hNXLj1[3SNcL`2P3\\MnLR3j2jLTMY3o2dLnL_3U3]LhLg3[3ULcLn3`3nK]LV4d3hKYL[4g3eKVL_4i3bKSLb4l3^KPLg4o3YKnKk4Q4UKlKo4S4QKjKS5U4X1O1O1O001O1O1O1O1O1O1OL4K5K5N22N1O1O2N1O2N1_OkH_LV7b2bHRNR8k1WHiMn7T2m0N10001O0O101O001N10001O1O1O2N1O1OWOWNeGh1P8cNQH\\1W7]OiHb0e61\\IMc65]IJb68]IHa6:_IG_6;`IE\\5ZNoJS2DCe4WO`KY1IBd4\\O]KT1N@c4CYKo03_Oa4IUKj0:]O_4a2`K`M]4c2aK^M]4e2bK\\M[4g2dKZMY4i2fKWMX4k2hKVMT4n2jKSMT4P3kKPMS4S3lKmLZ1QNR1X5cMgLZ1UNh0]5mM^LZ1ZN>`5VNWL[1^N4g7K_HIh76^H^Oi7a0]HSOj7l0a1O100N2O1O1N3N1N2O2N1N2O1O2L5J8I[ej0"}, "label": "a woman playing a wii in red shorts"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt and yellow shorts\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt and yellow shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 111, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 216, 217, 218, 219], "bbox": [0.0, 0.07845333333333333, 0.22178, 0.98816], "iscrowd": 0, "area": 26594.84245, "rle": {"size": [375, 500], "counts": "Y2;P7W1UKAe1c3fMTMUOP6g0UJkNW6T1lIhNV6W1W2N1O2N1O2N00O1000000000000000000000O10000001O00001O000O101O00001O0000001O00001O000hGQOh5o0WJSOh5n0WJROi5n0WJSOh5m0XJTOg5m0XJTOg5m0XHeNm1?j5m0XHeNm1?i5m0ZHeNl1?i5Q1UJQOj5P1UJQOj5Q1SJQOl5P1RJROm5o0QJSOn5o0oISOP6n0nITOQ6n0lITOS6m0]ICb6j2O001O1O1O1O2N2N3M2N3L4M3M3M3M4L3M3M3M4L6J9G2QLRJQ2P6kM]Jj1e5SNgJb1[5ZNRK[1o4cN]KR1b4mNkKh0U4VOYL>f3BfL3Y3LTMHl28`M]O_2b0nMSOQ2l0]400O2O1O2N100O2N1O1O1O2N10_X^4"}, "label": "a man in a blue shirt and yellow shorts"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt and yellow shorts watches on\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt and yellow shorts watches on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 111, 126, 127, 128, 129, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 198, 199, 200, 201, 216, 217, 218, 219], "bbox": [0.0, 0.07845333333333333, 0.22178, 0.98816], "iscrowd": 0, "area": 26594.84245, "rle": {"size": [375, 500], "counts": "Y2;P7W1UKAe1c3fMTMUOP6g0UJkNW6T1lIhNV6W1W2N1O2N1O2N00O1000000000000000000000O10000001O00001O000O101O00001O0000001O00001O000hGQOh5o0WJSOh5n0WJROi5n0WJSOh5m0XJTOg5m0XJTOg5m0XHeNm1?j5m0XHeNm1?i5m0ZHeNl1?i5Q1UJQOj5P1UJQOj5Q1SJQOl5P1RJROm5o0QJSOn5o0oISOP6n0nITOQ6n0lITOS6m0]ICb6j2O001O1O1O1O2N2N3M2N3L4M3M3M3M4L3M3M3M4L6J9G2QLRJQ2P6kM]Jj1e5SNgJb1[5ZNRK[1o4cN]KR1b4mNkKh0U4VOYL>f3BfL3Y3LTMHl28`M]O_2b0nMSOQ2l0]400O2O1O2N100O2N1O1O1O2N10_X^4"}, "label": "a man in a blue shirt and yellow shorts watches on"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"television screen showing a golf game\"."}], "task": "refering", "answer_template": "The \"television screen showing a golf game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 130], "bbox": [0.09122, 0.18765333333333334, 0.44367999999999996, 0.5474933333333334], "iscrowd": 0, "area": 18460.651549999995, "rle": {"size": [375, 500], "counts": "Zn`03c;4L4M1N2N2O1N2N2N2O1N1O2O1N2N2K4G:M2M4M201O001O002N3M3M4L6J6J5K6J5K6J4L2N2N2N2N1O2N2N2N2N1O000000O010O10000O100O100002N2N2N1O2N1O2N00000000000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O10O0100O1O1J6I700O1O11O1O1O1O3M7I4L0000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000D<00000O100000000003Mg0YOi0WOZ1fNgdU3"}, "label": "television screen showing a golf game"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"monitor displaying golf game\"."}], "task": "refering", "answer_template": "The \"monitor displaying golf game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 130], "bbox": [0.09122, 0.18765333333333334, 0.44367999999999996, 0.5474933333333334], "iscrowd": 0, "area": 18460.651549999995, "rle": {"size": [375, 500], "counts": "Zn`03c;4L4M1N2N2O1N2N2N2O1N1O2O1N2N2K4G:M2M4M201O001O002N3M3M4L6J6J5K6J5K6J4L2N2N2N2N1O2N2N2N2N1O000000O010O10000O100O100002N2N2N1O2N1O2N00000000000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O10O0100O1O1J6I700O1O11O1O1O1O3M7I4L0000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000D<00000O100000000003Mg0YOi0WOZ1fNgdU3"}, "label": "monitor displaying golf game"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with braided hair wearing a white tank top and a blue skirt\"."}], "task": "refering", "answer_template": "The \"a girl with braided hair wearing a white tank top and a blue skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 69, 70, 86, 87, 88, 105, 106, 123, 124, 141, 142, 159, 160, 177, 178], "bbox": [0.8140599999999999, 0.16621333333333332, 0.9358799999999999, 0.7245333333333333], "iscrowd": 0, "area": 5824.042549999998, "rle": {"size": [375, 500], "counts": "[Ue43e;0O1^DN0O[;3dD1\\;8M00000001OaHFa3:]LIb37TJGf04U55SJIh03T54QJMj00T53oI1l0MT52mI5n0IU53jI6Q1GU53gI:S1CV54dI<T1BW52bI?V1@X52_Ic0U1\\O[52]Ii0S1VO`52ZIP1n0POg51TIc1g0]NS63oHo1d0oM[6R3\\InLd6\\3RIeLl6T4O1O2L4M2VOQKiJS5n4oJ_J1`0R5l4^KnJc4P5`KnJa4Q5bKkJ`4T5aKkJa4S5`KlJa4T5P11jMoJfMP5Z2VK`Mk4`2XK\\Mi4c2ZKZMf4g2]KTMe4l2eKiL\\4j2cLdL]3Z3fLeLY3Z3jLdLX3Y3kLeLV3Y3lLfLU3>bJQ2_2\\MP3a0fJm1S8PNRHh1T8UNQHV1b8hNaGd0R9YOQGd0S9XOPGe0T9VOPGg0V:I7Efc;"}, "label": "a girl with braided hair wearing a white tank top and a blue skirt"}]}
{"id": 310360, "image": "COCO_train2014_000000310360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the white tank top with reflective shoes\"."}], "task": "refering", "answer_template": "The \"the woman in the white tank top with reflective shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 69, 70, 86, 87, 88, 105, 106, 123, 124, 141, 142, 159, 160, 177, 178], "bbox": [0.8140599999999999, 0.16621333333333332, 0.9358799999999999, 0.7245333333333333], "iscrowd": 0, "area": 5824.042549999998, "rle": {"size": [375, 500], "counts": "[Ue43e;0O1^DN0O[;3dD1\\;8M00000001OaHFa3:]LIb37TJGf04U55SJIh03T54QJMj00T53oI1l0MT52mI5n0IU53jI6Q1GU53gI:S1CV54dI<T1BW52bI?V1@X52_Ic0U1\\O[52]Ii0S1VO`52ZIP1n0POg51TIc1g0]NS63oHo1d0oM[6R3\\InLd6\\3RIeLl6T4O1O2L4M2VOQKiJS5n4oJ_J1`0R5l4^KnJc4P5`KnJa4Q5bKkJ`4T5aKkJa4S5`KlJa4T5P11jMoJfMP5Z2VK`Mk4`2XK\\Mi4c2ZKZMf4g2]KTMe4l2eKiL\\4j2cLdL]3Z3fLeLY3Z3jLdLX3Y3kLeLV3Y3lLfLU3>bJQ2_2\\MP3a0fJm1S8PNRHh1T8UNQHV1b8hNaGd0R9YOQGd0S9XOPGe0T9VOPGg0V:I7Efc;"}, "label": "the woman in the white tank top with reflective shoes"}]}
{"id": 244825, "image": "COCO_train2014_000000244825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in a white jersey hitting a ball\"."}], "task": "refering", "answer_template": "The \"a baseball player in a white jersey hitting a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 242, 243, 244, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 332, 333, 334, 335, 336, 337, 354, 355, 356, 358, 359, 360, 361, 362, 376, 377, 378, 379, 383, 384, 400, 401, 406, 407], "bbox": [0.355796875, 0.23512027491408935, 0.7464375000000001, 0.8502749140893471], "iscrowd": 0, "area": 31339.927450000014, "rle": {"size": [582, 640], "counts": "SiQ4:ja04L4L4N2M2O001O1iEWOo2i0QM[Ok2f0SM_Oi2a0WMBf2?YMEc2:^MGa28_MIa26`MJ`26`MK`24`MLa23^MNb23]MMd22\\MNe21[MOe22YMOh20XM0i20VM0k2OUM1j21TM0l20TM0k22TMNl22SMOm22RMNm23SMMm23RMNn23QMMn24RMLn24QMMo24PMLo25QMKo25PMLP34PMLP35nLLQ35mLMS33lLNT32jL0V30hL3W3NfL4Z3LdL6\\3JcL7]3IbL8^3HbL8^3I`L8`3H_L9a3G^L:b3F]L;c3F[L;d3F\\L:d3F[L;e3EZL<f3DYL=g3DWL=i3CVL>j3BUL?k3AUL?k3BSL?m3ARL`0n3@PLb0o3_OPLb0P4^OoKc0Q4^OmKc0S4]OlKd0S4]OkKe0U4[OjKf0V4[OhKf0X4ZOgKh0X4XOfKj0Y4WOfKi0[4WOdKe0a4[O^Kb0f4^OYK?j4BTK<P5DoJ9U5GjJ6Z5JdJ5^5L_J4d5LZJ4h5LUJ5m5KPJ5S6KkI5V6LgI4\\6LbI5_6K^I8b6H[I:f6FXI;i6ETI=l6DRI=o6CnH?S7AjHa0W7_OgHb0Z7^OcHd0^7\\O`He0a7_OYHb0h7BQH`0P8DjG=_7gM[FQ2P29e7jMXFQ2m16k7mMUFQ2j13Q8PNRFQ2f11X8QNPFR2b1N_8RNlEU2_1Je8RNlEW2Z1Ii8QNlE[2T1FP9PNkE\\2R1DS9RNjE\\2o0DW9QNiE]2k0D\\9QNhE]2f0Ea9QNhE]2`0Ef9_1UFdNj9_1RFbNP:_1kEfNS:i4M3N2M3M3N1N2N2O0O2O1O001O1O001O1O001O1O001O1O001O1O00100O00100O001O10O01O01\\NUGlIj8o5_GmIa8P6bGQJ]8m5fGSJZ8f5lGZJT8R5QGoIT1o0k7P5SGnIT1R1j7n4UGkIU1W1f7l4dHTK\\7k4fHUKZ7j4fHVK[7h4gHWKZ7g4hHXKY7g4gHZKY7d4iH[KX7c4iH]KX7]OlF]4n1ULW7UOTGe4e1WLW7kN\\Gc4B_Km1R1l7Y3cHgL^7V3eHiL\\7T3fHlL[7Q3hHoLX7o2jHPMW7m2kHSMW7j2kHUMV7j2jHWMW7f2kHYMV7f1VFhMe2b0W7b1\\FdM`2i0U7CfEV1m0WNZ2Q1T7oNXFe1g3[Oi6a0YI_Oi6<[IDf68\\IHf63^IMc6O_I1d6J_I6b6EaI;V<0O1O10O0100O100O00100O10O010O0100O001k_OPOS?U1e@mN[?X1\\@mNd?j101O01N101O00001N101O00001N1O2K4M3L5L3L5M2M3M4M2M4L3N2M4L3N3L3L5Kmdk2"}, "label": "a baseball player in a white jersey hitting a ball"}]}
{"id": 244825, "image": "COCO_train2014_000000244825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in a # 9 white uniform with a black cap swinging a bat\"."}], "task": "refering", "answer_template": "The \"a baseball player in a # 9 white uniform with a black cap swinging a bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 242, 243, 244, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 332, 333, 334, 335, 336, 337, 354, 355, 356, 358, 359, 360, 361, 362, 376, 377, 378, 379, 383, 384, 400, 401, 406, 407], "bbox": [0.355796875, 0.23512027491408935, 0.7464375000000001, 0.8502749140893471], "iscrowd": 0, "area": 31339.927450000014, "rle": {"size": [582, 640], "counts": "SiQ4:ja04L4L4N2M2O001O1iEWOo2i0QM[Ok2f0SM_Oi2a0WMBf2?YMEc2:^MGa28_MIa26`MJ`26`MK`24`MLa23^MNb23]MMd22\\MNe21[MOe22YMOh20XM0i20VM0k2OUM1j21TM0l20TM0k22TMNl22SMOm22RMNm23SMMm23RMNn23QMMn24RMLn24QMMo24PMLo25QMKo25PMLP34PMLP35nLLQ35mLMS33lLNT32jL0V30hL3W3NfL4Z3LdL6\\3JcL7]3IbL8^3HbL8^3I`L8`3H_L9a3G^L:b3F]L;c3F[L;d3F\\L:d3F[L;e3EZL<f3DYL=g3DWL=i3CVL>j3BUL?k3AUL?k3BSL?m3ARL`0n3@PLb0o3_OPLb0P4^OoKc0Q4^OmKc0S4]OlKd0S4]OkKe0U4[OjKf0V4[OhKf0X4ZOgKh0X4XOfKj0Y4WOfKi0[4WOdKe0a4[O^Kb0f4^OYK?j4BTK<P5DoJ9U5GjJ6Z5JdJ5^5L_J4d5LZJ4h5LUJ5m5KPJ5S6KkI5V6LgI4\\6LbI5_6K^I8b6H[I:f6FXI;i6ETI=l6DRI=o6CnH?S7AjHa0W7_OgHb0Z7^OcHd0^7\\O`He0a7_OYHb0h7BQH`0P8DjG=_7gM[FQ2P29e7jMXFQ2m16k7mMUFQ2j13Q8PNRFQ2f11X8QNPFR2b1N_8RNlEU2_1Je8RNlEW2Z1Ii8QNlE[2T1FP9PNkE\\2R1DS9RNjE\\2o0DW9QNiE]2k0D\\9QNhE]2f0Ea9QNhE]2`0Ef9_1UFdNj9_1RFbNP:_1kEfNS:i4M3N2M3M3N1N2N2O0O2O1O001O1O001O1O001O1O001O1O001O1O00100O00100O001O10O01O01\\NUGlIj8o5_GmIa8P6bGQJ]8m5fGSJZ8f5lGZJT8R5QGoIT1o0k7P5SGnIT1R1j7n4UGkIU1W1f7l4dHTK\\7k4fHUKZ7j4fHVK[7h4gHWKZ7g4hHXKY7g4gHZKY7d4iH[KX7c4iH]KX7]OlF]4n1ULW7UOTGe4e1WLW7kN\\Gc4B_Km1R1l7Y3cHgL^7V3eHiL\\7T3fHlL[7Q3hHoLX7o2jHPMW7m2kHSMW7j2kHUMV7j2jHWMW7f2kHYMV7f1VFhMe2b0W7b1\\FdM`2i0U7CfEV1m0WNZ2Q1T7oNXFe1g3[Oi6a0YI_Oi6<[IDf68\\IHf63^IMc6O_I1d6J_I6b6EaI;V<0O1O10O0100O100O00100O10O010O0100O001k_OPOS?U1e@mN[?X1\\@mNd?j101O01N101O00001N101O00001N1O2K4M3L5L3L5M2M3M4M2M4L3N2M4L3N3L3L5Kmdk2"}, "label": "a baseball player in a # 9 white uniform with a black cap swinging a bat"}]}
{"id": 154713, "image": "COCO_train2014_000000154713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hotdog on the left side of the plate\"."}], "task": "refering", "answer_template": "The \"the hotdog on the left side of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 88, 89, 90, 91, 92, 93, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 207, 208, 209, 210], "bbox": [0.11920833333333333, 0.197796875, 0.48314583333333333, 0.55309375], "iscrowd": 0, "area": 26896.46285, "rle": {"size": [640, 480], "counts": "clS12kc06K5J7J5K6I6K5J7J5J7J5J6K3L3N3M3L4M3L4M3L4M3L4M3M3L4M3L4M3L4M2M4M3L4M3M3L4M3L4M2M4M2M3O1O2O0O1O2N100O2N1O100O2N1O101N1O1O2O0O1O1O2O0O1O2O00000O2O00001N10001O0O01O1O1O001O1O1O001O1O16J3M0001O00001O0000001O00001O00001O00001OO1O100O1O100O0001O0O101O001N3N2N2M3M3N2M3M3M2N1O100O1O1O001O1O100O1O1O001K5J6I7I7I7J6I7I8L4K5K5K5L4K6J5K5L4K5K5K8H8I7H8H8H8I7H`kj4"}, "label": "the hotdog on the left side of the plate"}]}
{"id": 154713, "image": "COCO_train2014_000000154713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hot dog on the left\"."}], "task": "refering", "answer_template": "The \"the hot dog on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 88, 89, 90, 91, 92, 93, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 207, 208, 209, 210], "bbox": [0.11920833333333333, 0.197796875, 0.48314583333333333, 0.55309375], "iscrowd": 0, "area": 26896.46285, "rle": {"size": [640, 480], "counts": "clS12kc06K5J7J5K6I6K5J7J5J7J5J6K3L3N3M3L4M3L4M3L4M3L4M3M3L4M3L4M3L4M2M4M3L4M3M3L4M3L4M2M4M2M3O1O2O0O1O2N100O2N1O100O2N1O101N1O1O2O0O1O1O2O0O1O2O00000O2O00001N10001O0O01O1O1O001O1O1O001O1O16J3M0001O00001O0000001O00001O00001O00001OO1O100O1O100O0001O0O101O001N3N2N2M3M3N2M3M3M2N1O100O1O1O001O1O100O1O1O001K5J6I7I7I7J6I7I8L4K5K5K5L4K6J5K5L4K5K5K8H8I7H8H8H8I7H`kj4"}, "label": "the hot dog on the left"}]}
{"id": 203867, "image": "COCO_train2014_000000203867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"straps\"."}], "task": "refering", "answer_template": "The \"straps\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 230, 231, 322, 345, 368], "bbox": [0.0003125, 0.39477083333333335, 0.20414062499999996, 1.0], "iscrowd": 0, "area": 6691.96395, "rle": {"size": [480, 640], "counts": "^96U>e0lDWO_7Z1cH[Of6e0[I1n50SJe0V5[OlJk0l4UOVKj0i4WOXKh0g4XO[Kg0d4YO^Kf0a4[O`Kd0`4[ObKd0]4\\OfKa0Z4@hK>W4BlK<S4DQL9n3HTL6k3JXL4g3L\\L2c3Of501O01O00001O00001O00010O00001O00001O000001O01O000000000010O00000010O0001O0000010O0000001O01O01O00000010O0001O0001O01O00001O01O0001O1O1O100O001O1O100O1O1O10O01O100O100O100O010O100O100O01O0012M3M2O2M3N1N3N2M3N1N3N2M3N1N3NYk^7"}, "label": "straps"}]}
{"id": 203867, "image": "COCO_train2014_000000203867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the straps of a white and black backoack\"."}], "task": "refering", "answer_template": "The \"the straps of a white and black backoack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 230, 231, 322, 345, 368], "bbox": [0.0003125, 0.39477083333333335, 0.20414062499999996, 1.0], "iscrowd": 0, "area": 6691.96395, "rle": {"size": [480, 640], "counts": "^96U>e0lDWO_7Z1cH[Of6e0[I1n50SJe0V5[OlJk0l4UOVKj0i4WOXKh0g4XO[Kg0d4YO^Kf0a4[O`Kd0`4[ObKd0]4\\OfKa0Z4@hK>W4BlK<S4DQL9n3HTL6k3JXL4g3L\\L2c3Of501O01O00001O00001O00010O00001O00001O000001O01O000000000010O00000010O0001O0000010O0000001O01O01O00000010O0001O0001O01O00001O01O0001O1O1O100O001O1O100O1O1O10O01O100O100O100O010O100O100O01O0012M3M2O2M3N1N3N2M3N1N3N2M3N1N3NYk^7"}, "label": "the straps of a white and black backoack"}]}
{"id": 31838, "image": "COCO_train2014_000000031838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog see the mirror\"."}], "task": "refering", "answer_template": "The \"dog see the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 144, 145, 146, 161, 162, 163, 177, 178, 179, 180, 194, 195, 196, 197, 211, 212, 213, 214, 228, 229, 230, 231, 246, 247, 264, 265], "bbox": [0.41787500000000005, 0.3318125, 0.6186458333333335, 0.670359375], "iscrowd": 0, "area": 13435.760549999999, "rle": {"size": [640, 480], "counts": "Rom36dc06J7TOFk]O`0na0l0K6M2O1O1O2N1O1O2N1O1O2M2O2N1O1O2L3N3M3M3M3YOh0[O]Ld@j3V?\\Lc@j3X?c0L4M3L4N3L4L4M3L3M101N1O10O0100O10000000O1000000O1000000O10PB[Je=e5ZB\\Jf=c5ZB^Jf=b5XB`Jh=`5WBaJh=_5WBcJi=]5UBeJk=[5SBgJm=f5000000aBQJm<b6O0O2O010O001O00000000cNaCYK_<b4hC\\KY<]4oCaKQ<Y4XDcKi;X4^DeKd;l1jBKl1UN];m1mBBQ2]NS;o1RCXOe?f0`@nNh?P1f1M3^Oa0I8G9E]Ub3"}, "label": "dog see the mirror"}]}
{"id": 31838, "image": "COCO_train2014_000000031838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog ' s reflection in the mirror\"."}], "task": "refering", "answer_template": "The \"a dog ' s reflection in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 144, 145, 146, 161, 162, 163, 177, 178, 179, 180, 194, 195, 196, 197, 211, 212, 213, 214, 228, 229, 230, 231, 246, 247, 264, 265], "bbox": [0.41787500000000005, 0.3318125, 0.6186458333333335, 0.670359375], "iscrowd": 0, "area": 13435.760549999999, "rle": {"size": [640, 480], "counts": "Rom36dc06J7TOFk]O`0na0l0K6M2O1O1O2N1O1O2N1O1O2M2O2N1O1O2L3N3M3M3M3YOh0[O]Ld@j3V?\\Lc@j3X?c0L4M3L4N3L4L4M3L3M101N1O10O0100O10000000O1000000O1000000O10PB[Je=e5ZB\\Jf=c5ZB^Jf=b5XB`Jh=`5WBaJh=_5WBcJi=]5UBeJk=[5SBgJm=f5000000aBQJm<b6O0O2O010O001O00000000cNaCYK_<b4hC\\KY<]4oCaKQ<Y4XDcKi;X4^DeKd;l1jBKl1UN];m1mBBQ2]NS;o1RCXOe?f0`@nNh?P1f1M3^Oa0I8G9E]Ub3"}, "label": "a dog ' s reflection in the mirror"}]}
{"id": 31838, "image": "COCO_train2014_000000031838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog that is not the reflection\"."}], "task": "refering", "answer_template": "The \"the dog that is not the reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.5375156249999999, 0.5206041666666666, 0.9996875], "iscrowd": 0, "area": 60867.76649999999, "rle": {"size": [640, 480], "counts": "c=b0d0d4n<hKfBZ4X=gKhBZ4U=hKjBZ4T=gKkB[4R=hKlBZ4R=gKmB[4Q=fKnB\\4o<fKPC\\4n<\\1M3N2M3N2N2M3N2M3N2M3N2N2M3N2N2O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O10000O1000000O11O000000000000001O00000000O1O100O100O10000O100O100O100O10000O100O100O100O100O100000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000O10000000000000000O10000000000001O0000001O0000001O00001O001O001O00001O001O001O00001O001O001SO\\E\\Hd:`7bE^H^:^7gEaHZ:Z7lEdHT:Y7PFfHQ:W7SFgHm9W7VFhHk9T7ZFjHf9T7d1N3L3N2N3L3N2N3H7_Oa0G9N3M2O1N3M2O1N2N3M2N2O2M2N2N3^Nf@ZN\\?g1g@SN[?n1i@lMX?T2\\1O000O2O0O101N3M3M3M3M2N3M3M3M3M3M4L5J6I8I6I7J6I7J6I7JiY_4"}, "label": "the dog that is not the reflection"}]}
{"id": 31838, "image": "COCO_train2014_000000031838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the dog ' s head who is staring into the mirror\"."}], "task": "refering", "answer_template": "The \"the back of the dog ' s head who is staring into the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.5375156249999999, 0.5206041666666666, 0.9996875], "iscrowd": 0, "area": 60867.76649999999, "rle": {"size": [640, 480], "counts": "c=b0d0d4n<hKfBZ4X=gKhBZ4U=hKjBZ4T=gKkB[4R=hKlBZ4R=gKmB[4Q=fKnB\\4o<fKPC\\4n<\\1M3N2M3N2N2M3N2M3N2M3N2N2M3N2N2O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O10000O1000000O11O000000000000001O00000000O1O100O100O10000O100O100O100O10000O100O100O100O100O100000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000O10000000000000000O10000000000001O0000001O0000001O00001O001O001O00001O001O001O00001O001O001SO\\E\\Hd:`7bE^H^:^7gEaHZ:Z7lEdHT:Y7PFfHQ:W7SFgHm9W7VFhHk9T7ZFjHf9T7d1N3L3N2N3L3N2N3H7_Oa0G9N3M2O1N3M2O1N2N3M2N2O2M2N2N3^Nf@ZN\\?g1g@SN[?n1i@lMX?T2\\1O000O2O0O101N3M3M3M3M2N3M3M3M3M3M4L5J6I8I6I7J6I7J6I7JiY_4"}, "label": "the back of the dog ' s head who is staring into the mirror"}]}
{"id": 351328, "image": "COCO_train2014_000000351328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand holding the joystick closest to the camera\"."}], "task": "refering", "answer_template": "The \"the hand holding the joystick closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 225, 226, 227, 228, 229, 241, 242, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.49176562500000004, 0.44269320843091337, 1.0, 1.0], "iscrowd": 0, "area": 37799.50579999999, "rle": {"size": [427, 640], "counts": "XeS4<l<c0^O4K4M3L4M4K4M3L4N2O10000O010000O01000O01000O01000O010000O01000O01000O010000O012N0O2O1O0O2OO1O1O001O1O010O1O0000000000000000000000001O00000000000001O0000000001O000000000000000000000000001O0001O00000000000000000001O00000000000cLRN]Kn1b4VN[Kj1e4ZNWKf1h4_NTKa1l4`NTK_1k4bNUK^1k4cNTK]1k4dNUK]1j4cNVK]1i4dNWK\\1i4dNWK\\1h4eNXK[1h4eNXK[1g4gNXKY1h4gNXKY1g4hNYKX1g4hNYKX1f4iNZKW1f4iNZKW1f4iNZKW1f4jNYKW1f4iNZKW1e4jN[KV1e4jN[KV1e4jN[KV1e4jN[KV1e4jN\\KU1d4lN[KT1d4mN\\KS1d4mN\\KS1d4mN\\KS1d4mN\\KS1d4mN\\KS1d4mN\\KT1b4nN]KR1c4nN]KR1c4nN]KS1b4mN^KS1b4nN]KR1c4oN\\KR1c4nN]KR1b4PO]KP1c4QO\\KP1c4QO\\Ko0d4QO\\Ko0d4RO[Ko0d4RO[Kn0d4TO\\Kk0d4UO\\Kl0c4UO\\Kk0c4VO]Kj0c4VO]Kk0a4VO_Kj0a4VO_Kj0`4WO`Ki0`4WO`Ki0_4XOaKh0_4XOaKi0]4XOcKh0]4XOcKh0\\4YOdKg0\\4YOdKg0[4ZOeKf0[4ZOeKf0Z4[OfKe0Z4[OfKe0Z4[OfKe0Y4\\OgKe0X4[OhKe0W4\\OjKc0V4]OjKc0U4^OkKb0U4^OkKb0T4_OlKa0T4_OlKa0S4@mK`0S4@mKa0Q4@oK`0Q4AnK?Q4BoK>Q4BoK>P4CoK>Q4CnK=Q4EnK;R4FlK;T4FkK;T4FkK:U4GiK:W4GhK9X4HgK8Y4IeK8[4IdK8[4IdK7\\4JbK7^4JaK7^4J`K7`4J_K6a4K^K6b4J\\K7e4IZK7f4JYK6h4JVK8j4GVK9k4GTK9m4GQK;o4EPK;Q5EnJ;S5EkJ=U5ChJ?Y5AdJa0]5_O`Jc0a5]O[Jg0e5YOXJi0i5WOTJk0m5UOPJn0P6ROmIP1S6QOiIR1X6nNeIU1[6kNbIW1_6iN^IY1c6gNZI[1g6eNVI^1j6bNRIa1o6]2100O1O100O1O1O10O01O100O1O1O100O1O100O1O10O01O00100O00100O0010O01O1O010O1O001O001O1O001O1O001O1O0O2O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O0O2O1O001O1O1O2N2N1O2N2N1OWG"}, "label": "the hand holding the joystick closest to the camera"}]}
{"id": 351328, "image": "COCO_train2014_000000351328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurred man in a grey shirt behind people playing wii\"."}], "task": "refering", "answer_template": "The \"a blurred man in a grey shirt behind people playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 16, 17, 18, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 39, 40, 41, 42, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 234, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.0, 0.0022482435597189696, 0.8590937500000001, 0.9887587822014051], "iscrowd": 0, "area": 98573.63305000003, "rle": {"size": [427, 640], "counts": "g3[9o31O2N1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O2N1O1O1O1O1N2O1O1O1OnKcMXM\\2e2lMWMR2f2WNVMh1f2aNVM^1g2jNUMU1g2SOVMl0h2[OTMd0m2@oL>S3FiL9X3JeL5[3LeL3\\3NcL1^3ObL0_31`LO`31aLM`34_LKb35^LKa36_LIb38]LHc38]LGd3:[LFe3:[LEf3;ZLEe3=ZLBg3>YLBg3?XL@i3`0WL_Oj3a0VL_Oj3b0UL]Ol3c0TL]Ok3e0TLZOm3f0SLZOm3f0SLYOn3h0QLXOo3h0QLWOP4j0oKVOR4i0nKVOS4j0mKUOT4l0lKSOU4l0kKSOV4n0iKROW4n0iKQOY4o0fKQOZ4o0fKPO[4Q1dKoN]4P1cKoN^4R1aKnN_4R1aKmNa4S1^KlNc4T1]KlNc4T1]KkNe4U1ZKkNf4U1ZKjNg4W1XKiNi4V1WKiNj4X1SKjNm4V1QKkNQ5U1mJlNS5T1kJmNV5T1hJlNZ5S1dJoN\\5R1bJnN_5R1_JPOb5P1[JQOf5o0YJROh5m0VJTOk5m0SJTOn5l0oIUOR6l0lIUOT6k0jIWOW6i0fIYOZ6h0dIYO]6g0`I[O`6f0^I[Ob6f0[I\\Of6d0XI]Oh6c0VI_Ok6a0RIBm6?QIBP7>mHDS7=kHDU7=hHEY7:fHGZ7:cHH^78_HJa77]HJc7o210O01O001O00001O001O012M2N2NYOfJ`IW5^6TK[Ij4c6aKWI\\4h6nKQIP4m6X1N3M2N2O1N2N2N2O100O101N100O100O100O100O2O0O100O100O100O101N100O101N1N2N3L3N2N2N3O0O10001O000O2O0000001N10001O0O100O2O0OkNmJTJS5k5QKSJm4m5VKQJj4o5YKnIf4R6^KkIb4T6aKkI]4U6gKhIY4X6jKeIV4Z6mKeIR4Z6SLbIm3]6XL_Ih3a6\\L[Id3e6`LXI_3h6b1000000000000000000000000000000000000000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000001O00O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1hL]Gg1d8iMmGU2T8[M]Hb2X9N2N2N2N3M2N2N1O1O1O1N2O1O1O1O1O2N1O1O1O1000000001O0000001O0000001O0000001O0000001O000000001O00000000O100O100O100O100O100O100O100O10000O100OUNTNYHl1g7WNVHh1k7ZNSHf1m7]NPHb1Q8`NmG`1S8bNkG]1V8fNgGZ1Y8hNeGW1\\8lNaGT1_8nN_GQ1a8SO\\Gm0d8UOZGj0g8XOWGh0i8[OTGd0m8^OQGa0P9BmF=T9EjF;V9HgF7Z9KdF4]9NaF1`92]FNc94[FKf98WFGj9;TFEl9l11O1O100O1O1O100O1O100O1O1O100O1O1O100O1O10O01O10000O100O10000O100O100O10000O100O2O1N2O1O2M2O1N2O1O1N2O2M2O1N2O1O1N3N1N2O1O1N2O1N3N1O1N2O1N2O2N1N2O1N2O1O1N3N1N2O1N2O1O=BP_U1"}, "label": "a blurred man in a grey shirt behind people playing wii"}]}
{"id": 351328, "image": "COCO_train2014_000000351328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing wii while supporting his head with his other hand\"."}], "task": "refering", "answer_template": "The \"a man playing wii while supporting his head with his other hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 16, 17, 18, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 39, 40, 41, 42, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 234, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.0, 0.0022482435597189696, 0.8590937500000001, 0.9887587822014051], "iscrowd": 0, "area": 98573.63305000003, "rle": {"size": [427, 640], "counts": "g3[9o31O2N1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O2N1O1O1O1O1N2O1O1O1OnKcMXM\\2e2lMWMR2f2WNVMh1f2aNVM^1g2jNUMU1g2SOVMl0h2[OTMd0m2@oL>S3FiL9X3JeL5[3LeL3\\3NcL1^3ObL0_31`LO`31aLM`34_LKb35^LKa36_LIb38]LHc38]LGd3:[LFe3:[LEf3;ZLEe3=ZLBg3>YLBg3?XL@i3`0WL_Oj3a0VL_Oj3b0UL]Ol3c0TL]Ok3e0TLZOm3f0SLZOm3f0SLYOn3h0QLXOo3h0QLWOP4j0oKVOR4i0nKVOS4j0mKUOT4l0lKSOU4l0kKSOV4n0iKROW4n0iKQOY4o0fKQOZ4o0fKPO[4Q1dKoN]4P1cKoN^4R1aKnN_4R1aKmNa4S1^KlNc4T1]KlNc4T1]KkNe4U1ZKkNf4U1ZKjNg4W1XKiNi4V1WKiNj4X1SKjNm4V1QKkNQ5U1mJlNS5T1kJmNV5T1hJlNZ5S1dJoN\\5R1bJnN_5R1_JPOb5P1[JQOf5o0YJROh5m0VJTOk5m0SJTOn5l0oIUOR6l0lIUOT6k0jIWOW6i0fIYOZ6h0dIYO]6g0`I[O`6f0^I[Ob6f0[I\\Of6d0XI]Oh6c0VI_Ok6a0RIBm6?QIBP7>mHDS7=kHDU7=hHEY7:fHGZ7:cHH^78_HJa77]HJc7o210O01O001O00001O001O012M2N2NYOfJ`IW5^6TK[Ij4c6aKWI\\4h6nKQIP4m6X1N3M2N2O1N2N2N2O100O101N100O100O100O100O2O0O100O100O100O101N100O101N1N2N3L3N2N2N3O0O10001O000O2O0000001N10001O0O100O2O0OkNmJTJS5k5QKSJm4m5VKQJj4o5YKnIf4R6^KkIb4T6aKkI]4U6gKhIY4X6jKeIV4Z6mKeIR4Z6SLbIm3]6XL_Ih3a6\\L[Id3e6`LXI_3h6b1000000000000000000000000000000000000000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000001O00O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1hL]Gg1d8iMmGU2T8[M]Hb2X9N2N2N2N3M2N2N1O1O1O1N2O1O1O1O1O2N1O1O1O1000000001O0000001O0000001O0000001O0000001O000000001O00000000O100O100O100O100O100O100O100O10000O100OUNTNYHl1g7WNVHh1k7ZNSHf1m7]NPHb1Q8`NmG`1S8bNkG]1V8fNgGZ1Y8hNeGW1\\8lNaGT1_8nN_GQ1a8SO\\Gm0d8UOZGj0g8XOWGh0i8[OTGd0m8^OQGa0P9BmF=T9EjF;V9HgF7Z9KdF4]9NaF1`92]FNc94[FKf98WFGj9;TFEl9l11O1O100O1O1O100O1O100O1O1O100O1O1O100O1O10O01O10000O100O10000O100O100O10000O100O2O1N2O1O2M2O1N2O1O1N2O2M2O1N2O1O1N3N1N2O1O1N2O1N3N1O1N2O1N2O2N1N2O1N2O1O1N3N1N2O1N2O1O=BP_U1"}, "label": "a man playing wii while supporting his head with his other hand"}]}
{"id": 351328, "image": "COCO_train2014_000000351328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blurry face of the person to the right behind the controller\"."}], "task": "refering", "answer_template": "The \"the blurry face of the person to the right behind the controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182], "bbox": [0.8590937500000001, 0.0, 1.0, 0.5393208430913349], "iscrowd": 0, "area": 16381.483649999998, "rle": {"size": [427, 640], "counts": "[\\U7?X11e9d0\\E2Y:i1C<G9G:G8iFfLo7`4J6M3L4M4L3MAfHPKW7P5iHSKT7l4nHWKm6i4UIYKh6g4YI[Kd6d4]I_K`6a4aI`K]6_4eI`K[6`4eI`KZ6`4hI_KX6a4iI]KX6b4jI]KV6b4kI^KU6b4lI]KT6c4mI\\KS6d4nI[KR6e4nI[KR6e4oIYKR6g4oIXKQ6h4PJWKP6i4PJWKP6i4QJVKo5j4RJUKn5k4SJTKm5l4SJSKn5m4SJRKm5n4TJQKl5o4UJPKk5P5UJQKj5o4WJQKh5o4YJRKe5n4\\JRKc5n4^JSK`5m4`JUK^5k4cJUK\\5k4eJVKY5j4hJVKW5j4jJWKT5i4lJXKS5h4nJYKP5g4QKXKo4h4RKWKn4i4RKWKn4i4SKVKm4j4TKTKm4l4VKQKj4o4XKoJh4Q5[KlJe4T5^KiJb4W5`KgJ`4Y5i100000000002N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N"}, "label": "the blurry face of the person to the right behind the controller"}]}
{"id": 351328, "image": "COCO_train2014_000000351328.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"face on right cutoff\"."}], "task": "refering", "answer_template": "The \"face on right cutoff\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182], "bbox": [0.8590937500000001, 0.0, 1.0, 0.5393208430913349], "iscrowd": 0, "area": 16381.483649999998, "rle": {"size": [427, 640], "counts": "[\\U7?X11e9d0\\E2Y:i1C<G9G:G8iFfLo7`4J6M3L4M4L3MAfHPKW7P5iHSKT7l4nHWKm6i4UIYKh6g4YI[Kd6d4]I_K`6a4aI`K]6_4eI`K[6`4eI`KZ6`4hI_KX6a4iI]KX6b4jI]KV6b4kI^KU6b4lI]KT6c4mI\\KS6d4nI[KR6e4nI[KR6e4oIYKR6g4oIXKQ6h4PJWKP6i4PJWKP6i4QJVKo5j4RJUKn5k4SJTKm5l4SJSKn5m4SJRKm5n4TJQKl5o4UJPKk5P5UJQKj5o4WJQKh5o4YJRKe5n4\\JRKc5n4^JSK`5m4`JUK^5k4cJUK\\5k4eJVKY5j4hJVKW5j4jJWKT5i4lJXKS5h4nJYKP5g4QKXKo4h4RKWKn4i4RKWKn4i4SKVKm4j4TKTKm4l4VKQKj4o4XKoJh4Q5[KlJe4T5^KiJb4W5`KgJ`4Y5i100000000002N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N"}, "label": "face on right cutoff"}]}
{"id": 580706, "image": "COCO_train2014_000000580706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a calico cat sitting on a purple plaided blacket with a persons hand touching it\"."}], "task": "refering", "answer_template": "The \"a calico cat sitting on a purple plaided blacket with a persons hand touching it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 289, 308, 309, 310, 332, 333, 355, 356], "bbox": [0.38870312500000004, 0.2899581589958159, 0.7019062500000001, 0.9243723849372385], "iscrowd": 0, "area": 30821.7335, "rle": {"size": [478, 640], "counts": "n_d3:b>?A>B9G5K6J5K6I5L3L4L4M3L4M3L3`D]Mc:f2WEfMa:b3I5J3N3M2N2N2N2N2O1N2O1O1N2O1O1O1N2O1iFdJj8]5SGiJh8Y5TGmJi8g5M2N0O2O1O0O2O1O0O2O1O1O2M2O1O2N1O1O1ON21N3N2N0000001O00O100O1O100O1O2N2PM]HbNd7[1gH\\N[7b1iH\\L@X1m7Z2]IiL`NHY8^3[IeLU7]3mH]LV7c3R20O2O001N10001N10000O01000O2O1O2M2O2N1N2O2N1N3N1O1N3N1N3L4M3M3L4L5K4L4L4M4M2N2M4M2N2N2N3O00000O10N2N2O2M6J00OO2N2N1O2N2N101N2N2O2M3J5H9H7I8L4M2N3N1N3MhFYL^7g3]HcL_7]3_HfL`7Z3]HhLc7X3\\HiLe7W3XHkLj7T3SHlLS8Q3kGoLZ8n2cGRMd8j2ZGUMk8h2TGWMR9f2lFWM\\9f2bFWMe9g2YF[Mh9d2VF]Mk9c2SF^Mo9`2PFbMQ:]2mEdMT:\\2jEfMW:Y2gEiMY:W2eEmMZ:R2dERN[:l1dEXNZ:h1dE\\N[:c1cEaNOhNS:c40000O2O0OoKQFT3n9jLUFT3l9jLWFS3k9kLXFQ3l9lLWFQ3k9mLXFo2k9oLXFj2n9TMUFb2T:ZMRFY2Y:aMnEQ2g;E;`NoBi0k=A?I7Igmh2"}, "label": "a calico cat sitting on a purple plaided blacket with a persons hand touching it"}]}
{"id": 580706, "image": "COCO_train2014_000000580706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat staying on top of the purple and green blanket and touched by a hand\"."}], "task": "refering", "answer_template": "The \"a cat staying on top of the purple and green blanket and touched by a hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 289, 308, 309, 310, 332, 333, 355, 356], "bbox": [0.38870312500000004, 0.2899581589958159, 0.7019062500000001, 0.9243723849372385], "iscrowd": 0, "area": 30821.7335, "rle": {"size": [478, 640], "counts": "n_d3:b>?A>B9G5K6J5K6I5L3L4L4M3L4M3L3`D]Mc:f2WEfMa:b3I5J3N3M2N2N2N2N2O1N2O1O1N2O1O1O1N2O1iFdJj8]5SGiJh8Y5TGmJi8g5M2N0O2O1O0O2O1O0O2O1O1O2M2O1O2N1O1O1ON21N3N2N0000001O00O100O1O100O1O2N2PM]HbNd7[1gH\\N[7b1iH\\L@X1m7Z2]IiL`NHY8^3[IeLU7]3mH]LV7c3R20O2O001N10001N10000O01000O2O1O2M2O2N1N2O2N1N3N1O1N3N1N3L4M3M3L4L5K4L4L4M4M2N2M4M2N2N2N3O00000O10N2N2O2M6J00OO2N2N1O2N2N101N2N2O2M3J5H9H7I8L4M2N3N1N3MhFYL^7g3]HcL_7]3_HfL`7Z3]HhLc7X3\\HiLe7W3XHkLj7T3SHlLS8Q3kGoLZ8n2cGRMd8j2ZGUMk8h2TGWMR9f2lFWM\\9f2bFWMe9g2YF[Mh9d2VF]Mk9c2SF^Mo9`2PFbMQ:]2mEdMT:\\2jEfMW:Y2gEiMY:W2eEmMZ:R2dERN[:l1dEXNZ:h1dE\\N[:c1cEaNOhNS:c40000O2O0OoKQFT3n9jLUFT3l9jLWFS3k9kLXFQ3l9lLWFQ3k9mLXFo2k9oLXFj2n9TMUFb2T:ZMRFY2Y:aMnEQ2g;E;`NoBi0k=A?I7Igmh2"}, "label": "a cat staying on top of the purple and green blanket and touched by a hand"}]}
{"id": 310376, "image": "COCO_train2014_000000310376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of seasoning with a white lid to the left of a blender\"."}], "task": "refering", "answer_template": "The \"a bottle of seasoning with a white lid to the left of a blender\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 307, 308, 309], "bbox": [0.0, 0.5664374999999999, 0.23722916666666669, 0.8141874999999998], "iscrowd": 0, "area": 12344.153350000002, "rle": {"size": [640, 480], "counts": "T<3lc05K5Y]OH]a0>[^OGaa0?Y^OEda0a0U^OCha0b0R^OBka0\\1M4L3M3M3M3M4L3M3M3M3N2M4M2N2N2M3N3M2N2M3N2N3M2M3N2N2N2M4M2N2N2N1N2O00000O10001O0001O001O00100O001O001O010O1O001O002O0O2N2N1O2N2N2O0O2N2N1N3N2M2O2M3N1N3N2M2O2M3N2N1N3N2M2O2M3N1N3N2M2O2M3N1N3N2M3N1M4K7I8G9H8H8G8I8HWYT7"}, "label": "a bottle of seasoning with a white lid to the left of a blender"}]}
{"id": 7277, "image": "COCO_train2014_000000007277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white colored poka dot shirt next a man\"."}], "task": "refering", "answer_template": "The \"a woman in a white colored poka dot shirt next a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 86, 87, 88, 89, 99, 100, 101, 102, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 137, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 155, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 217, 218, 219, 220, 230, 231, 232, 233], "bbox": [0.55632, 0.30832, 0.9997333333333333, 1.0], "iscrowd": 0, "area": 38057.33714999999, "rle": {"size": [500, 375], "counts": "k\\V38T?9N2M3K7I7I7H8I7I6K5J5L5L4K4M4K4L5L4K4M4K4L5L4L3M3M4K4M4L3M3M4K4F:N3M2N3M3M2N3M3M3M2N3M3YNPJfIR6T6XJeIk5V6\\JeIg5V6^JhIc5S6cJjI`5R6dJlI_5P6dJnI^5n5eJbIjNNf6[6bJdIiN2j6S6bJeIgN7m6n5hJRJ]5g5fJXJ]5d5cJ\\J`5`5aJaJb5\\5`JaJe5Z5^JcJh5X5YJgJl5S5]2M2O0O10^ORK^Fm4d9ZKTFf4n9_KkE`4X:<3N2N2N2N1N3NATKXFj4i9XKTFh4n9YKoEg4R:[KkEe4V:]KgEc4[::1O1N3N1O2O0O1O000O2O0O1001O00001O00001O1O1O1O1O1O1O1O1O1O1O2N4L4VGfJR7^5cHiJ]7[5XHlJh7X5cGYK]8P6O001O1O1O1O1O1OO10000000000O1001O1O1O1O001O1O0dKaHc0`7PMaHe0;Y2U7RMbHMo0P3`6SMbHCX1X3W6TMdHYO^1b3o5UMdHROd1g3i5VMeHlNh1m3d5WMmJmNbM^3c7cMlJjNhM`3]7fMZKX2g4hMZKV2g4iM[KU2f4kM]KQ2c4oMaKm1`4SNbKj1_4VNdKf1\\4ZNgKc1Z4]NiK_1X4`NQM7o2JjM\\OW2d0\\6101O1NVOfH"}, "label": "a woman in a white colored poka dot shirt next a man"}]}
{"id": 7277, "image": "COCO_train2014_000000007277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"european young girl with long straight hair - bangs\"."}], "task": "refering", "answer_template": "The \"european young girl with long straight hair - bangs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 86, 87, 88, 89, 99, 100, 101, 102, 112, 113, 114, 115, 116, 125, 126, 127, 128, 129, 137, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 155, 163, 164, 165, 166, 167, 168, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 217, 218, 219, 220, 230, 231, 232, 233], "bbox": [0.55632, 0.30832, 0.9997333333333333, 1.0], "iscrowd": 0, "area": 38057.33714999999, "rle": {"size": [500, 375], "counts": "k\\V38T?9N2M3K7I7I7H8I7I6K5J5L5L4K4M4K4L5L4K4M4K4L5L4L3M3M4K4M4L3M3M4K4F:N3M2N3M3M2N3M3M3M2N3M3YNPJfIR6T6XJeIk5V6\\JeIg5V6^JhIc5S6cJjI`5R6dJlI_5P6dJnI^5n5eJbIjNNf6[6bJdIiN2j6S6bJeIgN7m6n5hJRJ]5g5fJXJ]5d5cJ\\J`5`5aJaJb5\\5`JaJe5Z5^JcJh5X5YJgJl5S5]2M2O0O10^ORK^Fm4d9ZKTFf4n9_KkE`4X:<3N2N2N2N1N3NATKXFj4i9XKTFh4n9YKoEg4R:[KkEe4V:]KgEc4[::1O1N3N1O2O0O1O000O2O0O1001O00001O00001O1O1O1O1O1O1O1O1O1O1O2N4L4VGfJR7^5cHiJ]7[5XHlJh7X5cGYK]8P6O001O1O1O1O1O1OO10000000000O1001O1O1O1O001O1O0dKaHc0`7PMaHe0;Y2U7RMbHMo0P3`6SMbHCX1X3W6TMdHYO^1b3o5UMdHROd1g3i5VMeHlNh1m3d5WMmJmNbM^3c7cMlJjNhM`3]7fMZKX2g4hMZKV2g4iM[KU2f4kM]KQ2c4oMaKm1`4SNbKj1_4VNdKf1\\4ZNgKc1Z4]NiK_1X4`NQM7o2JjM\\OW2d0\\6101O1NVOfH"}, "label": "european young girl with long straight hair - bangs"}]}
{"id": 7277, "image": "COCO_train2014_000000007277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing white colour shirt\"."}], "task": "refering", "answer_template": "The \"a person wearing white colour shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 42, 43, 44, 55, 56, 57, 68, 69, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227], "bbox": [0.0132, 0.1232, 0.5845066666666666, 0.99882], "iscrowd": 0, "area": 58027.590899999996, "rle": {"size": [500, 375], "counts": "Qf22`?6K4K6K5L3M2M4M3M2N3L3N2N1O2N2N2N2N1O1O1O2N1O1O1N3J5L4L4L5J5L4lFXMW5m2cJ^MS5f2iJbMn4`2QKgMh4[2VKlMc4V2[KYNV4i1fKXOZ3k0XLDe3>PLNm33SLOi34VLNg34YLMg32XL0g31WL1i3NVL4i3MUL5k3KhKb0W4^OcKi0]4WO\\KP1c4POmJa1S5_NZJT2f5m3O100O100O10000O100O100O100O100O1O100O100O1O100O1O100O100O1O1O1hM`F[Na9W1RGdNn8Z1XGaNi8]1^G]Nc8a1dGYN]8e1jGRNZ8l1mGlMV8R2g2N200O100O10000O100O100000000000000000001O000001O000D<2N2N2O2N1O1N01O1O001O1O1jDlMj8U2TGnMj8S2SGRNj8P2RGUNk8l1RGYNk8h1SGZNl8g1QG\\Nn8e1oF]NQ9e1kF^NT9d1iF^NV9d1fF_NY9c1cF`N\\9b1`F`N`9c1\\F_Nc9f1VF]Ni9k1mEXNR:P2dESN[:c301O1O1O001O1O1O1O001O1O00001O0000001O00001O00001O0000001O00001O00001O000eK\\JfNd5S1fJjN[5n0nJPOR5j0VKTOj4j0ZKTOh4i0_KQOd4j0jKjNZ4Q1SLcNP4X1^LVM[NjN[5k3gLXLZ5b3`3L5J5L5K5J5L5J5L5J6K4K6K5J8I7H8I8G8I7H8I[l[2"}, "label": "a person wearing white colour shirt"}]}
{"id": 7277, "image": "COCO_train2014_000000007277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall white man in white striped shirt and black pants\"."}], "task": "refering", "answer_template": "The \"a tall white man in white striped shirt and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 42, 43, 44, 55, 56, 57, 68, 69, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 104, 105, 106, 107, 108, 109, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227], "bbox": [0.0132, 0.1232, 0.5845066666666666, 0.99882], "iscrowd": 0, "area": 58027.590899999996, "rle": {"size": [500, 375], "counts": "Qf22`?6K4K6K5L3M2M4M3M2N3L3N2N1O2N2N2N2N1O1O1O2N1O1O1N3J5L4L4L5J5L4lFXMW5m2cJ^MS5f2iJbMn4`2QKgMh4[2VKlMc4V2[KYNV4i1fKXOZ3k0XLDe3>PLNm33SLOi34VLNg34YLMg32XL0g31WL1i3NVL4i3MUL5k3KhKb0W4^OcKi0]4WO\\KP1c4POmJa1S5_NZJT2f5m3O100O100O10000O100O100O100O100O1O100O100O1O100O1O100O100O1O1O1hM`F[Na9W1RGdNn8Z1XGaNi8]1^G]Nc8a1dGYN]8e1jGRNZ8l1mGlMV8R2g2N200O100O10000O100O100000000000000000001O000001O000D<2N2N2O2N1O1N01O1O001O1O1jDlMj8U2TGnMj8S2SGRNj8P2RGUNk8l1RGYNk8h1SGZNl8g1QG\\Nn8e1oF]NQ9e1kF^NT9d1iF^NV9d1fF_NY9c1cF`N\\9b1`F`N`9c1\\F_Nc9f1VF]Ni9k1mEXNR:P2dESN[:c301O1O1O001O1O1O1O001O1O00001O0000001O00001O00001O0000001O00001O00001O000eK\\JfNd5S1fJjN[5n0nJPOR5j0VKTOj4j0ZKTOh4i0_KQOd4j0jKjNZ4Q1SLcNP4X1^LVM[NjN[5k3gLXLZ5b3`3L5J5L5K5J5L5J5L5J6K4K6K5J8I7H8I8G8I7H8I[l[2"}, "label": "a tall white man in white striped shirt and black pants"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut with sprinkles to the top right of the other donuts\"."}], "task": "refering", "answer_template": "The \"donut with sprinkles to the top right of the other donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177], "bbox": [0.463484375, 0.02695833333333333, 0.77865625, 0.4763958333333333], "iscrowd": 0, "area": 31337.679650000002, "rle": {"size": [480, 640], "counts": "WZ[4:c>:F;F:E;E8H5K6K4K5K6J5K6L3M3M4L3M3M4K4M4L3L4M2N1O2M3N2N1O2N2O0O2N2N2O0O2N2O1N1O2N2O0O2N2O1N1O2N2O0O2N2O0O1O2N100O2N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101O00001O001O00001O00001O000O2O00000000O100O100O1O100O101N2N101N2O1N101N2N2O1N101N2N2O0O2O1N2O0O2N2O1N101N1O2O1N101N101N2N101N2O0O2N101N2O0O2O0O2N2O0O2N3L3M3N2M3M3N3L3M3N2M3M3N3L3N2M3M3N3L3M3N2L4L6I6J6K5J7I6K5J6J7J5J6J6K5Jo^R2"}, "label": "donut with sprinkles to the top right of the other donuts"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a doughnut with sprinkles in the upper right hand corner\"."}], "task": "refering", "answer_template": "The \"a doughnut with sprinkles in the upper right hand corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177], "bbox": [0.463484375, 0.02695833333333333, 0.77865625, 0.4763958333333333], "iscrowd": 0, "area": 31337.679650000002, "rle": {"size": [480, 640], "counts": "WZ[4:c>:F;F:E;E8H5K6K4K5K6J5K6L3M3M4L3M3M4K4M4L3L4M2N1O2M3N2N1O2N2O0O2N2N2O0O2N2O1N1O2N2O0O2N2O1N1O2N2O0O2N2O0O1O2N100O2N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101N10001N100O2O0O101O00001O001O00001O00001O000O2O00000000O100O100O1O100O101N2N101N2O1N101N2N2O1N101N2N2O0O2O1N2O0O2N2O1N101N1O2O1N101N101N2N101N2O0O2N101N2O0O2O0O2N2O0O2N3L3M3N2M3M3N3L3M3N2M3M3N3L3N2M3M3N3L3M3N2L4L6I6J6K5J7I6K5J6J7J5J6J6K5Jo^R2"}, "label": "a doughnut with sprinkles in the upper right hand corner"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut bottom left\"."}], "task": "refering", "answer_template": "The \"donut bottom left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351], "bbox": [0.030343750000000003, 0.5146041666666666, 0.36742187499999995, 0.9213541666666667], "iscrowd": 0, "area": 32585.184650000003, "rle": {"size": [480, 640], "counts": "SW91f>=D=I6K6I5L5K5J5L5K5J5L5K3L4M4L3L4M4L3N2M4M2M4M2M3N3L3N2M4M1N3N1O2M2O1N3N1N3N1N3N1N3N1N2O1N2O1N101N2O1N2O1O0O2O000O100O10000O100O10000O10000O100O10000O100O2O00000O10000000000O100000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000000000O10O10000O10000O10000O01000O100O100O2N101N2O0O2O1N101N2O0O2O1N1O2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2M3M3M3M3M3M3L6K5K4L5K5K4K6K5K4L5J6K4L8H;D<EV`m5"}, "label": "donut bottom left"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut with the least sprinkles\"."}], "task": "refering", "answer_template": "The \"donut with the least sprinkles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351], "bbox": [0.030343750000000003, 0.5146041666666666, 0.36742187499999995, 0.9213541666666667], "iscrowd": 0, "area": 32585.184650000003, "rle": {"size": [480, 640], "counts": "SW91f>=D=I6K6I5L5K5J5L5K5J5L5K3L4M4L3L4M4L3N2M4M2M4M2M3N3L3N2M4M1N3N1O2M2O1N3N1N3N1N3N1N3N1N2O1N2O1N101N2O1N2O1O0O2O000O100O10000O100O10000O10000O100O10000O100O2O00000O10000000000O100000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000000000O10O10000O10000O10000O01000O100O100O2N101N2O0O2O1N101N2O0O2O1N1O2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2N2N2M3M3M3M3M3M3L6K5K4L5K5K4K6K5K4L5J6K4L8H;D<EV`m5"}, "label": "donut with the least sprinkles"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a frosted sprinkled cupcake , one out of four to the right and front of the others\"."}], "task": "refering", "answer_template": "The \"a frosted sprinkled cupcake , one out of four to the right and front of the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360, 361, 381, 382, 383], "bbox": [0.45337500000000003, 0.5213541666666667, 0.7837031250000001, 0.9685416666666666], "iscrowd": 0, "area": 34513.808249999995, "rle": {"size": [480, 640], "counts": "iYX43e>:G8G:F9G:G8J7J5K6I7J5K6K4N3M2N3M2N3M3N1N3M2N3M2N3M2N3N2M2N3M2N3M2N3M2N3N2M2N2N2N1O2N1O2O0O1O2N1O2N101N1O2N1O2O0O2N1O2N101N1O1O2N101N1O2O0O2O001O001O001O001O00001O0O2O001O001O001O000000000O1000000000000000000000000O10001O00000000000000000O100000000001N1O100O2N100O1O2O0O1O100O2N100O1O2O0O100O1O2O0O1O101N1O100O2N101N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M4L3M4L4L4L4L4L4L4L3M4L4L4L4M3L4L4K1Lg0XOg0ZO\\ZP2"}, "label": "a frosted sprinkled cupcake , one out of four to the right and front of the others"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"doughnut in top left corner\"."}], "task": "refering", "answer_template": "The \"doughnut in top left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 165, 166, 167], "bbox": [0.10304687500000001, 0.04616666666666667, 0.393578125, 0.4358125], "iscrowd": 0, "area": 26731.711299999995, "rle": {"size": [480, 640], "counts": "ZRo08[>>F;D;F;E:M3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4M2M3N3L3M4M2N2N3N1O1N3N001N10001O0O2O000O2O001N10001N101O0O101O001N10001N101O0O101O001N10001N101O000O2O000O1000000O10000O10000O101O000O10000O1000000O10000O100001O00001O00001O00001O00001O00001N100O2N100O2O0O101N1O101N100O2N100O101N100O2N100O2O0O2N3N1N3N1N2O2M2N3N1N3N1N3N1N3M2O2M2O2M3M3N2K5L4L4L4L4K5L4L4L4K6K4L4L4L4K5L4L4L4K5Lehe5"}, "label": "doughnut in top left corner"}]}
{"id": 244846, "image": "COCO_train2014_000000244846.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut with sprinkles in left corner\"."}], "task": "refering", "answer_template": "The \"donut with sprinkles in left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 165, 166, 167], "bbox": [0.10304687500000001, 0.04616666666666667, 0.393578125, 0.4358125], "iscrowd": 0, "area": 26731.711299999995, "rle": {"size": [480, 640], "counts": "ZRo08[>>F;D;F;E:M3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4M2M3N3L3M4M2N2N3N1O1N3N001N10001O0O2O000O2O001N10001N101O0O101O001N10001N101O0O101O001N10001N101O000O2O000O1000000O10000O10000O101O000O10000O1000000O10000O100001O00001O00001O00001O00001O00001N100O2N100O2O0O101N1O101N100O2N100O101N100O2N100O2O0O2N3N1N3N1N2O2M2N3N1N3N1N3N1N3M2O2M2O2M3M3N2K5L4L4L4L4K5L4L4L4K6K4L4L4L4K5L4L4L4K5Lehe5"}, "label": "donut with sprinkles in left corner"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black duffle bag with a face on it\"."}], "task": "refering", "answer_template": "The \"a black duffle bag with a face on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 93, 94, 95, 114, 115, 116, 117, 136, 137, 138, 139, 157, 158, 159, 160, 161, 181, 182, 183], "bbox": [0.15666666666666665, 0.16315359477124183, 0.3645098039215686, 0.41449346405228754], "iscrowd": 0, "area": 12020.032150000001, "rle": {"size": [612, 612], "counts": "ebi12Pc04L4L4M1N2N2N2N2N2O1N2N2N2N2N101N2N2N2N2O0O2N2O1N2N2O1N2N2O1N2O1N2O1N2N2O1N2N3N2M2N3N2M2N3N2M3M2N3M3N1N3M3M2N3N2O1N1O2N2O0O2N2O1N1O2O1N1O2O0O101O0O10001O000000001N10001O001O001N101O00001O0O2O001O0O101O0O2O00@mAPLS>n3PBPLR>m3QBQLP>m3TBPLn=m3f0N3M8I5K4K4M3L4M3M4K4M4K4M3L7I;E;E`0AaYX7"}, "label": "a black duffle bag with a face on it"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black backpack with white writing\"."}], "task": "refering", "answer_template": "The \"a black backpack with white writing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 93, 94, 95, 114, 115, 116, 117, 136, 137, 138, 139, 157, 158, 159, 160, 161, 181, 182, 183], "bbox": [0.15666666666666665, 0.16315359477124183, 0.3645098039215686, 0.41449346405228754], "iscrowd": 0, "area": 12020.032150000001, "rle": {"size": [612, 612], "counts": "ebi12Pc04L4L4M1N2N2N2N2N2O1N2N2N2N2N101N2N2N2N2O0O2N2O1N2N2O1N2N2O1N2O1N2O1N2N2O1N2N3N2M2N3N2M2N3N2M3M2N3M3N1N3M3M2N3N2O1N1O2N2O0O2N2O1N1O2O1N1O2O0O101O0O10001O000000001N10001O001O001N101O00001O0O2O001O0O101O0O2O00@mAPLS>n3PBPLR>m3QBQLP>m3TBPLn=m3f0N3M8I5K4K4M3L4M3M4K4M4K4M3L7I;E;E`0AaYX7"}, "label": "a black backpack with white writing"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small black carry on luggage bag with a black neck pillow on top of it\"."}], "task": "refering", "answer_template": "The \"a small black carry on luggage bag with a black neck pillow on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337], "bbox": [0.10318627450980392, 0.35913398692810455, 0.4243300653594771, 0.6975980392156862], "iscrowd": 0, "area": 27395.9679, "rle": {"size": [612, 612], "counts": "m`V12Rc01N3N1N2O0b_OKl=7PBMo=3nA2l<:ZAFg13m<;XAFe14P=:XAJ]10Y=:VANX1L_=b1UBbNi=i1kAZNS>Q2`ARN_>[3N2N2N1O1O1O1O1O001O1O1O1O1O001O1O1O1O10O10000000000O10000000000002N2N2N2N2N2N1O2N2N2N2N2N2N2N1O001O0010O10O0100O10O01000O0100O00100O001O10O01O0010O0001O0010O01O00010O001O010O000010O01O001O01O01O001N101O001O000O2O001O001O0O101O1O001O1N101O001O1O0O2O1N101N1O2O1N101N2O0O2O2M6K5J6K5J6K5K5K4L2N3M2N3L4M2N3M3M2N3M2M4M3M2N3M2N3M2N2SOi^OJYa0Ok^O2Va0Ho^O8Sa0AR_O`0n`0ZOW_Of0ca01000N2N1N3M3N2M2N3N2M3Mebb6"}, "label": "a small black carry on luggage bag with a black neck pillow on top of it"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small left bag in the center\"."}], "task": "refering", "answer_template": "The \"the small left bag in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337], "bbox": [0.10318627450980392, 0.35913398692810455, 0.4243300653594771, 0.6975980392156862], "iscrowd": 0, "area": 27395.9679, "rle": {"size": [612, 612], "counts": "m`V12Rc01N3N1N2O0b_OKl=7PBMo=3nA2l<:ZAFg13m<;XAFe14P=:XAJ]10Y=:VANX1L_=b1UBbNi=i1kAZNS>Q2`ARN_>[3N2N2N1O1O1O1O1O001O1O1O1O1O001O1O1O1O10O10000000000O10000000000002N2N2N2N2N2N1O2N2N2N2N2N2N2N1O001O0010O10O0100O10O01000O0100O00100O001O10O01O0010O0001O0010O01O00010O001O010O000010O01O001O01O01O001N101O001O000O2O001O001O0O101O1O001O1N101O001O1O0O2O1N101N1O2O1N101N2O0O2O2M6K5J6K5J6K5K5K4L2N3M2N3L4M2N3M3M2N3M2M4M3M2N3M2N3M2N2SOi^OJYa0Ok^O2Va0Ho^O8Sa0AR_O`0n`0ZOW_Of0ca01000N2N1N3M3N2M2N3N2M3Mebb6"}, "label": "the small left bag in the center"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suitcase on the very right\"."}], "task": "refering", "answer_template": "The \"the suitcase on the very right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 210, 211, 212, 213, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 408, 409, 410, 411, 412, 413, 414, 415, 416, 431, 432, 433, 434, 435, 436, 437, 438, 454, 455, 456, 457, 458, 459, 478, 479, 480], "bbox": [0.5393300653594771, 0.3820261437908497, 1.0, 0.9887581699346405], "iscrowd": 0, "area": 75354.58865, "rle": {"size": [612, 612], "counts": "SfU64jb0;F9G:F9G9G6C=^Oa0^Ob0^Oc0]Ob0_Oa0^Ob0^Oc0]Ob0_Oa0M4N100O2N1O1O2N100O2N1O1O1O2O0O2N2N2N1O2O1N2N2N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O13M4L4L4L3M2N00O100O10000O100O100O100O10000O100O100O10000O100O100O10O00100O010O00010O01O00001O0010O0001O00001O010O00001O0010O0001O00001O010O00001O001O00001O0O101O0O2O000O2O001N100O2O001N1000O00100O00100O0010O01O1O010O1O010O1O0010O01O10O01O1O010O1O010O1O010O002N101N2N101N1O2N2O0O2N101N2N1O2O1N1O2O0O2N2O0O2N1O2O1N1O2O0O2N2N101N2N101N1O2N2O0O2N101N2D;E<L3L5L4K4M4L4K4M4K4M4K5L3M4K4M4K5L3M4K4M4K5L3L5L4L3L5L3L5H8^Oa0_Ob0_O`0_Ob0^Ob0^OfF"}, "label": "the suitcase on the very right"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suitcase with a small bag sitting on top of it\"."}], "task": "refering", "answer_template": "The \"black suitcase with a small bag sitting on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 210, 211, 212, 213, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 408, 409, 410, 411, 412, 413, 414, 415, 416, 431, 432, 433, 434, 435, 436, 437, 438, 454, 455, 456, 457, 458, 459, 478, 479, 480], "bbox": [0.5393300653594771, 0.3820261437908497, 1.0, 0.9887581699346405], "iscrowd": 0, "area": 75354.58865, "rle": {"size": [612, 612], "counts": "SfU64jb0;F9G:F9G9G6C=^Oa0^Ob0^Oc0]Ob0_Oa0^Ob0^Oc0]Ob0_Oa0M4N100O2N1O1O2N100O2N1O1O1O2O0O2N2N2N1O2O1N2N2N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O13M4L4L4L3M2N00O100O10000O100O100O100O10000O100O100O10000O100O100O10O00100O010O00010O01O00001O0010O0001O00001O010O00001O0010O0001O00001O010O00001O001O00001O0O101O0O2O000O2O001N100O2O001N1000O00100O00100O0010O01O1O010O1O010O1O0010O01O10O01O1O010O1O010O1O010O002N101N2N101N1O2N2O0O2N101N2N1O2O1N1O2O0O2N2O0O2N1O2O1N1O2O0O2N2N101N2N101N1O2N2O0O2N101N2D;E<L3L5L4K4M4L4K4M4K4M4K5L3M4K4M4K5L3M4K4M4K5L3L5L4L3L5L3L5H8^Oa0_Ob0_O`0_Ob0^Ob0^OfF"}, "label": "black suitcase with a small bag sitting on top of it"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black bag farthest to the left in the back corner\"."}], "task": "refering", "answer_template": "The \"black bag farthest to the left in the back corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 27, 44, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 115, 134, 135, 136], "bbox": [0.008741830065359477, 0.06243464052287582, 0.27369281045751637, 0.3058660130718954], "iscrowd": 0, "area": 15475.872950000003, "rle": {"size": [612, 612], "counts": "RS31Pc04M4K5K4L5K5K4L5K5M2M4M3M2N3L3N3M2M4M2N3M2M4M2N3N1N3M2N2N2O0O2N2N2O0O2N2N101N2N2N101N2O001N2O1N110O00100O00100O00100O00100O010O10O0100O010O010O0001O010O1O1O1O1O1O2O0O1O1O001O01O0O1000000M3^Oc0I6000000000O1O2O1N7I8I7H8H8H;E2O0O2N1O2O0O1OO2N1O1O2N1O2N1O1O2N100O2N101N1ON2OA?3M2M4M2N2N3M2N3N1O3M2N3M2N2N3M2O4L5K5K4L5K4L5K8H;E<E:EYXY8"}, "label": "black bag farthest to the left in the back corner"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black bag in the back left\"."}], "task": "refering", "answer_template": "The \"the black bag in the back left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 27, 44, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 115, 134, 135, 136], "bbox": [0.008741830065359477, 0.06243464052287582, 0.27369281045751637, 0.3058660130718954], "iscrowd": 0, "area": 15475.872950000003, "rle": {"size": [612, 612], "counts": "RS31Pc04M4K5K4L5K5K4L5K5M2M4M3M2N3L3N3M2M4M2N3M2M4M2N3N1N3M2N2N2O0O2N2N2O0O2N2N101N2N2N101N2O001N2O1N110O00100O00100O00100O00100O010O10O0100O010O010O0001O010O1O1O1O1O1O2O0O1O1O001O01O0O1000000M3^Oc0I6000000000O1O2O1N7I8I7H8H8H;E2O0O2N1O2O0O1OO2N1O1O2N1O2N1O1O2N100O2N101N1ON2OA?3M2M4M2N2N3M2N3N1O3M2N3M2N2N3M2O4L5K5K4L5K4L5K8H;E<E:EYXY8"}, "label": "the black bag in the back left"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black bag is resting against the two walls\"."}], "task": "refering", "answer_template": "The \"the black bag is resting against the two walls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 27, 44, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 72, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 134, 135, 136], "bbox": [0.01107843137254902, 0.05928104575163399, 0.2790359477124183, 0.30348039215686273], "iscrowd": 0, "area": 15951.066650000002, "rle": {"size": [612, 612], "counts": "bX4S1o`0S1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N101O1N2O1O1N2O1O1N2O1O1N2O1O1O100O1O100O100O100O1O10O0100O1O100O100O100O1O100O100O1O100O100O100O1O10O001O001O1O001O1O0O2O001O1O001O001O1O001O001O1O001O1N101O0O2O1O001N101O1N1N3N1N3N2M2O2M3M2O2M2O2M01N100O2N1O2N101N1O1O2O0O2N1O2N100O20O2O2M2O1N2N2O1N2O1O2N1O1O1O1O1O1O1O6K6I7I6J7I6J7I7I6J7I7JP_W8"}, "label": "the black bag is resting against the two walls"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the bag on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 27, 44, 45, 46, 47, 48, 49, 66, 67, 68, 69, 70, 71, 72, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 134, 135, 136], "bbox": [0.01107843137254902, 0.05928104575163399, 0.2790359477124183, 0.30348039215686273], "iscrowd": 0, "area": 15951.066650000002, "rle": {"size": [612, 612], "counts": "bX4S1o`0S1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N101O1N2O1O1N2O1O1N2O1O1N2O1O1O100O1O100O100O100O1O10O0100O1O100O100O100O1O100O100O1O100O100O100O1O10O001O001O1O001O1O0O2O001O1O001O001O1O001O001O1O001O1N101O0O2O1O001N101O1N1N3N1N3N2M2O2M3M2O2M2O2M01N100O2N1O2N101N1O1O2O0O2N1O2N100O20O2O2M2O1N2N2O1N2O1O2N1O1O1O1O1O1O1O6K6I7I6J7I6J7I7I6J7I7JP_W8"}, "label": "the bag on the left in the right hand picture"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suit case with white button handle\"."}], "task": "refering", "answer_template": "The \"black suit case with white button handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 183, 184, 185, 186, 187, 188, 189, 190, 205, 206, 207, 208, 209, 210, 211, 228, 229, 230, 231, 232, 250, 251, 252, 253, 254, 272, 273, 274, 275, 276, 295, 296, 297, 298, 318, 319, 320, 341, 342, 363], "bbox": [0.3325816993464052, 0.11423202614379084, 0.6876470588235295, 0.7434477124183007], "iscrowd": 0, "area": 48332.9364, "rle": {"size": [612, 612], "counts": "`Pj3:ib0=D<C=D;D<E8H9F9H9F:G8H9F9H9F:G8G:G9G8G:G8G:G9F9H9G2M2O1N2O1O1N3N1N2O1N2O1O1N3N1N2O1N2O1O1N3N1N2O1O1N2O2M2O1N2O1O1N2O2M2O100O10000O2O000O10000O101O0O10000O10001N10000O100O1O2N1O1O1N2O1O1O2N1O1N2O1O1O1O2M2O1O1O1O1O1N3N100O100O101N100O100O100O10000O100O100O100OK6\\Oc0\\Oe0\\Oc0\\Oe0\\Oc0\\Oe0\\Oc0\\Oe0\\Od0I6O2O0O2O0O2O0O2N101N101N101N101N2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O1N101N101N101N1O2O0O2O0O2O0O2O0O2O1N101N1O2M2M4M2M4L3N3L3N3L3N3L3N8G:G8G9H9F9H8G:FVoa3"}, "label": "black suit case with white button handle"}]}
{"id": 171120, "image": "COCO_train2014_000000171120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large suitcase in the middle\"."}], "task": "refering", "answer_template": "The \"large suitcase in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 183, 184, 185, 186, 187, 188, 189, 190, 205, 206, 207, 208, 209, 210, 211, 228, 229, 230, 231, 232, 250, 251, 252, 253, 254, 272, 273, 274, 275, 276, 295, 296, 297, 298, 318, 319, 320, 341, 342, 363], "bbox": [0.3325816993464052, 0.11423202614379084, 0.6876470588235295, 0.7434477124183007], "iscrowd": 0, "area": 48332.9364, "rle": {"size": [612, 612], "counts": "`Pj3:ib0=D<C=D;D<E8H9F9H9F:G8H9F9H9F:G8G:G9G8G:G8G:G9F9H9G2M2O1N2O1O1N3N1N2O1N2O1O1N3N1N2O1N2O1O1N3N1N2O1O1N2O2M2O1N2O1O1N2O2M2O100O10000O2O000O10000O101O0O10000O10001N10000O100O1O2N1O1O1N2O1O1O2N1O1N2O1O1O1O2M2O1O1O1O1O1N3N100O100O101N100O100O100O10000O100O100O100OK6\\Oc0\\Oe0\\Oc0\\Oe0\\Oc0\\Oe0\\Oc0\\Oe0\\Od0I6O2O0O2O0O2O0O2N101N101N101N101N2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O0O2O1N101N101N101N1O2O0O2O0O2O0O2O0O2O1N101N1O2M2M4M2M4L3N3L3N3L3N3L3N8G:G8G9H9F9H8G:FVoa3"}, "label": "large suitcase in the middle"}]}
{"id": 572529, "image": "COCO_train2014_000000572529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in grey shirt and blue shorts in background\"."}], "task": "refering", "answer_template": "The \"person in grey shirt and blue shorts in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 25, 26, 27, 28, 43, 44, 45, 61, 62, 78, 79, 95, 96, 112, 113], "bbox": [0.48527607361963193, 0.0, 0.6941104294478528, 0.294375], "iscrowd": 0, "area": 9344.030700000007, "rle": {"size": [640, 489], "counts": "PTd41oc07I7I7I7I8H7I5K001O0000001O00001O00010O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O01O00h@WNY<j1\\BBb=>iA9U>G`Ae0_>\\OTAQ1k>S2O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1OJ6H8G9H8H8G9H8G:G8G9H8F:D<C=D<D<C=DYhm2"}, "label": "person in grey shirt and blue shorts in background"}]}
{"id": 572529, "image": "COCO_train2014_000000572529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man riding on skate board through white cones\"."}], "task": "refering", "answer_template": "The \"a man riding on skate board through white cones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 41, 42, 43, 44, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 155, 156, 157, 158, 159, 160, 161, 164, 165, 166, 167, 168, 169, 172, 173, 174, 175, 176, 177, 184, 185, 186, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 227, 228, 229, 230, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349], "bbox": [0.09999999999999999, 0.06742187499999999, 1.0, 0.9011250000000001], "iscrowd": 0, "area": 90539.70415, "rle": {"size": [640, 489], "counts": "XPo02ic05K6J5L4K6J5L4K5K6J5L4K6J5L4K5K6J5L4_Ob0]Ob0^Ob0^Ob0F;N1O1O2N1N2O1O2N1O1N3N1O1O1O2M2O1O2N1N2O2N1O1O1N3N1O1O2N1N2O1O2N1N2O2N1O1O1NM4M3L4L4M2M4M3L4L3N33M2N3M2M4M2N2N3M2M4M2N3M2M4M2N2N3M2M4M2O2M2O2M;F:E;F:EQ2PNn1QN101N100O101N100O101N100O1O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O10O0100O100O100O1O1L4L4M3L4M3L3M4M3L4MSGfKZ2V4iMlKV2P4nMRLP2l3QNVLn1j3QNXLn1g3RN\\Ll1c3TN_Lk1a3TNaLk1^3VNdLh1\\3WNfLg1Z3YNhLf1W3ZNlLd1T3[NnLd1Q3\\NQMc1n2^NTM`1l2_NVM`1i2`NYM_1f2aN]M]1c2bN_M]1`2cNbM\\1]2eNeMY1[2fNgMY1X2gNjMW1V2iNmMU1S2jNoMU1P2kNRNT1m1lNVNR1j1nNWNQ1h1oNZNP1e1PO^Nn0b1QO`Nn0_1ROdNl0\\1SOfNl0Y1UOhNj0W1VOlNh0T1WOnNg0R1YOPOf0P1YOSOe0n0YOTOe0m0ZOUOd0l0[OWOc0i0\\OYOb0h0]OZOb0g0\\O\\Oa0e0_O\\O`0d0_O^OfMZMZNY3o3@dMiMkMh2`4AcMYN\\MW2o4BcM[NZMT2R5DaM[N[MQ2S5F`M\\NZMo1U5G_M]NYMm1W5I]M]NZMk1W5J]M^NYMi1Y5K\\M^NZMg1Y5NZM^NZMe1\\5NXM`NZMb1]50WMaNYMa1_50UMbNYM_1b51RMcNZM\\1d54nLcN[MZ1h55iLdN]MW1j58eLdN^MV1l59bLh2_3[M\\Lf2d3]MXLd2h3_MTLb2m3`MlKe2S4^MgKe2Y4^M`Kf2`4^MYKe2h4]MQKg2o4\\MjJi2U5ZMeJi2\\5YM]Jk2c5XMWJk2i5f510Ob0^On1SNc0\\O1O2O1N1O2O0O2N101N1O2O0O2N101N1O2O0O2N101N1O2N101N2N101N1O2O0O2N101N1O2O0O2N01O01O01O000010O0001O00010O000010O000001O01O01O000010O0001O01O01O000010O000001O01O0001O0001O01O0001O01O000001O01O0000010O00000010O00000010O00000010O000001O00000010O00000001O00000010O000001O0000000010O000001O0000000001O0000000O1000001O00000000000000001O00001O1O1O001O1O1O001O1O001O1O1O001O1N101O1O1I6H9G9H]H"}, "label": "a man riding on skate board through white cones"}]}
{"id": 572529, "image": "COCO_train2014_000000572529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in purple shirt skateboarding\"."}], "task": "refering", "answer_template": "The \"man in purple shirt skateboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 41, 42, 43, 44, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 155, 156, 157, 158, 159, 160, 161, 164, 165, 166, 167, 168, 169, 172, 173, 174, 175, 176, 177, 184, 185, 186, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 227, 228, 229, 230, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349], "bbox": [0.09999999999999999, 0.06742187499999999, 1.0, 0.9011250000000001], "iscrowd": 0, "area": 90539.70415, "rle": {"size": [640, 489], "counts": "XPo02ic05K6J5L4K6J5L4K5K6J5L4K6J5L4K5K6J5L4_Ob0]Ob0^Ob0^Ob0F;N1O1O2N1N2O1O2N1O1N3N1O1O1O2M2O1O2N1N2O2N1O1O1N3N1O1O2N1N2O1O2N1N2O2N1O1O1NM4M3L4L4M2M4M3L4L3N33M2N3M2M4M2N2N3M2M4M2N3M2M4M2N2N3M2M4M2O2M2O2M;F:E;F:EQ2PNn1QN101N100O101N100O101N100O1O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O10O0100O100O100O1O1L4L4M3L4M3L3M4M3L4MSGfKZ2V4iMlKV2P4nMRLP2l3QNVLn1j3QNXLn1g3RN\\Ll1c3TN_Lk1a3TNaLk1^3VNdLh1\\3WNfLg1Z3YNhLf1W3ZNlLd1T3[NnLd1Q3\\NQMc1n2^NTM`1l2_NVM`1i2`NYM_1f2aN]M]1c2bN_M]1`2cNbM\\1]2eNeMY1[2fNgMY1X2gNjMW1V2iNmMU1S2jNoMU1P2kNRNT1m1lNVNR1j1nNWNQ1h1oNZNP1e1PO^Nn0b1QO`Nn0_1ROdNl0\\1SOfNl0Y1UOhNj0W1VOlNh0T1WOnNg0R1YOPOf0P1YOSOe0n0YOTOe0m0ZOUOd0l0[OWOc0i0\\OYOb0h0]OZOb0g0\\O\\Oa0e0_O\\O`0d0_O^OfMZMZNY3o3@dMiMkMh2`4AcMYN\\MW2o4BcM[NZMT2R5DaM[N[MQ2S5F`M\\NZMo1U5G_M]NYMm1W5I]M]NZMk1W5J]M^NYMi1Y5K\\M^NZMg1Y5NZM^NZMe1\\5NXM`NZMb1]50WMaNYMa1_50UMbNYM_1b51RMcNZM\\1d54nLcN[MZ1h55iLdN]MW1j58eLdN^MV1l59bLh2_3[M\\Lf2d3]MXLd2h3_MTLb2m3`MlKe2S4^MgKe2Y4^M`Kf2`4^MYKe2h4]MQKg2o4\\MjJi2U5ZMeJi2\\5YM]Jk2c5XMWJk2i5f510Ob0^On1SNc0\\O1O2O1N1O2O0O2N101N1O2O0O2N101N1O2O0O2N101N1O2N101N2N101N1O2O0O2N101N1O2O0O2N01O01O01O000010O0001O00010O000010O000001O01O01O000010O0001O01O01O000010O000001O01O0001O0001O01O0001O01O000001O01O0000010O00000010O00000010O00000010O000001O00000010O00000001O00000010O000001O0000000010O000001O0000000001O0000000O1000001O00000000000000001O00001O1O1O001O1O1O001O1O001O1O1O001O1N101O1O1I6H9G9H]H"}, "label": "man in purple shirt skateboarding"}]}
{"id": 572529, "image": "COCO_train2014_000000572529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the red shirt and white helmet watches the guy skate\"."}], "task": "refering", "answer_template": "The \"the man in the red shirt and white helmet watches the guy skate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 116, 117, 133, 134, 135, 150, 151, 152, 167, 168, 169, 185, 186, 202, 203, 220, 237, 254], "bbox": [0.8401840490797546, 0.23303124999999997, 1.0, 0.6585], "iscrowd": 0, "area": 10048.615399999997, "rle": {"size": [640, 489], "counts": "hRQ82mc04M3L3eNOc^O4Ua08c^OLVa0:g^OJQa0<m^OGn`0<Q_OHk`0:R_OKk`07P_OOl`03P_O2n`0c1L4M3M2M4M1N2O2M2O1N3N100O101N1O102M4M2M3M3N2M3N2M4L3N2M3N2N201OYLSAX2l>aMZAa2f>XM^Al2a>lLdAX3[>`LjAe3T>TLPBP4n=iKXBZ4e>4L4L4L4L4M4K5K5K6J5L3L101N101O0O2O001N101OYLkBb0T=BnCZOQ<i0QERNR;o1b3POP1lNbI"}, "label": "the man in the red shirt and white helmet watches the guy skate"}]}
{"id": 572529, "image": "COCO_train2014_000000572529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboarder with the red shirt\"."}], "task": "refering", "answer_template": "The \"the skateboarder with the red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 116, 117, 133, 134, 135, 150, 151, 152, 167, 168, 169, 185, 186, 202, 203, 220, 237, 254], "bbox": [0.8401840490797546, 0.23303124999999997, 1.0, 0.6585], "iscrowd": 0, "area": 10048.615399999997, "rle": {"size": [640, 489], "counts": "hRQ82mc04M3L3eNOc^O4Ua08c^OLVa0:g^OJQa0<m^OGn`0<Q_OHk`0:R_OKk`07P_OOl`03P_O2n`0c1L4M3M2M4M1N2O2M2O1N3N100O101N1O102M4M2M3M3N2M3N2M4L3N2M3N2N201OYLSAX2l>aMZAa2f>XM^Al2a>lLdAX3[>`LjAe3T>TLPBP4n=iKXBZ4e>4L4L4L4L4M4K5K5K6J5L3L101N101O0O2O001N101OYLkBb0T=BnCZOQ<i0QERNR;o1b3POP1lNbI"}, "label": "the skateboarder with the red shirt"}]}
{"id": 310390, "image": "COCO_train2014_000000310390.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table the man is at\"."}], "task": "refering", "answer_template": "The \"the table the man is at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 195, 196, 210, 212, 213, 227, 229, 244, 246, 261], "bbox": [0.0059791666666666665, 0.435234375, 0.5799791666666666, 0.697578125], "iscrowd": 0, "area": 7243.625749999997, "rle": {"size": [640, 480], "counts": "iT27ic08H0000000001O000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O000000000000000000000001O00000000000000000000000O1000000000000000001O0O1000000000000000000O10000oS=MTlB00O10000000000000000O13Mj0VOj0VOk0UOj0VOj0VOb0^O00000000^KfB^2Z=bMZDj0f;VOb400000000001O0000000000000000000001O000000010O000000001O000000001O0000001O01O000001O0000000<EV1iNX1hNl0TOO1O2N1O1O1O1O2O0O1O1O1O2bL]300O10001O000000001O000000001O000000001O0O1000001O000000001b1]NO1O1O1O2Nb0^OXnm3"}, "label": "the table the man is at"}]}
{"id": 310390, "image": "COCO_train2014_000000310390.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the table in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 195, 196, 210, 212, 213, 227, 229, 244, 246, 261], "bbox": [0.0059791666666666665, 0.435234375, 0.5799791666666666, 0.697578125], "iscrowd": 0, "area": 7243.625749999997, "rle": {"size": [640, 480], "counts": "iT27ic08H0000000001O000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000001O000000000000000000000001O00000000000000000000000O1000000000000000001O0O1000000000000000000O10000oS=MTlB00O10000000000000000O13Mj0VOj0VOk0UOj0VOj0VOb0^O00000000^KfB^2Z=bMZDj0f;VOb400000000001O0000000000000000000001O000000010O000000001O000000001O0000001O01O000001O0000000<EV1iNX1hNl0TOO1O2N1O1O1O1O2O0O1O1O1O2bL]300O10001O000000001O000000001O000000001O0O1000001O000000001b1]NO1O1O1O2Nb0^OXnm3"}, "label": "the table in the right hand picture"}]}
{"id": 89208, "image": "COCO_train2014_000000089208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in green sweater with baby horse\"."}], "task": "refering", "answer_template": "The \"girl in green sweater with baby horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 31, 32, 33, 34, 35, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 169, 196, 215, 216, 218, 219, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334, 335], "bbox": [0.294421875, 0.0018837209302325583, 0.5918593750000001, 0.9906279069767443], "iscrowd": 0, "area": 31455.599049999997, "rle": {"size": [430, 640], "counts": "S__24Z=3M3L3N3L4M3L3N3L4M3L4M2N3L4M3L3N3L3N2M3N2M3I8F9H8H8H8M3M3L4M3M4L3L4M3M3mIcKP3`4mLbKP3b4mL`KQ3c4kL`KS3d4iL^KU3e4hL]KU3g4gL\\KW3g4fL[KX3S3bJPNS2oNX3R3gJPNn1POY3Q3jJPNi1SOZ3n2nJmMh1WOX3l2RKiMh1]OS3l2WKbMh1Dn2k2\\K]Mg1Li2j2`KWMi11c2j2eKQMl16[2k2jKlLm1;T2m2oKeLP2?m1n2TL`LQ2d0g1n2ZLZLQ2j0a1o2^LULS2m0[1P3cLPLT2Q1W1Q3eLkKV2U1S1Q3hLgKX2X1n0R3kLcKY2[1k0T3lL_KZ2^1h0T3OkL0V30kLNW31iLNX33hLKZ34gLJZ36fLI[38eLQO_NoMn4o2cLPObNoMl4R3bLlNdNQNk4S3aLkNeNQNl4S3`LiNfNSNk4U3^LfNiNTNj4V3\\LeNmNSNg4Z3[LaNPOUNe4[3YL_NTOVNc4\\3XL\\NWOXNa4]3VLZN[OYN^4_3VLVN_OZN[4a3ULTNA[NZ4c3RLQNF[NX4f3QLmMI]NV4g3oKkMN]NS4i3nKhM1_NQ4k3kKeM6`Nn3m3kKaM8cNm3m3iK_M;eNl3m3gK^M=eNm3n3dK\\M`0gNk3P4bKXMe0hNi3Q4_KSMm0mNc3R4^KbL^1]OR3Y4YKPLo1He2Z5[MfJd2Z5]MgJ`2[5`MfJ^2Z5dMfJY2[5hMfJU2\\5lMdJS2[5QNcJn1]5TNbJk1]5YNaJg1^5[NbJc1]5aNaJ^1_5dN`J^1\\5fNbJ\\1[5eNeJ]1W5eNjJ]1Q5eNoJ]1n4dNRK^1j4dNVK_1d4dN\\K^1`4dN`K^1\\4dNeK]1V4fNjKZ1T4hNlKY1Q4iNoKmNJ@T4f1RLgN7YOc3T2VLaN<XO[3Z2ZLZN`0ZOR3`2^LSNf0ZOh2g2cLlMj0ZO`2n2eLeMP1\\OV2S3kL^MT1\\Om1Z3oLXMc1QO\\1i3RMSMe1SOX1k3SMoLi1UOS1m3UMkLk1WOo0o3VMhLn1YOj0P4YMdLP2[Of0R4ZMaLS2\\Ob0T4\\M\\LV2_O=V4]MYLY2@9X4_MUL[2B5Z4`MTL\\2A3\\4cMRLZ2@4^4dMQLX2@3`4fMoKY2_O1b4hMnKW2_O0d4kMkKW2_ONf4mMjKW2]OKj4oMiKW2[OJl4QNgKX2ZOFP5TNeKW2XOFS5UNcKW2XODU5VNbKU5_4nJ[KU5e4oJTKS5m4QKlJR5T5c1100O100OQKPHb4P8]KSHa4m7]KWHb4g7^K\\H`4d7_K_H_4`7`KdH^4\\7aKhH\\4W7dKlHZ4T7eKoHY4P7fKTIX4l6gKWIW4h6iK[IU4e6jK_IS4P8N1O2RMfFd1[9[NgGc0[8ZOXH4j7Jn2O2M3N1O2M2O2M3N1O2M2O2N2M2O_f]3"}, "label": "girl in green sweater with baby horse"}]}
{"id": 89208, "image": "COCO_train2014_000000089208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in green sweater\"."}], "task": "refering", "answer_template": "The \"girl in green sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 31, 32, 33, 34, 35, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 169, 196, 215, 216, 218, 219, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334, 335], "bbox": [0.294421875, 0.0018837209302325583, 0.5918593750000001, 0.9906279069767443], "iscrowd": 0, "area": 31455.599049999997, "rle": {"size": [430, 640], "counts": "S__24Z=3M3L3N3L4M3L3N3L4M3L4M2N3L4M3L3N3L3N2M3N2M3I8F9H8H8H8M3M3L4M3M4L3L4M3M3mIcKP3`4mLbKP3b4mL`KQ3c4kL`KS3d4iL^KU3e4hL]KU3g4gL\\KW3g4fL[KX3S3bJPNS2oNX3R3gJPNn1POY3Q3jJPNi1SOZ3n2nJmMh1WOX3l2RKiMh1]OS3l2WKbMh1Dn2k2\\K]Mg1Li2j2`KWMi11c2j2eKQMl16[2k2jKlLm1;T2m2oKeLP2?m1n2TL`LQ2d0g1n2ZLZLQ2j0a1o2^LULS2m0[1P3cLPLT2Q1W1Q3eLkKV2U1S1Q3hLgKX2X1n0R3kLcKY2[1k0T3lL_KZ2^1h0T3OkL0V30kLNW31iLNX33hLKZ34gLJZ36fLI[38eLQO_NoMn4o2cLPObNoMl4R3bLlNdNQNk4S3aLkNeNQNl4S3`LiNfNSNk4U3^LfNiNTNj4V3\\LeNmNSNg4Z3[LaNPOUNe4[3YL_NTOVNc4\\3XL\\NWOXNa4]3VLZN[OYN^4_3VLVN_OZN[4a3ULTNA[NZ4c3RLQNF[NX4f3QLmMI]NV4g3oKkMN]NS4i3nKhM1_NQ4k3kKeM6`Nn3m3kKaM8cNm3m3iK_M;eNl3m3gK^M=eNm3n3dK\\M`0gNk3P4bKXMe0hNi3Q4_KSMm0mNc3R4^KbL^1]OR3Y4YKPLo1He2Z5[MfJd2Z5]MgJ`2[5`MfJ^2Z5dMfJY2[5hMfJU2\\5lMdJS2[5QNcJn1]5TNbJk1]5YNaJg1^5[NbJc1]5aNaJ^1_5dN`J^1\\5fNbJ\\1[5eNeJ]1W5eNjJ]1Q5eNoJ]1n4dNRK^1j4dNVK_1d4dN\\K^1`4dN`K^1\\4dNeK]1V4fNjKZ1T4hNlKY1Q4iNoKmNJ@T4f1RLgN7YOc3T2VLaN<XO[3Z2ZLZN`0ZOR3`2^LSNf0ZOh2g2cLlMj0ZO`2n2eLeMP1\\OV2S3kL^MT1\\Om1Z3oLXMc1QO\\1i3RMSMe1SOX1k3SMoLi1UOS1m3UMkLk1WOo0o3VMhLn1YOj0P4YMdLP2[Of0R4ZMaLS2\\Ob0T4\\M\\LV2_O=V4]MYLY2@9X4_MUL[2B5Z4`MTL\\2A3\\4cMRLZ2@4^4dMQLX2@3`4fMoKY2_O1b4hMnKW2_O0d4kMkKW2_ONf4mMjKW2]OKj4oMiKW2[OJl4QNgKX2ZOFP5TNeKW2XOFS5UNcKW2XODU5VNbKU5_4nJ[KU5e4oJTKS5m4QKlJR5T5c1100O100OQKPHb4P8]KSHa4m7]KWHb4g7^K\\H`4d7_K_H_4`7`KdH^4\\7aKhH\\4W7dKlHZ4T7eKoHY4P7fKTIX4l6gKWIW4h6iK[IU4e6jK_IS4P8N1O2RMfFd1[9[NgGc0[8ZOXH4j7Jn2O2M3N1O2M2O2M3N1O2M2O2N2M2O_f]3"}, "label": "girl in green sweater"}]}
{"id": 89208, "image": "COCO_train2014_000000089208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy with long blonde hair and a red jacket\"."}], "task": "refering", "answer_template": "The \"a little boy with long blonde hair and a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 197, 198, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341], "bbox": [0.5707187499999999, 0.4679767441860465, 0.88778125, 0.9893255813953488], "iscrowd": 0, "area": 31213.101100000003, "rle": {"size": [430, 640], "counts": "Z`i45Y=1O2N1O2N1N3N1O2N1O1O2N1O2N1O2N1O1O1O1O1TGYOo4h0gJBX5`0^JJ`56_JL`55^JMa54]JOa52]J0a52]J1a50]J2b5O\\J4b5M\\J5c5K\\J8b5I]J8b5I\\J:b5G\\J;c5F[J=c5D[J>d5CZJ`0d5@[Ja0e5@YJa0g5@WJa0i5@UJa0l5_OnIf0T6@\\Ih0f6AhHf0Y7DTHd0n7]22N2N2N1O2N00001O00000000000000001O00000000000000001O0000000000001O00000000001O00000000001O000000001O000000O100O100]O`KWHa4c7h0J6M3M3M3N2M3M3N2N2O1000000O10000O1000000O10000O1000000O10000O10000O10O1000O10000O1000000O1001N100O2O0O1O2O0O100O2N100O2O0O1O2O0O101N1O101N100O101N2N2O2M2O1N3M2kMbHWNa7`1iHZN[7]1PI^NT7]1RI^NQ7_1TI\\No6a1UI[No6a1VIZNm6c1XIXNl6d1YIVNl6f1YIUNm6f1WIUNP7d1UIWNR7b1b2J7I8H=C<Dgnm0"}, "label": "a little boy with long blonde hair and a red jacket"}]}
{"id": 89208, "image": "COCO_train2014_000000089208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing a red jacket\"."}], "task": "refering", "answer_template": "The \"a girl wearing a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 197, 198, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341], "bbox": [0.5707187499999999, 0.4679767441860465, 0.88778125, 0.9893255813953488], "iscrowd": 0, "area": 31213.101100000003, "rle": {"size": [430, 640], "counts": "Z`i45Y=1O2N1O2N1N3N1O2N1O1O2N1O2N1O2N1O1O1O1O1TGYOo4h0gJBX5`0^JJ`56_JL`55^JMa54]JOa52]J0a52]J1a50]J2b5O\\J4b5M\\J5c5K\\J8b5I]J8b5I\\J:b5G\\J;c5F[J=c5D[J>d5CZJ`0d5@[Ja0e5@YJa0g5@WJa0i5@UJa0l5_OnIf0T6@\\Ih0f6AhHf0Y7DTHd0n7]22N2N2N1O2N00001O00000000000000001O00000000000000001O0000000000001O00000000001O00000000001O000000001O000000O100O100]O`KWHa4c7h0J6M3M3M3N2M3M3N2N2O1000000O10000O1000000O10000O1000000O10000O10000O10O1000O10000O1000000O1001N100O2O0O1O2O0O100O2N100O2O0O1O2O0O101N1O101N100O101N2N2O2M2O1N3M2kMbHWNa7`1iHZN[7]1PI^NT7]1RI^NQ7_1TI\\No6a1UI[No6a1VIZNm6c1XIXNl6d1YIVNl6f1YIUNm6f1WIUNP7d1UIWNR7b1b2J7I8H=C<Dgnm0"}, "label": "a girl wearing a red jacket"}]}
{"id": 89208, "image": "COCO_train2014_000000089208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a foal\"."}], "task": "refering", "answer_template": "The \"a foal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 212, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.000171875, 0.3243488372093023, 0.662609375, 1.0], "iscrowd": 0, "area": 53042.606100000005, "rle": {"size": [430, 640], "counts": "k8c4j8100O100O100O100O100000000O100000000O1000000000000001O00000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000000000000000000000000000000000O100O100O100O100O100O100O1O100O100O100O1O1O1O1O1O1O1O1O100O100001O4L3M4L3M2N2N2N2N2N2N2N2N2N2N001O1O001O001O1O001O001O000000001O0000000000O1O1O1O1O1N2O1O1O1O1O1O1000000O100000000O1000000O1001O001O1O1O1O002N2N2N2N2NkNYLeHe3Y7_LgH_3V7dLkH[3S7hLmHW3Q7kLPIT3m6oLTIP3\\6aMdI^2W6gMjIX2S6kMnIT2P6oMPJP2n5RNSJm1j5VNWJh1h5[NXJc1g5_NZJ_1d5dN]JZ1b5iN^JU1a5mN`JQ1^5ROcJl0\\5WOcJh0\\5ZOeJd0Y5@gJ>X5DiJ:W5HiJ6W5LhJ3W5OiJOX53hJKX56hJIX59gJFX5=gJBY5?hJ_OX5c0gJ\\OY5e0gJYOY5k0eJTO[5n0dJPO]5S1^JoNa5T1^JjNc5X1\\JgNd5\\1ZJcNf5_1YJ_Ng5d1XJ[Nh5h1VJVNk5m1TJQNl5R2QJmMo5V2kIQL_Og1g6Y2jIhMY6Z2fI_LB<j6Y3aIZLO1c6f3]IWL;GZ6S4[IULR7IkHT42RLP77dHi3:oKP7m4nHSKQ7o4jHTKU7P5jHoJU7S5kHkJX7U5gHjJ[7U5eHiJ]7X5bHgJ`7X5aHeJb7Z5^HeJf7Y5YHeJl7X5THgJo7W55O000N2N3L6K4K4M4L3L3N3M2M4L3J6K5K5K4L5N200O1O100O1O2O0O1O100O1O101O1N3N000O100O10000O10001N2O1O1N3N1N101O0O2O001N101O0O101N101O0O2O000O101O00000O2O0000001N1000001O0O10001O000O2O1O1O001O1N101O00001O0O2O001OO1O1O1O100001O000O2O1N101N1O2O1N1O2O1K5N2L4OU`j2"}, "label": "a foal"}]}
{"id": 89208, "image": "COCO_train2014_000000089208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a foal with the mother close by\"."}], "task": "refering", "answer_template": "The \"a foal with the mother close by\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 212, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.000171875, 0.3243488372093023, 0.662609375, 1.0], "iscrowd": 0, "area": 53042.606100000005, "rle": {"size": [430, 640], "counts": "k8c4j8100O100O100O100O100000000O100000000O1000000000000001O00000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000000000000000000000000000000000O100O100O100O100O100O100O1O100O100O100O1O1O1O1O1O1O1O1O100O100001O4L3M4L3M2N2N2N2N2N2N2N2N2N2N001O1O001O001O1O001O001O000000001O0000000000O1O1O1O1O1N2O1O1O1O1O1O1000000O100000000O1000000O1001O001O1O1O1O002N2N2N2N2NkNYLeHe3Y7_LgH_3V7dLkH[3S7hLmHW3Q7kLPIT3m6oLTIP3\\6aMdI^2W6gMjIX2S6kMnIT2P6oMPJP2n5RNSJm1j5VNWJh1h5[NXJc1g5_NZJ_1d5dN]JZ1b5iN^JU1a5mN`JQ1^5ROcJl0\\5WOcJh0\\5ZOeJd0Y5@gJ>X5DiJ:W5HiJ6W5LhJ3W5OiJOX53hJKX56hJIX59gJFX5=gJBY5?hJ_OX5c0gJ\\OY5e0gJYOY5k0eJTO[5n0dJPO]5S1^JoNa5T1^JjNc5X1\\JgNd5\\1ZJcNf5_1YJ_Ng5d1XJ[Nh5h1VJVNk5m1TJQNl5R2QJmMo5V2kIQL_Og1g6Y2jIhMY6Z2fI_LB<j6Y3aIZLO1c6f3]IWL;GZ6S4[IULR7IkHT42RLP77dHi3:oKP7m4nHSKQ7o4jHTKU7P5jHoJU7S5kHkJX7U5gHjJ[7U5eHiJ]7X5bHgJ`7X5aHeJb7Z5^HeJf7Y5YHeJl7X5THgJo7W55O000N2N3L6K4K4M4L3L3N3M2M4L3J6K5K5K4L5N200O1O100O1O2O0O1O100O1O101O1N3N000O100O10000O10001N2O1O1N3N1N101O0O2O001N101O0O101N101O0O2O000O101O00000O2O0000001N1000001O0O10001O000O2O1O1O001O1N101O00001O0O2O001OO1O1O1O100001O000O2O1N101N1O2O1N1O2O1K5N2L4OU`j2"}, "label": "a foal with the mother close by"}]}
{"id": 228474, "image": "COCO_train2014_000000228474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side profile airplane\"."}], "task": "refering", "answer_template": "The \"the side profile airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 56, 57, 66, 67, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 125, 126, 128, 129, 130, 131, 132, 133], "bbox": [0.369765625, 0.0641225626740947, 1.0, 0.4206406685236769], "iscrowd": 0, "area": 20427.55315, "rle": {"size": [359, 640], "counts": "YVc24S;1N1O2O0O2N101N1O2O0O2O0O1O2O0O2N101N1O2O0O2O0O2N101N1O2O0O1O1000001N100000001O0000000O2O000O1L4O1aF`NW9a1hF`NW9`1iF`NV9a1jF`NU9`1kF`NU9`1kFaNT9`1kFaNT9_1lFaNT9_1lFbNS9^1mFcNR9]1nFcNR9^1mFcNR9]1nFcNQ9^1oFcNP9]1mFgNR9g101O00000000000O1000O10000000000000002N1N2O00000000000BlFiNT9X1mFfNS9Z1PGbNQ9^1RG_Nn8a1<0000000000001O00000000001O000010O0001O00001O00001O0000000000000000000000000000O10000000001O0000000O1000000000000000000O1000000000000000000001O1O0010O01O00000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000:F0000000000000000000000O1000000000000O1L4L40000O10000000001O00000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10000O10000O1000O10O10000O1000000O1000O010O10O010O010O010O010O010N101N101N101N101N101N10001N100O2O0O100O2O0O100O2O0O101N100O101N10[O`GdN_8Z1fGeNY8X1lGfNT8W1PHhNQ8T1THkNl7Q1YHmNh7o0\\HQOd7k0`HUO`7g0dHYO\\7c0hH]OX7?kHAV7;nHER77RIIn63VIMj6OZI1f6K^I4c6HaI8f81000O100000O10O1000000O10O100000O100M2MYM"}, "label": "the side profile airplane"}]}
{"id": 228474, "image": "COCO_train2014_000000228474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"airplane on runway\"."}], "task": "refering", "answer_template": "The \"airplane on runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 56, 57, 66, 67, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 125, 126, 128, 129, 130, 131, 132, 133], "bbox": [0.369765625, 0.0641225626740947, 1.0, 0.4206406685236769], "iscrowd": 0, "area": 20427.55315, "rle": {"size": [359, 640], "counts": "YVc24S;1N1O2O0O2N101N1O2O0O2O0O1O2O0O2N101N1O2O0O2O0O2N101N1O2O0O1O1000001N100000001O0000000O2O000O1L4O1aF`NW9a1hF`NW9`1iF`NV9a1jF`NU9`1kF`NU9`1kFaNT9`1kFaNT9_1lFaNT9_1lFbNS9^1mFcNR9]1nFcNR9^1mFcNR9]1nFcNQ9^1oFcNP9]1mFgNR9g101O00000000000O1000O10000000000000002N1N2O00000000000BlFiNT9X1mFfNS9Z1PGbNQ9^1RG_Nn8a1<0000000000001O00000000001O000010O0001O00001O00001O0000000000000000000000000000O10000000001O0000000O1000000000000000000O1000000000000000000001O1O0010O01O00000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000:F0000000000000000000000O1000000000000O1L4L40000O10000000001O00000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10000O10000O1000O10O10000O1000000O1000O010O10O010O010O010O010O010N101N101N101N101N101N10001N100O2O0O100O2O0O100O2O0O101N100O101N10[O`GdN_8Z1fGeNY8X1lGfNT8W1PHhNQ8T1THkNl7Q1YHmNh7o0\\HQOd7k0`HUO`7g0dHYO\\7c0hH]OX7?kHAV7;nHER77RIIn63VIMj6OZI1f6K^I4c6HaI8f81000O100000O10O1000000O10O100000O100M2MYM"}, "label": "airplane on runway"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cream colored leather sofa\"."}], "task": "refering", "answer_template": "The \"a cream colored leather sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 258, 259, 260], "bbox": [0.038546875, 0.366375, 0.44325000000000003, 0.7111250000000001], "iscrowd": 0, "area": 23397.641500000005, "rle": {"size": [480, 640], "counts": "Wn;5f>6J6J7M2O1N2O1O1O1N2O1O1N2O1O2N1N2O1O1O1N2O1O1O0O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000001N101O1O0O2O1O001N2O001O1N101O1N101O1O0O2O001O1N101O1O0O2O1O001N2O001O1N101O1O0O2O1O001N2O001N2O1O001N2O1O001N2OZ1fN00O010O100O010O100O010O100O010O100O010O100O01000O010O100O010O100O010O100O10000O100O100O100O100O102M3N1N3N2M2O2N2M2O2M3N2M2O2M3N1N3N1N100O100O10000O100O2O0O100O100O100O100O101N10000O100O100O100O100O2O0O100O10000O100O1O101N100O1O100O1O100O1O101N100O1O100O1O100O1O101N100O1OR1oN?@:FdcV5"}, "label": "a cream colored leather sofa"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green leather couch in a living room\"."}], "task": "refering", "answer_template": "The \"a green leather couch in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 258, 259, 260], "bbox": [0.038546875, 0.366375, 0.44325000000000003, 0.7111250000000001], "iscrowd": 0, "area": 23397.641500000005, "rle": {"size": [480, 640], "counts": "Wn;5f>6J6J7M2O1N2O1O1O1N2O1O1N2O1O2N1N2O1O1O1N2O1O1O0O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000001N101O1O0O2O1O001N2O001O1N101O1N101O1O0O2O001O1N101O1O0O2O1O001N2O001O1N101O1O0O2O1O001N2O001N2O1O001N2O1O001N2OZ1fN00O010O100O010O100O010O100O010O100O010O100O01000O010O100O010O100O010O100O10000O100O100O100O100O102M3N1N3N2M2O2N2M2O2M3N2M2O2M3N1N3N1N100O100O10000O100O2O0O100O100O100O100O101N10000O100O100O100O100O2O0O100O10000O100O1O101N100O1O100O1O100O1O101N100O1O100O1O100O1O101N100O1OR1oN?@:FdcV5"}, "label": "a green leather couch in a living room"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white single chair\"."}], "task": "refering", "answer_template": "The \"a white single chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295, 316, 317, 318], "bbox": [0.612671875, 0.43314583333333334, 0.9195781249999999, 0.8089375], "iscrowd": 0, "area": 16973.142700000004, "rle": {"size": [480, 640], "counts": "^Qh51k>4L5L3L4L4M3L5N100O1O1O2O0O1001O01O000000010O000000010O0000000010O00000001O0O100O101N1M3M3M3M4K4M3M3M3N3M2O1O1O1O2O000000010O000000010O000001O01O0001O0001O0001O0001O01O0000010O000000010O00PEgMY9Y2WFWNi9i1gEhNY:X1VEXOj:_2000010O000001O0001O0001O0001O01O0000010O000000010O00000010O000001O0001M2N2N2N2N3M2N2N2N2N3N1O1N2O1O2M2O1O1N02L3N3L3M3N3L3N3L3N3L3N3L3N3L3N3L3M4M2M4M2M3N3L3N3L4M2M4L3O2M2O2N3M2N3M2N3L3L5L4K4M;D<D>Cdeg0"}, "label": "a white single chair"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white chair facing the couch\"."}], "task": "refering", "answer_template": "The \"white chair facing the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295, 316, 317, 318], "bbox": [0.612671875, 0.43314583333333334, 0.9195781249999999, 0.8089375], "iscrowd": 0, "area": 16973.142700000004, "rle": {"size": [480, 640], "counts": "^Qh51k>4L5L3L4L4M3L5N100O1O1O2O0O1001O01O000000010O000000010O0000000010O00000001O0O100O101N1M3M3M3M4K4M3M3M3N3M2O1O1O1O2O000000010O000000010O000001O01O0001O0001O0001O0001O01O0000010O000000010O00PEgMY9Y2WFWNi9i1gEhNY:X1VEXOj:_2000010O000001O0001O0001O0001O01O0000010O000000010O00000010O000001O0001M2N2N2N2N3M2N2N2N2N3N1O1N2O1O2M2O1O1N02L3N3L3M3N3L3N3L3N3L3N3L3N3L3N3L3M4M2M4M2M3N3L3N3L4M2M4L3O2M2O2N3M2N3M2N3L3L5L4K4M;D<D>Cdeg0"}, "label": "white chair facing the couch"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white armchair opposite a white loveseat\"."}], "task": "refering", "answer_template": "The \"a white armchair opposite a white loveseat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 341], "bbox": [0.6094375000000001, 0.434375, 0.922265625, 0.8412083333333332], "iscrowd": 0, "area": 19324.856450000003, "rle": {"size": [480, 640], "counts": "QUg51W=6\\DL\\;8dDMS;7mDKm::REFi:?WEAe:c0[E]Oa:g0`EZO]:g0cE^OX:b0hECS:>lEEQ:;oEFP::PFFP::QFEo9;QFEo9;QFEo9<PFDP:<PFDP:<PFDP:<QFCo9=QFCo9=QFBP:?oE@S:?mE^OV:b0jEZOZ:f0gEVO\\:j0dETO^:l0bETO^:m0aETO^:l0bETO^:l0bETO^:l0cESO]:m0cETO\\:l0dETO\\:m0cESO]:m0cESO]:m0cESO]:m0dESO[:m0eESO[:m0eESO[:n0dERO\\:n0dERO\\:n0dESO[:m0fEROY:o0gEQOU:S1kEmNR:W1mEiNP:Z1PFgNk9]1UFcNc9e1^FZN[9m1eFSNZ9n1fFSNX9o1gFQNX9P2hFPNW9Q2iFoMW9Q2jFnMU9S2kFnMT9R2lFnMT9S2kFmMU9S2kFmMV9R2jFnMV9R2kFnMT9R2lFnMT9R2n10O00000001O01hDRN\\9n1cFSN]9m1bFTN_9k1_FXN`9h1_FYNa9g1^FZNb9g1\\FZNe9e1ZF\\Nf9d1XF^Nh9b1WF`Nh9`1WFaNi9_1VFbNk9^1SFcNm9]1QFeNo9[1PFgNo9Y1PFhNQ:W1nEjNR:V1mEhNV:Y1gE^Nd:b1\\ETNn:l1SEhMY;W2h000001PEgMX9X2[FUNe9k1nEbNR:^1aEoN_:Q1SE]Om:\\20O0000001O01O0001O0001O0001O0000010O000000010O0000001O01O000001O01O0O1N2N2N:G<C3M2N2O1N2N3M2O1N2N2N2ON2N3L3N2M4M2N3M2O2M2N3N1N3N1N2N3N1N3M2O2M2N3N1N2N3M2M4L3M4L4L4L4L4L4L5K4L5K5K5K5K5K5K5K5K5K5YOPB8U>EnA6W>FlA6a>LmVg0"}, "label": "a white armchair opposite a white loveseat"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small , dark brown , wooden end table with a can on top of it next to a beige leather armchair\"."}], "task": "refering", "answer_template": "The \"a small , dark brown , wooden end table with a can on top of it next to a beige leather armchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 314, 315, 316, 334, 338, 339, 357, 361, 384], "bbox": [0.51953125, 0.6211041666666667, 0.77734375, 0.9570416666666667], "iscrowd": 0, "area": 10633.59375, "rle": {"size": [480, 640], "counts": "l_l4<X><D=C=C<C>C=I6O2N2N1O2N1O2N2N^MZEl0e:UO^FI`98c21O1O1O1O2O0O1O100O1O101N1O100O1O100O2N100O1O100O101N1O100O1O100O10001N10000O1001O0001O0001O01O00010O000010O0000010O0001O01O01O000010O0000010O0001O00010O00001O01O01O00001O01O01O0000102Ma0_Ob0^OS1mN101N1O2N1O2O0O1O2N1O1OO2QMUF8Z:GSF[O_:d0o1O1N3N1O1O1NgCjNl:U1REoNm:P1PEUOo:j0nD[OR;c0kDAU;>hDGW;8gDLX;3eD2Z;McD8\\;GaD>_;@^De0a;ZO]Dj0b;UO[DP1d;oNYDV1f;iNWD\\1h;cNUDb1j;j01N1N2K5J6J6J7I6J6J6J6J9G:F;E:FhWR2"}, "label": "a small , dark brown , wooden end table with a can on top of it next to a beige leather armchair"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown table with a can on it beside a white chair\"."}], "task": "refering", "answer_template": "The \"a brown table with a can on it beside a white chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 314, 315, 316, 334, 338, 339, 357, 361, 384], "bbox": [0.51953125, 0.6211041666666667, 0.77734375, 0.9570416666666667], "iscrowd": 0, "area": 10633.59375, "rle": {"size": [480, 640], "counts": "l_l4<X><D=C=C<C>C=I6O2N2N1O2N1O2N2N^MZEl0e:UO^FI`98c21O1O1O1O2O0O1O100O1O101N1O100O1O100O2N100O1O100O101N1O100O1O100O10001N10000O1001O0001O0001O01O00010O000010O0000010O0001O01O01O000010O0000010O0001O00010O00001O01O01O00001O01O01O0000102Ma0_Ob0^OS1mN101N1O2N1O2O0O1O2N1O1OO2QMUF8Z:GSF[O_:d0o1O1N3N1O1O1NgCjNl:U1REoNm:P1PEUOo:j0nD[OR;c0kDAU;>hDGW;8gDLX;3eD2Z;McD8\\;GaD>_;@^De0a;ZO]Dj0b;UO[DP1d;oNYDV1f;iNWD\\1h;cNUDb1j;j01N1N2K5J6J6J7I6J6J6J6J9G:F;E:FhWR2"}, "label": "a brown table with a can on it beside a white chair"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair with a red cushion\"."}], "task": "refering", "answer_template": "The \"a brown chair with a red cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327, 349, 350], "bbox": [0.018015625, 0.5973958333333333, 0.232609375, 0.9096875], "iscrowd": 0, "area": 15441.451750000004, "rle": {"size": [480, 640], "counts": "Rm5S1m=7I7I7H8I8H7I7I7nCTMb;U301O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001N101O001N101O03Nk0TOd0]O1^NTDLn;0]DHc;5hDBZ;9RE_On:>]EYOe:c0o1L3M7I9FhiU7"}, "label": "a brown chair with a red cushion"}]}
{"id": 15485, "image": "COCO_train2014_000000015485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair with red seat cushion\"."}], "task": "refering", "answer_template": "The \"a wooden chair with red seat cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327, 349, 350], "bbox": [0.018015625, 0.5973958333333333, 0.232609375, 0.9096875], "iscrowd": 0, "area": 15441.451750000004, "rle": {"size": [480, 640], "counts": "Rm5S1m=7I7I7H8I8H7I7I7nCTMb;U301O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001N101O001N101O03Nk0TOd0]O1^NTDLn;0]DHc;5hDBZ;9RE_On:>]EYOe:c0o1L3M7I9FhiU7"}, "label": "a wooden chair with red seat cushion"}]}
{"id": 146561, "image": "COCO_train2014_000000146561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing black and white shirt\"."}], "task": "refering", "answer_template": "The \"woman wearing black and white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.023875, 0.1284065934065934, 0.44509374999999995, 0.9983241758241758], "iscrowd": 0, "area": 53751.6892, "rle": {"size": [364, 640], "counts": "_b53U;;F;E:F;D;F;E:C6J6K6L3L4M4L3N2M4L3N2N3M2O1O2M2O1O2N1O101N100O2O0O2O000O2O001N101O001O00001O001YNaK_L_4W3PLdLQ4W3XLdLh3Y3`LbLa3[3cLcL]3[3fLdLZ3Z3jLdLW3X3nLeLS3Y3PMfLQ3W3RMhLn2W3TMhLm2V3VMhLj2W3XMhLh2W3ZMhLg2V3[MiLe2V3]MiLd2U3^MjLb2U3`MjLa2T3bMiL_2W3bMhL_2V3cMhL^2W3dMhL\\2W3fMgL\\2X3eMgL[2X3n2N1O2O0O2N1O101N1O2N101N1O2N100O2N1O2O0O2N1O2O0O1O2O0O2O001N100O2O001N10001N101O0O101N101O000O2O001O00001O001O00001O001O01O01O010O01O01O010O00010O0010O000[JaN]1`1ZNgNg1Y1XNhNh1X1WNhNj1Y1TNhNl1X1SNhNn1Y1PNhNP2X1oMhNR2X1mMiNS2X1jMiNW2W1hMiNY2W1fMjNZ2W1dMiN]2X1aMiN_2X1_MhNb2Y1\\MhNd2X1[MhNf2Y1XMhNhM]OU4l1QNgNjM^OU4l1oMeNnM@R4l1nMdNPNAR4l1lMcNSNBP4l1jMcNWNAo3n1fMbN[NAo3o1bMbN_N@n3Q2^M`NeN_Om3T2YM_NjN]Om3\\2PMXNRO^Om3b2gLQN]O]Ol3i2]LmMG[Ok3n2SLkM2XOk3X3bKfMc0SOj3k4WLUKi3k4WLVKh3j4WLWKi3h4XLYKg3g4YLYKg3g4XL[Kg3e4YL[Kg3d4ZL\\Kf3d4ZL]Ke3c4ZL^Kf3a4[L`Kd3`4\\L`Kd3`4\\LaKc3_4\\LbKd3^4\\LcKc3\\4^LdKb3\\4]LfKb3Z4^LgKa3Y4_LhK`3X4`LiK_3V4aLlK^3U4aLkK_3U4aLlK^3T4bLmK]3S4bLoK]3Q4cLPL\\3Q4cLPL\\3P4dLQL[3o3eLSLY3m3fLULY3l3fLVLX3l3fLULY3l3fLVLX3l3fLULY3l3fLSL[3o3cLmKa3V4\\LfKh3\\4VLaKm3a4QL[KS4g4jKZKV4h4gKYKY4n5O00001O000000M3N20000000000000O107I4LO1M3N200O100000WLnHX3R7eLRIZ3n6bLVIc01e1l6cMUIa08j1Q7SNVIj1i6SN^Ij1b6SNcIk1\\6TNiIi1V6WNnIf1Q6ZNRJd1n5[NVJb1j5]NZJ`1e5`N^J^1b5aNbJ\\1^5cNfJZ1Y5fNjJX1V5gNPKT1P5kNUKQ1j4oN\\Kl0d4TOaKc0c4\\ObK;c4BeK3ZSn3"}, "label": "woman wearing black and white shirt"}]}
{"id": 146561, "image": "COCO_train2014_000000146561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black , white , and red patterned shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a black , white , and red patterned shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.023875, 0.1284065934065934, 0.44509374999999995, 0.9983241758241758], "iscrowd": 0, "area": 53751.6892, "rle": {"size": [364, 640], "counts": "_b53U;;F;E:F;D;F;E:C6J6K6L3L4M4L3N2M4L3N2N3M2O1O2M2O1O2N1O101N100O2O0O2O000O2O001N101O001O00001O001YNaK_L_4W3PLdLQ4W3XLdLh3Y3`LbLa3[3cLcL]3[3fLdLZ3Z3jLdLW3X3nLeLS3Y3PMfLQ3W3RMhLn2W3TMhLm2V3VMhLj2W3XMhLh2W3ZMhLg2V3[MiLe2V3]MiLd2U3^MjLb2U3`MjLa2T3bMiL_2W3bMhL_2V3cMhL^2W3dMhL\\2W3fMgL\\2X3eMgL[2X3n2N1O2O0O2N1O101N1O2N101N1O2N100O2N1O2O0O2N1O2O0O1O2O0O2O001N100O2O001N10001N101O0O101N101O000O2O001O00001O001O00001O001O01O01O010O01O01O010O00010O0010O000[JaN]1`1ZNgNg1Y1XNhNh1X1WNhNj1Y1TNhNl1X1SNhNn1Y1PNhNP2X1oMhNR2X1mMiNS2X1jMiNW2W1hMiNY2W1fMjNZ2W1dMiN]2X1aMiN_2X1_MhNb2Y1\\MhNd2X1[MhNf2Y1XMhNhM]OU4l1QNgNjM^OU4l1oMeNnM@R4l1nMdNPNAR4l1lMcNSNBP4l1jMcNWNAo3n1fMbN[NAo3o1bMbN_N@n3Q2^M`NeN_Om3T2YM_NjN]Om3\\2PMXNRO^Om3b2gLQN]O]Ol3i2]LmMG[Ok3n2SLkM2XOk3X3bKfMc0SOj3k4WLUKi3k4WLVKh3j4WLWKi3h4XLYKg3g4YLYKg3g4XL[Kg3e4YL[Kg3d4ZL\\Kf3d4ZL]Ke3c4ZL^Kf3a4[L`Kd3`4\\L`Kd3`4\\LaKc3_4\\LbKd3^4\\LcKc3\\4^LdKb3\\4]LfKb3Z4^LgKa3Y4_LhK`3X4`LiK_3V4aLlK^3U4aLkK_3U4aLlK^3T4bLmK]3S4bLoK]3Q4cLPL\\3Q4cLPL\\3P4dLQL[3o3eLSLY3m3fLULY3l3fLVLX3l3fLULY3l3fLVLX3l3fLULY3l3fLSL[3o3cLmKa3V4\\LfKh3\\4VLaKm3a4QL[KS4g4jKZKV4h4gKYKY4n5O00001O000000M3N20000000000000O107I4LO1M3N200O100000WLnHX3R7eLRIZ3n6bLVIc01e1l6cMUIa08j1Q7SNVIj1i6SN^Ij1b6SNcIk1\\6TNiIi1V6WNnIf1Q6ZNRJd1n5[NVJb1j5]NZJ`1e5`N^J^1b5aNbJ\\1^5cNfJZ1Y5fNjJX1V5gNPKT1P5kNUKQ1j4oN\\Kl0d4TOaKc0c4\\ObK;c4BeK3ZSn3"}, "label": "a woman in a black , white , and red patterned shirt"}]}
{"id": 146561, "image": "COCO_train2014_000000146561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a green shirt on the front row\"."}], "task": "refering", "answer_template": "The \"a lady in a green shirt on the front row\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 86, 87, 88, 109, 110, 111, 112, 133, 134, 135, 136, 137, 160], "bbox": [0.7634218749999999, 0.15010989010989012, 1.0, 0.4744780219780219], "iscrowd": 0, "area": 6267.326949999998, "rle": {"size": [364, 640], "counts": "cm]5;m:7aEC^9[1N3M2N2M3O2N1O1O1O2N1O1O1O2N1O1N3N1N2O1O2M20000O2O0O100O2O000O101O0000001O00001O03M6J01O000000UOZGQOe8n0]GQOc8n0^GROb8m0_GSOa8k0bGUO^8i0cGWO]8h0dGXO\\8f0gGYOZ8?mGBR8;QHEo77VHHk71[HOa91O01O010O000010O01O00010O001O01O01O0010O00010O01O0010O0001O010O0000100O001O010O1O0010O01O010O1O0010O01O1O010O001O010O1O0010O01O1OO2O0O1O2N100O1O2O0O1O1O2O0O1O100O2N10O0000001O0000dK"}, "label": "a lady in a green shirt on the front row"}]}
{"id": 146561, "image": "COCO_train2014_000000146561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady in the green sleeveless shirt with short hair\"."}], "task": "refering", "answer_template": "The \"the lady in the green sleeveless shirt with short hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 86, 87, 88, 109, 110, 111, 112, 133, 134, 135, 136, 137, 160], "bbox": [0.7634218749999999, 0.15010989010989012, 1.0, 0.4744780219780219], "iscrowd": 0, "area": 6267.326949999998, "rle": {"size": [364, 640], "counts": "cm]5;m:7aEC^9[1N3M2N2M3O2N1O1O1O2N1O1O1O2N1O1N3N1N2O1O2M20000O2O0O100O2O000O101O0000001O00001O03M6J01O000000UOZGQOe8n0]GQOc8n0^GROb8m0_GSOa8k0bGUO^8i0cGWO]8h0dGXO\\8f0gGYOZ8?mGBR8;QHEo77VHHk71[HOa91O01O010O000010O01O00010O001O01O01O0010O00010O01O0010O0001O010O0000100O001O010O1O0010O01O010O1O0010O01O1O010O001O010O1O0010O01O1OO2O0O1O2N100O1O2O0O1O1O2O0O1O100O2N10O0000001O0000dK"}, "label": "the lady in the green sleeveless shirt with short hair"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teady bear on right\"."}], "task": "refering", "answer_template": "The \"teady bear on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 361, 362], "bbox": [0.638765625, 0.5258333333333334, 0.95225, 0.8988750000000001], "iscrowd": 0, "area": 21591.6912, "rle": {"size": [480, 640], "counts": "nPP62m>3M4L3N3L4L3N0O2dC^OW:d0eE@Y:b0bED\\:<aEI\\:9_EK`:7\\ELd:5YENe:4YEMf:4YEMf:5XEMg:4WEMh:5VELi:6UELn9oNnEU13Mn9ROkES15Ln9WOhEo08Ln9YOfEm0;Jm9^OeEh0=Km9BaEe0`0Kl9X1QFiNm9[1PFfNo9]1nEeNP:^1mEcNS:_1jEbNV:U3O00001O0O10001O00001O000O2O00001O0O101O00001N2O1O1O1O2M2O1O1O000000O10000O1000000O1000000O1000000O101O000O1000000O1000O10O1000O10O10O1000O10O1000O01000O01001N1000000]NXFmMi9_1VF`M;k0`9d1QG[No8d1SGZNn8f1SGYNm8f1UGYNk8f1WGYNi8g1XGWNi8h1YGWNh8h1YGWNg8h1[GVNf8i1\\GVNd8j1\\GVNd8i1]GVNd8j1]GUNc8k1]GUNd8i1^GVNe8DkEe1`1fNi8YOTFm1S1jN\\9R1eFmN^9P1bFoNc9m0^FROc9l0^FTOb9l0^FTOc9k0^FSOc9m0]FSOc9m0]FSOc9l0^FTOb9l0^FTOc9k0\\FVOd9i0]FWOc9i0\\FXOe9g0[FZOd9e0]F[Od9d0[F]Of9b0ZF^Of9a0ZF@g9?YFAg9?XFBi9<XFDi9;WFFh9:WFGj97WFIk95TFLm93RFNo90QF1Q:MnE4S:KmE5T:IlE8V:FiE<W:ChE>Z:_OfEb0[:]OeEc0[<1O1O1O100O1O1001O001O00001O001O00001O001O001O00001N1O2M2O1N6K:FhT>"}, "label": "teady bear on right"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown teddy bear with a white stomach\"."}], "task": "refering", "answer_template": "The \"a light brown teddy bear with a white stomach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 361, 362], "bbox": [0.638765625, 0.5258333333333334, 0.95225, 0.8988750000000001], "iscrowd": 0, "area": 21591.6912, "rle": {"size": [480, 640], "counts": "nPP62m>3M4L3N3L4L3N0O2dC^OW:d0eE@Y:b0bED\\:<aEI\\:9_EK`:7\\ELd:5YENe:4YEMf:4YEMf:5XEMg:4WEMh:5VELi:6UELn9oNnEU13Mn9ROkES15Ln9WOhEo08Ln9YOfEm0;Jm9^OeEh0=Km9BaEe0`0Kl9X1QFiNm9[1PFfNo9]1nEeNP:^1mEcNS:_1jEbNV:U3O00001O0O10001O00001O000O2O00001O0O101O00001N2O1O1O1O2M2O1O1O000000O10000O1000000O1000000O1000000O101O000O1000000O1000O10O1000O10O10O1000O10O1000O01000O01001N1000000]NXFmMi9_1VF`M;k0`9d1QG[No8d1SGZNn8f1SGYNm8f1UGYNk8f1WGYNi8g1XGWNi8h1YGWNh8h1YGWNg8h1[GVNf8i1\\GVNd8j1\\GVNd8i1]GVNd8j1]GUNc8k1]GUNd8i1^GVNe8DkEe1`1fNi8YOTFm1S1jN\\9R1eFmN^9P1bFoNc9m0^FROc9l0^FTOb9l0^FTOc9k0^FSOc9m0]FSOc9m0]FSOc9l0^FTOb9l0^FTOc9k0\\FVOd9i0]FWOc9i0\\FXOe9g0[FZOd9e0]F[Od9d0[F]Of9b0ZF^Of9a0ZF@g9?YFAg9?XFBi9<XFDi9;WFFh9:WFGj97WFIk95TFLm93RFNo90QF1Q:MnE4S:KmE5T:IlE8V:FiE<W:ChE>Z:_OfEb0[:]OeEc0[<1O1O1O100O1O1001O001O00001O001O00001O001O001O00001N1O2M2O1N6K:FhT>"}, "label": "a light brown teddy bear with a white stomach"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"koala teddy bear dressed in a jacket\"."}], "task": "refering", "answer_template": "The \"koala teddy bear dressed in a jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 264, 265, 266, 267, 287, 288, 289, 290, 313], "bbox": [0.42471875, 0.4674166666666667, 0.6505625, 0.8134791666666668], "iscrowd": 0, "area": 10364.555249999998, "rle": {"size": [480, 640], "counts": "kgo32l>3L4M3M3N1O2O1O1O1O1O1O0O101O00001O001O0O110O0010O00QB[Oc=S1O1N2N101N2N101N2N0PD]Nc:d1iDPOV;P1eDUO[;l0bDWO];i0_D\\O`;e0[D@d;`0YDCh;k1O1O00000000000O1O1O1O100O001O1O1O100O1O1O100O10000O010O10000O10000O100O10000O01000O11O0000000O1000001N10000O100`MfDZ1Z;dNmDW1S;gNTET1l:kNYEQ1h:lN[ES1e:kN_ES1b:jNbET1_:iNeEU1\\:hNhEV1Y:eNmEY1T:cNQF[1o9aNVF^1g;O2N1O2N1O1O2N1O2N1O1O2N1O2N1O19HT1kN00O1O2N1O1O1M4K4M3[OUCPOo<l0e0N3M4L3M4I`eX3"}, "label": "koala teddy bear dressed in a jacket"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the kola in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the kola in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 264, 265, 266, 267, 287, 288, 289, 290, 313], "bbox": [0.42471875, 0.4674166666666667, 0.6505625, 0.8134791666666668], "iscrowd": 0, "area": 10364.555249999998, "rle": {"size": [480, 640], "counts": "kgo32l>3L4M3M3N1O2O1O1O1O1O1O0O101O00001O001O0O110O0010O00QB[Oc=S1O1N2N101N2N101N2N0PD]Nc:d1iDPOV;P1eDUO[;l0bDWO];i0_D\\O`;e0[D@d;`0YDCh;k1O1O00000000000O1O1O1O100O001O1O1O100O1O1O100O10000O010O10000O10000O100O10000O01000O11O0000000O1000001N10000O100`MfDZ1Z;dNmDW1S;gNTET1l:kNYEQ1h:lN[ES1e:kN_ES1b:jNbET1_:iNeEU1\\:hNhEV1Y:eNmEY1T:cNQF[1o9aNVF^1g;O2N1O2N1O1O2N1O2N1O1O2N1O2N1O19HT1kN00O1O2N1O1O1M4K4M3[OUCPOo<l0e0N3M4L3M4I`eX3"}, "label": "the kola in the right hand picture"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child not wearing glasses , holding a bear with a red shirt\"."}], "task": "refering", "answer_template": "The \"a child not wearing glasses , holding a bear with a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 346, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.027140625, 0.23633333333333334, 0.3830625, 0.9995208333333333], "iscrowd": 0, "area": 46001.814650000015, "rle": {"size": [480, 640], "counts": "cY81m>9G9G6J1O1O2N1O1O2N1O2N2N3M2N2N3M2N2N2M2N3M3M3M3M3M3M2N3M3N2N2N2N3M5K5\\FmLT7V3hHYMj6l2oH]Mj6j2mH]Mn6j2jH[MR7l2eH[MV7j2cH\\MY7j2`H\\M[7j2_HZM]7l2]HVMb7k2\\HWMb7l2[HVMc7l2[HVMc7l2[HUMd7T5O1O1O1O1O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100000000\\NRHULo7j3WHQLi7o3[HkKg7T4]HhKQ4YOgNn4[MeKn3@fNj4_M[KS4M]Nh4k2ZKTMe4l2\\KTMc4m2]KSMc4l2^KTMa4l2`KTM`4k2aKUM]4m2cKSM[4n2fKRMX4o2iKQMU4Q3kKoLS4R3oKmLn3U3SLkLk3W3ULiLi3Y3WLgLg3[3YLeLe3]3\\LbLc3^3^LbLa3_3_LZJ^OP1R4g4`LXJBn0m3k4aLUJIk0f3P5bLTJHl0e3P5dLSJIl0b3S5dLPJKm0`3T5eLoILl0^3V5gLlILn0]3V5gLkINn0Z3Y5gLhI0o0X3Z5hLgI1n0V3\\5iLfI1n0U3]5jLeI2m0T3^5jLRI@6b0Z1S3^5lLQI@7b0Y1Q3^5oLRI_O6a0Z1P3^5QMQI_O6b0Z1n2]5SMSI^O3c0]1k2]5UMRI_O0e0`1g2\\5XMQIc0c1T2\\5ZMPId0c1R2[5QOeJo0Y5TOfJl0Y5UOgJk0X5VOhJi0X5YOgJg0Y5YOgJg0X5ZOhJf0X5ZOhJe0X5]OgJc0Y5]OgJc0X5^OhJb0X5_OgJ`0Y5AgJ?Y5AgJ?X5BhJ>X5CgJ<Y5EgJ;Y5EgJ;X5GgJ9a4_LhJX3g09X4hLPKQ3g06n3VMZKe2g06m3VMZKf2i04k3YMYKf2k01k3ZMXKg2m0Oj3\\MVKh2o0Mi3]MUKh2R1Kg3aMSKg2U1Hg3b1XL^Ng3c1YL^Ne3c1[L]Nd3e1[L[Nc3g1]LYNc3h1\\LYNb3h1^LXNa3i1_LWNc2^LQM]5;UNc2`LPM\\5=UN`2cLPMY5`0TN_2dLQMY5?SN^2gLQMW5a0RN]2hLQMW5b0RNZ2jLRMV5c0PNZ2lLQMU5e0PNW2mLSMT5f0PNU2nLTMS5f0PNT2nLUMS5g0oMS2PMUMR5g0oMR2QMUMQ5i0oMP2RMVMo4j0PNn1TMVMl4l0PNm1VMUMl4m0oMl1XMUMi4o0PNj1X3VNiLh1Y3WNhLg1Y3YNgLf1[3YNfLf1[3YNfLe1\\3ZNeLc1]3]NcLb1_3]NbLa1_3_NcL]1_3cNbLX1c3gN_LU1c3kN_LQ1c3oN_Ln0b3RO_Lk0c3UO_Lh0c3WO_Le0c3[O_Lb0b3^O`L>b3B`L:b3FaL5b3JaL1b3NaLNa3_Oolh5"}, "label": "a child not wearing glasses , holding a bear with a red shirt"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy on the left\"."}], "task": "refering", "answer_template": "The \"the boy on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 346, 347, 348, 349, 350, 351, 352, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.027140625, 0.23633333333333334, 0.3830625, 0.9995208333333333], "iscrowd": 0, "area": 46001.814650000015, "rle": {"size": [480, 640], "counts": "cY81m>9G9G6J1O1O2N1O1O2N1O2N2N3M2N2N3M2N2N2M2N3M3M3M3M3M3M2N3M3N2N2N2N3M5K5\\FmLT7V3hHYMj6l2oH]Mj6j2mH]Mn6j2jH[MR7l2eH[MV7j2cH\\MY7j2`H\\M[7j2_HZM]7l2]HVMb7k2\\HWMb7l2[HVMc7l2[HVMc7l2[HUMd7T5O1O1O1O1O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100000000\\NRHULo7j3WHQLi7o3[HkKg7T4]HhKQ4YOgNn4[MeKn3@fNj4_M[KS4M]Nh4k2ZKTMe4l2\\KTMc4m2]KSMc4l2^KTMa4l2`KTM`4k2aKUM]4m2cKSM[4n2fKRMX4o2iKQMU4Q3kKoLS4R3oKmLn3U3SLkLk3W3ULiLi3Y3WLgLg3[3YLeLe3]3\\LbLc3^3^LbLa3_3_LZJ^OP1R4g4`LXJBn0m3k4aLUJIk0f3P5bLTJHl0e3P5dLSJIl0b3S5dLPJKm0`3T5eLoILl0^3V5gLlILn0]3V5gLkINn0Z3Y5gLhI0o0X3Z5hLgI1n0V3\\5iLfI1n0U3]5jLeI2m0T3^5jLRI@6b0Z1S3^5lLQI@7b0Y1Q3^5oLRI_O6a0Z1P3^5QMQI_O6b0Z1n2]5SMSI^O3c0]1k2]5UMRI_O0e0`1g2\\5XMQIc0c1T2\\5ZMPId0c1R2[5QOeJo0Y5TOfJl0Y5UOgJk0X5VOhJi0X5YOgJg0Y5YOgJg0X5ZOhJf0X5ZOhJe0X5]OgJc0Y5]OgJc0X5^OhJb0X5_OgJ`0Y5AgJ?Y5AgJ?X5BhJ>X5CgJ<Y5EgJ;Y5EgJ;X5GgJ9a4_LhJX3g09X4hLPKQ3g06n3VMZKe2g06m3VMZKf2i04k3YMYKf2k01k3ZMXKg2m0Oj3\\MVKh2o0Mi3]MUKh2R1Kg3aMSKg2U1Hg3b1XL^Ng3c1YL^Ne3c1[L]Nd3e1[L[Nc3g1]LYNc3h1\\LYNb3h1^LXNa3i1_LWNc2^LQM]5;UNc2`LPM\\5=UN`2cLPMY5`0TN_2dLQMY5?SN^2gLQMW5a0RN]2hLQMW5b0RNZ2jLRMV5c0PNZ2lLQMU5e0PNW2mLSMT5f0PNU2nLTMS5f0PNT2nLUMS5g0oMS2PMUMR5g0oMR2QMUMQ5i0oMP2RMVMo4j0PNn1TMVMl4l0PNm1VMUMl4m0oMl1XMUMi4o0PNj1X3VNiLh1Y3WNhLg1Y3YNgLf1[3YNfLf1[3YNfLe1\\3ZNeLc1]3]NcLb1_3]NbLa1_3_NcL]1_3cNbLX1c3gN_LU1c3kN_LQ1c3oN_Ln0b3RO_Lk0c3UO_Lh0c3WO_Le0c3[O_Lb0b3^O`L>b3B`L:b3FaL5b3JaL1b3NaLNa3_Oolh5"}, "label": "the boy on the left"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear with clothing\"."}], "task": "refering", "answer_template": "The \"a teddy bear with clothing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 327, 328, 329, 330, 351, 352, 353], "bbox": [0.198875, 0.5236041666666666, 0.453375, 0.9460625], "iscrowd": 0, "area": 18113.26515, "rle": {"size": [480, 640], "counts": "Wlk1>b>0O100O10000O100O100O10000O100O10000O101N1gB[OV<g0hC]OT<e0jC_OR<b0mCBo;`0PDCl;?RDEj;=TDDk;=TDEk;<SDEl;=RDDm;=RDEl;=SDCl;?RDCl;?RDBm;?RDBm;`0QDBm;`0QDAn;`0RD@m;b0QD@m;b0QD_Oo;j1N1O1O2N1jNmL^FT3a9oL\\FS3b9PM[FQ3d9RMYFP3e9RMYFo2f9TMWFm2h9VMUFl2i9WMUFi2j9ZMSFg2l9\\MQFf2m9]MPFd2o9_MnEc2P:W1O1O2N1O1O2N1O2N1O1O2N1O2O0O1O2O0O1O2O0O2N100O2N100O2N101N1O101N1O101N101O00001O00001O001O00001O001O00O1M3M3M3L4M3M3M3M3L4M3G9@`0@`0A?B>O100000000O1000000lNSDUOm;i0VDUOk;i0XDVOh;h0[DWOe;g0^DXOb;f0aDYO_;e0dDYO];f0eDYO[;e0hDZOX;d0kD[OU;c0nD\\OR;b0QE\\OP;b0SE]Om:a0VE^Oj:`0YE_Og:`0[E_Oe:?^E_Oc:?`E@`:>cEA_:;dEDP;FSE9k<N2O1O1O1O1O1N2OZgS5"}, "label": "a teddy bear with clothing"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear wearing a red jacket\"."}], "task": "refering", "answer_template": "The \"a teddy bear wearing a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 327, 328, 329, 330, 351, 352, 353], "bbox": [0.198875, 0.5236041666666666, 0.453375, 0.9460625], "iscrowd": 0, "area": 18113.26515, "rle": {"size": [480, 640], "counts": "Wlk1>b>0O100O10000O100O100O10000O100O10000O101N1gB[OV<g0hC]OT<e0jC_OR<b0mCBo;`0PDCl;?RDEj;=TDDk;=TDEk;<SDEl;=RDDm;=RDEl;=SDCl;?RDCl;?RDBm;?RDBm;`0QDBm;`0QDAn;`0RD@m;b0QD@m;b0QD_Oo;j1N1O1O2N1jNmL^FT3a9oL\\FS3b9PM[FQ3d9RMYFP3e9RMYFo2f9TMWFm2h9VMUFl2i9WMUFi2j9ZMSFg2l9\\MQFf2m9]MPFd2o9_MnEc2P:W1O1O2N1O1O2N1O2N1O1O2N1O2O0O1O2O0O1O2O0O2N100O2N100O2N101N1O101N1O101N101O00001O00001O001O00001O001O00O1M3M3M3L4M3M3M3M3L4M3G9@`0@`0A?B>O100000000O1000000lNSDUOm;i0VDUOk;i0XDVOh;h0[DWOe;g0^DXOb;f0aDYO_;e0dDYO];f0eDYO[;e0hDZOX;d0kD[OU;c0nD\\OR;b0QE\\OP;b0SE]Om:a0VE^Oj:`0YE_Og:`0[E_Oe:?^E_Oc:?`E@`:>cEA_:;dEDP;FSE9k<N2O1O1O1O1O1N2OZgS5"}, "label": "a teddy bear wearing a red jacket"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy holding two bears\"."}], "task": "refering", "answer_template": "The \"boy holding two bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 217, 220, 221, 222, 223, 226, 240, 243, 244, 245, 249, 263, 264, 267, 268, 286, 287, 290, 291, 294, 295, 296, 313, 314, 317, 318, 319, 336, 337, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.436515625, 0.15054166666666668, 0.995109375, 1.0], "iscrowd": 0, "area": 44345.969450000004, "rle": {"size": [480, 640], "counts": "[SS42\\>d0^Ob0K5K5L4K5K5K5L4K5K4M2M2N31O3M3M3M3M3L4M3M3M3M3aNoBi0T=nNWCm0b=K6K5K5K4L5K^T;MdkD1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O00100O010000O11O0O2O001O1O0O2O001SCAg4I;f0]J4T5XO?e0kIe0b5iNb0\\2YOgMg0Z2UOiMk0Y2POjMP1W2kNmMU1U2dNPN\\1Q2^NTNa1o1XNVNh1k1SNYNm1i1lM\\NT2e1fM`NZ2b1_McN`2_1ZMfNf2\\1SMiNm2X1mLmNQ1lMhMY3P1POR1PNkMQ3m0TOS1SNmMk2j0VOS1WNPNe2g0YOS1[NSN]2d0]OS1_NVNV2`0@T1cNYNn1=DT1fN]Nh18GV1iN^Nb16JV1mNaN[12MW1QOdNS13MX1SOcNR11M\\1TO`NP12M^1VO]No02L`1YOZNm04Kb1[OWNl03Kf1[OUNk03Kh1]ORNj03Jj1_OQNh02Km1_OnMi02IP2@lMh02IR2AjMh01HT2CiMf00IW2CgMf0OHZ2]8O10bJSNiMm1V2bN\\M^1c2dN\\M[1d2hNZMX1e2kN[JAi0d1k4mNZJAj0b1k4POXJ_Om0`1j4TOVJ_Oo0]1j4WOTJ]OR1\\1i4YOSJ]OS1Z1i4\\OPJ]OV1W1i4^OoI\\OX1V1h4AmI[OZ1T1h4CkI[O]1R1V4iMVJm14ZO_1P1U4mMUJk15YOa1o0R4RNUJi15XOc1m0P4WNUJg14WOg1k0m3\\NUJd15WOh1i0l3l0YJ\\Nk1h0i3P1YJ[Nm1e0g3i2YLWMg3j2XLVMh3j2XLVMi3j2VLVMj3k2ULUMk3k2ULUMk3l2TLTMl3l2TLTMl3m2SLSMm3m2SLSMn3m2QLSMo3m2QLSMo3n2PLRMP4n2PLRMP4o2oKQMQ4o2oKQMR4o2mKQMS4o2mKQMS4P3lKPMT4Q3kKoLU4Q3kKoLU4R3jKnLV4R3jKnLV4S3iKmLV4T3jKnLT4S3kKnLT4R3lKoLS4R3lKoLR4Q3oKPMP4P3PLRMn3n2RLSMm3l2TLUM_1bMUOZ5\\OUM]1cMVOW5^OWMY1fMWOS5@YMU1hMXOP5CYMS1iMYOm4E[Mo0lMZOi4G\\Ml0nM\\Of4H]Mj0PN[Oc4L^Mf0RN]O`4M`Mc0SN^O\\40bM?VN_OX42cM<XN@V44cM:YNAS46eM6ZNDQ46gM2[NGn37hM0[NHn38hMK^NMj38iMEcN2c3:kM_OfN7_3:mMXOiN>Z3:nMROnNc0T3;oMmNQOh0P3;PNgNUOn0j2<RN`NZOS1d2=SN[N]OX1`2=S1CmN=T1BlN>T1BkN?U1@lN`0U1_OkNa0U1_OkNa0U1_OkNa0V1^OiNc0W1]OiNc0W1]OiNc0X1\\OhNd0X1\\OhNd0X1\\OgNe0Z1ZOfNf0Z1ZOfNf0Z1ZOfNf0[1YOeNg0[1YOdNh0\\1XOdNh0]1WOcNi0]1WOcNi0]1WOcNi0^1VOaNk0_1UOaNk0_1UOaNk0`1TOmLPOlMl1W5TOlLTOjMh1Z5TOkLXOgMe1_5ROhL^OfM`1b5ROgLAeM]1d5ROeLFdMX1h5QOcLKbMT1k5QOULQO\\Mo0a0o0n5QOULVOVMl0e0n0Q6POSL=lMc0Q6QORL<mMc0R6ROoK;oMc0U6POkK=PNc0X6oNfK>QNd0\\6lNbK`0RNd0^6lN^K`0TNd0a6jNZKb0UNd0d6iNUKc0WNd0g6gNQKe0WNe0j6eNnJf0XNe0m6dNiJf0[Nf0o6eNbJe0_Nf0R7XOiI2VOe0S7WOfI5WOd0V7TOcI8WOd0Y7POaI<VOd0\\7mN^I?VOd0^7kN[Ib0WOc0a7gNYIf0VOc0d7dNVIi0VOc0g7aNRIm0WOb0i7_NPIo0WOb0l7[NnHS1VOb0o7XNkHU1WOc0Q8UNgHY1XOb0S8RNfH[1XOc0U8oMcH^1XOc0i9\\OXFd0g9]OYFc0g9\\OZFd0f9\\OZFd0e9\\O\\Fd0d9\\O\\Fd0d9[O]Fe0c9[O]Fe0b9[O_Fe0a9[O_Fe0a9ZO`Ff0`9ZO`Ff0b9WO_Fi0c9UO]Fk0d9SO]Fm0e9QO[Fo0g9nNZFR1g9mNYFS1i9jNXFV1j9hNVFX1k9fNUF[1m9cNSF]1n9aNSF_1o9_NPFb1R:[NoEe1R:ZNmEg1U:VNlEj1V:TNjEl1W:RNiEo1Y:oMgEQ2[:lMfET2[:kMdEV2c;000O10000O10000O11O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2K6J[n0"}, "label": "boy holding two bears"}]}
{"id": 556162, "image": "COCO_train2014_000000556162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy with glasses holding two teddy bears\"."}], "task": "refering", "answer_template": "The \"a young boy with glasses holding two teddy bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 217, 220, 221, 222, 223, 226, 240, 243, 244, 245, 249, 263, 264, 267, 268, 286, 287, 290, 291, 294, 295, 296, 313, 314, 317, 318, 319, 336, 337, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.436515625, 0.15054166666666668, 0.995109375, 1.0], "iscrowd": 0, "area": 44345.969450000004, "rle": {"size": [480, 640], "counts": "[SS42\\>d0^Ob0K5K5L4K5K5K5L4K5K4M2M2N31O3M3M3M3M3L4M3M3M3M3aNoBi0T=nNWCm0b=K6K5K5K4L5K^T;MdkD1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O00100O010000O11O0O2O001O1O0O2O001SCAg4I;f0]J4T5XO?e0kIe0b5iNb0\\2YOgMg0Z2UOiMk0Y2POjMP1W2kNmMU1U2dNPN\\1Q2^NTNa1o1XNVNh1k1SNYNm1i1lM\\NT2e1fM`NZ2b1_McN`2_1ZMfNf2\\1SMiNm2X1mLmNQ1lMhMY3P1POR1PNkMQ3m0TOS1SNmMk2j0VOS1WNPNe2g0YOS1[NSN]2d0]OS1_NVNV2`0@T1cNYNn1=DT1fN]Nh18GV1iN^Nb16JV1mNaN[12MW1QOdNS13MX1SOcNR11M\\1TO`NP12M^1VO]No02L`1YOZNm04Kb1[OWNl03Kf1[OUNk03Kh1]ORNj03Jj1_OQNh02Km1_OnMi02IP2@lMh02IR2AjMh01HT2CiMf00IW2CgMf0OHZ2]8O10bJSNiMm1V2bN\\M^1c2dN\\M[1d2hNZMX1e2kN[JAi0d1k4mNZJAj0b1k4POXJ_Om0`1j4TOVJ_Oo0]1j4WOTJ]OR1\\1i4YOSJ]OS1Z1i4\\OPJ]OV1W1i4^OoI\\OX1V1h4AmI[OZ1T1h4CkI[O]1R1V4iMVJm14ZO_1P1U4mMUJk15YOa1o0R4RNUJi15XOc1m0P4WNUJg14WOg1k0m3\\NUJd15WOh1i0l3l0YJ\\Nk1h0i3P1YJ[Nm1e0g3i2YLWMg3j2XLVMh3j2XLVMi3j2VLVMj3k2ULUMk3k2ULUMk3l2TLTMl3l2TLTMl3m2SLSMm3m2SLSMn3m2QLSMo3m2QLSMo3n2PLRMP4n2PLRMP4o2oKQMQ4o2oKQMR4o2mKQMS4o2mKQMS4P3lKPMT4Q3kKoLU4Q3kKoLU4R3jKnLV4R3jKnLV4S3iKmLV4T3jKnLT4S3kKnLT4R3lKoLS4R3lKoLR4Q3oKPMP4P3PLRMn3n2RLSMm3l2TLUM_1bMUOZ5\\OUM]1cMVOW5^OWMY1fMWOS5@YMU1hMXOP5CYMS1iMYOm4E[Mo0lMZOi4G\\Ml0nM\\Of4H]Mj0PN[Oc4L^Mf0RN]O`4M`Mc0SN^O\\40bM?VN_OX42cM<XN@V44cM:YNAS46eM6ZNDQ46gM2[NGn37hM0[NHn38hMK^NMj38iMEcN2c3:kM_OfN7_3:mMXOiN>Z3:nMROnNc0T3;oMmNQOh0P3;PNgNUOn0j2<RN`NZOS1d2=SN[N]OX1`2=S1CmN=T1BlN>T1BkN?U1@lN`0U1_OkNa0U1_OkNa0U1_OkNa0V1^OiNc0W1]OiNc0W1]OiNc0X1\\OhNd0X1\\OhNd0X1\\OgNe0Z1ZOfNf0Z1ZOfNf0Z1ZOfNf0[1YOeNg0[1YOdNh0\\1XOdNh0]1WOcNi0]1WOcNi0]1WOcNi0^1VOaNk0_1UOaNk0_1UOaNk0`1TOmLPOlMl1W5TOlLTOjMh1Z5TOkLXOgMe1_5ROhL^OfM`1b5ROgLAeM]1d5ROeLFdMX1h5QOcLKbMT1k5QOULQO\\Mo0a0o0n5QOULVOVMl0e0n0Q6POSL=lMc0Q6QORL<mMc0R6ROoK;oMc0U6POkK=PNc0X6oNfK>QNd0\\6lNbK`0RNd0^6lN^K`0TNd0a6jNZKb0UNd0d6iNUKc0WNd0g6gNQKe0WNe0j6eNnJf0XNe0m6dNiJf0[Nf0o6eNbJe0_Nf0R7XOiI2VOe0S7WOfI5WOd0V7TOcI8WOd0Y7POaI<VOd0\\7mN^I?VOd0^7kN[Ib0WOc0a7gNYIf0VOc0d7dNVIi0VOc0g7aNRIm0WOb0i7_NPIo0WOb0l7[NnHS1VOb0o7XNkHU1WOc0Q8UNgHY1XOb0S8RNfH[1XOc0U8oMcH^1XOc0i9\\OXFd0g9]OYFc0g9\\OZFd0f9\\OZFd0e9\\O\\Fd0d9\\O\\Fd0d9[O]Fe0c9[O]Fe0b9[O_Fe0a9[O_Fe0a9ZO`Ff0`9ZO`Ff0b9WO_Fi0c9UO]Fk0d9SO]Fm0e9QO[Fo0g9nNZFR1g9mNYFS1i9jNXFV1j9hNVFX1k9fNUF[1m9cNSF]1n9aNSF_1o9_NPFb1R:[NoEe1R:ZNmEg1U:VNlEj1V:TNjEl1W:RNiEo1Y:oMgEQ2[:lMfET2[:kMdEV2c;000O10000O10000O11O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2K6J[n0"}, "label": "a young boy with glasses holding two teddy bears"}]}
{"id": 253064, "image": "COCO_train2014_000000253064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the white shirt in the background right behind the pizza\"."}], "task": "refering", "answer_template": "The \"the man with the white shirt in the background right behind the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113], "bbox": [0.5682343750000001, 0.0, 0.9670468750000001, 0.30112412177985953], "iscrowd": 0, "area": 23431.138049999994, "rle": {"size": [427, 640], "counts": "Xig4<o<=C=C=C>B<DO1000000000000000O101O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O0000000000000O2O00000000000000001O000000000000001O00000000000000001O000000000000001O000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000006Jb0^O>B00O1000000000000000000000000000O10000000O100000000000000000000000000000000000000000001O1O1O1O1O1O001O1O1O1O1O10O01O1O1O1O1O1O1O001O1O1O2N1O2N1O2N1O2N1O2O1N1M4M2N3L3N3L3N3L4M2M4M2N3L3N3L3N3L5L6I7J6J6I7J\\e8"}, "label": "the man with the white shirt in the background right behind the pizza"}]}
{"id": 253064, "image": "COCO_train2014_000000253064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white onject behind pizza\"."}], "task": "refering", "answer_template": "The \"white onject behind pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113], "bbox": [0.5682343750000001, 0.0, 0.9670468750000001, 0.30112412177985953], "iscrowd": 0, "area": 23431.138049999994, "rle": {"size": [427, 640], "counts": "Xig4<o<=C=C=C>B<DO1000000000000000O101O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O0000000000000O2O00000000000000001O000000000000001O00000000000000001O000000000000001O000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000006Jb0^O>B00O1000000000000000000000000000O10000000O100000000000000000000000000000000000000000001O1O1O1O1O1O001O1O1O1O1O10O01O1O1O1O1O1O1O001O1O1O2N1O2N1O2N1O2N1O2O1N1M4M2N3L3N3L3N3L4M2M4M2N3L3N3L3N3L5L6I7J6J6I7J\\e8"}, "label": "white onject behind pizza"}]}
{"id": 253064, "image": "COCO_train2014_000000253064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"at the top left of the frame , it appears to be a person ' s blurry arm\"."}], "task": "refering", "answer_template": "The \"at the top left of the frame , it appears to be a person ' s blurry arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 24, 25, 26, 27, 28, 29, 30, 31, 32, 48, 49, 51, 52, 53, 54, 55], "bbox": [0.00465625, 0.0, 0.411765625, 0.16353629976580797], "iscrowd": 0, "area": 12466.68425, "rle": {"size": [427, 640], "counts": "\\e11Z=1O1O001O1O1O1O001O1O1O1OO12N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N1O1O1O001O1O001O1O001O1O001O1O001O1O001O1O0010O000000000000000001O00000000000000000000000000000000000000000000000000001O000000O2O0O1O100O1O100O100O1O100O2N100O10000001O0000000010O000001O000000001O00000010O00000001O0000001O0000010O00000000001O000000000O101O0000000000001O0000000O101O0000000000000000000001O000000001O000001O01O1O1O1O1O001O1O1O1O1O100O001O1O1O1O001O100O1O001O001O00010O001O00001O000010O01O00001O00001N101O00001O00001O001O00001O00001O002L5J5K6J6Jegl4"}, "label": "at the top left of the frame , it appears to be a person ' s blurry arm"}]}
{"id": 253064, "image": "COCO_train2014_000000253064.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a white shirt behind a table with white pizza on ti\"."}], "task": "refering", "answer_template": "The \"a person in a white shirt behind a table with white pizza on ti\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 24, 25, 26, 27, 28, 29, 30, 31, 32, 48, 49, 51, 52, 53, 54, 55], "bbox": [0.00465625, 0.0, 0.411765625, 0.16353629976580797], "iscrowd": 0, "area": 12466.68425, "rle": {"size": [427, 640], "counts": "\\e11Z=1O1O001O1O1O1O001O1O1O1OO12N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N1O1O1O001O1O001O1O001O1O001O1O001O1O001O1O0010O000000000000000001O00000000000000000000000000000000000000000000000000001O000000O2O0O1O100O1O100O100O1O100O2N100O10000001O0000000010O000001O000000001O00000010O00000001O0000001O0000010O00000000001O000000000O101O0000000000001O0000000O101O0000000000000000000001O000000001O000001O01O1O1O1O1O001O1O1O1O1O100O001O1O1O1O001O100O1O001O001O00010O001O00001O000010O01O00001O00001N101O00001O00001O001O00001O00001O002L5J5K6J6Jegl4"}, "label": "a person in a white shirt behind a table with white pizza on ti"}]}
{"id": 302216, "image": "COCO_train2014_000000302216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slices of red apple on a plate , with the peel still on\"."}], "task": "refering", "answer_template": "The \"slices of red apple on a plate , with the peel still on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 134, 135, 136, 137, 146, 147, 148, 149, 150, 159, 160, 161, 162, 163, 173, 174, 175, 176], "bbox": [0.2707629427792916, 0.5962409638554217, 0.6052043596730246, 0.9201927710843374], "iscrowd": 0, "area": 10254.30445, "rle": {"size": [415, 367], "counts": "m^X11h<6K6I7I7K5N2N1O2O1N2N2N2N1O2N2O1N2N2N1O2M3N2N2N2N1O2M3N2N2N2N1N3N2N2N2I6J6O0O2O1N2N2O1N2O1N2N2O1O0O2O1O1N2O1O0O101O0O2O000010O1O100O100O2N2O1N2N2O1N3N2M3M3N2M2N3N2M201N1O2N1O001O010O00001O00001O001O00001O00001O001O1O1O1O1O1O1N3N7I6I8G8I8H8G:G9Ff]j1"}, "label": "slices of red apple on a plate , with the peel still on"}]}
{"id": 302216, "image": "COCO_train2014_000000302216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cut apples to the furthest left under the cabbage\"."}], "task": "refering", "answer_template": "The \"the cut apples to the furthest left under the cabbage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 118, 119, 120, 121, 130, 131, 132, 133, 134, 144, 145, 146, 147, 157, 158, 159], "bbox": [0.06861035422343324, 0.5861445783132531, 0.3888010899182561, 0.8625542168674698], "iscrowd": 0, "area": 7494.40235, "rle": {"size": [415, 367], "counts": "[]:9S<d0M4M3M3L4M2N3M3L4M2N3L4M3M3O0O2N2N2N1O2N2O1N2N1O2N000001O01O000002N2N2N2N3N1N2N2N3M2N2N2N3N1N2N2NO1N2O1N2N3N1N2O1N2O1O1O1N3N1O2M3N2N1O2M3N2O1N2ON1N3L4M2M4M2M4M2N3M2O2O001N101O0O2O001N101O0O2O0O2O001N2O001N101O0O2M2M4M2N3L3N3L3N3L3N>BUaj2"}, "label": "the cut apples to the furthest left under the cabbage"}]}
{"id": 302216, "image": "COCO_train2014_000000302216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a carrot that is between another carrot and some celery\"."}], "task": "refering", "answer_template": "The \"a carrot that is between another carrot and some celery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 83, 84, 96, 97, 109, 110, 111, 123, 124, 137], "bbox": [0.39193460490463217, 0.34681927710843374, 0.5667302452316076, 0.7027951807228915], "iscrowd": 0, "area": 4868.778200000001, "rle": {"size": [415, 367], "counts": "``j15g<6J7I6I7J7L3M4L3M3M4L3M4L3M3M4L3M4L3M4L3M3M4L3M4L3M3M4L2N1O1O1O1O1O1O11O2N2N2N3M4L3M4L3M3L5L3M4L3M4L3M3M5K4L5K4L5K4L5K4L4L5F9CiUP2"}, "label": "a carrot that is between another carrot and some celery"}]}
{"id": 302216, "image": "COCO_train2014_000000302216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shortest carrot stick\"."}], "task": "refering", "answer_template": "The \"the shortest carrot stick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 83, 84, 96, 97, 109, 110, 111, 123, 124, 137], "bbox": [0.39193460490463217, 0.34681927710843374, 0.5667302452316076, 0.7027951807228915], "iscrowd": 0, "area": 4868.778200000001, "rle": {"size": [415, 367], "counts": "``j15g<6J7I6I7J7L3M4L3M3M4L3M4L3M3M4L3M4L3M4L3M3M4L3M4L3M3M4L2N1O1O1O1O1O1O11O2N2N2N3M4L3M4L3M3L5L3M4L3M4L3M3M5K4L5K4L5K4L5K4L4L5F9CiUP2"}, "label": "the shortest carrot stick"}]}
{"id": 302216, "image": "COCO_train2014_000000302216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrot touching the green lettuce\"."}], "task": "refering", "answer_template": "The \"carrot touching the green lettuce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 55, 56, 68, 69, 70, 82, 83, 95, 96, 109, 110], "bbox": [0.24517711171662127, 0.23481927710843375, 0.474741144414169, 0.6067710843373494], "iscrowd": 0, "area": 4598.990599999999, "rle": {"size": [415, 367], "counts": "_bT19f<3L4L4L4M2M2N2N3N1N2O1O1O2N1O1O1O1O2N2N2N3M3M2N3M2M4M3M2N2N2O1N2N3M2N2O11O0000000010O000001O0000002N6J7I2N1O001N2O1O2N1O1N3N1O1O2N0O2O0O2M2N3N1N3M2O2M2N2O2M2O2M2N3N1Iio]2"}, "label": "carrot touching the green lettuce"}]}
{"id": 302216, "image": "COCO_train2014_000000302216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two orange carrots between celery and lettuce\"."}], "task": "refering", "answer_template": "The \"two orange carrots between celery and lettuce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 55, 56, 68, 69, 70, 82, 83, 95, 96, 109, 110], "bbox": [0.24517711171662127, 0.23481927710843375, 0.474741144414169, 0.6067710843373494], "iscrowd": 0, "area": 4598.990599999999, "rle": {"size": [415, 367], "counts": "_bT19f<3L4L4L4M2M2N2N3N1N2O1O1O2N1O1O1O1O2N2N2N3M3M2N3M2M4M3M2N2N2O1N2N3M2N2O11O0000000010O000001O0000002N6J7I2N1O001N2O1O2N1O1N3N1O1O2N0O2O0O2M2N3N1N3M2O2M2N2O2M2O2M2N3N1Iio]2"}, "label": "two orange carrots between celery and lettuce"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pastry with peach filling\"."}], "task": "refering", "answer_template": "The \"the pastry with peach filling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349], "bbox": [0.47071895424836596, 0.19820261437908496, 0.977483660130719, 0.6959477124183007], "iscrowd": 0, "area": 72581.28180000001, "rle": {"size": [612, 612], "counts": "`\\\\51Qc03M2O2M2N3`N;]_OG``04f_OOV`0KQ@7l?DV@c0e?DQ@c0l?]1L4L3N3N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2O0O2O1N101N1O2O1N101N2O0O2N2O0O2O1N101N2N101N2O0O2O0O2O1N101O1N101O1N101N2O002M2O1N3N1O1N3N1N2O2N1N3N1N2O2N1N2O2M2O1O2N1O2N1O1N3N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N1O2M2O1O2N1O1O2N1N2O2N1O1O1O2M2O1O2N1O1O2N1N2O00001O0000000000000O2O00000000000000001O000O100000000000000000000O10000000000000000000000O1001O000000000000000000001O00000000000000000O2O0000000000000000001N1000000000000000001O000O101O0000000O2O0000001O0O10001O000O10001O000O101O00000O2O0000001N100000001N10001N2O0O2O0O2O1N1O2O0O2O1N101N101N2O0O2O0O2N2O0O2O0O2O9Fj2WMi2VMWQ8"}, "label": "the pastry with peach filling"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut with white frosting , and yellow - colored fruit filling in the hole\"."}], "task": "refering", "answer_template": "The \"a donut with white frosting , and yellow - colored fruit filling in the hole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 106, 107, 108, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349], "bbox": [0.47071895424836596, 0.19820261437908496, 0.977483660130719, 0.6959477124183007], "iscrowd": 0, "area": 72581.28180000001, "rle": {"size": [612, 612], "counts": "`\\\\51Qc03M2O2M2N3`N;]_OG``04f_OOV`0KQ@7l?DV@c0e?DQ@c0l?]1L4L3N3N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2O0O2O1N101N1O2O1N101N2O0O2N2O0O2O1N101N2N101N2O0O2O0O2O1N101O1N101O1N101N2O002M2O1N3N1O1N3N1N2O2N1N3N1N2O2N1N2O2M2O1O2N1O2N1O1N3N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N1O2M2O1O2N1O1O2N1N2O2N1O1O1O2M2O1O2N1O1O2N1N2O00001O0000000000000O2O00000000000000001O000O100000000000000000000O10000000000000000000000O1001O000000000000000000001O00000000000000000O2O0000000000000000001N1000000000000000001O000O101O0000000O2O0000001O0O10001O000O10001O000O101O00000O2O0000001N100000001N10001N2O0O2O0O2O1N1O2O0O2O1N101N101N2O0O2O0O2N2O0O2O0O2O9Fj2WMi2VMWQ8"}, "label": "a donut with white frosting , and yellow - colored fruit filling in the hole"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate donut with chocolate sprinkles\"."}], "task": "refering", "answer_template": "The \"a chocolate donut with chocolate sprinkles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 207, 223, 224, 225, 226, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476], "bbox": [0.022369281045751633, 0.4342156862745098, 0.7315359477124183, 0.9890359477124182], "iscrowd": 0, "area": 119095.49814999998, "rle": {"size": [612, 612], "counts": "nf8T1oa0Y1hNX1gNY1gN=C2O1N2N2N2O1N2N2N2O0O2N2N2O1N2N2N2O1N2N1O2O1N2N2O1N2N2N2O1N2N1O2O1N2N2N2O1N2N2N2O1N1O2N2O1N2N2N2O1N2N2N101N2N2N2O1N2N2O1N2N2N101N2N2N2O1N2N2O1N2O1N101N2O1N2O1O1N2O1N2O0O2O1N2O0O10000O100O100O100O10000O100O100O100O01000O100O100O100O10000O100O100O100O100O10000O100O100O1000O10O1000000O100000000O1000000O100000000O1000000O100000000O0100000O100000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O1000000000001O001O001O00001O001O001O001O00001O001O001O000O2O001O001O00001O001O001O001O00001O001O001O00001O001O001O0O2O00001O001O001O00001O001O001O001O00001O001O001O00001O0O2O001O00001O001O001O001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O2M3N2N1O2N2M3N2N1O2L4K5K5L4K4L5L4K5K5K5J5K6J6J6J6J5K6J6I7J6J6J5K6J6J6J6J6J5J7J6J6J6J5K6J6J6JbbQ3"}, "label": "a chocolate donut with chocolate sprinkles"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chocolat covered doughnut with chocolate sprinkles\"."}], "task": "refering", "answer_template": "The \"chocolat covered doughnut with chocolate sprinkles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 207, 223, 224, 225, 226, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476], "bbox": [0.022369281045751633, 0.4342156862745098, 0.7315359477124183, 0.9890359477124182], "iscrowd": 0, "area": 119095.49814999998, "rle": {"size": [612, 612], "counts": "nf8T1oa0Y1hNX1gNY1gN=C2O1N2N2N2O1N2N2N2O0O2N2N2O1N2N2N2O1N2N1O2O1N2N2O1N2N2N2O1N2N1O2O1N2N2N2O1N2N2N2O1N1O2N2O1N2N2N2O1N2N2N101N2N2N2O1N2N2O1N2N2N101N2N2N2O1N2N2O1N2O1N101N2O1N2O1O1N2O1N2O0O2O1N2O0O10000O100O100O100O10000O100O100O100O01000O100O100O100O10000O100O100O100O100O10000O100O100O1000O10O1000000O100000000O1000000O100000000O1000000O100000000O0100000O100000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000O1000000000001O001O001O00001O001O001O001O00001O001O001O000O2O001O001O00001O001O001O001O00001O001O001O00001O001O001O0O2O00001O001O001O00001O001O001O001O00001O001O001O00001O0O2O001O00001O001O001O001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O2M3N2N1O2N2M3N2N1O2L4K5K5L4K4L5L4K5K5K5J5K6J6J6J6J5K6J6I7J6J6J5K6J6J6J6J6J5J7J6J6J6J5K6J6J6JbbQ3"}, "label": "chocolat covered doughnut with chocolate sprinkles"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut with nuts\"."}], "task": "refering", "answer_template": "The \"a donut with nuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 242, 243, 244, 264, 265], "bbox": [0.02366013071895425, 0.1820424836601307, 0.31130718954248365, 0.586454248366013], "iscrowd": 0, "area": 25793.929550000004, "rle": {"size": [612, 612], "counts": "[_8d7`;O00010O00002O0O1O1O101N1O1O101N1O100O1O2N100O1O1O2O0O2N1O101N1O1O2O0O1O2N100O2N100O2N1O101N1O1O2O0O1O1O100O1O1O100O1O1O101N1O101N1O2N101N1O2N101N1O2N101N100O2O0O2O001N101N101N10O0100O01000O010O10O02O1N2O1N2O2M2O1O1N3N3L3N3L3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2N2M2O1N3N1N2O2M2O1N3N1N3N1O1O2N1N2O2N1O1N3N1O2N1N2O2N1O4L5J7J5K1O1N3N1O1O2N1N3N1O1O2M2O1O1O2Mank7"}, "label": "a donut with nuts"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut with the brown icing and nuts on top\"."}], "task": "refering", "answer_template": "The \"the donut with the brown icing and nuts on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 242, 243, 244, 264, 265], "bbox": [0.02366013071895425, 0.1820424836601307, 0.31130718954248365, 0.586454248366013], "iscrowd": 0, "area": 25793.929550000004, "rle": {"size": [612, 612], "counts": "[_8d7`;O00010O00002O0O1O1O101N1O1O101N1O100O1O2N100O1O1O2O0O2N1O101N1O1O2O0O1O2N100O2N100O2N1O101N1O1O2O0O1O1O100O1O1O100O1O1O101N1O101N1O2N101N1O2N101N1O2N101N100O2O0O2O001N101N101N10O0100O01000O010O10O02O1N2O1N2O2M2O1O1N3N3L3N3L3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2N2M2O1N3N1N2O2M2O1N3N1N3N1O1O2N1N2O2N1O1N3N1O2N1N2O2N1O4L5J7J5K1O1N3N1O1O2N1N3N1O1O2M2O1O1O2Mank7"}, "label": "the donut with the brown icing and nuts on top"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a doughnut with white frosting and red jelly\"."}], "task": "refering", "answer_template": "The \"a doughnut with white frosting and red jelly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 114, 115, 116, 117, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 141, 142, 143, 159, 160, 161, 162, 163, 164, 182, 183, 184, 185, 186], "bbox": [0.11058823529411765, 0.021094771241830065, 0.6327287581699347, 0.401421568627451], "iscrowd": 0, "area": 54089.50765, "rle": {"size": [612, 612], "counts": "hgX1;bb0;J5J4M2IPOX^OS1ca0PO[^OS1ba0oN]^OS1`a0;K4M4L3N3M2N3N1N2N3D;N3M2N3N10001N101N101O0O2O0O101O0O2N1O0O2N2O1N1O2N2O1N1100001O2N1O1O1O1O1O2N2N2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O1O2N1O1O1O2O1N2N2N00001O00000000001O00000000001O00000000001O00000000001O000000000000001O00000000000000001O0000000000000000001O00000000000000000000000000000000000000O1000000000000O1000000000000O100000000O10000O100O100O10000O10000O10000O10000O100ZOmC[JS<b5VDWJk;h5XDVJh;j5ZDSJg;l5\\DQJe;n5^DPJb;o5j0O1O101N1O1O1N2O1O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3L4N2O100O100O2O1N2O1N2O1N2O1N2O2M2O1N2O1N2O1L4L4M3L4L4N3O0O1O1O1O2N1O1O1O1O2N1O2M3N1N3N2M2O2M2O2M7J5J7H>UOc\\V4"}, "label": "a doughnut with white frosting and red jelly"}]}
{"id": 121994, "image": "COCO_train2014_000000121994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a donut with red jelly in the middle , next to a donut with orange jelly\"."}], "task": "refering", "answer_template": "The \"a donut with red jelly in the middle , next to a donut with orange jelly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 114, 115, 116, 117, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 141, 142, 143, 159, 160, 161, 162, 163, 164, 182, 183, 184, 185, 186], "bbox": [0.11058823529411765, 0.021094771241830065, 0.6327287581699347, 0.401421568627451], "iscrowd": 0, "area": 54089.50765, "rle": {"size": [612, 612], "counts": "hgX1;bb0;J5J4M2IPOX^OS1ca0PO[^OS1ba0oN]^OS1`a0;K4M4L3N3M2N3N1N2N3D;N3M2N3N10001N101N101O0O2O0O101O0O2N1O0O2N2O1N1O2N2O1N1100001O2N1O1O1O1O1O2N2N2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O1O2N1O1O1O2O1N2N2N00001O00000000001O00000000001O00000000001O00000000001O000000000000001O00000000000000001O0000000000000000001O00000000000000000000000000000000000000O1000000000000O1000000000000O100000000O10000O100O100O10000O10000O10000O10000O100ZOmC[JS<b5VDWJk;h5XDVJh;j5ZDSJg;l5\\DQJe;n5^DPJb;o5j0O1O101N1O1O1N2O1O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3L4N2O100O100O2O1N2O1N2O1N2O1N2O2M2O1N2O1N2O1L4L4M3L4L4N3O0O1O1O1O2N1O1O1O1O2N1O2M3N1N3N2M2O2M2O2M7J5J7H>UOc\\V4"}, "label": "a donut with red jelly in the middle , next to a donut with orange jelly"}]}
{"id": 416907, "image": "COCO_train2014_000000416907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a short man behind an asian girkl\"."}], "task": "refering", "answer_template": "The \"a short man behind an asian girkl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 95, 96, 97, 113, 114, 115, 131, 132, 133, 149, 150, 151, 152, 167, 168, 169, 170, 186, 187, 204, 205, 222, 223], "bbox": [0.29836, 0.30549333333333334, 0.45330000000000004, 0.9868], "iscrowd": 0, "area": 12871.106300000003, "rle": {"size": [375, 500], "counts": "Yff1;c0e0V9\\OfFh0Y9YObFl0]9VO]Fn0c9c000000000O2Ol0TO9G7I8H>B9G8H3SKYKb2k4WM^KZ2o4`MYK[2k4^M]K]2f4\\M`Ka2c4ZMaKe2b4UMcKj2_4RMcKn2^4oLdKQ3^4kLdKU3^4gLdKY3]60000000000000000001O001O1O1O2N3M2SIPLo5T4jIPLV6V4aInK]6g4O1O1O1O1O1O1O0UOjJUKW5h4lJWKU5f4nJXKS5f4PKYKQ5c4SK[Kn4c4VKZKk4c4YK[Ki4c0bJl2h0_Lg4d0eJh2o0ZL^4l0hJd2b6ZMcIa2`6\\MdI_2a6\\MeI\\2b6`McIY2R8J6I8H6J7H8G8I=Ae0\\ORgS3"}, "label": "a short man behind an asian girkl"}]}
{"id": 416907, "image": "COCO_train2014_000000416907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black t - shirt in the bathroom\"."}], "task": "refering", "answer_template": "The \"a man in a black t - shirt in the bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 95, 96, 97, 113, 114, 115, 131, 132, 133, 149, 150, 151, 152, 167, 168, 169, 170, 186, 187, 204, 205, 222, 223], "bbox": [0.29836, 0.30549333333333334, 0.45330000000000004, 0.9868], "iscrowd": 0, "area": 12871.106300000003, "rle": {"size": [375, 500], "counts": "Yff1;c0e0V9\\OfFh0Y9YObFl0]9VO]Fn0c9c000000000O2Ol0TO9G7I8H>B9G8H3SKYKb2k4WM^KZ2o4`MYK[2k4^M]K]2f4\\M`Ka2c4ZMaKe2b4UMcKj2_4RMcKn2^4oLdKQ3^4kLdKU3^4gLdKY3]60000000000000000001O001O1O1O2N3M2SIPLo5T4jIPLV6V4aInK]6g4O1O1O1O1O1O1O0UOjJUKW5h4lJWKU5f4nJXKS5f4PKYKQ5c4SK[Kn4c4VKZKk4c4YK[Ki4c0bJl2h0_Lg4d0eJh2o0ZL^4l0hJd2b6ZMcIa2`6\\MdI_2a6\\MeI\\2b6`McIY2R8J6I8H6J7H8G8I=Ae0\\ORgS3"}, "label": "a man in a black t - shirt in the bathroom"}]}
{"id": 416907, "image": "COCO_train2014_000000416907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in a bikini top\"."}], "task": "refering", "answer_template": "The \"the girl in a bikini top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 98, 99, 100, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 153, 154, 155, 171, 172, 173, 174, 189, 190, 191, 207, 208, 209], "bbox": [0.44695999999999997, 0.30304, 0.68736, 0.9012533333333334], "iscrowd": 0, "area": 15305.850099999996, "rle": {"size": [375, 500], "counts": "UWb24a;4L3N2M4M2N2N3M2N2N2N00000000000O2N2N1O2O1N1O2O0M4K4010O10O010O0VImNR3S1iLBf2>VMGj29QMLn24nL1R3NkL6T3KhL9X3GcL>]3B]Ld0c3]OSLk0m3VOoKn0Q4XOgKj0c3gMXKe1m0f0k3eMXKh1h0e0o3dMYKk1b0c0T4cMZKm1;d0Z4`MZKT21`0e4\\MZKa5e4aJZK^5f4gJVKY5j4kJRKU5m4k0O1O100O1O1O010O1000000000O100O101N100O2N100O101N100O2O0O101N1N2N3N1O1O1cJhJ[4Y5dKmJU4T5jKmJV4T5iKmJV4T5iKlJW4V5fKlJY4V5eKjJZ4Z5bKhJY4_5dKaJY4e5cK]JZ4h5cKXJ[4d6N1O2N1O2iM`H4a7LcHO_70iHGX79nH@T7?oH]OR7d0PIXOR7g0RITOo6l0TIQOm6n0VInNk6S1XIiNP7o0SImNX7h0R2E:F:F:Ehmh1"}, "label": "the girl in a bikini top"}]}
{"id": 416907, "image": "COCO_train2014_000000416907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the bikini top brushing her teeth down to the bottom of her skirt\"."}], "task": "refering", "answer_template": "The \"the girl in the bikini top brushing her teeth down to the bottom of her skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 98, 99, 100, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 153, 154, 155, 171, 172, 173, 174, 189, 190, 191, 207, 208, 209], "bbox": [0.44695999999999997, 0.30304, 0.68736, 0.9012533333333334], "iscrowd": 0, "area": 15305.850099999996, "rle": {"size": [375, 500], "counts": "UWb24a;4L3N2M4M2N2N3M2N2N2N00000000000O2N2N1O2O1N1O2O0M4K4010O10O010O0VImNR3S1iLBf2>VMGj29QMLn24nL1R3NkL6T3KhL9X3GcL>]3B]Ld0c3]OSLk0m3VOoKn0Q4XOgKj0c3gMXKe1m0f0k3eMXKh1h0e0o3dMYKk1b0c0T4cMZKm1;d0Z4`MZKT21`0e4\\MZKa5e4aJZK^5f4gJVKY5j4kJRKU5m4k0O1O100O1O1O010O1000000000O100O101N100O2N100O101N100O2O0O101N1N2N3N1O1O1cJhJ[4Y5dKmJU4T5jKmJV4T5iKmJV4T5iKlJW4V5fKlJY4V5eKjJZ4Z5bKhJY4_5dKaJY4e5cK]JZ4h5cKXJ[4d6N1O2N1O2iM`H4a7LcHO_70iHGX79nH@T7?oH]OR7d0PIXOR7g0RITOo6l0TIQOm6n0VInNk6S1XIiNP7o0SImNX7h0R2E:F:F:Ehmh1"}, "label": "the girl in the bikini top brushing her teeth down to the bottom of her skirt"}]}
{"id": 48267, "image": "COCO_train2014_000000048267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table and cake the priest is cutting\"."}], "task": "refering", "answer_template": "The \"a table and cake the priest is cutting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.228125, 0.6383098591549297, 0.9147343750000001, 0.9997887323943663], "iscrowd": 0, "area": 41222.29055, "rle": {"size": [426, 640], "counts": "fTm11X=3N2M2O1N100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O10000O100O100O100O100O100O100O100O100O100O10001N100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O10000O100O100O10000O100001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O000O1O2N100O1O2N1O1O2N1O1O1O2N1O101N1O1O1O2N1O1O2N1O1O1O2O0O1O2N1O1O1O2N1O1O2N1O100O2N1O1O2N1O1O1O2N1O1O2N100O1O2N1O1O2N1O1O1O2N1O101N2N1O2N2N2N1O5K5K5K6J5L5J5KZaf0"}, "label": "a table and cake the priest is cutting"}]}
{"id": 48267, "image": "COCO_train2014_000000048267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table with a red tablecloth\"."}], "task": "refering", "answer_template": "The \"the table with a red tablecloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.228125, 0.6383098591549297, 0.9147343750000001, 0.9997887323943663], "iscrowd": 0, "area": 41222.29055, "rle": {"size": [426, 640], "counts": "fTm11X=3N2M2O1N100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O10000O100O100O100O100O100O100O100O100O100O10001N100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O10000O100O100O10000O100001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O000O1O2N100O1O2N1O1O2N1O1O1O2N1O101N1O1O1O2N1O1O2N1O1O1O2O0O1O2N1O1O1O2N1O1O2N1O100O2N1O1O2N1O1O1O2N1O1O2N100O1O2N1O1O2N1O1O1O2N1O101N2N1O2N2N2N1O5K5K5K6J5L5J5KZaf0"}, "label": "the table with a red tablecloth"}]}
{"id": 121997, "image": "COCO_train2014_000000121997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cop car with the door open\"."}], "task": "refering", "answer_template": "The \"a cop car with the door open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 114, 124, 125, 126, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 316, 317, 318], "bbox": [0.39584375, 0.26238611713665944, 0.999359375, 0.8389587852494577], "iscrowd": 0, "area": 74481.16775, "rle": {"size": [461, 640], "counts": "YTb3T2o;`0@?A?A`0@?nNS1J5K6J5K1OO101O00000000001O0000000000001O000000000O2O0000000000001O00000000001O000000000O2O0000000000001O00000000001O000000000O101O00000000001O0000000000001O001O1O1O2M2O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N12N9G9G9GO2N1O1O1O1N3N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1N2O2N1O1OO1O1N2O1O1O1O1N2O1O1O1N2O1O001N2O001O0O2O1O001N101O001N2O001O001N2O001O0O2O1O001N101O1O1N2O001O1N2O1O1O001N2O1O1O1N101O1O1N2O1O001N2O1O1O1N101O1O2N3M4M3L3N3L4M0O100O100O100O1O100O100O100O10_HWJ^6h5cIRLc4n3]KkMj2T2WMEP1;PO`710O0100O100O100O100O1O100O100Ol0UOX1gNX1iNX1gNX1iNX1gNe0\\O0O100000O1000000000001O00000000000N2O1N2O1O1N2O1N2O1N2O2M2O1O10000000000000000O10000000001O0000000000000000000000000O1000001O00000000000000000000000000000O101O000000000000000000000000000000001N10000000000000000000000001O2N2N101N2N2N2N2N]HfIU7X6jH\\Le4b3ZKSOT2k0kMl7O1N_:"}, "label": "a cop car with the door open"}]}
{"id": 121997, "image": "COCO_train2014_000000121997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a police car with an open door\"."}], "task": "refering", "answer_template": "The \"a police car with an open door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 114, 124, 125, 126, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 316, 317, 318], "bbox": [0.39584375, 0.26238611713665944, 0.999359375, 0.8389587852494577], "iscrowd": 0, "area": 74481.16775, "rle": {"size": [461, 640], "counts": "YTb3T2o;`0@?A?A`0@?nNS1J5K6J5K1OO101O00000000001O0000000000001O000000000O2O0000000000001O00000000001O000000000O2O0000000000001O00000000001O000000000O101O00000000001O0000000000001O001O1O1O2M2O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N12N9G9G9GO2N1O1O1O1N3N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1N2O2N1O1OO1O1N2O1O1O1O1N2O1O1O1N2O1O001N2O001O0O2O1O001N101O001N2O001O001N2O001O0O2O1O001N101O1O1N2O001O1N2O1O1O001N2O1O1O1N101O1O1N2O1O001N2O1O1O1N101O1O2N3M4M3L3N3L4M0O100O100O100O1O100O100O100O10_HWJ^6h5cIRLc4n3]KkMj2T2WMEP1;PO`710O0100O100O100O100O1O100O100Ol0UOX1gNX1iNX1gNX1iNX1gNe0\\O0O100000O1000000000001O00000000000N2O1N2O1O1N2O1N2O1N2O2M2O1O10000000000000000O10000000001O0000000000000000000000000O1000001O00000000000000000000000000000O101O000000000000000000000000000000001N10000000000000000000000001O2N2N101N2N2N2N2N]HfIU7X6jH\\Le4b3ZKSOT2k0kMl7O1N_:"}, "label": "a police car with an open door"}]}
{"id": 121997, "image": "COCO_train2014_000000121997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front vehicle\"."}], "task": "refering", "answer_template": "The \"the front vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 214, 215, 237, 238, 260, 261, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309], "bbox": [0.308375, 0.5192624728850326, 0.5607031250000001, 0.8391973969631237], "iscrowd": 0, "area": 6309.452599999998, "rle": {"size": [461, 640], "counts": "UPi22W><C<D=C<E8I6K6K4K6J8I1N2N2O1N3M2O1N2N2O2M2N2N2O2M2N2O1N2N3N1N2N2O2M2N2O1N2N3N0O03Ll0TOl0VOj0WOi0VOj0WOS`4Ml_K4L4000001O01O0001O00<D000000000000000000000000000O10000000000000000000000000000000000000B_B0`=1jBEV=;>000000000000O1000000000000000000O01000000000000000O1000000000000000000O0100000000000000000O10000000000000000O010HYUn3"}, "label": "the front vehicle"}]}
{"id": 359565, "image": "COCO_train2014_000000359565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small shaded space between the top two white manuals on the right\"."}], "task": "refering", "answer_template": "The \"the small shaded space between the top two white manuals on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 243, 244, 257, 258, 271, 272, 285], "bbox": [0.7210106382978724, 0.7530625, 1.0, 0.9627812499999999], "iscrowd": 0, "area": 1426.1680499999943, "rle": {"size": [640, 376], "counts": "ZkY51mc04L3M3M30010O000000010O000000010O000000010O00000001O01O000001O01O000001O01O000001O01O000001O01O000001O0001O0001O0001O0001O0001O0001O0001O0001O0001O0001O000001O01O000001O01O000001O01O000001O01O000001N1O1O1O1N3[]O"}, "label": "the small shaded space between the top two white manuals on the right"}]}
{"id": 359565, "image": "COCO_train2014_000000359565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white hardback book stacked on top of 2 other hardbacks\"."}], "task": "refering", "answer_template": "The \"a white hardback book stacked on top of 2 other hardbacks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 244, 245, 246, 258, 259, 272, 285], "bbox": [0.7218351063829788, 0.743703125, 1.0, 0.932765625], "iscrowd": 0, "area": 6268.21555, "rle": {"size": [640, 376], "counts": "nR[51oc01O1O1O1N2O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1N2O1O1O2N1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O00O100O1OUA"}, "label": "a white hardback book stacked on top of 2 other hardbacks"}]}
{"id": 359565, "image": "COCO_train2014_000000359565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white notepad on top of pink notebooks\"."}], "task": "refering", "answer_template": "The \"a black and white notepad on top of pink notebooks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 244, 245, 246, 258, 259, 272, 285], "bbox": [0.7218351063829788, 0.743703125, 1.0, 0.932765625], "iscrowd": 0, "area": 6268.21555, "rle": {"size": [640, 376], "counts": "nR[51oc01O1O1O1N2O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1N2O1O1O2N1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O00O100O1OUA"}, "label": "a black and white notepad on top of pink notebooks"}]}
{"id": 162963, "image": "COCO_train2014_000000162963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby wearing a pink dress and sitting on a table\"."}], "task": "refering", "answer_template": "The \"a baby wearing a pink dress and sitting on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 126, 127, 128, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 240, 241, 242], "bbox": [0.35617187499999997, 0.23820093457943925, 0.674765625, 0.7191121495327103], "iscrowd": 0, "area": 20552.09295, "rle": {"size": [428, 640], "counts": "c`o23U=5K6M2N3M1O1O100O1O1O1O1O010O1O1O1O1O100O001O1O100OL4I7N22N1O100O1O1O1O1O10000O103L4M3M3L4M3L4M3M3L4M3M3L3N3L5L3M4K5L1N3N1O1N3N1N3N1O1N3N1O1N2O0O2O000O2O0O101O0jN\\LcHe3X7dLcH^3V7kLdHX3[7lL`HW3]7nL^HT3a7PMZHS3c7\\101N1O101N1O2N101N100O2N101O0O2O0O2O0O101O0O2OO0O2O1O001O1O0O2O1O1O1O1O1N101O1O1O1O2M2O2N1O2N1O2N1O2N1O2N1O1O2O0O2N1O2N1O3M3M3M3M4L3M4M2M3M4L3Mk0UO4L2N3M2N2N3M2O2M2[OoDlNS;U1TEaNo:^1`001O01O01O0010O01O1O2O0O2N1O2O1N1O1O10O01N2M2O2N2N1O2N2N1N3N1O2N2Naff2"}, "label": "a baby wearing a pink dress and sitting on a table"}]}
{"id": 162963, "image": "COCO_train2014_000000162963.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl sitting on the table\"."}], "task": "refering", "answer_template": "The \"a little girl sitting on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 126, 127, 128, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 240, 241, 242], "bbox": [0.35617187499999997, 0.23820093457943925, 0.674765625, 0.7191121495327103], "iscrowd": 0, "area": 20552.09295, "rle": {"size": [428, 640], "counts": "c`o23U=5K6M2N3M1O1O100O1O1O1O1O010O1O1O1O1O100O001O1O100OL4I7N22N1O100O1O1O1O1O10000O103L4M3M3L4M3L4M3M3L4M3M3L3N3L5L3M4K5L1N3N1O1N3N1N3N1O1N3N1O1N2O0O2O000O2O0O101O0jN\\LcHe3X7dLcH^3V7kLdHX3[7lL`HW3]7nL^HT3a7PMZHS3c7\\101N1O101N1O2N101N100O2N101O0O2O0O2O0O101O0O2OO0O2O1O001O1O0O2O1O1O1O1O1N101O1O1O1O2M2O2N1O2N1O2N1O2N1O2N1O1O2O0O2N1O2N1O3M3M3M3M4L3M4M2M3M4L3Mk0UO4L2N3M2N2N3M2O2M2[OoDlNS;U1TEaNo:^1`001O01O01O0010O01O1O2O0O2N1O2O1N1O1O10O01N2M2O2N2N1O2N2N1N3N1O2N2Naff2"}, "label": "a little girl sitting on the table"}]}
{"id": 351384, "image": "COCO_train2014_000000351384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bus hood\"."}], "task": "refering", "answer_template": "The \"a yellow bus hood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.703734375, 0.6227578947368421, 1.0, 0.9831157894736843], "iscrowd": 0, "area": 18455.98390000001, "rle": {"size": [475, 640], "counts": "`Ua61j>7I8G8I1O0O10000O10000O10000O10000O1000000O10000O10000O10000O10000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000000000O10000L4I7J6J6I7J6J6I7J6J6I7J6J6N3N1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O100O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O10000hF"}, "label": "a yellow bus hood"}]}
{"id": 351384, "image": "COCO_train2014_000000351384.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the frontmost part and headlight of a yellow school bus\"."}], "task": "refering", "answer_template": "The \"the frontmost part and headlight of a yellow school bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [251, 252, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.703734375, 0.6227578947368421, 1.0, 0.9831157894736843], "iscrowd": 0, "area": 18455.98390000001, "rle": {"size": [475, 640], "counts": "`Ua61j>7I8G8I1O0O10000O10000O10000O10000O1000000O10000O10000O10000O10000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000000000O10000L4I7J6J6I7J6J6I7J6J6I7J6J6N3N1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1N2O1O1O100O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O10000hF"}, "label": "the frontmost part and headlight of a yellow school bus"}]}
{"id": 253087, "image": "COCO_train2014_000000253087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bride in her wedding dress and vail cutting the cake with her groom\"."}], "task": "refering", "answer_template": "The \"bride in her wedding dress and vail cutting the cake with her groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 173, 174, 175, 196, 197, 219, 221, 222, 243, 244, 245, 268, 269, 291, 292], "bbox": [0.51875, 0.1344496487119438, 0.707671875, 0.8468149882903983], "iscrowd": 0, "area": 16955.52865, "rle": {"size": [427, 640], "counts": "VcZ43W=6I8H7J6I7J6I7J6I7J6I7I:UF]M]8i3N100001O00YNVLnIj3Q6\\LkIc3U6bLfI^3Y6hLcIW3]6nL^IR3b6SMYIn2e6XMWIg2i6_MQIa2n6eMmH[2S7jMiHU2V7QNeHo1[7VNaHi1^7R21O000010O0001O1O1ORJ[JX4d5hK]JX4b5iK^JW4a5iKaJU4_5kKbJU4^5kKbJU4]5kKeJT4Z5mKfJS4Z5lKgJT4X5mKhJS4X5lKjJR4V5nKkJR4U5nKkJS4S5mKoJR4P5oKPKQ4P5nKQKR4n4oKSKP4m4oKTKQ4k4oKVKR4i4nKWKR4h4nKZKQ4f4oKZKQ4e4oK\\KQ4c4oK_KQ4`4oK`KQ4_4oKbKR4]4nKdKQ4[4oKfKR4Y4nKgKR4X4nKjKR4T4nKmKR4S4nKmKS4Q4mKPLS4P4mKQLS4m3mKTLS4j3oKVLR4h3nKZLQ4e3oK\\LR4b3oK^LV4[3kKgLZ4R3gKnL_4k2aKVMe4c2\\K]Mj4[2WKgMn4R2RKPNo4m1RKTNP5h1PKZNQ5b1QK^NR5^1nJdNS5Y1nJhNT5T1lJnNV5n0jJTOW5h0kJYOW5c0iJ_OX5>iJET5:lJNn4OSK8g4GZK?`4^ObKf0\\4WOeKk0Z4ROhKP1W4nNjKT1V4iNkKY1T4dNnK^1R4_NoKc1P4ZNRLh1m3VNTLl1l3RNULo1j3PN_Li1`3VNiLd1V3ZNSM_1l2`N]MY1e2cNcMW1^2fNkMS1V2jNYNi0h1TOUO1l0L]6L3M3M3Maj]2"}, "label": "bride in her wedding dress and vail cutting the cake with her groom"}]}
{"id": 253087, "image": "COCO_train2014_000000253087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a wedding dress cutting a cake\"."}], "task": "refering", "answer_template": "The \"a woman in a wedding dress cutting a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 81, 82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 173, 174, 175, 196, 197, 219, 221, 222, 243, 244, 245, 268, 269, 291, 292], "bbox": [0.51875, 0.1344496487119438, 0.707671875, 0.8468149882903983], "iscrowd": 0, "area": 16955.52865, "rle": {"size": [427, 640], "counts": "VcZ43W=6I8H7J6I7J6I7J6I7J6I7I:UF]M]8i3N100001O00YNVLnIj3Q6\\LkIc3U6bLfI^3Y6hLcIW3]6nL^IR3b6SMYIn2e6XMWIg2i6_MQIa2n6eMmH[2S7jMiHU2V7QNeHo1[7VNaHi1^7R21O000010O0001O1O1ORJ[JX4d5hK]JX4b5iK^JW4a5iKaJU4_5kKbJU4^5kKbJU4]5kKeJT4Z5mKfJS4Z5lKgJT4X5mKhJS4X5lKjJR4V5nKkJR4U5nKkJS4S5mKoJR4P5oKPKQ4P5nKQKR4n4oKSKP4m4oKTKQ4k4oKVKR4i4nKWKR4h4nKZKQ4f4oKZKQ4e4oK\\KQ4c4oK_KQ4`4oK`KQ4_4oKbKR4]4nKdKQ4[4oKfKR4Y4nKgKR4X4nKjKR4T4nKmKR4S4nKmKS4Q4mKPLS4P4mKQLS4m3mKTLS4j3oKVLR4h3nKZLQ4e3oK\\LR4b3oK^LV4[3kKgLZ4R3gKnL_4k2aKVMe4c2\\K]Mj4[2WKgMn4R2RKPNo4m1RKTNP5h1PKZNQ5b1QK^NR5^1nJdNS5Y1nJhNT5T1lJnNV5n0jJTOW5h0kJYOW5c0iJ_OX5>iJET5:lJNn4OSK8g4GZK?`4^ObKf0\\4WOeKk0Z4ROhKP1W4nNjKT1V4iNkKY1T4dNnK^1R4_NoKc1P4ZNRLh1m3VNTLl1l3RNULo1j3PN_Li1`3VNiLd1V3ZNSM_1l2`N]MY1e2cNcMW1^2fNkMS1V2jNYNi0h1TOUO1l0L]6L3M3M3Maj]2"}, "label": "a woman in a wedding dress cutting a cake"}]}
{"id": 253087, "image": "COCO_train2014_000000253087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the groom\"."}], "task": "refering", "answer_template": "The \"the groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 61, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 343], "bbox": [0.509765625, 0.1078688524590164, 0.9490624999999999, 1.0], "iscrowd": 0, "area": 53224.3366, "rle": {"size": [427, 640], "counts": "nVX43W=9G3M2N2N1O2N2N2N2N100O1N2O001O1O00010O01O010O001O0010O01O001O001O010O010O01O1O001O001O001O00001O001O001O00010O0010O01O010O001O10O01O001O010O1O001O010O1O001O010O001O1O010O001O10O01N101N101N2N1N3L3O2O0000001N100000001[LWNgKi1Y4[NcKe1]4`N]Ka1c4dNoJf1Q5]NkJc1U5dNcJ]1]5SORJn0o5WOkIj0T6]OdIj0V6@`If0[6ZOdIl0V6VOiIo0Q6SOmIV1j5lNTJe1\\5\\NbJR2P5oMPK[2e4gMYKd2\\4^MbKX4h2jKWMY4f2hKXM[4e2hKXM\\4d2gKYM\\4e2fKYM]4c2X3N2N3M2M4M5K5J7J5K5K0000000O1000000000000000000O1000000001O001O001O001O001O001O001O1O1O1O1O001O1O6J7I=C3M1O2mEbMU9_2dFhM[9T301O1O001O1O001O1O1O1O1O1O001O1O10O01O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O000000001O001O001O00001O001O001O001O00002N2N2N1O2N2N1O2N2^NnJgJT5k4eKhJ\\4g4^2A?J5K6J6J6J8H9G9F:Eb0_Oc0\\O>A>ZOVa="}, "label": "the groom"}]}
{"id": 253087, "image": "COCO_train2014_000000253087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the groom\"."}], "task": "refering", "answer_template": "The \"the groom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 61, 62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342, 343], "bbox": [0.509765625, 0.1078688524590164, 0.9490624999999999, 1.0], "iscrowd": 0, "area": 53224.3366, "rle": {"size": [427, 640], "counts": "nVX43W=9G3M2N2N1O2N2N2N2N100O1N2O001O1O00010O01O010O001O0010O01O001O001O010O010O01O1O001O001O001O00001O001O001O00010O0010O01O010O001O10O01O001O010O1O001O010O1O001O010O001O1O010O001O10O01N101N101N2N1N3L3O2O0000001N100000001[LWNgKi1Y4[NcKe1]4`N]Ka1c4dNoJf1Q5]NkJc1U5dNcJ]1]5SORJn0o5WOkIj0T6]OdIj0V6@`If0[6ZOdIl0V6VOiIo0Q6SOmIV1j5lNTJe1\\5\\NbJR2P5oMPK[2e4gMYKd2\\4^MbKX4h2jKWMY4f2hKXM[4e2hKXM\\4d2gKYM\\4e2fKYM]4c2X3N2N3M2M4M5K5J7J5K5K0000000O1000000000000000000O1000000001O001O001O001O001O001O001O1O1O1O1O001O1O6J7I=C3M1O2mEbMU9_2dFhM[9T301O1O001O1O001O1O1O1O1O1O001O1O10O01O001O00001O001O001O001O001O00001O001O001O001O001O00001O001O000000001O001O001O00001O001O001O001O00002N2N2N1O2N2N1O2N2^NnJgJT5k4eKhJ\\4g4^2A?J5K6J6J6J8H9G9F:Eb0_Oc0\\O>A>ZOVa="}, "label": "the groom"}]}
{"id": 56480, "image": "COCO_train2014_000000056480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the knife to the right of the bowl\"."}], "task": "refering", "answer_template": "The \"the knife to the right of the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 249, 250, 251, 273, 274, 296, 297, 298, 320, 321, 344], "bbox": [0.831609375, 0.5328125, 1.0, 0.9032916666666667], "iscrowd": 0, "area": 5607.538499999995, "rle": {"size": [480, 640], "counts": "^di71o>2M2O2M3M2O2M2O2M3M2O2M3N1N3M3N1N3N1N3N2M2N3N2M2O2O10O010O01O10O01O10O01O10O01O010O10O01O10O01O10O01O010O10O01O10O01O10O01O10O01O010O10O000O2O0O1O2O0O1O2O0O1O200O100000000O10001O00O010O100O1O010O1O100O010O1O100OA`0ROTE"}, "label": "the knife to the right of the bowl"}]}
{"id": 56480, "image": "COCO_train2014_000000056480.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark silver knife\"."}], "task": "refering", "answer_template": "The \"a dark silver knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 249, 250, 251, 273, 274, 296, 297, 298, 320, 321, 344], "bbox": [0.831609375, 0.5328125, 1.0, 0.9032916666666667], "iscrowd": 0, "area": 5607.538499999995, "rle": {"size": [480, 640], "counts": "^di71o>2M2O2M3M2O2M2O2M3M2O2M3N1N3M3N1N3N1N3N2M2N3N2M2O2O10O010O01O10O01O10O01O10O01O010O10O01O10O01O10O01O010O10O01O10O01O10O01O10O01O010O10O000O2O0O1O2O0O1O2O0O1O200O100000000O10001O00O010O100O1O010O1O100O010O1O100OA`0ROTE"}, "label": "a dark silver knife"}]}
{"id": 449701, "image": "COCO_train2014_000000449701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rear most bird\"."}], "task": "refering", "answer_template": "The \"rear most bird\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 113, 114, 115, 133, 134, 135, 152, 153, 154, 155, 173, 174, 175, 196], "bbox": [0.2369639794168096, 0.29462242562929064, 0.48842195540308747, 0.6064530892448512], "iscrowd": 0, "area": 4873.10135, "rle": {"size": [437, 583], "counts": "bSk11d=2M3N2M3N2M3N2M1O10O0100O100O100O1O100O010O100O100O100O1O010O100O100O101N3N1N3M100O10O0100O10O01O100O010O100O010O10<C4L4M2M4M3L2O0O^OkDfNU;V1REgNm:U1jD_N?8g:X1`EeN`:Y1eEdNZ:[1Q1N1O2O001N101O0O101O0O2M2N3N1N3M2O2M2N2O2M2N3N1NL5D>B>K50O10000O0100000O1000O2O001O001N2O001O10O101N100\\OQDAP<=e0O1O2N1O1O2N1O1O2N1O1O001O000000000000000L4JPQo3"}, "label": "rear most bird"}]}
{"id": 449701, "image": "COCO_train2014_000000449701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger bird in the background\"."}], "task": "refering", "answer_template": "The \"the larger bird in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 113, 114, 115, 133, 134, 135, 152, 153, 154, 155, 173, 174, 175, 196], "bbox": [0.2369639794168096, 0.29462242562929064, 0.48842195540308747, 0.6064530892448512], "iscrowd": 0, "area": 4873.10135, "rle": {"size": [437, 583], "counts": "bSk11d=2M3N2M3N2M3N2M1O10O0100O100O100O1O100O010O100O100O100O1O010O100O100O101N3N1N3M100O10O0100O10O01O100O010O100O010O10<C4L4M2M4M3L2O0O^OkDfNU;V1REgNm:U1jD_N?8g:X1`EeN`:Y1eEdNZ:[1Q1N1O2O001N101O0O101O0O2M2N3N1N3M2O2M2N2O2M2N3N1NL5D>B>K50O10000O0100000O1000O2O001O001N2O001O10O101N100\\OQDAP<=e0O1O2N1O1O2N1O1O2N1O1O001O000000000000000L4JPQo3"}, "label": "the larger bird in the background"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza being cut by fork and knife\"."}], "task": "refering", "answer_template": "The \"pizza being cut by fork and knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 78, 89, 90, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146], "bbox": [0.2816458333333333, 0.192578125, 0.6501875, 0.37459375], "iscrowd": 0, "area": 14777.88625, "rle": {"size": [640, 480], "counts": "g`d22mc08I8H8H8H7I8H8G8I8H8H7I2N00O00100O100O101O1O1O001O1O1O1O1O1O1O1O1OTOgMR@X2n?jMQ@U2o?mMP@R2o?RNn_On1R`0TNm_Ok1S`0WNl_Oh1T`0ZNk_Oe1U`0]Nj_Ob1V`0`Ni_O_1W`0cNh_O\\1X`0fNg_OY1Y`0iNf_OV1Z`0lNe_OS1[`0oNd_OP1\\`0ROc_Om0]`0UOa_Ok0_`0WO`_Oh0_`0[O`_Od0``0^O__Oa0a`0A^_O>b`0D]_O;c`0G\\_O8e`0IZ_O6f`0LY_O3g`0c1000000000000000000000000000000000O10000000000WNX_O:h`0D[_O;e`0D\\_O<d`0B__O=a`0Aa_O?_`0@c_O?]`0_Oe_Oa0[`0]Oh_Ob0X`0]Oi_Oc0W`0[Ol_Od0T`0[Om_Oe0S`0YOP@f0P`0XOR@h0n?WOT@h0l?VOV@j0j?UOV@l0k?QOW@o0i?oNY@Q1g?nNY@S1g?kN[@U1e?iN\\@X1d?gN]@Y1c?eN_@[1a?dN_@]1a?aNa@_1_?_Nb@b1^?]Nc@c1]?[Ne@e1j`010O0100M3N2N2M2O2N2N2M3N2N1N3N2N2M3O10O010000O100O010O100O100O010O100O1000O0100O100O10O0100O100O10O0100O10000O010O100O100O010O1a0_Og0YOh0XOg0XO_iX3"}, "label": "pizza being cut by fork and knife"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza being cut into with a knife and fork\"."}], "task": "refering", "answer_template": "The \"the pizza being cut into with a knife and fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 78, 89, 90, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 111, 112, 124, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146], "bbox": [0.2816458333333333, 0.192578125, 0.6501875, 0.37459375], "iscrowd": 0, "area": 14777.88625, "rle": {"size": [640, 480], "counts": "g`d22mc08I8H8H8H7I8H8G8I8H8H7I2N00O00100O100O101O1O1O001O1O1O1O1O1O1O1O1OTOgMR@X2n?jMQ@U2o?mMP@R2o?RNn_On1R`0TNm_Ok1S`0WNl_Oh1T`0ZNk_Oe1U`0]Nj_Ob1V`0`Ni_O_1W`0cNh_O\\1X`0fNg_OY1Y`0iNf_OV1Z`0lNe_OS1[`0oNd_OP1\\`0ROc_Om0]`0UOa_Ok0_`0WO`_Oh0_`0[O`_Od0``0^O__Oa0a`0A^_O>b`0D]_O;c`0G\\_O8e`0IZ_O6f`0LY_O3g`0c1000000000000000000000000000000000O10000000000WNX_O:h`0D[_O;e`0D\\_O<d`0B__O=a`0Aa_O?_`0@c_O?]`0_Oe_Oa0[`0]Oh_Ob0X`0]Oi_Oc0W`0[Ol_Od0T`0[Om_Oe0S`0YOP@f0P`0XOR@h0n?WOT@h0l?VOV@j0j?UOV@l0k?QOW@o0i?oNY@Q1g?nNY@S1g?kN[@U1e?iN\\@X1d?gN]@Y1c?eN_@[1a?dN_@]1a?aNa@_1_?_Nb@b1^?]Nc@c1]?[Ne@e1j`010O0100M3N2N2M2O2N2N2M3N2N1N3N2N2M3O10O010000O100O010O100O100O010O100O1000O0100O100O10O0100O100O10O0100O10000O010O100O100O010O1a0_Og0YOh0XOg0XO_iX3"}, "label": "the pizza being cut into with a knife and fork"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza closest to the camera\"."}], "task": "refering", "answer_template": "The \"the pizza closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.0, 0.42921875, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 112786.37554999998, "rle": {"size": [640, 480], "counts": "\\:R1nb0m1SN0O101O000O1000001O0O1000000O101O000O1000000O2O0000000O10001O0O1000000O101O00000O10001O0O1000000O101O001N101O0O2O00001N101O0O2O001O0O101O0O2O001O0O2O00001N101O0O2O001O0O101O0O2O001O0O2O00001N101O0O2O001O000O2O001O001O0O2O00001O0O2O001O1O0O2O001O001N101O001O001N101O001O1N101O001O0O2O001O001O0O2O001O1O0O2O001O001N101O001O001N101O001O1N101O001O0O2O001O001N101O001O001N2O001O001N101O001O001O0O2O001O1O001O001O0O2O001O001O001O001N101O1O001O001O001N101O001O001O001O0O2O1O001O001O001O0O2O001O001O001O001N2O001O001O001O0O2O001O001O001O1O0O2O001O001O001O001N101O001O001O1O001O001O001O0O2O001O001O001O001O1O001O001O001O001N101O001O001O1O001O001O001O001O001O0O2O001O001O1O001O001O001O001O001O001N101O001O1O1O001O1O1O001O1O1O10O01O100O00100O1O010O1O10O01O100O00100O1O10O01O100O00100O1O010O1O10O0100O1O01O01O01O00001O00010O00001O00001O01O01O00001O00010O00001O00001O01O01O00001O001O00010O00001O00001O01O01O00001O000010O0001O00001O00010O00001O00001O00001O00001O001O0000TF"}, "label": "the pizza closest to the camera"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza closest to the camera\"."}], "task": "refering", "answer_template": "The \"the pizza closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.0, 0.42921875, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 112786.37554999998, "rle": {"size": [640, 480], "counts": "\\:R1nb0m1SN0O101O000O1000001O0O1000000O101O000O1000000O2O0000000O10001O0O1000000O101O00000O10001O0O1000000O101O001N101O0O2O00001N101O0O2O001O0O101O0O2O001O0O2O00001N101O0O2O001O0O101O0O2O001O0O2O00001N101O0O2O001O000O2O001O001O0O2O00001O0O2O001O1O0O2O001O001N101O001O001N101O001O1N101O001O0O2O001O001O0O2O001O1O0O2O001O001N101O001O001N101O001O1N101O001O0O2O001O001N101O001O001N2O001O001N101O001O001O0O2O001O1O001O001O0O2O001O001O001O001N101O1O001O001O001N101O001O001O001O0O2O1O001O001O001O0O2O001O001O001O001N2O001O001O001O0O2O001O001O001O1O0O2O001O001O001O001N101O001O001O1O001O001O001O0O2O001O001O001O001O1O001O001O001O001N101O001O001O1O001O001O001O001O001O0O2O001O001O1O001O001O001O001O001O001N101O001O1O1O001O1O1O001O1O1O10O01O100O00100O1O010O1O10O01O100O00100O1O10O01O100O00100O1O010O1O10O0100O1O01O01O01O00001O00010O00001O00001O01O01O00001O00010O00001O00001O01O01O00001O001O00010O00001O00001O01O01O00001O000010O0001O00001O00010O00001O00001O00001O00001O001O0000TF"}, "label": "the pizza closest to the camera"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass on the left\"."}], "task": "refering", "answer_template": "The \"the glass on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 17, 18, 19, 20, 21, 34, 35, 36, 37, 38, 51, 52, 53, 54, 55, 68, 69, 70, 71, 72, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174], "bbox": [0.006, 0.0045, 0.3086041666666667, 0.48090625], "iscrowd": 0, "area": 39753.4217, "rle": {"size": [640, 480], "counts": "Wl1T3l`0R4nK8H8H9G8H8H9G8H9G4L0000O100000000000O10000000O100000000000000000000O100000000000000000000O1000O1000000000000000O10000000000000000000000O10000000000000O100000O100000000000000000000O100000000000000000000O100000O10000000000000O100000000000000O10000O101N>B=VNZDaJT<R5mClJ`<g4bCTKo<^4SC_K\\=T4eBiKi=k3XBRLW>a3lAXLe>[3]A_LS?U3o@fLa?l2V1D=B>B=D=B>B=DiX_6"}, "label": "the glass on the left"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest glass of water\"."}], "task": "refering", "answer_template": "The \"the tallest glass of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 17, 18, 19, 20, 21, 34, 35, 36, 37, 38, 51, 52, 53, 54, 55, 68, 69, 70, 71, 72, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174], "bbox": [0.006, 0.0045, 0.3086041666666667, 0.48090625], "iscrowd": 0, "area": 39753.4217, "rle": {"size": [640, 480], "counts": "Wl1T3l`0R4nK8H8H9G8H8H9G8H9G4L0000O100000000000O10000000O100000000000000000000O100000000000000000000O1000O1000000000000000O10000000000000000000000O10000000000000O100000O100000000000000000000O100000000000000000000O100000O10000000000000O100000000000000O10000O101N>B=VNZDaJT<R5mClJ`<g4bCTKo<^4SC_K\\=T4eBiKi=k3XBRLW>a3lAXLe>[3]A_LS?U3o@fLa?l2V1D=B>B=D=B>B=DiX_6"}, "label": "the tallest glass of water"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear glass filled with a clear liquid\"."}], "task": "refering", "answer_template": "The \"a clear glass filled with a clear liquid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 183, 184, 185], "bbox": [0.6310208333333333, 0.160921875, 0.976625, 0.45696875], "iscrowd": 0, "area": 24614.434099999995, "rle": {"size": [640, 480], "counts": "c_m58gc0=D;D<D=D;D=D;D=C<E<C<E;D7J001O00000000001O000O101O0000001O000000001O0O2O1O001O001O1O001O00000O2O0000001O0000001O00000O2O000000001O0000001O0000001O0001O0001O0000001O0000001O000000001O00000010O000001O000000001O0000001O00000000O101O0O100O100O100O100O100O1L4K5L4L4L4K6K4L4K5L5K4L4K6K4L4K6K4L4L4K6K4L4L5J5L4L4K6K4L4L5J5L4L4L5J5L4L5J5L`h6"}, "label": "a clear glass filled with a clear liquid"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass\"."}], "task": "refering", "answer_template": "The \"glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 164, 165, 166, 167, 168, 183, 184, 185], "bbox": [0.6310208333333333, 0.160921875, 0.976625, 0.45696875], "iscrowd": 0, "area": 24614.434099999995, "rle": {"size": [640, 480], "counts": "c_m58gc0=D;D<D=D;D=D;D=C<E<C<E;D7J001O00000000001O000O101O0000001O000000001O0O2O1O001O001O1O001O00000O2O0000001O0000001O00000O2O000000001O0000001O0000001O0001O0001O0000001O0000001O000000001O00000010O000001O000000001O0000001O00000000O101O0O100O100O100O100O100O1L4K5L4L4L4K6K4L4K5L5K4L4K6K4L4K6K4L4L4K6K4L4L5J5L4L4K6K4L4L5J5L4L4L5J5L4L5J5L`h6"}, "label": "glass"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white tablelcoth is covering the table\"."}], "task": "refering", "answer_template": "The \"a white tablelcoth is covering the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 134, 135, 151, 152, 158, 159, 160, 161, 162, 163, 164, 167, 168, 169, 175, 176, 177, 178, 184, 185, 186, 203], "bbox": [0.30052083333333335, 0.264125, 0.9982708333333333, 0.49987499999999996], "iscrowd": 0, "area": 10055.697050000006, "rle": {"size": [640, 480], "counts": "[Yj21kc04L4L3M4L4L4L4L3M4L4L4O10O100000000O100000000O10000000000O100000000O2O0000000O100000000O100000000O10000000000O100000000O101O00000O100000000O100000000O10000000000O100000000O101O00000O100000000O100000000O10000000000O100000000O101O00000O100000000O100000000O10000000000O100000000O101O0000O010O100O10O01O100O10O0100O100O010O100O100O010O1002N5J5L5K4LdPg10[oXN2N1O1O2N1O2N1O2N1O2N1O1O2N101N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1G:G8H9F9H9G8J6L5K4L5K4L5K4O2O00001O0O2O001O001O0O101O001O001N101O0d0\\Og0ZOf0YOg0ZOf0YOg0ZO_:"}, "label": "a white tablelcoth is covering the table"}]}
{"id": 326823, "image": "COCO_train2014_000000326823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with food on it\"."}], "task": "refering", "answer_template": "The \"a table with food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 134, 135, 151, 152, 158, 159, 160, 161, 162, 163, 164, 167, 168, 169, 175, 176, 177, 178, 184, 185, 186, 203], "bbox": [0.30052083333333335, 0.264125, 0.9982708333333333, 0.49987499999999996], "iscrowd": 0, "area": 10055.697050000006, "rle": {"size": [640, 480], "counts": "[Yj21kc04L4L3M4L4L4L4L3M4L4L4O10O100000000O100000000O10000000000O100000000O2O0000000O100000000O100000000O10000000000O100000000O101O00000O100000000O100000000O10000000000O100000000O101O00000O100000000O100000000O10000000000O100000000O101O00000O100000000O100000000O10000000000O100000000O101O0000O010O100O10O01O100O10O0100O100O010O100O100O010O1002N5J5L5K4LdPg10[oXN2N1O1O2N1O2N1O2N1O2N1O1O2N101N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O1O2N1G:G8H9F9H9G8J6L5K4L5K4L5K4O2O00001O0O2O001O001O0O101O001O001N101O0d0\\Og0ZOf0YOg0ZOf0YOg0ZO_:"}, "label": "a table with food on it"}]}
{"id": 130215, "image": "COCO_train2014_000000130215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"computer monitor\"."}], "task": "refering", "answer_template": "The \"computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 215, 216], "bbox": [0.21392187499999998, 0.051569086651053864, 0.6148281250000001, 0.6233255269320842], "iscrowd": 0, "area": 50424.90795, "rle": {"size": [427, 640], "counts": "_Ui1:Q=c0\\Oc0]Od0\\Oc0]Od0\\Oc0^Oc0]Ob0^O5K4L4L4L4L4L4K5L2N00000000000000000000000O10000000O100000000000000000O1000000000000000O1000000000O1000000000000000000000O1000O10000000O1000000000000O1000O1000002N5K5K5K5K5K1O001O1O0O2O3M0000O100000000O10000000O0100000000O100000000O100000O10O100000000O100000000O1000O1000O100000000O100000000O10O1000O10000ROhIiJX6W5jIgJV6Y5kIeJV6[5lIcJT6\\5oIbJQ6^5PJaJP6_5RJ_Jn5a5TJ\\Jm5d5TJ[Jk5f5WJXJi5h5YJVJg5j5ZJUJf5j5]JSJd5m5^JQJb5o5_JPJa5P6aJnI_5R6m000000000O10O100000000000O10000000000000O10O100000000000000O10000000O10000000O100000000000000O10O10000000000000O100000000000O10O1M3C=C=D<C>C=B?B=B?B=B?B=B?BVoV3"}, "label": "computer monitor"}]}
{"id": 130215, "image": "COCO_train2014_000000130215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white apple computer monitor featuring a picture of gears\"."}], "task": "refering", "answer_template": "The \"white apple computer monitor featuring a picture of gears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 215, 216], "bbox": [0.21392187499999998, 0.051569086651053864, 0.6148281250000001, 0.6233255269320842], "iscrowd": 0, "area": 50424.90795, "rle": {"size": [427, 640], "counts": "_Ui1:Q=c0\\Oc0]Od0\\Oc0]Od0\\Oc0^Oc0]Ob0^O5K4L4L4L4L4L4K5L2N00000000000000000000000O10000000O100000000000000000O1000000000000000O1000000000O1000000000000000000000O1000O10000000O1000000000000O1000O1000002N5K5K5K5K5K1O001O1O0O2O3M0000O100000000O10000000O0100000000O100000000O100000O10O100000000O100000000O1000O1000O100000000O100000000O10O1000O10000ROhIiJX6W5jIgJV6Y5kIeJV6[5lIcJT6\\5oIbJQ6^5PJaJP6_5RJ_Jn5a5TJ\\Jm5d5TJ[Jk5f5WJXJi5h5YJVJg5j5ZJUJf5j5]JSJd5m5^JQJb5o5_JPJa5P6aJnI_5R6m000000000O10O100000000000O10000000000000O10O100000000000000O10000000O10000000O100000000000000O10O10000000000000O100000000000O10O1M3C=C=D<C>C=B?B=B?B=B?B=B?BVoV3"}, "label": "white apple computer monitor featuring a picture of gears"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lawn chair being occupied by a man\"."}], "task": "refering", "answer_template": "The \"lawn chair being occupied by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 251, 252, 275, 308, 309, 310, 331, 332, 353, 354, 355, 376, 377], "bbox": [0.380640625, 0.5634374999999999, 1.0, 0.9569791666666667], "iscrowd": 0, "area": 4157.5073, "rle": {"size": [480, 640], "counts": "Sjb38c>5L4K4L5L4K5L3L5K4M4K4L5M2O2N101N1O2N2O10O10000O10000O100O1O1O1O1O1O1O1N2O1O2N2N2N2O1O1N2O0O3N2M3J6K5K5L4L5KWV4FSjK4M3M3N2N2N1N0100O4M8GbUk3MbjTL0O2N1010O01O001O001O001O001O001O1O2N1O2O1O0O2O001O1N101O1O0O101O0O101N10001N100O2O0O101N101N2O0O2O1N101NK6H7I8HRG"}, "label": "lawn chair being occupied by a man"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue chair with a man sitting in it\"."}], "task": "refering", "answer_template": "The \"a blue chair with a man sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 251, 252, 275, 308, 309, 310, 331, 332, 353, 354, 355, 376, 377], "bbox": [0.380640625, 0.5634374999999999, 1.0, 0.9569791666666667], "iscrowd": 0, "area": 4157.5073, "rle": {"size": [480, 640], "counts": "Sjb38c>5L4K4L5L4K5L3L5K4M4K4L5M2O2N101N1O2N2O10O10000O10000O100O1O1O1O1O1O1O1N2O1O2N2N2N2O1O1N2O0O3N2M3J6K5K5L4L5KWV4FSjK4M3M3N2N2N1N0100O4M8GbUk3MbjTL0O2N1010O01O001O001O001O001O001O1O2N1O2O1O0O2O001O1N101O1O0O101O0O101N10001N100O2O0O101N101N2O0O2O1N101NK6H7I8HRG"}, "label": "a blue chair with a man sitting in it"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green tank top\"."}], "task": "refering", "answer_template": "The \"a woman in a green tank top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322], "bbox": [0.0033750000000000004, 0.0015625, 0.27809375, 0.8397708333333332], "iscrowd": 0, "area": 40578.81185, "rle": {"size": [480, 640], "counts": "[n0h0X>g1cBQN[:P2cEFg8<WGZ1T7hNiHn2a5TM]Jb4n3`KoKV6[2lIcMY6Y2iIeMW6[2jIcMV6]2lIaMU6_2mI^MS6b2oI\\MR6d2T3010O010O10O01O010O010O100O100O100O100O100O1O100O100O100O100O100O100O2N101N101N100O2O0O2O0O2O0O2O0O2XG^Lo5c3oI_LQ6b3lIaLQ6a3lIbLS6`3jIbLT6a3iIbLT6`3jIbLT6a3iIaLV6`3iIbLT6a3jI`LT6b3kI_LS6d3kI^LS6c3lI^LR6e3lI]LQ6f3mI[LQ6g3nIZLQ6h3mIZLP6j3mIWLQ6n3iIULU6Q4bIUL[6P4]IULb6P4TIVLj6o3nHWLo6n5N2O2M2N3M2O1N3M2N2O2M2K5C>C<01N1O3N1N2N2O1N2O1N2N2O1L4@`0N2O1O2M2O1O1N2O1O1N2O1O1N2O1O1N3N1O1N2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M3N5K4K5L4L5K4L4L4L4L5K4L4L_Zh6"}, "label": "a woman in a green tank top"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the green shirt watching the boy eat a sandwich\"."}], "task": "refering", "answer_template": "The \"the girl in the green shirt watching the boy eat a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322], "bbox": [0.0033750000000000004, 0.0015625, 0.27809375, 0.8397708333333332], "iscrowd": 0, "area": 40578.81185, "rle": {"size": [480, 640], "counts": "[n0h0X>g1cBQN[:P2cEFg8<WGZ1T7hNiHn2a5TM]Jb4n3`KoKV6[2lIcMY6Y2iIeMW6[2jIcMV6]2lIaMU6_2mI^MS6b2oI\\MR6d2T3010O010O10O01O010O010O100O100O100O100O100O1O100O100O100O100O100O100O2N101N101N100O2O0O2O0O2O0O2O0O2XG^Lo5c3oI_LQ6b3lIaLQ6a3lIbLS6`3jIbLT6a3iIbLT6`3jIbLT6a3iIaLV6`3iIbLT6a3jI`LT6b3kI_LS6d3kI^LS6c3lI^LR6e3lI]LQ6f3mI[LQ6g3nIZLQ6h3mIZLP6j3mIWLQ6n3iIULU6Q4bIUL[6P4]IULb6P4TIVLj6o3nHWLo6n5N2O2M2N3M2O1N3M2N2O2M2K5C>C<01N1O3N1N2N2O1N2O1N2N2O1L4@`0N2O1O2M2O1O1N2O1O1N2O1O1N2O1O1N3N1O1N2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M3N5K4K5L4L5K4L4L4L4L5K4L4L_Zh6"}, "label": "the girl in the green shirt watching the boy eat a sandwich"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the seat the woman is in\"."}], "task": "refering", "answer_template": "The \"the seat the woman is in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 71, 72, 73, 96, 189, 190, 210, 211, 212, 233, 234, 256, 278, 279, 301, 302, 303, 323, 324, 325, 326, 345, 346, 348, 371], "bbox": [0.0, 0.13039583333333335, 0.279640625, 0.9881458333333335], "iscrowd": 0, "area": 7837.459150000003, "rle": {"size": [480, 640], "counts": "X>1l>2O2O00000000001O0000000000001O00000000001O00000000001O0000000000001O000000005KO1000O100O100O010O10000O10O0100O1O100O1O001O1O1O1O100O002N1O1O1iGSO[2n0dMTO[2l0dMWOY2j0gMWOW2j0hMXOV2j0hMYOU2i0iMYOU2i0iMYOU2i0hMZOW2g0eM^OX2d0eM_OY2b0dMBZ2`0bME[2=bMF\\2<cME[2=eMDX2>hMBT2a0mM@o1c0QN]Ok1f0WNZOe1i0[NWOb1k0_NVO\\1m0eNSOX1P1hNQOS1R1oNmNn0U1SOlNi0W1WOiNf0Z1ZOgNb0_1[OaNc0d1ZO\\Nc0i1PMkLJ\\1S3n1`LVM=l0R3R2mKdMQ1:o2W2\\KPNf1Hm2\\2jJ\\NZ2XOj2d4WM[Kh20VLP2Q1QNf2NfLg1e0ZNd2NRM_1;cNa2NXM\\18eN_2N]Md0l0^Od1ObMGe1:h0NgM]Ok1e0<NlMZOj1g09OmMZOl1g04OQNYOn1g000RNYOP2g0LOTN[OQ2f0JOUNYOU2h0COXNWOZ2i0]O0YNTO_2l0VOO[NROe2n0oNO^NoNg2S1jNHX48hKB_4:cKCa40lKOU4GUL9X90100O10O10O100O010O1000O0100O010O1000O0100O10O01O100O00100O00100O10O01O100O00100O1O2O0O2N101N2O0OVZh6"}, "label": "the seat the woman is in"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green camping chair a young woman is sitting in\"."}], "task": "refering", "answer_template": "The \"a green camping chair a young woman is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 71, 72, 73, 96, 189, 190, 210, 211, 212, 233, 234, 256, 278, 279, 301, 302, 303, 323, 324, 325, 326, 345, 346, 348, 371], "bbox": [0.0, 0.13039583333333335, 0.279640625, 0.9881458333333335], "iscrowd": 0, "area": 7837.459150000003, "rle": {"size": [480, 640], "counts": "X>1l>2O2O00000000001O0000000000001O00000000001O00000000001O0000000000001O000000005KO1000O100O100O010O10000O10O0100O1O100O1O001O1O1O1O100O002N1O1O1iGSO[2n0dMTO[2l0dMWOY2j0gMWOW2j0hMXOV2j0hMYOU2i0iMYOU2i0iMYOU2i0hMZOW2g0eM^OX2d0eM_OY2b0dMBZ2`0bME[2=bMF\\2<cME[2=eMDX2>hMBT2a0mM@o1c0QN]Ok1f0WNZOe1i0[NWOb1k0_NVO\\1m0eNSOX1P1hNQOS1R1oNmNn0U1SOlNi0W1WOiNf0Z1ZOgNb0_1[OaNc0d1ZO\\Nc0i1PMkLJ\\1S3n1`LVM=l0R3R2mKdMQ1:o2W2\\KPNf1Hm2\\2jJ\\NZ2XOj2d4WM[Kh20VLP2Q1QNf2NfLg1e0ZNd2NRM_1;cNa2NXM\\18eN_2N]Md0l0^Od1ObMGe1:h0NgM]Ok1e0<NlMZOj1g09OmMZOl1g04OQNYOn1g000RNYOP2g0LOTN[OQ2f0JOUNYOU2h0COXNWOZ2i0]O0YNTO_2l0VOO[NROe2n0oNO^NoNg2S1jNHX48hKB_4:cKCa40lKOU4GUL9X90100O10O10O100O010O1000O0100O010O1000O0100O10O01O100O00100O00100O10O01O100O00100O1O2O0O2N101N2O0OVZh6"}, "label": "a green camping chair a young woman is sitting in"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an orange shirt is sitting and eating , he has sun glasses on top of his head\"."}], "task": "refering", "answer_template": "The \"a man in an orange shirt is sitting and eating , he has sun glasses on top of his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 35, 36, 37, 38, 39, 40, 41, 42, 43, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 279, 280, 281, 282, 283, 284, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.156390625, 0.006604166666666666, 1.0, 0.9819583333333334], "iscrowd": 0, "area": 141810.53710000002, "rle": {"size": [480, 640], "counts": "lU_12n>4K5L4L5J5L4K5L4K5L4L4K5L4K5L4L3L2O1N2O1O1N2O1N2O1N2O1O1N3N1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O00000O0100000O100000O10O100000000O10O1000O100000bLWEe2i:ZMYEd2g:\\M[Ec2e:\\M]Ec2c:\\M_Ec2a:\\MaEc2^:\\MeEc2[:\\MgEc2Y:\\MiEb2X:^MiEa2W:_MjE`2U:aMlE^2T:aMnE^2R:bMoE]2Q:cMPF\\2o9eMSFY2m9fMUFX2l9hMUFW2k9iMVFV2j9jMWFU2h9kMZFT2f9lM[FS2e9mM\\FR2d9nM]FP2d9oM^FP2a9QN`Fn1`9RNaFm1_9SNbFl1^9SNdFl1[9UNfFj1Z9VNgFh1Z9WNiFg1W9YNjFf1V9ZNkFe1T9\\NmFc1S9\\NoFc1Q9]NPGb1KPNa7>dHb1KPN`7?fH_1JSN`7=gH`1HTNa7<hH_1GUNa7<hH_1FVNb7;iH^1DXNb7:kH^1CXNb7:lH]1AZNc79lH\\1B[Nb79mH[1@]Nb78oH[1^O_Nb76QIZ1]O`Nb76QIZ1\\OaNc75RIY1ZOcNd73SIZ1YOcNb75VIV1XOfNa75WIU1WOgNa75YIS1VOhN_76\\IR1TOiN_76^Io0SOlN^76_Im0TOmN\\77`Il0SOnN[77dIi0QOQOZ77eIh0QOQOY78gIe0POTOW79iIb0QOUOU79lIa0oNVOT7:mI?oNXOR7;PJ;oNZOP7<QJ:oNZOo6<TJ8mN]On6<UJ7mN]Ol6>XJ3lN@k6>YJ1mNAi6>\\J0kNBg6`0^JMkNDf6`0`JJkNFd6a0aJIkNFc6a0dJGiNIa6b0fJEiNI`6c0hJBiNK^6d0kJ^OgNO\\6d0oJ\\OeN0[6e0QKYOdN3Z6e0SKVOdN5W6f0WKTOaN7W6f0ZKPO`N:U6g0\\KmN`N<S6h0^KkN^N?Q6g0cKhN]Na0o5h0fKeNZNd0o5h0hKbNZNf0l5j0kK^NYNi0k5i0nK]NWNj0j5j0PLZNVNm0i5j0dMVOZ2l0eMUOZ2k0gMUOX2l0gMUOW2m0iMSOU2o0kMQOS2P1mMQOR2P1nMPOP2R1oMoNo1S1QNmNn1S1RNnNl1T1TNlNj1V1UNkNi1W1WNiNh1W1UMeMZMU1_5X1PMlM_Ml0_5Z1jLTNdMc0`5[1jLUNeM`0`5\\1kLVNcM>`5^1lLWNaM<a5_1mLWNaM:a5`1nLXN^M9b5a1oLXN^M7a5c1PMYN\\M5c5c1PM[N[M2d5d1QM[NYM2f5c1PM^NXMOh5c1oLaNVMMj5c1PMaNUMLk5c1oLdNSMJn5b1nL4Q3MoL3Q3MnL5Q3KnL6Q3KnL6R3JnL6R3JmL7R3JmL7S3ImL7R3JmL7S3IlL8T3HlL8S3IlL8T3HkL9U3GkL9T3HkL9U3GjL:V3FjL:U3GjL:V3FiL;W3EiL;V3FiL;W3EhL<W3EiL;W3EhL<X3DgL=X3DhL<X3DgL=Y3CfL>Y3CgL=Y3CfL>Z3BeL?Z3BfL>Z3BeL?[3AdL`0[3AdL`0\\3@dL`0[3AdL`0\\3@dL`0\\3@dL`0[3AdL`0\\3@dL`0\\3@dL`0[3AdL`0\\3@dL`0\\3@dL`0[3AdL`0\\3@dL`0[1^LVNR3?`0S1XMkMY2Q1`0l0RN`M^1d1`0d0X2\\OhM<`2CaM5g2KYMMo23QMEW3:jL^O^3b0bLVOf3j0ZLnNn3R1RLlNP4S1QLkNQ4U1oKiNS4W1mKhNT4W1mKgNU4Y1kKeNW4[1iKcNY4\\1hKcNY4]1gKcNY4]1gKcNY4\\1hKdNX4\\1hKdNX4\\1hKdNX4[1iKeNW4[1iKeNW4[1iKeNX4Y1iKgNZ4V1fKjN\\4T1dKlN^4Q1cKoN`4n0`KROb4l0^KTOd4i0]KWOf4f0ZKZOh4c0YK]Oj4`0VK@l4=UKCm4;SKEP57QKIQ55oJKS52nJNU5NlJ2V5LjJ4X5IiJ7Z5FfJ:\\5CeJ=o9000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000001O001O00001O00001bHB8>FD:=CF<:BH>8@J`07]OLb04\\ONd03YO0f00XO2h0NVO4j0MSO6l0JRO8n0HPO;o0FoN;Q1EmN=S1DjN?U1CgN?Y1EaN=_1G[N<d1HQN?o1EaMh0^2[OTMP1l2TOfLV1Z3nN^LW1a3mNXLV1h3nNPLW1o3mNiKW1W4lNbKX1^4lNZKY1e4kNTKX1l4lNlJX1T5lNdJY1[5jN^JZ1b5jNVJ[1i5iNPJZ1P6jNhIZ1X6jN`I[1_6hNZI\\1f6`3O00001O1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N5K5I7I6I8I7I7H7J7I7H8_Oa0eNZE"}, "label": "a man in an orange shirt is sitting and eating , he has sun glasses on top of his head"}]}
{"id": 97450, "image": "COCO_train2014_000000097450.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man eating a sandwhich\"."}], "task": "refering", "answer_template": "The \"a man eating a sandwhich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 35, 36, 37, 38, 39, 40, 41, 42, 43, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 279, 280, 281, 282, 283, 284, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.156390625, 0.006604166666666666, 1.0, 0.9819583333333334], "iscrowd": 0, "area": 141810.53710000002, "rle": {"size": [480, 640], "counts": "lU_12n>4K5L4L5J5L4K5L4K5L4L4K5L4K5L4L3L2O1N2O1O1N2O1N2O1N2O1O1N3N1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O00000O0100000O100000O10O100000000O10O1000O100000bLWEe2i:ZMYEd2g:\\M[Ec2e:\\M]Ec2c:\\M_Ec2a:\\MaEc2^:\\MeEc2[:\\MgEc2Y:\\MiEb2X:^MiEa2W:_MjE`2U:aMlE^2T:aMnE^2R:bMoE]2Q:cMPF\\2o9eMSFY2m9fMUFX2l9hMUFW2k9iMVFV2j9jMWFU2h9kMZFT2f9lM[FS2e9mM\\FR2d9nM]FP2d9oM^FP2a9QN`Fn1`9RNaFm1_9SNbFl1^9SNdFl1[9UNfFj1Z9VNgFh1Z9WNiFg1W9YNjFf1V9ZNkFe1T9\\NmFc1S9\\NoFc1Q9]NPGb1KPNa7>dHb1KPN`7?fH_1JSN`7=gH`1HTNa7<hH_1GUNa7<hH_1FVNb7;iH^1DXNb7:kH^1CXNb7:lH]1AZNc79lH\\1B[Nb79mH[1@]Nb78oH[1^O_Nb76QIZ1]O`Nb76QIZ1\\OaNc75RIY1ZOcNd73SIZ1YOcNb75VIV1XOfNa75WIU1WOgNa75YIS1VOhN_76\\IR1TOiN_76^Io0SOlN^76_Im0TOmN\\77`Il0SOnN[77dIi0QOQOZ77eIh0QOQOY78gIe0POTOW79iIb0QOUOU79lIa0oNVOT7:mI?oNXOR7;PJ;oNZOP7<QJ:oNZOo6<TJ8mN]On6<UJ7mN]Ol6>XJ3lN@k6>YJ1mNAi6>\\J0kNBg6`0^JMkNDf6`0`JJkNFd6a0aJIkNFc6a0dJGiNIa6b0fJEiNI`6c0hJBiNK^6d0kJ^OgNO\\6d0oJ\\OeN0[6e0QKYOdN3Z6e0SKVOdN5W6f0WKTOaN7W6f0ZKPO`N:U6g0\\KmN`N<S6h0^KkN^N?Q6g0cKhN]Na0o5h0fKeNZNd0o5h0hKbNZNf0l5j0kK^NYNi0k5i0nK]NWNj0j5j0PLZNVNm0i5j0dMVOZ2l0eMUOZ2k0gMUOX2l0gMUOW2m0iMSOU2o0kMQOS2P1mMQOR2P1nMPOP2R1oMoNo1S1QNmNn1S1RNnNl1T1TNlNj1V1UNkNi1W1WNiNh1W1UMeMZMU1_5X1PMlM_Ml0_5Z1jLTNdMc0`5[1jLUNeM`0`5\\1kLVNcM>`5^1lLWNaM<a5_1mLWNaM:a5`1nLXN^M9b5a1oLXN^M7a5c1PMYN\\M5c5c1PM[N[M2d5d1QM[NYM2f5c1PM^NXMOh5c1oLaNVMMj5c1PMaNUMLk5c1oLdNSMJn5b1nL4Q3MoL3Q3MnL5Q3KnL6Q3KnL6R3JnL6R3JmL7R3JmL7S3ImL7R3JmL7S3IlL8T3HlL8S3IlL8T3HkL9U3GkL9T3HkL9U3GjL:V3FjL:U3GjL:V3FiL;W3EiL;V3FiL;W3EhL<W3EiL;W3EhL<X3DgL=X3DhL<X3DgL=Y3CfL>Y3CgL=Y3CfL>Z3BeL?Z3BfL>Z3BeL?[3AdL`0[3AdL`0\\3@dL`0[3AdL`0\\3@dL`0\\3@dL`0[3AdL`0\\3@dL`0\\3@dL`0[3AdL`0\\3@dL`0\\3@dL`0[3AdL`0\\3@dL`0[1^LVNR3?`0S1XMkMY2Q1`0l0RN`M^1d1`0d0X2\\OhM<`2CaM5g2KYMMo23QMEW3:jL^O^3b0bLVOf3j0ZLnNn3R1RLlNP4S1QLkNQ4U1oKiNS4W1mKhNT4W1mKgNU4Y1kKeNW4[1iKcNY4\\1hKcNY4]1gKcNY4]1gKcNY4\\1hKdNX4\\1hKdNX4\\1hKdNX4[1iKeNW4[1iKeNW4[1iKeNX4Y1iKgNZ4V1fKjN\\4T1dKlN^4Q1cKoN`4n0`KROb4l0^KTOd4i0]KWOf4f0ZKZOh4c0YK]Oj4`0VK@l4=UKCm4;SKEP57QKIQ55oJKS52nJNU5NlJ2V5LjJ4X5IiJ7Z5FfJ:\\5CeJ=o9000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000001O001O00001O00001bHB8>FD:=CF<:BH>8@J`07]OLb04\\ONd03YO0f00XO2h0NVO4j0MSO6l0JRO8n0HPO;o0FoN;Q1EmN=S1DjN?U1CgN?Y1EaN=_1G[N<d1HQN?o1EaMh0^2[OTMP1l2TOfLV1Z3nN^LW1a3mNXLV1h3nNPLW1o3mNiKW1W4lNbKX1^4lNZKY1e4kNTKX1l4lNlJX1T5lNdJY1[5jN^JZ1b5jNVJ[1i5iNPJZ1P6jNhIZ1X6jN`I[1_6hNZI\\1f6`3O00001O1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N5K5I7I6I8I7I7H7J7I7H8_Oa0eNZE"}, "label": "a man eating a sandwhich"}]}
{"id": 351403, "image": "COCO_train2014_000000351403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with a grey shirt and white shorts\"."}], "task": "refering", "answer_template": "The \"a lady with a grey shirt and white shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 29, 30, 31, 32, 47, 48, 49, 50, 51, 66, 67, 68, 69, 84, 85, 86, 87, 101, 102, 103, 104, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 213, 224, 225, 226, 227, 228, 229, 230, 231, 232, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304], "bbox": [0.43189349112426034, 0.029208333333333333, 0.9978106508875739, 0.9595416666666666], "iscrowd": 0, "area": 60871.59169999999, "rle": {"size": [480, 507], "counts": "[aW31o>0O100O2O0O100O100O100O100O100O100O2O0O100O100O101N100O1[MBgF?W9KaF5\\97ZFJd9a0SF_Oj9m0lEUOQ:U1gEkNV:Z1gEgNW:\\1hEdNU:_1kEaNS:b1kE_NR:e1mE[NQ:h1nEYNn9j1QFWNm9l1RFTNk9P2TFPNj9S2TFnMi9U2WFkMg9X2XFhMe9\\2YFfMd9\\2\\FdMb9_2]FaM`9c2^F^M`9e2_F[M^9h2bFXM\\9k2bFVM[9n2dFSMY9P3fFPMW9S3hFnLV9U3iFkLT9Y3kFgLS9\\3kFeLR9^3nFbLP9a3oF_Ln8e3PG]Lm8e3SG[Lj8i3UGWLi8Q5M3N2M3N3L3N2M3N2M3K5J7I6K5J6N2N2O1N3N1O1N2O1O1N2O2M2O1O1ZLbGh0`8VOcGh0^8WOdGg0]8XOfGd0\\8ZOgGd0[8ZOhGb0Z8]OhGa0Y8]OjGa0W8^OlG>V8@mG>T8AnG=S8BPH:S8CPH;Q8DQH9Q8FRH7o7GTH7m7HVH6j7IXH5i7IZH5h7I[H4f7J]H4d7K^H3c7L`H1a7MbH1_7NdHO]70eHO\\7NgH0Z70hHMY73hHKY75hHIY77iHFX7:iHDX7<iHBX7>jH@W7?jH_OW7a0kH\\OV7d0kHZOV7f0kHXOV7h0lHUOU7k0lHSOU7m0mHPOU7o0lHPOT7P1mHnNT7R1nHkNS7U1nHiNS7W1nHgNS7Y1oHdNR7\\1oHbNS7]1X3000000000001O0000000000001O0001O001O1O1O100O001O1O1O1O010O1O1O1O1O00100O1O1O001O1O100O1O001O1O1]FTNP6m1nI^Nh5b1WJiNe0^Oc0i1fNTOGJb1S1eN^OiN5b2>cNHkMa0a3HbNU3]1lLaNV3^1jLaNW3_1jL_NX3a1hL]NZ3b1gL]NZ3b1gL\\N[3c1eL\\N\\3d1eLZN]3e1eLXN]3g1fLUN\\3[OSJi0c2H[3_ORJi0f2DZ3ARJj0f2BZ3CQJj0h2_OY3EPJl0j2[OW3IPJk0l2XOV3LnIl0o2TOU3NnIn0o2POU31mIn0P3nNT33mIo0R3jNS36lIo0T3gNR38lIP1e:oN[EQ1e:nN]EQ1d:nN]EQ1d:nN\\ES1d:kN^ET1b:lN^ET1c:jN_EU1b:jN_EU1b:iN_EW1a:iN`EV1a:iN`EW1`:gNaEY1`:fNaEY1_:gNbEX1_:iN_EW1b:iN^EV1c:jN]EU1c:lN\\EU1d:lN[ES1f:mNYES1h:mNXER1h:oNXEP1i:QOUEo0l:QOTEn0m:ROSEn0l:SOSEl0o:UOPEi0R;WOnDh0S;XOlDg0U;[OjDc0X;]OgDc0Z;]OfDa0\\;_OdD?];CaD=`;C`D;b;E^D9c;H\\D8e;IZD5h;KXD3j;MWD1i;0YDLi;5XDGj;9XDDi;<XDAi;`0YD\\Oi;e0XDXOi;h0YDTOi;l0YDPOh;Q1ZDlNg;U1ZDgNh;Y1P1M2O2M3M3M2N3M3M3M2N3N2M3M2N3M3MVb0"}, "label": "a lady with a grey shirt and white shorts"}]}
{"id": 351403, "image": "COCO_train2014_000000351403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with shoulders exposed with glasses on\"."}], "task": "refering", "answer_template": "The \"a woman with shoulders exposed with glasses on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 29, 30, 31, 32, 47, 48, 49, 50, 51, 66, 67, 68, 69, 84, 85, 86, 87, 101, 102, 103, 104, 118, 119, 120, 121, 122, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 213, 224, 225, 226, 227, 228, 229, 230, 231, 232, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304], "bbox": [0.43189349112426034, 0.029208333333333333, 0.9978106508875739, 0.9595416666666666], "iscrowd": 0, "area": 60871.59169999999, "rle": {"size": [480, 507], "counts": "[aW31o>0O100O2O0O100O100O100O100O100O100O2O0O100O100O101N100O1[MBgF?W9KaF5\\97ZFJd9a0SF_Oj9m0lEUOQ:U1gEkNV:Z1gEgNW:\\1hEdNU:_1kEaNS:b1kE_NR:e1mE[NQ:h1nEYNn9j1QFWNm9l1RFTNk9P2TFPNj9S2TFnMi9U2WFkMg9X2XFhMe9\\2YFfMd9\\2\\FdMb9_2]FaM`9c2^F^M`9e2_F[M^9h2bFXM\\9k2bFVM[9n2dFSMY9P3fFPMW9S3hFnLV9U3iFkLT9Y3kFgLS9\\3kFeLR9^3nFbLP9a3oF_Ln8e3PG]Lm8e3SG[Lj8i3UGWLi8Q5M3N2M3N3L3N2M3N2M3K5J7I6K5J6N2N2O1N3N1O1N2O1O1N2O2M2O1O1ZLbGh0`8VOcGh0^8WOdGg0]8XOfGd0\\8ZOgGd0[8ZOhGb0Z8]OhGa0Y8]OjGa0W8^OlG>V8@mG>T8AnG=S8BPH:S8CPH;Q8DQH9Q8FRH7o7GTH7m7HVH6j7IXH5i7IZH5h7I[H4f7J]H4d7K^H3c7L`H1a7MbH1_7NdHO]70eHO\\7NgH0Z70hHMY73hHKY75hHIY77iHFX7:iHDX7<iHBX7>jH@W7?jH_OW7a0kH\\OV7d0kHZOV7f0kHXOV7h0lHUOU7k0lHSOU7m0mHPOU7o0lHPOT7P1mHnNT7R1nHkNS7U1nHiNS7W1nHgNS7Y1oHdNR7\\1oHbNS7]1X3000000000001O0000000000001O0001O001O1O1O100O001O1O1O1O010O1O1O1O1O00100O1O1O001O1O100O1O001O1O1]FTNP6m1nI^Nh5b1WJiNe0^Oc0i1fNTOGJb1S1eN^OiN5b2>cNHkMa0a3HbNU3]1lLaNV3^1jLaNW3_1jL_NX3a1hL]NZ3b1gL]NZ3b1gL\\N[3c1eL\\N\\3d1eLZN]3e1eLXN]3g1fLUN\\3[OSJi0c2H[3_ORJi0f2DZ3ARJj0f2BZ3CQJj0h2_OY3EPJl0j2[OW3IPJk0l2XOV3LnIl0o2TOU3NnIn0o2POU31mIn0P3nNT33mIo0R3jNS36lIo0T3gNR38lIP1e:oN[EQ1e:nN]EQ1d:nN]EQ1d:nN\\ES1d:kN^ET1b:lN^ET1c:jN_EU1b:jN_EU1b:iN_EW1a:iN`EV1a:iN`EW1`:gNaEY1`:fNaEY1_:gNbEX1_:iN_EW1b:iN^EV1c:jN]EU1c:lN\\EU1d:lN[ES1f:mNYES1h:mNXER1h:oNXEP1i:QOUEo0l:QOTEn0m:ROSEn0l:SOSEl0o:UOPEi0R;WOnDh0S;XOlDg0U;[OjDc0X;]OgDc0Z;]OfDa0\\;_OdD?];CaD=`;C`D;b;E^D9c;H\\D8e;IZD5h;KXD3j;MWD1i;0YDLi;5XDGj;9XDDi;<XDAi;`0YD\\Oi;e0XDXOi;h0YDTOi;l0YDPOh;Q1ZDlNg;U1ZDgNh;Y1P1M2O2M3M3M2N3M3M3M2N3N2M3M2N3M3MVb0"}, "label": "a woman with shoulders exposed with glasses on"}]}
{"id": 367792, "image": "COCO_train2014_000000367792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the highchair that the baby is sitting on\"."}], "task": "refering", "answer_template": "The \"the back of the highchair that the baby is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 86, 87, 103, 104, 105, 106, 110, 126, 127, 128, 129, 133, 148, 149, 150, 151, 152, 156, 171, 172, 173, 174, 194, 195, 217, 218], "bbox": [0.44399686520376175, 0.1315, 0.8039968652037618, 0.40365625000000005], "iscrowd": 0, "area": 10508.757300000001, "rle": {"size": [640, 638], "counts": "iQa51oc04K5L5J5L4K5L4K5L4K5L4K5L4K5L4K5I7IjMg^On1Sa0VNQ_Oe1l`0_NY_O[1d`0hNa_OS1]`0POh_Oj0U`0ZOo_Oa0n?CW@7f?L_@O^?5j@BS?b0S2201N2O1N101N2O001N2O1N101N2O0O2O1N2O0O2O1N2O0O2O1TOk001N2N2O0O2O1N101O1O1O001O1O001O1OO1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O100O1O010O100O1O100O100O1O100O1O100O100O1O010O100O1O100K5cN]1A?O1N2N2O1N2N2N2O0O2N2N2O1NV`_1Om_`N00001O1O2I6M4L3N3M2N3k^O^Og>d0n@FP?>k@FS?<h@HW?;d@H]?8Y@1h?0m_O:S`0Fc_Oc0^`0S1=C=B?AR1nNlP^2"}, "label": "the back of the highchair that the baby is sitting on"}]}
{"id": 367792, "image": "COCO_train2014_000000367792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a back of a highchair\"."}], "task": "refering", "answer_template": "The \"a back of a highchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 86, 87, 103, 104, 105, 106, 110, 126, 127, 128, 129, 133, 148, 149, 150, 151, 152, 156, 171, 172, 173, 174, 194, 195, 217, 218], "bbox": [0.44399686520376175, 0.1315, 0.8039968652037618, 0.40365625000000005], "iscrowd": 0, "area": 10508.757300000001, "rle": {"size": [640, 638], "counts": "iQa51oc04K5L5J5L4K5L4K5L4K5L4K5L4K5L4K5I7IjMg^On1Sa0VNQ_Oe1l`0_NY_O[1d`0hNa_OS1]`0POh_Oj0U`0ZOo_Oa0n?CW@7f?L_@O^?5j@BS?b0S2201N2O1N101N2O001N2O1N101N2O0O2O1N2O0O2O1N2O0O2O1TOk001N2N2O0O2O1N101O1O1O001O1O001O1OO1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O100O1O010O100O1O100O100O1O100O1O100O100O1O010O100O1O100K5cN]1A?O1N2N2O1N2N2N2O0O2N2N2O1NV`_1Om_`N00001O1O2I6M4L3N3M2N3k^O^Og>d0n@FP?>k@FS?<h@HW?;d@H]?8Y@1h?0m_O:S`0Fc_Oc0^`0S1=C=B?AR1nNlP^2"}, "label": "a back of a highchair"}]}
{"id": 367792, "image": "COCO_train2014_000000367792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older lady crouched near a little boy in a highchair\"."}], "task": "refering", "answer_template": "The \"an older lady crouched near a little boy in a highchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329], "bbox": [0.11673981191222571, 0.014015625, 0.5448119122257054, 0.6433125000000001], "iscrowd": 0, "area": 66528.45735000003, "rle": {"size": [640, 638], "counts": "ec^1i0hb0>K6M3M2N3M2N3M2O2N2M3N1N3I7]Oc0L4L4M2N3M4L3M3M3M3M3J6K5L4N2M3N2M3N2N2O1O2M2O1O1N2O1N3N1O1N3N1N3N1N3N1N2O2M2O2M2O2N1O1O2O0O2N1O101N1O2N101N1O1O2O0O2N1O2N100O2N101N100O2OO00100O010O010O10O010O010UM[IZHe6d7RJhGm5U8aJaG_5[8oJZGQ5d8ZKTGf4h8aKTG`4j8eKSGZ4m8jKjF\\4T9X301O1N101O1N101N20O01O1O001O10O01O1O001O010O1O001O10O01000O0100O010O10O010O0100O01O0010O01O1O010O0010O01O0010O01O10O01O0010O01O1O010O001O0010O01O1O00100O001O1O010O1O010O10O1000O0100000O010000O01eMfFPJZ9o5iFoIV9P6nFnIR9P6QGnIP9Q6SGmIl8R6XGlIh8R6\\GlId8S6_GjIc8T6`GjI`8V6bGgI_8X6dGeI^8Z6dGdI\\8\\6eGbI\\8^6fG_I\\8`6eG_I[8a6gG\\IZ8d6gGZIZ8e6hGYIZ8f6hGXIX8h6iGVIX8j6jGSIX8l6iGSIW8m6jGQIW8n6lGnHW8Q7jGhH\\8X7fG_Ha8a7P2O1O001N2O1O1O1O002aM_CZMc<a2cC]M_<]2gCaM[<Z2jCdMW<W2oCgMS<T2RDjMo;T2UDhMn;T2VDjMk;T2WDkMk;Q2XDnMj;o1WDQNk;k1WDUNj;i1XDVNj;f1XDYNk;c1XD\\Nj;a1WD_Nk;]1WDcNj;[1XDdNj;X1XDgNk;V1WDhNl;T1VDlNl;Q1UDnNm;o0VDoNn;l0TDTOP<e0TDYOQ<=UDCn;4YDJl;LZD3k;C[D=i;ZO^Dc0g;TOaDg0_?J7I6K5JTee5"}, "label": "an older lady crouched near a little boy in a highchair"}]}
{"id": 367792, "image": "COCO_train2014_000000367792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on left side\"."}], "task": "refering", "answer_template": "The \"woman on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329], "bbox": [0.11673981191222571, 0.014015625, 0.5448119122257054, 0.6433125000000001], "iscrowd": 0, "area": 66528.45735000003, "rle": {"size": [640, 638], "counts": "ec^1i0hb0>K6M3M2N3M2N3M2O2N2M3N1N3I7]Oc0L4L4M2N3M4L3M3M3M3M3J6K5L4N2M3N2M3N2N2O1O2M2O1O1N2O1N3N1O1N3N1N3N1N3N1N2O2M2O2M2O2N1O1O2O0O2N1O101N1O2N101N1O1O2O0O2N1O2N100O2N101N100O2OO00100O010O010O10O010O010UM[IZHe6d7RJhGm5U8aJaG_5[8oJZGQ5d8ZKTGf4h8aKTG`4j8eKSGZ4m8jKjF\\4T9X301O1N101O1N101N20O01O1O001O10O01O1O001O010O1O001O10O01000O0100O010O10O010O0100O01O0010O01O1O010O0010O01O0010O01O10O01O0010O01O1O010O001O0010O01O1O00100O001O1O010O1O010O10O1000O0100000O010000O01eMfFPJZ9o5iFoIV9P6nFnIR9P6QGnIP9Q6SGmIl8R6XGlIh8R6\\GlId8S6_GjIc8T6`GjI`8V6bGgI_8X6dGeI^8Z6dGdI\\8\\6eGbI\\8^6fG_I\\8`6eG_I[8a6gG\\IZ8d6gGZIZ8e6hGYIZ8f6hGXIX8h6iGVIX8j6jGSIX8l6iGSIW8m6jGQIW8n6lGnHW8Q7jGhH\\8X7fG_Ha8a7P2O1O001N2O1O1O1O002aM_CZMc<a2cC]M_<]2gCaM[<Z2jCdMW<W2oCgMS<T2RDjMo;T2UDhMn;T2VDjMk;T2WDkMk;Q2XDnMj;o1WDQNk;k1WDUNj;i1XDVNj;f1XDYNk;c1XD\\Nj;a1WD_Nk;]1WDcNj;[1XDdNj;X1XDgNk;V1WDhNl;T1VDlNl;Q1UDnNm;o0VDoNn;l0TDTOP<e0TDYOQ<=UDCn;4YDJl;LZD3k;C[D=i;ZO^Dc0g;TOaDg0_?J7I6K5JTee5"}, "label": "woman on left side"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the trailor on the right closest to the stairs\"."}], "task": "refering", "answer_template": "The \"the trailor on the right closest to the stairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 192, 193, 194, 195, 216, 218, 219, 240, 241, 242, 264, 265], "bbox": [0.11821875, 0.28385964912280703, 0.5447656249999999, 0.7300438596491228], "iscrowd": 0, "area": 16560.827999999998, "rle": {"size": [456, 640], "counts": "UoQ11W>000000000O2O00000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001N10000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O000O100000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O00000O7J5K5K5K6J5K5K5K6J5K5K5K6J5K2N000000001O00000000001O00000000001O00000000001O002N2N2N2N1O2N2N2N2N2N2N2N1O001O1O001O1O00^OVFgLi9Q3aFnL^9Z2QFoMf0FY9Y2UFmMf0IT9X2ZFlMe0HS9Z2\\FkMg:R2^EkMc:S2aEjM`:U2j000O1000000O10000O1000010O01O1O0010O01O001O1O0fDjMc:V2iD^NV;Z201O1O100O1jFTMn6m2jH]MS7d2dHgMY7Z2dHjM[7V2aHSNY7n1dHZNV7g1fHcNS7]1kHgNS7Z1lHfNT7Z1mHeNT7Y1mHfNT7Y1nHfNR7Y1PIfNP7Y1RIfNn6X1UIgNl6W1UIiNk6V1WIhNj6W1XIhNh6W1ZIhNf6W1\\IhNe6V1gHWMFc1c7T1iH\\MB_1e7T1jH`M^O\\1h7S1kHeMWOY1o7Q1jHjMSOV1S8o0kHEV79kHGU78lHHT76nHJR75oHJR75oHKR73nHNR71oHOR7OoH1R7MoH3R7KoH5R7HPI7R7GnH:T7CmH=_7VObH3bNROX9;ZHJYOIi8NPHB1>[8@dI=U:O2N2N1OiXQ4"}, "label": "the trailor on the right closest to the stairs"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus hidden by other vehicle\"."}], "task": "refering", "answer_template": "The \"a bus hidden by other vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 192, 193, 194, 195, 216, 218, 219, 240, 241, 242, 264, 265], "bbox": [0.11821875, 0.28385964912280703, 0.5447656249999999, 0.7300438596491228], "iscrowd": 0, "area": 16560.827999999998, "rle": {"size": [456, 640], "counts": "UoQ11W>000000000O2O00000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001N10000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O000O100000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O00000O7J5K5K5K6J5K5K5K6J5K5K5K6J5K2N000000001O00000000001O00000000001O00000000001O002N2N2N2N1O2N2N2N2N2N2N2N1O001O1O001O1O00^OVFgLi9Q3aFnL^9Z2QFoMf0FY9Y2UFmMf0IT9X2ZFlMe0HS9Z2\\FkMg:R2^EkMc:S2aEjM`:U2j000O1000000O10000O1000010O01O1O0010O01O001O1O0fDjMc:V2iD^NV;Z201O1O100O1jFTMn6m2jH]MS7d2dHgMY7Z2dHjM[7V2aHSNY7n1dHZNV7g1fHcNS7]1kHgNS7Z1lHfNT7Z1mHeNT7Y1mHfNT7Y1nHfNR7Y1PIfNP7Y1RIfNn6X1UIgNl6W1UIiNk6V1WIhNj6W1XIhNh6W1ZIhNf6W1\\IhNe6V1gHWMFc1c7T1iH\\MB_1e7T1jH`M^O\\1h7S1kHeMWOY1o7Q1jHjMSOV1S8o0kHEV79kHGU78lHHT76nHJR75oHJR75oHKR73nHNR71oHOR7OoH1R7MoH3R7KoH5R7HPI7R7GnH:T7CmH=_7VObH3bNROX9;ZHJYOIi8NPHB1>[8@dI=U:O2N2N1OiXQ4"}, "label": "a bus hidden by other vehicle"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman that is walking the white horse\"."}], "task": "refering", "answer_template": "The \"a woman that is walking the white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240, 262, 263, 285, 286, 287, 308, 310], "bbox": [0.38996875000000003, 0.4201535087719298, 0.5194062500000001, 0.8574122807017545], "iscrowd": 0, "area": 7038.651900000003, "rle": {"size": [456, 640], "counts": "`f_32T>3N3NO10000000bMd1PE]Nl:T2fDlMX;c2M4bNRM^GP3`8SM\\GQ3R8mLUG4f0R3T8PMoF1j0P3W8_MfGd2Y8^MdGd2c7eLWHj02c2g7jLPHd00k2o7gLkG`01l2T8lMhGX2W8b1:eN[1M3M3N2O1O1O2N2N3MgEbLm9[3mEPMP:f3O1O2M2O001O0O2O001N15L9F7JN101O0000\\MoEMJHW:9WF@4Nd9a0`FRO5<\\9a0hG_OX8?jGAV8=lGCT8;nGDR8;PHER87PHIS82oGMT8OnG1T8KnG5T8GnG8U8ElG;Q;00OhB@i<`0SCDm<<nBJQ=5lB0S=b01N2O001N1N3M2N3DaBL`=3dBH_=2WjX4"}, "label": "a woman that is walking the white horse"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing a jean inserting a hand into her pocket is walking with a horse\"."}], "task": "refering", "answer_template": "The \"a lady wearing a jean inserting a hand into her pocket is walking with a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240, 262, 263, 285, 286, 287, 308, 310], "bbox": [0.38996875000000003, 0.4201535087719298, 0.5194062500000001, 0.8574122807017545], "iscrowd": 0, "area": 7038.651900000003, "rle": {"size": [456, 640], "counts": "`f_32T>3N3NO10000000bMd1PE]Nl:T2fDlMX;c2M4bNRM^GP3`8SM\\GQ3R8mLUG4f0R3T8PMoF1j0P3W8_MfGd2Y8^MdGd2c7eLWHj02c2g7jLPHd00k2o7gLkG`01l2T8lMhGX2W8b1:eN[1M3M3N2O1O1O2N2N3MgEbLm9[3mEPMP:f3O1O2M2O001O0O2O001N15L9F7JN101O0000\\MoEMJHW:9WF@4Nd9a0`FRO5<\\9a0hG_OX8?jGAV8=lGCT8;nGDR8;PHER87PHIS82oGMT8OnG1T8KnG5T8GnG8U8ElG;Q;00OhB@i<`0SCDm<<nBJQ=5lB0S=b01N2O001N1N3M2N3DaBL`=3dBH_=2WjX4"}, "label": "a lady wearing a jean inserting a hand into her pocket is walking with a horse"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white tour bus with the word alaska on its side\"."}], "task": "refering", "answer_template": "The \"a white tour bus with the word alaska on its side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321], "bbox": [0.79078125, 0.16278508771929825, 0.9984531249999999, 0.8421491228070176], "iscrowd": 0, "area": 26640.2343, "rle": {"size": [456, 640], "counts": "QaQ72m=:G9I601N2N101N2N101N10000000000013L3M4L4L4L3M47ZD^OUO1^8=[HGQO3b84ZHOnN3f8NXH4mN5i8EWH<iN6o8\\OUHd0fN7S9SOTHk0dN6c8DeH<bN3l7>_IT1c5IZJ:h4e0TK^Oc4k0ZKYOc4i0^KVO`4l0aKSO_4n0bKQO\\4P1fKnNZ4S1gKlNX4U1iKiNW4X1jKfNU4[1mKdNR4]1nKbNR4_1oK_NQ4b1PL[NP4h1PLVNP4l1PLRNP4P2PLnMo3U2QLiMo3X2QLgMo3[2QLbMP4`2PL^Mo3e2QLYMo3i2QLUMo3l2RLQMo3Q3PLlLQ4W3oKdLT4^3lK]LW4e3iKVLZ4k3gKPL[4R4fKiK]4X4dKcK_4^4aK^Ka4d4`KWKc4i4_KUKa4l4`KRK`4o4aKoJ^4S5cKkJ]4U5eKiJ[4X5bKjJ^4W5`KjJ_4X5^KjJb4W5\\KjJd4V5ZKlJf4U5XKlJg4V5VKlJj4U5TKlJl4U5QKmJo4S5oJoJP5S5nJnJR5S5kJoJU5R5iJoJV5S5gJoJY5Q5fJPKZ5Q5cJQK]5P5aJQK^5Q5_JQKa5o4^JRKb5o4[JVKb5k4\\JZK_5h4^J_K]5b4aJdKZ5\\4dJkKW5h6J6K3M000O1000O01000000O1000000O0100000O10O1O001N2O1O001J6H8H7N300000O1U2kMY4gK`5"}, "label": "a white tour bus with the word alaska on its side"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white colored bus with ' alaska ' written over it\"."}], "task": "refering", "answer_template": "The \"a white colored bus with ' alaska ' written over it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321], "bbox": [0.79078125, 0.16278508771929825, 0.9984531249999999, 0.8421491228070176], "iscrowd": 0, "area": 26640.2343, "rle": {"size": [456, 640], "counts": "QaQ72m=:G9I601N2N101N2N101N10000000000013L3M4L4L4L3M47ZD^OUO1^8=[HGQO3b84ZHOnN3f8NXH4mN5i8EWH<iN6o8\\OUHd0fN7S9SOTHk0dN6c8DeH<bN3l7>_IT1c5IZJ:h4e0TK^Oc4k0ZKYOc4i0^KVO`4l0aKSO_4n0bKQO\\4P1fKnNZ4S1gKlNX4U1iKiNW4X1jKfNU4[1mKdNR4]1nKbNR4_1oK_NQ4b1PL[NP4h1PLVNP4l1PLRNP4P2PLnMo3U2QLiMo3X2QLgMo3[2QLbMP4`2PL^Mo3e2QLYMo3i2QLUMo3l2RLQMo3Q3PLlLQ4W3oKdLT4^3lK]LW4e3iKVLZ4k3gKPL[4R4fKiK]4X4dKcK_4^4aK^Ka4d4`KWKc4i4_KUKa4l4`KRK`4o4aKoJ^4S5cKkJ]4U5eKiJ[4X5bKjJ^4W5`KjJ_4X5^KjJb4W5\\KjJd4V5ZKlJf4U5XKlJg4V5VKlJj4U5TKlJl4U5QKmJo4S5oJoJP5S5nJnJR5S5kJoJU5R5iJoJV5S5gJoJY5Q5fJPKZ5Q5cJQK]5P5aJQK^5Q5_JQKa5o4^JRKb5o4[JVKb5k4\\JZK_5h4^J_K]5b4aJdKZ5\\4dJkKW5h6J6K3M000O1000O01000000O1000000O0100000O10O1O001N2O1O001J6H8H7N300000O1U2kMY4gK`5"}, "label": "a white colored bus with ' alaska ' written over it"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and green bus , with truck following\"."}], "task": "refering", "answer_template": "The \"white and green bus , with truck following\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 253, 254, 255, 276, 277], "bbox": [0.002015625, 0.29383771929824565, 0.33504687499999997, 0.7914254385964912], "iscrowd": 0, "area": 27317.468200000003, "rle": {"size": [456, 640], "counts": "^b0S4k1]Lo5U6I7I10O01O00001O00001O001O00001O00001OO100O10000O100O10000O1000000O10000O1O1O1N2N2N2O1M3N2N2M3N200O1O1O001O1O1O1O1O1O1N2M3M3M4\\Oc0N2M3N2M3B>0000O10001N1000000O10000O1000001N100L4L4N2O1O1N2O1O1O2N1N2O1O1O1N2O1N2N2O2N100O100O100O100O100O2O000000000000000000000000000000000000000000000O10001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00001N10001O1O1O001O1O1O8H8H7I8H=C=Ck_m5"}, "label": "white and green bus , with truck following"}]}
{"id": 515252, "image": "COCO_train2014_000000515252.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a standard commuter bus with a green and blue emblem , being closely followed by a blue chevrolet pickup truck\"."}], "task": "refering", "answer_template": "The \"a standard commuter bus with a green and blue emblem , being closely followed by a blue chevrolet pickup truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 253, 254, 255, 276, 277], "bbox": [0.002015625, 0.29383771929824565, 0.33504687499999997, 0.7914254385964912], "iscrowd": 0, "area": 27317.468200000003, "rle": {"size": [456, 640], "counts": "^b0S4k1]Lo5U6I7I10O01O00001O00001O001O00001O00001OO100O10000O100O10000O1000000O10000O1O1O1N2N2N2O1M3N2N2M3N200O1O1O001O1O1O1O1O1O1N2M3M3M4\\Oc0N2M3N2M3B>0000O10001N1000000O10000O1000001N100L4L4N2O1O1N2O1O1O2N1N2O1O1O1N2O1N2N2O2N100O100O100O100O100O2O000000000000000000000000000000000000000000000O10001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00001N10001O1O1O001O1O1O8H8H7I8H=C=Ck_m5"}, "label": "a standard commuter bus with a green and blue emblem , being closely followed by a blue chevrolet pickup truck"}]}
{"id": 425141, "image": "COCO_train2014_000000425141.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large bottle of red salsa in the side door of a white refridgerator\"."}], "task": "refering", "answer_template": "The \"a large bottle of red salsa in the side door of a white refridgerator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 133, 134, 135, 150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 216, 217, 218, 219, 233, 234, 235, 236, 250, 251, 252], "bbox": [0.733375, 0.29075, 0.9997708333333333, 0.65721875], "iscrowd": 0, "area": 17099.96125, "rle": {"size": [640, 480], "counts": "l[l61dc02d\\O2Xc0Oh\\O4Uc0Lj\\O8fb0OU]OI5;^b00]]OE5?Wb00d]OA5b0Pb00k]O_O4c0ka03Q^OZO4b0ha07U^OWO2c0ea0:X^OSO4b0`a0?]^OoN2c0]a0b0`^OkN4b0Ya0f0d^OhN2c0Ua0j0i^ObN3c0Qa0n0l^O`N2b0o`0R1o^O[N3c0i`0V1T_OXN2b0g`0Z1W_OSN3c0b`0]1[_OQN2b0_`0a1`_OlM2c0Z`0e1c_OiM2b0X`0h1f_OfM3a0S`0n1j_OaM2b0P`0P2n_O^M3a0l?T2Q@\\M2a0h?X2V@VM3a0d?\\2Y@TM2a0a?_2]@oL3a0^?b2_@nL2a0\\?d2a@kL4`0Y?g2d@iL2`0W?j2g@fL3`0S?m2i@dL3?R?o2l@aL2a0o>Q3n@^L4`0l>T3PA]L3`0j>U3TAZL2a0h>X3UAXL3`0e>Z3XAVL3>e>_3WASL4<e>c3XAQL39e>i3WAnK48d>l3XAmK44d>Q4YAjK33d>V4XAhK30e>Z4XAfK4Md>_4XAdK4Lc>b4ZAbK2Jd>f4>2N2N2N2N3M2N2N2N2N2N2O1O2N1O1O1O1O1O2N1O1O100O2NO1O2N1O1O1O2N1O1O1O2N1O0000002N3M2N3M2N2N3M2N3M2N2N3M2N3M2M3N3M2N3M2N2N3M2M4M2N2N3M2N3M2N2N3L3N3M2N2N3M2N3fNoJ"}, "label": "a large bottle of red salsa in the side door of a white refridgerator"}]}
{"id": 425141, "image": "COCO_train2014_000000425141.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the jug in the fridge\"."}], "task": "refering", "answer_template": "The \"the jug in the fridge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 133, 134, 135, 150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 216, 217, 218, 219, 233, 234, 235, 236, 250, 251, 252], "bbox": [0.733375, 0.29075, 0.9997708333333333, 0.65721875], "iscrowd": 0, "area": 17099.96125, "rle": {"size": [640, 480], "counts": "l[l61dc02d\\O2Xc0Oh\\O4Uc0Lj\\O8fb0OU]OI5;^b00]]OE5?Wb00d]OA5b0Pb00k]O_O4c0ka03Q^OZO4b0ha07U^OWO2c0ea0:X^OSO4b0`a0?]^OoN2c0]a0b0`^OkN4b0Ya0f0d^OhN2c0Ua0j0i^ObN3c0Qa0n0l^O`N2b0o`0R1o^O[N3c0i`0V1T_OXN2b0g`0Z1W_OSN3c0b`0]1[_OQN2b0_`0a1`_OlM2c0Z`0e1c_OiM2b0X`0h1f_OfM3a0S`0n1j_OaM2b0P`0P2n_O^M3a0l?T2Q@\\M2a0h?X2V@VM3a0d?\\2Y@TM2a0a?_2]@oL3a0^?b2_@nL2a0\\?d2a@kL4`0Y?g2d@iL2`0W?j2g@fL3`0S?m2i@dL3?R?o2l@aL2a0o>Q3n@^L4`0l>T3PA]L3`0j>U3TAZL2a0h>X3UAXL3`0e>Z3XAVL3>e>_3WASL4<e>c3XAQL39e>i3WAnK48d>l3XAmK44d>Q4YAjK33d>V4XAhK30e>Z4XAfK4Md>_4XAdK4Lc>b4ZAbK2Jd>f4>2N2N2N2N3M2N2N2N2N2N2O1O2N1O1O1O1O1O2N1O1O100O2NO1O2N1O1O1O2N1O1O1O2N1O0000002N3M2N3M2N2N3M2N3M2N2N3M2N3M2M3N3M2N3M2N2N3M2M4M2N2N3M2N3M2N2N3L3N3M2N2N3M2N3fNoJ"}, "label": "the jug in the fridge"}]}
{"id": 425148, "image": "COCO_train2014_000000425148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person riding on back of a motorcycle\"."}], "task": "refering", "answer_template": "The \"person riding on back of a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 216, 217, 218, 241, 263, 264], "bbox": [0.337671875, 0.31843096234309626, 0.519578125, 0.6969665271966526], "iscrowd": 0, "area": 6725.962850000002, "rle": {"size": [478, 640], "counts": "]QU35`>a0\\Oa0G5M3N2O1N2N2N2O1O1N101OIRC_Nl<]1\\C`Nd<[1a0O02O1ZOe0N3L400O011O0O100000001N1001O01O010O010O001[DbMi:^2TEeMk:\\2REfMo:^2iDfMW;^2bDeM_;l2101XM`Dn1b;mMbDS2`;gMcDZ2_;aMdD^2n;10O010O010O010O1O010O1bNeC7[<FPD2Q<KTD3l;KVD4k;KVD4k;KUD5l;JTD7l;ISD7n;HSD7o;GQD9P<GoC:Q<EoC;Q<EPD:Q<EoC;R<CoC>P<BPD>Q<APD>Q<@PD`0o;AQD`0n;@SD?m;@TD`0l;_OVD`0j;]OYDd0;]Oo:c0nD@S;?iDEW;;fDH[;:^DIc;i11O001O001O001O000F]CYNe<b1_C]Ne<\\1^CdNg<T1\\ClNh<m0\\CQOi<g0[CYOj<?ZC@j<8ZCHj<1YCOl=OO0100000O10O3N1Obo^4"}, "label": "person riding on back of a motorcycle"}]}
{"id": 425148, "image": "COCO_train2014_000000425148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a passenger on a motorcycle with his driver\"."}], "task": "refering", "answer_template": "The \"a passenger on a motorcycle with his driver\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 216, 217, 218, 241, 263, 264], "bbox": [0.337671875, 0.31843096234309626, 0.519578125, 0.6969665271966526], "iscrowd": 0, "area": 6725.962850000002, "rle": {"size": [478, 640], "counts": "]QU35`>a0\\Oa0G5M3N2O1N2N2N2O1O1N101OIRC_Nl<]1\\C`Nd<[1a0O02O1ZOe0N3L400O011O0O100000001N1001O01O010O010O001[DbMi:^2TEeMk:\\2REfMo:^2iDfMW;^2bDeM_;l2101XM`Dn1b;mMbDS2`;gMcDZ2_;aMdD^2n;10O010O010O010O1O010O1bNeC7[<FPD2Q<KTD3l;KVD4k;KVD4k;KUD5l;JTD7l;ISD7n;HSD7o;GQD9P<GoC:Q<EoC;Q<EPD:Q<EoC;R<CoC>P<BPD>Q<APD>Q<@PD`0o;AQD`0n;@SD?m;@TD`0l;_OVD`0j;]OYDd0;]Oo:c0nD@S;?iDEW;;fDH[;:^DIc;i11O001O001O001O000F]CYNe<b1_C]Ne<\\1^CdNg<T1\\ClNh<m0\\CQOi<g0[CYOj<?ZC@j<8ZCHj<1YCOl=OO0100000O10O3N1Obo^4"}, "label": "a passenger on a motorcycle with his driver"}]}
{"id": 425148, "image": "COCO_train2014_000000425148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red hood of a car\"."}], "task": "refering", "answer_template": "The \"red hood of a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.7210669456066946, 1.0, 0.9834100418410043], "iscrowd": 0, "area": 70563.25385, "rle": {"size": [478, 640], "counts": "P>f0k<]1QOo00000O10000O1000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000O100000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O2O0000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O001O001O001O001O1O001O001O001O0000001O000000000000001O000000000O10001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000000RD"}, "label": "red hood of a car"}]}
{"id": 425148, "image": "COCO_train2014_000000425148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hood of a car\"."}], "task": "refering", "answer_template": "The \"the hood of a car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.7210669456066946, 1.0, 0.9834100418410043], "iscrowd": 0, "area": 70563.25385, "rle": {"size": [478, 640], "counts": "P>f0k<]1QOo00000O10000O1000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000O100000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10000000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O2O0000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O001O001O001O001O1O001O001O001O0000001O000000000000001O000000000O10001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000000RD"}, "label": "the hood of a car"}]}
{"id": 425148, "image": "COCO_train2014_000000425148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person driving a bike with a pillion rider sitting behind\"."}], "task": "refering", "answer_template": "The \"a person driving a bike with a pillion rider sitting behind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 147, 148, 171, 172, 173, 174, 194, 195, 196, 218, 219, 241, 242, 265], "bbox": [0.415328125, 0.33780334728033473, 0.57596875, 0.6893933054393305], "iscrowd": 0, "area": 4332.150950000001, "rle": {"size": [478, 640], "counts": "Vjl31m>001O001O1BOlA2R>?M2N3M3O0O2N2O1N2O1N2O1O001O2M2O1O1eCdNQ;^1lDeNS;[1lDfNU;Z1iDgNW;Z1gDhNX;X1]DaNF7n;W1YDhNF1Q<V1XDmNB1V<Q1UD_Ok;a0SDBm;=QDEP<HjCP1W<lNmCT1S<`NoCHOh1S<^NYDb1h;[NYDP1]O_O]<^OWDP1B_Om<?^CXOb<f0bCXO^<g0eCXO[<h0eCWO\\<h0fCWOY<i0hCVOY<i0gCXOX<h0hCXOY<h0fCYOZ<f0gCYOY<g0gCZOY<e0gC\\OY<c0gC^OX<c0hC\\OY<c0gCOg;1YD0f;0ZD0g;0XD0h;0YDOg;1YDOh;0XD3e;M\\D7`;I_D>Z;BfDe0T;ZOlDm0m:SOTEU1d:jN\\EV1d:jN\\EV1e:jNZEV1f:jN[EU1f:jNZEV1f:jNZE[1b:dN_E^1^:bNbE_1]:`NeE`1Z:`NgE_1Y:aNhE^1X:aNkE]1U:cNnEY1S:fNRFV1o9iNVFQ1k9nN\\Fl0e9SOQG6R9HUG1k8OVGOk81UGOj82VGMk83UGMj83WGMi83VGMk82VGNi8N]G0e8N\\G2d8M^G1c80U301O1O14L01JYYn3"}, "label": "a person driving a bike with a pillion rider sitting behind"}]}
{"id": 425148, "image": "COCO_train2014_000000425148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person driving a black motorcycle with another person sitting behind them\"."}], "task": "refering", "answer_template": "The \"a person driving a black motorcycle with another person sitting behind them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 147, 148, 171, 172, 173, 174, 194, 195, 196, 218, 219, 241, 242, 265], "bbox": [0.415328125, 0.33780334728033473, 0.57596875, 0.6893933054393305], "iscrowd": 0, "area": 4332.150950000001, "rle": {"size": [478, 640], "counts": "Vjl31m>001O001O1BOlA2R>?M2N3M3O0O2N2O1N2O1N2O1O001O2M2O1O1eCdNQ;^1lDeNS;[1lDfNU;Z1iDgNW;Z1gDhNX;X1]DaNF7n;W1YDhNF1Q<V1XDmNB1V<Q1UD_Ok;a0SDBm;=QDEP<HjCP1W<lNmCT1S<`NoCHOh1S<^NYDb1h;[NYDP1]O_O]<^OWDP1B_Om<?^CXOb<f0bCXO^<g0eCXO[<h0eCWO\\<h0fCWOY<i0hCVOY<i0gCXOX<h0hCXOY<h0fCYOZ<f0gCYOY<g0gCZOY<e0gC\\OY<c0gC^OX<c0hC\\OY<c0gCOg;1YD0f;0ZD0g;0XD0h;0YDOg;1YDOh;0XD3e;M\\D7`;I_D>Z;BfDe0T;ZOlDm0m:SOTEU1d:jN\\EV1d:jN\\EV1e:jNZEV1f:jN[EU1f:jNZEV1f:jNZE[1b:dN_E^1^:bNbE_1]:`NeE`1Z:`NgE_1Y:aNhE^1X:aNkE]1U:cNnEY1S:fNRFV1o9iNVFQ1k9nN\\Fl0e9SOQG6R9HUG1k8OVGOk81UGOj82VGMk83UGMj83WGMi83VGMk82VGNi8N]G0e8N\\G2d8M^G1c80U301O1O14L01JYYn3"}, "label": "a person driving a black motorcycle with another person sitting behind them"}]}
{"id": 375996, "image": "COCO_train2014_000000375996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the negro woman is looking somewhere\"."}], "task": "refering", "answer_template": "The \"the negro woman is looking somewhere\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 161, 162, 163, 180, 181, 182, 199, 200, 201, 218, 219, 220, 237, 238, 239, 256, 257, 258, 275, 276, 277, 294, 295, 296, 313, 314], "bbox": [0.4812428298279159, 0.332578125, 0.621472275334608, 0.7258437499999999], "iscrowd": 0, "area": 13573.936549999999, "rle": {"size": [640, 523], "counts": "jhm4:cc0;E;E<D<D>Ba0_OW1iN<e@iKX>d4XAgK`>S5H7I8H8H8H8H8BQI^CT7c<4K401N1O2O0O101N100O01N101O1O001N101O001O0O2O1O1O1O1N2O1O1O1N2O1O1O1O1N101M3M3M3M3M3M3M3L4K5YL_Bj0f=lNjBj0Z=lNUCj0P=lN`Cj0d<lNkCj0n?D;D=C<D^]k3"}, "label": "the negro woman is looking somewhere"}]}
{"id": 375996, "image": "COCO_train2014_000000375996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color girl in middle of two white couple\"."}], "task": "refering", "answer_template": "The \"a black color girl in middle of two white couple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 161, 162, 163, 180, 181, 182, 199, 200, 201, 218, 219, 220, 237, 238, 239, 256, 257, 258, 275, 276, 277, 294, 295, 296, 313, 314], "bbox": [0.4812428298279159, 0.332578125, 0.621472275334608, 0.7258437499999999], "iscrowd": 0, "area": 13573.936549999999, "rle": {"size": [640, 523], "counts": "jhm4:cc0;E;E<D<D>Ba0_OW1iN<e@iKX>d4XAgK`>S5H7I8H8H8H8H8BQI^CT7c<4K401N1O2O0O101N100O01N101O1O001N101O001O0O2O1O1O1O1N2O1O1O1N2O1O1O1O1N101M3M3M3M3M3M3M3L4K5YL_Bj0f=lNjBj0Z=lNUCj0P=lN`Cj0d<lNkCj0n?D;D=C<D^]k3"}, "label": "a black color girl in middle of two white couple"}]}
{"id": 375996, "image": "COCO_train2014_000000375996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy playing wii\"."}], "task": "refering", "answer_template": "The \"a guy playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 268, 269, 270, 271, 272, 273, 274, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 310, 311, 324, 325, 326, 327, 328, 343, 344, 345, 346, 347], "bbox": [0.06359464627151051, 0.305546875, 0.5041491395793499, 0.7975156250000001], "iscrowd": 0, "area": 32053.4441, "rle": {"size": [640, 523], "counts": "fSe02ic05J6J6J6J6J6K5J6M3M3M3N2M3M3M3N3L3M3N2M3M3N2M3M3N200O100O100O100O100O10000O100O100O100000jMP@<P`0Eo_O;i;fNdGo0cL;h;nN^Gg0jL;g;WOVG?TM9e;_OQG8ZM9c;BRG5[M9b;DRG3]M8`;GRG1^M8_;IRGO_M5a;MPGN`M2a;2nFLaMOb;8lFIbMMc;<jFGdMId;a0hFFdMGe;e0fFEdMCi;i0bFDeMBj;k0_FDhM@J]OW;_1WGDiM_OEB\\;[1UGDkM^OAF_;Y1TGCmM]O\\OKc;V1SGBPN[OWO0f;T1eD]Ni1T1fNZOSO4j;Q1bD_Nj1R1hNXOoN:m;m0`DbNk1o0jN0k:`0^DbNm1n0kNOj:b0\\DcNn1l0mNNj:c0XDfNP2i0oNMi:e0VDgNQ2g0QOLi:f0SDiNR2d0TOLi:e0oCmNS2b0WOKj:X1nEmNXOKl:W1kEmN[OKm:V1gEoN\\OKo:T1eEPO^OKP;S1aERO@JQ;S1^EROBKS;Q1ZESOEKT;o0WEVOEKV;n0TEVOHKX;l0oDYOJJ[;j0jD\\OKJ_;g0eD_OLJc;c0`DDNHd;c0]DEOHd;d0\\DD0Hc;f0\\DB1Hc;f0\\DB1Hc;g0[DA2Hc;h0ZD@3Hc;i0YD_O4Hc;i0YD_O4Hc;j0XD^O5Hc;j0XD^O5Hc;k0WD]O6Hb;m0WD[O7Hb;m0WD[O7Ha;o0WDYO9G`;P1WDYO9G_;R1WDVO;H^;R1WDVO;H^;S1VDUO<H];U1VDSO=H];U1VDSO=H\\;W1VDQO>H\\;W1VDQO>H\\;X1UDPO?H[;Y1VDoN?H[;Z1UDnN`0HZ;\\1UDlNa0HZ;\\1UDlNa0HZ;]1TDkNc0GX;_1UDjNc0GW;a1UDhNd0GW;b1TDhNd0FW;c1UDgNd0FV;e1UDfNd0EW;e1UDfNd0DW;h1TDeNd0CW;i1UDdNd0Ce9@oE[2GcNd0Bd9EnEW2IbNe0Bb9ImET2LbNd0Aa9NkEQ2O`Ne0A_92mDI<T2c0aNd0ZOd9m0SEZ1d0_Nf0ROh9Y1kDW1g0_NX<=PCU1g0^NW<`0PCT1h0]NU<b0QCR1j0\\NT<d0QCQ1j0\\NS<e0QCP1j0]NT<c3iC`LU<a3iCaLV<b3fCaLX<b3dC`L[<c3aC`L]<c3_C_L`<d3\\C_Lb<`5O2N1O2N2N2N1O2N2N2N101O1O001O1O1O001O10N1L4M3M3M3M3M3M3M3M4L3O1O1O1O1O1O1O100O2N1O2N2N2N1O2N2O1A`B[Jb=d5=^KnAU3T>hLoAV3S>gLQBV3Q>gLRBW3P>fLSBX3o=eLUBY3k=eLXBY3l=bLWB\\3[>QLiAl3W?N101N1O1O2N1O110O01O10O01O0103M3L4M5K9F;F9G9F;F9G9F:Gh0WOh_Q5"}, "label": "a guy playing wii"}]}
{"id": 375996, "image": "COCO_train2014_000000375996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in glasses playing a video game with two other people\"."}], "task": "refering", "answer_template": "The \"a man in glasses playing a video game with two other people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 268, 269, 270, 271, 272, 273, 274, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 310, 311, 324, 325, 326, 327, 328, 343, 344, 345, 346, 347], "bbox": [0.06359464627151051, 0.305546875, 0.5041491395793499, 0.7975156250000001], "iscrowd": 0, "area": 32053.4441, "rle": {"size": [640, 523], "counts": "fSe02ic05J6J6J6J6J6K5J6M3M3M3N2M3M3M3N3L3M3N2M3M3N2M3M3N200O100O100O100O100O10000O100O100O100000jMP@<P`0Eo_O;i;fNdGo0cL;h;nN^Gg0jL;g;WOVG?TM9e;_OQG8ZM9c;BRG5[M9b;DRG3]M8`;GRG1^M8_;IRGO_M5a;MPGN`M2a;2nFLaMOb;8lFIbMMc;<jFGdMId;a0hFFdMGe;e0fFEdMCi;i0bFDeMBj;k0_FDhM@J]OW;_1WGDiM_OEB\\;[1UGDkM^OAF_;Y1TGCmM]O\\OKc;V1SGBPN[OWO0f;T1eD]Ni1T1fNZOSO4j;Q1bD_Nj1R1hNXOoN:m;m0`DbNk1o0jN0k:`0^DbNm1n0kNOj:b0\\DcNn1l0mNNj:c0XDfNP2i0oNMi:e0VDgNQ2g0QOLi:f0SDiNR2d0TOLi:e0oCmNS2b0WOKj:X1nEmNXOKl:W1kEmN[OKm:V1gEoN\\OKo:T1eEPO^OKP;S1aERO@JQ;S1^EROBKS;Q1ZESOEKT;o0WEVOEKV;n0TEVOHKX;l0oDYOJJ[;j0jD\\OKJ_;g0eD_OLJc;c0`DDNHd;c0]DEOHd;d0\\DD0Hc;f0\\DB1Hc;f0\\DB1Hc;g0[DA2Hc;h0ZD@3Hc;i0YD_O4Hc;i0YD_O4Hc;j0XD^O5Hc;j0XD^O5Hc;k0WD]O6Hb;m0WD[O7Hb;m0WD[O7Ha;o0WDYO9G`;P1WDYO9G_;R1WDVO;H^;R1WDVO;H^;S1VDUO<H];U1VDSO=H];U1VDSO=H\\;W1VDQO>H\\;W1VDQO>H\\;X1UDPO?H[;Y1VDoN?H[;Z1UDnN`0HZ;\\1UDlNa0HZ;\\1UDlNa0HZ;]1TDkNc0GX;_1UDjNc0GW;a1UDhNd0GW;b1TDhNd0FW;c1UDgNd0FV;e1UDfNd0EW;e1UDfNd0DW;h1TDeNd0CW;i1UDdNd0Ce9@oE[2GcNd0Bd9EnEW2IbNe0Bb9ImET2LbNd0Aa9NkEQ2O`Ne0A_92mDI<T2c0aNd0ZOd9m0SEZ1d0_Nf0ROh9Y1kDW1g0_NX<=PCU1g0^NW<`0PCT1h0]NU<b0QCR1j0\\NT<d0QCQ1j0\\NS<e0QCP1j0]NT<c3iC`LU<a3iCaLV<b3fCaLX<b3dC`L[<c3aC`L]<c3_C_L`<d3\\C_Lb<`5O2N1O2N2N2N1O2N2N2N101O1O001O1O1O001O10N1L4M3M3M3M3M3M3M3M4L3O1O1O1O1O1O1O100O2N1O2N2N2N1O2N2O1A`B[Jb=d5=^KnAU3T>hLoAV3S>gLQBV3Q>gLRBW3P>fLSBX3o=eLUBY3k=eLXBY3l=bLWB\\3[>QLiAl3W?N101N1O1O2N1O110O01O10O01O0103M3L4M5K9F;F9G9F;F9G9F:Gh0WOh_Q5"}, "label": "a man in glasses playing a video game with two other people"}]}
{"id": 375996, "image": "COCO_train2014_000000375996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady with brown hair with a remote in her hands\"."}], "task": "refering", "answer_template": "The \"a lady with brown hair with a remote in her hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 187, 201, 202, 203, 204, 205, 206, 207, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 273, 274, 277, 278, 279, 280, 281, 282, 283, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359], "bbox": [0.38124282982791585, 0.30103125, 0.9421223709369024, 0.793109375], "iscrowd": 0, "area": 57121.38799999999, "rle": {"size": [640, 523], "counts": "Yjl31kc05K6H7I7L5J5M3M4K4M3M4M2N2N3M2N2N3M2N2N3M3M4L4L2N2N3M2O10000O100O100O0100O1O1O00100O1O1O1O1O001O1O1O1O1O001O1O1O1O1O2N3M3M3M3M4L5K6J5K5K4LO10000000000O100O100O10000O1O100O1O100O1O1N2N2N2O1N2N2O1N2N2N2M3N2M3N2[KfMaG]2]8eMaG]2]8fM_G]2^8fM_G]2_8fM^G\\2`8fM]G]2a8fM\\G\\2c8eMZG^2f8cMWG_2h8bMUGa2j8aMSGa2l8`MUDBo1Q3l9^MlC1n1c2U:aNZE`NYNR3\\<`NXEP2g:RNVEP2i:RNSEQ2k:QNSEQ2k:RNQEQ2`:_N]Ec1_:aN_Ea1\\:eN`E^1\\:gNnDPM3[4k:jNkDXM1P4o:Z4N2N2N2N2O1O1O100O1O100O1O100O100O100O10O010000O100O100O1000000O1000000O1000000O1000000O1000000O100000000O1000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O001O001O001O1O1O2N1O2N1O2N1O2N1O2N101N1O2N1O2N1O2N1O2N1O2N1O2N1O2N4L3M4L4L3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M2N1O1O2N1O1O1O2N1O1O100O2N1OP1POl1eL_\\b0"}, "label": "a lady with brown hair with a remote in her hands"}]}
{"id": 375996, "image": "COCO_train2014_000000375996.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blond female in the picture\"."}], "task": "refering", "answer_template": "The \"the blond female in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 187, 201, 202, 203, 204, 205, 206, 207, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 273, 274, 277, 278, 279, 280, 281, 282, 283, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359], "bbox": [0.38124282982791585, 0.30103125, 0.9421223709369024, 0.793109375], "iscrowd": 0, "area": 57121.38799999999, "rle": {"size": [640, 523], "counts": "Yjl31kc05K6H7I7L5J5M3M4K4M3M4M2N2N3M2N2N3M2N2N3M3M4L4L2N2N3M2O10000O100O100O0100O1O1O00100O1O1O1O1O001O1O1O1O1O001O1O1O1O1O2N3M3M3M3M4L5K6J5K5K4LO10000000000O100O100O10000O1O100O1O100O1O1N2N2N2O1N2N2O1N2N2N2M3N2M3N2[KfMaG]2]8eMaG]2]8fM_G]2^8fM_G]2_8fM^G\\2`8fM]G]2a8fM\\G\\2c8eMZG^2f8cMWG_2h8bMUGa2j8aMSGa2l8`MUDBo1Q3l9^MlC1n1c2U:aNZE`NYNR3\\<`NXEP2g:RNVEP2i:RNSEQ2k:QNSEQ2k:RNQEQ2`:_N]Ec1_:aN_Ea1\\:eN`E^1\\:gNnDPM3[4k:jNkDXM1P4o:Z4N2N2N2N2O1O1O100O1O100O1O100O100O100O10O010000O100O100O1000000O1000000O1000000O1000000O1000000O100000000O1000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O001O001O001O1O1O2N1O2N1O2N1O2N1O2N101N1O2N1O2N1O2N1O2N1O2N1O2N1O2N4L3M4L4L3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M2N1O1O2N1O1O1O2N1O1O100O2N1OP1POl1eL_\\b0"}, "label": "the blond female in the picture"}]}
{"id": 523455, "image": "COCO_train2014_000000523455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy wearing long gray sleeves and pink shirt in the background\"."}], "task": "refering", "answer_template": "The \"boy wearing long gray sleeves and pink shirt in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 109, 110, 111, 112, 113, 128, 129, 130, 131, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 216, 217, 218, 234], "bbox": [0.00198, 0.39012500000000006, 0.38320000000000004, 0.9777750000000001], "iscrowd": 0, "area": 21726.847100000003, "rle": {"size": [400, 500], "counts": "lg0k0o9e1VOk00O010O010O010O010O010O010O010O010O00010O010O0010O010O01O010O0010O010O01O010O0010O010mLmLhLS3Y3RMbLm2_3XM[Li2f3[MULd2l3]MRLc2o3^MPLb2P4_MnKa2S4`MlK`2T4aMjK_2W4bMhK]2Y4eMdK\\2]4dMbK[2_4fM_K[2a4fM^KY2c4fM]KZ2d4fM\\KZ2d4fM[KZ2f4fMZKY2g4gMXKZ2i4eMWKZ2j4fMUK[2k4eMUKZ2l4fMSKZ2n4hMPKX2P5oMhJQ2Y5WN_Jh1b5_NVJb1k5dNnI[1S6iNhIX1X6iNgIV1Z6jNeIV1\\6jNdIV1[6kNdIU1]6lNbIT1]6mNbIS1_6mNaIR1`6oN^IR1a6oN_IP1b6PO]IP1c6RO\\In0d6RO[In0e6SO[Im0e6SOZIm0g6TOXIk0h6VOWIk0i6UOWIj0i6XOUIh0l6XOTIh0l6XOSIh0m6ZORIf0n6]21O10O0100O00100O00100O2O0O1O2O0O2N100O2O0O2N101N2N2O1N101N2O1N2O1N2O1N2O1N3N3L3M3N3L3N3L3N2M4M2M3N3L3N2M4M2M4M2M3M2O2M3N1N3N1N3N2M2O2M2O2M3N1N3N1N3N1N3M3N1N3N1N3N2M2O2M2O2M2O2M3N1N3N1N3M3Nh]h3"}, "label": "boy wearing long gray sleeves and pink shirt in the background"}]}
{"id": 523455, "image": "COCO_train2014_000000523455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man wearing a baseball cap\"."}], "task": "refering", "answer_template": "The \"a young man wearing a baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 109, 110, 111, 112, 113, 128, 129, 130, 131, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 216, 217, 218, 234], "bbox": [0.00198, 0.39012500000000006, 0.38320000000000004, 0.9777750000000001], "iscrowd": 0, "area": 21726.847100000003, "rle": {"size": [400, 500], "counts": "lg0k0o9e1VOk00O010O010O010O010O010O010O010O010O00010O010O0010O010O01O010O0010O010O01O010O0010O010mLmLhLS3Y3RMbLm2_3XM[Li2f3[MULd2l3]MRLc2o3^MPLb2P4_MnKa2S4`MlK`2T4aMjK_2W4bMhK]2Y4eMdK\\2]4dMbK[2_4fM_K[2a4fM^KY2c4fM]KZ2d4fM\\KZ2d4fM[KZ2f4fMZKY2g4gMXKZ2i4eMWKZ2j4fMUK[2k4eMUKZ2l4fMSKZ2n4hMPKX2P5oMhJQ2Y5WN_Jh1b5_NVJb1k5dNnI[1S6iNhIX1X6iNgIV1Z6jNeIV1\\6jNdIV1[6kNdIU1]6lNbIT1]6mNbIS1_6mNaIR1`6oN^IR1a6oN_IP1b6PO]IP1c6RO\\In0d6RO[In0e6SO[Im0e6SOZIm0g6TOXIk0h6VOWIk0i6UOWIj0i6XOUIh0l6XOTIh0l6XOSIh0m6ZORIf0n6]21O10O0100O00100O00100O2O0O1O2O0O2N100O2O0O2N101N2N2O1N101N2O1N2O1N2O1N2O1N3N3L3M3N3L3N3L3N2M4M2M3N3L3N2M4M2M4M2M3M2O2M3N1N3N1N3N2M2O2M2O2M3N1N3N1N3N1N3M3N1N3N1N3N2M2O2M2O2M2O2M3N1N3N1N3M3Nh]h3"}, "label": "a young man wearing a baseball cap"}]}
{"id": 523455, "image": "COCO_train2014_000000523455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the young person in blue who is on the phone\"."}], "task": "refering", "answer_template": "The \"the young person in blue who is on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 22, 23, 24, 25, 26, 27, 28, 29, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.0018, 0.022475000000000002, 0.99056, 1.0], "iscrowd": 0, "area": 108617.935, "rle": {"size": [400, 500], "counts": "[U12]<1N2O1N2O1N2O1N200O1000000O10000O101O000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000N2N2M3M3N2N2O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100cH]Nl3d1SLdNe3]1ZLkN_3U1_LTOY3m0fL_On2b0PMJf26YMKf26YMKf26XMLg25XMLh24WMNg23WMOh22WMOh22VM0j20UM2i2OUM4i2MUM5k2KTM7j2JTM9j2HTM:k2GTM;k2ESM>k2CSM`0k2ATMa0j2@UMb0i2_OUMd0i2]OVMe0h2\\OWMf0g2[OXMg0f2ZOXMi0f2XOYMi0e2YOZMh0f2XOYMi0f2XOZMi0e2WOZMj0e2WOZMj0f2VOYMk0g2UOYMk0f2VOYMl0f2TOYMo0d2RO[MQ1c2oN\\MT1b2lN^MV1_2kN`MX1^2hNaM[1\\2fNcM]1[2cNeM_1Y2aNfMb1W2_NhMe1U2[NjMh1S2YNmMi1Q2WNmMm1Q2SNnMP2o1QNPNR2n1nMQNU2l1lMRNX2l1hMSN[2k1eMTN^2i1cMVN`2h1`MVNd2h1\\MWNg2f1ZMYNh2f1XMYNk2d1VMZNn2d1RM[NP3d1PM[NS3b1nL]NU3a1kL_NV3`1jL_NY3^1hLaN[3]1eLcN\\3[1eLdN^3Z1bLfN_3X1bLgNa3W1_LhNd3U1]LkNd3T1\\LkNg3R1ZLnNh3P1XLoNj3o0WLPOl3n0TLROn3l0RLSOP4k0QLUOQ4i0oKVOS4h0nKWOU4g0kKYOW4d0jK[OX4d0hK\\OZ4a0gK^O\\4`0dK@]4>dKA_4=aKCa4:`KEb49_KGc46^KJb45_KJc44^KLc42^KMc42^KNc40^K0c4N^K1c4N^K2c4L^K4c4J^K6b4I_K7b4G_K9\\800000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000001O000000001O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O00001O00001O0000001O00001O00001O00001O1O1O1O001O001O00001O001O001O001O001O001O00001O001O002N2N3M2N2N2N3M2N2N2N2N3M2N4L5K6J5K5K4L5dIWLg3n3oK]Lk3g3mKbLn3c3hKhLR4\\3fKkLW4Z3_KkLa4Z3UKlLj4W3oJmLQ5W3fJnLZ5T50000000001O0000000000001O0000000000001O0000000001O0001O00O100000000000000000000001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O5K5K5K4L5K5K5K5K4L5K5K5K5J5L5K5K5JVb1"}, "label": "the young person in blue who is on the phone"}]}
{"id": 523455, "image": "COCO_train2014_000000523455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in a blue hoodie holding a cell phone\"."}], "task": "refering", "answer_template": "The \"person in a blue hoodie holding a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 22, 23, 24, 25, 26, 27, 28, 29, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.0018, 0.022475000000000002, 0.99056, 1.0], "iscrowd": 0, "area": 108617.935, "rle": {"size": [400, 500], "counts": "[U12]<1N2O1N2O1N2O1N200O1000000O10000O101O000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000N2N2M3M3N2N2O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100cH]Nl3d1SLdNe3]1ZLkN_3U1_LTOY3m0fL_On2b0PMJf26YMKf26YMKf26XMLg25XMLh24WMNg23WMOh22WMOh22VM0j20UM2i2OUM4i2MUM5k2KTM7j2JTM9j2HTM:k2GTM;k2ESM>k2CSM`0k2ATMa0j2@UMb0i2_OUMd0i2]OVMe0h2\\OWMf0g2[OXMg0f2ZOXMi0f2XOYMi0e2YOZMh0f2XOYMi0f2XOZMi0e2WOZMj0e2WOZMj0f2VOYMk0g2UOYMk0f2VOYMl0f2TOYMo0d2RO[MQ1c2oN\\MT1b2lN^MV1_2kN`MX1^2hNaM[1\\2fNcM]1[2cNeM_1Y2aNfMb1W2_NhMe1U2[NjMh1S2YNmMi1Q2WNmMm1Q2SNnMP2o1QNPNR2n1nMQNU2l1lMRNX2l1hMSN[2k1eMTN^2i1cMVN`2h1`MVNd2h1\\MWNg2f1ZMYNh2f1XMYNk2d1VMZNn2d1RM[NP3d1PM[NS3b1nL]NU3a1kL_NV3`1jL_NY3^1hLaN[3]1eLcN\\3[1eLdN^3Z1bLfN_3X1bLgNa3W1_LhNd3U1]LkNd3T1\\LkNg3R1ZLnNh3P1XLoNj3o0WLPOl3n0TLROn3l0RLSOP4k0QLUOQ4i0oKVOS4h0nKWOU4g0kKYOW4d0jK[OX4d0hK\\OZ4a0gK^O\\4`0dK@]4>dKA_4=aKCa4:`KEb49_KGc46^KJb45_KJc44^KLc42^KMc42^KNc40^K0c4N^K1c4N^K2c4L^K4c4J^K6b4I_K7b4G_K9\\800000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000001O000000001O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O00001O00001O0000001O00001O00001O00001O1O1O1O001O001O00001O001O001O001O001O001O00001O001O002N2N3M2N2N2N3M2N2N2N2N3M2N4L5K6J5K5K4L5dIWLg3n3oK]Lk3g3mKbLn3c3hKhLR4\\3fKkLW4Z3_KkLa4Z3UKlLj4W3oJmLQ5W3fJnLZ5T50000000001O0000000000001O0000000000001O0000000001O0001O00O100000000000000000000001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O5K5K5K4L5K5K5K5K4L5K5K5K5J5L5K5K5JVb1"}, "label": "person in a blue hoodie holding a cell phone"}]}
{"id": 269504, "image": "COCO_train2014_000000269504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in direct sunlight interacting with a giraffe in the shade\"."}], "task": "refering", "answer_template": "The \"a giraffe in direct sunlight interacting with a giraffe in the shade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 31, 32, 44, 45, 57, 58, 71, 84, 85, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 135, 136, 137, 138, 139, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 174, 175, 176, 178, 187, 188, 189, 191, 192, 200, 201, 202, 204, 205], "bbox": [0.3855733333333333, 0.09072, 0.7961066666666665, 0.87634], "iscrowd": 0, "area": 23920.500200000006, "rle": {"size": [500, 375], "counts": "fVW2`0R?2M3M3M3N2M2]ERO_6Q1aIUOW6m0iIWOP6l0oIXOe5Q1[JQOb40dGY1j3jNX4T2hKoMQ4U2nKnMb03lMQ2b1QN<5iMa1V2`NLa4;dK@Y4c0lKXOR4j0SLPOk3V1WLeNg3a1]LVNa3R2dLbM\\3d2kLPMS3W3SM^Ll2d3\\MSLc2o3eMgK[2Z4mM^KR2d4UNmJQ2U5VN]Jo1d5n35I6I7I7J7I6Jn0RO6RIiGY6]8gIjGl5\\8TJjGa5[8^JlG[5V8dJnGY5Q8gJSHY5i7gJ[HY5h0_Jl47aJ\\59dJS50gJa5JeJ]5JmJd5[OhJf5DSKh5dNQKW6VOZKP7d4PI`KQ7[4oHiKW7n3hHVL^7QO]HX34lMf7[N\\Hi3NPNl7gM\\HY4HTNY8m1cGXN\\8b4N1O1N2O2UNiHUKX7g4fI^J[6^5oIZJT6a5UJYJl5^4UIfKX1Ed5\\4c3F9@`0L5N1O1O1O2N1O1O2N1001O1N101N1O2N2O0O2N101N1O2N2O0O2N101N2N1O2O0O2N2N101N1O2O0O2N2N101N1O2N2O1N2N200000001O0000010O5K5K4gFmJg7X5mGmJU8X5`GlJb8Y5SGkJo8l53M2N2M3N2N2N2mMZFTNh9f1`FVNb9e1fFVN]9ISGYOHj0W9hNVH9jNk0b;POgDk0Z;QOmDk0T;QOTEi0o:QOYEk0m<K4L5K4L5K4LmeT1"}, "label": "a giraffe in direct sunlight interacting with a giraffe in the shade"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza on a plate in front of a woman\"."}], "task": "refering", "answer_template": "The \"a pizza on a plate in front of a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204], "bbox": [0.23608899297423888, 0.45696875, 0.7347072599531617, 0.6000781249999999], "iscrowd": 0, "area": 15284.758450000003, "rle": {"size": [640, 427], "counts": "S^o13mc03L4L6K6I7I7J6I3N1N1O2O0O1O2O0O1O2O0O101N1O2O0O1O2O0O1O2O0O101N101O00001N10001O0O101O001O0O101O00001N1000000O1000001O0O1000000O10001O00000O100000001O0000000000001O00000O100000001O00000000000000O100000O10000000O100000000000000O1000O100001O00000O100000000000001N100000000000000O101O00000O1000000O1000001N100000000O1000000O2O0000000O101O000O1000001N1000000O101N100O1O100O2N101N1O101N1O2O0O2N100O2N101N1O2O0O1O3N3L3N2M4J5J7I6J6JkYV2"}, "label": "a pizza on a plate in front of a woman"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza in front of a woman\"."}], "task": "refering", "answer_template": "The \"a pizza in front of a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204], "bbox": [0.23608899297423888, 0.45696875, 0.7347072599531617, 0.6000781249999999], "iscrowd": 0, "area": 15284.758450000003, "rle": {"size": [640, 427], "counts": "S^o13mc03L4L6K6I7I7J6I3N1N1O2O0O1O2O0O1O2O0O101N1O2O0O1O2O0O1O2O0O101N101O00001N10001O0O101O001O0O101O00001N1000000O1000001O0O1000000O10001O00000O100000001O0000000000001O00000O100000001O00000000000000O100000O10000000O100000000000000O1000O100001O00000O100000000000001N100000000000000O101O00000O1000000O1000001N100000000O1000000O2O0000000O101O000O1000001N1000000O101N100O1O100O2N101N1O101N1O2O0O2N100O2N101N1O2O0O1O3N3L3N2M4J5J7I6J6JkYV2"}, "label": "a pizza in front of a woman"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman sitting directly behind the pizza\"."}], "task": "refering", "answer_template": "The \"the woman sitting directly behind the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 21, 22, 23, 24, 36, 37, 38, 39, 51, 52, 53, 54, 66, 67, 68, 69, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 167, 168, 169, 175, 176, 177, 191], "bbox": [0.13442622950819672, 0.02840625, 0.8569555035128804, 0.5396093750000001], "iscrowd": 0, "area": 39778.006299999994, "rle": {"size": [640, 427], "counts": "`aT15ic06K4K6K2N0O2O0O2O001O10O100O2O0O100O10000O100O100O100O10000O100O100O10000O100O1=C7IO1O100O1O100O1O100O1O2O0O1O100O1O1O100O1O10N1O1O2M2O2M2O2M3N2M2O2M3N2M3N1O2M3kNaNg_Oa1V`0YOP_Oj0n`0T1O2N2N2M3N1O2N2N2N2M2O2N2N2N2M2O2N2N2N1010000000O10000000bMPLlDP4S;TLjDl3T;XLjDh3U;[LiDe3U;_LhDb3W;aLgD_3W;fLfDZ3Y;iLeDW3Y;mLeDS3Z;PMdDP3Z;TMdDl2Z;YMcDg2\\;\\MbDd2\\;`MbD`2];cMaD]2];gMaDY2^;kM_DU2_;oM_DP2a;SN]Dm1a;WN]Di1b;ZN\\Df1b;_N[Da1d;bNZD^1d;hNXDX1g;lNVDT1h;QOUDo0j;VORDj0l;[OQDe0n;_OoCa0o;i3O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N21O1O001O1O1O001O1N2O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O3L7J6J6J6J6J6J6J6J6J6J6J6F7I5J5L5K4K6K5J5L5K5J5013M3L4M3M2M4M3M3L4M3L4M3M3M3O100O100ON3M2N3M2N3N1N3M2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2M3M3L4L4M4K4M3L4M3L4M3L4L3N1N2O1N2O1N3N1N2O1N2N2O1N2O2M2O1L4L4K5L4L4K6KbjU1"}, "label": "the woman sitting directly behind the pizza"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting at a restaurant who is ready to eat a pizza\"."}], "task": "refering", "answer_template": "The \"a woman sitting at a restaurant who is ready to eat a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 21, 22, 23, 24, 36, 37, 38, 39, 51, 52, 53, 54, 66, 67, 68, 69, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 167, 168, 169, 175, 176, 177, 191], "bbox": [0.13442622950819672, 0.02840625, 0.8569555035128804, 0.5396093750000001], "iscrowd": 0, "area": 39778.006299999994, "rle": {"size": [640, 427], "counts": "`aT15ic06K4K6K2N0O2O0O2O001O10O100O2O0O100O10000O100O100O100O10000O100O100O10000O100O1=C7IO1O100O1O100O1O100O1O2O0O1O100O1O1O100O1O10N1O1O2M2O2M2O2M3N2M2O2M3N2M3N1O2M3kNaNg_Oa1V`0YOP_Oj0n`0T1O2N2N2M3N1O2N2N2N2M2O2N2N2N2M2O2N2N2N1010000000O10000000bMPLlDP4S;TLjDl3T;XLjDh3U;[LiDe3U;_LhDb3W;aLgD_3W;fLfDZ3Y;iLeDW3Y;mLeDS3Z;PMdDP3Z;TMdDl2Z;YMcDg2\\;\\MbDd2\\;`MbD`2];cMaD]2];gMaDY2^;kM_DU2_;oM_DP2a;SN]Dm1a;WN]Di1b;ZN\\Df1b;_N[Da1d;bNZD^1d;hNXDX1g;lNVDT1h;QOUDo0j;VORDj0l;[OQDe0n;_OoCa0o;i3O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N21O1O001O1O1O001O1N2O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O3L7J6J6J6J6J6J6J6J6J6J6J6F7I5J5L5K4K6K5J5L5K5J5013M3L4M3M2M4M3M3L4M3L4M3M3M3O100O100ON3M2N3M2N3N1N3M2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2M3M3L4L4M4K4M3L4M3L4M3L4L3N1N2O1N2O1N3N1N2O1N2N2O1N2O2M2O1L4L4K5L4L4K6KbjU1"}, "label": "a woman sitting at a restaurant who is ready to eat a pizza"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a man leaning to his right\"."}], "task": "refering", "answer_template": "The \"the back of a man leaning to his right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 60, 61, 75, 76, 90, 91, 92, 106, 107, 121, 122, 123, 137], "bbox": [0.0034660421545667447, 0.131453125, 0.2159016393442623, 0.397296875], "iscrowd": 0, "area": 6177.79985, "rle": {"size": [640, 427], "counts": "og0]1db0O0000001O0000001O000000O100iNW100O1O100O1O10000O10000O10101N2N101N2N2O0O2N2O2M4L4L4M3MRNo_OOn?NU@5i?FZ@>c?_O`@d0]?XOf@k0X?ROk@Q1S?jNPAZ1m>cNVA`1g>^NZAf1d>VN_Am1^>PNeAS2X>kMkAS2]>cMgA[2Q`0O1N1O2N2N2O1N2N3M4M3L3M4L4L3M4K5L2N3M2M4M2N0O10000N1O2O1N2N101N2M3M5KQUa6"}, "label": "the back of a man leaning to his right"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in the background looking away\"."}], "task": "refering", "answer_template": "The \"a person in the background looking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 60, 61, 75, 76, 90, 91, 92, 106, 107, 121, 122, 123, 137], "bbox": [0.0034660421545667447, 0.131453125, 0.2159016393442623, 0.397296875], "iscrowd": 0, "area": 6177.79985, "rle": {"size": [640, 427], "counts": "og0]1db0O0000001O0000001O000000O100iNW100O1O100O1O10000O10000O10101N2N101N2N2O0O2N2O2M4L4L4M3MRNo_OOn?NU@5i?FZ@>c?_O`@d0]?XOf@k0X?ROk@Q1S?jNPAZ1m>cNVA`1g>^NZAf1d>VN_Am1^>PNeAS2X>kMkAS2]>cMgA[2Q`0O1N1O2N2N2O1N2N3M4M3L3M4L4L3M4K5L2N3M2M4M2N0O10000N1O2O1N2N101N2M3M5KQUa6"}, "label": "a person in the background looking away"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with bacon and spinach\"."}], "task": "refering", "answer_template": "The \"a pizza with bacon and spinach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.04585480093676814, 0.614125, 0.9310304449648712, 0.9858750000000001], "iscrowd": 0, "area": 73012.10045, "rle": {"size": [640, 427], "counts": "QQ=5hc08I6J6J6I7J7I6J6J6I8I6J6J6I7J7I6J6J6I6K2N2N2M3N2N2N2N2M3N2N2N2M3N2O1O1O100O1O101N1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000001O000000001O0000001O0001O0001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O00001O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2M2L5L3M3L5L3L4M3M4K4M3M4K4M4L5J5L5J5L5K4K6K4L:E<E:E?B>B>Aaba0"}, "label": "a pizza with bacon and spinach"}]}
{"id": 507073, "image": "COCO_train2014_000000507073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza on plate at bottom of picture\"."}], "task": "refering", "answer_template": "The \"pizza on plate at bottom of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.04585480093676814, 0.614125, 0.9310304449648712, 0.9858750000000001], "iscrowd": 0, "area": 73012.10045, "rle": {"size": [640, 427], "counts": "QQ=5hc08I6J6J6I7J7I6J6J6I8I6J6J6I7J7I6J6J6I6K2N2N2M3N2N2N2N2M3N2N2N2M3N2O1O1O100O1O101N1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000001O000000001O0000001O0001O0001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O00001O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2M2L5L3M3L5L3L4M3M4K4M3M4K4M4L5J5L5J5L5K4K6K4L:E<E:E?B>B>Aaba0"}, "label": "pizza on plate at bottom of picture"}]}
{"id": 105666, "image": "COCO_train2014_000000105666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sweet and savory dessert next to three donut balls\"."}], "task": "refering", "answer_template": "The \"a sweet and savory dessert next to three donut balls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248], "bbox": [0.43792187499999996, 0.19295039164490863, 0.8702031250000001, 0.7600000000000001], "iscrowd": 0, "area": 46151.39899999998, "rle": {"size": [383, 640], "counts": "QmX34h;=D<C>C<C=D4K4M2M4M2M4M2M4M3N1O2M2O2N1N3N2N1N3N1O2M2O2N2M2O2M2O2N1N3O1N101N101N1O2O1N101N101N1O2O1N101N101N1O2O0O100O2O0O101N100O101N100O101N10000O2O0O101N1N2N2N3M2N2N2N3M2N2N2O2O0O101O0O10001N10000O2O0O100010O1O10O01O1O010O1O1O01N100O101O0O101N100O101O00000O2O0000001O0000000000000000O100000001O001O001O001O001O0O10000000000000000000O100O1000O10001O1O1O1O001O1O1O1O001O1O0O2O001O1O001O001O001O00001O0O2O00001O00001O001O1O001O1O0O2O1N101N2O1O0O2O1N101N2O0O2O1N2O1O1N2O1N2O1N2O1N2O5J2O1N2O2N1N2O1N2O2M2O1N2O2M2O1O1N2O2M2O1N2O1N3N1N2O1O2K4K5K5J6K6J:Ei0XOg0YOljn0"}, "label": "a sweet and savory dessert next to three donut balls"}]}
{"id": 105666, "image": "COCO_train2014_000000105666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut\"."}], "task": "refering", "answer_template": "The \"donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248], "bbox": [0.43792187499999996, 0.19295039164490863, 0.8702031250000001, 0.7600000000000001], "iscrowd": 0, "area": 46151.39899999998, "rle": {"size": [383, 640], "counts": "QmX34h;=D<C>C<C=D4K4M2M4M2M4M2M4M3N1O2M2O2N1N3N2N1N3N1O2M2O2N2M2O2M2O2N1N3O1N101N101N1O2O1N101N101N1O2O1N101N101N1O2O0O100O2O0O101N100O101N100O101N10000O2O0O101N1N2N2N3M2N2N2N3M2N2N2O2O0O101O0O10001N10000O2O0O100010O1O10O01O1O010O1O1O01N100O101O0O101N100O101O00000O2O0000001O0000000000000000O100000001O001O001O001O001O0O10000000000000000000O100O1000O10001O1O1O1O001O1O1O1O001O1O0O2O001O1O001O001O001O00001O0O2O00001O00001O001O1O001O1O0O2O1N101N2O1O0O2O1N101N2O0O2O1N2O1O1N2O1N2O1N2O1N2O5J2O1N2O2N1N2O1N2O2M2O1N2O2M2O1O1N2O2M2O1N2O1N3N1N2O1O2K4K5K5J6K6J:Ei0XOg0YOljn0"}, "label": "donut"}]}
{"id": 113857, "image": "COCO_train2014_000000113857.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady tennis player without a hat\"."}], "task": "refering", "answer_template": "The \"the lady tennis player without a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 100, 101, 116, 117, 118, 119, 136, 137, 154, 155, 172, 173, 190, 191, 192, 208, 209, 210], "bbox": [0.45095999999999997, 0.3859159159159159, 0.6854199999999999, 0.9912912912912913], "iscrowd": 0, "area": 7768.46805, "rle": {"size": [333, 500], "counts": "UeY22[:2O0O2N2N2N2N10O0001O001O01O01O00001O010O00001O00010O00001O001O00001O00001O000000O100O100O01O0000001N10000000000UOHbG8Z82`GNn3]O;f0eKOn3]O;f0eKOn3]O<e0cK1P4ZO=c1A_N>b1@`N?a1cLZNh17c1`1PL\\Oo1UOP2Z3fMhLY2\\3bMeL^2`3\\M`Le2d3VM]Lj2g3RMYLn2l3lLULT3o3gLQLZ3a5000O101O1O1N2O1O1O1N2O1O1dNaKhL`4Y3hK\\LZ4e3oKPLR4Q4\\LeK_3[4_LgKa3\\4ZLeKg38\\Ke3g0ULn3OeKk39XLh4i3SKXLQ5h3iJ[LZ5e3aJ\\Li5e1QJD2hN[6T1hId0[6WOiIf0\\6TOhIj0[6QOiIl0[6oNiIn0[6lNjIP1[6kNiIQ1\\6jNhIR1m7M2M3M4L3N2M3M2N3N2M2N]hb1"}, "label": "the lady tennis player without a hat"}]}
{"id": 113857, "image": "COCO_train2014_000000113857.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a white tank top shaking hands with one wearing a cap\"."}], "task": "refering", "answer_template": "The \"a woman wearing a white tank top shaking hands with one wearing a cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 100, 101, 116, 117, 118, 119, 136, 137, 154, 155, 172, 173, 190, 191, 192, 208, 209, 210], "bbox": [0.45095999999999997, 0.3859159159159159, 0.6854199999999999, 0.9912912912912913], "iscrowd": 0, "area": 7768.46805, "rle": {"size": [333, 500], "counts": "UeY22[:2O0O2N2N2N2N10O0001O001O01O01O00001O010O00001O00010O00001O001O00001O00001O000000O100O100O01O0000001N10000000000UOHbG8Z82`GNn3]O;f0eKOn3]O;f0eKOn3]O<e0cK1P4ZO=c1A_N>b1@`N?a1cLZNh17c1`1PL\\Oo1UOP2Z3fMhLY2\\3bMeL^2`3\\M`Le2d3VM]Lj2g3RMYLn2l3lLULT3o3gLQLZ3a5000O101O1O1N2O1O1O1N2O1O1dNaKhL`4Y3hK\\LZ4e3oKPLR4Q4\\LeK_3[4_LgKa3\\4ZLeKg38\\Ke3g0ULn3OeKk39XLh4i3SKXLQ5h3iJ[LZ5e3aJ\\Li5e1QJD2hN[6T1hId0[6WOiIf0\\6TOhIj0[6QOiIl0[6oNiIn0[6lNjIP1[6kNiIQ1\\6jNhIR1m7M2M3M4L3N2M3M2N3N2M2N]hb1"}, "label": "a woman wearing a white tank top shaking hands with one wearing a cap"}]}
{"id": 113857, "image": "COCO_train2014_000000113857.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the women is hand shaking\"."}], "task": "refering", "answer_template": "The \"the women is hand shaking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 95, 96, 113, 114, 115, 116, 131, 132, 149, 150, 167, 168, 185, 186, 187, 205], "bbox": [0.29204, 0.35576576576576574, 0.47374, 0.9257957957957957], "iscrowd": 0, "area": 5521.1793, "rle": {"size": [333, 500], "counts": "ng_11[:2N101UMMYK5a40^K1^44`KN]45aKKW3OTK9J_OR1:j3:QK1IKT1JT49mJ4C7U1]O[48jJa1d0WNWOFP5a0SKj1a0oM[OGQ5`0QKl1b0mM]OHo4?oJo1d0kM^OGo4>PKQ2a0jM@Hn4>PKQ2a0iMBIk4>PKR2b0hMCHj4?PKR2b0gMDJh4=RKS2a0fMFJf4>RKS2a0eMHKc4>SKS2a0dMKJa4?RKU2a0aMNHa4`3`KhL1F_4c3_KgLX5Z3fJfL[5[3cJeL]5^3`JbLa5_3]JaLd5`3ZJ`Lg5b3VJ^Lk5f0TJh1ObMn5e0UJU2l5iMXJT2j5jMXJT2i5kMZJR2P1SMX3i0kK1A[1U1lM`3g0kK2EW1j0UNe3b0nK1GS10PO[4LoK1Jm0MVO[4KPL1Kj0I\\O[4IRL1Me0GAZ4HTL2Na0CG[4EVL2O=ALZ4WO^KIR1g0G:_OO[4UOgL>E9YO5[4ROjL=H8SO9[4QOmL:K7mN?\\4nNnL9N7hNb0T6VOZK1bN7HN\\6JgL4Y3LiL2W3NkL1S30nLOR31oLNQ31RMMm24TMKl25UMJk26WMHi28XMGh29YMFg2:f3O0100000000000O100000000000O10O1001O1O2N1O0000O5L1N2O1N2OR[e2"}, "label": "the women is hand shaking"}]}
{"id": 113857, "image": "COCO_train2014_000000113857.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female tennis player , wearing a visor and holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a female tennis player , wearing a visor and holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 95, 96, 113, 114, 115, 116, 131, 132, 149, 150, 167, 168, 185, 186, 187, 205], "bbox": [0.29204, 0.35576576576576574, 0.47374, 0.9257957957957957], "iscrowd": 0, "area": 5521.1793, "rle": {"size": [333, 500], "counts": "ng_11[:2N101UMMYK5a40^K1^44`KN]45aKKW3OTK9J_OR1:j3:QK1IKT1JT49mJ4C7U1]O[48jJa1d0WNWOFP5a0SKj1a0oM[OGQ5`0QKl1b0mM]OHo4?oJo1d0kM^OGo4>PKQ2a0jM@Hn4>PKQ2a0iMBIk4>PKR2b0hMCHj4?PKR2b0gMDJh4=RKS2a0fMFJf4>RKS2a0eMHKc4>SKS2a0dMKJa4?RKU2a0aMNHa4`3`KhL1F_4c3_KgLX5Z3fJfL[5[3cJeL]5^3`JbLa5_3]JaLd5`3ZJ`Lg5b3VJ^Lk5f0TJh1ObMn5e0UJU2l5iMXJT2j5jMXJT2i5kMZJR2P1SMX3i0kK1A[1U1lM`3g0kK2EW1j0UNe3b0nK1GS10PO[4LoK1Jm0MVO[4KPL1Kj0I\\O[4IRL1Me0GAZ4HTL2Na0CG[4EVL2O=ALZ4WO^KIR1g0G:_OO[4UOgL>E9YO5[4ROjL=H8SO9[4QOmL:K7mN?\\4nNnL9N7hNb0T6VOZK1bN7HN\\6JgL4Y3LiL2W3NkL1S30nLOR31oLNQ31RMMm24TMKl25UMJk26WMHi28XMGh29YMFg2:f3O0100000000000O100000000000O10O1001O1O2N1O0000O5L1N2O1N2OR[e2"}, "label": "a female tennis player , wearing a visor and holding a tennis racket"}]}
{"id": 138436, "image": "COCO_train2014_000000138436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding scissors and cutting hair\"."}], "task": "refering", "answer_template": "The \"a hand holding scissors and cutting hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.106203125, 0.4022535211267606, 1.0, 0.9842723004694836], "iscrowd": 0, "area": 83938.46144999997, "rle": {"size": [426, 640], "counts": "bSm01X=1O100O1O1O1O1O100O1O1O1O1O101N1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O2N1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1000000000000O100000000000O100O00100O100O100O100O010O100O100O100O10O0100O1O100O100O10O0101N100O100O100O100O101O01O1O1O1O1O1O1O100O1O1O1O1O0000O1O1O2O0O1O100O1O1O100O2N1O100O1O1O100O1O1O2O0O1O1O100O1O1gN`IgKa6Y4mIYKT6e4[JlJf5T5X1O1O100O100O100O1O100O100O100O1O100O100O100O00100O100O100O1O100O100O100O1O100O100O100O1000000000000000000000000000000000O1000O100000000000000000001O001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O010O00001O00001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O010O1O1N2N2N1N3N2N2N2N1N3N2N2N2M9H:F]H"}, "label": "a hand holding scissors and cutting hair"}]}
{"id": 138436, "image": "COCO_train2014_000000138436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hand holding scissors cutting brown hair\"."}], "task": "refering", "answer_template": "The \"hand holding scissors cutting brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.106203125, 0.4022535211267606, 1.0, 0.9842723004694836], "iscrowd": 0, "area": 83938.46144999997, "rle": {"size": [426, 640], "counts": "bSm01X=1O100O1O1O1O1O100O1O1O1O1O101N1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O2N1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1000000000000O100000000000O100O00100O100O100O100O010O100O100O100O10O0100O1O100O100O10O0101N100O100O100O100O101O01O1O1O1O1O1O1O100O1O1O1O1O0000O1O1O2O0O1O100O1O1O100O2N1O100O1O1O100O1O1O2O0O1O1O100O1O1gN`IgKa6Y4mIYKT6e4[JlJf5T5X1O1O100O100O100O1O100O100O100O1O100O100O100O00100O100O100O1O100O100O100O1O100O100O100O1000000000000000000000000000000000O1000O100000000000000000001O001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O010O00001O00001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O010O1O1N2N2N1N3N2N2N2N1N3N2N2N2M9H:F]H"}, "label": "hand holding scissors cutting brown hair"}]}
{"id": 564421, "image": "COCO_train2014_000000564421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis player\"."}], "task": "refering", "answer_template": "The \"the tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 55, 56, 57, 58, 59, 63, 64, 65, 68, 69, 70, 71, 72, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 135, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 287, 288, 289, 290, 291, 292, 301, 302, 303, 304, 305, 306, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335], "bbox": [0.05427230046948357, 0.103265625, 0.9970422535211267, 0.985890625], "iscrowd": 0, "area": 72308.47150000004, "rle": {"size": [640, 426], "counts": "Wd>2jc07J7H7J7J4K6J5M4M2M4M3L5L3N2M4M2M3N3M2M3N3L3iG\\MmNf2f0dMZO`28jMIX2LPN5R2_OWNa0l1TO\\Nm0f1hNbNX1`1`NfNa1Z1YNkNg1V1SNoNn1Q1lMTOW2j0dMZOP33jL2^3G_L;d3C[L=h3AWL?k3@SLa0P4]OnKd0U4XOiKk0Z4QOcKS1_4jN_KY1d4cNZK`1i4\\NTKh1o4TNoJo1S5nMkJU2S5lMkJW2T5iMiJ[2U5fMiJ]2V5cMhJ`2V5aMgJc2X5]MfJg2W5ZMgJi2X5WMfJl2V4XNgKk1i3dNUL_1j3aNTLb1k3^NSLe1l3[NSLg1k3ZNSLi1l3WNRLl1m3TNRLn1l3SNRLP2m3PNQLg1]LYKa7P3QLg1bLWK\\7R3PLi1gLSKW7U3PLi1mLPKR7W3oKj1RMnJn6X3oKj1e4VNYKl1g4YNRKh1P5_NfJb1[5gNYJ[1i5lNmIU1U6SO`Io0a6YOSIi0h4gKjJh3DR1`5aKeJo9Y5[FaJd9]5eF\\JZ9HnDZ5R2hJn8L]Eo4^>D9G6J400O1O1O100O1O00100O1O100O1O1O100O1O100O100001O0O10001O0000001O0000001O0000001O000O101O0CRLh@n3i0QLY=4lAk3j0WLT=1PBi3j0]Lo<MUBf3k0cLj<JYBc3l0hLf<j3XCWLh<k3VCULj<l3UCTLj<n3UCRLk<o3TCQLl<Q4RCPLm<Q4QCPLo<Q4PCPLo<Q4PCPLo<R4oBoKo<S4PCmKP=T4oBmKP=T4nBnKQ=T4lBnKS=S4kBoKS=S4jBoKU=S4iBoKU=_5M3N1O2M3N2N2M2O2N2M3N2N1O2O1N2N1O2O1N2N2O0O2N2N2O1N1O2`FjHU6W7hIQIQ6Q7lIXIl5i6SJ^If5d6WJeIb5\\6[JmI]5U6aJRJX5o5fJZJS5g5jJfJj4[5TKRK`4P5^K\\KV4e4hKhKm3Y4PLPLh3Q4WLULc3l3[L\\L_3e3^LbL\\3_3cLhLW3Y3iLlLQ3T3oLQMl2o2UMVMe2k2ZMZM`2g2`M_MZ2b2fMbMU2^2kMfMQ2Z2PNiMl1X2SNlMi1T2XNoMd1R2[NSN_1n1bNUNZ1l1eNXNW1h1iN\\NS1d1nN_Nn0b1QOcNj0]1WOfNc0]1\\OgN1h10[N^OS2b0f80O010O00010O010O010O0010O0010O0010O010O01O01O010O010O01O010O010O00010O010O0010O010O0010O010O010O010O010O01O010O010O10O010O010O010O01O010O10O010O010O010O010O01O010O10O010O010O010O010O01O010O10O010O010O010O010O01O10O010O010O010O010O010O00001O001O001O001O001O001O01O010O100O100O100O010O100O1O2O1N101N2O0O2O1N101N2O0O2N2O0O2O1N101O0001O001O00001O00001O001O00001O02N3N2M2O2M3MQa0"}, "label": "the tennis player"}]}
{"id": 564421, "image": "COCO_train2014_000000564421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman about to hit the ball\"."}], "task": "refering", "answer_template": "The \"the woman about to hit the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 55, 56, 57, 58, 59, 63, 64, 65, 68, 69, 70, 71, 72, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 135, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 228, 229, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 287, 288, 289, 290, 291, 292, 301, 302, 303, 304, 305, 306, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335], "bbox": [0.05427230046948357, 0.103265625, 0.9970422535211267, 0.985890625], "iscrowd": 0, "area": 72308.47150000004, "rle": {"size": [640, 426], "counts": "Wd>2jc07J7H7J7J4K6J5M4M2M4M3L5L3N2M4M2M3N3M2M3N3L3iG\\MmNf2f0dMZO`28jMIX2LPN5R2_OWNa0l1TO\\Nm0f1hNbNX1`1`NfNa1Z1YNkNg1V1SNoNn1Q1lMTOW2j0dMZOP33jL2^3G_L;d3C[L=h3AWL?k3@SLa0P4]OnKd0U4XOiKk0Z4QOcKS1_4jN_KY1d4cNZK`1i4\\NTKh1o4TNoJo1S5nMkJU2S5lMkJW2T5iMiJ[2U5fMiJ]2V5cMhJ`2V5aMgJc2X5]MfJg2W5ZMgJi2X5WMfJl2V4XNgKk1i3dNUL_1j3aNTLb1k3^NSLe1l3[NSLg1k3ZNSLi1l3WNRLl1m3TNRLn1l3SNRLP2m3PNQLg1]LYKa7P3QLg1bLWK\\7R3PLi1gLSKW7U3PLi1mLPKR7W3oKj1RMnJn6X3oKj1e4VNYKl1g4YNRKh1P5_NfJb1[5gNYJ[1i5lNmIU1U6SO`Io0a6YOSIi0h4gKjJh3DR1`5aKeJo9Y5[FaJd9]5eF\\JZ9HnDZ5R2hJn8L]Eo4^>D9G6J400O1O1O100O1O00100O1O100O1O1O100O1O100O100001O0O10001O0000001O0000001O0000001O000O101O0CRLh@n3i0QLY=4lAk3j0WLT=1PBi3j0]Lo<MUBf3k0cLj<JYBc3l0hLf<j3XCWLh<k3VCULj<l3UCTLj<n3UCRLk<o3TCQLl<Q4RCPLm<Q4QCPLo<Q4PCPLo<Q4PCPLo<R4oBoKo<S4PCmKP=T4oBmKP=T4nBnKQ=T4lBnKS=S4kBoKS=S4jBoKU=S4iBoKU=_5M3N1O2M3N2N2M2O2N2M3N2N1O2O1N2N1O2O1N2N2O0O2N2N2O1N1O2`FjHU6W7hIQIQ6Q7lIXIl5i6SJ^If5d6WJeIb5\\6[JmI]5U6aJRJX5o5fJZJS5g5jJfJj4[5TKRK`4P5^K\\KV4e4hKhKm3Y4PLPLh3Q4WLULc3l3[L\\L_3e3^LbL\\3_3cLhLW3Y3iLlLQ3T3oLQMl2o2UMVMe2k2ZMZM`2g2`M_MZ2b2fMbMU2^2kMfMQ2Z2PNiMl1X2SNlMi1T2XNoMd1R2[NSN_1n1bNUNZ1l1eNXNW1h1iN\\NS1d1nN_Nn0b1QOcNj0]1WOfNc0]1\\OgN1h10[N^OS2b0f80O010O00010O010O010O0010O0010O0010O010O01O01O010O010O01O010O010O00010O010O0010O010O0010O010O010O010O010O01O010O010O10O010O010O010O01O010O10O010O010O010O010O01O010O10O010O010O010O010O01O010O10O010O010O010O010O01O10O010O010O010O010O010O00001O001O001O001O001O001O01O010O100O100O100O010O100O1O2O1N101N2O0O2O1N101N2O0O2N2O0O2O1N101O0001O001O00001O00001O001O00001O02N3N2M2O2M3MQa0"}, "label": "the woman about to hit the ball"}]}
{"id": 31943, "image": "COCO_train2014_000000031943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a black over shirt and black jeans standing behind the fence and other people\"."}], "task": "refering", "answer_template": "The \"a woman wearing a black over shirt and black jeans standing behind the fence and other people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 151, 152, 153, 175, 176, 177, 198, 199, 200, 222, 245, 268], "bbox": [0.593265625, 0.257375, 0.726984375, 0.6733541666666666], "iscrowd": 0, "area": 6233.4953, "rle": {"size": [480, 640], "counts": "QYb5`0^>7J2M2O2M3M3N1N3N2M3M2O2M3M2O2M3N2M2N3N2M3mCnMV;R2iDQNT;Q2iDTNT;m1iDXNT;h1jD]NS;d1jD`NS;b1kDbNR;_1kDeNS;[1kDhNT;Y1iDjNU;a201O1O001O1O001OcMXEg1c9UNkFe1Q9XNXGf1d8VNbGl1Z8RNiGP2T8nMPHT2l7kMVHX2f7fM^H\\2^7cMdH`2Z7\\MjHe2T7YMnHf2T7VMPIi2Q7SMRIl2o6RMSIm2o6oLSIQ3o6lLRIT3o6iLSIX3n6eLSI[3Y9SOQEiMo:S2Q1L4K5L5K4L4L4L4N2N2N2N2N3M2N3M2N3M2O2M2N3M2N3N1N3M3MRja2"}, "label": "a woman wearing a black over shirt and black jeans standing behind the fence and other people"}]}
{"id": 31943, "image": "COCO_train2014_000000031943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"partial view of a woman wearing black and white\"."}], "task": "refering", "answer_template": "The \"partial view of a woman wearing black and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 151, 152, 153, 175, 176, 177, 198, 199, 200, 222, 245, 268], "bbox": [0.593265625, 0.257375, 0.726984375, 0.6733541666666666], "iscrowd": 0, "area": 6233.4953, "rle": {"size": [480, 640], "counts": "QYb5`0^>7J2M2O2M3M3N1N3N2M3M2O2M3M2O2M3N2M2N3N2M3mCnMV;R2iDQNT;Q2iDTNT;m1iDXNT;h1jD]NS;d1jD`NS;b1kDbNR;_1kDeNS;[1kDhNT;Y1iDjNU;a201O1O001O1O001OcMXEg1c9UNkFe1Q9XNXGf1d8VNbGl1Z8RNiGP2T8nMPHT2l7kMVHX2f7fM^H\\2^7cMdH`2Z7\\MjHe2T7YMnHf2T7VMPIi2Q7SMRIl2o6RMSIm2o6oLSIQ3o6lLRIT3o6iLSIX3n6eLSI[3Y9SOQEiMo:S2Q1L4K5L5K4L4L4L4N2N2N2N2N3M2N3M2N3M2O2M2N3M2N3N1N3M3MRja2"}, "label": "partial view of a woman wearing black and white"}]}
{"id": 31943, "image": "COCO_train2014_000000031943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in blue sleveless dress is giving food to brown horse\"."}], "task": "refering", "answer_template": "The \"a girl in blue sleveless dress is giving food to brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 241, 242, 243, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387, 388], "bbox": [0.47325, 0.09079166666666666, 0.911234375, 1.0], "iscrowd": 0, "area": 56243.299050000016, "rle": {"size": [480, 640], "counts": "RZ^42n>2N3M2N2N3M2N2N3M2N2O2O0O1O01O000O2O00CXOaBg0_=ZObBe0]=]OaBd0^=\\OcBd0\\=]OcBc0]=]OcBd0\\=\\OdBe0[=\\OeBc0[=]OeBd0Y=^OfBc0Y=]OgBd0X=\\OhBd0X=]OhBc0W=]OiBd0V=]OiBc0W=]OiBd0V=]OjBc0U=]OkBc0U=^OjBc0V=\\OhBf0X=[OeBg0[=YOcBi0]=<100000000O2O00000O10001O000O2O00001O0O101O00001N10001O000O2O01O010jBmN]<T1`CnN`<R1]CQOd<o0YCSOg<n0VCTOk<k0SCWOn<i0nBZOS=f0jB[OX=V11O1O1O1O1N101O1O1O1O1O1N2O1O001O1O1N2O1O100000O100000000N101O1O1N2O1O1N2aHROj0o0SOWOi0j0UOZOh0g0VO^Oe0d0YOAc0`0[OEa0;]OK?6_OO=2A3<1^O2`0;ROHm0e0dN]O[1P1VNTOg1[1jMgNR2i1`MZNU1lMgLU4k1QNZ1SNcLQ4o1oMZ1YN]Ln3T2lM[1_NRLR4^2bM\\1c3`NaLZ1e3bN_LY1f3cN^LW1h3eN\\LV1g3gN^LS1e3lN`Ln0b3ROaLh0b3WOcLc0_3]OdL=_3BfL8\\3GhL5Y3KkLOW30lLMU33kLLV33kLKW35iLJX35iLIY37gLHZ37gLHZ38fLF\\39fLE[3:fLD\\3<dLC]3<dLC]3=cLA_3>bLA_3?aL_Oa3`0`L_Oa3a0_L^Ob3a0_L]Oc3c0^L[Oc3d0^L[Oc3d0^LZOd3f0\\LYOe3g0[LWOg3i0YLVOh3j0XLUOi3k0XLSOi3n0VLQOk3o0ULPOl3P1TLoNm3Q1SLoNm3R1RLmNo3S1QLlNP4T1QLjNP4V1PLjNP4W1l40000001O0000001O0000001O0000001O0000001O0000001O0000001O001O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1YK^MnLc2o2bMnL_2P3dMnL]2o2hMnLZ2n2jMPMW2m2nMPMT2m2oMQMS2k2SNQMo1k2XNPMj1i2`NRMb1g2gNUM[1d2oNWMS1b2SOZMR1WOfLc0_23n0XOgL=a27m0YOgL6d2=i0ZOgL1f2a0h0[OfLKi2g0e0[OfLFl2j0c0]OeL@o2P1`0]OeL[OQ3W1<ZOiLUOR3_18YOjLPOU3d14YOkLjNX3k10XOlLdNk3b1\\OG]26fMG^25eMH^25eMH_24dMHa25`MIc24`MId23_MIf23]MJf22^MJg22\\MKg22\\MJi22ZMKj20ZMMi20^_j0"}, "label": "a girl in blue sleveless dress is giving food to brown horse"}]}
{"id": 31943, "image": "COCO_train2014_000000031943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl feeding a horse\"."}], "task": "refering", "answer_template": "The \"a girl feeding a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 178, 179, 180, 181, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 241, 242, 243, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342, 360, 361, 362, 363, 364, 365, 383, 384, 385, 386, 387, 388], "bbox": [0.47325, 0.09079166666666666, 0.911234375, 1.0], "iscrowd": 0, "area": 56243.299050000016, "rle": {"size": [480, 640], "counts": "RZ^42n>2N3M2N2N3M2N2N3M2N2O2O0O1O01O000O2O00CXOaBg0_=ZObBe0]=]OaBd0^=\\OcBd0\\=]OcBc0]=]OcBd0\\=\\OdBe0[=\\OeBc0[=]OeBd0Y=^OfBc0Y=]OgBd0X=\\OhBd0X=]OhBc0W=]OiBd0V=]OiBc0W=]OiBd0V=]OjBc0U=]OkBc0U=^OjBc0V=\\OhBf0X=[OeBg0[=YOcBi0]=<100000000O2O00000O10001O000O2O00001O0O101O00001N10001O000O2O01O010jBmN]<T1`CnN`<R1]CQOd<o0YCSOg<n0VCTOk<k0SCWOn<i0nBZOS=f0jB[OX=V11O1O1O1O1N101O1O1O1O1O1N2O1O001O1O1N2O1O100000O100000000N101O1O1N2O1O1N2aHROj0o0SOWOi0j0UOZOh0g0VO^Oe0d0YOAc0`0[OEa0;]OK?6_OO=2A3<1^O2`0;ROHm0e0dN]O[1P1VNTOg1[1jMgNR2i1`MZNU1lMgLU4k1QNZ1SNcLQ4o1oMZ1YN]Ln3T2lM[1_NRLR4^2bM\\1c3`NaLZ1e3bN_LY1f3cN^LW1h3eN\\LV1g3gN^LS1e3lN`Ln0b3ROaLh0b3WOcLc0_3]OdL=_3BfL8\\3GhL5Y3KkLOW30lLMU33kLLV33kLKW35iLJX35iLIY37gLHZ37gLHZ38fLF\\39fLE[3:fLD\\3<dLC]3<dLC]3=cLA_3>bLA_3?aL_Oa3`0`L_Oa3a0_L^Ob3a0_L]Oc3c0^L[Oc3d0^L[Oc3d0^LZOd3f0\\LYOe3g0[LWOg3i0YLVOh3j0XLUOi3k0XLSOi3n0VLQOk3o0ULPOl3P1TLoNm3Q1SLoNm3R1RLmNo3S1QLlNP4T1QLjNP4V1PLjNP4W1l40000001O0000001O0000001O0000001O0000001O0000001O0000001O001O1O2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1YK^MnLc2o2bMnL_2P3dMnL]2o2hMnLZ2n2jMPMW2m2nMPMT2m2oMQMS2k2SNQMo1k2XNPMj1i2`NRMb1g2gNUM[1d2oNWMS1b2SOZMR1WOfLc0_23n0XOgL=a27m0YOgL6d2=i0ZOgL1f2a0h0[OfLKi2g0e0[OfLFl2j0c0]OeL@o2P1`0]OeL[OQ3W1<ZOiLUOR3_18YOjLPOU3d14YOkLjNX3k10XOlLdNk3b1\\OG]26fMG^25eMH^25eMH_24dMHa25`MIc24`MId23_MIf23]MJf22^MJg22\\MKg22\\MJi22ZMKj20ZMMi20^_j0"}, "label": "a girl feeding a horse"}]}
{"id": 31943, "image": "COCO_train2014_000000031943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow shirt\"."}], "task": "refering", "answer_template": "The \"man in yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 40, 41, 42, 64, 65, 88, 89, 112, 113, 130, 131, 135, 136, 153, 154, 155, 158, 159, 176, 177, 178, 181, 182, 204, 227, 249, 250, 272, 273, 295, 296], "bbox": [0.65165625, 0.029041666666666667, 0.956421875, 0.7597916666666666], "iscrowd": 0, "area": 9633.159650000001, "rle": {"size": [480, 640], "counts": "bdS68e>7I7J5J7I4M3L4M3M4L3N2M3M3N2M3N1O1O2N11O2N1O1O1O4L5K5K2N001O1O001000O010O10oNgB9Y=@mBb0R=WOUCj0`=001O001O1O001OM300O1N2O1O2N1OYK]OSKb0k4AVK>i4CWK<h4FXK9h4HYK7e4K\\K3d4N]K0c41^KN`44`KK`46aKG_4;aKD`4<`KC`4>aK@_4a0aK^O_4c0aK]O^4d0cK[O\\4f0n42N1O1000000O2O00001O0O101O00010O001O010O001O10O01O00100O0101O00001O0O101O000XOSB<m=AVB?k=]OWBd0S>00^OhA;_>00N4Lba3Hf^L3L4L3N2M4eADk=l0K5lHkN6Z1HhN3]1KfNN`11aNJe13]NGg19[NAg1b0ZNRLI<o1d3ZNlKK9n1m3YNgKL5[1gMbN`68`K00Y1V5iNgJ1KV1d5kN]J3Io0R6oNRJ5Gh0^6UOgI7F<n6@XI9BMe7KfH<_OZO]8<PH`0UOcN[Oe0e99hGc0XOPOS9?bGf0ZOeNW9g0[Gg0@\\NY9n0TGf0j9\\OSFd0n9^OnEd0S:]OjEc0W:_OfEb0Z:@cE`0_:A]E`0e:AXE`0i:ATE`0m:APE`0Q;AkD`0W;d11O1N2O1O1N4M4L3M3L4M3M3L4M3M3L4M4L3M4K4M4L3L5L3M4I7J6I:G8Gnm<"}, "label": "man in yellow shirt"}]}
{"id": 31943, "image": "COCO_train2014_000000031943.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a yellow shirt and a white hat\"."}], "task": "refering", "answer_template": "The \"a person wearing a yellow shirt and a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 40, 41, 42, 64, 65, 88, 89, 112, 113, 130, 131, 135, 136, 153, 154, 155, 158, 159, 176, 177, 178, 181, 182, 204, 227, 249, 250, 272, 273, 295, 296], "bbox": [0.65165625, 0.029041666666666667, 0.956421875, 0.7597916666666666], "iscrowd": 0, "area": 9633.159650000001, "rle": {"size": [480, 640], "counts": "bdS68e>7I7J5J7I4M3L4M3M4L3N2M3M3N2M3N1O1O2N11O2N1O1O1O4L5K5K2N001O1O001000O010O10oNgB9Y=@mBb0R=WOUCj0`=001O001O1O001OM300O1N2O1O2N1OYK]OSKb0k4AVK>i4CWK<h4FXK9h4HYK7e4K\\K3d4N]K0c41^KN`44`KK`46aKG_4;aKD`4<`KC`4>aK@_4a0aK^O_4c0aK]O^4d0cK[O\\4f0n42N1O1000000O2O00001O0O101O00010O001O010O001O10O01O00100O0101O00001O0O101O000XOSB<m=AVB?k=]OWBd0S>00^OhA;_>00N4Lba3Hf^L3L4L3N2M4eADk=l0K5lHkN6Z1HhN3]1KfNN`11aNJe13]NGg19[NAg1b0ZNRLI<o1d3ZNlKK9n1m3YNgKL5[1gMbN`68`K00Y1V5iNgJ1KV1d5kN]J3Io0R6oNRJ5Gh0^6UOgI7F<n6@XI9BMe7KfH<_OZO]8<PH`0UOcN[Oe0e99hGc0XOPOS9?bGf0ZOeNW9g0[Gg0@\\NY9n0TGf0j9\\OSFd0n9^OnEd0S:]OjEc0W:_OfEb0Z:@cE`0_:A]E`0e:AXE`0i:ATE`0m:APE`0Q;AkD`0W;d11O1N2O1O1N4M4L3M3L4M3M3L4M3M3L4M4L3M4K4M4L3L5L3M4I7J6I:G8Gnm<"}, "label": "a person wearing a yellow shirt and a white hat"}]}
{"id": 441544, "image": "COCO_train2014_000000441544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white stool with a maroon cushion in a living room\"."}], "task": "refering", "answer_template": "The \"a small white stool with a maroon cushion in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 352, 353, 354, 374, 375, 376, 377, 378], "bbox": [0.2765625, 0.7005833333333332, 0.48046875, 0.9818333333333332], "iscrowd": 0, "area": 10839.375, "rle": {"size": [480, 640], "counts": "cZc22l>3L5L4L3N3kA@f=a0VBCh=l0O1]DlNW9U1hFlNW9V1gFkNY9V1eFkNZ9W1dFjN\\9W1bFjN]9X1aFiN_9X1_FiN`9Y1^FhNa9Y1^FhNb9Y1\\FhNc9Y1\\FhNd9Y1ZFhNe9Y1ZFhNf9X1YFiNf9Y1XFhNg9Y1XFhNh9Y1VFhNi9Y1VFhNj9Y1UFgNj9Z1VFfNj9[1TFfNk9[1UFeNk9\\1TFdNl9\\1TFdNk9^1TFbNl9^1TFbNk9`1SFaNm9_1SFaNl9`1TF`Nl9`1TF`Nk9a1UF_Nk9a1TF`Nk9a1UF_Nk9a1UF_Nj9b1VF^Nj9b1VF^Nj9b1VF^Nj9b1UF_Nk9a1UF_Nk9a1UF_Nk9a1UF_Nk9a1UF_Nk9a1UF_Nk9a1TF`Nl9`1TF`Nl9Z30000000000000000000000000000O1000000O1000000O10000O[NgEeNY:Z1QF]No9c1QF]No9c1RF\\Nm9d1TF\\Nl9d1UF[Nl9c1UF]Nk9c1VF\\Nj9d1VF\\Nk9b1VF^Nj9b1WF]Nj9a1WF_Ni9a1XF^Nh9b1XF^Ni9`1YF_Ng9a1YF_Nh9_1YFaNg9_1ZF`Nf9`1ZF`Ng9^1[FaNe9_1[FaNf9]1\\FbNd9^1\\FbNd9]1^FbNc9]1^FbNb9]1`FbNa9]1`FbN`9]1bFbN^9^1cFaN^9]1dFbN\\9]1fFbN[9]1fFbNZ9]1hFbNX9^1iFaNX9]1jFbNV9^1kFaNV9]1lFbNU9\\1mFcNS9]1nFbNS9\\1oFcNR9\\1oFcNR9[1[2O0O2O1N2L4J5K:G<CbXk4"}, "label": "a small white stool with a maroon cushion in a living room"}]}
{"id": 441544, "image": "COCO_train2014_000000441544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white stool with a brown cushion\"."}], "task": "refering", "answer_template": "The \"a white stool with a brown cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 352, 353, 354, 374, 375, 376, 377, 378], "bbox": [0.2765625, 0.7005833333333332, 0.48046875, 0.9818333333333332], "iscrowd": 0, "area": 10839.375, "rle": {"size": [480, 640], "counts": "cZc22l>3L5L4L3N3kA@f=a0VBCh=l0O1]DlNW9U1hFlNW9V1gFkNY9V1eFkNZ9W1dFjN\\9W1bFjN]9X1aFiN_9X1_FiN`9Y1^FhNa9Y1^FhNb9Y1\\FhNc9Y1\\FhNd9Y1ZFhNe9Y1ZFhNf9X1YFiNf9Y1XFhNg9Y1XFhNh9Y1VFhNi9Y1VFhNj9Y1UFgNj9Z1VFfNj9[1TFfNk9[1UFeNk9\\1TFdNl9\\1TFdNk9^1TFbNl9^1TFbNk9`1SFaNm9_1SFaNl9`1TF`Nl9`1TF`Nk9a1UF_Nk9a1TF`Nk9a1UF_Nk9a1UF_Nj9b1VF^Nj9b1VF^Nj9b1VF^Nj9b1UF_Nk9a1UF_Nk9a1UF_Nk9a1UF_Nk9a1UF_Nk9a1UF_Nk9a1TF`Nl9`1TF`Nl9Z30000000000000000000000000000O1000000O1000000O10000O[NgEeNY:Z1QF]No9c1QF]No9c1RF\\Nm9d1TF\\Nl9d1UF[Nl9c1UF]Nk9c1VF\\Nj9d1VF\\Nk9b1VF^Nj9b1WF]Nj9a1WF_Ni9a1XF^Nh9b1XF^Ni9`1YF_Ng9a1YF_Nh9_1YFaNg9_1ZF`Nf9`1ZF`Ng9^1[FaNe9_1[FaNf9]1\\FbNd9^1\\FbNd9]1^FbNc9]1^FbNb9]1`FbNa9]1`FbN`9]1bFbN^9^1cFaN^9]1dFbN\\9]1fFbN[9]1fFbNZ9]1hFbNX9^1iFaNX9]1jFbNV9^1kFaNV9]1lFbNU9\\1mFcNS9]1nFbNS9\\1oFcNR9\\1oFcNR9[1[2O0O2O1N2L4J5K:G<CbXk4"}, "label": "a white stool with a brown cushion"}]}
{"id": 441544, "image": "COCO_train2014_000000441544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a gray shirt and black pants sitting on a brown couch reading a magazine\"."}], "task": "refering", "answer_template": "The \"a woman in a gray shirt and black pants sitting on a brown couch reading a magazine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 268, 269, 271, 272, 291, 292, 311, 312, 313, 314, 335, 336], "bbox": [0.5570468749999999, 0.3997083333333334, 0.8624218750000001, 0.8605416666666666], "iscrowd": 0, "area": 13340.053350000002, "rle": {"size": [480, 640], "counts": "lfW55k>5K1O0O2O001N10001N101O0O101O0O2O00001O001O00001O001N2O1N2O0O2N2O1N101N2O1N2O0O2O1N01O01O01O0010O01dNbX4o0VgKd0L310O10O10O10O010000O101O0O100O10000MUNQCl1Q=01O1O1O2O0O1O1O2NmM`NhF_1W9fNfFY1Y9mNcFR1[9TObFk0]9ZOaFd0^9A_F>_9H^F7a9N\\F2b94[FJc9<YFDf9a0WF^Og9h0VFWOi9n0TFQOk9R1TFmNk9U1UFjNk9W1UFhNj9Z1VFeNi9]1WFbNh9`1XF_Nh9b1YF]Ne9e1]FXNb9j1`FSN_9o1cFnM]9S2eFjM\\9V2gFfMY9Z2jFcMW9]2kF`MU9`2nF]MS9c2oFZMQ9f2RGWMo8i2c110O010O0100O101O0O2O001O0O2O001O0O2O001O001N101O0010O010O10O010O010O2O001N2O1N101O1N2O0O2O1N2O001N10N1N3N1O2N2N1N3N1O2N2M2O2N1N3K5J9H8H8H9F9H8E<B=C_PY1"}, "label": "a woman in a gray shirt and black pants sitting on a brown couch reading a magazine"}]}
{"id": 441544, "image": "COCO_train2014_000000441544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person sitting on the sofa\"."}], "task": "refering", "answer_template": "The \"person sitting on the sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 249, 268, 269, 271, 272, 291, 292, 311, 312, 313, 314, 335, 336], "bbox": [0.5570468749999999, 0.3997083333333334, 0.8624218750000001, 0.8605416666666666], "iscrowd": 0, "area": 13340.053350000002, "rle": {"size": [480, 640], "counts": "lfW55k>5K1O0O2O001N10001N101O0O101O0O2O00001O001O00001O001N2O1N2O0O2N2O1N101N2O1N2O0O2O1N01O01O01O0010O01dNbX4o0VgKd0L310O10O10O10O010000O101O0O100O10000MUNQCl1Q=01O1O1O2O0O1O1O2NmM`NhF_1W9fNfFY1Y9mNcFR1[9TObFk0]9ZOaFd0^9A_F>_9H^F7a9N\\F2b94[FJc9<YFDf9a0WF^Og9h0VFWOi9n0TFQOk9R1TFmNk9U1UFjNk9W1UFhNj9Z1VFeNi9]1WFbNh9`1XF_Nh9b1YF]Ne9e1]FXNb9j1`FSN_9o1cFnM]9S2eFjM\\9V2gFfMY9Z2jFcMW9]2kF`MU9`2nF]MS9c2oFZMQ9f2RGWMo8i2c110O010O0100O101O0O2O001O0O2O001O0O2O001O001N101O0010O010O10O010O010O2O001N2O1N101O1N2O0O2O1N2O001N10N1N3N1O2N2N1N3N1O2N2M2O2N1N3K5J9H8H8H9F9H8E<B=C_PY1"}, "label": "person sitting on the sofa"}]}
{"id": 441544, "image": "COCO_train2014_000000441544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the end of the brown couch sits unoccupied\"."}], "task": "refering", "answer_template": "The \"the end of the brown couch sits unoccupied\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 241, 242, 243], "bbox": [0.47767187499999997, 0.40858333333333335, 0.7255625, 0.6457083333333333], "iscrowd": 0, "area": 8733.0721, "rle": {"size": [480, 640], "counts": "_e_4U1k=4L4L3M3M1O0O101O001O00001O00001N101O00001O00001O0O1000J6K5O1O1O1O1L41N10001O0dBfNR=Z1jBkNU=a1O001QCVNf<k1TC^Ng<P2N001O0000001O00O1F9I8O1N2O1O1O1O100O10O100000O100000000O0100000O100000000O10O100000O100000N1O2N2O1N2N1O2N2O1N2N1O2N2N2O1N1O2O100O0100000000O0100001O0O101O000O2O00001O0O101N1N2O2M2O1N3M2N2N3L3N2N3M2M3N3M2N2M3N3M2M3N3M2N2M4M3M`Yb2"}, "label": "the end of the brown couch sits unoccupied"}]}
{"id": 203982, "image": "COCO_train2014_000000203982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black couch on the far left\"."}], "task": "refering", "answer_template": "The \"black couch on the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 184, 185, 186, 203, 204, 205, 206, 223, 224, 225, 226, 243, 244, 245, 246], "bbox": [0.1492857142857143, 0.6919672131147541, 0.35311355311355314, 0.985327868852459], "iscrowd": 0, "area": 10558.23035, "rle": {"size": [366, 546], "counts": "bbm04X;6J5K6J5L5J5QGTOT7R1\\HZOb7k0WHYOh7l0oGZOn7k0jGZOT8Q2O10000000000000000000000000000O1000000000000000000000000000001O00000000000O10000000000000000000000000000000000000000O10000000001O0000000000000000000000000000000O100000000001O4L4L4L1O2N1O2N1O2N1[OXGcNi83`GK;JV8J^INYSn3"}, "label": "black couch on the far left"}]}
{"id": 203982, "image": "COCO_train2014_000000203982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color single sofa placed left side of the double sofa\"."}], "task": "refering", "answer_template": "The \"a black color single sofa placed left side of the double sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 184, 185, 186, 203, 204, 205, 206, 223, 224, 225, 226, 243, 244, 245, 246], "bbox": [0.1492857142857143, 0.6919672131147541, 0.35311355311355314, 0.985327868852459], "iscrowd": 0, "area": 10558.23035, "rle": {"size": [366, 546], "counts": "bbm04X;6J5K6J5L5J5QGTOT7R1\\HZOb7k0WHYOh7l0oGZOn7k0jGZOT8Q2O10000000000000000000000000000O1000000000000000000000000000001O00000000000O10000000000000000000000000000000000000000O10000000001O0000000000000000000000000000000O100000000001O4L4L4L1O2N1O2N1O2N1[OXGcNi83`GK;JV8J^INYSn3"}, "label": "a black color single sofa placed left side of the double sofa"}]}
{"id": 203982, "image": "COCO_train2014_000000203982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty black chair facing the window\"."}], "task": "refering", "answer_template": "The \"an empty black chair facing the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 183, 184, 185, 186, 203, 204, 205, 206, 207, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247], "bbox": [0.14816849816849817, 0.684535519125683, 0.3586630036630037, 1.0], "iscrowd": 0, "area": 9842.847450000003, "rle": {"size": [366, 546], "counts": "SWm03Y;7I7I7I8H7I7ZGfNl6a1mHdNn6e1iH\\NV7l1bHUN]7R2\\HoMe7U2UHlMl7X2nGiMS8g22M201000000O10O10000000O10O10000000O1000O100000O10000000000O1000000000000O10O10000000O10000000000O10000000000O10000000000O01000000000O10000000000O10000000000O1000004L4L4K5L1O1O00001N13M7J6I8LM4J6dNmGSOX8a0d1C<Emil3"}, "label": "an empty black chair facing the window"}]}
{"id": 203982, "image": "COCO_train2014_000000203982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black leather chair to the left of the fireplace\"."}], "task": "refering", "answer_template": "The \"the black leather chair to the left of the fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 183, 184, 185, 186, 203, 204, 205, 206, 207, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247], "bbox": [0.14816849816849817, 0.684535519125683, 0.3586630036630037, 1.0], "iscrowd": 0, "area": 9842.847450000003, "rle": {"size": [366, 546], "counts": "SWm03Y;7I7I7I8H7I7ZGfNl6a1mHdNn6e1iH\\NV7l1bHUN]7R2\\HoMe7U2UHlMl7X2nGiMS8g22M201000000O10O10000000O10O10000000O1000O100000O10000000000O1000000000000O10O10000000O10000000000O10000000000O10000000000O01000000000O10000000000O10000000000O1000004L4L4K5L1O1O00001N13M7J6I8LM4J6dNmGSOX8a0d1C<Emil3"}, "label": "the black leather chair to the left of the fireplace"}]}
{"id": 203982, "image": "COCO_train2014_000000203982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch with the yellow pillow\"."}], "task": "refering", "answer_template": "The \"the couch with the yellow pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 177, 178, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 252, 253, 254, 255, 256, 257, 258, 259], "bbox": [0.6100732600732601, 0.6490710382513661, 1.0, 0.9883879781420765], "iscrowd": 0, "area": 19171.010400000003, "rle": {"size": [366, 546], "counts": "fZg38V;j0UO9G100O1O100O1O100O1O100O1O100O1000000000000000000000000O1N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2O1000000000000000000000000001O0000000oLUH_2k7`MVH`2j7_MXH`2h7_MYH`2h7_MZH`2f7_M\\H`2d7_M]Ha2c7_M^H`2b7_M_Ha2a7^MaHa2`7]MaHc2_7\\McHc2]7\\MeHc2[7\\MfHd2Z7[MhHd2X7[MiHe2W7ZMkHe2o701O001O1O001O001O001O1O001O001O001O1O001O001O001O00N2M3N2N2N2N2N2N2N2N2M3N2N2N2N2M3M3M3M3M3M3O1000000000000000000000000000000000000000000001O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2NiF"}, "label": "the couch with the yellow pillow"}]}
{"id": 203982, "image": "COCO_train2014_000000203982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black leather couch with two pillows\"."}], "task": "refering", "answer_template": "The \"black leather couch with two pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 177, 178, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 252, 253, 254, 255, 256, 257, 258, 259], "bbox": [0.6100732600732601, 0.6490710382513661, 1.0, 0.9883879781420765], "iscrowd": 0, "area": 19171.010400000003, "rle": {"size": [366, 546], "counts": "fZg38V;j0UO9G100O1O100O1O100O1O100O1O100O1000000000000000000000000O1N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2O1000000000000000000000000001O0000000oLUH_2k7`MVH`2j7_MXH`2h7_MYH`2h7_MZH`2f7_M\\H`2d7_M]Ha2c7_M^H`2b7_M_Ha2a7^MaHa2`7]MaHc2_7\\McHc2]7\\MeHc2[7\\MfHd2Z7[MhHd2X7[MiHe2W7ZMkHe2o701O001O1O001O001O001O1O001O001O001O1O001O001O001O00N2M3N2N2N2N2N2N2N2N2M3N2N2N2N2M3M3M3M3M3M3O1000000000000000000000000000000000000000000001O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2NiF"}, "label": "black leather couch with two pillows"}]}
{"id": 97492, "image": "COCO_train2014_000000097492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver laptop with a password prompt on the screen sits beside several black computers\"."}], "task": "refering", "answer_template": "The \"a silver laptop with a password prompt on the screen sits beside several black computers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366], "bbox": [0.77534375, 0.6542708333333334, 0.9983125000000002, 0.9448124999999999], "iscrowd": 0, "area": 11525.696149999994, "rle": {"size": [480, 640], "counts": "blX71n>1O2N1O2N2M2O2N1O2N1O2M2O2N1O2N1O2N1N3N1O2N1O2N1N3N1O2N2L3M4L3M4L3M4M2M4L3M4L3M4L3M4L3M4M2M4L3O20O01O1O010O0010O01O010O0010O01O010O0010O01O010O001O010O1O010O0010O01O010O0010O01O010O0010O01O010O00O1O2M2N2O2M2O1N3N1N2N3N1N2O2M2O1N3N1N2N3N1N2O2M2O1N2N3L3M3M4L3M3M4K4M3M4L3M3M4K4M3M4L3M3L\\4"}, "label": "a silver laptop with a password prompt on the screen sits beside several black computers"}]}
{"id": 97492, "image": "COCO_train2014_000000097492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop with many other moniter in room\"."}], "task": "refering", "answer_template": "The \"a laptop with many other moniter in room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 273, 274, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366], "bbox": [0.77534375, 0.6542708333333334, 0.9983125000000002, 0.9448124999999999], "iscrowd": 0, "area": 11525.696149999994, "rle": {"size": [480, 640], "counts": "blX71n>1O2N1O2N2M2O2N1O2N1O2M2O2N1O2N1O2N1N3N1O2N1O2N1N3N1O2N2L3M4L3M4L3M4M2M4L3M4L3M4L3M4L3M4M2M4L3O20O01O1O010O0010O01O010O0010O01O010O0010O01O010O001O010O1O010O0010O01O010O0010O01O010O0010O01O010O00O1O2M2N2O2M2O1N3N1N2N3N1N2O2M2O1N3N1N2N3N1N2O2M2O1N2N3L3M3M4L3M3M4K4M3M4L3M3M4K4M3M4L3M3L\\4"}, "label": "a laptop with many other moniter in room"}]}
{"id": 425175, "image": "COCO_train2014_000000425175.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass filled with a dark beverage\"."}], "task": "refering", "answer_template": "The \"a glass filled with a dark beverage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 292, 293, 294, 295], "bbox": [0.69240625, 0.15694444444444444, 0.997625, 0.9887777777777778], "iscrowd": 0, "area": 44786.86805, "rle": {"size": [360, 640], "counts": "gQl4;a50_NT1j0DQNa1X1UOdMn1f1jNUMY2V2h3H7N2O0O2O0O2O0O2O0O2O1N101O001O001O001O001O0000000000000000000000000000000000000000000000000000000000001O000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000O1N2N2N2L4M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M4L3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M3M4L3L4M3M3M3M4L3M3M3M3M3M3M4L3M3M3M3M3M4K4M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3L5L3N3M3K4L5K5K4L5K7IYd0"}, "label": "a glass filled with a dark beverage"}]}
{"id": 425175, "image": "COCO_train2014_000000425175.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large glass of tea sits next to the meatlof\"."}], "task": "refering", "answer_template": "The \"a large glass of tea sits next to the meatlof\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 292, 293, 294, 295], "bbox": [0.69240625, 0.15694444444444444, 0.997625, 0.9887777777777778], "iscrowd": 0, "area": 44786.86805, "rle": {"size": [360, 640], "counts": "gQl4;a50_NT1j0DQNa1X1UOdMn1f1jNUMY2V2h3H7N2O0O2O0O2O0O2O0O2O1N101O001O001O001O001O0000000000000000000000000000000000000000000000000000000000001O000000000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000O1N2N2N2L4M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M4L3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M3M4L3L4M3M3M3M4L3M3M3M3M3M3M4L3M3M3M3M3M4K4M3M3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3L5L3N3M3K4L5K5K4L5K7IYd0"}, "label": "a large glass of tea sits next to the meatlof"}]}
{"id": 425175, "image": "COCO_train2014_000000425175.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver chair outside\"."}], "task": "refering", "answer_template": "The \"a silver chair outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 49, 50, 51, 52, 73, 74, 75, 96, 97, 98, 119, 120, 121], "bbox": [0.139046875, 0.07641666666666667, 0.30084374999999997, 0.42697222222222225], "iscrowd": 0, "area": 7804.394700000002, "rle": {"size": [360, 640], "counts": "ZZo04o:6N3N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N3M3M2N3M3M2O2M3M2N3M3N1N3M3M2ZH^Mb6e2TIbMl6a2kHdMW7]2aHjM^7Q3O1O1ON3N1N2O1N2O20O001O0010O10O10000O2O000O10O001O1O0O2O1O001O1O001O001O1O2M3N2N3M2M3N2N2N2M3N2N2N3L5L4L4K5L4L4L8G8I8H7I>AXPm4"}, "label": "a silver chair outside"}]}
{"id": 425175, "image": "COCO_train2014_000000425175.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair sitting outside\"."}], "task": "refering", "answer_template": "The \"a white chair sitting outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 49, 50, 51, 52, 73, 74, 75, 96, 97, 98, 119, 120, 121], "bbox": [0.139046875, 0.07641666666666667, 0.30084374999999997, 0.42697222222222225], "iscrowd": 0, "area": 7804.394700000002, "rle": {"size": [360, 640], "counts": "ZZo04o:6N3N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N2N3M3M2N3M3M2O2M3M2N3M3N1N3M3M2ZH^Mb6e2TIbMl6a2kHdMW7]2aHjM^7Q3O1O1ON3N1N2O1N2O20O001O0010O10O10000O2O000O10O001O1O0O2O1O001O1O001O001O1O2M3N2N3M2M3N2N2N2M3N2N2N3L5L4L4K5L4L4L8G8I8H7I>AXPm4"}, "label": "a white chair sitting outside"}]}
{"id": 408792, "image": "COCO_train2014_000000408792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the backpack including the straps wrap around the front of the skier\"."}], "task": "refering", "answer_template": "The \"the backpack including the straps wrap around the front of the skier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 137, 139, 140, 141, 154, 155, 156, 168, 169, 183, 184], "bbox": [0.14179906542056075, 0.353015625, 0.470607476635514, 0.557109375], "iscrowd": 0, "area": 4523.41735, "rle": {"size": [640, 428], "counts": "a\\V15dc05K6N12O0O100O2N100O101N100O101N1O100O20O001O010W_OOb=1YB7d=JVB>g=BUBf0e=\\OWBk0b=ZO^Be0b=[O^Bf0a=[O^Bf0b=YO^Bh0b=XO]Bi0d=UO\\Bl0d=SO\\Bn0d=RO[Bo0e=PO[BQ1f=nNYBS1g=lNYBU1g=kNYBU1g=jNYBW1g=hNYBY1h=fNWBZ1j=fNUB[1k=eNTB\\1l=eNRB\\1m=eNRB\\1n=eNPB\\1P>dNoA]1Q>dNmA]1R>dNnA\\1R>eNlA\\1T>eNjA\\1V>fNgAZ1Y>iNdAU1_>lN_AR1d>POYAn0j>TOSAf0S?]Oj@2h?0U@Om?3Q@Mn?6o_OKQ`06m_OKS`07j_OJX`06e_OK\\`06a_OJb`06[_OJh`06U_OKl`06Q_OJRa05l^OKWa05f^OK\\a06b^OJ`a06]^OJfa06W^OJka0T13M3N1M4M3M2N3L4M?A5K1N101O1O0O2O1O1O0O2O1O001000O10000O01000O1O100O10O0100O100O100O10O01O100O100O100O022M0001O00001O00001O00005Kec]4"}, "label": "the backpack including the straps wrap around the front of the skier"}]}
{"id": 376024, "image": "COCO_train2014_000000376024.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest cow\"."}], "task": "refering", "answer_template": "The \"the closest cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.24821875000000002, 0.4301170960187354, 0.891015625, 0.8881733021077284], "iscrowd": 0, "area": 49756.993700000006, "rle": {"size": [427, 640], "counts": "U`R2`0d<<I201O0O101O00000O100000000O1000000O100000000O1000000O1000000O1000006I9HU1kN3M4K5L3M000O101O000O101O00001N100000000O100000000O1000O10000O0100000O1000000O100001O3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N000000000001O0001O1O1O001O1O001O1O1O001O0000000000O10001O0000000O2O000000001O0O10001O0000000O2O0O10001N101O001O001O001N2O001O0010O01O010O0010O01O02O2M3N1N3NO0100O1000O0100O100O10O0100O100O010O1O10O00000000001O01O000001O000000001O000000001O01O000001O000000001O00001O000010O01O00001O00001O00001O001O00001O001O000010O01O00001O00001O001O00001O001O00001O001O00001O010O00001O00000001O01O00000001O0001O000001O000001O0001O000001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O001O1O1O001O1O001O1O001O1O001O2N1O1O1O1N3N1O1N2N2N3N1N2N2O2M2N2O1N2N3M2O1N2N3N1N2N2N2O2M2N2N2O2M2N2O1N2N3M2K5J6JUll0"}, "label": "the closest cow"}]}
{"id": 376024, "image": "COCO_train2014_000000376024.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front cow in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the front cow in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.24821875000000002, 0.4301170960187354, 0.891015625, 0.8881733021077284], "iscrowd": 0, "area": 49756.993700000006, "rle": {"size": [427, 640], "counts": "U`R2`0d<<I201O0O101O00000O100000000O1000000O100000000O1000000O1000000O1000006I9HU1kN3M4K5L3M000O101O000O101O00001N100000000O100000000O1000O10000O0100000O1000000O100001O3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N000000000001O0001O1O1O001O1O001O1O1O001O0000000000O10001O0000000O2O000000001O0O10001O0000000O2O0O10001N101O001O001O001N2O001O0010O01O010O0010O01O02O2M3N1N3NO0100O1000O0100O100O10O0100O100O010O1O10O00000000001O01O000001O000000001O000000001O01O000001O000000001O00001O000010O01O00001O00001O00001O001O00001O001O000010O01O00001O00001O001O00001O001O00001O001O00001O010O00001O00000001O01O00000001O0001O000001O000001O0001O000001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O001O1O1O001O1O001O1O001O1O001O2N1O1O1O1N3N1O1N2N2N3N1N2N2O2M2N2O1N2N3M2O1N2N3N1N2N2N2O2M2N2N2O2M2N2O1N2N3M2K5J6JUll0"}, "label": "the front cow in the right hand picture"}]}
{"id": 335076, "image": "COCO_train2014_000000335076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white teddy bear being held by a young boy\"."}], "task": "refering", "answer_template": "The \"a white teddy bear being held by a young boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 349, 350, 351, 352, 353, 354, 372, 373, 374, 375, 376, 377], "bbox": [0.13921875, 0.6963333333333334, 0.4262499999999999, 1.0], "iscrowd": 0, "area": 14710.8294, "rle": {"size": [480, 640], "counts": "SSZ12l>6K4K6J6K5J3N2N110O00100O00100O0OlNWBT1g=4O2M2N3M2N2N3M2N3N1N2N3M2N3M2N3M2N2O2M2N4L3M3M3M3M3M4L3M3M2O1N2N2N3M2N2N2O1N101N100O100O100O2O0O100O100000001N100000000000001N100000000000O10O01O1O2O2M2O2N1N3N1O2M2O2N2M2N2O1N2N2N2N2N2N3M2N2N2N2N200001O0001O3N3L3M3M2ON1N2O1O2N1O1[1dN001O00001O0000001O00001O00001O00001O01O0001O00001O00001O00001O00001O001O3M2N3M2N3M3M2N3M2N3M1O2N2N1OQb[5"}, "label": "a white teddy bear being held by a young boy"}]}
{"id": 335076, "image": "COCO_train2014_000000335076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teddy bear in the hands of a little boy with a yellow cap and gray shirt\"."}], "task": "refering", "answer_template": "The \"teddy bear in the hands of a little boy with a yellow cap and gray shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 280, 281, 282, 283, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 349, 350, 351, 352, 353, 354, 372, 373, 374, 375, 376, 377], "bbox": [0.13921875, 0.6963333333333334, 0.4262499999999999, 1.0], "iscrowd": 0, "area": 14710.8294, "rle": {"size": [480, 640], "counts": "SSZ12l>6K4K6J6K5J3N2N110O00100O00100O0OlNWBT1g=4O2M2N3M2N2N3M2N3N1N2N3M2N3M2N3M2N2O2M2N4L3M3M3M3M3M4L3M3M2O1N2N2N3M2N2N2O1N101N100O100O100O2O0O100O100000001N100000000000001N100000000000O10O01O1O2O2M2O2N1N3N1O2M2O2N2M2N2O1N2N2N2N2N2N3M2N2N2N2N200001O0001O3N3L3M3M2ON1N2O1O2N1O1[1dN001O00001O0000001O00001O00001O00001O01O0001O00001O00001O00001O00001O001O3M2N3M2N3M3M2N3M2N3M1O2N2N1OQb[5"}, "label": "teddy bear in the hands of a little boy with a yellow cap and gray shirt"}]}
{"id": 474342, "image": "COCO_train2014_000000474342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat whose tail is hiding behind the curtain\"."}], "task": "refering", "answer_template": "The \"a cat whose tail is hiding behind the curtain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 112, 113, 114, 115, 129, 130, 131, 132, 133, 147, 148, 149, 150, 165, 166, 167, 168, 184, 185, 186], "bbox": [0.2084, 0.16628019323671497, 0.51354, 0.698864734299517], "iscrowd": 0, "area": 17940.687799999996, "rle": {"size": [414, 500], "counts": "eYZ14^<`0@`0@?B=B=C<D<D<fNiLbHb3^7P1O0000001N10000000\\OQHZLn7e3SH[Lm7d3TH\\Lk7d3VH]Li7b3YH]Lg7c3YH^Lf7a3[H_Le7`3\\H`Ld7_3^HaL`7_3aHaL_7^3bHcL]7\\3dHdL\\7[3fHdLZ7[3gHfLX7X3jHhLU7X3lHiLS7V3oHiLQ7V3PIjLP7T3RImLm6R3UImLk6W2_HjMg0Oi6X2cHeMe03h6Y2fH^Md09f6Y2jHYMb0=d6Z2nHTM?a0d6[2QJeMo5[2QJeMo5\\2PJdMo5]2QJbMP6^2PJbMP6^2PJbMP6^2PJbMP6^2PJbMP6_2oI`MQ6b2oI]MQ6g2lIXMT6l2iISMW6P3gInL[6U3bIjL^6Z3_IeLa6_3bIZL^6j3Y14L3L5L4K5K5L4K4L4M1N2N3M2O1N2N3NO01O1O10]OiH_KV7Z4RIfKm6S4\\IlKd6Q4_IoK`6R4aImK^6T4bIlK]6U4dIiK\\6^3SIhLa0JZ6^3]IbL:OX6]3gI^L33T6_3PK^Ln4b3XKZLg4d3Y200O2O00010O010O0010O0100O100O2O1N2O1N2O1N2O1N2O1N4M3L3M2M3J6K5J6J6O11O001N101O001O001O0O10O1O1O1O1O1O1O1O1O1cNQEj0P;ROXEj0h:QOaE:Q;AWE8n;IaTR3"}, "label": "a cat whose tail is hiding behind the curtain"}]}
{"id": 474342, "image": "COCO_train2014_000000474342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shorter cat on left side\"."}], "task": "refering", "answer_template": "The \"shorter cat on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 112, 113, 114, 115, 129, 130, 131, 132, 133, 147, 148, 149, 150, 165, 166, 167, 168, 184, 185, 186], "bbox": [0.2084, 0.16628019323671497, 0.51354, 0.698864734299517], "iscrowd": 0, "area": 17940.687799999996, "rle": {"size": [414, 500], "counts": "eYZ14^<`0@`0@?B=B=C<D<D<fNiLbHb3^7P1O0000001N10000000\\OQHZLn7e3SH[Lm7d3TH\\Lk7d3VH]Li7b3YH]Lg7c3YH^Lf7a3[H_Le7`3\\H`Ld7_3^HaL`7_3aHaL_7^3bHcL]7\\3dHdL\\7[3fHdLZ7[3gHfLX7X3jHhLU7X3lHiLS7V3oHiLQ7V3PIjLP7T3RImLm6R3UImLk6W2_HjMg0Oi6X2cHeMe03h6Y2fH^Md09f6Y2jHYMb0=d6Z2nHTM?a0d6[2QJeMo5[2QJeMo5\\2PJdMo5]2QJbMP6^2PJbMP6^2PJbMP6^2PJbMP6^2PJbMP6_2oI`MQ6b2oI]MQ6g2lIXMT6l2iISMW6P3gInL[6U3bIjL^6Z3_IeLa6_3bIZL^6j3Y14L3L5L4K5K5L4K4L4M1N2N3M2O1N2N3NO01O1O10]OiH_KV7Z4RIfKm6S4\\IlKd6Q4_IoK`6R4aImK^6T4bIlK]6U4dIiK\\6^3SIhLa0JZ6^3]IbL:OX6]3gI^L33T6_3PK^Ln4b3XKZLg4d3Y200O2O00010O010O0010O0100O100O2O1N2O1N2O1N2O1N2O1N4M3L3M2M3J6K5J6J6O11O001N101O001O001O0O10O1O1O1O1O1O1O1O1O1cNQEj0P;ROXEj0h:QOaE:Q;AWE8n;IaTR3"}, "label": "shorter cat on left side"}]}
{"id": 474342, "image": "COCO_train2014_000000474342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat on the right side\"."}], "task": "refering", "answer_template": "The \"the cat on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 30, 31, 32, 48, 49, 50, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 206, 207, 208, 225, 226, 244], "bbox": [0.44872, 0.0607487922705314, 0.8155399999999999, 0.8844202898550724], "iscrowd": 0, "area": 29682.420950000003, "rle": {"size": [414, 500], "counts": "gkj22`<=C<D=C=B>C=D<K5L3M4K5L4L4K5L3M4L4L4L4L4M3L3M4L4L4M3L4L3M4MRNmHVNo6m1]IhM_6[2cIdMY6_2hIaMT6b2mI^MP6d2QJ[Ml5h2UJXMh5i2ZJWMd5j2\\JWMb5j2_JVM_5k2bJUM\\5l2eJSMZ5n2gJRMW5o2jJQMT5o2nJQMP5P3QKPMm4Q3TKoLj4R3WKnLg4S3ZKlLe4V3[KjLc4W3^KiL`4Y3`KgL^4[3bKeL\\4]3dKcLZ4^3gKaLX4a3hK_LU4d3kK\\LS4f3mKZLQ4g3PLYLo3h3PLYLn3i3RLVLn3k3RLULm3k3TLULj3m3VLSLi3n3WLQLh3P4YLoKg3R4YLlKh3T4YLkKf3V4[LiKe3X4[LfKf3Z4[LeKd3]4\\LbKd3^4]L`Kd3`4]L_Kb3c4^L\\Kb3d4_LZKa3h4_LWKa3i4`LVK`3j4aLQKb3Q5^LjJf3V5ZLfJj3[5VL_Jn3b5g1001N100O10000O10000O10000OZOSI]Km6^4XIaKh6\\4\\IdKd6W4aIiK_6R4fInKY6n3lIRLT6j3PJVLP6e3UJ[Lj5a3[J_Le5\\3`JdL`5W3eJiLZ5S3kJmLU5P3nJPMQ5Q3oJoLQ5Q3oJnLR5R3nJnLQ5S3oJmLQ5S3oJmLQ5S3oJmLo4T3RKlLj4X3VKhLf4\\3oI\\L`08^5_3dIhLm0I[5`4dJ`KX5d4iJ[KS5j4mJTKn4S5RKjJi4^5WK_Je4g5\\KWJo3N[KR6f0mIQ45oJU6U6;N101O1N101O1OO2O1O001O1O001O1O1O000O0100000000O100O^MYJoMg5o1lJ`MS5`2^KPMb4n2lKeLT4Z3oKeLP4Z3SLeLk3\\3WLcLh3\\3\\LbLc3]3`LbL_3^3cLaL]3]3fL^L^3`3Q3N2O1N2N2N2O1oNbF`N`9\\1lFZNV9b1UGVNl8f1_GQNc8k1^1H9H7I8H4L4L6J5DPUU1"}, "label": "the cat on the right side"}]}
{"id": 474342, "image": "COCO_train2014_000000474342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat looking out of window on right\"."}], "task": "refering", "answer_template": "The \"cat looking out of window on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 30, 31, 32, 48, 49, 50, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 152, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 206, 207, 208, 225, 226, 244], "bbox": [0.44872, 0.0607487922705314, 0.8155399999999999, 0.8844202898550724], "iscrowd": 0, "area": 29682.420950000003, "rle": {"size": [414, 500], "counts": "gkj22`<=C<D=C=B>C=D<K5L3M4K5L4L4K5L3M4L4L4L4L4M3L3M4L4L4M3L4L3M4MRNmHVNo6m1]IhM_6[2cIdMY6_2hIaMT6b2mI^MP6d2QJ[Ml5h2UJXMh5i2ZJWMd5j2\\JWMb5j2_JVM_5k2bJUM\\5l2eJSMZ5n2gJRMW5o2jJQMT5o2nJQMP5P3QKPMm4Q3TKoLj4R3WKnLg4S3ZKlLe4V3[KjLc4W3^KiL`4Y3`KgL^4[3bKeL\\4]3dKcLZ4^3gKaLX4a3hK_LU4d3kK\\LS4f3mKZLQ4g3PLYLo3h3PLYLn3i3RLVLn3k3RLULm3k3TLULj3m3VLSLi3n3WLQLh3P4YLoKg3R4YLlKh3T4YLkKf3V4[LiKe3X4[LfKf3Z4[LeKd3]4\\LbKd3^4]L`Kd3`4]L_Kb3c4^L\\Kb3d4_LZKa3h4_LWKa3i4`LVK`3j4aLQKb3Q5^LjJf3V5ZLfJj3[5VL_Jn3b5g1001N100O10000O10000O10000OZOSI]Km6^4XIaKh6\\4\\IdKd6W4aIiK_6R4fInKY6n3lIRLT6j3PJVLP6e3UJ[Lj5a3[J_Le5\\3`JdL`5W3eJiLZ5S3kJmLU5P3nJPMQ5Q3oJoLQ5Q3oJnLR5R3nJnLQ5S3oJmLQ5S3oJmLQ5S3oJmLo4T3RKlLj4X3VKhLf4\\3oI\\L`08^5_3dIhLm0I[5`4dJ`KX5d4iJ[KS5j4mJTKn4S5RKjJi4^5WK_Je4g5\\KWJo3N[KR6f0mIQ45oJU6U6;N101O1N101O1OO2O1O001O1O001O1O1O000O0100000000O100O^MYJoMg5o1lJ`MS5`2^KPMb4n2lKeLT4Z3oKeLP4Z3SLeLk3\\3WLcLh3\\3\\LbLc3]3`LbL_3^3cLaL]3]3fL^L^3`3Q3N2O1N2N2N2O1oNbF`N`9\\1lFZNV9b1UGVNl8f1_GQNc8k1^1H9H7I8H4L4L6J5DPUU1"}, "label": "cat looking out of window on right"}]}
{"id": 195816, "image": "COCO_train2014_000000195816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench closest to the scissors\"."}], "task": "refering", "answer_template": "The \"the bench closest to the scissors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 176, 177, 178, 179, 193, 194, 195, 206, 207, 208, 209, 221, 222, 223, 236, 237, 251], "bbox": [0.588140703517588, 0.5540666666666667, 0.9903768844221106, 0.83435], "iscrowd": 0, "area": 9083.0579, "rle": {"size": [600, 398], "counts": "^^Y41fb02O1O1N2O2N1N2O1O1N3N1O1N2O2M2O1O1N10000O10O10O10000O01000O100O01000O10O10O10000O01000O1000O0100O11O0O1000001O`_OSOU?l0l@UOS?k0m@VOR?j0m@XOS?g0m@YOS?g0l@[OS?d0n@]OQ?c0o@^OQ?a0n@AQ??o@BP?>PACo><QAFn>:RAGn>8RAIm>7RAKm>4TAMk>3TAOl>0TA1k>OUA2j>NUA4j>KWA6i>IWA8h>HWA:h>FXA;g>EYA<f>CZA?f>@ZAa0e>_OZAc0e>]O[Ad0d>\\O\\Ae0d>YO\\Ai0c>WO]Aj0b>VO^Ak0a>UO^Am0a>SO_An0a>PO`AQ1_>oN`AS1_>mNaAT1^>lNbAT1_>kN`AW1_>hNbAY1]>gNbA[1]>eNcA\\1]>cNcA^1\\>bNcA`1\\>_NeAb1Z>^NfAc1Y>]NfAe1Z>ZNfAg1Y>YNgAh1X>WNhAk1W>UNiAl1W>SNhAo1W>QNiAP2_?1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O001N2O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1OL4@`0A?@`0@`0A?@?Aio1"}, "label": "the bench closest to the scissors"}]}
{"id": 163054, "image": "COCO_train2014_000000163054.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest airplanes tail\"."}], "task": "refering", "answer_template": "The \"the closest airplanes tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 19, 20, 21, 22, 23, 33, 34, 35, 36, 37, 47, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 67, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 135, 136, 137, 138, 139, 140, 141, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 240], "bbox": [0.0, 0.039765625, 0.656791569086651, 0.7139218749999999], "iscrowd": 0, "area": 70566.16115000001, "rle": {"size": [640, 427], "counts": "^;k2S<R5O0O2O1O001O1N101O1O001N2O1O001O1N101O1O001N2O1O001N2O001O1O0O2O1O1O001N2O001O1O0O2O1O001O1N2O001O1N101O1O001N2O1O001O1N101O1O001N2O1O001O1N101O1O001N2O001O1N2O001O1O0O2O1O001O1N2O001O1O0O2O1O001O1N2O001O1N101O1O001N2O001O1O1N101O1O001N2O001O1O1N101O1O0O2O1O001O1N2O001O1O0O2O1O001O1N101O1O1O001000O10000O1000000O10000O10000O1000O01000000O10000ZMoG\\IR8_6TH`Il7Z6ZHfIf7U6_HjIb7Q6cHoI]7l5iHSJW7h5nHXJR7c5SI\\Jn6_5WIaJi6Z5]IeJc6V5bIiJ_6R5fInJZ6m4kISKU6h4QJVKP6e4UJ[Kk5`4ZJ`Kf5[4_JeKa5V4eJhK\\5S4iJmKW5n3nJRLR5i3SKVLn4e3WK[Li4`3]K_Lc4[3cKdL^4W3gKiLY4R3lKnLT4m2RLQMo3j2VLVMj3e2[L[Me3`2`L`M`3[2fLcM[3X2jLhMV3S2oLmMQ3n1TMQNm2j1YMUNg2f1^MZNb2a1cM^N^2]1gMcNY2X1mMgNS2T1RNlNn1o0WNPOj1k0[NUOe1f0aNYO_1b0fN]O[1=kNCU18POHP13VOKk00ZO0f0K_O5a0FD:<AJ=7^ONb02YO3g0LUO9j0HQO=o0ClNc0S1]OhNh0W1YOdNl0\\1TO_NQ1a1h8N1000001N10000O2O000O101O000O101O0O10001N1000001N10000O101O0O10001N1000001N10000O3N3M3L5L3M3M4K``k2"}, "label": "the closest airplanes tail"}]}
{"id": 163054, "image": "COCO_train2014_000000163054.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tail where the full airline name can be read\"."}], "task": "refering", "answer_template": "The \"the tail where the full airline name can be read\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 19, 20, 21, 22, 23, 33, 34, 35, 36, 37, 47, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 67, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 135, 136, 137, 138, 139, 140, 141, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 240], "bbox": [0.0, 0.039765625, 0.656791569086651, 0.7139218749999999], "iscrowd": 0, "area": 70566.16115000001, "rle": {"size": [640, 427], "counts": "^;k2S<R5O0O2O1O001O1N101O1O001N2O1O001O1N101O1O001N2O1O001N2O001O1O0O2O1O1O001N2O001O1O0O2O1O001O1N2O001O1N101O1O001N2O1O001O1N101O1O001N2O1O001O1N101O1O001N2O001O1N2O001O1O0O2O1O001O1N2O001O1O0O2O1O001O1N2O001O1N101O1O001N2O001O1O1N101O1O001N2O001O1O1N101O1O0O2O1O001O1N2O001O1O0O2O1O001O1N101O1O1O001000O10000O1000000O10000O10000O1000O01000000O10000ZMoG\\IR8_6TH`Il7Z6ZHfIf7U6_HjIb7Q6cHoI]7l5iHSJW7h5nHXJR7c5SI\\Jn6_5WIaJi6Z5]IeJc6V5bIiJ_6R5fInJZ6m4kISKU6h4QJVKP6e4UJ[Kk5`4ZJ`Kf5[4_JeKa5V4eJhK\\5S4iJmKW5n3nJRLR5i3SKVLn4e3WK[Li4`3]K_Lc4[3cKdL^4W3gKiLY4R3lKnLT4m2RLQMo3j2VLVMj3e2[L[Me3`2`L`M`3[2fLcM[3X2jLhMV3S2oLmMQ3n1TMQNm2j1YMUNg2f1^MZNb2a1cM^N^2]1gMcNY2X1mMgNS2T1RNlNn1o0WNPOj1k0[NUOe1f0aNYO_1b0fN]O[1=kNCU18POHP13VOKk00ZO0f0K_O5a0FD:<AJ=7^ONb02YO3g0LUO9j0HQO=o0ClNc0S1]OhNh0W1YOdNl0\\1TO_NQ1a1h8N1000001N10000O2O000O101O000O101O0O10001N1000001N10000O101O0O10001N1000001N10000O3N3M3L5L3M3M4K``k2"}, "label": "the tail where the full airline name can be read"}]}
{"id": 163054, "image": "COCO_train2014_000000163054.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle planes ' tail\"."}], "task": "refering", "answer_template": "The \"the middle planes ' tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 82, 83, 84, 85, 97, 98, 99, 111, 112, 113, 114, 126, 127, 128, 129, 141, 142, 143, 144, 156, 157, 158, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191], "bbox": [0.40962529274004683, 0.19609375, 0.7511007025761125, 0.55215625], "iscrowd": 0, "area": 18080.98355, "rle": {"size": [640, 427], "counts": "Rf]33gc06K5J7I6K5J6K6I6K5J7J5J6J6K6I6K5J7J5J6J7J5J6K5J7J5J6K6M200O1O2N1N2O2N1O1O1O2M2O1O2N1O1O1N3N1O1O2N1O1O1N3N1O1O2N1O1N3N1O1O1O2N1N2O2N1O10001O00001O0000001O00aNRClKn<P4VCQLi<j3\\CVLd<e3aC[L_<a3fC^LZ<]3kCdLT<X3PDhLP<S3UDmLk;o2YDRMf;i2`DVM`;f2dDZM\\;a2iD`MV;[2oDeMQ;W2TEhMl:S2YEmMg:o1]ERNb:i1cEWN]:e1hEZNX:a1mE`NR:[1SFeNm9W1WFiNi9R1\\FnNd9n0aFQO_9j0fFVOZ9f0jFZOV9a0oF_OQ9>SG@n8<VGDj89YGGg85]GKc81bGN^8OeG0\\8LhG4X8IkG7U8EoG;Q8ATH>l7_OWHa0i7[O[Hd0f7YO]Hg0c7UObHj0^7ROfHn0Z7oNiHQ1S=O2O0000000O2O0000001N1000001O000O101O00000O3N7I7I7IZmQ2"}, "label": "the middle planes ' tail"}]}
{"id": 163054, "image": "COCO_train2014_000000163054.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plane tail in the center\"."}], "task": "refering", "answer_template": "The \"the plane tail in the center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 82, 83, 84, 85, 97, 98, 99, 111, 112, 113, 114, 126, 127, 128, 129, 141, 142, 143, 144, 156, 157, 158, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191], "bbox": [0.40962529274004683, 0.19609375, 0.7511007025761125, 0.55215625], "iscrowd": 0, "area": 18080.98355, "rle": {"size": [640, 427], "counts": "Rf]33gc06K5J7I6K5J6K6I6K5J7J5J6J6K6I6K5J7J5J6J7J5J6K5J7J5J6K6M200O1O2N1N2O2N1O1O1O2M2O1O2N1O1O1N3N1O1O2N1O1O1N3N1O1O2N1O1N3N1O1O1O2N1N2O2N1O10001O00001O0000001O00aNRClKn<P4VCQLi<j3\\CVLd<e3aC[L_<a3fC^LZ<]3kCdLT<X3PDhLP<S3UDmLk;o2YDRMf;i2`DVM`;f2dDZM\\;a2iD`MV;[2oDeMQ;W2TEhMl:S2YEmMg:o1]ERNb:i1cEWN]:e1hEZNX:a1mE`NR:[1SFeNm9W1WFiNi9R1\\FnNd9n0aFQO_9j0fFVOZ9f0jFZOV9a0oF_OQ9>SG@n8<VGDj89YGGg85]GKc81bGN^8OeG0\\8LhG4X8IkG7U8EoG;Q8ATH>l7_OWHa0i7[O[Hd0f7YO]Hg0c7UObHj0^7ROfHn0Z7oNiHQ1S=O2O0000000O2O0000001N1000001O000O101O00000O3N7I7I7IZmQ2"}, "label": "the plane tail in the center"}]}
{"id": 523505, "image": "COCO_train2014_000000523505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the remote that is to the left of the other remote\"."}], "task": "refering", "answer_template": "The \"the remote that is to the left of the other remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 327], "bbox": [0.097796875, 0.033749999999999995, 0.493078125, 0.8355625], "iscrowd": 0, "area": 46050.06355, "rle": {"size": [480, 640], "counts": "Qlm06e>6J6J6J7I6I7J6J6J7I6M3N2M3M4L3M3M3N1N2N3M2N3M2O2M2N2N3M2N3N1N3M2N2N3M2O2M2M3N3M2N2N2N2M3N3M2N2N2M3N2N2N2N2N2M4M2N2N2N2M3N2O1N2N2O1N3M2N2O1N2N2O1N2N2O1N3M2N2O2M2N2O2M2N2O2M2N2M4M2N2M3N3L3N2M4M2M3N3L3N3L3N2M4M2M101N3N1N2O2M2O1N3N101N100O10000O10O0100O01M3M2N3N1N3M2N3N1N3M3M2N3N1N2N110O1O1O1O1O1O100O1O1O1O2N2N3M3N2M2N3M3M2N3M3N1N3N2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N3M2N2N3M2N3M2N3M2N2N3M2N3M2N2M4M2N3M2N3L3N2N3M2M4M2N2N2N2M3N2N2N2M3N3M2N2N2M3N2N2N2N2M3N2N3M2M3N2N2N2N3L3N2N2N2N3L3M3L4L4L4M4K4L4L4L4J8H8HRZh4"}, "label": "the remote that is to the left of the other remote"}]}
{"id": 523505, "image": "COCO_train2014_000000523505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the controller with the wire not chewed around\"."}], "task": "refering", "answer_template": "The \"the controller with the wire not chewed around\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 327], "bbox": [0.097796875, 0.033749999999999995, 0.493078125, 0.8355625], "iscrowd": 0, "area": 46050.06355, "rle": {"size": [480, 640], "counts": "Qlm06e>6J6J6J7I6I7J6J6J7I6M3N2M3M4L3M3M3N1N2N3M2N3M2O2M2N2N3M2N3N1N3M2N2N3M2O2M2M3N3M2N2N2N2M3N3M2N2N2M3N2N2N2N2N2M4M2N2N2N2M3N2O1N2N2O1N3M2N2O1N2N2O1N2N2O1N3M2N2O2M2N2O2M2N2O2M2N2M4M2N2M3N3L3N2M4M2M3N3L3N3L3N2M4M2M101N3N1N2O2M2O1N3N101N100O10000O10O0100O01M3M2N3N1N3M2N3N1N3M3M2N3N1N2N110O1O1O1O1O1O100O1O1O1O2N2N3M3N2M2N3M3M2N3M3N1N3N2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N1O2N3M2N2N3M2N3M2N3M2N2N3M2N3M2N2M4M2N3M2N3L3N2N3M2M4M2N2N2N2M3N2N2N2M3N3M2N2N2M3N2N2N2N2M3N2N3M2M3N2N2N2N3L3N2N2N2N3L3M3L4L4L4M4K4L4L4L4J8H8HRZh4"}, "label": "the controller with the wire not chewed around"}]}
{"id": 72947, "image": "COCO_train2014_000000072947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing shorts plays frisbee with two others\"."}], "task": "refering", "answer_template": "The \"a boy wearing shorts plays frisbee with two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 144, 145, 146, 167, 168, 169, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 282, 283, 284, 285, 305, 306, 307, 308, 329, 331, 351, 352, 354], "bbox": [0.257234375, 0.3011261261261261, 0.43340625, 0.9752702702702702], "iscrowd": 0, "area": 14921.940149999993, "rle": {"size": [444, 640], "counts": "jhW25f=101O001O1O001PM4YF:0Cf95YFR1WOjN^:6YF_1S1YN]6:^He1g0\\Nf61bHk19_NQ7GdHR2LcN[7]OhHX2^OeNg7TOiHS4R7QLkHS4Q7oKlHV4o6mKoHV4n6`1L4L4K4M4L4L4RKaHW3b7cLbH\\3b7^L`Hb3d7XL_Hg3e7RL^Hn3e7mK^HR4f7hK\\HY4g7`K]H_4_8100QM`KaLa4^3dK]L^4`3hK\\LZ4`3mKZLV4b3QLXLR4d3ULWLn3d3YLVLj3f3]LTLf3h3aLRLe3j3^LRLf3l3]LnKi3n3[LlKj3R4YLiKl3S4XLgKj3Y4ZLaKh3_4\\L\\Kd3d4`LWKb3i4h2100O100O10O0100O2O1O1000O01000000000O1`LSGo1m8nM_Gi1a8UNiGc1W8ZNUH]1k7`N`HX1`7eNkHR1V7kNVIl0j6QOaIg0_6WOlI`0T6]OVJ<j5AXJ`0h5]OZJc0g5ZO[Jg0e5WO\\Jj0d5SO^Jn0b5oN`JR1`5kNbJV1R9002N7L3M4L3MM3L3N3M3L4M3M3M2M2O1O2M2O1O1N2N2EPPm4"}, "label": "a boy wearing shorts plays frisbee with two others"}]}
{"id": 72947, "image": "COCO_train2014_000000072947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt\"."}], "task": "refering", "answer_template": "The \"man in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 121, 122, 144, 145, 146, 167, 168, 169, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 282, 283, 284, 285, 305, 306, 307, 308, 329, 331, 351, 352, 354], "bbox": [0.257234375, 0.3011261261261261, 0.43340625, 0.9752702702702702], "iscrowd": 0, "area": 14921.940149999993, "rle": {"size": [444, 640], "counts": "jhW25f=101O001O1O001PM4YF:0Cf95YFR1WOjN^:6YF_1S1YN]6:^He1g0\\Nf61bHk19_NQ7GdHR2LcN[7]OhHX2^OeNg7TOiHS4R7QLkHS4Q7oKlHV4o6mKoHV4n6`1L4L4K4M4L4L4RKaHW3b7cLbH\\3b7^L`Hb3d7XL_Hg3e7RL^Hn3e7mK^HR4f7hK\\HY4g7`K]H_4_8100QM`KaLa4^3dK]L^4`3hK\\LZ4`3mKZLV4b3QLXLR4d3ULWLn3d3YLVLj3f3]LTLf3h3aLRLe3j3^LRLf3l3]LnKi3n3[LlKj3R4YLiKl3S4XLgKj3Y4ZLaKh3_4\\L\\Kd3d4`LWKb3i4h2100O100O10O0100O2O1O1000O01000000000O1`LSGo1m8nM_Gi1a8UNiGc1W8ZNUH]1k7`N`HX1`7eNkHR1V7kNVIl0j6QOaIg0_6WOlI`0T6]OVJ<j5AXJ`0h5]OZJc0g5ZO[Jg0e5WO\\Jj0d5SO^Jn0b5oN`JR1`5kNbJV1R9002N7L3M4L3MM3L3N3M3L4M3M3M2M2O1O2M2O1O1N2N2EPPm4"}, "label": "man in white shirt"}]}
{"id": 72947, "image": "COCO_train2014_000000072947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy in green shirt\"."}], "task": "refering", "answer_template": "The \"guy in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 180, 181, 202, 203, 204, 205, 224, 225, 226, 227, 228, 249, 250, 251, 272, 273, 295, 296, 318, 319], "bbox": [0.775109375, 0.39878378378378376, 0.9365625, 0.8319369369369369], "iscrowd": 0, "area": 8594.998899999999, "rle": {"size": [444, 640], "counts": "\\Zg6<`=1N100O1O100O01O01O010O010O01O01O010O0\\OBlC=T<EjC<U<EkC:U<HiC9V<IhC7X<KgC4Y<MfC4Y<NfC1Z<1dC0Z<3eCL[<5dCL[<6cCJ]<7cCH]<8cCI\\<7eCH[<8eCIZ<7gCHY<8gCHY<8gCIW<8jCGV<9jCHg2FZ6b0PGGU2:i6ORGHc1k0Z7]OTGGQ1^1d7XNSGc08Gd0m1]7[NgG18H`0P2_7XNiGO:H:W2b7QNjG0:I6Z2f7lMjG1;H1a2h7^NVHROOd2k7YNTHVOMf2n7SNRHZOLh2R8mMPHo2n7QMPHR3P8^1O001N1000000O100000O0100000\\NUHWMj7`NVHT44ZMl7a2YH]Mi7_2[H^Mh7]2]HaMe7\\2]HcMe7Z2^HdMd7Y2_HeMd7X2^HfMd7W2_HfMd7W2_HgMc7V2`HhMd7S2^HlMe7P2^HnMe7o1]HoMf7m1]HQNg7j1\\HSNh7i1[HUNh7g1ZHXNi7d1ZHZNj7a1YH]Nj7`1XH^Nk7^1XH`Nk7\\1XHaNl7[1_2K4L4M3L5K4L4L4L5L3LX`a0"}, "label": "guy in green shirt"}]}
{"id": 72947, "image": "COCO_train2014_000000072947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player in green shirt\"."}], "task": "refering", "answer_template": "The \"player in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 180, 181, 202, 203, 204, 205, 224, 225, 226, 227, 228, 249, 250, 251, 272, 273, 295, 296, 318, 319], "bbox": [0.775109375, 0.39878378378378376, 0.9365625, 0.8319369369369369], "iscrowd": 0, "area": 8594.998899999999, "rle": {"size": [444, 640], "counts": "\\Zg6<`=1N100O1O100O01O01O010O010O01O01O010O0\\OBlC=T<EjC<U<EkC:U<HiC9V<IhC7X<KgC4Y<MfC4Y<NfC1Z<1dC0Z<3eCL[<5dCL[<6cCJ]<7cCH]<8cCI\\<7eCH[<8eCIZ<7gCHY<8gCHY<8gCIW<8jCGV<9jCHg2FZ6b0PGGU2:i6ORGHc1k0Z7]OTGGQ1^1d7XNSGc08Gd0m1]7[NgG18H`0P2_7XNiGO:H:W2b7QNjG0:I6Z2f7lMjG1;H1a2h7^NVHROOd2k7YNTHVOMf2n7SNRHZOLh2R8mMPHo2n7QMPHR3P8^1O001N1000000O100000O0100000\\NUHWMj7`NVHT44ZMl7a2YH]Mi7_2[H^Mh7]2]HaMe7\\2]HcMe7Z2^HdMd7Y2_HeMd7X2^HfMd7W2_HfMd7W2_HgMc7V2`HhMd7S2^HlMe7P2^HnMe7o1]HoMf7m1]HQNg7j1\\HSNh7i1[HUNh7g1ZHXNi7d1ZHZNj7a1YH]Nj7`1XH^Nk7^1XH`Nk7\\1XHaNl7[1_2K4L4M3L5K4L4L4L5L3LX`a0"}, "label": "player in green shirt"}]}
{"id": 72947, "image": "COCO_train2014_000000072947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in navy blue jacket and tan pants running\"."}], "task": "refering", "answer_template": "The \"a man in navy blue jacket and tan pants running\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 291, 292, 311, 312, 314, 315, 334, 335, 357, 358], "bbox": [0.49731249999999994, 0.32135135135135134, 0.715578125, 0.9595495495495496], "iscrowd": 0, "area": 14764.873850000007, "rle": {"size": [444, 640], "counts": "aSZ42e=8I6I8I5J7J5K6L3M6J5K5K4N4K5L0O2N101N2O0O2ZO^MPFb2n3ZMW1a0iJW2n3YMU1e0kJT2o3WMR1k0mJn1P4YMR1j0lJo1Q4WMR1l0lJo1P4VMR1m0lJS2l3QMX1m0kJW2h3lL\\1P1iJ\\2d3eLb1h4WNYKi1P5nMQKQ2b5\\M^Je2V6eLlIZ3W6cLiI]3[6_LeIa3a6XLaIh3l6jKTIV4Z7\\KgHc4c7RK^Ho4b7PK^HP5d7nJ]HQ5e7lJ\\HT5W80O00001O001O0AkJWHV5L^Kh6DYIP5J`Kg6I[Ie5a6bJ\\I_5^6S1K5J5K5L5J5L4L5O00001N10001O001mN`GgLa8U3dGjL\\8S3iGkLX8P3mGoLS8k2UHSMk7h2[HVMg7o1TIPNl6k1YIUNg6f1^IZNc6`1bI`N^6\\1fIcN[6[1gIeNZ6X1hIhNY6V1hIjNZ6S1gImNZ6R1eInN^6o0cIQO^6m0cISO_6j0bIVO`6h0`IXOb6e0_I[Oc6b0^I^Od6>^IAe6<\\IDf69[IGg65[IKg62ZINi6MXI4j6IWI7l6EUI:n6BTI>o6^ORIb0T:10O01O00001O001O010O00001O001O010O001O1O101N1000001N1000000OO2O1N2N6K9F8Hia^2"}, "label": "a man in navy blue jacket and tan pants running"}]}
{"id": 72947, "image": "COCO_train2014_000000072947.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man running in the middle\"."}], "task": "refering", "answer_template": "The \"the man running in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 291, 292, 311, 312, 314, 315, 334, 335, 357, 358], "bbox": [0.49731249999999994, 0.32135135135135134, 0.715578125, 0.9595495495495496], "iscrowd": 0, "area": 14764.873850000007, "rle": {"size": [444, 640], "counts": "aSZ42e=8I6I8I5J7J5K6L3M6J5K5K4N4K5L0O2N101N2O0O2ZO^MPFb2n3ZMW1a0iJW2n3YMU1e0kJT2o3WMR1k0mJn1P4YMR1j0lJo1Q4WMR1l0lJo1P4VMR1m0lJS2l3QMX1m0kJW2h3lL\\1P1iJ\\2d3eLb1h4WNYKi1P5nMQKQ2b5\\M^Je2V6eLlIZ3W6cLiI]3[6_LeIa3a6XLaIh3l6jKTIV4Z7\\KgHc4c7RK^Ho4b7PK^HP5d7nJ]HQ5e7lJ\\HT5W80O00001O001O0AkJWHV5L^Kh6DYIP5J`Kg6I[Ie5a6bJ\\I_5^6S1K5J5K5L5J5L4L5O00001N10001O001mN`GgLa8U3dGjL\\8S3iGkLX8P3mGoLS8k2UHSMk7h2[HVMg7o1TIPNl6k1YIUNg6f1^IZNc6`1bI`N^6\\1fIcN[6[1gIeNZ6X1hIhNY6V1hIjNZ6S1gImNZ6R1eInN^6o0cIQO^6m0cISO_6j0bIVO`6h0`IXOb6e0_I[Oc6b0^I^Od6>^IAe6<\\IDf69[IGg65[IKg62ZINi6MXI4j6IWI7l6EUI:n6BTI>o6^ORIb0T:10O01O00001O001O010O00001O001O010O001O1O101N1000001N1000000OO2O1N2N6K9F8Hia^2"}, "label": "the man running in the middle"}]}
{"id": 122099, "image": "COCO_train2014_000000122099.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase with white flowers in the very center of other vases\"."}], "task": "refering", "answer_template": "The \"a vase with white flowers in the very center of other vases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 78, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 148, 149, 150, 172, 173, 195, 196, 218, 219, 241, 242, 264, 265], "bbox": [0.416296875, 0.17079326923076923, 0.58134375, 0.7910096153846153], "iscrowd": 0, "area": 11032.275699999997, "rle": {"size": [416, 640], "counts": "nT\\31o<3L3M4L4M2M2N1O100O1O100O1O10001N1000000O1000000O4M3\\EoNR9V1jFjNU9Z1hFgNW9\\1eFfNY9_1cFbN[9b1aF`N^9d1XFcNf9]2N2O1N2N2O1O1O1O1N2O1O1fJfLl0\\3kNmLS1U3dNUMY1l2^N^M`1d2WNeMg1]2PNjMo1X2iMoMW2S2_MUNa2m1VMZNj2g1nL`NR3b1eLeN[3]1\\LjNd3X1RLPOn3Q1jKVOW4i4000000001O001O001O001O001O001O0XLnJBS5QNlJd09S1k4UNPKOMUOa0^2c4YNSKI4@>V2[4]NWKB<J:o1T4bNXK[Od049f1k3kNXKWOh08=^1d3ROWKWOj08c0W1\\3ZOWKWOj07m0o0S3AWKXOj06V1i0i2IWKWOk04`1b0a22TKWOm02i19Y2>QKVOn00S22Q2h0mJUOP1NR8m0nFTOR1LS8P1[2O0O10001N10000O2O0N2O2M3M3M3M2Kgh\\3"}, "label": "a vase with white flowers in the very center of other vases"}]}
{"id": 122099, "image": "COCO_train2014_000000122099.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase of white flowers immediately under an outlet\"."}], "task": "refering", "answer_template": "The \"a vase of white flowers immediately under an outlet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 78, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 148, 149, 150, 172, 173, 195, 196, 218, 219, 241, 242, 264, 265], "bbox": [0.416296875, 0.17079326923076923, 0.58134375, 0.7910096153846153], "iscrowd": 0, "area": 11032.275699999997, "rle": {"size": [416, 640], "counts": "nT\\31o<3L3M4L4M2M2N1O100O1O100O1O10001N1000000O1000000O4M3\\EoNR9V1jFjNU9Z1hFgNW9\\1eFfNY9_1cFbN[9b1aF`N^9d1XFcNf9]2N2O1N2N2O1O1O1O1N2O1O1fJfLl0\\3kNmLS1U3dNUMY1l2^N^M`1d2WNeMg1]2PNjMo1X2iMoMW2S2_MUNa2m1VMZNj2g1nL`NR3b1eLeN[3]1\\LjNd3X1RLPOn3Q1jKVOW4i4000000001O001O001O001O001O001O0XLnJBS5QNlJd09S1k4UNPKOMUOa0^2c4YNSKI4@>V2[4]NWKB<J:o1T4bNXK[Od049f1k3kNXKWOh08=^1d3ROWKWOj08c0W1\\3ZOWKWOj07m0o0S3AWKXOj06V1i0i2IWKWOk04`1b0a22TKWOm02i19Y2>QKVOn00S22Q2h0mJUOP1NR8m0nFTOR1LS8P1[2O0O10001N10000O2O0N2O2M3M3M3M2Kgh\\3"}, "label": "a vase of white flowers immediately under an outlet"}]}
{"id": 433398, "image": "COCO_train2014_000000433398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a roof over the boat\"."}], "task": "refering", "answer_template": "The \"a roof over the boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 50, 52, 53, 54, 55, 56, 57, 58, 59, 73, 76, 77, 78, 79, 80, 81, 96, 99, 100, 101, 102, 103, 104, 119, 122, 123, 124, 126, 127, 142, 145, 165, 168, 188, 191, 192, 214, 215, 231, 253, 254, 276, 277, 283, 284, 285, 299, 300, 306, 307, 308, 322, 323, 330, 331, 345, 346, 353, 354, 368, 369], "bbox": [0.001953125, 0.0008541666666666666, 0.71565625, 0.9818125000000001], "iscrowd": 0, "area": 41516.6562, "rle": {"size": [480, 640], "counts": "nh0V1j=b3]L10000O10000O10000O10000O2O000O1000000O10000O10000O10000O10000OG:E;F9J7J6I7I6J7I7I6JgZS1`MPhlN1O1nG3o0NPNX1j1iNmILI]4T6hKlI6F]4X6^KkIa0C\\4[6TKkIk0@P4k6VK]IV1^Og2W8TLXH\\4l8eKRGR4X9oKfFn2JXMj9KZFX1V<iNhCl0d<UOZCe0m<\\OQC>V=CiB6^=K`BOg=2WBIo=d000000001O001O2N1O1O2N1O2N1O1O2N1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UHcMQ3^2aLUNY3l1[LfN^3[1VLVOd3k0QLEi3<lK4n3MfK?Y4B\\Kh0d4YOQKFTOQNk5[2]JL:aMY5d2[JBn0bMg4n2XJWOc1cMU4X3UJlNP2mMj3X3TJbN\\2VNIOQ1[3gLUNj2bNlNb0a1i2fLgMX3POaNb0_1h2fLZMe3^OVNc0\\1_3d0oKPNd0Z1\\3e0RLQNd0X1Y3f0ULRNd0V1h2PM]Le3<UN`0T1h2VMVL_3e0WN?R1g2o0mLoM>P1f2l0PMTN=m0d2k0QMYN=j0c2l0nL\\Na0f0b2n0iL_Ng0a0a2P1eLbNk0<a2R1`LeNQ17`2T1\\LgNV13_2U1XLkN[1N^2V1ULnN`1I\\2X1RLQOd1E[2Y1nKVOh1_O[2Z1kKYOl1[OY2\\1hK]OP2UOY2]4iMaKX2_4iM_KW2b4kM[KV2d4lMZKT2g4kMYKV2f4kMYKV2g4jMXKV2i4jMVKW2i4jMVKW2j4hMVKY2i4hMVKY2j4gMUKZ2k4fMTKZ2l4gMSKX2o4hMPKX2P5hMPKW2R5iMmJV2U5jMjJV2V5kMiJT2Y5lMfJS2[5mMeJR2]5nMbJR2_5nM`JQ2a5PN^Jo1d5PN\\Jo1e5QN[Jo1f5PNZJP2g5oMYJP2h5PNXJP2i5oMWJP2j5PNVJP2k5oMUJQ2l5nMTJQ2m5nMTJR2m5dM\\J[2e5\\MdJd2]5RMlJn2U5gLUKX3l4_L]Ka3d4ULeKj3\\4mKmKS4`700O100O100000000000000000000000000000000000000001O001O001O1O001O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O0000F:G9F:G9F:F:G9N2O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O100O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O10000001OAoAOQ>0RBNn=0VBOi=0YBOg=0[B0e=N\\B2f=J]B5X>O1O001N2O1O001OShe2"}, "label": "a roof over the boat"}]}
{"id": 433398, "image": "COCO_train2014_000000433398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat with multi colored roof carrying three passengers\"."}], "task": "refering", "answer_template": "The \"the boat with multi colored roof carrying three passengers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 50, 52, 53, 54, 55, 56, 57, 58, 59, 73, 76, 77, 78, 79, 80, 81, 96, 99, 100, 101, 102, 103, 104, 119, 122, 123, 124, 126, 127, 142, 145, 165, 168, 188, 191, 192, 214, 215, 231, 253, 254, 276, 277, 283, 284, 285, 299, 300, 306, 307, 308, 322, 323, 330, 331, 345, 346, 353, 354, 368, 369], "bbox": [0.001953125, 0.0008541666666666666, 0.71565625, 0.9818125000000001], "iscrowd": 0, "area": 41516.6562, "rle": {"size": [480, 640], "counts": "nh0V1j=b3]L10000O10000O10000O10000O2O000O1000000O10000O10000O10000O10000OG:E;F9J7J6I7I6J7I7I6JgZS1`MPhlN1O1nG3o0NPNX1j1iNmILI]4T6hKlI6F]4X6^KkIa0C\\4[6TKkIk0@P4k6VK]IV1^Og2W8TLXH\\4l8eKRGR4X9oKfFn2JXMj9KZFX1V<iNhCl0d<UOZCe0m<\\OQC>V=CiB6^=K`BOg=2WBIo=d000000001O001O2N1O1O2N1O2N1O1O2N1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1UHcMQ3^2aLUNY3l1[LfN^3[1VLVOd3k0QLEi3<lK4n3MfK?Y4B\\Kh0d4YOQKFTOQNk5[2]JL:aMY5d2[JBn0bMg4n2XJWOc1cMU4X3UJlNP2mMj3X3TJbN\\2VNIOQ1[3gLUNj2bNlNb0a1i2fLgMX3POaNb0_1h2fLZMe3^OVNc0\\1_3d0oKPNd0Z1\\3e0RLQNd0X1Y3f0ULRNd0V1h2PM]Le3<UN`0T1h2VMVL_3e0WN?R1g2o0mLoM>P1f2l0PMTN=m0d2k0QMYN=j0c2l0nL\\Na0f0b2n0iL_Ng0a0a2P1eLbNk0<a2R1`LeNQ17`2T1\\LgNV13_2U1XLkN[1N^2V1ULnN`1I\\2X1RLQOd1E[2Y1nKVOh1_O[2Z1kKYOl1[OY2\\1hK]OP2UOY2]4iMaKX2_4iM_KW2b4kM[KV2d4lMZKT2g4kMYKV2f4kMYKV2g4jMXKV2i4jMVKW2i4jMVKW2j4hMVKY2i4hMVKY2j4gMUKZ2k4fMTKZ2l4gMSKX2o4hMPKX2P5hMPKW2R5iMmJV2U5jMjJV2V5kMiJT2Y5lMfJS2[5mMeJR2]5nMbJR2_5nM`JQ2a5PN^Jo1d5PN\\Jo1e5QN[Jo1f5PNZJP2g5oMYJP2h5PNXJP2i5oMWJP2j5PNVJP2k5oMUJQ2l5nMTJQ2m5nMTJR2m5dM\\J[2e5\\MdJd2]5RMlJn2U5gLUKX3l4_L]Ka3d4ULeKj3\\4mKmKS4`700O100O100000000000000000000000000000000000000001O001O001O1O001O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O0000F:G9F:G9F:F:G9N2O1O1O1N2O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O100O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O10000001OAoAOQ>0RBNn=0VBOi=0YBOg=0[B0e=N\\B2f=J]B5X>O1O001N2O1O001OShe2"}, "label": "the boat with multi colored roof carrying three passengers"}]}
{"id": 433398, "image": "COCO_train2014_000000433398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color handbag holded by a lady\"."}], "task": "refering", "answer_template": "The \"a black color handbag holded by a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 243, 286, 287, 308, 309, 310, 332, 333, 355, 356], "bbox": [0.413375, 0.5698958333333334, 0.5866875, 0.907], "iscrowd": 0, "area": 3987.5894500000027, "rle": {"size": [480, 640], "counts": "nbl37g>3M3M3M3MIfAL\\>0fA0\\>MeA3]>IdA8`>2N2N2N2K5H7J5^BOY<OgC3X<MgC4Z<JfC8Y<]OQDe0o;YOQDh0o;XOPDj0P<WOmCk0R<WOkCk0S<XOiCj0W<XOeCk0Z<n00O10000O10000001O1O10O01O1O001O1O00N2N2N2M3N2N2N2M3M3M3N2M3M3M4L3M3M3M3M3M3L4M3L4M3L4L4M3Lhj80\\UG2N3N1N2N3M2O2M2O2O00N2N3L3N2N2N3L3Ngmk3"}, "label": "a black color handbag holded by a lady"}]}
{"id": 433398, "image": "COCO_train2014_000000433398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black bag on the woman ' s lap\"."}], "task": "refering", "answer_template": "The \"the black bag on the woman ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 243, 286, 287, 308, 309, 310, 332, 333, 355, 356], "bbox": [0.413375, 0.5698958333333334, 0.5866875, 0.907], "iscrowd": 0, "area": 3987.5894500000027, "rle": {"size": [480, 640], "counts": "nbl37g>3M3M3M3MIfAL\\>0fA0\\>MeA3]>IdA8`>2N2N2N2K5H7J5^BOY<OgC3X<MgC4Z<JfC8Y<]OQDe0o;YOQDh0o;XOPDj0P<WOmCk0R<WOkCk0S<XOiCj0W<XOeCk0Z<n00O10000O10000001O1O10O01O1O001O1O00N2N2N2M3N2N2N2M3M3M3N2M3M3M4L3M3M3M3M3M3L4M3L4M3L4L4M3Lhj80\\UG2N3N1N2N3M2O2M2O2O00N2N3L3N2N2N3L3Ngmk3"}, "label": "the black bag on the woman ' s lap"}]}
{"id": 433398, "image": "COCO_train2014_000000433398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a strap across mans chest\"."}], "task": "refering", "answer_template": "The \"a strap across mans chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 250, 251, 273, 274, 295, 296, 297, 318, 319, 341, 342, 364, 365, 386, 387, 388], "bbox": [0.817421875, 0.5550625, 0.9438124999999999, 0.9842708333333334], "iscrowd": 0, "area": 6466.071999999994, "rle": {"size": [480, 640], "counts": "Wce71l>4L3M4M2M4L3M3M4L3M4L3M4K4L5L3L5K4M3L5K4M4K4L5K4M4K4L5L3L4L5L3L5K4L5L3L1O01O0O101N10000O2O000O^OPL[Fo3i9TLPFl3T:XLdEg3b:74K5L5K4K5L5K4K5L5K4K5L5K4K5L4K6K4L4K6K4L4K6K4L4L5K4L4K6K4L_c`0"}, "label": "a strap across mans chest"}]}
{"id": 433398, "image": "COCO_train2014_000000433398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black strap that is on a young man wearing a purple and yellow hat\"."}], "task": "refering", "answer_template": "The \"a black strap that is on a young man wearing a purple and yellow hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 250, 251, 273, 274, 295, 296, 297, 318, 319, 341, 342, 364, 365, 386, 387, 388], "bbox": [0.817421875, 0.5550625, 0.9438124999999999, 0.9842708333333334], "iscrowd": 0, "area": 6466.071999999994, "rle": {"size": [480, 640], "counts": "Wce71l>4L3M4M2M4L3M3M4L3M4L3M4K4L5L3L5K4M3L5K4M4K4L5K4M4K4L5L3L4L5L3L5K4L5L3L1O01O0O101N10000O2O000O^OPL[Fo3i9TLPFl3T:XLdEg3b:74K5L5K4K5L5K4K5L5K4K5L5K4K5L4K6K4L4K6K4L4K6K4L4L5K4L4K6K4L_c`0"}, "label": "a black strap that is on a young man wearing a purple and yellow hat"}]}
{"id": 105719, "image": "COCO_train2014_000000105719.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear standing on four legs\"."}], "task": "refering", "answer_template": "The \"a bear standing on four legs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 153, 154], "bbox": [0.53690625, 0.10760416666666667, 0.763421875, 0.3939583333333333], "iscrowd": 0, "area": 11191.700999999997, "rle": {"size": [480, 640], "counts": "S[Q51m>102M3L4K5]OA_Bd0`=]O[Bf0f=[OWBg0i=YOUBh0l=60O100001M2O2M2N3M2O2M2O;F1O1O1^CSNo;n1PDTNn;m1QDVNl;k1SDVNl;j1TDWNj;k1UDVNj;k1UDVNj;k1UDVNj;l1TDUNj;o1SDRNl;Q2PDPNP<`200O10000O100000000001N1O1O2O0O1O2M2H8I80O100O00100O1O100O1O100O1O1kCcMe;^2oCmMR<a2O1O1O1O1OO1O1O1O2N1O1O1K5K51O1O10O0100O1O1dD\\M[:d2dE]M[:d2cE^M]:b2bE_M]:b2bE_M]:k2XEWMh:P3PEQMo:a30O00001O001O00001O001O00001N101O1O1O1O1O1O1O1CaDWM`;f2?M2N3M3M3M3M3M3N2M3M3M3M3M3M4L6J7I6H9G9G8HXeV2"}, "label": "a bear standing on four legs"}]}
{"id": 105719, "image": "COCO_train2014_000000105719.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large black bear on all fours\"."}], "task": "refering", "answer_template": "The \"large black bear on all fours\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 153, 154], "bbox": [0.53690625, 0.10760416666666667, 0.763421875, 0.3939583333333333], "iscrowd": 0, "area": 11191.700999999997, "rle": {"size": [480, 640], "counts": "S[Q51m>102M3L4K5]OA_Bd0`=]O[Bf0f=[OWBg0i=YOUBh0l=60O100001M2O2M2N3M2O2M2O;F1O1O1^CSNo;n1PDTNn;m1QDVNl;k1SDVNl;j1TDWNj;k1UDVNj;k1UDVNj;k1UDVNj;l1TDUNj;o1SDRNl;Q2PDPNP<`200O10000O100000000001N1O1O2O0O1O2M2H8I80O100O00100O1O100O1O100O1O1kCcMe;^2oCmMR<a2O1O1O1O1OO1O1O1O2N1O1O1K5K51O1O10O0100O1O1dD\\M[:d2dE]M[:d2cE^M]:b2bE_M]:b2bE_M]:k2XEWMh:P3PEQMo:a30O00001O001O00001O001O00001N101O1O1O1O1O1O1O1CaDWM`;f2?M2N3M3M3M3M3M3N2M3M3M3M3M3M4L6J7I6H9G9G8HXeV2"}, "label": "large black bear on all fours"}]}
{"id": 105719, "image": "COCO_train2014_000000105719.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear is looking into the distance\"."}], "task": "refering", "answer_template": "The \"a bear is looking into the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 234, 235, 236, 237], "bbox": [0.15471875, 0.26666666666666666, 0.40160937499999994, 0.6434166666666667], "iscrowd": 0, "area": 17510.97775, "rle": {"size": [480, 640], "counts": "Qd^1n0U=n0N3M3M2N3M3M2N3L4M2N3M2N2N1O2N2N2O1O0O2O1O1N101O001N101O0O2O001O0O10001O0O10000O2O00000O10001N10000O1O101N1O1O100O2N100O1O100O2N1O100O1O2O0BiKRFX4m9jKQFW4m9kKSFU4k9nKSFT4k9nKSFS4k9oKTFR4k9PLTFQ4k9b00001O001O00001O00001O00001O00001O00001O001O00001O0000001OhNbFnL^9o2iFmLW9Q3oFkLQ9R3VGjLj8V3ZGgLe8Y3_GcLa8]3`GbL`8^3`GbL`8^3aGaL_8_3bG`L^8`3cG`L[8a3fG^LZ8a3hG^LX8b3iG]LW8S2eFnNU1nNU8S2kFkNS1QOQ8R2RGiNc:T1cEfN`:X1d101N2O1N2O1N101N2N2N2M3N2N2N2N2N2N2N101N2N2N2N2O1N3N1N2O2M2O1N3Nn\\c5"}, "label": "a bear is looking into the distance"}]}
{"id": 105719, "image": "COCO_train2014_000000105719.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear looking over their shoulder\"."}], "task": "refering", "answer_template": "The \"bear looking over their shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 234, 235, 236, 237], "bbox": [0.15471875, 0.26666666666666666, 0.40160937499999994, 0.6434166666666667], "iscrowd": 0, "area": 17510.97775, "rle": {"size": [480, 640], "counts": "Qd^1n0U=n0N3M3M2N3M3M2N3L4M2N3M2N2N1O2N2N2O1O0O2O1O1N101O001N101O0O2O001O0O10001O0O10000O2O00000O10001N10000O1O101N1O1O100O2N100O1O100O2N1O100O1O2O0BiKRFX4m9jKQFW4m9kKSFU4k9nKSFT4k9nKSFS4k9oKTFR4k9PLTFQ4k9b00001O001O00001O00001O00001O00001O00001O001O00001O0000001OhNbFnL^9o2iFmLW9Q3oFkLQ9R3VGjLj8V3ZGgLe8Y3_GcLa8]3`GbL`8^3`GbL`8^3aGaL_8_3bG`L^8`3cG`L[8a3fG^LZ8a3hG^LX8b3iG]LW8S2eFnNU1nNU8S2kFkNS1QOQ8R2RGiNc:T1cEfN`:X1d101N2O1N2O1N101N2N2N2M3N2N2N2N2N2N2N101N2N2N2N2O1N3N1N2O2M2O1N3Nn\\c5"}, "label": "bear looking over their shoulder"}]}
{"id": 441599, "image": "COCO_train2014_000000441599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe walking in the savanna\"."}], "task": "refering", "answer_template": "The \"a giraffe walking in the savanna\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 56, 57, 58, 78, 79, 80, 81, 103, 104, 126, 127, 128, 149, 150, 151, 152, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 268, 269, 288, 289, 291, 292, 311, 312, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361], "bbox": [0.4181875, 0.10689583333333334, 0.7398906249999999, 0.9155416666666666], "iscrowd": 0, "area": 28335.34215, "rle": {"size": [480, 640], "counts": "gfm36i>3M3N1N1O2N2O0O1O1O100O2N001O1O10O01O1O1N2O1O1O001O100O1000000O0O2O1N2O2N2O1N3N003M?@a0@c0\\O3N8GjLcDP3Y;mLiDX3R;gLPE]3k:aLWEa3\\1`Lb7MUGd3T1fLc7EZGg3o0jLP8X3lGlLS8V3hGoLU8S3gGQMX8P3dGTM[8m2aGWM]8l2^GYM`8h2\\G\\Mc8e2YG_Mf8b2SGoMa8^4K2N2N1ULYIYOi6c0YI^Oh6=[ID2YLY6o3hIHM^L]6a3iI2I`L`6g0\\I[1`0]1DeLa7e1RIi1YOgLd7V1ZIl2d6kLbIZ3\\6bKfJb4i75L4L6I3N3L3N2M4@?]Oc0\\Oe0[Od0^Ob0@a0]Ob0_Ob04J5K4L2M<E:F4L2N1O2M3N1O?@4TIYIW5n6`JTI`5_7jIdHV6S81O1O10OUKkIk0V6POZJd0g5[OdJ:]5ElJ\\M_Nn1f6e0kKSOV4k0RLnNn3R1WLiNj3V1ZLgNh3V1\\LfNl3R1YLiNh3V1\\LdNg3[1^LPNR4P2`4O001O001O1O001O00001O0100O2O0XH^MP3d2lLbMP3_2mLcMT3^2hLeMW3\\2fLfMZ3\\2bLfM^3[2[LkMe3W2WLkMi3V2TLkMm3W2oKkMQ4V2lKlMT4V2aKSN_4n1]KUNc4m1TKZNl4g1kJ`NV5a2aIcMc6X5500O2N1O1O1UOl0I6J6K5J6K6J5N2N2N24M3M4L4L4L4oG`JY6e5]I_Jd6g5QI^Jo6l5bHXJ`7k5WHYJj7_60O2N1O1O2N1UM^HVNc7`1SISNo6]1jGoLb1^1g6a1gIYN[6f1iIQN]6m1fIgMe6X2]I^Ml6`2WIYMP7d2i2jMZDm0d=XO5L2M3N2M3N2M3M3Nen]2"}, "label": "a giraffe walking in the savanna"}]}
{"id": 441599, "image": "COCO_train2014_000000441599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger of two giraffes\"."}], "task": "refering", "answer_template": "The \"the larger of two giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 56, 57, 58, 78, 79, 80, 81, 103, 104, 126, 127, 128, 149, 150, 151, 152, 172, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 268, 269, 288, 289, 291, 292, 311, 312, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361], "bbox": [0.4181875, 0.10689583333333334, 0.7398906249999999, 0.9155416666666666], "iscrowd": 0, "area": 28335.34215, "rle": {"size": [480, 640], "counts": "gfm36i>3M3N1N1O2N2O0O1O1O100O2N001O1O10O01O1O1N2O1O1O001O100O1000000O0O2O1N2O2N2O1N3N003M?@a0@c0\\O3N8GjLcDP3Y;mLiDX3R;gLPE]3k:aLWEa3\\1`Lb7MUGd3T1fLc7EZGg3o0jLP8X3lGlLS8V3hGoLU8S3gGQMX8P3dGTM[8m2aGWM]8l2^GYM`8h2\\G\\Mc8e2YG_Mf8b2SGoMa8^4K2N2N1ULYIYOi6c0YI^Oh6=[ID2YLY6o3hIHM^L]6a3iI2I`L`6g0\\I[1`0]1DeLa7e1RIi1YOgLd7V1ZIl2d6kLbIZ3\\6bKfJb4i75L4L6I3N3L3N2M4@?]Oc0\\Oe0[Od0^Ob0@a0]Ob0_Ob04J5K4L2M<E:F4L2N1O2M3N1O?@4TIYIW5n6`JTI`5_7jIdHV6S81O1O10OUKkIk0V6POZJd0g5[OdJ:]5ElJ\\M_Nn1f6e0kKSOV4k0RLnNn3R1WLiNj3V1ZLgNh3V1\\LfNl3R1YLiNh3V1\\LdNg3[1^LPNR4P2`4O001O001O1O001O00001O0100O2O0XH^MP3d2lLbMP3_2mLcMT3^2hLeMW3\\2fLfMZ3\\2bLfM^3[2[LkMe3W2WLkMi3V2TLkMm3W2oKkMQ4V2lKlMT4V2aKSN_4n1]KUNc4m1TKZNl4g1kJ`NV5a2aIcMc6X5500O2N1O1O1UOl0I6J6K5J6K6J5N2N2N24M3M4L4L4L4oG`JY6e5]I_Jd6g5QI^Jo6l5bHXJ`7k5WHYJj7_60O2N1O1O2N1UM^HVNc7`1SISNo6]1jGoLb1^1g6a1gIYN[6f1iIQN]6m1fIgMe6X2]I^Ml6`2WIYMP7d2i2jMZDm0d=XO5L2M3N2M3N2M3M3Nen]2"}, "label": "the larger of two giraffes"}]}
{"id": 441599, "image": "COCO_train2014_000000441599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe standing next to an older giraffe\"."}], "task": "refering", "answer_template": "The \"a baby giraffe standing next to an older giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 121, 122, 143, 144, 145, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 258, 259, 279, 281, 282, 302, 304, 305], "bbox": [0.14109375, 0.26125000000000004, 0.35021874999999997, 0.7910208333333333], "iscrowd": 0, "area": 13433.581649999996, "rle": {"size": [480, 640], "counts": "S^Z11j>9H8H9aD@o7i0iGZOT8n0dGUOZ8S1]GPOc8U1UGnNk8W1kEeNO7U:[1aEL_:Z200000000000O1M3N2M3N2N2M3N2M4M2M3N2O1O1O1O2N1N2O1O1O1O1N3N1O1O1O1O1O1O1N10O1000012N4L4SFWM[7l3\\GeLZ8`3ZGQM\\8T3XGTMg8c4O1O1O100O101N10dNRHfKm7m3iHkKW7S4TIdKk6Z4o1N15L4Km0TOP1oN010O100O10[NiGaLV8Y3YH_Lg7Z3iH]LW7\\3YIZLg6`3hIXLX6`3YJWLf5c3m2K4M4L4M3_N[DFi;9]D]Oh;c0]DTOg;l0^DkNe;U1`DcNd;]1o00O10O10O10O1000O01O1O1O1O1O1O1O1O1O2N2N2F;N1O2N1O2N1O1N201O0O100O101O0O1000mkR6"}, "label": "a baby giraffe standing next to an older giraffe"}]}
{"id": 441599, "image": "COCO_train2014_000000441599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating leaves off the tree\"."}], "task": "refering", "answer_template": "The \"a giraffe eating leaves off the tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 121, 122, 143, 144, 145, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 258, 259, 279, 281, 282, 302, 304, 305], "bbox": [0.14109375, 0.26125000000000004, 0.35021874999999997, 0.7910208333333333], "iscrowd": 0, "area": 13433.581649999996, "rle": {"size": [480, 640], "counts": "S^Z11j>9H8H9aD@o7i0iGZOT8n0dGUOZ8S1]GPOc8U1UGnNk8W1kEeNO7U:[1aEL_:Z200000000000O1M3N2M3N2N2M3N2M4M2M3N2O1O1O1O2N1N2O1O1O1O1N3N1O1O1O1O1O1O1N10O1000012N4L4SFWM[7l3\\GeLZ8`3ZGQM\\8T3XGTMg8c4O1O1O100O101N10dNRHfKm7m3iHkKW7S4TIdKk6Z4o1N15L4Km0TOP1oN010O100O10[NiGaLV8Y3YH_Lg7Z3iH]LW7\\3YIZLg6`3hIXLX6`3YJWLf5c3m2K4M4L4M3_N[DFi;9]D]Oh;c0]DTOg;l0^DkNe;U1`DcNd;]1o00O10O10O10O1000O01O1O1O1O1O1O1O1O1O2N2N2F;N1O2N1O2N1O1N201O0O100O101O0O1000mkR6"}, "label": "a giraffe eating leaves off the tree"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a bench\"."}], "task": "refering", "answer_template": "The \"a man sitting on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 109, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 220, 221, 222, 223, 242, 243, 244, 245, 265, 266, 267, 289, 290], "bbox": [0.53221875, 0.22261904761904763, 0.8109531249999999, 0.9192592592592593], "iscrowd": 0, "area": 17000.368850000003, "rle": {"size": [378, 640], "counts": "jUn38a;1O1O0O2O1O1O1N2N1O2N2N2O1O1O010O1O1O1O010O1O1O1O1N1XFoNg8S1XGnNe8U1[GkNc8V1]GkNa8V1_GkN^8X1aGiN]8X1cGiNZ8Z1eGgNZ8Y1gGgNW8Z1iGgNV8Z1iGgNV8Y1jGhNT8Z1kGgNT8Y1lGhNS8X1nGhNP8Z1oGgNo7Z1QHgNn7Z1QHgNm7Z1SHgNl7Y1THhNj7Z1VHfNh7[1XHfNf7[1ZHfNd7[1\\HfNb7\\1]HeNa7\\1]HgNb7Y1\\HjNc7V1ZHnNe7S1XHPOg7P1WHSOh7m0UHWOj7i0THYOm7f0QH]On7X2O001O1O001O1O001O1ON2K5KXNfM[JU2i5PNSJj1R6ZNkIa1Y6bNdI[1_6hN^IV1d6mNZIP1h6SOUIk0l6YOQIf0P7o13M3M301O0000001O0000010O00N2XOfKnI]4R6hKeI\\4Z6b000O2O0O100O11O1O1O010O1O001O10O0103L3M3N2M3N3L3N3L3M4M2M4M2M3M4M2M4M2M4L3N0O101N1O101N101N1O101N2O1N2O1O1N2O1O1N2O1O1O1N2O1O1O1N3N1O2M3N2N1O2M3N2N1O2M3Nm0SO8G8I3M^^\\1"}, "label": "a man sitting on a bench"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the far right\"."}], "task": "refering", "answer_template": "The \"the man on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 109, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 220, 221, 222, 223, 242, 243, 244, 245, 265, 266, 267, 289, 290], "bbox": [0.53221875, 0.22261904761904763, 0.8109531249999999, 0.9192592592592593], "iscrowd": 0, "area": 17000.368850000003, "rle": {"size": [378, 640], "counts": "jUn38a;1O1O0O2O1O1O1N2N1O2N2N2O1O1O010O1O1O1O010O1O1O1O1N1XFoNg8S1XGnNe8U1[GkNc8V1]GkNa8V1_GkN^8X1aGiN]8X1cGiNZ8Z1eGgNZ8Y1gGgNW8Z1iGgNV8Z1iGgNV8Y1jGhNT8Z1kGgNT8Y1lGhNS8X1nGhNP8Z1oGgNo7Z1QHgNn7Z1QHgNm7Z1SHgNl7Y1THhNj7Z1VHfNh7[1XHfNf7[1ZHfNd7[1\\HfNb7\\1]HeNa7\\1]HgNb7Y1\\HjNc7V1ZHnNe7S1XHPOg7P1WHSOh7m0UHWOj7i0THYOm7f0QH]On7X2O001O1O001O1O001O1ON2K5KXNfM[JU2i5PNSJj1R6ZNkIa1Y6bNdI[1_6hN^IV1d6mNZIP1h6SOUIk0l6YOQIf0P7o13M3M301O0000001O0000010O00N2XOfKnI]4R6hKeI\\4Z6b000O2O0O100O11O1O1O010O1O001O10O0103L3M3N2M3N3L3N3L3M4M2M4M2M3M4M2M4M2M4L3N0O101N1O101N101N1O101N2O1N2O1O1N2O1O1N2O1O1O1N2O1O1O1N3N1O2M3N2N1O2M3N2N1O2M3Nm0SO8G8I3M^^\\1"}, "label": "the man on the far right"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit , with his arm across his chest\"."}], "task": "refering", "answer_template": "The \"a man in a suit , with his arm across his chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 101, 121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 189, 190, 191, 212, 213, 214, 236, 237, 259, 260], "bbox": [0.22398437499999999, 0.2432275132275132, 0.430640625, 0.8053174603174602], "iscrowd": 0, "area": 12008.736800000008, "rle": {"size": [378, 640], "counts": "SQe12g;?B1N1O100O100O1O101N1O100O1O100O1N2N2O1N2N2N3M2M3N2M3J6XOh0O01L3L5LRGRNW8k1fG`NV8\\1gGnNV8o0fG[ONmNn7f1PHNP8OlG8S8h1_N^LjJa3W5cLeJ]3Z5hLaJY3_5jL^JV3a5oLZJQ3g5SMUJm2j5VMSJk2m5XMPJh2o5[MoId2R6^MjId2U6`MgIa2Y6aMcIa2\\6_10O0100O1QOSKRKk4m4YKQKg4o4\\KnJd4Q5_KmJa4S5aKkJ^4U5fKhJZ4X5hKfJX4Z5kKcJU4]5P14F:H8H7J7H8I7H8I7H701O1O1O2O1N2N2O0O2N2N2O1N1O2N2O1N2N101N2N2N2N2N1O2M3N2N2N1N3N2N2N2M2N4L3M3M3M2N3M3J5K6J5L5J6L3O2N2O0O2N1O2N2N1M4L`cV4"}, "label": "a man in a suit , with his arm across his chest"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting in the middle with the light colored hat\"."}], "task": "refering", "answer_template": "The \"the man sitting in the middle with the light colored hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 169, 170, 171, 172, 192, 193, 194, 195, 215, 216, 238, 239, 260, 261, 262], "bbox": [0.332125, 0.22079365079365076, 0.554984375, 0.8260052910052911], "iscrowd": 0, "area": 12357.035799999998, "rle": {"size": [378, 640], "counts": "im^21h;1N2O100O100O100O100UOKQF5i91VF0d96\\FJ_9;`FFZ9a0eF_OU9g0jFZOP9l0jFZOQ9k0hF\\OS9h1M3L4M3L4M3M2M4O1O1O1O1O001O1O0O2O1O001G`LUHb3k7_LSHc3l7]LSHd3m7]LSHc3l7]LTHc3k7_LTHa3l7_LTHb3j7`LVH_3i7bLWH_3g7:O1O1O001O1O001O1O00hNnHkMQ7Q2_IcMa6\\2hI]MW6e2]13M2N2N3M2O2M2O2OO010O10O10O10O10O10O01000O010WOdLTI\\3k6lLmHT3T7mLkHS3P7dLeH9:S3o6jLdH4:T3o6WMPIj2n6YMPIg2P7\\MmHe2R7_MjHb2V7P101OO2L3N3M2O2N2N2N2N2N2N2N2N3M4L3M4L3aMdG\\1f8PNfGn1U9M2N3M2M4M3L3M3M3M3M3M3M3M3N2O1N2O1O1O1O1O101N2N3M2M3JcRY3"}, "label": "the man sitting in the middle with the light colored hat"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on the roadside bench with a white hat on his lap\"."}], "task": "refering", "answer_template": "The \"a man sitting on the roadside bench with a white hat on his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 169, 170, 171, 172, 192, 193, 194, 195, 215, 216, 238, 239, 260, 261, 262], "bbox": [0.332125, 0.22079365079365076, 0.554984375, 0.8260052910052911], "iscrowd": 0, "area": 12357.035799999998, "rle": {"size": [378, 640], "counts": "im^21h;1N2O100O100O100O100UOKQF5i91VF0d96\\FJ_9;`FFZ9a0eF_OU9g0jFZOP9l0jFZOQ9k0hF\\OS9h1M3L4M3L4M3M2M4O1O1O1O1O001O1O0O2O1O001G`LUHb3k7_LSHc3l7]LSHd3m7]LSHc3l7]LTHc3k7_LTHa3l7_LTHb3j7`LVH_3i7bLWH_3g7:O1O1O001O1O001O1O00hNnHkMQ7Q2_IcMa6\\2hI]MW6e2]13M2N2N3M2O2M2O2OO010O10O10O10O10O10O01000O010WOdLTI\\3k6lLmHT3T7mLkHS3P7dLeH9:S3o6jLdH4:T3o6WMPIj2n6YMPIg2P7\\MmHe2R7_MjHb2V7P101OO2L3N3M2O2N2N2N2N2N2N2N2N3M4L3M4L3aMdG\\1f8PNfGn1U9M2N3M2M4M3L3M3M3M3M3M3M3M3N2O1N2O1O1O1O1O101N2N3M2M3JcRY3"}, "label": "a man sitting on the roadside bench with a white hat on his lap"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third man from the left , holding a black hat on his lap\"."}], "task": "refering", "answer_template": "The \"the third man from the left , holding a black hat on his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 105, 106, 127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 217, 218, 219, 220, 240, 241, 242, 263, 264], "bbox": [0.43075, 0.23063492063492066, 0.65140625, 0.8100529100529101], "iscrowd": 0, "area": 12966.092849999997, "rle": {"size": [378, 640], "counts": "fUV32f;2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N3M2N2N2O1N2N2N2N2N2N2N2N2M3N2N2N2N2N2N3M2N2O1N2N2O1N2O1N2O1N2O1N100O10000O100O10000O100O100N2N2N2N2N2N1O2N2N2N2O0O2N2N2O010O1O1O1gNYL_Jg3]5bL^J_3_5hL\\JY3e5kLVJU3j5oLQJR3P6QMlIn2V6UMeIl2[6UMdIk2]6TMbIm2^6SMaIn2`6T12O001O1N101O2N1O1N2O2N1O1O1N2O2N1O1O1N3N1O2N2N2M2O9G9G=B5L5K2N1O2N2M3N2N2N2N2N2M3N2N1E<ZOe0H9H7J7HnUb2"}, "label": "the third man from the left , holding a black hat on his lap"}]}
{"id": 261381, "image": "COCO_train2014_000000261381.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit second from the right\"."}], "task": "refering", "answer_template": "The \"a man in a suit second from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 105, 106, 127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 217, 218, 219, 220, 240, 241, 242, 263, 264], "bbox": [0.43075, 0.23063492063492066, 0.65140625, 0.8100529100529101], "iscrowd": 0, "area": 12966.092849999997, "rle": {"size": [378, 640], "counts": "fUV32f;2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N3M2N2N2O1N2N2N2N2N2N2N2N2M3N2N2N2N2N2N3M2N2O1N2N2O1N2O1N2O1N2O1N100O10000O100O10000O100O100N2N2N2N2N2N1O2N2N2N2O0O2N2N2O010O1O1O1gNYL_Jg3]5bL^J_3_5hL\\JY3e5kLVJU3j5oLQJR3P6QMlIn2V6UMeIl2[6UMdIk2]6TMbIm2^6SMaIn2`6T12O001O1N101O2N1O1N2O2N1O1O1N2O2N1O1O1N3N1O2N2N2M2O9G9G=B5L5K2N1O2N2M3N2N2N2N2N2M3N2N1E<ZOe0H9H7J7HnUb2"}, "label": "a man in a suit second from the right"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three brown bananas with two stickers\"."}], "task": "refering", "answer_template": "The \"three brown bananas with two stickers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332], "bbox": [0.312359375, 0.23222916666666665, 0.884125, 0.8581041666666667], "iscrowd": 0, "area": 49226.00004999999, "rle": {"size": [480, 640], "counts": "]Sn23k>6J6J6J6J6K2M3N2M3M3N2M3N2M3M3N2M3M3N2M3M4L3L4M3M3M3N1O1O2N1O101N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O10000O10O10O10000O010O10000O10O10O10000O010O1O10O0100O010O1O010O10O0100O00100O010O10O01O100O00100O0010O01O10O01O10O01O10O01O10O01O10O01O10O01O10O01O010O1O010O10O01O010O10O01O010O1O010O010O1O010O10O01O010O10O01O00100O0010O01O0010O01O10O01O0010O01O0010O01O10O01O0010O01O001O1O0010O01O001O001O00100O001O001O0010O01O001O001O001O001O010O001O00001O001O001O001O0010O0001O00001O00001O00010O00010O00010O000010O00010O00010O0000010O0010O01O10O01O10O01O010O1O00100O0010O01O10O01O1O010O00100O001O10O01O010O1O010O1O00100O0010O01O10OO2O1O001O0O2O1O001O1N101O001O1N101N2M2O2M2N3N2M2N3N2M2O2N1N3N2N1N3N2L3M4L4L3M4L3M6J7I6J6J7I6J7J8G:F:G8G^aR1"}, "label": "three brown bananas with two stickers"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plantain that is in the middle of the two that have stickers on them\"."}], "task": "refering", "answer_template": "The \"the plantain that is in the middle of the two that have stickers on them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 329, 330, 331, 332], "bbox": [0.312359375, 0.23222916666666665, 0.884125, 0.8581041666666667], "iscrowd": 0, "area": 49226.00004999999, "rle": {"size": [480, 640], "counts": "]Sn23k>6J6J6J6J6K2M3N2M3M3N2M3N2M3M3N2M3M3N2M3M4L3L4M3M3M3N1O1O2N1O101N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O10000O10O10O10000O010O10000O10O10O10000O010O1O10O0100O010O1O010O10O0100O00100O010O10O01O100O00100O0010O01O10O01O10O01O10O01O10O01O10O01O10O01O10O01O010O1O010O10O01O010O10O01O010O1O010O010O1O010O10O01O010O10O01O00100O0010O01O0010O01O10O01O0010O01O0010O01O10O01O0010O01O001O1O0010O01O001O001O00100O001O001O0010O01O001O001O001O001O010O001O00001O001O001O001O0010O0001O00001O00001O00010O00010O00010O000010O00010O00010O0000010O0010O01O10O01O10O01O010O1O00100O0010O01O10O01O1O010O00100O001O10O01O010O1O010O1O00100O0010O01O10OO2O1O001O0O2O1O001O1N101O001O1N101N2M2O2M2N3N2M2N3N2M2O2N1N3N2N1N3N2L3M4L4L3M4L3M6J7I6J6J7I6J7J8G:F:G8G^aR1"}, "label": "the plantain that is in the middle of the two that have stickers on them"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a second banana is shown with full sticker\"."}], "task": "refering", "answer_template": "The \"a second banana is shown with full sticker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 60, 61, 62, 63, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 258, 259, 260], "bbox": [0.22178124999999999, 0.10327083333333334, 0.794015625, 0.6788958333333334], "iscrowd": 0, "area": 37603.2703, "rle": {"size": [480, 640], "counts": "`kR26_><F:N3M2N2N2N2M3N2N3M2M3N2N2M101O001N101O001O001N101O001O1O1O1O1O1O1O1O1N2O1O001O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1O001O1O001O010O1O001O00100O001O1O010O001O1O00100O001O010O1O001O010O1O00100O001O00100O001O10O01O00100O00001O010O00001O0010O0001O00001O001O00001O001O00001O001O00001O001O01O01O0010O0010O010O00010O010O01O010O0010O0010O0010O010O01O010O00010O0010O00010O00010O00010O000010O00010O00010O0001O01O01O01O01O01O01O00010O00010O00010O00010O01O0010O01O0010O00010O01O0010O01O00010O001O010O001O01O01O01O0001O01O0001O01O0001O01O01O01O0001O01O0001O01O01O0001O01O0000010O1O001O1O10O01O1O001O10O01O1O001O100O001O1O00100O001O1O00100O001O1O1O100O2N1O1O2N1O1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O3M2N2N2N2N2N2N3N1N2N2N2N2N2N3M2N2N2N2N2N2N3M4L8H8Hmhm1"}, "label": "a second banana is shown with full sticker"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second brown banana from the left\"."}], "task": "refering", "answer_template": "The \"the second brown banana from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 60, 61, 62, 63, 81, 82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 258, 259, 260], "bbox": [0.22178124999999999, 0.10327083333333334, 0.794015625, 0.6788958333333334], "iscrowd": 0, "area": 37603.2703, "rle": {"size": [480, 640], "counts": "`kR26_><F:N3M2N2N2N2M3N2N3M2M3N2N2M101O001N101O001O001N101O001O1O1O1O1O1O1O1O1N2O1O001O1O1O1O1O1O1O1O1N2O1O1O001O1O1O1O1O001O1O001O010O1O001O00100O001O1O010O001O1O00100O001O010O1O001O010O1O00100O001O00100O001O10O01O00100O00001O010O00001O0010O0001O00001O001O00001O001O00001O001O00001O001O01O01O0010O0010O010O00010O010O01O010O0010O0010O0010O010O01O010O00010O0010O00010O00010O00010O000010O00010O00010O0001O01O01O01O01O01O01O00010O00010O00010O00010O01O0010O01O0010O00010O01O0010O01O00010O001O010O001O01O01O01O0001O01O0001O01O0001O01O01O01O0001O01O0001O01O01O0001O01O0000010O1O001O1O10O01O1O001O10O01O1O001O100O001O1O00100O001O1O00100O001O1O1O100O2N1O1O2N1O1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O3M2N2N2N2N2N2N3N1N2N2N2N2N2N3M2N2N2N2N2N2N3M4L8H8Hmhm1"}, "label": "the second brown banana from the left"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of apple touching the plantains\"."}], "task": "refering", "answer_template": "The \"slice of apple touching the plantains\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 209, 210, 211, 212, 232, 233, 234], "bbox": [0.10940625, 0.24470833333333333, 0.33529687499999994, 0.6235208333333334], "iscrowd": 0, "area": 12150.430000000002, "rle": {"size": [480, 640], "counts": "PSQ12m>1N3M3L4M3L4M3L4M2N3L4M2M3N2N1O2O1N2N2O0O2N2O1N2N101N2N2O1N2N101N2N2O1N2N2O1N1O2O1N2N2O1N2N101N2N2N2N2O1N1O2N1O1O2O0O1O101O0O101O0O101N10000O2O000O2O000O2O0O10001O000000O100000O1000O10000000000O100000000O10000000001N101N2O1N101O1N101N2O1O0O2O1N2O1N3N2N2M3N2kNU1I7I7J6J7L3N2M2N3M3N1N3M2N3M2OjRW6"}, "label": "slice of apple touching the plantains"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first piece of cut apple right next to bananas\"."}], "task": "refering", "answer_template": "The \"the first piece of cut apple right next to bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 209, 210, 211, 212, 232, 233, 234], "bbox": [0.10940625, 0.24470833333333333, 0.33529687499999994, 0.6235208333333334], "iscrowd": 0, "area": 12150.430000000002, "rle": {"size": [480, 640], "counts": "PSQ12m>1N3M3L4M3L4M3L4M2N3L4M2M3N2N1O2O1N2N2O0O2N2O1N2N101N2N2O1N2N101N2N2O1N2N2O1N1O2O1N2N2O1N2N101N2N2N2N2O1N1O2N1O1O2O0O1O101O0O101O0O101N10000O2O000O2O000O2O0O10001O000000O100000O1000O10000000000O100000000O10000000001N101N2O1N101O1N101N2O1O0O2O1N2O1N3N2N2M3N2kNU1I7I7J6J7L3N2M2N3M3N1N3M2N3M2OjRW6"}, "label": "the first piece of cut apple right next to bananas"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the 3 left most apple slices\"."}], "task": "refering", "answer_template": "The \"the 3 left most apple slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140], "bbox": [0.0, 0.0, 0.24718749999999998, 0.41479166666666667], "iscrowd": 0, "area": 22534.8958, "rle": {"size": [480, 640], "counts": "d0j4U::E2O2N2M2O2O001O1O0O2O1N1O2N2N1O2N2M2O2N2N101O1O0000000000001O0000000000oJXGj3h8TL_Gg3a8WLfGc3[8[LkG`3V8^LQH\\3P8bLWHX3j7fL[HV3f7hLZHZ3f7dLZH]3g7`L[H`3f7^LZHc3g7[LYHf3h7XLYHh3h7VLXHk3i7SLWHn3j7PLVHQ4k7mKVHS4Z92N1O100O1O1O100O1O1O101N1O1O100O1O1O100O1O12O0O2N2N1O2N2N1OO1M3M3M4O01O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O2N3M3M3N2M3M3M2N3N2M3M3M3N2J6I7J8H8H9G8I7H9G8IWlQ7"}, "label": "the 3 left most apple slices"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plantain that is touching an apple slice without a sticker on it\"."}], "task": "refering", "answer_template": "The \"plantain that is touching an apple slice without a sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 32, 33, 34, 35, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 122, 123, 124, 125, 145, 146, 147, 168, 169], "bbox": [0.3040625, 0.0, 0.749640625, 0.43795833333333334], "iscrowd": 0, "area": 25624.14565, "rle": {"size": [480, 640], "counts": "dak2<c><C=D3L4M2M4L3N3L3N3L3N3L3N2M4L3N3L4M3L3N0O2O0O1O2O0O101N101N10001N10001O001O0O101O00001O0O10000000001N10000000000O10000000000O10000000000O2O000000000000000001O0000001O00001O01O0001O0001O01O01O01O01O00010O0001O01O00010O00010O00010O000010O00010O000010O00010O00010O00010O00010O0001O01O010O0100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O1O100O1O100O1O1O3N2M2N3N1N3M3N1N4L6K6I6JQoZ2"}, "label": "plantain that is touching an apple slice without a sticker on it"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana by the knife\"."}], "task": "refering", "answer_template": "The \"the banana by the knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 32, 33, 34, 35, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 122, 123, 124, 125, 145, 146, 147, 168, 169], "bbox": [0.3040625, 0.0, 0.749640625, 0.43795833333333334], "iscrowd": 0, "area": 25624.14565, "rle": {"size": [480, 640], "counts": "dak2<c><C=D3L4M2M4L3N3L3N3L3N3L3N2M4L3N3L4M3L3N0O2O0O1O2O0O101N101N10001N10001O001O0O101O00001O0O10000000001N10000000000O10000000000O10000000000O2O000000000000000001O0000001O00001O01O0001O0001O01O01O01O01O00010O0001O01O00010O00010O00010O000010O00010O000010O00010O00010O00010O00010O0001O01O010O0100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O1O100O1O100O1O1O3N2M2N3N1N3M3N1N4L6K6I6JQoZ2"}, "label": "the banana by the knife"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana on the far right\"."}], "task": "refering", "answer_template": "The \"the banana on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.42000000000000004, 0.19058333333333335, 1.0, 0.9835208333333334], "iscrowd": 0, "area": 57161.765949999986, "rle": {"size": [480, 640], "counts": "[`n32m>4M3M3M3L4M3M3M3L4M3M3L4M3M000O10000O1000000O10000O10000O100O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O101N1O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O101N100O1O100O1O100O10O01O10O01O10O0100O00100O001O1O010O1O001O00100O001O1O00100O001O1O010O1O001O1O010O1O001O00100O001O1O010O1O001O1O001N2O0O2N2O0O2O1N1O2O0O2O0O2N101N101N1O2O1N101N1O2O0O2O0O2N101N2M2N3M2O2M2N3M2N3N1N3M3M2O2M2N3M2N3N1N3M2N3M3L3L5K4L5K4L5L3L5K4L5K4L5L40O01O00010O000001O00010O0000001O0001O01O000000100O1O1O2N1O2O0O2N1O1O2N101N1O2N1O1O2O0O2N1O2N1O2O0O1O2N1O2N101N1O2N1O2bN]1QNP2PNo1PNdN"}, "label": "the banana on the far right"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first bottom banana in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the first bottom banana in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.42000000000000004, 0.19058333333333335, 1.0, 0.9835208333333334], "iscrowd": 0, "area": 57161.765949999986, "rle": {"size": [480, 640], "counts": "[`n32m>4M3M3M3L4M3M3M3L4M3M3L4M3M000O10000O1000000O10000O10000O100O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O101N1O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O101N100O1O100O1O100O10O01O10O01O10O0100O00100O001O1O010O1O001O00100O001O1O00100O001O1O010O1O001O1O010O1O001O00100O001O1O010O1O001O1O001N2O0O2N2O0O2O1N1O2O0O2O0O2N101N101N1O2O1N101N1O2O0O2O0O2N101N2M2N3M2O2M2N3M2N3N1N3M3M2O2M2N3M2N3N1N3M2N3M3L3L5K4L5K4L5L3L5K4L5K4L5L40O01O00010O000001O00010O0000001O0001O01O000000100O1O1O2N1O2O0O2N1O1O2N101N1O2N1O1O2O0O2N1O2N1O2O0O1O2N1O2N101N1O2N1O2bN]1QNP2PNo1PNdN"}, "label": "the first bottom banana in the right hand picture"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second from the right apple slice\"."}], "task": "refering", "answer_template": "The \"the second from the right apple slice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 74, 75, 76, 95, 96, 97, 98, 117, 118, 119, 120, 139, 140, 141, 142, 162, 163, 164], "bbox": [0.05475, 0.15095833333333333, 0.309328125, 0.4608541666666666], "iscrowd": 0, "area": 8160.788050000003, "rle": {"size": [480, 640], "counts": "Zc`01n>3M3M2N3N1N3M2N2N2N1O2N2N2N2N2O0O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O001O001O00001O0010O01O00001O001O00001O010O00001O001O001O00010O00001O00001O00010O0000001O000010O000001O0000O101O000O10000O10001O0O10000O1000001N10000O1000000O2O000O10000000001O0000000000000000001O0000000000000000002N2N2N2N2N1O2N2N2M3M3M2O2M3M3N2J6GaS_6"}, "label": "the second from the right apple slice"}]}
{"id": 138507, "image": "COCO_train2014_000000138507.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two upright slices of apple next to the red bananas\"."}], "task": "refering", "answer_template": "The \"the two upright slices of apple next to the red bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 74, 75, 76, 95, 96, 97, 98, 117, 118, 119, 120, 139, 140, 141, 142, 162, 163, 164], "bbox": [0.05475, 0.15095833333333333, 0.309328125, 0.4608541666666666], "iscrowd": 0, "area": 8160.788050000003, "rle": {"size": [480, 640], "counts": "Zc`01n>3M3M2N3N1N3M2N2N2N1O2N2N2N2N2O0O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O001O001O00001O0010O01O00001O001O00001O010O00001O001O001O00010O00001O00001O00010O0000001O000010O000001O0000O101O000O10000O10001O0O10000O1000001N10000O1000000O2O000O10000000001O0000000000000000001O0000000000000000002N2N2N2N2N1O2N2N2M3M3M2O2M3M3N2J6GaS_6"}, "label": "the two upright slices of apple next to the red bananas"}]}
{"id": 302353, "image": "COCO_train2014_000000302353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed closest to the back wall\"."}], "task": "refering", "answer_template": "The \"the bed closest to the back wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 191, 192], "bbox": [0.3118125, 0.49154088050314465, 0.5417343750000001, 0.7827358490566038], "iscrowd": 0, "area": 5637.118099999999, "rle": {"size": [318, 640], "counts": "bYn1<_9>B4L4N101O1O1O001O1O1O1O001O1N2O008H9G1O001O00001O0000000O1000000000N2A?00000000^Ob00O1000000000O100000000000000O010000000000000O1000000000O1000O100000000000000O100000O10000000O100000EXO`Gh0^8^O^Gb0a8>O1O001O1O1O1O1001N3N1O2N2N1O00000O010000O1000000O02O001O001O00001N101O001O00001EUGAk8?XG]Oj8b0XG\\Oi8b0:01O0O2BeF7[9HgF7a9N2O1O1Nnij2"}, "label": "the bed closest to the back wall"}]}
{"id": 302353, "image": "COCO_train2014_000000302353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"furthest bed from camera\"."}], "task": "refering", "answer_template": "The \"furthest bed from camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 191, 192], "bbox": [0.3118125, 0.49154088050314465, 0.5417343750000001, 0.7827358490566038], "iscrowd": 0, "area": 5637.118099999999, "rle": {"size": [318, 640], "counts": "bYn1<_9>B4L4N101O1O1O001O1O1O1O001O1N2O008H9G1O001O00001O0000000O1000000000N2A?00000000^Ob00O1000000000O100000000000000O010000000000000O1000000000O1000O100000000000000O100000O10000000O100000EXO`Gh0^8^O^Gb0a8>O1O001O1O1O1O1001N3N1O2N2N1O00000O010000O1000000O02O001O001O00001N101O001O00001EUGAk8?XG]Oj8b0XG\\Oi8b0:01O0O2BeF7[9HgF7a9N2O1O1Nnij2"}, "label": "furthest bed from camera"}]}
{"id": 302353, "image": "COCO_train2014_000000302353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed closest to the photographer\"."}], "task": "refering", "answer_template": "The \"the bed closest to the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252], "bbox": [0.558984375, 0.5710062893081761, 1.0, 0.9861949685534591], "iscrowd": 0, "area": 21847.947950000005, "rle": {"size": [318, 640], "counts": "Z__34f94K5L4M300O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O1N2O1O1O1O1N2O10000O1000000O10000O1000000O1000000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100N2O1O1N2O1O1N2O1O1O100O100O1O100O1O100O100O1O100O10000O1000000000000000000000000O1O1O1N2O1O1N2O1O1N2O1O1N2001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O00001O00000000001O00000000001O00000000001O00000000001O00000000001O0UNVK"}, "label": "the bed closest to the photographer"}]}
{"id": 302353, "image": "COCO_train2014_000000302353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed that is made with two beds to the left of it\"."}], "task": "refering", "answer_template": "The \"a bed that is made with two beds to the left of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252], "bbox": [0.558984375, 0.5710062893081761, 1.0, 0.9861949685534591], "iscrowd": 0, "area": 21847.947950000005, "rle": {"size": [318, 640], "counts": "Z__34f94K5L4M300O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O100O100O1N2O1O1O1O1N2O10000O1000000O10000O1000000O1000000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100N2O1O1N2O1O1N2O1O1O100O100O1O100O1O100O100O1O100O10000O1000000000000000000000000O1O1O1N2O1O1N2O1O1N2O1O1N2001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O00001O00000000001O00000000001O00000000001O00000000001O00000000001O0UNVK"}, "label": "a bed that is made with two beds to the left of it"}]}
{"id": 539924, "image": "COCO_train2014_000000539924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 127, 128, 142, 143, 157, 158, 159, 172, 173, 174, 187, 188, 202, 203, 217, 218, 232, 233, 246, 247, 248, 261, 262, 263], "bbox": [0.4473365617433414, 0.306640625, 0.6577481840193704, 0.7704375], "iscrowd": 0, "area": 12462.671949999998, "rle": {"size": [640, 413], "counts": "eRd35hc05K5J5L4L5oJVOeFn0U9YOeFl0V9ZOeFj0U9]OfFg0U9_OgFd0S9CiF`0R9GiF<R9NnCcNd1b1X:5gCdNk1[1\\:8^CdNR2W1]:X1^EkN_:X1^EjNa:Y1[EiNc:Z1ZEgNe:]1VEfNi:]1SEeNk:b1oD_NP;g1iD[Nk:`M]D\\4a0VNP;b5O200O10000^OjGZEU8f:oGVER8h:SHUEm7i:LoDeHP;Z7REeHm:[7UEeHi:[7ZEdHf:[7[EeHe:Y7]EfHd:X7_EgH`:X7bEhH^:V7eEhH\\:V7fEjHZ:T7hElHW:S7lElHT:R7nEmHS:P7PFPIP:n6SFQIm9l6VFTIi9j6[FTIf9j6\\FVId9g6_FYIb9c6bF\\I^9g5REeJb1C]9U1TEi2a0lKR14Z9R1]Ef2Z2ULZ8S1cE_2Y2\\LT8T1iEX2Y2`Lo7V1nER2]=kMjBm1Y=RNmB`0\\>^OkA;W>DoA6R>HTB2n=IZB2i=F]B9i=AXB?m=ZOUBe0b`010O1O101O1O100O2N100O10001O2N2M4M2Ndjg2"}, "label": "man"}]}
{"id": 539924, "image": "COCO_train2014_000000539924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a purple headband posing in the snow on a pair of skis\"."}], "task": "refering", "answer_template": "The \"a man wearing a purple headband posing in the snow on a pair of skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 127, 128, 142, 143, 157, 158, 159, 172, 173, 174, 187, 188, 202, 203, 217, 218, 232, 233, 246, 247, 248, 261, 262, 263], "bbox": [0.4473365617433414, 0.306640625, 0.6577481840193704, 0.7704375], "iscrowd": 0, "area": 12462.671949999998, "rle": {"size": [640, 413], "counts": "eRd35hc05K5J5L4L5oJVOeFn0U9YOeFl0V9ZOeFj0U9]OfFg0U9_OgFd0S9CiF`0R9GiF<R9NnCcNd1b1X:5gCdNk1[1\\:8^CdNR2W1]:X1^EkN_:X1^EjNa:Y1[EiNc:Z1ZEgNe:]1VEfNi:]1SEeNk:b1oD_NP;g1iD[Nk:`M]D\\4a0VNP;b5O200O10000^OjGZEU8f:oGVER8h:SHUEm7i:LoDeHP;Z7REeHm:[7UEeHi:[7ZEdHf:[7[EeHe:Y7]EfHd:X7_EgH`:X7bEhH^:V7eEhH\\:V7fEjHZ:T7hElHW:S7lElHT:R7nEmHS:P7PFPIP:n6SFQIm9l6VFTIi9j6[FTIf9j6\\FVId9g6_FYIb9c6bF\\I^9g5REeJb1C]9U1TEi2a0lKR14Z9R1]Ef2Z2ULZ8S1cE_2Y2\\LT8T1iEX2Y2`Lo7V1nER2]=kMjBm1Y=RNmB`0\\>^OkA;W>DoA6R>HTB2n=IZB2i=F]B9i=AXB?m=ZOUBe0b`010O1O101O1O100O2N100O10001O2N2M4M2Ndjg2"}, "label": "a man wearing a purple headband posing in the snow on a pair of skis"}]}
{"id": 204053, "image": "COCO_train2014_000000204053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a train with the numbers 70 - 079 written in white on the front\"."}], "task": "refering", "answer_template": "The \"the front of a train with the numbers 70 - 079 written in white on the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.46928125, 0.14156908665105386, 0.974546875, 0.9236065573770492], "iscrowd": 0, "area": 89697.6003, "rle": {"size": [427, 640], "counts": "\\\\m36\\;k1UNl1TNk1fNZ1M4M2M4L3M4M4K6J6J6K5J6J6K5J6J6J6K5J10000O10000000000O101O0000000O10000000000O1000001O000O10000000000O10000000001N100000000O10000000000O1000001O000O10000000000O10000000001N100001O0000000000000000000001O0000000000000001O0000000001O0000000000000000000000000010O0000000000000000000000000001O000000000001O00000000000001O000000000000000000000001O0001O000000000000000000000000O1O1O1O1N3N1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O2N1O1O2N1O1N2O2N1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O2M2O1N2O2N1N2O2M2O1N2O2N1N2O2M2O2M3N1O2M3N1N3N1O2M3N1O2M3N1O2N2M2O2N2M2O2N2M2O2N2N1N3N2N1N3N2N1O2M3N1O2M2O2N2N1N3N2N1N3N2N1N3N2N1O2M3N1O2M3N1O2N2M2O2N2oN\\1XNfb6"}, "label": "the front of a train with the numbers 70 - 079 written in white on the front"}]}
{"id": 204053, "image": "COCO_train2014_000000204053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"subway train number 70 - 079\"."}], "task": "refering", "answer_template": "The \"subway train number 70 - 079\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.46928125, 0.14156908665105386, 0.974546875, 0.9236065573770492], "iscrowd": 0, "area": 89697.6003, "rle": {"size": [427, 640], "counts": "\\\\m36\\;k1UNl1TNk1fNZ1M4M2M4L3M4M4K6J6J6K5J6J6K5J6J6J6K5J10000O10000000000O101O0000000O10000000000O1000001O000O10000000000O10000000001N100000000O10000000000O1000001O000O10000000000O10000000001N100001O0000000000000000000001O0000000000000001O0000000001O0000000000000000000000000010O0000000000000000000000000001O000000000001O00000000000001O000000000000000000000001O0001O000000000000000000000000O1O1O1O1N3N1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O2N1O1O2N1O1N2O2N1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O1O2N1O1O2M2O1O2M2O1N2O2N1N2O2M2O1N2O2N1N2O2M2O2M3N1O2M3N1N3N1O2M3N1O2M3N1O2N2M2O2N2M2O2N2M2O2N2N1N3N2N1N3N2N1O2M3N1O2M2O2N2N1N3N2N1N3N2N1N3N2N1O2M3N1O2M3N1O2N2M2O2N2oN\\1XNfb6"}, "label": "subway train number 70 - 079"}]}
{"id": 204053, "image": "COCO_train2014_000000204053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train on the railway station platform with number 70 - 039\"."}], "task": "refering", "answer_template": "The \"a train on the railway station platform with number 70 - 039\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.1244375, 0.1618032786885246, 0.47828125000000005, 0.7685480093676814], "iscrowd": 0, "area": 54161.32005000002, "rle": {"size": [427, 640], "counts": "X^Q1b2g:l2UM;D3N2M3N2N4Lj0WOb0^O1O0O10000O10001O0000000000001N10000000001O0000000000001O00000000000000000O1000001O0000000000000000000000000000000000000O1000000000001O0000000000000000000000000000000O100000000000000000001O000000000000000000000O10000000000000000000000000000000@hJmHX5R7jJmHV5Q7nJmHR5R7PKmHP5Q7TKmHl4Q7XKmHh4R7[KlHd4S7`KkH`4T7cKjH]4T7n0O1O10000001O1O1fJkH\\4V7bKkH^4V7`KkH`4V7^KkHc4U7ZKmHf4T7WKnHi4S7UKnHk4S7RKoHn4R7oJPIR5P7lJQIT5P7iJRIW5a7001O000000001O0O1N2N2O2M2N2O1M4I6N25K5K5L1NO1O100O2N1O100O2N1O100O1O2N100O1O101N1O1O101N1O1O100O2N1O100O101O0O10001N1000000H9cM\\2dMWS\\4"}, "label": "a train on the railway station platform with number 70 - 039"}]}
{"id": 204053, "image": "COCO_train2014_000000204053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train closest to the man standing\"."}], "task": "refering", "answer_template": "The \"the train closest to the man standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.1244375, 0.1618032786885246, 0.47828125000000005, 0.7685480093676814], "iscrowd": 0, "area": 54161.32005000002, "rle": {"size": [427, 640], "counts": "X^Q1b2g:l2UM;D3N2M3N2N4Lj0WOb0^O1O0O10000O10001O0000000000001N10000000001O0000000000001O00000000000000000O1000001O0000000000000000000000000000000000000O1000000000001O0000000000000000000000000000000O100000000000000000001O000000000000000000000O10000000000000000000000000000000@hJmHX5R7jJmHV5Q7nJmHR5R7PKmHP5Q7TKmHl4Q7XKmHh4R7[KlHd4S7`KkH`4T7cKjH]4T7n0O1O10000001O1O1fJkH\\4V7bKkH^4V7`KkH`4V7^KkHc4U7ZKmHf4T7WKnHi4S7UKnHk4S7RKoHn4R7oJPIR5P7lJQIT5P7iJRIW5a7001O000000001O0O1N2N2O2M2N2O1M4I6N25K5K5L1NO1O100O2N1O100O2N1O100O1O2N100O1O101N1O1O101N1O1O100O2N1O100O101O0O10001N1000000H9cM\\2dMWS\\4"}, "label": "the train closest to the man standing"}]}
{"id": 212247, "image": "COCO_train2014_000000212247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large elephants behind\"."}], "task": "refering", "answer_template": "The \"a large elephants behind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 227, 228, 229, 244, 245, 246, 250, 251, 252, 267, 268, 269, 274, 275, 289, 290, 291, 292, 297, 298, 312, 313, 314, 315, 320, 321], "bbox": [0.590828125, 0.004483568075117371, 1.0, 0.9235915492957748], "iscrowd": 0, "area": 68263.25030000001, "rle": {"size": [426, 640], "counts": "dcm4>i<4M2M4L4aJUO[Nm0d1UOXNo0f1SOWNo0h1SOTNQ1j1QOSNR1j1POSNS1l1oNPNU1n1mNoMU1P2mNlMW1R2kNkMX1S2jNiMY1U2jNgMZ1W2hNfMZ1Y2gNdM]1Z2eNbM^1]2dN_M_1`2cN]M_1b2cNZM`1d2cNXM`1g2bNUMa1f2dNXM^1`2jN]MY1Z2oNdMT1S2TOkMo0^1FaN=b0?]OC2l0MVOJS14POAZ1=iNYO_1g0cNVO_1h0dNVOM]NeMW2e2ZOAhNfMj1o2\\OVOSOhM[1Y3@jN^OiMm0d3C^NIjM?o3FRN4kM1Z4IgM>jM\\NFj0P5:\\Mh0jMSN;6f4m0QMR1iMjM_6R1dK\\1iM`Mf6S1]Kg1g4WNUKl1n4SNmJo1W5QNcJQ2a5oM[JR2h5nMVJP2n5PNPJn1T6RNkIk1Y6UNeIj1^6VN`Ih1d6XN^Ib1f6^N\\I\\1h6dN[IT1i6mNYIm0k6SOWIf0n6ZOUI?o6ASI8R7HPIUOWN>n8=cIUOa6k0o20000000jMoNSHQ1j7YOoGg0n7CkG=R8LgG5V82fGNW86hGJU8;iGET8?kGAR8d0lG\\OR8h0lGXOQ8l0nGTOo7Q1oGoNn7U1QHkNl7Z1RHfNk7^1THbNi7b1VH^Ng7f1WH[Nf7i1YHWNf7j1ZHVNe7l1ZHTNd7o1[HQNd7Q2[HoMd7S2[HmMc7V2\\HjMc7X2\\HhMb7[2]HeMb7]2UHkMj7k3N2O1N2O1O1N2O1O1N2O1N2O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O11O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2O2M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3M3M2N3M3M3M3M3M3MS1mN0000001O0000001O000000001O000000001OlJC[M=b2LXM4i20RM0n25nLJS3:hLFX3?cLA]3a0aL_O`3b0^L^Ob3d0]L[Oc3h0ZLXOg3i0WLWOi3k0ULUOk3m0SLSOm3P1QLoNo3S1oKmNP4V1nKjNR4X1lKhNT4[1jKdNV4^1hKbNX4`1fK`NZ4c1cK]N]4e1aK[N^4i1`KVN`4l1^KTNe0^1\\N`0m0RNgN[4_OQN\\1aM"}, "label": "a large elephants behind"}]}
{"id": 212247, "image": "COCO_train2014_000000212247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back end of a walking elephant\"."}], "task": "refering", "answer_template": "The \"the back end of a walking elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 227, 228, 229, 244, 245, 246, 250, 251, 252, 267, 268, 269, 274, 275, 289, 290, 291, 292, 297, 298, 312, 313, 314, 315, 320, 321], "bbox": [0.590828125, 0.004483568075117371, 1.0, 0.9235915492957748], "iscrowd": 0, "area": 68263.25030000001, "rle": {"size": [426, 640], "counts": "dcm4>i<4M2M4L4aJUO[Nm0d1UOXNo0f1SOWNo0h1SOTNQ1j1QOSNR1j1POSNS1l1oNPNU1n1mNoMU1P2mNlMW1R2kNkMX1S2jNiMY1U2jNgMZ1W2hNfMZ1Y2gNdM]1Z2eNbM^1]2dN_M_1`2cN]M_1b2cNZM`1d2cNXM`1g2bNUMa1f2dNXM^1`2jN]MY1Z2oNdMT1S2TOkMo0^1FaN=b0?]OC2l0MVOJS14POAZ1=iNYO_1g0cNVO_1h0dNVOM]NeMW2e2ZOAhNfMj1o2\\OVOSOhM[1Y3@jN^OiMm0d3C^NIjM?o3FRN4kM1Z4IgM>jM\\NFj0P5:\\Mh0jMSN;6f4m0QMR1iMjM_6R1dK\\1iM`Mf6S1]Kg1g4WNUKl1n4SNmJo1W5QNcJQ2a5oM[JR2h5nMVJP2n5PNPJn1T6RNkIk1Y6UNeIj1^6VN`Ih1d6XN^Ib1f6^N\\I\\1h6dN[IT1i6mNYIm0k6SOWIf0n6ZOUI?o6ASI8R7HPIUOWN>n8=cIUOa6k0o20000000jMoNSHQ1j7YOoGg0n7CkG=R8LgG5V82fGNW86hGJU8;iGET8?kGAR8d0lG\\OR8h0lGXOQ8l0nGTOo7Q1oGoNn7U1QHkNl7Z1RHfNk7^1THbNi7b1VH^Ng7f1WH[Nf7i1YHWNf7j1ZHVNe7l1ZHTNd7o1[HQNd7Q2[HoMd7S2[HmMc7V2\\HjMc7X2\\HhMb7[2]HeMb7]2UHkMj7k3N2O1N2O1O1N2O1O1N2O1N2O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O11O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2O2M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3M3M2N3M3M3M3M3M3MS1mN0000001O0000001O000000001O000000001OlJC[M=b2LXM4i20RM0n25nLJS3:hLFX3?cLA]3a0aL_O`3b0^L^Ob3d0]L[Oc3h0ZLXOg3i0WLWOi3k0ULUOk3m0SLSOm3P1QLoNo3S1oKmNP4V1nKjNR4X1lKhNT4[1jKdNV4^1hKbNX4`1fK`NZ4c1cK]N]4e1aK[N^4i1`KVN`4l1^KTNe0^1\\N`0m0RNgN[4_OQN\\1aM"}, "label": "the back end of a walking elephant"}]}
{"id": 212247, "image": "COCO_train2014_000000212247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant on a road\"."}], "task": "refering", "answer_template": "The \"a baby elephant on a road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 242, 243, 253, 254, 255, 256, 257, 259, 260, 261, 262, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 288, 289, 299, 300, 302, 303, 304, 305, 306, 307, 308, 322, 325, 326, 327, 328, 329], "bbox": [0.0, 0.2292018779342723, 0.616265625, 0.9887558685446008], "iscrowd": 0, "area": 83970.07260000004, "rle": {"size": [426, 640], "counts": "U5P8Y5100O1O1O100O1O1O100O100O00100O100O1ON3E:L5O0O2O00001O001O00001O001O00001O0O2O00001O001O00001O001O00001O0O2O001O00001O001O00001O001O1O1N2O1O1O1O1O1O1O1O001O1N2N2O1N2N2N2O1O14L3M4K4M4L3M4L3M4L4L3M4L3M4L3M4L3L5L3M4L3M4L3M4L2N000000000O101O0000000000001O000000000O101O00fLQLYMo3g2VLTMj3l2ZLPMf3P3_LkLa3U3cLgL]3Y3gLdLX3\\3mL_LS3a3QM[Lo2e3VMVLj2j3ZMRLe2o3_MmKa2R4eMiK[2W4iMfKV2Z4oMaKQ2_4SN]Km1c4XNXKh1h4\\NTKd1l4`NPK`1P5dNlJ\\1T5eNlJZ1T5fNlJZ1T5gNkJY1U5hNjJW1W5jNhJV1X5jNhJV1X5kNgJU1Y5lNfJT1Z5mNfJR1Y5oNgJQ1Y5POfJP1Z5QOeJo0[5ROdJn0\\5ROdJn0_5PO_JQ1d5mNZJT1i5jNVJV1n5fNQJZ1S6dNlI\\1W6bNgI_1\\6_NbIb1a6[N_Ie1e6XNYIi1j6UNUIk1n6SNPIn1S7oMlHR2X7kMgHU2\\7k13M3M3L5L3M3M3M3M4L3M3M3M3L5K4M3L4L4M4K2N010O0010O0001O010O00010O00010O0010O00100O101N1O101N1O101N100O2N100O2O0O1O2O0O1O2O0O101N1O101N1kK`K@`4`0bK]O`4b0cKZO^4e0fKXO[4g0gKVOZ4j0iKROX4m0kKQOV4n0mKnNT4Q1PLlNP4T1RLiNP4V1SLfNn3Y1ULeNk3[1XLaNj3^1YL_Ng3`1\\L]Ne3c1^LYNd3f1^LXNb3g1bLUN_3k1dLQN^3n1dLPN\\3o1hLmMY3S2iLkMX3S2lLhMV3X2mLaMW3_2kL\\MY3b2kLXMX3h2jLRMZ3n2iLlL[3R3iLhLZ3X3hLbL\\3^3gL\\L]3b3fLYL]3g3S301O00001O00001O00001O00001O0000001O00001O0000001O0000001O000010O000001O00001O0000001O0000100O2N1O1O1O1O1O1O1O1O1O2N1O100O1O1O1O1O1O2NhH]Mj3b2TLfMg3Y2VLjMk3U2SLlMo3S2nKPNS4o1kKSNV4l1gKWNZ4i1cKYN^4f1`K[Nb4d1[K_Ng4_1WKcNj4\\1SKgNn4X1PKiNR5V1hI0Y6`31O1O1N2O2N1L4L4M3L4L4L4L4L5K4L4L5K9H7H9G8H9G8H9G8H9G8I8G8H9G8H9G]RV3"}, "label": "a baby elephant on a road"}]}
{"id": 212247, "image": "COCO_train2014_000000212247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant walking behind a bigger elephant on a dirt surface\"."}], "task": "refering", "answer_template": "The \"a baby elephant walking behind a bigger elephant on a dirt surface\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 242, 243, 253, 254, 255, 256, 257, 259, 260, 261, 262, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 288, 289, 299, 300, 302, 303, 304, 305, 306, 307, 308, 322, 325, 326, 327, 328, 329], "bbox": [0.0, 0.2292018779342723, 0.616265625, 0.9887558685446008], "iscrowd": 0, "area": 83970.07260000004, "rle": {"size": [426, 640], "counts": "U5P8Y5100O1O1O100O1O1O100O100O00100O100O1ON3E:L5O0O2O00001O001O00001O001O00001O0O2O00001O001O00001O001O00001O0O2O001O00001O001O00001O001O1O1N2O1O1O1O1O1O1O1O001O1N2N2O1N2N2N2O1O14L3M4K4M4L3M4L3M4L4L3M4L3M4L3M4L3L5L3M4L3M4L3M4L2N000000000O101O0000000000001O000000000O101O00fLQLYMo3g2VLTMj3l2ZLPMf3P3_LkLa3U3cLgL]3Y3gLdLX3\\3mL_LS3a3QM[Lo2e3VMVLj2j3ZMRLe2o3_MmKa2R4eMiK[2W4iMfKV2Z4oMaKQ2_4SN]Km1c4XNXKh1h4\\NTKd1l4`NPK`1P5dNlJ\\1T5eNlJZ1T5fNlJZ1T5gNkJY1U5hNjJW1W5jNhJV1X5jNhJV1X5kNgJU1Y5lNfJT1Z5mNfJR1Y5oNgJQ1Y5POfJP1Z5QOeJo0[5ROdJn0\\5ROdJn0_5PO_JQ1d5mNZJT1i5jNVJV1n5fNQJZ1S6dNlI\\1W6bNgI_1\\6_NbIb1a6[N_Ie1e6XNYIi1j6UNUIk1n6SNPIn1S7oMlHR2X7kMgHU2\\7k13M3M3L5L3M3M3M3M4L3M3M3M3L5K4M3L4L4M4K2N010O0010O0001O010O00010O00010O0010O00100O101N1O101N1O101N100O2N100O2O0O1O2O0O1O2O0O101N1O101N1kK`K@`4`0bK]O`4b0cKZO^4e0fKXO[4g0gKVOZ4j0iKROX4m0kKQOV4n0mKnNT4Q1PLlNP4T1RLiNP4V1SLfNn3Y1ULeNk3[1XLaNj3^1YL_Ng3`1\\L]Ne3c1^LYNd3f1^LXNb3g1bLUN_3k1dLQN^3n1dLPN\\3o1hLmMY3S2iLkMX3S2lLhMV3X2mLaMW3_2kL\\MY3b2kLXMX3h2jLRMZ3n2iLlL[3R3iLhLZ3X3hLbL\\3^3gL\\L]3b3fLYL]3g3S301O00001O00001O00001O00001O0000001O00001O0000001O0000001O000010O000001O00001O0000001O0000100O2N1O1O1O1O1O1O1O1O1O2N1O100O1O1O1O1O1O2NhH]Mj3b2TLfMg3Y2VLjMk3U2SLlMo3S2nKPNS4o1kKSNV4l1gKWNZ4i1cKYN^4f1`K[Nb4d1[K_Ng4_1WKcNj4\\1SKgNn4X1PKiNR5V1hI0Y6`31O1O1N2O2N1L4L4M3L4L4L4L4L5K4L4L5K9H7H9G8H9G8H9G8H9G8I8G8H9G8H9G]RV3"}, "label": "a baby elephant walking behind a bigger elephant on a dirt surface"}]}
{"id": 376090, "image": "COCO_train2014_000000376090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra with rear to camera\"."}], "task": "refering", "answer_template": "The \"zebra with rear to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 235, 236, 237, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 292, 293, 294, 295, 307, 308, 309, 310, 322, 323, 325], "bbox": [0.46372365339578453, 0.6490625, 0.8651053864168617, 0.9460625], "iscrowd": 0, "area": 15667.230300000005, "rle": {"size": [640, 427], "counts": "hVl38^c0=I8K4L5K5J5L5K5K4L5K5J5L5K5K4L5L4L3L5L3M4L4L3L5L3j@TLR>P4`A[L_>d4O0O01kNYAPMh>o2_AjLb>T3fAeL[>Z3mA^LS>b3TBXLl=i3U1001N8I>B>C1N0001O0aN`A^M`>_2QBSMP>j2WBRMh=l2`BmLa=Q3fBiL\\=S3j1N2N3M2Nf@`MX=]2hBcM\\=[2bBeMa=[2[BfMh=Y2VBhMm=W2PBjMR>V2hAnM[>Q2g@kN[?k22N3M2N2K6K4K6J5L4K6J5L3L5N2000O010O10O01O001O1O101_@eMZ=]2VBSNh=T4N2O2M10000O100OWO^AVLb>a3mAYLS>_3[B[Ld=^3c1H8J5K6L4N1O2N2N1O2N1O2N2M2O2M3N1N3N1N3N1O2N2O0O2N101N2O0010O10000O100O101O2M4M3M3L4M3L10000O100O14L5K2N1N2N1POX^OFja0DT^O4?L^a0OU^O2j0BRa0<V^OOkb01V]OLlb03V]OJlb05V]OHlb07>10O0O2N1O2N2N1OkjS1"}, "label": "zebra with rear to camera"}]}
{"id": 376090, "image": "COCO_train2014_000000376090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra in front\"."}], "task": "refering", "answer_template": "The \"the zebra in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 235, 236, 237, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 292, 293, 294, 295, 307, 308, 309, 310, 322, 323, 325], "bbox": [0.46372365339578453, 0.6490625, 0.8651053864168617, 0.9460625], "iscrowd": 0, "area": 15667.230300000005, "rle": {"size": [640, 427], "counts": "hVl38^c0=I8K4L5K5J5L5K5K4L5K5J5L5K5K4L5L4L3L5L3M4L4L3L5L3j@TLR>P4`A[L_>d4O0O01kNYAPMh>o2_AjLb>T3fAeL[>Z3mA^LS>b3TBXLl=i3U1001N8I>B>C1N0001O0aN`A^M`>_2QBSMP>j2WBRMh=l2`BmLa=Q3fBiL\\=S3j1N2N3M2Nf@`MX=]2hBcM\\=[2bBeMa=[2[BfMh=Y2VBhMm=W2PBjMR>V2hAnM[>Q2g@kN[?k22N3M2N2K6K4K6J5L4K6J5L3L5N2000O010O10O01O001O1O101_@eMZ=]2VBSNh=T4N2O2M10000O100OWO^AVLb>a3mAYLS>_3[B[Ld=^3c1H8J5K6L4N1O2N2N1O2N1O2N2M2O2M3N1N3N1N3N1O2N2O0O2N101N2O0010O10000O100O101O2M4M3M3L4M3L10000O100O14L5K2N1N2N1POX^OFja0DT^O4?L^a0OU^O2j0BRa0<V^OOkb01V]OLlb03V]OJlb05V]OHlb07>10O0O2N1O2N2N1OkjS1"}, "label": "the zebra in front"}]}
{"id": 376090, "image": "COCO_train2014_000000376090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra , within a group of at least 4 other zebras , facing left\"."}], "task": "refering", "answer_template": "The \"a zebra , within a group of at least 4 other zebras , facing left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 226, 227, 228, 229, 232, 233, 234, 235, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 288, 289, 290, 291, 292, 303, 304, 305, 319], "bbox": [0.10407494145199063, 0.6355000000000001, 0.7016393442622951, 0.9286406250000001], "iscrowd": 0, "area": 16419.737350000003, "rle": {"size": [640, 427], "counts": "Qmk01fc0;M3N101N1O2O?@3M3M4Q^OiNn`0Y1k^OlNTa0W1e^OnNZa0V1^^OPOaa0h1O101N100N2M0002N2N2N2N2N2N2N3M2N21O1O1O2O0O1O1O2N1000000O2M2WAfMh;]2SDgMl;[2nCjMQ<Y2hCmMZ<Q2`CTNc<j1WC[Nk<d1oBaNQ=a1hBdNY=\\1aBiN_=X1]BlNb=U1\\BlNd=U1ZBlNg=U1UBmNk=T1SBmNm=T1QBnNo=R1oAoNQ>S1lAnNT>S1iAoNX>Q1fAPOZ>Z1[AgNf>U2U1N2N101N2O1N2N2O1N2O1N2N2O1Nh@mLf=R3PBXMQ>h2jA\\MW>d2dA`M]>`2^AdMd>[2WAiMj>W2QAmMP?S2k@QNV?\\300000000001O0XNm@cNS?W1WAeNi>U1aAgN_>X1fAdNZ>\\1jA`NW>`1lA\\NT>e1oAWNQ>i1dBbM\\=_2R20010O01O1O001O010O1O001O0010O01O00000001O01O00000001O01O0000000O101O000O10000O2O00000O10001N1000000O101O0OkM]_OP1b`0mNk_Oi0U`0TOW@c0i?YO`@c0^?[Oh@b0X?[Om@d0R?YOTAd0k>ZOZAd0f>\\O]Ab0b>]ObA`0^>@gA<W>EnA6R>ITB3k=MZBNf=1`BJ`=6T3010000O10O0100O1000O0100O100O01000O100O02O0000000O101O0000000O2O00000000001O00000000001O000000001O000000000O2N1O100O001O1O0100O01000O100O01000O100O010O1O1O001N2O1O1O001Oln^2"}, "label": "a zebra , within a group of at least 4 other zebras , facing left"}]}
{"id": 376090, "image": "COCO_train2014_000000376090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that is standing behind the zebra at front\"."}], "task": "refering", "answer_template": "The \"a zebra that is standing behind the zebra at front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 226, 227, 228, 229, 232, 233, 234, 235, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 277, 288, 289, 290, 291, 292, 303, 304, 305, 319], "bbox": [0.10407494145199063, 0.6355000000000001, 0.7016393442622951, 0.9286406250000001], "iscrowd": 0, "area": 16419.737350000003, "rle": {"size": [640, 427], "counts": "Qmk01fc0;M3N101N1O2O?@3M3M4Q^OiNn`0Y1k^OlNTa0W1e^OnNZa0V1^^OPOaa0h1O101N100N2M0002N2N2N2N2N2N2N3M2N21O1O1O2O0O1O1O2N1000000O2M2WAfMh;]2SDgMl;[2nCjMQ<Y2hCmMZ<Q2`CTNc<j1WC[Nk<d1oBaNQ=a1hBdNY=\\1aBiN_=X1]BlNb=U1\\BlNd=U1ZBlNg=U1UBmNk=T1SBmNm=T1QBnNo=R1oAoNQ>S1lAnNT>S1iAoNX>Q1fAPOZ>Z1[AgNf>U2U1N2N101N2O1N2N2O1N2O1N2N2O1Nh@mLf=R3PBXMQ>h2jA\\MW>d2dA`M]>`2^AdMd>[2WAiMj>W2QAmMP?S2k@QNV?\\300000000001O0XNm@cNS?W1WAeNi>U1aAgN_>X1fAdNZ>\\1jA`NW>`1lA\\NT>e1oAWNQ>i1dBbM\\=_2R20010O01O1O001O010O1O001O0010O01O00000001O01O00000001O01O0000000O101O000O10000O2O00000O10001N1000000O101O0OkM]_OP1b`0mNk_Oi0U`0TOW@c0i?YO`@c0^?[Oh@b0X?[Om@d0R?YOTAd0k>ZOZAd0f>\\O]Ab0b>]ObA`0^>@gA<W>EnA6R>ITB3k=MZBNf=1`BJ`=6T3010000O10O0100O1000O0100O100O01000O100O02O0000000O101O0000000O2O00000000001O00000000001O000000001O000000000O2N1O100O001O1O0100O01000O100O01000O100O010O1O1O001N2O1O1O001Oln^2"}, "label": "a zebra that is standing behind the zebra at front"}]}
{"id": 343322, "image": "COCO_train2014_000000343322.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue classic car driving down the street\"."}], "task": "refering", "answer_template": "The \"a blue classic car driving down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [274, 275, 276, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.07158333333333333, 0.70234375, 0.8363750000000001, 0.9134062500000001], "iscrowd": 0, "area": 33508.29110000001, "rle": {"size": [640, 480], "counts": "o\\f02mc02N2O1N2N2N2N2kNFa^O<^a0E`^O=_a0E^^O=aa0D^^O=aa0D]^O>ba0C\\^O?ca0B\\^O?ca0B[^O`0ca0CZ^O>fa0CX^O>ha0CW^O=ia0HQ^O9oa0l00001O00000000000000000O101O0000000QORNl_On1S`0TNl_Ol1T`0UNj_Ol1V`0UNi_Ok1W`0VNg_Ok1X`0WNf_Ok1Y`0VNf_Oj1Z`0XNc_Oi1]`0XNa_Oi1_`0XN__Oi1a`0XN]_Oi1c`0XN[_Oi1d`0YN[_Og1e`0[NX_Of1h`0[NV_Of1j`0\\NS_Oe1m`0h0O0000000O10000000001O0000000O101O1O1O1O1O1O2N1O1N2O1O1O001O000000000O101O00000000000000O001O1N2O1O1O1O1N2O001O1O1N2O1O100000O10001O000000000O10001O00000000000O101O00000000000O101O0000000000000O2O0000000000001O0O1000001O1O1O1O1N2O1O1O001O1O1O1O1N2O1O000000001O0O1000001O0000000000O100O1O1O1O1N3N1O1N2O1O1N2O1O2M2O1N2O100000001O000000000000001O000000000000001O000000000000001O000O10000000001O000000000000001O000000000000001O00001O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O0000000000000000000000000000000000000000000001O1O1O100O001O1O1O1O1O1O0000000000000001O0000000000000O2M2N2N2O1N2N2O2M2N2N2O1N2O1O2M2O1O1N2O1N2B?I6J6Icl`1"}, "label": "a blue classic car driving down the street"}]}
{"id": 343322, "image": "COCO_train2014_000000343322.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old blue car with a surfboard on top driving down the street\"."}], "task": "refering", "answer_template": "The \"an old blue car with a surfboard on top driving down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [274, 275, 276, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.07158333333333333, 0.70234375, 0.8363750000000001, 0.9134062500000001], "iscrowd": 0, "area": 33508.29110000001, "rle": {"size": [640, 480], "counts": "o\\f02mc02N2O1N2N2N2N2kNFa^O<^a0E`^O=_a0E^^O=aa0D^^O=aa0D]^O>ba0C\\^O?ca0B\\^O?ca0B[^O`0ca0CZ^O>fa0CX^O>ha0CW^O=ia0HQ^O9oa0l00001O00000000000000000O101O0000000QORNl_On1S`0TNl_Ol1T`0UNj_Ol1V`0UNi_Ok1W`0VNg_Ok1X`0WNf_Ok1Y`0VNf_Oj1Z`0XNc_Oi1]`0XNa_Oi1_`0XN__Oi1a`0XN]_Oi1c`0XN[_Oi1d`0YN[_Og1e`0[NX_Of1h`0[NV_Of1j`0\\NS_Oe1m`0h0O0000000O10000000001O0000000O101O1O1O1O1O1O2N1O1N2O1O1O001O000000000O101O00000000000000O001O1N2O1O1O1O1N2O001O1O1N2O1O100000O10001O000000000O10001O00000000000O101O00000000000O101O0000000000000O2O0000000000001O0O1000001O1O1O1O1N2O1O1O001O1O1O1O1N2O1O000000001O0O1000001O0000000000O100O1O1O1O1N3N1O1N2O1O1N2O1O2M2O1N2O100000001O000000000000001O000000000000001O000000000000001O000O10000000001O000000000000001O000000000000001O00001O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O0000000000000000000000000000000000000000000001O1O1O100O001O1O1O1O1O1O0000000000000001O0000000000000O2M2N2N2O1N2N2O2M2N2N2O1N2O1O2M2O1O1N2O1N2B?I6J6Icl`1"}, "label": "an old blue car with a surfboard on top driving down the street"}]}
{"id": 564508, "image": "COCO_train2014_000000564508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra with its face looking toward the camera\"."}], "task": "refering", "answer_template": "The \"the zebra with its face looking toward the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 84, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 131, 132, 133, 134, 146, 147, 148, 149, 162, 163, 164, 176, 177, 178, 179, 191, 192, 193, 206, 207, 208, 222, 223, 237, 238, 251, 252, 253, 267, 268], "bbox": [0.6022482435597191, 0.175515625, 1.0, 0.7691874999999999], "iscrowd": 0, "area": 25241.607200000006, "rle": {"size": [640, 427], "counts": "Y\\Q51oc01O1O2N100O1O1O1O1O1O2N1O100O1O1O1O2N1O1O1O10O0001O01O01O01O0010O0001O010O000100O10O01O100O00100O10=B`0@6K5J6K5J6J4MO100OXO[MV@e2g?`MW@`2g?eMV@Z2h?Q1N1N3N1O1N2O1N2O2M3]B\\Lh:e3WE^Lg:b3XEaLg:^3XEeLf:[3YEhLf:W3ZEkLe:T3ZEoLd:Q3[EUMa0UOo6e3_H^M8POY7e3[HdM1jNe7d3VHiMLfNn7d3RHjMJeNU8c3mGlMHdN[8c3iGnMEbNe8`3_GUNE_NR9Y3kFgNHWNc9o2WFXOMoMP:Y8O1O1N2O1O1N2O1O1O1N2O1N2O1O1N2O1N2O1N3N4LM3M2N3M3M3gLhHgIZ7_6hHWI[7n6gHhH\\7]7a25K5J6K4L5KW1hN01O1O1O1O001J6D<E;E:kNV1]Oc0I7M2N3M3M3M=C9H3L5K4L4L5K4L4L5K4L3N3L3N3M2M7J5J7J6J6I7J6J6Ah0TOiI"}, "label": "the zebra with its face looking toward the camera"}]}
{"id": 564508, "image": "COCO_train2014_000000564508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with his head up standing in the grass\"."}], "task": "refering", "answer_template": "The \"a zebra with his head up standing in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 84, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 131, 132, 133, 134, 146, 147, 148, 149, 162, 163, 164, 176, 177, 178, 179, 191, 192, 193, 206, 207, 208, 222, 223, 237, 238, 251, 252, 253, 267, 268], "bbox": [0.6022482435597191, 0.175515625, 1.0, 0.7691874999999999], "iscrowd": 0, "area": 25241.607200000006, "rle": {"size": [640, 427], "counts": "Y\\Q51oc01O1O2N100O1O1O1O1O1O2N1O100O1O1O1O2N1O1O1O10O0001O01O01O01O0010O0001O010O000100O10O01O100O00100O10=B`0@6K5J6K5J6J4MO100OXO[MV@e2g?`MW@`2g?eMV@Z2h?Q1N1N3N1O1N2O1N2O2M3]B\\Lh:e3WE^Lg:b3XEaLg:^3XEeLf:[3YEhLf:W3ZEkLe:T3ZEoLd:Q3[EUMa0UOo6e3_H^M8POY7e3[HdM1jNe7d3VHiMLfNn7d3RHjMJeNU8c3mGlMHdN[8c3iGnMEbNe8`3_GUNE_NR9Y3kFgNHWNc9o2WFXOMoMP:Y8O1O1N2O1O1N2O1O1O1N2O1N2O1O1N2O1N2O1N3N4LM3M2N3M3M3gLhHgIZ7_6hHWI[7n6gHhH\\7]7a25K5J6K4L5KW1hN01O1O1O1O001J6D<E;E:kNV1]Oc0I7M2N3M3M3M=C9H3L5K4L4L5K4L4L5K4L3N3L3N3M2M7J5J7J6J6I7J6J6Ah0TOiI"}, "label": "a zebra with his head up standing in the grass"}]}
{"id": 564508, "image": "COCO_train2014_000000564508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra farthest to the back\"."}], "task": "refering", "answer_template": "The \"zebra farthest to the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 51, 54, 55, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 87, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115], "bbox": [0.054145199063231855, 0.080703125, 0.8729039812646371, 0.34254687500000003], "iscrowd": 0, "area": 27532.052349999998, "rle": {"size": [640, 427], "counts": "f^>b0^c0010O1O001O1O001O001O1O1^^OAa?`0[@Ce?a0S@Cm?a0l_OCS`0a0f_OBZ`0a0`_OB``0b0Y_OAh`0b0Q_OAo`0h101O00000M3M4L3M3M3M3M4M2N2O1N2N100O1O1001O1O2O0O1O2N1O1O2N1O101N1O1O2N1O1O2O0O101O0O101O000O2O00001N101O001N101O001N1012M2O2M3N2M3N1N3N2M3N1N3N2M010O01O001O001O001O001O0010O01O0000001O0000001O01O0001O0000001O0000001O01O0001O0000001O0000001O01O0001O001O001O001O001O001O001O001O1O010O1O10O0100O00100O10O01O10O0100O1O010O1O010O100O00100O10O01O10O010O000O2O00001N101O0O10001O0O100000001N1000000O101O00000O10001O0O100000001N1000000O2O0000000O2OaM]_Oa1c`0_N]_Oa1c`0_N^_O`1b`0`N^_O`1a`0aN__O_1a`0aN`_O^1``0bN`_O^1``0bNa_O]1^`0dNb_O\\1^`0dNf_OX1Z`0hNj_OT1V`0mNm_Oo0S`0QOQ@k0o?UOU@g0k?YOY@c0g?]O]@?c?A^@>b?B_@=a?C`@<`?Da@;_?Ea@;_?Eb@:^?Fc@9]?Gd@7]?Ic@7]?Id@6\\?Kd@4\\?Le@3[?Me@3[?Ne@1[?Of@O[?1e@O[?1f@NZ?3f@LZ?4g@KY?5g@KZ?5f@JZ?6g@IY?7g@HZ?8g@GY?:g@DZ?<g@BZ?>h@_OY?a0j@\\OV?e0l@WOU?i0n@SOS?m0PAoNQ?Q1o11O000000M3\\Od000O100O100O2O0O100O100O100O100O2O0O2O0O1100O00100O100O00100O1O10O0100O1O10O10O10000l\\OG_b09_]OHcb08Z]OIhb07U]OJlb07Q]OKQc04l\\OMVc02h\\OO[c0:1N3M2N3M3M2NfWR1"}, "label": "zebra farthest to the back"}]}
{"id": 564508, "image": "COCO_train2014_000000564508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra that is behind the other two\"."}], "task": "refering", "answer_template": "The \"zebra that is behind the other two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 51, 54, 55, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 87, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115], "bbox": [0.054145199063231855, 0.080703125, 0.8729039812646371, 0.34254687500000003], "iscrowd": 0, "area": 27532.052349999998, "rle": {"size": [640, 427], "counts": "f^>b0^c0010O1O001O1O001O001O1O1^^OAa?`0[@Ce?a0S@Cm?a0l_OCS`0a0f_OBZ`0a0`_OB``0b0Y_OAh`0b0Q_OAo`0h101O00000M3M4L3M3M3M3M4M2N2O1N2N100O1O1001O1O2O0O1O2N1O1O2N1O101N1O1O2N1O1O2O0O101O0O101O000O2O00001N101O001N101O001N1012M2O2M3N2M3N1N3N2M3N1N3N2M010O01O001O001O001O001O0010O01O0000001O0000001O01O0001O0000001O0000001O01O0001O0000001O0000001O01O0001O001O001O001O001O001O001O001O1O010O1O10O0100O00100O10O01O10O0100O1O010O1O010O100O00100O10O01O10O010O000O2O00001N101O0O10001O0O100000001N1000000O101O00000O10001O0O100000001N1000000O2O0000000O2OaM]_Oa1c`0_N]_Oa1c`0_N^_O`1b`0`N^_O`1a`0aN__O_1a`0aN`_O^1``0bN`_O^1``0bNa_O]1^`0dNb_O\\1^`0dNf_OX1Z`0hNj_OT1V`0mNm_Oo0S`0QOQ@k0o?UOU@g0k?YOY@c0g?]O]@?c?A^@>b?B_@=a?C`@<`?Da@;_?Ea@;_?Eb@:^?Fc@9]?Gd@7]?Ic@7]?Id@6\\?Kd@4\\?Le@3[?Me@3[?Ne@1[?Of@O[?1e@O[?1f@NZ?3f@LZ?4g@KY?5g@KZ?5f@JZ?6g@IY?7g@HZ?8g@GY?:g@DZ?<g@BZ?>h@_OY?a0j@\\OV?e0l@WOU?i0n@SOS?m0PAoNQ?Q1o11O000000M3\\Od000O100O100O2O0O100O100O100O100O2O0O2O0O1100O00100O100O00100O1O10O0100O1O10O10O10000l\\OG_b09_]OHcb08Z]OIhb07U]OJlb07Q]OKQc04l\\OMVc02h\\OO[c0:1N3M2N3M3M2NfWR1"}, "label": "zebra that is behind the other two"}]}
{"id": 564508, "image": "COCO_train2014_000000564508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing in front of two others from the camera ' s view\"."}], "task": "refering", "answer_template": "The \"a zebra standing in front of two others from the camera ' s view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 204, 205, 206, 211, 212, 213, 214, 215, 219, 220, 221, 227, 228, 229, 230, 235, 236, 242, 243, 244, 245, 256, 257, 258, 259, 260, 271, 272, 273, 274, 275, 289, 290], "bbox": [0.006744730679156909, 0.26965625000000004, 0.8184543325526933, 0.8561875000000001], "iscrowd": 0, "area": 66930.2532, "rle": {"size": [640, 427], "counts": "aS2e0cb0k0L4L4L4L4L8H8H7I8H6J4L4L4L4L4L3M4L4L4L4L3M3M3M3M3M3M4L3M3M3M3M3M3N3L3N2N2M3N3L4M3L4M3L4M3L5]FTIj5o6nIVIR6m6gIWIX6m6`IXI`6k6YIYIf6k6SIYIm6j6kH[IT7i6eH[I[7h6^H\\Ia7b900O10000O100000fJbGQO_8P1cGkN_8U1cGgN_8Y1dGbN^8_1cG^N^8b1eGYN]8h1dGTN^8l1eG_LkN\\Ob9V4dGoKC]Ok8d4eG^K:@S8S5dGnJZ1@m6b5lG]Je15[6^5UKhJf4Y5YKkJc4U5\\KPK`4Q5_KTK\\4l4cKWK[4j4dKWK[4i4dKXK\\4i4cKWK]4i4bKYK]4h4bKXK^4h4dHaJU1h0W6g4WHQK_18Z6h4jG^Ki1L\\6f4^GnKS2\\O`6f6]IZId6i6YIVIi6k6TITIo6m6oHRIR7Q7fHSI\\7n6ZGXJh8_82M3N1N3N2M2O2M2O2N2M2M4M3M2M4M2N2N2M3N3M2M3N2N2NoETGR9i8jFbGR9\\8hFQHS9m7gF_HU9_7fFmHV9o6eF^IV9j8L5K4L2N0000000001O00000000000000001O0WMRI`Hn6Y7[IeHe6U7cIiH]6P7lInHT6k6UJTIj5e6^JZIb5_6gJ_IY5Z6PKdIQ5U6WKiIi4U6[KiIe4T6`KjI`4T6cKkI]4S6gKkIY4S6kKkIU4S6oKkIQ4S6o4M4M2N2N2N2M3N2N2N2N3L3N2N2N2N2N2O101N100O2O0O2O0O101N100O2O000O2O0O101N101N100O2O0O101N100O2O0O101N101N100O2O00001O00001O00001O001O00001O0O101O0000001O0000000000001O00000000000011N2N3M3M2N3M2N3N1N3M3M2N3M2N3N1N3M3M2N3M2N3N2M2N3M2N3M2N3N2M2N3M001O00100O001O001O1O0010O01O1O0100O10000O1O1O1N2O1O1N2O1N2O1N2O1O1N2O0O2O1N2O1O1N2J6F:E;F:E;A?A?A?A?L4M3N2M3N2M5K6K4K5L5J5L4K5L5Jn^`1"}, "label": "a zebra standing in front of two others from the camera ' s view"}]}
{"id": 564508, "image": "COCO_train2014_000000564508.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in front of two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra in front of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 204, 205, 206, 211, 212, 213, 214, 215, 219, 220, 221, 227, 228, 229, 230, 235, 236, 242, 243, 244, 245, 256, 257, 258, 259, 260, 271, 272, 273, 274, 275, 289, 290], "bbox": [0.006744730679156909, 0.26965625000000004, 0.8184543325526933, 0.8561875000000001], "iscrowd": 0, "area": 66930.2532, "rle": {"size": [640, 427], "counts": "aS2e0cb0k0L4L4L4L4L8H8H7I8H6J4L4L4L4L4L3M4L4L4L4L3M3M3M3M3M3M4L3M3M3M3M3M3N3L3N2N2M3N3L4M3L4M3L4M3L5]FTIj5o6nIVIR6m6gIWIX6m6`IXI`6k6YIYIf6k6SIYIm6j6kH[IT7i6eH[I[7h6^H\\Ia7b900O10000O100000fJbGQO_8P1cGkN_8U1cGgN_8Y1dGbN^8_1cG^N^8b1eGYN]8h1dGTN^8l1eG_LkN\\Ob9V4dGoKC]Ok8d4eG^K:@S8S5dGnJZ1@m6b5lG]Je15[6^5UKhJf4Y5YKkJc4U5\\KPK`4Q5_KTK\\4l4cKWK[4j4dKWK[4i4dKXK\\4i4cKWK]4i4bKYK]4h4bKXK^4h4dHaJU1h0W6g4WHQK_18Z6h4jG^Ki1L\\6f4^GnKS2\\O`6f6]IZId6i6YIVIi6k6TITIo6m6oHRIR7Q7fHSI\\7n6ZGXJh8_82M3N1N3N2M2O2M2O2N2M2M4M3M2M4M2N2N2M3N3M2M3N2N2NoETGR9i8jFbGR9\\8hFQHS9m7gF_HU9_7fFmHV9o6eF^IV9j8L5K4L2N0000000001O00000000000000001O0WMRI`Hn6Y7[IeHe6U7cIiH]6P7lInHT6k6UJTIj5e6^JZIb5_6gJ_IY5Z6PKdIQ5U6WKiIi4U6[KiIe4T6`KjI`4T6cKkI]4S6gKkIY4S6kKkIU4S6oKkIQ4S6o4M4M2N2N2N2M3N2N2N2N3L3N2N2N2N2N2O101N100O2O0O2O0O101N100O2O000O2O0O101N101N100O2O0O101N100O2O0O101N101N100O2O00001O00001O00001O001O00001O0O101O0000001O0000000000001O00000000000011N2N3M3M2N3M2N3N1N3M3M2N3M2N3N1N3M3M2N3M2N3N2M2N3M2N3M2N3N2M2N3M001O00100O001O001O1O0010O01O1O0100O10000O1O1O1N2O1O1N2O1N2O1N2O1O1N2O0O2O1N2O1O1N2J6F:E;F:E;A?A?A?A?L4M3N2M3N2M5K6K4K5L5J5L4K5L5Jn^`1"}, "label": "a zebra in front of two other zebras"}]}
{"id": 72987, "image": "COCO_train2014_000000072987.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a blue and green jacket riding a back\"."}], "task": "refering", "answer_template": "The \"a man with a blue and green jacket riding a back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 83, 84, 85, 101, 102, 103, 119, 120, 121, 138, 139, 156, 157], "bbox": [0.6200800000000001, 0.31763473053892216, 0.76536, 0.7458982035928143], "iscrowd": 0, "area": 4940.534850000001, "rle": {"size": [334, 500], "counts": "fXU34U:6J5M3M3M3N2N2N2N2O1N201N2O0O3M2O1N2O1N100O2N1O3N5J5^OSNZHR2b7QNZHU2`7a0K5L2N0O2O00001O01O10O01O010O1O01`0`0POO010O1O10O0100_MTI^1m6bNSI]1n6cNSI\\1n6cNRI\\1o6dNRI[1o6dNRIX1Q7hNoHo0[7POfHf0d7YO]H=m7BSH5W8JoGKW84oGBV8=PHXOV8f0k0O1O1N2K5DZPV1"}, "label": "a man with a blue and green jacket riding a back"}]}
{"id": 72987, "image": "COCO_train2014_000000072987.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green and blue jacket riding a bike\"."}], "task": "refering", "answer_template": "The \"a man in a green and blue jacket riding a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 83, 84, 85, 101, 102, 103, 119, 120, 121, 138, 139, 156, 157], "bbox": [0.6200800000000001, 0.31763473053892216, 0.76536, 0.7458982035928143], "iscrowd": 0, "area": 4940.534850000001, "rle": {"size": [334, 500], "counts": "fXU34U:6J5M3M3M3N2N2N2N2O1N201N2O0O3M2O1N2O1N100O2N1O3N5J5^OSNZHR2b7QNZHU2`7a0K5L2N0O2O00001O01O10O01O010O1O01`0`0POO010O1O10O0100_MTI^1m6bNSI]1n6cNSI\\1n6cNRI\\1o6dNRI[1o6dNRIX1Q7hNoHo0[7POfHf0d7YO]H=m7BSH5W8JoGKW84oGBV8=PHXOV8f0k0O1O1N2K5DZPV1"}, "label": "a man in a green and blue jacket riding a bike"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe , to the right of another giraffe , leaning forward toward a person\"."}], "task": "refering", "answer_template": "The \"a giraffe , to the right of another giraffe , leaning forward toward a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 262, 263, 265, 266, 267, 268, 289, 290, 291, 313, 314, 315, 336, 337, 338, 360], "bbox": [0.427359375, 0.30293749999999997, 0.7111562499999999, 0.9470625], "iscrowd": 0, "area": 23207.797799999997, "rle": {"size": [480, 640], "counts": "lgP46g>5K3L4M2N3L3J7gNWOQDR1n;o000O1O1O1O1O2N2N3M2H8@a0E:O1O1010O1O101N1O100O1O101N1Of0[O4K4L4M00000000000000N2N2O1N2O1N2O1N3N1I7F;L3O1O1O2O0O1O2N1O1O1O2N1O1O2O1N3M3M3M3M3M3M3L4L3N2M3M4M2M3O1000010O00000001O0001O01O001O01O010O010O10O01O010O010O0010O001C<101M3N1O2N1N3N1O2N1N3N2N4K6K5K5K5J3N2N1O1N3N1O1N3N101N1O101N101N1O10YNf1ZNdFoLU:k2RFTMm9f2ZF[Me9_2aFaMb9U2eFkM]9m1jFRNY9d1nF\\NU9[1QGeNR9Q1VGnNm8i0YGWOi87gGI\\8AWH?V;O0O2N1O2M2O2N1O1N3N1O2M2O2N1OdZf2"}, "label": "a giraffe , to the right of another giraffe , leaning forward toward a person"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bent over giraffe to the right of other bent over giraffe\"."}], "task": "refering", "answer_template": "The \"bent over giraffe to the right of other bent over giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 262, 263, 265, 266, 267, 268, 289, 290, 291, 313, 314, 315, 336, 337, 338, 360], "bbox": [0.427359375, 0.30293749999999997, 0.7111562499999999, 0.9470625], "iscrowd": 0, "area": 23207.797799999997, "rle": {"size": [480, 640], "counts": "lgP46g>5K3L4M2N3L3J7gNWOQDR1n;o000O1O1O1O1O2N2N3M2H8@a0E:O1O1010O1O101N1O100O1O101N1Of0[O4K4L4M00000000000000N2N2O1N2O1N2O1N3N1I7F;L3O1O1O2O0O1O2N1O1O1O2N1O1O2O1N3M3M3M3M3M3M3L4L3N2M3M4M2M3O1000010O00000001O0001O01O001O01O010O010O10O01O010O010O0010O001C<101M3N1O2N1N3N1O2N1N3N2N4K6K5K5K5J3N2N1O1N3N1O1N3N101N1O101N101N1O10YNf1ZNdFoLU:k2RFTMm9f2ZF[Me9_2aFaMb9U2eFkM]9m1jFRNY9d1nF\\NU9[1QGeNR9Q1VGnNm8i0YGWOi87gGI\\8AWH?V;O0O2N1O2M2O2N1O1N3N1O2M2O2N1OdZf2"}, "label": "bent over giraffe to the right of other bent over giraffe"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the left of the other three giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe to the left of the other three giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 238, 239, 256, 257, 258, 259, 261, 262, 279, 280, 281, 282, 302, 303, 304, 325, 326, 327, 348, 349, 350, 371, 372, 373], "bbox": [0.12471874999999999, 0.3004375, 0.431453125, 0.9768333333333333], "iscrowd": 0, "area": 30086.91494999999, "rle": {"size": [480, 640], "counts": "ogU1>Q>d0C=L4K4M4K5K5L3L5K4M3L4M4K4L4M4K4L4M5K7H9nFlKi6\\4hHmKU7[4]HoK`7Y4PHRLm7V4dGSLc1mNj4`6nJfIQ5]6gJhIX5[6QJ[Jn5[701N101O0O101N10001O002N3M3M3M3M2N1OJ`JoF^5b8SK^Gk4S8gKlGW4d7[L\\Hc3i4eL_Kf5b4[21O001O001O1O001O1O001O00o0POa0@a0_Og2[M0O01O0010O01O0011N10^NXKkHg4o6kKgHT4S7]LbHb3Y7oL\\HQ3^7[2K6J6J5K6J5K6J5J7J5N3L3M4M2M4L3N3QMYF`0k9_OZF9i9H[F1h9O\\FIh96]F@i9`0\\FdN\\:Z1iEkMn:U2o0O2O00001O00001O00001O00001O00010O00010O0002O0O1O100O1O2O0O1BmMlCT2n;SNQDm1i;ZNUDf1g;aNVD_1j;m0N1O100O1O2N100O1O2O0O1O2O1N2N2O1O100O1O100O100O100O01N1O2N1OL4M3M3L3N3M200101N101N1O101N101N2O0O2N2J5I8G^1cNk\\Z5"}, "label": "a giraffe to the left of the other three giraffes"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the left side closest to the fence\"."}], "task": "refering", "answer_template": "The \"the giraffe on the left side closest to the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 238, 239, 256, 257, 258, 259, 261, 262, 279, 280, 281, 282, 302, 303, 304, 325, 326, 327, 348, 349, 350, 371, 372, 373], "bbox": [0.12471874999999999, 0.3004375, 0.431453125, 0.9768333333333333], "iscrowd": 0, "area": 30086.91494999999, "rle": {"size": [480, 640], "counts": "ogU1>Q>d0C=L4K4M4K5K5L3L5K4M3L4M4K4L4M4K4L4M5K7H9nFlKi6\\4hHmKU7[4]HoK`7Y4PHRLm7V4dGSLc1mNj4`6nJfIQ5]6gJhIX5[6QJ[Jn5[701N101O0O101N10001O002N3M3M3M3M2N1OJ`JoF^5b8SK^Gk4S8gKlGW4d7[L\\Hc3i4eL_Kf5b4[21O001O001O1O001O1O001O00o0POa0@a0_Og2[M0O01O0010O01O0011N10^NXKkHg4o6kKgHT4S7]LbHb3Y7oL\\HQ3^7[2K6J6J5K6J5K6J5J7J5N3L3M4M2M4L3N3QMYF`0k9_OZF9i9H[F1h9O\\FIh96]F@i9`0\\FdN\\:Z1iEkMn:U2o0O2O00001O00001O00001O00001O00010O00010O0002O0O1O100O1O2O0O1BmMlCT2n;SNQDm1i;ZNUDf1g;aNVD_1j;m0N1O100O1O2N100O1O2O0O1O2O1N2N2O1O100O1O100O100O100O01N1O2N1OL4M3M3L3N3M200101N101N1O101N101N2O0O2N2J5I8G^1cNk\\Z5"}, "label": "the giraffe on the left side closest to the fence"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"long black hair from back of head of girl looking at giraffes\"."}], "task": "refering", "answer_template": "The \"long black hair from back of head of girl looking at giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 358, 359, 360, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.295609375, 0.6407708333333333, 0.68075, 0.9853541666666668], "iscrowd": 0, "area": 26566.778800000004, "rle": {"size": [480, 640], "counts": "hQi21m>2O1N2O1N2N2O1N2O1N2N20000000000001O00000000001OO100O1O1L4K5L4K5L4L4K5M3L4M3L4M3L4M3M3L4M3L4M3M3N2N2N2M3N2N2N2M2O2N2N2M3N2N2O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O10000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O1000000O10000000000000000001O00000000001O0000000000001O00000000001O001N2O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M2NZUo2"}, "label": "long black hair from back of head of girl looking at giraffes"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a woman ` s head with dark brown hair\"."}], "task": "refering", "answer_template": "The \"the back of a woman ` s head with dark brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 358, 359, 360, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.295609375, 0.6407708333333333, 0.68075, 0.9853541666666668], "iscrowd": 0, "area": 26566.778800000004, "rle": {"size": [480, 640], "counts": "hQi21m>2O1N2O1N2N2O1N2O1N2N20000000000001O00000000001OO100O1O1L4K5L4K5L4L4K5M3L4M3L4M3L4M3M3L4M3L4M3M3N2N2N2M3N2N2N2M2O2N2N2M3N2N2O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O10000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O1000000O10000000000000000001O00000000001O0000000000001O00000000001O001N2O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M2NZUo2"}, "label": "the back of a woman ` s head with dark brown hair"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the right of three other giraffes looking away from the people\"."}], "task": "refering", "answer_template": "The \"a giraffe to the right of three other giraffes looking away from the people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 110, 111, 112, 134, 135, 136, 157, 158, 159, 180, 181, 182, 204, 205, 227, 228, 250, 251, 273, 274, 295, 296, 297, 318, 319, 320], "bbox": [0.7923281249999999, 0.1755, 0.9463906249999999, 0.827875], "iscrowd": 0, "area": 11352.923850000016, "rle": {"size": [480, 640], "counts": "hh]7>^>5K3B?N101N1O2N101N11O2O1O1O100O01O0010O100O1001010001N100M4L3N2M4L3M4M2M3M4M2O1O2N1O2M5K5^HoMj1U2lMQNT2T2dMPN[2U2\\MoMe2U2SMoMl2V2lLjMW3[2`LcMf3e2lK\\MY4U3iJTM[5]3gIkL^6h54J6I7J6I8H7J6I7J6I7gNY1N2N2OPHgKZ5Z4_JQL\\5P4]J\\L]5e3]JcL`5^3YJhLg5X61O100O1O1N2M3H7J7UOk0@`0@?M4N2M3N2N1N3N2M9Ei0WOi0WOi0XOh0WOi0WO]e?"}, "label": "a giraffe to the right of three other giraffes looking away from the people"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right side last zebra which is facing to the other zebra\"."}], "task": "refering", "answer_template": "The \"right side last zebra which is facing to the other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 110, 111, 112, 134, 135, 136, 157, 158, 159, 180, 181, 182, 204, 205, 227, 228, 250, 251, 273, 274, 295, 296, 297, 318, 319, 320], "bbox": [0.7923281249999999, 0.1755, 0.9463906249999999, 0.827875], "iscrowd": 0, "area": 11352.923850000016, "rle": {"size": [480, 640], "counts": "hh]7>^>5K3B?N101N1O2N101N11O2O1O1O100O01O0010O100O1001010001N100M4L3N2M4L3M4M2M3M4M2O1O2N1O2M5K5^HoMj1U2lMQNT2T2dMPN[2U2\\MoMe2U2SMoMl2V2lLjMW3[2`LcMf3e2lK\\MY4U3iJTM[5]3gIkL^6h54J6I7J6I8H7J6I7J6I7gNY1N2N2OPHgKZ5Z4_JQL\\5P4]J\\L]5e3]JcL`5^3YJhLg5X61O100O1O1N2M3H7J7UOk0@`0@?M4N2M3N2N1N3N2M9Ei0WOi0WOi0XOh0WOi0WO]e?"}, "label": "right side last zebra which is facing to the other zebra"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spec woman\"."}], "task": "refering", "answer_template": "The \"a spec woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 276, 277, 278, 299, 300, 322, 323, 345, 346, 347, 348, 368, 369, 370, 371, 372], "bbox": [0.0, 0.5858541666666667, 0.19415625, 1.0], "iscrowd": 0, "area": 11238.420650000002, "rle": {"size": [480, 640], "counts": "m8T4k:P2QN0O10000O100O100000000000000001O0000000000000000001O001O1O001O001O1O001O1O0`NPGPMS9e2VGZMm8h1XGnL3Z1g8b1QH]NR8]1QHcNQ8X1RHhNQ8Q1THnNo7l0THTOn7XOhF=]1;n7iNRGh0T1>j8@XG`0h8^O[Ga0e8^O\\Gb0d8]O]Gc0d8ZO_Ge0a8ZO`Gf0`8YOaGg0`8VObGj0^8UOcGk0^8ROdGn0]8nNfGR1[8jNhGV1R;O00O100000000000000000000001O001O001O001O001O001O001O001O001O001O002N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O001O1O1O1O1O1O1O1O001OQ]a7"}, "label": "a spec woman"}]}
{"id": 130339, "image": "COCO_train2014_000000130339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady with short curly hair and glasses\"."}], "task": "refering", "answer_template": "The \"lady with short curly hair and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 276, 277, 278, 299, 300, 322, 323, 345, 346, 347, 348, 368, 369, 370, 371, 372], "bbox": [0.0, 0.5858541666666667, 0.19415625, 1.0], "iscrowd": 0, "area": 11238.420650000002, "rle": {"size": [480, 640], "counts": "m8T4k:P2QN0O10000O100O100000000000000001O0000000000000000001O001O1O001O001O1O001O1O0`NPGPMS9e2VGZMm8h1XGnL3Z1g8b1QH]NR8]1QHcNQ8X1RHhNQ8Q1THnNo7l0THTOn7XOhF=]1;n7iNRGh0T1>j8@XG`0h8^O[Ga0e8^O\\Gb0d8]O]Gc0d8ZO_Ge0a8ZO`Gf0`8YOaGg0`8VObGj0^8UOcGk0^8ROdGn0]8nNfGR1[8jNhGV1R;O00O100000000000000000000001O001O001O001O001O001O001O001O001O001O002N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O001O1O1O1O1O1O1O1O001OQ]a7"}, "label": "lady with short curly hair and glasses"}]}
{"id": 335140, "image": "COCO_train2014_000000335140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop screen\"."}], "task": "refering", "answer_template": "The \"a laptop screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 200, 201, 202, 203], "bbox": [0.13169999999999998, 0.6898798798798799, 0.36966, 0.9752852852852852], "iscrowd": 0, "area": 8808.72645, "rle": {"size": [333, 500], "counts": "jhe05R:5K6I7J6J6I7J5K6I7J6J6I6K6M300000O1000000000O0100000000O100000000O100000000O10000000O10O100000000O100000000O10000000000O10O100000O100000000O10000000000O10000000O0100000000O10000000000O100000000O1000O1000O10000000000O100000000O=D>B>B>B>AfTV3"}, "label": "a laptop screen"}]}
{"id": 335140, "image": "COCO_train2014_000000335140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop computer in the foreground\"."}], "task": "refering", "answer_template": "The \"the laptop computer in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 200, 201, 202, 203], "bbox": [0.13169999999999998, 0.6898798798798799, 0.36966, 0.9752852852852852], "iscrowd": 0, "area": 8808.72645, "rle": {"size": [333, 500], "counts": "jhe05R:5K6I7J6J6I7J5K6I7J6J6I6K6M300000O1000000000O0100000000O100000000O100000000O10000000O10O100000000O100000000O10000000000O10O100000O100000000O10000000000O10000000O0100000000O10000000000O100000000O1000O1000O10000000000O100000000O=D>B>B>B>AfTV3"}, "label": "the laptop computer in the foreground"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black color stripped coat and sandal color pant with sports bat in his hands sitting in a chair\"."}], "task": "refering", "answer_template": "The \"a man with black color stripped coat and sandal color pant with sports bat in his hands sitting in a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 148, 149, 165, 166, 167, 181, 182, 183, 184, 198, 199, 200, 201, 215, 216, 218, 232, 233, 234, 235, 249, 250, 252, 265, 266, 267, 282, 283, 284, 299, 300], "bbox": [0.5905208333333333, 0.327859375, 0.8810625, 0.76475], "iscrowd": 0, "area": 14860.48065, "rle": {"size": [640, 480], "counts": "nna52nc01N1O2N2N101N2N2N101N2N1O2O0O1O1O010O1O1PO]Ob^Oc0Va0Gi^O9Sa0Ll^O4Pa00o^O1n`03Q_OLl`08T_OHh`0=W_OCe`0b0Z_O^Ob`0f0^_OYO_`0l0`_OTO\\`0Q1c_OnNZ`0V1f_OjNU`0\\1j_OcNR`0b1n_O^Nn?g1Q@YNj?m1U@RNg?S2Y@mMX?c2f@]M[?d2d@\\M\\?d2d@[M]?f2b@ZMl>_OYAW3KYMi>E[AS3OTMc>L^AQ32nL^>5^Ao27hLh=l0QB]29cL`=X1VBV2Z>PNfAR2S>TNlAm1R>TNnAm1P>UNnAm1P>TNlAQ2Q>Q2N3M2N2O1N2N2O1N2N01O010O00001O00bNQJlDP6S;[JcDf1YO`1T<TMRDc1NU1P<3TDIl;6YDFg;:]DBc;>aDYM:2U;d2cDYM;OS;g2dDYM;MR;i2dDYM>Jo:l2dDZM?Gn:m2fDZM`0Ek:P3fD[Mb0Bh:R3hD[Mb0@g:U3gDZMf0]Oe:X3eD[Mh0ZOd:Z3fD[Mi0WOb:^3eDZM\\=f2eBYM[=g2eBYM[=g2eBXM\\=h2eBWM[=h2fBWM\\=h2dBXM]=g2dBXM\\=h2dBWM^=h2cBWM^=h2bBXM_=f2bBYM`=f2aBYMc=c2]B]Me>a1[A^Nh>`1WAbNi>]1WAeNh>Y1YAiNg>U1XAnNg>Q1YAQOe>o0[ASOc>m0]AUOb>j0]AYOa>g0_A[O_>e0aA^O]>b0bA_O]>a0bA@^>`0bA_O`>`0`A_Oa>b0]A^Od>b0\\A]Of>b0ZA]Og>d0XA[Oi>e0VA[Oe>k0[AUOg>j0XAUOk>i0TAWOo>g0QAXOR?f0n@YOU?f0j@YOY?e0k@VOW?j0Q2N4M2N3M2N2N3M9F;F:FgYS1"}, "label": "a man with black color stripped coat and sandal color pant with sports bat in his hands sitting in a chair"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a man sitting and holding a racquet\"."}], "task": "refering", "answer_template": "The \"this is a man sitting and holding a racquet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 148, 149, 165, 166, 167, 181, 182, 183, 184, 198, 199, 200, 201, 215, 216, 218, 232, 233, 234, 235, 249, 250, 252, 265, 266, 267, 282, 283, 284, 299, 300], "bbox": [0.5905208333333333, 0.327859375, 0.8810625, 0.76475], "iscrowd": 0, "area": 14860.48065, "rle": {"size": [640, 480], "counts": "nna52nc01N1O2N2N101N2N2N101N2N1O2O0O1O1O010O1O1PO]Ob^Oc0Va0Gi^O9Sa0Ll^O4Pa00o^O1n`03Q_OLl`08T_OHh`0=W_OCe`0b0Z_O^Ob`0f0^_OYO_`0l0`_OTO\\`0Q1c_OnNZ`0V1f_OjNU`0\\1j_OcNR`0b1n_O^Nn?g1Q@YNj?m1U@RNg?S2Y@mMX?c2f@]M[?d2d@\\M\\?d2d@[M]?f2b@ZMl>_OYAW3KYMi>E[AS3OTMc>L^AQ32nL^>5^Ao27hLh=l0QB]29cL`=X1VBV2Z>PNfAR2S>TNlAm1R>TNnAm1P>UNnAm1P>TNlAQ2Q>Q2N3M2N2O1N2N2O1N2N01O010O00001O00bNQJlDP6S;[JcDf1YO`1T<TMRDc1NU1P<3TDIl;6YDFg;:]DBc;>aDYM:2U;d2cDYM;OS;g2dDYM;MR;i2dDYM>Jo:l2dDZM?Gn:m2fDZM`0Ek:P3fD[Mb0Bh:R3hD[Mb0@g:U3gDZMf0]Oe:X3eD[Mh0ZOd:Z3fD[Mi0WOb:^3eDZM\\=f2eBYM[=g2eBYM[=g2eBXM\\=h2eBWM[=h2fBWM\\=h2dBXM]=g2dBXM\\=h2dBWM^=h2cBWM^=h2bBXM_=f2bBYM`=f2aBYMc=c2]B]Me>a1[A^Nh>`1WAbNi>]1WAeNh>Y1YAiNg>U1XAnNg>Q1YAQOe>o0[ASOc>m0]AUOb>j0]AYOa>g0_A[O_>e0aA^O]>b0bA_O]>a0bA@^>`0bA_O`>`0`A_Oa>b0]A^Od>b0\\A]Of>b0ZA]Og>d0XA[Oi>e0VA[Oe>k0[AUOg>j0XAUOk>i0TAWOo>g0QAXOR?f0n@YOU?f0j@YOY?e0k@VOW?j0Q2N4M2N3M2N2N3M9F;F:FgYS1"}, "label": "this is a man sitting and holding a racquet"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest man standing up\"."}], "task": "refering", "answer_template": "The \"the tallest man standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 91, 92, 107, 108, 109, 110, 124, 125, 126, 127, 142, 143, 144, 159, 160, 161, 176, 177, 178, 194, 195, 212], "bbox": [0.299625, 0.192578125, 0.5093541666666667, 0.543140625], "iscrowd": 0, "area": 10933.7138, "rle": {"size": [640, 480], "counts": "bVj22lc02M3N2M4M2M4M2O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N4L<D<D;E<D2O0ROZMb@f2U?dMj@\\2Q?jMm@W2S?jMl@V2S?lMl@T2T?mMk@S2U?nMj@S2T?oMk@R2T?oMk@S2R?oMm@S2Q?oMm@R2Q?PNm@S2Q?nMn@S2Q?`1O1N1O2N1O2N2O0O2N101N2N101N1O2O1O010O00gKeAm2[>PMgAQ3X>mLiAT3\\>eLeA]3\\>_LfAa3[>\\LfAf3[>VLfAk3\\>RLeAo3e>dK\\A^4R?101N1O2O1N1O2O0O2O000O101OjNU1N3M=Cf0ZO8H9H6I4L3J7H7I8\\Oc0I_Xc4"}, "label": "the tallest man standing up"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man standing up with his hands in his pocket and no tennis racket\"."}], "task": "refering", "answer_template": "The \"a young man standing up with his hands in his pocket and no tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 91, 92, 107, 108, 109, 110, 124, 125, 126, 127, 142, 143, 144, 159, 160, 161, 176, 177, 178, 194, 195, 212], "bbox": [0.299625, 0.192578125, 0.5093541666666667, 0.543140625], "iscrowd": 0, "area": 10933.7138, "rle": {"size": [640, 480], "counts": "bVj22lc02M3N2M4M2M4M2O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N4L<D<D;E<D2O0ROZMb@f2U?dMj@\\2Q?jMm@W2S?jMl@V2S?lMl@T2T?mMk@S2U?nMj@S2T?oMk@R2T?oMk@S2R?oMm@S2Q?oMm@R2Q?PNm@S2Q?nMn@S2Q?`1O1N1O2N1O2N2O0O2N101N2N101N1O2O1O010O00gKeAm2[>PMgAQ3X>mLiAT3\\>eLeA]3\\>_LfAa3[>\\LfAf3[>VLfAk3\\>RLeAo3e>dK\\A^4R?101N1O2O1N1O2O0O2O000O101OjNU1N3M=Cf0ZO8H9H6I4L3J7H7I8\\Oc0I_Xc4"}, "label": "a young man standing up with his hands in his pocket and no tennis racket"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with mustache sitting down\"."}], "task": "refering", "answer_template": "The \"man with mustache sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 140, 141, 157, 158, 159, 173, 174, 175, 176, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 230, 243, 245, 246, 247, 262, 263, 264, 265, 280, 281, 282, 283, 297, 298, 314, 315], "bbox": [0.22222916666666667, 0.3359375, 0.6756666666666666, 0.791671875], "iscrowd": 0, "area": 20389.98405, "rle": {"size": [640, 480], "counts": "PUS2i1Vb06J6K5J6J6J6J3M1O010O1O1O001O1O10O01O1O1O1O0kNlL_AU3_>QM]Ao2a>WMZAj2d>\\MWAf2f>`MUAc2g>cMTA_2e>aLTAW11[2j>a1N1O2N2O0O2N2N101N2N1O2O0O1O00O2O000O101O0O100O2O000O2O0O100kKgAc2Z>YLiBf3i>0000O100O10000O102N1N3N1N3N1O2M3N1O2M2O2M2O2N2M2O2M2O2N1N3N2O0010O100O10O10O100O010O10000O010O100O10O10O100O1N101N2N7I6K5J7I6J6K6I6J6J6L5J10N1O2N1O2N1XNb@lN`?Q1d@lN]?S1f@jN[?U1h@iNY?U1j@hNW?V1m@gNU?W1n@fNS?Y1o@eNS?X1QAeNP?Z1QAeNQ?i0T@gNl0?Q?i0U@fNj0a0S?g0U@eNj0c0R?g0V@dNi0d0S?e0X@dNf0f0S?e0]@bN>h0W?d0Z@fN?f0W?c0X@jNa0b0Y?b0T@oNc0>Z?e0n_OPOh0:\\?g0h_OQOl07]?Z1b@fN`?Z1_@eNb?[1^@dNd?[1\\@dNe?[1\\@dNe?[1\\@dNf?Z1[@eNf?Z1[@eNg?Y1Z@gNf?X1[@gNg?W1Z@gNj?V1W@iNl?T1U@kNn?R1S@mNP`0P1Q@oNP`0P1Q@oNP`0P1Q@oNP`0P1Q@oNo?Q1R@mNP`0R1Q@mNP`0S1_1O001O1O001O1O001O1O1O1O001O1O1O1O00100O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10jQQ3"}, "label": "man with mustache sitting down"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting in the chair on the left\"."}], "task": "refering", "answer_template": "The \"the man sitting in the chair on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 140, 141, 157, 158, 159, 173, 174, 175, 176, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 230, 243, 245, 246, 247, 262, 263, 264, 265, 280, 281, 282, 283, 297, 298, 314, 315], "bbox": [0.22222916666666667, 0.3359375, 0.6756666666666666, 0.791671875], "iscrowd": 0, "area": 20389.98405, "rle": {"size": [640, 480], "counts": "PUS2i1Vb06J6K5J6J6J6J3M1O010O1O1O001O1O10O01O1O1O1O0kNlL_AU3_>QM]Ao2a>WMZAj2d>\\MWAf2f>`MUAc2g>cMTA_2e>aLTAW11[2j>a1N1O2N2O0O2N2N101N2N1O2O0O1O00O2O000O101O0O100O2O000O2O0O100kKgAc2Z>YLiBf3i>0000O100O10000O102N1N3N1N3N1O2M3N1O2M2O2M2O2N2M2O2M2O2N1N3N2O0010O100O10O10O100O010O10000O010O100O10O10O100O1N101N2N7I6K5J7I6J6K6I6J6J6L5J10N1O2N1O2N1XNb@lN`?Q1d@lN]?S1f@jN[?U1h@iNY?U1j@hNW?V1m@gNU?W1n@fNS?Y1o@eNS?X1QAeNP?Z1QAeNQ?i0T@gNl0?Q?i0U@fNj0a0S?g0U@eNj0c0R?g0V@dNi0d0S?e0X@dNf0f0S?e0]@bN>h0W?d0Z@fN?f0W?c0X@jNa0b0Y?b0T@oNc0>Z?e0n_OPOh0:\\?g0h_OQOl07]?Z1b@fN`?Z1_@eNb?[1^@dNd?[1\\@dNe?[1\\@dNe?[1\\@dNf?Z1[@eNf?Z1[@eNg?Y1Z@gNf?X1[@gNg?W1Z@gNj?V1W@iNl?T1U@kNn?R1S@mNP`0P1Q@oNP`0P1Q@oNP`0P1Q@oNP`0P1Q@oNo?Q1R@mNP`0R1Q@mNP`0S1_1O001O1O001O1O001O1O1O1O001O1O1O1O00100O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10jQQ3"}, "label": "the man sitting in the chair on the left"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man third from left\"."}], "task": "refering", "answer_template": "The \"man third from left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 112, 113, 128, 129, 130, 131, 145, 146, 147, 148, 162, 163, 164, 165, 179, 180, 181, 196, 197, 198, 213, 214, 215, 230, 231, 232, 247, 248, 249, 265], "bbox": [0.5411458333333333, 0.2248125, 0.7390416666666667, 0.66815625], "iscrowd": 0, "area": 15371.941200000001, "rle": {"size": [640, 480], "counts": "bgR5m0mb0<E:Eo0ROa0@?A`0@`0SAbK^=l4SBZKh=d5J3M4L4L4N1N3N2M2O2M3N2M2O2M100O1O001O001O10QOfCWJZ<d5XDPJ`;S6T1O3M2N3M40lC^Jl9c5QFgJf9[5VFmJe9T5WFQKg9P5RFXKl9i4nE^KQ:b4iEfKV:c4VEfKk:e63N2M2O2J5L5J6L4K5K6D;kNV1L3L4M4L3M3L5K4B>M4WLi@e2Y?UMn@h2T?QMUAk2n>nLZAm2i>mL_An2m?J5K6L3N2O2M2N2O1N1O2O1N1O2O2M3M3M2N4L4L;E6J6Iel]2"}, "label": "man third from left"}]}
{"id": 269605, "image": "COCO_train2014_000000269605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ban wearing a piece of black fabric around his waist\"."}], "task": "refering", "answer_template": "The \"a ban wearing a piece of black fabric around his waist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 112, 113, 128, 129, 130, 131, 145, 146, 147, 148, 162, 163, 164, 165, 179, 180, 181, 196, 197, 198, 213, 214, 215, 230, 231, 232, 247, 248, 249, 265], "bbox": [0.5411458333333333, 0.2248125, 0.7390416666666667, 0.66815625], "iscrowd": 0, "area": 15371.941200000001, "rle": {"size": [640, 480], "counts": "bgR5m0mb0<E:Eo0ROa0@?A`0@`0SAbK^=l4SBZKh=d5J3M4L4L4N1N3N2M2O2M3N2M2O2M100O1O001O001O10QOfCWJZ<d5XDPJ`;S6T1O3M2N3M40lC^Jl9c5QFgJf9[5VFmJe9T5WFQKg9P5RFXKl9i4nE^KQ:b4iEfKV:c4VEfKk:e63N2M2O2J5L5J6L4K5K6D;kNV1L3L4M4L3M3L5K4B>M4WLi@e2Y?UMn@h2T?QMUAk2n>nLZAm2i>mL_An2m?J5K6L3N2O2M2N2O1N1O2O1N1O2O2M3M3M2N4L4L;E6J6Iel]2"}, "label": "a ban wearing a piece of black fabric around his waist"}]}
{"id": 228647, "image": "COCO_train2014_000000228647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman riding a horse over a hurdle\"."}], "task": "refering", "answer_template": "The \"a woman riding a horse over a hurdle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 34, 35, 36, 57, 58, 59, 79, 80, 81, 102, 103, 104, 125, 147, 148, 170, 171, 192, 193], "bbox": [0.379140625, 0.019933774834437083, 0.609703125, 0.5608609271523178], "iscrowd": 0, "area": 9231.019299999996, "rle": {"size": [453, 640], "counts": "Xg[32Q>4M2N2O1O1N2O10000O1000O10O1N1M4L4L3M4L4LQOSCl0j<VOYCi0c<YO_Cf0^<\\OcCd0Z<]OfCe0W<]OiCd0T<^OlCc0P<@PD`0n;ASD`0i;CVD?f;DZD<c;G]D6c;L^D0c;3]DKb;8]DGc;;]DCc;?]D_Od;a0]D]Oc;e0\\DZOe;g0[DWOe;k0[DTOT:_OUG]1hNROT:FnFZ1nNkNY:OdFW1ROgN]:7\\FS1WOaNc:?PFQ1[:QObEP1_:QO]EQ1d:POYEQ1g:POVEQ1l:POQEo0R;ROjDn0Y;SOdDl0^;k03M3M2M4M3M2O2NL400GPN_DQ2`;:00O2N1001O01O1O1O1O1O100O100O2N01M2O2O0100O10O01_OcMZE]2_:kM_ES2L]Mc:c0`Eo1b:TN\\Ek1d:XN[Eg1e:k00100010O01O010O0100O1O10O01O1O100O1O2N1O1lN\\EfNe:W1bEdN_:Z1gEaNZ:^1kE]NW:a1W1N3O100O1O100O101N6J5lNYCa0Y=L100O0100O010O010O010O02N6J6Jl`^3"}, "label": "a woman riding a horse over a hurdle"}]}
{"id": 228647, "image": "COCO_train2014_000000228647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white horse\"."}], "task": "refering", "answer_template": "The \"the white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 104, 105, 106, 107, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 263, 284, 285, 286, 306, 307, 308, 329, 330, 331], "bbox": [0.3213125, 0.14805739514348784, 0.6775, 0.9210816777041942], "iscrowd": 0, "area": 27718.268949999987, "rle": {"size": [453, 640], "counts": "o`k23P>3M3M3M3M2N2N2N2N3N1O1000000000000ZOYOnCg0i;AWD`0X;1hDL[;4eDEb;;^D@g;`0ZD[Oj;e07TOoCl0n;WORDf0m;^OSD?V:6QF^OI9T:`0mE[OO2R:V1^EkN`0LS:P2mEmMS:V2mEgMT:[2lEbMa1K`6e2oG]Md1N[6g2ZJYMd53bGa2i2]Mb54gG^2g2^M`54kG]2f2_M]55oGY2f2bMY56SHV2d2eMX55UHU2a2iMX53YHR2]2nMY51ZHo1\\2SNX51\\Hj1\\2VNW51]Hh1[2YNW51^Hd1Z2]NV51aH_1Y2bNU50bH]1X2eNU50cHY1X2hNS50fHW1V2kNS5MjHU1S2POQ5JPIT1o1SOP5HSIT1m1UOP5EVIT1T1^NfNi0o6DYIS1R1dNdNd0Q7C]IT1k0hNfN?U7C\\IU1h0kNgN9X7F\\IU1c0nNhN4\\7H\\IT1?ROTOCT75\\IV1:TOX7DaHV16YOW7@eHW12[OX7]OiHV1N_OY7XOmHX1DFW99jFGS9<mFDP9>QGBl8b0SG^Oj8g0TGYOi8l0aFXNMj0a9S1^FUN1e0a9Z1[FSN4`0`9c1WFPN9;_9j1TFmM>6]9a2bF^M\\9e2dFYM[9j2eFUMY9n2gFPMX9S3hFlLV9W3jFgLV9[3jFcLV9_3jF`LU9b3kF\\LT9g3lFXLS9j3b02N2N2N2N1O1O1N2O2N1O1O1O1000O010O100O100O10O0100O1O1M3M3M2N3O11O001O1O1O1N2O1O1O1O1O100O100O1^FnKQ9R4kFTLR9m3mFULR9_4O2O1N2O1NO2O001O001O001O000O01jNoJXIR5g6QKVIP5h6TKUIm4j6VKSIk4k6YKRIh4m6[KPIf4n6^KoHc4P7`KmHb4R7Y11O001N10001^OcHZJ]7b5lHYJU7c5h0L5L3MjHmJ`5P5`JSKa5m4[JVKe5k4XJWKh5j4TJYKl5h4QJZKo5g4nI[KR6e4lI]KT6c4jI_KV6`4iIbKW6^4fIeKZ6[4dIgK\\6X4cIjK]6V4aIlK_6T4_InKa6R4\\IQLe6P4WIRLi6o3TISLl6o3PISLP7\\55I7M3O1O1O1N2M3K5L4L4L4`NhGoL]8a2THXMQ8e2SHVMQ8h2SHRMQ8k2SHQMP8m2THnLQ8n2RHnLR8n2RHnLS8b2]HYMg7b2U2F4M3L5L3L4K5L4K5F;O0J6I7K5L4L5KV`k2"}, "label": "the white horse"}]}
{"id": 228647, "image": "COCO_train2014_000000228647.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray horse in mid jump over two white and blue poles\"."}], "task": "refering", "answer_template": "The \"a gray horse in mid jump over two white and blue poles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 82, 83, 84, 104, 105, 106, 107, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 263, 284, 285, 286, 306, 307, 308, 329, 330, 331], "bbox": [0.3213125, 0.14805739514348784, 0.6775, 0.9210816777041942], "iscrowd": 0, "area": 27718.268949999987, "rle": {"size": [453, 640], "counts": "o`k23P>3M3M3M3M2N2N2N2N3N1O1000000000000ZOYOnCg0i;AWD`0X;1hDL[;4eDEb;;^D@g;`0ZD[Oj;e07TOoCl0n;WORDf0m;^OSD?V:6QF^OI9T:`0mE[OO2R:V1^EkN`0LS:P2mEmMS:V2mEgMT:[2lEbMa1K`6e2oG]Md1N[6g2ZJYMd53bGa2i2]Mb54gG^2g2^M`54kG]2f2_M]55oGY2f2bMY56SHV2d2eMX55UHU2a2iMX53YHR2]2nMY51ZHo1\\2SNX51\\Hj1\\2VNW51]Hh1[2YNW51^Hd1Z2]NV51aH_1Y2bNU50bH]1X2eNU50cHY1X2hNS50fHW1V2kNS5MjHU1S2POQ5JPIT1o1SOP5HSIT1m1UOP5EVIT1T1^NfNi0o6DYIS1R1dNdNd0Q7C]IT1k0hNfN?U7C\\IU1h0kNgN9X7F\\IU1c0nNhN4\\7H\\IT1?ROTOCT75\\IV1:TOX7DaHV16YOW7@eHW12[OX7]OiHV1N_OY7XOmHX1DFW99jFGS9<mFDP9>QGBl8b0SG^Oj8g0TGYOi8l0aFXNMj0a9S1^FUN1e0a9Z1[FSN4`0`9c1WFPN9;_9j1TFmM>6]9a2bF^M\\9e2dFYM[9j2eFUMY9n2gFPMX9S3hFlLV9W3jFgLV9[3jFcLV9_3jF`LU9b3kF\\LT9g3lFXLS9j3b02N2N2N2N1O1O1N2O2N1O1O1O1000O010O100O100O10O0100O1O1M3M3M2N3O11O001O1O1O1N2O1O1O1O1O100O100O1^FnKQ9R4kFTLR9m3mFULR9_4O2O1N2O1NO2O001O001O001O000O01jNoJXIR5g6QKVIP5h6TKUIm4j6VKSIk4k6YKRIh4m6[KPIf4n6^KoHc4P7`KmHb4R7Y11O001N10001^OcHZJ]7b5lHYJU7c5h0L5L3MjHmJ`5P5`JSKa5m4[JVKe5k4XJWKh5j4TJYKl5h4QJZKo5g4nI[KR6e4lI]KT6c4jI_KV6`4iIbKW6^4fIeKZ6[4dIgK\\6X4cIjK]6V4aIlK_6T4_InKa6R4\\IQLe6P4WIRLi6o3TISLl6o3PISLP7\\55I7M3O1O1O1N2M3K5L4L4L4`NhGoL]8a2THXMQ8e2SHVMQ8h2SHRMQ8k2SHQMP8m2THnLQ8n2RHnLR8n2RHnLS8b2]HYMg7b2U2F4M3L5L3L4K5L4K5F;O0J6I7K5L4L5KV`k2"}, "label": "a gray horse in mid jump over two white and blue poles"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis player wearing a white cap backwards\"."}], "task": "refering", "answer_template": "The \"tennis player wearing a white cap backwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 122, 123, 124, 125, 144, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 213, 214, 215, 216, 237, 238, 239, 240, 260, 261, 262, 263, 264, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332, 352, 353, 355, 376, 378], "bbox": [0.282515625, 0.1950625, 0.48431250000000003, 0.9865416666666668], "iscrowd": 0, "area": 25833.705700000002, "rle": {"size": [480, 640], "counts": "WRe2m0o=5J6J6K5J6J6K5J6J6K5J6N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1_EjLU9W3kF]Ma8d2]GRNm7P2RH]Nb5cNlJQ3@dN\\5_NPKo2CiNU5]NSKk2GPOo4XNWKi2IUOi4WNYKf2MZOb4TN^Kc2NA\\4QNaK`22FU4nMfK]23Nn3jMjKY28Nl3mMiKV29Mo3V2PLiMP4Y2oKfMQ4\\2mKeMb0SLb2Y6lLcMb0YL]2V6oLaMd0ZL]2V6nL`Mc0\\L_2U6lL_Me0]L_2U6kL]Mf0_L`2T:_MmEa2R:_MoEb2o9]MSFc2l9]MUFc2k9\\MVFe2h9[MYFe2f9[M[Fe2e9YM]Fh2`9YMaFg2\\9[MeFf2W9\\MjFd2S9^MnFb2o8_MSGb2i8`MXG`2i4hLPNk0VM]2f4mLPNX1kLl1Q5PMPNf1`LZ1^5RMoMV5n1mJnMW5o1lJmMY5o1iJnM[5o1hJmM\\5P2gJmM]5P2eJjMa5S2aJgMT6h1nIQNY7m0iHmN^7P1dHiNc7U1_HfNf7X1\\HoN^7n0dHYOU7f0lH@n6>TIIe65]I1^6McI5[6JfI5[6KeI4]6JdI5]6KcI4^6KcI4_6JbI5_6KaI4QO^Mf6]2ZJ4PO_Mh6\\2XJ5nNaMj6Y2YJ5mNjMc6P2aJ5kNUN]6e1hJ5jN_NW6[1PK5iNhNP6S1WK4hNROj5i0_K4gN\\Oc5>gK5eNF]55nK4eNOV5LVL4cN9Q5B\\L5bN`0l4ZOcL5`Nc0m4WOdL5_Ne0n4UOcL5^Nh0o4ROdL5]Nj0o4QOdL4\\Nm0Q5mNdL5[No0Q5kNeL5YNR1R5iNeL4XNU1T5eNeL5WNW1T5dNdL6VNX1V5aNeL6UNZ1V5_NfL6SN]1X5\\NeL5TN`1X5YNeL4YN`1V5XNaL5jNU1i4aN^L7[Of0^4nNXL9K1\\4AiK;<]OX43]K=m0gNT4h0oJ>`8]OgG:`8AgG6`8FgG1_8JiGL^80abj4"}, "label": "tennis player wearing a white cap backwards"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second man from the left side of the image\"."}], "task": "refering", "answer_template": "The \"second man from the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 122, 123, 124, 125, 144, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 213, 214, 215, 216, 237, 238, 239, 240, 260, 261, 262, 263, 264, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332, 352, 353, 355, 376, 378], "bbox": [0.282515625, 0.1950625, 0.48431250000000003, 0.9865416666666668], "iscrowd": 0, "area": 25833.705700000002, "rle": {"size": [480, 640], "counts": "WRe2m0o=5J6J6K5J6J6K5J6J6K5J6N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1_EjLU9W3kF]Ma8d2]GRNm7P2RH]Nb5cNlJQ3@dN\\5_NPKo2CiNU5]NSKk2GPOo4XNWKi2IUOi4WNYKf2MZOb4TN^Kc2NA\\4QNaK`22FU4nMfK]23Nn3jMjKY28Nl3mMiKV29Mo3V2PLiMP4Y2oKfMQ4\\2mKeMb0SLb2Y6lLcMb0YL]2V6oLaMd0ZL]2V6nL`Mc0\\L_2U6lL_Me0]L_2U6kL]Mf0_L`2T:_MmEa2R:_MoEb2o9]MSFc2l9]MUFc2k9\\MVFe2h9[MYFe2f9[M[Fe2e9YM]Fh2`9YMaFg2\\9[MeFf2W9\\MjFd2S9^MnFb2o8_MSGb2i8`MXG`2i4hLPNk0VM]2f4mLPNX1kLl1Q5PMPNf1`LZ1^5RMoMV5n1mJnMW5o1lJmMY5o1iJnM[5o1hJmM\\5P2gJmM]5P2eJjMa5S2aJgMT6h1nIQNY7m0iHmN^7P1dHiNc7U1_HfNf7X1\\HoN^7n0dHYOU7f0lH@n6>TIIe65]I1^6McI5[6JfI5[6KeI4]6JdI5]6KcI4^6KcI4_6JbI5_6KaI4QO^Mf6]2ZJ4PO_Mh6\\2XJ5nNaMj6Y2YJ5mNjMc6P2aJ5kNUN]6e1hJ5jN_NW6[1PK5iNhNP6S1WK4hNROj5i0_K4gN\\Oc5>gK5eNF]55nK4eNOV5LVL4cN9Q5B\\L5bN`0l4ZOcL5`Nc0m4WOdL5_Ne0n4UOcL5^Nh0o4ROdL5]Nj0o4QOdL4\\Nm0Q5mNdL5[No0Q5kNeL5YNR1R5iNeL4XNU1T5eNeL5WNW1T5dNdL6VNX1V5aNeL6UNZ1V5_NfL6SN]1X5\\NeL5TN`1X5YNeL4YN`1V5XNaL5jNU1i4aN^L7[Of0^4nNXL9K1\\4AiK;<]OX43]K=m0gNT4h0oJ>`8]OgG:`8AgG6`8FgG1_8JiGL^80abj4"}, "label": "second man from the left side of the image"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a short balding man on a tennis court holding a tennis racket to his chest\"."}], "task": "refering", "answer_template": "The \"a short balding man on a tennis court holding a tennis racket to his chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 340, 341, 362, 363, 364, 365, 385, 386, 387, 388], "bbox": [0.684265625, 0.2625416666666667, 0.884828125, 0.997375], "iscrowd": 0, "area": 29165.294249999995, "rle": {"size": [480, 640], "counts": "ha]6d0U>i0WO9F:G9F:G9G9M3N2M3N2M3M?\\NSLYG`4n7hKoGm41UKi54UJU6e5SJXJo5b5WJ\\Jk5c5WJ[Jj5c5XJ[Jj5c5YJZJh5f5YJYJh5e5[JXJg5f5[JYJf5e5]JXJe5g5\\JWJe5h5^JVJc5h5_JVJc5i5_JUJb5j5j110O1O1O2O;D=C3N2M2N2N2O1N2N3N1O1O1O100O00O1O2N1O1O1O2N1O2XF[KQ9f4kF_KW7JUJi4_NaKY7MSJc4`NdKZ71PJP5n5XKlIh4T5gJYJh0>a4Y5oJoIU15l3k5XKfIS1:f3o5ZMlIf2T6_MgIa2Z6R32N2O2M2N3M2N2O2M2N3M2N2O1N2N2N2N2N2O1N2N2N2N2O2M2N3M2N3N2M3M6LiJaIo1[6mMmIR2P6kMXJP2g5mMbJn1^5oMjJk1W5QNQKk1o4RNYKi1f4TNcKg1]4VNlKc1U4YNTL`1m3^N[L[1g3aNbLX1`3eNhLT1Z3hNPMP1Q3nNWMk0k2QO^Mh0e2TOcMe0a2UOhMd0[2iN]No0f1lN`NP1c1lNbNP1a1mNcNo0a1mNcNP1_1lNfNP1]1mNgNo0\\1nNhNn0i7LnWR1"}, "label": "a short balding man on a tennis court holding a tennis racket to his chest"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing to the far right of a group of men\"."}], "task": "refering", "answer_template": "The \"a man standing to the far right of a group of men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 340, 341, 362, 363, 364, 365, 385, 386, 387, 388], "bbox": [0.684265625, 0.2625416666666667, 0.884828125, 0.997375], "iscrowd": 0, "area": 29165.294249999995, "rle": {"size": [480, 640], "counts": "ha]6d0U>i0WO9F:G9F:G9G9M3N2M3N2M3M?\\NSLYG`4n7hKoGm41UKi54UJU6e5SJXJo5b5WJ\\Jk5c5WJ[Jj5c5XJ[Jj5c5YJZJh5f5YJYJh5e5[JXJg5f5[JYJf5e5]JXJe5g5\\JWJe5h5^JVJc5h5_JVJc5i5_JUJb5j5j110O1O1O2O;D=C3N2M2N2N2O1N2N3N1O1O1O100O00O1O2N1O1O1O2N1O2XF[KQ9f4kF_KW7JUJi4_NaKY7MSJc4`NdKZ71PJP5n5XKlIh4T5gJYJh0>a4Y5oJoIU15l3k5XKfIS1:f3o5ZMlIf2T6_MgIa2Z6R32N2O2M2N3M2N2O2M2N3M2N2O1N2N2N2N2N2O1N2N2N2N2O2M2N3M2N3N2M3M6LiJaIo1[6mMmIR2P6kMXJP2g5mMbJn1^5oMjJk1W5QNQKk1o4RNYKi1f4TNcKg1]4VNlKc1U4YNTL`1m3^N[L[1g3aNbLX1`3eNhLT1Z3hNPMP1Q3nNWMk0k2QO^Mh0e2TOcMe0a2UOhMd0[2iN]No0f1lN`NP1c1lNbNP1a1mNcNo0a1mNcNP1_1lNfNP1]1mNgNo0\\1nNhNn0i7LnWR1"}, "label": "a man standing to the far right of a group of men"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and shorts , holding a yellow hat and a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and shorts , holding a yellow hat and a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 278, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328, 348, 349, 350, 351, 371, 372, 373, 374], "bbox": [0.0830625, 0.2504791666666667, 0.302390625, 1.0], "iscrowd": 0, "area": 31544.176900000002, "rle": {"size": [480, 640], "counts": "TSi01l>3M3M4L3M4L3N2M4L3M4L3M3M6K5J6J6J6J6J6J5L4K4M4N2M2O2bEiLo8Y3gFSMT9o2dF\\MX9f2_FeM\\9^2[FlMb9Q6hM1N2O1O2M2O1O1N2O1O0O2O4L3L5`IXH`5c810O1kNnF\\LR9c3UGWLk8g3]GRLd8m3cGmK]8Q4fGnKY8Q4jGmKW8R4jGlKX8R4jGlKX8S4jGjKX8T4jGkKW8T4jGjKX8T4Z101O1O1O1O1O1O010O10kNT110O01000O0100O11N101O1N101O1_N`110001Om0ROa0@O1O001O1O2fGfJj6]5QIgJm6\\5nHfJR7]5iHeJW7^5dHeJ[7^5THPKl7R601O001O000015J6J6K5J6K5K5KO001O1O001O1O1N101O1O001oHoIQ5R6bJ[J]5f5VJgJi5_5fImJY6n6O1QJ^I^3c6_LRKd0Y6ZOh4M4L3M3N2M`Xa6"}, "label": "a man in a white shirt and shorts , holding a yellow hat and a tennis racket"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing white clothes holding a yellow cap and tennis racquet\"."}], "task": "refering", "answer_template": "The \"a man wearing white clothes holding a yellow cap and tennis racquet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 278, 279, 280, 281, 282, 302, 303, 304, 305, 325, 326, 327, 328, 348, 349, 350, 351, 371, 372, 373, 374], "bbox": [0.0830625, 0.2504791666666667, 0.302390625, 1.0], "iscrowd": 0, "area": 31544.176900000002, "rle": {"size": [480, 640], "counts": "TSi01l>3M3M4L3M4L3N2M4L3M4L3M3M6K5J6J6J6J6J6J5L4K4M4N2M2O2bEiLo8Y3gFSMT9o2dF\\MX9f2_FeM\\9^2[FlMb9Q6hM1N2O1O2M2O1O1N2O1O0O2O4L3L5`IXH`5c810O1kNnF\\LR9c3UGWLk8g3]GRLd8m3cGmK]8Q4fGnKY8Q4jGmKW8R4jGlKX8R4jGlKX8S4jGjKX8T4jGkKW8T4jGjKX8T4Z101O1O1O1O1O1O010O10kNT110O01000O0100O11N101O1N101O1_N`110001Om0ROa0@O1O001O1O2fGfJj6]5QIgJm6\\5nHfJR7]5iHeJW7^5dHeJ[7^5THPKl7R601O001O000015J6J6K5J6K5K5KO001O1O001O1O1N101O1O001oHoIQ5R6bJ[J]5f5VJgJi5_5fImJY6n6O1QJ^I^3c6_LRKd0Y6ZOh4M4L3M3N2M`Xa6"}, "label": "a man wearing white clothes holding a yellow cap and tennis racquet"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis racket held by the man who is second from the left\"."}], "task": "refering", "answer_template": "The \"a tennis racket held by the man who is second from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 193, 194, 195, 196, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 260], "bbox": [0.29725, 0.4565416666666666, 0.528296875, 0.6793958333333333], "iscrowd": 0, "area": 5271.878050000002, "rle": {"size": [480, 640], "counts": "o[i21n>2N2N3M2O1N0010O0010O01O01O01O010O00010O0010O01O01O010O01O010O002O1N2N2O1NXP70hoH2N2M2O3L3N2M3N5J3N2M2N3N1N3M101N2N2N2O1N1O2O1N2N2O1O1O1O1O1O100O1O1O1O1O2N100O100O100O100O01000O10O0100O10O010000000O10000O100O100O2O0O100O100O2O0O100O1O1O2N2O1N2N1O2N2O1N2N2N3L4L4K5L7I8E\\Z]4"}, "label": "a tennis racket held by the man who is second from the left"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis racquet being held by a man wearing a white shirt , white shorts , and a white backwards baseball cap\"."}], "task": "refering", "answer_template": "The \"a tennis racquet being held by a man wearing a white shirt , white shorts , and a white backwards baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 193, 194, 195, 196, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 260], "bbox": [0.29725, 0.4565416666666666, 0.528296875, 0.6793958333333333], "iscrowd": 0, "area": 5271.878050000002, "rle": {"size": [480, 640], "counts": "o[i21n>2N2N3M2O1N0010O0010O01O01O01O010O00010O0010O01O01O010O01O010O002O1N2N2O1NXP70hoH2N2M2O3L3N2M3N5J3N2M2N3N1N3M101N2N2N2O1N1O2O1N2N2O1O1O1O1O1O100O1O1O1O1O2N100O100O100O100O01000O10O0100O10O010000000O10000O100O100O2O0O100O100O2O0O100O1O1O2N2O1N2N1O2N2O1N2N2N3L4L4K5L7I8E\\Z]4"}, "label": "a tennis racquet being held by a man wearing a white shirt , white shorts , and a white backwards baseball cap"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man wearing a horizontally striped shirt , holding a red tennis racket\"."}], "task": "refering", "answer_template": "The \"an older man wearing a horizontally striped shirt , holding a red tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 334, 336, 337, 356, 357, 359, 360, 380, 382, 383], "bbox": [0.48, 0.162375, 0.695140625, 1.0], "iscrowd": 0, "area": 32903.77205, "rle": {"size": [480, 640], "counts": "iS`44f>>C<C>C<C8I7H7J7H7L5L4M2M4M2N3M3M2O2Mk0VOd1\\Nl0SO2O1N2O2WI]IQ5d6eJUJj4m5mJnJ`4S5VKYKf4d7M0O10000O100O100lNfEaM[:\\2kEaMU:]2QF^MP:`2UF\\Mj9c2\\FZM`9h2eFTMY9k2mFRMP9n2WGoLe8R3`GkL]8U3iGhLU8Z3e13M4L4M3L3nNS1M3M3N1N3M3N2H7F;D<D<I66K5K4M4L4L4Kc0^O000000O01001O2N1O1O2M3dG`JS7b5hHbJX7_5bHdJb7[5VHkJm7R64L4M3L4L4L4M3L4L4L4M3L4L5L3L4M4K5L4K01O001O001kLYKoKg4Q4aKgK`4X4hK`KX4_2eK`L<i0P4Z1XKcMX1@H[1h3a1]K_MU3i0_1f1bK[MX3g0V1l1hKWM[3f0n0Q2mKSM^3d0e0X2RLoLd3?;_2XLkLm38Ni2l6M4L3L4M4L3M4K4M4L3L5I6I8G8I7He0\\ObVk2"}, "label": "an older man wearing a horizontally striped shirt , holding a red tennis racket"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a striped t - shirt holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man wearing a striped t - shirt holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 334, 336, 337, 356, 357, 359, 360, 380, 382, 383], "bbox": [0.48, 0.162375, 0.695140625, 1.0], "iscrowd": 0, "area": 32903.77205, "rle": {"size": [480, 640], "counts": "iS`44f>>C<C>C<C8I7H7J7H7L5L4M2M4M2N3M3M2O2Mk0VOd1\\Nl0SO2O1N2O2WI]IQ5d6eJUJj4m5mJnJ`4S5VKYKf4d7M0O10000O100O100lNfEaM[:\\2kEaMU:]2QF^MP:`2UF\\Mj9c2\\FZM`9h2eFTMY9k2mFRMP9n2WGoLe8R3`GkL]8U3iGhLU8Z3e13M4L4M3L3nNS1M3M3N1N3M3N2H7F;D<D<I66K5K4M4L4L4Kc0^O000000O01001O2N1O1O2M3dG`JS7b5hHbJX7_5bHdJb7[5VHkJm7R64L4M3L4L4L4M3L4L4L4M3L4L5L3L4M4K5L4K01O001O001kLYKoKg4Q4aKgK`4X4hK`KX4_2eK`L<i0P4Z1XKcMX1@H[1h3a1]K_MU3i0_1f1bK[MX3g0V1l1hKWM[3f0n0Q2mKSM^3d0e0X2RLoLd3?;_2XLkLm38Ni2l6M4L3L4M4L3M4K4M4L3L5I6I8G8I7He0\\ObVk2"}, "label": "a man wearing a striped t - shirt holding a tennis racket"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis racket in the hands of the man in the striped shirt\"."}], "task": "refering", "answer_template": "The \"tennis racket in the hands of the man in the striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 197, 198, 199, 220, 221, 234, 257, 258, 266, 280, 281, 303, 304, 326, 327, 328, 349, 350, 351, 372, 373], "bbox": [0.177265625, 0.37402083333333336, 0.695453125, 0.9532499999999999], "iscrowd": 0, "area": 8961.524150000001, "rle": {"size": [480, 640], "counts": "j[e12i>c0]Ob0@9I2M3N1O2M3N2N1N3L4L3L5K5M2N2N3M2PNWMgGk2Y8bMXG`2h8bMTGa2k8_MmFi2S9XMlFh2T9XMlFi2S9WMmFi2W9TMgFm2]9oLcFQ3a9lL_FT3d9hL^FU3j:M3M2N3L4M2N3M3M2M4M3M2N2M2O001O1N2N2N2N1O2N2N2N2N2N1O:_OnZk2TOUfTM2J6M21M3N2M2OfU3U1TiL8Gd0]O4L4K5L2N1N2O100O010O100O10O10001O1O001O010O001O0010O01O001O10O01O1O2N1O2O0O2N1O2N101N1O2N1O2O0O2N1N3L3N3L3N3L3N3L3M4Ma0^OcVk2"}, "label": "tennis racket in the hands of the man in the striped shirt"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red tennis racket the tallest man is holding\"."}], "task": "refering", "answer_template": "The \"red tennis racket the tallest man is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 174, 175, 176, 197, 198, 199, 220, 221, 234, 257, 258, 266, 280, 281, 303, 304, 326, 327, 328, 349, 350, 351, 372, 373], "bbox": [0.177265625, 0.37402083333333336, 0.695453125, 0.9532499999999999], "iscrowd": 0, "area": 8961.524150000001, "rle": {"size": [480, 640], "counts": "j[e12i>c0]Ob0@9I2M3N1O2M3N2N1N3L4L3L5K5M2N2N3M2PNWMgGk2Y8bMXG`2h8bMTGa2k8_MmFi2S9XMlFh2T9XMlFi2S9WMmFi2W9TMgFm2]9oLcFQ3a9lL_FT3d9hL^FU3j:M3M2N3L4M2N3M3M2M4M3M2N2M2O001O1N2N2N2N1O2N2N2N2N2N1O:_OnZk2TOUfTM2J6M21M3N2M2OfU3U1TiL8Gd0]O4L4K5L2N1N2O100O010O100O10O10001O1O001O010O001O0010O01O001O10O01O1O2N1O2O0O2N1O2N101N1O2N1O2O0O2N1N3L3N3L3N3L3N3L3M4Ma0^OcVk2"}, "label": "red tennis racket the tallest man is holding"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white tennis racket being held by the rightmost man\"."}], "task": "refering", "answer_template": "The \"a blue and white tennis racket being held by the rightmost man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 199, 200, 201, 202, 222, 223, 224, 225, 246, 247, 248, 271, 294, 295, 318], "bbox": [0.683796875, 0.46360416666666665, 0.85603125, 0.7790833333333335], "iscrowd": 0, "area": 5471.759400000001, "rle": {"size": [480, 640], "counts": "ga]6?]>9J3L4L4M3L4M3M2N3M2N2N3N1N2O1N2O1N2O1N2O1O001N2O1O001O1N2O001O1O1O001O1O000001O01O0001O001O001O001O001O001N2O1N2O1N2O1N2O1N3N1N2N3N1N3N1N3M3N2MYChNj;U1VDnNj;o0UDTOo;g0oC\\OU<KZD<T=2M22O3M00O100O010O0100O010O1O010000O010O2O0O01O0O2M2O2N1N3N0O2N4LahZ1"}, "label": "a blue and white tennis racket being held by the rightmost man"}]}
{"id": 548136, "image": "COCO_train2014_000000548136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tennis racket that the man in the white shirt and black shorts is holding\"."}], "task": "refering", "answer_template": "The \"the tennis racket that the man in the white shirt and black shorts is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 199, 200, 201, 202, 222, 223, 224, 225, 246, 247, 248, 271, 294, 295, 318], "bbox": [0.683796875, 0.46360416666666665, 0.85603125, 0.7790833333333335], "iscrowd": 0, "area": 5471.759400000001, "rle": {"size": [480, 640], "counts": "ga]6?]>9J3L4L4M3L4M3M2N3M2N2N3N1N2O1N2O1N2O1N2O1O001N2O1O001O1N2O001O1O1O001O1O000001O01O0001O001O001O001O001O001N2O1N2O1N2O1N2O1N3N1N2N3N1N3N1N3M3N2MYChNj;U1VDnNj;o0UDTOo;g0oC\\OU<KZD<T=2M22O3M00O100O010O0100O010O1O010000O010O2O0O01O0O2M2O2N1N3N0O2N4LahZ1"}, "label": "the tennis racket that the man in the white shirt and black shorts is holding"}]}
{"id": 253229, "image": "COCO_train2014_000000253229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the old lady selling kites\"."}], "task": "refering", "answer_template": "The \"the old lady selling kites\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 176, 177, 199, 200, 201, 222, 223, 224, 244, 245, 292, 313, 314, 315, 336, 337, 338], "bbox": [0.618546875, 0.40449999999999997, 0.77190625, 0.8853958333333334], "iscrowd": 0, "area": 7117.422000000002, "rle": {"size": [480, 640], "counts": "iPj56f>7J7H7J2M3N3M2M3N2M41N1O100O1O1O01`MjNVGU1e8TOXGj0d8@WG`0V88fGFn7k0oGTOm7Q1SHnNj7T1WHkNf7Y1[HfNb7]1_HaN^7c1bH]N]7d1dHWNa7g1aHWNa7h1`HVNb7i1^HVNc7j1^HTNd7k1]HSNe7k1\\HTNe7k1]HRNUN4`8j1]IPNnM=d8b1`IoMfMg0h8[1bInMcMk0j8V1eISOY6m0iISOV6l0kIUOS6l0nIUOo5l0RJTOl5m0UJSOj5m0VJTOh5m0YJSOe5m0]JSOa5n0_JSOa5l0`JTO_5l0bJTO]5l0dJTO[5l0eJUOZ5k0gJUOY5j0hJVOX5e0lJ\\OT5`0PKAo4>RKBn4?QKAo4`0oJoNnL^OS8d1nJkNVM]Ok7j1nJfN[M_Og7k1mJcNaMAb7m1lJ\\NjMEZ7P2kJUNRNJS7R2jJnM[NNk6T2iJiMk6X2n2100O001O\\N_Cj0c<QOeCj0b<lNaCS1W=000000O10000O1000002M2O1O1N2N2N2N2N6J5K5K9G2OTUT2"}, "label": "the old lady selling kites"}]}
{"id": 253229, "image": "COCO_train2014_000000253229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a red jacket holding a striped kite\"."}], "task": "refering", "answer_template": "The \"a lady in a red jacket holding a striped kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 176, 177, 199, 200, 201, 222, 223, 224, 244, 245, 292, 313, 314, 315, 336, 337, 338], "bbox": [0.618546875, 0.40449999999999997, 0.77190625, 0.8853958333333334], "iscrowd": 0, "area": 7117.422000000002, "rle": {"size": [480, 640], "counts": "iPj56f>7J7H7J2M3N3M2M3N2M41N1O100O1O1O01`MjNVGU1e8TOXGj0d8@WG`0V88fGFn7k0oGTOm7Q1SHnNj7T1WHkNf7Y1[HfNb7]1_HaN^7c1bH]N]7d1dHWNa7g1aHWNa7h1`HVNb7i1^HVNc7j1^HTNd7k1]HSNe7k1\\HTNe7k1]HRNUN4`8j1]IPNnM=d8b1`IoMfMg0h8[1bInMcMk0j8V1eISOY6m0iISOV6l0kIUOS6l0nIUOo5l0RJTOl5m0UJSOj5m0VJTOh5m0YJSOe5m0]JSOa5n0_JSOa5l0`JTO_5l0bJTO]5l0dJTO[5l0eJUOZ5k0gJUOY5j0hJVOX5e0lJ\\OT5`0PKAo4>RKBn4?QKAo4`0oJoNnL^OS8d1nJkNVM]Ok7j1nJfN[M_Og7k1mJcNaMAb7m1lJ\\NjMEZ7P2kJUNRNJS7R2jJnM[NNk6T2iJiMk6X2n2100O001O\\N_Cj0c<QOeCj0b<lNaCS1W=000000O10000O1000002M2O1O1N2N2N2N2N6J5K5K9G2OTUT2"}, "label": "a lady in a red jacket holding a striped kite"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate dish with a candle on it sits on a plate that says happy 90th birthday in front of a smiling man\"."}], "task": "refering", "answer_template": "The \"a chocolate dish with a candle on it sits on a plate that says happy 90th birthday in front of a smiling man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 164, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.0, 0.5170422535211268, 0.7895, 1.0], "iscrowd": 0, "area": 72022.8937, "rle": {"size": [426, 640], "counts": "W7a1i;Q3nLb1^N1O100O1O1O1O100O1O1O100O11O00001O001O001O00001O001O00001O001O00001O001O00001O001O00000000000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O1001O1O1O2N1O1O1O1O1O2N1O00001O00001O00001O00001O00001O003M3M4L3M0000001O0000001O0000001O0000001O0000001O0000001O0000001O000000I700O1O100O100O1O10000O10000O1000000O10000O1001O1O1O1O1O1O1O1O1O1O1O3M4L3M3M001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O0000001O00001O00001O0000001O00001O0000000000000000000000O100000000000000000000000000O100000000000000000000000000O1001O0000000000000000000000001O0000000000000000001O0000001O0000001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O001O1ZLcF]3]9aLeF_3\\9_LfF`3[9^LfFb3a90001O0000001O0000001O0000001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001N1N2M4L3L4L4K6K4L4L5K4L4K5L5K4L4L5KVjg1"}, "label": "a chocolate dish with a candle on it sits on a plate that says happy 90th birthday in front of a smiling man"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that is set for the birthday celebration\"."}], "task": "refering", "answer_template": "The \"the table that is set for the birthday celebration\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 164, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.0, 0.5170422535211268, 0.7895, 1.0], "iscrowd": 0, "area": 72022.8937, "rle": {"size": [426, 640], "counts": "W7a1i;Q3nLb1^N1O100O1O1O1O100O1O1O100O11O00001O001O001O00001O001O00001O001O00001O001O00001O001O00000000000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O1001O1O1O2N1O1O1O1O1O2N1O00001O00001O00001O00001O00001O003M3M4L3M0000001O0000001O0000001O0000001O0000001O0000001O0000001O000000I700O1O100O100O1O10000O10000O1000000O10000O1001O1O1O1O1O1O1O1O1O1O1O3M4L3M3M001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O0000001O00001O00001O0000001O00001O0000000000000000000000O100000000000000000000000000O100000000000000000000000000O1001O0000000000000000000000001O0000000000000000001O0000001O0000001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O001O1ZLcF]3]9aLeF_3\\9_LfF`3[9^LfFb3a90001O0000001O0000001O0000001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O0000001N1N2M4L3L4L4K6K4L4L5K4L4K5L5K4L4L5KVjg1"}, "label": "the table that is set for the birthday celebration"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty table behind older couple\"."}], "task": "refering", "answer_template": "The \"empty table behind older couple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 155, 156], "bbox": [0.6384375, 0.13870892018779343, 0.8020156249999999, 0.46326291079812204], "iscrowd": 0, "area": 5971.625949999998, "rle": {"size": [426, 640], "counts": "gVZ5g0c<5K4K2O1N21OoDmNP:n0kEUO\\:e0`E_Of:;WEGl:7PEMk:[1K6J5N3O001O010O00aNPEe0o:ZOSEe0m:ZOUEe0k:[OVEe0i:ZOYEe0h:ZOYEe0g:ZOZEf0f:YO[Eg0e:YO[Eg0d;0000001O000TDfNY;;hDn0OgNV;l1jDTNS;o1mDRNn:R2REnMl:T2TEmMk:S2UEmMk:T2TElMl:T2UElMk:S2UEmMk:S2UEnMj:R2VEnMj:R2VEnMj:R2VEoMj:P2VEPNm:m1SETNS;f1lDZN[;_1eDaNf;T1ZDmNe;S1[DmNf;R1ZDnNf;R1ZDoNe;Q1[DoNe;Q1[DPOd;Q1[DoNe;Q1[DoNe;Q1\\DoNc;Q1]DoNc;Q1]DPOb;P1^DQOa;o0_DRO`;n0`DRO`;n0`DSO_;n0`DSO_;m0aDTO^;l0bDUO^;j0bDWO^;h0bDXO_;g0aDZO`;d0`D]O`;b0`D_O`;`0`DA_;U110OZOaDB_;=aDD^;<bDE];;cDF];9cDG];9cDH];7cDJ\\;S1O010O100O00100O3N2M3M2O0O1O00B?_O`0@`0@[gd1"}, "label": "empty table behind older couple"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small table set with wine glasses and a menu\"."}], "task": "refering", "answer_template": "The \"a small table set with wine glasses and a menu\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 155, 156], "bbox": [0.6384375, 0.13870892018779343, 0.8020156249999999, 0.46326291079812204], "iscrowd": 0, "area": 5971.625949999998, "rle": {"size": [426, 640], "counts": "gVZ5g0c<5K4K2O1N21OoDmNP:n0kEUO\\:e0`E_Of:;WEGl:7PEMk:[1K6J5N3O001O010O00aNPEe0o:ZOSEe0m:ZOUEe0k:[OVEe0i:ZOYEe0h:ZOYEe0g:ZOZEf0f:YO[Eg0e:YO[Eg0d;0000001O000TDfNY;;hDn0OgNV;l1jDTNS;o1mDRNn:R2REnMl:T2TEmMk:S2UEmMk:T2TElMl:T2UElMk:S2UEmMk:S2UEnMj:R2VEnMj:R2VEnMj:R2VEoMj:P2VEPNm:m1SETNS;f1lDZN[;_1eDaNf;T1ZDmNe;S1[DmNf;R1ZDnNf;R1ZDoNe;Q1[DoNe;Q1[DPOd;Q1[DoNe;Q1[DoNe;Q1\\DoNc;Q1]DoNc;Q1]DPOb;P1^DQOa;o0_DRO`;n0`DRO`;n0`DSO_;n0`DSO_;m0aDTO^;l0bDUO^;j0bDWO^;h0bDXO_;g0aDZO`;d0`D]O`;b0`D_O`;`0`DA_;U110OZOaDB_;=aDD^;<bDE];;cDF];9cDG];9cDH];7cDJ\\;S1O010O100O00100O3N2M3M2O0O1O00B?_O`0@`0@[gd1"}, "label": "a small table set with wine glasses and a menu"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older man getting ready to blow out candle on cake\"."}], "task": "refering", "answer_template": "The \"older man getting ready to blow out candle on cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 293], "bbox": [0.43624999999999997, 0.15291079812206573, 0.8135937500000001, 0.8197652582159624], "iscrowd": 0, "area": 42515.155849999996, "rle": {"size": [426, 640], "counts": "kZd38l<6J6J7I6K5N2M3M3M3M4L3M3M3M3M3M4L3N2M3M3M3M3M2N3M2N3M2N2OIQFWMl9j2ZFQMd9P392N2O2M3M5bFcLj8a3oFcLQ9`3iFdLU9i3O101N101XNiKZJX4d5kK[JV4c5mKZJT4d5`LiIb3T6aLkI_3T6cL_IQOJ^4d6dL_IQOK\\4d6fL^IQOLZ4e6gL^IoNM[4c6hL_ImNM\\4c6iL^IlNN]4c6hL^IlNM]4d6iL^Ic3`6_L_Ia3a6aL]I`3a6e1O2O0O101N1O2O0O2N101N1O2O0O1SJlHc5V7ZJlHf5T7XJnHg5[7O2O0O1O2N100O1O2O0000001O0001O01O0000001O00001O0000001O0000001O00000100O1O00100O1O1O100O00100O1O2N2O1N2N2N2O1N2N2N2O1nImIi4T6nJUJQ5l5fJ\\JZ5o5YJTJg5g600010O01O0010O0001O010O001O01O01O0010O01O00010O00010O001O01O01O0010O0001O010O000010O01O01O01O00010O001O01O01O0010O0001O010O1O00100O0011ON101O1N2O1O2N1N3M3M2N3M3M2N3M3M2N3M3N1N3M5K6J6J5K6E:E<C=D;D<E;F:G9F:FWXa1"}, "label": "older man getting ready to blow out candle on cake"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older gentlemen at a table smiling at a dessert with a candle in it\"."}], "task": "refering", "answer_template": "The \"older gentlemen at a table smiling at a dessert with a candle in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 80, 81, 82, 83, 103, 104, 105, 106, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 293], "bbox": [0.43624999999999997, 0.15291079812206573, 0.8135937500000001, 0.8197652582159624], "iscrowd": 0, "area": 42515.155849999996, "rle": {"size": [426, 640], "counts": "kZd38l<6J6J7I6K5N2M3M3M3M4L3M3M3M3M3M4L3N2M3M3M3M3M2N3M2N3M2N2OIQFWMl9j2ZFQMd9P392N2O2M3M5bFcLj8a3oFcLQ9`3iFdLU9i3O101N101XNiKZJX4d5kK[JV4c5mKZJT4d5`LiIb3T6aLkI_3T6cL_IQOJ^4d6dL_IQOK\\4d6fL^IQOLZ4e6gL^IoNM[4c6hL_ImNM\\4c6iL^IlNN]4c6hL^IlNM]4d6iL^Ic3`6_L_Ia3a6aL]I`3a6e1O2O0O101N1O2O0O2N101N1O2O0O1SJlHc5V7ZJlHf5T7XJnHg5[7O2O0O1O2N100O1O2O0000001O0001O01O0000001O00001O0000001O0000001O00000100O1O00100O1O1O100O00100O1O2N2O1N2N2N2O1N2N2N2O1nImIi4T6nJUJQ5l5fJ\\JZ5o5YJTJg5g600010O01O0010O0001O010O001O01O01O0010O01O00010O00010O001O01O01O0010O0001O010O000010O01O01O01O00010O001O01O01O0010O0001O010O1O00100O0011ON101O1N2O1O2N1N3M3M2N3M3M2N3M3M2N3M3N1N3M5K6J6J5K6E:E<C=D;D<E;F:G9F:FWXa1"}, "label": "older gentlemen at a table smiling at a dessert with a candle in it"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old lady\"."}], "task": "refering", "answer_template": "The \"an old lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 50, 51, 52, 73, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 212, 213, 214, 215, 238], "bbox": [0.136109375, 0.11699530516431926, 0.38442187499999997, 0.681056338028169], "iscrowd": 0, "area": 22813.32425, "rle": {"size": [426, 640], "counts": "X]T11W=4L4L4L4L4L2M4M2N3M2N3M2N3L3N3M2M4K4M4L3M3L5L3M3M4K4M3L5K4K5L5N1O101N101N101mM]L\\Jd3c5lLnIU3P6TMgIm2X6ZMbIg2\\6[MbIf2^6\\M`Ie2^6^M_Ic2`6_M_Ib2_6aM^I`2a6bM^I^2a6eM\\I\\2d6eM[I\\2c6fM[I[2d6hMYIY2f6iMXIX2g6jMWIX2g6jMWIW2i6kMTIV2k6mMSIS2l6W2N1O1O101O0001O01O0001O01O0001O010O000010O01O0010O01O00010O001O010O001O010O2N100O2O0O2O0O2O0O2O0UKhHb3Z7XLkHh3\\7kKjHT4]7aKhH_4Q80100O100O100O100O0010O00000010O0001O000010O0001O00010O01O2O1N3M2O1N2N3M3N2M2N3N2M2N3N2]N`FVOc9a0nFTOT9d0^GQOd8f0nGnNX8f0g2\\OQgS5"}, "label": "an old lady"}]}
{"id": 81200, "image": "COCO_train2014_000000081200.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the left\"."}], "task": "refering", "answer_template": "The \"the woman on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 50, 51, 52, 73, 74, 75, 76, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 212, 213, 214, 215, 238], "bbox": [0.136109375, 0.11699530516431926, 0.38442187499999997, 0.681056338028169], "iscrowd": 0, "area": 22813.32425, "rle": {"size": [426, 640], "counts": "X]T11W=4L4L4L4L4L2M4M2N3M2N3M2N3L3N3M2M4K4M4L3M3L5L3M3M4K4M3L5K4K5L5N1O101N101N101mM]L\\Jd3c5lLnIU3P6TMgIm2X6ZMbIg2\\6[MbIf2^6\\M`Ie2^6^M_Ic2`6_M_Ib2_6aM^I`2a6bM^I^2a6eM\\I\\2d6eM[I\\2c6fM[I[2d6hMYIY2f6iMXIX2g6jMWIX2g6jMWIW2i6kMTIV2k6mMSIS2l6W2N1O1O101O0001O01O0001O01O0001O010O000010O01O0010O01O00010O001O010O001O010O2N100O2O0O2O0O2O0O2O0UKhHb3Z7XLkHh3\\7kKjHT4]7aKhH_4Q80100O100O100O100O0010O00000010O0001O000010O0001O00010O01O2O1N3M2O1N2N3M3N2M2N3N2M2N3N2]N`FVOc9a0nFTOT9d0^GQOd8f0nGnNX8f0g2\\OQgS5"}, "label": "the woman on the left"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the black shirt\"."}], "task": "refering", "answer_template": "The \"the woman in the black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 92, 93, 94, 95, 114, 115, 116, 117, 136, 137, 138, 139, 159, 160, 161, 162, 163, 180, 181, 182, 183, 184, 185, 186, 201, 202, 203, 204, 205, 206, 207, 208, 223, 224, 225, 226, 227, 228, 229, 230, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 292, 293, 294, 315], "bbox": [0.16472131147540983, 0.16262992125984252, 0.4850327868852459, 0.6373700787401574], "iscrowd": 0, "area": 32615.204150000016, "rle": {"size": [635, 610], "counts": "UYn11dc06L4L5K4L4L4M3M3L5L3M3O100O2O000O101N10001N1gL_NXDa1d;eNYD]1c;iNZDW1e;lNYDT1f;oNWDS1h;POUDP1j;SOSDn0m;_OfCb0X<K\\C5c<0XC2g<1VCOi<5SCLm<7PCJn<:oBFQ=>jBCU=a0hB@V=d0fB]OZ=f0cB[O[=i0bBXO]=l0^BWO_=m0^BTOa=o0[BSOc=R1YBPOe=V1UBkNi=e3N3N2M3N3L3N2M3M3N2M3N3M2N2N3M2N3M5K5K4L5K5K5K5K1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O001O1O1O001O1O1O0100010O001O0010M2O1N2O2M2O1N2`IPES4Q;iKREV4Q;eKREZ4P;bKSE]4o:_KTE`4Q;YKREf4X;nJkDQ5`;bJcD]5h;VJ\\Dh5l<O1O1O1O1O1O1O1O10000O2O0000000O1000000O1N2M3M4L3M3M3M3M3M3M3M3M3M3M3N3L3M3M3N2M3N3L5L3M4K4M4L3L5L4L3M4K4M4L3M4L4M2M4L3M4M2M4L4M2M4L3M4M2M4M2L5I7IX_R6"}, "label": "the woman in the black shirt"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black graphic shirt watching the decoration over a boy ' s shoulder\"."}], "task": "refering", "answer_template": "The \"a woman in a black graphic shirt watching the decoration over a boy ' s shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 92, 93, 94, 95, 114, 115, 116, 117, 136, 137, 138, 139, 159, 160, 161, 162, 163, 180, 181, 182, 183, 184, 185, 186, 201, 202, 203, 204, 205, 206, 207, 208, 223, 224, 225, 226, 227, 228, 229, 230, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 292, 293, 294, 315], "bbox": [0.16472131147540983, 0.16262992125984252, 0.4850327868852459, 0.6373700787401574], "iscrowd": 0, "area": 32615.204150000016, "rle": {"size": [635, 610], "counts": "UYn11dc06L4L5K4L4L4M3M3L5L3M3O100O2O000O101N10001N1gL_NXDa1d;eNYD]1c;iNZDW1e;lNYDT1f;oNWDS1h;POUDP1j;SOSDn0m;_OfCb0X<K\\C5c<0XC2g<1VCOi<5SCLm<7PCJn<:oBFQ=>jBCU=a0hB@V=d0fB]OZ=f0cB[O[=i0bBXO]=l0^BWO_=m0^BTOa=o0[BSOc=R1YBPOe=V1UBkNi=e3N3N2M3N3L3N2M3M3N2M3N3M2N2N3M2N3M5K5K4L5K5K5K5K1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O001O1O1O001O1O1O0100010O001O0010M2O1N2O2M2O1N2`IPES4Q;iKREV4Q;eKREZ4P;bKSE]4o:_KTE`4Q;YKREf4X;nJkDQ5`;bJcD]5h;VJ\\Dh5l<O1O1O1O1O1O1O1O10000O2O0000000O1000000O1N2M3M4L3M3M3M3M3M3M3M3M3M3M3N3L3M3M3N2M3N3L5L3M4K4M4L3L5L4L3M4K4M4L3M4L4M2M4L3M4M2M4L4M2M4L3M4M2M4M2L5I7IX_R6"}, "label": "a woman in a black graphic shirt watching the decoration over a boy ' s shoulder"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in a red hoodie sitting at a table with scissors in his hand held against his chin\"."}], "task": "refering", "answer_template": "The \"a young boy in a red hoodie sitting at a table with scissors in his hand held against his chin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 244, 245, 246, 247, 266, 267, 268, 269, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 356, 357, 358, 359, 374, 375, 376, 377, 378, 379, 380, 381, 399, 400, 401], "bbox": [0.030475409836065573, 0.4459527559055118, 0.3587213114754098, 0.7882834645669291], "iscrowd": 0, "area": 27804.848849999995, "rle": {"size": [635, 610], "counts": "iW<:]c0;E:F4M3N2N2M3N2N2M3N2N2M3N3L3N2M3N2M3N2N2O100O1O1O1O100O1O2N100O1O1O100O1O1O1O100O1O1O100O2N1dMnLmCT3P<VMhCk2T<aMcC`2Z<jM^CW2_<TNXCm1e<^NQCd1m<_NQCb1m<aNQC`1n<bNoB`1P=bNnB_1Q=cNmB^1R=dNmB\\1R=eNoBZ1P=hNoBX1o<kNPCV1n<lNQCT1n<nNRCQ1n<POQCP1n<ROoBP1P=QOoBP1Q=QOmBP1R=ROlBo0S=SOjBo0V=k2O1O100O1O1001O000000001O0000001O01O000001O0000001O000000001O0000001O000000001O0000001O001O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O101N4L5K5K5I7I8H7K5J6K1O2N1N3N1O2N2M2O2N1O2N1N3N1O2N1N101O000O2O0010O0001O00010O001O01O01O000010O01O001L4L3M4L4O0`Nd^Oa0]a0ZOj^Oc0Wa0WOR_Oe0n`0VOZ_Og0g`0TOa_Oh0``0ROi_Oj0[bb7"}, "label": "a young boy in a red hoodie sitting at a table with scissors in his hand held against his chin"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child with scissors\"."}], "task": "refering", "answer_template": "The \"child with scissors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 244, 245, 246, 247, 266, 267, 268, 269, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 356, 357, 358, 359, 374, 375, 376, 377, 378, 379, 380, 381, 399, 400, 401], "bbox": [0.030475409836065573, 0.4459527559055118, 0.3587213114754098, 0.7882834645669291], "iscrowd": 0, "area": 27804.848849999995, "rle": {"size": [635, 610], "counts": "iW<:]c0;E:F4M3N2N2M3N2N2M3N2N2M3N3L3N2M3N2M3N2N2O100O1O1O1O100O1O2N100O1O1O100O1O1O1O100O1O1O100O2N1dMnLmCT3P<VMhCk2T<aMcC`2Z<jM^CW2_<TNXCm1e<^NQCd1m<_NQCb1m<aNQC`1n<bNoB`1P=bNnB_1Q=cNmB^1R=dNmB\\1R=eNoBZ1P=hNoBX1o<kNPCV1n<lNQCT1n<nNRCQ1n<POQCP1n<ROoBP1P=QOoBP1Q=QOmBP1R=ROlBo0S=SOjBo0V=k2O1O100O1O1001O000000001O0000001O01O000001O0000001O000000001O0000001O000000001O0000001O001O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O101N4L5K5K5I7I8H7K5J6K1O2N1N3N1O2N2M2O2N1O2N1N3N1O2N1N101O000O2O0010O0001O00010O001O01O01O000010O01O001L4L3M4L4O0`Nd^Oa0]a0ZOj^Oc0Wa0WOR_Oe0n`0VOZ_Og0g`0TOa_Oh0``0ROi_Oj0[bb7"}, "label": "child with scissors"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown - haired woman bends to add green frosting to a cupcake\"."}], "task": "refering", "answer_template": "The \"a brown - haired woman bends to add green frosting to a cupcake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 235, 236, 237, 251, 252, 253, 254, 255, 256, 257, 258, 273, 274, 275, 277, 278, 279, 294, 295, 296, 300, 301, 315, 316, 317, 320, 321, 322, 323, 340, 341, 342, 343, 344, 345, 361, 362, 363, 364, 365, 382, 383, 384, 405], "bbox": [0.356688524590164, 0.008724409448818898, 0.8959508196721311, 0.8059685039370079], "iscrowd": 0, "area": 71493.61985000003, "rle": {"size": [635, 610], "counts": "RbW41V30\\=?UBCj=`0RBAm=`0RBBl=?SBBm==TBDj==UBDj==UBEi=;WBFi=:VBHh=9WBHh=8YBIf=7YBJf=7YBKe=5[BLe=4ZBMe=4ZBLf=4ZBLg=4YBLf=5YBKg=5YBLg=4XBLh=5WBKi=5WBLi=4WBKi=6VBKi=5WBKj=5UBKk=7SBJl=7SBIm=9RBGk=<TBDj=`0TB@j=d0TB]Oi=f0VBZOh=j0WBVOaND\\>Z1QCROcNDY>^1SCoNcNDW>a1UCjNdNFT>c1WCgNeNGQ>f1XCdNgNFn=j1ZC_NhNGl=m1[C]NhNGj=P2\\CYNjNHg=R2_CUNjNJd=U2`CRNlNIa=Y2aCPNlNGa=[2cCjNZ<W1hCiNU<Y1kChNR<Y1oChNo;Y1RDfNm;Z1TDgNj;Y1WDhNg;Y1ZDgNh6fNPLP2lMZOi5SOXLd1oMZOb5ZO\\L]1RNZO\\5@_LW1VNYOT5HcLP1YNVOQ51dLi0[NTOm4:fLd0]NoNi4e0gL=`NlNg4m0fL8dNhNg4S1bL6hNdNf4Z1_L3kNaNf4`1\\L0oN\\Ng4f1XL0POXNh4l1ULMTOTNg4S2RLJXORNe4W2PLH[OQNc4[2oKE_OoMa4a2kKADoM^4f2iK\\OJmM[4m2fKWO0kMY4S3bKSO5kMV4Y3_KmN<iMT4_3WKmNe0dMT4c3jI2S2[LR4f8oKYGQ4g8oKYGR4g8nKYGQ4g8oKYGQ4g8PLXGP4h8QLXGo3g8QLYGo3g8RLXGn3h8RLYGm3g8TLXGl3h8ULXGk3g8ULYGk3g8VLXGj3h8VLYGi3g8XLYGg3g8ZLYGf3f8ZLZGf3f8[LZGd3g8[LZGd3f8]LZGc3e8^LZGb3f8^L[Ga3f8_LZG`3f8`L[G_3e8bLZG_3f8aLZG^3f8bL[G]3e8dL[G[3f8dL[G\\3d8eL[G[3e8fL[GY3e8gL\\GX3e8hL[GW3e8jLZGW3e8iL\\GV3c8lL]GS3b8nL_GQ3`8QM_GP3_8QMbGn2]8SMdGm2Z8UMfGk2X8VMiGi2V8XMjGi2T8YMlGg2S8YMnGg2P8[MPHd2o7]MQHd2m7]MTHc2j7_MVH`2k7_MVHa2k7]MUHd2m7ZMSHf2o7WMRHh2Q8UMPHk2Q8TMoGl2S8QMmGo2V8nLkGR3V8mLjGS3X8jLiGV3Y8hLfGX3\\8fLeGZ3]8cLdG]3]8bLcG^3^8`LbG`3`8^LaGb3`8]L`Gc3`8\\LaGc3a8[L`Ge3a8ZL^Gg3c8WL^Gi3c8VL]Gi3e8UL\\Gk3e8SL[Gn3e8RL[Gn3f8PL[GP4f8nK[GS4e8lK[GT4f8jKZGW4g8gKZGZ4e8fK[GZ4f8dK[G]4e8bKZG_4g8_KZGa4g8]KZGd4f8[KZGe4g8YKYGi4f8VK[Gj4f8UKZGk4g8SKZGn4f8PK[GP5f8oJYGS5g8kJZGU5f8jJ[GV5f8hJ[GY5e8dJ\\G]5e8aJ\\G`5d8^J]Gb5d8[J^Ge5c8YJ]Gi5b8VJ_Gj5b8SJ`Gn5`8PJaGP6`8nIaGS6_8kIaGW6_8fIcGZ6]8eIdG\\6\\8bIdG`6\\8]IfGd6Z8ZIgGg6Y8WIgGk6d:1N2O0O2O1O1N2O1N2O1N101O1N2O1N3N3L3XNWHZGk7_8]H^Gf7[8bH`Ga7Y8gHdG[7U8mHgGW7Q8RIkGP7n7XInGk6l7\\IQHf6h7bIUHa6c7lITHW6f7S3L5L3M3M4K4M4L4L4VJ]CT4h<eK\\CZ4h<_K]C_4g<[K\\Cd4h<UK]Ch4h<QK]Cm4g<lJ]CS5g=N2M2O2N2N1N3N1O2N2N1N3N2N1O2M3N1O2N2M2O2O0O2O1O0O3N1O1N2O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O2N1N2O2N1N3N1O2M2O1O2M2O2N1N3N1N2N3M2N3M2O2M2N3M2N2N3M2N3N1N3M2N3MZ^W1"}, "label": "a brown - haired woman bends to add green frosting to a cupcake"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman leaning over table putting icing on a cupcake for a child\"."}], "task": "refering", "answer_template": "The \"woman leaning over table putting icing on a cupcake for a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 229, 230, 231, 232, 233, 234, 235, 236, 237, 251, 252, 253, 254, 255, 256, 257, 258, 273, 274, 275, 277, 278, 279, 294, 295, 296, 300, 301, 315, 316, 317, 320, 321, 322, 323, 340, 341, 342, 343, 344, 345, 361, 362, 363, 364, 365, 382, 383, 384, 405], "bbox": [0.356688524590164, 0.008724409448818898, 0.8959508196721311, 0.8059685039370079], "iscrowd": 0, "area": 71493.61985000003, "rle": {"size": [635, 610], "counts": "RbW41V30\\=?UBCj=`0RBAm=`0RBBl=?SBBm==TBDj==UBDj==UBEi=;WBFi=:VBHh=9WBHh=8YBIf=7YBJf=7YBKe=5[BLe=4ZBMe=4ZBLf=4ZBLg=4YBLf=5YBKg=5YBLg=4XBLh=5WBKi=5WBLi=4WBKi=6VBKi=5WBKj=5UBKk=7SBJl=7SBIm=9RBGk=<TBDj=`0TB@j=d0TB]Oi=f0VBZOh=j0WBVOaND\\>Z1QCROcNDY>^1SCoNcNDW>a1UCjNdNFT>c1WCgNeNGQ>f1XCdNgNFn=j1ZC_NhNGl=m1[C]NhNGj=P2\\CYNjNHg=R2_CUNjNJd=U2`CRNlNIa=Y2aCPNlNGa=[2cCjNZ<W1hCiNU<Y1kChNR<Y1oChNo;Y1RDfNm;Z1TDgNj;Y1WDhNg;Y1ZDgNh6fNPLP2lMZOi5SOXLd1oMZOb5ZO\\L]1RNZO\\5@_LW1VNYOT5HcLP1YNVOQ51dLi0[NTOm4:fLd0]NoNi4e0gL=`NlNg4m0fL8dNhNg4S1bL6hNdNf4Z1_L3kNaNf4`1\\L0oN\\Ng4f1XL0POXNh4l1ULMTOTNg4S2RLJXORNe4W2PLH[OQNc4[2oKE_OoMa4a2kKADoM^4f2iK\\OJmM[4m2fKWO0kMY4S3bKSO5kMV4Y3_KmN<iMT4_3WKmNe0dMT4c3jI2S2[LR4f8oKYGQ4g8oKYGR4g8nKYGQ4g8oKYGQ4g8PLXGP4h8QLXGo3g8QLYGo3g8RLXGn3h8RLYGm3g8TLXGl3h8ULXGk3g8ULYGk3g8VLXGj3h8VLYGi3g8XLYGg3g8ZLYGf3f8ZLZGf3f8[LZGd3g8[LZGd3f8]LZGc3e8^LZGb3f8^L[Ga3f8_LZG`3f8`L[G_3e8bLZG_3f8aLZG^3f8bL[G]3e8dL[G[3f8dL[G\\3d8eL[G[3e8fL[GY3e8gL\\GX3e8hL[GW3e8jLZGW3e8iL\\GV3c8lL]GS3b8nL_GQ3`8QM_GP3_8QMbGn2]8SMdGm2Z8UMfGk2X8VMiGi2V8XMjGi2T8YMlGg2S8YMnGg2P8[MPHd2o7]MQHd2m7]MTHc2j7_MVH`2k7_MVHa2k7]MUHd2m7ZMSHf2o7WMRHh2Q8UMPHk2Q8TMoGl2S8QMmGo2V8nLkGR3V8mLjGS3X8jLiGV3Y8hLfGX3\\8fLeGZ3]8cLdG]3]8bLcG^3^8`LbG`3`8^LaGb3`8]L`Gc3`8\\LaGc3a8[L`Ge3a8ZL^Gg3c8WL^Gi3c8VL]Gi3e8UL\\Gk3e8SL[Gn3e8RL[Gn3f8PL[GP4f8nK[GS4e8lK[GT4f8jKZGW4g8gKZGZ4e8fK[GZ4f8dK[G]4e8bKZG_4g8_KZGa4g8]KZGd4f8[KZGe4g8YKYGi4f8VK[Gj4f8UKZGk4g8SKZGn4f8PK[GP5f8oJYGS5g8kJZGU5f8jJ[GV5f8hJ[GY5e8dJ\\G]5e8aJ\\G`5d8^J]Gb5d8[J^Ge5c8YJ]Gi5b8VJ_Gj5b8SJ`Gn5`8PJaGP6`8nIaGS6_8kIaGW6_8fIcGZ6]8eIdG\\6\\8bIdG`6\\8]IfGd6Z8ZIgGg6Y8WIgGk6d:1N2O0O2O1O1N2O1N2O1N101O1N2O1N3N3L3XNWHZGk7_8]H^Gf7[8bH`Ga7Y8gHdG[7U8mHgGW7Q8RIkGP7n7XInGk6l7\\IQHf6h7bIUHa6c7lITHW6f7S3L5L3M3M4K4M4L4L4VJ]CT4h<eK\\CZ4h<_K]C_4g<[K\\Cd4h<UK]Ch4h<QK]Cm4g<lJ]CS5g=N2M2O2N2N1N3N1O2N2N1N3N2N1O2M3N1O2N2M2O2O0O2O1O0O3N1O1N2O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1N2O2N1N2O2N1N3N1O2M2O1O2M2O2N1N3N1N2N3M2N3M2O2M2N3M2N2N3M2N3N1N3M2N3MZ^W1"}, "label": "woman leaning over table putting icing on a cupcake for a child"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with soccer glasses licking icing off of his finger\"."}], "task": "refering", "answer_template": "The \"a child with soccer glasses licking icing off of his finger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 367, 368, 369, 370, 371, 372, 373, 389, 390, 391, 392, 393, 394, 395, 410, 411, 412, 413, 414, 415, 416, 417, 432, 433, 434, 435, 436, 437, 438, 439, 454, 455, 456, 457, 458, 459, 460, 461, 476, 477, 478, 479, 480, 481, 482, 483, 498, 499, 500, 501, 502, 503, 504, 505], "bbox": [0.6433114754098361, 0.40899212598425194, 0.9988688524590164, 0.9797795275590551], "iscrowd": 0, "area": 61504.90299999999, "rle": {"size": [635, 610], "counts": "Ucc74fc0S1lNU1lNi0VO2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2eMgKSEZ4\\:YL`Ei3n9lLnEU3a9_MZFc2V9PNfFQ2Y9SNbFo1]9TN`Fm1_9WN\\Fk1c9YNYFh1f9\\NVFe1i9_NRFc1m9aNoE`1P:V4O100O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O10000001O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2RIfEW4\\:YKTFe4n9kJbFS5V<M2N2N2N2N2N2_Oa0ZOf0YOg0YOg0YOg0YOg0YOg0YOh6"}, "label": "a child with soccer glasses licking icing off of his finger"}]}
{"id": 490801, "image": "COCO_train2014_000000490801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in front of the dining table wearing white and black specs and white dress\"."}], "task": "refering", "answer_template": "The \"a boy in front of the dining table wearing white and black specs and white dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 367, 368, 369, 370, 371, 372, 373, 389, 390, 391, 392, 393, 394, 395, 410, 411, 412, 413, 414, 415, 416, 417, 432, 433, 434, 435, 436, 437, 438, 439, 454, 455, 456, 457, 458, 459, 460, 461, 476, 477, 478, 479, 480, 481, 482, 483, 498, 499, 500, 501, 502, 503, 504, 505], "bbox": [0.6433114754098361, 0.40899212598425194, 0.9988688524590164, 0.9797795275590551], "iscrowd": 0, "area": 61504.90299999999, "rle": {"size": [635, 610], "counts": "Ucc74fc0S1lNU1lNi0VO2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2eMgKSEZ4\\:YL`Ei3n9lLnEU3a9_MZFc2V9PNfFQ2Y9SNbFo1]9TN`Fm1_9WN\\Fk1c9YNYFh1f9\\NVFe1i9_NRFc1m9aNoE`1P:V4O100O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O10000001O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2RIfEW4\\:YKTFe4n9kJbFS5V<M2N2N2N2N2N2_Oa0ZOf0YOg0YOg0YOg0YOg0YOg0YOh6"}, "label": "a boy in front of the dining table wearing white and black specs and white dress"}]}
{"id": 105777, "image": "COCO_train2014_000000105777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane\"."}], "task": "refering", "answer_template": "The \"an airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 133, 134, 135, 136, 151, 152, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.00671875, 0.24310416666666665, 0.9496718749999999, 0.6681666666666667], "iscrowd": 0, "area": 42130.3404, "rle": {"size": [480, 640], "counts": "VU2`0`>00000O100000000000001O0000000O1000000000000000000000000O100000001O0000000000000O10000000000000000000000O10001O00000000000000000O10000000000000000000001O0O10000000000000000000000O10000000000000001O00000O100000000000000000000KSOWBm0i=TOVBl0i=UOVBl0j=UOUBk0k=5000001N1000000000000O10000000000O10000000000O101O000000000O10000000000O1000000000000O1000001O000O10000000000O1000000000000O10000000000O2O000000000O1000000O1O1O1O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1000000O100000000000000000001O000000000000000000O10000000000000000000000O100000O1000000000000000O1000000000000000000000000O10000000000000000000000O1000O100000000000000000O10000000000000000000000O10000000000000000000000O10O10000000000000000000O1000000000000000000000000O10000000000000000000000O01000000000000bNmMoES2o9oMQFQ2n9QNQFo1n9RNRFn1l9UNSFj1n9WNQFi1o9WNQFi1o9XNPFh1Q:WNoEi1Q:XNnEh1S:WNmEi1S:XNlEh1T:YNkEg1V:XNjEh1V:YNiEg1W:YNiEg1X:YNgEf1Z:ZNfEf1Z:[NeEe1\\:[NcEe1]:[NcEe1^:[NaEe1_:[NaEe1_:\\N`Ed1a:[N_Ee1a:\\N^Ed1b:]N]Ec1d:\\N\\Ed1d:]NZEc1g:]NYEc1h:]NWEc1i:]NWEc1i:^NVEb1k:^NTEb1l:^NTEb1m:^NREb1n:^NREb1n:_NQEa1P;^NPEb1P;_NoDa1Q;`NnD`1S;_NmD`1T;aNkD_1U;aNkD_1V;aNiD_1W;aNiD_1X;aNgD_1Y;bNfD^1Z;bNfD^1[;bNdD^1\\;bNdD^1\\;cNcD]1^;bNbD^1^;cNaD\\1`;eN_D[1b;dN^D\\1b;eN]D[1c;eN]D[1d;eN[D[1e;eN[D[1f;eNYD[1g;fNXDZ1h;fNXDZ1i;fNVDZ1j;fNUD[1k;fNTDY1n;fNRDZ1n;gNQDY1o;hNPDX1Q<gNoCY1Q<hNnCX1S<gNmCY1S<hNlCX1T<hNlCX1U<i000001N10001O001O00001O00001O0O101O00001O001O00001O00N101O1N2O1O1N2O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1N10000O101O0O100O10001N10000O100O2O000O1001O100O1O1O1O1O100O1O1O1O1M3M3M3L4M3M3M3L4M3M3M3L5L3M3M3L4M3M3M3L4M3M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M4K4M3M3M3L4M3M3M\\l>"}, "label": "an airplane"}]}
{"id": 105777, "image": "COCO_train2014_000000105777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small aircraft on the other side of the fence\"."}], "task": "refering", "answer_template": "The \"small aircraft on the other side of the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 133, 134, 135, 136, 151, 152, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.00671875, 0.24310416666666665, 0.9496718749999999, 0.6681666666666667], "iscrowd": 0, "area": 42130.3404, "rle": {"size": [480, 640], "counts": "VU2`0`>00000O100000000000001O0000000O1000000000000000000000000O100000001O0000000000000O10000000000000000000000O10001O00000000000000000O10000000000000000000001O0O10000000000000000000000O10000000000000001O00000O100000000000000000000KSOWBm0i=TOVBl0i=UOVBl0j=UOUBk0k=5000001N1000000000000O10000000000O10000000000O101O000000000O10000000000O1000000000000O1000001O000O10000000000O1000000000000O10000000000O2O000000000O1000000O1O1O1O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1000000O100000000000000000001O000000000000000000O10000000000000000000000O100000O1000000000000000O1000000000000000000000000O10000000000000000000000O1000O100000000000000000O10000000000000000000000O10000000000000000000000O10O10000000000000000000O1000000000000000000000000O10000000000000000000000O01000000000000bNmMoES2o9oMQFQ2n9QNQFo1n9RNRFn1l9UNSFj1n9WNQFi1o9WNQFi1o9XNPFh1Q:WNoEi1Q:XNnEh1S:WNmEi1S:XNlEh1T:YNkEg1V:XNjEh1V:YNiEg1W:YNiEg1X:YNgEf1Z:ZNfEf1Z:[NeEe1\\:[NcEe1]:[NcEe1^:[NaEe1_:[NaEe1_:\\N`Ed1a:[N_Ee1a:\\N^Ed1b:]N]Ec1d:\\N\\Ed1d:]NZEc1g:]NYEc1h:]NWEc1i:]NWEc1i:^NVEb1k:^NTEb1l:^NTEb1m:^NREb1n:^NREb1n:_NQEa1P;^NPEb1P;_NoDa1Q;`NnD`1S;_NmD`1T;aNkD_1U;aNkD_1V;aNiD_1W;aNiD_1X;aNgD_1Y;bNfD^1Z;bNfD^1[;bNdD^1\\;bNdD^1\\;cNcD]1^;bNbD^1^;cNaD\\1`;eN_D[1b;dN^D\\1b;eN]D[1c;eN]D[1d;eN[D[1e;eN[D[1f;eNYD[1g;fNXDZ1h;fNXDZ1i;fNVDZ1j;fNUD[1k;fNTDY1n;fNRDZ1n;gNQDY1o;hNPDX1Q<gNoCY1Q<hNnCX1S<gNmCY1S<hNlCX1T<hNlCX1U<i000001N10001O001O00001O00001O0O101O00001O001O00001O00N101O1N2O1O1N2O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1N10000O101O0O100O10001N10000O100O2O000O1001O100O1O1O1O1O100O1O1O1O1M3M3M3L4M3M3M3L4M3M3M3L5L3M3M3L4M3M3M3L4M3M3M3L4M3M3M3L4M3M3M3L4M3M3L4M3M3M4K4M3M3M3L4M3M3M\\l>"}, "label": "small aircraft on the other side of the fence"}]}
{"id": 56632, "image": "COCO_train2014_000000056632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"piece of fruit tart\"."}], "task": "refering", "answer_template": "The \"piece of fruit tart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221], "bbox": [0.5095468750000001, 0.3109111617312073, 0.8070937499999999, 0.6070159453302961], "iscrowd": 0, "area": 16076.018599999994, "rle": {"size": [439, 640], "counts": "jo[44[=>A>C=I8I1O1O1O1O001O1O1O1O010O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O100O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O10O10O100000O10O100000O10O11O00001N10001O000O2O00001O0O101O00001N10001O000O2O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O000_OPE\\NP;c1QE\\No:c1RE]Nn:c1SE\\Nn:c1SE\\Nn:b1SE^Nn:a1SE^Nn:`1TE_Nm:`1TE_Nm:`1SE`Nn:^1TEaNl:_1UE`Nl:_1ZE[Ng:c1h0O1N2O1O1N2O1J6K4K6J6K5J6J6Kkad1"}, "label": "piece of fruit tart"}]}
{"id": 56632, "image": "COCO_train2014_000000056632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow slice of dessert\"."}], "task": "refering", "answer_template": "The \"the yellow slice of dessert\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221], "bbox": [0.5095468750000001, 0.3109111617312073, 0.8070937499999999, 0.6070159453302961], "iscrowd": 0, "area": 16076.018599999994, "rle": {"size": [439, 640], "counts": "jo[44[=>A>C=I8I1O1O1O1O001O1O1O1O010O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O100O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O10O01O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O10O10O100000O10O100000O10O11O00001N10001O000O2O00001O0O101O00001N10001O000O2O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O000_OPE\\NP;c1QE\\No:c1RE]Nn:c1SE\\Nn:c1SE\\Nn:b1SE^Nn:a1SE^Nn:`1TE_Nm:`1TE_Nm:`1SE`Nn:^1TEaNl:_1UE`Nl:_1ZE[Ng:c1h0O1N2O1O1N2O1J6K4K6J6K5J6J6Kkad1"}, "label": "the yellow slice of dessert"}]}
{"id": 458042, "image": "COCO_train2014_000000458042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe standing between two fences\"."}], "task": "refering", "answer_template": "The \"giraffe standing between two fences\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 158, 159, 180, 181, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250], "bbox": [0.75946875, 0.271598440545809, 0.9512031249999999, 0.6131384015594542], "iscrowd": 0, "area": 6127.5217999999995, "rle": {"size": [513, 640], "counts": "Qhc73^?Nk@8T?Ig@=X?Ch@>V?Cj@=U?Dk@=S?Dm@<R?En@<Q?Do@<P?EPA<n>ERA<l>ETA;k>FUA;j>EVA;i>FWA;g>FYA;f>EZA;e>F[A;d>E\\A;c>F]A;a>F_A:a>F_A;_>FaA;]>FcA:]>FcA;[>FeA:[>FeA;Y>FgA:X>GhA:W>FiA;U>FkA:U>FkA:T>GlA8T>HmA7S>JmA5R>MmA3S>NmA2R>OnA0Q>2oANP>3PBMo=4QBLm=6SBJl=7TBIj=9VBHh=9XBGg=:YBK`=7`BM[=4eBOW=2iBNU=4kBLT=5lBKS=6mBIS=8mBHS=8mBHS=8mBGS=:mBDU=<kB@X=a0hB[O\\=e0dBYO^=g0bBYO]=h0cBXO]=h0cBXO\\=i0dBWO\\=i0dBIi<8WCHg<:YCFf<;ZCEd<=\\CCc<>]CBa<`0_C@`<a0`C_O^<3`CoN2h0c<8_CnNNc0h<?^ClNJ=o<g0YCkNH7T=n0XCiND4X=S1WCgNA4X=U1[CeN]O3Z=W1\\CeNZO2Z=Z1cCcN_<\\1R101O001O001OROgNVCY1l<hNRCW1o<kNnBU1T=lNjBS1X=nNeBQ1^=PO_BP1b=QO]Bn0e=SOXBl0k=TOTBk0n=VOnAk0T>UOhAl0Y>=O10O5L4L4K4M2O1N1O2NN2O14L4K5L3M3N1N2O1N2O1O1O3MZ\\?"}, "label": "giraffe standing between two fences"}]}
{"id": 458042, "image": "COCO_train2014_000000458042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe in distance\"."}], "task": "refering", "answer_template": "The \"giraffe in distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 158, 159, 180, 181, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250], "bbox": [0.75946875, 0.271598440545809, 0.9512031249999999, 0.6131384015594542], "iscrowd": 0, "area": 6127.5217999999995, "rle": {"size": [513, 640], "counts": "Qhc73^?Nk@8T?Ig@=X?Ch@>V?Cj@=U?Dk@=S?Dm@<R?En@<Q?Do@<P?EPA<n>ERA<l>ETA;k>FUA;j>EVA;i>FWA;g>FYA;f>EZA;e>F[A;d>E\\A;c>F]A;a>F_A:a>F_A;_>FaA;]>FcA:]>FcA;[>FeA:[>FeA;Y>FgA:X>GhA:W>FiA;U>FkA:U>FkA:T>GlA8T>HmA7S>JmA5R>MmA3S>NmA2R>OnA0Q>2oANP>3PBMo=4QBLm=6SBJl=7TBIj=9VBHh=9XBGg=:YBK`=7`BM[=4eBOW=2iBNU=4kBLT=5lBKS=6mBIS=8mBHS=8mBHS=8mBGS=:mBDU=<kB@X=a0hB[O\\=e0dBYO^=g0bBYO]=h0cBXO]=h0cBXO\\=i0dBWO\\=i0dBIi<8WCHg<:YCFf<;ZCEd<=\\CCc<>]CBa<`0_C@`<a0`C_O^<3`CoN2h0c<8_CnNNc0h<?^ClNJ=o<g0YCkNH7T=n0XCiND4X=S1WCgNA4X=U1[CeN]O3Z=W1\\CeNZO2Z=Z1cCcN_<\\1R101O001O001OROgNVCY1l<hNRCW1o<kNnBU1T=lNjBS1X=nNeBQ1^=PO_BP1b=QO]Bn0e=SOXBl0k=TOTBk0n=VOnAk0T>UOhAl0Y>=O10O5L4L4K4M2O1N1O2NN2O14L4K5L3M3N1N2O1N2O1O1O3MZ\\?"}, "label": "giraffe in distance"}]}
{"id": 458042, "image": "COCO_train2014_000000458042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest giraffe that is looking down\"."}], "task": "refering", "answer_template": "The \"the closest giraffe that is looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 212, 213, 217, 218, 219, 220, 221, 222, 223, 224, 235, 236, 241, 242, 243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390, 409, 410, 411, 412, 413], "bbox": [0.21974999999999997, 0.22922027290448344, 1.0, 0.9887719298245614], "iscrowd": 0, "area": 69905.99265, "rle": {"size": [513, 640], "counts": "WmV23k?4L4L4L4L3M4L4L4L3M4L4L4L4L3M4L3fLRNgGh2X8[MmF]3T9eLiF\\3V9hLfFX3Z9kLcFV3]9lL`FT3`9oL^FQ3b9QM[Fo2e9UMWFl2i9VMTFk2k9XMRFh2o9ZMnEg2Q:\\MlEd2V:^MgEb2Z:_McEa2`:_M]Eb2d:_MYEa2j:_MSEb2n:W12O2N1N2L5K4K5K6J5K5O2M2O2M00N3L3N3L3M4M2M4N1100O00100O100O1O100O100O;F>A00100O010O00100O010O10O01O10O010O01O10O0100O010O1O101N2O0O2O00100O00100O00100O010O1O010O1O010O00100O00100O010O1O010O1O010O001001N10000O10000O10000O10000O2O0O10000O10000O10000O10001N10000O100O10000O10000O101O0O10000O10000O100O10000O2O000O10000O10000O100O10001N10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10O010000O10000O10000O10000O100O10000O10000O10000O10000O10000O10O010000O10000O10000O10000O10000O100O1000O010000O10000O100O10O0100O100O100O100O010O100O100O10000O10O0100O100O100O100O10O0100O100O100O100O010O100O100O100O1000O0100O100O100O100O010O100O100O100O2O1N2O0O2O1N2O0O2O1N2O001N`0A4K01O001O001O001O001O001O001O001O001O001O001O00001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000_E"}, "label": "the closest giraffe that is looking down"}]}
{"id": 572732, "image": "COCO_train2014_000000572732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair with nothing on it\"."}], "task": "refering", "answer_template": "The \"the chair with nothing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 231, 232, 254, 255, 256, 277, 278, 279, 300, 301, 302, 303, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 351, 369, 370, 371, 372, 373, 374], "bbox": [0.043484375, 0.5305416666666667, 0.278015625, 0.9987291666666667], "iscrowd": 0, "area": 17789.085349999998, "rle": {"size": [480, 640], "counts": "o[=;g>?@`0@`0@?B?@`0@`0A>A`0@`0@`0A?@:FO1OO10000O10000O10O10O100O10000OD=000O2O000OjJ\\GR4d8lK`GR4`8mKcGQ4\\8PLeGo3[8PLhGn3W8RLlGl3T8TLnGk3Q8TLPHl3o7TLRHl3n7TLRHl3n7SLSHm3l7SLUHm3k7RLUHo3k7nKXHR4[900001O0000000000000004L6K5J5K6J6J6J6J6J5K6K5J6J5K00000000000010O00000000000000001O0000000000000000001O0001O000000000001O0000000000000000001O0000000001O000001O0000000000000000001O0000000000000001mMZDk0e;kNUEe0T=_O[Sh6"}, "label": "the chair with nothing on it"}]}
{"id": 572732, "image": "COCO_train2014_000000572732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty chair in the left side of the image\"."}], "task": "refering", "answer_template": "The \"empty chair in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 231, 232, 254, 255, 256, 277, 278, 279, 300, 301, 302, 303, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 351, 369, 370, 371, 372, 373, 374], "bbox": [0.043484375, 0.5305416666666667, 0.278015625, 0.9987291666666667], "iscrowd": 0, "area": 17789.085349999998, "rle": {"size": [480, 640], "counts": "o[=;g>?@`0@`0@?B?@`0@`0A>A`0@`0@`0A?@:FO1OO10000O10000O10O10O100O10000OD=000O2O000OjJ\\GR4d8lK`GR4`8mKcGQ4\\8PLeGo3[8PLhGn3W8RLlGl3T8TLnGk3Q8TLPHl3o7TLRHl3n7TLRHl3n7SLSHm3l7SLUHm3k7RLUHo3k7nKXHR4[900001O0000000000000004L6K5J5K6J6J6J6J6J5K6K5J6J5K00000000000010O00000000000000001O0000000000000000001O0001O000000000001O0000000000000000001O0000000001O000001O0000000000000000001O0000000000000001mMZDk0e;kNUEe0T=_O[Sh6"}, "label": "empty chair in the left side of the image"}]}
{"id": 572732, "image": "COCO_train2014_000000572732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair that a girl is sitting in\"."}], "task": "refering", "answer_template": "The \"a chair that a girl is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 335, 336, 351, 352, 358, 374, 375, 381], "bbox": [0.275171875, 0.5348958333333333, 0.630875, 0.9867916666666666], "iscrowd": 0, "area": 28495.41385000001, "rle": {"size": [480, 640], "counts": "Shb2<d>Y1gNY1fNZ1gNZ1fNU1kNO01000000001O1O1O003L8I7I7I000000000dM\\G\\Nd8d1SHeMm7[2kHlLV7T3[20000000000000000000000000000000000000000000000000000000000000001O000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000^FlLX7T3WH^Mh7b2fGPNZ8P2UGaNk8j30000010O000000000001O0000000YOZJTHf5l7eJiG[5V8h0OG:F9G:G8G:F;E:G9F;E:F:G:E:F:F;F9F:Fi0WOb[^3"}, "label": "a chair that a girl is sitting in"}]}
{"id": 572732, "image": "COCO_train2014_000000572732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair she is sitting on\"."}], "task": "refering", "answer_template": "The \"the chair she is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 335, 336, 351, 352, 358, 374, 375, 381], "bbox": [0.275171875, 0.5348958333333333, 0.630875, 0.9867916666666666], "iscrowd": 0, "area": 28495.41385000001, "rle": {"size": [480, 640], "counts": "Shb2<d>Y1gNY1fNZ1gNZ1fNU1kNO01000000001O1O1O003L8I7I7I000000000dM\\G\\Nd8d1SHeMm7[2kHlLV7T3[20000000000000000000000000000000000000000000000000000000000000001O000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000^FlLX7T3WH^Mh7b2fGPNZ8P2UGaNk8j30000010O000000000001O0000000YOZJTHf5l7eJiG[5V8h0OG:F9G:G8G:F;E:G9F;E:F:G:E:F:F;F9F:Fi0WOb[^3"}, "label": "the chair she is sitting on"}]}
{"id": 466242, "image": "COCO_train2014_000000466242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich in the right side of the image\"."}], "task": "refering", "answer_template": "The \"sandwich in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 79, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 223], "bbox": [0.470046875, 0.16322916666666665, 0.7958125, 0.5398333333333334], "iscrowd": 0, "area": 26071.549850000003, "rle": {"size": [480, 640], "counts": "TV]44l>5K5K6I6K5K5J6K6J5K5J5L4L4L3L5L4L4L1N2O1O1N3N1O1O1N2O2N1O1N2O1O2M2O1O1O1N3N1O1O1N2O2N1O0O2O1O1N2O1O1O1N2O1O1O1N2O1O001N2O1O1O1O1N2O1O1O1O1O1N2O001O1O1N2O1O1O1O1N2O1O1O1O1N101O1O1O1N200OO1N2N3N1N2N2N3N1N200010O000001O01O0001O01O0001O0001O01O0000100O001O100O1O1O00100O1O1O1O10O01O1O100O1O001O1O100O1O1O00100O1O1O1O00100O1O1O1O10O01O1O2N100O1O1O1N3M2O1N2N2O1N3M2O1N2N2O1N3M2O1N2N2O5J>dMcCZ1S=O1N3N1N3M2O1N3M2O1N3M2O2M2M3N<CWVm1"}, "label": "sandwich in the right side of the image"}]}
{"id": 466242, "image": "COCO_train2014_000000466242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich half to the right of other half\"."}], "task": "refering", "answer_template": "The \"sandwich half to the right of other half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 79, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 223], "bbox": [0.470046875, 0.16322916666666665, 0.7958125, 0.5398333333333334], "iscrowd": 0, "area": 26071.549850000003, "rle": {"size": [480, 640], "counts": "TV]44l>5K5K6I6K5K5J6K6J5K5J5L4L4L3L5L4L4L1N2O1O1N3N1O1O1N2O2N1O1N2O1O2M2O1O1O1N3N1O1O1N2O2N1O0O2O1O1N2O1O1O1N2O1O1O1N2O1O001N2O1O1O1O1N2O1O1O1O1O1N2O001O1O1N2O1O1O1O1N2O1O1O1O1N101O1O1O1N200OO1N2N3N1N2N2N3N1N200010O000001O01O0001O01O0001O0001O01O0000100O001O100O1O1O00100O1O1O1O10O01O1O100O1O001O1O100O1O1O00100O1O1O1O00100O1O1O1O10O01O1O2N100O1O1O1N3M2O1N2N2O1N3M2O1N2N2O1N3M2O1N2N2O5J>dMcCZ1S=O1N3N1N3M2O1N3M2O1N3M2O2M2M3N<CWVm1"}, "label": "sandwich half to the right of other half"}]}
{"id": 466242, "image": "COCO_train2014_000000466242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwhich with onions and tomatoes by the potato chips\"."}], "task": "refering", "answer_template": "The \"half of a sandwhich with onions and tomatoes by the potato chips\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 260], "bbox": [0.2323125, 0.21575, 0.4928906250000001, 0.6594166666666667], "iscrowd": 0, "area": 26932.9555, "rle": {"size": [480, 640], "counts": "YRV28d>9G9G;E=C=C=C=C:G4L3N2M3M4M2M3M4M2M3N3L3M3N2M4M2N2N3M2O1N3M2N2N2N3M2O1N3M2N2N3M2N2N2O2M2N2N3M2O1O2O0000001N10001O0O101O00000O10N2N2N2N1O2O1N2O1O001N2O1O10O2O0000001N10001O0000001O00001O000O2O0000001O00001O000000K5G9J6N2N2N2N2N2N1O2N2N2N200000000O100000000O100000000O010000000O10000001O1O001O001O1O001O03N5K5J6K5J6K5K5J6K5J6K5K5J[1fNQTh4"}, "label": "half of a sandwhich with onions and tomatoes by the potato chips"}]}
{"id": 466242, "image": "COCO_train2014_000000466242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwich on the left\"."}], "task": "refering", "answer_template": "The \"the half of the sandwich on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 260], "bbox": [0.2323125, 0.21575, 0.4928906250000001, 0.6594166666666667], "iscrowd": 0, "area": 26932.9555, "rle": {"size": [480, 640], "counts": "YRV28d>9G9G;E=C=C=C=C:G4L3N2M3M4M2M3M4M2M3N3L3M3N2M4M2N2N3M2O1N3M2N2N2N3M2O1N3M2N2N3M2N2N2O2M2N2N3M2O1O2O0000001N10001O0O101O00000O10N2N2N2N1O2O1N2O1O001N2O1O10O2O0000001N10001O0000001O00001O000O2O0000001O00001O000000K5G9J6N2N2N2N2N2N1O2N2N2N200000000O100000000O100000000O010000000O10000001O1O001O001O1O001O03N5K5J6K5J6K5K5J6K5J6K5K5J[1fNQTh4"}, "label": "the half of the sandwich on the left"}]}
{"id": 220485, "image": "COCO_train2014_000000220485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown wooden chair against a white wall\"."}], "task": "refering", "answer_template": "The \"brown wooden chair against a white wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278], "bbox": [0.0, 0.1438125, 0.265171875, 0.557296875], "iscrowd": 0, "area": 35109.226650000004, "rle": {"size": [640, 640], "counts": "Y3`0`c0o0QOo0QOP1QA]Na9Q6000O1000O100000000000000000000000000000000O1000000000000000000000000000O100000O10000000000000O1O1O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O10O1000001O00000O2O000000001N100000001O0O10001O00000O101OO10000O01000O10000O10O10O101O1N3N1O1O2M2O1O2N1O1O4Ld0\\Oe0[Oe0ZOe0\\Oh0XOc0]O001O1O1O1O0O2O1O1O1M2K6I7J6I7J5J7J6J6I6K6I[`U9"}, "label": "brown wooden chair against a white wall"}]}
{"id": 220485, "image": "COCO_train2014_000000220485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty black chair\"."}], "task": "refering", "answer_template": "The \"an empty black chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278], "bbox": [0.0, 0.1438125, 0.265171875, 0.557296875], "iscrowd": 0, "area": 35109.226650000004, "rle": {"size": [640, 640], "counts": "Y3`0`c0o0QOo0QOP1QA]Na9Q6000O1000O100000000000000000000000000000000O1000000000000000000000000000O100000O10000000000000O1O1O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O10O1000001O00000O2O000000001N100000001O0O10001O00000O101OO10000O01000O10000O10O10O101O1N3N1O1O2M2O1O2N1O1O4Ld0\\Oe0[Oe0ZOe0\\Oh0XOc0]O001O1O1O1O0O2O1O1O1M2K6I7J6I7J5J7J6J6I6K6I[`U9"}, "label": "an empty black chair"}]}
{"id": 220485, "image": "COCO_train2014_000000220485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on which the child is sitting\"."}], "task": "refering", "answer_template": "The \"the chair on which the child is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 328, 341, 342, 343, 344, 351, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390, 409, 410, 411, 412, 413, 433, 434, 435, 436], "bbox": [0.25842187499999997, 0.4539375, 1.0, 0.808984375], "iscrowd": 0, "area": 23329.886699999992, "rle": {"size": [640, 640], "counts": "haW32`c0=XOBb]Ol0^b0;0000O10O10000000O01000000000O02O001O1O1O0O4M4L4L4L4K5L3M4LWXT68QgkId0L5J5L5K4L4L5J5L5K4L4L5J3N00001O0oK\\MeFe2[9^MbFb2^9`M`F`2`9cM]F]2c9fMZFZ2f9hMWFY2j9iMSFX2l9kMQFU2o9nMnER2R:PNlEP2T:SNhEn1X:UNeEk1[:WNbEj1^:YN[Ek1f:UNVEn1j:SNQER2n:oMnDT2R;mMiDW2W;jMeDY2[;hM`D\\2`;eM\\D^2e;bMVDb2j;_MRDd2n;]MmCg2S<ZMiCj2V<WMeCm2[<TMaCo2_<VMXCn2h<WMPCl2Q=Q2000001O000001O0000000001O0001O0000000000010O000000000000010O0000000000001O01O000000000001O0001O00000001O000001O00000001O00000001O0001L3M3L4M3M3M3M3L5L3M3M3M3L4MbF"}, "label": "the chair on which the child is sitting"}]}
{"id": 220485, "image": "COCO_train2014_000000220485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black chair girl is sitting on\"."}], "task": "refering", "answer_template": "The \"black chair girl is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 328, 341, 342, 343, 344, 351, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390, 409, 410, 411, 412, 413, 433, 434, 435, 436], "bbox": [0.25842187499999997, 0.4539375, 1.0, 0.808984375], "iscrowd": 0, "area": 23329.886699999992, "rle": {"size": [640, 640], "counts": "haW32`c0=XOBb]Ol0^b0;0000O10O10000000O01000000000O02O001O1O1O0O4M4L4L4L4K5L3M4LWXT68QgkId0L5J5L5K4L4L5J5L5K4L4L5J3N00001O0oK\\MeFe2[9^MbFb2^9`M`F`2`9cM]F]2c9fMZFZ2f9hMWFY2j9iMSFX2l9kMQFU2o9nMnER2R:PNlEP2T:SNhEn1X:UNeEk1[:WNbEj1^:YN[Ek1f:UNVEn1j:SNQER2n:oMnDT2R;mMiDW2W;jMeDY2[;hM`D\\2`;eM\\D^2e;bMVDb2j;_MRDd2n;]MmCg2S<ZMiCj2V<WMeCm2[<TMaCo2_<VMXCn2h<WMPCl2Q=Q2000001O000001O0000000001O0001O0000000000010O000000000000010O0000000000001O01O000000000001O0001O00000001O000001O00000001O00000001O0001L3M3L4M3M3M3M3L5L3M3M3M3L4MbF"}, "label": "black chair girl is sitting on"}]}
{"id": 138567, "image": "COCO_train2014_000000138567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black cat with eyes showing\"."}], "task": "refering", "answer_template": "The \"black cat with eyes showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390, 405, 406, 407, 408, 409, 410, 411, 412, 413, 428, 429, 430, 431, 432, 433, 434, 435, 436, 451, 452, 453, 454, 455, 456, 457, 458, 459, 474, 475, 476, 477, 478, 479, 480, 481, 482, 497, 498, 499, 500, 501, 502, 503, 504, 505, 520, 521, 522, 523, 524, 525, 526, 527, 528], "bbox": [0.42307812499999997, 0.04903125, 1.0, 0.98653125], "iscrowd": 0, "area": 148092.32305000004, "rle": {"size": [640, 640], "counts": "VRZ51oc01O00100O1O001O1O1O010O1O1O001O100O001O1O1O00100O001O1O1O00100O1O002N2N2N2O2M2N2N2N3M2a^OiNQ`0Y1a_OZOY`0g0Y_OL``0Q2I7I7I7I6J5L1N2N3M2N2N2O1N3M2N2N2N2M4M2N2N1O2M2O1O2N1N3N1O1O2N1N3N1O1O2N1N3N1O1O2M2O2N1O1O2M2O2N1O1O2M:Gk0UO9G2N3M2N2M3N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2M3N3M2N2N2N2N2N2M3N2N2N8oIhHjNe7:jHcNhLVOj:[1nH_Ob8EmG;`8hNPHX1\\=1O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00000000001O000000001O00000000001O000000001O00000000001O000000010O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1SOm0VGS3"}, "label": "black cat with eyes showing"}]}
{"id": 138567, "image": "COCO_train2014_000000138567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a close - up of a shaggy black cat\"."}], "task": "refering", "answer_template": "The \"a close - up of a shaggy black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390, 405, 406, 407, 408, 409, 410, 411, 412, 413, 428, 429, 430, 431, 432, 433, 434, 435, 436, 451, 452, 453, 454, 455, 456, 457, 458, 459, 474, 475, 476, 477, 478, 479, 480, 481, 482, 497, 498, 499, 500, 501, 502, 503, 504, 505, 520, 521, 522, 523, 524, 525, 526, 527, 528], "bbox": [0.42307812499999997, 0.04903125, 1.0, 0.98653125], "iscrowd": 0, "area": 148092.32305000004, "rle": {"size": [640, 640], "counts": "VRZ51oc01O00100O1O001O1O1O010O1O1O001O100O001O1O1O00100O001O1O1O00100O1O002N2N2N2O2M2N2N2N3M2a^OiNQ`0Y1a_OZOY`0g0Y_OL``0Q2I7I7I7I6J5L1N2N3M2N2N2O1N3M2N2N2N2M4M2N2N1O2M2O1O2N1N3N1O1O2N1N3N1O1O2N1N3N1O1O2M2O2N1O1O2M2O2N1O1O2M:Gk0UO9G2N3M2N2M3N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2M3N3M2N2N2N2N2N2M3N2N2N8oIhHjNe7:jHcNhLVOj:[1nH_Ob8EmG;`8hNPHX1\\=1O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00000000001O000000001O00000000001O000000001O00000000001O000000010O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1SOm0VGS3"}, "label": "a close - up of a shaggy black cat"}]}
{"id": 138567, "image": "COCO_train2014_000000138567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog staring at a woman on the couch\"."}], "task": "refering", "answer_template": "The \"a dog staring at a woman on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380, 381, 382, 400, 401, 402, 403, 404, 405, 423, 424, 425, 426, 427, 428, 446, 447, 448, 449, 450, 451, 470], "bbox": [0.3116875, 0.345453125, 0.636359375, 0.8805156249999999], "iscrowd": 0, "area": 45358.01285, "rle": {"size": [640, 640], "counts": "fcl31oc07H9H8H7H9H8H7H9H7I7H1000000O10000O10000O10000O10000O1000000O100__O_Mc?a2T@kMh?V2P@UNm?R3M3L4M3M3L4M3M3L4M3M3M3L4M3M3L4M3M3L4M3M3L4M3M3L4M3M3M3L4M3M3dDPIa9T7VFTIh9o6PFYIm9j6kE^IQ:f6gEaIW:b6`EgI]:\\6[ElIa:l7N3M3L4M2N3M3L3N3M3M2O2M3M2N3N0O000001O0O1000001N1000001O0O1000001N1000001O0O2O1O1O1N2O1O1O1O1N3N1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O2N1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O2N1O1N2O5K7I6I8I6J6J7I6I6K2N2N2M3N1O2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N1O2F:^Ob0^Ob0_Oa0^Ob0_Oa0^Ob0^OgWa4"}, "label": "a dog staring at a woman on the couch"}]}
{"id": 138567, "image": "COCO_train2014_000000138567.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of a cats back\"."}], "task": "refering", "answer_template": "The \"the reflection of a cats back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 377, 378, 379, 380, 381, 382, 400, 401, 402, 403, 404, 405, 423, 424, 425, 426, 427, 428, 446, 447, 448, 449, 450, 451, 470], "bbox": [0.3116875, 0.345453125, 0.636359375, 0.8805156249999999], "iscrowd": 0, "area": 45358.01285, "rle": {"size": [640, 640], "counts": "fcl31oc07H9H8H7H9H8H7H9H7I7H1000000O10000O10000O10000O10000O1000000O100__O_Mc?a2T@kMh?V2P@UNm?R3M3L4M3M3L4M3M3L4M3M3M3L4M3M3L4M3M3L4M3M3L4M3M3L4M3M3M3L4M3M3dDPIa9T7VFTIh9o6PFYIm9j6kE^IQ:f6gEaIW:b6`EgI]:\\6[ElIa:l7N3M3L4M2N3M3L3N3M3M2O2M3M2N3N0O000001O0O1000001N1000001O0O1000001N1000001O0O2O1O1O1N2O1O1O1O1N3N1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O2N1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O2N1O1N2O5K7I6I8I6J6J7I6I6K2N2N2M3N1O2N2N2M3N2N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2N2M3N2N1O2F:^Ob0^Ob0_Oa0^Ob0_Oa0^Ob0^OgWa4"}, "label": "the reflection of a cats back"}]}
{"id": 556360, "image": "COCO_train2014_000000556360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a notebook under a laptop\"."}], "task": "refering", "answer_template": "The \"a notebook under a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 27, 49, 50, 72, 73, 95, 96, 118, 119, 140, 141, 142, 161, 162, 164], "bbox": [0.002984375, 0.0045047169811320755, 0.19949999999999998, 0.503372641509434], "iscrowd": 0, "area": 5362.0062, "rle": {"size": [424, 640], "counts": "mP15S=3M000000000001O000O10000000000000001O000O1000000000000000001N1000000000000000000On`30S_L0O1000000000001O00000000000000001O000000000000001O0fH1[1O`N:[1G`N>`1B[Nc0e1]OWNf0j1ZOQNk0o1UOlMP1T2QOfMT1Z2lNaMY1_2gN]M\\1d2dNWMa1i2_NRMf1n2ZNmLk1S3UNhLP2X3PNdLT2\\3mM^LW2c3iMXL\\2h3dMSLa2m3_MoKe2Q4[MjKj2V4VMeKn2\\4RM_KS3a4mLZKX3f4iLUK[3k4eLPK`3P5`LkJe3U5[LfJi3[5WLaJm3^5TL]JQ4c5oKXJV4h5jKSJ[4m5fKmI^4T6bKhIb4X6^KcIg4]6YK^Il4b6TKYIQ5g6oJTIV5l6jJPIY5Q7?0000`0@Q1nNR1oNQ1oNQ1oNQ1oN;EVWd6"}, "label": "a notebook under a laptop"}]}
{"id": 556360, "image": "COCO_train2014_000000556360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book underneath a laptop\"."}], "task": "refering", "answer_template": "The \"a book underneath a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 27, 49, 50, 72, 73, 95, 96, 118, 119, 140, 141, 142, 161, 162, 164], "bbox": [0.002984375, 0.0045047169811320755, 0.19949999999999998, 0.503372641509434], "iscrowd": 0, "area": 5362.0062, "rle": {"size": [424, 640], "counts": "mP15S=3M000000000001O000O10000000000000001O000O1000000000000000001N1000000000000000000On`30S_L0O1000000000001O00000000000000001O000000000000001O0fH1[1O`N:[1G`N>`1B[Nc0e1]OWNf0j1ZOQNk0o1UOlMP1T2QOfMT1Z2lNaMY1_2gN]M\\1d2dNWMa1i2_NRMf1n2ZNmLk1S3UNhLP2X3PNdLT2\\3mM^LW2c3iMXL\\2h3dMSLa2m3_MoKe2Q4[MjKj2V4VMeKn2\\4RM_KS3a4mLZKX3f4iLUK[3k4eLPK`3P5`LkJe3U5[LfJi3[5WLaJm3^5TL]JQ4c5oKXJV4h5jKSJ[4m5fKmI^4T6bKhIb4X6^KcIg4]6YK^Il4b6TKYIQ5g6oJTIV5l6jJPIY5Q7?0000`0@Q1nNR1oNQ1oNQ1oNQ1oN;EVWd6"}, "label": "a book underneath a laptop"}]}
{"id": 556360, "image": "COCO_train2014_000000556360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the picture on the bottom left\"."}], "task": "refering", "answer_template": "The \"the picture on the bottom left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.002984375, 0.555188679245283, 0.196953125, 0.9898584905660378], "iscrowd": 0, "area": 16026.747949999999, "rle": {"size": [424, 640], "counts": "^V1T1R91VHd0X7MiH3S70nH0n64QIMk66VIJf69[IGb6<]IE_6>bIBZ6a0gI_OU6d0kI]OV6b0jI^OV6a0kI_OU6`0lI@T6`0kIAU6>lIBU6<lIDT6<kIEU6;kIEU6;kIEV6:iIGW69iIGW69iIGW69hIHX68hIHY67gIIY67gIIY67fIJZ66fIJZ66fIJ[65dIL\\64dIL\\64dIL\\64cIM]63cIM^62bIN^62bIN^62aIO_61aIO`60`I0`60_I1a6O_I1a6O_I1a6O^I2c6M]I4b6L^I4b6L^I4b6L]I5c6K\\I6e6IZI8f6HYI9g6GXI:h6FXI:h6FWI;j6ETI<l6ERI<n6EPI<P7EnH<S7DkH=U7DiH>V7ChH>X7i2001O000000001O000000001O0000001O000000001O000000001O0M3L4L4L5K4L4K5L5K4L4L4L4L5K4L4L4L4L5K4L4L4K5L5K4L4L4L5K4L4L4L4L5K4L4L9G\\bd6"}, "label": "the picture on the bottom left"}]}
{"id": 556360, "image": "COCO_train2014_000000556360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a page in a magazine showing different types of fruits and vegetables sits on a table next to a glass of ice water\"."}], "task": "refering", "answer_template": "The \"a page in a magazine showing different types of fruits and vegetables sits on a table next to a glass of ice water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 255, 256, 257, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.002984375, 0.555188679245283, 0.196953125, 0.9898584905660378], "iscrowd": 0, "area": 16026.747949999999, "rle": {"size": [424, 640], "counts": "^V1T1R91VHd0X7MiH3S70nH0n64QIMk66VIJf69[IGb6<]IE_6>bIBZ6a0gI_OU6d0kI]OV6b0jI^OV6a0kI_OU6`0lI@T6`0kIAU6>lIBU6<lIDT6<kIEU6;kIEU6;kIEV6:iIGW69iIGW69iIGW69hIHX68hIHY67gIIY67gIIY67fIJZ66fIJZ66fIJ[65dIL\\64dIL\\64dIL\\64cIM]63cIM^62bIN^62bIN^62aIO_61aIO`60`I0`60_I1a6O_I1a6O_I1a6O^I2c6M]I4b6L^I4b6L^I4b6L]I5c6K\\I6e6IZI8f6HYI9g6GXI:h6FXI:h6FWI;j6ETI<l6ERI<n6EPI<P7EnH<S7DkH=U7DiH>V7ChH>X7i2001O000000001O000000001O0000001O000000001O000000001O0M3L4L4L5K4L4K5L5K4L4L4L4L5K4L4L4L4L5K4L4L4K5L5K4L4L4L5K4L4L4L4L5K4L4L9G\\bd6"}, "label": "a page in a magazine showing different types of fruits and vegetables sits on a table next to a glass of ice water"}]}
{"id": 458057, "image": "COCO_train2014_000000458057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with pizza , drinks , and seasonings on it\"."}], "task": "refering", "answer_template": "The \"a table with pizza , drinks , and seasonings on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.087703125, 0.3976525821596244, 0.99765625, 1.0], "iscrowd": 0, "area": 72364.39894999999, "rle": {"size": [426, 640], "counts": "f_g02U=4L4L4L4M3O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O2N1N2O1O1O1N2O1O1O1N100iIUNo0k1oNWNQ1i1lNZNS1g1jN\\NV1d1QKTN`2:_2b1QKXNY29f2_1QK[NS29k2\\1SK^Nm18P3Z1RKbNh17V3W1RKeNc16[3U1RKhN]16`3S1SKjNW16f3P1RKnNS14k3n0RKQOm04Q4k0RKUOg02V4j0SKVOb03[4g0RKVOa06]4c0SKWO>8_4a0SKVO<<`4?TKUO:>b4=SKUO9a0d4:SKUO8b0e49SKTO7e0e48TKRO7g0e47SKRO7h0g45SKSO5i0h41VKUO2k0g4NZKVONn0i4IZKZOLn0j4G[KZOJQ1l4C[K[OIS1l4A\\K\\OGT1n4_OZK]OHU1n4]O[K]OHV1m4\\O\\K^OFW1o4YO\\K_OFX1n4XO\\K@FY1o4UO\\KBEY1o4TO]KBD[1P5QO]KCD\\1o4PO^KDB]1P5nN_KDB^1P5lN_KEA`1P5kN_KEA`1Q5jN^KEAb1Q5iN]KECb1Q5hN\\KFBc1R5gN\\KECd1Q5gN\\KDDa1U5jNWKEC_1Y5mNSKCE]1\\5oNoJDD[1`5QOlJCEY1b5TOiJCDW1g5UOdJDFS1j5YOeIQO2c0>Q1o5ZO`ISO3b0>n0R6^OZITO6??l0U6ATIUO8>?i0X6FnHTO;<?h0[6IhHVO=9`0e0_6MaHVO`08`0b0b62ZHVOd05`0f0=kN\\4T1QJWOf04`0g09lNdNOd5d1nJiNb0j04lNfNOg5a1mJjNa0m01jNhN1i5^1nJiN`0P1MTOe4S1nJhNa0R1IVOh4o0nJiNa0T1EVOl4n0mJgNb0X1AUOQ5k0mJgNa0[1\\OVOV5h0mJgNa0\\1YOWOZ5e0lJhN`0_1VOVO^5c0mJhN>a1SOSOe5e0iJhN>b1POPOm5e0eJjN;d1POmNR6e0dJiN9f2W5`N`JkN6d2]5aN]JlN4b2a5cN[JkN2`2g5dNWJmNO^2m5eNTJQOJY2U6eNRJUODT2]6gNoIXO_OP2f6gNkI]OZOk1m6iNiIZ2V6fMjI[2V6dMjI]2U6cMlI]2T6bMlI_2S6bMmI]2T6bMlI_2S6aMmI`2R6`MoI`2Q6_MoIb2Q6^MnIc2Q6]MPJc2P6]MoId2P6]MoId2Q6[MoIf2Q6ZMnIg2Q6ZMnIf2S6YMmIh2R6YMnIg2R6YMmIh2R6XMnIi2R6WMmIj2S6VMlIk2S6UMmIl2S6TMlIm2S6TMmIl2S6SMlIo2T6QMkIo2U6RMiIP3W6oLiIR3V6oLhIS3X6lLgIV3Y6jLbI[3]6fL_I^3a6aL[Id3d6a11O1O001O001O1O001O001O1O001O001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O001O1O001O1O001O1O1O001O1O002N2N1O2N1O1O1O001O00000000001O00000000001O0000001O000000000000001O00001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O000000O10000000000000000001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O_1"}, "label": "a table with pizza , drinks , and seasonings on it"}]}
{"id": 458057, "image": "COCO_train2014_000000458057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table of food , with plates , a pizza , pitchers , and glasses\"."}], "task": "refering", "answer_template": "The \"a table of food , with plates , a pizza , pitchers , and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 194, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.087703125, 0.3976525821596244, 0.99765625, 1.0], "iscrowd": 0, "area": 72364.39894999999, "rle": {"size": [426, 640], "counts": "f_g02U=4L4L4L4M3O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O2N1N2O1O1O1N2O1O1O1N100iIUNo0k1oNWNQ1i1lNZNS1g1jN\\NV1d1QKTN`2:_2b1QKXNY29f2_1QK[NS29k2\\1SK^Nm18P3Z1RKbNh17V3W1RKeNc16[3U1RKhN]16`3S1SKjNW16f3P1RKnNS14k3n0RKQOm04Q4k0RKUOg02V4j0SKVOb03[4g0RKVOa06]4c0SKWO>8_4a0SKVO<<`4?TKUO:>b4=SKUO9a0d4:SKUO8b0e49SKTO7e0e48TKRO7g0e47SKRO7h0g45SKSO5i0h41VKUO2k0g4NZKVONn0i4IZKZOLn0j4G[KZOJQ1l4C[K[OIS1l4A\\K\\OGT1n4_OZK]OHU1n4]O[K]OHV1m4\\O\\K^OFW1o4YO\\K_OFX1n4XO\\K@FY1o4UO\\KBEY1o4TO]KBD[1P5QO]KCD\\1o4PO^KDB]1P5nN_KDB^1P5lN_KEA`1P5kN_KEA`1Q5jN^KEAb1Q5iN]KECb1Q5hN\\KFBc1R5gN\\KECd1Q5gN\\KDDa1U5jNWKEC_1Y5mNSKCE]1\\5oNoJDD[1`5QOlJCEY1b5TOiJCDW1g5UOdJDFS1j5YOeIQO2c0>Q1o5ZO`ISO3b0>n0R6^OZITO6??l0U6ATIUO8>?i0X6FnHTO;<?h0[6IhHVO=9`0e0_6MaHVO`08`0b0b62ZHVOd05`0f0=kN\\4T1QJWOf04`0g09lNdNOd5d1nJiNb0j04lNfNOg5a1mJjNa0m01jNhN1i5^1nJiN`0P1MTOe4S1nJhNa0R1IVOh4o0nJiNa0T1EVOl4n0mJgNb0X1AUOQ5k0mJgNa0[1\\OVOV5h0mJgNa0\\1YOWOZ5e0lJhN`0_1VOVO^5c0mJhN>a1SOSOe5e0iJhN>b1POPOm5e0eJjN;d1POmNR6e0dJiN9f2W5`N`JkN6d2]5aN]JlN4b2a5cN[JkN2`2g5dNWJmNO^2m5eNTJQOJY2U6eNRJUODT2]6gNoIXO_OP2f6gNkI]OZOk1m6iNiIZ2V6fMjI[2V6dMjI]2U6cMlI]2T6bMlI_2S6bMmI]2T6bMlI_2S6aMmI`2R6`MoI`2Q6_MoIb2Q6^MnIc2Q6]MPJc2P6]MoId2P6]MoId2Q6[MoIf2Q6ZMnIg2Q6ZMnIf2S6YMmIh2R6YMnIg2R6YMmIh2R6XMnIi2R6WMmIj2S6VMlIk2S6UMmIl2S6TMlIm2S6TMmIl2S6SMlIo2T6QMkIo2U6RMiIP3W6oLiIR3V6oLhIS3X6lLgIV3Y6jLbI[3]6fL_I^3a6aL[Id3d6a11O1O001O001O1O001O001O1O001O001O00001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O001O1O001O1O001O1O1O001O1O002N2N1O2N1O1O1O001O00000000001O00000000001O0000001O000000000000001O00001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O000000O10000000000000000001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O_1"}, "label": "a table of food , with plates , a pizza , pitchers , and glasses"}]}
{"id": 253260, "image": "COCO_train2014_000000253260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver sedan , parked in a driveway\"."}], "task": "refering", "answer_template": "The \"a silver sedan , parked in a driveway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86], "bbox": [0.463953125, 0.03617728531855956, 0.9186093750000002, 0.2998060941828255], "iscrowd": 0, "area": 15677.965749999998, "rle": {"size": [361, 640], "counts": "mhX3<e:9G8K6]Ob00010O010O010O00010O010O000010O01O001O00000001O00000000O100O100O10000O100O100O10000O100O10000O100O100O10O10O100O10000O100O100O10000O100O101O0O100O100O101O0O1O100O2O1N2N2O1N2O1N2O1N100000000000O100000000O100O1O1O1O1O1O100O1O1O100000000000O1000000000O1000000000000000000000000000000000000000000000000000O1000000000000000000000O10000000001O001O001O001O001O00001O0DUGZNl8d1VG[Nj8d1XG[Ni8c1YG\\Ng8d1ZG[Ng8c1[G\\Ne8e1[GZNf8e1[GZNe8f1\\GYNe8g1>01O0SO`F2`9MdFO\\91hFKX95mFGS98QGDo8<TGAm8>TG@m8?TGAm8>TG@m8`0SG_Oo8`0RG_On8a0RG^OP9a0QG]OP9c0l0N101LZECf:6bc4`0W\\K100O2O0O1O101N10000O10O1O001O001O010O1O1O1O1O1O1O1O1O1O2N1O1O1O100O1O100O1O100O2N1O100O1O100O1Ohf02UYO3O2O1N2N10N1O2N1O2N1OZdb0"}, "label": "a silver sedan , parked in a driveway"}]}
{"id": 253260, "image": "COCO_train2014_000000253260.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver car with a spoiler on the back\"."}], "task": "refering", "answer_template": "The \"a silver car with a spoiler on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86], "bbox": [0.463953125, 0.03617728531855956, 0.9186093750000002, 0.2998060941828255], "iscrowd": 0, "area": 15677.965749999998, "rle": {"size": [361, 640], "counts": "mhX3<e:9G8K6]Ob00010O010O010O00010O010O000010O01O001O00000001O00000000O100O100O10000O100O100O10000O100O10000O100O100O10O10O100O10000O100O100O10000O100O101O0O100O100O101O0O1O100O2O1N2N2O1N2O1N2O1N100000000000O100000000O100O1O1O1O1O1O100O1O1O100000000000O1000000000O1000000000000000000000000000000000000000000000000000O1000000000000000000000O10000000001O001O001O001O001O00001O0DUGZNl8d1VG[Nj8d1XG[Ni8c1YG\\Ng8d1ZG[Ng8c1[G\\Ne8e1[GZNf8e1[GZNe8f1\\GYNe8g1>01O0SO`F2`9MdFO\\91hFKX95mFGS98QGDo8<TGAm8>TG@m8?TGAm8>TG@m8`0SG_Oo8`0RG_On8a0RG^OP9a0QG]OP9c0l0N101LZECf:6bc4`0W\\K100O2O0O1O101N10000O10O1O001O001O010O1O1O1O1O1O1O1O1O1O2N1O1O1O100O1O100O1O100O2N1O100O1O100O1Ohf02UYO3O2O1N2N10N1O2N1O2N1OZdb0"}, "label": "a silver car with a spoiler on the back"}]}
{"id": 351566, "image": "COCO_train2014_000000351566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat with its head turned away from the camera\"."}], "task": "refering", "answer_template": "The \"a cat with its head turned away from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 20, 21, 23, 24, 25, 26, 27, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 71, 72, 73, 88, 89], "bbox": [0.17027083333333334, 0.03646875, 0.6607083333333333, 0.24060937500000001], "iscrowd": 0, "area": 12902.494299999998, "rle": {"size": [640, 480], "counts": "b\\c11nc02N2O1N2N1O1O1O1O2N1O1O100O1O2N1000000O1YM[OSBf0l=\\ORBd0n=]OQBc0P>^OmAb0T>@jA?W>BhA=Z>DdA;]>G`A9a>I]A6e>JZA5g>MWA2j>0PA3Q?Og@6[?J_@:b?HW@<j?Fn_O`0S`0@h_Od0X`0^Oa_Oh0^`0ZO[_Ok0f`0VOT_On0l`0SOn^OS1Qa0n01O00010O00001O000O2O0000001O0O101O000O1O1O1O1N2O1O1O1O1O1O1O1O1O1O001O1O1K4O2M2O2M2O2M2O2N1N3N1N3N1N3N1100O100O2O0O11O010O1O1O10O01O00O2N1O1O1O2M2O1O1O2N1O1N2O2N1N2O1O2M2O1O1N3N1001O01O0000001O00000001O000000000000000000001O00000001O0001O000000001O00000000001O000000001O01O00000001O00000000000O100O100O1O100O1O100O100O1O100O1O100O100O1O2O2M2O2M2N3N1N3M2O2M2O1N3M2O2M2O2M2N3N1N3M3N2M3NliU3"}, "label": "a cat with its head turned away from the camera"}]}
{"id": 351566, "image": "COCO_train2014_000000351566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat in the background sleeping near the edge of a bed\"."}], "task": "refering", "answer_template": "The \"cat in the background sleeping near the edge of a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 20, 21, 23, 24, 25, 26, 27, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 71, 72, 73, 88, 89], "bbox": [0.17027083333333334, 0.03646875, 0.6607083333333333, 0.24060937500000001], "iscrowd": 0, "area": 12902.494299999998, "rle": {"size": [640, 480], "counts": "b\\c11nc02N2O1N2N1O1O1O1O2N1O1O100O1O2N1000000O1YM[OSBf0l=\\ORBd0n=]OQBc0P>^OmAb0T>@jA?W>BhA=Z>DdA;]>G`A9a>I]A6e>JZA5g>MWA2j>0PA3Q?Og@6[?J_@:b?HW@<j?Fn_O`0S`0@h_Od0X`0^Oa_Oh0^`0ZO[_Ok0f`0VOT_On0l`0SOn^OS1Qa0n01O00010O00001O000O2O0000001O0O101O000O1O1O1O1N2O1O1O1O1O1O1O1O1O1O001O1O1K4O2M2O2M2O2M2O2N1N3N1N3N1N3N1100O100O2O0O11O010O1O1O10O01O00O2N1O1O1O2M2O1O1O2N1O1N2O2N1N2O1O2M2O1O1N3N1001O01O0000001O00000001O000000000000000000001O00000001O0001O000000001O00000000001O000000001O01O00000001O00000000000O100O100O1O100O1O100O100O1O100O1O100O100O1O2O2M2O2M2N3N1N3M2O2M2O1N3M2O2M2O2M2N3N1N3M3N2M3NliU3"}, "label": "cat in the background sleeping near the edge of a bed"}]}
{"id": 351566, "image": "COCO_train2014_000000351566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black cat furtherest away from the wardrobe\"."}], "task": "refering", "answer_template": "The \"the black cat furtherest away from the wardrobe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 57, 64, 65, 74, 75, 76, 77, 78, 79, 80, 81, 82, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344, 345, 346, 357, 358, 359, 360, 361, 362, 363, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.0861875, 0.8713333333333334, 1.0], "iscrowd": 0, "area": 123301.18690000002, "rle": {"size": [640, 480], "counts": "\\?n2\\`0n0[O=D=G8L5M2M4M3M3L3N3L4M2N3L4M2N3L4M2N3L4M2N3M3M2N3M3M2N3M3M2M4K4L4M4K4M3L5K4M3L5L3K6J5K5M4M2N2N3N3L3N2M4M2M4M2M3N3L3N2M4M2M3N2N1O1N2O1O1O1N2O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O100O100O100O100O100O100O100O100O1O1O1O1O1N2O1O1O1O1N2O1O1N2O1N200O1O1O1O1O1O100O1O1O100O1O1O1O100O1O1O1O010O00O1O2O0O10dMbBWN]=h1jBSNU=l1RCPNm<o1ZCkMe<V2`CeM_<\\2fC_MZ<a2kC[MS<f2PDWMo;i2UDTMk;m2WDPM[;_3gD^LX;d3kDYLS;j3nDSLQ;o3QEnKn:T4TEiKk:Z4VEdKh:^4ZE_Ke:c4]EZK`:j4cERKZ:S5iEiJT:Z5U25K4L4M4N1N2O2N1N2O2N1N3N1O2M3M2O2M3M3M2N3M3N2M2N3M3M3N2M2N3M3L4L3N3L4L4L4L3M4M3L3M3M2N3M2N3M2O1N3M2N3M2N2N3M2N3N1N3M2N3M2N3N2M2O2M3N1N3N1N3N2M2O2M3N1O2M2O2N2M3N2M3N2N2M3N2M3L4M3L4M3L4M2M4M3L4L4M3M3O10000O100O1O1N2O1N3N1N3N100010O10O01O10O010O00001O001O00001O001O00000O10O10000O0100000O01000O1O010O100OK5010O010O01O000010O01O0000001O0000000000001N001N101O001O001O001O001O001O0010O010O010O010O0100O10O0100O100O100O1O1L4M3M3L6K4K6K5J5J7I6J7J5J7iNV_OPOQa0f0__OiNk`0n0[1YOcfV1"}, "label": "the black cat furtherest away from the wardrobe"}]}
{"id": 351566, "image": "COCO_train2014_000000351566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat looking toward the camera\"."}], "task": "refering", "answer_template": "The \"a black and white cat looking toward the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 57, 64, 65, 74, 75, 76, 77, 78, 79, 80, 81, 82, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344, 345, 346, 357, 358, 359, 360, 361, 362, 363, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.0861875, 0.8713333333333334, 1.0], "iscrowd": 0, "area": 123301.18690000002, "rle": {"size": [640, 480], "counts": "\\?n2\\`0n0[O=D=G8L5M2M4M3M3L3N3L4M2N3L4M2N3L4M2N3L4M2N3M3M2N3M3M2N3M3M2M4K4L4M4K4M3L5K4M3L5L3K6J5K5M4M2N2N3N3L3N2M4M2M4M2M3N3L3N2M4M2M3N2N1O1N2O1O1O1N2O1O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O100O100O100O100O100O100O100O100O1O1O1O1O1N2O1O1O1O1N2O1O1N2O1N200O1O1O1O1O1O100O1O1O100O1O1O1O100O1O1O1O010O00O1O2O0O10dMbBWN]=h1jBSNU=l1RCPNm<o1ZCkMe<V2`CeM_<\\2fC_MZ<a2kC[MS<f2PDWMo;i2UDTMk;m2WDPM[;_3gD^LX;d3kDYLS;j3nDSLQ;o3QEnKn:T4TEiKk:Z4VEdKh:^4ZE_Ke:c4]EZK`:j4cERKZ:S5iEiJT:Z5U25K4L4M4N1N2O2N1N2O2N1N3N1O2M3M2O2M3M3M2N3M3N2M2N3M3M3N2M2N3M3L4L3N3L4L4L4L3M4M3L3M3M2N3M2N3M2O1N3M2N3M2N2N3M2N3N1N3M2N3M2N3N2M2O2M3N1N3N1N3N2M2O2M3N1O2M2O2N2M3N2M3N2N2M3N2M3L4M3L4M3L4M2M4M3L4L4M3M3O10000O100O1O1N2O1N3N1N3N100010O10O01O10O010O00001O001O00001O001O00000O10O10000O0100000O01000O1O010O100OK5010O010O01O000010O01O0000001O0000000000001N001N101O001O001O001O001O001O0010O010O010O010O0100O10O0100O100O100O1O1L4M3M3L6K4K6K5J5J7I6J7J5J7iNV_OPOQa0f0__OiNk`0n0[1YOcfV1"}, "label": "a black and white cat looking toward the camera"}]}
{"id": 548184, "image": "COCO_train2014_000000548184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green plant is beside the girl\"."}], "task": "refering", "answer_template": "The \"the green plant is beside the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 238, 239], "bbox": [0.23595312499999999, 0.3591666666666667, 0.46854687500000003, 0.6400833333333333], "iscrowd": 0, "area": 11571.132400000002, "rle": {"size": [480, 640], "counts": "kPW21m>3M2M3N3M2L4L5L3M4L3M3M4N1N2N3M2O2M20001O001O00100O1O001O2N8H8H8H7H10001N101O0O2O000010O010O00010O010O010O00O2N1O1O1N2G900O100000O010000000000000000000000000O101O0000001N1M3N3L3M4L3O101N101N102M4L3N3L100O2O0B^C]Nc<d1bCVNo;2RDj1V<ZNiCf1S<_NkCb1S<bNjC_1U<g0O9G1O1O1O1OO@XMjDf2Y;ZMfDe2[;\\MdDc2];]McD_15lN\\;F^DZ1`0fNV;1YDW1g0`NT;:TDT1l<?1010O2N2O1NO101N100O2N100O2O1N3N4K8TO`BJ`=4eBH\\=6iBGX=6nBES=9RCDo<9VCCk<3TWo4"}, "label": "the green plant is beside the girl"}]}
{"id": 548184, "image": "COCO_train2014_000000548184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bush in a black pot\"."}], "task": "refering", "answer_template": "The \"a green bush in a black pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 238, 239], "bbox": [0.23595312499999999, 0.3591666666666667, 0.46854687500000003, 0.6400833333333333], "iscrowd": 0, "area": 11571.132400000002, "rle": {"size": [480, 640], "counts": "kPW21m>3M2M3N3M2L4L5L3M4L3M3M4N1N2N3M2O2M20001O001O00100O1O001O2N8H8H8H7H10001N101O0O2O000010O010O00010O010O010O00O2N1O1O1N2G900O100000O010000000000000000000000000O101O0000001N1M3N3L3M4L3O101N101N102M4L3N3L100O2O0B^C]Nc<d1bCVNo;2RDj1V<ZNiCf1S<_NkCb1S<bNjC_1U<g0O9G1O1O1O1OO@XMjDf2Y;ZMfDe2[;\\MdDc2];]McD_15lN\\;F^DZ1`0fNV;1YDW1g0`NT;:TDT1l<?1010O2N2O1NO101N100O2N100O2O1N3N4K8TO`BJ`=4eBH\\=6iBGX=6nBES=9RCDo<9VCCk<3TWo4"}, "label": "a green bush in a black pot"}]}
{"id": 548184, "image": "COCO_train2014_000000548184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red suit case in the hatchback of the blue car with other clothes\"."}], "task": "refering", "answer_template": "The \"a red suit case in the hatchback of the blue car with other clothes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 366, 367], "bbox": [0.766859375, 0.6400833333333333, 1.0, 0.90075], "iscrowd": 0, "area": 14828.376749999996, "rle": {"size": [480, 640], "counts": "`_V79e><D=C<D;E7J6I6J2N201O0000001O0000001O0000001O0000001O00000O2N1O1O1O2N1O1O1O2N100O101N10000010O000001O00000010O000001O0000001O01O0001O0000001O0001O01O00001O00000010O000001O00000010O000001O0000001O01O01O0000001O0000010O0000001O0000010O0000001O00001O01O000001O0000000001O01O000000000000010O000000000QF"}, "label": "a red suit case in the hatchback of the blue car with other clothes"}]}
{"id": 548184, "image": "COCO_train2014_000000548184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red suitcase in the trunk\"."}], "task": "refering", "answer_template": "The \"the red suitcase in the trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 366, 367], "bbox": [0.766859375, 0.6400833333333333, 1.0, 0.90075], "iscrowd": 0, "area": 14828.376749999996, "rle": {"size": [480, 640], "counts": "`_V79e><D=C<D;E7J6I6J2N201O0000001O0000001O0000001O0000001O00000O2N1O1O1O2N1O1O1O2N100O101N10000010O000001O00000010O000001O0000001O01O0001O0000001O0001O01O00001O00000010O000001O00000010O000001O0000001O01O01O0000001O0000010O0000001O0000010O0000001O00001O01O000001O0000000001O01O000000000000010O000000000QF"}, "label": "the red suitcase in the trunk"}]}
{"id": 474461, "image": "COCO_train2014_000000474461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back half of a red and white fire truck\"."}], "task": "refering", "answer_template": "The \"the back half of a red and white fire truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 144, 145], "bbox": [0.37212418300653594, 0.11145424836601306, 0.6981372549019608, 0.2957679738562091], "iscrowd": 0, "area": 16961.4449, "rle": {"size": [612, 612], "counts": "Q[X49ab0<K6N1N3N1N2N2N1O1O1O1O1O001O1m^OiNW`0X1h_OlNT`0U1k_OoNQ`0R1m_OTOn?m0Q@WOk?j0S@\\Oh?e0W@^Of?c0Y@]Og?d0W@]Oi?d0V@[Ok?f0T@ZOl?h0Q@YOo?i0o_OWOQ`0l0l_OTOT`0n1000000000O10000000000000000O1000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O2N1O1O2N1O2N2N2N1O2N1OO10000O10000O10000O10000O10000O10000O10000O10000O101O0O100O1N2O1O1O1O1N2O1O1O1N2000000000000000000000000001OO1I7I7N2N2N2N2O1N2N2N2N2N2N2O1N2N2L4K5L>Ah_^3"}, "label": "the back half of a red and white fire truck"}]}
{"id": 474461, "image": "COCO_train2014_000000474461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is red bus back of a women in pink t - shirt\"."}], "task": "refering", "answer_template": "The \"there is red bus back of a women in pink t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 312, 313, 314, 315, 316, 317, 318, 332, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 379, 380, 381, 382, 383, 384, 400, 401, 402, 403, 404, 405, 406, 422, 423, 424, 425, 426, 427, 428, 444, 445, 446, 447, 448, 449, 450, 467, 468, 469, 470, 471, 472], "bbox": [0.08315359477124183, 0.5146078431372549, 0.4853921568627451, 0.9865196078431373], "iscrowd": 0, "area": 50378.02870000002, "rle": {"size": [612, 612], "counts": "okn0=Vb0a0_Ob0^Oa0^Oc0@?01O00001O001O001O001O001O001O001O00_Na1O100O1O100O100O100O10O0100O1000000000000000000000000000000001O2N1O2N1O2N1O2N2N2N3M4L3M4L4L2N2N2N2N2iDVMP6l2hIZMX6i2^I^Mb6d2VIcMi6_2oHgMQ7Z2gHmMY7U2_HQNa7Q2VHWNi7k1oG[NQ8f1hG`NX8b1_GeNa8]1[GfNd8\\1ZGdNf8]1XGdNh8^1VGbNj8`1SGbNl8_1SGaNm8a1PG`NP9a1oF_NQ9c1lF_NS9b1lF^NT9d1iF]NW9d1hF\\NW9g1gFZNX9g1fFZNZ9h1dFXN\\9i1bFXN^9j1`FWN_9k1^FVNb9k1]FUNc9l1[FVNd9k1[FUNe9l1YFUNg9l1XFTNh9m1WFTNh9m1VFTNj9m1UFSNk9n1SFSNm9n1RFSNm9n1QFSNo9n1PFRNP:P2mEQNS:P2lEQNS:P2kEQNU:P2cEWN]:X501O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0O10001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000O2O00000000000000000000001O00000000000000000000001O0000K5aN_1aNPPl5"}, "label": "there is red bus back of a women in pink t - shirt"}]}
{"id": 556383, "image": "COCO_train2014_000000556383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with longish black hair sits behind a black laptop\"."}], "task": "refering", "answer_template": "The \"a person with longish black hair sits behind a black laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 106, 107, 108, 129, 130, 131, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 201, 202, 203, 224, 225, 226, 247], "bbox": [0.607796875, 0.2975833333333333, 0.851515625, 0.831361111111111], "iscrowd": 0, "area": 11369.320100000004, "rle": {"size": [360, 640], "counts": "clX45S;8G9H9G4K100O1000O0100N2N2N2N2RGeNm7]1RHdNl7]1SHeNk7]1THdNj7^1UHcNi7_1VHcNg7_1XHbNe7a1YHaNf7b1WH_Nh7c1VH^Nj7c1TH_Nk7c1QH_No7b1oG_NQ8b1mG_NR8d1kG^NT8c1jG^NV8Z2000000O2O00000000001N100001O1O00100O001O1O001O10OaMlGj1U8RNSHi1n7SNYHi1h7SN_Hi1a7TNfHh1`8M3N3L3M5K4L5K0001O0000000000001O000001O01O000PH]O[5d0SIl0n6TOlHR1T7nNjHS1W7nNfHS1[7mNcHT1^7mN_HU1a7kN]HV1d7jNZHW1g7U10O1O2O0O1O1O100O1O100O1O100O1O2N100O1O101N1O100O1O1O010O1O100O1O100O001O100O1O100O1O011N1O1O2O0O2N1N2M4L3N3L7I6J6J6J7I6K5J7I6JSVQ1"}, "label": "a person with longish black hair sits behind a black laptop"}]}
{"id": 556383, "image": "COCO_train2014_000000556383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a black computer\"."}], "task": "refering", "answer_template": "The \"a man with a black computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 106, 107, 108, 129, 130, 131, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 201, 202, 203, 224, 225, 226, 247], "bbox": [0.607796875, 0.2975833333333333, 0.851515625, 0.831361111111111], "iscrowd": 0, "area": 11369.320100000004, "rle": {"size": [360, 640], "counts": "clX45S;8G9H9G4K100O1000O0100N2N2N2N2RGeNm7]1RHdNl7]1SHeNk7]1THdNj7^1UHcNi7_1VHcNg7_1XHbNe7a1YHaNf7b1WH_Nh7c1VH^Nj7c1TH_Nk7c1QH_No7b1oG_NQ8b1mG_NR8d1kG^NT8c1jG^NV8Z2000000O2O00000000001N100001O1O00100O001O1O001O10OaMlGj1U8RNSHi1n7SNYHi1h7SN_Hi1a7TNfHh1`8M3N3L3M5K4L5K0001O0000000000001O000001O01O000PH]O[5d0SIl0n6TOlHR1T7nNjHS1W7nNfHS1[7mNcHT1^7mN_HU1a7kN]HV1d7jNZHW1g7U10O1O2O0O1O1O100O1O100O1O100O1O2N100O1O101N1O100O1O1O010O1O100O1O100O001O100O1O100O1O011N1O1O2O0O2N1N2M4L3N3L7I6J6J6J7I6K5J7I6JSVQ1"}, "label": "a man with a black computer"}]}
{"id": 556383, "image": "COCO_train2014_000000556383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white clothes using a laptop on a couch\"."}], "task": "refering", "answer_template": "The \"a woman in white clothes using a laptop on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 102, 103, 104, 125, 126, 147, 148, 149, 151, 152, 171, 172, 173, 174, 175], "bbox": [0.418125, 0.28641666666666665, 0.653421875, 0.5831666666666666], "iscrowd": 0, "area": 5933.6365000000005, "rle": {"size": [360, 640], "counts": "n\\n27j:7J6M3M3M3M2O2O1O000O2O00000O101O000O02M2O21N8I6J7H4M0O100O10000ZOVNmGl1l7^NPHb1l7fNoG[1P8kNkGU1U8o0O100O1O101N100O1000000O10001O0000nNXHdNg7Y1`HdN`7X1eHgN[7T1jHlNW7o0nHPOR7d0VHhNm0d0m6c0[I^Od6a0^I^Oc6`0^I@b6?_IAa6<bIDa65dIJ`6NdI2`6FdI:`6^OeIa0a81O03M3M3M1O1O001O1O000000O1O100O1O00100O1O100O1O1O100O1O100O1O1O001O1O1O0O2O1O1O0O20000O100O10000O100O100000O10000000000O10O10O10000O2O0O2O0O2N1O2N2M2O2N2M3N2L4Lmk]2"}, "label": "a woman in white clothes using a laptop on a couch"}]}
{"id": 556383, "image": "COCO_train2014_000000556383.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a seated woman holding a grey laptop that has a heart sticker on it\"."}], "task": "refering", "answer_template": "The \"a seated woman holding a grey laptop that has a heart sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 102, 103, 104, 125, 126, 147, 148, 149, 151, 152, 171, 172, 173, 174, 175], "bbox": [0.418125, 0.28641666666666665, 0.653421875, 0.5831666666666666], "iscrowd": 0, "area": 5933.6365000000005, "rle": {"size": [360, 640], "counts": "n\\n27j:7J6M3M3M3M2O2O1O000O2O00000O101O000O02M2O21N8I6J7H4M0O100O10000ZOVNmGl1l7^NPHb1l7fNoG[1P8kNkGU1U8o0O100O1O101N100O1000000O10001O0000nNXHdNg7Y1`HdN`7X1eHgN[7T1jHlNW7o0nHPOR7d0VHhNm0d0m6c0[I^Od6a0^I^Oc6`0^I@b6?_IAa6<bIDa65dIJ`6NdI2`6FdI:`6^OeIa0a81O03M3M3M1O1O001O1O000000O1O100O1O00100O1O100O1O1O100O1O100O1O1O001O1O1O0O2O1O1O0O20000O100O10000O100O100000O10000000000O10O10O10000O2O0O2O0O2N1O2N2M2O2N2M3N2L4Lmk]2"}, "label": "a seated woman holding a grey laptop that has a heart sticker on it"}]}
{"id": 523617, "image": "COCO_train2014_000000523617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and red single propeller plane flying below a second plane further away\"."}], "task": "refering", "answer_template": "The \"a white and red single propeller plane flying below a second plane further away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 103, 104, 105, 116, 117, 118, 119, 120, 135, 136], "bbox": [0.25416289592760183, 0.49521212121212116, 0.6371040723981901, 0.7487878787878788], "iscrowd": 0, "area": 4499.924500000001, "rle": {"size": [330, 442], "counts": "dcT11Y:1N101N1O2O0O2O0O2N101N010POP11N1000000O10001N10WGjN^8W1aGiN_8W1aGiNa8U1_GkNe8Q1ZGQOi8l0VGTOn8h0RGXOR9d0mF^OV9>jFBZ9;eFE[9;dFF\\9:dFG[99eFG[9:cFG]99cFG]99cFG]9c00O10000O10000O101O1N10000O1000000000001O0001O001O001O00001O010O001O00001O000000010O00000O101N100O10001N100O10001VGSOT8n0kGUOS8k0kGWOT8j0iGYOW8g0fG\\OZ8e0bG^O]8X10DgGgNY8U1nGgNR8V1c0O0101O0O100O10000O101N2O001N2O1O6ICbGPO^8P1cGoN\\8Q1fGnNZ8>cGJ4GY8=eGK3HW8<hGL2GV8;jGM1A[8a0gGMc81_GOa8OaG0_8OZG:f8E]G8d8GfG0Y81jGKW84kG8g7IZHMo72THGQ89o0010O0100O0010O0100O010O10O02O0O100O100O100O100O101N10ZQd1"}, "label": "a white and red single propeller plane flying below a second plane further away"}]}
{"id": 523617, "image": "COCO_train2014_000000523617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest airplane\"."}], "task": "refering", "answer_template": "The \"the biggest airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 103, 104, 105, 116, 117, 118, 119, 120, 135, 136], "bbox": [0.25416289592760183, 0.49521212121212116, 0.6371040723981901, 0.7487878787878788], "iscrowd": 0, "area": 4499.924500000001, "rle": {"size": [330, 442], "counts": "dcT11Y:1N101N1O2O0O2O0O2N101N010POP11N1000000O10001N10WGjN^8W1aGiN_8W1aGiNa8U1_GkNe8Q1ZGQOi8l0VGTOn8h0RGXOR9d0mF^OV9>jFBZ9;eFE[9;dFF\\9:dFG[99eFG[9:cFG]99cFG]99cFG]9c00O10000O10000O101O1N10000O1000000000001O0001O001O001O00001O010O001O00001O000000010O00000O101N100O10001N100O10001VGSOT8n0kGUOS8k0kGWOT8j0iGYOW8g0fG\\OZ8e0bG^O]8X10DgGgNY8U1nGgNR8V1c0O0101O0O100O10000O101N2O001N2O1O6ICbGPO^8P1cGoN\\8Q1fGnNZ8>cGJ4GY8=eGK3HW8<hGL2GV8;jGM1A[8a0gGMc81_GOa8OaG0_8OZG:f8E]G8d8GfG0Y81jGKW84kG8g7IZHMo72THGQ89o0010O0100O0010O0100O010O10O02O0O100O100O100O100O101N10ZQd1"}, "label": "the biggest airplane"}]}
{"id": 277858, "image": "COCO_train2014_000000277858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing sunglasses and a black shirt eats a piece of pizza while holding a rubbermaid container\"."}], "task": "refering", "answer_template": "The \"a man wearing sunglasses and a black shirt eats a piece of pizza while holding a rubbermaid container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 43, 44, 45, 61, 62, 63, 79, 80, 81, 82, 97, 98, 99, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 157, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209], "bbox": [0.21102, 0.15954954954954956, 0.78274, 1.0], "iscrowd": 0, "area": 39040.69165, "rle": {"size": [333, 500], "counts": "ahR15S:6M3bFGd8;UGLi87nF1P9g0N2N2N2J6oN[NTIk1i6ZNoHl1o6XNiHn1U7UNeHP2X7d0N2M3N2N2M3N2M3N2N2O10O01O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O010O1O100O1O100O100O1O100O1YMYKHg45_KnM1h0a4V1dKPN0d0]4Y1iKPN06ZODm4U2nKmM05YODj4X2TLjMK7\\4n1oKfME<\\4l1QMTNo2k1RMTNo2j1SMVNm2i1TMVNm2h1UMXNk2g1n2N2O10000O100O10000O100O10000O100O100O10O10O100O10000001O0000000000000000001O000000000000001O001O1O1O001O1O001O1O1O002N1O2QKiMd1X2YNlMe1V2XNmMf1T2XNoMf1S2VNQNh1Q2UNPNk1Q2SNQNk1R2QNQNn1P2oMRNQ2P2iMUNV2n1cMWN\\2m1nKVMm0R1T3k1cK_M5Be0X1c3`2ZLdMe3\\2WLiMh3X2SLlMm3U2nKPNQ4Q2jKTNU4R401O0N3K4O2O0O2N1O2N1O2N1O2O0O2O001N2O1O1N3N1O1O2M2O3M4L3H9K5L4L4K5L3L8I9G3L3N2N1N3N1O1[OcF8_9@jF=e9N3M1O2N1O1N10N2O1N2N2O1N2O1N1N3M3L4M3M300O100O01000O11N2O1O1N2O1O1N2O1O1O1N10000O10O1O1N101\\O^O[Gd0c8@QGi0o8;O10O01O100O00POTGa0l8^OUGa0k8_OVGa0k8]OVG3MOo8LUG4ONn8KTG61HQ90PG4`9JbF2S`S1"}, "label": "a man wearing sunglasses and a black shirt eats a piece of pizza while holding a rubbermaid container"}]}
{"id": 277858, "image": "COCO_train2014_000000277858.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in sunglasses eating pizza\"."}], "task": "refering", "answer_template": "The \"a man in sunglasses eating pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 43, 44, 45, 61, 62, 63, 79, 80, 81, 82, 97, 98, 99, 114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 157, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209], "bbox": [0.21102, 0.15954954954954956, 0.78274, 1.0], "iscrowd": 0, "area": 39040.69165, "rle": {"size": [333, 500], "counts": "ahR15S:6M3bFGd8;UGLi87nF1P9g0N2N2N2J6oN[NTIk1i6ZNoHl1o6XNiHn1U7UNeHP2X7d0N2M3N2N2M3N2M3N2N2O10O01O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O010O1O100O1O100O100O1O100O1YMYKHg45_KnM1h0a4V1dKPN0d0]4Y1iKPN06ZODm4U2nKmM05YODj4X2TLjMK7\\4n1oKfME<\\4l1QMTNo2k1RMTNo2j1SMVNm2i1TMVNm2h1UMXNk2g1n2N2O10000O100O10000O100O10000O100O100O10O10O100O10000001O0000000000000000001O000000000000001O001O1O1O001O1O001O1O1O002N1O2QKiMd1X2YNlMe1V2XNmMf1T2XNoMf1S2VNQNh1Q2UNPNk1Q2SNQNk1R2QNQNn1P2oMRNQ2P2iMUNV2n1cMWN\\2m1nKVMm0R1T3k1cK_M5Be0X1c3`2ZLdMe3\\2WLiMh3X2SLlMm3U2nKPNQ4Q2jKTNU4R401O0N3K4O2O0O2N1O2N1O2N1O2O0O2O001N2O1O1N3N1O1O2M2O3M4L3H9K5L4L4K5L3L8I9G3L3N2N1N3N1O1[OcF8_9@jF=e9N3M1O2N1O1N10N2O1N2N2O1N2O1N1N3M3L4M3M300O100O01000O11N2O1O1N2O1O1N2O1O1O1N10000O10O1O1N101\\O^O[Gd0c8@QGi0o8;O10O01O100O00POTGa0l8^OUGa0k8_OVGa0k8]OVG3MOo8LUG4ONn8KTG61HQ90PG4`9JbF2S`S1"}, "label": "a man in sunglasses eating pizza"}]}
{"id": 400740, "image": "COCO_train2014_000000400740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bus\"."}], "task": "refering", "answer_template": "The \"a white bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 255, 258], "bbox": [0.047984375, 0.37327083333333333, 0.351953125, 0.6545], "iscrowd": 0, "area": 22467.7486, "rle": {"size": [480, 640], "counts": "ch>j1\\<j0M4L3M3N2M3M4O0O100O100O2O0O100O100O2O0O100O2O1N2O1O1O1O1O1N10001O0000O100O1000O01N2N2N200000000000O100000000000000000000O100000001O000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O10002N2N2N1O000000000O1000000000O100000000N2N2N2M3001O001O00001O000O2O0000000000000000000000000000000000000000000000^NPETOP;j0SEUOm:i0VEVOj:h0YEWOf:h0]EVOd:h0_EWOa:h0aEWO_:h0cEWO]:i0eETO\\:l0eESO[:m0fEROZ:m0hEQOZ:n0gEQO^:j0cEUOb:f0_EYOf:a0\\E]Oj:>XE@m:;YE[OP;`0VEVOS;d0j1I7H8IlYR6"}, "label": "a white bus"}]}
{"id": 400740, "image": "COCO_train2014_000000400740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bus\"."}], "task": "refering", "answer_template": "The \"the white bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 144, 145, 162, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 255, 258], "bbox": [0.047984375, 0.37327083333333333, 0.351953125, 0.6545], "iscrowd": 0, "area": 22467.7486, "rle": {"size": [480, 640], "counts": "ch>j1\\<j0M4L3M3N2M3M4O0O100O100O2O0O100O100O2O0O100O2O1N2O1O1O1O1O1N10001O0000O100O1000O01N2N2N200000000000O100000000000000000000O100000001O000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O10002N2N2N1O000000000O1000000000O100000000N2N2N2M3001O001O00001O000O2O0000000000000000000000000000000000000000000000^NPETOP;j0SEUOm:i0VEVOj:h0YEWOf:h0]EVOd:h0_EWOa:h0aEWO_:h0cEWO]:i0eETO\\:l0eESO[:m0fEROZ:m0hEQOZ:n0gEQO^:j0cEUOb:f0_EYOf:a0\\E]Oj:>XE@m:;YE[OP;`0VEVOS;d0j1I7H8IlYR6"}, "label": "the white bus"}]}
{"id": 400740, "image": "COCO_train2014_000000400740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red tour bus in the city\"."}], "task": "refering", "answer_template": "The \"a red tour bus in the city\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 270, 271], "bbox": [0.40112500000000006, 0.26066666666666666, 0.9033750000000002, 0.6696666666666666], "iscrowd": 0, "area": 51028.30855, "rle": {"size": [480, 640], "counts": "Yeh3e0V>m0SOm0SOm0SOa0_O5N2N2N2N2N2N2O100O100O010O100O100O1000000000000O0100000000000O100000000000O10O10001O2N2N2N2M101O1O001O1O0O2O0000000000000O1000000000000O01O1O1O1O1O1O0100000O10000000O1000000000000O1000000000000O10000000000000O01000000000000O1000000000000O100000000000000O100000O100000O1000000000000O100000000000000O100000000000O01000000000000O100000000000000O1000000000000O100000O100000O100000000000000O1000000000000O1000000000000O010000000000000O2O001O1O001N1O2N2M2O2O1O001O1O001O1O001O1O1O001O1O02N2N00000000000000000000O1000O10000O1O1O1O1O1O001O1O1O1O2N1O1O1O2N11O10O01O1O010O00100O010O010O01O001oLeGZO\\8f0eGYO\\8f0dGZO\\8e0fGZO[8e0eG[O`8`0aG_Oe8;[GEk85VGJP90PG0V9JkF5[9DfF;a9_O`F>g9]OYFa0o9YORFd0V:VOjEi0]:QOdEl0d:nN^En0j:lNXEP1_<M3L3M4L4M3K5I_il0"}, "label": "a red tour bus in the city"}]}
{"id": 400740, "image": "COCO_train2014_000000400740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange bus\"."}], "task": "refering", "answer_template": "The \"an orange bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 270, 271], "bbox": [0.40112500000000006, 0.26066666666666666, 0.9033750000000002, 0.6696666666666666], "iscrowd": 0, "area": 51028.30855, "rle": {"size": [480, 640], "counts": "Yeh3e0V>m0SOm0SOm0SOa0_O5N2N2N2N2N2N2O100O100O010O100O100O1000000000000O0100000000000O100000000000O10O10001O2N2N2N2M101O1O001O1O0O2O0000000000000O1000000000000O01O1O1O1O1O1O0100000O10000000O1000000000000O1000000000000O10000000000000O01000000000000O1000000000000O100000000000000O100000O100000O1000000000000O100000000000000O100000000000O01000000000000O100000000000000O1000000000000O100000O100000O100000000000000O1000000000000O1000000000000O010000000000000O2O001O1O001N1O2N2M2O2O1O001O1O001O1O001O1O1O001O1O02N2N00000000000000000000O1000O10000O1O1O1O1O1O001O1O1O1O2N1O1O1O2N11O10O01O1O010O00100O010O010O01O001oLeGZO\\8f0eGYO\\8f0dGZO\\8e0fGZO[8e0eG[O`8`0aG_Oe8;[GEk85VGJP90PG0V9JkF5[9DfF;a9_O`F>g9]OYFa0o9YORFd0V:VOjEi0]:QOdEl0d:nN^En0j:lNXEP1_<M3L3M4L4M3K5I_il0"}, "label": "an orange bus"}]}
{"id": 138604, "image": "COCO_train2014_000000138604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat looking in a mirror\"."}], "task": "refering", "answer_template": "The \"a cat looking in a mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 265, 266, 267, 288, 289, 290, 311, 312, 313, 334, 335, 336, 357, 358, 359, 380, 381, 382], "bbox": [0.481859375, 0.3892468619246862, 0.647296875, 0.987092050209205], "iscrowd": 0, "area": 18867.500799999998, "rle": {"size": [478, 640], "counts": "d_`45i>5K4K5L4K2O1O1N2O1N2O1O010O100O100O100O1kNVOUDi0h;CoC>m;GRD9j;LTD5h;0WD0f;4XDMh;4WDLi;5UDLk;5TDJR2_Ol6h0PGeNMQ4P9\\MRGaNOU4m8\\MWGf2e8]MYGe2e8\\M[Ge2b8^M\\Ge2`8^M_Gc2_8Q2L3N3L3N2M4L3N`0YHTI[6h701N100O2O00O100000O01000000O1000O10O01O1O1N101O1O1O1O04M4L3M3L4M3M3M3L4M3M2N3L3N3M2N3L4M2N2N2M3N2N2N2M3N2N2hKPH^1Q;G:E:G?Aj0VOPSY3"}, "label": "a cat looking in a mirror"}]}
{"id": 138604, "image": "COCO_train2014_000000138604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat seeing it ' s reflection in the mirror\"."}], "task": "refering", "answer_template": "The \"a cat seeing it ' s reflection in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 265, 266, 267, 288, 289, 290, 311, 312, 313, 334, 335, 336, 357, 358, 359, 380, 381, 382], "bbox": [0.481859375, 0.3892468619246862, 0.647296875, 0.987092050209205], "iscrowd": 0, "area": 18867.500799999998, "rle": {"size": [478, 640], "counts": "d_`45i>5K4K5L4K2O1O1N2O1N2O1O010O100O100O100O1kNVOUDi0h;CoC>m;GRD9j;LTD5h;0WD0f;4XDMh;4WDLi;5UDLk;5TDJR2_Ol6h0PGeNMQ4P9\\MRGaNOU4m8\\MWGf2e8]MYGe2e8\\M[Ge2b8^M\\Ge2`8^M_Gc2_8Q2L3N3L3N2M4L3N`0YHTI[6h701N100O2O00O100000O01000000O1000O10O01O1O1N101O1O1O1O04M4L3M3L4M3M3M3L4M3M2N3L3N3M2N3L4M2N2N2M3N2N2N2M3N2N2hKPH^1Q;G:E:G?Aj0VOPSY3"}, "label": "a cat seeing it ' s reflection in the mirror"}]}
{"id": 220529, "image": "COCO_train2014_000000220529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing without his shirt\"."}], "task": "refering", "answer_template": "The \"a man standing without his shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 21, 22, 23, 24, 39, 40, 41, 42, 56, 57, 58, 59, 60, 73, 74, 75, 76, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 235, 236, 237, 238, 239], "bbox": [0.04354, 0.015714285714285715, 0.38416, 1.0], "iscrowd": 0, "area": 39543.69879999999, "rle": {"size": [392, 500], "counts": "Wd85m;7H7J7H7J7H7J7M2N3L3N3M2M3N3M2M4M2N3L3N3M2M4M2N3kKlLQOW3n0jLPOY3m0iLQOY3n0hLPO[3m0gLPO\\3o0eLoN^3n0dLPO^3n0dLoN`3>RM@P3LdM2_2]OQN`0R2ZOTNd0n1UOYNi0j1oN]Nn0f1kNaNT1a1eNeNZ1\\1_NkNa1V1WNQOi1o0RNVOm1l0mMYOS2g0iM]OV2e0dM@\\2`0`MD`2=ZMHe29WMKi2j400O100O10000O100O10000O10000O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1jNXFKi93fF@Z9>VGSOk8j0[1N2N2M3N2N20000O10000O1000000O1000000O10000O`GHR58iJ3Q5MjJ>P5BoJe0k4ZOVKl0d4TO\\KR1^4nNbKW1Y4hNhK^1R4bNnKd1l3\\NTLj1f3VNZLP2`3oMaLW2Y3iMgL]2S3cMmLb2n2^MRMf2j2YMWMk2e2UMZMP3b2PM^MU3]2jLdMZ3X2fLhM^3T2cLkMa3Q2_LoMe3m1\\LQNi3k1XLTNk3i1ULWNm3g1TLXNn3f1RLXNR4f1oKXNT4f1lKYNW4e1jKXNZ4f1gKXN\\4_1XKQL<^2`4X1]KXL4^2a4Q1dK`LK_2c4g0kKiLC^2e4e0[L[Og3c0ZL[Oi3c0XL\\Oi3c0WL\\Ol3a0VL^Ol3nNcJ7a1j0o3lNbJ9`1h0R4lNaJ:]1:c4ZORJ;\\19e4YORJ<Z19g4XOQJ>X18T6EnI9T6FlI8X6FiI8Y6GgI7]6GdI7^6HdI4`6IoIHS67[JZOh5d0o2O2N1H9Ffbe3"}, "label": "a man standing without his shirt"}]}
{"id": 220529, "image": "COCO_train2014_000000220529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with no shirt\"."}], "task": "refering", "answer_template": "The \"a man with no shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 21, 22, 23, 24, 39, 40, 41, 42, 56, 57, 58, 59, 60, 73, 74, 75, 76, 91, 92, 93, 94, 95, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 235, 236, 237, 238, 239], "bbox": [0.04354, 0.015714285714285715, 0.38416, 1.0], "iscrowd": 0, "area": 39543.69879999999, "rle": {"size": [392, 500], "counts": "Wd85m;7H7J7H7J7H7J7M2N3L3N3M2M3N3M2M4M2N3L3N3M2M4M2N3kKlLQOW3n0jLPOY3m0iLQOY3n0hLPO[3m0gLPO\\3o0eLoN^3n0dLPO^3n0dLoN`3>RM@P3LdM2_2]OQN`0R2ZOTNd0n1UOYNi0j1oN]Nn0f1kNaNT1a1eNeNZ1\\1_NkNa1V1WNQOi1o0RNVOm1l0mMYOS2g0iM]OV2e0dM@\\2`0`MD`2=ZMHe29WMKi2j400O100O10000O100O10000O10000O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1jNXFKi93fF@Z9>VGSOk8j0[1N2N2M3N2N20000O10000O1000000O1000000O10000O`GHR58iJ3Q5MjJ>P5BoJe0k4ZOVKl0d4TO\\KR1^4nNbKW1Y4hNhK^1R4bNnKd1l3\\NTLj1f3VNZLP2`3oMaLW2Y3iMgL]2S3cMmLb2n2^MRMf2j2YMWMk2e2UMZMP3b2PM^MU3]2jLdMZ3X2fLhM^3T2cLkMa3Q2_LoMe3m1\\LQNi3k1XLTNk3i1ULWNm3g1TLXNn3f1RLXNR4f1oKXNT4f1lKYNW4e1jKXNZ4f1gKXN\\4_1XKQL<^2`4X1]KXL4^2a4Q1dK`LK_2c4g0kKiLC^2e4e0[L[Og3c0ZL[Oi3c0XL\\Oi3c0WL\\Ol3a0VL^Ol3nNcJ7a1j0o3lNbJ9`1h0R4lNaJ:]1:c4ZORJ;\\19e4YORJ<Z19g4XOQJ>X18T6EnI9T6FlI8X6FiI8Y6GgI7]6GdI7^6HdI4`6IoIHS67[JZOh5d0o2O2N1H9Ffbe3"}, "label": "a man with no shirt"}]}
{"id": 220529, "image": "COCO_train2014_000000220529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing the black color shirt holding the paper\"."}], "task": "refering", "answer_template": "The \"a man wearing the black color shirt holding the paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 65, 66, 82, 83, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248], "bbox": [0.44044, 0.20186224489795918, 0.7981, 0.9883928571428571], "iscrowd": 0, "area": 37218.89080000001, "rle": {"size": [392, 500], "counts": "c^d29e;:F9G:F:K5L4L3M4ZFTNk8Q2UGVN`8n1_GTN\\8P2cGRN[8o1dGRN[8n1fGSNY8m1fGUNX8l1gGUNX8l1gGVNW8k1iGVNV8i1iGYNV8h1hG[NV8g1hG[NW8e1gG]NX8e1eG^NY8d1eG^Nj1PO`4h3RKkLh4W3aJdMY5^2bJjMY5W2cJRNX5o1cJZNW5h1dJ`NW5]4L0O100O100O10000O100O10000O100O100O10000O100O1N2N2N2O1QO^HVMd7b2jHUMY7c2UITMn6g2\\IRMf6m2[100O1O1O2N100O1O1O100O10000000000O1000000000000000000000000O10000001O1O001O1O1O001O1jHUMm4l2PKWMo4i2nJZMR5g2lJ[MS5f2kJ[MU5f2hJ\\MX5d2gJ^MX5c2dJ`M\\5a2[JgMe5\\2iITNV6Q2bIRN^6S2YIQNg6S2QIRNo6c30001O00001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N3M3M3M4L>B2N2N2N2N3M2N2N2N3M2N2N2N2N3I6M5K5K5K5L4K5K5K5K5K8I9F:F:PNZFk0f:K5I6DaZV1"}, "label": "a man wearing the black color shirt holding the paper"}]}
{"id": 220529, "image": "COCO_train2014_000000220529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in dark blue jeans and black long sleeved shirt holding the hand of another man\"."}], "task": "refering", "answer_template": "The \"a man in dark blue jeans and black long sleeved shirt holding the hand of another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 65, 66, 82, 83, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 206, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248], "bbox": [0.44044, 0.20186224489795918, 0.7981, 0.9883928571428571], "iscrowd": 0, "area": 37218.89080000001, "rle": {"size": [392, 500], "counts": "c^d29e;:F9G:F:K5L4L3M4ZFTNk8Q2UGVN`8n1_GTN\\8P2cGRN[8o1dGRN[8n1fGSNY8m1fGUNX8l1gGUNX8l1gGVNW8k1iGVNV8i1iGYNV8h1hG[NV8g1hG[NW8e1gG]NX8e1eG^NY8d1eG^Nj1PO`4h3RKkLh4W3aJdMY5^2bJjMY5W2cJRNX5o1cJZNW5h1dJ`NW5]4L0O100O100O10000O100O10000O100O100O10000O100O1N2N2N2O1QO^HVMd7b2jHUMY7c2UITMn6g2\\IRMf6m2[100O1O1O2N100O1O1O100O10000000000O1000000000000000000000000O10000001O1O001O1O1O001O1jHUMm4l2PKWMo4i2nJZMR5g2lJ[MS5f2kJ[MU5f2hJ\\MX5d2gJ^MX5c2dJ`M\\5a2[JgMe5\\2iITNV6Q2bIRN^6S2YIQNg6S2QIRNo6c30001O00001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1O2N1O2N3M3M3M4L>B2N2N2N2N3M2N2N2N3M2N2N2N2N3I6M5K5K5K5L4K5K5K5K5K8I9F:F:PNZFk0f:K5I6DaZV1"}, "label": "a man in dark blue jeans and black long sleeved shirt holding the hand of another man"}]}
{"id": 572786, "image": "COCO_train2014_000000572786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a minivan that serves as a taxi cab\"."}], "task": "refering", "answer_template": "The \"a minivan that serves as a taxi cab\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 300, 301, 302, 303, 304, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 410, 411, 412, 413, 414, 415, 416, 417, 434, 435, 436, 437, 438, 439, 456, 457, 458, 459, 460], "bbox": [0.46666666666666673, 0.5867320261437908, 0.9720424836601309, 0.9458823529411765], "iscrowd": 0, "area": 44123.81975, "rle": {"size": [612, 612], "counts": "T[[5<ab0=D<J5J3M1O2O1N2N2O1N2N2N2O0O2N2O1N2N2N2O1N2N2O0O2N2N2O1N1O100O2N1O100O100O100O1O100O101N100O1O100O100O100O100O1O2O0O100O101N101N1O2O001N101O001O001O001O001O001O001O001N101O001O001O00001O001O001O001O0O2O001O001O001O001O001O001O001O0O2O001O001O001O001O001O00001N101O001N101O001N101O001N101O0O2O001O0O2O001O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O100O100O101N1O100O2O0O100O100O010O01O01O010O010O010O010O010O010O0010O010O0010O010O010O0001O010O001O001O001O010O00001O001O001O010O001N101O0O101O0O2O001N101O0O2O001O0O101O0O2O001N101O0O2O000O2O001N101N101N1O2O0O2O0O101N101N101N101N101N101N100O2O0O2O0O2O1N3N2M3M3N3Ld0]Oc0\\Oc0^Oc0\\Od0]Ob0]OXd9"}, "label": "a minivan that serves as a taxi cab"}]}
{"id": 572786, "image": "COCO_train2014_000000572786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow hatch back taxi\"."}], "task": "refering", "answer_template": "The \"yellow hatch back taxi\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 300, 301, 302, 303, 304, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 410, 411, 412, 413, 414, 415, 416, 417, 434, 435, 436, 437, 438, 439, 456, 457, 458, 459, 460], "bbox": [0.46666666666666673, 0.5867320261437908, 0.9720424836601309, 0.9458823529411765], "iscrowd": 0, "area": 44123.81975, "rle": {"size": [612, 612], "counts": "T[[5<ab0=D<J5J3M1O2O1N2N2O1N2N2N2O0O2N2O1N2N2N2O1N2N2O0O2N2N2O1N1O100O2N1O100O100O100O1O100O101N100O1O100O100O100O100O1O2O0O100O101N101N1O2O001N101O001O001O001O001O001O001O001N101O001O001O00001O001O001O001O0O2O001O001O001O001O001O001O001O0O2O001O001O001O001O001O00001N101O001N101O001N101O001N101O0O2O001O0O2O001O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O100O100O101N1O100O2O0O100O100O010O01O01O010O010O010O010O010O010O0010O010O0010O010O010O0001O010O001O001O001O010O00001O001O001O010O001N101O0O101O0O2O001N101O0O2O001O0O101O0O2O001N101O0O2O000O2O001N101N101N1O2O0O2O0O101N101N101N101N101N101N100O2O0O2O0O2O1N3N2M3M3N3Ld0]Oc0\\Oc0^Oc0\\Od0]Ob0]OXd9"}, "label": "yellow hatch back taxi"}]}
{"id": 572786, "image": "COCO_train2014_000000572786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow taxi cab with advertising sign on roof\"."}], "task": "refering", "answer_template": "The \"yellow taxi cab with advertising sign on roof\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 207, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273], "bbox": [0.04478758169934641, 0.32929738562091504, 0.4869117647058824, 0.5724346405228758], "iscrowd": 0, "area": 23467.515, "rle": {"size": [612, 612], "counts": "Z\\`01Pc05J7J6J4N2N2O0O2O1N2O2M4M4K4M4K4M1N2O1O1N2O1O1N2O1O1N1000O010000O01000OO2N2N1O2O11N101O001N101O001N101O001O0O2O001O0O2O001O0O2O001O001O001O001O1O001O001N101O001O001O001O001O001O010O001O001O001O0010O01O001O1O001N1N3M2M4M2N3M2N3N101N5K6K3M1O2O1N101N0010O0100O010O0001O001N101O000O2O001N101N1G901O010O010O00010O0010O01O01O01O010O01O01O101N1O2O0O2N010O1O010O0010O01O001O1O01O00000000O2O001N101O0O2O001N10001N101O0O2O1O1N2O1O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1O1N101O0O2O1O0O2O001N2O001N101O1N101O0O2O1O0O2O001N101O1N101O0O2O1O0O4M4L3L5L2N1N2O1O1N2Gl[k5"}, "label": "yellow taxi cab with advertising sign on roof"}]}
{"id": 572786, "image": "COCO_train2014_000000572786.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow cab next to the parked truck\"."}], "task": "refering", "answer_template": "The \"a yellow cab next to the parked truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 207, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273], "bbox": [0.04478758169934641, 0.32929738562091504, 0.4869117647058824, 0.5724346405228758], "iscrowd": 0, "area": 23467.515, "rle": {"size": [612, 612], "counts": "Z\\`01Pc05J7J6J4N2N2O0O2O1N2O2M4M4K4M4K4M1N2O1O1N2O1O1N2O1O1N1000O010000O01000OO2N2N1O2O11N101O001N101O001N101O001O0O2O001O0O2O001O0O2O001O001O001O001O1O001O001N101O001O001O001O001O001O010O001O001O001O0010O01O001O1O001N1N3M2M4M2N3M2N3N101N5K6K3M1O2O1N101N0010O0100O010O0001O001N101O000O2O001N101N1G901O010O010O00010O0010O01O01O01O010O01O01O101N1O2O0O2N010O1O010O0010O01O001O1O01O00000000O2O001N101O0O2O001N10001N101O0O2O1O1N2O1O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1O1N101O0O2O1O0O2O001N2O001N101O1N101O0O2O1O0O2O001N101O1N101O0O2O1O0O4M4L3L5L2N1N2O1O1N2Gl[k5"}, "label": "a yellow cab next to the parked truck"}]}
{"id": 482675, "image": "COCO_train2014_000000482675.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black bat\"."}], "task": "refering", "answer_template": "The \"black bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 27, 28, 49, 50, 51, 52, 53, 54, 76, 77, 78, 81], "bbox": [0.005984375, 0.02653301886792453, 0.538375, 0.2838679245283019], "iscrowd": 0, "area": 4764.6948, "rle": {"size": [424, 640], "counts": "me13S=3L4L4L4N1O2N1O1O1O2N11O0001O01O00010O000010O0001O01O01O0001O01O00010O00001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O000010O0000010O0001O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O0001O01O01O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O00010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O00010O0001O01O01O0001O01O00001O1N2O1N2O1N2OZjm03dURO5J6KO0N2N3M2N3Mg^j3"}, "label": "black bat"}]}
{"id": 482675, "image": "COCO_train2014_000000482675.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black baseball bat\"."}], "task": "refering", "answer_template": "The \"black baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 27, 28, 49, 50, 51, 52, 53, 54, 76, 77, 78, 81], "bbox": [0.005984375, 0.02653301886792453, 0.538375, 0.2838679245283019], "iscrowd": 0, "area": 4764.6948, "rle": {"size": [424, 640], "counts": "me13S=3L4L4L4N1O2N1O1O1O2N11O0001O01O00010O000010O0001O01O01O0001O01O00010O00001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O000010O0000010O0001O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O0001O01O01O00010O000010O0001O01O0001O01O01O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O00010O0001O01O01O00010O0000010O000010O0001O01O01O00010O0000010O000010O00010O0001O01O01O0001O01O00001O1N2O1N2O1N2OZjm03dURO5J6KO0N2N3M2N3Mg^j3"}, "label": "black baseball bat"}]}
{"id": 417141, "image": "COCO_train2014_000000417141.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with cheese , spinache , and tomatoes\"."}], "task": "refering", "answer_template": "The \"a pizza with cheese , spinache , and tomatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340], "bbox": [0.483109375, 0.28829166666666667, 0.9307500000000001, 0.8491041666666667], "iscrowd": 0, "area": 60445.27365, "rle": {"size": [480, 640], "counts": "aRa44g>>B=C>B>B>B8I8G8H8H8H8J6M3L4L4M3L4L4M2M2N2O2M2N2O1N2N3N1N2N2O2M2N2O1N3M2O1N2O2M2N2O1N3M2O1N2N2O0O2N2O1N2N101N2N2O1N2N101N2N2O1N1O2O1N2O001O0O2O001O001N101O001N101O001O0O2O001O001N101O001O0O2O1O001O0O2O001O001N101O001O0O2O001O000O1010O000000000000001O00000000000000001O00000000000000010O00000000000000001O000000000000001O00000000000000010O00000000000000001O000000001O1O001O1O001O1O1O00100O001N2O0O2O1O0O2O1N101O1N101N2O0O2O1O1N101N2O001N2O0O2O1O0O2O1N101N2O001N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2M3M3N2M3N2M3N2M3N2M2O2M3N2M3N2M3M3N2M4M5J5L5J5L5J5L5J5J7G8I8G9H7H9H7H9H7H9H7H9HV[d0"}, "label": "a pizza with cheese , spinache , and tomatoes"}]}
{"id": 417141, "image": "COCO_train2014_000000417141.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza on tabule\"."}], "task": "refering", "answer_template": "The \"a pizza on tabule\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340], "bbox": [0.483109375, 0.28829166666666667, 0.9307500000000001, 0.8491041666666667], "iscrowd": 0, "area": 60445.27365, "rle": {"size": [480, 640], "counts": "aRa44g>>B=C>B>B>B8I8G8H8H8H8J6M3L4L4M3L4L4M2M2N2O2M2N2O1N2N3N1N2N2O2M2N2O1N3M2O1N2O2M2N2O1N3M2O1N2N2O0O2N2O1N2N101N2N2O1N2N101N2N2O1N1O2O1N2O001O0O2O001O001N101O001N101O001O0O2O001O001N101O001O0O2O1O001O0O2O001O001N101O001O0O2O001O000O1010O000000000000001O00000000000000001O00000000000000010O00000000000000001O000000000000001O00000000000000010O00000000000000001O000000001O1O001O1O001O1O1O00100O001N2O0O2O1O0O2O1N101O1N101N2O0O2O1O1N101N2O001N2O0O2O1O0O2O1N101N2O001N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2M3M3N2M3N2M3N2M3N2M2O2M3N2M3N2M3M3N2M4M5J5L5J5L5J5L5J5J7G8I8G9H7H9H7H9H7H9H7H9HV[d0"}, "label": "a pizza on tabule"}]}
{"id": 417141, "image": "COCO_train2014_000000417141.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plain cheese and pepperoni pizza\"."}], "task": "refering", "answer_template": "The \"a plain cheese and pepperoni pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329], "bbox": [0.02190625, 0.28695833333333337, 0.490453125, 0.8712291666666666], "iscrowd": 0, "area": 64304.39119999999, "rle": {"size": [480, 640], "counts": "ij6?]>f0YO6K6J5J7J5J6K6J5J7J5L4N3L3N3L3N3L3N2M4M2M4M9F5L3L3N2N2N2N2N3M2N2N2M3N3M2N2N2N2N3M2N2N2N2M4M2N2N2O1N3M2N2O1N2N2N3M101N1O2N1O2O0O2N1O2O0O2N1O2N101N2O0O2O0O2O0O2O0O2O0O2O001N101N101N2O0O2O0O100O101N10000O100O101O0O10000O101N10000O100O101O00000O100000000O100000000O10000000000O10000000000O0100000000000O10000000000O01000000000O010000000000O101O00000O101N100O100O101N100O100O2O0O100O10001N100O10001N10000O101O0O100O10001N10000O101N10000O10001N100O10001N100O1O1O2O0O2N2N2O1N1O2N2O1N2N2O1N2N1O2O1N2N2N2O1M2O2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N1N3M3N2M3M3M4L3M4M3L3M4L3M4L4M2M4L4L3M4K4L5J6J5L6I8H9H7H9E:G:E:G:EXbh4"}, "label": "a plain cheese and pepperoni pizza"}]}
{"id": 417141, "image": "COCO_train2014_000000417141.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza on the left\"."}], "task": "refering", "answer_template": "The \"the pizza on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329], "bbox": [0.02190625, 0.28695833333333337, 0.490453125, 0.8712291666666666], "iscrowd": 0, "area": 64304.39119999999, "rle": {"size": [480, 640], "counts": "ij6?]>f0YO6K6J5J7J5J6K6J5J7J5L4N3L3N3L3N3L3N2M4M2M4M9F5L3L3N2N2N2N2N3M2N2N2M3N3M2N2N2N2N3M2N2N2N2M4M2N2N2O1N3M2N2O1N2N2N3M101N1O2N1O2O0O2N1O2O0O2N1O2N101N2O0O2O0O2O0O2O0O2O0O2O001N101N101N2O0O2O0O100O101N10000O100O101O0O10000O101N10000O100O101O00000O100000000O100000000O10000000000O10000000000O0100000000000O10000000000O01000000000O010000000000O101O00000O101N100O100O101N100O100O2O0O100O10001N100O10001N10000O101O0O100O10001N10000O101N10000O10001N100O10001N100O1O1O2O0O2N2N2O1N1O2N2O1N2N2O1N2N1O2O1N2N2N2O1M2O2N2M3N2N2M3N2N1N3N2N2M3N2N2M3N1N3M3N2M3M3M4L3M4M3L3M4L3M4L4M2M4L4L3M4K4L5J6J5L6I8H9H7H9E:G:E:G:EXbh4"}, "label": "the pizza on the left"}]}
{"id": 548215, "image": "COCO_train2014_000000548215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sunglass wear woman bite something\"."}], "task": "refering", "answer_template": "The \"sunglass wear woman bite something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.099, 0.0011411411411411411, 0.9855, 0.9876276276276276], "iscrowd": 0, "area": 122729.4375, "rle": {"size": [333, 500], "counts": "Yb`03V:4L4L4L4L4M3L4L4L4L4L4L4M3L4eKXN0i1N\\NOd1O`NNb1OcNN]10gNMZ12jNKV13nNJS14QOJo04VOHk06YOHg07]OFc08AE`09CF=9EE<9FG:7IG88IH77K_MnNc0X1m1KZMTOh0Q1m1MXMTOk0o0m1MVMWOk0m0n1MTMYOn0j0n1NQM[Oo0h0o1NPM\\OQ1f0n10mL^OS1c0P2OkL@U1a0o1g1oMZNP2g1PNYNP2g1PNYNo1i1oMXNQ2h1oMXNP2j1nMWNQ2j1oMVNQ2k1mMVNR2k1nMUNR2l1lMUNS2l1mMTNR2m1mMTNS2m1jMUNU2l1hMWNW2k1fMWNZ2i1dMYN[2i1cMXN]2h1cMXN\\2j1bMWN]2j1bMWN^2j1aMVN^2k1aMVN^2k1aMVN_2k1_MVN`2k1`MUNP2\\2nMeMR2[2mMfMR2\\2lMeMT2[2lMeMS2[500O1M3M3N2M4M2M3M3N2M3N2M3N2N200O100O100O100O100O100O10000O100O100O100O100O100O1O1N2N2O1N2N2O10000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000001O00000000000000000000000000000001O1M4M2M3M4L3M3N3L3M3M4L3M3WMfN\\L]1a3eN\\L^1b3dN[L_1c3cNZLa1b3bN[La1c3aN[La1c3aNZLc1c3_NZLd1c3_NZLd1d3_NXLe1e3]NXLf1f3\\NXLf1e3]NXLg1e3[NXLh1f3ZNWLi1g3YNVLk1f3XNWLk1g3WNWLk1g3WNVLm1g3UNVLn1h3TNULo1h3TNULP2h3RNVLP2h3SNTLP2j3RNSLR2i3QNTLR2j3PNSLS2k3oMRLU2k3mMSLU2j3nMSLU2k3Y2M2J6G:F9G9G:F9G9G:F9G9G:F9GfV2"}, "label": "sunglass wear woman bite something"}]}
{"id": 548215, "image": "COCO_train2014_000000548215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman wearing the tortoise shell glasses\"."}], "task": "refering", "answer_template": "The \"the woman wearing the tortoise shell glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.099, 0.0011411411411411411, 0.9855, 0.9876276276276276], "iscrowd": 0, "area": 122729.4375, "rle": {"size": [333, 500], "counts": "Yb`03V:4L4L4L4L4M3L4L4L4L4L4L4M3L4eKXN0i1N\\NOd1O`NNb1OcNN]10gNMZ12jNKV13nNJS14QOJo04VOHk06YOHg07]OFc08AE`09CF=9EE<9FG:7IG88IH77K_MnNc0X1m1KZMTOh0Q1m1MXMTOk0o0m1MVMWOk0m0n1MTMYOn0j0n1NQM[Oo0h0o1NPM\\OQ1f0n10mL^OS1c0P2OkL@U1a0o1g1oMZNP2g1PNYNP2g1PNYNo1i1oMXNQ2h1oMXNP2j1nMWNQ2j1oMVNQ2k1mMVNR2k1nMUNR2l1lMUNS2l1mMTNR2m1mMTNS2m1jMUNU2l1hMWNW2k1fMWNZ2i1dMYN[2i1cMXN]2h1cMXN\\2j1bMWN]2j1bMWN^2j1aMVN^2k1aMVN^2k1aMVN_2k1_MVN`2k1`MUNP2\\2nMeMR2[2mMfMR2\\2lMeMT2[2lMeMS2[500O1M3M3N2M4M2M3M3N2M3N2M3N2N200O100O100O100O100O100O10000O100O100O100O100O100O1O1N2N2O1N2N2O10000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000001O00000000000000000000000000000001O1M4M2M3M4L3M3N3L3M3M4L3M3WMfN\\L]1a3eN\\L^1b3dN[L_1c3cNZLa1b3bN[La1c3aN[La1c3aNZLc1c3_NZLd1c3_NZLd1d3_NXLe1e3]NXLf1f3\\NXLf1e3]NXLg1e3[NXLh1f3ZNWLi1g3YNVLk1f3XNWLk1g3WNWLk1g3WNVLm1g3UNVLn1h3TNULo1h3TNULP2h3RNVLP2h3SNTLP2j3RNSLR2i3QNTLR2j3PNSLS2k3oMRLU2k3mMSLU2j3nMSLU2k3Y2M2J6G:F9G9G:F9G9G:F9G9G:F9GfV2"}, "label": "the woman wearing the tortoise shell glasses"}]}
{"id": 548215, "image": "COCO_train2014_000000548215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a pice of food in his mouth\"."}], "task": "refering", "answer_template": "The \"a man with a pice of food in his mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 198, 199], "bbox": [0.0, 0.0028228228228228225, 0.19996, 0.9892792792792793], "iscrowd": 0, "area": 22121.205449999994, "rle": {"size": [333, 500], "counts": "b9g0V8`1oGQ80000000000000000000000000000000000000000000000000000000000000000000000001O0000000N2N2M3N3L4L4M4K4L5L4K5]Oc0M3L5G8M3M3M3M5K4^KQM_1W3YNkLf1X3VNkLg1Y3UNjLh1Y3UNjL@H:a32jLAI9`32lLmNZOE>Y1_31mLnN[OC<\\1^3NQMoNYOC<\\1]3MSMQOXOA;_1\\3KWMPOVOC9a1\\3HZMPONW1P3_OWMWOJZ1b3iNjLIF\\1\\701N101O1N101N2O001N1O2O1N101N1O3N2M2N3M2N3M3H[kQ4"}, "label": "a man with a pice of food in his mouth"}]}
{"id": 548215, "image": "COCO_train2014_000000548215.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses and a piece of fried food in his mouth tries to pass it to the other girls mouth without dropping it\"."}], "task": "refering", "answer_template": "The \"a man with glasses and a piece of fried food in his mouth tries to pass it to the other girls mouth without dropping it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 198, 199], "bbox": [0.0, 0.0028228228228228225, 0.19996, 0.9892792792792793], "iscrowd": 0, "area": 22121.205449999994, "rle": {"size": [333, 500], "counts": "b9g0V8`1oGQ80000000000000000000000000000000000000000000000000000000000000000000000001O0000000N2N2M3N3L4L4M4K4L5L4K5]Oc0M3L5G8M3M3M3M5K4^KQM_1W3YNkLf1X3VNkLg1Y3UNjLh1Y3UNjL@H:a32jLAI9`32lLmNZOE>Y1_31mLnN[OC<\\1^3NQMoNYOC<\\1]3MSMQOXOA;_1\\3KWMPOVOC9a1\\3HZMPONW1P3_OWMWOJZ1b3iNjLIF\\1\\701N101O1N101N2O001N1O2O1N101N1O3N2M2N3M2N3M3H[kQ4"}, "label": "a man with glasses and a piece of fried food in his mouth tries to pass it to the other girls mouth without dropping it"}]}
{"id": 23929, "image": "COCO_train2014_000000023929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock facing a waving uk flag\"."}], "task": "refering", "answer_template": "The \"a clock facing a waving uk flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 193, 194, 195, 196, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 263, 264, 265], "bbox": [0.38146874999999997, 0.5192740046838408, 0.578921875, 0.7973067915690866], "iscrowd": 0, "area": 11026.401399999997, "rle": {"size": [427, 640], "counts": "SPV3l0]<4K4M4L3M4K4M4M2N3M2N3M3L2O001O1O1O1O001O1O100O10000O2O000O100O2O000O10001N2O8G8I000O10001O0O100000000000001O000000001O1O1O1O1O1O1O1O1O1O00000000000010O00001O001O001O0O2O0O2O1N101O0O2O0O2O0O2O1N101N101N101N2N2N3L3N2N3M2N2N3M2N2N3M2M3N3M4L4L3J7G9H7I8G9Hml_3"}, "label": "a clock facing a waving uk flag"}]}
{"id": 23929, "image": "COCO_train2014_000000023929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clock face 11 : 35\"."}], "task": "refering", "answer_template": "The \"clock face 11 : 35\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 193, 194, 195, 196, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 263, 264, 265], "bbox": [0.38146874999999997, 0.5192740046838408, 0.578921875, 0.7973067915690866], "iscrowd": 0, "area": 11026.401399999997, "rle": {"size": [427, 640], "counts": "SPV3l0]<4K4M4L3M4K4M4M2N3M2N3M3L2O001O1O1O1O001O1O100O10000O2O000O100O2O000O10001N2O8G8I000O10001O0O100000000000001O000000001O1O1O1O1O1O1O1O1O1O00000000000010O00001O001O001O0O2O0O2O1N101O0O2O0O2O0O2O1N101N101N101N2N2N3L3N2N3M2N2N3M2N2N3M2M3N3M4L4L3J7G9H7I8G9Hml_3"}, "label": "clock face 11 : 35"}]}
{"id": 531834, "image": "COCO_train2014_000000531834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green love seat facing the other two chairs\"."}], "task": "refering", "answer_template": "The \"a green love seat facing the other two chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 207, 208, 230, 231, 232, 233, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.5725233644859813, 0.285578125, 0.997429906542056], "iscrowd": 0, "area": 21987.753399999998, "rle": {"size": [428, 640], "counts": "i7b5j7000O100000000000000000000O100000000000O10000000O10000000000000000000000O1000000001O2N3M3M2N3M3M2N3M3M2N3M3M2N2N1O1O001N2O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001N2O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O000000O1000000O1000000O100000000O0100000O1000000O1000000O10000003M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3L4M4L3MWcn5"}, "label": "a green love seat facing the other two chairs"}]}
{"id": 531834, "image": "COCO_train2014_000000531834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch on the far left but only in the picture on the right\"."}], "task": "refering", "answer_template": "The \"the couch on the far left but only in the picture on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 207, 208, 230, 231, 232, 233, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.5725233644859813, 0.285578125, 0.997429906542056], "iscrowd": 0, "area": 21987.753399999998, "rle": {"size": [428, 640], "counts": "i7b5j7000O100000000000000000000O100000000000O10000000O10000000000000000000000O1000000001O2N3M3M2N3M3M2N3M3M2N3M3M2N2N1O1O001N2O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001N2O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O000000O1000000O1000000O100000000O0100000O1000000O1000000O10000003M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3M3M4L3M4L3L4M4L3MWcn5"}, "label": "the couch on the far left but only in the picture on the right"}]}
{"id": 531834, "image": "COCO_train2014_000000531834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a patterned arm chair\"."}], "task": "refering", "answer_template": "The \"a patterned arm chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221], "bbox": [0.45218749999999996, 0.39149532710280377, 0.6756874999999999, 0.6712850467289719], "iscrowd": 0, "area": 13133.784800000001, "rle": {"size": [428, 640], "counts": "dPi31Z=>C=B=D4L0H9GjNcDm0\\;b000O100O100O2N100O100O100O101N1O100O100O101N100O100O1O100O2O0O100O1K5K5K5K6J5N2000000000001O0000000000001O0000000001O0001O000000000000001O0000000000001O000001O00000001O00000000000000O1O100O1O1O100O1O1O2N100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O2N1O1L4SOm0O1O1O1O1ROcDL^;GRE5U<M4L4LYgf2"}, "label": "a patterned arm chair"}]}
{"id": 531834, "image": "COCO_train2014_000000531834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair next to a tv\"."}], "task": "refering", "answer_template": "The \"a chair next to a tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221], "bbox": [0.45218749999999996, 0.39149532710280377, 0.6756874999999999, 0.6712850467289719], "iscrowd": 0, "area": 13133.784800000001, "rle": {"size": [428, 640], "counts": "dPi31Z=>C=B=D4L0H9GjNcDm0\\;b000O100O100O2N100O100O100O101N1O100O100O101N100O100O1O100O2O0O100O1K5K5K5K6J5N2000000000001O0000000000001O0000000001O0001O000000000000001O0000000000001O000001O00000001O00000000000000O1O100O1O1O100O1O1O2N100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O2N1O1L4SOm0O1O1O1O1ROcDL^;GRE5U<M4L4LYgf2"}, "label": "a chair next to a tv"}]}
{"id": 531834, "image": "COCO_train2014_000000531834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark brown couch on the right\"."}], "task": "refering", "answer_template": "The \"the dark brown couch on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.6932031249999999, 0.49462616822429906, 1.0, 0.982803738317757], "iscrowd": 0, "area": 37123.28675000001, "rle": {"size": [428, 640], "counts": "gli53T=`0_Ob0_Ob0^Oa0_Oa0_Oa0_Oa0^O=D5B>]Oc0^Ob0N2000000000O10000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000O100O100O100O1O100O100O100O100O100O100O100O100O100O1O100O10000000000001O00000000001O0000000000001O0000000000001O00000000001O00000000WI"}, "label": "the dark brown couch on the right"}]}
{"id": 531834, "image": "COCO_train2014_000000531834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the floral couch couch facing to the left\"."}], "task": "refering", "answer_template": "The \"the floral couch couch facing to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.6932031249999999, 0.49462616822429906, 1.0, 0.982803738317757], "iscrowd": 0, "area": 37123.28675000001, "rle": {"size": [428, 640], "counts": "gli53T=`0_Ob0_Ob0^Oa0_Oa0_Oa0_Oa0^O=D5B>]Oc0^Ob0N2000000000O10000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000O100O100O100O1O100O100O100O100O100O100O100O100O100O1O100O10000000000001O00000000001O0000000000001O0000000000001O00000000001O00000000WI"}, "label": "the floral couch couch facing to the left"}]}
{"id": 228732, "image": "COCO_train2014_000000228732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gray hooded man sitting behind a couch\"."}], "task": "refering", "answer_template": "The \"gray hooded man sitting behind a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 39, 40, 41, 62, 63, 64, 65, 85, 86, 87, 88, 89, 109, 110, 111, 112, 133, 134, 135], "bbox": [0.707484375, 0.049102296450939456, 0.90215625, 0.33430062630480173], "iscrowd": 0, "area": 8874.29165, "rle": {"size": [479, 640], "counts": "Soc67f>3M2N3N1N3M2N3M2M3M4M3N1N3M3L4M3M2N3L4M3O1N2O1N2N2O0eCZN\\;g1cD[N\\;e1cD]N\\;d1cD]N[;d1dD^N[;b1dD`N[;`1dDaN\\;`1aDbN^;a2O001O000100O00100O00100O10O01O2O1N3N001N2O0O2O1O0O2O2N2M3N0O100O100O100O010O010O10O01000O010O010O010O0100O010O0O2O1N101N101N2O0O2O0O2O0O2O0O2O0O2N4M2M3N2N2N3M3M3M2M4M3M3M3L4J6L3OmZm0"}, "label": "gray hooded man sitting behind a couch"}]}
{"id": 228732, "image": "COCO_train2014_000000228732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark man wearing a gray hooded sweatshirt covering most of his face\"."}], "task": "refering", "answer_template": "The \"a dark man wearing a gray hooded sweatshirt covering most of his face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 39, 40, 41, 62, 63, 64, 65, 85, 86, 87, 88, 89, 109, 110, 111, 112, 133, 134, 135], "bbox": [0.707484375, 0.049102296450939456, 0.90215625, 0.33430062630480173], "iscrowd": 0, "area": 8874.29165, "rle": {"size": [479, 640], "counts": "Soc67f>3M2N3N1N3M2N3M2M3M4M3N1N3M3L4M3M2N3L4M3O1N2O1N2N2O0eCZN\\;g1cD[N\\;e1cD]N\\;d1cD]N[;d1dD^N[;b1dD`N[;`1dDaN\\;`1aDbN^;a2O001O000100O00100O00100O10O01O2O1N3N001N2O0O2O1O0O2O2N2M3N0O100O100O100O010O010O10O01000O010O010O010O0100O010O0O2O1N101N101N2O0O2O0O2O0O2O0O2O0O2N4M2M3N2N2N3M3M3M2M4M3M3M3L4J6L3OmZm0"}, "label": "a dark man wearing a gray hooded sweatshirt covering most of his face"}]}
{"id": 245118, "image": "COCO_train2014_000000245118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey backpack attached to an unseen person\"."}], "task": "refering", "answer_template": "The \"a grey backpack attached to an unseen person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 391, 392, 393, 394, 395], "bbox": [0.0, 0.57458984375, 0.21214062500000003, 0.981328125], "iscrowd": 0, "area": 21143.8128, "rle": {"size": [512, 640], "counts": "m<e2d9g31O01O01O01O01O0010O0001O01O01O00001O001O00001O00001O00001O001O00001O00001O00001O0O2O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001N2O001O00001O001O001O001O001O0010O01O002N2N2N2N2N2N3M2N2N2N2N3M2N2N5K5K5K5K5K5K2N101N2N1O2N2N1O2N2N1O2N2N3M6J7I6J7I3M2N2N2N2N2N2I7Iaak7"}, "label": "a grey backpack attached to an unseen person"}]}
{"id": 245118, "image": "COCO_train2014_000000245118.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cream colored backpack\"."}], "task": "refering", "answer_template": "The \"a cream colored backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 255, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 391, 392, 393, 394, 395], "bbox": [0.0, 0.57458984375, 0.21214062500000003, 0.981328125], "iscrowd": 0, "area": 21143.8128, "rle": {"size": [512, 640], "counts": "m<e2d9g31O01O01O01O01O0010O0001O01O01O00001O001O00001O00001O00001O001O00001O00001O00001O0O2O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O001N2O001O00001O001O001O001O001O0010O01O002N2N2N2N2N2N3M2N2N2N2N3M2N2N5K5K5K5K5K5K2N101N2N1O2N2N1O2N2N1O2N2N3M6J7I6J7I3M2N2N2N2N2N2I7Iaak7"}, "label": "a cream colored backpack"}]}
{"id": 572801, "image": "COCO_train2014_000000572801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus that says afghan post closest to the camera\"."}], "task": "refering", "answer_template": "The \"bus that says afghan post closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.2283125, 0.004494117647058823, 0.99534375, 0.9842588235294117], "iscrowd": 0, "area": 194444.22235, "rle": {"size": [425, 640], "counts": "oll12V=7J7H7I8I6I7I8I2M3kJoNnMR1n1JXM8d2d0bL^O[3]1lKdNP4V2XKlMd4]2UKeMh4c2RK^Mj4k2oJWMm4R3mJoLP5Y3iJiLS5i5L5L3L4L4M3L4L4M3L4L4M3L4L5L3L4L4M3L4L4M3L4L4M3L4M4K4L4M3L4L4M3L4L4M3L4L4M2M3M3N2M3M3N2M3O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000`6`IYQ1"}, "label": "bus that says afghan post closest to the camera"}]}
{"id": 572801, "image": "COCO_train2014_000000572801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus right in front of the camera\"."}], "task": "refering", "answer_template": "The \"the bus right in front of the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.2283125, 0.004494117647058823, 0.99534375, 0.9842588235294117], "iscrowd": 0, "area": 194444.22235, "rle": {"size": [425, 640], "counts": "oll12V=7J7H7I8I6I7I8I2M3kJoNnMR1n1JXM8d2d0bL^O[3]1lKdNP4V2XKlMd4]2UKeMh4c2RK^Mj4k2oJWMm4R3mJoLP5Y3iJiLS5i5L5L3L4L4M3L4L4M3L4L4M3L4L5L3L4L4M3L4L4M3L4L4M3L4M4K4L4M3L4L4M3L4L4M3L4L4M2M3M3N2M3M3N2M3O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000`6`IYQ1"}, "label": "the bus right in front of the camera"}]}
{"id": 572801, "image": "COCO_train2014_000000572801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck front in the background\"."}], "task": "refering", "answer_template": "The \"a truck front in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 26, 27, 28, 29, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 279, 280], "bbox": [0.03290625, 0.010141176470588235, 0.3230625, 0.8141882352941178], "iscrowd": 0, "area": 37622.73029999998, "rle": {"size": [425, 640], "counts": "jl89i2LV7i0oGAm7m0^G]O`8k0UG[Og8n0RGTOj8U1oFmNn8[1kFgNQ9b1hF_NU9m2L4L4L4L3L5L4L4L4M2N3L4M3M3M2N3M3M3L4M2N3M3M3M3L3N3M3M3M3M2N3L4M3M3M2N3M3L4M3M2O2N2N2O1N1O2O1N2N2O0O2N1O100O2N100O1O2O0O1O1O2O0O1O101N1O101N1O1O101N1O100O2N100O1O2O0O1O1O10O01O100O1O100O001O100O1O10O01O100O1O1O10O010O1N101N101N1O2O0@a000O0100O010O10oJoKi0Q4UO_L>a3@oL2P3LaME`29PNYOP2d0`NoN`1o0PObNQ1[1@WN`0g1OlM1Q2`0`MA]2P1UMoNj2`1iL`NT3l5N2M3N2M3N2M3N2M3M3N2M3N2M2O2M2N2O1N3M2N2N2N3M2N2N3M3M2N3M3M3M3N1N3M3M3M2N3M3M3M2Nhfc5"}, "label": "a truck front in the background"}]}
{"id": 572801, "image": "COCO_train2014_000000572801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of the lesser seen truck\"."}], "task": "refering", "answer_template": "The \"the front of the lesser seen truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 26, 27, 28, 29, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 279, 280], "bbox": [0.03290625, 0.010141176470588235, 0.3230625, 0.8141882352941178], "iscrowd": 0, "area": 37622.73029999998, "rle": {"size": [425, 640], "counts": "jl89i2LV7i0oGAm7m0^G]O`8k0UG[Og8n0RGTOj8U1oFmNn8[1kFgNQ9b1hF_NU9m2L4L4L4L3L5L4L4L4M2N3L4M3M3M2N3M3M3L4M2N3M3M3M3L3N3M3M3M3M2N3L4M3M3M2N3M3L4M3M2O2N2N2O1N1O2O1N2N2O0O2N1O100O2N100O1O2O0O1O1O2O0O1O101N1O101N1O1O101N1O100O2N100O1O2O0O1O1O10O01O100O1O100O001O100O1O10O01O100O1O1O10O010O1N101N101N1O2O0@a000O0100O010O10oJoKi0Q4UO_L>a3@oL2P3LaME`29PNYOP2d0`NoN`1o0PObNQ1[1@WN`0g1OlM1Q2`0`MA]2P1UMoNj2`1iL`NT3l5N2M3N2M3N2M3N2M3M3N2M3N2M2O2M2N2O1N3M2N2N2N3M2N2N3M3M2N3M3M3M3N1N3M3M3M2N3M3M3M2Nhfc5"}, "label": "the front of the lesser seen truck"}]}
{"id": 507266, "image": "COCO_train2014_000000507266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat taking a nap\"."}], "task": "refering", "answer_template": "The \"a black and white cat taking a nap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 66, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176], "bbox": [0.162, 0.2576, 1.0, 0.7617333333333333], "iscrowd": 0, "area": 52766.701550000005, "rle": {"size": [375, 500], "counts": "nkm06_;;E3M2N3M3M2N3M2M4M3M2N3M3L3N2N2N3M2N2N2O2N1O1O1O2N1O1O1O2M2O1O1ZO_MUHc2j7_MSHb2m7_MPHc2P8^MnGd2P8^MnGc2R8^MjGe2V8]MfGf2Y8<00O2O00001O0000001O00001O00001O00001O000O2O001O001N101O001N101O1O0O2O001N101N101N100O2O0O2O0O101N101N100O2O0O2O0O2O0O2O000O2O0O2O0O2O001N101O00001O001O001O001O0O101O00001O00001O00001N10001O00001O00001O00O100000000000O1000000000000000O1000000000000000000000O100000000000O1000000000O100000000000000O100O1O010O100O1O11O001O1O1O001O1O1O001O1N101O1O1O0000001O00001O00001O0000001O1O1O1O1O1O1N2O1O0O1000000O10000O1000000O10O10O01000O0100O1O1O1O1O1O1B>GiJmI\\5S6eJlI[5U631O1O2N1O1N2O1O1N2O1N2O1N2O1N101O1N101O001O000O2O001O00001O000000000000000000000O1000000000000000001O0000001O0O10001N10000O101N10000O2O000O10001N10000O2N100O101N1O100O100O1O2O0O100O1N2O1dNVHjNk7T1[HfNg7Y1^HbNc7^1bH\\N`7b1eHXN]7h1hHRNY7m1Y1O1O1O1O1O100O1O1O1O1N2O001N101N1N3N2N1O2@?O2O10000O100O100O1O100O00100O1O010O1O1O001O1O1O1O2N1O2N1O2N1O2N1OSL"}, "label": "a black and white cat taking a nap"}]}
{"id": 507266, "image": "COCO_train2014_000000507266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white cat laying on a blanket\"."}], "task": "refering", "answer_template": "The \"black and white cat laying on a blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 66, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176], "bbox": [0.162, 0.2576, 1.0, 0.7617333333333333], "iscrowd": 0, "area": 52766.701550000005, "rle": {"size": [375, 500], "counts": "nkm06_;;E3M2N3M3M2N3M2M4M3M2N3M3L3N2N2N3M2N2N2O2N1O1O1O2N1O1O1O2M2O1O1ZO_MUHc2j7_MSHb2m7_MPHc2P8^MnGd2P8^MnGc2R8^MjGe2V8]MfGf2Y8<00O2O00001O0000001O00001O00001O00001O000O2O001O001N101O001N101O1O0O2O001N101N101N100O2O0O2O0O101N101N100O2O0O2O0O2O0O2O000O2O0O2O0O2O001N101O00001O001O001O001O0O101O00001O00001O00001N10001O00001O00001O00O100000000000O1000000000000000O1000000000000000000000O100000000000O1000000000O100000000000000O100O1O010O100O1O11O001O1O1O001O1O1O001O1N101O1O1O0000001O00001O00001O0000001O1O1O1O1O1O1N2O1O0O1000000O10000O1000000O10O10O01000O0100O1O1O1O1O1O1B>GiJmI\\5S6eJlI[5U631O1O2N1O1N2O1O1N2O1N2O1N2O1N101O1N101O001O000O2O001O00001O000000000000000000000O1000000000000000001O0000001O0O10001N10000O101N10000O2O000O10001N10000O2N100O101N1O100O100O1O2O0O100O1N2O1dNVHjNk7T1[HfNg7Y1^HbNc7^1bH\\N`7b1eHXN]7h1hHRNY7m1Y1O1O1O1O1O100O1O1O1O1N2O001N101N1N3N2N1O2@?O2O10000O100O100O1O100O00100O1O010O1O1O001O1O1O1O2N1O2N1O2N1O2N1OSL"}, "label": "black and white cat laying on a blanket"}]}
{"id": 507266, "image": "COCO_train2014_000000507266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat behind a black and white cat\"."}], "task": "refering", "answer_template": "The \"a black cat behind a black and white cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 20, 21, 22, 23, 24, 25, 26, 27, 28, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93], "bbox": [0.01932, 0.02146666666666667, 0.57136, 0.4034400000000001], "iscrowd": 0, "area": 25391.55864999999, "rle": {"size": [375, 500], "counts": "Th3>X;?A8I0O2N1O1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O101N1O1O2N1O101O00001O0O101O000O10O10O01O010O10O0100O00100O10O0100O10O10O10O100O100O10O01O1O1O1O1O1O00100000000000O1000000000000O1000N2O1N101000000O10O100000O1000O10O100O1000O2O001O0O101N101O0O2O0O101N101O1N2O1O1O1N2O1O1O0O1000O10O10000O0100000O2O00000O2O000O101O0O10001N10000O100O1000001O000000000000O1000O100000O10000000000000000000O10000000000O1000000000000000000O100000O1000000O1O1O1O1O001O1O1O001O100O0010ZNjGKU83oGLP81UHNk7OYH0g7N\\H1c7N`H1`7NdHN\\72fHMY7CYI8TN^Oo:;UEFn:5TEK\\;0O1dX^2"}, "label": "a black cat behind a black and white cat"}]}
{"id": 507266, "image": "COCO_train2014_000000507266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color cat\"."}], "task": "refering", "answer_template": "The \"a black color cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 20, 21, 22, 23, 24, 25, 26, 27, 28, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93], "bbox": [0.01932, 0.02146666666666667, 0.57136, 0.4034400000000001], "iscrowd": 0, "area": 25391.55864999999, "rle": {"size": [375, 500], "counts": "Th3>X;?A8I0O2N1O1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O101N1O1O2N1O101O00001O0O101O000O10O10O01O010O10O0100O00100O10O0100O10O10O10O100O100O10O01O1O1O1O1O1O00100000000000O1000000000000O1000N2O1N101000000O10O100000O1000O10O100O1000O2O001O0O101N101O0O2O0O101N101O1N2O1O1O1N2O1O1O0O1000O10O10000O0100000O2O00000O2O000O101O0O10001N10000O100O1000001O000000000000O1000O100000O10000000000000000000O10000000000O1000000000000000000O100000O1000000O1O1O1O1O001O1O1O001O100O0010ZNjGKU83oGLP81UHNk7OYH0g7N\\H1c7N`H1`7NdHN\\72fHMY7CYI8TN^Oo:;UEFn:5TEK\\;0O1dX^2"}, "label": "a black color cat"}]}
{"id": 556424, "image": "COCO_train2014_000000556424.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange between a green leaf and another orange\"."}], "task": "refering", "answer_template": "The \"an orange between a green leaf and another orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 144, 145, 146, 168], "bbox": [0.26321875, 0.14152224824355972, 0.465140625, 0.48046838407494147], "iscrowd": 0, "area": 10596.927600000003, "rle": {"size": [427, 640], "counts": "\\UV22Q=d0@;G4M3M3L4M3L4M3M3L5L4L4K4M4L4L4M3L4M1N3M2O1N3N1N2O2M2O1N3M2O1N3N1N2O2M2N2O1N10O0100O10O0100O1O010O100ON3K4L5K4L5L301O0O3N1O1O1O2M2O1O1O2N1N2O1O2N1O1N2O2M2N2O1N3N1N2O1O2O001N101O1N101O1N101O1N2O001N2O001N2O1O0O2O1O0O2O1O1N101O1O2M4M3M3L3N3M3L4M3M2M4M3M3LSm^4"}, "label": "an orange between a green leaf and another orange"}]}
{"id": 556424, "image": "COCO_train2014_000000556424.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange on the right side\"."}], "task": "refering", "answer_template": "The \"the orange on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 144, 145, 146, 168], "bbox": [0.26321875, 0.14152224824355972, 0.465140625, 0.48046838407494147], "iscrowd": 0, "area": 10596.927600000003, "rle": {"size": [427, 640], "counts": "\\UV22Q=d0@;G4M3M3L4M3L4M3M3L5L4L4K4M4L4L4M3L4M1N3M2O1N3N1N2O2M2O1N3M2O1N3N1N2O2M2N2O1N10O0100O10O0100O1O010O100ON3K4L5K4L5L301O0O3N1O1O1O2M2O1O1O2N1N2O1O2N1O1N2O2M2N2O1N3N1N2O1O2O001N101O1N101O1N101O1N2O001N2O001N2O1O0O2O1O0O2O1O1N101O1O2M4M3M3L3N3M3L4M3M2M4M3M3LSm^4"}, "label": "the orange on the right side"}]}
{"id": 556424, "image": "COCO_train2014_000000556424.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange fruit\"."}], "task": "refering", "answer_template": "The \"an orange fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 186, 187, 188], "bbox": [0.04684375, 0.21911007025761126, 0.28121874999999996, 0.5652693208430913], "iscrowd": 0, "area": 15598.518849999999, "rle": {"size": [427, 640], "counts": "`e<3T=6J6J6J6J7I6J6J6J6J6J6J6K5M3M3N2M3N3L3N2M101N2N2O0O2O1N2O1N101N2N2O0O2O1N2O1N101N2N2O0O2O1N2O1N1O2O1O1O001O1O1O1O00100O00000000000001O00000O100000001O000000000O10001O000000000000001N1000000000001UO_GjLa8V3k0N1O2O1N1O2N101N2N1O2O0O2N101N2N101N101N2O0O2O0O2O1N101N101N2O1N2O1N2O2M2O1N2O1N3N1N2O1N5L8G9H8G9H9Fbdo5"}, "label": "an orange fruit"}]}
{"id": 556424, "image": "COCO_train2014_000000556424.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange on the left\"."}], "task": "refering", "answer_template": "The \"the orange on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 186, 187, 188], "bbox": [0.04684375, 0.21911007025761126, 0.28121874999999996, 0.5652693208430913], "iscrowd": 0, "area": 15598.518849999999, "rle": {"size": [427, 640], "counts": "`e<3T=6J6J6J6J7I6J6J6J6J6J6J6K5M3M3N2M3N3L3N2M101N2N2O0O2O1N2O1N101N2N2O0O2O1N2O1N101N2N2O0O2O1N2O1N1O2O1O1O001O1O1O1O00100O00000000000001O00000O100000001O000000000O10001O000000000000001N1000000000001UO_GjLa8V3k0N1O2O1N1O2N101N2N1O2O0O2N101N2N101N101N2O0O2O0O2O1N101N101N2O1N2O1N2O2M2O1N2O1N3N1N2O1N5L8G9H8G9H9Fbdo5"}, "label": "the orange on the left"}]}
{"id": 458124, "image": "COCO_train2014_000000458124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the striped shirt\"."}], "task": "refering", "answer_template": "The \"the man with the striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 121, 122, 123, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 176, 187, 188, 190, 191, 192, 193, 207, 208], "bbox": [0.033041666666666664, 0.28153125, 0.4144166666666667, 0.5608125], "iscrowd": 0, "area": 17223.00815, "rle": {"size": [640, 480], "counts": "dY:3jc04L5K4L5K4M4M2N3M1N2O2N1O1N2O2N1O1O1N3N1O1O1N3N1O1O0O1000000O0100000O10O1000O100000O10O100000O1000001O0O10001O0000lNbNg_O]1Y`0oN[_OR1d`0U1O100O00102O2N2M3d@kLi=W3RBnLl=T3nAQMQ>P3kAUMR>n2iAVMV>l2eAXMZ>j2aAZM]>V4O1O2M2O1O1O1O001O00001O00001O00001O00001O01O100O1N2N2O1N2N101N2N2O1N2N2O1M3L4L4K8I7K5J6K2N2N2M3N2N1O2M3N2N101O1O1O001O1O1O001N2O001O1O1O001O1O001O1O1N101O1O00001O001O001O0O2O00001O001O001O00001N101O003M3N3L3M\\1dN5K5KTZ_5"}, "label": "the man with the striped shirt"}]}
{"id": 458124, "image": "COCO_train2014_000000458124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with his back away from us in a blue and white striped shirt eating\"."}], "task": "refering", "answer_template": "The \"man with his back away from us in a blue and white striped shirt eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 121, 122, 123, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 176, 187, 188, 190, 191, 192, 193, 207, 208], "bbox": [0.033041666666666664, 0.28153125, 0.4144166666666667, 0.5608125], "iscrowd": 0, "area": 17223.00815, "rle": {"size": [640, 480], "counts": "dY:3jc04L5K4L5K4M4M2N3M1N2O2N1O1N2O2N1O1O1N3N1O1O1N3N1O1O0O1000000O0100000O10O1000O100000O10O100000O1000001O0O10001O0000lNbNg_O]1Y`0oN[_OR1d`0U1O100O00102O2N2M3d@kLi=W3RBnLl=T3nAQMQ>P3kAUMR>n2iAVMV>l2eAXMZ>j2aAZM]>V4O1O2M2O1O1O1O001O00001O00001O00001O00001O01O100O1N2N2O1N2N101N2N2O1N2N2O1M3L4L4K8I7K5J6K2N2N2M3N2N1O2M3N2N101O1O1O001O1O1O001N2O001O1O1O001O1O001O1O1N101O1O00001O001O001O0O2O00001O001O001O00001N101O003M3N3L3M\\1dN5K5KTZ_5"}, "label": "man with his back away from us in a blue and white striped shirt eating"}]}
{"id": 458124, "image": "COCO_train2014_000000458124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"the man wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 126, 127, 128, 143, 144, 145, 146, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198], "bbox": [0.39939583333333334, 0.269140625, 0.6636666666666666, 0.505625], "iscrowd": 0, "area": 13206.202800000003, "rle": {"size": [640, 480], "counts": "cXh35gc07I8H8H7I8H7I=D>A4N200O2O0O10001kNUNR@l1j?XNV@h1f?\\NY@e1c?_N]@a1_?cN`@_1^?bNb@^1^?bNa@_1^?bNb@^1^?bNb@^1^?bNa@_1^?bNb@]1_?cN`@^1_?bNb@^1^?bNa@_1^?bNb@^1^?bNa@_1_?aNa@_1^?bNb@^1^?b1O10000O100O100O10000O100O1001O0000001O0000001O00001O0000001O0000001O00000lLc@e1^?YNc@g1]?XNe@g1[?XNf@h1[?VNg@i1]?RNd@n1`?mMb@R2b?iM`@V2c?fM^@Z2f?aM\\@^2h?]MY@c2j?YMX@f2l?UMV@i2a`001O001O001O1O001O001O1O001O001O1O001N3N2N2N2M3N2N2M3N2N2N3L3N2N2M3N2N2N2M5L8H7I8G8IVZT3"}, "label": "the man wearing a white shirt"}]}
{"id": 458124, "image": "COCO_train2014_000000458124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man facing away wearing a white business shirt and conversing with others at a table\"."}], "task": "refering", "answer_template": "The \"a man facing away wearing a white business shirt and conversing with others at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 126, 127, 128, 143, 144, 145, 146, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198], "bbox": [0.39939583333333334, 0.269140625, 0.6636666666666666, 0.505625], "iscrowd": 0, "area": 13206.202800000003, "rle": {"size": [640, 480], "counts": "cXh35gc07I8H8H7I8H7I=D>A4N200O2O0O10001kNUNR@l1j?XNV@h1f?\\NY@e1c?_N]@a1_?cN`@_1^?bNb@^1^?bNa@_1^?bNb@^1^?bNb@^1^?bNa@_1^?bNb@]1_?cN`@^1_?bNb@^1^?bNa@_1^?bNb@^1^?bNa@_1_?aNa@_1^?bNb@^1^?b1O10000O100O100O10000O100O1001O0000001O0000001O00001O0000001O0000001O00000lLc@e1^?YNc@g1]?XNe@g1[?XNf@h1[?VNg@i1]?RNd@n1`?mMb@R2b?iM`@V2c?fM^@Z2f?aM\\@^2h?]MY@c2j?YMX@f2l?UMV@i2a`001O001O001O1O001O001O1O001O001O1O001N3N2N2N2M3N2N2M3N2N2N3L3N2N2M3N2N2N2M5L8H7I8G8IVZT3"}, "label": "a man facing away wearing a white business shirt and conversing with others at a table"}]}
{"id": 458124, "image": "COCO_train2014_000000458124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sipping a glass of wine\"."}], "task": "refering", "answer_template": "The \"a woman sipping a glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 99, 100, 101, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 216, 217, 218, 219], "bbox": [0.7221041666666667, 0.18426562500000002, 1.0, 0.5640468750000001], "iscrowd": 0, "area": 22497.192999999996, "rle": {"size": [640, 480], "counts": "nVi65ec06K5J7I6K5J6J6K5J6J6K5J6J6K5J7I6K5J6N2N2N2M3N2N2N2M3N3M2N2M3N2O1N2N2O1N2N2N2O2M2N2O1N2N2O1N2N2N2O1N2N3L3M3N2M3M3M3M3M3M3N2M3M4L3M3M3M3N2M3N200O10POXChJg<U5]CjJd<V5\\CjJc<W5^ChJb<X5^ChJb<X5^ChJa<Y5_CgJa<Y5`CfJ_<[5aCeJ_<[5aCeJ^<\\5bCdJ^<\\5cCcJ\\<^5dCaJ]<_5cCaJ\\<`5dC`J\\<`5dC`J\\<`5eC_JZ<b5fC^JZ<a5gC_JX<b5hC^JX<b5iC]JV<d5jC\\JV<d5jC\\JU<e5kC[JU<e5lCYJT<h5lCXJT<h5lCXJT<h5lCWJT<j5mCSJU<m5kCQJX<n5hCPJZ<P6fCnI\\<R6eCkI^<T6bCjI`<V6`ChIb<X6^CeIf<Z6`00001O001O00001fLZB8f=G[B9e=F]B9d=E]B:d=E^B:c=D]B<f=@[B`0h=]OXBc0j=[OWBd0l=YOTBh0n=UORBk0P>SOPBm0S>POmAP1W>kNjABIOi>GiAFe00P>@gAM_10e7"}, "label": "a woman sipping a glass of wine"}]}
{"id": 458124, "image": "COCO_train2014_000000458124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with dark brown hair in a black and white dress\"."}], "task": "refering", "answer_template": "The \"a woman with dark brown hair in a black and white dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 99, 100, 101, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 216, 217, 218, 219], "bbox": [0.7221041666666667, 0.18426562500000002, 1.0, 0.5640468750000001], "iscrowd": 0, "area": 22497.192999999996, "rle": {"size": [640, 480], "counts": "nVi65ec06K5J7I6K5J6J6K5J6J6K5J6J6K5J7I6K5J6N2N2N2M3N2N2N2M3N3M2N2M3N2O1N2N2O1N2N2N2O2M2N2O1N2N2O1N2N2N2O1N2N3L3M3N2M3M3M3M3M3M3N2M3M4L3M3M3M3N2M3N200O10POXChJg<U5]CjJd<V5\\CjJc<W5^ChJb<X5^ChJb<X5^ChJa<Y5_CgJa<Y5`CfJ_<[5aCeJ_<[5aCeJ^<\\5bCdJ^<\\5cCcJ\\<^5dCaJ]<_5cCaJ\\<`5dC`J\\<`5dC`J\\<`5eC_JZ<b5fC^JZ<a5gC_JX<b5hC^JX<b5iC]JV<d5jC\\JV<d5jC\\JU<e5kC[JU<e5lCYJT<h5lCXJT<h5lCXJT<h5lCWJT<j5mCSJU<m5kCQJX<n5hCPJZ<P6fCnI\\<R6eCkI^<T6bCjI`<V6`ChIb<X6^CeIf<Z6`00001O001O00001fLZB8f=G[B9e=F]B9d=E]B:d=E^B:c=D]B<f=@[B`0h=]OXBc0j=[OWBd0l=YOTBh0n=UORBk0P>SOPBm0S>POmAP1W>kNjABIOi>GiAFe00P>@gAM_10e7"}, "label": "a woman with dark brown hair in a black and white dress"}]}
{"id": 114060, "image": "COCO_train2014_000000114060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left hand side of sandwich on plate\"."}], "task": "refering", "answer_template": "The \"left hand side of sandwich on plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238], "bbox": [0.003359375, 0.09641666666666666, 0.48490625, 0.6556875000000001], "iscrowd": 0, "area": 60963.4411, "rle": {"size": [480, 640], "counts": "RQ12j>6K5_DDV8c0gG^OT8h0kGXOQ8m0mGTOn7S1PHmN_78fEQ1i2hN`7f1_HZN`7i1]HXNb7j1PHcNo7`1mGbNR8`1kGbNT8a1gGbNX8a1dGaN[8b1`GaN_8b1]GaNa8b1ZGaNe8c1VG`Nh8c1SG`Nl8c1PG`Nn8c1mF`NR9c1jF`NT9d1gF^NX9e1bF_N]9d1^F_Na9`30000000000000000000001O00001O001O00001O00001O001O00001O00002N4L4L3M3M3M3M1O1O1O2N1O1O1O1O1O00001O000000001O000000001O00002N1O1O1O1O2N1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1OO1000000O1000000O100000000O1000O10O1000000O1000000O10000000000O100000000000O10O10O100O10000O010O100O10000O01000O100000O10000O10000O1000000O10O10O1000000O10000O10000O100000O010000O1O100O1O1O1O100O1O1O1O1O10O01O1O1O1O1O1N2O1O1N2O1O1O1N2O1O1N10100O1O100O1O1O100O1O1O100O1O100O1O001O1iInG_5S8`JnH_4S7`KZIS4k8N2N2N2N3M2N2M3nL\\Ee1f:YN\\Ef1e:XN]Ef1e:WN_Ef1c:XN_Eg1b:WNaEf1a:WNbEg1`:WNcEf1m;O1N2N2N2O1N2N2O1N2N2N2O1L4I7J6J6IV`j4"}, "label": "left hand side of sandwich on plate"}]}
{"id": 114060, "image": "COCO_train2014_000000114060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwich with a lot of corned beef\"."}], "task": "refering", "answer_template": "The \"half of a sandwich with a lot of corned beef\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.290921875, 0.18610416666666665, 0.9854218750000001, 0.7466458333333333], "iscrowd": 0, "area": 68660.47605, "rle": {"size": [480, 640], "counts": "gog21n>1O100O1O101N100O1O100O2O0O1O100O1O101N100O1O100O2O0O100000001O00000000001O0000000000001O000000000O2O00000000001O00000000001O000oNZOhCf0P<BPD>m;ERD<k;GSD;i;IVD9f;JXD8e;KZD6b;N\\D4b;N\\D4b;N]D3a;:RDIk;g100O10000O2O0O10000O101N100O10000O1O1O1O1gLQM`JP3_5RM_Jo2_5UM^Jl2`5XM]Ji2a5[M]Je2b5]M\\Jd2b5`M[Ja2a5[LXHX1T2^2d5fMYJ[2f5gMXJZ2h5gMVJZ2i5iMTJX2k5jMRJX2n5iMPJX2o5jMoIW2Q6kMkIW2T6kMjIV2U6lMiIU2W6lMfIV2Y6mMdIT2\\6mMbIT2]6nMaIS2^6oM_IS2a6nM]IS2b6oM\\IR2d6oMlH`2S7g2O100O100O1O100O100O1O100O100O1O100O1O100O1O100O1O10000000000O100000000000000O10000000000O10000000000O0100O10000O1000O0100O10000O10000O010O10000O100O10000O10O10O100O10000O100O10000O010O10000O100O10000O100O10000O100O10000O100O1000O0100O10000O100O10000O100O10O10O100O100O10000O100O1000O0101N10000O2O0O10001N100O101O0O100O2O000O101N10000O2O0O10001N100O101O0O101N10001N100O2O000O2O0O101O0O100O2O000O2O0O101O0O101N10001N100O101N1O101N1O1O2N1O1O2N100O2N1O1O1O001O100O1O001TLYF^2h9bMXF]2i9bMYF\\2h9dMXF\\2g9eMYFZ2h9eMZFY2g9gMYFX2h9gMYFX2h9hMYFV2g9jMZFU2g9kMYFT2h9kMZFT2g9kMYFT2P:dMPF[2Y:\\MhEc2b:TM_Ej2j:mLWER3\\;O2N1O1O2O0O1O2N1O1O2N1O1O2N1O101N100O2O001N101O0O2O1O0O2O1O0O2O0O2O1O0O2O0O2L4L3L5N1O2N2N1O2N1O2N2M2O2N1O2N2N1O2N2N2O1N2O1N2O1O1N2O1O1N2O1Oe_4"}, "label": "half of a sandwich with a lot of corned beef"}]}
{"id": 114060, "image": "COCO_train2014_000000114060.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich half with meat facing out\"."}], "task": "refering", "answer_template": "The \"sandwich half with meat facing out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.290921875, 0.18610416666666665, 0.9854218750000001, 0.7466458333333333], "iscrowd": 0, "area": 68660.47605, "rle": {"size": [480, 640], "counts": "gog21n>1O100O1O101N100O1O100O2O0O1O100O1O101N100O1O100O2O0O100000001O00000000001O0000000000001O000000000O2O00000000001O00000000001O000oNZOhCf0P<BPD>m;ERD<k;GSD;i;IVD9f;JXD8e;KZD6b;N\\D4b;N\\D4b;N]D3a;:RDIk;g100O10000O2O0O10000O101N100O10000O1O1O1O1gLQM`JP3_5RM_Jo2_5UM^Jl2`5XM]Ji2a5[M]Je2b5]M\\Jd2b5`M[Ja2a5[LXHX1T2^2d5fMYJ[2f5gMXJZ2h5gMVJZ2i5iMTJX2k5jMRJX2n5iMPJX2o5jMoIW2Q6kMkIW2T6kMjIV2U6lMiIU2W6lMfIV2Y6mMdIT2\\6mMbIT2]6nMaIS2^6oM_IS2a6nM]IS2b6oM\\IR2d6oMlH`2S7g2O100O100O1O100O100O1O100O100O1O100O1O100O1O100O1O10000000000O100000000000000O10000000000O10000000000O0100O10000O1000O0100O10000O10000O010O10000O100O10000O10O10O100O10000O100O10000O010O10000O100O10000O100O10000O100O10000O100O1000O0100O10000O100O10000O100O10O10O100O100O10000O100O1000O0101N10000O2O0O10001N100O101O0O100O2O000O101N10000O2O0O10001N100O101O0O101N10001N100O2O000O2O0O101O0O100O2O000O2O0O101O0O101N10001N100O101N1O101N1O1O2N1O1O2N100O2N1O1O1O001O100O1O001TLYF^2h9bMXF]2i9bMYF\\2h9dMXF\\2g9eMYFZ2h9eMZFY2g9gMYFX2h9gMYFX2h9hMYFV2g9jMZFU2g9kMYFT2h9kMZFT2g9kMYFT2P:dMPF[2Y:\\MhEc2b:TM_Ej2j:mLWER3\\;O2N1O1O2O0O1O2N1O1O2N1O1O2N1O101N100O2O001N101O0O2O1O0O2O1O0O2O0O2O1O0O2O0O2L4L3L5N1O2N2N1O2N1O2N2M2O2N1O2N2N1O2N2N2O1N2O1N2O1O1N2O1O1N2O1Oe_4"}, "label": "sandwich half with meat facing out"}]}
{"id": 368014, "image": "COCO_train2014_000000368014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fighter plane with a w on the side\"."}], "task": "refering", "answer_template": "The \"the fighter plane with a w on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 177, 191, 192, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 307, 308, 310, 311, 314, 315, 316, 337, 338], "bbox": [0.0, 0.4584232365145228, 0.964671875, 0.8674066390041495], "iscrowd": 0, "area": 49855.16229999999, "rle": {"size": [482, 640], "counts": "_92P?5K5K3M000O10000000000000000O01000000000000000O10000000O10000000O1000000000000000O010000000000000000O100000O1000000000O10000000000000O1000O10000000000000000O10O10000000000001N3N1O2N00000000000O1000O100000000000O100000O1000000000O100000O1000000000O10000000O10000000O10000000O10000000O10000000O10000000O1000000000O100000cBYO^<h0[CAc<?TCKk<W1O1O1O2N1O1O1N2O2N1O1O1O1O2N000O10000000000000000O100000000000O1000O10O1000000O100000O010000O1000000O10O01OaNgMWFW2j9kMUFT2l9mMSFR2m9PNRFo1n9SNQFl1P:UNoEi1R:YNmEf1S:gN`EY1`:d111N1O2O0O2N1102M2O1N2O1N2O2M2<D=CO1O1O1N3N1O1O1N2N2M4M2M3N0O1O100O1O1O2N1M3@`000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O101O1O001O0UDRMb;o2\\DRMd;o2ZDSMd;n2[DSM9Lo:Z3jDlLV;`3000000O100000O10O100000000O100000O10O10_Oa0K5000O100000000O1000000O100000000O100000000O100000000O100000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O10000O100O100O10000O100O100O1YFRLZ8n3aGXL^8h3]G_L`8b3\\GcLc8]3\\GfLa8[3^GgLa8Y3^GjL_8W3_GmL_8S3`GnL`8R3_GoLa8Q3^GPMb8P3]GRMb8n2\\GTMd8e40001O0O1000001O000000O100gNZGULg8k3aGkK`8V4iG_KY8`4U100000000000000O10O10000001O0O2O001O001O001O0O2O001O001O001N101O001O001O001N101O001O001O001N101O001O001O0O2O000000001O000O10001O0000000WOhDgMY;X2SE]Mn:a2^ETMb:l2j0O1O1O1O2O0O100O100O2O0O100O100O2O0O100O100O2O0O100O101N100O2O0O101N1O101N100O101N100O2O0O101N100O2O0O101N100O2O0O1O1O2O0O1O2N1O1O2N1O1O2N1O1O2N1O101N1O1O1O2N1O1O3M3M4L3M3Mha:"}, "label": "the fighter plane with a w on the side"}]}
{"id": 368014, "image": "COCO_train2014_000000368014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broken image\"."}], "task": "refering", "answer_template": "The \"broken image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 177, 191, 192, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 307, 308, 310, 311, 314, 315, 316, 337, 338], "bbox": [0.0, 0.4584232365145228, 0.964671875, 0.8674066390041495], "iscrowd": 0, "area": 49855.16229999999, "rle": {"size": [482, 640], "counts": "_92P?5K5K3M000O10000000000000000O01000000000000000O10000000O10000000O1000000000000000O010000000000000000O100000O1000000000O10000000000000O1000O10000000000000000O10O10000000000001N3N1O2N00000000000O1000O100000000000O100000O1000000000O100000O1000000000O10000000O10000000O10000000O10000000O10000000O10000000O1000000000O100000cBYO^<h0[CAc<?TCKk<W1O1O1O2N1O1O1N2O2N1O1O1O1O2N000O10000000000000000O100000000000O1000O10O1000000O100000O010000O1000000O10O01OaNgMWFW2j9kMUFT2l9mMSFR2m9PNRFo1n9SNQFl1P:UNoEi1R:YNmEf1S:gN`EY1`:d111N1O2O0O2N1102M2O1N2O1N2O2M2<D=CO1O1O1N3N1O1O1N2N2M4M2M3N0O1O100O1O1O2N1M3@`000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O101O1O001O0UDRMb;o2\\DRMd;o2ZDSMd;n2[DSM9Lo:Z3jDlLV;`3000000O100000O10O100000000O100000O10O10_Oa0K5000O100000000O1000000O100000000O100000000O100000000O100000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O10000O100O100O10000O100O100O1YFRLZ8n3aGXL^8h3]G_L`8b3\\GcLc8]3\\GfLa8[3^GgLa8Y3^GjL_8W3_GmL_8S3`GnL`8R3_GoLa8Q3^GPMb8P3]GRMb8n2\\GTMd8e40001O0O1000001O000000O100gNZGULg8k3aGkK`8V4iG_KY8`4U100000000000000O10O10000001O0O2O001O001O001O0O2O001O001O001N101O001O001O001N101O001O001O001N101O001O001O0O2O000000001O000O10001O0000000WOhDgMY;X2SE]Mn:a2^ETMb:l2j0O1O1O1O2O0O100O100O2O0O100O100O2O0O100O100O2O0O100O101N100O2O0O101N1O101N100O101N100O2O0O101N100O2O0O101N100O2O0O1O1O2O0O1O2N1O1O2N1O1O2N1O1O2N1O101N1O1O1O2N1O1O3M3M4L3M3Mha:"}, "label": "broken image"}]}
{"id": 482706, "image": "COCO_train2014_000000482706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl brushing her teeth holding a baby also brushing its teeth\"."}], "task": "refering", "answer_template": "The \"a girl brushing her teeth holding a baby also brushing its teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 25, 26, 27, 28, 29, 30, 31, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.001875, 0.013639774859287053, 0.720515625, 0.9889118198874297], "iscrowd": 0, "area": 164589.78865, "rle": {"size": [533, 640], "counts": "PQ11d`00O1nMR2fMZ2fMZ2\\Nd1M3N2[NcGoJ`8n4dGoJ^8n4gGmJ\\8P5iGlJY8R5jGjJY8S5lGiJV8T5oGgJT8V5PHgJR8W5RHdJQ8T5YHhJi7c4mHYKV7S4`IiKb6c3b3I7N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O101N100O100O100O100O1001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O00001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O010O001O001O1O001O001O001O001O001O001O1O2N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2dDSLl9^5M3M3lGVJW5m5cJZJY5i5aJ^J[5e5_JbJ]5a5]JfJ_5]5[JjJa5Y5YJmJd5U5XJoJf5S5WJPKg5R5UJQKj5Q5SJRKk5P5QJTKm5o4oITKo5n4nIUKP6m4lIWKR6k4kIXKS6j4iIZKU6h4hI[KV6g4fI]KX6e4eI^KY6a7N2N2N1O2N2N2N00001O00001O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0000O100O100O100O100O10000O100O20O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O002N2N2N2N2N2N2N2N1O1O001O001gNgCTNY<l1iCRNX<l1kCRNV<m1lCQNT<n1oCPNR<o1PDoMP<P2RDoMo;o1TDoMm;P2UDnMk;Q2XDmMi;R2YDlMg;S2\\DkMe;T2]DjMc;U2`DiMa;V2aDhM`;V2cDhM];X2eDfM\\;X2gDfMY;Z2iDdMX;Z2kDdMU;n0[C>X>\\OnAb0T>XOQBg0P>UOTBk0m=oNXBP1i=kN]BT1b>O100O2O2M5L4K5L4K5L4K4Mofl2"}, "label": "a girl brushing her teeth holding a baby also brushing its teeth"}]}
{"id": 482706, "image": "COCO_train2014_000000482706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with blonde hair bushing teeth\"."}], "task": "refering", "answer_template": "The \"woman with blonde hair bushing teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 25, 26, 27, 28, 29, 30, 31, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429], "bbox": [0.001875, 0.013639774859287053, 0.720515625, 0.9889118198874297], "iscrowd": 0, "area": 164589.78865, "rle": {"size": [533, 640], "counts": "PQ11d`00O1nMR2fMZ2fMZ2\\Nd1M3N2[NcGoJ`8n4dGoJ^8n4gGmJ\\8P5iGlJY8R5jGjJY8S5lGiJV8T5oGgJT8V5PHgJR8W5RHdJQ8T5YHhJi7c4mHYKV7S4`IiKb6c3b3I7N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O101N100O100O100O100O1001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O00001O00001O00001O001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O001O010O001O001O1O001O001O001O001O001O001O1O2N2N2N3M2N2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2N2N2N2N3M2dDSLl9^5M3M3lGVJW5m5cJZJY5i5aJ^J[5e5_JbJ]5a5]JfJ_5]5[JjJa5Y5YJmJd5U5XJoJf5S5WJPKg5R5UJQKj5Q5SJRKk5P5QJTKm5o4oITKo5n4nIUKP6m4lIWKR6k4kIXKS6j4iIZKU6h4hI[KV6g4fI]KX6e4eI^KY6a7N2N2N1O2N2N2N00001O00001O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0000O100O100O100O100O10000O100O20O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O002N2N2N2N2N2N2N2N1O1O001O001gNgCTNY<l1iCRNX<l1kCRNV<m1lCQNT<n1oCPNR<o1PDoMP<P2RDoMo;o1TDoMm;P2UDnMk;Q2XDmMi;R2YDlMg;S2\\DkMe;T2]DjMc;U2`DiMa;V2aDhM`;V2cDhM];X2eDfM\\;X2gDfMY;Z2iDdMX;Z2kDdMU;n0[C>X>\\OnAb0T>XOQBg0P>UOTBk0m=oNXBP1i=kN]BT1b>O100O2O2M5L4K5L4K5L4K4Mofl2"}, "label": "woman with blonde hair bushing teeth"}]}
{"id": 482706, "image": "COCO_train2014_000000482706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chaild was doing rush\"."}], "task": "refering", "answer_template": "The \"a chaild was doing rush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389, 406, 407, 408, 409, 410, 411, 412, 429, 430, 431, 432, 433, 434, 435], "bbox": [0.46764062500000003, 0.1523452157598499, 0.9800000000000001, 0.9867917448405253], "iscrowd": 0, "area": 76925.78315, "rle": {"size": [533, 640], "counts": "\\lk41c`03N3L4M2M4M2M10O01O100O100O010O10000O10O12N3M2N2N0O01000000000O010000000000O01000000000O10O10000000O1000O101O001O1O0010O0100O0010O01O010O1O010O010O00100O0010O01O010O10O01O010O00100O010O0010O01O10O01O010O010O1O010O0010O01O10O010O01O0jBdNT;\\1eDkNZ;V1`DoN`;Q1[DUOe;k0UD[Oj;e0oCCP<>gCKY<4_C4a<MWC;i<i1O1O1SF[Li6e3UI^Li6b3VIaLh6`3VIcLi6]3UIeLj6\\3TIgLd6`3ZIcL]6d3bI_LU6i3iIYLo5o3oITLY4WOcJl4R1PLl3HeJ_4]1lK_39gJR4h1gKS3k0gJe3U2cKk2i6SMZIj2h6TM[Ii2g6UM[Ii2g6UM\\Ih2f6VM]Ig2d6XM]Ig2e6WM\\Ih2b7ZL_He3d7XL]Hg3g7UL[Hi3h7TLYHk3k7QLVHn3n7nKSHQ4P8lKQHS4S8iKnGV4U8gKlGX4Y:O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O1O100O100O1O100O101N1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1001O0000001O0000001O0000001O00000010O000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001ZHQMd1o2VNXMi1h2UN\\Mi1d2VN^Mi1c2VN^Mi1b2WN_Mh1a2WNaMh1_2XNbMg1_2XNbMg1^2XNdMg1\\2YNeMf1[2ZNfMe1[2YNhMe1X2[NiMd1W2\\NjMd1U2[NmMd1T2[NmMd1S2[NoMd1Q2[NQNd1o1\\NRNc1o1[NSNd1m1[NUN^OUOPMg2a3VNoND]MW2b3WNlNJ^MQ2c3WNjNO`Mk1d3[NdN1eMf1e3]N^N5iM`1f3`NYN7nMZ1g3dNSN9SNU1h3eNnM>VNn0j3iNhM`0[Ni0j3lNcMb0`Nd0k3nN]Mf0dN=m3ROWMh0iN8m3UORMk0mN1o3YOlLm0ROLP4[OfLS5W3QKbLT5\\3QK\\LU5b3oJVLV5h3n3M3N1N3M2N3N2M2J7@?_Ob0^Ob0^Oa0_O\\d6"}, "label": "a chaild was doing rush"}]}
{"id": 482706, "image": "COCO_train2014_000000482706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby with toothbrush in its mouth and being carried by its mother who is brushing her teeth\"."}], "task": "refering", "answer_template": "The \"a baby with toothbrush in its mouth and being carried by its mother who is brushing her teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389, 406, 407, 408, 409, 410, 411, 412, 429, 430, 431, 432, 433, 434, 435], "bbox": [0.46764062500000003, 0.1523452157598499, 0.9800000000000001, 0.9867917448405253], "iscrowd": 0, "area": 76925.78315, "rle": {"size": [533, 640], "counts": "\\lk41c`03N3L4M2M4M2M10O01O100O100O010O10000O10O12N3M2N2N0O01000000000O010000000000O01000000000O10O10000000O1000O101O001O1O0010O0100O0010O01O010O1O010O010O00100O0010O01O010O10O01O010O00100O010O0010O01O10O01O010O010O1O010O0010O01O10O010O01O0jBdNT;\\1eDkNZ;V1`DoN`;Q1[DUOe;k0UD[Oj;e0oCCP<>gCKY<4_C4a<MWC;i<i1O1O1SF[Li6e3UI^Li6b3VIaLh6`3VIcLi6]3UIeLj6\\3TIgLd6`3ZIcL]6d3bI_LU6i3iIYLo5o3oITLY4WOcJl4R1PLl3HeJ_4]1lK_39gJR4h1gKS3k0gJe3U2cKk2i6SMZIj2h6TM[Ii2g6UM[Ii2g6UM\\Ih2f6VM]Ig2d6XM]Ig2e6WM\\Ih2b7ZL_He3d7XL]Hg3g7UL[Hi3h7TLYHk3k7QLVHn3n7nKSHQ4P8lKQHS4S8iKnGV4U8gKlGX4Y:O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O1O100O100O1O100O101N1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1001O0000001O0000001O0000001O00000010O000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001ZHQMd1o2VNXMi1h2UN\\Mi1d2VN^Mi1c2VN^Mi1b2WN_Mh1a2WNaMh1_2XNbMg1_2XNbMg1^2XNdMg1\\2YNeMf1[2ZNfMe1[2YNhMe1X2[NiMd1W2\\NjMd1U2[NmMd1T2[NmMd1S2[NoMd1Q2[NQNd1o1\\NRNc1o1[NSNd1m1[NUN^OUOPMg2a3VNoND]MW2b3WNlNJ^MQ2c3WNjNO`Mk1d3[NdN1eMf1e3]N^N5iM`1f3`NYN7nMZ1g3dNSN9SNU1h3eNnM>VNn0j3iNhM`0[Ni0j3lNcMb0`Nd0k3nN]Mf0dN=m3ROWMh0iN8m3UORMk0mN1o3YOlLm0ROLP4[OfLS5W3QKbLT5\\3QK\\LU5b3oJVLV5h3n3M3N1N3M2N3N2M2J7@?_Ob0^Ob0^Oa0_O\\d6"}, "label": "a baby with toothbrush in its mouth and being carried by its mother who is brushing her teeth"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman touching a lapop wearing a white tank top under a black shirt\"."}], "task": "refering", "answer_template": "The \"a woman touching a lapop wearing a white tank top under a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 340, 341, 342], "bbox": [0.643703125, 0.27725352112676055, 0.91934375, 0.9690140845070423], "iscrowd": 0, "area": 29499.076950000002, "rle": {"size": [426, 640], "counts": "da[5n0[<;E;D9bH[Nh2h1fLhNY3\\1TLTOk3o0cKA\\4c0QKLo48^J8a5KTJ>k5FoI=P6FkI<U6HeI;Z6HbI:]6J]I9b6X3O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1000001N100000001N100000004K6K1O1O1O1N2O1O1O1O1N2O1O2N1O1N3N1O2N1O2NEYKTHe4f7hKVHV4c7m0G9H7K6O001O001O1O001O001O1O0kKaI_1`6^NcIa1]6]NfIb1[6[NhId1Y6[NhId1X6[NjId1W6[NjId1V6[NlId1U6[NlId1U6[NlId1T6[NnId1S6[NnId1S6ZNoIe1R6ZNoIf1P6YNRJf1o5YNRJf1o5YNRJf1n5YNTJf1m5YNTJf1m5WNVJh1k5TNYJk1g5RN\\Jn1`0`L[3^1YLQ2:cL^3W1\\LV24eL`3R1_LZ2NfLd3l0bL]2HiLg3f0dLa2ClLi3>hLf2]OnLl38kLi2VORMR4OkLo2QOTMV4HlLT3lNVM[4_OmL[3fNXMo6g2oH[MQ7e2kH_MU7a2dHgM\\7X2\\HPNd7Q2SHWNm7c31O0000010O00001O000010O0000PLYH]2h7bM[H[2e7eM^HX2b7hMbHT2_7kMdHS2[7mMiHo1W7QNlHl1U7SNoHi1Q7VNSIg1m6YNWIc1i6]NZI`1g6_N\\I^1e6aN_I\\1a6cNbIZ1^6fNfIV1[6iNhIT1Y6kNkIQ1V6mNnIP1R6PORJl0o5SOTJk0l5TOWJi0j5VOZJf0g5YO\\Jd0d5\\O`J`0a5_ObJ>_5AeJ;\\5ChJ;W5ElJ8U5GoJ5R5JQK3P5LTK0l40WKMj42ZKJg45\\KId46`KF`49Z4O1O1O002N2N1O2Mg[e0"}, "label": "a woman touching a lapop wearing a white tank top under a black shirt"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with black shirt using a silver laptop\"."}], "task": "refering", "answer_template": "The \"a woman with black shirt using a silver laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 340, 341, 342], "bbox": [0.643703125, 0.27725352112676055, 0.91934375, 0.9690140845070423], "iscrowd": 0, "area": 29499.076950000002, "rle": {"size": [426, 640], "counts": "da[5n0[<;E;D9bH[Nh2h1fLhNY3\\1TLTOk3o0cKA\\4c0QKLo48^J8a5KTJ>k5FoI=P6FkI<U6HeI;Z6HbI:]6J]I9b6X3O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1000001N100000001N100000004K6K1O1O1O1N2O1O1O1O1N2O1O2N1O1N3N1O2N1O2NEYKTHe4f7hKVHV4c7m0G9H7K6O001O001O1O001O001O1O0kKaI_1`6^NcIa1]6]NfIb1[6[NhId1Y6[NhId1X6[NjId1W6[NjId1V6[NlId1U6[NlId1U6[NlId1T6[NnId1S6[NnId1S6ZNoIe1R6ZNoIf1P6YNRJf1o5YNRJf1o5YNRJf1n5YNTJf1m5YNTJf1m5WNVJh1k5TNYJk1g5RN\\Jn1`0`L[3^1YLQ2:cL^3W1\\LV24eL`3R1_LZ2NfLd3l0bL]2HiLg3f0dLa2ClLi3>hLf2]OnLl38kLi2VORMR4OkLo2QOTMV4HlLT3lNVM[4_OmL[3fNXMo6g2oH[MQ7e2kH_MU7a2dHgM\\7X2\\HPNd7Q2SHWNm7c31O0000010O00001O000010O0000PLYH]2h7bM[H[2e7eM^HX2b7hMbHT2_7kMdHS2[7mMiHo1W7QNlHl1U7SNoHi1Q7VNSIg1m6YNWIc1i6]NZI`1g6_N\\I^1e6aN_I\\1a6cNbIZ1^6fNfIV1[6iNhIT1Y6kNkIQ1V6mNnIP1R6PORJl0o5SOTJk0l5TOWJi0j5VOZJf0g5YO\\Jd0d5\\O`J`0a5_ObJ>_5AeJ;\\5ChJ;W5ElJ8U5GoJ5R5JQK3P5LTK0l40WKMj42ZKJg45\\KId46`KF`49Z4O1O1O002N2N1O2Mg[e0"}, "label": "a woman with black shirt using a silver laptop"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with large round earrings\"."}], "task": "refering", "answer_template": "The \"a woman with large round earrings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 217, 234, 235, 240, 241, 256, 257, 258, 259, 263, 264, 280, 281, 282, 287], "bbox": [0.16928125, 0.34075117370892016, 0.5259218750000001, 0.869131455399061], "iscrowd": 0, "area": 18064.409799999994, "rle": {"size": [426, 640], "counts": "XX]18n<7I6K6I6J4L5K4L4M3L4L4L4N2N2O1O1O1O1O1O2N1O1O1O1O1O1O100O1O1O1O100O2N1O2O1N3M2N2N2O0O1O1O2O0O1J6IQM`Fb2Y9dMlFX2m8oMVGo1T8eNQHW1_7YOdHd0Y7^OlH>Q7ESI7j6LYI1d61aIK]67fIGV6<nI@o5c0TJZOi5i0[JSOb5P1aJmN\\5V1hJfNU5^1mJ`NQ5a1SK[Nl4f1WKWNi4i1[KSNd4n1_KoM`4R2dKjM[4W2hKgMV4Z2nKbMR4_2PL^Mo3c2ULYMj3h2ZLTMe3m2^LPMa3Q3cLlL\\3T3gLiLX3X3lLdLS3]3\\32N100O1O100000001O000000000000001O0000000O100000001O00000001O000010O0001O00001O0001O01O00001O2N1O100O1O2N1O1O1O1O2O0O1O1_OnEhMS:P2TFPNn9g1YFYNh9]1bFbN_9U1iFkNX9k0QGVOo8a0YG_OZ;O00001O0000010O00001O00000010O000001O00000010O1O1O102M3M3M3N2M3M3M3M3N2M3M3O1N2O1N1O2O1N2O1N2N2O1N2O1N2O1N2N2O1N2ON1O2N2N3M2N3M2N3M3M2N3M2N3M2N3M3M2L5L3L5L3L5L4K4Mdfm3"}, "label": "a woman with large round earrings"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a woman with blonde hear , and she is wearing earrings\"."}], "task": "refering", "answer_template": "The \"there is a woman with blonde hear , and she is wearing earrings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 217, 234, 235, 240, 241, 256, 257, 258, 259, 263, 264, 280, 281, 282, 287], "bbox": [0.16928125, 0.34075117370892016, 0.5259218750000001, 0.869131455399061], "iscrowd": 0, "area": 18064.409799999994, "rle": {"size": [426, 640], "counts": "XX]18n<7I6K6I6J4L5K4L4M3L4L4L4N2N2O1O1O1O1O1O2N1O1O1O1O1O1O100O1O1O1O100O2N1O2O1N3M2N2N2O0O1O1O2O0O1J6IQM`Fb2Y9dMlFX2m8oMVGo1T8eNQHW1_7YOdHd0Y7^OlH>Q7ESI7j6LYI1d61aIK]67fIGV6<nI@o5c0TJZOi5i0[JSOb5P1aJmN\\5V1hJfNU5^1mJ`NQ5a1SK[Nl4f1WKWNi4i1[KSNd4n1_KoM`4R2dKjM[4W2hKgMV4Z2nKbMR4_2PL^Mo3c2ULYMj3h2ZLTMe3m2^LPMa3Q3cLlL\\3T3gLiLX3X3lLdLS3]3\\32N100O1O100000001O000000000000001O0000000O100000001O00000001O000010O0001O00001O0001O01O00001O2N1O100O1O2N1O1O1O1O2O0O1O1_OnEhMS:P2TFPNn9g1YFYNh9]1bFbN_9U1iFkNX9k0QGVOo8a0YG_OZ;O00001O0000010O00001O00000010O000001O00000010O1O1O102M3M3M3N2M3M3M3M3N2M3M3O1N2O1N1O2O1N2O1N2N2O1N2O1N2O1N2N2O1N2ON1O2N2N3M2N3M2N3M3M2N3M2N3M2N3M3M2L5L3L5L3L5L4K4Mdfm3"}, "label": "there is a woman with blonde hear , and she is wearing earrings"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the rightmost laptop\"."}], "task": "refering", "answer_template": "The \"the back of the rightmost laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.84553125, 0.6497652582159624, 0.9999375, 0.9853521126760563], "iscrowd": 0, "area": 10580.974800000005, "rle": {"size": [426, 640], "counts": "kZQ71Y=4K4M3L5L3L4L4M4L3M3M4L3M3M3M4M2M3M3M4L3M3M4L3M3M3M4M2M3M3M4L3M3M4L3M3M3N3L3M3M3M4L3M0000000001O0000000000000000001O0O1000000000000000001O0000000000000000001O000000000000000000001O00000000000YOg0dN\\1dN_H"}, "label": "the back of the rightmost laptop"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back side of laptop computer , far right of picture\"."}], "task": "refering", "answer_template": "The \"the back side of laptop computer , far right of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.84553125, 0.6497652582159624, 0.9999375, 0.9853521126760563], "iscrowd": 0, "area": 10580.974800000005, "rle": {"size": [426, 640], "counts": "kZQ71Y=4K4M3L5L3L4L4M4L3M3M4L3M3M3M4M2M3M3M4L3M3M4L3M3M3M4M2M3M3M4L3M3M4L3M3M3N3L3M3M3M4L3M0000000001O0000000000000000001O0O1000000000000000001O0000000000000000001O000000000000000000001O00000000000YOg0dN\\1dN_H"}, "label": "the back side of laptop computer , far right of picture"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the woman on the left is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the woman on the left is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 290, 312, 313], "bbox": [0.3940625, 0.6033098591549295, 0.62253125, 0.8853990610328639], "iscrowd": 0, "area": 6957.615900000004, "rle": {"size": [426, 640], "counts": "T`Y31Y=1O001O001O001O00001N101O1O001O001O1O010O001O1O0010O01O001O10O01O001O1O010O001O1O0010O01O1O001O010O1O001O1O100O001O1O100O1O1O3M3M3N2M3M3Md0\\O1O2O0O2N1O2N1O2O0O2N1O2N2N2O0O001O001O01O0001O00010O0000001O01O0001O000010O000001O0001O012M3M2O2M2N2N3N1N2N3M2O2M2N2O2M3M4L4M3L5K4M3L4L5L3L4L4M4K4L4MnYT3"}, "label": "the chair the woman on the left is sitting in"}]}
{"id": 286116, "image": "COCO_train2014_000000286116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown jacket hanging on the chair in which a woman with white earrings is sitting\"."}], "task": "refering", "answer_template": "The \"a brown jacket hanging on the chair in which a woman with white earrings is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 290, 312, 313], "bbox": [0.3940625, 0.6033098591549295, 0.62253125, 0.8853990610328639], "iscrowd": 0, "area": 6957.615900000004, "rle": {"size": [426, 640], "counts": "T`Y31Y=1O001O001O001O00001N101O1O001O001O1O010O001O1O0010O01O001O10O01O001O1O010O001O1O0010O01O1O001O010O1O001O1O100O001O1O100O1O1O3M3M3N2M3M3Md0\\O1O2O0O2N1O2N1O2O0O2N1O2N2N2O0O001O001O01O0001O00010O0000001O01O0001O000010O000001O0001O012M3M2O2M2N2N3N1N2N3M2O2M2N2O2M3M4L4M3L5K4M3L4L5L3L4L4M4K4L4MnYT3"}, "label": "a brown jacket hanging on the chair in which a woman with white earrings is sitting"}]}
{"id": 474545, "image": "COCO_train2014_000000474545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zeebra following another zeebra\"."}], "task": "refering", "answer_template": "The \"the zeebra following another zeebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236], "bbox": [0.09796875000000001, 0.40654166666666663, 0.35134375, 0.6385208333333333], "iscrowd": 0, "area": 7967.53545, "rle": {"size": [480, 640], "counts": "Yhm02j>5L3N2O2N1O1O20O1O100O00VCGn:8lDOT;D\\C;Y17\\;J^D=`;D[Db0d;_OUDh0i;YOTDk0j;VOUDl0i;VOUD;_O@\\<8SD6G[OY<a0mC3b<0[COf<3YCKh<8UCHl<:QCFo<=oBBQ=V100010O0001O011O6J4L5K4L2N1O1OiNnCAQ<<ZD\\Of;d0`DWO_;j0bDUO\\;l0fDROZ;o0fDPO[;o0gDoNY;R1hDlNX;T1jDiNW;X1jDfNV;Z1kDeNU;\\1lDbNT;_1mD_NS;a1V11O010O1O001O00000000O2O00O10O10000O1000000O100000O01000000O10000O100000ZClN`;S1YDUOf;l0RD]Om;c0kCEU<c101N@lCVNT<h1QDUNo;h1VDWNi;g1\\DVNc;h1bDVN^;h1gDVNX;g1Q1O200000O3N^OmBYOQ=g0QCZOk<g0XCXOf<h0\\CXOa<i0aCXO\\<h0fCXOW<i0kCWOS<i0PDWOl;j0VDVOh;i0[D\\O];d0gDYOY;e0kDZOS;f0QEWOo:h0UEVOi:i0\\ESOe:l0_EoNb:Q1aEkNa:T1e1O2O001N101O001N3N1001O2N100O2N1O1O100O2N1O11N3M4L3M4K4M3M4L3L5L3M3LdYR6"}, "label": "the zeebra following another zeebra"}]}
{"id": 474545, "image": "COCO_train2014_000000474545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra to the left of two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra to the left of two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236], "bbox": [0.09796875000000001, 0.40654166666666663, 0.35134375, 0.6385208333333333], "iscrowd": 0, "area": 7967.53545, "rle": {"size": [480, 640], "counts": "Yhm02j>5L3N2O2N1O1O20O1O100O00VCGn:8lDOT;D\\C;Y17\\;J^D=`;D[Db0d;_OUDh0i;YOTDk0j;VOUDl0i;VOUD;_O@\\<8SD6G[OY<a0mC3b<0[COf<3YCKh<8UCHl<:QCFo<=oBBQ=V100010O0001O011O6J4L5K4L2N1O1OiNnCAQ<<ZD\\Of;d0`DWO_;j0bDUO\\;l0fDROZ;o0fDPO[;o0gDoNY;R1hDlNX;T1jDiNW;X1jDfNV;Z1kDeNU;\\1lDbNT;_1mD_NS;a1V11O010O1O001O00000000O2O00O10O10000O1000000O100000O01000000O10000O100000ZClN`;S1YDUOf;l0RD]Om;c0kCEU<c101N@lCVNT<h1QDUNo;h1VDWNi;g1\\DVNc;h1bDVN^;h1gDVNX;g1Q1O200000O3N^OmBYOQ=g0QCZOk<g0XCXOf<h0\\CXOa<i0aCXO\\<h0fCXOW<i0kCWOS<i0PDWOl;j0VDVOh;i0[D\\O];d0gDYOY;e0kDZOS;f0QEWOo:h0UEVOi:i0\\ESOe:l0_EoNb:Q1aEkNa:T1e1O2O001N101O001N3N1001O2N100O2N1O1O100O2N1O11N3M4L3M4K4M3M4L3L5L3M3LdYR6"}, "label": "a zebra to the left of two other zebras"}]}
{"id": 474545, "image": "COCO_train2014_000000474545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this zebra is to the right of two others\"."}], "task": "refering", "answer_template": "The \"this zebra is to the right of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 223, 225, 226], "bbox": [0.61575, 0.3453541666666667, 0.83040625, 0.5874583333333334], "iscrowd": 0, "area": 5884.552900000002, "rle": {"size": [480, 640], "counts": "Zlh53i>5K4L4O1N2O2M2O1O1O0O2O1O1O001O1O0O2O1N1010O100O0100011N4M1N3M13M2N1N10000O101N100O2N2O0O2O1N2O0O2O1N101bC_NZ;a1cDaN^;`1_DaNa;c1ZD^Ng;j1iC\\NY<U2000001O00O1iN\\DQOe;o0V100001O00000000001O001O000000001O0000000000000000O100000000O100O1O100O1O100O1O100O1O10000O12N2N2N2O0bCgNS;[1jDeNW;\\1gDeNY;]1dDdN\\;_1_DcNa;`1[D`Ng;b1TD`Nl;c1oC_NQ<f1iCZNX<V22M2O1N3F9LbC[Nj;e0eC2>]Ol;`0PDJ1Io;<VDGE2V<4[D?f;_O_D=c;_OcD=_;^OgD=];]OjD=kWc1"}, "label": "this zebra is to the right of two others"}]}
{"id": 474545, "image": "COCO_train2014_000000474545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra facing left\"."}], "task": "refering", "answer_template": "The \"a zebra facing left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 199, 200, 201, 202, 223, 225, 226], "bbox": [0.61575, 0.3453541666666667, 0.83040625, 0.5874583333333334], "iscrowd": 0, "area": 5884.552900000002, "rle": {"size": [480, 640], "counts": "Zlh53i>5K4L4O1N2O2M2O1O1O0O2O1O1O001O1O0O2O1N1010O100O0100011N4M1N3M13M2N1N10000O101N100O2N2O0O2O1N2O0O2O1N101bC_NZ;a1cDaN^;`1_DaNa;c1ZD^Ng;j1iC\\NY<U2000001O00O1iN\\DQOe;o0V100001O00000000001O001O000000001O0000000000000000O100000000O100O1O100O1O100O1O100O1O10000O12N2N2N2O0bCgNS;[1jDeNW;\\1gDeNY;]1dDdN\\;_1_DcNa;`1[D`Ng;b1TD`Nl;c1oC_NQ<f1iCZNX<V22M2O1N3F9LbC[Nj;e0eC2>]Ol;`0PDJ1Io;<VDGE2V<4[D?f;_O_D=c;_OcD=_;^OgD=];]OjD=kWc1"}, "label": "a zebra facing left"}]}
{"id": 499122, "image": "COCO_train2014_000000499122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chocolate glazed donut\"."}], "task": "refering", "answer_template": "The \"a chocolate glazed donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285], "bbox": [0.229734375, 0.26239583333333333, 0.43918750000000006, 0.7578750000000001], "iscrowd": 0, "area": 19423.074650000002, "rle": {"size": [480, 640], "counts": "VTU29[>a0_Oa0B>H7J7H8I7H6K5K4K6K4K6J6I6K6J5J7J5K6I7I6J6L3O2N1N2O2N1O2M2O1O2M2O2N1N2O2N1O1O0O2O1O001O1O0O2O1O001O1N101O1O001N2O001O1O0O2O1O001O10O01O1O0010O000000001O0kLfHeNY7=WHhMd0j1U7=ZHdMf0m1Q7<\\HdMg0n1m6=^HcMh0n1j6>aHbMf0o1i6>cHaMg0P2e6=gHbMf0o1d6=iHaMe0Q2b6=kHaMe0P2`6>mH`Mf0P2]6>QI`Md0P2[6?SI_Mg0o1V6`0mJ]OS5b0QK[Oo4c0VKZOj4e0ZKXOf4f0_KWOb4f0cKWO]4g0hKWOW4h0lKVOT4h0oKWOR4f0RLXOn3g0TLXOm3e0WLYOj3f0WLZOh3e0[LYOf3e0\\LZOe3d0^LZOb3e0`LZOa3d0bLZO_3d0cL[O^3c0eL\\OZ3c0hL\\OY3c0iL[OX3c0jL[OW3d0kL[OV3c0kL]OV3a0lL^OU3`0mL^OT3a0nL^OS3`0oL_OR3?PM@P3?RM_OP3=SMCn28WMGj25ZMJf23^MKd21`MNa2NcM1]2NdM2]2LeM2f8O1O1OjmW5"}, "label": "a chocolate glazed donut"}]}
{"id": 499122, "image": "COCO_train2014_000000499122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the darkest donut shown\"."}], "task": "refering", "answer_template": "The \"the darkest donut shown\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285], "bbox": [0.229734375, 0.26239583333333333, 0.43918750000000006, 0.7578750000000001], "iscrowd": 0, "area": 19423.074650000002, "rle": {"size": [480, 640], "counts": "VTU29[>a0_Oa0B>H7J7H8I7H6K5K4K6K4K6J6I6K6J5J7J5K6I7I6J6L3O2N1N2O2N1O2M2O1O2M2O2N1N2O2N1O1O0O2O1O001O1O0O2O1O001O1N101O1O001N2O001O1O0O2O1O001O10O01O1O0010O000000001O0kLfHeNY7=WHhMd0j1U7=ZHdMf0m1Q7<\\HdMg0n1m6=^HcMh0n1j6>aHbMf0o1i6>cHaMg0P2e6=gHbMf0o1d6=iHaMe0Q2b6=kHaMe0P2`6>mH`Mf0P2]6>QI`Md0P2[6?SI_Mg0o1V6`0mJ]OS5b0QK[Oo4c0VKZOj4e0ZKXOf4f0_KWOb4f0cKWO]4g0hKWOW4h0lKVOT4h0oKWOR4f0RLXOn3g0TLXOm3e0WLYOj3f0WLZOh3e0[LYOf3e0\\LZOe3d0^LZOb3e0`LZOa3d0bLZO_3d0cL[O^3c0eL\\OZ3c0hL\\OY3c0iL[OX3c0jL[OW3d0kL[OV3c0kL]OV3a0lL^OU3`0mL^OT3a0nL^OS3`0oL_OR3?PM@P3?RM_OP3=SMCn28WMGj25ZMJf23^MKd21`MNa2NcM1]2NdM2]2LeM2f8O1O1OjmW5"}, "label": "the darkest donut shown"}]}
{"id": 499122, "image": "COCO_train2014_000000499122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glazed donut with a hole in it\"."}], "task": "refering", "answer_template": "The \"a glazed donut with a hole in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 289, 290, 291, 292], "bbox": [0.519234375, 0.24008333333333332, 0.746421875, 0.7566041666666667], "iscrowd": 0, "area": 26066.597650000014, "rle": {"size": [480, 640], "counts": "Pjk41i>>C:E8H8I7I7L5K6J6VCaNi;l2^O6J5K6J5K6J5M3L3M4L3N3L3M4M2M4L3M4L4L3M2N3L4M3M2O2M3N2M3N1N3N1N2O0O2O1N2O0O2O1O001O1O1O001O1O1O001N101O1O001O00100O010O01O001O0010O01O1O001O010O001O2N1O1O1O1N2O2N1O1O1O1N2O2N1O1O3M2M4M2N3M2N3M2M4M2N3M2N3M3L3N3M3L5L3L4M4K4M4K4M4K4M4L3L5L3L5J5L5J5L4K5L4K5L4L5J5L4I7I:Fc0]Ob0^O_d[2"}, "label": "a glazed donut with a hole in it"}]}
{"id": 499122, "image": "COCO_train2014_000000499122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far right donut\"."}], "task": "refering", "answer_template": "The \"the far right donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 289, 290, 291, 292], "bbox": [0.519234375, 0.24008333333333332, 0.746421875, 0.7566041666666667], "iscrowd": 0, "area": 26066.597650000014, "rle": {"size": [480, 640], "counts": "Pjk41i>>C:E8H8I7I7L5K6J6VCaNi;l2^O6J5K6J5K6J5M3L3M4L3N3L3M4M2M4L3M4L4L3M2N3L4M3M2O2M3N2M3N1N3N1N2O0O2O1N2O0O2O1O001O1O1O001O1O1O001N101O1O001O00100O010O01O001O0010O01O1O001O010O001O2N1O1O1O1N2O2N1O1O1O1N2O2N1O1O3M2M4M2N3M2N3M2M4M2N3M2N3M3L3N3M3L5L3L4M4K4M4K4M4K4M4L3L5L3L5J5L5J5L4K5L4K5L4L5J5L4I7I:Fc0]Ob0^O_d[2"}, "label": "the far right donut"}]}
{"id": 359865, "image": "COCO_train2014_000000359865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half of a sandwich on the far right side of a blue plate\"."}], "task": "refering", "answer_template": "The \"a half of a sandwich on the far right side of a blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 199, 200, 201, 202, 222], "bbox": [0.510671875, 0.11461111111111111, 0.867140625, 0.701111111111111], "iscrowd": 0, "area": 23748.228200000005, "rle": {"size": [360, 640], "counts": "URc33R;3M4L4N1O2N1O1O1O1O001O001O001O010O10O010O11O0O2O00001O0O2O001O001O0O2O001O001O0O101O001O0O2O002N2N2M3N2N2M3N2N2M4M2N2M3N2N2M3N2N2M3N2N2M2O2N2M3N2N2N2M3M3M2UOcLhI`3V6cLeIa3X6cLdI`3Z6dLaI_3]6dL_I_3_6g0N2N2N7Hg0ZO7I101N101N1O2O0O2O0O2O0O2OO001O1O1O1N101O1O1O1O001O1O1N101O1O1O1O001O1N2O1O1O1O1O1O1O1N101O1O1O1O1O1O1N2N2N1O2N2N2N2N2N2N2O100O100O100O100O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N1N3N2M3N2N2M2O2M3N2N2M2J7N2O1O1O0O2O1O1O1L4L3L5N2N2N2O0O2N2O1N2N2N101N2N3CUFXOn9e0=M3M2N3M3M4L3M4LPgm0"}, "label": "a half of a sandwich on the far right side of a blue plate"}]}
{"id": 359865, "image": "COCO_train2014_000000359865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich half in the center of the plate\"."}], "task": "refering", "answer_template": "The \"the sandwich half in the center of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.36828125, 0.2816111111111111, 0.663390625, 0.9878888888888888], "iscrowd": 0, "area": 39448.92845000001, "rle": {"size": [360, 640], "counts": "nSc22k3c0i2@aL\\1V3gNTLV2c3kMgKR3P4QMeKa3Q4bLhKl3n3WLlKU4k3nKoK[4k3gKoK_4o3dKkKa4S4bKfKd4X4_KbKf4\\4T1O1N2N2N2N2O1000000000000010O00001O001O00001O001O001O001O00001O001O00O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N200001O1O1O1O1O1O1O1O00001O001O000010O01O00001O001O00001O00000000O100O100O10000O100O100O100O1001O001O00001O001O00001O001O001O00001O001O00001O2N2N2N2N2N2N2N2O1N2N2N2N2N2M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3L6K5K5I7I6J7I7I7H8I6J7I7I7G9E;E:F;E;E;EoZ[2"}, "label": "the sandwich half in the center of the plate"}]}
{"id": 359865, "image": "COCO_train2014_000000359865.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hoggie sandwich that is closest to purple bowl\"."}], "task": "refering", "answer_template": "The \"hoggie sandwich that is closest to purple bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.36828125, 0.2816111111111111, 0.663390625, 0.9878888888888888], "iscrowd": 0, "area": 39448.92845000001, "rle": {"size": [360, 640], "counts": "nSc22k3c0i2@aL\\1V3gNTLV2c3kMgKR3P4QMeKa3Q4bLhKl3n3WLlKU4k3nKoK[4k3gKoK_4o3dKkKa4S4bKfKd4X4_KbKf4\\4T1O1N2N2N2N2O1000000000000010O00001O001O00001O001O001O001O00001O001O00O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N200001O1O1O1O1O1O1O1O00001O001O000010O01O00001O001O00001O00000000O100O100O10000O100O100O100O1001O001O00001O001O00001O001O001O00001O001O00001O2N2N2N2N2N2N2N2O1N2N2N2N2N2M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3L6K5K5I7I6J7I7I7H8I6J7I7I7G9E;E:F;E;E;EoZ[2"}, "label": "hoggie sandwich that is closest to purple bowl"}]}
{"id": 73146, "image": "COCO_train2014_000000073146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train cover in artistic graffiti\"."}], "task": "refering", "answer_template": "The \"a train cover in artistic graffiti\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167], "bbox": [0.00078125, 0.3204225352112676, 0.47329687500000006, 0.5054460093896713], "iscrowd": 0, "area": 18171.21595, "rle": {"size": [426, 640], "counts": "ca0^2l:00000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000001O00000000000O10000000000000000000000O1000000000000000000000000O100000000000000000001O1N2O1O1O1O1O1O00000000000O10000000000000000000O1O010O1O1O1000000000000000O10000000000000000000000O1000001O000000000000000O10000000000000000000001N10000000000000000000000O100000000000001O0000000O10000000000000000000000O100000001O0000000000000O10000000000000000000000O2O000000000000000000000O10000000000000001O00000O1000000000000000000000000O1000001O000000000000000O1000000000001O1O001O001N101O001O00001O001O006Jc0\\OfP\\4"}, "label": "a train cover in artistic graffiti"}]}
{"id": 73146, "image": "COCO_train2014_000000073146.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train painted with different scenes\"."}], "task": "refering", "answer_template": "The \"train painted with different scenes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167], "bbox": [0.00078125, 0.3204225352112676, 0.47329687500000006, 0.5054460093896713], "iscrowd": 0, "area": 18171.21595, "rle": {"size": [426, 640], "counts": "ca0^2l:00000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000001O00000000000O10000000000000000000000O1000000000000000000000000O100000000000000000001O1N2O1O1O1O1O1O00000000000O10000000000000000000O1O010O1O1O1000000000000000O10000000000000000000000O1000001O000000000000000O10000000000000000000001N10000000000000000000000O100000000000001O0000000O10000000000000000000000O100000001O0000000000000O10000000000000000000000O2O000000000000000000000O10000000000000001O00000O1000000000000000000000000O1000001O000000000000000O1000000000001O1O001O001N101O001O00001O001O006Jc0\\OfP\\4"}, "label": "train painted with different scenes"}]}
{"id": 515518, "image": "COCO_train2014_000000515518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"happy to see the fruit cake at the dining table\"."}], "task": "refering", "answer_template": "The \"happy to see the fruit cake at the dining table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 34, 35, 36, 37, 38, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 231, 232, 233, 234, 235, 236, 237], "bbox": [0.08100702576112413, 0.082203125, 0.8101873536299767, 0.697078125], "iscrowd": 0, "area": 74362.58840000001, "rle": {"size": [640, 427], "counts": "bUf0U1hb03N3M3N101O001O001O010O001O00000010O000000B^N^^Oc1aa0^N^^Ob1ba0_N]^Oa1ca0`N]^O_1ca0bN\\^O^1da0cN[^O]1da0eN\\^OZ1da0gN[^OY1ea0hN[^OW1a1aNY>9U@V1b1bNX>9V@T1a1eNX>8V@S1b1fNW>8W@Q1a1hNX>8V@P1a1iNY>8V@n0a1jNY>9U@l0b1lNX>:V@i0a1nNY>:U@h0b1nNY>;U@e0b1QOY>;T@d0b1ROZ>;T@b0b1SOZ><S@`0c1UOZ><S@>b1VO\\>d1dA[N]>d1cA\\N^>d1aA\\N_>e1aAZN`>e1`A\\N`>d1_A\\Nb>d1^A[N^>i1bAWN[>m1dASNY>P2hAoMU>U2jAlMR>X2mAjMm=Z2TBfMh=^2WBcMe=a2ZB`Me<`3[CaLc<a3\\C`La<c3]C_La<b3_C_L^<d3`C^L^<d3aC]L\\<e3cC^LZ<d3eC]LX<e3gC]LW<d3iC]LS<f3lC\\Ln;i3QDYLh;m3WDULc;P4]DQL];T4bDnKV:^OUEj4d0jKS:Z5mEhJm9\\5RFfJj9]5UFeJf9c5UF_Jg9i5SFYJi9P6oESJn9T6lEnIS:P8N2N3N1N2O1N2N2O2M2N2O1N2O1O1O2N100O1O1O2N1O1O101N1O1O1O2N100O1O101O00000O2O0000001O0O10001O00000O101O0000000O10000000001O00000000000000001N100000000000001O000000001O000000001O0000001O000000001O00001O001O001O001O001O001O001O01O1O1O010XDiGe;X8[DiGc;[8OM^DiGa;X8_DiG_;]800OKbDjG\\;_8N2N2N2N2O1N2N2N2N3M5L5J5K5K6J7I6J6K7H9G9G5K4L3M4L3M3M4L6J9G001O1N101O1O001O1O001O1O001O1N2O1O001O1O1O1O00001N101O0O2O001O0O2O000O2O001N3N2N2oJjCS2Y<gMlCX2V<bMoC]2S<TM[Dk2^>0000O1000000O10000O10001N100O2O000O2O0WNY_O7i`0Af_O7Z`0Fk_O7W`0Em_O:S`0CR@;o?AU@=m?_OX@?j?\\OZ@?l?\\OY@;ka0IjVb1"}, "label": "happy to see the fruit cake at the dining table"}]}
{"id": 515518, "image": "COCO_train2014_000000515518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian woman sitting at a table\"."}], "task": "refering", "answer_template": "The \"an asian woman sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 34, 35, 36, 37, 38, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 231, 232, 233, 234, 235, 236, 237], "bbox": [0.08100702576112413, 0.082203125, 0.8101873536299767, 0.697078125], "iscrowd": 0, "area": 74362.58840000001, "rle": {"size": [640, 427], "counts": "bUf0U1hb03N3M3N101O001O001O010O001O00000010O000000B^N^^Oc1aa0^N^^Ob1ba0_N]^Oa1ca0`N]^O_1ca0bN\\^O^1da0cN[^O]1da0eN\\^OZ1da0gN[^OY1ea0hN[^OW1a1aNY>9U@V1b1bNX>9V@T1a1eNX>8V@S1b1fNW>8W@Q1a1hNX>8V@P1a1iNY>8V@n0a1jNY>9U@l0b1lNX>:V@i0a1nNY>:U@h0b1nNY>;U@e0b1QOY>;T@d0b1ROZ>;T@b0b1SOZ><S@`0c1UOZ><S@>b1VO\\>d1dA[N]>d1cA\\N^>d1aA\\N_>e1aAZN`>e1`A\\N`>d1_A\\Nb>d1^A[N^>i1bAWN[>m1dASNY>P2hAoMU>U2jAlMR>X2mAjMm=Z2TBfMh=^2WBcMe=a2ZB`Me<`3[CaLc<a3\\C`La<c3]C_La<b3_C_L^<d3`C^L^<d3aC]L\\<e3cC^LZ<d3eC]LX<e3gC]LW<d3iC]LS<f3lC\\Ln;i3QDYLh;m3WDULc;P4]DQL];T4bDnKV:^OUEj4d0jKS:Z5mEhJm9\\5RFfJj9]5UFeJf9c5UF_Jg9i5SFYJi9P6oESJn9T6lEnIS:P8N2N3N1N2O1N2N2O2M2N2O1N2O1O1O2N100O1O1O2N1O1O101N1O1O1O2N100O1O101O00000O2O0000001O0O10001O00000O101O0000000O10000000001O00000000000000001N100000000000001O000000001O000000001O0000001O000000001O00001O001O001O001O001O001O001O01O1O1O010XDiGe;X8[DiGc;[8OM^DiGa;X8_DiG_;]800OKbDjG\\;_8N2N2N2N2O1N2N2N2N3M5L5J5K5K6J7I6J6K7H9G9G5K4L3M4L3M3M4L6J9G001O1N101O1O001O1O001O1O001O1N2O1O001O1O1O1O00001N101O0O2O001O0O2O000O2O001N3N2N2oJjCS2Y<gMlCX2V<bMoC]2S<TM[Dk2^>0000O1000000O10000O10001N100O2O000O2O0WNY_O7i`0Af_O7Z`0Fk_O7W`0Em_O:S`0CR@;o?AU@=m?_OX@?j?\\OZ@?l?\\OY@;ka0IjVb1"}, "label": "an asian woman sitting at a table"}]}
{"id": 515518, "image": "COCO_train2014_000000515518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of 3 - layer cake with frosting and various fruits on top\"."}], "task": "refering", "answer_template": "The \"a piece of 3 - layer cake with frosting and various fruits on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 339, 340], "bbox": [0.5839578454332552, 0.729875, 0.965480093676815, 0.97403125], "iscrowd": 0, "area": 18402.8548, "rle": {"size": [640, 427], "counts": "RUl41jc08H7H9H7I8G9H7I8G8I8H7I8G8I8H8M2O2M2N3M2N3M3M2O2M2N3M2N3M3N0O0000O10001O0O1000000O2O000O1000000O2O000O1000001N1000000O101O0O1000000O2O000O101O000O2O00001N10000O2O00001N10001O0O10001N10001O0O101O0O10001O0O101O000O2O000O101O000O2O000O2O001O0O2O001O0O2O001N101O001N101O0O2O001O0O2O001O0O2O001N101O001N101O001N101O0O2F;C<E;D<D=DTl8"}, "label": "a piece of 3 - layer cake with frosting and various fruits on top"}]}
{"id": 515518, "image": "COCO_train2014_000000515518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mixed fruit cake is on the plate with spoon and fork\"."}], "task": "refering", "answer_template": "The \"mixed fruit cake is on the plate with spoon and fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 339, 340], "bbox": [0.5839578454332552, 0.729875, 0.965480093676815, 0.97403125], "iscrowd": 0, "area": 18402.8548, "rle": {"size": [640, 427], "counts": "RUl41jc08H7H9H7I8G9H7I8G8I8H7I8G8I8H8M2O2M2N3M2N3M3M2O2M2N3M2N3M3N0O0000O10001O0O1000000O2O000O1000000O2O000O1000001N1000000O101O0O1000000O2O000O101O000O2O00001N10000O2O00001N10001O0O10001N10001O0O101O0O10001O0O101O000O2O000O101O000O2O000O2O001O0O2O001O0O2O001N101O001N101O0O2O001O0O2O001O0O2O001N101O001N101O001N101O0O2F;C<E;D<D=DTl8"}, "label": "mixed fruit cake is on the plate with spoon and fork"}]}
{"id": 376258, "image": "COCO_train2014_000000376258.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black smudge in the top left corner\"."}], "task": "refering", "answer_template": "The \"a black smudge in the top left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 26, 27, 28, 46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186], "bbox": [0.0012968749999999999, 0.0, 0.24210937500000002, 0.3861875], "iscrowd": 0, "area": 15595.263049999996, "rle": {"size": [640, 640], "counts": "Rd0T6l=m0SO1O1O1O001O1O1O1O1O1O1O1O1O1O00001O0000001O000000001O0000001OhHeCk6[<UIgCi6Y<WIjCf6V<ZIlCd6T<\\IoCb6P<^ISD_6m;aIWD[6i;dI\\DX6d;hI`DU6_;kIeDQ6[;oInEh4R:WKiFo3W9QLcGU3]8jLiGR3V8nLlGP3T8PMnGn2R8QMQHm2o7SMSHk2m7TMWHi2i7WMYHf2h7YM[He2e7[M]Hc2c7\\M`Hb2`7^MaH`2`7`MbH^2^7cMbH\\2^7eMcHY2]7hMdHV2\\7kMdHS2]7nMdHP2\\7QNeHm1[7TNeHk1[7WNeHf1\\7[NeHc1[7^NeHa1[7`NfH^1Z7cNfH\\1Z7eNgHX1Z7iNgHU1Y7lNgHS1Y7nNhHP1X7POjHm0W7TOjHj0V7WOmHe0S7[OPI`0R7API8T7InH0V71lHGY79kH_OY7b0iHWO[7j0m5001O1O001O1O001O1O000000001O00000000001O000000001O00000000001O00000000000000001O00000000000000000000000000001O00000000O1O1O1O1O100OOPOT]Oo0nb001O2M3@\\T_9"}, "label": "a black smudge in the top left corner"}]}
{"id": 376258, "image": "COCO_train2014_000000376258.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s leg only visible\"."}], "task": "refering", "answer_template": "The \"a man ' s leg only visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 26, 27, 28, 46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186], "bbox": [0.0012968749999999999, 0.0, 0.24210937500000002, 0.3861875], "iscrowd": 0, "area": 15595.263049999996, "rle": {"size": [640, 640], "counts": "Rd0T6l=m0SO1O1O1O001O1O1O1O1O1O1O1O1O1O00001O0000001O000000001O0000001OhHeCk6[<UIgCi6Y<WIjCf6V<ZIlCd6T<\\IoCb6P<^ISD_6m;aIWD[6i;dI\\DX6d;hI`DU6_;kIeDQ6[;oInEh4R:WKiFo3W9QLcGU3]8jLiGR3V8nLlGP3T8PMnGn2R8QMQHm2o7SMSHk2m7TMWHi2i7WMYHf2h7YM[He2e7[M]Hc2c7\\M`Hb2`7^MaH`2`7`MbH^2^7cMbH\\2^7eMcHY2]7hMdHV2\\7kMdHS2]7nMdHP2\\7QNeHm1[7TNeHk1[7WNeHf1\\7[NeHc1[7^NeHa1[7`NfH^1Z7cNfH\\1Z7eNgHX1Z7iNgHU1Y7lNgHS1Y7nNhHP1X7POjHm0W7TOjHj0V7WOmHe0S7[OPI`0R7API8T7InH0V71lHGY79kH_OY7b0iHWO[7j0m5001O1O001O1O001O1O000000001O00000000001O000000001O00000000001O00000000000000001O00000000000000000000000000001O00000000O1O1O1O1O100OOPOT]Oo0nb001O2M3@\\T_9"}, "label": "a man ' s leg only visible"}]}
{"id": 376258, "image": "COCO_train2014_000000376258.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an arm holding a pizza\"."}], "task": "refering", "answer_template": "The \"an arm holding a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 274, 275], "bbox": [0.46275000000000005, 0.25225, 1.0, 0.50734375], "iscrowd": 0, "area": 25720.408600000002, "rle": {"size": [640, 640], "counts": "PYi51oc01N1O2O1N2N2O1N1O2O0O1O100O1O100O1O100O1O100O1O100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100O1O010O1O10O00010O00001O01O01O000010O00013L4L4M3L2O000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O100O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O2O00000000000O1000000000000O10000000000O100O10000O100O100O10000O100O10000O100O100O2O001N2O1N101O1N2O0O2O1N2O001N2O1N101O0O100O101N10000O2O0O10000O2O0O100O101O0O100O2O000O100O2O000O101N100O10001N100O10001N100O101N10000O101N10000O2O0O100O101O0O100O101O0O100O2O0O10001N101N2O1O0nJ"}, "label": "an arm holding a pizza"}]}
{"id": 376258, "image": "COCO_train2014_000000376258.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white mans arm holding a pizza\"."}], "task": "refering", "answer_template": "The \"a white mans arm holding a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 274, 275], "bbox": [0.46275000000000005, 0.25225, 1.0, 0.50734375], "iscrowd": 0, "area": 25720.408600000002, "rle": {"size": [640, 640], "counts": "PYi51oc01N1O2O1N2N2O1N1O2O0O1O100O1O100O1O100O1O100O1O100O1O010O1O10O01O100O00100O1O010O1O10O01O100O00100O1O010O1O10O00010O00001O01O01O000010O00013L4L4M3L2O000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O100O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O2O00000000000O1000000000000O10000000000O100O10000O100O100O10000O100O10000O100O100O2O001N2O1N101O1N2O0O2O1N2O001N2O1N101O0O100O101N10000O2O0O10000O2O0O100O101O0O100O2O000O100O2O000O101N100O10001N100O10001N100O101N10000O101N10000O2O0O100O101O0O100O101O0O100O2O0O10001N101N2O1O0nJ"}, "label": "a white mans arm holding a pizza"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby bull with huge horns\"."}], "task": "refering", "answer_template": "The \"a baby bull with huge horns\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 40, 61, 63, 64, 84, 85, 86, 87, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 155, 156, 157, 158], "bbox": [0.66871875, 0.14642857142857144, 0.8984687499999999, 0.8056302521008403], "iscrowd": 0, "area": 7642.2067, "rle": {"size": [238, 640], "counts": "baS34U77J6I7I7K5K5J6K5j0VO2M3M1O1O00001O010O00000000_J@S4`0iKGT49jKIV47hKKX44gKNY42fKOZ40eK1\\4ObK3^4LaK5`4KUK?l4@RKb0n4_OoJb0S5]OkJe0U5\\OhJf0Y5ZOdJh0\\5XObJj0^5WO_Jk0b5a0O10O01O100oJUN^4j1bKWN_4h1_KZNb4e1]K\\Nd4a1\\KaNe4]1ZKcNf4]1[KcNe4]1ZKcNg4\\1YKeNf4\\1ZKdNf4\\1YKdNh4_1TKbNk4b1RK]No4f1mJ[NS5P2O100PKfMk4=SK^11WNl4:TKd1k4]NUKc1k4\\NVKc1j4]NWKc1Q3^NUNNkNd1m2hNPNDTOd1i2oNoM\\OXOf1f2TOnMVO]Of1b2[OmMoNAf1_2UNlMe0O@Ge1\\2QNaNb0TOL8Y1[3mNULMa0V1[3^OfLb0[3]OeLc0\\3[OeLd0]3ZOcLg0a3TO`Ll0d3oN]LQ1f3lNZLS1k3hNVLX1m3dNSL]1S4\\NnKd1n400O010000O01000O1000O101O001O010O00001O001O001O00000001O0000001O0000000001O001O1O001O1O1O001O1O1O001N2O001N2O1O1N2O2N1M3M3M3Ldn>"}, "label": "a baby bull with huge horns"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long horn bull sitting on the right side of a larger bull\"."}], "task": "refering", "answer_template": "The \"the long horn bull sitting on the right side of a larger bull\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 40, 61, 63, 64, 84, 85, 86, 87, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 155, 156, 157, 158], "bbox": [0.66871875, 0.14642857142857144, 0.8984687499999999, 0.8056302521008403], "iscrowd": 0, "area": 7642.2067, "rle": {"size": [238, 640], "counts": "baS34U77J6I7I7K5K5J6K5j0VO2M3M1O1O00001O010O00000000_J@S4`0iKGT49jKIV47hKKX44gKNY42fKOZ40eK1\\4ObK3^4LaK5`4KUK?l4@RKb0n4_OoJb0S5]OkJe0U5\\OhJf0Y5ZOdJh0\\5XObJj0^5WO_Jk0b5a0O10O01O100oJUN^4j1bKWN_4h1_KZNb4e1]K\\Nd4a1\\KaNe4]1ZKcNf4]1[KcNe4]1ZKcNg4\\1YKeNf4\\1ZKdNf4\\1YKdNh4_1TKbNk4b1RK]No4f1mJ[NS5P2O100PKfMk4=SK^11WNl4:TKd1k4]NUKc1k4\\NVKc1j4]NWKc1Q3^NUNNkNd1m2hNPNDTOd1i2oNoM\\OXOf1f2TOnMVO]Of1b2[OmMoNAf1_2UNlMe0O@Ge1\\2QNaNb0TOL8Y1[3mNULMa0V1[3^OfLb0[3]OeLc0\\3[OeLd0]3ZOcLg0a3TO`Ll0d3oN]LQ1f3lNZLS1k3hNVLX1m3dNSL]1S4\\NnKd1n400O010000O01000O1000O101O001O010O00001O001O001O00000001O0000001O0000000001O001O1O001O1O1O001O1O1O001N2O001N2O1O1N2O2N1M3M3M3Ldn>"}, "label": "the long horn bull sitting on the right side of a larger bull"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a forest buffalo sitting separately near by another two\"."}], "task": "refering", "answer_template": "The \"a forest buffalo sitting separately near by another two\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 28, 47, 48, 50, 51, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122], "bbox": [0.056765625, 0.1815546218487395, 0.341796875, 0.7113445378151261], "iscrowd": 0, "area": 9745.35935, "rle": {"size": [238, 640], "counts": "n]81Z76K5J7I4L4M2N3M3M]OhI5V6DUJ9i5BbJ;Z6N01000O010O010O0010O010O010XIFZ6:fIGZ69eIG[69eIH[68dIH]67cIJ\\66cIK\\1J^3;UKLZ1Ka39UKLY1Lc37TKMW1Ne36SKLW10f33SKMW10f34RKK:2GL65g41RKL2g01]Ok40RKLOm01WOo41oJJ0`1P5gNoJa1Q5`NoJ_1R5>001O00001O001N101O000000000001N101N1O1O100O10QNQK_1o4`NRK`1m4`NUK_1k4`NVK`1i4aNWK_1i4`NYK_1g4aNYK_1f4bNZK^1f4aN\\K^1d4bN\\K^1c4bN_K]1a4cN_K]1a4cN_K]1`4cNaK]1_4cNaK]1_4cN`K^1`4aNaK_1^4bNbK^1^4aNcK_1]4aNcK_1\\4bNdK^1\\4aNeK_1[4aNeK_1Z4aNfK`1Z4`NfK`1Z4`NfK`1Z4_NgKa1X4`NhK`1X4_NiKa1W4_NiKa1V4`NjK`1U4`NlK`1T4`NlK`1S4aNmK_1R4aNoK_1Q4`NPL`1o3aNQL_1o3`NRL`1m3`NTL`1k3`NVL`1i3aNWL_1h3aNYL_1f3aN[L_1d3aN]L_1b3bN^L^1a3bN`L^1_3aNcL_1[3bNfL^1Y3bNhL^1V3cNkL]1S3cNoL]1o2cNSM]1k2aNYM_1a2dNbM\\1[2cNiM]1T4000000001O0000001O000001O01O000001O000O2O0O101O000010O000001O1O1O1O1O1O001O1O1O1O001O001O0O2O0O2O001N2N2N2NlNSOjKk0T4ZOgKh0Z4\\OWKg0Q58cJTOa5?kJ^OX56SKIP5KZK3V6JffQ3"}, "label": "a forest buffalo sitting separately near by another two"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ox relaxing on the ground away from other two oxens\"."}], "task": "refering", "answer_template": "The \"an ox relaxing on the ground away from other two oxens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 28, 47, 48, 50, 51, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122], "bbox": [0.056765625, 0.1815546218487395, 0.341796875, 0.7113445378151261], "iscrowd": 0, "area": 9745.35935, "rle": {"size": [238, 640], "counts": "n]81Z76K5J7I4L4M2N3M3M]OhI5V6DUJ9i5BbJ;Z6N01000O010O010O0010O010O010XIFZ6:fIGZ69eIG[69eIH[68dIH]67cIJ\\66cIK\\1J^3;UKLZ1Ka39UKLY1Lc37TKMW1Ne36SKLW10f33SKMW10f34RKK:2GL65g41RKL2g01]Ok40RKLOm01WOo41oJJ0`1P5gNoJa1Q5`NoJ_1R5>001O00001O001N101O000000000001N101N1O1O100O10QNQK_1o4`NRK`1m4`NUK_1k4`NVK`1i4aNWK_1i4`NYK_1g4aNYK_1f4bNZK^1f4aN\\K^1d4bN\\K^1c4bN_K]1a4cN_K]1a4cN_K]1`4cNaK]1_4cNaK]1_4cN`K^1`4aNaK_1^4bNbK^1^4aNcK_1]4aNcK_1\\4bNdK^1\\4aNeK_1[4aNeK_1Z4aNfK`1Z4`NfK`1Z4`NfK`1Z4_NgKa1X4`NhK`1X4_NiKa1W4_NiKa1V4`NjK`1U4`NlK`1T4`NlK`1S4aNmK_1R4aNoK_1Q4`NPL`1o3aNQL_1o3`NRL`1m3`NTL`1k3`NVL`1i3aNWL_1h3aNYL_1f3aN[L_1d3aN]L_1b3bN^L^1a3bN`L^1_3aNcL_1[3bNfL^1Y3bNhL^1V3cNkL]1S3cNoL]1o2cNSM]1k2aNYM_1a2dNbM\\1[2cNiM]1T4000000001O0000001O000001O01O000001O000O2O0O101O000010O000001O1O1O1O1O1O001O1O1O1O001O001O0O2O0O2O001N2N2N2NlNSOjKk0T4ZOgKh0Z4\\OWKg0Q58cJTOa5?kJ^OX56SKIP5KZK3V6JffQ3"}, "label": "an ox relaxing on the ground away from other two oxens"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow on most left\"."}], "task": "refering", "answer_template": "The \"a cow on most left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 28, 47, 48, 50, 51, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122], "bbox": [0.056765625, 0.1815546218487395, 0.341796875, 0.7113445378151261], "iscrowd": 0, "area": 9745.35935, "rle": {"size": [238, 640], "counts": "n]81Z76K5J7I4L4M2N3M3M]OhI5V6DUJ9i5BbJ;Z6N01000O010O010O0010O010O010XIFZ6:fIGZ69eIG[69eIH[68dIH]67cIJ\\66cIK\\1J^3;UKLZ1Ka39UKLY1Lc37TKMW1Ne36SKLW10f33SKMW10f34RKK:2GL65g41RKL2g01]Ok40RKLOm01WOo41oJJ0`1P5gNoJa1Q5`NoJ_1R5>001O00001O001N101O000000000001N101N1O1O100O10QNQK_1o4`NRK`1m4`NUK_1k4`NVK`1i4aNWK_1i4`NYK_1g4aNYK_1f4bNZK^1f4aN\\K^1d4bN\\K^1c4bN_K]1a4cN_K]1a4cN_K]1`4cNaK]1_4cNaK]1_4cN`K^1`4aNaK_1^4bNbK^1^4aNcK_1]4aNcK_1\\4bNdK^1\\4aNeK_1[4aNeK_1Z4aNfK`1Z4`NfK`1Z4`NfK`1Z4_NgKa1X4`NhK`1X4_NiKa1W4_NiKa1V4`NjK`1U4`NlK`1T4`NlK`1S4aNmK_1R4aNoK_1Q4`NPL`1o3aNQL_1o3`NRL`1m3`NTL`1k3`NVL`1i3aNWL_1h3aNYL_1f3aN[L_1d3aN]L_1b3bN^L^1a3bN`L^1_3aNcL_1[3bNfL^1Y3bNhL^1V3cNkL]1S3cNoL]1o2cNSM]1k2aNYM_1a2dNbM\\1[2cNiM]1T4000000001O0000001O000001O01O000001O000O2O0O101O000010O000001O1O1O1O1O1O001O1O1O1O001O001O0O2O0O2O001N2N2N2NlNSOjKk0T4ZOgKh0Z4\\OWKg0Q58cJTOa5?kJ^OX56SKIP5KZK3V6JffQ3"}, "label": "a cow on most left"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the animal with the longest , thickest horns\"."}], "task": "refering", "answer_template": "The \"the animal with the longest , thickest horns\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 37, 40, 60, 62, 63, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131], "bbox": [0.44503125, 0.04689075630252101, 0.76321875, 0.7461764705882353], "iscrowd": 0, "area": 11830.325100000002, "rle": {"size": [238, 640], "counts": "_\\R22[73L4M3L5L3M3L4M2M3N1O2M2O2M2O2O0O2N2N1O100O1O001O10O01O10O10O101O0O1000000O101OO010000O1000O10000000O101O001O001O0000O1000000O100000000000O100000O1000O100000K_J]N`5d1`J\\N`5d1`J\\N`5d1aJ\\N^5d1bJ\\N^5d1cJ[N\\5f1dJ[N[5e1fJZNZ5f1fJZNZ5f1fJ[NY5e1gJ[NX5f1hJZNX5f1gJ[NY5e1gJ\\NX5d1hJ[NX5f1hJZNX5f1hJZNX5f1hJZNX5f1hJZNX5f1hJZNW5g1iJYNW5h1hJXNX5h1hJWNY5i1gJWNY5j1fJVNY5k1hJTNX5l1iJSNW5m1iJSNW5n1iJQNV5P2jJPNV5P24001N100VORNTLn1\\3cNcL^1V3iNiLW1R3oNlLR1a2QNVMR19m0e2^OZMa0j2]OUMc0n2[OPMf0U3VOjLj0_3nN`LR1h3fNXLZ1n3aNQL_1Q4`NmKa1U4^NjKb1X4]NgKc1[4\\NdKd1^4ZNaKg1a4XN_Kg1c4WN]Ki1c4WN^Kh1c4XN\\Kh1d4XN]Kg1c4ZN\\Kf1e4YN\\Kf1d4ZN\\Kf1d4[N\\Kd1d4\\N\\Kd1e4[N\\Kd1d4]N[Kc1e4]N\\Kb1d4^N\\Kb1e4^N[Ka1e4_N[Ka1e4_NZKb1g4^NWKc1k4[NTKe1o4YNPKh1Q593N2M3N1N2O1N2O1N2O1O0O2OYOYNhKZ1[OnNl4KiKT1@nNe41kKn0EmN^4:lKf0c4\\O\\Kb0e4@[K?d4B]K=b4E]K;a4H_K7`4K_K5`4M`K2_40aK@m4b0RK]On4d0SKYOm4j0RKUOn4m0RKPOo4R1PKmNP5f0cJA=GP5h0fJ@j5`0ZJ_Od5a0`J\\O^5e0eJZOX5f0kJZOR5g0PKYOm4g0VKYOg4h0\\KWOb4h0aKXO\\4i0fKVOX4j0kKVOR4k0QLTOo22_Lm0e0QOn27nKm0W1lNl2a1c1L5K4M3L5K4MbYS1"}, "label": "the animal with the longest , thickest horns"}]}
{"id": 163267, "image": "COCO_train2014_000000163267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large bull with white horns laying on the ground between two smaller bulls\"."}], "task": "refering", "answer_template": "The \"a large bull with white horns laying on the ground between two smaller bulls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 37, 40, 60, 62, 63, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131], "bbox": [0.44503125, 0.04689075630252101, 0.76321875, 0.7461764705882353], "iscrowd": 0, "area": 11830.325100000002, "rle": {"size": [238, 640], "counts": "_\\R22[73L4M3L5L3M3L4M2M3N1O2M2O2M2O2O0O2N2N1O100O1O001O10O01O10O10O101O0O1000000O101OO010000O1000O10000000O101O001O001O0000O1000000O100000000000O100000O1000O100000K_J]N`5d1`J\\N`5d1`J\\N`5d1aJ\\N^5d1bJ\\N^5d1cJ[N\\5f1dJ[N[5e1fJZNZ5f1fJZNZ5f1fJ[NY5e1gJ[NX5f1hJZNX5f1gJ[NY5e1gJ\\NX5d1hJ[NX5f1hJZNX5f1hJZNX5f1hJZNX5f1hJZNX5f1hJZNW5g1iJYNW5h1hJXNX5h1hJWNY5i1gJWNY5j1fJVNY5k1hJTNX5l1iJSNW5m1iJSNW5n1iJQNV5P2jJPNV5P24001N100VORNTLn1\\3cNcL^1V3iNiLW1R3oNlLR1a2QNVMR19m0e2^OZMa0j2]OUMc0n2[OPMf0U3VOjLj0_3nN`LR1h3fNXLZ1n3aNQL_1Q4`NmKa1U4^NjKb1X4]NgKc1[4\\NdKd1^4ZNaKg1a4XN_Kg1c4WN]Ki1c4WN^Kh1c4XN\\Kh1d4XN]Kg1c4ZN\\Kf1e4YN\\Kf1d4ZN\\Kf1d4[N\\Kd1d4\\N\\Kd1e4[N\\Kd1d4]N[Kc1e4]N\\Kb1d4^N\\Kb1e4^N[Ka1e4_N[Ka1e4_NZKb1g4^NWKc1k4[NTKe1o4YNPKh1Q593N2M3N1N2O1N2O1N2O1O0O2OYOYNhKZ1[OnNl4KiKT1@nNe41kKn0EmN^4:lKf0c4\\O\\Kb0e4@[K?d4B]K=b4E]K;a4H_K7`4K_K5`4M`K2_40aK@m4b0RK]On4d0SKYOm4j0RKUOn4m0RKPOo4R1PKmNP5f0cJA=GP5h0fJ@j5`0ZJ_Od5a0`J\\O^5e0eJZOX5f0kJZOR5g0PKYOm4g0VKYOg4h0\\KWOb4h0aKXO\\4i0fKVOX4j0kKVOR4k0QLTOo22_Lm0e0QOn27nKm0W1lNl2a1c1L5K4M3L5K4MbYS1"}, "label": "a large bull with white horns laying on the ground between two smaller bulls"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in the middle\"."}], "task": "refering", "answer_template": "The \"man in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 289, 290, 291, 310, 312, 313, 333, 336], "bbox": [0.41792187500000005, 0.2246586345381526, 0.7008750000000001, 0.8089156626506023], "iscrowd": 0, "area": 29820.163100000005, "rle": {"size": [498, 640], "counts": "lTR42_?`0@`0@3mMRO]EP1^:YO\\Eh0`:@[Eb0`:F[E;a:NYE4b:4XENd:;VEGe:a0VE@f:i0TEYOg:o0TEROh:W1REkNi:Q3O2O0O2O0O2O1N101N101O0O2O0O2O0O2O0O2O001N101eFgJZ8Z5]GPKb8Q5TGXKk8i5O1N2O001N2O1WNSJjIo5T6SJkIm5U6UJhIm5V6UJiIk5V6WJhIh5Z6YJeIb3]O\\Mo6UOaI`3BWMn6[O_I^3ESMn6@\\I]3GPMn6FXI[3LkLm6LVIY3OgLm61RIY32cLm67oHV36_Ll6=lHU39[Lm6a0iHT3<WLl6g0fHS3`0RLk6i5VITJl6l5VIPJk6Q6WIkIj6V6WIgIj6[6WIaIk6_6WI]Ij6d6WIYIj6h6P1000001O00001O00001O00001O01O01N10001N10000O2O0O101O1N4M2N3L4M3M3L3RKZFl3i9mK\\FR4g9fK^FY4f9_K_Fa4Z:N101O0O2O001OmEdKP9\\4nFhKQ9W4mFnKP9R4nFRLQ9m3mFXLP9h3mF]LR9b3lFcLQ9]3mFgLQ9Y3mFlLQ9S3lFRMR9n2lFWMR9h2lF\\MR9d2lFaMR9^2lFfMR9^40O001O001O002N2N2N2N2N2N2hN`FZLc9c3`FZLb9c3bFZL`9c3dFZL^9c3fFZL\\9d3gFYL[9d3jFXLX9e3lFXLV9e3nFXLU9e3nFXLT9e3PGYLQ9d3SGYLo8d3b1N2M3M6J;F:E;E?A2O1N1O2N2O1N2N1O2O1N2N1O2O1N2N2N101F:E;F^cl2"}, "label": "man in the middle"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shortest player crouched in the middle with cupped hands\"."}], "task": "refering", "answer_template": "The \"the shortest player crouched in the middle with cupped hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 289, 290, 291, 310, 312, 313, 333, 336], "bbox": [0.41792187500000005, 0.2246586345381526, 0.7008750000000001, 0.8089156626506023], "iscrowd": 0, "area": 29820.163100000005, "rle": {"size": [498, 640], "counts": "lTR42_?`0@`0@3mMRO]EP1^:YO\\Eh0`:@[Eb0`:F[E;a:NYE4b:4XENd:;VEGe:a0VE@f:i0TEYOg:o0TEROh:W1REkNi:Q3O2O0O2O0O2O1N101N101O0O2O0O2O0O2O0O2O001N101eFgJZ8Z5]GPKb8Q5TGXKk8i5O1N2O001N2O1WNSJjIo5T6SJkIm5U6UJhIm5V6UJiIk5V6WJhIh5Z6YJeIb3]O\\Mo6UOaI`3BWMn6[O_I^3ESMn6@\\I]3GPMn6FXI[3LkLm6LVIY3OgLm61RIY32cLm67oHV36_Ll6=lHU39[Lm6a0iHT3<WLl6g0fHS3`0RLk6i5VITJl6l5VIPJk6Q6WIkIj6V6WIgIj6[6WIaIk6_6WI]Ij6d6WIYIj6h6P1000001O00001O00001O00001O01O01N10001N10000O2O0O101O1N4M2N3L4M3M3L3RKZFl3i9mK\\FR4g9fK^FY4f9_K_Fa4Z:N101O0O2O001OmEdKP9\\4nFhKQ9W4mFnKP9R4nFRLQ9m3mFXLP9h3mF]LR9b3lFcLQ9]3mFgLQ9Y3mFlLQ9S3lFRMR9n2lFWMR9h2lF\\MR9d2lFaMR9^2lFfMR9^40O001O001O002N2N2N2N2N2N2hN`FZLc9c3`FZLb9c3bFZL`9c3dFZL^9c3fFZL\\9d3gFYL[9d3jFXLX9e3lFXLV9e3nFXLU9e3nFXLT9e3PGYLQ9d3SGYLo8d3b1N2M3M6J;F:E;E?A2O1N1O2N2O1N2N1O2O1N2N1O2O1N2N2N101F:E;F^cl2"}, "label": "the shortest player crouched in the middle with cupped hands"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ballplayer on far right\"."}], "task": "refering", "answer_template": "The \"ballplayer on far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 319, 320, 342, 343], "bbox": [0.6955, 0.1485542168674699, 0.930515625, 0.825281124497992], "iscrowd": 0, "area": 33424.65885000001, "rle": {"size": [498, 640], "counts": "fch6o0]>;F9F:F9G8I7I8K5L5J6J<EV1iN;E:F=D=B5L4L5K4M3L5K3N1N2N2O1N1O2O1N100O1O100O2N1O1O1O1_OXHWIi7d6e0M3M3M3M2N3N2N2O1O1O1O1O1N101O1O1O1O10N101O1O1O1O001RNjI_JW6]5oI`JR6[5VJWJT6f5RJXJn5c5YJ\\Jf5a5_J]Ja5c5bJZJ]5g5eJXJZ5i5gJUJY5k5jJSJU5n5lJPJU5o5mJoIS5Q6PKlIP5U6QKfIR5Z6QK`IS5_6oJ[IU5f6k1010O00010O001O10O01001O001O001O0010d0\\O:FO1N3N7H9H7H9Hj0VO3L4M3L4M3L3N0O100O100O100O10O001^JPI`2P7^MUI_2l6^MYI_2i6^MZI`2g6^M\\I`2e6]M^Ib2c6\\M_Ic2d6XM_Ig2n6kLUIS3l6kLVIT3o6TKeJk4Q801_NmEaMT:R2\\FjMe9\\O`Eo1\\1aNT9]OfEc1d1nNY9c0UG[On8=ZG_Oj89]GEe86aGFc85aGHd81aGMc8MUWe0"}, "label": "ballplayer on far right"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a baseball outfit crouching to the right of a man who is squatting\"."}], "task": "refering", "answer_template": "The \"a man wearing a baseball outfit crouching to the right of a man who is squatting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 319, 320, 342, 343], "bbox": [0.6955, 0.1485542168674699, 0.930515625, 0.825281124497992], "iscrowd": 0, "area": 33424.65885000001, "rle": {"size": [498, 640], "counts": "fch6o0]>;F9F:F9G8I7I8K5L5J6J<EV1iN;E:F=D=B5L4L5K4M3L5K3N1N2N2O1N1O2O1N100O1O100O2N1O1O1O1_OXHWIi7d6e0M3M3M3M2N3N2N2O1O1O1O1O1N101O1O1O1O10N101O1O1O1O001RNjI_JW6]5oI`JR6[5VJWJT6f5RJXJn5c5YJ\\Jf5a5_J]Ja5c5bJZJ]5g5eJXJZ5i5gJUJY5k5jJSJU5n5lJPJU5o5mJoIS5Q6PKlIP5U6QKfIR5Z6QK`IS5_6oJ[IU5f6k1010O00010O001O10O01001O001O001O0010d0\\O:FO1N3N7H9H7H9Hj0VO3L4M3L4M3L3N0O100O100O100O10O001^JPI`2P7^MUI_2l6^MYI_2i6^MZI`2g6^M\\I`2e6]M^Ib2c6\\M_Ic2d6XM_Ig2n6kLUIS3l6kLVIT3o6TKeJk4Q801_NmEaMT:R2\\FjMe9\\O`Eo1\\1aNT9]OfEc1d1nNY9c0UG[On8=ZG_Oj89]GEe86aGFc85aGHd81aGMc8MUWe0"}, "label": "a man wearing a baseball outfit crouching to the right of a man who is squatting"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball bat on the far left\"."}], "task": "refering", "answer_template": "The \"the baseball bat on the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 190, 191, 213, 236, 258, 259, 281, 282, 304, 326, 327, 349, 350], "bbox": [0.19584375, 0.3977510040160643, 0.321734375, 0.8921285140562248], "iscrowd": 0, "area": 3411.9707999999996, "rle": {"size": [498, 640], "counts": "fVm12]?`0@3L4M3M3L4M4L3L4M3M3M3L4M3M3L4M3M3M3LO2O1O001N2O001O1N2O001N2O001O1N101O1O1N101O1O3L6K5K4K6K5K4L5J6K4L5KkN0U15K4L5K5K5K5K2N1O1O1O1N2O1O13M4L3M3M4L3M4L3M3M4L3M3Mmkb6"}, "label": "the baseball bat on the far left"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball bat of person on the left\"."}], "task": "refering", "answer_template": "The \"baseball bat of person on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 190, 191, 213, 236, 258, 259, 281, 282, 304, 326, 327, 349, 350], "bbox": [0.19584375, 0.3977510040160643, 0.321734375, 0.8921285140562248], "iscrowd": 0, "area": 3411.9707999999996, "rle": {"size": [498, 640], "counts": "fVm12]?`0@3L4M3M3L4M4L3L4M3M3M3L4M3M3L4M3M3M3LO2O1O001N2O001O1N2O001N2O001O1N101O1O1N101O1O3L6K5K4K6K5K4L5J6K4L5KkN0U15K4L5K5K5K5K2N1O1O1O1N2O1O13M4L3M3M4L3M4L3M3M4L3M3Mmkb6"}, "label": "baseball bat of person on the left"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball bat in the hands of the man with his right knee down\"."}], "task": "refering", "answer_template": "The \"a baseball bat in the hands of the man with his right knee down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 180, 181, 203, 247, 248, 270, 292, 293, 315, 337, 338], "bbox": [0.682296875, 0.3647791164658635, 0.8921250000000001, 0.832289156626506], "iscrowd": 0, "area": 2286.1451000000006, "rle": {"size": [498, 640], "counts": "Umd64[?4N3M2M3N2M3N3M2M3N2M4M2N0O01O1O001O001O10O01O0000O100O10000O1002N2O2M3M2N3M2N3M3M2N3M3M2Nc>1\\A6J6J6J6J3M000000000000000002N2N2N3M2N2N2N3M2N2O2M2N2Nkb<NV]C2N1O2N2O0O2N2N1O2N2N1O2N2N1O2O1O0101O1N2O1O1N2O1O1N2N2M3N2M3N2NV\\Q1"}, "label": "a baseball bat in the hands of the man with his right knee down"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bat held by the man on the right\"."}], "task": "refering", "answer_template": "The \"bat held by the man on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 180, 181, 203, 247, 248, 270, 292, 293, 315, 337, 338], "bbox": [0.682296875, 0.3647791164658635, 0.8921250000000001, 0.832289156626506], "iscrowd": 0, "area": 2286.1451000000006, "rle": {"size": [498, 640], "counts": "Umd64[?4N3M2M3N2M3N3M2M3N2M4M2N0O01O1O001O001O10O01O0000O100O10000O1002N2O2M3M2N3M2N3M3M2N3M3M2Nc>1\\A6J6J6J6J3M000000000000000002N2N2N3M2N2N2N3M2N2O2M2N2Nkb<NV]C2N1O2N2O0O2N2N1O2N2N1O2N2N1O2O1O0101O1N2O1O1N2O1O1N2N2M3N2M3N2NV\\Q1"}, "label": "bat held by the man on the right"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boston baseball player wearing one leather glove , kneeling to the left of two other players\"."}], "task": "refering", "answer_template": "The \"a boston baseball player wearing one leather glove , kneeling to the left of two other players\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 73, 74, 75, 96, 97, 98, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 279, 280, 282, 283, 284, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349], "bbox": [0.108171875, 0.14036144578313253, 0.40825000000000006, 0.8914859437751005], "iscrowd": 0, "area": 38775.3606, "rle": {"size": [498, 640], "counts": "XkQ12^?5K4L5K5aN0TC5Q<e0kC_Od;Q1YDTO];T1`DQOW;V1fDoNP;Y1mDlNj:[1TEiNc:Y3F:L5K4aH`Je4d5WK^Jh4e5TK^Ji4f5SK\\Jk4i5PKYJn4k5nJXJP5k5lJWJR5m5jJVJS5n5iJRJW5S6dJnI\\5U6_JlIa5X6[JiId5[6XJeIh5^6UJcIk5_6RJaIn5b6oI_IP6d6mI\\IS6g6jIZIV6h6gIXIY6k6dIVI[6m6bISI^6Q7^IPIb6R7[InHe6P80O01O010O001gNWGiKh8V4]GfKd8X4aGdK_8[4gGaKX8^4mG^KS8a4QH]Kn7b4VH[Kj7c4\\HYKd7e4aHYK^7e4gHXK`42^Le4WOVK\\4d0mKT4MUKV4U1^Ke3a6c2300O100O100O100O100O100O1001O001N10001N101N101O0O101N101O0O2O0L5K4L4L5K4LXJSIT3h6iL`IV3\\6gLkIZ3P6dLWJ[3f5aLbJ^3Z5_LmJb3Q5XLWKg3j4PL^K^4^OnI\\3]1^MY5[N`IW4P1fMU6Y2cIoM]6Q2\\IWNc6i1UI_Nk6a1nHgNQ7Y1mHjNS7U1kHmNU7S1iHPOV7P1hHSOW7d4000010O000000010O0000000010O000000001N10000O10001N100fJfH\\2[7`MlH[2V7bMoH[2Q7bMUI[2l6bMYIZ2i6fMZIV2g6jM[IS2e6mMaIl1a6TNdIf1]6ZNhI`1X6`NnIY1T6gNQJS1P6mNUJm0k5SOZJb0m5^O_JLk54PKfN[5Y1_410O1O001O1O001O1O1O001O1O001N2O001O1O1O001O1O002N2N2N2N2J6J6J6J6J6K5J6JTTh5"}, "label": "a boston baseball player wearing one leather glove , kneeling to the left of two other players"}]}
{"id": 564676, "image": "COCO_train2014_000000564676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man kneeling with his hand on the handle of the bat\"."}], "task": "refering", "answer_template": "The \"a man kneeling with his hand on the handle of the bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 73, 74, 75, 96, 97, 98, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 279, 280, 282, 283, 284, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349], "bbox": [0.108171875, 0.14036144578313253, 0.40825000000000006, 0.8914859437751005], "iscrowd": 0, "area": 38775.3606, "rle": {"size": [498, 640], "counts": "XkQ12^?5K4L5K5aN0TC5Q<e0kC_Od;Q1YDTO];T1`DQOW;V1fDoNP;Y1mDlNj:[1TEiNc:Y3F:L5K4aH`Je4d5WK^Jh4e5TK^Ji4f5SK\\Jk4i5PKYJn4k5nJXJP5k5lJWJR5m5jJVJS5n5iJRJW5S6dJnI\\5U6_JlIa5X6[JiId5[6XJeIh5^6UJcIk5_6RJaIn5b6oI_IP6d6mI\\IS6g6jIZIV6h6gIXIY6k6dIVI[6m6bISI^6Q7^IPIb6R7[InHe6P80O01O010O001gNWGiKh8V4]GfKd8X4aGdK_8[4gGaKX8^4mG^KS8a4QH]Kn7b4VH[Kj7c4\\HYKd7e4aHYK^7e4gHXK`42^Le4WOVK\\4d0mKT4MUKV4U1^Ke3a6c2300O100O100O100O100O100O1001O001N10001N101N101O0O101N101O0O2O0L5K4L4L5K4LXJSIT3h6iL`IV3\\6gLkIZ3P6dLWJ[3f5aLbJ^3Z5_LmJb3Q5XLWKg3j4PL^K^4^OnI\\3]1^MY5[N`IW4P1fMU6Y2cIoM]6Q2\\IWNc6i1UI_Nk6a1nHgNQ7Y1mHjNS7U1kHmNU7S1iHPOV7P1hHSOW7d4000010O000000010O0000000010O000000001N10000O10001N100fJfH\\2[7`MlH[2V7bMoH[2Q7bMUI[2l6bMYIZ2i6fMZIV2g6jM[IS2e6mMaIl1a6TNdIf1]6ZNhI`1X6`NnIY1T6gNQJS1P6mNUJm0k5SOZJb0m5^O_JLk54PKfN[5Y1_410O1O001O1O001O1O1O001O1O001N2O001O1O1O001O1O002N2N2N2N2J6J6J6J6J6K5J6JTTh5"}, "label": "a man kneeling with his hand on the handle of the bat"}]}
{"id": 7621, "image": "COCO_train2014_000000007621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest cow\"."}], "task": "refering", "answer_template": "The \"the closest cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.5000312499999999, 0.2756440281030445, 0.8914843749999999, 0.85807962529274], "iscrowd": 0, "area": 45657.96725, "rle": {"size": [427, 640], "counts": "`gU44Q=7I7J7H7I8H7J6I8I6L4M4K4M4K4L4M4K4L4M4K4L5L3L4L5L3L5K4M3L5K4M3L5oNYKXIj4b6gKoH^4j6U1J6N3N0O10000O2O000O1000000O10000O100000000000010O000001O0000001O0000001O0000000010O000001O0000001O001O2N2N2N3M2N2N2O1N2N3M1O00000000[NQJnKo5S4QJlKo5T4RJkKn5U4SJjKm5W4SJhKm5X4TJgKl5Z4UJdKk5\\4VJcKj5^4VJaKk5^4VJaKj5_4WJ`Ki5a4WJ^Ki5b4XJ]Kh5d4XJ[Kh5e4ZJXKg5i4YJUKh5k4YJTKg5l4ZJRKg5o4YJoJh5Q5YJnJg5S5YJkJh5U5ZJhJg5X5ZJgJf5Z5ZJdJg5\\5ZJbJg5_5YJ`Jh5_5YJ_Jh5b5XJ\\Ji5d5XJ[Jh5e5ZJXJg5i5YJUJh5k5YJTJg5l5ZJRJg5o5YJoIh5Q6i000000001O000000000000000001O000000000000000O10000000000O2O00000000000O10000000O10O01O1O1O1O100O001O1O100O1O0104L4L4L4L4L1N2O1O001M3N2N2N1O2M3N2N2N1O2N2M3N2N1O2N2M3N101O1N2O1O001aKWHZ3j7XLeHh1C]Oj7]NRIS2BSOd8k0iGhNZ8U1SH^No7_1Q2N2M3N2N2M4M2M3N2N1N3N2M2O2M3K4L5L4K4L5LPal0"}, "label": "the closest cow"}]}
{"id": 7621, "image": "COCO_train2014_000000007621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bull seated on the ground facing the camera\"."}], "task": "refering", "answer_template": "The \"the bull seated on the ground facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.5000312499999999, 0.2756440281030445, 0.8914843749999999, 0.85807962529274], "iscrowd": 0, "area": 45657.96725, "rle": {"size": [427, 640], "counts": "`gU44Q=7I7J7H7I8H7J6I8I6L4M4K4M4K4L4M4K4L4M4K4L5L3L4L5L3L5K4M3L5K4M3L5oNYKXIj4b6gKoH^4j6U1J6N3N0O10000O2O000O1000000O10000O100000000000010O000001O0000001O0000001O0000000010O000001O0000001O001O2N2N2N3M2N2N2O1N2N3M1O00000000[NQJnKo5S4QJlKo5T4RJkKn5U4SJjKm5W4SJhKm5X4TJgKl5Z4UJdKk5\\4VJcKj5^4VJaKk5^4VJaKj5_4WJ`Ki5a4WJ^Ki5b4XJ]Kh5d4XJ[Kh5e4ZJXKg5i4YJUKh5k4YJTKg5l4ZJRKg5o4YJoJh5Q5YJnJg5S5YJkJh5U5ZJhJg5X5ZJgJf5Z5ZJdJg5\\5ZJbJg5_5YJ`Jh5_5YJ_Jh5b5XJ\\Ji5d5XJ[Jh5e5ZJXJg5i5YJUJh5k5YJTJg5l5ZJRJg5o5YJoIh5Q6i000000001O000000000000000001O000000000000000O10000000000O2O00000000000O10000000O10O01O1O1O1O100O001O1O100O1O0104L4L4L4L4L1N2O1O001M3N2N2N1O2M3N2N2N1O2N2M3N2N1O2N2M3N101O1N2O1O001aKWHZ3j7XLeHh1C]Oj7]NRIS2BSOd8k0iGhNZ8U1SH^No7_1Q2N2M3N2N2M4M2M3N2N1N3N2M2O2M3K4L5L4K4L5LPal0"}, "label": "the bull seated on the ground facing the camera"}]}
{"id": 7621, "image": "COCO_train2014_000000007621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red cow standing up\"."}], "task": "refering", "answer_template": "The \"the red cow standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 171, 172, 189, 190, 191, 193, 194, 195, 213], "bbox": [0.23304687500000001, 0.190327868852459, 0.6026875, 0.6111709601873535], "iscrowd": 0, "area": 22818.281950000004, "rle": {"size": [427, 640], "counts": "g[n12R<?[DIc;P1O2bMEiG>m7e2NO1M3M2M4M3M3M47J4K5L4K:G8H3L4M3L4M2M2O1O1N2O]NkHQMS7m2YIjLg6S3_IjL`6T3gIhLY6V3mIgLQ6W3VJeLj5X3\\JeLd5X3X2N3N1`0A5K2N2N3M2N2N3M2N2N3M2N2N3N1NPN`GXO]8g0mGQOR8o0VHkNh7U1aHdN]7\\1lH]NR7d1VITNi6l1]IPNa6P2aIPN]6P2fInMY6R2iInMU6R2mImMR6S2PJlMo5T2SJkMl5U2VJjMi5W2XJhMg5X2\\JeMd5[2^JdMa5\\2`JdM`5[2aJdM_5\\2bJcM^5^2bJaM^5_2cJ`M]5`2dJ_M\\5a2hJ[MX5f2f201O01O0000001O0001O0001O000000010O000000001O0001O000O10000000001O000000001O0O10001O000000001O0000^GgMR6Y2jIkMV6U2gInMZ6Q2bITN]6l1`IWN`6i1\\I[Nd6e1]HZOd7e0XH@g7`0VHCi7>THDm7<PHER8g23^Ob0L4M3L4L3N3L46J3M3M2N3ZGkKP8X4hGVLn7f4N000000O100mMUHhMI0R8X2[H_ML3h7_2cHUMM6`7e2jHlLM:Y7k2jInLW6R3Q20000O10O10000000O2O0O2O0O101O0O2O1N3N1N3N1O2N2N1O2M2O00000O1O100O1O1O100O100O100_Oa0J6O0O2O1N2N2nNmCk0S<TOPDi0P<XORDd0o;]OSD5W<MkC2S<OoC1n;LWD4g;J]D6a;GdD:V<1O1O1O10N101N101O1N101O0O2O0O2O1OhVZ3"}, "label": "the red cow standing up"}]}
{"id": 7621, "image": "COCO_train2014_000000007621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow that is standing\"."}], "task": "refering", "answer_template": "The \"a cow that is standing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 78, 79, 80, 81, 82, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 171, 172, 189, 190, 191, 193, 194, 195, 213], "bbox": [0.23304687500000001, 0.190327868852459, 0.6026875, 0.6111709601873535], "iscrowd": 0, "area": 22818.281950000004, "rle": {"size": [427, 640], "counts": "g[n12R<?[DIc;P1O2bMEiG>m7e2NO1M3M2M4M3M3M47J4K5L4K:G8H3L4M3L4M2M2O1O1N2O]NkHQMS7m2YIjLg6S3_IjL`6T3gIhLY6V3mIgLQ6W3VJeLj5X3\\JeLd5X3X2N3N1`0A5K2N2N3M2N2N3M2N2N3M2N2N3N1NPN`GXO]8g0mGQOR8o0VHkNh7U1aHdN]7\\1lH]NR7d1VITNi6l1]IPNa6P2aIPN]6P2fInMY6R2iInMU6R2mImMR6S2PJlMo5T2SJkMl5U2VJjMi5W2XJhMg5X2\\JeMd5[2^JdMa5\\2`JdM`5[2aJdM_5\\2bJcM^5^2bJaM^5_2cJ`M]5`2dJ_M\\5a2hJ[MX5f2f201O01O0000001O0001O0001O000000010O000000001O0001O000O10000000001O000000001O0O10001O000000001O0000^GgMR6Y2jIkMV6U2gInMZ6Q2bITN]6l1`IWN`6i1\\I[Nd6e1]HZOd7e0XH@g7`0VHCi7>THDm7<PHER8g23^Ob0L4M3L4L3N3L46J3M3M2N3ZGkKP8X4hGVLn7f4N000000O100mMUHhMI0R8X2[H_ML3h7_2cHUMM6`7e2jHlLM:Y7k2jInLW6R3Q20000O10O10000000O2O0O2O0O101O0O2O1N3N1N3N1O2N2N1O2M2O00000O1O100O1O1O100O100O100_Oa0J6O0O2O1N2N2nNmCk0S<TOPDi0P<XORDd0o;]OSD5W<MkC2S<OoC1n;LWD4g;J]D6a;GdD:V<1O1O1O10N101N101O1N101O0O2O0O2O1OhVZ3"}, "label": "a cow that is standing"}]}
{"id": 7621, "image": "COCO_train2014_000000007621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow with the horns facing the water\"."}], "task": "refering", "answer_template": "The \"the cow with the horns facing the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 74, 94, 95, 96, 117, 118, 119, 120, 139, 140, 141, 142, 143, 164, 165, 166, 187, 188, 189], "bbox": [0.08321875, 0.21934426229508197, 0.25396874999999997, 0.6], "iscrowd": 0, "area": 8969.40155, "rle": {"size": [427, 640], "counts": "fXf09P=4K5M2N2M3N1O1N2N2O1N1O2N2N2N2N2J5N3O101N1O101N1O1UFWNT8j1jGXNT8i1lGXNR8j1lGXNS8i1kGYNS8i1lGXNR8i1mGYNQ8i1nGWNg7FUGU2R1VNc7NWGm1V1UN^75YGh1W1TN\\7d2bH]M[7f2eHZMY7i2eHXMa7c2^H]Mh7^2VHcMP8X2lGkMW8T2dGoM^8P2_GRNc8m1[GTNg8l1VGUNl8j1RGWNP9h1nFYNS9g1kFZNT9h1jFYNU9j1hFVNX9l1gFTNY9m1eFTNZ9n1eFRNZ9o21O10O01O1O001O010dLfFf2[9YMgFf2X9ZMjFd2W9[MkFb2V9^MlF`2U9_MmF_2S9_MQG_2P9_MRGa2n8^MSGa2m8_MTGa2l8_MTG`2l8`MUGa2j8_MWGa2g8`MYGb2e8]M\\Gd2b8]M^Gc2b8]M^Gd2`8\\MaGe2[OTMk87jGf2VOXMn82mGP3Q8PMoGQ3o7oLRHQ3n7nLTHQ3k7oLVHQ3i7oLXHQ3f7oL\\HQ3b7nLaHR3]7mLfHT3X7jLkHV3S7iLPIW3n6hLUIX3_800ZNQF0o9NVFOj9OZFOf9N`F0_9LhF1X9KQG0Z;K2N2N1O4L5K3L3L8HWeV6"}, "label": "the cow with the horns facing the water"}]}
{"id": 7621, "image": "COCO_train2014_000000007621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bull sitting on beach looking at ocean\"."}], "task": "refering", "answer_template": "The \"bull sitting on beach looking at ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 74, 94, 95, 96, 117, 118, 119, 120, 139, 140, 141, 142, 143, 164, 165, 166, 187, 188, 189], "bbox": [0.08321875, 0.21934426229508197, 0.25396874999999997, 0.6], "iscrowd": 0, "area": 8969.40155, "rle": {"size": [427, 640], "counts": "fXf09P=4K5M2N2M3N1O1N2N2O1N1O2N2N2N2N2J5N3O101N1O101N1O1UFWNT8j1jGXNT8i1lGXNR8j1lGXNS8i1kGYNS8i1lGXNR8i1mGYNQ8i1nGWNg7FUGU2R1VNc7NWGm1V1UN^75YGh1W1TN\\7d2bH]M[7f2eHZMY7i2eHXMa7c2^H]Mh7^2VHcMP8X2lGkMW8T2dGoM^8P2_GRNc8m1[GTNg8l1VGUNl8j1RGWNP9h1nFYNS9g1kFZNT9h1jFYNU9j1hFVNX9l1gFTNY9m1eFTNZ9n1eFRNZ9o21O10O01O1O001O010dLfFf2[9YMgFf2X9ZMjFd2W9[MkFb2V9^MlF`2U9_MmF_2S9_MQG_2P9_MRGa2n8^MSGa2m8_MTGa2l8_MTG`2l8`MUGa2j8_MWGa2g8`MYGb2e8]M\\Gd2b8]M^Gc2b8]M^Gd2`8\\MaGe2[OTMk87jGf2VOXMn82mGP3Q8PMoGQ3o7oLRHQ3n7nLTHQ3k7oLVHQ3i7oLXHQ3f7oL\\HQ3b7nLaHR3]7mLfHT3X7jLkHV3S7iLPIW3n6hLUIX3_800ZNQF0o9NVFOj9OZFOf9N`F0_9LhF1X9KQG0Z;K2N2N1O4L5K3L3L8HWeV6"}, "label": "bull sitting on beach looking at ocean"}]}
{"id": 351686, "image": "COCO_train2014_000000351686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unusually straight , slightly green banana next to a more ripe banana\"."}], "task": "refering", "answer_template": "The \"an unusually straight , slightly green banana next to a more ripe banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 178, 180, 181, 182, 183], "bbox": [0.16704687499999998, 0.03966666666666666, 0.9911562499999998, 0.47281249999999997], "iscrowd": 0, "area": 34873.78070000001, "rle": {"size": [480, 640], "counts": "gVb11o>1N2O2N1N2O1O1N3N1O1N2O1O2M2O00O01000000O1000O100000O1000O10000000O10O1000000000O10O1000000000O0100000000000O010000000000O01000000000O10O1000001O000O2O0O1O2O0O101N1O101N1O2O0O101N1O101N1O101N100O2O0O101O0O101N101N100O2O0O101N10001O00001O001O00001O00001O001O00001O000010O01O00001O00001O001O00001O00001O00001O001O00001O00001O001O00010O00010O010O00010O00010O0010O010O01O010O0010O010O01O010O0010O010O01O010O0010O010O01O010O0010O01O010O0010O010O01O010O0010O01O01O010O01O010O0010O01O0100O010O10O0100O10O001O010O001O010O001O010O001O010O001O010O001O010O001O010O001O010O001O010O0100O010O010O010O010O010O010O10O010O010O010O010O010O0100O010O010O010O010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O0010O010O010O010O010O010O010O01O010O010O010O010O10O010O010O010O01OO2N101N1O2O0O2N101N1O2O0O2N1O101N1O2O0O20O0001O010O0010O00010O01O010O00010O0010O00010O01O010O00010O0010O01O01O01O010O010O0010O010O01O010O0010O010O010O00010O010O010O010O01O010O010O010O00100O010O100O00100O10O0100O10O01O100O0O2O1O1N101O1N2O001N2O1O0O2O1N2O001N2O1O0O2O1O1N2O2L3M3L4L5K4L4Mnc2"}, "label": "an unusually straight , slightly green banana next to a more ripe banana"}]}
{"id": 351686, "image": "COCO_train2014_000000351686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the straight banana\"."}], "task": "refering", "answer_template": "The \"the straight banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 178, 180, 181, 182, 183], "bbox": [0.16704687499999998, 0.03966666666666666, 0.9911562499999998, 0.47281249999999997], "iscrowd": 0, "area": 34873.78070000001, "rle": {"size": [480, 640], "counts": "gVb11o>1N2O2N1N2O1O1N3N1O1N2O1O2M2O00O01000000O1000O100000O1000O10000000O10O1000000000O10O1000000000O0100000000000O010000000000O01000000000O10O1000001O000O2O0O1O2O0O101N1O101N1O2O0O101N1O101N1O101N100O2O0O101O0O101N101N100O2O0O101N10001O00001O001O00001O00001O001O00001O000010O01O00001O00001O001O00001O00001O00001O001O00001O00001O001O00010O00010O010O00010O00010O0010O010O01O010O0010O010O01O010O0010O010O01O010O0010O010O01O010O0010O01O010O0010O010O01O010O0010O01O01O010O01O010O0010O01O0100O010O10O0100O10O001O010O001O010O001O010O001O010O001O010O001O010O001O010O001O010O001O010O0100O010O010O010O010O010O010O10O010O010O010O010O010O0100O010O010O010O010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O0010O010O010O010O010O010O010O01O010O010O010O010O10O010O010O010O01OO2N101N1O2O0O2N101N1O2O0O2N1O101N1O2O0O20O0001O010O0010O00010O01O010O00010O0010O00010O01O010O00010O0010O01O01O01O010O010O0010O010O01O010O0010O010O010O00010O010O010O010O01O010O010O010O00100O010O100O00100O10O0100O10O01O100O0O2O1O1N101O1N2O001N2O1O0O2O1N2O001N2O1O0O2O1O1N2O2L3M3L4L5K4L4Mnc2"}, "label": "the straight banana"}]}
{"id": 351686, "image": "COCO_train2014_000000351686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana with a sticker on it\"."}], "task": "refering", "answer_template": "The \"a banana with a sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267, 268, 269, 270, 291, 292, 293], "bbox": [0.3, 0.2, 0.7623593750000001, 0.7647083333333333], "iscrowd": 0, "area": 29517.460800000004, "rle": {"size": [480, 640], "counts": "jTj21m>3L4M3M3M3M3M3M2N3M3M3M3N2N100O1O101N1O100O1O100O1O100O1O100O2N100O100O100O10000O100O101N100O100O100O100O100O100O101O0O101N101N101N101N101N101N101O01O01O0010O01O0010O01O0010O01O0010O01O00010O0010O01O0010O01O0010O01O0010O01O00010O0010O01O0010O01O1O10O01O10O0100O10O0100O00100O10O0100O010O1O10O0100O100O00100O100O100O100O1O100O100O100O010O1O100O100O10000O10000O10000O10O11N4M2N3M2M4M2N3L3N3M2M3N3M2M2O000O10001N10000O1000000O10000O10000O10000O2O000O100000000001O001O00001O00001O00001O00001N10001O00001O00001N2O1O1O1O1O1O1O1N2O1O1O1O1M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M6J8I7H8H8H8H8H8H8H8HbmV2"}, "label": "a banana with a sticker on it"}]}
{"id": 351686, "image": "COCO_train2014_000000351686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow and brown banana sitting next to a green banana\"."}], "task": "refering", "answer_template": "The \"a yellow and brown banana sitting next to a green banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267, 268, 269, 270, 291, 292, 293], "bbox": [0.3, 0.2, 0.7623593750000001, 0.7647083333333333], "iscrowd": 0, "area": 29517.460800000004, "rle": {"size": [480, 640], "counts": "jTj21m>3L4M3M3M3M3M3M2N3M3M3M3N2N100O1O101N1O100O1O100O1O100O1O100O2N100O100O100O10000O100O101N100O100O100O100O100O100O101O0O101N101N101N101N101N101N101O01O01O0010O01O0010O01O0010O01O0010O01O00010O0010O01O0010O01O0010O01O0010O01O00010O0010O01O0010O01O1O10O01O10O0100O10O0100O00100O10O0100O010O1O10O0100O100O00100O100O100O100O1O100O100O100O010O1O100O100O10000O10000O10000O10O11N4M2N3M2M4M2N3L3N3M2M3N3M2M2O000O10001N10000O1000000O10000O10000O10000O2O000O100000000001O001O00001O00001O00001O00001N10001O00001O00001N2O1O1O1O1O1O1O1N2O1O1O1O1M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M6J8I7H8H8H8H8H8H8H8HbmV2"}, "label": "a yellow and brown banana sitting next to a green banana"}]}
{"id": 425415, "image": "COCO_train2014_000000425415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic light on the image\"."}], "task": "refering", "answer_template": "The \"traffic light on the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 331, 332, 333, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361], "bbox": [0.34665625, 0.382781954887218, 0.7704375000000001, 0.8226503759398496], "iscrowd": 0, "area": 35958.00480000001, "rle": {"size": [532, 640], "counts": "hec35[`05J7L4N2M3N1N3N2M3N1O2N2N1O1O1O100O1O1O1O100O1O1O1O1O100O3M9G1O000001O0001O1O001O10OO01O1O1000O1000000O10O100000TNWNlDj1S;_NdDb1Z;aNdD`1`9oMPGc0^O_1`9SNnFa0_O]1a9VNmF?@[1b9[NjF<BZ1b9_NgF:EX1b9cNeF7GW1b9fNdF5HV1c9gNcF6GT1f9hN`F6HS1h9hN_F6GR1k9iN]F6FR1m9iN\\F6ER1o9iN[F6EQ1P:jNZF7Do0R:lNXF6En0T:lNVF7Em0U:mNTF8El0W:mNSF8El0W:mNSFm2m9UMPFm2o9TMjER3V:f101O01O01O000000001O01O0000000001O0000000000000001O0001O00000000000000001O00001O0000<D5K3M0000000000001O001000O001O1C^FeId9Z6^FcId9\\6]FbId9^6^F`Ic9_6:O1O001O1O1O001N2M3O0010O00001O001O00001O00001O00O100O010JhEoIY:Q65001O1O1N1000O010000O1000O01000O0100001aKeEb2\\:]MgEa2Z:^MhEa2W:^MlEa2T:]MmEd2R:\\MoE6K<W:\\OPF10c0P:ZOQF02g0l9WOUFO2j0i9TOWF03l0e9SOZFN4o0b9PO]FM5R1_9nN`FL3V1]9mNbFJ4X1[9kNeFI3[1X9kNgFF5^1U9iNjFE5`1Q9hNoF_O:f1h8WNcIf1i:MO2O0O10000O2O000O100O10000O10000O10001N101N101O1N100000000O10M300O10O100000O010J6F:N2O2N2N1N3N2N1O2M3N1O2M3N3L4L5G8H^P\\2"}, "label": "traffic light on the image"}]}
{"id": 237002, "image": "COCO_train2014_000000237002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"you can see part of the umbrella on the left\"."}], "task": "refering", "answer_template": "The \"you can see part of the umbrella on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 121, 122], "bbox": [0.009093750000000001, 0.01574074074074074, 0.634046875, 0.35956018518518523], "iscrowd": 0, "area": 41057.69069999999, "rle": {"size": [432, 640], "counts": "^d2P1S;]1cN^1B=000001O00001OO10000000000O100000000O1000000O1M3N2N2N2N2K5J6J6J6J6N20000000000000000000000000000001O2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N1O1O001O001O001O001O1O1O001O1O1O1O1O1O1O001O1O1O1O1O00000000000000O1000000000000O1000000000000O100N2N2O1N2N2O1N2N2N2O1N2N2O1QO]FoMe9Q2_FiMc9W2`FeMa9[2cF_M_9a2dFZM^9f2f00000000000O100000000000000000000002N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O000000000000000000O10000000000000000O1000000O10000O1000000O10000O1000000O100O1N2N2M3N2[N`FXOb9g0dFRO^9n0gFkN[9U1jFdNX9\\1nF[NU9e1QGSNQ9m1UGkMm8U2W1O10000000000002N2N3M2N2N3M2N3M2N3M2N2N3M2N3M2N3M2N2N3M2N3M1OO10000O100O10000O100O10000O100O10001N100O10000O100O10000O10000O100O10000O100O10000mMaF4`9GoFOQ9K`GIa81PHDP8:\\H\\Od7c0e2N2N2O1O13M3M3M3M3M3M3M3M3M3M3M4L1O1O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O0000O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O10000O100O10000O100O10000O100O10000O1N2N2N2O1N2N2N2N2I7F:GQgR3"}, "label": "you can see part of the umbrella on the left"}]}
{"id": 237002, "image": "COCO_train2014_000000237002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella to the left\"."}], "task": "refering", "answer_template": "The \"the umbrella to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 121, 122], "bbox": [0.009093750000000001, 0.01574074074074074, 0.634046875, 0.35956018518518523], "iscrowd": 0, "area": 41057.69069999999, "rle": {"size": [432, 640], "counts": "^d2P1S;]1cN^1B=000001O00001OO10000000000O100000000O1000000O1M3N2N2N2N2K5J6J6J6J6N20000000000000000000000000000001O2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N1O1O001O001O001O001O1O1O001O1O1O1O1O1O1O001O1O1O1O1O00000000000000O1000000000000O1000000000000O100N2N2O1N2N2O1N2N2N2O1N2N2O1QO]FoMe9Q2_FiMc9W2`FeMa9[2cF_M_9a2dFZM^9f2f00000000000O100000000000000000000002N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O000000000000000000O10000000000000000O1000000O10000O1000000O10000O1000000O100O1N2N2M3N2[N`FXOb9g0dFRO^9n0gFkN[9U1jFdNX9\\1nF[NU9e1QGSNQ9m1UGkMm8U2W1O10000000000002N2N3M2N2N3M2N3M2N3M2N2N3M2N3M2N3M2N2N3M2N3M1OO10000O100O10000O100O10000O100O10001N100O10000O100O10000O10000O100O10000O100O10000mMaF4`9GoFOQ9K`GIa81PHDP8:\\H\\Od7c0e2N2N2O1O13M3M3M3M3M3M3M3M3M3M3M4L1O1O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O0000O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O10000O100O10000O100O10000O100O10000O1N2N2N2O1N2N2N2N2I7F:GQgR3"}, "label": "the umbrella to the left"}]}
{"id": 237002, "image": "COCO_train2014_000000237002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella with red , green blue , yellow and red showing in that order\"."}], "task": "refering", "answer_template": "The \"umbrella with red , green blue , yellow and red showing in that order\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137], "bbox": [0.638765625, 0.019166666666666665, 0.98903125, 0.38240740740740736], "iscrowd": 0, "area": 28938.20815000001, "rle": {"size": [432, 640], "counts": "ha\\5a1o;2N3M2N3M2N3M2N3M2N3M1O001O1O001O002N2N2N4L4L4L3M4LAhEgMW:Y2kEeMU:\\2mEbMR:^2PFaMo9`2SF^Ml9b2VF]Mi9d2YFYMg9g2[FXMd9i2^FUMa9l2a03M3M3M3M3M3M4L4L1O1O1O1O1O001O1O1O1O001O00000000000000001O002N1O2N1O2N1OWOnFUMQ9k2XGkLi8U3YGhLh8Y3YGcLi8]3e000000000001[FkLm8m300O1O2N1O1O1O1O00001O000000000000001O000000000000000000000000001O000000000000O1N2N2N2ZO`GZLb8f3d000000000001O00000000000000000^FiLk8o3O1O1O1O001O1O1O1O1O0000000000001O0000000000000000000000000000000000000000001O0000O1O1O1O1O1O1O1O1O1O1O1O1[Oe0000000000000000000000000000000d0\\O2N1O1O1O1O1O1O1O1O1i0XO[m2"}, "label": "umbrella with red , green blue , yellow and red showing in that order"}]}
{"id": 237002, "image": "COCO_train2014_000000237002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella on the right close to the photographer\"."}], "task": "refering", "answer_template": "The \"umbrella on the right close to the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137], "bbox": [0.638765625, 0.019166666666666665, 0.98903125, 0.38240740740740736], "iscrowd": 0, "area": 28938.20815000001, "rle": {"size": [432, 640], "counts": "ha\\5a1o;2N3M2N3M2N3M2N3M2N3M1O001O1O001O002N2N2N4L4L4L3M4LAhEgMW:Y2kEeMU:\\2mEbMR:^2PFaMo9`2SF^Ml9b2VF]Mi9d2YFYMg9g2[FXMd9i2^FUMa9l2a03M3M3M3M3M3M4L4L1O1O1O1O1O001O1O1O1O001O00000000000000001O002N1O2N1O2N1OWOnFUMQ9k2XGkLi8U3YGhLh8Y3YGcLi8]3e000000000001[FkLm8m300O1O2N1O1O1O1O00001O000000000000001O000000000000000000000000001O000000000000O1N2N2N2ZO`GZLb8f3d000000000001O00000000000000000^FiLk8o3O1O1O1O001O1O1O1O1O0000000000001O0000000000000000000000000000000000000000001O0000O1O1O1O1O1O1O1O1O1O1O1O1[Oe0000000000000000000000000000000d0\\O2N1O1O1O1O1O1O1O1O1i0XO[m2"}, "label": "umbrella on the right close to the photographer"}]}
{"id": 556492, "image": "COCO_train2014_000000556492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus passing by\"."}], "task": "refering", "answer_template": "The \"bus passing by\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114], "bbox": [0.796109375, 0.01761006289308176, 0.9993281249999999, 0.42660377358490564], "iscrowd": 0, "area": 14210.942100000006, "rle": {"size": [318, 640], "counts": "`_n4=Y87YHK^7_1G9G9H8L5K400O2O001O00001O0O101O001O1O1N2O1O2N001O1N2OO1000000O1N1O2M3N2O1000O10000000000O1000001O00000O10000000000O1000001O000O1000000000000O10001O00000O1000000000000O10001O00000O10000000000O1000001O00000O10000000000O2O00001O000O2O0000001O000O2O00001O00_L[3"}, "label": "bus passing by"}]}
{"id": 556492, "image": "COCO_train2014_000000556492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a city bus with let windows stopped behind a dark car\"."}], "task": "refering", "answer_template": "The \"a city bus with let windows stopped behind a dark car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114], "bbox": [0.796109375, 0.01761006289308176, 0.9993281249999999, 0.42660377358490564], "iscrowd": 0, "area": 14210.942100000006, "rle": {"size": [318, 640], "counts": "`_n4=Y87YHK^7_1G9G9H8L5K400O2O001O00001O0O101O001O1O1N2O1O2N001O1N2OO1000000O1N1O2M3N2O1000O10000000000O1000001O00000O10000000000O1000001O000O1000000000000O10001O00000O1000000000000O10001O00000O10000000000O1000001O00000O10000000000O2O00001O000O2O0000001O000O2O00001O00_L[3"}, "label": "a city bus with let windows stopped behind a dark car"}]}
{"id": 556492, "image": "COCO_train2014_000000556492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person standing by the street\"."}], "task": "refering", "answer_template": "The \"a person standing by the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 53, 75, 76, 98, 99, 100, 121, 122, 123, 144, 145, 146, 167, 168, 191, 214, 215, 237, 238], "bbox": [0.29167187499999997, 0.16364779874213836, 0.378140625, 0.9454402515723271], "iscrowd": 0, "area": 7687.8157999999985, "rle": {"size": [318, 640], "counts": "XVj1;`9<C<E<C=TNXNmJg2W4\\MbKg3^3]L\\Le4e2_KQMW5^2mJXM[5g2o0N2N2O0O2O1N2O1N2N2O1N2O010ObNaM^K^2\\4lMaKT2Y4VNcKi1W4bNeK^1U4hNkKW1R4mNnKT1CgM_3Y1nLU1ROTNf3k0XMT2]2PNbMX2S2lMmMU2n1nMRNT2i1PNVNR2f1QNYNQ2b1SN]No1^1UNaNm1[1UNeNm1X1UNgNm1U1VNjNl1R1WNmNk1o0XNPOk1l0WNSOl1h0WNWOl1d0VN\\Om1`0VN^Om1=WNBk19YNEj17YNDn16WNCQ27P4J6KQ^k3"}, "label": "a person standing by the street"}]}
{"id": 556492, "image": "COCO_train2014_000000556492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person holding a bag standing on a curb\"."}], "task": "refering", "answer_template": "The \"a person holding a bag standing on a curb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 53, 75, 76, 98, 99, 100, 121, 122, 123, 144, 145, 146, 167, 168, 191, 214, 215, 237, 238], "bbox": [0.29167187499999997, 0.16364779874213836, 0.378140625, 0.9454402515723271], "iscrowd": 0, "area": 7687.8157999999985, "rle": {"size": [318, 640], "counts": "XVj1;`9<C<E<C=TNXNmJg2W4\\MbKg3^3]L\\Le4e2_KQMW5^2mJXM[5g2o0N2N2O0O2O1N2O1N2N2O1N2O010ObNaM^K^2\\4lMaKT2Y4VNcKi1W4bNeK^1U4hNkKW1R4mNnKT1CgM_3Y1nLU1ROTNf3k0XMT2]2PNbMX2S2lMmMU2n1nMRNT2i1PNVNR2f1QNYNQ2b1SN]No1^1UNaNm1[1UNeNm1X1UNgNm1U1VNjNl1R1WNmNk1o0XNPOk1l0WNSOl1h0WNWOl1d0VN\\Om1`0VN^Om1=WNBk19YNEj17YNDn16WNCQ27P4J6KQ^k3"}, "label": "a person holding a bag standing on a curb"}]}
{"id": 507342, "image": "COCO_train2014_000000507342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car to the left of another car\"."}], "task": "refering", "answer_template": "The \"a car to the left of another car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120], "bbox": [0.012718750000000001, 0.13118393234672304, 0.2368125, 0.3569978858350952], "iscrowd": 0, "area": 10696.42395, "rle": {"size": [473, 640], "counts": "Zi3;\\>l0SO3N2N2N1N3N2N2N2N2M2OF:3O1N2O0O2O4Kl0UO0O2O0O2O1N101N2O001O001O1O001O001O1O001O1O00O100O10000O10000O10000O10000O10000O100J6A?001O0000000000000000000000001O000001O0000001O00001O0000010O00001O0000001O00001O0001O01O002N4L3M4L4L4L1OO101N2N3N1N2N3N1N3N10TN^C^1e<_N^C_1d<^N_C`1b<^NaC`1m<N3N2N2N2M2O2N2N2N2M3N1O2N2D<D^YQ7"}, "label": "a car to the left of another car"}]}
{"id": 507342, "image": "COCO_train2014_000000507342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan car by a blue box\"."}], "task": "refering", "answer_template": "The \"a tan car by a blue box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120], "bbox": [0.012718750000000001, 0.13118393234672304, 0.2368125, 0.3569978858350952], "iscrowd": 0, "area": 10696.42395, "rle": {"size": [473, 640], "counts": "Zi3;\\>l0SO3N2N2N1N3N2N2N2N2M2OF:3O1N2O0O2O4Kl0UO0O2O0O2O1N101N2O001O001O1O001O001O1O001O1O00O100O10000O10000O10000O10000O10000O100J6A?001O0000000000000000000000001O000001O0000001O00001O0000010O00001O0000001O00001O0001O01O002N4L3M4L4L4L1OO101N2N3N1N2N3N1N3N10TN^C^1e<_N^C_1d<^N_C`1b<^NaC`1m<N3N2N2N2M2O2N2N2N2M3N1O2N2D<D^YQ7"}, "label": "a tan car by a blue box"}]}
{"id": 507342, "image": "COCO_train2014_000000507342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beige car with two people looking at us\"."}], "task": "refering", "answer_template": "The \"a beige car with two people looking at us\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 77, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 152, 153, 154, 155, 156], "bbox": [0.34965625, 0.13334038054968286, 0.8360156249999999, 0.4021564482029598], "iscrowd": 0, "area": 25506.670649999993, "rle": {"size": [473, 640], "counts": "PbW37b>:F:F:VBlNZ=Y11N2O1O02O00000O2O00000O101O0O1O2N3N3L3M3M4L3M3M4L3N0O1O1O1O1O2N1O100O1O1O2N1O1000000O2O0000000O101O000O1000001O0000000000001O00000000001N10000000001O00000000001O00000000001O000000000O2O00000000001O0000000000001O00000000001O000O1000001O00000000001O0000M3K5K5M3N2O1N2O1N2O0O2O1N2O1001O0000000000000001O0000010O0000001O0000001O0002O0O2O000O2O0O2O001N103M3L4M3M3L4M1N0100O001O1O001O00100O001O1O001O1O010O001O1O001O1O001O001O1O001O1O0000001O0O101O00001O0000001O00001O00001O0000001O000O2O00001O0000001N10001N100O101O0O101O0O101N10000O2O000O2O000O2O0O10001N10001N100O2O000O101O0O101O0O2O1N2O1O1N2O1O1N2O1O2M2O2M2O2N1N>CX[`1"}, "label": "a beige car with two people looking at us"}]}
{"id": 507342, "image": "COCO_train2014_000000507342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large bus which the person with the bagpack is staring at\"."}], "task": "refering", "answer_template": "The \"the large bus which the person with the bagpack is staring at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 282, 283, 286, 287, 288, 305, 306], "bbox": [0.285640625, 0.4921987315010571, 0.545703125, 0.7884143763213531], "iscrowd": 0, "area": 15956.642900000004, "rle": {"size": [473, 640], "counts": "Qkd2R1[<_1C<H8H8H8H8H8N2NO1O001O1O1O0D=1O0O2O00001N10001O1N3N1O1O2N1O1O2N1OM3QO_EcMd:]2\\E`Mg:_2[E]Mh:b2ZEZMi:e2XEXMk:g2WEUMl:j2VETMk:l2UESMl:W2kDWN;_Ol:Y2lDUN9Al:Y2nDSNb;l1aDQN`;n1cDoM_;o1cDPN];o1fDnM[;Q2g0O100O100O100O100O1002N1O1O2N1O1O2N1O1O2TDdMW;]2fDdM[;^2aDdM_;]2^DeMb;\\2[DfMe;\\2WDfMj;d20000000O10000000000001O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2O0O1O2N101O0O101O00001O0O01N2N2O1N2N101N2N2O1N2N101N2N2O1N2N101N2N2O1N2N102M2N2O1N2N2N2O1N2N2O1N2N2O1N2N3N1N2N2O1N2N2O1N2N8Ie0ZO_SV4"}, "label": "the large bus which the person with the bagpack is staring at"}]}
{"id": 507342, "image": "COCO_train2014_000000507342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front of bus on bottom pic\"."}], "task": "refering", "answer_template": "The \"front of bus on bottom pic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 282, 283, 286, 287, 288, 305, 306], "bbox": [0.285640625, 0.4921987315010571, 0.545703125, 0.7884143763213531], "iscrowd": 0, "area": 15956.642900000004, "rle": {"size": [473, 640], "counts": "Qkd2R1[<_1C<H8H8H8H8H8N2NO1O001O1O1O0D=1O0O2O00001N10001O1N3N1O1O2N1O1O2N1OM3QO_EcMd:]2\\E`Mg:_2[E]Mh:b2ZEZMi:e2XEXMk:g2WEUMl:j2VETMk:l2UESMl:W2kDWN;_Ol:Y2lDUN9Al:Y2nDSNb;l1aDQN`;n1cDoM_;o1cDPN];o1fDnM[;Q2g0O100O100O100O100O1002N1O1O2N1O1O2N1O1O2TDdMW;]2fDdM[;^2aDdM_;]2^DeMb;\\2[DfMe;\\2WDfMj;d20000000O10000000000001O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2O0O1O2N101O0O101O00001O0O01N2N2O1N2N101N2N2O1N2N101N2N2O1N2N101N2N2O1N2N102M2N2O1N2N2N2O1N2N2O1N2N2O1N2N3N1N2N2O1N2N2O1N2N8Ie0ZO_SV4"}, "label": "front of bus on bottom pic"}]}
{"id": 392657, "image": "COCO_train2014_000000392657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a blue star - patterned shirt lays on his side on a white sheet\"."}], "task": "refering", "answer_template": "The \"a boy in a blue star - patterned shirt lays on his side on a white sheet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253], "bbox": [0.0015, 0.35295081967213116, 0.6641874999999999, 0.7439812646370023], "iscrowd": 0, "area": 36707.77769999999, "rle": {"size": [427, 640], "counts": "ef0c0T<d000000O10000O10000O1000000O10000O10000O1000000O10000O0100000O10000O01000O0100000O01000O10O10O10O1000O10O10O1000O1000000O100000O01000000O1000000O1000O10O1000000O1000000O0100000O1000000O100000O01000000000000O10001O1O1O1O1O1O2M2O1O1O1O1O1O000O1000000000O1000O1XNnNWGR1h8oNXGQ1h8POWGP1i8QOVGo0j8QOVGo0i8SOVGm0j8TOUGl0k8TOUGl0k8UOSGl0m8UORGk0m8VOSGj0m8WORGh0o8XOQGh0o8WORGi0m8XOSGh0m8XOSGh0m8XORGi0n8WORGi0m8XOSGh0m8XORGi0n8WORGi0n8WOQGj0o8VOQGj0n8WOQGj0o8VOQGj0o8VOPGk0P9UOPGk0o8VOPGj0Q9VOoFj0Q9VOoFj0Q9VOnFk0R9UOnFk0Q9VOnFk0R9TOoFl0Q9TOnFm0R9SOnFm0Q9TOnFm0R9SOnFm0R9ROnFo0R9QOnFo0Q9ROnFo0R9QOnFo0R9QOnFn0S9QOmFP1R9QOnFo0R9QOmFP1S9POmFP1S9oNmFR1R9oNnFQ1R9oNmFR1S9nNmFR1S9nNkFT1T9lNkFV1U9jNiFX1W9hNgFZ1Y9fNfF[1Y9fNeF\\1[9cNdF_1\\9aNbF`1_9`N_Fb1`9_N_Fb1a9T10000O1000000O1000000O1000O10O10000001O00000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000000O1000O10000000001O0000000000000000000000000000000O1000000000000000000000000000000000000000000O100000000000N2N2M3M3M3N2M3O100O1O100O1O1O100O1O100O1OTOWGXMi8g2XGYMg8f2[GZMd8e2^G[Ma8e2`G[M`8c2bG]M^8a2dG_M\\8`2eG`M[8^2gGbMZ8[2hGeMX8W2lGiMT8S2PHmMP8k1XG]Ml0h0m7j1WHVNi7j1WHVNi7j1WHVNi7j1WHVNj7i1VHWNj7i1VHWNj7i1VHWNj7i1VHWNj7i1VHWNk7h1UHXNk7h1UHXNk7h1UHXNk7h1UHXNl7g1THYNl7g1THYNl7g1THYNl7g1THYNl7g1THYNm7f1SHZNm7f1SHZNm7e1TH[Nl7e1UHZNl7d1UH\\Nl7c1TH]Nl7b1UH^Nl7a1TH_Nl7`1UH`Nl7_1THaNl7^1UHbNl7]1THcNl7\\1VHVNX8i1f1N101N101N101N3N1N3L4K5K5K4M4K5K5K5L3L5K5KP_i2"}, "label": "a boy in a blue star - patterned shirt lays on his side on a white sheet"}]}
{"id": 392657, "image": "COCO_train2014_000000392657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child in blue and star shirt\"."}], "task": "refering", "answer_template": "The \"child in blue and star shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253], "bbox": [0.0015, 0.35295081967213116, 0.6641874999999999, 0.7439812646370023], "iscrowd": 0, "area": 36707.77769999999, "rle": {"size": [427, 640], "counts": "ef0c0T<d000000O10000O10000O1000000O10000O10000O1000000O10000O0100000O10000O01000O0100000O01000O10O10O10O1000O10O10O1000O1000000O100000O01000000O1000000O1000O10O1000000O1000000O0100000O1000000O100000O01000000000000O10001O1O1O1O1O1O2M2O1O1O1O1O1O000O1000000000O1000O1XNnNWGR1h8oNXGQ1h8POWGP1i8QOVGo0j8QOVGo0i8SOVGm0j8TOUGl0k8TOUGl0k8UOSGl0m8UORGk0m8VOSGj0m8WORGh0o8XOQGh0o8WORGi0m8XOSGh0m8XOSGh0m8XORGi0n8WORGi0m8XOSGh0m8XORGi0n8WORGi0n8WOQGj0o8VOQGj0n8WOQGj0o8VOQGj0o8VOPGk0P9UOPGk0o8VOPGj0Q9VOoFj0Q9VOoFj0Q9VOnFk0R9UOnFk0Q9VOnFk0R9TOoFl0Q9TOnFm0R9SOnFm0Q9TOnFm0R9SOnFm0R9ROnFo0R9QOnFo0Q9ROnFo0R9QOnFo0R9QOnFn0S9QOmFP1R9QOnFo0R9QOmFP1S9POmFP1S9oNmFR1R9oNnFQ1R9oNmFR1S9nNmFR1S9nNkFT1T9lNkFV1U9jNiFX1W9hNgFZ1Y9fNfF[1Y9fNeF\\1[9cNdF_1\\9aNbF`1_9`N_Fb1`9_N_Fb1a9T10000O1000000O1000000O1000O10O10000001O00000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000000O1000O10000000001O0000000000000000000000000000000O1000000000000000000000000000000000000000000O100000000000N2N2M3M3M3N2M3O100O1O100O1O1O100O1O100O1OTOWGXMi8g2XGYMg8f2[GZMd8e2^G[Ma8e2`G[M`8c2bG]M^8a2dG_M\\8`2eG`M[8^2gGbMZ8[2hGeMX8W2lGiMT8S2PHmMP8k1XG]Ml0h0m7j1WHVNi7j1WHVNi7j1WHVNi7j1WHVNj7i1VHWNj7i1VHWNj7i1VHWNj7i1VHWNj7i1VHWNk7h1UHXNk7h1UHXNk7h1UHXNk7h1UHXNl7g1THYNl7g1THYNl7g1THYNl7g1THYNl7g1THYNm7f1SHZNm7f1SHZNm7e1TH[Nl7e1UHZNl7d1UH\\Nl7c1TH]Nl7b1UH^Nl7a1TH_Nl7`1UH`Nl7_1THaNl7^1UHbNl7]1THcNl7\\1VHVNX8i1f1N101N101N101N3N1N3L4K5K5K4M4K5K5K5L3L5K5KP_i2"}, "label": "child in blue and star shirt"}]}
{"id": 392657, "image": "COCO_train2014_000000392657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy looking like he is about to start running\"."}], "task": "refering", "answer_template": "The \"a little boy looking like he is about to start running\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 27, 28, 29, 49, 50, 51, 52, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 164, 165, 187, 188, 189], "bbox": [0.026984374999999998, 0.0022482435597189696, 0.313359375, 0.6022482435597188], "iscrowd": 0, "area": 22570.00785, "rle": {"size": [427, 640], "counts": "`X71X=3N2M2N3N1N3M3N1M4M2M4L3M4L3M4L3O2M2O1O0O2O1O1O1N2O001N2O1O1O0O2O1O1N2O1O001N2OD<C<010000O01000000O10O10O1000000O100^ElNn8T1aF_O\\9b0]FGa99^FJ`96^FN`92^F2_9O_F5_9K_F9_9G`F<]9EaF?]9i1N2N2VHRLT6P4gITLY6m3cIVL]6k3_IXLa6i3ZI[Le6g3WI\\Li6e3SI^Lm6c3oH`LP7b3kHbLU7_3gHdLY7]3cHfL]7[3_HhL`7b4000000O10000O10O1000O10000SOZJTJg5a5iJZJW5Z5ZKaJf4X5fKcJZ4\\5lK^JU4b5j10000O1000000O1000000O10lNdJYJ\\5b5iJ^JW5]5oJbJP5Z5UKfJk4V5ZKiJf4V5[KjJe4V5\\KiJd4W5\\KiJc4W5^KhJc4X5^KgJb4Y5^KgJb4Y5_KfJ`4Z5aKfJ_4Z5bKeJ^4[5bKeJ^4[5cKdJ\\4\\5eKdJ[4\\5j100O0O2O1N3N1N3N2M3N1N3N2M2O2N2N2M2lLjGR1X8lNjGR1X8lNiGS1Y8kNiGR1Y8mNiGQ1Y8mNhGS1Y8jNjGT1X8jNjGU1V8jNkGU1W8iNkGV1V8ZN\\G\\O?Y2W8WN_G]O<[2U8VNbG]O;[2U8TNeG^O7^2U8QNaHm1`7PNdHo1]7nMeHQ2]7kMgHT2Z7iMiHU2X7iMmHT2X9L4M4K4M3L4M5J5L4K5L5J5L4K6K4K5LUmf5"}, "label": "a little boy looking like he is about to start running"}]}
{"id": 392657, "image": "COCO_train2014_000000392657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child with a rainbow shirt runs toward another child\"."}], "task": "refering", "answer_template": "The \"a small child with a rainbow shirt runs toward another child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 27, 28, 29, 49, 50, 51, 52, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 164, 165, 187, 188, 189], "bbox": [0.026984374999999998, 0.0022482435597189696, 0.313359375, 0.6022482435597188], "iscrowd": 0, "area": 22570.00785, "rle": {"size": [427, 640], "counts": "`X71X=3N2M2N3N1N3M3N1M4M2M4L3M4L3M4L3O2M2O1O0O2O1O1O1N2O001N2O1O1O0O2O1O1N2O1O001N2OD<C<010000O01000000O10O10O1000000O100^ElNn8T1aF_O\\9b0]FGa99^FJ`96^FN`92^F2_9O_F5_9K_F9_9G`F<]9EaF?]9i1N2N2VHRLT6P4gITLY6m3cIVL]6k3_IXLa6i3ZI[Le6g3WI\\Li6e3SI^Lm6c3oH`LP7b3kHbLU7_3gHdLY7]3cHfL]7[3_HhL`7b4000000O10000O10O1000O10000SOZJTJg5a5iJZJW5Z5ZKaJf4X5fKcJZ4\\5lK^JU4b5j10000O1000000O1000000O10lNdJYJ\\5b5iJ^JW5]5oJbJP5Z5UKfJk4V5ZKiJf4V5[KjJe4V5\\KiJd4W5\\KiJc4W5^KhJc4X5^KgJb4Y5^KgJb4Y5_KfJ`4Z5aKfJ_4Z5bKeJ^4[5bKeJ^4[5cKdJ\\4\\5eKdJ[4\\5j100O0O2O1N3N1N3N2M3N1N3N2M2O2N2N2M2lLjGR1X8lNjGR1X8lNiGS1Y8kNiGR1Y8mNiGQ1Y8mNhGS1Y8jNjGT1X8jNjGU1V8jNkGU1W8iNkGV1V8ZN\\G\\O?Y2W8WN_G]O<[2U8VNbG]O;[2U8TNeG^O7^2U8QNaHm1`7PNdHo1]7nMeHQ2]7kMgHT2Z7iMiHU2X7iMmHT2X9L4M4K4M3L4M5J5L4K5L5J5L4K6K4K5LUmf5"}, "label": "a small child with a rainbow shirt runs toward another child"}]}
{"id": 499155, "image": "COCO_train2014_000000499155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"counter between two women\"."}], "task": "refering", "answer_template": "The \"counter between two women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 309, 310, 311, 312, 313, 314, 335, 336, 337], "bbox": [0.336625, 0.7111943793911008, 0.6870625, 0.9997423887587822], "iscrowd": 0, "area": 11186.403899999998, "rle": {"size": [427, 640], "counts": "Woi21U=5FMVC:i<5O001O0O2O001O1O001O001O1O0O2O001O001O1O001O010O10O010O010O01O1O001O001N2O001O001N101O1O001N101O001O0O2O00001N1O2O0O2M2N3M201O0010O0001O010O001O010O001O01O01O0010O01O0010O01O010O000010O01O0010O01O0010O01O01O01O010O0010O01O010O00010O0010O01O010O0010O01O01O01O010O00010O00010O00010O0001O01O01O01O01O01O01O01O01O01O01O01O01O01m0RO8H0010O000010O00010O0001O01O01O00010O00010O00010O0001O01O01O01O01O01O000100O2N2O0O2N2O0O2N2O0O2N1O2O1N1O2O1N3M3N2M3M3N]ob2"}, "label": "counter between two women"}]}
{"id": 499155, "image": "COCO_train2014_000000499155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and brown bar in a room\"."}], "task": "refering", "answer_template": "The \"a red and brown bar in a room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 260, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 309, 310, 311, 312, 313, 314, 335, 336, 337], "bbox": [0.336625, 0.7111943793911008, 0.6870625, 0.9997423887587822], "iscrowd": 0, "area": 11186.403899999998, "rle": {"size": [427, 640], "counts": "Woi21U=5FMVC:i<5O001O0O2O001O1O001O001O1O0O2O001O001O1O001O010O10O010O010O01O1O001O001N2O001O001N101O1O001N101O001O0O2O00001N1O2O0O2M2N3M201O0010O0001O010O001O010O001O01O01O0010O01O0010O01O010O000010O01O0010O01O0010O01O01O01O010O0010O01O010O00010O0010O01O010O0010O01O01O01O010O00010O00010O00010O0001O01O01O01O01O01O01O01O01O01O01O01O01O01m0RO8H0010O000010O00010O0001O01O01O00010O00010O00010O0001O01O01O01O01O01O000100O2N2O0O2N2O0O2N2O0O2N1O2O1N1O2O1N3M3N2M3M3N]ob2"}, "label": "a red and brown bar in a room"}]}
{"id": 433619, "image": "COCO_train2014_000000433619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car with a dog inside\"."}], "task": "refering", "answer_template": "The \"a car with a dog inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 59, 60, 61, 62, 63, 64, 65, 66, 67, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0015833333333333333, 0.11140625, 1.0, 1.0], "iscrowd": 0, "area": 237922.1843, "rle": {"size": [640, 480], "counts": "PS1o4R:o4N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O1M3N2M3O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O100000000O10000000000000000O1000000000000000000O1000000000001O000O1000000000000000000O100000000000000000000000000000000000000000000000000O1O100O1O1O100O1O1O1O100O1O1O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000OmNT1mDl8"}, "label": "a car with a dog inside"}]}
{"id": 433619, "image": "COCO_train2014_000000433619.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car with a dog sitting on the front seat\"."}], "task": "refering", "answer_template": "The \"a white car with a dog sitting on the front seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 59, 60, 61, 62, 63, 64, 65, 66, 67, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0015833333333333333, 0.11140625, 1.0, 1.0], "iscrowd": 0, "area": 237922.1843, "rle": {"size": [640, 480], "counts": "PS1o4R:o4N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O1M3N2M3O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O100000000O10000000000000000O1000000000000000000O1000000000001O000O1000000000000000000O100000000000000000000000000000000000000000000000000O1O100O1O1O100O1O1O1O100O1O1O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000OmNT1mDl8"}, "label": "a white car with a dog sitting on the front seat"}]}
{"id": 114132, "image": "COCO_train2014_000000114132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed covered in white sheets with pillows\"."}], "task": "refering", "answer_template": "The \"a bed covered in white sheets with pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 179, 180], "bbox": [0.12440384615384614, 0.4268096514745308, 0.5105192307692308, 0.7083378016085791], "iscrowd": 0, "area": 16008.476599999996, "rle": {"size": [373, 520], "counts": "Ulg0j0j:100O2O0O1\\OTO_Fl0a9UO]Fl0b9VO]Fj0c9VO\\Fk0c9WO\\Fj0c9WO[Fj0d9WO\\Fi0d9XO[Fh0d9ZOZFg0f9YOZFg0e9[OYFg0f9ZOYFf0f9\\OXFe0h9[OXFe0g9]OWFd0i9d0O101N100O100O100O2O0O100O10001N100O100O100O2O00000000001O0O10000000001O0000000000001O00000000001O0000000O10001O00000000001O00000000000010O000000000001O01O00000001O00000001O0001O0000000000001N10000O10001N10000O10000O2O000O10000O2O000O10000O101O01O0000000001O0001O000001O000001O000001O00000001O0001O00000001O01O0000000O10010O01O001O0010O01O1O001O00100O001O1Omfl2"}, "label": "a bed covered in white sheets with pillows"}]}
{"id": 114132, "image": "COCO_train2014_000000114132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white mattress with two white pillows\"."}], "task": "refering", "answer_template": "The \"a white mattress with two white pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 123, 135, 136, 137, 138, 139, 140, 141, 142, 154, 155, 156, 157, 158, 159, 160, 161, 179, 180], "bbox": [0.12440384615384614, 0.4268096514745308, 0.5105192307692308, 0.7083378016085791], "iscrowd": 0, "area": 16008.476599999996, "rle": {"size": [373, 520], "counts": "Ulg0j0j:100O2O0O1\\OTO_Fl0a9UO]Fl0b9VO]Fj0c9VO\\Fk0c9WO\\Fj0c9WO[Fj0d9WO\\Fi0d9XO[Fh0d9ZOZFg0f9YOZFg0e9[OYFg0f9ZOYFf0f9\\OXFe0h9[OXFe0g9]OWFd0i9d0O101N100O100O100O2O0O100O10001N100O100O100O2O00000000001O0O10000000001O0000000000001O00000000001O0000000O10001O00000000001O00000000000010O000000000001O01O00000001O00000001O0001O0000000000001N10000O10001N10000O10000O2O000O10000O2O000O10000O101O01O0000000001O0001O000001O000001O000001O00000001O0001O00000001O01O0000000O10010O01O001O0010O01O1O001O00100O001O1Omfl2"}, "label": "a white mattress with two white pillows"}]}
{"id": 114132, "image": "COCO_train2014_000000114132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather bed with a white sheet and pillows\"."}], "task": "refering", "answer_template": "The \"a black leather bed with a white sheet and pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 65, 69, 70, 71, 72, 77, 78, 79, 80, 81, 82, 83, 84, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 104, 106, 107, 108, 109, 110, 111, 112, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 217, 218, 219, 220, 221, 238], "bbox": [0.06609615384615385, 0.1572922252010724, 0.9204038461538462, 0.9415817694369975], "iscrowd": 0, "area": 92055.92729999998, "rle": {"size": [373, 520], "counts": "U^<`0U;U4kKg1YN`0@0000000000000000O100000aKWIR4V7000000000001OJ6M3O1O10000001O000000001O01O0001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000010O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O0001O0001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000010O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O0001O0001O000000001O0000000d0]Ol0SOm0SOd0\\O10O01O001O010O1O001O010O1O001O02O2M10OO10000000007I>B00000000000000001O00000O10000000O10000O100O100O100O10000O10O0100N2O1N2D<00O100000000O100000000O10O10000N101BZKkIh4R6?N2O00100O10O10O100C=lNS1100000O10O100000000O100000000O1000O10O100000000O100000000O10O100000O100000000O1000000O10O100000O100000000O10000000O01000000O100000000O101O1O001N2O1O001Oo0PO2O2N1O1O2M2O1O2N1N2O2N1O000O10000000O0100000000O010000009F<E0000lHkKh6a4000001O000000000000O1000000O10O1000O10000A?O10001OZ1fNa2_MRe>"}, "label": "a black leather bed with a white sheet and pillows"}]}
{"id": 114132, "image": "COCO_train2014_000000114132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed in home\"."}], "task": "refering", "answer_template": "The \"a bed in home\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 65, 69, 70, 71, 72, 77, 78, 79, 80, 81, 82, 83, 84, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 104, 106, 107, 108, 109, 110, 111, 112, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 217, 218, 219, 220, 221, 238], "bbox": [0.06609615384615385, 0.1572922252010724, 0.9204038461538462, 0.9415817694369975], "iscrowd": 0, "area": 92055.92729999998, "rle": {"size": [373, 520], "counts": "U^<`0U;U4kKg1YN`0@0000000000000000O100000aKWIR4V7000000000001OJ6M3O1O10000001O000000001O01O0001O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000010O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O0001O0001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O0000010O000000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O0001O0001O000000001O0000000d0]Ol0SOm0SOd0\\O10O01O001O010O1O001O010O1O001O02O2M10OO10000000007I>B00000000000000001O00000O10000000O10000O100O100O100O10000O10O0100N2O1N2D<00O100000000O100000000O10O10000N101BZKkIh4R6?N2O00100O10O10O100C=lNS1100000O10O100000000O100000000O1000O10O100000000O100000000O10O100000O100000000O1000000O10O100000O100000000O10000000O01000000O100000000O101O1O001N2O1O001Oo0PO2O2N1O1O2M2O1O2N1N2O2N1O000O10000000O0100000000O010000009F<E0000lHkKh6a4000001O000000000000O1000000O10O1000O10000A?O10001OZ1fNa2_MRe>"}, "label": "a bed in home"}]}
{"id": 130518, "image": "COCO_train2014_000000130518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty black chair at a desk\"."}], "task": "refering", "answer_template": "The \"an empty black chair at a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 288, 289, 290, 291, 311, 312, 313, 314, 333, 334, 335, 336, 337, 358, 360, 361, 381], "bbox": [0.483296875, 0.460125, 0.7192812500000001, 0.9579791666666667], "iscrowd": 0, "area": 18073.48295, "rle": {"size": [480, 640], "counts": "PXa42^>`0O100O100O100O100O1000DPBHP>8mLa0\\G@`8P1TGPOi8T1VGlNf8X1ZGhNc8\\1\\GdNa8`1^G`N^8e1aG[N\\8i1cGWNY8m1gGSNV8Q2iGoMT8U2kGlMP8Y2oGgMn7]2QHcMk7a2UH_Mh7e2WH[Mf7i2YHWMc7n2\\HRMa7R3^HnL_7V3`HjL`7V3`HjL`7W3_HiL`7Y3_HgLa7Z3^HfLb7[3]HeLc7[3]HeLd7[3[HeLe7\\3ZHdLd7DkFi3`1cLe7DkFj3^1cLi7]3WHcLi7^3VHbLj7_3UHaLk7`3TH`Ll7a3SH_Lm7c3QH]Lo7e3oG[L;EQ6R4bIZL6JX6n3`IXL3O]6j3_IWL03a6h3]IUL05c6f3]IULM8f6`3`IXLH:h6Z3dI\\LB<j6V3fI^L^O>l6X3bIZL@`0n6Z3^IVLAc0Q7[3ZIRLCe0S7]3VInKEg0U7_3RIjKGi0W7a3nHfKH<HDb7^4jHbKJ<1_O[7g4fH^KL<9[OV7n4aH[KN<b0VOo6W5]HSL[8W52N2N2N2N2N2N2N2N2N2N2N2O11O001O00001O00POQG[KN;P9U4WG`KJ:o8k3bGkK_O:o8a3kGVLWO9m8`3mGWLVO9m8_3nGXLVO9k8^3PHYLUO9l8\\3PH[LTO9l8Z3QH^LTO8k8X3RH`LSO8k8X3RH`LTO8j8W3QHbLUO7j8W3QHaLWO7i8W3oGcLXO7h8V3oGdLZO5h8V3nGeLZO6g8U3nGfL\\O4g8U3mGgL\\O4g8U3lGhL^O3f8T3lGiL^O3f8T3kGjL@2e8S3kGkL@2e8S3jGlLB0e8R3iGoLB0d8Q3jGnLD0c8Q3hGPME0d8n2gGRMFOi8i2aGXMFOP9b2ZG_MF0U9Z2VGfMFOZ9U2oFmMGO_9m1kFTNGNd9h1eFZNGNj9a1aF`NFOo9Z1dF^N]O8U:S1\\GmNi8m0WGSOo8f0SGZOl8f0TGZOk8f0WGYOi8f0XG[Og8e0ZGHX82`ESOY2k0W82bEQOX2l0V83oGMQ83PHLP84PHLP84PHLP84QHKo75QHKo75RHJn76RHJn76SHHn78RH\\OZ8d0Q301O00001O00000MbRd2"}, "label": "an empty black chair at a desk"}]}
{"id": 130518, "image": "COCO_train2014_000000130518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black office chair with a solid plastic back\"."}], "task": "refering", "answer_template": "The \"a black office chair with a solid plastic back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 288, 289, 290, 291, 311, 312, 313, 314, 333, 334, 335, 336, 337, 358, 360, 361, 381], "bbox": [0.483296875, 0.460125, 0.7192812500000001, 0.9579791666666667], "iscrowd": 0, "area": 18073.48295, "rle": {"size": [480, 640], "counts": "PXa42^>`0O100O100O100O100O1000DPBHP>8mLa0\\G@`8P1TGPOi8T1VGlNf8X1ZGhNc8\\1\\GdNa8`1^G`N^8e1aG[N\\8i1cGWNY8m1gGSNV8Q2iGoMT8U2kGlMP8Y2oGgMn7]2QHcMk7a2UH_Mh7e2WH[Mf7i2YHWMc7n2\\HRMa7R3^HnL_7V3`HjL`7V3`HjL`7W3_HiL`7Y3_HgLa7Z3^HfLb7[3]HeLc7[3]HeLd7[3[HeLe7\\3ZHdLd7DkFi3`1cLe7DkFj3^1cLi7]3WHcLi7^3VHbLj7_3UHaLk7`3TH`Ll7a3SH_Lm7c3QH]Lo7e3oG[L;EQ6R4bIZL6JX6n3`IXL3O]6j3_IWL03a6h3]IUL05c6f3]IULM8f6`3`IXLH:h6Z3dI\\LB<j6V3fI^L^O>l6X3bIZL@`0n6Z3^IVLAc0Q7[3ZIRLCe0S7]3VInKEg0U7_3RIjKGi0W7a3nHfKH<HDb7^4jHbKJ<1_O[7g4fH^KL<9[OV7n4aH[KN<b0VOo6W5]HSL[8W52N2N2N2N2N2N2N2N2N2N2N2O11O001O00001O00POQG[KN;P9U4WG`KJ:o8k3bGkK_O:o8a3kGVLWO9m8`3mGWLVO9m8_3nGXLVO9k8^3PHYLUO9l8\\3PH[LTO9l8Z3QH^LTO8k8X3RH`LSO8k8X3RH`LTO8j8W3QHbLUO7j8W3QHaLWO7i8W3oGcLXO7h8V3oGdLZO5h8V3nGeLZO6g8U3nGfL\\O4g8U3mGgL\\O4g8U3lGhL^O3f8T3lGiL^O3f8T3kGjL@2e8S3kGkL@2e8S3jGlLB0e8R3iGoLB0d8Q3jGnLD0c8Q3hGPME0d8n2gGRMFOi8i2aGXMFOP9b2ZG_MF0U9Z2VGfMFOZ9U2oFmMGO_9m1kFTNGNd9h1eFZNGNj9a1aF`NFOo9Z1dF^N]O8U:S1\\GmNi8m0WGSOo8f0SGZOl8f0TGZOk8f0WGYOi8f0XG[Og8e0ZGHX82`ESOY2k0W82bEQOX2l0V83oGMQ83PHLP84PHLP84PHLP84QHKo75QHKo75RHJn76RHJn76SHHn78RH\\OZ8d0Q301O00001O00000MbRd2"}, "label": "a black office chair with a solid plastic back"}]}
{"id": 130518, "image": "COCO_train2014_000000130518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair rests at a desk in front of a computer monitor\"."}], "task": "refering", "answer_template": "The \"a chair rests at a desk in front of a computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 256, 257, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 324, 325], "bbox": [0.05, 0.31612500000000004, 0.3, 0.8602083333333334], "iscrowd": 0, "area": 18411.1268, "rle": {"size": [480, 640], "counts": "hT?5k>8H3N2M3M3N2M2N3M3N1N5K4nDoNn7T1jGQOMAl5c1nIPO6]Ol5f1fISO<XOn5h1_IUOa0TOQ6i1VIXOi0oNQ6l1mH\\OP1iNS6n1iH[OR1hNV6n1dH\\OU1gNW6P2`H[OW1fNY6Q2\\H\\OY1dN[6S2WH[O]1cN]6T2QH\\O`1aN_6V2lG[Od1`N`6X2gGZOh1_Nb6U4]IkKc6U4]IkKc6U4\\IlKd6U4[IkKe6U4[IkKe6U4[IlKd6U4ZIlKf6T4ZIkKg6V4XIjKh6V4XIiKi6W4VIjKj6W4UIhKl6X4TIhKl6X4TIgKn6Y4PIhKAD^6d4QJgK^OJ`6_4RJgK[ONb6[4SJfKYO4b6V4TJgKYO5b6T4UJfKYO8a6S4UJeKZO9`6R4VJdKZO=^6o3WJeKZO>^6m3XJdKZO?_6m3WJdKYO?a6m3UJdK[O?`6m3UJdKZO?b6n3SJcKZO?d6n3RJbK[O?c6P4RJaKZO?e6P4PJbKZO>g6P4oIaKZO`0g6o3oIaKZO?h6Q4mI`KZO?j6Q4lI_KZO`0k6Q4kI_KYO`0m6Q4jI_KXO`0n6R4iI^KYO`0o6R4hI^KXO`0Q7R4gI^KWO`0S7R4fI]KWO`0U7S4dI]KUO>[7U4`I]KTO<_7W4]I^KRO8e7Z4XI_KRO6g7\\4WI^KPO6k7\\4UI^KoN5n7]4SI_KmN3R8^4QI_KkN2W8_4nHjKT7V4lHiKU7W4jHhKY7W4gHgK[7Y4eHeK]7[4cHeK]7[4eGWK`0d0k7U4gGYK4h0V8n3hG\\KFm0c8g3hG^LX8b3hG^LX8b3iG\\LY8c3hG\\LX8d3hG\\LX8e3hGZLX8f3kGVLW8i3lGTLU8k3nGRLS8m3QHoKo7Q4SHlKo7S4QHnKo7Q4QHPLn7P4RHRLm7m3SHTLm7k3SHVLl7j3THXLk7f3VH[Lj7c3WH^Li7`3XHbLf7]3[HdLe7Z3\\HhLc7V3]HTLoN=c8]3_HVLnN>d8Y3_HYLnN?c8S3bH^LkNa0d8`2oHnL_Na0d8W2UIXMWNb0e8l1\\IbMoMb0g8b1bIlMhMa0h8NSKAULb0j8VOeK8bKa0^=@cB?]=AeB>Z=BgB=Y=ChB=W=DiB;W=EjB:V=FkB:T=FmB9S=GmB9R=HoB8P=HQC7o<IRC6n<JSC6l<KSC5l<LUC3k<MVC3i<MWC3i<MXC3g<MYC3f<N[C1e<O\\C0d<0\\C0d<O^C0b<0_COa<1`CN`<2`CN`<2aCM_<3bCL^<4bCL^<4cCK]<4eCCc<=Q1O000000000000000N2L5K`ea6"}, "label": "a chair rests at a desk in front of a computer monitor"}]}
{"id": 130518, "image": "COCO_train2014_000000130518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray office chair sits behind a computer screen with green letters\"."}], "task": "refering", "answer_template": "The \"a gray office chair sits behind a computer screen with green letters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 256, 257, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 324, 325], "bbox": [0.05, 0.31612500000000004, 0.3, 0.8602083333333334], "iscrowd": 0, "area": 18411.1268, "rle": {"size": [480, 640], "counts": "hT?5k>8H3N2M3M3N2M2N3M3N1N5K4nDoNn7T1jGQOMAl5c1nIPO6]Ol5f1fISO<XOn5h1_IUOa0TOQ6i1VIXOi0oNQ6l1mH\\OP1iNS6n1iH[OR1hNV6n1dH\\OU1gNW6P2`H[OW1fNY6Q2\\H\\OY1dN[6S2WH[O]1cN]6T2QH\\O`1aN_6V2lG[Od1`N`6X2gGZOh1_Nb6U4]IkKc6U4]IkKc6U4\\IlKd6U4[IkKe6U4[IkKe6U4[IlKd6U4ZIlKf6T4ZIkKg6V4XIjKh6V4XIiKi6W4VIjKj6W4UIhKl6X4TIhKl6X4TIgKn6Y4PIhKAD^6d4QJgK^OJ`6_4RJgK[ONb6[4SJfKYO4b6V4TJgKYO5b6T4UJfKYO8a6S4UJeKZO9`6R4VJdKZO=^6o3WJeKZO>^6m3XJdKZO?_6m3WJdKYO?a6m3UJdK[O?`6m3UJdKZO?b6n3SJcKZO?d6n3RJbK[O?c6P4RJaKZO?e6P4PJbKZO>g6P4oIaKZO`0g6o3oIaKZO?h6Q4mI`KZO?j6Q4lI_KZO`0k6Q4kI_KYO`0m6Q4jI_KXO`0n6R4iI^KYO`0o6R4hI^KXO`0Q7R4gI^KWO`0S7R4fI]KWO`0U7S4dI]KUO>[7U4`I]KTO<_7W4]I^KRO8e7Z4XI_KRO6g7\\4WI^KPO6k7\\4UI^KoN5n7]4SI_KmN3R8^4QI_KkN2W8_4nHjKT7V4lHiKU7W4jHhKY7W4gHgK[7Y4eHeK]7[4cHeK]7[4eGWK`0d0k7U4gGYK4h0V8n3hG\\KFm0c8g3hG^LX8b3hG^LX8b3iG\\LY8c3hG\\LX8d3hG\\LX8e3hGZLX8f3kGVLW8i3lGTLU8k3nGRLS8m3QHoKo7Q4SHlKo7S4QHnKo7Q4QHPLn7P4RHRLm7m3SHTLm7k3SHVLl7j3THXLk7f3VH[Lj7c3WH^Li7`3XHbLf7]3[HdLe7Z3\\HhLc7V3]HTLoN=c8]3_HVLnN>d8Y3_HYLnN?c8S3bH^LkNa0d8`2oHnL_Na0d8W2UIXMWNb0e8l1\\IbMoMb0g8b1bIlMhMa0h8NSKAULb0j8VOeK8bKa0^=@cB?]=AeB>Z=BgB=Y=ChB=W=DiB;W=EjB:V=FkB:T=FmB9S=GmB9R=HoB8P=HQC7o<IRC6n<JSC6l<KSC5l<LUC3k<MVC3i<MWC3i<MXC3g<MYC3f<N[C1e<O\\C0d<0\\C0d<O^C0b<0_COa<1`CN`<2`CN`<2aCM_<3bCL^<4bCL^<4cCK]<4eCCc<=Q1O000000000000000N2L5K`ea6"}, "label": "a gray office chair sits behind a computer screen with green letters"}]}
{"id": 171479, "image": "COCO_train2014_000000171479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young skier in red pants with white writing , racing another skier\"."}], "task": "refering", "answer_template": "The \"a young skier in red pants with white writing , racing another skier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 174, 175, 176, 197, 198, 199, 200, 220, 221, 222, 223], "bbox": [0.5389375000000001, 0.18358490566037736, 0.711625, 0.6330188679245282], "iscrowd": 0, "area": 8697.428249999999, "rle": {"size": [424, 640], "counts": "RP_41U=3M3M2N3N2M2N3M3M1O2N1O2O0O2N101N1O2O0O2O0O2N101O0O2O002M2H9_Oa0N101N1fFkNa6V1XIYO^6i0[IE]6<\\IO_64XI2i60nH6R7LfH:[7i21O010O1O001O001PO_HXLa7d3iHULW7h3TIPLn6l3XIPLj6k3^IPLc6l3dIPL^6l3e1H7J7H5K3N2M3M3N2M3M33N2N2O1N2N2N2O2M1O2O0O10001N1000O019F6K2M2O2Mc0^O3L4M3L4M3M2M<E0O2O0O100O1N3N1N2O1N5K6J5KR[\\2"}, "label": "a young skier in red pants with white writing , racing another skier"}]}
{"id": 171479, "image": "COCO_train2014_000000171479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right skier in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the right skier in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 174, 175, 176, 197, 198, 199, 200, 220, 221, 222, 223], "bbox": [0.5389375000000001, 0.18358490566037736, 0.711625, 0.6330188679245282], "iscrowd": 0, "area": 8697.428249999999, "rle": {"size": [424, 640], "counts": "RP_41U=3M3M2N3N2M2N3M3M1O2N1O2O0O2N101N1O2O0O2O0O2N101O0O2O002M2H9_Oa0N101N1fFkNa6V1XIYO^6i0[IE]6<\\IO_64XI2i60nH6R7LfH:[7i21O010O1O001O001PO_HXLa7d3iHULW7h3TIPLn6l3XIPLj6k3^IPLc6l3dIPL^6l3e1H7J7H5K3N2M3M3N2M3M33N2N2O1N2N2N2O2M1O2O0O10001N1000O019F6K2M2O2Mc0^O3L4M3L4M3M2M<E0O2O0O100O1N3N1N2O1N5K6J5KR[\\2"}, "label": "the right skier in the right hand picture"}]}
{"id": 24026, "image": "COCO_train2014_000000024026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottle of water for the guy in the yellow short\"."}], "task": "refering", "answer_template": "The \"the bottle of water for the guy in the yellow short\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 249, 250, 251, 272, 273, 274, 295, 296, 297, 318, 319, 320, 341, 342, 343], "bbox": [0.822140625, 0.5956908665105387, 0.94840625, 1.0], "iscrowd": 0, "area": 10032.055499999999, "rle": {"size": [427, 640], "counts": "kgk62S=8I6hNKRE<b::nDKh:b1D;O10000[OZMdFg2n8iMPGW2f8UNXGk1f8YNXGg1e8^NYGb1d8bN[G^1b8gN[GZ1c8h1O10001N100000000O10000O10000O100L4I8N10000O1000000O10000O1000000O2O0001O<D2N1O1O2N1N2O2M1SMYGU1i8jNYGT1g8lNZGR1h8mNZGQ1f8oN[Gk0k8TOWG=V9CkFNc92_F_Oo9`0RFRO[:n0gEbNh:]1d0O2N2N2N2M3D;F;E[_="}, "label": "the bottle of water for the guy in the yellow short"}]}
{"id": 24026, "image": "COCO_train2014_000000024026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray water bottle with a black lid\"."}], "task": "refering", "answer_template": "The \"a gray water bottle with a black lid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 249, 250, 251, 272, 273, 274, 295, 296, 297, 318, 319, 320, 341, 342, 343], "bbox": [0.822140625, 0.5956908665105387, 0.94840625, 1.0], "iscrowd": 0, "area": 10032.055499999999, "rle": {"size": [427, 640], "counts": "kgk62S=8I6hNKRE<b::nDKh:b1D;O10000[OZMdFg2n8iMPGW2f8UNXGk1f8YNXGg1e8^NYGb1d8bN[G^1b8gN[GZ1c8h1O10001N100000000O10000O10000O100L4I8N10000O1000000O10000O1000000O2O0001O<D2N1O1O2N1N2O2M1SMYGU1i8jNYGT1g8lNZGR1h8mNZGQ1f8oN[Gk0k8TOWG=V9CkFNc92_F_Oo9`0RFRO[:n0gEbNh:]1d0O2N2N2N2M3D;F;E[_="}, "label": "a gray water bottle with a black lid"}]}
{"id": 286174, "image": "COCO_train2014_000000286174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table covered in wine bottles and glasses\"."}], "task": "refering", "answer_template": "The \"the table covered in wine bottles and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.0, 0.5488235294117647, 0.515625, 1.0], "iscrowd": 0, "area": 45129.5, "rle": {"size": [425, 640], "counts": "]<l0d:i1WNi1SOm0O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O11O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O00001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O00000kKlGU3T8iLQHT3o7jLVHS3j7kL[HR3f7kL_HR3a7lLcHR3]7mLfHQ3Z7oLgHP3Y7PMiHn2W7RMkHl2U7TMmHj2T7UMmHj2S7VMoHh2Q7XMPIg2P7YMRIe2n6[MSId2m6\\MTIc2l6]MTIc2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6[MTIe2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMRIg2n6YMRIg2o6XMQIh2o6YMPIg2P7YMPIg2P7YMPIg2P7YMPIg2P7ZMoHf2Q7ZMoHf2Q7ZMoHf2R7ZMmHf2S7ZMmHf2S7ZMmHf2S7ZMmHf2S7[MlHe2T7[MlHe2T7[MlHe2U7[MjHe2V7[MjHe2V7[MjHe2V7[MjHe2V7\\MhHe2X7[MhHe2X7[MhHe2Y7[MfHe2Z7[MfHe2Z7[MfHe2Z7[MfHe2Z7\\MeHd2[7\\MeHd2[7\\MeHc2\\7^McHb2^7]MbHc2^7]MbHc2^7]MbHc2^7^MaHb2_7^MaHb2_7^MaHb2_7_M`H`2b7_M^Ha2b7_M^Ha2b7_M^Ha2b7`M\\Ha2d7_M\\Ha2d7_M\\Ha2d7_M\\H`2f7^M[Hb2e7^M[Hb2e7^M[Hb2e7^M[Hb2e7]M\\Hc2d7]M\\Hc2d7]M\\Hb2e7]M\\Hc2e7\\M[Hd2e7\\M[Hd2e7[M[Hf2e7ZM[Hf2e7ZM[Hf2e7ZM[He2f7ZM[Hf2f7YMZHg2f7YMZHg2f7XM[Hh2e7XM[Hh2e7XM[Hh2e7XM[Hg2f7XMZHi2g7VMYHj2g7VMYHj2g7UMZHk2f7UMZHk2f7UMZHk2f7TM[Hk2f7UMZHk2f7UMZHk2g7TMYHl2g7SMZHm2f7SMZHm2f7SMYHn2g7QMZHn2g7RMYHn2g7RMYHn2h7QMXHo2h7PMYHP3g7PMYHP3g7PMYHP3P9O1000O10000000O100000000000O10O1000000000000O1000000000000O100L4H8H8H7I8H8H8H8J6M3M3L4M2N2M3N2N2M4Ma[P4"}, "label": "the table covered in wine bottles and glasses"}]}
{"id": 286174, "image": "COCO_train2014_000000286174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a camera sits on a table next to a man ' s hand gripping a wine glass\"."}], "task": "refering", "answer_template": "The \"a camera sits on a table next to a man ' s hand gripping a wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.0, 0.5488235294117647, 0.515625, 1.0], "iscrowd": 0, "area": 45129.5, "rle": {"size": [425, 640], "counts": "]<l0d:i1WNi1SOm0O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O11O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O00001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O00000kKlGU3T8iLQHT3o7jLVHS3j7kL[HR3f7kL_HR3a7lLcHR3]7mLfHQ3Z7oLgHP3Y7PMiHn2W7RMkHl2U7TMmHj2T7UMmHj2S7VMoHh2Q7XMPIg2P7YMRIe2n6[MSId2m6\\MTIc2l6]MTIc2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6[MTIe2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMRIg2n6YMRIg2o6XMQIh2o6YMPIg2P7YMPIg2P7YMPIg2P7YMPIg2P7ZMoHf2Q7ZMoHf2Q7ZMoHf2R7ZMmHf2S7ZMmHf2S7ZMmHf2S7ZMmHf2S7[MlHe2T7[MlHe2T7[MlHe2U7[MjHe2V7[MjHe2V7[MjHe2V7[MjHe2V7\\MhHe2X7[MhHe2X7[MhHe2Y7[MfHe2Z7[MfHe2Z7[MfHe2Z7[MfHe2Z7\\MeHd2[7\\MeHd2[7\\MeHc2\\7^McHb2^7]MbHc2^7]MbHc2^7]MbHc2^7^MaHb2_7^MaHb2_7^MaHb2_7_M`H`2b7_M^Ha2b7_M^Ha2b7_M^Ha2b7`M\\Ha2d7_M\\Ha2d7_M\\Ha2d7_M\\H`2f7^M[Hb2e7^M[Hb2e7^M[Hb2e7^M[Hb2e7]M\\Hc2d7]M\\Hc2d7]M\\Hb2e7]M\\Hc2e7\\M[Hd2e7\\M[Hd2e7[M[Hf2e7ZM[Hf2e7ZM[Hf2e7ZM[He2f7ZM[Hf2f7YMZHg2f7YMZHg2f7XM[Hh2e7XM[Hh2e7XM[Hh2e7XM[Hg2f7XMZHi2g7VMYHj2g7VMYHj2g7UMZHk2f7UMZHk2f7UMZHk2f7TM[Hk2f7UMZHk2f7UMZHk2g7TMYHl2g7SMZHm2f7SMZHm2f7SMYHn2g7QMZHn2g7RMYHn2g7RMYHn2h7QMXHo2h7PMYHP3g7PMYHP3g7PMYHP3P9O1000O10000000O100000000000O10O1000000000000O1000000000000O100L4H8H8H7I8H8H8H8J6M3M3L4M2N2M3N2N2M4Ma[P4"}, "label": "a camera sits on a table next to a man ' s hand gripping a wine glass"}]}
{"id": 286174, "image": "COCO_train2014_000000286174.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black table with wine bottles , wine glasses , and a camera on it\"."}], "task": "refering", "answer_template": "The \"a black table with wine bottles , wine glasses , and a camera on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.0, 0.5488235294117647, 0.515625, 1.0], "iscrowd": 0, "area": 45129.5, "rle": {"size": [425, 640], "counts": "]<l0d:i1WNi1SOm0O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O11O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O00001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O00000kKlGU3T8iLQHT3o7jLVHS3j7kL[HR3f7kL_HR3a7lLcHR3]7mLfHQ3Z7oLgHP3Y7PMiHn2W7RMkHl2U7TMmHj2T7UMmHj2S7VMoHh2Q7XMPIg2P7YMRIe2n6[MSId2m6\\MTIc2l6]MTIc2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6\\MSId2m6[MTIe2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMSIf2m6ZMRIg2n6YMRIg2o6XMQIh2o6YMPIg2P7YMPIg2P7YMPIg2P7YMPIg2P7ZMoHf2Q7ZMoHf2Q7ZMoHf2R7ZMmHf2S7ZMmHf2S7ZMmHf2S7ZMmHf2S7[MlHe2T7[MlHe2T7[MlHe2U7[MjHe2V7[MjHe2V7[MjHe2V7[MjHe2V7\\MhHe2X7[MhHe2X7[MhHe2Y7[MfHe2Z7[MfHe2Z7[MfHe2Z7[MfHe2Z7\\MeHd2[7\\MeHd2[7\\MeHc2\\7^McHb2^7]MbHc2^7]MbHc2^7]MbHc2^7^MaHb2_7^MaHb2_7^MaHb2_7_M`H`2b7_M^Ha2b7_M^Ha2b7_M^Ha2b7`M\\Ha2d7_M\\Ha2d7_M\\Ha2d7_M\\H`2f7^M[Hb2e7^M[Hb2e7^M[Hb2e7^M[Hb2e7]M\\Hc2d7]M\\Hc2d7]M\\Hb2e7]M\\Hc2e7\\M[Hd2e7\\M[Hd2e7[M[Hf2e7ZM[Hf2e7ZM[Hf2e7ZM[He2f7ZM[Hf2f7YMZHg2f7YMZHg2f7XM[Hh2e7XM[Hh2e7XM[Hh2e7XM[Hg2f7XMZHi2g7VMYHj2g7VMYHj2g7UMZHk2f7UMZHk2f7UMZHk2f7TM[Hk2f7UMZHk2f7UMZHk2g7TMYHl2g7SMZHm2f7SMZHm2f7SMYHn2g7QMZHn2g7RMYHn2g7RMYHn2h7QMXHo2h7PMYHP3g7PMYHP3g7PMYHP3P9O1000O10000000O100000000000O10O1000000000000O1000000000000O100L4H8H8H7I8H8H8H8J6M3M3L4M2N2M3N2N2M4Ma[P4"}, "label": "a black table with wine bottles , wine glasses , and a camera on it"}]}
{"id": 155105, "image": "COCO_train2014_000000155105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of pasta and meat\"."}], "task": "refering", "answer_template": "The \"a bowl of pasta and meat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 365, 366, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.399625, 0.676078125, 1.0, 1.0], "iscrowd": 0, "area": 50723.986549999994, "rle": {"size": [640, 480], "counts": "n_h36fc09H9F9H9F5L5J5N3M2N3M3M2O2M3M2N3M2N3M3M2N3M3M2O2M2N3M2N1O2O1N2O0O2O1N2O0O2O1N2O0O2N2O0O2O1N2O0O2O1N2O0O2O1N2O0O2O1N101N2O1N2O1N101O1N2O1O1O1N101O1O1N2O1O1N101O1O1N2O1O1O0O2O1O001N101O0O2O1O001O001O001N101O1O001O001N101O001O001O1O0O2O001O001O001O0O100000000000000O100000000000000O100000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O001O001O001O001O001O001O001L3MnA"}, "label": "a bowl of pasta and meat"}]}
{"id": 155105, "image": "COCO_train2014_000000155105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the plate with the rice on it\"."}], "task": "refering", "answer_template": "The \"the plate with the rice on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 365, 366, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.399625, 0.676078125, 1.0, 1.0], "iscrowd": 0, "area": 50723.986549999994, "rle": {"size": [640, 480], "counts": "n_h36fc09H9F9H9F5L5J5N3M2N3M3M2O2M3M2N3M2N3M3M2N3M3M2O2M2N3M2N1O2O1N2O0O2O1N2O0O2O1N2O0O2N2O0O2O1N2O0O2O1N2O0O2O1N2O0O2O1N101N2O1N2O1N101O1N2O1O1O1N101O1O1N2O1O1N101O1O1N2O1O1O0O2O1O001N101O0O2O1O001O001O001N101O1O001O001N101O001O001O1O0O2O001O001O001O0O100000000000000O100000000000000O100000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O001O001O001O001O001O001O001L3MnA"}, "label": "the plate with the rice on it"}]}
{"id": 310759, "image": "COCO_train2014_000000310759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white man in a plaid dress shirt\"."}], "task": "refering", "answer_template": "The \"a white man in a plaid dress shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.6627500000000001, 0.06941724941724942, 1.0, 0.9836363636363636], "iscrowd": 0, "area": 57660.263050000016, "rle": {"size": [429, 640], "counts": "]ka52T=8G9H8H8G9H9G;D;F:F:J6M3M4K4M3M3L4M3M4L3L2N2oJjKi0X4ROkKm0W4POkKo0X4mNjKR1^4gNbKY1e4`N[K_1m4ZNSKe1U5TNkJl1\\5mMdJS2a5hM_JW2b5iM^JW2b5iM^JV2d5iM\\JW2d5iM\\JW2d5iM\\JV2e5jM[JV2f5iMZJL3Dc5`0ZJH=AZ5g0YJGc0]OT5l0YJFg0[OP5o0YJEl0WOm4S1WJFP1SOi4W1WJES1ROf4Y1WJDV1QOc4[1WJBY1QOa4]1VJA\\1PO_4^1UJ@^1RO]4^1UJXOHaNg1i0\\4^1UJWOI`Nh1j0[4_1TJVOJ_Ni1l0Z4^1SJVOK_Ni1m0Y4^1SJUOj1]OS4^1SJTOk1]OS4_1RJSOl1^OR4_1RJSOl1]OT4_1PJSOm1]OT4`1oISOm1\\OU4_1PJUOl1[OT4^1RJVOm1YOR4_1SJXOm1VOR4_1SJZOn1UOo3_1UJ\\On1lNT4f1PJ]Oj8a0XG_Oh8?ZGAf8=\\GBf8;\\GEd89^GGP9ISG5T9CnF=R9APG?o8@SG?n8_OTGa0l8]OVGc0j8[OXGd0h8[OZGe0f8ZO[Gf0e8XO]Gh0c8VO_Gj0a8TOaGl0^8SOdGm0\\8QOfGo0Z8oNhGQ1X8mNjGS1U8lNmGT1V:0000000000000000000000000000001O1O1O1O1O003M2N2N3M2N2N3M2N2N2N2N3M2N2TFPNW8e3N1O1O1O2N1O1O1O2N1O1O0000001O00001O00001O0000001O00001O0000000000000000000000000000000001O000000000000000000000001O2N2N3M2N2N2N3M2N2N3M2N2PLeJJ]54eJJ^53dJK^53dJK^52fJK\\53fJK]52eJL]52eJL]5TNTJ:a0`1^5RNSJ=a0_1^5QNSJ?a0^1^5PNTJ`0`0^1^5oMTJb0`0]1_5mMSJe0?]1`5lMRJf0`0\\1`5kMRJh0`0[1`5jMSJi0?[1a5hMRJl0?Z1a5gMRJn0?Y1a5fMRJQ1>W1c5dMQJV1;U1f5bMQJY1c0j0^5jMRJ[1]1Mc4eNRJ^1X2oNmL"}, "label": "a white man in a plaid dress shirt"}]}
{"id": 310759, "image": "COCO_train2014_000000310759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a plaid shirt\"."}], "task": "refering", "answer_template": "The \"a man in a plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.6627500000000001, 0.06941724941724942, 1.0, 0.9836363636363636], "iscrowd": 0, "area": 57660.263050000016, "rle": {"size": [429, 640], "counts": "]ka52T=8G9H8H8G9H9G;D;F:F:J6M3M4K4M3M3L4M3M4L3L2N2oJjKi0X4ROkKm0W4POkKo0X4mNjKR1^4gNbKY1e4`N[K_1m4ZNSKe1U5TNkJl1\\5mMdJS2a5hM_JW2b5iM^JW2b5iM^JV2d5iM\\JW2d5iM\\JW2d5iM\\JV2e5jM[JV2f5iMZJL3Dc5`0ZJH=AZ5g0YJGc0]OT5l0YJFg0[OP5o0YJEl0WOm4S1WJFP1SOi4W1WJES1ROf4Y1WJDV1QOc4[1WJBY1QOa4]1VJA\\1PO_4^1UJ@^1RO]4^1UJXOHaNg1i0\\4^1UJWOI`Nh1j0[4_1TJVOJ_Ni1l0Z4^1SJVOK_Ni1m0Y4^1SJUOj1]OS4^1SJTOk1]OS4_1RJSOl1^OR4_1RJSOl1]OT4_1PJSOm1]OT4`1oISOm1\\OU4_1PJUOl1[OT4^1RJVOm1YOR4_1SJXOm1VOR4_1SJZOn1UOo3_1UJ\\On1lNT4f1PJ]Oj8a0XG_Oh8?ZGAf8=\\GBf8;\\GEd89^GGP9ISG5T9CnF=R9APG?o8@SG?n8_OTGa0l8]OVGc0j8[OXGd0h8[OZGe0f8ZO[Gf0e8XO]Gh0c8VO_Gj0a8TOaGl0^8SOdGm0\\8QOfGo0Z8oNhGQ1X8mNjGS1U8lNmGT1V:0000000000000000000000000000001O1O1O1O1O003M2N2N3M2N2N3M2N2N2N2N3M2N2TFPNW8e3N1O1O1O2N1O1O1O2N1O1O0000001O00001O00001O0000001O00001O0000000000000000000000000000000001O000000000000000000000001O2N2N3M2N2N2N3M2N2N3M2N2PLeJJ]54eJJ^53dJK^53dJK^52fJK\\53fJK]52eJL]52eJL]5TNTJ:a0`1^5RNSJ=a0_1^5QNSJ?a0^1^5PNTJ`0`0^1^5oMTJb0`0]1_5mMSJe0?]1`5lMRJf0`0\\1`5kMRJh0`0[1`5jMSJi0?[1a5hMRJl0?Z1a5gMRJn0?Y1a5fMRJQ1>W1c5dMQJV1;U1f5bMQJY1c0j0^5jMRJ[1]1Mc4eNRJ^1X2oNmL"}, "label": "a man in a plaid shirt"}]}
{"id": 310759, "image": "COCO_train2014_000000310759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with beard wearing blur shirt and tie\"."}], "task": "refering", "answer_template": "The \"a man with beard wearing blur shirt and tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 31, 32, 33, 34, 54, 55, 56, 57, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.114484375, 0.0, 0.77275, 1.0], "iscrowd": 0, "area": 102080.14545000005, "rle": {"size": [429, 640], "counts": "gln01Z=3L3M4M3L3N3L4M2M4L4M2M4M3L3N3L4M2M4L4L3M4L4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L4L4L3M4M3M3M3N2M3M3N2M3M3M2O2M3M3M3N2M3M3M3N2M3MV1kNa0^O2N2N2O1N2N2N2O1O1O1O100O1O1O1O100O1O1O100O1O1O001NiFnKW9Q4iFoKX9Q42O100O1O100O100O1O100O100O100O1000000O1000000O1000000O1000000O100000000O1000000O1000000O0100000O1000000O10000O1000000O100cLPGZ2S9fMQGS2R9mMQGm1R9SNRGg1P9YNTG`1o8`N[GS1h8mNbGf0a8ZOi2000000000000000000001O0000000000000000000000000O1000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000O10000000000000000aEKf75XH0e70YH5d7KYH;d7EZH`0c7AZHc0d7]OWHk0f7UOQHW1l7iNkGc1R8]NjGi1T8WNhGn1W8RNfGS2X8mMeGX2Y8hMdG\\2[8dMbGa2\\8_M`Gg2^8YM_Gk2`8UM\\GP3c8PMZGS3f8mLVGW3j8iLRG[3n8eLnF_3R9aLjFc3V9;O000000001O000000001O1O001O1O001N2O001O1O001O001O1O001O1O001O1O001O1O001O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2oL[JVNg5h1PKaMR5^2aKnLa4P3aKnLa4P3aKnLb4P3_KnLd4o2]KPMf4n2[KPMh4m2ZKQMi4m2XKQMk4l2WKRMl4k2UKTMm4k2TKSMo4j2SKSMQ5k2PKSMS5j2oJTMT5j2mJTMV5i2kJVMX5g2jJWMY5g2hJWM[5f2gJXM\\5f2eJXM^5e2jJSMY5k2gJTM[5j2fJUM]5h2cJXM`5f2aJXMb5e2e2O2M3N2XNREk0P;ROREm0Q;oNQEQ1Q;lNPES1S;iNnDX1S;fNnDY1U;cNmD]1U;`NlD_1b;000O10000000000O10008H9G7I0O01O1O1O1O10O01N2L4L4L3N3L3N3L3N3LmDkNT:R1lEQOU:n0hETOY:m0dEUO]:j0`EXOa:i0[EZOf:f0UE]Ol:c0QE@P;`0kDCV;=eDH\\;S11N3M2N2N3M2N3M3M2N3M3M2N3M3M2N3M3O001O1N2N1O2O1N1OZ]l1"}, "label": "a man with beard wearing blur shirt and tie"}]}
{"id": 310759, "image": "COCO_train2014_000000310759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bearded man in a blue shirt with a biblical tie\"."}], "task": "refering", "answer_template": "The \"a bearded man in a blue shirt with a biblical tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 31, 32, 33, 34, 54, 55, 56, 57, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.114484375, 0.0, 0.77275, 1.0], "iscrowd": 0, "area": 102080.14545000005, "rle": {"size": [429, 640], "counts": "gln01Z=3L3M4M3L3N3L4M2M4L4M2M4M3L3N3L4M2M4L4L3M4L4L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4L4L4L4L4L4L3M4M3M3M3N2M3M3N2M3M3M2O2M3M3M3N2M3M3M3N2M3MV1kNa0^O2N2N2O1N2N2N2O1O1O1O100O1O1O1O100O1O1O100O1O1O001NiFnKW9Q4iFoKX9Q42O100O1O100O100O1O100O100O100O1000000O1000000O1000000O1000000O100000000O1000000O1000000O0100000O1000000O10000O1000000O100cLPGZ2S9fMQGS2R9mMQGm1R9SNRGg1P9YNTG`1o8`N[GS1h8mNbGf0a8ZOi2000000000000000000001O0000000000000000000000000O1000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000O10000000000000000aEKf75XH0e70YH5d7KYH;d7EZH`0c7AZHc0d7]OWHk0f7UOQHW1l7iNkGc1R8]NjGi1T8WNhGn1W8RNfGS2X8mMeGX2Y8hMdG\\2[8dMbGa2\\8_M`Gg2^8YM_Gk2`8UM\\GP3c8PMZGS3f8mLVGW3j8iLRG[3n8eLnF_3R9aLjFc3V9;O000000001O000000001O1O001O1O001N2O001O1O001O001O1O001O1O001O1O001O1O001O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2oL[JVNg5h1PKaMR5^2aKnLa4P3aKnLa4P3aKnLb4P3_KnLd4o2]KPMf4n2[KPMh4m2ZKQMi4m2XKQMk4l2WKRMl4k2UKTMm4k2TKSMo4j2SKSMQ5k2PKSMS5j2oJTMT5j2mJTMV5i2kJVMX5g2jJWMY5g2hJWM[5f2gJXM\\5f2eJXM^5e2jJSMY5k2gJTM[5j2fJUM]5h2cJXM`5f2aJXMb5e2e2O2M3N2XNREk0P;ROREm0Q;oNQEQ1Q;lNPES1S;iNnDX1S;fNnDY1U;cNmD]1U;`NlD_1b;000O10000000000O10008H9G7I0O01O1O1O1O10O01N2L4L4L3N3L3N3L3N3LmDkNT:R1lEQOU:n0hETOY:m0dEUO]:j0`EXOa:i0[EZOf:f0UE]Ol:c0QE@P;`0kDCV;=eDH\\;S11N3M2N2N3M2N3M3M2N3M3M2N3M3M2N3M3O001O1N2N1O2O1N1OZ]l1"}, "label": "a bearded man in a blue shirt with a biblical tie"}]}
{"id": 318955, "image": "COCO_train2014_000000318955.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow\"."}], "task": "refering", "answer_template": "The \"a cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 192, 193, 194, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245], "bbox": [0.18925, 0.45740625, 0.5075208333333333, 0.6315937500000001], "iscrowd": 0, "area": 8800.888249999998, "rle": {"size": [640, 480], "counts": "aXi11nc01O1O2N1O1O2N1O100O001O1O0O2M2O2M3N1O2M2O2M3N101O1O001O1O1O1O1O1O1O1O1O001O100O1O1O1O1O10O01O1O1O1O100O1O1O1O101N1O1O1O1O2O0000001O0O1000001O00000O2O000000000000O010000000O1000O1000O10nN^Nf_Ob1Z`0`Nd_O`1\\`0aNc_O_1]`0cN`_O^1``0cN__O]1a`0eN]_OZ1d`0hNZ_OX1f`0iNY_OW1g`0Q1000000000O1000000000000O10000000003M2M3N2N2N3M`0_O2O1O1N2O1O1N2O1O1N2O1O1O1N2O1N2N2M3N2M4L4M3L4M2M4M5J<DcTc4"}, "label": "a cow"}]}
{"id": 318955, "image": "COCO_train2014_000000318955.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cow sitting outside of the gate of its habitat\"."}], "task": "refering", "answer_template": "The \"a white cow sitting outside of the gate of its habitat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 192, 193, 194, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245], "bbox": [0.18925, 0.45740625, 0.5075208333333333, 0.6315937500000001], "iscrowd": 0, "area": 8800.888249999998, "rle": {"size": [640, 480], "counts": "aXi11nc01O1O2N1O1O2N1O100O001O1O0O2M2O2M3N1O2M2O2M3N101O1O001O1O1O1O1O1O1O1O1O001O100O1O1O1O1O10O01O1O1O1O100O1O1O1O101N1O1O1O1O2O0000001O0O1000001O00000O2O000000000000O010000000O1000O1000O10nN^Nf_Ob1Z`0`Nd_O`1\\`0aNc_O_1]`0cN`_O^1``0cN__O]1a`0eN]_OZ1d`0hNZ_OX1f`0iNY_OW1g`0Q1000000000O1000000000000O10000000003M2M3N2N2N3M`0_O2O1O1N2O1O1N2O1O1N2O1O1O1N2O1N2N2M3N2M4L4M3L4M2M4M5J<DcTc4"}, "label": "a white cow sitting outside of the gate of its habitat"}]}
{"id": 15851, "image": "COCO_train2014_000000015851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy pushing a bike\"."}], "task": "refering", "answer_template": "The \"a boy pushing a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 174, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 265, 284, 285, 287, 288, 307, 310, 311], "bbox": [0.34954687500000003, 0.23256756756756758, 0.614875, 0.8700225225225224], "iscrowd": 0, "area": 13228.42265, "rle": {"size": [444, 640], "counts": "k_Q32h=2M3N2N2M3N2N2M3N2N3L2O1O1N2O1O1000O1N2N7I7IbT2CnkM1O2N1O2N2]OG_C<S;DaE7[O7P;D^E=COl:EZEe0JFg:HXEj01@b:HWEP17YO\\:J\\EP18UOX:MaEP16POX:3aEo08kNU:9cEn07gNU:=dEn07aNV:b0cEn08]NU94TGc0_On0j9dNdF?Bo0d9hNfF:FQ1`9hNhF8HQ1^9iNgF8Ko0\\9mNeF6Om0Z9PO_F:7f0X9:gFGX9:hFFV9;kFES9=mFCQ9?oF^OS9c0mF[OS9g0lFWOU9k0kFSOU9o0kFnNW9R1jFlNV9V1jFhNW9Y1hFfNX9\\1hFbNX9`1hF^NY9c1gFZNZ9h1fFVN[9j1fFWNW9k1hFWNV9j1jFXNS9i1mFXNQ9i1oFXNP9h1PGWNP9j1oFVNR9i1oFVNQ9k1oFTNQ9m1oFRNR9n1nFQNR9P2nFnMT9R2kFnMU9S2kFlMV9S2kFlMU9U2kFjMU9W2kFiMU9W2jFjMU9W2kFiMQ1oNk6Y3THgMQ1WOd6S3[HfMo0_O^6l2dHeMm0FX6f2kHeMk0LQ6b2SIcMj00f5g2`IZMh02^5l2jISMf04]5j2mIRMd08\\5g2PJRMb0:[5d2SJSM`0=Z5a2VJRM?`0X5_2YJQM>c0V5l1XJgL6j0:g0U5c1iJbLJS18k0R5BeJLk0F^Oo0o4\\OQK@k05TOR1n4ZOVKYOS11jN^1l4YOXKTO\\1JbNk1h4XO]KnNd1DWNZ2f4TO`KmN[2Q2S2SObKlNZ2S2S2QOdKkNW2W2S2oNfKjNV2X2T2nNfKjNU2Y2T2oNgKgNV2Y2R2POiKhN\\2P2k1WOkKhNa2j1c1^OmKhNf3d0=c0nKiNj3>8i0PLiNl395l0PLkNo352n0PLmNT4OMR1QLnNY4IGW1QLQOd9n0\\FROd9m0^FROc9l0^FUOb9j0_FUOb9i0_FWOb9h0^FXOc9f0_FZOa9d0`F\\Oa9b0`F^Oa9a0`F_O`9?aFA`9<bFD`99bFF_98bFI^95cFKb9O`F0f9I[F8d;0001O01O000010O0001O00010O00000010O0O100O100L4M3O1O1O100O10klZ3"}, "label": "a boy pushing a bike"}]}
{"id": 15851, "image": "COCO_train2014_000000015851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy rolling the cycle by walk\"."}], "task": "refering", "answer_template": "The \"a boy rolling the cycle by walk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 174, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 265, 284, 285, 287, 288, 307, 310, 311], "bbox": [0.34954687500000003, 0.23256756756756758, 0.614875, 0.8700225225225224], "iscrowd": 0, "area": 13228.42265, "rle": {"size": [444, 640], "counts": "k_Q32h=2M3N2N2M3N2N2M3N2N3L2O1O1N2O1O1000O1N2N7I7IbT2CnkM1O2N1O2N2]OG_C<S;DaE7[O7P;D^E=COl:EZEe0JFg:HXEj01@b:HWEP17YO\\:J\\EP18UOX:MaEP16POX:3aEo08kNU:9cEn07gNU:=dEn07aNV:b0cEn08]NU94TGc0_On0j9dNdF?Bo0d9hNfF:FQ1`9hNhF8HQ1^9iNgF8Ko0\\9mNeF6Om0Z9PO_F:7f0X9:gFGX9:hFFV9;kFES9=mFCQ9?oF^OS9c0mF[OS9g0lFWOU9k0kFSOU9o0kFnNW9R1jFlNV9V1jFhNW9Y1hFfNX9\\1hFbNX9`1hF^NY9c1gFZNZ9h1fFVN[9j1fFWNW9k1hFWNV9j1jFXNS9i1mFXNQ9i1oFXNP9h1PGWNP9j1oFVNR9i1oFVNQ9k1oFTNQ9m1oFRNR9n1nFQNR9P2nFnMT9R2kFnMU9S2kFlMV9S2kFlMU9U2kFjMU9W2kFiMU9W2jFjMU9W2kFiMQ1oNk6Y3THgMQ1WOd6S3[HfMo0_O^6l2dHeMm0FX6f2kHeMk0LQ6b2SIcMj00f5g2`IZMh02^5l2jISMf04]5j2mIRMd08\\5g2PJRMb0:[5d2SJSM`0=Z5a2VJRM?`0X5_2YJQM>c0V5l1XJgL6j0:g0U5c1iJbLJS18k0R5BeJLk0F^Oo0o4\\OQK@k05TOR1n4ZOVKYOS11jN^1l4YOXKTO\\1JbNk1h4XO]KnNd1DWNZ2f4TO`KmN[2Q2S2SObKlNZ2S2S2QOdKkNW2W2S2oNfKjNV2X2T2nNfKjNU2Y2T2oNgKgNV2Y2R2POiKhN\\2P2k1WOkKhNa2j1c1^OmKhNf3d0=c0nKiNj3>8i0PLiNl395l0PLkNo352n0PLmNT4OMR1QLnNY4IGW1QLQOd9n0\\FROd9m0^FROc9l0^FUOb9j0_FUOb9i0_FWOb9h0^FXOc9f0_FZOa9d0`F\\Oa9b0`F^Oa9a0`F_O`9?aFA`9<bFD`99bFF_98bFI^95cFKb9O`F0f9I[F8d;0001O01O000010O0001O00010O00000010O0O100O100L4M3O1O1O100O10klZ3"}, "label": "a boy rolling the cycle by walk"}]}
{"id": 15851, "image": "COCO_train2014_000000015851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike\"."}], "task": "refering", "answer_template": "The \"bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 191, 192, 193, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 334, 335, 336, 337, 350, 351, 352], "bbox": [0.18009375, 0.4599324324324324, 0.7141093749999999, 0.9610585585585586], "iscrowd": 0, "area": 41277.96264999999, "rle": {"size": [444, 640], "counts": "nVb14^=?A?C=I7J5J4M2M4L4M2M4M3L3M4M3L3N3L4L2O2M3O0O2O1N2N101N2O1N101N2N101N2O1N1O2O1N100O100O1O10O0100O100O1O100O100000000O100000000O1000000O010000000O100000000O100000000O10O100000O02O001OfNcKVI[4j6fKVIZ4i6hKWIW4i6iKWIV4i6kKXIT4h6mKWIS4h6nKYIQ4h6oKWIP4j6PLWIo3i6RLVIn3j6RLWIm3j6RLWIl3j6ULUIk3k6ULVIj3k6VLTIi3m6WLTIh3l6YLSIg3n6XLSIg3m6ZLRIe3o6[LlHj3T7VLfHP4Z7QL_HU4a7kKYHZ4h7iKoG]4Q8fKfG`4Z8`00100000O1N2O1O1O1O1lJdGg4\\8ZKkG^4V8bKjG]4W8cKiG\\4X8dKhG[4Y8eKgGZ4Z8fKfGY4[8fKfGY4[8gKdGY4]8gKcGX4^8hKbGW4^8jKbGU4_8kKaGi3k8WLUGi3k8WLUGi3k8WLTGi3m8VLTGj3l8VLTGj3l8VLTGj3l8VLTGi3m8WLSGi3l8XLTGh3l8XLSGi3m8WLSGh3P9ULQGk3Q9SLoFm3`9O1000O1000000000001O0O1000001O00000O1hFSLa8m3WG[Li8\\40O100000O0100_OXGnKh8i2oFWN:oNh8i2PGUN:ROf8h2RGTN:SOd8h2TGSN9TOc8h2WGRN7VOb8g2XGRN7WOa8f2[GPN6YO_8e2_GoM3[O_8e2`GnM2]O^8d2cGlMV9S2lFkMT9T2PGiMQ9V2QGhMP9W2SGfMn8Y2[1N2O1O001N2O1O1N2O1O001N2O1O1O1N2O1O001N2O1O1NnN]NoEb1n9bNRF]1n9dNQF\\1P:dNPFZ1R:fNnEY1T:fNlEY1U:gNkEW1c0POc7IjGV14CP8WOlGU1OMR8nNnGT103o7iNQHS1O9m7dNTHQ1O`0j7_N]GN`0R18f0h7ZN_G4<k0<l0f7UNaG:8c0b0R1b7QNdG?3=f0X1`7lMfG=OCOl0k0]1_7fMgG=2e0h0]1\\7aMjG;4f0e0c1Z7\\MlG98g0a0i1X7dMYHa0?n1W7_M\\Hb0<R2W7YM`Hc09V2W7UMbHd06Z2W7PMeHe04]2V7kLiHk0L]2Z7fLlHS1BY2b7cLmHZ1YOV2j7_LmHR5R7nJnHR5R7oJlHS5S7mJmHT5R7lJnHT5R7lJnHU5Q7kJoHV5P7jJPIW5P7hJPI4MY4l8M000O100N3N1O1O1O1O000000000000000O1000000000000000000000000O100000000000000000001O0O10001O000000001O0000000O2M2O1N2O2M3M4M2M3N3L3N3L3N2M4L3N2M4M2M4M3L5L4K5K5L4K6I6H8H8H8HkP_2"}, "label": "bike"}]}
{"id": 15851, "image": "COCO_train2014_000000015851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle that the boy is pushing through the grass\"."}], "task": "refering", "answer_template": "The \"the bicycle that the boy is pushing through the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 191, 192, 193, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 334, 335, 336, 337, 350, 351, 352], "bbox": [0.18009375, 0.4599324324324324, 0.7141093749999999, 0.9610585585585586], "iscrowd": 0, "area": 41277.96264999999, "rle": {"size": [444, 640], "counts": "nVb14^=?A?C=I7J5J4M2M4L4M2M4M3L3M4M3L3N3L4L2O2M3O0O2O1N2N101N2O1N101N2N101N2O1N1O2O1N100O100O1O10O0100O100O1O100O100000000O100000000O1000000O010000000O100000000O100000000O10O100000O02O001OfNcKVI[4j6fKVIZ4i6hKWIW4i6iKWIV4i6kKXIT4h6mKWIS4h6nKYIQ4h6oKWIP4j6PLWIo3i6RLVIn3j6RLWIm3j6RLWIl3j6ULUIk3k6ULVIj3k6VLTIi3m6WLTIh3l6YLSIg3n6XLSIg3m6ZLRIe3o6[LlHj3T7VLfHP4Z7QL_HU4a7kKYHZ4h7iKoG]4Q8fKfG`4Z8`00100000O1N2O1O1O1O1lJdGg4\\8ZKkG^4V8bKjG]4W8cKiG\\4X8dKhG[4Y8eKgGZ4Z8fKfGY4[8fKfGY4[8gKdGY4]8gKcGX4^8hKbGW4^8jKbGU4_8kKaGi3k8WLUGi3k8WLUGi3k8WLTGi3m8VLTGj3l8VLTGj3l8VLTGj3l8VLTGi3m8WLSGi3l8XLTGh3l8XLSGi3m8WLSGh3P9ULQGk3Q9SLoFm3`9O1000O1000000000001O0O1000001O00000O1hFSLa8m3WG[Li8\\40O100000O0100_OXGnKh8i2oFWN:oNh8i2PGUN:ROf8h2RGTN:SOd8h2TGSN9TOc8h2WGRN7VOb8g2XGRN7WOa8f2[GPN6YO_8e2_GoM3[O_8e2`GnM2]O^8d2cGlMV9S2lFkMT9T2PGiMQ9V2QGhMP9W2SGfMn8Y2[1N2O1O001N2O1O1N2O1O001N2O1O1O1N2O1O001N2O1O1NnN]NoEb1n9bNRF]1n9dNQF\\1P:dNPFZ1R:fNnEY1T:fNlEY1U:gNkEW1c0POc7IjGV14CP8WOlGU1OMR8nNnGT103o7iNQHS1O9m7dNTHQ1O`0j7_N]GN`0R18f0h7ZN_G4<k0<l0f7UNaG:8c0b0R1b7QNdG?3=f0X1`7lMfG=OCOl0k0]1_7fMgG=2e0h0]1\\7aMjG;4f0e0c1Z7\\MlG98g0a0i1X7dMYHa0?n1W7_M\\Hb0<R2W7YM`Hc09V2W7UMbHd06Z2W7PMeHe04]2V7kLiHk0L]2Z7fLlHS1BY2b7cLmHZ1YOV2j7_LmHR5R7nJnHR5R7oJlHS5S7mJmHT5R7lJnHT5R7lJnHU5Q7kJoHV5P7jJPIW5P7hJPI4MY4l8M000O100N3N1O1O1O1O000000000000000O1000000000000000000000000O100000000000000000001O0O10001O000000001O0000000O2M2O1N2O2M3M4M2M3N3L3N3L3N2M4L3N2M4M2M4M3L5L4K5K5L4K6I6H8H8H8HkP_2"}, "label": "the bicycle that the boy is pushing through the grass"}]}
{"id": 302582, "image": "COCO_train2014_000000302582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snow boarder\"."}], "task": "refering", "answer_template": "The \"a snow boarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 147, 148, 149, 150, 172, 173, 195, 196, 218, 219, 220], "bbox": [0.411609375, 0.3288758782201405, 0.573421875, 0.65096018735363], "iscrowd": 0, "area": 6532.105750000001, "rle": {"size": [427, 640], "counts": "lW^38o<6L4M3L3N2O1O00000O11O01O000000000M4O0O100O1O100O1001N100O00010007J0O001O001M2O2M2O2N101ZD`Na0OX:h1cEbNZ:\\1aEPOX:n0gEYO8UOW9a1_F^O8ROZ9k2cFWM\\9k2`FXM_9i2_FYMa9h2\\FZMc9X3O01O000O2O001NOO2O3N2N2N3N0O101N1M40O1O10O0100O1VFmLZ9Q2\\FmN5WO_9j1bFkNM]O`9g1fFkNGBa9a1kFlNCDb9_1lF]OT9b0nF]OR9a0QG^Oo8`0SGAm8oN^F=1Nh0d0j8oNeF6L4h0f0h8nNeF8M1h0h0g8nNeF8NOh0j0f8mNfF8ONg0m0R9UOYFIi0Q1P9TObGk0a:N1N2O1N2L4L4M3M3N2Moba3"}, "label": "a snow boarder"}]}
{"id": 302582, "image": "COCO_train2014_000000302582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black jacket snowboarding\"."}], "task": "refering", "answer_template": "The \"man in black jacket snowboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 147, 148, 149, 150, 172, 173, 195, 196, 218, 219, 220], "bbox": [0.411609375, 0.3288758782201405, 0.573421875, 0.65096018735363], "iscrowd": 0, "area": 6532.105750000001, "rle": {"size": [427, 640], "counts": "lW^38o<6L4M3L3N2O1O00000O11O01O000000000M4O0O100O1O100O1001N100O00010007J0O001O001M2O2M2O2N101ZD`Na0OX:h1cEbNZ:\\1aEPOX:n0gEYO8UOW9a1_F^O8ROZ9k2cFWM\\9k2`FXM_9i2_FYMa9h2\\FZMc9X3O01O000O2O001NOO2O3N2N2N3N0O101N1M40O1O10O0100O1VFmLZ9Q2\\FmN5WO_9j1bFkNM]O`9g1fFkNGBa9a1kFlNCDb9_1lF]OT9b0nF]OR9a0QG^Oo8`0SGAm8oN^F=1Nh0d0j8oNeF6L4h0f0h8nNeF8M1h0h0g8nNeF8NOh0j0f8mNfF8ONg0m0R9UOYFIi0Q1P9TObGk0a:N1N2O1N2L4L4M3M3N2Moba3"}, "label": "man in black jacket snowboarding"}]}
{"id": 105976, "image": "COCO_train2014_000000105976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a full glass of water\"."}], "task": "refering", "answer_template": "The \"a full glass of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 40, 41, 42, 43, 63, 64, 65, 66, 86, 87, 88, 89, 109, 110, 111, 132, 133, 155, 156, 157], "bbox": [0.73828125, 0.040375, 0.890625, 0.3763125], "iscrowd": 0, "area": 10818.75, "rle": {"size": [480, 640], "counts": "Vjm66^1:Z;LdD7V;OgD4R;3kDOo:7oDIm:>PECl:c0QE]Ok:i0SEWOi:P1TEPOi:U1UEkNg:[1WEeNe:a1YE`Nb:f1\\EZNa:k1\\EVN`:P2ZETNb:R2WESNe:V3N3M2O1O1N2O1O1N2O1N200O100000O10O100000000O100000000O0100TOoEiLQ:U3RFjLm9U3VFjLj9T3YFkLg9S3\\FlLd9S3^FlLb9T3_FkL`9U3bFjL^9U3dFjL\\9U3fFjLZ9V3gFhLZ9W3hFhLW9X3kFfLV9Y3lFeLU9[3lFcLU9\\3mFbLS9^3oF`LS9^3nFaLS9_3mF`LU9^3kFbLV9]3W1OdEdLW9[3W1O2N1O2O0O2N1O2O0O2N1O1O2M2N3N1N3M2O2M2O2K4H9G8H8H9G8H9G8HUiP1"}, "label": "a full glass of water"}]}
{"id": 105976, "image": "COCO_train2014_000000105976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass with ice water and freezing outside\"."}], "task": "refering", "answer_template": "The \"a glass with ice water and freezing outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 40, 41, 42, 43, 63, 64, 65, 66, 86, 87, 88, 89, 109, 110, 111, 132, 133, 155, 156, 157], "bbox": [0.73828125, 0.040375, 0.890625, 0.3763125], "iscrowd": 0, "area": 10818.75, "rle": {"size": [480, 640], "counts": "Vjm66^1:Z;LdD7V;OgD4R;3kDOo:7oDIm:>PECl:c0QE]Ok:i0SEWOi:P1TEPOi:U1UEkNg:[1WEeNe:a1YE`Nb:f1\\EZNa:k1\\EVN`:P2ZETNb:R2WESNe:V3N3M2O1O1N2O1O1N2O1N200O100000O10O100000000O100000000O0100TOoEiLQ:U3RFjLm9U3VFjLj9T3YFkLg9S3\\FlLd9S3^FlLb9T3_FkL`9U3bFjL^9U3dFjL\\9U3fFjLZ9V3gFhLZ9W3hFhLW9X3kFfLV9Y3lFeLU9[3lFcLU9\\3mFbLS9^3oF`LS9^3nFaLS9_3mF`LU9^3kFbLV9]3W1OdEdLW9[3W1O2N1O2O0O2N1O2O0O2N1O1O2M2N3N1N3M2O2M2O2K4H9G8H8H9G8H9G8HUiP1"}, "label": "a glass with ice water and freezing outside"}]}
{"id": 572923, "image": "COCO_train2014_000000572923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall artificial tree\"."}], "task": "refering", "answer_template": "The \"a tall artificial tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 187, 210, 233, 256], "bbox": [0.001515625, 0.02427570093457944, 0.341171875, 0.7669392523364486], "iscrowd": 0, "area": 31869.051700000007, "rle": {"size": [428, 640], "counts": "c`03W=7J7I6J7I6I8I6J7I3M2N2M4M2N2N1O2M2O2N1O2N2L3J7ROm0O2N1O1O1O1O1O100O1O1O1O10000O:G;E2hHUKk5m4oIWKQ6j4kIYKU6h4fI\\KZ6e4bI^K^6d4]I_Kb6f500000000000001O000000000O01O1O001O1O001O001O1O001O1O001O001003L5L3M3M3M3M3M3M4L3M3Mc1]Na0_O9G3L1000O001N2O0O2O1O0O2O1O0VMk2O1N101N2O0O2O1O1000000000000001O001N2O1O001O1O001O1O1O001OO1000O02O001N2O001N101O1O0O2O1iNV1O2M4L4M3L5L3L3M2N3K5A>01N2O1N101N2O001O1O1O0O2O1O1O0PObFRN_9b1\\1J5J7O1O10O0100O1O010O10O01O100O010O10O01O100O010DlC^OU<=RDAn;:YDCg;9i0N2O001O1O1O001N2O00PZ`5"}, "label": "a tall artificial tree"}]}
{"id": 556544, "image": "COCO_train2014_000000556544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in the black coat with the scarf\"."}], "task": "refering", "answer_template": "The \"man in the black coat with the scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.663265625, 0.07710416666666667, 0.98809375, 0.9863958333333332], "iscrowd": 0, "area": 61722.984500000006, "rle": {"size": [480, 640], "counts": "cVW61k>4K5K5K5K5K5K5K5L4K5K5bK^NYKg1f4\\NSKi1l4ZNmJk1R5XNgJm1X5WNaJn1]5UN\\JP2c5SNVJR2i5QNPJT2o5PNiIU2V6nMcIW2\\6lM]IY2b6jMWI[2h6iMQI[2n6lMgHY2X7nM]HW2b7QNRHT2m7SNhGR2W8\\2O1O1O1O1O1O1O1O1M3L4M3L4M3L4M3L4M3L5K4M3L4M3L4M3L4M3L4N2O1O1gMkFgNU9X1RGaNo8^1XG[Ni8d1]GVNd8i1cGPN^8n1jGkMW8T2PHeMQ8Z2VH`Mj7_2ZH]Mg7b2ZH]Mg7b2ZH]Mg7a2[H^Mf7a2[H^Mf7a2[H^Mf7a2[H_Me7\\2`HcMa7X2dHgM]7S2iHlMX7o1mHPNU7i1QIVNP7e1UIZNl6c1WI]Ni6c1WI\\Nj6c1WI\\Nj6d1VI[Nk6e1UIZNl6e1UIZNl6f1W300O10000O10000O10000O10000O10000O10000O1N201O000000000O100000000000000O100000000000000O100000000000000O101O000002N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1N2fEjMb7]2lGVNm7P2aGdNV8d1WGPOa8V1mF^Ok8`3H9H8G9H7I8G9H8G8I8G9H8H7H9H8G9H7I8G9H8G8I8H8G9H7H9H8G9H7I8G9H8G8I8HX`3"}, "label": "man in the black coat with the scarf"}]}
{"id": 556544, "image": "COCO_train2014_000000556544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding wine glass wearing black coat and scarf\"."}], "task": "refering", "answer_template": "The \"man holding wine glass wearing black coat and scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.663265625, 0.07710416666666667, 0.98809375, 0.9863958333333332], "iscrowd": 0, "area": 61722.984500000006, "rle": {"size": [480, 640], "counts": "cVW61k>4K5K5K5K5K5K5K5L4K5K5bK^NYKg1f4\\NSKi1l4ZNmJk1R5XNgJm1X5WNaJn1]5UN\\JP2c5SNVJR2i5QNPJT2o5PNiIU2V6nMcIW2\\6lM]IY2b6jMWI[2h6iMQI[2n6lMgHY2X7nM]HW2b7QNRHT2m7SNhGR2W8\\2O1O1O1O1O1O1O1O1M3L4M3L4M3L4M3L4M3L5K4M3L4M3L4M3L4M3L4N2O1O1gMkFgNU9X1RGaNo8^1XG[Ni8d1]GVNd8i1cGPN^8n1jGkMW8T2PHeMQ8Z2VH`Mj7_2ZH]Mg7b2ZH]Mg7b2ZH]Mg7a2[H^Mf7a2[H^Mf7a2[H^Mf7a2[H_Me7\\2`HcMa7X2dHgM]7S2iHlMX7o1mHPNU7i1QIVNP7e1UIZNl6c1WI]Ni6c1WI\\Nj6c1WI\\Nj6d1VI[Nk6e1UIZNl6e1UIZNl6f1W300O10000O10000O10000O10000O10000O10000O1N201O000000000O100000000000000O100000000000000O100000000000000O101O000002N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1N2fEjMb7]2lGVNm7P2aGdNV8d1WGPOa8V1mF^Ok8`3H9H8G9H7I8G9H8G8I8G9H8H7H9H8G9H7I8G9H8G8I8H8G9H7H9H8G9H7I8G9H8G8I8HX`3"}, "label": "man holding wine glass wearing black coat and scarf"}]}
{"id": 556544, "image": "COCO_train2014_000000556544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mostly bald man in a beige coat tasting wine\"."}], "task": "refering", "answer_template": "The \"a mostly bald man in a beige coat tasting wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 377, 378, 379, 380], "bbox": [0.3581875, 0.12929166666666667, 0.676015625, 0.9902916666666666], "iscrowd": 0, "area": 53101.018000000004, "rle": {"size": [480, 640], "counts": "jb[31o>1O1O1N2O1O2lEOU52cJ5\\5M\\J;c5FTJc0k5^OmIj0R6WOfIQ1Y6PO_IW1`6QOQIX1m6XO]HP1^7VOYHR1c7ROTHW1g7nNQHY1k7mNlG[1o7jNhG_1T8fNfG_1?lLX6j1SI`1O[Mj6Z1RIl2j6g2M3L4L4M3N2N2M3N2N3M2M3N2N2M3N2N2N2M3N2N2N2N2O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1OJ6J6I7J6J6NVOjFoKX9l3QGnKP9n3YGlKh8P4aGiKa8S4X1K5L5K=C<D<G:G2N000O101O000O1000000O1000000O10001N1000000O10000001O001O001O010O001O001O001O001O001O0N3L3L5L7H=D=oD[LP:]4K5L4L4K5L4N2N2O2M4L4L4M3L4L101N1O2O0O2N101N1O2O0O2N1O2O0J7_O`0_Ob0_O`0_Ob0^Oa0@a0^Oa0K6N1O2N1O2O0O2N1O2N1O2O0O2N3M4L4L4M3L4L4L4L4L4M3L4L4L5K4M3L4L4L4L4L4M3L4L4K5G9GiiP3"}, "label": "a mostly bald man in a beige coat tasting wine"}]}
{"id": 556544, "image": "COCO_train2014_000000556544.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses , a brown jacket , black shirt , and khaki pants\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses , a brown jacket , black shirt , and khaki pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 354, 355, 356, 357, 358, 377, 378, 379, 380], "bbox": [0.3581875, 0.12929166666666667, 0.676015625, 0.9902916666666666], "iscrowd": 0, "area": 53101.018000000004, "rle": {"size": [480, 640], "counts": "jb[31o>1O1O1N2O1O2lEOU52cJ5\\5M\\J;c5FTJc0k5^OmIj0R6WOfIQ1Y6PO_IW1`6QOQIX1m6XO]HP1^7VOYHR1c7ROTHW1g7nNQHY1k7mNlG[1o7jNhG_1T8fNfG_1?lLX6j1SI`1O[Mj6Z1RIl2j6g2M3L4L4M3N2N2M3N2N3M2M3N2N2M3N2N2N2M3N2N2N2N2O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1OJ6J6I7J6J6NVOjFoKX9l3QGnKP9n3YGlKh8P4aGiKa8S4X1K5L5K=C<D<G:G2N000O101O000O1000000O1000000O10001N1000000O10000001O001O001O010O001O001O001O001O001O0N3L3L5L7H=D=oD[LP:]4K5L4L4K5L4N2N2O2M4L4L4M3L4L101N1O2O0O2N101N1O2O0O2N1O2O0J7_O`0_Ob0_O`0_Ob0^Oa0@a0^Oa0K6N1O2N1O2O0O2N1O2N1O2O0O2N3M4L4L4M3L4L4L4L4L4M3L4L4L5K4M3L4L4L4L4L4M3L4L4K5G9GiiP3"}, "label": "a man wearing glasses , a brown jacket , black shirt , and khaki pants"}]}
{"id": 335362, "image": "COCO_train2014_000000335362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of sandwich in front of guiness\"."}], "task": "refering", "answer_template": "The \"half of sandwich in front of guiness\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 361, 362, 363], "bbox": [0.48539062499999996, 0.37527083333333333, 0.9202187500000001, 0.9168541666666666], "iscrowd": 0, "area": 44058.3403, "rle": {"size": [480, 640], "counts": "[Qb42n>1O2N1N3N1O2N1N3N2N1TD_OW9c0bFNQ98eFOY96]FOc96SF0l96iEOV:^201O000O2O00001N1000001O0O101O000O2O00001N10001O0O10001O0O101O0O101O0O101O0O101O0O10001N10001N10001N10001N10001N10001N10000O2O000O2O000O2O000O2O000O2O00000O2O000O2O000O2O000O2O000O2O000O2O000O101O0O101O0O101O0O101O0O101O0O10001N10001N10001N10001N10001N1000001N10001N10001N10001N10001N10001N10000O2O000O2O000O2O000O2O000O2O000O101O0O101O0O101O0O101O0O101O0O10001N10O100O00100O010O1O010O10O0100O00100O0O2M3N1N3N2N1N3N2M2O2M3N1O2M3N1N3N2M2O2N2M2O2M3N1\\NfFaM]9[2lF^MV9_2SGYMn8d2[GTMh8h2aGQMa8l2hGkLZ8R3oGgLS8U3n1M2M4L4M3L3M4M3L3M4L4M2M4L4M2M4L4M2M4L4M2M4L4M2M4L4M2M4L4M2M`gg0"}, "label": "half of sandwich in front of guiness"}]}
{"id": 335362, "image": "COCO_train2014_000000335362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich\"."}], "task": "refering", "answer_template": "The \"sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 361, 362, 363], "bbox": [0.48539062499999996, 0.37527083333333333, 0.9202187500000001, 0.9168541666666666], "iscrowd": 0, "area": 44058.3403, "rle": {"size": [480, 640], "counts": "[Qb42n>1O2N1N3N1O2N1N3N2N1TD_OW9c0bFNQ98eFOY96]FOc96SF0l96iEOV:^201O000O2O00001N1000001O0O101O000O2O00001N10001O0O10001O0O101O0O101O0O101O0O101O0O10001N10001N10001N10001N10001N10001N10000O2O000O2O000O2O000O2O000O2O00000O2O000O2O000O2O000O2O000O2O000O2O000O101O0O101O0O101O0O101O0O101O0O10001N10001N10001N10001N10001N1000001N10001N10001N10001N10001N10001N10000O2O000O2O000O2O000O2O000O2O000O101O0O101O0O101O0O101O0O101O0O10001N10O100O00100O010O1O010O10O0100O00100O0O2M3N1N3N2N1N3N2M2O2M3N1O2M3N1N3N2M2O2N2M2O2M3N1\\NfFaM]9[2lF^MV9_2SGYMn8d2[GTMh8h2aGQMa8l2hGkLZ8R3oGgLS8U3n1M2M4L4M3L3M4M3L3M4L4M2M4L4M2M4L4M2M4L4M2M4L4M2M4L4M2M4L4M2M`gg0"}, "label": "sandwich"}]}
{"id": 335362, "image": "COCO_train2014_000000335362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwich behind a pint of guinness\"."}], "task": "refering", "answer_template": "The \"half of a sandwich behind a pint of guinness\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 199], "bbox": [0.64546875, 0.24622916666666667, 0.8831875, 0.48277083333333337], "iscrowd": 0, "area": 10205.925150000001, "rle": {"size": [480, 640], "counts": "WjQ61k>4L4L3TMNhF6W9;VFHj9g200O10000O10O1000O1000O010000O01000O0100000O010000O2O000O1000000O10001N10000O1000000O101O0O101O0O101O000O2O000O101O0O101O000O2O000O2O1O1O0O2O1O1N101O1N2O1O001N2O1O0O2O1O1N101O1O1N2O001N2O1O0O2O1O1O0O2O1O1N2O001N2O1O001N2O1O0O2O1O1O1N101O1N2O001N2O1O001N2O1O0O2O1O1N2O001O1N2O001N2O1O0O2O1O1OSoR1"}, "label": "half of a sandwich behind a pint of guinness"}]}
{"id": 335362, "image": "COCO_train2014_000000335362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far away piece of sandwich\"."}], "task": "refering", "answer_template": "The \"the far away piece of sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 199], "bbox": [0.64546875, 0.24622916666666667, 0.8831875, 0.48277083333333337], "iscrowd": 0, "area": 10205.925150000001, "rle": {"size": [480, 640], "counts": "WjQ61k>4L4L3TMNhF6W9;VFHj9g200O10000O10O1000O1000O010000O01000O0100000O010000O2O000O1000000O10001N10000O1000000O101O0O101O0O101O000O2O000O101O0O101O000O2O000O2O1O1O0O2O1O1N101O1N2O1O001N2O1O0O2O1O1N101O1O1N2O001N2O1O0O2O1O1O0O2O1O1N2O001N2O1O001N2O1O0O2O1O1O1N101O1N2O001N2O1O001N2O1O0O2O1O1N2O001O1N2O001N2O1O0O2O1O1OSoR1"}, "label": "the far away piece of sandwich"}]}
{"id": 499204, "image": "COCO_train2014_000000499204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hot dog in the front with the most ketchup\"."}], "task": "refering", "answer_template": "The \"the hot dog in the front with the most ketchup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.178234375, 0.6023541666666666, 0.57, 0.7670625], "iscrowd": 0, "area": 15168.12985, "rle": {"size": [480, 640], "counts": "ehe16f>:F4L5K4L5K4N3N1N3N1O2M3N1N20000O1000000O1000000O1000000O101O0O1000000O101O000O1000001N1000000O10001N10000O10001O0O10000O10001N1000000000001O00000000001O00000000000O2O0000000000000000000000000000000000000O10000000O1000001O00000000001O000000001O000000001O00000000001O0000000O2O000000001O0000000000001O0000000000000O2O00000000000000001N100000000000001O00000O100000001O000000000O10001O00001O001O00001N1010O01O00001O001O001O001O00010O001O001O00001O001O001O010N101O6J6I7J2N1O1N2O1O1O1N2O1O1O1N2O1N2O1N2O1NnaP4"}, "label": "the hot dog in the front with the most ketchup"}]}
{"id": 499204, "image": "COCO_train2014_000000499204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hot dog with ketchup at the edge of the counter\"."}], "task": "refering", "answer_template": "The \"hot dog with ketchup at the edge of the counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.178234375, 0.6023541666666666, 0.57, 0.7670625], "iscrowd": 0, "area": 15168.12985, "rle": {"size": [480, 640], "counts": "ehe16f>:F4L5K4L5K4N3N1N3N1O2M3N1N20000O1000000O1000000O1000000O101O0O1000000O101O000O1000001N1000000O10001N10000O10001O0O10000O10001N1000000000001O00000000001O00000000000O2O0000000000000000000000000000000000000O10000000O1000001O00000000001O000000001O000000001O00000000001O0000000O2O000000001O0000000000001O0000000000000O2O00000000000000001N100000000000001O00000O100000001O000000000O10001O00001O001O00001N1010O01O00001O001O001O001O00010O001O001O00001O001O001O010N101O6J6I7J2N1O1N2O1O1O1N2O1O1O1N2O1N2O1N2O1NnaP4"}, "label": "hot dog with ketchup at the edge of the counter"}]}
{"id": 499204, "image": "COCO_train2014_000000499204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hot dog furthest away\"."}], "task": "refering", "answer_template": "The \"the hot dog furthest away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 265, 266], "bbox": [0.222359375, 0.5148958333333333, 0.57353125, 0.6866666666666668], "iscrowd": 0, "area": 8921.33545, "rle": {"size": [480, 640], "counts": "UkR21n>5K5K0O2O1O1O00100O001O1M3M2M4M3M2N3M3O10O01000O10O1000O010000O010000O0100000O10O1000000O100000000O1000000O100000000O1000000O10000000000000000001O000000000000000000000000000000000000000000000001O00000001O01O0000000000001O00000000001O00000000001O0000000000010O00000000010O000000000010O0000000001O01O00000000001O01O0000000001O0000001O001O010O010O010O010O0100O010O010O010O010O010O0100O001O001O001O0100O100<D00000000000O101N100O10000O100O2O0O1M3M5K5K5KUeo3"}, "label": "the hot dog furthest away"}]}
{"id": 499204, "image": "COCO_train2014_000000499204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog sitting in the back of a carton behind another hot dog\"."}], "task": "refering", "answer_template": "The \"a hot dog sitting in the back of a carton behind another hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 265, 266], "bbox": [0.222359375, 0.5148958333333333, 0.57353125, 0.6866666666666668], "iscrowd": 0, "area": 8921.33545, "rle": {"size": [480, 640], "counts": "UkR21n>5K5K0O2O1O1O00100O001O1M3M2M4M3M2N3M3O10O01000O10O1000O010000O010000O0100000O10O1000000O100000000O1000000O100000000O1000000O10000000000000000001O000000000000000000000000000000000000000000000001O00000001O01O0000000000001O00000000001O00000000001O0000000000010O00000000010O000000000010O0000000001O01O00000000001O01O0000000001O0000001O001O010O010O010O010O0100O010O010O010O010O010O0100O001O001O001O0100O100<D00000000000O101N100O10000O100O2O0O1M3M5K5K5KUeo3"}, "label": "a hot dog sitting in the back of a carton behind another hot dog"}]}
{"id": 491018, "image": "COCO_train2014_000000491018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the basket which contains papaya fruits\"."}], "task": "refering", "answer_template": "The \"the basket which contains papaya fruits\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 265, 266, 272, 273, 274, 275, 276, 277, 279, 280, 281, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 319, 320, 321, 322, 323, 324, 325], "bbox": [0.16864801864801862, 0.7652187500000001, 0.7642657342657343, 0.9350156249999999], "iscrowd": 0, "area": 18073.038799999995, "rle": {"size": [640, 429], "counts": "]_]11Pd04K5K5L4K5L4K5K5L4K5L4K5K5L4K2O0O00100O10O01O100O00100O10O01O100O010O1O100O100O1O100O100O1O1O1O100O1O1O000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O000000000010O000000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O000000000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O01OPOiMY@U2i?nMT@P2n?SNo_Ol1R`0WNj_Oh1X`0[Ne_Oc1]`0`N`_O_1a`0dN\\_OZ1f`0iNW_OU1k`0nNQ_OR1Pa0ROl^Ol0Va0m03N1N2N2O1N2N2N2O1N2N2O1N2N2O1N6J9H8G9G9H8G:FXcn1"}, "label": "the basket which contains papaya fruits"}]}
{"id": 491018, "image": "COCO_train2014_000000491018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fruit basket from the front\"."}], "task": "refering", "answer_template": "The \"the fruit basket from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 265, 266, 272, 273, 274, 275, 276, 277, 279, 280, 281, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 319, 320, 321, 322, 323, 324, 325], "bbox": [0.16864801864801862, 0.7652187500000001, 0.7642657342657343, 0.9350156249999999], "iscrowd": 0, "area": 18073.038799999995, "rle": {"size": [640, 429], "counts": "]_]11Pd04K5K5L4K5L4K5K5L4K5L4K5K5L4K2O0O00100O10O01O100O00100O10O01O100O010O1O100O100O1O100O100O1O1O1O100O1O1O000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O000000000010O000000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O000000000000O10000000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O01OPOiMY@U2i?nMT@P2n?SNo_Ol1R`0WNj_Oh1X`0[Ne_Oc1]`0`N`_O_1a`0dN\\_OZ1f`0iNW_OU1k`0nNQ_OR1Pa0ROl^Ol0Va0m03N1N2N2O1N2N2N2O1N2N2O1N2N2O1N6J9H8G9G9H8G:FXcn1"}, "label": "the fruit basket from the front"}]}
{"id": 24076, "image": "COCO_train2014_000000024076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a painting of scissors that is sitting on a chair\"."}], "task": "refering", "answer_template": "The \"a painting of scissors that is sitting on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 210, 211, 232, 233, 234, 235, 255, 256, 257, 258, 259, 280, 281, 282, 283, 303, 304, 305, 306, 307, 327], "bbox": [0.105375, 0.5870023419203747, 0.35920312499999996, 0.9688524590163935], "iscrowd": 0, "area": 7160.992200000005, "rle": {"size": [427, 640], "counts": "mdl02Y=2N2N2O1N2N100O010O1O010O1O010O10O01O010O010O00100O001QN<VFCj9>UFCj9>VFAj9`0UFAj9`0UFAi9a0WF^Oi9c0VF^Oi9c0WF\\Oi9e0VF\\Oi9e0VF[Oj9g0UFYOj9f0\\EiNi0b0j9e0XF[Oh9e0YF[Of9e0ZF[Of9e0[F[Od9e0\\F[Od9e0\\F\\Oc9d0^F\\Oa9d0_F\\Oa9d0`F\\O_9d0aF\\O_9d0aF]O_9c0aF\\O_9d0aF]O^9h0^FXOa9n0YFROh9o0VFROi9P1VFoNj9R1UFoNk9R1SFnNm9S1SFmNl9U1RFlNn9U1QFjNo9W1PFjNo9X1oEhNR:Y1mEgNR:Z1mEfNV:Y1iEgNb:o0\\EQOn:f0QE[OU;>jD1n:IRER1Z:gNeEm1k9PNTFQ2n9d03N2N1O2N2N2N1N2O1O1O1O2N1O1J6K5J6J6K5J6N101N2N1O2O1N1O2N2O0O2O1O001O1O010O1O00100O010000O10O0100O100O01000O100O010O10N101O001N101O1N101N1O2O0O2N3N1N3M2O2M2N100O1O010O0O2O1N1OjP[5"}, "label": "a painting of scissors that is sitting on a chair"}]}
{"id": 196111, "image": "COCO_train2014_000000196111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle vase\"."}], "task": "refering", "answer_template": "The \"middle vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 333, 334, 355, 356, 357, 378, 379, 380, 381, 400, 401, 402, 403, 404, 423, 424, 425, 426, 427, 447, 448, 449, 471, 472, 494, 495, 517, 518], "bbox": [0.4260501567398119, 0.382015625, 0.5815987460815047, 0.986515625], "iscrowd": 0, "area": 24762.76755000001, "rle": {"size": [640, 638], "counts": "T`Z58]c0;D=D;H8L5L3M4L3M3oIPNnIU2o5SNhIP2V6XNaIl1]6[NZIh1d6gNlH\\1S7DjGa0S8`0kFCS9^1kEfNS:Z2kDiMT;b501O0O2O1O1O1N2O1O001N2O1O1N2O1O1O2M3N2N3L3nG[Fe5g9UJ]Fk5f9mI_FS6c9gIaFX6b9bIbF^6b9YIcFh6`9PIdFP7`9hHdFX7_9`HfF`7^9XHfFh7i:00000000000000001O00000M3M3M3L4M3M3L4M3M3L5L3L4L4L4L4L4K5L4L4L4L:]GkEj6k:YHcFa6b<jNV1jNV1jN;D9H8H8H8H8H8G9H8H8H8HR]V5"}, "label": "middle vase"}]}
{"id": 196111, "image": "COCO_train2014_000000196111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it ' s a white vase which is between two other vases\"."}], "task": "refering", "answer_template": "The \"it ' s a white vase which is between two other vases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 309, 310, 311, 332, 333, 334, 355, 356, 357, 378, 379, 380, 381, 400, 401, 402, 403, 404, 423, 424, 425, 426, 427, 447, 448, 449, 471, 472, 494, 495, 517, 518], "bbox": [0.4260501567398119, 0.382015625, 0.5815987460815047, 0.986515625], "iscrowd": 0, "area": 24762.76755000001, "rle": {"size": [640, 638], "counts": "T`Z58]c0;D=D;H8L5L3M4L3M3oIPNnIU2o5SNhIP2V6XNaIl1]6[NZIh1d6gNlH\\1S7DjGa0S8`0kFCS9^1kEfNS:Z2kDiMT;b501O0O2O1O1O1N2O1O001N2O1O1N2O1O1O2M3N2N3L3nG[Fe5g9UJ]Fk5f9mI_FS6c9gIaFX6b9bIbF^6b9YIcFh6`9PIdFP7`9hHdFX7_9`HfF`7^9XHfFh7i:00000000000000001O00000M3M3M3L4M3M3L4M3M3L5L3L4L4L4L4L4K5L4L4L4L:]GkEj6k:YHcFa6b<jNV1jNV1jN;D9H8H8H8H8H8G9H8H8H8HR]V5"}, "label": "it ' s a white vase which is between two other vases"}]}
{"id": 196111, "image": "COCO_train2014_000000196111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a trianguar vase that is white with lines in the pattern\"."}], "task": "refering", "answer_template": "The \"a trianguar vase that is white with lines in the pattern\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 289, 290, 291, 292, 312, 313, 314, 335, 336, 337, 358, 359, 360, 381, 382, 383, 404, 405, 406, 407, 427, 428, 429, 430, 449, 450, 451, 452, 453, 472, 473, 474, 475, 476, 495, 496, 497, 498, 499, 518, 519, 520, 521, 522, 523], "bbox": [0.5320062695924765, 0.5071718749999999, 0.7574294670846395, 0.9858281249999999], "iscrowd": 0, "area": 24711.40344999999, "rle": {"size": [640, 638], "counts": "bSe62lc03L3K6J5K6I6K5K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5]J\\LeJi3[5WL`Jn3_5SL[JS4e5VLmIo3S6SLeIS4Z6PL^IV4b6lKWIY4h6iKQI]4o6dKkHa4U7`KdHf4[7\\K_Hi4a7WKYHn4g7TKRHR5n7oJlGV5S8kJgG[5Y8fJaG_5^8cJ[Gc5d8_JTGh5k8YJoFm5P9b200000000000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0kG^Fl5b9PJiFi5W9SJSGg5n8UJYGg5g8WJ^Ge5c8ZJbGb5^8]JgG_5Z8`JiG]5W8bJnGZ5S8dJQHY5o7gJUHU5l7iJYHT5f7kJ^HR5c7mJaHo4`7oJeHm4[7RKiHk4X7SKmHi4T7VKPIf4Q7VK_J]3a5^LjJ\\3R;K5K5K5K6J5K2O2M3M3M3M3M2N3M3M3M`0@2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2M3M^XP3"}, "label": "a trianguar vase that is white with lines in the pattern"}]}
{"id": 196111, "image": "COCO_train2014_000000196111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shortest white vase\"."}], "task": "refering", "answer_template": "The \"shortest white vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 289, 290, 291, 292, 312, 313, 314, 335, 336, 337, 358, 359, 360, 381, 382, 383, 404, 405, 406, 407, 427, 428, 429, 430, 449, 450, 451, 452, 453, 472, 473, 474, 475, 476, 495, 496, 497, 498, 499, 518, 519, 520, 521, 522, 523], "bbox": [0.5320062695924765, 0.5071718749999999, 0.7574294670846395, 0.9858281249999999], "iscrowd": 0, "area": 24711.40344999999, "rle": {"size": [640, 638], "counts": "bSe62lc03L3K6J5K6I6K5K5K5K5K5K5K5K5K5K5J6K5K5K5K5K5K5]J\\LeJi3[5WL`Jn3_5SL[JS4e5VLmIo3S6SLeIS4Z6PL^IV4b6lKWIY4h6iKQI]4o6dKkHa4U7`KdHf4[7\\K_Hi4a7WKYHn4g7TKRHR5n7oJlGV5S8kJgG[5Y8fJaG_5^8cJ[Gc5d8_JTGh5k8YJoFm5P9b200000000000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0kG^Fl5b9PJiFi5W9SJSGg5n8UJYGg5g8WJ^Ge5c8ZJbGb5^8]JgG_5Z8`JiG]5W8bJnGZ5S8dJQHY5o7gJUHU5l7iJYHT5f7kJ^HR5c7mJaHo4`7oJeHm4[7RKiHk4X7SKmHi4T7VKPIf4Q7VK_J]3a5^LjJ\\3R;K5K5K5K6J5K2O2M3M3M3M3M2N3M3M3M`0@2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2M3M^XP3"}, "label": "shortest white vase"}]}
{"id": 196111, "image": "COCO_train2014_000000196111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beautiful flower vase on the table\"."}], "task": "refering", "answer_template": "The \"a beautiful flower vase on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 239, 260, 261, 262, 283, 284, 285, 306, 307, 308, 329, 330, 331, 352, 353, 354, 374, 375, 376, 377, 397, 398, 399, 400, 420, 421, 422, 423, 424, 442, 443, 444, 445, 446, 447, 464, 465, 466, 467, 468, 469, 470, 471, 487, 488, 489, 490, 491, 492, 493, 494, 511, 512, 513, 514, 515, 516, 517], "bbox": [0.205141065830721, 0.4449375, 0.5004388714733542, 0.9887656249999999], "iscrowd": 0, "area": 34844.55539999999, "rle": {"size": [640, 638], "counts": "`^b25fc07I7I7I7I8H7J6I7M3M3N2M4M2M3N2M3N2M100O100O100O100O100O010O100O100O100O100O100O100O100O100O100O100O100O100O100O1N2@`0@`0@`0_Oa0@`0@`0@?C>C=B>C=C=B>C=G9I7I7J6K500000000001O000000000000001O000000000000001O00000O100000001O000000000000001O000000000000001O000000000000001O000000=C>B?Ah0WOi0XOi0WOi0WOi0WOi0WOi0WO8H1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O4K5L3M4L4K5J5K6I7J6J5K6J6J6J5JZiV6"}, "label": "a beautiful flower vase on the table"}]}
{"id": 196111, "image": "COCO_train2014_000000196111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white vase with circular bumps all over holding a flower\"."}], "task": "refering", "answer_template": "The \"a white vase with circular bumps all over holding a flower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 239, 260, 261, 262, 283, 284, 285, 306, 307, 308, 329, 330, 331, 352, 353, 354, 374, 375, 376, 377, 397, 398, 399, 400, 420, 421, 422, 423, 424, 442, 443, 444, 445, 446, 447, 464, 465, 466, 467, 468, 469, 470, 471, 487, 488, 489, 490, 491, 492, 493, 494, 511, 512, 513, 514, 515, 516, 517], "bbox": [0.205141065830721, 0.4449375, 0.5004388714733542, 0.9887656249999999], "iscrowd": 0, "area": 34844.55539999999, "rle": {"size": [640, 638], "counts": "`^b25fc07I7I7I7I8H7J6I7M3M3N2M4M2M3N2M3N2M100O100O100O100O100O010O100O100O100O100O100O100O100O100O100O100O100O100O100O1N2@`0@`0@`0_Oa0@`0@`0@?C>C=B>C=C=B>C=G9I7I7J6K500000000001O000000000000001O000000000000001O00000O100000001O000000000000001O000000000000001O000000000000001O000000=C>B?Ah0WOi0XOi0WOi0WOi0WOi0WOi0WO8H1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O4K5L3M4L4K5J5K6I7J6J5K6J6J6J5JZiV6"}, "label": "a white vase with circular bumps all over holding a flower"}]}
{"id": 196112, "image": "COCO_train2014_000000196112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with colby jack cheese , tomato , and lettuce , on fresh cut bread\"."}], "task": "refering", "answer_template": "The \"a sandwich with colby jack cheese , tomato , and lettuce , on fresh cut bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.11290625000000001, 0.3533690987124463, 0.5645, 0.8028111587982832], "iscrowd": 0, "area": 41112.91620000001, "rle": {"size": [466, 640], "counts": "mPQ11`>4M2M4L3N3L3M4L3N3L3M4M3L3M4L3N3L3M4M2M2N1O2O1N2N2N101N2N2O1N1O2N2O1N2N101N2N1O2O1N1O2O1N1O2N2O0O2N2O0O2N2N2O0O2N2O0O2N2N101N1O2O0O2N1O101N1O2O0O2N1O2O0O2N101O0O2O000O2O0O10001N10000O2O0O10001N10000O2O0O10000O2O000O101N10000O2O000O101N10000O101O0O100O2O000O101O0O100O2O000O10001N100O101O0O10001N100O10000O10000O100O10000O10000O100O10000O1000O0100O10000O10000O100O10000O100O10O10O10000O010O10000O01000O100O01000O10O100000000O101O00000O101O00000O10001O000O1000001O0O1000001O0O101O3M3M3L5L3M3M3L4M3M3M3M3L4M3M3M3M3L4L4L4K5K5L4K5K5L4K5L4K5K5L4K5L4K5K4M3L4L4M3L4M3L4L4M3L4M3L4L4M3L`en3"}, "label": "a sandwich with colby jack cheese , tomato , and lettuce , on fresh cut bread"}]}
{"id": 196112, "image": "COCO_train2014_000000196112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller part of sandwich to left of knife with tomato , cheese and lettuce\"."}], "task": "refering", "answer_template": "The \"smaller part of sandwich to left of knife with tomato , cheese and lettuce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.11290625000000001, 0.3533690987124463, 0.5645, 0.8028111587982832], "iscrowd": 0, "area": 41112.91620000001, "rle": {"size": [466, 640], "counts": "mPQ11`>4M2M4L3N3L3M4L3N3L3M4M3L3M4L3N3L3M4M2M2N1O2O1N2N2N101N2N2O1N1O2N2O1N2N101N2N1O2O1N1O2O1N1O2N2O0O2N2O0O2N2N2O0O2N2O0O2N2N101N1O2O0O2N1O101N1O2O0O2N1O2O0O2N101O0O2O000O2O0O10001N10000O2O0O10001N10000O2O0O10000O2O000O101N10000O2O000O101N10000O101O0O100O2O000O101O0O100O2O000O10001N100O101O0O10001N100O10000O10000O100O10000O10000O100O10000O1000O0100O10000O10000O100O10000O100O10O10O10000O010O10000O01000O100O01000O10O100000000O101O00000O101O00000O10001O000O1000001O0O1000001O0O101O3M3M3L5L3M3M3L4M3M3M3M3L4M3M3M3M3L4L4L4K5K5L4K5K5L4K5L4K5K5L4K5L4K5K4M3L4L4M3L4M3L4L4M3L4M3L4L4M3L`en3"}, "label": "smaller part of sandwich to left of knife with tomato , cheese and lettuce"}]}
{"id": 196112, "image": "COCO_train2014_000000196112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long slice of bread\"."}], "task": "refering", "answer_template": "The \"a long slice of bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 31, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 223, 224, 225, 247], "bbox": [0.346015625, 0.023648068669527896, 0.915078125, 0.6092274678111588], "iscrowd": 0, "area": 45245.763300000006, "rle": {"size": [466, 640], "counts": "hQU34[>4L4N1O2N2O1N5K6K6I6J7J5J5K2N2O1O1O1O2N1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O10O0100O100O00100O100O001O1O1O10O01O1O1O001O1O1O1O001O1O100O010O1O100O00100O1O10O01O100O00100O100O00100O1O10O01O100O1O010O100O001000000O0100000O10O1000O10O1000O1000O10O100000O0100000O01000000O0100000O10O1000O10O1000O1000O10O100000O010000O10O1000O10O10O10000O01000O1000O10O1000O010000O10O10O10000O0100000O01000O1000O010000O10O1000O10O10O10000O0100000O10N2N2O0O2N2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N2O0O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N1N2O2M2O1N3N1N3N1N2O2N1N2O2M2O2M2O1N3N1N3N1N2O2N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1O2M2O1N3N1N3N1N2O2M2O1N3N1O2M2O2M2Onkh0"}, "label": "a long slice of bread"}]}
{"id": 196112, "image": "COCO_train2014_000000196112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"piece of plain baguette\"."}], "task": "refering", "answer_template": "The \"piece of plain baguette\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 31, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 86, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 150, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 223, 224, 225, 247], "bbox": [0.346015625, 0.023648068669527896, 0.915078125, 0.6092274678111588], "iscrowd": 0, "area": 45245.763300000006, "rle": {"size": [466, 640], "counts": "hQU34[>4L4N1O2N2O1N5K6K6I6J7J5J5K2N2O1O1O1O2N1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O10O0100O100O00100O100O001O1O1O10O01O1O1O001O1O1O1O001O1O100O010O1O100O00100O1O10O01O100O00100O100O00100O1O10O01O100O1O010O100O001000000O0100000O10O1000O10O1000O1000O10O100000O0100000O01000000O0100000O10O1000O10O1000O1000O10O100000O010000O10O1000O10O10O10000O01000O1000O10O1000O010000O10O10O10000O0100000O01000O1000O010000O10O1000O10O10O10000O0100000O10N2N2O0O2N2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N2O0O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N1N2O2M2O1N3N1N3N1N2O2N1N2O2M2O2M2O1N3N1N3N1N2O2N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1O2M2O1N3N1N3N1N2O2M2O1N3N1O2M2O2M2Onkh0"}, "label": "piece of plain baguette"}]}
{"id": 48665, "image": "COCO_train2014_000000048665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange closest to the banana\"."}], "task": "refering", "answer_template": "The \"the orange closest to the banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307], "bbox": [0.33141156462585036, 0.3366875, 0.6902551020408163, 0.653296875], "iscrowd": 0, "area": 32593.46660000001, "rle": {"size": [640, 588], "counts": "aUj35gc07I8H7I8H7K5K6J5K6J5K4L4L3M4L4L3M4L4M2N3M3M2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M3N1N3M101N101N2N101N2N101N101N2N101N2O0O2N101N2N101N2O0O2N101N2O001N2O0O2O001N2O0O2O1O0O2O0O1001O0000000000000000000000001O0000000000000000000000001O000000000000001O000000001O00000O2O0O100O2O000O100O2O0O100O2O000O101N100O10001N100O101N100O10001N101N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1O1M3N3L3M3N2M3N2M3N2M3N2M3N3L3N2M3M3N2M3N2M5L4K4M4K5L3L5L4K5K4H9G9F9Hg^a3"}, "label": "the orange closest to the banana"}]}
{"id": 48665, "image": "COCO_train2014_000000048665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange between other oranges and a banana\"."}], "task": "refering", "answer_template": "The \"orange between other oranges and a banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307], "bbox": [0.33141156462585036, 0.3366875, 0.6902551020408163, 0.653296875], "iscrowd": 0, "area": 32593.46660000001, "rle": {"size": [640, 588], "counts": "aUj35gc07I8H7I8H7K5K6J5K6J5K4L4L3M4L4L3M4L4M2N3M3M2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M3N1N3M101N101N2N101N2N101N101N2N101N2O0O2N101N2N101N2O0O2N101N2O001N2O0O2O001N2O0O2O1O0O2O0O1001O0000000000000000000000001O0000000000000000000000001O000000000000001O000000001O00000O2O0O100O2O000O100O2O0O100O2O000O101N100O10001N100O101N100O10001N101N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1O1M3N3L3M3N2M3N2M3N2M3N2M3N3L3N2M3M3N2M3N2M5L4K4M4K5L3L5L4K5K4H9G9F9Hg^a3"}, "label": "orange between other oranges and a banana"}]}
{"id": 48665, "image": "COCO_train2014_000000048665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange in top right corner of oranges\"."}], "task": "refering", "answer_template": "The \"orange in top right corner of oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 160, 161, 162, 163, 164, 165, 182, 183, 184, 185], "bbox": [0.6058333333333333, 0.087265625, 0.9369727891156463, 0.37971875], "iscrowd": 0, "area": 28636.806849999997, "rle": {"size": [640, 588], "counts": "^en66Vc0g0YOf0@a0K4L5K4L3M3M3M3M3M3M3M3M3M2N3M3M2N3M2N3M3M2N3M3M2O2N1O2O1O001O001O001O001N101N101N101N101O0O2O000O2O001N101O0O101N10001N101O00001O001N101O1O001O001O001O001O001O001O1N1000000000000000000000000000001N100000001O000000001O0000O100O101N100O100O100O2O0O100O100O2O0O100O100O100O100O100O2O0O100O101N5L1N101N2O1N101N2N1O2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2M3N2M3M2N3N2M3M3M3N2M3M5K7J5J7I7I6E<C<E<Cnof0"}, "label": "orange in top right corner of oranges"}]}
{"id": 48665, "image": "COCO_train2014_000000048665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"upper right hand orange in a group of three\"."}], "task": "refering", "answer_template": "The \"upper right hand orange in a group of three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 160, 161, 162, 163, 164, 165, 182, 183, 184, 185], "bbox": [0.6058333333333333, 0.087265625, 0.9369727891156463, 0.37971875], "iscrowd": 0, "area": 28636.806849999997, "rle": {"size": [640, 588], "counts": "^en66Vc0g0YOf0@a0K4L5K4L3M3M3M3M3M3M3M3M3M2N3M3M2N3M2N3M3M2N3M3M2O2N1O2O1O001O001O001O001N101N101N101N101O0O2O000O2O001N101O0O101N10001N101O00001O001N101O1O001O001O001O001O001O001O1N1000000000000000000000000000001N100000001O000000001O0000O100O101N100O100O100O2O0O100O100O2O0O100O100O100O100O100O2O0O100O101N5L1N101N2O1N101N2N1O2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2M3N2M3M2N3N2M3M3M3N2M3M5K7J5J7I7I6E<C<E<Cnof0"}, "label": "upper right hand orange in a group of three"}]}
{"id": 48665, "image": "COCO_train2014_000000048665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange on top left side\"."}], "task": "refering", "answer_template": "The \"orange on top left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 44, 45, 46, 47, 48, 49, 64, 65, 66, 67, 68, 69, 70, 85, 86, 87, 88, 89, 90, 91, 92, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 172, 173], "bbox": [0.07683673469387756, 0.058828124999999995, 0.38671768707482995, 0.36], "iscrowd": 0, "area": 27365.8003, "rle": {"size": [640, 588], "counts": "`Xl09mb0l0J7J5K6K4K6J5L5J5K6K4K5K5L3L4N2N2N2M3N2N2M4M2N2N2M3N2N2N2M4M2O1N2O1O1N2O1O1N101O0O2O000O2O000O2O000O2O001N10001N10001N101O0O101O0O101O0O2O000O2O000O2O001N10001O0O101O0O101O0O2O000O10000O10000000000000000000O10000000001O000000000O100O1O100O100O1O2O0O1O100O100O1O100O100O1O100O2N101N2O2M2N2O1N3N1N2N2O2M2O1N2N3M2N2N2N3M2M3N2N3M2N2N2N3M2N2N3L6I7H8I7I7H8I7H8F:D<E;D]`Q7"}, "label": "orange on top left side"}]}
{"id": 48665, "image": "COCO_train2014_000000048665.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange on the top left\"."}], "task": "refering", "answer_template": "The \"the orange on the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 44, 45, 46, 47, 48, 49, 64, 65, 66, 67, 68, 69, 70, 85, 86, 87, 88, 89, 90, 91, 92, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 172, 173], "bbox": [0.07683673469387756, 0.058828124999999995, 0.38671768707482995, 0.36], "iscrowd": 0, "area": 27365.8003, "rle": {"size": [640, 588], "counts": "`Xl09mb0l0J7J5K6K4K6J5L5J5K6K4K5K5L3L4N2N2N2M3N2N2M4M2N2N2M3N2N2N2M4M2O1N2O1O1N2O1O1N101O0O2O000O2O000O2O000O2O001N10001N10001N101O0O101O0O101O0O2O000O2O000O2O001N10001O0O101O0O101O0O2O000O10000O10000000000000000000O10000000001O000000000O100O1O100O100O1O2O0O1O100O100O1O100O100O1O100O2N101N2O2M2N2O1N3N1N2N2O2M2O1N2N3M2N2N2N3M2M3N2N3M2N2N2N3M2N2N3L6I7H8I7I7H8I7H8F:D<E;D]`Q7"}, "label": "the orange on the top left"}]}
{"id": 114202, "image": "COCO_train2014_000000114202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller giraffe eats while in the shade\"."}], "task": "refering", "answer_template": "The \"the smaller giraffe eats while in the shade\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 143, 144, 160, 161, 178, 179, 195, 196, 197, 212, 213, 214, 229, 230, 231, 246, 247, 248], "bbox": [0.432625, 0.319890625, 0.6261041666666666, 0.639859375], "iscrowd": 0, "area": 6620.074949999997, "rle": {"size": [640, 480], "counts": "mVR43ic04M4K4M3N3O01O04M3M3L4M3L5L3L4N4Q_O\\NY?f1`@`N_?c1X@cNg?a1Q@eNn?^1i_OhNW`0[1a_OkN]`0X1[_OoNd`0X201N101O0O3N2N2h@YLo=k3jAVLY>l3aATLb>o3XARLk>Q4n@PLU?R4d@oK^?Y41O1O2cN\\1:F:F:F9HO0O1kM`@Ia?5QAYOQ?f0aAhN`>V1SBWNn=h1W2N2O0O2N2N101N20N3M2O2M2N3M2O2M2N3M2m^OTNo?n1Z_OgNe`0[2O1O01N1ZOf0[Of0L35K6J[NZ^OU1ea0dNc^O\\1Tb0O10O01O1O1J5J7I7H7I8I7He__3"}, "label": "the smaller giraffe eats while in the shade"}]}
{"id": 114202, "image": "COCO_train2014_000000114202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smaller giraffe standing to the left of a larger giraffe\"."}], "task": "refering", "answer_template": "The \"a smaller giraffe standing to the left of a larger giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 143, 144, 160, 161, 178, 179, 195, 196, 197, 212, 213, 214, 229, 230, 231, 246, 247, 248], "bbox": [0.432625, 0.319890625, 0.6261041666666666, 0.639859375], "iscrowd": 0, "area": 6620.074949999997, "rle": {"size": [640, 480], "counts": "mVR43ic04M4K4M3N3O01O04M3M3L4M3L5L3L4N4Q_O\\NY?f1`@`N_?c1X@cNg?a1Q@eNn?^1i_OhNW`0[1a_OkN]`0X1[_OoNd`0X201N101O0O3N2N2h@YLo=k3jAVLY>l3aATLb>o3XARLk>Q4n@PLU?R4d@oK^?Y41O1O2cN\\1:F:F:F9HO0O1kM`@Ia?5QAYOQ?f0aAhN`>V1SBWNn=h1W2N2O0O2N2N101N20N3M2O2M2N3M2O2M2N3M2m^OTNo?n1Z_OgNe`0[2O1O01N1ZOf0[Of0L35K6J[NZ^OU1ea0dNc^O\\1Tb0O10O01O1O1J5J7I7H7I8I7He__3"}, "label": "a smaller giraffe standing to the left of a larger giraffe"}]}
{"id": 425497, "image": "COCO_train2014_000000425497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the entire blue board\"."}], "task": "refering", "answer_template": "The \"the entire blue board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 162, 163, 164, 165, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 285, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.20879166666666665, 0.062859375, 0.7815, 1.0], "iscrowd": 0, "area": 141247.24725000001, "rle": {"size": [640, 480], "counts": "Zdn1=bc0[1dN<D7J7H8H7J7mIiLoJ_3k4jLjJ]3P5lLgJ[3R5oLcJY3V5PM`JW3Z5SM[JU3^5UMXJR3a5WMUJP3e5RMXJV3a5kL^J\\3\\5dLbJc3X5^LfJj3S5WLlJP4m4QLQKV4i4kKUK]4d4dK[Kc4^4^K`Ki4Z4XKdKP5U4QKjKV5P4jJnK]5l3dJRLd5g3]JXLj5a3WJ]LP6]3QJaLW6X3jIfL^6S3cIlLc6n2^IPMj6i2WIUMQ7e2oHZMW7`2jH^M^7[2cHcMd7W2]HhMj7Q2WHmMQ8l1PHRNW8h1jGWN]8b1dG\\Nd8]1]GaNj8Z1VGeNQ9T1PGkNW9n0jFQO\\9i0eFWOa9b0`F^Of9;[FEj95WFJP:OQF1U:HlE8Y:CgE=_:\\ObEd0d:UO]Ej0Q;hNPEX1l>0000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10O1000O100000000O1000000O1001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O004L;E;E;E;E;E<D;E;E;E;E<D;E;E;E;E;E<D;E;E;E;ET5lJdTQ2"}, "label": "the entire blue board"}]}
{"id": 425497, "image": "COCO_train2014_000000425497.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue snowboard with the letters gmr written on it in yellow\"."}], "task": "refering", "answer_template": "The \"a blue snowboard with the letters gmr written on it in yellow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 157, 158, 159, 160, 161, 162, 163, 164, 165, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 285, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.20879166666666665, 0.062859375, 0.7815, 1.0], "iscrowd": 0, "area": 141247.24725000001, "rle": {"size": [640, 480], "counts": "Zdn1=bc0[1dN<D7J7H8H7J7mIiLoJ_3k4jLjJ]3P5lLgJ[3R5oLcJY3V5PM`JW3Z5SM[JU3^5UMXJR3a5WMUJP3e5RMXJV3a5kL^J\\3\\5dLbJc3X5^LfJj3S5WLlJP4m4QLQKV4i4kKUK]4d4dK[Kc4^4^K`Ki4Z4XKdKP5U4QKjKV5P4jJnK]5l3dJRLd5g3]JXLj5a3WJ]LP6]3QJaLW6X3jIfL^6S3cIlLc6n2^IPMj6i2WIUMQ7e2oHZMW7`2jH^M^7[2cHcMd7W2]HhMj7Q2WHmMQ8l1PHRNW8h1jGWN]8b1dG\\Nd8]1]GaNj8Z1VGeNQ9T1PGkNW9n0jFQO\\9i0eFWOa9b0`F^Of9;[FEj95WFJP:OQF1U:HlE8Y:CgE=_:\\ObEd0d:UO]Ej0Q;hNPEX1l>0000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O10O1000O100000000O1000000O1001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O004L;E;E;E;E;E<D;E;E;E;E<D;E;E;E;E;E<D;E;E;E;ET5lJdTQ2"}, "label": "a blue snowboard with the letters gmr written on it in yellow"}]}
{"id": 97818, "image": "COCO_train2014_000000097818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and brown shorts riding a bicycle and pulling a cart behind\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and brown shorts riding a bicycle and pulling a cart behind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 86, 87, 97, 98, 99, 100, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 133, 134, 135, 136, 145, 146, 147, 148, 158, 169, 170], "bbox": [0.03036036036036036, 0.34832, 0.41165165165165163, 0.8337], "iscrowd": 0, "area": 11600.469350000003, "rle": {"size": [500, 333], "counts": "VU51`?3M3N2O2O0000001O00001N1N2N2N3M2O0O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1\\EkNi6V1WIlNf6U1ZImNc6T1^ImN`6S1`IoN]6R1\\GiNa06Q8R1]GFEZOk8Q1_GN_OROo8Q1bG2ZOoNR9o0cG8WOjNS9o0eG=UOdNS9P1gGc0RO]NT9Q1iGh0POWNT9R1lGl0mNRNU9Q1nGT1iNkMV9Q1QHZ1fNdMX9R1QHn1l7RNUHP2i7oMWHT2g7mMWHV2g7kMWHX2g7jMVHY2g7jMWH;nNJi8LWH;WOD`83VH:@_OX88VH:IUOS8c0QH93kNm7n0hG>b0\\Ng7X1[Gf0U1iMb7]4eHZK]7f4h10O10001O000O100000000O1003M3N2M3M3M3M4L3M3M3M3M3N2M3M3M3MhDbL^:Y43M2N2N2N2N2N2QOZDlMh;Q2^DjMd;T2`DhMb;U2cDgM_;W2fDdM];X2hDdMZ;Y2kDcMW;[2S1M3N2M3M3N2M3N3L3N2M3M3N2M3N2M3M3N2@Z^o2"}, "label": "a man in a white shirt and brown shorts riding a bicycle and pulling a cart behind"}]}
{"id": 97818, "image": "COCO_train2014_000000097818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man is wearing a pair of khaki shorts\"."}], "task": "refering", "answer_template": "The \"the man is wearing a pair of khaki shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 86, 87, 97, 98, 99, 100, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 133, 134, 135, 136, 145, 146, 147, 148, 158, 169, 170], "bbox": [0.03036036036036036, 0.34832, 0.41165165165165163, 0.8337], "iscrowd": 0, "area": 11600.469350000003, "rle": {"size": [500, 333], "counts": "VU51`?3M3N2O2O0000001O00001N1N2N2N3M2O0O1O1O1O1O1O1O1O1O1O100O1O1O1O100O1\\EkNi6V1WIlNf6U1ZImNc6T1^ImN`6S1`IoN]6R1\\GiNa06Q8R1]GFEZOk8Q1_GN_OROo8Q1bG2ZOoNR9o0cG8WOjNS9o0eG=UOdNS9P1gGc0RO]NT9Q1iGh0POWNT9R1lGl0mNRNU9Q1nGT1iNkMV9Q1QHZ1fNdMX9R1QHn1l7RNUHP2i7oMWHT2g7mMWHV2g7kMWHX2g7jMVHY2g7jMWH;nNJi8LWH;WOD`83VH:@_OX88VH:IUOS8c0QH93kNm7n0hG>b0\\Ng7X1[Gf0U1iMb7]4eHZK]7f4h10O10001O000O100000000O1003M3N2M3M3M3M4L3M3M3M3M3N2M3M3M3MhDbL^:Y43M2N2N2N2N2N2QOZDlMh;Q2^DjMd;T2`DhMb;U2cDgM_;W2fDdM];X2hDdMZ;Y2kDcMW;[2S1M3N2M3M3N2M3N3L3N2M3M3N2M3N2M3M3N2@Z^o2"}, "label": "the man is wearing a pair of khaki shorts"}]}
{"id": 97818, "image": "COCO_train2014_000000097818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing on bike\"."}], "task": "refering", "answer_template": "The \"man standing on bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 50, 51, 52, 62, 63, 64, 65, 74, 75, 76, 77, 86, 87, 88, 89, 99, 100, 101, 111, 112, 113, 114, 123, 124, 125, 126, 134, 135, 137, 138, 146, 147, 149, 150, 158, 159], "bbox": [0.1585885885885886, 0.1955, 0.5601201201201201, 0.7527999999999999], "iscrowd": 0, "area": 15338.042900000004, "rle": {"size": [500, 333], "counts": "dWj05_?3L10000O1oJGgJ9W5IhJ8V5JjJ6U5KjJ6]3OQILc35\\33kHJi33^2E_Ic04Gn32^2F^Ig00BT41]2I]Ih0L_OZ41[2L\\Ii0I\\O`4OZ2N[Im0EWOe4OZ20ZIn0AUOk4LZ24WI_1_4^NY25VI^1a4]NX28UI[1c4]NW2;SIY1e4]NX2;QIX1h4]NV2>oHV1i4^NX2b2`MfM_2Z2ZMnMf2R2RMVNm2k1kL]NT3d1dLdN\\3[1]LmNb3T1VLTOj3h0SL_Om34ZL2g3JWL;i3BTLd0l3YORLk0P4ROlKT1T4iNjK\\1W4`NfKf1Z4YNaKm1`4VNWKm0l5VOjIi0_6ZOWId0R7^OdHf0`7\\OWHg0l7\\OjGh0Z8ZO]Gh0h8ZOnFj0V9g23M4L4L3M4L3M2O1O100O100O101N100O100O100O1K5L4K5L5J5L4K5L4L44O001O1O001N1000N2N2O1N2N1O2N2N2N2N2cLnCU3Z<O1O2N1O1O2O0O1O2N1O101NPDUMV;j2fDYM];e2_D]Md;b2XDaMj;_2QDdMR<Z2jCiMX<h24M2N3dM`C]1b<_NbC]1c<]NaC`1c<[NbCa1c<YNaCd1Z=M4K4M4WOPBGS>KYB4d>O1O2N1N2O1O\\^W2"}, "label": "man standing on bike"}]}
{"id": 97818, "image": "COCO_train2014_000000097818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man riding on the back\"."}], "task": "refering", "answer_template": "The \"man riding on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 50, 51, 52, 62, 63, 64, 65, 74, 75, 76, 77, 86, 87, 88, 89, 99, 100, 101, 111, 112, 113, 114, 123, 124, 125, 126, 134, 135, 137, 138, 146, 147, 149, 150, 158, 159], "bbox": [0.1585885885885886, 0.1955, 0.5601201201201201, 0.7527999999999999], "iscrowd": 0, "area": 15338.042900000004, "rle": {"size": [500, 333], "counts": "dWj05_?3L10000O1oJGgJ9W5IhJ8V5JjJ6U5KjJ6]3OQILc35\\33kHJi33^2E_Ic04Gn32^2F^Ig00BT41]2I]Ih0L_OZ41[2L\\Ii0I\\O`4OZ2N[Im0EWOe4OZ20ZIn0AUOk4LZ24WI_1_4^NY25VI^1a4]NX28UI[1c4]NW2;SIY1e4]NX2;QIX1h4]NV2>oHV1i4^NX2b2`MfM_2Z2ZMnMf2R2RMVNm2k1kL]NT3d1dLdN\\3[1]LmNb3T1VLTOj3h0SL_Om34ZL2g3JWL;i3BTLd0l3YORLk0P4ROlKT1T4iNjK\\1W4`NfKf1Z4YNaKm1`4VNWKm0l5VOjIi0_6ZOWId0R7^OdHf0`7\\OWHg0l7\\OjGh0Z8ZO]Gh0h8ZOnFj0V9g23M4L4L3M4L3M2O1O100O100O101N100O100O100O1K5L4K5L5J5L4K5L4L44O001O1O001N1000N2N2O1N2N1O2N2N2N2N2cLnCU3Z<O1O2N1O1O2O0O1O2N1O101NPDUMV;j2fDYM];e2_D]Md;b2XDaMj;_2QDdMR<Z2jCiMX<h24M2N3dM`C]1b<_NbC]1c<]NaC`1c<[NbCa1c<YNaCd1Z=M4K4M4WOPBGS>KYB4d>O1O2N1N2O1O\\^W2"}, "label": "man riding on the back"}]}
{"id": 278045, "image": "COCO_train2014_000000278045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest zebra\"."}], "task": "refering", "answer_template": "The \"the closest zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 265, 266, 283, 284, 285, 288, 289, 308, 311, 312], "bbox": [0.30615625, 0.47679166666666667, 0.643515625, 0.8119791666666667], "iscrowd": 0, "area": 17286.595499999996, "rle": {"size": [480, 640], "counts": "iUl2<^>8K6J5K6I6K6K4K5K7I9H8G8H3N3M2N2O1O01O0001O0001O0000001O0O100000001O000^OSDRNm;l1XDRNh;l1\\DQNd;m1aDPN`;m1eDPN\\;m1iDPNW;n1o0M2N3O0010O010O010O010O01001N101O0O2O001Oe1ZNf0[O5K000O1000000O10000O]Od0H8H8H7XOi0N2N2O0O2N2001N10000O10000O100000000000000O100000001O00000000000000000000001O0000000000000000001O00000000000000O2O000000001N10aEgMW8Y2fGjMZ8V2cGmM]8S2`GPNa8o1\\GTNd8l1YGWNg8h1XGZNh8f1VG\\Nk8e1QG]No8e1mF]NS9e1iF]NW9d1fF^NZ9d1cF]N^9d1^F[Ne9g1WFXNl9j1PFUNS:m1iERNZ:P2_ERNd:P2TESNo:k24M3M3M3M3M3L4O10fEnLf8R3UGSMk8m2QGWMo8i2mF[MS9e2cFeM]9b2RFhMn9e30YMlEb0U:[OPFb0P:\\OUFa0l9\\OXFc0g9[O`F`0a9]OgF=Y9@oF;R9BUG7m8GZG2i8nNlE5`1h0g8SORFI\\1o0d8XOXHd0k7\\OWH?k7AWH<k7DWH0R80`30O01O010O010O01O01O010O0000010O00000O1O1O1O2N2KYdZ3"}, "label": "the closest zebra"}]}
{"id": 278045, "image": "COCO_train2014_000000278045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"closest zebra\"."}], "task": "refering", "answer_template": "The \"closest zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 265, 266, 283, 284, 285, 288, 289, 308, 311, 312], "bbox": [0.30615625, 0.47679166666666667, 0.643515625, 0.8119791666666667], "iscrowd": 0, "area": 17286.595499999996, "rle": {"size": [480, 640], "counts": "iUl2<^>8K6J5K6I6K6K4K5K7I9H8G8H3N3M2N2O1O01O0001O0001O0000001O0O100000001O000^OSDRNm;l1XDRNh;l1\\DQNd;m1aDPN`;m1eDPN\\;m1iDPNW;n1o0M2N3O0010O010O010O010O01001N101O0O2O001Oe1ZNf0[O5K000O1000000O10000O]Od0H8H8H7XOi0N2N2O0O2N2001N10000O10000O100000000000000O100000001O00000000000000000000001O0000000000000000001O00000000000000O2O000000001N10aEgMW8Y2fGjMZ8V2cGmM]8S2`GPNa8o1\\GTNd8l1YGWNg8h1XGZNh8f1VG\\Nk8e1QG]No8e1mF]NS9e1iF]NW9d1fF^NZ9d1cF]N^9d1^F[Ne9g1WFXNl9j1PFUNS:m1iERNZ:P2_ERNd:P2TESNo:k24M3M3M3M3M3L4O10fEnLf8R3UGSMk8m2QGWMo8i2mF[MS9e2cFeM]9b2RFhMn9e30YMlEb0U:[OPFb0P:\\OUFa0l9\\OXFc0g9[O`F`0a9]OgF=Y9@oF;R9BUG7m8GZG2i8nNlE5`1h0g8SORFI\\1o0d8XOXHd0k7\\OWH?k7AWH<k7DWH0R80`30O01O010O010O01O01O010O0000010O00000O1O1O1O2N2KYdZ3"}, "label": "closest zebra"}]}
{"id": 278045, "image": "COCO_train2014_000000278045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra hidden behind two other zebras ' bodies\"."}], "task": "refering", "answer_template": "The \"a zebra hidden behind two other zebras ' bodies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 241, 242, 264, 265, 287, 288], "bbox": [0.387734375, 0.40658333333333335, 0.566015625, 0.7575416666666667], "iscrowd": 0, "area": 3714.1269500000003, "rle": {"size": [480, 640], "counts": "emd32m>1O1N2O1O1N2O1O1N2O1O1O1000000O100000000O10000000000O10000000000000000000000000O100000000000000000000000000000000000dD^OZ8b0eG@Z8`0eGBnM0Z9>hHDhM4^99iHEdM6b95jHLZM3k91kHh0T7XOlHi0T7VOlHk0S7UOmHk0S7UOmHl0R7UOmHj0T7VOkHi0W7WOiHg0Y7YOgHd0\\7]OcHa0_7_OaH`0cMoN[9a0RI>bMTO\\9?QI;bMYO]9<QI:aM\\O^9:QI8`MB^97QI7^MEa94QI9ZMDf93PIo0Q7ROmHo0S7QOmHn0T7ROkHn0V7SOiHl0X7TOgHm0Z7ROfHn0[7QOeHo0[7QOdHo0^7QOaHo0`7PO`HP1`7PO`Ho0b7PO]HP1e7oN[Ho0h7POXHo0i7QOYH9_MEZ:1[IEj6:n310O1O00100O002N101N2N2N2O0OSSR4"}, "label": "a zebra hidden behind two other zebras ' bodies"}]}
{"id": 278045, "image": "COCO_train2014_000000278045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra who ' s head and neck isn ' t visible\"."}], "task": "refering", "answer_template": "The \"the zebra who ' s head and neck isn ' t visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 241, 242, 264, 265, 287, 288], "bbox": [0.387734375, 0.40658333333333335, 0.566015625, 0.7575416666666667], "iscrowd": 0, "area": 3714.1269500000003, "rle": {"size": [480, 640], "counts": "emd32m>1O1N2O1O1N2O1O1N2O1O1O1000000O100000000O10000000000O10000000000000000000000000O100000000000000000000000000000000000dD^OZ8b0eG@Z8`0eGBnM0Z9>hHDhM4^99iHEdM6b95jHLZM3k91kHh0T7XOlHi0T7VOlHk0S7UOmHk0S7UOmHl0R7UOmHj0T7VOkHi0W7WOiHg0Y7YOgHd0\\7]OcHa0_7_OaH`0cMoN[9a0RI>bMTO\\9?QI;bMYO]9<QI:aM\\O^9:QI8`MB^97QI7^MEa94QI9ZMDf93PIo0Q7ROmHo0S7QOmHn0T7ROkHn0V7SOiHl0X7TOgHm0Z7ROfHn0[7QOeHo0[7QOdHo0^7QOaHo0`7PO`HP1`7PO`Ho0b7PO]HP1e7oN[Ho0h7POXHo0i7QOYH9_MEZ:1[IEj6:n310O1O00100O002N101N2N2N2O0OSSR4"}, "label": "the zebra who ' s head and neck isn ' t visible"}]}
{"id": 278045, "image": "COCO_train2014_000000278045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zebras forage with heads down in tall grass\"."}], "task": "refering", "answer_template": "The \"two zebras forage with heads down in tall grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 235, 236, 237], "bbox": [0.23776562499999998, 0.4449166666666667, 0.60271875, 0.6557499999999999], "iscrowd": 0, "area": 6202.259749999999, "rle": {"size": [480, 640], "counts": "i`W22f>9H8M3N2N2N2N2N2N2N2N2N2N2N2M4M2N2N2N2N2N2N3N2M3M2N3M00N2N2N2N2N201O0000000000000010O2O00000O1000N1O2N2N1O2N2N1O2N2N1O200O1O001O10O01O1O00100O001O1O10O01O1O00100O001O1O2O0O1O1O10000O100O10000O100O10001N10000O100O10000O100O10000O10000O100O10000O100O10000O01000O1000000000000000O100000000000000000000000000000000001O0000000000001O0000000000000000000000001O001O001O001O00010O001O0000001O01O0001O0000000010O000001O00000010O000001O0000100O010O1O010O1O010O10O01O10O0Q?0`jf3"}, "label": "two zebras forage with heads down in tall grass"}]}
{"id": 278045, "image": "COCO_train2014_000000278045.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle zebra in a group of three zebras\"."}], "task": "refering", "answer_template": "The \"the middle zebra in a group of three zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 235, 236, 237], "bbox": [0.23776562499999998, 0.4449166666666667, 0.60271875, 0.6557499999999999], "iscrowd": 0, "area": 6202.259749999999, "rle": {"size": [480, 640], "counts": "i`W22f>9H8M3N2N2N2N2N2N2N2N2N2N2N2M4M2N2N2N2N2N2N3N2M3M2N3M00N2N2N2N2N201O0000000000000010O2O00000O1000N1O2N2N1O2N2N1O2N2N1O200O1O001O10O01O1O00100O001O1O10O01O1O00100O001O1O2O0O1O1O10000O100O10000O100O10001N10000O100O10000O100O10000O10000O100O10000O100O10000O01000O1000000000000000O100000000000000000000000000000000001O0000000000001O0000000000000000000000001O001O001O001O00010O001O0000001O01O0001O0000000010O000001O00000010O000001O0000100O010O1O010O1O010O10O01O10O0Q?0`jf3"}, "label": "the middle zebra in a group of three zebras"}]}
{"id": 24095, "image": "COCO_train2014_000000024095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young giraffe standing behind an adult giraffe in captivity ; fencing is present\"."}], "task": "refering", "answer_template": "The \"young giraffe standing behind an adult giraffe in captivity ; fencing is present\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 114, 115, 116, 130, 131, 132, 148, 149, 150, 166, 167, 168, 184, 185, 186], "bbox": [0.2528, 0.42696000000000006, 0.49213999999999997, 0.8449333333333334], "iscrowd": 0, "area": 4479.762299999999, "rle": {"size": [375, 500], "counts": "j[^12c;c0\\O<D5L4K5L3M4M2UFfNV9\\2hF`M]8l20@eGeM\\8V2kGgMU8U2QHeMT8m211O1000001OEaG^M^8^2iG`MW8[2e0K5J6L4M3L4M2M4M2001O1O1O1O100O1O1O1O1O1O1jGWN]6n1RI_Nn6f1`HgN`7]1QHnNo7]2O10mNQH^No7T1^HkNd7h0iHTO[7>RI^OS75YIFgNAf7f0eIE]N2n75hI`0V6^OlId0S6YOoIGcN9]7MQKNo41WKKh44^KHa47`3O1O1O001O1000O1O010O1O10O0100O1O010O1O01O01O01O010O00010O00010O00010O00010O00010O00010O00010O00010O0001O01O100O1O1Mekl2"}, "label": "young giraffe standing behind an adult giraffe in captivity ; fencing is present"}]}
{"id": 24095, "image": "COCO_train2014_000000024095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe in the rear whose head is not visible\"."}], "task": "refering", "answer_template": "The \"giraffe in the rear whose head is not visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 114, 115, 116, 130, 131, 132, 148, 149, 150, 166, 167, 168, 184, 185, 186], "bbox": [0.2528, 0.42696000000000006, 0.49213999999999997, 0.8449333333333334], "iscrowd": 0, "area": 4479.762299999999, "rle": {"size": [375, 500], "counts": "j[^12c;c0\\O<D5L4K5L3M4M2UFfNV9\\2hF`M]8l20@eGeM\\8V2kGgMU8U2QHeMT8m211O1000001OEaG^M^8^2iG`MW8[2e0K5J6L4M3L4M2M4M2001O1O1O1O100O1O1O1O1O1O1jGWN]6n1RI_Nn6f1`HgN`7]1QHnNo7]2O10mNQH^No7T1^HkNd7h0iHTO[7>RI^OS75YIFgNAf7f0eIE]N2n75hI`0V6^OlId0S6YOoIGcN9]7MQKNo41WKKh44^KHa47`3O1O1O001O1000O1O010O1O10O0100O1O010O1O01O01O01O010O00010O00010O00010O00010O00010O00010O00010O00010O0001O01O100O1O1Mekl2"}, "label": "giraffe in the rear whose head is not visible"}]}
{"id": 32289, "image": "COCO_train2014_000000032289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first bench on the right\"."}], "task": "refering", "answer_template": "The \"the first bench on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 313, 314, 317, 318], "bbox": [0.428890625, 0.34189583333333334, 0.890375, 0.8229166666666666], "iscrowd": 0, "area": 39888.07014999999, "rle": {"size": [480, 640], "counts": "mfP42n>7I0O1000000000000bCLP::hEHX:;dEF\\::cEG\\::cEG]:9bEH^:8`EJ`:6^ELb:4\\ENd:3YEOg:1WE1i:OUE3k:MSE5l:LQE7o:IoD9Q;HlD:T;GgD=Y;C_De0a;EmCc0S<U10000O1000000000000000000O1000000000E;O1O10O01O1O100O1O1O1O100O010O100O10000O1000O010000O10000O100000O1000001O000000001N100000000M4L3L4M3L5K4M4K5L4K5L4K4L5L4K5L4K5K5L3L5L4K7K5M2N00000O100nFhJ\\8X5bGkJ]8U5bGlJ^8T5`GnJ`8R5^GPKa8Q5^GPKb8P5\\GSKc8m4\\GTKd8l4ZGVKf8b5O101O0000000O10000000000O100000000O10000000000O10000000000O10000000_OdG^J[8`5iG_JW8`5kG_JU8`5f000O1O00100O1O1O10O01O100O10O0100O100O10O0100O10000O0100000000O10000000000O10000000000O100000000O2O001O001O0O2O001O00001N101O0YG\\JR8e5mG]JQ8c5oG]JP8e5PHZJP8f5PHZJP8h5nGXJR8i5mGWJS8j5lGWJS8j5lGVJS8l5mGSJS8n5lGRJT8o5kGQJU8P6jGPJV8^6O0O100000000O10O10000000ORNnGPMR8h2[HRMf7g2gHSMY7f2QIWMo6b2[I[Me6^2eI_MZ6^2mI_MS6_2RJ^Mn5a2VJ[Mk5c2YJ[Mg5d2[J[Md5d2`JZM`5e2bJZM^5d2fJZMZ5e2hJYMY5e2kJYMU5f2mJYMR5f2RKXMn4g2TKXMl4d2YK[Mg4`2^K_Mc4]2aKcM^4Y2hKfMX4V2lKjMT4Q2QLoMo3m1ULSNk3h1ZLWNg3e1_LYN`3c1gL[NY3a1mL]NS3^1TM`Nl2[1_5L4K4M4K6K6I6J6ImdP1"}, "label": "the first bench on the right"}]}
{"id": 32289, "image": "COCO_train2014_000000032289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger bench\"."}], "task": "refering", "answer_template": "The \"the larger bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 313, 314, 317, 318], "bbox": [0.428890625, 0.34189583333333334, 0.890375, 0.8229166666666666], "iscrowd": 0, "area": 39888.07014999999, "rle": {"size": [480, 640], "counts": "mfP42n>7I0O1000000000000bCLP::hEHX:;dEF\\::cEG\\::cEG]:9bEH^:8`EJ`:6^ELb:4\\ENd:3YEOg:1WE1i:OUE3k:MSE5l:LQE7o:IoD9Q;HlD:T;GgD=Y;C_De0a;EmCc0S<U10000O1000000000000000000O1000000000E;O1O10O01O1O100O1O1O1O100O010O100O10000O1000O010000O10000O100000O1000001O000000001N100000000M4L3L4M3L5K4M4K5L4K5L4K4L5L4K5L4K5K5L3L5L4K7K5M2N00000O100nFhJ\\8X5bGkJ]8U5bGlJ^8T5`GnJ`8R5^GPKa8Q5^GPKb8P5\\GSKc8m4\\GTKd8l4ZGVKf8b5O101O0000000O10000000000O100000000O10000000000O10000000000O10000000_OdG^J[8`5iG_JW8`5kG_JU8`5f000O1O00100O1O1O10O01O100O10O0100O100O10O0100O10000O0100000000O10000000000O10000000000O100000000O2O001O001O0O2O001O00001N101O0YG\\JR8e5mG]JQ8c5oG]JP8e5PHZJP8f5PHZJP8h5nGXJR8i5mGWJS8j5lGWJS8j5lGVJS8l5mGSJS8n5lGRJT8o5kGQJU8P6jGPJV8^6O0O100000000O10O10000000ORNnGPMR8h2[HRMf7g2gHSMY7f2QIWMo6b2[I[Me6^2eI_MZ6^2mI_MS6_2RJ^Mn5a2VJ[Mk5c2YJ[Mg5d2[J[Md5d2`JZM`5e2bJZM^5d2fJZMZ5e2hJYMY5e2kJYMU5f2mJYMR5f2RKXMn4g2TKXMl4d2YK[Mg4`2^K_Mc4]2aKcM^4Y2hKfMX4V2lKjMT4Q2QLoMo3m1ULSNk3h1ZLWNg3e1_LYN`3c1gL[NY3a1mL]NS3^1TM`Nl2[1_5L4K4M4K6K6I6J6ImdP1"}, "label": "the larger bench"}]}
{"id": 32289, "image": "COCO_train2014_000000032289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench with green moss growing on it that is in the left backround\"."}], "task": "refering", "answer_template": "The \"a wooden bench with green moss growing on it that is in the left backround\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 236, 237, 238, 239, 240, 259, 261, 262, 263, 282, 284, 285, 286], "bbox": [0.25078125, 0.4278541666666667, 0.54309375, 0.7619375], "iscrowd": 0, "area": 12694.90005, "rle": {"size": [480, 640], "counts": "ng[21n>2O1O1O1O1O1O2N1O1iBFf;;UDIk;8PDLP<5lCNT<4fC0Z<1aC3^<O^C4b<MYC7g<JTC:l<GPC<P=j0000000000000000eNSCc0m<]O\\C:d<FeC0[<1nCFR<:VD^Oj;b0Z1000000000000000000000O10000000000000000000000000000O10000_C^Od:b0XEBh:>TEFl::PEJP;6lDNT;2hD2X;NeD6Z;KaD9_;G]D=c;CYDa0g;_OUDe0i;]OSDg0i;]OSDg0j;\\ORDh0j;\\ORDh0k;Y1K4M3M3L4M3L4M3L5L3L4M0ON3N4O2000O0100000000O10O100000O100000O0100000000O1000OWNSEAm:?UE_Ok:`0XE]Oi:c0YE[Og:e0ZEZOf:f0\\EXOd:h0^EVOa:k0`ESOa:l0bERO^:n0dEPO\\:P1eEoN[:Q1gElNZ:T1hEjNX:U1kEiNT:X1mEgNS:Y1oEeNQ:[1QFbNP:^1QFaNo9_1SF_Nm9`1VF]Nk9c1VF[Nj9f1XFWNi9i1YFTNh9l1ZFQNg9n1[FPNf9P2\\FlMf9T2c10O10O1000000000O0100000000O10O10000000O1000O1000O1000000N2K4M4L4K5L4L4K5L3M4L4K5L4L4K5L3MgeX4"}, "label": "a wooden bench with green moss growing on it that is in the left backround"}]}
{"id": 32289, "image": "COCO_train2014_000000032289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench on the left\"."}], "task": "refering", "answer_template": "The \"the bench on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 236, 237, 238, 239, 240, 259, 261, 262, 263, 282, 284, 285, 286], "bbox": [0.25078125, 0.4278541666666667, 0.54309375, 0.7619375], "iscrowd": 0, "area": 12694.90005, "rle": {"size": [480, 640], "counts": "ng[21n>2O1O1O1O1O1O2N1O1iBFf;;UDIk;8PDLP<5lCNT<4fC0Z<1aC3^<O^C4b<MYC7g<JTC:l<GPC<P=j0000000000000000eNSCc0m<]O\\C:d<FeC0[<1nCFR<:VD^Oj;b0Z1000000000000000000000O10000000000000000000000000000O10000_C^Od:b0XEBh:>TEFl::PEJP;6lDNT;2hD2X;NeD6Z;KaD9_;G]D=c;CYDa0g;_OUDe0i;]OSDg0i;]OSDg0j;\\ORDh0j;\\ORDh0k;Y1K4M3M3L4M3L4M3L5L3L4M0ON3N4O2000O0100000000O10O100000O100000O0100000000O1000OWNSEAm:?UE_Ok:`0XE]Oi:c0YE[Og:e0ZEZOf:f0\\EXOd:h0^EVOa:k0`ESOa:l0bERO^:n0dEPO\\:P1eEoN[:Q1gElNZ:T1hEjNX:U1kEiNT:X1mEgNS:Y1oEeNQ:[1QFbNP:^1QFaNo9_1SF_Nm9`1VF]Nk9c1VF[Nj9f1XFWNi9i1YFTNh9l1ZFQNg9n1[FPNf9P2\\FlMf9T2c10O10O1000000000O0100000000O10O10000000O1000O1000O1000000N2K4M4L4K5L4L4K5L3M4L4K5L4L4K5L3MgeX4"}, "label": "the bench on the left"}]}
{"id": 433704, "image": "COCO_train2014_000000433704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the bat wearing his shirt untucked\"."}], "task": "refering", "answer_template": "The \"the man with the bat wearing his shirt untucked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 142, 143, 144, 165, 166, 167, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 305], "bbox": [0.176078125, 0.31997916666666665, 0.29628125, 0.7804791666666666], "iscrowd": 0, "area": 10454.56415, "rle": {"size": [480, 640], "counts": "iYe11n>1IO^A2a>1\\A0d>600O1UL2dHMY7<bHDZ7Y1lGhNP8`1lG`NP8h1lGXNQ8P2jGPNU8U2fGlMY8Y2cGgM[7ElGi2e0aM`7HhGP3?YMh7JfGZ36lLT8KdGe3M`L_8MaGP4CTLl8N^Gd4b8^K[Gc4e8_KXGb4h8`KUG`4k8bKSG_4m8cKPG^4P9dKmF]4S9h00000O10O100O1J6K5K7I8H8G8I8H7I5O24K5K5L4K4L5L4K5K5L3L5K5L4K5L3L5L4OK5L4L5K4L4L4L4L7I1O2N1YLUFU2o9cMYFX2k9`M^FZ2h9]MaF^2c9ZMSkb6"}, "label": "the man with the bat wearing his shirt untucked"}]}
{"id": 433704, "image": "COCO_train2014_000000433704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player holding the bat away from his body\"."}], "task": "refering", "answer_template": "The \"a baseball player holding the bat away from his body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 142, 143, 144, 165, 166, 167, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 305], "bbox": [0.176078125, 0.31997916666666665, 0.29628125, 0.7804791666666666], "iscrowd": 0, "area": 10454.56415, "rle": {"size": [480, 640], "counts": "iYe11n>1IO^A2a>1\\A0d>600O1UL2dHMY7<bHDZ7Y1lGhNP8`1lG`NP8h1lGXNQ8P2jGPNU8U2fGlMY8Y2cGgM[7ElGi2e0aM`7HhGP3?YMh7JfGZ36lLT8KdGe3M`L_8MaGP4CTLl8N^Gd4b8^K[Gc4e8_KXGb4h8`KUG`4k8bKSG_4m8cKPG^4P9dKmF]4S9h00000O10O100O1J6K5K7I8H8G8I8H7I5O24K5K5L4K4L5L4K5K5L3L5K5L4K5L3L5L4OK5L4L5K4L4L4L4L7I1O2N1YLUFU2o9cMYFX2k9`M^FZ2h9]MaF^2c9ZMSkb6"}, "label": "a baseball player holding the bat away from his body"}]}
{"id": 433704, "image": "COCO_train2014_000000433704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball player on the right\"."}], "task": "refering", "answer_template": "The \"the baseball player on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 85, 86, 108, 109, 110, 130, 131, 132, 153, 154, 155, 177, 178, 199, 200, 201, 222, 223, 224, 246, 247], "bbox": [0.663640625, 0.15839583333333335, 0.79328125, 0.5992916666666666], "iscrowd": 0, "area": 8431.688749999998, "rle": {"size": [480, 640], "counts": "U_W63k>3M3M2N3M2M3N2N2N3kMXOnEj0o9\\OlEf0R:_OiEc0T:CfE`0X:EcE<[:IaE9]:L^E5`:1[E1c:4XEMf:9TEJg:>TECc:j0XEXO^:T1^EmNY:R3L30100O1[F\\Km8\\5O100O1O100O1O1O100O1O12OG8N2N2O1N2N2LAcGcJ\\8X5kGgJT8X5nGiJP8V5SHiJl7i2cGVOe0PNg7j2iGoNc0WNc7k2oGfNb0^N_7k2TH_Nb0eNZ7l2XHWNb0mNW7e2bHVN;UOR7^2nHVN4ZOo6Z2XISNNAk6U2bISNGGh6o1kIRNT8j1c21O01O0000101N1hNmBa0X=YOkBe0V=YOlBe0U=ZOnBd0R=\\OPCa0R=^OQC?o<@TC=m<BVC<l<@YC<h=N3M2M4L3Mkgm1"}, "label": "the baseball player on the right"}]}
{"id": 433704, "image": "COCO_train2014_000000433704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the batter in the right side a 6 in his jearsy\"."}], "task": "refering", "answer_template": "The \"the batter in the right side a 6 in his jearsy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 85, 86, 108, 109, 110, 130, 131, 132, 153, 154, 155, 177, 178, 199, 200, 201, 222, 223, 224, 246, 247], "bbox": [0.663640625, 0.15839583333333335, 0.79328125, 0.5992916666666666], "iscrowd": 0, "area": 8431.688749999998, "rle": {"size": [480, 640], "counts": "U_W63k>3M3M2N3M2M3N2N2N3kMXOnEj0o9\\OlEf0R:_OiEc0T:CfE`0X:EcE<[:IaE9]:L^E5`:1[E1c:4XEMf:9TEJg:>TECc:j0XEXO^:T1^EmNY:R3L30100O1[F\\Km8\\5O100O1O100O1O1O100O1O12OG8N2N2O1N2N2LAcGcJ\\8X5kGgJT8X5nGiJP8V5SHiJl7i2cGVOe0PNg7j2iGoNc0WNc7k2oGfNb0^N_7k2TH_Nb0eNZ7l2XHWNb0mNW7e2bHVN;UOR7^2nHVN4ZOo6Z2XISNNAk6U2bISNGGh6o1kIRNT8j1c21O01O0000101N1hNmBa0X=YOkBe0V=YOlBe0U=ZOnBd0R=\\OPCa0R=^OQC?o<@TC=m<BVC<l<@YC<h=N3M2M4L3Mkgm1"}, "label": "the batter in the right side a 6 in his jearsy"}]}
{"id": 343598, "image": "COCO_train2014_000000343598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white macbook with the firefox sticker\"."}], "task": "refering", "answer_template": "The \"the white macbook with the firefox sticker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 230, 231], "bbox": [0.0, 0.1438262910798122, 0.495109375, 0.7258450704225352], "iscrowd": 0, "area": 37203.18315, "rle": {"size": [426, 640], "counts": "o4k0_<e1ZNg1ZN`0@O100O010O10N1O2N101N1O2N2N1100O10O0100O10O0100O010O100O010O100O010O10O0100O10O0100O10O0100O01000O10O0100O010O100O010O100O010O10O0100O10O0100O10O0100O010O100O01000O10O0100O010O100O010O10O0100O10O0100O10O0100O010O100O010O100O010O10O10O100O010O10O0100O10O0100O10O0100O010O100O010O100O010O10O0100O10O10O100O010O10O0100O10O0100O010O100O010O100O010O10O0100O10O0100O10O01000O0100O10O0100O10O0100O010O100O010O10O0100O10O0100O10O0100O010O1000O0100O10O0100O010O100O010O1O0O2O1N2O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O0O2O1N2O0O2O1N2O0O2N2O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O0O2O1N2O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O0O2N2N2N1N4M3M3L4MoYV4"}, "label": "the white macbook with the firefox sticker"}]}
{"id": 343598, "image": "COCO_train2014_000000343598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white computer\"."}], "task": "refering", "answer_template": "The \"the white computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 230, 231], "bbox": [0.0, 0.1438262910798122, 0.495109375, 0.7258450704225352], "iscrowd": 0, "area": 37203.18315, "rle": {"size": [426, 640], "counts": "o4k0_<e1ZNg1ZN`0@O100O010O10N1O2N101N1O2N2N1100O10O0100O10O0100O010O100O010O100O010O10O0100O10O0100O10O0100O01000O10O0100O010O100O010O100O010O10O0100O10O0100O10O0100O010O100O01000O10O0100O010O100O010O10O0100O10O0100O10O0100O010O100O010O100O010O10O10O100O010O10O0100O10O0100O10O0100O010O100O010O100O010O10O0100O10O10O100O010O10O0100O10O0100O010O100O010O100O010O10O0100O10O0100O10O01000O0100O10O0100O10O0100O010O100O010O10O0100O10O0100O10O0100O010O1000O0100O10O0100O010O100O010O1O0O2O1N2O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O0O2O1N2O0O2O1N2O0O2N2O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O0O2O1N2O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O0O2N2N2N1N4M3M3L4MoYV4"}, "label": "the white computer"}]}
{"id": 343598, "image": "COCO_train2014_000000343598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black samsung laptop next to a white apple computer\"."}], "task": "refering", "answer_template": "The \"a black samsung laptop next to a white apple computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158], "bbox": [0.4793125, 0.003380281690140845, 0.9991875, 0.4548591549295775], "iscrowd": 0, "area": 50852.34360000001, "rle": {"size": [426, 640], "counts": "Xlo38g<;D=D;D<E;D<E;D<E;D<E;D<E;D<H80000010O000000000000000000000000000001O0000000000000000000000000000001O00000000000000000001O000000000001O0000000000000000000000000000001O0000000000000000000000000000001O000001O00000000000000000000000001O0000000000000000000000000000001O000000000000000001O000000000001O01O0000000000001O0000000000000010O00000000000001O0000000000001O01O000000000001O000000000000001O01O0000000001O000000000000001O0001O0000000001O0000000000001O00000001O000001O0000000000001O0000000001O0001O000N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2M3N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N_<"}, "label": "a black samsung laptop next to a white apple computer"}]}
{"id": 343598, "image": "COCO_train2014_000000343598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black samsung laptop on a table with a green circular sticker on it\"."}], "task": "refering", "answer_template": "The \"a black samsung laptop on a table with a green circular sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158], "bbox": [0.4793125, 0.003380281690140845, 0.9991875, 0.4548591549295775], "iscrowd": 0, "area": 50852.34360000001, "rle": {"size": [426, 640], "counts": "Xlo38g<;D=D;D<E;D<E;D<E;D<E;D<E;D<H80000010O000000000000000000000000000001O0000000000000000000000000000001O00000000000000000001O000000000001O0000000000000000000000000000001O0000000000000000000000000000001O000001O00000000000000000000000001O0000000000000000000000000000001O000000000000000001O000000000001O01O0000000000001O0000000000000010O00000000000001O0000000000001O01O000000000001O000000000000001O01O0000000001O000000000000001O0001O0000000001O0000000000001O00000001O000001O0000000000001O0000000001O0001O000N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2M3N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N_<"}, "label": "a black samsung laptop on a table with a green circular sticker on it"}]}
{"id": 458286, "image": "COCO_train2014_000000458286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black keyboard\"."}], "task": "refering", "answer_template": "The \"the black keyboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [303, 304, 305, 324, 325, 326, 327, 328, 329, 348, 349, 350, 351, 352, 353, 372, 373, 374, 375, 376, 377, 378, 396, 397, 398, 399, 400, 401, 402, 420, 421, 422, 423, 424, 425, 444, 445, 446], "bbox": [0.11134796238244515, 0.5836718750000001, 0.5209874608150471, 0.8586406249999999], "iscrowd": 0, "area": 20183.658099999986, "rle": {"size": [640, 638], "counts": "fh\\11oc01O1N2O1N2O001N2O1N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O1N2O1O1N101N2O1O1N2O1N2O1O1N2O1N101O1N2O1N2O1O1N2O1N2O1O0O2O1N2O1O1N2O1N2O1O1N2O0O2O1O1N2O1N2O1O1N2O1N2O001N2O1O100O10000O10000O0100000O10000O10000O1000O10O10000O10000O1000000O01000O10000O1000000O1000O010000O1000000O10000O10O10O1000000O10000O10000O0100000O10000OO2O1O001O1O001N2O1O001O1O001O1N101O1O1O001O1N101O1O001O1O1O0O2O1O0O2O1N1O2O1N2N101N2N101N2O0O2N2O1N1O2O1N1O2O1N101N2N2O0O2N2O0O2N2O0O2N2O1N101N2N101N2N101N2N2O0O2O1N1O2O1N1O2O1N2N101N2O0O2N2O0O2N2O1N1O2O1N10^hn5"}, "label": "the black keyboard"}]}
{"id": 458286, "image": "COCO_train2014_000000458286.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black keyboard for a pc sitting on a wooden computer desk\"."}], "task": "refering", "answer_template": "The \"a black keyboard for a pc sitting on a wooden computer desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [303, 304, 305, 324, 325, 326, 327, 328, 329, 348, 349, 350, 351, 352, 353, 372, 373, 374, 375, 376, 377, 378, 396, 397, 398, 399, 400, 401, 402, 420, 421, 422, 423, 424, 425, 444, 445, 446], "bbox": [0.11134796238244515, 0.5836718750000001, 0.5209874608150471, 0.8586406249999999], "iscrowd": 0, "area": 20183.658099999986, "rle": {"size": [640, 638], "counts": "fh\\11oc01O1N2O1N2O001N2O1N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O1N2O1O1N101N2O1O1N2O1N2O1O1N2O1N101O1N2O1N2O1O1N2O1N2O1O0O2O1N2O1O1N2O1N2O1O1N2O0O2O1O1N2O1N2O1O1N2O1N2O001N2O1O100O10000O10000O0100000O10000O10000O1000O10O10000O10000O1000000O01000O10000O1000000O1000O010000O1000000O10000O10O10O1000000O10000O10000O0100000O10000OO2O1O001O1O001N2O1O001O1O001O1N101O1O1O001O1N101O1O001O1O1O0O2O1O0O2O1N1O2O1N2N101N2N101N2O0O2N2O1N1O2O1N1O2O1N101N2N2O0O2N2O0O2N2O0O2N2O1N101N2N101N2N101N2N2O0O2O1N1O2O1N1O2O1N2N101N2O0O2N2O0O2N2O1N1O2O1N10^hn5"}, "label": "a black keyboard for a pc sitting on a wooden computer desk"}]}
{"id": 179763, "image": "COCO_train2014_000000179763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter giraffe with its next in front of the other giraffe\"."}], "task": "refering", "answer_template": "The \"the shorter giraffe with its next in front of the other giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 102, 103, 115, 116, 117, 129, 130, 131, 132, 144, 145, 146, 159, 160, 173, 174, 175, 187, 188, 189, 202, 203, 204, 216, 217, 218, 231, 232, 233, 245, 246, 247, 248, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 319, 320, 321, 322, 323, 333, 334, 335, 336, 337, 338], "bbox": [0.24381733021077284, 0.25440625, 0.895456674473068, 1.0], "iscrowd": 0, "area": 34434.45479999999, "rle": {"size": [640, 427], "counts": "ocQ21mc02M3N2N2M3N2M3N2N2eN[OX_Oh0d`0DR_O>j`0Ml^O6o`07g^OKUa0b1L4M3L4L4M2M4M3J6I7K5M3N2M3M3M3M3N2M3N2N2N2O1N2O1N3M10O0N2N3N1O2N2O1O001O1O1O1O1O001O1O1O10O4M2N2N1O2N1N3N1O2N1O2N1N3N1O1N3M3]BdJ^<_5ZChJd<Z5UCmJj<U5oBQKo<R5gBWKW=i5OM3L4L4M3M32O0O1O1O1O1O2M2O1O001O001O005K8H7I7I7I8H7I7I5K2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2O1N101N2OmIcFe1[9jMeGi1Z8VNSHY1Q8gNXHl0m7SO[Ha0i7_O_H5e7KdHHa77gH^O\\7b0lHROY7m0QIaNW7_1[IbMS7]2W501N101O2N1O2N1O2N2N1O2N1O1O1N101O1OUOiMo_OW2P`0lMo_OS2Q`0PNm_Oo1T`0RNk_Om1V`0UNg_Oj1[`0XNc_Og1^`0ZNa_Oe1_`0^N__Oa1b`0`N\\_O`1e`0bNY_O]1h`0dNW_O[1j`0gNT_OY1k`0iNT_OW1l`0kNR_OT1o`0mNP_OS1Pa0n00O10O0EVMc_Oj2^`0VMa_Oi2``0YM^_Of2c`0[M\\_Oe2d`0]MZ_Ob2g`0`MW_O`2i`0aMV_O^2k`0:1M3N2N2M4M2O0100O010O100O0101O1N2O1O1O2M2O1O1N2O1O2M101O000010O01O0010O0001O001O001N100O2O000O100O001O1N2N2O0O1O010O1O101N2N2O2M2N21O0010O010O01000O1O1N2O1N2N2O1O2N2N1O2N2M3N5Gink0"}, "label": "the shorter giraffe with its next in front of the other giraffe"}]}
{"id": 179763, "image": "COCO_train2014_000000179763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smaller giraffe rubbing it ' s head against the neck of a larger giraffe\"."}], "task": "refering", "answer_template": "The \"a smaller giraffe rubbing it ' s head against the neck of a larger giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 102, 103, 115, 116, 117, 129, 130, 131, 132, 144, 145, 146, 159, 160, 173, 174, 175, 187, 188, 189, 202, 203, 204, 216, 217, 218, 231, 232, 233, 245, 246, 247, 248, 259, 260, 261, 262, 263, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 319, 320, 321, 322, 323, 333, 334, 335, 336, 337, 338], "bbox": [0.24381733021077284, 0.25440625, 0.895456674473068, 1.0], "iscrowd": 0, "area": 34434.45479999999, "rle": {"size": [640, 427], "counts": "ocQ21mc02M3N2N2M3N2M3N2N2eN[OX_Oh0d`0DR_O>j`0Ml^O6o`07g^OKUa0b1L4M3L4L4M2M4M3J6I7K5M3N2M3M3M3M3N2M3N2N2N2O1N2O1N3M10O0N2N3N1O2N2O1O001O1O1O1O1O001O1O1O10O4M2N2N1O2N1N3N1O2N1O2N1N3N1O1N3M3]BdJ^<_5ZChJd<Z5UCmJj<U5oBQKo<R5gBWKW=i5OM3L4L4M3M32O0O1O1O1O1O2M2O1O001O001O005K8H7I7I7I8H7I7I5K2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2O1N101N2OmIcFe1[9jMeGi1Z8VNSHY1Q8gNXHl0m7SO[Ha0i7_O_H5e7KdHHa77gH^O\\7b0lHROY7m0QIaNW7_1[IbMS7]2W501N101O2N1O2N1O2N2N1O2N1O1O1N101O1OUOiMo_OW2P`0lMo_OS2Q`0PNm_Oo1T`0RNk_Om1V`0UNg_Oj1[`0XNc_Og1^`0ZNa_Oe1_`0^N__Oa1b`0`N\\_O`1e`0bNY_O]1h`0dNW_O[1j`0gNT_OY1k`0iNT_OW1l`0kNR_OT1o`0mNP_OS1Pa0n00O10O0EVMc_Oj2^`0VMa_Oi2``0YM^_Of2c`0[M\\_Oe2d`0]MZ_Ob2g`0`MW_O`2i`0aMV_O^2k`0:1M3N2N2M4M2O0100O010O100O0101O1N2O1O1O2M2O1O1N2O1O2M101O000010O01O0010O0001O001O001N100O2O000O100O001O1N2N2O0O1O010O1O101N2N2O2M2N21O0010O010O01000O1O1N2O1N2N2O1O2N2N1O2N2M3N5Gink0"}, "label": "a smaller giraffe rubbing it ' s head against the neck of a larger giraffe"}]}
{"id": 179763, "image": "COCO_train2014_000000179763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller giraffe\"."}], "task": "refering", "answer_template": "The \"the taller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 35, 36, 37, 38, 51, 52, 53, 54, 66, 67, 68, 69, 82, 83, 84, 98, 99, 113, 114, 115, 128, 129, 130, 144, 159, 160, 161, 175, 176, 189, 190, 191, 204, 205, 206, 207, 219, 220, 221, 222, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 338, 339, 340, 341, 342, 343], "bbox": [0.31660421545667444, 0.0561875, 0.8992974238875878, 0.9887656250000001], "iscrowd": 0, "area": 45028.14364999998, "rle": {"size": [640, 427], "counts": "_]d21nc02M3N2N2M4M2N2M3O2N1O1O1O1O2OO01O00001O001O00001O010O001O01O01O1O3N1N2N2N2O2M2N2N2O1N3M2N101N1O1O101N1O1O100O2N1O100O3M7I7J3L1O01N1N3L3M4M2M4M2N31O1N2O1N101N2O1N2O1N101N2N2O1N2N1O2O1N2N2N2O0O2N2N2O1N1O2N2O1N2NlLf_Od2X`0\\Mj_Og2Q`0XMR@j2c8cMaMBoIm2U8PNVKiNeN:PNQ3i7[NQKdNVOORNU3[7hNkJ^NGEUNX3Q7QObJ[N8\\OVN\\3j6UOWJ]Ni0QOXN`3c6[OjI^N[1gNYNc3]6@^I`Nl1\\N[Ng3U6k1`K]J]Nk3n5i1fK[J]No3h5g1mKWJ]NV4`5d1ULSJ]N\\4Y5c1[LoI]Na4S5a1^MbN\\2`1eMcNV2^1jMgNP2Z1QNiNj1X1WNkNd1V1]NmN^1T1cNoNX1S1hNPOR1R1nNSOk0o0VOTOc0o0]OXN`LlKP4n5ATN`LmKm3R6CPNbLlKj3V6LeM\\LSLg3Z6<SMoK`Ld3a6a0hLmKfLa3d6g0`LjKjL]3i6n0WLgKnLZ3m6S1PLeKQMW3Q7X1hKcKUMT3V7\\1aKbKWMQ3Z7a1ZK`KZMm2_7f1SK_K\\M[2R8Y2^JPNT6S2gI^Mj6e2RIZMP7i2lHVMV7n2eHRM\\7Q3_HoLc7T3YHkLi7X3eGnIYOi2T9\\3`GmIYOf2Y9a3ZGkIYOe2^9c3VGjIYOb2c9g3QGaLQ9b3lF\\LV9g3gFWL[9m3aFQL_8T5^GkJ_8\\5^GbJ`8e5]GYJa8n5\\GQJa8g8N3M2N3M2011N2O1O1N2O1N3N3L4O11O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O17I4KN3M3M3M3M3M2N3M3M3M3M3M3M2M4YN[CmKh<T4YCgKj<Z4XC`Kj<b4WCYKk<i4VCRKl<P5ZCfJi<[5dCTJ^<n5j01O2N2O0O2N1O00N2N3M2O1N2N2O2M2N2N4Mb0]Of1ZN3M2O2QOn0UOl0_Oa0_O^[j0"}, "label": "the taller giraffe"}]}
{"id": 179763, "image": "COCO_train2014_000000179763.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 zebras\"."}], "task": "refering", "answer_template": "The \"2 zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 35, 36, 37, 38, 51, 52, 53, 54, 66, 67, 68, 69, 82, 83, 84, 98, 99, 113, 114, 115, 128, 129, 130, 144, 159, 160, 161, 175, 176, 189, 190, 191, 204, 205, 206, 207, 219, 220, 221, 222, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 338, 339, 340, 341, 342, 343], "bbox": [0.31660421545667444, 0.0561875, 0.8992974238875878, 0.9887656250000001], "iscrowd": 0, "area": 45028.14364999998, "rle": {"size": [640, 427], "counts": "_]d21nc02M3N2N2M4M2N2M3O2N1O1O1O1O2OO01O00001O001O00001O010O001O01O01O1O3N1N2N2N2O2M2N2N2O1N3M2N101N1O1O101N1O1O100O2N1O100O3M7I7J3L1O01N1N3L3M4M2M4M2N31O1N2O1N101N2O1N2O1N101N2N2O1N2N1O2O1N2N2N2O0O2N2N2O1N1O2N2O1N2NlLf_Od2X`0\\Mj_Og2Q`0XMR@j2c8cMaMBoIm2U8PNVKiNeN:PNQ3i7[NQKdNVOORNU3[7hNkJ^NGEUNX3Q7QObJ[N8\\OVN\\3j6UOWJ]Ni0QOXN`3c6[OjI^N[1gNYNc3]6@^I`Nl1\\N[Ng3U6k1`K]J]Nk3n5i1fK[J]No3h5g1mKWJ]NV4`5d1ULSJ]N\\4Y5c1[LoI]Na4S5a1^MbN\\2`1eMcNV2^1jMgNP2Z1QNiNj1X1WNkNd1V1]NmN^1T1cNoNX1S1hNPOR1R1nNSOk0o0VOTOc0o0]OXN`LlKP4n5ATN`LmKm3R6CPNbLlKj3V6LeM\\LSLg3Z6<SMoK`Ld3a6a0hLmKfLa3d6g0`LjKjL]3i6n0WLgKnLZ3m6S1PLeKQMW3Q7X1hKcKUMT3V7\\1aKbKWMQ3Z7a1ZK`KZMm2_7f1SK_K\\M[2R8Y2^JPNT6S2gI^Mj6e2RIZMP7i2lHVMV7n2eHRM\\7Q3_HoLc7T3YHkLi7X3eGnIYOi2T9\\3`GmIYOf2Y9a3ZGkIYOe2^9c3VGjIYOb2c9g3QGaLQ9b3lF\\LV9g3gFWL[9m3aFQL_8T5^GkJ_8\\5^GbJ`8e5]GYJa8n5\\GQJa8g8N3M2N3M2011N2O1O1N2O1N3N3L4O11O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O17I4KN3M3M3M3M3M2N3M3M3M3M3M3M2M4YN[CmKh<T4YCgKj<Z4XC`Kj<b4WCYKk<i4VCRKl<P5ZCfJi<[5dCTJ^<n5j01O2N2O0O2N1O00N2N3M2O1N2N2O2M2N2N4Mb0]Of1ZN3M2O2QOn0UOl0_Oa0_O^[j0"}, "label": "2 zebras"}]}
{"id": 237110, "image": "COCO_train2014_000000237110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shoppa in the hall\"."}], "task": "refering", "answer_template": "The \"a shoppa in the hall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 291, 292], "bbox": [0.573265625, 0.4451041666666667, 0.8035625, 0.7348958333333333], "iscrowd": 0, "area": 13258.483549999999, "rle": {"size": [480, 640], "counts": "iZ\\54h>4L4L5K4L4L5N1O101N1WOPOaCQ1_<PO`CQ1_<PO_CQ1`<RO]CP1b<QO]Co0c<SOZCn0f<SOXCo0g<e000O2O00001N100O2O0O101N100O2O001N100O2O000O2O0O101N10001N100O2O001N1O101N1O2O2M3L4K5\\OiLYE[3b:kLXEY3e:b001O1O1O1O00000000000000001O02N1O00001O00000000001N1O100O100O1O2O0O1O100O101N1O100O100O1O2O0O100O1O101N1O100O100O2N100O2O0O1O2O0_OZDgMg;U2]DkMd;P2aDnM`;n1dDQN];k1gDUNZ;f1kDXNV;d1nDZNT;b1QE\\NQ;_1SE_No:^1TEaNm:]1]1M2N3C<L5L3L4L5K4Lkjj1"}, "label": "a shoppa in the hall"}]}
{"id": 237110, "image": "COCO_train2014_000000237110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair that is on the right side of the table that has a floral design on the seat\"."}], "task": "refering", "answer_template": "The \"the chair that is on the right side of the table that has a floral design on the seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 291, 292], "bbox": [0.573265625, 0.4451041666666667, 0.8035625, 0.7348958333333333], "iscrowd": 0, "area": 13258.483549999999, "rle": {"size": [480, 640], "counts": "iZ\\54h>4L4L5K4L4L5N1O101N1WOPOaCQ1_<PO`CQ1_<PO_CQ1`<RO]CP1b<QO]Co0c<SOZCn0f<SOXCo0g<e000O2O00001N100O2O0O101N100O2O001N100O2O000O2O0O101N10001N100O2O001N1O101N1O2O2M3L4K5\\OiLYE[3b:kLXEY3e:b001O1O1O1O00000000000000001O02N1O00001O00000000001N1O100O100O1O2O0O1O100O101N1O100O100O1O2O0O100O1O101N1O100O100O2N100O2O0O1O2O0_OZDgMg;U2]DkMd;P2aDnM`;n1dDQN];k1gDUNZ;f1kDXNV;d1nDZNT;b1QE\\NQ;_1SE_No:^1TEaNm:]1]1M2N3C<L5L3L4L5K4Lkjj1"}, "label": "the chair that is on the right side of the table that has a floral design on the seat"}]}
{"id": 237110, "image": "COCO_train2014_000000237110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green chair with its back to the viewer\"."}], "task": "refering", "answer_template": "The \"the green chair with its back to the viewer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 383, 384, 385], "bbox": [0.5567968750000001, 0.6609375, 0.8214375, 0.9974791666666666], "iscrowd": 0, "area": 13739.023899999998, "rle": {"size": [480, 640], "counts": "kWW5m0S>n0RO9G1O1O2N1O00O1QOo0L41N2O1O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O001O1O0O2O1O001N101O1O0O2J6K4K6K4K6L4O0O20O0100O010O01O001O001O001O001OaDZMd:a31O001O00000ZOUEWMk:i2f00001N1000000000000000000000000O1000000000001O0000000000000O1000000000000000000O101N100O100O100O100O100O1001O3M3M3M3M3M3M3N2M3M2NQOiDUNW;^1ZE^Nf:\\1eE_NZ:_1nE\\NQ:b1WFYNh9e1_FWN`9g1hFTNW9j1P2N2N2N101N2N2NWOcNTD[1P<eNnCY1U<iNhCU1[<mNbCQ1a<QO\\Cm0g<UOVCi0n<WOPCg0S=[OjBc0Y=_OdB?_==5L5J6J7J5J6J7IfSe1"}, "label": "the green chair with its back to the viewer"}]}
{"id": 237110, "image": "COCO_train2014_000000237110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the turquoise chair where the back is visible\"."}], "task": "refering", "answer_template": "The \"the turquoise chair where the back is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 383, 384, 385], "bbox": [0.5567968750000001, 0.6609375, 0.8214375, 0.9974791666666666], "iscrowd": 0, "area": 13739.023899999998, "rle": {"size": [480, 640], "counts": "kWW5m0S>n0RO9G1O1O2N1O00O1QOo0L41N2O1O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O001O1O0O2O1O001N101O1O0O2J6K4K6K4K6L4O0O20O0100O010O01O001O001O001O001OaDZMd:a31O001O00000ZOUEWMk:i2f00001N1000000000000000000000000O1000000000001O0000000000000O1000000000000000000O101N100O100O100O100O100O1001O3M3M3M3M3M3M3N2M3M2NQOiDUNW;^1ZE^Nf:\\1eE_NZ:_1nE\\NQ:b1WFYNh9e1_FWN`9g1hFTNW9j1P2N2N2N101N2N2NWOcNTD[1P<eNnCY1U<iNhCU1[<mNbCQ1a<QO\\Cm0g<UOVCi0n<WOPCg0S=[OjBc0Y=_OdB?_==5L5J6J7J5J6J7IfSe1"}, "label": "the turquoise chair where the back is visible"}]}
{"id": 302656, "image": "COCO_train2014_000000302656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffes\"."}], "task": "refering", "answer_template": "The \"two giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 246, 247, 248, 262, 263, 264, 278, 279, 280, 295, 296, 297, 312, 314, 329, 331], "bbox": [0.36033333333333334, 0.59046875, 0.6036041666666667, 0.845125], "iscrowd": 0, "area": 4038.207200000001, "rle": {"size": [640, 480], "counts": "kd\\31dc05_\\ON\\c0=M3M3M3@YOd]O`0O^O\\b0m0;N3N2N2N2NXOi0O1O0S^OlNSa0S1g^OZOSa0h0c^OA\\a0a10De^OUNZa0e1S_OTNm`0g1i0J5O2N1O2N1O2N1O2O0O2XO`]OO`b0Of]OL\\b02i]OJ[b01R]4JkcK2N2N2U]OKka07T^OLka03T^O0la0OQ^O4oa0KP^O8oa0Go]O=Pb0Bm]Ob0Rb0_Ok]Oe0Tb0ZOj]Oj0Ub0UOl]Ol0Sb0SOm]On0T1VOW?Lf_Oo0D50XOd`0Ch_Ok2V`0UMj_On2a`02O1N2O1OQO[_OdNd`0Y1l_O\\NR`0b1\\@TNW?T2[1N3N2M2O2M3O1\\OQ^OVOPb0j0U^OPOma0n0Y^OlNia0R1c0O1N1O2N2O1N1O2O1N2N101O1O100O1O100O1O1010O2K401O1O001O11N3N3L4M0O1O1O101N1O2N`[f3"}, "label": "two giraffes"}]}
{"id": 302656, "image": "COCO_train2014_000000302656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 giraffes behind a tree\"."}], "task": "refering", "answer_template": "The \"2 giraffes behind a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 246, 247, 248, 262, 263, 264, 278, 279, 280, 295, 296, 297, 312, 314, 329, 331], "bbox": [0.36033333333333334, 0.59046875, 0.6036041666666667, 0.845125], "iscrowd": 0, "area": 4038.207200000001, "rle": {"size": [640, 480], "counts": "kd\\31dc05_\\ON\\c0=M3M3M3@YOd]O`0O^O\\b0m0;N3N2N2N2NXOi0O1O0S^OlNSa0S1g^OZOSa0h0c^OA\\a0a10De^OUNZa0e1S_OTNm`0g1i0J5O2N1O2N1O2N1O2O0O2XO`]OO`b0Of]OL\\b02i]OJ[b01R]4JkcK2N2N2U]OKka07T^OLka03T^O0la0OQ^O4oa0KP^O8oa0Go]O=Pb0Bm]Ob0Rb0_Ok]Oe0Tb0ZOj]Oj0Ub0UOl]Ol0Sb0SOm]On0T1VOW?Lf_Oo0D50XOd`0Ch_Ok2V`0UMj_On2a`02O1N2O1OQO[_OdNd`0Y1l_O\\NR`0b1\\@TNW?T2[1N3N2M2O2M3O1\\OQ^OVOPb0j0U^OPOma0n0Y^OlNia0R1c0O1N1O2N2O1N1O2O1N2N101O1O100O1O100O1O1010O2K401O1O001O11N3N3L4M0O1O1O101N1O2N`[f3"}, "label": "2 giraffes behind a tree"}]}
{"id": 24129, "image": "COCO_train2014_000000024129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right half of the sandwich\"."}], "task": "refering", "answer_template": "The \"the right half of the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 84, 85, 86, 87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 344], "bbox": [0.469796875, 0.16129166666666667, 1.0, 0.841375], "iscrowd": 0, "area": 85506.45885000002, "rle": {"size": [480, 640], "counts": "[Z]4;]><D<D<D<D<D<D<D<D<L3L5L4K5L4K5L4L4K5L4K5L3L100O2O0O101N100O2O0O2O000O2O0O101N100O2O0O101N100O2O0O2O0O101N100O2O0O101N100O2O000O2O0O101N101N100O2O0O101N100O2O0O101N100O101N100O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2N100O100O2O0O100O2O0O100O2O0O100O2N100O101N100O101N100O101N100O101N1O100O101N100O101O000O101O0O10001O0O10001N1000001N1000001N10000O2O00000O101O000O101O0O10001O0O10001O0O10001N11O00010O00000010O0001O0001O01O000010O0001O0000010O00001O01O01O0000010O00001O0001O01O000010O0001O0001O01O00001O01O01O000010O01O001O01O01O001O0010O0001O0010O0001O0010O01O00001O010O001O00010O001O0010O0001O001O010O00001O010O001O00010O001O1O00100O001O1O010O1O001O10O01O1O001O10jK"}, "label": "the right half of the sandwich"}]}
{"id": 24129, "image": "COCO_train2014_000000024129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a burger with lettuce a mayo to the right of the photo\"."}], "task": "refering", "answer_template": "The \"half of a burger with lettuce a mayo to the right of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 84, 85, 86, 87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 344], "bbox": [0.469796875, 0.16129166666666667, 1.0, 0.841375], "iscrowd": 0, "area": 85506.45885000002, "rle": {"size": [480, 640], "counts": "[Z]4;]><D<D<D<D<D<D<D<D<L3L5L4K5L4K5L4L4K5L4K5L3L100O2O0O101N100O2O0O2O000O2O0O101N100O2O0O101N100O2O0O2O0O101N100O2O0O101N100O2O000O2O0O101N101N100O2O0O101N100O2O0O101N100O101N100O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2O0O100O2N100O100O2O0O100O2O0O100O2O0O100O2N100O101N100O101N100O101N100O101N1O100O101N100O101O000O101O0O10001O0O10001N1000001N1000001N10000O2O00000O101O000O101O0O10001O0O10001O0O10001N11O00010O00000010O0001O0001O01O000010O0001O0000010O00001O01O01O0000010O00001O0001O01O000010O0001O0001O01O00001O01O01O000010O01O001O01O01O001O0010O0001O0010O0001O0010O01O00001O010O001O00010O001O0010O0001O001O010O00001O010O001O00010O001O1O00100O001O1O010O1O001O10O01O1O001O10jK"}, "label": "half of a burger with lettuce a mayo to the right of the photo"}]}
{"id": 327234, "image": "COCO_train2014_000000327234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man surfing on a green surfboard\"."}], "task": "refering", "answer_template": "The \"a man surfing on a green surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267, 268, 269, 270, 292], "bbox": [0.58703125, 0.40247916666666667, 0.7666718749999999, 0.7160416666666667], "iscrowd": 0, "area": 11232.176950000001, "rle": {"size": [480, 640], "counts": "X``5;`>`0@<E6L5L3M4L3M4L3M3M3M3M3L4L4L4L4L4L4L3M2N2N3L3N2N2N3M2N2O1O2N1O1O1O2N1O1O1O2N1O1O10001O0001O1O00100O001O00100O001O00100O001O00100O1O3M3M3N2M2N1O2N1POhDWNY;g1jDVNW;i1lDTNV;j1mDSNT;l1oDQNS;n1oDPNQ;P2RElMP;R2SEiMP;W2REeMQ;Z2k0O2O2N13MO1O1O1N2O1O1O1O1N3N1O1O1N2O1O1N2O1N2N4L3L5L4K4L5B=DbaU2"}, "label": "a man surfing on a green surfboard"}]}
{"id": 327234, "image": "COCO_train2014_000000327234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a green surfboard in the water\"."}], "task": "refering", "answer_template": "The \"a man with a green surfboard in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 267, 268, 269, 270, 292], "bbox": [0.58703125, 0.40247916666666667, 0.7666718749999999, 0.7160416666666667], "iscrowd": 0, "area": 11232.176950000001, "rle": {"size": [480, 640], "counts": "X``5;`>`0@<E6L5L3M4L3M4L3M3M3M3M3L4L4L4L4L4L4L3M2N2N3L3N2N2N3M2N2O1O2N1O1O1O2N1O1O1O2N1O1O10001O0001O1O00100O001O00100O001O00100O001O00100O1O3M3M3N2M2N1O2N1POhDWNY;g1jDVNW;i1lDTNV;j1mDSNT;l1oDQNS;n1oDPNQ;P2RElMP;R2SEiMP;W2REeMQ;Z2k0O2O2N13MO1O1O1N2O1O1O1O1N3N1O1O1N2O1O1N2O1N2N4L3L5L4K4L5B=DbaU2"}, "label": "a man with a green surfboard in the water"}]}
{"id": 310851, "image": "COCO_train2014_000000310851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red london bridge bus number 521\"."}], "task": "refering", "answer_template": "The \"red london bridge bus number 521\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274], "bbox": [0.529796875, 0.3032291666666667, 0.9857031249999999, 0.6768333333333333], "iscrowd": 0, "area": 40426.4344, "rle": {"size": [480, 640], "counts": "^Uo4`0^<S2N101N101N1O2O0O2O0O2O0O3M3N2M3N2M4M0O1O010O100O010O1O100O100O100O2O0O1O100O100O2O0O1O100O100O101N1O100O100O102M2N2O2M2O1N3N1N2N3N1N100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O10000001O00000000000000000000001N10N2O1O1O100000000000000000O1000000000000000001O00000O100000000000000000000000000O1000001O0000000000000000000O1001O001O1O001O0010O01O001O001O001OO10000O1000000O10000O10001N10000O10000O100000000000000O10000000001O00000000000000000000000O10000000000000001O00000000000000000O2O1O1O2N1O1O1O2jKYGo1g8PNV4M3N2O001O100O001O1O000001O01O000001O5K3N2M3M3M2N1OO101O3M4L3M4K5HUQ4"}, "label": "red london bridge bus number 521"}]}
{"id": 310851, "image": "COCO_train2014_000000310851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red bus 521\"."}], "task": "refering", "answer_template": "The \"red bus 521\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274], "bbox": [0.529796875, 0.3032291666666667, 0.9857031249999999, 0.6768333333333333], "iscrowd": 0, "area": 40426.4344, "rle": {"size": [480, 640], "counts": "^Uo4`0^<S2N101N101N1O2O0O2O0O2O0O3M3N2M3N2M4M0O1O010O100O010O1O100O100O100O2O0O1O100O100O2O0O1O100O100O101N1O100O100O102M2N2O2M2O1N3N1N2N3N1N100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O10000001O00000000000000000000001N10N2O1O1O100000000000000000O1000000000000000001O00000O100000000000000000000000000O1000001O0000000000000000000O1001O001O1O001O0010O01O001O001O001OO10000O1000000O10000O10001N10000O10000O100000000000000O10000000001O00000000000000000000000O10000000000000001O00000000000000000O2O1O1O2N1O1O1O2jKYGo1g8PNV4M3N2O001O100O001O1O000001O01O000001O5K3N2M3M3M2N1OO101O3M4L3M4K5HUQ4"}, "label": "red bus 521"}]}
{"id": 556617, "image": "COCO_train2014_000000556617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table where a man is working\"."}], "task": "refering", "answer_template": "The \"a table where a man is working\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 218, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 284, 288, 289, 307, 311, 334], "bbox": [0.304359375, 0.6, 0.60721875, 0.9438173302107729], "iscrowd": 0, "area": 8222.982700000002, "rle": {"size": [427, 640], "counts": "cca21Y=2N1O1O2N1O2N1O1O2N1O1O2N1N3N1O1O2N1O2N1O1O2N1O1Ob0^O;E6J6J6J6J6JaNjE0P:1UFNf93`FF_9<fF_OY9c0lFVOU9j0PGQOP9o0VGjNl8U1YGeNh8[1^G_Nb8a1e10O000001O01O02N1O2N2O0O2N2N1O2N2O0O001O0010O01O000010O01O001O0010N100000001O0000000O2O00000000001O000O10001O000000001M2O1N2O1N2O2M2O1N2O1N3M2O1N2O1N3N1N2O1N2O1N3N1N22N2N101N2N2TNREW1o:eNVEY1l:bNZE\\1\\;N2N3M2N3M100O1O001O010O1O001O10On0SOn0QO01O1O0O2O001O1VNi101O0000O10000000O10000000000000000O10000000000000000N2M3M3M3M3M3M3MjhX3"}, "label": "a table where a man is working"}]}
{"id": 556617, "image": "COCO_train2014_000000556617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white table at which man is sitting\"."}], "task": "refering", "answer_template": "The \"white table at which man is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 218, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 266, 284, 288, 289, 307, 311, 334], "bbox": [0.304359375, 0.6, 0.60721875, 0.9438173302107729], "iscrowd": 0, "area": 8222.982700000002, "rle": {"size": [427, 640], "counts": "cca21Y=2N1O1O2N1O2N1O1O2N1O1O2N1N3N1O1O2N1O2N1O1O2N1O1Ob0^O;E6J6J6J6J6JaNjE0P:1UFNf93`FF_9<fF_OY9c0lFVOU9j0PGQOP9o0VGjNl8U1YGeNh8[1^G_Nb8a1e10O000001O01O02N1O2N2O0O2N2N1O2N2O0O001O0010O01O000010O01O001O0010N100000001O0000000O2O00000000001O000O10001O000000001M2O1N2O1N2O2M2O1N2O1N3M2O1N2O1N3N1N2O1N2O1N3N1N22N2N101N2N2TNREW1o:eNVEY1l:bNZE\\1\\;N2N3M2N3M100O1O001O010O1O001O10On0SOn0QO01O1O0O2O001O1VNi101O0000O10000000O10000000000000000O10000000000000000N2M3M3M3M3M3M3MjhX3"}, "label": "white table at which man is sitting"}]}
{"id": 556617, "image": "COCO_train2014_000000556617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table on right\"."}], "task": "refering", "answer_template": "The \"table on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 247, 250, 266, 267, 268, 269, 270, 271, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 336, 340, 341, 342], "bbox": [0.5945, 0.6718501170960187, 0.937484375, 1.0], "iscrowd": 0, "area": 12224.465999999999, "rle": {"size": [427, 640], "counts": "[in41W=8H3L4M4L3M3M3M4L3L4O1O2N1O1O1O2M2K5N2O2M2N2OSNjDg1_;O2MfEbNi8\\1iFROZ9k0XFDk99fE4^:Z14L1N3N2N2O001N2O1O0O2O1N101O1N101O1N2PNmDf1d;H1N2O1N2O1O0O2O1O0O101O0O10000O10O010O10O0100O010N2K401N1H900O010O1O010O10O0100O010O010O10O0100O010O101N1O2N2J5J7K5L310O10O01O13L5L4L2M2O000O1000000O01O10O01O1O010OdNlDb0S;VOVEi0k:oN]EQ1b:gNgEY1V;O01O10O01O1O0O2O1N1M4M3O01000O010000O01000O0100O001O1O001N2O001O1N101O1O001K5L3001O1O1O0000000NROnMhFm1Z9VNeFe1^9]NbF_1a9bN_FZ1e8`NZG71T1j8fNSG73o0n8kNmF75j0R9POgF77d0W9VO`F79?[9[OZF8:9`9AQF9?6`9DlE9c03a9b0^F_Ob9b0]F^Oc9b0[F@f9`0WFBl9<QFFS:7jEKZ:2cE1`:K^E7g:EWE<m:CoD>U;j04M4K4L4L4L5L3L4L4L5K4M3K6H7ISZ`0"}, "label": "table on right"}]}
{"id": 556617, "image": "COCO_train2014_000000556617.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right side of two tables setup\"."}], "task": "refering", "answer_template": "The \"the right side of two tables setup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 247, 250, 266, 267, 268, 269, 270, 271, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 336, 340, 341, 342], "bbox": [0.5945, 0.6718501170960187, 0.937484375, 1.0], "iscrowd": 0, "area": 12224.465999999999, "rle": {"size": [427, 640], "counts": "[in41W=8H3L4M4L3M3M3M4L3L4O1O2N1O1O1O2M2K5N2O2M2N2OSNjDg1_;O2MfEbNi8\\1iFROZ9k0XFDk99fE4^:Z14L1N3N2N2O001N2O1O0O2O1N101O1N101O1N2PNmDf1d;H1N2O1N2O1O0O2O1O0O101O0O10000O10O010O10O0100O010N2K401N1H900O010O1O010O10O0100O010O010O10O0100O010O101N1O2N2J5J7K5L310O10O01O13L5L4L2M2O000O1000000O01O10O01O1O010OdNlDb0S;VOVEi0k:oN]EQ1b:gNgEY1V;O01O10O01O1O0O2O1N1M4M3O01000O010000O01000O0100O001O1O001N2O001O1N101O1O001K5L3001O1O1O0000000NROnMhFm1Z9VNeFe1^9]NbF_1a9bN_FZ1e8`NZG71T1j8fNSG73o0n8kNmF75j0R9POgF77d0W9VO`F79?[9[OZF8:9`9AQF9?6`9DlE9c03a9b0^F_Ob9b0]F^Oc9b0[F@f9`0WFBl9<QFFS:7jEKZ:2cE1`:K^E7g:EWE<m:CoD>U;j04M4K4L4L4L5L3L4L4L5K4M3K6H7ISZ`0"}, "label": "the right side of two tables setup"}]}
{"id": 253513, "image": "COCO_train2014_000000253513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"computer screen of a news headline\"."}], "task": "refering", "answer_template": "The \"computer screen of a news headline\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241], "bbox": [0.19790625, 0.42262295081967216, 0.49176562500000004, 0.7259953161592506], "iscrowd": 0, "area": 18932.183600000008, "rle": {"size": [427, 640], "counts": "^Te1d0g<3M001O00001O001O001O001O00001O001O001O1O001O001O1O0O2O1O001O001O1O001O001O2N2N2N3M2N3M2N3M3M4L3M4L4L3M4L4L0000001O0bFUM\\8k2`GYM`8g2]G\\Mc8e2XG_Mh8a2UGbMk8^2QGfMo8Z2nFiMR9X3O10000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000O1000cN]1oMT]W4"}, "label": "computer screen of a news headline"}]}
{"id": 253513, "image": "COCO_train2014_000000253513.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a portable monitor\"."}], "task": "refering", "answer_template": "The \"a portable monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241], "bbox": [0.19790625, 0.42262295081967216, 0.49176562500000004, 0.7259953161592506], "iscrowd": 0, "area": 18932.183600000008, "rle": {"size": [427, 640], "counts": "^Te1d0g<3M001O00001O001O001O001O00001O001O001O1O001O001O1O0O2O1O001O001O1O001O001O2N2N2N3M2N3M2N3M3M4L3M4L4L3M4L4L0000001O0bFUM\\8k2`GYM`8g2]G\\Mc8e2XG_Mh8a2UGbMk8^2QGfMo8Z2nFiMR9X3O10000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000O1000cN]1oMT]W4"}, "label": "a portable monitor"}]}
{"id": 474699, "image": "COCO_train2014_000000474699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cooked chicken\"."}], "task": "refering", "answer_template": "The \"cooked chicken\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 147, 148, 149], "bbox": [0.246078125, 0.08311475409836067, 0.7434375, 0.41557377049180333], "iscrowd": 0, "area": 31218.1116, "rle": {"size": [427, 640], "counts": "hQR22X=1N1O2O1N2N101N2N1O2O1`NBQF`0m9EnE<R:HjE9U:KfE7Y:MbE4^:0]E2b:2ZEOe:5VEMi:7REJn::nDGQ;Z1O0O2O1O1O1O1O1O1O1O1O1O1AaMPFa2o9>00O10000O10000O100O10000O10000O100O1O1O100O1O100O2N100O1O1O100O1O100O100O2O0000000O100000001N100000000O1000001N100000000O10001O000O1000000001O01O00000000000O10001O0O10000O1000000O2O000O1000000O10000O1000000O02O001O001O001N101O00001O001O0O2O001O001O000O2O001OO100000O10000000O1000O1000000000O1000O1000000000O10O100000001O0000000O100000000000001O1N2O1O001O1O1O1O1O1N2O1O1O1O1O1O0000000O100O10000O100O100O100O100O100O100O100O1O1K5K5O1O1O1O1O1O1O2N1O1O2N1O101N101N100O2O000O2O0O101N100O2O0O101N100O2O000O2O0O101N101N100O2O0O101O2M3N1N3N2M2O2M3N1N3N2M2O2M3N1O2M3N1N3NQYT2"}, "label": "cooked chicken"}]}
{"id": 474699, "image": "COCO_train2014_000000474699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of fish with vegetables is on a table\"."}], "task": "refering", "answer_template": "The \"a piece of fish with vegetables is on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 147, 148, 149], "bbox": [0.246078125, 0.08311475409836067, 0.7434375, 0.41557377049180333], "iscrowd": 0, "area": 31218.1116, "rle": {"size": [427, 640], "counts": "hQR22X=1N1O2O1N2N101N2N1O2O1`NBQF`0m9EnE<R:HjE9U:KfE7Y:MbE4^:0]E2b:2ZEOe:5VEMi:7REJn::nDGQ;Z1O0O2O1O1O1O1O1O1O1O1O1O1AaMPFa2o9>00O10000O10000O100O10000O10000O100O1O1O100O1O100O2N100O1O1O100O1O100O100O2O0000000O100000001N100000000O1000001N100000000O10001O000O1000000001O01O00000000000O10001O0O10000O1000000O2O000O1000000O10000O1000000O02O001O001O001N101O00001O001O0O2O001O001O000O2O001OO100000O10000000O1000O1000000000O1000O1000000000O10O100000001O0000000O100000000000001O1N2O1O001O1O1O1O1O1N2O1O1O1O1O1O0000000O100O10000O100O100O100O100O100O100O100O1O1K5K5O1O1O1O1O1O1O2N1O1O2N1O101N101N100O2O000O2O0O101N100O2O0O101N100O2O000O2O0O101N101N100O2O0O101O2M3N1N3N2M2O2M3N1N3N2M2O2M3N1O2M3N1N3NQYT2"}, "label": "a piece of fish with vegetables is on a table"}]}
{"id": 196170, "image": "COCO_train2014_000000196170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the part of the table to the right of the man in a checkered shirt with the chopsticks\"."}], "task": "refering", "answer_template": "The \"the part of the table to the right of the man in a checkered shirt with the chopsticks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 215, 216, 219, 220, 221, 237, 238, 239, 240, 243, 244, 260, 261, 262, 264, 265, 266, 267, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 379, 380, 381, 382], "bbox": [0.302984375, 0.5081458333333333, 0.8856718750000001, 0.9956458333333332], "iscrowd": 0, "area": 27410.50105, "rle": {"size": [480, 640], "counts": "gWk22n>2M2N3M3M2O1N3M2N2O1N2N2N2O1N2N2N3M2O1N2N2N2N2O1N2N2N3M2N2N2N2O1N200N2N2N3M2N2O1N2N2N2N2N3M2O1N2O1N2O1N2O2M2O1N2O1N2O1N2O1NeMfDS1Y;lNjDS1U;lNoDR1o:nNTEQ1k:mNZEQ1d:oN_EP1`:oNcEP1[:POhEo0X:oNjEQ1V:mNmER1S:lNPFS1P:jNTFV1k9iNWFV1i9hNYFX1g9gN[FX1e9gN[FZ1e9dN]F\\1c9cN]F^1c9aN^F_1b9_N`Fa1`9^N`Fc1`9\\NaFd1^9[NcFf1]9YNcFh1\\9XNeFh1[9VNfFk1Y9UNgFl1Y9SNgFn1X9QNiFP2W9oMiFR2V9nMjFS2V9kMkFW2T9hMmFX2R9gMoFZ2Q9eMoF\\2P9dMPG]2i:1O1O1O100O1O1O1O1O1O1O100N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1a0_O00O1O1O1SOQMgEP3W:QMjEo2S:SMmEn2P:TMPFm2m9UMSFm2j9TMVFm2g9TM[Fl2b9VM^Fk2_9WMaFj2_9UMaFl2^9TMbFm2]9SMdFm2[9SMeFn2Z9RMfFo2Y9QMgFP3X9oLiFR3V9nLkFR3U9mLkFT3T9lLlFU3S9kLmFV3R9jLnFW3o8kLRGS3m8oLSGQ3m8oLSGP3m6bLRJ<ROR3k6eLRJ8SOS3k6hLPJ4VOR3j6XMcIFCR3j6ZMbICDR3j6\\McIACS3j6\\MdI@BS3j6^MdI_OBS3i6_MfI]OAS3j6`MfI[OAU3h6aMfI[OBS3i6cMcI[ODR3h6hLTHk0_1ZOFS3j6cM^I[OHQ3l6dMUHYOj016R3k6PNoHnN6Q3m6PNlHoN7R3m6oMlHoN7Q3o6nMjHQO8Q3o6mMiHQO8R3P7mMgHRO8R3R7lMeHQO:R3R7nMbHQO;R3T7mM`HPO<S3V7mM\\HQO=S3W7mM[HoN?T3W7mMYHnN`0U3X7nMVHnNa0U3Z7mMTHmNb0V3\\7lMQHoNc0U3\\7mMkGROi0Q3^7mMcGWOn0m2_7`NaH`1a7^N`Hb1a7]N^Hc1c7\\N]He1d7YN\\Hh1d7WN]Hh1e7WNZHj1f7UNZHk1h7SNXHn1h7QNYHn1i7PNWHQ2i7nMWHR2k7mMUHS2k7lMUHT2m7kMRHV2n7kMPHU2R8jMnGV2R8kMTHm1n7RN]Hc1c7^NgHV1Z7kNoHk0R7TOYI`0h6A_I7b6H^I7c6J\\I5f6KYI5g6KYI4j6JWI5j6KUI4n6JRI6o6IQI6R7ImH7U7]OVF@e2R1W7[O[F_O]2V1Z7XO^F^OX2Y1_7VOnHi0\\7PObHP1h:N100O100O100O100O101N1O100O10L3I8H7I8L3100O010O00100O01O0100O5L3L4M3L100O2N100O101N100O101N100O101N1O100O2OO010O010O010O0CVOfBj0Z==1O10O0100O10SOmNmCR1n;TOQDm0i;YOWDf0f;^OZDb0g;\\OZDc0g;]OYDc0h;\\OWDd0j;[OWDe0j;ZOVDe0k;[OUDe0l;YOTDg0m;YOSDg0n;XORDg0P<WOQDi0o;WOPDi0R<VOnCj0R<VOmCj0U<VOjCj0V<VOiCj0Y<VOfCj0Z<WOdCi0^<VObCi0`<WO^Cj0b<VO]Cj0e<VOZCj0f<VOYCj0i<VOVCj0j<WOTCi0n<VORCj0n<WOPCi0R=WOmBi0S=XOkBh0W=XOgBi0Z=WOeBh0\\=YObBh0_=;3N1N2O2M2O1N2O1N101N2N2O0O2O1N101N100O100O101N100O100O101N100O100O101N100O1O100O2O0O2OgkQ1"}, "label": "the part of the table to the right of the man in a checkered shirt with the chopsticks"}]}
{"id": 196170, "image": "COCO_train2014_000000196170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown table top\"."}], "task": "refering", "answer_template": "The \"a brown table top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 215, 216, 219, 220, 221, 237, 238, 239, 240, 243, 244, 260, 261, 262, 264, 265, 266, 267, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 293, 294, 295, 296, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 379, 380, 381, 382], "bbox": [0.302984375, 0.5081458333333333, 0.8856718750000001, 0.9956458333333332], "iscrowd": 0, "area": 27410.50105, "rle": {"size": [480, 640], "counts": "gWk22n>2M2N3M3M2O1N3M2N2O1N2N2N2O1N2N2N3M2O1N2N2N2N2O1N2N2N3M2N2N2N2O1N200N2N2N3M2N2O1N2N2N2N2N3M2O1N2O1N2O1N2O2M2O1N2O1N2O1N2O1NeMfDS1Y;lNjDS1U;lNoDR1o:nNTEQ1k:mNZEQ1d:oN_EP1`:oNcEP1[:POhEo0X:oNjEQ1V:mNmER1S:lNPFS1P:jNTFV1k9iNWFV1i9hNYFX1g9gN[FX1e9gN[FZ1e9dN]F\\1c9cN]F^1c9aN^F_1b9_N`Fa1`9^N`Fc1`9\\NaFd1^9[NcFf1]9YNcFh1\\9XNeFh1[9VNfFk1Y9UNgFl1Y9SNgFn1X9QNiFP2W9oMiFR2V9nMjFS2V9kMkFW2T9hMmFX2R9gMoFZ2Q9eMoF\\2P9dMPG]2i:1O1O1O100O1O1O1O1O1O1O100N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1a0_O00O1O1O1SOQMgEP3W:QMjEo2S:SMmEn2P:TMPFm2m9UMSFm2j9TMVFm2g9TM[Fl2b9VM^Fk2_9WMaFj2_9UMaFl2^9TMbFm2]9SMdFm2[9SMeFn2Z9RMfFo2Y9QMgFP3X9oLiFR3V9nLkFR3U9mLkFT3T9lLlFU3S9kLmFV3R9jLnFW3o8kLRGS3m8oLSGQ3m8oLSGP3m6bLRJ<ROR3k6eLRJ8SOS3k6hLPJ4VOR3j6XMcIFCR3j6ZMbICDR3j6\\McIACS3j6\\MdI@BS3j6^MdI_OBS3i6_MfI]OAS3j6`MfI[OAU3h6aMfI[OBS3i6cMcI[ODR3h6hLTHk0_1ZOFS3j6cM^I[OHQ3l6dMUHYOj016R3k6PNoHnN6Q3m6PNlHoN7R3m6oMlHoN7Q3o6nMjHQO8Q3o6mMiHQO8R3P7mMgHRO8R3R7lMeHQO:R3R7nMbHQO;R3T7mM`HPO<S3V7mM\\HQO=S3W7mM[HoN?T3W7mMYHnN`0U3X7nMVHnNa0U3Z7mMTHmNb0V3\\7lMQHoNc0U3\\7mMkGROi0Q3^7mMcGWOn0m2_7`NaH`1a7^N`Hb1a7]N^Hc1c7\\N]He1d7YN\\Hh1d7WN]Hh1e7WNZHj1f7UNZHk1h7SNXHn1h7QNYHn1i7PNWHQ2i7nMWHR2k7mMUHS2k7lMUHT2m7kMRHV2n7kMPHU2R8jMnGV2R8kMTHm1n7RN]Hc1c7^NgHV1Z7kNoHk0R7TOYI`0h6A_I7b6H^I7c6J\\I5f6KYI5g6KYI4j6JWI5j6KUI4n6JRI6o6IQI6R7ImH7U7]OVF@e2R1W7[O[F_O]2V1Z7XO^F^OX2Y1_7VOnHi0\\7PObHP1h:N100O100O100O100O101N1O100O10L3I8H7I8L3100O010O00100O01O0100O5L3L4M3L100O2N100O101N100O101N100O101N1O100O2OO010O010O010O0CVOfBj0Z==1O10O0100O10SOmNmCR1n;TOQDm0i;YOWDf0f;^OZDb0g;\\OZDc0g;]OYDc0h;\\OWDd0j;[OWDe0j;ZOVDe0k;[OUDe0l;YOTDg0m;YOSDg0n;XORDg0P<WOQDi0o;WOPDi0R<VOnCj0R<VOmCj0U<VOjCj0V<VOiCj0Y<VOfCj0Z<WOdCi0^<VObCi0`<WO^Cj0b<VO]Cj0e<VOZCj0f<VOYCj0i<VOVCj0j<WOTCi0n<VORCj0n<WOPCi0R=WOmBi0S=XOkBh0W=XOgBi0Z=WOeBh0\\=YObBh0_=;3N1N2O2M2O1N2O1N101N2N2O0O2O1N101N100O100O101N100O100O101N100O100O101N100O1O100O2O0O2OgkQ1"}, "label": "a brown table top"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse wearing a mask getting rode by a jockey\"."}], "task": "refering", "answer_template": "The \"a brown horse wearing a mask getting rode by a jockey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 181, 182, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 284, 286, 287, 289, 290, 291, 292, 293, 307, 310, 311, 312, 313, 315, 316, 330, 333, 334, 335, 338, 339], "bbox": [0.355828125, 0.18925058548009369, 0.9355, 0.9892505854800937], "iscrowd": 0, "area": 42005.292400000006, "rle": {"size": [427, 640], "counts": "Q]o22Y=4M3L4M3L8bDYOi9S1hEROU:R1dESO[:P1\\EVOc:i12N2N2O2M2N2N4M7H8H8I7H8H8H\\id0MfV[O3M2M4M3M2M4M3M2M4M3M3M2H9XOh0L4K4M4L4L3L5L4LfNSF\\Oh9e0_FWO^9j0fFUOU9l0oFVOk8j0YGYO_8i0eGYOT8g0PH\\Oh7e0\\HYOa7h0dHUOZ7l0iHROT7Q1nHmNP7T1SIjNk6X1XIeNf6\\1]IbNa6`1aI^N\\6d1hIYNV6i1lIVNQ6l1QJRNm5o1VJoMh5S2ZJkMd5V2_JhM_5Z2cJdM[5^2fJaMX5`2kJ^MS5c2PKZMo4h2SKQMo4Q3d21O1O1M4K4M3UI_Lj3d3SL`Lh3e3SL`Lj3b3RLcLo3[3nKiLS4U3iKQMV4o2fKUM[4i2cKZM_4d2_K^Mc4`2[KbMg4\\2WKfMl4W2RKlMo4R2oJPNS5n1lJSNU5l1iJVNX5i1eJZN]5e1]J`Nd5`1WJdNk5Z1PJlNP6T1kIPOW6o0eITO\\6k0bIWO`6h0]IZOd6e0YI^Oh6b0UI\\Oo6e0nHYOV7g0gHYO\\7h0aHVOc7k0ZHSOj7m0SHROQ8o25L4K5O1O1N2O1O1O1O1O\\N_K_Ja4b5aK[J^4g5cKWJ\\4k5fKQJZ4Q6hKlIX4U6iKiIV4Y6lKcIT4_6mK_IR4c6PLZIo3h6RLVIm3l6ULRIi3P7YLmHg3T7ZLjHe3X7]LZHm3h7k00O10000O10000O1000000O10000O1000000O10000O1000000O1000O2O2N2M3M3M3M2O2M2N2N2N2N2N4K4cJfId3^6YLbIe3b6XL_Ig3e6VL[Ih3i6VLWIi3m6SLTIm3o6oKRIQ4P7lKQIT4Q7iKPIW4S7eKnH[4T7bKmH^4X7\\KiHd4]7UKdHk4_7QKbH`3FjMW8V2oGdMQ8[2VH_Mj7a2]HWMd7h2aHSM`7m2`HRMa7n2_HPMc7o2^HPMc7P3]HoLd7P3]HoLd7Q3[HoLg4FXM[3QNnLg4LTMU3VNnLf42PMP3ZNmLf48lLj2\\4eMVK[2i4_N^Ja1a5bN]J^1b5fN[JY1e5jNYJV1f5nNUJT1j5oNSJR1m5QOnIP1R6SOkIn0T6VOgIl0X6WOdIk0[6YOaIg0_6S3O1L3`NZIULn6i3fI^Ka6`4V1N1O2NN2N2O1002N3M3M3M2N3M3N2M3M2N3M3L4K4M4K5K4M30O1000000O1N2J6J6O11O1O1O1O1O1O001O1O1O1O1O1O101N2O1N2O1N2O1O1N2O0O100O1000O0100O100O100O10000O2O3L4M3L4M3M1N101N1N3N2N1O2N1N3N1O2N1O2O1N1O2N101M2\\OmCMV<NoCNT<MSDNQ<MUDOWYa0"}, "label": "a brown horse wearing a mask getting rode by a jockey"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a jockey racing a horse on the outside of the pack\"."}], "task": "refering", "answer_template": "The \"a jockey racing a horse on the outside of the pack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 116, 117, 118, 121, 139, 140, 141, 162, 163, 184, 185], "bbox": [0.026921875, 0.1677049180327869, 0.30965624999999997, 0.6094145199063232], "iscrowd": 0, "area": 13090.120300000002, "rle": {"size": [427, 640], "counts": "UZ71X=3N2N2N2M3N3L6K2M3N2M2N3N1O2N1O3M2N3M3M2jM^N`Hc1\\7dN`H^1\\7iN_HX1]7oN_HR1\\7VO`Hj0V7BhH>V7FhH;V7GiH9V7JhH6W7MgH3Y7NfH3Y7OdH2\\70bHMa74]HSOBTOT8l1XHoN_8S1_GmN`8V1]GkNb8X1\\GiNc8Y1[GgNd8[1[GfNd8\\1ZGdNf8]1YGcNh8^1WGbNh8_1WGaNi8a1UG`Nk8`1TG`Nl8b1RG^No8b1PG_NP9b1nF^NR9c1mF^NS9c1kF]NV9c1iF^NV9d1hF\\NY9k20O10O010O010O0100O010O10O0100O10O10O1WNcFD]9:fFD[9:hFEX9:jFDW9:lFET99oFER99QGFo89SGEm8<TGCl8>TG@m8`0TG]On8c0SG[On8f0RGXOo8h0RGVOo8j0RGSOP9m0QGQOo8Q1QGmNP9S1QGjNQ9V1PGhNQ9Y1oFcNS9^1PG]NR9d1oFXNS9i1oFQNT9o1S12OnNiD1V;OlDOT;2mDLS;5mDJS;6mDJS;7mDIR;8mDHS;8mDHS;9lDGT;9mDFS;:mDGS;9lDGT;9lDGT;9mDFS;;lDET;;lDET;;mDDS;=lDCT;=lDCT;=lDCU;=kDBU;>kDBU;>kDBU;?jDAW;>jDAV;?jDAW;>iDBW;?iD@X;?hDAX;?hDAX;`0gD@Z;?gD@Y;`0gD@Y;a0fD_OY;b0gD^OY;b0hD]OX;d0gD\\OX;e0hD\\OW;d0iD\\OV;e0kDZOU;f0kDZOU;f0kDZOU;f0lDYOT;g0lDZOT;e0lD[OU;d0kD\\OU;d0lD[OU;d0kD]OU;b0kD^OV;a0kD^OV;a0jD_OV;a0jD@V;?kD@V;?jDAW;>iDBX;=iDCW;<iDDW;<iDEW;9kDFV;9kDGU;8kDHV;6kDJU;4nDLR;1PEOV;IlD7T<0001O00001O000O101O0O10001O1N2O1O1NfUh5"}, "label": "a jockey racing a horse on the outside of the pack"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with blue hat that has a green cover on his horse\"."}], "task": "refering", "answer_template": "The \"man with blue hat that has a green cover on his horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 116, 117, 118, 121, 139, 140, 141, 162, 163, 184, 185], "bbox": [0.026921875, 0.1677049180327869, 0.30965624999999997, 0.6094145199063232], "iscrowd": 0, "area": 13090.120300000002, "rle": {"size": [427, 640], "counts": "UZ71X=3N2N2N2M3N3L6K2M3N2M2N3N1O2N1O3M2N3M3M2jM^N`Hc1\\7dN`H^1\\7iN_HX1]7oN_HR1\\7VO`Hj0V7BhH>V7FhH;V7GiH9V7JhH6W7MgH3Y7NfH3Y7OdH2\\70bHMa74]HSOBTOT8l1XHoN_8S1_GmN`8V1]GkNb8X1\\GiNc8Y1[GgNd8[1[GfNd8\\1ZGdNf8]1YGcNh8^1WGbNh8_1WGaNi8a1UG`Nk8`1TG`Nl8b1RG^No8b1PG_NP9b1nF^NR9c1mF^NS9c1kF]NV9c1iF^NV9d1hF\\NY9k20O10O010O010O0100O010O10O0100O10O10O1WNcFD]9:fFD[9:hFEX9:jFDW9:lFET99oFER99QGFo89SGEm8<TGCl8>TG@m8`0TG]On8c0SG[On8f0RGXOo8h0RGVOo8j0RGSOP9m0QGQOo8Q1QGmNP9S1QGjNQ9V1PGhNQ9Y1oFcNS9^1PG]NR9d1oFXNS9i1oFQNT9o1S12OnNiD1V;OlDOT;2mDLS;5mDJS;6mDJS;7mDIR;8mDHS;8mDHS;9lDGT;9mDFS;:mDGS;9lDGT;9lDGT;9mDFS;;lDET;;lDET;;mDDS;=lDCT;=lDCT;=lDCU;=kDBU;>kDBU;>kDBU;?jDAW;>jDAV;?jDAW;>iDBW;?iD@X;?hDAX;?hDAX;`0gD@Z;?gD@Y;`0gD@Y;a0fD_OY;b0gD^OY;b0hD]OX;d0gD\\OX;e0hD\\OW;d0iD\\OV;e0kDZOU;f0kDZOU;f0kDZOU;f0lDYOT;g0lDZOT;e0lD[OU;d0kD\\OU;d0lD[OU;d0kD]OU;b0kD^OV;a0kD^OV;a0jD_OV;a0jD@V;?kD@V;?jDAW;>iDBX;=iDCW;<iDDW;<iDEW;9kDFV;9kDGU;8kDHV;6kDJU;4nDLR;1PEOV;IlD7T<0001O00001O000O101O0O10001O1N2O1O1NfUh5"}, "label": "man with blue hat that has a green cover on his horse"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse wearing green colour cloth\"."}], "task": "refering", "answer_template": "The \"the horse wearing green colour cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 119, 120, 121, 122, 123, 124, 138, 139, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 216, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 305, 306, 328, 329], "bbox": [0.0, 0.2672599531615925, 0.439140625, 1.0], "iscrowd": 0, "area": 43121.3824, "rle": {"size": [427, 640], "counts": "i5l4_8N2O1O001O1O1N2O001N2N2O1O100O100O1TMSGZ1m8eNWGX1i8fN\\GW1d8hN^GW1b8gNaGX1_8fNdGY1\\8fNfGY1Z8eNiGZ1W8dNkG]1T8bNnG^1Q8`NRH_1n7_NTHb1j7]NYHc1f7\\N\\Hc1d7\\N]He1b7[N_He1_7[NbHe1^7ZNdHf1[7YNgHg1X7XNiHh1W7XNjHh1U7WNlHj1S7UNoHj1Q7UNQIk1n6TN[HBEZ2P8RN[HJBT2R8QN]HM_OR2T8PN]H0^OP2U8oM^H3ZOP2W8lM_H7WOn1Z8jM_H<ROm1^8fMaHT3_7kLaHV3_7iLbHX3]7gLeHX3[7gLgHX3Y7gLhHZ3W7dLlH[3T7dLnH[3R7dLoH\\3P7dLRI[3m6eLUI[3j6dLWI\\3h6dLZI[3e6eL\\I[3c6eL\\I]3c6cL[I`3d6`LZId3d6\\LZIg3S82N2N2L5M2O1N200O1O2O0O1O100O1O2O0O1O100O1O2O0O1O100O2N1O1O1O1O1O2N1O1O1O1O1O1O1M31N100010O010O0100O01N101N100O2N101L3N3K4N3N1O1O2N101N1O2O0O101O001O001O000010OO2N1O2N101N1O1O1O1O2O0O2O0O2O3L2O1O1N3N2N1O2M3N1N3M3N1O2O1N3N1O1N1O000O100O1O4M3M3L3N2N2N2N2N2N2M4M^JhKR2]3eJUMU3EXND[2c3ZLQMQ3MTNH[2Z3hLiLf2;oML^2Q39YMmL3e2e2;UO@k0>\\O]Oe0?CUM^Mk0o2n1KhLbMW1d2n1Y1nMgNP2^1mMcNQ2a1lM`NR2f1iMZNV2j1gMWNW2m1fMSNY2Q2dMPNY2V2cM_Mc2h2[MYMe2h2YMYMh2g2VMYMk2h2SMYMm2h2QMYMP3f2oL[MR3d2mL]MU3b2iL_MW3[61O00O2O0O1000O001O1O0O2O1O001O2O1N2N3M2N3M2O3L2N10O010O01FTLTGk3j8;01O100001O2N3M4L4L4L6J7H3M3M4L3L6J3M3nM\\EY1h:cN[EX1k:bNZEb0d;XO`Df0U<L3M3M3M4L5KW`e4"}, "label": "the horse wearing green colour cloth"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse on the left\"."}], "task": "refering", "answer_template": "The \"a brown horse on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 119, 120, 121, 122, 123, 124, 138, 139, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 216, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 305, 306, 328, 329], "bbox": [0.0, 0.2672599531615925, 0.439140625, 1.0], "iscrowd": 0, "area": 43121.3824, "rle": {"size": [427, 640], "counts": "i5l4_8N2O1O001O1O1N2O001N2N2O1O100O100O1TMSGZ1m8eNWGX1i8fN\\GW1d8hN^GW1b8gNaGX1_8fNdGY1\\8fNfGY1Z8eNiGZ1W8dNkG]1T8bNnG^1Q8`NRH_1n7_NTHb1j7]NYHc1f7\\N\\Hc1d7\\N]He1b7[N_He1_7[NbHe1^7ZNdHf1[7YNgHg1X7XNiHh1W7XNjHh1U7WNlHj1S7UNoHj1Q7UNQIk1n6TN[HBEZ2P8RN[HJBT2R8QN]HM_OR2T8PN]H0^OP2U8oM^H3ZOP2W8lM_H7WOn1Z8jM_H<ROm1^8fMaHT3_7kLaHV3_7iLbHX3]7gLeHX3[7gLgHX3Y7gLhHZ3W7dLlH[3T7dLnH[3R7dLoH\\3P7dLRI[3m6eLUI[3j6dLWI\\3h6dLZI[3e6eL\\I[3c6eL\\I]3c6cL[I`3d6`LZId3d6\\LZIg3S82N2N2L5M2O1N200O1O2O0O1O100O1O2O0O1O100O1O2O0O1O100O2N1O1O1O1O1O2N1O1O1O1O1O1O1M31N100010O010O0100O01N101N100O2N101L3N3K4N3N1O1O2N101N1O2O0O101O001O001O000010OO2N1O2N101N1O1O1O1O2O0O2O0O2O3L2O1O1N3N2N1O2M3N1N3M3N1O2O1N3N1O1N1O000O100O1O4M3M3L3N2N2N2N2N2N2M4M^JhKR2]3eJUMU3EXND[2c3ZLQMQ3MTNH[2Z3hLiLf2;oML^2Q39YMmL3e2e2;UO@k0>\\O]Oe0?CUM^Mk0o2n1KhLbMW1d2n1Y1nMgNP2^1mMcNQ2a1lM`NR2f1iMZNV2j1gMWNW2m1fMSNY2Q2dMPNY2V2cM_Mc2h2[MYMe2h2YMYMh2g2VMYMk2h2SMYMm2h2QMYMP3f2oL[MR3d2mL]MU3b2iL_MW3[61O00O2O0O1000O001O1O0O2O1O001O2O1N2N3M2N3M2O3L2N10O010O01FTLTGk3j8;01O100001O2N3M4L4L4L6J7H3M3M4L3L6J3M3nM\\EY1h:cN[EX1k:bNZEb0d;XO`Df0U<L3M3M3M4L5KW`e4"}, "label": "a brown horse on the left"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"jockey in dark red\"."}], "task": "refering", "answer_template": "The \"jockey in dark red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 39, 40, 41, 60, 61, 62, 63, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 129, 130, 131, 132, 152, 173, 174, 196], "bbox": [0.546015625, 0.05646370023419204, 0.831421875, 0.5470960187353631], "iscrowd": 0, "area": 11901.877999999997, "rle": {"size": [427, 640], "counts": "Sha41X=2N2M4M2N2M4N12N1O10fLBeI>[6BfI=X6FhI:V6GkI8T6JlI5R6MoI2P60QJLP65SJEn5>^31O2M2O2N3L3O2M2N3N1N3N2M3M2O2M3M2O1N2O1N1O2O1N2O5Jd0]O1N2O2M2N2O1N10O01O001O001O000010O01O001O01O010O1O010O001000O10O010000O10O1000O10000O01000O1000O010000O10O1000O1000O0lNoEPOQ:P1PFoNP:R1PFmNo9T1RFkNn9U1SFiNn9W1SFhNl9Y1UFfNk9Z1VFeNi9]1WFbNi9^1XFaNg9`1ZF_Nf9a1[F^Nd9c1]F\\Nc9d1^F[Na9f1`FYN`9g1aFXN^9h1dFWN\\9i1eFVNZ9j1gFUNZ9k1fFSN[9m1fFRN[9n1eFQN[9P2fFmM\\9R2eFmM[9T2eFkM\\9T2eFjM\\9W2dFhM]9X2j0010O0010O0100O`NVE`0k:@VE?i:AZE<g:DZE;e:E]E7f:I]E1e:O_ELc:4aEEa:;cE_O`:a0cEYO_:g0eESO^:m0P1O010O10O0100O010O10O0100O010O10O0100O011N101M2N3N2M2N3N2M2N3N1NWP]1"}, "label": "jockey in dark red"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse jockey wearing a burgundy - blue shirt and riding a horse during a competition\"."}], "task": "refering", "answer_template": "The \"horse jockey wearing a burgundy - blue shirt and riding a horse during a competition\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 39, 40, 41, 60, 61, 62, 63, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 129, 130, 131, 132, 152, 173, 174, 196], "bbox": [0.546015625, 0.05646370023419204, 0.831421875, 0.5470960187353631], "iscrowd": 0, "area": 11901.877999999997, "rle": {"size": [427, 640], "counts": "Sha41X=2N2M4M2N2M4N12N1O10fLBeI>[6BfI=X6FhI:V6GkI8T6JlI5R6MoI2P60QJLP65SJEn5>^31O2M2O2N3L3O2M2N3N1N3N2M3M2O2M3M2O1N2O1N1O2O1N2O5Jd0]O1N2O2M2N2O1N10O01O001O001O000010O01O001O01O010O1O010O001000O10O010000O10O1000O10000O01000O1000O010000O10O1000O1000O0lNoEPOQ:P1PFoNP:R1PFmNo9T1RFkNn9U1SFiNn9W1SFhNl9Y1UFfNk9Z1VFeNi9]1WFbNi9^1XFaNg9`1ZF_Nf9a1[F^Nd9c1]F\\Nc9d1^F[Na9f1`FYN`9g1aFXN^9h1dFWN\\9i1eFVNZ9j1gFUNZ9k1fFSN[9m1fFRN[9n1eFQN[9P2fFmM\\9R2eFmM[9T2eFkM\\9T2eFjM\\9W2dFhM]9X2j0010O0010O0100O`NVE`0k:@VE?i:AZE<g:DZE;e:E]E7f:I]E1e:O_ELc:4aEEa:;cE_O`:a0cEYO_:g0eESO^:m0P1O010O10O0100O010O10O0100O010O10O0100O011N101M2N3N2M2N3N2M2N3N1NWP]1"}, "label": "horse jockey wearing a burgundy - blue shirt and riding a horse during a competition"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the middle\"."}], "task": "refering", "answer_template": "The \"the horse in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 280, 281, 283, 284, 285, 286, 303, 304, 305, 308, 309, 327, 328, 331, 332], "bbox": [0.20421874999999998, 0.24412177985948477, 0.6397031249999999, 1.0], "iscrowd": 0, "area": 29289.65130000001, "rle": {"size": [427, 640], "counts": "Qof1T1V<4N2M1O100O100O1O101N1O100O100O1O100O2N10]N^D\\1a;dNaD[1_;eNcDY1\\;gNfDX1Y;hNjDV1V;jNkDOIm0[;TOYEi0g:WO\\Ef0c:[O`Ec0^:_ObE`0^:@cE?]:BbE>^:CbE;_:EaE:`:G_E8b:H_E6b:K]E4d:M\\E1e:O[E0g:0YENh:2XEMi:4VEKk:6UEHl:8TEGm::SEDn:<RECo:>QE@Q;`0o0N3N2M3N1O00O100000000000000000NdLMbI1Y67eIHW6?hI_OT6g0]2^OgDi0S;BaDf0X;k0J7H7J<B5L3000000000000001O1O2O0O1O1O1O1O1001O110O00O2O00000O10O100O1O2O0O2N1O2O0O2N1O2N100O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1N2nGTLb6n3[ITLc6o3ZISLd6R4VIQLh6T4SInKl6V4nHlKP7Z4kHhKQ7_4jHcKR7d4hH_Kn6n4mHTKl6i5N2XNaIcL`6]3bI_L`6a3bI[L`6e3bIdK1D_6h4bI]K:CV6o4`1N2O1N2N2N2N2N2000G`KgG`4T8fKkGZ4P8lKoGT4m7QLRHo3i7k0L5K5L4K5K5L3L5ZOf0O100O1O10O01O01O001OVOjH`KV7_4QI[Ko6d4XIWKg6h4`IRK`6m4gImJY6Q5R1N1O[OQKRIm4Q7SKmHl4U7VKhHi4[7WKdHg4^7[K_Hd4c7]K[Ha4i7`KTH_4n7=2N3M2N2N3M2N3L3N2N3M2N3M2N2N2M2O2N2N2N2N2N00N1O2N22N3M3M3CVFUMm9j2UFRMm9n29N2O1O1O1O1O1O0011O2M4M2N0O2O0O1100O1O100O1O1N2N2N2N1O1O1O1O001dM\\Eo1e:mMcEn1^:nMjEl1P;K4L2N1O2N1O2N1L5K5K4N3M2O2M2K6F^TP3"}, "label": "the horse in the middle"}]}
{"id": 245326, "image": "COCO_train2014_000000245326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the middle\"."}], "task": "refering", "answer_template": "The \"the horse in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 260, 261, 262, 263, 264, 280, 281, 283, 284, 285, 286, 303, 304, 305, 308, 309, 327, 328, 331, 332], "bbox": [0.20421874999999998, 0.24412177985948477, 0.6397031249999999, 1.0], "iscrowd": 0, "area": 29289.65130000001, "rle": {"size": [427, 640], "counts": "Qof1T1V<4N2M1O100O100O1O101N1O100O100O1O100O2N10]N^D\\1a;dNaD[1_;eNcDY1\\;gNfDX1Y;hNjDV1V;jNkDOIm0[;TOYEi0g:WO\\Ef0c:[O`Ec0^:_ObE`0^:@cE?]:BbE>^:CbE;_:EaE:`:G_E8b:H_E6b:K]E4d:M\\E1e:O[E0g:0YENh:2XEMi:4VEKk:6UEHl:8TEGm::SEDn:<RECo:>QE@Q;`0o0N3N2M3N1O00O100000000000000000NdLMbI1Y67eIHW6?hI_OT6g0]2^OgDi0S;BaDf0X;k0J7H7J<B5L3000000000000001O1O2O0O1O1O1O1O1001O110O00O2O00000O10O100O1O2O0O2N1O2O0O2N1O2N100O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1N2nGTLb6n3[ITLc6o3ZISLd6R4VIQLh6T4SInKl6V4nHlKP7Z4kHhKQ7_4jHcKR7d4hH_Kn6n4mHTKl6i5N2XNaIcL`6]3bI_L`6a3bI[L`6e3bIdK1D_6h4bI]K:CV6o4`1N2O1N2N2N2N2N2000G`KgG`4T8fKkGZ4P8lKoGT4m7QLRHo3i7k0L5K5L4K5K5L3L5ZOf0O100O1O10O01O01O001OVOjH`KV7_4QI[Ko6d4XIWKg6h4`IRK`6m4gImJY6Q5R1N1O[OQKRIm4Q7SKmHl4U7VKhHi4[7WKdHg4^7[K_Hd4c7]K[Ha4i7`KTH_4n7=2N3M2N2N3M2N3L3N2N3M2N3M2N2N2M2O2N2N2N2N2N00N1O2N22N3M3M3CVFUMm9j2UFRMm9n29N2O1O1O1O1O1O0011O2M4M2N0O2O0O1100O1O100O1O1N2N2N2N1O1O1O1O001dM\\Eo1e:mMcEn1^:nMjEl1P;K4L2N1O2N1O2N1L5K5K4N3M2O2M2K6F^TP3"}, "label": "the horse in the middle"}]}
{"id": 212566, "image": "COCO_train2014_000000212566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bench\"."}], "task": "refering", "answer_template": "The \"a green bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 250, 255, 256, 257, 258, 259, 260, 261, 270, 271, 276], "bbox": [0.0, 0.6486875000000001, 0.7447775175644028, 0.812828125], "iscrowd": 0, "area": 7469.284449999999, "rle": {"size": [640, 427], "counts": "`?d0\\c04K10O1000O0100000O010000O010000O010000O0100000O01000O0100000O01000O10O1000O10O10O10O1000O10O10O1000O10O10O10O1000O0100000O01000O0100000O010000O0100000O01000O0100000O01000O10O1000O01000O10O1000O10O10O1000O10O10O10O1000O10O1000O01000O0100000O010000O0100000O01000O0100000O01000O10O1000O01000O10O1000O10O10O10O1000O10O10O1000O10O11O7H7J7I7H5L00000O1000000O10N2L3M4L4K5L301O1O1@S]OLmb05?O1000O1000O0100000O0100000O010000O0100000O0100000O10O10O1000O10O1000O10O1000O10O1000O0100000O01000000O01000O10O1000O10O1000O10O1000O10O10O100000O0100000O010000O0100000O0100000O01000O1000O10O1000O10O1000O10O1000O010001O5J6K5L4K2OO001O001N1XO\\]O4eb0Kc]OM]b03k]OEUb0;i0O001O001O0OjfS2"}, "label": "a green bench"}]}
{"id": 212566, "image": "COCO_train2014_000000212566.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small green bench used for sitting\"."}], "task": "refering", "answer_template": "The \"a small green bench used for sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 235, 236, 243, 244, 245, 246, 247, 248, 250, 255, 256, 257, 258, 259, 260, 261, 270, 271, 276], "bbox": [0.0, 0.6486875000000001, 0.7447775175644028, 0.812828125], "iscrowd": 0, "area": 7469.284449999999, "rle": {"size": [640, 427], "counts": "`?d0\\c04K10O1000O0100000O010000O010000O010000O0100000O01000O0100000O01000O10O1000O10O10O10O1000O10O10O1000O10O10O10O1000O0100000O01000O0100000O010000O0100000O01000O0100000O01000O10O1000O01000O10O1000O10O10O1000O10O10O10O1000O10O1000O01000O0100000O010000O0100000O01000O0100000O01000O10O1000O01000O10O1000O10O10O10O1000O10O10O1000O10O11O7H7J7I7H5L00000O1000000O10N2L3M4L4K5L301O1O1@S]OLmb05?O1000O1000O0100000O0100000O010000O0100000O0100000O10O10O1000O10O1000O10O1000O10O1000O0100000O01000000O01000O10O1000O10O1000O10O1000O10O10O100000O0100000O010000O0100000O0100000O01000O1000O10O1000O10O1000O10O1000O010001O5J6K5L4K2OO001O001N1XO\\]O4eb0Kc]OM]b03k]OEUb0;i0O001O001O0OjfS2"}, "label": "a small green bench used for sitting"}]}
{"id": 319062, "image": "COCO_train2014_000000319062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white dog with two brown eyes\"."}], "task": "refering", "answer_template": "The \"the white dog with two brown eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 329], "bbox": [0.0668125, 0.39355371900826447, 0.34910937499999994, 0.8825826446280991], "iscrowd": 0, "area": 30047.285750000003, "rle": {"size": [484, 640], "counts": "_`d04P?7I6J7I7H8I7I7I7I7H6K4L5K4K6K4L4L5J5L4L5K4K5L5K4L5K4L5K4L5K4Lk0UOj0VO00000000000000000000000000000000000001O00000000000001O0000000000000000000000000000000000000001O00000001O01O001O00001O00001O0000100O1O1O1O1O1O1O1O1O2N1O100O2N100O2O0O2N100O2N100O2N100O2N100O1O100O1O2O0O1O100O1O2O0O2N101N1O2O0O2N100O2N100O1O101N3M3N1N3L4Li0WO<D3M4L4M2M4L3M4L4L3M4L4L3M4L3M4K5I6K6JjgT6"}, "label": "the white dog with two brown eyes"}]}
{"id": 319062, "image": "COCO_train2014_000000319062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dog laying head on car seat\"."}], "task": "refering", "answer_template": "The \"dog laying head on car seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328, 329], "bbox": [0.0668125, 0.39355371900826447, 0.34910937499999994, 0.8825826446280991], "iscrowd": 0, "area": 30047.285750000003, "rle": {"size": [484, 640], "counts": "_`d04P?7I6J7I7H8I7I7I7I7H6K4L5K4K6K4L4L5J5L4L5K4K5L5K4L5K4L5K4L5K4Lk0UOj0VO00000000000000000000000000000000000001O00000000000001O0000000000000000000000000000000000000001O00000001O01O001O00001O00001O0000100O1O1O1O1O1O1O1O1O2N1O100O2N100O2O0O2N100O2N100O2N100O2N100O1O100O1O2O0O1O100O1O2O0O2N101N1O2O0O2N100O2N100O1O101N3M3N1N3L4Li0WO<D3M4L4M2M4L3M4L4L3M4L4L3M4L3M4K5I6K6JjgT6"}, "label": "dog laying head on car seat"}]}
{"id": 106072, "image": "COCO_train2014_000000106072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red car with black rims , next to a motorcycle\"."}], "task": "refering", "answer_template": "The \"a red car with black rims , next to a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272], "bbox": [0.305609375, 0.36789227166276345, 0.8824062500000001, 0.795831381733021], "iscrowd": 0, "area": 45541.3754, "rle": {"size": [427, 640], "counts": "ila22Y=2N2M3bEId7\\1nFnNQ9Y1^FQOb9]2O00001O001O00001N101O00001O001O0000001O0O10001O00001O000O101O000O101O000O2O00000O2O0000001N1000001N10001O0O10001O0O10001O000O2O00000O2O00000O2O00001N1000001O0O10001O0000001O000010O000001O0000001O00001O0000001O01O0001O0000001O00001O0000001O0001O01O00001O0000001O0000001O1O1O1O100O1O1O1O1O1O1O1O1O001O1O1O1O1O10O01O1O1O1O1O1O001O1O000000000001O00001O0O2O1O001O1O001N2O0O2O1O0O2O1N101O1N101N2O001N2O001O001O10O01O1O001O10O01O1O00100O001O1O00100O001O1O00100O001O001O10O01O1O001O010O0000001O01O01O00000010O0001O0000010O00001O01O01O00000010O0001N100O101O0O101O0O101O0O100O2O000O2O000O101N10001N10001N100O101O0O101O0O10001N100O2O000O101O0O101N10001N10000O2O1O1N101N2O1O1N2O1O0O2O1N2O1O1N101O1N2O1N2O001N3N1O2M3N2N1N3N2M2O1O0O101D;[OSQo0"}, "label": "a red car with black rims , next to a motorcycle"}]}
{"id": 106072, "image": "COCO_train2014_000000106072.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a shiny red acura integra with black rims\"."}], "task": "refering", "answer_template": "The \"a shiny red acura integra with black rims\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272], "bbox": [0.305609375, 0.36789227166276345, 0.8824062500000001, 0.795831381733021], "iscrowd": 0, "area": 45541.3754, "rle": {"size": [427, 640], "counts": "ila22Y=2N2M3bEId7\\1nFnNQ9Y1^FQOb9]2O00001O001O00001N101O00001O001O0000001O0O10001O00001O000O101O000O101O000O2O00000O2O0000001N1000001N10001O0O10001O0O10001O000O2O00000O2O00000O2O00001N1000001O0O10001O0000001O000010O000001O0000001O00001O0000001O01O0001O0000001O00001O0000001O0001O01O00001O0000001O0000001O1O1O1O100O1O1O1O1O1O1O1O1O001O1O1O1O1O10O01O1O1O1O1O1O001O1O000000000001O00001O0O2O1O001O1O001N2O0O2O1O0O2O1N101O1N101N2O001N2O001O001O10O01O1O001O10O01O1O00100O001O1O00100O001O1O00100O001O001O10O01O1O001O010O0000001O01O01O00000010O0001O0000010O00001O01O01O00000010O0001N100O101O0O101O0O101O0O100O2O000O2O000O101N10001N10001N100O101O0O101O0O10001N100O2O000O101O0O101N10001N10000O2O1O1N101N2O1O1N2O1O0O2O1N2O1O1N101O1N2O1N2O001N3N1O2M3N2N1N3N2M2O1O0O101D;[OSQo0"}, "label": "a shiny red acura integra with black rims"}]}
{"id": 204378, "image": "COCO_train2014_000000204378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed animal being held by a little girl\"."}], "task": "refering", "answer_template": "The \"a stuffed animal being held by a little girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 41, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 96, 97, 98, 114, 115, 116, 117, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 170, 171], "bbox": [0.29418, 0.10537537537537539, 0.666, 0.7892492492492492], "iscrowd": 0, "area": 18417.11955, "rle": {"size": [333, 500], "counts": "_P`17T:8I4L3L5L3M3mL^OULe0P39kLIR3<jLFT3>gLFU3=iLET3=jLFS3<lLFR3;mLFQ3<mLER3=mLDR3>kLCT3?kLBS3`0kLAU3`0jLAV3?hLBY3>fLB[3>WKhMZ1l1_3>`LBa3>]LDc3<\\LDd3=ZLCh3>UL@o3`0oK_OT4a0jKBU4>iKFU4:iKIV47hKLW44gKNY42^K7b4k2O00000O10001O0000000000N2N2O1N2L3M4O1O10000O1000000O10O10O10000O1001O3M3M4L3M3L4M3NO01OWKZKl3e4TL\\Kl3c4SL_KNNS3b4oLaKH3Y3[4nLdKF4\\3X4mLeKG2]3X4lLgKG1]3W4lLiKF1^3U4lLkKF0^3T4lLnKDN[2F\\N]4EQLCLZ2H^NZ4EULAIZ2K^NX4FVL@H[2K_NV4FZL^OE[2N`NR4G]L\\OEZ2NcNo3G`L[O5h1^OUOl3JbLWO5g1_OXOi3LeMj0dNZOk3I`Ml0gNZOP4DXMP1jN\\OZ4YOlLX1lN_OU4]OnLR1PO_Oo3DSMi0oNBl3HXMa0nNFh3M]M7nNJc33aMMoNO]39cMFSOOX3?dM_OWO1S3d0eMYOZO2o2i0fMSO^O2h2Q1hMkNC3b2W1RN\\N^O<\\2^10_NMf14WNHo19mMHU29fMH]2R42NO1N2O2N1O1O1O1O1O2N1N2N2N2N2O2M2N3M4L3N3N1N3N1O1N3N^JROg1m0XNYOe1d0]N_O`1?cNDZ1:hNIV13mNOR1MQO6n0EUO=j0_OYOd0e0XO^Ok0`0ROBP1l44MO1O1O1O1O1O1O1O003M2N1OO01O001O1O001O1O1O2O0O1O2N1O1O1O2N1O1LZSf1"}, "label": "a stuffed animal being held by a little girl"}]}
{"id": 204378, "image": "COCO_train2014_000000204378.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stuffed animal that the little girl is holding\"."}], "task": "refering", "answer_template": "The \"stuffed animal that the little girl is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 41, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 96, 97, 98, 114, 115, 116, 117, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 170, 171], "bbox": [0.29418, 0.10537537537537539, 0.666, 0.7892492492492492], "iscrowd": 0, "area": 18417.11955, "rle": {"size": [333, 500], "counts": "_P`17T:8I4L3L5L3M3mL^OULe0P39kLIR3<jLFT3>gLFU3=iLET3=jLFS3<lLFR3;mLFQ3<mLER3=mLDR3>kLCT3?kLBS3`0kLAU3`0jLAV3?hLBY3>fLB[3>WKhMZ1l1_3>`LBa3>]LDc3<\\LDd3=ZLCh3>UL@o3`0oK_OT4a0jKBU4>iKFU4:iKIV47hKLW44gKNY42^K7b4k2O00000O10001O0000000000N2N2O1N2L3M4O1O10000O1000000O10O10O10000O1001O3M3M4L3M3L4M3NO01OWKZKl3e4TL\\Kl3c4SL_KNNS3b4oLaKH3Y3[4nLdKF4\\3X4mLeKG2]3X4lLgKG1]3W4lLiKF1^3U4lLkKF0^3T4lLnKDN[2F\\N]4EQLCLZ2H^NZ4EULAIZ2K^NX4FVL@H[2K_NV4FZL^OE[2N`NR4G]L\\OEZ2NcNo3G`L[O5h1^OUOl3JbLWO5g1_OXOi3LeMj0dNZOk3I`Ml0gNZOP4DXMP1jN\\OZ4YOlLX1lN_OU4]OnLR1PO_Oo3DSMi0oNBl3HXMa0nNFh3M]M7nNJc33aMMoNO]39cMFSOOX3?dM_OWO1S3d0eMYOZO2o2i0fMSO^O2h2Q1hMkNC3b2W1RN\\N^O<\\2^10_NMf14WNHo19mMHU29fMH]2R42NO1N2O2N1O1O1O1O1O2N1N2N2N2N2O2M2N3M4L3N3N1N3N1O1N3N^JROg1m0XNYOe1d0]N_O`1?cNDZ1:hNIV13mNOR1MQO6n0EUO=j0_OYOd0e0XO^Ok0`0ROBP1l44MO1O1O1O1O1O1O1O003M2N1OO01O001O1O001O1O1O2O0O1O2N1O1O1O2N1O1LZSf1"}, "label": "stuffed animal that the little girl is holding"}]}
{"id": 474725, "image": "COCO_train2014_000000474725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bikini girl standing in sand\"."}], "task": "refering", "answer_template": "The \"the bikini girl standing in sand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 96, 97, 98, 113, 114, 115, 116, 131, 132, 133, 134, 150, 151, 152, 167, 168, 169, 170, 185, 186, 187, 188, 203, 204, 205, 206, 221, 222, 223, 224, 239, 240, 241, 242, 257, 258, 259, 260, 275, 276, 277, 278, 295, 296, 313], "bbox": [0.31236, 0.22696, 0.50338, 0.9685400000000001], "iscrowd": 0, "area": 22004.7516, "rle": {"size": [500, 500], "counts": "S\\\\22^?4oBL^:9]E1Y:4bE7S:NhE<P:GkEd0m9^OnE^2X8eMcGY4b6kKWIl4S6ZKfIi4Y6\\KaIf4^6^K\\If4b6_KXIc4h6aKQIb4o6aKkHb4U7n101O0`NiGdKW8\\4_HnJa7R5`11O0J7I6K6K4L4L5K4L5K4TOl0N3N1D=G8101N1O11O01O0gKTIJl64VILk61XINk6MWI3S7API>Y7WOiHj0`7jNbHV1h7^N[Ha1n7TNTHl1V8gMnGX2\\8\\MfGd2d8oL_GQ3j8dLYG\\3P9WLSGi3Z:2M3N2N2N2M4M2N2N2M3N2N2dEUKm9m4RFWKS9b5jFcJf0Nm6i5\\H]J5;]7X5\\HgK`7Z6M2M4M2M4M3L7J8G9H8G9GW1cNg0YOg0YO>C8G:F:F9G:F:G8G:Flhh3"}, "label": "the bikini girl standing in sand"}]}
{"id": 474725, "image": "COCO_train2014_000000474725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cartoon woman standing next to a surfboard\"."}], "task": "refering", "answer_template": "The \"a cartoon woman standing next to a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 96, 97, 98, 113, 114, 115, 116, 131, 132, 133, 134, 150, 151, 152, 167, 168, 169, 170, 185, 186, 187, 188, 203, 204, 205, 206, 221, 222, 223, 224, 239, 240, 241, 242, 257, 258, 259, 260, 275, 276, 277, 278, 295, 296, 313], "bbox": [0.31236, 0.22696, 0.50338, 0.9685400000000001], "iscrowd": 0, "area": 22004.7516, "rle": {"size": [500, 500], "counts": "S\\\\22^?4oBL^:9]E1Y:4bE7S:NhE<P:GkEd0m9^OnE^2X8eMcGY4b6kKWIl4S6ZKfIi4Y6\\KaIf4^6^K\\If4b6_KXIc4h6aKQIb4o6aKkHb4U7n101O0`NiGdKW8\\4_HnJa7R5`11O0J7I6K6K4L4L5K4L5K4TOl0N3N1D=G8101N1O11O01O0gKTIJl64VILk61XINk6MWI3S7API>Y7WOiHj0`7jNbHV1h7^N[Ha1n7TNTHl1V8gMnGX2\\8\\MfGd2d8oL_GQ3j8dLYG\\3P9WLSGi3Z:2M3N2N2N2M4M2N2N2M3N2N2dEUKm9m4RFWKS9b5jFcJf0Nm6i5\\H]J5;]7X5\\HgK`7Z6M2M4M2M4M3L7J8G9H8G9GW1cNg0YOg0YO>C8G:F:F9G:F:G8G:Flhh3"}, "label": "a cartoon woman standing next to a surfboard"}]}
{"id": 474725, "image": "COCO_train2014_000000474725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfboard in the sand with a girl in a bikini\"."}], "task": "refering", "answer_template": "The \"a surfboard in the sand with a girl in a bikini\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 19, 20, 21, 37, 38, 39, 40, 55, 56, 57, 58, 73, 74, 75, 76, 91, 92, 93, 94, 95, 110, 111, 112, 113, 128, 129, 130, 131, 146, 147, 148, 149, 164, 165, 166, 167, 182, 183, 184, 185, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 274, 275, 276, 277], "bbox": [0.09418000000000001, 0.0, 0.41704, 0.86234], "iscrowd": 0, "area": 33993.683300000004, "rle": {"size": [500, 500], "counts": "_of0Z2Q=b0^Oa0_Ob0E:E<E;F9H9G8H9G8H9G9G4L4L3M4L4L3M4L4N2M2O2N2M3N1N3N2N1N3N2N2M2O2M3N1O2M3N2N1N3N2N2M2O2M3N1O2M3N2N1N3N2M2O2N2M3N1O2M3N2O0O2N2N101N1OJ7M2N2N2O2M2N2N3M2N2N2O2M2N2N3M2N2N3N1N2N2N3M2N2N3M2O1N2N3M2N2N3M2O1N2N3M2N2N3M2O1N2dK\\Hd0g7fL\\HZ1<g1Z7lL\\HV1i0e1m6QM\\HT1V1b1a6UM\\HR1b1`1T6ZM\\HP1o1^1a6WNoIc1R6QN_Ja1i5TNiJY1U:_Oa0^OU`2F^`M3M3N2M3M3N2M4L3jN^OXCd0b<GVC<e<NSC5h<5QCMi<b1H8I7H8H9M20000003M;E;E;E:F;E;F:E9G8H8H7I8H^e]4"}, "label": "a surfboard in the sand with a girl in a bikini"}]}
{"id": 474725, "image": "COCO_train2014_000000474725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"surf board on the left side\"."}], "task": "refering", "answer_template": "The \"surf board on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 19, 20, 21, 37, 38, 39, 40, 55, 56, 57, 58, 73, 74, 75, 76, 91, 92, 93, 94, 95, 110, 111, 112, 113, 128, 129, 130, 131, 146, 147, 148, 149, 164, 165, 166, 167, 182, 183, 184, 185, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 274, 275, 276, 277], "bbox": [0.09418000000000001, 0.0, 0.41704, 0.86234], "iscrowd": 0, "area": 33993.683300000004, "rle": {"size": [500, 500], "counts": "_of0Z2Q=b0^Oa0_Ob0E:E<E;F9H9G8H9G8H9G9G4L4L3M4L4L3M4L4N2M2O2N2M3N1N3N2N1N3N2N2M2O2M3N1O2M3N2N1N3N2N2M2O2M3N1O2M3N2N1N3N2M2O2N2M3N1O2M3N2O0O2N2N101N1OJ7M2N2N2O2M2N2N3M2N2N2O2M2N2N3M2N2N3N1N2N2N3M2N2N3M2O1N2N3M2N2N3M2O1N2N3M2N2N3M2O1N2dK\\Hd0g7fL\\HZ1<g1Z7lL\\HV1i0e1m6QM\\HT1V1b1a6UM\\HR1b1`1T6ZM\\HP1o1^1a6WNoIc1R6QN_Ja1i5TNiJY1U:_Oa0^OU`2F^`M3M3N2M3M3N2M4L3jN^OXCd0b<GVC<e<NSC5h<5QCMi<b1H8I7H8H9M20000003M;E;E;E:F;E;F:E9G8H8H7I8H^e]4"}, "label": "surf board on the left side"}]}
{"id": 245367, "image": "COCO_train2014_000000245367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in white tshirt , slipped off from the skating slidder\"."}], "task": "refering", "answer_template": "The \"a boy in white tshirt , slipped off from the skating slidder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 169, 170, 171, 182, 183, 184, 185, 195, 196, 197, 198, 199, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 235, 236, 238, 239, 248, 249, 251, 252, 253, 254, 261, 262, 265, 266, 267, 274, 275, 276, 288], "bbox": [0.0, 0.35285937500000003, 0.5694444444444444, 0.97265625], "iscrowd": 0, "area": 30927.93059999999, "rle": {"size": [640, 360], "counts": "Z9c0\\c0T1lNT1kNU1lNT1lNb0]O3N1O1N3N1O1N3N1N2O2N1N2O2N1N2O2N1N2O2N1N2O2N10005K:F:F9G4L5L3L4L5K4L4L5K4L5K5K4L5LTLhETNQ:n1PFVNi9k1YFZN^9h1cF]NV9e1kF_Nl8d1VGaN`8c1`GbNW8a1kG]NR8g1oGXNm7k1THTNi7o1YHPNc7T2^HjM_7Y2bHgM[7\\2fHbMZ7^2hHaMW7`2iH_MW7a2kH]MT7M]HUNa0l1R7LbHWN<k1S7KgHWN8k1R7KkHYNn0o0W6f0QIXNc16]5_1UIZNa12[5a1YI[N_10X5c1_IZN\\10U5c1cI]N[1KS5e1fI_NZ1HQ5f1iIbNX1Do4h1mIcNW1Am4i1PJfNV1\\Ok4j1SJjNU1XOi4k1VJlNT1VOe4l1[JnNS1QOc4n0oGPO_2P1R1nNa4R1nGmNc2S1P1jN_4V1oGmNc2S1o0hN`4X1nGmNd2R1o0gN`4Y1nGnNd2R1m0fNb4Z1mGnNe2R1l0eNb4Z1nGnNf2S1i0dNe4Z1lGoNg2S1h0bNf4[1mGoNe2U1h0_Ng4]1lGoNf2U1f0^Nj4\\1kGQOf2U1e0\\Nk4]1kGROf2U1c0[Nm4^1jGQOh2V1a0ZNm4^1kGROh2V1`0XNo4_1iGSOi2V1>WNQ5_1iGTOi2V1=UNR5a1hGTOj2V1;TNU5`1gGVOj2V1:RNV5b1fGUOk2X1n8AZDVOi2Y1m8_O[DXOi2Y1k8^O]DYOi2X1j8^O^DZOi2X1i8\\O_D\\Oi2X1g8ZObD]Oi2Y1d8YOdD^Oi2Y1c8WOeD@i2Y1c8TOeDCi2Y1d8POdDGi2Y1d8mNdDJh2Z1f8gNdDNh2[1f8cNdD1g2\\1f8`NdD4g2\\1g8\\NcD8g2\\1h8XNbD<g2]1g8SNdD`0f2^1j9bNWF_1g9`N[Fa1c9_N^Fb1`9^NaFc1]9]NcFe1[9[NfFf1Y9YNhFh1V9WNlFj1R9VNoFk1o8UNSGk1k8UNWGk1g8UN[Gk1c8UN_Gk1_8TNdGl1Z8TNiGk1U8UNmGk1Q8UNQHk1n7TNTHl1j7TNXHl1f7TN]Hk1a7UNaHk1]7UNeHk1Y7VNhHj1V7VNlHj1R7VNQIi1m6WNUIi1i6WNYIi1e6WN]Ii1b6WN_Ii1_6WNcIi1[6WNhIh1V6XNlIh1R6XNPJh1n5XNTJh1j5XNXJh1f5YN\\Jf1b5ZN`Jf1^5ZNdJf1Z5ZNhJf1V5ZNlJf1S5YNPKf1n4[NSKe1k4[NWKe1g4[N[Kb1f4^N\\K_1e4aN^K\\1b4dN`KY1a4hN`KV1`4jNbKS1_4mNcKP1^4oNfKn0Z4ROhKk0Y4TOjKj0V4VOlKg0U4YOmKd0T4[OPLb0P4^ORL?o3ASL=m3BVL;k3EWL8j3GYL7g3IZL5g3KXL6h3IYL6h3JXL5i3KWL5i3JXL5i3KWL5i3JWL6j3JVL5k3KUL5k3JVL5k3KUL5k3KUL4l3KTL5m3KSL5m3GWL8i;O10O0100O1O100O010O1O10iZP3"}, "label": "a boy in white tshirt , slipped off from the skating slidder"}]}
{"id": 245367, "image": "COCO_train2014_000000245367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black and white outfit kicking the skateboard\"."}], "task": "refering", "answer_template": "The \"a man wearing a black and white outfit kicking the skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 169, 170, 171, 182, 183, 184, 185, 195, 196, 197, 198, 199, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 235, 236, 238, 239, 248, 249, 251, 252, 253, 254, 261, 262, 265, 266, 267, 274, 275, 276, 288], "bbox": [0.0, 0.35285937500000003, 0.5694444444444444, 0.97265625], "iscrowd": 0, "area": 30927.93059999999, "rle": {"size": [640, 360], "counts": "Z9c0\\c0T1lNT1kNU1lNT1lNb0]O3N1O1N3N1O1N3N1N2O2N1N2O2N1N2O2N1N2O2N1N2O2N10005K:F:F9G4L5L3L4L5K4L4L5K4L5K5K4L5LTLhETNQ:n1PFVNi9k1YFZN^9h1cF]NV9e1kF_Nl8d1VGaN`8c1`GbNW8a1kG]NR8g1oGXNm7k1THTNi7o1YHPNc7T2^HjM_7Y2bHgM[7\\2fHbMZ7^2hHaMW7`2iH_MW7a2kH]MT7M]HUNa0l1R7LbHWN<k1S7KgHWN8k1R7KkHYNn0o0W6f0QIXNc16]5_1UIZNa12[5a1YI[N_10X5c1_IZN\\10U5c1cI]N[1KS5e1fI_NZ1HQ5f1iIbNX1Do4h1mIcNW1Am4i1PJfNV1\\Ok4j1SJjNU1XOi4k1VJlNT1VOe4l1[JnNS1QOc4n0oGPO_2P1R1nNa4R1nGmNc2S1P1jN_4V1oGmNc2S1o0hN`4X1nGmNd2R1o0gN`4Y1nGnNd2R1m0fNb4Z1mGnNe2R1l0eNb4Z1nGnNf2S1i0dNe4Z1lGoNg2S1h0bNf4[1mGoNe2U1h0_Ng4]1lGoNf2U1f0^Nj4\\1kGQOf2U1e0\\Nk4]1kGROf2U1c0[Nm4^1jGQOh2V1a0ZNm4^1kGROh2V1`0XNo4_1iGSOi2V1>WNQ5_1iGTOi2V1=UNR5a1hGTOj2V1;TNU5`1gGVOj2V1:RNV5b1fGUOk2X1n8AZDVOi2Y1m8_O[DXOi2Y1k8^O]DYOi2X1j8^O^DZOi2X1i8\\O_D\\Oi2X1g8ZObD]Oi2Y1d8YOdD^Oi2Y1c8WOeD@i2Y1c8TOeDCi2Y1d8POdDGi2Y1d8mNdDJh2Z1f8gNdDNh2[1f8cNdD1g2\\1f8`NdD4g2\\1g8\\NcD8g2\\1h8XNbD<g2]1g8SNdD`0f2^1j9bNWF_1g9`N[Fa1c9_N^Fb1`9^NaFc1]9]NcFe1[9[NfFf1Y9YNhFh1V9WNlFj1R9VNoFk1o8UNSGk1k8UNWGk1g8UN[Gk1c8UN_Gk1_8TNdGl1Z8TNiGk1U8UNmGk1Q8UNQHk1n7TNTHl1j7TNXHl1f7TN]Hk1a7UNaHk1]7UNeHk1Y7VNhHj1V7VNlHj1R7VNQIi1m6WNUIi1i6WNYIi1e6WN]Ii1b6WN_Ii1_6WNcIi1[6WNhIh1V6XNlIh1R6XNPJh1n5XNTJh1j5XNXJh1f5YN\\Jf1b5ZN`Jf1^5ZNdJf1Z5ZNhJf1V5ZNlJf1S5YNPKf1n4[NSKe1k4[NWKe1g4[N[Kb1f4^N\\K_1e4aN^K\\1b4dN`KY1a4hN`KV1`4jNbKS1_4mNcKP1^4oNfKn0Z4ROhKk0Y4TOjKj0V4VOlKg0U4YOmKd0T4[OPLb0P4^ORL?o3ASL=m3BVL;k3EWL8j3GYL7g3IZL5g3KXL6h3IYL6h3JXL5i3KWL5i3JXL5i3KWL5i3JWL6j3JVL5k3KUL5k3JVL5k3KUL5k3KUL4l3KTL5m3KSL5m3GWL8i;O10O0100O1O100O010O1O10iZP3"}, "label": "a man wearing a black and white outfit kicking the skateboard"}]}
{"id": 523907, "image": "COCO_train2014_000000523907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leg of a man sitting down behind the hotdog\"."}], "task": "refering", "answer_template": "The \"the leg of a man sitting down behind the hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 157, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 249, 250, 272, 273, 275, 295, 296, 297, 298, 317, 318, 319, 320, 321, 339, 340, 343, 344, 366, 367, 389, 390], "bbox": [0.766890625, 0.0011250000000000001, 1.0, 1.0], "iscrowd": 0, "area": 31081.48854999999, "rle": {"size": [480, 640], "counts": "QbV75h>3N2N2M3N2N2N2N2M2O2N2N2N2M3N100000000001O000000001O000000hIoNbNQ1]1QOcNn0[1UOdNk0[1XOcNh0[1[OdNe0[1^OcNb0[1AeN>Z1EdN;[1GdN9Z1JeN1UJXOU7j0dNNZJUOP7P1fNK]18]MnN^Lk0R6:]MPO^Lf0T6=[MQO^Lc0T6?\\MRO]L`0U6`0[MTO_L;U6d0YMUO_L8U6f0ZMVO^L5V6h0XMXO_L0W6k0XMYO^LMX6l0XM[O^LIW6P1WM\\O_LDY6R1VM^O^LAY6U1VM^O^L^OZ6V1UMc0i2@TMa0j2BTM`0j2BSMa0j2CSM>m2CQM?m2DoL?o2CoL>Q3DkL?S3CkL?T3BjL?U3DgL?W3CgL?X3CeL>Z3DcL?[3DbL>]3CaL>^3D_L?_3D^L>a3C]L?a3D[L>e3CYL?e3DXL>f3DWL>h3DVL>h3EUL=i3ETL=k3FRL<l3FRL<e2_K]MV4K=c2dK_MR4K<d2dK_MR4K<d2eK^MR4J;g2eK]MR4J;g2fK]MQ4I;h2fK^MP4G<V2QKnMd03P4G=U2SKQN?2S4E[N4`NR2gNRN;1T4DTNk0[N[1VOUN60W4BfMb1`Ne0AVN3OW4AgMi1[N?GXNOMn6`1YI8OYNKMm6f1UI36[NGJn6l1RIL=ZNoNGe04m6T2nHEd0[NlNHf02l6Z2iH@n0YNiNKd00l6f4jI^KgNLW8d4VI^KeNNR8d4]I\\KbN0o7b4cI]K_N0m7b4fI^K^N0j7`4kI`K\\N0g7_4oIaK[N0d7]4TJcKYN0a7T3[HjMm1ROXNO]7Q3iHgMe1YOVNOW7U3QI]Me1_OTNOQ7X3[IRMd1GQNOk6\\3ZLeLkL0f6^3`LbLkLOa6c3dL^LlLO[6f3jL[LlLOU6i3PMXLlLOo5m3UMTL_3l3bLTL]3m3cLSL\\3m3eLTLZ3k3gLULX3l3hLTLX3k3iLULV3l3jLTLV3k3kLULT3k3mLULR3k3oLULQ3j3PMVLo2j3RMVLn2i3SMWLl2j3TMVLl2i3UMWLj2i3WMWLh2i3YMWL\\2S4eMmKZ2S4gMmKX2R4jMnKU2Q4mMoKR2P4PNPLo1o3SNRLl1l3SNWLl1h3SN[Ll1d3SN_Ll0_4UOcKH]58fJ_LLk0^6g2TKTLV8c1^F"}, "label": "the leg of a man sitting down behind the hotdog"}]}
{"id": 523907, "image": "COCO_train2014_000000523907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"floral designed top\"."}], "task": "refering", "answer_template": "The \"floral designed top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 209, 231, 232, 233, 254, 255], "bbox": [0.013546875, 0.003958333333333333, 0.14221875, 0.6585833333333333], "iscrowd": 0, "area": 18331.59135, "rle": {"size": [480, 640], "counts": "c[4\\2f8l4RO9G2M2O2N2N1N3N2N1O2M3N2N1O2M3N1O2M3N1O2N2N2N1O2N2N1O2N2N1O2N2M3M2N1O1O1O100O1O1O1O1O1OgKoJeNQ5m0`KPO_4P1fKlNY4T1kKiNW4T1nKhNT4U1QLgNQ4W1RLfNo3X1VLdNl3Y1XLdNj3Y1[LcNg3Z1^LbNd3[1aL`Na3^1dL^N^3_1gL\\N\\3a1iLZNZ3d1jLXNX3e1mLVNU3h1PMTNR3i1SMRNP3k1UMPNn2m1WMoMk2n1ZMmMV3c1oLYN`3W1eLdNc3S1aLjNg3n0\\LPOi3j0ZLSOl3g0WLWOn3b0VL[OP4?SL^OS4<PLBU48nKEV46nKGV45mKIV43mKJX40lKNW4NlKOX[Q8"}, "label": "floral designed top"}]}
{"id": 523907, "image": "COCO_train2014_000000523907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a dress with black tights\"."}], "task": "refering", "answer_template": "The \"a woman wearing a dress with black tights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 209, 231, 232, 233, 254, 255], "bbox": [0.013546875, 0.003958333333333333, 0.14221875, 0.6585833333333333], "iscrowd": 0, "area": 18331.59135, "rle": {"size": [480, 640], "counts": "c[4\\2f8l4RO9G2M2O2N2N1N3N2N1O2M3N2N1O2M3N1O2M3N1O2N2N2N1O2N2N1O2N2N1O2N2M3M2N1O1O1O100O1O1O1O1O1OgKoJeNQ5m0`KPO_4P1fKlNY4T1kKiNW4T1nKhNT4U1QLgNQ4W1RLfNo3X1VLdNl3Y1XLdNj3Y1[LcNg3Z1^LbNd3[1aL`Na3^1dL^N^3_1gL\\N\\3a1iLZNZ3d1jLXNX3e1mLVNU3h1PMTNR3i1SMRNP3k1UMPNn2m1WMoMk2n1ZMmMV3c1oLYN`3W1eLdNc3S1aLjNg3n0\\LPOi3j0ZLSOl3g0WLWOn3b0VL[OP4?SL^OS4<PLBU48nKEV46nKGV45mKIV43mKJX40lKNW4NlKOX[Q8"}, "label": "a woman wearing a dress with black tights"}]}
{"id": 523907, "image": "COCO_train2014_000000523907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding food\"."}], "task": "refering", "answer_template": "The \"a hand holding food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.053937500000000006, 0.6629166666666666, 0.5309062499999999, 0.9887708333333334], "iscrowd": 0, "area": 34601.3859, "rle": {"size": [480, 640], "counts": "ek`01m>2N2O1N2O1N2O1N2N3N1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O2M2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2000000O101O0000000000000000000000000000000000000000000000000000000000000O2O0000000000O10000000000000000000000000000000000000000000000000O10000001O00001O000O101O0000001O0000001O00001O0000001O0000001O0000001O0000O10000O10000O10O10O100O10000O10000O10000O10000O10000O10000O10000O1001O000000001O000O101O000O1O2N1N2O1O2M3N3M2M4M3M2M4M2N3M3M2N3M2N3L4M2N3M2N3M3M2M4M2N3M3M3M3M3M3L4M3M3M3M3M3M3L4M3M3M3M3M3M3M3L4M3M3MZW\\4"}, "label": "a hand holding food"}]}
{"id": 523907, "image": "COCO_train2014_000000523907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"thumb finger of a person\"."}], "task": "refering", "answer_template": "The \"thumb finger of a person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.053937500000000006, 0.6629166666666666, 0.5309062499999999, 0.9887708333333334], "iscrowd": 0, "area": 34601.3859, "rle": {"size": [480, 640], "counts": "ek`01m>2N2O1N2O1N2O1N2N3N1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O2M2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2000000O101O0000000000000000000000000000000000000000000000000000000000000O2O0000000000O10000000000000000000000000000000000000000000000000O10000001O00001O000O101O0000001O0000001O00001O0000001O0000001O0000001O0000O10000O10000O10O10O100O10000O10000O10000O10000O10000O10000O10000O1001O000000001O000O101O000O1O2N1N2O1O2M3N3M2M4M3M2M4M2N3M3M2N3M2N3L4M2N3M2N3M3M2M4M2N3M3M3M3M3M3L4M3M3M3M3M3M3L4M3M3M3M3M3M3M3L4M3M3MZW\\4"}, "label": "thumb finger of a person"}]}
{"id": 155269, "image": "COCO_train2014_000000155269.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra eating some hay\"."}], "task": "refering", "answer_template": "The \"a zebra eating some hay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 246, 247, 248, 250, 251, 266, 267, 269, 271, 273, 274, 289, 290, 292, 294, 296, 297], "bbox": [0.583109375, 0.4392037470725995, 0.9382187500000001, 0.8409836065573771], "iscrowd": 0, "area": 21127.4728, "rle": {"size": [427, 640], "counts": "Zhk42V=6J6jDER9c0hFCQ9c0jFBP9d0jFCo8c0lFBP9c0jFg0m7\\OnGg0R8\\OiGe0W8_OdGc0[8A_Gb0`8A\\G`0d8T200O2N2N101NkNdG]M[8b2lGZMR8e2UHVMk7h2]HRMb7m2eHoLY7P3nHkLQ7U3SIhLm6V3VIjLh6U3g100O2N1O101N2N2O1N2N2N2O1N2O1O1O1O1O1O1O100O1O1O1O1O1O1O1O101N1001O00001O00001O001O00001O000010O01O00001O0000000UGhMc6X2ZImMd6S2YIQNf6o1XIUNf6k1XIYNf6g1XI\\Ng6c1XI_Nh6a1VIaNj6_1TIcNm6\\1RIeNn6[1PIgNP7Y1nHiNR7W1lHkNT7U1jHbNiN\\O]8R2hH[NXO_OP8V2gHUNDAe7Z2eHTNY8l1eGTN]8l1aGSNb8m1\\GSNf8m1QGYNR9g1bFcN`9a23N2O2M2O1N2N2O1N2O1N2O2M2N2O1N2O1O1000001O00000:Fb1^N001O1O010VGeKb8d4N2N1O000001jNRHmLn7U2SIhMm6n1`IoMa6P2bImM^6R2fIkM[6S2iIjMW6V2_2O2O0O2N100O2O0O2N10001O00001O000010O0001O01O2O001N101O001N101O0O2O1O0nEhMo8Z2jFmMU9S2eFSN[9n1]FYNb9h1XF_Ng9c20001N10001N10001N10003M2N2OO101N1O1O1N2N2M3N2N2N3M2M3N2N3M3L3QN_EP1d:mNdEj0`:SOhE?`:@gE3a:JfEIc:5\\1M2Ni\\`0"}, "label": "a zebra eating some hay"}]}
{"id": 155269, "image": "COCO_train2014_000000155269.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra eating grass facing the right\"."}], "task": "refering", "answer_template": "The \"a zebra eating grass facing the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 246, 247, 248, 250, 251, 266, 267, 269, 271, 273, 274, 289, 290, 292, 294, 296, 297], "bbox": [0.583109375, 0.4392037470725995, 0.9382187500000001, 0.8409836065573771], "iscrowd": 0, "area": 21127.4728, "rle": {"size": [427, 640], "counts": "Zhk42V=6J6jDER9c0hFCQ9c0jFBP9d0jFCo8c0lFBP9c0jFg0m7\\OnGg0R8\\OiGe0W8_OdGc0[8A_Gb0`8A\\G`0d8T200O2N2N101NkNdG]M[8b2lGZMR8e2UHVMk7h2]HRMb7m2eHoLY7P3nHkLQ7U3SIhLm6V3VIjLh6U3g100O2N1O101N2N2O1N2N2N2O1N2O1O1O1O1O1O1O100O1O1O1O1O1O1O1O101N1001O00001O00001O001O00001O000010O01O00001O0000000UGhMc6X2ZImMd6S2YIQNf6o1XIUNf6k1XIYNf6g1XI\\Ng6c1XI_Nh6a1VIaNj6_1TIcNm6\\1RIeNn6[1PIgNP7Y1nHiNR7W1lHkNT7U1jHbNiN\\O]8R2hH[NXO_OP8V2gHUNDAe7Z2eHTNY8l1eGTN]8l1aGSNb8m1\\GSNf8m1QGYNR9g1bFcN`9a23N2O2M2O1N2N2O1N2O1N2O2M2N2O1N2O1O1000001O00000:Fb1^N001O1O010VGeKb8d4N2N1O000001jNRHmLn7U2SIhMm6n1`IoMa6P2bImM^6R2fIkM[6S2iIjMW6V2_2O2O0O2N100O2O0O2N10001O00001O000010O0001O01O2O001N101O001N101O0O2O1O0nEhMo8Z2jFmMU9S2eFSN[9n1]FYNb9h1XF_Ng9c20001N10001N10001N10003M2N2OO101N1O1O1N2N2M3N2N2N3M2M3N2N3M3L3QN_EP1d:mNdEj0`:SOhE?`:@gE3a:JfEIc:5\\1M2Ni\\`0"}, "label": "a zebra eating grass facing the right"}]}
{"id": 155269, "image": "COCO_train2014_000000155269.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra with head up\"."}], "task": "refering", "answer_template": "The \"zebra with head up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 212, 213, 214, 232, 233, 235, 236, 258, 259], "bbox": [0.0015, 0.2780327868852459, 0.32610937500000003, 0.773536299765808], "iscrowd": 0, "area": 19715.37639999999, "rle": {"size": [427, 640], "counts": "db06T=9F5K2N2N2O1N1O2N2N2O1N2N2N200O1O100O1O1O1@eNPE[1P;jNjDW1U;nNfDS1Y;`00O01O10000001O0000001O0000001O00001O00000011N2N1O2O0O2N101N4L4L5L3L4L4L3N1N3N1O2SHhLc5Y3ZJhLf5Y3WJhLk5Y3RJgLP6Z3lIgLU6Z3hIgLZ6Z3cIgL^6Z3^IgLc6Z3ZIgLh6Z3UIfLm6[3oHfLR7[3lHeLV7\\3]HoLc7S3PHWMR8P42N1O2O1N1O0000000O101N100O100O100O100O2G8F:G9G9F:G9I70001O00000000000000001O0000O10000000000000000000000000000ZGbM_6^2_IgM^6Y2`IkM^6U2`IoM^6Q2`ITN]6l1aIVN^6k1`IWN`6i1^IXNc6k1XIWNh6l1SIUNn6n1mHTNS7o1hHSNX7Q2bHPN_7S2\\HoMd7T2WHmMj7V2QHkMP8X2kGhMW8[2dGeM^8b32N2N2M3N2N2N2N1N3N2N2N2N2N2lNeFUN^9i1hFQNZ9n1kFlMW9S2mFhMU9W2PGcMR9\\2o0O1N2O20O6K6I6K5[GnLf6X3oHlLQ7Z3eHkL[7Z3[HjLf7[3PHiLP8^3eGgL[8T4010O10O01O10O010OiMhH[NY7e1W20O100003M8I6I8H8H7Ib0^Of0ZOi^c5"}, "label": "zebra with head up"}]}
{"id": 155269, "image": "COCO_train2014_000000155269.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing mostly in light\"."}], "task": "refering", "answer_template": "The \"a zebra standing mostly in light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 212, 213, 214, 232, 233, 235, 236, 258, 259], "bbox": [0.0015, 0.2780327868852459, 0.32610937500000003, 0.773536299765808], "iscrowd": 0, "area": 19715.37639999999, "rle": {"size": [427, 640], "counts": "db06T=9F5K2N2N2O1N1O2N2N2O1N2N2N200O1O100O1O1O1@eNPE[1P;jNjDW1U;nNfDS1Y;`00O01O10000001O0000001O0000001O00001O00000011N2N1O2O0O2N101N4L4L5L3L4L4L3N1N3N1O2SHhLc5Y3ZJhLf5Y3WJhLk5Y3RJgLP6Z3lIgLU6Z3hIgLZ6Z3cIgL^6Z3^IgLc6Z3ZIgLh6Z3UIfLm6[3oHfLR7[3lHeLV7\\3]HoLc7S3PHWMR8P42N1O2O1N1O0000000O101N100O100O100O100O2G8F:G9G9F:G9I70001O00000000000000001O0000O10000000000000000000000000000ZGbM_6^2_IgM^6Y2`IkM^6U2`IoM^6Q2`ITN]6l1aIVN^6k1`IWN`6i1^IXNc6k1XIWNh6l1SIUNn6n1mHTNS7o1hHSNX7Q2bHPN_7S2\\HoMd7T2WHmMj7V2QHkMP8X2kGhMW8[2dGeM^8b32N2N2M3N2N2N2N1N3N2N2N2N2N2lNeFUN^9i1hFQNZ9n1kFlMW9S2mFhMU9W2PGcMR9\\2o0O1N2O20O6K6I6K5[GnLf6X3oHlLQ7Z3eHkL[7Z3[HjLf7[3PHiLP8^3eGgL[8T4010O10O01O10O010OiMhH[NY7e1W20O100003M8I6I8H8H7Ib0^Of0ZOi^c5"}, "label": "a zebra standing mostly in light"}]}
{"id": 89734, "image": "COCO_train2014_000000089734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby horse to the right of an older horse\"."}], "task": "refering", "answer_template": "The \"a baby horse to the right of an older horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 64, 65, 66, 82, 83, 84, 99, 100, 101, 102, 103, 117, 118, 119, 135, 136, 137, 153, 154, 155, 171, 172, 173, 189, 190, 191, 208, 209], "bbox": [0.50806, 0.20874666666666666, 0.73064, 0.8775733333333334], "iscrowd": 0, "area": 12141.338650000002, "rle": {"size": [375, 500], "counts": "UZm24];6J7UMCgH07b0Q7IWH3=9Z7OhG6b00d7T1QHQOm7]2N3]OfLfH[3U7e0L4L4N2000O10000WO[IXLn6f3b0M3M3K5nNiG`N]8]1RHQNS8m1l0N2O16J7H7I>jH^M]4P3cJeMZ5^4M3M3M4L3M2N3M2N2NcLjJ=T5FQK4l40XKKf48^KD_4>fKeNH[N_4T3mKiM=mNd3`3PLlLT5AgJ`4i5mKgIY4W6f0N3L3N8H>A`0A:F2M2O2NOO1000[NoKcKR4V4]LbKc3U4nLcKR3\\1fKlN]3@m0a1mKeN_3Fe0c1`0[N_Od1e0ZNZOd1k0ZNTOe1o0ZNPOg1R1VNnNl1S1RNlNo1]52N1O2M20YNdFS1\\9hNjFX1U9dNPG[1P9`NVG`1a9O2O0O102M2N3N1N101N101N101N2N101N1O2N1O2N2N2N2N2N2N2K`Ya1"}, "label": "a baby horse to the right of an older horse"}]}
{"id": 89734, "image": "COCO_train2014_000000089734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young , small horse beside a much larger , darker horse\"."}], "task": "refering", "answer_template": "The \"a young , small horse beside a much larger , darker horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 64, 65, 66, 82, 83, 84, 99, 100, 101, 102, 103, 117, 118, 119, 135, 136, 137, 153, 154, 155, 171, 172, 173, 189, 190, 191, 208, 209], "bbox": [0.50806, 0.20874666666666666, 0.73064, 0.8775733333333334], "iscrowd": 0, "area": 12141.338650000002, "rle": {"size": [375, 500], "counts": "UZm24];6J7UMCgH07b0Q7IWH3=9Z7OhG6b00d7T1QHQOm7]2N3]OfLfH[3U7e0L4L4N2000O10000WO[IXLn6f3b0M3M3K5nNiG`N]8]1RHQNS8m1l0N2O16J7H7I>jH^M]4P3cJeMZ5^4M3M3M4L3M2N3M2N2NcLjJ=T5FQK4l40XKKf48^KD_4>fKeNH[N_4T3mKiM=mNd3`3PLlLT5AgJ`4i5mKgIY4W6f0N3L3N8H>A`0A:F2M2O2NOO1000[NoKcKR4V4]LbKc3U4nLcKR3\\1fKlN]3@m0a1mKeN_3Fe0c1`0[N_Od1e0ZNZOd1k0ZNTOe1o0ZNPOg1R1VNnNl1S1RNlNo1]52N1O2M20YNdFS1\\9hNjFX1U9dNPG[1P9`NVG`1a9O2O0O102M2N3N1N101N101N101N2N101N1O2N1O2N2N2N2N2N2N2K`Ya1"}, "label": "a young , small horse beside a much larger , darker horse"}]}
{"id": 89734, "image": "COCO_train2014_000000089734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown adult horse with long tail\"."}], "task": "refering", "answer_template": "The \"brown adult horse with long tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 113, 114, 115, 116, 117, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 185, 186, 187, 203, 204], "bbox": [0.25956, 0.19325333333333333, 0.57134, 0.9123466666666666], "iscrowd": 0, "area": 25314.556650000006, "rle": {"size": [375, 500], "counts": "]k_1=[7E_Lb0V3KbL:S34cL0X3;[LKd3=oKIo3`0gKCW4V1PKmNn4[1iJgNU5b1TJnNl5^3010O06K4K5L2M100O2RLPJS2P6kMRJT2P6iLWJ1LU3S6ZLdJ7[O_3e7O100O010O100O10000O100O100O10O010000O100O10SOeL^IZ3Y6UMbIk2c5UNXJj1f5\\NVJe1g5`NUJb1g5eNUJ[1i5[2M2O2N1O3M2N3M2N2N3N1N2O1O1O1N2O2N1O1O1N\\K]KX2b4eMcKZ2\\4cMhK]2W4bMkK_2R4aMQL^2n3aMTL_2m3^MULb2l3[MWLd2j3YMXLg2i3VMYLj2h3SM[Lm2e3hKeMX4k42O1O1N2O2M2O1N2O1M4L3M3M3N3J5K6K4hN_JULg5i3T1O2O0O2N100O1O010002M4M2N2M4M2N2N2M4M3N2M3NaNnIZMo5e2lJXM^4f2hKZMS4e2SLYMj3f2[LYMc3e2bLWM^3h2hLSMY3k2oLmLR3R3k2N1VOQHkMS8P2n0J6J7J5J6J7K4M3L3M2O0O1O2O0O2N1O101N3M6K5JeW^2"}, "label": "brown adult horse with long tail"}]}
{"id": 89734, "image": "COCO_train2014_000000089734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown donkey with long black tail standing showind its back in grass\"."}], "task": "refering", "answer_template": "The \"a brown donkey with long black tail standing showind its back in grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 113, 114, 115, 116, 117, 131, 132, 133, 134, 135, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 185, 186, 187, 203, 204], "bbox": [0.25956, 0.19325333333333333, 0.57134, 0.9123466666666666], "iscrowd": 0, "area": 25314.556650000006, "rle": {"size": [375, 500], "counts": "]k_1=[7E_Lb0V3KbL:S34cL0X3;[LKd3=oKIo3`0gKCW4V1PKmNn4[1iJgNU5b1TJnNl5^3010O06K4K5L2M100O2RLPJS2P6kMRJT2P6iLWJ1LU3S6ZLdJ7[O_3e7O100O010O100O10000O100O100O10O010000O100O10SOeL^IZ3Y6UMbIk2c5UNXJj1f5\\NVJe1g5`NUJb1g5eNUJ[1i5[2M2O2N1O3M2N3M2N2N3N1N2O1O1O1N2O2N1O1O1N\\K]KX2b4eMcKZ2\\4cMhK]2W4bMkK_2R4aMQL^2n3aMTL_2m3^MULb2l3[MWLd2j3YMXLg2i3VMYLj2h3SM[Lm2e3hKeMX4k42O1O1N2O2M2O1N2O1M4L3M3M3N3J5K6K4hN_JULg5i3T1O2O0O2N100O1O010002M4M2N2M4M2N2N2M4M3N2M3NaNnIZMo5e2lJXM^4f2hKZMS4e2SLYMj3f2[LYMc3e2bLWM^3h2hLSMY3k2oLmLR3R3k2N1VOQHkMS8P2n0J6J7J5J6J7K4M3L3M2O0O1O2O0O2N1O101N3M6K5JeW^2"}, "label": "a brown donkey with long black tail standing showind its back in grass"}]}
{"id": 245392, "image": "COCO_train2014_000000245392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white train\"."}], "task": "refering", "answer_template": "The \"a red and white train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218], "bbox": [0.141578125, 0.17358333333333334, 0.9771093750000001, 0.7691111111111111], "iscrowd": 0, "area": 66088.50080000001, "rle": {"size": [360, 640], "counts": "bVP11V;3N1O2N1N3N1O2N1N3RM_OQKb0n4@PKb0n4@PK`0P5BmJ?S5CkJ=U5EiJ;X5GeJ9[5IbJ8^5J`J6`5L^J4b5N\\J2d50ZJ0f54UJMk5;mIES6c0eI\\O\\6k0]IUOd6P1UIQOl6U1mHkNS7\\1fHdN[7j1VHVNj7d2N2O1N6K7H9H7H8I2M2TO_KoJc4P5aKlJ`4T5cKgJ_4X5dKeJ^4Z5eKaJ]4^5g000O100O100O2O0O100O1M3J6N3N1O1O1O11O000001O000001O000001O000001O00000001O0001O0000000001O01O0000000001O01O0000001O0001O0001O0000001O01O000001O0000010O000000001O0001O01O0O10001O000O10001O000O101O000O10001O000O101O00000O101O000O101O00000O2O0000000O2O00000O2O0000000O2O0000001N1000000O2O0000001N1000001O0O1000001O0O10001O0O1000001O0O10001O000O10001O0O10001O000O101O00000O101O000O101O00000O101O00000O2O00000O101O00000O2O0000001N1000000O2O0000001N100000001N1000001N100000001N1000001O0O1000001N1000001O0O10001O000O10001O0O10001O000O10001O000O101O00000O101O000O101O00000O2O0000000O2O00000O2O0000000O2O0000001N1000000O2O0000001N1000001O0O1000001N1000001O0O1000001O0O10001O0O1000001O0O10001O000O101O000O10001O000O101O00000O101O000O101O00000O101O00000O2O0000001N1000000O2O0000001N100000001N1000001N100000001N1000001O0O1000001N1000001O0O10001O000O10001O0O10001O000O10001O000O101O000O10001O000O101O00000O2O00000O101O00000O2O0000000O2O00000O2O0000000O2O0000001N1000001N100O10001N100O1J7I6JhS5"}, "label": "a red and white train"}]}
{"id": 245392, "image": "COCO_train2014_000000245392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large white and maroon train in the front of the photo\"."}], "task": "refering", "answer_template": "The \"large white and maroon train in the front of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 54, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218], "bbox": [0.141578125, 0.17358333333333334, 0.9771093750000001, 0.7691111111111111], "iscrowd": 0, "area": 66088.50080000001, "rle": {"size": [360, 640], "counts": "bVP11V;3N1O2N1N3N1O2N1N3RM_OQKb0n4@PKb0n4@PK`0P5BmJ?S5CkJ=U5EiJ;X5GeJ9[5IbJ8^5J`J6`5L^J4b5N\\J2d50ZJ0f54UJMk5;mIES6c0eI\\O\\6k0]IUOd6P1UIQOl6U1mHkNS7\\1fHdN[7j1VHVNj7d2N2O1N6K7H9H7H8I2M2TO_KoJc4P5aKlJ`4T5cKgJ_4X5dKeJ^4Z5eKaJ]4^5g000O100O100O2O0O100O1M3J6N3N1O1O1O11O000001O000001O000001O000001O00000001O0001O0000000001O01O0000000001O01O0000001O0001O0001O0000001O01O000001O0000010O000000001O0001O01O0O10001O000O10001O000O101O000O10001O000O101O00000O101O000O101O00000O2O0000000O2O00000O2O0000000O2O0000001N1000000O2O0000001N1000001O0O1000001O0O10001O0O1000001O0O10001O000O10001O0O10001O000O101O00000O101O000O101O00000O101O00000O2O00000O101O00000O2O0000001N1000000O2O0000001N100000001N1000001N100000001N1000001O0O1000001N1000001O0O10001O000O10001O0O10001O000O10001O000O101O00000O101O000O101O00000O2O0000000O2O00000O2O0000000O2O0000001N1000000O2O0000001N1000001O0O1000001N1000001O0O1000001O0O10001O0O1000001O0O10001O000O101O000O10001O000O101O00000O101O000O101O00000O101O00000O2O0000001N1000000O2O0000001N100000001N1000001N100000001N1000001O0O1000001N1000001O0O10001O000O10001O0O10001O000O10001O000O101O000O10001O000O101O00000O2O00000O101O00000O2O0000000O2O00000O2O0000000O2O0000001N1000001N100O10001N100O1J7I6JhS5"}, "label": "large white and maroon train in the front of the photo"}]}
{"id": 302740, "image": "COCO_train2014_000000302740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sub sitting on table\"."}], "task": "refering", "answer_template": "The \"sub sitting on table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [293, 294, 309, 310, 311, 326, 327, 328, 342, 343, 344, 345, 359, 360, 361, 362, 376, 377, 378], "bbox": [0.11085416666666667, 0.761421875, 0.3595416666666667, 0.9883906249999999], "iscrowd": 0, "area": 9465.09165, "rle": {"size": [640, 480], "counts": "cgQ11mc02N2N3M2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2N3M2N2N2M3N2N2N2N2N2N3M2N2N2N2N2O1N2N2N2N3M1O2O1N2N2N2N2N2N2N2N1O2M300O1O1O100O1O001O100O1O1O100O1O1O010O1O1O1O1O1O1O1O010O1OO1N2M3N2M5L4K4M4K4M4K5K4M4K5L3L5L3L5L4K4M4K4M4K5L3L5L4K4L5L3L5L4K4MZPP6"}, "label": "sub sitting on table"}]}
{"id": 302740, "image": "COCO_train2014_000000302740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bun on the plate\"."}], "task": "refering", "answer_template": "The \"a bun on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [293, 294, 309, 310, 311, 326, 327, 328, 342, 343, 344, 345, 359, 360, 361, 362, 376, 377, 378], "bbox": [0.11085416666666667, 0.761421875, 0.3595416666666667, 0.9883906249999999], "iscrowd": 0, "area": 9465.09165, "rle": {"size": [640, 480], "counts": "cgQ11mc02N2N3M2N2N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2N3M2N2N2M3N2N2N2N2N2N3M2N2N2N2N2O1N2N2N2N3M1O2O1N2N2N2N2N2N2N2N1O2M300O1O1O100O1O001O100O1O1O100O1O1O010O1O1O1O1O1O1O1O010O1OO1N2M3N2M5L4K4M4K4M4K5K4M4K5L3L5L3L5L4K4M4K4M4K5L3L5L4K4L5L3L5L4K4MZPP6"}, "label": "a bun on the plate"}]}
{"id": 302740, "image": "COCO_train2014_000000302740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"arm of the chair\"."}], "task": "refering", "answer_template": "The \"arm of the chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 269, 270, 286, 338, 339, 355, 356, 371, 372, 389], "bbox": [0.8277083333333334, 0.64415625, 0.9800833333333333, 0.97921875], "iscrowd": 0, "area": 2982.6850999999997, "rle": {"size": [640, 480], "counts": "ebh71^c0`0H90O0000001O01O01O00000010O000001O0001YAYOT:g0kE]OS:b0lEAR:?mEDQ:<nEHP:7oELo95oENo92PF2m9NRF5m9JRF9l9GRF>k9BTFa0j9@TFc0k9\\OTFh0i9XOVFk0h9UOWFn0g9ROXFR1f9mNYFV1e9jNZFY1d9hNYF]1e9bNZF`1e9`NZF^1i9bNVF\\1m9dNRFZ1R:eNmEZ1U:gNiEW1\\:gNcEW1b:gN]EW1g:hNWEX1m:fNREX1l>M3M4N2M2N3N2M2O2M3N1N3N2M2O2M2O2M3M2O2M3N1N3Nag5"}, "label": "arm of the chair"}]}
{"id": 302740, "image": "COCO_train2014_000000302740.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"handle of the chair were the woman is sitting\"."}], "task": "refering", "answer_template": "The \"handle of the chair were the woman is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 269, 270, 286, 338, 339, 355, 356, 371, 372, 389], "bbox": [0.8277083333333334, 0.64415625, 0.9800833333333333, 0.97921875], "iscrowd": 0, "area": 2982.6850999999997, "rle": {"size": [640, 480], "counts": "ebh71^c0`0H90O0000001O01O01O00000010O000001O0001YAYOT:g0kE]OS:b0lEAR:?mEDQ:<nEHP:7oELo95oENo92PF2m9NRF5m9JRF9l9GRF>k9BTFa0j9@TFc0k9\\OTFh0i9XOVFk0h9UOWFn0g9ROXFR1f9mNYFV1e9jNZFY1d9hNYF]1e9bNZF`1e9`NZF^1i9bNVF\\1m9dNRFZ1R:eNmEZ1U:gNiEW1\\:gNcEW1b:gN]EW1g:hNWEX1m:fNREX1l>M3M4N2M2N3N2M2O2M3N1N3N2M2O2M2O2M3M2O2M3N1N3Nag5"}, "label": "handle of the chair were the woman is sitting"}]}
{"id": 114326, "image": "COCO_train2014_000000114326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dining table of the kid\"."}], "task": "refering", "answer_template": "The \"dining table of the kid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 308, 309, 310, 311, 312, 325, 326, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 384, 385, 389, 390], "bbox": [0.006734374999999999, 0.7910208333333333, 0.9994375, 0.9887708333333334], "iscrowd": 0, "area": 26172.436499999993, "rle": {"size": [480, 640], "counts": "Xi24k>4M3M3L5L1N01000O100O10000O100O10000O10O010000O10000O100O10000O100O10000O100O10000O100O10000O10000O100O2O000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O1001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00mNTOSDk0m;WOQDi0n;YOQDg0o;ZOPDf0o;\\OPDd0P<^OnCb0R<@lC`0S<CkC=U<EiC;V<IgC7X<LfC4Y<OeC1Y<4dCL[<7cCI[<;cCE\\<>bCB]<b0_C_O`<a100O100O100O100O100O100O10000O100O1000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000001O0000000O10000000000000000000000000000000000000000000000l0TOfaU1XNS`jN0O10O1O1N2O001O1O00000000O1000000000000O1000000000000O10O10iWY13ThfN4L2N0O10000O10000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O100001O4L4L4M3L4L4LVA"}, "label": "dining table of the kid"}]}
{"id": 114326, "image": "COCO_train2014_000000114326.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table near a boy\"."}], "task": "refering", "answer_template": "The \"a wooden table near a boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [306, 307, 308, 309, 310, 311, 312, 325, 326, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 384, 385, 389, 390], "bbox": [0.006734374999999999, 0.7910208333333333, 0.9994375, 0.9887708333333334], "iscrowd": 0, "area": 26172.436499999993, "rle": {"size": [480, 640], "counts": "Xi24k>4M3M3L5L1N01000O100O10000O100O10000O10O010000O10000O100O10000O100O10000O100O10000O100O10000O10000O100O2O000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O1001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O00mNTOSDk0m;WOQDi0n;YOQDg0o;ZOPDf0o;\\OPDd0P<^OnCb0R<@lC`0S<CkC=U<EiC;V<IgC7X<LfC4Y<OeC1Y<4dCL[<7cCI[<;cCE\\<>bCB]<b0_C_O`<a100O100O100O100O100O100O10000O100O1000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000001O0000000O10000000000000000000000000000000000000000000000l0TOfaU1XNS`jN0O10O1O1N2O001O1O00000000O1000000000000O1000000000000O10O10iWY13ThfN4L2N0O10000O10000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O100001O4L4L4M3L4L4LVA"}, "label": "a wooden table near a boy"}]}
{"id": 286359, "image": "COCO_train2014_000000286359.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second commode from the left\"."}], "task": "refering", "answer_template": "The \"the second commode from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 327, 328, 329, 330], "bbox": [0.22753125000000002, 0.4651666666666667, 0.37584375000000003, 0.8786458333333333], "iscrowd": 0, "area": 15531.153600000001, "rle": {"size": [480, 640], "counts": "UhT2o0R;YOcGm0U8_ObGe0Z8E[G>d8OmF4R9?ZFCe9g0oE\\OP:e0nE]OQ:d0mE^OR:b0mE@R:a0lEAS:`0kEAU:`0iEBV:>iEDV:=hEEY1nM`7^2UGFZ1lMa7f4]H[Kc7h4iG^KCJd8k4[GgK0_Od8g5ZGZJf8g5YGYJg8h5WGYJi8h5VGYJi8g5WGYJi8o50000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1O1O1aLTGC3kNj8b1TG_O9jNd8g1TG[O`0gN^8n1SGWOf0fNX8S2SGSOm0dNQ8Y2SGoNS1cNk7^2SGlNX1aNe7d2TGgN_:Y1cEbN`:^1aE_Na:`1aE\\Nb:c1`EZNb:e1`EWNc:h1_EUNc:i1cEPN`:o1gEhM\\:W2W1N3N:F;D_ij5"}, "label": "the second commode from the left"}]}
{"id": 286359, "image": "COCO_train2014_000000286359.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toilet with no lid in between two others\"."}], "task": "refering", "answer_template": "The \"a toilet with no lid in between two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 327, 328, 329, 330], "bbox": [0.22753125000000002, 0.4651666666666667, 0.37584375000000003, 0.8786458333333333], "iscrowd": 0, "area": 15531.153600000001, "rle": {"size": [480, 640], "counts": "UhT2o0R;YOcGm0U8_ObGe0Z8E[G>d8OmF4R9?ZFCe9g0oE\\OP:e0nE]OQ:d0mE^OR:b0mE@R:a0lEAS:`0kEAU:`0iEBV:>iEDV:=hEEY1nM`7^2UGFZ1lMa7f4]H[Kc7h4iG^KCJd8k4[GgK0_Od8g5ZGZJf8g5YGYJg8h5WGYJi8h5VGYJi8g5WGYJi8o50000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1O1O1aLTGC3kNj8b1TG_O9jNd8g1TG[O`0gN^8n1SGWOf0fNX8S2SGSOm0dNQ8Y2SGoNS1cNk7^2SGlNX1aNe7d2TGgN_:Y1cEbN`:^1aE_Na:`1aE\\Nb:c1`EZNb:e1`EWNc:h1_EUNc:i1cEPN`:o1gEhM\\:W2W1N3N:F;D_ij5"}, "label": "a toilet with no lid in between two others"}]}
{"id": 286359, "image": "COCO_train2014_000000286359.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toilet that is very wide\"."}], "task": "refering", "answer_template": "The \"a toilet that is very wide\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361], "bbox": [0.535625, 0.42116666666666663, 0.76484375, 0.9219375], "iscrowd": 0, "area": 27697.757400000006, "rle": {"size": [480, 640], "counts": "`RQ5e0W>:H7I4M2O1eDgNl8\\1kFlNT9U1dFTOY9n0_FZO`9h0XF@e9b0SFGk9:mENQ:4gE5W:M^E<a:R201N100O2O0_MdKXH3g1[4o5SLoIm3Q6ZLWIAoNV4i7aL_Hb4_7`10001N101N10001O00001O00001O010O000000001O0000000000001O00000000001O00000000001O00000000010O00O100000000000000000000O10001O000000000000000O10001O000000000000000O101O00000000000O100O1O2O0O100O100O100_KRHm1Q8mMVHo1m7kM[HP2h7jM_H<IlNl79jHi0m8jN`GU1a8TNWHj1a:N1O2N2N1O2M3N1O2N2N1O2N2N4J8D=B=CUPV2"}, "label": "a toilet that is very wide"}]}
{"id": 286359, "image": "COCO_train2014_000000286359.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biggest toilet they ' ve got !\"."}], "task": "refering", "answer_template": "The \"the biggest toilet they ' ve got !\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361], "bbox": [0.535625, 0.42116666666666663, 0.76484375, 0.9219375], "iscrowd": 0, "area": 27697.757400000006, "rle": {"size": [480, 640], "counts": "`RQ5e0W>:H7I4M2O1eDgNl8\\1kFlNT9U1dFTOY9n0_FZO`9h0XF@e9b0SFGk9:mENQ:4gE5W:M^E<a:R201N100O2O0_MdKXH3g1[4o5SLoIm3Q6ZLWIAoNV4i7aL_Hb4_7`10001N101N10001O00001O00001O010O000000001O0000000000001O00000000001O00000000001O00000000010O00O100000000000000000000O10001O000000000000000O10001O000000000000000O101O00000000000O100O1O2O0O100O100O100_KRHm1Q8mMVHo1m7kM[HP2h7jM_H<IlNl79jHi0m8jN`GU1a8TNWHj1a:N1O2N2N1O2M3N1O2N2N1O2N2N4J8D=B=CUPV2"}, "label": "the biggest toilet they ' ve got !"}]}
{"id": 89754, "image": "COCO_train2014_000000089754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white pickup truck\"."}], "task": "refering", "answer_template": "The \"a white pickup truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 296], "bbox": [0.746296875, 0.48364583333333333, 1.0, 0.7139791666666667], "iscrowd": 0, "area": 11130.124300000001, "rle": {"size": [480, 640], "counts": "kZP76=Og=9RBKj=g0M4M3N2N1O2N1O2N1O2O00000000000000000000O110O010O010O010O010O01O001O001O00001O00001O001O1O1O1O1O1O001O1O001O001O000000O1O1O1N2O11O1O001O1O1O1ERNbCo1]<VN^Ck1a<;O1O1O1O1N2O1O001O00000000000O1000O10000O10000O1000O10O10000O1O1O1O1O00O100O1O1O2O0OH8011N2N2N2000000O10000000000000000000000000000000000000000000000000000O1000000000000001OgH"}, "label": "a white pickup truck"}]}
{"id": 89754, "image": "COCO_train2014_000000089754.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white pickup truck\"."}], "task": "refering", "answer_template": "The \"white pickup truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 296], "bbox": [0.746296875, 0.48364583333333333, 1.0, 0.7139791666666667], "iscrowd": 0, "area": 11130.124300000001, "rle": {"size": [480, 640], "counts": "kZP76=Og=9RBKj=g0M4M3N2N1O2N1O2N1O2O00000000000000000000O110O010O010O010O010O01O001O001O00001O00001O001O1O1O1O1O1O001O1O001O001O000000O1O1O1N2O11O1O001O1O1O1ERNbCo1]<VN^Ck1a<;O1O1O1O1N2O1O001O00000000000O1000O10000O10000O1000O10O10000O1O1O1O1O00O100O1O1O2O0OH8011N2N2N2000000O10000000000000000000000000000000000000000000000000000O1000000000000001OgH"}, "label": "white pickup truck"}]}
{"id": 425628, "image": "COCO_train2014_000000425628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow next to the people\"."}], "task": "refering", "answer_template": "The \"the cow next to the people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 370, 371, 372], "bbox": [0.016125, 0.23225, 0.530640625, 0.9526875], "iscrowd": 0, "area": 73274.63825000002, "rle": {"size": [480, 640], "counts": "Wl41f><D;E<E:N3M2N1O001O00001O00001O00001PI`N>`1\\OmN=S1^OUO?l0ZOZOf0f0TOAk0?POFP1;kNIU17gNNX12dN2\\1O_N5a1K]N8b1H[N;e1FXN<h1DUN`0j1@TNb0l1_OPNd0P2\\OnMg0Q2YOlMj0T2VOhMn0X2ROQMf1n2ZNnLj1R3VNkLm1U3SNgLR2X3nMdLV2\\3jMaLY2_3gM]L^2b3bMULg2k3YMkKQ3U4oLaK\\3_4cLmJQ4S5oKdJZ4\\5fKaJ^4^5bK`J`4`5bK[Ja4e5bKUJb4k5aKoIa4Q6W201O000010O01O00001O01O0001O000001O0001O000000001O000O1O100O1O100O1O1O100O1O100N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001O1O100O00100O100O010O100O10O01O100O10O0100O100O010O1O100O010O10I6J7O1O1O2N1O2M3N1O2N1O2N2O0O2O0O2O0O2O1N101N101N2O0O2O0O2O0O2O1N101N2K4I8J6O1O001N2O1O001N2O1O001N2O1O001N101O000O2O001O001O00001O0O2O001O00001O001O001O0O101O001O001O001O001O0O2O001O0O2N1O1O2O0O2N2N3M3N1N3M2N1O1O1M2M4M3M3M2O2N2M3N1O2N2000O010000RMfF6Y9FmF9S9EPG:P9FQG8P9GRG8m8ITG6l8IVG6j8JWG5i8JYG5h8JYG5g8J[G4f8K\\G4d8L]G3c8L_G3a8M`G2`8MbG2^8NcG1]8NeG1[8OfG0Z8OiGOX80iGOW80lGNT82mGMS82oGMQ83QHKo74SHKm75UHIk76WHIi77YHGg78[HGf78\\HFd79^HFb7:`HD`7;bHD^7<dHB\\7=fHBZ7=iHAW7?jH@V7?mH_OS7a0nH^OR7a0QI]OP7b0QI]Oo6b0TI\\Ol6d0UI[Ok6d0XIZOh6e0ZIZOf6e0]IYOc6f0`IXO`6f0cIXO^6g0eISO`6k0bIoNh6j0\\IPOm6j0i3K6I7J6Jhl\\4"}, "label": "the cow next to the people"}]}
{"id": 425628, "image": "COCO_train2014_000000425628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow peeking over the fence\"."}], "task": "refering", "answer_template": "The \"cow peeking over the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 370, 371, 372], "bbox": [0.016125, 0.23225, 0.530640625, 0.9526875], "iscrowd": 0, "area": 73274.63825000002, "rle": {"size": [480, 640], "counts": "Wl41f><D;E<E:N3M2N1O001O00001O00001O00001PI`N>`1\\OmN=S1^OUO?l0ZOZOf0f0TOAk0?POFP1;kNIU17gNNX12dN2\\1O_N5a1K]N8b1H[N;e1FXN<h1DUN`0j1@TNb0l1_OPNd0P2\\OnMg0Q2YOlMj0T2VOhMn0X2ROQMf1n2ZNnLj1R3VNkLm1U3SNgLR2X3nMdLV2\\3jMaLY2_3gM]L^2b3bMULg2k3YMkKQ3U4oLaK\\3_4cLmJQ4S5oKdJZ4\\5fKaJ^4^5bK`J`4`5bK[Ja4e5bKUJb4k5aKoIa4Q6W201O000010O01O00001O01O0001O000001O0001O000000001O000O1O100O1O100O1O1O100O1O100N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001O1O100O00100O100O010O100O10O01O100O10O0100O100O010O1O100O010O10I6J7O1O1O2N1O2M3N1O2N1O2N2O0O2O0O2O0O2O1N101N101N2O0O2O0O2O0O2O1N101N2K4I8J6O1O001N2O1O001N2O1O001N2O1O001N101O000O2O001O001O00001O0O2O001O00001O001O001O0O101O001O001O001O001O0O2O001O0O2N1O1O2O0O2N2N3M3N1N3M2N1O1O1M2M4M3M3M2O2N2M3N1O2N2000O010000RMfF6Y9FmF9S9EPG:P9FQG8P9GRG8m8ITG6l8IVG6j8JWG5i8JYG5h8JYG5g8J[G4f8K\\G4d8L]G3c8L_G3a8M`G2`8MbG2^8NcG1]8NeG1[8OfG0Z8OiGOX80iGOW80lGNT82mGMS82oGMQ83QHKo74SHKm75UHIk76WHIi77YHGg78[HGf78\\HFd79^HFb7:`HD`7;bHD^7<dHB\\7=fHBZ7=iHAW7?jH@V7?mH_OS7a0nH^OR7a0QI]OP7b0QI]Oo6b0TI\\Ol6d0UI[Ok6d0XIZOh6e0ZIZOf6e0]IYOc6f0`IXO`6f0cIXO^6g0eISO`6k0bIoNh6j0\\IPOm6j0i3K6I7J6Jhl\\4"}, "label": "cow peeking over the fence"}]}
{"id": 425628, "image": "COCO_train2014_000000425628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child in rainbow hoodie being held\"."}], "task": "refering", "answer_template": "The \"a small child in rainbow hoodie being held\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 250, 251, 265, 266, 267, 268, 269, 270, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 387, 388, 389], "bbox": [0.497890625, 0.19568750000000001, 0.9528906249999999, 1.0], "iscrowd": 0, "area": 53895.460050000016, "rle": {"size": [480, 640], "counts": "Yie43l>3M4L3M4L3M4M2M4L3M4L3N3M4L3M3M4L2N1O1O1N2O0O2O1O1N2O0O2001O000O100000RGaNZ4^1nJZOS5f0oI7S6HaIc0`6\\O^If0d6YOZIh0h6WOUIk0m6SOQIP1o6POoHQ1S7mNkHU1W7jNfHX1\\7fNcH[1^7eN`H\\1^7fNaH[1Z7kNeHU1U7QOkHo0o6XOoHi0k6]OUIc0f6CXI?a6F`I:V3iLjNQ3PN6P3RMlNi2SN6m2WMnNc2UN6l2[MZM[O8T3C6j2_MUMZO>Q3B7j2bMPMZOb0n2D6j2hMfLYOk0h2F7h2PN]LTOT1f2F7h2iN`MQ1H6g2jNaMP1H7f2jN`Mo0J8e2kN`Mm0K8d2mN_Ml0M7c2PO^Mi0O7c2RO[Mg037a2UOZMd053d2\\OTMb09Ne2CPM?;Ji2IiL>>El20dL;`0Ao28]L9d0ZOR3`0XL7e0UOW3f0QL8f0nN\\3n0jK6i0hN`3T1fK6i0bNc3[1aK5k0\\Nh3a1[K3m0XNk3\\5VL`Jm3a5TLZJo3g5QLWJQ4i5oKXJP4h5QLWJo3h5RLXJm3i5SLXJl3h5TLXJl3h5ULWJk3h5VLXJj3h5VLYJh3g5YLYJg3g5ZLXJf3h5ZLYJe3f5\\LZJc3g5]LYJc3f5^LZJb3f5_LZJ`3f5`LZJ_3f5bLZJ^3f5bL[J]3e5dLZJ\\3g5cLYJ]3h5bLYJ\\3h5dLXJ\\3i5dLVJ\\3k5cLUJ]3l5bLUJ\\3m5cLSJ]3o5aLQJ_3Q6`LoI_3S6_LmIa3U6]LkIb3X6\\LiIc3Y6\\LfId3\\6ZLdIf3\\6ZLdIf3]6YLdIe3^6[L`If3a6YL_Ig3b6XL^Ih3c6WL]Ii3d6VL\\Ii3f6WLYIi3g6WLXIj3h6VLXIj3g6WLYIh3h6YLWIg3h6ZLXIf3h6ZLXIf3g6[LXIf3h6[LWId3i6]LWIc3i6]LWIc3h6^LXIb3h6^LXIc3f6_LYIa3f6`LYIb3f6^LZIb3e6_L[Ib3d6^L\\Ib3c6_L]Ia3c6_L]Ib3a6_L^Ib3b6^L^Ic3`6^L`Ib3`6^L`Ic3^6^LbIb3^6^LbIc3\\6^LdIb3\\6^LeIa3Z6`LfIa3Y6_LgIb3W6_LiIa3V6`LjIa3T6`LlIa3R6`LnIa3o5aLQJ`3m5aLSJ`3X1TKm1]1kL`3W1SKl1_1mL^3V1TKl1_1nL^3T1UKl1^1PM^3R1UKl1_1RM]3Q1TKl1`1SM]3o0TKm1`1TM]3m0UKm1_1VM\\3l0VKl1`1XM[3i0WKn1_1YM[3f0XKP2^1ZM[3c0ZKP2]1]MZ3`0[KR2\\1^MZ3=\\KT2[1_MZ3:]KU2[1aMX38`KU2Y1cM]30\\K]2W1cMd3GWKe2V1dMi3_OSKl2U1eMo3VOoJS3S1gMT4nNkJZ3R1hMZ4eNfJb3Q1iMa5V2`JjMa5T2`JlM`5S2aJmM_5S2aJmM_5R2bJnM_5Q2aJoM_5P2bJPN^5n1dJRN\\5j1hJVNX5e1mJ[NT5`1PK`NP5\\1TKdNl4X1XKhNh4U1[KkNe4S1\\KnNd4P1^KPOb4n0`KROa4k0`KVO`4h0bKXO^4f0dKZO\\4d0fK]OY4a0hK@Y4=iKDV4<jKDW4;hKGW49iKGX48hKIW47hKJY45gKLX44hKLY43fKOY41gKOZ40fK1Y4OfK2[4MeK4Z4MeK3\\4MbK6\\4JdK8[4HdK:Z4GeK;Z4DeK`0X4AgKa0X4_OgKc0W4^OgKe0X4ZOhKh0V4YOiKi0V4WOjKk0S4UOnKl0Q4TOoKm0o3TOQLm0n3SORLn0m3QOTLP1j3QOVLQ1h3oNXLR1g3mNZLT1d3mN\\LR1e3nN[LQ1f3nN[LQ1e3PO[Lo0f3QOZLn0g3ROYLm0g3SOZLl0g3TOYLk0h3UOXLj0h3XOWLg0j3YOVLf0k3ZOULe0k3\\OULc0l3]OTLb0m3^ORLb0o3^OQLa0o3APL>P4CPL<P4EPL:P4GPL8Q4HoK7Q4JoK5Q4JQL5o3KRL4n3KUL3l3LUL3k3MWL1i3NZL0f30\\LNd31_LMa33aLK`34bLJ_34cLK`32bLLa31aLMb3OaLOb3N`L0d3K_L3d3FbL8b3@cL?`3ZOfLd0n8M4L3M3M4L3M3M4L3M3M4L3MQc="}, "label": "a small child in rainbow hoodie being held"}]}
{"id": 425628, "image": "COCO_train2014_000000425628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child\"."}], "task": "refering", "answer_template": "The \"the child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 250, 251, 265, 266, 267, 268, 269, 270, 291, 292, 293, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 387, 388, 389], "bbox": [0.497890625, 0.19568750000000001, 0.9528906249999999, 1.0], "iscrowd": 0, "area": 53895.460050000016, "rle": {"size": [480, 640], "counts": "Yie43l>3M4L3M4L3M4M2M4L3M4L3N3M4L3M3M4L2N1O1O1N2O0O2O1O1N2O0O2001O000O100000RGaNZ4^1nJZOS5f0oI7S6HaIc0`6\\O^If0d6YOZIh0h6WOUIk0m6SOQIP1o6POoHQ1S7mNkHU1W7jNfHX1\\7fNcH[1^7eN`H\\1^7fNaH[1Z7kNeHU1U7QOkHo0o6XOoHi0k6]OUIc0f6CXI?a6F`I:V3iLjNQ3PN6P3RMlNi2SN6m2WMnNc2UN6l2[MZM[O8T3C6j2_MUMZO>Q3B7j2bMPMZOb0n2D6j2hMfLYOk0h2F7h2PN]LTOT1f2F7h2iN`MQ1H6g2jNaMP1H7f2jN`Mo0J8e2kN`Mm0K8d2mN_Ml0M7c2PO^Mi0O7c2RO[Mg037a2UOZMd053d2\\OTMb09Ne2CPM?;Ji2IiL>>El20dL;`0Ao28]L9d0ZOR3`0XL7e0UOW3f0QL8f0nN\\3n0jK6i0hN`3T1fK6i0bNc3[1aK5k0\\Nh3a1[K3m0XNk3\\5VL`Jm3a5TLZJo3g5QLWJQ4i5oKXJP4h5QLWJo3h5RLXJm3i5SLXJl3h5TLXJl3h5ULWJk3h5VLXJj3h5VLYJh3g5YLYJg3g5ZLXJf3h5ZLYJe3f5\\LZJc3g5]LYJc3f5^LZJb3f5_LZJ`3f5`LZJ_3f5bLZJ^3f5bL[J]3e5dLZJ\\3g5cLYJ]3h5bLYJ\\3h5dLXJ\\3i5dLVJ\\3k5cLUJ]3l5bLUJ\\3m5cLSJ]3o5aLQJ_3Q6`LoI_3S6_LmIa3U6]LkIb3X6\\LiIc3Y6\\LfId3\\6ZLdIf3\\6ZLdIf3]6YLdIe3^6[L`If3a6YL_Ig3b6XL^Ih3c6WL]Ii3d6VL\\Ii3f6WLYIi3g6WLXIj3h6VLXIj3g6WLYIh3h6YLWIg3h6ZLXIf3h6ZLXIf3g6[LXIf3h6[LWId3i6]LWIc3i6]LWIc3h6^LXIb3h6^LXIc3f6_LYIa3f6`LYIb3f6^LZIb3e6_L[Ib3d6^L\\Ib3c6_L]Ia3c6_L]Ib3a6_L^Ib3b6^L^Ic3`6^L`Ib3`6^L`Ic3^6^LbIb3^6^LbIc3\\6^LdIb3\\6^LeIa3Z6`LfIa3Y6_LgIb3W6_LiIa3V6`LjIa3T6`LlIa3R6`LnIa3o5aLQJ`3m5aLSJ`3X1TKm1]1kL`3W1SKl1_1mL^3V1TKl1_1nL^3T1UKl1^1PM^3R1UKl1_1RM]3Q1TKl1`1SM]3o0TKm1`1TM]3m0UKm1_1VM\\3l0VKl1`1XM[3i0WKn1_1YM[3f0XKP2^1ZM[3c0ZKP2]1]MZ3`0[KR2\\1^MZ3=\\KT2[1_MZ3:]KU2[1aMX38`KU2Y1cM]30\\K]2W1cMd3GWKe2V1dMi3_OSKl2U1eMo3VOoJS3S1gMT4nNkJZ3R1hMZ4eNfJb3Q1iMa5V2`JjMa5T2`JlM`5S2aJmM_5S2aJmM_5R2bJnM_5Q2aJoM_5P2bJPN^5n1dJRN\\5j1hJVNX5e1mJ[NT5`1PK`NP5\\1TKdNl4X1XKhNh4U1[KkNe4S1\\KnNd4P1^KPOb4n0`KROa4k0`KVO`4h0bKXO^4f0dKZO\\4d0fK]OY4a0hK@Y4=iKDV4<jKDW4;hKGW49iKGX48hKIW47hKJY45gKLX44hKLY43fKOY41gKOZ40fK1Y4OfK2[4MeK4Z4MeK3\\4MbK6\\4JdK8[4HdK:Z4GeK;Z4DeK`0X4AgKa0X4_OgKc0W4^OgKe0X4ZOhKh0V4YOiKi0V4WOjKk0S4UOnKl0Q4TOoKm0o3TOQLm0n3SORLn0m3QOTLP1j3QOVLQ1h3oNXLR1g3mNZLT1d3mN\\LR1e3nN[LQ1f3nN[LQ1e3PO[Lo0f3QOZLn0g3ROYLm0g3SOZLl0g3TOYLk0h3UOXLj0h3XOWLg0j3YOVLf0k3ZOULe0k3\\OULc0l3]OTLb0m3^ORLb0o3^OQLa0o3APL>P4CPL<P4EPL:P4GPL8Q4HoK7Q4JoK5Q4JQL5o3KRL4n3KUL3l3LUL3k3MWL1i3NZL0f30\\LNd31_LMa33aLK`34bLJ_34cLK`32bLLa31aLMb3OaLOb3N`L0d3K_L3d3FbL8b3@cL?`3ZOfLd0n8M4L3M3M4L3M3M4L3M3M4L3MQc="}, "label": "the child"}]}
{"id": 425628, "image": "COCO_train2014_000000425628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman holds a girl by a cow\"."}], "task": "refering", "answer_template": "The \"a young woman holds a girl by a cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 83, 84, 85, 86, 87, 109, 110, 111, 133, 134, 155, 156, 157, 158, 179, 180, 181, 204, 225, 226, 248, 249, 250, 265, 266, 267, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 313, 314, 315, 316, 317, 318], "bbox": [0.42409375000000005, 0.13839583333333336, 0.9066562499999999, 0.8054791666666666], "iscrowd": 0, "area": 21663.7819, "rle": {"size": [480, 640], "counts": "k[o31o>:F:G2M1O1O010O1O00100O1O00100O00100O00O101N100O101O0O100O2O0O10001N10O001O00100O001O001O001O10O01O001O001O001O10O01O001O001O1O010O001O001O1O0010O2O0O2O001N2O0O2O0O2O001N101N101N2O001N101N101N101O1N101N101N101O0O01O3N9F:F3N0O2O0O2N2O0O2N20O1O1O1O010O1O1O]H[Ol0d0TO[On0e0POZOS1e0lN[OU1e0jN[OW1e0hN\\OX1d0gN]OX1d0hN\\OX1e0fN\\OZ1d0eN\\O[1e0dN\\O\\1d0cN]O]1c0bN^O]1d0bN\\O^1d0aN]O_1c0`N^O_1c0`N^O`1b0_N_Oa1a0_N_O`1c0^N^Ob1b0]N_Oc1a0\\N@d1`0[NAd1`0[NAe1?[N@f1a0XN@i1?VNBk1=UNCl1<TNDn1:RNFo1:PNFQ29nMHS27mMIU25kMKV24jMLW23iMMY22fMM\\22cMO^20bM0`2N`M2a2M_M3b2M]M3d2L[M5g2IYM7h2HXM8i2GWM9k2EUM;l2ERM<o2CQM=n2DRM;o2EQM;n2FRM:m2GRM:n2GQM9n2HRM8n2HRM8m2ISM7l2JSM7m2JRM6m2KSM5m2KSM5l2LTM4k2MTM3m2MSM3l2OSM1l20TM0l20TM1j20VM0j20UM2i2OWM1h21WM0h20XM1f20ZM0f20YM2e2O\\M0c22]MNb22_MM`24aML^24cMK\\26dMKZ26gMIX29hMGV2:kMFR2;PNDo1=RNCl1>TNBk1?VNAg1a0ZN^OXMQOh1b1Q1]OTMSOj1a1S1[OQMVOk1`1U1ZOlLZOl1^1X1YOiL[On1]1Z1WOfL^Oo1\\1\\1VObL@Q2[1^1TO_LDP2Z1b1RO[LFR2Y1d1POXLIS2X1f1oNTLKU2W1g1nNRLMV2V1i1mNnK0W2S1m1lNjK3X2R1o1VO[KJf2P1P2AlJAS3o0R2K]JYO_3m0T2X1kMiNV2V1iMkNX2T1hMlNY2T1eMmN\\2R1cMoN^2P1^MTOb2l0YMYOh2g0RM^Oo2a0lLDU3;fLJ[35`L0a3O[L5e3LYL4i3KVL6k3ITL8m3GRL:o3FoK:S4ElK<T4DkK=V4BiK>Y4BeK?\\4@cKa0^4^OaKb0a4]O^Kd0b4\\O]Kd0f4[OYKe0j4XOVKh0m4UOSKj0Q5SOnJn0U5POjJP1Y5mNgJR1\\5lNdJT1_5jN`JV1c5gN\\JY1h5eNWJ[1l5bNUJ]1n5aNRJ]1Q6bNoI]1T6`NmI_1V6_NjI_1Z6_NfI`1\\6^NdIa1_6^NaIa1`6^NaIa1a6^N_I`1c6`N]I_1e6_N\\I`1e6`N[I^1h6aNWI_1j6`NWI_1k6`NUI^1m6aNTI^1n6aNRI^1P7aNPI]1R7bNoH]1S7bNmH\\1U7cNkH]1W7bNiH]1X7cNhH[1[7cNfH\\1\\7cNdH\\1^7cNbH[1b7bN_H\\1e7bNZH]1j7`NWH]1m7bNSH\\1Q8bNoG\\1U8aNlG\\1Y8bNgG\\1\\8cNdG[1`8bN`G\\1e8bN[G\\1h8bNYG\\1k8bNUG8S:Ff2M2Oglk0"}, "label": "a young woman holds a girl by a cow"}]}
{"id": 425628, "image": "COCO_train2014_000000425628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue shirt holding a toddler\"."}], "task": "refering", "answer_template": "The \"a woman in a blue shirt holding a toddler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 83, 84, 85, 86, 87, 109, 110, 111, 133, 134, 155, 156, 157, 158, 179, 180, 181, 204, 225, 226, 248, 249, 250, 265, 266, 267, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 309, 310, 313, 314, 315, 316, 317, 318], "bbox": [0.42409375000000005, 0.13839583333333336, 0.9066562499999999, 0.8054791666666666], "iscrowd": 0, "area": 21663.7819, "rle": {"size": [480, 640], "counts": "k[o31o>:F:G2M1O1O010O1O00100O1O00100O00100O00O101N100O101O0O100O2O0O10001N10O001O00100O001O001O001O10O01O001O001O001O10O01O001O001O1O010O001O001O1O0010O2O0O2O001N2O0O2O0O2O001N101N101N2O001N101N101N101O1N101N101N101O0O01O3N9F:F3N0O2O0O2N2O0O2N20O1O1O1O010O1O1O]H[Ol0d0TO[On0e0POZOS1e0lN[OU1e0jN[OW1e0hN\\OX1d0gN]OX1d0hN\\OX1e0fN\\OZ1d0eN\\O[1e0dN\\O\\1d0cN]O]1c0bN^O]1d0bN\\O^1d0aN]O_1c0`N^O_1c0`N^O`1b0_N_Oa1a0_N_O`1c0^N^Ob1b0]N_Oc1a0\\N@d1`0[NAd1`0[NAe1?[N@f1a0XN@i1?VNBk1=UNCl1<TNDn1:RNFo1:PNFQ29nMHS27mMIU25kMKV24jMLW23iMMY22fMM\\22cMO^20bM0`2N`M2a2M_M3b2M]M3d2L[M5g2IYM7h2HXM8i2GWM9k2EUM;l2ERM<o2CQM=n2DRM;o2EQM;n2FRM:m2GRM:n2GQM9n2HRM8n2HRM8m2ISM7l2JSM7m2JRM6m2KSM5m2KSM5l2LTM4k2MTM3m2MSM3l2OSM1l20TM0l20TM1j20VM0j20UM2i2OWM1h21WM0h20XM1f20ZM0f20YM2e2O\\M0c22]MNb22_MM`24aML^24cMK\\26dMKZ26gMIX29hMGV2:kMFR2;PNDo1=RNCl1>TNBk1?VNAg1a0ZN^OXMQOh1b1Q1]OTMSOj1a1S1[OQMVOk1`1U1ZOlLZOl1^1X1YOiL[On1]1Z1WOfL^Oo1\\1\\1VObL@Q2[1^1TO_LDP2Z1b1RO[LFR2Y1d1POXLIS2X1f1oNTLKU2W1g1nNRLMV2V1i1mNnK0W2S1m1lNjK3X2R1o1VO[KJf2P1P2AlJAS3o0R2K]JYO_3m0T2X1kMiNV2V1iMkNX2T1hMlNY2T1eMmN\\2R1cMoN^2P1^MTOb2l0YMYOh2g0RM^Oo2a0lLDU3;fLJ[35`L0a3O[L5e3LYL4i3KVL6k3ITL8m3GRL:o3FoK:S4ElK<T4DkK=V4BiK>Y4BeK?\\4@cKa0^4^OaKb0a4]O^Kd0b4\\O]Kd0f4[OYKe0j4XOVKh0m4UOSKj0Q5SOnJn0U5POjJP1Y5mNgJR1\\5lNdJT1_5jN`JV1c5gN\\JY1h5eNWJ[1l5bNUJ]1n5aNRJ]1Q6bNoI]1T6`NmI_1V6_NjI_1Z6_NfI`1\\6^NdIa1_6^NaIa1`6^NaIa1a6^N_I`1c6`N]I_1e6_N\\I`1e6`N[I^1h6aNWI_1j6`NWI_1k6`NUI^1m6aNTI^1n6aNRI^1P7aNPI]1R7bNoH]1S7bNmH\\1U7cNkH]1W7bNiH]1X7cNhH[1[7cNfH\\1\\7cNdH\\1^7cNbH[1b7bN_H\\1e7bNZH]1j7`NWH]1m7bNSH\\1Q8bNoG\\1U8aNlG\\1Y8bNgG\\1\\8cNdG[1`8bN`G\\1e8bN[G\\1h8bNYG\\1k8bNUG8S:Ff2M2Oglk0"}, "label": "a woman in a blue shirt holding a toddler"}]}
{"id": 138910, "image": "COCO_train2014_000000138910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white couch with cream pillows\"."}], "task": "refering", "answer_template": "The \"a white couch with cream pillows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318], "bbox": [0.5964375000000001, 0.5603541666666667, 0.8647656250000001, 0.8189583333333333], "iscrowd": 0, "area": 14681.525199999998, "rle": {"size": [480, 640], "counts": "h[c54i>5K6J5ZOf0@a0M2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2NF:G90000000000000001O00000000000000000000000000000000000000000O100000000000000000000000000000000000001O001O1O001O001O001O001O001N101O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O010O010O010O010O010O010O010O0101N100O100O2O0O100OM4I6I8H7I8H7J6I7I7I7I__X1"}, "label": "a white couch with cream pillows"}]}
{"id": 138910, "image": "COCO_train2014_000000138910.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white couch on the right\"."}], "task": "refering", "answer_template": "The \"the white couch on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318], "bbox": [0.5964375000000001, 0.5603541666666667, 0.8647656250000001, 0.8189583333333333], "iscrowd": 0, "area": 14681.525199999998, "rle": {"size": [480, 640], "counts": "h[c54i>5K6J5ZOf0@a0M2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2NF:G90000000000000001O00000000000000000000000000000000000000000O100000000000000000000000000000000000001O001O1O001O001O001O001O001N101O001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O010O010O010O010O010O010O010O0101N100O100O2O0O100OM4I6I8H7I8H7J6I7I7I7I__X1"}, "label": "the white couch on the right"}]}
{"id": 212641, "image": "COCO_train2014_000000212641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man\"."}], "task": "refering", "answer_template": "The \"a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 28, 29, 30, 31, 32, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.042333333333333334, 0.5722343750000001, 0.9858958333333334], "iscrowd": 0, "area": 119913.89785000001, "rle": {"size": [480, 640], "counts": "i8n0P>n1SNm1SN[1eN1M3N2N2N2M3N2N2N2M3N2N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2N2M3N200O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O1K5K5J6K5K5J6K5J6K5O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N200O100000000000000000000O10000000000000000000000O100000000000000000000O1001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O1O1O2N1O1O1O2N1O1O1O2L3L4L5J5L4L6J6J5K6J6I6M4L4L3M4M3L3M4L4L3M4L4M2M4Ld0\\O3M3M3M4M2M3M3M3M3M4L3N2M3M3M4L3M3M3N2M4L3M3M3Ne0ZO9G2O2M3M3N1N3M3N2M2N3N2M3M2O2M3M3N1N3M3N2M2O2M3M3N1N3M3N2M2N3L4L4M3L4M:E;E;F:E;F:E:F;F:E;FoRP4"}, "label": "a man"}]}
{"id": 212641, "image": "COCO_train2014_000000212641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a hood on his head\"."}], "task": "refering", "answer_template": "The \"a man wearing a hood on his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 28, 29, 30, 31, 32, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.042333333333333334, 0.5722343750000001, 0.9858958333333334], "iscrowd": 0, "area": 119913.89785000001, "rle": {"size": [480, 640], "counts": "i8n0P>n1SNm1SN[1eN1M3N2N2N2M3N2N2N2M3N2N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2N2M3N200O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O10000O100O100O100O10000O1K5K5J6K5K5J6K5J6K5O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N200O100000000000000000000O10000000000000000000000O100000000000000000000O1001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O1O1O2N1O1O1O2N1O1O1O2L3L4L5J5L4L6J6J5K6J6I6M4L4L3M4M3L3M4L4L3M4L4M2M4Ld0\\O3M3M3M4M2M3M3M3M3M4L3N2M3M3M4L3M3M3N2M4L3M3M3Ne0ZO9G2O2M3M3N1N3M3N2M2N3N2M3M2O2M3M3N1N3M3N2M2O2M3M3N1N3M3N2M2N3L4L4M3L4M:E;E;F:E;F:E:F;F:E;FoRP4"}, "label": "a man wearing a hood on his head"}]}
{"id": 212641, "image": "COCO_train2014_000000212641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey haired woman\"."}], "task": "refering", "answer_template": "The \"a grey haired woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.42953125, 0.29774999999999996, 1.0, 1.0], "iscrowd": 0, "area": 80616.7478, "rle": {"size": [480, 640], "counts": "][Q4a0m=b0O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N200O10000O100O10000O100O10000O10000O100O10000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O100O100O100O1ZN`JlI`5R6QK^IP5`6_KSIa4k6k1N2M3N2N2N2N2N2N2N2M3N2N2N1O2N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1kGfMb3jK"}, "label": "a grey haired woman"}]}
{"id": 212641, "image": "COCO_train2014_000000212641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman eating a sandwich thats wrapper in paper\"."}], "task": "refering", "answer_template": "The \"a woman eating a sandwich thats wrapper in paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.42953125, 0.29774999999999996, 1.0, 1.0], "iscrowd": 0, "area": 80616.7478, "rle": {"size": [480, 640], "counts": "][Q4a0m=b0O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1N200O10000O100O10000O100O10000O10000O100O10000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O100O100O100O1ZN`JlI`5R6QK^IP5`6_KSIa4k6k1N2M3N2N2N2N2N2N2N2M3N2N2N1O2N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1kGfMb3jK"}, "label": "a woman eating a sandwich thats wrapper in paper"}]}
{"id": 106148, "image": "COCO_train2014_000000106148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long sofa against the wall\"."}], "task": "refering", "answer_template": "The \"a long sofa against the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260], "bbox": [0.08159374999999999, 0.541924882629108, 0.37360937499999997, 0.8042957746478873], "iscrowd": 0, "area": 14162.289149999997, "rle": {"size": [426, 640], "counts": "_le01Y=2M3N2N2ZDGW:<eEFY:=eEC[:?cEB\\:`0aEB^:>bEB]:?cEB\\:>dEC[:>cEC\\:>aEF^::`EH`:8^EJb:6\\ELc:5[EMe:3[EMe:3[EMe:3\\ELc:5]EKc:4_EKa:5_EKa:6^EK`:8]EIc:9ZEHf::WEGh:=TEDl:>QECo:?nDBR;[1O101O001N1000000O1000000O10000O1000000O1000000O100000O01000000O10001N1000000O1000000O10001O0O10000O10001O1N2O1O001N2O1O1O0O2O1O1O0O2O000O2O001O0O101O001N010000O010000O010000O0100000O2O001N2O001N2O001O1N101O1N110O10O010O0100O01000O010O01ON2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N102N:E8I2N1N3N2N1N3N1O2M3N3M8GgaV5"}, "label": "a long sofa against the wall"}]}
{"id": 106148, "image": "COCO_train2014_000000106148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch with the deer next to it\"."}], "task": "refering", "answer_template": "The \"the couch with the deer next to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260], "bbox": [0.08159374999999999, 0.541924882629108, 0.37360937499999997, 0.8042957746478873], "iscrowd": 0, "area": 14162.289149999997, "rle": {"size": [426, 640], "counts": "_le01Y=2M3N2N2ZDGW:<eEFY:=eEC[:?cEB\\:`0aEB^:>bEB]:?cEB\\:>dEC[:>cEC\\:>aEF^::`EH`:8^EJb:6\\ELc:5[EMe:3[EMe:3[EMe:3\\ELc:5]EKc:4_EKa:5_EKa:6^EK`:8]EIc:9ZEHf::WEGh:=TEDl:>QECo:?nDBR;[1O101O001N1000000O1000000O10000O1000000O1000000O100000O01000000O10001N1000000O1000000O10001O0O10000O10001O1N2O1O001N2O1O1O0O2O1O1O0O2O000O2O001O0O101O001N010000O010000O010000O0100000O2O001N2O001N2O001O1N101O1N110O10O010O0100O01000O010O01ON2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N102N:E8I2N1N3N2N1N3N1O2M3N3M8GgaV5"}, "label": "the couch with the deer next to it"}]}
{"id": 106148, "image": "COCO_train2014_000000106148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couch by the window\"."}], "task": "refering", "answer_template": "The \"a couch by the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 317, 320, 321], "bbox": [0.701640625, 0.5612910798122066, 0.9991875, 0.8944600938967137], "iscrowd": 0, "area": 19810.02495, "rle": {"size": [426, 640], "counts": "\\Rk53<7P<0gC:S<a0M3L4WOeN]E`1b:aNYEd1e:_NTEg1k:=N2O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1001N1OO001N101O001O001O001O001O010O0100O100O010O101N100O20O0010O01O010O0010O01O010O0010O01O01L3M3M4N1N2O2N1O1O2O0O101O00001O00001O00010O01O010O0010O01O010O1O010O0010O01O010O0010O01O10O01O0010O01O010O00100O0010O01O010O0010O01O10O01O010ON2N3N1O1O2I6O101N10001N100O2O0O3Ne4"}, "label": "a couch by the window"}]}
{"id": 360110, "image": "COCO_train2014_000000360110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light blue back pack\"."}], "task": "refering", "answer_template": "The \"a light blue back pack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 140, 141, 142, 154, 155, 156, 168, 169, 170, 182, 183, 184, 196, 197, 198, 210, 211, 212], "bbox": [0.0, 0.5004, 0.197025, 0.84886], "iscrowd": 0, "area": 11638.643200000002, "rle": {"size": [500, 400], "counts": "j7k38ZMU:S4O01O001O001O0000001O0001O00000000000001O00001O00001O00001O001O00001O00001O001O00001O1O1O001O1O1O010O1O1O00100O1O001O100O1O3M2N3N1N3M2N3M2M4J5L5H7G9F:dM]C\\1m<_NX2Lf]l4"}, "label": "a light blue back pack"}]}
{"id": 360110, "image": "COCO_train2014_000000360110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and black everest back pack\"."}], "task": "refering", "answer_template": "The \"a blue and black everest back pack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 140, 141, 142, 154, 155, 156, 168, 169, 170, 182, 183, 184, 196, 197, 198, 210, 211, 212], "bbox": [0.0, 0.5004, 0.197025, 0.84886], "iscrowd": 0, "area": 11638.643200000002, "rle": {"size": [500, 400], "counts": "j7k38ZMU:S4O01O001O001O0000001O0001O00000000000001O00001O00001O00001O001O00001O00001O001O00001O1O1O001O1O1O010O1O1O00100O1O001O100O1O3M2N3N1N3M2N3M2M4J5L5H7G9F:dM]C\\1m<_NX2Lf]l4"}, "label": "a blue and black everest back pack"}]}
{"id": 188087, "image": "COCO_train2014_000000188087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fishing boat parked next to the light house\"."}], "task": "refering", "answer_template": "The \"fishing boat parked next to the light house\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262], "bbox": [0.34228125, 0.5332932692307693, 0.8597187500000001, 0.7426442307692308], "iscrowd": 0, "area": 14195.068750000008, "rle": {"size": [416, 640], "counts": "iVi2b1^;:F0001O000000001N1000001O000000001O000000001O0000001N100000001O0000001O000000001O00000O101O0000001O00000000000eN`DV1e;000L4KdNfD\\1S;kNmDU1T;jNlDLNT1W;`01O1O1O1OTO^NbFZ1^9jNPF1GU1V:mNSF2CQ1Y:oNSF1Ij0T:XOPFOIk0W:0dE4\\:X100001O000O10000000001O00000000000000001O00000O10000000001O000000000000001O00000O10000000001O00000000000hM[E51g1Q;O001O00000O10001O0000000000000000001O0000000O100000000000001O000000000eNPEf0e;00000000000000O10000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O10001O00000000001O0000000VOnCf0V<000001O1O1O002N00000000000O1000000000000000000000000000O1000000000000000000O1000O100000000000O1000000000000000000O10000000000000000O0100O1O100O2O0O100O1O100O100O100O101N101N5L0O101N10XYT1"}, "label": "fishing boat parked next to the light house"}]}
{"id": 188087, "image": "COCO_train2014_000000188087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat in the water in front of a building\"."}], "task": "refering", "answer_template": "The \"a boat in the water in front of a building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262], "bbox": [0.34228125, 0.5332932692307693, 0.8597187500000001, 0.7426442307692308], "iscrowd": 0, "area": 14195.068750000008, "rle": {"size": [416, 640], "counts": "iVi2b1^;:F0001O000000001N1000001O000000001O000000001O0000001N100000001O0000001O000000001O00000O101O0000001O00000000000eN`DV1e;000L4KdNfD\\1S;kNmDU1T;jNlDLNT1W;`01O1O1O1OTO^NbFZ1^9jNPF1GU1V:mNSF2CQ1Y:oNSF1Ij0T:XOPFOIk0W:0dE4\\:X100001O000O10000000001O00000000000000001O00000O10000000001O000000000000001O00000O10000000001O00000000000hM[E51g1Q;O001O00000O10001O0000000000000000001O0000000O100000000000001O000000000eNPEf0e;00000000000000O10000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O10001O00000000001O0000000VOnCf0V<000001O1O1O002N00000000000O1000000000000000000000000000O1000000000000000000O1000O100000000000O1000000000000000000O10000000000000000O0100O1O100O2O0O100O1O100O100O100O101N101N5L0O101N10XYT1"}, "label": "a boat in the water in front of a building"}]}
{"id": 188087, "image": "COCO_train2014_000000188087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat with a large white sail\"."}], "task": "refering", "answer_template": "The \"a boat with a large white sail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 272, 273, 274, 275, 297], "bbox": [0.8413437500000001, 0.24959134615384615, 1.0, 0.8046394230769232], "iscrowd": 0, "area": 12905.083550000001, "rle": {"size": [416, 640], "counts": "Slj61n<1O2O001N101O001O00001O0O1000001O1O001O0O100I7K5K5K5`K;QK9n4MeJ;_5FVJ`0n5AfIg0]6l2e0[OL4K5K]OTH\\Lh7h3ZHdKO=e7P4cHoK[7R4fHnKX7T4hHlKV7U4lHjKR7X4oHgKo6[4QIeKl6^4TIbKj6_4WIaKg6a4YI_Ke6c4\\I\\Kb6f4^IZK`6h4`IXK]6k4cIUKY6o4gIRKU6Q5kIoJQ6T5PJlJm5W5TJhJh5\\5XJdJc5a5]J_Ja5c5_J^J_5c5aJ]J]5e5cJ[J[5f5fJZJY5g5gJXJX5EWJR6b0XJ[5g5eJYJ\\5f5dJZJ]5e5cJ[J^5c5cJ]J_5a5aJ_J`5`5`J`Ja5_5_JaJb5^5^JaJd5]5\\JdJe5[5[JeJg5Y5YJgJi5W5WJiJk5U5UJkJn5R5RJnJP6P5PJPKR6m4oISKS6k4mITKU6k4jIVKX6h4hIXKY6g4gIYKZ6f4fIZK\\6d4dI\\K^6b4bI^K`6_4aIaK`6^4`IaKc6]4]IcKe6[4[IeKg6X4ZIhKh6V4XIjKi6U4WIkKk6S4UIlKm6R4TInKn6P4RIPLo6o3QIQLQ7l3PITLQ7k3oHULT7h3lHXLW7\\1eInNUOE[7R1PJPOfNM^7i0WKWOWO"}, "label": "a boat with a large white sail"}]}
{"id": 188087, "image": "COCO_train2014_000000188087.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sailboat on the water\"."}], "task": "refering", "answer_template": "The \"a sailboat on the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 272, 273, 274, 275, 297], "bbox": [0.8413437500000001, 0.24959134615384615, 1.0, 0.8046394230769232], "iscrowd": 0, "area": 12905.083550000001, "rle": {"size": [416, 640], "counts": "Slj61n<1O2O001N101O001O00001O0O1000001O1O001O0O100I7K5K5K5`K;QK9n4MeJ;_5FVJ`0n5AfIg0]6l2e0[OL4K5K]OTH\\Lh7h3ZHdKO=e7P4cHoK[7R4fHnKX7T4hHlKV7U4lHjKR7X4oHgKo6[4QIeKl6^4TIbKj6_4WIaKg6a4YI_Ke6c4\\I\\Kb6f4^IZK`6h4`IXK]6k4cIUKY6o4gIRKU6Q5kIoJQ6T5PJlJm5W5TJhJh5\\5XJdJc5a5]J_Ja5c5_J^J_5c5aJ]J]5e5cJ[J[5f5fJZJY5g5gJXJX5EWJR6b0XJ[5g5eJYJ\\5f5dJZJ]5e5cJ[J^5c5cJ]J_5a5aJ_J`5`5`J`Ja5_5_JaJb5^5^JaJd5]5\\JdJe5[5[JeJg5Y5YJgJi5W5WJiJk5U5UJkJn5R5RJnJP6P5PJPKR6m4oISKS6k4mITKU6k4jIVKX6h4hIXKY6g4gIYKZ6f4fIZK\\6d4dI\\K^6b4bI^K`6_4aIaK`6^4`IaKc6]4]IcKe6[4[IeKg6X4ZIhKh6V4XIjKi6U4WIkKk6S4UIlKm6R4TInKn6P4RIPLo6o3QIQLQ7l3PITLQ7k3oHULT7h3lHXLW7\\1eInNUOE[7R1PJPOfNM^7i0WKWOWO"}, "label": "a sailboat on the water"}]}
{"id": 48824, "image": "COCO_train2014_000000048824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white fire hydrant\"."}], "task": "refering", "answer_template": "The \"a red and white fire hydrant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 273, 274, 275, 276, 277, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 358, 359, 360, 361, 362, 375, 376, 377, 378, 379], "bbox": [0.0796875, 0.663015625, 0.3676041666666666, 1.0], "iscrowd": 0, "area": 20146.46195, "rle": {"size": [640, 480], "counts": "YWh04`c0>J6M2N3N2M2N3k]OUOPa0n0k^OVOSa0l0j^OWOUa0k0f^OYOXa0i0e^O[OXa0f0h^O[OVa0g0h^O[OWa0f0h^O\\OUa0f0j^Oo1`>TN]AT2[>nMdAR2[>j1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1H8000000000000000000001O0000001OM3L4M3000000002N3M2N3M00000000000000001O2N2N2N1OJ6N200000000000000:F00000000001O:F1O1O1O1O001O1O1O1oKbA^2_>aMcA]2]>bMgA[2Z>cMiAo0d?PO]@b0Q`0\\OP@8]`0Fc_O2g`0MY_O2i`0LX_O4h`0KY_O5h`0IX_O8l`0]O[_Ob0k`0RO]_Om0ka0N3N1O2M2O1O2N1O2M3J8M3M1O001O001O001O0O2L[_m5"}, "label": "a red and white fire hydrant"}]}
{"id": 48824, "image": "COCO_train2014_000000048824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white fire hydrant\"."}], "task": "refering", "answer_template": "The \"a red and white fire hydrant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 273, 274, 275, 276, 277, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 341, 342, 343, 344, 345, 358, 359, 360, 361, 362, 375, 376, 377, 378, 379], "bbox": [0.0796875, 0.663015625, 0.3676041666666666, 1.0], "iscrowd": 0, "area": 20146.46195, "rle": {"size": [640, 480], "counts": "YWh04`c0>J6M2N3N2M2N3k]OUOPa0n0k^OVOSa0l0j^OWOUa0k0f^OYOXa0i0e^O[OXa0f0h^O[OVa0g0h^O[OWa0f0h^O\\OUa0f0j^Oo1`>TN]AT2[>nMdAR2[>j1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1H8000000000000000000001O0000001OM3L4M3000000002N3M2N3M00000000000000001O2N2N2N1OJ6N200000000000000:F00000000001O:F1O1O1O1O001O1O1O1oKbA^2_>aMcA]2]>bMgA[2Z>cMiAo0d?PO]@b0Q`0\\OP@8]`0Fc_O2g`0MY_O2i`0LX_O4h`0KY_O5h`0IX_O8l`0]O[_Ob0k`0RO]_Om0ka0N3N1O2M2O1O2N1O2M3J8M3M1O001O001O001O0O2L[_m5"}, "label": "a red and white fire hydrant"}]}
{"id": 122560, "image": "COCO_train2014_000000122560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on left side\"."}], "task": "refering", "answer_template": "The \"elephant on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 46, 47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 75, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 241, 242, 256], "bbox": [0.027018779342723002, 0.103375, 0.48276995305164316, 0.768546875], "iscrowd": 0, "area": 45932.45744999999, "rle": {"size": [640, 426], "counts": "de74hc06K5K4L5K5K5K5J5L5J6J6K5J5L5J6VA_MP<d2kCaMR<a2iCdMU<c2bCbM[<m2SCYMj<l2oBYMm<k2mB[MP=h2kB\\MS=g2gB]MX=f2cB\\M]=g2^B\\MW5cNc2T4PH\\MS5kNl2l3lG\\Mo4POU3\\5bLnJ\\3S5\\LVKc3k4SL_Kl3a4lKhKS4Y4cKPL]4P4[KYLd4g3SKcLl4^3fJPMY5P3mHjNS7R7N200O10O01O100O1OJ7I6J7I6J6J7I6J6L50O1O100O10O01O100O1O100O1O010O100O1O100O1O010O1O100O100OPNRGPIm8k6\\GQIe8i6dGTI[8g6nGVIQ8e6YHWIg7c6bHZI]7a6lH\\IT7`6SI]Il6e6VIXIi6j6YISIg6n6\\ImHd6U7^IhHa6[7`IbH`6_7bI^H]6d7eIYH[6h7gIUHY6l7jIPHV6P8mImGS6T8l2000000000000001O0001O00001O000010O01OZFmG_7S8hGnHP8R7WGnIc8n8I7I7JQK`G`N\\8_1gG_NZ8_1hG`NX8_1jG`NV8_1mG_NT8_1nG`NR8_1PH`NP8_1SH_Nn7_1TH`Nl7_1VH`Nk7^1XHhMXOlL`8[5ZHcM[ORM[8Z5\\H_M]OXMW8W5^H[M@^MR8V5aHVMAdMn7U5cHQMDjMj7S5dHnLFoMf7R5gHhLHVNb7P5hHeLJ[N^7o4jH`LMaNY7n4mH[LNgNV7l4nHWL1mNQ7k4PISL3ROn6i4RIoK4XOj6h4TIjK7^Of6f4UIgK9Cb6n3`GkKh1=;J^6j3hGhKa1?=OZ6f3SHdKX1a0`05V6a3[JULC:R6]3aJRLBa0n5X3eJRLAf0j5T3jMlLW2P3lMkLY2Q3kM^Lg2_3n6O2N1O2M2O2N1O2N1O2M2O2N1O2N1O2N101O0O2O001O1N3N2N2M3N2N2M2@a0XOh0XOR]Y4"}, "label": "elephant on left side"}]}
{"id": 122560, "image": "COCO_train2014_000000122560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with small trunk is walking together with another elephant\"."}], "task": "refering", "answer_template": "The \"an elephant with small trunk is walking together with another elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 46, 47, 48, 49, 50, 51, 61, 62, 63, 64, 65, 66, 67, 75, 76, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 151, 152, 153, 154, 155, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 241, 242, 256], "bbox": [0.027018779342723002, 0.103375, 0.48276995305164316, 0.768546875], "iscrowd": 0, "area": 45932.45744999999, "rle": {"size": [640, 426], "counts": "de74hc06K5K4L5K5K5K5J5L5J6J6K5J5L5J6VA_MP<d2kCaMR<a2iCdMU<c2bCbM[<m2SCYMj<l2oBYMm<k2mB[MP=h2kB\\MS=g2gB]MX=f2cB\\M]=g2^B\\MW5cNc2T4PH\\MS5kNl2l3lG\\Mo4POU3\\5bLnJ\\3S5\\LVKc3k4SL_Kl3a4lKhKS4Y4cKPL]4P4[KYLd4g3SKcLl4^3fJPMY5P3mHjNS7R7N200O10O01O100O1OJ7I6J7I6J6J7I6J6L50O1O100O10O01O100O1O100O1O010O100O1O100O1O010O1O100O100OPNRGPIm8k6\\GQIe8i6dGTI[8g6nGVIQ8e6YHWIg7c6bHZI]7a6lH\\IT7`6SI]Il6e6VIXIi6j6YISIg6n6\\ImHd6U7^IhHa6[7`IbH`6_7bI^H]6d7eIYH[6h7gIUHY6l7jIPHV6P8mImGS6T8l2000000000000001O0001O00001O000010O01OZFmG_7S8hGnHP8R7WGnIc8n8I7I7JQK`G`N\\8_1gG_NZ8_1hG`NX8_1jG`NV8_1mG_NT8_1nG`NR8_1PH`NP8_1SH_Nn7_1TH`Nl7_1VH`Nk7^1XHhMXOlL`8[5ZHcM[ORM[8Z5\\H_M]OXMW8W5^H[M@^MR8V5aHVMAdMn7U5cHQMDjMj7S5dHnLFoMf7R5gHhLHVNb7P5hHeLJ[N^7o4jH`LMaNY7n4mH[LNgNV7l4nHWL1mNQ7k4PISL3ROn6i4RIoK4XOj6h4TIjK7^Of6f4UIgK9Cb6n3`GkKh1=;J^6j3hGhKa1?=OZ6f3SHdKX1a0`05V6a3[JULC:R6]3aJRLBa0n5X3eJRLAf0j5T3jMlLW2P3lMkLY2Q3kM^Lg2_3n6O2N1O2M2O2N1O2N1O2M2O2N1O2N1O2N101O0O2O001O1N3N2N2M3N2N2M2@a0XOh0XOR]Y4"}, "label": "an elephant with small trunk is walking together with another elephant"}]}
{"id": 122560, "image": "COCO_train2014_000000122560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with ears flapping out from its head\"."}], "task": "refering", "answer_template": "The \"elephant with ears flapping out from its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 85, 86, 87, 88, 89, 96, 97, 98, 99, 100, 101, 102, 103, 104, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 215, 216, 217, 218, 219, 220, 221, 222, 223, 231, 232, 233, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 281, 282, 290, 291, 292, 293, 295, 296, 297, 310, 311, 312, 326], "bbox": [0.3443661971830986, 0.206734375, 0.9891784037558685, 0.9213437500000001], "iscrowd": 0, "area": 83685.12499999997, "rle": {"size": [640, 426], "counts": "hTl23lc03[@KY;9cDI[;<_DFa;=ZDFd;>WDEg;?UDBj;b0QDAn;b0mC@R<d0iC_OU<e0fC^OY<f0bC[O]<i0^CZO`<j0[CXOc1WOn6e1ZGWOe1XOo6d1WGWOg1XOP7e1QGXOl1XOP7d1kF[OR2TOR7e1bF]OZ2ROQ7e1\\F@_2POf5P3bGWNe2lNg5Q3[GYNl2jNf5Q3UG\\NQ3gNi5e5PJ_Jm5j5iIZJU6n5`IVJ^6R6YIRJd6W6RIlIm6[6iHiIU7`6aHcI^7e6XH^If7j6PHZIn7o6hGTIW8[9N2N2O1N2O1001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O0000O100O10000O10000O100OL5K5K4L5K4L5O1O010O00100O00100O0010O01O1O010O0`MhFYJY9\\5TGbJl8S5aGjJ_8l4nGQKS8j4THTKk7n4VHoJk7Q5WHmJi7T5XHiJh7X5ZHeJg7[5[HcJe7^5\\H_Jd7b5^H[Jc7f5^HXJa7i5aHTJ`7m5aHQJ_7o5cHnI]7T6dHiI]7W6eH_Ib7b6n21O0O2O0O2O001N10001N101N101WHdHl2\\7RMgHl2[7RMgHm2Y7QMjHm2X7PMkHo2U7oLnHP3R7nLQIP3Q7mLRIR3n6lLUIS3l6jLWIT3j6kLXIT3R7`LQI^3]7SLfHl3g7eK\\HZ4Q8WKRHg4\\8jJgGU5g8[J\\Gd5Q9mIRGQ6\\9aIeF_6_;N2O1O1N2O1O1N2O1N10000O10000O010O10000O01000O100O01000O1000O0100O1000O010000O10O010000O1000O0100O1000QLkJRIU5l6QLQHo3l7[LmGe3o7eLkGZ3S8PMfGP3W8YMbGh2[8bM^G]2`8mMYGS2d8VNVGi1g8bNRG^1k8m4M2N3M3L3N3M2N3M3L3N3M3M2O20O02N101N1O2O1N1O2O0O2N101N1O2O1N1O2O0O2N101N1O2O0O2N2N101N1O2O0O2N101N1O2O1N1O2OP2oMX2hMPk2"}, "label": "elephant with ears flapping out from its head"}]}
{"id": 122560, "image": "COCO_train2014_000000122560.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant with its ears extended\"."}], "task": "refering", "answer_template": "The \"the elephant with its ears extended\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 85, 86, 87, 88, 89, 96, 97, 98, 99, 100, 101, 102, 103, 104, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 215, 216, 217, 218, 219, 220, 221, 222, 223, 231, 232, 233, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 281, 282, 290, 291, 292, 293, 295, 296, 297, 310, 311, 312, 326], "bbox": [0.3443661971830986, 0.206734375, 0.9891784037558685, 0.9213437500000001], "iscrowd": 0, "area": 83685.12499999997, "rle": {"size": [640, 426], "counts": "hTl23lc03[@KY;9cDI[;<_DFa;=ZDFd;>WDEg;?UDBj;b0QDAn;b0mC@R<d0iC_OU<e0fC^OY<f0bC[O]<i0^CZO`<j0[CXOc1WOn6e1ZGWOe1XOo6d1WGWOg1XOP7e1QGXOl1XOP7d1kF[OR2TOR7e1bF]OZ2ROQ7e1\\F@_2POf5P3bGWNe2lNg5Q3[GYNl2jNf5Q3UG\\NQ3gNi5e5PJ_Jm5j5iIZJU6n5`IVJ^6R6YIRJd6W6RIlIm6[6iHiIU7`6aHcI^7e6XH^If7j6PHZIn7o6hGTIW8[9N2N2O1N2O1001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O0000O100O10000O10000O100OL5K5K4L5K4L5O1O010O00100O00100O0010O01O1O010O0`MhFYJY9\\5TGbJl8S5aGjJ_8l4nGQKS8j4THTKk7n4VHoJk7Q5WHmJi7T5XHiJh7X5ZHeJg7[5[HcJe7^5\\H_Jd7b5^H[Jc7f5^HXJa7i5aHTJ`7m5aHQJ_7o5cHnI]7T6dHiI]7W6eH_Ib7b6n21O0O2O0O2O001N10001N101N101WHdHl2\\7RMgHl2[7RMgHm2Y7QMjHm2X7PMkHo2U7oLnHP3R7nLQIP3Q7mLRIR3n6lLUIS3l6jLWIT3j6kLXIT3R7`LQI^3]7SLfHl3g7eK\\HZ4Q8WKRHg4\\8jJgGU5g8[J\\Gd5Q9mIRGQ6\\9aIeF_6_;N2O1O1N2O1O1N2O1N10000O10000O010O10000O01000O100O01000O1000O0100O1000O010000O10O010000O1000O0100O1000QLkJRIU5l6QLQHo3l7[LmGe3o7eLkGZ3S8PMfGP3W8YMbGh2[8bM^G]2`8mMYGS2d8VNVGi1g8bNRG^1k8m4M2N3M3L3N3M2N3M3L3N3M3M2O20O02N101N1O2O1N1O2O0O2N101N1O2O1N1O2O0O2N101N1O2O0O2N2N101N1O2O0O2N101N1O2O1N1O2OP2oMX2hMPk2"}, "label": "the elephant with its ears extended"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bedspread the man is laying on top of\"."}], "task": "refering", "answer_template": "The \"the white bedspread the man is laying on top of\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.6482500000000001, 0.9960625000000001, 0.9988125000000001], "iscrowd": 0, "area": 91678.77354999997, "rle": {"size": [480, 640], "counts": "i93f1g0g:_OnDV2a9QNcEf3\\9o0O1001O000000001O0000001O000000001O000000001O00000000000000001O00000000MQKUFo4k9QKUFo4k9QKUFo4n90001O00MRKTFn4l9RKTFn4l9RKTFn4l9RKTFn4l9RKTFn4l9RKTFn4l9RKTFn4m9RKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9TKRFl4n9TKRFl4n9TKRFl4n9TKRFl4o9SKQFm4o9SKQFm4o9SKQFm4P:10000000000000000000O100000000000000000O10O100000000001O00000000000000001O000000000000001O0000000000000000000000000000000000000000000000000000O1000000000000000000O1000000000O10000000000000O1000000000000000000000000000000000000000000000000001O0000000000001O0000000000001O000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O2N1O1O2N1O2N1O1O2N001O001O001N10001O001O001O001O001O00001O001O001O001O001O00001O001O001O0000000000001O000000000001O00000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O0001O0000000000000001O0000000001O0000000001O000000001O0001O01O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001N100000001O0001O01O0000001O0hLPE`2h;01O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O000O2O0000VP1"}, "label": "the white bedspread the man is laying on top of"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hotel bed man is laying on\"."}], "task": "refering", "answer_template": "The \"hotel bed man is laying on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.6482500000000001, 0.9960625000000001, 0.9988125000000001], "iscrowd": 0, "area": 91678.77354999997, "rle": {"size": [480, 640], "counts": "i93f1g0g:_OnDV2a9QNcEf3\\9o0O1001O000000001O0000001O000000001O000000001O00000000000000001O00000000MQKUFo4k9QKUFo4k9QKUFo4n90001O00MRKTFn4l9RKTFn4l9RKTFn4l9RKTFn4l9RKTFn4l9RKTFn4l9RKTFn4m9RKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9SKSFm4m9TKRFl4n9TKRFl4n9TKRFl4n9TKRFl4o9SKQFm4o9SKQFm4o9SKQFm4P:10000000000000000000O100000000000000000O10O100000000001O00000000000000001O000000000000001O0000000000000000000000000000000000000000000000000000O1000000000000000000O1000000000O10000000000000O1000000000000000000000000000000000000000000000000001O0000000000001O0000000000001O000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O2N1O1O2N1O2N1O1O2N001O001O001N10001O001O001O001O001O00001O001O001O001O001O00001O001O001O0000000000001O000000000001O00000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O0001O0000000000000001O0000000001O0000000001O000000001O0001O01O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001N100000001O0001O01O0000001O0hLPE`2h;01O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O001O000O2O0000VP1"}, "label": "hotel bed man is laying on"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection of a bed in the mirror\"."}], "task": "refering", "answer_template": "The \"a reflection of a bed in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.05871875, 0.44987499999999997, 0.593953125, 0.6579375], "iscrowd": 0, "area": 25011.795799999996, "rle": {"size": [480, 640], "counts": "hPb0l0T>o0QOo0QO:F000000000000000000000000000000000000000000000000000000000000O100000000000000000000000002N6J5K5K4L001O1O1O1O001O1O1O1O00000000001O0000000000000000001O0000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O1000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001N1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000[OUNUDk1k;[NoCe1Q<bNhC^1W<f0000000000000000000000000000000000000000000000000000000000000000kNSfi3"}, "label": "a reflection of a bed in the mirror"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed in the reflection\"."}], "task": "refering", "answer_template": "The \"the bed in the reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.05871875, 0.44987499999999997, 0.593953125, 0.6579375], "iscrowd": 0, "area": 25011.795799999996, "rle": {"size": [480, 640], "counts": "hPb0l0T>o0QOo0QO:F000000000000000000000000000000000000000000000000000000000000O100000000000000000000000002N6J5K5K4L001O1O1O1O001O1O1O1O00000000001O0000000000000000001O0000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O1000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001N1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000[OUNUDk1k;[NoCe1Q<bNhC^1W<f0000000000000000000000000000000000000000000000000000000000000000kNSfi3"}, "label": "the bed in the reflection"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the head portion of a guy turned left and lying\"."}], "task": "refering", "answer_template": "The \"the head portion of a guy turned left and lying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321, 343, 344], "bbox": [0.61046875, 0.40077083333333335, 1.0, 0.8555416666666666], "iscrowd": 0, "area": 35283.49545, "rle": {"size": [480, 640], "counts": "iag5;Z>?C=J7H7J6I7J6I4M4K5M2N3M3L3N3M3M2N3M1O1N3N1O1O1O2N1O1O1O2N1O100O2N1O1O1O2O1N2N1O1O1O100O1O1O100O1000000O10000O1000000O10000O1000000O10000O10001O0O1000000O1000000O1000000O1000000O100000000O11O000001O0001O00000000010O00000000000010O000000000000010O000000000001O010O001O1O001O001O001O1O001O001O001O1O001O001O00000000000000000001O00001O000010O000001O00010O00001O01O01O00000010O1O100O1O100O001O100O1O100O2N2OhE]LS9h41N2N2N1O2N100O1O2N1O101N1O1O1O2O0OO1O1O1O2O0O1O1O1O101N1O1O100O2N1O100O1O2N10000000000000lLdK"}, "label": "the head portion of a guy turned left and lying"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back side of a person ' s head who is lying on bed and seen through the mirror\"."}], "task": "refering", "answer_template": "The \"the back side of a person ' s head who is lying on bed and seen through the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 320, 321, 343, 344], "bbox": [0.61046875, 0.40077083333333335, 1.0, 0.8555416666666666], "iscrowd": 0, "area": 35283.49545, "rle": {"size": [480, 640], "counts": "iag5;Z>?C=J7H7J6I7J6I4M4K5M2N3M3L3N3M3M2N3M1O1N3N1O1O1O2N1O1O1O2N1O100O2N1O1O1O2O1N2N1O1O1O100O1O1O100O1000000O10000O1000000O10000O1000000O10000O10001O0O1000000O1000000O1000000O1000000O100000000O11O000001O0001O00000000010O00000000000010O000000000000010O000000000001O010O001O1O001O001O001O1O001O001O001O1O001O001O00000000000000000001O00001O000010O000001O00010O00001O01O01O00000010O1O100O1O100O001O100O1O100O2N2OhE]LS9h41N2N2N1O2N100O1O2N1O101N1O1O1O2O0OO1O1O1O2O0O1O1O1O101N1O1O100O2N1O100O1O2N10000000000000lLdK"}, "label": "the back side of a person ' s head who is lying on bed and seen through the mirror"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection in a mirror of a man lying down on a bed\"."}], "task": "refering", "answer_template": "The \"a reflection in a mirror of a man lying down on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.140203125, 0.34572916666666664, 0.5388437500000001, 0.5101249999999999], "iscrowd": 0, "area": 12961.955899999999, "rle": {"size": [480, 640], "counts": "g\\Z13l>4L4L4L5K4L2NN1012N2N2N2N2N2N2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O01O01O001O001O00001O001O001O00001O010O001O001O00001O001O001O00010O001O001O001O000000O100O101N100O100O100O100O100O100O101N100O100O100O100O100O100O101N100O100O100O100O100O100O1O100O100O100O100O2O0O100O100O100000000000000O010000000000000000000O10000000000001O1O001O1O1O001N2O001O1O001O1O001O1O001O1O1O001O1O001O1O001O001O001fNhBj0X=TOjBl0W=ROjBn0V=POlBP1U=mNmBS1S=lNnBT1_=0001O0000000000000000000O2O000000000000000000001O0000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0K6IQRZ4"}, "label": "a reflection in a mirror of a man lying down on a bed"}]}
{"id": 483015, "image": "COCO_train2014_000000483015.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"reflection of man lying on bed\"."}], "task": "refering", "answer_template": "The \"reflection of man lying on bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.140203125, 0.34572916666666664, 0.5388437500000001, 0.5101249999999999], "iscrowd": 0, "area": 12961.955899999999, "rle": {"size": [480, 640], "counts": "g\\Z13l>4L4L4L5K4L2NN1012N2N2N2N2N2N2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O01O01O001O001O00001O001O001O00001O010O001O001O00001O001O001O00010O001O001O001O000000O100O101N100O100O100O100O100O100O101N100O100O100O100O100O100O101N100O100O100O100O100O100O1O100O100O100O100O2O0O100O100O100000000000000O010000000000000000000O10000000000001O1O001O1O1O001N2O001O1O001O1O001O1O001O1O1O001O1O001O1O001O001O001fNhBj0X=TOjBl0W=ROjBn0V=POlBP1U=mNmBS1S=lNnBT1_=0001O0000000000000000000O2O000000000000000000001O0000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0K6IQRZ4"}, "label": "reflection of man lying on bed"}]}
{"id": 286411, "image": "COCO_train2014_000000286411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front end of accordian bus\"."}], "task": "refering", "answer_template": "The \"front end of accordian bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 120, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 219, 220], "bbox": [0.0, 0.32727083333333334, 0.58246875, 0.5662291666666667], "iscrowd": 0, "area": 34234.71975, "rle": {"size": [480, 640], "counts": "R5d0[>Y1gNU1kN1O1O100000010O0000000001O000000001O0001O0001O00000000001O000001O0000000000000000000000001O0000000000000000000000001O0O10000000001O1O1O1O1O001O1O1O000000000000000000000000000000001O00000000O1O1O1O1O1O1O1K5N2000000001O0000000000000O10001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000000000001O000000000008Hh2XMmX:P1ZdEi100O100O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000O10000O1000000O10000O10000O10000O1000000O10000O10000O100000000001O000000001O000000001O0000001O000000001O000000000000000000001O00000000000000000000001O0000000000000000000000D<VOj0WOi0WOfPm3"}, "label": "front end of accordian bus"}]}
{"id": 286411, "image": "COCO_train2014_000000286411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back half of extended bus\"."}], "task": "refering", "answer_template": "The \"back half of extended bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.581453125, 0.3299583333333333, 1.0, 0.5681666666666667], "iscrowd": 0, "area": 26988.240949999996, "rle": {"size": [480, 640], "counts": "Va^5Q3o;0000O0100000000000O1000O10000000O100000O1000000000000000O1000000000000000000000000000O100000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000O100000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000000O101O000000000000000000001O1O1O001O1O1O001O001O00000000000000000000001O00000000000000000000001O000000000000000O1000001O000000O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O11O0000001O00001O0000001O0000001O0000001O0000001O0000001N1000001O00lMVM"}, "label": "back half of extended bus"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light cement colour fur toy in a group of toys\"."}], "task": "refering", "answer_template": "The \"a light cement colour fur toy in a group of toys\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 258, 259, 260, 277, 278, 279, 281, 282, 283], "bbox": [0.021953125, 0.17566666666666667, 0.35304687500000004, 0.759], "iscrowd": 0, "area": 39513.84409999999, "rle": {"size": [480, 640], "counts": "`i63k>7kLHUG`0`8G]G`0X8GeG`0P8GmG`0h7GTH`0g7BVHe0d7]OYHj0b7VO\\HQ1^7QO^HS1`7oN]HS1b7oN[HS1d7oNYHS1g7nNUHU1j7mNSHU1l7mNQHT1o7nNnGT1R8POgGS1X8QOcGQ1\\8TO^Gn0a8VOZGl0f8XOSGj0m8ZOnFh0Q9]OiFe0V9c200O001O100VIYJZ4h5mJnJU5S5RJfKm5Z4[I_Le6n5O1O1O100O1O1O10000000000000000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000K5M30L4G9G9H8G9G9H8G9G9H8O10000000000000000000O100001O1O001O1O001O1O1O001O1O1O1O2N1O1O2N1O1005L4K5K5L4K5K5K5L4K5K5L]JSHU4h7hK[H\\4a7aKaH^4b7_K`H`4a7]KbHb4_7\\KcHc4^7ZKdHf4^7VKeHi4\\7UKfHj4[7SKhHk4[7RKfHn4[7oJhHP5Y7mJjHR5W7lJkHS5W7iJkHW5V7gJlHX5U7eJnHZ5\\8O2O1O1O001O1O1N101O1O1O001O1O1N1aKRFl3o9SLVFh3k9VLZFf3g9YL]Fc3c9\\LbF`3_9_LfF\\3[9bLjFY3W9gLmFU3T9iLRGR3o8mLUGo2l8oLYGm2g8SM]Gi2b:K4K5L4L5K4L4L5K4L4K5L5K4L4L4L5K4K5L4L5KWgQ6"}, "label": "a light cement colour fur toy in a group of toys"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray stuffed bear with a blue cloth and gray hat\"."}], "task": "refering", "answer_template": "The \"a gray stuffed bear with a blue cloth and gray hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 258, 259, 260, 277, 278, 279, 281, 282, 283], "bbox": [0.021953125, 0.17566666666666667, 0.35304687500000004, 0.759], "iscrowd": 0, "area": 39513.84409999999, "rle": {"size": [480, 640], "counts": "`i63k>7kLHUG`0`8G]G`0X8GeG`0P8GmG`0h7GTH`0g7BVHe0d7]OYHj0b7VO\\HQ1^7QO^HS1`7oN]HS1b7oN[HS1d7oNYHS1g7nNUHU1j7mNSHU1l7mNQHT1o7nNnGT1R8POgGS1X8QOcGQ1\\8TO^Gn0a8VOZGl0f8XOSGj0m8ZOnFh0Q9]OiFe0V9c200O001O100VIYJZ4h5mJnJU5S5RJfKm5Z4[I_Le6n5O1O1O100O1O1O10000000000000000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000K5M30L4G9G9H8G9G9H8G9G9H8O10000000000000000000O100001O1O001O1O001O1O1O001O1O1O1O2N1O1O2N1O1005L4K5K5L4K5K5K5L4K5K5L]JSHU4h7hK[H\\4a7aKaH^4b7_K`H`4a7]KbHb4_7\\KcHc4^7ZKdHf4^7VKeHi4\\7UKfHj4[7SKhHk4[7RKfHn4[7oJhHP5Y7mJjHR5W7lJkHS5W7iJkHW5V7gJlHX5U7eJnHZ5\\8O2O1O1O001O1O1N101O1O1O001O1O1N1aKRFl3o9SLVFh3k9VLZFf3g9YL]Fc3c9\\LbF`3_9_LfF\\3[9bLjFY3W9gLmFU3T9iLRGR3o8mLUGo2l8oLYGm2g8SM]Gi2b:K4K5L4L5K4L4L5K4L4K5L5K4L4L4L5K4K5L4L5KWgQ6"}, "label": "a gray stuffed bear with a blue cloth and gray hat"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large black bear with white patch on front\"."}], "task": "refering", "answer_template": "The \"large black bear with white patch on front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271], "bbox": [0.6505625, 0.18427083333333333, 0.98090625, 0.703375], "iscrowd": 0, "area": 29861.430450000003, "rle": {"size": [480, 640], "counts": "hUS61l>5J6K5K5\\N^ObDh0\\;\\O]Di0`;]OXDh0f;\\OSDi0j;\\OnCj0o;[OjCj0T<S1M3N2M3N2M3N2M3M3N2M3N2M3N2^FWLh7m3THWLi7k3UHWLh7l3VHVLg7m3VHWLg7k3WHWLf7l3WHXLf7j3XHXLd7l3YHXLa7n3\\HTL`7P4^HSL]7Q4`HRL[7S4cHPLX7T4eHoKX7T4fHoKV7T4hHnKV7T4gHoKW7S4gHmKY7V4cHkK]7W4aHjK^7X4`HhK`7l5000001O0000001N1000001O0000001O000000001N1000001O0000001O000000001O0O11O010O1O1O10OO2N2N2O1N1O2N2N2O1N1O2N2O1N2N1O2N2O1N2N2N101N2YOcGhJ_8W5cGfJ_8Y5dGcJ]8\\5gGaJZ8[1]G]2;ULZ8[1^G_2;RLY8[1`Gb2S9[MPGd2P9ZMSGe2n8WMVGh2k8UMXGj2i8SMYGm2h8PM[Go2e8nL_GQ3b8lLaGS3`8jLcGU3^8gLfGX3[8eLhG[3W8cLlG\\3U8`LoG_3R8^LQHb3g9O1O10O01O100O1O100O1O100O1O100O1O010O1O100O1O100O1O101N3M3N2M4L3N2M3M3N2M3M3N1N1N101O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O0I8F\\[5"}, "label": "large black bear with white patch on front"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big bear with red behind him\"."}], "task": "refering", "answer_template": "The \"the big bear with red behind him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271], "bbox": [0.6505625, 0.18427083333333333, 0.98090625, 0.703375], "iscrowd": 0, "area": 29861.430450000003, "rle": {"size": [480, 640], "counts": "hUS61l>5J6K5K5\\N^ObDh0\\;\\O]Di0`;]OXDh0f;\\OSDi0j;\\OnCj0o;[OjCj0T<S1M3N2M3N2M3N2M3M3N2M3N2M3N2^FWLh7m3THWLi7k3UHWLh7l3VHVLg7m3VHWLg7k3WHWLf7l3WHXLf7j3XHXLd7l3YHXLa7n3\\HTL`7P4^HSL]7Q4`HRL[7S4cHPLX7T4eHoKX7T4fHoKV7T4hHnKV7T4gHoKW7S4gHmKY7V4cHkK]7W4aHjK^7X4`HhK`7l5000001O0000001N1000001O0000001O000000001N1000001O0000001O000000001O0O11O010O1O1O10OO2N2N2O1N1O2N2N2O1N1O2N2O1N2N1O2N2O1N2N2N101N2YOcGhJ_8W5cGfJ_8Y5dGcJ]8\\5gGaJZ8[1]G]2;ULZ8[1^G_2;RLY8[1`Gb2S9[MPGd2P9ZMSGe2n8WMVGh2k8UMXGj2i8SMYGm2h8PM[Go2e8nL_GQ3b8lLaGS3`8jLcGU3^8gLfGX3[8eLhG[3W8cLlG\\3U8`LoG_3R8^LQHb3g9O1O10O01O100O1O100O1O100O1O100O1O010O1O100O1O100O1O101N3M3N2M4L3N2M3M3N2M3M3N1N1N101O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O0I8F\\[5"}, "label": "the big bear with red behind him"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teddy bear wearing the red shirt\"."}], "task": "refering", "answer_template": "The \"the teddy bear wearing the red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 286], "bbox": [0.318265625, 0.25395833333333334, 0.5635312499999999, 0.7150416666666667], "iscrowd": 0, "area": 20612.74485, "rle": {"size": [480, 640], "counts": "Vjo2c0X>8K5L4M2M4L4M1N3M3N1N3M>C2POWNdDl1W;YNgDj1T;YNjDj1Q;\\NlDe1R;^NkDc1c:_MhEg3U:^LgEd3U:aLhE`3X:aLfEa3Y:`LeEb3Z:_LdEc3[:a001O10O01O1O1O001O000000WNjEhNV:V1mEiNS:X1mEgNS:Y1oEeNQ:\\1oEcNQ:^1oEaNQ:_1PF`NP:a1QF]No9c1RF\\Nn9e1SFYNR:d1mE[NS:g1mEWNS:k1lESNU:o1kEnMV:S2jEkMW:[2dEcM]:^2cE`M^:a2bE]M_:d2n01UEZMZ9g2VFiMi9i3N1O2N1O2N3M2Nb0^O4L4L4L101N1O1O2N1O2M10000O1000O10001O00000000000O10000^IiGX6Z8fIiGV6d8\\NjGXLY8g3kGSLX8k3mGPLU8o3QHiKR8U4UHdKl7\\4]1N2O1`LZEc2k:XMVEh2m:SMUEl2o:PMSEo2P;mLREQ3_;01O11iDnL\\:S3_ERMa:n2ZEWMf:i2VE[Mj:^30000000YLWEV3h:iL\\EU3d:jL_ET3`:lLbES3X;N0O2O001O001O1O000O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1N2UOfCnN\\<l0hCTOZ<e0lCXOX<d0mCWOY<c0lCUO\\<f0T1J8I6IdoR4"}, "label": "the teddy bear wearing the red shirt"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear in a red shirt sitting with other teddy bears\"."}], "task": "refering", "answer_template": "The \"a teddy bear in a red shirt sitting with other teddy bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 286], "bbox": [0.318265625, 0.25395833333333334, 0.5635312499999999, 0.7150416666666667], "iscrowd": 0, "area": 20612.74485, "rle": {"size": [480, 640], "counts": "Vjo2c0X>8K5L4M2M4L4M1N3M3N1N3M>C2POWNdDl1W;YNgDj1T;YNjDj1Q;\\NlDe1R;^NkDc1c:_MhEg3U:^LgEd3U:aLhE`3X:aLfEa3Y:`LeEb3Z:_LdEc3[:a001O10O01O1O1O001O000000WNjEhNV:V1mEiNS:X1mEgNS:Y1oEeNQ:\\1oEcNQ:^1oEaNQ:_1PF`NP:a1QF]No9c1RF\\Nn9e1SFYNR:d1mE[NS:g1mEWNS:k1lESNU:o1kEnMV:S2jEkMW:[2dEcM]:^2cE`M^:a2bE]M_:d2n01UEZMZ9g2VFiMi9i3N1O2N1O2N3M2Nb0^O4L4L4L101N1O1O2N1O2M10000O1000O10001O00000000000O10000^IiGX6Z8fIiGV6d8\\NjGXLY8g3kGSLX8k3mGPLU8o3QHiKR8U4UHdKl7\\4]1N2O1`LZEc2k:XMVEh2m:SMUEl2o:PMSEo2P;mLREQ3_;01O11iDnL\\:S3_ERMa:n2ZEWMf:i2VE[Mj:^30000000YLWEV3h:iL\\EU3d:jL_ET3`:lLbES3X;N0O2O001O001O1O000O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1N2UOfCnN\\<l0hCTOZ<e0lCXOX<d0mCWOY<c0lCUO\\<f0T1J8I6IdoR4"}, "label": "a teddy bear in a red shirt sitting with other teddy bears"}]}
{"id": 442062, "image": "COCO_train2014_000000442062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest brown bear\"."}], "task": "refering", "answer_template": "The \"the smallest brown bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 223, 243, 244, 245], "bbox": [0.5109843749999999, 0.2526041666666667, 0.7209687499999999, 0.6059375], "iscrowd": 0, "area": 13515.3993, "rle": {"size": [480, 640], "counts": "[li42m>2N1O2N2N1O2fAMd0Ll;8]C75OY<L^C?O8R<h1L4M2N3L3N1O2N2M101O001N101O001N101N101O1N2O1N2O1O0WEZLW:[4M3M1O10O01O1O001O1O1O1O1O100O1O2N1O1O1O000000000001O1O001O100O1O1O2M2N2O0O1O101N100O1hLgEh05dNU:a0iEg09eNo9a0kEf0;fNk9a0mEh0<cNh9a0QFZ1JSNU:<XFd1^:UNhEm1Y:kMkEW2W:aMnE`2X;2O2N1O10O000O001O2O0O2N101N1O2O1N2N?B4K1O1O2O0O2N1O2N1O2N2N1O2N2N3M5K2N2N2M2O1O2N1N3M2M4M2M`dc2"}, "label": "the smallest brown bear"}]}
{"id": 433874, "image": "COCO_train2014_000000433874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse with light colored saddle\"."}], "task": "refering", "answer_template": "The \"horse with light colored saddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 247, 267, 268, 270, 293], "bbox": [0.573390625, 0.3462295081967213, 0.78340625, 0.8234192037470727], "iscrowd": 0, "area": 9838.101349999997, "rle": {"size": [427, 640], "counts": "iZi41Y=3M3M2`Ni0hD[OU;]1O10ObNZNYGf1g8[OYFe0g9^11UO`FkM`9S2m0N2N3M2N3N101O0010O010O3N1N3N1N01\\OPElNQ;S1QEkNP;W1oDfNS;\\1lDbNU;`1jD^NW;k0gDG1\\O=l0^9NYFILZO:]1W9]OiFLLm0Y9TOQGNAR1]9mNXGk1f8RN^Gh0ZOFW9@aGi0AAm8BdGm0G[Od8EhGn0MXOY8GmGP1MXOU8EPHe0QO]Om09P8CTH>XOEg09V9J\\FLa08U:HmE5U:JmE4S:LoE1R:OPFOP:1e10C3]CL^<<_CE\\<a0bC_OY<?dCH0IW<>SDHf0Ln9`0_EAa0?]95UFYO=i0S94^G2Z81eG2W81TGc0j8@_FU1_9i0dFYMZ9c2nF[Mo8a2[G\\Mc8e2`GYM^8g2fGWMW8j2nGSMQ8l2SHRMl7n2W101]Oc0N12O0O2O001N101O0O101O0O2O0O2O001N101O1N3N3L4M2N3_O`0O2O1O0O2O1O0O2O1O1N2O1O2MjNVGmMi8o1^GoM_8n1iGPNT8n1RHQNl7U1dGYNg0i0Z7d0jIGh5OfJ5Q5GWK<b4CcK?X4`NhHe0V3k0n3_NoHc0`3d0_3hNTIa0k3<P3QOXI?n3?g2QO^I=P4?S3^ORM=Q3_OUM=m2@XM=i2@\\M=k7N3M3M3MPdi1"}, "label": "horse with light colored saddle"}]}
{"id": 433874, "image": "COCO_train2014_000000433874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a horse riding in front of a man on a horse\"."}], "task": "refering", "answer_template": "The \"a man on a horse riding in front of a man on a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 247, 267, 268, 270, 293], "bbox": [0.573390625, 0.3462295081967213, 0.78340625, 0.8234192037470727], "iscrowd": 0, "area": 9838.101349999997, "rle": {"size": [427, 640], "counts": "iZi41Y=3M3M2`Ni0hD[OU;]1O10ObNZNYGf1g8[OYFe0g9^11UO`FkM`9S2m0N2N3M2N3N101O0010O010O3N1N3N1N01\\OPElNQ;S1QEkNP;W1oDfNS;\\1lDbNU;`1jD^NW;k0gDG1\\O=l0^9NYFILZO:]1W9]OiFLLm0Y9TOQGNAR1]9mNXGk1f8RN^Gh0ZOFW9@aGi0AAm8BdGm0G[Od8EhGn0MXOY8GmGP1MXOU8EPHe0QO]Om09P8CTH>XOEg09V9J\\FLa08U:HmE5U:JmE4S:LoE1R:OPFOP:1e10C3]CL^<<_CE\\<a0bC_OY<?dCH0IW<>SDHf0Ln9`0_EAa0?]95UFYO=i0S94^G2Z81eG2W81TGc0j8@_FU1_9i0dFYMZ9c2nF[Mo8a2[G\\Mc8e2`GYM^8g2fGWMW8j2nGSMQ8l2SHRMl7n2W101]Oc0N12O0O2O001N101O0O101O0O2O0O2O001N101O1N3N3L4M2N3_O`0O2O1O0O2O1O0O2O1O1N2O1O2MjNVGmMi8o1^GoM_8n1iGPNT8n1RHQNl7U1dGYNg0i0Z7d0jIGh5OfJ5Q5GWK<b4CcK?X4`NhHe0V3k0n3_NoHc0`3d0_3hNTIa0k3<P3QOXI?n3?g2QO^I=P4?S3^ORM=Q3_OUM=m2@XM=i2@\\M=k7N3M3M3MPdi1"}, "label": "a man on a horse riding in front of a man on a horse"}]}
{"id": 171736, "image": "COCO_train2014_000000171736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a panasonic mobile front view\"."}], "task": "refering", "answer_template": "The \"a panasonic mobile front view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 44, 45, 46, 47, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 188, 189, 190], "bbox": [0.4011, 0.12135135135135135, 0.6645, 0.844954954954955], "iscrowd": 0, "area": 25193.58765000001, "rle": {"size": [333, 500], "counts": "ibQ2j0[9c0]O;E9G8G:@?nNS1B=L5K4K6K4L5K5K4L5J5M4N1O2N1N3N1O1O1O1O1O10000O2O0O10000O100O1M3L4L4M3L4M3L4M3O11O000000000000000001O0001O2N5K?A000000000000000000000000O10000000000000000000O100000001O0O100000001O0O10000fN_LhKa3W4`LiK`3V4bLiK_3U4bLjK_3U4bLkK^3T4cLlK]3S4eLlK[3S4fLmKZ3R4gLlK\\3Q4fLmK\\3R4eLlK]3S4eLiK^3U4dLiK^3V4cLhK_3W4bLfKb3X4_LfKc3X4`LeKc3Z4]LdKe3\\4[LaKi3^4WL`Kk3`4R1O1A>D=I6K6J5J7J5KRbf1"}, "label": "a panasonic mobile front view"}]}
{"id": 171736, "image": "COCO_train2014_000000171736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink panasonic cell phone\"."}], "task": "refering", "answer_template": "The \"a pink panasonic cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 44, 45, 46, 47, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 188, 189, 190], "bbox": [0.4011, 0.12135135135135135, 0.6645, 0.844954954954955], "iscrowd": 0, "area": 25193.58765000001, "rle": {"size": [333, 500], "counts": "ibQ2j0[9c0]O;E9G8G:@?nNS1B=L5K4K6K4L5K5K4L5J5M4N1O2N1N3N1O1O1O1O1O10000O2O0O10000O100O1M3L4L4M3L4M3L4M3O11O000000000000000001O0001O2N5K?A000000000000000000000000O10000000000000000000O100000001O0O100000001O0O10000fN_LhKa3W4`LiK`3V4bLiK_3U4bLjK_3U4bLkK^3T4cLlK]3S4eLlK[3S4fLmKZ3R4gLlK\\3Q4fLmK\\3R4eLlK]3S4eLiK^3U4dLiK^3V4cLhK_3W4bLfKb3X4_LfKc3X4`LeKc3Z4]LdKe3\\4[LaKi3^4WL`Kk3`4R1O1A>D=I6K6J5J7J5KRbf1"}, "label": "a pink panasonic cell phone"}]}
{"id": 171736, "image": "COCO_train2014_000000171736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of the pink cellphone\"."}], "task": "refering", "answer_template": "The \"the reflection of the pink cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 46, 47, 48, 49, 65, 66, 67, 83, 84, 85, 86, 101, 102, 103, 104, 120, 121, 122, 123, 138, 139, 140, 141, 157, 158, 159, 176], "bbox": [0.5417799999999999, 0.13483483483483483, 0.87852, 0.7730330330330331], "iscrowd": 0, "area": 12174.691100000002, "rle": {"size": [333, 500], "counts": "`Uh25Y:1N1O2O0O2O0O101N0001N2O1O1O1N2O1O1N100010O00001O01O01O000001O0001O01O0001O0010O0100O0010O01O010O10O010O10O01000iG>a5DnIm0Q6UOgIR1X6PO`IW1^6lNZIZ1f6POjHW1T7S1N2N2N1O2N3M2N2N2N2M3M4M2M3GQLoIR4P6nKoIT4o5mKoIV4P65O1N2O1N101N2O1N2O0OO2O1N101N1010O01O001O0010O01O1O001O101N3M2N2N210O01O00101N1O2ON1O100O2N1O101N2N1N3M3L4M3M3M3L4M3M3L4M3M3L4M3M3M3L4M2N3L4M3M3L4M2N2N2M3N2N2M3N2N2M3N2N2M3N2N9G\\cc0"}, "label": "the reflection of the pink cellphone"}]}
{"id": 171736, "image": "COCO_train2014_000000171736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of the phone in the mirror\"."}], "task": "refering", "answer_template": "The \"the reflection of the phone in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 46, 47, 48, 49, 65, 66, 67, 83, 84, 85, 86, 101, 102, 103, 104, 120, 121, 122, 123, 138, 139, 140, 141, 157, 158, 159, 176], "bbox": [0.5417799999999999, 0.13483483483483483, 0.87852, 0.7730330330330331], "iscrowd": 0, "area": 12174.691100000002, "rle": {"size": [333, 500], "counts": "`Uh25Y:1N1O2O0O2O0O101N0001N2O1O1O1N2O1O1N100010O00001O01O01O000001O0001O01O0001O0010O0100O0010O01O010O10O010O10O01000iG>a5DnIm0Q6UOgIR1X6PO`IW1^6lNZIZ1f6POjHW1T7S1N2N2N1O2N3M2N2N2N2M3M4M2M3GQLoIR4P6nKoIT4o5mKoIV4P65O1N2O1N101N2O1N2O0OO2O1N101N1010O01O001O0010O01O1O001O101N3M2N2N210O01O00101N1O2ON1O100O2N1O101N2N1N3M3L4M3M3M3L4M3M3L4M3M3L4M3M3M3L4M2N3L4M3M3L4M2N2N2M3N2N2M3N2N2M3N2N2M3N2N9G\\cc0"}, "label": "the reflection of the phone in the mirror"}]}
{"id": 237273, "image": "COCO_train2014_000000237273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with a white mane standing behind another similar horse\"."}], "task": "refering", "answer_template": "The \"a horse with a white mane standing behind another similar horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151], "bbox": [0.48167187499999997, 0.22839912280701755, 1.0, 0.4119956140350877], "iscrowd": 0, "area": 14258.646849999997, "rle": {"size": [456, 640], "counts": "_ZY43S>9H1O2M2O1O2M2O1O1O1N2O1O1N2O1O1O1N2O1O1000001N100000000O100O1N101O1O1O1O1O1O001O1O1O1O1O1O001O100O1O100O010O100O0010O0010O01O01O01O100O1O101N1O1O100O1O100O2N100O1O100O1O100O2N100O1O010O1O10O01O100O00100O1O010O1O10O01O11O0O10001O00000O2O0000001N10000000000O10000000O10O100000000O1000O100000O10000000O010000000000O1000O1000O1000000000O0100001O00001O000O2O00001O00001O000001O00001O0000001O0000001O00000000000001O00000000000000001O0000000000000000001O00000000000000001O0000000000O1000000O10000O10000O1000000O10000O1000000O10000O1000000000001O0000001N10001O001O001O1O001O1O001O0O2O001O001O01O01O001O01O01O001O01O01O0010O0001O0O101O000O101O000O2O03M2lNhC4Z<\\OVDc0i<N2M3L5J5L4K5LUK"}, "label": "a horse with a white mane standing behind another similar horse"}]}
{"id": 237273, "image": "COCO_train2014_000000237273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of the horse ' s head in the back\"."}], "task": "refering", "answer_template": "The \"the top of the horse ' s head in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151], "bbox": [0.48167187499999997, 0.22839912280701755, 1.0, 0.4119956140350877], "iscrowd": 0, "area": 14258.646849999997, "rle": {"size": [456, 640], "counts": "_ZY43S>9H1O2M2O1O2M2O1O1O1N2O1O1N2O1O1O1N2O1O1000001N100000000O100O1N101O1O1O1O1O1O001O1O1O1O1O1O001O100O1O100O010O100O0010O0010O01O01O01O100O1O101N1O1O100O1O100O2N100O1O100O1O100O2N100O1O010O1O10O01O100O00100O1O010O1O10O01O11O0O10001O00000O2O0000001N10000000000O10000000O10O100000000O1000O100000O10000000O010000000000O1000O1000O1000000000O0100001O00001O000O2O00001O00001O000001O00001O0000001O0000001O00000000000001O00000000000000001O0000000000000000001O00000000000000001O0000000000O1000000O10000O10000O1000000O10000O1000000O10000O1000000000001O0000001N10001O001O001O1O001O1O001O0O2O001O001O01O01O001O01O01O001O01O01O0010O0001O0O101O000O101O000O2O03M2lNhC4Z<\\OVDc0i<N2M3L5J5L4K5LUK"}, "label": "the top of the horse ' s head in the back"}]}
{"id": 237273, "image": "COCO_train2014_000000237273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse with black straps on it\"."}], "task": "refering", "answer_template": "The \"a white horse with black straps on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 136, 137, 145, 146, 147, 148, 149, 150, 151, 155, 156, 157, 158, 159, 160, 161, 162, 163, 166, 167, 168, 169, 170, 171, 172, 173, 174, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.006109375, 0.31508771929824564, 1.0, 0.9896052631578948], "iscrowd": 0, "area": 103971.84654999996, "rle": {"size": [456, 640], "counts": "go1\\7l6000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000O1000000O10000O10000O10000O1000000O10000O10000O10000O10000O1000000O10000O10000O100O1O100O100O100O100O100O100O100O1O100O100O100O100O1O1O1O1O1O1O1O1O1N2O1O100O1000000000O1000000000O100000000000000000000O100000000000000000000O100000000000000O100000000000000O100000000000000O100000000000000000000O10000000000000000000000000000000000000000000O1N2N2M3N2M3N2M3N3M2M3N2M3N2M3N2N2M3N2M3N3L3N2N2M3N2M3N2M3N2N2M3N2M3N2M3N2N1N3N2M3N1O2N2N2O0O2N2O1N2N1O2O1O1O01000O100O10O10O100O1000O0100O10000O10O0100O100O1O10O01VNWFUOi9f0bFSO_9h0lFQOU9k0TGPOl8k0aGiNb8T1lG\\NZ8c1V2N1O2O1N1ROo0N2M2N3N0O1O100O1O3M2N2M40O01O01O01O010O00010O010O0010O0X>0_l00ZeN0O0100000O10000O10000O10000000000O10000000O100000000000000O1000000000000O100000000000000001O1O001O1O001O00100O001O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O100O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O2N2N2N101N2N2N2N2N2N2N1O2N1O2N1O2N1O2N1O2_EYN]8h1ZGaNe8a1RGfNn8[1jFmNV9S1bFTO^9l2O001O001O001O001O001O001O001O001O10O01O000000000000001O000000000000000000010O00000001O001O000010O01O00001O001O00001O01O01O001O00001O001O00010O01O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O2k0TOP1POP1POo0QOP1PO`E"}, "label": "a white horse with black straps on it"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with glasses sitting in a chair who is wearing a headset , a black jacket and a blue lanyard\"."}], "task": "refering", "answer_template": "The \"a woman with glasses sitting in a chair who is wearing a headset , a black jacket and a blue lanyard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 110, 111, 112, 133, 134, 135, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248], "bbox": [0.73765625, 0.22023419203747074, 0.98953125, 0.7056206088992975], "iscrowd": 0, "area": 15348.944050000002, "rle": {"size": [427, 640], "counts": "eQU6c0e<2N2M4M3M3M4L3M4K4M3M3M3M3M4O0O100000000O1000001O1N2O1O3M4L3L5L<D>B000OO2nM\\F:S8dNnHn0XO6k7XOcH?J1c7K[H0;L[78XHJe0FS7d0WHCn0Ak6o0WH]OW1[Ob6\\1VHUOk9o0TFnNl9V1RFhNo9Z1QFcNP:`1nE^NS:e1kE[NT:h1jEXNU:j1jEWNT:l1jEUNT:n1jESNT:P2kEPNT:R2jEnMW:R2hEoMX:b20000000O10O100000O1001O5K4L5K4L4Ln0RO:F:F`0@9G1OO001O1N2O1O1O1O1M3N4K5N1O2N2N2N2N1O2N2gLiGZ1Y8bNkG[1X8`NlG_1V8]NmGb1U8ZNoGd1S8WNQHh1Q8TNQHl1Q8oMSHP2o7kMTHU2n7fMUHZ2m7aMVH_2k7]MXHc2j7YMXHg2V900000010O001O001O010O001O0010O01O001O0010O01O001O2N1O2N2N1O2N1O2N2N1O2N1O7I9G8H6J5K5K5K5K5K5K2M3Njd2"}, "label": "a woman with glasses sitting in a chair who is wearing a headset , a black jacket and a blue lanyard"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman sitting on the end in a black and white polka dot shirt\"."}], "task": "refering", "answer_template": "The \"the woman sitting on the end in a black and white polka dot shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 110, 111, 112, 133, 134, 135, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248], "bbox": [0.73765625, 0.22023419203747074, 0.98953125, 0.7056206088992975], "iscrowd": 0, "area": 15348.944050000002, "rle": {"size": [427, 640], "counts": "eQU6c0e<2N2M4M3M3M4L3M4K4M3M3M3M3M4O0O100000000O1000001O1N2O1O3M4L3L5L<D>B000OO2nM\\F:S8dNnHn0XO6k7XOcH?J1c7K[H0;L[78XHJe0FS7d0WHCn0Ak6o0WH]OW1[Ob6\\1VHUOk9o0TFnNl9V1RFhNo9Z1QFcNP:`1nE^NS:e1kE[NT:h1jEXNU:j1jEWNT:l1jEUNT:n1jESNT:P2kEPNT:R2jEnMW:R2hEoMX:b20000000O10O100000O1001O5K4L5K4L4Ln0RO:F:F`0@9G1OO001O1N2O1O1O1O1M3N4K5N1O2N2N2N2N1O2N2gLiGZ1Y8bNkG[1X8`NlG_1V8]NmGb1U8ZNoGd1S8WNQHh1Q8TNQHl1Q8oMSHP2o7kMTHU2n7fMUHZ2m7aMVH_2k7]MXHc2j7YMXHg2V900000010O001O001O010O001O0010O01O001O0010O01O001O2N1O2N2N1O2N1O2N2N1O2N1O7I9G8H6J5K5K5K5K5K5K2M3Njd2"}, "label": "the woman sitting on the end in a black and white polka dot shirt"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black chair that the woman is sitting in\"."}], "task": "refering", "answer_template": "The \"the black chair that the woman is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 200, 223], "bbox": [0.716234375, 0.3402810304449649, 0.9425000000000001, 0.6876814988290397], "iscrowd": 0, "area": 4416.5126, "rle": {"size": [427, 640], "counts": "mUo51T=a0@`0_Oa0@?@a0@`0@`0_O?C13M2H9@?A`0K4N3M2N3M2N2N3B=L5M2N3M210O0010O01O010O100000000O01000000000O10000000000O10O1002N1O2M2O2N1O2N2M2O2N2N2M2O2N2M3N1O2N1N101O1N2O1OjZ=1TeB3L3N3M2M4M3L3N2M2O1N2O1O1O10000000001O0000000003M3M3M3M3M4L3M3M3M3MmW?"}, "label": "the black chair that the woman is sitting in"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is black chair on which a women sat\"."}], "task": "refering", "answer_template": "The \"there is black chair on which a women sat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 200, 223], "bbox": [0.716234375, 0.3402810304449649, 0.9425000000000001, 0.6876814988290397], "iscrowd": 0, "area": 4416.5126, "rle": {"size": [427, 640], "counts": "mUo51T=a0@`0_Oa0@?@a0@`0@`0_O?C13M2H9@?A`0K4N3M2N3M2N2N3B=L5M2N3M210O0010O01O010O100000000O01000000000O10000000000O10O1002N1O2M2O2N1O2N2M2O2N2N2M2O2N2M3N1O2N1N101O1N2O1OjZ=1TeB3L3N3M2M4M3L3N2M2O1N2O1O1O10000000001O0000000003M3M3M3M3M4L3M3M3M3MmW?"}, "label": "there is black chair on which a women sat"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair that the man in the middle is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair that the man in the middle is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 104, 105, 122, 123, 124, 125, 128, 145, 146, 168, 169, 191, 214, 237], "bbox": [0.32534375, 0.2314519906323185, 0.58921875, 0.6876346604215456], "iscrowd": 0, "area": 5636.1371500000005, "rle": {"size": [427, 640], "counts": "cPg22U=4L4K5]NT1lDgN`9j2jFoLQ9T3`01N3N1N@a01O1N2O1N2O1N2O1O13J5K6K4K6J5K6J5N3O001O001O001O0010O010O10O100O01000O100O01000O100O10O10O100O10000O10000O100O01000O100O10000O2O5J5L5K5J5L_fj0XOTZUO6L3N2O1O2M2O1C>O01O10O01O010O001O100O5K5L4K6J5Lm`]3"}, "label": "the chair that the man in the middle is sitting in"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair that a man with black hair is sitting in\"."}], "task": "refering", "answer_template": "The \"a chair that a man with black hair is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 104, 105, 122, 123, 124, 125, 128, 145, 146, 168, 169, 191, 214, 237], "bbox": [0.32534375, 0.2314519906323185, 0.58921875, 0.6876346604215456], "iscrowd": 0, "area": 5636.1371500000005, "rle": {"size": [427, 640], "counts": "cPg22U=4L4K5]NT1lDgN`9j2jFoLQ9T3`01N3N1N@a01O1N2O1N2O1N2O1O13J5K6K4K6J5K6J5N3O001O001O001O0010O010O10O100O01000O100O01000O100O10O10O100O10000O10000O100O01000O100O10000O2O5J5L5K5J5L_fj0XOTZUO6L3N2O1O2M2O1C>O01O10O01O010O001O100O5K5L4K6J5Lm`]3"}, "label": "a chair that a man with black hair is sitting in"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man in a suit with headphones on at a meeting\"."}], "task": "refering", "answer_template": "The \"an older man in a suit with headphones on at a meeting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236], "bbox": [0.0015, 0.1636768149882904, 0.286359375, 0.7254566744730678], "iscrowd": 0, "area": 27937.201249999995, "rle": {"size": [427, 640], "counts": "Vb0i3a92O0O100O2O0O1O101N100O101N100O2O0O1O101N100POiKkHX4Q7RLiHo3S7ZLfHg3W7aLdH_3Y7jL`HW3]7^1L3M3L4O2O0O1O2N1O100O2N1O1O101N100O101O0000001N10001O00000O2O0000001O0O10001O00000O2O00001O01O00100O001O010O1O0010O01O00100O0010O01O2N2O1N3M2N3N1cJgId3\\6SLlIk3V6mKPJS4R6fKTJZ4n5]KXJc4Y7O0010O00010O01O0010O01O001O001O00001O001O001O001O001O001O001O00001O001O001O005K6J6J6J6J7J5J6J6J6J3M2N3M3N1N3M2N3M3M2N3M2O2M3L2M4M2N2M4M2M3N3L3N2N3L3N3LTYn5"}, "label": "an older man in a suit with headphones on at a meeting"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue tie and glasses\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue tie and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236], "bbox": [0.0015, 0.1636768149882904, 0.286359375, 0.7254566744730678], "iscrowd": 0, "area": 27937.201249999995, "rle": {"size": [427, 640], "counts": "Vb0i3a92O0O100O2O0O1O101N100O101N100O2O0O1O101N100POiKkHX4Q7RLiHo3S7ZLfHg3W7aLdH_3Y7jL`HW3]7^1L3M3L4O2O0O1O2N1O100O2N1O1O101N100O101O0000001N10001O00000O2O0000001O0O10001O00000O2O00001O01O00100O001O010O1O0010O01O00100O0010O01O2N2O1N3M2N3N1cJgId3\\6SLlIk3V6mKPJS4R6fKTJZ4n5]KXJc4Y7O0010O00010O01O0010O01O001O001O00001O001O001O001O001O001O001O00001O001O001O005K6J6J6J6J7J5J6J6J6J3M2N3M3N1N3M2N3M3M2N3M2O2M3L2M4M2N2M4M2M3N3L3N2N3L3N3LTYn5"}, "label": "a man wearing a blue tie and glasses"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit at a table\"."}], "task": "refering", "answer_template": "The \"a man in a suit at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241], "bbox": [0.33284375, 0.1393208430913349, 0.6731875, 0.730327868852459], "iscrowd": 0, "area": 32445.021449999997, "rle": {"size": [427, 640], "counts": "QSi2?Z<c0K4L4M3L4L4M3L4L4L5L3L4L4L4M3L4L4M3L4L3M4M2M4L4M2N300O010O00100O00100O0010O010O0O1O2N1O1M4I6J6M4N1O2M2O1O1O2N1O1NfFgMg7X2kGXNU8f1hG_NW8`1hGcNW8\\1gGhNY8V1fGmNY8R1eGROZ8l0fGWOY8h0fG[OZ8c0dGBZ8=fGEY8:hGHV87kGKT83nGMe6RNRJo1[O1\\6ZNTJd1A4S6bNYJX1F6k5lNZJm0L9b5TO^Jb01<^5XO\\J;7>\\5[OZJ5<a0Y5^OVJ0b0c0X5@RJLg0f0U5[1nJfNQ5X1PKiNP5U1QKmNm4T1RKmNn4T1QKmNm4T1RKmNn4S1QKoNn4Q1RKoNm4S1PKPOo4P1mJTOS5l0hJZOW5f0eJ^OZ5d0`JB_5>\\JGd5Q4O001N101O00O010O10000RObJnI^5R6iJhIW5X6oJbIQ5]6VK\\Ik4d6m0010000O10O02O01O0001O000000001O000000001O0000001O00001O4M3L4L5K5K4L5K5bKVHQ3n7jLSHV3R8eLnG[3X8^LjGa3\\8XLeGh3b8QL^Go3h8jKYGV4n80000000001O000000000000001O000000001O001O00001O001O001O0O2O00001O001O001N101O001O000O20O01O0010O01O001O01O01O001O001O0O2O00001O0O2cMQRg2"}, "label": "a man in a suit at a table"}]}
{"id": 523995, "image": "COCO_train2014_000000523995.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a red tie\"."}], "task": "refering", "answer_template": "The \"a man with a red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 79, 80, 81, 102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241], "bbox": [0.33284375, 0.1393208430913349, 0.6731875, 0.730327868852459], "iscrowd": 0, "area": 32445.021449999997, "rle": {"size": [427, 640], "counts": "QSi2?Z<c0K4L4M3L4L4M3L4L4L5L3L4L4L4M3L4L4M3L4L3M4M2M4L4M2N300O010O00100O00100O0010O010O0O1O2N1O1M4I6J6M4N1O2M2O1O1O2N1O1NfFgMg7X2kGXNU8f1hG_NW8`1hGcNW8\\1gGhNY8V1fGmNY8R1eGROZ8l0fGWOY8h0fG[OZ8c0dGBZ8=fGEY8:hGHV87kGKT83nGMe6RNRJo1[O1\\6ZNTJd1A4S6bNYJX1F6k5lNZJm0L9b5TO^Jb01<^5XO\\J;7>\\5[OZJ5<a0Y5^OVJ0b0c0X5@RJLg0f0U5[1nJfNQ5X1PKiNP5U1QKmNm4T1RKmNn4T1QKmNm4T1RKmNn4S1QKoNn4Q1RKoNm4S1PKPOo4P1mJTOS5l0hJZOW5f0eJ^OZ5d0`JB_5>\\JGd5Q4O001N101O00O010O10000RObJnI^5R6iJhIW5X6oJbIQ5]6VK\\Ik4d6m0010000O10O02O01O0001O000000001O000000001O0000001O00001O4M3L4L5K5K4L5K5bKVHQ3n7jLSHV3R8eLnG[3X8^LjGa3\\8XLeGh3b8QL^Go3h8jKYGV4n80000000001O000000000000001O000000001O001O00001O001O001O0O2O00001O001O001N101O001O000O20O01O0010O01O001O01O01O001O001O0O2O00001O0O2cMQRg2"}, "label": "a man with a red tie"}]}
{"id": 24287, "image": "COCO_train2014_000000024287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe with a close up head shot\"."}], "task": "refering", "answer_template": "The \"the giraffe with a close up head shot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336], "bbox": [0.152921875, 0.10562060889929743, 0.8456093750000001, 0.9887587822014052], "iscrowd": 0, "area": 83290.15590000001, "rle": {"size": [427, 640], "counts": "gRY15T=6I6J7L4K4M4L4K2O1O2QOjNdEX1Z:nNaES1\\:SO`En0^:WO]El0_:ZO]Eg0a:S1N2M4M2M3N2N3L3N1N2O2N100O2N1O1O2N1O101N1O1O2N1O1O1O101N1O1O1O1O1O1O1O2N1O100O1O1O1O1O100000000000000000000000001O1O2N2N1O2N1O2O0O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1N3N2N1N3N1N3N1O2M101N101N10001N101N101N100O2O0O2O0O2O0O2O0O101N100O101N100N2O2N1N2O1O2M2O1O1N3N1O1N2O1O2M2O1O1N2O1O1O1O1O1O1N2O1O1O1O2N1O2N1O1O2N1O2N1O1N3N1O2N1O1O3M4L5J5L4L5K4L5K4L5L3M3L3N2M4M2N2M4M2N2M3N1N10000O10000O2O000O10000O10001N10lE`N]8`1_GdNa8[1\\GjNc8V1[GlNf8T1XGmNh8T1VGmNj8S1UGoNj8R1TGoNl8R1QGPOo8Q1oFPOQ9P1nFROQ9o0nFQOR9P1lFQOT9o0kFROU9o0jFQOV9P1hFQOX9P1fFPO[9P1eFPO[9W1]FiNd9_1SFbNm9]21O100O100O100O100O100O100O1O100O100O100O100O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1OlKdM[M[2f2jMUMU2k2PNQMo1P3VNjLj1W3YNfLf1[3]NbLb1_3aN^L]1d3fNYLW1j3lNSLR1o3QOnKm03SNh0T1ROg06YNf0S1QOb07`Ng0Q1oN=9gNf0n0oN99nNg0k0nN5:UOf0h0nN1:\\Og0e0mNO9@ZNQOV1b15M8GUNnN]1a14J71nMhNf1_13H6;fMcNP2\\12F6V1HVO0D5Z1JUONA5]1MTOL_O5`1NTOJ\\O5d10SOHYO6f11TOGVO5j13CUOcNe0n15BSO`Nf0P27CPO]Nf0T29BlN\\Ni0U2;BdN]Nn0T2?A\\N_NR1T2b0_OVN`NV1T2d0c1YO`Nh0_1VOdNi0]1UOeNk0[1SOgNm0X1ROiNo0W1POjNP1U1oNmNQ1S1nNmNR1S1mNoNS1Q1kNQOU1n0kNROV1n0hNTOX1m0eNUOZ1m0dNTO\\1m0aNTO`1m0\\NVOd1k0XNXOh1i0SNZOn1g0nM[OR2P60O1O100O00100O1O1O10O01O100O1O10O01O1O100O00100O001O01O00001O01O01O0000001O01O01O000000000O10001O0O100000000O101O000O100000001N2O001O1N2O1O1O1O1N2O1O1O1N101O1O1N2O1O1O1O1N2O1O1O1N101O1O1O1N2O1O1N2N2M4M2N3M2M3N3M2N3L3NhoX1"}, "label": "the giraffe with a close up head shot"}]}
{"id": 24287, "image": "COCO_train2014_000000024287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giarrafe loking at you\"."}], "task": "refering", "answer_template": "The \"giarrafe loking at you\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 37, 38, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336], "bbox": [0.152921875, 0.10562060889929743, 0.8456093750000001, 0.9887587822014052], "iscrowd": 0, "area": 83290.15590000001, "rle": {"size": [427, 640], "counts": "gRY15T=6I6J7L4K4M4L4K2O1O2QOjNdEX1Z:nNaES1\\:SO`En0^:WO]El0_:ZO]Eg0a:S1N2M4M2M3N2N3L3N1N2O2N100O2N1O1O2N1O101N1O1O2N1O1O1O101N1O1O1O1O1O1O1O2N1O100O1O1O1O1O100000000000000000000000001O1O2N2N1O2N1O2O0O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1N3N2N1N3N1N3N1O2M101N101N10001N101N101N100O2O0O2O0O2O0O2O0O101N100O101N100N2O2N1N2O1O2M2O1O1N3N1O1N2O1O2M2O1O1N2O1O1O1O1O1O1N2O1O1O1O2N1O2N1O1O2N1O2N1O1N3N1O2N1O1O3M4L5J5L4L5K4L5K4L5L3M3L3N2M4M2N2M4M2N2M3N1N10000O10000O2O000O10000O10001N10lE`N]8`1_GdNa8[1\\GjNc8V1[GlNf8T1XGmNh8T1VGmNj8S1UGoNj8R1TGoNl8R1QGPOo8Q1oFPOQ9P1nFROQ9o0nFQOR9P1lFQOT9o0kFROU9o0jFQOV9P1hFQOX9P1fFPO[9P1eFPO[9W1]FiNd9_1SFbNm9]21O100O100O100O100O100O100O1O100O100O100O100O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1OlKdM[M[2f2jMUMU2k2PNQMo1P3VNjLj1W3YNfLf1[3]NbLb1_3aN^L]1d3fNYLW1j3lNSLR1o3QOnKm03SNh0T1ROg06YNf0S1QOb07`Ng0Q1oN=9gNf0n0oN99nNg0k0nN5:UOf0h0nN1:\\Og0e0mNO9@ZNQOV1b15M8GUNnN]1a14J71nMhNf1_13H6;fMcNP2\\12F6V1HVO0D5Z1JUONA5]1MTOL_O5`1NTOJ\\O5d10SOHYO6f11TOGVO5j13CUOcNe0n15BSO`Nf0P27CPO]Nf0T29BlN\\Ni0U2;BdN]Nn0T2?A\\N_NR1T2b0_OVN`NV1T2d0c1YO`Nh0_1VOdNi0]1UOeNk0[1SOgNm0X1ROiNo0W1POjNP1U1oNmNQ1S1nNmNR1S1mNoNS1Q1kNQOU1n0kNROV1n0hNTOX1m0eNUOZ1m0dNTO\\1m0aNTO`1m0\\NVOd1k0XNXOh1i0SNZOn1g0nM[OR2P60O1O100O00100O1O1O10O01O100O1O10O01O1O100O00100O001O01O00001O01O01O0000001O01O01O000000000O10001O0O100000000O101O000O100000001N2O001O1N2O1O1O1O1N2O1O1O1N101O1O1N2O1O1O1O1N2O1O1O1N101O1O1O1N2O1O1N2N2M4M2N3M2M3N3M2N3L3NhoX1"}, "label": "giarrafe loking at you"}]}
{"id": 409316, "image": "COCO_train2014_000000409316.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man talking on a cell phone and getting ready to take notes while sitting next to a wall\"."}], "task": "refering", "answer_template": "The \"a man talking on a cell phone and getting ready to take notes while sitting next to a wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 174, 175, 176, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268], "bbox": [0.47334375, 0.435807962529274, 0.764859375, 0.791662763466042], "iscrowd": 0, "area": 11151.799250000002, "rle": {"size": [427, 640], "counts": "Ven31Z=1O1O2N1N2O1O1O1N2O2N1O1N2O00000O10000000000000001O0000000000000000000000O10000O1000001N10000O101O0O10O100O01000O01000O010N2N1O2O1N2N1O2O1N1O2N2N101N2N2`MnNeHS1[7TO^Hm0a7ZOWHh0h7^OQHc0P8ChG?W8HaG:^8L\\G4d83TGOl87lFKS9<eFFZ9U20100O001O01O00000010O2N2N2O1N1O00100O11O1O1O1N101O1O1O1O1O1N2O1O1O1O1O1O1N2O001O1O1O1O1N2O1O1dMnE]1S:`NoE`1R:^NPFa1R:[NPFd1R:YNPFg1R:UNPFk1Q:RNRFm1P:PNQFP2P:mMRFS2c:01N1000001O00000O2O00001O0000001O0000001O0000001O0000001O00001O000O1O5K7I6J7J5J7I7I6Jfen1"}, "label": "a man talking on a cell phone and getting ready to take notes while sitting next to a wall"}]}
{"id": 409316, "image": "COCO_train2014_000000409316.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a cell phone\"."}], "task": "refering", "answer_template": "The \"a man on a cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 174, 175, 176, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268], "bbox": [0.47334375, 0.435807962529274, 0.764859375, 0.791662763466042], "iscrowd": 0, "area": 11151.799250000002, "rle": {"size": [427, 640], "counts": "Ven31Z=1O1O2N1N2O1O1O1N2O2N1O1N2O00000O10000000000000001O0000000000000000000000O10000O1000001N10000O101O0O10O100O01000O01000O010N2N1O2O1N2N1O2O1N1O2N2N101N2N2`MnNeHS1[7TO^Hm0a7ZOWHh0h7^OQHc0P8ChG?W8HaG:^8L\\G4d83TGOl87lFKS9<eFFZ9U20100O001O01O00000010O2N2N2O1N1O00100O11O1O1O1N101O1O1O1O1O1N2O1O1O1O1O1O1N2O001O1O1O1O1N2O1O1dMnE]1S:`NoE`1R:^NPFa1R:[NPFd1R:YNPFg1R:UNPFk1Q:RNRFm1P:PNQFP2P:mMRFS2c:01N1000001O00000O2O00001O0000001O0000001O0000001O0000001O00001O000O1O5K7I6J7J5J7I7I6Jfen1"}, "label": "a man on a cell phone"}]}
{"id": 392936, "image": "COCO_train2014_000000392936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the entire blue and grey train\"."}], "task": "refering", "answer_template": "The \"the entire blue and grey train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 130, 131, 132, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179], "bbox": [0.0045, 0.37992974238875876, 0.806625, 0.5012646370023419], "iscrowd": 0, "area": 18659.74005, "rle": {"size": [427, 640], "counts": "b]1T1W<000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000^OiP:e0SoE100O100O100O100O100O100O10000O100O100O100O100O12N5K2N0000000000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000O1000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O10O100000000000000000000000000000000000000000000000000O10000000000000000000000000N2K51O0000001O0000001O000O2O0000001O0000001O00001O001O1O1O2N1O1O2N1O1O2N1O1O1N3N1O1N3L3N2M4Mb`c1"}, "label": "the entire blue and grey train"}]}
{"id": 392936, "image": "COCO_train2014_000000392936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray and blue train on the track\"."}], "task": "refering", "answer_template": "The \"the gray and blue train on the track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 130, 131, 132, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179], "bbox": [0.0045, 0.37992974238875876, 0.806625, 0.5012646370023419], "iscrowd": 0, "area": 18659.74005, "rle": {"size": [427, 640], "counts": "b]1T1W<000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000^OiP:e0SoE100O100O100O100O100O100O10000O100O100O100O100O12N5K2N0000000000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000O1000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000O10O100000000000000000000000000000000000000000000000000O10000000000000000000000000N2K51O0000001O0000001O000O2O0000001O0000001O00001O001O1O1O2N1O1O2N1O1O2N1O1O1N3N1O1N3L3N2M4Mb`c1"}, "label": "the gray and blue train on the track"}]}
{"id": 392936, "image": "COCO_train2014_000000392936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white train\"."}], "task": "refering", "answer_template": "The \"the white train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 130, 131, 133, 134, 135, 156, 157, 158, 159, 179, 180, 181, 182], "bbox": [0.374640625, 0.37756440281030446, 0.9255156250000001, 0.48882903981264636], "iscrowd": 0, "area": 4761.12235, "rle": {"size": [427, 640], "counts": "jWT31Y=1N2O1O1000000000000O10000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000O10000000000000000000000000000000000000N]P:KjoE001O001O1O001O1O001O1O001O1O2N2N2N2N2lC]O\\;e0aD\\O_;f0]D\\Oc;f0YD\\Og;f0VD[Oj;g0RD[OS<l0000001O0000000000000000000000000000000000000000000000000000000000001O0000000000000000000O100000000000000000000000000000000O1001O010O001O00001O<D5K1Hjjc0"}, "label": "the white train"}]}
{"id": 392936, "image": "COCO_train2014_000000392936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white train behind the blue one\"."}], "task": "refering", "answer_template": "The \"the white train behind the blue one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 128, 129, 130, 131, 133, 134, 135, 156, 157, 158, 159, 179, 180, 181, 182], "bbox": [0.374640625, 0.37756440281030446, 0.9255156250000001, 0.48882903981264636], "iscrowd": 0, "area": 4761.12235, "rle": {"size": [427, 640], "counts": "jWT31Y=1N2O1O1000000000000O10000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000O10000000000000000000000000000000000000N]P:KjoE001O001O1O001O1O001O1O001O1O2N2N2N2N2lC]O\\;e0aD\\O_;f0]D\\Oc;f0YD\\Og;f0VD[Oj;g0RD[OS<l0000001O0000000000000000000000000000000000000000000000000000000000001O0000000000000000000O100000000000000000000000000000000O1001O010O001O00001O<D5K1Hjjc0"}, "label": "the white train behind the blue one"}]}
{"id": 188140, "image": "COCO_train2014_000000188140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady standing in jeans and a pink jacket with long brown hair\"."}], "task": "refering", "answer_template": "The \"lady standing in jeans and a pink jacket with long brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 111, 126, 127, 128, 144, 145, 146, 147, 163, 164, 181, 182, 199, 200, 216, 217, 218], "bbox": [0.018539999999999997, 0.2561866666666667, 0.18202, 0.9887733333333334], "iscrowd": 0, "area": 14352.702350000001, "rle": {"size": [375, 500], "counts": "Z_3:e:n0C<G:F9J7H8H7J7H4M3L3M3M4WK^LW1e3fNaLT1c3hNaLS1d3hNaLT1f3eN]LX1g3dN[LZ1j3^N[L_1n3VNWLg1n3QNXLl1l3mMZLQ2i3hM]LU2g3fM\\LX2h3cM\\L\\2P4TMULj2]6O2N1O101O000O10000000000001O0001O0001O00000jHnLg5S3VJoLj5Q3UJQMi5P3WJQMh5o2XJQMg5Q3WJRMg5n2WJUMg5l2VJXMk4@PK]3MXMo4_OQKm3ZOhLb5_OcJG6R4DjL`5@bJH8n4T5fKkJ[4R5gKmJZ4Q5iKnJX4o4kKlJY4R5W1XO`IWLc6e3cIULa6g3fIRL^6i3k0[M]Hn0j7PO[H4_8IgGLa81cGK_82gGI[85iGHY85jGHX86kGGW87lGF[84gGI\\84gGI[84g1I]ce4"}, "label": "lady standing in jeans and a pink jacket with long brown hair"}]}
{"id": 188140, "image": "COCO_train2014_000000188140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing jeans and a pink and gray north face jacket\"."}], "task": "refering", "answer_template": "The \"a lady wearing jeans and a pink and gray north face jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110, 111, 126, 127, 128, 144, 145, 146, 147, 163, 164, 181, 182, 199, 200, 216, 217, 218], "bbox": [0.018539999999999997, 0.2561866666666667, 0.18202, 0.9887733333333334], "iscrowd": 0, "area": 14352.702350000001, "rle": {"size": [375, 500], "counts": "Z_3:e:n0C<G:F9J7H8H7J7H4M3L3M3M4WK^LW1e3fNaLT1c3hNaLS1d3hNaLT1f3eN]LX1g3dN[LZ1j3^N[L_1n3VNWLg1n3QNXLl1l3mMZLQ2i3hM]LU2g3fM\\LX2h3cM\\L\\2P4TMULj2]6O2N1O101O000O10000000000001O0001O0001O00000jHnLg5S3VJoLj5Q3UJQMi5P3WJQMh5o2XJQMg5Q3WJRMg5n2WJUMg5l2VJXMk4@PK]3MXMo4_OQKm3ZOhLb5_OcJG6R4DjL`5@bJH8n4T5fKkJ[4R5gKmJZ4Q5iKnJX4o4kKlJY4R5W1XO`IWLc6e3cIULa6g3fIRL^6i3k0[M]Hn0j7PO[H4_8IgGLa81cGK_82gGI[85iGHY85jGHX86kGGW87lGF[84gGI\\84gGI[84g1I]ce4"}, "label": "a lady wearing jeans and a pink and gray north face jacket"}]}
{"id": 556783, "image": "COCO_train2014_000000556783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white lamb\"."}], "task": "refering", "answer_template": "The \"a black and white lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 37, 38, 39, 40, 48, 49, 50, 51, 52, 60, 61, 62, 63, 72, 73, 74, 75, 85, 86], "bbox": [0.04808571428571428, 0.2838812785388128, 0.37422857142857147, 0.8904566210045661], "iscrowd": 0, "area": 8884.1768, "rle": {"size": [219, 350], "counts": "ah3o0j5;D7J4K4M4L3N2N2N2N1N3N1O1O1O2N1N2O1O2O0O1O1O2N101N1O2O0O2O00000O100000001O00000000000O1O1O100O1N2O1O001O1O1O1O1N2O1O1O1O1O1O001O1O1O10O000010OlMUM9k2D]M7d2FcM5]2MeMO\\21jMHV29QN_Oo1c0SNWOP2i0RNoNS2R1nMkNT2V1lMgNV2Y1lMcNU2^1lM_NV2b1jM[NX2e1^113N4K1O01O001O01O0000001O1O001O1O3M3M3L5K4L4L3L4L3M4K5Lcg^1"}, "label": "a black and white lamb"}]}
{"id": 556783, "image": "COCO_train2014_000000556783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spotted black & white baby sheep lying in straw\"."}], "task": "refering", "answer_template": "The \"a spotted black & white baby sheep lying in straw\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 37, 38, 39, 40, 48, 49, 50, 51, 52, 60, 61, 62, 63, 72, 73, 74, 75, 85, 86], "bbox": [0.04808571428571428, 0.2838812785388128, 0.37422857142857147, 0.8904566210045661], "iscrowd": 0, "area": 8884.1768, "rle": {"size": [219, 350], "counts": "ah3o0j5;D7J4K4M4L3N2N2N2N1N3N1O1O1O2N1N2O1O2O0O1O1O2N101N1O2O0O2O00000O100000001O00000000000O1O1O100O1N2O1O001O1O1O1O1N2O1O1O1O1O1O001O1O1O10O000010OlMUM9k2D]M7d2FcM5]2MeMO\\21jMHV29QN_Oo1c0SNWOP2i0RNoNS2R1nMkNT2V1lMgNV2Y1lMcNU2^1lM_NV2b1jM[NX2e1^113N4K1O01O001O01O0000001O1O001O1O3M3M3L5K4L4L3L4L3M4K5Lcg^1"}, "label": "a spotted black & white baby sheep lying in straw"}]}
{"id": 556783, "image": "COCO_train2014_000000556783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white baby lamb on the right of a black and white spotted lamb\"."}], "task": "refering", "answer_template": "The \"a white baby lamb on the right of a black and white spotted lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 41, 42, 43, 44, 45, 46, 53, 54, 55, 56, 57, 58, 65, 66, 67, 68, 69, 70, 71, 79], "bbox": [0.4257142857142857, 0.3453881278538813, 0.9842857142857143, 0.7746118721461187], "iscrowd": 0, "area": 10346.499999999998, "rle": {"size": [219, 350], "counts": "doo04`67O2N2N2N2N2N2N2N2N2O1N2N101N2O1N2N2O3L2O0O2N101O0O101N101N101O0O101N100O2O001N100O2O00001N10001O0O101O000O2O00001N100010O0001OCaK^N_4`1dK_N]4^1eKcNZ4]1fKcNZ4]1gKbNY4^1gKcNX4]1iKbNW4^1iKbNX4]1hKdNW4\\1jKcNV4]1jKcNV4]1jKdNU4\\1lKcNT4]1lKcNU4]1kKbNU4^1kKbNU4^1kKbNU4^1lKaNU4_1jK`NW4`1jK_NV4a1jK_NW4`1iK_NX4a1iK^NW4c1hK]NX4c1hK]NY4b1hK\\NY4d1gK[NZ4e1>10O0000JZNWKf1h4\\NXKc1g4^NYKb1g4_NXKa1g4`NYK`1g4aNYK^1f4cNZK]1e4eNZK[1f4eN[KZ1e4fN[KZ1d4hN[KY1d4gN\\KY1d4hN\\KW1d4iN\\KW1d4iN\\KW1d4jN[KW1d4iN]KV1c4kN\\KV1c4jN]KV1c4jN^KV1`4kN`KU1`4kN`KV1_4jNaKV1_4jNbKV1]4jNcKV1]4jNcKW1\\4jNcKV1]4kNcKU1\\4lNcKT1]4mNbKT1\\4nNdKQ1[4h0O10O01O0010O01O10O100000000O1000000O1000001O1N101O01O0000cNlK4S4GUL6k3F[L8e3EkLNU3NRM0m2OWMNi22WMNi21XMOg22YMOf20\\MOd20]MOd2O^M1b2M`M3`2KbM5^2JcM6a4000000000001O000000000000001O1O1O100O1O1O100O2N10Wm0"}, "label": "a white baby lamb on the right of a black and white spotted lamb"}]}
{"id": 556783, "image": "COCO_train2014_000000556783.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small white lamb laying in the straw\"."}], "task": "refering", "answer_template": "The \"the small white lamb laying in the straw\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 41, 42, 43, 44, 45, 46, 53, 54, 55, 56, 57, 58, 65, 66, 67, 68, 69, 70, 71, 79], "bbox": [0.4257142857142857, 0.3453881278538813, 0.9842857142857143, 0.7746118721461187], "iscrowd": 0, "area": 10346.499999999998, "rle": {"size": [219, 350], "counts": "doo04`67O2N2N2N2N2N2N2N2N2O1N2N101N2O1N2N2O3L2O0O2N101O0O101N101N101O0O101N100O2O001N100O2O00001N10001O0O101O000O2O00001N100010O0001OCaK^N_4`1dK_N]4^1eKcNZ4]1fKcNZ4]1gKbNY4^1gKcNX4]1iKbNW4^1iKbNX4]1hKdNW4\\1jKcNV4]1jKcNV4]1jKdNU4\\1lKcNT4]1lKcNU4]1kKbNU4^1kKbNU4^1kKbNU4^1lKaNU4_1jK`NW4`1jK_NV4a1jK_NW4`1iK_NX4a1iK^NW4c1hK]NX4c1hK]NY4b1hK\\NY4d1gK[NZ4e1>10O0000JZNWKf1h4\\NXKc1g4^NYKb1g4_NXKa1g4`NYK`1g4aNYK^1f4cNZK]1e4eNZK[1f4eN[KZ1e4fN[KZ1d4hN[KY1d4gN\\KY1d4hN\\KW1d4iN\\KW1d4iN\\KW1d4jN[KW1d4iN]KV1c4kN\\KV1c4jN]KV1c4jN^KV1`4kN`KU1`4kN`KV1_4jNaKV1_4jNbKV1]4jNcKV1]4jNcKW1\\4jNcKV1]4kNcKU1\\4lNcKT1]4mNbKT1\\4nNdKQ1[4h0O10O01O0010O01O10O100000000O1000000O1000001O1N101O01O0000cNlK4S4GUL6k3F[L8e3EkLNU3NRM0m2OWMNi22WMNi21XMOg22YMOf20\\MOd20]MOd2O^M1b2M`M3`2KbM5^2JcM6a4000000000001O000000000000001O1O1O100O1O1O100O2N10Wm0"}, "label": "the small white lamb laying in the straw"}]}
{"id": 401136, "image": "COCO_train2014_000000401136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike stands in street\"."}], "task": "refering", "answer_template": "The \"bike stands in street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 49, 50, 51, 65, 79, 80, 81, 94, 95, 96, 97, 110, 111, 112, 126, 127, 128, 138, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.0033806146572104017, 0.076234375, 1.0, 0.9865468749999999], "iscrowd": 0, "area": 100860.4818, "rle": {"size": [640, 423], "counts": "WT1`3^`02O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2^NVJlDl5S;\\JeDe5Z;cJ]D_5a;jJVDX5i;PKnCR5Q<_1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1000000O1000000O1fMmGgHS8R7UHmHk7m6[HSIe7l6]HSIc7l6_HRIb7m6`HRI`7m6bHRI^7m6dHRI\\7m6eHRI\\7m6fHRIZ7m6hHRIX7n6iHQIW7o6jHPIV7P7kHnHV7R7kHmHU7S7kHmHU7S7lHlHT7T7mHjHT7V7mHiHS7W7mHiHS7W7mHiHS7W7mHiHS7W7mHhHT7X7lHhHT7X7lHhHT7X7lHhHT7X7lHgHU7c3aGdNZ1iMU7Y7kHgHU7d3aGbNZ1jMU7e3aG`NZ1jMV7f3aG_NY1kMV7g3aG]NY1lMV7g3aG]NY1lMV7h3aG[NY1mMV7h3bG\\NV1kMY7j3aG[NU1kMZ7j3bG[NS1kM[7k3aG\\NR1iM]7k3bG\\NP1iM^7l3bG\\No0fM`7n3bG\\Nn0eM`7P4bG[Nm0eMa7P4bG]Nl0bMb7R4bG\\Nk0bMc7R4cG\\Nj0`Md7U4bG]Nh0]Mf7V4cG`Nc0ZMj7W4bGbNb0VMl7X4cGdN>TMo7X4dGgN:oLS8e1\\GN7_19mLT8d1_GO5`17mLU8b1bG11`19lLT8a1eG2Oa18kLT8`1hG4L`18lLT8^1kG5Ja17jLU8^1mG7Fa18jLU8\\1oG9Ea17gLW8\\1QH<A`18TLi8n1aG=_Oa1g9PNmF?\\O`1g9oMPG`0YOb1f9hMXGf0ROa1f9jMXGd0SOb1d9kMYGc0SOa1d9mMYGa0TOb1b9nMZG`0SOb1c9nMZG`0TOb1a9oM[G?TOa1a9QN[G=UOb1_9RN\\G<UOa1_9TN\\G:UOc1^9TN]G9UOb1^9UN^G8UOc1\\9VN_G7UOb1\\9XN_G5VOc1Z9YN_G5VOb1[9ZN_G3WOc1Y9[N`G2WOb1Y9\\NcGOUOd1X9^NgGJQOe1Z9bNiGDnNg1[9fNkG_OjNh1]9jNmGYOgNj1^9mNPHUObNk1`9QORHlNaNQ2_9TOQGcMLP1HV2]9XOmFfMOd0L[2Z9\\OhFjM46Oa2W9@dFmM8J1f2U9C`FRNh1X2j7G\\FUNh1Q2n7KXFXNh1j1R8OSF\\Ni1b1V83oE_Nh1\\1[86jEcNi1T1_89gEgNh1m0b8>cEiNj1f0e8b0_ElNj1?i8f0[EoNj18m8j0VESOk10Q9n0REVOk1IU9Q1nD[OP3\\NT8e4oGSKS8o4QHiJQ8Y5Z33M2N2N2N001O1O001O001O1O001O001O001O1O001O001O1O001O001O001Of0ZO2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N5K5K5K5K6J3M001O1O1O001O1O1O001O1O1O001O1O1O1O001O00001O00001O00001O00001O00001O0lN\\FlGd9S8^FlGb9R8bFlG_9R8cFmG]9Q8gFmGY9R8iFmGX9P8kFoGU9P8nFnGR9P8QGoGP9n7TGPHl8n7WGQHj8l7ZGQHh8l7[GSHf8j7^GTHd8k4TFaM[1cMb8j4XF]MZ1gM_8k4\\FWMX1mM^8i4_FTMV1RN\\8h4dFoLT1WNY8>mEc3k0bMR1\\NW8>nEb3n0^MP1`NT8`0PF_3Q1[Mn0eNo7c0TFZ3U1WMP;_OmCX3U1VMn:DoCT3T1UMm:IQDo2T1UMk:MTDk2S1UMk:0TDh2S1UMk:3TDf2R1TMl:6SDd2S1SMl:8TDb2R1SMl:;TDd1VO\\Nl1Bl:>TDa1YO[Nh1Cl:b0UD]1\\OYNe1El:e0UDZ1_OWNb1Gl:i0SDW1DTN_>g0mAQ1JTNY>l0nAk0S?VOm@f0V?[Ok@a0W?Ai@:Z?Gf@6\\?Ke@0^?2b@J_?8a@Da?>`@^Oa?e0^@WOe?j0\\@ROe?P1\\@kNf?W1Z@eNh?^1X@^Ni?d1Y11N3M1O1O100O1O10bNR^Oj0na0VOS^Oi0na0VOT^Oh0la0XOV^Of0ja0ZOX^Od0ia0[OY^Oc0ga0]OZ^Ob0ga0]O[^Oa0ea0_O]^O?ca0A_^O=ba0B_^O=aa0Ca^O<^a0Dd^O;\\a0Df^O;Ya0Eh^O;Wa0Ek^O:Ua0El^O;Sa0En^O<Qa0CQ_O<n`0DS_O<l`0DU_O<k`0CV_O=i`0CY_O<f`0D[_O=d`0B]_O>b`0B__O>``0Bb_O=^`0Bc_O>\\`0Be_O>Z`0Bg_O?X`0@j_O?U`0Al_O?S`0Bm_O>R`0Bo_O>o?CS@<l?ET@<j?DW@<h?DZ21N10001O0O101O00001N10001O0O101O000O2O00001O0O101O0O101N101N101N101N101N2O0O2O0O2O0105J6K5J6K5K5JjE"}, "label": "bike stands in street"}]}
{"id": 401136, "image": "COCO_train2014_000000401136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike in which a man is sitting\"."}], "task": "refering", "answer_template": "The \"a bike in which a man is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 49, 50, 51, 65, 79, 80, 81, 94, 95, 96, 97, 110, 111, 112, 126, 127, 128, 138, 141, 142, 143, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.0033806146572104017, 0.076234375, 1.0, 0.9865468749999999], "iscrowd": 0, "area": 100860.4818, "rle": {"size": [640, 423], "counts": "WT1`3^`02O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2^NVJlDl5S;\\JeDe5Z;cJ]D_5a;jJVDX5i;PKnCR5Q<_1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1000000O1000000O1fMmGgHS8R7UHmHk7m6[HSIe7l6]HSIc7l6_HRIb7m6`HRI`7m6bHRI^7m6dHRI\\7m6eHRI\\7m6fHRIZ7m6hHRIX7n6iHQIW7o6jHPIV7P7kHnHV7R7kHmHU7S7kHmHU7S7lHlHT7T7mHjHT7V7mHiHS7W7mHiHS7W7mHiHS7W7mHiHS7W7mHhHT7X7lHhHT7X7lHhHT7X7lHhHT7X7lHgHU7c3aGdNZ1iMU7Y7kHgHU7d3aGbNZ1jMU7e3aG`NZ1jMV7f3aG_NY1kMV7g3aG]NY1lMV7g3aG]NY1lMV7h3aG[NY1mMV7h3bG\\NV1kMY7j3aG[NU1kMZ7j3bG[NS1kM[7k3aG\\NR1iM]7k3bG\\NP1iM^7l3bG\\No0fM`7n3bG\\Nn0eM`7P4bG[Nm0eMa7P4bG]Nl0bMb7R4bG\\Nk0bMc7R4cG\\Nj0`Md7U4bG]Nh0]Mf7V4cG`Nc0ZMj7W4bGbNb0VMl7X4cGdN>TMo7X4dGgN:oLS8e1\\GN7_19mLT8d1_GO5`17mLU8b1bG11`19lLT8a1eG2Oa18kLT8`1hG4L`18lLT8^1kG5Ja17jLU8^1mG7Fa18jLU8\\1oG9Ea17gLW8\\1QH<A`18TLi8n1aG=_Oa1g9PNmF?\\O`1g9oMPG`0YOb1f9hMXGf0ROa1f9jMXGd0SOb1d9kMYGc0SOa1d9mMYGa0TOb1b9nMZG`0SOb1c9nMZG`0TOb1a9oM[G?TOa1a9QN[G=UOb1_9RN\\G<UOa1_9TN\\G:UOc1^9TN]G9UOb1^9UN^G8UOc1\\9VN_G7UOb1\\9XN_G5VOc1Z9YN_G5VOb1[9ZN_G3WOc1Y9[N`G2WOb1Y9\\NcGOUOd1X9^NgGJQOe1Z9bNiGDnNg1[9fNkG_OjNh1]9jNmGYOgNj1^9mNPHUObNk1`9QORHlNaNQ2_9TOQGcMLP1HV2]9XOmFfMOd0L[2Z9\\OhFjM46Oa2W9@dFmM8J1f2U9C`FRNh1X2j7G\\FUNh1Q2n7KXFXNh1j1R8OSF\\Ni1b1V83oE_Nh1\\1[86jEcNi1T1_89gEgNh1m0b8>cEiNj1f0e8b0_ElNj1?i8f0[EoNj18m8j0VESOk10Q9n0REVOk1IU9Q1nD[OP3\\NT8e4oGSKS8o4QHiJQ8Y5Z33M2N2N2N001O1O001O001O1O001O001O001O1O001O001O1O001O001O001Of0ZO2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N5K5K5K5K6J3M001O1O1O001O1O1O001O1O1O001O1O1O1O001O00001O00001O00001O00001O00001O0lN\\FlGd9S8^FlGb9R8bFlG_9R8cFmG]9Q8gFmGY9R8iFmGX9P8kFoGU9P8nFnGR9P8QGoGP9n7TGPHl8n7WGQHj8l7ZGQHh8l7[GSHf8j7^GTHd8k4TFaM[1cMb8j4XF]MZ1gM_8k4\\FWMX1mM^8i4_FTMV1RN\\8h4dFoLT1WNY8>mEc3k0bMR1\\NW8>nEb3n0^MP1`NT8`0PF_3Q1[Mn0eNo7c0TFZ3U1WMP;_OmCX3U1VMn:DoCT3T1UMm:IQDo2T1UMk:MTDk2S1UMk:0TDh2S1UMk:3TDf2R1TMl:6SDd2S1SMl:8TDb2R1SMl:;TDd1VO\\Nl1Bl:>TDa1YO[Nh1Cl:b0UD]1\\OYNe1El:e0UDZ1_OWNb1Gl:i0SDW1DTN_>g0mAQ1JTNY>l0nAk0S?VOm@f0V?[Ok@a0W?Ai@:Z?Gf@6\\?Ke@0^?2b@J_?8a@Da?>`@^Oa?e0^@WOe?j0\\@ROe?P1\\@kNf?W1Z@eNh?^1X@^Ni?d1Y11N3M1O1O100O1O10bNR^Oj0na0VOS^Oi0na0VOT^Oh0la0XOV^Of0ja0ZOX^Od0ia0[OY^Oc0ga0]OZ^Ob0ga0]O[^Oa0ea0_O]^O?ca0A_^O=ba0B_^O=aa0Ca^O<^a0Dd^O;\\a0Df^O;Ya0Eh^O;Wa0Ek^O:Ua0El^O;Sa0En^O<Qa0CQ_O<n`0DS_O<l`0DU_O<k`0CV_O=i`0CY_O<f`0D[_O=d`0B]_O>b`0B__O>``0Bb_O=^`0Bc_O>\\`0Be_O>Z`0Bg_O?X`0@j_O?U`0Al_O?S`0Bm_O>R`0Bo_O>o?CS@<l?ET@<j?DW@<h?DZ21N10001O0O101O00001N10001O0O101O000O2O00001O0O101O0O101N101N101N101N101N2O0O2O0O2O0105J6K5J6K5K5JjE"}, "label": "a bike in which a man is sitting"}]}
{"id": 155379, "image": "COCO_train2014_000000155379.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with gray hair in a black top with white splotches\"."}], "task": "refering", "answer_template": "The \"a woman with gray hair in a black top with white splotches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 213, 234, 235, 236], "bbox": [0.141578125, 0.23370833333333335, 0.291578125, 0.6202291666666667], "iscrowd": 0, "area": 12395.049299999999, "rle": {"size": [480, 640], "counts": "dkZ16e>8K5KCgA9U>FmA>n=ATBc0m;FZEFkNh0b;7REPO^Ol0Z;=REfNEQ1T;?_ED]:>aEE\\:>aEE[:?bEC\\:?bEDZ:`0cECZ:?dEDX:`0fEBW:a0fEBW:b0eE@X:d0eE_OX:f2O1O2N1O2O0O1O2N1O2N1O2N2N1O2N1O1O1O00100O1O1O1O1O1O1O001O0000000011N1O2N2N1O2O1N1O2N2N1O2O1N1O2N2N1O2N1O2N2N7H:E;E<D:F4L3M4L4L3M4L3M4L4L3M4L4L3M4L4LP1PO[Sd6"}, "label": "a woman with gray hair in a black top with white splotches"}]}
{"id": 155379, "image": "COCO_train2014_000000155379.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with short white hair , wearing a white and black pattern blouse , smiling into the camera\"."}], "task": "refering", "answer_template": "The \"a woman with short white hair , wearing a white and black pattern blouse , smiling into the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 118, 119, 120, 141, 142, 143, 144, 164, 165, 166, 167, 187, 188, 189, 190, 211, 212, 213, 234, 235, 236], "bbox": [0.141578125, 0.23370833333333335, 0.291578125, 0.6202291666666667], "iscrowd": 0, "area": 12395.049299999999, "rle": {"size": [480, 640], "counts": "dkZ16e>8K5KCgA9U>FmA>n=ATBc0m;FZEFkNh0b;7REPO^Ol0Z;=REfNEQ1T;?_ED]:>aEE\\:>aEE[:?bEC\\:?bEDZ:`0cECZ:?dEDX:`0fEBW:a0fEBW:b0eE@X:d0eE_OX:f2O1O2N1O2O0O1O2N1O2N1O2N2N1O2N1O1O1O00100O1O1O1O1O1O1O001O0000000011N1O2N2N1O2O1N1O2N2N1O2O1N1O2N2N1O2N1O2N2N7H:E;E<D:F4L3M4L4L3M4L3M4L4L3M4L4L3M4L4LP1PO[Sd6"}, "label": "a woman with short white hair , wearing a white and black pattern blouse , smiling into the camera"}]}
{"id": 155379, "image": "COCO_train2014_000000155379.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two old women smiling\"."}], "task": "refering", "answer_template": "The \"two old women smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0016875000000000002, 0.2736458333333333, 0.41384374999999995, 0.989875], "iscrowd": 0, "area": 45725.896050000025, "rle": {"size": [480, 640], "counts": "\\f0a6^83M2N2N2N2N2N2N2O0O2N2N1O2jMXIRLi6k3gIfKZ6W4kIgKV6U4oIhKS6V4PJgKR6W4QJgKo5X4TJeKn5Y4VJbKm5\\4ZJ]Kg5b4aJUKb5i4\\2O1N1O2O1N1O2N2O0O10000O10000O10000O10000000001O000001O0000000000000000001O000000001O00000010O01O001O00000000001O0PHcK]5^4aJdK^5\\4aJfK^5[4_JhK`5X4_JiKa5X4]JjKb5V4\\JnKb5S4\\JoKc5Q4[JRLd5o3ZJSLe5m3YJVLf5k3XJWLg5i3WJZLi5_6O1O1O1O1O1O1O1O6J7I3M1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O2oLnGPOV8k0lGTOX8f0kGYOY8a0jG^OW8?lG@U8>mGAT8<nGDS89oGGQ87QHIP85QHKP82RHNo7OSH1n7LTH4m7JTH6m7GUH9l7DVH<l7@VH`0k7^OVHb0k7[OWHf0j7VOXHj0i7SOYHm0h7QOYHo0h7nNZHR1h7jNZHV1g7gN[HY1f7eN[H[1f7bN\\H^1f7^N\\Hb1e7[N]He1d7YN]Hg1e7UN]Hk1d7RN^Hn1c7oM_HQ2b7mM_HS2W:0000O10000O10000O10000O10000O100O10000O10000O10000O10000O11O0000001N2O1O1O1O1O1O1O00000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000004L6J6J7I6J6J6J6J6J4L3M3M3MXZ_5"}, "label": "two old women smiling"}]}
{"id": 155379, "image": "COCO_train2014_000000155379.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with brown hair wearing a beige and brown top and light pants\"."}], "task": "refering", "answer_template": "The \"a woman with brown hair wearing a beige and brown top and light pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0016875000000000002, 0.2736458333333333, 0.41384374999999995, 0.989875], "iscrowd": 0, "area": 45725.896050000025, "rle": {"size": [480, 640], "counts": "\\f0a6^83M2N2N2N2N2N2N2O0O2N2N1O2jMXIRLi6k3gIfKZ6W4kIgKV6U4oIhKS6V4PJgKR6W4QJgKo5X4TJeKn5Y4VJbKm5\\4ZJ]Kg5b4aJUKb5i4\\2O1N1O2O1N1O2N2O0O10000O10000O10000O10000000001O000001O0000000000000000001O000000001O00000010O01O001O00000000001O0PHcK]5^4aJdK^5\\4aJfK^5[4_JhK`5X4_JiKa5X4]JjKb5V4\\JnKb5S4\\JoKc5Q4[JRLd5o3ZJSLe5m3YJVLf5k3XJWLg5i3WJZLi5_6O1O1O1O1O1O1O1O6J7I3M1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O2oLnGPOV8k0lGTOX8f0kGYOY8a0jG^OW8?lG@U8>mGAT8<nGDS89oGGQ87QHIP85QHKP82RHNo7OSH1n7LTH4m7JTH6m7GUH9l7DVH<l7@VH`0k7^OVHb0k7[OWHf0j7VOXHj0i7SOYHm0h7QOYHo0h7nNZHR1h7jNZHV1g7gN[HY1f7eN[H[1f7bN\\H^1f7^N\\Hb1e7[N]He1d7YN]Hg1e7UN]Hk1d7RN^Hn1c7oM_HQ2b7mM_HS2W:0000O10000O10000O10000O10000O100O10000O10000O10000O10000O11O0000001N2O1O1O1O1O1O1O00000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000004L6J6J7I6J6J6J6J6J4L3M3M3MXZ_5"}, "label": "a woman with brown hair wearing a beige and brown top and light pants"}]}
{"id": 155379, "image": "COCO_train2014_000000155379.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white eating at a table with other woman\"."}], "task": "refering", "answer_template": "The \"a woman in white eating at a table with other woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5309062499999999, 0.2674166666666667, 0.964046875, 1.0], "iscrowd": 0, "area": 52417.4, "rle": {"size": [480, 640], "counts": "Sdo42m>2N3M3L3N3M3N1O2O0O2N1O2N1O1O000000000000000000000000000YGAn2>PMMg23WMOi21TM2l2NRM4o2KoL7Q3IoL7Q3InL8S3GmL9S3CoG]Ol4Q1U3CoG[Ol4R1V3BoG[Ok4S1V3BmL?S3BkL?V3@hLb0X3^OgLc0Y3^OdLe0[3[OcLg0^3XOaLi0_3XO^Lj0b3VO]Lk0d3TOZLn0f3SOXLn0h3ROVLP1j3QOTLP1l3PORLR1n3nNQLS1o3nNnKT1S4kNlKV1T4jNjKX1V4iNhKX1X4hNfKZ1Z4fNeK[1[4fNcK[1]4eNcK[1]4fNaK[1_4eNaK[1_4eNaK\\1_4dN`K\\1`4dN`K\\1`4dN`K\\1`4eN^K\\1b4dN^K\\1b4eN]K[1c4eN]K[1c4eN]K[1c4fN\\KZ1d4fN[K[1e4eN[K[1f4eNYK\\1f4dNZK\\1f4dNZK\\1f4eNYK[1g4eNXK\\1h4eNWK[1i4eNWK[1i4eNWK[1i4fNVKZ1j4fNVKZ1j4fNUK[1k4fNTKZ1l4fNTKZ1l4fNTKZ1l4gNTKX1l4hNTKX1l4iNSKW1m4iNSKW1m4iNSKW1m4jNRKV1n4jNbHmNe1Y2i5jN`HPOf1V2j5kN]HROh1S2k5kN[HUOi1P2l5lNXHWOk1m1m5lNVHZOl1j1n5mNRH@l1c1R62jINV66fIJZ69cIG]6=_ICa6a0[I_Od6f0XIZOg6j0VIVOi6o0SIQOk6U1QIkNn6Z1nHfNQ7^1lHbNR7`1nH`NQ7b1nH^NQ7d1nH\\N]5fM[KQ4WOYNU5QNaKg3ZOXNl4\\NgK^3\\OVNd4gNlKV3_OSN\\4SOQLl2BQNZ4XOQLh2EPNY4[OnKh2HmMY4^OlKg2JkMY4BiKe2MiMX4FgKd20fMX4IeKb24dMV4MbKb27aMV40`Ka29_MV44\\K`2=\\MV4W5hKjJW4Y5gKgJW4]5gKcJX4`5fK_JZ4d5dK\\J[4g5cKYJ\\4j5aKWJ^4V8O1O100O100O10000O100O10000O100O10000O10000O100O10000O100O10O10O11O00000000000O2O00000000001O0O10000000001O001O2M2O1O1O1O1O2N1O1O1N2O1O4L5K5K5K5K5J6K5K5K5K5K5K5K5K6I5L1O2N1O1O2N1O1O2M2O1N3N1N2O2M2O2M3N2M4M2M4M2M3N3L3N2M4M2M4M2M3N3LW1iN5J6J6K5J6J6K5J6J6K5J6J6K5J6J6K7H8H8I6I8H8In[:"}, "label": "a woman in white eating at a table with other woman"}]}
{"id": 155379, "image": "COCO_train2014_000000155379.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a white coat and eating a salad\"."}], "task": "refering", "answer_template": "The \"a woman wearing a white coat and eating a salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5309062499999999, 0.2674166666666667, 0.964046875, 1.0], "iscrowd": 0, "area": 52417.4, "rle": {"size": [480, 640], "counts": "Sdo42m>2N3M3L3N3M3N1O2O0O2N1O2N1O1O000000000000000000000000000YGAn2>PMMg23WMOi21TM2l2NRM4o2KoL7Q3IoL7Q3InL8S3GmL9S3CoG]Ol4Q1U3CoG[Ol4R1V3BoG[Ok4S1V3BmL?S3BkL?V3@hLb0X3^OgLc0Y3^OdLe0[3[OcLg0^3XOaLi0_3XO^Lj0b3VO]Lk0d3TOZLn0f3SOXLn0h3ROVLP1j3QOTLP1l3PORLR1n3nNQLS1o3nNnKT1S4kNlKV1T4jNjKX1V4iNhKX1X4hNfKZ1Z4fNeK[1[4fNcK[1]4eNcK[1]4fNaK[1_4eNaK[1_4eNaK\\1_4dN`K\\1`4dN`K\\1`4dN`K\\1`4eN^K\\1b4dN^K\\1b4eN]K[1c4eN]K[1c4eN]K[1c4fN\\KZ1d4fN[K[1e4eN[K[1f4eNYK\\1f4dNZK\\1f4dNZK\\1f4eNYK[1g4eNXK\\1h4eNWK[1i4eNWK[1i4eNWK[1i4fNVKZ1j4fNVKZ1j4fNUK[1k4fNTKZ1l4fNTKZ1l4fNTKZ1l4gNTKX1l4hNTKX1l4iNSKW1m4iNSKW1m4iNSKW1m4jNRKV1n4jNbHmNe1Y2i5jN`HPOf1V2j5kN]HROh1S2k5kN[HUOi1P2l5lNXHWOk1m1m5lNVHZOl1j1n5mNRH@l1c1R62jINV66fIJZ69cIG]6=_ICa6a0[I_Od6f0XIZOg6j0VIVOi6o0SIQOk6U1QIkNn6Z1nHfNQ7^1lHbNR7`1nH`NQ7b1nH^NQ7d1nH\\N]5fM[KQ4WOYNU5QNaKg3ZOXNl4\\NgK^3\\OVNd4gNlKV3_OSN\\4SOQLl2BQNZ4XOQLh2EPNY4[OnKh2HmMY4^OlKg2JkMY4BiKe2MiMX4FgKd20fMX4IeKb24dMV4MbKb27aMV40`Ka29_MV44\\K`2=\\MV4W5hKjJW4Y5gKgJW4]5gKcJX4`5fK_JZ4d5dK\\J[4g5cKYJ\\4j5aKWJ^4V8O1O100O100O10000O100O10000O100O10000O10000O100O10000O100O10O10O11O00000000000O2O00000000001O0O10000000001O001O2M2O1O1O1O1O2N1O1O1N2O1O4L5K5K5K5K5J6K5K5K5K5K5K5K5K6I5L1O2N1O1O2N1O1O2M2O1N3N1N2O2M2O2M3N2M4M2M4M2M3N3L3N2M4M2M4M2M3N3LW1iN5J6J6K5J6J6K5J6J6K5J6J6K5J6J6K7H8H8I6I8H8In[:"}, "label": "a woman wearing a white coat and eating a salad"}]}
{"id": 401147, "image": "COCO_train2014_000000401147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant driver / handler\"."}], "task": "refering", "answer_template": "The \"an elephant driver / handler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 23, 24, 41, 42, 60, 61, 78, 79, 80, 97, 98], "bbox": [0.29054, 0.05856, 0.49324, 0.4459466666666667], "iscrowd": 0, "area": 4511.878499999998, "rle": {"size": [375, 500], "counts": "bTe12d;2N2N2N2N3L3N2N1O2N2M3M2N3N2N2N2iEoNe9S1YFPOe9Q1XFSOg9l0WFXOg9i0UF\\Oj9^1J5L2M2N3N0O100O100O\\NlFe0U9WOPGh0o8WOSGi0n8SOUGm0l8POWGo0k8lNXGS1k8iNXGV1f9O01O000O10000010O001O0100O00100O00101M2O2N1N3N000O10000O10000O11O1O001O001N2N2N3M1O2N2O1N1O2N2N2N2N2N2N2O1N2N2N1O1O1O1O1N5K5K5Li_l2"}, "label": "an elephant driver / handler"}]}
{"id": 401147, "image": "COCO_train2014_000000401147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with bare feet on elephant\"."}], "task": "refering", "answer_template": "The \"man with bare feet on elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 23, 24, 41, 42, 60, 61, 78, 79, 80, 97, 98], "bbox": [0.29054, 0.05856, 0.49324, 0.4459466666666667], "iscrowd": 0, "area": 4511.878499999998, "rle": {"size": [375, 500], "counts": "bTe12d;2N2N2N2N3L3N2N1O2N2M3M2N3N2N2N2iEoNe9S1YFPOe9Q1XFSOg9l0WFXOg9i0UF\\Oj9^1J5L2M2N3N0O100O100O\\NlFe0U9WOPGh0o8WOSGi0n8SOUGm0l8POWGo0k8lNXGS1k8iNXGV1f9O01O000O10000010O001O0100O00100O00101M2O2N1N3N000O10000O10000O11O1O001O001N2N2N3M1O2N2O1N1O2N2N2N2N2N2N2O1N2N2N1O1O1O1O1N5K5K5Li_l2"}, "label": "man with bare feet on elephant"}]}
{"id": 401147, "image": "COCO_train2014_000000401147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue hat riding an elephant\"."}], "task": "refering", "answer_template": "The \"a woman in a blue hat riding an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 18, 19, 20, 36, 37, 38, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110], "bbox": [0.02388, 0.01872, 0.1573, 0.5205866666666666], "iscrowd": 0, "area": 8065.504800000001, "rle": {"size": [375, 500], "counts": "R_44Q36o42mJ2P51mJ3o40TJBjN`0o61SJFiN;Q72SJHgN9S72RJP1j5SOSJP1i5TOSJP1i5TOTJP1f5UOVJo0S5nMRJY1i0k0U5FiJ<V5EhJ>V5ChJ?W5GbJ;]5K[J9c5NUJ4k5l2O1O1OM4K5K4M4I7E;E;E<F:5K5K5K5K4bIbKZ5c4VJgKk5S50O101N2O2N1N2O1N1000O01J5J7J?_O8D<E;\\Od0E;L4M4K4M3L4M4J6J6J6@`0@PTj4"}, "label": "a woman in a blue hat riding an elephant"}]}
{"id": 401147, "image": "COCO_train2014_000000401147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue beanie hat\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue beanie hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 18, 19, 20, 36, 37, 38, 55, 56, 72, 73, 74, 90, 91, 92, 108, 109, 110], "bbox": [0.02388, 0.01872, 0.1573, 0.5205866666666666], "iscrowd": 0, "area": 8065.504800000001, "rle": {"size": [375, 500], "counts": "R_44Q36o42mJ2P51mJ3o40TJBjN`0o61SJFiN;Q72SJHgN9S72RJP1j5SOSJP1i5TOSJP1i5TOTJP1f5UOVJo0S5nMRJY1i0k0U5FiJ<V5EhJ>V5ChJ?W5GbJ;]5K[J9c5NUJ4k5l2O1O1OM4K5K4M4I7E;E;E<F:5K5K5K5K4bIbKZ5c4VJgKk5S50O101N2O2N1N2O1N1000O01J5J7J?_O8D<E;\\Od0E;L4M4K4M3L4M4J6J6J6@`0@PTj4"}, "label": "a woman wearing a blue beanie hat"}]}
{"id": 401147, "image": "COCO_train2014_000000401147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large rustic man with a beard riding a gray elephant\"."}], "task": "refering", "answer_template": "The \"a large rustic man with a beard riding a gray elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 20, 21, 22, 23, 38, 39, 40, 41, 56, 57, 58, 59, 74, 75, 76, 77, 92, 93, 94, 110, 111, 112, 128], "bbox": [0.1264, 0.011386666666666665, 0.32192000000000004, 0.55296], "iscrowd": 0, "area": 10651.436500000002, "rle": {"size": [375, 500], "counts": "nWg04X;=C=C=C=B=D=H8L4L4L4M3L3M2N2QOo0J6K5D<M3N]OfImKW6k3\\JmKb5i3QKPLl4m3`101O2N2M3M3M3M3M3M3M3M3M3M3M3M3M3001O1O1O1O1^OWMTHj2Y1WMg42mIh2U1]Mm4NkIf2R1bMR5KiId2n0hMW5GiIb2i0mM^5BhIb2d0RNc5\\OiIc2=XNi5VOiIc28]Nm5W2mInMS6l3O1O1N101O1O1O0O01O1O1N2O1O1N2N2UMkIOZ63nI^OW6e0PJmNT6U1PJ`NU6c1mISNX6NQIP1l0gNX67QIn0m0aNW6`0PIl0o0ZNW6g0nHk0`8QOeGl0]8QOgGk0\\8QOiGl0Y8QOiGm0[8oNgGP1[8lNgGS1_901N101N101N101N100O2J5BVSl3"}, "label": "a large rustic man with a beard riding a gray elephant"}]}
{"id": 401147, "image": "COCO_train2014_000000401147.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man riding a elephant wearing a gray tank top\"."}], "task": "refering", "answer_template": "The \"a man riding a elephant wearing a gray tank top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 20, 21, 22, 23, 38, 39, 40, 41, 56, 57, 58, 59, 74, 75, 76, 77, 92, 93, 94, 110, 111, 112, 128], "bbox": [0.1264, 0.011386666666666665, 0.32192000000000004, 0.55296], "iscrowd": 0, "area": 10651.436500000002, "rle": {"size": [375, 500], "counts": "nWg04X;=C=C=C=B=D=H8L4L4L4M3L3M2N2QOo0J6K5D<M3N]OfImKW6k3\\JmKb5i3QKPLl4m3`101O2N2M3M3M3M3M3M3M3M3M3M3M3M3M3001O1O1O1O1^OWMTHj2Y1WMg42mIh2U1]Mm4NkIf2R1bMR5KiId2n0hMW5GiIb2i0mM^5BhIb2d0RNc5\\OiIc2=XNi5VOiIc28]Nm5W2mInMS6l3O1O1N101O1O1O0O01O1O1N2O1O1N2N2UMkIOZ63nI^OW6e0PJmNT6U1PJ`NU6c1mISNX6NQIP1l0gNX67QIn0m0aNW6`0PIl0o0ZNW6g0nHk0`8QOeGl0]8QOgGk0\\8QOiGl0Y8QOiGm0[8oNgGP1[8lNgGS1_901N101N101N101N100O2J5BVSl3"}, "label": "a man riding a elephant wearing a gray tank top"}]}
{"id": 98044, "image": "COCO_train2014_000000098044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"all of the broccoli except the left most piece\"."}], "task": "refering", "answer_template": "The \"all of the broccoli except the left most piece\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 31, 32, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 166, 167, 186], "bbox": [0.4038697318007663, 0.035296875, 0.9346934865900383, 0.40235937499999996], "iscrowd": 0, "area": 33435.9426, "rle": {"size": [640, 522], "counts": "bnS42lc03N3M3L3N3M3L3N3`_OXOb=k0[BXOc=k0ZBVOe=m0WBVOf=m0XBTOg=o0UBTOi=o0SBTOj=o0TBROk=Q1QBROl=Q1RBPOm=S1oAPOo=S1nAnNP>U1mAnNQ>U1kAmNT>V1iAlNT>W1iAkNV>X1gAjNV>Y1gAiNX>Z1eAhNY>[1cAgN\\>[1aAhN^>Z1_AgNa>[1[AgNe>Z1YAgNg>[1UAfNl>[1QAfNQ?[1l@fNT?\\1h@eNY?P31O100O1O101N1O1O100O1O100O2N1O100O1O100O1O1O2O0O1O100O1O1O101O01O00000000001O01XMo_O_1Q`0^NQ@c1o?[NR@f1n?WNT@j1l?SNV@o1i?oMX@R2h?kMZ@V2f?gM\\@Z2e?cM\\@^2d?_M^@c2a?ZMa@g2_?VMc@k2]?SMd@n2\\?oLf@R3S`01O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O100O1O2N1O1O100O1O2N1O1O1O100O2N1O1O2N2O0O2N2N1O2N2N2O0O2N2N1O2N2O0O2N2N2N1O2O10O2O1O1O0O2O1O1N101O1N101O1O1N101O1N2O001O10O0O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O101N1O1O100O1O100O1O00M4M2M3N2M4L3N2M4M2M3N2M4M2M3N2M4F9D<D=K4eMX_O^1h`0`NZ_O_1g`0\\N__Oc1b`0WNc_Oi1]`0RNi_Om1W`0nMn_OQ2Ra0O001O0O2O1O001N101O0O2]Ol]OXOTb0g0o]OWORb0g0P^OXOQb0f0R^OWOoa0h0S^OWOna0g0U^OWOka0h0W^OWOja0g0X^OWOia0h0Z^OVOga0h0[^OWOea0h0k0N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1OPTe0"}, "label": "all of the broccoli except the left most piece"}]}
{"id": 98044, "image": "COCO_train2014_000000098044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli on top of chicken\"."}], "task": "refering", "answer_template": "The \"broccoli on top of chicken\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 31, 32, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 166, 167, 186], "bbox": [0.4038697318007663, 0.035296875, 0.9346934865900383, 0.40235937499999996], "iscrowd": 0, "area": 33435.9426, "rle": {"size": [640, 522], "counts": "bnS42lc03N3M3L3N3M3L3N3`_OXOb=k0[BXOc=k0ZBVOe=m0WBVOf=m0XBTOg=o0UBTOi=o0SBTOj=o0TBROk=Q1QBROl=Q1RBPOm=S1oAPOo=S1nAnNP>U1mAnNQ>U1kAmNT>V1iAlNT>W1iAkNV>X1gAjNV>Y1gAiNX>Z1eAhNY>[1cAgN\\>[1aAhN^>Z1_AgNa>[1[AgNe>Z1YAgNg>[1UAfNl>[1QAfNQ?[1l@fNT?\\1h@eNY?P31O100O1O101N1O1O100O1O100O2N1O100O1O100O1O1O2O0O1O100O1O1O101O01O00000000001O01XMo_O_1Q`0^NQ@c1o?[NR@f1n?WNT@j1l?SNV@o1i?oMX@R2h?kMZ@V2f?gM\\@Z2e?cM\\@^2d?_M^@c2a?ZMa@g2_?VMc@k2]?SMd@n2\\?oLf@R3S`01O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O100O1O2N1O1O100O1O2N1O1O1O100O2N1O1O2N2O0O2N2N1O2N2N2O0O2N2N1O2N2O0O2N2N2N1O2O10O2O1O1O0O2O1O1N101O1N101O1O1N101O1N2O001O10O0O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O101N1O1O100O1O100O1O00M4M2M3N2M4L3N2M4M2M3N2M4M2M3N2M4F9D<D=K4eMX_O^1h`0`NZ_O_1g`0\\N__Oc1b`0WNc_Oi1]`0RNi_Om1W`0nMn_OQ2Ra0O001O0O2O1O001N101O0O2]Ol]OXOTb0g0o]OWORb0g0P^OXOQb0f0R^OWOoa0h0S^OWOna0g0U^OWOka0h0W^OWOja0g0X^OWOia0h0Z^OVOga0h0[^OWOea0h0k0N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1OPTe0"}, "label": "broccoli on top of chicken"}]}
{"id": 98044, "image": "COCO_train2014_000000098044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an individual piece of brocolli with a tear in the stem , next to a piece of cooked tomato\"."}], "task": "refering", "answer_template": "The \"an individual piece of brocolli with a tear in the stem , next to a piece of cooked tomato\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 59, 60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 118, 119, 120, 138, 139, 157, 158], "bbox": [0.1443295019157088, 0.102515625, 0.43735632183908046, 0.370796875], "iscrowd": 0, "area": 12817.0114, "rle": {"size": [640, 522], "counts": "ho^1?[c07I6N2O2O000O2O0O2O000O2O0O101O0O1O2N1O101N1O10001O0JaNk]O_1Tb0eNi]O\\1Ub08O1O1O2N1O1O2N1O1O1O2M201N101O001N1o^OcM_`0^2__OdM``0\\2__OeMa`0\\2]_OeMb`0^2[_OdMd`0^2[_OaMd`0b2\\_OcM]`0_2b_OeMX`0^2h_OaMW`0b2h_O]MV`0g2i_OYMU`0j2i_OVMU`0n2j_OQMU`0^3N3N3L5L3M3M3M3M3M3M4L3N2M3M3M3M1O0001O00000000001O0000001O000O101O0O2O1_NTAkMo>P2XAlMi>o1_AmMb>n1eAnM]>n1jAnMW>o1nAmMT>Q2QBkMQ>Q2UBkMl=R2cBaM]=^2R2O2N2O1N2N3M3L3N3O1O1O00100O001O10O01O1O1O2N4L3M1O1O00100O1O1O1O1O1O1O1O1N101O1O1O1O1O1N2[OR]O6ob0Hf0M2OUdg5"}, "label": "an individual piece of brocolli with a tear in the stem , next to a piece of cooked tomato"}]}
{"id": 98044, "image": "COCO_train2014_000000098044.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the single left most piece of broccoli\"."}], "task": "refering", "answer_template": "The \"the single left most piece of broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 59, 60, 61, 62, 63, 64, 78, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 118, 119, 120, 138, 139, 157, 158], "bbox": [0.1443295019157088, 0.102515625, 0.43735632183908046, 0.370796875], "iscrowd": 0, "area": 12817.0114, "rle": {"size": [640, 522], "counts": "ho^1?[c07I6N2O2O000O2O0O2O000O2O0O101O0O1O2N1O101N1O10001O0JaNk]O_1Tb0eNi]O\\1Ub08O1O1O2N1O1O2N1O1O1O2M201N101O001N1o^OcM_`0^2__OdM``0\\2__OeMa`0\\2]_OeMb`0^2[_OdMd`0^2[_OaMd`0b2\\_OcM]`0_2b_OeMX`0^2h_OaMW`0b2h_O]MV`0g2i_OYMU`0j2i_OVMU`0n2j_OQMU`0^3N3N3L5L3M3M3M3M3M3M4L3N2M3M3M3M1O0001O00000000001O0000001O000O101O0O2O1_NTAkMo>P2XAlMi>o1_AmMb>n1eAnM]>n1jAnMW>o1nAmMT>Q2QBkMQ>Q2UBkMl=R2cBaM]=^2R2O2N2O1N2N3M3L3N3O1O1O00100O001O10O01O1O1O2N4L3M1O1O00100O1O1O1O1O1O1O1O1N101O1O1O1O1O1N2[OR]O6ob0Hf0M2OUdg5"}, "label": "the single left most piece of broccoli"}]}
{"id": 327421, "image": "COCO_train2014_000000327421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing black shirt walking on sidewalk with hands in the air\"."}], "task": "refering", "answer_template": "The \"woman wearing black shirt walking on sidewalk with hands in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 130, 150, 151, 152, 153, 173, 174, 175, 197, 198, 220, 221, 243, 244, 266, 267], "bbox": [0.534109375, 0.33652680652680655, 0.6731093749999999, 0.773076923076923], "iscrowd": 0, "area": 6601.997000000003, "rle": {"size": [429, 640], "counts": "T__41\\=2N3M2N2N3M21O10O0M3M4L3L3N001O000OG:1O1O0O2N2O0O2O1N101N2N10m0RO:G9F:G9F:G9F:F8I4K6L4M3M2N3N2M00O1O1O3N3VO^GaLf8X3`GeLc8U3cGhL]8U3jGhLU8U3QHbLT8k2[GZMe9o28:F9G8H3M4K5L\\L\\Gb2a8^MaGa2^8_MdG_2\\8`MgG_2X8aMjG]2U8dMnGf1ZOSNg86RHa1]OXN`87UHHWOm0<ROX89WH_O[17S6d0UKZOa4o0cKoNW4V1mKgNP4[1TLaNj3`1Q41N2N2O0:G8F8G9H8G8IV\\g2"}, "label": "woman wearing black shirt walking on sidewalk with hands in the air"}]}
{"id": 327421, "image": "COCO_train2014_000000327421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in black walking away\"."}], "task": "refering", "answer_template": "The \"a person in black walking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 130, 150, 151, 152, 153, 173, 174, 175, 197, 198, 220, 221, 243, 244, 266, 267], "bbox": [0.534109375, 0.33652680652680655, 0.6731093749999999, 0.773076923076923], "iscrowd": 0, "area": 6601.997000000003, "rle": {"size": [429, 640], "counts": "T__41\\=2N3M2N2N3M21O10O0M3M4L3L3N001O000OG:1O1O0O2N2O0O2O1N101N2N10m0RO:G9F:G9F:G9F:F8I4K6L4M3M2N3N2M00O1O1O3N3VO^GaLf8X3`GeLc8U3cGhL]8U3jGhLU8U3QHbLT8k2[GZMe9o28:F9G8H3M4K5L\\L\\Gb2a8^MaGa2^8_MdG_2\\8`MgG_2X8aMjG]2U8dMnGf1ZOSNg86RHa1]OXN`87UHHWOm0<ROX89WH_O[17S6d0UKZOa4o0cKoNW4V1mKgNP4[1TLaNj3`1Q41N2N2O0:G8F8G9H8G8IV\\g2"}, "label": "a person in black walking away"}]}
{"id": 24319, "image": "COCO_train2014_000000024319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girrafe leaning to the left\"."}], "task": "refering", "answer_template": "The \"a girrafe leaning to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 141, 142, 143, 144, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 310, 311, 312, 331, 334, 335, 354, 357, 377, 380], "bbox": [0.12471874999999999, 0.31910416666666663, 0.62528125, 0.9865208333333333], "iscrowd": 0, "area": 28380.885350000015, "rle": {"size": [480, 640], "counts": "XgU12l>5L3M3L4M0O2O001N101O1N101O0O2O001N101O1N101O1N101O1N1_Ob0000O2O000O10000O101O0O10O1O0O2O1N2O001N2O0O2O1N101O1N101N2O0101O2M3N2O0O2N2O0O2O1N2O0O2N2O1N100O010O00010O010O010O0010O010O00010O010O010O01O010O010O01O01O010O010O0010O010O0010O0010O010O010O01O01O010O010O01O010O01O01O010O010O010O0010O0010_1`N2O1N101N2O1N2N2O1N101N2O1N2O0O100O1O100O100O10O0100O100O1O100dGQMe4P3VJTNj5m1QIXOn6]4000001O00O1G9G9G9G9G9G9G9G9G9G9G9G9G9H80000001O000000000000001O0000000000001O000000000000001O000001O1O001O100O1O001O1O10O4M3M4L3L4M4L3M3L5L3M3M4K4M3M5K4K5L5K4L5JkMPG_Nl8`1[G^Na8a1eG_NV8`1PH_Nl7`1[H^N`7c1eHYNZ7f1mHUNR7j1TIoMm6Q2YIbLo7]3j101O1O001O001O001O001O001O10O01O001O001O001O001O1O0014K6K5K5L4MN1L4L4J6F:G9F:G9K5J6J7J5J8I9F:F9HlU`3"}, "label": "a girrafe leaning to the left"}]}
{"id": 24319, "image": "COCO_train2014_000000024319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a giraffe in front of another\"."}], "task": "refering", "answer_template": "The \"this is a giraffe in front of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 141, 142, 143, 144, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 310, 311, 312, 331, 334, 335, 354, 357, 377, 380], "bbox": [0.12471874999999999, 0.31910416666666663, 0.62528125, 0.9865208333333333], "iscrowd": 0, "area": 28380.885350000015, "rle": {"size": [480, 640], "counts": "XgU12l>5L3M3L4M0O2O001N101O1N101O0O2O001N101O1N101O1N101O1N1_Ob0000O2O000O10000O101O0O10O1O0O2O1N2O001N2O0O2O1N101O1N101N2O0101O2M3N2O0O2N2O0O2O1N2O0O2N2O1N100O010O00010O010O010O0010O010O00010O010O010O01O010O010O01O01O010O010O0010O010O0010O0010O010O010O01O01O010O010O01O010O01O01O010O010O010O0010O0010_1`N2O1N101N2O1N2N2O1N101N2O1N2O0O100O1O100O100O10O0100O100O1O100dGQMe4P3VJTNj5m1QIXOn6]4000001O00O1G9G9G9G9G9G9G9G9G9G9G9G9G9H80000001O000000000000001O0000000000001O000000000000001O000001O1O001O100O1O001O1O10O4M3M4L3L4M4L3M3L5L3M3M4K4M3M5K4K5L5K4L5JkMPG_Nl8`1[G^Na8a1eG_NV8`1PH_Nl7`1[H^N`7c1eHYNZ7f1mHUNR7j1TIoMm6Q2YIbLo7]3j101O1O001O001O001O001O001O10O01O001O001O001O001O1O0014K6K5K5L4MN1L4L4J6F:G9F:G9K5J6J7J5J8I9F:F9HlU`3"}, "label": "this is a giraffe in front of another"}]}
{"id": 179969, "image": "COCO_train2014_000000179969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black suit wearing a helmet sitting on a motorbike\"."}], "task": "refering", "answer_template": "The \"a man in black suit wearing a helmet sitting on a motorbike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 284, 285, 307, 308], "bbox": [0.362828125, 0.0922950819672131, 0.572734375, 0.9259953161592506], "iscrowd": 0, "area": 17347.30935, "rle": {"size": [427, 640], "counts": "nRQ33R=;E;E;E;E9G7J6I6J7J6I6J7RM\\MmJk2P5]MeJh2Z5^M[Ji2c5]MRJi2m5]MhIj2U6W2O1O2N1N3[OdIXJ]6m3jImLP7S3XI]LQ7a39kKRIT4l6QLRIo3k6VLSIi3k6\\LRIe3l6_LRI`3l6fLPIZ3n6kLPIU3m6PMPIP3n6UMPIj2n6[MoHf2o6_MnH`2P7eMnH[2o6jMnHV2P7oMnHQ2o6V2N3M2N3L3N3L2N3N1N3M01N1M4K4K5013L3N2M4L3N2M4M2M3M4M2bN`HoLc7S3]HkLd7U3]HiLc7X3^HeLc7\\3^HbLd7]3]H`Le7`3\\H^Le7b3[H]Lf7c3[HZLg7e3ZHZLg7f3YHYLh7j2QH\\M7Hj7l2SHXM4Jj7m2VHSM3Oh7n2YHlL25g7n2[HfL1:f7o2QIoLQ7P3oHoLS7o2oHnLT7Q3mHmLT7S3nHjLT7T3c1ZNlE3T:JoE6S:FoE9T:BoE>S:^OoEb0T:ZOmEe0W:VOkEj0Y:POjEo0Z:hNkEX1T;00001O000001O1O1O1O2N2N3N2M3M2N3M3L4M2N1O1N2O001O1N2O1O1O1O1N101O1O1N2OiSb3"}, "label": "a man in black suit wearing a helmet sitting on a motorbike"}]}
{"id": 179969, "image": "COCO_train2014_000000179969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man in a suit on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 284, 285, 307, 308], "bbox": [0.362828125, 0.0922950819672131, 0.572734375, 0.9259953161592506], "iscrowd": 0, "area": 17347.30935, "rle": {"size": [427, 640], "counts": "nRQ33R=;E;E;E;E9G7J6I6J7J6I6J7RM\\MmJk2P5]MeJh2Z5^M[Ji2c5]MRJi2m5]MhIj2U6W2O1O2N1N3[OdIXJ]6m3jImLP7S3XI]LQ7a39kKRIT4l6QLRIo3k6VLSIi3k6\\LRIe3l6_LRI`3l6fLPIZ3n6kLPIU3m6PMPIP3n6UMPIj2n6[MoHf2o6_MnH`2P7eMnH[2o6jMnHV2P7oMnHQ2o6V2N3M2N3L3N3L2N3N1N3M01N1M4K4K5013L3N2M4L3N2M4M2M3M4M2bN`HoLc7S3]HkLd7U3]HiLc7X3^HeLc7\\3^HbLd7]3]H`Le7`3\\H^Le7b3[H]Lf7c3[HZLg7e3ZHZLg7f3YHYLh7j2QH\\M7Hj7l2SHXM4Jj7m2VHSM3Oh7n2YHlL25g7n2[HfL1:f7o2QIoLQ7P3oHoLS7o2oHnLT7Q3mHmLT7S3nHjLT7T3c1ZNlE3T:JoE6S:FoE9T:BoE>S:^OoEb0T:ZOmEe0W:VOkEj0Y:POjEo0Z:hNkEX1T;00001O000001O1O1O1O2N2N3N2M3M2N3M3L4M2N1O1N2O001O1N2O1O1O1O1N101O1O1N2OiSb3"}, "label": "a man in a suit on a motorcycle"}]}
{"id": 179969, "image": "COCO_train2014_000000179969.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a city bus painted red and white\"."}], "task": "refering", "answer_template": "The \"a city bus painted red and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 136, 137], "bbox": [0.69415625, 0.16540983606557377, 1.0, 0.36110070257611243], "iscrowd": 0, "area": 11880.0163, "rle": {"size": [427, 640], "counts": "UXi52V=6I8I4K4N2M3M3N3L3M3N2M3N3M2O10000O2O0O1000002N1O2N1O000000000001O0000000O100O100O100O2O000000000001O00000000001O00000000010O000000001O00000000001O00000000001O01O000001O00000000001O00000000001O0001O0001O00000000001O00000000001O000000010O00000000001O00000000001O000001O01O00000000001O000000010O2N1O001O001O1O001O0001O01O000000O100O1N2M30001O01O0000001O00001O00000010O000001O0000001O00001O01O]M"}, "label": "a city bus painted red and white"}]}
{"id": 483078, "image": "COCO_train2014_000000483078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the child in the outer most right side of the photo , he is standing upright and facing towards the left side of the photo\"."}], "task": "refering", "answer_template": "The \"the child in the outer most right side of the photo , he is standing upright and facing towards the left side of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 39, 40, 41, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 316, 317, 318, 338, 339, 340, 341], "bbox": [0.69296875, 0.004674418604651162, 0.851140625, 0.9934651162790697], "iscrowd": 0, "area": 29097.706099999992, "rle": {"size": [430, 640], "counts": "icj53X=3M3M3M4gD8g8KWG7f8LXG6`4UOoL0QNi0>3`4WOoLOoMj0`02_4XOPM0lMi0c00`4ZOnL0kMi0e0O_4[OPM0gMj0g0M`4\\OoL0fMj0j0K`4\\OoL2bMj0m0Ja4\\OmL3aMk0n0Ge4\\OjLg1_NoNf4[OiLh1_NoNh4ZOgLj1^NmNj4ZOfLm1\\NkNl3TOmK5i1P2YNiNP4SOmK5h1S2WNfNU4ROlK6f1U2UNeNY4POkK7e1Y2RNbN]4POkK6d1[2QN`N`4oNkKd0T1g3Q3eKjKT1e0]3[3`KPLU1a0c3Y3XKULV1a0h3U3SKYLT1a0Q4o2lJ_LT1a0W4j2eJeLT1?_4e2^JlLS1=g4`2XJRMQ1=m4\\2RJWMQ1;U5W2kI^MP1:\\5Q2eIeMo08d5l1_IkMl09l5e1YIRNk07T6`1RIYNj06Z6\\1lH^Nj04b6W1fHdNh04i6Q1`HkNj0On6o0YHROP:g0QFYOT:a0mE_OY:;gEE_:4aEMe:L\\E4o;000O1000000O1000000O1000000O1000000O1000000O1000000O1aDEk9;SFHl98RFLl94RFOl92RF2l9NRF5m9KQF8n9HPF<m9FPF=o9EmE?Q:BlEa0S:AiEb0U:@hEd0V:^OfEe0Y:\\OdEg0[:[OaEi0U9dNaGe0WOj0b8_OnGI\\Ol0o78]Hj0]3fNRN]1YNN`3l6J5K6J5K5KQ1oN\\1dNjgW1"}, "label": "the child in the outer most right side of the photo , he is standing upright and facing towards the left side of the photo"}]}
{"id": 483078, "image": "COCO_train2014_000000483078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy standing straight in the playground\"."}], "task": "refering", "answer_template": "The \"a small boy standing straight in the playground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 39, 40, 41, 63, 64, 65, 85, 86, 87, 88, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 247, 248, 249, 270, 271, 272, 293, 294, 295, 316, 317, 318, 338, 339, 340, 341], "bbox": [0.69296875, 0.004674418604651162, 0.851140625, 0.9934651162790697], "iscrowd": 0, "area": 29097.706099999992, "rle": {"size": [430, 640], "counts": "icj53X=3M3M3M4gD8g8KWG7f8LXG6`4UOoL0QNi0>3`4WOoLOoMj0`02_4XOPM0lMi0c00`4ZOnL0kMi0e0O_4[OPM0gMj0g0M`4\\OoL0fMj0j0K`4\\OoL2bMj0m0Ja4\\OmL3aMk0n0Ge4\\OjLg1_NoNf4[OiLh1_NoNh4ZOgLj1^NmNj4ZOfLm1\\NkNl3TOmK5i1P2YNiNP4SOmK5h1S2WNfNU4ROlK6f1U2UNeNY4POkK7e1Y2RNbN]4POkK6d1[2QN`N`4oNkKd0T1g3Q3eKjKT1e0]3[3`KPLU1a0c3Y3XKULV1a0h3U3SKYLT1a0Q4o2lJ_LT1a0W4j2eJeLT1?_4e2^JlLS1=g4`2XJRMQ1=m4\\2RJWMQ1;U5W2kI^MP1:\\5Q2eIeMo08d5l1_IkMl09l5e1YIRNk07T6`1RIYNj06Z6\\1lH^Nj04b6W1fHdNh04i6Q1`HkNj0On6o0YHROP:g0QFYOT:a0mE_OY:;gEE_:4aEMe:L\\E4o;000O1000000O1000000O1000000O1000000O1000000O1000000O1aDEk9;SFHl98RFLl94RFOl92RF2l9NRF5m9KQF8n9HPF<m9FPF=o9EmE?Q:BlEa0S:AiEb0U:@hEd0V:^OfEe0Y:\\OdEg0[:[OaEi0U9dNaGe0WOj0b8_OnGI\\Ol0o78]Hj0]3fNRN]1YNN`3l6J5K6J5K5KQ1oN\\1dNjgW1"}, "label": "a small boy standing straight in the playground"}]}
{"id": 483078, "image": "COCO_train2014_000000483078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boyw tiha black glove is bent over while in the baseball field\"."}], "task": "refering", "answer_template": "The \"a boyw tiha black glove is bent over while in the baseball field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 215, 216, 217, 218, 219, 238, 239, 241, 242, 261, 262, 264, 265, 285, 287, 288], "bbox": [0.388171875, 0.3264418604651163, 0.5506562500000001, 0.8400930232558139], "iscrowd": 0, "area": 12529.002550000007, "rle": {"size": [430, 640], "counts": "b_X34l9^3N2O1N2N2O2M2QOULYHm3f7VLWHk3g7YLUHi3j7ZLRHh3m7[LoGg3o7]LnGe3P8^LlGd3S8_LiGc3V8k0N2O1O1O^OTHeKi7X4`HdK_7X4iHeKV7X4RIdKm6X4U1M2N1O2N2N2N2N2N102M3M32N2N2N1O2O11O00000O11O2O0O1O2N1O1O1O2O0O1O1O2N1O4M2M4L3M4L3M3M4M2M4LcKUHZ3o83Nn0QO00O2N1O2bLmG]1T8[NUHc1l7VN[Hi1g7oM`HQ2`7gMhHX2Y7aMoH]2S7ZMUIe2l6TM[Ik2f8N1O100O2N1O1Ol1TN2N2M3N2N2N2M3N2N2M3N2Nlbh3"}, "label": "a boyw tiha black glove is bent over while in the baseball field"}]}
{"id": 483078, "image": "COCO_train2014_000000483078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with the black glove\"."}], "task": "refering", "answer_template": "The \"boy with the black glove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 125, 126, 147, 148, 149, 170, 171, 172, 193, 194, 195, 215, 216, 217, 218, 219, 238, 239, 241, 242, 261, 262, 264, 265, 285, 287, 288], "bbox": [0.388171875, 0.3264418604651163, 0.5506562500000001, 0.8400930232558139], "iscrowd": 0, "area": 12529.002550000007, "rle": {"size": [430, 640], "counts": "b_X34l9^3N2O1N2N2O2M2QOULYHm3f7VLWHk3g7YLUHi3j7ZLRHh3m7[LoGg3o7]LnGe3P8^LlGd3S8_LiGc3V8k0N2O1O1O^OTHeKi7X4`HdK_7X4iHeKV7X4RIdKm6X4U1M2N1O2N2N2N2N2N102M3M32N2N2N1O2O11O00000O11O2O0O1O2N1O1O1O2O0O1O1O2N1O4M2M4L3M4L3M3M4M2M4LcKUHZ3o83Nn0QO00O2N1O2bLmG]1T8[NUHc1l7VN[Hi1g7oM`HQ2`7gMhHX2Y7aMoH]2S7ZMUIe2l6TM[Ik2f8N1O100O2N1O1Ol1TN2N2M3N2N2N2M3N2N2M3N2Nlbh3"}, "label": "boy with the black glove"}]}
{"id": 7945, "image": "COCO_train2014_000000007945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lemon on right front\"."}], "task": "refering", "answer_template": "The \"lemon on right front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 106, 107, 108, 109], "bbox": [0.569796875, 0.0, 0.7869375, 0.26153846153846155], "iscrowd": 0, "area": 14535.403449999998, "rle": {"size": [520, 640], "counts": "Y]i5a0V?e0I6K5K5K3M2N3M2N3M2N2O2M2N3M3M2O2N1N3N2N1N3N1O2M3N1O2N1O2O000O2O001N101O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O1O1O100O001O000000O10000000001O0O10000000001O1N2O1O001O1O1O1O1N2O1O0O2N2N2N2N2N2O1N2N2N2N2N2N2M3M3M3L4M3M3M3M6I7J6J6G:E:F:F:FTPU2"}, "label": "lemon on right front"}]}
{"id": 7945, "image": "COCO_train2014_000000007945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange in front of a group of three oranges\"."}], "task": "refering", "answer_template": "The \"an orange in front of a group of three oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 106, 107, 108, 109], "bbox": [0.569796875, 0.0, 0.7869375, 0.26153846153846155], "iscrowd": 0, "area": 14535.403449999998, "rle": {"size": [520, 640], "counts": "Y]i5a0V?e0I6K5K5K3M2N3M2N3M2N2O2M2N3M3M2O2N1N3N2N1N3N1O2M3N1O2N1O2O000O2O001N101O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O1O1O100O001O000000O10000000001O0O10000000001O1N2O1O001O1O1O1O1N2O1O0O2N2N2N2N2N2O1N2N2N2N2N2N2M3M3M3L4M3M3M3M6I7J6J6G:E:F:F:FTPU2"}, "label": "an orange in front of a group of three oranges"}]}
{"id": 7945, "image": "COCO_train2014_000000007945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white table\"."}], "task": "refering", "answer_template": "The \"a white table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 205, 206, 227, 228, 229, 251, 252, 274, 275, 297, 298, 321, 344, 367, 390, 413, 436], "bbox": [0.9024218749999999, 0.38157692307692304, 0.99921875, 1.0], "iscrowd": 0, "area": 8588.450450000017, "rle": {"size": [520, 640], "counts": "chU93T`03M3N2M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M4L4L4L4L3M5K4L5K5K5K5K5K5K5K4L5K5hHgKY2]4\\MjKc2[4RMmKl2X4iLoKV3V4_LQL`3S4VLTLi3Q4lKVLS4o3bKXL]4R4SKULl4W4bJPL]5]4PJjKo5b4_IeK`6R7N2O1O1O1O2N1O1N2O1O1OiG^KR3ROVLm:C\\H"}, "label": "a white table"}]}
{"id": 7945, "image": "COCO_train2014_000000007945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white color table\"."}], "task": "refering", "answer_template": "The \"a white color table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 205, 206, 227, 228, 229, 251, 252, 274, 275, 297, 298, 321, 344, 367, 390, 413, 436], "bbox": [0.9024218749999999, 0.38157692307692304, 0.99921875, 1.0], "iscrowd": 0, "area": 8588.450450000017, "rle": {"size": [520, 640], "counts": "chU93T`03M3N2M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M4L4L4L4L3M5K4L5K5K5K5K5K5K5K4L5K5hHgKY2]4\\MjKc2[4RMmKl2X4iLoKV3V4_LQL`3S4VLTLi3Q4lKVLS4o3bKXL]4R4SKULl4W4bJPL]5]4PJjKo5b4_IeK`6R7N2O1O1O1O2N1O1N2O1O1OiG^KR3ROVLm:C\\H"}, "label": "a white color table"}]}
{"id": 7945, "image": "COCO_train2014_000000007945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an expanse of table above the bowl\"."}], "task": "refering", "answer_template": "The \"an expanse of table above the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 46, 47, 48, 49, 50, 51, 52, 53, 54], "bbox": [0.0041875, 0.0, 0.39259375, 0.15946153846153846], "iscrowd": 0, "area": 16642.4868, "rle": {"size": [520, 640], "counts": "h`1e1c>n0RO0000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000001O000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O104K8H8H8H8I6IQ`U6"}, "label": "an expanse of table above the bowl"}]}
{"id": 163598, "image": "COCO_train2014_000000163598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a yellow shirt sitting on a table watching a tv\"."}], "task": "refering", "answer_template": "The \"a guy in a yellow shirt sitting on a table watching a tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 132, 133, 155, 156, 157, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 360, 361, 362, 383, 384, 385], "bbox": [0.6520312500000001, 0.27139583333333334, 0.94425, 0.9853541666666668], "iscrowd": 0, "area": 32226.271700000005, "rle": {"size": [480, 640], "counts": "fjS62k>?B>A`0A>A?C>B5L1N2O1N2N2O1N2N2O1N2N2O0O2O1N2N2O1N2N2O0OC>C<C=D=O01O10O01O100O100O1O100O2N2O2kDaMa9d2VFeMe9a2oEiMl9^2gEmMT:b3L4K5L4K5YLiJ`M\\5\\2iJ_M\\5[2jJ]K_1SOl3[5kJ`KX7i3cIQL_6n3fImK[6S4iIgKY6X4kIdKV6\\4kIaKW6^4kI_KW6a4jI]KW6b4kI[KX6d4R2O100O100O100O101N100OM4E:E<E:eN^JoHm5Q7dJRHg5n7g0O001O001O1O001O001O001O1N101O001O001O1O001O001_IcH]5]7_JhH`5Y7ZJlHf5U7TJQIj5P7QJVIn5k6lI[IS6e6hI`IX6e7O1O001O1O001O1O1O001N2O001O1O001O1O1O001O1O001O1N101O1O1O001O2N2N2N2N1O2N2M3N2N1O2N2N2N2N1O2N2N2N2N2N1N3N2N2N2N1O2N2K7@R1POQ1XOg0YO4L3M2N3L3N3M2N2N3M[a`0"}, "label": "a guy in a yellow shirt sitting on a table watching a tv"}]}
{"id": 163598, "image": "COCO_train2014_000000163598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a yellow t - shirt watching tv\"."}], "task": "refering", "answer_template": "The \"a man in a yellow t - shirt watching tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 132, 133, 155, 156, 157, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 337, 338, 339, 360, 361, 362, 383, 384, 385], "bbox": [0.6520312500000001, 0.27139583333333334, 0.94425, 0.9853541666666668], "iscrowd": 0, "area": 32226.271700000005, "rle": {"size": [480, 640], "counts": "fjS62k>?B>A`0A>A?C>B5L1N2O1N2N2O1N2N2O1N2N2O0O2O1N2N2O1N2N2O0OC>C<C=D=O01O10O01O100O100O1O100O2N2O2kDaMa9d2VFeMe9a2oEiMl9^2gEmMT:b3L4K5L4K5YLiJ`M\\5\\2iJ_M\\5[2jJ]K_1SOl3[5kJ`KX7i3cIQL_6n3fImK[6S4iIgKY6X4kIdKV6\\4kIaKW6^4kI_KW6a4jI]KW6b4kI[KX6d4R2O100O100O100O101N100OM4E:E<E:eN^JoHm5Q7dJRHg5n7g0O001O001O1O001O001O001O1N101O001O001O1O001O001_IcH]5]7_JhH`5Y7ZJlHf5U7TJQIj5P7QJVIn5k6lI[IS6e6hI`IX6e7O1O001O1O001O1O1O001N2O001O1O001O1O1O001O1O001O1N101O1O1O001O2N2N2N2N1O2N2M3N2N1O2N2N2N2N1O2N2N2N2N2N1N3N2N2N2N1O2N2K7@R1POQ1XOg0YO4L3M2N3L3N3M2N2N3M[a`0"}, "label": "a man in a yellow t - shirt watching tv"}]}
{"id": 163598, "image": "COCO_train2014_000000163598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chubby kid in a grey shirt sitting on the couch watching tv with his dad\"."}], "task": "refering", "answer_template": "The \"a chubby kid in a grey shirt sitting on the couch watching tv with his dad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 195, 196, 197, 218, 219, 220, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 355, 356, 357, 358, 359, 381, 382], "bbox": [0.40695312499999997, 0.4478125, 0.6524687499999999, 0.95875], "iscrowd": 0, "area": 26189.9334, "rle": {"size": [480, 640], "counts": "TYj31d><D;F:E;E;E;J7K4M3L5K4L5K4M3L5K4N2O2M2O2N1N2O2M2O1N3O0O1O2N1O2N1O1O2N1O1O2N100O2N1O2N1O101N100O2O0O2O0lNXKoGi4m7cKkG^4Q8nKeGS4W8YLaGg3\\8a1K4M3N3O0O2O0O1O2O0O1O2O0O2O00001O00001O00001O00001O0000000000001O000O100000001O0000000000001O01O1O1cJ\\H\\3e7cL\\H\\3e7cL[H]3f7aL\\H^3d7bL]H^3c7aL]H_3d7_L^H`3c7_L^H`3c7^L_Ha3b7^L^Hb3c7]L^Hb3c7[L`He3a7XL`Hh3e7QL^Hn3f7mK\\HR4i7gKZHX4k7bKVH^4X8RKkGm4U90O01O1O1O001O10O01O1O1O001O10O02N5K4L5K5K4M4K4L5fNgDcN^;W1gDeN];U1iDgN`;o0dDnNP<=UD_O^<MgC0h=L3M4LhRX3"}, "label": "a chubby kid in a grey shirt sitting on the couch watching tv with his dad"}]}
{"id": 163598, "image": "COCO_train2014_000000163598.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a gray shirt watching tv with another person\"."}], "task": "refering", "answer_template": "The \"a person wearing a gray shirt watching tv with another person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 195, 196, 197, 218, 219, 220, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 355, 356, 357, 358, 359, 381, 382], "bbox": [0.40695312499999997, 0.4478125, 0.6524687499999999, 0.95875], "iscrowd": 0, "area": 26189.9334, "rle": {"size": [480, 640], "counts": "TYj31d><D;F:E;E;E;J7K4M3L5K4L5K4M3L5K4N2O2M2O2N1N2O2M2O1N3O0O1O2N1O2N1O1O2N1O1O2N100O2N1O2N1O101N100O2O0O2O0lNXKoGi4m7cKkG^4Q8nKeGS4W8YLaGg3\\8a1K4M3N3O0O2O0O1O2O0O1O2O0O2O00001O00001O00001O00001O0000000000001O000O100000001O0000000000001O01O1O1cJ\\H\\3e7cL\\H\\3e7cL[H]3f7aL\\H^3d7bL]H^3c7aL]H_3d7_L^H`3c7_L^H`3c7^L_Ha3b7^L^Hb3c7]L^Hb3c7[L`He3a7XL`Hh3e7QL^Hn3f7mK\\HR4i7gKZHX4k7bKVH^4X8RKkGm4U90O01O1O1O001O10O01O1O1O001O10O02N5K4L5K5K4M4K4L5fNgDcN^;W1gDeN];U1iDgN`;o0dDnNP<=UD_O^<MgC0h=L3M4LhRX3"}, "label": "a person wearing a gray shirt watching tv with another person"}]}
{"id": 401173, "image": "COCO_train2014_000000401173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra facing a forest\"."}], "task": "refering", "answer_template": "The \"a zebra facing a forest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 93, 94, 95, 111, 112, 113, 128, 129, 130, 131, 146, 147, 148, 149, 165, 166, 183, 184, 201, 202, 203], "bbox": [0.15456, 0.3905105105105105, 0.33198, 1.0], "iscrowd": 0, "area": 9959.885350000002, "rle": {"size": [333, 500], "counts": "bWi07T:f0ZOa0@2M4M2N2N3M2O2M2N2N3N1N3M3M2O2M2N3M3K4L5L4H7@a0D`0Ia0@`0_O`0A0O2N100O2O0O10O001O010O00001O010O100000000O10O10000000O100PNPLcMP4Y2gLTMY3h2oLTMQ3h2VMQMn2Z1kKUO_1YOl2c1UL^NV1Gj2l1`31O010O15K7\\N[GQ1R9O1O1N2O001O1N2N2N1O2N2N2N2N2N1L9Dk]\\3"}, "label": "a zebra facing a forest"}]}
{"id": 401173, "image": "COCO_train2014_000000401173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small zebra behind the big zebra\"."}], "task": "refering", "answer_template": "The \"a small zebra behind the big zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 93, 94, 95, 111, 112, 113, 128, 129, 130, 131, 146, 147, 148, 149, 165, 166, 183, 184, 201, 202, 203], "bbox": [0.15456, 0.3905105105105105, 0.33198, 1.0], "iscrowd": 0, "area": 9959.885350000002, "rle": {"size": [333, 500], "counts": "bWi07T:f0ZOa0@2M4M2N2N3M2O2M2N2N3N1N3M3M2O2M2N3M3K4L5L4H7@a0D`0Ia0@`0_O`0A0O2N100O2O0O10O001O010O00001O010O100000000O10O10000000O100PNPLcMP4Y2gLTMY3h2oLTMQ3h2VMQMn2Z1kKUO_1YOl2c1UL^NV1Gj2l1`31O010O15K7\\N[GQ1R9O1O1N2O001O1N2N2N1O2N2N2N2N2N1L9Dk]\\3"}, "label": "a small zebra behind the big zebra"}]}
{"id": 401173, "image": "COCO_train2014_000000401173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult zebra standing near a baby zebra\"."}], "task": "refering", "answer_template": "The \"an adult zebra standing near a baby zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 119, 120, 121, 132, 133, 134, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 187, 189, 190, 191], "bbox": [0.33076, 0.30492492492492496, 0.77226, 0.9071771771771773], "iscrowd": 0, "area": 24686.632399999995, "rle": {"size": [333, 500], "counts": "cje1d0h95J7J5J5L4K5K5L4K5L3L5L3M3N2M3M3N2M3M2O2M3M3N2M3M5L5J5L4K6J5L5J5K2O001O001O001O001O001O001O001OO1N2M2O2M3N2N2M3N2M3N2M3N3M2M4M2M3N3M2O2O010O0010O010O01O001O0O2O0000001N100000O10O0100lIjLP5W3lJnLS5R3jJQMV5n2hJUMW5l2eJXM[5h2bJ[M]5f2`J^M_5b2]JbMb5_2[JdMe5[2YJhMg5X2UJlMj5]3O0O100O10N1N3M3M2O2M2N3M3M2O2M2N3M3M2O2O0O2N1O2O1N1O2N12O1O2M3N1N3N1N[1eN01O01O01O00010O0010O00M4E:G9H9N1O1O2N1N3N1O2N1O2M2O2N2N1O2O001N2O000O1000000O10O1000O10000O2O1O1O1N2O1003M2N2O1N2N000010O00010O0010O00010N10L4L4K4M4K5L4L3M4L4hNX1I6J7IZoT1"}, "label": "an adult zebra standing near a baby zebra"}]}
{"id": 401173, "image": "COCO_train2014_000000401173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large zebra in front of a baby zebra\"."}], "task": "refering", "answer_template": "The \"a large zebra in front of a baby zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 119, 120, 121, 132, 133, 134, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 187, 189, 190, 191], "bbox": [0.33076, 0.30492492492492496, 0.77226, 0.9071771771771773], "iscrowd": 0, "area": 24686.632399999995, "rle": {"size": [333, 500], "counts": "cje1d0h95J7J5J5L4K5K5L4K5L3L5L3M3N2M3M3N2M3M2O2M3M3N2M3M5L5J5L4K6J5L5J5K2O001O001O001O001O001O001O001OO1N2M2O2M3N2N2M3N2M3N2M3N3M2M4M2M3N3M2O2O010O0010O010O01O001O0O2O0000001N100000O10O0100lIjLP5W3lJnLS5R3jJQMV5n2hJUMW5l2eJXM[5h2bJ[M]5f2`J^M_5b2]JbMb5_2[JdMe5[2YJhMg5X2UJlMj5]3O0O100O10N1N3M3M2O2M2N3M3M2O2M2N3M3M2O2O0O2N1O2O1N1O2N12O1O2M3N1N3N1N[1eN01O01O01O00010O0010O00M4E:G9H9N1O1O2N1N3N1O2N1O2M2O2N2N1O2O001N2O000O1000000O10O1000O10000O2O1O1O1N2O1003M2N2O1N2N000010O00010O0010O00010N10L4L4K4M4K5L4L3M4L4hNX1I6J7IZoT1"}, "label": "a large zebra in front of a baby zebra"}]}
{"id": 384790, "image": "COCO_train2014_000000384790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car bottom left corner\"."}], "task": "refering", "answer_template": "The \"black car bottom left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.001546875, 0.6224772727272727, 0.37696875, 0.9887727272727271], "iscrowd": 0, "area": 32684.2677, "rle": {"size": [440, 640], "counts": "Sj0X1o9a2hNX1001O0000001O0000000000000000001O0000000000000000000000001O0000000000000000000000001O00000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O00000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0lKnFb3R9XLUGg3k8SL]Gk3Y9O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N3M4L3M3M4L3M3J7BeQ[5"}, "label": "black car bottom left corner"}]}
{"id": 384790, "image": "COCO_train2014_000000384790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car in front of a red bus\"."}], "task": "refering", "answer_template": "The \"black car in front of a red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353], "bbox": [0.001546875, 0.6224772727272727, 0.37696875, 0.9887727272727271], "iscrowd": 0, "area": 32684.2677, "rle": {"size": [440, 640], "counts": "Sj0X1o9a2hNX1001O0000001O0000000000000000001O0000000000000000000000001O0000000000000000000000001O00000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O00000000000000000000001O0000000000000000000000001O0000000000000000000000001O0000000000000000000000001O0lKnFb3R9XLUGg3k8SL]Gk3Y9O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N3M4L3M3M4L3M3J7BeQ[5"}, "label": "black car in front of a red bus"}]}
{"id": 368409, "image": "COCO_train2014_000000368409.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus on the road with four square on it\"."}], "task": "refering", "answer_template": "The \"bus on the road with four square on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 248, 249, 250, 251, 252], "bbox": [0.606984375, 0.44279166666666664, 1.0, 0.6447708333333334], "iscrowd": 0, "area": 17232.17155, "rle": {"size": [480, 640], "counts": "hTf51c>=C>C<C>C<I8M2O2M2O0O100000O1000000000000O100O1N2M3M3O1000001O0000000000000000000O10000000000000001O0000000000000O10000000000000000000001O0000000O100001O00000000000000001O0001O000000000001O0000000000000000001O00000000000000001O100O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N00N2O1N2O1N2N2O1N201O00000000000000000^O_CeNa<Y1dCdN\\<Z1iCcNW<[1mCcNS<\\1PDbNP<]1j000O1O100O100O1O100O11O00100O001O1O1O001O1O2N2N2N2N2N3M2N001O1O001O1O001O4L4L1O00010O000000000000000000000000000000000000000000000000000001O00<D6Km0ROo0QOdZ1]2obN00000000000000000000000000000oH"}, "label": "bus on the road with four square on it"}]}
{"id": 368409, "image": "COCO_train2014_000000368409.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus with gold brown and red trapezoids on side\"."}], "task": "refering", "answer_template": "The \"bus with gold brown and red trapezoids on side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 248, 249, 250, 251, 252], "bbox": [0.606984375, 0.44279166666666664, 1.0, 0.6447708333333334], "iscrowd": 0, "area": 17232.17155, "rle": {"size": [480, 640], "counts": "hTf51c>=C>C<C>C<I8M2O2M2O0O100000O1000000000000O100O1N2M3M3O1000001O0000000000000000000O10000000000000001O0000000000000O10000000000000000000001O0000000O100001O00000000000000001O0001O000000000001O0000000000000000001O00000000000000001O100O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N00N2O1N2O1N2N2O1N201O00000000000000000^O_CeNa<Y1dCdN\\<Z1iCcNW<[1mCcNS<\\1PDbNP<]1j000O1O100O100O1O100O11O00100O001O1O1O001O1O2N2N2N2N2N3M2N001O1O001O1O001O4L4L1O00010O000000000000000000000000000000000000000000000000000001O00<D6Km0ROo0QOdZ1]2obN00000000000000000000000000000oH"}, "label": "bus with gold brown and red trapezoids on side"}]}
{"id": 237340, "image": "COCO_train2014_000000237340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table on the right\"."}], "task": "refering", "answer_template": "The \"a table on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 321], "bbox": [0.777203125, 0.5103541666666667, 1.0, 0.810875], "iscrowd": 0, "area": 11004.207649999998, "rle": {"size": [480, 640], "counts": "UWY71o>5K5K3L2O1O1O1O0O2O1O1O1O1O1N101O1O1O1N2O1O001O1N2O1O1O1O0O2O1O1O1O1O1N101O1O1O1N2O1O001O1N2O1O1O1O001N2O1O1O1O0O2O1O1O1O1N2O001O1O1O1N2O1O001O1N2O1O1O1O0O2O1O1O1O1N2O001O1O100O100O00100O100O1O10O0100O1O100O10O01O100O100O1O010O100O1O100O010O1O100O010O0010O0100O0010O0100O0010O01Y1gNT1kNm0TOYD"}, "label": "a table on the right"}]}
{"id": 237340, "image": "COCO_train2014_000000237340.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of the table in front of the woman in the white shirt\"."}], "task": "refering", "answer_template": "The \"the top of the table in front of the woman in the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 321], "bbox": [0.777203125, 0.5103541666666667, 1.0, 0.810875], "iscrowd": 0, "area": 11004.207649999998, "rle": {"size": [480, 640], "counts": "UWY71o>5K5K3L2O1O1O1O0O2O1O1O1O1O1N101O1O1O1N2O1O001O1N2O1O1O1O0O2O1O1O1O1O1N101O1O1O1N2O1O001O1N2O1O1O1O001N2O1O1O1O0O2O1O1O1O1N2O001O1O1O1N2O1O001O1N2O1O1O1O0O2O1O1O1O1N2O001O1O100O100O00100O100O1O10O0100O1O100O10O01O100O100O1O010O100O1O100O010O1O100O010O0010O0100O0010O0100O0010O01Y1gNT1kNm0TOYD"}, "label": "the top of the table in front of the woman in the white shirt"}]}
{"id": 425758, "image": "COCO_train2014_000000425758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden chairs at the table\"."}], "task": "refering", "answer_template": "The \"the wooden chairs at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 256, 257, 258, 279, 280, 281, 300, 301, 302, 303, 304, 324, 325, 326, 327, 348, 349, 350, 371, 372], "bbox": [0.07096875, 0.6086041666666666, 0.25321875, 0.9978541666666667], "iscrowd": 0, "area": 9636.205600000003, "rle": {"size": [480, 640], "counts": "[ne02l>0O4O0O1O100O1O100O101N1O100O100O1O100O1O2O0O100O100O100O101N100O100O100O101O1N101\\OPOUCR1i<POVCQ1h<POWCQ1R;0PEQOMP1R;f0mD\\OS;c0mD]OT;c0jD;Y:EgE?U:BiEc0S:]OmEg0o9YOPFm0k9TOTFQ1g9oNXFV1d9kN[FZ1`9fN_F`1\\9aNcFd1X9\\NhFg1U9YNjFh1V9YNiFf1X9Q200000O10000bNkFQMU9n2QGlLP9T3VGfLj8Y3\\GbLd8^3aG]L_8b3hGWLX8j3mGQLS8n3SHmKm7S4`1O100O100O100O100O100O1O100O100O100N101O13M3M2N3M32N2N2M^1cNQ1oN9F3L4M3M3L3L4K5K4L5KVdo6"}, "label": "the wooden chairs at the table"}]}
{"id": 425758, "image": "COCO_train2014_000000425758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small black table\"."}], "task": "refering", "answer_template": "The \"the small black table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 270, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 319, 336, 339, 340, 341, 342, 359], "bbox": [0.5725625, 0.678, 0.8821875, 0.9111666666666667], "iscrowd": 0, "area": 11055.4531, "rle": {"size": [480, 640], "counts": "e\\\\53k>2O2M3M3N1N3M2O1N2N3N1N2N2O1N3M2O1N2N2O2M2N2O1N3M2O1N2N2O20O00000001O00c1]N000001O0O1ZOf0WOi0N200000010O00000000000001O01O000000000001O000001O0000000001O000001O00000001O0000000001O0001O000000000001O0001O000000000001O01O0000000000000010O00000000000001O01O00000000000001O0001O0000000001O000001O00000001O00000001O000001O000000000001O0001O000000000001O01O000000000000010O0001O4L5K4L5K4L5L3L4L5K4L5K4L5K4MZhR1"}, "label": "the small black table"}]}
{"id": 425758, "image": "COCO_train2014_000000425758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black topped table with a white rag on top of it\"."}], "task": "refering", "answer_template": "The \"a black topped table with a white rag on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 270, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 319, 336, 339, 340, 341, 342, 359], "bbox": [0.5725625, 0.678, 0.8821875, 0.9111666666666667], "iscrowd": 0, "area": 11055.4531, "rle": {"size": [480, 640], "counts": "e\\\\53k>2O2M3M3N1N3M2O1N2N3N1N2N2O1N3M2O1N2N2O2M2N2O1N3M2O1N2N2O20O00000001O00c1]N000001O0O1ZOf0WOi0N200000010O00000000000001O01O000000000001O000001O0000000001O000001O00000001O0000000001O0001O000000000001O0001O000000000001O01O0000000000000010O00000000000001O01O00000000000001O0001O0000000001O000001O00000001O00000001O000001O000000000001O0001O000000000001O01O000000000000010O0001O4L5K4L5K4L5L3L4L5K4L5K4L5K4MZhR1"}, "label": "a black topped table with a white rag on top of it"}]}
{"id": 270111, "image": "COCO_train2014_000000270111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden wired chair on the right side\"."}], "task": "refering", "answer_template": "The \"a wooden wired chair on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [296, 297, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.7919375, 0.7308333333333333, 0.9983906250000001, 0.9867291666666667], "iscrowd": 0, "area": 13258.115150000009, "rle": {"size": [480, 640], "counts": "dS^75e>6I7J6I7J6I7J6I7J6J6I7J6I7L4O1O1O100O1O1O10000000000001OO100000000O1000000O100O1O100O1O100O1O1O100O1O1000000000000000000000000O1N2N2O1N2N2O1O10000001O00000000001O00000000001O001O1O1O001O1O1O001O1O001O001O001O00001O00000000001O000000002O0O1O2N1O2N1O2N1O2N2N2N1O2dNU4"}, "label": "a wooden wired chair on the right side"}]}
{"id": 270111, "image": "COCO_train2014_000000270111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair back , bottom right of picture\"."}], "task": "refering", "answer_template": "The \"chair back , bottom right of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [296, 297, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.7919375, 0.7308333333333333, 0.9983906250000001, 0.9867291666666667], "iscrowd": 0, "area": 13258.115150000009, "rle": {"size": [480, 640], "counts": "dS^75e>6I7J6I7J6I7J6I7J6J6I7J6I7L4O1O1O100O1O1O10000000000001OO100000000O1000000O100O1O100O1O100O1O1O100O1O1000000000000000000000000O1N2N2O1N2N2O1O10000001O00000000001O00000000001O001O1O1O001O1O1O001O1O001O001O001O00001O00000000001O000000002O0O1O2N1O2N1O2N1O2N2N2N1O2dNU4"}, "label": "chair back , bottom right of picture"}]}
{"id": 40735, "image": "COCO_train2014_000000040735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat on a chair in front of it ' s reflection\"."}], "task": "refering", "answer_template": "The \"a cat on a chair in front of it ' s reflection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 20, 34, 35, 36, 37, 51, 52, 53, 54, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144], "bbox": [0.014270833333333333, 0.1613061224489796, 0.793625, 0.935469387755102], "iscrowd": 0, "area": 37657.727000000006, "rle": {"size": [245, 480], "counts": "[j12a73M3M3M4ZODmI?l5LmI7l54mI0k5Q1M4M2J7L3M4M3K5XO_MSLh2a3l0K6K4L4L4M3L4O1001O001N101N1O100O100O1O100O2N1O1O10O01O00000000000001O00000000000001O01O00000000000000001O0000000000001O00000O100O100O2O0OoKlLc3S3]LnLc3R3[LQMd3n2\\LSMd3l2\\LUMd3j2[LYMe3e2[L\\Me3b2\\L_Md3a2ZLbMe3]2ZLfMe3Y2[LkMb3T2^LSN\\3m1cLUN]3i1cLZN[3e1eL^NY3b1fLaNX3^1iLdNU3[1iLkNT3n2PMbKd2h4O001O00001O0000001O0000000001O000000000000001O0000001N100000000000001O0000000000001O000O1000001O001O001O1O0010O01O0010O01O00010O001O0000001O01O0001O00000010O0001O001O1O1O1O1O00100O0000O1000001O0O1000O10O100000000O10O1000O10O1000O0100O1000O10O1000O10O100000O010000O01000O10O010000O10O1000000O100000000O1000000O110O00001O01O0001O000001O000001O01O00000000000000000O1000O10000000000000O010000O1000O10O1O2M2000001O0O1000001O0000001N100000001O000O101O0O1O100O2N1O100O2O0O1O101N1N3N2N2N1O2O1O1O1N101O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N101N101O<Cc0^O_`g0"}, "label": "a cat on a chair in front of it ' s reflection"}]}
{"id": 40735, "image": "COCO_train2014_000000040735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat closest to you\"."}], "task": "refering", "answer_template": "The \"the cat closest to you\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 20, 34, 35, 36, 37, 51, 52, 53, 54, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144], "bbox": [0.014270833333333333, 0.1613061224489796, 0.793625, 0.935469387755102], "iscrowd": 0, "area": 37657.727000000006, "rle": {"size": [245, 480], "counts": "[j12a73M3M3M4ZODmI?l5LmI7l54mI0k5Q1M4M2J7L3M4M3K5XO_MSLh2a3l0K6K4L4L4M3L4O1001O001N101N1O100O100O1O100O2N1O1O10O01O00000000000001O00000000000001O01O00000000000000001O0000000000001O00000O100O100O2O0OoKlLc3S3]LnLc3R3[LQMd3n2\\LSMd3l2\\LUMd3j2[LYMe3e2[L\\Me3b2\\L_Md3a2ZLbMe3]2ZLfMe3Y2[LkMb3T2^LSN\\3m1cLUN]3i1cLZN[3e1eL^NY3b1fLaNX3^1iLdNU3[1iLkNT3n2PMbKd2h4O001O00001O0000001O0000000001O000000000000001O0000001N100000000000001O0000000000001O000O1000001O001O001O1O0010O01O0010O01O00010O001O0000001O01O0001O00000010O0001O001O1O1O1O1O00100O0000O1000001O0O1000O10O100000000O10O1000O10O1000O0100O1000O10O1000O10O100000O010000O01000O10O010000O10O1000000O100000000O1000000O110O00001O01O0001O000001O000001O01O00000000000000000O1000O10000000000000O010000O1000O10O1O2M2000001O0O1000001O0000001N100000001O000O101O0O1O100O2N1O100O2O0O1O101N1N3N2N2N1O2O1O1O1N101O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N101N101O<Cc0^O_`g0"}, "label": "the cat closest to you"}]}
{"id": 40735, "image": "COCO_train2014_000000040735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a greyish colored cat laying on his right side\"."}], "task": "refering", "answer_template": "The \"a greyish colored cat laying on his right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 22, 23, 24, 25, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 72, 73, 74, 75, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 115, 116], "bbox": [0.251375, 0.017224489795918365, 0.9922916666666666, 0.7010612244897959], "iscrowd": 0, "area": 22983.90215, "rle": {"size": [245, 480], "counts": "nQm01c7100O2O000O100O100O100O100O10000O100O1O1N2O1O1N2O1O1O1N200O100\\MYOZNg0f1BPN?P2DmM<S2HiM8W2KeM6Z2ObM1_21]LZO5f0_33SLA8=f35hKG>5k3P1QLROP4o0mKROT4o0hKSOY4T2100O001O001O00001O00000000000000001O000000000000000000001O000000000000001O0000000O100O1O100O10O1O1O10O01O1O1O1000O1000000000000OO2O1O1O1O1O00100O10fMdK[1[4dNlKW1U4fNSLS1o3jNVLS1k3iNZLU1g3gN_LV1T5L5K5K000O0100O2O0O100000000000000000000000001O01O0000001O01O01O00001O01O000O0100000O1O1OlNgIn0X6QOjIo0V6oNlIQ1S6nNoIR1P6mNQJT1X60000001O0O10000O100O1O100O100O100000000000000000000000001O000000001O001O00001O00001O00001O001O00001N10001O000O2O001O000O2O001O001N2O1O1O1O1O1O1O001O1O00100O1O001O1O1O1O00100oJoM\\4R2_KSN`4c2N1O1O2N1O1O01O0O100O100O100O1000000O1000003L5L1O0000O0100O10000O10000O100O1000O010000O100O10000O1000O0100O010O10O01O10O02O2N1N2O2M2O1N1O2N1O2N1O2N1O2M2O2N101N2O1N2N2O1N2O1N3M3M3L5L4L3M5J[k0"}, "label": "a greyish colored cat laying on his right side"}]}
{"id": 40735, "image": "COCO_train2014_000000040735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection of a tabby cat in a mirror\"."}], "task": "refering", "answer_template": "The \"a reflection of a tabby cat in a mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 22, 23, 24, 25, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 72, 73, 74, 75, 79, 80, 81, 82, 83, 84, 98, 99, 100, 101, 115, 116], "bbox": [0.251375, 0.017224489795918365, 0.9922916666666666, 0.7010612244897959], "iscrowd": 0, "area": 22983.90215, "rle": {"size": [245, 480], "counts": "nQm01c7100O2O000O100O100O100O100O10000O100O1O1N2O1O1N2O1O1O1N200O100\\MYOZNg0f1BPN?P2DmM<S2HiM8W2KeM6Z2ObM1_21]LZO5f0_33SLA8=f35hKG>5k3P1QLROP4o0mKROT4o0hKSOY4T2100O001O001O00001O00000000000000001O000000000000000000001O000000000000001O0000000O100O1O100O10O1O1O10O01O1O1O1000O1000000000000OO2O1O1O1O1O00100O10fMdK[1[4dNlKW1U4fNSLS1o3jNVLS1k3iNZLU1g3gN_LV1T5L5K5K000O0100O2O0O100000000000000000000000001O01O0000001O01O01O00001O01O000O0100000O1O1OlNgIn0X6QOjIo0V6oNlIQ1S6nNoIR1P6mNQJT1X60000001O0O10000O100O1O100O100O100000000000000000000000001O000000001O001O00001O00001O00001O001O00001N10001O000O2O001O000O2O001O001N2O1O1O1O1O1O1O001O1O00100O1O001O1O1O1O00100oJoM\\4R2_KSN`4c2N1O1O2N1O1O01O0O100O100O100O1000000O1000003L5L1O0000O0100O10000O10000O100O1000O010000O100O10000O1000O0100O010O10O01O10O02O2N1N2O2M2O1N1O2N1O2N1O2N1O2M2O2N101N2O1N2N2O1N2O1N3M3M3L5L4L3M5J[k0"}, "label": "a reflection of a tabby cat in a mirror"}]}
{"id": 327462, "image": "COCO_train2014_000000327462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"snowboard of fallen person\"."}], "task": "refering", "answer_template": "The \"snowboard of fallen person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [250, 264, 265, 266, 267, 268, 273, 284, 285, 286, 287, 288, 289, 290, 291, 292, 296, 307, 308, 309, 310, 311, 334], "bbox": [0.38292187499999997, 0.60224609375, 0.9114531249999999, 0.7932617187500001], "iscrowd": 0, "area": 8267.593099999996, "rle": {"size": [512, 640], "counts": "Zkj35i?;F:F:F2N1N2O1O2N1O1N2O2N1O1O001000O10O10O10000O10000O10000O100O01000O10000O100001O1O1O1O1O1N3N1O1OO100O100O1O100O1O10O0100O1O10000001O00001O001N10001O1O1O001O1O1O1O1N101O1O1O1O1O1O000O1000O10000bNXBd0h=[O[Bc0e=\\O_B`0b=^ObB`0^=_OfB>Y=BjB<V=CnB:R=DRC:n<EUC8l<GWC7i<HZC6f<H^C6b<IaC5_<KdC1\\<0fCNZ<1iCMW<3kCKU<5mCIS<7k1O1000000O0100000O1000000O11N101O1O1O001O1O001O1N2O002N2N1O0000000O1000000000000000000O10000000000000000O0100000000000000000O01000000000000001N1000001O0000000010O2N2N1O2O1N2N2N2N2O1N3M4L3N`na10_Q^N8H6J3M2N2N2N2M201O001O001O001O2N2N2N2N2OK6K4K6I7I7IdUl0"}, "label": "snowboard of fallen person"}]}
{"id": 327462, "image": "COCO_train2014_000000327462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person sitting in the snow\"."}], "task": "refering", "answer_template": "The \"the person sitting in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 200, 201, 202, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 265, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342, 361, 362, 363, 364], "bbox": [0.532140625, 0.439765625, 0.913265625, 0.86673828125], "iscrowd": 0, "area": 27794.068999999996, "rle": {"size": [512, 640], "counts": "kjZ5=^?6O1N3M2O1N3N1N2O2M2O2O0O2O0O2O0O2O0O2O0O3N;E1OO100O100O1O10O0100O100O1O1O1O1O1O001O1O1O00O10000O10O010000O100O10000O100O10000O10000O10O010O100O1O010O100O1O100O1O100O010O1001O1O1O1O001O1O1O1\\N\\NjDe1S;cNgD]1W;kNcDV1Y;SOaDn0\\;ZO^Dg0_;A\\D?a;F\\D;e:d0ZE]O\\:m0cETOT:U1kElNS:V1lEkNR:W1mEjNQ:X1oEiNn9Z1PFgNn9[1QFfNm9\\1RFeNl9]1SFdNk9^1TFcNj9_1UFbNi9`1VFaNh9a1XF_Nf9c1YF^Ne9d1ZF\\Ne9Q4N101O00001O001O001O00001O001O001O001O00001O001O001O0100O1O101N100O100O1O1N2O1O1O2N1O1N2O1O1O1O1O1N3N1O1O1O1O1N101O001O001N101O001O001N101O001O001N101O001O2N2M4M2N2N2M3N3L3N2M3N3L3N2M3N3L7J6I7J6I5L2M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N2I7H8H8G9HYdk0"}, "label": "the person sitting in the snow"}]}
{"id": 327462, "image": "COCO_train2014_000000327462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboarder that is sitting on the ground\"."}], "task": "refering", "answer_template": "The \"a snowboarder that is sitting on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 200, 201, 202, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 265, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 342, 361, 362, 363, 364], "bbox": [0.532140625, 0.439765625, 0.913265625, 0.86673828125], "iscrowd": 0, "area": 27794.068999999996, "rle": {"size": [512, 640], "counts": "kjZ5=^?6O1N3M2O1N3N1N2O2M2O2O0O2O0O2O0O2O0O2O0O3N;E1OO100O100O1O10O0100O100O1O1O1O1O1O001O1O1O00O10000O10O010000O100O10000O100O10000O10000O10O010O100O1O010O100O1O100O1O100O010O1001O1O1O1O001O1O1O1\\N\\NjDe1S;cNgD]1W;kNcDV1Y;SOaDn0\\;ZO^Dg0_;A\\D?a;F\\D;e:d0ZE]O\\:m0cETOT:U1kElNS:V1lEkNR:W1mEjNQ:X1oEiNn9Z1PFgNn9[1QFfNm9\\1RFeNl9]1SFdNk9^1TFcNj9_1UFbNi9`1VFaNh9a1XF_Nf9c1YF^Ne9d1ZF\\Ne9Q4N101O00001O001O001O00001O001O001O001O00001O001O001O0100O1O101N100O100O1O1N2O1O1O2N1O1N2O1O1O1O1O1N3N1O1O1O1O1N101O001O001N101O001O001N101O001O001N101O001O2N2M4M2N2N2M3N3L3N2M3N3L3N2M3N3L7J6I7J6I5L2M3N2M3N2M3N2M3N2M3N1N3N2M3N2M3N2M3N2M3N2I7H8H8G9HYdk0"}, "label": "a snowboarder that is sitting on the ground"}]}
{"id": 327462, "image": "COCO_train2014_000000327462.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large man in blue , black and white standing on a snowboard\"."}], "task": "refering", "answer_template": "The \"large man in blue , black and white standing on a snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 31, 32, 33, 34, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 146, 147, 148, 149], "bbox": [0.368546875, 0.0, 0.526734375, 0.3662890625], "iscrowd": 0, "area": 12274.027600000003, "rle": {"size": [512, 640], "counts": "kPf38g?9F;E:oClNd8_1PGPOb8Z1SGVO^8U1UGPOi8T1QGoNm8S1PGPOn8R1oFQOo8Q1nFROP9P1lFTOR9n0kFUOS9m0cF]O\\9e0WFGh9:mE1S:n2000000000000000O10000^OhE[KY:b4kE]KU:_4PF_KQ:]4TFbKl9[4XF_Km9]4m0M3L4L4O1N2O1N2O1N2O1N2J6I6K6M33M2N2N2N2N2N3M2N2N2N2N2N3M8H<Dk0UO5K00000010O4L3M3M3M3N2M3M4L3TMnDd0V;XOnDd0V;WOoDf0T;UOREf0R;UOSEg0Q;TOTEh0Q;SOSEi0R;QOWEg0n:SO]E1^NF[<2bEJXN4Z<LRF3T:GnE8W:CjE<m<01O001N101O2N2N2M3NXZg4"}, "label": "large man in blue , black and white standing on a snowboard"}]}
{"id": 16166, "image": "COCO_train2014_000000016166.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald head of a man\"."}], "task": "refering", "answer_template": "The \"a bald head of a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 287, 288, 289, 290, 291, 302, 303, 304, 305, 306, 307, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.0899067599067599, 0.8050937499999999, 0.5411188811188811, 0.9948750000000001], "iscrowd": 0, "area": 16910.807300000004, "rle": {"size": [640, 429], "counts": "goh02kc04L3N2M4M2M3N3L3M4M2M3N2M3N2M3M3N2M3N2M3N2N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O10000O100000000O100000000O10000000O0100000000O100000000O1000000O10000001O000000001O000000001O0000001O000000001O0000001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O100O001O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O3M3M3M3M3M3M4L3N2M3M3M3M3M4L4L6J7IX`j3"}, "label": "a bald head of a man"}]}
{"id": 16166, "image": "COCO_train2014_000000016166.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald head\"."}], "task": "refering", "answer_template": "The \"a bald head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 287, 288, 289, 290, 291, 302, 303, 304, 305, 306, 307, 316, 317, 318, 319, 320, 321, 322, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.0899067599067599, 0.8050937499999999, 0.5411188811188811, 0.9948750000000001], "iscrowd": 0, "area": 16910.807300000004, "rle": {"size": [640, 429], "counts": "goh02kc04L3N2M4M2M3N3L3M4M2M3N2M3N2M3M3N2M3N2M3N2N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O10000O100000000O100000000O10000000O0100000000O100000000O1000000O10000001O000000001O000000001O0000001O000000001O0000001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O100O001O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O3M3M3M3M3M3M4L3N2M3M3M3M3M4L4L6J7IX`j3"}, "label": "a bald head"}]}
{"id": 48937, "image": "COCO_train2014_000000048937.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a grey sweater holding up a toothbrush next to a bookcase of files\"."}], "task": "refering", "answer_template": "The \"a woman in a grey sweater holding up a toothbrush next to a bookcase of files\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 60, 64, 65, 66, 67, 68, 83, 84, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321], "bbox": [0.605859375, 0.0015529411764705883, 0.999828125, 0.9184235294117647], "iscrowd": 0, "area": 59619.27980000001, "rle": {"size": [425, 640], "counts": "WTQ52U=4L4L4L5K4L4K5L4M3N2M3N2M3N2N2N2N2O1N2N2N2N2N3N1N2N2N2N2N2N2O1N2N21N1L4M3L4L4L5L]EcN^9Y1dFiN\\9S1eFPO[9l0fFWOZ9g0eF\\O[9b0dFA]9<cFF]98bFK^93aF0_9N`F4a9J_F6c9H\\F:e9DZF=g9BYF>h9AWF`0j9_OUFb0l9]OSFd0n9[ORFe0o9ZOPFg0Q:X11O100O1O1OjKbNbL]1^3cNcL\\1^3cNcL\\1]3dNdL[1\\3fNcLZ1]3fNdLY1\\3gNeLX1[3hNfLX1Y3hNhLW1X3iNiLV1W3jNjLU1V3lNjLS1V3mNkLR1U3nNlLQ1S3POnLo0R3QOoLn0Q3ROPMm0P3SORMk0n2VOQMk0n2UORMk0n2UOQMl0o2TOQMl0o2TOQMl0o2TOQMl0o2UOoLl0Q3TOoLl0Q3UOmLl0S3TOjLo0V3ROeLR1[3nNaLW1^3jN^LY1b3gNZL]1f3dNUL`1k3`NRLc1n3^NmKf1S4ZNjKi1V4XNeKl1[4TNaKP2_4QN]KR2c4nMYKV2g4kMUKY2j4gMRK]2n4dMmJ`2S5`MjJc2V5^MeJf2[5[MaJh2_5YM]Jj2c5WMZJk2f5VMVJo2h5QMUJT3i5mLSJX3k5iLQJ\\3m5eLPJ^3o5P2N5K:F:F;E7I3M4L4L4L4L4L3M4L4L4L4L3M4L4L4L4L4L3M=Cf0ZO:F0000000000000000000000000000000000O100000000000000000000000000000000000001O6J5K3M00001O000000001O00000000001O000000001O000000000000O1000000000000000000000000000000000000O1000000000000000000000000001O001N1L5Ja1`NT7kHo1RNYE"}, "label": "a woman in a grey sweater holding up a toothbrush next to a bookcase of files"}]}
{"id": 48937, "image": "COCO_train2014_000000048937.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in grey hoodie holding toothbrush\"."}], "task": "refering", "answer_template": "The \"woman in grey hoodie holding toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 60, 64, 65, 66, 67, 68, 83, 84, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321], "bbox": [0.605859375, 0.0015529411764705883, 0.999828125, 0.9184235294117647], "iscrowd": 0, "area": 59619.27980000001, "rle": {"size": [425, 640], "counts": "WTQ52U=4L4L4L5K4L4K5L4M3N2M3N2M3N2N2N2N2O1N2N2N2N2N3N1N2N2N2N2N2N2O1N2N21N1L4M3L4L4L5L]EcN^9Y1dFiN\\9S1eFPO[9l0fFWOZ9g0eF\\O[9b0dFA]9<cFF]98bFK^93aF0_9N`F4a9J_F6c9H\\F:e9DZF=g9BYF>h9AWF`0j9_OUFb0l9]OSFd0n9[ORFe0o9ZOPFg0Q:X11O100O1O1OjKbNbL]1^3cNcL\\1^3cNcL\\1]3dNdL[1\\3fNcLZ1]3fNdLY1\\3gNeLX1[3hNfLX1Y3hNhLW1X3iNiLV1W3jNjLU1V3lNjLS1V3mNkLR1U3nNlLQ1S3POnLo0R3QOoLn0Q3ROPMm0P3SORMk0n2VOQMk0n2UORMk0n2UOQMl0o2TOQMl0o2TOQMl0o2TOQMl0o2UOoLl0Q3TOoLl0Q3UOmLl0S3TOjLo0V3ROeLR1[3nNaLW1^3jN^LY1b3gNZL]1f3dNUL`1k3`NRLc1n3^NmKf1S4ZNjKi1V4XNeKl1[4TNaKP2_4QN]KR2c4nMYKV2g4kMUKY2j4gMRK]2n4dMmJ`2S5`MjJc2V5^MeJf2[5[MaJh2_5YM]Jj2c5WMZJk2f5VMVJo2h5QMUJT3i5mLSJX3k5iLQJ\\3m5eLPJ^3o5P2N5K:F:F;E7I3M4L4L4L4L4L3M4L4L4L4L3M4L4L4L4L4L3M=Cf0ZO:F0000000000000000000000000000000000O100000000000000000000000000000000000001O6J5K3M00001O000000001O00000000001O000000001O000000000000O1000000000000000000000000000000000000O1000000000000000000000000001O001N1L5Ja1`NT7kHo1RNYE"}, "label": "woman in grey hoodie holding toothbrush"}]}
{"id": 48939, "image": "COCO_train2014_000000048939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the right near the door\"."}], "task": "refering", "answer_template": "The \"the chair on the right near the door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 203, 204, 205, 225, 226, 227, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296], "bbox": [0.762046875, 0.47775700934579435, 0.929484375, 0.8734579439252337], "iscrowd": 0, "area": 8031.0205000000005, "rle": {"size": [428, 640], "counts": "nY\\62X=2N3M2N3M2N3M2N3M2N3M2N3M1OM1O10O01O0102M5L4L5YOf0O2M2O1O2N10001O00001O00001O000ZFcM]8^2WGmMi8S2mFWNR9k1cF_N]9h2000000000D<C=C=O100O100000O1000000L3D=XOh0O1N2O1000O100000000002N5K5K5J7J5K5K4L00O1SMRHROOHo7V1UHkN:Db7a1WH\\Nm0Il6k1l2O1000O10000000ZO[EdNf:Y1h0L3N3M3M3M3M3M3L4H7Ihcb0"}, "label": "the chair on the right near the door"}]}
{"id": 48939, "image": "COCO_train2014_000000048939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bar stool that is closest to the doorway\"."}], "task": "refering", "answer_template": "The \"the bar stool that is closest to the doorway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 203, 204, 205, 225, 226, 227, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296], "bbox": [0.762046875, 0.47775700934579435, 0.929484375, 0.8734579439252337], "iscrowd": 0, "area": 8031.0205000000005, "rle": {"size": [428, 640], "counts": "nY\\62X=2N3M2N3M2N3M2N3M2N3M2N3M1OM1O10O01O0102M5L4L5YOf0O2M2O1O2N10001O00001O00001O000ZFcM]8^2WGmMi8S2mFWNR9k1cF_N]9h2000000000D<C=C=O100O100000O1000000L3D=XOh0O1N2O1000O100000000002N5K5K5J7J5K5K4L00O1SMRHROOHo7V1UHkN:Db7a1WH\\Nm0Il6k1l2O1000O10000000ZO[EdNf:Y1h0L3N3M3M3M3M3M3L4H7Ihcb0"}, "label": "the bar stool that is closest to the doorway"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the handle of some luggage\"."}], "task": "refering", "answer_template": "The \"the handle of some luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 275, 293, 294, 295, 298, 316, 317, 320, 321, 339, 340, 343, 344, 361, 362, 365, 366, 384, 385, 388, 389], "bbox": [0.707859375, 0.48089583333333336, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 15091.735600000009, "rle": {"size": [480, 640], "counts": "gid61l>4L3N2M3M3N2M3M3M3N3L3M3N2M3M3M3N2M3M3M100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O10O01O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O010O1O1O100O1O100YOg01O2N2N2N1O2N2N1O2N2N2N2N3M2N2N2N3M2N2M3N3M2N2N2N3O0O100O2N100O101NYHVNZ2j1dMYN[2f1dM\\N\\2d1aM_N`2_1_McNa2\\1^MfNb2Z1[MiNe2V1ZMlNg2S1WMoNi2P1UMSOk2m0SMUOm2j0RMXOo2g0nL\\OR3c0mL_OS3a0kLAU3>iLEX39gLIY37eLK[34cLO]31aL1`3M_L5a3K\\L8d3G[L;e3EYL>g3@XLb0h3^OULc0m3]OQLc0Q4]OmKb0W4]OfKd0\\4\\ObKc0a4]O]Kc0e4]OXKc0l4\\ORKd0P5\\OnJc0U5]OhJd0[5[OcJd0`5\\O^Jc0e5]OXJd0j5\\OTJc0P6\\OnId0T6]OhIc0[6]OcIc0_6]O_Ib0e6]OXId0j6\\OTIc0o6]OoHc0S7]OjHc0Z7]OcHc0_7]O_Hb0d7_OXHb0j7^OTHa0P8_OmGa0U8_OhGa0[8@bG`0`8@^G?f8AVG`0l8@RG?Q9BlF=W9DeF=^9IYF6j92lENV:9aEFb:a0TE@o:U23N2M3N3L3N2M3N3L3N2M3NSH"}, "label": "the handle of some luggage"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the handle to the man ' s luggage\"."}], "task": "refering", "answer_template": "The \"the handle to the man ' s luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 275, 293, 294, 295, 298, 316, 317, 320, 321, 339, 340, 343, 344, 361, 362, 365, 366, 384, 385, 388, 389], "bbox": [0.707859375, 0.48089583333333336, 1.0, 0.9887708333333334], "iscrowd": 0, "area": 15091.735600000009, "rle": {"size": [480, 640], "counts": "gid61l>4L3N2M3M3N2M3M3M3N3L3M3N2M3M3M3N2M3M3M100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O10O01O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O010O1O1O100O1O100YOg01O2N2N2N1O2N2N1O2N2N2N2N3M2N2N2N3M2N2M3N3M2N2N2N3O0O100O2N100O101NYHVNZ2j1dMYN[2f1dM\\N\\2d1aM_N`2_1_McNa2\\1^MfNb2Z1[MiNe2V1ZMlNg2S1WMoNi2P1UMSOk2m0SMUOm2j0RMXOo2g0nL\\OR3c0mL_OS3a0kLAU3>iLEX39gLIY37eLK[34cLO]31aL1`3M_L5a3K\\L8d3G[L;e3EYL>g3@XLb0h3^OULc0m3]OQLc0Q4]OmKb0W4]OfKd0\\4\\ObKc0a4]O]Kc0e4]OXKc0l4\\ORKd0P5\\OnJc0U5]OhJd0[5[OcJd0`5\\O^Jc0e5]OXJd0j5\\OTJc0P6\\OnId0T6]OhIc0[6]OcIc0_6]O_Ib0e6]OXId0j6\\OTIc0o6]OoHc0S7]OjHc0Z7]OcHc0_7]O_Hb0d7_OXHb0j7^OTHa0P8_OmGa0U8_OhGa0[8@bG`0`8@^G?f8AVG`0l8@RG?Q9BlF=W9DeF=^9IYF6j92lENV:9aEFb:a0TE@o:U23N2M3N3L3N2M3N3L3N2M3NSH"}, "label": "the handle to the man ' s luggage"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the headrest of the man in the background\"."}], "task": "refering", "answer_template": "The \"the headrest of the man in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 55, 56, 57, 58, 59, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 147, 148, 149, 150, 171, 172, 173, 194, 195], "bbox": [0.39437500000000003, 0.096625, 0.5983125, 0.49437499999999995], "iscrowd": 0, "area": 14604.6739, "rle": {"size": [480, 640], "counts": "nXf36d>7I8I6I8H7J7H7I8H7H9E:TOXMUER3i:e0N1O2N1O2M2O2N1O1O2N1N3N1O2N1O2N1N3N1O2N1O2M3N2N1O2N2N2M2O2N2N2N1O2M3N0000000O10O10000000O0100dMgFQOY9m0iFSOW9j0mFUOS9i0oFVOQ9i0QGWOo8i0RGVOn8i0SGWOm8h0UGWOk8h0VGXOi8h0XGWOi8h0YGWOg8h0ZGXOf8g0[GYOe8f0]GYOc8f0^GYOb8h0^GXOb8g0`GXO`8g0aGYO_8g0bGWO_8i0aGVO`8i0bGUO^8l0cGSO]8l0eGRO\\8n0fGoN[8P1gGoNY8Q1hGmNX8S1kGjNV8V1kGiNU8W1mGfNT8Y1nGeNS8[1oGcNQ8\\1RHaNn7`1TH]Nm7b1UH\\Nl7d1VHZNj7e1YHXNh7h1ZHUNg7X1mHgNR7W1SIfNn6W1VIgNk6W1XIhNh6U1\\IiNe6T1_IjNa6U1aIjNa6R1cImN]6P1gInN[6o0hIoNY6n0jIROW6j0mITOT6j0nIUOT6i0nIVOS6h0oIVOR6i0T4N1O2O0O2N1O2N2N1O2O0O2N1O2N1O2N2O0OT]h3"}, "label": "the headrest of the man in the background"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reflection of man\"."}], "task": "refering", "answer_template": "The \"a reflection of man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213], "bbox": [0.15328124999999998, 0.15360416666666668, 0.50565625, 0.5392708333333334], "iscrowd": 0, "area": 28626.363, "rle": {"size": [480, 640], "counts": "RV^11m>2N2N2QM0lF2n85oFMo86nFLR95kFMT95iFLX95eFM[94bFN]94`FN`93]FOc91[F1d91YF1g90VF2j9NTF4k9NRF4n9MoE5Q:KmE7R:KkE7U:JhE8X:IdE:[:I_E;a:GZE<f:FTE>k:EgDg0Y;`100O100N2N2O1N2N2L3N3L4M3M3L4N2O1N2N2O1N2O1N2N2O1N2O100O1O1O1O1O1O1O100O001O1O1O100O1000000O1000000O1000000O1000000O10000O100000O10000000001O00000000000000001O000000000000001O0000000O100000001O00000000000000001O00001O001O1O001O1O001O001O1N101O1O001O1O001O001O2N3M2N3M2N2N3M2N3M9G<D4L2N1O2N1O2M2O2N2N1O2N1O2N1O2N1O2N2N1O2N1O1O2N1O1O1O1O2N1O1O1O2M2O1O1O2N1O1O1O1N3M2N2N2M4M2N3M3M3M2N3M3M3M2N3L4M5KQ\\d4"}, "label": "a reflection of man"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of the man\"."}], "task": "refering", "answer_template": "The \"the reflection of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213], "bbox": [0.15328124999999998, 0.15360416666666668, 0.50565625, 0.5392708333333334], "iscrowd": 0, "area": 28626.363, "rle": {"size": [480, 640], "counts": "RV^11m>2N2N2QM0lF2n85oFMo86nFLR95kFMT95iFLX95eFM[94bFN]94`FN`93]FOc91[F1d91YF1g90VF2j9NTF4k9NRF4n9MoE5Q:KmE7R:KkE7U:JhE8X:IdE:[:I_E;a:GZE<f:FTE>k:EgDg0Y;`100O100N2N2O1N2N2L3N3L4M3M3L4N2O1N2N2O1N2O1N2N2O1N2O100O1O1O1O1O1O1O100O001O1O1O100O1000000O1000000O1000000O1000000O10000O100000O10000000001O00000000000000001O000000000000001O0000000O100000001O00000000000000001O00001O001O1O001O1O001O001O1N101O1O001O1O001O001O2N3M2N3M2N2N3M2N3M9G<D4L2N1O2N1O2M2O2N2N1O2N1O2N1O2N1O2N2N1O2N1O1O2N1O1O1O1O2N1O1O1O2M2O1O1O2N1O1O1O1N3M2N2N2M4M2N3M3M3M2N3M3M3M2N3L4M5KQ\\d4"}, "label": "the reflection of the man"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the old man sitting with the black hat and book on his lap\"."}], "task": "refering", "answer_template": "The \"the old man sitting with the black hat and book on his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0050625, 0.11910416666666666, 0.9842656250000001, 0.9842708333333333], "iscrowd": 0, "area": 135643.3477, "rle": {"size": [480, 640], "counts": "aj1T1_:]3O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000001O00000001O0001O00000000001O0000000000001O00000000001O0000000000001O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O002N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N3M2N3M1O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O000000\\LgNbIX1]6jNbIV1]6lNbIT1]6nNbIR1]6oNcIQ1]6PObIP1]6RObIn0]6SOhGF6V1Q8VOhGE6U1Q8XOgGE7S1Q8ZOfGE8Q1Q8[OfGG7n0R8]OeGG8l0R8_OdGG:i0R8@dGH9h0R8BcGI9e0S8DbGI:c0S8FaGI;b0R8FbGJ;`0R8I`GJ<=S8L^GI?:R80]GG`09S83YGFc07S86WGFd04T89UGEf02T8=RGCi00T8`0PGBl0MS8d0nFAn0KS8h0lF_Oo0IT8k0jF^OQ1GT8]2kGcMU8^2jGbMU8a2jG^MU8d2jG\\MU8f2jGZMU8h2jGXMU8k2iGUMV8m2iGSMV8o2iGQMW8P3iGoLV8T3hGlLW8V3hGjLW8X3hGhLW8Y3iGgLV8[3iGeLV8]3jGbLU8_3kGaLT8a3kG_LU8b3jG^LU8c3kG]LT8e3kG[LT8g3kGYLT8h3mGWLR8k3mGULR8m3mGSLR8o3mGQLS8o3mGQLR8Q4mGoKR8S4nGlKQ8U4oGkKP8W4oGiKP8Y4oGgKP8Z4PHfKo7\\4PHdKo7^4QHaKo7_4QHaKn7a4QH_Kn7c4QH]Kn7d4RH\\Km7f4RHZKm7h4RHXKm7i4THVKk7l4THTKl7m4SHSKa4ROeNl5jLRK`4VObNj5mLPK`4YO`Ni5oLnJa4\\O\\Ng5TMlJ_4@ZNf5VMjJ_4CXNe5XMhJ_4GUNb5\\MgJ_4IRN`6m1dIPN\\6o1gInMZ6Q2iIlMX6T2jIiMW6V2kIhMV6W2lIgMU6X2mIfMT6Z2mIdMT6[2nIdMR6[2PJcMQ6\\2QJbMP6^2QJ`MP6_2RJ_Mo5`2SJ^Mn5a2TJ]Mm5c2TJ[Mm5d2UJZMl5e2VJZMj5e2XJYMi5g2XJWMi5h2YJVMh5i2ZJUMg5j2[JTMf5l2[JRMf5m2\\JRMd5m2^JQMc5n2_JPMb5k2dJSM]5h2iJVMX5e2oJXMR5c2TK\\Ml4_2ZK_Mg4\\2_KbMb4Y2dKeM]4X2gKfMZ4Y2hKeMY4[2hKcMY4\\2iKcMW4\\2kKbMV4^2kK`MV4_2V4O100O1O1O100O1O100O1O1O100O1O100O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000001O01O1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N3M4L4L4L4L3M4L4L4L3M4L4L4L4L2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N4L6J6J5K6J6J5K6J6J6J5K6J6J5K6J6J5K6J6J5K6J6J5K6J6J5K[W4"}, "label": "the old man sitting with the black hat and book on his lap"}]}
{"id": 89902, "image": "COCO_train2014_000000089902.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0050625, 0.11910416666666666, 0.9842656250000001, 0.9842708333333333], "iscrowd": 0, "area": 135643.3477, "rle": {"size": [480, 640], "counts": "aj1T1_:]3O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000001O00000001O0001O00000000001O0000000000001O00000000001O0000000000001O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O002N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N3M2N3M1O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O000000\\LgNbIX1]6jNbIV1]6lNbIT1]6nNbIR1]6oNcIQ1]6PObIP1]6RObIn0]6SOhGF6V1Q8VOhGE6U1Q8XOgGE7S1Q8ZOfGE8Q1Q8[OfGG7n0R8]OeGG8l0R8_OdGG:i0R8@dGH9h0R8BcGI9e0S8DbGI:c0S8FaGI;b0R8FbGJ;`0R8I`GJ<=S8L^GI?:R80]GG`09S83YGFc07S86WGFd04T89UGEf02T8=RGCi00T8`0PGBl0MS8d0nFAn0KS8h0lF_Oo0IT8k0jF^OQ1GT8]2kGcMU8^2jGbMU8a2jG^MU8d2jG\\MU8f2jGZMU8h2jGXMU8k2iGUMV8m2iGSMV8o2iGQMW8P3iGoLV8T3hGlLW8V3hGjLW8X3hGhLW8Y3iGgLV8[3iGeLV8]3jGbLU8_3kGaLT8a3kG_LU8b3jG^LU8c3kG]LT8e3kG[LT8g3kGYLT8h3mGWLR8k3mGULR8m3mGSLR8o3mGQLS8o3mGQLR8Q4mGoKR8S4nGlKQ8U4oGkKP8W4oGiKP8Y4oGgKP8Z4PHfKo7\\4PHdKo7^4QHaKo7_4QHaKn7a4QH_Kn7c4QH]Kn7d4RH\\Km7f4RHZKm7h4RHXKm7i4THVKk7l4THTKl7m4SHSKa4ROeNl5jLRK`4VObNj5mLPK`4YO`Ni5oLnJa4\\O\\Ng5TMlJ_4@ZNf5VMjJ_4CXNe5XMhJ_4GUNb5\\MgJ_4IRN`6m1dIPN\\6o1gInMZ6Q2iIlMX6T2jIiMW6V2kIhMV6W2lIgMU6X2mIfMT6Z2mIdMT6[2nIdMR6[2PJcMQ6\\2QJbMP6^2QJ`MP6_2RJ_Mo5`2SJ^Mn5a2TJ]Mm5c2TJ[Mm5d2UJZMl5e2VJZMj5e2XJYMi5g2XJWMi5h2YJVMh5i2ZJUMg5j2[JTMf5l2[JRMf5m2\\JRMd5m2^JQMc5n2_JPMb5k2dJSM]5h2iJVMX5e2oJXMR5c2TK\\Ml4_2ZK_Mg4\\2_KbMb4Y2dKeM]4X2gKfMZ4Y2hKeMY4[2hKcMY4\\2iKcMW4\\2kKbMV4^2kK`MV4_2V4O100O1O1O100O1O100O1O1O100O1O100O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000001O01O1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N3M4L4L4L4L3M4L4L4L3M4L4L4L4L2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N4L6J6J5K6J6J5K6J6J6J5K6J6J5K6J6J5K6J6J5K6J6J5K6J6J5K[W4"}, "label": "a man wearing glasses"}]}
{"id": 294702, "image": "COCO_train2014_000000294702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of bread with meat between a knife and cup of sauce\"."}], "task": "refering", "answer_template": "The \"slice of bread with meat between a knife and cup of sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 116, 131, 132, 133, 134, 135, 150, 151, 152, 153, 168, 169, 170, 171], "bbox": [0.29238, 0.5175301204819277, 0.52298, 0.8352710843373494], "iscrowd": 0, "area": 7878.504750000001, "rle": {"size": [332, 500], "counts": "aj_12Z:3N1N3N1O2M2O2N1N3N1N3N1O1N3N1N3M2N2N3M2N3M2N2N3M2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1N10001O00000O101O0000000O2O000O101N100O1O101N1001O0010O01O00001O001O00001O001N101O1O1O1O2N1O1N3N1O1O2N1O1N3M2N2M5L3M4K4M4L3L5L3M3L5L3M4BhX]2"}, "label": "slice of bread with meat between a knife and cup of sauce"}]}
{"id": 294702, "image": "COCO_train2014_000000294702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"open sandwich closest to camera\"."}], "task": "refering", "answer_template": "The \"open sandwich closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 116, 131, 132, 133, 134, 135, 150, 151, 152, 153, 168, 169, 170, 171], "bbox": [0.29238, 0.5175301204819277, 0.52298, 0.8352710843373494], "iscrowd": 0, "area": 7878.504750000001, "rle": {"size": [332, 500], "counts": "aj_12Z:3N1N3N1O2M2O2N1N3N1N3N1O1N3N1N3M2N2N3M2N3M2N2N3M2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1N10001O00000O101O0000000O2O000O101N100O1O101N1001O0010O01O00001O001O00001O001N101O1O1O1O2N1O1N3N1O1O2N1O1N3M2N2M5L3M4K4M4L3L5L3M3L5L3M4BhX]2"}, "label": "open sandwich closest to camera"}]}
{"id": 311088, "image": "COCO_train2014_000000311088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the open hood next to the truck\"."}], "task": "refering", "answer_template": "The \"the open hood next to the truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 132, 133, 134, 155, 156, 178, 179], "bbox": [0.7280625, 0.21784753363228698, 0.93596875, 0.46141255605381165], "iscrowd": 0, "area": 5625.803599999999, "rle": {"size": [446, 640], "counts": "lR[68e=2M2O2M2O2N1N3N1N3N1N3N2O0O2_DoNS:R1hEVOT:l0gEXOW:k0cEYO]:h0_E\\O_:h0VEgNDd0V;h0QE@m:i101N100O2O001N101N101N10001O001O0OG[E\\Md:a2bE\\M^:`2hE_MW:^2nE`MR:\\2j0M3M2A^D`Nf;Z1cD`N`;^1a000000000O0100SDbNZ;^1eDoNo:Q1gDcNO;];T1`DSOc;a13N101N2N2O1N2O1O001M3L4L4N3N1N2N2N2O1N2N2N2N3N1O100O1O1O100O1O2O0O1O100O100O11O00010O000010O01O00010O00I701O01O01O00000010O000001O1O1N2O1O1N101O00kUb0"}, "label": "the open hood next to the truck"}]}
{"id": 311088, "image": "COCO_train2014_000000311088.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hood of a vehicle , propped open\"."}], "task": "refering", "answer_template": "The \"the hood of a vehicle , propped open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 132, 133, 134, 155, 156, 178, 179], "bbox": [0.7280625, 0.21784753363228698, 0.93596875, 0.46141255605381165], "iscrowd": 0, "area": 5625.803599999999, "rle": {"size": [446, 640], "counts": "lR[68e=2M2O2M2O2N1N3N1N3N1N3N2O0O2_DoNS:R1hEVOT:l0gEXOW:k0cEYO]:h0_E\\O_:h0VEgNDd0V;h0QE@m:i101N100O2O001N101N101N10001O001O0OG[E\\Md:a2bE\\M^:`2hE_MW:^2nE`MR:\\2j0M3M2A^D`Nf;Z1cD`N`;^1a000000000O0100SDbNZ;^1eDoNo:Q1gDcNO;];T1`DSOc;a13N101N2N2O1N2O1O001M3L4L4N3N1N2N2N2O1N2N2N2N3N1O100O1O1O100O1O2O0O1O100O100O11O00010O000010O01O00010O00I701O01O01O00000010O000001O1O1N2O1O1N101O00kUb0"}, "label": "the hood of a vehicle , propped open"}]}
{"id": 515893, "image": "COCO_train2014_000000515893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sink with its circular metal drain visible at its center\"."}], "task": "refering", "answer_template": "The \"a white sink with its circular metal drain visible at its center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 133, 134, 135, 136, 137, 138, 145, 146, 147, 148, 149, 150, 151, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 200, 201], "bbox": [0.12978666666666666, 0.54596, 0.6674666666666667, 0.8528600000000001], "iscrowd": 0, "area": 19700.862950000002, "rle": {"size": [500, 375], "counts": "VXh0<W?3M2O2M2N3M2N3N2M2N3M2O2M2N3M2O2M3M2N2N2O2M2N2N2O1N2N2N2O1N2N2N2O1N2N2N1O101N1O1O2O0O1O2N101N1O1O2N100O2N1O101N1O1O2O0O1O2N1O2O0O1O2N100O2N1O101N1O1O2N100O2N1O101N1O2N100010O00010O00010O00010O0001O010O000010O00010O0001O01O01O01O01O00010O0010O0000001O00001O0000001O000O2M2O1N2O2N1N2O2M2O1O1N3N1O1N3N1N2O1O2M2O1N3N1O1N2O2N1N2O2M2O1O1N3N1N2O1O2M2O1O2M2O1N2O2N1N2O2M2O1O1N3N1O1N3N1N2O1O2M2O1N3K4K6K5J6K5Jdgl1"}, "label": "a white sink with its circular metal drain visible at its center"}]}
{"id": 515893, "image": "COCO_train2014_000000515893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink furthest from the hair drier\"."}], "task": "refering", "answer_template": "The \"the sink furthest from the hair drier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 133, 134, 135, 136, 137, 138, 145, 146, 147, 148, 149, 150, 151, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 200, 201], "bbox": [0.12978666666666666, 0.54596, 0.6674666666666667, 0.8528600000000001], "iscrowd": 0, "area": 19700.862950000002, "rle": {"size": [500, 375], "counts": "VXh0<W?3M2O2M2N3M2N3N2M2N3M2O2M2N3M2O2M3M2N2N2O2M2N2N2O1N2N2N2O1N2N2N2O1N2N2N1O101N1O1O2O0O1O2N101N1O1O2N100O2N1O101N1O1O2O0O1O2N1O2O0O1O2N100O2N1O101N1O1O2N100O2N1O101N1O2N100010O00010O00010O00010O0001O010O000010O00010O0001O01O01O01O01O00010O0010O0000001O00001O0000001O000O2M2O1N2O2N1N2O2M2O1O1N3N1O1N3N1N2O1O2M2O1N3N1O1N2O2N1N2O2M2O1O1N3N1N2O1O2M2O1O2M2O1N2O2N1N2O2M2O1O1N3N1O1N3N1N2O1O2M2O1N3K4K6K5J6K5Jdgl1"}, "label": "the sink furthest from the hair drier"}]}
{"id": 515893, "image": "COCO_train2014_000000515893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink bowl on the right , next to the wall\"."}], "task": "refering", "answer_template": "The \"the sink bowl on the right , next to the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 111, 112, 113, 114, 115, 124, 125, 126, 127, 139], "bbox": [0.5324533333333333, 0.38445999999999997, 0.8927466666666666, 0.5655], "iscrowd": 0, "area": 8340.301700000002, "rle": {"size": [500, 375], "counts": "RlQ37\\?3M3M2N3M2N3L4M2N3M3M2O2M3M2N3N2M2N3N1N3M2N2O0O2N1O2O0O2N1O2O0O2N101N10000O101N100O100O2O0O100010O000000010O000000010O000001O0001O0001O0001O01O000000010O00000001O000000001O000O1O100O2N1O1N2O1O2N1O1N2O1O1N3N1O1O1N2O2N1N2O1O1O2M2O1O1N2O2N1O1N2O1N3L3N2M3N2M4M2M3N2M3N3LVjc0"}, "label": "the sink bowl on the right , next to the wall"}]}
{"id": 515893, "image": "COCO_train2014_000000515893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink nearest the cord\"."}], "task": "refering", "answer_template": "The \"the sink nearest the cord\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 111, 112, 113, 114, 115, 124, 125, 126, 127, 139], "bbox": [0.5324533333333333, 0.38445999999999997, 0.8927466666666666, 0.5655], "iscrowd": 0, "area": 8340.301700000002, "rle": {"size": [500, 375], "counts": "RlQ37\\?3M3M2N3M2N3L4M2N3M3M2O2M3M2N3N2M2N3N1N3M2N2O0O2N1O2O0O2N1O2O0O2N101N10000O101N100O100O2O0O100010O000000010O000000010O000001O0001O0001O0001O01O000000010O00000001O000000001O000O1O100O2N1O1N2O1O2N1O1N2O1O1N3N1O1O1N2O2N1N2O1O1O2M2O1O1N2O2N1O1N2O1N3L3N2M3N2M4M2M3N2M3N3LVjc0"}, "label": "the sink nearest the cord"}]}
{"id": 180021, "image": "COCO_train2014_000000180021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bigger mother sheep\"."}], "task": "refering", "answer_template": "The \"bigger mother sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 263], "bbox": [0.272125, 0.20429245283018868, 0.53, 0.7784905660377359], "iscrowd": 0, "area": 25363.12585, "rle": {"size": [424, 640], "counts": "[WX23Q=7H8I7H8I7H8I8H7H7J5J7J5K5K5J7J5N2N2N2N3M2N2N2N2O2M2N2N2N2N3M2N2N2N2N3M2N2O1iH^Ke5d4TJfKg5]4QJlKk5V4oISLl5o3mI[Ln5g3kIcLP6]5M1O1O1O101O0000O1N2UOYJRJj5b5fJXJ\\5f5kJRJX5P6o02N2N2N3M2N2N2N2N00001O00000000O10000000001N1O1N2O1N2O1N2O1N21O0000001O0000001O0000001O0000001O00U1kN0001O100O1O1O100O1O1_NWJbKi5\\4cJYK^5e4iJUKX5i4mJSKT5j4SKRKl4m4^KjJc4T5j1N2N2N2N2N2M3N2N2M3VNcGdNa8Y1dGbN_8\\1dG`N_8^1bG_Nb8_1_G_Nd8^1^G`Ne8^1\\G_Nh8`1YG]Ni8b1XG\\Nl8b1TG[NR9a1oF]NV9`1jF_NZ9]1gFaN^9\\1bFcNb9Y1_FfNe9W1[FgNk9T1VFkNn9R1SFlNQ:P1PFoNT:n0mEoNX:n0hEQO\\:k0fESO]:k0cESOb:i0Q1N3L4M2M4M2M4M2Mb^l3"}, "label": "bigger mother sheep"}]}
{"id": 180021, "image": "COCO_train2014_000000180021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep with ram and having heavy wool on it\"."}], "task": "refering", "answer_template": "The \"a sheep with ram and having heavy wool on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 263], "bbox": [0.272125, 0.20429245283018868, 0.53, 0.7784905660377359], "iscrowd": 0, "area": 25363.12585, "rle": {"size": [424, 640], "counts": "[WX23Q=7H8I7H8I7H8I8H7H7J5J7J5K5K5J7J5N2N2N2N3M2N2N2N2O2M2N2N2N2N3M2N2N2N2N3M2N2O1iH^Ke5d4TJfKg5]4QJlKk5V4oISLl5o3mI[Ln5g3kIcLP6]5M1O1O1O101O0000O1N2UOYJRJj5b5fJXJ\\5f5kJRJX5P6o02N2N2N3M2N2N2N2N00001O00000000O10000000001N1O1N2O1N2O1N2O1N21O0000001O0000001O0000001O0000001O00U1kN0001O100O1O1O100O1O1_NWJbKi5\\4cJYK^5e4iJUKX5i4mJSKT5j4SKRKl4m4^KjJc4T5j1N2N2N2N2N2M3N2N2M3VNcGdNa8Y1dGbN_8\\1dG`N_8^1bG_Nb8_1_G_Nd8^1^G`Ne8^1\\G_Nh8`1YG]Ni8b1XG\\Nl8b1TG[NR9a1oF]NV9`1jF_NZ9]1gFaN^9\\1bFcNb9Y1_FfNe9W1[FgNk9T1VFkNn9R1SFlNQ:P1PFoNT:n0mEoNX:n0hEQO\\:k0fESO]:k0cESOb:i0Q1N3L4M2M4M2M4M2Mb^l3"}, "label": "a sheep with ram and having heavy wool on it"}]}
{"id": 180021, "image": "COCO_train2014_000000180021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby lamb stands near his mother\"."}], "task": "refering", "answer_template": "The \"a baby lamb stands near his mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268], "bbox": [0.4771875, 0.4357311320754717, 0.6664843749999999, 0.7991981132075472], "iscrowd": 0, "area": 9070.197499999993, "rle": {"size": [424, 640], "counts": "oln35P=5M1O2N10000O1O1O1N2O1O1O1O1O1O1O1O1O2O000cDTO]:n0_EZO\\:g0`E@[:c0`EA_:h0VE[Oo1[Oc6Y2YIiMg6`2oHaMQ7h2jGWMM:Y8R4001O00001M2O1O1O1M2nKQGl24bMT9BiFg2=XMS91`FV2d:iM^EV2b:jM^E57^1Z:]N_E56`1Z:ZNaE74_1>`Nl8KbF63a1:eNn8DeF63d12iNT9\\OhF72h1HmN\\9TOjF71j20WMa8H^G71]3i8cLWG^3h8bLXG`3f8`LZGa3f8^LZGd3d8\\L\\Ge3c8ZL]Gh3c8WL]Gj3b8VL^Gk3a8TL`G^2n9aMSFU2X:hMjEc1X;O100O100O1002N3M3M1O1O2N1O2N100O1O1O1TFmMb8T2YGPNf8Q2VGRNj8o1RGUNm8l1oFXNQ9i1kFZNT9i1fF[NY9h1aF[N_9i1ZF[Nf9d2O0001O1O1L3M4L4L4L4L4M3J6J6L4L4002N2N2Om0SOhMmEW1U:eNRFV1Q:`NXF^1k:N2N2N2N2N3M6mN\\DMR<Iimg2"}, "label": "a baby lamb stands near his mother"}]}
{"id": 180021, "image": "COCO_train2014_000000180021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby sheep\"."}], "task": "refering", "answer_template": "The \"a baby sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268], "bbox": [0.4771875, 0.4357311320754717, 0.6664843749999999, 0.7991981132075472], "iscrowd": 0, "area": 9070.197499999993, "rle": {"size": [424, 640], "counts": "oln35P=5M1O2N10000O1O1O1N2O1O1O1O1O1O1O1O1O2O000cDTO]:n0_EZO\\:g0`E@[:c0`EA_:h0VE[Oo1[Oc6Y2YIiMg6`2oHaMQ7h2jGWMM:Y8R4001O00001M2O1O1O1M2nKQGl24bMT9BiFg2=XMS91`FV2d:iM^EV2b:jM^E57^1Z:]N_E56`1Z:ZNaE74_1>`Nl8KbF63a1:eNn8DeF63d12iNT9\\OhF72h1HmN\\9TOjF71j20WMa8H^G71]3i8cLWG^3h8bLXG`3f8`LZGa3f8^LZGd3d8\\L\\Ge3c8ZL]Gh3c8WL]Gj3b8VL^Gk3a8TL`G^2n9aMSFU2X:hMjEc1X;O100O100O1002N3M3M1O1O2N1O2N100O1O1O1TFmMb8T2YGPNf8Q2VGRNj8o1RGUNm8l1oFXNQ9i1kFZNT9i1fF[NY9h1aF[N_9i1ZF[Nf9d2O0001O1O1L3M4L4L4L4L4M3J6J6L4L4002N2N2Om0SOhMmEW1U:eNRFV1Q:`NXF^1k:N2N2N2N2N3M6mN\\DMR<Iimg2"}, "label": "a baby sheep"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a woman holding a wineglass and is wearing a white tshirt\"."}], "task": "refering", "answer_template": "The \"this is a woman holding a wineglass and is wearing a white tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 360, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.360609375, 0.25364583333333335, 0.66196875, 0.9647083333333334], "iscrowd": 0, "area": 45904.8667, "rle": {"size": [480, 640], "counts": "cb\\3b1]=<D;D<E;E;D<E8H8G2O1O2N1N2O2N1N3N1O1N3N1O1M4L3N3L3N2M4M2M3M4M2M4nMTJlJn5P5aJbJb5Y5oJXJU5c5ZKPJh4k5iKfIZ4V6[2J5L5J5hJkFa4Y9]KiF^4]9_KeF]4^9cKcFX4b9fK`FV4d9hK^FX4b9fK`FZ4`9dKbF\\4^9cKcF\\4^9bKdF^4o900000000O100000O1000O10000000000O100000000O10000000000O100000000001O00001O00001O00001O0O2O00001O00001O00001O001O00001O00001O00002N2N2N2N2N2N2N2N2N3M2N2N2N2N2NX2gM2O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1^JbG`4^8RKRHk4T9N2N2N2N2N2hLjEb1X:nM[Fm1g9cMlFX2S;K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5KfiT3"}, "label": "this is a woman holding a wineglass and is wearing a white tshirt"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white blouse holding a glass of wine\"."}], "task": "refering", "answer_template": "The \"a woman in a white blouse holding a glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336, 353, 354, 355, 356, 357, 358, 359, 360, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.360609375, 0.25364583333333335, 0.66196875, 0.9647083333333334], "iscrowd": 0, "area": 45904.8667, "rle": {"size": [480, 640], "counts": "cb\\3b1]=<D;D<E;E;D<E8H8G2O1O2N1N2O2N1N3N1O1N3N1O1M4L3N3L3N2M4M2M3M4M2M4nMTJlJn5P5aJbJb5Y5oJXJU5c5ZKPJh4k5iKfIZ4V6[2J5L5J5hJkFa4Y9]KiF^4]9_KeF]4^9cKcFX4b9fK`FV4d9hK^FX4b9fK`FZ4`9dKbF\\4^9cKcF\\4^9bKdF^4o900000000O100000O1000O10000000000O100000000O10000000000O100000000001O00001O00001O00001O0O2O00001O00001O00001O001O00001O00001O00002N2N2N2N2N2N2N2N2N3M2N2N2N2N2NX2gM2O1O001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1^JbG`4^8RKRHk4T9N2N2N2N2N2hLjEb1X:nM[Fm1g9cMlFX2S;K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5KfiT3"}, "label": "a woman in a white blouse holding a glass of wine"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yhe big wine glass the man on the left is holding\"."}], "task": "refering", "answer_template": "The \"yhe big wine glass the man on the left is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 282, 305, 306, 328, 329], "bbox": [0.2009375, 0.4716875, 0.335703125, 0.8821875], "iscrowd": 0, "area": 8989.3463, "rle": {"size": [480, 640], "counts": "egl1?Y>`0_Oa0_O`0D:I3M2N3M3M3M3O1O001N2O1O1O0O2O1O1O1N2O001O1N2O1O001N101O001N101O1O0lF`Lg6b3TIbLl6^3QIfLn6[3PIfLP7Z3PIfLP7Z3oHhLP7X3oHiLQ7X3nHhLS7W3lHjLT7V3kHlLT7U3kHkLV7T3iHmLW7S3WH[L[Ob0^8S3VH^LYO`0b8R3RHaLZO>d8Q3QHcMo7Z2SHfMo7W2RHjMn7T2THlMm7Q2UHoMk7P2VHoMk7Q2UHoMl7o1UHQNk7o1UHQNk7n1VHQNl7n1THRNl7m1UHSNk7m1VHRNl7k1UHUNo7f1RHYNS8c1mG]NV8_1kGaNY8[1gGeN]8V1dGiN`8S1aGmNc8l0`GSOe8f0_GXOe8b0_G\\Of8=^GAf8:]GDh85\\GIi80[GNi8L[G2j8GZG6k8EXG9m8@XG=i;L5K4L`jV6"}, "label": "yhe big wine glass the man on the left is holding"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass holding by a left side man\"."}], "task": "refering", "answer_template": "The \"a wine glass holding by a left side man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 282, 305, 306, 328, 329], "bbox": [0.2009375, 0.4716875, 0.335703125, 0.8821875], "iscrowd": 0, "area": 8989.3463, "rle": {"size": [480, 640], "counts": "egl1?Y>`0_Oa0_O`0D:I3M2N3M3M3M3O1O001N2O1O1O0O2O1O1O1N2O001O1N2O1O001N101O001N101O1O0lF`Lg6b3TIbLl6^3QIfLn6[3PIfLP7Z3PIfLP7Z3oHhLP7X3oHiLQ7X3nHhLS7W3lHjLT7V3kHlLT7U3kHkLV7T3iHmLW7S3WH[L[Ob0^8S3VH^LYO`0b8R3RHaLZO>d8Q3QHcMo7Z2SHfMo7W2RHjMn7T2THlMm7Q2UHoMk7P2VHoMk7Q2UHoMl7o1UHQNk7o1UHQNk7n1VHQNl7n1THRNl7m1UHSNk7m1VHRNl7k1UHUNo7f1RHYNS8c1mG]NV8_1kGaNY8[1gGeN]8V1dGiN`8S1aGmNc8l0`GSOe8f0_GXOe8b0_G\\Of8=^GAf8:]GDh85\\GIi80[GNi8L[G2j8GZG6k8EXG9m8@XG=i;L5K4L`jV6"}, "label": "a wine glass holding by a left side man"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the blue ring\"."}], "task": "refering", "answer_template": "The \"the man with the blue ring\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0, 0.1505625, 0.313484375, 0.9865208333333333], "iscrowd": 0, "area": 57500.01299999999, "rle": {"size": [480, 640], "counts": "ae0W8h61O100O1O1O100O1O100O1O10000O2O000O10000O100O10000O10000O1D]G[Jc8Y5iGgJW8W5lGgJU8W5nGhJR8V5PHjJP8T5SHjJn7T5THlJl7R5WHlJj7R5S1N2N2N2M3N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O100O100O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O1000000000000001O00000000000000000000000000000000000oJkMWMU2c02\\OOa04^OM`05_OKN`MoKg2R4JMaMnKh2T4HKdMnKf2V4GJeMnKf2W4EJgMlKg2Y4CHjMlKe2[4BGQ19POER1:nNDU1;lNCV1<kNAX1>hNAZ1>gN@[1?fN^O]1a0dN]O^1b0bN]O`1b0aN[Ob1d0_NZOc1e0^NXOe1g0[NXOg1g0ZNWOh1h0YNUOj1j0VNUOl1j0UNTOl1l0UNQOm1o0TNoNl1R1TNjNP2V1QNeNS2[1nM_NW2a1jMZNY2g1jMQN[2o1kMgMY2Y2mM\\MW2P1SJYOl3]OU2X1TJXOm3UOT2b1RJWOP4mNR2j1PJYOT4bNP2U2mIYOc:e0^E\\Ob:b0`E^O_:b0bE^O^:`0dE@\\:>fEBZ:=gEDX::iEGW:8jEHV:6lEJU:3mEMS:NRF2n9IWF7i9EZF<f9_O_Fa0a9]OaFc0a9ZO`Ff0c9UO_Fk0c9RO^Fn0d9nN]FS1f9iN[FW1h9dNZF\\1i9`NXF]1g;L3M4L4L4K5J6IR\\]6"}, "label": "the man with the blue ring"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a blue ring on his finger\"."}], "task": "refering", "answer_template": "The \"a man with a blue ring on his finger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0, 0.1505625, 0.313484375, 0.9865208333333333], "iscrowd": 0, "area": 57500.01299999999, "rle": {"size": [480, 640], "counts": "ae0W8h61O100O1O1O100O1O100O1O10000O2O000O10000O100O10000O10000O1D]G[Jc8Y5iGgJW8W5lGgJU8W5nGhJR8V5PHjJP8T5SHjJn7T5THlJl7R5WHlJj7R5S1N2N2N2M3N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O100O100O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O1000000000000001O00000000000000000000000000000000000oJkMWMU2c02\\OOa04^OM`05_OKN`MoKg2R4JMaMnKh2T4HKdMnKf2V4GJeMnKf2W4EJgMlKg2Y4CHjMlKe2[4BGQ19POER1:nNDU1;lNCV1<kNAX1>hNAZ1>gN@[1?fN^O]1a0dN]O^1b0bN]O`1b0aN[Ob1d0_NZOc1e0^NXOe1g0[NXOg1g0ZNWOh1h0YNUOj1j0VNUOl1j0UNTOl1l0UNQOm1o0TNoNl1R1TNjNP2V1QNeNS2[1nM_NW2a1jMZNY2g1jMQN[2o1kMgMY2Y2mM\\MW2P1SJYOl3]OU2X1TJXOm3UOT2b1RJWOP4mNR2j1PJYOT4bNP2U2mIYOc:e0^E\\Ob:b0`E^O_:b0bE^O^:`0dE@\\:>fEBZ:=gEDX::iEGW:8jEHV:6lEJU:3mEMS:NRF2n9IWF7i9EZF<f9_O_Fa0a9]OaFc0a9ZO`Ff0c9UO_Fk0c9RO^Fn0d9nN]FS1f9iN[FW1h9dNZF\\1i9`NXF]1g;L3M4L4L4K5J6IR\\]6"}, "label": "a man with a blue ring on his finger"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blazer holding a wine glass next to a woman and man\"."}], "task": "refering", "answer_template": "The \"a man wearing a blazer holding a wine glass next to a woman and man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 39, 40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 109, 110, 111, 112, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6286562499999999, 0.03820833333333333, 1.0, 0.9842708333333333], "iscrowd": 0, "area": 76876.8551, "rle": {"size": [480, 640], "counts": "Zel5R1Y2oNZ9X1XFPOg9W1kEROT:T1_ESO`:T1REUOl:R1fDVOY;Y201N1O101N1O2M2N2N3M2N3M2N2N3M2O1N3M201N100O2O0O1O100O2O0O100O1O101N100O1THoJh5R5oIgKa5Y4VJ`LY5b3\\JYMS5g2dJcMZ5^2aJgM_5Y2^JjMa5W2\\JlMc5U2ZJnMf5S2VJPNi5Q2SJSNW2dLX1Y5^LVNX2eLW1i8gN\\GV1d8gNbGU1_8iNfGS1[8kNiGS1W8kNnGQ1S8lNSHQ1m7mNXHo0i7oN[Hn0f7PO_Hm0a7POeHl0\\7ROiHk0W7SOnHi0S7UORIh0n6UOXIg0i6WO\\Ie0e6[O]Ic0c6]O`I?a6@bI>^6BdI;]6EeI9[6GfI7[6HhI5Y6KiI3W6MjI1W6NlI0T60nIMS63nILR64PJIQ66RJGo59RJFn5:TJCm5<VJ@l5`0UJ\\On5d0TJXOn5h0TJSOo5l0SJPOP6P1P400O1000000O10000O1000000O10000O1000000O10000O1000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M3M2N3M3M3QEiMo8Y2iFoMU9T2aFTN^9o1XFYNg9j1oE^NP:d1gEeNW:^1_EjN`:l2O2N1O1O2N1O1O1O1O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O000O2O00001O00001O00001O00001O00001O00001O0mIl0"}, "label": "a man wearing a blazer holding a wine glass next to a woman and man"}]}
{"id": 237367, "image": "COCO_train2014_000000237367.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a grey suit\"."}], "task": "refering", "answer_template": "The \"a man wearing a grey suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 39, 40, 41, 42, 62, 63, 64, 65, 85, 86, 87, 88, 109, 110, 111, 112, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6286562499999999, 0.03820833333333333, 1.0, 0.9842708333333333], "iscrowd": 0, "area": 76876.8551, "rle": {"size": [480, 640], "counts": "Zel5R1Y2oNZ9X1XFPOg9W1kEROT:T1_ESO`:T1REUOl:R1fDVOY;Y201N1O101N1O2M2N2N3M2N3M2N2N3M2O1N3M201N100O2O0O1O100O2O0O100O1O101N100O1THoJh5R5oIgKa5Y4VJ`LY5b3\\JYMS5g2dJcMZ5^2aJgM_5Y2^JjMa5W2\\JlMc5U2ZJnMf5S2VJPNi5Q2SJSNW2dLX1Y5^LVNX2eLW1i8gN\\GV1d8gNbGU1_8iNfGS1[8kNiGS1W8kNnGQ1S8lNSHQ1m7mNXHo0i7oN[Hn0f7PO_Hm0a7POeHl0\\7ROiHk0W7SOnHi0S7UORIh0n6UOXIg0i6WO\\Ie0e6[O]Ic0c6]O`I?a6@bI>^6BdI;]6EeI9[6GfI7[6HhI5Y6KiI3W6MjI1W6NlI0T60nIMS63nILR64PJIQ66RJGo59RJFn5:TJCm5<VJ@l5`0UJ\\On5d0TJXOn5h0TJSOo5l0SJPOP6P1P400O1000000O10000O1000000O10000O1000000O10000O1000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M3M2N3M3M3QEiMo8Y2iFoMU9T2aFTN^9o1XFYNg9j1oE^NP:d1gEeNW:^1_EjN`:l2O2N1O1O2N1O1O1O1O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00001O00001O000O2O00001O00001O00001O00001O00001O00001O0mIl0"}, "label": "a man wearing a grey suit"}]}
{"id": 130869, "image": "COCO_train2014_000000130869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a blue shirt on a skateboard\"."}], "task": "refering", "answer_template": "The \"a boy in a blue shirt on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 219, 220, 221, 241, 242, 243, 244, 264, 265, 286, 287, 288, 309, 310, 332, 333], "bbox": [0.3913125, 0.12149882903981266, 0.805125, 0.952950819672131], "iscrowd": 0, "area": 31859.325799999995, "rle": {"size": [427, 640], "counts": "PkX34T=4L3L4N2O1O1O2O0O1O1O101N1O1O100O1O2O0O1O1O100O2N1O10O00001O01O010O0010O01O010O0010O01O010O0010OnIZO@g09EA;91BO7=CC9e0C\\O;e0FZO:f0F[O8g0GYO9g0HYO6h0JXO6h0JYO4i0LVO4j0LWO2j0NVO2k0NUO0l00TO0l00UONm02RONn02SOLn04ROLn05ROIP16POJP16POIQ17oNIn0;QODn0>ROBk0a0UO^Oi0f0VOZOg0i0XOWOg0k0YOROg0Q1YOmNf0W1YOfNh0\\1XObNg0a1YO\\Nh0f1XOWNk0j1TOTNm0m1SOPNo0Q2QOlMQ1V2nNgMU1Y2kNdMW1]2iN_MZ1b2fN[M]1f2bNWM`1j2`NSMb1n2^NoLd1R3]NjLf1W3ZNeLh1\\3YN`Li1a3YNZLj1g3VNULl1l3UNPLm1Q4TNjKo1W4RN_KW2b4iMTK_2m4bMiJf2X5[M^Jn2b5SMTJU3n5kLhI]3Y6dL`Ia3a6`L[Ic3e6`12N100O2N1O100O2N100O1O2N100O2N100O1O20O3M4M2M3N2N2M3N3L4M3L4MN101N1O2N2N1O2O0O2N1O2N1O2O0O2N2N1O2O001N10000O2O00001N100010O0001O01O01O00010O0001O01O1O100O1O1O100O1O100O1RMbHN_71kHEV7:TI]On6`0ZIXOh6f0ZIYOh6e0XI[Oi6c0YI[Oj6b0XI]Oi6a0YI^Oi6?XI@k6>VIAk6>VIAl6>TIAm6`0QI_OR7`0nH_OS7a0mH^OU7a0kH]OX7c0fH]O[7c0eH\\O]7c0cH\\O_7d0_H[Oc7e0]HZOe7e0[HZOg7e0YHZOh7g0VHXOm7g0SHWOP8h0PHVOT8i0kGVOX8i0fGVO^8i0aGUOc8k0[GSOi8l0VGROo8l0PGSOS9m0lFPOX9U1bFhNb9`1UF]No9d1oEYNU:e1lEYNW:e1nETNV:j1oEoMT:Q2d0O001O1O1O010O1O001O2M3M2N3M2N3M2N3M2N3L3N2N3M2N2L4M4K4L5Klmc1"}, "label": "a boy in a blue shirt on a skateboard"}]}
{"id": 130869, "image": "COCO_train2014_000000130869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a blue tee shirt doing tricks on a skateboard\"."}], "task": "refering", "answer_template": "The \"a boy in a blue tee shirt doing tricks on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 59, 60, 61, 62, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 219, 220, 221, 241, 242, 243, 244, 264, 265, 286, 287, 288, 309, 310, 332, 333], "bbox": [0.3913125, 0.12149882903981266, 0.805125, 0.952950819672131], "iscrowd": 0, "area": 31859.325799999995, "rle": {"size": [427, 640], "counts": "PkX34T=4L3L4N2O1O1O2O0O1O1O101N1O1O100O1O2O0O1O1O100O2N1O10O00001O01O010O0010O01O010O0010O01O010O0010OnIZO@g09EA;91BO7=CC9e0C\\O;e0FZO:f0F[O8g0GYO9g0HYO6h0JXO6h0JYO4i0LVO4j0LWO2j0NVO2k0NUO0l00TO0l00UONm02RONn02SOLn04ROLn05ROIP16POJP16POIQ17oNIn0;QODn0>ROBk0a0UO^Oi0f0VOZOg0i0XOWOg0k0YOROg0Q1YOmNf0W1YOfNh0\\1XObNg0a1YO\\Nh0f1XOWNk0j1TOTNm0m1SOPNo0Q2QOlMQ1V2nNgMU1Y2kNdMW1]2iN_MZ1b2fN[M]1f2bNWM`1j2`NSMb1n2^NoLd1R3]NjLf1W3ZNeLh1\\3YN`Li1a3YNZLj1g3VNULl1l3UNPLm1Q4TNjKo1W4RN_KW2b4iMTK_2m4bMiJf2X5[M^Jn2b5SMTJU3n5kLhI]3Y6dL`Ia3a6`L[Ic3e6`12N100O2N1O100O2N100O1O2N100O2N100O1O20O3M4M2M3N2N2M3N3L4M3L4MN101N1O2N2N1O2O0O2N1O2N1O2O0O2N2N1O2O001N10000O2O00001N100010O0001O01O01O00010O0001O01O1O100O1O1O100O1O100O1RMbHN_71kHEV7:TI]On6`0ZIXOh6f0ZIYOh6e0XI[Oi6c0YI[Oj6b0XI]Oi6a0YI^Oi6?XI@k6>VIAk6>VIAl6>TIAm6`0QI_OR7`0nH_OS7a0mH^OU7a0kH]OX7c0fH]O[7c0eH\\O]7c0cH\\O_7d0_H[Oc7e0]HZOe7e0[HZOg7e0YHZOh7g0VHXOm7g0SHWOP8h0PHVOT8i0kGVOX8i0fGVO^8i0aGUOc8k0[GSOi8l0VGROo8l0PGSOS9m0lFPOX9U1bFhNb9`1UF]No9d1oEYNU:e1lEYNW:e1nETNV:j1oEoMT:Q2d0O001O1O1O010O1O001O2M3M2N3M2N3M2N3M2N3L3N2N3M2N2L4M4K4L5Klmc1"}, "label": "a boy in a blue tee shirt doing tricks on a skateboard"}]}
{"id": 352061, "image": "COCO_train2014_000000352061.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cow in the grass\"."}], "task": "refering", "answer_template": "The \"a white cow in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 231, 232, 235, 236, 237, 254, 255, 258, 259, 260], "bbox": [0.046140625000000005, 0.2525393258426966, 0.4582812499999999, 0.7577977528089889], "iscrowd": 0, "area": 30618.859799999995, "rle": {"size": [445, 640], "counts": "aX=U1V<Y1PNZ1L4L4K5L4L4K5LZMjFm0T9VO\\G6d8OaHgN_7]1gHXN[7k1lHjMT7[2T26dEhMh8^2kFnMQ9V2dFTNX9U4SO3M4M0O1O100O1O1O100O1000000O1000000O100000O01000000O10O1000O1001F9H8G9I7I7I7J6K6J5K5L4O1O1O100O2N1O1O101O01O010O0010O02N2O1N2O1N2N0001O00000O2O00001O0000001N100O101N100O100O2O000O1000000O1000001O0O100000000O10001O000O1001O1O1O1cEWMd9j2TFgMa0Jf7k400000000O10O100000000O1J6J5M4M3M3L4^Ob0K5L4K5K4L41O1O1O1O1O1O8H7I9H<C<D:F;E:F6J2O0O2O001N2OM3I7D;E<H8G8I8H8J5O1000000O1000000000000000O2O0001O1O2O0N2M3M3M3F:K5K5K5K5K5K5J6O2N101N1O2N2N1O2^Oa0I8K5L7I6K4O00O1O00100O1O0000O2O0000001O00000O2O0000001O1O1O1N101O1O2NQef4"}, "label": "a white cow in the grass"}]}
{"id": 352061, "image": "COCO_train2014_000000352061.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light tan cow stands in front of a fence\"."}], "task": "refering", "answer_template": "The \"a light tan cow stands in front of a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 231, 232, 235, 236, 237, 254, 255, 258, 259, 260], "bbox": [0.046140625000000005, 0.2525393258426966, 0.4582812499999999, 0.7577977528089889], "iscrowd": 0, "area": 30618.859799999995, "rle": {"size": [445, 640], "counts": "aX=U1V<Y1PNZ1L4L4K5L4L4K5LZMjFm0T9VO\\G6d8OaHgN_7]1gHXN[7k1lHjMT7[2T26dEhMh8^2kFnMQ9V2dFTNX9U4SO3M4M0O1O100O1O1O100O1000000O1000000O100000O01000000O10O1000O1001F9H8G9I7I7I7J6K6J5K5L4O1O1O100O2N1O1O101O01O010O0010O02N2O1N2O1N2N0001O00000O2O00001O0000001N100O101N100O100O2O000O1000000O1000001O0O100000000O10001O000O1001O1O1O1cEWMd9j2TFgMa0Jf7k400000000O10O100000000O1J6J5M4M3M3L4^Ob0K5L4K5K4L41O1O1O1O1O1O8H7I9H<C<D:F;E:F6J2O0O2O001N2OM3I7D;E<H8G8I8H8J5O1000000O1000000000000000O2O0001O1O2O0N2M3M3M3F:K5K5K5K5K5K5J6O2N101N1O2N2N1O2^Oa0I8K5L7I6K4O00O1O00100O1O0000O2O0000001O00000O2O0000001O1O1O1N101O1O2NQef4"}, "label": "a light tan cow stands in front of a fence"}]}
{"id": 122688, "image": "COCO_train2014_000000122688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white document with a blue border\"."}], "task": "refering", "answer_template": "The \"a white document with a blue border\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.38739583333333333, 0.827703125, 0.7417499999999999, 0.9932500000000001], "iscrowd": 0, "area": 14887.815099999994, "rle": {"size": [640, 480], "counts": "Zkd34ec08H8H8H7I8H8H8H7I7H9H7I7M30000000000000001O00000001O00000000000001O000000000000000000010O00000000000000000000001O0000000001O0000000001O000000000000000000001O01O00000000000000000001O00O10000000001O0000000000000000000O100000000000001O000000000000000O1000000000000000001O00000000000O10003M3M4L3M4L3M3M4L3M4L3M4L3M3L5L3M4L3M3M4L3M4L3M4L3M3M4L[l\\2"}, "label": "a white document with a blue border"}]}
{"id": 122688, "image": "COCO_train2014_000000122688.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batch of documents sitting on a table\"."}], "task": "refering", "answer_template": "The \"a batch of documents sitting on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [330, 331, 332, 333, 334, 346, 347, 348, 349, 350, 351, 363, 364, 365, 366, 367, 368, 369, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.38739583333333333, 0.827703125, 0.7417499999999999, 0.9932500000000001], "iscrowd": 0, "area": 14887.815099999994, "rle": {"size": [640, 480], "counts": "Zkd34ec08H8H8H7I8H8H8H7I7H9H7I7M30000000000000001O00000001O00000000000001O000000000000000000010O00000000000000000000001O0000000001O0000000001O000000000000000000001O01O00000000000000000001O00O10000000001O0000000000000000000O100000000000001O000000000000000O1000000000000000001O00000000000O10003M3M4L3M4L3M3M4L3M4L3M4L3M3L5L3M4L3M3M4L3M4L3M4L3M3M4L[l\\2"}, "label": "a batch of documents sitting on a table"}]}
{"id": 114500, "image": "COCO_train2014_000000114500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a parking meter in front of a black car\"."}], "task": "refering", "answer_template": "The \"a parking meter in front of a black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 123, 124, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 346, 347, 348, 349, 350, 363, 364, 365, 366, 367, 381, 382, 383], "bbox": [0.20029166666666667, 0.29809375, 0.7922291666666667, 0.979703125], "iscrowd": 0, "area": 84322.9355, "rle": {"size": [640, 480], "counts": "SXl12mc06J6K5J6J6J6J6J6J5K6J6J6K5J6J6J6J6J6J6J6J6K5J6J6J5K5K5J6K4K6K5J6K5J6K5J6K5J6K5J6K5J6K5J6J6K5J6K5K5L4L4L3M4L3M4L4L3M4L3M4K5L3M4L3M4L3M4L4L3M4L3M4L4L3M4L3M4K5L2N001O1O001O1O001O1O001O1O001N2O010000O10000O10000O1000000O10000O10000O10001N10000O10000OO2M3N1N3N2N1N3N2M2O2N2M2O2M3N1O2M3N1N101O000100O10O0100O100O100O100O010O100O100O100O00100O100O100O100O010O100O100O100O10O01O100O100O100O010O100O100O100O10OM4J6J6J6J6I7J7I7I7I7H8I7I7I6J7I7L4N2N2N2N2O1N1O2N2N2N2N2N2N2N1O2O1N2N2N2N2N2N2N1O2N2N2O1N5K5K5K5K5D=A>B>A>C:F9F;I7M3M3L3N3M3L4M2N3L4M3M2M4M3M3L3N3M3L4M3M2M4M3M3L3N3TOnXn1"}, "label": "a parking meter in front of a black car"}]}
{"id": 114500, "image": "COCO_train2014_000000114500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black parking meter in front of a black car\"."}], "task": "refering", "answer_template": "The \"a black parking meter in front of a black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 123, 124, 125, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 146, 147, 148, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 346, 347, 348, 349, 350, 363, 364, 365, 366, 367, 381, 382, 383], "bbox": [0.20029166666666667, 0.29809375, 0.7922291666666667, 0.979703125], "iscrowd": 0, "area": 84322.9355, "rle": {"size": [640, 480], "counts": "SXl12mc06J6K5J6J6J6J6J6J5K6J6J6K5J6J6J6J6J6J6J6J6K5J6J6J5K5K5J6K4K6K5J6K5J6K5J6K5J6K5J6K5J6K5J6J6K5J6K5K5L4L4L3M4L3M4L4L3M4L3M4K5L3M4L3M4L3M4L4L3M4L3M4L4L3M4L3M4K5L2N001O1O001O1O001O1O001O1O001N2O010000O10000O10000O1000000O10000O10000O10001N10000O10000OO2M3N1N3N2N1N3N2M2O2N2M2O2M3N1O2M3N1N101O000100O10O0100O100O100O100O010O100O100O100O00100O100O100O100O010O100O100O100O10O01O100O100O100O010O100O100O100O10OM4J6J6J6J6I7J7I7I7I7H8I7I7I6J7I7L4N2N2N2N2O1N1O2N2N2N2N2N2N2N1O2O1N2N2N2N2N2N2N1O2N2N2O1N5K5K5K5K5D=A>B>A>C:F9F;I7M3M3L3N3M3L4M2N3L4M3M2M4M3M3L3N3M3L4M3M2M4M3M3L3N3TOnXn1"}, "label": "a black parking meter in front of a black car"}]}
{"id": 89931, "image": "COCO_train2014_000000089931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman sitting in the chair\"."}], "task": "refering", "answer_template": "The \"woman sitting in the chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 33, 34, 35, 48, 49, 50, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 91, 92, 93, 94, 95, 96, 97, 98, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 136, 137, 138, 142, 150, 151, 152, 153, 157, 158, 165, 166, 167, 168, 172, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212, 225, 226], "bbox": [0.01938679245283019, 0.05695312500000001, 0.5534905660377358, 0.69009375], "iscrowd": 0, "area": 37564.17325000001, "rle": {"size": [640, 424], "counts": "PZ52hc07M4l]OH[`0:b_O1T`02i_O7n?KP@?e?DX@g0^?]O^@j0[?[O`@f0_?@Z@b0d?DW@=h?IR@8m?Mm_O5Q`0n100O100O100O100O100O100O1O100O102bNnKiBV4R=SLgBQ4S=YLdBl3V=]LcBg3W=bLaBb3Z=gL_BX3`=b11ND=C<2000O2O1O1N2O1N2L4L3M4L4L5K400010O01O010O010O010O010O010O0001O010O001O010^NTDoJl;P5XDmJi;P5\\DmJd;i3UESK[OR1_;\\3\\FaLc7XOiGh3U1mLR7I[GV3j1nLj61YGP3P2mLf67VGl2W2jLb6?TGg2[2hLa6f0oFb2c2eL]6n0mF\\2i2dLY6R1lF[2m2`LV6X1lFW2Q3^LS6^1iFT2W3[Lo5d1hFR2[3XLl5h1hFo1_3VLh5o1fFk1e3SLe5T2eFh1Z=[NdBf1[=`N`B_1`=`2O2O01O10O01O1O10O01O100O00100000O10O1000O10000O10O1000O11O00001N10001O00001O0010O0001O01O01O00010O001O2O0O1O2N100O2_KoAo2R>lL_Bg2b=TMcBl2`=mLdBR3`=gLcBY3b=_LaBa3c=XL`Bi3k>O1O1O100O1O1O100O1O2N101N3M3M101O1O001O001N2O001O001O001Om_OlL[?S3d@QMY?o2g@SMX?k2i@XMT?h2k@[MT?d2l@]MT?b2l@_MU?_2j@cMW?[2i@gMW?W2i@jMX?T2g@nMY?Q2g@PNZ?n1e@TN\\?j1d@WN]?g1c@[NNZOg>X2[A`N]ONT?`1_AE]>9cAI]>5bAKa>3`AKc>3_AJd>4^AHe>7]AFf>8\\AEg>9[ADh>:ZABj><YA@j>=YA@j>>YA^Oj>`0fAnN^>P1\\2L4K5L3M;Eb0]OVme3"}, "label": "woman sitting in the chair"}]}
{"id": 89931, "image": "COCO_train2014_000000089931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the tan shirt sitting down\"."}], "task": "refering", "answer_template": "The \"the woman in the tan shirt sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 33, 34, 35, 48, 49, 50, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 91, 92, 93, 94, 95, 96, 97, 98, 106, 107, 108, 109, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 127, 128, 136, 137, 138, 142, 150, 151, 152, 153, 157, 158, 165, 166, 167, 168, 172, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212, 225, 226], "bbox": [0.01938679245283019, 0.05695312500000001, 0.5534905660377358, 0.69009375], "iscrowd": 0, "area": 37564.17325000001, "rle": {"size": [640, 424], "counts": "PZ52hc07M4l]OH[`0:b_O1T`02i_O7n?KP@?e?DX@g0^?]O^@j0[?[O`@f0_?@Z@b0d?DW@=h?IR@8m?Mm_O5Q`0n100O100O100O100O100O100O1O100O102bNnKiBV4R=SLgBQ4S=YLdBl3V=]LcBg3W=bLaBb3Z=gL_BX3`=b11ND=C<2000O2O1O1N2O1N2L4L3M4L4L5K400010O01O010O010O010O010O010O0001O010O001O010^NTDoJl;P5XDmJi;P5\\DmJd;i3UESK[OR1_;\\3\\FaLc7XOiGh3U1mLR7I[GV3j1nLj61YGP3P2mLf67VGl2W2jLb6?TGg2[2hLa6f0oFb2c2eL]6n0mF\\2i2dLY6R1lF[2m2`LV6X1lFW2Q3^LS6^1iFT2W3[Lo5d1hFR2[3XLl5h1hFo1_3VLh5o1fFk1e3SLe5T2eFh1Z=[NdBf1[=`N`B_1`=`2O2O01O10O01O1O10O01O100O00100000O10O1000O10000O10O1000O11O00001N10001O00001O0010O0001O01O01O00010O001O2O0O1O2N100O2_KoAo2R>lL_Bg2b=TMcBl2`=mLdBR3`=gLcBY3b=_LaBa3c=XL`Bi3k>O1O1O100O1O1O100O1O2N101N3M3M101O1O001O001N2O001O001O001Om_OlL[?S3d@QMY?o2g@SMX?k2i@XMT?h2k@[MT?d2l@]MT?b2l@_MU?_2j@cMW?[2i@gMW?W2i@jMX?T2g@nMY?Q2g@PNZ?n1e@TN\\?j1d@WN]?g1c@[NNZOg>X2[A`N]ONT?`1_AE]>9cAI]>5bAKa>3`AKc>3_AJd>4^AHe>7]AFf>8\\AEg>9[ADh>:ZABj><YA@j>=YA@j>>YA^Oj>`0fAnN^>P1\\2L4K5L3M;Eb0]OVme3"}, "label": "the woman in the tan shirt sitting down"}]}
{"id": 89931, "image": "COCO_train2014_000000089931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little baby boy who is smashing cake in his own face\"."}], "task": "refering", "answer_template": "The \"the little baby boy who is smashing cake in his own face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 310, 311, 312, 313, 317, 318, 319, 320, 321, 322, 323, 324, 328, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.0033962264150943396, 0.37275, 0.9246933962264151, 0.9876093749999999], "iscrowd": 0, "area": 70725.12294999999, "rle": {"size": [640, 424], "counts": "iS11nc06K6I7I2N2O2M2N2N3N1N2N2O1O1O1N2O1O1N2O000O1000O01O1G9G9G9F:L301N2N2N2O1N2N2N100O1O1O1O101N1O1O100O1O1O100O1O11O001O1O001O010O02O2M3N2M4M2M3N2nAeLZ;^3`DgL_;\\3YDkLd;X3UDnLj;T3nCSMP<P3eC[MY<g2[CdMc<n4O2N1[NjH]FX7^9oH_FS7[9SIbFn6Z9XIcFi6Y9^IcFc6W9eIfF\\6T9lIiFU6Y8lJcGU5h7bKUH_4e7iKXHX4g7jKWHW4i7jKVHV4i7lKUHU4j7nKSHS4m7nKQHS4n7oKPHR4P8PLmGQ4R8QLlGP4S8RLkGo3U8n3O100O1O100O100O1O100O100O100O1O100O100O1O1000000000000O100000000000000O1000000000000000000001O000000001O0000001O000000001O0000001O0000001O0O2O1O001O1O0O2O1O001O1O0O2O1O001O1N101O1O1O1O2M2O2N1O1O2QHXGa4j8XK]Gg4e8RKaGm4`8nJeGQ5]8jJfGV5[8fJiGY5Y8bJkG\\5W8aJlG^5V8]JmGc5T8YJPHf5R8UJRHj5o7RJUHm5m7nIVHQ6l7jIYHU6i7fI[HY6g7bI]H]6d7^I`Hb6b7YIbHe6`7VIfHh6\\7SIiHk6X7QIlHn6U:N2N1N3N1O2N2N1O2N2N1O2N2N1O2N1O2RO`BXKb=e4_B[Kb=b4`B^Kb=^4`BbKb=[4_BeKd=V4^BjKg=P4ZBPLj=j3XBVLm=c3UB]Lo=^3RBbLS>W3oAiLU>R3lAnLY>k2iAUM[>e2gA[M[>a2gA_M[>]2gAcM[>Y2gAgM[>V2fAjM\\>R2fAnM\\>n1fARNZ>l1hATNX>j1jAVNW>g1kAYNU>f1lAYNU>e1mA[NS>c1oA]NR>`1PB`NP>^1RBaNo=]1SBcNm=\\1TBdNm=\\1RBdNn=]1QBcNo=^1PBaNR>_1mAaNS>`1lA_NU>b1jA\\NX>e1gAZN[>f1dAYN]>h1bAVN`>k1_ATNb>m1]AQNf>o1YAPNh>Q2WAnMj>S2UAkMm>V2RAiMP?W2\\110000O10O100000000000O100000000000000O1000O1O1O1O1O1O001N2O1O1O1O1N2O001O1O1N2O1O1O0O2O1O1O1O1N2O1O001O1N200O1O1O10O01O100O1O1O100O001O100O1O1O100O001O100O1O1O10O100000O10000000O01000001O0O1O1O1N2O2N1N2O1O1O1N2O2N1N2O1Od]c0"}, "label": "the little baby boy who is smashing cake in his own face"}]}
{"id": 89931, "image": "COCO_train2014_000000089931.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an infant with a fake tie on its onesie rubs a white substance on itself\"."}], "task": "refering", "answer_template": "The \"an infant with a fake tie on its onesie rubs a white substance on itself\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 310, 311, 312, 313, 317, 318, 319, 320, 321, 322, 323, 324, 328, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.0033962264150943396, 0.37275, 0.9246933962264151, 0.9876093749999999], "iscrowd": 0, "area": 70725.12294999999, "rle": {"size": [640, 424], "counts": "iS11nc06K6I7I2N2O2M2N2N3N1N2N2O1O1O1N2O1O1N2O000O1000O01O1G9G9G9F:L301N2N2N2O1N2N2N100O1O1O1O101N1O1O100O1O1O100O1O11O001O1O001O010O02O2M3N2M4M2M3N2nAeLZ;^3`DgL_;\\3YDkLd;X3UDnLj;T3nCSMP<P3eC[MY<g2[CdMc<n4O2N1[NjH]FX7^9oH_FS7[9SIbFn6Z9XIcFi6Y9^IcFc6W9eIfF\\6T9lIiFU6Y8lJcGU5h7bKUH_4e7iKXHX4g7jKWHW4i7jKVHV4i7lKUHU4j7nKSHS4m7nKQHS4n7oKPHR4P8PLmGQ4R8QLlGP4S8RLkGo3U8n3O100O1O100O100O1O100O100O100O1O100O100O1O1000000000000O100000000000000O1000000000000000000001O000000001O0000001O000000001O0000001O0000001O0O2O1O001O1O0O2O1O001O1O0O2O1O001O1N101O1O1O1O2M2O2N1O1O2QHXGa4j8XK]Gg4e8RKaGm4`8nJeGQ5]8jJfGV5[8fJiGY5Y8bJkG\\5W8aJlG^5V8]JmGc5T8YJPHf5R8UJRHj5o7RJUHm5m7nIVHQ6l7jIYHU6i7fI[HY6g7bI]H]6d7^I`Hb6b7YIbHe6`7VIfHh6\\7SIiHk6X7QIlHn6U:N2N1N3N1O2N2N1O2N2N1O2N2N1O2N1O2RO`BXKb=e4_B[Kb=b4`B^Kb=^4`BbKb=[4_BeKd=V4^BjKg=P4ZBPLj=j3XBVLm=c3UB]Lo=^3RBbLS>W3oAiLU>R3lAnLY>k2iAUM[>e2gA[M[>a2gA_M[>]2gAcM[>Y2gAgM[>V2fAjM\\>R2fAnM\\>n1fARNZ>l1hATNX>j1jAVNW>g1kAYNU>f1lAYNU>e1mA[NS>c1oA]NR>`1PB`NP>^1RBaNo=]1SBcNm=\\1TBdNm=\\1RBdNn=]1QBcNo=^1PBaNR>_1mAaNS>`1lA_NU>b1jA\\NX>e1gAZN[>f1dAYN]>h1bAVN`>k1_ATNb>m1]AQNf>o1YAPNh>Q2WAnMj>S2UAkMm>V2RAiMP?W2\\110000O10O100000000000O100000000000000O1000O1O1O1O1O1O001N2O1O1O1O1N2O001O1O1N2O1O1O0O2O1O1O1O1N2O1O001O1N200O1O1O10O01O100O1O1O100O001O100O1O1O100O001O100O1O1O10O100000O10000000O01000001O0O1O1O1N2O2N1N2O1O1O1N2O2N1N2O1Od]c0"}, "label": "an infant with a fake tie on its onesie rubs a white substance on itself"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in yellow and black shirt\"."}], "task": "refering", "answer_template": "The \"man in yellow and black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 122, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 256, 257, 258, 259, 280, 281], "bbox": [0.125203125, 0.251125, 0.38021875, 0.72775], "iscrowd": 0, "area": 19773.38375, "rle": {"size": [480, 640], "counts": "_iU18c><D6J5K6J5K6J5M4L3N3L3M4L3M4L3N3L3M4L3M4L3M4M2N3N101N101N2O0O2O0O2O0O2O0O2O0O2N1O2OO0000010O0001O3N1N2N2N3N1N2N2N3N1N2N1OHjEgKW:Z4kEbKU:`442]NXLkGj3b7[KUHU14b3c7^KRHX16]3d7TMWHo2g7TMUHm2k7T200O100000000000O1O2O0O101N1O1N3POlGPKV8o4lGoJU8Q5kGnJW8P5kGnJV8Q5lGaJL0Y8_5kG_JO1W8^5lG^J11T8`5mG]J20S8b5l0O1O2O000O2O001N101N2O001N101N2O001N101N107I6I8I6J1N3N1N3eLoDc2R;ZMREd2P;XMTEe2n:YMUEe2m:WMXEe2j:XMZEf2g:WM]Eg2`;M3N2N3L3N2M3N2N3L3N7H9H7I5J2O1N2O2N1N2O1N2N2N2O1N2N2O1N2N2O1N2N2O_ji5"}, "label": "man in yellow and black shirt"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black and yellow shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a black and yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 122, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 256, 257, 258, 259, 280, 281], "bbox": [0.125203125, 0.251125, 0.38021875, 0.72775], "iscrowd": 0, "area": 19773.38375, "rle": {"size": [480, 640], "counts": "_iU18c><D6J5K6J5K6J5M4L3N3L3M4L3M4L3N3L3M4L3M4L3M4M2N3N101N101N2O0O2O0O2O0O2O0O2O0O2N1O2OO0000010O0001O3N1N2N2N3N1N2N2N3N1N2N1OHjEgKW:Z4kEbKU:`442]NXLkGj3b7[KUHU14b3c7^KRHX16]3d7TMWHo2g7TMUHm2k7T200O100000000000O1O2O0O101N1O1N3POlGPKV8o4lGoJU8Q5kGnJW8P5kGnJV8Q5lGaJL0Y8_5kG_JO1W8^5lG^J11T8`5mG]J20S8b5l0O1O2O000O2O001N101N2O001N101N2O001N101N107I6I8I6J1N3N1N3eLoDc2R;ZMREd2P;XMTEe2n:YMUEe2m:WMXEe2j:XMZEf2g:WM]Eg2`;M3N2N3L3N2M3N2N3L3N7H9H7I5J2O1N2O2N1N2O1N2N2N2O1N2N2O1N2N2O1N2N2O_ji5"}, "label": "a man wearing a black and yellow shirt"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver apple laptop being used by a person in a plaid shirt\"."}], "task": "refering", "answer_template": "The \"a silver apple laptop being used by a person in a plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 236, 237, 238, 239, 240, 257, 259, 260, 261, 280, 281, 282, 283, 284, 304, 305], "bbox": [0.179921875, 0.5471666666666667, 0.45785937499999996, 0.7823958333333333], "iscrowd": 0, "area": 6448.236349999999, "rle": {"size": [480, 640], "counts": "aWf11o>0O2O1N2O0O2N2O0O2O1O11N10000O1000000O10000O01O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001N2O0O2N2SOROfCo0Y<QOfCP1Y<ROeCP1Y<QOgCP1X<POgCQ1X<QOfCP1Z<POfCP1Y<ROeCn0\\<j0010O01O010O010O010O010O0010O0010O010O010O0010O010O010O0010O0100O100O100O100O100O100O1O100O100O100O100O100O100O10O01O001N101O001N101O001O010O010O0100O010O^OiB\\OX=c0b001O001O001O001O001O001O1O0010O2O0O101N101N100O2O001N100O2ON2M2N3N2M2N3M3N1N3M2N3N2Ml\\R5"}, "label": "a silver apple laptop being used by a person in a plaid shirt"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver laptop sitting alone on one side of the table\"."}], "task": "refering", "answer_template": "The \"silver laptop sitting alone on one side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 313, 314], "bbox": [0.53040625, 0.5349166666666666, 0.78378125, 0.8006666666666666], "iscrowd": 0, "area": 6799.044700000002, "rle": {"size": [480, 640], "counts": "Qgo42m>1O2M2O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1O1O2N1O1O2M2O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1O1O2N10001O01O01O01O01O00010O00010O000010O00010O0001O01O00010O0001O01O01O0001O000N3M2O1N2N3M2N2O2M2N2N2N3M2O1N3M2N2N2O2M2N2N3M2N2O1N3M2N2N3N1N1O001O010O001O1O0010O01O001O010O001O001O010O001O00100O001O001O010O001O0010O0PmP2"}, "label": "silver laptop sitting alone on one side of the table"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop that no one is using\"."}], "task": "refering", "answer_template": "The \"the laptop that no one is using\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 313, 314], "bbox": [0.53040625, 0.5349166666666666, 0.78378125, 0.8006666666666666], "iscrowd": 0, "area": 6799.044700000002, "rle": {"size": [480, 640], "counts": "Qgo42m>1O2M2O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1O1O2N1O1O2M2O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1O1O2N10001O01O01O01O01O00010O00010O000010O00010O0001O01O00010O0001O01O01O0001O000N3M2O1N2N3M2N2O2M2N2N2N3M2O1N3M2N2N2O2M2N2N3M2N2O1N3M2N2N3N1N1O001O010O001O1O0010O01O001O010O001O001O010O001O00100O001O001O010O001O0010O0PmP2"}, "label": "the laptop that no one is using"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with child in her lap sitting in front of the table\"."}], "task": "refering", "answer_template": "The \"a woman with child in her lap sitting in front of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 222, 223], "bbox": [0.53275, 0.26325, 0.7269375, 0.5790416666666667], "iscrowd": 0, "area": 8855.608849999997, "rle": {"size": [480, 640], "counts": "\\QP52l>5L3M4K5L4L3M4K4M3M4K4M4L3L4NO10O10O10O01000O01000OCkBPOV=n0PCnNP=P1VCkNj<T1b0O1L3I8M3O010O\\O]OmBc0o<DnB;o<LmB4P=2nBNo<8nBGR=<kBET=?iB@W=U1O1O2N1O1O1O1O1O100O2O1N101N2O1N2O1N20O01O1O1O100dD`MV:a2hEaMW:_2gEdMX:]2fEeMY:\\2VE]M196Gn9c2eEdM445HR:a2`EhM723HU:o2fEZM2IX:o2eEXMOL\\:m2dEXMMM`:l2bEfM_:[2_EfMa:Z2_EeMb:\\31O1O1O100O1kLYEo1h:oMYEQ2h:lM[ES2f:kM\\EU2d:iM]EW2d:fM_EY2b:eM`EZ2a:dM`E\\2i:YMZEg2_;01O1O001O010O1O001O010O1O1N3M2N3N2M3M3M3M5K4L5K4L:F?A?BRia2"}, "label": "a woman with child in her lap sitting in front of the table"}]}
{"id": 188239, "image": "COCO_train2014_000000188239.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holds a child on her lap at a kitchen table\"."}], "task": "refering", "answer_template": "The \"a woman holds a child on her lap at a kitchen table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 128, 129, 130, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 222, 223], "bbox": [0.53275, 0.26325, 0.7269375, 0.5790416666666667], "iscrowd": 0, "area": 8855.608849999997, "rle": {"size": [480, 640], "counts": "\\QP52l>5L3M4K5L4L3M4K4M3M4K4M4L3L4NO10O10O10O01000O01000OCkBPOV=n0PCnNP=P1VCkNj<T1b0O1L3I8M3O010O\\O]OmBc0o<DnB;o<LmB4P=2nBNo<8nBGR=<kBET=?iB@W=U1O1O2N1O1O1O1O1O100O2O1N101N2O1N2O1N20O01O1O1O100dD`MV:a2hEaMW:_2gEdMX:]2fEeMY:\\2VE]M196Gn9c2eEdM445HR:a2`EhM723HU:o2fEZM2IX:o2eEXMOL\\:m2dEXMMM`:l2bEfM_:[2_EfMa:Z2_EeMb:\\31O1O1O100O1kLYEo1h:oMYEQ2h:lM[ES2f:kM\\EU2d:iM]EW2d:fM_EY2b:eM`EZ2a:dM`E\\2i:YMZEg2_;01O1O001O010O1O001O010O1O1N3M2N3N2M3M3M3M5K4L5K4L:F?A?BRia2"}, "label": "a woman holds a child on her lap at a kitchen table"}]}
{"id": 343892, "image": "COCO_train2014_000000343892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and orange wetsuit on woman\"."}], "task": "refering", "answer_template": "The \"blue and orange wetsuit on woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 176, 177, 178, 196, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 293, 294, 295, 310, 311, 312, 313, 317, 318, 319, 334, 335, 336, 337, 358, 359, 360, 361, 381, 382, 383, 404, 405, 406, 427, 428, 429, 450, 451], "bbox": [0.496890625, 0.30093749999999997, 0.917015625, 0.849796875], "iscrowd": 0, "area": 27469.09175, "rle": {"size": [640, 640], "counts": "^nV62kc04M3L4L4TABR:a0XE7e:JYE:e:FYE?c:B[Eb0c:^O[Ef0b:[O\\Ei0b:WO\\Em0a:TO]EP1a:PO]ES1b:mN[EV1d:kNZEW1f:iNXEX1h:iNVEY1k:fNSE\\1m:dNQE^1o:bNoD_1S;`NkDb1U;^NiDd1X;[NfDf1[;ZNcDh1];XNaDj1`;UN^Dm1b;SN\\Dn1e;RNYDP2h;PNUDR2k;nMSDS2o;lMoCV2P<lMmCV2S<jMkCW2V<iMhCY2W<hMgCZ2Y<gMdC[2\\<eMbC\\2^<eM`C]2`<dM]C^2b<cM[C_2f<aMXCa2h<`MUCb2j<^MUCd2k<\\MSCe2n<ZMQCh2n<YMPCi2P=VMoBl2Q=TMmBn2R=RMmBP3i2WL[6i0jFR3e2YLc6d0gFU3a2[Lh6a0eFV3]2^Lo6;cFY3Y2_LV78_FZ3W2bLZ75]F[3S2eLa7O[F^3o1gLg7KXF`3j1jLo7FVFb3e1mLV8ASFd3a1PM\\8\\ORFf3[1SMe8WOnEh3W1VMk8ROmEj3R1YMR9mNiEm3o0[MX9hNgEo3k0^M_9dNbEl1L\\Oo07d9`N`Ek11^Og0:h9^N^Eh17@`0;m9]N[Ee1<C9=P:[NZEd1`0D3?S:i0lEhNNa0W:d0nEkNHb0[:a0oEnNBd0`:<PFoN^Og0b:8SFQOXOh0g:5SFSOSOj0j:2UFS2k9lMWFS2j9lMWFR2j9nMWFQ2j9mMXFR2h9nMXFQ2i9oMXFP2i9nMYFQ2g9oMZFP2g9oMZFo1g9PN[Fo1f9PN[Fo1e9QN\\Fm1e9RN]Fm1d9RN]Fm1c9RN^Fn1b9RN_Fl1b9TN_Fk1a9TNaFk1_9UNbFi1^9XNcFg1]9XNeFg1[9YNfF_1a9aN`FX1f9gN\\FT1h9lNYFQ1i9oNYFm0i9ROYFk0i9UOXFh0i9YOXFd0j9[OXFb0j9^OWF?k9AVF<l9CVF:l9EVF8l9GVF6k9IXF3k9LWF1j9OXFNj90ZFLU9SNXDQ2e2In8o0TGnNi8U1ZGhNg8X1[GeNe8\\1]GaNd8a1\\G[Nf8f1[GUNg8n1YGnMi8S2XGhMk8Z2VGaMk8b2UGZMm8g2TGTMo8n2RGmLR9S3nFhLV9X3lFcLW9^3iF]L[9d3`3001O1O010O1O001O1O001O1OdNQ@gNP`0W1R@iNm?W1U@hNk?V1X@hNi?V1Y@jNg?T1\\@kNd?T1]@kNd?T1]@lNc?R1_@nNa?P1a@oN_?P1c@PO]?n0e@RO[?l0g@SOZ?k0h@UOX?i0j@WOV?g0l@XOU?f0m@ZOR?e0o@[OS?c0n@]OR?a0PA^OQ?`0QA@o>=TACl>9XAFj>6YAJg>3\\ALe>0_A0a>MbA3^>IeA7Sa000O100O10000O100O10O10O100O100O10O10O100O100O01000O100O01000O100O10O010000O100O10O10O100O100O100O100O100O01O0010O01O001O010O001O01O01O001O010O001O001O00H80000O1000O01000000O2O1O1O1N101O1N2O00ZhP1"}, "label": "blue and orange wetsuit on woman"}]}
{"id": 343892, "image": "COCO_train2014_000000343892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman surfing\"."}], "task": "refering", "answer_template": "The \"a woman surfing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 176, 177, 178, 196, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 293, 294, 295, 310, 311, 312, 313, 317, 318, 319, 334, 335, 336, 337, 358, 359, 360, 361, 381, 382, 383, 404, 405, 406, 427, 428, 429, 450, 451], "bbox": [0.496890625, 0.30093749999999997, 0.917015625, 0.849796875], "iscrowd": 0, "area": 27469.09175, "rle": {"size": [640, 640], "counts": "^nV62kc04M3L4L4TABR:a0XE7e:JYE:e:FYE?c:B[Eb0c:^O[Ef0b:[O\\Ei0b:WO\\Em0a:TO]EP1a:PO]ES1b:mN[EV1d:kNZEW1f:iNXEX1h:iNVEY1k:fNSE\\1m:dNQE^1o:bNoD_1S;`NkDb1U;^NiDd1X;[NfDf1[;ZNcDh1];XNaDj1`;UN^Dm1b;SN\\Dn1e;RNYDP2h;PNUDR2k;nMSDS2o;lMoCV2P<lMmCV2S<jMkCW2V<iMhCY2W<hMgCZ2Y<gMdC[2\\<eMbC\\2^<eM`C]2`<dM]C^2b<cM[C_2f<aMXCa2h<`MUCb2j<^MUCd2k<\\MSCe2n<ZMQCh2n<YMPCi2P=VMoBl2Q=TMmBn2R=RMmBP3i2WL[6i0jFR3e2YLc6d0gFU3a2[Lh6a0eFV3]2^Lo6;cFY3Y2_LV78_FZ3W2bLZ75]F[3S2eLa7O[F^3o1gLg7KXF`3j1jLo7FVFb3e1mLV8ASFd3a1PM\\8\\ORFf3[1SMe8WOnEh3W1VMk8ROmEj3R1YMR9mNiEm3o0[MX9hNgEo3k0^M_9dNbEl1L\\Oo07d9`N`Ek11^Og0:h9^N^Eh17@`0;m9]N[Ee1<C9=P:[NZEd1`0D3?S:i0lEhNNa0W:d0nEkNHb0[:a0oEnNBd0`:<PFoN^Og0b:8SFQOXOh0g:5SFSOSOj0j:2UFS2k9lMWFS2j9lMWFR2j9nMWFQ2j9mMXFR2h9nMXFQ2i9oMXFP2i9nMYFQ2g9oMZFP2g9oMZFo1g9PN[Fo1f9PN[Fo1e9QN\\Fm1e9RN]Fm1d9RN]Fm1c9RN^Fn1b9RN_Fl1b9TN_Fk1a9TNaFk1_9UNbFi1^9XNcFg1]9XNeFg1[9YNfF_1a9aN`FX1f9gN\\FT1h9lNYFQ1i9oNYFm0i9ROYFk0i9UOXFh0i9YOXFd0j9[OXFb0j9^OWF?k9AVF<l9CVF:l9EVF8l9GVF6k9IXF3k9LWF1j9OXFNj90ZFLU9SNXDQ2e2In8o0TGnNi8U1ZGhNg8X1[GeNe8\\1]GaNd8a1\\G[Nf8f1[GUNg8n1YGnMi8S2XGhMk8Z2VGaMk8b2UGZMm8g2TGTMo8n2RGmLR9S3nFhLV9X3lFcLW9^3iF]L[9d3`3001O1O010O1O001O1O001O1OdNQ@gNP`0W1R@iNm?W1U@hNk?V1X@hNi?V1Y@jNg?T1\\@kNd?T1]@kNd?T1]@lNc?R1_@nNa?P1a@oN_?P1c@PO]?n0e@RO[?l0g@SOZ?k0h@UOX?i0j@WOV?g0l@XOU?f0m@ZOR?e0o@[OS?c0n@]OR?a0PA^OQ?`0QA@o>=TACl>9XAFj>6YAJg>3\\ALe>0_A0a>MbA3^>IeA7Sa000O100O10000O100O10O10O100O100O10O10O100O100O01000O100O01000O100O10O010000O100O10O10O100O100O100O100O100O01O0010O01O001O010O001O01O01O001O010O001O001O00H80000O1000O01000000O2O1O1O1N101O1N2O00ZhP1"}, "label": "a woman surfing"}]}
{"id": 556888, "image": "COCO_train2014_000000556888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy in the white shorts closet to the net\"."}], "task": "refering", "answer_template": "The \"guy in the white shorts closet to the net\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 117, 118, 134, 135, 136, 137, 152, 153, 154, 155, 169, 170, 171, 172, 173, 189, 190, 191, 206, 207, 208, 209, 224, 225, 226, 227, 242, 243, 244, 245], "bbox": [0.42094, 0.3727365728900256, 0.64638, 0.987621483375959], "iscrowd": 0, "area": 12816.550800000003, "rle": {"size": [391, 500], "counts": "Zj`24R<4L4M3L3M4L1O01O0001O000001O01O000001O01O0RH]OY4c0dKBZ4>bKG]4:_KJ_47]KN`44]K0a41[K4b4N[K6c4LXK9e4IXK;f4FVK?g4CVKb0g4_OUKf0h4]OTKg0j4ZOSKj0j4WOSKn0k4ROSKR1k4nNRKW1k4jNSKY1l4fNRK\\1Q5bNmJ_1U5`NhJc1Y5\\NeJe1]5ZN`Jh1b5WN[Jk1g5TNWJm1k4RMoJ_1Da1]5WMkJi4U5_KeJ_4[5hK`Jj3m5\\LnIc3Q6eLiIZ3V6[11O2N2M3N2M4L3L4M3N21O0001O00000001O3M3N101N1O2N2O5J8H9K4M4K2NK5K2N001O1O1O001O1O001O1O0_M[J`Nf5m0ULjMl3h1`3E<J6M2N3M4L4L3M4K5L4L3MW\\S2"}, "label": "guy in the white shorts closet to the net"}]}
{"id": 556888, "image": "COCO_train2014_000000556888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing yellow t - shirt\"."}], "task": "refering", "answer_template": "The \"man wearing yellow t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 117, 118, 134, 135, 136, 137, 152, 153, 154, 155, 169, 170, 171, 172, 173, 189, 190, 191, 206, 207, 208, 209, 224, 225, 226, 227, 242, 243, 244, 245], "bbox": [0.42094, 0.3727365728900256, 0.64638, 0.987621483375959], "iscrowd": 0, "area": 12816.550800000003, "rle": {"size": [391, 500], "counts": "Zj`24R<4L4M3L3M4L1O01O0001O000001O01O000001O01O0RH]OY4c0dKBZ4>bKG]4:_KJ_47]KN`44]K0a41[K4b4N[K6c4LXK9e4IXK;f4FVK?g4CVKb0g4_OUKf0h4]OTKg0j4ZOSKj0j4WOSKn0k4ROSKR1k4nNRKW1k4jNSKY1l4fNRK\\1Q5bNmJ_1U5`NhJc1Y5\\NeJe1]5ZN`Jh1b5WN[Jk1g5TNWJm1k4RMoJ_1Da1]5WMkJi4U5_KeJ_4[5hK`Jj3m5\\LnIc3Q6eLiIZ3V6[11O2N2M3N2M4L3L4M3N21O0001O00000001O3M3N101N1O2N2O5J8H9K4M4K2NK5K2N001O1O1O001O1O001O1O0_M[J`Nf5m0ULjMl3h1`3E<J6M2N3M4L4L3M4K5L4L3MW\\S2"}, "label": "man wearing yellow t - shirt"}]}
{"id": 556888, "image": "COCO_train2014_000000556888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with white shorts and dark top standing between two tennis players at a tennis court\"."}], "task": "refering", "answer_template": "The \"a man with white shorts and dark top standing between two tennis players at a tennis court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 115, 116, 133, 134, 150, 151, 152, 168, 169, 170, 186, 187, 188, 204, 205, 206, 222, 223, 224, 240, 241, 242], "bbox": [0.35928, 0.3697186700767264, 0.50626, 0.9886956521739132], "iscrowd": 0, "area": 10315.172349999999, "rle": {"size": [391, 500], "counts": "RSU29e;9G9G9F:G:F9G9SOgMcGb2o7oMdG[2P8T1F;E:F:F:F:eN]JnKg5o3bJgKc5U4fJbK]5\\4lJZKW5c4X1M3M4OJ6G9H9G8H8G9H9G8H8GbJkK\\3l2VKXMh0j0S4l1XKPM0Q2h4n0[KJe45^KJa45aKJ`44bKK^44bKM^41cK0^4NbK3^4LcK4]4KcK6]4IcK8^4FbK;^4DbK>i4UOXKg0[5iNeJS1a5kN_JQ1g5mNYJo0m5oNSJm0S6POoIk0W6ROjIj0\\6SOeIi0a6TO`Ih0f6UO[Ig0k6VOVIf0P7WORId0T7YOmHc0Y7ZOmH=Z7_OlH8Z7FkH1[7LlHJZ73lHDZ79mH]OY7`0Y2M3M3M3M3M3MhSn2"}, "label": "a man with white shorts and dark top standing between two tennis players at a tennis court"}]}
{"id": 556888, "image": "COCO_train2014_000000556888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white shorts and a black shirt with a necklace on\"."}], "task": "refering", "answer_template": "The \"a man in white shorts and a black shirt with a necklace on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 115, 116, 133, 134, 150, 151, 152, 168, 169, 170, 186, 187, 188, 204, 205, 206, 222, 223, 224, 240, 241, 242], "bbox": [0.35928, 0.3697186700767264, 0.50626, 0.9886956521739132], "iscrowd": 0, "area": 10315.172349999999, "rle": {"size": [391, 500], "counts": "RSU29e;9G9G9F:G:F9G9SOgMcGb2o7oMdG[2P8T1F;E:F:F:F:eN]JnKg5o3bJgKc5U4fJbK]5\\4lJZKW5c4X1M3M4OJ6G9H9G8H8G9H9G8H8GbJkK\\3l2VKXMh0j0S4l1XKPM0Q2h4n0[KJe45^KJa45aKJ`44bKK^44bKM^41cK0^4NbK3^4LcK4]4KcK6]4IcK8^4FbK;^4DbK>i4UOXKg0[5iNeJS1a5kN_JQ1g5mNYJo0m5oNSJm0S6POoIk0W6ROjIj0\\6SOeIi0a6TO`Ih0f6UO[Ig0k6VOVIf0P7WORId0T7YOmHc0Y7ZOmH=Z7_OlH8Z7FkH1[7LlHJZ73lHDZ79mH]OY7`0Y2M3M3M3M3M3MhSn2"}, "label": "a man in white shorts and a black shirt with a necklace on"}]}
{"id": 556888, "image": "COCO_train2014_000000556888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a red / orange shirt holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a boy in a red / orange shirt holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 130, 131, 148, 149, 150, 166, 167, 168, 184, 185, 202, 203, 220, 221, 238, 239], "bbox": [0.23488, 0.42744245524296676, 0.34202, 0.9777493606138107], "iscrowd": 0, "area": 6079.357099999999, "rle": {"size": [391, 500], "counts": "`Q]11P<6J6K5J6\\MBYId0`6CZIc0d6_OVIg0h6[ORIk0l6WOSIj0k6WOTHYO`0`1R6b0^IoM?_1R6c0_IPN<^1S6d0aITN5X1Y6f0aIYN0m0b6k0^I]NJf0k6m0[IbND>T7P1YIJj67VIFm6:SIBP7?oH^OT7c0lHZOW7f0iHVO?`Nk5Z2gIROLZO\\6e1hI5X6KgI7d5^M^J\\2N6c5aM\\JZ215b5dM[JW235`5gMZJV263_5jMVJU2;2^5a0bJ@[5b0eJ^OZ5c0fJROJgM_5X3gJ_N<YNk4Z3iJnMl0gNZ4\\3iJiMR1lNS4[3lJdMW1POk3^3nJ\\M]1VOd3_3oJVMT7i2lHRMY7n2fHoL_7P3aHmLb7R3_HkLe7T3[HiLh7W3XHfLl7Y3b0O0dMXGY1i8bN^G[1b8bNdGg0ASOm82hG?V9\\OQG7a9]OfF6igm3"}, "label": "a boy in a red / orange shirt holding a tennis racket"}]}
{"id": 556888, "image": "COCO_train2014_000000556888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy holding a tennis racket with two men\"."}], "task": "refering", "answer_template": "The \"a boy holding a tennis racket with two men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 130, 131, 148, 149, 150, 166, 167, 168, 184, 185, 202, 203, 220, 221, 238, 239], "bbox": [0.23488, 0.42744245524296676, 0.34202, 0.9777493606138107], "iscrowd": 0, "area": 6079.357099999999, "rle": {"size": [391, 500], "counts": "`Q]11P<6J6K5J6\\MBYId0`6CZIc0d6_OVIg0h6[ORIk0l6WOSIj0k6WOTHYO`0`1R6b0^IoM?_1R6c0_IPN<^1S6d0aITN5X1Y6f0aIYN0m0b6k0^I]NJf0k6m0[IbND>T7P1YIJj67VIFm6:SIBP7?oH^OT7c0lHZOW7f0iHVO?`Nk5Z2gIROLZO\\6e1hI5X6KgI7d5^M^J\\2N6c5aM\\JZ215b5dM[JW235`5gMZJV263_5jMVJU2;2^5a0bJ@[5b0eJ^OZ5c0fJROJgM_5X3gJ_N<YNk4Z3iJnMl0gNZ4\\3iJiMR1lNS4[3lJdMW1POk3^3nJ\\M]1VOd3_3oJVMT7i2lHRMY7n2fHoL_7P3aHmLb7R3_HkLe7T3[HiLh7W3XHfLl7Y3b0O0dMXGY1i8bN^G[1b8bNdGg0ASOm82hG?V9\\OQG7a9]OfF6igm3"}, "label": "a boy holding a tennis racket with two men"}]}
{"id": 352091, "image": "COCO_train2014_000000352091.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red , black , and white kite lying on the grass\"."}], "task": "refering", "answer_template": "The \"a red , black , and white kite lying on the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 298, 299, 300, 301, 302, 303, 304, 316, 317, 318, 319, 320, 321, 322, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356], "bbox": [0.510625, 0.6821875000000001, 0.9993124999999999, 0.90565625], "iscrowd": 0, "area": 20556.97345, "rle": {"size": [640, 480], "counts": "QVj42nc01O2N1O1O2N1O2N1O1O2N1O2N1N2O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N2N1O2N1O2N1O2N1N3N1O00000000000O100000O100000O1000000000O1000O1000000000O1000O100000000000O01000000000000O10O10000000O100000O1000O10000000000O01000000000001O1O2M2O1O1O1O1O1O2N1O2M2O2N1O2N1O2N1O1O2N2N1O2M2O2N2N1O2N1O2N1O2N2N1N3N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N2N1N9H^^O"}, "label": "a red , black , and white kite lying on the grass"}]}
{"id": 352091, "image": "COCO_train2014_000000352091.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black , white , and red kite\"."}], "task": "refering", "answer_template": "The \"a black , white , and red kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [267, 268, 269, 280, 281, 282, 283, 284, 285, 286, 298, 299, 300, 301, 302, 303, 304, 316, 317, 318, 319, 320, 321, 322, 333, 334, 335, 336, 337, 338, 339, 351, 352, 353, 354, 355, 356], "bbox": [0.510625, 0.6821875000000001, 0.9993124999999999, 0.90565625], "iscrowd": 0, "area": 20556.97345, "rle": {"size": [640, 480], "counts": "QVj42nc01O2N1O1O2N1O2N1O1O2N1O2N1N2O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N2N1O2N1O2N1O2N1N3N1O00000000000O100000O100000O1000000000O1000O1000000000O1000O100000000000O01000000000000O10O10000000O100000O1000O10000000000O01000000000001O1O2M2O1O1O1O1O1O2N1O2M2O2N1O2N1O2N1O1O2N2N1O2M2O2N2N1O2N1O2N1O2N2N1N3N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1N3N1O2N1O2N1O2N1O2N2N1N9H^^O"}, "label": "a black , white , and red kite"}]}
{"id": 253796, "image": "COCO_train2014_000000253796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking down\"."}], "task": "refering", "answer_template": "The \"a giraffe looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 176, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 356, 357, 359, 360, 361, 362, 379, 380, 382, 383, 384, 385, 402, 403], "bbox": [0.399578125, 0.30293172690763054, 0.7570937499999999, 0.9650803212851405], "iscrowd": 0, "area": 39972.64805, "rle": {"size": [498, 640], "counts": "ohl32`?1iL3fF0S9>dFCW9j0_FYO[9S1]FoN`9T1]FnNb9S1\\FPOa9S1jDdNQ1:T:T1hDgNo08V:T1hDgNP16W:T1gDiNo06W:h1fE[NY:g1bE\\N^:f1]E^Nb:d1ZE^Nf:d1UE`Nj:a1REbNn:`1mDdNR;n210O1O12N1O100O2N1O1O1O2O0O1O1O2NO2O0O101N4L3N2M4M2M3M4M2M3M4L3M5Kb0^Oa0_Ob0^O5K3M3M3M3M3L4M3L4M3L4M3L5L2M100O2M2M3[Oe0N3M2N2N3M2O1N2N3N1O1O1O2N1O10001O00001O0002N1O2L3YMcJVK_5g4dJVK`5UO\\Jl37kL`5VO\\Jl38jL`5UO\\Jn37jL`5UO\\Jn3?bLY5[O\\JQ4m0QLm4HYJT4\\1aKi4MoI_4P9TKSGi4l9O1N2N2102M2N2N2O2M2N2N2N3N1N2N3M2N2N2N3M2N2N2N3M2N2NPKfFk3W9ULkFk3S9ULnFl3P9ULQGk3l8VLVGk3g8UL[Gk3c8UL_Gk3_8ULbGl3\\8TLfGm3W8TLjGl3T8TLnGl3o7ULSHl3j7TLWHm3g7SL[Hm3c7SL_Hm3_7SLcHn3Z7SLgHm3V7TLkHm3S7SLoHk3Q7ULQIi3o6WLSIg3m6YLUIe3k6[LVId3i6\\LZIb3f6^L[Ia3e6^L^I`3b6`L`I^3`6bLaI]3_6bLdI]3Z6dLgI[3Y6eLiIY3W6fLkIY3U6gLmIW3S6iLoIU3Q6jLQJU3n5lLTJR3l5nLUJQ3k5nLXJP3i5oLXJP3h5PMZJo2e5PM^Jn2c5QM^Jn2b5RM`Jl2`5SMbJl2_5SMcJk2]5UMeJi2[5VMgJi2Y5WMiJg2X5XMiJg2W5XMlJf2T5ZMmJf2S5XMPKf2P5YMSKe2m4ZMUKe2l4ZMVKd2j4[MXKd2h4[M[Kc2f4[M\\Kd2d4[M_Kc2a4]MaKa2`4]MdK`2\\4_MgK_2Y4`MjK_2V4`MmK]2S4bMPL\\2P4cMSL[2n3cMVLZ2j3fMXLX2h3gM\\LV2e3hM^LV2b3iMaLU2_3jMeLS2\\3lMfLR2Z3mMiLR2V3mMnLP2S3nMPMP2P3PNSMm1n2QNUMm1k2RNXMl1Z8M2N2N3L3M4L3M3M4L3N4K6J7I6J6JomZ2"}, "label": "a giraffe looking down"}]}
{"id": 253796, "image": "COCO_train2014_000000253796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the right side has his head where you can see all of it\"."}], "task": "refering", "answer_template": "The \"the giraffe on the right side has his head where you can see all of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 176, 193, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 356, 357, 359, 360, 361, 362, 379, 380, 382, 383, 384, 385, 402, 403], "bbox": [0.399578125, 0.30293172690763054, 0.7570937499999999, 0.9650803212851405], "iscrowd": 0, "area": 39972.64805, "rle": {"size": [498, 640], "counts": "ohl32`?1iL3fF0S9>dFCW9j0_FYO[9S1]FoN`9T1]FnNb9S1\\FPOa9S1jDdNQ1:T:T1hDgNo08V:T1hDgNP16W:T1gDiNo06W:h1fE[NY:g1bE\\N^:f1]E^Nb:d1ZE^Nf:d1UE`Nj:a1REbNn:`1mDdNR;n210O1O12N1O100O2N1O1O1O2O0O1O1O2NO2O0O101N4L3N2M4M2M3M4M2M3M4L3M5Kb0^Oa0_Ob0^O5K3M3M3M3M3L4M3L4M3L4M3L5L2M100O2M2M3[Oe0N3M2N2N3M2O1N2N3N1O1O1O2N1O10001O00001O0002N1O2L3YMcJVK_5g4dJVK`5UO\\Jl37kL`5VO\\Jl38jL`5UO\\Jn37jL`5UO\\Jn3?bLY5[O\\JQ4m0QLm4HYJT4\\1aKi4MoI_4P9TKSGi4l9O1N2N2102M2N2N2O2M2N2N2N3N1N2N3M2N2N2N3M2N2N2N3M2N2NPKfFk3W9ULkFk3S9ULnFl3P9ULQGk3l8VLVGk3g8UL[Gk3c8UL_Gk3_8ULbGl3\\8TLfGm3W8TLjGl3T8TLnGl3o7ULSHl3j7TLWHm3g7SL[Hm3c7SL_Hm3_7SLcHn3Z7SLgHm3V7TLkHm3S7SLoHk3Q7ULQIi3o6WLSIg3m6YLUIe3k6[LVId3i6\\LZIb3f6^L[Ia3e6^L^I`3b6`L`I^3`6bLaI]3_6bLdI]3Z6dLgI[3Y6eLiIY3W6fLkIY3U6gLmIW3S6iLoIU3Q6jLQJU3n5lLTJR3l5nLUJQ3k5nLXJP3i5oLXJP3h5PMZJo2e5PM^Jn2c5QM^Jn2b5RM`Jl2`5SMbJl2_5SMcJk2]5UMeJi2[5VMgJi2Y5WMiJg2X5XMiJg2W5XMlJf2T5ZMmJf2S5XMPKf2P5YMSKe2m4ZMUKe2l4ZMVKd2j4[MXKd2h4[M[Kc2f4[M\\Kd2d4[M_Kc2a4]MaKa2`4]MdK`2\\4_MgK_2Y4`MjK_2V4`MmK]2S4bMPL\\2P4cMSL[2n3cMVLZ2j3fMXLX2h3gM\\LV2e3hM^LV2b3iMaLU2_3jMeLS2\\3lMfLR2Z3mMiLR2V3mMnLP2S3nMPMP2P3PNSMm1n2QNUMm1k2RNXMl1Z8M2N2N3L3M4L3M3M4L3N4K6J7I6J6JomZ2"}, "label": "the giraffe on the right side has his head where you can see all of it"}]}
{"id": 253796, "image": "COCO_train2014_000000253796.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking up while another giraffe next to it looks down\"."}], "task": "refering", "answer_template": "The \"a giraffe looking up while another giraffe next to it looks down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 118, 119, 120, 139, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 210, 211, 212, 213, 214, 234, 235, 236, 237, 238, 259, 260, 261, 262, 282, 283, 284, 285, 306, 307, 308, 329, 330, 331, 332, 353, 354, 355, 376, 377, 378], "bbox": [0.064703125, 0.2561847389558233, 0.47212499999999996, 0.9325903614457831], "iscrowd": 0, "area": 29263.015650000005, "rle": {"size": [498, 640], "counts": "oSd07[?2N1O2O1N1O2N2N1O2N2N1O2N2O0O3M4L3M4L1O2N1O2N100O2N1O0000000000000O1O1O2N1O1O1O1O1O100O1O1O2N3M2WOi0O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1XDYLc;k301O00001O0O101O01O00001O000O2O0000O1M3N2M3M3N2M3N2O001O1O100O1O1O1O100O2002N2N1N3N2N2N1O2M3O0O100O01O0010O01O0011N2N2O2M2ON10001O0O\\MaDW1^;hNcDX1];fNeDZ1[;dNgD]1X;bNiD^1V;aNlD`1S;^NoDb1Q;]NPEd1o:ZNSEf1l:YNVEg1i:YNXEh1f:WN\\Ei1b:WN`Ej1]:WNdEi1[:VNgEj1W:VNkEk1R:VNoEj1P:UNRFl1k9TNWFl1h9TNYFl1e9TN]Fm1`9SNbFn1\\9RNeFo1X9QNjFQ2R9PNoFR2n8nMSGT2i8mMXGT2e8mM\\GU2a8kM`GW2\\8iMfGY2V8hMkGY2S8gMnG[2n7fMSH\\2j7dMWH^2f7bM[H_2c7aM^Ha2_7^McHd2Z7\\MgHf2V7ZMkHg2S7YMnHi2o6WMRIk2j6VMWIl2f6TM[Im2c6SM^Io2_6QMcI\\3n5dLSJk3]5ULeJY4k4gKVK`4b4`K`K^4`4bKaK]4_4cKcK[4]4eKdKZ4\\4fKeKY4[4gKgKW4Y4hKiKW4W4hKlKV4T4iKnKV4R4iKQLU4o3jKSLU4m3iKWLU4i3jKYLU4g3jK\\LT4d3kK^LT4b3kK`LT4`3kKcLS4]3lKeLS4[3lKhLR4X3mKjLR4V3mKmLQ4S3mKPMR4P3mKSMQ4m2nKUMQ4k2nKXMP4h2oKZMP4f2oK\\MP4d2oK_Mo3a2nKcMQ4]2lKhMR4X2jKmMU4j6N1O2N1O2N3M2N3M3M2N3M2N3L4M2N3M2N3M2N3M3M2N3M2N3M3M4L5K4L4L4L4L4L4L4L4L4L4L5K4L4L2N2N3M2NfmS5"}, "label": "a giraffe looking up while another giraffe next to it looks down"}]}
{"id": 409454, "image": "COCO_train2014_000000409454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in the back\"."}], "task": "refering", "answer_template": "The \"the pizza in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200], "bbox": [0.44564062499999996, 0.2977916666666667, 0.799203125, 0.5074583333333333], "iscrowd": 0, "area": 15495.888299999997, "rle": {"size": [480, 640], "counts": "^iU42h>7I7I7N1O2O1N2N3N1N2N3N1N2O2M2N2O0O1O2O0O101O000010O01O00001O00010N1N2N3M2N3M2N2N3N1000001O0O1000001O000O10001O000O11O0001O01O0000010O00O100O2O0O100O101N100O100O2O0O100O100O100O1O100O10O0100O100O010O100O10O0100001O000000001O00000000001O000000001O000000001O000000001O000000001O00000O101O0000001O00001O0O10001O00001O1O2N1O1N2ON2O0O2O1N101O11O1N101lMaCe1`<XNcCf1_<XNbCh1_<UNcCk1]<TNeCj1]<TNdCl1h<00O2O000O2O0O101N1O101N1O101N100O2N100O2N100O101N101N2O0O2O1N2O0O2O1N101N2O0O2O1N2O3L4M3L3N3LeXl1"}, "label": "the pizza in the back"}]}
{"id": 409454, "image": "COCO_train2014_000000409454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza behind other pizza\"."}], "task": "refering", "answer_template": "The \"pizza behind other pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200], "bbox": [0.44564062499999996, 0.2977916666666667, 0.799203125, 0.5074583333333333], "iscrowd": 0, "area": 15495.888299999997, "rle": {"size": [480, 640], "counts": "^iU42h>7I7I7N1O2O1N2N3N1N2N3N1N2O2M2N2O0O1O2O0O101O000010O01O00001O00010N1N2N3M2N3M2N2N3N1000001O0O1000001O000O10001O000O11O0001O01O0000010O00O100O2O0O100O101N100O100O2O0O100O100O100O1O100O10O0100O100O010O100O10O0100001O000000001O00000000001O000000001O000000001O000000001O000000001O00000O101O0000001O00001O0O10001O00001O1O2N1O1N2ON2O0O2O1N101O11O1N101lMaCe1`<XNcCf1_<XNbCh1_<UNcCk1]<TNeCj1]<TNdCl1h<00O2O000O2O0O101N1O101N1O101N100O2N100O2N100O101N101N2O0O2O1N2O0O2O1N101N2O0O2O1N2O3L4M3L3N3LeXl1"}, "label": "pizza behind other pizza"}]}
{"id": 409454, "image": "COCO_train2014_000000409454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza closest to us\"."}], "task": "refering", "answer_template": "The \"the pizza closest to us\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360], "bbox": [0.22971875000000003, 0.5829791666666666, 0.742, 0.9136249999999999], "iscrowd": 0, "area": 39655.6546, "rle": {"size": [480, 640], "counts": "fWU2;b>9G4L4L4L4O1N101N2O1N2N2O1N101N2O1N2N2O1N101N2O1N2O1N2O0O2O1N2N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N2O001O1O1N2O001O0O1000001O0O1000001N100000001N1000000O2O0000000O2O00000O2O0000000O1000000O100000000O1000000O100000000O11O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O000000000000O1000000O10001O0O10000O2O00000O101O0O1000000O2O000O10001O0O100000001O00000O101O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O0000000000000000O100O2O0O100O100O100O10000O100O100O100O100O10000O100O100O2O1N2O1N2O1O0O2O1N2O1N2O1N101N2O1O1N2O1N103L4M3L4M4K4M3L5J5K5K6I6K5K9G8H4L5K4L4L4L5K4L4L4L\\o\\2"}, "label": "the pizza closest to us"}]}
{"id": 409454, "image": "COCO_train2014_000000409454.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"prominent pizza slice\"."}], "task": "refering", "answer_template": "The \"prominent pizza slice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360], "bbox": [0.22971875000000003, 0.5829791666666666, 0.742, 0.9136249999999999], "iscrowd": 0, "area": 39655.6546, "rle": {"size": [480, 640], "counts": "fWU2;b>9G4L4L4L4O1N101N2O1N2N2O1N101N2O1N2N2O1N101N2O1N2O1N2O0O2O1N2N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N2O0O2O1N2O1N2O1N101N2O1N2O1N2O001O1O1N2O001O0O1000001O0O1000001N100000001N1000000O2O0000000O2O00000O2O0000000O1000000O100000000O1000000O100000000O11O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O000000000000O1000000O10001O0O10000O2O00000O101O0O1000000O2O000O10001O0O100000001O00000O101O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O0000000000000000O100O2O0O100O100O100O10000O100O100O100O100O10000O100O100O2O1N2O1N2O1O0O2O1N2O1N2O1N101N2O1O1N2O1N103L4M3L4M4K4M3L5J5K5K6I6K5K9G8H4L5K4L4L4L5K4L4L4L\\o\\2"}, "label": "prominent pizza slice"}]}
{"id": 532335, "image": "COCO_train2014_000000532335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person walking and wearing a coat and a light blue shirt underneath\"."}], "task": "refering", "answer_template": "The \"a person walking and wearing a coat and a light blue shirt underneath\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 100, 101, 115, 116, 117, 130, 131, 132, 145, 146, 147, 159, 160, 161, 162, 175, 176, 177, 190, 191], "bbox": [0.656791569086651, 0.25842187499999997, 0.8487822014051523, 0.5438124999999999], "iscrowd": 0, "area": 10296.419550000004, "rle": {"size": [640, 427], "counts": "hY_52dc09F;E;F:E;E;E;F:E;E;O1O10O0100O100O102M7J5J7J6I1E<ZOe001O001O001O001O001O001O001O001O001O001OO1L4L4N2M3N3L4M3L4M3L4M3L5Lc0\\O8I3L4M3L5L3L4M3003M3M3M3M4LO1O1N2N2N2O1QOP1mNW^X1"}, "label": "a person walking and wearing a coat and a light blue shirt underneath"}]}
{"id": 532335, "image": "COCO_train2014_000000532335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a hat and jacket walking down the street\"."}], "task": "refering", "answer_template": "The \"a person in a hat and jacket walking down the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 100, 101, 115, 116, 117, 130, 131, 132, 145, 146, 147, 159, 160, 161, 162, 175, 176, 177, 190, 191], "bbox": [0.656791569086651, 0.25842187499999997, 0.8487822014051523, 0.5438124999999999], "iscrowd": 0, "area": 10296.419550000004, "rle": {"size": [640, 427], "counts": "hY_52dc09F;E;F:E;E;E;F:E;E;O1O10O0100O100O102M7J5J7J6I1E<ZOe001O001O001O001O001O001O001O001O001O001OO1L4L4N2M3N3L4M3L4M3L4M3L5Lc0\\O8I3L4M3L5L3L4M3003M3M3M3M4LO1O1N2N2N2O1QOP1mNW^X1"}, "label": "a person in a hat and jacket walking down the street"}]}
{"id": 532335, "image": "COCO_train2014_000000532335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in theater mask with gold tie\"."}], "task": "refering", "answer_template": "The \"person in theater mask with gold tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 65, 66, 67, 68, 69, 70, 79, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0033489461358313815, 0.147984375, 1.0, 1.0], "iscrowd": 0, "area": 151785.14685000002, "rle": {"size": [640, 427], "counts": "_U1b0[c0K4[2eM4M3L4M3L4L4M3L4M3G9C=C=C=C=C<D=F:I7I7M3O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O10O0100O1O100O1O100O100O1O10000O10000O100O10000O100O10000O10000O10O010000O100O10000O1O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O001]LjF[LW9a3oF\\LR9a3SG\\Ln8`3YG]Lg8`3^G]Lc8_3cG^L^8_3hG^LX8^3nG_LS8]3SH`Ln7]3XH`Lh7\\3^HaLc7\\3bHaL_7[3hHbLX7[3mHbLT7Z3RIcLo6Z3WIcLi6Y3]IdLd6Y3bIcL_6Y3gIeLY6X3mIdLT6X3SJdLn5Y3WJeLi5W3^JdLd5Y3aJdL`5X3gJcL[5Z3kJbLV5Z3PKcLQ5Z3UKaLm4[3ZKaLg4\\3^K`Ld4\\3cK`L]4]3iK`LX4]3nK^LT4a3V6000000O10000000000000000O100000000000000O10000000000O10O1000O100000000O1000000O1000000O100000000O1000000O1000000O100000O0100000000O1000000O1000000001O00001O00001O00001O00001O00001O000O2O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001N1000001O00001O00001O00001O00001_E^Lb4b3WKSM[4m2^K[Ma4h2UK`Mj4b2mJeMS5^2cJjM\\5^2YJeMg5d2mI`MR6i2cIYM]6P3WITMh6S3nHPMR7T3hHoLW7U3bHoL]7T3^HnLb7V3WHnLh7V3RHlLn7X3kGlLS8Y3gGiLY8[3`GiL_8[3[GhLd8[3VGhLj8\\3PGgLo8]3jFfLV9S7O001O001O1O001O001O001O1O001O001O001O001O1O001O0O2O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N7Ic0]Od0\\Oc0]Oc0]Od0\\Oc0]Od0\\Oc0]Oc0]Od0\\Oe\\O"}, "label": "person in theater mask with gold tie"}]}
{"id": 532335, "image": "COCO_train2014_000000532335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a mask and a shiney gold necktie\"."}], "task": "refering", "answer_template": "The \"a man wearing a mask and a shiney gold necktie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 54, 65, 66, 67, 68, 69, 70, 79, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0033489461358313815, 0.147984375, 1.0, 1.0], "iscrowd": 0, "area": 151785.14685000002, "rle": {"size": [640, 427], "counts": "_U1b0[c0K4[2eM4M3L4M3L4L4M3L4M3G9C=C=C=C=C<D=F:I7I7M3O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O10O0100O1O100O1O100O100O1O10000O10000O100O10000O100O10000O10000O10O010000O100O10000O1O1O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1N2O001]LjF[LW9a3oF\\LR9a3SG\\Ln8`3YG]Lg8`3^G]Lc8_3cG^L^8_3hG^LX8^3nG_LS8]3SH`Ln7]3XH`Lh7\\3^HaLc7\\3bHaL_7[3hHbLX7[3mHbLT7Z3RIcLo6Z3WIcLi6Y3]IdLd6Y3bIcL_6Y3gIeLY6X3mIdLT6X3SJdLn5Y3WJeLi5W3^JdLd5Y3aJdL`5X3gJcL[5Z3kJbLV5Z3PKcLQ5Z3UKaLm4[3ZKaLg4\\3^K`Ld4\\3cK`L]4]3iK`LX4]3nK^LT4a3V6000000O10000000000000000O100000000000000O10000000000O10O1000O100000000O1000000O1000000O100000000O1000000O1000000O100000O0100000000O1000000O1000000001O00001O00001O00001O00001O00001O000O2O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001N1000001O00001O00001O00001O00001_E^Lb4b3WKSM[4m2^K[Ma4h2UK`Mj4b2mJeMS5^2cJjM\\5^2YJeMg5d2mI`MR6i2cIYM]6P3WITMh6S3nHPMR7T3hHoLW7U3bHoL]7T3^HnLb7V3WHnLh7V3RHlLn7X3kGlLS8Y3gGiLY8[3`GiL_8[3[GhLd8[3VGhLj8\\3PGgLo8]3jFfLV9S7O001O001O1O001O001O001O1O001O001O001O001O1O001O0O2O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N7Ic0]Od0\\Oc0]Oc0]Od0\\Oc0]Od0\\Oc0]Oc0]Od0\\Oe\\O"}, "label": "a man wearing a mask and a shiney gold necktie"}]}
{"id": 532335, "image": "COCO_train2014_000000532335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person that ' s whole body is facing the right\"."}], "task": "refering", "answer_template": "The \"a person that ' s whole body is facing the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 103, 104, 118, 119, 132, 133, 134, 147, 148, 149, 162, 163, 177, 178, 192, 193, 207, 208], "bbox": [0.8220140515222483, 0.255078125, 0.9798829039812645, 0.6005156250000001], "iscrowd": 0, "area": 6990.002700000003, "rle": {"size": [640, 427], "counts": "Wgk61mc02N3M3M2N3M3N1N3M2oM@`@c0R?Mi@6i>:RAHa>g0\\A[O]>n0^AUO_>P1\\AROa>S1[APOa>V1ZAlNc>f1l@]NP?\\3L3M4L4M2N3M3N1N3N1N3N2M2C>H8M2O2M2N3O1O00]Ob0E<E;E;D<I7I6J7J6XOh0gNY1I7N1N3M3N2M3N2M2N3N5J5L4K5K6K4Kb^5"}, "label": "a person that ' s whole body is facing the right"}]}
{"id": 507765, "image": "COCO_train2014_000000507765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl standing behind the girls sitting\"."}], "task": "refering", "answer_template": "The \"the girl standing behind the girls sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 97, 98, 99, 112, 113, 114, 127, 128, 129, 143, 144, 158, 173, 293, 294, 308, 309], "bbox": [0.4817990654205608, 0.22275, 0.6641355140186916, 0.9068125], "iscrowd": 0, "area": 8666.114300000001, "rle": {"size": [640, 428], "counts": "QnP43kc08H8I7H5K3K5J5L5K5R^O[NZa0j1f^OWNWa0l1g^OVNWa0l1g^OVNWa0l1g^OVNWa0k1h^OWNVa0k1h^OWNVa0k1h^OWNWa0i1h^OYNVa0Z2O1N2_@[Me=f2YB\\Mf=e2YB\\Me=f2YB[Mg=f2XB[Mf=g2YBZMf=i2VBYMi=k2SBUMl=P3oARMP>R3lAoLR>W3eAmL[>W3]AnLb>W3UAnLi>U400O1cHoJKQ53SKKm42YKJh44]KIb46gKBZ4<kKAU4=mKCS4:PLEQ4;oKEP4<QLBo3`0PL@o3a0QL_On3b0SL^Og3h0XLZO`3l0`LVOZ3o0fLROS3T1lLnN^2h1`MZN]2h1cMYN[2i1cMYNZ2j1dMXNY2k1eMXNW2h1kMZNQ2c1SN`Ni1\\1\\NgN`1U1fNmNV1S1kNPOa0a1_ObN;a1CeN7]1FhN6Z1HjN4X1LjNOY11jNKW16jNGW19kNDV1<CjN>2ZG5V8Ib16^NIe15[NJg15YNJj13WNLk12S9OU^i2"}, "label": "the girl standing behind the girls sitting"}]}
{"id": 507765, "image": "COCO_train2014_000000507765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing in the back with her hands on two girls ' shoulders\"."}], "task": "refering", "answer_template": "The \"woman standing in the back with her hands on two girls ' shoulders\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 97, 98, 99, 112, 113, 114, 127, 128, 129, 143, 144, 158, 173, 293, 294, 308, 309], "bbox": [0.4817990654205608, 0.22275, 0.6641355140186916, 0.9068125], "iscrowd": 0, "area": 8666.114300000001, "rle": {"size": [640, 428], "counts": "QnP43kc08H8I7H5K3K5J5L5K5R^O[NZa0j1f^OWNWa0l1g^OVNWa0l1g^OVNWa0l1g^OVNWa0k1h^OWNVa0k1h^OWNVa0k1h^OWNWa0i1h^OYNVa0Z2O1N2_@[Me=f2YB\\Mf=e2YB\\Me=f2YB[Mg=f2XB[Mf=g2YBZMf=i2VBYMi=k2SBUMl=P3oARMP>R3lAoLR>W3eAmL[>W3]AnLb>W3UAnLi>U400O1cHoJKQ53SKKm42YKJh44]KIb46gKBZ4<kKAU4=mKCS4:PLEQ4;oKEP4<QLBo3`0PL@o3a0QL_On3b0SL^Og3h0XLZO`3l0`LVOZ3o0fLROS3T1lLnN^2h1`MZN]2h1cMYN[2i1cMYNZ2j1dMXNY2k1eMXNW2h1kMZNQ2c1SN`Ni1\\1\\NgN`1U1fNmNV1S1kNPOa0a1_ObN;a1CeN7]1FhN6Z1HjN4X1LjNOY11jNKW16jNGW19kNDV1<CjN>2ZG5V8Ib16^NIe15[NJg15YNJj13WNLk12S9OU^i2"}, "label": "woman standing in the back with her hands on two girls ' shoulders"}]}
{"id": 565115, "image": "COCO_train2014_000000565115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female tennis player on court , jumping to hit the ball , wearing a red blouse and black shorts\"."}], "task": "refering", "answer_template": "The \"a female tennis player on court , jumping to hit the ball , wearing a red blouse and black shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 126, 146, 147, 148, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 263], "bbox": [0.31706249999999997, 0.2966510538641686, 0.48659374999999994, 0.7518501170960188], "iscrowd": 0, "area": 7158.9573, "rle": {"size": [427, 640], "counts": "omd22W=5K4L4L5K4L4L4L2N2N2N2J6H8I7M3NUOkDDR;=SE@j:a0\\E[Ob:e0cEXOZ:i0kETOR:m0TFoNj9Q1[FlNb9T1eFhNX83YGV1R:nNkER1W:oNeER1]:nNaER1`:PO\\EQ1f:POVER1k:oNREQ1P;c0101N2N101N2N2O0O2N2O1N2N2O1N2N2O1000O003M4L3L5L3M4L3M4K4M4L3M4hF\\Mb7U3lGnLQ8W3kGkLS8Y3hGjLV8Z3fGhLX8\\3cGgL[8\\3bGgLZ8^3aGeL]8V4N2N2N2N2N1NPM\\H=c7AnH2P77TI]Oj6g0\\IQO`6U1gIbNU6b1SJVNh5V1`HaNR2OX5e1iHZNX2Il4n1PIVNY2Gi4P2QIWN\\2Ce4S2SIWNj9e1ZFZNe9e1`FWN`9h1R1O1NN31bNeDk0\\;SOfD=O@h;GYDj01\\Om;e0TDXOm;h0UDUOl;k0:00000O1010L5J6I7H`QY4"}, "label": "a female tennis player on court , jumping to hit the ball , wearing a red blouse and black shorts"}]}
{"id": 565115, "image": "COCO_train2014_000000565115.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"serena williams serving a tennis ball\"."}], "task": "refering", "answer_template": "The \"serena williams serving a tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 126, 146, 147, 148, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 263], "bbox": [0.31706249999999997, 0.2966510538641686, 0.48659374999999994, 0.7518501170960188], "iscrowd": 0, "area": 7158.9573, "rle": {"size": [427, 640], "counts": "omd22W=5K4L4L5K4L4L4L2N2N2N2J6H8I7M3NUOkDDR;=SE@j:a0\\E[Ob:e0cEXOZ:i0kETOR:m0TFoNj9Q1[FlNb9T1eFhNX83YGV1R:nNkER1W:oNeER1]:nNaER1`:PO\\EQ1f:POVER1k:oNREQ1P;c0101N2N101N2N2O0O2N2O1N2N2O1N2N2O1000O003M4L3L5L3M4L3M4K4M4L3M4hF\\Mb7U3lGnLQ8W3kGkLS8Y3hGjLV8Z3fGhLX8\\3cGgL[8\\3bGgLZ8^3aGeL]8V4N2N2N2N2N1NPM\\H=c7AnH2P77TI]Oj6g0\\IQO`6U1gIbNU6b1SJVNh5V1`HaNR2OX5e1iHZNX2Il4n1PIVNY2Gi4P2QIWN\\2Ce4S2SIWNj9e1ZFZNe9e1`FWN`9h1R1O1NN31bNeDk0\\;SOfD=O@h;GYDj01\\Om;e0TDXOm;h0UDUOl;k0:00000O1010L5J6I7H`QY4"}, "label": "serena williams serving a tennis ball"}]}
{"id": 49022, "image": "COCO_train2014_000000049022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large brown suitcase beside chair\"."}], "task": "refering", "answer_template": "The \"large brown suitcase beside chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 301, 311, 312, 313, 314, 315, 316, 329, 330], "bbox": [0.3116041666666666, 0.557296875, 0.8209791666666667, 0.8471875000000001], "iscrowd": 0, "area": 29624.331949999996, "rle": {"size": [640, 480], "counts": "RWn25ic04K6K4K6K4K6K4K6K4K6K4K6K4K6K4K6K4L5J5L5J5L5J3N2M2O2M3N2M2O2M3N2M2O2M3N2M2O2M3N2N11000O01000O10O1000O10O1000O10O10O1000O10O1000O10O10O10O1000O10O1000O10O1000O100000O01000000O1000000O10O10O1000000O1000000O0100000O1000000O10O1000O10000O1000000O01000O10000O100O10O10O10000O100O10000O010O10000O10000O10O010000O10000O10O010000O100O1000O010000O100O10000O01000O100O10000O0O2O1O1N2O1N2O1N101N2O1O1N2O1N2O0O2O1N2O1O1N2O1N3N2M3N2M4M2N2M4M2M3N2M4M2M3N2N3L3N2M4M2M3N2M4M2N2M3N3L3N2M4M2M3N2N3L3N2M3N^[e1"}, "label": "large brown suitcase beside chair"}]}
{"id": 49022, "image": "COCO_train2014_000000049022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suitcase to the left of the woman\"."}], "task": "refering", "answer_template": "The \"the suitcase to the left of the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 189, 190, 191, 192, 206, 207, 208, 223, 224, 225, 239, 240, 241, 242, 256, 257, 258, 273, 274, 275], "bbox": [0.08389583333333334, 0.43820312499999997, 0.3116041666666666, 0.737078125], "iscrowd": 0, "area": 11206.501000000002, "rle": {"size": [640, 480], "counts": "f^i02hc06J6J6J6J6J6J6J6J6J6J6J6J6J6M3N2M3N2N2M3N1O2M2O1O2M2O2N1N2O2N1N3N1N3N1O1N3N1O2M2O2N1N2O2N1N3N1N3N1O1N2O001N101O0O2O0012M2N2O1N3N1N2O1N3N1N2O1N2N3N1N2O1N3N1N2O1N3N1M3L4L4L5K4L4K5L5K4L4L4L5K4K5L4L4L5K4L4K5L5K4L4L4L4L5JQc^6"}, "label": "the suitcase to the left of the woman"}]}
{"id": 49022, "image": "COCO_train2014_000000049022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan suitcase is sitting on the left side of the woman\"."}], "task": "refering", "answer_template": "The \"a tan suitcase is sitting on the left side of the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 189, 190, 191, 192, 206, 207, 208, 223, 224, 225, 239, 240, 241, 242, 256, 257, 258, 273, 274, 275], "bbox": [0.08389583333333334, 0.43820312499999997, 0.3116041666666666, 0.737078125], "iscrowd": 0, "area": 11206.501000000002, "rle": {"size": [640, 480], "counts": "f^i02hc06J6J6J6J6J6J6J6J6J6J6J6J6J6M3N2M3N2N2M3N1O2M2O1O2M2O2N1N2O2N1N3N1N3N1O1N3N1O2M2O2N1N2O2N1N3N1N3N1O1N2O001N101O0O2O0012M2N2O1N3N1N2O1N3N1N2O1N2N3N1N2O1N3N1N2O1N3N1M3L4L4L5K4L4K5L5K4L4L4L5K4K5L4L4L5K4L4K5L5K4L4L4L4L5JQc^6"}, "label": "a tan suitcase is sitting on the left side of the woman"}]}
{"id": 212863, "image": "COCO_train2014_000000212863.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plane called the overnight delivery system\"."}], "task": "refering", "answer_template": "The \"a plane called the overnight delivery system\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.003015625, 0.004541176470588235, 1.0, 0.8027294117647059], "iscrowd": 0, "area": 107641.54865, "rle": {"size": [425, 640], "counts": "To0Z1o;00001O01O000001O0001O0001O0001O0001O000001O0001O000001O01O000001O01O000000010O000000010O00000000010O0000000010O000000010OcDjNg:V1XEoNd:Q1[ETOa:l0^EZO]:g0aEZO_:f0_E\\Ob:c0]E_Ob:a0]E@c:`0\\EAd:?[EBe:?YEBg:>XECi:<VEFi::VEGj:\\100001O01hGcMb5]2[JgMd5Y2ZJjMe5V2XJnMg5R2WJQNi5n1UJUNj5k1SJYNl5h1QJ[Nn5e1VHgM`1f0Y6n1fITNZ6k1eIWNZ6i1fIXNY6h1fIYNZ6g1eI[NZ6e1eI]NZ6c1fI]N[6b1fI]NZ6c1gI\\NY6d1hI[NX6e1iIZNW6f1jIYNV6g1kIXNV6g1kIYNT6U2_IjMa6T3bHkL^7`4001O00O101N1O100O1O100O101N1O100O10iN_HfL`7R3iHnLW7i2SIVMl6d2\\I\\Mc6b2_I^M`6a2cI^M\\6a2fI_MZ6_2nH_L>Q1c6_2mHfL?j0d6^2lHlL?f0d6]2kHSM`0`0d6[2kHZM`0:d6[2jH`Ma05e6Y2iHgMa0Oe6`3ZI_Le6d3YI\\Lg6e3YIXLh6k3YIoKj6S4XIfKj6[4ZI^Ki6c4ZIVKh6k4k01N101N1O2O0O2O0O2O0O2O0O2O0O2N101N101N101N3N3L4M2M2N2O0O2O0O2O0O2O0O2O0O1OO2N1O2N1O2N1O2O02O0O2N101N100O2O0O2O0O1O2O0O2O0OO2O001N101O0002O0O100O100O100O100O2O0O100O100O100O100O101N100O100O100O100O100O2O0O100O100O10PLlIk0S6VOmIj0S6UOnIk0Q6VOPJi0P6WOPJj0n5VOSJj0m5VOTJi0k5WOVJi0j5WOWJh0i5WOXJi0g5XOYJh0g5XOZJg0e5YO\\Jg0d5YO\\Jg0c5YO_Jf0a5ZO_Jf0`5[O`Jf0_5YOcJf0\\5[OdJe0\\5ZOfJe0Y5\\OgJd0Y5\\OgJd0X5\\OjJc0V5]OjJc0U5]OlJc0T5]OmJb0R5^OoJb0Q5^OoJc0o4^ORKa0n4^OSKb0l4_OUK`0k4_OVKa0i4@WK`0i4@XK?g4AZK?f4AZK?e4A]K>c4B]K>b4C^K>a4AaK>^4CbK=^4BdK=[4DeK<[4CfK=Y4DgK<Y4DgK<X4DhK=X4ChK=W4CjK=V4CjK=U4DjK>U4AlK?S4BmK>S4AmK`0R4AnK?R4AnK?Q4AoK`0Q4@oK`0P4@QL`0o3@PLa0o3_ORLa0n3_ORLb0l3_OSLb0m3]OTLc0k3^OULb0k3]OULd0j3]OVLc0j3]OVLc0i3]OWLd0i3\\OWLd0h3\\OYLd0g3\\OXLe0g3\\OYLd0g3[OZLf0d3[O[Lf0e3YO\\Lg0c3ZO]Lf0c3YO]Lh0c3XO]Lh0b3YO^Lg0b3XO^Li0a3XO_Lh0a3WO`Li0_3XO`Li0`3WO`Lj0^3VOcLj0]3VObLk0]3UOdLk0\\3UOdLk0[3VOdLk0\\3TOeLl0Z3UOfLk0Z3TOgLl0X3UOhLk0X3TOhLn0V3SOjLm0V3SOjLm0U3SOlLm0T3SOlLm0S3SOnLm0R3SOnLm0Q3TOnLm0R3ROoLn0P3SOPMm0P3ROQMo0m2ROSMn0m2ROSMn0l2ROUMn0k2ROTMo0k2QOVMo0j2QOVMo0i2QOXMo0h2QOXMo0g2ROYMn0g2QOZMP1d2QO[MP1e2oN\\MQ1c2PO]MP1c2PO]MP1b2PO_MP1a2PO_MP1kNeKY3[3lMQ1jNeKZ3Z3lMQ1hNgK[3Y3mMP1gNhK\\3W3nMR1cNiK^3V3oMQ1bNjK_3T3PN]1o1dNQN\\1o1cNQN^1n1cNRN]1n1cNRN]1m1cNTN]1l1cNTN\\1l1dNUN\\1k1dNUN\\1j1eNUN\\1k1cNVN]1i1dNWN\\1i1cNXN]1g1dNYN\\1g1dNYN\\1f1dN[N\\1e1dNZN]1e1cN\\N]1d1cN\\N]1c1cN^N]1b1cN^N\\1c1dN]N\\1b1dN_N\\1a1dN_N\\1`1dNaN\\1_1dNaN\\1^1eNbN[1^1dNcN\\1\\1eNcN\\1]1cNdN]1[1dNeN\\1[1dNeN\\1Z1dNgN\\1Y1dNgN\\1X1dNiN[1X1eNhNm0e1SO\\Nm0d1SO\\Nl0d1UO\\Nk0d1TO]Nl0b1UO^Nk0b1TO_Nk0a1VO^Nk0b1UO^Nk0a1UO`Nj0a1VO_Nj0`1VOaNj0_1VOaNj0_1UObNj0_1VOaNj0_1VOaNj0_1UObNj0_1VOaNj0_1UObNk0^1UObNk0^1UObNj0_1UOaNl0_1TOaNl0_1SOcNk0^1UObNk0^1UObNk0^1TOcNl0]1TOcNk0^1TOdNk0\\1UOdNk0\\1TOeNk0\\1UOdNk0\\1UOeNj0[1UOfNk0Z1UOfNj0[1UOfNk0Z1UOfNk0Z1UOgNi0Z1VOgNj0Y1VOgNj0Y1UOhNk0X1UOiNi0X1WOhNi0X1VOiNj0W1VOiNi0X1VOiNj0W1VOjNi0V1VOkNj0U1VOkNi0V1WOjNi0V1VOlNi0T1WOlNh0U1WOlNi0T1WOlNi0T1WOlNi0T1VOnNh0S1XOmNh0S1WOnNi0R1WOnNh0S1XOnNg0R1XOoNh0Q1XOoNh0Q1WOPOh0Q1XOoNh0Q1WOQOh0o0XOQOg0P1YOPOg0P1XOROg0n0YOROg0n0XOSOg0n0YOSOf0m0ZOSOf0m0YOUOf0k0ZOUOe0l0ZOVOe0j0[OVOe0j0[OVOd0k0[OWOd0i0\\OWOd0i0[OYOd0g0\\OYOc0h0\\OZOc0f0]OZOc0f0]O[Oa0f0^O[Ob0e0^O[Ob0e0]O]Ob0c0^O]Oa0d0_O]O`0c0_O^Oa0b0_O_O?b0@@?`0AA>?BA>?AC=>CC<=CD=<CE;<DF;:EF;:EG:9EI98GI87GJ78II49LH1:NG0;0FM<2EM<3CL?4AJa05@Ib07^OGd08\\OGf09ZOEh0:YODi0<WOBk0>TOAn0>SO@o0`0QO^OQ1a0oN_OR1a0nN]OT1c0lN[OV1d0kNZOW1f0hNYOZ1f0gNXO[1h0eNXO[1h0eNXO[1g0eNZO[1f0eNYO\\1f0eNZO[1f0dN[O\\1d0eN\\O[1d0eN\\O[1d0eN\\O[1c0eN^O[1b0eN^O[1a0fN_OZ1a0fN_OZ1a0eN@[1?fNAZ1?fNAZ1>gNAZ1?eNB[1>eNB[1=fNC[1<eND[1;fNEZ1;fNEZ1:fNGZ19fNGZ19fNGZ18gNHY18gNHY17hNIX17gNIZ17fNIZ16gNJY16gNJY15hNKX15hNKX15gNLY13hNMX13hNMX12iNNW12iNNW11jNOV11jNOV11iNOX10iN0W10iN0W1OjN1V1OjN1V1OjN1V1NjN3V1MjN3V1LkN4U1LkN4U1LkN4U1KlN5T1KkN6U1IlN6U1JkN6U1IlN7T1IlN7T1IlN7T1HmN8S1HlN9T1FmN:S1FmN:S1FmN:S1EnN;R1EnN;R1DnN=R1CnN<S1DmN<S1CnN=R1CnN=R1BoN>Q1BnN?R1@oN`0Q1@oN`0Q1@oN`0Q1_OPOa0P1_OPOa0P1^OQOb0o0^OQOb0o0^OQOa0P1^OQOb0o0^OQOb0o0]OROc0n0]OROc0n0]OROc0n0\\OSOd0m0\\OSOd0m0[OTOe0l0[OTOe0l0ZOUOf0k0ZOUOf0k0ZOVOd0k0[OVOe0j0[OVOe0j0ZOWOf0i0ZOWOf0i0ZOWOf0i0YOXOg0e0"}, "label": "a plane called the overnight delivery system"}]}
{"id": 212863, "image": "COCO_train2014_000000212863.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane\"."}], "task": "refering", "answer_template": "The \"an airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.003015625, 0.004541176470588235, 1.0, 0.8027294117647059], "iscrowd": 0, "area": 107641.54865, "rle": {"size": [425, 640], "counts": "To0Z1o;00001O01O000001O0001O0001O0001O0001O000001O0001O000001O01O000001O01O000000010O000000010O00000000010O0000000010O000000010OcDjNg:V1XEoNd:Q1[ETOa:l0^EZO]:g0aEZO_:f0_E\\Ob:c0]E_Ob:a0]E@c:`0\\EAd:?[EBe:?YEBg:>XECi:<VEFi::VEGj:\\100001O01hGcMb5]2[JgMd5Y2ZJjMe5V2XJnMg5R2WJQNi5n1UJUNj5k1SJYNl5h1QJ[Nn5e1VHgM`1f0Y6n1fITNZ6k1eIWNZ6i1fIXNY6h1fIYNZ6g1eI[NZ6e1eI]NZ6c1fI]N[6b1fI]NZ6c1gI\\NY6d1hI[NX6e1iIZNW6f1jIYNV6g1kIXNV6g1kIYNT6U2_IjMa6T3bHkL^7`4001O00O101N1O100O1O100O101N1O100O10iN_HfL`7R3iHnLW7i2SIVMl6d2\\I\\Mc6b2_I^M`6a2cI^M\\6a2fI_MZ6_2nH_L>Q1c6_2mHfL?j0d6^2lHlL?f0d6]2kHSM`0`0d6[2kHZM`0:d6[2jH`Ma05e6Y2iHgMa0Oe6`3ZI_Le6d3YI\\Lg6e3YIXLh6k3YIoKj6S4XIfKj6[4ZI^Ki6c4ZIVKh6k4k01N101N1O2O0O2O0O2O0O2O0O2O0O2N101N101N101N3N3L4M2M2N2O0O2O0O2O0O2O0O2O0O1OO2N1O2N1O2N1O2O02O0O2N101N100O2O0O2O0O1O2O0O2O0OO2O001N101O0002O0O100O100O100O100O2O0O100O100O100O100O101N100O100O100O100O100O2O0O100O100O10PLlIk0S6VOmIj0S6UOnIk0Q6VOPJi0P6WOPJj0n5VOSJj0m5VOTJi0k5WOVJi0j5WOWJh0i5WOXJi0g5XOYJh0g5XOZJg0e5YO\\Jg0d5YO\\Jg0c5YO_Jf0a5ZO_Jf0`5[O`Jf0_5YOcJf0\\5[OdJe0\\5ZOfJe0Y5\\OgJd0Y5\\OgJd0X5\\OjJc0V5]OjJc0U5]OlJc0T5]OmJb0R5^OoJb0Q5^OoJc0o4^ORKa0n4^OSKb0l4_OUK`0k4_OVKa0i4@WK`0i4@XK?g4AZK?f4AZK?e4A]K>c4B]K>b4C^K>a4AaK>^4CbK=^4BdK=[4DeK<[4CfK=Y4DgK<Y4DgK<X4DhK=X4ChK=W4CjK=V4CjK=U4DjK>U4AlK?S4BmK>S4AmK`0R4AnK?R4AnK?Q4AoK`0Q4@oK`0P4@QL`0o3@PLa0o3_ORLa0n3_ORLb0l3_OSLb0m3]OTLc0k3^OULb0k3]OULd0j3]OVLc0j3]OVLc0i3]OWLd0i3\\OWLd0h3\\OYLd0g3\\OXLe0g3\\OYLd0g3[OZLf0d3[O[Lf0e3YO\\Lg0c3ZO]Lf0c3YO]Lh0c3XO]Lh0b3YO^Lg0b3XO^Li0a3XO_Lh0a3WO`Li0_3XO`Li0`3WO`Lj0^3VOcLj0]3VObLk0]3UOdLk0\\3UOdLk0[3VOdLk0\\3TOeLl0Z3UOfLk0Z3TOgLl0X3UOhLk0X3TOhLn0V3SOjLm0V3SOjLm0U3SOlLm0T3SOlLm0S3SOnLm0R3SOnLm0Q3TOnLm0R3ROoLn0P3SOPMm0P3ROQMo0m2ROSMn0m2ROSMn0l2ROUMn0k2ROTMo0k2QOVMo0j2QOVMo0i2QOXMo0h2QOXMo0g2ROYMn0g2QOZMP1d2QO[MP1e2oN\\MQ1c2PO]MP1c2PO]MP1b2PO_MP1a2PO_MP1kNeKY3[3lMQ1jNeKZ3Z3lMQ1hNgK[3Y3mMP1gNhK\\3W3nMR1cNiK^3V3oMQ1bNjK_3T3PN]1o1dNQN\\1o1cNQN^1n1cNRN]1n1cNRN]1m1cNTN]1l1cNTN\\1l1dNUN\\1k1dNUN\\1j1eNUN\\1k1cNVN]1i1dNWN\\1i1cNXN]1g1dNYN\\1g1dNYN\\1f1dN[N\\1e1dNZN]1e1cN\\N]1d1cN\\N]1c1cN^N]1b1cN^N\\1c1dN]N\\1b1dN_N\\1a1dN_N\\1`1dNaN\\1_1dNaN\\1^1eNbN[1^1dNcN\\1\\1eNcN\\1]1cNdN]1[1dNeN\\1[1dNeN\\1Z1dNgN\\1Y1dNgN\\1X1dNiN[1X1eNhNm0e1SO\\Nm0d1SO\\Nl0d1UO\\Nk0d1TO]Nl0b1UO^Nk0b1TO_Nk0a1VO^Nk0b1UO^Nk0a1UO`Nj0a1VO_Nj0`1VOaNj0_1VOaNj0_1UObNj0_1VOaNj0_1VOaNj0_1UObNj0_1VOaNj0_1UObNk0^1UObNk0^1UObNj0_1UOaNl0_1TOaNl0_1SOcNk0^1UObNk0^1UObNk0^1TOcNl0]1TOcNk0^1TOdNk0\\1UOdNk0\\1TOeNk0\\1UOdNk0\\1UOeNj0[1UOfNk0Z1UOfNj0[1UOfNk0Z1UOfNk0Z1UOgNi0Z1VOgNj0Y1VOgNj0Y1UOhNk0X1UOiNi0X1WOhNi0X1VOiNj0W1VOiNi0X1VOiNj0W1VOjNi0V1VOkNj0U1VOkNi0V1WOjNi0V1VOlNi0T1WOlNh0U1WOlNi0T1WOlNi0T1WOlNi0T1VOnNh0S1XOmNh0S1WOnNi0R1WOnNh0S1XOnNg0R1XOoNh0Q1XOoNh0Q1WOPOh0Q1XOoNh0Q1WOQOh0o0XOQOg0P1YOPOg0P1XOROg0n0YOROg0n0XOSOg0n0YOSOf0m0ZOSOf0m0YOUOf0k0ZOUOe0l0ZOVOe0j0[OVOe0j0[OVOd0k0[OWOd0i0\\OWOd0i0[OYOd0g0\\OYOc0h0\\OZOc0f0]OZOc0f0]O[Oa0f0^O[Ob0e0^O[Ob0e0]O]Ob0c0^O]Oa0d0_O]O`0c0_O^Oa0b0_O_O?b0@@?`0AA>?BA>?AC=>CC<=CD=<CE;<DF;:EF;:EG:9EI98GI87GJ78II49LH1:NG0;0FM<2EM<3CL?4AJa05@Ib07^OGd08\\OGf09ZOEh0:YODi0<WOBk0>TOAn0>SO@o0`0QO^OQ1a0oN_OR1a0nN]OT1c0lN[OV1d0kNZOW1f0hNYOZ1f0gNXO[1h0eNXO[1h0eNXO[1g0eNZO[1f0eNYO\\1f0eNZO[1f0dN[O\\1d0eN\\O[1d0eN\\O[1d0eN\\O[1c0eN^O[1b0eN^O[1a0fN_OZ1a0fN_OZ1a0eN@[1?fNAZ1?fNAZ1>gNAZ1?eNB[1>eNB[1=fNC[1<eND[1;fNEZ1;fNEZ1:fNGZ19fNGZ19fNGZ18gNHY18gNHY17hNIX17gNIZ17fNIZ16gNJY16gNJY15hNKX15hNKX15gNLY13hNMX13hNMX12iNNW12iNNW11jNOV11jNOV11iNOX10iN0W10iN0W1OjN1V1OjN1V1OjN1V1NjN3V1MjN3V1LkN4U1LkN4U1LkN4U1KlN5T1KkN6U1IlN6U1JkN6U1IlN7T1IlN7T1IlN7T1HmN8S1HlN9T1FmN:S1FmN:S1FmN:S1EnN;R1EnN;R1DnN=R1CnN<S1DmN<S1CnN=R1CnN=R1BoN>Q1BnN?R1@oN`0Q1@oN`0Q1@oN`0Q1_OPOa0P1_OPOa0P1^OQOb0o0^OQOb0o0^OQOa0P1^OQOb0o0^OQOb0o0]OROc0n0]OROc0n0]OROc0n0\\OSOd0m0\\OSOd0m0[OTOe0l0[OTOe0l0ZOUOf0k0ZOUOf0k0ZOVOd0k0[OVOe0j0[OVOe0j0ZOWOf0i0ZOWOf0i0ZOWOf0i0YOXOg0e0"}, "label": "an airplane"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back end of the blue and white truck\"."}], "task": "refering", "answer_template": "The \"the back end of the blue and white truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.5116458333333334, 0.389734375, 1.0], "iscrowd": 0, "area": 55069.20769999999, "rle": {"size": [480, 640], "counts": "h7R5n9U2kM000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001O00000000001O01O00000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000>B`1`N`1`NU1kNo0QOl0TORjf5"}, "label": "the back end of the blue and white truck"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and white ambulance\"."}], "task": "refering", "answer_template": "The \"blue and white ambulance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.5116458333333334, 0.389734375, 1.0], "iscrowd": 0, "area": 55069.20769999999, "rle": {"size": [480, 640], "counts": "h7R5n9U2kM000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001O00000000001O01O00000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000>B`1`N`1`NU1kNo0QOl0TORjf5"}, "label": "blue and white ambulance"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ladder that is raised the tallest\"."}], "task": "refering", "answer_template": "The \"the ladder that is raised the tallest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 29, 30, 53, 54, 76, 77, 100, 101, 123, 124, 147, 148, 241, 264, 265, 287, 288, 310, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.27703125, 0.0011250000000000001, 0.6655468750000001, 0.8502291666666667], "iscrowd": 0, "area": 7006.673849999996, "rle": {"size": [480, 640], "counts": "Q^c24l>3M2N2N3N1N2N2N2N2N2N2O1N2000000000O10O10001O1O1O1O1O001O1O1O1O1OO10O10000000O10000000000O10000000O10O10000000000O10000000000O1000001O0O1010O1O100O1O10O01O10N1O1O2N1O100O1O2N1O1O100N2O1N2O2M2O1O1N2O1N2O1O1N101N101O0O2Oe_90Z`F1O2N2N1O2N2N1N3N1O2N2N1M4K5L4K5K6J6J5M400O001oBPOo;Q200010O000N3M2N2O2M2N2N3M2N2N3M2M3N3M2M3N3M2M2O2N2M3N2N2N2N2N2N1O2N2N2N2N2N2N00000000000000000000000000O1000000000000002M3N1O00000000000000001O00000000000000000000001O00000000001O00001O0JmmS3"}, "label": "the ladder that is raised the tallest"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the crane arm going straight up without the person in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the crane arm going straight up without the person in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 29, 30, 53, 54, 76, 77, 100, 101, 123, 124, 147, 148, 241, 264, 265, 287, 288, 310, 311, 312, 313, 314, 334, 335, 336, 337], "bbox": [0.27703125, 0.0011250000000000001, 0.6655468750000001, 0.8502291666666667], "iscrowd": 0, "area": 7006.673849999996, "rle": {"size": [480, 640], "counts": "Q^c24l>3M2N2N3N1N2N2N2N2N2N2O1N2000000000O10O10001O1O1O1O1O001O1O1O1O1OO10O10000000O10000000000O10000000O10O10000000000O10000000000O1000001O0O1010O1O100O1O10O01O10N1O1O2N1O100O1O2N1O1O100N2O1N2O2M2O1O1N2O1N2O1O1N101N101O0O2Oe_90Z`F1O2N2N1O2N2N1N3N1O2N2N1M4K5L4K5K6J6J5M400O001oBPOo;Q200010O000N3M2N2O2M2N2N3M2N2N3M2M3N3M2M3N3M2M2O2N2M3N2N2N2N2N2N1O2N2N2N2N2N2N00000000000000000000000000O1000000000000002M3N1O00000000000000001O00000000000000000000001O00000000001O00001O0JmmS3"}, "label": "the crane arm going straight up without the person in the right hand picture"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fire truck platform\"."}], "task": "refering", "answer_template": "The \"a fire truck platform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 96, 97, 98, 99, 119, 120, 121, 122, 123, 144, 145, 146, 147, 148, 169, 170, 171, 172, 194, 195, 196, 218, 219, 220, 221, 242, 243, 244, 245, 248, 249, 250, 251, 252, 266, 267, 268, 269, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 367, 390], "bbox": [0.174078125, 0.20493750000000002, 1.0, 0.950625], "iscrowd": 0, "area": 36011.93174999999, "rle": {"size": [480, 640], "counts": "ZUd11i>7I8H7H9H8I6O2N2N1O2M2O2N000000000000000O2O00000000000000001N100000001O001O1O1O1N200O1O1O010O0001O0001O01O00001O01O00O2N100O2N100O1O2O0O10100O2O001O1O0O2O1OO10O100000O10O100000O0100000O010000000O01000000O0100000O010000000O0100000O10O10O100O10O010000O010O10000O010O10000O010O100O01000O100O01000O100O010O10000O010O1000O0100O010O1000O0100O01000O10O0100O10O10O10O010000O010O1O00100O1O001O100O001O10O01O1O1O010O1O001O100O001O1004K6K4L5J4MO1000O10O10O1000O10O1000O10O1000O0100000O010000O0100000O01000O10O1000O0100000O0100000O0100000O0100000O0100000O0100000O0100000O010000O0100000O0100000O0100000O0100000O10O1000O10O1000O10O1000O10O1000O10O10O1000O10O1000O10O1000O10O1000O10O1001N2O1O2N1N2O1O2N1N2O1O1O2N1N2O1O2N1N2O1O1OO001O1O1O001O1O1O001O1O1O00100O1O001O1O1O001O1O1O1O1O1O2N1O1O100O2N1O1O000000O1J6J6K5J6J6K5J6J6N2N2N2N2N2N2N2N2N3M2O10000000000O10000000000000000000000000000000000O10001O0000000000000000000000000000000O1000O1K5K5L4O10000000000000000001O000O100000000000001O00000000001O1O1O2N1O1N2O2N1O1O1O2N1O2N2N3M3M2N3M2N3M2N3M3M2M4M2N3M2N3M000000O1000000OR3oLPD"}, "label": "a fire truck platform"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck says number 9\"."}], "task": "refering", "answer_template": "The \"truck says number 9\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.38528125, 0.8324583333333333, 0.8971406249999999, 1.0], "iscrowd": 0, "area": 19795.98895, "rle": {"size": [480, 640], "counts": "SVd3m1S=00000000000O100000000000000000000000000000000000000000000000O01000000000000000000000000000000000000000000O1000000000000000000000000O100000000000000000000000000O100000000O1O100O100O1O10O01O100O100000000O1000000000000O10000000000O1000000000000O10000000000000000000000O100000000000000000000000000O01N2O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O2N1O100000000001O01O00000001O00000001O01O00000001O01O1O2N2N2N2O0O2N2N2N2N0001O00000000001O0000000000010O0000000000001O00000000000010O000000000001O000O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1000000O100000000O10000000000O100000000O2O0000000O10000O1O101N1O1O100O1O1O100O1O2N10000O100O10000O100O2O000O10000O100O100HlADU>8PBEQ>8TBFl=7`0M3MSan0"}, "label": "truck says number 9"}]}
{"id": 507776, "image": "COCO_train2014_000000507776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red vehicle with a large number 9 on the window\"."}], "task": "refering", "answer_template": "The \"the red vehicle with a large number 9 on the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.38528125, 0.8324583333333333, 0.8971406249999999, 1.0], "iscrowd": 0, "area": 19795.98895, "rle": {"size": [480, 640], "counts": "SVd3m1S=00000000000O100000000000000000000000000000000000000000000000O01000000000000000000000000000000000000000000O1000000000000000000000000O100000000000000000000000000O100000000O1O100O100O1O10O01O100O100000000O1000000000000O10000000000O1000000000000O10000000000000000000000O100000000000000000000000000O01N2O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O2N1O100000000001O01O00000001O00000001O01O00000001O01O1O2N2N2N2O0O2N2N2N2N0001O00000000001O0000000000010O0000000000001O00000000000010O000000000001O000O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1000000O100000000O10000000000O100000000O2O0000000O10000O1O101N1O1O100O1O1O100O1O2N10000O100O10000O100O2O000O10000O100O100HlADU>8PBEQ>8TBFl=7`0M3MSan0"}, "label": "the red vehicle with a large number 9 on the window"}]}
{"id": 393095, "image": "COCO_train2014_000000393095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person standing in the background wearing a black shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"a person standing in the background wearing a black shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 110, 111, 112, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 270, 271, 272, 294, 295, 317, 318, 340, 341], "bbox": [0.754140625, 0.22023419203747074, 0.87859375, 0.9797658079625293], "iscrowd": 0, "area": 16915.865599999997, "rle": {"size": [427, 640], "counts": "UdY65R=5K4J7UDJl9j0QFE`9>]FH]9;`FK[96cFOX98aFNZ9V2J6K5J6K4K6J6K5J6K5J6K5K4L5K5L4K5K4L4L4L3M4L4M3L4L4L3M4kN]H[Lg7_3aH[Lc7`3cH[La7`3cH^L`7\\3eHaL^7[3fHbL\\7[3hHaLZ7]3jH^LY7`3Z1N2O100O1O100O1O100O1OO2G9G9G9G9G9G8H9G9H8G9K5L4L4L5K4L5K4Li0WO9C>hNW1iNg]P1"}, "label": "a person standing in the background wearing a black shirt and blue jeans"}]}
{"id": 393095, "image": "COCO_train2014_000000393095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with black shirt leaning against the wall in the background\"."}], "task": "refering", "answer_template": "The \"man with black shirt leaning against the wall in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 110, 111, 112, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 270, 271, 272, 294, 295, 317, 318, 340, 341], "bbox": [0.754140625, 0.22023419203747074, 0.87859375, 0.9797658079625293], "iscrowd": 0, "area": 16915.865599999997, "rle": {"size": [427, 640], "counts": "UdY65R=5K4J7UDJl9j0QFE`9>]FH]9;`FK[96cFOX98aFNZ9V2J6K5J6K4K6J6K5J6K5J6K5K4L5K5L4K5K4L4L4L3M4L4M3L4L4L3M4kN]H[Lg7_3aH[Lc7`3cH[La7`3cH^L`7\\3eHaL^7[3fHbL\\7[3hHaLZ7]3jH^LY7`3Z1N2O100O1O100O1O100O1OO2G9G9G9G9G9G8H9G9H8G9K5L4L4L5K4L5K4Li0WO9C>hNW1iNg]P1"}, "label": "man with black shirt leaning against the wall in the background"}]}
{"id": 393095, "image": "COCO_train2014_000000393095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man makes an odd face as he bites into a glazed donut\"."}], "task": "refering", "answer_template": "The \"a man makes an odd face as he bites into a glazed donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.1424375, 0.06531615925058548, 0.8081250000000001, 0.9844262295081967], "iscrowd": 0, "area": 96471.74089999998, "rle": {"size": [427, 640], "counts": "U[V12T=5K5J6K5K5K5K5J6K5J6J6J6I7J6J6L4L4M4K4M3L4M3L4M3O1O1O1O1N2O1O1O1O1O1N2O1O1O2N1O1N2O1O1O1O1O1N2O1O100O100O100O100O101N100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O100O100jKiJ7X5FVKOk4LdKH]45RL_On3>`LVOa3g0mLnNS3o0TMlNm2Q1WMnNi2o0\\MoNd2n0`MPOa2l0dMSO\\2j0hMTOY2i0kMVOU2i0mMUOT2k0mMTOS2k0oMSOR2l0PNSOP2l0SNQOn1o0SNPOm1o0UNoNl1P1VNoNj1P1XNnNj1Q1WNnNi1Q1YNmNh1R1ZNmNf1S1[NlNe1S1^NjNc1U1_NjNa1U1aNiN`1W1aNfNa1Y1aNeN`1Z1bNcN`1\\1bNbN_1^1bN_N`1`1bN^N_1a1dN\\N]1d1dNYN^1f1dNXN]1g1eNVN]1i1S500O1O1O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O010O10000000000O100000000O100000000O10000000000O100000000O1000000000000000O100000000000000000000000000000001O00001O00001O00001O00001O00001O00001O000O2O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O002N2N2N2N2N2N2N3bFZNQ7h1eHTOg6g1VHhNc7f3J6J6aImIZ5T7N2N3L2O2N2N2N2N1O2N2N2N1O2N1O1O1O1O001O1O1O1O001O1O1O001O1O1O1O2N2N1O2N2N2N2N1N3N2N2N2N1O2N2N2N2N1O]OVLgGh3Y8_LbG_3^8gL^GW3b8oLZGo2g8VMTGi2l8k000001O0000001O0000001O0000001O0000001O000O101O0000001O00001O001O001O001O001O001O001O002N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2M3N1O2N2N2N1O2N2N2N2N5K5K4L3M2N3M2N2N2N2N2N2N2N4L4L4L5K4L4L8H7I6J]lb1"}, "label": "a man makes an odd face as he bites into a glazed donut"}]}
{"id": 393095, "image": "COCO_train2014_000000393095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black jacket eating some sweets\"."}], "task": "refering", "answer_template": "The \"a man with black jacket eating some sweets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.1424375, 0.06531615925058548, 0.8081250000000001, 0.9844262295081967], "iscrowd": 0, "area": 96471.74089999998, "rle": {"size": [427, 640], "counts": "U[V12T=5K5J6K5K5K5K5J6K5J6J6J6I7J6J6L4L4M4K4M3L4M3L4M3O1O1O1O1N2O1O1O1O1O1N2O1O1O2N1O1N2O1O1O1O1O1N2O1O100O100O100O100O101N100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O100O100jKiJ7X5FVKOk4LdKH]45RL_On3>`LVOa3g0mLnNS3o0TMlNm2Q1WMnNi2o0\\MoNd2n0`MPOa2l0dMSO\\2j0hMTOY2i0kMVOU2i0mMUOT2k0mMTOS2k0oMSOR2l0PNSOP2l0SNQOn1o0SNPOm1o0UNoNl1P1VNoNj1P1XNnNj1Q1WNnNi1Q1YNmNh1R1ZNmNf1S1[NlNe1S1^NjNc1U1_NjNa1U1aNiN`1W1aNfNa1Y1aNeN`1Z1bNcN`1\\1bNbN_1^1bN_N`1`1bN^N_1a1dN\\N]1d1dNYN^1f1dNXN]1g1eNVN]1i1S500O1O1O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O010O10000000000O100000000O100000000O10000000000O100000000O1000000000000000O100000000000000000000000000000001O00001O00001O00001O00001O00001O00001O000O2O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O002N2N2N2N2N2N2N3bFZNQ7h1eHTOg6g1VHhNc7f3J6J6aImIZ5T7N2N3L2O2N2N2N2N1O2N2N2N1O2N1O1O1O1O001O1O1O1O001O1O1O001O1O1O1O2N2N1O2N2N2N2N1N3N2N2N2N1O2N2N2N2N1O]OVLgGh3Y8_LbG_3^8gL^GW3b8oLZGo2g8VMTGi2l8k000001O0000001O0000001O0000001O0000001O000O101O0000001O00001O001O001O001O001O001O001O002N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2M3N1O2N2N2N1O2N2N2N2N5K5K4L3M2N3M2N2N2N2N2N2N2N4L4L4L5K4L4L8H7I6J]lb1"}, "label": "a man with black jacket eating some sweets"}]}
{"id": 204680, "image": "COCO_train2014_000000204680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow tray with different foods\"."}], "task": "refering", "answer_template": "The \"a yellow tray with different foods\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 220, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.17987999999999998, 0.6052266666666667, 0.71454, 0.96168], "iscrowd": 0, "area": 31185.6968, "rle": {"size": [375, 500], "counts": "iWQ15^;;F;E:E;F8G5M2O2N2M3N2N2N2N2M3N2N2N2N2N2N1O2N2N2N2N2N2O1N2N2N2N2N2O1N101O0O100O10000O2O000O100O10000O100O10000O100O2O0000000000000000000000000O2O000000000000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O00001O0000001O000O2O0O101N100O101N100O2O0O101N101N100O2N1O2O0O1O2N101N1O101N2N2N3N2M2N3M3N1N2N2O1N2N101N5J5K6J5K7I7I6J7I:F`Rd1"}, "label": "a yellow tray with different foods"}]}
{"id": 204680, "image": "COCO_train2014_000000204680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"thebox of food nearest to the camers\"."}], "task": "refering", "answer_template": "The \"thebox of food nearest to the camers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 220, 221, 222, 223, 224, 225, 226, 227], "bbox": [0.17987999999999998, 0.6052266666666667, 0.71454, 0.96168], "iscrowd": 0, "area": 31185.6968, "rle": {"size": [375, 500], "counts": "iWQ15^;;F;E:E;F8G5M2O2N2M3N2N2N2N2M3N2N2N2N2N2N1O2N2N2N2N2N2O1N2N2N2N2N2O1N101O0O100O10000O2O000O100O10000O100O10000O100O2O0000000000000000000000000O2O000000000000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000001O00001O0000001O000O2O0O101N100O101N100O2O0O101N101N100O2N1O2O0O1O2N101N1O101N2N2N3N2M2N3M3N1N2N2O1N2N101N5J5K6J5K7I7I6J7I:F`Rd1"}, "label": "thebox of food nearest to the camers"}]}
{"id": 491400, "image": "COCO_train2014_000000491400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt holding a umbrella\"."}], "task": "refering", "answer_template": "The \"man in blue shirt holding a umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 421, 423, 424, 441, 444, 446, 447, 448, 464, 467, 468, 469, 470, 471, 492], "bbox": [0.1763125, 0.30653125000000003, 0.7455937499999999, 0.9208124999999999], "iscrowd": 0, "area": 49343.97564999999, "rle": {"size": [640, 640], "counts": "_UW23jc03N3L4L3O2O1O001O1O1O001N2O00Em\\OJRc05V]OEib0;>0000O1000O11N3N2N2N2MkSj0h0SkUO;N2N2N3M2N2N2N3M2M3N2N3c^OWN_`0k1a_OjNh?X1X@\\OR?f0n@O[>4dAa0e=A[BS1o<oNQCV1h<lNXCS1f<POZCn0f<TOZCj0f<XOZCg0e<[OZCd0h<\\OXCb0j<^OVC`0l<@TC?m<ASC=o<CQC;P=FPC9Q=GoBFd=:\\BQOY>o0gAbMh?^2X@bMh?^2X@bMg?_2Y@aMg?_2Y@aMg?_2Y@bMf?]2[@cMe?]2[@cMe?]2[@dMc?]2]@cMc?]2]@dMb?\\2^@eMa?[2_@eM\\2CT:h2`CfMZ2DV:e2_CiMZ2BW:e2_CiMY2CX:d2_CjMX2BY:d2_CkMV2CZ:b2`CkMV2CZ:b2`ClMT2C\\:a2`CmMR2C_:_2_CoMQ2B`:_2_CoMP2D`:]2`CPNo1Ca:]2`CQNm1Cc:\\2`CQNm1Cc:\\2`CRNk1Dd:Z2aCSNi1Df:Y2aCSNi1Dg:X2`CUNg1Di:W2`CVNf1Di:V2aCVNe1Ej:U2`CXNe1Ck:U2`CYNc1Cm:T2`CYNb1Em:R2aCZNa1Do:Q2`C\\N_1DQ;P2`C\\N_1DQ;P2`C]N]1DS;o1`C_N[1CT;n1aC`NY1CV;m1aCbNW1AX;m1aCdNT1@\\;k1`CfNR1A];i1aChNP1_O_;i1aCiNn0_Oa;h1aCkNl0]Oc;h1aCmNi0]Oe;f1bCnNh0\\Of;f1bCPOe0[Oi;e1cCQOa0[Om;c1dCQO>\\On;c1eCRO:\\OQ<b1fCSO7ZOT<c1gCRO3\\OV<b1hCSO0[OW<c1jCROM[OZ<c1kCROH\\O]<b1lCSOEZO`<c1lCSOB[Ob<b1nCSO^O[Od<b1oCTOZOZOh<b1oCTOXOZOi<b1QDTOSO[Ol<a1RDTOPO[Oo<a1RDUOmNZOQ=a1TDUOhNZOU=a1TD8k;IVD6j;JXD3i;MXD2h;NYD1g;O[DNf;2[DMe;3\\DKe;5]DIc;6_DIa;7`DGa;9aDE_;;bDC^;>cDA];?eD_O[;a0fD]O[;d0eD[O[;e0gDYOX;h0iDVOX;j0iDUOW;l0jDQOW;o0jDPOV;P1kDoNT;R1nDkNS;V1oDgNQ;Y1QEeNo:[1SEbNn:^1TE`Nk:b1VE[Nk:e1XEXNh:h1ZEVNf:j1\\ESNe:n1\\EPN^7WOjHj2KlMX7@jHd20iMU7GiHa23fMR7MiH]27cMn66hHW2=`Mi6=gHT2b0]Md6e0gHo1f0YMb6l0fHk1j0WM^6S1eHf1o0TMZ6[1dHb1U1PMW6`1bHa1m:bNPE^1o:eNoD[1Q;hNlDX1T;jNiDX1V;jNhDV1W;mNfDT1Z;oNbDR1^;PO`DQ1^;RO_Do0a;SO\\Dn0g;ROUDP1m;oNQDQ1Q<oNlCR1W<mNfCT1]<lN_CV1b<jN\\CV1g<iNVCX1m<gNPCZ1S=eNjB]1W=dNfB\\1]=cN`B^1c=aNZB`1h=`NUBb1m=^NPBb1S>T22N3N2M2N3N2M2N3MmLiJ`GU5_8mJbGP5^8RKbGl4^8VKcGg4]8[KdGc4[8_KeG_4[8cKeG[4[8gKaG[4_8gK\\G]4c8eKXG^4h8dKTG^4l8dKoF_4Q9cKjFa4U9bKfF`4Z9cK`F`4`9dKYF_4g9dKSF_4m9eKmE]4S:gKfE[4[:gK`E\\4`:dK^E]4c:cKZE`4f:`KWEb4j:_KREd4n:\\KPEe4Q;[KlDh4T;XKiDj4X;WKeDk4[;UKbDm4`;RK]DQ5c;oJZDS5g;c10O100O10000O100O100O100O100O100O100O100O100O100O100O2O1N2O1N3N1N2O1N3N1N2O1O2M2O2M2O2M3N1N3N2M3N1N3N2M2O2M2O2M2N3K4L5K4L5K4M4K4L5K4L5K4L5K4L5K4L7I9G9H8G:F9G9G9G9G9G:F9G9G]aU3"}, "label": "man in blue shirt holding a umbrella"}]}
{"id": 491400, "image": "COCO_train2014_000000491400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man holding blue umbrella\"."}], "task": "refering", "answer_template": "The \"the man holding blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 421, 423, 424, 441, 444, 446, 447, 448, 464, 467, 468, 469, 470, 471, 492], "bbox": [0.1763125, 0.30653125000000003, 0.7455937499999999, 0.9208124999999999], "iscrowd": 0, "area": 49343.97564999999, "rle": {"size": [640, 640], "counts": "_UW23jc03N3L4L3O2O1O001O1O1O001N2O00Em\\OJRc05V]OEib0;>0000O1000O11N3N2N2N2MkSj0h0SkUO;N2N2N3M2N2N2N3M2M3N2N3c^OWN_`0k1a_OjNh?X1X@\\OR?f0n@O[>4dAa0e=A[BS1o<oNQCV1h<lNXCS1f<POZCn0f<TOZCj0f<XOZCg0e<[OZCd0h<\\OXCb0j<^OVC`0l<@TC?m<ASC=o<CQC;P=FPC9Q=GoBFd=:\\BQOY>o0gAbMh?^2X@bMh?^2X@bMg?_2Y@aMg?_2Y@aMg?_2Y@bMf?]2[@cMe?]2[@cMe?]2[@dMc?]2]@cMc?]2]@dMb?\\2^@eMa?[2_@eM\\2CT:h2`CfMZ2DV:e2_CiMZ2BW:e2_CiMY2CX:d2_CjMX2BY:d2_CkMV2CZ:b2`CkMV2CZ:b2`ClMT2C\\:a2`CmMR2C_:_2_CoMQ2B`:_2_CoMP2D`:]2`CPNo1Ca:]2`CQNm1Cc:\\2`CQNm1Cc:\\2`CRNk1Dd:Z2aCSNi1Df:Y2aCSNi1Dg:X2`CUNg1Di:W2`CVNf1Di:V2aCVNe1Ej:U2`CXNe1Ck:U2`CYNc1Cm:T2`CYNb1Em:R2aCZNa1Do:Q2`C\\N_1DQ;P2`C\\N_1DQ;P2`C]N]1DS;o1`C_N[1CT;n1aC`NY1CV;m1aCbNW1AX;m1aCdNT1@\\;k1`CfNR1A];i1aChNP1_O_;i1aCiNn0_Oa;h1aCkNl0]Oc;h1aCmNi0]Oe;f1bCnNh0\\Of;f1bCPOe0[Oi;e1cCQOa0[Om;c1dCQO>\\On;c1eCRO:\\OQ<b1fCSO7ZOT<c1gCRO3\\OV<b1hCSO0[OW<c1jCROM[OZ<c1kCROH\\O]<b1lCSOEZO`<c1lCSOB[Ob<b1nCSO^O[Od<b1oCTOZOZOh<b1oCTOXOZOi<b1QDTOSO[Ol<a1RDTOPO[Oo<a1RDUOmNZOQ=a1TDUOhNZOU=a1TD8k;IVD6j;JXD3i;MXD2h;NYD1g;O[DNf;2[DMe;3\\DKe;5]DIc;6_DIa;7`DGa;9aDE_;;bDC^;>cDA];?eD_O[;a0fD]O[;d0eD[O[;e0gDYOX;h0iDVOX;j0iDUOW;l0jDQOW;o0jDPOV;P1kDoNT;R1nDkNS;V1oDgNQ;Y1QEeNo:[1SEbNn:^1TE`Nk:b1VE[Nk:e1XEXNh:h1ZEVNf:j1\\ESNe:n1\\EPN^7WOjHj2KlMX7@jHd20iMU7GiHa23fMR7MiH]27cMn66hHW2=`Mi6=gHT2b0]Md6e0gHo1f0YMb6l0fHk1j0WM^6S1eHf1o0TMZ6[1dHb1U1PMW6`1bHa1m:bNPE^1o:eNoD[1Q;hNlDX1T;jNiDX1V;jNhDV1W;mNfDT1Z;oNbDR1^;PO`DQ1^;RO_Do0a;SO\\Dn0g;ROUDP1m;oNQDQ1Q<oNlCR1W<mNfCT1]<lN_CV1b<jN\\CV1g<iNVCX1m<gNPCZ1S=eNjB]1W=dNfB\\1]=cN`B^1c=aNZB`1h=`NUBb1m=^NPBb1S>T22N3N2M2N3N2M2N3MmLiJ`GU5_8mJbGP5^8RKbGl4^8VKcGg4]8[KdGc4[8_KeG_4[8cKeG[4[8gKaG[4_8gK\\G]4c8eKXG^4h8dKTG^4l8dKoF_4Q9cKjFa4U9bKfF`4Z9cK`F`4`9dKYF_4g9dKSF_4m9eKmE]4S:gKfE[4[:gK`E\\4`:dK^E]4c:cKZE`4f:`KWEb4j:_KREd4n:\\KPEe4Q;[KlDh4T;XKiDj4X;WKeDk4[;UKbDm4`;RK]DQ5c;oJZDS5g;c10O100O10000O100O100O100O100O100O100O100O100O100O100O2O1N2O1N3N1N2O1N3N1N2O1O2M2O2M2O2M3N1N3N2M3N1N3N2M2O2M2O2M2N3K4L5K4L5K4M4K4L5K4L5K4L5K4L5K4L7I9G9H8G:F9G9G9G9G9G:F9G9G]aU3"}, "label": "the man holding blue umbrella"}]}
{"id": 491400, "image": "COCO_train2014_000000491400.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt\"."}], "task": "refering", "answer_template": "The \"man in blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 375, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 421, 423, 424, 441, 444, 446, 447, 448, 464, 467, 468, 469, 470, 471, 492], "bbox": [0.1763125, 0.30653125000000003, 0.7455937499999999, 0.9208124999999999], "iscrowd": 0, "area": 49343.97564999999, "rle": {"size": [640, 640], "counts": "_UW23jc03N3L4L3O2O1O001O1O1O001N2O00Em\\OJRc05V]OEib0;>0000O1000O11N3N2N2N2MkSj0h0SkUO;N2N2N3M2N2N2N3M2M3N2N3c^OWN_`0k1a_OjNh?X1X@\\OR?f0n@O[>4dAa0e=A[BS1o<oNQCV1h<lNXCS1f<POZCn0f<TOZCj0f<XOZCg0e<[OZCd0h<\\OXCb0j<^OVC`0l<@TC?m<ASC=o<CQC;P=FPC9Q=GoBFd=:\\BQOY>o0gAbMh?^2X@bMh?^2X@bMg?_2Y@aMg?_2Y@aMg?_2Y@bMf?]2[@cMe?]2[@cMe?]2[@dMc?]2]@cMc?]2]@dMb?\\2^@eMa?[2_@eM\\2CT:h2`CfMZ2DV:e2_CiMZ2BW:e2_CiMY2CX:d2_CjMX2BY:d2_CkMV2CZ:b2`CkMV2CZ:b2`ClMT2C\\:a2`CmMR2C_:_2_CoMQ2B`:_2_CoMP2D`:]2`CPNo1Ca:]2`CQNm1Cc:\\2`CQNm1Cc:\\2`CRNk1Dd:Z2aCSNi1Df:Y2aCSNi1Dg:X2`CUNg1Di:W2`CVNf1Di:V2aCVNe1Ej:U2`CXNe1Ck:U2`CYNc1Cm:T2`CYNb1Em:R2aCZNa1Do:Q2`C\\N_1DQ;P2`C\\N_1DQ;P2`C]N]1DS;o1`C_N[1CT;n1aC`NY1CV;m1aCbNW1AX;m1aCdNT1@\\;k1`CfNR1A];i1aChNP1_O_;i1aCiNn0_Oa;h1aCkNl0]Oc;h1aCmNi0]Oe;f1bCnNh0\\Of;f1bCPOe0[Oi;e1cCQOa0[Om;c1dCQO>\\On;c1eCRO:\\OQ<b1fCSO7ZOT<c1gCRO3\\OV<b1hCSO0[OW<c1jCROM[OZ<c1kCROH\\O]<b1lCSOEZO`<c1lCSOB[Ob<b1nCSO^O[Od<b1oCTOZOZOh<b1oCTOXOZOi<b1QDTOSO[Ol<a1RDTOPO[Oo<a1RDUOmNZOQ=a1TDUOhNZOU=a1TD8k;IVD6j;JXD3i;MXD2h;NYD1g;O[DNf;2[DMe;3\\DKe;5]DIc;6_DIa;7`DGa;9aDE_;;bDC^;>cDA];?eD_O[;a0fD]O[;d0eD[O[;e0gDYOX;h0iDVOX;j0iDUOW;l0jDQOW;o0jDPOV;P1kDoNT;R1nDkNS;V1oDgNQ;Y1QEeNo:[1SEbNn:^1TE`Nk:b1VE[Nk:e1XEXNh:h1ZEVNf:j1\\ESNe:n1\\EPN^7WOjHj2KlMX7@jHd20iMU7GiHa23fMR7MiH]27cMn66hHW2=`Mi6=gHT2b0]Md6e0gHo1f0YMb6l0fHk1j0WM^6S1eHf1o0TMZ6[1dHb1U1PMW6`1bHa1m:bNPE^1o:eNoD[1Q;hNlDX1T;jNiDX1V;jNhDV1W;mNfDT1Z;oNbDR1^;PO`DQ1^;RO_Do0a;SO\\Dn0g;ROUDP1m;oNQDQ1Q<oNlCR1W<mNfCT1]<lN_CV1b<jN\\CV1g<iNVCX1m<gNPCZ1S=eNjB]1W=dNfB\\1]=cN`B^1c=aNZB`1h=`NUBb1m=^NPBb1S>T22N3N2M2N3N2M2N3MmLiJ`GU5_8mJbGP5^8RKbGl4^8VKcGg4]8[KdGc4[8_KeG_4[8cKeG[4[8gKaG[4_8gK\\G]4c8eKXG^4h8dKTG^4l8dKoF_4Q9cKjFa4U9bKfF`4Z9cK`F`4`9dKYF_4g9dKSF_4m9eKmE]4S:gKfE[4[:gK`E\\4`:dK^E]4c:cKZE`4f:`KWEb4j:_KREd4n:\\KPEe4Q;[KlDh4T;XKiDj4X;WKeDk4[;UKbDm4`;RK]DQ5c;oJZDS5g;c10O100O10000O100O100O100O100O100O100O100O100O100O100O2O1N2O1N3N1N2O1N3N1N2O1O2M2O2M2O2M3N1N3N2M3N1N3N2M2O2M2O2M2N3K4L5K4L5K4M4K4L5K4L5K4L5K4L5K4L7I9G9H8G:F9G9G9G9G9G:F9G9G]aU3"}, "label": "man in blue shirt"}]}
{"id": 335758, "image": "COCO_train2014_000000335758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with glasses in the grey sirt\"."}], "task": "refering", "answer_template": "The \"the man with glasses in the grey sirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 210, 211, 212, 233, 234, 235, 256, 257, 258, 279, 280, 281, 302, 303, 304, 327], "bbox": [0.1483125, 0.2584166666666667, 0.2528125, 0.8292083333333334], "iscrowd": 0, "area": 11346.465150000002, "rle": {"size": [480, 640], "counts": "Rh\\19_>c0]Oc0]Oc0]Oa0_O:cFTM[6V3\\IQMa6S3XITMc6R3TIUMi6o2oHWMo6n2iHWMU7m2cHYM[7l2\\HZMb7k2VHZMh7j2PH\\Mn7i2iG\\MV8h4N2M3N2N2N2N2N2N2N1O2N2N2N2N3M7K4L4L5K4K5L1O00000000gHbIP6]6PJhIl5X6TJlIh5T6WJRJd5n5\\J[J[5e5eJcJ_1oNn0^6bMmJT1oNS1S6jMVKi0oNW1k5oM`K>nN]1b5UNhK3POa1X5[NRLHoNf1V5\\NSLCPOk1U5ZNTL_OPOQ2T5WNXNd1P2TNVNf1h7J6J5K6J6J5K6J6J5K6Jeeo6"}, "label": "the man with glasses in the grey sirt"}]}
{"id": 335758, "image": "COCO_train2014_000000335758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far most man waring a gray t - shirt\"."}], "task": "refering", "answer_template": "The \"the far most man waring a gray t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 210, 211, 212, 233, 234, 235, 256, 257, 258, 279, 280, 281, 302, 303, 304, 327], "bbox": [0.1483125, 0.2584166666666667, 0.2528125, 0.8292083333333334], "iscrowd": 0, "area": 11346.465150000002, "rle": {"size": [480, 640], "counts": "Rh\\19_>c0]Oc0]Oc0]Oa0_O:cFTM[6V3\\IQMa6S3XITMc6R3TIUMi6o2oHWMo6n2iHWMU7m2cHYM[7l2\\HZMb7k2VHZMh7j2PH\\Mn7i2iG\\MV8h4N2M3N2N2N2N2N2N2N1O2N2N2N2N3M7K4L4L5K4K5L1O00000000gHbIP6]6PJhIl5X6TJlIh5T6WJRJd5n5\\J[J[5e5eJcJ_1oNn0^6bMmJT1oNS1S6jMVKi0oNW1k5oM`K>nN]1b5UNhK3POa1X5[NRLHoNf1V5\\NSLCPOk1U5ZNTL_OPOQ2T5WNXNd1P2TNVNf1h7J6J5K6J6J5K6J6J5K6Jeeo6"}, "label": "the far most man waring a gray t - shirt"}]}
{"id": 335758, "image": "COCO_train2014_000000335758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a grey shirt\"."}], "task": "refering", "answer_template": "The \"a man in a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376], "bbox": [0.22009375000000003, 0.24097916666666666, 0.5553125, 0.9881458333333334], "iscrowd": 0, "area": 38723.60315000001, "rle": {"size": [480, 640], "counts": "eXR23>=`=c0O2M2N4XEiN`7\\1SHPOi7W1kGROR8S1cGUOZ8Q1[GXOb8j0TGAi8a0mFIP9d0ZFN^9Q3H9H7I6I8I6I8I7I<QJaHd3m7_KcHa4i7cJgH\\5_800O10000O100O10000O1000000O1000000O10O1000O100O100O100O1O2O0O100O10N1J7I7I6PNQ2K4L5L4I7L42N2N2M3N2N1O2N2N2M3N2N1O4M4K5K5L4K6J5L4K3M3N3L3M3N2M4L4M2M21OM1N2O1O1K5J6L3L5L4LRK^KFa40mKLT42PLLo32VLLi33[LKd34aLH_37eLGZ39iLEU3;oLCP3=TM_Ok2d0YMWOg2j0]MQOc2Q1_MkNa2W1aMdN`2]1cM_N]2c1eMYN[2i1gMSNY2o1iMlMX2U2kMgMU2WOaKo0Z2GU2YOdKQ1V2DV2YOgKU1S2_OV2[OiKW1Q2\\OV2\\OlKY1n1XOV2^OnK[1l1UOV2_OQL[1j1TOU2@TLZ1j1SOS2AVL[1h1QOS2CXLZ1h1POP2E[LZ1f1oNo1G]LX1g1nNm1H_LY1e1mNl1J`LY1n1bNb15bLW1X2XNV1`0gLU1\\2QNn0i0jLR1a2kMf0S1mLo0d2cMb0]1oLl0_7SOeHi0[7VOPJAP6=QJCP6;RJDo5:SJEn59SJGn56TJJm5OZJ0g5\\OlJe0T5ZOlJf0U5XOmJg0T5XOlJh0W5TOjJl0f9O00010O00001O00010O0O101O00001O0000000O1000001O000000000O1001O01O000000000001O01O0000001O0010O01O001O0O2M3N3M4KjYU4"}, "label": "a man in a grey shirt"}]}
{"id": 335758, "image": "COCO_train2014_000000335758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man swinging in a gray tshirt\"."}], "task": "refering", "answer_template": "The \"the man swinging in a gray tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376], "bbox": [0.22009375000000003, 0.24097916666666666, 0.5553125, 0.9881458333333334], "iscrowd": 0, "area": 38723.60315000001, "rle": {"size": [480, 640], "counts": "eXR23>=`=c0O2M2N4XEiN`7\\1SHPOi7W1kGROR8S1cGUOZ8Q1[GXOb8j0TGAi8a0mFIP9d0ZFN^9Q3H9H7I6I8I6I8I7I<QJaHd3m7_KcHa4i7cJgH\\5_800O10000O100O10000O1000000O1000000O10O1000O100O100O100O1O2O0O100O10N1J7I7I6PNQ2K4L5L4I7L42N2N2M3N2N1O2N2N2M3N2N1O4M4K5K5L4K6J5L4K3M3N3L3M3N2M4L4M2M21OM1N2O1O1K5J6L3L5L4LRK^KFa40mKLT42PLLo32VLLi33[LKd34aLH_37eLGZ39iLEU3;oLCP3=TM_Ok2d0YMWOg2j0]MQOc2Q1_MkNa2W1aMdN`2]1cM_N]2c1eMYN[2i1gMSNY2o1iMlMX2U2kMgMU2WOaKo0Z2GU2YOdKQ1V2DV2YOgKU1S2_OV2[OiKW1Q2\\OV2\\OlKY1n1XOV2^OnK[1l1UOV2_OQL[1j1TOU2@TLZ1j1SOS2AVL[1h1QOS2CXLZ1h1POP2E[LZ1f1oNo1G]LX1g1nNm1H_LY1e1mNl1J`LY1n1bNb15bLW1X2XNV1`0gLU1\\2QNn0i0jLR1a2kMf0S1mLo0d2cMb0]1oLl0_7SOeHi0[7VOPJAP6=QJCP6;RJDo5:SJEn59SJGn56TJJm5OZJ0g5\\OlJe0T5ZOlJf0U5XOmJg0T5XOlJh0W5TOjJl0f9O00010O00001O00010O0O101O00001O0000000O1000001O000000000O1001O01O000000000001O01O0000001O0010O01O001O0O2M3N3M4KjYU4"}, "label": "the man swinging in a gray tshirt"}]}
{"id": 335758, "image": "COCO_train2014_000000335758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with green shirt playing video games\"."}], "task": "refering", "answer_template": "The \"a person with green shirt playing video games\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 93, 94, 95, 116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325, 347, 348, 370, 371], "bbox": [0.035312500000000004, 0.21256250000000002, 0.21189062500000003, 0.9816041666666666], "iscrowd": 0, "area": 20988.46985, "rle": {"size": [480, 640], "counts": "mn:i2k;?J5K5L5J5K5L5L3M3M4K4M3M4L3M3M7I8H8H8H8H8B>ZOf0F:F:F:F:F:H8H8I8G8I7H8I7H8I5J101N110O01O1O1O1O001O1O100O001^FULm7l3QHVLn7k3QHVLn7k3PHWLP8h3PHYLo7h3oGYLQ8h3nG\\Ln7e3PH`Ll7`3THdLh7]3VHhLg7W3XHnLd7S3XHTMd7l2ZHZMb7g2ZH`Mb7`2ZHhMb7Y2YHkMg1RNf3R4_JPNg1[Na3e3dJSNg1fN\\3X3hJVNh1oNW3k2oJXNe1ZOT3_2UKXNh0UN_Oa1Z4S2^KXNh0^1m2h2SM^Ng1c1XNQOJcJ`0\\6EROJgJ<X6ISOHkJ;Q6MUOGXKNd5:VOEZKNa5<VOD]KN\\5>YOA_KNX5a0ZO_ObKNS5c0f70000000000000000000001O0000000000000O10001O3M3M2N3M3MVa\\7"}, "label": "a person with green shirt playing video games"}]}
{"id": 335758, "image": "COCO_train2014_000000335758.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green shirt and white hat playing wii\"."}], "task": "refering", "answer_template": "The \"a man in a green shirt and white hat playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 93, 94, 95, 116, 117, 118, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325, 347, 348, 370, 371], "bbox": [0.035312500000000004, 0.21256250000000002, 0.21189062500000003, 0.9816041666666666], "iscrowd": 0, "area": 20988.46985, "rle": {"size": [480, 640], "counts": "mn:i2k;?J5K5L5J5K5L5L3M3M4K4M3M4L3M3M7I8H8H8H8H8B>ZOf0F:F:F:F:F:H8H8I8G8I7H8I7H8I5J101N110O01O1O1O1O001O1O100O001^FULm7l3QHVLn7k3QHVLn7k3PHWLP8h3PHYLo7h3oGYLQ8h3nG\\Ln7e3PH`Ll7`3THdLh7]3VHhLg7W3XHnLd7S3XHTMd7l2ZHZMb7g2ZH`Mb7`2ZHhMb7Y2YHkMg1RNf3R4_JPNg1[Na3e3dJSNg1fN\\3X3hJVNh1oNW3k2oJXNe1ZOT3_2UKXNh0UN_Oa1Z4S2^KXNh0^1m2h2SM^Ng1c1XNQOJcJ`0\\6EROJgJ<X6ISOHkJ;Q6MUOGXKNd5:VOEZKNa5<VOD]KN\\5>YOA_KNX5a0ZO_ObKNS5c0f70000000000000000000001O0000000000000O10001O3M3M2N3M3MVa\\7"}, "label": "a man in a green shirt and white hat playing wii"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of chicken ball below another chicken ball in a bowl with paper cover\"."}], "task": "refering", "answer_template": "The \"a piece of chicken ball below another chicken ball in a bowl with paper cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 122, 123], "bbox": [0.3204575163398693, 0.07217320261437908, 0.6330555555555556, 0.2473039215686275], "iscrowd": 0, "area": 11231.09815, "rle": {"size": [612, 612], "counts": "hie32Qc02O1O1N2O2N1N2O1O2N1N2O1O2M2O1O1N2O2N1N2O1O2N1N2O1O2M101O0O2O001N101O001N101O0O2O001N101O0O2O001O0O2O000O2O001N101O001N101O0O2O001N101O0O2O001O0O2O001N101O001O001O001O001O010O0000001O000000001O0000001O0000000100O001O00100O001O0010O01O1O010O001O1O010O001O00100O002N100O1O1O1O101N1O1O1O100O1O2N100O1O1O1O101N1O1O1O100O1O2N1O100O1O1O101N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2N2M3N2N2N2M3N2NaZV4"}, "label": "a piece of chicken ball below another chicken ball in a bowl with paper cover"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half seen cake that is inside the cone\"."}], "task": "refering", "answer_template": "The \"a half seen cake that is inside the cone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 122, 123], "bbox": [0.3204575163398693, 0.07217320261437908, 0.6330555555555556, 0.2473039215686275], "iscrowd": 0, "area": 11231.09815, "rle": {"size": [612, 612], "counts": "hie32Qc02O1O1N2O2N1N2O1O2N1N2O1O2M2O1O1N2O2N1N2O1O2N1N2O1O2M101O0O2O001N101O001N101O0O2O001N101O0O2O001O0O2O000O2O001N101O001N101O0O2O001N101O0O2O001O0O2O001N101O001O001O001O001O010O0000001O000000001O0000001O0000000100O001O00100O001O0010O01O1O010O001O1O010O001O00100O002N100O1O1O1O101N1O1O1O100O1O2N100O1O1O1O101N1O1O1O100O1O2N1O100O1O1O101N2N2N2N2O1N2N2N2N2O1N2N2N2N2N2N2M3N2N2N2M3N2NaZV4"}, "label": "a half seen cake that is inside the cone"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pastries in a paper container inside of the coffee cup on a yellow plate with other pastries\"."}], "task": "refering", "answer_template": "The \"pastries in a paper container inside of the coffee cup on a yellow plate with other pastries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 33, 34, 35, 36, 37, 38, 39, 40, 41, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 103, 104], "bbox": [0.5190196078431373, 0.022581699346405228, 0.8970915032679738, 0.19931372549019608], "iscrowd": 0, "area": 17644.4765, "rle": {"size": [612, 612], "counts": "VRn51Sc02N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N001O1O001O1O001O1O00100O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O00001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O000000001O0000O100O10001N100O100O10000O100O100O10000O100O10000O100O1O100O100O100O1O100O100O100O1O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100N2O1O1N2O1O1O1N2O1M3L4M3L4L4L4L5K4L4L4L4L4L4L4L_dU1"}, "label": "pastries in a paper container inside of the coffee cup on a yellow plate with other pastries"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a doughnut almost falling out of a cup\"."}], "task": "refering", "answer_template": "The \"a doughnut almost falling out of a cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 33, 34, 35, 36, 37, 38, 39, 40, 41, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 103, 104], "bbox": [0.5190196078431373, 0.022581699346405228, 0.8970915032679738, 0.19931372549019608], "iscrowd": 0, "area": 17644.4765, "rle": {"size": [612, 612], "counts": "VRn51Sc02N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N001O1O001O1O001O1O00100O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O00001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O000000001O0000O100O10001N100O100O10000O100O100O10000O100O10000O100O1O100O100O100O1O100O100O100O1O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100N2O1O1N2O1O1O1N2O1M3L4M3L4L4L4L5K4L4L4L4L4L4L4L_dU1"}, "label": "a doughnut almost falling out of a cup"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small white bowl with white stuff in it\"."}], "task": "refering", "answer_template": "The \"small white bowl with white stuff in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 298, 299, 300, 301, 302, 303, 304, 305, 306, 321, 322, 323, 324, 325, 326, 327, 328, 344, 345, 346, 347, 348, 349, 350, 367, 368, 369, 370, 371], "bbox": [0.5829575163398693, 0.4775816993464052, 0.9439542483660132, 0.7757843137254902], "iscrowd": 0, "area": 28848.232400000008, "rle": {"size": [612, 612], "counts": "^fe6R1Qb02N3M3L3N3M2N3M3L3N3M2N3M3L3N3M2N3M2M4M3M1O2N2M3N1O2N2M3N1O2N2N2N101N2O1N101N2O1N2O0O2O1O1N101N2O1N101N2O1N101N2O1O0O2O1N101O1O001O0O2O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O1O001O0O2O001O001O001O001O001O10O01O010OO101O0O101O0O101N10001N100O2O000O2O000O2O0O101O0O101N10001N10001N101N10001N101O0O2O0O2O001N101N101O0O2O000O2O0O2O1N2N2N3M2M3N2N2N2M3N2N2N2M3N2N3L3N2N2N2N2N2N2N2N2M3N2N2N3M2N2M3M3M3N2M5K4L5L3L4L5H7E<Dknc0"}, "label": "small white bowl with white stuff in it"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small dish of butter served next to some baked dish\"."}], "task": "refering", "answer_template": "The \"a small dish of butter served next to some baked dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 298, 299, 300, 301, 302, 303, 304, 305, 306, 321, 322, 323, 324, 325, 326, 327, 328, 344, 345, 346, 347, 348, 349, 350, 367, 368, 369, 370, 371], "bbox": [0.5829575163398693, 0.4775816993464052, 0.9439542483660132, 0.7757843137254902], "iscrowd": 0, "area": 28848.232400000008, "rle": {"size": [612, 612], "counts": "^fe6R1Qb02N3M3L3N3M2N3M3L3N3M2N3M3L3N3M2N3M2M4M3M1O2N2M3N1O2N2M3N1O2N2N2N101N2O1N101N2O1N2O0O2O1O1N101N2O1N101N2O1N101N2O1O0O2O1N101O1O001O0O2O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O1O001O0O2O001O001O001O001O001O10O01O010OO101O0O101O0O101N10001N100O2O000O2O000O2O0O101O0O101N10001N10001N101N10001N101O0O2O0O2O001N101N101O0O2O000O2O0O2O1N2N2N3M2M3N2N2N2M3N2N2N2M3N2N3L3N2N2N2N2N2N2N2N2M3N2N2N3M2N2M3M3M3N2M5K4L5L3L4L5H7E<Dknc0"}, "label": "a small dish of butter served next to some baked dish"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white ceramic bowl holding a paper cone full of fresh rolls\"."}], "task": "refering", "answer_template": "The \"a white ceramic bowl holding a paper cone full of fresh rolls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 158, 159, 160, 168, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 208, 209, 210, 211, 225, 226, 227, 228, 229, 230, 231, 232, 233, 248, 249, 250, 251, 252, 253, 254, 255, 270, 271, 272, 273, 274, 275, 276, 293, 294, 295, 296, 297, 298, 317, 318], "bbox": [0.2112418300653595, 0.26965686274509804, 0.6696568627450981, 0.6494444444444445], "iscrowd": 0, "area": 31150.0627, "rle": {"size": [612, 612], "counts": "lX]21Rc05K6J5K5K5L4K5K5K6J5K6J5L5J5K6J5K5K4L3M4M3M3000O2O0000001O0O100000N2M3N1O2N2N2M3N2N1O1N2O1O1O1N2O1O1O1N2O1O1O1N0100O10O10O1000O0100O1000O010000O01000O10O010000O01000O10O10O01O001O1O001O1O001O0010O01O00001O001O00001O00001O0010O0001O001O00001O00001O001O01O01O00001O001O00001O00001O010O00000000O10000O10000O10001N100O10000O10000O10000O10001N10000O10000O100O10000O10001N10O10O10000O1000O010000O10000O010O10000O10O10O1O1N2N2N2N1O2N2N2O1N2N2N1O2N2M3N2N2N1O2M3N2N2N2N1O2M2O1O1O2N1N2O1O2N1O1N2O2N1O1O1N2O2N1O1O1O2M2O1N2N3M2N2M3N3M2N2M3N2N3N1O1O1O2N1O1O1O2N1O1O1O3M3M3Mj`h3"}, "label": "a white ceramic bowl holding a paper cone full of fresh rolls"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup with paper and food in it\"."}], "task": "refering", "answer_template": "The \"a white cup with paper and food in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 158, 159, 160, 168, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 203, 204, 205, 206, 207, 208, 209, 210, 211, 225, 226, 227, 228, 229, 230, 231, 232, 233, 248, 249, 250, 251, 252, 253, 254, 255, 270, 271, 272, 273, 274, 275, 276, 293, 294, 295, 296, 297, 298, 317, 318], "bbox": [0.2112418300653595, 0.26965686274509804, 0.6696568627450981, 0.6494444444444445], "iscrowd": 0, "area": 31150.0627, "rle": {"size": [612, 612], "counts": "lX]21Rc05K6J5K5K5L4K5K5K6J5K6J5L5J5K6J5K5K4L3M4M3M3000O2O0000001O0O100000N2M3N1O2N2N2M3N2N1O1N2O1O1O1N2O1O1O1N2O1O1O1N0100O10O10O1000O0100O1000O010000O01000O10O010000O01000O10O10O01O001O1O001O1O001O0010O01O00001O001O00001O00001O0010O0001O001O00001O00001O001O01O01O00001O001O00001O00001O010O00000000O10000O10000O10001N100O10000O10000O10000O10001N10000O10000O100O10000O10001N10O10O10000O1000O010000O10000O010O10000O10O10O1O1N2N2N2N1O2N2N2O1N2N2N1O2N2M3N2N2N1O2M3N2N2N2N1O2M2O1O1O2N1N2O1O2N1O1N2O2N1O1O1N2O2N1O1O1O2M2O1N2N3M2N2M3N3M2N2M3N2N3N1O1O1O2N1O1O1O2N1O1O1O3M3M3Mj`h3"}, "label": "a white cup with paper and food in it"}]}
{"id": 262031, "image": "COCO_train2014_000000262031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pastry filled with jelly that is not spilling\"."}], "task": "refering", "answer_template": "The \"a pastry filled with jelly that is not spilling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 310, 311, 312, 313, 314, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 360, 375, 376, 377, 378, 379, 380, 381, 382, 396, 397, 398, 399, 400, 401, 402, 403, 404, 419, 420, 421, 422, 423, 424, 425, 426, 441, 442, 443, 444, 445, 446, 447, 448, 463, 464, 465, 466, 467, 468, 469, 470], "bbox": [0.040539215686274506, 0.6204901960784314, 0.4031535947712418, 0.9786111111111111], "iscrowd": 0, "area": 37243.243249999985, "rle": {"size": [612, 612], "counts": "[^?8gb0<D<E;D<E<C<E;D<D<E4K6L4M2O1O1O1N3N1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N3N1O1N2O1O1O1N2O1O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O2N100O100O100O1O100O100001O001O001O00001O001O001O001O00001O001O001O0010O01O00001O001O001O001O001O00001O001O001O001O1O001O1O1O1O1O001O1O1O1O1O010O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O1PMfAT1Z>fNPBV1Q>cNYBY1h=`NbB\\1_=\\NmB_1T=ZNVCb1k<WN_Ce1b<TNhCh1o>K4L4J6J6J7I6K5J6J6J`bi6"}, "label": "a pastry filled with jelly that is not spilling"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man writing while sitting on a park bench\"."}], "task": "refering", "answer_template": "The \"man writing while sitting on a park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.07246875, 0.3296458333333333, 0.5528125, 0.9835833333333334], "iscrowd": 0, "area": 53290.55520000001, "rle": {"size": [480, 640], "counts": "ble02c>=B?B>J6K5J5L5K5K5J6K4L5K5J6A?L3L5L4L4K5L4K4M4L4K5L4L4K5L3M4L9F;F:F:fNWIjIS7l5nHSJW7h5jHWJW7g5lHWJU7h5lHWJU7h5lHWJU7g5mHXJT7h5lHWJU7i5kHVJV7j5jHVJV7i5lHVJT7j5lHVJT7j5lHWJS7i5mHWJS7i5mHWJS7h5nHXJR7h5oHXJP7h5PIWJQ7i5oHVJR7j5nHUJS7k5mHTJT7k5mHTJT7l5mHQJU7o5kHPJV7P6jHoIW7n5lHQJU7k5oHTJR7h5WIQJk6k5W1L4M3L4M300O100000000O100000000O1000000O100000000O100000000O1000000000000O1000000000000O100000000000000O1000000000000O100001O00001O00001O00001gIlJd2T5\\MPK`2P5_MUK]2l4aMYK[2g4dM^KX2b4hM_KW2b4gMaKW2_4hMcKW2^4gMdKX2\\4fMgKY2Y4eMjKZ2W4dMkK[2U4bMoK]2Q4aMSL]2n3`MUL_2k3_MXL`2h3]M\\Lb2e3[LaJ@m1T4d3RLhJGg1V4[4QKVJg0c1V4m4dKYK[4g4`K_K_4a4_KbK`4_4_KbK`4^4`KcK_4^4`KcK_4]4aKeK]4\\4bKeK]4[4cKfK\\4Z4dKgK[4Z4cKhK\\4X4dKiK[4X4dKgK]4Y4cKgK]4Y4cKfK^4[4aKdK`4\\4`KcKa4^4^KaKc4_4\\KbKd4^4\\KaKe4`4ZK_Kg4a4YK^Kh4c4WK\\Kj4d4VK[Kk4e4UK[Kk4f4TKYKm4\\71O1O100N2O1N2O1N2O1O1N2OlLdFf0Z9ZOjFd0V9ZOnFd0R9YOTGd0l8ZOXGd0h8YO]Ge0c8ZOaGc0_8]OcGa0]8^OfG`0Z8_OgGa0Y8_OgGa0Y8^OhGb0W8^OkGa0U8_OkGa0U8^OlGb0T8]OmGc0T8\\OmGc0T8[OmGe0T8YOmGg0S8YOmGg0T8WOnGh0T8UOmGk0V8ROjGn0X8oNjGP1Y8XNVF0b1h1[8nM\\F8Y1j1P9SNQGm1P9QNRGn1P9PNPGP2R9mMoFS2S9jMnFV2S9iMnFV2n:000001O00001O0000001O0000001O00001O0000001O0000001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O1O4L5K4L5K4L4L5CmcU4"}, "label": "man writing while sitting on a park bench"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man reading\"."}], "task": "refering", "answer_template": "The \"man reading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.07246875, 0.3296458333333333, 0.5528125, 0.9835833333333334], "iscrowd": 0, "area": 53290.55520000001, "rle": {"size": [480, 640], "counts": "ble02c>=B?B>J6K5J5L5K5K5J6K4L5K5J6A?L3L5L4L4K5L4K4M4L4K5L4L4K5L3M4L9F;F:F:fNWIjIS7l5nHSJW7h5jHWJW7g5lHWJU7h5lHWJU7h5lHWJU7g5mHXJT7h5lHWJU7i5kHVJV7j5jHVJV7i5lHVJT7j5lHVJT7j5lHWJS7i5mHWJS7i5mHWJS7h5nHXJR7h5oHXJP7h5PIWJQ7i5oHVJR7j5nHUJS7k5mHTJT7k5mHTJT7l5mHQJU7o5kHPJV7P6jHoIW7n5lHQJU7k5oHTJR7h5WIQJk6k5W1L4M3L4M300O100000000O100000000O1000000O100000000O100000000O1000000000000O1000000000000O100000000000000O1000000000000O100001O00001O00001O00001gIlJd2T5\\MPK`2P5_MUK]2l4aMYK[2g4dM^KX2b4hM_KW2b4gMaKW2_4hMcKW2^4gMdKX2\\4fMgKY2Y4eMjKZ2W4dMkK[2U4bMoK]2Q4aMSL]2n3`MUL_2k3_MXL`2h3]M\\Lb2e3[LaJ@m1T4d3RLhJGg1V4[4QKVJg0c1V4m4dKYK[4g4`K_K_4a4_KbK`4_4_KbK`4^4`KcK_4^4`KcK_4]4aKeK]4\\4bKeK]4[4cKfK\\4Z4dKgK[4Z4cKhK\\4X4dKiK[4X4dKgK]4Y4cKgK]4Y4cKfK^4[4aKdK`4\\4`KcKa4^4^KaKc4_4\\KbKd4^4\\KaKe4`4ZK_Kg4a4YK^Kh4c4WK\\Kj4d4VK[Kk4e4UK[Kk4f4TKYKm4\\71O1O100N2O1N2O1N2O1O1N2OlLdFf0Z9ZOjFd0V9ZOnFd0R9YOTGd0l8ZOXGd0h8YO]Ge0c8ZOaGc0_8]OcGa0]8^OfG`0Z8_OgGa0Y8_OgGa0Y8^OhGb0W8^OkGa0U8_OkGa0U8^OlGb0T8]OmGc0T8\\OmGc0T8[OmGe0T8YOmGg0S8YOmGg0T8WOnGh0T8UOmGk0V8ROjGn0X8oNjGP1Y8XNVF0b1h1[8nM\\F8Y1j1P9SNQGm1P9QNRGn1P9PNPGP2R9mMoFS2S9jMnFV2S9iMnFV2n:000001O00001O0000001O0000001O00001O0000001O0000001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O1O4L5K4L5K4L4L5CmcU4"}, "label": "man reading"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty bench with black metal arm rests\"."}], "task": "refering", "answer_template": "The \"an empty bench with black metal arm rests\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 131, 132, 133, 134, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 270, 271, 273, 274, 293, 297], "bbox": [0.71103125, 0.283125, 0.97596875, 0.7454583333333333], "iscrowd": 0, "area": 20085.526649999996, "rle": {"size": [480, 640], "counts": "m]e62m>4M3M4L3L5L3M3L5L3M4K4M3M4K4M4L3L4M4L3L5L3M3M4K4M4L3L4[GkLY5Y3ZJoLg5T3SJnLm5V3nIkLS6Y3gIhLZ6[3aIfL_6_3\\IaLe6`3XIaLi6`3TIaLl6a3RI^LP7c3mH^LT7c3iH^LW7c3gH^LZ7c3dH]L]7d3`H]L`7e3]H\\Ld7e3ZH[Lg7e3WH\\Lj7e3SH\\Lm7f3PH[LQ8f3mGZLT8g3iGYLX8h3fGYL[8h3cGXL^8i3_GXLa8j3\\GWLe8j3XGWLi8i3VGWLj8k3SGVLn8k3oFVLR9k3lFULT9k41O1N101O1O1O001O3M3M3M3M3M3L4M3N2N2N2N2N2N2N2N2N2N2N2N1O0000000000000000000000000000000000000001O001O1O1O1O001O1O1O001O1O\\O]ElLb:S3`ElL`:S3aEnL^:R3cEnL\\:Q3fEoLY:P3iEoLX:o2iERMW:m2jESMV:k2lEUMT:i2nEVMR:j2nEWMR:g2PFYMP:e2RF[Mn9d2SF_Mj9_2WFmM]9R2eFZNo8d1SG_Nj8`1WG_Nj8_1WG`Nk8^1WGaNi8_1XG`Ni8^1YGaNh8]1YGcNh8[1ZGdNg8[1ZGdNf8[1\\GdNe8Z1\\GeNf8Z1[GeNf8Y1]GeNd8Y1dG`N\\8`1jGZNW8d1PHVNQ8h1VHRNk7l1]HmMd7R2f2N1O4M4K4L5K5L4K5K5L4K2N100O1O2Nei6"}, "label": "an empty bench with black metal arm rests"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench that is a little ways from where the man is sitting\"."}], "task": "refering", "answer_template": "The \"the bench that is a little ways from where the man is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 131, 132, 133, 134, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 270, 271, 273, 274, 293, 297], "bbox": [0.71103125, 0.283125, 0.97596875, 0.7454583333333333], "iscrowd": 0, "area": 20085.526649999996, "rle": {"size": [480, 640], "counts": "m]e62m>4M3M4L3L5L3M3L5L3M4K4M3M4K4M4L3L4M4L3L5L3M3M4K4M4L3L4[GkLY5Y3ZJoLg5T3SJnLm5V3nIkLS6Y3gIhLZ6[3aIfL_6_3\\IaLe6`3XIaLi6`3TIaLl6a3RI^LP7c3mH^LT7c3iH^LW7c3gH^LZ7c3dH]L]7d3`H]L`7e3]H\\Ld7e3ZH[Lg7e3WH\\Lj7e3SH\\Lm7f3PH[LQ8f3mGZLT8g3iGYLX8h3fGYL[8h3cGXL^8i3_GXLa8j3\\GWLe8j3XGWLi8i3VGWLj8k3SGVLn8k3oFVLR9k3lFULT9k41O1N101O1O1O001O3M3M3M3M3M3L4M3N2N2N2N2N2N2N2N2N2N2N2N1O0000000000000000000000000000000000000001O001O1O1O1O001O1O1O001O1O\\O]ElLb:S3`ElL`:S3aEnL^:R3cEnL\\:Q3fEoLY:P3iEoLX:o2iERMW:m2jESMV:k2lEUMT:i2nEVMR:j2nEWMR:g2PFYMP:e2RF[Mn9d2SF_Mj9_2WFmM]9R2eFZNo8d1SG_Nj8`1WG_Nj8_1WG`Nk8^1WGaNi8_1XG`Ni8^1YGaNh8]1YGcNh8[1ZGdNg8[1ZGdNf8[1\\GdNe8Z1\\GeNf8Z1[GeNf8Y1]GeNd8Y1dG`N\\8`1jGZNW8d1PHVNQ8h1VHRNk7l1]HmMd7R2f2N1O4M4K4L5K5L4K5K5L4K2N100O1O2Nei6"}, "label": "the bench that is a little ways from where the man is sitting"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the bench the man is sitting on\"."}], "task": "refering", "answer_template": "The \"the back of the bench the man is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 230, 231, 253, 254, 276, 277, 299, 300, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 330, 346, 347, 348, 349, 350, 351, 352, 353, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.018546874999999997, 0.5707916666666667, 0.37921875000000005, 0.9842708333333334], "iscrowd": 0, "area": 22884.9715, "rle": {"size": [480, 640], "counts": "nm59_>=C>B=E;K6I6K5K5J7J5K5J7J5K5J6K6J5K5J6K6J5J6K6J5J6K5K6I3N00000O10000O1000nJUHo2j7oL`Hj2`7SMiHg2W7VMSIc2l6[M^I^2b6_MdI^2\\6_MgIa2X6]MkIc2U6ZMoIe2Q6XMRJh2n5UMUJk2j5SMYJm2m800000000000000000000000000000000000000000O100O100O100O100O100O100O100O100O100O100O100O10000001O001O001O1O001O001O001O001O001O001O001O001O001O001O001O001O001O0000001O000000001O0000001O000000001O000000O100O100O10000O100O100O100O10000O100O100O10000O100O100O10000O100O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2Nd0\\Og0YO^di5"}, "label": "the back of the bench the man is sitting on"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting on a park bench\"."}], "task": "refering", "answer_template": "The \"a woman sitting on a park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 182, 183, 206, 229], "bbox": [0.8584375, 0.176625, 0.98890625, 0.6025208333333334], "iscrowd": 0, "area": 7457.167449999997, "rle": {"size": [480, 640], "counts": "^`Q85o=l0I7J6J600O2N1O1O100O1O1O100O1O10000O101O0000000000000O10001O1O1O1O1O1O1N2G9C=N2N2N2N2N2N2N2N2N2O1N3N1O1O2N1O1O1O2N1O1O1OiMTE<j:CYE<c0PNU9d1YF<=[NS9X1cF<5gNQ9m0kF=OPOo8b0UG=G\\Om86^GT2`8lMbGS2]8lMeGU2Y8kMiGT2V8kMlGU2S8kMoGT2P8kMSHc0GhNT8e0VHO?WOZ7i0ZH]OT1Da6o0\\HkNQ;P1g1I6I8H7J7H7JPU3"}, "label": "a woman sitting on a park bench"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurred black and white woman sitting\"."}], "task": "refering", "answer_template": "The \"a blurred black and white woman sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 182, 183, 206, 229], "bbox": [0.8584375, 0.176625, 0.98890625, 0.6025208333333334], "iscrowd": 0, "area": 7457.167449999997, "rle": {"size": [480, 640], "counts": "^`Q85o=l0I7J6J600O2N1O1O100O1O1O100O1O10000O101O0000000000000O10001O1O1O1O1O1O1N2G9C=N2N2N2N2N2N2N2N2N2O1N3N1O1O2N1O1O1O2N1O1O1OiMTE<j:CYE<c0PNU9d1YF<=[NS9X1cF<5gNQ9m0kF=OPOo8b0UG=G\\Om86^GT2`8lMbGS2]8lMeGU2Y8kMiGT2V8kMlGU2S8kMoGT2P8kMSHc0GhNT8e0VHO?WOZ7i0ZH]OT1Da6o0\\HkNQ;P1g1I6I8H7J7H7JPU3"}, "label": "a blurred black and white woman sitting"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty side of the bench\"."}], "task": "refering", "answer_template": "The \"empty side of the bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 241, 262, 263, 264, 265, 284, 285, 286, 287, 288, 307, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 357], "bbox": [0.372703125, 0.601125, 0.603421875, 0.9096666666666666], "iscrowd": 0, "area": 10918.025150000001, "rle": {"size": [480, 640], "counts": "]\\`36i>5J6K6I6K1000001N100QNROlEn0T:ROlEn0T:ROlEo0R:ROnEn0R:ROnEn0Q:TOmEm0S:SOmEm0S:XOhEi0W:\\OdEd0[:]OeEc0[:^OdEb0\\:_OcEb0\\:_ObEc0]:\\OdEe0Z:[OgEf0X:YOdEm0[:ROaET1^:lN^EY1a:fN]E]1b:cN_E^1`:aN`Ea1_:_N`Eb1`:]N`Ee1_:ZNaEg1_:YN`Eh1`:WN`Ek1_:TNaEm1_:SN`En1`:QN`EQ2_:nMaES2_:mM`EU2_:jMaEW2_:hMaEY2_:fMaE\\2^:dMaE]2_:bMaE_2_:`MbEa2]:_MbEb2^:]MbEd2\\;2N1UOXM]Ei2c:XM[Ej2d:WMZEj2g:VMWEl2h:UMVEl2j:UMTEl2l:UMREm2m:TMQEm2P;RMoDo2Q;RMmDP3R;QMlDQ3S;PMkDR3T;oLkDR3U;nLiDT3V;<0O1O100O1O010O1O1O100O1O00100O1O1O100O001O100O1O10O01O1O100O1O1O2O0O2N1O101N1O101O0O2O00oCVMk;i2SDXMP<k23L7J6I7I6K6I7I4M0O2O0O1O101N100O2N100O101N1O101N1O4M2M4M2M4L4M2M4M3LUef3"}, "label": "empty side of the bench"}]}
{"id": 409488, "image": "COCO_train2014_000000409488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 handrails on a bench behind a man\"."}], "task": "refering", "answer_template": "The \"2 handrails on a bench behind a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 241, 262, 263, 264, 265, 284, 285, 286, 287, 288, 307, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 357], "bbox": [0.372703125, 0.601125, 0.603421875, 0.9096666666666666], "iscrowd": 0, "area": 10918.025150000001, "rle": {"size": [480, 640], "counts": "]\\`36i>5J6K6I6K1000001N100QNROlEn0T:ROlEn0T:ROlEo0R:ROnEn0R:ROnEn0Q:TOmEm0S:SOmEm0S:XOhEi0W:\\OdEd0[:]OeEc0[:^OdEb0\\:_OcEb0\\:_ObEc0]:\\OdEe0Z:[OgEf0X:YOdEm0[:ROaET1^:lN^EY1a:fN]E]1b:cN_E^1`:aN`Ea1_:_N`Eb1`:]N`Ee1_:ZNaEg1_:YN`Eh1`:WN`Ek1_:TNaEm1_:SN`En1`:QN`EQ2_:nMaES2_:mM`EU2_:jMaEW2_:hMaEY2_:fMaE\\2^:dMaE]2_:bMaE_2_:`MbEa2]:_MbEb2^:]MbEd2\\;2N1UOXM]Ei2c:XM[Ej2d:WMZEj2g:VMWEl2h:UMVEl2j:UMTEl2l:UMREm2m:TMQEm2P;RMoDo2Q;RMmDP3R;QMlDQ3S;PMkDR3T;oLkDR3U;nLiDT3V;<0O1O100O1O010O1O1O100O1O00100O1O1O100O001O100O1O10O01O1O100O1O1O2O0O2N1O101N1O101O0O2O00oCVMk;i2SDXMP<k23L7J6I7I6K6I7I4M0O2O0O1O101N100O2N100O101N1O101N1O4M2M4M2M4L4M2M4M3LUef3"}, "label": "2 handrails on a bench behind a man"}]}
{"id": 16273, "image": "COCO_train2014_000000016273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing glasses and a red shirt sitting at a table\"."}], "task": "refering", "answer_template": "The \"a woman wearing glasses and a red shirt sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 263, 264, 265, 266], "bbox": [0.45, 0.1479375, 0.711234375, 0.6827708333333333], "iscrowd": 0, "area": 21007.177900000006, "rle": {"size": [480, 640], "counts": "bYW49f>2N2N3M2N2^O]OeBd0X=_OfBc0W=@gBb0U=BiB`0T=CjB?S=ClB=R=GlB:S=HlB8T=IjB8V=i00000000000O001N2N2N1O2N1O2N2N1O2N2N1O10O2O1OdChMR<W2oCiMQ<U2QDkMo;T2RDlMm;U2QDmMn;T2oCoMo;S2nCPNo;S2nCPNP<a2M3N2M4M2M3N2M3N2N2N2N2N2M3N2N3M2N2N2N2N1O2O1O1O001O1O10O010000O010mN`KfG`4Y8cKeG]4Y8gKdGY4\\8iKcGW4[8mKcGS4\\8oKbGR4\\8RLbGn3]8YL]Gf3b8\\1POTJmHl5R7PKRHQ5n7P10O10O10000O2O1O1N101O1O1N10000001O1O1O1O1O1N101O1O1O1O1O1O1O001N2O1O1O1N2O001N2O1O1N2O1O1N1VLaGT1`8kNkGj0V8TORHe0o7ZOXH>j7A[HfNnN`0i8h0[H]N\\Od0Z8m0]HYNCd0R8Q1]HTNJg0j7T1^HoM1g0d7W1RJeNo5Z1YJ^Ni5_1aJXN`5f1o3L3N3M2M4M2N3M2O2M2O2M2N3N1N3N1N3N5JZdf2"}, "label": "a woman wearing glasses and a red shirt sitting at a table"}]}
{"id": 16273, "image": "COCO_train2014_000000016273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older lady in red and glasses sitting a table listening and talking with friends\"."}], "task": "refering", "answer_template": "The \"older lady in red and glasses sitting a table listening and talking with friends\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 105, 106, 107, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 263, 264, 265, 266], "bbox": [0.45, 0.1479375, 0.711234375, 0.6827708333333333], "iscrowd": 0, "area": 21007.177900000006, "rle": {"size": [480, 640], "counts": "bYW49f>2N2N3M2N2^O]OeBd0X=_OfBc0W=@gBb0U=BiB`0T=CjB?S=ClB=R=GlB:S=HlB8T=IjB8V=i00000000000O001N2N2N1O2N1O2N2N1O2N2N1O10O2O1OdChMR<W2oCiMQ<U2QDkMo;T2RDlMm;U2QDmMn;T2oCoMo;S2nCPNo;S2nCPNP<a2M3N2M4M2M3N2M3N2N2N2N2N2M3N2N3M2N2N2N2N1O2O1O1O001O1O10O010000O010mN`KfG`4Y8cKeG]4Y8gKdGY4\\8iKcGW4[8mKcGS4\\8oKbGR4\\8RLbGn3]8YL]Gf3b8\\1POTJmHl5R7PKRHQ5n7P10O10O10000O2O1O1N101O1O1N10000001O1O1O1O1O1N101O1O1O1O1O1O1O001N2O1O1O1N2O001N2O1O1N2O1O1N1VLaGT1`8kNkGj0V8TORHe0o7ZOXH>j7A[HfNnN`0i8h0[H]N\\Od0Z8m0]HYNCd0R8Q1]HTNJg0j7T1^HoM1g0d7W1RJeNo5Z1YJ^Ni5_1aJXN`5f1o3L3N3M2M4M2N3M2O2M2O2M2N3N1N3N1N3N5JZdf2"}, "label": "older lady in red and glasses sitting a table listening and talking with friends"}]}
{"id": 434067, "image": "COCO_train2014_000000434067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench the man in red is sitting in\"."}], "task": "refering", "answer_template": "The \"the bench the man in red is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 169, 170, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 245, 257, 258, 268], "bbox": [0.101, 0.402387706855792, 0.793125, 0.8023877068557919], "iscrowd": 0, "area": 30072.204100000003, "rle": {"size": [423, 640], "counts": "dPk01V=6J6J7I7I6J8H<D;E9G2N2N2N2N2N2N2N2N2N2N2N2N000000000000000000000000000000000000000001O000001O0000000000000000000000000000]OlESNT:l1mETNS:k1nEUNR:j1oEVNQ:i1PFWNP:h1QFXNo9f1SFZNm9e1TF[Nl9d1UF\\Nk9c1VFfNa9Y1`FYOn8f0SGL[82gG`0g7_OZHf0a7YO`Hg0`7XOaHh0_7WObHi0^7VOcHj0]7YNVHE?R2[7UN^HD8W2Z7PNgHD0\\2Y7lMoHCIa2X7gMSJX2m5dMWJ\\2i5_M\\Ja2U8WOi0fNZ100O100000000O1000000O1000000000000O1000000000000000000O2O00000000000000000O2O001O1O1O1O001O1N2O00]c?0c\\@2N9G9G9G8H7I2N1O2N2N1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O3M2N3M2N2N3M2N2N00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O1000000000000000000000000004L<D<D<D<D=C6J00O100000000000000O1nMmGPOS8P1^H_Nb7`1PIoMP7Q2bI]M^6c2S200O10000000000O10000000000O1000000000002M5L5K4L4L4L5K4K5L4L7I8H7I7I0O101O00001O000O2O00001O00001N10001O00001O0O101O00001O00001N10001O00RYf1"}, "label": "the bench the man in red is sitting in"}]}
{"id": 434067, "image": "COCO_train2014_000000434067.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench that man is sitting on\"."}], "task": "refering", "answer_template": "The \"the bench that man is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 146, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 169, 170, 171, 172, 173, 174, 175, 176, 177, 186, 187, 188, 189, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 245, 257, 258, 268], "bbox": [0.101, 0.402387706855792, 0.793125, 0.8023877068557919], "iscrowd": 0, "area": 30072.204100000003, "rle": {"size": [423, 640], "counts": "dPk01V=6J6J7I7I6J8H<D;E9G2N2N2N2N2N2N2N2N2N2N2N2N000000000000000000000000000000000000000001O000001O0000000000000000000000000000]OlESNT:l1mETNS:k1nEUNR:j1oEVNQ:i1PFWNP:h1QFXNo9f1SFZNm9e1TF[Nl9d1UF\\Nk9c1VFfNa9Y1`FYOn8f0SGL[82gG`0g7_OZHf0a7YO`Hg0`7XOaHh0_7WObHi0^7VOcHj0]7YNVHE?R2[7UN^HD8W2Z7PNgHD0\\2Y7lMoHCIa2X7gMSJX2m5dMWJ\\2i5_M\\Ja2U8WOi0fNZ100O100000000O1000000O1000000000000O1000000000000000000O2O00000000000000000O2O001O1O1O1O001O1N2O00]c?0c\\@2N9G9G9G8H7I2N1O2N2N1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O3M2N3M2N2N3M2N2N00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000O1000000000000000000000000004L<D<D<D<D=C6J00O100000000000000O1nMmGPOS8P1^H_Nb7`1PIoMP7Q2bI]M^6c2S200O10000000000O10000000000O1000000000002M5L5K4L4L4L5K4K5L4L7I8H7I7I0O101O00001O000O2O00001O00001N10001O00001O0O101O00001O00001N10001O00RYf1"}, "label": "the bench that man is sitting on"}]}
{"id": 335766, "image": "COCO_train2014_000000335766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a faucet sits above the sink in the forground\"."}], "task": "refering", "answer_template": "The \"a faucet sits above the sink in the forground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 256, 257, 279, 280, 302, 303, 325, 326, 327, 328, 329], "bbox": [0.15632812499999998, 0.6727336448598131, 0.350875, 0.9870093457943925], "iscrowd": 0, "area": 4529.565500000002, "rle": {"size": [428, 640], "counts": "`RZ1:R=e0[Oe0\\Od0[Oe0[O7I5K5_FfLg8^3mFiLS9h3000000001O000000010O00000fMhF`0X9PNVGi0Io0Q9XNXGh0Mi0l8^NYGh02b0e8gNYGg08;_8nN[Gf0=4X8VO]Ge0a0NR8]O_Gd0f0Gl7D`Gd0j0Af7KaGd0n0[Oa71cGa0l0_Oa70eG>k0B`70gG<i0Da70gG9i0G`70hG8h0H`70jG5f0L`7OlG2e0O_7OnG0c01_7OPHMb04_7NPHM`06`7MQJ3o5MQJ3o5NPJ2P6NPJ2Q6MnI4R6LnI4S6KmI5T6JlI6V6HiI9X6FhI:e900000O1000000O100000000O1000000O1000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O0000000000000O1M3M3MlQ]5"}, "label": "a faucet sits above the sink in the forground"}]}
{"id": 335766, "image": "COCO_train2014_000000335766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the faucet on the left sink\"."}], "task": "refering", "answer_template": "The \"the faucet on the left sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [233, 234, 235, 256, 257, 279, 280, 302, 303, 325, 326, 327, 328, 329], "bbox": [0.15632812499999998, 0.6727336448598131, 0.350875, 0.9870093457943925], "iscrowd": 0, "area": 4529.565500000002, "rle": {"size": [428, 640], "counts": "`RZ1:R=e0[Oe0\\Od0[Oe0[O7I5K5_FfLg8^3mFiLS9h3000000001O000000010O00000fMhF`0X9PNVGi0Io0Q9XNXGh0Mi0l8^NYGh02b0e8gNYGg08;_8nN[Gf0=4X8VO]Ge0a0NR8]O_Gd0f0Gl7D`Gd0j0Af7KaGd0n0[Oa71cGa0l0_Oa70eG>k0B`70gG<i0Da70gG9i0G`70hG8h0H`70jG5f0L`7OlG2e0O_7OnG0c01_7OPHMb04_7NPHM`06`7MQJ3o5MQJ3o5NPJ2P6NPJ2Q6MnI4R6LnI4S6KmI5T6JlI6V6HiI9X6FhI:e900000O1000000O100000000O1000000O1000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O0000000000000O1M3M3MlQ]5"}, "label": "the faucet on the left sink"}]}
{"id": 40855, "image": "COCO_train2014_000000040855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female wearing a red shirt is holding a blue cell phone in an outdoor setting\"."}], "task": "refering", "answer_template": "The \"a female wearing a red shirt is holding a blue cell phone in an outdoor setting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 278, 279, 284, 285, 286, 287, 295, 296, 301, 302, 305, 306, 307, 308, 309, 310, 318, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 370, 371], "bbox": [0.09953125, 0.09112704918032787, 0.8883281249999999, 0.9531967213114754], "iscrowd": 0, "area": 55004.60185, "rle": {"size": [488, 640], "counts": "ckn0:h>?Bb0bBiNk;c2I6J6I7J6I6K1O1N2O1000O2O1O1O001O1O1O007H8I7I5K2N1O1N3N1O2M2O2N1N3N1O1N101O1N2O1O1N101O1N2O1O1N101O0O10000O2O000O100O2O000O10001N10000O101O0000000000O10000O10000O1000000O10000O10000O1000000O100O10000O100O10000O100O010O01000O010O0100O0100O0100O010O10O10O10O010O10000O10000O10000O1000000O1000000O1000000O1000000O1000000O1O1O1O100O1O1O1O1O100O1O1O1O10000000000000000000000000000000000O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O10000O101N100O10000O100O100O100000000001O00001O0000001O0000001O000HjC^MW<o1\\DPNd;]1]1^Ob0^OZRV1G\\niN;F:F:E9H5K5J6K6I6L4K5L5K4L5J5L5K5J6VMcLbIb3Y6aLjG=b0V3PO`LP80\\Hl5Z7lIYH<:j5Y7bJeH_5W7fJfH\\5V7iJgHX5V7mJgHU5V7oJhHR5W7`1N2N2N2O1N2N2N2O1N2N2O1N2N1O2O1O100O10000O1000000O10000O10O10O1000000O10000O10000O1000000O01000O10000O100001O0000000000000000000000000O100000000000001O00002N1O2O001O001jIaI`3_6]LdIc3]6XLgIh3Y6ULjIj3X6RLkIn3U6oKnIQ4S6kKPJU4P6gKTJY4m5cKVJ]4j5`KYJ`4h5\\K[Jd4e5YK^Jg4c5UK`Jl4_5QKdJR5Z5jJiJY5b73M2N3M3M3M3M3M3N2O100O1O100OO1I8I6J6I8H7I7I8H7I7I8H7I7I8H;A?@`0A?@`0^Oc0dNXcQ1"}, "label": "a female wearing a red shirt is holding a blue cell phone in an outdoor setting"}]}
{"id": 40855, "image": "COCO_train2014_000000040855.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman sitting with her her black purse and water bottle beside her , and she is holding a blue cell phone up to her ear\"."}], "task": "refering", "answer_template": "The \"the woman sitting with her her black purse and water bottle beside her , and she is holding a blue cell phone up to her ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 107, 108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 278, 279, 284, 285, 286, 287, 295, 296, 301, 302, 305, 306, 307, 308, 309, 310, 318, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 341, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 370, 371], "bbox": [0.09953125, 0.09112704918032787, 0.8883281249999999, 0.9531967213114754], "iscrowd": 0, "area": 55004.60185, "rle": {"size": [488, 640], "counts": "ckn0:h>?Bb0bBiNk;c2I6J6I7J6I6K1O1N2O1000O2O1O1O001O1O1O007H8I7I5K2N1O1N3N1O2M2O2N1N3N1O1N101O1N2O1O1N101O1N2O1O1N101O0O10000O2O000O100O2O000O10001N10000O101O0000000000O10000O10000O1000000O10000O10000O1000000O100O10000O100O10000O100O010O01000O010O0100O0100O0100O010O10O10O10O010O10000O10000O10000O1000000O1000000O1000000O1000000O1000000O1O1O1O100O1O1O1O1O100O1O1O1O10000000000000000000000000000000000O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O10000O101N100O10000O100O100O100000000001O00001O0000001O0000001O000HjC^MW<o1\\DPNd;]1]1^Ob0^OZRV1G\\niN;F:F:E9H5K5J6K6I6L4K5L5K4L5J5L5K5J6VMcLbIb3Y6aLjG=b0V3PO`LP80\\Hl5Z7lIYH<:j5Y7bJeH_5W7fJfH\\5V7iJgHX5V7mJgHU5V7oJhHR5W7`1N2N2N2O1N2N2N2O1N2N2O1N2N1O2O1O100O10000O1000000O10000O10O10O1000000O10000O10000O1000000O01000O10000O100001O0000000000000000000000000O100000000000001O00002N1O2O001O001jIaI`3_6]LdIc3]6XLgIh3Y6ULjIj3X6RLkIn3U6oKnIQ4S6kKPJU4P6gKTJY4m5cKVJ]4j5`KYJ`4h5\\K[Jd4e5YK^Jg4c5UK`Jl4_5QKdJR5Z5jJiJY5b73M2N3M3M3M3M3M3N2O100O1O100OO1I8I6J6I8H7I7I8H7I7I8H7I7I8H;A?@`0A?@`0^Oc0dNXcQ1"}, "label": "the woman sitting with her her black purse and water bottle beside her , and she is holding a blue cell phone up to her ear"}]}
{"id": 57242, "image": "COCO_train2014_000000057242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"floral pattern blanket below everything\"."}], "task": "refering", "answer_template": "The \"floral pattern blanket below everything\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 21, 22, 23, 24, 25, 40, 41, 42, 43, 44, 59, 60, 61, 62, 79, 80, 81, 98, 99, 137, 138, 144, 145, 146, 154, 155, 156, 157, 162, 163, 164, 165, 171, 172, 173, 180, 181, 182, 183, 184, 189, 190, 191, 198, 199, 200, 201, 202, 203, 204, 206, 207, 208, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.0019399999999999999, 0.0, 0.74416, 0.9844266666666668], "iscrowd": 0, "area": 36290.6167, "rle": {"size": [375, 500], "counts": "Pc0T2c9T2lM00000000000001O00000000000000001O00000000000000001O00000000000000001O001O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1N2O001O1ORHbMd6]2YIgMf6Y2VIlMj6S2SIQNm6n1PIVNP7i1lH\\NS7d1jH`NV7_1gHeNY7Z1\\HROd7m0\\HTOc7l0]HUOc7j0]HWOc7h0]HYOc7f0]H[Ob7e0^H\\Ob7c0^H^Ob7a0^H@b7?^HBa7>_HCa7<_HEa7:_HGa78_HI`77`HJ`75aHK_74aHM_72aHO_70aH1^7ObH2^7MbH4^7KbH6]7JcH6^7IbH8^7GbH:]7FcH;]7DcH=]7BcH?\\7AdH`0\\7_OdHb0\\7]OdHd0[7\\OeHe0[7ZOeHg0[7XOeHi0Z7WOfHj0Z7UOfHl0Z7SOfHn0Y7ROgHo0Y7POgHQ1Y7nNgHS1X7mNhHX1T7gNlH`1n6_NRIb1m6^NSIb1n6]NSIc1m6\\NSId1m6\\NSIe1m6ZNSIg1m6XNSIh1m6XNSIi1m6VNSIj1n6UNRIl1m6TNSIm1m6RNSIn1m6RNSIo1m6PNSIP2n6oMRIR2m6nMSIR2n6mMRIT2n6kMRIV2m6jMSIV2n6iMRIX2n6gMRIY2n6gMRIZ2n6eMRI\\2n6cMRI]2n6cMRI^2n6aMRI_2o6`MQIa2n6_MRIb2n6]MRIc2o6\\MQIe2n6[MRIe2o6ZMQIg2o6XMRIh2m6XMSIh2n6WMRIj2n6UMRIk2n6UMRIl2n6SMRIm2o6RMQIo2n6QMRIP3n6oLRIQ3o6nLQIS3n6mLRIS3o6lLQIU3n6kLRIV3n6iLRIW3o6hLQIY3n6gLRIY3o6fLQI[3o6dLQI]3n6cLRI]3o6bLQI_3o6`LQI`3o6`LQIa3n6_LRIb3m6^LSIb3m6^LSIc3l6]LTIc3l6]LTId0Ll1o6`MUI`03m1h6cMUI;;P2_6eMVI6d0R2U6hMXI0k0U2m5kMXIKS1X2d5mMoJP2Q5PNRKn1m4RNWKk1f4WN]Kf1b4[N`Kd1]4^NfK_1X4cNjK\\1T4eNQLW1l3kNoL:P3G[MOb23]MNa24aMK]26dMJY28hMGW2:jMFS2<nMCQ2>PNBm1`0TN@i1b0XN]Og1d0[N[Ob1g0_NXO`1i0aNWO\\1k0dNVOY1k0hNUOW1l0iNUOT1m0lNSOS1n0mNSOP1o0POROm0P1TOoNk0R1UOoNh0S1XOmNg0T1YOmNd0U1\\OlNa0V1_OjN`0W1@jN=X1ChN<Y1EgN8[1HfN6[1JeN4]1LdN1^1ObN0_10bNM`13aNK`15`NIb18^NEd1;\\NDe1<\\NAf1?[N_Of1a0ZN]Oh1c0_MeKe0e3n1f0[MiKe0`3Q2Q1mMlNU2W1iMgNX2[1fMcN\\2`1bM]N`2f1]MYNd2i1[MTNg2o1WMoMj2S2TMkMn2X2PMfMQ3]2lLaMV3^52O1N2O1N00O2O001O0O101O001O0O101O001N10002N3L4M3M3M2N3L4M3M3M3L]OSNcGj1]8ZNaGd1_8`N_G^1a8fN\\GY1d8kNZGS1i8nNTGR1n8oNPGo0S9ROjFm0Y9f02N3M3M3M3M3M2O2M3M3M3M3M3M2N3M3M3M3N2M2N3M3O1O1O1O010O1O100O1O10O01O1O100O1O010O1O100O1O010O1O1O10ho^1"}, "label": "floral pattern blanket below everything"}]}
{"id": 57242, "image": "COCO_train2014_000000057242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flowered tablecloth under the pizza box\"."}], "task": "refering", "answer_template": "The \"flowered tablecloth under the pizza box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 21, 22, 23, 24, 25, 40, 41, 42, 43, 44, 59, 60, 61, 62, 79, 80, 81, 98, 99, 137, 138, 144, 145, 146, 154, 155, 156, 157, 162, 163, 164, 165, 171, 172, 173, 180, 181, 182, 183, 184, 189, 190, 191, 198, 199, 200, 201, 202, 203, 204, 206, 207, 208, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.0019399999999999999, 0.0, 0.74416, 0.9844266666666668], "iscrowd": 0, "area": 36290.6167, "rle": {"size": [375, 500], "counts": "Pc0T2c9T2lM00000000000001O00000000000000001O00000000000000001O00000000000000001O001O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1N2O001O1ORHbMd6]2YIgMf6Y2VIlMj6S2SIQNm6n1PIVNP7i1lH\\NS7d1jH`NV7_1gHeNY7Z1\\HROd7m0\\HTOc7l0]HUOc7j0]HWOc7h0]HYOc7f0]H[Ob7e0^H\\Ob7c0^H^Ob7a0^H@b7?^HBa7>_HCa7<_HEa7:_HGa78_HI`77`HJ`75aHK_74aHM_72aHO_70aH1^7ObH2^7MbH4^7KbH6]7JcH6^7IbH8^7GbH:]7FcH;]7DcH=]7BcH?\\7AdH`0\\7_OdHb0\\7]OdHd0[7\\OeHe0[7ZOeHg0[7XOeHi0Z7WOfHj0Z7UOfHl0Z7SOfHn0Y7ROgHo0Y7POgHQ1Y7nNgHS1X7mNhHX1T7gNlH`1n6_NRIb1m6^NSIb1n6]NSIc1m6\\NSId1m6\\NSIe1m6ZNSIg1m6XNSIh1m6XNSIi1m6VNSIj1n6UNRIl1m6TNSIm1m6RNSIn1m6RNSIo1m6PNSIP2n6oMRIR2m6nMSIR2n6mMRIT2n6kMRIV2m6jMSIV2n6iMRIX2n6gMRIY2n6gMRIZ2n6eMRI\\2n6cMRI]2n6cMRI^2n6aMRI_2o6`MQIa2n6_MRIb2n6]MRIc2o6\\MQIe2n6[MRIe2o6ZMQIg2o6XMRIh2m6XMSIh2n6WMRIj2n6UMRIk2n6UMRIl2n6SMRIm2o6RMQIo2n6QMRIP3n6oLRIQ3o6nLQIS3n6mLRIS3o6lLQIU3n6kLRIV3n6iLRIW3o6hLQIY3n6gLRIY3o6fLQI[3o6dLQI]3n6cLRI]3o6bLQI_3o6`LQI`3o6`LQIa3n6_LRIb3m6^LSIb3m6^LSIc3l6]LTIc3l6]LTId0Ll1o6`MUI`03m1h6cMUI;;P2_6eMVI6d0R2U6hMXI0k0U2m5kMXIKS1X2d5mMoJP2Q5PNRKn1m4RNWKk1f4WN]Kf1b4[N`Kd1]4^NfK_1X4cNjK\\1T4eNQLW1l3kNoL:P3G[MOb23]MNa24aMK]26dMJY28hMGW2:jMFS2<nMCQ2>PNBm1`0TN@i1b0XN]Og1d0[N[Ob1g0_NXO`1i0aNWO\\1k0dNVOY1k0hNUOW1l0iNUOT1m0lNSOS1n0mNSOP1o0POROm0P1TOoNk0R1UOoNh0S1XOmNg0T1YOmNd0U1\\OlNa0V1_OjN`0W1@jN=X1ChN<Y1EgN8[1HfN6[1JeN4]1LdN1^1ObN0_10bNM`13aNK`15`NIb18^NEd1;\\NDe1<\\NAf1?[N_Of1a0ZN]Oh1c0_MeKe0e3n1f0[MiKe0`3Q2Q1mMlNU2W1iMgNX2[1fMcN\\2`1bM]N`2f1]MYNd2i1[MTNg2o1WMoMj2S2TMkMn2X2PMfMQ3]2lLaMV3^52O1N2O1N00O2O001O0O101O001O0O101O001N10002N3L4M3M3M2N3L4M3M3M3L]OSNcGj1]8ZNaGd1_8`N_G^1a8fN\\GY1d8kNZGS1i8nNTGR1n8oNPGo0S9ROjFm0Y9f02N3M3M3M3M3M2O2M3M3M3M3M3M2N3M3M3M3N2M2N3M3O1O1O1O010O1O100O1O10O01O1O100O1O010O1O100O1O010O1O1O10ho^1"}, "label": "flowered tablecloth under the pizza box"}]}
{"id": 57242, "image": "COCO_train2014_000000057242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the part of the bedspread that is to the right of sewing machine and above the blue and white thread\"."}], "task": "refering", "answer_template": "The \"the part of the bedspread that is to the right of sewing machine and above the blue and white thread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 51, 52, 53, 69, 70, 71, 88, 89, 105, 106, 107, 123, 124, 125, 141, 142, 143, 159, 160, 161, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.74806, 0.0077866666666666666, 0.9974, 0.9844266666666668], "iscrowd": 0, "area": 26766.21445, "rle": {"size": [375, 500], "counts": "ZZY41c;3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M2O2M3N2M3N2O1N2O1O1N2O1O1N2O1O1O1_MnL[LT3d3oLYLR3f3PMXLQ3g3RMULP3j3RMTLo2k3TMRLm2m3UMPLm2[1eLZOa0XOk2[1iLZO?XOi2[1lL[O=XOh2Z1oL[O=WOf2[1RM[O:XOe2mNhLf1=39WOc2POlL]1=:7VOa2SOPMU1=?4VOa2VOSMm0<e03UO_2YOXMd0:l01UO^2[O\\M<9R10TO\\2^O`M49W1MTO\\2ABY1TNTO[2CBW1UNSO[2FAU1WNPO[2K@R1WNPO[2N_OP1XNoN[21^Om0[NnNX26^Oi0]NmNX2:\\Ob0o2^ORM<S3DoL4W3LjLN[32fLG`39o30000000000000000000000000000O10000000000000000000000000000000000000O1000000000000000000000000RHbc0"}, "label": "the part of the bedspread that is to the right of sewing machine and above the blue and white thread"}]}
{"id": 442268, "image": "COCO_train2014_000000442268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy sitting on a park bench\"."}], "task": "refering", "answer_template": "The \"a little boy sitting on a park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146, 149, 150, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281, 282, 296, 297, 298, 299, 300, 314, 315, 316, 317, 318, 332, 333, 334, 335, 349, 350, 351, 366, 367, 368, 384], "bbox": [0.2486875, 0.10300000000000001, 0.8838958333333333, 0.97265625], "iscrowd": 0, "area": 74061.37455, "rle": {"size": [640, 480], "counts": "^fZ21jc09G8H9F9H9H8J5L5J5L5J3M2O2M3N1N3N2M3M2O2M3N2M2N3N2M3N1N3M3N1N3N2M2N3N1N3N2[LPLVGR4c8YLWGj3a8aLZG`3`8jLZGY3^8QM^GP3[8[M_Gh2^8^M\\Gd2c8`MYGb2d8aMZG`2e8cMXG_2e8eMXG]2e8gMXGZ2g8hMXGY2e8kMXGW2BWKc7e2hHV2^O`Kc7^2lHU2XOhKg7U2oHV2QOPLk7m1QIV2lNVLo7g1RIo3l6TLQIm3n6ULPIm3n6VLoHl3o6WLnHk3o6XLoHi3P7ZLmHh3Q7[LlHg3R7\\LkHe3T7]LjHe3T7^LiHd3U7_LhHb3W7bLeH`3Y7cLeH^3Y7fLcH[3\\7f4N2N1O2O1O001O1O1O001O1O001O001O1O0O2O001O1O001O1O001O001O10O01O001O1O010O001O1O010O1O0010O0100O0010O0100O010O00100O010O010O10O0100O010O010O101N2O1O1N2O2M2O1N2O1N3N1N3dKiAj2Y>kLRBS3o=cL\\BZ3g=[LdBc3]=SLnBk3f>N2N2N2N[K\\LmHc3R7`LlHa3R7aLmH`3_OdLY3MWM`3QOQMg3BVM^3cN`MU4TOXM\\3]NfMZ4POWM[3ZNiM]ObNj2<^OY3VNoMAeNa25GX3RNRNFgN[21KW3oMVNJgN\\N]OX3a0c0U3jMZNOWOa1\\Oe0T3fM^N3TO`1[Of0^3XMWNb0RO]1ZOi0m3dLlMU1oN[1YOl0\\4PL`Mi1kN[1ZOl0X7hM^I[1[Ok0Z7iM[I[1\\Ok0Z7jMZIZ1]Oj0]7jMVI\\1^Oh0_7jMSI]1_Og0b7jMoH^1@g0e7hMkHa1@f0h7gMhHb1Bd0j7hMdHc1Cd0l7gMaHe1Cb0P8gM]Hf1Ea0Q8gMZHg1F`0T8gMVHi1F?V8gMTHi1G>W8iMRHh1I=W8jMPHh1J<X8lMnGh1J;Y8mMmGg1L9Y8PNkGf1M9Z8PNiGg1M7\\8RNgGf1N7\\8SNfGe105\\8VNdGd115]8VNbGd133]8YN`Gc143]8ZN_Ga172\\8]N]G`182[8^N^G_190[8aN[G_1;O[8aN[G_1<M[8dNYG]1>NZ8eNXG\\1`0LZ8gNWG\\1`0LZ8hNVG[1b0JZ8jNUGZ1c0KY8kNTGY1e0IX8oNSGW1g0HW8POSGW1g0GX8ROQGV1i0FW8SOQGU1j0FW8UOoFT1l0EU8WOPGS1l0DV8YOnFR1n0CU8[OmFQ1o0BV8\\OlFP1Q1BS8_OlFn0R1AT8@lFm0R1AS8BkFl0U1^OR8EjFl0W1[OP8JjFg0Z1\\On7MhFe0^1ZOl70gFd0`1YOj74gF`0c1XOh77fF>f1WOg7;cF<j1UOe7>cF:l1TOb7c0bF6P2TO`7f0`F4T2RO]7j0aF1U2RO\\7l0`FOm=1UBKl=5]BAe=>cBXO`=h0k201O010O1O010O10O01OZO[OR^Of0ma0Am]O>Tb0Hf]O7[b0g000O1000O10O1000000O1000000O10O100000000000000000000O1000000000001O0O10000O10001N101N5I6J7J5J6K6IWhR1"}, "label": "a little boy sitting on a park bench"}]}
{"id": 442268, "image": "COCO_train2014_000000442268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby\"."}], "task": "refering", "answer_template": "The \"a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146, 149, 150, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 249, 250, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281, 282, 296, 297, 298, 299, 300, 314, 315, 316, 317, 318, 332, 333, 334, 335, 349, 350, 351, 366, 367, 368, 384], "bbox": [0.2486875, 0.10300000000000001, 0.8838958333333333, 0.97265625], "iscrowd": 0, "area": 74061.37455, "rle": {"size": [640, 480], "counts": "^fZ21jc09G8H9F9H9H8J5L5J5L5J3M2O2M3N1N3N2M3M2O2M3N2M2N3N2M3N1N3M3N1N3N2M2N3N1N3N2[LPLVGR4c8YLWGj3a8aLZG`3`8jLZGY3^8QM^GP3[8[M_Gh2^8^M\\Gd2c8`MYGb2d8aMZG`2e8cMXG_2e8eMXG]2e8gMXGZ2g8hMXGY2e8kMXGW2BWKc7e2hHV2^O`Kc7^2lHU2XOhKg7U2oHV2QOPLk7m1QIV2lNVLo7g1RIo3l6TLQIm3n6ULPIm3n6VLoHl3o6WLnHk3o6XLoHi3P7ZLmHh3Q7[LlHg3R7\\LkHe3T7]LjHe3T7^LiHd3U7_LhHb3W7bLeH`3Y7cLeH^3Y7fLcH[3\\7f4N2N1O2O1O001O1O1O001O1O001O001O1O0O2O001O1O001O1O001O001O10O01O001O1O010O001O1O010O1O0010O0100O0010O0100O010O00100O010O010O10O0100O010O010O101N2O1O1N2O2M2O1N2O1N3N1N3dKiAj2Y>kLRBS3o=cL\\BZ3g=[LdBc3]=SLnBk3f>N2N2N2N[K\\LmHc3R7`LlHa3R7aLmH`3_OdLY3MWM`3QOQMg3BVM^3cN`MU4TOXM\\3]NfMZ4POWM[3ZNiM]ObNj2<^OY3VNoMAeNa25GX3RNRNFgN[21KW3oMVNJgN\\N]OX3a0c0U3jMZNOWOa1\\Oe0T3fM^N3TO`1[Of0^3XMWNb0RO]1ZOi0m3dLlMU1oN[1YOl0\\4PL`Mi1kN[1ZOl0X7hM^I[1[Ok0Z7iM[I[1\\Ok0Z7jMZIZ1]Oj0]7jMVI\\1^Oh0_7jMSI]1_Og0b7jMoH^1@g0e7hMkHa1@f0h7gMhHb1Bd0j7hMdHc1Cd0l7gMaHe1Cb0P8gM]Hf1Ea0Q8gMZHg1F`0T8gMVHi1F?V8gMTHi1G>W8iMRHh1I=W8jMPHh1J<X8lMnGh1J;Y8mMmGg1L9Y8PNkGf1M9Z8PNiGg1M7\\8RNgGf1N7\\8SNfGe105\\8VNdGd115]8VNbGd133]8YN`Gc143]8ZN_Ga172\\8]N]G`182[8^N^G_190[8aN[G_1;O[8aN[G_1<M[8dNYG]1>NZ8eNXG\\1`0LZ8gNWG\\1`0LZ8hNVG[1b0JZ8jNUGZ1c0KY8kNTGY1e0IX8oNSGW1g0HW8POSGW1g0GX8ROQGV1i0FW8SOQGU1j0FW8UOoFT1l0EU8WOPGS1l0DV8YOnFR1n0CU8[OmFQ1o0BV8\\OlFP1Q1BS8_OlFn0R1AT8@lFm0R1AS8BkFl0U1^OR8EjFl0W1[OP8JjFg0Z1\\On7MhFe0^1ZOl70gFd0`1YOj74gF`0c1XOh77fF>f1WOg7;cF<j1UOe7>cF:l1TOb7c0bF6P2TO`7f0`F4T2RO]7j0aF1U2RO\\7l0`FOm=1UBKl=5]BAe=>cBXO`=h0k201O010O1O010O10O01OZO[OR^Of0ma0Am]O>Tb0Hf]O7[b0g000O1000O10O1000000O1000000O10O100000000000000000000O1000000000001O0O10000O10001N101N5I6J7J5J6K6IWhR1"}, "label": "a baby"}]}
{"id": 130972, "image": "COCO_train2014_000000130972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman wearing black , sitting at a table next to the window\"."}], "task": "refering", "answer_template": "The \"a blonde woman wearing black , sitting at a table next to the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 86, 87, 88, 108, 109, 110, 131, 132, 133, 134, 154, 155, 156, 157, 158, 178, 179, 180, 181], "bbox": [0.70746875, 0.16285416666666666, 0.899234375, 0.43985416666666666], "iscrowd": 0, "area": 9675.815849999999, "rle": {"size": [480, 640], "counts": "l`d65e>7I7I7I6J7I6L4L4H8K401O100O1O1O1M3I9L3M3M4K3N3N2M3M3M2O1G9N3N1N2O1N2O2N1000000000000000O1000000000000000000000000000000000000000000000001O0hMXE8h:HZE6g:H\\E6d:I^E6c:H_E7a:HaE7`:YOPFf0R:XOoEg0T:UOoEi0U:QOnEn0U:nNmEQ1V:jNmEU1W:_NRFa1g;O1O100O100O1O1O100O1O1O1O1M3O001O1O1O001O1O1O1O1O1O00100O001O002N2N1O2M3N2M2N3GPjm0"}, "label": "a blonde woman wearing black , sitting at a table next to the window"}]}
{"id": 130972, "image": "COCO_train2014_000000130972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde haired woman wearing a black tank top is sitting at a table\"."}], "task": "refering", "answer_template": "The \"a blonde haired woman wearing a black tank top is sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 86, 87, 88, 108, 109, 110, 131, 132, 133, 134, 154, 155, 156, 157, 158, 178, 179, 180, 181], "bbox": [0.70746875, 0.16285416666666666, 0.899234375, 0.43985416666666666], "iscrowd": 0, "area": 9675.815849999999, "rle": {"size": [480, 640], "counts": "l`d65e>7I7I7I6J7I6L4L4H8K401O100O1O1O1M3I9L3M3M4K3N3N2M3M3M2O1G9N3N1N2O1N2O2N1000000000000000O1000000000000000000000000000000000000000000000001O0hMXE8h:HZE6g:H\\E6d:I^E6c:H_E7a:HaE7`:YOPFf0R:XOoEg0T:UOoEi0U:QOnEn0U:nNmEQ1V:jNmEU1W:_NRFa1g;O1O100O100O1O1O100O1O1O1O1M3O001O1O1O001O1O1O1O1O1O00100O001O002N2N1O2M3N2M2N3GPjm0"}, "label": "a blonde haired woman wearing a black tank top is sitting at a table"}]}
{"id": 130972, "image": "COCO_train2014_000000130972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the picture of the man ' s face and neck and shoulder who is eating the hot dog , but not his hands\"."}], "task": "refering", "answer_template": "The \"the picture of the man ' s face and neck and shoulder who is eating the hot dog , but not his hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.004020833333333333, 0.42112499999999997, 0.9746041666666667], "iscrowd": 0, "area": 101903.82174999997, "rle": {"size": [480, 640], "counts": "V=]1k:h2WMi2WMi2XMh2RNn10000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000001O000000000000000000000000000O100000000000000000000000QOFXC:g<IWC7g<MWC3h<1UCOj<4TCLk<7SCIl<;QCEn<>PCBn<b0PC_On<d0PC\\Oo<h0nBXOQ=k0mBVOQ=]1O1N2O2N1O1O2M2O1O2N1N2O1O2M2O1O2M2O1O1N3N1O2M2O1O2M2O1O2N1N3N1O1O2N1N2O2N1O1N3M2M4L3M3N3O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2OR1nN6J5K5K6I6K6J5K6J4\\M_HaMc7Y2bHfM`7U2eHhM^7S2fHlM\\7o1hHPNZ7k1kHSNW7h1mHWNU7c1QI[NQ7`1SI_No6\\1UIcNm6X1XIfNj6W1WIiNk6T1WIkNk6R1VInNl6o0VIPOj6o0WIQOj6m0XIROh6m0YISOh6k0YIUOg6i0\\IVOe6h0\\IXOd6g0^IXOc6f0^IZOc6d0_I[Oa6d0`I\\Oa6>dIB\\6:hIFY64lILT6NRJ2o5HVJ8j5FXJ:i5DXJ<h5BZJ>g5@ZJ`0f5_O[J`0g5]O[Jb0f5]O[Ja0h5\\OZJc0g5\\OZJc0n5UO^J?i5YOeJ:c5^OiJ6]:CZS]5"}, "label": "the picture of the man ' s face and neck and shoulder who is eating the hot dog , but not his hands"}]}
{"id": 130972, "image": "COCO_train2014_000000130972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man eating hotdog\"."}], "task": "refering", "answer_template": "The \"man eating hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.004020833333333333, 0.42112499999999997, 0.9746041666666667], "iscrowd": 0, "area": 101903.82174999997, "rle": {"size": [480, 640], "counts": "V=]1k:h2WMi2WMi2XMh2RNn10000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000001O000000000000000000000000000O100000000000000000000000QOFXC:g<IWC7g<MWC3h<1UCOj<4TCLk<7SCIl<;QCEn<>PCBn<b0PC_On<d0PC\\Oo<h0nBXOQ=k0mBVOQ=]1O1N2O2N1O1O2M2O1O2N1N2O1O2M2O1O2M2O1O1N3N1O2M2O1O2M2O1O2N1N3N1O1O2N1N2O2N1O1N3M2M4L3M3N3O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2OR1nN6J5K5K6I6K6J5K6J4\\M_HaMc7Y2bHfM`7U2eHhM^7S2fHlM\\7o1hHPNZ7k1kHSNW7h1mHWNU7c1QI[NQ7`1SI_No6\\1UIcNm6X1XIfNj6W1WIiNk6T1WIkNk6R1VInNl6o0VIPOj6o0WIQOj6m0XIROh6m0YISOh6k0YIUOg6i0\\IVOe6h0\\IXOd6g0^IXOc6f0^IZOc6d0_I[Oa6d0`I\\Oa6>dIB\\6:hIFY64lILT6NRJ2o5HVJ8j5FXJ:i5DXJ<h5BZJ>g5@ZJ`0f5_O[J`0g5]O[Jb0f5]O[Ja0h5\\OZJc0g5\\OZJc0n5UO^J?i5YOeJ:c5^OiJ6]:CZS]5"}, "label": "man eating hotdog"}]}
{"id": 343968, "image": "COCO_train2014_000000343968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"shortere giraffe\"."}], "task": "refering", "answer_template": "The \"shortere giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 53, 54, 67, 68, 69, 82, 83, 84, 97, 98, 99, 112, 113, 127, 128, 142, 143, 157, 158, 171, 172, 173, 186, 187, 188, 201, 202, 203, 215, 216, 217, 218, 229, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 272, 273, 274, 275, 276, 277, 278, 279, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 317, 318, 319, 320, 321, 322, 323, 333, 334, 335, 336, 337, 338], "bbox": [0.18600461893764436, 0.116484375, 0.6609699769053118, 1.0], "iscrowd": 0, "area": 52416.71305000001, "rle": {"size": [640, 433], "counts": "lTc1j1Va0V1F9F:G9F:G9F;F7H6K4M3O1O100O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1OPOQ12M2N3M3M2N3M3M2N3M3N1N3M2N3M3M2N3M3M2N3N0O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1M3M3L4M3M3L4M3L4M3M3L4M3L4M3kIeFb1^9]NcFa1_9^NbF`1`9`N`F^1a9bN`F\\1_9fNbFX1]9jNdFT1\\9mNeFQ1Z9QOgFm0X9UOiFi0W9XOkFe0T9RKbFm3:o0T9PKiFo34o0R9oJPGQ4On0P9nJWGT4Il0j9SOWFk0k9UOUFi0m9VOSFi0o9VORFh0P:XOoEg0T:WOmEg0U:XOlEf0V:ZOiEe0Y:ZOhEd0Z:\\OeEc0]:\\OdE8h:GXEMU;3kD@c;>^DTOP<k0PDiN]<W1cC\\Nj<c1VCQNW=n1jBeMc=[2]BXMP>g2PBmL]>n1VA[N>YOk>[2k@WNQ`0h1T@TNl?k1Y@PNh?o1^@kMc?T2b@gM_?X2f@dMZ?[2V101N13L6I6J7I5J3L5K4L5K5K4L5lJSLoIQ4\\5gL]J^3l4\\MlJh2^4RNXKU2Q4dNbKf1<nL^MV2i1?X1W3[NRM^1Z3UNoLd1[3oMnLk1]3hMlLP2_3cMjLW2a3\\MhL\\2c3XMeLe2]:03N2M4M2M4L4RN]^OU1ha0]Nd^Oa1oa0N2UOf]OM[b03T^O\\Ooa0b0i0O1N2O1O1N2M3M3N2M2NRhk2"}, "label": "shortere giraffe"}]}
{"id": 343968, "image": "COCO_train2014_000000343968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe facing other animals\"."}], "task": "refering", "answer_template": "The \"giraffe facing other animals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 53, 54, 67, 68, 69, 82, 83, 84, 97, 98, 99, 112, 113, 127, 128, 142, 143, 157, 158, 171, 172, 173, 186, 187, 188, 201, 202, 203, 215, 216, 217, 218, 229, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 272, 273, 274, 275, 276, 277, 278, 279, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 317, 318, 319, 320, 321, 322, 323, 333, 334, 335, 336, 337, 338], "bbox": [0.18600461893764436, 0.116484375, 0.6609699769053118, 1.0], "iscrowd": 0, "area": 52416.71305000001, "rle": {"size": [640, 433], "counts": "lTc1j1Va0V1F9F:G9F:G9F;F7H6K4M3O1O100O2N1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1OPOQ12M2N3M3M2N3M3M2N3M3N1N3M2N3M3M2N3M3M2N3N0O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1M3M3L4M3M3L4M3L4M3M3L4M3L4M3kIeFb1^9]NcFa1_9^NbF`1`9`N`F^1a9bN`F\\1_9fNbFX1]9jNdFT1\\9mNeFQ1Z9QOgFm0X9UOiFi0W9XOkFe0T9RKbFm3:o0T9PKiFo34o0R9oJPGQ4On0P9nJWGT4Il0j9SOWFk0k9UOUFi0m9VOSFi0o9VORFh0P:XOoEg0T:WOmEg0U:XOlEf0V:ZOiEe0Y:ZOhEd0Z:\\OeEc0]:\\OdE8h:GXEMU;3kD@c;>^DTOP<k0PDiN]<W1cC\\Nj<c1VCQNW=n1jBeMc=[2]BXMP>g2PBmL]>n1VA[N>YOk>[2k@WNQ`0h1T@TNl?k1Y@PNh?o1^@kMc?T2b@gM_?X2f@dMZ?[2V101N13L6I6J7I5J3L5K4L5K5K4L5lJSLoIQ4\\5gL]J^3l4\\MlJh2^4RNXKU2Q4dNbKf1<nL^MV2i1?X1W3[NRM^1Z3UNoLd1[3oMnLk1]3hMlLP2_3cMjLW2a3\\MhL\\2c3XMeLe2]:03N2M4M2M4L4RN]^OU1ha0]Nd^Oa1oa0N2UOf]OM[b03T^O\\Ooa0b0i0O1N2O1O1N2M3M3N2M2NRhk2"}, "label": "giraffe facing other animals"}]}
{"id": 343968, "image": "COCO_train2014_000000343968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe near another giraffe and facing the camera\"."}], "task": "refering", "answer_template": "The \"a giraffe near another giraffe and facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 55, 56, 57, 69, 70, 71, 72, 84, 85, 86, 87, 99, 100, 101, 114, 115, 116, 129, 130, 144, 145, 158, 159, 160, 173, 174, 175, 188, 189, 190, 203, 204, 205, 218, 219, 220, 233, 234, 235, 248, 249, 250, 264, 265, 279, 280, 294, 309, 324, 339], "bbox": [0.5737413394919169, 0.099234375, 0.8702771362586607, 0.99875], "iscrowd": 0, "area": 26107.339849999993, "rle": {"size": [640, 433], "counts": "f]k43Vc0j0VOj0[Oe0[Oe0[Oe0ZOf0[Oe0C=J6K5K6YDfJ^8_5]FkKX9[4bEQMS:g6F:E;A?A?B:E8H7I7J7H7I8H7J6I8H7I01O000O101N2M5D=D<D;H9O001O1O0O2O001O00001O000_JWJnLj5R3`J^Lg5a3cJoKe5o3dJcKe5Y4eJWKd5e4gJkJc5Q5fJ`Jd5\\5fJUJc5g5]4L4K4M4L4L4K5L4L4L3M4K5L4L4L4L3L5L4L4L4K5L3M4L4L4K5L4L4L4L400001O000001O0001O00000000010O000000XN__O1Zb0N100O2N1O2N1O2O0O2N1O2N101N1O2N1O2N10ZkR1"}, "label": "a giraffe near another giraffe and facing the camera"}]}
{"id": 343968, "image": "COCO_train2014_000000343968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe facing the camera\"."}], "task": "refering", "answer_template": "The \"the giraffe facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 55, 56, 57, 69, 70, 71, 72, 84, 85, 86, 87, 99, 100, 101, 114, 115, 116, 129, 130, 144, 145, 158, 159, 160, 173, 174, 175, 188, 189, 190, 203, 204, 205, 218, 219, 220, 233, 234, 235, 248, 249, 250, 264, 265, 279, 280, 294, 309, 324, 339], "bbox": [0.5737413394919169, 0.099234375, 0.8702771362586607, 0.99875], "iscrowd": 0, "area": 26107.339849999993, "rle": {"size": [640, 433], "counts": "f]k43Vc0j0VOj0[Oe0[Oe0[Oe0ZOf0[Oe0C=J6K5K6YDfJ^8_5]FkKX9[4bEQMS:g6F:E;A?A?B:E8H7I7J7H7I8H7J6I8H7I01O000O101N2M5D=D<D;H9O001O1O0O2O001O00001O000_JWJnLj5R3`J^Lg5a3cJoKe5o3dJcKe5Y4eJWKd5e4gJkJc5Q5fJ`Jd5\\5fJUJc5g5]4L4K4M4L4L4K5L4L4L3M4K5L4L4L4L3L5L4L4L4K5L3M4L4L4K5L4L4L4L400001O000001O0001O00000000010O000000XN__O1Zb0N100O2N1O2N1O2O0O2N1O2N101N1O2N1O2N10ZkR1"}, "label": "the giraffe facing the camera"}]}
{"id": 139173, "image": "COCO_train2014_000000139173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands holding a phone\"."}], "task": "refering", "answer_template": "The \"hands holding a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 26, 27, 36, 37, 38, 39, 40, 41, 42, 45, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 165, 166, 167, 168, 169, 170, 171, 172, 180, 182, 183, 184, 185, 186, 187, 188, 189, 190, 198, 200, 201, 202, 203, 204, 205, 206, 207, 208, 216, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261], "bbox": [0.0, 0.0872591006423983, 0.5751000000000001, 0.8445610278372592], "iscrowd": 0, "area": 66022.04515, "rle": {"size": [467, 500], "counts": "R:W2m7_4lLT30000O1000000ObJ^Km0b4SOmL^OS3c0[NnMd1S2dNeM\\1[2gNbMY1^2jN_MU1b2nN[MR1f2POWMP1i2TOSMk0n2XOoLh0Q3[OlLe0T3^OiLa0X3BeL>\\3DaL<_3H]L7d3LYL4g3NXL1h30WLOk31TLOl32SLNm33RLLo35PLKP46oKJQ47nKHT48kKHU49jKGV4:iKEX4<gKDY4=fKC[4=dKB]4?bKA^4`0aK@_4a0`K^Oa4c0^K]Oc4c0\\K]Od4d0[K[Of4f0YKZOg4g0XKYOh4h0WKWOj4i0VKWOk4e0XK[Oh4b0[K]Of4`0]K@c4=`KC`4:cKF^45fKKZ42iKNW4OlK1T4LoK4Q4IRL7n3FUL:l3BWL>i3WNRKbNX1W3f3WN\\KTNR1e3b3VNgKhMj0R4_3VNcMj1^2TNcMm1\\2SNdMm1\\2RNeMn1[2RNeMn1[2QNfMo1[2PNeMP2[2oMfMQ2Z2oMfMQ2Z2nMgMR2Y2nMgMR2Y2kMjMU2W2gMlMY2T2dMoM\\2Q2aMRN_2n1^MUNb2k1[MXNe2i1WMZNi2f1TM]Nm2b1PMaNP3_1mLdNS3\\1jLgNV3Z1gLhNY3X1eLjN[3V1dLkN\\3U1cLlN]3T1aLnN_3R1`LoN`3R1^LoNb3Q1]LPOc3P1[LROe3n0ZLSOf3m0YLTOg3m0VLUOj3k0ULVOk3j0TLWOm3HVLTLNU4l3EaLjKDc4k3ASM`0m2^OTMc0l2\\OSMf0m2XOTMi0l2UOUMl0k2ROVMo0j2oNVMS1j2kNWMV1h2iNXMY1h2fNWM\\1i2bNXM_1h2_NXMc1h2[NYMf1g2XNYMj1f2UNZMm1f2QN[MP2e2oMZMS2f2kM[MV2e2hM[MZ2d2eM]M\\2c2bM]M`2c2^M]Md2c2ZM^Mg2b2XM]Mj2b2UM_Ml2a2RM_MP3a2nL`MS3`2kL`MW3`2gL`M[3_2dLbM]3^2bLaM`3_2^LbMc3^2[LcMf3]2XLcMj3\\2ULeMl3[2RLfMo3Z2oKgMR4Y2mKgMT4Y2jKhMW4W2hKkMX4U2fKmMZ4S2dKnM]4R2aKPN_4P2_KQNb4n1^KSNb4m1\\KTNe4l1YKVNg4j1WKXNi4h1UKYNl4f1SK\\Nm4d1RK]Nn4c1PK^NQ5b1mJ`NS5`1kJbNU5]1jJeNV5[1iJfNW5Z1gJhNY5X1eJjN[5V1cJlN]5T1bJmN^5R1aJPO_5P1_JQOb5o0\\JSOd5m0ZJTOg5l0XJTOi5k0VJWOj5i0TJXOm5h0QJZOo5f0oI[OR6e0mI\\OS6c0lI^OU6b0iI@f1cLZ2m3nKAg1cL\\2l3lKBf1cL_2k3iKCg1bLb2j3fKFf1aLe2i3cKGh1`Lf2i3`KIh1_Li2h3^KIh1`Lk2g3[KLg1]LP3f3XK0e1[LT3d7iL]HX3c7fL^H[3c7bL^H_3c7^L]Hd3d7YL]Hh3d7UL]Hl3d7QL]HP4d7mK]HT4d7iK\\HY4e7dK\\H]4e7`K\\Ha4f7\\KZHe4i7VKXHk4j7QKVHQ5l7kJUHV5m7gJSHZ5b81O1N2O1O1O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1RLRI5P7KQI3P7LRI1P7OPI0Q70PIMR73oHKR75nHIT77mHGT78nHET7;lHDU7<lHAV7?kH_OV7a0kH\\OW7d0iH[OX7e0iHXOY7h0hHVOY7j0gHTO[7l0fHROi0\\MX5b3PJoNj0`MS5c3TJkNj0cMP5b3WJjNi0fMn4a3ZJhNh0hMl4a3]JfNg0jMj4a3_JeNg0kMh4a3bJcNe0mMh4a3dJaNd0oMf4a3fJbNb0nMf4a3iJbN>oMg4`3lJbN9QNi4_3nJaN6RNj4_3oJaN4RNj4_3SK_N0UNk4^3UK^NMVNl4^3VK^NJWNn4]3XK]NGWNP5^3UK`NGUNR5]3SKcNHRNS5\\3RKgNGPNU5[3PKjNHmMV5[3nJ4P5NlJ6R5g4N2M2O2N2N2N2N4L8H8H9[NXEnNP;9kE]O]:K]FKk9]OoF8P<FnZP3"}, "label": "hands holding a phone"}]}
{"id": 139173, "image": "COCO_train2014_000000139173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands holding an old cell phone\"."}], "task": "refering", "answer_template": "The \"hands holding an old cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 26, 27, 36, 37, 38, 39, 40, 41, 42, 45, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 165, 166, 167, 168, 169, 170, 171, 172, 180, 182, 183, 184, 185, 186, 187, 188, 189, 190, 198, 200, 201, 202, 203, 204, 205, 206, 207, 208, 216, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261], "bbox": [0.0, 0.0872591006423983, 0.5751000000000001, 0.8445610278372592], "iscrowd": 0, "area": 66022.04515, "rle": {"size": [467, 500], "counts": "R:W2m7_4lLT30000O1000000ObJ^Km0b4SOmL^OS3c0[NnMd1S2dNeM\\1[2gNbMY1^2jN_MU1b2nN[MR1f2POWMP1i2TOSMk0n2XOoLh0Q3[OlLe0T3^OiLa0X3BeL>\\3DaL<_3H]L7d3LYL4g3NXL1h30WLOk31TLOl32SLNm33RLLo35PLKP46oKJQ47nKHT48kKHU49jKGV4:iKEX4<gKDY4=fKC[4=dKB]4?bKA^4`0aK@_4a0`K^Oa4c0^K]Oc4c0\\K]Od4d0[K[Of4f0YKZOg4g0XKYOh4h0WKWOj4i0VKWOk4e0XK[Oh4b0[K]Of4`0]K@c4=`KC`4:cKF^45fKKZ42iKNW4OlK1T4LoK4Q4IRL7n3FUL:l3BWL>i3WNRKbNX1W3f3WN\\KTNR1e3b3VNgKhMj0R4_3VNcMj1^2TNcMm1\\2SNdMm1\\2RNeMn1[2RNeMn1[2QNfMo1[2PNeMP2[2oMfMQ2Z2oMfMQ2Z2nMgMR2Y2nMgMR2Y2kMjMU2W2gMlMY2T2dMoM\\2Q2aMRN_2n1^MUNb2k1[MXNe2i1WMZNi2f1TM]Nm2b1PMaNP3_1mLdNS3\\1jLgNV3Z1gLhNY3X1eLjN[3V1dLkN\\3U1cLlN]3T1aLnN_3R1`LoN`3R1^LoNb3Q1]LPOc3P1[LROe3n0ZLSOf3m0YLTOg3m0VLUOj3k0ULVOk3j0TLWOm3HVLTLNU4l3EaLjKDc4k3ASM`0m2^OTMc0l2\\OSMf0m2XOTMi0l2UOUMl0k2ROVMo0j2oNVMS1j2kNWMV1h2iNXMY1h2fNWM\\1i2bNXM_1h2_NXMc1h2[NYMf1g2XNYMj1f2UNZMm1f2QN[MP2e2oMZMS2f2kM[MV2e2hM[MZ2d2eM]M\\2c2bM]M`2c2^M]Md2c2ZM^Mg2b2XM]Mj2b2UM_Ml2a2RM_MP3a2nL`MS3`2kL`MW3`2gL`M[3_2dLbM]3^2bLaM`3_2^LbMc3^2[LcMf3]2XLcMj3\\2ULeMl3[2RLfMo3Z2oKgMR4Y2mKgMT4Y2jKhMW4W2hKkMX4U2fKmMZ4S2dKnM]4R2aKPN_4P2_KQNb4n1^KSNb4m1\\KTNe4l1YKVNg4j1WKXNi4h1UKYNl4f1SK\\Nm4d1RK]Nn4c1PK^NQ5b1mJ`NS5`1kJbNU5]1jJeNV5[1iJfNW5Z1gJhNY5X1eJjN[5V1cJlN]5T1bJmN^5R1aJPO_5P1_JQOb5o0\\JSOd5m0ZJTOg5l0XJTOi5k0VJWOj5i0TJXOm5h0QJZOo5f0oI[OR6e0mI\\OS6c0lI^OU6b0iI@f1cLZ2m3nKAg1cL\\2l3lKBf1cL_2k3iKCg1bLb2j3fKFf1aLe2i3cKGh1`Lf2i3`KIh1_Li2h3^KIh1`Lk2g3[KLg1]LP3f3XK0e1[LT3d7iL]HX3c7fL^H[3c7bL^H_3c7^L]Hd3d7YL]Hh3d7UL]Hl3d7QL]HP4d7mK]HT4d7iK\\HY4e7dK\\H]4e7`K\\Ha4f7\\KZHe4i7VKXHk4j7QKVHQ5l7kJUHV5m7gJSHZ5b81O1N2O1O1O1N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1RLRI5P7KQI3P7LRI1P7OPI0Q70PIMR73oHKR75nHIT77mHGT78nHET7;lHDU7<lHAV7?kH_OV7a0kH\\OW7d0iH[OX7e0iHXOY7h0hHVOY7j0gHTO[7l0fHROi0\\MX5b3PJoNj0`MS5c3TJkNj0cMP5b3WJjNi0fMn4a3ZJhNh0hMl4a3]JfNg0jMj4a3_JeNg0kMh4a3bJcNe0mMh4a3dJaNd0oMf4a3fJbNb0nMf4a3iJbN>oMg4`3lJbN9QNi4_3nJaN6RNj4_3oJaN4RNj4_3SK_N0UNk4^3UK^NMVNl4^3VK^NJWNn4]3XK]NGWNP5^3UK`NGUNR5]3SKcNHRNS5\\3RKgNGPNU5[3PKjNHmMV5[3nJ4P5NlJ6R5g4N2M2O2N2N2N2N4L8H8H9[NXEnNP;9kE]O]:K]FKk9]OoF8P<FnZP3"}, "label": "hands holding an old cell phone"}]}
{"id": 139173, "image": "COCO_train2014_000000139173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man ' s arm next to a wooden object\"."}], "task": "refering", "answer_template": "The \"man ' s arm next to a wooden object\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 67, 68, 69, 86], "bbox": [0.65016, 0.007794432548179872, 0.99708, 0.25194860813704495], "iscrowd": 0, "area": 10224.198700000003, "rle": {"size": [467, 500], "counts": "ZWd41b>1O1O2N1O1O2N1O2N1O1O2M2O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2M2O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O00O10O01O100O1O100O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O001O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O001O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O9FY>"}, "label": "man ' s arm next to a wooden object"}]}
{"id": 139173, "image": "COCO_train2014_000000139173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hairy body part next to or under a piece of wood\"."}], "task": "refering", "answer_template": "The \"a hairy body part next to or under a piece of wood\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 67, 68, 69, 86], "bbox": [0.65016, 0.007794432548179872, 0.99708, 0.25194860813704495], "iscrowd": 0, "area": 10224.198700000003, "rle": {"size": [467, 500], "counts": "ZWd41b>1O1O2N1O1O2N1O2N1O1O2M2O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1N3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2M2O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O00O10O01O100O1O100O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O001O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O001O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O9FY>"}, "label": "a hairy body part next to or under a piece of wood"}]}
{"id": 507815, "image": "COCO_train2014_000000507815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reddish - brown old looking suitcase\"."}], "task": "refering", "answer_template": "The \"a reddish - brown old looking suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267], "bbox": [0.22815625, 0.002256267409470752, 0.72984375, 0.9011142061281338], "iscrowd": 0, "area": 42561.165599999986, "rle": {"size": [359, 640], "counts": "RWc13n:6J7J500O101N100O2N100O101O00001O00001O0000001O1O2N2N3M2N1O00001O000000000000000O1O1N3M2O1000000000000000001O0000000O1000000000001O000000000000000000000O100O100O100O010O1000O1000000O100000000O1000000O1000000O1000000O1001O3M2N2N3M2O1N0000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000001O0001O0001O001O001O001O001O001O001O00001O001O001O00000000000000000000000000000001O1O2N1O1O1O2Nn0ROZ1fN[1eNZ1fN[1eNV1jN10O01O001M2L4M4O000000001O0O10001O000000000O10O100O100O2N100O2O0O101N100O2N100O2O0O100O2O0O1O101N100O101N100O1O2O0O100O2O0O100O1O101N100O100O1O2O0O100O101N100O1O100O2O0O100O101N1O100O100O2O0O100O2N2K5cN]1dN]1cN\\1cN]1dNZel1"}, "label": "a reddish - brown old looking suitcase"}]}
{"id": 507815, "image": "COCO_train2014_000000507815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan suitcase slightly shorter than the other suitcase\"."}], "task": "refering", "answer_template": "The \"a tan suitcase slightly shorter than the other suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.0, 0.11649025069637883, 0.565984375, 1.0], "iscrowd": 0, "area": 104923.40985000001, "rle": {"size": [359, 640], "counts": "V8W1b7_2aM^2ZNg1N101O001O1N101O001O1N101O001O001N2O001O001N101O1O001N101O1O001O0O2O001O1O0O2O001O001N20O010O01O010O10O010O0100O0010O010O0100O001O0O2O001N2O0O2O001N101O1N101N101O1N101N101O0O2O1O001O001O001O1O001O00000000000000000000000000000O10000000000000000000000000000000000000000000000001O000000000000001O000000000000001O000000000000001O002N3M4L3M000000001O00000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000000000000001OO10000O1000000O10000O10000O1000O10O10000O1000000O11O0000001O0000001O0000001O002N1O1O2N00000000000000001O000000000000000000000000000000001O0000000O10000000000000000000001O00000000000000000000000000001O000000000000000000001O000000000000000000001O001O1O1O1O1O1N2O1O4Lo0QOo0QOn0ROo0POP1POP1QOn0QOP1POaTQ3"}, "label": "a tan suitcase slightly shorter than the other suitcase"}]}
{"id": 507815, "image": "COCO_train2014_000000507815.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the light brown suitcase furthest to the left\"."}], "task": "refering", "answer_template": "The \"the light brown suitcase furthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.0, 0.11649025069637883, 0.565984375, 1.0], "iscrowd": 0, "area": 104923.40985000001, "rle": {"size": [359, 640], "counts": "V8W1b7_2aM^2ZNg1N101O001O1N101O001O1N101O001O001N2O001O001N101O1O001N101O1O001O0O2O001O1O0O2O001O001N20O010O01O010O10O010O0100O0010O010O0100O001O0O2O001N2O0O2O001N101O1N101N101O1N101N101O0O2O1O001O001O001O1O001O00000000000000000000000000000O10000000000000000000000000000000000000000000000001O000000000000001O000000000000001O000000000000001O002N3M4L3M000000001O00000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000000000000001OO10000O1000000O10000O10000O1000O10O10000O1000000O11O0000001O0000001O0000001O002N1O1O2N00000000000000001O000000000000000000000000000000001O0000000O10000000000000000000001O00000000000000000000000000001O000000000000000000001O000000000000000000001O001O1O1O1O1O1N2O1O4Lo0QOo0QOn0ROo0POP1POP1QOn0QOP1POaTQ3"}, "label": "the light brown suitcase furthest to the left"}]}
{"id": 49073, "image": "COCO_train2014_000000049073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big giraffe\"."}], "task": "refering", "answer_template": "The \"the big giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 39, 40, 41, 42, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 150, 151, 152, 153, 155, 156, 157, 158, 159, 160, 161, 165, 169, 171, 172, 173, 174, 175, 180, 184, 185, 188, 189, 190, 199, 200, 201, 204, 214, 215, 216, 230, 231, 245, 246, 260, 261, 276], "bbox": [0.0, 0.0959375, 0.8925526932084309, 0.7948124999999999], "iscrowd": 0, "area": 84483.06095000003, "rle": {"size": [640, 427], "counts": "l8`2o<_42N1O2N1O2N2N101N2N1O2N1O2N2N1O2O0O2N2N1O2N2N1O2N1O2O1N1O2N2N1O2N1O2N2O0O2N2N1011N2O1N2O2M2O1N2O1N2O1N3N1N2O1N2N2O2M10O001O0010O01O001O010O00001O010O001O001O010O001O00010O001O0010O01O001O0010O01O000010O01O001O010O001O001O01O100O100O100O100O10000O1jD\\JP8d5lGcJQ8]5kGjJQ8W5kGPKR8P5jGWKS8i4hG`KS8a4iGfKT8Z4hGmKU8S4gGTLU8m3fG[LW8e3iG^LT8b3nG_Ln7b3TH_Li7a3YHbLb7^3`HdL[7]3fHgLU7Y3mHiLn6Y3SIjLh6W3XIlLd6U3\\IoL^6S3bIPMZ6R3eIRMV6o2jITMQ6n2oIVMl5l2TJVMh5k2XJYMb5i2^JZM^5h2aJ\\MZ5e2fJ^MU5d2kJ`MP5b2oJ_Mo4b2RK]Ml4e2TK[Mk4g2TKYMk4h2UKXMi4j2WKVMh4l2WKUMg4l2YKTMe4n2[KRMd4o2]KPMb4Q3^KoL`4S3`KmL_4T3aKlL^4U3bKjL]4W3dKhL\\4Y3dKgL[4Z3fKdLY4]3hKbLX4_3hKaLW4`3jK^LU4c3nKZLR4g3PLWLn3k3TLRLl3n3VLQLi3P4YLmKf3U4\\LhKd3X4_LfK`3[4aLcK`3]4bL`K^3`4hL[KW3f4oLSKQ3n4UMkJk2U5\\MeJc2\\5cM]J]2d5hMVJX2j5oMPJP2Q6WNgIi1Z6^N_Ia1a6hNVIX1k6oNmHQ1T7WOdHh0]7_O[Ha0e7GSH9n7l51O00100O001O1O001O1O001O1O001O1O001O1O001O100O00J6J6K5J6K5J6J6K5J6K5J6J6K5J6XOh0SOn0M2OiASL]<l3aCeLQ<Y3nCWMe;h2YD]Me;a2[DbMd;]2ZDgMe;X2YDlMf;S2XDPNh;o1VDUNi;j1UDZN9PNP:e3eE_N7SNP:^3fEbN7WNo9V3hEgN6YNn9o2kEkN3]Nn9g2mEPO1`Nn9`2QFPONgNm9X2UFSOJiNP:S2WFUOFkNS:n1WFYOBlNW:j1XFZO^OoNZ:g1WF\\O[OQO]:b1YFa1g9^NYFd1f9[N[Fe1e9[NZFg1e9XN\\Fh1d9WN\\Fj1d9UN]Fl1b9SN^Fn1b9RN^Fo1a9PN_FQ2a9nM_FT2`9kMaFU2_9kM`FW2_9hMbFX2^9gMbFZ2^9eMcF\\2]9bMcF_2]9aMcF`2\\9_MdF`2^9_McF_2_9`MaF`2`9`M`F^2b9aM^F^2d9aM]F]2e9bMZF^2h9aMWF`2j9`MTF`2n9_MRF`2P:_MoEa2S:^MlEb2V:^MiEb2X:]MgEc2[:\\MdEd2^:[MaEd2b:\\M]Ec2f:[MYEe2i:ZMVEe2m:ZMSEe2o:ZMPEf2R;ZMlDe2W;ZMiDe2Y;ZMfDf2\\;YMcDf2`;ZM^Df2d;YM\\Df2f;YMYDf2j;YMUDg2m;XMSDg2o;YMoCf2T<YMkCg2W<XMiCg2Y<XMfCg2]<YMaCg2a<XM^Ch2e<VM[Ch2h<WMWCi2k<T22M3N2N2M3N2N2N2O1N2O1N2O1N2N2O1N3N1N2O1N2O1N2O1O1SOjAoKW>o3kAPLX>k3jAULX>h3jAWLY>d3jA[LX>b3jA]LY>^3jAaLX>\\3jAcLY>X3jAgLX>V3iAjLZ>Q3iAnLY>o2iAPMZ>l2hASMZ>i2iAVMZ>f2hAYMZ>c2hA]MZ>`2hA^M\\>]2gAbM[>[2gAdM\\>W2gAhM[>U2gAjM\\>Q2gAnM[>o1fAQN]>k1eATN]>h1fAWN]>e1eAZN]>b1fA]N]>_1eA`N]>\\1fAcN]>Y1dAgN^>U1Z2L3L5L3L5L3M4K4M3LUdl0"}, "label": "the big giraffe"}]}
{"id": 49073, "image": "COCO_train2014_000000049073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big giraffe taking small giraffe nearly\"."}], "task": "refering", "answer_template": "The \"a big giraffe taking small giraffe nearly\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 38, 39, 40, 41, 42, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 150, 151, 152, 153, 155, 156, 157, 158, 159, 160, 161, 165, 169, 171, 172, 173, 174, 175, 180, 184, 185, 188, 189, 190, 199, 200, 201, 204, 214, 215, 216, 230, 231, 245, 246, 260, 261, 276], "bbox": [0.0, 0.0959375, 0.8925526932084309, 0.7948124999999999], "iscrowd": 0, "area": 84483.06095000003, "rle": {"size": [640, 427], "counts": "l8`2o<_42N1O2N1O2N2N101N2N1O2N1O2N2N1O2O0O2N2N1O2N2N1O2N1O2O1N1O2N2N1O2N1O2N2O0O2N2N1011N2O1N2O2M2O1N2O1N2O1N3N1N2O1N2N2O2M10O001O0010O01O001O010O00001O010O001O001O010O001O00010O001O0010O01O001O0010O01O000010O01O001O010O001O001O01O100O100O100O100O10000O1jD\\JP8d5lGcJQ8]5kGjJQ8W5kGPKR8P5jGWKS8i4hG`KS8a4iGfKT8Z4hGmKU8S4gGTLU8m3fG[LW8e3iG^LT8b3nG_Ln7b3TH_Li7a3YHbLb7^3`HdL[7]3fHgLU7Y3mHiLn6Y3SIjLh6W3XIlLd6U3\\IoL^6S3bIPMZ6R3eIRMV6o2jITMQ6n2oIVMl5l2TJVMh5k2XJYMb5i2^JZM^5h2aJ\\MZ5e2fJ^MU5d2kJ`MP5b2oJ_Mo4b2RK]Ml4e2TK[Mk4g2TKYMk4h2UKXMi4j2WKVMh4l2WKUMg4l2YKTMe4n2[KRMd4o2]KPMb4Q3^KoL`4S3`KmL_4T3aKlL^4U3bKjL]4W3dKhL\\4Y3dKgL[4Z3fKdLY4]3hKbLX4_3hKaLW4`3jK^LU4c3nKZLR4g3PLWLn3k3TLRLl3n3VLQLi3P4YLmKf3U4\\LhKd3X4_LfK`3[4aLcK`3]4bL`K^3`4hL[KW3f4oLSKQ3n4UMkJk2U5\\MeJc2\\5cM]J]2d5hMVJX2j5oMPJP2Q6WNgIi1Z6^N_Ia1a6hNVIX1k6oNmHQ1T7WOdHh0]7_O[Ha0e7GSH9n7l51O00100O001O1O001O1O001O1O001O1O001O1O001O100O00J6J6K5J6K5J6J6K5J6K5J6J6K5J6XOh0SOn0M2OiASL]<l3aCeLQ<Y3nCWMe;h2YD]Me;a2[DbMd;]2ZDgMe;X2YDlMf;S2XDPNh;o1VDUNi;j1UDZN9PNP:e3eE_N7SNP:^3fEbN7WNo9V3hEgN6YNn9o2kEkN3]Nn9g2mEPO1`Nn9`2QFPONgNm9X2UFSOJiNP:S2WFUOFkNS:n1WFYOBlNW:j1XFZO^OoNZ:g1WF\\O[OQO]:b1YFa1g9^NYFd1f9[N[Fe1e9[NZFg1e9XN\\Fh1d9WN\\Fj1d9UN]Fl1b9SN^Fn1b9RN^Fo1a9PN_FQ2a9nM_FT2`9kMaFU2_9kM`FW2_9hMbFX2^9gMbFZ2^9eMcF\\2]9bMcF_2]9aMcF`2\\9_MdF`2^9_McF_2_9`MaF`2`9`M`F^2b9aM^F^2d9aM]F]2e9bMZF^2h9aMWF`2j9`MTF`2n9_MRF`2P:_MoEa2S:^MlEb2V:^MiEb2X:]MgEc2[:\\MdEd2^:[MaEd2b:\\M]Ec2f:[MYEe2i:ZMVEe2m:ZMSEe2o:ZMPEf2R;ZMlDe2W;ZMiDe2Y;ZMfDf2\\;YMcDf2`;ZM^Df2d;YM\\Df2f;YMYDf2j;YMUDg2m;XMSDg2o;YMoCf2T<YMkCg2W<XMiCg2Y<XMfCg2]<YMaCg2a<XM^Ch2e<VM[Ch2h<WMWCi2k<T22M3N2N2M3N2N2N2O1N2O1N2O1N2N2O1N3N1N2O1N2O1N2O1O1SOjAoKW>o3kAPLX>k3jAULX>h3jAWLY>d3jA[LX>b3jA]LY>^3jAaLX>\\3jAcLY>X3jAgLX>V3iAjLZ>Q3iAnLY>o2iAPMZ>l2hASMZ>i2iAVMZ>f2hAYMZ>c2hA]MZ>`2hA^M\\>]2gAbM[>[2gAdM\\>W2gAhM[>U2gAjM\\>Q2gAnM[>o1fAQN]>k1eATN]>h1fAWN]>e1eAZN]>b1fA]N]>_1eA`N]>\\1fAcN]>Y1dAgN^>U1Z2L3L5L3L5L3M4K4M3LUdl0"}, "label": "a big giraffe taking small giraffe nearly"}]}
{"id": 49073, "image": "COCO_train2014_000000049073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby giraffe and the head of its mama\"."}], "task": "refering", "answer_template": "The \"baby giraffe and the head of its mama\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 312, 313, 314, 322, 323, 327, 328, 329, 337, 338], "bbox": [0.13473067915690867, 0.413109375, 0.9531850117096019, 0.9838906250000001], "iscrowd": 0, "area": 39246.54945000001, "rle": {"size": [640, 427], "counts": "QbT14kc04L5K4L2O0O1O1O1O1O001O1O1O1O1O001N101O001O001O1O001O0O2O001O001O001O001O001O0O2O1O1O1O1N2O001O1N2O1O1O0011O000000000O10000000000O2O000001O1O2N1O1O1O2N1O1O2O0O1O1O2N1O1O1O2N11O01O0000001O00001O0000010O0000001O00001O000010O0001O00001O00001O00001O000O101O00001O00001N10001O00001O000O2O0U@UM`>k2]AXMc>i2YA[Mf>e2WA^Mi>b2TAaMl>_2QAeMm>]2o@fMQ?Z2l@iMT?W2i@mMV?T2f@oM^4UOn5l2aESNX4VOZ6g2[EVNS4YOd6b2UEXNo3[Oo6i3iH\\LX7e3_HaLd7_3QHiLR8X3cGPM_8P3eFiM^9X2VFPNm9P2hEYNZ:g1\\E`Ng:a1mDhNT;Y1bDnNa;f43N2M3M3O11O001O1N101O1O001O1O1O001N2OO1O1O1O1O1O1O1L4K4L5L4K5K5K5K5K5K5K4M4K5L4N2M2N3N1N3M3M2O2M2N3N2M2N3N1N3M2N3M3I6J7I6I8N20O10O010000O010O100O01000O10O010000O010O100O01000O10O010000O010O1000O0100O10O010000O01000O100000000O10000000000O10000b@dLW>\\3gAeLZ>[3eAeL[>\\3cAeL^>[3`AfLa>Z3^AfLc>Z3[AgLf>Y3XAhLh>X41O1N2O1O1O001O1O1O1O001O1oMl@TOU?g0PAXOQ?b0UA]Ok>?ZA@g>:_AEb>5dAJ]>1hANY>LmA3T>HQB7o=DWB:k=AZB>g=\\O_Bc0b=XOcBg0^=SOhBl0X=POmBo0T=kNRCT1o<gNVCX1j?O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O>B`m;"}, "label": "baby giraffe and the head of its mama"}]}
{"id": 49073, "image": "COCO_train2014_000000049073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe\"."}], "task": "refering", "answer_template": "The \"a baby giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 312, 313, 314, 322, 323, 327, 328, 329, 337, 338], "bbox": [0.13473067915690867, 0.413109375, 0.9531850117096019, 0.9838906250000001], "iscrowd": 0, "area": 39246.54945000001, "rle": {"size": [640, 427], "counts": "QbT14kc04L5K4L2O0O1O1O1O1O001O1O1O1O1O001N101O001O001O1O001O0O2O001O001O001O001O001O0O2O1O1O1O1N2O001O1N2O1O1O0011O000000000O10000000000O2O000001O1O2N1O1O1O2N1O1O2O0O1O1O2N1O1O1O2N11O01O0000001O00001O0000010O0000001O00001O000010O0001O00001O00001O00001O000O101O00001O00001N10001O00001O000O2O0U@UM`>k2]AXMc>i2YA[Mf>e2WA^Mi>b2TAaMl>_2QAeMm>]2o@fMQ?Z2l@iMT?W2i@mMV?T2f@oM^4UOn5l2aESNX4VOZ6g2[EVNS4YOd6b2UEXNo3[Oo6i3iH\\LX7e3_HaLd7_3QHiLR8X3cGPM_8P3eFiM^9X2VFPNm9P2hEYNZ:g1\\E`Ng:a1mDhNT;Y1bDnNa;f43N2M3M3O11O001O1N101O1O001O1O1O001N2OO1O1O1O1O1O1O1L4K4L5L4K5K5K5K5K5K5K4M4K5L4N2M2N3N1N3M3M2O2M2N3N2M2N3N1N3M2N3M3I6J7I6I8N20O10O010000O010O100O01000O10O010000O010O100O01000O10O010000O010O1000O0100O10O010000O01000O100000000O10000000000O10000b@dLW>\\3gAeLZ>[3eAeL[>\\3cAeL^>[3`AfLa>Z3^AfLc>Z3[AgLf>Y3XAhLh>X41O1N2O1O1O001O1O1O1O001O1oMl@TOU?g0PAXOQ?b0UA]Ok>?ZA@g>:_AEb>5dAJ]>1hANY>LmA3T>HQB7o=DWB:k=AZB>g=\\O_Bc0b=XOcBg0^=SOhBl0X=POmBo0T=kNRCT1o<gNVCX1j?O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O>B`m;"}, "label": "a baby giraffe"}]}
{"id": 65457, "image": "COCO_train2014_000000065457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink the cat is in\"."}], "task": "refering", "answer_template": "The \"the sink the cat is in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 74, 75, 82, 84, 92, 110, 139, 140, 157, 158, 169, 174, 175, 187, 188, 189, 190], "bbox": [0.12057999999999999, 0.16405099150141642, 0.79008, 0.813484419263456], "iscrowd": 0, "area": 10229.3763, "rle": {"size": [353, 500], "counts": "kjd02i::G9G9F:D<@`0B>K0O100000O001O1O1O1O0H9C=F:N2N1O2N2N2N2H8H9G800100O1O100O1O100O1O100O001O100O1O100O1O100O1O100O1O10O10000000000000000000000000010O00000001O01O000001O0001O0001O0001O0001O0000010O0000000010O0O1O010O010000O100O100O10O0100O100O1O1O1O10O1000000000000000001N100000000000000001^KQOPOo0P1ROoNn0Q1SOnNm0S1SOlNm0T1TOkNl0U1UOjNk0V1UOjNk0W1TOiNl0W1UOhNk0X1UOhNk0X1UOhNk0X1UOhNk0Y1UOgNj0Y1VOgNj0Y1VOgNj0Y1WOfNi0[1VOeNj0[1VOeNj0[1WOdNj0[1VOeNj0\\1UOdNk0\\1UOdNk0\\1VOcNj0]1VOcNj0]1VOcNj0^1VOaNj0_1VOaNj0_1VOaNj0_1VOaNj0`1VO`Ni0`1WO`Ni0`1WO`Ni0`1XO_Nh0a1XO_Nh0b1WO_Nh0a1XO_Nh0a1YO^Ng0b1YO^Ng0c1XO^Nh0a1YO^Ng0b1YO^Ng0b1YO^Ng0b1YO_Nf0b1ZO]Nf0c1ZO]Nf0c1ZO]Nf0c1[O\\Ne0e1ZO\\Ne0d1[O\\Ne0d1[O\\Ne0d1[O\\Ne0d1[O]Nd0d1[O\\Ne0d1[O\\Ne0d1[O\\Ne0d1[O\\Na0h1_OYN;l1ETN5R2KnMOX21hMJ]26W400000001O000000000000000001O0cJJO6[O`0f0_OZOa0f0_OZOb0e0^OZOc0f0]OZOc0f0]OZOc0f0]OZOd0e0\\O[Od0e0\\O[Od0e0\\O[Od0e0\\O[Oe0d0[O\\Oe0d0[O\\Oe0d0[O\\Oe0e0ZO[Og0d0YO\\Og0d0YO\\Og0d0YO\\Oh0c0XO]Oh0c0XO]Oh0c0XO^Og0b0YO^Og0b0YO^Og0b0YO^Og0b0YO^Of0d0YO\\Og0e0XO[Oh0f0WOZOi0g0VOYOi0`5000000O1000000O1000000O1000000O100000000O10000000000001O0000001O00001O0000001O0000001O000\\JXOo0h0QOXOo0h0POYOP1h0nNYOS1f0lN[OT1e0kN]OT1b0mN^OS1b0lN_OT1`0lNAT1?kNBU1>jNDU1;kNFW18gNJ\\12cN0e1FZN=S62N2M2O0O2O1O0O2O1N2O001N2N1N3M3M3M21N4L4L5L3N3L7J6J7H7J_oS1"}, "label": "the sink the cat is in"}]}
{"id": 65457, "image": "COCO_train2014_000000065457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sink under cat\"."}], "task": "refering", "answer_template": "The \"sink under cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 74, 75, 82, 84, 92, 110, 139, 140, 157, 158, 169, 174, 175, 187, 188, 189, 190], "bbox": [0.12057999999999999, 0.16405099150141642, 0.79008, 0.813484419263456], "iscrowd": 0, "area": 10229.3763, "rle": {"size": [353, 500], "counts": "kjd02i::G9G9F:D<@`0B>K0O100000O001O1O1O1O0H9C=F:N2N1O2N2N2N2H8H9G800100O1O100O1O100O1O100O001O100O1O100O1O100O1O100O1O10O10000000000000000000000000010O00000001O01O000001O0001O0001O0001O0001O0000010O0000000010O0O1O010O010000O100O100O10O0100O100O1O1O1O10O1000000000000000001N100000000000000001^KQOPOo0P1ROoNn0Q1SOnNm0S1SOlNm0T1TOkNl0U1UOjNk0V1UOjNk0W1TOiNl0W1UOhNk0X1UOhNk0X1UOhNk0X1UOhNk0Y1UOgNj0Y1VOgNj0Y1VOgNj0Y1WOfNi0[1VOeNj0[1VOeNj0[1WOdNj0[1VOeNj0\\1UOdNk0\\1UOdNk0\\1VOcNj0]1VOcNj0]1VOcNj0^1VOaNj0_1VOaNj0_1VOaNj0_1VOaNj0`1VO`Ni0`1WO`Ni0`1WO`Ni0`1XO_Nh0a1XO_Nh0b1WO_Nh0a1XO_Nh0a1YO^Ng0b1YO^Ng0c1XO^Nh0a1YO^Ng0b1YO^Ng0b1YO^Ng0b1YO_Nf0b1ZO]Nf0c1ZO]Nf0c1ZO]Nf0c1[O\\Ne0e1ZO\\Ne0d1[O\\Ne0d1[O\\Ne0d1[O\\Ne0d1[O]Nd0d1[O\\Ne0d1[O\\Ne0d1[O\\Ne0d1[O\\Na0h1_OYN;l1ETN5R2KnMOX21hMJ]26W400000001O000000000000000001O0cJJO6[O`0f0_OZOa0f0_OZOb0e0^OZOc0f0]OZOc0f0]OZOc0f0]OZOd0e0\\O[Od0e0\\O[Od0e0\\O[Od0e0\\O[Oe0d0[O\\Oe0d0[O\\Oe0d0[O\\Oe0e0ZO[Og0d0YO\\Og0d0YO\\Og0d0YO\\Oh0c0XO]Oh0c0XO]Oh0c0XO^Og0b0YO^Og0b0YO^Og0b0YO^Og0b0YO^Of0d0YO\\Og0e0XO[Oh0f0WOZOi0g0VOYOi0`5000000O1000000O1000000O1000000O100000000O10000000000001O0000001O00001O0000001O0000001O000\\JXOo0h0QOXOo0h0POYOP1h0nNYOS1f0lN[OT1e0kN]OT1b0mN^OS1b0lN_OT1`0lNAT1?kNBU1>jNDU1;kNFW18gNJ\\12cN0e1FZN=S62N2M2O0O2O1O0O2O1N2O001N2N1N3M3M3M21N4L4L5L3N3L7J6J7H7J_oS1"}, "label": "sink under cat"}]}
{"id": 65457, "image": "COCO_train2014_000000065457.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beige counter top in the bathroom\"."}], "task": "refering", "answer_template": "The \"a beige counter top in the bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 20, 22, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 72, 73, 74, 86, 90, 91, 92, 103, 104, 105, 108, 109, 110, 122, 123, 126, 127, 128, 140, 141, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 164, 165, 166, 167, 175, 176, 177, 178, 179, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.0, 0.0, 1.0, 0.9870254957507083], "iscrowd": 0, "area": 55066.6333, "rle": {"size": [353, 500], "counts": "]5Q1P8Q2[Od001N2N2N2O0O2N2O1N2N1O2O1N2N2O1N1O2N2O1N2N101N2N2O1N1O2N2O1N2N2O0O2N2N2O1N1O2O1N2N2N101N2N2O1N2N1O2O1N2N2O0O2N2O1N2N1O2O1N2N2O1N1O21O1O1O2gMjKaMX4W2PLhMR4o1VLPNk3h1]LWNe3_1dL`N]3W1lLgNV3P1TMnNm2i0\\MVOc2b0fM^OY2;oMEP24WNMi1K_N4a1DhN<W1]OPOd0o0UOYOk0f0PO@P1`0oN@Q1?POBP1=PODP1;POEQ1:oNGQ17POJo06QOKo04QOLP13PONP11PO0P1NQO2P1MPO4o0LQO5o0JQO6P1HQO9o0FQO;o0DQO=n0CRO=o0BQO?o0_OROb0n0]OROc0o0\\OQOe0n0ZOSOg0m0XOSOh0n0VOSOk0m0TOSOm0m0ROSOn0n0POSOQ1l0oNTOR1l0lNUOU1k0jNUOV1l0iNTOX1l0fNUO[1j0eNVO[1k0cNVO^1j0aNVO`1j0_NVOa1k0]NVOd1i0\\NWOe1i0eL[NLm0`3h0cL_NIi0d3i0bLcNEd0j3i0`LgNAa0P4h0^LlN]O<U4h0^LPOYO8Z4h0[LVOWO2^4h0ZLWOWO1_4i0YLUOZO1^4j0WLSO]O3]4i0VLRO_O5\\4i0SLQOD5Y4k0RLoNF6Y4k0PLmNI8X4k0nKkNM9U4l0nKiNO;T4l0lKhN1<T4l0iKgN6<Q4n0hKeN8=Q4n0fKeN9>Q4m0eKeN;=P4n0eKeN;=Q4n0bKeN?<P4o0`KeN`0<P4P1_KdNa0<P4Q1^KcNc0;P4Q1]KdNc0;P4R1\\KbNe0<o3S1ZKbNh0:n3U1YKaNi0:o3U1WKaNj0:o3U1WKaNj0:o3V1WK^Nk0=m3V1WK]Nl0=n3V1VK\\Nk0?o3V1UK[Nl0>P4X1TKYNl0?P4X1UKWNl0a0P4X1SKWNm0a0P4Y1SKUNm0b0P4Z1RKTNn0b0P4[1RKRNm0d0Q4Z1RKQNn0e0Q4Z1QKPNn0f0Q4[1PKoMo0f0Q4\\1PKmMo0g0Q4]1oJlMP1g0R4\\1oJkMo0j0R4\\1nJjMP1i0S4^1mJhMP1j0S4_1mJfMP1k0T4^1lJmMj0e0Z4^1mJZN;8h4^1mJgNNKU5^1nJSO_O@d5\\1mJ4S5LnJ3R5MnJ3R5MoJ2Q5NoJ2R5MoJ2Q5NoJ1R5OoJ0Q50oJ0R5OoJ0Q50PKOP51PKOP51PKOQ50oJ0Q50PKOP51PKOP51PKOQ50oJ0Q50oJ0Q50oJOR51nJOS50mJ0S50nJOR51nJOR51nJOR51nJOS50mJ0S50nJOR51oJNQ52PKMQ52PKMP53QKLo44RKKn45SKJn45TKIl47UKHk48VKGj49WKFj49WKFi4:XKEh4;YKDg4<ZKCf4=[KBf4=[KBe4>\\KAd4?]K@c4`0^K_Oc4`0^K_Ob4a0_K^Oa4b0`K]O`4c0aK\\O`4c0`K]O`4c0`K]O`4c0_K^Oa4b0_K^Oa4b0^K_Ob4a0^K_Ob4a0]K@d4?[KBe4>[KBe4>ZKCf4=ZKCf4=YKDg4<YKDg4<XKEh4;WKFi4:WKFi4:VKGj49VKGj49UKHk48UKHk48TKIl47SKJn45RKKn45QKfNmN`0R6j0QKfNPO=o5m0PKfNTO;l5o0oJfNYO8h5R1oJeN\\O5g5V1lJeN\\6[1dIdN]6\\1cIcN^6]1eI_N\\6a1gI[NZ6e1hIXNY6h1jITNW6l1_101O00000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000001O0000000000001O000000001O00000000001O00000000001O00000000001O000000O1O1O1O1O1O1O1N2O1O1G9E;N200O100O100O100O1O100O100O1cLUMdNk2[1WMcNj2\\1WMcNj2\\1XMcNh2\\1ZMbNg2]1[MbNe2]1]MaNd2^1^M`Nc2_1^M_Nd2`1^M]Nd2b1^M\\Nc2c1_MZNc2e1_MXNc2g1_MVNc2j1]MUNd2k1]MRNe2n1\\MoMf2Q2[MmMf2S2[MjMg2V2YM]LWO?b3T3XMZL\\O=]3Y3XMXL@9Z3_3WMULD7W3d3VMSLH4S3i3VMQLL0P3o3TMoK1Mm2T4SMmK5Ji2Y4WMfKU4Z4Y10000000000O100O100O13M3M3M3M3M3K4M2N2M3N2M3N1O2M3aKdJV3^5fLfJW3]5eLgJY3[5cLhJ\\3Z5`LjJ]3X5`LlJ^3V5^LnJ_3U5]LnJb3T5ZLPKd3X6M3N2M2O2N2M3N2M2O2N2M3N1N2O1O1N2O1N2O1O1N2O1N2O1O1N3N1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O2M2O1O1N2O1N2O1O1N2O1N2O1O1NK6I62O00jI"}, "label": "a beige counter top in the bathroom"}]}
{"id": 204728, "image": "COCO_train2014_000000204728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"its a white coior flight\"."}], "task": "refering", "answer_template": "The \"its a white coior flight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 219, 240, 242, 243, 244, 245, 246, 247, 253, 254, 255, 256, 257, 258, 259, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 334, 335, 336, 337, 338, 339, 340, 341, 351, 352, 357, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 405, 406, 407, 408, 409], "bbox": [0.008125, 0.35210110584518167, 0.8425156250000001, 0.7729857819905214], "iscrowd": 0, "area": 43163.77350000001, "rle": {"size": [633, 640], "counts": "e\\36cc000001O000000000000010O000000000000001O01O000000000001O000001O00000001O000000000001O01O0000000000000010O00000000000001O0001O0000000001O00000001O000001O000000000001O01O0000000000000010O00000000000001O0001O00000001O00000000010O0000000000001O0001O00000001O0000000001O01O0000000000001O01O0000000001O00000001O0001O00000000000010O00000000U^O4i?LT@:i?GT@=j?CT@`0k?@S@c0l?]OS@f0l?YOS@i0l?WOR@l0m?TOR@n0m?SOQ@o0n?QOQ@P1o?POP@R1o?nNP@S1P`0mNo_OU1P`0kNP@V1o?jNP@W1Q`0iNm_OY1R`0gNm_O[1R`0eNn_O[1R`0eNm_O\\1S`0dNl_O]1T`0cNl_O^1S`0bNl_O_1T`0bNk_O^1U`0bNk_O^1V`0aNi_O`1W`0`Ni_Oa1V`0_Nj_Oa1V`0_Nj_Oa1V`0_Ni_Ob1W`0^Ni_Oa1Y`0_Nf_Oa1Z`0_Ne_Ob1[`0^Ne_Ob1[`0^Nd_Oc1\\`0]Nc_Od1]`0\\Na_Of1_`0ZN__Oh1b`0XN[_Oi1f`0WNX_Ok1h`0UNW_Ol1i`0TNU_Om1l`0SNR_Oo1n`0>000O2N100O1O100O1YNl^Oh0Va0TOn^Oj0Sa0TOo^Ol0Qa0QOS_Ol0o`0QOU_Om0l`0ROV_Ol0k`0TOV_Oj0k`0VOV_Oh0l`0WOU_Of0m`0ZOS_Od0o`0]OR_O?Pa0AQ_O<Qa0DQ_O6Sa0Jo^O0Va0OQ_OETa0<T1000000010O0000000010O00000001O0V]O@\\b0?d]OA\\b0`0b]OB]b0>b]OD]b0=a]OE^b0;a]OF_b0:`]OFbb0:\\]OGdb09[]OHeb08Z]OIfb08X]OHib08W]OHib0d01N10cNTOm_Ol0S`0TOm_Ol0S`0UOl_Ok0T`0UOm_Oj0S`0VOm_Oj0T`0UOl_Ok0T`0VOk_Oj0U`0VOl_Oi0T`0WOl_Oi0T`0WOl_Oi0T`0WOl_Oi0U`0WOj_Oh0W`0XOj_Og0V`0YOj_Og0V`0YOj_Og0V`0YOj_Og0V`0YOj_Og0V`0ZOj_Oe0W`0ZOi_Of0W`0ZOi_Of0W`0ZOi_Of0W`0ZOi_Of0W`0ZOj_Of0U`0[Oj_Oe0V`0[Oj_Oe0W`0ZOi_Of0W`0ZOi_Of0W`0ZOi_Of0W`0[Oh_Oe0X`0[Oi_Oe0V`0[Oj_Oe0V`0[Oj_Oe0W`0ZOi_Of0W`0ZOi_Of0W`0[Oh_Oe0X`0[Oh_Oe0X`0[Oi_Oe0V`0[OU_ONMg0n`0[OT_OONf0o`0ZOS_O0Ng0n`0ZOS_OOOg0n`0ZOS_OOOh0m`0YOT_OOOh0m`0YOT_OOOh0m`0YOT_OOOi0l`0XOU_OO0h0l`0YOS_OO1h0l`0YOS_OO1i0k`0XOT_OO1i0k`0XOT_OO1j0j`0WOV_OOOj0k`0WOV_OOOj0k`0WOV_OOOk0k`0VOU_O0Oj0l`0VOU_O00j0j`0VOV_O1Nj0l`0UOW_O0Mk0l`0UOW_O1Lk0l`0TOY_O0Jm0m`0SOY_O1_OW1Wa0iNV_Od1i`0\\NP_On1m`0RNQ_OS2m`0lMP_OZ2m`0fMQ_Oa2j`0_MS_Oh2i`08L2N2N2N2M3M3M3M3N1N3M3N1N3M2O1O2M2O1fNZ1nNR1O1O1g0YOk0VOa0^OO1O1O1O1N2O1O2N1O1O1O1O1N2O1O1O1N101O001N2O001N101O1O0O2O001N2O001N101O001N101N1O2O0O2O0O2O00001O001O001O001O010CoIPCQ6Q=<O02N8IO0001O0001O000001O0001O000000010O0000000010O00000000010O00001O01O01O001O00001O001O00001O0O2O00001O001O001O001N101O1O001O001O1N101N101O0O2O1N101N2O1N2O2M2O1N3N1N2O2M2O1O2M2O2M2O1N3N1N2N3L3N2N3M2N2N3M2M3N3M2M4L3M3M4L3M3M4L3M6J9G9G9D<_Oa0_OiQn1"}, "label": "its a white coior flight"}]}
{"id": 204728, "image": "COCO_train2014_000000204728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"airplane parked at an airport terminal gate\"."}], "task": "refering", "answer_template": "The \"airplane parked at an airport terminal gate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 219, 240, 242, 243, 244, 245, 246, 247, 253, 254, 255, 256, 257, 258, 259, 264, 265, 266, 267, 268, 269, 270, 271, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 334, 335, 336, 337, 338, 339, 340, 341, 351, 352, 357, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 405, 406, 407, 408, 409], "bbox": [0.008125, 0.35210110584518167, 0.8425156250000001, 0.7729857819905214], "iscrowd": 0, "area": 43163.77350000001, "rle": {"size": [633, 640], "counts": "e\\36cc000001O000000000000010O000000000000001O01O000000000001O000001O00000001O000000000001O01O0000000000000010O00000000000001O0001O0000000001O00000001O000001O000000000001O01O0000000000000010O00000000000001O0001O00000001O00000000010O0000000000001O0001O00000001O0000000001O01O0000000000001O01O0000000001O00000001O0001O00000000000010O00000000U^O4i?LT@:i?GT@=j?CT@`0k?@S@c0l?]OS@f0l?YOS@i0l?WOR@l0m?TOR@n0m?SOQ@o0n?QOQ@P1o?POP@R1o?nNP@S1P`0mNo_OU1P`0kNP@V1o?jNP@W1Q`0iNm_OY1R`0gNm_O[1R`0eNn_O[1R`0eNm_O\\1S`0dNl_O]1T`0cNl_O^1S`0bNl_O_1T`0bNk_O^1U`0bNk_O^1V`0aNi_O`1W`0`Ni_Oa1V`0_Nj_Oa1V`0_Nj_Oa1V`0_Ni_Ob1W`0^Ni_Oa1Y`0_Nf_Oa1Z`0_Ne_Ob1[`0^Ne_Ob1[`0^Nd_Oc1\\`0]Nc_Od1]`0\\Na_Of1_`0ZN__Oh1b`0XN[_Oi1f`0WNX_Ok1h`0UNW_Ol1i`0TNU_Om1l`0SNR_Oo1n`0>000O2N100O1O100O1YNl^Oh0Va0TOn^Oj0Sa0TOo^Ol0Qa0QOS_Ol0o`0QOU_Om0l`0ROV_Ol0k`0TOV_Oj0k`0VOV_Oh0l`0WOU_Of0m`0ZOS_Od0o`0]OR_O?Pa0AQ_O<Qa0DQ_O6Sa0Jo^O0Va0OQ_OETa0<T1000000010O0000000010O00000001O0V]O@\\b0?d]OA\\b0`0b]OB]b0>b]OD]b0=a]OE^b0;a]OF_b0:`]OFbb0:\\]OGdb09[]OHeb08Z]OIfb08X]OHib08W]OHib0d01N10cNTOm_Ol0S`0TOm_Ol0S`0UOl_Ok0T`0UOm_Oj0S`0VOm_Oj0T`0UOl_Ok0T`0VOk_Oj0U`0VOl_Oi0T`0WOl_Oi0T`0WOl_Oi0T`0WOl_Oi0U`0WOj_Oh0W`0XOj_Og0V`0YOj_Og0V`0YOj_Og0V`0YOj_Og0V`0YOj_Og0V`0ZOj_Oe0W`0ZOi_Of0W`0ZOi_Of0W`0ZOi_Of0W`0ZOi_Of0W`0ZOj_Of0U`0[Oj_Oe0V`0[Oj_Oe0W`0ZOi_Of0W`0ZOi_Of0W`0ZOi_Of0W`0[Oh_Oe0X`0[Oi_Oe0V`0[Oj_Oe0V`0[Oj_Oe0W`0ZOi_Of0W`0ZOi_Of0W`0[Oh_Oe0X`0[Oh_Oe0X`0[Oi_Oe0V`0[OU_ONMg0n`0[OT_OONf0o`0ZOS_O0Ng0n`0ZOS_OOOg0n`0ZOS_OOOh0m`0YOT_OOOh0m`0YOT_OOOh0m`0YOT_OOOi0l`0XOU_OO0h0l`0YOS_OO1h0l`0YOS_OO1i0k`0XOT_OO1i0k`0XOT_OO1j0j`0WOV_OOOj0k`0WOV_OOOj0k`0WOV_OOOk0k`0VOU_O0Oj0l`0VOU_O00j0j`0VOV_O1Nj0l`0UOW_O0Mk0l`0UOW_O1Lk0l`0TOY_O0Jm0m`0SOY_O1_OW1Wa0iNV_Od1i`0\\NP_On1m`0RNQ_OS2m`0lMP_OZ2m`0fMQ_Oa2j`0_MS_Oh2i`08L2N2N2N2M3M3M3M3N1N3M3N1N3M2O1O2M2O1fNZ1nNR1O1O1g0YOk0VOa0^OO1O1O1O1N2O1O2N1O1O1O1O1N2O1O1O1N101O001N2O001N101O1O0O2O001N2O001N101O001N101N1O2O0O2O0O2O00001O001O001O001O010CoIPCQ6Q=<O02N8IO0001O0001O000001O0001O000000010O0000000010O00000000010O00001O01O01O001O00001O001O00001O0O2O00001O001O001O001N101O1O001O001O1N101N101O0O2O1N101N2O1N2O2M2O1N3N1N2O2M2O1O2M2O2M2O1N3N1N2N3L3N2N3M2N2N3M2M3N3M2M4L3M3M4L3M3M4L3M6J9G9G9D<_Oa0_OiQn1"}, "label": "airplane parked at an airport terminal gate"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"coffee table\"."}], "task": "refering", "answer_template": "The \"coffee table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [365, 366, 388, 410, 431, 432, 454, 455, 476, 477, 478, 479], "bbox": [0.6283986928104575, 0.7238398692810457, 0.8122712418300654, 1.0], "iscrowd": 0, "area": 4737.133300000002, "rle": {"size": [612, 612], "counts": "o`V7:jb0:F;]_OWOV>S1^A@P>k0dAk0g<_OnBb0R=cMXBo19`0_=aM[BW2J9k=_M`Bh4`=WKcBg4]=YKdBo2BeN\\>Y1hAdNY>[1jAbNW>FaA`0;HV>E`Ab0<HU>VO^AJ4V1:JU>RO`AL3W1:JT>POaAN3V19LT>lNdAO1Y18KT>jNeA10Y18Ld>hNWAZ17Mb>hNXAZ18MR?1o@OQ?KVA4j>G[A9f>@aA?_>[OgAe0Z>UOlAj0T>POSBo0Q`001O001O1O001N101O001O1O001O001O1O000000001O00000000000000001O000000000000001O000000000000001O000000000O1000001O0000000000000000O100000000000000000000O10Cd]O2\\b0Mg]O1Yb0Mj]O2Vb0Ln]O2Rb0LQ^O3ab0O2N[TT2"}, "label": "coffee table"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two men sitting behind red table\"."}], "task": "refering", "answer_template": "The \"two men sitting behind red table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 270, 271], "bbox": [0.013366013071895425, 0.3583660130718954, 0.3569117647058824, 0.5833660130718954], "iscrowd": 0, "area": 16289.718349999996, "rle": {"size": [612, 612], "counts": "gR5a0`b09Ge0[O5K6L0O2N101N1O1O2O010O01O010O01O010O01O010O001O01O01O00010O010O01O100O00100O1O010O1O10O01O100O00100O1O100O1O1O1O1O1O1O1O1N2O1L4O1O1O10O01O1O1O100O1O1O1O1O100O1O1O1O100O1O11O000000000000001O1O1O0`Nh@SOZ?b0o@^OS?`0m@@U?>k@nNk?P1V1N`^OROm`0l0Q_OYOn`0e0Q_OBk`06i^O^O:e0h`0HT_O@2Q1e`0[Oe_OP1T`0kNn_OW1R`0dNQ@^1i`0010000M3N4E:L2N2Dn]OFTb07=N2N2N21O4L5K4L4L4L3M001O00004L2N3M4I8F:G=D>D3L4M3M3M3N2O1N2N2O1N2N3N100O1O1O1O1O0000001O01O01O00010O000010O0001O01O0010O01O1O100O001VMm@b1S?SNXAm1P`0O1O100O100O1O101N1O1O1O1OR1mN5L3L4L4L5KiT[7"}, "label": "two men sitting behind red table"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two men , one in a red and blue plaid shirt looking at the camera and another in a blue shirt talking to a woman\"."}], "task": "refering", "answer_template": "The \"two men , one in a red and blue plaid shirt looking at the camera and another in a blue shirt talking to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 270, 271], "bbox": [0.013366013071895425, 0.3583660130718954, 0.3569117647058824, 0.5833660130718954], "iscrowd": 0, "area": 16289.718349999996, "rle": {"size": [612, 612], "counts": "gR5a0`b09Ge0[O5K6L0O2N101N1O1O2O010O01O010O01O010O01O010O001O01O01O00010O010O01O100O00100O1O010O1O10O01O100O00100O1O100O1O1O1O1O1O1O1O1N2O1L4O1O1O10O01O1O1O100O1O1O1O1O100O1O1O1O100O1O11O000000000000001O1O1O0`Nh@SOZ?b0o@^OS?`0m@@U?>k@nNk?P1V1N`^OROm`0l0Q_OYOn`0e0Q_OBk`06i^O^O:e0h`0HT_O@2Q1e`0[Oe_OP1T`0kNn_OW1R`0dNQ@^1i`0010000M3N4E:L2N2Dn]OFTb07=N2N2N21O4L5K4L4L4L3M001O00004L2N3M4I8F:G=D>D3L4M3M3M3N2O1N2N2O1N2N3N100O1O1O1O1O0000001O01O01O00010O000010O0001O01O0010O01O1O100O001VMm@b1S?SNXAm1P`0O1O100O100O1O101N1O1O1O1OR1mN5L3L4L4L5KiT[7"}, "label": "two men , one in a red and blue plaid shirt looking at the camera and another in a blue shirt talking to a woman"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in aqua blue v necked long sleeved shirt with purple writing on it seated on couch and appearing to be clapping\"."}], "task": "refering", "answer_template": "The \"man in aqua blue v necked long sleeved shirt with purple writing on it seated on couch and appearing to be clapping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 212, 213, 233, 234, 235, 236, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 298, 299, 300, 301, 302, 320, 321, 322, 323, 324, 342, 343, 365, 387, 409], "bbox": [0.561797385620915, 0.39738562091503266, 0.761797385620915, 0.8355882352941176], "iscrowd": 0, "area": 17385.486900000004, "rle": {"size": [612, 612], "counts": "To]65lb07I7H8I6J7E;E;F:E;F:E;E:G9F:F:H7M4M3L4M3L4\\CjJT;X5hDjJW;Z5dDiJ[;Z5`DhJ_;\\5\\DfJd;]5WDeJh;^5TDdJl;_5oCcJP<Y600O10O0100O01O0001M2N2MfNUDbKi;k2RD`Mb0[OY;X3^DQMf0\\Oi:g3kD`Li0^OZ:U4VEQL^<Q4lCcKT<`4V12O1O1O00100O1O00100O1O1O01000000O10O1000O1000O10O10000000O01000000O10O102O201N100O10M4L4L4L5K4L3M1O100O1O1O1O1O2N1O1N2N2M3N3L3N3M3L4M3L4M3L4M6JZ1eNi0XO\\lf2"}, "label": "man in aqua blue v necked long sleeved shirt with purple writing on it seated on couch and appearing to be clapping"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a blue t - shirt looking at the other person\"."}], "task": "refering", "answer_template": "The \"a man with a blue t - shirt looking at the other person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 212, 213, 233, 234, 235, 236, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 298, 299, 300, 301, 302, 320, 321, 322, 323, 324, 342, 343, 365, 387, 409], "bbox": [0.561797385620915, 0.39738562091503266, 0.761797385620915, 0.8355882352941176], "iscrowd": 0, "area": 17385.486900000004, "rle": {"size": [612, 612], "counts": "To]65lb07I7H8I6J7E;E;F:E;F:E;E:G9F:F:H7M4M3L4M3L4\\CjJT;X5hDjJW;Z5dDiJ[;Z5`DhJ_;\\5\\DfJd;]5WDeJh;^5TDdJl;_5oCcJP<Y600O10O0100O01O0001M2N2MfNUDbKi;k2RD`Mb0[OY;X3^DQMf0\\Oi:g3kD`Li0^OZ:U4VEQL^<Q4lCcKT<`4V12O1O1O00100O1O00100O1O1O01000000O10O1000O1000O10O10000000O01000000O10O102O201N100O10M4L4L4L5K4L3M1O100O1O1O1O1O2N1O1N2N2M3N3L3N3M3L4M3L4M3L4M6JZ1eNi0XO\\lf2"}, "label": "a man with a blue t - shirt looking at the other person"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a red polo shirt with blue lettering\"."}], "task": "refering", "answer_template": "The \"man in a red polo shirt with blue lettering\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 239, 240, 241, 260, 261, 262, 263, 282, 283, 284, 285, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 346, 347, 350], "bbox": [0.7540849673202614, 0.4294117647058824, 1.0, 0.6995098039215687], "iscrowd": 0, "area": 13512.482749999996, "rle": {"size": [612, 612], "counts": "Z`d8>db04L3L4M4K4M3M2O1O1O1O1O1O1O1O100O2N1O1O1O2N1O1O2O0O1O1OBi^OROUa0n0U_OiNj`0X1`01O001O1O0O2M3M3M3N2M2N3N2O1N2O1N2O1N2O0O2O2M2O1N2N2O2N1O1O101N1O2N101N1O2N101N1O2O0O2N1O2O0O2N10000000L4L301000000000O1000001O1O001O1O0@dLXA]3e>iLXAW3e>oLWAR3f>g0M4M3M3L3N3N101O1OROkAmLU>R3mAnLR>Q3PBnLP>R3QBnLm=R3UBmLk=R3WBnLh=Q3ZBnLf=Q3\\BnLd=R3]BmLc=R3_BmL`=S3bBkL_=T3cBjL^=U3dBhL^=X3cBfL^=Y3W101O1O1N2O1O1O1O0O2O1bMZ@\\1g?aN[@^1h?^N[@`1i?ZNZ@f1h?VN[@h1i?SNY@m1i?oMY@P2_`0O0O2J5E<E>A`0A[F"}, "label": "man in a red polo shirt with blue lettering"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on the far right on a couch\"."}], "task": "refering", "answer_template": "The \"a man on the far right on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 239, 240, 241, 260, 261, 262, 263, 282, 283, 284, 285, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 346, 347, 350], "bbox": [0.7540849673202614, 0.4294117647058824, 1.0, 0.6995098039215687], "iscrowd": 0, "area": 13512.482749999996, "rle": {"size": [612, 612], "counts": "Z`d8>db04L3L4M4K4M3M2O1O1O1O1O1O1O1O100O2N1O1O1O2N1O1O2O0O1O1OBi^OROUa0n0U_OiNj`0X1`01O001O1O0O2M3M3M3N2M2N3N2O1N2O1N2O1N2O0O2O2M2O1N2N2O2N1O1O101N1O2N101N1O2N101N1O2O0O2N1O2O0O2N10000000L4L301000000000O1000001O1O001O1O0@dLXA]3e>iLXAW3e>oLWAR3f>g0M4M3M3L3N3N101O1OROkAmLU>R3mAnLR>Q3PBnLP>R3QBnLm=R3UBmLk=R3WBnLh=Q3ZBnLf=Q3\\BnLd=R3]BmLc=R3_BmL`=S3bBkL_=T3cBjL^=U3dBhL^=X3cBfL^=Y3W101O1O1N2O1O1O1O0O2O1bMZ@\\1g?aN[@^1h?^N[@`1i?ZNZ@f1h?VN[@h1i?SNY@m1i?oMY@P2_`0O0O2J5E<E>A`0A[F"}, "label": "a man on the far right on a couch"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with dark brown hair sitting between two men on a black couch\"."}], "task": "refering", "answer_template": "The \"a women with dark brown hair sitting between two men on a black couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 208, 209, 210, 211, 229, 230, 231, 232, 233, 251, 252, 253, 273, 274, 275, 295, 296, 317, 318, 339, 340, 361, 362, 383, 384], "bbox": [0.41142156862745094, 0.39540849673202616, 0.6460457516339869, 0.8086274509803921], "iscrowd": 0, "area": 9718.525349999998, "rle": {"size": [612, 612], "counts": "_nf4e0Wb0;L2N2N3M2N2N3MUO\\^O9k`0UOj_Ob0\\O:d`0[Om_O;A;Z`0CQ@1G=l2TOU:a0UCNK?g2VOV:a0TCJ1?b2YOW:a0SCF6a0f0XO<3c;b0QCC;a0=D;Ij;S1^CA605CU<m0aC@N>L]Od<f0bCAFh1g<hNdC`2[<aMeC^2[<cMfC\\2Y<eMgC[2X<fMiCY2V<hMjCX2U<iMlCV2S<kMmCU2S<kMoCS2P<nMRDP2m;QNUDm1j;TNWDk1h;VNZDn0`<RObCg0b<ZO`C>e<C]C6a<1aCEW<k0SD\\N[<e1`22N3M2N2N2O100O1O1O10000000001O1O001O100O10O10O1000O10O10N2O001O1N2O2N2N2YOU@PNP`0e1[@WNj?]1`@`Ne?S1g@iN[?l0c1G9I610000O1000000O0100000O1000000O1000O10O10000O1000000O1000O10O100000000001O00001N101O00001O00001O000O2O001O00001O000O2N1N2O2Mb]Q4"}, "label": "a women with dark brown hair sitting between two men on a black couch"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady sitting on a couch\"."}], "task": "refering", "answer_template": "The \"a lady sitting on a couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 208, 209, 210, 211, 229, 230, 231, 232, 233, 251, 252, 253, 273, 274, 275, 295, 296, 317, 318, 339, 340, 361, 362, 383, 384], "bbox": [0.41142156862745094, 0.39540849673202616, 0.6460457516339869, 0.8086274509803921], "iscrowd": 0, "area": 9718.525349999998, "rle": {"size": [612, 612], "counts": "_nf4e0Wb0;L2N2N3M2N2N3MUO\\^O9k`0UOj_Ob0\\O:d`0[Om_O;A;Z`0CQ@1G=l2TOU:a0UCNK?g2VOV:a0TCJ1?b2YOW:a0SCF6a0f0XO<3c;b0QCC;a0=D;Ij;S1^CA605CU<m0aC@N>L]Od<f0bCAFh1g<hNdC`2[<aMeC^2[<cMfC\\2Y<eMgC[2X<fMiCY2V<hMjCX2U<iMlCV2S<kMmCU2S<kMoCS2P<nMRDP2m;QNUDm1j;TNWDk1h;VNZDn0`<RObCg0b<ZO`C>e<C]C6a<1aCEW<k0SD\\N[<e1`22N3M2N2N2O100O1O1O10000000001O1O001O100O10O10O1000O10O10N2O001O1N2O2N2N2YOU@PNP`0e1[@WNj?]1`@`Ne?S1g@iN[?l0c1G9I610000O1000000O0100000O1000000O1000O10O10000O1000000O1000O10O100000000001O00001N101O00001O00001O000O2O001O00001O000O2N1N2O2Mb]Q4"}, "label": "a lady sitting on a couch"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red table with dirty white plates and a pizza tin\"."}], "task": "refering", "answer_template": "The \"a red table with dirty white plates and a pizza tin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 380, 381, 398, 399, 400, 403, 420, 421, 442, 443, 464, 465], "bbox": [0.0, 0.5303431372549019, 0.35055555555555556, 1.0], "iscrowd": 0, "area": 30258.718800000002, "rle": {"size": [612, 612], "counts": "l:c1`a03N1O1N2O2N1N2O2N1N2O1N3N1O1N3N1O1N2O2N1N2O2N1N2O1N3N1O1N3N1O1N2O2N1N2O2N1N2O1O2N1O1N3N1O1O1O2N1O1N3N1O1O1O2N1O1N3N1O1O1O2N1O4K:G:F9G:F:F9G:E:G:F9G:F:F9G5J1000000000000O10000000000O1000000000000O1000000000]M]FeKc9Z4YHkIg7U6e2N1O2O0O2O0O2N1ZOTCTKm<j4[CnJf<Q5aChJa<V5g0O2N1O2N1O2N1O3M3L4M4L3M3M4L3M3M4L3M3M3M4L3M3M4L3M2N1N2O1O2N1O1O1O2N1O1OK5K5K5K5J6K5K55K4K3N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O1O2N11O`0@a0_Oa0_Oa0@?@9G00000001O000000000mMT2jNU1O1O1O1O1O2N1O<DV1jNVl\\7"}, "label": "a red table with dirty white plates and a pizza tin"}]}
{"id": 442298, "image": "COCO_train2014_000000442298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a red cloth drapped over it\"."}], "task": "refering", "answer_template": "The \"a table with a red cloth drapped over it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 334, 335, 336, 337, 353, 354, 355, 356, 357, 358, 359, 376, 377, 378, 380, 381, 398, 399, 400, 403, 420, 421, 442, 443, 464, 465], "bbox": [0.0, 0.5303431372549019, 0.35055555555555556, 1.0], "iscrowd": 0, "area": 30258.718800000002, "rle": {"size": [612, 612], "counts": "l:c1`a03N1O1N2O2N1N2O2N1N2O1N3N1O1N3N1O1N2O2N1N2O2N1N2O1N3N1O1N3N1O1N2O2N1N2O2N1N2O1O2N1O1N3N1O1O1O2N1O1N3N1O1O1O2N1O1N3N1O1O1O2N1O4K:G:F9G:F:F9G:E:G:F9G:F:F9G5J1000000000000O10000000000O1000000000000O1000000000]M]FeKc9Z4YHkIg7U6e2N1O2O0O2O0O2N1ZOTCTKm<j4[CnJf<Q5aChJa<V5g0O2N1O2N1O2N1O3M3L4M4L3M3M4L3M3M4L3M3M3M4L3M3M4L3M2N1N2O1O2N1O1O1O2N1O1OK5K5K5K5J6K5K55K4K3N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O1O2N11O`0@a0_Oa0_Oa0@?@9G00000001O000000000mMT2jNU1O1O1O1O1O2N1O<DV1jNVl\\7"}, "label": "a table with a red cloth drapped over it"}]}
{"id": 278461, "image": "COCO_train2014_000000278461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow color tie in the alternate persons\"."}], "task": "refering", "answer_template": "The \"yellow color tie in the alternate persons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 199, 200, 208, 216, 217, 222, 223, 230, 231, 245, 246, 261, 262, 268, 284, 285, 307, 330], "bbox": [0.0408125, 0.4716458333333333, 0.72959375, 0.8888958333333332], "iscrowd": 0, "area": 4503.46445, "rle": {"size": [480, 640], "counts": "X_<c0X>4K4L5K4L5L301O0010O01O02mNaBd0c=VOfBb0S>I7H8Ihhg24oVXM7J7H8I7I6I8I7H7J3M2M4M2N3L10O001O3L;F9G:E:G:F9G9Fgn6<SQI10O010O010O10O10O010O010O10O10O010O10O10O010O010O10O10N1Jncj1]1kZUNa0^O10O1O010O010O010O00100O010O010O00CnMoCR2S<RNfCn1\\<:1N2O2M2N2O1N2O2M2O1`N`1O1N3N1N2O1N2OkZa2"}, "label": "yellow color tie in the alternate persons"}]}
{"id": 204735, "image": "COCO_train2014_000000204735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bag\"."}], "task": "refering", "answer_template": "The \"a bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 27, 28, 29, 30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 100, 101], "bbox": [0.161265625, 0.0043124999999999995, 0.544046875, 0.2530208333333333], "iscrowd": 0, "area": 19401.487699999998, "rle": {"size": [480, 640], "counts": "^Y`11n>3M3M2N3L4M3M2N3N2N2N2N4L5K5eBgNe<^1SCeNm<h1O0O2O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O000000O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O1N2M3N2N2M3N2M3N2N2M3N2N2M3N1O2M3N2O1O100O1O1O100O1O1O100O1O1O100O1O1O100O10000O1000000000000O1000000000000O1O1N2N2M3N3M3M2N3L4M2NlkX4"}, "label": "a bag"}]}
{"id": 204735, "image": "COCO_train2014_000000204735.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black travel bag with blue and white graphic design\"."}], "task": "refering", "answer_template": "The \"black travel bag with blue and white graphic design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 27, 28, 29, 30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 100, 101], "bbox": [0.161265625, 0.0043124999999999995, 0.544046875, 0.2530208333333333], "iscrowd": 0, "area": 19401.487699999998, "rle": {"size": [480, 640], "counts": "^Y`11n>3M3M2N3L4M3M2N3N2N2N2N4L5K5eBgNe<^1SCeNm<h1O0O2O001O00001O001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O000000O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O1N2M3N2N2M3N2M3N2N2M3N2N2M3N1O2M3N2O1O100O1O1O100O1O1O100O1O1O100O1O1O100O10000O1000000000000O1000000000000O1O1N2N2M3N3M3M2N3L4M2NlkX4"}, "label": "black travel bag with blue and white graphic design"}]}
{"id": 221119, "image": "COCO_train2014_000000221119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a twirling woman smoking a cigarette\"."}], "task": "refering", "answer_template": "The \"a twirling woman smoking a cigarette\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 197, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 337, 338, 359, 360, 361, 381, 382, 383, 384, 403, 404, 405, 406, 407, 427, 429, 430, 431, 450], "bbox": [0.5571830985915494, 0.170046875, 0.8053051643192489, 0.84796875], "iscrowd": 0, "area": 33789.69050000001, "rle": {"size": [640, 639], "counts": "Pon62nc03L3N3L4M2M4ZG\\OiMh0R2\\OlMg0P2]OlMf0P2^OnMe0n1^OoMf0l1^ORNe0i1_OTNe0h1\\OgFXOP7_1T2[OiF[OP7]1S2ZOkF]OP7\\1Q2XOmF@o6\\1o1VOPGBn6\\1n1SORGEn6[1k1ROUGGm6[1i1POXGIm6Z1g1lN\\GMk6X1g1kN^GOk6V1f1jN_G2j6V1e10ZN0e11ZN0e12YNOf12YNOf12YNOf13XNZO^IgNY8P2XNVOhIeNn7X2XNoNTJeN_7`2\\NhN]JdNT7g2]NbNiJbNf6Q3TKeMDg0[6j1aIoM37U6S2_InM;OS6W2ZISN`0FT6T2[I^N`0^OR6o1_IlN=UOS6Y3PHjMl1lNT6]6iIdIV6^6iIbIW6`6gI_IZ6b6dI_I\\6b6cI^I]6d6aI[I`6f6_IZIa6h6\\IYIe6g6ZIXIg6j6WIVIj6j6UIUIm6l6PIUIP7l6oHTIR7l6mHSIU7n6hHSIX7n6gHQI[7P7cHoH_7Q7`HoH[OjNY7Y8YImH^O[Oj6i7fIlH_O@h6f7fIkH@Ai6g7dIlH^O@m6e7bIPI\\O^OQ7d7^ITI\\O[OT7d7YIbJf6_5TIgJk6[5nHkJP7T9O0O2O000001O00001O001O00001O001O0010O00VNjHjFW7P9WIgFi6S9eIeF\\6U9RJbFn5Y9_J_Fb5[9fJbFZ5X9mJgFT5S9RKlFP5_1eH_5\\2RIQ5]1eH]5]2UIo4\\1gH\\5\\2WIo4[1gH[5[2ZIP5Y1gHY5\\2]Io4W1hHY5[2_Io4V1hHX5Z2bIP5T1hHV5[2eIo4T1gHQ5_2kIk4S1gHd4k2XJ`4S1fHW4V3fJV4Q1fHn3^3PKm3R1fHl3]3RKo3Q1eHj3]3UKP4P1dHi3\\3WKQ4o0eHh3Z3YKR4o0eHf3Y3[KS4o0eHc3Y3^KS4o0eHa3X3`KT4n0hH^3T3dKU4n0jHY3R3iKV4m0kHV3o2mKW4m0mHR3l2QLX4m0oHm2j2VLX4l0RIj2c;VM`Df2a;YMbDd2_;iLiBMl1V3];mLgBLn1U3\\;oLfBKP2T3[;PMfBJS2R3Y;TMdBIU2Q3Y;UMaBIZ2o2V;XM`BH\\2m2V;ZM_BH^2k2T;\\M_BG`2k2R;]M_BGb2i2P;^M`BGc2h2o:`M_BGe2f2m:bM_BGf2e2l:cM_EZ2c:eM`EX2a:fMcEW2^:hMdEU2^:jMeES2\\:lMfER2\\:lMgEP2W>N3M3M2M4M2N3L3N3M3L3\\OV]ONmb0N^]OHdb04h0L3MZU]2"}, "label": "a twirling woman smoking a cigarette"}]}
{"id": 221119, "image": "COCO_train2014_000000221119.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing on a platform while wearing leopard skin tights and smoking a cigarrette\"."}], "task": "refering", "answer_template": "The \"a woman standing on a platform while wearing leopard skin tights and smoking a cigarrette\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 197, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 271, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 337, 338, 359, 360, 361, 381, 382, 383, 384, 403, 404, 405, 406, 407, 427, 429, 430, 431, 450], "bbox": [0.5571830985915494, 0.170046875, 0.8053051643192489, 0.84796875], "iscrowd": 0, "area": 33789.69050000001, "rle": {"size": [640, 639], "counts": "Pon62nc03L3N3L4M2M4ZG\\OiMh0R2\\OlMg0P2]OlMf0P2^OnMe0n1^OoMf0l1^ORNe0i1_OTNe0h1\\OgFXOP7_1T2[OiF[OP7]1S2ZOkF]OP7\\1Q2XOmF@o6\\1o1VOPGBn6\\1n1SORGEn6[1k1ROUGGm6[1i1POXGIm6Z1g1lN\\GMk6X1g1kN^GOk6V1f1jN_G2j6V1e10ZN0e11ZN0e12YNOf12YNOf12YNOf13XNZO^IgNY8P2XNVOhIeNn7X2XNoNTJeN_7`2\\NhN]JdNT7g2]NbNiJbNf6Q3TKeMDg0[6j1aIoM37U6S2_InM;OS6W2ZISN`0FT6T2[I^N`0^OR6o1_IlN=UOS6Y3PHjMl1lNT6]6iIdIV6^6iIbIW6`6gI_IZ6b6dI_I\\6b6cI^I]6d6aI[I`6f6_IZIa6h6\\IYIe6g6ZIXIg6j6WIVIj6j6UIUIm6l6PIUIP7l6oHTIR7l6mHSIU7n6hHSIX7n6gHQI[7P7cHoH_7Q7`HoH[OjNY7Y8YImH^O[Oj6i7fIlH_O@h6f7fIkH@Ai6g7dIlH^O@m6e7bIPI\\O^OQ7d7^ITI\\O[OT7d7YIbJf6_5TIgJk6[5nHkJP7T9O0O2O000001O00001O001O00001O001O0010O00VNjHjFW7P9WIgFi6S9eIeF\\6U9RJbFn5Y9_J_Fb5[9fJbFZ5X9mJgFT5S9RKlFP5_1eH_5\\2RIQ5]1eH]5]2UIo4\\1gH\\5\\2WIo4[1gH[5[2ZIP5Y1gHY5\\2]Io4W1hHY5[2_Io4V1hHX5Z2bIP5T1hHV5[2eIo4T1gHQ5_2kIk4S1gHd4k2XJ`4S1fHW4V3fJV4Q1fHn3^3PKm3R1fHl3]3RKo3Q1eHj3]3UKP4P1dHi3\\3WKQ4o0eHh3Z3YKR4o0eHf3Y3[KS4o0eHc3Y3^KS4o0eHa3X3`KT4n0hH^3T3dKU4n0jHY3R3iKV4m0kHV3o2mKW4m0mHR3l2QLX4m0oHm2j2VLX4l0RIj2c;VM`Df2a;YMbDd2_;iLiBMl1V3];mLgBLn1U3\\;oLfBKP2T3[;PMfBJS2R3Y;TMdBIU2Q3Y;UMaBIZ2o2V;XM`BH\\2m2V;ZM_BH^2k2T;\\M_BG`2k2R;]M_BGb2i2P;^M`BGc2h2o:`M_BGe2f2m:bM_BGf2e2l:cM_EZ2c:eM`EX2a:fMcEW2^:hMdEU2^:jMeES2\\:lMfER2\\:lMgEP2W>N3M3M2M4M2N3L3N3M3L3\\OV]ONmb0N^]OHdb04h0L3MZU]2"}, "label": "a woman standing on a platform while wearing leopard skin tights and smoking a cigarrette"}]}
{"id": 532419, "image": "COCO_train2014_000000532419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in the background looking off camera\"."}], "task": "refering", "answer_template": "The \"a zebra in the background looking off camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 186, 187, 188, 209, 210, 211, 232, 233, 234, 235, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327, 348, 349, 350, 371, 372, 373], "bbox": [0.1014375, 0.4205625, 0.25990625, 0.9839791666666666], "iscrowd": 0, "area": 13081.286499999995, "rle": {"size": [480, 640], "counts": "len05i>;F2N2O0O2O1O0SDIQ97mFMQ94lF0R90kF5S9K`E^O;n0T:DZEF<l0X:^OXEN8h0`:[OVE4Of0k:WOTEP2l:PNSEQ2m:QNPEP2R;n01O1O001O001O0O10O4K8He0\\O4K5K4L5L3N3L4L4L4M3M5ZG]Jc0Lf5g6oIbIP6a6fIfIZ6]6\\IjId6Y7O1001O001O00001N2lNPJUIR6i6UJPIm5m6U1[OPHVJS8d5SHZJQ8]5n0]Od0F9J7I30O2O0O2N2O1O1O1O2N1O`LTFn1k9hMjFn1U9hMaGm1^8jMVHl1i7nMhHh1X7UNPIf1o6YNYIa1f6]NbI^1]6aNjIZ1U6dNPJ[1l9O2O0000fCjNh:h21O00001b0^OXMaDQ2Z;_MVEd2c;M3M2M4Mb1^N9G9Fegm6"}, "label": "a zebra in the background looking off camera"}]}
{"id": 532419, "image": "COCO_train2014_000000532419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra looking away\"."}], "task": "refering", "answer_template": "The \"a zebra looking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 186, 187, 188, 209, 210, 211, 232, 233, 234, 235, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327, 348, 349, 350, 371, 372, 373], "bbox": [0.1014375, 0.4205625, 0.25990625, 0.9839791666666666], "iscrowd": 0, "area": 13081.286499999995, "rle": {"size": [480, 640], "counts": "len05i>;F2N2O0O2O1O0SDIQ97mFMQ94lF0R90kF5S9K`E^O;n0T:DZEF<l0X:^OXEN8h0`:[OVE4Of0k:WOTEP2l:PNSEQ2m:QNPEP2R;n01O1O001O001O0O10O4K8He0\\O4K5K4L5L3N3L4L4L4M3M5ZG]Jc0Lf5g6oIbIP6a6fIfIZ6]6\\IjId6Y7O1001O001O00001N2lNPJUIR6i6UJPIm5m6U1[OPHVJS8d5SHZJQ8]5n0]Od0F9J7I30O2O0O2N2O1O1O1O2N1O`LTFn1k9hMjFn1U9hMaGm1^8jMVHl1i7nMhHh1X7UNPIf1o6YNYIa1f6]NbI^1]6aNjIZ1U6dNPJ[1l9O2O0000fCjNh:h21O00001b0^OXMaDQ2Z;_MVEd2c;M3M2M4Mb1^N9G9Fegm6"}, "label": "a zebra looking away"}]}
{"id": 466885, "image": "COCO_train2014_000000466885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sky chefs truck is seen in the bottom right corner of the screen\"."}], "task": "refering", "answer_template": "The \"a sky chefs truck is seen in the bottom right corner of the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6859531249999999, 0.7701041666666666, 0.9994375, 0.9858333333333333], "iscrowd": 0, "area": 18284.045399999995, "rle": {"size": [480, 640], "counts": "WV^6f0Q>9000001O00000000000`BTOl<l0PCYOo<g0lB_OS=a0hBEW=S1N2O0O10000O1000000O10001POVNgDk1X;]NaDc1^;eN[D[1d;lNVDT1i;R1O1O1O1O1O100000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0O10001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000\\MaF"}, "label": "a sky chefs truck is seen in the bottom right corner of the screen"}]}
{"id": 466885, "image": "COCO_train2014_000000466885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skychefs van standing left to the other van of skychefs\"."}], "task": "refering", "answer_template": "The \"a skychefs van standing left to the other van of skychefs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.7166041666666667, 0.39221875, 0.9982291666666667], "iscrowd": 0, "area": 27713.13054999999, "rle": {"size": [480, 640], "counts": "Q=P1\\=e0M3N1N3hNZNSEh1k:]NoDf1o:`NkDa1T;eNeD^1Y;gNbD[1\\;kN^DW1`;U101N1O2N1000001O0000001O00000010O000001O0000001O0000010O000000010O00000010O0000010O000001O01O0001O0001O01O0000010O00000010O000001O0000010O0000000010O000001O0001O0000000000001O0001O000001O0000000000001O001O00001O00001O00000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000010O0000000000001O00000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O00000000001O00Roe5"}, "label": "a skychefs van standing left to the other van of skychefs"}]}
{"id": 466885, "image": "COCO_train2014_000000466885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sky chefs vehicle on the left\"."}], "task": "refering", "answer_template": "The \"the sky chefs vehicle on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.7166041666666667, 0.39221875, 0.9982291666666667], "iscrowd": 0, "area": 27713.13054999999, "rle": {"size": [480, 640], "counts": "Q=P1\\=e0M3N1N3hNZNSEh1k:]NoDf1o:`NkDa1T;eNeD^1Y;gNbD[1\\;kN^DW1`;U101N1O2N1000001O0000001O00000010O000001O0000001O0000010O000000010O00000010O0000010O000001O01O0001O0001O01O0000010O00000010O000001O0000010O0000000010O000001O0001O0000000000001O0001O000001O0000000000001O001O00001O00001O00000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O0000000000001O00000000010O0000000000001O00000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O00000000001O00Roe5"}, "label": "the sky chefs vehicle on the left"}]}
{"id": 466885, "image": "COCO_train2014_000000466885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tail end of an airplane at night\"."}], "task": "refering", "answer_template": "The \"the tail end of an airplane at night\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 66, 67, 76, 77, 78, 79, 80, 81, 82, 83, 89, 90, 91, 99, 100, 101, 102, 103, 104, 105, 106, 107, 113, 114, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.142859375, 0.0030208333333333333, 1.0, 0.5608541666666667], "iscrowd": 0, "area": 73210.9032, "rle": {"size": [480, 640], "counts": "TW\\11o>00001O001O00001O001O00001O00001N101O00001O00001O001O00001O00001O001O00001O001N10001O00001O001O00001O00001O001O00001O00001O0O2O00001O001O00001O00001O001O00001O00001O001O0PKoNPLR1P4QOmKQ1Q4ROlKo0S4TOjKn0U4TOhKm0W4VOfKl0X4WOeKj0Z4XOdKj0Z4YOcKi0[4ZObKg0]4\\O`Kf0^4]O_Kd0`4_O]Kc0a4@\\Ka0c4BZK`0d4CYK>f4EWK=g4FVK<h4GUK:j4ISK9k4JRK7m4LPK6n4MoJ4P5NmJ5Q5NlJ3S50jJ2T51iJ0V53gJOW54fJNX55eJLZ57cJJ\\59aJG_5<^JDb5?[JAe5b0XJ^Oh5e0UJ[Ok5h0RJXOn5k0oIUOQ6n0lISOS6o0kIQOU6R1hInNX6U1eIkN[6X1bIhN^6[1_IeNa6^1\\IbNd6a1YI_Ng6d1UI]Nk6f1RIZNo6h1nHXNR7k1kHUNU7n1hHRNX7Q2eHoM[7T2bHlM^7l4O0000000000000000000000000000001O00000000000000000000000000001O00000000000001O00000000001O1O100O1O1O1O1O1O001O1O1O1O1O1O1O100O001O1O1O1O1O1O1O1O1O001O1O1O100O1O1O1O1O001O1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O1O1O001O1O1O100O1O1O1O1O001O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O001O1O1O1O1O100O1O1O001O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O001O1O1O1O1O1O100O1O001O1O1O1O1O1O1O1O1O001O1O100O1O1O1O1O1O001O1O1O1O1O1O1O100O001O1O1O1O1O1O1O1O1O001O1O0001O000000000001O0000000000000001O0000000000000000000000000000001O00000001O00000000000000000001O0000000000000000000000000000001O01O00000000000000000000000001O000000000000000000000000000010O000000000000000000000000000001O0000000000000mLjL\\JV3d5lLZJT3g5nLVJR3j5QMSJo2m5SMQJm2o5VMnIj2R6YMlIf2T6]MiId2V6^MhIb2X6aMeI_2[6dMbI\\2^6fM`IZ2`6iM]IW2c6lMZIT2f6oMWIQ2i6QNUIo1k6TNRIl1n6WNoHi1Q7YNmHg1S7\\NjHd1W7^NfHb1Z7aNcH_1]7cNaH]1`7eN]H\\1c7fNZHZ1g7gNWHY1j7iNSHW1n7kNoGU1R8lNlGT1U8nNhGR1Y8POdGP1]8RO`Gn0a8SO]Gm0d8UOYGk0h8WOVGh0k8YOSGg0n8[OoFe0R9d2100O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1nNR1aN_1aNjM"}, "label": "the tail end of an airplane at night"}]}
{"id": 466885, "image": "COCO_train2014_000000466885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flight in the airport for dispatch\"."}], "task": "refering", "answer_template": "The \"a flight in the airport for dispatch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 66, 67, 76, 77, 78, 79, 80, 81, 82, 83, 89, 90, 91, 99, 100, 101, 102, 103, 104, 105, 106, 107, 113, 114, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.142859375, 0.0030208333333333333, 1.0, 0.5608541666666667], "iscrowd": 0, "area": 73210.9032, "rle": {"size": [480, 640], "counts": "TW\\11o>00001O001O00001O001O00001O00001N101O00001O00001O001O00001O00001O001O00001O001N10001O00001O001O00001O00001O001O00001O00001O0O2O00001O001O00001O00001O001O00001O00001O001O0PKoNPLR1P4QOmKQ1Q4ROlKo0S4TOjKn0U4TOhKm0W4VOfKl0X4WOeKj0Z4XOdKj0Z4YOcKi0[4ZObKg0]4\\O`Kf0^4]O_Kd0`4_O]Kc0a4@\\Ka0c4BZK`0d4CYK>f4EWK=g4FVK<h4GUK:j4ISK9k4JRK7m4LPK6n4MoJ4P5NmJ5Q5NlJ3S50jJ2T51iJ0V53gJOW54fJNX55eJLZ57cJJ\\59aJG_5<^JDb5?[JAe5b0XJ^Oh5e0UJ[Ok5h0RJXOn5k0oIUOQ6n0lISOS6o0kIQOU6R1hInNX6U1eIkN[6X1bIhN^6[1_IeNa6^1\\IbNd6a1YI_Ng6d1UI]Nk6f1RIZNo6h1nHXNR7k1kHUNU7n1hHRNX7Q2eHoM[7T2bHlM^7l4O0000000000000000000000000000001O00000000000000000000000000001O00000000000001O00000000001O1O100O1O1O1O1O1O001O1O1O1O1O1O1O100O001O1O1O1O1O1O1O1O1O001O1O1O100O1O1O1O1O001O1O1O1O1O1O1O1O10O01O1O1O1O1O1O1O1O1O001O1O1O100O1O1O1O1O001O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O001O1O1O1O1O100O1O1O001O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O001O1O1O1O1O1O100O1O001O1O1O1O1O1O1O1O1O001O1O100O1O1O1O1O1O001O1O1O1O1O1O1O100O001O1O1O1O1O1O1O1O1O001O1O0001O000000000001O0000000000000001O0000000000000000000000000000001O00000001O00000000000000000001O0000000000000000000000000000001O01O00000000000000000000000001O000000000000000000000000000010O000000000000000000000000000001O0000000000000mLjL\\JV3d5lLZJT3g5nLVJR3j5QMSJo2m5SMQJm2o5VMnIj2R6YMlIf2T6]MiId2V6^MhIb2X6aMeI_2[6dMbI\\2^6fM`IZ2`6iM]IW2c6lMZIT2f6oMWIQ2i6QNUIo1k6TNRIl1n6WNoHi1Q7YNmHg1S7\\NjHd1W7^NfHb1Z7aNcH_1]7cNaH]1`7eN]H\\1c7fNZHZ1g7gNWHY1j7iNSHW1n7kNoGU1R8lNlGT1U8nNhGR1Y8POdGP1]8RO`Gn0a8SO]Gm0d8UOYGk0h8WOVGh0k8YOSGg0n8[OoFe0R9d2100O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1nNR1aN_1aNjM"}, "label": "a flight in the airport for dispatch"}]}
{"id": 393159, "image": "COCO_train2014_000000393159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flight parked at an air port\"."}], "task": "refering", "answer_template": "The \"flight parked at an air port\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 113, 114], "bbox": [0.33089583333333333, 0.156640625, 0.7812083333333334, 0.30153125000000003], "iscrowd": 0, "area": 6908.612650000002, "rle": {"size": [640, 480], "counts": "o`S3b0Yc05000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000O10000001O001O1O001O1O00001O00001O000000000000001O00000000001O0000001O0T]OSOeb0m0Y]OUOgb0k0X]OWOgb0Q100000000000000000000000O100HX]OXOhb0i0[]OSOeb0m0700001N10000010O000001O00001O01O0001O0000000001N1000000O1000000O2O0000000O1000000O10O1000O1000000O1000O1O1O1O1O0O2O1O1O1N2O1O001N2O1O1O1N101O1O1N101O1O0O2O1O1O0O2O1O001N2O001OCQ^OkNma0U1U^OkNja0T1Y^OjNga0U1[^OkNda0T1_^OkN`a0o0g^OPOXa0o0l^OPOSa0n0P_OROo`0l0U_OSOk`0k0W_OTOj`0i0Y_OWOg`0g0\\_OXOd`0f0^_OYOc`0f0__OXOb`0f0]1N2N1O2N2M3K5JRaQ2"}, "label": "flight parked at an air port"}]}
{"id": 393159, "image": "COCO_train2014_000000393159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane with the word martinair on the side\"."}], "task": "refering", "answer_template": "The \"an airplane with the word martinair on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 107, 108, 109, 110, 111, 112, 113, 114], "bbox": [0.33089583333333333, 0.156640625, 0.7812083333333334, 0.30153125000000003], "iscrowd": 0, "area": 6908.612650000002, "rle": {"size": [640, 480], "counts": "o`S3b0Yc05000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000O10000001O001O1O001O1O00001O00001O000000000000001O00000000001O0000001O0T]OSOeb0m0Y]OUOgb0k0X]OWOgb0Q100000000000000000000000O100HX]OXOhb0i0[]OSOeb0m0700001N10000010O000001O00001O01O0001O0000000001N1000000O1000000O2O0000000O1000000O10O1000O1000000O1000O1O1O1O1O0O2O1O1O1N2O1O001N2O1O1O1N101O1O1N101O1O0O2O1O1O0O2O1O001N2O001OCQ^OkNma0U1U^OkNja0T1Y^OjNga0U1[^OkNda0T1_^OkN`a0o0g^OPOXa0o0l^OPOSa0n0P_OROo`0l0U_OSOk`0k0W_OTOj`0i0Y_OWOg`0g0\\_OXOd`0f0^_OYOc`0f0__OXOb`0f0]1N2N1O2N2M3K5JRaQ2"}, "label": "an airplane with the word martinair on the side"}]}
{"id": 393159, "image": "COCO_train2014_000000393159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver sink on the right\"."}], "task": "refering", "answer_template": "The \"a silver sink on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 318, 319, 320, 321, 322], "bbox": [0.6544791666666666, 0.660578125, 1.0, 0.8180624999999999], "iscrowd": 0, "area": 13602.633249999999, "rle": {"size": [640, 480], "counts": "\\fT63lc0:E7J2M3N2M2N3N2N2N2O1N2O0O2O1N2O1N2O1N2O0O2N2O1N2O1N2O0O2O1O1N2O1O1O1N101O1O1N101O0O1000000O2O00000O101O000O10000O2O00000O101O000O10000O2O00000O101O0O100000001O000000001O00000000001O000000001O00000000010O00000000001O000000001O01O00000001OO100000000000001O0000000000000O10000000001O00000000000000000O1000001O0000000000000000000O10dB"}, "label": "a silver sink on the right"}]}
{"id": 393159, "image": "COCO_train2014_000000393159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sink basin to the right of the hole\"."}], "task": "refering", "answer_template": "The \"the sink basin to the right of the hole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 318, 319, 320, 321, 322], "bbox": [0.6544791666666666, 0.660578125, 1.0, 0.8180624999999999], "iscrowd": 0, "area": 13602.633249999999, "rle": {"size": [640, 480], "counts": "\\fT63lc0:E7J2M3N2M2N3N2N2N2O1N2O0O2O1N2O1N2O1N2O0O2N2O1N2O1N2O0O2O1O1N2O1O1O1N101O1O1N101O0O1000000O2O00000O101O000O10000O2O00000O101O000O10000O2O00000O101O0O100000001O000000001O00000000001O000000001O00000000010O00000000001O000000001O01O00000001OO100000000000001O0000000000000O10000000001O00000000000000000O1000001O0000000000000000000O10dB"}, "label": "the sink basin to the right of the hole"}]}
{"id": 204746, "image": "COCO_train2014_000000204746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rear end of a baby cow on the other side of the cow that the woman is milking\"."}], "task": "refering", "answer_template": "The \"rear end of a baby cow on the other side of the cow that the woman is milking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 105, 106, 107, 108, 109, 128, 131, 132, 151, 154, 155, 174, 177], "bbox": [0.5619375, 0.13491803278688524, 0.7508593749999999, 0.4922248243559719], "iscrowd": 0, "area": 8498.339350000004, "rle": {"size": [427, 640], "counts": "TZf41X=2N2O1QNKfF7Z9KbF7]9M^F5b93TFOk9i100O100O100O10O010F90010O010O0010O010OH9ROZOXOdFf0_9[O`Fb0b9A[F=g9EWF<i9EVF;j9FTF<j9FUF:k9GSF:l9HSF8m9JPF7o9KPF5P:LnE5Q:MnE3R:OkE2T:2hE0W:2gENX:6dEK\\:7aEJ^:a0VEAj:`100000O10000000000O01000000O10O010000O1000O01000000000O100002N1O1O2N1O1O2ZGWMf6j2WIXMj6h2SIZMn6g2nH[MR7f2dHbM^7U4001N101N1O2L4K4L6H8G9H8G9L3M4K5L3L\\FeM^9Y1aFjNc9S1XFROk9n0nEUOV:k0cEYO^:i12N2N2L4I7F9H9G9F:G\\UR2"}, "label": "rear end of a baby cow on the other side of the cow that the woman is milking"}]}
{"id": 204746, "image": "COCO_train2014_000000204746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow closest to log\"."}], "task": "refering", "answer_template": "The \"cow closest to log\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 105, 106, 107, 108, 109, 128, 131, 132, 151, 154, 155, 174, 177], "bbox": [0.5619375, 0.13491803278688524, 0.7508593749999999, 0.4922248243559719], "iscrowd": 0, "area": 8498.339350000004, "rle": {"size": [427, 640], "counts": "TZf41X=2N2O1QNKfF7Z9KbF7]9M^F5b93TFOk9i100O100O100O10O010F90010O010O0010O010OH9ROZOXOdFf0_9[O`Fb0b9A[F=g9EWF<i9EVF;j9FTF<j9FUF:k9GSF:l9HSF8m9JPF7o9KPF5P:LnE5Q:MnE3R:OkE2T:2hE0W:2gENX:6dEK\\:7aEJ^:a0VEAj:`100000O10000000000O01000000O10O010000O1000O01000000000O100002N1O1O2N1O1O2ZGWMf6j2WIXMj6h2SIZMn6g2nH[MR7f2dHbM^7U4001N101N1O2L4K4L6H8G9H8G9L3M4K5L3L\\FeM^9Y1aFjNc9S1XFROk9n0nEUOV:k0cEYO^:i12N2N2L4I7F9H9G9F:G\\UR2"}, "label": "cow closest to log"}]}
{"id": 204746, "image": "COCO_train2014_000000204746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow ' s ribs are exposed\"."}], "task": "refering", "answer_template": "The \"a cow ' s ribs are exposed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 50, 69, 70, 92, 93, 115, 116], "bbox": [0.001171875, 0.1304215456674473, 0.22939062500000001, 0.3888290398126464], "iscrowd": 0, "area": 3723.49335, "rle": {"size": [427, 640], "counts": "R`0Y2P;20000000O10000001N10001O001O00001O00001N10001O0O10001O00O0M4M300O010O10000O010SORE]On:<aE[O`:>nE[OR:b0VFXOj9f0[FXOe9e0`FYO`9d0e1N2N2N2O1N2O1N1O2O100O100O1O100O10O010000O100O100O1000O010bb>0^]A000000000O100000000O100000001O0O100000000O1000000000000000O1O1O1O1Oh`]6"}, "label": "a cow ' s ribs are exposed"}]}
{"id": 204746, "image": "COCO_train2014_000000204746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the body of a brown horse behind two cows\"."}], "task": "refering", "answer_template": "The \"the body of a brown horse behind two cows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 50, 69, 70, 92, 93, 115, 116], "bbox": [0.001171875, 0.1304215456674473, 0.22939062500000001, 0.3888290398126464], "iscrowd": 0, "area": 3723.49335, "rle": {"size": [427, 640], "counts": "R`0Y2P;20000000O10000001N10001O001O00001O00001N10001O0O10001O00O0M4M300O010O10000O010SORE]On:<aE[O`:>nE[OR:b0VFXOj9f0[FXOe9e0`FYO`9d0e1N2N2N2O1N2O1N1O2O100O100O1O100O10O010000O100O100O1000O010bb>0^]A000000000O100000000O100000001O0O100000000O1000000000000000O1O1O1O1Oh`]6"}, "label": "the body of a brown horse behind two cows"}]}
{"id": 237515, "image": "COCO_train2014_000000237515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair above tabby cat\"."}], "task": "refering", "answer_template": "The \"chair above tabby cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 148, 149, 150, 151, 152, 156, 157, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 178, 179, 184, 185, 187, 188, 189, 194, 195, 196, 200, 201, 202, 203, 204, 206, 207, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 222, 223, 224, 225, 226, 227, 228, 229, 231, 232, 233, 236, 237, 238, 239, 240, 244, 245, 246, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 266, 267, 268, 272, 273, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 288, 289, 290, 291, 294, 295, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 310, 311, 312, 313, 314, 315, 316, 317, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 332, 333, 334, 338, 339, 341, 342, 349, 350, 354, 355, 356, 360, 361, 363, 364, 371, 372, 376, 377, 378, 382, 383, 384, 385, 386, 393, 394, 398, 399, 400, 406, 407, 408, 415, 416, 420, 421, 422, 428, 429, 430, 437, 438, 439, 442, 443, 450, 451, 452, 460], "bbox": [0.05558966074313408, 0.004296875, 0.9627786752827139, 0.9010781250000001], "iscrowd": 0, "area": 161197.01085000008, "rle": {"size": [640, 619], "counts": "Sle07ic06J7I7I6J7I6J7I6J7I6J7I6J7I6J7I;Ea0_O`0@`0@`0@`0@a0_Ob0^Od0\\Od0\\Od0\\Od0\\Od0\\Od0\\Od0\\OP1PO]1cN\\1dNd0\\O9G:F:F9G9G00000000000000O100000000000000000000000000O100O100O100O100O100RDaMZ6`2fImMm5S2SJ[N^5f1cJgNo4Y1QKUO`4l0`KAS4?mKOd32]L=S3CmLLfJXNk7m1_MKTKZNm6k1PNJTKkNZ6\\1bNIUK^Od5j0WOITK3o44MITKj0W4^Of0GSKa1a3hN\\1GSKW2j2SNS2FTK`0]OVOf2d0i2GSK?ABS29Z3ERKa0CNa1Lj3ERKa0H8l0CZ4DRKa0Lc09XOi4DSK`0Oo0EnNZ5BRKa01W4m4YKQK`03V4l4ZKQK`03V4l4ZKQK`03V4m4YKQK`02W4m4YKQK`02W4m4ZKPK?3W4n4YKoJ`03W4n4YKoJ`03W4n4YKPK?2X4n4YKPK?2X4o4YKnJ`02W4P5YKnJ`02W4P5YKnJ`02W4P5YKoJ?1X4Q5XKnJ`02W4P5ZKmJ?3W4P5ZKmJ?3W4P5ZKnJ>2X4Q5YKmJ?2X4Q5YKmJ?2X4Q5YKmJ?2X4Q5ZKlJ>3X4R5YKlJ>2Y4R5YKlJ>2Y4R5YKlJ>2Y4R5YKlJ?1X4T5YKjJ?2X4T5YKkJ>1Y4T5YKkJ>2X4T5YKjJ?2X4T5YKjJ?2X4T5ZKiJ>3X4T5ZKjJ=2Y4U5YKiJ>2Y4U5YKiJ>2Y4U5YKiJ>2Y4U5ZKhJ=3Y4V5YKhJ=2Z4V5YKhJ=2Z4V5YKhJ>1Y4W5YKhJ>1Y4X5XKgJ?1Y4X5YKgJ=2Y4W5ZKgJ=2Y4W5ZKgJ=2Y4X5YKfJ>2Y4X5YKfJ>2Y4X5ZKfJ<2Z4Y5YKeJ=2Z4Y5YKeJ=2Z4Y5YKeJ=2Z4Y5YKeJ=2Z4Z5YKdJ<2[4Z5YKdJ<2[4Z5YKdJ=1Z4[5YKdJ=1Z4\\5XKdJ=1Z4[5YKdJ=1Z4[5ZKcJ<2Z4[5ZKcJ<2Z4\\5YKbJ=2Z4\\5YKcJ<1[4\\5YKcJ<1[4\\5ZKbJ;2[4]5YKaJ<2[4]5YKaJ<2[4]5YKbJ;1\\4]5YKbJ;1\\4]5ZKaJ:2\\4^5YK`J<1[4_5YK`J<2Z4^5ZKaJ;1[4^5ZKaJ;1[4^5[K`J:2[4_5ZK_J;2[4_5ZK_J;2[4_5ZK`J:1\\4_5ZK`J:1\\4_5ZK`J:1\\4`5ZK^J:2\\4`5ZK^J:2\\4`5ZK_J91]4`5ZK_J91]4`5ZK_J91^4`5YK^J:2\\4`5ZK^J:2\\4`5ZK_J91]4`5ZK_J91]4`5ZK_J91]4a5ZK]J92]4a5ZK]J92]4a5ZK^J81^4a5ZK^J81^4a5ZK^J81^4b5ZK\\JT8d5lG\\JT3_OYOU6cM]JX295Z5cM]J\\1U1Q1_4bM\\JIi2e2k2UM`0j2@cKS2]4h6000000000000000001O000000000000000000001O000000000000O100O100O100O100O100O1_IXKjMh4i1gMTLZ2_3CQK=a4Y1VJhN]5n2ZIRMY6a8L4000000000000O1000000000000000000000000O1000ZMYJSGg5e8aJ[G_5^8hJbGX5V8PKjGP5o7WKQHi4h7^KXHb4`7fK`HZ4Y7mKgHS4Q7VLnHj3k6]LUIc3c6eL]I[3\\6lLdIS3U6UMkIk2h5bMXJ^2[5oMeJQ2n4\\NRKd1`4jN`KV1S4WOmKi0f3DZL<Y31gLOl2>TMB_2k0aMUOR2X1nMhNd1f1\\NZNX1S2gNmMl0`2TO_M`0n2@RM3[3MeLFh3:XLYOU4g0kKlNb4T1^K_No4a1QKRN\\5n1[600000000000001O00000000000000000000G9E;D<E;D<D<E;D<D<E;D<E;D<D<E;D<D<E;F:0[HoD[6f7mJSI]NBf6U7SKYIRNGk6i6ZK`IiMIm6a6`KfIcMIm6Z6gKmI\\MIm6T6[IZIb1i0VNIm6n5aIfIV1c0\\NIm6g5iIQJi0?aNIm6a5oI]J=9hNHl6\\5UJeJ57jNHl6U5\\JlJN6kNIk6o4bJRKH6kNIk6i4hJXKB6kNJj6a4PK_K[O6kNJj6[4VKeKUO6kNJj6U4\\KkKoN6kNJj6d1hHNe2^NnN6kNJj6[1QI8[2]NoN6kNJj6S1YI`0T2\\NnN7kNJj6j0bIi0k1\\NnN7kNJj6b0jIQ1c1]NmN6lNJj69SJZ1Z1]NmN6lNJj61\\Ja1Q1^NmN6lNJj6HeJk1g0]NnN6lNJj6@mJS2?]NnN6lNJ]9c1iH]NnN6lNJ]9c1iH]NnN6lNJ^9b1hH^NnN6lNJ^9b1hH^NnN6lNJ^9c1hH\\NnN7mNI]9d1hH\\NnN7mNI]9d1hH\\NnN7mNI^9c1gH]NnN7mNI^9c1gH^NmN6nNI^9c1gH^NmN6nNI^9d1fH]NnN6nNI^9d1fH]NnN6nNI^9d1fH]NnN6nNI_9c1eH^NnN6nNI_9c1eH^NnN6nNI_9c1fH]NmN7nNI_9d1eH\\NnN7nNI_9d1eH\\NnN7nNI`9c1dH]NnN7nNI`9c1dH]NnN7nNI`9c1dH^NmN6oNI`9c1dH^NmN6oNI`9d1cH]NnN6oNIa9c1bH^NnN6oNIa9c1bH^NnN6POH`9d1bH^NnN6POH`9d1cH]NmN7POH`9d1cH]NmN7POH`9e1bH\\NnN7POI`9c1bH]NnN7POI`9c1bH]NnN7POI`9c1bH]NnN7POI`9c1bH]NnN7POI`9c1bH^NmN6QOIa9c1`H^NnN6QOIa9c1`H^NnN6QOIa9c1`H^NnN6QOIa9c1aH]NmN7QOIa9c1aH]NmN7QOIb9b1`H^NmN7QOIb9c1_H]NnN7QOIb9c1_H]NnN7QOIb9c1_H]NnN7QOIb9c1_H]NnN7QOIb9c1_H]NnN7QOIc9b1^H_NmN6SOHb9d1]H^NnN6SOHb9d1^H]NmN7SOHb9d1^H]NmN7SOHb9d1^H]NmN7SOHc9c1]H^NmN7SOHc9c1]H^NmN7SOHc9d1\\H]NnN7SOHc9d1\\H]NnN7SOHc9d1\\H]NnN7SOHc9d1\\H]NnN7SOHd9c1[H^NnN7SOHd9c1[H^NnN7SOHd9d1[H]NmN7TOHd9d1[H]NmN7TOHd9d1[H]NmN7TOHe9c1ZH^NmN7TOHe9c1ZH^NmN7TOHe9c1ZH^NmN7TOHe9d1YH]NnN7TOHe9d1YH]NnN7UOGe9d1XH^NnN7UOGe9d1XH^NnN7UOGe9d1XH^NnN7UOGe9d1YH]NmN8UOGe9e1XH]NmN7VOHd9d1YH]NmN7VOHe9c1XH^NmN7VOHe9c1XH^NmN7VOHe9c1XH^NmN7VOHe9c1XH^NmN7VOHe9d1WH]NnN7VOHf9c1VH^NnN7VOHf9c1VH^NnN7VOHf9c1VH^NnN7VOHf9c1WH]NmN8VOHf9c1WH]NmN8VOHg9c1UH]NnN8VOHg9c1UH^NmN7WOHg9c1UH^NmN7WOHg9c1UH^NmN7WOHg9c1UH^NmN7XOGf9d1UH^NmN7XOGg9d1SH^NnN7XOGg9d1SH^NnN7XOGg9d1SH^NnN7XOGg9d1TH]NmN8XOGg9d1TH]NmN8XOGh9c1SH^NmN8XOGh9d1RH]NnN8XOGh9d1RH^NmN7YOGh9d1RH^NmN7YOGh9d1RH^NmN7YOGi9c1QH_NmN7YOGi9c1QH_NmN7YOGi9d1PH^NnN7YOGi9d1PH^NnN7YOGi9d1QH]NmN8YOGi9d1QH]NmN8YOGj9c1PH^NmN8YOGj9d1oG]NnN7ZOHi9d1oG]NnN7[OGh9e1oG]NnN7[OGh9e1oG^NmN6\\OGi9d1nG_NmN6\\OGi9e1mG^NnN6\\OGi9e1mG^NnN6\\OGi9e1mG^NnN6\\OHh9d1oG]NmN7\\OHh9d1oG]NmN7\\OHi9d1mG]NnN7\\OHi9d1mG]NnN7\\OHi9d1mG]NnN7\\OHi9d1mG]NnN7\\OHi9d1mG]NnN7\\OHj9d1kG]NoN7\\OHj9d1kG^NnN6]OHj9d1kG^NnN6]OHj9d1lG]NmN7]OHj9d1lG]NmN7]OHk9c1kG^NmN7]OHk9d1jG]NnN7]OHk9d1jG]NnN7^OGj9e1jG]NnN7^OGj9e1jG]NnN7^OGj9e1jG]NnN7^OGk9e1hG]NoN7^OGk9e1hG]NoN7^OGk9e1hG^NnN6_OGk9e1iG]NmN7_OGk9e1iG]NmN7_OGl9e1gG]NnN7_OGl9e1gG]NnN7_OGl9e1gG]NUO0XONl9e1gGTOgNPOF7l9e1_Em1h0mKMa0m9b6dF^I\\9T7RFlHn9e7aE[H_:c804La0@a0^Oi0WOi0WOj0VOj0WO2M00000010O0000010O001O00100O001O010O001O1O010O001O010O001O1O0UGaIV4_6dJ^Kd4b5VI\\LR6`9YOh0YOg0XO>B:F:F:G9F:F:F:F:G9F:F;E:F:G9Dhj="}, "label": "chair above tabby cat"}]}
{"id": 237515, "image": "COCO_train2014_000000237515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange chair on left\"."}], "task": "refering", "answer_template": "The \"orange chair on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 88, 89, 110, 111, 132, 133, 135, 136, 137, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 198, 199, 200, 202, 203, 220, 221, 222, 224, 225, 242, 243, 244, 246, 247, 248, 249, 250, 264, 265, 266, 268, 269, 270, 271, 286, 287, 288, 290, 291, 308, 309, 310, 312, 313, 334, 335, 356, 357, 378, 379, 400], "bbox": [0.0, 0.16284375, 0.3705977382875606, 0.7888125], "iscrowd": 0, "area": 31816.4208, "rle": {"size": [640, 619], "counts": "X3e9\\:O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001N10001O0000001O00001O0000001O007If0ZOf0ZOf0ZOf0ZOg0YOf0ZOf0ZOf0ZOf0ZOf0ZOg0YOf0ZOSih0@]WWOa0@`0_O`0@a0@?@a0@?@a0@?@a0_O`0A`0_Oa0@?@a0@?@a0@?@a0_O?BO001N101O00001O001O0O2O001O00001XJQF]1o9YN[Fg1f9nMdFQ2]9eMmF[2T9ZMVGf2j8PM`GP3`8fLjGZ3W8[LSHeNoNU3n8lM]HoN4f1`7POfHZOY17Q65PIDQ<0ZD0f;FdD:a?000O100000O1000000000000000000000000000O1000000000000000O1000000000000000O1000000000000000000000000000O1000O10000000000000000000000000000000000OIX^c7"}, "label": "orange chair on left"}]}
{"id": 237515, "image": "COCO_train2014_000000237515.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front right leg of a partially obsucred chair , the furthest to the left\"."}], "task": "refering", "answer_template": "The \"the front right leg of a partially obsucred chair , the furthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 88, 89, 110, 111, 132, 133, 135, 136, 137, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 198, 199, 200, 202, 203, 220, 221, 222, 224, 225, 242, 243, 244, 246, 247, 248, 249, 250, 264, 265, 266, 268, 269, 270, 271, 286, 287, 288, 290, 291, 308, 309, 310, 312, 313, 334, 335, 356, 357, 378, 379, 400], "bbox": [0.0, 0.16284375, 0.3705977382875606, 0.7888125], "iscrowd": 0, "area": 31816.4208, "rle": {"size": [640, 619], "counts": "X3e9\\:O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001N10001O0000001O00001O0000001O007If0ZOf0ZOf0ZOf0ZOg0YOf0ZOf0ZOf0ZOf0ZOf0ZOg0YOf0ZOSih0@]WWOa0@`0_O`0@a0@?@a0@?@a0@?@a0_O`0A`0_Oa0@?@a0@?@a0@?@a0_O?BO001N101O00001O001O0O2O001O00001XJQF]1o9YN[Fg1f9nMdFQ2]9eMmF[2T9ZMVGf2j8PM`GP3`8fLjGZ3W8[LSHeNoNU3n8lM]HoN4f1`7POfHZOY17Q65PIDQ<0ZD0f;FdD:a?000O100000O1000000000000000000000000000O1000000000000000O1000000000000000O1000000000000000000000000000O1000O10000000000000000000000000000000000OIX^c7"}, "label": "the front right leg of a partially obsucred chair , the furthest to the left"}]}
{"id": 253904, "image": "COCO_train2014_000000253904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower vase with a red ribbon\"."}], "task": "refering", "answer_template": "The \"a flower vase with a red ribbon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 47, 48, 49, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 118, 119, 120, 121, 137, 138, 139, 155, 156], "bbox": [0.56576, 0.09315315315315315, 0.8152999999999999, 0.7386786786786788], "iscrowd": 0, "area": 15545.021700000001, "rle": {"size": [333, 500], "counts": "PVl2a0k93L5K4M3L4bNmN_IU1Z6QOdIQ1V6UOhIm0S6XOlIi0Q6ZOmId0S6@kI7\\6McI2\\61aI1\\62bI0\\63aIO]63`I0]64`IN^63aIO]62bI0\\61cI2Z6OeI3Z65\\IMc68XIJf6o1\\J]L^4j4N101O0O010QOUKZLm4Y3oKVLT4h3RLQLP4o3X1O001N101O001N101O00010O0010O010O000O2O001O0O101O001N10001O0O2O00001N2O2N1O1N2O1O1O2M2O1O1C=B>A?nNYJoLV6o2g0L5L2N1N2O1O1N2O1O1N2O1N2O1N2N2O1N2O2M2O1N3M4M3L4M3K6K7H3N2M3M3N2M3N2M2N3N2M2O2J5Jiim0"}, "label": "a flower vase with a red ribbon"}]}
{"id": 253904, "image": "COCO_train2014_000000253904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase of flowers at the very right of the group of flower vases\"."}], "task": "refering", "answer_template": "The \"the vase of flowers at the very right of the group of flower vases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 47, 48, 49, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 118, 119, 120, 121, 137, 138, 139, 155, 156], "bbox": [0.56576, 0.09315315315315315, 0.8152999999999999, 0.7386786786786788], "iscrowd": 0, "area": 15545.021700000001, "rle": {"size": [333, 500], "counts": "PVl2a0k93L5K4M3L4bNmN_IU1Z6QOdIQ1V6UOhIm0S6XOlIi0Q6ZOmId0S6@kI7\\6McI2\\61aI1\\62bI0\\63aIO]63`I0]64`IN^63aIO]62bI0\\61cI2Z6OeI3Z65\\IMc68XIJf6o1\\J]L^4j4N101O0O010QOUKZLm4Y3oKVLT4h3RLQLP4o3X1O001N101O001N101O00010O0010O010O000O2O001O0O101O001N10001O0O2O00001N2O2N1O1N2O1O1O2M2O1O1C=B>A?nNYJoLV6o2g0L5L2N1N2O1O1N2O1O1N2O1N2O1N2N2O1N2O2M2O1N3M4M3L4M3K6K7H3N2M3M3N2M3N2M2N3N2M2O2J5Jiim0"}, "label": "the vase of flowers at the very right of the group of flower vases"}]}
{"id": 253904, "image": "COCO_train2014_000000253904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vase with green ribbon around it\"."}], "task": "refering", "answer_template": "The \"vase with green ribbon around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 54, 55, 72, 73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 162, 163, 180, 181, 198, 199], "bbox": [0.00298, 0.2210810810810811, 0.16426, 0.9856456456456457], "iscrowd": 0, "area": 11672.7034, "rle": {"size": [333, 500], "counts": "j<j7c20O1O1O101O000001O00001O00001O00001O00001O001O001jH[Mb6e2]I\\Mc6e2ZI^Me6b2ZI_Mf6c2WI_Mh6c2TI_Ml6b2RI`Me3MVOS3i0kMZNV2d1hN_MX1`2EdL<Z3\\301N1O2N1O2N1O2N100O2N1O2N1O2N=C2N100O1O1O1O1O1O1O2N100O2N3M2N3N1N2N3M2N5L3L4K6I6K5K6I6K6JdhW4"}, "label": "vase with green ribbon around it"}]}
{"id": 253904, "image": "COCO_train2014_000000253904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bouquet of carnations with purple flowers in the center above yellow ones\"."}], "task": "refering", "answer_template": "The \"bouquet of carnations with purple flowers in the center above yellow ones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 54, 55, 72, 73, 74, 90, 91, 92, 108, 109, 110, 126, 127, 128, 144, 145, 162, 163, 180, 181, 198, 199], "bbox": [0.00298, 0.2210810810810811, 0.16426, 0.9856456456456457], "iscrowd": 0, "area": 11672.7034, "rle": {"size": [333, 500], "counts": "j<j7c20O1O1O101O000001O00001O00001O00001O00001O001O001jH[Mb6e2]I\\Mc6e2ZI^Me6b2ZI_Mf6c2WI_Mh6c2TI_Ml6b2RI`Me3MVOS3i0kMZNV2d1hN_MX1`2EdL<Z3\\301N1O2N1O2N1O2N100O2N1O2N1O2N=C2N100O1O1O1O1O1O1O2N100O2N3M2N3N1N2N3M2N5L3L4K6I6K5K6I6K6JdhW4"}, "label": "bouquet of carnations with purple flowers in the center above yellow ones"}]}
{"id": 253904, "image": "COCO_train2014_000000253904.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase of flowers tied with a purple ribbon between vases tied with green and brown ribbons\"."}], "task": "refering", "answer_template": "The \"a vase of flowers tied with a purple ribbon between vases tied with green and brown ribbons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 75, 76, 77, 78, 79, 93, 94, 95, 96, 111, 112, 113, 129, 130, 131, 147, 148, 149, 165, 166, 167, 183, 184, 185], "bbox": [0.17212, 0.2763963963963964, 0.39362, 0.8763963963963964], "iscrowd": 0, "area": 12840.1229, "rle": {"size": [333, 500], "counts": "UTl03V:7J5J6L4L5J5L3L4N1O2N101N1O2O0O2NP1nH`MX4\\4N2O1N2N2@\\KRKh4m4=O001N2O1O1O001O000O10000000000O1000001O000O10O1000N2H7J7K5000O2O0000001N1000001N10001O1N3N1O1O2M2O1aL[KT1f4jN]KT1e4iN_KT1b4jNaKm0g4QO]K8Y5EjJFj58ZJQO[6m0hI]Nl6b1P1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1N2O1O1O1O1M3^OT^R3"}, "label": "a vase of flowers tied with a purple ribbon between vases tied with green and brown ribbons"}]}
{"id": 204759, "image": "COCO_train2014_000000204759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a white hat and red shirt next to a large animal\"."}], "task": "refering", "answer_template": "The \"a person in a white hat and red shirt next to a large animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.77453125, 0.4297916666666667, 1.0, 1.0], "iscrowd": 0, "area": 21169.95865, "rle": {"size": [480, 640], "counts": "onX71o>0O1O100O1O1O100O1O1O100O1O1O100O1H8O1N2N2N2O1O100O100O100O1O100O100O100O100O1O100O100O100O100K5K5J6J6K5L4K5L4VK`MQMc2n2_MnLd2Q3_MjLd2V3^MdLf2[3]M`Lf2`3[M\\Lh2c3ZMZLh2f3XMXLj2g3WMXLj2g3WMWLk2i3UMULm2j3TMTLn2l3RMSLo2l3RMRLP3n3PMPLR3o3oLoKS3P4nLnKT3R4lLmKU3R4lLlKV3T4jLjKX3U4iLiKY3W4gLfK\\3Y4eLeK]3Z4eLcK]3\\4dLbK^3]4dL`K^3_4dLVIMR1b3g5bLVI1m0_3k5cLVI2j0]3o5cLTI2k0]3P6bLTI2j0^3Q6kMmIW2R6jMmIW2R6jMnIV2Q6kMnIV2Q6kMoIU2P6lMoIU2P6lMPJT2o5mMPJT2n5nMRJR2n5nMQJS2o5mMQJS2o5mMPJT2o5mMQJS2o5mMPJT2P6oMlIR2T6QNiIo1W6TNeIm1Z6WNcIi1]6YN`Ih1`6\\30000O10000000000O100000000O100000000O11O00001O00001O00001O00001O00001O0000"}, "label": "a person in a white hat and red shirt next to a large animal"}]}
{"id": 204759, "image": "COCO_train2014_000000204759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white cap and brown shirt standing next to an elephant\"."}], "task": "refering", "answer_template": "The \"a man with a white cap and brown shirt standing next to an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.77453125, 0.4297916666666667, 1.0, 1.0], "iscrowd": 0, "area": 21169.95865, "rle": {"size": [480, 640], "counts": "onX71o>0O1O100O1O1O100O1O1O100O1O1O100O1H8O1N2N2N2O1O100O100O100O1O100O100O100O100O1O100O100O100O100K5K5J6J6K5L4K5L4VK`MQMc2n2_MnLd2Q3_MjLd2V3^MdLf2[3]M`Lf2`3[M\\Lh2c3ZMZLh2f3XMXLj2g3WMXLj2g3WMWLk2i3UMULm2j3TMTLn2l3RMSLo2l3RMRLP3n3PMPLR3o3oLoKS3P4nLnKT3R4lLmKU3R4lLlKV3T4jLjKX3U4iLiKY3W4gLfK\\3Y4eLeK]3Z4eLcK]3\\4dLbK^3]4dL`K^3_4dLVIMR1b3g5bLVI1m0_3k5cLVI2j0]3o5cLTI2k0]3P6bLTI2j0^3Q6kMmIW2R6jMmIW2R6jMnIV2Q6kMnIV2Q6kMoIU2P6lMoIU2P6lMPJT2o5mMPJT2n5nMRJR2n5nMQJS2o5mMQJS2o5mMPJT2o5mMQJS2o5mMPJT2P6oMlIR2T6QNiIo1W6TNeIm1Z6WNcIi1]6YN`Ih1`6\\30000O10000000000O100000000O100000000O11O00001O00001O00001O00001O00001O0000"}, "label": "a man with a white cap and brown shirt standing next to an elephant"}]}
{"id": 204759, "image": "COCO_train2014_000000204759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue striped shirt\"."}], "task": "refering", "answer_template": "The \"a man in a blue striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.10304687500000001, 0.25, 0.48817187500000003, 0.9819791666666667], "iscrowd": 0, "area": 52890.82170000001, "rle": {"size": [480, 640], "counts": "R[o0o0m<\\1UOl0SOi0XOc0_Oa0M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1QNjGSMW8m2QHkLo7T3[HbLf7]3dHZL\\7f3lHPLV7o3n1O100O100O1O100O1O100O1O100O100O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100001O001O001O001O1O001XHjKd4V4\\KkKc4V4[KlKd4T4\\KmKb1M]NW4OnK_13^No33oKZ19`Ni34PLW1>bNc36PLS1e0cN[39RLo0j0eNU3;RLk0Q1fNm2>TLg0V1hNg2`0TLc0\\1jN`2b0VL?b1kNX2f0WL<f1kNT2d0[La0d1gNQ2a0cLh0^1dNo1<mLn0W1cNm17UMU1Q1`Nj14^MZ1k0_Nh1JkMf1`0\\Ne1DWNn17[Nb1GYNm17YNa1H[Nn17VN^1L]Nl18UN\\1M_Nm1a5SNaJk1_5UNcJj1]5TNfJj1Z5VNhJi1X5UNkJi1U5WNmJg1S5XNPKf1Q5YNQKe1o4[NVK`1k4^N[K\\1T5VNQKe1Z9K5K5K2N000000000O10000000000000O1N2N2N1O2N2N2N2N2N2O1N2O00O101N10001N10001N100O2L3K5L5K4K5L5K4K5L5Kh_i4"}, "label": "a man in a blue striped shirt"}]}
{"id": 204759, "image": "COCO_train2014_000000204759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue standing next to an elephant\"."}], "task": "refering", "answer_template": "The \"a man in blue standing next to an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 143, 144, 145, 146, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 278, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 353, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.10304687500000001, 0.25, 0.48817187500000003, 0.9819791666666667], "iscrowd": 0, "area": 52890.82170000001, "rle": {"size": [480, 640], "counts": "R[o0o0m<\\1UOl0SOi0XOc0_Oa0M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2N2M3N2O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1QNjGSMW8m2QHkLo7T3[HbLf7]3dHZL\\7f3lHPLV7o3n1O100O100O1O100O1O100O1O100O100O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100001O001O001O001O1O001XHjKd4V4\\KkKc4V4[KlKd4T4\\KmKb1M]NW4OnK_13^No33oKZ19`Ni34PLW1>bNc36PLS1e0cN[39RLo0j0eNU3;RLk0Q1fNm2>TLg0V1hNg2`0TLc0\\1jN`2b0VL?b1kNX2f0WL<f1kNT2d0[La0d1gNQ2a0cLh0^1dNo1<mLn0W1cNm17UMU1Q1`Nj14^MZ1k0_Nh1JkMf1`0\\Ne1DWNn17[Nb1GYNm17YNa1H[Nn17VN^1L]Nl18UN\\1M_Nm1a5SNaJk1_5UNcJj1]5TNfJj1Z5VNhJi1X5UNkJi1U5WNmJg1S5XNPKf1Q5YNQKe1o4[NVK`1k4^N[K\\1T5VNQKe1Z9K5K5K2N000000000O10000000000000O1N2N2N1O2N2N2N2N2N2O1N2O00O101N10001N10001N100O2L3K5L5K4K5L5K4K5L5Kh_i4"}, "label": "a man in blue standing next to an elephant"}]}
{"id": 425945, "image": "COCO_train2014_000000425945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the whole brown suitcase underneath the suitcase with stickers on the side\"."}], "task": "refering", "answer_template": "The \"the whole brown suitcase underneath the suitcase with stickers on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 180, 181, 182, 183, 184, 185, 186, 187, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 289, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 320, 321, 322, 323, 336, 337], "bbox": [0.0033723653395784543, 0.35954687500000004, 1.0, 0.9707812499999999], "iscrowd": 0, "area": 84233.8973, "rle": {"size": [640, 427], "counts": "ik0U5j>2O1O1N2O2N1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O2M2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O2M2O1O1N2O1N2O1O1O1N2O1O1O1O1O1O1O2M2O1O10000O100000000O1000000O10001O0O1000000O1000000O100000000O10001O0O1000000O1000000O1000000O1000001N100000000O1000000O1000000O10001O0O1000000O100000000O1000000O10001O0O1000000O1000000O100000000O10001O0O1000000O1000000O1000000O1000001N100000000O1000000O1000000O1000001N1000000O01N2O1N2N101N2N2O0O2N2N2O0O2N2O1N1O2O1N2N101N2N2O1N1O2O1N2N101N2N2N101N2N2O0O2N2O1N1O2O1N2N2O0O2N2O1N1O2O1N2N1O2O1N2N101N2N2O0O2N2O1N1O2O1N2N2O0O2N2O1N1O2N2O1N1O2O1N2N101N2N2O0O2N2O1N2N101N2N2O0O2N2N2O0O2N2O1N1O2O1N2N101N2N2O1N1O2O1N2N101N2N2O0O001O010O0010O00010O01O010O0010O00010O01O010O001O01O01O010O0010O00010O01O010O0010O00010O01O0010O01O01O01O010O0010O00010O01O010O0010O0001O010O0010O01O01O01O010O0010O00010O01O010O001O01O01O010O0010O01O01O01O010O0010O0bD"}, "label": "the whole brown suitcase underneath the suitcase with stickers on the side"}]}
{"id": 425945, "image": "COCO_train2014_000000425945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second from the bottom luggage\"."}], "task": "refering", "answer_template": "The \"the second from the bottom luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 180, 181, 182, 183, 184, 185, 186, 187, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 289, 290, 291, 292, 293, 294, 295, 296, 305, 306, 307, 308, 309, 320, 321, 322, 323, 336, 337], "bbox": [0.0033723653395784543, 0.35954687500000004, 1.0, 0.9707812499999999], "iscrowd": 0, "area": 84233.8973, "rle": {"size": [640, 427], "counts": "ik0U5j>2O1O1N2O2N1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O2M2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O2M2O1O1N2O1N2O1O1O1N2O1O1O1O1O1O1O2M2O1O10000O100000000O1000000O10001O0O1000000O1000000O100000000O10001O0O1000000O1000000O1000000O1000001N100000000O1000000O1000000O10001O0O1000000O100000000O1000000O10001O0O1000000O1000000O100000000O10001O0O1000000O1000000O1000000O1000001N100000000O1000000O1000000O1000001N1000000O01N2O1N2N101N2N2O0O2N2N2O0O2N2O1N1O2O1N2N101N2N2O1N1O2O1N2N101N2N2N101N2N2O0O2N2O1N1O2O1N2N2O0O2N2O1N1O2O1N2N1O2O1N2N101N2N2O0O2N2O1N1O2O1N2N2O0O2N2O1N1O2N2O1N1O2O1N2N101N2N2O0O2N2O1N2N101N2N2O0O2N2N2O0O2N2O1N1O2O1N2N101N2N2O1N1O2O1N2N101N2N2O0O001O010O0010O00010O01O010O0010O00010O01O010O001O01O01O010O0010O00010O01O010O0010O00010O01O0010O01O01O01O010O0010O00010O01O010O0010O0001O010O0010O01O01O01O010O0010O00010O01O010O001O01O01O010O0010O01O01O01O010O0010O0bD"}, "label": "the second from the bottom luggage"}]}
{"id": 425945, "image": "COCO_train2014_000000425945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a box with a sticker on the side\"."}], "task": "refering", "answer_template": "The \"a box with a sticker on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 18, 19, 20, 21, 22, 31, 32, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 51, 52, 53, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 219, 220, 221, 222, 223, 235, 236], "bbox": [0.1246135831381733, 0.038203125, 1.0, 0.67865625], "iscrowd": 0, "area": 94632.86070000002, "rle": {"size": [640, 427], "counts": "RVQ16jc0b0^Oc0\\Oc0^Oc0]Ob0]Od0]Oc0]Ob0^O6I2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O001O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O001O1N2O1O001N2O1O1O0O2O1O1N2O001O1N2O1O001N2O1O1O0O2O1O1O1O001O1O1O1O001O1O1O1O001;E;E1O2M2O1O2N1O1N3N2N3M2N3L4M5K4L3L2O2N1O1O1N2OO100O10O010000O10O0100O1000O0100O100O01000O100O010O10000O010O100O10O10O100O10O01O100O1O001O1O1O10O01O1O1O1O00100O1O001O1O100O001O1O1O1O010O1O1O1O001O01N1O100O1O100O2N100O1O100O1O2N100O1O100O2N100O1O100O1O2O0O1O100O1O101N1O1O100O1O101N1O100O1O101N1O100O1O100O2N1O100O1O100O2N100O1O100O1O2O0O1O100O1O2O0O1O1O100O1O2O0O1O100O1O101N1O100O1O100O2N1O100O1O10l0SOZJ"}, "label": "a box with a sticker on the side"}]}
{"id": 425945, "image": "COCO_train2014_000000425945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown suitcase with stickers on it\"."}], "task": "refering", "answer_template": "The \"a brown suitcase with stickers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 18, 19, 20, 21, 22, 31, 32, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 51, 52, 53, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 219, 220, 221, 222, 223, 235, 236], "bbox": [0.1246135831381733, 0.038203125, 1.0, 0.67865625], "iscrowd": 0, "area": 94632.86070000002, "rle": {"size": [640, 427], "counts": "RVQ16jc0b0^Oc0\\Oc0^Oc0]Ob0]Od0]Oc0]Ob0^O6I2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O001O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O1O001N2O1O1O0O2O1O1O1N101O1N2O001O1N2O1O001N2O1O1O0O2O1O1N2O001O1N2O1O001N2O1O1O0O2O1O1O1O001O1O1O1O001O1O1O1O001;E;E1O2M2O1O2N1O1N3N2N3M2N3L4M5K4L3L2O2N1O1O1N2OO100O10O010000O10O0100O1000O0100O100O01000O100O010O10000O010O100O10O10O100O10O01O100O1O001O1O1O10O01O1O1O1O00100O1O001O1O100O001O1O1O1O010O1O1O1O001O01N1O100O1O100O2N100O1O100O1O2N100O1O100O2N100O1O100O1O2O0O1O100O1O101N1O1O100O1O101N1O100O1O101N1O100O1O100O2N1O100O1O100O2N100O1O100O1O2O0O1O100O1O2O0O1O1O100O1O2O0O1O100O1O101N1O100O1O100O2N1O100O1O10l0SOZJ"}, "label": "a brown suitcase with stickers on it"}]}
{"id": 278491, "image": "COCO_train2014_000000278491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"park bench\"."}], "task": "refering", "answer_template": "The \"park bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 336, 337], "bbox": [0.556453125, 0.7285421412300683, 1.0, 0.9142369020501138], "iscrowd": 0, "area": 5147.656049999998, "rle": {"size": [439, 640], "counts": "`nh4:]=0000001O000000000O101O0000000000001O0000000O101O0000000000001O000000000O101O0000000000001O00001O;E;D<E:F000000000000000000000O100000O1N2J6J6I7K5O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1O1O1000O1000O10000000000000000O1000O10000000000000O1000000000O100000O10000000000000000O10O1000000000000000O1000O100000000000O100000O100000000000O100000O1000000000O1000000000O12N3M3M2M4Maf4NaYK4L3M3M3M1O000O10O100000000000000000000000000O100000000000000000000M2N3M3LTU80ojG1N2O1N2O1N2N2O1O1000000000000O1000000000000O1000000000000O100000000PF"}, "label": "park bench"}]}
{"id": 278491, "image": "COCO_train2014_000000278491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden picnic bench located at the edge of a lake\"."}], "task": "refering", "answer_template": "The \"a wooden picnic bench located at the edge of a lake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 273, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 336, 337], "bbox": [0.556453125, 0.7285421412300683, 1.0, 0.9142369020501138], "iscrowd": 0, "area": 5147.656049999998, "rle": {"size": [439, 640], "counts": "`nh4:]=0000001O000000000O101O0000000000001O0000000O101O0000000000001O000000000O101O0000000000001O00001O;E;D<E:F000000000000000000000O100000O1N2J6J6I7K5O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1O1O1000O1000O10000000000000000O1000O10000000000000O1000000000O100000O10000000000000000O10O1000000000000000O1000O100000000000O100000O100000000000O100000O1000000000O1000000000O12N3M3M2M4Maf4NaYK4L3M3M3M1O000O10O100000000000000000000000000O100000000000000000000M2N3M3LTU80ojG1N2O1N2O1N2N2O1O1000000000000O1000000000000O1000000000000O100000000PF"}, "label": "a wooden picnic bench located at the edge of a lake"}]}
{"id": 73694, "image": "COCO_train2014_000000073694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder in a jersey and camoflague pants\"."}], "task": "refering", "answer_template": "The \"a skateboarder in a jersey and camoflague pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 248, 249, 250, 251, 264, 265, 266, 279, 280, 281, 294, 295, 296, 309, 310, 323, 324, 325, 339, 340], "bbox": [0.3355220417633411, 0.324796875, 0.9002552204176334, 0.989234375], "iscrowd": 0, "area": 42158.880250000024, "rle": {"size": [640, 431], "counts": "hlj21nc02N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N0000O10i@ZOU;e0mDDi:<XEN]:1dE4X:KiE5W:IkE6U:LiE5W:LhE3Y:NfE2Y:0fEO[:2cEO]:2bEM^:5aEK_:6`EIa:8]EIb:9]EFd:;[EDe:OfAFc3<g:MiAF_3<i:NjAE\\3=i:NnADX3=k:NQBDQ3?n:MSBCn2`0n:MWBBj2a0o:MYBAg2b0P;L\\BAb2d0Q;KaB_O]2g0Q;JdB^OZ2h0R;IgB^OV2i0R;JjB\\OR2k0T;IoE7Q:IoE8k9MTF4h90XF0d94cEoN[Mm0m<9gESOTMd0Q==jEXOlL<V=`0mE<R:DmE=S:DTDeN2h1k;BRDlNMd1P<@RDROH_1V<_OPDZOCY1\\<]OPD@^OT1c<[OnCGYOP1h<YOnCNSOj0o<XOlC5oNe0U=UOkCe2U<[MjCf2V<ZMiCh2V<XMhCj2Y<UMfCm2Y<SMfCn2Z<RMeCP3[<oLcCS3]<mLbCT3^<mL`CU3_<lL^CV3c<kLZCW3e<jLYCW3g<kLVCW3j<iLSCZ3l<hLQC[3m<fLQC\\3n<fLoB]3o<dLnB_3Q=j1O0O1O2N1OO1O100O2N1O100O1O1O1O101N1O1O1O\\G]Im3b6RLbIl3]6RLgIm3X6RLkIm3T6QLPJo3n5PLVJn3i5PL[Jo3d5PL_Jo3`5oKdJP4[5oKhJP4W5nKnJP4Q5mKTKR4l4^KeKb4Y4oJXLP5g3`JlL^5S3SJ^Ml5a2dIQN[6n1VIcNi6\\1gHWOW7h0ZHIe7Z600O2N1O1O001O0001O01O00001O0000010O0000001O000010O000001O00N2K5J7I6J6K5J6J5K5L4K5K5K6K4K5K5K5L4K6J5ZKXE1l:OWEIn:6VEBo:>TEZOQ;f0REQOT;o0PEVNh;i1[D]M\\<b2hCeLn<[3UCkKb=U4T10001O0O100000001N100000001O2M4M2N2_NP@oNR`0o0V@iNl?V1[@aNi?\\1_@\\Nc?b1f@UN\\?i1l@oMV?o1_1N3M2O1N2N2N3M2N2N2O1N3M2N2N2N2N3N1N2N2N2N3M2N2O1N2N3M2N2N2N2O2Mncj0"}, "label": "a skateboarder in a jersey and camoflague pants"}]}
{"id": 73694, "image": "COCO_train2014_000000073694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skater in a white shirt\"."}], "task": "refering", "answer_template": "The \"the skater in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 248, 249, 250, 251, 264, 265, 266, 279, 280, 281, 294, 295, 296, 309, 310, 323, 324, 325, 339, 340], "bbox": [0.3355220417633411, 0.324796875, 0.9002552204176334, 0.989234375], "iscrowd": 0, "area": 42158.880250000024, "rle": {"size": [640, 431], "counts": "hlj21nc02N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N0000O10i@ZOU;e0mDDi:<XEN]:1dE4X:KiE5W:IkE6U:LiE5W:LhE3Y:NfE2Y:0fEO[:2cEO]:2bEM^:5aEK_:6`EIa:8]EIb:9]EFd:;[EDe:OfAFc3<g:MiAF_3<i:NjAE\\3=i:NnADX3=k:NQBDQ3?n:MSBCn2`0n:MWBBj2a0o:MYBAg2b0P;L\\BAb2d0Q;KaB_O]2g0Q;JdB^OZ2h0R;IgB^OV2i0R;JjB\\OR2k0T;IoE7Q:IoE8k9MTF4h90XF0d94cEoN[Mm0m<9gESOTMd0Q==jEXOlL<V=`0mE<R:DmE=S:DTDeN2h1k;BRDlNMd1P<@RDROH_1V<_OPDZOCY1\\<]OPD@^OT1c<[OnCGYOP1h<YOnCNSOj0o<XOlC5oNe0U=UOkCe2U<[MjCf2V<ZMiCh2V<XMhCj2Y<UMfCm2Y<SMfCn2Z<RMeCP3[<oLcCS3]<mLbCT3^<mL`CU3_<lL^CV3c<kLZCW3e<jLYCW3g<kLVCW3j<iLSCZ3l<hLQC[3m<fLQC\\3n<fLoB]3o<dLnB_3Q=j1O0O1O2N1OO1O100O2N1O100O1O1O1O101N1O1O1O\\G]Im3b6RLbIl3]6RLgIm3X6RLkIm3T6QLPJo3n5PLVJn3i5PL[Jo3d5PL_Jo3`5oKdJP4[5oKhJP4W5nKnJP4Q5mKTKR4l4^KeKb4Y4oJXLP5g3`JlL^5S3SJ^Ml5a2dIQN[6n1VIcNi6\\1gHWOW7h0ZHIe7Z600O2N1O1O001O0001O01O00001O0000010O0000001O000010O000001O00N2K5J7I6J6K5J6J5K5L4K5K5K6K4K5K5K5L4K6J5ZKXE1l:OWEIn:6VEBo:>TEZOQ;f0REQOT;o0PEVNh;i1[D]M\\<b2hCeLn<[3UCkKb=U4T10001O0O100000001N100000001O2M4M2N2_NP@oNR`0o0V@iNl?V1[@aNi?\\1_@\\Nc?b1f@UN\\?i1l@oMV?o1_1N3M2O1N2N2N3M2N2N2O1N3M2N2N2N2N3N1N2N2N2N3M2N2O1N2N3M2N2N2N2O2Mncj0"}, "label": "the skater in a white shirt"}]}
{"id": 73694, "image": "COCO_train2014_000000073694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skate board with teal wheels that a young man is holding in his left hand and getting ready to ride\"."}], "task": "refering", "answer_template": "The \"a skate board with teal wheels that a young man is holding in his left hand and getting ready to ride\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 261, 262, 272, 273, 274], "bbox": [0.12039443155452437, 0.665546875, 0.5651972157772622, 0.827703125], "iscrowd": 0, "area": 8797.50905, "rle": {"size": [640, 431], "counts": "SoP11oc02M2O1N3N0O2O1O0O2O1a\\OCXc0>g\\ODXc0c0N3N1N2O2M2O1N10O1O010O0011N101N101N101N10001N101N101N101N101N101N101O0O2O:E100O101N100O10001N1000000O2O00D;01O1O001O1O00100O001000O0100O01000O01000O01000O010O10O10O10O10O10O10O10O010000O01000O01000O01000O01000O01000O010000O01000O010O10O0100O01000O010O10O0100O010O10O3N2N2M3N2N2M3N2N1N2O1O001N101N1O1O0000001O0000000010O000001O1O001O1O001O1O1O3M3M3M3M3M2N3MQRe3"}, "label": "a skate board with teal wheels that a young man is holding in his left hand and getting ready to ride"}]}
{"id": 73694, "image": "COCO_train2014_000000073694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard with green wheels and stickers on the bottom of it\"."}], "task": "refering", "answer_template": "The \"a skateboard with green wheels and stickers on the bottom of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 261, 262, 272, 273, 274], "bbox": [0.12039443155452437, 0.665546875, 0.5651972157772622, 0.827703125], "iscrowd": 0, "area": 8797.50905, "rle": {"size": [640, 431], "counts": "SoP11oc02M2O1N3N0O2O1O0O2O1a\\OCXc0>g\\ODXc0c0N3N1N2O2M2O1N10O1O010O0011N101N101N101N10001N101N101N101N101N101N101O0O2O:E100O101N100O10001N1000000O2O00D;01O1O001O1O00100O001000O0100O01000O01000O01000O010O10O10O10O10O10O10O10O010000O01000O01000O01000O01000O01000O010000O01000O010O10O0100O01000O010O10O0100O010O10O3N2N2M3N2N2M3N2N1N2O1O001N101N1O1O0000001O0000000010O000001O1O001O1O001O1O1O3M3M3M3M3M2N3MQRe3"}, "label": "a skateboard with green wheels and stickers on the bottom of it"}]}
{"id": 499682, "image": "COCO_train2014_000000499682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man in a black suit\"."}], "task": "refering", "answer_template": "The \"a smiling man in a black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 355, 356, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 373, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 390], "bbox": [0.006270833333333333, 0.11529687500000001, 1.0, 0.9858749999999998], "iscrowd": 0, "area": 162788.81845000002, "rle": {"size": [640, 480], "counts": "j\\27[c0=B>C=C>A>J6000001O0000010O0000001O000001O01O0000001O0001O0001O0000001O0000001O000000001O0000001O0000000O2O0O1O1O2O0O1O1O2O0O1O1M3L5L3L4L4L5L3L4L4L4M4K4L4L4YEVKW6m4`I\\K\\6g4[IcKa6`4WIiKe6Z4RIPLj6T4mHULo6n3hH\\LU7f3bHeLY7a3^HfL^7a3XHfLd7a3RHfLj7j7M3L4L4M3O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O10000O2O0O10000O10000O100PNbD]K_;_4RETKn:g4fEjJZ:Q5oEjJR:Q5UFmJk9o4\\FnJd9m4cFPK^9l4iFQKW9m4o2N2N2M3N2N2N2N2N2N2N2N2N2N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O101N1O1O1O1000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O00100O1O1O2N3M2N2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M5K6J7I6J6jCVIj:P7[DfId;[701O001O001O001O001O001O001O001O001O001O001O001O0010O01O00001O001O001O001O001O001O001O001O001O001O001hK`InJ`6T5eIeJ[6\\5jI^JW6b5oIWJQ6k5TJnIm5R6YJgIg5Z6`J^Ia5c6fJTIZ5m6mJkHT5U7k3101N100O2N100O2O0O100O2O0O101N101N101N1O2O0O2O1O04M2M3N2M3N3L3N2N2M3N3L3N3L3N3M3L3N3L4M4K4M3M4K4M3L2O0O2N100O100O1O100O1O101N100O1O100O100O1O10O01O100O100O1O010O1J6TNl1TNoB"}, "label": "a smiling man in a black suit"}]}
{"id": 499682, "image": "COCO_train2014_000000499682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a well dressed gentleman wearing a sport jacket and tie at a get together\"."}], "task": "refering", "answer_template": "The \"a well dressed gentleman wearing a sport jacket and tie at a get together\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 355, 356, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 373, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 390], "bbox": [0.006270833333333333, 0.11529687500000001, 1.0, 0.9858749999999998], "iscrowd": 0, "area": 162788.81845000002, "rle": {"size": [640, 480], "counts": "j\\27[c0=B>C=C>A>J6000001O0000010O0000001O000001O01O0000001O0001O0001O0000001O0000001O000000001O0000001O0000000O2O0O1O1O2O0O1O1O2O0O1O1M3L5L3L4L4L5L3L4L4L4M4K4L4L4YEVKW6m4`I\\K\\6g4[IcKa6`4WIiKe6Z4RIPLj6T4mHULo6n3hH\\LU7f3bHeLY7a3^HfL^7a3XHfLd7a3RHfLj7j7M3L4L4M3O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O10000O2O0O10000O10000O100PNbD]K_;_4RETKn:g4fEjJZ:Q5oEjJR:Q5UFmJk9o4\\FnJd9m4cFPK^9l4iFQKW9m4o2N2N2M3N2N2N2N2N2N2N2N2N2N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O101N1O1O1O1000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O00100O1O1O2N3M2N2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M5K6J7I6J6jCVIj:P7[DfId;[701O001O001O001O001O001O001O001O001O001O001O001O0010O01O00001O001O001O001O001O001O001O001O001O001O001hK`InJ`6T5eIeJ[6\\5jI^JW6b5oIWJQ6k5TJnIm5R6YJgIg5Z6`J^Ia5c6fJTIZ5m6mJkHT5U7k3101N100O2N100O2O0O100O2O0O101N101N101N1O2O0O2O1O04M2M3N2M3N3L3N2N2M3N3L3N3L3N3M3L3N3L4M4K4M3M4K4M3L2O0O2N100O100O1O100O1O101N100O1O100O100O1O10O01O100O100O1O010O1J6TNl1TNoB"}, "label": "a well dressed gentleman wearing a sport jacket and tie at a get together"}]}
{"id": 147429, "image": "COCO_train2014_000000147429.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black pick up truck with chrome wheels\"."}], "task": "refering", "answer_template": "The \"a black pick up truck with chrome wheels\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 47, 60, 61, 62, 75, 76, 77, 90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 123], "bbox": [0.0, 0.144, 0.21718969555035128, 0.389953125], "iscrowd": 0, "area": 11575.780549999998, "rle": {"size": [640, 427], "counts": "o2l3T`0i0WO0000000000001O00000000000O101O0000000000001O000000000000001O000000000O10mKn@X3R?fLPA[3o>cLSA]3m>aLVA^3j>_LYAa3g>_LYAa3g>^L[Aa3e>_L[Aa3e>^L]Aa3c>_L]Ab3b>^L^Aa3c>^L_Aa3a>_L_Aa3a>^LaAa3_>_LaAa3_>^LbAa3_>_LbA`3^>`LbA`3^>_LdA`3\\>`LdA`3[>`LfA_3[>aLfA^3Z>aLgA_3Y>aLhA^3X>bLhA]3Y>bLhA^3X>bLiA]3\\>]LeAc3]?O0O101O00=C6I2O1O2N1O1N2O1O1O1N2O1O2N1O1M3LU1kN7I8H7I8HWa`6"}, "label": "a black pick up truck with chrome wheels"}]}
{"id": 147429, "image": "COCO_train2014_000000147429.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car directly behind man in white shirt\"."}], "task": "refering", "answer_template": "The \"black car directly behind man in white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 46, 47, 60, 61, 62, 75, 76, 77, 90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 123], "bbox": [0.0, 0.144, 0.21718969555035128, 0.389953125], "iscrowd": 0, "area": 11575.780549999998, "rle": {"size": [640, 427], "counts": "o2l3T`0i0WO0000000000001O00000000000O101O0000000000001O000000000000001O000000000O10mKn@X3R?fLPA[3o>cLSA]3m>aLVA^3j>_LYAa3g>_LYAa3g>^L[Aa3e>_L[Aa3e>^L]Aa3c>_L]Ab3b>^L^Aa3c>^L_Aa3a>_L_Aa3a>^LaAa3_>_LaAa3_>^LbAa3_>_LbA`3^>`LbA`3^>_LdA`3\\>`LdA`3[>`LfA_3[>aLfA^3Z>aLgA_3Y>aLhA^3X>bLhA]3Y>bLhA^3X>bLiA]3\\>]LeAc3]?O0O101O00=C6I2O1O2N1O1N2O1O1O1N2O1O2N1O1M3LU1kN7I8H7I8HWa`6"}, "label": "black car directly behind man in white shirt"}]}
{"id": 507881, "image": "COCO_train2014_000000507881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the silver bowl in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the silver bowl in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252], "bbox": [0.65728125, 0.1660041841004184, 0.999640625, 0.6462133891213389], "iscrowd": 0, "area": 37484.51069999999, "rle": {"size": [478, 640], "counts": "ocT66h>7H8I7I7I7I7H5L2N2N3L3N2N2N3L3N2N2N2M4M2N3M2N3L3N3M2N3L3N3M2N3L3N3M2N3L4M001O0O1000000O1000000O1000000O2O0000000O1000000O1000000O101O000O100000000O10000000000eFbKV8_4iGbKV8^4iGcKW8]4iGdKU8]4jGeKU8[4kGeKU8[4jGgKU8Y4kGhKT8X4kGiKU8W4kGjKT8V4kGlKT8T4lGlKS8U4lGmKS8S4mGnKR8S4lGnKT8R4lGoK1_OU7b4iHPL2^OU7b4iHQL0^OW7a4iHRLN^OY7`4hHSLO]OX7a4iHSLM]OZ7`4hHULL\\O\\7_4hHULK]O]7^4gHWLK[O^7^4gHXLI[O`7]4fHYLI[Oa7\\4fHZLHZOb7\\4eH\\LGYOc7]4eHZLGZOd7\\4dH\\LGYOd7[4eH\\LFZOe7Z4dH^LEYOg7Y4dH_LDXOh7Y4cH`LDXOi7X4cHaLBXOk7W4bHcLAWOm7V4bHcLAWOl7W4bHTM^7l2bHTM^7l2bHTM^7l2aHUM_7k2aHUM_7l2_HUMa7k2_HUMa7k2^HWM`7j2`HVM`7j2_HWMa7i2_HWMa7i2^HXMb7h2^HXMb7h2]HYMc7g2]HYMc7g2\\HZMc7g2]HYMc7g2\\HZMd7f2\\HZMd7g2ZHZMf7f2ZHZMf7f2YH[Mg7e2YH\\Mf7d2YH]Mf7d2ZH\\Mf7d2ZH\\Mf7d2YH]Mg7c2YH]Mg7m41O000000000000000O100000000000000000000000000000000AXHhIh7X6XHhIh7X6XHhIh7X6YHgIg7X6ZHhIf7X6[HgIe7Y6[HgIe7Y6[HgIe7X6]HgIc7Y6]HgIc7Y6]HgIc7X6_HgIa7Y6_HgIa7Y6_HgIb7X6_HgIa7X6`HhI`7X6aHgI_7Y6aHfI`7Z6`HfI`7Y6bHfI^7Z6bHfI^7Z6bHfI^7Y6dHfI\\7Z6dHfI\\7Z6eHeI[7[6eHeI[7Z6fHfIZ7Z6gHeIY7[6gHeIY7Z6hHfIX7Z6iHeIW7[6iHeIW7[6iHeIW7Z6g0000000O10000O1000000O1000000O`M"}, "label": "the silver bowl in the right hand picture"}]}
{"id": 507881, "image": "COCO_train2014_000000507881.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chrome bowl\"."}], "task": "refering", "answer_template": "The \"a chrome bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252], "bbox": [0.65728125, 0.1660041841004184, 0.999640625, 0.6462133891213389], "iscrowd": 0, "area": 37484.51069999999, "rle": {"size": [478, 640], "counts": "ocT66h>7H8I7I7I7I7H5L2N2N3L3N2N2N3L3N2N2N2M4M2N3M2N3L3N3M2N3L3N3M2N3L3N3M2N3L4M001O0O1000000O1000000O1000000O2O0000000O1000000O1000000O101O000O100000000O10000000000eFbKV8_4iGbKV8^4iGcKW8]4iGdKU8]4jGeKU8[4kGeKU8[4jGgKU8Y4kGhKT8X4kGiKU8W4kGjKT8V4kGlKT8T4lGlKS8U4lGmKS8S4mGnKR8S4lGnKT8R4lGoK1_OU7b4iHPL2^OU7b4iHQL0^OW7a4iHRLN^OY7`4hHSLO]OX7a4iHSLM]OZ7`4hHULL\\O\\7_4hHULK]O]7^4gHWLK[O^7^4gHXLI[O`7]4fHYLI[Oa7\\4fHZLHZOb7\\4eH\\LGYOc7]4eHZLGZOd7\\4dH\\LGYOd7[4eH\\LFZOe7Z4dH^LEYOg7Y4dH_LDXOh7Y4cH`LDXOi7X4cHaLBXOk7W4bHcLAWOm7V4bHcLAWOl7W4bHTM^7l2bHTM^7l2bHTM^7l2aHUM_7k2aHUM_7l2_HUMa7k2_HUMa7k2^HWM`7j2`HVM`7j2_HWMa7i2_HWMa7i2^HXMb7h2^HXMb7h2]HYMc7g2]HYMc7g2\\HZMc7g2]HYMc7g2\\HZMd7f2\\HZMd7g2ZHZMf7f2ZHZMf7f2YH[Mg7e2YH\\Mf7d2YH]Mf7d2ZH\\Mf7d2ZH\\Mf7d2YH]Mg7c2YH]Mg7m41O000000000000000O100000000000000000000000000000000AXHhIh7X6XHhIh7X6XHhIh7X6YHgIg7X6ZHhIf7X6[HgIe7Y6[HgIe7Y6[HgIe7X6]HgIc7Y6]HgIc7Y6]HgIc7X6_HgIa7Y6_HgIa7Y6_HgIb7X6_HgIa7X6`HhI`7X6aHgI_7Y6aHfI`7Z6`HfI`7Y6bHfI^7Z6bHfI^7Z6bHfI^7Y6dHfI\\7Z6dHfI\\7Z6eHeI[7[6eHeI[7Z6fHfIZ7Z6gHeIY7[6gHeIY7Z6hHfIX7Z6iHeIW7[6iHeIW7[6iHeIW7Z6g0000000O10000O1000000O1000000O`M"}, "label": "a chrome bowl"}]}
{"id": 212974, "image": "COCO_train2014_000000212974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table under the spotted cake\"."}], "task": "refering", "answer_template": "The \"table under the spotted cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 206, 207, 208, 226, 227, 246, 265, 283, 284, 301, 302, 303, 319, 320, 321, 322, 337, 338, 339, 340, 341, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360], "bbox": [0.5012962962962962, 0.4953148148148149, 1.0, 0.9810185185185185], "iscrowd": 0, "area": 16587.88465000001, "rle": {"size": [540, 540], "counts": "c]_43f`0300001O0000000000000000001O00000000000000001O00000000000000000000O10000000000000000000000O10000000000000000000000O1000000000000000000O10000000000000000O10000000000000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O100O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O1O100O1O1O1O1O1O1O1O1O1O1O1ZJRN]Lo1\\3YNbLh1^3YN`Lh1a3XN]Li1c3XN\\Lh1d3XN[Li1e3XNYLi1g3XNWLi1i3XNULi1k3XNSLi1m3WNRLj1n3WNPLj1P4WNnKj1R4WNlKj1T4WNjKj1V4WNhKj1X4VNgKk1Y4VNeKk1\\4UNbKl1^4UNaKk1_4VN_Kk1a4UN^Kl1b4UN\\Kl1d4VNYKk1g4VNWKk1i4VNUKk1k4WNRKj1n4WNPKj1P5WNnJj1R5XNkJi1U5XNiJi1W5XNgJi1Y5YNdJh1\\5YNbJh1_5XN[Jm1e5WNRJn1n5VNiIo1W6VN`In1`6VNWIo1i6VNmHo1S7ROgGS1Y8TO[GQ1e8c3000000000000001O0000000000000000000000000000001O`G"}, "label": "table under the spotted cake"}]}
{"id": 212974, "image": "COCO_train2014_000000212974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"surface below iced cake\"."}], "task": "refering", "answer_template": "The \"surface below iced cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 206, 207, 208, 226, 227, 246, 265, 283, 284, 301, 302, 303, 319, 320, 321, 322, 337, 338, 339, 340, 341, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360], "bbox": [0.5012962962962962, 0.4953148148148149, 1.0, 0.9810185185185185], "iscrowd": 0, "area": 16587.88465000001, "rle": {"size": [540, 540], "counts": "c]_43f`0300001O0000000000000000001O00000000000000001O00000000000000000000O10000000000000000000000O10000000000000000000000O1000000000000000000O10000000000000000O10000000000000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O100O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O1O100O1O1O1O1O1O1O1O1O1O1O1ZJRN]Lo1\\3YNbLh1^3YN`Lh1a3XN]Li1c3XN\\Lh1d3XN[Li1e3XNYLi1g3XNWLi1i3XNULi1k3XNSLi1m3WNRLj1n3WNPLj1P4WNnKj1R4WNlKj1T4WNjKj1V4WNhKj1X4VNgKk1Y4VNeKk1\\4UNbKl1^4UNaKk1_4VN_Kk1a4UN^Kl1b4UN\\Kl1d4VNYKk1g4VNWKk1i4VNUKk1k4WNRKj1n4WNPKj1P5WNnJj1R5XNkJi1U5XNiJi1W5XNgJi1Y5YNdJh1\\5YNbJh1_5XN[Jm1e5WNRJn1n5VNiIo1W6VN`In1`6VNWIo1i6VNmHo1S7ROgGS1Y8TO[GQ1e8c3000000000000001O0000000000000000000000000000001O`G"}, "label": "surface below iced cake"}]}
{"id": 212974, "image": "COCO_train2014_000000212974.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table\"."}], "task": "refering", "answer_template": "The \"table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 180, 197, 198, 199, 216, 217, 218, 234, 235, 236, 237, 251, 252, 253, 254, 255, 256, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 342, 343, 346, 347, 348, 349, 350, 351], "bbox": [0.0012222222222222222, 0.5027037037037037, 0.5042222222222222, 0.9884259259259259], "iscrowd": 0, "area": 40956.505950000006, "rle": {"size": [540, 540], "counts": "Zo0Q1k?S1lN5L0O1O100O010O100O1O100O100O100O1O100O100O1O10O0100O100O1O100O100O1O100O100O100O1O010O100O100O1O100O01M2N2O2M2N3N1N30O0010O010O010O01O01O010O01O010O010O010O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O2O2M3N2M4M2M3M3N3L3N2M4L3N2M3N0O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O100O1\\NPJ[HQ6\\7\\J`Hd5X7hJcHY5]7kJ_HU5a7oJZHR5f7RKUHo4k7UKQHk4o7YKlGh4T8\\KhGd4X8`KcGa4]8cK_G]4a8o10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000:Fm2SMm2SMmk\\4"}, "label": "table"}]}
{"id": 401392, "image": "COCO_train2014_000000401392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the farther away red chair\"."}], "task": "refering", "answer_template": "The \"the farther away red chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 19, 20, 21, 22, 42, 43, 44, 45, 68, 91], "bbox": [0.699109375, 0.0026010101010101013, 1.0, 0.2883080808080808], "iscrowd": 0, "area": 5176.0075000000015, "rle": {"size": [396, 640], "counts": "gk\\5f0f;00000000000000001O0O10000000000000001O0000000000000000001O00000O1000000000001O0000000f0ZO2N3Maha06YW^O=C7I0000000000000000001O00000000000000000O100000001O000000000000000000000000001O00000000000000000000000O101O000000000000000000000000001O000000000000001O1O1O1O1O1O1O1N3N2N3M?A?A?A`0@1O0000000000O1000N"}, "label": "the farther away red chair"}]}
{"id": 450551, "image": "COCO_train2014_000000450551.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe that is in front of the other giraffes\"."}], "task": "refering", "answer_template": "The \"the giraffe that is in front of the other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 177, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 309, 313, 314, 315, 316, 324, 325, 326, 327, 328, 329, 330, 331, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.006734374999999999, 0.3251458333333333, 0.766859375, 0.9858333333333333], "iscrowd": 0, "area": 57786.68265, "rle": {"size": [480, 640], "counts": "\\i21n>2M2O1O1N2O1O1O100O1O100O1O100O1O100O2N100O1O1O100O1O100O1O100O1O100O1O2O0O1O100O1O100O1O1O100O1O100O1O100O2N100O1O100O1O100O1O100O1O1O100O1O2O0O1O100O1O100O1O100O1O100O1O100O1O2N100O1O100O1O100O1O1O100O1O1O100O2N1O1O100O1O1O1O100O1O1O100O1O1O1O2O0O1O1O100O1O1O1O100O1O1O100O1O1O1O2O0O1O1O1O100O1O1O100O1O1O1O100O1O2N100O1O1O1O100O1O1O1O100O1O1O100O1O2N1O100O1O100O100O1O100O1O100O1O101N1O100O1O100O1O100O1O100O0001O0O101O00001O000O2O00001O000O2O00001O00001N10001O00001O0O101O001O00001N10001O00001O0O101O00001O0O101O00001O000O2O0000001O000O1000000000000O1000001O000O1000000000000O10001O00O100O010O100O100O10O01O100O100O10O0100O1001001O0000001O00001O0001O01O00001O0000001O0000001O000L5mNR1O1O1O2N1O1N3N1O1O1O2N1O1N3N1O1O12O2M2N3N1N2N3M2O2M2N3M2O2M2N5L5J5K6J10L3N2M4L3M3L4A`0_O`0O110O01O1O010O00100O001O010O1O010O001O3N5XKSGQ3S9_LWGa3o8nK\\GS4h90O2O00001N101O0O101O001N101O001N101O0O2O001O1N101OlLSFU1m9jNWFT1g9mN\\FP1d9oNaFn0^9QOfFm0Y9ROkFk0T9VOmFj0R9UOQGi0o8VOTGi0j8WOXGh0h8XOZGg0e8XO^Gf0b8YO`Gg0^8YOeGf0Z8ZOhGd0X8[OjGe0U8ZOnGd0Q8\\OQHd0n7\\OTHb0l7]OWHb0g7^O[Ha0e7^O^Ha0a7_OaH`0^7_OdHa0[7^OhHa0X7]OkHa0V7^OkHa0U7^OnH`0S7^OoHa0R7]OQIa0P7]OSIa0n6ZOWIe0i6WO]Ig0d6TObIj0_6ROeIm0\\6nNjIP1W6kNoIR1R6jNSJU1n5fNXJX1i9O2N2N1O2N2N1O2N2N7I9G8Hc^U2"}, "label": "the giraffe that is in front of the other giraffes"}]}
{"id": 450551, "image": "COCO_train2014_000000450551.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest giraffe in the image\"."}], "task": "refering", "answer_template": "The \"tallest giraffe in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 149, 168, 287, 309, 310, 311, 331, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381, 382], "bbox": [0.266296875, 0.22472916666666667, 0.6235937500000001, 0.9842708333333333], "iscrowd": 0, "area": 17017.458099999996, "rle": {"size": [480, 640], "counts": "^Y`22n>2N2N2N1O2N2N2N2N1O2M3N2O1N1O2N2N2O1N1O2N2N101N1O1O2N101N101O00001O001O0O101O1O4L4L5K4LCoNcNXEU1k:QOSEh0T;ZOkD=];EbD7`;M\\D4d;NWD5i;MRD6n;LmC7S<LgC8X<X100O1000000001O2O1N2N2N2N2N2N2N2N2N2O0O2N2N1O001O0N2O2M2O1O2M2OM2K6K4L5M2O21N2N2O2M2O1N200000000010O0UJ^NXNb1g1hNPNX1o1QOiMo0U2\\OcMc0\\2G_M5`25\\MFc2d0XMXOg2Q1VMjNi2X1YMeNf2\\1\\MbNb2`1`M^N_2c1dMZN[2g1gMWNX2j1iMUNV2l1jMTNU2m1kMSNT2n1lMRNR2P2mMQNR2P2nMPNQ2Q2oMoMP2R2PNmMP2T2oMmMP2T2PNlMo1U2QNkMm1W2SNkMj1V2UNmMh1T2XNoMd1R2\\NPNa1Q2^NRN_1o1aNRN]1o1cNSNY1o1gNSNV1n1iNTNU1m1kNUNR1l1nNVNo0k1QOVNm0k1ROXNk0i1UOXNh0j1XOVNg0k1XOUNh0k1YOTNg0m1XORNi0o1VOQNj0P2VOoMj0R2UOnMj0T2VOjMk0W2TOiMl0X2TOgMl0Z2SOfMm0[2SOcMn0^2QObMP1^2PO`MT1^2lN`MX1^2hN`M[1_2fN^M^1`2cN]Ma1a2_N]Md1b2]N[Mg1c2YN[Mk1c2VNZMm1e2TNXMP2f2PNXMT2f2mMVMW2i2iMUM[2i2fMTM^2j2cMSM`2l2`MRMd2l2]MQMg2m2g4O2N1O2N2N1O2N2N1N3N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1N3N2N1O2N2N1O2N2N1O2N\\``3"}, "label": "tallest giraffe in the image"}]}
{"id": 475129, "image": "COCO_train2014_000000475129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white horse up close with black markings on it\"."}], "task": "refering", "answer_template": "The \"white horse up close with black markings on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409], "bbox": [0.22784374999999998, 0.06677734375, 0.8633906249999999, 0.99826171875], "iscrowd": 0, "area": 130976.39610000001, "rle": {"size": [512, 640], "counts": "_^Y27a?:E:G:E;E:G:E:G:E:G:I6M4K4M4UOTL\\EP4_:XLYEl3c:[LWEi3c:m0L5K4L5J5L5K5J6K3M3M3L4M3M3O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2L4L4L4L4L4L4L4K5L4L4M3L4L4L4L4L4M3O1O1O1O1O100O1O1O1O1O1O1O100000000O10000000000O1002N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N00N2N2N2N2N2N200000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000N2N2N2N2N2N2N2N2N2N21O001O1O001O001O1O001O001O1O001O1UJdNPM^1m2gNoL[1n2iNRIGW2b1c4mNPIKU2Z1h4POlH1T2Q1m4SOhH6T2i0Q5VOdH<S2`0U5ZOaHa0R27Z5]O\\Hg0S2N^5@XHm0R2DVOSNo4U4cKjMWOXNR5W4`KcMWO\\NW5Z4ZK\\MXOaN[5_4RKRM\\OfN_5k4cJaLGjNc5R9gLiEd0Y:ZORF;Q:B[F2h9LbFHa96kF^OX9?iG`NZ8^1U3N3M3M2O2M2N3M3L3M4L4L3N3L4L3L5I6J7IT[[1"}, "label": "white horse up close with black markings on it"}]}
{"id": 475129, "image": "COCO_train2014_000000475129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse with black spots , has long hair , looks to be in a pasture\"."}], "task": "refering", "answer_template": "The \"a white horse with black spots , has long hair , looks to be in a pasture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409], "bbox": [0.22784374999999998, 0.06677734375, 0.8633906249999999, 0.99826171875], "iscrowd": 0, "area": 130976.39610000001, "rle": {"size": [512, 640], "counts": "_^Y27a?:E:G:E;E:G:E:G:E:G:I6M4K4M4UOTL\\EP4_:XLYEl3c:[LWEi3c:m0L5K4L5J5L5K5J6K3M3M3L4M3M3O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2L4L4L4L4L4L4L4K5L4L4M3L4L4L4L4L4M3O1O1O1O1O100O1O1O1O1O1O1O100000000O10000000000O1002N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N00N2N2N2N2N2N200000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000N2N2N2N2N2N2N2N2N2N21O001O1O001O001O1O001O001O1O001O1UJdNPM^1m2gNoL[1n2iNRIGW2b1c4mNPIKU2Z1h4POlH1T2Q1m4SOhH6T2i0Q5VOdH<S2`0U5ZOaHa0R27Z5]O\\Hg0S2N^5@XHm0R2DVOSNo4U4cKjMWOXNR5W4`KcMWO\\NW5Z4ZK\\MXOaN[5_4RKRM\\OfN_5k4cJaLGjNc5R9gLiEd0Y:ZORF;Q:B[F2h9LbFHa96kF^OX9?iG`NZ8^1U3N3M3M2O2M2N3M3L3M4L4L3N3L4L3L5I6J7IT[[1"}, "label": "a white horse with black spots , has long hair , looks to be in a pasture"}]}
{"id": 475129, "image": "COCO_train2014_000000475129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black horse with a white stripe on it ' s nose\"."}], "task": "refering", "answer_template": "The \"a black horse with a white stripe on it ' s nose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 277, 278, 279, 280, 300, 301, 302, 303, 323, 324, 325, 326, 347, 348], "bbox": [0.020640625000000003, 0.29158203125, 0.22623437500000004, 0.84666015625], "iscrowd": 0, "area": 25201.40335, "rle": {"size": [512, 640], "counts": "ng6m0a>d0E;J6L3L5K5L3M4N2M2O=B5L4K5L4jEbLa7c3VHcLh7e3lGaLS8g3_G_L`8l3QGZLn8R4`FUL^9X5O2N1N3N0O10001N10001O00000O2O0000001O00001O1O1O2N2N00001O0001O01O00001O000000000000000000000N2OiNdGjJ\\8o4dHVJ^7]4aG`LQ:^3^FnKQ1Gm6Z4RJmKk5T4PJSLn5n3lIYLQ6h3hI`LW6`3cIgL\\6Y3^InL`6S3YIUMf6k2UI[Mi6f2TI^Mi6f2RI^Me6k2XIXM]6S3_IQM]6U3]IoLa6U3YIoLg6S3SIQMm6Q3mHSMT7l52N2N1O000000000000OmNgH`IZ7X6WI^Ij6Z6fI]I[6]6mIaIS6Z6TJcIm5Y6YJdIh5W6S2K5L4L4O1O1O1O1O1O1O1M3M3M1N3N2N3L6_KVE01o2l:mLXE07g2l:mLSE8<`2V;]McEj1a:RNdEh1`<I6I8H8ZOVXg7"}, "label": "a black horse with a white stripe on it ' s nose"}]}
{"id": 475129, "image": "COCO_train2014_000000475129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white horse behind a mainly white horse\"."}], "task": "refering", "answer_template": "The \"a black and white horse behind a mainly white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 277, 278, 279, 280, 300, 301, 302, 303, 323, 324, 325, 326, 347, 348], "bbox": [0.020640625000000003, 0.29158203125, 0.22623437500000004, 0.84666015625], "iscrowd": 0, "area": 25201.40335, "rle": {"size": [512, 640], "counts": "ng6m0a>d0E;J6L3L5K5L3M4N2M2O=B5L4K5L4jEbLa7c3VHcLh7e3lGaLS8g3_G_L`8l3QGZLn8R4`FUL^9X5O2N1N3N0O10001N10001O00000O2O0000001O00001O1O1O2N2N00001O0001O01O00001O000000000000000000000N2OiNdGjJ\\8o4dHVJ^7]4aG`LQ:^3^FnKQ1Gm6Z4RJmKk5T4PJSLn5n3lIYLQ6h3hI`LW6`3cIgL\\6Y3^InL`6S3YIUMf6k2UI[Mi6f2TI^Mi6f2RI^Me6k2XIXM]6S3_IQM]6U3]IoLa6U3YIoLg6S3SIQMm6Q3mHSMT7l52N2N1O000000000000OmNgH`IZ7X6WI^Ij6Z6fI]I[6]6mIaIS6Z6TJcIm5Y6YJdIh5W6S2K5L4L4O1O1O1O1O1O1O1M3M3M1N3N2N3L6_KVE01o2l:mLXE07g2l:mLSE8<`2V;]McEj1a:RNdEh1`<I6I8H8ZOVXg7"}, "label": "a black and white horse behind a mainly white horse"}]}
{"id": 180220, "image": "COCO_train2014_000000180220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the portion of the train to the right of the sign\"."}], "task": "refering", "answer_template": "The \"the portion of the train to the right of the sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 125, 144, 148, 149, 150, 151, 152, 153, 167, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 194, 195, 196, 197, 198, 199, 200, 201, 202, 223, 224, 225], "bbox": [0.255609375, 0.2959583333333333, 0.81053125, 0.5426041666666667], "iscrowd": 0, "area": 15443.87405, "rle": {"size": [480, 640], "counts": "lP]22m>4M2M4M2M4M3L8Hc0^O=B>Ba0_O100O1O1O2O00000O10SOln0lMUTO00000O10000000000001O0000000000000000iZX19mdgN=C=D=C<D=D8G001O01O000000001O0000001O0000001O000000001O0000001O0000001O000000001O0001O01O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000000001O0001O01O0000001O000000001O0000001O00000000001O01O000001O00000000001O000000010O00000000001O000000001O000001O0001O000000001O000000001O0001O000001O000000001O00000000010O000000001O000000001O000000010O000000001O0000001O0000000010O00000001O000000001O000000010O000000001O00001O00001O0O2G8G9H9F9Hb`h1"}, "label": "the portion of the train to the right of the sign"}]}
{"id": 180220, "image": "COCO_train2014_000000180220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the last traincar , closest to a stop sign\"."}], "task": "refering", "answer_template": "The \"the last traincar , closest to a stop sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190], "bbox": [0.0033750000000000004, 0.19775, 0.26965625, 0.5101041666666667], "iscrowd": 0, "area": 21111.944849999996, "rle": {"size": [480, 640], "counts": "TQ1]1^=Y3lL01O000000001O0O10001O0000001O000000001O0000001O0000001O000000001O0O10001O0000001O0000001O000000001O0000001O0000001O000O10001O0000001O0000001O000000001O0000001O0000001O000O10001O0000001O0000001O0000001O000000001O0000001O000O101O000000001O0000001O0000001O000000001O0000001O000O101O0000001O000000001O000000?A1O1O2N1O2N1O1O2N1O2N1N2O2M4@`0@R1oN^gj6"}, "label": "the last traincar , closest to a stop sign"}]}
{"id": 180220, "image": "COCO_train2014_000000180220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the 3431 yellow train\"."}], "task": "refering", "answer_template": "The \"the 3431 yellow train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190], "bbox": [0.0033750000000000004, 0.19775, 0.26965625, 0.5101041666666667], "iscrowd": 0, "area": 21111.944849999996, "rle": {"size": [480, 640], "counts": "TQ1]1^=Y3lL01O000000001O0O10001O0000001O000000001O0000001O0000001O000000001O0O10001O0000001O0000001O000000001O0000001O0000001O000O10001O0000001O0000001O000000001O0000001O0000001O000O10001O0000001O0000001O0000001O000000001O0000001O000O101O000000001O0000001O0000001O000000001O0000001O000O101O0000001O000000001O000000?A1O1O2N1O2N1O1O2N1O2N1N2O2M4@`0@R1oN^gj6"}, "label": "the 3431 yellow train"}]}
{"id": 368637, "image": "COCO_train2014_000000368637.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock on the left side of the tower\"."}], "task": "refering", "answer_template": "The \"a clock on the left side of the tower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 284], "bbox": [0.264796875, 0.3655625, 0.456546875, 0.712625], "iscrowd": 0, "area": 15865.723150000002, "rle": {"size": [480, 640], "counts": "n^_24h>f0YO9H9H8G7J6J6J6K4K5L4L4L3M4M3L3M3N3L3N3L3N2M4M2N1O2N2N2N2N2O1N2N2N2O1N1O2O0O1O2O0O101N1O2O000O2O0O10000O1000O010000O1000000000000000O1000O10O2O000O10000O2O000O101N101N2O0O2N101N101N101N2N2N2N2N2M3N2N2N2N2N2M3M4M3L3M4L4M2M4L4L4M3L5J6I8H8H8G:E;FclR5"}, "label": "a clock on the left side of the tower"}]}
{"id": 368637, "image": "COCO_train2014_000000368637.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clock face on the left reads the same time as the right\"."}], "task": "refering", "answer_template": "The \"the clock face on the left reads the same time as the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 284], "bbox": [0.264796875, 0.3655625, 0.456546875, 0.712625], "iscrowd": 0, "area": 15865.723150000002, "rle": {"size": [480, 640], "counts": "n^_24h>f0YO9H9H8G7J6J6J6K4K5L4L4L3M4M3L3M3N3L3N3L3N2M4M2N1O2N2N2N2N2O1N2N2N2O1N1O2O0O1O2O0O101N1O2O000O2O0O10000O1000O010000O1000000000000000O1000O10O2O000O10000O2O000O101N101N2O0O2N101N101N101N2N2N2N2N2M3N2N2N2N2N2M3M4M3L3M4L4M2M4L4L4M3L5J6I8H8H8G:E;FclR5"}, "label": "the clock face on the left reads the same time as the right"}]}
{"id": 368637, "image": "COCO_train2014_000000368637.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the face of the clock tower is in shadow\"."}], "task": "refering", "answer_template": "The \"the face of the clock tower is in shadow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 243, 244, 245], "bbox": [0.507125, 0.2990625, 0.6934843749999999, 0.6237916666666667], "iscrowd": 0, "area": 13042.137649999999, "rle": {"size": [480, 640], "counts": "l`h4:^><D<G8K6I7J5K6J5J203M4M2M4M2M4M2M4M1N3N2N1N3N2M2O2M3N2N1O2N2N1O1O1O1O2N100O1O1O1O2N1O1O1O1O2OO01O001O001O0100O10O01O100O10O0100O100O010O100O00100O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O2N2N2N2N1O2M3M3M3M2N3M3M3M4L4L4L4L4K5L4L5I8F`0@`0Aock2"}, "label": "the face of the clock tower is in shadow"}]}
